diff --git a/models/parakeet-ctc-110m-coreml/AudioEncoder.mlmodelc/analytics/coremldata.bin b/AudioEncoder.mlmodelc/analytics/coremldata.bin similarity index 100% rename from models/parakeet-ctc-110m-coreml/AudioEncoder.mlmodelc/analytics/coremldata.bin rename to AudioEncoder.mlmodelc/analytics/coremldata.bin diff --git a/models/parakeet-ctc-110m-coreml/AudioEncoder.mlmodelc/coremldata.bin b/AudioEncoder.mlmodelc/coremldata.bin similarity index 100% rename from models/parakeet-ctc-110m-coreml/AudioEncoder.mlmodelc/coremldata.bin rename to AudioEncoder.mlmodelc/coremldata.bin diff --git a/models/parakeet-ctc-110m-coreml/AudioEncoder.mlmodelc/metadata.json b/AudioEncoder.mlmodelc/metadata.json similarity index 100% rename from models/parakeet-ctc-110m-coreml/AudioEncoder.mlmodelc/metadata.json rename to AudioEncoder.mlmodelc/metadata.json diff --git a/models/parakeet-ctc-110m-coreml/AudioEncoder.mlmodelc/model.mil b/AudioEncoder.mlmodelc/model.mil similarity index 100% rename from models/parakeet-ctc-110m-coreml/AudioEncoder.mlmodelc/model.mil rename to AudioEncoder.mlmodelc/model.mil diff --git a/models/parakeet-ctc-110m-coreml/AudioEncoder.mlmodelc/weights/weight.bin b/AudioEncoder.mlmodelc/weights/weight.bin similarity index 100% rename from models/parakeet-ctc-110m-coreml/AudioEncoder.mlmodelc/weights/weight.bin rename to AudioEncoder.mlmodelc/weights/weight.bin diff --git a/models/parakeet-ctc-110m-coreml/MelSpectrogram.mlmodelc/analytics/coremldata.bin b/MelSpectrogram.mlmodelc/analytics/coremldata.bin similarity index 100% rename from models/parakeet-ctc-110m-coreml/MelSpectrogram.mlmodelc/analytics/coremldata.bin rename to MelSpectrogram.mlmodelc/analytics/coremldata.bin diff --git a/models/parakeet-ctc-110m-coreml/MelSpectrogram.mlmodelc/coremldata.bin b/MelSpectrogram.mlmodelc/coremldata.bin similarity index 100% rename from models/parakeet-ctc-110m-coreml/MelSpectrogram.mlmodelc/coremldata.bin rename to MelSpectrogram.mlmodelc/coremldata.bin diff --git a/models/parakeet-ctc-110m-coreml/MelSpectrogram.mlmodelc/metadata.json b/MelSpectrogram.mlmodelc/metadata.json similarity index 100% rename from models/parakeet-ctc-110m-coreml/MelSpectrogram.mlmodelc/metadata.json rename to MelSpectrogram.mlmodelc/metadata.json diff --git a/models/parakeet-ctc-110m-coreml/MelSpectrogram.mlmodelc/model.mil b/MelSpectrogram.mlmodelc/model.mil similarity index 100% rename from models/parakeet-ctc-110m-coreml/MelSpectrogram.mlmodelc/model.mil rename to MelSpectrogram.mlmodelc/model.mil diff --git a/models/parakeet-ctc-110m-coreml/MelSpectrogram.mlmodelc/weights/weight.bin b/MelSpectrogram.mlmodelc/weights/weight.bin similarity index 100% rename from models/parakeet-ctc-110m-coreml/MelSpectrogram.mlmodelc/weights/weight.bin rename to MelSpectrogram.mlmodelc/weights/weight.bin diff --git a/cli/CtcEarningsBenchmark.swift b/cli/CtcEarningsBenchmark.swift deleted file mode 100644 index 1366881f21e31fd699568575d0e3d519179391fc..0000000000000000000000000000000000000000 --- a/cli/CtcEarningsBenchmark.swift +++ /dev/null @@ -1,1048 +0,0 @@ -#if os(macOS) -import AVFoundation -import CoreML -import FluidAudio -import Foundation - -/// Earnings22 benchmark using TDT for transcription + CTC for keyword spotting. -/// TDT provides low WER transcription, CTC provides high recall dictionary detection. -public enum CtcEarningsBenchmark { - - private enum KeywordMode: String { - case chunk - case file - } - - /// Default CTC model directory - private static func defaultCtcModelPath() -> String? { - let appSupport = FileManager.default.urls( - for: .applicationSupportDirectory, in: .userDomainMask - ).first! - let modelPath = appSupport.appendingPathComponent("FluidAudio/Models/parakeet-ctc-110m-coreml") - if FileManager.default.fileExists(atPath: modelPath.path) { - return modelPath.path - } - return nil - } - - /// Default data directory (from download command) - private static func defaultDataDir() -> String? { - let dataDir = DatasetDownloader.getEarnings22Directory().appendingPathComponent("test-dataset") - if FileManager.default.fileExists(atPath: dataDir.path) { - return dataDir.path - } - return nil - } - - public static func runCLI(arguments: [String]) async { - // Check for help - if arguments.contains("--help") || arguments.contains("-h") { - printUsage() - return - } - - // Parse arguments - var dataDir: String? = nil - var outputFile = "ctc_earnings_benchmark.json" - var maxFiles: Int? = nil - var ctcModelPath: String? = nil - // Note: Using v2 by default because v3 has issues with certain audio files - // (returns empty transcription for ~7 files in Earnings22 dataset) - var tdtVersion: AsrModelVersion = .v2 - var autoDownload = false - var keywordMode: KeywordMode = .chunk - - var i = 0 - while i < arguments.count { - switch arguments[i] { - case "--data-dir": - if i + 1 < arguments.count { - dataDir = arguments[i + 1] - i += 1 - } - case "--output", "-o": - if i + 1 < arguments.count { - outputFile = arguments[i + 1] - i += 1 - } - case "--max-files": - if i + 1 < arguments.count { - maxFiles = Int(arguments[i + 1]) - i += 1 - } - case "--ctc-model": - if i + 1 < arguments.count { - ctcModelPath = arguments[i + 1] - i += 1 - } - case "--tdt-version": - if i + 1 < arguments.count { - if arguments[i + 1] == "v2" || arguments[i + 1] == "2" { - tdtVersion = .v2 - } - i += 1 - } - case "--auto-download": - autoDownload = true - case "--keyword-mode": - if i + 1 < arguments.count, let mode = parseKeywordMode(arguments[i + 1]) { - keywordMode = mode - i += 1 - } - default: - break - } - i += 1 - } - - // Use defaults if not specified - if dataDir == nil { - dataDir = defaultDataDir() - } - if ctcModelPath == nil { - ctcModelPath = defaultCtcModelPath() - } - - // Handle auto-download for dataset - if autoDownload && dataDir == nil { - print("đŸ“„ Downloading earnings22-kws dataset...") - await DatasetDownloader.downloadEarnings22KWS(force: false) - dataDir = defaultDataDir() - } - - // Handle auto-download for CTC models - if autoDownload && ctcModelPath == nil { - print("đŸ“„ Downloading CTC models...") - do { - _ = try await CtcModels.download() - ctcModelPath = defaultCtcModelPath() - } catch { - print("ERROR: Failed to download CTC models: \(error)") - } - } - - print("Earnings Benchmark (TDT transcription + CTC keyword spotting)") - print(" Data directory: \(dataDir ?? "not found")") - print(" Output file: \(outputFile)") - print(" TDT version: \(tdtVersion == .v2 ? "v2" : "v3")") - print(" CTC model: \(ctcModelPath ?? "not found")") - print(" Keyword mode: \(keywordMode.rawValue)") - - guard let finalDataDir = dataDir else { - print("ERROR: Data directory not found") - print("💡 Download with: fluidaudio download --dataset earnings22-kws") - print(" Or specify: --data-dir ") - printUsage() - return - } - - guard let modelPath = ctcModelPath else { - print("ERROR: CTC model not found") - print("💡 Download parakeet-ctc-110m-coreml model to:") - print(" ~/Library/Application Support/FluidAudio/Models/parakeet-ctc-110m-coreml/") - print(" Or specify: --ctc-model ") - printUsage() - return - } - - let dataDirResolved = finalDataDir - - do { - // Load TDT models for transcription - print("Loading TDT models (\(tdtVersion == .v2 ? "v2" : "v3")) for transcription...") - let tdtModels = try await AsrModels.downloadAndLoad(version: tdtVersion) - let asrManager = AsrManager(config: .default) - try await asrManager.initialize(models: tdtModels) - print("TDT models loaded successfully") - - // Load CTC models for keyword spotting - print("Loading CTC models from: \(modelPath)") - let modelDir = URL(fileURLWithPath: modelPath) - let ctcModels = try await CtcModels.loadDirect(from: modelDir) - print("Loaded CTC vocabulary with \(ctcModels.vocabulary.count) tokens") - - // Create keyword spotter - let vocabSize = ctcModels.vocabulary.count - let blankId = vocabSize // Blank is at index = vocab_size - let spotter = CtcKeywordSpotter(models: ctcModels, blankId: blankId) - print("Created CTC spotter with blankId=\(blankId)") - - // Collect test files - let dataDirURL = URL(fileURLWithPath: dataDirResolved) - let fileIds = try collectFileIds(from: dataDirURL, maxFiles: maxFiles) - let keywordIndex = try buildKeywordIndex(dataDir: dataDirURL, keywordMode: keywordMode) - - if fileIds.isEmpty { - print("ERROR: No test files found in \(dataDirResolved)") - return - } - - print("Processing \(fileIds.count) test files...") - - var results: [[String: Any]] = [] - var totalWer = 0.0 - var totalKeywordReference = 0 - var totalKeywordPredicted = 0 - var totalKeywordTruePositives = 0 - var totalKeywordFalsePositives = 0 - var totalKeywordFalseNegatives = 0 - var totalAudioDuration = 0.0 - var totalProcessingTime = 0.0 - - for (index, fileId) in fileIds.enumerated() { - print("[\(index + 1)/\(fileIds.count)] \(fileId)") - - if let result = try await processFile( - fileId: fileId, - dataDir: dataDirURL, - asrManager: asrManager, - ctcModels: ctcModels, - spotter: spotter, - keywordMode: keywordMode, - keywordIndex: keywordIndex - ) { - results.append(result) - totalWer += result["wer"] as? Double ?? 0 - totalKeywordReference += result["keywordReference"] as? Int ?? 0 - totalKeywordPredicted += result["keywordPredicted"] as? Int ?? 0 - totalKeywordTruePositives += result["keywordTruePositives"] as? Int ?? 0 - totalKeywordFalsePositives += result["keywordFalsePositives"] as? Int ?? 0 - totalKeywordFalseNegatives += result["keywordFalseNegatives"] as? Int ?? 0 - totalAudioDuration += result["audioLength"] as? Double ?? 0 - totalProcessingTime += result["processingTime"] as? Double ?? 0 - - let wer = result["wer"] as? Double ?? 0 - let precision = result["keywordPrecision"] as? Double ?? 0 - let recall = result["keywordRecall"] as? Double ?? 0 - let fscore = result["keywordFscore"] as? Double ?? 0 - print( - " WER: \(String(format: "%.1f", wer))%, " + - "KW P/R/F: \(String(format: "%.2f", precision))/" + - "\(String(format: "%.2f", recall))/" + - "\(String(format: "%.2f", fscore))" - ) - } - } - - // Calculate summary - let avgWer = results.isEmpty ? 0.0 : totalWer / Double(results.count) - let keywordPrecision = - totalKeywordPredicted > 0 - ? Double(totalKeywordTruePositives) / Double(totalKeywordPredicted) - : 0 - let keywordRecall = - totalKeywordReference > 0 - ? Double(totalKeywordTruePositives) / Double(totalKeywordReference) - : 0 - let keywordFscore = - (keywordPrecision + keywordRecall) > 0 - ? 2 * keywordPrecision * keywordRecall / (keywordPrecision + keywordRecall) - : 0 - - // Print summary - print("\n" + String(repeating: "=", count: 60)) - print("EARNINGS22 BENCHMARK (TDT + CTC)") - print(String(repeating: "=", count: 60)) - print("Model: \(modelPath)") - print("Total tests: \(results.count)") - print("Average WER: \(String(format: "%.2f", avgWer))%") - print( - "Keyword Precision/Recall/F1: " + - "\(String(format: "%.2f", keywordPrecision))/" + - "\(String(format: "%.2f", keywordRecall))/" + - "\(String(format: "%.2f", keywordFscore))" - ) - print("Total audio: \(String(format: "%.1f", totalAudioDuration))s") - print("Total processing: \(String(format: "%.1f", totalProcessingTime))s") - if totalProcessingTime > 0 { - print("RTFx: \(String(format: "%.2f", totalAudioDuration / totalProcessingTime))x") - } - print(String(repeating: "=", count: 60)) - - // Sort results by WER descending (worst first) - let sortedResults = results.sorted { r1, r2 in - let wer1 = r1["wer"] as? Double ?? 0 - let wer2 = r2["wer"] as? Double ?? 0 - return wer1 > wer2 - } - - // Save to JSON - let summaryDict: [String: Any] = [ - "totalTests": results.count, - "avgWer": round(avgWer * 100) / 100, - "keywordTruePositives": totalKeywordTruePositives, - "keywordFalsePositives": totalKeywordFalsePositives, - "keywordFalseNegatives": totalKeywordFalseNegatives, - "keywordPredicted": totalKeywordPredicted, - "keywordReference": totalKeywordReference, - "keywordPrecision": round(keywordPrecision * 1000) / 1000, - "keywordRecall": round(keywordRecall * 1000) / 1000, - "keywordFscore": round(keywordFscore * 1000) / 1000, - "totalAudioDuration": round(totalAudioDuration * 100) / 100, - "totalProcessingTime": round(totalProcessingTime * 100) / 100, - ] - - let output: [String: Any] = [ - "model": modelPath, - "keywordMode": keywordMode.rawValue, - "summary": summaryDict, - "results": sortedResults, - ] - - let jsonData = try JSONSerialization.data(withJSONObject: output, options: [.prettyPrinted, .sortedKeys]) - try jsonData.write(to: URL(fileURLWithPath: outputFile)) - print("\nResults written to: \(outputFile)") - - } catch { - print("ERROR: Benchmark failed: \(error)") - } - } - - private static func collectFileIds(from dataDir: URL, maxFiles: Int?) throws -> [String] { - var fileIds: [String] = [] - let suffix = ".dictionary.txt" - - let fileManager = FileManager.default - let contents = try fileManager.contentsOfDirectory(at: dataDir, includingPropertiesForKeys: nil) - - for url in contents.sorted(by: { $0.path < $1.path }) { - let name = url.lastPathComponent - if name.hasSuffix(suffix) { - let data = try? Data(contentsOf: url) - if let data = data, !data.isEmpty { - let fileId = String(name.dropLast(suffix.count)) - fileIds.append(fileId) - } - } - } - - if let maxFiles = maxFiles { - return Array(fileIds.prefix(maxFiles)) - } - return fileIds - } - - private static func processFile( - fileId: String, - dataDir: URL, - asrManager: AsrManager, - ctcModels: CtcModels, - spotter: CtcKeywordSpotter, - keywordMode: KeywordMode, - keywordIndex: [String: [String]] - ) async throws -> [String: Any]? { - let wavFile = dataDir.appendingPathComponent("\(fileId).wav") - let dictionaryFile = dataDir.appendingPathComponent("\(fileId).dictionary.txt") - let textFile = dataDir.appendingPathComponent("\(fileId).text.txt") - - let fm = FileManager.default - guard fm.fileExists(atPath: wavFile.path), - fm.fileExists(atPath: dictionaryFile.path) - else { - return nil - } - - // Load dictionary words (chunk or file keywords) - let dictionaryWords = try loadDictionaryWords( - fileId: fileId, - dictionaryFile: dictionaryFile, - keywordMode: keywordMode, - keywordIndex: keywordIndex - ) - - // Load reference text - let referenceRaw = - (try? String(contentsOf: textFile, encoding: .utf8))?.trimmingCharacters(in: .whitespacesAndNewlines) ?? "" - - // Get audio samples - let audioFile = try AVAudioFile(forReading: wavFile) - let audioLength = Double(audioFile.length) / audioFile.processingFormat.sampleRate - let format = audioFile.processingFormat - let frameCount = AVAudioFrameCount(audioFile.length) - - guard let buffer = AVAudioPCMBuffer(pcmFormat: format, frameCapacity: frameCount) else { - throw NSError( - domain: "CtcEarningsBenchmark", code: 1, - userInfo: [NSLocalizedDescriptionKey: "Failed to create audio buffer"]) - } - try audioFile.read(into: buffer) - - // Resample to 16kHz - let converter = AudioConverter() - let samples = try converter.resampleBuffer(buffer) - - let startTime = Date() - - // 1. TDT transcription for low WER - let tdtResult = try await asrManager.transcribe(wavFile) - - // Skip files where TDT returns empty (some audio files cause model issues) - if tdtResult.text.isEmpty { - print(" SKIPPED: TDT returned empty transcription") - return nil - } - - // 2. Build custom vocabulary for CTC keyword spotting - var vocabTerms: [CustomVocabularyTerm] = [] - for word in dictionaryWords { - let tokenIds = tokenize(word, vocabulary: ctcModels.vocabulary) - if !tokenIds.isEmpty { - let term = CustomVocabularyTerm( - text: word, - weight: nil, - aliases: nil, - tokenIds: nil, - ctcTokenIds: tokenIds - ) - vocabTerms.append(term) - } - } - let customVocab = CustomVocabularyContext(terms: vocabTerms) - - // 3. CTC keyword spotting for high recall dictionary detection - let spotResult = try await spotter.spotKeywordsWithLogProbs( - audioSamples: samples, - customVocabulary: customVocab, - minScore: nil - ) - - // 4. Post-process: Use VocabularyRescorer with Argmax-style parameters - // Argmax uses cbw=3.0 (context-biasing weight) for boosting vocab terms - let useRescorer = ProcessInfo.processInfo.environment["NO_CTC_RESCORING"] != "1" - let hypothesis: String - if useRescorer { - let rescorerConfig = VocabularyRescorer.Config( - minScoreAdvantage: 1.0, // Lower threshold - rely more on CTC scoring - minVocabScore: -15.0, // Permissive to include more detections - maxOriginalScoreForReplacement: -2.0, // Don't replace very confident words - vocabBoostWeight: 3.0 // Argmax cbw=3.0 - ) - let rescorer = VocabularyRescorer( - spotter: spotter, - vocabulary: customVocab, - config: rescorerConfig - ) - let rescoreResult = rescorer.rescore(transcript: tdtResult.text, spotResult: spotResult) - hypothesis = rescoreResult.text - } else { - hypothesis = tdtResult.text // Baseline: no CTC corrections - } - - let processingTime = Date().timeIntervalSince(startTime) - - // Normalize texts - let referenceNormalized = TextNormalizer.normalize(referenceRaw) - let hypothesisNormalized = TextNormalizer.normalize(hypothesis) - - // Keyword sets for precision/recall - let referenceKeywords = keywordsInText(referenceNormalized, dictionaryWords: dictionaryWords) - let predictedKeywords = keywordsInText(hypothesisNormalized, dictionaryWords: dictionaryWords) - let truePositives = referenceKeywords.intersection(predictedKeywords) - let falsePositives = predictedKeywords.subtracting(referenceKeywords) - let falseNegatives = referenceKeywords.subtracting(predictedKeywords) - let keywordPrecision = predictedKeywords.isEmpty ? 0 : Double(truePositives.count) / Double(predictedKeywords.count) - let keywordRecall = referenceKeywords.isEmpty ? 0 : Double(truePositives.count) / Double(referenceKeywords.count) - let keywordFscore = - (keywordPrecision + keywordRecall) > 0 - ? 2 * keywordPrecision * keywordRecall / (keywordPrecision + keywordRecall) - : 0 - - let referenceWords = referenceNormalized.components(separatedBy: CharacterSet.whitespacesAndNewlines).filter { - !$0.isEmpty - } - let hypothesisWords = hypothesisNormalized.components(separatedBy: CharacterSet.whitespacesAndNewlines).filter { - !$0.isEmpty - } - - // Calculate WER - let wer: Double - if referenceWords.isEmpty { - wer = hypothesisWords.isEmpty ? 0.0 : 1.0 - } else { - wer = calculateWER(reference: referenceWords, hypothesis: hypothesisWords) - } - - // Count dictionary detections (debug only) - let minCtcScore: Float = -15.0 // Permissive threshold for detection - var detectionDetails: [[String: Any]] = [] - var ctcFoundWords: Set = [] - - // 1. CTC detections - for detection in spotResult.detections { - let inRef = referenceKeywords.contains(detection.term.text.lowercased()) - let detail: [String: Any] = [ - "word": detection.term.text, - "score": round(Double(detection.score) * 100) / 100, - "startTime": round(detection.startTime * 100) / 100, - "endTime": round(detection.endTime * 100) / 100, - "source": "ctc", - "inReference": inRef, - ] - detectionDetails.append(detail) - - if detection.score >= minCtcScore { // Use >= to include edge cases - ctcFoundWords.insert(detection.term.text.lowercased()) - } - } - - // 2. Fallback: check hypothesis for dictionary words not found by CTC - let hypothesisLower = hypothesis.lowercased() - for word in dictionaryWords { - let wordLower = word.lowercased() - if !ctcFoundWords.contains(wordLower) { - // Check if word appears as whole word in hypothesis (avoid substring false positives) - let pattern = "\\b\(NSRegularExpression.escapedPattern(for: wordLower))\\b" - if let regex = try? NSRegularExpression(pattern: pattern, options: []), - regex.firstMatch( - in: hypothesisLower, options: [], - range: NSRange(hypothesisLower.startIndex..., in: hypothesisLower)) != nil - { - ctcFoundWords.insert(wordLower) - let inRef = referenceKeywords.contains(wordLower) - let detail: [String: Any] = [ - "word": word, - "score": 0.0, - "startTime": 0.0, - "endTime": 0.0, - "source": "hypothesis", - "inReference": inRef, - ] - detectionDetails.append(detail) - } - } - } - - let result: [String: Any] = [ - "fileId": fileId, - "reference": referenceNormalized, - "hypothesis": hypothesisNormalized, - "wer": round(wer * 10000) / 100, - "dictFound": predictedKeywords.count, - "dictTotal": referenceKeywords.count, - "keywordPredicted": predictedKeywords.count, - "keywordReference": referenceKeywords.count, - "keywordTruePositives": truePositives.count, - "keywordFalsePositives": falsePositives.count, - "keywordFalseNegatives": falseNegatives.count, - "keywordPrecision": round(keywordPrecision * 1000) / 1000, - "keywordRecall": round(keywordRecall * 1000) / 1000, - "keywordFscore": round(keywordFscore * 1000) / 1000, - "audioLength": round(audioLength * 100) / 100, - "processingTime": round(processingTime * 1000) / 1000, - "ctcDetections": detectionDetails, - ] - return result - } - - /// Simple tokenization using vocabulary lookup - private static func tokenize(_ text: String, vocabulary: [Int: String]) -> [Int] { - // Build reverse vocabulary (token -> id) - var tokenToId: [String: Int] = [:] - for (id, token) in vocabulary { - tokenToId[token] = id - } - - let normalizedText = text.lowercased() - var result: [Int] = [] - var position = normalizedText.startIndex - var isWordStart = true - - while position < normalizedText.endIndex { - var matched = false - let remaining = normalizedText.distance(from: position, to: normalizedText.endIndex) - var matchLength = min(20, remaining) - - while matchLength > 0 { - let endPos = normalizedText.index(position, offsetBy: matchLength) - let substring = String(normalizedText[position.. String { - // Filter detections by score - let validDetections = detections.filter { $0.score >= minScore } - guard !validDetections.isEmpty else { - return tdtResult.text - } - - var text = tdtResult.text - var usedDetections: Set = [] - - // PASS 1: Fuzzy matching for phonetically similar words - for detection in validDetections { - let keyword = detection.term.text - let keywordLower = keyword.lowercased() - let keywordParts = keywordLower.components(separatedBy: " ").filter { !$0.isEmpty } - - let words = text.components(separatedBy: .whitespacesAndNewlines).filter { !$0.isEmpty } - - // Handle multi-word keywords - if keywordParts.count > 1 { - for i in 0..<(words.count - keywordParts.count + 1) { - var allMatch = true - var matchedWords: [String] = [] - - for j in 0..= 3 - && !stopWords.contains(keywordLower) - - guard isProperNoun else { continue } - - // Look for "this is X" pattern specifically for names - let thisIsPattern = try? NSRegularExpression(pattern: "this is ([A-Z][a-z]+)", options: []) - if let regex = thisIsPattern { - let textRange = NSRange(text.startIndex..., in: text) - if let match = regex.firstMatch(in: text, options: [], range: textRange), - match.numberOfRanges > 1, - let captureRange = Range(match.range(at: 1), in: text) - { - let capturedWord = String(text[captureRange]) - let capturedLower = capturedWord.lowercased() - - // Skip if captured word is already a detected keyword - let isOtherKeyword = validDetections.contains { det in - det.term.text.lowercased() == capturedLower - } - - if !isOtherKeyword && !stopWords.contains(capturedLower) { - // Similar length check - if abs(capturedWord.count - keyword.count) <= 3 { - text = text.replacingOccurrences(of: capturedWord, with: keyword) - usedDetections.insert(keyword) - } - } - } - } - } - - return text - } - - /// Build word timings by merging subword tokens (tokens starting with "▁" begin new words) - private static func buildWordTimings( - from tokenTimings: [TokenTiming] - ) -> [(word: String, startTime: Double, endTime: Double)] { - var wordTimings: [(word: String, startTime: Double, endTime: Double)] = [] - var currentWord = "" - var wordStart: Double = 0 - var wordEnd: Double = 0 - - for timing in tokenTimings { - let token = timing.token - - // Skip special tokens - if token.isEmpty || token == "" || token == "" { - continue - } - - // Check if this starts a new word (has ▁ prefix or is first token) - let startsNewWord = token.hasPrefix("▁") || currentWord.isEmpty - - if startsNewWord && !currentWord.isEmpty { - // Save previous word - wordTimings.append((word: currentWord, startTime: wordStart, endTime: wordEnd)) - currentWord = "" - } - - if startsNewWord { - currentWord = token.hasPrefix("▁") ? String(token.dropFirst()) : token - wordStart = timing.startTime - } else { - currentWord += token - } - wordEnd = timing.endTime - } - - // Save final word - if !currentWord.isEmpty { - wordTimings.append((word: currentWord, startTime: wordStart, endTime: wordEnd)) - } - - return wordTimings - } - - /// Common English words that should never be replaced by keyword matching - private static let stopWords: Set = [ - // Pronouns - "i", "you", "he", "she", "it", "we", "they", "me", "him", "her", "us", "them", - "my", "your", "his", "its", "our", "their", "mine", "yours", "hers", "ours", "theirs", - "this", "that", "these", "those", "who", "whom", "what", "which", "whose", - // Common verbs - "is", "are", "was", "were", "be", "been", "being", "am", - "have", "has", "had", "having", "do", "does", "did", "doing", "done", - "will", "would", "shall", "should", "may", "might", "must", "can", "could", - "get", "got", "getting", "go", "goes", "went", "going", "gone", - "come", "came", "coming", "see", "saw", "seen", "know", "knew", "known", - "think", "thought", "make", "made", "take", "took", "taken", "give", "gave", "given", - "say", "said", "tell", "told", "ask", "asked", "use", "used", "want", "wanted", - "need", "needed", "try", "tried", "let", "put", "keep", "kept", "look", "looked", - // Articles and determiners - "a", "an", "the", "some", "any", "no", "every", "each", "all", "both", "few", "many", - "much", "more", "most", "other", "another", "such", - // Prepositions - "in", "on", "at", "to", "for", "of", "with", "by", "from", "up", "down", "out", - "about", "into", "over", "after", "before", "between", "under", "through", "during", - // Conjunctions - "and", "or", "but", "so", "yet", "nor", "if", "then", "than", "because", "while", - "although", "unless", "since", "when", "where", "as", - // Adverbs - "not", "very", "just", "also", "only", "even", "still", "already", "always", "never", - "often", "sometimes", "usually", "really", "well", "now", "here", "there", "how", "why", - // Common words - "yes", "no", "okay", "ok", "thank", "thanks", "please", "sorry", "hello", "hi", "bye", - "good", "great", "bad", "new", "old", "first", "last", "long", "short", "big", "small", - "high", "low", "right", "left", "next", "back", "same", "different", "own", "able", - "way", "thing", "things", "time", "times", "year", "years", "day", "days", "week", "weeks", - "part", "place", "case", "point", "fact", "end", "kind", "lot", "set", - // Numbers - "one", "two", "three", "four", "five", "six", "seven", "eight", "nine", "ten", - "hundred", "thousand", "million", "billion", - ] - - /// Check if two words are similar (edit distance / length ratio) - private static func isSimilar(_ a: String, _ b: String) -> Bool { - // Never match stop words - they're too common to be proper nouns - if stopWords.contains(a) || stopWords.contains(b) { - return false - } - - let maxLen = max(a.count, b.count) - let minLen = min(a.count, b.count) - guard maxLen > 0, minLen >= 3 else { return false } - - // Allow more length difference for longer words - let lenDiff = abs(a.count - b.count) - if lenDiff > max(3, maxLen / 2) { return false } - - // Calculate edit distance - let distance = editDistance(a, b) - - // More aggressive threshold: allow up to 40% of max length as edits - let threshold = max(2, Int(Double(maxLen) * 0.4)) - - // Also check if one is substring of other (handles "Erik" in "Ririek") - if a.contains(b) || b.contains(a) { - return true - } - - // Check common prefix/suffix (handles "Heri" vs "Harry") - let commonPrefix = commonPrefixLength(a, b) - let commonSuffix = commonSuffixLength(a, b) - if commonPrefix >= 2 || commonSuffix >= 2 { - return distance <= threshold + 1 - } - - return distance <= threshold - } - - /// Get length of common prefix - private static func commonPrefixLength(_ a: String, _ b: String) -> Int { - let aChars = Array(a) - let bChars = Array(b) - var count = 0 - for i in 0.. Int { - let aChars = Array(a.reversed()) - let bChars = Array(b.reversed()) - var count = 0 - for i in 0.. Int { - let a = Array(a) - let b = Array(b) - let m = a.count - let n = b.count - - if m == 0 { return n } - if n == 0 { return m } - - var dp = Array(repeating: Array(repeating: 0, count: n + 1), count: m + 1) - - for i in 0...m { dp[i][0] = i } - for j in 0...n { dp[0][j] = j } - - for i in 1...m { - for j in 1...n { - if a[i - 1] == b[j - 1] { - dp[i][j] = dp[i - 1][j - 1] - } else { - dp[i][j] = 1 + min(dp[i - 1][j - 1], min(dp[i - 1][j], dp[i][j - 1])) - } - } - } - - return dp[m][n] - } - - /// Match the case pattern of the original word - private static func matchCase(_ keyword: String, to original: String) -> String { - let origClean = original.trimmingCharacters(in: .punctuationCharacters) - - // Check case pattern - if origClean.first?.isUppercase == true { - // Capitalize first letter - return keyword.prefix(1).uppercased() + keyword.dropFirst() - } - return keyword - } - - private static func calculateWER(reference: [String], hypothesis: [String]) -> Double { - if reference.isEmpty { - return hypothesis.isEmpty ? 0.0 : 1.0 - } - - let m = reference.count - let n = hypothesis.count - var dp = Array(repeating: Array(repeating: 0, count: n + 1), count: m + 1) - - for i in 0...m { dp[i][0] = i } - for j in 0...n { dp[0][j] = j } - - for i in 1...m { - for j in 1...n { - if reference[i - 1] == hypothesis[j - 1] { - dp[i][j] = dp[i - 1][j - 1] - } else { - dp[i][j] = min(dp[i - 1][j - 1], min(dp[i - 1][j], dp[i][j - 1])) + 1 - } - } - } - - return Double(dp[m][n]) / Double(m) - } - - private static func printUsage() { - print( - """ - CTC Earnings Benchmark (TDT + CTC keyword spotting) - - Usage: fluidaudio ctc-earnings-benchmark [options] - - Options: - --data-dir Path to earnings test dataset (auto-detected if downloaded) - --ctc-model Path to CTC model directory (auto-detected if in standard location) - --max-files Maximum number of files to process - --output, -o Output JSON file (default: ctc_earnings_benchmark.json) - --auto-download Download earnings22-kws dataset if not found - --keyword-mode Keyword mode: chunk or file (default: chunk) - - Default locations: - Dataset: ~/Library/Application Support/FluidAudio/earnings22-kws/test-dataset/ - CTC Model: ~/Library/Application Support/FluidAudio/Models/parakeet-ctc-110m-coreml/ - - Setup: - 1. Download dataset: fluidaudio download --dataset earnings22-kws - 2. Place CTC model in standard location - 3. Run: fluidaudio ctc-earnings-benchmark - - Examples: - # Run with auto-detected paths - fluidaudio ctc-earnings-benchmark - - # Run with auto-download - fluidaudio ctc-earnings-benchmark --auto-download - - # Run with explicit paths - fluidaudio ctc-earnings-benchmark \\ - --data-dir /path/to/test-dataset \\ - --ctc-model /path/to/parakeet-ctc-110m-coreml \\ - --max-files 100 - """) - } - - private static func parseKeywordMode(_ value: String) -> KeywordMode? { - switch value.lowercased() { - case "chunk", "chunk-keywords": - return .chunk - case "file", "file-keywords": - return .file - default: - return nil - } - } - - private static func parentId(from fileId: String) -> String { - guard let range = fileId.range(of: "_chunk") else { - return fileId - } - return String(fileId[.. [String: [String]] { - guard keywordMode == .file else { - return [:] - } - - var index: [String: Set] = [:] - let suffix = ".dictionary.txt" - let fileManager = FileManager.default - let contents = try fileManager.contentsOfDirectory(at: dataDir, includingPropertiesForKeys: nil) - - for url in contents { - let name = url.lastPathComponent - guard name.hasSuffix(suffix) else { continue } - let fileId = String(name.dropLast(suffix.count)) - let parent = parentId(from: fileId) - let words = try loadDictionaryWords(from: url) - var set = index[parent] ?? Set() - set.formUnion(words) - index[parent] = set - } - - return index.mapValues { Array($0).sorted() } - } - - private static func loadDictionaryWords( - fileId: String, - dictionaryFile: URL, - keywordMode: KeywordMode, - keywordIndex: [String: [String]] - ) throws -> [String] { - switch keywordMode { - case .chunk: - return try loadDictionaryWords(from: dictionaryFile) - case .file: - let parent = parentId(from: fileId) - if let words = keywordIndex[parent] { - return words - } - return try loadDictionaryWords(from: dictionaryFile) - } - } - - private static func loadDictionaryWords(from url: URL) throws -> [String] { - let dictionaryContent = try String(contentsOf: url, encoding: .utf8) - return dictionaryContent - .components(separatedBy: .newlines) - .map { $0.trimmingCharacters(in: .whitespacesAndNewlines) } - .filter { !$0.isEmpty } - } - - private static func keywordsInText(_ text: String, dictionaryWords: [String]) -> Set { - let textLower = text.lowercased() - var result: Set = [] - - for word in dictionaryWords { - let wordLower = word.lowercased() - let pattern = "\\b\(NSRegularExpression.escapedPattern(for: wordLower))\\b" - guard let regex = try? NSRegularExpression(pattern: pattern, options: []) else { continue } - let range = NSRange(textLower.startIndex..., in: textLower) - if regex.firstMatch(in: textLower, options: [], range: range) != nil { - result.insert(wordLower) - } - } - return result - } -} -#endif diff --git a/cli/HybridEarningsBenchmark.swift b/cli/HybridEarningsBenchmark.swift deleted file mode 100644 index dcd8023757245b94deed24d634195358afc82c7d..0000000000000000000000000000000000000000 --- a/cli/HybridEarningsBenchmark.swift +++ /dev/null @@ -1,554 +0,0 @@ -#if os(macOS) -import AVFoundation -import FluidAudio -import Foundation - -/// Earnings22 benchmark using ONLY the Hybrid 110M model (single encoder). -/// CTC head provides both transcription AND keyword spotting from the same encoder. -public enum HybridEarningsBenchmark { - - private enum KeywordMode: String { - case chunk - case file - } - - public static func runCLI(arguments: [String]) async { - if arguments.contains("--help") || arguments.contains("-h") { - printUsage() - return - } - - // Parse arguments - var outputFile = "hybrid_earnings_benchmark.json" - var maxFiles: Int? = nil - var decodingMode: HybridDecodingMode = .tdt - var useRescoring = false - var keywordMode: KeywordMode = .chunk - - var i = 0 - while i < arguments.count { - switch arguments[i] { - case "--output", "-o": - if i + 1 < arguments.count { - outputFile = arguments[i + 1] - i += 1 - } - case "--max-files": - if i + 1 < arguments.count { - maxFiles = Int(arguments[i + 1]) - i += 1 - } - case "--ctc": - decodingMode = .ctc - case "--tdt": - decodingMode = .tdt - case "--rescore": - useRescoring = true - case "--keyword-mode": - if i + 1 < arguments.count, let mode = parseKeywordMode(arguments[i + 1]) { - keywordMode = mode - i += 1 - } - default: - break - } - i += 1 - } - - let dataDir = DatasetDownloader.getEarnings22Directory().appendingPathComponent("test-dataset") - guard FileManager.default.fileExists(atPath: dataDir.path) else { - print("ERROR: Earnings dataset not found at \(dataDir.path)") - print("Download with: fluidaudio download --dataset earnings22-kws") - return - } - - let modeStr = decodingMode == .ctc ? "CTC" : "TDT" - let rescoringStr = useRescoring ? " + Rescoring" : "" - print("Hybrid 110M Earnings Benchmark (Decoding: \(modeStr)\(rescoringStr))") - print(" Output file: \(outputFile)") - print(" Decoding mode: \(modeStr)") - print(" Rescoring: \(useRescoring ? "enabled" : "disabled")") - print(" Keyword mode: \(keywordMode.rawValue)") - - do { - // Load Hybrid 110M model (single encoder with CTC head) - print("Loading Hybrid 110M model...") - let hybridModels = try await HybridAsrModels.downloadAndLoad() - let hybridManager = HybridAsrManager(models: hybridModels, decodingMode: decodingMode) - let spotter = HybridKeywordSpotter(vocabulary: hybridModels.vocabulary, blankId: hybridModels.blankId) - print(" Vocab size: \(hybridModels.vocabSize)") - - // Collect test files - let fileIds = try collectFileIds(from: dataDir, maxFiles: maxFiles) - let keywordIndex = try buildKeywordIndex(dataDir: dataDir, keywordMode: keywordMode) - - if fileIds.isEmpty { - print("ERROR: No test files found") - return - } - - print("Processing \(fileIds.count) test files...") - - var results: [[String: Any]] = [] - var totalWer = 0.0 - var totalKeywordReference = 0 - var totalKeywordPredicted = 0 - var totalKeywordTruePositives = 0 - var totalKeywordFalsePositives = 0 - var totalKeywordFalseNegatives = 0 - var totalAudioDuration = 0.0 - var totalProcessingTime = 0.0 - - for (index, fileId) in fileIds.enumerated() { - print("[\(index + 1)/\(fileIds.count)] \(fileId)") - - if let result = try await processFile( - fileId: fileId, - dataDir: dataDir, - hybridManager: hybridManager, - spotter: spotter, - useRescoring: useRescoring, - keywordMode: keywordMode, - keywordIndex: keywordIndex - ) { - results.append(result) - totalWer += result["wer"] as? Double ?? 0 - totalKeywordReference += result["keywordReference"] as? Int ?? 0 - totalKeywordPredicted += result["keywordPredicted"] as? Int ?? 0 - totalKeywordTruePositives += result["keywordTruePositives"] as? Int ?? 0 - totalKeywordFalsePositives += result["keywordFalsePositives"] as? Int ?? 0 - totalKeywordFalseNegatives += result["keywordFalseNegatives"] as? Int ?? 0 - totalAudioDuration += result["audioLength"] as? Double ?? 0 - totalProcessingTime += result["processingTime"] as? Double ?? 0 - - let wer = result["wer"] as? Double ?? 0 - let precision = result["keywordPrecision"] as? Double ?? 0 - let recall = result["keywordRecall"] as? Double ?? 0 - let fscore = result["keywordFscore"] as? Double ?? 0 - print( - " WER: \(String(format: "%.1f", wer))%, " + - "KW P/R/F: \(String(format: "%.2f", precision))/" + - "\(String(format: "%.2f", recall))/" + - "\(String(format: "%.2f", fscore))" - ) - } - } - - // Calculate summary - let avgWer = results.isEmpty ? 0.0 : totalWer / Double(results.count) - let keywordPrecision = - totalKeywordPredicted > 0 - ? Double(totalKeywordTruePositives) / Double(totalKeywordPredicted) - : 0 - let keywordRecall = - totalKeywordReference > 0 - ? Double(totalKeywordTruePositives) / Double(totalKeywordReference) - : 0 - let keywordFscore = - (keywordPrecision + keywordRecall) > 0 - ? 2 * keywordPrecision * keywordRecall / (keywordPrecision + keywordRecall) - : 0 - - // Print summary - print("\n" + String(repeating: "=", count: 60)) - print("HYBRID 110M BENCHMARK (\(modeStr)\(rescoringStr))") - print(String(repeating: "=", count: 60)) - print("Model: parakeet-tdt-ctc-110m-hybrid") - print("Decoding: \(modeStr), Rescoring: \(useRescoring ? "yes" : "no")") - print("Total tests: \(results.count)") - print("Average WER: \(String(format: "%.2f", avgWer))%") - print( - "Keyword Precision/Recall/F1: " + - "\(String(format: "%.2f", keywordPrecision))/" + - "\(String(format: "%.2f", keywordRecall))/" + - "\(String(format: "%.2f", keywordFscore))" - ) - print("Total audio: \(String(format: "%.1f", totalAudioDuration))s") - print("Total processing: \(String(format: "%.1f", totalProcessingTime))s") - if totalProcessingTime > 0 { - print("RTFx: \(String(format: "%.2f", totalAudioDuration / totalProcessingTime))x") - } - print(String(repeating: "=", count: 60)) - - // Sort results by WER descending (worst first) - let sortedResults = results.sorted { r1, r2 in - let wer1 = r1["wer"] as? Double ?? 0 - let wer2 = r2["wer"] as? Double ?? 0 - return wer1 > wer2 - } - - // Save to JSON - let summaryDict: [String: Any] = [ - "totalTests": results.count, - "avgWer": round(avgWer * 100) / 100, - "keywordTruePositives": totalKeywordTruePositives, - "keywordFalsePositives": totalKeywordFalsePositives, - "keywordFalseNegatives": totalKeywordFalseNegatives, - "keywordPredicted": totalKeywordPredicted, - "keywordReference": totalKeywordReference, - "keywordPrecision": round(keywordPrecision * 1000) / 1000, - "keywordRecall": round(keywordRecall * 1000) / 1000, - "keywordFscore": round(keywordFscore * 1000) / 1000, - "totalAudioDuration": round(totalAudioDuration * 100) / 100, - "totalProcessingTime": round(totalProcessingTime * 100) / 100, - ] - - let output: [String: Any] = [ - "model": "parakeet-tdt-ctc-110m-hybrid", - "approach": "single-encoder", - "decodingMode": modeStr, - "rescoring": useRescoring, - "keywordMode": keywordMode.rawValue, - "summary": summaryDict, - "results": sortedResults, - ] - - let jsonData = try JSONSerialization.data(withJSONObject: output, options: [.prettyPrinted, .sortedKeys]) - try jsonData.write(to: URL(fileURLWithPath: outputFile)) - print("\nResults written to: \(outputFile)") - - } catch { - print("ERROR: \(error)") - } - } - - private static func collectFileIds(from dataDir: URL, maxFiles: Int?) throws -> [String] { - var fileIds: [String] = [] - let suffix = ".dictionary.txt" - - let fileManager = FileManager.default - let contents = try fileManager.contentsOfDirectory(at: dataDir, includingPropertiesForKeys: nil) - - for url in contents.sorted(by: { $0.path < $1.path }) { - let name = url.lastPathComponent - if name.hasSuffix(suffix) { - let data = try? Data(contentsOf: url) - if let data = data, !data.isEmpty { - let fileId = String(name.dropLast(suffix.count)) - fileIds.append(fileId) - } - } - } - - if let maxFiles = maxFiles { - return Array(fileIds.prefix(maxFiles)) - } - return fileIds - } - - private static func processFile( - fileId: String, - dataDir: URL, - hybridManager: HybridAsrManager, - spotter: HybridKeywordSpotter, - useRescoring: Bool, - keywordMode: KeywordMode, - keywordIndex: [String: [String]] - ) async throws -> [String: Any]? { - let wavFile = dataDir.appendingPathComponent("\(fileId).wav") - let dictionaryFile = dataDir.appendingPathComponent("\(fileId).dictionary.txt") - let textFile = dataDir.appendingPathComponent("\(fileId).text.txt") - - let fm = FileManager.default - guard fm.fileExists(atPath: wavFile.path), - fm.fileExists(atPath: dictionaryFile.path) - else { - return nil - } - - // Load dictionary words (chunk or file keywords) - let dictionaryWords = try loadDictionaryWords( - fileId: fileId, - dictionaryFile: dictionaryFile, - keywordMode: keywordMode, - keywordIndex: keywordIndex - ) - - // Load reference text - let referenceRaw = - (try? String(contentsOf: textFile, encoding: .utf8))?.trimmingCharacters(in: .whitespacesAndNewlines) ?? "" - - // Get audio samples - let audioFile = try AVAudioFile(forReading: wavFile) - let audioLength = Double(audioFile.length) / audioFile.processingFormat.sampleRate - let format = audioFile.processingFormat - let frameCount = AVAudioFrameCount(audioFile.length) - - guard let buffer = AVAudioPCMBuffer(pcmFormat: format, frameCapacity: frameCount) else { - return nil - } - try audioFile.read(into: buffer) - - // Resample to 16kHz - let converter = AudioConverter() - let samples = try converter.resampleBuffer(buffer) - - // Build custom vocabulary for keyword spotting - var vocabTerms: [CustomVocabularyTerm] = [] - for word in dictionaryWords { - let term = CustomVocabularyTerm( - text: word, - weight: nil, - aliases: nil, - tokenIds: nil, - ctcTokenIds: nil - ) - vocabTerms.append(term) - } - let customVocab = CustomVocabularyContext(terms: vocabTerms) - - // Run Hybrid 110M using new API (TDT transcription + CTC keyword detection) - let rescorerConfig: HybridTextRescorer.Config? = useRescoring ? .default : nil - let hybridResult = try await hybridManager.transcribeHybrid( - audioSamples: samples, - customVocabulary: customVocab, - rescorerConfig: rescorerConfig - ) - - // Skip if empty transcription - if hybridResult.text.isEmpty { - print(" SKIPPED: Empty transcription") - return nil - } - - let detections = hybridResult.keywordDetections - let processingTime = hybridResult.processingTime - - // Use hybrid transcription as hypothesis (may be rescored if enabled) - let hypothesis = hybridResult.text - - // Normalize texts - let referenceNormalized = TextNormalizer.normalize(referenceRaw) - let hypothesisNormalized = TextNormalizer.normalize(hypothesis) - - // Keyword sets for precision/recall - let referenceKeywords = keywordsInText(referenceNormalized, dictionaryWords: dictionaryWords) - let predictedKeywords = keywordsInText(hypothesisNormalized, dictionaryWords: dictionaryWords) - let truePositives = referenceKeywords.intersection(predictedKeywords) - let falsePositives = predictedKeywords.subtracting(referenceKeywords) - let falseNegatives = referenceKeywords.subtracting(predictedKeywords) - let keywordPrecision = predictedKeywords.isEmpty ? 0 : Double(truePositives.count) / Double(predictedKeywords.count) - let keywordRecall = referenceKeywords.isEmpty ? 0 : Double(truePositives.count) / Double(referenceKeywords.count) - let keywordFscore = - (keywordPrecision + keywordRecall) > 0 - ? 2 * keywordPrecision * keywordRecall / (keywordPrecision + keywordRecall) - : 0 - - let referenceWords = referenceNormalized.components(separatedBy: CharacterSet.whitespacesAndNewlines).filter { - !$0.isEmpty - } - let hypothesisWords = hypothesisNormalized.components(separatedBy: CharacterSet.whitespacesAndNewlines).filter { - !$0.isEmpty - } - - // Calculate WER - let wer: Double - if referenceWords.isEmpty { - wer = hypothesisWords.isEmpty ? 0.0 : 1.0 - } else { - wer = calculateWER(reference: referenceWords, hypothesis: hypothesisWords) - } - - // Count dictionary detections for debugging - let minCtcScore: Float = -15.0 - var detectionDetails: [[String: Any]] = [] - var foundWords: Set = [] - - // CTC detections - for detection in detections { - let inRef = referenceKeywords.contains(detection.term.text.lowercased()) - let detail: [String: Any] = [ - "word": detection.term.text, - "score": round(Double(detection.score) * 100) / 100, - "startTime": round(detection.startTime * 100) / 100, - "endTime": round(detection.endTime * 100) / 100, - "source": "ctc", - "inReference": inRef, - ] - detectionDetails.append(detail) - - if detection.score >= minCtcScore { - foundWords.insert(detection.term.text.lowercased()) - } - } - - // Fallback: check hypothesis for dictionary words not found by CTC - let hypothesisLower = hypothesis.lowercased() - for word in dictionaryWords { - let wordLower = word.lowercased() - if !foundWords.contains(wordLower) { - let pattern = "\\b\(NSRegularExpression.escapedPattern(for: wordLower))\\b" - if let regex = try? NSRegularExpression(pattern: pattern, options: []), - regex.firstMatch( - in: hypothesisLower, options: [], - range: NSRange(hypothesisLower.startIndex..., in: hypothesisLower)) != nil - { - foundWords.insert(wordLower) - let inRef = referenceKeywords.contains(wordLower) - let detail: [String: Any] = [ - "word": word, - "score": 0.0, - "startTime": 0.0, - "endTime": 0.0, - "source": "hypothesis", - "inReference": inRef, - ] - detectionDetails.append(detail) - } - } - } - - let result: [String: Any] = [ - "fileId": fileId, - "reference": referenceNormalized, - "hypothesis": hypothesisNormalized, - "wer": round(wer * 10000) / 100, - "dictFound": predictedKeywords.count, - "dictTotal": referenceKeywords.count, - "keywordPredicted": predictedKeywords.count, - "keywordReference": referenceKeywords.count, - "keywordTruePositives": truePositives.count, - "keywordFalsePositives": falsePositives.count, - "keywordFalseNegatives": falseNegatives.count, - "keywordPrecision": round(keywordPrecision * 1000) / 1000, - "keywordRecall": round(keywordRecall * 1000) / 1000, - "keywordFscore": round(keywordFscore * 1000) / 1000, - "audioLength": round(audioLength * 100) / 100, - "processingTime": round(processingTime * 1000) / 1000, - "ctcDetections": detectionDetails, - ] - return result - } - - private static func calculateWER(reference: [String], hypothesis: [String]) -> Double { - if reference.isEmpty { - return hypothesis.isEmpty ? 0.0 : 1.0 - } - - let m = reference.count - let n = hypothesis.count - var dp = Array(repeating: Array(repeating: 0, count: n + 1), count: m + 1) - - for i in 0...m { dp[i][0] = i } - for j in 0...n { dp[0][j] = j } - - for i in 1...m { - for j in 1...n { - if reference[i - 1] == hypothesis[j - 1] { - dp[i][j] = dp[i - 1][j - 1] - } else { - dp[i][j] = min(dp[i - 1][j - 1], min(dp[i - 1][j], dp[i][j - 1])) + 1 - } - } - } - - return Double(dp[m][n]) / Double(m) - } - - private static func printUsage() { - print( - """ - Hybrid 110M Earnings Benchmark (Single Encoder) - - Usage: fluidaudio hybrid-earnings-benchmark [options] - - This benchmark uses ONLY the Hybrid 110M model: - - Single encoder provides CTC log-probs - - CTC greedy decode for transcription - - CTC keyword spotting from same encoder output - - Options: - --max-files Maximum number of files to process - --output, -o Output JSON file (default: hybrid_earnings_benchmark.json) - --keyword-mode Keyword mode: chunk or file (default: chunk) - - Compare with: - fluidaudio ctc-earnings-benchmark (Canary-CTC + TDT 0.6B, two encoders) - """) - } - - private static func parseKeywordMode(_ value: String) -> KeywordMode? { - switch value.lowercased() { - case "chunk", "chunk-keywords": - return .chunk - case "file", "file-keywords": - return .file - default: - return nil - } - } - - private static func parentId(from fileId: String) -> String { - guard let range = fileId.range(of: "_chunk") else { - return fileId - } - return String(fileId[.. [String: [String]] { - guard keywordMode == .file else { - return [:] - } - - var index: [String: Set] = [:] - let suffix = ".dictionary.txt" - let fileManager = FileManager.default - let contents = try fileManager.contentsOfDirectory(at: dataDir, includingPropertiesForKeys: nil) - - for url in contents { - let name = url.lastPathComponent - guard name.hasSuffix(suffix) else { continue } - let fileId = String(name.dropLast(suffix.count)) - let parent = parentId(from: fileId) - let words = try loadDictionaryWords(from: url) - var set = index[parent] ?? Set() - set.formUnion(words) - index[parent] = set - } - - return index.mapValues { Array($0).sorted() } - } - - private static func loadDictionaryWords( - fileId: String, - dictionaryFile: URL, - keywordMode: KeywordMode, - keywordIndex: [String: [String]] - ) throws -> [String] { - switch keywordMode { - case .chunk: - return try loadDictionaryWords(from: dictionaryFile) - case .file: - let parent = parentId(from: fileId) - if let words = keywordIndex[parent] { - return words - } - return try loadDictionaryWords(from: dictionaryFile) - } - } - - private static func loadDictionaryWords(from url: URL) throws -> [String] { - let dictionaryContent = try String(contentsOf: url, encoding: .utf8) - return dictionaryContent - .components(separatedBy: .newlines) - .map { $0.trimmingCharacters(in: .whitespacesAndNewlines) } - .filter { !$0.isEmpty } - } - - private static func keywordsInText(_ text: String, dictionaryWords: [String]) -> Set { - let textLower = text.lowercased() - var result: Set = [] - - for word in dictionaryWords { - let wordLower = word.lowercased() - let pattern = "\\b\(NSRegularExpression.escapedPattern(for: wordLower))\\b" - guard let regex = try? NSRegularExpression(pattern: pattern, options: []) else { continue } - let range = NSRange(textLower.startIndex..., in: textLower) - if regex.firstMatch(in: textLower, options: [], range: range) != nil { - result.insert(wordLower) - } - } - return result - } -} -#endif diff --git a/config.json b/config.json deleted file mode 100644 index 9e26dfeeb6e641a33dae4961196235bdb965b21b..0000000000000000000000000000000000000000 --- a/config.json +++ /dev/null @@ -1 +0,0 @@ -{} \ No newline at end of file diff --git a/convert/.DS_Store b/convert/.DS_Store deleted file mode 100644 index 75ff82f86c29bd5d5eabcec4b41e4db80ef621da..0000000000000000000000000000000000000000 Binary files a/convert/.DS_Store and /dev/null differ diff --git a/convert/parakeet-tdt-ctc-110m/convert_tdt_decoder.py b/convert/parakeet-tdt-ctc-110m/convert_tdt_decoder.py deleted file mode 100644 index 4ed15a38318bb788204bfdbd2bc466cff9ff6e5c..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/convert_tdt_decoder.py +++ /dev/null @@ -1,323 +0,0 @@ -#!/usr/bin/env python3 -""" -Convert Parakeet TDT-CTC 110M decoder components to CoreML. - -This script exports the TDT decoder (prediction network) and joint network -with the SAME format as the working 0.6B model: -- JointDecision outputs token_id, token_prob, duration (argmax done inside) -- Uses shape [1, dim, 1] for encoder/decoder steps -- Matches the interface expected by TdtDecoderV3 -""" - -import argparse -import os -import torch -import torch.nn.functional as F -import coremltools as ct -import numpy as np -from pathlib import Path - -# NeMo imports -import nemo.collections.asr as nemo_asr - - -def get_model_config(model): - """Extract model configuration.""" - encoder_dim = None - pred_hidden = 640 # Default for parakeet models - num_layers = 1 - vocab_size = 1024 - num_durations = 5 - - # Get encoder dimension - if hasattr(model, 'encoder'): - encoder = model.encoder - if hasattr(encoder, 'd_model'): - encoder_dim = encoder.d_model - elif hasattr(encoder, '_feat_out'): - encoder_dim = encoder._feat_out - - # Get decoder config - if hasattr(model, 'decoder'): - decoder = model.decoder - if hasattr(decoder, 'pred_hidden'): - pred_hidden = decoder.pred_hidden - if hasattr(decoder, 'pred_rnn_layers'): - num_layers = decoder.pred_rnn_layers - - # Get joint config - if hasattr(model, 'joint'): - joint = model.joint - if hasattr(joint, 'num_extra_outputs'): - num_durations = joint.num_extra_outputs - if hasattr(joint, 'num_classes'): - vocab_size = joint.num_classes - num_durations - - return { - 'encoder_dim': encoder_dim, - 'pred_hidden': pred_hidden, - 'num_layers': num_layers, - 'vocab_size': vocab_size, - 'num_durations': num_durations, - } - - -class DecoderWrapper(torch.nn.Module): - """ - Wrapper for the RNNT/TDT decoder (prediction network). - - Matches 0.6B format: - - Input: targets[1,1], target_lengths[1], h_in[num_layers,1,pred_hidden], c_in[...] - - Output: decoder_output[1,pred_hidden,2], h_out[...], c_out[...] - """ - - def __init__(self, decoder, pred_hidden): - super().__init__() - self.decoder = decoder - self.pred_hidden = pred_hidden - - def forward(self, targets, target_lengths, h_in, c_in): - """ - Args: - targets: [1, 1] - previous token ID - target_lengths: [1] - always 1 - h_in: [num_layers, 1, pred_hidden] - c_in: [num_layers, 1, pred_hidden] - Returns: - decoder_output: [1, pred_hidden, 2] - prediction network output (transposed) - h_out: [num_layers, 1, pred_hidden] - c_out: [num_layers, 1, pred_hidden] - """ - state = (h_in, c_in) - # pred_output shape: [batch, time, pred_hidden] = [1, 1, pred_hidden] - pred_output, new_state = self.decoder.predict(targets, state=state, add_sos=False) - h_out, c_out = new_state - - # Transpose to [batch, pred_hidden, time] and concat two time steps - # (0.6B outputs [1, 640, 2] - we match this by duplicating) - pred_transposed = pred_output.transpose(1, 2) # [1, pred_hidden, 1] - decoder_output = torch.cat([pred_transposed, pred_transposed], dim=2) # [1, pred_hidden, 2] - - return decoder_output, h_out, c_out - - -class JointWrapper(torch.nn.Module): - """ - Wrapper for the TDT joint network with internal argmax. - - Matches 0.6B format: - - Input: encoder_step[1,encoder_dim,1], decoder_step[1,pred_hidden,1] - - Output: token_id[1,1,1], token_prob[1,1,1], duration[1,1,1] - """ - - def __init__(self, joint, vocab_size, num_durations=5): - super().__init__() - self.joint = joint - self.vocab_size = vocab_size - self.num_durations = num_durations - - def forward(self, encoder_step, decoder_step): - """ - Args: - encoder_step: [1, encoder_dim, 1] - decoder_step: [1, pred_hidden, 1] - Returns: - token_id: [1, 1, 1] - argmax token ID - token_prob: [1, 1, 1] - probability of selected token - duration: [1, 1, 1] - argmax duration bin - """ - # Transpose to [batch, 1, dim] for joint network - enc = encoder_step.transpose(1, 2) # [1, 1, encoder_dim] - dec = decoder_step.transpose(1, 2) # [1, 1, pred_hidden] - - # Run joint network - # Joint output: [1, 1, 1, vocab_size + 1 (blank) + num_durations] - joint_out = self.joint.joint(enc, dec) - - # Debug: print shape on first call - if not hasattr(self, '_debug_printed'): - self._debug_printed = True - print(f" Joint output shape: {joint_out.shape}") - print(f" Expected: vocab={self.vocab_size} + blank=1 + durations={self.num_durations} = {self.vocab_size + 1 + self.num_durations}") - - # Split: token logits include vocab + blank, durations are separate - # vocab_size = 1024 tokens (0-1023), blank = index 1024, durations = indices 1025+ - num_tokens = self.vocab_size + 1 # Include blank at vocab_size - logits = joint_out[..., :num_tokens] # [1, 1, 1, vocab_size + 1] - duration_logits = joint_out[..., num_tokens:] # [1, 1, 1, num_durations] - - # Apply softmax and get probabilities - probs = F.softmax(logits, dim=-1) - - # Argmax for token - token_id = torch.argmax(logits, dim=-1, keepdim=True) # [1, 1, 1, 1] - token_id = token_id.squeeze(-1) # [1, 1, 1] - - # Get probability of selected token - token_prob = torch.gather(probs, -1, token_id.unsqueeze(-1)) # [1, 1, 1, 1] - token_prob = token_prob.squeeze(-1) # [1, 1, 1] - - # Argmax for duration - duration = torch.argmax(duration_logits, dim=-1, keepdim=False) # [1, 1, 1] - - return token_id.int(), token_prob, duration.int() - - -def convert_decoder(model, config, output_dir: Path): - """Convert decoder to CoreML.""" - print(f"Converting Decoder...") - print(f" pred_hidden={config['pred_hidden']}, num_layers={config['num_layers']}") - - wrapper = DecoderWrapper(model.decoder, config['pred_hidden']) - wrapper.eval() - - # Create example inputs - targets = torch.zeros(1, 1, dtype=torch.long) - target_lengths = torch.ones(1, dtype=torch.long) - h_in = torch.zeros(config['num_layers'], 1, config['pred_hidden']) - c_in = torch.zeros(config['num_layers'], 1, config['pred_hidden']) - - # Trace the model - with torch.no_grad(): - traced = torch.jit.trace(wrapper, (targets, target_lengths, h_in, c_in)) - - # Convert to CoreML - mlmodel = ct.convert( - traced, - inputs=[ - ct.TensorType(name="targets", shape=(1, 1), dtype=np.int32), - ct.TensorType(name="target_lengths", shape=(1,), dtype=np.int32), - ct.TensorType(name="h_in", shape=(config['num_layers'], 1, config['pred_hidden']), dtype=np.float32), - ct.TensorType(name="c_in", shape=(config['num_layers'], 1, config['pred_hidden']), dtype=np.float32), - ], - outputs=[ - ct.TensorType(name="decoder_output"), - ct.TensorType(name="h_out"), - ct.TensorType(name="c_out"), - ], - minimum_deployment_target=ct.target.iOS17, - compute_precision=ct.precision.FLOAT16, - ) - - # Add metadata - mlmodel.author = "Fluid Inference" - mlmodel.short_description = "Hybrid TDT Decoder (110M)" - - # Save - output_path = output_dir / "Decoder.mlpackage" - mlmodel.save(str(output_path)) - print(f" Saved to {output_path}") - - return mlmodel - - -def convert_joint(model, config, output_dir: Path): - """Convert joint network to CoreML.""" - print(f"Converting JointDecision...") - print(f" encoder_dim={config['encoder_dim']}, pred_hidden={config['pred_hidden']}") - print(f" vocab_size={config['vocab_size']}, num_durations={config['num_durations']}") - - wrapper = JointWrapper( - model.joint, - vocab_size=config['vocab_size'], - num_durations=config['num_durations'] - ) - wrapper.eval() - - # Create example inputs - shape [1, dim, 1] - encoder_step = torch.randn(1, config['encoder_dim'], 1) - decoder_step = torch.randn(1, config['pred_hidden'], 1) - - # Trace the model - with torch.no_grad(): - traced = torch.jit.trace(wrapper, (encoder_step, decoder_step)) - - # Convert to CoreML - mlmodel = ct.convert( - traced, - inputs=[ - ct.TensorType(name="encoder_step", shape=(1, config['encoder_dim'], 1), dtype=np.float32), - ct.TensorType(name="decoder_step", shape=(1, config['pred_hidden'], 1), dtype=np.float32), - ], - outputs=[ - ct.TensorType(name="token_id"), - ct.TensorType(name="token_prob"), - ct.TensorType(name="duration"), - ], - minimum_deployment_target=ct.target.iOS17, - compute_precision=ct.precision.FLOAT16, - ) - - # Add metadata - mlmodel.author = "Fluid Inference" - mlmodel.short_description = "Hybrid Joint Decision (110M)" - - # Save - output_path = output_dir / "JointDecision.mlpackage" - mlmodel.save(str(output_path)) - print(f" Saved to {output_path}") - - return mlmodel - - -def main(): - parser = argparse.ArgumentParser(description="Convert TDT decoder to CoreML (0.6B format)") - parser.add_argument( - "--model-name", - default="nvidia/parakeet-tdt_ctc-110m", - help="NeMo model name or path" - ) - parser.add_argument( - "--output-dir", - type=Path, - default=Path("./output"), - help="Output directory for CoreML models" - ) - args = parser.parse_args() - - # Create output directory - args.output_dir.mkdir(parents=True, exist_ok=True) - - # Load model - print(f"Loading model: {args.model_name}") - model = nemo_asr.models.EncDecRNNTBPEModel.from_pretrained(args.model_name) - model.eval() - - # Get model configuration - config = get_model_config(model) - - # Auto-detect encoder dim if not found - if config['encoder_dim'] is None: - print("Auto-detecting encoder dimension...") - dummy_audio = torch.randn(1, 16000) - dummy_length = torch.tensor([16000]) - with torch.no_grad(): - enc_out, enc_len = model.encoder( - audio_signal=dummy_audio, - length=dummy_length - ) - config['encoder_dim'] = enc_out.shape[-1] - - print(f"\nModel config:") - for k, v in config.items(): - print(f" {k}: {v}") - - # Convert components - print() - convert_decoder(model, config, args.output_dir) - convert_joint(model, config, args.output_dir) - - print("\nConversion complete!") - print(f"Models saved to: {args.output_dir}") - print("\nNext steps:") - print("1. Compile to .mlmodelc:") - print(f" cd {args.output_dir}") - print(" xcrun coremlcompiler compile Decoder.mlpackage .") - print(" xcrun coremlcompiler compile JointDecision.mlpackage .") - print("2. Copy to model cache:") - print(" cp -r Decoder.mlmodelc JointDecision.mlmodelc ~/Library/Application\\ Support/FluidAudio/Models/parakeet-ctc-110m-coreml/") - print("3. Test with: swift run fluidaudio hybrid-earnings-benchmark --max-files 1") - - -if __name__ == "__main__": - main() diff --git a/convert/parakeet-tdt-ctc-110m/coreml/audio/yc_first_minute_16k_15s.wav b/convert/parakeet-tdt-ctc-110m/coreml/audio/yc_first_minute_16k_15s.wav deleted file mode 100644 index 0b8040f5f124d20a1d7e812c576c831a7573eddf..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/audio/yc_first_minute_16k_15s.wav +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:c79c8bc763b4efccb3e12f199ec0a59aa2edc5e9e4d21ca70fde8f36762d4147 -size 480078 diff --git a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/CTCHead.mlmodelc/analytics/coremldata.bin b/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/CTCHead.mlmodelc/analytics/coremldata.bin deleted file mode 100644 index 4093e2df453ec508a533d293a574198b2372845b..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/CTCHead.mlmodelc/analytics/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:fc681823d92eca3dbece3a30c975afa7251eedae0e718b07ffbf1a8b4313b87e -size 243 diff --git a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/CTCHead.mlmodelc/coremldata.bin b/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/CTCHead.mlmodelc/coremldata.bin deleted file mode 100644 index db4ae8de920867977480465fffece41bc606c0a2..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/CTCHead.mlmodelc/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:2ebec8fc38c063de4b2159e21b1f981309fa5947c24d7e4883aca20f7c15fbb9 -size 377 diff --git a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/CTCHead.mlmodelc/metadata.json b/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/CTCHead.mlmodelc/metadata.json deleted file mode 100644 index 4abe30bf0b35b72d2c9c013b61cc78b4f26a85e1..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/CTCHead.mlmodelc/metadata.json +++ /dev/null @@ -1,66 +0,0 @@ -[ - { - "metadataOutputVersion" : "3.0", - "shortDescription" : "Parakeet 110M CTC decoder head", - "outputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 188 × 1025)", - "shortDescription" : "", - "shape" : "[1, 188, 1025]", - "name" : "ctc_logits", - "type" : "MultiArray" - } - ], - "storagePrecision" : "Float16", - "modelParameters" : [ - - ], - "author" : "Fluid Inference", - "specificationVersion" : 8, - "mlProgramOperationTypeHistogram" : { - "Ios17.cast" : 2, - "Ios17.conv" : 1, - "Ios17.transpose" : 1, - "Ios16.softmax" : 1, - "Ios17.log" : 1 - }, - "computePrecision" : "Mixed (Float16, Float32, Int32)", - "isUpdatable" : "0", - "stateSchema" : [ - - ], - "availability" : { - "macOS" : "14.0", - "tvOS" : "17.0", - "visionOS" : "1.0", - "watchOS" : "10.0", - "iOS" : "17.0", - "macCatalyst" : "17.0" - }, - "modelType" : { - "name" : "MLModelType_mlProgram" - }, - "inputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 512 × 188)", - "shortDescription" : "", - "shape" : "[1, 512, 188]", - "name" : "encoder_output", - "type" : "MultiArray" - } - ], - "userDefinedMetadata" : { - "com.github.apple.coremltools.source_dialect" : "TorchScript", - "com.github.apple.coremltools.source" : "torch==2.9.0", - "com.github.apple.coremltools.version" : "8.3.0" - }, - "generatedClassName" : "parakeet_ctc_head", - "method" : "predict" - } -] \ No newline at end of file diff --git a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/CTCHead.mlmodelc/model.mil b/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/CTCHead.mlmodelc/model.mil deleted file mode 100644 index 67b3b5f87ce33caed71f9233828f90775bb8ac9d..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/CTCHead.mlmodelc/model.mil +++ /dev/null @@ -1,24 +0,0 @@ -program(1.0) -[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3510.2.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.9.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "8.3.0"}})] -{ - func main(tensor encoder_output) { - tensor var_4 = const()[name = tensor("op_4"), val = tensor(-1)]; - tensor var_18_pad_type_0 = const()[name = tensor("op_18_pad_type_0"), val = tensor("valid")]; - tensor var_18_strides_0 = const()[name = tensor("op_18_strides_0"), val = tensor([1])]; - tensor var_18_pad_0 = const()[name = tensor("op_18_pad_0"), val = tensor([0, 0])]; - tensor var_18_dilations_0 = const()[name = tensor("op_18_dilations_0"), val = tensor([1])]; - tensor var_18_groups_0 = const()[name = tensor("op_18_groups_0"), val = tensor(1)]; - tensor encoder_output_to_fp16_dtype_0 = const()[name = tensor("encoder_output_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor module_decoder_layers_0_weight_to_fp16 = const()[name = tensor("module_decoder_layers_0_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; - tensor module_decoder_layers_0_bias_to_fp16 = const()[name = tensor("module_decoder_layers_0_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1049728)))]; - tensor encoder_output_to_fp16 = cast(dtype = encoder_output_to_fp16_dtype_0, x = encoder_output)[name = tensor("cast_1")]; - tensor var_18_cast_fp16 = conv(bias = module_decoder_layers_0_bias_to_fp16, dilations = var_18_dilations_0, groups = var_18_groups_0, pad = var_18_pad_0, pad_type = var_18_pad_type_0, strides = var_18_strides_0, weight = module_decoder_layers_0_weight_to_fp16, x = encoder_output_to_fp16)[name = tensor("op_18_cast_fp16")]; - tensor input_perm_0 = const()[name = tensor("input_perm_0"), val = tensor([0, 2, 1])]; - tensor input_cast_fp16 = transpose(perm = input_perm_0, x = var_18_cast_fp16)[name = tensor("transpose_0")]; - tensor out_objects_softmax_cast_fp16 = softmax(axis = var_4, x = input_cast_fp16)[name = tensor("out_objects_softmax_cast_fp16")]; - tensor out_objects_epsilon_0 = const()[name = tensor("out_objects_epsilon_0"), val = tensor(0x1p-149)]; - tensor out_objects_cast_fp16 = log(epsilon = out_objects_epsilon_0, x = out_objects_softmax_cast_fp16)[name = tensor("out_objects_cast_fp16")]; - tensor out_objects_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("out_objects_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; - tensor ctc_logits = cast(dtype = out_objects_cast_fp16_to_fp32_dtype_0, x = out_objects_cast_fp16)[name = tensor("cast_0")]; - } -> (ctc_logits); -} \ No newline at end of file diff --git a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/CTCHead.mlmodelc/weights/weight.bin b/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/CTCHead.mlmodelc/weights/weight.bin deleted file mode 100644 index 23cedd570125191064c6111997f08c48898245f7..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/CTCHead.mlmodelc/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:fb9bead064427ffcb7529c0e3f378e421b4dde8e6d81447b6d1ca3352ca850e1 -size 1051842 diff --git a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/Decoder.mlmodelc/analytics/coremldata.bin b/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/Decoder.mlmodelc/analytics/coremldata.bin deleted file mode 100644 index dd015548a03cc453e6477a24374e8fee955e75ef..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/Decoder.mlmodelc/analytics/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:990455f6431342750254f66edf27bfb41be62a7ba17a18e1dd6afd4f5f56e9eb -size 243 diff --git a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/Decoder.mlmodelc/coremldata.bin b/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/Decoder.mlmodelc/coremldata.bin deleted file mode 100644 index cbea0affef0c3bb63be750e134dc0c9ed70aadb1..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/Decoder.mlmodelc/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:29009727821ad8551ab5fe9271e93c597d92a9714f64b94aa533a9ceb6e22b93 -size 498 diff --git a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/Decoder.mlmodelc/metadata.json b/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/Decoder.mlmodelc/metadata.json deleted file mode 100644 index 786f4bb2823b1aa7ebb55fe335c9385ee0065b7c..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/Decoder.mlmodelc/metadata.json +++ /dev/null @@ -1,118 +0,0 @@ -[ - { - "metadataOutputVersion" : "3.0", - "shortDescription" : "Parakeet 110M decoder (RNNT prediction network)", - "outputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 640 × 1)", - "shortDescription" : "", - "shape" : "[1, 640, 1]", - "name" : "decoder", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 1 × 640)", - "shortDescription" : "", - "shape" : "[1, 1, 640]", - "name" : "h_out", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 1 × 640)", - "shortDescription" : "", - "shape" : "[1, 1, 640]", - "name" : "c_out", - "type" : "MultiArray" - } - ], - "storagePrecision" : "Float16", - "modelParameters" : [ - - ], - "author" : "Fluid Inference", - "specificationVersion" : 8, - "mlProgramOperationTypeHistogram" : { - "Ios17.squeeze" : 2, - "Ios17.gather" : 1, - "Ios17.cast" : 6, - "Ios17.lstm" : 1, - "Ios17.transpose" : 2, - "Identity" : 1, - "Ios17.expandDims" : 2 - }, - "computePrecision" : "Mixed (Float16, Float32, Int16, Int32)", - "isUpdatable" : "0", - "stateSchema" : [ - - ], - "availability" : { - "macOS" : "14.0", - "tvOS" : "17.0", - "visionOS" : "1.0", - "watchOS" : "10.0", - "iOS" : "17.0", - "macCatalyst" : "17.0" - }, - "modelType" : { - "name" : "MLModelType_mlProgram" - }, - "inputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1 × 1)", - "shortDescription" : "", - "shape" : "[1, 1]", - "name" : "targets", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1)", - "shortDescription" : "", - "shape" : "[1]", - "name" : "target_length", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 1 × 640)", - "shortDescription" : "", - "shape" : "[1, 1, 640]", - "name" : "h_in", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 1 × 640)", - "shortDescription" : "", - "shape" : "[1, 1, 640]", - "name" : "c_in", - "type" : "MultiArray" - } - ], - "userDefinedMetadata" : { - "com.github.apple.coremltools.version" : "8.3.0", - "com.github.apple.coremltools.source_dialect" : "TorchScript", - "com.github.apple.coremltools.source" : "torch==2.9.0" - }, - "generatedClassName" : "parakeet_decoder", - "method" : "predict" - } -] \ No newline at end of file diff --git a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/Decoder.mlmodelc/model.mil b/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/Decoder.mlmodelc/model.mil deleted file mode 100644 index f69c7247525bc9d95f9341acc4f11eca709c7d00..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/Decoder.mlmodelc/model.mil +++ /dev/null @@ -1,45 +0,0 @@ -program(1.0) -[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3510.2.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.9.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "8.3.0"}})] -{ - func main(tensor c_in, tensor h_in, tensor target_length, tensor targets) { - tensor y_axis_0 = const()[name = tensor("y_axis_0"), val = tensor(0)]; - tensor y_batch_dims_0 = const()[name = tensor("y_batch_dims_0"), val = tensor(0)]; - tensor y_validate_indices_0 = const()[name = tensor("y_validate_indices_0"), val = tensor(false)]; - tensor module_prediction_embed_weight_to_fp16 = const()[name = tensor("module_prediction_embed_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; - tensor targets_to_int16_dtype_0 = const()[name = tensor("targets_to_int16_dtype_0"), val = tensor("int16")]; - tensor targets_to_int16 = cast(dtype = targets_to_int16_dtype_0, x = targets)[name = tensor("cast_8")]; - tensor y_cast_fp16_cast_uint16 = gather(axis = y_axis_0, batch_dims = y_batch_dims_0, indices = targets_to_int16, validate_indices = y_validate_indices_0, x = module_prediction_embed_weight_to_fp16)[name = tensor("y_cast_fp16_cast_uint16")]; - tensor input_3_perm_0 = const()[name = tensor("input_3_perm_0"), val = tensor([1, 0, 2])]; - tensor input_lstm_h0_squeeze_axes_0 = const()[name = tensor("input_lstm_h0_squeeze_axes_0"), val = tensor([0])]; - tensor h_in_to_fp16_dtype_0 = const()[name = tensor("h_in_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor h_in_to_fp16 = cast(dtype = h_in_to_fp16_dtype_0, x = h_in)[name = tensor("cast_7")]; - tensor input_lstm_h0_squeeze_cast_fp16 = squeeze(axes = input_lstm_h0_squeeze_axes_0, x = h_in_to_fp16)[name = tensor("input_lstm_h0_squeeze_cast_fp16")]; - tensor input_lstm_c0_squeeze_axes_0 = const()[name = tensor("input_lstm_c0_squeeze_axes_0"), val = tensor([0])]; - tensor c_in_to_fp16_dtype_0 = const()[name = tensor("c_in_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor c_in_to_fp16 = cast(dtype = c_in_to_fp16_dtype_0, x = c_in)[name = tensor("cast_6")]; - tensor input_lstm_c0_squeeze_cast_fp16 = squeeze(axes = input_lstm_c0_squeeze_axes_0, x = c_in_to_fp16)[name = tensor("input_lstm_c0_squeeze_cast_fp16")]; - tensor input_direction_0 = const()[name = tensor("input_direction_0"), val = tensor("forward")]; - tensor input_output_sequence_0 = const()[name = tensor("input_output_sequence_0"), val = tensor(true)]; - tensor input_recurrent_activation_0 = const()[name = tensor("input_recurrent_activation_0"), val = tensor("sigmoid")]; - tensor input_cell_activation_0 = const()[name = tensor("input_cell_activation_0"), val = tensor("tanh")]; - tensor input_activation_0 = const()[name = tensor("input_activation_0"), val = tensor("tanh")]; - tensor concat_1_to_fp16 = const()[name = tensor("concat_1_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1312128)))]; - tensor concat_2_to_fp16 = const()[name = tensor("concat_2_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4588992)))]; - tensor concat_0_to_fp16 = const()[name = tensor("concat_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7865856)))]; - tensor input_3_cast_fp16 = transpose(perm = input_3_perm_0, x = y_cast_fp16_cast_uint16)[name = tensor("transpose_2")]; - tensor input_cast_fp16_0, tensor input_cast_fp16_1, tensor input_cast_fp16_2 = lstm(activation = input_activation_0, bias = concat_0_to_fp16, cell_activation = input_cell_activation_0, direction = input_direction_0, initial_c = input_lstm_c0_squeeze_cast_fp16, initial_h = input_lstm_h0_squeeze_cast_fp16, output_sequence = input_output_sequence_0, recurrent_activation = input_recurrent_activation_0, weight_hh = concat_2_to_fp16, weight_ih = concat_1_to_fp16, x = input_3_cast_fp16)[name = tensor("input_cast_fp16")]; - tensor obj_3_axes_0 = const()[name = tensor("obj_3_axes_0"), val = tensor([0])]; - tensor obj_3_cast_fp16 = expand_dims(axes = obj_3_axes_0, x = input_cast_fp16_1)[name = tensor("obj_3_cast_fp16")]; - tensor obj_3_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("obj_3_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; - tensor obj_axes_0 = const()[name = tensor("obj_axes_0"), val = tensor([0])]; - tensor obj_cast_fp16 = expand_dims(axes = obj_axes_0, x = input_cast_fp16_2)[name = tensor("obj_cast_fp16")]; - tensor obj_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("obj_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; - tensor transpose_0_perm_0 = const()[name = tensor("transpose_0_perm_0"), val = tensor([1, 2, 0])]; - tensor transpose_0_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("transpose_0_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; - tensor transpose_0_cast_fp16 = transpose(perm = transpose_0_perm_0, x = input_cast_fp16_0)[name = tensor("transpose_1")]; - tensor decoder = cast(dtype = transpose_0_cast_fp16_to_fp32_dtype_0, x = transpose_0_cast_fp16)[name = tensor("cast_3")]; - tensor c_out = cast(dtype = obj_cast_fp16_to_fp32_dtype_0, x = obj_cast_fp16)[name = tensor("cast_4")]; - tensor h_out = cast(dtype = obj_3_cast_fp16_to_fp32_dtype_0, x = obj_3_cast_fp16)[name = tensor("cast_5")]; - tensor target_length_tmp = identity(x = target_length)[name = tensor("target_length_tmp")]; - } -> (decoder, h_out, c_out); -} \ No newline at end of file diff --git a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/Decoder.mlmodelc/weights/weight.bin b/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/Decoder.mlmodelc/weights/weight.bin deleted file mode 100644 index 4543ad4ea89b21ebed0de0d5ec3975d07eace7fb..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/Decoder.mlmodelc/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:dd90b58597ee2c172c672dffe13b1110898ba07394c1a15efc96cc8c6b18411b -size 7871040 diff --git a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/Encoder.mlmodelc/analytics/coremldata.bin b/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/Encoder.mlmodelc/analytics/coremldata.bin deleted file mode 100644 index d2787d4e5376d7f894d2b0a5a09f497a9b486c73..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/Encoder.mlmodelc/analytics/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:b7ae65e2af616df46066b7efca2d7c19941666ac0685f4ed005666890a052b0d -size 243 diff --git a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/Encoder.mlmodelc/coremldata.bin b/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/Encoder.mlmodelc/coremldata.bin deleted file mode 100644 index 9d9cdbfd392e8f7391b47c1c96f0fcbaab440389..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/Encoder.mlmodelc/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:0713c2d6ac5f8f6fb9582be250351ebd8efc925f71f4261191165f1406f2ee5d -size 437 diff --git a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/Encoder.mlmodelc/metadata.json b/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/Encoder.mlmodelc/metadata.json deleted file mode 100644 index 8e3f6cc9ddca206205987b2c828d2e595e19bfc2..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/Encoder.mlmodelc/metadata.json +++ /dev/null @@ -1,105 +0,0 @@ -[ - { - "metadataOutputVersion" : "3.0", - "shortDescription" : "Parakeet 110M encoder (15 s window)", - "outputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 512 × 188)", - "shortDescription" : "", - "shape" : "[1, 512, 188]", - "name" : "encoder_output", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1)", - "shortDescription" : "", - "shape" : "[1]", - "name" : "encoder_length", - "type" : "MultiArray" - } - ], - "storagePrecision" : "Float16", - "modelParameters" : [ - - ], - "author" : "Fluid Inference", - "specificationVersion" : 8, - "mlProgramOperationTypeHistogram" : { - "Ios17.logicalAnd" : 2, - "Ios17.reshape" : 103, - "Ios16.softmax" : 17, - "Ios17.matmul" : 51, - "Ios17.transpose" : 123, - "Split" : 17, - "Ios17.expandDims" : 17, - "Select" : 51, - "Ios17.add" : 128, - "Tile" : 8, - "Ios17.sliceByIndex" : 34, - "Ios16.sigmoid" : 17, - "Pad" : 34, - "Ios17.logicalNot" : 2, - "Ios17.layerNorm" : 85, - "Ios16.silu" : 51, - "Ios17.less" : 5, - "Ios17.sub" : 3, - "Ios17.conv" : 56, - "Ios16.relu" : 3, - "Ios17.linear" : 137, - "Ios17.cast" : 11, - "Ios17.floorDiv" : 3, - "Ios17.mul" : 77 - }, - "computePrecision" : "Mixed (Float16, Float32, Int32)", - "isUpdatable" : "0", - "stateSchema" : [ - - ], - "availability" : { - "macOS" : "14.0", - "tvOS" : "17.0", - "visionOS" : "1.0", - "watchOS" : "10.0", - "iOS" : "17.0", - "macCatalyst" : "17.0" - }, - "modelType" : { - "name" : "MLModelType_mlProgram" - }, - "inputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 80 × 1501)", - "shortDescription" : "", - "shape" : "[1, 80, 1501]", - "name" : "mel_features", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1)", - "shortDescription" : "", - "shape" : "[1]", - "name" : "mel_length", - "type" : "MultiArray" - } - ], - "userDefinedMetadata" : { - "com.github.apple.coremltools.source_dialect" : "TorchScript", - "com.github.apple.coremltools.source" : "torch==2.9.0", - "com.github.apple.coremltools.version" : "8.3.0" - }, - "generatedClassName" : "parakeet_encoder", - "method" : "predict" - } -] \ No newline at end of file diff --git a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/Encoder.mlmodelc/model.mil b/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/Encoder.mlmodelc/model.mil deleted file mode 100644 index 344bd1abf313c03c9ed9287c00893739f5d361bd..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/Encoder.mlmodelc/model.mil +++ /dev/null @@ -1,2690 +0,0 @@ -program(1.0) -[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3510.2.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.9.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "8.3.0"}})] -{ - func main(tensor mel_features, tensor mel_length) { - tensor var_23 = const()[name = tensor("op_23"), val = tensor(-1)]; - tensor x_1_perm_0 = const()[name = tensor("x_1_perm_0"), val = tensor([0, 2, 1])]; - tensor mel_features_to_fp16_dtype_0 = const()[name = tensor("mel_features_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor tensor_1_axes_0 = const()[name = tensor("tensor_1_axes_0"), val = tensor([1])]; - tensor mel_features_to_fp16 = cast(dtype = mel_features_to_fp16_dtype_0, x = mel_features)[name = tensor("cast_182")]; - tensor x_1_cast_fp16 = transpose(perm = x_1_perm_0, x = mel_features_to_fp16)[name = tensor("transpose_207")]; - tensor tensor_1_cast_fp16 = expand_dims(axes = tensor_1_axes_0, x = x_1_cast_fp16)[name = tensor("tensor_1_cast_fp16")]; - tensor expand_dims_0 = const()[name = tensor("expand_dims_0"), val = tensor([[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390, 391, 392, 393, 394, 395, 396, 397, 398, 399, 400, 401, 402, 403, 404, 405, 406, 407, 408, 409, 410, 411, 412, 413, 414, 415, 416, 417, 418, 419, 420, 421, 422, 423, 424, 425, 426, 427, 428, 429, 430, 431, 432, 433, 434, 435, 436, 437, 438, 439, 440, 441, 442, 443, 444, 445, 446, 447, 448, 449, 450, 451, 452, 453, 454, 455, 456, 457, 458, 459, 460, 461, 462, 463, 464, 465, 466, 467, 468, 469, 470, 471, 472, 473, 474, 475, 476, 477, 478, 479, 480, 481, 482, 483, 484, 485, 486, 487, 488, 489, 490, 491, 492, 493, 494, 495, 496, 497, 498, 499, 500, 501, 502, 503, 504, 505, 506, 507, 508, 509, 510, 511, 512, 513, 514, 515, 516, 517, 518, 519, 520, 521, 522, 523, 524, 525, 526, 527, 528, 529, 530, 531, 532, 533, 534, 535, 536, 537, 538, 539, 540, 541, 542, 543, 544, 545, 546, 547, 548, 549, 550, 551, 552, 553, 554, 555, 556, 557, 558, 559, 560, 561, 562, 563, 564, 565, 566, 567, 568, 569, 570, 571, 572, 573, 574, 575, 576, 577, 578, 579, 580, 581, 582, 583, 584, 585, 586, 587, 588, 589, 590, 591, 592, 593, 594, 595, 596, 597, 598, 599, 600, 601, 602, 603, 604, 605, 606, 607, 608, 609, 610, 611, 612, 613, 614, 615, 616, 617, 618, 619, 620, 621, 622, 623, 624, 625, 626, 627, 628, 629, 630, 631, 632, 633, 634, 635, 636, 637, 638, 639, 640, 641, 642, 643, 644, 645, 646, 647, 648, 649, 650, 651, 652, 653, 654, 655, 656, 657, 658, 659, 660, 661, 662, 663, 664, 665, 666, 667, 668, 669, 670, 671, 672, 673, 674, 675, 676, 677, 678, 679, 680, 681, 682, 683, 684, 685, 686, 687, 688, 689, 690, 691, 692, 693, 694, 695, 696, 697, 698, 699, 700, 701, 702, 703, 704, 705, 706, 707, 708, 709, 710, 711, 712, 713, 714, 715, 716, 717, 718, 719, 720, 721, 722, 723, 724, 725, 726, 727, 728, 729, 730, 731, 732, 733, 734, 735, 736, 737, 738, 739, 740, 741, 742, 743, 744, 745, 746, 747, 748, 749, 750, 751, 752, 753, 754, 755, 756, 757, 758, 759, 760, 761, 762, 763, 764, 765, 766, 767, 768, 769, 770, 771, 772, 773, 774, 775, 776, 777, 778, 779, 780, 781, 782, 783, 784, 785, 786, 787, 788, 789, 790, 791, 792, 793, 794, 795, 796, 797, 798, 799, 800, 801, 802, 803, 804, 805, 806, 807, 808, 809, 810, 811, 812, 813, 814, 815, 816, 817, 818, 819, 820, 821, 822, 823, 824, 825, 826, 827, 828, 829, 830, 831, 832, 833, 834, 835, 836, 837, 838, 839, 840, 841, 842, 843, 844, 845, 846, 847, 848, 849, 850, 851, 852, 853, 854, 855, 856, 857, 858, 859, 860, 861, 862, 863, 864, 865, 866, 867, 868, 869, 870, 871, 872, 873, 874, 875, 876, 877, 878, 879, 880, 881, 882, 883, 884, 885, 886, 887, 888, 889, 890, 891, 892, 893, 894, 895, 896, 897, 898, 899, 900, 901, 902, 903, 904, 905, 906, 907, 908, 909, 910, 911, 912, 913, 914, 915, 916, 917, 918, 919, 920, 921, 922, 923, 924, 925, 926, 927, 928, 929, 930, 931, 932, 933, 934, 935, 936, 937, 938, 939, 940, 941, 942, 943, 944, 945, 946, 947, 948, 949, 950, 951, 952, 953, 954, 955, 956, 957, 958, 959, 960, 961, 962, 963, 964, 965, 966, 967, 968, 969, 970, 971, 972, 973, 974, 975, 976, 977, 978, 979, 980, 981, 982, 983, 984, 985, 986, 987, 988, 989, 990, 991, 992, 993, 994, 995, 996, 997, 998, 999, 1000, 1001, 1002, 1003, 1004, 1005, 1006, 1007, 1008, 1009, 1010, 1011, 1012, 1013, 1014, 1015, 1016, 1017, 1018, 1019, 1020, 1021, 1022, 1023, 1024, 1025, 1026, 1027, 1028, 1029, 1030, 1031, 1032, 1033, 1034, 1035, 1036, 1037, 1038, 1039, 1040, 1041, 1042, 1043, 1044, 1045, 1046, 1047, 1048, 1049, 1050, 1051, 1052, 1053, 1054, 1055, 1056, 1057, 1058, 1059, 1060, 1061, 1062, 1063, 1064, 1065, 1066, 1067, 1068, 1069, 1070, 1071, 1072, 1073, 1074, 1075, 1076, 1077, 1078, 1079, 1080, 1081, 1082, 1083, 1084, 1085, 1086, 1087, 1088, 1089, 1090, 1091, 1092, 1093, 1094, 1095, 1096, 1097, 1098, 1099, 1100, 1101, 1102, 1103, 1104, 1105, 1106, 1107, 1108, 1109, 1110, 1111, 1112, 1113, 1114, 1115, 1116, 1117, 1118, 1119, 1120, 1121, 1122, 1123, 1124, 1125, 1126, 1127, 1128, 1129, 1130, 1131, 1132, 1133, 1134, 1135, 1136, 1137, 1138, 1139, 1140, 1141, 1142, 1143, 1144, 1145, 1146, 1147, 1148, 1149, 1150, 1151, 1152, 1153, 1154, 1155, 1156, 1157, 1158, 1159, 1160, 1161, 1162, 1163, 1164, 1165, 1166, 1167, 1168, 1169, 1170, 1171, 1172, 1173, 1174, 1175, 1176, 1177, 1178, 1179, 1180, 1181, 1182, 1183, 1184, 1185, 1186, 1187, 1188, 1189, 1190, 1191, 1192, 1193, 1194, 1195, 1196, 1197, 1198, 1199, 1200, 1201, 1202, 1203, 1204, 1205, 1206, 1207, 1208, 1209, 1210, 1211, 1212, 1213, 1214, 1215, 1216, 1217, 1218, 1219, 1220, 1221, 1222, 1223, 1224, 1225, 1226, 1227, 1228, 1229, 1230, 1231, 1232, 1233, 1234, 1235, 1236, 1237, 1238, 1239, 1240, 1241, 1242, 1243, 1244, 1245, 1246, 1247, 1248, 1249, 1250, 1251, 1252, 1253, 1254, 1255, 1256, 1257, 1258, 1259, 1260, 1261, 1262, 1263, 1264, 1265, 1266, 1267, 1268, 1269, 1270, 1271, 1272, 1273, 1274, 1275, 1276, 1277, 1278, 1279, 1280, 1281, 1282, 1283, 1284, 1285, 1286, 1287, 1288, 1289, 1290, 1291, 1292, 1293, 1294, 1295, 1296, 1297, 1298, 1299, 1300, 1301, 1302, 1303, 1304, 1305, 1306, 1307, 1308, 1309, 1310, 1311, 1312, 1313, 1314, 1315, 1316, 1317, 1318, 1319, 1320, 1321, 1322, 1323, 1324, 1325, 1326, 1327, 1328, 1329, 1330, 1331, 1332, 1333, 1334, 1335, 1336, 1337, 1338, 1339, 1340, 1341, 1342, 1343, 1344, 1345, 1346, 1347, 1348, 1349, 1350, 1351, 1352, 1353, 1354, 1355, 1356, 1357, 1358, 1359, 1360, 1361, 1362, 1363, 1364, 1365, 1366, 1367, 1368, 1369, 1370, 1371, 1372, 1373, 1374, 1375, 1376, 1377, 1378, 1379, 1380, 1381, 1382, 1383, 1384, 1385, 1386, 1387, 1388, 1389, 1390, 1391, 1392, 1393, 1394, 1395, 1396, 1397, 1398, 1399, 1400, 1401, 1402, 1403, 1404, 1405, 1406, 1407, 1408, 1409, 1410, 1411, 1412, 1413, 1414, 1415, 1416, 1417, 1418, 1419, 1420, 1421, 1422, 1423, 1424, 1425, 1426, 1427, 1428, 1429, 1430, 1431, 1432, 1433, 1434, 1435, 1436, 1437, 1438, 1439, 1440, 1441, 1442, 1443, 1444, 1445, 1446, 1447, 1448, 1449, 1450, 1451, 1452, 1453, 1454, 1455, 1456, 1457, 1458, 1459, 1460, 1461, 1462, 1463, 1464, 1465, 1466, 1467, 1468, 1469, 1470, 1471, 1472, 1473, 1474, 1475, 1476, 1477, 1478, 1479, 1480, 1481, 1482, 1483, 1484, 1485, 1486, 1487, 1488, 1489, 1490, 1491, 1492, 1493, 1494, 1495, 1496, 1497, 1498, 1499, 1500]])]; - tensor var_87_axes_0 = const()[name = tensor("op_87_axes_0"), val = tensor([1])]; - tensor var_87 = expand_dims(axes = var_87_axes_0, x = mel_length)[name = tensor("op_87")]; - tensor time_mask_1 = less(x = expand_dims_0, y = var_87)[name = tensor("time_mask_1")]; - tensor var_89_axes_0 = const()[name = tensor("op_89_axes_0"), val = tensor([-1])]; - tensor var_89 = expand_dims(axes = var_89_axes_0, x = time_mask_1)[name = tensor("op_89")]; - tensor var_91_reps_0 = const()[name = tensor("op_91_reps_0"), val = tensor([1, 1, 80])]; - tensor var_91 = tile(reps = var_91_reps_0, x = var_89)[name = tensor("op_91")]; - tensor var_97_axes_0 = const()[name = tensor("op_97_axes_0"), val = tensor([1])]; - tensor cast_3_to_fp16_dtype_0 = const()[name = tensor("cast_3_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor var_91_to_fp16 = cast(dtype = cast_3_to_fp16_dtype_0, x = var_91)[name = tensor("cast_181")]; - tensor var_97_cast_fp16 = expand_dims(axes = var_97_axes_0, x = var_91_to_fp16)[name = tensor("op_97_cast_fp16")]; - tensor input_1_cast_fp16 = mul(x = tensor_1_cast_fp16, y = var_97_cast_fp16)[name = tensor("input_1_cast_fp16")]; - tensor tensor_3_pad_type_0 = const()[name = tensor("tensor_3_pad_type_0"), val = tensor("custom")]; - tensor tensor_3_pad_0 = const()[name = tensor("tensor_3_pad_0"), val = tensor([1, 1, 1, 1])]; - tensor tensor_3_strides_0 = const()[name = tensor("tensor_3_strides_0"), val = tensor([2, 2])]; - tensor tensor_3_dilations_0 = const()[name = tensor("tensor_3_dilations_0"), val = tensor([1, 1])]; - tensor tensor_3_groups_0 = const()[name = tensor("tensor_3_groups_0"), val = tensor(1)]; - tensor module_pre_encode_conv_0_weight_to_fp16 = const()[name = tensor("module_pre_encode_conv_0_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; - tensor module_pre_encode_conv_0_bias_to_fp16 = const()[name = tensor("module_pre_encode_conv_0_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4736)))]; - tensor tensor_3_cast_fp16 = conv(bias = module_pre_encode_conv_0_bias_to_fp16, dilations = tensor_3_dilations_0, groups = tensor_3_groups_0, pad = tensor_3_pad_0, pad_type = tensor_3_pad_type_0, strides = tensor_3_strides_0, weight = module_pre_encode_conv_0_weight_to_fp16, x = input_1_cast_fp16)[name = tensor("tensor_3_cast_fp16")]; - tensor cast_1_to_fp16_dtype_0 = const()[name = tensor("cast_1_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor var_108_promoted_to_fp16 = const()[name = tensor("op_108_promoted_to_fp16"), val = tensor(0x1p+0)]; - tensor mel_length_to_fp16 = cast(dtype = cast_1_to_fp16_dtype_0, x = mel_length)[name = tensor("cast_180")]; - tensor var_109_cast_fp16 = add(x = mel_length_to_fp16, y = var_108_promoted_to_fp16)[name = tensor("op_109_cast_fp16")]; - tensor var_110_promoted_to_fp16 = const()[name = tensor("op_110_promoted_to_fp16"), val = tensor(0x1p+0)]; - tensor var_111_cast_fp16 = add(x = var_109_cast_fp16, y = var_110_promoted_to_fp16)[name = tensor("op_111_cast_fp16")]; - tensor var_112_promoted_to_fp16 = const()[name = tensor("op_112_promoted_to_fp16"), val = tensor(0x1.8p+1)]; - tensor var_113_cast_fp16 = sub(x = var_111_cast_fp16, y = var_112_promoted_to_fp16)[name = tensor("op_113_cast_fp16")]; - tensor var_21_promoted_to_fp16 = const()[name = tensor("op_21_promoted_to_fp16"), val = tensor(0x1p+1)]; - tensor floor_div_0_cast_fp16 = floor_div(x = var_113_cast_fp16, y = var_21_promoted_to_fp16)[name = tensor("floor_div_0_cast_fp16")]; - tensor var_115_promoted_to_fp16 = const()[name = tensor("op_115_promoted_to_fp16"), val = tensor(0x1p+0)]; - tensor current_lengths_3_cast_fp16 = add(x = floor_div_0_cast_fp16, y = var_115_promoted_to_fp16)[name = tensor("current_lengths_3_cast_fp16")]; - tensor cast_4_dtype_0 = const()[name = tensor("cast_4_dtype_0"), val = tensor("int32")]; - tensor expand_dims_1 = const()[name = tensor("expand_dims_1"), val = tensor([[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390, 391, 392, 393, 394, 395, 396, 397, 398, 399, 400, 401, 402, 403, 404, 405, 406, 407, 408, 409, 410, 411, 412, 413, 414, 415, 416, 417, 418, 419, 420, 421, 422, 423, 424, 425, 426, 427, 428, 429, 430, 431, 432, 433, 434, 435, 436, 437, 438, 439, 440, 441, 442, 443, 444, 445, 446, 447, 448, 449, 450, 451, 452, 453, 454, 455, 456, 457, 458, 459, 460, 461, 462, 463, 464, 465, 466, 467, 468, 469, 470, 471, 472, 473, 474, 475, 476, 477, 478, 479, 480, 481, 482, 483, 484, 485, 486, 487, 488, 489, 490, 491, 492, 493, 494, 495, 496, 497, 498, 499, 500, 501, 502, 503, 504, 505, 506, 507, 508, 509, 510, 511, 512, 513, 514, 515, 516, 517, 518, 519, 520, 521, 522, 523, 524, 525, 526, 527, 528, 529, 530, 531, 532, 533, 534, 535, 536, 537, 538, 539, 540, 541, 542, 543, 544, 545, 546, 547, 548, 549, 550, 551, 552, 553, 554, 555, 556, 557, 558, 559, 560, 561, 562, 563, 564, 565, 566, 567, 568, 569, 570, 571, 572, 573, 574, 575, 576, 577, 578, 579, 580, 581, 582, 583, 584, 585, 586, 587, 588, 589, 590, 591, 592, 593, 594, 595, 596, 597, 598, 599, 600, 601, 602, 603, 604, 605, 606, 607, 608, 609, 610, 611, 612, 613, 614, 615, 616, 617, 618, 619, 620, 621, 622, 623, 624, 625, 626, 627, 628, 629, 630, 631, 632, 633, 634, 635, 636, 637, 638, 639, 640, 641, 642, 643, 644, 645, 646, 647, 648, 649, 650, 651, 652, 653, 654, 655, 656, 657, 658, 659, 660, 661, 662, 663, 664, 665, 666, 667, 668, 669, 670, 671, 672, 673, 674, 675, 676, 677, 678, 679, 680, 681, 682, 683, 684, 685, 686, 687, 688, 689, 690, 691, 692, 693, 694, 695, 696, 697, 698, 699, 700, 701, 702, 703, 704, 705, 706, 707, 708, 709, 710, 711, 712, 713, 714, 715, 716, 717, 718, 719, 720, 721, 722, 723, 724, 725, 726, 727, 728, 729, 730, 731, 732, 733, 734, 735, 736, 737, 738, 739, 740, 741, 742, 743, 744, 745, 746, 747, 748, 749, 750]])]; - tensor var_124_axes_0 = const()[name = tensor("op_124_axes_0"), val = tensor([1])]; - tensor current_lengths_3_cast_fp16_to_int32 = cast(dtype = cast_4_dtype_0, x = current_lengths_3_cast_fp16)[name = tensor("cast_179")]; - tensor var_124 = expand_dims(axes = var_124_axes_0, x = current_lengths_3_cast_fp16_to_int32)[name = tensor("op_124")]; - tensor time_mask_3 = less(x = expand_dims_1, y = var_124)[name = tensor("time_mask_3")]; - tensor var_126_axes_0 = const()[name = tensor("op_126_axes_0"), val = tensor([-1])]; - tensor var_126 = expand_dims(axes = var_126_axes_0, x = time_mask_3)[name = tensor("op_126")]; - tensor var_128_reps_0 = const()[name = tensor("op_128_reps_0"), val = tensor([1, 1, 40])]; - tensor var_128 = tile(reps = var_128_reps_0, x = var_126)[name = tensor("op_128")]; - tensor var_134_axes_0 = const()[name = tensor("op_134_axes_0"), val = tensor([1])]; - tensor cast_5_to_fp16_dtype_0 = const()[name = tensor("cast_5_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor var_128_to_fp16 = cast(dtype = cast_5_to_fp16_dtype_0, x = var_128)[name = tensor("cast_178")]; - tensor var_134_cast_fp16 = expand_dims(axes = var_134_axes_0, x = var_128_to_fp16)[name = tensor("op_134_cast_fp16")]; - tensor expanded_mask_3_reps_0 = const()[name = tensor("expanded_mask_3_reps_0"), val = tensor([1, 256, 1, 1])]; - tensor expanded_mask_3_cast_fp16 = tile(reps = expanded_mask_3_reps_0, x = var_134_cast_fp16)[name = tensor("expanded_mask_3_cast_fp16")]; - tensor input_3_cast_fp16 = mul(x = tensor_3_cast_fp16, y = expanded_mask_3_cast_fp16)[name = tensor("input_3_cast_fp16")]; - tensor tensor_5_cast_fp16 = relu(x = input_3_cast_fp16)[name = tensor("tensor_5_cast_fp16")]; - tensor input_5_cast_fp16 = mul(x = tensor_5_cast_fp16, y = expanded_mask_3_cast_fp16)[name = tensor("input_5_cast_fp16")]; - tensor tensor_7_pad_type_0 = const()[name = tensor("tensor_7_pad_type_0"), val = tensor("custom")]; - tensor tensor_7_pad_0 = const()[name = tensor("tensor_7_pad_0"), val = tensor([1, 1, 1, 1])]; - tensor tensor_7_strides_0 = const()[name = tensor("tensor_7_strides_0"), val = tensor([2, 2])]; - tensor tensor_7_groups_0 = const()[name = tensor("tensor_7_groups_0"), val = tensor(256)]; - tensor tensor_7_dilations_0 = const()[name = tensor("tensor_7_dilations_0"), val = tensor([1, 1])]; - tensor module_pre_encode_conv_2_weight_to_fp16 = const()[name = tensor("module_pre_encode_conv_2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(5312)))]; - tensor module_pre_encode_conv_2_bias_to_fp16 = const()[name = tensor("module_pre_encode_conv_2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9984)))]; - tensor tensor_7_cast_fp16 = conv(bias = module_pre_encode_conv_2_bias_to_fp16, dilations = tensor_7_dilations_0, groups = tensor_7_groups_0, pad = tensor_7_pad_0, pad_type = tensor_7_pad_type_0, strides = tensor_7_strides_0, weight = module_pre_encode_conv_2_weight_to_fp16, x = input_5_cast_fp16)[name = tensor("tensor_7_cast_fp16")]; - tensor var_154_promoted_to_fp16 = const()[name = tensor("op_154_promoted_to_fp16"), val = tensor(0x1p+0)]; - tensor var_155_cast_fp16 = add(x = current_lengths_3_cast_fp16, y = var_154_promoted_to_fp16)[name = tensor("op_155_cast_fp16")]; - tensor var_156_promoted_to_fp16 = const()[name = tensor("op_156_promoted_to_fp16"), val = tensor(0x1p+0)]; - tensor var_157_cast_fp16 = add(x = var_155_cast_fp16, y = var_156_promoted_to_fp16)[name = tensor("op_157_cast_fp16")]; - tensor var_158_promoted_to_fp16 = const()[name = tensor("op_158_promoted_to_fp16"), val = tensor(0x1.8p+1)]; - tensor var_159_cast_fp16 = sub(x = var_157_cast_fp16, y = var_158_promoted_to_fp16)[name = tensor("op_159_cast_fp16")]; - tensor var_21_promoted_1_to_fp16 = const()[name = tensor("op_21_promoted_1_to_fp16"), val = tensor(0x1p+1)]; - tensor floor_div_1_cast_fp16 = floor_div(x = var_159_cast_fp16, y = var_21_promoted_1_to_fp16)[name = tensor("floor_div_1_cast_fp16")]; - tensor var_161_promoted_to_fp16 = const()[name = tensor("op_161_promoted_to_fp16"), val = tensor(0x1p+0)]; - tensor current_lengths_5_cast_fp16 = add(x = floor_div_1_cast_fp16, y = var_161_promoted_to_fp16)[name = tensor("current_lengths_5_cast_fp16")]; - tensor cast_6_dtype_0 = const()[name = tensor("cast_6_dtype_0"), val = tensor("int32")]; - tensor expand_dims_2 = const()[name = tensor("expand_dims_2"), val = tensor([[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375]])]; - tensor var_170_axes_0 = const()[name = tensor("op_170_axes_0"), val = tensor([1])]; - tensor current_lengths_5_cast_fp16_to_int32 = cast(dtype = cast_6_dtype_0, x = current_lengths_5_cast_fp16)[name = tensor("cast_177")]; - tensor var_170 = expand_dims(axes = var_170_axes_0, x = current_lengths_5_cast_fp16_to_int32)[name = tensor("op_170")]; - tensor time_mask_5 = less(x = expand_dims_2, y = var_170)[name = tensor("time_mask_5")]; - tensor var_172_axes_0 = const()[name = tensor("op_172_axes_0"), val = tensor([-1])]; - tensor var_172 = expand_dims(axes = var_172_axes_0, x = time_mask_5)[name = tensor("op_172")]; - tensor var_174_reps_0 = const()[name = tensor("op_174_reps_0"), val = tensor([1, 1, 20])]; - tensor var_174 = tile(reps = var_174_reps_0, x = var_172)[name = tensor("op_174")]; - tensor var_180_axes_0 = const()[name = tensor("op_180_axes_0"), val = tensor([1])]; - tensor cast_7_to_fp16_dtype_0 = const()[name = tensor("cast_7_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor var_174_to_fp16 = cast(dtype = cast_7_to_fp16_dtype_0, x = var_174)[name = tensor("cast_176")]; - tensor var_180_cast_fp16 = expand_dims(axes = var_180_axes_0, x = var_174_to_fp16)[name = tensor("op_180_cast_fp16")]; - tensor expanded_mask_7_reps_0 = const()[name = tensor("expanded_mask_7_reps_0"), val = tensor([1, 256, 1, 1])]; - tensor expanded_mask_7_cast_fp16 = tile(reps = expanded_mask_7_reps_0, x = var_180_cast_fp16)[name = tensor("expanded_mask_7_cast_fp16")]; - tensor input_7_cast_fp16 = mul(x = tensor_7_cast_fp16, y = expanded_mask_7_cast_fp16)[name = tensor("input_7_cast_fp16")]; - tensor tensor_9_pad_type_0 = const()[name = tensor("tensor_9_pad_type_0"), val = tensor("valid")]; - tensor tensor_9_strides_0 = const()[name = tensor("tensor_9_strides_0"), val = tensor([1, 1])]; - tensor tensor_9_pad_0 = const()[name = tensor("tensor_9_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor tensor_9_dilations_0 = const()[name = tensor("tensor_9_dilations_0"), val = tensor([1, 1])]; - tensor tensor_9_groups_0 = const()[name = tensor("tensor_9_groups_0"), val = tensor(1)]; - tensor module_pre_encode_conv_3_weight_to_fp16 = const()[name = tensor("module_pre_encode_conv_3_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10560)))]; - tensor module_pre_encode_conv_3_bias_to_fp16 = const()[name = tensor("module_pre_encode_conv_3_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141696)))]; - tensor tensor_9_cast_fp16 = conv(bias = module_pre_encode_conv_3_bias_to_fp16, dilations = tensor_9_dilations_0, groups = tensor_9_groups_0, pad = tensor_9_pad_0, pad_type = tensor_9_pad_type_0, strides = tensor_9_strides_0, weight = module_pre_encode_conv_3_weight_to_fp16, x = input_7_cast_fp16)[name = tensor("tensor_9_cast_fp16")]; - tensor input_9_cast_fp16 = mul(x = tensor_9_cast_fp16, y = expanded_mask_7_cast_fp16)[name = tensor("input_9_cast_fp16")]; - tensor tensor_11_cast_fp16 = relu(x = input_9_cast_fp16)[name = tensor("tensor_11_cast_fp16")]; - tensor input_11_cast_fp16 = mul(x = tensor_11_cast_fp16, y = expanded_mask_7_cast_fp16)[name = tensor("input_11_cast_fp16")]; - tensor tensor_13_pad_type_0 = const()[name = tensor("tensor_13_pad_type_0"), val = tensor("custom")]; - tensor tensor_13_pad_0 = const()[name = tensor("tensor_13_pad_0"), val = tensor([1, 1, 1, 1])]; - tensor tensor_13_strides_0 = const()[name = tensor("tensor_13_strides_0"), val = tensor([2, 2])]; - tensor tensor_13_groups_0 = const()[name = tensor("tensor_13_groups_0"), val = tensor(256)]; - tensor tensor_13_dilations_0 = const()[name = tensor("tensor_13_dilations_0"), val = tensor([1, 1])]; - tensor module_pre_encode_conv_5_weight_to_fp16 = const()[name = tensor("module_pre_encode_conv_5_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(142272)))]; - tensor module_pre_encode_conv_5_bias_to_fp16 = const()[name = tensor("module_pre_encode_conv_5_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146944)))]; - tensor tensor_13_cast_fp16 = conv(bias = module_pre_encode_conv_5_bias_to_fp16, dilations = tensor_13_dilations_0, groups = tensor_13_groups_0, pad = tensor_13_pad_0, pad_type = tensor_13_pad_type_0, strides = tensor_13_strides_0, weight = module_pre_encode_conv_5_weight_to_fp16, x = input_11_cast_fp16)[name = tensor("tensor_13_cast_fp16")]; - tensor var_215_promoted_to_fp16 = const()[name = tensor("op_215_promoted_to_fp16"), val = tensor(0x1p+0)]; - tensor var_216_cast_fp16 = add(x = current_lengths_5_cast_fp16, y = var_215_promoted_to_fp16)[name = tensor("op_216_cast_fp16")]; - tensor var_217_promoted_to_fp16 = const()[name = tensor("op_217_promoted_to_fp16"), val = tensor(0x1p+0)]; - tensor var_218_cast_fp16 = add(x = var_216_cast_fp16, y = var_217_promoted_to_fp16)[name = tensor("op_218_cast_fp16")]; - tensor var_219_promoted_to_fp16 = const()[name = tensor("op_219_promoted_to_fp16"), val = tensor(0x1.8p+1)]; - tensor var_220_cast_fp16 = sub(x = var_218_cast_fp16, y = var_219_promoted_to_fp16)[name = tensor("op_220_cast_fp16")]; - tensor var_21_promoted_2_to_fp16 = const()[name = tensor("op_21_promoted_2_to_fp16"), val = tensor(0x1p+1)]; - tensor floor_div_2_cast_fp16 = floor_div(x = var_220_cast_fp16, y = var_21_promoted_2_to_fp16)[name = tensor("floor_div_2_cast_fp16")]; - tensor var_222_promoted_to_fp16 = const()[name = tensor("op_222_promoted_to_fp16"), val = tensor(0x1p+0)]; - tensor current_lengths_cast_fp16 = add(x = floor_div_2_cast_fp16, y = var_222_promoted_to_fp16)[name = tensor("current_lengths_cast_fp16")]; - tensor cast_8_dtype_0 = const()[name = tensor("cast_8_dtype_0"), val = tensor("int32")]; - tensor expand_dims_3 = const()[name = tensor("expand_dims_3"), val = tensor([[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187]])]; - tensor var_231_axes_0 = const()[name = tensor("op_231_axes_0"), val = tensor([1])]; - tensor current_lengths_cast_fp16_to_int32 = cast(dtype = cast_8_dtype_0, x = current_lengths_cast_fp16)[name = tensor("cast_175")]; - tensor var_231 = expand_dims(axes = var_231_axes_0, x = current_lengths_cast_fp16_to_int32)[name = tensor("op_231")]; - tensor time_mask = less(x = expand_dims_3, y = var_231)[name = tensor("time_mask")]; - tensor var_233_axes_0 = const()[name = tensor("op_233_axes_0"), val = tensor([-1])]; - tensor var_233 = expand_dims(axes = var_233_axes_0, x = time_mask)[name = tensor("op_233")]; - tensor var_235_reps_0 = const()[name = tensor("op_235_reps_0"), val = tensor([1, 1, 10])]; - tensor var_235 = tile(reps = var_235_reps_0, x = var_233)[name = tensor("op_235")]; - tensor var_241_axes_0 = const()[name = tensor("op_241_axes_0"), val = tensor([1])]; - tensor cast_9_to_fp16_dtype_0 = const()[name = tensor("cast_9_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor var_235_to_fp16 = cast(dtype = cast_9_to_fp16_dtype_0, x = var_235)[name = tensor("cast_174")]; - tensor var_241_cast_fp16 = expand_dims(axes = var_241_axes_0, x = var_235_to_fp16)[name = tensor("op_241_cast_fp16")]; - tensor expanded_mask_13_reps_0 = const()[name = tensor("expanded_mask_13_reps_0"), val = tensor([1, 256, 1, 1])]; - tensor expanded_mask_13_cast_fp16 = tile(reps = expanded_mask_13_reps_0, x = var_241_cast_fp16)[name = tensor("expanded_mask_13_cast_fp16")]; - tensor input_13_cast_fp16 = mul(x = tensor_13_cast_fp16, y = expanded_mask_13_cast_fp16)[name = tensor("input_13_cast_fp16")]; - tensor tensor_15_pad_type_0 = const()[name = tensor("tensor_15_pad_type_0"), val = tensor("valid")]; - tensor tensor_15_strides_0 = const()[name = tensor("tensor_15_strides_0"), val = tensor([1, 1])]; - tensor tensor_15_pad_0 = const()[name = tensor("tensor_15_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor tensor_15_dilations_0 = const()[name = tensor("tensor_15_dilations_0"), val = tensor([1, 1])]; - tensor tensor_15_groups_0 = const()[name = tensor("tensor_15_groups_0"), val = tensor(1)]; - tensor module_pre_encode_conv_6_weight_to_fp16 = const()[name = tensor("module_pre_encode_conv_6_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147520)))]; - tensor module_pre_encode_conv_6_bias_to_fp16 = const()[name = tensor("module_pre_encode_conv_6_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(278656)))]; - tensor tensor_15_cast_fp16 = conv(bias = module_pre_encode_conv_6_bias_to_fp16, dilations = tensor_15_dilations_0, groups = tensor_15_groups_0, pad = tensor_15_pad_0, pad_type = tensor_15_pad_type_0, strides = tensor_15_strides_0, weight = module_pre_encode_conv_6_weight_to_fp16, x = input_13_cast_fp16)[name = tensor("tensor_15_cast_fp16")]; - tensor input_15_cast_fp16 = mul(x = tensor_15_cast_fp16, y = expanded_mask_13_cast_fp16)[name = tensor("input_15_cast_fp16")]; - tensor tensor_cast_fp16 = relu(x = input_15_cast_fp16)[name = tensor("tensor_cast_fp16")]; - tensor x_3_cast_fp16 = mul(x = tensor_cast_fp16, y = expanded_mask_13_cast_fp16)[name = tensor("x_3_cast_fp16")]; - tensor var_275_perm_0 = const()[name = tensor("op_275_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_276 = const()[name = tensor("op_276"), val = tensor([1, 188, -1])]; - tensor var_275_cast_fp16 = transpose(perm = var_275_perm_0, x = x_3_cast_fp16)[name = tensor("transpose_206")]; - tensor input_17_cast_fp16 = reshape(shape = var_276, x = var_275_cast_fp16)[name = tensor("input_17_cast_fp16")]; - tensor module_pre_encode_out_weight_to_fp16 = const()[name = tensor("module_pre_encode_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(279232)))]; - tensor module_pre_encode_out_bias_to_fp16 = const()[name = tensor("module_pre_encode_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2900736)))]; - tensor linear_0_cast_fp16 = linear(bias = module_pre_encode_out_bias_to_fp16, weight = module_pre_encode_out_weight_to_fp16, x = input_17_cast_fp16)[name = tensor("linear_0_cast_fp16")]; - tensor cast_12_dtype_0 = const()[name = tensor("cast_12_dtype_0"), val = tensor("int32")]; - tensor var_314_axes_0 = const()[name = tensor("op_314_axes_0"), val = tensor([-1])]; - tensor encoder_length = cast(dtype = cast_12_dtype_0, x = current_lengths_cast_fp16)[name = tensor("cast_173")]; - tensor var_314 = expand_dims(axes = var_314_axes_0, x = encoder_length)[name = tensor("op_314")]; - tensor pad_mask_1 = less(x = expand_dims_3, y = var_314)[name = tensor("pad_mask_1")]; - tensor var_316_axes_0 = const()[name = tensor("op_316_axes_0"), val = tensor([1])]; - tensor var_316 = expand_dims(axes = var_316_axes_0, x = pad_mask_1)[name = tensor("op_316")]; - tensor var_317 = const()[name = tensor("op_317"), val = tensor([1, 188, 1])]; - tensor pad_mask_for_att_mask_1 = tile(reps = var_317, x = var_316)[name = tensor("pad_mask_for_att_mask_1")]; - tensor var_319_perm_0 = const()[name = tensor("op_319_perm_0"), val = tensor([0, 2, 1])]; - tensor var_319 = transpose(perm = var_319_perm_0, x = pad_mask_for_att_mask_1)[name = tensor("transpose_205")]; - tensor pad_mask_for_att_mask = logical_and(x = pad_mask_for_att_mask_1, y = var_319)[name = tensor("pad_mask_for_att_mask")]; - tensor const_63 = const()[name = tensor("const_63"), val = tensor([[[true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true]]])]; - tensor att_mask = logical_and(x = pad_mask_for_att_mask, y = const_63)[name = tensor("att_mask")]; - tensor mask_9 = logical_not(x = att_mask)[name = tensor("mask_9")]; - tensor pad_mask = logical_not(x = pad_mask_1)[name = tensor("pad_mask")]; - tensor input_21_axes_0 = const()[name = tensor("input_21_axes_0"), val = tensor([-1])]; - tensor module_layers_0_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_0_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2901824)))]; - tensor module_layers_0_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_0_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2902912)))]; - tensor var_9_to_fp16 = const()[name = tensor("op_9_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_21_cast_fp16 = layer_norm(axes = input_21_axes_0, beta = module_layers_0_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_0_norm_feed_forward1_weight_to_fp16, x = linear_0_cast_fp16)[name = tensor("input_21_cast_fp16")]; - tensor module_layers_0_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_0_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2904000)))]; - tensor module_layers_0_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_0_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(5001216)))]; - tensor linear_1_cast_fp16 = linear(bias = module_layers_0_feed_forward1_linear1_bias_to_fp16, weight = module_layers_0_feed_forward1_linear1_weight_to_fp16, x = input_21_cast_fp16)[name = tensor("linear_1_cast_fp16")]; - tensor input_25_cast_fp16 = silu(x = linear_1_cast_fp16)[name = tensor("input_25_cast_fp16")]; - tensor module_layers_0_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_0_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(5005376)))]; - tensor module_layers_0_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_0_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7102592)))]; - tensor linear_2_cast_fp16 = linear(bias = module_layers_0_feed_forward1_linear2_bias_to_fp16, weight = module_layers_0_feed_forward1_linear2_weight_to_fp16, x = input_25_cast_fp16)[name = tensor("linear_2_cast_fp16")]; - tensor var_352_to_fp16 = const()[name = tensor("op_352_to_fp16"), val = tensor(0x1p-1)]; - tensor var_353_cast_fp16 = mul(x = linear_2_cast_fp16, y = var_352_to_fp16)[name = tensor("op_353_cast_fp16")]; - tensor input_31_cast_fp16 = add(x = linear_0_cast_fp16, y = var_353_cast_fp16)[name = tensor("input_31_cast_fp16")]; - tensor query_1_axes_0 = const()[name = tensor("query_1_axes_0"), val = tensor([-1])]; - tensor module_layers_0_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_0_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7103680)))]; - tensor module_layers_0_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_0_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7104768)))]; - tensor query_1_cast_fp16 = layer_norm(axes = query_1_axes_0, beta = module_layers_0_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_0_norm_self_att_weight_to_fp16, x = input_31_cast_fp16)[name = tensor("query_1_cast_fp16")]; - tensor module_layers_0_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_0_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7105856)))]; - tensor module_layers_0_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_0_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7630208)))]; - tensor linear_3_cast_fp16 = linear(bias = module_layers_0_self_attn_linear_q_bias_to_fp16, weight = module_layers_0_self_attn_linear_q_weight_to_fp16, x = query_1_cast_fp16)[name = tensor("linear_3_cast_fp16")]; - tensor var_370 = const()[name = tensor("op_370"), val = tensor([1, -1, 8, 64])]; - tensor q_1_cast_fp16 = reshape(shape = var_370, x = linear_3_cast_fp16)[name = tensor("q_1_cast_fp16")]; - tensor module_layers_0_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_0_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7631296)))]; - tensor module_layers_0_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_0_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8155648)))]; - tensor linear_4_cast_fp16 = linear(bias = module_layers_0_self_attn_linear_k_bias_to_fp16, weight = module_layers_0_self_attn_linear_k_weight_to_fp16, x = query_1_cast_fp16)[name = tensor("linear_4_cast_fp16")]; - tensor var_375 = const()[name = tensor("op_375"), val = tensor([1, -1, 8, 64])]; - tensor k_1_cast_fp16 = reshape(shape = var_375, x = linear_4_cast_fp16)[name = tensor("k_1_cast_fp16")]; - tensor module_layers_0_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_0_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8156736)))]; - tensor module_layers_0_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_0_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8681088)))]; - tensor linear_5_cast_fp16 = linear(bias = module_layers_0_self_attn_linear_v_bias_to_fp16, weight = module_layers_0_self_attn_linear_v_weight_to_fp16, x = query_1_cast_fp16)[name = tensor("linear_5_cast_fp16")]; - tensor var_380 = const()[name = tensor("op_380"), val = tensor([1, -1, 8, 64])]; - tensor v_1_cast_fp16 = reshape(shape = var_380, x = linear_5_cast_fp16)[name = tensor("v_1_cast_fp16")]; - tensor value_3_perm_0 = const()[name = tensor("value_3_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_0_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_0_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8682176)))]; - tensor var_392_cast_fp16 = add(x = q_1_cast_fp16, y = module_layers_0_self_attn_pos_bias_u_to_fp16)[name = tensor("op_392_cast_fp16")]; - tensor module_layers_0_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_0_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8683264)))]; - tensor var_394_cast_fp16 = add(x = q_1_cast_fp16, y = module_layers_0_self_attn_pos_bias_v_to_fp16)[name = tensor("op_394_cast_fp16")]; - tensor q_with_bias_v_1_perm_0 = const()[name = tensor("q_with_bias_v_1_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_7_transpose_x_0 = const()[name = tensor("x_7_transpose_x_0"), val = tensor(false)]; - tensor x_7_transpose_y_0 = const()[name = tensor("x_7_transpose_y_0"), val = tensor(false)]; - tensor var_396_to_fp16 = const()[name = tensor("op_396_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8684352)))]; - tensor q_with_bias_v_1_cast_fp16 = transpose(perm = q_with_bias_v_1_perm_0, x = var_394_cast_fp16)[name = tensor("transpose_203")]; - tensor x_7_cast_fp16 = matmul(transpose_x = x_7_transpose_x_0, transpose_y = x_7_transpose_y_0, x = q_with_bias_v_1_cast_fp16, y = var_396_to_fp16)[name = tensor("x_7_cast_fp16")]; - tensor x_9_pad_0 = const()[name = tensor("x_9_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_9_mode_0 = const()[name = tensor("x_9_mode_0"), val = tensor("constant")]; - tensor const_70_to_fp16 = const()[name = tensor("const_70_to_fp16"), val = tensor(0x0p+0)]; - tensor x_9_cast_fp16 = pad(constant_val = const_70_to_fp16, mode = x_9_mode_0, pad = x_9_pad_0, x = x_7_cast_fp16)[name = tensor("x_9_cast_fp16")]; - tensor var_404 = const()[name = tensor("op_404"), val = tensor([1, 8, -1, 188])]; - tensor x_11_cast_fp16 = reshape(shape = var_404, x = x_9_cast_fp16)[name = tensor("x_11_cast_fp16")]; - tensor var_408_begin_0 = const()[name = tensor("op_408_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_408_end_0 = const()[name = tensor("op_408_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_408_end_mask_0 = const()[name = tensor("op_408_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_408_cast_fp16 = slice_by_index(begin = var_408_begin_0, end = var_408_end_0, end_mask = var_408_end_mask_0, x = x_11_cast_fp16)[name = tensor("op_408_cast_fp16")]; - tensor var_409 = const()[name = tensor("op_409"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_1_cast_fp16 = reshape(shape = var_409, x = var_408_cast_fp16)[name = tensor("matrix_bd_1_cast_fp16")]; - tensor matrix_ac_1_transpose_x_0 = const()[name = tensor("matrix_ac_1_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_1_transpose_y_0 = const()[name = tensor("matrix_ac_1_transpose_y_0"), val = tensor(false)]; - tensor transpose_51_perm_0 = const()[name = tensor("transpose_51_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_52_perm_0 = const()[name = tensor("transpose_52_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_52 = transpose(perm = transpose_52_perm_0, x = k_1_cast_fp16)[name = tensor("transpose_201")]; - tensor transpose_51 = transpose(perm = transpose_51_perm_0, x = var_392_cast_fp16)[name = tensor("transpose_202")]; - tensor matrix_ac_1_cast_fp16 = matmul(transpose_x = matrix_ac_1_transpose_x_0, transpose_y = matrix_ac_1_transpose_y_0, x = transpose_51, y = transpose_52)[name = tensor("matrix_ac_1_cast_fp16")]; - tensor matrix_bd_3_begin_0 = const()[name = tensor("matrix_bd_3_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_3_end_0 = const()[name = tensor("matrix_bd_3_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_3_end_mask_0 = const()[name = tensor("matrix_bd_3_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_3_cast_fp16 = slice_by_index(begin = matrix_bd_3_begin_0, end = matrix_bd_3_end_0, end_mask = matrix_bd_3_end_mask_0, x = matrix_bd_1_cast_fp16)[name = tensor("matrix_bd_3_cast_fp16")]; - tensor var_418_cast_fp16 = add(x = matrix_ac_1_cast_fp16, y = matrix_bd_3_cast_fp16)[name = tensor("op_418_cast_fp16")]; - tensor _inversed_scores_1_y_0_to_fp16 = const()[name = tensor("_inversed_scores_1_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_1_cast_fp16 = mul(x = var_418_cast_fp16, y = _inversed_scores_1_y_0_to_fp16)[name = tensor("_inversed_scores_1_cast_fp16")]; - tensor mask_11_axes_0 = const()[name = tensor("mask_11_axes_0"), val = tensor([1])]; - tensor mask_11 = expand_dims(axes = mask_11_axes_0, x = mask_9)[name = tensor("mask_11")]; - tensor var_12_to_fp16 = const()[name = tensor("op_12_to_fp16"), val = tensor(-0x1.388p+13)]; - tensor scores_3_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_1_cast_fp16, cond = mask_11)[name = tensor("scores_3_cast_fp16")]; - tensor var_424_cast_fp16 = softmax(axis = var_23, x = scores_3_cast_fp16)[name = tensor("op_424_cast_fp16")]; - tensor var_11_to_fp16 = const()[name = tensor("op_11_to_fp16"), val = tensor(0x0p+0)]; - tensor input_33_cast_fp16 = select(a = var_11_to_fp16, b = var_424_cast_fp16, cond = mask_11)[name = tensor("input_33_cast_fp16")]; - tensor x_13_transpose_x_0 = const()[name = tensor("x_13_transpose_x_0"), val = tensor(false)]; - tensor x_13_transpose_y_0 = const()[name = tensor("x_13_transpose_y_0"), val = tensor(false)]; - tensor value_3_cast_fp16 = transpose(perm = value_3_perm_0, x = v_1_cast_fp16)[name = tensor("transpose_204")]; - tensor x_13_cast_fp16 = matmul(transpose_x = x_13_transpose_x_0, transpose_y = x_13_transpose_y_0, x = input_33_cast_fp16, y = value_3_cast_fp16)[name = tensor("x_13_cast_fp16")]; - tensor var_428_perm_0 = const()[name = tensor("op_428_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_429 = const()[name = tensor("op_429"), val = tensor([1, -1, 512])]; - tensor var_428_cast_fp16 = transpose(perm = var_428_perm_0, x = x_13_cast_fp16)[name = tensor("transpose_200")]; - tensor input_35_cast_fp16 = reshape(shape = var_429, x = var_428_cast_fp16)[name = tensor("input_35_cast_fp16")]; - tensor module_layers_0_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_0_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9068416)))]; - tensor module_layers_0_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_0_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9592768)))]; - tensor linear_7_cast_fp16 = linear(bias = module_layers_0_self_attn_linear_out_bias_to_fp16, weight = module_layers_0_self_attn_linear_out_weight_to_fp16, x = input_35_cast_fp16)[name = tensor("linear_7_cast_fp16")]; - tensor input_39_cast_fp16 = add(x = input_31_cast_fp16, y = linear_7_cast_fp16)[name = tensor("input_39_cast_fp16")]; - tensor x_17_axes_0 = const()[name = tensor("x_17_axes_0"), val = tensor([-1])]; - tensor module_layers_0_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_0_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9593856)))]; - tensor module_layers_0_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_0_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9594944)))]; - tensor x_17_cast_fp16 = layer_norm(axes = x_17_axes_0, beta = module_layers_0_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_0_norm_conv_weight_to_fp16, x = input_39_cast_fp16)[name = tensor("x_17_cast_fp16")]; - tensor input_41_perm_0 = const()[name = tensor("input_41_perm_0"), val = tensor([0, 2, 1])]; - tensor input_43_pad_type_0 = const()[name = tensor("input_43_pad_type_0"), val = tensor("valid")]; - tensor input_43_strides_0 = const()[name = tensor("input_43_strides_0"), val = tensor([1])]; - tensor input_43_pad_0 = const()[name = tensor("input_43_pad_0"), val = tensor([0, 0])]; - tensor input_43_dilations_0 = const()[name = tensor("input_43_dilations_0"), val = tensor([1])]; - tensor input_43_groups_0 = const()[name = tensor("input_43_groups_0"), val = tensor(1)]; - tensor module_layers_0_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_0_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9596032)))]; - tensor module_layers_0_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_0_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10644672)))]; - tensor input_41_cast_fp16 = transpose(perm = input_41_perm_0, x = x_17_cast_fp16)[name = tensor("transpose_199")]; - tensor input_43_cast_fp16 = conv(bias = module_layers_0_conv_pointwise_conv1_bias_to_fp16, dilations = input_43_dilations_0, groups = input_43_groups_0, pad = input_43_pad_0, pad_type = input_43_pad_type_0, strides = input_43_strides_0, weight = module_layers_0_conv_pointwise_conv1_weight_to_fp16, x = input_41_cast_fp16)[name = tensor("input_43_cast_fp16")]; - tensor x_19_split_num_splits_0 = const()[name = tensor("x_19_split_num_splits_0"), val = tensor(2)]; - tensor x_19_split_axis_0 = const()[name = tensor("x_19_split_axis_0"), val = tensor(1)]; - tensor x_19_split_cast_fp16_0, tensor x_19_split_cast_fp16_1 = split(axis = x_19_split_axis_0, num_splits = x_19_split_num_splits_0, x = input_43_cast_fp16)[name = tensor("x_19_split_cast_fp16")]; - tensor x_19_split_1_sigmoid_cast_fp16 = sigmoid(x = x_19_split_cast_fp16_1)[name = tensor("x_19_split_1_sigmoid_cast_fp16")]; - tensor x_19_cast_fp16 = mul(x = x_19_split_cast_fp16_0, y = x_19_split_1_sigmoid_cast_fp16)[name = tensor("x_19_cast_fp16")]; - tensor var_453_axes_0 = const()[name = tensor("op_453_axes_0"), val = tensor([1])]; - tensor var_453 = expand_dims(axes = var_453_axes_0, x = pad_mask)[name = tensor("op_453")]; - tensor input_45_cast_fp16 = select(a = var_11_to_fp16, b = x_19_cast_fp16, cond = var_453)[name = tensor("input_45_cast_fp16")]; - tensor input_47_pad_0 = const()[name = tensor("input_47_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_47_mode_0 = const()[name = tensor("input_47_mode_0"), val = tensor("constant")]; - tensor const_73_to_fp16 = const()[name = tensor("const_73_to_fp16"), val = tensor(0x0p+0)]; - tensor input_47_cast_fp16 = pad(constant_val = const_73_to_fp16, mode = input_47_mode_0, pad = input_47_pad_0, x = input_45_cast_fp16)[name = tensor("input_47_cast_fp16")]; - tensor input_49_pad_type_0 = const()[name = tensor("input_49_pad_type_0"), val = tensor("valid")]; - tensor input_49_groups_0 = const()[name = tensor("input_49_groups_0"), val = tensor(512)]; - tensor input_49_strides_0 = const()[name = tensor("input_49_strides_0"), val = tensor([1])]; - tensor input_49_pad_0 = const()[name = tensor("input_49_pad_0"), val = tensor([0, 0])]; - tensor input_49_dilations_0 = const()[name = tensor("input_49_dilations_0"), val = tensor([1])]; - tensor const_234_to_fp16 = const()[name = tensor("const_234_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10646784)))]; - tensor const_235_to_fp16 = const()[name = tensor("const_235_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10656064)))]; - tensor input_51_cast_fp16 = conv(bias = const_235_to_fp16, dilations = input_49_dilations_0, groups = input_49_groups_0, pad = input_49_pad_0, pad_type = input_49_pad_type_0, strides = input_49_strides_0, weight = const_234_to_fp16, x = input_47_cast_fp16)[name = tensor("input_51_cast_fp16")]; - tensor input_53_cast_fp16 = silu(x = input_51_cast_fp16)[name = tensor("input_53_cast_fp16")]; - tensor x_21_pad_type_0 = const()[name = tensor("x_21_pad_type_0"), val = tensor("valid")]; - tensor x_21_strides_0 = const()[name = tensor("x_21_strides_0"), val = tensor([1])]; - tensor x_21_pad_0 = const()[name = tensor("x_21_pad_0"), val = tensor([0, 0])]; - tensor x_21_dilations_0 = const()[name = tensor("x_21_dilations_0"), val = tensor([1])]; - tensor x_21_groups_0 = const()[name = tensor("x_21_groups_0"), val = tensor(1)]; - tensor module_layers_0_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_0_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10657152)))]; - tensor module_layers_0_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_0_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11181504)))]; - tensor x_21_cast_fp16 = conv(bias = module_layers_0_conv_pointwise_conv2_bias_to_fp16, dilations = x_21_dilations_0, groups = x_21_groups_0, pad = x_21_pad_0, pad_type = x_21_pad_type_0, strides = x_21_strides_0, weight = module_layers_0_conv_pointwise_conv2_weight_to_fp16, x = input_53_cast_fp16)[name = tensor("x_21_cast_fp16")]; - tensor input_55_perm_0 = const()[name = tensor("input_55_perm_0"), val = tensor([0, 2, 1])]; - tensor input_55_cast_fp16 = transpose(perm = input_55_perm_0, x = x_21_cast_fp16)[name = tensor("transpose_198")]; - tensor input_57_cast_fp16 = add(x = input_39_cast_fp16, y = input_55_cast_fp16)[name = tensor("input_57_cast_fp16")]; - tensor input_59_axes_0 = const()[name = tensor("input_59_axes_0"), val = tensor([-1])]; - tensor module_layers_0_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_0_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11182592)))]; - tensor module_layers_0_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_0_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11183680)))]; - tensor input_59_cast_fp16 = layer_norm(axes = input_59_axes_0, beta = module_layers_0_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_0_norm_feed_forward2_weight_to_fp16, x = input_57_cast_fp16)[name = tensor("input_59_cast_fp16")]; - tensor module_layers_0_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_0_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11184768)))]; - tensor module_layers_0_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_0_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(13281984)))]; - tensor linear_8_cast_fp16 = linear(bias = module_layers_0_feed_forward2_linear1_bias_to_fp16, weight = module_layers_0_feed_forward2_linear1_weight_to_fp16, x = input_59_cast_fp16)[name = tensor("linear_8_cast_fp16")]; - tensor input_63_cast_fp16 = silu(x = linear_8_cast_fp16)[name = tensor("input_63_cast_fp16")]; - tensor module_layers_0_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_0_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(13286144)))]; - tensor module_layers_0_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_0_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15383360)))]; - tensor linear_9_cast_fp16 = linear(bias = module_layers_0_feed_forward2_linear2_bias_to_fp16, weight = module_layers_0_feed_forward2_linear2_weight_to_fp16, x = input_63_cast_fp16)[name = tensor("linear_9_cast_fp16")]; - tensor var_495_to_fp16 = const()[name = tensor("op_495_to_fp16"), val = tensor(0x1p-1)]; - tensor var_496_cast_fp16 = mul(x = linear_9_cast_fp16, y = var_495_to_fp16)[name = tensor("op_496_cast_fp16")]; - tensor input_69_cast_fp16 = add(x = input_57_cast_fp16, y = var_496_cast_fp16)[name = tensor("input_69_cast_fp16")]; - tensor input_71_axes_0 = const()[name = tensor("input_71_axes_0"), val = tensor([-1])]; - tensor module_layers_0_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_0_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15384448)))]; - tensor module_layers_0_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_0_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15385536)))]; - tensor input_71_cast_fp16 = layer_norm(axes = input_71_axes_0, beta = module_layers_0_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_0_norm_out_weight_to_fp16, x = input_69_cast_fp16)[name = tensor("input_71_cast_fp16")]; - tensor input_73_axes_0 = const()[name = tensor("input_73_axes_0"), val = tensor([-1])]; - tensor module_layers_1_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_1_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15386624)))]; - tensor module_layers_1_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_1_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15387712)))]; - tensor input_73_cast_fp16 = layer_norm(axes = input_73_axes_0, beta = module_layers_1_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_1_norm_feed_forward1_weight_to_fp16, x = input_71_cast_fp16)[name = tensor("input_73_cast_fp16")]; - tensor module_layers_1_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_1_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15388800)))]; - tensor module_layers_1_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_1_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(17486016)))]; - tensor linear_10_cast_fp16 = linear(bias = module_layers_1_feed_forward1_linear1_bias_to_fp16, weight = module_layers_1_feed_forward1_linear1_weight_to_fp16, x = input_73_cast_fp16)[name = tensor("linear_10_cast_fp16")]; - tensor input_77_cast_fp16 = silu(x = linear_10_cast_fp16)[name = tensor("input_77_cast_fp16")]; - tensor module_layers_1_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_1_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(17490176)))]; - tensor module_layers_1_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_1_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19587392)))]; - tensor linear_11_cast_fp16 = linear(bias = module_layers_1_feed_forward1_linear2_bias_to_fp16, weight = module_layers_1_feed_forward1_linear2_weight_to_fp16, x = input_77_cast_fp16)[name = tensor("linear_11_cast_fp16")]; - tensor var_526_to_fp16 = const()[name = tensor("op_526_to_fp16"), val = tensor(0x1p-1)]; - tensor var_527_cast_fp16 = mul(x = linear_11_cast_fp16, y = var_526_to_fp16)[name = tensor("op_527_cast_fp16")]; - tensor input_83_cast_fp16 = add(x = input_71_cast_fp16, y = var_527_cast_fp16)[name = tensor("input_83_cast_fp16")]; - tensor query_3_axes_0 = const()[name = tensor("query_3_axes_0"), val = tensor([-1])]; - tensor module_layers_1_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_1_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19588480)))]; - tensor module_layers_1_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_1_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19589568)))]; - tensor query_3_cast_fp16 = layer_norm(axes = query_3_axes_0, beta = module_layers_1_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_1_norm_self_att_weight_to_fp16, x = input_83_cast_fp16)[name = tensor("query_3_cast_fp16")]; - tensor module_layers_1_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_1_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19590656)))]; - tensor module_layers_1_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_1_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(20115008)))]; - tensor linear_12_cast_fp16 = linear(bias = module_layers_1_self_attn_linear_q_bias_to_fp16, weight = module_layers_1_self_attn_linear_q_weight_to_fp16, x = query_3_cast_fp16)[name = tensor("linear_12_cast_fp16")]; - tensor var_544 = const()[name = tensor("op_544"), val = tensor([1, -1, 8, 64])]; - tensor q_7_cast_fp16 = reshape(shape = var_544, x = linear_12_cast_fp16)[name = tensor("q_7_cast_fp16")]; - tensor module_layers_1_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_1_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(20116096)))]; - tensor module_layers_1_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_1_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(20640448)))]; - tensor linear_13_cast_fp16 = linear(bias = module_layers_1_self_attn_linear_k_bias_to_fp16, weight = module_layers_1_self_attn_linear_k_weight_to_fp16, x = query_3_cast_fp16)[name = tensor("linear_13_cast_fp16")]; - tensor var_549 = const()[name = tensor("op_549"), val = tensor([1, -1, 8, 64])]; - tensor k_5_cast_fp16 = reshape(shape = var_549, x = linear_13_cast_fp16)[name = tensor("k_5_cast_fp16")]; - tensor module_layers_1_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_1_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(20641536)))]; - tensor module_layers_1_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_1_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21165888)))]; - tensor linear_14_cast_fp16 = linear(bias = module_layers_1_self_attn_linear_v_bias_to_fp16, weight = module_layers_1_self_attn_linear_v_weight_to_fp16, x = query_3_cast_fp16)[name = tensor("linear_14_cast_fp16")]; - tensor var_554 = const()[name = tensor("op_554"), val = tensor([1, -1, 8, 64])]; - tensor v_3_cast_fp16 = reshape(shape = var_554, x = linear_14_cast_fp16)[name = tensor("v_3_cast_fp16")]; - tensor value_5_perm_0 = const()[name = tensor("value_5_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_1_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_1_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21166976)))]; - tensor var_566_cast_fp16 = add(x = q_7_cast_fp16, y = module_layers_1_self_attn_pos_bias_u_to_fp16)[name = tensor("op_566_cast_fp16")]; - tensor module_layers_1_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_1_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21168064)))]; - tensor var_568_cast_fp16 = add(x = q_7_cast_fp16, y = module_layers_1_self_attn_pos_bias_v_to_fp16)[name = tensor("op_568_cast_fp16")]; - tensor q_with_bias_v_3_perm_0 = const()[name = tensor("q_with_bias_v_3_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_29_transpose_x_0 = const()[name = tensor("x_29_transpose_x_0"), val = tensor(false)]; - tensor x_29_transpose_y_0 = const()[name = tensor("x_29_transpose_y_0"), val = tensor(false)]; - tensor var_570_to_fp16 = const()[name = tensor("op_570_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21169152)))]; - tensor q_with_bias_v_3_cast_fp16 = transpose(perm = q_with_bias_v_3_perm_0, x = var_568_cast_fp16)[name = tensor("transpose_196")]; - tensor x_29_cast_fp16 = matmul(transpose_x = x_29_transpose_x_0, transpose_y = x_29_transpose_y_0, x = q_with_bias_v_3_cast_fp16, y = var_570_to_fp16)[name = tensor("x_29_cast_fp16")]; - tensor x_31_pad_0 = const()[name = tensor("x_31_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_31_mode_0 = const()[name = tensor("x_31_mode_0"), val = tensor("constant")]; - tensor const_80_to_fp16 = const()[name = tensor("const_80_to_fp16"), val = tensor(0x0p+0)]; - tensor x_31_cast_fp16 = pad(constant_val = const_80_to_fp16, mode = x_31_mode_0, pad = x_31_pad_0, x = x_29_cast_fp16)[name = tensor("x_31_cast_fp16")]; - tensor var_578 = const()[name = tensor("op_578"), val = tensor([1, 8, -1, 188])]; - tensor x_33_cast_fp16 = reshape(shape = var_578, x = x_31_cast_fp16)[name = tensor("x_33_cast_fp16")]; - tensor var_582_begin_0 = const()[name = tensor("op_582_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_582_end_0 = const()[name = tensor("op_582_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_582_end_mask_0 = const()[name = tensor("op_582_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_582_cast_fp16 = slice_by_index(begin = var_582_begin_0, end = var_582_end_0, end_mask = var_582_end_mask_0, x = x_33_cast_fp16)[name = tensor("op_582_cast_fp16")]; - tensor var_583 = const()[name = tensor("op_583"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_5_cast_fp16 = reshape(shape = var_583, x = var_582_cast_fp16)[name = tensor("matrix_bd_5_cast_fp16")]; - tensor matrix_ac_3_transpose_x_0 = const()[name = tensor("matrix_ac_3_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_3_transpose_y_0 = const()[name = tensor("matrix_ac_3_transpose_y_0"), val = tensor(false)]; - tensor transpose_53_perm_0 = const()[name = tensor("transpose_53_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_54_perm_0 = const()[name = tensor("transpose_54_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_54 = transpose(perm = transpose_54_perm_0, x = k_5_cast_fp16)[name = tensor("transpose_194")]; - tensor transpose_53 = transpose(perm = transpose_53_perm_0, x = var_566_cast_fp16)[name = tensor("transpose_195")]; - tensor matrix_ac_3_cast_fp16 = matmul(transpose_x = matrix_ac_3_transpose_x_0, transpose_y = matrix_ac_3_transpose_y_0, x = transpose_53, y = transpose_54)[name = tensor("matrix_ac_3_cast_fp16")]; - tensor matrix_bd_7_begin_0 = const()[name = tensor("matrix_bd_7_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_7_end_0 = const()[name = tensor("matrix_bd_7_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_7_end_mask_0 = const()[name = tensor("matrix_bd_7_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_7_cast_fp16 = slice_by_index(begin = matrix_bd_7_begin_0, end = matrix_bd_7_end_0, end_mask = matrix_bd_7_end_mask_0, x = matrix_bd_5_cast_fp16)[name = tensor("matrix_bd_7_cast_fp16")]; - tensor var_592_cast_fp16 = add(x = matrix_ac_3_cast_fp16, y = matrix_bd_7_cast_fp16)[name = tensor("op_592_cast_fp16")]; - tensor _inversed_scores_5_y_0_to_fp16 = const()[name = tensor("_inversed_scores_5_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_5_cast_fp16 = mul(x = var_592_cast_fp16, y = _inversed_scores_5_y_0_to_fp16)[name = tensor("_inversed_scores_5_cast_fp16")]; - tensor scores_7_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_5_cast_fp16, cond = mask_11)[name = tensor("scores_7_cast_fp16")]; - tensor var_598_cast_fp16 = softmax(axis = var_23, x = scores_7_cast_fp16)[name = tensor("op_598_cast_fp16")]; - tensor input_85_cast_fp16 = select(a = var_11_to_fp16, b = var_598_cast_fp16, cond = mask_11)[name = tensor("input_85_cast_fp16")]; - tensor x_35_transpose_x_0 = const()[name = tensor("x_35_transpose_x_0"), val = tensor(false)]; - tensor x_35_transpose_y_0 = const()[name = tensor("x_35_transpose_y_0"), val = tensor(false)]; - tensor value_5_cast_fp16 = transpose(perm = value_5_perm_0, x = v_3_cast_fp16)[name = tensor("transpose_197")]; - tensor x_35_cast_fp16 = matmul(transpose_x = x_35_transpose_x_0, transpose_y = x_35_transpose_y_0, x = input_85_cast_fp16, y = value_5_cast_fp16)[name = tensor("x_35_cast_fp16")]; - tensor var_602_perm_0 = const()[name = tensor("op_602_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_603 = const()[name = tensor("op_603"), val = tensor([1, -1, 512])]; - tensor var_602_cast_fp16 = transpose(perm = var_602_perm_0, x = x_35_cast_fp16)[name = tensor("transpose_193")]; - tensor input_87_cast_fp16 = reshape(shape = var_603, x = var_602_cast_fp16)[name = tensor("input_87_cast_fp16")]; - tensor module_layers_1_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_1_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21553216)))]; - tensor module_layers_1_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_1_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22077568)))]; - tensor linear_16_cast_fp16 = linear(bias = module_layers_1_self_attn_linear_out_bias_to_fp16, weight = module_layers_1_self_attn_linear_out_weight_to_fp16, x = input_87_cast_fp16)[name = tensor("linear_16_cast_fp16")]; - tensor input_91_cast_fp16 = add(x = input_83_cast_fp16, y = linear_16_cast_fp16)[name = tensor("input_91_cast_fp16")]; - tensor x_39_axes_0 = const()[name = tensor("x_39_axes_0"), val = tensor([-1])]; - tensor module_layers_1_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_1_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22078656)))]; - tensor module_layers_1_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_1_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22079744)))]; - tensor x_39_cast_fp16 = layer_norm(axes = x_39_axes_0, beta = module_layers_1_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_1_norm_conv_weight_to_fp16, x = input_91_cast_fp16)[name = tensor("x_39_cast_fp16")]; - tensor input_93_perm_0 = const()[name = tensor("input_93_perm_0"), val = tensor([0, 2, 1])]; - tensor input_95_pad_type_0 = const()[name = tensor("input_95_pad_type_0"), val = tensor("valid")]; - tensor input_95_strides_0 = const()[name = tensor("input_95_strides_0"), val = tensor([1])]; - tensor input_95_pad_0 = const()[name = tensor("input_95_pad_0"), val = tensor([0, 0])]; - tensor input_95_dilations_0 = const()[name = tensor("input_95_dilations_0"), val = tensor([1])]; - tensor input_95_groups_0 = const()[name = tensor("input_95_groups_0"), val = tensor(1)]; - tensor module_layers_1_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_1_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22080832)))]; - tensor module_layers_1_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_1_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23129472)))]; - tensor input_93_cast_fp16 = transpose(perm = input_93_perm_0, x = x_39_cast_fp16)[name = tensor("transpose_192")]; - tensor input_95_cast_fp16 = conv(bias = module_layers_1_conv_pointwise_conv1_bias_to_fp16, dilations = input_95_dilations_0, groups = input_95_groups_0, pad = input_95_pad_0, pad_type = input_95_pad_type_0, strides = input_95_strides_0, weight = module_layers_1_conv_pointwise_conv1_weight_to_fp16, x = input_93_cast_fp16)[name = tensor("input_95_cast_fp16")]; - tensor x_41_split_num_splits_0 = const()[name = tensor("x_41_split_num_splits_0"), val = tensor(2)]; - tensor x_41_split_axis_0 = const()[name = tensor("x_41_split_axis_0"), val = tensor(1)]; - tensor x_41_split_cast_fp16_0, tensor x_41_split_cast_fp16_1 = split(axis = x_41_split_axis_0, num_splits = x_41_split_num_splits_0, x = input_95_cast_fp16)[name = tensor("x_41_split_cast_fp16")]; - tensor x_41_split_1_sigmoid_cast_fp16 = sigmoid(x = x_41_split_cast_fp16_1)[name = tensor("x_41_split_1_sigmoid_cast_fp16")]; - tensor x_41_cast_fp16 = mul(x = x_41_split_cast_fp16_0, y = x_41_split_1_sigmoid_cast_fp16)[name = tensor("x_41_cast_fp16")]; - tensor input_97_cast_fp16 = select(a = var_11_to_fp16, b = x_41_cast_fp16, cond = var_453)[name = tensor("input_97_cast_fp16")]; - tensor input_99_pad_0 = const()[name = tensor("input_99_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_99_mode_0 = const()[name = tensor("input_99_mode_0"), val = tensor("constant")]; - tensor const_83_to_fp16 = const()[name = tensor("const_83_to_fp16"), val = tensor(0x0p+0)]; - tensor input_99_cast_fp16 = pad(constant_val = const_83_to_fp16, mode = input_99_mode_0, pad = input_99_pad_0, x = input_97_cast_fp16)[name = tensor("input_99_cast_fp16")]; - tensor input_101_pad_type_0 = const()[name = tensor("input_101_pad_type_0"), val = tensor("valid")]; - tensor input_101_groups_0 = const()[name = tensor("input_101_groups_0"), val = tensor(512)]; - tensor input_101_strides_0 = const()[name = tensor("input_101_strides_0"), val = tensor([1])]; - tensor input_101_pad_0 = const()[name = tensor("input_101_pad_0"), val = tensor([0, 0])]; - tensor input_101_dilations_0 = const()[name = tensor("input_101_dilations_0"), val = tensor([1])]; - tensor const_236_to_fp16 = const()[name = tensor("const_236_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23131584)))]; - tensor const_237_to_fp16 = const()[name = tensor("const_237_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23140864)))]; - tensor input_103_cast_fp16 = conv(bias = const_237_to_fp16, dilations = input_101_dilations_0, groups = input_101_groups_0, pad = input_101_pad_0, pad_type = input_101_pad_type_0, strides = input_101_strides_0, weight = const_236_to_fp16, x = input_99_cast_fp16)[name = tensor("input_103_cast_fp16")]; - tensor input_105_cast_fp16 = silu(x = input_103_cast_fp16)[name = tensor("input_105_cast_fp16")]; - tensor x_43_pad_type_0 = const()[name = tensor("x_43_pad_type_0"), val = tensor("valid")]; - tensor x_43_strides_0 = const()[name = tensor("x_43_strides_0"), val = tensor([1])]; - tensor x_43_pad_0 = const()[name = tensor("x_43_pad_0"), val = tensor([0, 0])]; - tensor x_43_dilations_0 = const()[name = tensor("x_43_dilations_0"), val = tensor([1])]; - tensor x_43_groups_0 = const()[name = tensor("x_43_groups_0"), val = tensor(1)]; - tensor module_layers_1_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_1_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23141952)))]; - tensor module_layers_1_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_1_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23666304)))]; - tensor x_43_cast_fp16 = conv(bias = module_layers_1_conv_pointwise_conv2_bias_to_fp16, dilations = x_43_dilations_0, groups = x_43_groups_0, pad = x_43_pad_0, pad_type = x_43_pad_type_0, strides = x_43_strides_0, weight = module_layers_1_conv_pointwise_conv2_weight_to_fp16, x = input_105_cast_fp16)[name = tensor("x_43_cast_fp16")]; - tensor input_107_perm_0 = const()[name = tensor("input_107_perm_0"), val = tensor([0, 2, 1])]; - tensor input_107_cast_fp16 = transpose(perm = input_107_perm_0, x = x_43_cast_fp16)[name = tensor("transpose_191")]; - tensor input_109_cast_fp16 = add(x = input_91_cast_fp16, y = input_107_cast_fp16)[name = tensor("input_109_cast_fp16")]; - tensor input_111_axes_0 = const()[name = tensor("input_111_axes_0"), val = tensor([-1])]; - tensor module_layers_1_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_1_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23667392)))]; - tensor module_layers_1_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_1_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23668480)))]; - tensor input_111_cast_fp16 = layer_norm(axes = input_111_axes_0, beta = module_layers_1_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_1_norm_feed_forward2_weight_to_fp16, x = input_109_cast_fp16)[name = tensor("input_111_cast_fp16")]; - tensor module_layers_1_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_1_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23669568)))]; - tensor module_layers_1_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_1_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(25766784)))]; - tensor linear_17_cast_fp16 = linear(bias = module_layers_1_feed_forward2_linear1_bias_to_fp16, weight = module_layers_1_feed_forward2_linear1_weight_to_fp16, x = input_111_cast_fp16)[name = tensor("linear_17_cast_fp16")]; - tensor input_115_cast_fp16 = silu(x = linear_17_cast_fp16)[name = tensor("input_115_cast_fp16")]; - tensor module_layers_1_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_1_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(25770944)))]; - tensor module_layers_1_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_1_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27868160)))]; - tensor linear_18_cast_fp16 = linear(bias = module_layers_1_feed_forward2_linear2_bias_to_fp16, weight = module_layers_1_feed_forward2_linear2_weight_to_fp16, x = input_115_cast_fp16)[name = tensor("linear_18_cast_fp16")]; - tensor var_669_to_fp16 = const()[name = tensor("op_669_to_fp16"), val = tensor(0x1p-1)]; - tensor var_670_cast_fp16 = mul(x = linear_18_cast_fp16, y = var_669_to_fp16)[name = tensor("op_670_cast_fp16")]; - tensor input_121_cast_fp16 = add(x = input_109_cast_fp16, y = var_670_cast_fp16)[name = tensor("input_121_cast_fp16")]; - tensor input_123_axes_0 = const()[name = tensor("input_123_axes_0"), val = tensor([-1])]; - tensor module_layers_1_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_1_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27869248)))]; - tensor module_layers_1_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_1_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27870336)))]; - tensor input_123_cast_fp16 = layer_norm(axes = input_123_axes_0, beta = module_layers_1_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_1_norm_out_weight_to_fp16, x = input_121_cast_fp16)[name = tensor("input_123_cast_fp16")]; - tensor input_125_axes_0 = const()[name = tensor("input_125_axes_0"), val = tensor([-1])]; - tensor module_layers_2_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_2_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27871424)))]; - tensor module_layers_2_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_2_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27872512)))]; - tensor input_125_cast_fp16 = layer_norm(axes = input_125_axes_0, beta = module_layers_2_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_2_norm_feed_forward1_weight_to_fp16, x = input_123_cast_fp16)[name = tensor("input_125_cast_fp16")]; - tensor module_layers_2_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_2_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27873600)))]; - tensor module_layers_2_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_2_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(29970816)))]; - tensor linear_19_cast_fp16 = linear(bias = module_layers_2_feed_forward1_linear1_bias_to_fp16, weight = module_layers_2_feed_forward1_linear1_weight_to_fp16, x = input_125_cast_fp16)[name = tensor("linear_19_cast_fp16")]; - tensor input_129_cast_fp16 = silu(x = linear_19_cast_fp16)[name = tensor("input_129_cast_fp16")]; - tensor module_layers_2_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_2_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(29974976)))]; - tensor module_layers_2_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_2_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32072192)))]; - tensor linear_20_cast_fp16 = linear(bias = module_layers_2_feed_forward1_linear2_bias_to_fp16, weight = module_layers_2_feed_forward1_linear2_weight_to_fp16, x = input_129_cast_fp16)[name = tensor("linear_20_cast_fp16")]; - tensor var_700_to_fp16 = const()[name = tensor("op_700_to_fp16"), val = tensor(0x1p-1)]; - tensor var_701_cast_fp16 = mul(x = linear_20_cast_fp16, y = var_700_to_fp16)[name = tensor("op_701_cast_fp16")]; - tensor input_135_cast_fp16 = add(x = input_123_cast_fp16, y = var_701_cast_fp16)[name = tensor("input_135_cast_fp16")]; - tensor query_5_axes_0 = const()[name = tensor("query_5_axes_0"), val = tensor([-1])]; - tensor module_layers_2_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_2_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32073280)))]; - tensor module_layers_2_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_2_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32074368)))]; - tensor query_5_cast_fp16 = layer_norm(axes = query_5_axes_0, beta = module_layers_2_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_2_norm_self_att_weight_to_fp16, x = input_135_cast_fp16)[name = tensor("query_5_cast_fp16")]; - tensor module_layers_2_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_2_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32075456)))]; - tensor module_layers_2_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_2_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32599808)))]; - tensor linear_21_cast_fp16 = linear(bias = module_layers_2_self_attn_linear_q_bias_to_fp16, weight = module_layers_2_self_attn_linear_q_weight_to_fp16, x = query_5_cast_fp16)[name = tensor("linear_21_cast_fp16")]; - tensor var_718 = const()[name = tensor("op_718"), val = tensor([1, -1, 8, 64])]; - tensor q_13_cast_fp16 = reshape(shape = var_718, x = linear_21_cast_fp16)[name = tensor("q_13_cast_fp16")]; - tensor module_layers_2_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_2_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32600896)))]; - tensor module_layers_2_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_2_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33125248)))]; - tensor linear_22_cast_fp16 = linear(bias = module_layers_2_self_attn_linear_k_bias_to_fp16, weight = module_layers_2_self_attn_linear_k_weight_to_fp16, x = query_5_cast_fp16)[name = tensor("linear_22_cast_fp16")]; - tensor var_723 = const()[name = tensor("op_723"), val = tensor([1, -1, 8, 64])]; - tensor k_9_cast_fp16 = reshape(shape = var_723, x = linear_22_cast_fp16)[name = tensor("k_9_cast_fp16")]; - tensor module_layers_2_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_2_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33126336)))]; - tensor module_layers_2_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_2_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33650688)))]; - tensor linear_23_cast_fp16 = linear(bias = module_layers_2_self_attn_linear_v_bias_to_fp16, weight = module_layers_2_self_attn_linear_v_weight_to_fp16, x = query_5_cast_fp16)[name = tensor("linear_23_cast_fp16")]; - tensor var_728 = const()[name = tensor("op_728"), val = tensor([1, -1, 8, 64])]; - tensor v_5_cast_fp16 = reshape(shape = var_728, x = linear_23_cast_fp16)[name = tensor("v_5_cast_fp16")]; - tensor value_7_perm_0 = const()[name = tensor("value_7_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_2_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_2_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33651776)))]; - tensor var_740_cast_fp16 = add(x = q_13_cast_fp16, y = module_layers_2_self_attn_pos_bias_u_to_fp16)[name = tensor("op_740_cast_fp16")]; - tensor module_layers_2_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_2_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33652864)))]; - tensor var_742_cast_fp16 = add(x = q_13_cast_fp16, y = module_layers_2_self_attn_pos_bias_v_to_fp16)[name = tensor("op_742_cast_fp16")]; - tensor q_with_bias_v_5_perm_0 = const()[name = tensor("q_with_bias_v_5_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_51_transpose_x_0 = const()[name = tensor("x_51_transpose_x_0"), val = tensor(false)]; - tensor x_51_transpose_y_0 = const()[name = tensor("x_51_transpose_y_0"), val = tensor(false)]; - tensor var_744_to_fp16 = const()[name = tensor("op_744_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33653952)))]; - tensor q_with_bias_v_5_cast_fp16 = transpose(perm = q_with_bias_v_5_perm_0, x = var_742_cast_fp16)[name = tensor("transpose_189")]; - tensor x_51_cast_fp16 = matmul(transpose_x = x_51_transpose_x_0, transpose_y = x_51_transpose_y_0, x = q_with_bias_v_5_cast_fp16, y = var_744_to_fp16)[name = tensor("x_51_cast_fp16")]; - tensor x_53_pad_0 = const()[name = tensor("x_53_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_53_mode_0 = const()[name = tensor("x_53_mode_0"), val = tensor("constant")]; - tensor const_90_to_fp16 = const()[name = tensor("const_90_to_fp16"), val = tensor(0x0p+0)]; - tensor x_53_cast_fp16 = pad(constant_val = const_90_to_fp16, mode = x_53_mode_0, pad = x_53_pad_0, x = x_51_cast_fp16)[name = tensor("x_53_cast_fp16")]; - tensor var_752 = const()[name = tensor("op_752"), val = tensor([1, 8, -1, 188])]; - tensor x_55_cast_fp16 = reshape(shape = var_752, x = x_53_cast_fp16)[name = tensor("x_55_cast_fp16")]; - tensor var_756_begin_0 = const()[name = tensor("op_756_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_756_end_0 = const()[name = tensor("op_756_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_756_end_mask_0 = const()[name = tensor("op_756_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_756_cast_fp16 = slice_by_index(begin = var_756_begin_0, end = var_756_end_0, end_mask = var_756_end_mask_0, x = x_55_cast_fp16)[name = tensor("op_756_cast_fp16")]; - tensor var_757 = const()[name = tensor("op_757"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_9_cast_fp16 = reshape(shape = var_757, x = var_756_cast_fp16)[name = tensor("matrix_bd_9_cast_fp16")]; - tensor matrix_ac_5_transpose_x_0 = const()[name = tensor("matrix_ac_5_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_5_transpose_y_0 = const()[name = tensor("matrix_ac_5_transpose_y_0"), val = tensor(false)]; - tensor transpose_55_perm_0 = const()[name = tensor("transpose_55_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_56_perm_0 = const()[name = tensor("transpose_56_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_56 = transpose(perm = transpose_56_perm_0, x = k_9_cast_fp16)[name = tensor("transpose_187")]; - tensor transpose_55 = transpose(perm = transpose_55_perm_0, x = var_740_cast_fp16)[name = tensor("transpose_188")]; - tensor matrix_ac_5_cast_fp16 = matmul(transpose_x = matrix_ac_5_transpose_x_0, transpose_y = matrix_ac_5_transpose_y_0, x = transpose_55, y = transpose_56)[name = tensor("matrix_ac_5_cast_fp16")]; - tensor matrix_bd_11_begin_0 = const()[name = tensor("matrix_bd_11_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_11_end_0 = const()[name = tensor("matrix_bd_11_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_11_end_mask_0 = const()[name = tensor("matrix_bd_11_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_11_cast_fp16 = slice_by_index(begin = matrix_bd_11_begin_0, end = matrix_bd_11_end_0, end_mask = matrix_bd_11_end_mask_0, x = matrix_bd_9_cast_fp16)[name = tensor("matrix_bd_11_cast_fp16")]; - tensor var_766_cast_fp16 = add(x = matrix_ac_5_cast_fp16, y = matrix_bd_11_cast_fp16)[name = tensor("op_766_cast_fp16")]; - tensor _inversed_scores_9_y_0_to_fp16 = const()[name = tensor("_inversed_scores_9_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_9_cast_fp16 = mul(x = var_766_cast_fp16, y = _inversed_scores_9_y_0_to_fp16)[name = tensor("_inversed_scores_9_cast_fp16")]; - tensor scores_11_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_9_cast_fp16, cond = mask_11)[name = tensor("scores_11_cast_fp16")]; - tensor var_772_cast_fp16 = softmax(axis = var_23, x = scores_11_cast_fp16)[name = tensor("op_772_cast_fp16")]; - tensor input_137_cast_fp16 = select(a = var_11_to_fp16, b = var_772_cast_fp16, cond = mask_11)[name = tensor("input_137_cast_fp16")]; - tensor x_57_transpose_x_0 = const()[name = tensor("x_57_transpose_x_0"), val = tensor(false)]; - tensor x_57_transpose_y_0 = const()[name = tensor("x_57_transpose_y_0"), val = tensor(false)]; - tensor value_7_cast_fp16 = transpose(perm = value_7_perm_0, x = v_5_cast_fp16)[name = tensor("transpose_190")]; - tensor x_57_cast_fp16 = matmul(transpose_x = x_57_transpose_x_0, transpose_y = x_57_transpose_y_0, x = input_137_cast_fp16, y = value_7_cast_fp16)[name = tensor("x_57_cast_fp16")]; - tensor var_776_perm_0 = const()[name = tensor("op_776_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_777 = const()[name = tensor("op_777"), val = tensor([1, -1, 512])]; - tensor var_776_cast_fp16 = transpose(perm = var_776_perm_0, x = x_57_cast_fp16)[name = tensor("transpose_186")]; - tensor input_139_cast_fp16 = reshape(shape = var_777, x = var_776_cast_fp16)[name = tensor("input_139_cast_fp16")]; - tensor module_layers_2_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_2_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34038016)))]; - tensor module_layers_2_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_2_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34562368)))]; - tensor linear_25_cast_fp16 = linear(bias = module_layers_2_self_attn_linear_out_bias_to_fp16, weight = module_layers_2_self_attn_linear_out_weight_to_fp16, x = input_139_cast_fp16)[name = tensor("linear_25_cast_fp16")]; - tensor input_143_cast_fp16 = add(x = input_135_cast_fp16, y = linear_25_cast_fp16)[name = tensor("input_143_cast_fp16")]; - tensor x_61_axes_0 = const()[name = tensor("x_61_axes_0"), val = tensor([-1])]; - tensor module_layers_2_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_2_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34563456)))]; - tensor module_layers_2_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_2_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34564544)))]; - tensor x_61_cast_fp16 = layer_norm(axes = x_61_axes_0, beta = module_layers_2_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_2_norm_conv_weight_to_fp16, x = input_143_cast_fp16)[name = tensor("x_61_cast_fp16")]; - tensor input_145_perm_0 = const()[name = tensor("input_145_perm_0"), val = tensor([0, 2, 1])]; - tensor input_147_pad_type_0 = const()[name = tensor("input_147_pad_type_0"), val = tensor("valid")]; - tensor input_147_strides_0 = const()[name = tensor("input_147_strides_0"), val = tensor([1])]; - tensor input_147_pad_0 = const()[name = tensor("input_147_pad_0"), val = tensor([0, 0])]; - tensor input_147_dilations_0 = const()[name = tensor("input_147_dilations_0"), val = tensor([1])]; - tensor input_147_groups_0 = const()[name = tensor("input_147_groups_0"), val = tensor(1)]; - tensor module_layers_2_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_2_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34565632)))]; - tensor module_layers_2_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_2_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35614272)))]; - tensor input_145_cast_fp16 = transpose(perm = input_145_perm_0, x = x_61_cast_fp16)[name = tensor("transpose_185")]; - tensor input_147_cast_fp16 = conv(bias = module_layers_2_conv_pointwise_conv1_bias_to_fp16, dilations = input_147_dilations_0, groups = input_147_groups_0, pad = input_147_pad_0, pad_type = input_147_pad_type_0, strides = input_147_strides_0, weight = module_layers_2_conv_pointwise_conv1_weight_to_fp16, x = input_145_cast_fp16)[name = tensor("input_147_cast_fp16")]; - tensor x_63_split_num_splits_0 = const()[name = tensor("x_63_split_num_splits_0"), val = tensor(2)]; - tensor x_63_split_axis_0 = const()[name = tensor("x_63_split_axis_0"), val = tensor(1)]; - tensor x_63_split_cast_fp16_0, tensor x_63_split_cast_fp16_1 = split(axis = x_63_split_axis_0, num_splits = x_63_split_num_splits_0, x = input_147_cast_fp16)[name = tensor("x_63_split_cast_fp16")]; - tensor x_63_split_1_sigmoid_cast_fp16 = sigmoid(x = x_63_split_cast_fp16_1)[name = tensor("x_63_split_1_sigmoid_cast_fp16")]; - tensor x_63_cast_fp16 = mul(x = x_63_split_cast_fp16_0, y = x_63_split_1_sigmoid_cast_fp16)[name = tensor("x_63_cast_fp16")]; - tensor input_149_cast_fp16 = select(a = var_11_to_fp16, b = x_63_cast_fp16, cond = var_453)[name = tensor("input_149_cast_fp16")]; - tensor input_151_pad_0 = const()[name = tensor("input_151_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_151_mode_0 = const()[name = tensor("input_151_mode_0"), val = tensor("constant")]; - tensor const_93_to_fp16 = const()[name = tensor("const_93_to_fp16"), val = tensor(0x0p+0)]; - tensor input_151_cast_fp16 = pad(constant_val = const_93_to_fp16, mode = input_151_mode_0, pad = input_151_pad_0, x = input_149_cast_fp16)[name = tensor("input_151_cast_fp16")]; - tensor input_153_pad_type_0 = const()[name = tensor("input_153_pad_type_0"), val = tensor("valid")]; - tensor input_153_groups_0 = const()[name = tensor("input_153_groups_0"), val = tensor(512)]; - tensor input_153_strides_0 = const()[name = tensor("input_153_strides_0"), val = tensor([1])]; - tensor input_153_pad_0 = const()[name = tensor("input_153_pad_0"), val = tensor([0, 0])]; - tensor input_153_dilations_0 = const()[name = tensor("input_153_dilations_0"), val = tensor([1])]; - tensor const_238_to_fp16 = const()[name = tensor("const_238_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35616384)))]; - tensor const_239_to_fp16 = const()[name = tensor("const_239_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35625664)))]; - tensor input_155_cast_fp16 = conv(bias = const_239_to_fp16, dilations = input_153_dilations_0, groups = input_153_groups_0, pad = input_153_pad_0, pad_type = input_153_pad_type_0, strides = input_153_strides_0, weight = const_238_to_fp16, x = input_151_cast_fp16)[name = tensor("input_155_cast_fp16")]; - tensor input_157_cast_fp16 = silu(x = input_155_cast_fp16)[name = tensor("input_157_cast_fp16")]; - tensor x_65_pad_type_0 = const()[name = tensor("x_65_pad_type_0"), val = tensor("valid")]; - tensor x_65_strides_0 = const()[name = tensor("x_65_strides_0"), val = tensor([1])]; - tensor x_65_pad_0 = const()[name = tensor("x_65_pad_0"), val = tensor([0, 0])]; - tensor x_65_dilations_0 = const()[name = tensor("x_65_dilations_0"), val = tensor([1])]; - tensor x_65_groups_0 = const()[name = tensor("x_65_groups_0"), val = tensor(1)]; - tensor module_layers_2_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_2_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35626752)))]; - tensor module_layers_2_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_2_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36151104)))]; - tensor x_65_cast_fp16 = conv(bias = module_layers_2_conv_pointwise_conv2_bias_to_fp16, dilations = x_65_dilations_0, groups = x_65_groups_0, pad = x_65_pad_0, pad_type = x_65_pad_type_0, strides = x_65_strides_0, weight = module_layers_2_conv_pointwise_conv2_weight_to_fp16, x = input_157_cast_fp16)[name = tensor("x_65_cast_fp16")]; - tensor input_159_perm_0 = const()[name = tensor("input_159_perm_0"), val = tensor([0, 2, 1])]; - tensor input_159_cast_fp16 = transpose(perm = input_159_perm_0, x = x_65_cast_fp16)[name = tensor("transpose_184")]; - tensor input_161_cast_fp16 = add(x = input_143_cast_fp16, y = input_159_cast_fp16)[name = tensor("input_161_cast_fp16")]; - tensor input_163_axes_0 = const()[name = tensor("input_163_axes_0"), val = tensor([-1])]; - tensor module_layers_2_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_2_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36152192)))]; - tensor module_layers_2_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_2_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36153280)))]; - tensor input_163_cast_fp16 = layer_norm(axes = input_163_axes_0, beta = module_layers_2_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_2_norm_feed_forward2_weight_to_fp16, x = input_161_cast_fp16)[name = tensor("input_163_cast_fp16")]; - tensor module_layers_2_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_2_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36154368)))]; - tensor module_layers_2_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_2_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(38251584)))]; - tensor linear_26_cast_fp16 = linear(bias = module_layers_2_feed_forward2_linear1_bias_to_fp16, weight = module_layers_2_feed_forward2_linear1_weight_to_fp16, x = input_163_cast_fp16)[name = tensor("linear_26_cast_fp16")]; - tensor input_167_cast_fp16 = silu(x = linear_26_cast_fp16)[name = tensor("input_167_cast_fp16")]; - tensor module_layers_2_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_2_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(38255744)))]; - tensor module_layers_2_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_2_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40352960)))]; - tensor linear_27_cast_fp16 = linear(bias = module_layers_2_feed_forward2_linear2_bias_to_fp16, weight = module_layers_2_feed_forward2_linear2_weight_to_fp16, x = input_167_cast_fp16)[name = tensor("linear_27_cast_fp16")]; - tensor var_843_to_fp16 = const()[name = tensor("op_843_to_fp16"), val = tensor(0x1p-1)]; - tensor var_844_cast_fp16 = mul(x = linear_27_cast_fp16, y = var_843_to_fp16)[name = tensor("op_844_cast_fp16")]; - tensor input_173_cast_fp16 = add(x = input_161_cast_fp16, y = var_844_cast_fp16)[name = tensor("input_173_cast_fp16")]; - tensor input_175_axes_0 = const()[name = tensor("input_175_axes_0"), val = tensor([-1])]; - tensor module_layers_2_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_2_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40354048)))]; - tensor module_layers_2_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_2_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40355136)))]; - tensor input_175_cast_fp16 = layer_norm(axes = input_175_axes_0, beta = module_layers_2_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_2_norm_out_weight_to_fp16, x = input_173_cast_fp16)[name = tensor("input_175_cast_fp16")]; - tensor input_177_axes_0 = const()[name = tensor("input_177_axes_0"), val = tensor([-1])]; - tensor module_layers_3_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_3_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40356224)))]; - tensor module_layers_3_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_3_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40357312)))]; - tensor input_177_cast_fp16 = layer_norm(axes = input_177_axes_0, beta = module_layers_3_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_3_norm_feed_forward1_weight_to_fp16, x = input_175_cast_fp16)[name = tensor("input_177_cast_fp16")]; - tensor module_layers_3_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_3_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40358400)))]; - tensor module_layers_3_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_3_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(42455616)))]; - tensor linear_28_cast_fp16 = linear(bias = module_layers_3_feed_forward1_linear1_bias_to_fp16, weight = module_layers_3_feed_forward1_linear1_weight_to_fp16, x = input_177_cast_fp16)[name = tensor("linear_28_cast_fp16")]; - tensor input_181_cast_fp16 = silu(x = linear_28_cast_fp16)[name = tensor("input_181_cast_fp16")]; - tensor module_layers_3_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_3_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(42459776)))]; - tensor module_layers_3_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_3_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44556992)))]; - tensor linear_29_cast_fp16 = linear(bias = module_layers_3_feed_forward1_linear2_bias_to_fp16, weight = module_layers_3_feed_forward1_linear2_weight_to_fp16, x = input_181_cast_fp16)[name = tensor("linear_29_cast_fp16")]; - tensor var_874_to_fp16 = const()[name = tensor("op_874_to_fp16"), val = tensor(0x1p-1)]; - tensor var_875_cast_fp16 = mul(x = linear_29_cast_fp16, y = var_874_to_fp16)[name = tensor("op_875_cast_fp16")]; - tensor input_187_cast_fp16 = add(x = input_175_cast_fp16, y = var_875_cast_fp16)[name = tensor("input_187_cast_fp16")]; - tensor query_7_axes_0 = const()[name = tensor("query_7_axes_0"), val = tensor([-1])]; - tensor module_layers_3_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_3_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44558080)))]; - tensor module_layers_3_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_3_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44559168)))]; - tensor query_7_cast_fp16 = layer_norm(axes = query_7_axes_0, beta = module_layers_3_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_3_norm_self_att_weight_to_fp16, x = input_187_cast_fp16)[name = tensor("query_7_cast_fp16")]; - tensor module_layers_3_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_3_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44560256)))]; - tensor module_layers_3_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_3_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45084608)))]; - tensor linear_30_cast_fp16 = linear(bias = module_layers_3_self_attn_linear_q_bias_to_fp16, weight = module_layers_3_self_attn_linear_q_weight_to_fp16, x = query_7_cast_fp16)[name = tensor("linear_30_cast_fp16")]; - tensor var_892 = const()[name = tensor("op_892"), val = tensor([1, -1, 8, 64])]; - tensor q_19_cast_fp16 = reshape(shape = var_892, x = linear_30_cast_fp16)[name = tensor("q_19_cast_fp16")]; - tensor module_layers_3_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_3_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45085696)))]; - tensor module_layers_3_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_3_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45610048)))]; - tensor linear_31_cast_fp16 = linear(bias = module_layers_3_self_attn_linear_k_bias_to_fp16, weight = module_layers_3_self_attn_linear_k_weight_to_fp16, x = query_7_cast_fp16)[name = tensor("linear_31_cast_fp16")]; - tensor var_897 = const()[name = tensor("op_897"), val = tensor([1, -1, 8, 64])]; - tensor k_13_cast_fp16 = reshape(shape = var_897, x = linear_31_cast_fp16)[name = tensor("k_13_cast_fp16")]; - tensor module_layers_3_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_3_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45611136)))]; - tensor module_layers_3_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_3_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46135488)))]; - tensor linear_32_cast_fp16 = linear(bias = module_layers_3_self_attn_linear_v_bias_to_fp16, weight = module_layers_3_self_attn_linear_v_weight_to_fp16, x = query_7_cast_fp16)[name = tensor("linear_32_cast_fp16")]; - tensor var_902 = const()[name = tensor("op_902"), val = tensor([1, -1, 8, 64])]; - tensor v_7_cast_fp16 = reshape(shape = var_902, x = linear_32_cast_fp16)[name = tensor("v_7_cast_fp16")]; - tensor value_9_perm_0 = const()[name = tensor("value_9_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_3_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_3_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46136576)))]; - tensor var_914_cast_fp16 = add(x = q_19_cast_fp16, y = module_layers_3_self_attn_pos_bias_u_to_fp16)[name = tensor("op_914_cast_fp16")]; - tensor module_layers_3_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_3_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46137664)))]; - tensor var_916_cast_fp16 = add(x = q_19_cast_fp16, y = module_layers_3_self_attn_pos_bias_v_to_fp16)[name = tensor("op_916_cast_fp16")]; - tensor q_with_bias_v_7_perm_0 = const()[name = tensor("q_with_bias_v_7_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_73_transpose_x_0 = const()[name = tensor("x_73_transpose_x_0"), val = tensor(false)]; - tensor x_73_transpose_y_0 = const()[name = tensor("x_73_transpose_y_0"), val = tensor(false)]; - tensor var_918_to_fp16 = const()[name = tensor("op_918_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46138752)))]; - tensor q_with_bias_v_7_cast_fp16 = transpose(perm = q_with_bias_v_7_perm_0, x = var_916_cast_fp16)[name = tensor("transpose_182")]; - tensor x_73_cast_fp16 = matmul(transpose_x = x_73_transpose_x_0, transpose_y = x_73_transpose_y_0, x = q_with_bias_v_7_cast_fp16, y = var_918_to_fp16)[name = tensor("x_73_cast_fp16")]; - tensor x_75_pad_0 = const()[name = tensor("x_75_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_75_mode_0 = const()[name = tensor("x_75_mode_0"), val = tensor("constant")]; - tensor const_100_to_fp16 = const()[name = tensor("const_100_to_fp16"), val = tensor(0x0p+0)]; - tensor x_75_cast_fp16 = pad(constant_val = const_100_to_fp16, mode = x_75_mode_0, pad = x_75_pad_0, x = x_73_cast_fp16)[name = tensor("x_75_cast_fp16")]; - tensor var_926 = const()[name = tensor("op_926"), val = tensor([1, 8, -1, 188])]; - tensor x_77_cast_fp16 = reshape(shape = var_926, x = x_75_cast_fp16)[name = tensor("x_77_cast_fp16")]; - tensor var_930_begin_0 = const()[name = tensor("op_930_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_930_end_0 = const()[name = tensor("op_930_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_930_end_mask_0 = const()[name = tensor("op_930_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_930_cast_fp16 = slice_by_index(begin = var_930_begin_0, end = var_930_end_0, end_mask = var_930_end_mask_0, x = x_77_cast_fp16)[name = tensor("op_930_cast_fp16")]; - tensor var_931 = const()[name = tensor("op_931"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_13_cast_fp16 = reshape(shape = var_931, x = var_930_cast_fp16)[name = tensor("matrix_bd_13_cast_fp16")]; - tensor matrix_ac_7_transpose_x_0 = const()[name = tensor("matrix_ac_7_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_7_transpose_y_0 = const()[name = tensor("matrix_ac_7_transpose_y_0"), val = tensor(false)]; - tensor transpose_57_perm_0 = const()[name = tensor("transpose_57_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_58_perm_0 = const()[name = tensor("transpose_58_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_58 = transpose(perm = transpose_58_perm_0, x = k_13_cast_fp16)[name = tensor("transpose_180")]; - tensor transpose_57 = transpose(perm = transpose_57_perm_0, x = var_914_cast_fp16)[name = tensor("transpose_181")]; - tensor matrix_ac_7_cast_fp16 = matmul(transpose_x = matrix_ac_7_transpose_x_0, transpose_y = matrix_ac_7_transpose_y_0, x = transpose_57, y = transpose_58)[name = tensor("matrix_ac_7_cast_fp16")]; - tensor matrix_bd_15_begin_0 = const()[name = tensor("matrix_bd_15_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_15_end_0 = const()[name = tensor("matrix_bd_15_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_15_end_mask_0 = const()[name = tensor("matrix_bd_15_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_15_cast_fp16 = slice_by_index(begin = matrix_bd_15_begin_0, end = matrix_bd_15_end_0, end_mask = matrix_bd_15_end_mask_0, x = matrix_bd_13_cast_fp16)[name = tensor("matrix_bd_15_cast_fp16")]; - tensor var_940_cast_fp16 = add(x = matrix_ac_7_cast_fp16, y = matrix_bd_15_cast_fp16)[name = tensor("op_940_cast_fp16")]; - tensor _inversed_scores_13_y_0_to_fp16 = const()[name = tensor("_inversed_scores_13_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_13_cast_fp16 = mul(x = var_940_cast_fp16, y = _inversed_scores_13_y_0_to_fp16)[name = tensor("_inversed_scores_13_cast_fp16")]; - tensor scores_15_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_13_cast_fp16, cond = mask_11)[name = tensor("scores_15_cast_fp16")]; - tensor var_946_cast_fp16 = softmax(axis = var_23, x = scores_15_cast_fp16)[name = tensor("op_946_cast_fp16")]; - tensor input_189_cast_fp16 = select(a = var_11_to_fp16, b = var_946_cast_fp16, cond = mask_11)[name = tensor("input_189_cast_fp16")]; - tensor x_79_transpose_x_0 = const()[name = tensor("x_79_transpose_x_0"), val = tensor(false)]; - tensor x_79_transpose_y_0 = const()[name = tensor("x_79_transpose_y_0"), val = tensor(false)]; - tensor value_9_cast_fp16 = transpose(perm = value_9_perm_0, x = v_7_cast_fp16)[name = tensor("transpose_183")]; - tensor x_79_cast_fp16 = matmul(transpose_x = x_79_transpose_x_0, transpose_y = x_79_transpose_y_0, x = input_189_cast_fp16, y = value_9_cast_fp16)[name = tensor("x_79_cast_fp16")]; - tensor var_950_perm_0 = const()[name = tensor("op_950_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_951 = const()[name = tensor("op_951"), val = tensor([1, -1, 512])]; - tensor var_950_cast_fp16 = transpose(perm = var_950_perm_0, x = x_79_cast_fp16)[name = tensor("transpose_179")]; - tensor input_191_cast_fp16 = reshape(shape = var_951, x = var_950_cast_fp16)[name = tensor("input_191_cast_fp16")]; - tensor module_layers_3_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_3_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46522816)))]; - tensor module_layers_3_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_3_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47047168)))]; - tensor linear_34_cast_fp16 = linear(bias = module_layers_3_self_attn_linear_out_bias_to_fp16, weight = module_layers_3_self_attn_linear_out_weight_to_fp16, x = input_191_cast_fp16)[name = tensor("linear_34_cast_fp16")]; - tensor input_195_cast_fp16 = add(x = input_187_cast_fp16, y = linear_34_cast_fp16)[name = tensor("input_195_cast_fp16")]; - tensor x_83_axes_0 = const()[name = tensor("x_83_axes_0"), val = tensor([-1])]; - tensor module_layers_3_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_3_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47048256)))]; - tensor module_layers_3_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_3_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47049344)))]; - tensor x_83_cast_fp16 = layer_norm(axes = x_83_axes_0, beta = module_layers_3_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_3_norm_conv_weight_to_fp16, x = input_195_cast_fp16)[name = tensor("x_83_cast_fp16")]; - tensor input_197_perm_0 = const()[name = tensor("input_197_perm_0"), val = tensor([0, 2, 1])]; - tensor input_199_pad_type_0 = const()[name = tensor("input_199_pad_type_0"), val = tensor("valid")]; - tensor input_199_strides_0 = const()[name = tensor("input_199_strides_0"), val = tensor([1])]; - tensor input_199_pad_0 = const()[name = tensor("input_199_pad_0"), val = tensor([0, 0])]; - tensor input_199_dilations_0 = const()[name = tensor("input_199_dilations_0"), val = tensor([1])]; - tensor input_199_groups_0 = const()[name = tensor("input_199_groups_0"), val = tensor(1)]; - tensor module_layers_3_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_3_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47050432)))]; - tensor module_layers_3_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_3_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48099072)))]; - tensor input_197_cast_fp16 = transpose(perm = input_197_perm_0, x = x_83_cast_fp16)[name = tensor("transpose_178")]; - tensor input_199_cast_fp16 = conv(bias = module_layers_3_conv_pointwise_conv1_bias_to_fp16, dilations = input_199_dilations_0, groups = input_199_groups_0, pad = input_199_pad_0, pad_type = input_199_pad_type_0, strides = input_199_strides_0, weight = module_layers_3_conv_pointwise_conv1_weight_to_fp16, x = input_197_cast_fp16)[name = tensor("input_199_cast_fp16")]; - tensor x_85_split_num_splits_0 = const()[name = tensor("x_85_split_num_splits_0"), val = tensor(2)]; - tensor x_85_split_axis_0 = const()[name = tensor("x_85_split_axis_0"), val = tensor(1)]; - tensor x_85_split_cast_fp16_0, tensor x_85_split_cast_fp16_1 = split(axis = x_85_split_axis_0, num_splits = x_85_split_num_splits_0, x = input_199_cast_fp16)[name = tensor("x_85_split_cast_fp16")]; - tensor x_85_split_1_sigmoid_cast_fp16 = sigmoid(x = x_85_split_cast_fp16_1)[name = tensor("x_85_split_1_sigmoid_cast_fp16")]; - tensor x_85_cast_fp16 = mul(x = x_85_split_cast_fp16_0, y = x_85_split_1_sigmoid_cast_fp16)[name = tensor("x_85_cast_fp16")]; - tensor input_201_cast_fp16 = select(a = var_11_to_fp16, b = x_85_cast_fp16, cond = var_453)[name = tensor("input_201_cast_fp16")]; - tensor input_203_pad_0 = const()[name = tensor("input_203_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_203_mode_0 = const()[name = tensor("input_203_mode_0"), val = tensor("constant")]; - tensor const_103_to_fp16 = const()[name = tensor("const_103_to_fp16"), val = tensor(0x0p+0)]; - tensor input_203_cast_fp16 = pad(constant_val = const_103_to_fp16, mode = input_203_mode_0, pad = input_203_pad_0, x = input_201_cast_fp16)[name = tensor("input_203_cast_fp16")]; - tensor input_205_pad_type_0 = const()[name = tensor("input_205_pad_type_0"), val = tensor("valid")]; - tensor input_205_groups_0 = const()[name = tensor("input_205_groups_0"), val = tensor(512)]; - tensor input_205_strides_0 = const()[name = tensor("input_205_strides_0"), val = tensor([1])]; - tensor input_205_pad_0 = const()[name = tensor("input_205_pad_0"), val = tensor([0, 0])]; - tensor input_205_dilations_0 = const()[name = tensor("input_205_dilations_0"), val = tensor([1])]; - tensor const_240_to_fp16 = const()[name = tensor("const_240_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48101184)))]; - tensor const_241_to_fp16 = const()[name = tensor("const_241_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48110464)))]; - tensor input_207_cast_fp16 = conv(bias = const_241_to_fp16, dilations = input_205_dilations_0, groups = input_205_groups_0, pad = input_205_pad_0, pad_type = input_205_pad_type_0, strides = input_205_strides_0, weight = const_240_to_fp16, x = input_203_cast_fp16)[name = tensor("input_207_cast_fp16")]; - tensor input_209_cast_fp16 = silu(x = input_207_cast_fp16)[name = tensor("input_209_cast_fp16")]; - tensor x_87_pad_type_0 = const()[name = tensor("x_87_pad_type_0"), val = tensor("valid")]; - tensor x_87_strides_0 = const()[name = tensor("x_87_strides_0"), val = tensor([1])]; - tensor x_87_pad_0 = const()[name = tensor("x_87_pad_0"), val = tensor([0, 0])]; - tensor x_87_dilations_0 = const()[name = tensor("x_87_dilations_0"), val = tensor([1])]; - tensor x_87_groups_0 = const()[name = tensor("x_87_groups_0"), val = tensor(1)]; - tensor module_layers_3_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_3_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48111552)))]; - tensor module_layers_3_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_3_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48635904)))]; - tensor x_87_cast_fp16 = conv(bias = module_layers_3_conv_pointwise_conv2_bias_to_fp16, dilations = x_87_dilations_0, groups = x_87_groups_0, pad = x_87_pad_0, pad_type = x_87_pad_type_0, strides = x_87_strides_0, weight = module_layers_3_conv_pointwise_conv2_weight_to_fp16, x = input_209_cast_fp16)[name = tensor("x_87_cast_fp16")]; - tensor input_211_perm_0 = const()[name = tensor("input_211_perm_0"), val = tensor([0, 2, 1])]; - tensor input_211_cast_fp16 = transpose(perm = input_211_perm_0, x = x_87_cast_fp16)[name = tensor("transpose_177")]; - tensor input_213_cast_fp16 = add(x = input_195_cast_fp16, y = input_211_cast_fp16)[name = tensor("input_213_cast_fp16")]; - tensor input_215_axes_0 = const()[name = tensor("input_215_axes_0"), val = tensor([-1])]; - tensor module_layers_3_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_3_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48636992)))]; - tensor module_layers_3_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_3_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48638080)))]; - tensor input_215_cast_fp16 = layer_norm(axes = input_215_axes_0, beta = module_layers_3_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_3_norm_feed_forward2_weight_to_fp16, x = input_213_cast_fp16)[name = tensor("input_215_cast_fp16")]; - tensor module_layers_3_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_3_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48639168)))]; - tensor module_layers_3_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_3_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(50736384)))]; - tensor linear_35_cast_fp16 = linear(bias = module_layers_3_feed_forward2_linear1_bias_to_fp16, weight = module_layers_3_feed_forward2_linear1_weight_to_fp16, x = input_215_cast_fp16)[name = tensor("linear_35_cast_fp16")]; - tensor input_219_cast_fp16 = silu(x = linear_35_cast_fp16)[name = tensor("input_219_cast_fp16")]; - tensor module_layers_3_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_3_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(50740544)))]; - tensor module_layers_3_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_3_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52837760)))]; - tensor linear_36_cast_fp16 = linear(bias = module_layers_3_feed_forward2_linear2_bias_to_fp16, weight = module_layers_3_feed_forward2_linear2_weight_to_fp16, x = input_219_cast_fp16)[name = tensor("linear_36_cast_fp16")]; - tensor var_1017_to_fp16 = const()[name = tensor("op_1017_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1018_cast_fp16 = mul(x = linear_36_cast_fp16, y = var_1017_to_fp16)[name = tensor("op_1018_cast_fp16")]; - tensor input_225_cast_fp16 = add(x = input_213_cast_fp16, y = var_1018_cast_fp16)[name = tensor("input_225_cast_fp16")]; - tensor input_227_axes_0 = const()[name = tensor("input_227_axes_0"), val = tensor([-1])]; - tensor module_layers_3_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_3_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52838848)))]; - tensor module_layers_3_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_3_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52839936)))]; - tensor input_227_cast_fp16 = layer_norm(axes = input_227_axes_0, beta = module_layers_3_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_3_norm_out_weight_to_fp16, x = input_225_cast_fp16)[name = tensor("input_227_cast_fp16")]; - tensor input_229_axes_0 = const()[name = tensor("input_229_axes_0"), val = tensor([-1])]; - tensor module_layers_4_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_4_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52841024)))]; - tensor module_layers_4_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_4_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52842112)))]; - tensor input_229_cast_fp16 = layer_norm(axes = input_229_axes_0, beta = module_layers_4_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_4_norm_feed_forward1_weight_to_fp16, x = input_227_cast_fp16)[name = tensor("input_229_cast_fp16")]; - tensor module_layers_4_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_4_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52843200)))]; - tensor module_layers_4_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_4_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(54940416)))]; - tensor linear_37_cast_fp16 = linear(bias = module_layers_4_feed_forward1_linear1_bias_to_fp16, weight = module_layers_4_feed_forward1_linear1_weight_to_fp16, x = input_229_cast_fp16)[name = tensor("linear_37_cast_fp16")]; - tensor input_233_cast_fp16 = silu(x = linear_37_cast_fp16)[name = tensor("input_233_cast_fp16")]; - tensor module_layers_4_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_4_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(54944576)))]; - tensor module_layers_4_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_4_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57041792)))]; - tensor linear_38_cast_fp16 = linear(bias = module_layers_4_feed_forward1_linear2_bias_to_fp16, weight = module_layers_4_feed_forward1_linear2_weight_to_fp16, x = input_233_cast_fp16)[name = tensor("linear_38_cast_fp16")]; - tensor var_1048_to_fp16 = const()[name = tensor("op_1048_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1049_cast_fp16 = mul(x = linear_38_cast_fp16, y = var_1048_to_fp16)[name = tensor("op_1049_cast_fp16")]; - tensor input_239_cast_fp16 = add(x = input_227_cast_fp16, y = var_1049_cast_fp16)[name = tensor("input_239_cast_fp16")]; - tensor query_9_axes_0 = const()[name = tensor("query_9_axes_0"), val = tensor([-1])]; - tensor module_layers_4_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_4_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57042880)))]; - tensor module_layers_4_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_4_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57043968)))]; - tensor query_9_cast_fp16 = layer_norm(axes = query_9_axes_0, beta = module_layers_4_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_4_norm_self_att_weight_to_fp16, x = input_239_cast_fp16)[name = tensor("query_9_cast_fp16")]; - tensor module_layers_4_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_4_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57045056)))]; - tensor module_layers_4_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_4_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57569408)))]; - tensor linear_39_cast_fp16 = linear(bias = module_layers_4_self_attn_linear_q_bias_to_fp16, weight = module_layers_4_self_attn_linear_q_weight_to_fp16, x = query_9_cast_fp16)[name = tensor("linear_39_cast_fp16")]; - tensor var_1066 = const()[name = tensor("op_1066"), val = tensor([1, -1, 8, 64])]; - tensor q_25_cast_fp16 = reshape(shape = var_1066, x = linear_39_cast_fp16)[name = tensor("q_25_cast_fp16")]; - tensor module_layers_4_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_4_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57570496)))]; - tensor module_layers_4_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_4_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58094848)))]; - tensor linear_40_cast_fp16 = linear(bias = module_layers_4_self_attn_linear_k_bias_to_fp16, weight = module_layers_4_self_attn_linear_k_weight_to_fp16, x = query_9_cast_fp16)[name = tensor("linear_40_cast_fp16")]; - tensor var_1071 = const()[name = tensor("op_1071"), val = tensor([1, -1, 8, 64])]; - tensor k_17_cast_fp16 = reshape(shape = var_1071, x = linear_40_cast_fp16)[name = tensor("k_17_cast_fp16")]; - tensor module_layers_4_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_4_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58095936)))]; - tensor module_layers_4_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_4_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58620288)))]; - tensor linear_41_cast_fp16 = linear(bias = module_layers_4_self_attn_linear_v_bias_to_fp16, weight = module_layers_4_self_attn_linear_v_weight_to_fp16, x = query_9_cast_fp16)[name = tensor("linear_41_cast_fp16")]; - tensor var_1076 = const()[name = tensor("op_1076"), val = tensor([1, -1, 8, 64])]; - tensor v_9_cast_fp16 = reshape(shape = var_1076, x = linear_41_cast_fp16)[name = tensor("v_9_cast_fp16")]; - tensor value_11_perm_0 = const()[name = tensor("value_11_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_4_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_4_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58621376)))]; - tensor var_1088_cast_fp16 = add(x = q_25_cast_fp16, y = module_layers_4_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1088_cast_fp16")]; - tensor module_layers_4_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_4_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58622464)))]; - tensor var_1090_cast_fp16 = add(x = q_25_cast_fp16, y = module_layers_4_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1090_cast_fp16")]; - tensor q_with_bias_v_9_perm_0 = const()[name = tensor("q_with_bias_v_9_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_95_transpose_x_0 = const()[name = tensor("x_95_transpose_x_0"), val = tensor(false)]; - tensor x_95_transpose_y_0 = const()[name = tensor("x_95_transpose_y_0"), val = tensor(false)]; - tensor var_1092_to_fp16 = const()[name = tensor("op_1092_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58623552)))]; - tensor q_with_bias_v_9_cast_fp16 = transpose(perm = q_with_bias_v_9_perm_0, x = var_1090_cast_fp16)[name = tensor("transpose_175")]; - tensor x_95_cast_fp16 = matmul(transpose_x = x_95_transpose_x_0, transpose_y = x_95_transpose_y_0, x = q_with_bias_v_9_cast_fp16, y = var_1092_to_fp16)[name = tensor("x_95_cast_fp16")]; - tensor x_97_pad_0 = const()[name = tensor("x_97_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_97_mode_0 = const()[name = tensor("x_97_mode_0"), val = tensor("constant")]; - tensor const_110_to_fp16 = const()[name = tensor("const_110_to_fp16"), val = tensor(0x0p+0)]; - tensor x_97_cast_fp16 = pad(constant_val = const_110_to_fp16, mode = x_97_mode_0, pad = x_97_pad_0, x = x_95_cast_fp16)[name = tensor("x_97_cast_fp16")]; - tensor var_1100 = const()[name = tensor("op_1100"), val = tensor([1, 8, -1, 188])]; - tensor x_99_cast_fp16 = reshape(shape = var_1100, x = x_97_cast_fp16)[name = tensor("x_99_cast_fp16")]; - tensor var_1104_begin_0 = const()[name = tensor("op_1104_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_1104_end_0 = const()[name = tensor("op_1104_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_1104_end_mask_0 = const()[name = tensor("op_1104_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_1104_cast_fp16 = slice_by_index(begin = var_1104_begin_0, end = var_1104_end_0, end_mask = var_1104_end_mask_0, x = x_99_cast_fp16)[name = tensor("op_1104_cast_fp16")]; - tensor var_1105 = const()[name = tensor("op_1105"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_17_cast_fp16 = reshape(shape = var_1105, x = var_1104_cast_fp16)[name = tensor("matrix_bd_17_cast_fp16")]; - tensor matrix_ac_9_transpose_x_0 = const()[name = tensor("matrix_ac_9_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_9_transpose_y_0 = const()[name = tensor("matrix_ac_9_transpose_y_0"), val = tensor(false)]; - tensor transpose_59_perm_0 = const()[name = tensor("transpose_59_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_60_perm_0 = const()[name = tensor("transpose_60_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_60 = transpose(perm = transpose_60_perm_0, x = k_17_cast_fp16)[name = tensor("transpose_173")]; - tensor transpose_59 = transpose(perm = transpose_59_perm_0, x = var_1088_cast_fp16)[name = tensor("transpose_174")]; - tensor matrix_ac_9_cast_fp16 = matmul(transpose_x = matrix_ac_9_transpose_x_0, transpose_y = matrix_ac_9_transpose_y_0, x = transpose_59, y = transpose_60)[name = tensor("matrix_ac_9_cast_fp16")]; - tensor matrix_bd_19_begin_0 = const()[name = tensor("matrix_bd_19_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_19_end_0 = const()[name = tensor("matrix_bd_19_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_19_end_mask_0 = const()[name = tensor("matrix_bd_19_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_19_cast_fp16 = slice_by_index(begin = matrix_bd_19_begin_0, end = matrix_bd_19_end_0, end_mask = matrix_bd_19_end_mask_0, x = matrix_bd_17_cast_fp16)[name = tensor("matrix_bd_19_cast_fp16")]; - tensor var_1114_cast_fp16 = add(x = matrix_ac_9_cast_fp16, y = matrix_bd_19_cast_fp16)[name = tensor("op_1114_cast_fp16")]; - tensor _inversed_scores_17_y_0_to_fp16 = const()[name = tensor("_inversed_scores_17_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_17_cast_fp16 = mul(x = var_1114_cast_fp16, y = _inversed_scores_17_y_0_to_fp16)[name = tensor("_inversed_scores_17_cast_fp16")]; - tensor scores_19_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_17_cast_fp16, cond = mask_11)[name = tensor("scores_19_cast_fp16")]; - tensor var_1120_cast_fp16 = softmax(axis = var_23, x = scores_19_cast_fp16)[name = tensor("op_1120_cast_fp16")]; - tensor input_241_cast_fp16 = select(a = var_11_to_fp16, b = var_1120_cast_fp16, cond = mask_11)[name = tensor("input_241_cast_fp16")]; - tensor x_101_transpose_x_0 = const()[name = tensor("x_101_transpose_x_0"), val = tensor(false)]; - tensor x_101_transpose_y_0 = const()[name = tensor("x_101_transpose_y_0"), val = tensor(false)]; - tensor value_11_cast_fp16 = transpose(perm = value_11_perm_0, x = v_9_cast_fp16)[name = tensor("transpose_176")]; - tensor x_101_cast_fp16 = matmul(transpose_x = x_101_transpose_x_0, transpose_y = x_101_transpose_y_0, x = input_241_cast_fp16, y = value_11_cast_fp16)[name = tensor("x_101_cast_fp16")]; - tensor var_1124_perm_0 = const()[name = tensor("op_1124_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_1125 = const()[name = tensor("op_1125"), val = tensor([1, -1, 512])]; - tensor var_1124_cast_fp16 = transpose(perm = var_1124_perm_0, x = x_101_cast_fp16)[name = tensor("transpose_172")]; - tensor input_243_cast_fp16 = reshape(shape = var_1125, x = var_1124_cast_fp16)[name = tensor("input_243_cast_fp16")]; - tensor module_layers_4_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_4_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59007616)))]; - tensor module_layers_4_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_4_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59531968)))]; - tensor linear_43_cast_fp16 = linear(bias = module_layers_4_self_attn_linear_out_bias_to_fp16, weight = module_layers_4_self_attn_linear_out_weight_to_fp16, x = input_243_cast_fp16)[name = tensor("linear_43_cast_fp16")]; - tensor input_247_cast_fp16 = add(x = input_239_cast_fp16, y = linear_43_cast_fp16)[name = tensor("input_247_cast_fp16")]; - tensor x_105_axes_0 = const()[name = tensor("x_105_axes_0"), val = tensor([-1])]; - tensor module_layers_4_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_4_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59533056)))]; - tensor module_layers_4_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_4_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59534144)))]; - tensor x_105_cast_fp16 = layer_norm(axes = x_105_axes_0, beta = module_layers_4_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_4_norm_conv_weight_to_fp16, x = input_247_cast_fp16)[name = tensor("x_105_cast_fp16")]; - tensor input_249_perm_0 = const()[name = tensor("input_249_perm_0"), val = tensor([0, 2, 1])]; - tensor input_251_pad_type_0 = const()[name = tensor("input_251_pad_type_0"), val = tensor("valid")]; - tensor input_251_strides_0 = const()[name = tensor("input_251_strides_0"), val = tensor([1])]; - tensor input_251_pad_0 = const()[name = tensor("input_251_pad_0"), val = tensor([0, 0])]; - tensor input_251_dilations_0 = const()[name = tensor("input_251_dilations_0"), val = tensor([1])]; - tensor input_251_groups_0 = const()[name = tensor("input_251_groups_0"), val = tensor(1)]; - tensor module_layers_4_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_4_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59535232)))]; - tensor module_layers_4_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_4_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60583872)))]; - tensor input_249_cast_fp16 = transpose(perm = input_249_perm_0, x = x_105_cast_fp16)[name = tensor("transpose_171")]; - tensor input_251_cast_fp16 = conv(bias = module_layers_4_conv_pointwise_conv1_bias_to_fp16, dilations = input_251_dilations_0, groups = input_251_groups_0, pad = input_251_pad_0, pad_type = input_251_pad_type_0, strides = input_251_strides_0, weight = module_layers_4_conv_pointwise_conv1_weight_to_fp16, x = input_249_cast_fp16)[name = tensor("input_251_cast_fp16")]; - tensor x_107_split_num_splits_0 = const()[name = tensor("x_107_split_num_splits_0"), val = tensor(2)]; - tensor x_107_split_axis_0 = const()[name = tensor("x_107_split_axis_0"), val = tensor(1)]; - tensor x_107_split_cast_fp16_0, tensor x_107_split_cast_fp16_1 = split(axis = x_107_split_axis_0, num_splits = x_107_split_num_splits_0, x = input_251_cast_fp16)[name = tensor("x_107_split_cast_fp16")]; - tensor x_107_split_1_sigmoid_cast_fp16 = sigmoid(x = x_107_split_cast_fp16_1)[name = tensor("x_107_split_1_sigmoid_cast_fp16")]; - tensor x_107_cast_fp16 = mul(x = x_107_split_cast_fp16_0, y = x_107_split_1_sigmoid_cast_fp16)[name = tensor("x_107_cast_fp16")]; - tensor input_253_cast_fp16 = select(a = var_11_to_fp16, b = x_107_cast_fp16, cond = var_453)[name = tensor("input_253_cast_fp16")]; - tensor input_255_pad_0 = const()[name = tensor("input_255_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_255_mode_0 = const()[name = tensor("input_255_mode_0"), val = tensor("constant")]; - tensor const_113_to_fp16 = const()[name = tensor("const_113_to_fp16"), val = tensor(0x0p+0)]; - tensor input_255_cast_fp16 = pad(constant_val = const_113_to_fp16, mode = input_255_mode_0, pad = input_255_pad_0, x = input_253_cast_fp16)[name = tensor("input_255_cast_fp16")]; - tensor input_257_pad_type_0 = const()[name = tensor("input_257_pad_type_0"), val = tensor("valid")]; - tensor input_257_groups_0 = const()[name = tensor("input_257_groups_0"), val = tensor(512)]; - tensor input_257_strides_0 = const()[name = tensor("input_257_strides_0"), val = tensor([1])]; - tensor input_257_pad_0 = const()[name = tensor("input_257_pad_0"), val = tensor([0, 0])]; - tensor input_257_dilations_0 = const()[name = tensor("input_257_dilations_0"), val = tensor([1])]; - tensor const_242_to_fp16 = const()[name = tensor("const_242_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60585984)))]; - tensor const_243_to_fp16 = const()[name = tensor("const_243_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60595264)))]; - tensor input_259_cast_fp16 = conv(bias = const_243_to_fp16, dilations = input_257_dilations_0, groups = input_257_groups_0, pad = input_257_pad_0, pad_type = input_257_pad_type_0, strides = input_257_strides_0, weight = const_242_to_fp16, x = input_255_cast_fp16)[name = tensor("input_259_cast_fp16")]; - tensor input_261_cast_fp16 = silu(x = input_259_cast_fp16)[name = tensor("input_261_cast_fp16")]; - tensor x_109_pad_type_0 = const()[name = tensor("x_109_pad_type_0"), val = tensor("valid")]; - tensor x_109_strides_0 = const()[name = tensor("x_109_strides_0"), val = tensor([1])]; - tensor x_109_pad_0 = const()[name = tensor("x_109_pad_0"), val = tensor([0, 0])]; - tensor x_109_dilations_0 = const()[name = tensor("x_109_dilations_0"), val = tensor([1])]; - tensor x_109_groups_0 = const()[name = tensor("x_109_groups_0"), val = tensor(1)]; - tensor module_layers_4_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_4_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60596352)))]; - tensor module_layers_4_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_4_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61120704)))]; - tensor x_109_cast_fp16 = conv(bias = module_layers_4_conv_pointwise_conv2_bias_to_fp16, dilations = x_109_dilations_0, groups = x_109_groups_0, pad = x_109_pad_0, pad_type = x_109_pad_type_0, strides = x_109_strides_0, weight = module_layers_4_conv_pointwise_conv2_weight_to_fp16, x = input_261_cast_fp16)[name = tensor("x_109_cast_fp16")]; - tensor input_263_perm_0 = const()[name = tensor("input_263_perm_0"), val = tensor([0, 2, 1])]; - tensor input_263_cast_fp16 = transpose(perm = input_263_perm_0, x = x_109_cast_fp16)[name = tensor("transpose_170")]; - tensor input_265_cast_fp16 = add(x = input_247_cast_fp16, y = input_263_cast_fp16)[name = tensor("input_265_cast_fp16")]; - tensor input_267_axes_0 = const()[name = tensor("input_267_axes_0"), val = tensor([-1])]; - tensor module_layers_4_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_4_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61121792)))]; - tensor module_layers_4_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_4_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61122880)))]; - tensor input_267_cast_fp16 = layer_norm(axes = input_267_axes_0, beta = module_layers_4_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_4_norm_feed_forward2_weight_to_fp16, x = input_265_cast_fp16)[name = tensor("input_267_cast_fp16")]; - tensor module_layers_4_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_4_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61123968)))]; - tensor module_layers_4_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_4_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(63221184)))]; - tensor linear_44_cast_fp16 = linear(bias = module_layers_4_feed_forward2_linear1_bias_to_fp16, weight = module_layers_4_feed_forward2_linear1_weight_to_fp16, x = input_267_cast_fp16)[name = tensor("linear_44_cast_fp16")]; - tensor input_271_cast_fp16 = silu(x = linear_44_cast_fp16)[name = tensor("input_271_cast_fp16")]; - tensor module_layers_4_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_4_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(63225344)))]; - tensor module_layers_4_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_4_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65322560)))]; - tensor linear_45_cast_fp16 = linear(bias = module_layers_4_feed_forward2_linear2_bias_to_fp16, weight = module_layers_4_feed_forward2_linear2_weight_to_fp16, x = input_271_cast_fp16)[name = tensor("linear_45_cast_fp16")]; - tensor var_1191_to_fp16 = const()[name = tensor("op_1191_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1192_cast_fp16 = mul(x = linear_45_cast_fp16, y = var_1191_to_fp16)[name = tensor("op_1192_cast_fp16")]; - tensor input_277_cast_fp16 = add(x = input_265_cast_fp16, y = var_1192_cast_fp16)[name = tensor("input_277_cast_fp16")]; - tensor input_279_axes_0 = const()[name = tensor("input_279_axes_0"), val = tensor([-1])]; - tensor module_layers_4_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_4_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65323648)))]; - tensor module_layers_4_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_4_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65324736)))]; - tensor input_279_cast_fp16 = layer_norm(axes = input_279_axes_0, beta = module_layers_4_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_4_norm_out_weight_to_fp16, x = input_277_cast_fp16)[name = tensor("input_279_cast_fp16")]; - tensor input_281_axes_0 = const()[name = tensor("input_281_axes_0"), val = tensor([-1])]; - tensor module_layers_5_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_5_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65325824)))]; - tensor module_layers_5_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_5_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65326912)))]; - tensor input_281_cast_fp16 = layer_norm(axes = input_281_axes_0, beta = module_layers_5_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_5_norm_feed_forward1_weight_to_fp16, x = input_279_cast_fp16)[name = tensor("input_281_cast_fp16")]; - tensor module_layers_5_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_5_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65328000)))]; - tensor module_layers_5_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_5_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67425216)))]; - tensor linear_46_cast_fp16 = linear(bias = module_layers_5_feed_forward1_linear1_bias_to_fp16, weight = module_layers_5_feed_forward1_linear1_weight_to_fp16, x = input_281_cast_fp16)[name = tensor("linear_46_cast_fp16")]; - tensor input_285_cast_fp16 = silu(x = linear_46_cast_fp16)[name = tensor("input_285_cast_fp16")]; - tensor module_layers_5_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_5_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67429376)))]; - tensor module_layers_5_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_5_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(69526592)))]; - tensor linear_47_cast_fp16 = linear(bias = module_layers_5_feed_forward1_linear2_bias_to_fp16, weight = module_layers_5_feed_forward1_linear2_weight_to_fp16, x = input_285_cast_fp16)[name = tensor("linear_47_cast_fp16")]; - tensor var_1222_to_fp16 = const()[name = tensor("op_1222_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1223_cast_fp16 = mul(x = linear_47_cast_fp16, y = var_1222_to_fp16)[name = tensor("op_1223_cast_fp16")]; - tensor input_291_cast_fp16 = add(x = input_279_cast_fp16, y = var_1223_cast_fp16)[name = tensor("input_291_cast_fp16")]; - tensor query_11_axes_0 = const()[name = tensor("query_11_axes_0"), val = tensor([-1])]; - tensor module_layers_5_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_5_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(69527680)))]; - tensor module_layers_5_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_5_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(69528768)))]; - tensor query_11_cast_fp16 = layer_norm(axes = query_11_axes_0, beta = module_layers_5_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_5_norm_self_att_weight_to_fp16, x = input_291_cast_fp16)[name = tensor("query_11_cast_fp16")]; - tensor module_layers_5_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_5_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(69529856)))]; - tensor module_layers_5_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_5_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70054208)))]; - tensor linear_48_cast_fp16 = linear(bias = module_layers_5_self_attn_linear_q_bias_to_fp16, weight = module_layers_5_self_attn_linear_q_weight_to_fp16, x = query_11_cast_fp16)[name = tensor("linear_48_cast_fp16")]; - tensor var_1240 = const()[name = tensor("op_1240"), val = tensor([1, -1, 8, 64])]; - tensor q_31_cast_fp16 = reshape(shape = var_1240, x = linear_48_cast_fp16)[name = tensor("q_31_cast_fp16")]; - tensor module_layers_5_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_5_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70055296)))]; - tensor module_layers_5_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_5_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70579648)))]; - tensor linear_49_cast_fp16 = linear(bias = module_layers_5_self_attn_linear_k_bias_to_fp16, weight = module_layers_5_self_attn_linear_k_weight_to_fp16, x = query_11_cast_fp16)[name = tensor("linear_49_cast_fp16")]; - tensor var_1245 = const()[name = tensor("op_1245"), val = tensor([1, -1, 8, 64])]; - tensor k_21_cast_fp16 = reshape(shape = var_1245, x = linear_49_cast_fp16)[name = tensor("k_21_cast_fp16")]; - tensor module_layers_5_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_5_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70580736)))]; - tensor module_layers_5_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_5_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71105088)))]; - tensor linear_50_cast_fp16 = linear(bias = module_layers_5_self_attn_linear_v_bias_to_fp16, weight = module_layers_5_self_attn_linear_v_weight_to_fp16, x = query_11_cast_fp16)[name = tensor("linear_50_cast_fp16")]; - tensor var_1250 = const()[name = tensor("op_1250"), val = tensor([1, -1, 8, 64])]; - tensor v_11_cast_fp16 = reshape(shape = var_1250, x = linear_50_cast_fp16)[name = tensor("v_11_cast_fp16")]; - tensor value_13_perm_0 = const()[name = tensor("value_13_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_5_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_5_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71106176)))]; - tensor var_1262_cast_fp16 = add(x = q_31_cast_fp16, y = module_layers_5_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1262_cast_fp16")]; - tensor module_layers_5_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_5_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71107264)))]; - tensor var_1264_cast_fp16 = add(x = q_31_cast_fp16, y = module_layers_5_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1264_cast_fp16")]; - tensor q_with_bias_v_11_perm_0 = const()[name = tensor("q_with_bias_v_11_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_117_transpose_x_0 = const()[name = tensor("x_117_transpose_x_0"), val = tensor(false)]; - tensor x_117_transpose_y_0 = const()[name = tensor("x_117_transpose_y_0"), val = tensor(false)]; - tensor var_1266_to_fp16 = const()[name = tensor("op_1266_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71108352)))]; - tensor q_with_bias_v_11_cast_fp16 = transpose(perm = q_with_bias_v_11_perm_0, x = var_1264_cast_fp16)[name = tensor("transpose_168")]; - tensor x_117_cast_fp16 = matmul(transpose_x = x_117_transpose_x_0, transpose_y = x_117_transpose_y_0, x = q_with_bias_v_11_cast_fp16, y = var_1266_to_fp16)[name = tensor("x_117_cast_fp16")]; - tensor x_119_pad_0 = const()[name = tensor("x_119_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_119_mode_0 = const()[name = tensor("x_119_mode_0"), val = tensor("constant")]; - tensor const_120_to_fp16 = const()[name = tensor("const_120_to_fp16"), val = tensor(0x0p+0)]; - tensor x_119_cast_fp16 = pad(constant_val = const_120_to_fp16, mode = x_119_mode_0, pad = x_119_pad_0, x = x_117_cast_fp16)[name = tensor("x_119_cast_fp16")]; - tensor var_1274 = const()[name = tensor("op_1274"), val = tensor([1, 8, -1, 188])]; - tensor x_121_cast_fp16 = reshape(shape = var_1274, x = x_119_cast_fp16)[name = tensor("x_121_cast_fp16")]; - tensor var_1278_begin_0 = const()[name = tensor("op_1278_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_1278_end_0 = const()[name = tensor("op_1278_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_1278_end_mask_0 = const()[name = tensor("op_1278_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_1278_cast_fp16 = slice_by_index(begin = var_1278_begin_0, end = var_1278_end_0, end_mask = var_1278_end_mask_0, x = x_121_cast_fp16)[name = tensor("op_1278_cast_fp16")]; - tensor var_1279 = const()[name = tensor("op_1279"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_21_cast_fp16 = reshape(shape = var_1279, x = var_1278_cast_fp16)[name = tensor("matrix_bd_21_cast_fp16")]; - tensor matrix_ac_11_transpose_x_0 = const()[name = tensor("matrix_ac_11_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_11_transpose_y_0 = const()[name = tensor("matrix_ac_11_transpose_y_0"), val = tensor(false)]; - tensor transpose_61_perm_0 = const()[name = tensor("transpose_61_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_62_perm_0 = const()[name = tensor("transpose_62_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_62 = transpose(perm = transpose_62_perm_0, x = k_21_cast_fp16)[name = tensor("transpose_166")]; - tensor transpose_61 = transpose(perm = transpose_61_perm_0, x = var_1262_cast_fp16)[name = tensor("transpose_167")]; - tensor matrix_ac_11_cast_fp16 = matmul(transpose_x = matrix_ac_11_transpose_x_0, transpose_y = matrix_ac_11_transpose_y_0, x = transpose_61, y = transpose_62)[name = tensor("matrix_ac_11_cast_fp16")]; - tensor matrix_bd_23_begin_0 = const()[name = tensor("matrix_bd_23_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_23_end_0 = const()[name = tensor("matrix_bd_23_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_23_end_mask_0 = const()[name = tensor("matrix_bd_23_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_23_cast_fp16 = slice_by_index(begin = matrix_bd_23_begin_0, end = matrix_bd_23_end_0, end_mask = matrix_bd_23_end_mask_0, x = matrix_bd_21_cast_fp16)[name = tensor("matrix_bd_23_cast_fp16")]; - tensor var_1288_cast_fp16 = add(x = matrix_ac_11_cast_fp16, y = matrix_bd_23_cast_fp16)[name = tensor("op_1288_cast_fp16")]; - tensor _inversed_scores_21_y_0_to_fp16 = const()[name = tensor("_inversed_scores_21_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_21_cast_fp16 = mul(x = var_1288_cast_fp16, y = _inversed_scores_21_y_0_to_fp16)[name = tensor("_inversed_scores_21_cast_fp16")]; - tensor scores_23_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_21_cast_fp16, cond = mask_11)[name = tensor("scores_23_cast_fp16")]; - tensor var_1294_cast_fp16 = softmax(axis = var_23, x = scores_23_cast_fp16)[name = tensor("op_1294_cast_fp16")]; - tensor input_293_cast_fp16 = select(a = var_11_to_fp16, b = var_1294_cast_fp16, cond = mask_11)[name = tensor("input_293_cast_fp16")]; - tensor x_123_transpose_x_0 = const()[name = tensor("x_123_transpose_x_0"), val = tensor(false)]; - tensor x_123_transpose_y_0 = const()[name = tensor("x_123_transpose_y_0"), val = tensor(false)]; - tensor value_13_cast_fp16 = transpose(perm = value_13_perm_0, x = v_11_cast_fp16)[name = tensor("transpose_169")]; - tensor x_123_cast_fp16 = matmul(transpose_x = x_123_transpose_x_0, transpose_y = x_123_transpose_y_0, x = input_293_cast_fp16, y = value_13_cast_fp16)[name = tensor("x_123_cast_fp16")]; - tensor var_1298_perm_0 = const()[name = tensor("op_1298_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_1299 = const()[name = tensor("op_1299"), val = tensor([1, -1, 512])]; - tensor var_1298_cast_fp16 = transpose(perm = var_1298_perm_0, x = x_123_cast_fp16)[name = tensor("transpose_165")]; - tensor input_295_cast_fp16 = reshape(shape = var_1299, x = var_1298_cast_fp16)[name = tensor("input_295_cast_fp16")]; - tensor module_layers_5_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_5_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71492416)))]; - tensor module_layers_5_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_5_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(72016768)))]; - tensor linear_52_cast_fp16 = linear(bias = module_layers_5_self_attn_linear_out_bias_to_fp16, weight = module_layers_5_self_attn_linear_out_weight_to_fp16, x = input_295_cast_fp16)[name = tensor("linear_52_cast_fp16")]; - tensor input_299_cast_fp16 = add(x = input_291_cast_fp16, y = linear_52_cast_fp16)[name = tensor("input_299_cast_fp16")]; - tensor x_127_axes_0 = const()[name = tensor("x_127_axes_0"), val = tensor([-1])]; - tensor module_layers_5_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_5_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(72017856)))]; - tensor module_layers_5_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_5_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(72018944)))]; - tensor x_127_cast_fp16 = layer_norm(axes = x_127_axes_0, beta = module_layers_5_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_5_norm_conv_weight_to_fp16, x = input_299_cast_fp16)[name = tensor("x_127_cast_fp16")]; - tensor input_301_perm_0 = const()[name = tensor("input_301_perm_0"), val = tensor([0, 2, 1])]; - tensor input_303_pad_type_0 = const()[name = tensor("input_303_pad_type_0"), val = tensor("valid")]; - tensor input_303_strides_0 = const()[name = tensor("input_303_strides_0"), val = tensor([1])]; - tensor input_303_pad_0 = const()[name = tensor("input_303_pad_0"), val = tensor([0, 0])]; - tensor input_303_dilations_0 = const()[name = tensor("input_303_dilations_0"), val = tensor([1])]; - tensor input_303_groups_0 = const()[name = tensor("input_303_groups_0"), val = tensor(1)]; - tensor module_layers_5_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_5_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(72020032)))]; - tensor module_layers_5_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_5_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73068672)))]; - tensor input_301_cast_fp16 = transpose(perm = input_301_perm_0, x = x_127_cast_fp16)[name = tensor("transpose_164")]; - tensor input_303_cast_fp16 = conv(bias = module_layers_5_conv_pointwise_conv1_bias_to_fp16, dilations = input_303_dilations_0, groups = input_303_groups_0, pad = input_303_pad_0, pad_type = input_303_pad_type_0, strides = input_303_strides_0, weight = module_layers_5_conv_pointwise_conv1_weight_to_fp16, x = input_301_cast_fp16)[name = tensor("input_303_cast_fp16")]; - tensor x_129_split_num_splits_0 = const()[name = tensor("x_129_split_num_splits_0"), val = tensor(2)]; - tensor x_129_split_axis_0 = const()[name = tensor("x_129_split_axis_0"), val = tensor(1)]; - tensor x_129_split_cast_fp16_0, tensor x_129_split_cast_fp16_1 = split(axis = x_129_split_axis_0, num_splits = x_129_split_num_splits_0, x = input_303_cast_fp16)[name = tensor("x_129_split_cast_fp16")]; - tensor x_129_split_1_sigmoid_cast_fp16 = sigmoid(x = x_129_split_cast_fp16_1)[name = tensor("x_129_split_1_sigmoid_cast_fp16")]; - tensor x_129_cast_fp16 = mul(x = x_129_split_cast_fp16_0, y = x_129_split_1_sigmoid_cast_fp16)[name = tensor("x_129_cast_fp16")]; - tensor input_305_cast_fp16 = select(a = var_11_to_fp16, b = x_129_cast_fp16, cond = var_453)[name = tensor("input_305_cast_fp16")]; - tensor input_307_pad_0 = const()[name = tensor("input_307_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_307_mode_0 = const()[name = tensor("input_307_mode_0"), val = tensor("constant")]; - tensor const_123_to_fp16 = const()[name = tensor("const_123_to_fp16"), val = tensor(0x0p+0)]; - tensor input_307_cast_fp16 = pad(constant_val = const_123_to_fp16, mode = input_307_mode_0, pad = input_307_pad_0, x = input_305_cast_fp16)[name = tensor("input_307_cast_fp16")]; - tensor input_309_pad_type_0 = const()[name = tensor("input_309_pad_type_0"), val = tensor("valid")]; - tensor input_309_groups_0 = const()[name = tensor("input_309_groups_0"), val = tensor(512)]; - tensor input_309_strides_0 = const()[name = tensor("input_309_strides_0"), val = tensor([1])]; - tensor input_309_pad_0 = const()[name = tensor("input_309_pad_0"), val = tensor([0, 0])]; - tensor input_309_dilations_0 = const()[name = tensor("input_309_dilations_0"), val = tensor([1])]; - tensor const_244_to_fp16 = const()[name = tensor("const_244_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73070784)))]; - tensor const_245_to_fp16 = const()[name = tensor("const_245_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73080064)))]; - tensor input_311_cast_fp16 = conv(bias = const_245_to_fp16, dilations = input_309_dilations_0, groups = input_309_groups_0, pad = input_309_pad_0, pad_type = input_309_pad_type_0, strides = input_309_strides_0, weight = const_244_to_fp16, x = input_307_cast_fp16)[name = tensor("input_311_cast_fp16")]; - tensor input_313_cast_fp16 = silu(x = input_311_cast_fp16)[name = tensor("input_313_cast_fp16")]; - tensor x_131_pad_type_0 = const()[name = tensor("x_131_pad_type_0"), val = tensor("valid")]; - tensor x_131_strides_0 = const()[name = tensor("x_131_strides_0"), val = tensor([1])]; - tensor x_131_pad_0 = const()[name = tensor("x_131_pad_0"), val = tensor([0, 0])]; - tensor x_131_dilations_0 = const()[name = tensor("x_131_dilations_0"), val = tensor([1])]; - tensor x_131_groups_0 = const()[name = tensor("x_131_groups_0"), val = tensor(1)]; - tensor module_layers_5_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_5_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73081152)))]; - tensor module_layers_5_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_5_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73605504)))]; - tensor x_131_cast_fp16 = conv(bias = module_layers_5_conv_pointwise_conv2_bias_to_fp16, dilations = x_131_dilations_0, groups = x_131_groups_0, pad = x_131_pad_0, pad_type = x_131_pad_type_0, strides = x_131_strides_0, weight = module_layers_5_conv_pointwise_conv2_weight_to_fp16, x = input_313_cast_fp16)[name = tensor("x_131_cast_fp16")]; - tensor input_315_perm_0 = const()[name = tensor("input_315_perm_0"), val = tensor([0, 2, 1])]; - tensor input_315_cast_fp16 = transpose(perm = input_315_perm_0, x = x_131_cast_fp16)[name = tensor("transpose_163")]; - tensor input_317_cast_fp16 = add(x = input_299_cast_fp16, y = input_315_cast_fp16)[name = tensor("input_317_cast_fp16")]; - tensor input_319_axes_0 = const()[name = tensor("input_319_axes_0"), val = tensor([-1])]; - tensor module_layers_5_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_5_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73606592)))]; - tensor module_layers_5_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_5_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73607680)))]; - tensor input_319_cast_fp16 = layer_norm(axes = input_319_axes_0, beta = module_layers_5_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_5_norm_feed_forward2_weight_to_fp16, x = input_317_cast_fp16)[name = tensor("input_319_cast_fp16")]; - tensor module_layers_5_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_5_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73608768)))]; - tensor module_layers_5_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_5_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(75705984)))]; - tensor linear_53_cast_fp16 = linear(bias = module_layers_5_feed_forward2_linear1_bias_to_fp16, weight = module_layers_5_feed_forward2_linear1_weight_to_fp16, x = input_319_cast_fp16)[name = tensor("linear_53_cast_fp16")]; - tensor input_323_cast_fp16 = silu(x = linear_53_cast_fp16)[name = tensor("input_323_cast_fp16")]; - tensor module_layers_5_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_5_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(75710144)))]; - tensor module_layers_5_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_5_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(77807360)))]; - tensor linear_54_cast_fp16 = linear(bias = module_layers_5_feed_forward2_linear2_bias_to_fp16, weight = module_layers_5_feed_forward2_linear2_weight_to_fp16, x = input_323_cast_fp16)[name = tensor("linear_54_cast_fp16")]; - tensor var_1365_to_fp16 = const()[name = tensor("op_1365_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1366_cast_fp16 = mul(x = linear_54_cast_fp16, y = var_1365_to_fp16)[name = tensor("op_1366_cast_fp16")]; - tensor input_329_cast_fp16 = add(x = input_317_cast_fp16, y = var_1366_cast_fp16)[name = tensor("input_329_cast_fp16")]; - tensor input_331_axes_0 = const()[name = tensor("input_331_axes_0"), val = tensor([-1])]; - tensor module_layers_5_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_5_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(77808448)))]; - tensor module_layers_5_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_5_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(77809536)))]; - tensor input_331_cast_fp16 = layer_norm(axes = input_331_axes_0, beta = module_layers_5_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_5_norm_out_weight_to_fp16, x = input_329_cast_fp16)[name = tensor("input_331_cast_fp16")]; - tensor input_333_axes_0 = const()[name = tensor("input_333_axes_0"), val = tensor([-1])]; - tensor module_layers_6_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_6_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(77810624)))]; - tensor module_layers_6_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_6_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(77811712)))]; - tensor input_333_cast_fp16 = layer_norm(axes = input_333_axes_0, beta = module_layers_6_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_6_norm_feed_forward1_weight_to_fp16, x = input_331_cast_fp16)[name = tensor("input_333_cast_fp16")]; - tensor module_layers_6_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_6_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(77812800)))]; - tensor module_layers_6_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_6_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(79910016)))]; - tensor linear_55_cast_fp16 = linear(bias = module_layers_6_feed_forward1_linear1_bias_to_fp16, weight = module_layers_6_feed_forward1_linear1_weight_to_fp16, x = input_333_cast_fp16)[name = tensor("linear_55_cast_fp16")]; - tensor input_337_cast_fp16 = silu(x = linear_55_cast_fp16)[name = tensor("input_337_cast_fp16")]; - tensor module_layers_6_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_6_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(79914176)))]; - tensor module_layers_6_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_6_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82011392)))]; - tensor linear_56_cast_fp16 = linear(bias = module_layers_6_feed_forward1_linear2_bias_to_fp16, weight = module_layers_6_feed_forward1_linear2_weight_to_fp16, x = input_337_cast_fp16)[name = tensor("linear_56_cast_fp16")]; - tensor var_1396_to_fp16 = const()[name = tensor("op_1396_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1397_cast_fp16 = mul(x = linear_56_cast_fp16, y = var_1396_to_fp16)[name = tensor("op_1397_cast_fp16")]; - tensor input_343_cast_fp16 = add(x = input_331_cast_fp16, y = var_1397_cast_fp16)[name = tensor("input_343_cast_fp16")]; - tensor query_13_axes_0 = const()[name = tensor("query_13_axes_0"), val = tensor([-1])]; - tensor module_layers_6_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_6_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82012480)))]; - tensor module_layers_6_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_6_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82013568)))]; - tensor query_13_cast_fp16 = layer_norm(axes = query_13_axes_0, beta = module_layers_6_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_6_norm_self_att_weight_to_fp16, x = input_343_cast_fp16)[name = tensor("query_13_cast_fp16")]; - tensor module_layers_6_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_6_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82014656)))]; - tensor module_layers_6_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_6_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82539008)))]; - tensor linear_57_cast_fp16 = linear(bias = module_layers_6_self_attn_linear_q_bias_to_fp16, weight = module_layers_6_self_attn_linear_q_weight_to_fp16, x = query_13_cast_fp16)[name = tensor("linear_57_cast_fp16")]; - tensor var_1414 = const()[name = tensor("op_1414"), val = tensor([1, -1, 8, 64])]; - tensor q_37_cast_fp16 = reshape(shape = var_1414, x = linear_57_cast_fp16)[name = tensor("q_37_cast_fp16")]; - tensor module_layers_6_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_6_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82540096)))]; - tensor module_layers_6_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_6_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83064448)))]; - tensor linear_58_cast_fp16 = linear(bias = module_layers_6_self_attn_linear_k_bias_to_fp16, weight = module_layers_6_self_attn_linear_k_weight_to_fp16, x = query_13_cast_fp16)[name = tensor("linear_58_cast_fp16")]; - tensor var_1419 = const()[name = tensor("op_1419"), val = tensor([1, -1, 8, 64])]; - tensor k_25_cast_fp16 = reshape(shape = var_1419, x = linear_58_cast_fp16)[name = tensor("k_25_cast_fp16")]; - tensor module_layers_6_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_6_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83065536)))]; - tensor module_layers_6_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_6_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83589888)))]; - tensor linear_59_cast_fp16 = linear(bias = module_layers_6_self_attn_linear_v_bias_to_fp16, weight = module_layers_6_self_attn_linear_v_weight_to_fp16, x = query_13_cast_fp16)[name = tensor("linear_59_cast_fp16")]; - tensor var_1424 = const()[name = tensor("op_1424"), val = tensor([1, -1, 8, 64])]; - tensor v_13_cast_fp16 = reshape(shape = var_1424, x = linear_59_cast_fp16)[name = tensor("v_13_cast_fp16")]; - tensor value_15_perm_0 = const()[name = tensor("value_15_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_6_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_6_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83590976)))]; - tensor var_1436_cast_fp16 = add(x = q_37_cast_fp16, y = module_layers_6_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1436_cast_fp16")]; - tensor module_layers_6_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_6_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83592064)))]; - tensor var_1438_cast_fp16 = add(x = q_37_cast_fp16, y = module_layers_6_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1438_cast_fp16")]; - tensor q_with_bias_v_13_perm_0 = const()[name = tensor("q_with_bias_v_13_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_139_transpose_x_0 = const()[name = tensor("x_139_transpose_x_0"), val = tensor(false)]; - tensor x_139_transpose_y_0 = const()[name = tensor("x_139_transpose_y_0"), val = tensor(false)]; - tensor var_1440_to_fp16 = const()[name = tensor("op_1440_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83593152)))]; - tensor q_with_bias_v_13_cast_fp16 = transpose(perm = q_with_bias_v_13_perm_0, x = var_1438_cast_fp16)[name = tensor("transpose_161")]; - tensor x_139_cast_fp16 = matmul(transpose_x = x_139_transpose_x_0, transpose_y = x_139_transpose_y_0, x = q_with_bias_v_13_cast_fp16, y = var_1440_to_fp16)[name = tensor("x_139_cast_fp16")]; - tensor x_141_pad_0 = const()[name = tensor("x_141_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_141_mode_0 = const()[name = tensor("x_141_mode_0"), val = tensor("constant")]; - tensor const_130_to_fp16 = const()[name = tensor("const_130_to_fp16"), val = tensor(0x0p+0)]; - tensor x_141_cast_fp16 = pad(constant_val = const_130_to_fp16, mode = x_141_mode_0, pad = x_141_pad_0, x = x_139_cast_fp16)[name = tensor("x_141_cast_fp16")]; - tensor var_1448 = const()[name = tensor("op_1448"), val = tensor([1, 8, -1, 188])]; - tensor x_143_cast_fp16 = reshape(shape = var_1448, x = x_141_cast_fp16)[name = tensor("x_143_cast_fp16")]; - tensor var_1452_begin_0 = const()[name = tensor("op_1452_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_1452_end_0 = const()[name = tensor("op_1452_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_1452_end_mask_0 = const()[name = tensor("op_1452_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_1452_cast_fp16 = slice_by_index(begin = var_1452_begin_0, end = var_1452_end_0, end_mask = var_1452_end_mask_0, x = x_143_cast_fp16)[name = tensor("op_1452_cast_fp16")]; - tensor var_1453 = const()[name = tensor("op_1453"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_25_cast_fp16 = reshape(shape = var_1453, x = var_1452_cast_fp16)[name = tensor("matrix_bd_25_cast_fp16")]; - tensor matrix_ac_13_transpose_x_0 = const()[name = tensor("matrix_ac_13_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_13_transpose_y_0 = const()[name = tensor("matrix_ac_13_transpose_y_0"), val = tensor(false)]; - tensor transpose_63_perm_0 = const()[name = tensor("transpose_63_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_64_perm_0 = const()[name = tensor("transpose_64_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_64 = transpose(perm = transpose_64_perm_0, x = k_25_cast_fp16)[name = tensor("transpose_159")]; - tensor transpose_63 = transpose(perm = transpose_63_perm_0, x = var_1436_cast_fp16)[name = tensor("transpose_160")]; - tensor matrix_ac_13_cast_fp16 = matmul(transpose_x = matrix_ac_13_transpose_x_0, transpose_y = matrix_ac_13_transpose_y_0, x = transpose_63, y = transpose_64)[name = tensor("matrix_ac_13_cast_fp16")]; - tensor matrix_bd_27_begin_0 = const()[name = tensor("matrix_bd_27_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_27_end_0 = const()[name = tensor("matrix_bd_27_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_27_end_mask_0 = const()[name = tensor("matrix_bd_27_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_27_cast_fp16 = slice_by_index(begin = matrix_bd_27_begin_0, end = matrix_bd_27_end_0, end_mask = matrix_bd_27_end_mask_0, x = matrix_bd_25_cast_fp16)[name = tensor("matrix_bd_27_cast_fp16")]; - tensor var_1462_cast_fp16 = add(x = matrix_ac_13_cast_fp16, y = matrix_bd_27_cast_fp16)[name = tensor("op_1462_cast_fp16")]; - tensor _inversed_scores_25_y_0_to_fp16 = const()[name = tensor("_inversed_scores_25_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_25_cast_fp16 = mul(x = var_1462_cast_fp16, y = _inversed_scores_25_y_0_to_fp16)[name = tensor("_inversed_scores_25_cast_fp16")]; - tensor scores_27_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_25_cast_fp16, cond = mask_11)[name = tensor("scores_27_cast_fp16")]; - tensor var_1468_cast_fp16 = softmax(axis = var_23, x = scores_27_cast_fp16)[name = tensor("op_1468_cast_fp16")]; - tensor input_345_cast_fp16 = select(a = var_11_to_fp16, b = var_1468_cast_fp16, cond = mask_11)[name = tensor("input_345_cast_fp16")]; - tensor x_145_transpose_x_0 = const()[name = tensor("x_145_transpose_x_0"), val = tensor(false)]; - tensor x_145_transpose_y_0 = const()[name = tensor("x_145_transpose_y_0"), val = tensor(false)]; - tensor value_15_cast_fp16 = transpose(perm = value_15_perm_0, x = v_13_cast_fp16)[name = tensor("transpose_162")]; - tensor x_145_cast_fp16 = matmul(transpose_x = x_145_transpose_x_0, transpose_y = x_145_transpose_y_0, x = input_345_cast_fp16, y = value_15_cast_fp16)[name = tensor("x_145_cast_fp16")]; - tensor var_1472_perm_0 = const()[name = tensor("op_1472_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_1473 = const()[name = tensor("op_1473"), val = tensor([1, -1, 512])]; - tensor var_1472_cast_fp16 = transpose(perm = var_1472_perm_0, x = x_145_cast_fp16)[name = tensor("transpose_158")]; - tensor input_347_cast_fp16 = reshape(shape = var_1473, x = var_1472_cast_fp16)[name = tensor("input_347_cast_fp16")]; - tensor module_layers_6_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_6_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83977216)))]; - tensor module_layers_6_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_6_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84501568)))]; - tensor linear_61_cast_fp16 = linear(bias = module_layers_6_self_attn_linear_out_bias_to_fp16, weight = module_layers_6_self_attn_linear_out_weight_to_fp16, x = input_347_cast_fp16)[name = tensor("linear_61_cast_fp16")]; - tensor input_351_cast_fp16 = add(x = input_343_cast_fp16, y = linear_61_cast_fp16)[name = tensor("input_351_cast_fp16")]; - tensor x_149_axes_0 = const()[name = tensor("x_149_axes_0"), val = tensor([-1])]; - tensor module_layers_6_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_6_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84502656)))]; - tensor module_layers_6_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_6_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84503744)))]; - tensor x_149_cast_fp16 = layer_norm(axes = x_149_axes_0, beta = module_layers_6_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_6_norm_conv_weight_to_fp16, x = input_351_cast_fp16)[name = tensor("x_149_cast_fp16")]; - tensor input_353_perm_0 = const()[name = tensor("input_353_perm_0"), val = tensor([0, 2, 1])]; - tensor input_355_pad_type_0 = const()[name = tensor("input_355_pad_type_0"), val = tensor("valid")]; - tensor input_355_strides_0 = const()[name = tensor("input_355_strides_0"), val = tensor([1])]; - tensor input_355_pad_0 = const()[name = tensor("input_355_pad_0"), val = tensor([0, 0])]; - tensor input_355_dilations_0 = const()[name = tensor("input_355_dilations_0"), val = tensor([1])]; - tensor input_355_groups_0 = const()[name = tensor("input_355_groups_0"), val = tensor(1)]; - tensor module_layers_6_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_6_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84504832)))]; - tensor module_layers_6_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_6_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85553472)))]; - tensor input_353_cast_fp16 = transpose(perm = input_353_perm_0, x = x_149_cast_fp16)[name = tensor("transpose_157")]; - tensor input_355_cast_fp16 = conv(bias = module_layers_6_conv_pointwise_conv1_bias_to_fp16, dilations = input_355_dilations_0, groups = input_355_groups_0, pad = input_355_pad_0, pad_type = input_355_pad_type_0, strides = input_355_strides_0, weight = module_layers_6_conv_pointwise_conv1_weight_to_fp16, x = input_353_cast_fp16)[name = tensor("input_355_cast_fp16")]; - tensor x_151_split_num_splits_0 = const()[name = tensor("x_151_split_num_splits_0"), val = tensor(2)]; - tensor x_151_split_axis_0 = const()[name = tensor("x_151_split_axis_0"), val = tensor(1)]; - tensor x_151_split_cast_fp16_0, tensor x_151_split_cast_fp16_1 = split(axis = x_151_split_axis_0, num_splits = x_151_split_num_splits_0, x = input_355_cast_fp16)[name = tensor("x_151_split_cast_fp16")]; - tensor x_151_split_1_sigmoid_cast_fp16 = sigmoid(x = x_151_split_cast_fp16_1)[name = tensor("x_151_split_1_sigmoid_cast_fp16")]; - tensor x_151_cast_fp16 = mul(x = x_151_split_cast_fp16_0, y = x_151_split_1_sigmoid_cast_fp16)[name = tensor("x_151_cast_fp16")]; - tensor input_357_cast_fp16 = select(a = var_11_to_fp16, b = x_151_cast_fp16, cond = var_453)[name = tensor("input_357_cast_fp16")]; - tensor input_359_pad_0 = const()[name = tensor("input_359_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_359_mode_0 = const()[name = tensor("input_359_mode_0"), val = tensor("constant")]; - tensor const_133_to_fp16 = const()[name = tensor("const_133_to_fp16"), val = tensor(0x0p+0)]; - tensor input_359_cast_fp16 = pad(constant_val = const_133_to_fp16, mode = input_359_mode_0, pad = input_359_pad_0, x = input_357_cast_fp16)[name = tensor("input_359_cast_fp16")]; - tensor input_361_pad_type_0 = const()[name = tensor("input_361_pad_type_0"), val = tensor("valid")]; - tensor input_361_groups_0 = const()[name = tensor("input_361_groups_0"), val = tensor(512)]; - tensor input_361_strides_0 = const()[name = tensor("input_361_strides_0"), val = tensor([1])]; - tensor input_361_pad_0 = const()[name = tensor("input_361_pad_0"), val = tensor([0, 0])]; - tensor input_361_dilations_0 = const()[name = tensor("input_361_dilations_0"), val = tensor([1])]; - tensor const_246_to_fp16 = const()[name = tensor("const_246_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85555584)))]; - tensor const_247_to_fp16 = const()[name = tensor("const_247_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85564864)))]; - tensor input_363_cast_fp16 = conv(bias = const_247_to_fp16, dilations = input_361_dilations_0, groups = input_361_groups_0, pad = input_361_pad_0, pad_type = input_361_pad_type_0, strides = input_361_strides_0, weight = const_246_to_fp16, x = input_359_cast_fp16)[name = tensor("input_363_cast_fp16")]; - tensor input_365_cast_fp16 = silu(x = input_363_cast_fp16)[name = tensor("input_365_cast_fp16")]; - tensor x_153_pad_type_0 = const()[name = tensor("x_153_pad_type_0"), val = tensor("valid")]; - tensor x_153_strides_0 = const()[name = tensor("x_153_strides_0"), val = tensor([1])]; - tensor x_153_pad_0 = const()[name = tensor("x_153_pad_0"), val = tensor([0, 0])]; - tensor x_153_dilations_0 = const()[name = tensor("x_153_dilations_0"), val = tensor([1])]; - tensor x_153_groups_0 = const()[name = tensor("x_153_groups_0"), val = tensor(1)]; - tensor module_layers_6_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_6_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85565952)))]; - tensor module_layers_6_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_6_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86090304)))]; - tensor x_153_cast_fp16 = conv(bias = module_layers_6_conv_pointwise_conv2_bias_to_fp16, dilations = x_153_dilations_0, groups = x_153_groups_0, pad = x_153_pad_0, pad_type = x_153_pad_type_0, strides = x_153_strides_0, weight = module_layers_6_conv_pointwise_conv2_weight_to_fp16, x = input_365_cast_fp16)[name = tensor("x_153_cast_fp16")]; - tensor input_367_perm_0 = const()[name = tensor("input_367_perm_0"), val = tensor([0, 2, 1])]; - tensor input_367_cast_fp16 = transpose(perm = input_367_perm_0, x = x_153_cast_fp16)[name = tensor("transpose_156")]; - tensor input_369_cast_fp16 = add(x = input_351_cast_fp16, y = input_367_cast_fp16)[name = tensor("input_369_cast_fp16")]; - tensor input_371_axes_0 = const()[name = tensor("input_371_axes_0"), val = tensor([-1])]; - tensor module_layers_6_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_6_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86091392)))]; - tensor module_layers_6_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_6_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86092480)))]; - tensor input_371_cast_fp16 = layer_norm(axes = input_371_axes_0, beta = module_layers_6_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_6_norm_feed_forward2_weight_to_fp16, x = input_369_cast_fp16)[name = tensor("input_371_cast_fp16")]; - tensor module_layers_6_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_6_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86093568)))]; - tensor module_layers_6_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_6_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(88190784)))]; - tensor linear_62_cast_fp16 = linear(bias = module_layers_6_feed_forward2_linear1_bias_to_fp16, weight = module_layers_6_feed_forward2_linear1_weight_to_fp16, x = input_371_cast_fp16)[name = tensor("linear_62_cast_fp16")]; - tensor input_375_cast_fp16 = silu(x = linear_62_cast_fp16)[name = tensor("input_375_cast_fp16")]; - tensor module_layers_6_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_6_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(88194944)))]; - tensor module_layers_6_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_6_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90292160)))]; - tensor linear_63_cast_fp16 = linear(bias = module_layers_6_feed_forward2_linear2_bias_to_fp16, weight = module_layers_6_feed_forward2_linear2_weight_to_fp16, x = input_375_cast_fp16)[name = tensor("linear_63_cast_fp16")]; - tensor var_1539_to_fp16 = const()[name = tensor("op_1539_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1540_cast_fp16 = mul(x = linear_63_cast_fp16, y = var_1539_to_fp16)[name = tensor("op_1540_cast_fp16")]; - tensor input_381_cast_fp16 = add(x = input_369_cast_fp16, y = var_1540_cast_fp16)[name = tensor("input_381_cast_fp16")]; - tensor input_383_axes_0 = const()[name = tensor("input_383_axes_0"), val = tensor([-1])]; - tensor module_layers_6_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_6_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90293248)))]; - tensor module_layers_6_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_6_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90294336)))]; - tensor input_383_cast_fp16 = layer_norm(axes = input_383_axes_0, beta = module_layers_6_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_6_norm_out_weight_to_fp16, x = input_381_cast_fp16)[name = tensor("input_383_cast_fp16")]; - tensor input_385_axes_0 = const()[name = tensor("input_385_axes_0"), val = tensor([-1])]; - tensor module_layers_7_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_7_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90295424)))]; - tensor module_layers_7_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_7_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90296512)))]; - tensor input_385_cast_fp16 = layer_norm(axes = input_385_axes_0, beta = module_layers_7_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_7_norm_feed_forward1_weight_to_fp16, x = input_383_cast_fp16)[name = tensor("input_385_cast_fp16")]; - tensor module_layers_7_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_7_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90297600)))]; - tensor module_layers_7_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_7_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92394816)))]; - tensor linear_64_cast_fp16 = linear(bias = module_layers_7_feed_forward1_linear1_bias_to_fp16, weight = module_layers_7_feed_forward1_linear1_weight_to_fp16, x = input_385_cast_fp16)[name = tensor("linear_64_cast_fp16")]; - tensor input_389_cast_fp16 = silu(x = linear_64_cast_fp16)[name = tensor("input_389_cast_fp16")]; - tensor module_layers_7_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_7_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92398976)))]; - tensor module_layers_7_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_7_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(94496192)))]; - tensor linear_65_cast_fp16 = linear(bias = module_layers_7_feed_forward1_linear2_bias_to_fp16, weight = module_layers_7_feed_forward1_linear2_weight_to_fp16, x = input_389_cast_fp16)[name = tensor("linear_65_cast_fp16")]; - tensor var_1570_to_fp16 = const()[name = tensor("op_1570_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1571_cast_fp16 = mul(x = linear_65_cast_fp16, y = var_1570_to_fp16)[name = tensor("op_1571_cast_fp16")]; - tensor input_395_cast_fp16 = add(x = input_383_cast_fp16, y = var_1571_cast_fp16)[name = tensor("input_395_cast_fp16")]; - tensor query_15_axes_0 = const()[name = tensor("query_15_axes_0"), val = tensor([-1])]; - tensor module_layers_7_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_7_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(94497280)))]; - tensor module_layers_7_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_7_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(94498368)))]; - tensor query_15_cast_fp16 = layer_norm(axes = query_15_axes_0, beta = module_layers_7_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_7_norm_self_att_weight_to_fp16, x = input_395_cast_fp16)[name = tensor("query_15_cast_fp16")]; - tensor module_layers_7_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_7_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(94499456)))]; - tensor module_layers_7_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_7_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95023808)))]; - tensor linear_66_cast_fp16 = linear(bias = module_layers_7_self_attn_linear_q_bias_to_fp16, weight = module_layers_7_self_attn_linear_q_weight_to_fp16, x = query_15_cast_fp16)[name = tensor("linear_66_cast_fp16")]; - tensor var_1588 = const()[name = tensor("op_1588"), val = tensor([1, -1, 8, 64])]; - tensor q_43_cast_fp16 = reshape(shape = var_1588, x = linear_66_cast_fp16)[name = tensor("q_43_cast_fp16")]; - tensor module_layers_7_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_7_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95024896)))]; - tensor module_layers_7_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_7_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95549248)))]; - tensor linear_67_cast_fp16 = linear(bias = module_layers_7_self_attn_linear_k_bias_to_fp16, weight = module_layers_7_self_attn_linear_k_weight_to_fp16, x = query_15_cast_fp16)[name = tensor("linear_67_cast_fp16")]; - tensor var_1593 = const()[name = tensor("op_1593"), val = tensor([1, -1, 8, 64])]; - tensor k_29_cast_fp16 = reshape(shape = var_1593, x = linear_67_cast_fp16)[name = tensor("k_29_cast_fp16")]; - tensor module_layers_7_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_7_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95550336)))]; - tensor module_layers_7_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_7_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96074688)))]; - tensor linear_68_cast_fp16 = linear(bias = module_layers_7_self_attn_linear_v_bias_to_fp16, weight = module_layers_7_self_attn_linear_v_weight_to_fp16, x = query_15_cast_fp16)[name = tensor("linear_68_cast_fp16")]; - tensor var_1598 = const()[name = tensor("op_1598"), val = tensor([1, -1, 8, 64])]; - tensor v_15_cast_fp16 = reshape(shape = var_1598, x = linear_68_cast_fp16)[name = tensor("v_15_cast_fp16")]; - tensor value_17_perm_0 = const()[name = tensor("value_17_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_7_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_7_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96075776)))]; - tensor var_1610_cast_fp16 = add(x = q_43_cast_fp16, y = module_layers_7_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1610_cast_fp16")]; - tensor module_layers_7_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_7_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96076864)))]; - tensor var_1612_cast_fp16 = add(x = q_43_cast_fp16, y = module_layers_7_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1612_cast_fp16")]; - tensor q_with_bias_v_15_perm_0 = const()[name = tensor("q_with_bias_v_15_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_161_transpose_x_0 = const()[name = tensor("x_161_transpose_x_0"), val = tensor(false)]; - tensor x_161_transpose_y_0 = const()[name = tensor("x_161_transpose_y_0"), val = tensor(false)]; - tensor var_1614_to_fp16 = const()[name = tensor("op_1614_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96077952)))]; - tensor q_with_bias_v_15_cast_fp16 = transpose(perm = q_with_bias_v_15_perm_0, x = var_1612_cast_fp16)[name = tensor("transpose_154")]; - tensor x_161_cast_fp16 = matmul(transpose_x = x_161_transpose_x_0, transpose_y = x_161_transpose_y_0, x = q_with_bias_v_15_cast_fp16, y = var_1614_to_fp16)[name = tensor("x_161_cast_fp16")]; - tensor x_163_pad_0 = const()[name = tensor("x_163_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_163_mode_0 = const()[name = tensor("x_163_mode_0"), val = tensor("constant")]; - tensor const_140_to_fp16 = const()[name = tensor("const_140_to_fp16"), val = tensor(0x0p+0)]; - tensor x_163_cast_fp16 = pad(constant_val = const_140_to_fp16, mode = x_163_mode_0, pad = x_163_pad_0, x = x_161_cast_fp16)[name = tensor("x_163_cast_fp16")]; - tensor var_1622 = const()[name = tensor("op_1622"), val = tensor([1, 8, -1, 188])]; - tensor x_165_cast_fp16 = reshape(shape = var_1622, x = x_163_cast_fp16)[name = tensor("x_165_cast_fp16")]; - tensor var_1626_begin_0 = const()[name = tensor("op_1626_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_1626_end_0 = const()[name = tensor("op_1626_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_1626_end_mask_0 = const()[name = tensor("op_1626_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_1626_cast_fp16 = slice_by_index(begin = var_1626_begin_0, end = var_1626_end_0, end_mask = var_1626_end_mask_0, x = x_165_cast_fp16)[name = tensor("op_1626_cast_fp16")]; - tensor var_1627 = const()[name = tensor("op_1627"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_29_cast_fp16 = reshape(shape = var_1627, x = var_1626_cast_fp16)[name = tensor("matrix_bd_29_cast_fp16")]; - tensor matrix_ac_15_transpose_x_0 = const()[name = tensor("matrix_ac_15_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_15_transpose_y_0 = const()[name = tensor("matrix_ac_15_transpose_y_0"), val = tensor(false)]; - tensor transpose_65_perm_0 = const()[name = tensor("transpose_65_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_66_perm_0 = const()[name = tensor("transpose_66_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_66 = transpose(perm = transpose_66_perm_0, x = k_29_cast_fp16)[name = tensor("transpose_152")]; - tensor transpose_65 = transpose(perm = transpose_65_perm_0, x = var_1610_cast_fp16)[name = tensor("transpose_153")]; - tensor matrix_ac_15_cast_fp16 = matmul(transpose_x = matrix_ac_15_transpose_x_0, transpose_y = matrix_ac_15_transpose_y_0, x = transpose_65, y = transpose_66)[name = tensor("matrix_ac_15_cast_fp16")]; - tensor matrix_bd_31_begin_0 = const()[name = tensor("matrix_bd_31_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_31_end_0 = const()[name = tensor("matrix_bd_31_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_31_end_mask_0 = const()[name = tensor("matrix_bd_31_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_31_cast_fp16 = slice_by_index(begin = matrix_bd_31_begin_0, end = matrix_bd_31_end_0, end_mask = matrix_bd_31_end_mask_0, x = matrix_bd_29_cast_fp16)[name = tensor("matrix_bd_31_cast_fp16")]; - tensor var_1636_cast_fp16 = add(x = matrix_ac_15_cast_fp16, y = matrix_bd_31_cast_fp16)[name = tensor("op_1636_cast_fp16")]; - tensor _inversed_scores_29_y_0_to_fp16 = const()[name = tensor("_inversed_scores_29_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_29_cast_fp16 = mul(x = var_1636_cast_fp16, y = _inversed_scores_29_y_0_to_fp16)[name = tensor("_inversed_scores_29_cast_fp16")]; - tensor scores_31_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_29_cast_fp16, cond = mask_11)[name = tensor("scores_31_cast_fp16")]; - tensor var_1642_cast_fp16 = softmax(axis = var_23, x = scores_31_cast_fp16)[name = tensor("op_1642_cast_fp16")]; - tensor input_397_cast_fp16 = select(a = var_11_to_fp16, b = var_1642_cast_fp16, cond = mask_11)[name = tensor("input_397_cast_fp16")]; - tensor x_167_transpose_x_0 = const()[name = tensor("x_167_transpose_x_0"), val = tensor(false)]; - tensor x_167_transpose_y_0 = const()[name = tensor("x_167_transpose_y_0"), val = tensor(false)]; - tensor value_17_cast_fp16 = transpose(perm = value_17_perm_0, x = v_15_cast_fp16)[name = tensor("transpose_155")]; - tensor x_167_cast_fp16 = matmul(transpose_x = x_167_transpose_x_0, transpose_y = x_167_transpose_y_0, x = input_397_cast_fp16, y = value_17_cast_fp16)[name = tensor("x_167_cast_fp16")]; - tensor var_1646_perm_0 = const()[name = tensor("op_1646_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_1647 = const()[name = tensor("op_1647"), val = tensor([1, -1, 512])]; - tensor var_1646_cast_fp16 = transpose(perm = var_1646_perm_0, x = x_167_cast_fp16)[name = tensor("transpose_151")]; - tensor input_399_cast_fp16 = reshape(shape = var_1647, x = var_1646_cast_fp16)[name = tensor("input_399_cast_fp16")]; - tensor module_layers_7_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_7_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96462016)))]; - tensor module_layers_7_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_7_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96986368)))]; - tensor linear_70_cast_fp16 = linear(bias = module_layers_7_self_attn_linear_out_bias_to_fp16, weight = module_layers_7_self_attn_linear_out_weight_to_fp16, x = input_399_cast_fp16)[name = tensor("linear_70_cast_fp16")]; - tensor input_403_cast_fp16 = add(x = input_395_cast_fp16, y = linear_70_cast_fp16)[name = tensor("input_403_cast_fp16")]; - tensor x_171_axes_0 = const()[name = tensor("x_171_axes_0"), val = tensor([-1])]; - tensor module_layers_7_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_7_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96987456)))]; - tensor module_layers_7_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_7_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96988544)))]; - tensor x_171_cast_fp16 = layer_norm(axes = x_171_axes_0, beta = module_layers_7_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_7_norm_conv_weight_to_fp16, x = input_403_cast_fp16)[name = tensor("x_171_cast_fp16")]; - tensor input_405_perm_0 = const()[name = tensor("input_405_perm_0"), val = tensor([0, 2, 1])]; - tensor input_407_pad_type_0 = const()[name = tensor("input_407_pad_type_0"), val = tensor("valid")]; - tensor input_407_strides_0 = const()[name = tensor("input_407_strides_0"), val = tensor([1])]; - tensor input_407_pad_0 = const()[name = tensor("input_407_pad_0"), val = tensor([0, 0])]; - tensor input_407_dilations_0 = const()[name = tensor("input_407_dilations_0"), val = tensor([1])]; - tensor input_407_groups_0 = const()[name = tensor("input_407_groups_0"), val = tensor(1)]; - tensor module_layers_7_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_7_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96989632)))]; - tensor module_layers_7_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_7_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98038272)))]; - tensor input_405_cast_fp16 = transpose(perm = input_405_perm_0, x = x_171_cast_fp16)[name = tensor("transpose_150")]; - tensor input_407_cast_fp16 = conv(bias = module_layers_7_conv_pointwise_conv1_bias_to_fp16, dilations = input_407_dilations_0, groups = input_407_groups_0, pad = input_407_pad_0, pad_type = input_407_pad_type_0, strides = input_407_strides_0, weight = module_layers_7_conv_pointwise_conv1_weight_to_fp16, x = input_405_cast_fp16)[name = tensor("input_407_cast_fp16")]; - tensor x_173_split_num_splits_0 = const()[name = tensor("x_173_split_num_splits_0"), val = tensor(2)]; - tensor x_173_split_axis_0 = const()[name = tensor("x_173_split_axis_0"), val = tensor(1)]; - tensor x_173_split_cast_fp16_0, tensor x_173_split_cast_fp16_1 = split(axis = x_173_split_axis_0, num_splits = x_173_split_num_splits_0, x = input_407_cast_fp16)[name = tensor("x_173_split_cast_fp16")]; - tensor x_173_split_1_sigmoid_cast_fp16 = sigmoid(x = x_173_split_cast_fp16_1)[name = tensor("x_173_split_1_sigmoid_cast_fp16")]; - tensor x_173_cast_fp16 = mul(x = x_173_split_cast_fp16_0, y = x_173_split_1_sigmoid_cast_fp16)[name = tensor("x_173_cast_fp16")]; - tensor input_409_cast_fp16 = select(a = var_11_to_fp16, b = x_173_cast_fp16, cond = var_453)[name = tensor("input_409_cast_fp16")]; - tensor input_411_pad_0 = const()[name = tensor("input_411_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_411_mode_0 = const()[name = tensor("input_411_mode_0"), val = tensor("constant")]; - tensor const_143_to_fp16 = const()[name = tensor("const_143_to_fp16"), val = tensor(0x0p+0)]; - tensor input_411_cast_fp16 = pad(constant_val = const_143_to_fp16, mode = input_411_mode_0, pad = input_411_pad_0, x = input_409_cast_fp16)[name = tensor("input_411_cast_fp16")]; - tensor input_413_pad_type_0 = const()[name = tensor("input_413_pad_type_0"), val = tensor("valid")]; - tensor input_413_groups_0 = const()[name = tensor("input_413_groups_0"), val = tensor(512)]; - tensor input_413_strides_0 = const()[name = tensor("input_413_strides_0"), val = tensor([1])]; - tensor input_413_pad_0 = const()[name = tensor("input_413_pad_0"), val = tensor([0, 0])]; - tensor input_413_dilations_0 = const()[name = tensor("input_413_dilations_0"), val = tensor([1])]; - tensor const_248_to_fp16 = const()[name = tensor("const_248_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98040384)))]; - tensor const_249_to_fp16 = const()[name = tensor("const_249_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98049664)))]; - tensor input_415_cast_fp16 = conv(bias = const_249_to_fp16, dilations = input_413_dilations_0, groups = input_413_groups_0, pad = input_413_pad_0, pad_type = input_413_pad_type_0, strides = input_413_strides_0, weight = const_248_to_fp16, x = input_411_cast_fp16)[name = tensor("input_415_cast_fp16")]; - tensor input_417_cast_fp16 = silu(x = input_415_cast_fp16)[name = tensor("input_417_cast_fp16")]; - tensor x_175_pad_type_0 = const()[name = tensor("x_175_pad_type_0"), val = tensor("valid")]; - tensor x_175_strides_0 = const()[name = tensor("x_175_strides_0"), val = tensor([1])]; - tensor x_175_pad_0 = const()[name = tensor("x_175_pad_0"), val = tensor([0, 0])]; - tensor x_175_dilations_0 = const()[name = tensor("x_175_dilations_0"), val = tensor([1])]; - tensor x_175_groups_0 = const()[name = tensor("x_175_groups_0"), val = tensor(1)]; - tensor module_layers_7_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_7_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98050752)))]; - tensor module_layers_7_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_7_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98575104)))]; - tensor x_175_cast_fp16 = conv(bias = module_layers_7_conv_pointwise_conv2_bias_to_fp16, dilations = x_175_dilations_0, groups = x_175_groups_0, pad = x_175_pad_0, pad_type = x_175_pad_type_0, strides = x_175_strides_0, weight = module_layers_7_conv_pointwise_conv2_weight_to_fp16, x = input_417_cast_fp16)[name = tensor("x_175_cast_fp16")]; - tensor input_419_perm_0 = const()[name = tensor("input_419_perm_0"), val = tensor([0, 2, 1])]; - tensor input_419_cast_fp16 = transpose(perm = input_419_perm_0, x = x_175_cast_fp16)[name = tensor("transpose_149")]; - tensor input_421_cast_fp16 = add(x = input_403_cast_fp16, y = input_419_cast_fp16)[name = tensor("input_421_cast_fp16")]; - tensor input_423_axes_0 = const()[name = tensor("input_423_axes_0"), val = tensor([-1])]; - tensor module_layers_7_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_7_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98576192)))]; - tensor module_layers_7_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_7_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98577280)))]; - tensor input_423_cast_fp16 = layer_norm(axes = input_423_axes_0, beta = module_layers_7_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_7_norm_feed_forward2_weight_to_fp16, x = input_421_cast_fp16)[name = tensor("input_423_cast_fp16")]; - tensor module_layers_7_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_7_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98578368)))]; - tensor module_layers_7_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_7_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(100675584)))]; - tensor linear_71_cast_fp16 = linear(bias = module_layers_7_feed_forward2_linear1_bias_to_fp16, weight = module_layers_7_feed_forward2_linear1_weight_to_fp16, x = input_423_cast_fp16)[name = tensor("linear_71_cast_fp16")]; - tensor input_427_cast_fp16 = silu(x = linear_71_cast_fp16)[name = tensor("input_427_cast_fp16")]; - tensor module_layers_7_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_7_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(100679744)))]; - tensor module_layers_7_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_7_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102776960)))]; - tensor linear_72_cast_fp16 = linear(bias = module_layers_7_feed_forward2_linear2_bias_to_fp16, weight = module_layers_7_feed_forward2_linear2_weight_to_fp16, x = input_427_cast_fp16)[name = tensor("linear_72_cast_fp16")]; - tensor var_1713_to_fp16 = const()[name = tensor("op_1713_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1714_cast_fp16 = mul(x = linear_72_cast_fp16, y = var_1713_to_fp16)[name = tensor("op_1714_cast_fp16")]; - tensor input_433_cast_fp16 = add(x = input_421_cast_fp16, y = var_1714_cast_fp16)[name = tensor("input_433_cast_fp16")]; - tensor input_435_axes_0 = const()[name = tensor("input_435_axes_0"), val = tensor([-1])]; - tensor module_layers_7_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_7_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102778048)))]; - tensor module_layers_7_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_7_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102779136)))]; - tensor input_435_cast_fp16 = layer_norm(axes = input_435_axes_0, beta = module_layers_7_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_7_norm_out_weight_to_fp16, x = input_433_cast_fp16)[name = tensor("input_435_cast_fp16")]; - tensor input_437_axes_0 = const()[name = tensor("input_437_axes_0"), val = tensor([-1])]; - tensor module_layers_8_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_8_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102780224)))]; - tensor module_layers_8_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_8_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102781312)))]; - tensor input_437_cast_fp16 = layer_norm(axes = input_437_axes_0, beta = module_layers_8_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_8_norm_feed_forward1_weight_to_fp16, x = input_435_cast_fp16)[name = tensor("input_437_cast_fp16")]; - tensor module_layers_8_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_8_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102782400)))]; - tensor module_layers_8_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_8_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(104879616)))]; - tensor linear_73_cast_fp16 = linear(bias = module_layers_8_feed_forward1_linear1_bias_to_fp16, weight = module_layers_8_feed_forward1_linear1_weight_to_fp16, x = input_437_cast_fp16)[name = tensor("linear_73_cast_fp16")]; - tensor input_441_cast_fp16 = silu(x = linear_73_cast_fp16)[name = tensor("input_441_cast_fp16")]; - tensor module_layers_8_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_8_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(104883776)))]; - tensor module_layers_8_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_8_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(106980992)))]; - tensor linear_74_cast_fp16 = linear(bias = module_layers_8_feed_forward1_linear2_bias_to_fp16, weight = module_layers_8_feed_forward1_linear2_weight_to_fp16, x = input_441_cast_fp16)[name = tensor("linear_74_cast_fp16")]; - tensor var_1744_to_fp16 = const()[name = tensor("op_1744_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1745_cast_fp16 = mul(x = linear_74_cast_fp16, y = var_1744_to_fp16)[name = tensor("op_1745_cast_fp16")]; - tensor input_447_cast_fp16 = add(x = input_435_cast_fp16, y = var_1745_cast_fp16)[name = tensor("input_447_cast_fp16")]; - tensor query_17_axes_0 = const()[name = tensor("query_17_axes_0"), val = tensor([-1])]; - tensor module_layers_8_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_8_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(106982080)))]; - tensor module_layers_8_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_8_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(106983168)))]; - tensor query_17_cast_fp16 = layer_norm(axes = query_17_axes_0, beta = module_layers_8_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_8_norm_self_att_weight_to_fp16, x = input_447_cast_fp16)[name = tensor("query_17_cast_fp16")]; - tensor module_layers_8_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_8_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(106984256)))]; - tensor module_layers_8_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_8_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(107508608)))]; - tensor linear_75_cast_fp16 = linear(bias = module_layers_8_self_attn_linear_q_bias_to_fp16, weight = module_layers_8_self_attn_linear_q_weight_to_fp16, x = query_17_cast_fp16)[name = tensor("linear_75_cast_fp16")]; - tensor var_1762 = const()[name = tensor("op_1762"), val = tensor([1, -1, 8, 64])]; - tensor q_49_cast_fp16 = reshape(shape = var_1762, x = linear_75_cast_fp16)[name = tensor("q_49_cast_fp16")]; - tensor module_layers_8_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_8_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(107509696)))]; - tensor module_layers_8_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_8_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108034048)))]; - tensor linear_76_cast_fp16 = linear(bias = module_layers_8_self_attn_linear_k_bias_to_fp16, weight = module_layers_8_self_attn_linear_k_weight_to_fp16, x = query_17_cast_fp16)[name = tensor("linear_76_cast_fp16")]; - tensor var_1767 = const()[name = tensor("op_1767"), val = tensor([1, -1, 8, 64])]; - tensor k_33_cast_fp16 = reshape(shape = var_1767, x = linear_76_cast_fp16)[name = tensor("k_33_cast_fp16")]; - tensor module_layers_8_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_8_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108035136)))]; - tensor module_layers_8_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_8_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108559488)))]; - tensor linear_77_cast_fp16 = linear(bias = module_layers_8_self_attn_linear_v_bias_to_fp16, weight = module_layers_8_self_attn_linear_v_weight_to_fp16, x = query_17_cast_fp16)[name = tensor("linear_77_cast_fp16")]; - tensor var_1772 = const()[name = tensor("op_1772"), val = tensor([1, -1, 8, 64])]; - tensor v_17_cast_fp16 = reshape(shape = var_1772, x = linear_77_cast_fp16)[name = tensor("v_17_cast_fp16")]; - tensor value_19_perm_0 = const()[name = tensor("value_19_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_8_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_8_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108560576)))]; - tensor var_1784_cast_fp16 = add(x = q_49_cast_fp16, y = module_layers_8_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1784_cast_fp16")]; - tensor module_layers_8_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_8_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108561664)))]; - tensor var_1786_cast_fp16 = add(x = q_49_cast_fp16, y = module_layers_8_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1786_cast_fp16")]; - tensor q_with_bias_v_17_perm_0 = const()[name = tensor("q_with_bias_v_17_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_183_transpose_x_0 = const()[name = tensor("x_183_transpose_x_0"), val = tensor(false)]; - tensor x_183_transpose_y_0 = const()[name = tensor("x_183_transpose_y_0"), val = tensor(false)]; - tensor var_1788_to_fp16 = const()[name = tensor("op_1788_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108562752)))]; - tensor q_with_bias_v_17_cast_fp16 = transpose(perm = q_with_bias_v_17_perm_0, x = var_1786_cast_fp16)[name = tensor("transpose_147")]; - tensor x_183_cast_fp16 = matmul(transpose_x = x_183_transpose_x_0, transpose_y = x_183_transpose_y_0, x = q_with_bias_v_17_cast_fp16, y = var_1788_to_fp16)[name = tensor("x_183_cast_fp16")]; - tensor x_185_pad_0 = const()[name = tensor("x_185_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_185_mode_0 = const()[name = tensor("x_185_mode_0"), val = tensor("constant")]; - tensor const_150_to_fp16 = const()[name = tensor("const_150_to_fp16"), val = tensor(0x0p+0)]; - tensor x_185_cast_fp16 = pad(constant_val = const_150_to_fp16, mode = x_185_mode_0, pad = x_185_pad_0, x = x_183_cast_fp16)[name = tensor("x_185_cast_fp16")]; - tensor var_1796 = const()[name = tensor("op_1796"), val = tensor([1, 8, -1, 188])]; - tensor x_187_cast_fp16 = reshape(shape = var_1796, x = x_185_cast_fp16)[name = tensor("x_187_cast_fp16")]; - tensor var_1800_begin_0 = const()[name = tensor("op_1800_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_1800_end_0 = const()[name = tensor("op_1800_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_1800_end_mask_0 = const()[name = tensor("op_1800_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_1800_cast_fp16 = slice_by_index(begin = var_1800_begin_0, end = var_1800_end_0, end_mask = var_1800_end_mask_0, x = x_187_cast_fp16)[name = tensor("op_1800_cast_fp16")]; - tensor var_1801 = const()[name = tensor("op_1801"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_33_cast_fp16 = reshape(shape = var_1801, x = var_1800_cast_fp16)[name = tensor("matrix_bd_33_cast_fp16")]; - tensor matrix_ac_17_transpose_x_0 = const()[name = tensor("matrix_ac_17_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_17_transpose_y_0 = const()[name = tensor("matrix_ac_17_transpose_y_0"), val = tensor(false)]; - tensor transpose_67_perm_0 = const()[name = tensor("transpose_67_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_68_perm_0 = const()[name = tensor("transpose_68_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_68 = transpose(perm = transpose_68_perm_0, x = k_33_cast_fp16)[name = tensor("transpose_145")]; - tensor transpose_67 = transpose(perm = transpose_67_perm_0, x = var_1784_cast_fp16)[name = tensor("transpose_146")]; - tensor matrix_ac_17_cast_fp16 = matmul(transpose_x = matrix_ac_17_transpose_x_0, transpose_y = matrix_ac_17_transpose_y_0, x = transpose_67, y = transpose_68)[name = tensor("matrix_ac_17_cast_fp16")]; - tensor matrix_bd_35_begin_0 = const()[name = tensor("matrix_bd_35_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_35_end_0 = const()[name = tensor("matrix_bd_35_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_35_end_mask_0 = const()[name = tensor("matrix_bd_35_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_35_cast_fp16 = slice_by_index(begin = matrix_bd_35_begin_0, end = matrix_bd_35_end_0, end_mask = matrix_bd_35_end_mask_0, x = matrix_bd_33_cast_fp16)[name = tensor("matrix_bd_35_cast_fp16")]; - tensor var_1810_cast_fp16 = add(x = matrix_ac_17_cast_fp16, y = matrix_bd_35_cast_fp16)[name = tensor("op_1810_cast_fp16")]; - tensor _inversed_scores_33_y_0_to_fp16 = const()[name = tensor("_inversed_scores_33_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_33_cast_fp16 = mul(x = var_1810_cast_fp16, y = _inversed_scores_33_y_0_to_fp16)[name = tensor("_inversed_scores_33_cast_fp16")]; - tensor scores_35_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_33_cast_fp16, cond = mask_11)[name = tensor("scores_35_cast_fp16")]; - tensor var_1816_cast_fp16 = softmax(axis = var_23, x = scores_35_cast_fp16)[name = tensor("op_1816_cast_fp16")]; - tensor input_449_cast_fp16 = select(a = var_11_to_fp16, b = var_1816_cast_fp16, cond = mask_11)[name = tensor("input_449_cast_fp16")]; - tensor x_189_transpose_x_0 = const()[name = tensor("x_189_transpose_x_0"), val = tensor(false)]; - tensor x_189_transpose_y_0 = const()[name = tensor("x_189_transpose_y_0"), val = tensor(false)]; - tensor value_19_cast_fp16 = transpose(perm = value_19_perm_0, x = v_17_cast_fp16)[name = tensor("transpose_148")]; - tensor x_189_cast_fp16 = matmul(transpose_x = x_189_transpose_x_0, transpose_y = x_189_transpose_y_0, x = input_449_cast_fp16, y = value_19_cast_fp16)[name = tensor("x_189_cast_fp16")]; - tensor var_1820_perm_0 = const()[name = tensor("op_1820_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_1821 = const()[name = tensor("op_1821"), val = tensor([1, -1, 512])]; - tensor var_1820_cast_fp16 = transpose(perm = var_1820_perm_0, x = x_189_cast_fp16)[name = tensor("transpose_144")]; - tensor input_451_cast_fp16 = reshape(shape = var_1821, x = var_1820_cast_fp16)[name = tensor("input_451_cast_fp16")]; - tensor module_layers_8_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_8_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108946816)))]; - tensor module_layers_8_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_8_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109471168)))]; - tensor linear_79_cast_fp16 = linear(bias = module_layers_8_self_attn_linear_out_bias_to_fp16, weight = module_layers_8_self_attn_linear_out_weight_to_fp16, x = input_451_cast_fp16)[name = tensor("linear_79_cast_fp16")]; - tensor input_455_cast_fp16 = add(x = input_447_cast_fp16, y = linear_79_cast_fp16)[name = tensor("input_455_cast_fp16")]; - tensor x_193_axes_0 = const()[name = tensor("x_193_axes_0"), val = tensor([-1])]; - tensor module_layers_8_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_8_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109472256)))]; - tensor module_layers_8_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_8_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109473344)))]; - tensor x_193_cast_fp16 = layer_norm(axes = x_193_axes_0, beta = module_layers_8_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_8_norm_conv_weight_to_fp16, x = input_455_cast_fp16)[name = tensor("x_193_cast_fp16")]; - tensor input_457_perm_0 = const()[name = tensor("input_457_perm_0"), val = tensor([0, 2, 1])]; - tensor input_459_pad_type_0 = const()[name = tensor("input_459_pad_type_0"), val = tensor("valid")]; - tensor input_459_strides_0 = const()[name = tensor("input_459_strides_0"), val = tensor([1])]; - tensor input_459_pad_0 = const()[name = tensor("input_459_pad_0"), val = tensor([0, 0])]; - tensor input_459_dilations_0 = const()[name = tensor("input_459_dilations_0"), val = tensor([1])]; - tensor input_459_groups_0 = const()[name = tensor("input_459_groups_0"), val = tensor(1)]; - tensor module_layers_8_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_8_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109474432)))]; - tensor module_layers_8_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_8_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110523072)))]; - tensor input_457_cast_fp16 = transpose(perm = input_457_perm_0, x = x_193_cast_fp16)[name = tensor("transpose_143")]; - tensor input_459_cast_fp16 = conv(bias = module_layers_8_conv_pointwise_conv1_bias_to_fp16, dilations = input_459_dilations_0, groups = input_459_groups_0, pad = input_459_pad_0, pad_type = input_459_pad_type_0, strides = input_459_strides_0, weight = module_layers_8_conv_pointwise_conv1_weight_to_fp16, x = input_457_cast_fp16)[name = tensor("input_459_cast_fp16")]; - tensor x_195_split_num_splits_0 = const()[name = tensor("x_195_split_num_splits_0"), val = tensor(2)]; - tensor x_195_split_axis_0 = const()[name = tensor("x_195_split_axis_0"), val = tensor(1)]; - tensor x_195_split_cast_fp16_0, tensor x_195_split_cast_fp16_1 = split(axis = x_195_split_axis_0, num_splits = x_195_split_num_splits_0, x = input_459_cast_fp16)[name = tensor("x_195_split_cast_fp16")]; - tensor x_195_split_1_sigmoid_cast_fp16 = sigmoid(x = x_195_split_cast_fp16_1)[name = tensor("x_195_split_1_sigmoid_cast_fp16")]; - tensor x_195_cast_fp16 = mul(x = x_195_split_cast_fp16_0, y = x_195_split_1_sigmoid_cast_fp16)[name = tensor("x_195_cast_fp16")]; - tensor input_461_cast_fp16 = select(a = var_11_to_fp16, b = x_195_cast_fp16, cond = var_453)[name = tensor("input_461_cast_fp16")]; - tensor input_463_pad_0 = const()[name = tensor("input_463_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_463_mode_0 = const()[name = tensor("input_463_mode_0"), val = tensor("constant")]; - tensor const_153_to_fp16 = const()[name = tensor("const_153_to_fp16"), val = tensor(0x0p+0)]; - tensor input_463_cast_fp16 = pad(constant_val = const_153_to_fp16, mode = input_463_mode_0, pad = input_463_pad_0, x = input_461_cast_fp16)[name = tensor("input_463_cast_fp16")]; - tensor input_465_pad_type_0 = const()[name = tensor("input_465_pad_type_0"), val = tensor("valid")]; - tensor input_465_groups_0 = const()[name = tensor("input_465_groups_0"), val = tensor(512)]; - tensor input_465_strides_0 = const()[name = tensor("input_465_strides_0"), val = tensor([1])]; - tensor input_465_pad_0 = const()[name = tensor("input_465_pad_0"), val = tensor([0, 0])]; - tensor input_465_dilations_0 = const()[name = tensor("input_465_dilations_0"), val = tensor([1])]; - tensor const_250_to_fp16 = const()[name = tensor("const_250_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110525184)))]; - tensor const_251_to_fp16 = const()[name = tensor("const_251_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110534464)))]; - tensor input_467_cast_fp16 = conv(bias = const_251_to_fp16, dilations = input_465_dilations_0, groups = input_465_groups_0, pad = input_465_pad_0, pad_type = input_465_pad_type_0, strides = input_465_strides_0, weight = const_250_to_fp16, x = input_463_cast_fp16)[name = tensor("input_467_cast_fp16")]; - tensor input_469_cast_fp16 = silu(x = input_467_cast_fp16)[name = tensor("input_469_cast_fp16")]; - tensor x_197_pad_type_0 = const()[name = tensor("x_197_pad_type_0"), val = tensor("valid")]; - tensor x_197_strides_0 = const()[name = tensor("x_197_strides_0"), val = tensor([1])]; - tensor x_197_pad_0 = const()[name = tensor("x_197_pad_0"), val = tensor([0, 0])]; - tensor x_197_dilations_0 = const()[name = tensor("x_197_dilations_0"), val = tensor([1])]; - tensor x_197_groups_0 = const()[name = tensor("x_197_groups_0"), val = tensor(1)]; - tensor module_layers_8_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_8_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110535552)))]; - tensor module_layers_8_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_8_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(111059904)))]; - tensor x_197_cast_fp16 = conv(bias = module_layers_8_conv_pointwise_conv2_bias_to_fp16, dilations = x_197_dilations_0, groups = x_197_groups_0, pad = x_197_pad_0, pad_type = x_197_pad_type_0, strides = x_197_strides_0, weight = module_layers_8_conv_pointwise_conv2_weight_to_fp16, x = input_469_cast_fp16)[name = tensor("x_197_cast_fp16")]; - tensor input_471_perm_0 = const()[name = tensor("input_471_perm_0"), val = tensor([0, 2, 1])]; - tensor input_471_cast_fp16 = transpose(perm = input_471_perm_0, x = x_197_cast_fp16)[name = tensor("transpose_142")]; - tensor input_473_cast_fp16 = add(x = input_455_cast_fp16, y = input_471_cast_fp16)[name = tensor("input_473_cast_fp16")]; - tensor input_475_axes_0 = const()[name = tensor("input_475_axes_0"), val = tensor([-1])]; - tensor module_layers_8_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_8_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(111060992)))]; - tensor module_layers_8_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_8_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(111062080)))]; - tensor input_475_cast_fp16 = layer_norm(axes = input_475_axes_0, beta = module_layers_8_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_8_norm_feed_forward2_weight_to_fp16, x = input_473_cast_fp16)[name = tensor("input_475_cast_fp16")]; - tensor module_layers_8_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_8_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(111063168)))]; - tensor module_layers_8_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_8_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(113160384)))]; - tensor linear_80_cast_fp16 = linear(bias = module_layers_8_feed_forward2_linear1_bias_to_fp16, weight = module_layers_8_feed_forward2_linear1_weight_to_fp16, x = input_475_cast_fp16)[name = tensor("linear_80_cast_fp16")]; - tensor input_479_cast_fp16 = silu(x = linear_80_cast_fp16)[name = tensor("input_479_cast_fp16")]; - tensor module_layers_8_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_8_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(113164544)))]; - tensor module_layers_8_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_8_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(115261760)))]; - tensor linear_81_cast_fp16 = linear(bias = module_layers_8_feed_forward2_linear2_bias_to_fp16, weight = module_layers_8_feed_forward2_linear2_weight_to_fp16, x = input_479_cast_fp16)[name = tensor("linear_81_cast_fp16")]; - tensor var_1887_to_fp16 = const()[name = tensor("op_1887_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1888_cast_fp16 = mul(x = linear_81_cast_fp16, y = var_1887_to_fp16)[name = tensor("op_1888_cast_fp16")]; - tensor input_485_cast_fp16 = add(x = input_473_cast_fp16, y = var_1888_cast_fp16)[name = tensor("input_485_cast_fp16")]; - tensor input_487_axes_0 = const()[name = tensor("input_487_axes_0"), val = tensor([-1])]; - tensor module_layers_8_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_8_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(115262848)))]; - tensor module_layers_8_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_8_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(115263936)))]; - tensor input_487_cast_fp16 = layer_norm(axes = input_487_axes_0, beta = module_layers_8_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_8_norm_out_weight_to_fp16, x = input_485_cast_fp16)[name = tensor("input_487_cast_fp16")]; - tensor input_489_axes_0 = const()[name = tensor("input_489_axes_0"), val = tensor([-1])]; - tensor module_layers_9_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_9_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(115265024)))]; - tensor module_layers_9_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_9_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(115266112)))]; - tensor input_489_cast_fp16 = layer_norm(axes = input_489_axes_0, beta = module_layers_9_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_9_norm_feed_forward1_weight_to_fp16, x = input_487_cast_fp16)[name = tensor("input_489_cast_fp16")]; - tensor module_layers_9_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_9_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(115267200)))]; - tensor module_layers_9_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_9_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(117364416)))]; - tensor linear_82_cast_fp16 = linear(bias = module_layers_9_feed_forward1_linear1_bias_to_fp16, weight = module_layers_9_feed_forward1_linear1_weight_to_fp16, x = input_489_cast_fp16)[name = tensor("linear_82_cast_fp16")]; - tensor input_493_cast_fp16 = silu(x = linear_82_cast_fp16)[name = tensor("input_493_cast_fp16")]; - tensor module_layers_9_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_9_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(117368576)))]; - tensor module_layers_9_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_9_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(119465792)))]; - tensor linear_83_cast_fp16 = linear(bias = module_layers_9_feed_forward1_linear2_bias_to_fp16, weight = module_layers_9_feed_forward1_linear2_weight_to_fp16, x = input_493_cast_fp16)[name = tensor("linear_83_cast_fp16")]; - tensor var_1918_to_fp16 = const()[name = tensor("op_1918_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1919_cast_fp16 = mul(x = linear_83_cast_fp16, y = var_1918_to_fp16)[name = tensor("op_1919_cast_fp16")]; - tensor input_499_cast_fp16 = add(x = input_487_cast_fp16, y = var_1919_cast_fp16)[name = tensor("input_499_cast_fp16")]; - tensor query_19_axes_0 = const()[name = tensor("query_19_axes_0"), val = tensor([-1])]; - tensor module_layers_9_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_9_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(119466880)))]; - tensor module_layers_9_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_9_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(119467968)))]; - tensor query_19_cast_fp16 = layer_norm(axes = query_19_axes_0, beta = module_layers_9_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_9_norm_self_att_weight_to_fp16, x = input_499_cast_fp16)[name = tensor("query_19_cast_fp16")]; - tensor module_layers_9_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_9_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(119469056)))]; - tensor module_layers_9_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_9_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(119993408)))]; - tensor linear_84_cast_fp16 = linear(bias = module_layers_9_self_attn_linear_q_bias_to_fp16, weight = module_layers_9_self_attn_linear_q_weight_to_fp16, x = query_19_cast_fp16)[name = tensor("linear_84_cast_fp16")]; - tensor var_1936 = const()[name = tensor("op_1936"), val = tensor([1, -1, 8, 64])]; - tensor q_55_cast_fp16 = reshape(shape = var_1936, x = linear_84_cast_fp16)[name = tensor("q_55_cast_fp16")]; - tensor module_layers_9_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_9_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(119994496)))]; - tensor module_layers_9_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_9_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(120518848)))]; - tensor linear_85_cast_fp16 = linear(bias = module_layers_9_self_attn_linear_k_bias_to_fp16, weight = module_layers_9_self_attn_linear_k_weight_to_fp16, x = query_19_cast_fp16)[name = tensor("linear_85_cast_fp16")]; - tensor var_1941 = const()[name = tensor("op_1941"), val = tensor([1, -1, 8, 64])]; - tensor k_37_cast_fp16 = reshape(shape = var_1941, x = linear_85_cast_fp16)[name = tensor("k_37_cast_fp16")]; - tensor module_layers_9_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_9_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(120519936)))]; - tensor module_layers_9_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_9_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121044288)))]; - tensor linear_86_cast_fp16 = linear(bias = module_layers_9_self_attn_linear_v_bias_to_fp16, weight = module_layers_9_self_attn_linear_v_weight_to_fp16, x = query_19_cast_fp16)[name = tensor("linear_86_cast_fp16")]; - tensor var_1946 = const()[name = tensor("op_1946"), val = tensor([1, -1, 8, 64])]; - tensor v_19_cast_fp16 = reshape(shape = var_1946, x = linear_86_cast_fp16)[name = tensor("v_19_cast_fp16")]; - tensor value_21_perm_0 = const()[name = tensor("value_21_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_9_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_9_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121045376)))]; - tensor var_1958_cast_fp16 = add(x = q_55_cast_fp16, y = module_layers_9_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1958_cast_fp16")]; - tensor module_layers_9_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_9_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121046464)))]; - tensor var_1960_cast_fp16 = add(x = q_55_cast_fp16, y = module_layers_9_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1960_cast_fp16")]; - tensor q_with_bias_v_19_perm_0 = const()[name = tensor("q_with_bias_v_19_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_205_transpose_x_0 = const()[name = tensor("x_205_transpose_x_0"), val = tensor(false)]; - tensor x_205_transpose_y_0 = const()[name = tensor("x_205_transpose_y_0"), val = tensor(false)]; - tensor var_1962_to_fp16 = const()[name = tensor("op_1962_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121047552)))]; - tensor q_with_bias_v_19_cast_fp16 = transpose(perm = q_with_bias_v_19_perm_0, x = var_1960_cast_fp16)[name = tensor("transpose_140")]; - tensor x_205_cast_fp16 = matmul(transpose_x = x_205_transpose_x_0, transpose_y = x_205_transpose_y_0, x = q_with_bias_v_19_cast_fp16, y = var_1962_to_fp16)[name = tensor("x_205_cast_fp16")]; - tensor x_207_pad_0 = const()[name = tensor("x_207_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_207_mode_0 = const()[name = tensor("x_207_mode_0"), val = tensor("constant")]; - tensor const_160_to_fp16 = const()[name = tensor("const_160_to_fp16"), val = tensor(0x0p+0)]; - tensor x_207_cast_fp16 = pad(constant_val = const_160_to_fp16, mode = x_207_mode_0, pad = x_207_pad_0, x = x_205_cast_fp16)[name = tensor("x_207_cast_fp16")]; - tensor var_1970 = const()[name = tensor("op_1970"), val = tensor([1, 8, -1, 188])]; - tensor x_209_cast_fp16 = reshape(shape = var_1970, x = x_207_cast_fp16)[name = tensor("x_209_cast_fp16")]; - tensor var_1974_begin_0 = const()[name = tensor("op_1974_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_1974_end_0 = const()[name = tensor("op_1974_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_1974_end_mask_0 = const()[name = tensor("op_1974_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_1974_cast_fp16 = slice_by_index(begin = var_1974_begin_0, end = var_1974_end_0, end_mask = var_1974_end_mask_0, x = x_209_cast_fp16)[name = tensor("op_1974_cast_fp16")]; - tensor var_1975 = const()[name = tensor("op_1975"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_37_cast_fp16 = reshape(shape = var_1975, x = var_1974_cast_fp16)[name = tensor("matrix_bd_37_cast_fp16")]; - tensor matrix_ac_19_transpose_x_0 = const()[name = tensor("matrix_ac_19_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_19_transpose_y_0 = const()[name = tensor("matrix_ac_19_transpose_y_0"), val = tensor(false)]; - tensor transpose_69_perm_0 = const()[name = tensor("transpose_69_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_70_perm_0 = const()[name = tensor("transpose_70_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_70 = transpose(perm = transpose_70_perm_0, x = k_37_cast_fp16)[name = tensor("transpose_138")]; - tensor transpose_69 = transpose(perm = transpose_69_perm_0, x = var_1958_cast_fp16)[name = tensor("transpose_139")]; - tensor matrix_ac_19_cast_fp16 = matmul(transpose_x = matrix_ac_19_transpose_x_0, transpose_y = matrix_ac_19_transpose_y_0, x = transpose_69, y = transpose_70)[name = tensor("matrix_ac_19_cast_fp16")]; - tensor matrix_bd_39_begin_0 = const()[name = tensor("matrix_bd_39_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_39_end_0 = const()[name = tensor("matrix_bd_39_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_39_end_mask_0 = const()[name = tensor("matrix_bd_39_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_39_cast_fp16 = slice_by_index(begin = matrix_bd_39_begin_0, end = matrix_bd_39_end_0, end_mask = matrix_bd_39_end_mask_0, x = matrix_bd_37_cast_fp16)[name = tensor("matrix_bd_39_cast_fp16")]; - tensor var_1984_cast_fp16 = add(x = matrix_ac_19_cast_fp16, y = matrix_bd_39_cast_fp16)[name = tensor("op_1984_cast_fp16")]; - tensor _inversed_scores_37_y_0_to_fp16 = const()[name = tensor("_inversed_scores_37_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_37_cast_fp16 = mul(x = var_1984_cast_fp16, y = _inversed_scores_37_y_0_to_fp16)[name = tensor("_inversed_scores_37_cast_fp16")]; - tensor scores_39_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_37_cast_fp16, cond = mask_11)[name = tensor("scores_39_cast_fp16")]; - tensor var_1990_cast_fp16 = softmax(axis = var_23, x = scores_39_cast_fp16)[name = tensor("op_1990_cast_fp16")]; - tensor input_501_cast_fp16 = select(a = var_11_to_fp16, b = var_1990_cast_fp16, cond = mask_11)[name = tensor("input_501_cast_fp16")]; - tensor x_211_transpose_x_0 = const()[name = tensor("x_211_transpose_x_0"), val = tensor(false)]; - tensor x_211_transpose_y_0 = const()[name = tensor("x_211_transpose_y_0"), val = tensor(false)]; - tensor value_21_cast_fp16 = transpose(perm = value_21_perm_0, x = v_19_cast_fp16)[name = tensor("transpose_141")]; - tensor x_211_cast_fp16 = matmul(transpose_x = x_211_transpose_x_0, transpose_y = x_211_transpose_y_0, x = input_501_cast_fp16, y = value_21_cast_fp16)[name = tensor("x_211_cast_fp16")]; - tensor var_1994_perm_0 = const()[name = tensor("op_1994_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_1995 = const()[name = tensor("op_1995"), val = tensor([1, -1, 512])]; - tensor var_1994_cast_fp16 = transpose(perm = var_1994_perm_0, x = x_211_cast_fp16)[name = tensor("transpose_137")]; - tensor input_503_cast_fp16 = reshape(shape = var_1995, x = var_1994_cast_fp16)[name = tensor("input_503_cast_fp16")]; - tensor module_layers_9_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_9_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121431616)))]; - tensor module_layers_9_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_9_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121955968)))]; - tensor linear_88_cast_fp16 = linear(bias = module_layers_9_self_attn_linear_out_bias_to_fp16, weight = module_layers_9_self_attn_linear_out_weight_to_fp16, x = input_503_cast_fp16)[name = tensor("linear_88_cast_fp16")]; - tensor input_507_cast_fp16 = add(x = input_499_cast_fp16, y = linear_88_cast_fp16)[name = tensor("input_507_cast_fp16")]; - tensor x_215_axes_0 = const()[name = tensor("x_215_axes_0"), val = tensor([-1])]; - tensor module_layers_9_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_9_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121957056)))]; - tensor module_layers_9_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_9_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121958144)))]; - tensor x_215_cast_fp16 = layer_norm(axes = x_215_axes_0, beta = module_layers_9_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_9_norm_conv_weight_to_fp16, x = input_507_cast_fp16)[name = tensor("x_215_cast_fp16")]; - tensor input_509_perm_0 = const()[name = tensor("input_509_perm_0"), val = tensor([0, 2, 1])]; - tensor input_511_pad_type_0 = const()[name = tensor("input_511_pad_type_0"), val = tensor("valid")]; - tensor input_511_strides_0 = const()[name = tensor("input_511_strides_0"), val = tensor([1])]; - tensor input_511_pad_0 = const()[name = tensor("input_511_pad_0"), val = tensor([0, 0])]; - tensor input_511_dilations_0 = const()[name = tensor("input_511_dilations_0"), val = tensor([1])]; - tensor input_511_groups_0 = const()[name = tensor("input_511_groups_0"), val = tensor(1)]; - tensor module_layers_9_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_9_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121959232)))]; - tensor module_layers_9_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_9_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123007872)))]; - tensor input_509_cast_fp16 = transpose(perm = input_509_perm_0, x = x_215_cast_fp16)[name = tensor("transpose_136")]; - tensor input_511_cast_fp16 = conv(bias = module_layers_9_conv_pointwise_conv1_bias_to_fp16, dilations = input_511_dilations_0, groups = input_511_groups_0, pad = input_511_pad_0, pad_type = input_511_pad_type_0, strides = input_511_strides_0, weight = module_layers_9_conv_pointwise_conv1_weight_to_fp16, x = input_509_cast_fp16)[name = tensor("input_511_cast_fp16")]; - tensor x_217_split_num_splits_0 = const()[name = tensor("x_217_split_num_splits_0"), val = tensor(2)]; - tensor x_217_split_axis_0 = const()[name = tensor("x_217_split_axis_0"), val = tensor(1)]; - tensor x_217_split_cast_fp16_0, tensor x_217_split_cast_fp16_1 = split(axis = x_217_split_axis_0, num_splits = x_217_split_num_splits_0, x = input_511_cast_fp16)[name = tensor("x_217_split_cast_fp16")]; - tensor x_217_split_1_sigmoid_cast_fp16 = sigmoid(x = x_217_split_cast_fp16_1)[name = tensor("x_217_split_1_sigmoid_cast_fp16")]; - tensor x_217_cast_fp16 = mul(x = x_217_split_cast_fp16_0, y = x_217_split_1_sigmoid_cast_fp16)[name = tensor("x_217_cast_fp16")]; - tensor input_513_cast_fp16 = select(a = var_11_to_fp16, b = x_217_cast_fp16, cond = var_453)[name = tensor("input_513_cast_fp16")]; - tensor input_515_pad_0 = const()[name = tensor("input_515_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_515_mode_0 = const()[name = tensor("input_515_mode_0"), val = tensor("constant")]; - tensor const_163_to_fp16 = const()[name = tensor("const_163_to_fp16"), val = tensor(0x0p+0)]; - tensor input_515_cast_fp16 = pad(constant_val = const_163_to_fp16, mode = input_515_mode_0, pad = input_515_pad_0, x = input_513_cast_fp16)[name = tensor("input_515_cast_fp16")]; - tensor input_517_pad_type_0 = const()[name = tensor("input_517_pad_type_0"), val = tensor("valid")]; - tensor input_517_groups_0 = const()[name = tensor("input_517_groups_0"), val = tensor(512)]; - tensor input_517_strides_0 = const()[name = tensor("input_517_strides_0"), val = tensor([1])]; - tensor input_517_pad_0 = const()[name = tensor("input_517_pad_0"), val = tensor([0, 0])]; - tensor input_517_dilations_0 = const()[name = tensor("input_517_dilations_0"), val = tensor([1])]; - tensor const_252_to_fp16 = const()[name = tensor("const_252_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123009984)))]; - tensor const_253_to_fp16 = const()[name = tensor("const_253_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123019264)))]; - tensor input_519_cast_fp16 = conv(bias = const_253_to_fp16, dilations = input_517_dilations_0, groups = input_517_groups_0, pad = input_517_pad_0, pad_type = input_517_pad_type_0, strides = input_517_strides_0, weight = const_252_to_fp16, x = input_515_cast_fp16)[name = tensor("input_519_cast_fp16")]; - tensor input_521_cast_fp16 = silu(x = input_519_cast_fp16)[name = tensor("input_521_cast_fp16")]; - tensor x_219_pad_type_0 = const()[name = tensor("x_219_pad_type_0"), val = tensor("valid")]; - tensor x_219_strides_0 = const()[name = tensor("x_219_strides_0"), val = tensor([1])]; - tensor x_219_pad_0 = const()[name = tensor("x_219_pad_0"), val = tensor([0, 0])]; - tensor x_219_dilations_0 = const()[name = tensor("x_219_dilations_0"), val = tensor([1])]; - tensor x_219_groups_0 = const()[name = tensor("x_219_groups_0"), val = tensor(1)]; - tensor module_layers_9_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_9_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123020352)))]; - tensor module_layers_9_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_9_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123544704)))]; - tensor x_219_cast_fp16 = conv(bias = module_layers_9_conv_pointwise_conv2_bias_to_fp16, dilations = x_219_dilations_0, groups = x_219_groups_0, pad = x_219_pad_0, pad_type = x_219_pad_type_0, strides = x_219_strides_0, weight = module_layers_9_conv_pointwise_conv2_weight_to_fp16, x = input_521_cast_fp16)[name = tensor("x_219_cast_fp16")]; - tensor input_523_perm_0 = const()[name = tensor("input_523_perm_0"), val = tensor([0, 2, 1])]; - tensor input_523_cast_fp16 = transpose(perm = input_523_perm_0, x = x_219_cast_fp16)[name = tensor("transpose_135")]; - tensor input_525_cast_fp16 = add(x = input_507_cast_fp16, y = input_523_cast_fp16)[name = tensor("input_525_cast_fp16")]; - tensor input_527_axes_0 = const()[name = tensor("input_527_axes_0"), val = tensor([-1])]; - tensor module_layers_9_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_9_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123545792)))]; - tensor module_layers_9_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_9_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123546880)))]; - tensor input_527_cast_fp16 = layer_norm(axes = input_527_axes_0, beta = module_layers_9_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_9_norm_feed_forward2_weight_to_fp16, x = input_525_cast_fp16)[name = tensor("input_527_cast_fp16")]; - tensor module_layers_9_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_9_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123547968)))]; - tensor module_layers_9_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_9_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(125645184)))]; - tensor linear_89_cast_fp16 = linear(bias = module_layers_9_feed_forward2_linear1_bias_to_fp16, weight = module_layers_9_feed_forward2_linear1_weight_to_fp16, x = input_527_cast_fp16)[name = tensor("linear_89_cast_fp16")]; - tensor input_531_cast_fp16 = silu(x = linear_89_cast_fp16)[name = tensor("input_531_cast_fp16")]; - tensor module_layers_9_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_9_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(125649344)))]; - tensor module_layers_9_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_9_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127746560)))]; - tensor linear_90_cast_fp16 = linear(bias = module_layers_9_feed_forward2_linear2_bias_to_fp16, weight = module_layers_9_feed_forward2_linear2_weight_to_fp16, x = input_531_cast_fp16)[name = tensor("linear_90_cast_fp16")]; - tensor var_2061_to_fp16 = const()[name = tensor("op_2061_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2062_cast_fp16 = mul(x = linear_90_cast_fp16, y = var_2061_to_fp16)[name = tensor("op_2062_cast_fp16")]; - tensor input_537_cast_fp16 = add(x = input_525_cast_fp16, y = var_2062_cast_fp16)[name = tensor("input_537_cast_fp16")]; - tensor input_539_axes_0 = const()[name = tensor("input_539_axes_0"), val = tensor([-1])]; - tensor module_layers_9_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_9_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127747648)))]; - tensor module_layers_9_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_9_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127748736)))]; - tensor input_539_cast_fp16 = layer_norm(axes = input_539_axes_0, beta = module_layers_9_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_9_norm_out_weight_to_fp16, x = input_537_cast_fp16)[name = tensor("input_539_cast_fp16")]; - tensor input_541_axes_0 = const()[name = tensor("input_541_axes_0"), val = tensor([-1])]; - tensor module_layers_10_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_10_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127749824)))]; - tensor module_layers_10_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_10_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127750912)))]; - tensor input_541_cast_fp16 = layer_norm(axes = input_541_axes_0, beta = module_layers_10_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_10_norm_feed_forward1_weight_to_fp16, x = input_539_cast_fp16)[name = tensor("input_541_cast_fp16")]; - tensor module_layers_10_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_10_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127752000)))]; - tensor module_layers_10_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_10_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(129849216)))]; - tensor linear_91_cast_fp16 = linear(bias = module_layers_10_feed_forward1_linear1_bias_to_fp16, weight = module_layers_10_feed_forward1_linear1_weight_to_fp16, x = input_541_cast_fp16)[name = tensor("linear_91_cast_fp16")]; - tensor input_545_cast_fp16 = silu(x = linear_91_cast_fp16)[name = tensor("input_545_cast_fp16")]; - tensor module_layers_10_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_10_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(129853376)))]; - tensor module_layers_10_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_10_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(131950592)))]; - tensor linear_92_cast_fp16 = linear(bias = module_layers_10_feed_forward1_linear2_bias_to_fp16, weight = module_layers_10_feed_forward1_linear2_weight_to_fp16, x = input_545_cast_fp16)[name = tensor("linear_92_cast_fp16")]; - tensor var_2092_to_fp16 = const()[name = tensor("op_2092_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2093_cast_fp16 = mul(x = linear_92_cast_fp16, y = var_2092_to_fp16)[name = tensor("op_2093_cast_fp16")]; - tensor input_551_cast_fp16 = add(x = input_539_cast_fp16, y = var_2093_cast_fp16)[name = tensor("input_551_cast_fp16")]; - tensor query_21_axes_0 = const()[name = tensor("query_21_axes_0"), val = tensor([-1])]; - tensor module_layers_10_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_10_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(131951680)))]; - tensor module_layers_10_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_10_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(131952768)))]; - tensor query_21_cast_fp16 = layer_norm(axes = query_21_axes_0, beta = module_layers_10_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_10_norm_self_att_weight_to_fp16, x = input_551_cast_fp16)[name = tensor("query_21_cast_fp16")]; - tensor module_layers_10_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_10_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(131953856)))]; - tensor module_layers_10_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_10_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132478208)))]; - tensor linear_93_cast_fp16 = linear(bias = module_layers_10_self_attn_linear_q_bias_to_fp16, weight = module_layers_10_self_attn_linear_q_weight_to_fp16, x = query_21_cast_fp16)[name = tensor("linear_93_cast_fp16")]; - tensor var_2110 = const()[name = tensor("op_2110"), val = tensor([1, -1, 8, 64])]; - tensor q_61_cast_fp16 = reshape(shape = var_2110, x = linear_93_cast_fp16)[name = tensor("q_61_cast_fp16")]; - tensor module_layers_10_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_10_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132479296)))]; - tensor module_layers_10_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_10_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133003648)))]; - tensor linear_94_cast_fp16 = linear(bias = module_layers_10_self_attn_linear_k_bias_to_fp16, weight = module_layers_10_self_attn_linear_k_weight_to_fp16, x = query_21_cast_fp16)[name = tensor("linear_94_cast_fp16")]; - tensor var_2115 = const()[name = tensor("op_2115"), val = tensor([1, -1, 8, 64])]; - tensor k_41_cast_fp16 = reshape(shape = var_2115, x = linear_94_cast_fp16)[name = tensor("k_41_cast_fp16")]; - tensor module_layers_10_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_10_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133004736)))]; - tensor module_layers_10_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_10_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133529088)))]; - tensor linear_95_cast_fp16 = linear(bias = module_layers_10_self_attn_linear_v_bias_to_fp16, weight = module_layers_10_self_attn_linear_v_weight_to_fp16, x = query_21_cast_fp16)[name = tensor("linear_95_cast_fp16")]; - tensor var_2120 = const()[name = tensor("op_2120"), val = tensor([1, -1, 8, 64])]; - tensor v_21_cast_fp16 = reshape(shape = var_2120, x = linear_95_cast_fp16)[name = tensor("v_21_cast_fp16")]; - tensor value_23_perm_0 = const()[name = tensor("value_23_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_10_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_10_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133530176)))]; - tensor var_2132_cast_fp16 = add(x = q_61_cast_fp16, y = module_layers_10_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2132_cast_fp16")]; - tensor module_layers_10_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_10_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133531264)))]; - tensor var_2134_cast_fp16 = add(x = q_61_cast_fp16, y = module_layers_10_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2134_cast_fp16")]; - tensor q_with_bias_v_21_perm_0 = const()[name = tensor("q_with_bias_v_21_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_227_transpose_x_0 = const()[name = tensor("x_227_transpose_x_0"), val = tensor(false)]; - tensor x_227_transpose_y_0 = const()[name = tensor("x_227_transpose_y_0"), val = tensor(false)]; - tensor var_2136_to_fp16 = const()[name = tensor("op_2136_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133532352)))]; - tensor q_with_bias_v_21_cast_fp16 = transpose(perm = q_with_bias_v_21_perm_0, x = var_2134_cast_fp16)[name = tensor("transpose_133")]; - tensor x_227_cast_fp16 = matmul(transpose_x = x_227_transpose_x_0, transpose_y = x_227_transpose_y_0, x = q_with_bias_v_21_cast_fp16, y = var_2136_to_fp16)[name = tensor("x_227_cast_fp16")]; - tensor x_229_pad_0 = const()[name = tensor("x_229_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_229_mode_0 = const()[name = tensor("x_229_mode_0"), val = tensor("constant")]; - tensor const_170_to_fp16 = const()[name = tensor("const_170_to_fp16"), val = tensor(0x0p+0)]; - tensor x_229_cast_fp16 = pad(constant_val = const_170_to_fp16, mode = x_229_mode_0, pad = x_229_pad_0, x = x_227_cast_fp16)[name = tensor("x_229_cast_fp16")]; - tensor var_2144 = const()[name = tensor("op_2144"), val = tensor([1, 8, -1, 188])]; - tensor x_231_cast_fp16 = reshape(shape = var_2144, x = x_229_cast_fp16)[name = tensor("x_231_cast_fp16")]; - tensor var_2148_begin_0 = const()[name = tensor("op_2148_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_2148_end_0 = const()[name = tensor("op_2148_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_2148_end_mask_0 = const()[name = tensor("op_2148_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_2148_cast_fp16 = slice_by_index(begin = var_2148_begin_0, end = var_2148_end_0, end_mask = var_2148_end_mask_0, x = x_231_cast_fp16)[name = tensor("op_2148_cast_fp16")]; - tensor var_2149 = const()[name = tensor("op_2149"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_41_cast_fp16 = reshape(shape = var_2149, x = var_2148_cast_fp16)[name = tensor("matrix_bd_41_cast_fp16")]; - tensor matrix_ac_21_transpose_x_0 = const()[name = tensor("matrix_ac_21_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_21_transpose_y_0 = const()[name = tensor("matrix_ac_21_transpose_y_0"), val = tensor(false)]; - tensor transpose_71_perm_0 = const()[name = tensor("transpose_71_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_72_perm_0 = const()[name = tensor("transpose_72_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_72 = transpose(perm = transpose_72_perm_0, x = k_41_cast_fp16)[name = tensor("transpose_131")]; - tensor transpose_71 = transpose(perm = transpose_71_perm_0, x = var_2132_cast_fp16)[name = tensor("transpose_132")]; - tensor matrix_ac_21_cast_fp16 = matmul(transpose_x = matrix_ac_21_transpose_x_0, transpose_y = matrix_ac_21_transpose_y_0, x = transpose_71, y = transpose_72)[name = tensor("matrix_ac_21_cast_fp16")]; - tensor matrix_bd_43_begin_0 = const()[name = tensor("matrix_bd_43_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_43_end_0 = const()[name = tensor("matrix_bd_43_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_43_end_mask_0 = const()[name = tensor("matrix_bd_43_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_43_cast_fp16 = slice_by_index(begin = matrix_bd_43_begin_0, end = matrix_bd_43_end_0, end_mask = matrix_bd_43_end_mask_0, x = matrix_bd_41_cast_fp16)[name = tensor("matrix_bd_43_cast_fp16")]; - tensor var_2158_cast_fp16 = add(x = matrix_ac_21_cast_fp16, y = matrix_bd_43_cast_fp16)[name = tensor("op_2158_cast_fp16")]; - tensor _inversed_scores_41_y_0_to_fp16 = const()[name = tensor("_inversed_scores_41_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_41_cast_fp16 = mul(x = var_2158_cast_fp16, y = _inversed_scores_41_y_0_to_fp16)[name = tensor("_inversed_scores_41_cast_fp16")]; - tensor scores_43_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_41_cast_fp16, cond = mask_11)[name = tensor("scores_43_cast_fp16")]; - tensor var_2164_cast_fp16 = softmax(axis = var_23, x = scores_43_cast_fp16)[name = tensor("op_2164_cast_fp16")]; - tensor input_553_cast_fp16 = select(a = var_11_to_fp16, b = var_2164_cast_fp16, cond = mask_11)[name = tensor("input_553_cast_fp16")]; - tensor x_233_transpose_x_0 = const()[name = tensor("x_233_transpose_x_0"), val = tensor(false)]; - tensor x_233_transpose_y_0 = const()[name = tensor("x_233_transpose_y_0"), val = tensor(false)]; - tensor value_23_cast_fp16 = transpose(perm = value_23_perm_0, x = v_21_cast_fp16)[name = tensor("transpose_134")]; - tensor x_233_cast_fp16 = matmul(transpose_x = x_233_transpose_x_0, transpose_y = x_233_transpose_y_0, x = input_553_cast_fp16, y = value_23_cast_fp16)[name = tensor("x_233_cast_fp16")]; - tensor var_2168_perm_0 = const()[name = tensor("op_2168_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_2169 = const()[name = tensor("op_2169"), val = tensor([1, -1, 512])]; - tensor var_2168_cast_fp16 = transpose(perm = var_2168_perm_0, x = x_233_cast_fp16)[name = tensor("transpose_130")]; - tensor input_555_cast_fp16 = reshape(shape = var_2169, x = var_2168_cast_fp16)[name = tensor("input_555_cast_fp16")]; - tensor module_layers_10_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_10_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133916416)))]; - tensor module_layers_10_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_10_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134440768)))]; - tensor linear_97_cast_fp16 = linear(bias = module_layers_10_self_attn_linear_out_bias_to_fp16, weight = module_layers_10_self_attn_linear_out_weight_to_fp16, x = input_555_cast_fp16)[name = tensor("linear_97_cast_fp16")]; - tensor input_559_cast_fp16 = add(x = input_551_cast_fp16, y = linear_97_cast_fp16)[name = tensor("input_559_cast_fp16")]; - tensor x_237_axes_0 = const()[name = tensor("x_237_axes_0"), val = tensor([-1])]; - tensor module_layers_10_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_10_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134441856)))]; - tensor module_layers_10_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_10_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134442944)))]; - tensor x_237_cast_fp16 = layer_norm(axes = x_237_axes_0, beta = module_layers_10_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_10_norm_conv_weight_to_fp16, x = input_559_cast_fp16)[name = tensor("x_237_cast_fp16")]; - tensor input_561_perm_0 = const()[name = tensor("input_561_perm_0"), val = tensor([0, 2, 1])]; - tensor input_563_pad_type_0 = const()[name = tensor("input_563_pad_type_0"), val = tensor("valid")]; - tensor input_563_strides_0 = const()[name = tensor("input_563_strides_0"), val = tensor([1])]; - tensor input_563_pad_0 = const()[name = tensor("input_563_pad_0"), val = tensor([0, 0])]; - tensor input_563_dilations_0 = const()[name = tensor("input_563_dilations_0"), val = tensor([1])]; - tensor input_563_groups_0 = const()[name = tensor("input_563_groups_0"), val = tensor(1)]; - tensor module_layers_10_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_10_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134444032)))]; - tensor module_layers_10_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_10_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(135492672)))]; - tensor input_561_cast_fp16 = transpose(perm = input_561_perm_0, x = x_237_cast_fp16)[name = tensor("transpose_129")]; - tensor input_563_cast_fp16 = conv(bias = module_layers_10_conv_pointwise_conv1_bias_to_fp16, dilations = input_563_dilations_0, groups = input_563_groups_0, pad = input_563_pad_0, pad_type = input_563_pad_type_0, strides = input_563_strides_0, weight = module_layers_10_conv_pointwise_conv1_weight_to_fp16, x = input_561_cast_fp16)[name = tensor("input_563_cast_fp16")]; - tensor x_239_split_num_splits_0 = const()[name = tensor("x_239_split_num_splits_0"), val = tensor(2)]; - tensor x_239_split_axis_0 = const()[name = tensor("x_239_split_axis_0"), val = tensor(1)]; - tensor x_239_split_cast_fp16_0, tensor x_239_split_cast_fp16_1 = split(axis = x_239_split_axis_0, num_splits = x_239_split_num_splits_0, x = input_563_cast_fp16)[name = tensor("x_239_split_cast_fp16")]; - tensor x_239_split_1_sigmoid_cast_fp16 = sigmoid(x = x_239_split_cast_fp16_1)[name = tensor("x_239_split_1_sigmoid_cast_fp16")]; - tensor x_239_cast_fp16 = mul(x = x_239_split_cast_fp16_0, y = x_239_split_1_sigmoid_cast_fp16)[name = tensor("x_239_cast_fp16")]; - tensor input_565_cast_fp16 = select(a = var_11_to_fp16, b = x_239_cast_fp16, cond = var_453)[name = tensor("input_565_cast_fp16")]; - tensor input_567_pad_0 = const()[name = tensor("input_567_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_567_mode_0 = const()[name = tensor("input_567_mode_0"), val = tensor("constant")]; - tensor const_173_to_fp16 = const()[name = tensor("const_173_to_fp16"), val = tensor(0x0p+0)]; - tensor input_567_cast_fp16 = pad(constant_val = const_173_to_fp16, mode = input_567_mode_0, pad = input_567_pad_0, x = input_565_cast_fp16)[name = tensor("input_567_cast_fp16")]; - tensor input_569_pad_type_0 = const()[name = tensor("input_569_pad_type_0"), val = tensor("valid")]; - tensor input_569_groups_0 = const()[name = tensor("input_569_groups_0"), val = tensor(512)]; - tensor input_569_strides_0 = const()[name = tensor("input_569_strides_0"), val = tensor([1])]; - tensor input_569_pad_0 = const()[name = tensor("input_569_pad_0"), val = tensor([0, 0])]; - tensor input_569_dilations_0 = const()[name = tensor("input_569_dilations_0"), val = tensor([1])]; - tensor const_254_to_fp16 = const()[name = tensor("const_254_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(135494784)))]; - tensor const_255_to_fp16 = const()[name = tensor("const_255_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(135504064)))]; - tensor input_571_cast_fp16 = conv(bias = const_255_to_fp16, dilations = input_569_dilations_0, groups = input_569_groups_0, pad = input_569_pad_0, pad_type = input_569_pad_type_0, strides = input_569_strides_0, weight = const_254_to_fp16, x = input_567_cast_fp16)[name = tensor("input_571_cast_fp16")]; - tensor input_573_cast_fp16 = silu(x = input_571_cast_fp16)[name = tensor("input_573_cast_fp16")]; - tensor x_241_pad_type_0 = const()[name = tensor("x_241_pad_type_0"), val = tensor("valid")]; - tensor x_241_strides_0 = const()[name = tensor("x_241_strides_0"), val = tensor([1])]; - tensor x_241_pad_0 = const()[name = tensor("x_241_pad_0"), val = tensor([0, 0])]; - tensor x_241_dilations_0 = const()[name = tensor("x_241_dilations_0"), val = tensor([1])]; - tensor x_241_groups_0 = const()[name = tensor("x_241_groups_0"), val = tensor(1)]; - tensor module_layers_10_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_10_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(135505152)))]; - tensor module_layers_10_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_10_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136029504)))]; - tensor x_241_cast_fp16 = conv(bias = module_layers_10_conv_pointwise_conv2_bias_to_fp16, dilations = x_241_dilations_0, groups = x_241_groups_0, pad = x_241_pad_0, pad_type = x_241_pad_type_0, strides = x_241_strides_0, weight = module_layers_10_conv_pointwise_conv2_weight_to_fp16, x = input_573_cast_fp16)[name = tensor("x_241_cast_fp16")]; - tensor input_575_perm_0 = const()[name = tensor("input_575_perm_0"), val = tensor([0, 2, 1])]; - tensor input_575_cast_fp16 = transpose(perm = input_575_perm_0, x = x_241_cast_fp16)[name = tensor("transpose_128")]; - tensor input_577_cast_fp16 = add(x = input_559_cast_fp16, y = input_575_cast_fp16)[name = tensor("input_577_cast_fp16")]; - tensor input_579_axes_0 = const()[name = tensor("input_579_axes_0"), val = tensor([-1])]; - tensor module_layers_10_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_10_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136030592)))]; - tensor module_layers_10_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_10_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136031680)))]; - tensor input_579_cast_fp16 = layer_norm(axes = input_579_axes_0, beta = module_layers_10_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_10_norm_feed_forward2_weight_to_fp16, x = input_577_cast_fp16)[name = tensor("input_579_cast_fp16")]; - tensor module_layers_10_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_10_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136032768)))]; - tensor module_layers_10_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_10_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(138129984)))]; - tensor linear_98_cast_fp16 = linear(bias = module_layers_10_feed_forward2_linear1_bias_to_fp16, weight = module_layers_10_feed_forward2_linear1_weight_to_fp16, x = input_579_cast_fp16)[name = tensor("linear_98_cast_fp16")]; - tensor input_583_cast_fp16 = silu(x = linear_98_cast_fp16)[name = tensor("input_583_cast_fp16")]; - tensor module_layers_10_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_10_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(138134144)))]; - tensor module_layers_10_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_10_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140231360)))]; - tensor linear_99_cast_fp16 = linear(bias = module_layers_10_feed_forward2_linear2_bias_to_fp16, weight = module_layers_10_feed_forward2_linear2_weight_to_fp16, x = input_583_cast_fp16)[name = tensor("linear_99_cast_fp16")]; - tensor var_2235_to_fp16 = const()[name = tensor("op_2235_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2236_cast_fp16 = mul(x = linear_99_cast_fp16, y = var_2235_to_fp16)[name = tensor("op_2236_cast_fp16")]; - tensor input_589_cast_fp16 = add(x = input_577_cast_fp16, y = var_2236_cast_fp16)[name = tensor("input_589_cast_fp16")]; - tensor input_591_axes_0 = const()[name = tensor("input_591_axes_0"), val = tensor([-1])]; - tensor module_layers_10_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_10_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140232448)))]; - tensor module_layers_10_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_10_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140233536)))]; - tensor input_591_cast_fp16 = layer_norm(axes = input_591_axes_0, beta = module_layers_10_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_10_norm_out_weight_to_fp16, x = input_589_cast_fp16)[name = tensor("input_591_cast_fp16")]; - tensor input_593_axes_0 = const()[name = tensor("input_593_axes_0"), val = tensor([-1])]; - tensor module_layers_11_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_11_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140234624)))]; - tensor module_layers_11_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_11_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140235712)))]; - tensor input_593_cast_fp16 = layer_norm(axes = input_593_axes_0, beta = module_layers_11_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_11_norm_feed_forward1_weight_to_fp16, x = input_591_cast_fp16)[name = tensor("input_593_cast_fp16")]; - tensor module_layers_11_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_11_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140236800)))]; - tensor module_layers_11_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_11_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(142334016)))]; - tensor linear_100_cast_fp16 = linear(bias = module_layers_11_feed_forward1_linear1_bias_to_fp16, weight = module_layers_11_feed_forward1_linear1_weight_to_fp16, x = input_593_cast_fp16)[name = tensor("linear_100_cast_fp16")]; - tensor input_597_cast_fp16 = silu(x = linear_100_cast_fp16)[name = tensor("input_597_cast_fp16")]; - tensor module_layers_11_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_11_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(142338176)))]; - tensor module_layers_11_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_11_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(144435392)))]; - tensor linear_101_cast_fp16 = linear(bias = module_layers_11_feed_forward1_linear2_bias_to_fp16, weight = module_layers_11_feed_forward1_linear2_weight_to_fp16, x = input_597_cast_fp16)[name = tensor("linear_101_cast_fp16")]; - tensor var_2266_to_fp16 = const()[name = tensor("op_2266_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2267_cast_fp16 = mul(x = linear_101_cast_fp16, y = var_2266_to_fp16)[name = tensor("op_2267_cast_fp16")]; - tensor input_603_cast_fp16 = add(x = input_591_cast_fp16, y = var_2267_cast_fp16)[name = tensor("input_603_cast_fp16")]; - tensor query_23_axes_0 = const()[name = tensor("query_23_axes_0"), val = tensor([-1])]; - tensor module_layers_11_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_11_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(144436480)))]; - tensor module_layers_11_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_11_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(144437568)))]; - tensor query_23_cast_fp16 = layer_norm(axes = query_23_axes_0, beta = module_layers_11_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_11_norm_self_att_weight_to_fp16, x = input_603_cast_fp16)[name = tensor("query_23_cast_fp16")]; - tensor module_layers_11_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_11_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(144438656)))]; - tensor module_layers_11_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_11_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(144963008)))]; - tensor linear_102_cast_fp16 = linear(bias = module_layers_11_self_attn_linear_q_bias_to_fp16, weight = module_layers_11_self_attn_linear_q_weight_to_fp16, x = query_23_cast_fp16)[name = tensor("linear_102_cast_fp16")]; - tensor var_2284 = const()[name = tensor("op_2284"), val = tensor([1, -1, 8, 64])]; - tensor q_67_cast_fp16 = reshape(shape = var_2284, x = linear_102_cast_fp16)[name = tensor("q_67_cast_fp16")]; - tensor module_layers_11_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_11_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(144964096)))]; - tensor module_layers_11_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_11_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145488448)))]; - tensor linear_103_cast_fp16 = linear(bias = module_layers_11_self_attn_linear_k_bias_to_fp16, weight = module_layers_11_self_attn_linear_k_weight_to_fp16, x = query_23_cast_fp16)[name = tensor("linear_103_cast_fp16")]; - tensor var_2289 = const()[name = tensor("op_2289"), val = tensor([1, -1, 8, 64])]; - tensor k_45_cast_fp16 = reshape(shape = var_2289, x = linear_103_cast_fp16)[name = tensor("k_45_cast_fp16")]; - tensor module_layers_11_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_11_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145489536)))]; - tensor module_layers_11_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_11_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146013888)))]; - tensor linear_104_cast_fp16 = linear(bias = module_layers_11_self_attn_linear_v_bias_to_fp16, weight = module_layers_11_self_attn_linear_v_weight_to_fp16, x = query_23_cast_fp16)[name = tensor("linear_104_cast_fp16")]; - tensor var_2294 = const()[name = tensor("op_2294"), val = tensor([1, -1, 8, 64])]; - tensor v_23_cast_fp16 = reshape(shape = var_2294, x = linear_104_cast_fp16)[name = tensor("v_23_cast_fp16")]; - tensor value_25_perm_0 = const()[name = tensor("value_25_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_11_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_11_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146014976)))]; - tensor var_2306_cast_fp16 = add(x = q_67_cast_fp16, y = module_layers_11_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2306_cast_fp16")]; - tensor module_layers_11_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_11_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146016064)))]; - tensor var_2308_cast_fp16 = add(x = q_67_cast_fp16, y = module_layers_11_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2308_cast_fp16")]; - tensor q_with_bias_v_23_perm_0 = const()[name = tensor("q_with_bias_v_23_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_249_transpose_x_0 = const()[name = tensor("x_249_transpose_x_0"), val = tensor(false)]; - tensor x_249_transpose_y_0 = const()[name = tensor("x_249_transpose_y_0"), val = tensor(false)]; - tensor var_2310_to_fp16 = const()[name = tensor("op_2310_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146017152)))]; - tensor q_with_bias_v_23_cast_fp16 = transpose(perm = q_with_bias_v_23_perm_0, x = var_2308_cast_fp16)[name = tensor("transpose_126")]; - tensor x_249_cast_fp16 = matmul(transpose_x = x_249_transpose_x_0, transpose_y = x_249_transpose_y_0, x = q_with_bias_v_23_cast_fp16, y = var_2310_to_fp16)[name = tensor("x_249_cast_fp16")]; - tensor x_251_pad_0 = const()[name = tensor("x_251_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_251_mode_0 = const()[name = tensor("x_251_mode_0"), val = tensor("constant")]; - tensor const_180_to_fp16 = const()[name = tensor("const_180_to_fp16"), val = tensor(0x0p+0)]; - tensor x_251_cast_fp16 = pad(constant_val = const_180_to_fp16, mode = x_251_mode_0, pad = x_251_pad_0, x = x_249_cast_fp16)[name = tensor("x_251_cast_fp16")]; - tensor var_2318 = const()[name = tensor("op_2318"), val = tensor([1, 8, -1, 188])]; - tensor x_253_cast_fp16 = reshape(shape = var_2318, x = x_251_cast_fp16)[name = tensor("x_253_cast_fp16")]; - tensor var_2322_begin_0 = const()[name = tensor("op_2322_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_2322_end_0 = const()[name = tensor("op_2322_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_2322_end_mask_0 = const()[name = tensor("op_2322_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_2322_cast_fp16 = slice_by_index(begin = var_2322_begin_0, end = var_2322_end_0, end_mask = var_2322_end_mask_0, x = x_253_cast_fp16)[name = tensor("op_2322_cast_fp16")]; - tensor var_2323 = const()[name = tensor("op_2323"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_45_cast_fp16 = reshape(shape = var_2323, x = var_2322_cast_fp16)[name = tensor("matrix_bd_45_cast_fp16")]; - tensor matrix_ac_23_transpose_x_0 = const()[name = tensor("matrix_ac_23_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_23_transpose_y_0 = const()[name = tensor("matrix_ac_23_transpose_y_0"), val = tensor(false)]; - tensor transpose_73_perm_0 = const()[name = tensor("transpose_73_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_74_perm_0 = const()[name = tensor("transpose_74_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_74 = transpose(perm = transpose_74_perm_0, x = k_45_cast_fp16)[name = tensor("transpose_124")]; - tensor transpose_73 = transpose(perm = transpose_73_perm_0, x = var_2306_cast_fp16)[name = tensor("transpose_125")]; - tensor matrix_ac_23_cast_fp16 = matmul(transpose_x = matrix_ac_23_transpose_x_0, transpose_y = matrix_ac_23_transpose_y_0, x = transpose_73, y = transpose_74)[name = tensor("matrix_ac_23_cast_fp16")]; - tensor matrix_bd_47_begin_0 = const()[name = tensor("matrix_bd_47_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_47_end_0 = const()[name = tensor("matrix_bd_47_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_47_end_mask_0 = const()[name = tensor("matrix_bd_47_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_47_cast_fp16 = slice_by_index(begin = matrix_bd_47_begin_0, end = matrix_bd_47_end_0, end_mask = matrix_bd_47_end_mask_0, x = matrix_bd_45_cast_fp16)[name = tensor("matrix_bd_47_cast_fp16")]; - tensor var_2332_cast_fp16 = add(x = matrix_ac_23_cast_fp16, y = matrix_bd_47_cast_fp16)[name = tensor("op_2332_cast_fp16")]; - tensor _inversed_scores_45_y_0_to_fp16 = const()[name = tensor("_inversed_scores_45_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_45_cast_fp16 = mul(x = var_2332_cast_fp16, y = _inversed_scores_45_y_0_to_fp16)[name = tensor("_inversed_scores_45_cast_fp16")]; - tensor scores_47_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_45_cast_fp16, cond = mask_11)[name = tensor("scores_47_cast_fp16")]; - tensor var_2338_cast_fp16 = softmax(axis = var_23, x = scores_47_cast_fp16)[name = tensor("op_2338_cast_fp16")]; - tensor input_605_cast_fp16 = select(a = var_11_to_fp16, b = var_2338_cast_fp16, cond = mask_11)[name = tensor("input_605_cast_fp16")]; - tensor x_255_transpose_x_0 = const()[name = tensor("x_255_transpose_x_0"), val = tensor(false)]; - tensor x_255_transpose_y_0 = const()[name = tensor("x_255_transpose_y_0"), val = tensor(false)]; - tensor value_25_cast_fp16 = transpose(perm = value_25_perm_0, x = v_23_cast_fp16)[name = tensor("transpose_127")]; - tensor x_255_cast_fp16 = matmul(transpose_x = x_255_transpose_x_0, transpose_y = x_255_transpose_y_0, x = input_605_cast_fp16, y = value_25_cast_fp16)[name = tensor("x_255_cast_fp16")]; - tensor var_2342_perm_0 = const()[name = tensor("op_2342_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_2343 = const()[name = tensor("op_2343"), val = tensor([1, -1, 512])]; - tensor var_2342_cast_fp16 = transpose(perm = var_2342_perm_0, x = x_255_cast_fp16)[name = tensor("transpose_123")]; - tensor input_607_cast_fp16 = reshape(shape = var_2343, x = var_2342_cast_fp16)[name = tensor("input_607_cast_fp16")]; - tensor module_layers_11_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_11_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146401216)))]; - tensor module_layers_11_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_11_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146925568)))]; - tensor linear_106_cast_fp16 = linear(bias = module_layers_11_self_attn_linear_out_bias_to_fp16, weight = module_layers_11_self_attn_linear_out_weight_to_fp16, x = input_607_cast_fp16)[name = tensor("linear_106_cast_fp16")]; - tensor input_611_cast_fp16 = add(x = input_603_cast_fp16, y = linear_106_cast_fp16)[name = tensor("input_611_cast_fp16")]; - tensor x_259_axes_0 = const()[name = tensor("x_259_axes_0"), val = tensor([-1])]; - tensor module_layers_11_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_11_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146926656)))]; - tensor module_layers_11_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_11_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146927744)))]; - tensor x_259_cast_fp16 = layer_norm(axes = x_259_axes_0, beta = module_layers_11_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_11_norm_conv_weight_to_fp16, x = input_611_cast_fp16)[name = tensor("x_259_cast_fp16")]; - tensor input_613_perm_0 = const()[name = tensor("input_613_perm_0"), val = tensor([0, 2, 1])]; - tensor input_615_pad_type_0 = const()[name = tensor("input_615_pad_type_0"), val = tensor("valid")]; - tensor input_615_strides_0 = const()[name = tensor("input_615_strides_0"), val = tensor([1])]; - tensor input_615_pad_0 = const()[name = tensor("input_615_pad_0"), val = tensor([0, 0])]; - tensor input_615_dilations_0 = const()[name = tensor("input_615_dilations_0"), val = tensor([1])]; - tensor input_615_groups_0 = const()[name = tensor("input_615_groups_0"), val = tensor(1)]; - tensor module_layers_11_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_11_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146928832)))]; - tensor module_layers_11_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_11_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147977472)))]; - tensor input_613_cast_fp16 = transpose(perm = input_613_perm_0, x = x_259_cast_fp16)[name = tensor("transpose_122")]; - tensor input_615_cast_fp16 = conv(bias = module_layers_11_conv_pointwise_conv1_bias_to_fp16, dilations = input_615_dilations_0, groups = input_615_groups_0, pad = input_615_pad_0, pad_type = input_615_pad_type_0, strides = input_615_strides_0, weight = module_layers_11_conv_pointwise_conv1_weight_to_fp16, x = input_613_cast_fp16)[name = tensor("input_615_cast_fp16")]; - tensor x_261_split_num_splits_0 = const()[name = tensor("x_261_split_num_splits_0"), val = tensor(2)]; - tensor x_261_split_axis_0 = const()[name = tensor("x_261_split_axis_0"), val = tensor(1)]; - tensor x_261_split_cast_fp16_0, tensor x_261_split_cast_fp16_1 = split(axis = x_261_split_axis_0, num_splits = x_261_split_num_splits_0, x = input_615_cast_fp16)[name = tensor("x_261_split_cast_fp16")]; - tensor x_261_split_1_sigmoid_cast_fp16 = sigmoid(x = x_261_split_cast_fp16_1)[name = tensor("x_261_split_1_sigmoid_cast_fp16")]; - tensor x_261_cast_fp16 = mul(x = x_261_split_cast_fp16_0, y = x_261_split_1_sigmoid_cast_fp16)[name = tensor("x_261_cast_fp16")]; - tensor input_617_cast_fp16 = select(a = var_11_to_fp16, b = x_261_cast_fp16, cond = var_453)[name = tensor("input_617_cast_fp16")]; - tensor input_619_pad_0 = const()[name = tensor("input_619_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_619_mode_0 = const()[name = tensor("input_619_mode_0"), val = tensor("constant")]; - tensor const_183_to_fp16 = const()[name = tensor("const_183_to_fp16"), val = tensor(0x0p+0)]; - tensor input_619_cast_fp16 = pad(constant_val = const_183_to_fp16, mode = input_619_mode_0, pad = input_619_pad_0, x = input_617_cast_fp16)[name = tensor("input_619_cast_fp16")]; - tensor input_621_pad_type_0 = const()[name = tensor("input_621_pad_type_0"), val = tensor("valid")]; - tensor input_621_groups_0 = const()[name = tensor("input_621_groups_0"), val = tensor(512)]; - tensor input_621_strides_0 = const()[name = tensor("input_621_strides_0"), val = tensor([1])]; - tensor input_621_pad_0 = const()[name = tensor("input_621_pad_0"), val = tensor([0, 0])]; - tensor input_621_dilations_0 = const()[name = tensor("input_621_dilations_0"), val = tensor([1])]; - tensor const_256_to_fp16 = const()[name = tensor("const_256_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147979584)))]; - tensor const_257_to_fp16 = const()[name = tensor("const_257_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147988864)))]; - tensor input_623_cast_fp16 = conv(bias = const_257_to_fp16, dilations = input_621_dilations_0, groups = input_621_groups_0, pad = input_621_pad_0, pad_type = input_621_pad_type_0, strides = input_621_strides_0, weight = const_256_to_fp16, x = input_619_cast_fp16)[name = tensor("input_623_cast_fp16")]; - tensor input_625_cast_fp16 = silu(x = input_623_cast_fp16)[name = tensor("input_625_cast_fp16")]; - tensor x_263_pad_type_0 = const()[name = tensor("x_263_pad_type_0"), val = tensor("valid")]; - tensor x_263_strides_0 = const()[name = tensor("x_263_strides_0"), val = tensor([1])]; - tensor x_263_pad_0 = const()[name = tensor("x_263_pad_0"), val = tensor([0, 0])]; - tensor x_263_dilations_0 = const()[name = tensor("x_263_dilations_0"), val = tensor([1])]; - tensor x_263_groups_0 = const()[name = tensor("x_263_groups_0"), val = tensor(1)]; - tensor module_layers_11_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_11_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147989952)))]; - tensor module_layers_11_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_11_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(148514304)))]; - tensor x_263_cast_fp16 = conv(bias = module_layers_11_conv_pointwise_conv2_bias_to_fp16, dilations = x_263_dilations_0, groups = x_263_groups_0, pad = x_263_pad_0, pad_type = x_263_pad_type_0, strides = x_263_strides_0, weight = module_layers_11_conv_pointwise_conv2_weight_to_fp16, x = input_625_cast_fp16)[name = tensor("x_263_cast_fp16")]; - tensor input_627_perm_0 = const()[name = tensor("input_627_perm_0"), val = tensor([0, 2, 1])]; - tensor input_627_cast_fp16 = transpose(perm = input_627_perm_0, x = x_263_cast_fp16)[name = tensor("transpose_121")]; - tensor input_629_cast_fp16 = add(x = input_611_cast_fp16, y = input_627_cast_fp16)[name = tensor("input_629_cast_fp16")]; - tensor input_631_axes_0 = const()[name = tensor("input_631_axes_0"), val = tensor([-1])]; - tensor module_layers_11_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_11_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(148515392)))]; - tensor module_layers_11_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_11_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(148516480)))]; - tensor input_631_cast_fp16 = layer_norm(axes = input_631_axes_0, beta = module_layers_11_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_11_norm_feed_forward2_weight_to_fp16, x = input_629_cast_fp16)[name = tensor("input_631_cast_fp16")]; - tensor module_layers_11_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_11_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(148517568)))]; - tensor module_layers_11_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_11_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(150614784)))]; - tensor linear_107_cast_fp16 = linear(bias = module_layers_11_feed_forward2_linear1_bias_to_fp16, weight = module_layers_11_feed_forward2_linear1_weight_to_fp16, x = input_631_cast_fp16)[name = tensor("linear_107_cast_fp16")]; - tensor input_635_cast_fp16 = silu(x = linear_107_cast_fp16)[name = tensor("input_635_cast_fp16")]; - tensor module_layers_11_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_11_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(150618944)))]; - tensor module_layers_11_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_11_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(152716160)))]; - tensor linear_108_cast_fp16 = linear(bias = module_layers_11_feed_forward2_linear2_bias_to_fp16, weight = module_layers_11_feed_forward2_linear2_weight_to_fp16, x = input_635_cast_fp16)[name = tensor("linear_108_cast_fp16")]; - tensor var_2409_to_fp16 = const()[name = tensor("op_2409_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2410_cast_fp16 = mul(x = linear_108_cast_fp16, y = var_2409_to_fp16)[name = tensor("op_2410_cast_fp16")]; - tensor input_641_cast_fp16 = add(x = input_629_cast_fp16, y = var_2410_cast_fp16)[name = tensor("input_641_cast_fp16")]; - tensor input_643_axes_0 = const()[name = tensor("input_643_axes_0"), val = tensor([-1])]; - tensor module_layers_11_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_11_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(152717248)))]; - tensor module_layers_11_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_11_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(152718336)))]; - tensor input_643_cast_fp16 = layer_norm(axes = input_643_axes_0, beta = module_layers_11_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_11_norm_out_weight_to_fp16, x = input_641_cast_fp16)[name = tensor("input_643_cast_fp16")]; - tensor input_645_axes_0 = const()[name = tensor("input_645_axes_0"), val = tensor([-1])]; - tensor module_layers_12_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_12_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(152719424)))]; - tensor module_layers_12_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_12_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(152720512)))]; - tensor input_645_cast_fp16 = layer_norm(axes = input_645_axes_0, beta = module_layers_12_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_12_norm_feed_forward1_weight_to_fp16, x = input_643_cast_fp16)[name = tensor("input_645_cast_fp16")]; - tensor module_layers_12_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_12_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(152721600)))]; - tensor module_layers_12_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_12_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(154818816)))]; - tensor linear_109_cast_fp16 = linear(bias = module_layers_12_feed_forward1_linear1_bias_to_fp16, weight = module_layers_12_feed_forward1_linear1_weight_to_fp16, x = input_645_cast_fp16)[name = tensor("linear_109_cast_fp16")]; - tensor input_649_cast_fp16 = silu(x = linear_109_cast_fp16)[name = tensor("input_649_cast_fp16")]; - tensor module_layers_12_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_12_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(154822976)))]; - tensor module_layers_12_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_12_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(156920192)))]; - tensor linear_110_cast_fp16 = linear(bias = module_layers_12_feed_forward1_linear2_bias_to_fp16, weight = module_layers_12_feed_forward1_linear2_weight_to_fp16, x = input_649_cast_fp16)[name = tensor("linear_110_cast_fp16")]; - tensor var_2440_to_fp16 = const()[name = tensor("op_2440_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2441_cast_fp16 = mul(x = linear_110_cast_fp16, y = var_2440_to_fp16)[name = tensor("op_2441_cast_fp16")]; - tensor input_655_cast_fp16 = add(x = input_643_cast_fp16, y = var_2441_cast_fp16)[name = tensor("input_655_cast_fp16")]; - tensor query_25_axes_0 = const()[name = tensor("query_25_axes_0"), val = tensor([-1])]; - tensor module_layers_12_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_12_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(156921280)))]; - tensor module_layers_12_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_12_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(156922368)))]; - tensor query_25_cast_fp16 = layer_norm(axes = query_25_axes_0, beta = module_layers_12_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_12_norm_self_att_weight_to_fp16, x = input_655_cast_fp16)[name = tensor("query_25_cast_fp16")]; - tensor module_layers_12_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_12_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(156923456)))]; - tensor module_layers_12_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_12_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157447808)))]; - tensor linear_111_cast_fp16 = linear(bias = module_layers_12_self_attn_linear_q_bias_to_fp16, weight = module_layers_12_self_attn_linear_q_weight_to_fp16, x = query_25_cast_fp16)[name = tensor("linear_111_cast_fp16")]; - tensor var_2458 = const()[name = tensor("op_2458"), val = tensor([1, -1, 8, 64])]; - tensor q_73_cast_fp16 = reshape(shape = var_2458, x = linear_111_cast_fp16)[name = tensor("q_73_cast_fp16")]; - tensor module_layers_12_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_12_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157448896)))]; - tensor module_layers_12_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_12_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157973248)))]; - tensor linear_112_cast_fp16 = linear(bias = module_layers_12_self_attn_linear_k_bias_to_fp16, weight = module_layers_12_self_attn_linear_k_weight_to_fp16, x = query_25_cast_fp16)[name = tensor("linear_112_cast_fp16")]; - tensor var_2463 = const()[name = tensor("op_2463"), val = tensor([1, -1, 8, 64])]; - tensor k_49_cast_fp16 = reshape(shape = var_2463, x = linear_112_cast_fp16)[name = tensor("k_49_cast_fp16")]; - tensor module_layers_12_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_12_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157974336)))]; - tensor module_layers_12_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_12_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158498688)))]; - tensor linear_113_cast_fp16 = linear(bias = module_layers_12_self_attn_linear_v_bias_to_fp16, weight = module_layers_12_self_attn_linear_v_weight_to_fp16, x = query_25_cast_fp16)[name = tensor("linear_113_cast_fp16")]; - tensor var_2468 = const()[name = tensor("op_2468"), val = tensor([1, -1, 8, 64])]; - tensor v_25_cast_fp16 = reshape(shape = var_2468, x = linear_113_cast_fp16)[name = tensor("v_25_cast_fp16")]; - tensor value_27_perm_0 = const()[name = tensor("value_27_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_12_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_12_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158499776)))]; - tensor var_2480_cast_fp16 = add(x = q_73_cast_fp16, y = module_layers_12_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2480_cast_fp16")]; - tensor module_layers_12_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_12_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158500864)))]; - tensor var_2482_cast_fp16 = add(x = q_73_cast_fp16, y = module_layers_12_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2482_cast_fp16")]; - tensor q_with_bias_v_25_perm_0 = const()[name = tensor("q_with_bias_v_25_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_271_transpose_x_0 = const()[name = tensor("x_271_transpose_x_0"), val = tensor(false)]; - tensor x_271_transpose_y_0 = const()[name = tensor("x_271_transpose_y_0"), val = tensor(false)]; - tensor var_2484_to_fp16 = const()[name = tensor("op_2484_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158501952)))]; - tensor q_with_bias_v_25_cast_fp16 = transpose(perm = q_with_bias_v_25_perm_0, x = var_2482_cast_fp16)[name = tensor("transpose_119")]; - tensor x_271_cast_fp16 = matmul(transpose_x = x_271_transpose_x_0, transpose_y = x_271_transpose_y_0, x = q_with_bias_v_25_cast_fp16, y = var_2484_to_fp16)[name = tensor("x_271_cast_fp16")]; - tensor x_273_pad_0 = const()[name = tensor("x_273_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_273_mode_0 = const()[name = tensor("x_273_mode_0"), val = tensor("constant")]; - tensor const_190_to_fp16 = const()[name = tensor("const_190_to_fp16"), val = tensor(0x0p+0)]; - tensor x_273_cast_fp16 = pad(constant_val = const_190_to_fp16, mode = x_273_mode_0, pad = x_273_pad_0, x = x_271_cast_fp16)[name = tensor("x_273_cast_fp16")]; - tensor var_2492 = const()[name = tensor("op_2492"), val = tensor([1, 8, -1, 188])]; - tensor x_275_cast_fp16 = reshape(shape = var_2492, x = x_273_cast_fp16)[name = tensor("x_275_cast_fp16")]; - tensor var_2496_begin_0 = const()[name = tensor("op_2496_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_2496_end_0 = const()[name = tensor("op_2496_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_2496_end_mask_0 = const()[name = tensor("op_2496_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_2496_cast_fp16 = slice_by_index(begin = var_2496_begin_0, end = var_2496_end_0, end_mask = var_2496_end_mask_0, x = x_275_cast_fp16)[name = tensor("op_2496_cast_fp16")]; - tensor var_2497 = const()[name = tensor("op_2497"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_49_cast_fp16 = reshape(shape = var_2497, x = var_2496_cast_fp16)[name = tensor("matrix_bd_49_cast_fp16")]; - tensor matrix_ac_25_transpose_x_0 = const()[name = tensor("matrix_ac_25_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_25_transpose_y_0 = const()[name = tensor("matrix_ac_25_transpose_y_0"), val = tensor(false)]; - tensor transpose_75_perm_0 = const()[name = tensor("transpose_75_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_76_perm_0 = const()[name = tensor("transpose_76_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_76 = transpose(perm = transpose_76_perm_0, x = k_49_cast_fp16)[name = tensor("transpose_117")]; - tensor transpose_75 = transpose(perm = transpose_75_perm_0, x = var_2480_cast_fp16)[name = tensor("transpose_118")]; - tensor matrix_ac_25_cast_fp16 = matmul(transpose_x = matrix_ac_25_transpose_x_0, transpose_y = matrix_ac_25_transpose_y_0, x = transpose_75, y = transpose_76)[name = tensor("matrix_ac_25_cast_fp16")]; - tensor matrix_bd_51_begin_0 = const()[name = tensor("matrix_bd_51_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_51_end_0 = const()[name = tensor("matrix_bd_51_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_51_end_mask_0 = const()[name = tensor("matrix_bd_51_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_51_cast_fp16 = slice_by_index(begin = matrix_bd_51_begin_0, end = matrix_bd_51_end_0, end_mask = matrix_bd_51_end_mask_0, x = matrix_bd_49_cast_fp16)[name = tensor("matrix_bd_51_cast_fp16")]; - tensor var_2506_cast_fp16 = add(x = matrix_ac_25_cast_fp16, y = matrix_bd_51_cast_fp16)[name = tensor("op_2506_cast_fp16")]; - tensor _inversed_scores_49_y_0_to_fp16 = const()[name = tensor("_inversed_scores_49_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_49_cast_fp16 = mul(x = var_2506_cast_fp16, y = _inversed_scores_49_y_0_to_fp16)[name = tensor("_inversed_scores_49_cast_fp16")]; - tensor scores_51_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_49_cast_fp16, cond = mask_11)[name = tensor("scores_51_cast_fp16")]; - tensor var_2512_cast_fp16 = softmax(axis = var_23, x = scores_51_cast_fp16)[name = tensor("op_2512_cast_fp16")]; - tensor input_657_cast_fp16 = select(a = var_11_to_fp16, b = var_2512_cast_fp16, cond = mask_11)[name = tensor("input_657_cast_fp16")]; - tensor x_277_transpose_x_0 = const()[name = tensor("x_277_transpose_x_0"), val = tensor(false)]; - tensor x_277_transpose_y_0 = const()[name = tensor("x_277_transpose_y_0"), val = tensor(false)]; - tensor value_27_cast_fp16 = transpose(perm = value_27_perm_0, x = v_25_cast_fp16)[name = tensor("transpose_120")]; - tensor x_277_cast_fp16 = matmul(transpose_x = x_277_transpose_x_0, transpose_y = x_277_transpose_y_0, x = input_657_cast_fp16, y = value_27_cast_fp16)[name = tensor("x_277_cast_fp16")]; - tensor var_2516_perm_0 = const()[name = tensor("op_2516_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_2517 = const()[name = tensor("op_2517"), val = tensor([1, -1, 512])]; - tensor var_2516_cast_fp16 = transpose(perm = var_2516_perm_0, x = x_277_cast_fp16)[name = tensor("transpose_116")]; - tensor input_659_cast_fp16 = reshape(shape = var_2517, x = var_2516_cast_fp16)[name = tensor("input_659_cast_fp16")]; - tensor module_layers_12_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_12_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158886016)))]; - tensor module_layers_12_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_12_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159410368)))]; - tensor linear_115_cast_fp16 = linear(bias = module_layers_12_self_attn_linear_out_bias_to_fp16, weight = module_layers_12_self_attn_linear_out_weight_to_fp16, x = input_659_cast_fp16)[name = tensor("linear_115_cast_fp16")]; - tensor input_663_cast_fp16 = add(x = input_655_cast_fp16, y = linear_115_cast_fp16)[name = tensor("input_663_cast_fp16")]; - tensor x_281_axes_0 = const()[name = tensor("x_281_axes_0"), val = tensor([-1])]; - tensor module_layers_12_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_12_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159411456)))]; - tensor module_layers_12_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_12_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159412544)))]; - tensor x_281_cast_fp16 = layer_norm(axes = x_281_axes_0, beta = module_layers_12_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_12_norm_conv_weight_to_fp16, x = input_663_cast_fp16)[name = tensor("x_281_cast_fp16")]; - tensor input_665_perm_0 = const()[name = tensor("input_665_perm_0"), val = tensor([0, 2, 1])]; - tensor input_667_pad_type_0 = const()[name = tensor("input_667_pad_type_0"), val = tensor("valid")]; - tensor input_667_strides_0 = const()[name = tensor("input_667_strides_0"), val = tensor([1])]; - tensor input_667_pad_0 = const()[name = tensor("input_667_pad_0"), val = tensor([0, 0])]; - tensor input_667_dilations_0 = const()[name = tensor("input_667_dilations_0"), val = tensor([1])]; - tensor input_667_groups_0 = const()[name = tensor("input_667_groups_0"), val = tensor(1)]; - tensor module_layers_12_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_12_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159413632)))]; - tensor module_layers_12_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_12_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160462272)))]; - tensor input_665_cast_fp16 = transpose(perm = input_665_perm_0, x = x_281_cast_fp16)[name = tensor("transpose_115")]; - tensor input_667_cast_fp16 = conv(bias = module_layers_12_conv_pointwise_conv1_bias_to_fp16, dilations = input_667_dilations_0, groups = input_667_groups_0, pad = input_667_pad_0, pad_type = input_667_pad_type_0, strides = input_667_strides_0, weight = module_layers_12_conv_pointwise_conv1_weight_to_fp16, x = input_665_cast_fp16)[name = tensor("input_667_cast_fp16")]; - tensor x_283_split_num_splits_0 = const()[name = tensor("x_283_split_num_splits_0"), val = tensor(2)]; - tensor x_283_split_axis_0 = const()[name = tensor("x_283_split_axis_0"), val = tensor(1)]; - tensor x_283_split_cast_fp16_0, tensor x_283_split_cast_fp16_1 = split(axis = x_283_split_axis_0, num_splits = x_283_split_num_splits_0, x = input_667_cast_fp16)[name = tensor("x_283_split_cast_fp16")]; - tensor x_283_split_1_sigmoid_cast_fp16 = sigmoid(x = x_283_split_cast_fp16_1)[name = tensor("x_283_split_1_sigmoid_cast_fp16")]; - tensor x_283_cast_fp16 = mul(x = x_283_split_cast_fp16_0, y = x_283_split_1_sigmoid_cast_fp16)[name = tensor("x_283_cast_fp16")]; - tensor input_669_cast_fp16 = select(a = var_11_to_fp16, b = x_283_cast_fp16, cond = var_453)[name = tensor("input_669_cast_fp16")]; - tensor input_671_pad_0 = const()[name = tensor("input_671_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_671_mode_0 = const()[name = tensor("input_671_mode_0"), val = tensor("constant")]; - tensor const_193_to_fp16 = const()[name = tensor("const_193_to_fp16"), val = tensor(0x0p+0)]; - tensor input_671_cast_fp16 = pad(constant_val = const_193_to_fp16, mode = input_671_mode_0, pad = input_671_pad_0, x = input_669_cast_fp16)[name = tensor("input_671_cast_fp16")]; - tensor input_673_pad_type_0 = const()[name = tensor("input_673_pad_type_0"), val = tensor("valid")]; - tensor input_673_groups_0 = const()[name = tensor("input_673_groups_0"), val = tensor(512)]; - tensor input_673_strides_0 = const()[name = tensor("input_673_strides_0"), val = tensor([1])]; - tensor input_673_pad_0 = const()[name = tensor("input_673_pad_0"), val = tensor([0, 0])]; - tensor input_673_dilations_0 = const()[name = tensor("input_673_dilations_0"), val = tensor([1])]; - tensor const_258_to_fp16 = const()[name = tensor("const_258_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160464384)))]; - tensor const_259_to_fp16 = const()[name = tensor("const_259_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160473664)))]; - tensor input_675_cast_fp16 = conv(bias = const_259_to_fp16, dilations = input_673_dilations_0, groups = input_673_groups_0, pad = input_673_pad_0, pad_type = input_673_pad_type_0, strides = input_673_strides_0, weight = const_258_to_fp16, x = input_671_cast_fp16)[name = tensor("input_675_cast_fp16")]; - tensor input_677_cast_fp16 = silu(x = input_675_cast_fp16)[name = tensor("input_677_cast_fp16")]; - tensor x_285_pad_type_0 = const()[name = tensor("x_285_pad_type_0"), val = tensor("valid")]; - tensor x_285_strides_0 = const()[name = tensor("x_285_strides_0"), val = tensor([1])]; - tensor x_285_pad_0 = const()[name = tensor("x_285_pad_0"), val = tensor([0, 0])]; - tensor x_285_dilations_0 = const()[name = tensor("x_285_dilations_0"), val = tensor([1])]; - tensor x_285_groups_0 = const()[name = tensor("x_285_groups_0"), val = tensor(1)]; - tensor module_layers_12_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_12_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160474752)))]; - tensor module_layers_12_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_12_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160999104)))]; - tensor x_285_cast_fp16 = conv(bias = module_layers_12_conv_pointwise_conv2_bias_to_fp16, dilations = x_285_dilations_0, groups = x_285_groups_0, pad = x_285_pad_0, pad_type = x_285_pad_type_0, strides = x_285_strides_0, weight = module_layers_12_conv_pointwise_conv2_weight_to_fp16, x = input_677_cast_fp16)[name = tensor("x_285_cast_fp16")]; - tensor input_679_perm_0 = const()[name = tensor("input_679_perm_0"), val = tensor([0, 2, 1])]; - tensor input_679_cast_fp16 = transpose(perm = input_679_perm_0, x = x_285_cast_fp16)[name = tensor("transpose_114")]; - tensor input_681_cast_fp16 = add(x = input_663_cast_fp16, y = input_679_cast_fp16)[name = tensor("input_681_cast_fp16")]; - tensor input_683_axes_0 = const()[name = tensor("input_683_axes_0"), val = tensor([-1])]; - tensor module_layers_12_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_12_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161000192)))]; - tensor module_layers_12_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_12_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161001280)))]; - tensor input_683_cast_fp16 = layer_norm(axes = input_683_axes_0, beta = module_layers_12_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_12_norm_feed_forward2_weight_to_fp16, x = input_681_cast_fp16)[name = tensor("input_683_cast_fp16")]; - tensor module_layers_12_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_12_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161002368)))]; - tensor module_layers_12_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_12_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(163099584)))]; - tensor linear_116_cast_fp16 = linear(bias = module_layers_12_feed_forward2_linear1_bias_to_fp16, weight = module_layers_12_feed_forward2_linear1_weight_to_fp16, x = input_683_cast_fp16)[name = tensor("linear_116_cast_fp16")]; - tensor input_687_cast_fp16 = silu(x = linear_116_cast_fp16)[name = tensor("input_687_cast_fp16")]; - tensor module_layers_12_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_12_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(163103744)))]; - tensor module_layers_12_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_12_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165200960)))]; - tensor linear_117_cast_fp16 = linear(bias = module_layers_12_feed_forward2_linear2_bias_to_fp16, weight = module_layers_12_feed_forward2_linear2_weight_to_fp16, x = input_687_cast_fp16)[name = tensor("linear_117_cast_fp16")]; - tensor var_2583_to_fp16 = const()[name = tensor("op_2583_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2584_cast_fp16 = mul(x = linear_117_cast_fp16, y = var_2583_to_fp16)[name = tensor("op_2584_cast_fp16")]; - tensor input_693_cast_fp16 = add(x = input_681_cast_fp16, y = var_2584_cast_fp16)[name = tensor("input_693_cast_fp16")]; - tensor input_695_axes_0 = const()[name = tensor("input_695_axes_0"), val = tensor([-1])]; - tensor module_layers_12_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_12_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165202048)))]; - tensor module_layers_12_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_12_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165203136)))]; - tensor input_695_cast_fp16 = layer_norm(axes = input_695_axes_0, beta = module_layers_12_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_12_norm_out_weight_to_fp16, x = input_693_cast_fp16)[name = tensor("input_695_cast_fp16")]; - tensor input_697_axes_0 = const()[name = tensor("input_697_axes_0"), val = tensor([-1])]; - tensor module_layers_13_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_13_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165204224)))]; - tensor module_layers_13_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_13_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165205312)))]; - tensor input_697_cast_fp16 = layer_norm(axes = input_697_axes_0, beta = module_layers_13_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_13_norm_feed_forward1_weight_to_fp16, x = input_695_cast_fp16)[name = tensor("input_697_cast_fp16")]; - tensor module_layers_13_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_13_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165206400)))]; - tensor module_layers_13_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_13_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(167303616)))]; - tensor linear_118_cast_fp16 = linear(bias = module_layers_13_feed_forward1_linear1_bias_to_fp16, weight = module_layers_13_feed_forward1_linear1_weight_to_fp16, x = input_697_cast_fp16)[name = tensor("linear_118_cast_fp16")]; - tensor input_701_cast_fp16 = silu(x = linear_118_cast_fp16)[name = tensor("input_701_cast_fp16")]; - tensor module_layers_13_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_13_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(167307776)))]; - tensor module_layers_13_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_13_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169404992)))]; - tensor linear_119_cast_fp16 = linear(bias = module_layers_13_feed_forward1_linear2_bias_to_fp16, weight = module_layers_13_feed_forward1_linear2_weight_to_fp16, x = input_701_cast_fp16)[name = tensor("linear_119_cast_fp16")]; - tensor var_2614_to_fp16 = const()[name = tensor("op_2614_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2615_cast_fp16 = mul(x = linear_119_cast_fp16, y = var_2614_to_fp16)[name = tensor("op_2615_cast_fp16")]; - tensor input_707_cast_fp16 = add(x = input_695_cast_fp16, y = var_2615_cast_fp16)[name = tensor("input_707_cast_fp16")]; - tensor query_27_axes_0 = const()[name = tensor("query_27_axes_0"), val = tensor([-1])]; - tensor module_layers_13_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_13_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169406080)))]; - tensor module_layers_13_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_13_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169407168)))]; - tensor query_27_cast_fp16 = layer_norm(axes = query_27_axes_0, beta = module_layers_13_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_13_norm_self_att_weight_to_fp16, x = input_707_cast_fp16)[name = tensor("query_27_cast_fp16")]; - tensor module_layers_13_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_13_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169408256)))]; - tensor module_layers_13_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_13_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169932608)))]; - tensor linear_120_cast_fp16 = linear(bias = module_layers_13_self_attn_linear_q_bias_to_fp16, weight = module_layers_13_self_attn_linear_q_weight_to_fp16, x = query_27_cast_fp16)[name = tensor("linear_120_cast_fp16")]; - tensor var_2632 = const()[name = tensor("op_2632"), val = tensor([1, -1, 8, 64])]; - tensor q_79_cast_fp16 = reshape(shape = var_2632, x = linear_120_cast_fp16)[name = tensor("q_79_cast_fp16")]; - tensor module_layers_13_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_13_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169933696)))]; - tensor module_layers_13_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_13_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170458048)))]; - tensor linear_121_cast_fp16 = linear(bias = module_layers_13_self_attn_linear_k_bias_to_fp16, weight = module_layers_13_self_attn_linear_k_weight_to_fp16, x = query_27_cast_fp16)[name = tensor("linear_121_cast_fp16")]; - tensor var_2637 = const()[name = tensor("op_2637"), val = tensor([1, -1, 8, 64])]; - tensor k_53_cast_fp16 = reshape(shape = var_2637, x = linear_121_cast_fp16)[name = tensor("k_53_cast_fp16")]; - tensor module_layers_13_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_13_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170459136)))]; - tensor module_layers_13_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_13_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170983488)))]; - tensor linear_122_cast_fp16 = linear(bias = module_layers_13_self_attn_linear_v_bias_to_fp16, weight = module_layers_13_self_attn_linear_v_weight_to_fp16, x = query_27_cast_fp16)[name = tensor("linear_122_cast_fp16")]; - tensor var_2642 = const()[name = tensor("op_2642"), val = tensor([1, -1, 8, 64])]; - tensor v_27_cast_fp16 = reshape(shape = var_2642, x = linear_122_cast_fp16)[name = tensor("v_27_cast_fp16")]; - tensor value_29_perm_0 = const()[name = tensor("value_29_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_13_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_13_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170984576)))]; - tensor var_2654_cast_fp16 = add(x = q_79_cast_fp16, y = module_layers_13_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2654_cast_fp16")]; - tensor module_layers_13_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_13_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170985664)))]; - tensor var_2656_cast_fp16 = add(x = q_79_cast_fp16, y = module_layers_13_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2656_cast_fp16")]; - tensor q_with_bias_v_27_perm_0 = const()[name = tensor("q_with_bias_v_27_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_293_transpose_x_0 = const()[name = tensor("x_293_transpose_x_0"), val = tensor(false)]; - tensor x_293_transpose_y_0 = const()[name = tensor("x_293_transpose_y_0"), val = tensor(false)]; - tensor var_2658_to_fp16 = const()[name = tensor("op_2658_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170986752)))]; - tensor q_with_bias_v_27_cast_fp16 = transpose(perm = q_with_bias_v_27_perm_0, x = var_2656_cast_fp16)[name = tensor("transpose_112")]; - tensor x_293_cast_fp16 = matmul(transpose_x = x_293_transpose_x_0, transpose_y = x_293_transpose_y_0, x = q_with_bias_v_27_cast_fp16, y = var_2658_to_fp16)[name = tensor("x_293_cast_fp16")]; - tensor x_295_pad_0 = const()[name = tensor("x_295_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_295_mode_0 = const()[name = tensor("x_295_mode_0"), val = tensor("constant")]; - tensor const_200_to_fp16 = const()[name = tensor("const_200_to_fp16"), val = tensor(0x0p+0)]; - tensor x_295_cast_fp16 = pad(constant_val = const_200_to_fp16, mode = x_295_mode_0, pad = x_295_pad_0, x = x_293_cast_fp16)[name = tensor("x_295_cast_fp16")]; - tensor var_2666 = const()[name = tensor("op_2666"), val = tensor([1, 8, -1, 188])]; - tensor x_297_cast_fp16 = reshape(shape = var_2666, x = x_295_cast_fp16)[name = tensor("x_297_cast_fp16")]; - tensor var_2670_begin_0 = const()[name = tensor("op_2670_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_2670_end_0 = const()[name = tensor("op_2670_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_2670_end_mask_0 = const()[name = tensor("op_2670_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_2670_cast_fp16 = slice_by_index(begin = var_2670_begin_0, end = var_2670_end_0, end_mask = var_2670_end_mask_0, x = x_297_cast_fp16)[name = tensor("op_2670_cast_fp16")]; - tensor var_2671 = const()[name = tensor("op_2671"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_53_cast_fp16 = reshape(shape = var_2671, x = var_2670_cast_fp16)[name = tensor("matrix_bd_53_cast_fp16")]; - tensor matrix_ac_27_transpose_x_0 = const()[name = tensor("matrix_ac_27_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_27_transpose_y_0 = const()[name = tensor("matrix_ac_27_transpose_y_0"), val = tensor(false)]; - tensor transpose_77_perm_0 = const()[name = tensor("transpose_77_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_78_perm_0 = const()[name = tensor("transpose_78_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_78 = transpose(perm = transpose_78_perm_0, x = k_53_cast_fp16)[name = tensor("transpose_110")]; - tensor transpose_77 = transpose(perm = transpose_77_perm_0, x = var_2654_cast_fp16)[name = tensor("transpose_111")]; - tensor matrix_ac_27_cast_fp16 = matmul(transpose_x = matrix_ac_27_transpose_x_0, transpose_y = matrix_ac_27_transpose_y_0, x = transpose_77, y = transpose_78)[name = tensor("matrix_ac_27_cast_fp16")]; - tensor matrix_bd_55_begin_0 = const()[name = tensor("matrix_bd_55_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_55_end_0 = const()[name = tensor("matrix_bd_55_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_55_end_mask_0 = const()[name = tensor("matrix_bd_55_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_55_cast_fp16 = slice_by_index(begin = matrix_bd_55_begin_0, end = matrix_bd_55_end_0, end_mask = matrix_bd_55_end_mask_0, x = matrix_bd_53_cast_fp16)[name = tensor("matrix_bd_55_cast_fp16")]; - tensor var_2680_cast_fp16 = add(x = matrix_ac_27_cast_fp16, y = matrix_bd_55_cast_fp16)[name = tensor("op_2680_cast_fp16")]; - tensor _inversed_scores_53_y_0_to_fp16 = const()[name = tensor("_inversed_scores_53_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_53_cast_fp16 = mul(x = var_2680_cast_fp16, y = _inversed_scores_53_y_0_to_fp16)[name = tensor("_inversed_scores_53_cast_fp16")]; - tensor scores_55_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_53_cast_fp16, cond = mask_11)[name = tensor("scores_55_cast_fp16")]; - tensor var_2686_cast_fp16 = softmax(axis = var_23, x = scores_55_cast_fp16)[name = tensor("op_2686_cast_fp16")]; - tensor input_709_cast_fp16 = select(a = var_11_to_fp16, b = var_2686_cast_fp16, cond = mask_11)[name = tensor("input_709_cast_fp16")]; - tensor x_299_transpose_x_0 = const()[name = tensor("x_299_transpose_x_0"), val = tensor(false)]; - tensor x_299_transpose_y_0 = const()[name = tensor("x_299_transpose_y_0"), val = tensor(false)]; - tensor value_29_cast_fp16 = transpose(perm = value_29_perm_0, x = v_27_cast_fp16)[name = tensor("transpose_113")]; - tensor x_299_cast_fp16 = matmul(transpose_x = x_299_transpose_x_0, transpose_y = x_299_transpose_y_0, x = input_709_cast_fp16, y = value_29_cast_fp16)[name = tensor("x_299_cast_fp16")]; - tensor var_2690_perm_0 = const()[name = tensor("op_2690_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_2691 = const()[name = tensor("op_2691"), val = tensor([1, -1, 512])]; - tensor var_2690_cast_fp16 = transpose(perm = var_2690_perm_0, x = x_299_cast_fp16)[name = tensor("transpose_109")]; - tensor input_711_cast_fp16 = reshape(shape = var_2691, x = var_2690_cast_fp16)[name = tensor("input_711_cast_fp16")]; - tensor module_layers_13_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_13_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171370816)))]; - tensor module_layers_13_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_13_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171895168)))]; - tensor linear_124_cast_fp16 = linear(bias = module_layers_13_self_attn_linear_out_bias_to_fp16, weight = module_layers_13_self_attn_linear_out_weight_to_fp16, x = input_711_cast_fp16)[name = tensor("linear_124_cast_fp16")]; - tensor input_715_cast_fp16 = add(x = input_707_cast_fp16, y = linear_124_cast_fp16)[name = tensor("input_715_cast_fp16")]; - tensor x_303_axes_0 = const()[name = tensor("x_303_axes_0"), val = tensor([-1])]; - tensor module_layers_13_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_13_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171896256)))]; - tensor module_layers_13_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_13_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171897344)))]; - tensor x_303_cast_fp16 = layer_norm(axes = x_303_axes_0, beta = module_layers_13_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_13_norm_conv_weight_to_fp16, x = input_715_cast_fp16)[name = tensor("x_303_cast_fp16")]; - tensor input_717_perm_0 = const()[name = tensor("input_717_perm_0"), val = tensor([0, 2, 1])]; - tensor input_719_pad_type_0 = const()[name = tensor("input_719_pad_type_0"), val = tensor("valid")]; - tensor input_719_strides_0 = const()[name = tensor("input_719_strides_0"), val = tensor([1])]; - tensor input_719_pad_0 = const()[name = tensor("input_719_pad_0"), val = tensor([0, 0])]; - tensor input_719_dilations_0 = const()[name = tensor("input_719_dilations_0"), val = tensor([1])]; - tensor input_719_groups_0 = const()[name = tensor("input_719_groups_0"), val = tensor(1)]; - tensor module_layers_13_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_13_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171898432)))]; - tensor module_layers_13_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_13_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(172947072)))]; - tensor input_717_cast_fp16 = transpose(perm = input_717_perm_0, x = x_303_cast_fp16)[name = tensor("transpose_108")]; - tensor input_719_cast_fp16 = conv(bias = module_layers_13_conv_pointwise_conv1_bias_to_fp16, dilations = input_719_dilations_0, groups = input_719_groups_0, pad = input_719_pad_0, pad_type = input_719_pad_type_0, strides = input_719_strides_0, weight = module_layers_13_conv_pointwise_conv1_weight_to_fp16, x = input_717_cast_fp16)[name = tensor("input_719_cast_fp16")]; - tensor x_305_split_num_splits_0 = const()[name = tensor("x_305_split_num_splits_0"), val = tensor(2)]; - tensor x_305_split_axis_0 = const()[name = tensor("x_305_split_axis_0"), val = tensor(1)]; - tensor x_305_split_cast_fp16_0, tensor x_305_split_cast_fp16_1 = split(axis = x_305_split_axis_0, num_splits = x_305_split_num_splits_0, x = input_719_cast_fp16)[name = tensor("x_305_split_cast_fp16")]; - tensor x_305_split_1_sigmoid_cast_fp16 = sigmoid(x = x_305_split_cast_fp16_1)[name = tensor("x_305_split_1_sigmoid_cast_fp16")]; - tensor x_305_cast_fp16 = mul(x = x_305_split_cast_fp16_0, y = x_305_split_1_sigmoid_cast_fp16)[name = tensor("x_305_cast_fp16")]; - tensor input_721_cast_fp16 = select(a = var_11_to_fp16, b = x_305_cast_fp16, cond = var_453)[name = tensor("input_721_cast_fp16")]; - tensor input_723_pad_0 = const()[name = tensor("input_723_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_723_mode_0 = const()[name = tensor("input_723_mode_0"), val = tensor("constant")]; - tensor const_203_to_fp16 = const()[name = tensor("const_203_to_fp16"), val = tensor(0x0p+0)]; - tensor input_723_cast_fp16 = pad(constant_val = const_203_to_fp16, mode = input_723_mode_0, pad = input_723_pad_0, x = input_721_cast_fp16)[name = tensor("input_723_cast_fp16")]; - tensor input_725_pad_type_0 = const()[name = tensor("input_725_pad_type_0"), val = tensor("valid")]; - tensor input_725_groups_0 = const()[name = tensor("input_725_groups_0"), val = tensor(512)]; - tensor input_725_strides_0 = const()[name = tensor("input_725_strides_0"), val = tensor([1])]; - tensor input_725_pad_0 = const()[name = tensor("input_725_pad_0"), val = tensor([0, 0])]; - tensor input_725_dilations_0 = const()[name = tensor("input_725_dilations_0"), val = tensor([1])]; - tensor const_260_to_fp16 = const()[name = tensor("const_260_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(172949184)))]; - tensor const_261_to_fp16 = const()[name = tensor("const_261_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(172958464)))]; - tensor input_727_cast_fp16 = conv(bias = const_261_to_fp16, dilations = input_725_dilations_0, groups = input_725_groups_0, pad = input_725_pad_0, pad_type = input_725_pad_type_0, strides = input_725_strides_0, weight = const_260_to_fp16, x = input_723_cast_fp16)[name = tensor("input_727_cast_fp16")]; - tensor input_729_cast_fp16 = silu(x = input_727_cast_fp16)[name = tensor("input_729_cast_fp16")]; - tensor x_307_pad_type_0 = const()[name = tensor("x_307_pad_type_0"), val = tensor("valid")]; - tensor x_307_strides_0 = const()[name = tensor("x_307_strides_0"), val = tensor([1])]; - tensor x_307_pad_0 = const()[name = tensor("x_307_pad_0"), val = tensor([0, 0])]; - tensor x_307_dilations_0 = const()[name = tensor("x_307_dilations_0"), val = tensor([1])]; - tensor x_307_groups_0 = const()[name = tensor("x_307_groups_0"), val = tensor(1)]; - tensor module_layers_13_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_13_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(172959552)))]; - tensor module_layers_13_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_13_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(173483904)))]; - tensor x_307_cast_fp16 = conv(bias = module_layers_13_conv_pointwise_conv2_bias_to_fp16, dilations = x_307_dilations_0, groups = x_307_groups_0, pad = x_307_pad_0, pad_type = x_307_pad_type_0, strides = x_307_strides_0, weight = module_layers_13_conv_pointwise_conv2_weight_to_fp16, x = input_729_cast_fp16)[name = tensor("x_307_cast_fp16")]; - tensor input_731_perm_0 = const()[name = tensor("input_731_perm_0"), val = tensor([0, 2, 1])]; - tensor input_731_cast_fp16 = transpose(perm = input_731_perm_0, x = x_307_cast_fp16)[name = tensor("transpose_107")]; - tensor input_733_cast_fp16 = add(x = input_715_cast_fp16, y = input_731_cast_fp16)[name = tensor("input_733_cast_fp16")]; - tensor input_735_axes_0 = const()[name = tensor("input_735_axes_0"), val = tensor([-1])]; - tensor module_layers_13_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_13_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(173484992)))]; - tensor module_layers_13_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_13_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(173486080)))]; - tensor input_735_cast_fp16 = layer_norm(axes = input_735_axes_0, beta = module_layers_13_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_13_norm_feed_forward2_weight_to_fp16, x = input_733_cast_fp16)[name = tensor("input_735_cast_fp16")]; - tensor module_layers_13_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_13_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(173487168)))]; - tensor module_layers_13_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_13_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(175584384)))]; - tensor linear_125_cast_fp16 = linear(bias = module_layers_13_feed_forward2_linear1_bias_to_fp16, weight = module_layers_13_feed_forward2_linear1_weight_to_fp16, x = input_735_cast_fp16)[name = tensor("linear_125_cast_fp16")]; - tensor input_739_cast_fp16 = silu(x = linear_125_cast_fp16)[name = tensor("input_739_cast_fp16")]; - tensor module_layers_13_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_13_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(175588544)))]; - tensor module_layers_13_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_13_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(177685760)))]; - tensor linear_126_cast_fp16 = linear(bias = module_layers_13_feed_forward2_linear2_bias_to_fp16, weight = module_layers_13_feed_forward2_linear2_weight_to_fp16, x = input_739_cast_fp16)[name = tensor("linear_126_cast_fp16")]; - tensor var_2757_to_fp16 = const()[name = tensor("op_2757_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2758_cast_fp16 = mul(x = linear_126_cast_fp16, y = var_2757_to_fp16)[name = tensor("op_2758_cast_fp16")]; - tensor input_745_cast_fp16 = add(x = input_733_cast_fp16, y = var_2758_cast_fp16)[name = tensor("input_745_cast_fp16")]; - tensor input_747_axes_0 = const()[name = tensor("input_747_axes_0"), val = tensor([-1])]; - tensor module_layers_13_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_13_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(177686848)))]; - tensor module_layers_13_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_13_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(177687936)))]; - tensor input_747_cast_fp16 = layer_norm(axes = input_747_axes_0, beta = module_layers_13_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_13_norm_out_weight_to_fp16, x = input_745_cast_fp16)[name = tensor("input_747_cast_fp16")]; - tensor input_749_axes_0 = const()[name = tensor("input_749_axes_0"), val = tensor([-1])]; - tensor module_layers_14_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_14_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(177689024)))]; - tensor module_layers_14_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_14_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(177690112)))]; - tensor input_749_cast_fp16 = layer_norm(axes = input_749_axes_0, beta = module_layers_14_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_14_norm_feed_forward1_weight_to_fp16, x = input_747_cast_fp16)[name = tensor("input_749_cast_fp16")]; - tensor module_layers_14_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_14_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(177691200)))]; - tensor module_layers_14_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_14_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(179788416)))]; - tensor linear_127_cast_fp16 = linear(bias = module_layers_14_feed_forward1_linear1_bias_to_fp16, weight = module_layers_14_feed_forward1_linear1_weight_to_fp16, x = input_749_cast_fp16)[name = tensor("linear_127_cast_fp16")]; - tensor input_753_cast_fp16 = silu(x = linear_127_cast_fp16)[name = tensor("input_753_cast_fp16")]; - tensor module_layers_14_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_14_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(179792576)))]; - tensor module_layers_14_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_14_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(181889792)))]; - tensor linear_128_cast_fp16 = linear(bias = module_layers_14_feed_forward1_linear2_bias_to_fp16, weight = module_layers_14_feed_forward1_linear2_weight_to_fp16, x = input_753_cast_fp16)[name = tensor("linear_128_cast_fp16")]; - tensor var_2788_to_fp16 = const()[name = tensor("op_2788_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2789_cast_fp16 = mul(x = linear_128_cast_fp16, y = var_2788_to_fp16)[name = tensor("op_2789_cast_fp16")]; - tensor input_759_cast_fp16 = add(x = input_747_cast_fp16, y = var_2789_cast_fp16)[name = tensor("input_759_cast_fp16")]; - tensor query_29_axes_0 = const()[name = tensor("query_29_axes_0"), val = tensor([-1])]; - tensor module_layers_14_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_14_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(181890880)))]; - tensor module_layers_14_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_14_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(181891968)))]; - tensor query_29_cast_fp16 = layer_norm(axes = query_29_axes_0, beta = module_layers_14_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_14_norm_self_att_weight_to_fp16, x = input_759_cast_fp16)[name = tensor("query_29_cast_fp16")]; - tensor module_layers_14_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_14_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(181893056)))]; - tensor module_layers_14_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_14_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182417408)))]; - tensor linear_129_cast_fp16 = linear(bias = module_layers_14_self_attn_linear_q_bias_to_fp16, weight = module_layers_14_self_attn_linear_q_weight_to_fp16, x = query_29_cast_fp16)[name = tensor("linear_129_cast_fp16")]; - tensor var_2806 = const()[name = tensor("op_2806"), val = tensor([1, -1, 8, 64])]; - tensor q_85_cast_fp16 = reshape(shape = var_2806, x = linear_129_cast_fp16)[name = tensor("q_85_cast_fp16")]; - tensor module_layers_14_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_14_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182418496)))]; - tensor module_layers_14_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_14_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182942848)))]; - tensor linear_130_cast_fp16 = linear(bias = module_layers_14_self_attn_linear_k_bias_to_fp16, weight = module_layers_14_self_attn_linear_k_weight_to_fp16, x = query_29_cast_fp16)[name = tensor("linear_130_cast_fp16")]; - tensor var_2811 = const()[name = tensor("op_2811"), val = tensor([1, -1, 8, 64])]; - tensor k_57_cast_fp16 = reshape(shape = var_2811, x = linear_130_cast_fp16)[name = tensor("k_57_cast_fp16")]; - tensor module_layers_14_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_14_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182943936)))]; - tensor module_layers_14_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_14_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183468288)))]; - tensor linear_131_cast_fp16 = linear(bias = module_layers_14_self_attn_linear_v_bias_to_fp16, weight = module_layers_14_self_attn_linear_v_weight_to_fp16, x = query_29_cast_fp16)[name = tensor("linear_131_cast_fp16")]; - tensor var_2816 = const()[name = tensor("op_2816"), val = tensor([1, -1, 8, 64])]; - tensor v_29_cast_fp16 = reshape(shape = var_2816, x = linear_131_cast_fp16)[name = tensor("v_29_cast_fp16")]; - tensor value_31_perm_0 = const()[name = tensor("value_31_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_14_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_14_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183469376)))]; - tensor var_2828_cast_fp16 = add(x = q_85_cast_fp16, y = module_layers_14_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2828_cast_fp16")]; - tensor module_layers_14_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_14_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183470464)))]; - tensor var_2830_cast_fp16 = add(x = q_85_cast_fp16, y = module_layers_14_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2830_cast_fp16")]; - tensor q_with_bias_v_29_perm_0 = const()[name = tensor("q_with_bias_v_29_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_315_transpose_x_0 = const()[name = tensor("x_315_transpose_x_0"), val = tensor(false)]; - tensor x_315_transpose_y_0 = const()[name = tensor("x_315_transpose_y_0"), val = tensor(false)]; - tensor var_2832_to_fp16 = const()[name = tensor("op_2832_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183471552)))]; - tensor q_with_bias_v_29_cast_fp16 = transpose(perm = q_with_bias_v_29_perm_0, x = var_2830_cast_fp16)[name = tensor("transpose_105")]; - tensor x_315_cast_fp16 = matmul(transpose_x = x_315_transpose_x_0, transpose_y = x_315_transpose_y_0, x = q_with_bias_v_29_cast_fp16, y = var_2832_to_fp16)[name = tensor("x_315_cast_fp16")]; - tensor x_317_pad_0 = const()[name = tensor("x_317_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_317_mode_0 = const()[name = tensor("x_317_mode_0"), val = tensor("constant")]; - tensor const_210_to_fp16 = const()[name = tensor("const_210_to_fp16"), val = tensor(0x0p+0)]; - tensor x_317_cast_fp16 = pad(constant_val = const_210_to_fp16, mode = x_317_mode_0, pad = x_317_pad_0, x = x_315_cast_fp16)[name = tensor("x_317_cast_fp16")]; - tensor var_2840 = const()[name = tensor("op_2840"), val = tensor([1, 8, -1, 188])]; - tensor x_319_cast_fp16 = reshape(shape = var_2840, x = x_317_cast_fp16)[name = tensor("x_319_cast_fp16")]; - tensor var_2844_begin_0 = const()[name = tensor("op_2844_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_2844_end_0 = const()[name = tensor("op_2844_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_2844_end_mask_0 = const()[name = tensor("op_2844_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_2844_cast_fp16 = slice_by_index(begin = var_2844_begin_0, end = var_2844_end_0, end_mask = var_2844_end_mask_0, x = x_319_cast_fp16)[name = tensor("op_2844_cast_fp16")]; - tensor var_2845 = const()[name = tensor("op_2845"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_57_cast_fp16 = reshape(shape = var_2845, x = var_2844_cast_fp16)[name = tensor("matrix_bd_57_cast_fp16")]; - tensor matrix_ac_29_transpose_x_0 = const()[name = tensor("matrix_ac_29_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_29_transpose_y_0 = const()[name = tensor("matrix_ac_29_transpose_y_0"), val = tensor(false)]; - tensor transpose_79_perm_0 = const()[name = tensor("transpose_79_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_80_perm_0 = const()[name = tensor("transpose_80_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_80 = transpose(perm = transpose_80_perm_0, x = k_57_cast_fp16)[name = tensor("transpose_103")]; - tensor transpose_79 = transpose(perm = transpose_79_perm_0, x = var_2828_cast_fp16)[name = tensor("transpose_104")]; - tensor matrix_ac_29_cast_fp16 = matmul(transpose_x = matrix_ac_29_transpose_x_0, transpose_y = matrix_ac_29_transpose_y_0, x = transpose_79, y = transpose_80)[name = tensor("matrix_ac_29_cast_fp16")]; - tensor matrix_bd_59_begin_0 = const()[name = tensor("matrix_bd_59_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_59_end_0 = const()[name = tensor("matrix_bd_59_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_59_end_mask_0 = const()[name = tensor("matrix_bd_59_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_59_cast_fp16 = slice_by_index(begin = matrix_bd_59_begin_0, end = matrix_bd_59_end_0, end_mask = matrix_bd_59_end_mask_0, x = matrix_bd_57_cast_fp16)[name = tensor("matrix_bd_59_cast_fp16")]; - tensor var_2854_cast_fp16 = add(x = matrix_ac_29_cast_fp16, y = matrix_bd_59_cast_fp16)[name = tensor("op_2854_cast_fp16")]; - tensor _inversed_scores_57_y_0_to_fp16 = const()[name = tensor("_inversed_scores_57_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_57_cast_fp16 = mul(x = var_2854_cast_fp16, y = _inversed_scores_57_y_0_to_fp16)[name = tensor("_inversed_scores_57_cast_fp16")]; - tensor scores_59_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_57_cast_fp16, cond = mask_11)[name = tensor("scores_59_cast_fp16")]; - tensor var_2860_cast_fp16 = softmax(axis = var_23, x = scores_59_cast_fp16)[name = tensor("op_2860_cast_fp16")]; - tensor input_761_cast_fp16 = select(a = var_11_to_fp16, b = var_2860_cast_fp16, cond = mask_11)[name = tensor("input_761_cast_fp16")]; - tensor x_321_transpose_x_0 = const()[name = tensor("x_321_transpose_x_0"), val = tensor(false)]; - tensor x_321_transpose_y_0 = const()[name = tensor("x_321_transpose_y_0"), val = tensor(false)]; - tensor value_31_cast_fp16 = transpose(perm = value_31_perm_0, x = v_29_cast_fp16)[name = tensor("transpose_106")]; - tensor x_321_cast_fp16 = matmul(transpose_x = x_321_transpose_x_0, transpose_y = x_321_transpose_y_0, x = input_761_cast_fp16, y = value_31_cast_fp16)[name = tensor("x_321_cast_fp16")]; - tensor var_2864_perm_0 = const()[name = tensor("op_2864_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_2865 = const()[name = tensor("op_2865"), val = tensor([1, -1, 512])]; - tensor var_2864_cast_fp16 = transpose(perm = var_2864_perm_0, x = x_321_cast_fp16)[name = tensor("transpose_102")]; - tensor input_763_cast_fp16 = reshape(shape = var_2865, x = var_2864_cast_fp16)[name = tensor("input_763_cast_fp16")]; - tensor module_layers_14_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_14_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183855616)))]; - tensor module_layers_14_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_14_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184379968)))]; - tensor linear_133_cast_fp16 = linear(bias = module_layers_14_self_attn_linear_out_bias_to_fp16, weight = module_layers_14_self_attn_linear_out_weight_to_fp16, x = input_763_cast_fp16)[name = tensor("linear_133_cast_fp16")]; - tensor input_767_cast_fp16 = add(x = input_759_cast_fp16, y = linear_133_cast_fp16)[name = tensor("input_767_cast_fp16")]; - tensor x_325_axes_0 = const()[name = tensor("x_325_axes_0"), val = tensor([-1])]; - tensor module_layers_14_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_14_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184381056)))]; - tensor module_layers_14_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_14_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184382144)))]; - tensor x_325_cast_fp16 = layer_norm(axes = x_325_axes_0, beta = module_layers_14_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_14_norm_conv_weight_to_fp16, x = input_767_cast_fp16)[name = tensor("x_325_cast_fp16")]; - tensor input_769_perm_0 = const()[name = tensor("input_769_perm_0"), val = tensor([0, 2, 1])]; - tensor input_771_pad_type_0 = const()[name = tensor("input_771_pad_type_0"), val = tensor("valid")]; - tensor input_771_strides_0 = const()[name = tensor("input_771_strides_0"), val = tensor([1])]; - tensor input_771_pad_0 = const()[name = tensor("input_771_pad_0"), val = tensor([0, 0])]; - tensor input_771_dilations_0 = const()[name = tensor("input_771_dilations_0"), val = tensor([1])]; - tensor input_771_groups_0 = const()[name = tensor("input_771_groups_0"), val = tensor(1)]; - tensor module_layers_14_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_14_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184383232)))]; - tensor module_layers_14_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_14_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185431872)))]; - tensor input_769_cast_fp16 = transpose(perm = input_769_perm_0, x = x_325_cast_fp16)[name = tensor("transpose_101")]; - tensor input_771_cast_fp16 = conv(bias = module_layers_14_conv_pointwise_conv1_bias_to_fp16, dilations = input_771_dilations_0, groups = input_771_groups_0, pad = input_771_pad_0, pad_type = input_771_pad_type_0, strides = input_771_strides_0, weight = module_layers_14_conv_pointwise_conv1_weight_to_fp16, x = input_769_cast_fp16)[name = tensor("input_771_cast_fp16")]; - tensor x_327_split_num_splits_0 = const()[name = tensor("x_327_split_num_splits_0"), val = tensor(2)]; - tensor x_327_split_axis_0 = const()[name = tensor("x_327_split_axis_0"), val = tensor(1)]; - tensor x_327_split_cast_fp16_0, tensor x_327_split_cast_fp16_1 = split(axis = x_327_split_axis_0, num_splits = x_327_split_num_splits_0, x = input_771_cast_fp16)[name = tensor("x_327_split_cast_fp16")]; - tensor x_327_split_1_sigmoid_cast_fp16 = sigmoid(x = x_327_split_cast_fp16_1)[name = tensor("x_327_split_1_sigmoid_cast_fp16")]; - tensor x_327_cast_fp16 = mul(x = x_327_split_cast_fp16_0, y = x_327_split_1_sigmoid_cast_fp16)[name = tensor("x_327_cast_fp16")]; - tensor input_773_cast_fp16 = select(a = var_11_to_fp16, b = x_327_cast_fp16, cond = var_453)[name = tensor("input_773_cast_fp16")]; - tensor input_775_pad_0 = const()[name = tensor("input_775_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_775_mode_0 = const()[name = tensor("input_775_mode_0"), val = tensor("constant")]; - tensor const_213_to_fp16 = const()[name = tensor("const_213_to_fp16"), val = tensor(0x0p+0)]; - tensor input_775_cast_fp16 = pad(constant_val = const_213_to_fp16, mode = input_775_mode_0, pad = input_775_pad_0, x = input_773_cast_fp16)[name = tensor("input_775_cast_fp16")]; - tensor input_777_pad_type_0 = const()[name = tensor("input_777_pad_type_0"), val = tensor("valid")]; - tensor input_777_groups_0 = const()[name = tensor("input_777_groups_0"), val = tensor(512)]; - tensor input_777_strides_0 = const()[name = tensor("input_777_strides_0"), val = tensor([1])]; - tensor input_777_pad_0 = const()[name = tensor("input_777_pad_0"), val = tensor([0, 0])]; - tensor input_777_dilations_0 = const()[name = tensor("input_777_dilations_0"), val = tensor([1])]; - tensor const_262_to_fp16 = const()[name = tensor("const_262_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185433984)))]; - tensor const_263_to_fp16 = const()[name = tensor("const_263_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185443264)))]; - tensor input_779_cast_fp16 = conv(bias = const_263_to_fp16, dilations = input_777_dilations_0, groups = input_777_groups_0, pad = input_777_pad_0, pad_type = input_777_pad_type_0, strides = input_777_strides_0, weight = const_262_to_fp16, x = input_775_cast_fp16)[name = tensor("input_779_cast_fp16")]; - tensor input_781_cast_fp16 = silu(x = input_779_cast_fp16)[name = tensor("input_781_cast_fp16")]; - tensor x_329_pad_type_0 = const()[name = tensor("x_329_pad_type_0"), val = tensor("valid")]; - tensor x_329_strides_0 = const()[name = tensor("x_329_strides_0"), val = tensor([1])]; - tensor x_329_pad_0 = const()[name = tensor("x_329_pad_0"), val = tensor([0, 0])]; - tensor x_329_dilations_0 = const()[name = tensor("x_329_dilations_0"), val = tensor([1])]; - tensor x_329_groups_0 = const()[name = tensor("x_329_groups_0"), val = tensor(1)]; - tensor module_layers_14_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_14_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185444352)))]; - tensor module_layers_14_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_14_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185968704)))]; - tensor x_329_cast_fp16 = conv(bias = module_layers_14_conv_pointwise_conv2_bias_to_fp16, dilations = x_329_dilations_0, groups = x_329_groups_0, pad = x_329_pad_0, pad_type = x_329_pad_type_0, strides = x_329_strides_0, weight = module_layers_14_conv_pointwise_conv2_weight_to_fp16, x = input_781_cast_fp16)[name = tensor("x_329_cast_fp16")]; - tensor input_783_perm_0 = const()[name = tensor("input_783_perm_0"), val = tensor([0, 2, 1])]; - tensor input_783_cast_fp16 = transpose(perm = input_783_perm_0, x = x_329_cast_fp16)[name = tensor("transpose_100")]; - tensor input_785_cast_fp16 = add(x = input_767_cast_fp16, y = input_783_cast_fp16)[name = tensor("input_785_cast_fp16")]; - tensor input_787_axes_0 = const()[name = tensor("input_787_axes_0"), val = tensor([-1])]; - tensor module_layers_14_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_14_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185969792)))]; - tensor module_layers_14_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_14_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185970880)))]; - tensor input_787_cast_fp16 = layer_norm(axes = input_787_axes_0, beta = module_layers_14_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_14_norm_feed_forward2_weight_to_fp16, x = input_785_cast_fp16)[name = tensor("input_787_cast_fp16")]; - tensor module_layers_14_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_14_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185971968)))]; - tensor module_layers_14_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_14_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188069184)))]; - tensor linear_134_cast_fp16 = linear(bias = module_layers_14_feed_forward2_linear1_bias_to_fp16, weight = module_layers_14_feed_forward2_linear1_weight_to_fp16, x = input_787_cast_fp16)[name = tensor("linear_134_cast_fp16")]; - tensor input_791_cast_fp16 = silu(x = linear_134_cast_fp16)[name = tensor("input_791_cast_fp16")]; - tensor module_layers_14_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_14_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188073344)))]; - tensor module_layers_14_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_14_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190170560)))]; - tensor linear_135_cast_fp16 = linear(bias = module_layers_14_feed_forward2_linear2_bias_to_fp16, weight = module_layers_14_feed_forward2_linear2_weight_to_fp16, x = input_791_cast_fp16)[name = tensor("linear_135_cast_fp16")]; - tensor var_2931_to_fp16 = const()[name = tensor("op_2931_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2932_cast_fp16 = mul(x = linear_135_cast_fp16, y = var_2931_to_fp16)[name = tensor("op_2932_cast_fp16")]; - tensor input_797_cast_fp16 = add(x = input_785_cast_fp16, y = var_2932_cast_fp16)[name = tensor("input_797_cast_fp16")]; - tensor input_799_axes_0 = const()[name = tensor("input_799_axes_0"), val = tensor([-1])]; - tensor module_layers_14_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_14_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190171648)))]; - tensor module_layers_14_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_14_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190172736)))]; - tensor input_799_cast_fp16 = layer_norm(axes = input_799_axes_0, beta = module_layers_14_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_14_norm_out_weight_to_fp16, x = input_797_cast_fp16)[name = tensor("input_799_cast_fp16")]; - tensor input_801_axes_0 = const()[name = tensor("input_801_axes_0"), val = tensor([-1])]; - tensor module_layers_15_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_15_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190173824)))]; - tensor module_layers_15_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_15_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190174912)))]; - tensor input_801_cast_fp16 = layer_norm(axes = input_801_axes_0, beta = module_layers_15_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_15_norm_feed_forward1_weight_to_fp16, x = input_799_cast_fp16)[name = tensor("input_801_cast_fp16")]; - tensor module_layers_15_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_15_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190176000)))]; - tensor module_layers_15_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_15_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(192273216)))]; - tensor linear_136_cast_fp16 = linear(bias = module_layers_15_feed_forward1_linear1_bias_to_fp16, weight = module_layers_15_feed_forward1_linear1_weight_to_fp16, x = input_801_cast_fp16)[name = tensor("linear_136_cast_fp16")]; - tensor input_805_cast_fp16 = silu(x = linear_136_cast_fp16)[name = tensor("input_805_cast_fp16")]; - tensor module_layers_15_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_15_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(192277376)))]; - tensor module_layers_15_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_15_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194374592)))]; - tensor linear_137_cast_fp16 = linear(bias = module_layers_15_feed_forward1_linear2_bias_to_fp16, weight = module_layers_15_feed_forward1_linear2_weight_to_fp16, x = input_805_cast_fp16)[name = tensor("linear_137_cast_fp16")]; - tensor var_2962_to_fp16 = const()[name = tensor("op_2962_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2963_cast_fp16 = mul(x = linear_137_cast_fp16, y = var_2962_to_fp16)[name = tensor("op_2963_cast_fp16")]; - tensor input_811_cast_fp16 = add(x = input_799_cast_fp16, y = var_2963_cast_fp16)[name = tensor("input_811_cast_fp16")]; - tensor query_31_axes_0 = const()[name = tensor("query_31_axes_0"), val = tensor([-1])]; - tensor module_layers_15_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_15_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194375680)))]; - tensor module_layers_15_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_15_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194376768)))]; - tensor query_31_cast_fp16 = layer_norm(axes = query_31_axes_0, beta = module_layers_15_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_15_norm_self_att_weight_to_fp16, x = input_811_cast_fp16)[name = tensor("query_31_cast_fp16")]; - tensor module_layers_15_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_15_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194377856)))]; - tensor module_layers_15_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_15_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194902208)))]; - tensor linear_138_cast_fp16 = linear(bias = module_layers_15_self_attn_linear_q_bias_to_fp16, weight = module_layers_15_self_attn_linear_q_weight_to_fp16, x = query_31_cast_fp16)[name = tensor("linear_138_cast_fp16")]; - tensor var_2980 = const()[name = tensor("op_2980"), val = tensor([1, -1, 8, 64])]; - tensor q_91_cast_fp16 = reshape(shape = var_2980, x = linear_138_cast_fp16)[name = tensor("q_91_cast_fp16")]; - tensor module_layers_15_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_15_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194903296)))]; - tensor module_layers_15_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_15_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195427648)))]; - tensor linear_139_cast_fp16 = linear(bias = module_layers_15_self_attn_linear_k_bias_to_fp16, weight = module_layers_15_self_attn_linear_k_weight_to_fp16, x = query_31_cast_fp16)[name = tensor("linear_139_cast_fp16")]; - tensor var_2985 = const()[name = tensor("op_2985"), val = tensor([1, -1, 8, 64])]; - tensor k_61_cast_fp16 = reshape(shape = var_2985, x = linear_139_cast_fp16)[name = tensor("k_61_cast_fp16")]; - tensor module_layers_15_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_15_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195428736)))]; - tensor module_layers_15_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_15_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195953088)))]; - tensor linear_140_cast_fp16 = linear(bias = module_layers_15_self_attn_linear_v_bias_to_fp16, weight = module_layers_15_self_attn_linear_v_weight_to_fp16, x = query_31_cast_fp16)[name = tensor("linear_140_cast_fp16")]; - tensor var_2990 = const()[name = tensor("op_2990"), val = tensor([1, -1, 8, 64])]; - tensor v_31_cast_fp16 = reshape(shape = var_2990, x = linear_140_cast_fp16)[name = tensor("v_31_cast_fp16")]; - tensor value_33_perm_0 = const()[name = tensor("value_33_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_15_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_15_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195954176)))]; - tensor var_3002_cast_fp16 = add(x = q_91_cast_fp16, y = module_layers_15_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3002_cast_fp16")]; - tensor module_layers_15_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_15_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195955264)))]; - tensor var_3004_cast_fp16 = add(x = q_91_cast_fp16, y = module_layers_15_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3004_cast_fp16")]; - tensor q_with_bias_v_31_perm_0 = const()[name = tensor("q_with_bias_v_31_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_337_transpose_x_0 = const()[name = tensor("x_337_transpose_x_0"), val = tensor(false)]; - tensor x_337_transpose_y_0 = const()[name = tensor("x_337_transpose_y_0"), val = tensor(false)]; - tensor var_3006_to_fp16 = const()[name = tensor("op_3006_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195956352)))]; - tensor q_with_bias_v_31_cast_fp16 = transpose(perm = q_with_bias_v_31_perm_0, x = var_3004_cast_fp16)[name = tensor("transpose_98")]; - tensor x_337_cast_fp16 = matmul(transpose_x = x_337_transpose_x_0, transpose_y = x_337_transpose_y_0, x = q_with_bias_v_31_cast_fp16, y = var_3006_to_fp16)[name = tensor("x_337_cast_fp16")]; - tensor x_339_pad_0 = const()[name = tensor("x_339_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_339_mode_0 = const()[name = tensor("x_339_mode_0"), val = tensor("constant")]; - tensor const_220_to_fp16 = const()[name = tensor("const_220_to_fp16"), val = tensor(0x0p+0)]; - tensor x_339_cast_fp16 = pad(constant_val = const_220_to_fp16, mode = x_339_mode_0, pad = x_339_pad_0, x = x_337_cast_fp16)[name = tensor("x_339_cast_fp16")]; - tensor var_3014 = const()[name = tensor("op_3014"), val = tensor([1, 8, -1, 188])]; - tensor x_341_cast_fp16 = reshape(shape = var_3014, x = x_339_cast_fp16)[name = tensor("x_341_cast_fp16")]; - tensor var_3018_begin_0 = const()[name = tensor("op_3018_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_3018_end_0 = const()[name = tensor("op_3018_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_3018_end_mask_0 = const()[name = tensor("op_3018_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_3018_cast_fp16 = slice_by_index(begin = var_3018_begin_0, end = var_3018_end_0, end_mask = var_3018_end_mask_0, x = x_341_cast_fp16)[name = tensor("op_3018_cast_fp16")]; - tensor var_3019 = const()[name = tensor("op_3019"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_61_cast_fp16 = reshape(shape = var_3019, x = var_3018_cast_fp16)[name = tensor("matrix_bd_61_cast_fp16")]; - tensor matrix_ac_31_transpose_x_0 = const()[name = tensor("matrix_ac_31_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_31_transpose_y_0 = const()[name = tensor("matrix_ac_31_transpose_y_0"), val = tensor(false)]; - tensor transpose_81_perm_0 = const()[name = tensor("transpose_81_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_82_perm_0 = const()[name = tensor("transpose_82_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_82 = transpose(perm = transpose_82_perm_0, x = k_61_cast_fp16)[name = tensor("transpose_96")]; - tensor transpose_81 = transpose(perm = transpose_81_perm_0, x = var_3002_cast_fp16)[name = tensor("transpose_97")]; - tensor matrix_ac_31_cast_fp16 = matmul(transpose_x = matrix_ac_31_transpose_x_0, transpose_y = matrix_ac_31_transpose_y_0, x = transpose_81, y = transpose_82)[name = tensor("matrix_ac_31_cast_fp16")]; - tensor matrix_bd_63_begin_0 = const()[name = tensor("matrix_bd_63_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_63_end_0 = const()[name = tensor("matrix_bd_63_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_63_end_mask_0 = const()[name = tensor("matrix_bd_63_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_63_cast_fp16 = slice_by_index(begin = matrix_bd_63_begin_0, end = matrix_bd_63_end_0, end_mask = matrix_bd_63_end_mask_0, x = matrix_bd_61_cast_fp16)[name = tensor("matrix_bd_63_cast_fp16")]; - tensor var_3028_cast_fp16 = add(x = matrix_ac_31_cast_fp16, y = matrix_bd_63_cast_fp16)[name = tensor("op_3028_cast_fp16")]; - tensor _inversed_scores_61_y_0_to_fp16 = const()[name = tensor("_inversed_scores_61_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_61_cast_fp16 = mul(x = var_3028_cast_fp16, y = _inversed_scores_61_y_0_to_fp16)[name = tensor("_inversed_scores_61_cast_fp16")]; - tensor scores_63_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_61_cast_fp16, cond = mask_11)[name = tensor("scores_63_cast_fp16")]; - tensor var_3034_cast_fp16 = softmax(axis = var_23, x = scores_63_cast_fp16)[name = tensor("op_3034_cast_fp16")]; - tensor input_813_cast_fp16 = select(a = var_11_to_fp16, b = var_3034_cast_fp16, cond = mask_11)[name = tensor("input_813_cast_fp16")]; - tensor x_343_transpose_x_0 = const()[name = tensor("x_343_transpose_x_0"), val = tensor(false)]; - tensor x_343_transpose_y_0 = const()[name = tensor("x_343_transpose_y_0"), val = tensor(false)]; - tensor value_33_cast_fp16 = transpose(perm = value_33_perm_0, x = v_31_cast_fp16)[name = tensor("transpose_99")]; - tensor x_343_cast_fp16 = matmul(transpose_x = x_343_transpose_x_0, transpose_y = x_343_transpose_y_0, x = input_813_cast_fp16, y = value_33_cast_fp16)[name = tensor("x_343_cast_fp16")]; - tensor var_3038_perm_0 = const()[name = tensor("op_3038_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_3039 = const()[name = tensor("op_3039"), val = tensor([1, -1, 512])]; - tensor var_3038_cast_fp16 = transpose(perm = var_3038_perm_0, x = x_343_cast_fp16)[name = tensor("transpose_95")]; - tensor input_815_cast_fp16 = reshape(shape = var_3039, x = var_3038_cast_fp16)[name = tensor("input_815_cast_fp16")]; - tensor module_layers_15_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_15_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196340416)))]; - tensor module_layers_15_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_15_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196864768)))]; - tensor linear_142_cast_fp16 = linear(bias = module_layers_15_self_attn_linear_out_bias_to_fp16, weight = module_layers_15_self_attn_linear_out_weight_to_fp16, x = input_815_cast_fp16)[name = tensor("linear_142_cast_fp16")]; - tensor input_819_cast_fp16 = add(x = input_811_cast_fp16, y = linear_142_cast_fp16)[name = tensor("input_819_cast_fp16")]; - tensor x_347_axes_0 = const()[name = tensor("x_347_axes_0"), val = tensor([-1])]; - tensor module_layers_15_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_15_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196865856)))]; - tensor module_layers_15_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_15_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196866944)))]; - tensor x_347_cast_fp16 = layer_norm(axes = x_347_axes_0, beta = module_layers_15_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_15_norm_conv_weight_to_fp16, x = input_819_cast_fp16)[name = tensor("x_347_cast_fp16")]; - tensor input_821_perm_0 = const()[name = tensor("input_821_perm_0"), val = tensor([0, 2, 1])]; - tensor input_823_pad_type_0 = const()[name = tensor("input_823_pad_type_0"), val = tensor("valid")]; - tensor input_823_strides_0 = const()[name = tensor("input_823_strides_0"), val = tensor([1])]; - tensor input_823_pad_0 = const()[name = tensor("input_823_pad_0"), val = tensor([0, 0])]; - tensor input_823_dilations_0 = const()[name = tensor("input_823_dilations_0"), val = tensor([1])]; - tensor input_823_groups_0 = const()[name = tensor("input_823_groups_0"), val = tensor(1)]; - tensor module_layers_15_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_15_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196868032)))]; - tensor module_layers_15_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_15_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(197916672)))]; - tensor input_821_cast_fp16 = transpose(perm = input_821_perm_0, x = x_347_cast_fp16)[name = tensor("transpose_94")]; - tensor input_823_cast_fp16 = conv(bias = module_layers_15_conv_pointwise_conv1_bias_to_fp16, dilations = input_823_dilations_0, groups = input_823_groups_0, pad = input_823_pad_0, pad_type = input_823_pad_type_0, strides = input_823_strides_0, weight = module_layers_15_conv_pointwise_conv1_weight_to_fp16, x = input_821_cast_fp16)[name = tensor("input_823_cast_fp16")]; - tensor x_349_split_num_splits_0 = const()[name = tensor("x_349_split_num_splits_0"), val = tensor(2)]; - tensor x_349_split_axis_0 = const()[name = tensor("x_349_split_axis_0"), val = tensor(1)]; - tensor x_349_split_cast_fp16_0, tensor x_349_split_cast_fp16_1 = split(axis = x_349_split_axis_0, num_splits = x_349_split_num_splits_0, x = input_823_cast_fp16)[name = tensor("x_349_split_cast_fp16")]; - tensor x_349_split_1_sigmoid_cast_fp16 = sigmoid(x = x_349_split_cast_fp16_1)[name = tensor("x_349_split_1_sigmoid_cast_fp16")]; - tensor x_349_cast_fp16 = mul(x = x_349_split_cast_fp16_0, y = x_349_split_1_sigmoid_cast_fp16)[name = tensor("x_349_cast_fp16")]; - tensor input_825_cast_fp16 = select(a = var_11_to_fp16, b = x_349_cast_fp16, cond = var_453)[name = tensor("input_825_cast_fp16")]; - tensor input_827_pad_0 = const()[name = tensor("input_827_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_827_mode_0 = const()[name = tensor("input_827_mode_0"), val = tensor("constant")]; - tensor const_223_to_fp16 = const()[name = tensor("const_223_to_fp16"), val = tensor(0x0p+0)]; - tensor input_827_cast_fp16 = pad(constant_val = const_223_to_fp16, mode = input_827_mode_0, pad = input_827_pad_0, x = input_825_cast_fp16)[name = tensor("input_827_cast_fp16")]; - tensor input_829_pad_type_0 = const()[name = tensor("input_829_pad_type_0"), val = tensor("valid")]; - tensor input_829_groups_0 = const()[name = tensor("input_829_groups_0"), val = tensor(512)]; - tensor input_829_strides_0 = const()[name = tensor("input_829_strides_0"), val = tensor([1])]; - tensor input_829_pad_0 = const()[name = tensor("input_829_pad_0"), val = tensor([0, 0])]; - tensor input_829_dilations_0 = const()[name = tensor("input_829_dilations_0"), val = tensor([1])]; - tensor const_264_to_fp16 = const()[name = tensor("const_264_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(197918784)))]; - tensor const_265_to_fp16 = const()[name = tensor("const_265_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(197928064)))]; - tensor input_831_cast_fp16 = conv(bias = const_265_to_fp16, dilations = input_829_dilations_0, groups = input_829_groups_0, pad = input_829_pad_0, pad_type = input_829_pad_type_0, strides = input_829_strides_0, weight = const_264_to_fp16, x = input_827_cast_fp16)[name = tensor("input_831_cast_fp16")]; - tensor input_833_cast_fp16 = silu(x = input_831_cast_fp16)[name = tensor("input_833_cast_fp16")]; - tensor x_351_pad_type_0 = const()[name = tensor("x_351_pad_type_0"), val = tensor("valid")]; - tensor x_351_strides_0 = const()[name = tensor("x_351_strides_0"), val = tensor([1])]; - tensor x_351_pad_0 = const()[name = tensor("x_351_pad_0"), val = tensor([0, 0])]; - tensor x_351_dilations_0 = const()[name = tensor("x_351_dilations_0"), val = tensor([1])]; - tensor x_351_groups_0 = const()[name = tensor("x_351_groups_0"), val = tensor(1)]; - tensor module_layers_15_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_15_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(197929152)))]; - tensor module_layers_15_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_15_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198453504)))]; - tensor x_351_cast_fp16 = conv(bias = module_layers_15_conv_pointwise_conv2_bias_to_fp16, dilations = x_351_dilations_0, groups = x_351_groups_0, pad = x_351_pad_0, pad_type = x_351_pad_type_0, strides = x_351_strides_0, weight = module_layers_15_conv_pointwise_conv2_weight_to_fp16, x = input_833_cast_fp16)[name = tensor("x_351_cast_fp16")]; - tensor input_835_perm_0 = const()[name = tensor("input_835_perm_0"), val = tensor([0, 2, 1])]; - tensor input_835_cast_fp16 = transpose(perm = input_835_perm_0, x = x_351_cast_fp16)[name = tensor("transpose_93")]; - tensor input_837_cast_fp16 = add(x = input_819_cast_fp16, y = input_835_cast_fp16)[name = tensor("input_837_cast_fp16")]; - tensor input_839_axes_0 = const()[name = tensor("input_839_axes_0"), val = tensor([-1])]; - tensor module_layers_15_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_15_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198454592)))]; - tensor module_layers_15_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_15_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198455680)))]; - tensor input_839_cast_fp16 = layer_norm(axes = input_839_axes_0, beta = module_layers_15_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_15_norm_feed_forward2_weight_to_fp16, x = input_837_cast_fp16)[name = tensor("input_839_cast_fp16")]; - tensor module_layers_15_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_15_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198456768)))]; - tensor module_layers_15_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_15_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(200553984)))]; - tensor linear_143_cast_fp16 = linear(bias = module_layers_15_feed_forward2_linear1_bias_to_fp16, weight = module_layers_15_feed_forward2_linear1_weight_to_fp16, x = input_839_cast_fp16)[name = tensor("linear_143_cast_fp16")]; - tensor input_843_cast_fp16 = silu(x = linear_143_cast_fp16)[name = tensor("input_843_cast_fp16")]; - tensor module_layers_15_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_15_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(200558144)))]; - tensor module_layers_15_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_15_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202655360)))]; - tensor linear_144_cast_fp16 = linear(bias = module_layers_15_feed_forward2_linear2_bias_to_fp16, weight = module_layers_15_feed_forward2_linear2_weight_to_fp16, x = input_843_cast_fp16)[name = tensor("linear_144_cast_fp16")]; - tensor var_3105_to_fp16 = const()[name = tensor("op_3105_to_fp16"), val = tensor(0x1p-1)]; - tensor var_3106_cast_fp16 = mul(x = linear_144_cast_fp16, y = var_3105_to_fp16)[name = tensor("op_3106_cast_fp16")]; - tensor input_849_cast_fp16 = add(x = input_837_cast_fp16, y = var_3106_cast_fp16)[name = tensor("input_849_cast_fp16")]; - tensor input_851_axes_0 = const()[name = tensor("input_851_axes_0"), val = tensor([-1])]; - tensor module_layers_15_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_15_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202656448)))]; - tensor module_layers_15_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_15_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202657536)))]; - tensor input_851_cast_fp16 = layer_norm(axes = input_851_axes_0, beta = module_layers_15_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_15_norm_out_weight_to_fp16, x = input_849_cast_fp16)[name = tensor("input_851_cast_fp16")]; - tensor input_853_axes_0 = const()[name = tensor("input_853_axes_0"), val = tensor([-1])]; - tensor module_layers_16_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_16_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202658624)))]; - tensor module_layers_16_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_16_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202659712)))]; - tensor input_853_cast_fp16 = layer_norm(axes = input_853_axes_0, beta = module_layers_16_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_16_norm_feed_forward1_weight_to_fp16, x = input_851_cast_fp16)[name = tensor("input_853_cast_fp16")]; - tensor module_layers_16_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_16_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202660800)))]; - tensor module_layers_16_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_16_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(204758016)))]; - tensor linear_145_cast_fp16 = linear(bias = module_layers_16_feed_forward1_linear1_bias_to_fp16, weight = module_layers_16_feed_forward1_linear1_weight_to_fp16, x = input_853_cast_fp16)[name = tensor("linear_145_cast_fp16")]; - tensor input_857_cast_fp16 = silu(x = linear_145_cast_fp16)[name = tensor("input_857_cast_fp16")]; - tensor module_layers_16_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_16_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(204762176)))]; - tensor module_layers_16_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_16_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(206859392)))]; - tensor linear_146_cast_fp16 = linear(bias = module_layers_16_feed_forward1_linear2_bias_to_fp16, weight = module_layers_16_feed_forward1_linear2_weight_to_fp16, x = input_857_cast_fp16)[name = tensor("linear_146_cast_fp16")]; - tensor var_3136_to_fp16 = const()[name = tensor("op_3136_to_fp16"), val = tensor(0x1p-1)]; - tensor var_3137_cast_fp16 = mul(x = linear_146_cast_fp16, y = var_3136_to_fp16)[name = tensor("op_3137_cast_fp16")]; - tensor input_863_cast_fp16 = add(x = input_851_cast_fp16, y = var_3137_cast_fp16)[name = tensor("input_863_cast_fp16")]; - tensor query_axes_0 = const()[name = tensor("query_axes_0"), val = tensor([-1])]; - tensor module_layers_16_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_16_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(206860480)))]; - tensor module_layers_16_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_16_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(206861568)))]; - tensor query_cast_fp16 = layer_norm(axes = query_axes_0, beta = module_layers_16_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_16_norm_self_att_weight_to_fp16, x = input_863_cast_fp16)[name = tensor("query_cast_fp16")]; - tensor module_layers_16_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_16_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(206862656)))]; - tensor module_layers_16_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_16_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(207387008)))]; - tensor linear_147_cast_fp16 = linear(bias = module_layers_16_self_attn_linear_q_bias_to_fp16, weight = module_layers_16_self_attn_linear_q_weight_to_fp16, x = query_cast_fp16)[name = tensor("linear_147_cast_fp16")]; - tensor var_3154 = const()[name = tensor("op_3154"), val = tensor([1, -1, 8, 64])]; - tensor q_97_cast_fp16 = reshape(shape = var_3154, x = linear_147_cast_fp16)[name = tensor("q_97_cast_fp16")]; - tensor module_layers_16_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_16_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(207388096)))]; - tensor module_layers_16_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_16_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(207912448)))]; - tensor linear_148_cast_fp16 = linear(bias = module_layers_16_self_attn_linear_k_bias_to_fp16, weight = module_layers_16_self_attn_linear_k_weight_to_fp16, x = query_cast_fp16)[name = tensor("linear_148_cast_fp16")]; - tensor var_3159 = const()[name = tensor("op_3159"), val = tensor([1, -1, 8, 64])]; - tensor k_65_cast_fp16 = reshape(shape = var_3159, x = linear_148_cast_fp16)[name = tensor("k_65_cast_fp16")]; - tensor module_layers_16_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_16_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(207913536)))]; - tensor module_layers_16_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_16_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208437888)))]; - tensor linear_149_cast_fp16 = linear(bias = module_layers_16_self_attn_linear_v_bias_to_fp16, weight = module_layers_16_self_attn_linear_v_weight_to_fp16, x = query_cast_fp16)[name = tensor("linear_149_cast_fp16")]; - tensor var_3164 = const()[name = tensor("op_3164"), val = tensor([1, -1, 8, 64])]; - tensor v_cast_fp16 = reshape(shape = var_3164, x = linear_149_cast_fp16)[name = tensor("v_cast_fp16")]; - tensor value_perm_0 = const()[name = tensor("value_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_16_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_16_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208438976)))]; - tensor var_3176_cast_fp16 = add(x = q_97_cast_fp16, y = module_layers_16_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3176_cast_fp16")]; - tensor module_layers_16_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_16_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208440064)))]; - tensor var_3178_cast_fp16 = add(x = q_97_cast_fp16, y = module_layers_16_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3178_cast_fp16")]; - tensor q_with_bias_v_perm_0 = const()[name = tensor("q_with_bias_v_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_359_transpose_x_0 = const()[name = tensor("x_359_transpose_x_0"), val = tensor(false)]; - tensor x_359_transpose_y_0 = const()[name = tensor("x_359_transpose_y_0"), val = tensor(false)]; - tensor var_3180_to_fp16 = const()[name = tensor("op_3180_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208441152)))]; - tensor q_with_bias_v_cast_fp16 = transpose(perm = q_with_bias_v_perm_0, x = var_3178_cast_fp16)[name = tensor("transpose_91")]; - tensor x_359_cast_fp16 = matmul(transpose_x = x_359_transpose_x_0, transpose_y = x_359_transpose_y_0, x = q_with_bias_v_cast_fp16, y = var_3180_to_fp16)[name = tensor("x_359_cast_fp16")]; - tensor x_361_pad_0 = const()[name = tensor("x_361_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_361_mode_0 = const()[name = tensor("x_361_mode_0"), val = tensor("constant")]; - tensor const_230_to_fp16 = const()[name = tensor("const_230_to_fp16"), val = tensor(0x0p+0)]; - tensor x_361_cast_fp16 = pad(constant_val = const_230_to_fp16, mode = x_361_mode_0, pad = x_361_pad_0, x = x_359_cast_fp16)[name = tensor("x_361_cast_fp16")]; - tensor var_3188 = const()[name = tensor("op_3188"), val = tensor([1, 8, -1, 188])]; - tensor x_363_cast_fp16 = reshape(shape = var_3188, x = x_361_cast_fp16)[name = tensor("x_363_cast_fp16")]; - tensor var_3192_begin_0 = const()[name = tensor("op_3192_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_3192_end_0 = const()[name = tensor("op_3192_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_3192_end_mask_0 = const()[name = tensor("op_3192_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_3192_cast_fp16 = slice_by_index(begin = var_3192_begin_0, end = var_3192_end_0, end_mask = var_3192_end_mask_0, x = x_363_cast_fp16)[name = tensor("op_3192_cast_fp16")]; - tensor var_3193 = const()[name = tensor("op_3193"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_65_cast_fp16 = reshape(shape = var_3193, x = var_3192_cast_fp16)[name = tensor("matrix_bd_65_cast_fp16")]; - tensor matrix_ac_transpose_x_0 = const()[name = tensor("matrix_ac_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_transpose_y_0 = const()[name = tensor("matrix_ac_transpose_y_0"), val = tensor(false)]; - tensor transpose_83_perm_0 = const()[name = tensor("transpose_83_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_84_perm_0 = const()[name = tensor("transpose_84_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_84 = transpose(perm = transpose_84_perm_0, x = k_65_cast_fp16)[name = tensor("transpose_89")]; - tensor transpose_83 = transpose(perm = transpose_83_perm_0, x = var_3176_cast_fp16)[name = tensor("transpose_90")]; - tensor matrix_ac_cast_fp16 = matmul(transpose_x = matrix_ac_transpose_x_0, transpose_y = matrix_ac_transpose_y_0, x = transpose_83, y = transpose_84)[name = tensor("matrix_ac_cast_fp16")]; - tensor matrix_bd_begin_0 = const()[name = tensor("matrix_bd_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_end_0 = const()[name = tensor("matrix_bd_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_end_mask_0 = const()[name = tensor("matrix_bd_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_cast_fp16 = slice_by_index(begin = matrix_bd_begin_0, end = matrix_bd_end_0, end_mask = matrix_bd_end_mask_0, x = matrix_bd_65_cast_fp16)[name = tensor("matrix_bd_cast_fp16")]; - tensor var_3202_cast_fp16 = add(x = matrix_ac_cast_fp16, y = matrix_bd_cast_fp16)[name = tensor("op_3202_cast_fp16")]; - tensor _inversed_scores_65_y_0_to_fp16 = const()[name = tensor("_inversed_scores_65_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_65_cast_fp16 = mul(x = var_3202_cast_fp16, y = _inversed_scores_65_y_0_to_fp16)[name = tensor("_inversed_scores_65_cast_fp16")]; - tensor scores_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_65_cast_fp16, cond = mask_11)[name = tensor("scores_cast_fp16")]; - tensor var_3208_cast_fp16 = softmax(axis = var_23, x = scores_cast_fp16)[name = tensor("op_3208_cast_fp16")]; - tensor input_865_cast_fp16 = select(a = var_11_to_fp16, b = var_3208_cast_fp16, cond = mask_11)[name = tensor("input_865_cast_fp16")]; - tensor x_365_transpose_x_0 = const()[name = tensor("x_365_transpose_x_0"), val = tensor(false)]; - tensor x_365_transpose_y_0 = const()[name = tensor("x_365_transpose_y_0"), val = tensor(false)]; - tensor value_cast_fp16 = transpose(perm = value_perm_0, x = v_cast_fp16)[name = tensor("transpose_92")]; - tensor x_365_cast_fp16 = matmul(transpose_x = x_365_transpose_x_0, transpose_y = x_365_transpose_y_0, x = input_865_cast_fp16, y = value_cast_fp16)[name = tensor("x_365_cast_fp16")]; - tensor var_3212_perm_0 = const()[name = tensor("op_3212_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_3213 = const()[name = tensor("op_3213"), val = tensor([1, -1, 512])]; - tensor var_3212_cast_fp16 = transpose(perm = var_3212_perm_0, x = x_365_cast_fp16)[name = tensor("transpose_88")]; - tensor input_867_cast_fp16 = reshape(shape = var_3213, x = var_3212_cast_fp16)[name = tensor("input_867_cast_fp16")]; - tensor module_layers_16_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_16_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208825216)))]; - tensor module_layers_16_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_16_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(209349568)))]; - tensor linear_151_cast_fp16 = linear(bias = module_layers_16_self_attn_linear_out_bias_to_fp16, weight = module_layers_16_self_attn_linear_out_weight_to_fp16, x = input_867_cast_fp16)[name = tensor("linear_151_cast_fp16")]; - tensor input_871_cast_fp16 = add(x = input_863_cast_fp16, y = linear_151_cast_fp16)[name = tensor("input_871_cast_fp16")]; - tensor x_369_axes_0 = const()[name = tensor("x_369_axes_0"), val = tensor([-1])]; - tensor module_layers_16_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_16_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(209350656)))]; - tensor module_layers_16_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_16_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(209351744)))]; - tensor x_369_cast_fp16 = layer_norm(axes = x_369_axes_0, beta = module_layers_16_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_16_norm_conv_weight_to_fp16, x = input_871_cast_fp16)[name = tensor("x_369_cast_fp16")]; - tensor input_873_perm_0 = const()[name = tensor("input_873_perm_0"), val = tensor([0, 2, 1])]; - tensor input_875_pad_type_0 = const()[name = tensor("input_875_pad_type_0"), val = tensor("valid")]; - tensor input_875_strides_0 = const()[name = tensor("input_875_strides_0"), val = tensor([1])]; - tensor input_875_pad_0 = const()[name = tensor("input_875_pad_0"), val = tensor([0, 0])]; - tensor input_875_dilations_0 = const()[name = tensor("input_875_dilations_0"), val = tensor([1])]; - tensor input_875_groups_0 = const()[name = tensor("input_875_groups_0"), val = tensor(1)]; - tensor module_layers_16_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_16_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(209352832)))]; - tensor module_layers_16_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_16_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210401472)))]; - tensor input_873_cast_fp16 = transpose(perm = input_873_perm_0, x = x_369_cast_fp16)[name = tensor("transpose_87")]; - tensor input_875_cast_fp16 = conv(bias = module_layers_16_conv_pointwise_conv1_bias_to_fp16, dilations = input_875_dilations_0, groups = input_875_groups_0, pad = input_875_pad_0, pad_type = input_875_pad_type_0, strides = input_875_strides_0, weight = module_layers_16_conv_pointwise_conv1_weight_to_fp16, x = input_873_cast_fp16)[name = tensor("input_875_cast_fp16")]; - tensor x_371_split_num_splits_0 = const()[name = tensor("x_371_split_num_splits_0"), val = tensor(2)]; - tensor x_371_split_axis_0 = const()[name = tensor("x_371_split_axis_0"), val = tensor(1)]; - tensor x_371_split_cast_fp16_0, tensor x_371_split_cast_fp16_1 = split(axis = x_371_split_axis_0, num_splits = x_371_split_num_splits_0, x = input_875_cast_fp16)[name = tensor("x_371_split_cast_fp16")]; - tensor x_371_split_1_sigmoid_cast_fp16 = sigmoid(x = x_371_split_cast_fp16_1)[name = tensor("x_371_split_1_sigmoid_cast_fp16")]; - tensor x_371_cast_fp16 = mul(x = x_371_split_cast_fp16_0, y = x_371_split_1_sigmoid_cast_fp16)[name = tensor("x_371_cast_fp16")]; - tensor input_877_cast_fp16 = select(a = var_11_to_fp16, b = x_371_cast_fp16, cond = var_453)[name = tensor("input_877_cast_fp16")]; - tensor input_879_pad_0 = const()[name = tensor("input_879_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_879_mode_0 = const()[name = tensor("input_879_mode_0"), val = tensor("constant")]; - tensor const_233_to_fp16 = const()[name = tensor("const_233_to_fp16"), val = tensor(0x0p+0)]; - tensor input_879_cast_fp16 = pad(constant_val = const_233_to_fp16, mode = input_879_mode_0, pad = input_879_pad_0, x = input_877_cast_fp16)[name = tensor("input_879_cast_fp16")]; - tensor input_881_pad_type_0 = const()[name = tensor("input_881_pad_type_0"), val = tensor("valid")]; - tensor input_881_groups_0 = const()[name = tensor("input_881_groups_0"), val = tensor(512)]; - tensor input_881_strides_0 = const()[name = tensor("input_881_strides_0"), val = tensor([1])]; - tensor input_881_pad_0 = const()[name = tensor("input_881_pad_0"), val = tensor([0, 0])]; - tensor input_881_dilations_0 = const()[name = tensor("input_881_dilations_0"), val = tensor([1])]; - tensor const_266_to_fp16 = const()[name = tensor("const_266_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210403584)))]; - tensor const_267_to_fp16 = const()[name = tensor("const_267_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210412864)))]; - tensor input_883_cast_fp16 = conv(bias = const_267_to_fp16, dilations = input_881_dilations_0, groups = input_881_groups_0, pad = input_881_pad_0, pad_type = input_881_pad_type_0, strides = input_881_strides_0, weight = const_266_to_fp16, x = input_879_cast_fp16)[name = tensor("input_883_cast_fp16")]; - tensor input_885_cast_fp16 = silu(x = input_883_cast_fp16)[name = tensor("input_885_cast_fp16")]; - tensor x_373_pad_type_0 = const()[name = tensor("x_373_pad_type_0"), val = tensor("valid")]; - tensor x_373_strides_0 = const()[name = tensor("x_373_strides_0"), val = tensor([1])]; - tensor x_373_pad_0 = const()[name = tensor("x_373_pad_0"), val = tensor([0, 0])]; - tensor x_373_dilations_0 = const()[name = tensor("x_373_dilations_0"), val = tensor([1])]; - tensor x_373_groups_0 = const()[name = tensor("x_373_groups_0"), val = tensor(1)]; - tensor module_layers_16_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_16_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210413952)))]; - tensor module_layers_16_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_16_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210938304)))]; - tensor x_373_cast_fp16 = conv(bias = module_layers_16_conv_pointwise_conv2_bias_to_fp16, dilations = x_373_dilations_0, groups = x_373_groups_0, pad = x_373_pad_0, pad_type = x_373_pad_type_0, strides = x_373_strides_0, weight = module_layers_16_conv_pointwise_conv2_weight_to_fp16, x = input_885_cast_fp16)[name = tensor("x_373_cast_fp16")]; - tensor input_887_perm_0 = const()[name = tensor("input_887_perm_0"), val = tensor([0, 2, 1])]; - tensor input_887_cast_fp16 = transpose(perm = input_887_perm_0, x = x_373_cast_fp16)[name = tensor("transpose_86")]; - tensor input_889_cast_fp16 = add(x = input_871_cast_fp16, y = input_887_cast_fp16)[name = tensor("input_889_cast_fp16")]; - tensor input_891_axes_0 = const()[name = tensor("input_891_axes_0"), val = tensor([-1])]; - tensor module_layers_16_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_16_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210939392)))]; - tensor module_layers_16_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_16_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210940480)))]; - tensor input_891_cast_fp16 = layer_norm(axes = input_891_axes_0, beta = module_layers_16_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_16_norm_feed_forward2_weight_to_fp16, x = input_889_cast_fp16)[name = tensor("input_891_cast_fp16")]; - tensor module_layers_16_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_16_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210941568)))]; - tensor module_layers_16_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_16_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(213038784)))]; - tensor linear_152_cast_fp16 = linear(bias = module_layers_16_feed_forward2_linear1_bias_to_fp16, weight = module_layers_16_feed_forward2_linear1_weight_to_fp16, x = input_891_cast_fp16)[name = tensor("linear_152_cast_fp16")]; - tensor input_895_cast_fp16 = silu(x = linear_152_cast_fp16)[name = tensor("input_895_cast_fp16")]; - tensor module_layers_16_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_16_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(213042944)))]; - tensor module_layers_16_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_16_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(215140160)))]; - tensor linear_153_cast_fp16 = linear(bias = module_layers_16_feed_forward2_linear2_bias_to_fp16, weight = module_layers_16_feed_forward2_linear2_weight_to_fp16, x = input_895_cast_fp16)[name = tensor("linear_153_cast_fp16")]; - tensor var_3279_to_fp16 = const()[name = tensor("op_3279_to_fp16"), val = tensor(0x1p-1)]; - tensor var_3280_cast_fp16 = mul(x = linear_153_cast_fp16, y = var_3279_to_fp16)[name = tensor("op_3280_cast_fp16")]; - tensor input_cast_fp16 = add(x = input_889_cast_fp16, y = var_3280_cast_fp16)[name = tensor("input_cast_fp16")]; - tensor audio_signal_axes_0 = const()[name = tensor("audio_signal_axes_0"), val = tensor([-1])]; - tensor module_layers_16_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_16_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(215141248)))]; - tensor module_layers_16_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_16_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(215142336)))]; - tensor audio_signal_cast_fp16 = layer_norm(axes = audio_signal_axes_0, beta = module_layers_16_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_16_norm_out_weight_to_fp16, x = input_cast_fp16)[name = tensor("audio_signal_cast_fp16")]; - tensor obj_1_perm_0 = const()[name = tensor("obj_1_perm_0"), val = tensor([0, 2, 1])]; - tensor obj_1_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("obj_1_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; - tensor obj_1_cast_fp16 = transpose(perm = obj_1_perm_0, x = audio_signal_cast_fp16)[name = tensor("transpose_85")]; - tensor encoder_output = cast(dtype = obj_1_cast_fp16_to_fp32_dtype_0, x = obj_1_cast_fp16)[name = tensor("cast_172")]; - } -> (encoder_output, encoder_length); -} \ No newline at end of file diff --git a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/Encoder.mlmodelc/weights/weight.bin b/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/Encoder.mlmodelc/weights/weight.bin deleted file mode 100644 index c5bffa9667270bb21c093e55898f4fb2e299e426..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/Encoder.mlmodelc/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:cecf7994b2758397d992802a4f6e5d656e3a1aeb7bbedc2aa430b1316d62474c -size 215143424 diff --git a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecision.mlmodelc/analytics/coremldata.bin b/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecision.mlmodelc/analytics/coremldata.bin deleted file mode 100644 index 4b4b92e5e8e2b45c67e95237edd74d11f323a65a..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecision.mlmodelc/analytics/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:983ba26dd9276b8d2d4f75f3475aefb1817c542df87dbd0fdac95bd63647494f -size 243 diff --git a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecision.mlmodelc/coremldata.bin b/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecision.mlmodelc/coremldata.bin deleted file mode 100644 index 74e1d92e032200815d325fa2c7c84ee8562d6aad..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecision.mlmodelc/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:0800e3bdf4ecb1bd46fd27e1826d33125cd574f9ae1e15dd9ff70ea42944ca2d -size 476 diff --git a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecision.mlmodelc/metadata.json b/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecision.mlmodelc/metadata.json deleted file mode 100644 index fe49b70e7e98f7ec9524130b0027886d862f6f8a..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecision.mlmodelc/metadata.json +++ /dev/null @@ -1,102 +0,0 @@ -[ - { - "metadataOutputVersion" : "3.0", - "shortDescription" : "Parakeet 110M joint + decision head (split, softmax, argmax)", - "outputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1 × 188 × 1)", - "shortDescription" : "", - "shape" : "[1, 188, 1]", - "name" : "token_id", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 188 × 1)", - "shortDescription" : "", - "shape" : "[1, 188, 1]", - "name" : "token_prob", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1 × 188 × 1)", - "shortDescription" : "", - "shape" : "[1, 188, 1]", - "name" : "duration", - "type" : "MultiArray" - } - ], - "storagePrecision" : "Float16", - "modelParameters" : [ - - ], - "author" : "Fluid Inference", - "specificationVersion" : 8, - "mlProgramOperationTypeHistogram" : { - "Ios17.reduceArgmax" : 2, - "Ios17.squeeze" : 1, - "Ios17.cast" : 4, - "Ios17.linear" : 3, - "Ios17.transpose" : 2, - "Ios17.sliceByIndex" : 2, - "Ios17.add" : 1, - "Ios16.relu" : 1, - "Ios16.softmax" : 1, - "Ios17.gatherAlongAxis" : 1, - "Ios17.expandDims" : 3 - }, - "computePrecision" : "Mixed (Float16, Float32, Int16, Int32)", - "isUpdatable" : "0", - "stateSchema" : [ - - ], - "availability" : { - "macOS" : "14.0", - "tvOS" : "17.0", - "visionOS" : "1.0", - "watchOS" : "10.0", - "iOS" : "17.0", - "macCatalyst" : "17.0" - }, - "modelType" : { - "name" : "MLModelType_mlProgram" - }, - "inputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 512 × 188)", - "shortDescription" : "", - "shape" : "[1, 512, 188]", - "name" : "encoder", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 640 × 1)", - "shortDescription" : "", - "shape" : "[1, 640, 1]", - "name" : "decoder", - "type" : "MultiArray" - } - ], - "userDefinedMetadata" : { - "com.github.apple.coremltools.version" : "8.3.0", - "com.github.apple.coremltools.source_dialect" : "TorchScript", - "com.github.apple.coremltools.source" : "torch==2.9.0" - }, - "generatedClassName" : "parakeet_joint_decision", - "method" : "predict" - } -] \ No newline at end of file diff --git a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecision.mlmodelc/model.mil b/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecision.mlmodelc/model.mil deleted file mode 100644 index edfc5050f47ffbe1f5344799cef5dc0b21c392e0..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecision.mlmodelc/model.mil +++ /dev/null @@ -1,58 +0,0 @@ -program(1.0) -[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3510.2.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.9.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "8.3.0"}})] -{ - func main(tensor decoder, tensor encoder) { - tensor input_1_perm_0 = const()[name = tensor("input_1_perm_0"), val = tensor([0, 2, 1])]; - tensor encoder_to_fp16_dtype_0 = const()[name = tensor("encoder_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor input_3_perm_0 = const()[name = tensor("input_3_perm_0"), val = tensor([0, 2, 1])]; - tensor decoder_to_fp16_dtype_0 = const()[name = tensor("decoder_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor joint_module_enc_weight_to_fp16 = const()[name = tensor("joint_module_enc_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; - tensor joint_module_enc_bias_to_fp16 = const()[name = tensor("joint_module_enc_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(655488)))]; - tensor encoder_to_fp16 = cast(dtype = encoder_to_fp16_dtype_0, x = encoder)[name = tensor("cast_6")]; - tensor input_1_cast_fp16 = transpose(perm = input_1_perm_0, x = encoder_to_fp16)[name = tensor("transpose_1")]; - tensor linear_0_cast_fp16 = linear(bias = joint_module_enc_bias_to_fp16, weight = joint_module_enc_weight_to_fp16, x = input_1_cast_fp16)[name = tensor("linear_0_cast_fp16")]; - tensor joint_module_pred_weight_to_fp16 = const()[name = tensor("joint_module_pred_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(656832)))]; - tensor joint_module_pred_bias_to_fp16 = const()[name = tensor("joint_module_pred_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1476096)))]; - tensor decoder_to_fp16 = cast(dtype = decoder_to_fp16_dtype_0, x = decoder)[name = tensor("cast_5")]; - tensor input_3_cast_fp16 = transpose(perm = input_3_perm_0, x = decoder_to_fp16)[name = tensor("transpose_0")]; - tensor linear_1_cast_fp16 = linear(bias = joint_module_pred_bias_to_fp16, weight = joint_module_pred_weight_to_fp16, x = input_3_cast_fp16)[name = tensor("linear_1_cast_fp16")]; - tensor var_23_axes_0 = const()[name = tensor("op_23_axes_0"), val = tensor([2])]; - tensor var_23_cast_fp16 = expand_dims(axes = var_23_axes_0, x = linear_0_cast_fp16)[name = tensor("op_23_cast_fp16")]; - tensor var_24_axes_0 = const()[name = tensor("op_24_axes_0"), val = tensor([1])]; - tensor var_24_cast_fp16 = expand_dims(axes = var_24_axes_0, x = linear_1_cast_fp16)[name = tensor("op_24_cast_fp16")]; - tensor input_5_cast_fp16 = add(x = var_23_cast_fp16, y = var_24_cast_fp16)[name = tensor("input_5_cast_fp16")]; - tensor input_7_cast_fp16 = relu(x = input_5_cast_fp16)[name = tensor("input_7_cast_fp16")]; - tensor joint_module_joint_net_2_weight_to_fp16 = const()[name = tensor("joint_module_joint_net_2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1477440)))]; - tensor joint_module_joint_net_2_bias_to_fp16 = const()[name = tensor("joint_module_joint_net_2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2795904)))]; - tensor linear_2_cast_fp16 = linear(bias = joint_module_joint_net_2_bias_to_fp16, weight = joint_module_joint_net_2_weight_to_fp16, x = input_7_cast_fp16)[name = tensor("linear_2_cast_fp16")]; - tensor token_logits_begin_0 = const()[name = tensor("token_logits_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor token_logits_end_0 = const()[name = tensor("token_logits_end_0"), val = tensor([1, 188, 1, 1025])]; - tensor token_logits_end_mask_0 = const()[name = tensor("token_logits_end_mask_0"), val = tensor([true, true, true, false])]; - tensor token_logits_cast_fp16 = slice_by_index(begin = token_logits_begin_0, end = token_logits_end_0, end_mask = token_logits_end_mask_0, x = linear_2_cast_fp16)[name = tensor("token_logits_cast_fp16")]; - tensor duration_logits_begin_0 = const()[name = tensor("duration_logits_begin_0"), val = tensor([0, 0, 0, 1025])]; - tensor duration_logits_end_0 = const()[name = tensor("duration_logits_end_0"), val = tensor([1, 188, 1, 1030])]; - tensor duration_logits_end_mask_0 = const()[name = tensor("duration_logits_end_mask_0"), val = tensor([true, true, true, true])]; - tensor duration_logits_cast_fp16 = slice_by_index(begin = duration_logits_begin_0, end = duration_logits_end_0, end_mask = duration_logits_end_mask_0, x = linear_2_cast_fp16)[name = tensor("duration_logits_cast_fp16")]; - tensor var_43_axis_0 = const()[name = tensor("op_43_axis_0"), val = tensor(-1)]; - tensor var_43_keep_dims_0 = const()[name = tensor("op_43_keep_dims_0"), val = tensor(false)]; - tensor var_43_output_dtype_0 = const()[name = tensor("op_43_output_dtype_0"), val = tensor("int32")]; - tensor token_id = reduce_argmax(axis = var_43_axis_0, keep_dims = var_43_keep_dims_0, output_dtype = var_43_output_dtype_0, x = token_logits_cast_fp16)[name = tensor("op_43_cast_fp16")]; - tensor var_49 = const()[name = tensor("op_49"), val = tensor(-1)]; - tensor token_probs_all_cast_fp16 = softmax(axis = var_49, x = token_logits_cast_fp16)[name = tensor("token_probs_all_cast_fp16")]; - tensor var_58_axes_0 = const()[name = tensor("op_58_axes_0"), val = tensor([-1])]; - tensor var_58 = expand_dims(axes = var_58_axes_0, x = token_id)[name = tensor("op_58")]; - tensor var_59 = const()[name = tensor("op_59"), val = tensor(-1)]; - tensor var_61_validate_indices_0 = const()[name = tensor("op_61_validate_indices_0"), val = tensor(false)]; - tensor var_58_to_int16_dtype_0 = const()[name = tensor("op_58_to_int16_dtype_0"), val = tensor("int16")]; - tensor var_58_to_int16 = cast(dtype = var_58_to_int16_dtype_0, x = var_58)[name = tensor("cast_4")]; - tensor var_61_cast_fp16_cast_int16 = gather_along_axis(axis = var_59, indices = var_58_to_int16, validate_indices = var_61_validate_indices_0, x = token_probs_all_cast_fp16)[name = tensor("op_61_cast_fp16_cast_int16")]; - tensor var_63_axes_0 = const()[name = tensor("op_63_axes_0"), val = tensor([-1])]; - tensor var_63_cast_fp16 = squeeze(axes = var_63_axes_0, x = var_61_cast_fp16_cast_int16)[name = tensor("op_63_cast_fp16")]; - tensor var_63_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("op_63_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; - tensor var_66_axis_0 = const()[name = tensor("op_66_axis_0"), val = tensor(-1)]; - tensor var_66_keep_dims_0 = const()[name = tensor("op_66_keep_dims_0"), val = tensor(false)]; - tensor var_66_output_dtype_0 = const()[name = tensor("op_66_output_dtype_0"), val = tensor("int32")]; - tensor duration = reduce_argmax(axis = var_66_axis_0, keep_dims = var_66_keep_dims_0, output_dtype = var_66_output_dtype_0, x = duration_logits_cast_fp16)[name = tensor("op_66_cast_fp16")]; - tensor token_prob = cast(dtype = var_63_cast_fp16_to_fp32_dtype_0, x = var_63_cast_fp16)[name = tensor("cast_3")]; - } -> (token_id, token_prob, duration); -} \ No newline at end of file diff --git a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecision.mlmodelc/weights/weight.bin b/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecision.mlmodelc/weights/weight.bin deleted file mode 100644 index 747ec9d24e2d3e008bc869d2c404bb3d2653325c..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecision.mlmodelc/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:b3f771cb65b190f1873e39629676ed79b65a8361522f451b37bdba8b1106e6ff -size 2798028 diff --git a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecisionSingleStep.mlmodelc/analytics/coremldata.bin b/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecisionSingleStep.mlmodelc/analytics/coremldata.bin deleted file mode 100644 index ab3e67b973cf846c0af5e1f075f1b4b7ffb77a00..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecisionSingleStep.mlmodelc/analytics/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:c7c11c6bb985fab7f835ba687a575f1eb04f4c93b0783155d634adbc49f0e797 -size 243 diff --git a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecisionSingleStep.mlmodelc/coremldata.bin b/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecisionSingleStep.mlmodelc/coremldata.bin deleted file mode 100644 index bc71cc6aa73de6959f4c6718500197ebbf364633..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecisionSingleStep.mlmodelc/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:1af2cb9bcc13eec83ce006e4f1c2cf158393745cd9187428333fbcb6917da244 -size 535 diff --git a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecisionSingleStep.mlmodelc/metadata.json b/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecisionSingleStep.mlmodelc/metadata.json deleted file mode 100644 index b96db3615aa1d5e110c61c1f4694def827cb0a1f..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecisionSingleStep.mlmodelc/metadata.json +++ /dev/null @@ -1,123 +0,0 @@ -[ - { - "metadataOutputVersion" : "3.0", - "shortDescription" : "Parakeet 110M single-step joint decision (current frame)", - "outputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1 × 1 × 1)", - "shortDescription" : "", - "shape" : "[1, 1, 1]", - "name" : "token_id", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 1 × 1)", - "shortDescription" : "", - "shape" : "[1, 1, 1]", - "name" : "token_prob", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1 × 1 × 1)", - "shortDescription" : "", - "shape" : "[1, 1, 1]", - "name" : "duration", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1 × 1 × 1 × 64)", - "shortDescription" : "", - "shape" : "[1, 1, 1, 64]", - "name" : "top_k_ids", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 1 × 1 × 64)", - "shortDescription" : "", - "shape" : "[1, 1, 1, 64]", - "name" : "top_k_logits", - "type" : "MultiArray" - } - ], - "storagePrecision" : "Float16", - "modelParameters" : [ - - ], - "author" : "Fluid Inference", - "specificationVersion" : 8, - "mlProgramOperationTypeHistogram" : { - "Ios17.reduceArgmax" : 2, - "Ios17.linear" : 3, - "Ios17.transpose" : 2, - "Ios17.sliceByIndex" : 2, - "Ios17.add" : 1, - "Ios17.topk" : 1, - "Ios16.relu" : 1, - "Ios16.softmax" : 1, - "Ios17.expandDims" : 3, - "Ios17.squeeze" : 1, - "Ios17.cast" : 6, - "Ios17.gatherAlongAxis" : 1 - }, - "computePrecision" : "Mixed (Float16, Float32, Int16, Int32, UInt16)", - "isUpdatable" : "0", - "stateSchema" : [ - - ], - "availability" : { - "macOS" : "14.0", - "tvOS" : "17.0", - "visionOS" : "1.0", - "watchOS" : "10.0", - "iOS" : "17.0", - "macCatalyst" : "17.0" - }, - "modelType" : { - "name" : "MLModelType_mlProgram" - }, - "inputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 512 × 1)", - "shortDescription" : "", - "shape" : "[1, 512, 1]", - "name" : "encoder_step", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 640 × 1)", - "shortDescription" : "", - "shape" : "[1, 640, 1]", - "name" : "decoder_step", - "type" : "MultiArray" - } - ], - "userDefinedMetadata" : { - "com.github.apple.coremltools.source_dialect" : "TorchScript", - "com.github.apple.coremltools.source" : "torch==2.9.0", - "com.github.apple.coremltools.version" : "8.3.0" - }, - "generatedClassName" : "parakeet_joint_decision_single_step", - "method" : "predict" - } -] \ No newline at end of file diff --git a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecisionSingleStep.mlmodelc/model.mil b/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecisionSingleStep.mlmodelc/model.mil deleted file mode 100644 index 42bd4ac5def601f37ece871ff09bd6242de025aa..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecisionSingleStep.mlmodelc/model.mil +++ /dev/null @@ -1,69 +0,0 @@ -program(1.0) -[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3510.2.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.9.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "8.3.0"}})] -{ - func main(tensor decoder_step, tensor encoder_step) { - tensor input_1_perm_0 = const()[name = tensor("input_1_perm_0"), val = tensor([0, 2, 1])]; - tensor encoder_step_to_fp16_dtype_0 = const()[name = tensor("encoder_step_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor input_3_perm_0 = const()[name = tensor("input_3_perm_0"), val = tensor([0, 2, 1])]; - tensor decoder_step_to_fp16_dtype_0 = const()[name = tensor("decoder_step_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor joint_module_enc_weight_to_fp16 = const()[name = tensor("joint_module_enc_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; - tensor joint_module_enc_bias_to_fp16 = const()[name = tensor("joint_module_enc_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(655488)))]; - tensor encoder_step_to_fp16 = cast(dtype = encoder_step_to_fp16_dtype_0, x = encoder_step)[name = tensor("cast_9")]; - tensor input_1_cast_fp16 = transpose(perm = input_1_perm_0, x = encoder_step_to_fp16)[name = tensor("transpose_1")]; - tensor linear_0_cast_fp16 = linear(bias = joint_module_enc_bias_to_fp16, weight = joint_module_enc_weight_to_fp16, x = input_1_cast_fp16)[name = tensor("linear_0_cast_fp16")]; - tensor joint_module_pred_weight_to_fp16 = const()[name = tensor("joint_module_pred_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(656832)))]; - tensor joint_module_pred_bias_to_fp16 = const()[name = tensor("joint_module_pred_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1476096)))]; - tensor decoder_step_to_fp16 = cast(dtype = decoder_step_to_fp16_dtype_0, x = decoder_step)[name = tensor("cast_8")]; - tensor input_3_cast_fp16 = transpose(perm = input_3_perm_0, x = decoder_step_to_fp16)[name = tensor("transpose_0")]; - tensor linear_1_cast_fp16 = linear(bias = joint_module_pred_bias_to_fp16, weight = joint_module_pred_weight_to_fp16, x = input_3_cast_fp16)[name = tensor("linear_1_cast_fp16")]; - tensor var_23_axes_0 = const()[name = tensor("op_23_axes_0"), val = tensor([2])]; - tensor var_23_cast_fp16 = expand_dims(axes = var_23_axes_0, x = linear_0_cast_fp16)[name = tensor("op_23_cast_fp16")]; - tensor var_24_axes_0 = const()[name = tensor("op_24_axes_0"), val = tensor([1])]; - tensor var_24_cast_fp16 = expand_dims(axes = var_24_axes_0, x = linear_1_cast_fp16)[name = tensor("op_24_cast_fp16")]; - tensor input_5_cast_fp16 = add(x = var_23_cast_fp16, y = var_24_cast_fp16)[name = tensor("input_5_cast_fp16")]; - tensor input_7_cast_fp16 = relu(x = input_5_cast_fp16)[name = tensor("input_7_cast_fp16")]; - tensor joint_module_joint_net_2_weight_to_fp16 = const()[name = tensor("joint_module_joint_net_2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1477440)))]; - tensor joint_module_joint_net_2_bias_to_fp16 = const()[name = tensor("joint_module_joint_net_2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2795904)))]; - tensor linear_2_cast_fp16 = linear(bias = joint_module_joint_net_2_bias_to_fp16, weight = joint_module_joint_net_2_weight_to_fp16, x = input_7_cast_fp16)[name = tensor("linear_2_cast_fp16")]; - tensor token_logits_begin_0 = const()[name = tensor("token_logits_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor token_logits_end_0 = const()[name = tensor("token_logits_end_0"), val = tensor([1, 1, 1, 1025])]; - tensor token_logits_end_mask_0 = const()[name = tensor("token_logits_end_mask_0"), val = tensor([true, true, true, false])]; - tensor token_logits_cast_fp16 = slice_by_index(begin = token_logits_begin_0, end = token_logits_end_0, end_mask = token_logits_end_mask_0, x = linear_2_cast_fp16)[name = tensor("token_logits_cast_fp16")]; - tensor duration_logits_begin_0 = const()[name = tensor("duration_logits_begin_0"), val = tensor([0, 0, 0, 1025])]; - tensor duration_logits_end_0 = const()[name = tensor("duration_logits_end_0"), val = tensor([1, 1, 1, 1030])]; - tensor duration_logits_end_mask_0 = const()[name = tensor("duration_logits_end_mask_0"), val = tensor([true, true, true, true])]; - tensor duration_logits_cast_fp16 = slice_by_index(begin = duration_logits_begin_0, end = duration_logits_end_0, end_mask = duration_logits_end_mask_0, x = linear_2_cast_fp16)[name = tensor("duration_logits_cast_fp16")]; - tensor var_43_axis_0 = const()[name = tensor("op_43_axis_0"), val = tensor(-1)]; - tensor var_43_keep_dims_0 = const()[name = tensor("op_43_keep_dims_0"), val = tensor(false)]; - tensor var_43_output_dtype_0 = const()[name = tensor("op_43_output_dtype_0"), val = tensor("int32")]; - tensor token_id = reduce_argmax(axis = var_43_axis_0, keep_dims = var_43_keep_dims_0, output_dtype = var_43_output_dtype_0, x = token_logits_cast_fp16)[name = tensor("op_43_cast_fp16")]; - tensor var_49 = const()[name = tensor("op_49"), val = tensor(-1)]; - tensor token_probs_all_cast_fp16 = softmax(axis = var_49, x = token_logits_cast_fp16)[name = tensor("token_probs_all_cast_fp16")]; - tensor var_58_axes_0 = const()[name = tensor("op_58_axes_0"), val = tensor([-1])]; - tensor var_58 = expand_dims(axes = var_58_axes_0, x = token_id)[name = tensor("op_58")]; - tensor var_59 = const()[name = tensor("op_59"), val = tensor(-1)]; - tensor var_61_validate_indices_0 = const()[name = tensor("op_61_validate_indices_0"), val = tensor(false)]; - tensor var_58_to_int16_dtype_0 = const()[name = tensor("op_58_to_int16_dtype_0"), val = tensor("int16")]; - tensor var_58_to_int16 = cast(dtype = var_58_to_int16_dtype_0, x = var_58)[name = tensor("cast_7")]; - tensor var_61_cast_fp16_cast_int16 = gather_along_axis(axis = var_59, indices = var_58_to_int16, validate_indices = var_61_validate_indices_0, x = token_probs_all_cast_fp16)[name = tensor("op_61_cast_fp16_cast_int16")]; - tensor var_63_axes_0 = const()[name = tensor("op_63_axes_0"), val = tensor([-1])]; - tensor var_63_cast_fp16 = squeeze(axes = var_63_axes_0, x = var_61_cast_fp16_cast_int16)[name = tensor("op_63_cast_fp16")]; - tensor var_63_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("op_63_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; - tensor var_66_axis_0 = const()[name = tensor("op_66_axis_0"), val = tensor(-1)]; - tensor var_66_keep_dims_0 = const()[name = tensor("op_66_keep_dims_0"), val = tensor(false)]; - tensor var_66_output_dtype_0 = const()[name = tensor("op_66_output_dtype_0"), val = tensor("int32")]; - tensor duration = reduce_argmax(axis = var_66_axis_0, keep_dims = var_66_keep_dims_0, output_dtype = var_66_output_dtype_0, x = duration_logits_cast_fp16)[name = tensor("op_66_cast_fp16")]; - tensor var_72 = const()[name = tensor("op_72"), val = tensor(64)]; - tensor var_76_axis_0 = const()[name = tensor("op_76_axis_0"), val = tensor(-1)]; - tensor var_76_ascending_0 = const()[name = tensor("op_76_ascending_0"), val = tensor(false)]; - tensor var_76_sort_0 = const()[name = tensor("op_76_sort_0"), val = tensor(true)]; - tensor var_76_return_indices_0 = const()[name = tensor("op_76_return_indices_0"), val = tensor(true)]; - tensor var_76_cast_fp16_cast_int16_output_indices_dtype_0 = const()[name = tensor("op_76_cast_fp16_cast_int16_output_indices_dtype_0"), val = tensor("uint16")]; - tensor var_76_cast_fp16_cast_int16_0, tensor var_76_cast_fp16_cast_int16_1 = topk(ascending = var_76_ascending_0, axis = var_76_axis_0, k = var_72, output_indices_dtype = var_76_cast_fp16_cast_int16_output_indices_dtype_0, return_indices = var_76_return_indices_0, sort = var_76_sort_0, x = token_logits_cast_fp16)[name = tensor("op_76_cast_fp16_cast_int16")]; - tensor var_76_cast_fp16_cast_int16_1_to_int32_dtype_0 = const()[name = tensor("op_76_cast_fp16_cast_int16_1_to_int32_dtype_0"), val = tensor("int32")]; - tensor var_76_cast_fp16_0_to_fp32_dtype_0 = const()[name = tensor("op_76_cast_fp16_0_to_fp32_dtype_0"), val = tensor("fp32")]; - tensor top_k_logits = cast(dtype = var_76_cast_fp16_0_to_fp32_dtype_0, x = var_76_cast_fp16_cast_int16_0)[name = tensor("cast_4")]; - tensor top_k_ids = cast(dtype = var_76_cast_fp16_cast_int16_1_to_int32_dtype_0, x = var_76_cast_fp16_cast_int16_1)[name = tensor("cast_5")]; - tensor token_prob = cast(dtype = var_63_cast_fp16_to_fp32_dtype_0, x = var_63_cast_fp16)[name = tensor("cast_6")]; - } -> (token_id, token_prob, duration, top_k_ids, top_k_logits); -} \ No newline at end of file diff --git a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecisionSingleStep.mlmodelc/weights/weight.bin b/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecisionSingleStep.mlmodelc/weights/weight.bin deleted file mode 100644 index 747ec9d24e2d3e008bc869d2c404bb3d2653325c..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecisionSingleStep.mlmodelc/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:b3f771cb65b190f1873e39629676ed79b65a8361522f451b37bdba8b1106e6ff -size 2798028 diff --git a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/Preprocessor.mlmodelc/analytics/coremldata.bin b/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/Preprocessor.mlmodelc/analytics/coremldata.bin deleted file mode 100644 index d34457003aea6648633a14e7cd6134edc9c30f17..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/Preprocessor.mlmodelc/analytics/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:a1ac15543fbb9301fba5f018b147e44d767479dec352aaa91dfe7bcf65949693 -size 243 diff --git a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/Preprocessor.mlmodelc/coremldata.bin b/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/Preprocessor.mlmodelc/coremldata.bin deleted file mode 100644 index cd81f47cc6672a441cbe9556757a30ffc0ff0bc7..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/Preprocessor.mlmodelc/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:4940877938cc1b6d8830bbdd68ac8a49377cc57d75b61308883da5235b6a1914 -size 439 diff --git a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/Preprocessor.mlmodelc/metadata.json b/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/Preprocessor.mlmodelc/metadata.json deleted file mode 100644 index 6087d5e76f7dd060cf4f135f1f3ae0bbbeb2f30a..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/Preprocessor.mlmodelc/metadata.json +++ /dev/null @@ -1,112 +0,0 @@ -[ - { - "metadataOutputVersion" : "3.0", - "shortDescription" : "Parakeet 110M preprocessor (15 s window)", - "outputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32)", - "shortDescription" : "", - "shape" : "[]", - "name" : "mel_features", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1)", - "shortDescription" : "", - "shape" : "[1]", - "name" : "mel_length", - "type" : "MultiArray" - } - ], - "storagePrecision" : "Float16", - "modelParameters" : [ - - ], - "author" : "Fluid Inference", - "specificationVersion" : 8, - "mlProgramOperationTypeHistogram" : { - "Range1d" : 3, - "Ios17.equal" : 1, - "Ios17.notEqual" : 1, - "Ios17.reshape" : 2, - "Identity" : 1, - "Ios17.matmul" : 1, - "Select" : 6, - "Ios17.expandDims" : 12, - "Ios17.add" : 3, - "Tile" : 2, - "Ios17.sliceByIndex" : 3, - "Ios16.reduceSum" : 4, - "Shape" : 4, - "Ios17.gather" : 4, - "Ios17.logicalNot" : 1, - "Pad" : 1, - "Ios17.log" : 1, - "Ios17.less" : 2, - "Ios17.sub" : 4, - "Ios17.conv" : 2, - "Ios17.pow" : 2, - "Ios17.cast" : 10, - "Ios17.concat" : 3, - "Stack" : 1, - "Ios17.floorDiv" : 1, - "Ios17.realDiv" : 4, - "Ios17.sqrt" : 1, - "Ios17.greaterEqual" : 1, - "Ios17.mul" : 1 - }, - "computePrecision" : "Mixed (Float16, Float32, Int16, Int32, UInt16)", - "isUpdatable" : "0", - "stateSchema" : [ - - ], - "availability" : { - "macOS" : "14.0", - "tvOS" : "17.0", - "visionOS" : "1.0", - "watchOS" : "10.0", - "iOS" : "17.0", - "macCatalyst" : "17.0" - }, - "modelType" : { - "name" : "MLModelType_mlProgram" - }, - "inputSchema" : [ - { - "dataType" : "Float32", - "hasShapeFlexibility" : "1", - "isOptional" : "0", - "shapeFlexibility" : "1 × 1...240000", - "shapeRange" : "[[1, 1], [1, 240000]]", - "formattedType" : "MultiArray (Float32 1 × 1)", - "type" : "MultiArray", - "shape" : "[1, 1]", - "name" : "audio", - "shortDescription" : "" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1)", - "shortDescription" : "", - "shape" : "[1]", - "name" : "audio_length", - "type" : "MultiArray" - } - ], - "userDefinedMetadata" : { - "com.github.apple.coremltools.source_dialect" : "TorchScript", - "com.github.apple.coremltools.source" : "torch==2.9.0", - "com.github.apple.coremltools.version" : "8.3.0" - }, - "generatedClassName" : "parakeet_preprocessor", - "method" : "predict" - } -] \ No newline at end of file diff --git a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/Preprocessor.mlmodelc/model.mil b/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/Preprocessor.mlmodelc/model.mil deleted file mode 100644 index ae325b6071a8ee5dad74a2e516e60889d89771d8..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/Preprocessor.mlmodelc/model.mil +++ /dev/null @@ -1,191 +0,0 @@ -program(1.0) -[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3510.2.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.9.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "8.3.0"}})] -{ - func main(tensor audio, tensor audio_length) [FlexibleShapeInformation = tuple, dict, tensor>>, tuple, dict, list, ?>>>>((("DefaultShapes", {{"audio", [1, 1]}}), ("RangeDims", {{"audio", [[1, 1], [1, 240000]]}})))] { - tensor var_9 = const()[name = tensor("op_9"), val = tensor(1)]; - tensor var_10 = const()[name = tensor("op_10"), val = tensor(160)]; - tensor var_12 = const()[name = tensor("op_12"), val = tensor(0)]; - tensor var_34 = const()[name = tensor("op_34"), val = tensor(512)]; - tensor var_35 = add(x = audio_length, y = var_34)[name = tensor("op_35")]; - tensor var_36 = const()[name = tensor("op_36"), val = tensor(512)]; - tensor var_37 = sub(x = var_35, y = var_36)[name = tensor("op_37")]; - tensor floor_div_0 = floor_div(x = var_37, y = var_10)[name = tensor("floor_div_0")]; - tensor var_40 = equal(x = audio_length, y = var_12)[name = tensor("op_40")]; - tensor var_41 = const()[name = tensor("op_41"), val = tensor([0])]; - tensor mel_length = select(a = var_41, b = floor_div_0, cond = var_40)[name = tensor("seq_len")]; - tensor audio_to_fp16_dtype_0 = const()[name = tensor("audio_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor audio_to_fp16 = cast(dtype = audio_to_fp16_dtype_0, x = audio)[name = tensor("cast_27")]; - tensor var_43_shape_cast_fp16 = shape(x = audio_to_fp16)[name = tensor("op_43_shape_cast_fp16")]; - tensor gather_0_axis_0 = const()[name = tensor("gather_0_axis_0"), val = tensor(0)]; - tensor gather_0_batch_dims_0 = const()[name = tensor("gather_0_batch_dims_0"), val = tensor(0)]; - tensor gather_0_validate_indices_0 = const()[name = tensor("gather_0_validate_indices_0"), val = tensor(false)]; - tensor var_43_shape_cast_fp16_to_int16_dtype_0 = const()[name = tensor("op_43_shape_cast_fp16_to_int16_dtype_0"), val = tensor("int16")]; - tensor select_0_to_uint16 = const()[name = tensor("select_0_to_uint16"), val = tensor(1)]; - tensor var_43_shape_cast_fp16_to_int16 = cast(dtype = var_43_shape_cast_fp16_to_int16_dtype_0, x = var_43_shape_cast_fp16)[name = tensor("cast_26")]; - tensor gather_0_cast_uint16 = gather(axis = gather_0_axis_0, batch_dims = gather_0_batch_dims_0, indices = select_0_to_uint16, validate_indices = gather_0_validate_indices_0, x = var_43_shape_cast_fp16_to_int16)[name = tensor("gather_0_cast_uint16")]; - tensor gather_0_cast_uint16_to_int32_dtype_0 = const()[name = tensor("gather_0_cast_uint16_to_int32_dtype_0"), val = tensor("int32")]; - tensor const_0 = const()[name = tensor("const_0"), val = tensor(0)]; - tensor const_1 = const()[name = tensor("const_1"), val = tensor(1)]; - tensor gather_0_cast_uint16_to_int32 = cast(dtype = gather_0_cast_uint16_to_int32_dtype_0, x = gather_0_cast_uint16)[name = tensor("cast_25")]; - tensor var_44 = range_1d(end = gather_0_cast_uint16_to_int32, start = const_0, step = const_1)[name = tensor("op_44")]; - tensor var_45_axes_0 = const()[name = tensor("op_45_axes_0"), val = tensor([0])]; - tensor var_45 = expand_dims(axes = var_45_axes_0, x = var_44)[name = tensor("op_45")]; - tensor var_46_axes_0 = const()[name = tensor("op_46_axes_0"), val = tensor([1])]; - tensor var_46 = expand_dims(axes = var_46_axes_0, x = audio_length)[name = tensor("op_46")]; - tensor timemask = less(x = var_45, y = var_46)[name = tensor("timemask")]; - tensor var_49_begin_0 = const()[name = tensor("op_49_begin_0"), val = tensor([0, 0])]; - tensor var_49_end_0 = const()[name = tensor("op_49_end_0"), val = tensor([1, 1])]; - tensor var_49_end_mask_0 = const()[name = tensor("op_49_end_mask_0"), val = tensor([true, false])]; - tensor var_49_squeeze_mask_0 = const()[name = tensor("op_49_squeeze_mask_0"), val = tensor([false, true])]; - tensor var_49_cast_fp16 = slice_by_index(begin = var_49_begin_0, end = var_49_end_0, end_mask = var_49_end_mask_0, squeeze_mask = var_49_squeeze_mask_0, x = audio_to_fp16)[name = tensor("op_49_cast_fp16")]; - tensor var_50_axes_0 = const()[name = tensor("op_50_axes_0"), val = tensor([1])]; - tensor var_50_cast_fp16 = expand_dims(axes = var_50_axes_0, x = var_49_cast_fp16)[name = tensor("op_50_cast_fp16")]; - tensor var_52_begin_0 = const()[name = tensor("op_52_begin_0"), val = tensor([0, 1])]; - tensor var_52_end_0 = const()[name = tensor("op_52_end_0"), val = tensor([1, 0])]; - tensor var_52_end_mask_0 = const()[name = tensor("op_52_end_mask_0"), val = tensor([true, true])]; - tensor var_52_cast_fp16 = slice_by_index(begin = var_52_begin_0, end = var_52_end_0, end_mask = var_52_end_mask_0, x = audio_to_fp16)[name = tensor("op_52_cast_fp16")]; - tensor var_54_begin_0 = const()[name = tensor("op_54_begin_0"), val = tensor([0, 0])]; - tensor var_54_end_0 = const()[name = tensor("op_54_end_0"), val = tensor([1, -1])]; - tensor var_54_end_mask_0 = const()[name = tensor("op_54_end_mask_0"), val = tensor([true, false])]; - tensor var_54_cast_fp16 = slice_by_index(begin = var_54_begin_0, end = var_54_end_0, end_mask = var_54_end_mask_0, x = audio_to_fp16)[name = tensor("op_54_cast_fp16")]; - tensor var_55_to_fp16 = const()[name = tensor("op_55_to_fp16"), val = tensor(0x1.f0cp-1)]; - tensor var_56_cast_fp16 = mul(x = var_54_cast_fp16, y = var_55_to_fp16)[name = tensor("op_56_cast_fp16")]; - tensor var_57_cast_fp16 = sub(x = var_52_cast_fp16, y = var_56_cast_fp16)[name = tensor("op_57_cast_fp16")]; - tensor x_3_interleave_0 = const()[name = tensor("x_3_interleave_0"), val = tensor(false)]; - tensor x_3_cast_fp16 = concat(axis = var_9, interleave = x_3_interleave_0, values = (var_50_cast_fp16, var_57_cast_fp16))[name = tensor("x_3_cast_fp16")]; - tensor var_60 = logical_not(x = timemask)[name = tensor("op_60")]; - tensor var_16_to_fp16 = const()[name = tensor("op_16_to_fp16"), val = tensor(0x0p+0)]; - tensor input_1_cast_fp16 = select(a = var_16_to_fp16, b = x_3_cast_fp16, cond = var_60)[name = tensor("input_1_cast_fp16")]; - tensor concat_1x = const()[name = tensor("concat_1x"), val = tensor([1, 1, -1])]; - tensor input_3_cast_fp16 = reshape(shape = concat_1x, x = input_1_cast_fp16)[name = tensor("input_3_cast_fp16")]; - tensor input_5_pad_0 = const()[name = tensor("input_5_pad_0"), val = tensor([0, 0, 0, 0, 256, 256])]; - tensor input_5_mode_0 = const()[name = tensor("input_5_mode_0"), val = tensor("constant")]; - tensor const_3_to_fp16 = const()[name = tensor("const_3_to_fp16"), val = tensor(0x0p+0)]; - tensor input_5_cast_fp16 = pad(constant_val = const_3_to_fp16, mode = input_5_mode_0, pad = input_5_pad_0, x = input_3_cast_fp16)[name = tensor("input_5_cast_fp16")]; - tensor concat_2x = const()[name = tensor("concat_2x"), val = tensor([1, -1])]; - tensor input_cast_fp16 = reshape(shape = concat_2x, x = input_5_cast_fp16)[name = tensor("input_cast_fp16")]; - tensor expand_dims_3 = const()[name = tensor("expand_dims_3"), val = tensor([160])]; - tensor expand_dims_4_axes_0 = const()[name = tensor("expand_dims_4_axes_0"), val = tensor([1])]; - tensor expand_dims_4_cast_fp16 = expand_dims(axes = expand_dims_4_axes_0, x = input_cast_fp16)[name = tensor("expand_dims_4_cast_fp16")]; - tensor conv_0_pad_type_0 = const()[name = tensor("conv_0_pad_type_0"), val = tensor("valid")]; - tensor conv_0_pad_0 = const()[name = tensor("conv_0_pad_0"), val = tensor([0, 0])]; - tensor conv_0_dilations_0 = const()[name = tensor("conv_0_dilations_0"), val = tensor([1])]; - tensor conv_0_groups_0 = const()[name = tensor("conv_0_groups_0"), val = tensor(1)]; - tensor expand_dims_1_to_fp16 = const()[name = tensor("expand_dims_1_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; - tensor conv_0_cast_fp16 = conv(dilations = conv_0_dilations_0, groups = conv_0_groups_0, pad = conv_0_pad_0, pad_type = conv_0_pad_type_0, strides = expand_dims_3, weight = expand_dims_1_to_fp16, x = expand_dims_4_cast_fp16)[name = tensor("conv_0_cast_fp16")]; - tensor conv_1_pad_type_0 = const()[name = tensor("conv_1_pad_type_0"), val = tensor("valid")]; - tensor conv_1_pad_0 = const()[name = tensor("conv_1_pad_0"), val = tensor([0, 0])]; - tensor conv_1_dilations_0 = const()[name = tensor("conv_1_dilations_0"), val = tensor([1])]; - tensor conv_1_groups_0 = const()[name = tensor("conv_1_groups_0"), val = tensor(1)]; - tensor expand_dims_2_to_fp16 = const()[name = tensor("expand_dims_2_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(263296)))]; - tensor conv_1_cast_fp16 = conv(dilations = conv_1_dilations_0, groups = conv_1_groups_0, pad = conv_1_pad_0, pad_type = conv_1_pad_type_0, strides = expand_dims_3, weight = expand_dims_2_to_fp16, x = expand_dims_4_cast_fp16)[name = tensor("conv_1_cast_fp16")]; - tensor stack_0_axis_0 = const()[name = tensor("stack_0_axis_0"), val = tensor(-1)]; - tensor stack_0_cast_fp16 = stack(axis = stack_0_axis_0, values = (conv_0_cast_fp16, conv_1_cast_fp16))[name = tensor("stack_0_cast_fp16")]; - tensor var_19_promoted_to_fp16 = const()[name = tensor("op_19_promoted_to_fp16"), val = tensor(0x1p+1)]; - tensor var_75_cast_fp16 = pow(x = stack_0_cast_fp16, y = var_19_promoted_to_fp16)[name = tensor("op_75_cast_fp16")]; - tensor var_77_axes_0 = const()[name = tensor("op_77_axes_0"), val = tensor([-1])]; - tensor var_77_keep_dims_0 = const()[name = tensor("op_77_keep_dims_0"), val = tensor(false)]; - tensor var_77_cast_fp16 = reduce_sum(axes = var_77_axes_0, keep_dims = var_77_keep_dims_0, x = var_75_cast_fp16)[name = tensor("op_77_cast_fp16")]; - tensor x_11_cast_fp16 = identity(x = var_77_cast_fp16)[name = tensor("x_11_cast_fp16")]; - tensor x_13_transpose_x_0 = const()[name = tensor("x_13_transpose_x_0"), val = tensor(false)]; - tensor x_13_transpose_y_0 = const()[name = tensor("x_13_transpose_y_0"), val = tensor(false)]; - tensor const_4_to_fp16 = const()[name = tensor("const_4_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(526528)))]; - tensor x_13_cast_fp16 = matmul(transpose_x = x_13_transpose_x_0, transpose_y = x_13_transpose_y_0, x = const_4_to_fp16, y = x_11_cast_fp16)[name = tensor("x_13_cast_fp16")]; - tensor var_84_to_fp16 = const()[name = tensor("op_84_to_fp16"), val = tensor(0x1p-24)]; - tensor var_85_cast_fp16 = add(x = x_13_cast_fp16, y = var_84_to_fp16)[name = tensor("op_85_cast_fp16")]; - tensor x_15_epsilon_0 = const()[name = tensor("x_15_epsilon_0"), val = tensor(0x1p-149)]; - tensor x_15_cast_fp16 = log(epsilon = x_15_epsilon_0, x = var_85_cast_fp16)[name = tensor("x_15_cast_fp16")]; - tensor var_87_shape_cast_fp16 = shape(x = x_15_cast_fp16)[name = tensor("op_87_shape_cast_fp16")]; - tensor gather_5 = const()[name = tensor("gather_5"), val = tensor(1)]; - tensor gather_6_axis_0 = const()[name = tensor("gather_6_axis_0"), val = tensor(0)]; - tensor gather_6_batch_dims_0 = const()[name = tensor("gather_6_batch_dims_0"), val = tensor(0)]; - tensor gather_6_validate_indices_0 = const()[name = tensor("gather_6_validate_indices_0"), val = tensor(false)]; - tensor var_87_shape_cast_fp16_to_uint16_dtype_0 = const()[name = tensor("op_87_shape_cast_fp16_to_uint16_dtype_0"), val = tensor("uint16")]; - tensor select_6_to_uint16 = const()[name = tensor("select_6_to_uint16"), val = tensor(2)]; - tensor var_87_shape_cast_fp16_to_uint16 = cast(dtype = var_87_shape_cast_fp16_to_uint16_dtype_0, x = var_87_shape_cast_fp16)[name = tensor("cast_24")]; - tensor gather_6_cast_uint16 = gather(axis = gather_6_axis_0, batch_dims = gather_6_batch_dims_0, indices = select_6_to_uint16, validate_indices = gather_6_validate_indices_0, x = var_87_shape_cast_fp16_to_uint16)[name = tensor("gather_6_cast_uint16")]; - tensor gather_6_cast_uint16_to_int32_dtype_0 = const()[name = tensor("gather_6_cast_uint16_to_int32_dtype_0"), val = tensor("int32")]; - tensor const_5 = const()[name = tensor("const_5"), val = tensor(0)]; - tensor const_6 = const()[name = tensor("const_6"), val = tensor(1)]; - tensor gather_6_cast_uint16_to_int32 = cast(dtype = gather_6_cast_uint16_to_int32_dtype_0, x = gather_6_cast_uint16)[name = tensor("cast_23")]; - tensor var_89 = range_1d(end = gather_6_cast_uint16_to_int32, start = const_5, step = const_6)[name = tensor("op_89")]; - tensor var_90_axes_0 = const()[name = tensor("op_90_axes_0"), val = tensor([0])]; - tensor var_90 = expand_dims(axes = var_90_axes_0, x = var_89)[name = tensor("op_90")]; - tensor concat_3_axis_0 = const()[name = tensor("concat_3_axis_0"), val = tensor(0)]; - tensor concat_3_interleave_0 = const()[name = tensor("concat_3_interleave_0"), val = tensor(false)]; - tensor concat_3 = concat(axis = concat_3_axis_0, interleave = concat_3_interleave_0, values = (gather_5, gather_6_cast_uint16_to_int32))[name = tensor("concat_3")]; - tensor shape_8 = shape(x = var_90)[name = tensor("shape_8")]; - tensor real_div_0 = real_div(x = concat_3, y = shape_8)[name = tensor("real_div_0")]; - tensor time_steps = tile(reps = real_div_0, x = var_90)[name = tensor("time_steps")]; - tensor var_93_axes_0 = const()[name = tensor("op_93_axes_0"), val = tensor([1])]; - tensor var_93 = expand_dims(axes = var_93_axes_0, x = mel_length)[name = tensor("op_93")]; - tensor valid_mask = less(x = time_steps, y = var_93)[name = tensor("valid_mask")]; - tensor var_95_axes_0 = const()[name = tensor("op_95_axes_0"), val = tensor([1])]; - tensor var_95 = expand_dims(axes = var_95_axes_0, x = valid_mask)[name = tensor("op_95")]; - tensor var_96_cast_fp16 = select(a = x_15_cast_fp16, b = var_16_to_fp16, cond = var_95)[name = tensor("op_96_cast_fp16")]; - tensor x_mean_numerator_axes_0 = const()[name = tensor("x_mean_numerator_axes_0"), val = tensor([2])]; - tensor x_mean_numerator_keep_dims_0 = const()[name = tensor("x_mean_numerator_keep_dims_0"), val = tensor(false)]; - tensor x_mean_numerator_cast_fp16 = reduce_sum(axes = x_mean_numerator_axes_0, keep_dims = x_mean_numerator_keep_dims_0, x = var_96_cast_fp16)[name = tensor("x_mean_numerator_cast_fp16")]; - tensor x_mean_denominator_axes_0 = const()[name = tensor("x_mean_denominator_axes_0"), val = tensor([1])]; - tensor x_mean_denominator_keep_dims_0 = const()[name = tensor("x_mean_denominator_keep_dims_0"), val = tensor(false)]; - tensor cast_6_to_fp16_dtype_0 = const()[name = tensor("cast_6_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor valid_mask_to_fp16 = cast(dtype = cast_6_to_fp16_dtype_0, x = valid_mask)[name = tensor("cast_22")]; - tensor x_mean_denominator_cast_fp16 = reduce_sum(axes = x_mean_denominator_axes_0, keep_dims = x_mean_denominator_keep_dims_0, x = valid_mask_to_fp16)[name = tensor("x_mean_denominator_cast_fp16")]; - tensor var_101_axes_0 = const()[name = tensor("op_101_axes_0"), val = tensor([1])]; - tensor var_101_cast_fp16 = expand_dims(axes = var_101_axes_0, x = x_mean_denominator_cast_fp16)[name = tensor("op_101_cast_fp16")]; - tensor x_mean_cast_fp16 = real_div(x = x_mean_numerator_cast_fp16, y = var_101_cast_fp16)[name = tensor("x_mean_cast_fp16")]; - tensor var_104_axes_0 = const()[name = tensor("op_104_axes_0"), val = tensor([2])]; - tensor var_104_cast_fp16 = expand_dims(axes = var_104_axes_0, x = x_mean_cast_fp16)[name = tensor("op_104_cast_fp16")]; - tensor var_105_cast_fp16 = sub(x = x_15_cast_fp16, y = var_104_cast_fp16)[name = tensor("op_105_cast_fp16")]; - tensor var_106_cast_fp16 = select(a = var_105_cast_fp16, b = var_16_to_fp16, cond = var_95)[name = tensor("op_106_cast_fp16")]; - tensor var_19_promoted_1_to_fp16 = const()[name = tensor("op_19_promoted_1_to_fp16"), val = tensor(0x1p+1)]; - tensor var_107_cast_fp16 = pow(x = var_106_cast_fp16, y = var_19_promoted_1_to_fp16)[name = tensor("op_107_cast_fp16")]; - tensor var_109_axes_0 = const()[name = tensor("op_109_axes_0"), val = tensor([2])]; - tensor var_109_keep_dims_0 = const()[name = tensor("op_109_keep_dims_0"), val = tensor(false)]; - tensor var_109_cast_fp16 = reduce_sum(axes = var_109_axes_0, keep_dims = var_109_keep_dims_0, x = var_107_cast_fp16)[name = tensor("op_109_cast_fp16")]; - tensor var_111_to_fp16 = const()[name = tensor("op_111_to_fp16"), val = tensor(0x1p+0)]; - tensor var_112_cast_fp16 = sub(x = var_101_cast_fp16, y = var_111_to_fp16)[name = tensor("op_112_cast_fp16")]; - tensor var_113_cast_fp16 = real_div(x = var_109_cast_fp16, y = var_112_cast_fp16)[name = tensor("op_113_cast_fp16")]; - tensor x_std_1_cast_fp16 = sqrt(x = var_113_cast_fp16)[name = tensor("x_std_1_cast_fp16")]; - tensor var_115_cast_fp16 = not_equal(x = x_std_1_cast_fp16, y = x_std_1_cast_fp16)[name = tensor("op_115_cast_fp16")]; - tensor x_std_3_cast_fp16 = select(a = var_16_to_fp16, b = x_std_1_cast_fp16, cond = var_115_cast_fp16)[name = tensor("x_std_3_cast_fp16")]; - tensor var_25_to_fp16 = const()[name = tensor("op_25_to_fp16"), val = tensor(0x1.5p-17)]; - tensor x_std_cast_fp16 = add(x = x_std_3_cast_fp16, y = var_25_to_fp16)[name = tensor("x_std_cast_fp16")]; - tensor var_120_axes_0 = const()[name = tensor("op_120_axes_0"), val = tensor([2])]; - tensor var_120_cast_fp16 = expand_dims(axes = var_120_axes_0, x = x_std_cast_fp16)[name = tensor("op_120_cast_fp16")]; - tensor x_cast_fp16 = real_div(x = var_105_cast_fp16, y = var_120_cast_fp16)[name = tensor("x_cast_fp16")]; - tensor var_122_shape_cast_fp16 = shape(x = x_cast_fp16)[name = tensor("op_122_shape_cast_fp16")]; - tensor gather_7_axis_0 = const()[name = tensor("gather_7_axis_0"), val = tensor(0)]; - tensor gather_7_batch_dims_0 = const()[name = tensor("gather_7_batch_dims_0"), val = tensor(0)]; - tensor gather_7_validate_indices_0 = const()[name = tensor("gather_7_validate_indices_0"), val = tensor(false)]; - tensor var_122_shape_cast_fp16_to_uint16_dtype_0 = const()[name = tensor("op_122_shape_cast_fp16_to_uint16_dtype_0"), val = tensor("uint16")]; - tensor select_7_to_uint16 = const()[name = tensor("select_7_to_uint16"), val = tensor(2)]; - tensor var_122_shape_cast_fp16_to_uint16 = cast(dtype = var_122_shape_cast_fp16_to_uint16_dtype_0, x = var_122_shape_cast_fp16)[name = tensor("cast_21")]; - tensor gather_7_cast_uint16 = gather(axis = gather_7_axis_0, batch_dims = gather_7_batch_dims_0, indices = select_7_to_uint16, validate_indices = gather_7_validate_indices_0, x = var_122_shape_cast_fp16_to_uint16)[name = tensor("gather_7_cast_uint16")]; - tensor gather_7_cast_uint16_to_int32_dtype_0 = const()[name = tensor("gather_7_cast_uint16_to_int32_dtype_0"), val = tensor("int32")]; - tensor const_7 = const()[name = tensor("const_7"), val = tensor(0)]; - tensor const_8 = const()[name = tensor("const_8"), val = tensor(1)]; - tensor gather_7_cast_uint16_to_int32 = cast(dtype = gather_7_cast_uint16_to_int32_dtype_0, x = gather_7_cast_uint16)[name = tensor("cast_20")]; - tensor mask_1 = range_1d(end = gather_7_cast_uint16_to_int32, start = const_7, step = const_8)[name = tensor("mask_1")]; - tensor gather_8_axis_0 = const()[name = tensor("gather_8_axis_0"), val = tensor(0)]; - tensor gather_8_batch_dims_0 = const()[name = tensor("gather_8_batch_dims_0"), val = tensor(0)]; - tensor gather_8_validate_indices_0 = const()[name = tensor("gather_8_validate_indices_0"), val = tensor(false)]; - tensor select_8_to_uint16 = const()[name = tensor("select_8_to_uint16"), val = tensor(0)]; - tensor gather_8_cast_uint16 = gather(axis = gather_8_axis_0, batch_dims = gather_8_batch_dims_0, indices = select_8_to_uint16, validate_indices = gather_8_validate_indices_0, x = var_122_shape_cast_fp16_to_uint16)[name = tensor("gather_8_cast_uint16")]; - tensor gather_8_cast_uint16_to_int32_dtype_0 = const()[name = tensor("gather_8_cast_uint16_to_int32_dtype_0"), val = tensor("int32")]; - tensor concat_4_axis_0 = const()[name = tensor("concat_4_axis_0"), val = tensor(0)]; - tensor concat_4_interleave_0 = const()[name = tensor("concat_4_interleave_0"), val = tensor(false)]; - tensor gather_8_cast_uint16_to_int32 = cast(dtype = gather_8_cast_uint16_to_int32_dtype_0, x = gather_8_cast_uint16)[name = tensor("cast_19")]; - tensor concat_4 = concat(axis = concat_4_axis_0, interleave = concat_4_interleave_0, values = (gather_8_cast_uint16_to_int32, var_9))[name = tensor("concat_4")]; - tensor expand_dims_0_axes_0 = const()[name = tensor("expand_dims_0_axes_0"), val = tensor([0])]; - tensor expand_dims_0 = expand_dims(axes = expand_dims_0_axes_0, x = mask_1)[name = tensor("expand_dims_0")]; - tensor var_126 = tile(reps = concat_4, x = expand_dims_0)[name = tensor("op_126")]; - tensor mask = greater_equal(x = var_126, y = var_93)[name = tensor("mask")]; - tensor var_129_axes_0 = const()[name = tensor("op_129_axes_0"), val = tensor([1])]; - tensor var_129 = expand_dims(axes = var_129_axes_0, x = mask)[name = tensor("op_129")]; - tensor cast_15_to_fp16 = const()[name = tensor("cast_15_to_fp16"), val = tensor(0x0p+0)]; - tensor processed_signal_cast_fp16 = select(a = cast_15_to_fp16, b = x_cast_fp16, cond = var_129)[name = tensor("processed_signal_cast_fp16")]; - tensor processed_signal_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("processed_signal_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; - tensor mel_features = cast(dtype = processed_signal_cast_fp16_to_fp32_dtype_0, x = processed_signal_cast_fp16)[name = tensor("cast_18")]; - } -> (mel_features, mel_length); -} \ No newline at end of file diff --git a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/Preprocessor.mlmodelc/weights/weight.bin b/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/Preprocessor.mlmodelc/weights/weight.bin deleted file mode 100644 index d0c6ac87e5f78315662149af9ab4ea3658497dbe..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/Preprocessor.mlmodelc/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:c062338de852a26607ce4101f74e6895de3a4134a57b07232bd72bfc6f1d7f1a -size 567712 diff --git a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/metadata.json b/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/metadata.json deleted file mode 100644 index 6db872ae8a7d463c95dc180210ace0fce019f4a6..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/metadata.json +++ /dev/null @@ -1,247 +0,0 @@ -{ - "model_id": "nvidia/parakeet-tdt_ctc-110m", - "model_type": "hybrid_rnnt_ctc", - "sample_rate": 16000, - "max_audio_seconds": 15.0, - "max_audio_samples": 240000, - "max_symbol_steps": 1, - "vocab_size": 1024, - "joint_extra_outputs": 5, - "encoder_dim": 512, - "decoder_dim": 640, - "decoder_hidden": 640, - "decoder_layers": 1, - "blank_id": 1024, - "checkpoint": { - "type": "pretrained", - "model_id": "nvidia/parakeet-tdt_ctc-110m" - }, - "coreml": { - "compute_units": "CPU_ONLY", - "compute_precision": "FLOAT32" - }, - "components": { - "preprocessor": { - "inputs": { - "audio_signal": [ - 1, - 240000 - ], - "audio_length": [ - 1 - ] - }, - "outputs": { - "mel": [ - 1, - 80, - 1501 - ], - "mel_length": [ - 1 - ] - }, - "path": "parakeet_preprocessor.mlpackage" - }, - "encoder": { - "inputs": { - "mel": [ - 1, - 80, - 1501 - ], - "mel_length": [ - 1 - ] - }, - "outputs": { - "encoder": [ - 1, - 512, - 188 - ], - "encoder_length": [ - 1 - ] - }, - "path": "parakeet_encoder.mlpackage" - }, - "ctc_head": { - "inputs": { - "encoder": [ - 1, - 512, - 188 - ] - }, - "outputs": { - "log_probs": [ - 1, - 188, - 1025 - ] - }, - "path": "parakeet_ctc_head.mlpackage" - }, - "mel_encoder": { - "inputs": { - "audio_signal": [ - 1, - 240000 - ], - "audio_length": [ - 1 - ] - }, - "outputs": { - "encoder": [ - 1, - 512, - 188 - ], - "encoder_length": [ - 1 - ] - }, - "path": "parakeet_mel_encoder.mlpackage" - }, - "decoder": { - "inputs": { - "targets": [ - 1, - 1 - ], - "target_length": [ - 1 - ], - "h_in": [ - 1, - 1, - 640 - ], - "c_in": [ - 1, - 1, - 640 - ] - }, - "outputs": { - "decoder": [ - 1, - 640, - 1 - ], - "h_out": [ - 1, - 1, - 640 - ], - "c_out": [ - 1, - 1, - 640 - ] - }, - "path": "parakeet_decoder.mlpackage" - }, - "joint": { - "inputs": { - "encoder": [ - 1, - 512, - 188 - ], - "decoder": [ - 1, - 640, - 1 - ] - }, - "outputs": { - "logits": [ - 1, - 188, - 1, - 1030 - ] - }, - "path": "parakeet_joint.mlpackage" - }, - "joint_decision": { - "inputs": { - "encoder": [ - 1, - 512, - 188 - ], - "decoder": [ - 1, - 640, - 1 - ] - }, - "outputs": { - "token_id": [ - 1, - 188, - 1 - ], - "token_prob": [ - 1, - 188, - 1 - ], - "duration": [ - 1, - 188, - 1 - ] - }, - "path": "parakeet_joint_decision.mlpackage" - }, - "joint_decision_single_step": { - "inputs": { - "encoder_step": [ - 1, - 512, - 1 - ], - "decoder_step": [ - 1, - 640, - 1 - ] - }, - "outputs": { - "token_id": [ - 1, - 1, - 1 - ], - "token_prob": [ - 1, - 1, - 1 - ], - "duration": [ - 1, - 1, - 1 - ], - "top_k_ids": [ - 1, - 1, - 1, - 64 - ], - "top_k_logits": [ - 1, - 1, - 1, - 64 - ] - }, - "path": "parakeet_joint_decision_single_step.mlpackage" - } - } -} \ No newline at end of file diff --git a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/vocab.json b/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/vocab.json deleted file mode 100644 index 33b95831722563db5362641673b37885e25fb47a..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/vocab.json +++ /dev/null @@ -1 +0,0 @@ -{"0": "", "1": "▁t", "2": "▁th", "3": "▁a", "4": "in", "5": "re", "6": "▁the", "7": "▁w", "8": "▁s", "9": "▁o", "10": "er", "11": "ou", "12": "at", "13": "nd", "14": "it", "15": "▁h", "16": "▁c", "17": "▁b", "18": "is", "19": "en", "20": "on", "21": "ing", "22": "▁f", "23": "▁to", "24": "▁m", "25": "es", "26": "▁p", "27": "or", "28": "an", "29": "▁d", "30": "ll", "31": "▁I", "32": "ed", "33": "▁and", "34": "▁l", "35": "▁of", "36": "▁in", "37": "▁y", "38": "ar", "39": "▁g", "40": "▁you", "41": "as", "42": "om", "43": "▁n", "44": "ve", "45": "▁that", "46": "le", "47": "ic", "48": "us", "49": "ow", "50": "et", "51": "al", "52": "▁e", "53": "ut", "54": "▁it", "55": "ot", "56": "▁be", "57": "▁T", "58": "ion", "59": "▁is", "60": "▁wh", "61": "▁re", "62": "▁on", "63": "▁we", "64": "ent", "65": "▁A", "66": "ay", "67": "▁ha", "68": "▁Th", "69": "id", "70": "▁S", "71": "ac", "72": "gh", "73": "ver", "74": "ke", "75": "▁for", "76": "im", "77": "ly", "78": "ur", "79": "ld", "80": "▁he", "81": "▁st", "82": "all", "83": "ro", "84": "st", "85": "se", "86": "ct", "87": "ith", "88": "ir", "89": "am", "90": "▁this", "91": "if", "92": "▁W", "93": "oo", "94": "ri", "95": "▁was", "96": "ght", "97": "▁u", "98": "▁with", "99": "ad", "100": "ch", "101": "▁se", "102": "▁k", "103": "▁an", "104": "▁The", "105": "▁li", "106": "▁do", "107": "▁B", "108": "▁have", "109": "▁as", "110": "th", "111": "▁are", "112": "▁sh", "113": "ust", "114": "ce", "115": "ally", "116": "ill", "117": "▁H", "118": "▁j", "119": "ter", "120": "▁go", "121": "▁And", "122": "ation", "123": "▁C", "124": "▁so", "125": "ome", "126": "▁not", "127": "op", "128": "il", "129": "ore", "130": "▁ne", "131": "▁can", "132": "▁me", "133": "▁at", "134": "ould", "135": "ant", "136": "▁M", "137": "▁like", "138": "ere", "139": "▁they", "140": "ra", "141": "ers", "142": "▁ab", "143": "▁de", "144": "▁kn", "145": "ge", "146": "▁Y", "147": "▁ch", "148": "ul", "149": "pp", "150": "▁or", "151": "▁al", "152": "▁con", "153": "▁com", "154": "ess", "155": "▁su", "156": "out", "157": "▁your", "158": "▁So", "159": "ate", "160": "▁one", "161": "▁all", "162": "▁ex", "163": "est", "164": "▁fr", "165": "▁just", "166": "▁pro", "167": "▁know", "168": "▁O", "169": "ain", "170": "▁but", "171": "ol", "172": "ive", "173": "▁v", "174": "use", "175": "very", "176": "art", "177": "qu", "178": "▁my", "179": "el", "180": "▁N", "181": "nt", "182": "▁It", "183": "▁what", "184": "ab", "185": "▁P", "186": "▁wor", "187": "▁out", "188": "▁there", "189": "▁up", "190": "um", "191": "▁from", "192": "pe", "193": "▁tw", "194": "▁r", "195": "and", "196": "ight", "197": "ort", "198": "un", "199": "▁L", "200": "ist", "201": "▁about", "202": "ide", "203": "ig", "204": "ake", "205": "▁D", "206": "em", "207": "os", "208": "king", "209": "rou", "210": "ind", "211": "our", "212": "res", "213": "▁We", "214": "▁get", "215": "▁E", "216": "▁G", "217": "ack", "218": "▁le", "219": "ity", "220": "od", "221": "▁F", "222": "ard", "223": "▁pl", "224": "▁our", "225": "▁int", "226": "ment", "227": "▁will", "228": "ies", "229": "▁by", "230": "ink", "231": "ca", "232": "▁if", "233": "red", "234": "her", "235": "ie", "236": "▁us", "237": "▁some", "238": "▁don", "239": "ven", "240": "ood", "241": "ast", "242": "▁R", "243": "▁his", "244": "▁tim", "245": "▁tr", "246": "▁more", "247": "ich", "248": "ous", "249": "ame", "250": "▁going", "251": "▁had", "252": "▁them", "253": "ook", "254": "▁pe", "255": "▁Wh", "256": "▁You", "257": "▁But", "258": "ine", "259": "▁here", "260": "▁would", "261": "cause", "262": "right", "263": "so", "264": "ost", "265": "ure", "266": "▁has", "267": "ect", "268": "▁think", "269": "▁fe", "270": "ong", "271": "▁see", "272": "▁when", "273": "▁who", "274": "▁were", "275": "▁really", "276": "▁their", "277": "▁want", "278": "one", "279": "ople", "280": "▁then", "281": "▁time", "282": "▁sa", "283": "ap", "284": "▁te", "285": "▁He", "286": "▁ye", "287": "ck", "288": "▁her", "289": "▁thing", "290": "▁right", "291": "▁which", "292": "itt", "293": "ice", "294": "act", "295": "▁people", "296": "ty", "297": "▁two", "298": "▁J", "299": "▁im", "300": "ther", "301": "ci", "302": "ose", "303": "▁cl", "304": "▁qu", "305": "▁man", "306": "▁also", "307": "ree", "308": "▁en", "309": "ud", "310": "▁how", "311": "reat", "312": "ak", "313": "hing", "314": "ag", "315": "▁any", "316": "ff", "317": "ace", "318": "per", "319": "▁because", "320": "▁very", "321": "own", "322": "▁ad", "323": "▁act", "324": "▁been", "325": "▁now", "326": "▁ag", "327": "▁into", "328": "▁comp", "329": "ars", "330": "ions", "331": "are", "332": "ite", "333": "iv", "334": "▁these", "335": "ays", "336": "ep", "337": "▁This", "338": "▁she", "339": "ans", "340": "ah", "341": "een", "342": "▁over", "343": "ry", "344": "▁lo", "345": "age", "346": "▁pr", "347": "▁sp", "348": "ue", "349": "▁co", "350": "ick", "351": "ber", "352": "▁did", "353": "ip", "354": "ach", "355": "▁back", "356": "▁no", "357": "▁cont", "358": "▁other", "359": "▁every", "360": "pt", "361": "▁need", "362": "▁him", "363": "▁U", "364": "▁In", "365": "▁work", "366": "irst", "367": "▁part", "368": "▁look", "369": "ittle", "370": "ble", "371": "iz", "372": "▁un", "373": "▁make", "374": "omet", "375": "nder", "376": "ish", "377": "na", "378": "▁little", "379": "▁off", "380": "▁than", "381": "▁got", "382": "ually", "383": "▁per", "384": "▁good", "385": "▁way", "386": "▁could", "387": "▁ac", "388": "▁imp", "389": "able", "390": "▁where", "391": "iff", "392": "▁That", "393": "▁res", "394": "ount", "395": "pl", "396": "ance", "397": "▁first", "398": "▁ro", "399": "▁pre", "400": "ass", "401": "▁say", "402": "int", "403": "ated", "404": "ire", "405": "uch", "406": "ase", "407": "▁somet", "408": "ound", "409": "▁down", "410": "▁diff", "411": "sel", "412": "▁gu", "413": "▁am", "414": "ress", "415": "▁lot", "416": "ence", "417": "▁dis", "418": "orm", "419": "ix", "420": "▁po", "421": "ving", "422": "enty", "423": "▁K", "424": "▁spe", "425": "und", "426": "he", "427": "▁much", "428": "▁ar", "429": "round", "430": "▁app", "431": "co", "432": "ark", "433": "▁new", "434": "ater", "435": "ult", "436": "end", "437": "▁even", "438": "▁start", "439": "ations", "440": "rough", "441": "ile", "442": "fter", "443": "▁well", "444": "be", "445": "▁They", "446": "▁three", "447": "ign", "448": "ild", "449": "▁said", "450": "ough", "451": "ang", "452": "▁too", "453": "ade", "454": "▁bl", "455": "ens", "456": "▁inc", "457": "ia", "458": "▁those", "459": "▁mo", "460": "▁take", "461": "▁through", "462": "▁fl", "463": "▁kind", "464": "▁things", "465": "▁bet", "466": "▁only", "467": "▁St", "468": "▁let", "469": "cess", "470": "▁Ch", "471": "ary", "472": "vel", "473": "▁If", "474": "xt", "475": "other", "476": "av", "477": "ical", "478": "ord", "479": "▁again", "480": "▁something", "481": "onna", "482": "fore", "483": "▁may", "484": "ting", "485": "▁bu", "486": "▁differe", "487": "urn", "488": "▁gonna", "489": "▁does", "490": "uct", "491": "og", "492": "▁twenty", "493": "▁gr", "494": "▁Ye", "495": "wn", "496": "▁should", "497": "▁comm", "498": "ition", "499": "▁under", "500": "▁hel", "501": "ory", "502": "▁fo", "503": "▁use", "504": "igh", "505": "ife", "506": "▁actually", "507": "▁tal", "508": "▁call", "509": "ents", "510": "ious", "511": "ull", "512": "▁There", "513": "▁Yeah", "514": "▁most", "515": "▁ke", "516": "ors", "517": "ved", "518": "ys", "519": "▁sc", "520": "▁happ", "521": "ope", "522": "▁help", "523": "atch", "524": "▁What", "525": "▁rem", "526": "ple", "527": "▁Now", "528": "▁br", "529": "ool", "530": "oth", "531": "▁four", "532": "self", "533": "▁str", "534": "ne", "535": "thing", "536": "▁put", "537": "ial", "538": "▁great", "539": "ail", "540": "ub", "541": "ning", "542": "▁sm", "543": "▁feel", "544": "▁five", "545": "ody", "546": "undred", "547": "iss", "548": "ank", "549": "get", "550": "aking", "551": "▁many", "552": "▁hundred", "553": "▁years", "554": "▁being", "555": "▁come", "556": "▁mean", "557": "ily", "558": "▁different", "559": "▁after", "560": "▁ser", "561": "▁show", "562": "form", "563": "ful", "564": "oy", "565": "▁six", "566": "▁vide", "567": "▁V", "568": "▁its", "569": "▁point", "570": "▁day", "571": "▁des", "572": "ons", "573": "▁bit", "574": "▁bel", "575": "▁before", "576": "▁aw", "577": "▁end", "578": "▁Oh", "579": "▁still", "580": "ath", "581": "▁long", "582": "▁'", "583": "ise", "584": "ob", "585": "day", "586": "▁add", "587": "ft", "588": "ves", "589": "ces", "590": "ady", "591": "▁cr", "592": "▁around", "593": "▁try", "594": "les", "595": "vers", "596": "kay", "597": "ian", "598": "ates", "599": "▁find", "600": "ward", "601": "▁As", "602": "▁eight", "603": "lic", "604": "▁same", "605": "▁pos", "606": "▁em", "607": "▁made", "608": "▁supp", "609": "▁life", "610": "▁Be", "611": "pect", "612": "▁dec", "613": "▁play", "614": "ange", "615": "▁att", "616": "▁pers", "617": "ways", "618": "▁high", "619": "▁hand", "620": "▁next", "621": "▁cons", "622": "▁own", "623": "▁inv", "624": "ower", "625": "▁ind", "626": "ert", "627": "ng", "628": "ave", "629": "▁year", "630": "▁big", "631": "ating", "632": "▁world", "633": "▁rel", "634": "▁sure", "635": "▁tra", "636": "ew", "637": "ered", "638": "▁fin", "639": "▁Well", "640": "▁sl", "641": "▁doing", "642": "bs", "643": "▁set", "644": "▁rec", "645": "ual", "646": "cial", "647": "▁ph", "648": "erm", "649": "▁love", "650": "ph", "651": "▁real", "652": "▁last", "653": "ict", "654": "▁bo", "655": "▁ra", "656": "ible", "657": "▁wr", "658": "mer", "659": "▁count", "660": "ities", "661": "▁always", "662": "inet", "663": "ments", "664": "uc", "665": "▁might", "666": "▁inter", "667": "▁video", "668": "gin", "669": "▁tell", "670": "▁never", "671": "vent", "672": "▁import", "673": "ied", "674": "▁sy", "675": "▁How", "676": "ically", "677": "ought", "678": "▁thir", "679": "▁rep", "680": "ks", "681": "ib", "682": "▁fam", "683": "ject", "684": "▁bas", "685": "▁She", "686": "▁give", "687": "akes", "688": "▁ninet", "689": "▁reg", "690": "▁min", "691": "▁op", "692": "▁def", "693": "▁didn", "694": "te", "695": "▁cour", "696": "▁why", "697": "▁ent", "698": "▁place", "699": "▁ins", "700": "▁car", "701": "ather", "702": "▁person", "703": "ular", "704": "▁inst", "705": "▁prod", "706": "lect", "707": "▁Al", "708": "▁today", "709": "▁bec", "710": "▁sur", "711": "▁All", "712": "▁another", "713": "▁bus", "714": "▁keep", "715": "ell", "716": "ese", "717": "riend", "718": "▁quest", "719": "▁talk", "720": "als", "721": "ings", "722": "▁mon", "723": "cond", "724": "old", "725": "▁acc", "726": "▁la", "727": "▁num", "728": "ident", "729": "▁che", "730": "iness", "731": "▁turn", "732": "▁ear", "733": "▁No", "734": "ousand", "735": "▁better", "736": "ific", "737": "▁loo", "738": "▁gl", "739": "oc", "740": "▁important", "741": "ited", "742": "▁An", "743": "▁thousand", "744": "ility", "745": "llow", "746": "▁used", "747": "▁gen", "748": "▁sim", "749": "li", "750": "▁happen", "751": "▁Un", "752": "▁Let", "753": "air", "754": "ock", "755": "ably", "756": "gg", "757": "▁watch", "758": "▁For", "759": "▁sw", "760": "ren", "761": "ute", "762": "ever", "763": "▁pol", "764": "▁sch", "765": "▁When", "766": "▁such", "767": "▁fif", "768": "▁home", "769": "▁cle", "770": "▁contin", "771": "ouse", "772": "▁friend", "773": "uring", "774": "▁Okay", "775": "gr", "776": "▁able", "777": "▁stud", "778": "▁eff", "779": "hip", "780": "body", "781": "▁top", "782": "ness", "783": "▁exper", "784": "▁pret", "785": "▁both", "786": "▁done", "787": "cri", "788": "▁mark", "789": "▁while", "790": "▁old", "791": "ros", "792": "ont", "793": "▁second", "794": "ative", "795": "▁thought", "796": "▁best", "797": "▁found", "798": "iew", "799": "▁belie", "800": "▁each", "801": "erest", "802": "▁tri", "803": "▁eas", "804": "▁ca", "805": "▁fact", "806": "▁care", "807": "▁fun", "808": "atter", "809": "ures", "810": "▁head", "811": "▁lear", "812": "▁water", "813": "▁hard", "814": "▁few", "815": "▁side", "816": "ween", "817": "▁exp", "818": "▁away", "819": "its", "820": "▁ext", "821": "lud", "822": "▁run", "823": "▁trans", "824": "ince", "825": "▁sk", "826": "▁open", "827": "cus", "828": "▁between", "829": "▁called", "830": "▁wee", "831": "▁pretty", "832": "ason", "833": "▁far", "834": "ember", "835": "omm", "836": "▁interest", "837": "any", "838": "ner", "839": "uff", "840": "▁pres", "841": "▁cur", "842": "▁child", "843": "ee", "844": "▁toget", "845": "▁together", "846": "olog", "847": "▁God", "848": "ond", "849": "▁char", "850": "▁looking", "851": "stem", "852": "az", "853": "cent", "854": "▁ob", "855": "▁ass", "856": "land", "857": "▁doesn", "858": "▁business", "859": "▁course", "860": "▁ten", "861": "ps", "862": "arch", "863": "ced", "864": "ms", "865": "ize", "866": "nce", "867": "▁ref", "868": "▁name", "869": "ross", "870": "▁grow", "871": "oney", "872": "▁went", "873": "ics", "874": "teen", "875": "▁cou", "876": "▁prob", "877": "▁ret", "878": "▁guys", "879": "▁came", "880": "ash", "881": "led", "882": "▁Eur", "883": "ues", "884": "▁ide", "885": "gan", "886": "▁everything", "887": "▁getting", "888": "▁ask", "889": "▁cor", "890": "▁build", "891": "▁sign", "892": "▁small", "893": "uck", "894": "▁el", "895": "▁col", "896": "▁Is", "897": "ational", "898": "stand", "899": "cy", "900": "▁conf", "901": "der", "902": "▁bre", "903": "▁cap", "904": "▁mod", "905": "ets", "906": "ike", "907": "▁number", "908": "▁comple", "909": "ertain", "910": "▁ever", "911": "▁coll", "912": "▁hum", "913": "▁Europe", "914": "▁cre", "915": "▁met", "916": "▁exam", "917": "▁move", "918": "▁pass", "919": "▁left", "920": "▁system", "921": "▁includ", "922": "▁Thank", "923": "cept", "924": "▁wom", "925": "▁product", "926": "ten", "927": "▁rest", "928": "▁probably", "929": "▁dri", "930": "▁Do", "931": "▁gener", "932": "▁anything", "933": "▁lar", "934": "▁My", "935": "▁school", "936": "▁lead", "937": "▁sub", "938": "▁ty", "939": "▁plan", "940": "▁seem", "941": "▁whole", "942": "irect", "943": "▁light", "944": "▁must", "945": "▁mom", "946": "▁opp", "947": "▁support", "948": "▁family", "949": "ices", "950": "amp", "951": "▁proble", "952": "▁dr", "953": "ready", "954": "▁using", "955": "ense", "956": "▁prov", "957": "ush", "958": "ax", "959": "▁power", "960": "▁Re", "961": "alth", "962": "▁ev", "963": "▁stand", "964": "▁war", "965": "ts", "966": "▁", "967": "e", "968": "t", "969": "o", "970": "a", "971": "n", "972": "i", "973": "s", "974": "r", "975": "h", "976": "l", "977": "d", "978": "u", "979": "c", "980": "m", "981": "y", "982": "g", "983": "w", "984": "f", "985": "p", "986": ".", "987": "b", "988": ",", "989": "v", "990": "k", "991": "'", "992": "I", "993": "T", "994": "A", "995": "S", "996": "x", "997": "W", "998": "j", "999": "B", "1000": "C", "1001": "H", "1002": "?", "1003": "M", "1004": "O", "1005": "Y", "1006": "N", "1007": "P", "1008": "E", "1009": "q", "1010": "L", "1011": "D", "1012": "z", "1013": "G", "1014": "F", "1015": "R", "1016": "!", "1017": "J", "1018": "U", "1019": "K", "1020": "V", "1021": "Q", "1022": "Z", "1023": "X"} \ No newline at end of file diff --git a/convert/parakeet-tdt-ctc-110m/coreml/convert-parakeet.py b/convert/parakeet-tdt-ctc-110m/coreml/convert-parakeet.py deleted file mode 100644 index 6cd1c871c81f26b9449bf6313077afe8e27527d9..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/convert-parakeet.py +++ /dev/null @@ -1,697 +0,0 @@ -#!/usr/bin/env python3 -"""CLI for exporting Parakeet TDT-CTC 110M Hybrid components to CoreML.""" -from __future__ import annotations - -import json -from dataclasses import asdict -from pathlib import Path -from typing import Dict, Optional, Tuple - -import coremltools as ct -import numpy as np -import soundfile as sf -import torch -import typer - -import nemo.collections.asr as nemo_asr - -from individual_components import ( - CTCHeadWrapper, - DecoderWrapper, - EncoderWrapper, - ExportSettings, - JointWrapper, - JointDecisionWrapper, - JointDecisionSingleStep, - PreprocessorWrapper, - MelEncoderWrapper, - _coreml_convert, -) - -DEFAULT_MODEL_ID = "nvidia/parakeet-tdt_ctc-110m" -AUTHOR = "Fluid Inference" - - -def _compute_length(seconds: float, sample_rate: int) -> int: - return int(round(seconds * sample_rate)) - - -def _prepare_audio( - validation_audio: Optional[Path], - sample_rate: int, - max_samples: int, - seed: Optional[int], -) -> torch.Tensor: - if validation_audio is None: - if seed is not None: - torch.manual_seed(seed) - audio = torch.randn(1, max_samples, dtype=torch.float32) - return audio - - data, sr = sf.read(str(validation_audio), dtype="float32") - if sr != sample_rate: - raise typer.BadParameter( - f"Validation audio sample rate {sr} does not match model rate {sample_rate}" - ) - - if data.ndim > 1: - data = data[:, 0] - - if data.size == 0: - raise typer.BadParameter("Validation audio is empty") - - if data.size < max_samples: - pad_width = max_samples - data.size - data = np.pad(data, (0, pad_width)) - elif data.size > max_samples: - data = data[:max_samples] - - audio = torch.from_numpy(data).unsqueeze(0).to(dtype=torch.float32) - return audio - - -def _save_mlpackage(model: ct.models.MLModel, path: Path, description: str) -> None: - # Ensure iOS 17+ target for MLProgram ops and ANE readiness - try: - model.minimum_deployment_target = ct.target.iOS17 - except Exception: - pass - model.short_description = description - model.author = AUTHOR - path.parent.mkdir(parents=True, exist_ok=True) - model.save(str(path)) - - -def _tensor_shape(tensor: torch.Tensor) -> Tuple[int, ...]: - return tuple(int(dim) for dim in tensor.shape) - - -def _parse_compute_units(name: str) -> ct.ComputeUnit: - """Parse a human-friendly compute units string into ct.ComputeUnit. - - Accepted (case-insensitive): ALL, CPU_ONLY, CPU_AND_GPU, CPU_AND_NE. - """ - normalized = str(name).strip().upper() - mapping = { - "ALL": ct.ComputeUnit.ALL, - "CPU_ONLY": ct.ComputeUnit.CPU_ONLY, - "CPU_AND_GPU": ct.ComputeUnit.CPU_AND_GPU, - "CPU_AND_NE": ct.ComputeUnit.CPU_AND_NE, - "CPU_AND_NEURALENGINE": ct.ComputeUnit.CPU_AND_NE, - } - if normalized not in mapping: - raise typer.BadParameter( - f"Unknown compute units '{name}'. Choose from: " + ", ".join(mapping.keys()) - ) - return mapping[normalized] - - -def _parse_compute_precision(name: Optional[str]) -> Optional[ct.precision]: - """Parse compute precision string into ct.precision or None. - - Accepted (case-insensitive): FLOAT32, FLOAT16. If None/empty, returns None (tool default). - """ - if name is None: - return None - normalized = str(name).strip().upper() - if normalized == "": - return None - mapping = { - "FLOAT32": ct.precision.FLOAT32, - "FLOAT16": ct.precision.FLOAT16, - } - if normalized not in mapping: - raise typer.BadParameter( - f"Unknown compute precision '{name}'. Choose from: " + ", ".join(mapping.keys()) - ) - return mapping[normalized] - - -# Fixed export choices: CPU_ONLY + FP32, min target iOS17 - - -app = typer.Typer(add_completion=False, pretty_exceptions_show_locals=False) - - -@app.command() -def convert( - nemo_path: Optional[Path] = typer.Option( - None, - "--nemo-path", - exists=True, - resolve_path=True, - help="Path to parakeet-tdt_ctc-110m .nemo checkpoint (skip to auto-download)", - ), - model_id: str = typer.Option( - DEFAULT_MODEL_ID, - "--model-id", - help="Model identifier to download when --nemo-path is omitted", - ), - output_dir: Path = typer.Option(Path("parakeet_110m_coreml"), help="Directory where mlpackages and metadata will be written"), - preprocessor_cu: str = typer.Option( - "CPU_ONLY", - "--preprocessor-cu", - help="Compute units for preprocessor (default CPU_ONLY)", - ), - mel_encoder_cu: str = typer.Option( - "CPU_ONLY", - "--mel-encoder-cu", - help="Compute units for fused mel+encoder (default CPU_ONLY)", - ), - compute_precision: Optional[str] = typer.Option( - None, - "--compute-precision", - help="Export precision: FLOAT32 (default) or FLOAT16 to shrink non-quantized weights.", - ), -) -> None: - """Export all Parakeet TDT-CTC 110M Hybrid sub-modules to CoreML with a fixed 15-second window. - - This exports both CTC and TDT components from the hybrid model. - """ - # Runtime CoreML contract keeps U=1 so the prediction net matches the streaming decoder. - export_settings = ExportSettings( - output_dir=output_dir, - compute_units=ct.ComputeUnit.CPU_ONLY, # Default: CPU-only for all components - deployment_target=ct.target.iOS17, # iOS 17+ features and kernels - compute_precision=_parse_compute_precision(compute_precision), - max_audio_seconds=15.0, - max_symbol_steps=1, - ) - - typer.echo("Export configuration:") - typer.echo(asdict(export_settings)) - - output_dir.mkdir(parents=True, exist_ok=True) - pre_cu = _parse_compute_units(preprocessor_cu) - melenc_cu = _parse_compute_units(mel_encoder_cu) - - if nemo_path is not None: - typer.echo(f"Loading NeMo model from {nemo_path}
") - # 110M is a hybrid model: EncDecHybridRNNTCTCBPEModel - asr_model = nemo_asr.models.EncDecHybridRNNTCTCBPEModel.restore_from( - str(nemo_path), map_location="cpu" - ) - checkpoint_meta = { - "type": "file", - "path": str(nemo_path), - } - else: - typer.echo(f"Downloading NeMo model via {model_id}
") - # 110M is a hybrid model: EncDecHybridRNNTCTCBPEModel - asr_model = nemo_asr.models.EncDecHybridRNNTCTCBPEModel.from_pretrained( - model_id, map_location="cpu" - ) - checkpoint_meta = { - "type": "pretrained", - "model_id": model_id, - } - asr_model.eval() - - sample_rate = int(asr_model.cfg.preprocessor.sample_rate) - max_samples = _compute_length(export_settings.max_audio_seconds, sample_rate) - - # Look for a bundled 15s 16kHz audio file - default_audio = (Path(__file__).parent / "audio" / "yc_first_minute_16k_15s.wav").resolve() - if default_audio.exists(): - typer.echo(f"Using trace audio: {default_audio}") - audio_tensor = _prepare_audio(default_audio, sample_rate, max_samples, seed=None) - else: - typer.echo("No trace audio found, using random noise for tracing") - audio_tensor = _prepare_audio(None, sample_rate, max_samples, seed=42) - audio_length = torch.tensor([max_samples], dtype=torch.int32) - - preprocessor = PreprocessorWrapper(asr_model.preprocessor.eval()) - encoder = EncoderWrapper(asr_model.encoder.eval()) - decoder = DecoderWrapper(asr_model.decoder.eval()) - joint = JointWrapper(asr_model.joint.eval()) - # CTC head for hybrid model - ctc_head = CTCHeadWrapper(asr_model.ctc_decoder.eval()) - - decoder_export_flag = getattr(asr_model.decoder, "_rnnt_export", False) - asr_model.decoder._rnnt_export = True - - try: - with torch.inference_mode(): - mel_ref, mel_length_ref = preprocessor(audio_tensor, audio_length) - mel_length_ref = mel_length_ref.to(dtype=torch.int32) - encoder_ref, encoder_length_ref = encoder(mel_ref, mel_length_ref) - encoder_length_ref = encoder_length_ref.to(dtype=torch.int32) - # CTC log probs - ctc_log_probs_ref = ctc_head(encoder_ref) - - # Clone Tensors to drop the inference tensor flag before tracing - mel_ref = mel_ref.clone() - mel_length_ref = mel_length_ref.clone() - encoder_ref = encoder_ref.clone() - encoder_length_ref = encoder_length_ref.clone() - ctc_log_probs_ref = ctc_log_probs_ref.clone() - - vocab_size = int(asr_model.tokenizer.vocab_size) - num_extra = int(asr_model.joint.num_extra_outputs) - decoder_hidden = int(asr_model.decoder.pred_hidden) - decoder_layers = int(asr_model.decoder.pred_rnn_layers) - - typer.echo(f"Model info:") - typer.echo(f" Vocab size: {vocab_size}") - typer.echo(f" Num extra (duration bins): {num_extra}") - typer.echo(f" Decoder hidden: {decoder_hidden}") - typer.echo(f" Decoder layers: {decoder_layers}") - typer.echo(f" Encoder output shape: {_tensor_shape(encoder_ref)}") - - targets = torch.full( - (1, export_settings.max_symbol_steps), - fill_value=asr_model.decoder.blank_idx, - dtype=torch.int32, - ) - target_lengths = torch.tensor( - [export_settings.max_symbol_steps], dtype=torch.int32 - ) - zero_state = torch.zeros( - decoder_layers, - 1, - decoder_hidden, - dtype=torch.float32, - ) - - with torch.inference_mode(): - decoder_ref, h_ref, c_ref = decoder(targets, target_lengths, zero_state, zero_state) - joint_ref = joint(encoder_ref, decoder_ref) - - decoder_ref = decoder_ref.clone() - h_ref = h_ref.clone() - c_ref = c_ref.clone() - joint_ref = joint_ref.clone() - - typer.echo(f" Decoder output shape: {_tensor_shape(decoder_ref)}") - typer.echo(f" Joint output shape: {_tensor_shape(joint_ref)}") - typer.echo(f" CTC log probs shape: {_tensor_shape(ctc_log_probs_ref)}") - - typer.echo("Tracing and converting preprocessor
") - # Ensure tracing happens on CPU explicitly - preprocessor = preprocessor.cpu() - audio_tensor = audio_tensor.cpu() - audio_length = audio_length.cpu() - traced_preprocessor = torch.jit.trace( - preprocessor, (audio_tensor, audio_length), strict=False - ) - traced_preprocessor.eval() - preprocessor_inputs = [ - # Allow variable-length audio up to the fixed 15s window using RangeDim - ct.TensorType( - name="audio", - shape=(1, ct.RangeDim(1, max_samples)), - dtype=np.float32, - ), - ct.TensorType(name="audio_length", shape=(1,), dtype=np.int32), - ] - preprocessor_outputs = [ - ct.TensorType(name="mel_features", dtype=np.float32), - ct.TensorType(name="mel_length", dtype=np.int32), - ] - # Preprocessor compute units (parametrized; default CPU_ONLY) - preprocessor_model = _coreml_convert( - traced_preprocessor, - preprocessor_inputs, - preprocessor_outputs, - export_settings, - compute_units_override=pre_cu, - ) - preprocessor_path = output_dir / "parakeet_preprocessor.mlpackage" - _save_mlpackage( - preprocessor_model, - preprocessor_path, - "Parakeet 110M preprocessor (15 s window)", - ) - - typer.echo("Tracing and converting encoder
") - traced_encoder = torch.jit.trace( - encoder, (mel_ref, mel_length_ref), strict=False - ) - traced_encoder.eval() - encoder_inputs = [ - ct.TensorType(name="mel_features", shape=_tensor_shape(mel_ref), dtype=np.float32), - ct.TensorType(name="mel_length", shape=(1,), dtype=np.int32), - ] - encoder_outputs = [ - ct.TensorType(name="encoder_output", dtype=np.float32), - ct.TensorType(name="encoder_length", dtype=np.int32), - ] - # Encoder: CPU only - encoder_model = _coreml_convert( - traced_encoder, - encoder_inputs, - encoder_outputs, - export_settings, - compute_units_override=ct.ComputeUnit.CPU_ONLY, - ) - encoder_path = output_dir / "parakeet_encoder.mlpackage" - _save_mlpackage( - encoder_model, - encoder_path, - "Parakeet 110M encoder (15 s window)", - ) - - # CTC Head for hybrid model - typer.echo("Tracing and converting CTC head
") - traced_ctc_head = torch.jit.trace( - ctc_head, (encoder_ref,), strict=False - ) - traced_ctc_head.eval() - ctc_head_inputs = [ - ct.TensorType(name="encoder_output", shape=_tensor_shape(encoder_ref), dtype=np.float32), - ] - ctc_head_outputs = [ - ct.TensorType(name="ctc_logits", dtype=np.float32), - ] - ctc_head_model = _coreml_convert( - traced_ctc_head, - ctc_head_inputs, - ctc_head_outputs, - export_settings, - compute_units_override=ct.ComputeUnit.CPU_ONLY, - ) - ctc_head_path = output_dir / "parakeet_ctc_head.mlpackage" - _save_mlpackage( - ctc_head_model, - ctc_head_path, - "Parakeet 110M CTC decoder head", - ) - - # Optional fused export: Preprocessor + Encoder - typer.echo("Tracing and converting fused mel+encoder
") - mel_encoder = MelEncoderWrapper(preprocessor, encoder) - traced_mel_encoder = torch.jit.trace( - mel_encoder, (audio_tensor, audio_length), strict=False - ) - traced_mel_encoder.eval() - mel_encoder_inputs = [ - # Keep fixed 15s window for fused Mel+Encoder - ct.TensorType(name="audio", shape=(1, max_samples), dtype=np.float32), - ct.TensorType(name="audio_length", shape=(1,), dtype=np.int32), - ] - mel_encoder_outputs = [ - ct.TensorType(name="encoder_output", dtype=np.float32), - ct.TensorType(name="encoder_length", dtype=np.int32), - ] - # Fused mel+encoder compute units (parametrized; default CPU_ONLY) - mel_encoder_model = _coreml_convert( - traced_mel_encoder, - mel_encoder_inputs, - mel_encoder_outputs, - export_settings, - compute_units_override=melenc_cu, - ) - mel_encoder_path = output_dir / "parakeet_mel_encoder.mlpackage" - _save_mlpackage( - mel_encoder_model, - mel_encoder_path, - "Parakeet 110M fused Mel+Encoder (15 s window)", - ) - - typer.echo("Tracing and converting decoder
") - traced_decoder = torch.jit.trace( - decoder, - (targets, target_lengths, zero_state, zero_state), - strict=False, - ) - traced_decoder.eval() - decoder_inputs = [ - ct.TensorType(name="targets", shape=_tensor_shape(targets), dtype=np.int32), - ct.TensorType(name="target_length", shape=(1,), dtype=np.int32), - ct.TensorType(name="h_in", shape=_tensor_shape(zero_state), dtype=np.float32), - ct.TensorType(name="c_in", shape=_tensor_shape(zero_state), dtype=np.float32), - ] - decoder_outputs = [ - ct.TensorType(name="decoder", dtype=np.float32), - ct.TensorType(name="h_out", dtype=np.float32), - ct.TensorType(name="c_out", dtype=np.float32), - ] - # Decoder: CPU only - decoder_model = _coreml_convert( - traced_decoder, - decoder_inputs, - decoder_outputs, - export_settings, - compute_units_override=ct.ComputeUnit.CPU_ONLY, - ) - decoder_path = output_dir / "parakeet_decoder.mlpackage" - _save_mlpackage( - decoder_model, - decoder_path, - "Parakeet 110M decoder (RNNT prediction network)", - ) - - typer.echo("Tracing and converting joint
") - traced_joint = torch.jit.trace( - joint, - (encoder_ref, decoder_ref), - strict=False, - ) - traced_joint.eval() - joint_inputs = [ - ct.TensorType(name="encoder", shape=_tensor_shape(encoder_ref), dtype=np.float32), - ct.TensorType(name="decoder", shape=_tensor_shape(decoder_ref), dtype=np.float32), - ] - joint_outputs = [ - ct.TensorType(name="logits", dtype=np.float32), - ] - # Joint: CPU only - joint_model = _coreml_convert( - traced_joint, - joint_inputs, - joint_outputs, - export_settings, - compute_units_override=ct.ComputeUnit.CPU_ONLY, - ) - joint_path = output_dir / "parakeet_joint.mlpackage" - _save_mlpackage( - joint_model, - joint_path, - "Parakeet 110M joint network (RNNT)", - ) - - # Joint + decision head (split logits, softmax, argmax) - typer.echo("Tracing and converting joint decision head
") - vocab_size = int(asr_model.tokenizer.vocab_size) - num_extra = int(asr_model.joint.num_extra_outputs) - joint_decision = JointDecisionWrapper(joint, vocab_size=vocab_size, num_extra=num_extra) - traced_joint_decision = torch.jit.trace( - joint_decision, - (encoder_ref, decoder_ref), - strict=False, - ) - traced_joint_decision.eval() - joint_decision_inputs = [ - ct.TensorType(name="encoder", shape=_tensor_shape(encoder_ref), dtype=np.float32), - ct.TensorType(name="decoder", shape=_tensor_shape(decoder_ref), dtype=np.float32), - ] - joint_decision_outputs = [ - ct.TensorType(name="token_id", dtype=np.int32), - ct.TensorType(name="token_prob", dtype=np.float32), - ct.TensorType(name="duration", dtype=np.int32), - ] - # JointDecision: CPU only - joint_decision_model = _coreml_convert( - traced_joint_decision, - joint_decision_inputs, - joint_decision_outputs, - export_settings, - compute_units_override=ct.ComputeUnit.CPU_ONLY, - ) - joint_decision_path = output_dir / "parakeet_joint_decision.mlpackage" - _save_mlpackage( - joint_decision_model, - joint_decision_path, - "Parakeet 110M joint + decision head (split, softmax, argmax)", - ) - - # Single-step JointDecision for [1,512,1] x [1,640,1] -> [1,1,1] - # Note: 110M encoder dim is 512 (not 1024 like 0.6B) - typer.echo("Tracing and converting single-step joint decision
") - jd_single = JointDecisionSingleStep(joint, vocab_size=vocab_size, num_extra=num_extra) - # Create single-step slices from refs - enc_step = encoder_ref[:, :, :1].contiguous() - dec_step = decoder_ref[:, :, :1].contiguous() - traced_jd_single = torch.jit.trace( - jd_single, - (enc_step, dec_step), - strict=False, - ) - traced_jd_single.eval() - jd_single_inputs = [ - ct.TensorType(name="encoder_step", shape=(1, enc_step.shape[1], 1), dtype=np.float32), - ct.TensorType(name="decoder_step", shape=(1, dec_step.shape[1], 1), dtype=np.float32), - ] - jd_single_outputs = [ - ct.TensorType(name="token_id", dtype=np.int32), - ct.TensorType(name="token_prob", dtype=np.float32), - ct.TensorType(name="duration", dtype=np.int32), - ct.TensorType(name="top_k_ids", dtype=np.int32), - ct.TensorType(name="top_k_logits", dtype=np.float32), - ] - # Single-step JointDecision: CPU only - jd_single_model = _coreml_convert( - traced_jd_single, - jd_single_inputs, - jd_single_outputs, - export_settings, - compute_units_override=ct.ComputeUnit.CPU_ONLY, - ) - jd_single_path = output_dir / "parakeet_joint_decision_single_step.mlpackage" - _save_mlpackage( - jd_single_model, - jd_single_path, - "Parakeet 110M single-step joint decision (current frame)", - ) - - # Export vocabulary - typer.echo("Exporting vocabulary
") - vocab_path = output_dir / "vocab.json" - vocab_dict = { - "vocab_size": vocab_size, - "blank_id": int(asr_model.decoder.blank_idx), - "tokens": asr_model.tokenizer.vocab, - } - vocab_path.write_text(json.dumps(vocab_dict, indent=2, ensure_ascii=False)) - - metadata: Dict[str, object] = { - "model_id": model_id, - "model_type": "hybrid_rnnt_ctc", - "sample_rate": sample_rate, - "max_audio_seconds": export_settings.max_audio_seconds, - "max_audio_samples": max_samples, - "max_symbol_steps": export_settings.max_symbol_steps, - "vocab_size": vocab_size, - "joint_extra_outputs": num_extra, - "encoder_dim": int(encoder_ref.shape[1]), # 512 for 110M - "decoder_dim": int(decoder_ref.shape[1]), # 640 - "decoder_hidden": decoder_hidden, - "decoder_layers": decoder_layers, - "blank_id": int(asr_model.decoder.blank_idx), - "checkpoint": checkpoint_meta, - "coreml": { - "compute_units": export_settings.compute_units.name, - "compute_precision": ( - export_settings.compute_precision.name - if export_settings.compute_precision is not None - else "FLOAT32" - ), - }, - "components": { - "preprocessor": { - "inputs": { - "audio_signal": list(_tensor_shape(audio_tensor)), - "audio_length": [1], - }, - "outputs": { - "mel": list(_tensor_shape(mel_ref)), - "mel_length": [1], - }, - "path": preprocessor_path.name, - }, - "encoder": { - "inputs": { - "mel": list(_tensor_shape(mel_ref)), - "mel_length": [1], - }, - "outputs": { - "encoder": list(_tensor_shape(encoder_ref)), - "encoder_length": [1], - }, - "path": encoder_path.name, - }, - "ctc_head": { - "inputs": { - "encoder": list(_tensor_shape(encoder_ref)), - }, - "outputs": { - "log_probs": list(_tensor_shape(ctc_log_probs_ref)), - }, - "path": ctc_head_path.name, - }, - "mel_encoder": { - "inputs": { - "audio_signal": [1, max_samples], - "audio_length": [1], - }, - "outputs": { - "encoder": list(_tensor_shape(encoder_ref)), - "encoder_length": [1], - }, - "path": mel_encoder_path.name, - }, - "decoder": { - "inputs": { - "targets": list(_tensor_shape(targets)), - "target_length": [1], - "h_in": list(_tensor_shape(zero_state)), - "c_in": list(_tensor_shape(zero_state)), - }, - "outputs": { - "decoder": list(_tensor_shape(decoder_ref)), - "h_out": list(_tensor_shape(h_ref)), - "c_out": list(_tensor_shape(c_ref)), - }, - "path": decoder_path.name, - }, - "joint": { - "inputs": { - "encoder": list(_tensor_shape(encoder_ref)), - "decoder": list(_tensor_shape(decoder_ref)), - }, - "outputs": { - "logits": list(_tensor_shape(joint_ref)), - }, - "path": joint_path.name, - }, - "joint_decision": { - "inputs": { - "encoder": list(_tensor_shape(encoder_ref)), - "decoder": list(_tensor_shape(decoder_ref)), - }, - "outputs": { - "token_id": [ - _tensor_shape(encoder_ref)[0], - _tensor_shape(encoder_ref)[2], - _tensor_shape(decoder_ref)[2], - ], - "token_prob": [ - _tensor_shape(encoder_ref)[0], - _tensor_shape(encoder_ref)[2], - _tensor_shape(decoder_ref)[2], - ], - "duration": [ - _tensor_shape(encoder_ref)[0], - _tensor_shape(encoder_ref)[2], - _tensor_shape(decoder_ref)[2], - ], - }, - "path": joint_decision_path.name, - }, - "joint_decision_single_step": { - "inputs": { - "encoder_step": [1, int(encoder_ref.shape[1]), 1], - "decoder_step": [1, int(decoder_ref.shape[1]), 1], - }, - "outputs": { - "token_id": [1, 1, 1], - "token_prob": [1, 1, 1], - "duration": [1, 1, 1], - "top_k_ids": [1, 1, 1, 64], - "top_k_logits": [1, 1, 1, 64], - }, - "path": jd_single_path.name, - }, - }, - } - - metadata_path = output_dir / "metadata.json" - metadata_path.write_text(json.dumps(metadata, indent=2)) - typer.echo(f"Export complete. Metadata written to {metadata_path}") - - finally: - asr_model.decoder._rnnt_export = decoder_export_flag - - -if __name__ == "__main__": - app() diff --git a/convert/parakeet-tdt-ctc-110m/coreml/hybrid_earnings_benchmark.json b/convert/parakeet-tdt-ctc-110m/coreml/hybrid_earnings_benchmark.json deleted file mode 100644 index 3dd6bc7551d07ef8a966ba06590d2454e4324983..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/hybrid_earnings_benchmark.json +++ /dev/null @@ -1,35 +0,0 @@ -{ - "approach" : "single-encoder", - "model" : "parakeet-tdt-ctc-110m-hybrid", - "results" : [ - { - "audioLength" : 15, - "ctcDetections" : [ - { - "endTime" : 6.0800000000000001, - "inReference" : true, - "score" : -8.3699999999999992, - "source" : "ctc", - "startTime" : 4.96, - "word" : "LATAM Airlines" - } - ], - "dictFound" : 1, - "dictTotal" : 1, - "fileId" : "4329526_chunk0", - "hypothesis" : "goodday everyone and welcome to latam airlines group earnings release confonference call just as a reminder this conference is being recorded lat tam airlines group eararnings releaseed for the", - "processingTime" : 0.070000000000000007, - "reference" : "good day everyone and welcome to latam airlines group earnings release conference call just as a reminder this conference is being recorded latam airlines group earnings released for the", - "wer" : 24.140000000000001 - } - ], - "summary" : { - "avgWer" : 24.140000000000001, - "dictPass" : 1, - "dictRate" : 100, - "dictTotal" : 1, - "totalAudioDuration" : 15, - "totalProcessingTime" : 0.070000000000000007, - "totalTests" : 1 - } -} \ No newline at end of file diff --git a/convert/parakeet-tdt-ctc-110m/coreml/individual_components.py b/convert/parakeet-tdt-ctc-110m/coreml/individual_components.py deleted file mode 100644 index 0f6ae75f46ab359edc5c6dd6799dccf0b04a7ed7..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/individual_components.py +++ /dev/null @@ -1,265 +0,0 @@ -#!/usr/bin/env python3 -"""Export Parakeet TDT-CTC 110M Hybrid RNNT components into CoreML and validate outputs.""" -from __future__ import annotations - -from dataclasses import dataclass -from pathlib import Path -from typing import Optional, Tuple - -import coremltools as ct -import torch - - -@dataclass -class ExportSettings: - output_dir: Path - compute_units: ct.ComputeUnit - deployment_target: Optional[ct.target.iOS17] - compute_precision: Optional[ct.precision] - max_audio_seconds: float - max_symbol_steps: int - - -@dataclass -class ValidationSettings: - audio_path: Optional[Path] - seconds: float - seed: Optional[int] - rtol: float - atol: float - skip: bool - - -@dataclass -class ValidationDiff: - name: str - max_abs_diff: float - max_rel_diff: float - - -@dataclass -class ValidationResult: - source: str - audio_num_samples: int - audio_seconds: float - token_length: int - atol: float - rtol: float - diffs: Tuple[ValidationDiff, ...] - - -class PreprocessorWrapper(torch.nn.Module): - def __init__(self, module: torch.nn.Module) -> None: - super().__init__() - self.module = module - - def forward(self, audio_signal: torch.Tensor, length: torch.Tensor) -> Tuple[torch.Tensor, torch.Tensor]: - mel, mel_length = self.module(input_signal=audio_signal, length=length.to(dtype=torch.long)) - return mel, mel_length - - -class EncoderWrapper(torch.nn.Module): - def __init__(self, module: torch.nn.Module) -> None: - super().__init__() - self.module = module - - def forward(self, features: torch.Tensor, length: torch.Tensor) -> Tuple[torch.Tensor, torch.Tensor]: - encoded, encoded_lengths = self.module(audio_signal=features, length=length.to(dtype=torch.long)) - return encoded, encoded_lengths - - -class DecoderWrapper(torch.nn.Module): - def __init__(self, module: torch.nn.Module) -> None: - super().__init__() - self.module = module - - def forward( - self, - targets: torch.Tensor, - target_lengths: torch.Tensor, - h_in: torch.Tensor, - c_in: torch.Tensor, - ) -> Tuple[torch.Tensor, torch.Tensor, torch.Tensor]: - state = [h_in, c_in] - decoder_output, _, new_state = self.module( - targets=targets.to(dtype=torch.long), - target_length=target_lengths.to(dtype=torch.long), - states=state, - ) - return decoder_output, new_state[0], new_state[1] - - -class JointWrapper(torch.nn.Module): - """Joint network for 110M hybrid model. - - Note: The 110M model has encoder_dim=512 and decoder_dim=640. - The joint network projects both to 640, then combines them. - """ - def __init__(self, module: torch.nn.Module) -> None: - super().__init__() - self.module = module - - def forward(self, encoder_outputs: torch.Tensor, decoder_outputs: torch.Tensor) -> torch.Tensor: - # Input: encoder_outputs [B, D_enc, T], decoder_outputs [B, D_dec, U] - # For 110M: D_enc=512, D_dec=640 - # Transpose to match what projection layers expect - encoder_outputs = encoder_outputs.transpose(1, 2) # [B, T, D_enc] - decoder_outputs = decoder_outputs.transpose(1, 2) # [B, U, D_dec] - - # Apply projections - enc_proj = self.module.enc(encoder_outputs) # [B, T, 640] - dec_proj = self.module.pred(decoder_outputs) # [B, U, 640] - - # Explicit broadcasting along T and U to avoid converter ambiguity - x = enc_proj.unsqueeze(2) + dec_proj.unsqueeze(1) # [B, T, U, 640] - x = self.module.joint_net[0](x) # ReLU - x = self.module.joint_net[1](x) # Dropout (no-op in eval) - out = self.module.joint_net[2](x) # Linear -> logits [B, T, U, vocab+1+durations] - return out - - -class CTCHeadWrapper(torch.nn.Module): - """CTC decoder head for 110M hybrid model. - - Takes encoder output and produces log probabilities over vocabulary. - The NeMo CTC decoder (ConvASRDecoder) uses Conv1d so it expects [B, D, T] format. - """ - def __init__(self, module: torch.nn.Module) -> None: - super().__init__() - self.module = module - - def forward(self, encoder_outputs: torch.Tensor) -> torch.Tensor: - # Input: encoder_outputs [B, D_enc, T] - already in the format CTC decoder expects - # The NeMo CTC decoder uses Conv1d internally, so it expects [B, D, T] - # Output: log probabilities [B, T, vocab+1] - log_probs = self.module(encoder_output=encoder_outputs) - return log_probs - - -class MelEncoderWrapper(torch.nn.Module): - """Fused wrapper: waveform -> mel -> encoder. - - Inputs: - - audio_signal: [B, S] - - audio_length: [B] - - Outputs: - - encoder: [B, D, T_enc] - - encoder_length: [B] - """ - def __init__(self, preprocessor: PreprocessorWrapper, encoder: EncoderWrapper) -> None: - super().__init__() - self.preprocessor = preprocessor - self.encoder = encoder - - def forward(self, audio_signal: torch.Tensor, audio_length: torch.Tensor) -> Tuple[torch.Tensor, torch.Tensor]: - mel, mel_length = self.preprocessor(audio_signal, audio_length) - encoded, enc_len = self.encoder(mel, mel_length.to(dtype=torch.int32)) - return encoded, enc_len - - -class JointDecisionWrapper(torch.nn.Module): - """Joint + decision head: outputs label id, label prob, duration frames. - - Splits joint logits into token logits and duration logits, applies softmax - over tokens, argmax for both heads, and gathers probability of the chosen token. - - Inputs: - - encoder_outputs: [B, D, T] - - decoder_outputs: [B, D, U] - - Returns: - - token_id: [B, T, U] int32 - - token_prob: [B, T, U] float32 - - duration: [B, T, U] int32 (frames; for v3 bins=[0,1,2,3,4]) - """ - def __init__(self, joint: JointWrapper, vocab_size: int, num_extra: int) -> None: - super().__init__() - self.joint = joint - self.vocab_with_blank = int(vocab_size) + 1 - self.num_extra = int(num_extra) - - def forward(self, encoder_outputs: torch.Tensor, decoder_outputs: torch.Tensor): - logits = self.joint(encoder_outputs, decoder_outputs) - token_logits = logits[..., : self.vocab_with_blank] - duration_logits = logits[..., -self.num_extra :] - - # Token selection - token_ids = torch.argmax(token_logits, dim=-1).to(dtype=torch.int32) - token_probs_all = torch.softmax(token_logits, dim=-1) - # gather expects int64 (long) indices; cast only for gather - token_prob = torch.gather( - token_probs_all, dim=-1, index=token_ids.long().unsqueeze(-1) - ).squeeze(-1) - - # Duration prediction (bins are identity mapping to frames for v3) - duration = torch.argmax(duration_logits, dim=-1).to(dtype=torch.int32) - return token_ids, token_prob, duration - - -class JointDecisionSingleStep(torch.nn.Module): - """Single-step variant for streaming: encoder_step [1, 512, 1] -> [1,1,1]. - - Note: For 110M model, encoder_dim is 512 (not 1024 like 0.6B). - - Inputs: - - encoder_step: [B=1, D=512, T=1] - - decoder_step: [B=1, D=640, U=1] - - Returns: - - token_id: [1, 1, 1] int32 - - token_prob: [1, 1, 1] float32 - - duration: [1, 1, 1] int32 - - top_k_ids: [1, 1, 1, K] int32 - - top_k_logits: [1, 1, 1, K] float32 - """ - def __init__(self, joint: JointWrapper, vocab_size: int, num_extra: int, top_k: int = 64) -> None: - super().__init__() - self.joint = joint - self.vocab_with_blank = int(vocab_size) + 1 - self.num_extra = int(num_extra) - # Emit top-K candidates to enable host-side re-ranking with contextual biasing - self.top_k = int(top_k) - - def forward(self, encoder_step: torch.Tensor, decoder_step: torch.Tensor): - # Reuse JointWrapper which expects [B, D, T] and [B, D, U] - logits = self.joint(encoder_step, decoder_step) # [1, 1, 1, V+extra] - token_logits = logits[..., : self.vocab_with_blank] - duration_logits = logits[..., -self.num_extra :] - - token_ids = torch.argmax(token_logits, dim=-1, keepdim=False).to(dtype=torch.int32) - token_probs_all = torch.softmax(token_logits, dim=-1) - token_prob = torch.gather( - token_probs_all, dim=-1, index=token_ids.long().unsqueeze(-1) - ).squeeze(-1) - duration = torch.argmax(duration_logits, dim=-1, keepdim=False).to(dtype=torch.int32) - - # Also expose top-K candidates for host-side re-ranking. - # Shapes preserved as [1, 1, 1, K] to match CoreML broadcasting expectations. - # Note: topk expects last dimension; original shape is [1, 1, 1, V]. - topk_logits, topk_ids_long = torch.topk(token_logits, k=min(self.top_k, token_logits.shape[-1]), dim=-1) - topk_ids = topk_ids_long.to(dtype=torch.int32) - return token_ids, token_prob, duration, topk_ids, topk_logits - - -def _coreml_convert( - traced: torch.jit.ScriptModule, - inputs, - outputs, - settings: ExportSettings, - compute_units_override: Optional[ct.ComputeUnit] = None, -) -> ct.models.MLModel: - cu = compute_units_override if compute_units_override is not None else settings.compute_units - kwargs = { - "convert_to": "mlprogram", - "inputs": inputs, - "outputs": outputs, - "compute_units": cu, - } - print("Converting:", traced.__class__.__name__) - print("Conversion kwargs:", kwargs) - if settings.deployment_target is not None: - kwargs["minimum_deployment_target"] = settings.deployment_target - if settings.compute_precision is not None: - kwargs["compute_precision"] = settings.compute_precision - return ct.convert(traced, **kwargs) diff --git a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/metadata.json b/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/metadata.json deleted file mode 100644 index 6db872ae8a7d463c95dc180210ace0fce019f4a6..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/metadata.json +++ /dev/null @@ -1,247 +0,0 @@ -{ - "model_id": "nvidia/parakeet-tdt_ctc-110m", - "model_type": "hybrid_rnnt_ctc", - "sample_rate": 16000, - "max_audio_seconds": 15.0, - "max_audio_samples": 240000, - "max_symbol_steps": 1, - "vocab_size": 1024, - "joint_extra_outputs": 5, - "encoder_dim": 512, - "decoder_dim": 640, - "decoder_hidden": 640, - "decoder_layers": 1, - "blank_id": 1024, - "checkpoint": { - "type": "pretrained", - "model_id": "nvidia/parakeet-tdt_ctc-110m" - }, - "coreml": { - "compute_units": "CPU_ONLY", - "compute_precision": "FLOAT32" - }, - "components": { - "preprocessor": { - "inputs": { - "audio_signal": [ - 1, - 240000 - ], - "audio_length": [ - 1 - ] - }, - "outputs": { - "mel": [ - 1, - 80, - 1501 - ], - "mel_length": [ - 1 - ] - }, - "path": "parakeet_preprocessor.mlpackage" - }, - "encoder": { - "inputs": { - "mel": [ - 1, - 80, - 1501 - ], - "mel_length": [ - 1 - ] - }, - "outputs": { - "encoder": [ - 1, - 512, - 188 - ], - "encoder_length": [ - 1 - ] - }, - "path": "parakeet_encoder.mlpackage" - }, - "ctc_head": { - "inputs": { - "encoder": [ - 1, - 512, - 188 - ] - }, - "outputs": { - "log_probs": [ - 1, - 188, - 1025 - ] - }, - "path": "parakeet_ctc_head.mlpackage" - }, - "mel_encoder": { - "inputs": { - "audio_signal": [ - 1, - 240000 - ], - "audio_length": [ - 1 - ] - }, - "outputs": { - "encoder": [ - 1, - 512, - 188 - ], - "encoder_length": [ - 1 - ] - }, - "path": "parakeet_mel_encoder.mlpackage" - }, - "decoder": { - "inputs": { - "targets": [ - 1, - 1 - ], - "target_length": [ - 1 - ], - "h_in": [ - 1, - 1, - 640 - ], - "c_in": [ - 1, - 1, - 640 - ] - }, - "outputs": { - "decoder": [ - 1, - 640, - 1 - ], - "h_out": [ - 1, - 1, - 640 - ], - "c_out": [ - 1, - 1, - 640 - ] - }, - "path": "parakeet_decoder.mlpackage" - }, - "joint": { - "inputs": { - "encoder": [ - 1, - 512, - 188 - ], - "decoder": [ - 1, - 640, - 1 - ] - }, - "outputs": { - "logits": [ - 1, - 188, - 1, - 1030 - ] - }, - "path": "parakeet_joint.mlpackage" - }, - "joint_decision": { - "inputs": { - "encoder": [ - 1, - 512, - 188 - ], - "decoder": [ - 1, - 640, - 1 - ] - }, - "outputs": { - "token_id": [ - 1, - 188, - 1 - ], - "token_prob": [ - 1, - 188, - 1 - ], - "duration": [ - 1, - 188, - 1 - ] - }, - "path": "parakeet_joint_decision.mlpackage" - }, - "joint_decision_single_step": { - "inputs": { - "encoder_step": [ - 1, - 512, - 1 - ], - "decoder_step": [ - 1, - 640, - 1 - ] - }, - "outputs": { - "token_id": [ - 1, - 1, - 1 - ], - "token_prob": [ - 1, - 1, - 1 - ], - "duration": [ - 1, - 1, - 1 - ], - "top_k_ids": [ - 1, - 1, - 1, - 64 - ], - "top_k_logits": [ - 1, - 1, - 1, - 64 - ] - }, - "path": "parakeet_joint_decision_single_step.mlpackage" - } - } -} \ No newline at end of file diff --git a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_ctc_head.mlpackage/Data/com.apple.CoreML/model.mlmodel b/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_ctc_head.mlpackage/Data/com.apple.CoreML/model.mlmodel deleted file mode 100644 index b807d0d0234624c649b82a7bd3e36bd9095fa961..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_ctc_head.mlpackage/Data/com.apple.CoreML/model.mlmodel +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:6459b9564e0630f2eec300eb732fceccbc1d2d16f12cb0694ce310d84bfbecf2 -size 3366 diff --git a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_ctc_head.mlpackage/Data/com.apple.CoreML/weights/weight.bin b/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_ctc_head.mlpackage/Data/com.apple.CoreML/weights/weight.bin deleted file mode 100644 index 23cedd570125191064c6111997f08c48898245f7..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_ctc_head.mlpackage/Data/com.apple.CoreML/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:fb9bead064427ffcb7529c0e3f378e421b4dde8e6d81447b6d1ca3352ca850e1 -size 1051842 diff --git a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_ctc_head.mlpackage/Manifest.json b/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_ctc_head.mlpackage/Manifest.json deleted file mode 100644 index 6fa9ee67551a75a4b78fabbd02c58c0b39b247ed..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_ctc_head.mlpackage/Manifest.json +++ /dev/null @@ -1,18 +0,0 @@ -{ - "fileFormatVersion": "1.0.0", - "itemInfoEntries": { - "6651E3CE-C3ED-4267-AAC3-5A772FC3515A": { - "author": "com.apple.CoreML", - "description": "CoreML Model Specification", - "name": "model.mlmodel", - "path": "com.apple.CoreML/model.mlmodel" - }, - "A3F7798B-67CA-418C-B8BB-58731D3A413F": { - "author": "com.apple.CoreML", - "description": "CoreML Model Weights", - "name": "weights", - "path": "com.apple.CoreML/weights" - } - }, - "rootModelIdentifier": "6651E3CE-C3ED-4267-AAC3-5A772FC3515A" -} diff --git a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_decoder.mlpackage/Data/com.apple.CoreML/model.mlmodel b/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_decoder.mlpackage/Data/com.apple.CoreML/model.mlmodel deleted file mode 100644 index 44a413afd31f449824ec8abc34caed9f15d411c3..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_decoder.mlpackage/Data/com.apple.CoreML/model.mlmodel +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:a05548eb455c5cd564782b125a5f9279a789be1f4141e5f044453ea79cd68b47 -size 6729 diff --git a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_decoder.mlpackage/Data/com.apple.CoreML/weights/weight.bin b/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_decoder.mlpackage/Data/com.apple.CoreML/weights/weight.bin deleted file mode 100644 index 4543ad4ea89b21ebed0de0d5ec3975d07eace7fb..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_decoder.mlpackage/Data/com.apple.CoreML/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:dd90b58597ee2c172c672dffe13b1110898ba07394c1a15efc96cc8c6b18411b -size 7871040 diff --git a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_decoder.mlpackage/Manifest.json b/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_decoder.mlpackage/Manifest.json deleted file mode 100644 index f969ba20230983a15c01e6a386db029d1a892816..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_decoder.mlpackage/Manifest.json +++ /dev/null @@ -1,18 +0,0 @@ -{ - "fileFormatVersion": "1.0.0", - "itemInfoEntries": { - "83E7B87A-4EBE-48BF-BF3C-EE74DEA4C7AF": { - "author": "com.apple.CoreML", - "description": "CoreML Model Specification", - "name": "model.mlmodel", - "path": "com.apple.CoreML/model.mlmodel" - }, - "98BF03AC-26AF-410B-95AC-C9B99B3B240C": { - "author": "com.apple.CoreML", - "description": "CoreML Model Weights", - "name": "weights", - "path": "com.apple.CoreML/weights" - } - }, - "rootModelIdentifier": "83E7B87A-4EBE-48BF-BF3C-EE74DEA4C7AF" -} diff --git a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_encoder.mlpackage/Data/com.apple.CoreML/model.mlmodel b/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_encoder.mlpackage/Data/com.apple.CoreML/model.mlmodel deleted file mode 100644 index ecd864b1e23052bed6937545ee41e528f892980c..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_encoder.mlpackage/Data/com.apple.CoreML/model.mlmodel +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:70d7747b57beba0248fabb6cbfa5d276e3604d0d7e234f4ccb578ea0a4d25110 -size 508107 diff --git a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_encoder.mlpackage/Data/com.apple.CoreML/weights/weight.bin b/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_encoder.mlpackage/Data/com.apple.CoreML/weights/weight.bin deleted file mode 100644 index c5bffa9667270bb21c093e55898f4fb2e299e426..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_encoder.mlpackage/Data/com.apple.CoreML/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:cecf7994b2758397d992802a4f6e5d656e3a1aeb7bbedc2aa430b1316d62474c -size 215143424 diff --git a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_encoder.mlpackage/Manifest.json b/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_encoder.mlpackage/Manifest.json deleted file mode 100644 index 5da111c5105a7bcf84de61da5c96b86c946acf57..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_encoder.mlpackage/Manifest.json +++ /dev/null @@ -1,18 +0,0 @@ -{ - "fileFormatVersion": "1.0.0", - "itemInfoEntries": { - "82FE91E6-784B-4D60-921C-8CAE58385192": { - "author": "com.apple.CoreML", - "description": "CoreML Model Weights", - "name": "weights", - "path": "com.apple.CoreML/weights" - }, - "DCA53FEB-DFA1-45B9-9E10-E5D3F5C04877": { - "author": "com.apple.CoreML", - "description": "CoreML Model Specification", - "name": "model.mlmodel", - "path": "com.apple.CoreML/model.mlmodel" - } - }, - "rootModelIdentifier": "DCA53FEB-DFA1-45B9-9E10-E5D3F5C04877" -} diff --git a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint.mlpackage/Data/com.apple.CoreML/model.mlmodel b/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint.mlpackage/Data/com.apple.CoreML/model.mlmodel deleted file mode 100644 index 40cb45ad4abce1672fdfcd80332f638fe690d4bc..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint.mlpackage/Data/com.apple.CoreML/model.mlmodel +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:7d115951cfb00e12b26f4b1a153e39dceffad34f5c940ede2ddb92d7a21859aa -size 4548 diff --git a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint.mlpackage/Data/com.apple.CoreML/weights/weight.bin b/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint.mlpackage/Data/com.apple.CoreML/weights/weight.bin deleted file mode 100644 index 747ec9d24e2d3e008bc869d2c404bb3d2653325c..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint.mlpackage/Data/com.apple.CoreML/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:b3f771cb65b190f1873e39629676ed79b65a8361522f451b37bdba8b1106e6ff -size 2798028 diff --git a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint.mlpackage/Manifest.json b/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint.mlpackage/Manifest.json deleted file mode 100644 index 51666c08cb1568e77793f3f8b7df68a1ab931cd8..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint.mlpackage/Manifest.json +++ /dev/null @@ -1,18 +0,0 @@ -{ - "fileFormatVersion": "1.0.0", - "itemInfoEntries": { - "16695A4B-7F7A-41C7-9F44-8306983C0AA4": { - "author": "com.apple.CoreML", - "description": "CoreML Model Specification", - "name": "model.mlmodel", - "path": "com.apple.CoreML/model.mlmodel" - }, - "E3F73234-AB5F-41AB-9362-57ADB67197EA": { - "author": "com.apple.CoreML", - "description": "CoreML Model Weights", - "name": "weights", - "path": "com.apple.CoreML/weights" - } - }, - "rootModelIdentifier": "16695A4B-7F7A-41C7-9F44-8306983C0AA4" -} diff --git a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint_decision.mlpackage/Data/com.apple.CoreML/model.mlmodel b/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint_decision.mlpackage/Data/com.apple.CoreML/model.mlmodel deleted file mode 100644 index 49e6f595a1328fbcf9ffafd61b2de9c7b27f5b17..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint_decision.mlpackage/Data/com.apple.CoreML/model.mlmodel +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:02bb6ce108de2737a945f3d7aa71db5bd710ff75282565ba20741d82000347d6 -size 8711 diff --git a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint_decision.mlpackage/Data/com.apple.CoreML/weights/weight.bin b/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint_decision.mlpackage/Data/com.apple.CoreML/weights/weight.bin deleted file mode 100644 index 747ec9d24e2d3e008bc869d2c404bb3d2653325c..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint_decision.mlpackage/Data/com.apple.CoreML/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:b3f771cb65b190f1873e39629676ed79b65a8361522f451b37bdba8b1106e6ff -size 2798028 diff --git a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint_decision.mlpackage/Manifest.json b/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint_decision.mlpackage/Manifest.json deleted file mode 100644 index c00fe89cce0df115b26086dfee3c450663c06e4f..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint_decision.mlpackage/Manifest.json +++ /dev/null @@ -1,18 +0,0 @@ -{ - "fileFormatVersion": "1.0.0", - "itemInfoEntries": { - "0A2C8DEE-CF15-4176-9EF2-786EF0484092": { - "author": "com.apple.CoreML", - "description": "CoreML Model Weights", - "name": "weights", - "path": "com.apple.CoreML/weights" - }, - "2EF6A0AB-12F9-4ED0-84E2-94A0D744E07D": { - "author": "com.apple.CoreML", - "description": "CoreML Model Specification", - "name": "model.mlmodel", - "path": "com.apple.CoreML/model.mlmodel" - } - }, - "rootModelIdentifier": "2EF6A0AB-12F9-4ED0-84E2-94A0D744E07D" -} diff --git a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint_decision_single_step.mlpackage/Data/com.apple.CoreML/model.mlmodel b/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint_decision_single_step.mlpackage/Data/com.apple.CoreML/model.mlmodel deleted file mode 100644 index 24daa1ae01d93329a3e31b992186474b6521c4e6..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint_decision_single_step.mlpackage/Data/com.apple.CoreML/model.mlmodel +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:6e47db8f5e51f21028004c1fe76f4b2f1430ccf82b9613fe7dbe9d087478aeba -size 10620 diff --git a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint_decision_single_step.mlpackage/Data/com.apple.CoreML/weights/weight.bin b/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint_decision_single_step.mlpackage/Data/com.apple.CoreML/weights/weight.bin deleted file mode 100644 index 747ec9d24e2d3e008bc869d2c404bb3d2653325c..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint_decision_single_step.mlpackage/Data/com.apple.CoreML/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:b3f771cb65b190f1873e39629676ed79b65a8361522f451b37bdba8b1106e6ff -size 2798028 diff --git a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint_decision_single_step.mlpackage/Manifest.json b/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint_decision_single_step.mlpackage/Manifest.json deleted file mode 100644 index a3794d1952292c241c090386207789e999e901f5..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint_decision_single_step.mlpackage/Manifest.json +++ /dev/null @@ -1,18 +0,0 @@ -{ - "fileFormatVersion": "1.0.0", - "itemInfoEntries": { - "BDFCDB9F-5819-45ED-B191-3F50652482FC": { - "author": "com.apple.CoreML", - "description": "CoreML Model Weights", - "name": "weights", - "path": "com.apple.CoreML/weights" - }, - "CAC241D9-FB01-43BD-AFA9-1285CC5E9E24": { - "author": "com.apple.CoreML", - "description": "CoreML Model Specification", - "name": "model.mlmodel", - "path": "com.apple.CoreML/model.mlmodel" - } - }, - "rootModelIdentifier": "CAC241D9-FB01-43BD-AFA9-1285CC5E9E24" -} diff --git a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_mel_encoder.mlpackage/Data/com.apple.CoreML/model.mlmodel b/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_mel_encoder.mlpackage/Data/com.apple.CoreML/model.mlmodel deleted file mode 100644 index 64a97d13c860d2e7021389f6e3108ad2d16d8a1f..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_mel_encoder.mlpackage/Data/com.apple.CoreML/model.mlmodel +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:aea604d3edc6e879e525d8ec2e0eb8413c5d3d4a9fae86233cbe4ba23d89a258 -size 1241478 diff --git a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_mel_encoder.mlpackage/Data/com.apple.CoreML/weights/weight.bin b/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_mel_encoder.mlpackage/Data/com.apple.CoreML/weights/weight.bin deleted file mode 100644 index 6b1f64e4316b232d108b8ca7402ffb8071a98a93..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_mel_encoder.mlpackage/Data/com.apple.CoreML/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:a1c90b88b52667ee4f349b39994003cc2a9fb1b12c310752399d90c880286a7b -size 215951360 diff --git a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_mel_encoder.mlpackage/Manifest.json b/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_mel_encoder.mlpackage/Manifest.json deleted file mode 100644 index adbaa38b4e0eda5372704a0880329c85f2729354..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_mel_encoder.mlpackage/Manifest.json +++ /dev/null @@ -1,18 +0,0 @@ -{ - "fileFormatVersion": "1.0.0", - "itemInfoEntries": { - "BCB3EEA3-98D8-45A1-9DF8-D4B55B9897A5": { - "author": "com.apple.CoreML", - "description": "CoreML Model Weights", - "name": "weights", - "path": "com.apple.CoreML/weights" - }, - "E71B04CC-8B97-4E78-A14F-E65CA1517841": { - "author": "com.apple.CoreML", - "description": "CoreML Model Specification", - "name": "model.mlmodel", - "path": "com.apple.CoreML/model.mlmodel" - } - }, - "rootModelIdentifier": "E71B04CC-8B97-4E78-A14F-E65CA1517841" -} diff --git a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_preprocessor.mlpackage/Data/com.apple.CoreML/model.mlmodel b/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_preprocessor.mlpackage/Data/com.apple.CoreML/model.mlmodel deleted file mode 100644 index 888c5ddc608dc69393977dc0334df8848b29537f..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_preprocessor.mlpackage/Data/com.apple.CoreML/model.mlmodel +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:04c6e8c1d678969c477d45d0bf8b231aa83c3a7ca0892d7820f701d18392e777 -size 25277 diff --git a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_preprocessor.mlpackage/Data/com.apple.CoreML/weights/weight.bin b/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_preprocessor.mlpackage/Data/com.apple.CoreML/weights/weight.bin deleted file mode 100644 index d0c6ac87e5f78315662149af9ab4ea3658497dbe..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_preprocessor.mlpackage/Data/com.apple.CoreML/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:c062338de852a26607ce4101f74e6895de3a4134a57b07232bd72bfc6f1d7f1a -size 567712 diff --git a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_preprocessor.mlpackage/Manifest.json b/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_preprocessor.mlpackage/Manifest.json deleted file mode 100644 index f490745b3939f47130ed35b5ed897f1ec76acb45..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_preprocessor.mlpackage/Manifest.json +++ /dev/null @@ -1,18 +0,0 @@ -{ - "fileFormatVersion": "1.0.0", - "itemInfoEntries": { - "4EC5BFFD-7A9E-42E4-86D0-365412CFD8E5": { - "author": "com.apple.CoreML", - "description": "CoreML Model Weights", - "name": "weights", - "path": "com.apple.CoreML/weights" - }, - "7A405622-5F02-4930-AB77-D20FA3B49A18": { - "author": "com.apple.CoreML", - "description": "CoreML Model Specification", - "name": "model.mlmodel", - "path": "com.apple.CoreML/model.mlmodel" - } - }, - "rootModelIdentifier": "7A405622-5F02-4930-AB77-D20FA3B49A18" -} diff --git a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/vocab.json b/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/vocab.json deleted file mode 100644 index b3436e3ca1609c34360bbdf9fa4dd50cd3eb1a3d..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/vocab.json +++ /dev/null @@ -1,1030 +0,0 @@ -{ - "vocab_size": 1024, - "blank_id": 1024, - "tokens": [ - "", - "▁t", - "▁th", - "▁a", - "in", - "re", - "▁the", - "▁w", - "▁s", - "▁o", - "er", - "ou", - "at", - "nd", - "it", - "▁h", - "▁c", - "▁b", - "is", - "en", - "on", - "ing", - "▁f", - "▁to", - "▁m", - "es", - "▁p", - "or", - "an", - "▁d", - "ll", - "▁I", - "ed", - "▁and", - "▁l", - "▁of", - "▁in", - "▁y", - "ar", - "▁g", - "▁you", - "as", - "om", - "▁n", - "ve", - "▁that", - "le", - "ic", - "us", - "ow", - "et", - "al", - "▁e", - "ut", - "▁it", - "ot", - "▁be", - "▁T", - "ion", - "▁is", - "▁wh", - "▁re", - "▁on", - "▁we", - "ent", - "▁A", - "ay", - "▁ha", - "▁Th", - "id", - "▁S", - "ac", - "gh", - "ver", - "ke", - "▁for", - "im", - "ly", - "ur", - "ld", - "▁he", - "▁st", - "all", - "ro", - "st", - "se", - "ct", - "ith", - "ir", - "am", - "▁this", - "if", - "▁W", - "oo", - "ri", - "▁was", - "ght", - "▁u", - "▁with", - "ad", - "ch", - "▁se", - "▁k", - "▁an", - "▁The", - "▁li", - "▁do", - "▁B", - "▁have", - "▁as", - "th", - "▁are", - "▁sh", - "ust", - "ce", - "ally", - "ill", - "▁H", - "▁j", - "ter", - "▁go", - "▁And", - "ation", - "▁C", - "▁so", - "ome", - "▁not", - "op", - "il", - "ore", - "▁ne", - "▁can", - "▁me", - "▁at", - "ould", - "ant", - "▁M", - "▁like", - "ere", - "▁they", - "ra", - "ers", - "▁ab", - "▁de", - "▁kn", - "ge", - "▁Y", - "▁ch", - "ul", - "pp", - "▁or", - "▁al", - "▁con", - "▁com", - "ess", - "▁su", - "out", - "▁your", - "▁So", - "ate", - "▁one", - "▁all", - "▁ex", - "est", - "▁fr", - "▁just", - "▁pro", - "▁know", - "▁O", - "ain", - "▁but", - "ol", - "ive", - "▁v", - "use", - "very", - "art", - "qu", - "▁my", - "el", - "▁N", - "nt", - "▁It", - "▁what", - "ab", - "▁P", - "▁wor", - "▁out", - "▁there", - "▁up", - "um", - "▁from", - "pe", - "▁tw", - "▁r", - "and", - "ight", - "ort", - "un", - "▁L", - "ist", - "▁about", - "ide", - "ig", - "ake", - "▁D", - "em", - "os", - "king", - "rou", - "ind", - "our", - "res", - "▁We", - "▁get", - "▁E", - "▁G", - "ack", - "▁le", - "ity", - "od", - "▁F", - "ard", - "▁pl", - "▁our", - "▁int", - "ment", - "▁will", - "ies", - "▁by", - "ink", - "ca", - "▁if", - "red", - "her", - "ie", - "▁us", - "▁some", - "▁don", - "ven", - "ood", - "ast", - "▁R", - "▁his", - "▁tim", - "▁tr", - "▁more", - "ich", - "ous", - "ame", - "▁going", - "▁had", - "▁them", - "ook", - "▁pe", - "▁Wh", - "▁You", - "▁But", - "ine", - "▁here", - "▁would", - "cause", - "right", - "so", - "ost", - "ure", - "▁has", - "ect", - "▁think", - "▁fe", - "ong", - "▁see", - "▁when", - "▁who", - "▁were", - "▁really", - "▁their", - "▁want", - "one", - "ople", - "▁then", - "▁time", - "▁sa", - "ap", - "▁te", - "▁He", - "▁ye", - "ck", - "▁her", - "▁thing", - "▁right", - "▁which", - "itt", - "ice", - "act", - "▁people", - "ty", - "▁two", - "▁J", - "▁im", - "ther", - "ci", - "ose", - "▁cl", - "▁qu", - "▁man", - "▁also", - "ree", - "▁en", - "ud", - "▁how", - "reat", - "ak", - "hing", - "ag", - "▁any", - "ff", - "ace", - "per", - "▁because", - "▁very", - "own", - "▁ad", - "▁act", - "▁been", - "▁now", - "▁ag", - "▁into", - "▁comp", - "ars", - "ions", - "are", - "ite", - "iv", - "▁these", - "ays", - "ep", - "▁This", - "▁she", - "ans", - "ah", - "een", - "▁over", - "ry", - "▁lo", - "age", - "▁pr", - "▁sp", - "ue", - "▁co", - "ick", - "ber", - "▁did", - "ip", - "ach", - "▁back", - "▁no", - "▁cont", - "▁other", - "▁every", - "pt", - "▁need", - "▁him", - "▁U", - "▁In", - "▁work", - "irst", - "▁part", - "▁look", - "ittle", - "ble", - "iz", - "▁un", - "▁make", - "omet", - "nder", - "ish", - "na", - "▁little", - "▁off", - "▁than", - "▁got", - "ually", - "▁per", - "▁good", - "▁way", - "▁could", - "▁ac", - "▁imp", - "able", - "▁where", - "iff", - "▁That", - "▁res", - "ount", - "pl", - "ance", - "▁first", - "▁ro", - "▁pre", - "ass", - "▁say", - "int", - "ated", - "ire", - "uch", - "ase", - "▁somet", - "ound", - "▁down", - "▁diff", - "sel", - "▁gu", - "▁am", - "ress", - "▁lot", - "ence", - "▁dis", - "orm", - "ix", - "▁po", - "ving", - "enty", - "▁K", - "▁spe", - "und", - "he", - "▁much", - "▁ar", - "round", - "▁app", - "co", - "ark", - "▁new", - "ater", - "ult", - "end", - "▁even", - "▁start", - "ations", - "rough", - "ile", - "fter", - "▁well", - "be", - "▁They", - "▁three", - "ign", - "ild", - "▁said", - "ough", - "ang", - "▁too", - "ade", - "▁bl", - "ens", - "▁inc", - "ia", - "▁those", - "▁mo", - "▁take", - "▁through", - "▁fl", - "▁kind", - "▁things", - "▁bet", - "▁only", - "▁St", - "▁let", - "cess", - "▁Ch", - "ary", - "vel", - "▁If", - "xt", - "other", - "av", - "ical", - "ord", - "▁again", - "▁something", - "onna", - "fore", - "▁may", - "ting", - "▁bu", - "▁differe", - "urn", - "▁gonna", - "▁does", - "uct", - "og", - "▁twenty", - "▁gr", - "▁Ye", - "wn", - "▁should", - "▁comm", - "ition", - "▁under", - "▁hel", - "ory", - "▁fo", - "▁use", - "igh", - "ife", - "▁actually", - "▁tal", - "▁call", - "ents", - "ious", - "ull", - "▁There", - "▁Yeah", - "▁most", - "▁ke", - "ors", - "ved", - "ys", - "▁sc", - "▁happ", - "ope", - "▁help", - "atch", - "▁What", - "▁rem", - "ple", - "▁Now", - "▁br", - "ool", - "oth", - "▁four", - "self", - "▁str", - "ne", - "thing", - "▁put", - "ial", - "▁great", - "ail", - "ub", - "ning", - "▁sm", - "▁feel", - "▁five", - "ody", - "undred", - "iss", - "ank", - "get", - "aking", - "▁many", - "▁hundred", - "▁years", - "▁being", - "▁come", - "▁mean", - "ily", - "▁different", - "▁after", - "▁ser", - "▁show", - "form", - "ful", - "oy", - "▁six", - "▁vide", - "▁V", - "▁its", - "▁point", - "▁day", - "▁des", - "ons", - "▁bit", - "▁bel", - "▁before", - "▁aw", - "▁end", - "▁Oh", - "▁still", - "ath", - "▁long", - "▁'", - "ise", - "ob", - "day", - "▁add", - "ft", - "ves", - "ces", - "ady", - "▁cr", - "▁around", - "▁try", - "les", - "vers", - "kay", - "ian", - "ates", - "▁find", - "ward", - "▁As", - "▁eight", - "lic", - "▁same", - "▁pos", - "▁em", - "▁made", - "▁supp", - "▁life", - "▁Be", - "pect", - "▁dec", - "▁play", - "ange", - "▁att", - "▁pers", - "ways", - "▁high", - "▁hand", - "▁next", - "▁cons", - "▁own", - "▁inv", - "ower", - "▁ind", - "ert", - "ng", - "ave", - "▁year", - "▁big", - "ating", - "▁world", - "▁rel", - "▁sure", - "▁tra", - "ew", - "ered", - "▁fin", - "▁Well", - "▁sl", - "▁doing", - "bs", - "▁set", - "▁rec", - "ual", - "cial", - "▁ph", - "erm", - "▁love", - "ph", - "▁real", - "▁last", - "ict", - "▁bo", - "▁ra", - "ible", - "▁wr", - "mer", - "▁count", - "ities", - "▁always", - "inet", - "ments", - "uc", - "▁might", - "▁inter", - "▁video", - "gin", - "▁tell", - "▁never", - "vent", - "▁import", - "ied", - "▁sy", - "▁How", - "ically", - "ought", - "▁thir", - "▁rep", - "ks", - "ib", - "▁fam", - "ject", - "▁bas", - "▁She", - "▁give", - "akes", - "▁ninet", - "▁reg", - "▁min", - "▁op", - "▁def", - "▁didn", - "te", - "▁cour", - "▁why", - "▁ent", - "▁place", - "▁ins", - "▁car", - "ather", - "▁person", - "ular", - "▁inst", - "▁prod", - "lect", - "▁Al", - "▁today", - "▁bec", - "▁sur", - "▁All", - "▁another", - "▁bus", - "▁keep", - "ell", - "ese", - "riend", - "▁quest", - "▁talk", - "als", - "ings", - "▁mon", - "cond", - "old", - "▁acc", - "▁la", - "▁num", - "ident", - "▁che", - "iness", - "▁turn", - "▁ear", - "▁No", - "ousand", - "▁better", - "ific", - "▁loo", - "▁gl", - "oc", - "▁important", - "ited", - "▁An", - "▁thousand", - "ility", - "llow", - "▁used", - "▁gen", - "▁sim", - "li", - "▁happen", - "▁Un", - "▁Let", - "air", - "ock", - "ably", - "gg", - "▁watch", - "▁For", - "▁sw", - "ren", - "ute", - "ever", - "▁pol", - "▁sch", - "▁When", - "▁such", - "▁fif", - "▁home", - "▁cle", - "▁contin", - "ouse", - "▁friend", - "uring", - "▁Okay", - "gr", - "▁able", - "▁stud", - "▁eff", - "hip", - "body", - "▁top", - "ness", - "▁exper", - "▁pret", - "▁both", - "▁done", - "cri", - "▁mark", - "▁while", - "▁old", - "ros", - "ont", - "▁second", - "ative", - "▁thought", - "▁best", - "▁found", - "iew", - "▁belie", - "▁each", - "erest", - "▁tri", - "▁eas", - "▁ca", - "▁fact", - "▁care", - "▁fun", - "atter", - "ures", - "▁head", - "▁lear", - "▁water", - "▁hard", - "▁few", - "▁side", - "ween", - "▁exp", - "▁away", - "its", - "▁ext", - "lud", - "▁run", - "▁trans", - "ince", - "▁sk", - "▁open", - "cus", - "▁between", - "▁called", - "▁wee", - "▁pretty", - "ason", - "▁far", - "ember", - "omm", - "▁interest", - "any", - "ner", - "uff", - "▁pres", - "▁cur", - "▁child", - "ee", - "▁toget", - "▁together", - "olog", - "▁God", - "ond", - "▁char", - "▁looking", - "stem", - "az", - "cent", - "▁ob", - "▁ass", - "land", - "▁doesn", - "▁business", - "▁course", - "▁ten", - "ps", - "arch", - "ced", - "ms", - "ize", - "nce", - "▁ref", - "▁name", - "ross", - "▁grow", - "oney", - "▁went", - "ics", - "teen", - "▁cou", - "▁prob", - "▁ret", - "▁guys", - "▁came", - "ash", - "led", - "▁Eur", - "ues", - "▁ide", - "gan", - "▁everything", - "▁getting", - "▁ask", - "▁cor", - "▁build", - "▁sign", - "▁small", - "uck", - "▁el", - "▁col", - "▁Is", - "ational", - "stand", - "cy", - "▁conf", - "der", - "▁bre", - "▁cap", - "▁mod", - "ets", - "ike", - "▁number", - "▁comple", - "ertain", - "▁ever", - "▁coll", - "▁hum", - "▁Europe", - "▁cre", - "▁met", - "▁exam", - "▁move", - "▁pass", - "▁left", - "▁system", - "▁includ", - "▁Thank", - "cept", - "▁wom", - "▁product", - "ten", - "▁rest", - "▁probably", - "▁dri", - "▁Do", - "▁gener", - "▁anything", - "▁lar", - "▁My", - "▁school", - "▁lead", - "▁sub", - "▁ty", - "▁plan", - "▁seem", - "▁whole", - "irect", - "▁light", - "▁must", - "▁mom", - "▁opp", - "▁support", - "▁family", - "ices", - "amp", - "▁proble", - "▁dr", - "ready", - "▁using", - "ense", - "▁prov", - "ush", - "ax", - "▁power", - "▁Re", - "alth", - "▁ev", - "▁stand", - "▁war", - "ts", - "▁", - "e", - "t", - "o", - "a", - "n", - "i", - "s", - "r", - "h", - "l", - "d", - "u", - "c", - "m", - "y", - "g", - "w", - "f", - "p", - ".", - "b", - ",", - "v", - "k", - "'", - "I", - "T", - "A", - "S", - "x", - "W", - "j", - "B", - "C", - "H", - "?", - "M", - "O", - "Y", - "N", - "P", - "E", - "q", - "L", - "D", - "z", - "G", - "F", - "R", - "!", - "J", - "U", - "K", - "V", - "Q", - "Z", - "X" - ] -} \ No newline at end of file diff --git a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_ctc_coreml/.DS_Store b/convert/parakeet-tdt-ctc-110m/coreml/parakeet_ctc_coreml/.DS_Store deleted file mode 100644 index 7b37950bb8ce3a439220b30b68fcd5a01d75cb8d..0000000000000000000000000000000000000000 Binary files a/convert/parakeet-tdt-ctc-110m/coreml/parakeet_ctc_coreml/.DS_Store and /dev/null differ diff --git a/convert/parakeet-tdt-ctc-110m/get-pip.py b/convert/parakeet-tdt-ctc-110m/get-pip.py deleted file mode 100644 index 22f29f881487fe04461440ad6bc7b6a2e5e781de..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/get-pip.py +++ /dev/null @@ -1,27368 +0,0 @@ -#!/usr/bin/env python -# -# Hi There! -# -# You may be wondering what this giant blob of binary data here is, you might -# even be worried that we're up to something nefarious (good for you for being -# paranoid!). This is a base85 encoding of a zip file, this zip file contains -# an entire copy of pip (version 25.3). -# -# Pip is a thing that installs packages, pip itself is a package that someone -# might want to install, especially if they're looking to run this get-pip.py -# script. Pip has a lot of code to deal with the security of installing -# packages, various edge cases on various platforms, and other such sort of -# "tribal knowledge" that has been encoded in its code base. Because of this -# we basically include an entire copy of pip inside this blob. We do this -# because the alternatives are attempt to implement a "minipip" that probably -# doesn't do things correctly and has weird edge cases, or compress pip itself -# down into a single file. -# -# If you're wondering how this is created, it is generated using -# `scripts/generate.py` in https://github.com/pypa/get-pip. - -import sys - -this_python = sys.version_info[:2] -min_version = (3, 9) -if this_python < min_version: - message_parts = [ - "This script does not work on Python {}.{}.".format(*this_python), - "The minimum supported Python version is {}.{}.".format(*min_version), - "Please use https://bootstrap.pypa.io/pip/{}.{}/get-pip.py instead.".format(*this_python), - ] - print("ERROR: " + " ".join(message_parts)) - sys.exit(1) - - -import os.path -import pkgutil -import shutil -import tempfile -import argparse -import importlib -from base64 import b85decode - - -def include_setuptools(args): - """ - Install setuptools only if absent, not excluded and when using Python <3.12. - """ - cli = not args.no_setuptools - env = not os.environ.get("PIP_NO_SETUPTOOLS") - absent = not importlib.util.find_spec("setuptools") - python_lt_3_12 = this_python < (3, 12) - return cli and env and absent and python_lt_3_12 - - -def include_wheel(args): - """ - Install wheel only if absent, not excluded and when using Python <3.12. - """ - cli = not args.no_wheel - env = not os.environ.get("PIP_NO_WHEEL") - absent = not importlib.util.find_spec("wheel") - python_lt_3_12 = this_python < (3, 12) - return cli and env and absent and python_lt_3_12 - - -def determine_pip_install_arguments(): - pre_parser = argparse.ArgumentParser() - pre_parser.add_argument("--no-setuptools", action="store_true") - pre_parser.add_argument("--no-wheel", action="store_true") - pre, args = pre_parser.parse_known_args() - - args.append("pip") - - if include_setuptools(pre): - args.append("setuptools") - - if include_wheel(pre): - args.append("wheel") - - return ["install", "--upgrade", "--force-reinstall"] + args - - -def monkeypatch_for_cert(tmpdir): - """Patches `pip install` to provide default certificate with the lowest priority. - - This ensures that the bundled certificates are used unless the user specifies a - custom cert via any of pip's option passing mechanisms (config, env-var, CLI). - - A monkeypatch is the easiest way to achieve this, without messing too much with - the rest of pip's internals. - """ - from pip._internal.commands.install import InstallCommand - - # We want to be using the internal certificates. - cert_path = os.path.join(tmpdir, "cacert.pem") - with open(cert_path, "wb") as cert: - cert.write(pkgutil.get_data("pip._vendor.certifi", "cacert.pem")) - - install_parse_args = InstallCommand.parse_args - - def cert_parse_args(self, args): - if not self.parser.get_default_values().cert: - # There are no user provided cert -- force use of bundled cert - self.parser.defaults["cert"] = cert_path # calculated above - return install_parse_args(self, args) - - InstallCommand.parse_args = cert_parse_args - - -def bootstrap(tmpdir): - monkeypatch_for_cert(tmpdir) - - # Execute the included pip and use it to install the latest pip and - # any user-requested packages from PyPI. - from pip._internal.cli.main import main as pip_entry_point - args = determine_pip_install_arguments() - sys.exit(pip_entry_point(args)) - - -def main(): - tmpdir = None - try: - # Create a temporary working directory - tmpdir = tempfile.mkdtemp() - - # Unpack the zipfile into the temporary directory - pip_zip = os.path.join(tmpdir, "pip.zip") - with open(pip_zip, "wb") as fp: - fp.write(b85decode(DATA.replace(b"\n", b""))) - - # Add the zipfile to sys.path so that we can import it - sys.path.insert(0, pip_zip) - - # Run the bootstrap - bootstrap(tmpdir=tmpdir) - finally: - # Clean up our temporary working directory - if tmpdir: - shutil.rmtree(tmpdir, ignore_errors=True) - - -DATA = b""" -P)h>@6aWAK2ms3fSz9=Nlw0@!003bD000jF003}la4%n9X>MtBUtcb8c|DL%OT<77#qaYeLNB_YQ}7R -JLBWgQMLc*D8D`sbJJ4o^By}nH;Y}-R7|8JQc>H)h=LtgSXPp^CfHalN3Xv#l)Rak_3*j4C>~Hr+sIG -4Pb>*Dvu!kuoI*)vi2F4`%Dav2)18n<}T4!jWSs$bTq|)*=0iTP-{H3s6e~1QY-O00;of09jja_pjHy0RRA2 -0{{RI0001RX>c!JUu|J&ZeL$6aCu!)OK;mS48HqU5b43r;JP^vOMxACEp{6QLy+m1h%E`C9MAjpBNe- -8r;{H19{ebpf{zJ27j)n8%0=-6Z#elILRo@w9oRWWbO{z8ujDS!QAC@3T%nJCf;1rX6ghzu#Z}R@K&*?Hgj1WFD91+adaM4G`4Xs@*hA^t@nbDYdL)-aOjsW~3}QVVby(8=@7U$ -Fzj5Y{w!2hUUH`?e9j7WDA;>-1aos>7j{2$~BfyL8p@__Y98dsP#Bs7^lWF -=e_gr;(4^?am?Cp93+7b-!?~nb}-$cPSR1zckA*zNp!)$;YjlZrfn&RWNM}=QA7*cb8A{(9@{5!vBfq -rEMoeu5FvJZngI@N#4#(2v$WnMGCAVD?b9t8W^qDfcFBe5ZZF%dPAPaq#>ikclG~yPvCg`JUGb_W2#PdCXxx}7!|T*xc9qdnTILbO-nAJaF2 -~0snMFDU<%E01X4*yW9@|}F2;vY~;0|XQR000O8%K%whB^=tD)dBzjss#W56#xJLaA|NaUte%(a4 -m9mZf<3AUtcb8d3{vDPTN2bz56Q$bEqvD7eOzLnyL~CDi@L_;ZRYu+Sp^V)ZR6_YZ?pj@13#Z1R`1=l -J)M)n>TOXIt;_f2D8Q^;6`S?Y{9RUgUr+|m;!25C-6tno(2iIDhjlyJ)nM4*651XX%H+qrBEdT{cBla -4$^`0^qPP-6zv*|ge-jzUzxn2=uGMl9#)iA)y8^Cdr~rxdixH}OOJhxFbsp>7(O2Tg09*VTBnRAqE#) -uTB%a`7P2*FzrkVV`K)SOhdyilnqJR#!6l}Q6a+(^)-m{nsTFZ3tf`=GYik||DD|c)gW1pJ_vy8mPk! -87%_j>OLv)_N=Qs$09E*XCaNb7Sbvyz%2H(~=0(GyA#Q^BB=o_mcOvCiSC>?bfF%-ta6OhP5HUX=GiK -PR!(uKJlo!!9~IAAmCk)?77i`J23la2CGx64oXMzMaXkQ<~~8EU?%I}z$$rRNujEIu~M()ri%^Gi%ri -C!gNA@cLO=l6KV$(xV^&hZYbU&TCtOIChO4g;gfcAY_>ak~kgLxGa?L$cMXVJ{&`q`lnqv$j}Cr3vW0 -iSMRu8%^e>;b`+HM=<$xdKPpu@6SuMN-LR>$cFaIP$0f`iClb~O`=>COC -NvJms>bV(-KMn=QG5PXY-h9vs;@fOIZ_lmTi^Fg`mulO!YZVO^KOIvSWtj)HD-~+vPht4%90 -Btz&yv-M$^(udyl?*`G;BgAr}tWa5RRHyc3Sz7-4H^#tC)@T$~!*>z3|0?b+NOYH~Nw+WUjLL%ySwnL -u=yutnX)PsolsJX_Jx&=d9p7_`6i5S -Mvz$qHBvD4Gc2vqMK2J#u@ySRoi8HJ74pBUZpQaDYr)B{xbde@6aWAK2ms3fSz8Tl@oYE&00931000>P003}la4%nJZggdGZ -eeUMUtei%X>?y-E^v8uQNc>YKn%UlSA_Ml1)W|5bwQ7w#FICXGTWwYU^+7-sY}6+H?6MX!CdkPkC&I1 -p7R7v)6Y6HHVx2JGAo3XvIeD`#JPUu6e_-52zR!@h!P7KI~18)C%HP@fr -1dD$kQ8NRuGKz%ZZysu2=G*Ual7)rq;5cIqz_b_FXoN_lu6z|qu{_j%fT!+RBl=guKIY1=QS5bb04|v -WA;eKlsTs@d!Jvgx1?RGCC*+Bw@QIOxwu-SziJ7_J091)~tDk`9(d78okVj;x!LdG5$Q)?DBIa2R7@M -sdD>u3!!MCee1<#q{z2%~C|LtPJ~<9zgOp6arcP+QP7iOeYV&Gp@_KO5Zof3NVEl$Vli`umm>uNm@}6 --N7T`WbHNRPGZ{O9KQH000080LuVbTZC7+>nRWb0C_h602%-Q0B~t=FJEbHbY*gGVQepAb!lv5UuAA~ -E^v9>8(nkUw((uR0-+yFrX-X2q0_m^wd0Sot*I+%BspzHqhW9)PGZd?Spt-FQT5-uiw_b2d1s}4N^i! -#Be+;B_Inpl5Cm`fvMR``zAL+?-m+Sdp0k2%nvRsbsi-KMniPFk);EL~B^P9kGvF}@f}^8N*KA3aZF< -pnEXzo_ZJSOITGx$`bNSJc9;=$08<=Ju8*YBJRNPkO+C1`7u;KS^fD-IM+;_B9OXf{gv0N@-);#SB*0 -JJUnTrWbO4qr8I~J^?>xwBLv1{3Y;Zw_TGnJ}@t*Rh5my`=<)FZL^~625o}pcd%eCnr;^pd<} -22FJ)bz*=$^OTO1Mi%peDF_MmsfKy%=6SmI2LzL$gh5Wv3i9}I8-dl?KxJ)T=!kr?ud!sb^GqNCbxgo -FCSF2L}s<$GFj7AcbP!w}kNz=9M2dc{Q-6Zr4?=;#RL2UIVOmq0cx(1t+%uVZ;W$zvt^(sYKMB~${b-kW -x-&tzv2jRGzH^>eNbWH%qa&}-UAD1jVzMy!IIE_RJ(M5)H)8L3KfOXk*)g;PdB_^c~da`o_t4w3}q}u -!n!O=+g&i2n^edh2k(?&Xx4s)YlUFaj;|f9P|z}v69cYj_{@6{+fMBo-=m$|SYrvc?giNZAckTBy7Y? -`lEUaaD_VuoFUZ=yrgyZgWLQT8As|kHt-w-{zJNtR4M^TGj3Q*cvqpb>^}S}$xnvLr*>+o6mA?WcY>z -Y>pzlfc_cV5x+~Ks3dP6*Q%$OBoOmrOe-uwRv1qb?qHR73dDXhX-l$ClBLUT`s_xY9I7AQej%Cbo?)z8!yJ@}VN@Fa*2 -8)Mzo25I~&8(5GO>NJ!E&9r$@M`0l-&_4%pNkQw%Y<>s2XGF?UkA7jG|+lHvUhi@udTDW=A_*#;X<&Y -^u`cg{XR!{?(90NBCS~;>{ZcerVuY=VQlfe|YyKb7wBmh3^+~}>7xuvGleVI+D1HkDR=;YP)S7!P&&L -HwZEFN1qyFCH&iu{XAmu9wSgpHrK2q=xm&i?Odhxq}>=J+godwOzldH(wNKwu;!=&kIY+^mzy1#E~7HFB$=oWFm8rU+}WGY$DiI=`%B4X19=)p!K=`!ReK2BGl* -zCM-Ap9Ndn1VqV;L{Ji1;YU7s&j+56S&%g^VlW#G>XULNA$4cIPt(q-`KkLjzxeFv#Q&z7zajDDwq%X -5{?LOZ0nIOd;w96{(xPP~7&$) -u2qN<}jf~C#nEVO!+j$tc@VST0{5363I*t^aXj%$1=|L@aThCiv?>(tC*DHC0Y}tahtw^MAgFrM1zWN -TdYwTEVz>?Y)qK6cUg&0?bGqCE?qq&+otGYzI0M+CX3VWG|~dS(6-@1rN8M6V>IMsgO{*7#Z&EkO00S -B8&H%Q5*@mRXaR)F4EXrB$|@P9t4s%al+FPEslWv!NZ#BkQD|QlAQ#e-4B!FApK`k) -=CU!fESO!qDGe%JVi5;U9v4#CYy;_-&K6bI@58?R*P|TMnswiS3y$321d}U@2{^TpYOUmD4|ej -7WfZ~>&1k4Z-qooPvoowX2F6M!fRXU?9zI`G^i8WtDNJh6`14~q{}D-d}TOxc;gAeB@zxO!tZOF&l`g -dikFAoVN4PPkk~V>T^euE3yn02t_P=BWH|g4GBs#|ps)shT}Lx?<^a}gvz&F}FhYbFiM6K9qpeDYG4< -!<<1(@KEe$9)8dkY -_4Hw#8%I9b~vkbBhz9ZvFFL=rcXFOhh%h6?kS0V!~)*<@#(*)Gm+u4S^Yc_A8#=4f~uIpzq*+zL9qKb -re^ykZI#!C7J+s6JyTkvQWKEzO<&RO03u!Jxg9e*Cv@N)79!IEXCzp_NQTwW@rf?fjz657p0BKW?E-* -;AA+Z+Q81o`TMoh&xLwQG23EiTRG0SoB2eb01Dimq?xU_*Qatl|3Z(y0BT|y~TXkqQqB%`maM#nM0&?9gPzhxd&zGB^^jCbGhCGQx8FKabpfKit= -yr{_*R|-aLH338GF-CSJtjLL}O~$u_MXnn~dsc(9~SrVt+4rU&%Sl1qB&?47IcBEhohMphBo6 -a^b9lXBnoe3)>Qt=)G3;1>?5HB^pm4+O+#+Z(&E=H^>~+nZ&JC2h=PmmW|9t -u2TdkUhv?4&bX;acnVGke&9=a$!_5r+%R39f5PBBmJ)T|;B}{E(2Gs-C( -GTU``ol%71^fpaBk`{&HbgcVCV(+6obLEcQ@RZiitXB#7j{Vs2Ait=#Yp>(KW3WW1_?eX|Hr@xO=ZhA -^FmkP{7j)<(%nT3)qvwpI3|P3A{WMlz{RilZkMe2HB@+adF8s)|Oj;E9jlxyxVy(5beoq(F -YuHrHE;D~BI1B2X6e@mHpj8uwMKfHQ;+*62)$GdhrM>)Q>g+VIr9Fq9Xkz1er)1Iu!*mN~(`eEY0%jE -Of70WV(>ab-8MnDtqp&0ayOF0Go*194fm12YS!kqtM4_ -$eOZx%qe?{)njRmuP7?o2nJXY38#oi7zr+1j^-(jZfpu#Y_@DGPn^hP{~;_}(+%a0h9&wjrAczSYj`u -SpYesO%c8pkBWh{8&gjsI{$r0qo-s6QMavx6(OuG4umz^;IGq>wSxK%~5_N?{Tb&f3U|P_1n18k2Kf| -MduFF{^J2>O1krsV+#07i3&j1sAC~)1-CEP+4c}1>A>y*p0WwSRT51^l<0^oIT_j-ho}no3K}Jz-g== -a@Jx4|JpevG0>0}L1Jse&4P*b!o5%XwIdf)!aLe29Ywjy{A(=UjiRI=xh?k(C$+ZdIo!iNBjUDc!(q9 -pNlbf%?))n~7t_mO;gRj~RKqTFxF-NS`0@|DM$YKVi%eX3*BGzk;mVcO7zL-KE-#T+_}Z%K6M05ZmFm -FUJ#5pLo5JCYC?@w5oK6u*;uI0wLWQ@0_M*WlHbKL0Gt18vY{L9^rNZY5K)JbTT=(u4hlf0@Mm}8JlA -;I(;9v?ZtL#1ZnnT}x+fS&^c&ip;c?eUW#f*7-}9B<~SM4S -b0G~bwY^LOUxYW+`bj*BDJqj=+2M8Kzs)UJGPK^WXtkfT9dm=Tu2!0f@k)y6mUdt8MzT6- -lXqrJS89Bn>V2?KcWlr(22as^Y{z4gvhLaBrThzpC2lB{GAr>W!@5T=6`ID$tVSR;2tB`^)?rK!_6z4 -AVR&G|rKN!&LR`uGH9&0s4q(q-2lLE~HZ92}cpREU%jKhu?rEBz%=@zE;r=Re%{!c84l=tG*+2ogw^6 -GHuX+%BrE5Hc6Rw7w>2EfpInPwp-p^5Yi5XJ2H2-Xre6q$Q%ub?JyFy-H_b?83+EJ)}ZSOF~k(RztWb<=ufSF%TSJm8yk -0pPW_+M=Jvee&W`|qSVZT${h=nssSIbQ?02^7y*K_}8<57&mIGxuli?*Y!S9tipk+5H6kcap@pbmw5} -HU7!Lb2j<^W^5fTA#*rs#PJsLD4fWORo68gWDWlTP)h>@6aWAK2ms3fSzE;-+kFQN003zy000&M003} -la4%nJZggdGZeeUMV_{=xWiD`e-CAvr+r|<8e!pU?zz9{yJlAmCR)C5+&N+7Bm%w)54~GM>5?9jtC@# -Y#PgdIg_dYYbB$py}J~vHTR4~rszRb@1JhRkfGI=UjTP0q}TyE4(<<(>|IXhdIW+#$lT~~Ffl0@iTVa -iI#JU5lBw8`z+nIGFqyUp~ndHiTi-h9u@W~1{>^JuK2TgZxbG(>;EqnoG>1(rACPx6Cjq|im2+^9S?W -n9SBwIr%>B{#NN`(AElLg$q#i&EillFOaykKCxzg7MoZ)|Jj$k}H{;T(4xNe^yK`WQGanGKv -&Au1;4fdoTwn}Bsbf$Rg$j+TfRc7NcjV)%;<4Wy -{1RS+$#j|6_l!uw1Y0M_qI#2CsDv+hs2H85P49RzPM*g5mv1lA4-l*y&k3|WqI7y~wXK&uV`2NM~eWI7Wnw-hkau -C(TrQ>ItC<9(>a?)1pUDqq1!(S+Aerqg(y(~NuEH8B}9`exG2xx7e#4qbgC?T -mSQ>e7SkA6n^L5*l7g*8ZiC1rQgh;e=XQ|E=i)uKmc~@1l?vZ^DsIkoyIAqCx2}>TvMO88LJDAuHUHY -=o?%vIUBJN8xZO8xr@+2~lOU;dWCS=iHYUf3wS}xvJJoHZqvLMNqQ9Na5BPax -zUV|L^h&M92NVrtF7E5&k{9fCkFb*8I>pwb)%1h!RG*!lVS1`^fF7>mz~Lm|&mJYdJH8PKz8RhOg}B -x>ZsiV`!7Oys_KJcB7KM<3;Bi#h|psQ|MwFi;hki=h9|J+`$KnL_4gx@sDW=Vq^NXj{)N)*EdnuMuA; -x|G^O2k9(GWrD;dF;)_PkS>CVr=WuWEyAMPDtdY%!R07po};cFS%bL8B-DRHP7B8-pOYBNx$u}Gpq61 -#*4w`)#wba-B|$=TWejXnay+Vmu%N4>W5ku{$El?`~Ib|p#nGfkqeWHRhbMRqL!j*m>1_BlV<9ziF&$ydjJj#8iX^#672D}m -fop!j_mr6uB7di)RU4sgP0M}<|J^r&mH2|K!fo}z3ahSH-zxDN{;@n5D>{Yl}<_AFtiqQeXMc)sEr?t -cgT(BZR4Q9?~T;1_NH5}bb=Fx4(=bq*=W3DMt*F}(%h|E@3DD~|i^1sQm16f2dP{t+-7#!R76y#iZI5 -=opWDrhMQ`U6o4XwQ1EipvRm*c6!rY5f>uoNYSjx5(uf1+Psc3w*9ug>*PwTi_>yop78?qe$i%BlGoJ -sf#i<5P5`E0DUw%A~Uffi-vvdmh9O;y9*BhzM?JNyUr3jioajLx`@X1ExW-D88g&lE+fYc8xogNv|;I -p)OIZ31lo4hEZBDzMo_NU_tgGny&?li_?y_N1lZS)pxNRsk^&TG}1(7r;%tdZX>PT -{QqHFrG~r)pVmqQNYECNDPV-zZAMoGtFTM2!CsnJAl+@y)4GuC3k4isMf~+5%Y?geW#kjZT0v2{V2V) -oeCOC@|aTeAaX3@Ds{ETrs_~en|+Y<2DeYen?D*>itT5%juA7R?xS)w)9+f;bnVpE;1>G=oaN2gKdODPpqTo_C>ZR0<|A78<+M%? -I0;nMXX~(E7#B0ZMt|qTa#Ys?CU!%!rpunBUsA9Lg4l{KwZzhx%v -Yf!^)MJ%^m{;_=(c0Jd+0B6SMc-h^@J~q*uwX%&0@DM>E3H3zHOxUJWtbhN}>G_Q`RYMd@dds*fn -{<(srUBUE0q{yc{=YljH)Trp5ot~&H2S2hG`@ej)%#G&VvuSWMGsO6Hx^P5aE6RL=3?< -Zfm^>Z%go+MzbWNPt!qRwLJbcozssBj$Xn5=zO1tw?p}*t$v -NSD5u_vBrbC|H${WK`s6+qz1?@5>3?Ah?90!9!;|+&z0(|Pzv#?Bx;vfu-|rX5OfHUtE4N_S1M(iGJ3 -6R$9T!rB^A6vwgizyXEZ7@&P^2+W$kcDHwyuWy?I-TJf_46~{qr|pNQf;lA{}#69i~{I2Xqz;Xk3@gH -1^W=6T?VRMLJ@iRHe+qcdefB&;rQpZSQ>)JN>di5-o=CKUDPMbPv$i_M-n7qMg90Ja;}BW<3`U{HIfX -KBihXyhY+nR5O+glUkbx+MPMP#TM*W-2dJK!I!oEKbo?=dYTwwp=qD%@pKA9%(AAo(2>SE;Fmt2IeW~ -evMs=#S55)fhNYz^a#~SOs-@p0p&79xwb!t~X)f%4ZUmoWpyKZ}NTL4F$|KY(mj~`JmAdhY6F7lX0>n -HT<0J@FoIVTMcxuwrtj76Wx8)qT*nvp=%k;GZWR&!4C$X$If^RU~I -S9zIc3#JCvQN1K6J?hO{h+zCRw7q>0b78;3D=-cr^(Z6x*O4xO9mzwUpgJj5;3#x!m>}NGA@MqXHdr- -(Z2vtO9KQH000080LuVbTkxXRe@YPm0O&XX03HAU0B~t=FJEbHbY*gGVQepBZ*FF3XLWL6bZKvHE^v9 -pTmNs{IF|q2e+5^?pf=V~ckUL48w0-G%p~mulO{#dc4iP5TE(JcZDmOiWvAY3?r*=(dn6@Nl9Oz)zy( -EPOXTDG{*5m|5d2i+tMqPL#dTWbViV`_o!rR07E0FJax}UwHd55G+N61r6?gmD=t$)8MvjCyR^q-&>s -sa_&Bc-diB+0O6=KmY&bGAbGyN8^QpGZjYnh1qv|fu?13403d0y1KUyVlQnFd?Nm6DZxmKDo5llEJo>?gHX8uYTrGEIvLU~KRn0Lq -dekv~H{l|SN4T5D_hEbYhQyyob{JNASgS1=5V2lu+Y)`AIsHXkrZy~Usifxu$6!nktyoeK-Oh=QUbGC -JwHAlo{nWU9ExGsb^%ec17e?7Z6x%~C|@Ny{EokR4Utk3ZToISW>ld6V)GFK!oU^K<&?PfH(itja@A6 -P(Q?#V0cz80^<^%{`Yah{BRN%I6749~{_eGjHW)zxG)`swoG?b*BN$K&hMi{tmFSQij>X`TimU0Fhf% -#_q`=-gm*dht1;_1DqayNk=K(-Ydbla+!D06SRAtc3o(5+9+lUvQuj#mZM*McNFVjw>0C^pZJKtHu54 -`t)iZY(Wj+fu2u9*L5kE=_+0Fig)LiKOUdYgS)J_jWfOKdar7^5x=I1sD&o3^OWM|WT8O-;HgT5zLy{jN2ym8(T#2Z8!End)-;6a*9Q_ado-q&Mq -u$3~r177m9x8?Wz_v=>#Z>~fvEHa+TP>v+ONX&V#ol>BevV)c>H-lAm3?#*g+6+@|PCEr1@0EDR%Vz@z|qa(q&kv$}FwNpnod9pBVN*1tZTbL>zO+%) -BIB);Hv?rY#>0BQ?5atKI{a-848BDulmYhkq~9d`zFFqT}i&Vg((dS0oA(077yV7Aid$lhZSbvl#I&c -L^PgbG_3L9Yd{A{WBRr2IXGvp90Y%Lw+qGRPu)7;D3Y3v{}ID^*wb+kK0upH1rE&4f6geOe|mMXooL` -Ee8lRLVsw^cYV=pNOsfCQgT!?anoh>qPvb<9oF?ZI_(l>wmw4dmKhAz*Wbn;{R}pfE;+elH*keo4L(% -G4q-;Y+62e)63j~dQWiNL0hl>3 -ok1@JceOb0wXQ+%$fi%N2g)yZowX-$#!09XrTnF?v8kYmYk;YEAjFK&U5VpzIHiK}hV5>pnb&iCC(g!zM1rh_T30JLpa -(y`BD(_8T!iw0s9=?MHt>i}1WfzUA#J@uOVE6vyr<6_KT=t9bc^sL_QJ9^ZizjdpN}8;pwn5-373=d -L5B(m8f;ysRAVUZ>*ohz~gz%9QrOfh-EB^+V$7A=1Q3q?y1A>i(akaN6_?p1Tn2-VF+H4;*coe*o2(s -4iaxbJCzCsrX=)@h>5LC%=CrSV4C*7I~PbY_E%0bz8B@LI!XFPhVSEicMlqpin9w+px@x(u;dsetD;b -7(1fi@XZ4|mTENa?8;%AEn_dL`O&i)9Utp?-y%5F5mMDrHvNOqWu^mQvIPLDs%0vwa6xa8V#;FW1E4p -h+)cE_q`eE{OT5%3s&^HD@BrVuQpa33XNMLG2ZUo}icohr;*D#RsAJ!79XK;Ao&apn+4ZXfu-Wu;YEt -*4$|1P&(!5;|ml|tBGk?t>c}H8dv2eKZCxkdAlOGQ&HO9)>EFyOQ_5a$0v5ZJFD;1*8Ib==_rNa60yd ->w~6Jj8!@;L2Bsaij#Xriw!Pi~voVX~BJ%n@JJn5Wy;a=DnkWr0345zwUYpTR@D|lbhFnVD1 -(?8flP2@;7oqroAPHc|ZupLZqTKEFE)060dOqxqIdB3AeYhC7fZqq*+-_tH~x!buLB1vg{VfPqX -il|IWn?TwWq~`8LZIgYH(VX}=kJ{Ut4V7J|q>a3>&=En7n0rUR!z1@GCpl!iSJ#1=Ilz)%rBDF3NJqv -=d(KG;TxPj3^KXgRMNN}tYF`l_EsV64GgZKLcSd;iv$>ECqMU3htQ3ybUW2G!j3bPIy7`pMX!r}7%_7 -LPy7J1d6+i0_9Qd9wR@{hfL(#!dhAJ2kN}c5095x5ep)E_2g-D@VPzokD_{My0Fw&w@E?$wp+Gp;HyQQ;>oraOMI_obBCxwHG -P`>eY~HN`&<`Udl?uyOHT?No^fcZY`^e+E%P?<(D$eD&RVvbwzo)(ngh-zt_}B))fn2{kka1cQ!Z`vt -{XdO*bbT4vS51W=1F)bYdwpC)X}~jzeV7GNrv3iWpPtsSs3%qiG%6zr#!`5c_Qimf!axP;aW?fw?#3zKq|LY54|O{#L3tfBxaW{_`d2^w1RFaCQy4RFThDffAuF?TX+XGh-pg|6-QyJKvg9}H%IGz* -^SP5OIp%!bAr4!F;*Yi!Qo*VLRd*qd9w_Fy;0b)k6JO>NXLjd*stv%}Y1ytz|n&STnTOB%L?W`j0$U_s7A>L*I1UGo6NZ*HYR={>`7KDNkN!I+m0g?XIvgtmVo8;<1<=7EJcW}^`#_xnl+cF@v9~?jEO-*&}s-2 -&yIWR4Bp>8Et3B0fc!nR0*os{#$s}&R)1{(9wH~=6UuF?d7pZ3xOI{V+cqRIOc*iu -O?a48iTV*NZJYIBNFkim5WgUVF0k22PTyGK=>*r=fK;_ZNb*uX!Zx7VO<4QBpO6++C3u8}4ti_$% -SR&~Hd`sk>2V$=`)8XdG9XrAh}%E6Ym2&rxzZ_cY6LG6q)6NQq+U@1Mu -j@!n^D0vhU!fmkXbV`ENQ~o!biJ_Rnoz_<>1}T12UxsTiunpG!DsTEz^2m`*0oV!9GII5N|zQJutT@b -;$<-O -|G}1dyj1I_o)7bLC%90h5zx=i^7R(2cOKJb@V3Ey2YzkI;fkY?-$`zd -(jWR`i38V%5JrEwcRBv_1={c$+Y{@7b}NR6E(FYIvHy234qsQngFYhS0q`ga?VsX2jvu-+SbwMc!JX>N37a&BR4FJ*XRWpH$9Z*FrgaCzN5+j8 -T`b?^F$HYTYdZ4$G)zLblJwyf1?cBXbNilkZFrBx^pXp*o%07iq-a2Zym@(oVqH}Vhp6Msq0rMuC%QP -g;=QjThOL;~HXPoF-0?%nb{@9a}3^D@p-CDuwtqFff)da=ybrO1QuE?7wa=;&%0E3wL=bt1*PkC#{C{ -f8@278_A!B3|WLQHptytwfM+%4M7`#6yg~#cdwv{xnWYS)@U73(b$RToyqoeL*ncKlv$_VTmJWVkR}X -U`8bN3ijo{f@jf4WslSrglDI%H6G!hv#U-?I#=N%mp%;|K#F4u*eP7U@ -!3OxcM!_kp`7G~)@UgSZcWZ^t3<)<=*fg?Q -w+nuZvrsuqM>e~gt>Uj_8AL%Dn4aGRO1_L$S;Vt7QlhV4D$9qgxbvgzAx*L%vdF%I{qs^k?CT6z33#*h)JaO}m4G%KVx6u;#%Nf3B(U&txXuDB1jY{&ooTdktAqA6va4BkQg;H0y -v=VP(4L2pj&{9F2#5(9+n`zL{LbBA?Q!!cG8T58q81+rm=*v001%sh!N-kJiJMyOBnbMWKNE-Sd#$p7 -z9`Y`I>u>ZUN1%a=`4<7*(qr&(?7Q{B`S_l`>4LCr=xOpV9t=O&OQdX~m05Ci9W~_VeoF9Wdy`sI6;{ -&;e%k9)cnTRJm+~gzYz?k@}{=&tmLZ;rv?3rW3CH2>uJ6L3wWH -;2_+ck$4_H(_7Wgci^KR3W2H=v4zU(TB~Ec0OK+f6k0{&`Eyr&E%p)2VM%@NOcu^*Qgp_J;081k14i7 -}$u|3s`UB>?2Y-gW7`b8*bn=n%Z%X8+iPz*ukmDzHS%z$JNv*&NwPeIw~?sEojtROCHPH -q?|RE{c3_Jo(aqMv*j>N4nTTuWX~k6V%4yG=ZiP -(beUMr=Oq*k_eAC0#cBx|^)(5Px?X?)2KQ3pEI$MOz^>lxDm}7^`H~;frKt17gw`Wvp?f?3h|LEy3{7 -EKfY>%I0a)15L|NiS={{8Ex%@Cd*G{d?L{kP2dZ(dOE8ur{lOX-RI6KhY3C*l){8->WatOo)%l=xt%Gb+yLUzHYOa9FVD_^NL`Wq3Iu@>4Rs3l*7`EnlFp&>r^9(oO -iO?hw%x6xrHA}Kv9_NDg!@2jC4B)K=v-5Nm;Up~A;2DuesNe~^ji(f&TNmV@^|_7T$`ri^A&pmx4ov9 -wbUjfW?RX-@XkD;pRCRWNz6#<*MpqdqxE#E;Q*h|wC@AHiHCi>uMG#{^_F*Yg@gT(Bgxc+`0Wr_ltg$r|Mtjc|gQQ;M4E=W|*R*f%m~=ag8v?O=6S+ -q{+(u12?4Jm0$G?w<4Px6TMYm|6&B=M8H}`))@t+}y8CpQSMk@?9Fw>Rp`ip|SJLGUdBi!3rDvR)1`!q;CCoya0{Z6=OHH@W*cVvFYB -$8BK<$;){!9@e!k5{-9)o_%(`IxUVJ2ZCZY{;ZZQl5Gn8W%EDzC}2p&>w0skO9mCOd@{ixOZ6k_02BRh;_`vQ -*9$5yrceQkU9+_Qb0-SC9$F2aAP&Htict4X1r3U%9g=Bs~nOUCi#To2aF4H-+>(wRdIWZsigN=`6^qd5zXf -I*68#h3;URp9AsCJ-3CSm2`92&vS^LL$C5+wusGU0SOpt2E}}KJlLg5!o$DgYLLF`o*l!U`c+$$BxnW -?QtYbCw83UeUv`)YX3sxHXNu8|Oj2ODvwlNY)a8mr)yiWC*Bo-iMkWr0A2behk8uhWm(CSK_Ok!a3qrWj;&FXU -q(|G^Y63;n^DWTNABsiJ;j~F;K;+Yxg%!*jRS(uc9jti6D(YNU8|pcHvEYgkI{}5Hl?bB&J9Z{O@QT; ->3v*eDM~4lU08ZSeNK;N%?DUOiL{pc}*pH8Db{qG+WK$lxEU9t;Z3W$vG| -_=){rmIDv?n$6!d54v;;FzCp22*OT#=H^cVgol;2#v@6Ay;q7R>w^dK*xC2SCkiZj#n@X%!(uk=;{j8 -wI2k2oT60qTzQ9vvy0HNj}&RqN3b@iMyxk5>Kjs+wv{SQWBVf;(ZzL5_luo8uJL#;xCWHh+SvX)Q%1~ -^O?T%07JQKO9>Tt<&5AreJkHi>mUF0(QDmuAC(qzbHBMV-r_2~2F@*YOA<;3+gh9~l<*04fUD41#zh< -X(nxiT*BN4fPPUitHh^iNF*k|DXDgk#YjqX-(90M*v^_KF-#Ptjd>haqv*j6{X|2Hw7A!I4p7L*kQE0 -(jDzR7jB$pY0C9LeZ#%Jx;jUox*5;tNg9!QOjl~*Xgl6}T$D1gN-e@Brl94(Y^AG`-T;a*2%aB -h&(ztDW3NS`QbBIGH8|c{K&))l9>r(+kk72^ZW;}E7RP2{_1BhO(7=9ngI;wjxjv>_o#M8_*|31C5NI -_pNmv7o60y|Oi6jgsUBjCh)v##|)bb(h{6?Z0RQDP_Q4uMfKCTGZa@~HvlrljniF$dTeAMl6mq~Vuu} -4kRd`w|BSg<4+>^^W#TLxK!@RXM?iHusB?cu=eqKtMQ)s_(`HwbBDH+U`C4B?uJ7)3U;;l7a#Q;w0F5-_X4J9;J;i(~MJ2$K; -hv(%b@j^-xpo)F)JZhxM1)Z&c*ADKiZt~rHq_2-rS+#B7*B}Tx&L-?x^L8O@cF~e`XLeTgMe>c1d{fd -4+ktv&%1{c;;-da?56m+OD%(g;zH)-YcZGrK&LYr -3LZHwv_O3EAAyYU{}oVjuZg@)`0OkBP?Fbth9U*E1c0V$-G_pBuH4FRm-4EC4Xfu_T-#F+I_^s_|HU0 -!JzuqdhTZYw5E}(X_=l1?My@;=GdIm0<+~O~Wh#VIHSWK8~{|Q#3Wz1Z;IxUPHf|TB%sa(#G>z&+c+9 --2&#i%2D5Q=<0nYf_oqj1zos_@9j%6H^H<6e+bw|DLCc&>cO0o-q%}@h65MT;Z88ZnVG=N^^2PyQm^ZCirZ`+5$POwDb32o~`?%Ku?XN$2Xrw*Sa`RX|_4RO4%A -jd2PKTo8Ygd<0_)&A(`VcW(#1I@g~69qsZW}=5BhA0fm;09NLd7e%)5O0{fo~STb4*S@g0`{DYridOl -Zvy|xQwL9Vi|i_jY88cDMa?^SwuEaP_SzWhq>OVQu5B*N0R%9X@n{id+sjO#b3UgET6B7_aJttL>9a9 -70I$`T9S`fTuWu?)L5_#6v`M247u$G0XkYiGZwC{VH4P!r1yjLv2X}IFoCK>`6bKNp($9)W7L|HH!i0 -&^_^N3(pN4p5Cp8DG94{7HxsYrPY{LmBtjvS}n7_p2NJkT3bDFb1%c9MI1Ef@{U&EzVNW^AD4VK|7P% -o@VS%bDXueRoR*2X1kb@s;Tf)Cy458E8Nk3s4832`@8sRJj1D;87u)E?ejpyO7GyEJ=9$BQhZu*Xs=# -&r_j2dIZ&#U|neJ*iSC=3#2*)Zh?azznkuDB$@#4q&!}Mt$rzjotWE8umOHPbNQes6ui58{P7vG;=S{Qsl5oCLeuYop^1kVX&kS|W -~bjTe5Fq)WmE^%exKqxUwmc~@ILqzf&sv*;1a!Tn?A%4dh>wHI^1RE_c^Ni|&C{YLMZsVA=_IN0zsX$ -uViuy^6gOJ;TxvaM9!mf4+HpGI|K~*ATkAFI<$fiW*Fe%O-RRvp03pT(JnvI-f(sK&1SvYw7G311!HG -pMMfOBpR0XhCcr^8hbW>QNe9O_@@XU&YriN|96Wwobh4|j_xj1%*YLK6y1iUK(k^h<3q)>zj~Q`bx7R -|XLg3`PlntMO!BLddfAQ1~5BE5#`tK633P&MsE;;nhGkaDoxeI67JR=(6H@kL<3ts$8aBlT4WGqjEr< -DgaK)*9pa#V4LDh3#qaMb|tLi)VxFw;BGK32V*{DsAG4OVN -Y$5*_)9<@kNW1Fnxfy>T_1A#MG&F(Q8OG4R>biJC%LhU84joku_p)KshFaw_-Sf?bfjo_;cm+wP`n!y -RM4l4ccIP)aCAA;(PlFkk4P~8}9K(Js;^$eynBbR6tz9DcA5~MZCYOG-Z2w+|CKRBh#yl|vSZijtnj=ONf|g#}F5B+~TkNUE#21UCy*x;}B~Br?*)l}cHW%H5V+$pRV@t5n-F{iyl17mjXD5ia9 -1Gq -SppIK-sG(5wM$!kQ}m6P|gjma211lG4tIz?sgsgDjP2;$;>Ak;1qD0~!oGz6GGluXbDW+sBKyZOH}pv$3J5T%E&K2{vv(Q&Og;KB-+t -0p9qMEU+N0#Q~~xj{_-lOXGE67KSWKUP!5q}7V(Nwh? -vLU93*e~P~G-7UysFwhdud$032>HqXk;eQSp&!Qs*r69U-WPRR%HA6355rq@i%b&<5i*bd?}E -!!9bL7TknQ`Qf1Gz%ro>bAT~25#rI(#ZMZ6%`E>k`&cL5tot=nxOvd_BOfJwah93wL+tcs~2l^jZLr8 -EZMJmmh9^F+2xryKRNyGh!gMcLJFbp1->I@Q%J;oc+tkHy3B`&fZ^Lx{LPe&D%3 -yXx&Ncy3QkVtXvwIr5y?I9B|ijBNtF*4*22&iob^#i$3J1V1?d-3r*MSV*ag=l+pMn7m|@ESjfwqMq1 -vvn_}fZrex`Q5gAQDl7^Hf8&wIj8`8nbeMgt@+#IHfE$9;hK-Fx?0)%S0ToHr820+u2tlNx+RosaB)u -Ybx698G@9X}lkUmZGz-m(o%Zw;L^t2|`q1{Xb*8IcL|r}?h^Z)>=yybn<4KArZ>8yS_WVJ`Zr+lhmkz= -pO&6;e?Uxp`PDCf_3K0b=>+yIaDo280q2MQQ+M#4py8Y-GzF$gr&%$eFI5c^t^%DWqdnKO#$LgcK%cU -L;S&)dD?Q?5+V;O0L;r%=ox-XQlI$Z;{tajWI>#g&FbWF}D#a%i$Q!I_rAuE_-I6dDAt8c;Ym54gxdd -(CG2OkPlT2)7-fQ^+2BWVcacsr$yR*_aPc0U%8%#X?nIroS>-5-iU1kmEfM#N--gTFplrmyeR`}xeG{ -f;McR`u9>?h}a=u8RU8*G5qeDGYTUDLy?I4vYh9-%wLI;Ugq@-?!A?nsV*v_cHbH>89MEG>X#G45~tg*o~cGgw6tV`pz-yZ;96h~2}ILaqB9tj)d4#V0Pjr9uw*0T9z1l#D)EP!f)@>iVfx12*+PZ`8%&GSN9hnY|%}J%itYj-`i -nUR4F`@>5ntH$>L4~I5LmzwV -az6fTqX^m*0v+qV*Bz?byJ@k2zP0URZB -41ExtJrJ5}VA1-yp?Z-66g58%A|JBcGD|5~VZoBa7&4Epm%-+5J{n^DgFj_s_1ZYh0tdYJt8-|J18D% -`9S6MO8eH#8fH`I6D4xAf^$V&T16)>(Hl4?5HZG{=n7 -rS8_ng#E+cz&~=ra5C>Ec4)U|^d1_Ju3=NT=BzS>bjElwsk#`g%ruNs_m%!7hRazA)hUxc)*%B^uIl$ -rWLAMgbkk6)b9pjfQ5?it=kSX>Y~QrHRIY0-1du2c4I(4d#fyJyA9u!wwB4P`-Ks3lU-E;1>SHtGzus -w%|34pMAEkGnKA@(v_1FcnDN=_8Tp-ttIxZ*DTaGfsr>u7pHRq)$)>?b!eJ!Pd&&%ldu;GR72F3bl0S -gvB5jfM_AmRt12ggu_2voY#u^ms->B@I6M7t@sd4%BwTg>L=Wd3YQHXP2ac+({Z=l5aD$tR_TAeQI}dpd_v9=oVV4+q9aIO+I`qxV-8h1 -EU#rd?WbJ_3;qJt5C&G;6#H2L9f(?OGIQ-{=Q+T+SCw=(uNB+6grF3oLw>#s=8K!d&dYV|{mU}o;{@h ->n)vEh{+A=CrXQahT)G5%9X~~?Oo#3@ZSk1uVO!O^%D9uMthPsy^p*E^^CQgeO^q4S1cZ8=Ms18Z3S7?cr6bS-VX2RE0<-f+u7#NO7;#FMk -&BWju>OWyb*v3^Pg9}8XR$kbhlGlwWUa61qN*%rG!daJ}M%C!hBIlT&g<1W5h_-B}XiEzvbVz6zn)Na`N0|fYUm`dop@!SZ>x%ctn?ZjaRtTzOnhyqMLPuE4#yfd=T5XG -NI4_xZMS!FLZx{u3E-$_e*$(G6zvJbAZ?4^|XXNWj+`f6j3piqne6aRd1>Zf?M#)B|;`xf2tEc0^NBg -qPj(74Rd{gU}p4p<(ROnBKqnPc#bG@cvL7!#g7DTNOG2c$CgtR_S)z2%-#1M%0S-J%HcWoNt|cT*-uNt|Khp5v^tw}A+~Aoyy@dMu*a=)>@CYOSCbbnkfHNU9{6$gLZ&YiUD^Ah%co>Gl8!8p*zeT*&{W6X2~_CtBmwLwm=_a>h~*Y?x#y_+BL#mm3p6h9YgCPZj&KG@lA -+xHi8aMbX`kOzxS7G%Jn04kTF?ZZ&>h6OmxH=4Iw7Oa74w@U=&US|53FA{j>iCP)h>@6aWAK2ms3fSz -GL$&@br#002?}000#L003}la4%nJZggdGZeeUMZDDC{E^v80kU>iXF${(8{S{$OE@*oZ)B!zu^yE#Xb -f@V|pwpHlqYU`t?T)M=knjjEA1`rT!TYr7#^^oJ+A1rAQmO`}TErL=F~J(B+ytIwgiN9zqWlRz@Ky`L -^D$0v_X1ROG|+1phWCX8dN)Qv{=x{UGSnPl>b^B2$i&oCK$oEX!w#a9Gn^3^6Ec>G%6!<7T5~L~-6F( -D7q*Zgq6W%x3xWuXYLn5qnRMCCcB-pudkd~nb0o5%xQF|vM|%Z4X1XNMFRbtE$lcU#okA=Y<-OIe$~6 -}M&7n2qKNQ9uTKN}`bWY|MP)h>@6aWAK2ms3fSzC|3|9W-?007Go000^Q003}la4%nJZggdGZeeUMaC -vZYZ)#;@bS`jtl~-Mh+%^<_pI@Om4?BiglR`_EfeExU4}r8SN#4R@G`5^k)|OmJ-i-$$tYGq;kS2-WK9ZCHhq;+YQ6t8)Y4DTc5# -eW3<^#uf5gV@z&_i$l;Gcw$XLBq$yK(Dn4tep4Z>8Y<5*p=`sq#*Z%!z5 -!|1`aa4^R60C)z&Mu|09s8I~rKE@KNM2PA8!vUs4a@;Ilh$P*?3%sRl?OW0O)wWeU)iS&&t_ui-!taM#wy`^ifsY4*+}vys6(MjZ2-AE@s`OLtC@K3h%WQpLj -KbnQHjMdG46TYx5}aowGnp(aLre!HRoAudZsgTS>=EZ(rZCUw{6MO&HoeZ*qy%1CMc3LozM@H&ppe|Oei4am2L+B0N#MfAsbwCeUpL -Kf!O^ju)YtRaERllAUd<#1~Zn|+dQA8J*Z0Sc?sf^Vu2a7&HT6P@GT1}qP1fGoHF(0oY(4`wEp9w`2w -fovBi%T3HfH?YGmR|-^xn58L+k&m270-r5)NRF~O}Jy3qK({0>vl&YJFoA>3)}Sk-DPS{UyJpaBOBjl -_&Iyq091w^J7_bSJ)uY6z|gZ&ORN$CD5(y-A~C51{9XJ(GBR9Cqlb7pJEw#EILVsN~nxX)Ysdk9oG3e -s%ceWpPuy%)LA_wFmBlI;0-ot-LId*iiyy5DXS1c!3hdt5 -$qE`NkmOAf4ZrtCI?^Dxu8s7=HKJ171?dsNj9 -wfHY&{CJXpFR^he!ys6^yw~gTLKhzxLTmU1!5S)m>f! -6Qyi8vw5)FZ-}{}pkQ|EMY*Ia}#o;+;&i$q>TDfOwT6DT=MVhj5-^f-oUe{7{UCLT57Va$PTIg~w{Bz -OOnU=Cr{$R(|t}5^R<9j7*|GkpiZCP*qqil4;Th)`2DsH&m`A2nBi-H)#l{~AsQbMtT3wvZC^sxb5|M -KhlxMnYSRq?w@EG$G*HgWnO>Rh(CB)jLtNc*U>N=xSN2Q#@En1*%yw}GJ<3XW+<&&RY -n}n!^Y*@LJfu(=Z^Sd8T0CkG_!i&C0++QGZOyB=taI@Q2C7QHTCInFRo3?}{NeL>R(z2b^w=Z6v8Ugf -!#U)=(3}kA5w=37xhQy7=~UDYWh?8wsP!zW%tfVQC5Sa>G;LZ@^wz)j$8T^dT2W_0%~7d^KFIbSWYpM -bZ_4J{p3EV2y2Q-A#w{Pjgy351Il`|x(2zvmfvOZC1<3CLcT?L>B;Qmm!&P>a@`%I28nWow+_m^Px_mqtsbs^hri0W};F_`aAH^%j>^bu9nN?3yX(j&0b!=W^kK -592a<$5w&F8;KmCB{~fP$xC+xd;UOgwG7cXj7aH{WgRC7qeF!YJez0MAnduwAt*MaQ!mJ;j<^YYuPz(Wrs4aS+GxoMUl -IknTbE9hUSSn50G?+bLS&UfgPd{>xtB~<~WSS^IFoEQ0L03e;1QrxG=j7%BIOr;%5A09LUCO{QZw$OJnGmdpAx%?`ZQFy#y_LANm3X)+@j*{GipyA&tLp*iUX!3WsB+Zp9noDCUrLU%Q -G78^3)chB$P7U)voA1J7h3tg4*m7e~T>1hWR&G5APxD;p5>A~Ol85C$cr98ZtQW4!?N;S2UtW#+C2&J -mnR*uN=9YPvyL_QS=O&RiB+Y>F(KybXwkOPn#28#^6+N?Yi3s7YZB!pMgS=Hsjh=5$klw_s;OfkWgDX -3b272ArS`(%e!Y6CnWL-&S>cL{3C4f9oR9ljg|=O*yxcxXvh6PzGYHbGbnkWYiiW=wO%E${Y?vMNHsK -_``hCzZTyJ9-wFB{>wPl4!Z6zGq`v9#6;=;FHFF-d(Bo7NHvz@u4k|ors^li4E)P^Mk;gS=X!>ad*-) -oUvLP}9e@K&xe+s*M@rN#xirvWO|L*RvU -8(17bMTyrrM<{#+cnJC~CI=sY|^~G1iX2Q0pKQ-GHCG!pX+eTYz=$K9Bw4;MK*oRT`N^`wyUQO~en}W -?ahq!?>8HQHnf;@D!$81CS`No&r@MP_RDomP(PWM!Bho2;IC?2-Ks={+`c0T6Dg -ppcl2b5*oXLXEs3u)ZGfit*e;bBcRx+4g~|OhnfY&+gQ_v^u4HyyI-COXk7sfhrWWYM5<)9(QvTsV+i$0RI!i&CTt46mz -$08-+pT58^Q-loD!Uxu-M}jhS6}<<8*DD0qhMDd*A9?0h0&s~y^)#<{rbw$btj=@5ftQOK^&mti+Kwt -DZKMphasPj3(B$<<;;V~{HNy=8{VYH`Z*kCtIi#FqUH`=dgW9a5OXensAKf^4w^&v7hMBA-@IFg{yUZ -CK4*`i_2DqN+D3qJOvd&fU>_3>Bg|fZ2egB=x#Bd$DmEHH(nG?1ntJm2x0^icZijgcCN(Sz{3Pw~4Zc -B#MF=fW~)xFf#i?GC>Spj}jVKo@eSr@*qSbz$BE5LTj$6S4`$SY%$THOID=w(|i^EraC^^ne+=fwA-} -kx1FGQ8VvEO+?TeD7p?erhw61CH_dEYhLR&`4#PBunknvMX3`@WR-tmhs-iS~TWM8EbkJm#7}?MU>5f -5H#le0P<8^YTX;+CfPcb|dOGl|)Xu6S$Jf+OpAR+Y|O+gQNcTZfq1m&h -Bhp3{hW1jhk15^gt}|$K=NzFtj~{0SjeJ&z3-c-xIY;+@8jf|oc_*I%R?0Wc -J?1QS={MtL?ZD2B-Ep?%megkeM$QN6A;6vb;+x6viX&^cinf7K0JoWWF3b2M7N{vdoz;8B-8_MYcz$? -(j?9haSiM0mM0&UGY-D9}E^v9>SzC|W#ua|ouNcz^N+DV{HsW3cC_rnk3j;|U*iQ1WFo;=dB(Y18G-t -T$jhp}8@64SSX%_+dQUZJ7%$)mm4wEE#+Z6{PRmWA+u^oP{`?A`W^`6}wgs3Fjb`OjDw?Vig@(-db>R@cQ -KIA5l|7`yvE7xGaJwG_{pH>LD{xmpl6%WpTeew#E9!02fm(6HiG!?HB?Pmu1+SaBEO=G^Ci0pe#cH^8 -D>?~o&f&91Jg@aFK)1h`)fKb<#4MFH;>HiOOAf -*GlI5@!?TcrBeLRj_Y)eJXG>6RX&7teWC;GvHj)q`{ad6B5Yp2U>5IE4X-r{(U)zcuyQu;6uur1dP-&W$Pg7;lb75s6< -PVIbLEEZd_V|mfkio;Z%$5gHve0j+(zGK%-Q>_;a57O8x0VN#ET2RhY2Z)Vdbd)SrHO-9#wHtw~#vHC -}p@G*e@A#1c@rMAR8YsJyB4d}?K0{J%E}m!495Ma36Udjf<|2)Cm`TBrkW;x_Ix*y->EWF^QKT05(|X -3~Rtql~`w#oEsl|Fg^PQpTMC97-=e(ElAG;=Tnoextqgov;Wiez3csF4)(WtpvF)n3))S@3myu(V#_$ -@Chil}VBO}QgEG9QYlJSZ7@O*Va3XeT!U(DRhl0%_Kw^-x_@ko;Mz+}FI4Vi6P5If~a?|B{t5hquRMe -tG$xfkbiB^eSVo8qJla8Xf9uq)WsLqN+gzNisR7{D15XUx0hkcbZipbrcKv5qRbjZU -NpAC)#BhrvaGpjWa(U7)Q2H1Vh0`!c0U -|Pi#|n}$HBzfCN#waiz`K?qk48Z}4vT2v=4WwPvw9@Xv`8rB?mkDROi!8AQX}l4c!b>QJ4Ia*!Q7;q) -U4=D7&b`^Lw7i-T_d$!a7K=+j|X~7@91|1!IRC1UgYQ;9zEL+rh}N=C3=$*6Ox$t+`_NmH@cVB7@1%D -GNkpS`+dcd5wbnP=ZoPE`^n008`i0HgrxJmWIF0OW64(|K?6G7jw^Bl{7z}bb>GY2q0{a?jbN7pMaGB -jJg)z(W+P_Ksns$wr&-s(JJxfs*OUKgjf{7lW==q%NC%G&GLP>OY4SrUB~|$!ES?K{v$0;Sy8D1*Obf -i()9E}LHth8!XLN;%u*y9U?Gn2`;bhIe%wzvOc}O~-UY=4E`z~;wJjAk2tf_5Z==x@pl1V7;UlZKJwz3%&$zagKzl?D|Iobv7a%dP`Ii0~*_%ta}?<6Vu|7+TXvPa*S_)Wa3}qj_KB-zchM0fwutd<<8`I3bv -$y3CQfT(@J5Y}K|EEt=LET|($ELT{0hkG**IH(R1gbN0`Gn`>C(nAbx=0l1eOL{ZqW -_ori(`H`ODd)U=N_SLMo?92lQ{;?$bDn1=K^} -Nwx}n&{p@3Fu0>~1V>x(LNI;#yA}oSr5I^q-}8>w%3K0x(BWEW=Qxt|vh%$ -+g?5lILqP_%yAm}6cT(g$9H?jTan&5l0-TwIcXucYV-ZmZDXFl~NaVIbkA_=%(RVmm+~b7og`B|$X=> -IsZC`Qa21(JyJbqvttB^>At6;yWHf>JV6N1SmJ;}>XaB9uXUD~AjwQo!|$omo(h0Rx6CJ+4ii@$6p)! -RL%gNvTvKlA&>P0qpUk?9#bjk8c9(E(9>R!4J6f;ba -!=@(VpE)MngW04)IPI(qHWT3SZ7jScm05I?Y~0JPNdIIe`s^qMeBkb&Adk~v-$is*DurLn|kG$$*nqY -#bjR;NCb8sq>=Hco~lQnF#7iJlhLi6VFuNVwR!b%g&)_38j1XHSQ>n@VL5UD(|6gAg5bz}bQ5yq3h>^ -sSHzIIZ(*NuUn3S4#vMXG6h6wN(e2HYSFKyLEOKQRNK?2sNThZvvjaH6z%VXOHBa -=!L$A>oGHQVJ;xJ&%h8{g^CYQhXAvBXkPi4Ag9C*4unB=x7t59)8MpG|*$elns6LaQM0T2n2qHFs7! -Hk7b=y)eO?7*9ZzV)BLS_$c&(Sy!bwj&JnghLb1$UPWE$*0RoYyq@^ecg3Y%8y8}0WBUBr|D -{|#oNKM|+VC&@GzemPft0NBtpkPf&m&3_v3ixa@n5Su=1XhPG@cNpCnZ(DS|dX_ybb(0=EK)X~h0Ja$ -ro`%R_mcXCwQI}3f)F;D*Y3Gu>H-~q9Ro&Z($)q?YE-7BkWyKNEk0&xCJbYEXCaCs$+F%APE32D -@HWGjLU%^u!RzkgMlSf}&AV?Q?!R4AolHwi=L*~k~%|<45B(=K+?vKA2en_MwZ<6{UYm=PsncobVEX& -@@Bd>`$QA1k!#FQqqGqRTFwqb@#LAcO{HVxa8hYQ*5THcajceg9EEL$uNN}fqo9Xit~R#k+bTd548LP -$gLpk6Henrux=m1g~K9lq&{c8qEX@@h*~(uX>5duN!U4-IpNIQ%kFHacwAC(4D}GnbZ&W7v2jE?z@Wz -O1+aNVX!mRO$8RO|=o{&%(KwOIxxeszw)nHjpLh6V;lW~x0q(5^8R6OUX; -ac06S77JL7gTnHH-2O^_mx8Sp1U?XfJ7Ls?+)Ln24pK!CBJx1c6o|#1=g-&*R||ajY&ZDsfySic*9D|0SL1DclXtsKR$n`-hTM^{rkJmcNu{UfE!L#wj#BL{TVU9 -8cC8P4`Az;iITyFkSSwET-(GLip(=;Sy%`O*v(bUr&^Q83J*C2c%#a1B^+*wrXX<$T{D88}i2-mr@bvl9o3bPa_zQSne -<3EMbktZ;x{U-GZ?iN?EV(`ZRM(fcOl8;P`s8C6WaEi61YYn9fy|WW+T#X|S`z1md`P;gm?#fS -p+~0q^`}C>$=iU4JsZF7U{5V&K_{rf!1+}!a>;+XvT?938-Od!SQ#Dc?EBFSy+yaC}8M`uJyj?_~3X= -=eE|Hxb8y>mrG~lU)+5=~Ohah%#qmPu^=uWU5X`Os9<=OZBY@HRGpSM=w=eQ=xiVPimsK}_J=Fdy==u -Gurouzgky!{-?ycfKkG0LBZkv)t7T(q5;75JqWh~GxyzQo1ic@?olK<0SdgwdOkBY$MUW?Ni?`>wQL{ -0v+w%f+ES=6yBVN-NvRSx(o|rLQLTa;D5D_Tjv%JpwDZq)5e?b<}P*QgXM0DI#LD+x>#XzwUNR{0`*s -6(>;mfkwiw7+pr((h{=J0S6YZ@BEc%lY&%8>cOwkct1-(8g3KDS_6MK@MoovdEd$@ -(tZ#;V*C&{-|J!lfkB)i>nv%}=L -qm;W0%kHV&b|+nh~^ddF9{fao%Hl86T&nozFVeJW+U)d+9Q7AyV#HGZ*O*Kj9i?)4Wwb*yWZ#U^BoB7 -LD>WG$WO(%yk7*nu`~R|^t-SxDsPwD8Oh2&B^UYTdFaxy6l}G#y|3_B<#_tCZ!|A=D}TMUx^G~B*?Ql -s@JYZMdF^ek(THifJOn1HJG=D)GO$bpX~cUYFVhuuzp$E^Zu1j>Jo+eEDF -%agR>Y2aSL8o?kKl?(8OO_cj8TCyJZkDO%$c;~h1;EW71Ambt|BkJX2N&A%Ma7$KxO@I58)039z*e`o2BB%kEp+aMBD}+Bq|pcmB!jU7ctzWx)nXh9qbeo1P>-7_zfo7PMdn^F -cdyIzd{~!PDxnA*-*r{v_5`ma@h&gy>j^%-Qng%55rA)j#FHwDWB0z&R^aj|I^!^lQS!3I`ZIjXz|ir -aJ@6aWAK2ms3fSzFom@aG2$000ak001E -X003}la4%nJZggdGZeeUMV{BO4q%gVWiWQs#kTEXTn%n|F+ee-Pfd_IVw!_=r&~w#&h1RuQx-!@>H12pM3Ln -7XPIR0wAqRrXpW(^&}?DBS1fIXd^}=!uJ7+}@2Bqd?*97y{PyXAUp&okuK3mbp0{?N7QZZRKP=iy@8| -cAPvOZ;KlSxaQ*&bZfUXP4}9whF#fHaebsWnYMyuR#}ZRYk4Yt1)dn=^hht%r^kwJ -{ZrOoVp-HZASleml6CTqx>nyJ?F8GIPXGLikMX|&30eXeX~C6dk*`zQyyeWf11-LmEBopH@Zk-x)m$4 -3(GajH_Dd3*&69PtB+#QE$~v!kaVr}FcO){T{s}Adkxy4%QVs$iO%9|m4-tP`4izCw`^hBr8a3_;2aF -+8;MYaC@7MO`5FT26`3Y+hFkGjmWtyJao4h1kR@IZ)zwg&!=G;gRuND{%ESlkE;$2Q~gu@ia^nUZ$p&UqpTyJMsn=WOi8wbr2T%qv-$!IWU=9|&UhB_Npu2==)5*U8XHC(LnhT~)nc(`P%ppLuMYmP*kLUBK1Zef|e4*w -?Ore22>l=nD=9pvnJ=-DAx|{|4Q7f24@7o_=>wJ+Bq{^i2LQL^62)c7yX>T-N6Ihq8IBTj}z8!Y=ld{ -+KRI$&s5o?7+Ol+a5HwLQ(n7w$f6qNYzabhs{yg8@7~@C=j--BZ^D`j45FxRok$0C^PVTT$@-wWAJ53 -MN@fl2eB?4{}2Fma5J=Up{LRnBEnxg=(8W1p5f~H;_0Wu0NVsY#{!4A#s?(6Tgt!4{Zb1PD8MV5c -V6mF5MoN8G_K`=>VOeFxMW`aM~JcN}s(r+-`zdc&|R&}l&QdhGl-V}EuM0l;b%j9QVTo=lI?=R`WuW>(>@78+ygCmC-3(x`pal80BIlOeh?R8StTWUJ@lqS~F*$Gri1vn{zvODu%& -|KPZ^#0a5?;ol+sPchL!eHqjurNBr^lRAP3R%q7^HR2?rqntGAX9h?3I}HmZr+m> -y>ho<;$uMiTiil*cTbPOCknfN2?kL?RyMa>M2D4VAK&%g=MRq${QUB9e!K8UE8iW&HasV3&-~q8$(~# -U9{MQIYjx%bT+E$>6idLknXQ!x7H&j!#*$)Fk;WYyfFgp&K8g(0jRzI^WnFHeWJDl=)8N9-PKmAWFXM -{AN!aTyI)*y*6>wN|5#orX&qc$Dw%(?X7oE>q;{%X`K8|kF|28*p?5K2(<_LP6%U4y~p1plz-!e1~3f -7M*ih^^q3B?jG!~g=(FgiNx!KhOftx`8CjKfFyIAlcD@qP>uMbo8}Dd=05Z$ftIjLAhFX|~E*_esGOK -O_?>D~}*^xu?EJNX9B(WK_5yL1E0dDn+cPY_E$dM;yWCT6w^}we0j~wo1f097vz^9BAnf?9lv%jHdgI -rPD68r@ieGYsz|K*19?JQ&g&_vu5%itH||7)B68Q_&CsQP8}!huCqOB2SQQ;B?9-UrzFID(6O#1tq1( -2$d{+(?jXkUreD5}{sT};0|XQR000O8%K%wh4L++KLLmSEF?j$09{>OVaA|NaUukZ1WpZv|Y%gPMX)j -}KWN&bEX>V?GE^v9>J^gdr#&W;wueb+KhvX}X#PLhhVWUhHMX}ZA*cwU6Yev-|5P2jKfdC%>O6GI@zj -wdx3j}4Eil?1sCKd_Y?d|RD?bq!sqbNG47ip!Zs$4Z$S=1^mrfOc5>s5VlaJkTGRb|B_Tjg3Ubfr~Rt -EN;_JLx9kEN!%E7J7Lw&&!)M&o{Ap3*9Q6)@7k?bXLsaecD(+_gS-0 -=>)$V_R`f#7t@}aqy_w6RT~9R8+~x2n|rMbwMwg6*RfhzMOfz=jBIUS9lsc6NSpagn?_d -3`oiZwXoR_a!os`ErKcs-gNN&DXkSXl7&sZK|h0FnPCGAra-xQ2{G=3~Q0zlns=P0#uXEp3tv4h{}LE~zTbYMt=LNi?oSpO<&s;ze#Uwrsns=d|wzYV@6YQx~G;0voy&jA79ZEai+GB+~4CX=_W}EkaM2&!$<0PgUHtRAgNH)WG>$<^xoTm*+Bxgf_p^%~?)Kl35W -1voI8;nU;1{*-$#D41#&A%-)usgJ@6lgnKV0w>uj7HoDtrNB?)-?|8uUwldNe%4}VZAqfOaCkT$*o2m -v+M(11N$B1M0Xn@LEZg2y~%dG0a2H0G!#fPxHi}HX~5LG$%?evBxR9raG`uL--l`R2SfEf8Uw}0(Ff* -{nx;F|XhCVP-*lbGf&hU2K^u#fKKavOFeGjoB)P?TUX);d_|Sc9pQ%%nRiLfKIt0Ec?~A-lr?tpQ#7= -pOQL81e5)A@&6O63|QeCdAiKg~d1_q`zQSr&BUg=3T%O>i6p+Q-JxmKe{+zE-gG-AFS8 -NP&FV4&e=V7VcZel%*aSR0VeXj_d&MLB}$$x;B7VCDh=xnRT`aY$@t+q7UAVi7Rz;8vDFsy=}9hQK9h -K^Ye)fCjO9ES#0iLYZLS!wa85r@*<@y0DDWxuAXrv2vzaH4aeu*n$$e$DlJ1*qY3PjVBH(1*@Pemm2Z -7E^-ZiN3;zpH@pR0cEP%q^^=0j-sRu=Rc?3vkP=c(~PYTLfUPA>p;xPay^DfwY7F`2>cxfMjx0 -O01cea6|gRFVb!M|-dN}gv%C8Tjk)DUyCEkz22g2vc^Xg3)uzAQ`YHr>4ZU|0qq8?4CB0h91rvSd{yTVP1^Kh!`N2Zxt7L2g6<2W_oA?I;mQ -V_d8d(Jvi6bC~d>SDT`8EZubr^XMBr@hd!jFg1#iG2Yk1;$RAS>Lg=Lx#-d!aL;{s9&aVhb1W0KgV(s -)eKxo+xgO5sEBFJXt)>l^TNKp|=0f+3BmQ_v7I -6B?k1BMOqbnS!e-vUosrF=FHiq%j}@>I88A00NzZ@q%H^jGSI8{1Ez3k@-!LhvF`)%7Vg-cWIU3RM=Z^g9ztPQY{k(K6kSS~rzWZ%6e8cy~DxDOCrgd0@zoQHn)EKn -!4xh|R(7k=r_;$?ORLJC#^rPgZQ#5QIJyKy?r8hFnq4|tSetj9I2%nQ|I -OI^#{iW01f9HVsm&>Rd=>K)9z%1fx(j^o}^t3GP3qT_?KWc>a4{t;j@phJ83}aL8Qj -bEjWUZ!2r@882;U1NI(Mo{3nm)6;E0b*N`K)2iJJAof$}h8Ct-mIYH7qQ_$AY!lBzR!xL#$)J= -M};N(;Ln2*=z`o=>4_o3+7{mZwEg4VU-?9b!&rL!b9K4|1v`7#D4-7z1N5r#eHAuCIzG7P^@)teY^*} -_s&w3-X*G87NClcZjlcGpp$J`0erp(8$DA$knfn*x8%SP98c=e_qNpYAAY*j`ys0lMO@3}iozgvV42U -Y-k=qhK#Mz2)O&6Zfif~sf?2eO*tsXBjO`RZvsaIwz+@Ij`BV1DNkJ3Gtud+SNEUF1bu4c|_Fz;B-w0 -+Z&u%IV@?|r&If+0d!6f|ebHCet*<=9>HUSSHtI1Z!*qLk&Gnb*nCx*(+<2r!nHqJlEr$I}0z1!?9wB -Bvn(#k&JZ2)d^i;1F8Bgt%2MSbl-KSl;?@GF;k{KXenNVPGhA>2>@{hAua@Z=63!rX<`!2GE)UGHm|Q -|Ff9?LXuBWLt{sYkxXFrSG!kda2R|Y}QpnX&?X#u*qsLpjDyo)vL?PGmNejn)X}Vxk54{Jf=&YI5C8= -V}3sPr%schVF(uE$SEaUz;fd0K@NtgJ>7`K<%)Jr8&!3=mk5OUN@pj+1$OI&72LI+1Vjm&F*Rm@cbgv -11A<&&h#oETa&n6w8Hin3J3|G-f$f3z15K5}QDB7CO`Wj#?R$s3L4ZkuM|NjFGnNp=2vOaDI027ox?z -A$1dK*J!iYzR-1!+#IT+u2+hJ&gg0&_|ivQf>KR5VKYNV)7tSj%C(K4+4Jh*mAwh$3zBP8V4Jngh&;; ->OX`PZ)p;5SVcL)Fg)bR-9#?g!Zl|K1G1tHk`9emb}jNkXC#qjtJ_EDLU%71%joEAPRUeb8k@wDpa#9 -?njF8fn;6Q?PQ6j}(H%QxgmZ-K;TE6#J7uohi|ZpeT&9Os9WlX@ce#TC3j0#cOr!f&E2S{1N8YL-~xZ -=(L=y?O~Y@0`)&Rd0D~7rR`A4B+sB<@-*raj1D3@*t=+LI1l}45R<{+Q5Wb5ORAVwA`~d|C|OteUJ@{KqK^QL0FZmM+N?6~i3H}%fNh5S`TTW@ZkP -xa*K}Bl9ei?qpMrS;Juu3M)0Ei683%5_Nx*!nEr~d=`rs0?1;w@CtUt1rE4?b~3`|l!iO#dMka*PD$( -j24%NJIkkN_ss{7y$_rvo}<kDcw|QVoCLAIZfghx^At8qkju@- -6a&=QamU`yCg0z&j}8K>r>+bNjOc>pt?5JC58dXtA>ROj+}jNiq!?Ll1=jzqE%}Ot=7BxdD4|!4|u+J -42ODedo%9gr+YJ}q#i=js6K^A(Xhvzoi^an&fv``&x+fc$tltGK&NAygxY)pqs+X8jQnx7IL3kBmnke -`g$yZ_n5mSva>WW3cVeLy%`#Vfde}eas_>=l`HhNhvAt=zf -t82y)%lUuN498qfn9flENx;hAX<7|!kK;M9X~#FZlyEjqHzJvjhqVYZ~-_qCbOFUJ?CLotbtS6i}XD3 -}2z1A1?ct#Rpr)oMKDSd?}1DN;iEjj-R=g8M)pwdQtxU%kD!d~@{E$)kkCPib|_Y1=pu(i{JdCCNFfD -0hfa87h5;htnVm<_ncIKJ}(0&>4xeNbj_qapMvc8}fcG+9dNcFdk2_Nw{4ckS*%@xe`y^FJ6Gi(e2m_ -V4|xg9|<{ao*|*y2>}7Sk?xb^M?a9c1T^ScfsKFrWZO<(`2==r=WT2eZ9C265u|H-{dRa*d_yGUzFsuN^Mi6lt)l_2fYosV)#$pEv=3TNyNoxkEz%(s%uU8);32Q$p%(xf*v{xw-G*aT)tnQ4;oekXa;j7E%WB%>m-fZ*Q#9`tn{3 -@5v`zmGtjcRgbjAJ~a_cf_^c?Tx!OhIhi7C;dTaXkxbP4cLw*{w!Y)y(J1`=f?i9>0IcGSauR-KzpSr -%|`59I@diov)9EP29%?3F-w3L=({hFY_aNXcSLiDFPoY3?3z1I)>VaWE=*mz_nb+aP)=8n1I^sv<;;g -HNGx?!Y;5x6OUx=&tvqVJp%9ytM-dBiVXh9YYV$rA{+vb2{PVPJaM0QZ6jDMJbJ$6MPTuG35oDmpg6n -dPtCi??lY9+xaRP=QIPqhKjqT(*-l1y}ylqp)Yw-V(He2(5h;oKDj$5`gQj0NnTq2-9!6lOp>&j+!Z@ -X^6ARVf!B}Y1CNe5RvWO;>SZUt6Lo0%=y=qA}FR>!R9(b>tfEoINgzQQm9`A>BD%J2y-W -UP)Rn9EixNvfAi8Q9wCxt^SG(OGc|t)a<&r#^TBvEd*DCeFiD@H(r56j98r%)u}mP^x}e&@si*clI&) -QEt}aozxVs$nXW)(4k!700vqNdx0+R+RLhfAUTPcOMlDKy+$sC`IorHL}7!_9JWT2Lv~kqeTeUI@8n -Jo%*Pqk-uv3~oLRbuN6iN2)*(vRy3--pJ_Tr>x4gxuA94?0n({6|F_FJD!+aMjIheAn;nBw3a^aYGwJ -zv{fRh?y^>eL*-5#=HE1*`(S-EXwY2boU>W10IT=nx419?5zpu>y!Fzh`r$}$DNB?U&q8!Cm!KC8{A? -QWnK6(Dnr$GQRjGqD>LKh`)(nL}VgE2O0!Pjo!bxcf&Gmtd7)IBW}0bAPzfbBU#gfWkU?%N=bWVEo*2}zlT_no)y@}Nj!@>rl;5rOx+Of;)qNIm5^8lU1(1X=`+1t2oJPrO -V7m76rUrE1t(+@3EfN3mQX0A#>t=u?}4`cAzVZvjNz5P@b}VkEFxR9f2=VbIudf-=rrNkZK0n}9Z -{(6b@DR;oK>Gkc(cve}Osrp&`0V=7^{5>rB>mx2&R*liG2~EL)WEcFZMx53-5b7rMFVos -gN~9_`&OmTlBU;=J%+BrLvq7%r|XgN7I28+f9BqPR&aE=8>O*&Pk;YZeSnV^ot~#Dso -_e4Srd%=tF0%g?yyyxAX^9P#8SX~3=dg}N{Rzoihn-aa;H(RHnRS7ujBn{uTRPp3xFf67MbcQ!@@E)<|(?E8P3y_grDgpflilk&+%qXFlM1REt -`>1o>L1AHg&D~I>uuqkZm}d7x9_|tsc*yY-+_EWdi7ST-ca>^-{8L!{OspZ -0v&tVsu!zM*5j5X-9nZ(b6t5=T0TWE;I94j&dd#gDo8&Lgd+o#FkFA4Gd<5F0b4PU#)W^;6E!!LErKV=18o|O$*rt4GOHF64WmeZCLjwORB9>8#93_xU`G})5b5vV% --uGR_JULD*p#shD3O%C9Vm+_X_R=GwZzF@0V%fIfd1n4ucksne8maLg*}S0&MG>0MD&A8k50vgf@7Bp{t#Nqup^J?zk%w6oBmr1a0Nw_+rD$$LWwAH2xXEZ-)aZvvE6~Lg@+Yo^4g& -2-|W_hrYK?vTPBd<-r6jcujRL_Rd&D2qXVQiP91b+?D`sRXhuY@B>Fzl8cN)7Uyb$r#@+uqU-W4p6d8 -HN?tJ8VzSebc%l<=##a-J`|5lNv`-LMzL;moWR80d#9++mfXXX76~8qH(QL4O433WryWETi>PvSPxd7e#`JFmxw}qJwrcFjCbk@SA~ehXSavD2d7GM;knd7# -yfFtm%fA!cjZkU}mX6>ALJZ>cli9j!e1}J1f&aj_?eLy8T`!sfz;ceJ9;I8>DO(mfUc0EkSAd0CgP^; -&pgdlZy>Abhab17MxP$SHYo3!VYpE`Q9k@JceTd@0eKAei$CA9R;avhO!iEb)I&FL4>eKp|aVFtwJW4h5oQ(NL5$8 -b@rHOJOLpBj^S#CiV0)vb}1Mlmpo0(^}#M>Njw4IN>91%4PnWAH^`8VUM(%vB6JTf<-q41_aCB#=n_j -Oz5X!;iSS5ek1iD>Evk0C5S}Lf~1*guII+-^sI6}T^YAmvF_BzI60KUD=7aWMNCb0C{Wji2g&o -(E2k5PLWe2}3B&_{s~PAs&bC1$dtvh)WxerPnyFRFb%0H0Tt9vkpRFs6S48^wU=S15f7-?%R9hUkjF| -M;k-X6d@CYr}KPFtcwJb6F@%=t8glrCP1`6F}5AkTP3XDs`JNuKQYa4pYjK0FDj<0>JNM -7|^k{?pQ%QoGJB#@N7_Bu<#q@-I><~s2tY!%v+6rLcM<}i@14cQujEpQf+AI+2+hwYKGFzR8IzGQESG -?(noI(bUP5{6~cSL?0{Cf&n*On(GdFmN`lfNyxA%;>{*0lR%wW2-b>FrU^KofBb;FJxabs5;cTg~Y;} -!1Ku-=7fUV-BHfdUXuELI|JIqroj`SJ5kU7)g3X(%}CL|Hc}Olowb!_B9!&3Bkpf<5ELQpPE(^j)ZTW -+F$s0#ob4=eV`0D)la*e5{rsCT<|Tdve`r18)&a&2$YXJBaY9%Se#`XXGE|9pCH3bx+Pl>oQMl04s#; -Ie9y(16%}A{sM!vH3YFy8vBjeVhjCwDqeXu-e#2w~A{PpD5!(WbG3wR_b3_t&9BC*AWdH}PAE~>zJ|G -_?aOm1=j7%ym&c~plLQ?uIC6z}7^D`Lo`AVGUPsP;fZx9bk8#cvKxFe(-D$ -zFT(*!cqtZe+lPn=~9MdWv3d!D^}H))4k~jVsJ%-1S<#XbHz@Pv -WYj!&1iUbF-I11%pT{Ievd4MRUmA&035iF~KHl}#5qw)|r%TXstD*4YU`NE$~N8pK* -ja~Rcm{?Vkbz$gk-{YNdN6svqM*g0$+Rq-8rHtvPM63`!wu+w3^NWc2}U&gI&>CI%Sfy@2s@Ip4GWl8 -C5!SBsHW-44U6VM@q5mW_xZ&}mEeE)$?(%yfFBl)p}?TXnURePFhn&5pQnV63ICpm`4dYGexJ2M23G#Q6D6G^>S)YX|* -o_zabzkyzq`Waf_{8|K+xD%zEi7s13mq!)vhhQe6r{zg^?Yd_lvguZP&Csf6%bCqyfd!$VoLJI>f~LD -LBny2L5U-0$dW9eOj%pmvrnq08;~cgn%haa$Llg#MzcfhRXs}T^zO^5bePz+_yKHjzZSR$n{=wpz`*aVjmv_EuVlQxz9$UXTUW!pHp!B!w#piTF -BB%eBxtd~itp_lh3QWt0T)bv=b-z_gx2r*fh>Phj>;9*gIbjNbVS)GmH-mH{A%>w3otm+qR#jLgfC(I -2=ZZFnPVa@Epl&;MltMnqy$?&3S|q4&q(H63JgNN_6@zX9AXr9uRwr^a-e%me&x8 -4i=3K#A)N-%!=z+(}%M*kgWJH2f*ih%+IBAFhz2O -8<8&4i4l;QPWUSBXBT3~&-2*9(=^h02C%txF$s_j^*}?nfD*ky7!=wtqpPqVT-WB! -;J7G2V;7PYGaZOr6FnqDSy?i?P)mSaRe=zj`AIW%OxxGiQpDjD*FGx_2uEhrM)ID*9$UKby2T)4`1QY --O00;of09jiI3LIfQ0RR9o0{{Rd0001RX>c!JX>N37a&BR4FJo+JFJo_QZDDR?Ut@1>bY*ySE^v8;Qb -A6`Fc7@^6;@A;RPumIMW6~02e=S%sVd7&Hf4olN9%2Aevi`-hlbMb!IF1p$1@|BP8(AH#ORAPm(n?ylWCdWpdHq -rW2J-+U>n2)Rq?+?I00CZoONLR -{NJCpyIvhXRgQ`buJd98yNc0@&j>x-_^CfN%7b+XiCM?zm5QrI+885)=t@2Wyd!(YYkDeZEF}@&P) -h>@6aWAK2ms3fSzFfEZ*Hsy0031M001HY003}la4%nJZggdGZeeUMV{BMd?cwb|0ZEaz0WG--dt -ypVs<2Dxku3tfTK8%gEm1&D@vB7%LB9l&oO=kv^Y=L4h5EzNJxsgRJNyTv=|Gnq(p_k?9VzEMi$l`tR -^4u3^vsrSbD$8YMSg!VaUKK1axiP}9!%pTqw&zDCD=Re5EjviNW4ywLLOinzW_N;B!XA`7!ft3S?iOkBq2wG*eW$$HuCybG*V&saAmI5ZE$qbH4Hd^gjmT -el_Xpv=s4TP1l+=9m>N%&T(?H2OlPi5kpIT|l#(^R=tsx(ndshce4`A+zCRoj|t6OCg(@v;$t*B)y!H -muzL{Qf$-`r-QOpEqyc1S@MK{G(*l+Bus)XMLHiv^vpM)j@VCzP)zq9#QOZLU9g(soM+m#mLq1n -!7uchG+B^_j^kv)i -`Qk!wy_mu;+|vMpn6Shh>`DiKaM~@OicUuZ=o~&Q8z&$Jpq{y5x4F^gehnST2gs8Jyc^V5bw7cBeka_ -S5$t-rc{udUu=M-`-_cw>Q^s?-!9VEfp`YH!U~EtW>$r3-b_s>0u|dm@hnePz)}B(Pdp0ARHnHp(CsL -L^uUZQC6b=-}#)4!vHXJ3Ogx^ZDq*O>wybL+e*+jYiqp|waAP&Q3K6uA+UlG(M)GL -zc0SIL1Rjt*NET~3!`Dt{g3!c^n82#~6_kF{V?HSDE*6Mj+U9wrFyxp@U`<|CZ^unGVNP?hKwpI`LXd -jCK7qnq%C~0V2%Ig||G{<|#8q7Y`2K;E=^Y4jLuh0xN_OwAj&}KTO7wkh9LQHd1W{vX9sfIGKukFzt# -5mx3E(T&dU=R)n2KdjL(zakFbWMYy1KI!vZv=`n3u0D?2foE=2XVvG-VCuznQe%w6L`bF?zf=2)D{U( -G7Pd{@nU$iFQ|dh;w55JQ9&2IK7!#7Efcn9O&WH#Wax>ceQ8Egcw^NbTsbexBa2DG?NxvLt~!2ljm*$ -=BuBMJR>1P|)ToN>tgX%E#l;rwUh_a@{-SpE=wg4Yj~7V!Fzw=-umAQndzASjT9|4ehH?&7p0P_Qa@; -wOkT4+ZWul=U2Wh6WdjyDOUpU$~ETDB5CZLXn+CTJ)XKBvD>wjxZBDZTgvO-di60=YHt)&WXB(HT02o -+`08bOv)-9w6uON*qB=UFMXJU_0e@|c+ug4z(36{6OtkSK)N22Bt8S(AwhWmFU+%y%6SZ2hwqsbv+&f -S(i8MY})H7*d_H<^yiG$#vZLzZJ@2yKm$*0_^JP%h`f<-ua5G7#(;I80+w#c-q$vq@FuEED#I@IXA$C -I|Xm7BkADl@=PXOd(4l@T}&0rbFN$W-aDe+$-iJfAl@-qp5Hmf~f;g)kb -a+Sp`F<@eJ@xSw~C9X@de64Ept?0dR<)T@J-8r^(Fc>3*N~27wqU;R)(pKQnEjtTQLFcT(3<8ePwdN`~S?mCV;bO)ekjh?J;O&rS{P+MP?(?V7Na});>d@TlRwTUTHoo8rE4xG3W*aOZ!o;%SWKq%qEayT66Tk_Ev^1VJ7nBY!g-@rQ-{$a -ju^|S`C~k5JfQ2QyCAwR2y;ho;xPV>hz{20wJqss{p=0AryC%EiY|FHN>8nb(JIH*3PJ~o4HI-cP_Xx -Q9(G%t7+!WR_T2G|Pv%_ncrrb-@2}Og-Z%slZtxUE+M&hmOo3VhMaHp*{u52tjz38%-+^1Xk(GAbZDw -r33-%*_#L$g2a?YoWV-3~ZAa*(-p?BDEb8b!Yh-+ -!7TX)P)h>@6aWAK2ms3fSzF!Fyn+-30093A000>P003}la4%nJZggdGZeeUMV{B_ZR;v~-fV(4;_8cD$GWzUN5Eiqmufl1SveKKI<6@;v|K)Ly -!iki(!=ErYT~kO}TISY<+)=XsX3&hAB3wPP3^Rh3Zt!MY%%F*ZaEpJmgTwzbr>ymg&2-TZ3({NzvP$D -?!xGJbM1sG+RjnVgZjthFlT7;J6#gQghV?0h+;Z)xtOGSwiRCpW9bw+f1TX_{B{(v_bTYDY5suLeEUG -(;3;olk1`;GA`@NsmG4vtUDmqXc@EF<1z6JEK;16z<|#QB_9nf&41VvMcf43IM-Up%J4e!O0dY(I{99 -`t+Un-NT4b;d~^N7qSlOK>7qTjq@+M4!FwsFw$2ovm&;(gGwN4EhIuWCDa2jN(>@}HdA49wb)~ -@PC)zJ0T$PjNJx8X`}PJokx=$zb_F!ZID$kx{5o>P$;FyQt8RYnPw^@%ot<$P~P-H4-gkD>**hNO#i0 -&rYlH_1syqJ|c%xCm&Z0Y4=s@Y4$}MBsk0M*@;-aU}6*>@iuZ$#J%fbFh!Z2*bvL -aVc(FfqVfh;s7)z;Vf%z2G}MJY@R3{RGC7X4MF;%Na!xZ(Xr+Z!0&7;9~htP3TcBV2)_7d=Ds!ruY{N6x^(NG(0xWLJ3NA564Ty6% -Afz0V<-uNfXBghRg;J?Q)}iAJ*}{#JY2jGBr`o=>>lws}>>(N*QEX#b-Pl*$#eiB+N}m0+GpFI!BQxg -*xVRx5-oClbVXp@wAGbSd6?ZpGArf_sCn8bE7Kv&qP*Xmx5{7-8jg3S%NH>+gumsk|SF0U8{SnVNug) -=-=1$T-{xt$eum!99q7kd`69#s5}f^Aj|~O~wGJHloJS@nt&XpLn`GV}1YT<{^uhSK>D?$uSNQlgF;Y -3-CuC&!_CXZ6>7%?1npn%PmE!2h*nf@zMqlGmG|_6xoTfZZCDDfhgjTKY)zZl^2?;*cOnVG#g4%pnTG -lV#h6gsT76h%~-ZdqYCqDvA_{$?lxB)Vj(aHNm6IsY5RUj`H!Two|4<f4gmm*78n8ZL#n35z#p>^>4Ts4GwJ^c}=29AIFVq26R_% -d@KZ8Jhr%2aG)ICtYrvuRj#ZSt))6RSn^dzQ$?t<|(A7Vmi2cSG~QvTZ=rufM~j1U|K9)iL()~$?z1G -yoj-_YbSlry*XVoYe!TSyI5-V7d6kq;VPF6?Z&RiZ<(EO&PIRVV}SK!*iYPA$r{K@DaNU=DD>s0SFw; -md8>JGZ;O1hnO|~-aHh#>s)-|fvYOdu{2(mOGTDDnO9KQH000080LuVbTc;hRFJJ}$0O=3_03ZMW0B~ -t=FJEbHbY*gGVQepBY-ulTVQFq(aA9(DWpXZXd5u?5Z`(Eye)q2+R1CH;S!)Lj1B|)AmNi>|A`O}bd+ --c_mMDjdL<%I8#7OhscSn&T<*3;Vw2^qnV%;Cn&qH7wW+ANS1+1Uw!ItEidlabuYV>*#SfqCmMbes$xT~Q494%QqFrtDpR8` -rrD~`FGm~az(?H<#;Dme6OJ11NUP1GQ*HqfHs`d~EQl+g*v+eHU -o68GK^r*zP|H27ql3_wjSsiiUq-jeO0F%8ZT@r%|5%w*bTk9WxekM=O9)!P=neO049iNjJ2{yGUi4_ -TWJF4f58`faPaL;a(-jOc=&m{Xta*%_#%y -tARg<02KkE?V8)H_sLynxgJ6vsc;RBz0~Lr-*fRpFS0J&2>zIoYa3kc1aL;* -*8bw6^w0|yzNjPFOLJ?}N+C27j)q#oXAMZZ1K1r84_suUjc5=J!>&aVyg<@?d3w}r?NAmQ(5QPJz -CXm-L_S?6?Hv41q<4!pQqdJGYRb4aC3~De0rMov&7gH%xYB^V2*fa}EO-_=k_H71XQ3r -;E#rFH+WezVo|1Njt~JhQQJ;J|?#)0G0n^{KRq^EI^7JyY>0DvBv*~EUi`!Rt~o(wlY_*7^x(`y12?e --&|f^T>gO)boSx;IrX#`Wa2-!MyHm -5kM25Lij}Co9t{F54Qq}a}OR%_wDV~&)X)!lYkUn|bXjWIyV5v=EWP{?>D=bKcEv60h8vwC=4L>_I0x}ox0h3mxYku_~fV4LhgL82Dpadr7^MaW^z -1=FmZ@NMu2ul4n-{bPqnXd$hF3)=DB4P4+S7ERzMVA;j+wtWjQ036_k#8OVpL=Z1g_%7-BZ*jt9L6ll5_HQQhxIIve|PQo}u(CYLN+^ -%<{WvH`8-m}vnxS7PSri1?MUrc!JX>N37a&BR4F -Jo+JFK}UUb7gWaaCyyIZExJT5&pivf{;;Ix?3yp9r|T)IW*17UCw2qU0sSr3TX*e9t8=n~}Ns_-($%tqb7Y%7?Cs{`%Q+=Bx$zrkX#DNsWwpV?}ih}S%D>_AJ(+EX -zh+Hh}vs&zSyxF;1;HbH|d%5owul?Oox}S=DR<|8(&0M9ZYgTH`IlU^~guk$cbySJYQ8hTM=2wnD_HE -lSDb0IzY(W(F?s;=ul6T@Gtg375np)oG1qUu&L+iY(c`ji=y(~&mG3iErdiS??@BjI(c=hq)`;W5&MY -H9*zSFvJqh5p$m+|eZu5*isIiYrBe&EutxnmVARY7~TFK|H7clDA;)hTf$M7>y`#=%<8oa*x(Q*ZUNv -?v;S0E1h`th4phOksk2f3rl}YRQbn(`)q=nUe(;>MIGVHWNewzCH>|r7;5B<7 -CtF94zr5LV?LIiZVq5Sa-f&eEsbuwbNxq}qPOgkyzWeU_*5792;%Bf_!_3rpotar$A-UG^p_htWF+yv -;Ygk2Yx!QaFxt<$jkU3bf?6`Jo&&$lTP7d^Qfm{p6HQ9g`e|$P2&RK6mTP3oS32Mh*&dqjK7|tf|r|Ex2Je#|@YKNM@dN%&TS -%63LAGOfF~m@CG&)m0mgz$mZ6i=rxnY4Xt|!nwSZsF=kAJgb?70a7kb%(kUY%b!K?xsNDmTqTKkr7Pl -CMC)aH4nj3SbmWvyf1uL-el_dG^f?GXILU_+;1^!@T*WT$sm^weG#1fEa5G5w4WD={y&pMXF4gH=ZtK -XVrB1tdlj#)YM?-Npl{`jhqfC7lxGx!Ud7fpSHKa4c&w&w8v6;rp2HHM*%K%gaC-hrj8K;iN{qehS*U -FWltALI^`#c$0gBkkQiEX&zNXEmFwQ4*OtQ~_qjr>PbR;yD4XSvJ-dFrR2BkjUyOcw&;*f3)LGm39W9 -KyP5>eO+s;(utY~*fm2IL~yO^bG&Q-)9T>i7j<~1^GQ?PH6J!;M^?yY`C?QF88kp| -p`*>uT+4viyi`7zReR03EXj%LM7I|5aRPd2uQ -P08WB$HS|cTPufW}SQ$q*-)PR|f7n(YNCwkXN5qK&Kuk?uR08G~X}CFoK+K^n_qJ5>OC|Hs0|Y*o|G` -o`(=o3+AGNR#hzUe0}X!kcYG~Z$$_PW(aI81Iu2Td8ew^kImD+w4H!>Q%A2-%MJahfaI(5gw_o{S&tO -7iInpTbH0vsD-3qo6?eR{?rmBZW}C2Ou@yUsDg~cH0%_K(VT(hr<;Y~JfbrHWA(JSqn0Ux+suZ9cJaL -j()2VY#cEq?&U=zm7atNAu-_jOA?!W;5i(i;!!mFtFn7XM0v)J9ov-_6*l-Ba&w -(upnW=9LpAYlznKlp!Goa&CV(9do=WR1=A^ ->;d&ur0TKgh#d#2PvCGr0`KyOK>$GU$FPNltw7OLQ0Z1h#H8Sg9mdVH!Q8&%z1A(NwN@EuX|vRxVE%4 -GEFRff=Shj$mKea0N&Ci%L7*Z8-Z^=X_ZsaDB{!n?j#tdcj5}GftHl+qljYzfP`a!jVpd78%bk%A=+3 -yg+H{250TlLpyMG+Oiqo2z$`x=m=UH#$q{NWg3tiSOy}7g0_tw9E9&r# -54l+*l-A;c6Q32VId5uRgxMeELH)F>-LiJNH9Qnx*G4+1Jkq-!&S0A6j)Z)N(>)Aqd9CT8}j^Jxdq~o%PCup ->PvIpGB)_U&LU4++q;cx@eL^?3%&VYoB@Jl6m}PbD@^UfJ3l3feJpT`F9~+kymf^%FI$j5<_e0X-{iu -$!8p;WBYqTNhfhIGXpofOF7-i*1+eB24QT8};y216MZ-(GY3Agmz_$>%O4R_qTUKH>XbWIppC(kGmE&{vjuPb`ldjuw!|YKXVVjefrml-_~K%0t^`!34=3t5F2wD#LVOPFlW{{zr-83=bL3p-L|)mick?;u$9tG`8`o4=FJ3#Ysd&-6%c1> -0u@p7YX?uMa(8elu1l(jhKM&Fcp)w_IgMVw#X!}@?8!D&4t+&;6<-S++fJsyxkxMgLQQ1aDd)#X`_61 -0m2Vffjlq$gnY%3_xVb*;1jx9 -{dYn7UO(O*3E{kLa=yEE(fuS{OIlUollRKE=-Yu4*0KBsKA3gUD{NmoO*B!{sT?vFBelVbee0u~=C_k -fi`-2>C&B&3-CyjWJ4Jz=7)*BwlTt0b0sL+&s$9A!T0C@_I(DIKer> -lO098{2quslljh}yWnMEPVFt3-Cyjd0VlQK#1YL6Hj+%~X2*@nH+P+p*BXS-YgU-0PuB`%Xu#aAwa|} -2r8Cl)(sgExeSv{+Key_T*xN5$z>)-?wuDu3Y|goJH-|pH;3uf0xo)sr2PWGXr;N)TMA+b6-p;FeHb9 -2M!KA}Uqw9NW)Zj+MNGtLwe-dSQ4mcJaCGLC%yi0pwz6X;A3=;AZscqFv`(W8lj#mJ_;SU4T{&t{>b9Eu*5oL2sVX#jOf=^nfLDP3z5W{Dzfem91QY --O00;of09jk>>>*=*1poj%5&!@q0001RX>c!JX>N37a&BR4FJo+JFK}{iXL4n8b6;X%a&suZ` -?KzzUx;I>P71)LK+}F7*G#&n>0Wg6!xN*%|f81k(UjZR7lF(jpP5`p-AevuAKlu*YH{zlEZm?-*D(!$ -vr8Gx;3qWq9APFNM#5WLK`uShR+J%;cTz;Wthj4b@zH346)}#$hs`sL(a2s|J -P@t8pBdz!(Y2vy!t+I}7?iTv7w?*xrWm)Di`s>@Nzm!Jvm#Z{6!9d}^73JqUAZacdaY8J9OVqukwqDM -|!@wCS@iOjNkO!%!5hNgh&}#Ort!g(zNPGC_gD@sRz^CZJ`r-fjoP*tZB%YxUrC50OnQMVyjY*0uek= -po)lQB*-q3JZ~<^4Tu}R*GvhqKh%}`<(aq+*}_G=&eh^nfrs&3&bhT@|!!cM)aH$3<~)|-b>-7m`B4X -?>8IIK%?En6?y*izz(-(>c+b7?Rlq*Y7H3Kus_Lj;+o)#X%Sht88JH`=sj(@A$wU -t7Dvvqnjoj0*`2_C46>Y~T{T9?Bry{;<|Yi59qP~p(;_D}GvJdf%eJ^Q>E};jpTwIGgscKxD%MymZ0` -)!AIFmWc$H3cOxTpfoC!duwIxFSWAIPb=HE_& -&1E*-je2hP5B$?HrWJR8lm#freFz>$cW#(!ggQ0OWWFMI`r!_(TSvFa=4#4OQYMTn{ -Z{kf&39W9aC?TJ#arzI;?E1BXwpuDu?ppEDvbrx-lN=mvvfQMJ{82+Vh@@M#T0L(4GC)`)#wFIDB2TN -Z%QFUR{F%>5S1w(B?92X%1uhDwDEAOV!?mA9`u4$;2R;h1>%V~C!N{GZaGEAN -k3W690XB}cDL=MUs}R@pPzo&h~n?M~DL2(!-}6)d_+(1eElNO>&LRwIcl%T{@Pt-!OsQ`c!}8lVboq_ -@&-bPMWe4*vUN15o^N;(J2`tS -=I4}&7JLr1l_~^HXe2z!a8LePG0EuwCXk(s6h|k+;e>Y=E9nIP-soBEy1)ajDM$f%*RgSw9;YgpJ`Lc -)?}TMq>clW$(`48Ph?H*IyLtn#HgI|;sKCTl8J;AW>nQfrKB*TR9ao|8`S+#6{m2bKe54j -U9?gD`UUrzy{!{EHd_c(X1$8r1y>^+mn>%Z~th&2XlREh7QmBmWr>~tNicValDTuv_mFRsGYv9iBBcB -%~hXv^D>MXNAADB8qXuppf7^A3B+j_TA8SKq9m%6dQT@&4nYk^IUV}aNQTaT}WZ@6a<=>0#z -`*7DbIiVlPPxLQPO9KQH000080LuVbTbOBx{`3t10OvIT03ZMW0B~t=FJEbHbY*gGVQepBY-ulYWpQ6 -)Z*6U1Ze%WSdBt1ZkK?uxfA?QO=pJG(Y@vNMpf;C1C%5Q*Z1>y*!C@iL5^Xapiz+Ft&rS2ccZQ@sC{b -Sfk|KUtiJTdc!{2;x$PI$vR^%n+xgzDBkq7qgic5B2d3h@i2b!m3Llq-QMwJRDll7h}^FrW(=Y$f)3M -y&I$U&r4#)uMHQexOqEGw9|E45Wp0k80MJ3Npg -dYkaOB6hD`3^MM^PI^O!LiaTBfM1PndPaF@R5`i@W}FnGc_lyOu|VJ1e3{DiUWz`ZBu~hg~(LR3tA{S%}lQ-N_Ypw+0GwnRx!(2o(hmx_q@(eb8-sEO-&dJTIBdX%=oOo~1F%pKIPv)uxaFwe%f!t@39sMUIPim#Y$FGU=0(fXA+E@m*e-m{c885WNKq9QP*-t(g1NybE$e&!1tF(=%|5a0OTx^MYk8JqC8dyaaxd947n -RZ88!mOTPmms4m`TLW3k?BLkP4EDq_6)J}EVG?m6*;cihvCS}vEt3Fo= -3N*orKEDl%D6a{dy&^N*1GXtxA}N;Rvv|Xf+1*iHzVhTT>u}GD3~-RWB_YnTS}O3JA1A2wP|}n3QQYp -sR3-qMctbouK>yZ4GRWlAS+y*MsL#zfgd;a5d&3>%rQ%OsV&gooq>-T>C$$d;(I-;5kh=`;ehSWC$q6 -3BT?hDPVo*6r{iH18}xcty%x3Q8FSe?<2=i0hu@m9e(fmV&#XjN;j)D~Y}AVl&kxx3>|ZP3K!~zK5vN -?bE47AJFimY@r}?l69p(euanP*oMmZoRegexM&$9HR^Vz1_3qiSRA4B1lJs{Rr8Bg7kOrF`<=Wn)S$th -C2pV4oZXA3Iu2`#Cm%j>r6{WtY+RaiXjc7Gs_XsNLWJf9n!;xfUtmA&eE_Mync`jgyazGLS|wLYHS+B -_jmX4-G}(Qf8D)%6W{&t;p4+9zWw;&)B54{-G}w3U`C>d1b+o7lYgvM_piTuw~E)R_xEq^9^&=S_p48 -^2A_wE6dHl64?lWVFkIcgUA>Qgdb?V^i*H}wzFql<9#;P}!NAE>hZDuNtb3Q7w5Ge%p% -f^vKM(B_8t(#*{a&CF&Eu}NtBd5bIZ7160W8V3<(y)&Lb7DvI -XlcbX5xM#V})WPpa;|9>Nz&8q@N0R$?fseS-7Oxp}R93%nL+3YF-HK>tKGk&XZAZ0ebl*-_doarm!Di2sJ*lO1LLk5|o$FmDAk`pp$6*84*k2t|tyu;c+tK8-s6_?n -bvk+hfTC-RZ6s8&je}m3v%-Q_`*nvEpYe_D{8mGyr=J6Ta#E_;CaZGV2uw0HZ%rR_9Fq_OUYEY?27i4 -A*K|0!^wPCi8(;)FA@D6C2=Ud6Ju0VYpCcTDDkBK}DXs{M>K-8(wE7M&7NZ!GVMZMM#Tl=BI$LI?j22 -hQ;fN&kwx&U(x2`B&};mejb>srB3O5|YHMSDj7bBd3+(}%AGTdK+hSc-B?r*AKg0Qh1n!ts4SQ -kg>>m=5~&DnQPU8mknB3Xq0%of=VK`&x9}{dT>+CxKs%r1_2!G&-gri-iy(Hdw)Gi#-t1bAZJ`Zy>`l -IXXZ>v}2MVbo}up&*dK6>ux_LW3d2ZSy;ZPQLIkTd07JlaGahrP)Ec-UY}NaHy;iuI^iFH)LYmA+Sam -S<7;oQ*zHD-dKn?x{)usFXUg)OrD-?$nLMMayxSoy`V8$m93iFE#!s^0=qDL#kw=3F~3lcbBN^;G^!#-1 -AwpHw9sh%wGFhrnY?2w5fPQ+r{rTy5vT!kx#oe&56N`_pF_S+w8J9*+k^L%}1P6)IPc-O->OtPxEuWFzUVc -C!?wZ#@Z5h$eqb_o9BJoJeLESuZ-kXLB?df$E#8_&|Rp3LPIjF89?)|C2j7@h++zUZsSe^=|-6NK3$w$z!JaVikL*k%!j -)^!F8D@QWCfLdf9-3{T+1hUy=93aW55`p&I<@YuO`_{E`dG!gWvn0Nq~~*VRKG?*_%QG*4dAW!h>C$5 -sefmjIsQ6cBkDX43H!K>`VB*ti#@;Atnne;!`c?n!@f}gjjxLO4WleQzZLnLJ*ssbS=4*?D5U4oqJG1 -WrRUP3-g__SXo39#PEB0&VDu$?_83R7`2sG_VDch7UWvtvarnD1cqRUR2llSQ-IbWT7;l{#Jsv7|j}% -)h=-OyoMR6&5X_dO{{JnHwz-oKzw5)I^ka(*9`QrC -5%R2v5WwTN{L^gcXB}k0EEvK0FAX`Ra(=*>rX|hue5$YEQ9|ZFp4d-`v-q8u~8$wI7Y#oY56=Hb&vkP -NFtPKhfDu5dLPviQ^Abt5U=?#eMt*hx>j-;IF0ij}U*CCT#tmaavUY=@l^dvSnwMW0P#bibo%LI&CN3 -DLfz)EC#g2_BViMRE1_URE}30U}@mqkSlUGCSR*2S3Q2PnKPtVu@cX!R5UCE^(cb5-YU$-;_Ow5?VKk -L&TfV#?3m&3tY=h`2jk}#t=(AoaXkv=uL1HUt2D&#*I>6ERDG-GDj?TP7lQecSzW8GKNqX7$cK;X)sp -xtJdlB%@Z1hQji_EMZ1=);!*0*1L63zbCp-=;;QDyf#G+nE8HA61sBl$wtp!~gAzq3O|1X2WUms4F#I -Ud)hrapKhnqjl=y_~CLPn|aM8iEl>;7hcsDF4x?zV*Hr!I#+lR)6MH!e&?*EBX2h^`+85%_@+D6~n~( -{<{}2ZSi&f!$eJw2ZJv81VzvdNkPN+6rzi(){gX}{S-j -ynm8ULQ+xmx;;fC(a!CkFo*9e`3iQL!zX|ALDsQ2y?JC+qWZ`}d^4^T@31QY-O00;of09jkI0%$xQ2> -<}X8~^|v0001RX>c!JX>N37a&BR4FJo+JFLQ8dZf<3Ab1ras%~@@a+qe<_u3teYIK<9*wMnl%1dDFkL -$YZWO?QKJf_z900xi)tv9zd?(mK9g|9hVysTa%EZgM>o-4AO^91e%``pkG~^oFJBvaV_)(v-q%Yt6bTdEm2ErZ_pM3uHGV3E7u^e6O_Yy??e!NGQ7gHyF&`C$|_AK?HNv*_Gu%mfSD -;!mG>cPk8$?`_)MarO`)6MCwViH8=h`Sl_coNBD?cY)T2|;!5pi>^ZB;oL9o0S^@j#u&}5+EkdhHN(n -)@DfNs_R5DC1V<)f&e??fgFVti9jg@A2?xNX?My2UqTUkKu@l?HWek;?w9#u!UG^WVOYPEqHcoOP*0? -sWhwBw1|vR;KhuQBvcMMZbU0Ak?K2PoTP4(Rym{GH$&yg=}oc6nytq-))Bh$>f7?y^vh=whReA`^&&5Hc-D8tpqIhp=HZ*yg=`7op6&rB0Lr1^3*wi3 -$Ddd1x+F#j!>dCx=IM_}y6~kiR&8_;P*8C@4n!`QXHXoKf`Gt~E}=*9|WaIVK|Rx)g=0eU$q83C$hWH -ol*RS#jnuzx)5Qm*-Kx17U$DTWQpVTpG01s0jcl9cD&X12knQO9N5~0#gRBLymZa_jp{b9SCb$L|+tU -Y*AB-PqiZy2s}`RSEx6NU0%Juo&_#45eZfwqdFCkAeXjOda&Bk79Mk?3IyYJmDe&yJCSC!v1nhrD-M$BP72`eaF&h;l`&4Mb|B`|K{<2edxUMo^=DoX0W6wEJ?zM%?x1={s58$ZQ?- -_}EEBLHy -Whmc5@YeOKYl`8T>rkL+Gd`0FmGXD;jEnqA|<;q;Mrg-=;s%tr2Z -JNQ&tEV60b4RffQlp_q#~Z8yI#%1am@s4ao}qtQK7bci?iv&vE9AM -XOAq1-g{ObFL*7D`O|elY8K>Z}=u%#Iv3hWYMCdx)QTrAr^`Zsbsqp{cWRpsm4hA@L^hfj18>JTe;?0 -=%z*4$iF_{LlsSp-1fx3kT?;ha!Rz8|ZH@3Nm#58nvQd*m~9C5qqVJ$FTbvAgsxqU~_Wv&sqvsfLBOt -p{%x{GBB^VOf#!LZ@&Q!^ybmQo4XP?Na|h%p2$@@8o*Bi`3y%b0q53f30ruecYZLZY)(E8pS_#z71P6 -uJX|1qvfo!EomSc+*JFL8SRP|JESN|5?G?^rJlf*<=JM4=`trlg{QAa?)xJDieEQ2D{-ok){^t7Qk(G -7)u1uunGv~rKAQkAXV0V(!FH -qX>4NurygB$s^TZ8FaSQvg!IXmLdZZhMyZ8-`Q@OGwZLbHC9=_e)5Gf68(qpYtv0XRU277+(lVMOYti -f)uWqiB#Ge$c!JX>N37a&BR4FJo_QZDDR?b1z?CX>MtBUtcb8d7W2Fk -K;BBzWY}Y?xD#dllB~73T&s_c7Q$>=(ar!79Pc>V_{^;mE_F2kN;j$vJ*S;Z0pU%5ucw(ilWjqP2P(9 -NgNPzqrH$?>H`R+ppgxc>RRZ+B`L`zr8PBVS=sv5VwQod8)H2Pt&JC68tsy#r^Szw`>2!~r7#a7gO(O -G7QSs1MiJjN?78Q@)?M$9U~Ds~Jz6c)Hdk`H7Y?%?*&u%xmr2t9x!0wE+mVx1I^v`oNNZEH3bRJ|Y6H -0v&SfmU0oRh8?b3CUd7bASIuW5T=s>`})p-S`gbKy610_q(;7enPq8c!*Z(6@212^zSi@jp4qm -e}jZc)f!Pr)iT&Z}rZU{}x5N1mf?LwXVCKJh8pB0Q!1unDqd%ay@aYVkyy^5_Ck6i%Xrm1{7UrOc^-I(V)K1bLQa -M>1sTaWI>4NFPD!U;rFaj^GPg2wRwEz1}7h$z4dngY$L+qcpy!4F8n?{ZgrES4yS3F@F_!fj_^y?Vz? -fOWsFPZ)S!=+Yjf(ZF}ZAeo|%fr=&OkZwUc^#b#6NKDxe=f$(!U=+NW9b5yvtE&kC;i!~kQa8BvSUIHXVIx -+qe>NcoNtCLCdb<3P^ew&V(f(Fz_X^m!NS5++CVmLu(T*b7W0m9^&P+cN$m3dx@An1%-{X***1WYCbF -&ET=LF~dlja42NS%TAI!t^tZw -*s$>!@jPxetW1(W@~5_wC{)5tYVM02BUCgC%4%GX`wI$blxot3=Xf7AKN*#)+i$OL)R~Jc`aZdH*CyQ -eX+WMVhaMXT^%20y_u;=c`vw&%z$ujvpm!xcO=Tx{K)PWyIgRV1Q11Y~wmdE*rjc?}nW4BjY@FA0*G{ ->}WeN^j{AX8hR(g@dI<;k5ozE^!(b!e_7 -w_XO9KQH000080LuVbTYDi$q9_Rf045*+03QGV0B~t=FJEbHbY*gGVQepBZ*6U1Ze(*WV_{=xWiD`e% -^7QN<2Lepeg&b$B2r^3+3ghv81)_wd%0eLMH(dOeQ^y%mS|h8EUF@9Z!XAxznLK=k&>Lp^+PW}Y>S*3 -&ig@SyHZq&33-y~Uy6TKhbk^l)|J>XQK~YoBxh##-*H}ZsrS@FnH8I6=X0Fr@hayhdb7;RFv<$Wt0K< -BB+tTCEP0fO-7YTD7D+!ILy;<0wFIP;%Le80=H{oH3A_FI^XJRk+c8Z3HQ}YoL?L^}TvnB+ChR&Z>DL -%;t+KpDTW2|!4^r{nQP40Kn~kM~4Od@ag&!v#?lPIQCx~hnt4L=50lnpf$mSgpKZ=TY$D=4XIXOx4SV -~ZHvgJk|-+TwN69#{t=h63%g)DiZ80s0e;|)hXP9-~RInP;w6v7h%y{%WqB66lzrYy~>8N1zz15;bBp ->Yi;is~UWKv}UC1`2sM1z@B6hXmjw2Vu)Muo!IjkyP?_xEh{c1$_FRM?Xf6E*lM1 -BWSasQiz@zm4H}E~yx4<1@2Z;9b}bQa-dYO(-j|iwpn`iVsMg*Rs*x>PMwe_cooe0B7uy=FH~|Al*xo -LeJyfkl*5dZINms5YDPOZVO(UU0$(KA|2WY4kXxRYbMKbt7&7P<;K60xgeHr3qS(wBQDMG0GayGW^ -g43EiK`aMoK=_=@TrVmqD>|0(!ZFvnhQFmZy_3UUPZgO?3A8niCRbn*{jL3i-pd`6LJgOTP`PLoaM4J -PsLST73|Tq26&q{(+ohai6{^vSadbxp$iV-aJ2`B95jGrFawG4)&td!*Z}LNXTZrZ8)$maz#Z6Ar@5* -wNMjRboOZ?<9mUlKo?18$ -I5f`G}aNSb_`NwBy?GpC^!IE0NwsSI{Zy<`925ZJp6(xVQlKEJ2|a$ERbTr6n~@~tK0GXgj}WSA=s`5 -HvfhcW-X444wE*$~tl&=xloIGzNvXw#gUCV~Wc^j7{hjw7gD?Vu?$ -5_`iriK$4wj@;b;Htx7cA>GBzHtHVU5gL1WyiMeLE2V>#gg%7^vgpd#A^gMt4#;gy1x_3ZwUl9ql3Q? -kjqJ&K@({AQNDL41ojsX2wv}ap`>SpiIESpg+GXuV2o=5{j=FF>Rq%|s*I@l1Mh-%g_sIqrAV#4Myeo -iklUVSfJ*I`?7YVN#tW>IiL%y2E|SL7bpO#x4TyP}8S}mgecWs(aa47y2TG -35YGFoH#SZwwmzYV`V3uf$bMEa-yl`?1(1{9z#k@6AUzL7er(Pa}?#)(|jyA?11lf%F6?yj|U0N0J#a -nvusU7kO%Rh{L^*t(SrXp9>24Wx`111EY(kbHy~&@ovAM&G(nLmJ>6Smp}xdpu*8-(7rorg|FB9lF?v -?dBtE!%BX{@be?JZsc{oGFJFC0Xo3|5|NY(l7I88(;Q_GtYyS;VT -E{Vw|t)lIBZ2pF(W?86@|prduzjGb1q*6e_?_%T>0kR9Eo9zRo14Q62T(zG7d9LeTDT91Y9$wb=ZdH9 -M5#;1dWXBh6InZd9@Y;886tZinp*b6qIQGp=Gx=8X8?@ElO6P;*}3=(!*DPgol6)!sT0JH6t4K~dqEa -+RMITJEOvFV-X#ua9CP0ynM+Z>lM940RD-6_#RwIHsyj+-Qa?tpAY5X-KlMF3 -Pj&B5EuOLdOUX=H-g%g4ka)VZ}FqA0^B*)0(GO>;)n41y(=axDgy6;3Fi-yU5gVNJDfrv;v3yWl+KpE -ke3p5v@TjNou+2Mq7G=nz84|=s32y_7X`Gka7&?7kK@#7Sf|?sbn*#_VZqX@Fpd7+I#Z+9wU|Y#VK*z -xR2=c*>#|2$JK9pO$e54CszQ?gIN$ScXwBl=gops*{cpr{({sYz(U@KPkTQ^&4Q}*N9DG~>M3Z$tZQU -4QXz)EbGV*V5uMetezUzv3Dugfxu0Ow()*6&9Qk0IE&2J;^{N%m_E+qyMl_uruI?;T>wlwTqNI#M|NW --5Fw@2TLjI8}`dLlKds(A)J9YP(?<~U<3OA(m)^IROj4OWn83vfp~3oH1z -6^+BssF9Gf^Yg>u5Na@6aWAK2ms3fSzG*jYEi%f006`Y0018V003}la4%nJZggdGZeeUMV{dJ3VQyq|FJowBV{0 -yOd9_wsYuhjse)q391PV4}X3v4`CG@Rq&~AMRK`8dIqg9qXT~hMzcjT+mIPFU(0+DsD-*-Ma5lyR&1F -7p;s5+^PZa}xL<;Eh2-uTR=N1ODvSZW^x2F=X36-U_jfJ8vndL_3DuKR*6^T -*K4v?e_lZX(_Tn$4l<`7*V5RC05)^$55Sw(W*h^mO^!lhC88^&5I)zFJp={0b1P=YkBd)Yh)JPIy&~s -`ZhRYIUF104r6x~DpE0mC1r5SmU{3`NGh+JmO42IDcqSkUfUSXRO)tuM%ZEuc{PGM(cZP*B}o_}8u&I -&GpCFX-C4?*;tevEB}r09ZY^vBxM57`@SSZsfUDK&8I7pUFhPO=t%V~$A&ej-LWwP^5>+A4<`kK9DMY -QbL1yr;7xaCi@lapCjK|LGE#;zv;1AupYmKg9N74&J+E0mSzF@pcEy`*QgLyMZat++n^dyCKJL}9ogD ->ByX}1}&0>ndwplafUQLTfe@*hqz&89>|dx%{Vc2!AHIC=;v-byEp_H%S3qA(xoH=S&YQoA&VbwllGm -Xij|+%qdFyAcQ9QI%%v#Y}UdHx6^pSu$-aZGcGogN5saL>!NVY%ZctNDOO57$X{Xap*oDd@kZbfk{c{ -DE(#Z<_b%E1Fks>YjtOVK7$m)ehsfjQ5->7P}@h%EHqN+qU-#E8AJ81h?Nh1l(R91 -9`@qqglAlKZkxwHCs_mvg5{0w&{VSz&`%e>~!Z%0rPR^K{g|UshZfgAq0(;RVzU1qDt_Ge4P0nu%ny@ -?4X)C#-O3ea8L-E}3*BfX?k3}>y5>buU4hU2HqU#{WFHbtS3DltTM)g1FlL4j0clcQS&~PyN1yw8)OX -PY*7c*NbF#c8e2T)4`1QY-O00;of09jiI(q8hL1polD5dZ)r0001RX>c!JX>N37a&BR4FJo_QZDDR?b -1!3WZE$R5bZKvHE^v9JSZ#0HHW2>qUvW^3h3YJ&>9@I1vDR~e0BvK~?ux=OG)1BvF0!POl=I^Gzwby% -)~nmj(}zSN?;hU!9nCgL%q55?(PP=Gq>u7}dr5D_;|;`#Ogt_qqs93b1QwXpVcQg$5fWjpql`M_`Pv&wbe9jWNKT1VWbOqBZf!*0?dNrBP@PJ(p@G(&DJ7cJigZyz3&n -KyYh4KV!7WNaC1}RZakWNr0AV~`Thf@$<@St`+0s{* -lz4xLjMw8y}1bJ_99Bd5KJiL-|r8vxqcL^7y`QT(oll+zNtIxlz%cVG|?b?d2X5POtBygSQF)q#Y^iq -dwEt`RKXInZa>B11wqSl4>1n}n(Vm2s-_}j^m#fRf0|pDrg@a3(@xnzm&>}jx{SqEuvkDJ^>*pj3(-K -4O16cC%#H(L>;+yb<5e2@uw1?dc}Js)AZMU<6&2{bSS6mk8-VZOd5T9{I$)goGJLA=X|sXR|GhLc4F# -9~+;TYFdyG0051zdB9pc -rT2W5wlNP9wP@A&W3VEkEU)M$*Dy^r;ZIBIA1+vFB`4TAoxiBIcbTa0E?bv;dAvnBs&w;KOC^2h`S)pJn#~q5rIELx)|I52W9LE#P*)K;q*#@i^i@y~pdWfde!+s)l0f -gg3uFHto8AKJ0)1Q#hco4tg86*$&+Fy6smhr>=WH`m?Uk6#AS5M}tx3ceChv%mIRlKg_WJyBv>jFod< -pU$(!97=&eQUm_6HiR(G6B3gAUvsg}z-5e#zsiAr&;xt*lrsD~hdWJ;lz}QOv2!>9-V{s=KAp7GrO0CvjxpY_YNZA>l4~ -n9u+0Ci+jEC%&!?)^20K}^=l`pi9%7oBtxnaf1MOth8dxXjR_TOI4UH*|AKFsgG1Mv77zd|g03mj3FI`8q0TBbdgkqM42Qm8S$C3w8j%WH;J1 -v23)lZIXpSZvx?PWzrdiAt-{q(uo*tr`7hz^~c5womrP@EE)-(|}~)=CV511Y$*SP05YJQ+Yg*c+E&^aPp8+@A$*Nf*rY;2qI%SBe-jyncm|+94EyhxMWp7q?`qtT@n29&0|X -QR000O8%K%whX6zHCjbBdaA|NaUukZ1WpZv|Y%gPPZEaz0WOFZLZ*FF3XLWL6bZKvHE^ -v9h8f$Ocy7jw%1*f8j49L3eJ{s2-=+d+c*4q|MbN9hB1lgi&t+J$%QW7J}f8RNWq(thGG`;8ow6;W^_ -nUMo$|H-Sw5c1xqljh4suVSgi=wRKIx7n~o9TCXx!-5S-aMAle3Q+#66KCdIZF{;S=Ci6Bsa+Zi1UW) -wfa;64zu#6I0fuH&hvPi^O@RFWmOnuMa@MK=ix5T!fh;hv@4ItxJV4-xq6%ek-&T-0Vv_pfLwg}^65* -!uCC9|FRrdeAmum4FY=f`!F@pSV$1BpZq4qheE`% -t>*zTszav5qf7y!V|i#me0(;&x8?#zDV6|A)n*saKF9+gd9HT7(UXt`i3W9ASxexP@L6czCioMS$l&3 -1V#rK=5X3SWQDmn635dGz^vT*P}0Sa)&G(182A+`?B!U*U}QUNu{7X1P>v7A((Xy<{KYe^wtji+6Q)! -(C|Rmx3wpf*q)OCM&+n(u^l_jfN)~pt*qmX_E_2AXx$I(LT_=0s<8**?X=)ZqSPvNMG)X~NiumTn>=A -#&XN*>n?o!UFD3%`>R0T>kQlh6+ou|HWgs#IsR?QZTYZa17y6pahx5kd^`4HuZp9Qr;GM87E<#9XZ4Xd; -LoJagP9tVFe)IMZ(P9vdB+Fhp-w{UQ1HD`DMzAvf!*t*<#U1E&}+}VS&nBEF?rGi-Af7zQB_JsNH0uZ -sME}@B*I7;s}91>dBG;R!Eg8Lo)EFqXv_=;svWqf<=B6 -neB;ifM>-`UWaYe)0b>tlu%Fv|4);_=MAt7x#g?SzTZep3ad6t;NMG3OKklsB3~#6`)xz<|HPK{Vxhv -8{i;3}Ra_rn0iR}107HzA1=uuz900H`rl-Lu)}Kd-Lflm=D4WGe5~)4Xmpo7BZ1EfW2$o*9`0#bOJJL -QGb^w06gRmOCd4cmDYG%FiMwif$Ac=VrmiPdS9ASm5xGM*qS1Xq`(Z5hu1J44l!VhlpGb$ov*Ag<Cck%Y}>!&Zw8zQm;;wBauwi6QlAG#RGMCDDE@Fa8r=N*sp+4D80^V*o^)DraokEt)JAsQxh_ -%6=nm=e8fR1~J!ldSo1aSMWtr2x!C-k%c?1ASCHT===dQI7dJQekEtl?oSru2j252z1aVviB20QfDoo -9>V{%mNMFlvZ?%6fvbmkmopBKbaF7;F|N9tkz~7iEo;%W;On*O)dNbqxVG5@lv_5NF1rX0!$QICUELy -tw^SV9_egafu0|-Um|Z256=+o>m1%k#AupSK*s^w=M!Q3HJEY#BX!iC|<~2L%VGJ83TZeyTn}HY)omN -FtkTwI|56LAn`isRLo?0!EzM(7clxSH5>L5|Ijbxeq1$)b(oj*eVroFLU(iskkVS>cZ^{WlNgLc-!w{ -(P6y?)4X&=^U}A34V^PZMMQOF75nTyX)t%`xoxKY;|y3;$hapbnTZc;Z{6=W}=e+jUTax-3=S14{}AEJ5(PF{VW$iN5_bnDa&G^eO5kVbIY -yQ3TorGRwcc@UT*$BrqOTBX-1oq`d&DZpE?&@vTa%Bl%ViLgZ;PHOClE2Z-Qyp4sS3pI?p)#z`H{fAF~d-I_KvD&O@psUNT7ad4j3?r`Dx-DHo->?VA-l=j;^j9{yIdQNvSG66=^IYqeJ$NvJ|r9t?wURkMf7`3{LkEB -@;<_Jdb*>(~GJx>7PpM41FdoShcB@7)FDoZb>4g7JX0NS=#4^@LF0c#|%AqBneLPH6)8ml^phL(LmZm -F~{mp7EhZCd}7RDsLlh;XtMj;(3067sK)mQ-mA=L*pIn2c|@n;3jc$gbc1S!PoQE8~V6kx4N$X0q=QX -E+%AxLBU&(dn%|eFSC}%8FH-2|TkCC8U06L{l!hSy6Sk_PN=VV$?*_k99Rf3YbMA7(_t`VFc;a_JUp3 -Dz%k2?IdoUDHAwWI;cx0b%+?4!$k{8U!qEIiGrxM>)0YlN+%Q10ZsuJ6~U*nk+2KZFv;`-aY>O#6y}2 -zmZBFSMZz_bt?-n)yK(VH5*nexI-?3MIBu^36OCcn9pg372eeB@S&<+E8~ZpK(5tm)oW-6iSSmE`(3w -Fo+{VBpgo*?Ww^a9I3L3`hT8xJPUa#~W#l6cZQ#}4wSZ%*YgDH)4Mdy(OE%B&9tr^+orZDwf=4mTqmx -rzbt%eeC+z(0>Jc>*F)^h69#c2}_yq8fa89_QQP%pYCPKx^$kiA&Nk!9>~)i%Ufhg^t0{dg77Hy6V*1dZ5qwN|&LJa5X`DI)8V -;H{jFNkL^d+2a8JdJV95PPbh+$Ryp_Q&ou)d@D8+h-F*q#Lv7l%ALc2c)`9NGIO(7oYi_3Z|9p^4$|P -A8urY@WuAS*Q`5K(dx$PmdTplp0ASA&;*HatDIlquJZE_G#3fE@cR2jlKtFs;@ -raRcazu!31@j6-dSVS8-3`Kb-aE%jq+P?$;CH?!b57qFtQletLGV$UQJmqx8-K&>+wA6<(taP8r?-Lm -$ifUD*`L)Nz9zPI-vjkLRs!4P4B+4R_Smr)|lAN69?-te&R$=;=N+H$hh|9;|%r1vq|Z_*)Kk&qFoK; -nRa>#iQK-sbK8IUSRH-sTe_9w8R5hipV#`$;pU+=Xu5HUg{Ww8%n(IwND85rMjgO+5P~deskQyZY)z` --(Z?_vZEeRk()XD%L?c^syNC$D0;;Gy!Sw*DfnH*cc_UGGANQ>y)PcZk(;sf%x|Uk?#*vMApe=B?p`m -0F8eO~h{1sAxqBuuJ#?ER+}33VL%y}qvgb%6K2gJ7uz#YFcIY=L#Qq7`1$5mhZufHon&rW+h`iUW_I{ -9J7KZUUGA98xe-Mxt=EptbM}B5?bl%di3Mz(-pTH3B)&f0Xg1nw~9jFcQ@b$7d4P&{?_>tR)aE?QUoaWw7n+CXi3$kg@mbCGGNNB#FZi -YK?Px>48v15ir?1QY-O00;of09jikZx9Ik2mk<;8UO$v0001RX>c!JX>N37a&BR4FJo_QZDDR?b1!6N -Vs&ROaCx0s-EZ5v6@T|%!Rex?1h_(r`!u62&?PB$fv&f^Nw$Y9AZUrUxk{o&Qb~<0|NGA2heXPD8!Qj -8OrG=O`96p8R>_u7x*g1*IHg3iy;O#tl`tQ-A;77@Kx&Yqz`(rMO4|Fo+-@(=08 -|7aKC2`J=bC5Z?Ya5egjM6CLo91kKJJup%;BgA9z`(yT^+bC+}Cp~6KG&)**8uNQ#)>Ge=vIx>F_;bjm=(aM?Ug;HirL!eLPS2F!Lb_0pvEgVqYw-L>ctBpJLlEXq?jF!LD;cv -*poxt-+ug?t?#8At~~8}eB9yb{|321z42uFHt}W25DgDBj8k$nov~#H`|^S9~kJ6RC(Ew(tZglZ&ouh -8j^YX+d9IOh^_O%z8JjgjGm9B|zy?_l+=lMnHpi#ogO`vITgg08^7~38tzo3fM<%Nw!{RxPeO}=KW&f -#J*9i)4<;XiIF60ZKcf6ut94yu#-=5sNrrQcM>^C-~<#Uk;6(5+=sd0kSXR;ct#KB(X)+v>(r@WlmAA -6`itMNu}&)1kqsw9-*Z(lZI7KiYNgQ<=)fQvqOx>KMCXB-qi;|E9w9diFks&>jFvYTBh*Jf1jNUii@$ -=5V1?dD*+BKvgjHr>O>-c!gMahAg`4k~Ng9 -f`sZde1+uyhU9J@E(Ltlt(RocAqwiW#Hw2yhvT<{0?*iuZ>6W(BxXN~$<<{s{@tfHBEyqXuoXk&3@25 -M5g)8vOJXvmSWF1?wA*NilTn0SMU!{0!mu)ixvYYx!q_{f#r#2q;z~=L!2x$7YrXOO%ghibamRC;6`s -v2UY#6xjgdZL$j!nnz*mOHhD|!ufP0!)Fr0jtD&b!klmh54pG=9pp$bK1W$7I2#Wt7HQyF8^E*`P-0) -*sDt$k6zTC0NTZr>hOTaS9U#zblg*jd9A^f*SHf^78avu)drV?KXxd)csD0;mQp22N_GK+pr11z<+)t -5N3vB=b*s+6B{Hnxeu#N1h6rXus=TS2XYpYo|n)TT%%U}Q?X3 -u@u7^OlF^$JQbq?S?4nj7^nIAVfbEjuGPOEqu;_7k@|2dd=IIA?FdSj_7(oNhm{E=G1+!jaI|fQ&`~a -6-9K7~bl*hg(@`%e7`ZTePw%LW0erxP4*nB#uN)QmoXia}n}n*Q!nO+s+mfh;uA}W;B|QZ4bO8gNFK( -oj{1XikkrhAJe5s8H(iFks0Q;o^08Gw#mgHPniqyr}d0r4p#ghwz>HI%csx3UEHh^B?lM?p@dG8&mQC -I*fiS=%Q#bY`CAm~nY&bEBFe+XO828x0RvpJe!X -47=;mY-hr3xYe3Y0K73n}>^>QH@f#)QUYQ7!NMTt`560*vFW>qh3Out0{)VIq46Vu;DyZ<41uJw~SM! -+K|FFYJ9$*!lqK0!stu<74ss!Yk_)j?YYagH>VJ{#SSRNhoODPowxT6l-(4 -(jYO>T*9#ooZ|6JEibwjlFt>7+0g=|0x-%(2I6JHl*Ch_Eo7xJu>JrMoESfwG84FtkAVZz;Om5u?AW2 -BYN?#?qaCwUztNVw|n)XQ)y*!S3uFg#`fLN{nRoVHjtY5NofXVC~3zpUoe_hS+t`HQ+AxpFt?w1|CAT -+I@^fR@c5W^7)Jg42c=6pFF6gOg+GX4NF$r+=_8fGl3Hu9SZZ(gV*mo+J2WUkyq>UU?cpqCUMzp&AQTb`hQT`FWe#6w45!|sE>2>;#vck^`v&6P -$O81NT^*<98593;_4nJW>ziwO@sa-J_RHnxD*%K0v%gQH-Y1F9z+K$_@%e{PGlDZPCw> -j=d71k?wCtpI=Oyi;`zis+mA=*XwSEn@H@r-08mQ<1QY-O00;of09jjP4BpJJ1polk6951r0001RX>c -!JX>N37a&BR4FJo_QZDDR?b1!6XcW!KNVPr0Fd8JrwZ`(Ey{_bBvcqk?Vwzhp31_V_wAl-lg#ZW9=zX -SnGOJ|D>C8{LlRQvTiQj|yX)ddZ)y2U44}5O9f(mML*YB*! -GdCs81+cnp&WX~7nO38;0!Bsd|Gm5WmR@TonB$HHzg=paucff@ot-RBO109&zD`lmw{9(=v}P4hx>_ -t2xnNq8H}bg_k`*4#WQy?WWF+D4lxd4|n&ZQSuX7LRuxp^|vx-V%#absT@uSq(b?XG(fa$RStxyyyOQ<=I9Vh!i1JYw~Yl(l -foNgRdaAJ0|n7CE)xc_goALAxan-UAzl@n)n?)AWp_jf#Zz23{lR^&!`xr*2=O6_NLBpub4&oo2=UG!l< -+6|f_$D5$4TG>F}bY;MWlvW6QklMuUXi;tz#>5(Bh7Nj)C56oOlE}g7a7qRwf{~ZECFI>u7OIkp44cn -I2Om&ws%35?+vNj5Kpsr-rD!<~qD?LF=rc1aOR2#esw5D68u5pNqBd(=z?dDPmt;WUy&9A@wb}46GLj -L5Bzb2UY9$Obs51#g&+)DYnn32N1mYZU^K3Umdmp1dO7)Vye27AC)3Hp9`LS|{QrROm)H17=sV-6A%TK$4rY`UkpKvTb7A -Lti=xjwUw99qGnGHLG{9k3#U%#0UPo4zycc!Zv5@@6n*={XtY)J80I#n{(r5*EM$kJw6Gp%>$O{N7p6 -L7sn=X63nQJVM0KGK)zQkjC4N`D7ax;VcR(9`bUFv22yrzm_=pL2(rWJVJCfgR}+Ix6C_$H8o!vj{OBQP#r0#S4zh>dl(G2cnK%0`I -R-ySCs;92x^K|uAEW!}KpAwgPYbGa6i+qnYYifX)$W`E+H31Wf+sIA1!E^|nYk7vAt^UYy`Yrhy{IUt --3Kfc+*7xDeh404)v6t%UWW1rZ`k=J)=zjwY(T6Y3YiJIq5K++o#2k{tb8oNO@0+;$9iH-9$>5H_#!$ -LuyOTg~pvSTKiQ~!esUOypqm>aZ0yOWcHqq|9ZaLqI>0^?OS=&pSm8Xe`X;-9hIo|-i;{U&zT&T^$1C -pnE*-GXjA!u81wZ+J0nCkV%2|J$&Hw(eNs3GIl9W`{XIEEKB=S8#P_O#^k242Okd*e=8+k`ex?lPvGs -Jjd+awRV0uvXEEykA@{#rk^CQljFuJ3K7$)di>|=e^5&U1QY-O00;of09jl9R$of+0{{RU3;+Ni0001 -RX>c!JX>N37a&BR4FJo_QZDDR?b1!CcWo3G0E^v9RR!xuFHW0o0R}caU$fzx(?QK!@vQ2li-aYHxMNv~h-81SXQ;H!eNkDhZ7+MJ1F2tG -Fy`DF&^&YsgS&g;x}d9Y?5Y-E~|wgVk;Nyk>%qJC9`401hU1A8&78-`!m@(FQD!ITU*f@Yi^S^mWO^J -0Y>4tPFR9Ho8WNU1c6yy7j7oJ*)7by=*G<##lhCc3#luj@OT%8qC*h`wQD?EpJwg{wz3eRY0^jf<;P* -0`4{YndLui*bhIh2e%CzW0>Rs!QG>?56MY5XC?cRQNr2RRXh{VNF`a19{vt4$g?+&^22PyvPMu8vGWe -X*XIL(90K&2_gAkaM`Ujzczvbv{$$+z(xUwa0pinq*1cp=?5^CFkEymCZlvZgARUf4~f;tjzAFPKc_DB)vU`Um*?~K(gyYK1xFg -gk);Z0NZf*e8aV`1KD1t(a;aY1wsdRUwc@=+tEx3$@3ak8Btu@Dl$U71ahS}5>3#3(av4AuF}de6Od< -^99_yaUIN+?Ls1NLbh$K$>!%bIkH1y?l3o1a+uJ_-LsH%R-V%7N0VGWaoauOOJ{~FvVI>@olP>>ByCW -eK-hY2|B8`Vmr9x#v_%zqss|H@to)Yi@DnRbg%U{7fGHMN$Q#}k5hRrNrt+h3nX;Jx+}GL!zIG=O7M1 -{UcpgV=eZ1ob_eQD2ac?jddDS&t?BV^SfiQJuE+Eeeo0?GZO`B4|B2(4>a-+$0_}p}#h7L9m3$g5J!v -(=4GFyvc1z6zitL;2b(&*EXh;>!g+spn4}ef67x755s5%$_D=OFM@@wJ%J_-{aE4|-o_l<8kB6V|t9_ --84tVlgM57VIdLkpEp2Giu~5y2Oc1pch0Z(s|JaC~>xk`h3@Qgub?c`#+KlNUC488BgK$4NctWz^^s4 -eXd3$Mwk;VSZWHTUV{s-sG<~bN}np=&@eKk+bQWBHr|?*M`PpOJ5;qToB?=q7`N`lS1vwutDUL?d5bQ -S{!1VFw^ZQaPVdZU2s!)({Kkf4y>k*1}FXX@w$tp=ETng%lZKIV>Ldu0qtHqu_2lg1_}bJ(+%bXn~bM -&a6TPbbDG3vUTJD1qjEf&YlSv4R|+R{oscz-^6)j63)GU}XhxOcpz3I`-JHQ!hJW+p<(nuM(2jx$p%- -30`}c~zWYAkA=*wGM;dM$DGresmQvTXQ1!I5zim_*-4flojvEVPG~9F9gN+7bCpyZ6 -a$ppfw2*=xxCK_|w&kjvLavA9dm+|w*INuK!A>EY?e=b)k=?iJ>&J(VX^rSMk{PFAsFB#E1F(C*^Imx;ehGhb-jxsPY@WAPfXDo+}E&|Fs675jHOW&6_Q&^ZZN -O0>|}hW>xkeJ$8mV$F2VxgK?tHN{F1uhl61Uy;x$@K(THcDQSjemoj?k%$N@-ip(HQDHDt94;X+}!fu -FrF(1LQ7R6+uAnv@#%JK&BVBm^X(ka3BHAM*(xz4eFYB);g3uE_%mFSN3c#Dc6+vbrR^2@Sn!C~6GGjf}F|=)OxnWJ}J|SuWG7AX7$0(bi+1XLt*1TuEWwgLURH2#F0e>nJLsv>W -^uNuK7@8H{1!iX}K&u;)uKW?7U0*Bdpoesr|VB<53c^tgBQLN>8)+|UTGGC6JJHyFRL>+fo4q#VC5EN -bAvvib0gFmGnMV!j)aGxM1`$JJQUG{&OHll>ZLO_16XxUddPc!JX>N37a&BR4FJo_QZDDR?b1!IRY;Z1cd394wkDD+ -Mz4I$ZtSUgtqMoDfUZRz{=d`^<5prMvw_v+tv)QQXfA8>N7EtPov1Z^p_|3R!1uCoj((0a)6`{D3%Jl=!gdyd9Qt?Y*Xg-BBn1b -HK@g|FC{DG*P0vJZey7>37MU!mpAI$4 -EshMjzk5PK%I&$Hs!9jvX{M|JluhxieHVi?qnUE6L@?E?sGpC1-R$fB-S?{&&mDq(X5cU+pR-{!p*Z; -$0~H0m91<9`2_oskp%OnquYLke1M4c4PHiYm97S*hrU2ssF|n10xq|)3GPC=EK7%lVmIPst*|XJCEDK -kp5gTQ-1EB{eOquvP{@RSbTfQh3V2loz7Kn_R!QhOQu9{LFaMK;MD3oKy0mXD14J0K0zstWQ4&f2P^r -x08X3k^*@cN!M3$ti&lRA!*QbW9?SwEV1!MCygA5cpJ1QY-O00;of09jkAW6N7N1^@ti6aWAp0001RX ->c!JX>N37a&BR4FJo_QZDDR?b1!LbWMz0RaCwzjZHwDD5dNNDp>rwN;F|VpI2X#cJql%S+3g)1AsDT# -B&1@)o+85P5VGS2WkLx -?EPA~3>kZ`IOXM1lUFi>QEMS<$sRC4Dl|l=!vx -`bOm(gyqpATNX~z&*cz*qZ?iq(d0U0tI6wQ=iwp;Cp0qRiZ!b0q9jPE8XG9B=rqe-Bn9H|r7WtvzLbaY>XtGrj -xS~8tUU&-o>x0vGlCNX`VQX_E2i43@Jc<)iC2sSO_N77S}olI#I$kh9J3EX;CqMh^oW=6mAWzhYb|JsReB|rdQ>tg{r9bw%c5X<#sy-rXbKk8G4LtPF&O{Yomn@D8aaI+_~FfKrS*p*Gbt5V2>#R>ghmJGzd|1gRHu$Ig0L={Q%vj-3c$ZtY1QpW+op&z<$}CMx16_2JUR=a -$zj4RY)cn{-~l}78=jp#z}14oNToI*AeEPO-9dInD;A|enA}y<^i#%zPf_c`Qj3jG<XSRk|PTWS%yC>n&y#Th8h#`#J4!ajT}nF7RNu%ZkaT};00u5x*Ku@Tttj9lmdj58 -p@yo_IG%_jt=Jb90aqaZD#x0%8!Z=Nsezfrv5#Ei5+ry$Ds*T(yWSvzb^%+ -~>-Y`<$fIynooEN_kq8UgUT0oaH!wevCYAQqijmn60|qTkm>804zJKP>G(ZCL2Z~3`O$yLihw}t@9*?%;W(OlbPeLtKze% -YsXxTQdo+=CmkTe$`rLj^TsLuZh;X)~_Tj8Z9MicnPTK(D7*y}FHew-nYs9P=!&06%rC>YGubM+VKjj -=#&e#8_{vtZCRQ>lX@uE}T3*d!A=R*^*It#XqR=aTfqx8cP*!9pfPUVNUGy+X}eO}Biv5|$+M -`Q0(mSpGYlMntMfT;#OJ|%^LHk}6VIo4Sg3x4jusMLb5;mub615ir?1QY-O00;of09jk3koovT1ONbO -3;+Nj0001RX>c!JX>N37a&BR4FJo_QZDDR?b1!Lbb8uy2bS`jtrB>f>+cpq>_g`^PF-!s+VS6>;6^+7_C}V?$X+b>KF|=Ba~xbtX5N)sA -FVctk8m`>5WWDlI@W>%}BP{__?EjeI!6hS^OPj -Ib({q$XK!2Y%0mEg%8ShSh=w=`S@AvHb8$_mW9s|1n!7XX-Y#m&`0M`eBOJ78xhs?N@I^FA8)OE*_JTf%GB1?;C(~n4_m($Ef|v}zPvpOWhjIRo$q`WzcwMtmkz1690&a -eTk6PhwGzReElKA6swww&|gb3-~+(=z0M%P1kzhl6hr2W -q10x%BnBBeCJgEi_U;aYLtrnhJ=>b7p5t -tkF?$Fe&mPoR6(^)8hIQ-#z!ANIxyhh_Fdbxm8O^etun9p%NDVc46dq=0DLXvrF*`!4@wyYyX5L^Qn$ -1l+X2cnbF7{%E4v( -W;6BsT`F-k;7cJ-G`@RHu$r59qMYp%;^4V5fW#2gAJ$zfzZ9ECt4zgd*sFTrRWRre&KF)BdWnSwSQz|s{@yyMvv*c*y^yS6U;thPddv^zpH})-98fJ@c6X1LLHr>SBAC&#jim5dQU2 -jynMTum~A_ls7(eC0F1(JD`r`*;b`N!SNJ@Ij=yc|QkcD`_OoUUJpyBJkt8 -M-z8l$Er&9S*zj6>0aWpZ!vN@n2t+)VyZiALQJxgK(Y3IS2PMpTij{zzRiYNzj08*rOCln|jzG(?W^a -m$TOubE3Y4LMBN9iNhQ>Mp&MnvDpvV1k4_T%T#k5<}`eywYWQD)%&)mB5@wjlUHKHs}o=?9DkCj9kyJ -UM>V56`o){afyCmj0v2wl-s_&OmVd1U-!4#rNnX91|9_KTjd}L;k_!qHctU>Gg9KD`ODO-Z$GWfYm*P -2fBZbVwi71J!U4b0VG%IvTy3ndiAD-b~>9#nmullXotDr81jp5s|5+*qrNlDh1e#c!z2Alr?P)h>@6a -WAK2ms3fSzA?UNT#bE000Pg001EX003}la4%nJZggdGZeeUMV{dJ3VQyq|FKKRbbYX04E^v9xJ!_NOx -RKxIS0Hq`hDs#PI(563nu@m0Yflp2Irfwv*-On%aVZiSaZQn0f}F9=@%{Ge2S5TOsCn3xi)w2P94GSv5(MSEV{S($8dFmzDYQi>gZVceRss(o}W#U75esrfu?~`<6H -IXpI9_yJnZvN}A!`zCS;IbN2Ra{Fk$r=X3G5q-dq$rn{us6#2@uJcqAG`c1xz;%ixERUPe;^h0u$msb -%Es7&iLDXTJ1lOq2g8JEdcj(gVfR~V~Nk^Gp-9nIMEeoNN*2UhuRI*sdRoi4qBeOW&$n=9~Dv?}Qc_z|2Q5&fu@w(Z_`X*O$4iK76uEe&=+Cqw(sx30HGUSq(_ --(RoWc^#jM%*My)N)rOsm$i^muia!C;>3?(f}9v8X!th${Q~-W*`8Smua;Hs^_ah0*sry*-%#i46I$# -kj*A(j+ID?S|-`PLu)Ro8Xgwa4U8`m?7fk)z^TfQGHn|vl6sXlbyDv|T4hpHC5)#U=p$BpJqyO&gJom -{iFg3ymUWpFk#9cYktnNQlT-1hfBWW%#kj6hnm|Lto*xm*XZXb1w8%So+GZ8Y$V9`aZ(;gulMga(lKM -(E@owL2KtLw?f+)s7`)yKYrgL~C;PVD#G_U1Wmd!JMKd0x&n>a0GQntJK(QFDX(6?z-j0EqnI(zfx^_ -w{?yl2l~-T8Ws*8?M?F*vL1s%A7^Fb7Fpj2{zme`xc+ZN -~Dsv3rxRIQEnj{gTP_$ZoT)f-q0H#9M*P-YP7;x@&98G>hy!GZG@Z`w}4`?)t#SJ<+{kz<~hs~wAozH -+gPIxjO5o|Ap{{ZU`+d~9P|~;_1CEmBza~XI?UE2u>^2}f24LJ)*RikdhzP{>mR+s#m(1wA=O?ra%;L -Ypk(5kI&TutK5>?N$;O?F>IxOUMevvK#7}7MvCUQLPLJs-KXS8TAA@9r=GMzgmI^mTz=FTi3eI<+{c8 -nGx3uG=zBR9sh8qhG+BC2FxI -G!CH9i>I0lo2(pUc#}?o1GF)K@ixYDR}D#&4x&w9?`pnINzVY5Fu*P3U^gl3mI%dY9?BRLR0t9h1$s! -I0j6ObmzZ4dy9Pi1&Auzy6B$21SM5;Z!%5wTn-w*CJ3NGEQG4sI>k*d-M|`zW-KzWgC>Qlglcz4CS_J -CVn(D5Ec{?%xJ-gMY^G>$!0^|3CJ9nf3Yn-qh}9l9S)R#WB+#^i?cCCc_!ukRh -%Rk71lEaiKS}OYui9$-WmCn*Ix-@SC#X!P=24NrVCWp@8EgK&UT%BAsq{>lmpR6m~5fQ8pnjYsp_`-`tySu6r@QPsk5i -koqM$QL~c#b}vHc{@xJZpU+=NwnX(Q6%t@3Lx#dS%S5u({wT&0V$`Woz7qf+$o%Z>JY38H8}d$Gh9G49r1O>=Z -2C1Gu`u7*X5&f>LXG0QO^Le?-?euzLXVo~r#(}b;ZoqzOBlwW<_sZOsa7EvA(V>`7JMZTH~*mOj1Yd6 -(yQ2?0=zvQwyOrS5vtRk?L8c8uF6KLWHmau(wu>~8~0}l)rSOKL_X`KTZEi`aPL-Si{!x^v!gv*o{p9 -hp)kL1$89O3^u-ZdwvsN)Gqc=+?rhATh+?A;}BjgyU!V^M1Y{`1d*d29l5I9SOK6pE1tRPNoNl9Fc`? -kc776{saZbg{0bocVxQuE3!^;a=95s?)n!a07XMKeht&&ga)S-ggh)8s6(o7D;OSn5-wR-Z5Yk#ds&I -pu9kf(??Ui$->qE9X22$EiXpRKnO7^9$enhP{swg%}XHRrn>2A3gjU?k=ANsBI87Uzjb?|5kD1=^wTvJM1Rr)>@Trs=JG0H9??C{Su@jqoRg!tszRRu&x^T3 -SOj^Uq{JdYFsr{xj -IQ1j4NcCY-7T7GfjI|vbHDc#UyQZk&}Z|`nj6JajUH|~6I{JH8L>6=0E-tS7=x9#{GiR&yZgOEV<<)6 -6o~bKjcXI=Wx*NxHQtSn#4G_bYE3(L)2$w`emPZjz^Nr)$nI=L;uvrJo3FJEYDVbgp|t1pfG7$_Vsh} -q$>O;$1hmX|a`)|%f4_J1h{E7J{6j)HC=+!KJI6SZe(w+uCO7h{Sjh&(Lp{jASYRFkobQK)dijY0L5PSjl=}pm+1=h%H`qKHiPar+ksxo%cS1lT -d9DJ^1!Y;a`hhAS?lq5LxmIfAG*4NO+dw~q`vpa7_^|+3UlAleUufU-ePHo?5MnezrK~oSwiGugD -utZqw$WnhX@oG>p!&G)?DXSI9(3Ju}(YBP|wylx$2q#Xk?Ou-fJy$K3ceL!?OuRieMz~&ylnc^AQ;591=qz^Ig+K{^r;XeGOBAgzlfhY$$?IGc -}_^;q4#v6-Z2N_j|up($-W#1*4ld{o>47|?aA7SiOBZ`0l$u -MB*(ctq~Fjbmzk0h4FO@mI=ARB2pNWVc(gIS5PInj_lR)7%7Co*_OTeAh9Y`@65=n6fho -m3gT|HB!K%uoXh5<4xm}shn4w5X;VhOzKn)|r}dVmeucdvR3CY&}qd#~Sn^LyL}mQNF7IcJ@-ouOZwI -y2@k*^AcXxriqM%C2Btgu#0%qfI%dy+E -TM5h2+Hf1Pt1dIF`I22@=c(7A(gT}Pu2Z*&OZ^a8zPGK -l_(5QS;DAPytQ7tmjvKl9X0*`p^GK|>&8eu?>(MI%pQ1&|2OgLHs2iyQ-8EOx>{i5VJWAIDI2|?gY_y -{`_zB+8XjEdv0WT$1V07pmJn8d;^S+7z^I3V>%4lrRSQ5neKC30$oII^TN}DU)I(tdCVM_RxMps$5bm -cMp0M@;!#t3GIy0W5Q4j&UI#x94%L>ua^2Gnji=;L?;ilTl??nVq`3q97adZjm9E*y-5uH!Q7v8G(38 -&jKcv~Ja5!Tk3zk>WDE}I^PC5#v5H+4wmlVwa;01%aoTs7L1CP9S^jZuAzVooToxI -a%peJWyKb=J2Mt5f6$&wP5WJV5WKMYT<-^Fu>4ioqT-a*p*7wa>2uBSxnb2{fpJMIpA=YHhXgupLazq -Gil0Z~b(dhV4P}MS&FZ3%x94M4mobsjs->$KehloE5T^G6cEfKXCXNO*E(Edg()N=7%_*T`iloF1tg4LxW_^;WY|>m3y(pG6CK> -IN%KSOjha8nbaljP7U82-VcB(Og`OU2~|!Y72Pf6*!XFs6VE| -1Bt{I0b#ec~?xu|34bWjcBt{A(4F=%1CAiO$>B!Q#Aec3x5_$~Vr!F}aOz8j0*BddRh28eKUE0UTdk* -S}`;?>?8g}m#zywDk8>3kdSh@>vaS_E7!#$s@9or4;Ms2JsnH%bTSrHes4*M{pl$qP`)s6pw0e=F{g{ -Q|oTbSx0WlRO@80WNA&aTEWs;LzqTBea1K{gHSCbCHvac=&v;o_O6nvLea(#DCL5LM$_r0F7Xsn1QAm9l!du$#5=&WJgajcl@_`J5q1@0 -(3UaAppG~!uOJC&9<8sg}Z;!4|O?v5j_?4W$-CkEJbsh9~-!%HF3~Ms|!-J}=$X9VMo^*;iwx -7jUu)Qnk;TB?l*Xi7YD|cn|eYj_`7PqA_m_GJ5JX5|L-oOw#L>p^^T!GbqCzVyZ-Qsm1)!JKbO3sh`2 -res_dsvn_G$6AMl2P7Nb&yc3jju+b3tj-~itCt>2a$gfM(B=&ALM>%?zD-Lra_i~t_#@*8Tf7ne=r9w -{XmUlj-vDF^C-54*|<7iwVYWlk(^PEQCU<4_DirJ$C_q5%7xu0>TbZH^Y)jRd9z>ZP9vlYE%UM;VLZ5 -|J%L8sj5|ei(ax}~DcxOqP{y4+9NpCu+Tg(@#qHGVk!dh~FAM|QFz7n?WP|&>_@sM%9_Lx6U2zS}hBO -WqvOq-14tMGhyNpdG%kr2i=O9?L)B1L3zeCx*_IrHBa*D5$I#0m2C;0tFK78M0NXFoXQ|fq%MNgO)H? -FEC$H@9{dw90cTfOJtZJVZ`c<~&h?Qu-H(C(1FIhnt-@AUdD*`}yW1?<3>X6kpCMc1?^42pq45R4hSmkZC1p#lcj0N00~gc{S1Z -Oe{9e!C%a%VlU|G(F}ttUrPcV0OSN#I{M`;kh4E}dqOn}8elq_jvrHY)7L9J?yo9j?7FH`PuZ{o%DIz+a(d=w4n9de9jnLBfCkf -z{oKdTB+Kw)Q9V}}tiB`F@$;1_O4L6}qaFU3e}TfPGy`HI!3-Zr(Aqki6Q)BX*9Dm(&%)Eoozn4vT{wRSXy)=FAk%!dS=SvHU_jy -SY}x}cY?q#s{Pc-h-J{&;LZVNb;nnP(H1EQY@Puyzp6OiB1)Z_>uJdFwea!j8qw1{2AtF>IWebR^DMw -kg@1VAZ*5<;3MOu%3>UZlTLqPQ|8acIx!xc`1 -QD3#t$QuPbxauZ2M`zf*!WRzKPg^wLm21cegZpNxm;*Im!hBp;-{RV7WU7Do2WovxGWb3}^of=O!oQp -JA~B`tU6)$4a>r+5Jg=j9bRB_eBgg-&z|e4(SL`f6cy!j~!UdKZTxetaJy$MV3g)Jw%>R3F5%*5ifU_ -448yHw;Q4&VEuoWVf3qJ>3OM{Km;T_+&%)EV_uE$-OG^sUP{`$>rzyggFrK1wlF#*R#`Cf<1iV$f3r0 -;C~EBJE?UZPF{LBr@u@Ym?_${M*C8d2cM=XZDn0Q&#BD))4pmyb0x+E**_V~KNF|o4s6E{`!aGLJFw0 -UOh-Su2zhBw*>AV$QJ3EOIrFimw}ij_jm&hG9Y-dcnL%xic^d-}mtg?^4<@%6OGT1$71;c^C<1TXcKG -BL0oJxmiU%{nKo5hTN*&pyj#wamgjT`fq!@mN~euD{vYMs{ib=wIm^9sk3RnMd08x^Tlr^Ff5@Qj;I!yXL5Z>Tb{uh -a{X2=tktgKThp2ox@n8E4jmL=BPJ1&IyVtoD_kcB4bMEbea*P$#Y>&BwI86AC^4W`S>ZB8AMa3WbQB! -Coejbz5zXhpufhdKTu0$a*2Bz(t}x2ysHO00RrrSu*S^`%wll$DI!jQI;5tcvXtR0%q^+CIW)Km%}N(DX*;#woqyLUTRsQ@+-td_&Jd -UWrK1vL_|REh5%Lu>e3N!leFXQ@#WLkH^H}6V7wOR@d{=sck3^m5bI(}MM$9x=&C)dGBQzy;{#bVaEraJ< -#t7)o1VKQZEt9@IOZ-#=Wac;&d`rBEl1aK<=b81b$$A~pPZW{yRw8S4+V3|o}2H7`FE$Dt^f6l12fvm -8Vwd8csD1F8C?O)H@_}A55wn$=&H%<94~#8T(|~;)#E$qcRdo%P&(;;Gq7Zm90PxP&9HoLsuGtet~k1 -a_UzTuA6}ltbQA5{7w^t2P*l?jn^B7Z6;@vqcnfEDg{DRBP)1jH>d3-z7TL7M`FO|s*C4mPZ6MQ?%|L -9P!JeGGP!%s`oGC5ne%KKyy -Pl?Fb$s}h_4E*Ltf;3z#9Vi-%x_4IZH!KDAFvgdeKpzGR3nS$9y^q#gi8df>%j32QbB{J(@DK4{+ule -eL;*p7;73cVu?;+F}b<}y%G1QjLmLdgJhB(tkvMU5K}2K9~5SsYV&~(Aq>q|HWtCk>}OtX -QXOEZYJ2OGK(xk}Tvre6=pDe_aBlM?)cTS%44+%(K=(P1Fhh9Y0i%D6ugf7RC#6D-0e?^F>I@2W -EQ^x*BRiRk9s!!u50@QjYQSh1Bph^7~PKRYcJW-YqpbmOrd{8M2Glk*Cw8xr=~&2VFnE|kdJg7Zkc-Q -`7NPrb}cK#i=*gf4=TJ%>Wh)Xw_JUo4(nir;~E4*qfnwgstHq=46Lq3?L>t{f&GS+^A76Z2Cw9z_Beg -} -aLN_v(8I)!4~4TXLO)zNo)|FR+2V@}Lsh10y1%ux|oLD!`JdQib>5^J>`D^!^J0J -HNr?$3kDI4S3-$i_<*}pD@{(XfxOT7|B{y~bf7q9;I^d&O$?A7bDSMPeAsXOAjRu7WoqN02YG}5PHfxPH0YQrSbQ2E;2s$^f@c=A;0f -p#{V>Ie_qeYCkZIsdRzuc-7bW=Dpj`X|_^ldTPMG|O2X5QXMOIS`d*R(qrk62$TEuqXu{QwVS-z9%Aa -*)aX=9Q{gX==PsQ951|AGVCueUcEbIvgjQ#zuKU+D`=yq_`o6dq5OHQ`TtEzBS3q} -6ob(h7JAWGHXdTxQ2{)h(lJL84z6NQ?Y$O4%fjoREn)i}HOcwEt%w0l^`Wgbf#MvSrz1b@*#b(0d?^R -1-6+B5U&}u_NCZGH*6f_y|M?tB({^<;F(#@o>@wetMcOB)`i7r`DC)|C3d|41*eky%r*A`vigsmeig5 -SYL^E=+5{|``00|XQR000O8%K%whX){rSd<_5q$}|7~9smFUaA|NaUukZ1WpZv|Y%gPPZEaz0WOFZUX ->)WgaCzlg-EZSI5`Xt!L8vGq9UP&34d6aBm)$NdXcxD8Ee=7Tr7hZ4BTFhtC2{xmf4>=iiIgZOyV>5$ -`602%;cz%#{LM&P-X2I%Y_DqJ?Q7cQ^uSK7yk}g%0qxp1%{{Cuf7PQ)V!rsBSB -+%6p>@8gE4SsLgu^MV#tHMPDcNJOX%7e5l*4W?_?owUcc~}lRtvCMFJO7e+%ez%>+Epow(*AyLR+d*DC{1!$TN4JS; -OW869O?6b-wL+G4wJy55-H9cKN;n^tblulfvL$)om1ITR%dQ6>s= -f;2eX7$%mHTfi{HiU~jALogP`Jm7AIhK?>CMZ@< -?Du$Lv7wz%Nl64$|YU_#M5K9Qw^34#_CwjNCuKAWE&yT$o6_a4|QU-&Qq)EWs|o8@OIwyyY!;kzge%N -|2RHNH-CU-Yymk%x_@Gl}&Ux9P?U@rG?5BMUerSw#qfIFoBa!Le-c9PlErEmQ)P{D&)+P7{?Jy7~9l+ -i8^Iv1#Ve0YX;*+f>6##4eMcD)u8df^5?*dHKLMeX91oAlPxc;5}g98UDHB>=Wy-{{zHDdhts|1JyX- -YfE{4Kib=?~9dCE!#-yuT$EajmLd&u+6DdUTHY4Xhke}Oz4X%g7Iix$JgImEZNF9X|@4QWxqmgs#O`{ -pAlENwx5)UFl@(}?@M*CW>UeQ{xaB9!$Zj~tATkcNv9+&L1#qX!$_nmBUhrdRA%l+$!Kb8T+x8C>Hh; -O+Q@GeY2B7c$m^;G`4g&X~A3m@*Qc6zhbK?paU>}iehB=>L=?WrKUx?R&6w#B5gBxOqgO{vf&$&{0(u -8-%_w2+77*-LbF-UGmzoLk6M5}A0pxqoTuBhlBXJC2w5KzF$9Ib<&OnLaRps2o*V!2Q2ZvoBBxz>fcg -qo~A=9?#A>U>iu<1hHMo?pfV-2$G1o(?m{wJlZ{$F<$}fKw8u+A`JmIoC+ql{DixFzIwJTKTvr}Ls4& -1Hi@pa%1fYN=nst;O>g#X1p%{4hk250Ib(k^{D$R6l08r78bc~{VO&rPMChSyb{E8KKvE-8wj_F^C0w -4W(8*)k2QWm=G~t=$xKX7Nf+O(f5|EPIt9^%ijA;FSUE{!>=jJm1f{POE1RRPP4qn1`p{tS`A`m=n$v -P;I06$pTT_@Yc~jOHg}UC-h|K$jjpmI8)Nu|v4*Gt-OI@KW!tn+0P>(xW*-U@k*$-z~kZv(H%; -oHiRk?;etN%?ilCwgf3Qd$ze#2zY6NTwIuLMFbK|sMqldCuIw11S&&?giMbkk8=BlT+=^~u1b -e|x@YwMsBKCPZQ0jY5DEAy46>@4dw_nroAJ;#XOc7wG5OAm5D46Un%m3iW+^C?k0fJRm;2eF=%Y>I~zLNJ^hF8X3XG@QV@3RIFS9L~2 -l&omaq9YEt1*j=Vx7ch%y#avTjlxcwhmmNpit)^JEm -~`Zf0iRBvfpm~G>MAy@h&;^LI;pWt=ozVSs#^gv3Z916H__HjmKtN)u3ZUQJCpM6ufCUhYV96bMgzgJ -Ac6Vakrtz1bpjE?OnlAS#2u}PeF`msvKOi4Fw$&(H;=KaU8Rr<#THLx&E55kBb=Q_)gJ43292?{lqZE&)EW;vsDx>zYIdhYGIy1IsN(L9;7B8^HJcP -B7tMgUMwAz00m5X0u|7>18Wri2s^nk=DNL)pf^;KJoz_?vIptydo&2Cj??+Ae+fXIW6#ow@fy?wcItk -+Huha|Vr{azsF*ZJq?w;}W2BbNDiCjEJWt05l2v~>nsXDsSsh=0q3~L;}BK+hhW>+=9y^Tp15@37JNZ -EHa$mg}WmkPq*rAaD9NfjFNSi#sG16lOrCFK$L7znatBk-XKD -d-U}lsm;Lt*TZDP+2>C^04L!=lcpS;U2xmo^LHxU8?`y4&I`aqXNSekfs1}SPO3{TP22Gpo^YWuzFN} -8YZsJ^fIwkiKm1qTtV-;h&}%t-Tx{erc~rVqhj2CuvP)9l@;0?BSp)bv^^j$}#9$*CAXV;liIZB)3@=AE6UUd2GVlvF`!C$GaeT}|VKgh8!9L3 -ZV@ME4HDHE(BW|`u}siD^BY=bUVgH$;_M=iqnG*nd7?jw$snDIIMn5=>XQ~h8Tlx(peT)yQ;9o>VN`LB>pS`mVv`f9LGUU>1dL(nFq -t(#U=Ph@{S0ucRF?G%0$ghyeLpgl3tDg0T;o&aJU=*&NRZS1;lA0`fGGL+=^SEVTAJC{v*&Qp)=v2}r -cJ1`lpcf;L!YqCYWSs!$j)lL;Oxf_PZbuC0Osy~ZaW0tuD6rUWc6id{qf#)Z4@$M4R;!ObEc_G+GW8z -T=%unR+iEq8%M}i^afT8 -DfNE-3D?5P-bt~QEg^gPoLbX~z)>xy2da4wR+A^P!y?QLU}s4m0>d0@J8jexdk9~6ZEq>c~Tv@`^bE& -M>OS)+m>gSzayX&{|2Fs9B?si7&h|XXi%r5r5IL$PM*s(cMmI}FdRAJ5hnF4tmTDiVol7uAr? -`_qj$$3KlMjlm6KB{63Wy(Aua0Vhp1?#};gdYGgzO98;j^v?*5$ZX4Sy@uzJ99S^+YjHvSH3RW=H6P` -J{-~SP&Qy`h{t==i|gXUpRTU16EAJ(jQR -?we{SWgwl_`eQIp#PVWTnqK60ArhlL0ZR@f@f+Be4 -cX&f^-?ugrqnAyiI{)MV8Ttfu(X2upQ9%7wdNdZcE`Z}`7^J2#B=OOJRyACxPI+2f8CrR}bOdZ&aVd$ -V|Y{!f{k_pB^azO$(J9@X%Faj#`ZJ7V`c~(RbZW5O^ZmqbXYebefa}R6WP)vk)#G37c7*!0N>PPWAzu -E?^cZP6cFy3rbnT4|i{=sI6Vkev>R?0|XQR000O8%K%whn7mPi; -syW!r4|4H9smFUaA|NaUukZ1WpZv|Y%gPPZEaz0WOFZUZ)0mNaCxOzZExE)5dQ98L1-u-1G3tEST_V! -u_4QV0NENeX*VD!2u3=|Tqsc`DX04PJ5rP^I&x5F0uoE&-SJ)?-ivtMN@a-d^{P~|CbBgxQyK^#7Ctl -4fnRHx*$RF?E?wi}%CGP?rMxkq8Ya@5<=aWX-&+82Zr?1D3O*9%b%C36+2|vLse-RGmvzmWVq}U}1h0 -?qwc`qEXw3Prp1Ajq+-zwsfHhs4tk%cmYQs#YQFH-%6nyjP{QTnPW{IUnI-%3H7czgCoZ+4l!ZqD7!3 -#Pp(~=9I4XYQemn|sfOiu4RE{fqlHHzgAphnRZ7#b`riRuhTorJ6^-c&QrD}3`DN6mE}I@4~4lynD5q -iRR|HgFrLTUzi*)mwc29V3w&sruDQxXnqnS0tv+8I-P9t5q(T*5nG^Kd=fzXK`V+fa(;`LQ_DKb9`;JKU+EW4aVvElEju$A}K14WZyuqNvEyj)LsFiTgZ9 -I(G96*{(OEzS|$HQ`$(t6^d8k8+r#*o -4CA;C4f76WNHKR2@?cfq;vH0Ec)|qJ8L7l6Oj>@WG_sY-K|{mRy8d!3>ht=@eLZ(`k$$LS>9JA+OM=i -ro36@P*uM;0t*t0$q-B~4+WHj6$SP5 -u?A7D$>u%zOEz%kWfEdvrYGf@I$hIKbEV`NyYI6&TN9GytIwha5sBIibwJA{b^g!Nv_4MwjmuH -ThgCbs^#+&0mOzQjvtV2ck9|9UlF-@M#)+(?pC-FeR4Z9boGisHe%)_fjDxf|rnhq)G1>PAv%(WD*Mx -QF7(?DBjfBNy~lTon+>>P&e#}J(gtCCk#gE3e=PUpAA^MX~s~q2k>%erDWG0&ANjGM-yfnr)-uUtEF0B($$jL+h7f;lGy};^>oWuU^@X<)Fa~Jnq>ZnQD+ebZcb -$qoL>4W0Qg1h;4bkVI$3L2FBLFPY5ZALhegLYV3%xU9nBr4vN?LmNgYNF&rBcQg)_Ob`70?mEwTe`Uh -(U_8lAxh*vL-mLhlLAkQK#e?pAK#DAVKMZe7m@kjLqxiLGXEPcE+ -^GYKAUsrz#i|y3CF7rcB-%*GVsmBPv)HfA!B#lBXclrgr?$%1_rW58Zygj4&k9?*k2b}y9 -^dv90fjy-vRl=`;&DaG^*ne%Q(zCvbcPZiwq$hpOvMr1a~dUXGAZO^BP1)<8i -)c?v0M5X@uQU%OtgV1~H?z{%8m4N5l`eYMBXy74Rs0lrBGBj_lO4gzfaF&ukGU%r-qk-Ri*IW^i!u>z<~-Ug~Zi;Uc8>o5i-v03=&hq3rj@;cnuUeC1!;}yNqHtE^DnExUng2C``;5%m7Q>yhb&mFS -b)@%IqVJ~m1oj|R>hI!O}%o$M|7K+2nH-|@6aWAK2ms3fSzC4K2GVK=005R2001BW00 -3}la4%nJZggdGZeeUMV{dJ3VQyq|FLPyKa${&NaCxm*TW{Mo6n@XIAXF5PR$J(>?Zul5bZLqfNYbEAw -gRr9vKDP~l_`~^l6aZ_`_AD-5+$YG!xSKq$aCj+F3d)jdy-|FTGx_g8R7e?l$uZ>N=-E{g*rKLYel); -aU}sEl)pFAX=CEJr%hZacV -ywW0CA@pJTxkYH^Gg{GN-)xD|xIRv;i}O^hzzWfw26Pv4v^ve)l!E?-T^;^y@9Y_S*x)E;^g^%hxb=pX~{p=Xviod_*bZui=5qizN^F4+c8?f^ -j#^xq!v&OHLpIHg= -n9NQHJc91oj6e{~m#X0iHtUv%9p?RJ;{#dbMlaT2Pa#;#Cb2=c9u$->RytLs -^lc}+7v7aDFeT?b35D0vaaIilQ1wz8&g>e6Uzy4k)64UIw4eUyZAHFhAe1<%{K(vfu#%FlazyJ-MSJ- -7aVX0Ip$6{N46U41_AvBbp{Yg8YPV^vjQ+c4b7s7=#Faz8^$Zk9vkAujaG;w)jl5$Z`?@fgb@;a&mqO -LJD%$u@F2Fj-a?!^?i<9N^Rjav@5Lf`ef-_bznl-lFLRz2$z>Xd&U^X;ERK+CjkpCk7T~JBR9qG!v|1t5pXl2xd$I2#DjhmA#8#QlbExKTUt;*$VGz!l6O>4Hh9yH~QjKXjQJHAzSFA&j*H8H)9Ie2B1IMl<4o8M&(ISB|h$$ -`0CXXYSeb+T=gXBPJNLUJX*N%!cI+pH1}P;{5&9smV8}V!+P}$hsv?4m0v##HHr -`#eq<}1x46*-IV;v$Q^~+)a)5CZ9C+U-P#N7aDr7O7B?;mE?5eRK@YS^*av_Bhmq4aMd*@1sUdfatf_ -$J$V2s)#3Y=et^9AXCwNLU_BBcJ2q(xP_clV~j|YgB?tK-K9osZ46#JHq%U0SdrBwwNirB4mY7;(BF -cKOOz|Gl1z_vTgMMpjX1XcxIvnJ?xCs1oZP#}kI+Mv(95Pvm^AGfX->M>WEnZWClRTQhV<4{ -{5VNmn%cIN15IWUyeQ1w5I#0#4DVp38EGd3X=a1TpZsH5Y;TSOcaa=}h6oA=Jd=6tfZm6b<4rs9a&`< -t(Vg1k= -xoZT41jpkKA%yWp(KR8+KOds|qv34|ayN)69A`eA3(u;eKD#dtqn288!%Kp&LpqWjGR!k_}S}Qz~o*) -|Y0#uO9O-%)uc^e*)Rl`ym>aqvu;N5qJ7fLY}#$u*v}mMmlcG+k`BmDKI2MG~v_heFK_RuoB96Mt<&J -v;NOkidX2U%{Pl-GOX<1@VVP-dzX&UvWbs|YZPGBpIDD=I&!R2@=L$E20~ttAD<7i_6fm%^O;^|3w09 -+7;`pUH<|~WjxNL5$=uo8zYJ6S75nxc*f2c&HtKQoQABW<5&Zl>;-P=f$WKX<{#9}j2dX=L+%PpKw22 -;Y;Yi{KyLL=E=g3XsIU2__dl;xa9wD#~PGM)n9$YGd(udMRXHAowf3N6;JpZ9%+a@)*mfQ%ASBEoUe& -e{`ACi=x-upNn`Dd1?OYk!5R#U7Ro_0cDTAT;nFgf`bP)h>@6aWAK2ms3fSzAyCrI@@4004p?0015U0 -03}la4%nJZggdGZeeUMV{dJ3VQyq|FLP*bcP?;wof&Iy+c@&Oe+8jp5eX;jcK2R@aVpTurhCQq(jsYh -KLmk5OSGenENLk!iMMzE{bq(DMM|=JIHR>Ca$cPGFgw{CSeEU2)k~3OtUR=hRE*bkqqr)Yx?8P=wW`_ -g%X)8KJ0?45$AZ6$^$SkstIt6DBQVNibNRWCZdr;aUfFgw@$Ac{ -|YTZxt4Zp${wo-d8FU^;0REdU9w>j1HR|%FY_?S(ZvLzy0M;Ec``O)%vk+o@@3|lVqQ -7-h--Y_OC$+Q9^X`2XLZHH5r(s~(4iM4Rr3tGLLKE5?>ct-&^!-r5e~nR)(DtscDpo=Gv1aGkg}9= -Qnc(?@F;n8OiB|r(DGka5NC>HhgJ*l$^lkWU?5idkwJQ6Xv>CHOx0@k_#*E87yrF1!;M4b{Z;1O*-6h -7X(s14pbsov_Qp`)H-j*AFu+nv`uKJ^QGoe -}}S<)oY-Lx}tW#UhaEf`CKY@+DhbqbnE_trNft)`S=mfO=$KdTX8l}MPIY*eR5CRhxK-rUOE;^?H$iQyQInzu$xK{`ik<_l~>te&c>99+If61@l=ixvl -a>_)8s@FD$9qZF!-b)DDS5KN6Chy=OR0Gj5m1gR>%=pB(3*Xjo4H6vLt}hsS;=<_(ZUQM>bw-zZM0cdM#|Ba)2Co0@Vjpq69UD-kcPQWS;9S_lId26OqOLvyzQdqy$oz;RSn)^Ny2LAva8XFNOJTPPRTje3iD -X~CfaA3?W)`asmuOAC+^93&Wb{9(cpofmlEV4Ecp3y!jYIGTLq<9Lgm!&5)+jSz0 -D15gG4XJ{-~kj^#08m3xE-*@-p;Y{B&K?{R_!p0hRXfU01!?xe@j`Iq#_(jL?cA~*5&J~t&93-IuV1Y -pjm?Kckpa?L_ZbDCDPX_hBhrUZe{^(kfmvE*d;~fp~fBpVDRy3lkFBEizatFh37i#*IrK)ngn2;u`k -MRT^VKGbxSNxgVAt4k+!MxAsXzL$1aN3V}N%tEOjzqfi-NAKijVQ`StJowvgXr%7TOw~D0hOFYlvsh;!Qo{omQgBcN{c{efzbJfm~a#6$(zHmqgr%t-x#2D{Ls8o9mvlo(yznQ -h2-C40@j_0yVrPaFp8^*UInjV87-g;yUyn0~_pZ$n@ev&BF(89fQ2$f4;oyrzMWCLba~-TevJpUn-%2 -tKeng>h@{F-D9boKIocLh0g|Kt6@^X1MPdF?2YdhhGo(AyX_J#9m8q!Va%f7JeVIKSy*~p!5Cw4&{uw -=Yi^qGpFS=inCs8k@5n$Q%ZSn@4+0DVa-#<`Qn7V5Oz@WF4AK+tJlHRK+z;<=k9J4=p@?{laGJI8+)T -S8TE%wU++d=K`++Z<5>z=U)_iUotd0=#$x7AG3$jeoBCgLhR(b34ss-5EBZOiOXoszGTUG-9E-S}QQa -%dnMHl#OiMFjoRj1VtI>M6ZCntAs+E<(yPN#q#2^&YJD!uo>EstBaT*9K&-vziQ62}yi#4Z0o)Pjb0R -*$F9+F1SV2g9zoYE!z#af4+U^iEnA8vlYhHl{|b2|Ob5|U97PP|q(3uig27)18%G~Ad?itXyZP)h>@6 -aWAK2ms3fSz9p97vS9l000~h001KZ003}la4%nJZggdGZeeUMV{dJ3VQyq|FLiEdZgX^DY-}!Yd7W2T -Z`(K!e)q2!2qb}IAAT}Mn>jXv|Mc`B=9~^1X|O`&W}`v(EGI3_G`Z7e(6(FvJ+js5rEt&8OQ~=*e--og3ZuB(a?>a- -EH_#QYZe}uXdvqm%68~}4QERpu*Er1E%H)W3tv>{WS8G%_qm$Od%6yalOzeB1243wM#~nhAF>8%K|Gf -TPD3-$Roc2fs#u)a;1{Kj3Pyks2G^~yTlk|f2!*t+Et#HcBrO6!K -W5ZkJe_4{hUGvJ%)d3s>f+>fWWdEGH3XL>V*@V;MDBHDiI1f!BJQndN|!x8P0xMe4#H%KclLs3k7b=r -Z4OqRLF<3r9?y6x31XBzslxG2Qp<)(Op{og*USL^+!Nm0IQd>q;G(>etcuK*wJ!r%5zeOq(zC_so5O# -K!+*ldsJ3B`&--XwYKO?FFV03U`KZ;617Tlas|`>K-AYr1W(0aWD(eDE447icn_M~%G3L7$r;r)m+O| -yJ#;*@(j4W`bw??8u7dJHR=&$Q!@tx64y?no#YD>?hLv1#c%&N%u_ -XmC~UC^hKH@WrtnHRf`#71sCM$6B#=T4u<33{0b!P3r%ZKoddLy=})?4y5@&+0&0ig8u1;hh1wQ&b2A -+{P6i_(DTC=b_(_@U^CKZA1KuYpmlA!jd@5{+UleI?1)~G=08@su{_ihQxE6w-wJ5?@DoydL0AC``oX -32xr~rH2bqzp2J_fK+R|BOt<<04!;w%{`oE(J7fPpBuTo`E|zgok3!P7gJ{2Pep7ysk;cPgMlO`IH+s(%TrZjfrnx<$zYl~S(Mf)P%3GU$i4 -;pG@RSe~bg!x7#-7*~~_u9mTjCN4pK9i~BnXefmfri#wqlRyJ{{c`-0|XQR000O8%K%whIZn|;+6Djs -eHQ=#9{>OVaA|NaUukZ1WpZv|Y%gPPZEaz0WOFZfXk}$=E^v9RSYL13HV}XJr{G)^l>uAs9)edTsWuzYY-YlLD$~E2ZD)m?Ra( -_dwYKHj01LD~xa`2pJh2sRnv2GQZn#VfUTg=xKeqt#nnosBrRW)px=;)YQ+mC!;{LQ@DZ>C9>HvI6POUAt;AKzswFrRH|&DXHisoERbj5!*4 -oxwt=Dc0(}S1DMWj@yH8g9y8^VcD-1)zUzoV8)AHK_QWabdBD6@Jl6ezyj&!dF<^hNjD$JKM5B!M%SuNP=E&D|^0dIO}IK{F -^d?q~zBm;+ocl9`59*wMAuvfTHk4P9=o6Y^AJf+r{bvv%J`IKPAJ83gz(xTH;Ga~X9WShl)IZymLboL -A-nPP8H_U)2}w9D*XGc7gxoS!N|gPIcJQRxR-G-Y3K(wKR1@M-aF5D)uJ+Nk_qG+ghK#btu3k^5(gwr -@tHxmX^5aJR!`xLDL1{SOMD6s}-hDmSh$m&zJlT -q+_qxF16vteU*t+wJ8c#w2-8e5jP6QIp?uO|lo4qgA6;xUGq5KUsXxtw|~`NLe;Fmga7C2WLvqM=lLa -l%M%~)x5IID`%qe`*zqhvAXQA!70=ExJX=@xRu!C2CT&@WS)l&F8lv`-VQ{|VM2P$;9wKk<;e#<|IKT(69v&Ql*H(Q9&MZ56T$zt~~5~=wE1WoxytFD -cq7_TU&!STchI?vp}MxXcG(@_cj+SuUh<;c6mC1MdK!QxNR*(C&@>8^1t!jgsQT!qsYHlcchUs4!*i# -74iH4#`C4?V=0!ac5`fgVlTP;3!w=Omp{LW1OhG$BOU;ktPz6YPYhIrqd~YS~*DiT7R(y$NcHPbJ9-% -K`7hS!RlBV2cl7!BwE%mk_b=)g~uLvM3XW)&qs3vrlrgWF=V0@EE{G)!4-~PCx8t%FvG;t@@2r_jo4Q -=Xkn71k3C*R))zG#f4+Wp^7#uN{}IMmLW(}_}2 -f*#K&cKro#3&PT*3L|OE8yujT|v+zGAdN{IxN*qG<*hs~_hbK7|^HMW~e>yh;qdrlmu~^F?=jPKlQKXxO+H^^DR_WAhY>|Cno1?!>p0JcyuwWw;pst1!% -IR-Vaa@I+iK8MxUgS2k=%G5F-qYOS1a9vR< -x{zP39R176P>nq6$hgtsZ@2))AuQ%?Bldgp(vcccrrH@=U*b<*MqV) -JX06Mq(PMiRKX&A*FaCN*iM;l-NkkfPtI8B`xUYv}?85U)0&w>j_LX*Ut3-(>cUdPkLW(k|<@wIO<|L -EpG@$nh28eshmp3nYyR2FQfo8g-IA5cpJ1QY-O00;of09jkpbqy{%0RRA60{{Rg0001RX>c!JX>N37a -&BR4FJx(RbaH88b#!TOZgVeRUukY>bYEXCaCwzd&1%Ci48Hp*M0Z#VT^^u^!C*V|H0TC91}9NkQO9m& -XQ6LDC22N~7)lK}$ny82PpR28aH3O|lMTpLo2()qn2I)#vpy@dN1*29IKWFG9bJak=!L3pG5EfmC_Y@ -vCEiES9T3e#@YNT$@QmmFlM5fT`NeL>a);-Z7#p-fDOF(&jXq)?i{`cC37zxb`=tUoQDc)JU8@y5Jtx -!J4EV)jvBS`^^`IecgUCWSAnfTIZGu{m1H%KnRTOk#5F&RTI+AI%>GZ`zf#8R<;eVX)dB6*_vQ|;LvK -id#@&d-~sf5-BGSDHEj5?rrY+NP5g1E7j=4!IpWMdq4oJfXOXXC8>;bGXs?w1j6rLJDKLKhzUXokhFr -)s|WggE(pK+VKw=%n_$OvN2}T#I1hU%g|RA5cpJ1QY-O00;of09jk6tDgJS0ssIe2LJ#g0001RX>c!J -X>N37a&BR4FJx(RbaH88b#!TOZgVeUVRL0JaCxm&O>f&U488kT5WOr068``L9oBZi77Qzfp_iQ;*`gC -IvgAr~vjF??qom0BsN2&51c)v2y~jsVX^rlns@lN~2CE80*K6Z|DWx5ALMxkP>0RAqtq*sHUZ_n9efs -rdSv`DRK73m}e#v4)FZ!Yq%ArwA79BdqnPWxGET*)3`u$m0XGA@Co~cBG&nbMRtYcFC#OFa6>`*z<)n -1a98PI0}ryhjl6{?q!+`m|=7h$yWSVwd;S(Y`DSqpbHNn?#WmpJoVtzu8-tCLhbTK{DyIoFFp{CHNuRp_xzV*q}lal?+Ne4WbNQyM5Q6R^bHeV>=s%mCYwsVhAPOMrEi;wWA6_!Z>*r7)=%Zm{pKOGTNSlwHX7r8$9g<;2?>W!vkc -NkT~{XD9y1P^KgDJ8()E9$Um$$~9&M{{U!k8V(Z=|I>EL&&TNLq^b58(8K(NX2YU-KSl!v;?tnv4_*` -qm>!4*6X4-*weaJ{A_}%G$6Bly{6%ol8{5g8-Mrk!gnTN7$)c{dDL#z^)gg|d})j(&nFxuA2Zqv#Vkd?Y=!fpP1

i!2%O9K -QH000080LuVbTlPUWvYY_`0HFf_04M+e0B~t=FJEbHbY*gGVQepCX>)XPX<~JBX>V?GFKKRbbYX04Wn -?aJd3933irX*{z3VFmpR%y^2jo!NZo_sdET!zFCm~zoWQZ)OBYDBlkMG#3?Iff+*?Mo@ym_OJLl0VaL -mn8l2I^zr495GAO{U+?TGq*!b68QB|GTS}R!8D@>TF$tI2&y66a -&0jAV_Y``}#+q!Fx-X(}Rimi8HnLYB=Z;h}kh2>7`>Hiv9Q-?R@ICtc;!`%xAuFX?XHtUY#$LAE-Shr -X+E!Kd6C-01xg-GsEea4#dp1X`)l7UlQ34w}-f!E-%%rhNEc@}=YRI9`hvtD+iFW -01mrl!BXgwTC7Ix54cR+p{hti!oi|b8gq^M+r;)1yD-^1QY-O00;of09jjIsG<(11^@u#82|tz -0001RX>c!JX>N37a&BR4FJx(RbaH88b#!TOZgVelWNCABE^v9(SX*z~HWYsMuOL(mumM+FAKDEDGHmO -DF4$HSYk^@X3W1i69X1lBl2lS7%YWbDRiq@lZMqIvUs9Xr#_xRRLhnS|kfPZ2suxfcgf(3&6rr5AiYn -G}xmdXG>UO_pe81Rfm9njCD2+f)x24y+P$1~G1}mfv9V+;$)byWbr_9{Q=FV6Wj6D@Cmt2w2LNhR2}x3O%g1h4($7EfrYevTyOEWrB`I-?%0!KI>gZ2&kIq7`#$4 -JcYsMMJRY`a`bppq9Dpx1DH9kintoIR{bPFm+i-1WkWv@SN@e@!zG`Z59Oo-3LE+FA$4`&L#+Q=F^(* -LH%aFTNVYU4U%ECSS-q#Vw5l2UX;n@mUG~)Y-CxM{Q;c-i5?}PL>g@pCn{eU$yb*^P_U!IKux5Aj_g` -N)Fo;Y4p2!FaCQfzqvaI>BZ+Fs7L1`iAYxQOmHP;+gC7v*4um?GPZjKl#Yl&Kg&4DHDWTr2$iwf^xgd -X&=Pd_&4QJ@0!fUYI0bMgm1ib5+01YPTIeXVbeIVNdVR+M;@vB_B_Sy&?`X<^#j*7V!2!{f&l*H6bR>Ck{(?Mt@=cEAjbN2#mS}(~OxVaEP*F2%Lt7|mj}?Doi(Xf@G}uht;L$ -Y*U64$Ttc3{NRVt?RFj^Cioukw+hd79aTdZTpH>(D>&l7Wota?)IqfQJ6yfR{KYLzL6gMAv(8sF#RGALWVs~Ij+ -e7`0QlLr6DVMt|e!;aaoOf1a@pNOQS(@+fFHxpx)UV<_Sg%^~|%w*W*HikG`$L1!k+}%#QaYB~G?!#) -@$bAnB(-*j`{5T3uF&*~Dv<HHnYe_lhd0#3Hyk*!DEz`czRJDJb0+A3yr`>&K72^UV^{wt -uX&j+K?&@ADffIG#=0ME1MGWM-ZrlhZxQtlnSAJX_C~E^4n&beRvvV&4JAA5LemB$XYMdd$;wn5<1nW -m6;g7V^u+PFI_j4~vAt-BA*^S|7PeW-aG$TgI0S?UvX@a7B@7AH)ek`L*NOiS{VAtTr^APsov5DwVR{1cafW7=)GsxW-MppzK;S*T^=G -!h=oRCj()R!v58wy(1=I9$csH0SF9CZoJ~k0-Y)8`mDlz}^e2{wew)_S4%Mo!i4-oxADZ`%N5RM!Dk- -6m2PI;i$?tw3&VJf7>mXACx@AviSr95#mBFB(bPt=nodFn0lai)9=@LHcW@CTQlHO4=6FTU5>+a*Hp-QZVzQ?_Hinhy+&ZUwwUB$JmlsbNeZcH;U+NekmFUayI@V6^~* -(!!)|RWYUp$*==Yhce>v*Uban#uXe062Z8=+15c+*lg2V=uZfk@JOZDcXdF1H+{0FX+nz(JA -$kXP+gV-)jx2G?Tt0hw!;LIJ|ZL`=Z?N?xN(LSUT^I^gmEb0|XQR000O8%K%whdaIqnIsyOyR0RM4Bm -e*aaA|NaUukZ1WpZv|Y%gSKb98cPVs&(BZ*FrhcW7m0Y%XwlbyLBP(=ZUd=PO1%rBcZUR06>+1y-OHf -|XXBEGM3{Rvp{ePFF#EJY##4CZ)|O9?yGzGjFDKrUS;>KJ*S51KC;Q0&uNO;6WPgi=vG=VX#u~$NQ&W --*@c(%kKW`!{g`I3@dHLPN;=(mF3Nm?+H-#K`KAO8m^6&4Oj9HGR-@jL!}PrG*^|7J~&zT>G`VW9!LC -#nr?U#KF%_(F?1NX;DL|0vH<$-h-r@8r06H9&<{Rf_k2K9Ik?BbUgD971z=qivRzicO@J#R>Ru|rP@h -H>cQLTjC+UptP=^QY16K-NQ522h-a}F_msnn_c%7eItyVwKI`l;4fkV;@bG=oO5}yvzgD~iUOa+Z7$l -n17YPm}$1NRy3qeW07(Ku{^8lczFOBB`kX6Afvow4Y`kWK!6dXWT%I5U*OOg%j8O1w!)xy{Goc -ut_a+sD`)<>2g7B+^oXiJvA{JUCl*K|718d5MpT+C0EMOVz5ZVEDwhbtbSsM&X7w1OuH~l^e!q8=0z* -2`ZjCaGNO9U?sA7T04H?hYrL5Oa}{qd_o2L{Z++?I^_?kGX!HN-mkWvO^?3|(ow%Tx=AxXQjrh%Unzo -G=@O7=Ww>Rz*C%sXrv0JE(QRnOCfX$KRB!Sk?@6$jo$J$c+i+@l{0|XQR000O8%K%whaD#$%9{>OV9R -L6T9{>OVaA|NaUukZ1WpZv|Y%ghUWMz0SUtei%X>?y-E^v8MQd08FOG&Lz$jmEAElNx-$;{7FNX}15R -Z>#o0sv4;0|XQR000O8%K%wh8&`BAc@qEtIX?gZAOHXWaA|NaUukZ1WpZv|Y%ghUWMz0SV{dG1Wn*-2 -axQRr#N=MgXd!@60%rRq)L#M)hzq%dtLw#ASv5z_ieg -dEP=p-2haNp7zDxWVk5;iEh@3=nq94=sMn(16iRICyep-EZ*48)9X)$rR5vf{vXog{H*v<_5>8g~G_|UcZF*`KD!Osv_g~oqHiB(sWxi$~88EPen`_iaqd@bAeG!`aF1@f9CAe%m6s#2x1E~BP%45uj}^9vv;6St*q -Mcwb8ga10%wMG!$);gi4m{2y7xvO;84jypgdNNiJ2^Q@p1&epI)kGg)XZ93RBIUj#eG?nDp+_oqM6g# -i=YpGHPxrO+3pyCT_;V1HQz>kXCv@CM->;#|r5;vfn!765)6)IjBxTgss5qwwORQ0|J=Ccuuq#V%S -Na{=ruRqz9U(2b`6VjlSckL;9!65PjbpK<4GytvL0`taRWZ`&}Bhq;eUKb?d?dvaJ8No-H)TS&VyYY -a4sz(bbp2~_QfVUEP3n3fYXtH%aHfSgb&~FfNF&1%Sg~6$obru0_+f5{3$L0`;b(3Ciaoy63_+t2>Kv -Y_!w3S_LnsHmOryqk&+wM>w@UIF!IdWoUgSe~p1StypXpWJJA$XvC2}%OY&5H&?7PY)+Iyvu!%<=hTi -!`_+mTGR~22?j;hW!$`F+8=h^JNBZ9GG8e^F*At!hxDAzIuJZ;-lS%D?VB_E-+639;pIDLrNSPDS;QD -{R^J*A{Os@c`3p3DsTyES7b%qsq!Go;zlAKI+VdP^b)-;8j@$jgKXLs0~Q!Pqf73O6QLfYw>5}S4tA} -q(;Q>w)W9m!<{AVRBBz*zd_u-O*~}<`5i1Ga5u&7RQOVp&_YnavS|id1!PG&KMf&qte3`W!$jd>1uLQ -Z+YGSi}(byqk!ikI_?9>sZScCe3&?tuYhxlQiwM~KQWEJ*OT$9F42!b`5BnCZ=+(XtF{Ok)g&`&(OAD -@O88ItBvD2oA~V}y?xL@U!v{a5P=2a|CwSKW0Od?ni!nY)7!R6JEdG#Tg8=F@769?{~0+u9%6XD&bbJ -_i+O*9HrLrXv2aE-K%_;pcKRLQ8*{yZR+d(K-HA{Iq-?zc>cL`dyy)d}`vAH!}ApJQ3f60@T&BR_xP? -oXrZ7tsI1x4F!bBSxV6|=Dg?;2J255;Zw*DWR`XyRLG?oV*P7bL4u8XAn$gV=K)LGP1+6sw0lrJ0l6) -xwD{fcFmnOZ4Ysr_!F0iYR4o7+VFm&UxE`+G$iuT+ItA?QE!SB9`0-Z!=f!gYO5QfjPA5*vuEZc$AEkrhy64cpbjGjs~K+opFXAo9!SbO4Z4G{ -RMVLhvH0-2nLvA_EYrMdTnp*OeH^<&bD;Y3l%>eElhC`Tzuwr`Q%hLu}Ec>vfT#@g_H#r5H5qF)yNXk -Ya991wI#~w%GO<;K6_xdzUtKRUQV*>nf9I -cSXUP=5J`7Ql&JS;i5E`{qt*=8B=u2P`t*PyEA5Kc6phbWZr&xMN&R+bi9aL!&OIR3zY!@f$ey$swOIJkl} -8qlS`MeotT+(0IPnZXuGoItFaZo@zi2R(4iYI$(FywWzIem*~1WFghZya-EU8lj{06%?F>cvt1}PEVb -vreIAOj5W@U$;c;I=)_Z?Kk&;UB#BF!J99RbKiMrQc3X&g^&>c2+HFO#jo*$-e^g33k+mA#6b@1*MeB-(FB@6^9)(BMu4Bf^!W~gLC4}3FNRlH+V2|(Pr+7Z2A_nWHTU^WouwrmIwsGG8-zHW?M-ZpOO8ZwJ4y|#x%nVU5P3@P0h -1gTPF5;)^KeOp1w@8>hoO&xb<&#h_BWh4=-Tzkp92S4yp}Tes03jSERdmDXT+7vyge=rtsodq%s4@_B -e0p7MKLRj5cZsTA`dv;e?(&~ZXwTPLL=xUl?cp6OY_BZ8$$Lqk- -8+JnOOF|W6F!s1~HhpwDxqL^Pf!=+Rg4+R#(aQB{%+SDD;b|rPP8;i38gzaR}aJ|wq9)d&hC+&@49FT ->aR78R$?@}@blQw7pzS5-fqh6M$EDI+NHqzeDnxYbhO(3r*A8xDzOWaTgevTa7WLiL>j~Dd}l)S$yF! -$|w|A2#z8!2}j18^x76-2!K_VpXG>M6M*O5h;lX5C^+XzTtA(sD?zSA9}!sX7tCdVSZ2yMU0KwF$=1j -a-}ZxF5k9fLzD-pBYtmnFaGK&+J5FT-e#?Mb_FDvR-zDiNobpf36J>2SN_=J#Biej|bm$nH`B@0JeC? -VhnMD2AGhi6jud^e9%YE)=*AbsSVJFN|i7ZQx^Ok5(*;+7yTpt!I;rv -ysi{$!_vEukGTk0atzq(sAJl5B6i3()>c~YixZ%lh5}}P72jg>Vbt5Q49n?7+Nwj9Y(`LcqI_LSQ^9y -oHk&f9?h0PX5+;;o;Ayz9E_EOFzq!k?(6PM3DJ^3RHip#E -DQh1>K#1a0bs1Y0T&S4T>8%NO$xBhduJ>LG#BU=g2uafyh0@iD9SOORZg?p~FMHL}MO}pB$+P^fL`IO -Y^0Fgi(bB)0cmvNpyQd_q#Rvtli@CiSJONI>m|s4>GDX9ewqWSqc_9~( -*CaNk4vfXXZiiw0UarW$Ys=qXut6S9UAnO`d_U}jbS$^~w5fnDgWCWvxde!NVqFxp?7o+6f-io3etW(c=?>!#45hG|qMrbnIdzlq*7eVO -V_qg}%-*=5u!Mo4j?V6V`>zKCEx$ERoyL;Ny#(w8S9&2y~P7bU8qg3w|^#uKQj16Qp00}O0v5yi330} -;U1fw3;IxUM9%*j?|2U{DCL+{_u9BQeS>)`$0>AcG~ynMju@*Vqw*tmYBcvD3 -O##$CcgN~3kMK?RC_4<*xTOha(U)8S@hMd2{=IQw8=R+ly{(eblHzmn1@r*F+A+0Ji~qvWLYRGS-r(J -7AFIJeY@rwD#YlqYlbEShT4vl0R_E;IyFqcNy-#3_cn~29wPMD)V!j`DTkf7x@`^%&zKt9J#qhFqD_m -dxLs}UOgz#WoNISwYw3;^nCw+>jY9Vk0w35)cT2cdmU0{2?Me)7#*}Sv&e8Y2(el89-+CyN&8N!tNU$ -02`YCkb@$gXC1EN+>uMO@R1%#20k_OajtYzcz*>r9tMmP(HD>8V&24b@cOx}>ma -Jbq6@hu;g3`MG}Op}%IY9UVBwoXfEb`%=JVe<#{xJnY~JORRm|KiBN;y?x>5{W?A`%pE-Q0t1MaTb_IY6w^sJY$V&CdQ`CH^YoXura8`kV2l --V=RkKvvY>Vq&{UIl$!<7*H=sKUp1X-w2!FYtck-i&dwRE(2jL`SGMhI(*>#EjQK2A=m`h?4Sd^*qW? ->Ah)1bfz*i$2n2pBCA8@*utDVC6;=FSess;xj|h82nU|CjA08U^R2hQOL8I6vIEH@8l-009ZudArZJU -4of9v;S<-af=FK+J9;&MRGI2Rm2>35g{(ml$fL|vUr>7@KG-{dy^*5J6kG*d$u+VLFDdJSvli!h&jV- -S(o)YE1a#d!2TJB}Hy)~r4}IBxjgA~|*?;w?7?v%E+?`7J$n0lEYjn%OYLSoLAi|9X5BsAo-5>X)nf~ -V5u0U|GgCO8wjY?07zLHF}G<*kU6rJKSPxDAdi0i4Z*_+eftzfjdHtrzA -@Ivz6uQ$c1zji8a3o5Rri7yUbl4pd)8-8ZhrmI-glj|Qg>y_^~xFQqOV-PBJ( -fIYs%g-Vpn@ER4^!t0_Mua8Pbm{mqAGJ})c4Vf@)TR65K3j>_6-a{Cb*whMAfky$Y>*VX(0h3 -Em`cmaxt`Q3-v0>~vHtrgCQeWMpYYJJb`z{|oxjP3)awc>^!XdG_vX#SSFiu^-B;tjKXGt;t9u>%UVP -+NPCo_Xr+A5xJh*vfInFj<7{jrAL}SIDkKkch(CzVX8!0wjZZqtyyL$9VKH>|8G-3Y9dW2BTF&<$35E -DlX@Q}B%QT%4YT6}c*_i5oT!SIZ>I3835|EhFw#JD)D)1G3KiX96j__oc6*7Ku*1uYme!gb{T1yD-^1 -QY-O00;of09ji+J1c!JX>N37a&BR4FKKRMWq2=eVPk7yXJubzX>Md?axQRr -dv9Y;=F>o~4h+1!=a1w-Ue!kPjY0F9j^(m -QV9?Xk)6>(hX-p=QAM0ILR;o=JwJAEaO1e6Ear5HoEGes%dZ*gSWO8)0ZtBY<&)2(d*Qh*C%FAutbV* -TFbysv{UA0F?a!pmcOZ|PltCn3|Z(4n!>{R1j*wp9eWp%E9Hj2Ai)|-u565y;jTk8Iv!sEr+MjEY(u2 -^o0wgv2^fw{yjx|=Nk*EiFn1U{$#_2MLd{N2gpzde2a?Tjw{u*IH=4gdP7f*-S^sf>DCZnOMKRjazmw -#D*oaqfU@w`y6gOVzX^jkB&eHyEB4=VMJ_GB&2|^SZe#Hs!Bs^}M)Lvt(IRbyY6uQeMHOF-TW{6b!_W -uDXKfR)6QKvn2nzXqEiB1(Traar>v(M}9NxiG&jtt`G3#B$s%8f!A=t*mLTkJ&fU7WuNPQ=rha97=E#GaMa+sFFiOvbi>i){ -?ZWxZ0HHiOw%mB4Vyl9+};>*2A!IRapf)7$0Cy6T#GqoF*ZtB?8O$kT|Pi~mT|Ru#?iB5#-V)*$tjWQr)Ir4nNV8Wy=e-e495+w^n9r`1xIh5S-^<{`BI-4?musd}9gn5_W~PFLCF -zTpj70t9PBMT31e-0*$}j70v4CD9>RoVZ&J@uO{Bgn9Pz1N)ofsepKyl)8XY8G&1!B>DEkM9~~WSXkr -0~ndCGl0h;F^=D$XIfn0|BB>P`=yX{83N~f=934ev5&k_(yuU{XT{$2v6fd1dq4Ls*Gv;Vmwc4jsiYz -2UIr-=6mC+r;b+Eo|ABTM&J)J15s=caao?@%@(J$#QIu`N2s7f_$g4Bf2{W -fm>V#BYKp4c}$~9>NO4rH53p+Ei46OIz&Z;MZ>; -b(=Cl80X`q0PIbR?2$Xp=?wXsq0amuTL&!yZQKqK3W{fOoG4oougk88lY_lnS<1O3KYBJezU+OCsQ(j -ZqN}cV_eS{`Yk!&Ox_~|5|U6Ov%OcU7tgsDGiC$n1)X=B!bgFSGvL&w<_)3CyL>+PBbW3I;(O``)ZdMinZJ;Fenf&LbG60@tgYz8z-<&X -!p~wgdj0;Xb=mCKeQ``k$agMcUjH)mtD)|Ar`xJFDvsH_}V5-(U+GDmKbNXMmQ=i%6S*QbSE$QkHTFZrfM`CC@Y&29z8OO=35QD5gmXV`rU&>TdLgVet)-m0VpLpMoaLjR -0!Feg!!W!vW2rHQS;90t;FdXp9!{D0QJ$S#o*-nn%B;ZmdXie$O5!*8jNxcAuD7!qj3lUWNwdHjoIw?hSRCVkvV|B{S -z$!KEnDN>;;7==+HCy6G^Zld_VLg&w4Oi3_pRdZQ(;^{k6cDG_tT(bGx!{%^p8RV={xI9dRDnpbY`P<@z7@D? -O@(oMp)`J9{k;0Kh*EJ|;acq$deQ`j|<82qAPJ;q7%e}2*;Z0J|(iaV=%X=4s)`M8YZv;5$u4~FQng7Agr|Hc7RX#P)Ms5PyU61?)Cuc2rrCxN|z{l*6CoWXmk(bJ -oz%Y-FP`ZO9)m18h4`FsHb6g5s5Y%GW<7*cE%LU2gDN ->*g)*(`aw_C{8=d}Ee;%J_Z?G9BdI85y6xl++J_z|qnS{iW6UUy$>|u4W+jQDx~+V)cMAzJxl=U5GRHkrvdA4t}Rma}#d#ZOh*1`xe -41nc(9EZ5)#xXzZWutEm5OxXY-~IEF{`f%U{SX3!%Ib1X6prVFt95V7)lqnNPjGtoowg(7kx(RmR%Y- -7U{PFzo(!}f?^OW2g`ritwIz9%O?Ozt)}++BY>m&LMf-#z$zf}UDqQ>3h?`2Ug-8A90)$7-HJ&Dh%Ev$f=zeYOQ>3D{6@NAZj%U;NwWpW9hQk;)eFo-N`Jd -!(MfJut{y?+&-x8A5-y5NgPijnTBGfPS^F3?pquTj_J3(M0;K@IVAfY=B9K+oJRXMQB4c? -KE<@hl1JkC4<#g?$`x-U^w9Hh`TebC!#ou+??OI+LHuF6<{T}In -a`Y}B;<(md3B&vj2vys`>F;>7)fS^>=u7uA!CQ8dMwBgPIxa~)FU -jK*En1~NZf4mx9TRnBFC&_TT`Sr%d*#_dJ?rwp>5i7{DijBE4DY%Nj8}xUM-kytMVMKOe5mDKi0`i_Q -Pn`tRM_;b{J&y&~ZYJCf4{;b!pJ?bULiW0f`ndx*e*-O^3wRdx7|&7o&O~ptnzC4TV)A!pn=EO(+M7P -4rJ%6P@HB7_Xtkh_a)>?j3)HZB5w+pkH*VfgMQv&p9Wo@B~^9;T-yEyI0XfCvo)igLBDo5wASoio_U* -nMYvyv7{%2Wm3aV?|3b*FHobA3>#36FxDf(n9;^AbSRQaT@xJa5EU&8L_A68mFW_U;&kLlu?`vlZvsc -K&tq>hR7U8MhfUdC5G+-)C(C~b-5FKQUp3FKNTp*eF=CYbfKOhRtBJHQ9njB2y;{c(4a}#Xu*ieR -Z9i~4;i0dskUrn+NXZXhioshR^n^k4<39=^DmTCt;uKs3E;$0wPN*2GY{;7$*OFZMMKt}?X9XZ^xmi+x0&#_l(?#LPWq`bH@~X~Zqf#w8&R((NbuW?ZO@a; -@>#=4S$`_Dyy6vvnf~XK{uR{4NtqdT#kS{^nBJ|Jc8iH^s#;l!``aLTj>~GN0ARpM?-DPXU%D!t>d7CKylGDd5kg$=sF;UOrR9*i{P?iX@a>u(%~$jJfDQ;KTK~L6iRV{#64vUi&<~x -vu}72JBTqNs3-9Uf;p6Q120!d^Ac!Sf3?RFIA6vnI8Oewuj-ss%VOK^Hnbj&P5GDO}Z?ijw*6>Q#^%uuaN;G4F8kZyG91z;Q8OeT{guf9Pq*{)MTPwDhca;ZR%U- -@aQNF`TgegtOyFFTwu92u%R#Z^&m@EnUDe8<3`2M7~Tme|gNhgll-mZ!uz5Bs?CQ7GUcgI -#CA@f%nK@j7t=u_@w98F!wBGuB3Bqit;b9CO#@fd>VQ+hYissAb*HcN>~SA?#zK9@u -UDcCt6PSrs2KP0XZh1^HoMbJIXlwwWl~X9|Y9e?Z8E#AuoX3Czdrn%$E8?{uyNqeQ!k^g@2Tm(Ao%!= -uXPZ3$!y9g#_g5Z*ACzwuCDgYk_$CO6=}dNW0OFa$@^#T=e(4@o`YE{6U#>3SC2C{~aS6`a3{$(}Gl$ -R&R2`9;l0e?Y$9}lc4>pG8BCLL2D|RKoBCa6zJ6`xdXYnB|%(u`kcErTO`{g%`ii$<_ -s{f#8_w|(X- -EJV^s3Yh5;AN5q$=n-&#N!lV`k*Mn!ps9Fexg1g>CP%5AEyVrI&>(6f)UEyJyUM_)m!7=M!mPj=uHJ# -1XW)8H{uD}o&d2c15*a3tNT0pz38_F1lv5MMGp3SXn4F&l!sjENg#%vEbpb#?l&Y@{CGH_ONGpmS577 -UphhY-^zu5m{UGvW;rU@4Ou5V0 -Y%jL9)ou;`H6zXwA9OLjR!8w_U>##fTAi^Ss+;ar8KtCLZioHZ=d1a?1Z0zNA-Ah#8j&f6);qRGBzPM -`I(}VAAuj5!H^H09a`vHsaEKY=2V7mx1~}Ee0*@GJ;^Ix82+Gyit<1FnreEKHb{ -$N)8x00xgOC*ToH@4=CXXdC;^HVewr82=oZBdbqqOs&f}>c4?6Ucnve!)mwm}o$ns1D2=X4wA}mh<{s -M~7>1X};hChqu4V`om@HW}09j%R9c8|BP3lNw%ZNFA)8RvM3a3j-TTYjnG*j`L`EFnv{dXQl+c2(1WU -LSNqoe5+LCiqpM?q$Nmg0qHvhA_v6rUD#BfG~r<|1zaxcWHO`5H{RU)U>u7 -v6EQ1ghhO9LdZLR5(Q3jIUC&;#x^W@~Xw0<1MD+no<`V^4mToS0bQCt^1*$9_r!iE7XFZbW= -=Em12HIxs^O0v|AQ)?AZNF?C5ZuHyfFEQ)MbL2g^gFE{_vq3WnE+>KfX`J~aGJEvl+8&w%`Y)*6VavJ -@Z6V~q#)|d`PLncul--cGJq|~Nm>v@9NH9Tz`HuD(czEhH43M)BOsb9*pE)<==h}Kramv1to$I&p1A= -~nb!VdZaHz2kg|u~+Z8H_T!6<1VWR6T+9_rlIM7cFDW}>jmI}|ZL}ewp9Q|zP0Xs#{4JhdeM+WV)c1F -1H3aR)`O%ca^!gJvm2qwskpS>|;9TKLQ^W84@NY!Zy&@Q8y^j_nG)AOSV{$Mu2v&Zq|u2GX|I3LVXb2 -es9MW5f7$tC9t{*EExlcsU~q58vg3u{Ql(Opm8>p?QfKIrpSh2uVug9PRjucTmtw%ZM4Y)@bUNyN!&% -=*oG*`pCNzdyJb(>?H78NM~?*yf69N8tuatjDz(=u^xP&LXjxV!DH?+u@1Es!OBU4X(=OB+{e6&-rvd -bn_tj)Pum=!4y%0dq{EFAVS^EpUx03Uq!TC`qz83^l#gx8!bcugqp<(A9d-Go*2X&Kbn0%7BAat;3HC -wRkq2W16rcFsy;)CP|Qvw!}4BnA$jD-;zKR%e--B*t6q<2pDFXyi4~f=>b3=hMk##fbd -6~oWW3eNSDXj#pR|szKJ~k#!OC}z(e*K5op-m6peZIC{I9Q+#HXr_N=bRc!mtbz~3H&e~1m~$whC((?4o?QI)^!l;pA16hMRT1Ve+Ip+>4pe{Asz@ -q^PG3DMLbesDhcV7*51@CX7W^PAF%^Pp&=Y4)3|eE4M%=YB!Nq2Q>Q94=HikEYlRJ^YdDig4K -GXU)`LG=Y=&dm6kVn?$?8YHYe<)jq@jDz27sgm~-Sh4MgXhWhs3dqK;$r<;E$)i$V(q(!%TV=zGVq9o -GOa3VPpS{AxIAEx-Bz4}A3yyvIh2*aHL!7;7`#~y(I#Fb%47m+F_#*-A*sPT6u(x;sGi1-{mr|bDWG3kIPEnFYQH$KfL#4qzFWh58>xuXOd+A9ZJ#lA9K^ofD0(;fXYYApHyLFyR2w+d%%K^D -Q!Jrnw?|#WpyUG^+&)G0h7v8VfKv^ISj4yP}2fJo-cqbm|?DIoCz=d}=7C46dL%-4Z+_jWtur0Saff{7RX4j%8{-Oqx$M&aR0uc-d&D;xeI -djeQ{he$0ph=#C&F#z8n^Ixj38$2L`l$gwxO$cVvA)cgaB-YKpxg*1SE|7>B_Q1OiBYmU!AeY0+^Mv_ -#K|Fltm@?D%V!Q8dcItex;Ra!n5xu9p^`xy=q0q0$BzVx?1C9)v0xO;0ka6CVOtQfOCym{p1lz? -pb-7i@4SJ-&qpkKvrF?P5C)?4<>}2dm=({Pgn9Y9rVvBuPVjJs`u*vrhcGyIUz{93L!IP&Hk=2C(k{2 -i63sTf&I|&C(oeYd7x&hj8*>dc?B^7LR@vYSHB%=dwytA0ZJE#}bIhx(zjCoqZoyO*8EqY@b0Dj0=!r -VwZ1_T}+)NkLG+pVcdw@>C19=X&Bemj`53A$U>^O@BRD612*yt60 -c|o~HcC5JWZ&%N5fd3A1)yma4wk^wZ?H*Sc-H@J>sdxT1niHN*p0MI{jU1Dqle+0GRFNL7XHtee%CJK -@0fBK(VfOs%)mwTQOlp0`R9;`RaLmi1Xy!s^F?~Two8D1j>U~&d)vq{`FbsG8=XQuqa&|aW>*f8Vsg$ -bF&CVr%!w*smcu5dHsvKOUvdqucI4A{oh>d70IJ1aa{NNay=VH!H^WsGyMSmM#@XY{rxYZugb?9!q`R -iHl#0~q!K+!ZT2bIKS~P7>8A^97`oMJ@R{v<2{s;x)kIPEYRQim1a(Dv7kr9N^vaec_%^mL<%QPz~eA -x_xBn}Jj1%Fp`NX;!GhXEK&B)cm4P#Zd`B?~KnXJCnLMfvADS -csGhbdzlA%E<`Fk6*y$CDKhhEb(}=)NbLCG7H~G-kd$L-UpkitY$(BB1DQrPw35v%_)HnS-!3f1 -(YNq8_+Y}KfX3c`1niW3*wzIwv9D>k#+QyA^kQtK_3Qf3NN^!Jc$ -VlnGjHn6mKQ}?WrTbwdu?#AiO7=$V1mNXROwO3jYo;XIT&b6Toh?hn(P#3|BA@r04Xa7!y&u*99V3bm -g8MXVIllJs6hDi^ZLew_81%4Q&5&psOO@*Qe5*V%4lfg@@`zVClyj}^#qow)Mv}SIr;jhZ>J&={cnGf2u*m3Srb;Ts$x@tIlKk}JvmS`S68MB!Ngfip-$xeawS@E<*%@!?2lbN?M&0>`BqJOY+qh*Nko -{FUA_dESmt40m0iE<=x5lO|y{K&|6yx^Db3ae$cUghQVrkzeb(eH{*y%j&(@DOY$sA0rf*hvt4Qt3j% -U65)aJ1(B?jN(2;?bChtn}3GQCS-D@>ortJ1;N;;Z9WQy_X|0plsmGhkI_p;uGJV?lyZyZaB4uLmvv< -GN;FXw9RJ8ZhPJoE5*6&c6}f0-92q?c>T}|&(Mmr=a`{WJJ6GNTXccp#hGK7I>?Yj5PW7f%fUdfr^n; -i!J6Vah4u+Nw$~=#lr6tQ5t=`*JFb-uEQ;$^e36ehE%>Ra_}s6>;Vxf^R>D{^@9?Bf5+L)|YTTe8^(J -UanWT6UP)cCa3x$pXK`fm}(F00I$thd0Qr&`zAUTIeJ?_bM0N7c~LmP`8($H!{6z;m>dWU_YXlAjKRQ -NG9xXnv0k{mZOU4q2hpj`1vI0J{YSWrAayBYRl_PP+FyHEGuKSL2Oi9754ffhSievaEY?mg0pDPdfS^ -yNlvQES3i%c4K-$`)(k+9VPp(Qn~EGlUG00MXx|$r-urc2{4y>yrys^&E#b}8-QFpIAAFr5zUaw6-Rm34eU25EuLLOfml74rZV5I -RrYJxu~=bahu0_|R*km1wCI;-9pU^UITz7)`GeG#saqSH|Gp(7V`MWN`r!R9GGK4^uyMoIA -EHr1ahH5P2ekBtzHZnBOjrS#gHt@?IXG^@SN=zT0Q_>v$>a0;Lt!b -%PZTMs(8;lt63EP2>Rv-dtLQ3!~$qk|d{>SGh)qZ4`4+;B`tef7J3*4CUE3|mK*475lt9?J|CQlD$qjXzX&{0zE;gF9+Jtt# -|HkuJTLtd|xlX{LSxP3d7+NLJYi`+Xwf}sskmX$4(PmUKpsEd9SDpD_}Y^k<(B5!k+_$b^oP{{=FOVw -Kh9wlGNBM(3ON^pE=UW5wvjkKKLtGm=CCYl>6V4j}7@716W2Xwk0F+MS`YnWN(w>4!#-4xOln_{b4@j -g*0442V4Yl6lFfpoPX7|S`9&tg7BtnvME=q1Oy7>Uc!7p2}6=zW??ZK;P3ADYpD<#2yp*Li+cG{+CJ& -+nTSy-ZhZcMQ#JlRZoj&OdY6zw3!V;$`^4eLE~JJH2Qw!E~hJG0Ycyt{hcHluBqNfCZN&I*u*S>g96R -=)-kuYNOh}=-OjC)vZD^q9Qp~cE;=H&4s{YM@AR+p?E+tRcRdm#^t%F9EB@WMg?`m%J?|1RCXg4ef*FL@wv>-UmxMB@2Y{*K{5z7{Y@Oz$P -!yO7V)4$}CFt)v;C1Fd{2GSmALAa%wq-N)I-8zT^a*w73$!$R6yI43UMc-+96~6Jt4!2h6#|q%8m4T3 -$rXEHtVQy14ZkOg@uRNh3w1!s0X5%V4<7Lu1Zi%b|CshT!bs8QM|*2T)4`1QY-O00;of09jk#z}3a-2 -LJ%SApig#0001RX>c!JX>N37a&BR4FKKRMWq2=hZ*_8GWpgfYdF>h7Zre8W-Cset7-|=`wr6iHR&2`_ -ARC4@-5vr%Q6$>tLW=@P=d#Pc@0>$P6eZeCnzqA+1w~<#=YBu*TB{vP)Aio&HBVC}c9qhWWo4;sW`!! -vXjGp`wb_Vr(|p~D9k=%tH_d@Eqcu{?m6SXu+=g7i*V$g$LgaP;lVn$U19xh<&aNbPKtLV%HFw`+Qqr -BftVEThH@qyAPO2=w&QP{wZ-s2RPDTv=PRpB2isDTnj6D&Om)VZbx;MXWIhWy0o|UQ;Io$rvQ-54#+V -IpMrF0!2xSLy~)VY~l3vvi&h5%;EiW5qFSyT25J!iFZ&7t73NsYNVC?Fi0vN^0N|xz-D{eSQXe=K6aQ@>N!=Qe=c!-svrhUoC(}7UV_p -Lmp;<;FprD}N$D};ucA@3h9#|VW;A$GTR=Mr;5B0(rGn6a7!Lbp3nG`Y%Gwq&#C#-#YU#hW1>)UoABC -Jmone6eE7B^z6QXFo?S<2HcO6NtfYNoJ+*&(&MVLdOT>e*;cyG=f{OdjVA^--4Q&8>p2Am23@}LQ)Ew -reww$C9zi^3^{~ID5W9BDYN@GLC7q~&lf-o6k!*uKD1W9fru`(UD*H>EaZqX^T>$ZDf1J&nD&SnU|aG -c^=dB}nX#iN!gv55HiNwCTTqobCo@*6+DnS=D+NaA>D!VZ(1>;{vy`g_&X|Qz=rV0Tuc3rMtN~w9Lgz -~ianKWQ{i&B0|A2JukDAeh&=Ud;AVoi90k}u(pf{7Os-Pzl$^au*B_#B|ws!OPjA8F%yS=e&n?c@ZjM -^fjZpJ`QYIw{yJCCWbc->+G!nnY0p;2s+ixE)fD0RV_%ynlTpn~Q@jT&z>4V3IP?5nlm@FS*QK!)Gty -t2{GKI9h_;;0~D>pv#QP?%r -l|O^BMb|Tc}C(R3~ss#+C`O5VIoYg{a|8H(6-15M^i1@Scg)E#L8Qw2xq2;5T1AVYT58AQY2%f-uWrM -?J_i5Xk3fyvRIY+^_Q=51%rl<8XoVLSW@ayre(C1*+Z)E(=Dx+~Gi@LPI^`J!bWH2p4Np!@ym|q;941 -Rq{+hWMk7jQX!*-Q0X~PfX-t__Nig~(*k>%~{GuU{U*;kc>ay!FJpaJO$$0Jw -E7C;-|5{Kck!lN*&h#ZzRRpBa1L;wP6@D0Qjfz+(lLK~a7V*Or47`8UQr5&rd=UW`r%6ShuzH5B2dWn -v4LVvWXNRJ^);`H?V^(3E?#r0qj&hnmgG92q2`}XKzx$G&vM=}p~jmOK^HCB_hulY_{{>Z$2XTAn``k -ze}(%`0fr&_S!+3J(#R4558LEVFYO$Nra>88y;orH(;=oEwhAzQ++4(E0PN1~TM@XuCc*jwc+l{$A5wMszwdmksQ&~JTg%%%v0m+k -EK;|nI`>H!+;HmY>_RWv-&P*s#s^eGepdl9F60Y(%Wc@?Tc=A}zioGKBkLL|YxkApxLVdOnst5#sE_K -=wcP>hV+aBo`|+-MUXENrNm7O${SvoYO&Uv07r3LH;)@v`N#c;SCLg{Zwur%AN}#pPA=~BQ -~YdyjZYt2m$1amo))sG61t4dN0xO9KQH000080LuVbTb}pZ$L< -gS0BJV>03!eZ0B~t=FJEbHbY*gGVQepKZ)0I}X>V?GFJE72ZfSI1UoLQY&07C++qM<|U4I2ew>N2%sW -{2f##d!?61VZqNqup$w3qoPN`xfV6rmBM9j)vCzV`(HL5h-{uDdO_)iww`y!Y^VUx9g{Rw9b#Raq4>i -bT3vtD+Qfo~tr0QdEGSi`+E#*qu$(Ju)oxVPe{(Ac3?bCnZnYUS^-aX>4&p7FTekAxqEs -d*M*#s_;!sQlF9C6H)&B;aVGN{yQ--290ZK2Jds6Ie>Y-=c3#xdOE`_SleFIx@bCTQ>D%bV2}m8iJbO -E&v4t$7Rs54GqE(v1e{W==f#>?8E@lz%Jx@RJSgIqXKSlE*UddZkT=S?-OBt=>*>${-nujV4{NnhqH| -X`ED9$nf$5?!%uR)WUfY$`L<2239j_pw->5p|V?tflqaS0e|3;{%2LyU>Wh!fe4HpH`oPCyvA>h*evW -m$lP2Gk+YI;XLI6y>0gOAnKaZ=d93rgo>~(FZ0?1u#$|d}G6dpbJW5{c&(vBu6=pmg}r7E -0G^aD^dx&7^014aj0Y18MF(9C|b+H2o;GO3Je07T!tN1{i8Se#WL_SJn@jugS -BT6OW8Q&;XZdjQ_RZOw?|Kl^Wg3p>(Ku79Sh$guL9fjuD?NOh-xqNqWejafMzvKsTD -tB=<223JHwewxX3DgG)qr$wQPny@j`+M}3(jt$2m_yR{LNOT-QA(IHq03u5XK`5~cZItSBXgUUU%*v1 -8B8M0+u3E=ci^tR3f&lh{(HC`483kx?=I>gpnv6ZgLc(t|*>9~Lig_^l*^6Soj37X{onYFN$FK*2x@( -^;^;NkMzjJFtZpf7FPEO@!X}y40#ngw^e8ImB@F<*iObUP@`!6mb4QDZY-+4lK8rx>_7MOk&Q?d$+wb -Nm0(7;Iqy9WGwsaBHR2x`YL5;lVi23wVbZ@G;(;*6r9215c9EG6JS7G~2!t2K_HznKvk)xxyCJF{z)s -6E?$MCm>U)NLcYz8W}M7*1)ke&8t(PtxhnAYgEx(@W$ujdPRXb)4Tz`Yz%m0lnWU{3Ye=z@SBkznw+{ -;0D8PDOW*x{e^lUl!T2 -DA+oR1OOVdi+LUvY2m<%Rma#za^QB%{OFr?i=X2{`vBRCJJir5B4%#mRZM!{cWf4(MKEATfv?WN6*VO -Ktd*(y24QWrd?q0#pz^#=nxEu5zya{5mhVX6YzXush=YZJ%0p32c-&V!$o#v@#YP+(#1wCuNVG5E!tM -Z+5)t-Hx|QMi{?Yw+~fkZ8)?7{9br9Lnor1l{jE#`FwMkUo1E8?ukl+NLm$jzzGe1DHk38%Ke(;7>#+ -MJxa(*9HK1Y-3P_1-x07Yl~kE*VhWQ+J}I}}XLCCU)pw{ba{9mNs3bld+U~}|(?>^74gbtGeFwO+t -x##7fndcfc7R8?n_>rh1h`qHD+r?%IBlB91tF>rpu2+XpX4AJG28+fJU_<;I`;o!ojDx?k;WHyy7d=U828pf_|b1wxj -?kT-TPRz{MX6nHmk#r!&ZP~v$EI;gpDA{Z#g~OYh^%f6B8yvtHod5auX=X^1ppR=KY<7B6TgEY>1-&t -EW>?K=y^E(>T97Wf!5)a`dt4 -uw{KfU4jy|B2-Lu)T3=Bg?%ZC1#m)85@Kjbyd(dpQcgGL+E%Uh^xe}M@MM?^svQ&=+kadO11zRAQR$Y -}~zdX!C~_FcZic`*@AlvCQLW6l3##S}icM^ -2ph+k4I@(^j@IgPW3QY1fq(ljW3ptQ?C?3^!1 -Ly`j9g}6Fs%*JiWFQy!7-`M*Df5P?dgkx(i;Q;Ef0(N(BFn{lujO>L_XM)q@oQUP`AkM@(=7U)E?17K -VZmr#oPrnkIOd5BRRH*a_BVBlza?M%-?X(1F$wC|Qq4*$w4<&sjOOM -zKd$9!q!M3+c3FHgYF1Hv2Z!iaB!a?~^d-}MIsBO{q;|phO9ZdV$w!eLL*MvYL-Zlt*+Z8Tua2R68)UR30^~_Dq5A*E?K6P(fgCKs-TMuhuqq_DiY7d -@q6AUFy?64jX1rm+7`Gp=DXOrQ=16OgC*Ptb6<0Xm@uSjTm6r(hbenf{vOQ?%IR2GF9cW<^uhwX?=75 -j{9(679KsDE2Hh?Ek3O)B(=AU#ocP%&n%i}v+d4Q{nVrV&AR~+zZ)|%+mpI4-1Z>e&AJ;3b$6<(v*3Z -j^eG6qK5k)P=+Uzm6YlI_9UlUK_e`hHjHg8){mZ`z5qemu)qYEJvb|kg;teFW!D2@9f$>SD4|W3( -R|(=bNFYH1`(!;|?BCpYR4IPbmvJL@w0ew_OjJ`hz_=N~u2{9Y4B4EA2zg$S-23U%b5ti1l4Gg7Qv06 -GuA|6hoQl64|YJ{Vi(Wh({=fPvGGdTmBA)`g`wXfYzS7s(rBCc_ep3x7pFFn~$as-p-d0FRcccM+}u< -+kMpEcbAy;lK|8My%?BMJCnPW$VCfgepeQgqc*sV?ce1$w){a?w|9?fTWs8|P?PC_GB3WrJo~}k$3Xl -o(lsjbMqQ&bPS>J;lg2{k8hC|22R0nHN*j6*2v!6CWtZz3yh5w0m{C0gi76hFcb&$!7IXl9Qd^|h4uO -+qlwf-#ogqE1iM^!zJZgIuQe#?Y{0kS^9=Xo`p^=J0Ug;6az#;)1t1r7rwoALsX4DbW<4A1_#4>AqT?*dG_IS}%#X1`A!Z -^+&SUs#xHfS<2_32Ys2UTL*SZ1qck6fXdDl7n{-oxn!x#0nFy7~FM&&+6*b*PvpaW?0<3TPoASJm4WV -=3D`ia(wNGdAU$Iz!EBL%yUF9tpu3caC5uQOgt}HgE6z84kNJ<+6ee1&oX&2Lm;cwVg`1`Txw2OIxH*Xbf*0m`nw~?mS77y;oD7 -Su9S1FfZF2PR((+LI@z&8B*Obc`s@aN**|WLWsH%%5=BivN#Nrhk#k7#O7zn8xqd(Z44k6p2s|0-d0x -ZueO%Fi3NS0l0$r2we%`b-#kfgoj`0B}%heu6XgJDa{lx|^%W?VOB?Y{+#L}}joGT38?7fL(J)uzQnR -^iN1E51+02S4rt1#-qI1m4iK51Ven+%)a>2wnFM&Gq%&L*rthZkc+mQm;kY?0*g#4_!!X*FLXbbMrMT -RNQ;I)BbF0PXAGRxRc=4Jy4^I3Umg;W@8mEkyu@U_d4CP)`s#1`?Y~;ZRx_y?QYcv0bf*bfD1XZ&>hb -S>jy@p1ovRQPraP-bybxhuc>^rD@PM@7(4*x^Lana)_iB2R7as-A^ -OEc8QREj$`)WxaBbc@e0^#J*D6bmR>IL#`J24>D8|+>J(L{t2B!X`|JX9PIPGK>Izlmlyprd-Z22r_}VJ6(<2;WzO!+tlu3nCNOc^S9r^Q@Nt(^WBswt?s9N^L+kLFCn}qpK}6sm(M+ -c-Ov7apU>7_|NHYX_3ORQ=ig9E0|XQR000O8%K%wh@MJ@%-3R~xR~G;PB>(^baA|NaUukZ1WpZv|Y%g -qYV_|e@Z*FrhUu0=>baixTY;!JfdDU2LZ`-;R{_bBvC>SDVj*Mc!-Qd9+F529!Sl0qehjs`8ftDzntt -@IJ701ik-@fPYMK4Z#@33z(f>O2)l_n87ECtA^15cmc -2%`t!WK!A%w`wt<4%CGuGK*{ThQdv8eWfwMt@NHq6q&&URX4V6#8A#}&A1 -S=ejUl0Lj%|HNw}NXpxawX(f5Y^z9&Rjf0Dl@jV$S(ARf%tl#~+Zjg)A%$v>cpwFY?OLl;ZAhZr?kwB -!2LWxqh{8fisHWT*(L6}48nPvWiPA9qm8ubJbMz)f?(r-*nw{K3lM2c})3^cauy-p2+gaP1)pEI&cGq -n(_-ff6TfT&4OKD6e%<^}?{o$2!Jh4__iN`{;mZfjqi5snzUa%i6BDm?Sa -KGoCYs(FU1i`i1!l-ThsZ4w_N8^?4sL&7u<@rdr427$=CrW86K_JfMqh8r<9V$Z)Qo$qCX{|`b_HqK@gO`+L3hBX7_JGYkRdBg -X}xcO{0D@8SOp*J*jxk0Or&=vo}t)aB_ifvG*VC*2kukTbKWB+FFDbLsL%WrC>?Ac9eRLxO~FL7qgR2$7Iur=$2PBIEzBviIlKM?>?!OWSwj5oef|?t()Zs0KN=-GxMF}+AMG*qNn+7c3D~^- -i=&uq0~eJSX5(+e&63i)u)7fova4GfAi*5gOA+8gy~&k3N2gzBOyL_$@O)kt|?O2MYT=tgBsWxE0^{x -=1~m6^GnX`rS@RdtC6)_n<;rZ?@0m<{k@R|>gbK2{|}K1NF+Rl1+Jz8GU?CX9upSKGMma3(#A -q{wRu}OHx<+6|kZc)SyHIO{hxuhqLQ|A2kkBKbgIv%$GysG^JVo652s -li>740Q#{sxP~px0Bu14jicx*6t80#69jvHj%72O6PZ2S?7Yu?xd|Y6_xh!4Tg -BK5_$BG&T`!Ru3Yt&dXlZ{25y(0!XAMw_S~1-qo=U9|D}l~-q~U3M)id!UAqq -Z~naF6;wEU|u!c|THbFypZ>{dzo$NtKM$O -~`#7N4!osg>%4>`Dwhet}THp=~EIsgWfZ70OA)sbh`5`1QIHon^n@0JdjP(o<4n=_sL5i=;3SS4Dp}_ -5fR*sj+%2KCiOGQ*kmU-)b@=t0=+~17x+w39)2zAt`sM|BQesRu2qSjGM~=OYU%vj$%%dh7?y*_qt~g -tN7()P9$mx -biKl@LI$3FoZs5TRCN`fNQT9{TfWw;Uh+i2Tact^uL;E2iG{k(71-hNedbg^}96T09oTM82k--<`u-$ -3~Tmo?wup6R8AF)dzX!3m*pX_qOzUH)bn>IUDXc+7ToUtjRrT|}(79=B^Bu9eet91)UAPc#eAa|>AFT -Q^W4wps$9{n+5Uc14#?1g{+O*1uT@`4j0VU&a!ByrHEmpqYs;6r6);FBwnd;f&9Pm(S2%JQJGw$gL9u -!90VBLg{1gWy7a}d|tWgbAX2nzASc3Q5WiuOYr8J-mhE}`*O@+$UUIFd?2WFy<#%6QzJqns|*Ge>iFN6&o&#?Tn7Reb^?Nah8^n|Yv5&n`rWf0?YX;A@Mk<4t0(YTKfi?zDCm --6;ljLN9%iB=7jZ}%gUsy+Q(RU1Hnxrs7TenWfUltk4kV!`jz~h{uAc$nV&qb~j145=_m~7dm-qYhBW -}(|ptzAu?(Iq)>;R5dVQTpN`SXgsZJ5GvGP_< -q=8|V;Bbjai+w_Zssm-G`s27cQqe@_;t@$?QZIBqZIi#C3ARUCgv|11g71D(x){ClTy{n;YXu)JLFO_ -ZsM?fVmu%0A54Wr$@INU1HIB6=aQd-E9B{iOR^ONJBmzu0jjZUVPCF4^2Io}QW9IELgnFEXId$^R8&% ->pl@b8%?tbFbM53KwJ`ae)h0|XQR000O8%K%whf~Afk;|Tx&Bpv_&B>(^baA|NaUukZ1WpZv|Y%gqYV -_|e@Z*FrhUvqhLV{dL|X=g5Qd9_(>Z`-;R{_bBvs2^f?in4CNJ_N{s+%DbCZq@`%yKD#wnU?67jZ6xn -6vycL-}fAnk|B$jAZ)d3DP8kgYGo!9|5>3Q9;A9`-)hs -oGTZQFrqxnDOeRZ0E@hddQq`hTyhy}zCQ1w1?sKW0cp>xKY$ev>tzIvrvb)CYlM&UWGBqy>`-xvU2D{ -d|C`@83NQ->?_UDVgU!=cW{OdMm4u4FxsU<|8WK}o35bCKFuNtMGY}%+?RB8X*>y<3H>9)#NTs?@i(z -=e>gQ(Lr|DvmOEfs!0iON`oq)D}TPL~y5i;b=xClfNLsMuL+*#xpz_Oof4DlALW*<><##op^$7^W4&x -tBQ3Yqo4uW(}F(@h7#!F&bD-6k9f*(-Cs9ZC1R79<-PprDY|S^7)8)yW@CRY*-0@0ec-K?Bhxp0leCXNxU+d$X-qT0sx5e@{qqR)K -RWN9%-LE=dodIoO@&EZ4%pFPj2Z?Vk3CG24ivqF_r$oXYBaTY@u~=I$`*UqUh~PKr1r+c?tDl*7fPb8I2V#px7J1?q+<`(MxpZlSXoR@ -8TqiHq0?ELt?oVy;I=_4JS@q|0WQ`Vw9UnU}Xg%cxf@dv2Xkr64gQx`yHXYk?78A!aFyDHGw4+_Lk?e-0lg%`1ny3NQZuxD69_d|9-ZV4)lRmRhfjGq3ZtER2`vkvULb#V{*KsY(Xftym(E -P(+Nov~%~75OyT=|jkh-_ozzxL2Kyhj2zicJG7-q+9R<=lDVJ1JbUt4D@Za@}}&R`<2@49^byJ_sVqBd>T6MH`&|3F6Xw+cMD!*{! -lU5Bwtj7o&Q@0b7Yp7-vskXDn!6!;Wa94`F!V$IfjPp$n}^}__ -|*~Rt%r)z4x`wy%WzOOS_LgPh2JpI$ZZMed3i1CJQxdyJMQBoMHN7C%i0$%ERTJAVH&MWvIun)7>di# -hr+RPVLME_1|HZ5bkp_j$Ow7ZFj{}`r{tbJ;T9=(rB1Kk=XwIt8n!`@-roRUGUeG -TIIoLBNZkOR&q)awWy4*_^0-hP;**rQ@3vdN|b#5V0(O!3N74|@0q)d={k&#j#Am@EA+hBr(IX_dKK> -I?272a)Ie7!fN#5Tmn6ga(#trTkIPt!kH|Pz&~Ii_sUL;39RYG#iLBPtm)bG3630Z4JOt?7B_EW=>~3 -d1#7^nPCNiD9qJrJm&{8gBA?q!hZfF?ntYT*ckhiXAm~Ll~Hu8>*L{ygB4lK!*4@7G8YqaMqJW)%MaY -OjvY3)w+q^$Il%!Qr)x&ztkvDN_FF8P_Zx&}$l<^~63cizd*hi(X`(ebf$Wz3>>An$9RO}YOasP**_b -bz{-3~a{jFK-Z2gWBK=rPuU3BCe$vtF2?J+XmGg@D;laro~fsM>ERq-A9*yove*67WOzE%FmS)r=!ho -!fu85f!^7}qUJ=y)~U2K~Ob`=>IqQHrmR$9hNjUhEKz#CQA#1W`>~FDhZ5iW|&>; -21ct*D~+&m^8@J+&QWXX_&DBHPv8Zd&2eLID_d~G!#-A4vx)1v*My_)T(a+-Ht!|x3JsO*(VpB(LON1 -xvR^8Whn{ZYd1ga^ZqRyase0^qCn7m#CACISB#zDkXnafUVtd=be}$NZBOameLe79$8_e$*p9UKXxXx&$wl-T@@Pou)8q@R -G^IUb);-5ds#f^h?HFWGYs#*h==SJ7qOkWY@Dc9|)tPT@`YSQ&6p=#C{Q3Ob=yTk!Xetq|aBssEz6&` -UD733^+#hci2f3<}Wo=q;|270iE6v--^tyf4)i~RBjHVuB2e<-lMQWRHn?c%VHvfn%kv{l}Y{-tG{5^ -h=+k&sTrrmd4?wbf}+n8`?vp~h6>0iGd;tnP(-&^&qGfMLKUEwSCPFI<4KB;=0G^RQ!G{W%&LRLZ4Je -Bh*XKS8a-v(a~_;T0>qUE|D#qLl4^w|~~h;F2cToLMm;r{fEe|~anukuf>TnRdf_HT%WhZ@6aWAK2ms3fSz8W~Ai#J8008z00018V003}la4%nJZggdG -ZeeUMY;R*>bZKvHb1!0Hb7d}Yd7W0edkvUs0UH4OeM%e0Rb**I0|AEjsjnh7I8qV$fdLqwae} -gu-UO0FS*67UBMorR)~eO8Du;s -ce7#Nl}54;4M)@^*%c=*bmoPq5h9iQtF+RZ1N`TMUE!1&Vont}921Y -1ajo)Q{7Xm>#^UXpFg~%PYdhiNQ~dZ8#$aK?Z1fWxiW^%9TdrCs(iM|So@k|PB>B!SJPBiRUQ4#&e*4 -U+#-CyPYDha_S2S#u45`$-*cvW_%##DJ+!{`>>|HxkkI_xDkde -Ow>_weAml`f8()};f3ML38U(I;IEzeF7f*-yYHVW^cuXczIsv@R2fbbL*4aFErbpuF!%&Aa_xP}FYATV**8U;=GFDyFsqO!b~1)a85ga4|yy}Fyh!{55EtJr@|`05fSQD&dy8cE=?q^_?8UQPSlxr6BR>! -4mHTZV6v&wt*1oPYXob8$WUGQasf9F2m108mQ<1QY-O00;of09jk1nmd{v2LJ%T761Su0001RX>c!JX ->N37a&BR4FKuOXVPs)+VJ}}_X>MtBUtcb8d9_$=Z`(Ey{;pp^cpvI?ssJxtHyBN^q;Zs`pWWN~^FmqKo6!T)CYCEc(nN>f@C5SHSr`PatoJ=O)#!xM}Q8F|l5aE_9mXHq>B5gMw$K-0w6(oE~mdxxIcT7s2=L{q}LZR -a7s$fK{$d>NZ)FORM&Q~O$o!n8$>9RoY)A{*4{qXB7J-?X0|K;l4`B{1~yE=V!dUXn5z?fp-AJ;|^Av -w>W$idOAqw;{muo3vkW3mD!od?gsg{O%_^uEJLjQ_twR-B*iU6T$%TDL2zv$ex=z%)=s^BlOiDa4W%* -o2g(rEYfJ<1&=XYg#C_SR|I}4ZQ_tU3eP}{g{pG5~@X+go1SvNi$z4xX+MsNN1LZF>tz?NC4``SF&c0 -k6e0n2zm-BW3%-@P9izbe<3Xl?}^WN^9h{=N;!d4QGf#)-Aa%bCu9sfPo5r49TY;o35>otL9=8Fxl|T -AqUy@DnIYpzvt2Sc5lVty7}=KVb#MeEr~{y@@msc=NhxGZexpUr%<};lCTC*w0Nvl4!#+n5>XPjsGb! -MNrVRx!0Eg{h+_dwtP>IudJ-R0?Mjwxk4l?=N5;?iYgNZ>$3#>`gc{)*yNsvCn6hU>3vs-Gs9r$IkHt -O5nTUvl5TeIxeliDj2$EUU64wqrnZH|VdLJy}}%kgk+Oz!12HW>p1yQa=IT&Ep^(%=YB`6?|L%UN#w` -y~|C^z&kKS`?Pz+9u%8H+|q!gJGNdcAOs+T -Ga2ac*hu@`6xE)j_ktE=vh%TG~3zO>8xF|oprun9+HxrO^bjAU~hA%XDt*?sBV5Xh3-uEv!UH>V~7*YvgX)EQQKmHJszgOHOxNPc{g-t3J?< -nipZ#gnpA5lS)dDtRzfo=6wZ-Fp}G^2qr0D5NdPQyfX5;6A>jy0O-TW{ehumVk^Ey0r%%R#_SqN1VMk -Ta4Obl6A#@BF0?`!As9zrof;ZV5?J-SAcun8fhG9BC+<|gHvW%QG -mJdmS?UKXdav}^=|IzHcI_L_90ol`LxJGVkpkX5MU=R#ei>JyeiNia%is!57}Wz{HnE&}<%W_eCO%q-SbRFq>E -AT32f5cgvmj^YDrn%=!ZRtf=p#dGYWH`(&t_;RtOLkLtgOu{~z<_Pv?A&45vo#ej%KH@JID~wR>ugb;GP_xWmS!lCREDRR5(Z+VF<(8} -izjd%$Zu-Fk;xNf*5zdH)4aO9KQH000080LuVbTk>*lf%*dg0GA2?03QGV0B~t=FJEbHbY*gGVQepLW -prU=VRT_HUutu2ZZ2?njaE@_+c*q<_pcD#huMJb;@*k?Js?<8Yc^CaF&(#gEhtS-u%XGXps(l&DE*jMTH++Z{7yvrRHl-^!vELcy)v@6s9!&8eZDGfvZ@F^z& -WC9IZ08A3T1h8iZg_BtD=DXWqaXbx(y`sBVMl0znPNX!1Sru*UJbZ-q>(;W<(SKVi4VOq)fZcxp68TUs+>;lN;JP7SXbllrHI-qqg7{g3Qw091k1hTSl*S1Bh&<3rLm2f2cIQNR&3 -MsrTC)u0YtMQoia2)nI#I4D;|Bj=y6>pEviFf?!1dJ0ULG&sq{6znl%v7qfmJZh`NxXl -rh++_&NScQ?}i~{ykOc~ReozR-szF=dL$nxPOC|kDDeX_)2DWMv>VU!awy0 -TEqIDm3js!4^^ix0V;cC50jYutOfWK!0a~)*PGV_ -ACktOY>L-&0ji}V1tj||2{Y;7h8-oHTUA5UMu#MqQ1_Lyx24hAn3d6D35&%xNUHRa4OJ+zacob^qI{5 -y@xE>;kx8P|q#45f^E^YCb4S5`KBuWd8goXL@|WZPn9BLWQa1@ba!%AN?e~?L|{~%Rn0P%$bEWcAr)X7NB{W;LkeX4FS*ripZe8ZTSBD-o(I9`ro!`>`x?jU2u=o4&|#d|p4gS4*V9A -eSV2$z`OkgZvT$6%*=FaBt%=cKI;%iuYcpXDEoIJ&4;wh+4ZI8Y(BfGm)`q8m?$mRX1Fafx)J7^)*as -{NH}CWDbd5=L2wg;n-n(Q6CZ$!?pKI?C$1!5F9mn^=8QhUBfQ|5iN9e%#xc-P9{^$4@tIGx!!G0KJ(} -KGp^UPJ}R+tj=i32s(AgQAE{X=<9w8Tz!lwCluqZy0GF&+%IZ!)waRx#6(S9~}Tf!AoC63HdP`;l)m* -j>cMx*js}a!BxUm<&!5?{tZlcvyf);bwv`&F~oa6yvm@ZS}rj{x*Y6SJslYcF?2HzO4e+wTJufbwN)I -46m2W#<+fd@=&oS^o9d<9Gf~@?|)EB0|XQR000O8%K%whwVq`iB^m$#Ok)559smFUaA|NaUukZ1WpZv -|Y%gtPbYWy+bYU-IVRL0JaCy}{YjfL1lHdI+rr=AN^pS9~x5?$0(%oreIj23pPL`9@s&wFx7*dEp0E+ -=la~@61=R;a -MJUgp_rl^IR+miUu31d&tgTJ$*G}FUogvCIt3b+T@J6V -(lcp#3YKM@6{-bvs|qCa?k8=QyMOA-YfkLFP -f5}6xIa;S$lQ%l%B_G`sY`|Kv*H~p@fvRkkm$k%FjW}Kso?M15FVf0B#cRGXG$dArC!ovAmBz432WlMq-4&U;y+M?x;T!|ZeXKRST7XS1&7SQt7+vl+s5(H4Hdh(UwuC8zQAeH1=5xM~#5UvkoPlmeVDmX*4Q5Q -1`8$DcCTy{7SW{=v6>!{@^|gXs+3C#`B45SqCt9%iobUsvKrEa-wV%)97{P%hM1@H7`0JMEjBm -A_j5v_8na`16^Z9oyTiG9=(!l+w8ekC`f$^#W!-W=JjY8`rc&e!QhnW2tGzSgY>yPq3gUyEA(_LcL~44!1bC!q2>>Y18t -L~~W9Rtb_7vqCUh=D?x$C&4F`)Q0Wo1&j8)a|cz!`ViA9UTGpq|_UbL>0KZucOuS -JV{!*`Fh~qDE~#_W@(sF>BMTZyWPjMkrrnG!_KombAQUu%L`(M=tkEy%D79YQi`b3QOoRu@FlWZqA}l -<*pQ8?f|3LesNbSr}iM>8~&A$2iTbKn~ivk^Ylx5B$c?H_G+(k4lD;J18j0pN4yjTl(Arwq74UANmLJ -!qW3Cb%icSQ~pK$xg3w^>f#5p%w5PzMqCy=b~z(nIU>n`Ybs;TGc=su=uAs2X}I{YVt{h6QF -)I3v*6;|f78OtdvHR?7y1No0SMo|_|*M@;&Wx!*d7{K!V?nF; -wmjw62<0Q)&ciH2NnGi)#f`<24ffJ1-65UtCiKzC#WR5ZY>9`_t0Amk5<6|f!p2ZdN-a-`03a41L#K- -|`d4S7!Bxn#vM96&m4&kN)aZ!}f{IyIB!wq9aXOg4$viwq1jI0wK7@N0Pm(j(DU;gsfc&tPzfulpthV -#Lv-%!=s-9J9bqJ}qe;e!f6y88uDPTG}@H9*FNf#=yB2UZeFtN56DY=Gl^=cM7TmfN?Mb=mVqXsu{cJjs>Ng<6_3yEjYEmXz*b!q^dkr& -g%1w?YO@ee(akW=0y2KTAvq!AZXYE^l<;Uun28n19!)U9Yf+nv)%V~&(z|3+1zb{J(4mK_*MkU)uOc= -5kau6)re^^*v;^v5vB)ccRzwMs3d}jYuY%(2*nZ;-vHup4%m)971AmHC2AN?VFBT&w05kW{XTQj9>4Y -DEnFFT%JqbdMdkCmWviF4j@*@Befx-0}hqK>#U3CzIZ9tRzQJP<|$!+D`mpd-g^iUmEp1TyQp7EKTHq -ArpG3~-9VO9s_utSD>S1*iT-S4~sN*^?)0&>PzYX3m}fC-Wy&Rz1n2+zR>Rt1tiX$A1|QEm>eY+Li&i -d`N&@vgA&Z>kG>x$s!9Tvh-+6tOSQMPqh;ZW*45ZG- -WCw%q2to`eYtoh3-epG}El+;;ore{OKr9$$O`^ZgYmc-hY2z+$xGoHyAfgDENfH=p|}$Zz%(f*8*87@ -S%EM01XbB5ty1E|@jaq32s5%xj%6((JrIF2QgBcW(^1JiA6l_+Xv6Cm*a=HvM3nu`p=<2Xify&-L3o7 -*0Pe_6d$aX?Vw^i?Yny4l1E$V?Lh;oGu_T8ckuq>I7=wzJ?&e{t>Kd^3pLoDziFyKcORY48x0d9?UW -w(kGM3xbDtnL0fTIe@$W`8PFTCz!2geDzKV^ymy}?k`F$BtB;)D#jBSH51OxnS-<~q0DQ4gQSHp-lLl ->O2P|? -D1;>AX^OLtY5+w~R7m10e^(>0zjidXmK!htaB^tXf5fWO1*lst1vkh28@60q7Thq&SpB8)Ii|Mge@I^ -dN5{W;PS_4i^)>%S&p%|TdlbHDNfr3M{rU4nBjpkBK4^Hnx@M#}7X3;+)*3`~lF((}mjjr2fG&91igB -e7Lanv6sITJ14fgMi_BouDnJOljPd4x6zuZ{be)fEVwj986dP(MOeoAdI4EG+|K^4U)nolpFU6qh&mGgTWfoZmx_r+^65OK)MB&fMxHmm|P!>{U-oQhK!2&`%~TlL#bD?~57n2hTl -C8ZhCbp;9a_+orM`xK#pt_pIC&isO1knbJLjSm;KUm@%cta7rVMU}xA-0N^c1HR$)J0{^e8!2I_JD*p -4ES1$)!3|Etx7GrcITKPp)7ZvY#4YAES#EEE6FfeaID*OwI=HO!n|Ds8!`QiFE|BqHHRA=2rtw-a*g} -K3k*8_T?vyQd{@lcnkqSKde&X1ozKRJuK3;ODfA~^dIa|-%fFKsM2PU(#umngY2 -G9hcO=VfpR%Ksnm2v1iB{6L-KS#IkZb=bSCrs4omz_+q -MBj{&Ma)|pRJFmB-eRtv!1R2Ub!R=Yq7rG8OODp3awNulGQfMDM()4;TX4_*l0P%oFonUn>CZ~r;b~B -!6J8X^a-H{)F9;?zZa_qvg)sY=+d}v}p74iLM7l8UbW*7`O_O{}UupvlQS36^-=C+^Tzo*p2PRBu*5~ -5-<&FQUFc4h+G_|8QFfN|}D)y<;yFk8r`GP!Wp)~1#bJ%XVAqU(z~dg -B9|Tq3rVRG?KnYGBr4uxF_F!CTX-#S9g7-*BPg;TH%h0325`Wdei}a#uO}5&jkmcFjLJn=SAKbi;D_o -}Y-JeMu{rR;xtREc`n*eEWzW!nKZaTMq@&)<$hRI8$Y+)aW2C?WqAzKROelL{i`js$U;pjAHvge_2-&va?*LJxeoygLKPUA9vS3t34tX}% -1@ZmZzyr#3s6OR*p=caoYlz7enUC4wh!i#w{xTQf2zb#Xi)Rl!^J~8;WG}3^%wi7@2})%+QWdTH2NY+ -MUdMtSsX4kGHIb{_c#5@{!z1-LxWoS(3g0o?;B7u@$tbNz`D9Fcdi(e4x$Tc8+TSAx|6ru)A&Y -7Qhc)G?hB1v>1wcW1U2)9IlOIpcP;+lGY4VFys5)cULaBP3Zr0FV@>RL>;Ne0~*b0mCH}P|B#s!AZ-HI^Z}YmuJ-Dn#=a72<4x4xsX -UmjisbCSz&8MIR+WQ?RZ_S?4D$-Gz#|fb__&iKXdkicNy!+ScEqCB`hdNSOvfBb>}a5qF -u^!sqBo~9KvN?C@V4!U3p@`aie;wV?}<9fA!pvfS@ObRTBC937J27YL!Z*%NgIX0MCB4Ii#-=PEb=Gf1?hREQBdALywk%Pw( -stNH?7``S$CH+wrqiYHEaqK+AQ-ieQZ-o9>TSHuJ5u(kCxR%F(9o7SuW@-5g;YtW3Rnk1=# -C@O(Xs<+Q*~0an=}1{-1|f2HWE+LIjSit^Cbrfub&%Bpew@k2w*_TbIcRz -Kpms3-YOF?|w6F2EtYmtju{>RM!TsO5}V-yP>KpL1Q;I0q@l8Tq&ygJPZ -|M8b%j3dPNXD!&q1lDZ@DSOmRnSF;A$2luoA*7kdx!ARZ=0nL#v9YlVxUw~ -3@V)SY1*1doRHbi-tsi4qEtQF_;-7hpd$&x4_I$i*t#akF^}S(Of?+|W_0<8dD&t9~{j1Edddb8xJ=l -_bH;7M0%{Sc(9&LsRse#_uaEDA?)C(`hEJ%o?k5dcVKEiokaVVs`AMjFhuN1rFc}EnzW*v6!mF*xO;! -hpeT!{eUvdl*n|ti_mh1zjR6YNEM?&L^)N?OvST$?9xE$NUg=v!J~RM@BFt1&|otT-7-^72OJ1(x64N -5+|`vJsMw6S0^%1yd??7waw>% -;8S>zK4huQy3`{bTiLiG@2)!%QF6kPiKC@Azvu!Wk+G!1zoTdtb -q^2qBD>E-p0=(;@Fd^}2hlO}k;XwGaHKvu+RA2w2ZOUt*!Y8JbX4vR()Z{jpPwlqriTDhe2Ait0khc6 -Jx~Ak>d_YK7>4a!4K-3XjnMnNq1?EQ?+(s=#B|W-hcV6XL{SJH)hY(>6$}Jh#pI0bTL*TdG7s?GN#JoVSP(AqcFSo9IH -kxydC!1se>7Q={5Ar_miQqY+BQLQi`<1HDnCc;YSzrY#TCCS)(YW;`26b$S{IU)W`3&ULdEe&ZOre0f -ZG6@GJy1wZ4>ESZ!9)(xGS1;wk$g$x>u~ZFuh4Ae*;mx-8aI%!29WIo>=0J~U3>udVGjQ-3efwb$Il` -ev7o>{_k6dNlgrH~pv-yB-}Z+z0ylF?tkYmz%*3!oD4t+hOWAYXNDW<_-EqQJY9ZLn*zqiZnqy7X8S+ -V;?kuqFW5EHC!8vr|tT88dXZy$l9!b`&U;Yx1zk`G -sIJ{qv_ng``qg}K{!2~w+y&1VM5!r{GHD`VJX|sM1v(^xZnl@g+MRwNJx7V#57eNG|<2qD@K}4$#1T= -Mt8UKdXgdp@@*8>0eHgE8*7JDU2XKeht$89i_icq(~(AOD-vVMZ$o$88q`BzHp{5zwrZIEf -YpMYS%@_zQF}=2X1E2SIKPnp6Oy0mn(?ry}j8;Nz-O4!?UiAUG-`1vb^3__fH}FDEGw5Z7N+J%>Uem6 -n-+^?LB4=Q#vBLT@VXtAA@exQBk1_0nL?c?5(!c)pmiWc<1xZ%&3fCU+|`#0Y^_WGPq%BN>KsNaq|s8 -;YEe~8qe_cTr+@i!`f -r!NnlJh@PNtWS&9m`8^_26|4GuRI>92CWmPdfX=?^boot=Dt{O06P*}gWI-Fb8^-A#uR@Gm;Wqyc&Hd -TTsjoi8LCeIds^bG!BYY(sllMW@M~k^l&+CBZId9}0<=(gPD@f2liX$|B#-r~nr@0IR}X19_cQGj~n3 -*?9ME3wCX@VwrDKVXuLa12FCTaA4XCa;>OVdZ@u&RMMHQ2O|rQ8?m>v4)EWCV^+K@j4to$M^xzQT(Hy -GR+m|LrMg@?_Ky5&h(qgT;F5l<*IP0Sej4>Lu>NCx4JqB#F^r(OunUWq^sOM?jzF=>*4tVi(-iM%3GG -!Zbi0po$GUgg`OI^=QpX8Kx-}l#gcr8X-#-i*a-VV7TSphbtE3vU!6OM83B{mhUjk#=auql%p@LMNC} -jt~3nWV8Efq06{`Pkr>?OG!Mu+(;_S9L?TBsYaO^XhUwh(9nTC&~H-II>qp5f_RUBZl=oMhvfho0K#A -YQWR<{|F#qZf}KLOqR8M(U1B+9=DYoGv9bmyRQZ^taojPU`to*x)oRuD6~X7Tx3xrrTCn#| -9#K%kd#PCc6;4|qLC=_d4FeiS{+zX?E0?Pq9~X=v`Tl3H;w9eCskusE4L;NT-NzP7{m7>UTnEGLi@E^ -?f0_T`&VWB)3a=MvKFfyA}v*2i;_^}d|P_VKMl)kcHIfhx3zFjoziP|!`l{t6O8VJ&;Dq-V({~hA4K( -`Z{gS~TiJcBZ1dttG?mhRbNje2v@ojIr7*Kyz)|_gk&9f5fA><0gJ?SA5C3gkyL7bgWIddDr}Tl>@>x -_!;hL4aQH?BVsc2y75~L@g4T!|U`iFh}6(>kTj2rtlmh<8QP^rTo1yeT!R#-LwRjgprH=zNlsM=t2AUqUGH~fyatkm1KG6fgGe6?4^>Xt@(-0uUX@TO -z-C*9FfXMHpS6$rvT$H?G*gq7FQbXsnEvfE+xP1EVqdj+_|wI``ngQjB_ubZ8E -kHpiPl`$G_QSiD3kGx{H$-7rKZ;~}jLNxdu088#xD|9@e;e_+Rd@s7c&{|rMi4_F^Mqscfi98@#tyX2 -tjbS6v>F7$?hNwdr8wNDc8yH0&yzaH?Tl?x^T{k!4p!CU`i-nRT$+cg3uho;RfavalcO-nnyxbYe-7u -0n2Srm<;QExW2+W5EX&A4KVh7#^rz%mZ3d2HR$em)qC-YV?e!>tUa!-~8T!|foMA^u$C{iQp-Fg7N@$ -u&t*j%&isT1aI&8)o)(#hD%KaiVPe(?vvi1~gU>^c0<$Vo7gLWTf -_dv@^OR#pPcJziulgHF?2|hcGQE9XXe|Cu_V=$o65MlMxR6(CN?34J6QvF1)ggJ?zX3G-Bh$qaVoPEk -e`?{N{F>=}bBt=~Xu<}TY2saA|7?4X-N&WSv^A=o8EP$yY7M%lx-kOr|-@X3>@?Z4pQ+NMaXc+kYWQRQr -Ngvx5fh@}f+To1T?q5*t(I)_q9cB%h5hU_43#cmQ(lg{Bs9K03~+YS(|C`^}1Bh!)3I(~wj>}<)y%c0 -6O828fzE6JA%6I%c$nnC2&zBmuUNMxx4Ye}6aTF^G#XYOoSvOVSy&xaP?HdTOyW{6A6cRG -Yj^h=<4uT?j)m0v$p`KxGhws{z?q3SIe!Z0bn}2M=Ula-XQcE=mQ}88lgAVn(N$+2F8mW6Q5`2SOQh$ -Cx4lMuJaYaL)l9)}6Cxl-#`e@apxe53iEVoUjq~=c$~(tmC*W@x9QM=WQK*9&|A+w9|OZWyLPEkvVoBvH?Fyv)C%WSFe@{u;v-l&RKaeVkY-pnKJVhK_=E0dZ5g|$pvz02UB!~k&RkaeAh$7 -CCf`FN_Ct32@`t<4+2)ZD;-NW0{K33M(Hqj%^Dalqu!2gLcSAh1_&csU8(((GC(rEX@00nl(GFel?PQ -n#$g6+K_ANXv6yLUP4x9hf3g26$6QWH&Vv1v0aG%96tIUBNTTw%^!*~N?ArhEt(V^-Nc>G^IYr@jE`B -x0&l9&BT@0!(R9>-Xb0f(-nn^oIE4Kz$u&cKxo*i&7Ls`xq#!yxDC(hSG;C7K_L*xkT}{i)5OS^}^^! -u4gP-*xtEwCh5~c5PJTl)Zs{xrbl(}4XdpB<=wZM-Sg^5~(6-#Dg3U1cJU@I0AjX5QPN -$^CV88*DuO5A)2!;uPqxg%2$#jB7DK?Jd!~!)*fR=JT72Ri9i)7_W_JM;iCZZpmW0{O#DSe{nf -|tEjeN&O~rg5xwWvP#f7W8~n6w=|3EVk+BXF7zE9;nuRyfmB2}J(c-9isv(?!#S`>ESr`gv*t}*4 -LJBkK_mMdZ9Qyb)%K_WLq!ZyG6x-1RbRZxHbNQ6lAk>s=I{KAJ$3A1Ig7wIcBXb7_<+SL2mp -1%)!2*|Hnn*`ilM!EO4wf<)N)dNk$iT`HlTOvV-$(|GXiknz>dtXbv=c4RePUSgsj%A6aO2A94CYgK# -jU-$%xl-8-wW*z!esBnAvhTum8<#PwIyc@vEd!f}o_wr&^dg4v<>fBw_U>vwPeIv$$konS~}tDYe+M9 -^lLXXd{#*vmB!*?_$QgH%bSif!kIOR+RM9UJ4E@nv|RZ+!`ly?DxxTsJr@Ox~$*vnZH)o -CE_~*a;|m>plhk3mnCT@a@YRV-*TAIZH -EVw0(N23nJq{gVl6K^^@fHY)8owNLCCYiC{(x-Xr3bAfhPI!7_K~Jo}W4v8JcS~<~@oaTM?A=tbK`eI -H=O*=Z!~*susDreS2I2y0extBJbN(n!26Z`HX%5$<>EsUf{H?&zOz#9fD*Xu1e-_=@8-XK=@n|lLM%* -11&P9gtM?lZKK!h?sCgW(yD8LB5ID@$s$xu7h!sO6Hr!h*6kyZ2&m_T9TR -)`r((GIy_HQxSJ1(>1MkWQ+nxw^e~&=#?~!7B7o#xqao9>XL -zXUoul*f#X5UTjc=!{9IJEr*Xn2V<&M1)t+OZW6{7Hh?)DCEgRfX9%l0tTWznpdB9tJFV-kro5Y^EsE -JpSE6to>}3(kM1e@m)&Rp&r}YJ~OcH -$0Ws?Vx$jT3L|Ik4r-$MShxg+1_I3{b@N%el70X}8_VhIo*WeM3zLA@7JWgH=Wo3|&>?31WfOGFXg=> -`WE#|}rr7QoVdgVhO582ZLj;P4__o)oYh0|Gy&C*gX+cz_Ye$9w2bYS8fjbk3es4rGtT{FTwrIuCMFUNSd%vjDVi3qEQMEs4k3CUa*g- -@v8YD -+Ca&&BIVlQ7`X>MtBUtcb8c}pwG&sETiPsuDUDOSkLEyyn_QE&lriZYW*OEUBGxY8gBQuE5-@~(MhnM -L_|xv6<2TwL+-i8(p(@d~yI(Mo=fzOG6-3QEWZfcQunlw!HKK+1sXl_19DWF{$b0RT`-0|XQR000O8% -K%wheqW~rdjtRg^a=m~D*ylhaA|NaUukZ1WpZv|Y%gtPbYWy+bYU-PZE$aLbZlv2FJEJCZE#_9E^v93 -R@-jeHV}Q+R}A*W+CUZ(^v!^aI=!VRV)&t#haix-66Ix0kpfAry-D-$Jwr-bt-V29!?1UGIWu$S%nVy -+JF2R72!oSVMXKwq3nYxOK?G%u&$4kX{nKhwb~1=Y1TpPcpS6q=_PsK96ULR<7j)xnu(j0%)xw7?o$F -O!Ry%1L>&jl#_u>xM${{H2r_0yYbwaC8vbh$WH2o*tZ5p7UA2oHK7YxJJU2|^h_!?Qb|XzZQz(j>Q#rAR}}1wdcL6JwOG$|)1c_76x$FJ0Kr!eCKv}Ca7?e#^dA|M{S}z=q9q=Y;ybv^Fj5`@nI<2j=tz$39n{O7*kumyrQ8Nhr7lpKktxT!|* -ZjFz+&=RM(KpILl-BLXY`wes@%M|AGbgnI!O+EXZvfFdwG)ALQT3w`Loy*t4miRI$Y*JKlsbcA`Mw}; -PpMsTl{of{{ht&mAA1WjIh@%P33xJWw5m&b3lX8|o$=*HCZPk#>LykK%MYh(4G9(z5XK>#h8c=tCb6@ -sp{>|)gT;3zADn^_)_Et)epF>i?;wPB&=4#D&^42PT6M}WClSxv4&Y*c(Bz``AnP1M-W0To{H=MgYxN@xqHRKuORVRo0Qn6ZwjI;sjZ`4g1_kaarSJ~_aT>Cqhkgg)T1tPu) -iGA`{Of`s@{!C)s6-obo#_k8i&S6NNeZ0#6l&whsb>J0E2jrq|kGEG-{(3^P5(E?3IzV1Ds9!r(%IpE -1_isMv3K`yWtC0|XQR000O8%K%who`0Lp5DEYQx#^*R?X-NLA~zv;Iu%xva9i(3;-~cW6H=kD=~Y__rzXIms#sHzeLq0jiHa8$n^j#Ku|h?Z|E(SI|lP1$?iG_t-+*+0d{PSiz@O=* -YV3TXaa-uLI`h|um`e8%9f;M$~T$(e;{WRvBeL|v*jYj|IZ^|zw+f}R;!up76&2Uo+QBzqJdgU!f1j*~S`< -Z>HM_li2Kx4v6v0Sl&PX5G8DY4r%uaQ8yJaLO^aK$v(33U5_xXl^Zf~vl(#CT|cBbscq-I7b7FGPqF= -AWQq6uGWRtr&T79Rvg?MuV%g=m=B3HEQ<{D3Axd^=ELC75bNO;_}ksm-sT< -)jz#dM9UW2}rSRbtJ`TAFmUh4VGfyZVEvN*_hwi-4&ZZc;d%Rwo&Py=KwO`cmP -!#^#mZ8|)mGWE)(;F~_XXID@*u9TLa$phKu2z1utEpn<#6(rS)E#D&Ogq{BxGG;AS0RtHYgk~3(f$&{ -`9YBI-E81Vkg+XgORSvR}zxNI@%3D-=Wc7O^HS_z@p+%2_v`o7gJ;}Yn9$*x~ZgzqgPSNUMDx_uh&TuU_h@$Kwt($&hTpm%*G*<+%k*J@Yd*q1n -k@B!SVn2DjLC)&|!e(hX7S8Dr9$cGWU`2+pZ$8IkWCf>M$Yaq@4luaoPJw$4iQG0O??nk!&lVFRIZsV -v;8faVBH@NXYqcHKwRxRz_FO$P-)}vm83>(Jx-GpK1twURCNqXn_<|E3<1U6=E`&XMt=KdV4}n@WGgr -nxM|S;95;AL?R>M*aPx(!X;FAjEE1){GhtJeA~9F$2}^93~a<_)2jrr41q%$flv0|Nz6k(5VAvJUij= -TD=>EWll5@l_BHHBaU>2Gdf~1PoB&7RDOnH3-8|$6w$-xON97rFL1GtzMw -B@vHS3$1z~~I7g>D70s6vVe+-ESLi)I358@Fm``|S3NIH)~smXZLhAt*bGq*?SM}rcl!Un1D<<^X@^} -d0}HJNuX6GD03P4&)ZWL7GnvApj2y#S@uz#EFBI%u-$^S+q%-4f(aa{kM!cOJIBbJMKDN{)=m8q%4`C -jJV(`p(r@VLf(eaDv`u+O77@%-7UNEb^;`ktmyYzxX#Pj;@dEW5S1n#0!&A`5f@c8j6ge8^g!s&C?}x -3OnV4E=&)fE4)L_im~F3DQrv(QxmL~Rpw`O%m4$XB(MlF`LszgZsJK^pF{!{0K_%P|C7f)hLh(s5EML -U^1&lv)I8CN48RUR|9=E1l^;vAAoP<@cFZ7~n6#0&5e&zN -I%48mb{n%FEzPQgGG-^;qZh#WEKVD2il0arWbdtO6-TwYwaok%ho_Z*63t~cruE=CSsvFQEV>&rKn*O -$>UbVlOjq~>_0|HRYP_b@Y?wkwSFv<636D%RAI!xV9VN?t%Igq*WftiRqrLk69ropjSB@RV}Hv)*ZYP ->N0Zy6OjU7FO@mE;Os6*fI8V9oIwSonU>0K05&Qlrn3bv!HKxO>)m=HjmxDFuOJ$2Isz$UntDBNcd}y -I7HibwaW$oVVjmQbjloHZ{d#C5)h{8Rh;+AuoKC9v_w~ -yspShdY>lr5K5Jw2N_Y9ztCDR0p(76AcrHkuhBbLSmN)X=i&0~@_d0N#Mk+cb9ERg~?ob0wIb{m~kcF -Pa>OyS@uGlMQjK03Lcx97nSx>2fn3Zpl)dDEf>4>s1bgL#+k(2_IPMJpTIYv_^xVdll5!0$0G*>KfG< -`7!EI`E@r(H4#(79(kBJVbVvF>!Z?^T9qznN}V;iAy%P>TlUzYRXD1h}xg784nr|7UNif$USua?hr^x -gTV*I9P~XjLcBhDGFb6-FohH2nxv}>v6QURfR$t^pNvf&IDN5`RXGmtVTS${dJjr3QO``2WNs-jL4A} -0@DMgy`J5iw7hEXt$-XBV2K19ove|QMoa^8o57mcvY_Ig -monFgint%SKVM3+4gTs~jk;&o`%a0fi{8~ogsy|q?_GpIPu$5-^bhj -8f&dmy(ow)ou85BV>%LpmfWqY71_r8@9|)^s&=vbdZ=-{%@s3T}R&5nwO1_7=1+0LI|18@ft#H%*IYb -`Z2$$HA$nI$CBxg~DDCP++ZU__0#FjFXK05^@J*l^sAP%W^`D=XS47IRvbPqXIC|YYzELJ$xNB>J{r6 -Fsa~s3UEN%s!w)|aiV7|Zb-$i3j6uuyFYyQ$KTC85tI=xnk@BQz?o&TILwi-ETIp8sA>6<*M?<&Jxk& -@K@`n?YK#f-e4k3kLrNK@53~g6FOIG(WBU9p0>Th>vmTMGr{zF2qAo>U$i4`q7C(%v<4?O(3$Q^*LIJ -&BSl5#T^BIR;_nYrcDGP&0QVhp6qPlsAh)%R(xb#A?sALfBb&x_^@iE65fj|G96uM_cpX2mlWe$_n3n -VQYaNCV2cn1k<(ewkxgzfp!S&Fj?5+ri)pL&8-k|PNOQSg2Ta1db!2-QQ&n?{t`q(poK=_kKm#-oT?& -*glZ52XpfvORC_MLVCsY=HoXa2;9oSFK_xWJfHSfYzsj^bn8Mh@!{jo3KMwjh9gNZjjhAXdg&hy?Z=K -<$q910|XQR000O8%K%wh4G2><{{{d6)f4~#DF6TfaA|NaUukZ1WpZv|Y%gtPbYWy+bYU-PZE$aLbZlv -2FJEPDc5^Opd8JtGZre5#{qLvXJQS4ymk-b(!!|6%3T)}H7U-}b5NPSj=0b~_NF{Z#hu^uRzNKav79f -c%k%#x=+;gv7)b=dWv}k-&Q<@5Wu59gvDog8?*S2(Yf;TJnHh;eYCE-s+2!rL- -)~L`X`@tRcJ$-g3r&aF_eYVq&j4s5pO-!%G-J`-qT+Vm07#oUn-o|t$J7SMrAhfVWlgXK2Vw4T2?B%S -8SPVyf&`ek!%Eh-j)x_==|e}$h?6mr8?6-Jh`#;8IL|we$(H}RB20{h1=<{v{JRB^pP>rq6gWHZP6v% -fPH<`{lddFOKBX9EPO5r_s&uK$nG0jm-ihI#^!apFQ?rvtdF|Ev`CdNPzqt?j>n}O;f~ZlT# -r8YwulZ^nDx0q&w%e0OU@*`c@pQHLpXwbs<1Z(Jhh4vwPJ3l+1kJ#pV1{6R2H^H_A8~V^}Ado>DswWP -j>YJAHi_p01@n=|DNf;YJU;3mfTEMwMUX-V^7xH`dJztwh;cLorw6j!2UCxEh<42wkh)|%uB)ExV>;T;Mvk#@Mn4$!$hwr2S_bQ -WhW8ql>&ei-*cWl9}o6JBk~vw7|{Sf+LRn9Qw^7sXGjOg%7ZQLr4t7_sRwQA-5zq5hn(k@g~P)`had+DG(=>5picr`$h|R!zzAe*0sUQ;WH7E3TD#v#)2 -6vO)2)6*QQ`lw#@?ID`X$Se%kU_ALug3KDKE7{UlEXx9{sV^fice0I%bU%P`Hk@6f%v9O&f~KC -zez$%~k4MTlh|=ib4_4fl_7NJVGSAFK9T0Pw(=)56J>N%p22TruK?pV~WCX|csyEb_W+Xc~wjx`2=_EBAR2i+xh%dBRsy8!nnm< -m}AoK?Im&|6cy{aqB;7z@}f#T{3h4?jNW#!VtnT+869)FD!rl@Mx9cNzr!XllK)MKhe6~yVAwk`1hY`hBkVGj;~XR_-tSk?chh92hK?$`q?BuqPD$? -nX8$+VmT~y$czk7ac`ta>a$$&QC$R-bIQg_jk5-KjG^WNk!t#bkjY%VnF%dSjg%gj<*stA)AqHLEi4( -+im~b%6bW;#hC=U7hppEty84M_N4aI|~yrOgEhaLf+s5N@AB?6>0H*oHathri}?XCoBv%Z?O5 -X}!eMD#(CXS`te0lM*W39BdZXE2H8hI7IA?Q2W-0wC~gT>QX5>Z%ue>I!3&dO(}s4o*w#-^+m==dEu1 -TQ2GoqLwNg{Dm2K<4&d%{=HgFUS)_t+jq1>ALv`P)h>@6aWAK2ms3fSzDh{i=#RK002G!001BW003}l -a4%nJZggdGZeeUMZEs{{Y;!MPUukY>bYEXCaCuWwQgT!%NKDR7OixuP$w(|wNY2kINzBYER>;jyNzEy -S2o|Ll6r~oY=9MS_ab`(oYO$V@k`fmHP)h>@6aWAK2ms3fSzF8BYdc2)008j<001EX003}la4%nJZgg -dGZeeUMZEs{{Y;!MTVQyq;WMOn=E^v93QOi!lFc7@^D;A#`srUmT6(Mm#LL3m{f{?W&o66L&BYWEesQ --?gI7zt`U%Zc*-5I+JzK0eW8;dbg1k(>b@LCe0%aFkgN}tHJKBxhk1MZF6s}qH2yqigHQ~D3afFTlKo -C00CmphGP3J$I6eM>efYusUGKc!d*aleS&B;Gf<-D@~fCB_mwy2xlPavJ=At~CiwwSaV_lrJ;{isU!~ -!r(bY1}(^DMoQ>>0CO=!C7KxdT1V?SY7IMhmkxW9Tad*>4zpPKF@bZthR9*e3aDP*EEhApF0OLkQkQl -ZXISe^B-?G*Dcg0#n62CjZtvjPyJ9jQ|M^7?tE870nZq(}7M-u^3i#^n&xdWwU0pL{e3VIWb?GxMpwU -4}R|A3g7#wu+C-+sBdlSyE{1VIxU;>ov49KOD*PSaVegRNR0|XQR000O8%K%whn~;%qlm`F+n;8HAA^ --pYaA|NaUukZ1WpZv|Y%gtZWMyn~FJx(QWn*+-b#iPjaCx;@ZI9bF68^4V!BY#I)q0KO%Kchbj9RcW=0wvgKhXXz;cHIZ2Y|yskn01zRNye(6i8K!>-gPN$=V8(zzjOuJI5Rl) -wNrfS{vQWYDy7va)Pq3v{?Bufxwm6)W0`c&Q|BhB=C-WrboHBqfs#C`*g@QvwpVahlM -w5wT$93XZJZaXPAHqdVRdVnReC8?%5dms*^6si(tVl|B%+zoO#U4PFM>g>r50T)SO_-_wNm49B)=8+F6LCO(Z;^rA;&co)6)Ird{z?I-V^3Ar_9k_2V7&n~K^PUhNdC}E}9O -UF}HIZ;D_1pu@3U4XiO>5tp40DS{F*<^1fdh&V4IsD74wz#WG$jtPi86^qepR}pTd6f78@j%&7&l0hB -TI0qRCcs4G>yio0UFezF-B+XNA~<{)o7Jq-G(<`s|bR(vaVP?TZg_1Bj@&A@G4nDpz;RDCf|K`|ADJJ -2p(iFg(=Qw4;`=hy`VkF#G>v?^dIFe<%RY*VP!tt&)J^BdJp@9qPC&E(Iq6(7bJfxx=mY}i%yOM(9F>XlcL -tsS;PW{l!3=<7cbH_KIjb+=%^&%5)=WOx -PB8Av%3N+)RUD4(!hJziw%69ooX%YYx7)d;hHZCSEXbm#3yW8Q0^w#18bUG~LgF6@3-b+O6(s1?`DE%Esr8v48bmU>!0UeK_Aq*L|LCA4-QSw;0IfQ -lZT-k}CAKXf76AFiM6wZG-4?dOmA~XCdt0UNKkY5@G^rMP;4fi=NEB^RwEakYcwG9IlY`AJ) -(wV5C@MJ*o5Y%+mMNKAfPmAgsBP@!a;h>9+z?Q^tnv$|`u-n)nH0J7H$CcuD=(|B)HjXj))*?u!U2hC>QK! -=-AVZV9C|z04$&vj((l9`0o~8Nz+h|q_6pHzRAR%e@iDx_i}44Vqn2cKvS$krcotN!Q3;P8I@w6P`&s -<{1RGys&#|yz^$8aK1HdU3>uBZltMqr2IE3KyZ&D1ceQFM9^NDy8ZMZT$ne%!z0JCcvhZSF;kFE)v{9 -Jb9XK<6Bv-AGWSRb38GAH7(rUIdJ#&H+Anox4tA5 -P>btk&M&j&n9IyDtV4)@_$fE0|XQR000O8%K%wh|0GbUQUd@0rwIT6CIA2caA|NaUukZ1WpZv|Y%gtZ -WMyn~FJ^CYZDDj@V{dMBa&K%daCxm&OK;RL5WeSE4Cay*yFpwd6(J!IT)1+mD$7k~yQYqv+D^Bn^uII -q=3$4W2ZRs1@jSkHkN4Je!1;dggGJ6kcD=D4gx1CjFO7CZu@5r6>??kcx;3`yMRO}|q`s*JFO{2=8lj -DrjZpF@axFScRi*Z5wNMp4G}uSvN$;cSI-%Q7)*4$BMWcjs@Y&c-_>V?=YgAnT9md!fR6SbQ8w;coeF -ztlbP#PF53)IcfRl?De1dBPsht-}VOtSM5z#nz%6LaHU%(~PhF?o9>?7L(Gg>`P*Q=9si+j?LT6)gQN -D0TG+HdFEj8PNAUpjBE;1_%|8p%ZXTSOBJFAHyA_dZ_hSy%2COCa5fn>j*I1YT}Z-Nf{H?NngW&myi; -{P{Z<$w87^nohiHMS>foX4i&e3OS_k=j~QjQ(raE9{V -(uq{&1QGIHKR~IKw7SdV3&q6s|a7HRMSb2jqHr|d=-EQFG0w&t%F%d)m-cnPp%x}`x9w%cj58@uc+>~2BhZp-kNk(xQWRK)w#41uB*yE -z_a#NXp;jR=P&m7VsYJqd$c0DI7#BXbcf)C;JSJipA=HLlMUTJQH;+S8T| -c92z_{;#v7H%YjFuF=77+_(H@6aWAK2ms3fS -zF(ism;Ox000IA0012T003}la4%nJZggdGZeeUMZEs{{Y;!MZZe(S6E^v8;Q%!H9Fc7`-D@NW6NGZ8T -%HDgaqE&CJmF2=v+&Z@98QSLGcMKs3TdE=iqw%~qA8%}rF(if|IcJ~hXcA#j6pb@Pa4^lQdB%I+;>WH -4d6ZIrFeW7QOkm~+7(>AL7Ger-wnYLoWHEth8WbuHfWD2v3QkChkxgrIfk|uF!(SyOs0}EO>_U?*vd8 -F9pZ)O^?QC~e{KO$iZrz5Xv0nX~I9u$5*LurZmxQj{3#nu94i%cU6AIcABw_A1!XlWpCFq@&@!JP$+tSH(P(Z$SB2AQ#Lpch2cMAbhg$x(OTWGbrS6HrS71QY-O00; -of09ji@+@@_r1ONaR3jhEs0001RX>c!JX>N37a&BR4FKusRWo&aVX>N0LVQg$+bZKvHUvgz|Z*p`laC -wDRZExE)5dQ98ad5uGfF&m@nzaV@25bce3~P~XpMpT3rBlS2CRviQYZUqKy`wByV%%1MVTn9G_x9W!Y -iD*)8l|ud(rBNHO$l;mjSKJ(er!>fIICGG9Iez({oZf-IP2D}w0W^dT^W~KQEtT(iRC-tw&;8>USTa- -6$;dQ>5Sf?4(H>&P(d)xlNCiVf|fd<(?aDAKk|=L$aj@d55`Z{ -UA#bUYRs9Khhe`@ixjbS_-k$USx+jqnY@?!+;H-5`jKC#T@hppKxUxWPgAKAu=XM@teRSyWJpvcUvT= -qgIYJs=2oj3Cz7mUuxRn)7mYtC-&{8$Swb+UTF&GO9_f!QVh%~PK|NJ#-fzX~JsbW! -y!8~Aco_2#m+@0@!mkqt-f;uHltOCzoOVUZIXh?KbT~^^C-iYjQ|j0c=a3hj==_noxTg__)%D1d#i-A -R)@eRG8-`KerFg)7bIhC$Pp8EE)JrLh^V|6}%X2EZ;=Df^{fXR(tt9g>+2H5J+#Locd$@>{;WwiEH)i -w%_Aw>nH#z_M=FRGAbf?Kq*PaC9-}k@2UqjhCYSBR*m#13yMv>QC)4m&@-6aV|fZU33FH}pjDghl!oz -yNV!QJ4ZzH;FI1M7*`q=x20PPd1ssctDNw4}z^ct-@^mJK(LEb9W2^ -+VIR1dk+W%yLs*f<@N|>!H;=Yoi(5Y`N*^u{VVwI43LGm#C^=zLJn98Jc#JQ!%gX(TWujY9+lV(ywlB -Z{NO|(rAQFRJw2-9Z%xt*%wet0|XQR000O8%K%whK}Y*`-x&Y^AyohX8~^|SaA|NaUukZ1WpZv|Y%gt -ZWMyn~FKlUUYc6nk)ja)o+_sUw>#yKZmxRo%XiJWhRM&0ox=QV9V*6Rn8+xfE*I4|`6(w)b8YD)Se;7DaPYD9OugWG=E -c60;&*W{3q}e((K(c7!UZ-tM&~c0LJMI5ovGBLym?oG2T2#5bO*Avqje- -0Eo1fLtt7`Z5kmF<%%nj=$m$Lnq+M+bAU`x{!Hgj)3c-7?4ZzWoQu4aK0O<6#LChk&P%RmIUjWU;2CW -HDKo|dquw$B1^xLJx8fZfR8rYK85IJ*kkI_Q*4_R`buU%!6!=IzO!)8|i5pP!rpR#B*y;_=~8I1BLGq -p!cv-;N%CX}%qqwnO}Pl|K#<<0qD#BcRPsvYNkbIvMX#Q}At3T_Gnh@xm~~28F0(Oc_||hF2w57c$Rf -PRj_cMcGM}1aFsx3K$ki9*kYqxC#Pzj+8qOR-I~t3m{Cj304TF3F<{~K3eFx9_9eeM1>;(8=xZy2BId -aocH2i54B7*3blh^s^D@7Jt7b7bXGUO3~eP9a0@;HAXShpWp;%FZoph+(-2 -uB2QuOhA+LY~_%CNHMS!4p?!CyJA!slCmNa9$4xEoJxWq=9*FAe<>8oD-br97u_Y!T{$q7S1p#&nmjV!SV5Fkbd#wfC#S~Vt5+&}b79do_71I%UW?hO5gvN|$!gu?ho<$$BWnE --a-Np0h-QPYv`#cU4gdy5~qljI&+dTX^On$73DuNxAfcsiTwB^VaF>#~)iL;O7FSbx)R@X5zPA(T<+Q -Ig=pjAP)r;(E7A`TAz5WK7_d2ENF9R<~jPm`^)Fw7^k!7JeYI_C!K79^Q~9keqqK!J~0H_A#y7J(R?a -52g(_<4X9$j2!t>THR^#B7Gd5X@AnTLq%LtQ(N8RWHJzy=zo=b=R)DmW7EV%|L~10ElzoRFzDMR-$Oq -2%3A!g4yDMz>7YXY*oR1s_21jM~r4#qs)15LQj{YlYfMb7oPbc#R^3FLXP!Dxk!pyL>elEy%G@yg&V@_}m#3TdftObGH9JOK -Wq~svv6s&=2=Lg7!xpfj3$~D@1h)(qt6}-WC1}T?Hrc^#!>+pDJor!@2iHs%&?iV1!+5P+fd)dL)g&$ -N%$&F#%Yi-^&gU_;CU%;^5-ZZG6b68^)kgTISNZU-Wl{R~84dxq3$b5_Zl(ynfAcNH5kXWWG6}>2opF -jpS~LZ~(czO2^p9x3^n+I)CmzWC;0Fo2^S-OW{aH6fP2L0k#L@6|C9c4XL)eQeU24B0uRHWEX;hxLYj -gwEgq}xUlb(uJ8pl2@P{=GnK8_a1JBy-G7)xO+?NNk0FLGPdx@J!q9z(b$N)D%BKD7a_ZRJvM0P;5@= -zj3DRP_ue`~jSV`i5O}!jv#u6_?A_`0uEU8u08ua4>jdvtwjuS$AbloO)sB2wqj_Q*^iJS^_feXr&!v -Jzm}x=ns-Ds7t^{19X7Fnj7>gYs&$Rd=D{{0f8QAO%NKwniq}}TkZjI^oJ(6R8&_-&>jt*!OtV}+%S` -tNPv}q@oid#S~?0JJa`bsy>NPZ33&yBdGwDq==Gx}pONy2XaOYz2T}2&xW(DR*cA -x)heGSMY%;K-2io&G4#uAcd*D}r8ove&jQpiW5lk38`T3KhZ{iTuU-IHFFJHYmdG_@5B%W3(5&j_=nx -Tt8oS$+VdI0wc7PA^?U9^ivw0v$a76ZVRK3b%~?S&;w2{`qXVEcEE@f75qW~WmWM`SkQC`~g*CVRmnA -h?yPCVij*H1fJoxGeUVDPH-{XQ#M8aBv!w+!}Tac7JWaCnhFt(Q=G;Dj^*~B-fr9%?&-wn0IHab_g?G --mJ$SFfd7|$mR?afDR!KL6`;a!tgAP2lO%%Bw#{il}9}hBQr}JJKX;gU7fP{Nd}+X-Ub{=sUVUGSl>^ -M7RVFGV!J1?L;n*KYhFhO}k6omS6(Uq__uDbjNs48#83R9 -EFmY>0Ybq2nxjYVN^;YrI$sSD&A3QV;#7k+Z|Fyz1x10-xiN01fsfzjVJ$OrtT13OJB=Uu -0m2a%}?e`K7F6BT8Iumtk62LpY~|h&+DqfK!VPkfHtvO0U`MoAB%0C{5Lsj^5N3wkeqtZDh93UKJP|f -=G#i)wW0>)n(Vp6orK3>Ov^ex7q~k7PF -_pLMU_{Fn1-9&4f|>!v&*on-Bsv~>$SS7YKE&rSNwv8B!%Nj9KDDds$I&U%QV7ZJP`*_F7IoOAz7aq^G&&_;x5s ->-@mX^K{$kCjt~hNx-P!_C9r$TVI9M-pxmcnDI_k%Zfd-ckX`r|S)&xn{mcz`b*yU#P@cm6vYgu69=j@f0P&@Ju%W|Epeg<{b6CM!wHZ3u*CWHuo@2T -kd)b6-!+1UCZDg1Rd2*M}d_5A1Oz!%0S)cFMU-t*Xg#0`T)Uy5GTT>UZveck(r|%z=7qnH -BJzyemBYc=W|%oh`?sA8-Qt9@v+xSA1!K*9XF50M%{AQ-C{E0ylX<37^;zroOFT&CH~|N6YF4vHNHpOgVNVM}M=MGA -UJVGT6(X3$S`}c?Ha*c25d%~x5j~?t>Of+&M5CggnuH|^h -CLi5-_%hO8rE!1XwAp2*$7yXWA=d}?7Y6I%39=pM%@YD-clCEDCF5tZ}O&&nbWclg|`koN^#J> -Tx+T5+5;AqElEE31Px<5!_EKkYd#%;ktf;_&axJD7`H9rP*4e%mZJqYVSM8QG$~HW}`zrFJBy_hIqg@m8cm@ivB@m#kW7h_y}%C%Gx7J_9z -Z5rjx0S&vY>E@42kJH_;81t+e6e@dQt~{b7Ic#yhFuZ$4I;eo6;Z7Z2S7*!o}b-_WTw~SNcsv5ATeWV0lggK=OY3T_>b}KA{qZD;lY9T-m76ALKgx(48sjHb(K56LC8H -M(X|u`1cV39(>y$j%n==nx0mhv%bq3BCpgB9PDxV)KR(p{)~%mFaxmOtcw8A0PUNRk~E>+*mJ@8IIkm -Uj_H$6ljsDJ*fn6$RTN@q=drf!dhe0B8z*6EWt>NwNC6o7%4;gVVP&yCPNM; -#w#Z)sJHCy7)?UrEPB$92&?AW>$r -yz=HsS*$h}rpa!PlLp@m&u)=kPT?y%rPo6+4y7e9Ww`PbQE#w5__ex`c2o6d2zn&f*7GUd4DgLP^<1d -n{xO(h-B`VTh6eHI#lZ!yDD7)~Xkb{OQXAgpQrW!9F){&psJ>vI0F#&_#9lV-SI`#n~e5=pFw!?x-09 -Ehy3ZmSm#C^8UOs!hHTRXXAqfER9G+MC^%4e_Axri`uz!HmSM|g3A`WK}}r&9B%3uJ_;?BBu>NDxmW^26*Ud| -0qZDy;FuM1m*)L#&;%VvUFT>jW1OuZQ!y}!eaRGEa>Bzti2AOm)X_WZD%`@L0g$4~$}X4aw}C&{*t;W -Z2r+KzW?($@MM&zJVi#QJgnx_PJw5oJhX=nq``LWa -|0e0d+2^J${?}m*H0K)}o8}!uxO5)nBFtsC@$rRo(aq|Skp`yWACWBB2b&rI^nM4+y1v45>?=&yYI+5 -lQd*0Nvc{XVg)W$8I3Ta)$R0Vg0q5QnvNQ==ZmzGVqa-ZlU>Cg?Wp^|DYj@G%XnC5+{QXO4Oysr>tbc -#se;h|>&se*EcYZylOCVijSk={mArDTKOvDnp>Jo>|L%wdVOKIK*Pbg8NHZaRA$A?_B&!r;HgcEQtVG -kiTnTJPTCJ!f`7fN+U~(}Q7u`5l_9ia$MS!mWH#Z3uleHP)-+;NA5ZW&W=8Wjpr(-dwr6+se>3tlG-I)_(*`j8`ZkZvuX+Llno -ez)28H&y$Xv_+K{JeRHT~eRF>dvdw*A2aqgzxDR&lFn;&&%ryMbG;E!`P_(oC+&7|QfoRLEb5vLHy -^0I(74YqWl&;Xdp^F_+|Hh=)-Y78VU3`lGkGtM?1*j)kWgC7~fcvf(Q#V=I>uz;<&2RbO_lDTltlZRC ->v+)DH0vB5)%_7bbT#o(8RjTdqhYfG+0z_84(p7rElo9XjKn@=J` -5sP1xR-iW~P+q$N+kc>|P_!?R$=7*nN%#Xt0S_L6|`4k$$n6nQY!P8r0$^s9NyNZvJb;TVfAe5%xGk$ -deUE-u$w)6_0MoPYb3dV}M9VWq_L0X6!plNXtoxWw;zgwl1*|I7!svV`wdY}p{gvTn|2KOWD-?c(5K- -0xS(yJ59`-V`B=Xh@s;q!?MmS406Gttg8%VT+=frffUsY1Wc=du%x0C3Xmx3ZVj2x~mt>9=aeSQ=1N* -{B#gH`cgDf?JkS&wzHaI``Jo@uTsqW -8&A5E>idHiu94H9#ccFdQVK!Xj%ri>&c7d}Zrq*RsZwecU5DpP~|4l!l7vu39b0SObf}9dwe-qU`EUu -~H-^G|;SmpsLv$iSH4r6&DI%$$N2G=~CZSe>07~a60AKPoB`lPFCsEiTKGalqdl~{C38deBGVxUs^T& -LJj}&Hbp!5=a_TqXh&U(a_k_9uXMwjhhYjx-)5ZaZaOWOG}=Xc!Y<;$;h9d?JauoU;+4m*AAbG2Szk% -CL|%ryq}!1Y-3SQP>9Qt1(APp2X0oBrXe){si{2z#EJF*6O}Ufd@-~6~rtejxXEOcl)qiB@kPy{R`pL -Y5&F}^8KBY)N0{Z!2B%V_`*PvS1J}2?Qw^j5J`U-ZVK$dkk+sdrTKk&OGz!ymzIQj|_1iq3*@j`-3 -Z$P&G$1d`shjUhwLv)SlCN-;?4C`(_68W^%4G_EcZ>G>JZG&>kP(B;HJ{lVz09$5lQktJFb~5%CksE~ -J`2Rpcd{^bafWyn!TmA3C+Jm;4eqq^E5=Z#AR|MdX59f&EQhRFn9YwIHPgH6hByP2bQq&p4_lAoz5fG -HO9KQH000080LuVbTh=>&D%1r407Do603HAU0B~t=FJEbHbY*gGVQepLZ)9a`b1!gtY;R+0E^v9}7;A -6aHt@TD1>x?4gt@}@)qsZvZMFr?+M*pc6sb$l(m5tVi3&+Maku^V-BC|lbed#cF<=A|SmeEU@8jC&oN ->OcTxEcBCi7Ao$3#(RC!ExUolKg&OgND$VJ+Bp%fMt!1X8UOq#isGtF)zZ=D>)RYJdj#?2u#)Q3|(Fa -@C%`!O!}@?Mhi(w)?-ne{;&8|9txV<=LyBX6&TcO&WYzCVUG;rcDd$bgm@783M=x#VUkKa-D&)NhS@- -&MTu@q!W{Fe}_N~T+<(JSSM7+@~r-R3)dBjlSAP`$d!}G1|{t -@h`+hhoHU1qTJP&-K4G{98su>6ps-}8mcTnu>&fKJ`}5b&U-BQ%&QJNPlV48x+f%kc^CqdzOQ|3-(~s -p&>GbMD5?y>umfyx764LEA(|9tOWUyw^@~u!Z<24*#BN{J64iO1A_tuQr(;wJMYgJ2T0KKm8W={x`Ef -#EAqkH00hK3~Y2qnpdOE*zK&{Cq;j@t*Zg@BgD>(kpQZ=w3BbG!@Jl -O5(rJ!2W~%=_&ZVCoCe0wkpYlkj1Q4@#~U8P$;^_vbn%@1r|ZLfvqINkBFg#bLqIudY-mg)j@%J+k-z -Y{fiFXQ>f)>=|LtP<$<|3=8*X#I5ojJ{{q?>Lt6O>29Owd#~{&#w8*%V`49nR2HL8zHo|Ve&Ihj5v3| -tG-(~2#^|i^@d-Xm}xUZ$UBT6`a&)+uy>VI(0&=t2}4ucf)@EvmJ!d!v5pX8=4E*3Mi+YpocAZ1WuL4 -_ifSoXb$n5PCgDt9_yL+pB&J7S*>X-byVKlZi1V>e~40LXAD2C=Yj5G##X`tR!B3b -%+w3iUP-G3Kqa(4u?SkNwh+Y~lvnM-^GPnjarzz}>zl&IsKnN(h1=WgK&dmNo-lY=*49A^c&&(nWZ>U?=Vcqt~XDYfWk(i?7s*f(V_r}2Ay^^vQ -HczMGnMF07PLL*#a_%+E}{-Twn^V8Ha7uc42p1G+_7^>nILd>@jGVc%e_O_q@VabhE(>7;rVjr -~Dnbc+&=i<5ft1X?K?6077XPvQuNU2^g|<=B7OMuG|O~HsS-4M=fY5*6Wjh*#Z{EI(owFMpr6h1z;nD -5|-F|jG*qL#WDRZYu6nX>JzJL-hpSan-5MeJi#xJn-!L6 -dDYSgK_RQ%~fWE@c)vgD4al0Pp&9uTfbUE4$yDY6%XSV5uhF#}z#XLkQ4^+tGl<{>Vw0ryUx(bd-{vo -KA!M`Z&4Mz(${XY3_dZe3%Vk?a?VF@lpeVI=0oWyKO1UY4 -f(?Rdz795c1%K6`cxH7U6&(D%CxLEeIq22NMVl9N2x(>sn-STJU270?iNY+Pu3aS(TSI8VV*XO~@X5# -)dC+i}9l!z2_0p2Ms=hM!X(@0wDA5ZTez78!ZoqBTp|dAvJfUi6m5O?8180j&ndbD>Rl&j!twm(e$!~ -6BoTk5*)olhHgGC%;Riu1?VXvUuA-ob5E^v8eQASGIH9W^Am87?4fs3pb#aYwxKw1 --#&?$8HZ2Ri=^*jAv9JcEs`(m`h+M8>IhGB1iMTFJYky4dECH^p5sYNoZKL+-Q*R~VGIO}1_?4do_F= -1tA9feICENcNMWZ@S1jLB^~epO>%8@@4t4f(3kLT_Om6oy{QYwoK#b2rFXgN3rZX9mbv}`!dGd{}Em% -$scML@4TfOGeEE!;?51EFWGt+0dO%?0v)Oo*gJ0>AQNmQHH^|118Wq5$}?Gw?34dG@<4;EzDBG6p9SG -c!JX>N37a&BR4FKusRWo&aVb7f(2V`yJjHW2>4ze05==c -OB^O$mjAElEQQfukI`lzvDLM{8@>%Gs79$$Pgc&2R5B()!)k7HaacmPVs_dFGk1Dr;>VQ%)MEbzxU3E -EM{1P}QM*)!jy|t$L`1%iD#<6ouOHS{N%=d!tL12`AD*SSu|%3w2AYAX2NECr4Rj+Tdh*C+ITXI8}5< -Qc>wjrJ_)uWm1V!4v(q|23MlsrOsqwxvDbx(4M?LUB6k9_3C-5OPJrv=E5cpZa*qsDx0>#l{RIthXSx -+MrI;)iD+D&kb(rUSgj~@(y(nCLw=C%%CE(VcO*&HtJUgd7n2xgKg(+S&NOnp@=NSS3X|qHsjlVais6 -Z($p8Oym5SOng>ZvY@wY^5}5iv+9QNXp{hF@ZC -K>vN$&W!N+*6Hx(z3L*-KU^>)u;uZv9SwZo1RzxMVxvEa-XZeJ!QSb>eCmUVF2c7-oj)-A9=^CiZ|#hYqCNC+0(M><7N6^9lQ@QAIZJinBL+ -CMqV(p($nMn03N9%kGuS*_RFKl$GMdc5;wbjcz%2-~+ke6(RC0iX<_)a`o=?Y#&Ok*Dk8=u5O73wbojdw -(8RjENzp)%8psinZk=z(aK7R>OjFdRpaYsgcaoWiIL(bb7ve5L4jk9vINo1(6*x@$Tw|$x7@BnQbVw7 -5n|w+h55#1iLOoDskY#5V7dHF+uKIQK}2I!s1o~p|^iJRKc##R=4Pu(c;K8>&h_#Bmd=+QXa+c6dZK1dlK>=O-6*S^9b8Z5ZQFi -=QZyqyCbtCX$D;W^IcR>7#HaU#awK|Qz66Lw9Q_CWdV?bWdZ(s2Cf=o*b`dd&XihFUgfgd{eU<(zsp{@z9edn^dZqTTWP*yJJkq@b?L(LN -k&tXToV6!9_}Wt(fIt<-!=VE+3N;)KPoXL% -0Bw9bD7?SbJ*~A(cD4rB_;)yZV^u&j@dPNrf$<7If8$Fn=Zw87K&j0k4`Efrt8Ef&Ez^mjn0Sr4CVkCgK*V$}iB8 -EP-MDa!G_OJWlwoQVA&(PHzUk5z(HTPLPYD`Zs8Vt66voqxU@~;}!cJ7}B<5hH4IizaNf`;`k9(_0N@7z04D$d0B~t=F -JEbHbY*gGVQepLZ)9a`b1!pcY-M9~X>V>{aB^j4b1rasbyZ7m+b|5i`&SU|p)JsQFYK@Y9WeB9T!XE{ -4#Qw5Hr;5EB~O&o7TAv;CE0Q0COO23#K%W|AL+)L7Rs{eU2m~0LAIT-4usam2`7!FENi??C%dAQ+M(4 -#6|JdJQPCLN3RhM}J8RVFd>2ly<6)L%C-888bH9LxwItApHNo=5t&N5c#k;qsBcQ$uaX3 -2E!ih==A|#IE46G11-4?B5d&XlxOh!c(9`GF%Rv!J*VuKcSg|x_kKY5;q=dG?pN4*jb3EY~43SCPMgg -~P+bdz@kW1h4!qL%tG1YsA3Ku##dN6<)x3}H+Yew1aXi934*zrsnlW)JgJz6ia+I=`5o^LNR)Edvo;_ -`X5f@%@_!wp>Xq?92XzeAI@ywfMJ}7OCt)?+1(D5#|Zp7=7xvno9LGa%Hd&5o&BiuN-WJ>e18CiVb&) -H3KNBHFr;`ArMIyC?Z*-C3spRx7WfDXsihYyM`247f~5Pk>NnSr#gmmYvDL`CCRf9$!`V2!{<)PN_xB -ld69vNWqH9VB|}=H>c&7Fnogm|q8ULTF#555B#aAWuB4W(EYEhmV<|LKC-*zDfR!<7em5g78Zq;CTpL -ctDT#}8VBjBoH5%E=KPIPy-(!&vb3Ire2z(BgpTpn9q?bIu<{mML;@}*CS^eR|BHMW^I$=dSYtPSt_b -?R9n)U1p8g+0sFh{5fcai+ZG_br~!pBeV&x1%hw5@c0+K1^M2Q|**Fn`4y&Ai87W?ha|07vlU7x05(h -QT3$6oHOgfdP`~ZN<<;wGodDd%2^D-x8+n`1P<-DR=p|!TB;S_&hEQ@^c#NgqyfY$cGE!E%uc -+3As6f2Y6GYf<3PLaXzwSpo?U)T_w6bWH=k@oyY7vb)CH#uV#8ojT5rjHu;~L9xe3nMfM+1O9KQH000 -080LuVbTPB)h_Z|fR0Fw{^03`qb0B~t=FJEbHbY*gGVQepLZ)9a`b1!sZa%W|9UvPPJXm4&VaCyB~L2 -v6e48G@AaOq(Q;3r`3ZI=KU&_f68vK3ozMd7E|a-vn13`uU5y}kcFQgUs}cKUjn4~Z?46iIy_Nmtg?k -mptFTPyM$WZf9+fop9%_tI#$T6Jsgg7+24H?m>*i_oR9tl`BoKT3UMo*&(-{sTXbWu^AQYOYx8rE;t= -b%VwaQb~Ww$9hNMbJsR>FUkyFgmvfx>Q^TBCfBAk{<_^-y5iN6Hn!$U{v-0Sk=1HdDDE6OXmPLmi$58 -?TLE6{^*a3D>4G<|Ron{){v@EWrV-Xld^8nsh)NLlZmhsxFj{~_>0lId5e!wvqIAp;zShojr7|eZSu4 -XaI^he}Zyj#(+$rN7p2HpuiwV2kFMom;@ziq2n2j-!vE)dR8amgXQJb*?##8=C -%w+cV#(u!gsHpVsNFXFhHrL&+|>M^-Rw{H-M+UrlkFw6;R#G!~K>x(DbF>f -YstZ|q#XG7IBT7kOAKI8yw0@bpg|%n!L74)1b78ZTPLKaqgK5YtR}l)7V?=`kE0CuSlY;4{EtAbUI(N -2!Y+>afJKznpPQXU6BZJSFP&VH2arf!z^OhyX9!^*WD1|5fB0vd!^e6ZsgO3@V@+UZ-grA~7GckR -meQ07x(|`(yAX*?eABS#$Z`rA2kKg{sv>h>w%8nHN*w-*$fdUv7-)a|3^RgD58yh9WpU_JFC7)iXjLf -im5OC6$IPmr`W(WMwtffoirVhMqD$pnqm^2{65iQ0|)smV4M{BDh^Gltk2Cvo-t6NWH0fv%w?GraL_D -K0a+n+95%NK~$!{^hF8_6iF9S;h|#ehw3pA8T&dsTod}{Io72bg)ae7CFvU8>#|fWYg- -c3m}OYo#15XWL>Yok!I2>u4+*4!-gRimHJ*7>r4dXhD?N3f(NzSNq!090~0cYhyD;Ob014;Y#3BxG3S -?MMs+Voxlcktw7b|vYVqpZ>XK~5HnYv_NqH8lAbS9YGW) -snL%9+M86OY5zK$n@oboyE%&`4=03LQ?@u#~!L{RIR5y8jYF{?2|~hy3AM)*&0SI7$f~Ww$&WT5oVW; -4l=-fd~T>d)O}n@{i~&0dg2Z?(gp5$B#cj80v|GCs908ffkDTVkyfoSMwvMftU?v`9M@9B^qjC9>So3 -U8kf88>_GtmB7)az>ZuS?2vK6(qJL2^V?;>_lO?&xm>_+;GVt=x>I-iKrPhPW>%Wd*s@|fauWTSm`N> -QHFDq@QAnIiFyFilL(NeZ>FBV~Ki;ANM^o=kJMx`vp2?lNvLAn!#ru;#y5^L&V|Rb^!mBvXJ6;PKktQ_Fa6I@6aWAK2ms3fSzDBMO32y+0 -03wU0012T003}la4%nJZggdGZeeUMZEs{{Y;!MnXk}$=E^v9pR$Z&yHV}RHuMqnZFKkQ*EtG`>`jEhV -DVH>~50}Mg?eVS}TQZWo+5G#Sk^CLsq%DOQF8IpxbT5+B+e|=POYyDpiw^x)iN7H&qh -86a_gayb=CN;wG&!ofECJ@G=!2CR#Rx7wN@S-bgegwbGX?+3j{oQd!f8Jg++6S;}*vn$}p49NKv4mC> -%ZDNJ2c5l%9BE(XH~Pgb65>aAMUX8DbDX>8WY;*-2^Dlac?aQG-M-!kpIsz=*25qKW-P1dUXlNopq0s -Lf{uTB0lA^cX;Z>|bY14BYd1Nfqc6%O^L8&@z7@Y2O)(5VY?Td0;*FRp9xS~5GMV!ep($M%*l(dthnM~67qNbaW$ptlb6Kez -ZSs;*Chf&Ny*lkhAeV6^OI)FK0{yZMG#}#Oxl?tSV~P$F_&f~Nn;EoapgsaDz+yTC+)u1u$mf-@jr5+ -%d<+?y5LmC3^G|R-P55@KN+?AKwT09#AU>`{vPvCBg392@>Vq?Qa~XGgq$m1E%am(_JNPh`;|Qf!5N8 -#;uH~5n*plvfqjWzoY<7(0LS3g&JVy4pvTO61P+y1=_X66`FN+b=kRQ5l&Pw;IUDZ){0+sVC*HqA|#h6Qm=$*C39>qf?^fP3~ -N-&MZwFkb)$mVNaXYGAn`aKz%?mksv%oJ#1SfslK%iB{L<}5U1YyMUX;dgJ^u=RN)wxmA+PG`egVcv4 -4EFUT3_~}BmU(Iqj$i0D+GFJI7M^UOGeB;*mQ*i&@#1GB`@2k%7n14$L6S*rr{3{)Z&S4o$X7i|XgqV -T0h2?>3uu-mdfS~xww~WBj%MrnB>ONi}48_Ti7FD&iENE8Dv^+wq4OVy|T|^M(oW&q-MwOG;F)H_#v_ -sXvrb$}i3Y;RU6g5^ljgVJ@|F5#kHX)@dg+>q{?#dZ$Q6vBRplr~HpRg-t -GUZC;{xaAVap@(015ir?1QY-O00;of09jjO5aR(c0000n0000X0001RX>c!JX>N37a&BR4FK%UYcW-i -QFJE72ZfSI1UoLQYQ&LiL&d)1J%*-oRC@3vT&8bw#OD!qSFUr;hauQ2YQxr-|GIKIZGEc!JX>N37a&BR4FK%UYcW-iQFJX0bXfAMh#XM_s+s -KjM^(zp#u9n=RVLRT9W7RRWUiq=Ab+kU)n^elBfFMcr_ -cWNM=^5*uFcf`uy+6rUDMR7Y+r -PONqmY`sS$bdGjYTE9QfRoIilmDOET6LjU1mTRyy-u!X4I4&)C(%@MLzBz^4!w3w;5j_d{>L?m=8Z+A4XcV>^Dy -2&Wtu$bsYXiP$%FFCPHj4KJPN^B6LhrN}NpeLyf#wmaUqL_clU{OG8mtt3mkty!M8z{NbDA07St>>sa -B&tBwDPv1QG@#V9pFzz{M3^4boYyj=##H!KE@_v3y2m@xv2?(S7A`qz^vv>5qytX&{Q#i83m -iRWPwxK3P*UAHPO>b!ej!n4_TIeG!Y2}~=q%%t>}+TnL$U}N$0TfFh9O~$aF*X+nM?#viaXfSmTn$>T -@uBY@JaUS0IX?0k7(8dU2W@fi`WP-_hr#E2c=)&cit-szMmlL+iq|+oWn*zw{zNRWD81d)$}8CABSfa -sQ3I|!L2*Vn_=&w@{OQ1tT!EzN5}~@u1Dl7EY}BL)xe1Xe5t`IOWtm4u+}2sEl{?&Eb69MHPVq=APSM -cEFXNI{V}_rS(>ep)=XQVikrpwl*lO6!(iakI4_7j^n(BZ}@?Rn)y3Rd7I0cyEE}r ->I&(L<84g0eJeF6achv*{ULm=8XFw4Ny?^{dpvDL&x#7eTC`EfAdGf3v-U-KQEY(oFbp1p#1##1H -v%AII6)w;7)>>ExDJ?iI}Eo0rq=eW8iui!^b}?%J=e}(4Pck#tZ6M7;{#qAE=ZrF=5;4}7)k(MJ$d_6 -rZ@hX_7oIicPoAY7#R7lw4U0cA@yK|0x}r1?G4+3K?una7z+o?w(s}i_};w@7|mgo0Uz%DRzRr1@2RQ -xRt;a>d;I86e}4S%kKaLtu)hCbRXuq0_+j%icmNL?>VhQ5O!kJ -i+vOAZ4(T>vJG84&d!pfkYkcVj1;`Wv%*IyZA!dkJ$P1%->c<1NVY{xF-n&)=o5{_*ywU(QZ{d3JX4% -Tq{4(sP)T4*mM>pHqh=0)@`x(^)sPGv6QYREniPqOa<9=}dY0>~CkkzI-_v1%Zh>?A7V3XZ{#Ka|!?K -&ACfgKrNG$!$QO~MT2`dO)d^@7iZ4Yp6778tM@Y?8=79h_O5t|&lj(o1^%{jCceI$eSJ~6PDnyA7T1Z -fF9ejKIu-u&`pRO{!X;kC!$ZBqFRcSZOdz0~__MR8 -MoV|KwdNlO(_@1v}5va|@*J4!(S$00YE$7U(;s8)*U(a;y@xbRLspT+B-8hm!640q(J7ANH)oOqA)_!7K>s1f*O|m0T5_gn)wWjoJweiqcaP8=}s+5gRgA -CU15f6pfK@H*8!g_KZEdTnzvIcLH1lrctcQAri!zu33|PVCB$Lj0p570Eec*$R&;O4fh62ag?CBNCigR1N(!l>}D;`;fm#Cj=T|QZ7?R!Aq -baHbAp;z4ATD{+Pu8(XF6^fGn-1zo;ioz&QgR4M5S1IG>2kDtQ^N&hW?CnJ-#ZE7!aGKCvS3Ebi5pXl -m^b8^x$zW{}tO>DBJI3(XU7o@y~2;hJRU76=2y@ylrBon*D=cOQQL==<-Flc&7>rl)Mn -&OoxEa?`?!*vR(evgm5kqFO|tf>}e|16#s&`~HAj^h!_qql_Hk=EqG^W&wLVF}Ojjj{+muZ*k(gbMH3 -R737ZETM1uA74w-o;J`2`B8M7395R-HNofM2A_w=GfIwgk-fgBfbB#O@7Fn_1Lyk4Eo|3oc0Pu|W&-u -_+5hq~*g0mP=PRfhga`qdqTUty6M83{WEeOz3;uzShXkF+A_h>Dq_7%p2Qyc+cqpmYj5}~2Di&ki96E -w&Le867G+*V>@CyO(EP^B2`l^MKpf(_XK6VM%V0HMkh_Br@9_5lxY#BtKpqJIZUcCKxhf!C_!0> -kitT%5hT#4uc4cH`o3UiwlggU!sMQ+=^X5lRM{=K^a!7`uL?$EsFISxE_Pw0HfQm0fv;yvs7gAw -Xc0%6W1Ci4bIv7#x6L@&Ip)cA++J|>a95~N9lZ0<3YvSPK3LEOWbjfY~`QLZAh63DVU>KZ^J=F6v-6`K*NeoB}Q73wZQ@rDJ~ez$3qp4A`1S%1qKHBzz2xmT -arzAbS&i-G`Q{$h#?;pmV2SUs_iu`8*Fwc`mo79Dy! -0Dh+fVM2g6cDW5DQr)m0H?g;HjkB9DLU?@f^l1&Y{Ag<~i%w%pI<_-jlxXUeBPmE@0{1qkcLmTo`6QtLDnJ>9Hi+R^bc)w2*80^zD -kC@6ya`YTY37MsDLBEK9tOx;~$d!Y%5&iqK2riN-FRrDyY`zrA|hQ(ZM{x;z@H=9Pog7g9@_6)8!T*s -lE=(#VPmK;GziWBzP1=-V#Sy?HSF+vnzu#zeYI=s3rN6S3UV8s}t0XVLi4KLa3j -HK9##25?=8-BC5()tkDdb8di>{QCN3vMc&>tB=*JwSq^P)ru2DwL*ha9YDBi*)FgFUP_PS#uU5akWl+ -svB91(OJAh}u4G?PJIN2iPV1V9GYc{F63OlwKPi%_{6hYV%9`s0&lNEF@ctY5*qB@ueD!> -VWBfua|YTbPtj1R2avKbI3)rO~7M}e1vzT|<-{FS{jgqO>r=~`pPv_LY$5ao8@nSp|1l$Sw=axej4=J -GpY_z)ys2C?!j>>*BGM%%j3+|9F27W314=^P7R_w4G_HOiG8e47(D3t50em<2P?H?ZLJ06$@|FjJ;b|-!h@kN)(FO~!NYQ5j9=U -VMZh{Q$(noU5k_z!`G;0Az|!R1{qyj+wPjasb~F{HX*-sigJJXNTYCwXn_tP+(TTiXj_{ffbthdUxbs -}ibEpBS%MOu9NZi($aLNY -R8AxT*$f3nMlu~*MX5{zHi14^rfHFj%J&4Ikg&f)(x(U)NV_kXlQNGV*E2^@cj*3t^|X$XIx?xa9Xy) -sF+8lOp0L#=JD1?D74}fDJ^iw)-D&clxz=eUop#(8QAlEcTk5!P9(aNXdv|tSSPuHtPPfJ>Ya-d0I@$ -vTijSFq|RwV22PRYr+KA%c`y>yO~)ai)C=}uuWU0bU0TN7C=M1$RpW639Ugs0#bEjMKGrv6jsyr*0hveyZ#Regh5a%#bI{M)V;*l^|op?+IcH+SOax*ufc8pxY$hA` -(qTem>qK=Kw>(W$Yai)1zQK0c6DG5<{g&pyvwk4@g1&bIAlWH+%Bx9h~+o~ZfJLkkSN(yTLs-pSeLf% -!X?g?J7G!!uaY0(l@sRVpua=$I-hkA->>fHmYS@!c!~P*<9;*t2O-2(r-d#>iprEg{Zt$2K{h5D*=1e -7PYccRV(_lbrAtvgzJuzaLnU-8g&$WKx!t`)58{evFjM%9pM#gS@4Z?5Xo2YE`3rf~poPei>AmA!`^q -)QI4fNr)Smz{o+KWm`Uw8mcG6`3fr=9Z1NhBMt8nBd=_R+BH@w74xYea@M=2DIdvk|l -oWG9M}PIW5Z?7O4mGLFHb{Q7q~KI9ce2-m -qS$uljIApTMi*dphRQubY9q#WFetSymWswkEerwGcTNh&ZfJY2}HXDuhJHq7VTjJUx#IIg-L7!;`)Tt -p=;T<@^G74nchdrvAJ853sSE-jw#Y+p$e8ddSpp#w^u?sQVe1Am`^We_WG;>AC^OqqADXY2G=boPkWM+yDYagxqWbGkh -tW{qHeAG$sh&?mQ3<=Tb%D_@{>X44AJiMkI#rA=RyHC^J9w{4q8vJ8Kfb+Ue`cZ;(Mi~_S5?;^rxdny4jINn6kd^|AUw8N@{{llOU=T*g$S_Ug>AYY#Qs5Olo+L8dEcE^wr)d59I+=hW|~&swq6B$kXa4T{f1L-YPe -gD?f7TxA!hRwSSbN9=O_1==`xZV<4~+DOLP<1yA>Mwdo||AY`j`gv};gz2s)en9br!pt$W+LIcWEJXW -JjWoyJUXEnoU1B-M+WH~5E0>-KddI=N0;iR)dn^mZFRC)L|u#?xy3(HJlN--~fSo{&K5Du$i~W8UMEUqeu`GtR^h+2W!z_J^8mdbY`W+mn-G`1ym(x1)MIK#3 -m0lqbgZ}|3xbI6r^hGHB1|_!cVn!1FP>%J?k=N_uR11EmA?gVUbCMQ)>HYMLi2wGD6e#yw*j7BbLDL9 -3OP|X(nIj?tTl{}eO9KQH000080LuVbTY+Npkb(vP0R9pH03HAU0B~t=FJEbHbY*gGVQepMWpsCMa%( -SRVPj}zE^vA6SWS=HHW0o0R}jvja;wz>{kXVD(I#lx1&TIkQuMG`2$VEhF_A=-q}MUfAKy15^|kh*pU -3*Jw#DIa=FOWo^lrD?@87tJR#oA^xI@ -sn*1PPJigsoR!HZ<)!W936Z|JsIh=8LfIpA6%=7(I7tQWH0Y2T2<1KQyq;qJQ(`zK-Sr -;(-pPT?w`pHz-<+@0xYJNZN?DWtcFv2vW~iOQ?e0w$?BafQOewPL(WF2e_ZO}lvm22BHEntflQ~Y<>G -{OLqm&w>1Oxzk=QYnk$TI#fXB^dQ&BCH&3R+)VMt6&U_Yg@6IWk}LhE)P&jYV|aZezVGn0KPwzHRQDq -hJJTXDqKk7&V1+QTqlqopd?=VXjFHt|PIJiGl-_wYOay$cb5?kcer=BTTW#hsA-I-H}gkL8`ZRNkwhmr;-etPpQXER$25=U(| -Ws7(5M%FaZD!EA4-F|i2H-gC^3M^MK!(*e$289zA-h|WQU`fyer_H%D6iynHNKC{LojiGWPyta;L`@+ -%p^Z2i`Jj9ME^%ULWsTr8XN+j$ME`@>u#?UQ=bx|)q{6 -Bm2N2SVCOFF~u7|GuK{i6z#@<{355<+Dj$ZwfUw-}}jTF*{`^LqrLIZ(FY#h>s -mVgD`SEY}At7U~00q5(}>l-|R3Db%31XVgJ8`&${u|s>51KguMnKRU~*MK>8G#olVP5oi?LGm1f-XS9-srcAC -+|7ss@QCtzTJWPdIS~K%M?87Psu}e;C>e<2Tk}D-fCIC|3kfopz9$*10$GCoNMMFyH1JhMdOpm@#YaF -h5bR~NUhn=yY%^Y+Nr*6*oymxVKinp(Z95;^Ba7*f-EK~L0WG63JwHO2pPusT{oYm&)USg;Z<%7dHB-S9QGsZ0N;*2}zoMBGxL)%{< -M(EEq?E-;Y*%cLr?;bo2uG!P>oUQ)mtn~TG77ARF_qm^b8vVlxq=}vrAcy^!R@=&F%j0CewIn;Mp`C66eyYRW8*e^~pai8-Ir31%%FffY&dC@<>qDA^o|UU_8B{An -~3CD?two^8ad(VH841Ah@fc_V1FL!k?4$LGLzy08mQ<1QY-O00;of09jjik6iFU4*&p)F#rG|0001RX ->c!JX>N37a&BR4FK%UYcW-iQFJy0bZftL1WG--dom%U2+qe<`uD=3RCR3`V=_Gx8&{Q|gCAo{Id5!Js -eaLt?5D7`BDUwH!wv{yhdv_NP0wksQs+l+v2`sR?zkLGp`TTWtUltY5*+v#Zv3uF>Slv|HMkvK(nahl -~Ril#md_J3PnrhF|bknt6Bhr+~eO)yz<7HX3yp>g{W;63l?71wGJy79WVILONb}P%R{kfNW(H?4{?1M -__C9g@rBcC}tl-dtWuTud1@teCwS~5Y@U!0NLWfY4Y&t;>6fOmbZ5 -3JNP~|wA33&GM?>(9el+%;wLbxSNJeyQl&UEr7@?70FV{%V9C&QV6APc;$_Zj7#*`uvcA&8PEqPw*#q -zGo0qTOTwcYj;ZlioQ#Gj~Lx3j+Q)#yA%3C!>l1L?cshl|ifjw_iCBF!ykn^lfdDreVmjI}ksk)G@cx -Y!cG=0&qmGuD$;NcxTi_)~@dvGU<*=&}J4NDO!XtJr>7OoNDucb4e1v~oz98}Bx%RW@4Sk4%H+UB6Y; -HMGbrb-+`Q%Rm{K8t6h0vljHuMe$IH(C!MN8AG1R8eJYwIY9!V7vrkRR -i%2xPWk1_Q(M6Bh+!4+Ew>x3>ssF5|g9gYl^WKoGh3R-|hjQGtE94NuKJlQ&twizs1aMH4uSY310Zzy -XT_?qTm&NHuJCjh;JY07y5nX&c4_^5jS$tc}1HudUE0oA+KPOioEn$Tacv`YG&uXfT!%zG1I+5pW`m*bPP`L$qKFpk2j^NZKK%5 -tUxWbYQo0rUiNhXx)=tXmJ98oJI|8`j#OtQ>=J|7A-_hTmzhHl2&!4^c>xs2J621(lw@9ybFe9@G!kg -yO$rQN=sEj5K))o=$0cQvyeoVxGJ5T^9sL!0^f!1*8b2Y$g1*16LmKX1(D!7i?D2`?G#kai -sDExYcnkMOTF(HnLxq7;J3N38T%GnM{FZZYRu5sGz)nvsWg7zLp?YB{AVT6)Gx~)V$p#Yp%k5*d0S2t -@`ceJ=JL98ErmhFpqad&)f-~J|X%4lc5;mVeT`7WX#&kQX|JpVO8N{si6F++e0XfJs>8iijw#G|41u! --zO#rmiY~gpJ1QHx`kJ4kIomr+(d8t$H59}?12*C-c5&8X9$DiiI^iiNHV+xV+5a{1y; -9A=7Jafvbt^rmbc)3ODe6Li;3BWLnQ@1e9g04kfcsbU@(;tK4Q@CfT7<3^*n@%Cm+fV+;0#$l1I`rt% -0|o6aa8hRkw_{w9s)dW6BPL*P?|)hi#!cRoYRGVP?E$n8;atdgFSeO_}PH#($Tw$E^1iH(*JUg68ZPSBgUjF0`n2pPT) -w`hH)W6utB?ERh%3=9X1(-(?hg|#A>+y(%INmF4EZ3Ce05aG1mp6Dl}v~{N -qblLY;@C^n)KNF{ilzM~JW&(^W*RLT%H*9b~F1S|JKc^RH@{q_#*B^+KC^lry>#8b70UaOGkOcRD4jc -7Dh(ZO?n{XrSAIyzk^o!m%pdFo75WSdMPfN`Ot=|`P#PS4X&=aU7!ikG;peMS7Rx*R!@}L``-}elM3x -ZZ&3zP^jb^vdS3nCHXlcx9WK5+YAL{lZ!4NdMt32_@eJF(HW>BRW(CydYVZm7T{K1T7cJ#he -Ef(BsNi7j@$>O&T;)@agK=!Ha2CXd1gZDnK=w7{_gOP(>uLf-Nuv+sr+YJOyW(~4(H8L9b4OS@`cWGH -KL^rJMYBM-y`a7j>@aN{{Ccr7gEeNMll$l`QpLZ2yD6XE`LztJfaXTsiDTX!Or7O)gM;;NKY3yPYI&% -_ev55-Wnq^}V%sfa@Xn!5EXqt>?jVTmSMFEiQ1!fu-p_E8K16hO?YkB6HE%Hei=w-_~%=sO?Knxfn3| -EOxelUU|g`O){uaM(goou!l)_f6qk?B8`EF-&*_`4aGmWLNS>8PQq!SnTko!_tL3+#3K*{wM6GM=ytU4}Xn@spd|5}wBRzNET^b+sFa<%% -uglB1Bf)Iz2Af|F1Pidpc&w5#rUQ^FdAT5~|dT3{x}QylU>4neXKSV_+VO{rFp6)>@c~jN3@X}%Yge)d~9&6LeVQfb8d7Q5?`N>F2HldvczPAHNKcAmc+ -DOw@HiQ~09@g)$mrnFloj+AEGi$M@`C=ZA==m+(BPO?a@7xQpJ2z5D*B~uBY{HiAjWH7nS+&3sC@D48H;2hlGxcQ+i^8P5Xv)1q1FN{-JZH&mPxL -SxnpeJfDb5DSL0zGh!}@JjW9v4qC=#v6c0539f*0yhGUa(-I2A}$`T=zjA&vCoZ8tdU$lv)on}=ous- -gO>yu9}FRw3NzDr+y{BZr|!}W+LNYAG`*5CjNu5=nJB^derOv!G3nozI(kk#2}^kZ_8>9y?P<=_N79> -tBAeY_f;W7-eq39{@wAN~Yf2wJujJ{a3lwz2Rx8!Wl?y3p@+ev6Di0w8?mqMJQMdz& -Au;WKROB>xhLR=EPM_jXx@X(br}kM(-p;m+1Yu>)1Bl-4l&eqCDd#>V(}dG5A0be6%7+w*f^fN&zkayi06mXBhwxmMyp1GRrA83^HI^%eSaT{jcYg7?s!YVGth$ -FgYYG9tLVhbm624Y??=ESp~^qDhW?9Ze`nDra0U|UYm+iBoJLe^TqS6^_j~9UF?c!!(#@!ZUGD(XrY? -2b=XtJ-J1$L&81)QK!tEg@@N{>(!y1nAgq}M*bhkY<l@^P}ikI9Q_l>9aooT((63L{%^)YIYD^ -DLbJNESr6Bpu4(}7MI%scPISp)tR$p6;5082n>!A#s<(P7NOpN>n&F7lwWiaL-W6>#RDqlnv^FB1d@U -c)3Xxmb5o>C4rvq}yW_QO>^2hL*6j(V9V3VOr;OENjp5?PuVCsQ4NSC%6lQEUyJ_0WVz)TTLF*VWy{< -)pv~IkYdn%Qg|^9o3-mIx0XiIB*gr|%S+Xv1Rb?9OAW3-?Sp%5ej9H%I;r*8NSnoK{_~q_Euj@%|1lN -o>g%_~^qMA)nsID3xTfZ8DzAlga;D_7Xw4oAO=?RLE3E`I(+x*W6 -ut*up0nC;jaBmwJ2~a_`XkR4JqnI5B!=AUD~5LV{jIlxYv=pZh;@WWV$Y@_4G_II|vXUX>c0dzE+_GN -3f44)DW|^=XDoDcd9oB3$>BkW*ly9+@lL}olViHT{P7`JDM5+m>bhi+boZ}65LgaFwqx5WaV#h+(YsV?m{$!^>f0YkM)F -ISN3p~+O+7!2RS6spj)jdPMN_*g(kVrOGrf5Ff{X8MNjDRvRa;>=XlQt6y7BJ>I-DnJgJ)PxhT(mxjM -%==f8^eId1&?=T7EAozoRQi`l+T^qj4}5DJ!1A3S6^-$Vq+C2i(ZkFUNSsn5Ti`BH$4x~?E=psU)Lqh -K~%=5(M7NzzEJ{H~G^eXFMF0F}8%$+daF_UV4&`d}_v?5ox~j2+OUJfb`U;on2q8&nz~3n*)#2yWPNv -WlC&q3Ppi5K5eY>^(`6QNPZx6FX?ghmA{1Z3^=f2MK2M(q4h@PY`;A9>F&h#mM~kZXLSe!zM`3`S8-h -s3+Wqd>clLe_!G4t%*|2yM-KuRjsLLr*C#obU#hct-G4 -nuj2PBw5Coahsg2}9n=c=A5cpJ1QY-O00;of09jj8A5$OQ2><}z9RL6$0001RX>c!JX>N37a&BR4FK% -UYcW-iQFKl6Yd0%&EWo2wGaCwzkUvJwu5`WL9AhZuF@5w^4kG}Qda%pl6uGbW4iUn!}0xeNC8(PwdQW -Ixyzx&M$DUp)pB=tiqi}R1encvJXK@fc4&wKXu`V%X^i<-T^yZaOb!O2NpmphgudE2zLND?M@RarNT7 -e(3dMwW#>IY|<(6s#@TDnQ0tUTnmz_-`w8^QF*jg)AZqA{Jy)Hwns29=AfMq^(tOZ?&wY7HQLIFuxKd -@+Rtj)N-?J3~^d21+UP6n6J~0>tCV}HE+tgqbeHl)F`>`2<)33FZf1Ssz&UpTq@DgeB?Wk-3k1`>-`n -1Sh{^Bi%my$UhK{5XF1^hC(pjh3b`U?AM6GzS;fhBQDkKuS3G^-8xR+_ja0f`oh(os8cvZ -){G82yN;^yPs<;T0^;{BJ8|4u$%f4z*@t9#E%A -OcTP#kFQH1Ydahl!^+io-Ixod@5i7#!TdFA}B+mDc}qiqce#ZEH4;%A!2ru3x9hvHvj#=okk@j)UAL# -Di~+JA6dEnhLLZAY%ADC!f`V)hJCsH5T6kH-BxPWf==2exCJ)Kx`1F!$;v|QSuUD%OHy#-Bt55}B#s1 -O4scmuEO&7bv2U$zm?+A2vt{6GVzG9C=(RvDFu6^^nK6mE&Y5mOS8oEZ1p`=Mz0;$YPK0E@j_7!;NQ4 -~xew59YAtFHn-2fh;m=OHlyL`sDX3x3tD}H`rS73B4cs3_DAd+EL>OKjLdkP^kjoaRq@Qw75CqoPI5Y -b=01`b|i01gV&;YP$N&um_=QhW+^b;q1&f|h&<~g6Cd_WKgD7h)hT6k*Db4Y{^nwZ5VC -Y~ffABiO?9s{_ILv9t{Auco8CtL3s=2L-3?u@EUyiMnC_-L{lh0ppV8o9j>UG(8x$%j5HcD$-YhS9-$ -cf0EG)}U4VSq{P$#r&*|H;Trb&f`|XDoCQ}(MBu8WNx193CG7{lBD#Zht5J(TGU{!MVMD(HEU~v|3Wq -r@qjy5Q>i2xfRiLak|YWHQ_w=?ui!?+2+I-;IJ3Z$gh)C>b&^O3^RX<1i~cRdF)&!UG5$2^f={B>OAD -qm7Z@W@4Uy6*?uWV{xLpEtsD+Pw~<1$yrB^@V7gXrbMxU>|df2$g0?GVi -jDNn3#VclHh(eVxBu8mv&Vu#65jr6JkOlL%6 -GKvl!Fh^hX-;vcu|q20V1t=E+WxR!tZu;|u`p8xKpc$smTpD*kU-}Go}s0G0H9THA3d%)Fo+(EgALP5O_j%h;-IRt8m>!!ZFib_isOTfghOx}CHG-#g{00_1X=4VDP#{I~#fKhFQ$4G!c6t3s7MFmT))eUcVwu<2A9HI-W0<=a_6e -&a<Zo!DxYPGj=5&$s@dQ9(N}0_#SIbM~`YzhNSiZ5 --E%QZT8Pu`BA)UgYiJHoh(8U%KkPq -s8cf1ZS6dbjNti@X^bM>VrcEQJZG~P!pMXT2^)mckENv_y&d-tCnTbQjo>oRwP+}YspW$B>H_+d41(u -jvTg}*0jLCrai-LgdhKEOLbv(I_*$2$ZA4Gx#q?*efI4?l51a~`itmmT)MgS36}2qZuAciBcDi7n>DZ -vhx2A-s7k0;6)yyEaK@`*5vQ-%d@$ayozk34C7VM&}YbdHt=bU(K)E8nyCizUiF4!gP)9ExEt!^RL_C -3pVS{z0M9fR>e=X)-wp5up&R|qw( -8N3Z>uLC^t+aA_S3Uckw>l0(sf(W*Xt6LxGf(C1Ff=KHP}^IN$~kG6z>l?T-j={u9AIhcXlTK2q6OX# -Ko{MMoE%K>;~D7KCNXFr|B0G(bKJj(r*wf=IuoUNDBf@(eOH4=S4zDG0L8ttzQ37Tq4bEkzxPvq;s@j -4gAjE6WJdb~C^eH_@^?ul!?xj#!|0YM_YD7YED9b -^v7d2iN`a)NPG=!^NwXt>FF9DpHNf^e^L2RKnmczLM%9FDeS_M<~AU$gVfrY&jwbEDufqSA@&n7l^Pz -ix`HXx%L4LdGjT#Tely0-iL(A~xn34>>p8s#x^6q*g_NHrUxS?wCw?v=z -$M+q-*~hpuiNSkuD4HX4kNCewiXNuX|popcT?k}W@lnIQ+GF}Q~79aFD!w%^B9g?)X17(f>khL5Qx?a -(CQl%r)9PGlB*wdY0L}I$lOej%Rfo{i><-LfYSUIqtcVcj@HfEgrsG -Apq?>4f5Z&J^HKwt5@LF&5DDmm6LtveGT^Qyd&InfEfCJIU(%b&&C2HJzM}7@6?-MfQvW8g0OE$$TDh -D0%fEc{U9o`$D&xSHMT{SBHMk#&_E3G7sT{V4sxSBUBIw(?iANhA9{(`#~LEdiKfB|JiKki{fF)CJzG -w*80U^q@CD|kdIW58;Dw(!Uh%fChJ0ihf#%Fx;o+`^pnjX@)K2tHKKn2ccEuN0lx51`(?1u>HSwxobP -EQ%-vRwizh-sCR4OEQj$8139kAy;mWCeL&=eb?1cZ~FE@3pi8NYUo?legI&v)GrFwrT54^JEGF=omRzRTz>|dKGCx$g -h~!d6qg0gJtaxCB_(vsFiDQ>?RTaFn;1*)dA7xeqBhT|jqh*n8SQsv=vMNLvGP%LUU}Tz=G)av{rcG? -POoFPEiLy=0Dvio4YZ_!$f6I9s7qAc8@Vm-V`#s54E5P4AZsbOkyIj;`v-(?}@N$_I8~d!*h(P^RE%G -9ZoK@7W{#q4DA{Rl<3ngsZmKR7krS~0WNg^WJWxyAa-TZ5j3ZPh4OxPQq=P-W4Zjhj9B=kf;7RhhM`y -&Q_xBq>85uX3;;{2ES+t(9nI!kx@%Tk-OAetGfg?EU3!cztzs`Saz~`7aZOK))T(-)?WOyQn#B -pMkUi=1*!k*M)$C74bXWJz?*J%HbTvXCSDus3PQZ)euaHIZR5GEajtkDWkGK-yA^z#7;H0AWXiHM*};cEO%aJ;Xnj&yfq`cLthoxyMC~8Gf>WzC0)EF+*gt$DRXuu9$s`Wp; -6of9!uFFanlMsC=4n45m{xNLf+({P&|ZTUUba=tlynKfL7G{z-dv)Pi>Gqb^bheP^6-YId?jkFgADve -H+^Alkv;9Yld;Xt9*t1LL+8w@`T*GSlG^+*O!)?E4d}lc5`e2h!Vnl+R(T@sKujl$KCIfjBaL|nl%kI -N`}+}b%qqfo;c7o}B(QIL7~uejMtaP -EU?bPP<(|MPZkF3nFMb4uBs)M=0=4wtT0|MglVl`r0V!v%FmkB^YRQ=IAW;$P39A2`KCwv^KWT=@wN= -Ctjs7@E8utDg{8!=V<>+u~7kIufUuNP%bqP;2of_GYAw5D1-{uMGa$z0USg_rm;kBhyfHJ=lBZ1aEe7 -hNx+1(698G-#o*vLORLQSq~HLb@+_|sUf4A!3`Qq{;~9g}f&sYp@ScH$fLt69d8FYC)5yvUnd?!;LT} -69i})c9;OQmV7O1nxfJ%s;rcYV$B{+&W{d#vjr_EJ>l_bd71K%;Jak8s{BS^Zq5NouPj{@;Q;%-p*7C -d@b6}SeF*K()=7G245T`f@S9_727AAue@l1hOv9liMBr=LdW^YG&Buk&|TZ{J+Jy$ydodpAG(`SRihg -vU2*boK`d0Jy^fEEG70*y{cKet!8fq|f({vfj;8{&P9ghtSooT7r` -np3R5Z$P5yXnAT3115lB~Xz+nd-a_86N8XJ}(xiDbN`f(-{qGkCOh(c!l5xEzge9gXK3WTKH$Fk5Or* -~Lr+VFJeqb4QmuL5%lv?JTRkGb$D{;LjJf~2TIa!ULSe^T-fF*_O4QNf;z5G+|*>T?34Gk8oKd&u?(q -oD8sNsbU0L1>>2dTLYHSX62|2O$F5i;vVv1A&c8i0ee*P1#2ecLZNNIC|8W#uJ1V$1D6aX=dj;AT{>e -z$wrzcC>yFRcnI^+&K$_qT1DDMV5qxfaQaJJ1{i{tW%~vhZ+Jb+Zzl42r7K|coYdO3vAC5XWe~k_uB5J&+o -s9OkUsxzh^S_TWcZf`A=gKnJ1K&-a^eP-$u~Cch3@xPKFao!CqGwr~ms8c@Em4$7BB@ye!`n*Vs`Y&q -b-WpwoO0ELrvBjT=Pvr2t|uzeP5x$I--)TIHf5ML9Er0^BT!p?z#b-L_!jNDpfENj-GfYCs`jLz`)|& -)jY!#Q0_rbM`c4Pj@HxlV1NHGWX4SI0_8D%GkS~4D4&%hA|%Xzfo-V&qor#x;PScI`9LN_+_l%+;Mzc -u3b9Q3K6joesL$Zj-w5L7*h9p_w}Lcpk%7nwsm)W-&MTMQjPtA30F653p5C_Bw$QrAbpjP(VWf{Krq@Mt1uk`oN8D{RZe}Mtgx8On(K~-5T7_?h1Ealviya^~f5C$uhWeZBh -qgc8DY3pHT;HmV=+IlF-#S@ElxSU%Z9zKK%9K-Oc>!t)2$^ta9>ZNQPkOp8StL4yeZwG@wj9+Z)%M)P -ZP101WWD6mW*NkOk|~fns%%qai7g8*JAy0xrccJpv{jTnr-prFK};9e7$?f$sZ2Ga$k@fIqM$mk9(i3 -WCzKOm@aHg9ly1DT`fOv1&fagD;3fjR0y31w2<2GY8Ji?aQn8w-Z*%jR0t~ -}HKOx_5L}ovVaGe*j)@Y9676oFr@I$d0W37G%(XcUVKK7Ro;OAZeWFS=H0#6SnI3r(yf_U`nE%r -v`Z4<9bg^Kk$EPQU$1e_#e;WT!`s~BNWQb106NI5r$$keQIzW+eoCxM?cwh4}h};LU@d|^IQfDg*a~b0VG#H3Iq}f)PxEW_H1O -)t$_PoFp}@@$SEEt_WOW)P7&VlQB$4yB0AR8ot|HeF?KCYZM54U&%!$HK$M_WezQ4J67oNSw@Qc2_ -Ji9$>luR@GGwP)tM0nLtBR3fIKGN7aXu7{c>v8|-sMl^-=Kd4=h*wdcyg)O)-&%EsAC0NUku&@V;%bv -CpL#(>0)yfOk(&A5gc4-l7}HBd89s=enjsC1!+;@Ylz{lMOH9DMx-mB$zH@(KL=z)Vh#}(!BM2In4Kn -5;W7@m?XvUYI9ye3UD*(xev726_gr;nnWWW)BntUrgO?2Zuw32+x=$~dTP){y*2J)Y{kYSZ(NF)k?jX -i&MGP!%9Vj@`z@ITEx+vjIcC~oh0oTv?Cf&%}Un=tncwJ>Qt9yv5rISjN=9ETWtP;r)(3Q=jmsMF))b -~nI6h0{Q80QZawf>K&I@T+gp?5X^#n~S2z=sFBq+CUDhcBU+$wGB4a%~7SuEckg>3N^oC+2VH*m6Hwx -Py;q|0&-vC&6}VFH=x+~s$)t9$neAQ5B(v=4$mkK_5rnVU=#ZTgZUOlHkkTL@1q0d)0BM#2q2Oo)%w~ -YQqS0JvRE!9Y=7onNgHVf0Tn)x9ySqfkmV<+I#!!fOJxfb-k*ZE5aJ#HTzwB6tS}~|=H}eMy9L$FJiJ -^v0;dnW>i$mGg2#B>IUBlL?0XU)doB3qool=fZvg@Dv|Tj&u}f>lrASxh+MBYP2AJ2{GwG76^6-rzC( -N;w?GOLd@mlsBG5W4G#Xv6>3vb*X8zq?%BuILqZg7Q^*cH1+wPvI-K59hL18svr{*hI*VRH#)6zm)MB --1*BM^7jgJdR1QK72rbJJM4=d_a2ZyJD2eVpUOw7slg7HM2L`fkq8ID1amNO;s}->>6<7mKQN_s!n8r -Q(EfVOu9fe7ifV_?t_&=f1=WE9*m23Ktl;81Ezr+g9+0u8gkei94@-HFbI>g&<2)Vb2_=x?j1p$$k1S -6?aW-N!LI3cj9|g0F_RC*r%6}Vln+Bp(}v+byY!u7fEqeZox8|=`v1G{9)LnCFaUL_#K7U=YC-6L)&Y -@yYQ|MLMxjBCuLVDD_Ch_*-RHhNHJAN>99}%uJtkG$XqLyp0SXMdO( -Y2o^mY*3O!g*yfa9+4Sz`7G*3Et1;nnwj3kQdIGv|9-t=N>i`M2Fk -5PePpX!TcfI_33Q_%ExF0tMCliIDoVzM7=4OCB*(FFvyKqJhusz3p7fY_r02OecN3%=&A?Bx0Bk3arj -S;5)kKZO_iq@|WFJ;!@k#Oi0K-pAjgt0%q6Ambyq6#YO=(-!~^%oqY}Z~t9P`?!!*0;f0B&hh6|$+S} -^)bxTAshRRKuJ*V1*!#7pIYbv@M2#er3e9Hl|HLdm)3F0BIKz)|k2pir-%sEFr^ITCIsLj0%`_Oo)KM -AOleV4w1`FIVCPsD(imedtj2%04!9%iy8MMo$Y$n8V5rANeAMc2CGv;3T$ut&ESnQ-sDa_~My?=>&V6 -Rs3Z;FBWri;I>QLB?XVk4VMP)aOPRSEXeHF!&jqS)UAgNyb34Khls<96WR0ecG$vC6@0kk#O&&ccFE2 -fClGcSBF-!@xC-xDvX+Lp)m7O+(2uV4X!GsIWJIkV0ty2y=n6i0gtDn*<2lQpA)KnX_`FqT#-KDqysuZ|^{~c8qdMAQi^x-pVA=UOdC56;sLtTlfWL930R&J3(>GU1H9IDT&O2X -u9bRYX(|ts|9<;Hhddj;|oXgjSGQT^rYd);Nc3l_Becy(lWM8Zn9ie-0)EzTEou%djpT%G*SEIrjy8+h3!NI+G|D; -}ZRG4sD!NM&h6b+YIUw1P1)g47fod1P@1z~|TPO7-0JM}(E$t+i#;}VP(V^bLpYJ#GDcldbka6 -P&Uz9rTdDRDKHfujmS&uI{x835oqS72iY9O@x0(`m4eoROW<7I7V-Z?M{|2m4QauoKk@ -$*b(C>7gqFpT-SV52+zt=Q_Qh$uo?+Pu~>yyd>Owq{*fbm;d49M+o+n)QGO+6XlJ?UAV+JrcQ1LQH{s -8x6*UJ*u{E7r!AW)6319LUX5to$IqF`-S(@=m20*f3l;i>K2_~TOtm+(CrS+(k8>L+f0q_gjmdM0xe1 -*YD3cTgUI8E@JbQ&?(kWFzQusCN1b+#Q>}ndOAjvy(1|f-L{s7;{_~gOQ~MSPQ+yijy?H$g`FjM-e6k -H6^`uYZHZ=VjPJp#yK1^}sjowDmd@->PGal5eZC7;Yy)Ubqz}{@cj8v;PX5TaKs7}2Os -dgu`J+SQlmA(S>xXZ)NHV3FRu+eYd?9>}@3ie|7Ilth5y`caD9OF3RMNF?#X*ZT)1e99+w%4AWe1Ue; -FVL>mt`Z(YV;v0_HMf)XE5xlg!zMHJ-chj`Xw$sAHI?h6BsSMy0V9 -u+}WLK@6DJ>e2I*z)96Lf&O0pncwP!1%tod{zet34kwv2KBe{L4je=7B>bHo%?n$-jD5g3EqTk=fS_} -H(V{xIy!YLEbScc8}cBVn^+V*Pw)@EkCUZ8n)GSpSSx;NB`FnjkIS;>UKqZ+rl}R|r9YMev|EBgm*#oj!{R67yTg^t;m&o#wVjoUF)xLUs%hgnb1TqQ__&H^yA7Nhs#bdy;Vz2gB`I>Vz&Dv2@7X& -b4jU-(r)Z!$VGy+-_V=rh=lN9K$P=oL?Y?3}LKycdKs)|_~6=s>kF{?&ZoHufW7O`+d}rgN9Xh=M~~S -0UQp4Cw;{0-GN`!FRkNM3Op-Vt-s#y>RUzOb#1t8uPRx2?0uY<2auS+RetuIX`vFdHavb_LQ5 -A3B=uFegrE`%Paf#*~26S;JLUmLjDRg$eF9ZLzl5wN6A;MIm~b=(eUdNins~swB3>1Q?XJ;{~UEv>&I -^k2_F#D(H`bO#2TZujHhXxK?<5Q;-5m!=Mk1TD3fk01E)&Y$JW -6iz+^~rbv_eGLW&?2%+hr9w3wk8#H^F0YBCr~Li{Ty)z>!3X1ona%;WegRx*{aDt*E0C6V%@6aWA -K2ms3fSzBtT^GMJI008?B0015U003}la4%nJZggdGZeeUMZe?_LZ*prdb#!TLb1rasrB~Z-+eQ$5*H; -Xb2U~?C$nsr?3%Icx2Q8c;b`YRJAS}tDv=zl=c9&L^!2jNvU0y^=j+;~uGQG>0IdeaXwOT@`k_1C5Rc -eMJ2th70r7gV16t%EQcRDe@mdR{5eZVwQdZ_Vpj>ekdQbm}UQsf<)jI`jUgiI~!R3tfzt)<~y@2~ULQJsKN~SXcxLjgt4Tv;?oBR8}V2&a}ZD55OL1{UYWFc9Dh$4(2Sd)q?MMn|b -**uOd$P{MMI!*LvJPwaf#z*6GadHp@r>6&}y%LrT5`RNVz( -w52CMX`TmJ*3dvADH!dH`HzgqCt?DJoTzehdkzr3ggp%@~mlhvZ2Vea<>ZMQmB0!0fBcifY*bMX~Dz7 -Ll&)+^L~xkg3XZD;9KHhlD}z)ldV(w1v0&7pT_6OvtqFjt$yE13nA&w>7`+3dVA2SXEXQWuGdYbyNk2 -g%awAJu1by%6#Za2ot%a=CD3$q)HP!VQ>g=qcV*1#497QNTK0q0Dh5%OTx#A^sMGGoYOaQ7tshP!qLb -jx%y-RUCc<&TIGz7Z3weqz_HRBwMTFJyjHPilM!yjle7d#Qk+HK3}xfVxJ(Uu%09vy@`+Q-iCvKJ9;r -lbmw;25siN8`OiNC(3p7upVdv)h^6L8C-2@^T+K8 -qRfw|`9U-u!({V&mfz_yrC|y$%Dv^5&=2UWaTkhdA!04ROSV(y?r1cEm&KJYgQ-i(OE?~lY8_ -kI-*3Q+yWiRx9^q9YsV2dD1K+&iM;F>$C+>tq=<%^zDeH9#2MF5nlh987FSI+qvwj8|0c%>^2sAg;6P -#kdDb%kv?*%{f%9bn+r{O$L7sh*19@eed2X{X;)BwZ+D99JEqv<=Yg)0{I(e-2B^|b%~Im#(Bkx -d>2q+-8#GYT~P5i3DFAu6mH3{1O=`@=O!@EcSk1JRl?&37>m!aE$rxJO@G|PSz%*Sq=v($;+AA(kvGe -V_e9F;h?L0KZ`E*);lkw0mSE}5<|}?l1#HHjS9`3Dm$up_#}l(+dAeP$!Z6o0opm)S@%np82(1p -HV&@E9<7tIc`xDuOeanh3eQRdhZhixYLeHOVT{~uW8Pb1zowwti*K)XY<=Dyxo#*^ILAhnv{s#s-+;u-LJY;+b|JaY?50VlT>ZXjR|5%Q(AD{h -WM_DRS&mbvmMqF=|`lSf_^o6!fU#Ka<@J&_$Gl9HiH}^Wn$@Z)*h(a}# -Idl8<6z(-7Xk>8Q+;vkIf83;b(wlz)P)h>@6aWAK2ms3fSzFV*bqmP?001Tj0018V003}la4%nJZggd -GZeeUMZe?_LZ*prdcx`NQaAPiTd3{w)kJB&^z2{e0kyc8ithrYz2UZJ -eRB5&@#p=!&>=jYXtO*nZjEv%81;yASUl-$fy#7Byu;IWUS#3a1=uf3)pu9>Kk6^W#n{1-ak3H%6Olw -JKy%3)W&@53mvSQ;7VpsBuCLXBsXnuu>alk%-133FW-`RYdS<@n%!^2<$;yw>{U85%vi=I&vxbUP3h1 -G2{m95sJR%N&JH1O;eU&qR7(?hOB$u=>gowHkL4^cm<55Q55BpTkA*s+;%YtnasKx1x`cbDD1}QXHfv -*WOu~~y9+0fXTg2yNan5u();hI3hhC|cmcs2mdCZ06e;nd+%crj%tQWOfr|t#;GTu^wr^B21%#G%fm=?96#=ucZih0mh&I -JH`vcp?hyz7LgeTC-(}S9{-t@7(l-(3$0cY?P#s1Doy&!7D4nGUr&WtrgWXHyYFQJvIx9dZ%qU9&2-- -G#m$uT{m;|dQ99LevhM9-aMu_WGFx_84>e#?!(T?+a5ewqK1O%=gt;7@V=zv8$k~3l<7jS#6FHYlw&R -EGc0-^&;Kq`dqmV%GB6R*g<;KQK8mDK`KoQGz0n5yVl2?e1{sK@-0|XQR000O8%K%wh000000ssI200 -000Bme*aaA|NaUukZ1WpZv|Y%gzcWpZJ3X>V?GFJE72ZfSI1UoLQY0{~D<0|XQR000O8%K%whixY~tN -e2J`1{VMTApigXaA|NaUukZ1WpZv|Y%gzcWpZJ3X>V?GFJowBV{0yOd7W6>ZrnByec!Jj)(9d2iax-F -`_Nq4pt%%DkUqEzfh$qFtR+$*V3U#bFYuW!JIVZYb%#_Hw^6!4x%8738 -IXzxEyT`-NejOvm?rs~(3zF|DWftf#yy2dXqve#am7X?ov|qBc@?+P_q+8EdE9+z;TGf0Jzp{5+*yhv -p0W^oI7V`7jRnoV!yGd?>412#r+b6cbQVD)!IF73-NW!PdKZ?IBS^4?*MuJru+qr-J2&q(;N%Q@xR@h -va=`4SHOVnhGg+ZI>zmOAl$fWlCetovx45V|-c;Nl=1zWy7^md_UR*dV!wauqu92wNinMHW157oeVZ4kVGtSgV=68w8**Rsk -6wyFRZo4TV~1MqBxJOjTVH`xrH{L-Sy1^P0C7#HI*yCfaBejRzR1LYUw{So{-Qqky*v+YH~=14@FMk^O=!RL -{0tSc23XPY{~5V%!0Sq*ja?RyppgG@ug!<)OoxEGJKXKA~$5q_dmS4ulETA^E@z)SkxDYg$UeNO(DGD -3NASIx-DTfB8BsDhV9b<1t>)PBXg*y< -Wx2yM{$LH*#A#Co)|n{BZ!zkA-oJ!AmKMlnR)O^kAVZaqX}WwR>wsb&K({@eG -@mBOY2oo{6J@0?IgSnBTuc_zhU17M)~NPsCv0kCMWO%RCZUdbCKNT8!cb|R4Sjw%6I%(m8uWq2&HWsi -LX6!1uI3mevgu*K=;tbb7B6kx@m1B*B?McC&tjIfR&oGpQMNn+^gqa3@Y;s+rWy9^!i`WmpYbI2zDLe -CMHdHl7pMG%$G418Q8yajAKdjx^ZHFkClsj@7Kn_sr6VcatStFopb+_-mS&p=b{JR-n}h8So+P+0MqC -R74ny6|A#q1%j#(t9>p?E^Ww%Bm9rgvC%cK3-F5{8M9LhvgK-6M@_VcxM=0KzC`O!uN7C=O(MPYMf_S -4gd-SyeCfZ3zIlwBl0=g9y4YT2{L>cZj4s4lErT2X+_{z#sxqytYUPT~*r2V -|&7uwioVaXYvO8JXhJ`Fs%~($)K(>;eXS@_HVc*P7I+2vNA_h4L#(3POF- -V&{FJFl{Jd(P+mbd-mQ5H0OZi&aIyZ~hnIlu|G1i0!PO{UFgI8;|^}M=@xHR_IdSEqgIh-7wpwL&cz{ -D`LD+U#)nZCk!=j*dBaU7pD=$Q^3<2b35M5%jo9V6M+0*q78x?{nD+_3vAV7NS1y{fEb75KO3Oe_^dW -@7qwo)odlguVSV?ceLnNWbW&#`XBRn`k+md;&GUfQq%s?H$O%VrHg$a!_eR@9I+>#QBB(1QWOO7798(~`!Q^>xD -MY8R!Ry>)6@TM2Dc47Y6wv!-*IycUlpqc_By9OaozB^LIzax~hd;NRTDYffatC$hLp}_Wb#)Z -GK|+bOu1~XC{zK@esCA;HQXeMT22uISM?@x^K}T3*QI|_q`*_r_Ue|hHQ2+#|vu(5X!b-6$Z=eU7oKs -29w$1Wxk~P`9vkxFl+xU;p&p5`8q-E_nZNb3=SXXYtRaQ549EN=GO@v80PdI8DGZwVKgv$Aq)9#7k%J -z1Q$`;>`*z=+3$=EaJK=WMA*)xcMJzuU%W9>S`Lb4u?zBAHoU#Z8JFtVXvC!U@;6a&1|#v<<9i_qLxA -e}W{1+1o5G?XhHTKiMu&~P=Yb#J{pBX}nrA#X6^TjV?GFJ^LOWqM^UaCyxdU2o*J@qK>PuGa_G5V#WMRZc6aBxQTg^?%O -{KO{;kY4>_9P~vdbQp4eVb7rVq@I6VAyitu{NkWQ!&4nVgs<@)6;FVk~>|DyrlBH(OPL=%Op{O3*Pc9 -cZ;`1P5x=kI@D_$vDR7@<%-x&O%io@eO1y9?OnWY(3G%cx=OuA+5)RN={D>DPA4mIrTU~cH1WgnUv(3 -sbCQOC(6t1>R)nx>y9(iu0ZC|kTYTwmwhy|J`F_${ -8hA!*>lQjEfcrsFEqR2`#r5PkU}LDElVx(YazH8L$POyqBJ@LlKr)iN)+1$7cR*IQwggc3&E>BtJD}0 -!TuGeK!p-b3JPec2^R!LVdzKzD#67G%L;|+S&qLhL7&S-G_{e2D@xXyz^}jja((k5xxF4+q16*-g#hD -88c`-YR@ST|(R)K>QnUK|@4rp#q+mG{tV+QE&?=cA8)u#|Nabfe7baS9;~u#GS;ej3`Tw`>UHP6W}giSBTzh9{n$A(zFC@$_5b&4-)KxYS -%Vf+Wgy=Q{+U&`otzNu%j-g~%77&azXv8AYL2700GmlHW|k$H=LXWqzc1NM_qc3dBWD}r!%eiZ-n38V -Wu*VlxzKb0ONPF6yArgGQX6s_#!`&Jw)9M@x-k%T4-Z+l`5)+btrtf-f6#fAxxV&N)u* -{IAK5EAX_9CMC};iHZrLl1vlf>r9!zAI20U=0icD3X -ACNg&N`2&Rdm+&85vFm`v?JX$5e2m;Yxa`MDT#Q-bN^NF#VC;<_f@SHp$;X5i}YuL$d?o-DNZ-BtCha -d$0(FZVUB&|@Fl1N;wr`oj^S7Jp3QGhAu0`M3{-Y!F$igag)C)DT2vhC3Xiu2Gs6pv^Ogr;c&j>X6m6 -_q~i@&O`R*Nj4dbP^=0*fSaXR;lA}&+Xb^fh_oEHAFB9h!8R?d)A7E?{FpPfk^+TJ>hsN0D#x53VUlJ -0{U??{S|HxJWed?2`FpJ;pht-kho|xF$RR%Q_pbN?HnBZr9w%hJq#{_bJn%+ldw4E-{J^XB3PaQwelo ->)`I*zw0TiwE8yvqJPW=!4=W&L&nq~r&jYUoI}Z>N7YPoz+HrZF6aY0(nUt%HflX$RBpkkg+ZB5TkN# -3_;P_bOjAB-OV!^bbjlYS;eFBH}o{h0iamxI`i8MU=OHmz#xmlVAz%k9T5RpyqhU0+QyWVsXtUcqV{C -?-&?Dg+c0cKE@K_WRHH*Z4sB|A=u0WB*c=K{9`hGpASMNi -=TsVIpoGYOoNXTIxmws7VzCX`CU~J%Vrw}!BQiPoD~0ytWDXaJ9Uuk7?ZCVUUexF7g%r15*GXh>8Y-A -y<@RG`gK`#_eS1B%9B!vHzb(HrV|{bGc6%!Fc}jhKuaXv??|5q^^eytwHLkqY)Qc>3o(d -Xu!o1udtIb6Eafmy#FX>zw?E?}edI-ULP`J+*_lnaqG{l-vGSy!r(+g2>^+1%V>%hd$FSPK-mku9nZF -I{h(^4>+9p+uqI8>Xx-u0w$KhFgYYdKbpr!z1d!R@GbrK -awc;2BHJ)4Tz-*i5g8kv9T`))VF#DFlu$@H98RE)u%YllJK)^QZpl!w|=ms$;Bhs%%h5e8tIYk`NF_* -e{r~=m!0Q(-f -?zE|q8a?VNPpsy&`Pc&3E9xXRZ}u`AWL+4_AQdu*$dBZ^V@00h);%|VBPZEL=bmuHX|u;vGya_yRVcZ -DR3XF;fT1`VaV}3o#_pi9ydWRoM9QJ;=1+LG_oS#51t%GhyFFlc)OHltLDzut&CIkBG$lRjiyvv(Fc9 -&Qd}4=9N%z}~lIJyf*7g*`-b{pkri>jB4?{eDnCMnukIiv^^1Q}0Mw1-7?F#$0x&v$b=ejAjK5As^wj -%sATPbnq_QVBlH5{Dy(>3E=LzC6RhP>bID%k@~55!-CLojT1kvw?OC2rssEIJzCm1f4ZMQno)`N+n?QYFAE3GHHJz6^99RD5L -?buCKo(d)b^m4q<5o#Js+Xc5Mxr{Nu#T6J$3gQUe1l>yf8%|=1tjmquE1|d8oTR74o3Il=g3vU(N3|co{M1(CFZPV3X2YCRt#e2qFm-GM#=lr8Pd2Ia4*tQu&zsn -m=K3@F>Dh7tMqT8;rUo-@e-h=uf%@>#baAis9PdawG*=T6*{ri*NYH#DsF(w?&G&gs7a6ud<7l=$BvE -t5Gbt)!!Nx~OP?Q0a=`?R&1vT3zC)GYKg5XJ5SXmK1_O!?_SYYj`0CLSZLhL@!0=8+$KM$GF!*ZKcY& -M!_>uSX&mnuTdek8#`i4K{k&^ZyIa$DQUALh!gY0+Z^UJq!w?Nl}pMn|z=fg5Vyz2+o&y-bDz>8%^4<4TU@F$nbs?uOb -DebczYsan^KsJ#SK5Y+*-Ep#XKKf@cDF9cN=s>uk%^>lLR7v?hWVGD&yQat_iPX&@^~KFx#n?A?XAA+ -*R$Oh8ji4TW?1ZXZ;lJS3j}z7{hs{r-M6Is>_w7VFbP{PO&+U`7Mer+tPe%P4r=r;pXT3Ch>)9&i>G4 -bQoK*p{Ly!0T%`SADQW8({5?eQmq9*WU0y2oD#z4!xAO9KQH000080LuVbTa4r$p_Lr~0P=7E03!eZ0B~t=FJEbHbY -*gGVQepNaAk5~bZKvHb1!gmWpH6~WiD`e-CcQe8%K8kpP!-wmP$YZj@ElsmRrb1*0EAWl2tObPRiwi8 -$g2?YcMmO9*}|+fA_qj@96<4dpDH|t1J_kzTfqGM-L~H$=gn~vQt_}u_<-mm8(NvHnrF`ohWPF%c_$2 -#bh!$J3ALwI|Yxbs<|!eYq4(jdsX*B@0vrk5i2D+wQp|JMo9Rw*(=dDB|I-{(PPI$f9O;J!|adk@lxP -)>;9r?aRON_p36$Bvu)Sx#d67wmrGIZ+otPq=@3XJHz+uee1k`(0J8?9;dKksrVk@HEW#SO4|)+49MYXHWj)<(t2rnXYA9EN -@i3X}Y47>ksl8Iuy7YK7_YNW+WrkUz;y -+aDSiTCn@T82-SK>MYk{jyU(9!j7p41MQ_h3N$Jw?}Nhcp9|JjiWxURU0j9dM4oKDa^g@o9_6&t6oUG -dtd8Ai_g_&P7UzO*QMUee!cVGysv>GwQch<;Nxdq*L3{#Yt>66@ymJ*^TGmQkNo6~>TjFw!;=QaSSP* -yw$_7%%yL^Q>y7$osI8V;{MdQkp>Lnlm$}Y+vo=R&sa1G}b8)TurEL|FFFRbbQI!U2;b|@pU1i68+^u -)zje1#cn-l%2vi@Luyn^3PG`-y^p<6{{M0>ecf8BK)a!IVx~=4Ki3*)JfolR0wR_r?8^z3lpE|XXutRz1cf7?wC!XACul6m -7rlb57e0BRTW$MU&N97eheA{B^{u!~%2+)_`c0f%>Q3p<71XEcO4l4ZGp{?T}`Qh^549%vqPD -YXEwB;GW&;!WVsZQ1YQ125?RIO#m-J@Z`#n<)&TtlrkZ+#SXFi}xG|bF!5um|3Sg^0{(z2iaih9bqsty}DF_HLgg-wVV1)-%BR${~&27FRYHHZTUeN>v|IQ0XK*@s -X1_-E=4sK+BYS%UwpoJDz9r5V!%Pd&#!u^Ju{&$aA>0j|2lm_nswH@u1C84Pi?<^5M)3@TkX+^L-S{N -%zl<#HRfYmJPvR^KzT2Sl4(9SmEaBaNh6n;Kcl+?Gg+(i2z6s#9`vbj`s*fM=ch3yTOpUHYeFnJp(8{)&#lo%ML -%C2N(i8_CZ+8fk7$Zgytp85z+}Nf5wHGSb+!O&HAW2as|g7w4i4`xFq6K{;uHDlGhRyHBB>h6Pm*@IvDwAIe%Wo9tCe^tHVGtjO!nou3tu8 -6YNG@vujzl?W@VQGdN@XLbiv44^%uH1bZfH_0AAq5fsqO(C@jfRiM*FIo)AW4y9E41dc=Y$TOy+@;>G!mF!I6g=6rzVjc3Sd~wvHVN=Eh6Fr*iM6eQYOgRzHB1n5@Z@Gb6Sk=z}sN0*62g5d05H{O!rtxN -z{3Z+GR|+m~C4yEYx1h3raDd9)~z1gK`j`x2wJcueMm4e!>#%hsbuD2j+vTlOSzFEs*a&!ag|JrC%Ug -{FS{#J3(}8i0(5*v_Ut0+t8rlH!|LB}3Xcx45md1*EO8*&?<;sIPGG&$!g-|!D28Mxunz}lMa$|-wV`$RFN0^<*1Lo$5NXG( ->?SQAiF;f_AS(n(}U|$NOIf8^Vq$dNO*^;_~e+pJto@waI1WE+_La`4MaL8E4laSh}rxOZ^AKl8XUV= -kQ;h?<{c#aVuRtEHIG?v$OPGLPEuo2El0RnKCZ_4JNs|fuPE0IIKZaPHYv?>l(BQ~+$Qaey7+vWhu!k -h&H8ZyVoo*rmek#xURFw6$nNG+Q+)ABu86N1#Jl@@G%#03PzHz&n8IVMu89dbd@rW>Mo?QlhpxGd{2GJV^Nr-rF9iLpJ -eDZ+KOspiEsY1L{7lAAnlX-T?Xk)`Bv*?=!!;YlQN6sT9cJMjs0Yb`whalEx;%8kNP{X8S(2P=yv~mh -Nn1}=0s2}B}c>C(zZ~pYf@5Hok(6rV^L8e{-=F)zl=&$?T{$kVtfnTtuy^jgc&oBKO= -sLKVyN)c=Qv2Q$7kxRa^#X3;!?-{9f9KuYb(>s>?W2H4o&T$hNe~G+y07gWUzQz69H^(uY2KlRn8|Oz -NSo<@&gU)i%`)E-LN@-yil09`{{|YjJ%ufhxd63NS_C*vn`=%Q=OiHRw(ATm0BD5j1eW=bG50&mtyO+ -;nNvOAAY@hB{&8Ul>vw&QSvqu$dcd_8#_C&O5M=gmyRrx441j^HLAf%Y&+XK>+y#iw>w;;B{r8wi_FL_=UAf*cLL@2!lb -a%wMRk@gGZk;V>4F7;_&Gs>EfLUU7Y6bBhU_Ox&$Nw5l_FrJxBLAo#>{&0Ap2t7s8N~ompfpt_kue+8YR55Ubl|ilPV8epl*XD2`58Bu^2rb#*|NFQZ9J4MBCBrG&$WvgVbn8#ox++yID|ni -N;PR@1m83C3&)$%C=txK-Re>Ng}*Jkp?d;o@@BS-Rq}&7_q57ckfpK@mRj$#?pBHaQE=YHE099(JCIe -_Grvz483_687w5aU>-0>GG{pO5Z7oeddTG$#G*W}3 -%4<%p2*_&1mfIlWF%$|a>KV;v#E4){vo$I?tBxbJ1~wH$De6seZopMqv66Hly)c?sj48$>TElaEgO(Z --thM4mkcmFAAqUZFFvdv_*`Dxnl2JA`yBH^B*>`kbnJ6HxDc+QGHZ!Y^gg%*nxXwcfMB{(k%!0qi$_5xGggJ3(Fo5+HZ?V*r5RWt9KPI)zUqlCiDKPC$?FNC?R -TWDJZ15BTzEPj;xwo-Gu3LbG-NNU6ggZ=N{y$VL@B<0O$8WWL1?N&tgk{qug?7z$~E>YN!sqYR{-ocq -Glj4sVdmXqmun>)z(N6c`w+PEX&v4mptr4jCq*HQ?=o7-ct?H=*p1tplML|7|UE>)MZV0!ps45Lmvc3 -S!usu|K@D?d)Iyj$7iMo37VyYVGAj_J8z3e%u`w;7KI4s#R+}w;R$|w;+2RcH_=>W^zM{DtDCyJh@5p -8Ulv#>52Cwesbg#?wm=72_kk^mBPHF-=?py3>4?9j4HQ3!FxVuyMkG`d45J -h_IA=nxfNqI(O$mWCc=>hJI$8qR -Q#G~G@cW4c|xiK4L>?13&Rlo_pUJubqg04aXG+r~Bt<{ -XOF^2Ft?>wfm0gM~9I%MMV6MI46f{+!hpK{JzzenDSVeBMAkVNE9N72L8o>JYvezDRdUtH)YD5m)AC3 -7L)+h=dZAOG;L!rfOO%P?n#++rwcMG4D>VmQJS?PG0%a~lV2#y&{GE)X-zNwOcH*%Nvf`<6T -tLY==Oh92XEAD%~;yJ|hAnp1eR5a^V`_if}mevbLgaKnC$-W1nIA#Y|wS6frD9CzCo3 -Xjhp5j2Xl#8A>jj>QpzFI~p1%fXmT?*apUZvc;h}J_|ephG%i{Y#Ro_aV$}XDw_7q{4V>*XTAoSqPrB7)hN@?qM!8H|cQOls3Q&g72@dji$h5Vh!{Gva@d -H-rn1gDxsnfH=WDd)e;g6S#pI+FIBcdommSKwcNwlNi&xG;*k0xV3E&*w#nzyd%&zxJ@~YC|}qaDMmJI98xh0(Q|hPhenx#rxQF66Mk}>A@a)^vB5?!Mt$x -ySKh>pn5wj1=9gM0PLz22e#!=2paXZYRT3S6#a4nPqr4Xt-mO&L0dt;hoEW;>b|gh&2ntTrDAAdm23f -BH5EBEd2OU?q5VzB%XDZ^x`_R4OE}1%wp>iGvUS8jttQg&{oF)-X+la(z!EK9(wZbUr=bB2kkwY^%)P -p>+aW->`5HF#cdSdvs&}~)r)5$pq$`{d1i$N)mtvVUr<(#r;;l;f;zdkJM8NVrdf*&Riv?sFZ0QvxiA -*Bf8V3-;2=bR=YSFxc-t9lGBc6}+MdtiD@WKG*DD%*yR`qJHI|8@vIzl#PKhN*!L -FoQHGqkhtR^7b840HtZq2xd6Ti(88~~6jS(ecwX6S*$O4$U_zuFnohE2xMr}21xH<|b^+rm*j}qX818 -$59T>}Ax`_1maNb-B+Jf3_JFKjTrZ6FANgjC^sDV?pPJqT&H8hw^pepMx5o-8S6MsC^cG)+}?IC18lY -{1a9U0E!oU|PT=|mvDKEcV6>;&i^R}jl)eo5cj#^14H;;FL2&4o2JF%mB5`E+F1;)6OO0CtQri?+oj# -#GzNxOX|jZN^ic(k)MYMIZCb17ZWpT=C#`>{Pq_X!L0AEa)7f_UGcAS~negC26T4m5gGzoJuEVI$;gTSoUUeke_gA@nnr -@d4=O6v`^oK%T4B>3jmy}fU}}>4~6=Zv{`8dGN|2jaqncvFd22*m`jeB+B>u`Twhl9LJREqGq!jx9&Z -MOPYTI%FyC*2c94wesyz78CKm?a{j;cD9Er_9FUu!*Y07!4V?`he#**>`2;LGXQ&^=T`I -0GTu}I;0e9D2^mdiO*m}t<4HGW4jWMdz{$e4V-qH~G&F;bE*k(lowW_6-xj3quZg>A9HxQCF`>)5OE3 -ID7xxQ*~q88_ESSBMS9)IFb0rliPSn8%+xN(9aY*XWm9V~i_xBWqt|ZwpmiWOAtBLecSXC2AKcxX2kX -LNKgv3_@D2*yV(afX>^IfTi^#<8r>Jiz?OC#LaLvv}ANoR^Xc_>Y*|TrN@Y}fxV8%cVP#SVCJ2v4V4jGJeFAOm(C&`zV# -mmCw2TPJY;!skpL_7Aqtiv@JD~*L_b0d_J%W915doOp%(a}s_bAUPd^xHe3N#`ahFB*ok)4j8jiw3AD1GHcFLc}c%aaq*c)> -L7Nr#LGVjZu_@z|>-*|_To_=-dGqL2;^w=keA -JNWbz_s1v1wH&c);2NHZno;2^PfZ*YWo?VJyYgq>X*f{>ZQb{-gIax3w*B8Ealw^9Ys%h(YjZ^?o$x -_^BdfAE-|Wt13GLW^a)_%6fpJ;=!^n--~8QUsIGji+t25c9l*9r$&mMqBia-=szULl^__-Lg!yIjUfh -TkO_`c<;mb>*7;N0PM%Pq3=?S>(*Fb>es*FHa)v4K5h2)4xZM5|EbC_ifA8PD5}%MTKcA9YL!}hi{8Q#c6FJ{TwUdsw&bgAls(4Te7CTF9EHhoa!8Z@!+q}VOer2!gz{6xjuC7-RO_Mlknpb=N@;_fc -yQJ4g>B~otuFHORSm9mJNA0ndkJ_?*RBC-t`q39({OR|<``>|Hmav}(zr2v#1+0j43;*)$sqgCEm|E= -A$4z;yFejWrB;Vklh-eQLI$UmHq++lT5pcqWr~%q;RaT&!hY-Z1|A_i{5e`e080-!QR(ja~!~tOhU;+ -^Ew}1HK;uC)H^CVOWc(n_aConHg)7g|F=5>Q+RQuT7D4)wF1!!tFvbCD!HlRXU{glh3CkO9WX~MmCy` -(!BJyRf@yp?FqeM~mI#~8Xw3l2<2>Xt_^i^(|u;@rF=pC< -gvlAqQ3y+K-hrxIdmDCTs -eWMlR1H<5zKFd=)(b^Et3Q(c$=-UHguM0Jh#y5OJSdXLM4Z=Xjw1o;wobbC_+fT&WR9A?h&?oJX*CN- --K11S>j|8x?4ED`Q_lg^Y~F?}5IQ08&I&02NS%LeZ#?LLG9pAuaH2x5-C3(>%Sxs{ -+SrSEM=%a@4^B}n-h2-X*q%kT&r&eX9UZ&?J{T)le+YNxiT^U$lhwr1$0TU2b+??Rpv@(FVQ_agKp=jse`u&FM -}ymtjJ?5e;bG6hMo|yq3QUY+sWvu6*6(*eEt{g>RfzE#OM7wD7OAoCmLx!z%w#?M^St7SOLhNWd*O_t -4HHuGzmVM`@pve-`F{08mQ<1QY-O00;of09jiA00002000000000g0001RX>c!JX>N37a&BR4FK=*Va -$$67Z*FrhVs&Y3WG`P|X>MtBUtcb8c>@4YO9KQH000080LuVbTdQvEu#N@*0HYEB051Rl0B~t=FJEbH -bY*gGVQepNaAk5~bZKvHb1!0bX>4RKVs&Y3WM6c0VPk7$axQRrtytS`+c*$?_g4@uij@mj55+z#3}ms -rq>G?Mu}D+wLr^HRMB8j+QYEP*-fsVW&kS{ql{OD+AW1FG^>F6Qs2gYZBFi>)sGZ6(q4$+_K}ch4kU? -AHXS3#6ZcR{kq15a4$xiy+;A?5OTWz-OSLWJ-cu+S8JJHU* -PF13RJpEGmQ|Ok!9-CSNg#NA7oj!if8fbx%ygbr}oN(Ddjro(x*Y~tE|wjbBgb8pp_%0!)!*`C?`(Zw -9>5#@A+AhWk&9S)?zlB{n5jcHvim9BOx{`)JD`*A%n`4xyhQ-tT|4eJ`LgHTv{4{AWIeX@v|a5nDHf4T4Pk}Q$&m_z -*9NvpeDRYAN{hg0XQdxV*Zbo)RCE=aVEK4Lox^Pp8(#B4QcsT>?V5$DR|@|3VnZ|StX4}j3$M%pR#c( -n>WXu}+ik)w~^9I5%KEXP{(DO5R)DkEEZb*v)2SLhM#x#3K%lCke)tYKh4N$n5S;mE{2kjMHz? -spsZ7Qhb3E4kM)ujue1o=volvv%A#)X>TNud~0hSD((_zsWAnU%q~KdOnY(f2+f8KzZruoem!ZjKi(; -A`f>xNVo~M4*bdC7`?_X6yO*H)bcCBtrH?%2)`DCQ`+6N+ZB7+H)&mlkS -WZPyKw!87*d=4Mu7Re!>ETQj^FS5bmJ%2=|_sC@#_Ho)rldQq-8w=bw -!JTC4IPbb+k(=0Jf!g8o#_-rzB8S#GaExX1^}o$+CGo#|8KG4r4)i6Dh9&^03G$tN&c(091e;8nSyRH -X1K)D@pJqa}?P{Jm4AH-H+boT{W##S-dm+?rS*_+uk2Ik2hb{)F9#tE*wqDQA@|uC7vXhIob+gx}d)F -W78eJFjnm*Ah0hkA*L_hj&C6NfU#(iTna00Mk~Uy7pSE?SAcr+OJhXup+QMDk0alMpE!EiNzV$$RZnT -4jT26WeLF)E2E!g<6a<%G;6wn{@go3CQf%_rYnc##7rNIoZ|w1?G)Of7+1q8R&F1|Vwgkhr%zz|A|uMB~%imzP -SKx=M!V(Y+lq;;E(xuGL}Hw=W5EM@C(l -1tsB-Nt4lz6Gy$jEDCC$V|8MlF-i9AL7yiWPl35#3qjc -a3YowsDvu=hMPumzn4`;K2;l^23n=x??oncXdWH-1UF0QE?S&I2BIE-5n+9&fmY@Yu%pS&pC3|BM_?F -1eMtHmg|OGHhq4UyRceEO1S&QpJ{r|?qp)Xf>N2lyK~*&FqL^!YhkdZ5+0n||EvZ&axQzha5JnL^z6) -NK6FI&@EE4PP)L-d_b+tu|DF$z+rFz|!`94DLg)#+(?~`x*Qk+LrIT!8nGDR(<%6)r@*9K`qHBwD!&A --m44CH222d$22yb!Tmw&rUv>CULxA0;wtr+a(VP?c~d{bjq(Z>uG)37`(KJjtdsOvH9OmYR93_w(de! -08LHcX0lyTqb{cYbWAiv&?B%{{T=+0|XQR000O8%K%wh)&*O!T>=0AjRgPzDgXcgaA|NaUukZ1WpZv|Y%gzcWpZJ3X> -V?GFJg6RY-BHOWprU=VRT_GaCwE4O>f&U42JLi6@+(~4Y&q&7zPB$4&Azs!-it(-B2thW@}4=B&W^4A -0@?3+%}yqwm$fhdPpkI^IzzN6kfm%6W39GTJTNi%J|p-hs7G04Ef)n!AO*KKxA5`9r#g!Qf6$MX)8IRtgO?{MIo+?G+{N#--}u1T;JId%WSEc1fRL6jqqJM7_xRoG_Hm54t{IK;5%`tfJ5eb4AdP6*z#In -AKK?k#8@=Ko-)|*c9Yyac8Gn&tp<zG|;1jo -OJqx7LWB_l}Xp!H>AXA!;sVe4PuDu@mGB0P%L@d!b>xL$|DAK1qkqpDz!*U^v@Vb!N>SK;3$^RiM&U0 -HXRMSiFR^x+kfZ^w5I98Cui*G_4RKZDn*}WMOn+U -u9%zbYWs_WiD`eeN;_v+b|5h`&SU&VK(4;JM|DCJ9O(l4jYPXcSE+ASgkAxlAJdGev}ls@!HfETNe39 -zDKew%b(~OXP&_x6`|3!6dy() -o+9M4q?Wp&biVy-XRM1gW&->#L~EEIlVOnWjM4ZnJn!I_CK~+6E_HH{P-0-+0fQa=3*3h}Is-8 -l4YSV*c68mPaE~LR{E#(nX!zifW>2g49!7Sc*!m1;pgtGGK;+hF=R_~0@W+&)ABBT3edLcWAm9@Xak~ -JdW+f)x=-O(rj_TaS)Pwjr$7dz0W#(J2Cs~(^YrT9WW)o{?RzC -mqHxH@lxdIg-e$6Lz+m;($SFEg@CIJ~$=Dnl6(pWSKHwx@RuaV4RKcW7m0Y%XwleNx+w(=ZTy_g9P}t&~W`5)TNeQh8W*5l9GD`vaESOj7SUcCg2k3ix+ -?X|ibx_{Fz5Gv}O{&e;Kk=mG_YLVy}Z>j -@KKKm)%ZBV+smAmH{^)Y_Dms+D+;*}wEMrs~NY0rJw#zZ^xa|#!XK_?E1_QH=UUxxTEdt@w8LkF9QBk -z$u=eJr2BZsK0VT`rd!8TD5hdrVeBaUxgzpWX>m7><*$*uBrTga~eLA?YGNV1hAMXOxYY~68@p_I7%Y -2eknbb?>-#Ts1a(5k9(8Hg|N$#{x29`Rd<*Z~be@KF=c9^v8s0V3t10!|ijma{x!?4j{6O4=B|d!OlXvMoM(eL-g7OdtBtvo}MS=N{R_O+BaU1%&xd`cCjwlC&S6asbGn)O-D!avoBOeg -GO!p<4~Whyoo_th2P+WgJ+(>bhY!XG;v3GVY2&A;joI%i#7eUM6HTPEti -CmYroNQ4aH(N`08mQ<1QY-O00;of09jj}hMlxT0ssKU1pojr0001RX>c!JX>N37a&BR4FK=*Va$$67Z -*FrhVs&Y3WG{DUWo2w%Wn^h|VPb4$E^v8$RNHQxFc5v`D@I6>K-yY;%ELO%zTVPTu%Pd#YWZ7sFs_VGyrlayVZHONDKZ!539t(=E-h-w1aX!NLyW6DYK>;b(@sj%^vLajnt{ -hd@9$ZMSr+O<4L4K4q(l2P+6qb~>wTMLmSqjLP=&ha@qj2OHr%pGV3C3Ug5MsD)w_7CQtkVEfz`*PQtV#s{Z^5`axV27*Q;f6(et0x}o$QK@Op_M(bM_{<0gD* -9;Cs_bw&OWjpZ^*3l?l#nS6K2QJKvcY2bdQ|CxL-dxx)090sPQ$wI3%{0@`-3&pHdVpT87d=4(>T}U! -bej!4dYADiD?$6Smbv^^dzCvQ(!33DJcdokaz?qPni}P+XXg!~swP_t%o>}l$je}i2P(Oaab#LQzQER -2Qo5=6t-((c&Jd7&?71qodj8aaw`8LAH88ix$(hlq;g?W_lr^&ae2$XN%Y?#ZHB|U{#S&yXL$HfbHrS -Zf<{)uscO=@W+z^L?4*lWJaJmXmz5*{ykQd}07qhE-s%>DsTO9KQH000080LuVbTg?!!`ZNFl05Sjo0 -4x9i0B~t=FJEbHbY*gGVQepNaAk5~bZKvHb1!Lbb97;BY%gD5X>MtBUtcb8c~eqSa?3AL$jwhF%}Fg* -C`!#qEJ;mKD9KmI%quQQ%*n~jOIIjJOwLYBPc7C{Qc~gq08mQ<1QY-O00;of09jj>u5tpqApih0Z2$l -%0001RX>c!JX>N37a&BR4FK=*Va$$67Z*FrhX>N0LVQg$KcW7m0Y%Xwl%{^;#+eVV#^(!XO6z*7>Gw>cC_35Hb=7ogRW -&Lv+b%1LygXG|xm5E_UM%rPccxWzd8Ty{sd-*z%~sXf;ygRm*lnG4NfbqUd#k2et2AA0x=o|gROM@eo -RwwOWnEsC?cSa|TU14%7yQgVTGx4@VSM+b?DV^?$mjM+dttuvAMagJou0yE&CjY$>a066KO1enwP%|y -w_~(hI{_T)vdL=m`!Z`v0NC=n=2@%ne___NX^J*m>2!Yo3tcX%rRGN7vXX4RFs;AWrEapWYNqOO2jAv -}mT&azM8Dkt6oH}FSzaV--L{Bdf$$^$>nqRH#duI_|>{+ONXW_KT?sp)X4yHvl!ucbiW8OE!CPJz|Uu#X-;KYsWkefj) -{^vB~D4<0>u@gRNj;K$>sdPHqsHpLC_+Qpe(TYfls_?P1!kJG;%|NR7Z0KFRPOY})rc1>2OEPu@wK;5 -)k*XfEFtI>6lEpC9a&Re@BPr62D>z`n|*BZ81E!Xp4ts2>QE%9sDWQ$HOY2>u4(xqNyn*vw_u$R`wriC_ovFY5VgdLWGELf*`_->S_J -X_)cGgDYytC=nK_VyM<*0#hwvDx608PBg-q%h&CS*9pbXX-iqc!uH?-!FCB$s6D;UjA8Dp6caH%|TA> -NvoA=vkM)!x>!xr!8htjRcamqKQ#;w(?Bt<8+Bylc>=8a13imV+LLKI+1qn7fR-Yh;>bOXPyjZQJ=vL -h`S9d#&#Oyl+->SYzj7_9;gqG=E5H{0e*GF|f)j2uazktvH&t`k$LB2wh#9EM4glU&Uw^Ic-JumRaB7 -E8cyXW`pKDNwrHWZsEP(+|0cTTH=<>8Xn~?IwG1R6-PQo!s5T0fCzPuk#P`e_eGt&cxwOmS*^X4^SY| -{wPE@B`@k@{??wc@jc8^)7Kvefh_LOE96jV4KhlA?~IqljTPvI;YMcpb|MCW0?e&}m+!=b$@J+Zb7w& -_>8v=8KMT11e{Ln{&Or1d=G|xLL!_X;oLLx~X7$7Fz{>N|h;~B5FC!`>7p+(FW=}##?SE2lyR)iPxW(buA@(`!szZK -E&1I!8nhi+76)gmejMi+HaQge|=bwL8Aa9>Oe?-LAEx;}cRcgIN+|0Gw)SwjflG?-W(a!8i^49Xkg!d -m!4Dl!o6E*W_0^JlZ(rv_5-k?q`HE<8VOF&jG@ku>dQHb%hASc5%o6gZ|Cq6|bL06BQv9z0_p$nvnWi|P{O=|mlY -eCS}@NCqa$2~DQD0^!!iXe`VIc7`YF+@3~xH&x+p+2mlyw-qQgW7i$2t82DOHu>-&%0ik$N|V5Cj?+l -c>P@{DZS3(DaHIDsIc=&RirsY4o$cl}}x5dt3y_NNb^xk~La6;SOe5XD%A)z -=9X*RDmoq3nSR!yriefHqRUt)PSvwjB>>@^V;zNr85^fCyeM;h?H#(f2g&C0DZ1j`yt5&&c#f+dfUy+ -IJv;yWdt+-u!qWxbmfbk_Vzzff3;FfuFi_B5dlJM4*6{hnxKMw)L)hmh`tF -IxrsLK{CA=%{8`GQ4;Mk9;?;EX7TU&tJ> ->Dd=3sR&Qg9zE3^9l~|Ia>Kn&vgI=N;7ts^e$r}LmAB9xhRq0i2WiMF5i2$a&D5PC*#UJzGA0jl9X7` -iaEHIL6ESlIzh@0zqmKOt_$AIK)&PNbAhqF?UB^6lt`30I-99O}1BiKsNF) -5evM_{baFX2iiqANwG;W)6EU95aH!jdz?uK^+4`>EYkK$?)#P-WyvU{F8c1G-H{?d4>RZSFm;~XbaF^gj%#O=w0l=3a|q9bR2dkEqyms -B4lf&7_b6}tB6~wtFYs%JFl<7>5M#m)#@AIL}qPQ*SM9Vgr#{L^uP~pvE*0KP`#PHnjPE)J@T1~NEx6 -7)@uOo185>;0{jm+pGXnCfW30%=6f9oX?5$$5$gID7*P4*4A$;zk;EZFh6RVWqcPLf8WG@zj2^;{&6n -VZf>^|z3Mc{5M|ut;c0j0YwE_1;ub_Tw -oE>XMh2wf^`_GMKrvPG(@ufLmv2GP+&Rx@*bhdW_qv$pAGSB4Fo$TvzN%#-vhamc6qTxr3D=v+H+u8G -M|2r{y8C!Q^-S9h9kPA)6Z1MlL8_F1|VWSdHUjbrjYR@n!^l$Vd)IGkF16m$V1CSE}@0K`Tfn*V+n%9 -{J74x#E{^*>32135#U`gK!AxApufmPQv{5TK93uxJA`Y{P6RI#N2&|(LZrhrA2 -#TG|FV;&JyYmR44Nly~*XuR5=P@41iMVfsueVDWJrl3i -$>#87?F%)I3|9qk&c~(dyF&2h!}U>*{uPcz6nOY%@=Q28Z>w&JOFmKFr&8quawTzqu(a=RhIt432b;HbGuRY0pwcWVnbR>Cjc_s-5pYn1Es{zxoBFTRS{0`f_(HLXBV863PEk(ZPTo)Bnb(ycgu+WyHis}^8QWyw1Z;fU -%Dr=s+O^wdi>{HT(vAY5)`|yp3~KN&3>{;FuotHHQI6p!j36 -Tikj}O?o~WE+Y2S-gZB6Ufl4LPG#Ix~v!+n>=HcNr7&*1GTXKnGl9fjZc#P^k4Ujfbn={W-vHCy{Ld; -n^wR+OyCn#CD%glGj0NK~dVOCWw15ULjjF`#`6^(PhjH>A$W5VL)M1RhS*H|q1yci==#A{Gcxm1j;S2 -5f3pryXDz{%JtH838|2pSD1E1nQhy7~b1x7)u#)3iJSmb)8Cm+3RH2pVU8Px;nHjf*4=#kmH4X*a?>mCX2QV$h{=B1kp^tspXb8pZmWD{dh4(Hp;Qn?t`iadO~J^G? -!L~^b?(eXZ1t|srB=?&8pK@)ZaN7RL$V(*>5+K8PBzYr#Df{S(>MmD1fX}tCYSP-L4+gdN$SRT%L0dg -lgf^;GM{hq8~G#fxjQ27RzhuHAEqPQOjiG?ZEFsmmM5%$}Ku}Nfj3D78-Q{S|EMApxG(*aEhq;77t$` -mE|P>#$uun_8PEXu%ynt}3$HCWUL{1X^72OUO+AXpS+`+?xy)LhE~3$guVRIqF>IYZA%Z(l&6OGfqxT -H+LXY|&h-o9b8aJXGczwvuH&N11nDgZ0{dnd&M-K%w(S;cb%y60cc)zJQKwHVPvZrq6Z4zJxT(l{31e6sKqa -OG!PjoW%VPLK{6Uv}dB*>NtOp<`DztxKj!$jVuKnfH4B7fUU>eD$rK|@J -6V})q8?B}rfSYJ@dT{ffx+x1Uuua@kK)+4TC_5pf!({^fKe(8TZ9|9LU)89fRq9Vz=vnxlvwAbL -Ur^p2ZIlKBauQ(!w?kk#M6la~+nJDGRMj~#D3~B`-W)$UdHMYK+2d#F(`PRpKYeoY=8clL(%Ko?7g>` -N#o1{kI2i9~szM%_onwNfAo4w}5Ug@AS5y#3Qa~@xN|wARm_MBR=dw;nLuZv*^1I{xm#%O75NOF-E#$u -*cNZ0{WPb?X_+xH_{$vtWOz-k4+PW*S0A(cqrCL#r}^H()7XKEcKqUE{!Zjs}a5Od7)lHG2X1p%OJ#4 --Sk`dN5|(cwPXVJ;ESVIvk%(oKFCZLv{4QQ&`~~t}w48y8!UEswmHJq^{hot|M<=?}fLc4lf7<3hYK6nrDN5;G!i;myO4#8SR`63vG86= -qWd%4925FY@?2-`5zSXd+6(;NCmHWI=0y{BpId;;Yx<}atlteW0sr6>AbhjfyP@i@97++Ujn;M8)dBs -xv<$AkcfAZYR~#UTTm>lFri034;k6jryH+J;hJ@*xhi=DAV=lN3};ge9A -8=K-P+{#0Q^9V6px3TJ?TQ&#q@XDPH)RN0aY>-G!`V}CZ)9Q^|o1Ev!8m+LJqvuz6uLsgqoaN|34Yw; -Waz}PfJk?g-Nw_-5K{=WgdWbyPku+h~a{cZ7NDW^;eXrmt~Qc=LTB=|L -kL05JMLRpVcZrCt9OG6l!DufOoW@#B=4N6*}bQMjJx`h=mR3dQJApY+UERQFwD^k${+H51jI&RC+m?`UpNJYab0kJT-*bcCA{c2%9>61@6Od={=7&?M`kBMB*GExu8A!d4c1Z}ZY<(w_!c4g2B8mU08~Kr6$ -Au!0WR^9qbhLHwj~ZD%fnY3#M0%z6*~(<}F4=%)zab%?dijgR@DKomlotPhDQCH{VT-*`j2q!11$+9+ -wwckuOc4(X&csV9h!B*Wk%Luz|~dxkO7-$4%Vmx0}2{*YiYco1b*Yx46nYshZQnJ`58Aq?(FD8vB6lq -r-g=ga@Fu4vZ7OG8L#jAjtS=0=W_${J6YimbY3GRkUuecsdRyFosV?&e7nwTG-Qh0MuM+nh*+mulO`^ -2n)$Y6WqZVn3Sd76URc&^4dFX7kl3)HmhlZLvr%&R%_jHAv!9?0_{^4u_a}JY;L^LEykJU2-2HIzrm9 -k8;o9CWoF0%q8nAcT4imgo5V9x) -Qp0txe~w^DFi@-j_t4Ri#0$XRoGf%#Q1H26@&EiA^oH5rP0ueBV3kYBH62VaEsa$M4xck!n&mU>Lk7~ -eA?e9!*F>Xw0RJ>bWwgl*bc?`)3LfaB3BESn_YBbf^Pe8e8`k6N&8E5d|k8#J{uvs=x$NaV@H$Vp}OH -k+x3f4p;ds-FGu{LUZmerfApI3R@FQ=ZbONH$5=;!i$c0h-PF4*qlJV+hCL;l!C7Qsig#R*oz42{*tL -9^VikSaBw7Y6|>Dnz`eZGk1J4#Qap;thXDG_U1I5Q81Et4vO_f*qe~E)GbNknQ`z0lSb9b7>s1uy4ac -D-a`_33W6yTmK^t;YdZHB1R|xgK^mSL;WM^!{FADytcPISa<8_~hXPAaIbl#U7EieFWz({6%V)1-HKH -ppnmffwT6Qxh!#%(r{lFWdfB~i{z+gO(7hV=cg7$R&C_vCCp*?u8;>EcnsJx>MA;L&zlvLK^Dkr68l} -Dp3)`GVnfp5K}lf7;e-nOQ%2jUE@e5&-~6RLVCL?fw;xezR}&gY)Fq!htD_?wsW>w%~9)(2$R;~@o<= -A@4(5Yh`ZxT=Iz#Z@*oC3NJUJ&i4+O_g`_Ab;OEC{L71Zkdj6A(ti6mi26qWFIyKHG)k=L2<;v6x|u# -BHDwZy0GDPFF7q{C*~e%47%`1=%MSn2Xv -QaCp{n>pfp?n*PQUz&6h+=EjFr(YQlfyDDkOa_+>e9V@8uQp9}TmTd7}-499OX2D)SKP1FYNq2!Jr?6 -KjD0hX0fIolC>aoh3c1P(Azf2;pmI6?`yXwJAZ>0FauJDyC3^5_Ve2c>_DX;^~iBUYXBb_|$4|@fV7k -DLW4rjeHgZL3`%+9~DR;Y5zLgd4xX;~4_6x=>eCejM+U_yls%opuN>PAp6r!_`S#D@?AU_H~xoFVI%Z -4+r<*xh&aNpnp>ql{w<^cxHmsh>3A_8|z2inS&S%G20&-UH`rI8|AOskL_&NU(N^ej2vgHu4G6GZ*0AP~l+`E?>PD4!{?`sJ~(57w -zWZa5o7h{H&svCHVRpqh`t+1HE#a{(7dayaj@gNxqK|$q?8uoVZJr$yz-g-$@0J%dFwWJom~`=k*w}|Ejk;P|eP!>yGT(KbaQ$LrsQD`w()Sm(^vd -9P+I5(IcCi$+a4%(JG3mdXm&oNae|Z%#NPBfH^v2_1b3cpT}j&lc}F66OyA9?1F|Xi?0Sqy`KE+hWh5>wUZp|S5VtR -XbhsB=#$o2W+75Hk49#2(&kuAG+d}&fVf`zF!tSn$uB~mr>s4e4pi-A>5F3-$ftM=+UF{&}8Jvzj6xS --I8HWj=dN5h)VJRS{*cyAE@D(Fg)29Nk+nKsq>;m=WGd#~^EAxU$;1kahO9(vRnE+C|jv&Y94J3?JQ& -N#XQFk3!u!gyrkmy_YN*=zMk)`vA6S*AyQ+y!_QUQEBHBfuxFh@um(9MjRuVO@xSPnrbjhZm?h4EAna -^Xzypb2+$QE~a?Lbk9juDYc-RJxp$(cOi?th9zKD*D4iXKJI1BLN?u3$IqtlA4a8#w~n7lIs|K=Gw(J -I7eDZWYWqIQpfhG_UE!*?8%WIWNSy7uxng)c1D=|rO8(w}$@i}qqEM6o#y?>co$hSe0nM%d17P00^I< ->%TK2J1>>mp0xA%zf__U;(Q!M5W9J%_}HCWYUpn3I7m;m#i1N-@9{lO;MA!PU5bVs-Dh&1vTe`K7YS{ -gEWAFS;1-4Jo6F5`Sd_fd&5+M1(`QY@45MFedS`LbqTE@5PUA9^Mq7ClJ%qR4*R(!R`AfC28!*f61N? -oCzqSOM)$u-aNTI>?I`W&Fwej#B6r*3NoBHf67bnYZo@m(he9I@qgTKdC_M<0j;8clV-y<44aK2mw&w -Ypn5DCVdc79LxMqdm>7HOQ6jYwe@z9w$-Lth;}1Bj!D4UYhF2%Mj0IX_q9)Myii9UEH`}MUx#rPG+e&&@_2R*l3yRcS%@JG-WZRpjEYWS#C?|Cu4LxjoLnvlF$3@V -1F>K572=FpvAZgsZYChTE~!Dew@yie^Na>xoy`FjhI@{3p7M>irja-Y|Cq3Ny -$6&8r%`@e&7jjg*7UF&yTwJJtRB_tn3v--GMTO-=WIcv@G!uCo%>&{&hi;>M{S$r|6j&6h^cumY`TPT -T)8!MIJ$VebBg@S@iY+_8ddv7ziUkoSu-SZchYVCi%W3%*~$TEa9Avhd(Wn7Wr63{Y<1-P}dBCv>Q{4 -a^?Kp#8No?U}k_o9OYAlNS$u_~H0@4!6f9T9-?|<@c97uEJRgrQXanC(*y*?uPvT*vo}) -n{o`JmYJkuFEJhIe{P;g#9-q88e&k)0zVd1aJ@%DHAH+-+Y}CUO9cl<+$q|YWA9TI{&V_{l?D;ajQiZ -Rl^h2L?JvY5D_s-aRhNuf&`XP#D(OytAm -B-Qlo+f$xluN?0@MZ$*1FpCAmM>oO;YGBwnuv;g{JKU2t5sx_NAxFUGw)IQ6L)gFOux}-0CwRYG2cNX -vx=tvg1k-2jinFXjyQH>BK|{l_BZHAu77)*(h5^~84pTejsNb{qf?53Y78>*r0OJ4Ns0AAvKG98()3|s$yE4FEGNg+>A`9& -uo;Kltf*b7J6{MIvg-peXI0i2GY+5aHQkI0f4hAT|&?&!B&FLB;S-Vz4Qm|91zM=bwi$#Z1CjSMwHr= -zil!>*2YNAByenWfy3{T4Qw?EN=TO9KQH000080LuVbThGT>edq%K0O1P&0384T0B~t=FJEbHbY*gGV -QepQWpOWGUukY>bYEXCaCxm(ZL8Zh6#njCAyNi5)xLfl|V+gVL5TDV1fGD`|=%*jgn7w398EggT$OJQuz5lC9V5)J9j -ofdtL0(i5si_sX=z2L2t -PnEf=N^5V!xsKaZlbV3Dfs+ -;?yhcVqwOYA^oEJxa!2*;9~9g?@5p1dE&&XhZ8&mDWS_99LV4J_U|saxpNR*<^`XE3hPw-KLtEiDeIL -b|4c{p){6>yFRX)xo_}83pO}`^w3-jD@LQj~134)rru7j^3I8-T+(N!E+*@@Wus0UQtV<86J*RZ#XUU -2rWMsFSR0{wCH_&%`@!ctjy#Vn_c`tS0p`}&i1f-qPzzd({6}#zw*|WUn$-ErK%=!yw67G^uD8JUP*RB=xe3RI1k>uef_TTQVPfeR^%o5F@&&ib#Cxi=af*g -5ge%~%OdWqOq|x};@N4wlnv{QTCA}que?1xMrfnr${#juo&clM -78}zR4eSie^qhNuXhs^}%hTz@nn{sZd)wcWDBE!MD0vl!A;_E6!u2vqN(IEdMhtS -c1*^9g0mWNJ=l8u;?R>_|0Yq&siS0D_mgKP)z3N5f!-`AUS9+Ac7G2B3H+}D0{*-h5QeQhsKBet1-wj -o>7w%IanZ(j`KX7TnFN^wXx>1hr8LjTV#43@m|KkhPE;a|8yjKi|S8_^Z38QT`1!&<`Ob;7abpApigXaA|NaUukZ1WpZv|Y%g+UaW7+UZgX^Ubz^j -Ia&s9h?v{Z_NTQb_7=qljvfcmv-g^K@fTX -l?Qs12>$x0yb@bJDrzyv|?pFF+I%Pq_IRU!7GYs+>qr4pXjZQ -96M&1Ua+g5{auY$s~?rdYG%4g5zafuQyxW1Q{z2Y{OH+VYmk%`|wG|14%oF4$h)AzA=p&v=qW}SfOIxP}S8BEsMFptV0_G|@(^@o*s99TPqK56ujIZ+|Zw~YjAhBAgMne&_ -S4@Y4V{kawUhuNPJ=cP@3bq%868CiG`@GrZC9CoZpX)7Ijz=t0*jsL96LThrlWGah7qy!-5>(s)1-3{w$8E-g|3(z!Ex*>xcoGMQ4M48 -Dtu0UWoZXEA?Li6(73Ux_AML`ohqspt_8w#fznxfB&n_6&}x(9+u3kE~8S&{&bppPqdHDmAz4B83NAK)Zlh8jFXusAXS*xx;@kW228o7oI -?7|hYiDsv3y`;NxKBq{kG%t17hm8dz|S(w*Jc${>@@vS(hFq+L~nb@!d%yM2O;$u^D72?sC42B|h@dw -tlVC1jh*8+#wKiHpTDQ=d8-yWdX0+hN*cVQj;J&a#O;q}$k->+|OUR>Wq{~9cakP*$+;9&_uppksT_8 -Jm?h~%YAx_SF}Tg$cze;dU$VHif;;;--*P8mZ1g$#@PfsioJ7?a4u0yzsGNMa>G3>y_>cuv;=p|dImHSY}YiETE2 -W)i3+UB0gx(VS#R<8#TWna#qVD}OC?4?jkws>d=IXz=XMZYUEg0fEWWtuRpL3Q#2>@>d1TZZKY$LhE9 -yr!1>dR_AolgU*O!-iT#L*zmK%$V;Sy)$1wC3SX#$ch_#L&AZ2>`4anBD*(?L@T+&0*){fRJCx`2GtfpcKWR+N!6T~_3zO%_LYtvE$$Tg -A*fRaB8b3b6Vy;7_e;;_T7?x0t{xt5f*1o>4%)m3;tw(}Lj|5fWMQv{QAA?g{U*Z&LVMCUY=Q?k?c;- -9^7w$qd<;4(uDt~~M>rMwq|R}K_De#RVPx73!O6leI{RP@T2mifGxif4(lHtyXlv&&W}1l7T)V8)gt- -QapnwDF{g@W~{{kYCXzKtHY)}d@QmI3EhF%X~)=kXbBd_jr2v*iKM5d~32H+rk10b}XyQg7%%kwh!W* -emR#y(#7p}kiiStv6x{(59zkpP(`kAb(wKk)#H1>rq6!Vuzvg(Z;wVr0p9G@gS9{f!r`(1EEZ77)KA> -HxXg$A6Ak01kk#3t?uA6j*!CDAq*hifU-1?o9g#>ubrULz<&AC*v$E@Uc^d?fgImW|^Y{k0eHZ%`VOr -C_v}|hxFq##Jh&#FFk#|g&kiwf*gEnUg1)*@NRBW-S}lX -fRSyhdMPj4^lYjyASZuctF7ho!w2r6_;|^DVMpEbQvlu<~nJ{|%%OV#ajuX#~)S?T}92|#&p~ER!h9( -?LvjTITK-|{U67rQ2qjJx_0DvhTfcg#=KYASLC4OS&iwAf#Q1dki(ybS1S}yv$qa$?(A%*pW{$Hh}5C -l=82a;WOy6Sm4w1NFGFFT&k!==R=r`Uelh!3QICI$j -9F=o074C`K~UiY=*8WmQ2dj)#u;{RyFI~xX1Nq5YG0gTJW266&0$3OxCM(w9479io5zCzxE~!-tO0&j#-$v9ms0URU^#0^L|N6rvM1@@C#UXaN;y -H&;W6PO@eyt;20GmN_lQv_MfW~(2^JBgkPzd}1U@wgGKVXii33ZTdop`5sg#>nKM#vtq|HB*|y>zn4Y -vrfHc}ex95dVW~d((;LvNMmbaXFG!7snJ?Uy>sX2RxfaQ)10e3NzuKo#1!Cpgt%5@yF_R3y3HE$SO$k;A8?ugFGN4U! -4**M~ygruh5d2w5h#)jQ~?v9gEkv50$MZ5RRZ=bu$zqY3_@W=f3BALX -q&vuLR=Ny3{cX7q|<9|#K+;sf(kOgzXj2XQsPYKW1yi%S|=zQqmk64yuKa=StBGbzP#zU_JSZ2Cn5N= -$+G-P2Z7LG%_keaJd$kR{2n+spR>Bs;%wz_Wk;W<7w~>ipnH0(t0dv*ph!?w91J$# -t48QnZv0>3p^*NNMiN+1se24a+l3z5cHf{p;24T#po-rRFX1o-h1>CfDQ0NG0mY!iuI%Z7=v`~aNaXY -nkGjFtbOjt@>>cvJ++3HUzA?PxAoQZY24TIO3wi>_rZ`2$u&j_V}vgWBmF%sd$(4K4pqzSLLEqOlhTj -;^>a*&*HG(aSdNDBZ7)9~ZNX%l49YQ`Jg0u`ebC}az49g=FzFX}m!2yq9~I~Xl#h?V3whsifcZQjI3d -@WD`l;c&`$_pVT`B{{!_jL=JL_@BGzD1k3Gyoh25iI!dTDxr?xSJB+RKkd4V5E_ -(loxKXd;o2QPkKDUY$z7oB$aE7pKJ7%aSMZhB#A1ft`jX3xT>ljw4#Ys8pT~b~zH -h}Ema#7F1RsXer?2&Bc{pu%Kq{yK4!_(AUO%M0WP_$PXW0G6fCQ8^Qf`lL&ETXXlkX7l}W&!Zv$A>oC3Y{g}W}Jj@lgY+ee80=n9riLIFC@oLH-%bA`@#5hMW#XH(y -3|VZs53yKf@RgMR_=1qciG?4IiHY5hr-bPaKTH%x;1CP3Ff68M-*Ebdh-=G3)LZESN8Gy%F|X(QhzdT -wtV``eMlYc+9pMAbE6diuA;#Fe`72IbRr5l?j!AC|ZDE(uRy@AK=;<2D-VtqJ#5LwZK@a6vX$t!f+71 -*`_UZYtf)Q1k}fVWfQsjv%<*k^UYCejuax%Qt^n5~;=@S%Wv~DD~Xu+g-y<-c*WID8=i814n;Kh!S)VO?sC}lW9 -wyCDp;y-2jz6kWtD$Ej_a$_Kg*zE^M;Odxl2d#vAzJ_Q -nC4~IYHA;YGnULVkx5O9SCmXz`@;&3146kP8U5-=?pPX@GL$`O_OBa^BRk2y;KBl^!S^a^*+#Ow$3ls -ZPPizLGtZf0qMJMgTm*KLzC$bRylglO#!z^boFs`Q)xOE(prt%kQs22V#>%8^c5=>d}U;LkU0BuIuql -JkNk`#j!FCy$Oud*;R#$1pp(0PKY!I|%odDX!-bR|P4umny`1XYP#YiEpQw}8a0Ga=#xMqN8ob8K_Bb -PV-4k{TNJ&epwfgWGnCDd6MFg*{(%OLZzn(6NEXL7Y69a1 -Zh8Q`2Xl`g7Iwbx;o%7!xcoc*Com1~io>98>ZqU4O0x`ULCj!D=WMAL_a2W)@q4@Kc4!_+JRw&}xTua -2;RAFM~zbp^1*;?P~m!942@~%_wsHs=g9TS8Ip30%z%|kAYsW($|MISZl$k&W-hW>n5p=Koj4XDZ@gC -$NEk1K+yHInAOVsGQC@;V?I_fyn@%5%WOZHEBj1rx6C%qW5T@D3`UutQnkLUHGpKU17X^TA=SJs5ua@ -C#!V;BK}7LxTo~!{yXD>G^665_$kNRmeT1RO+orBbq3P%XFcdd7K5Ss#MA6#t-8L|+N0^Gz5W>&{0_S -k}k%g3SJvcBPqoBtm>AjK0GV~5QS%y1VDVx0a=E3}>^|5h2<)i9N#t+Z3#1zLb8{fqfl;G%!7po)Pg{ -f0$S;5ZDDi~qaMO2Wct@iC~;q{q9<-rl;2ev5O^)v&E@fyC?^_=P2K>?oqW`{OzH+8ogSbUo+_^ -07Llcf3)-(&4VzYm9_p$tJii|3nqH2QzhWXXjWl05ki#uTSkse{{Uw`-HljQ-eLD+G+v~>y4vhrBezW -yEx=Xxm1UclSR}2FXhrBA+HDF$&%{lDwJaJ%)Cs-r;4AtqQy;6tSd!2>Bat$jRyOh{9Q%E=1M4reZsk -L(hsl%)FCDyy3Od*>T5I-$O?#!j6i2XrFFihTP>7;quq^qwV5Tc7H+1C;^~W -zi94Q=kX^fY_{F5ukXVh+u?wDs*lX{o>zf-$?qJn(1v^$dYsY>`!d*GO*E(>w|_?fMgk?Nf?2Pf6?q+ -U&2OP>lGJi~f+yzr^Lw0Reg7I}Cih~Lz*-olPU+*F|T$jU+`-4L!{R)8IN8uX%GRf# -l;2A~@efmi=cETL}axq8j$a@Bm8z}}_<1~VzHv~;B-nQJhW6u$-m?8Tz$X2T&#aX*eTs -$V0WK3NSg?o3Xdp9Yo_Oz0J{Dt1Rb-l0~L$%8M&o|6F -zKz&dNFnNim|Ls_4x94!&ER*!bi-wxLm~lwzxKm2eVz0kZ;#LEa<;A2g+0#)jq@|5%C`*1UZw72Prsr -=mce}sw`;JIZSeXA(NTH*L?O+MP58f%RBOzvkRHoLu5LcBfaB-LJsnzd>axz1a$uYZ*358O#|cMtWHI -Z{xpe+lC=`l3o&D3_{%v~v-?5nNICnJ3lY6>suc%y|>8mQB81DO#-}772N43)-wFQdWg2!NTP5tG8UZ -}QcsL)&;V*C$DcjP>byJGaib1jkPMZ4$bm1B3VjsSW!dRm3NO}yUDluGp?=Bd`eVy1neSm=$ShUR|Kz3VXgS{ArLrYZ5mL^$}N-6~T_njd{*@|T>P_#b8v@{&PdGqE -mtr2gd3r(eMVI1Ovb4@Zt7m2 -|@vbhc&|7uQp-F1Grv!_M^%|+B+s3cjr|jA1yG|`#m+{QDe-r)#|aDfU*O(4&K3b*9qBd$43Pn3az#f -R!i8H1Si5ShwZ+zDpXvgja!6~$QpAk)3$P;cP@I0Xa$X^eI)mSA~q@Qg?412A4!=rJ-O?UUMb_D9dWl -rr7KMBi8Z7TIFSiKLJt>UJRmJzhlEU>cVfyo@fqn1&=ouIBdZbJDGk_{k&&Q+%^2cse~hP3FSu+X)B< -Zo_P>|mJY3KjDN!{niV;Csj)~$xLGdq>9S``IaNnD>yw&MHZjuJ!<8^m)uGcT``rV`6B* -Lb{W$2Ui5Pr#wzP;DL0#;2x7Gx&;jOH(N|YqsN)0JS{Lo}clUgVH`gNxOb&lRQxb&)EEfaE|6=LVq>W -q5K0=glS*^7FRLb0(3C%>P@?jW5AxzJs4RTD_hT$zd5BRP=P41SXJ|4HsLb#x%^!$F^T#(pqQ~Gk)&V -3(V15a27&ZkFsk|LCssb@N~1C(zZJRb#qmf+cfS`7_FNh?F-qd_A{+BmblH-B+<*S-cDTuIeB|8Cx+w -LMZbfa#&67j8sku%p@mb*5VYnKuuiWnP;90GC)9G0HaQ{~XwrEYHWw66*G^4+@zt){<0-O`4zMzvfkG -&2GY-wW!|J&lpp$QOi^pV%Pi6;EUb}yeFqv;=V!88&YFwa{{v7<0|XQR000O8% -K%whMrTO|ofiNAqD}w+9RL6TaA|NaUukZ1WpZv|Y%g+UaW8UZabIR>Y-KKRdA&UQd)vs3zw57Toz(Dao~icBq)rxDA_ -Nz@ttpZ%^!#-7FG<#LRafsyz(m&!dJLY6zn@Lx7k`?*`18rzpGQ>mEZ@l=FOn?7vE|#_;vEN4=A+(F5 -w}dsC{A;jVV-2sB1<*s#bycfAQA@}@=pscYjyjRT+}!g7EY@Q@pc#Y6|C(UB1Lr}kQ-RQOhkew)nXls -MNt|8oKx937QuS2e1yk@SpZ;~FZm}}sl~_Sn#U`Atu(|lS^A0!_W{Uxy)CMbFeM=XE7LbE&t*xk7sr4 -zFy@5K`m|LG_C?yu7v)_Y=jVOBRK{0n1^n5e5OXo*dTL>08t>o%KV?TE?d@Oz}#0ehSS4c< -)ScT?z4@n}(ON*HQT_5U$g*LC?T{v&=xu^Nv1BW!Xq?7cXB^Je?WyFoYck=Cl@PLVXz1B1AsGVaK(T-Dn5+Z#X1GG0RmWT1vNT5JwN#y -5H!h`Nws844$ID$IK7kUbjr@_B|Coj@NslJ0$7h89zSN&Wr{2t^`>us!P$rC^6zJVfcgJ)`Q7C${zep -jnms%G&xeOk5Tq69>$_9ir=i7M26Wm=!y(?sSRm#R?R_V5h4u -;ADW(l4>2obPY<+G&^VYj^xaH9CH>w}lVtX_QmN-*z>zedfo>>z@oG9iIYZ&EET?vCsx!C;T_hCl|1w! -yXke+O(IBwIfZE19oxx@^lQ-EVE<*Jb;@OZHXc)YC#~_cFl9PtcsG%y0cxq21Da<^ppwM5P?A>0u{&9 -g$xa8^WoEXlv00Wi&GJ6rR^32t+%9Fln#YuVBLFoSHcEAz-@heyfGX*0@XCJ{n87?qDtv#(9CilImLEUzisCb%> -^jp!-?8WbZL3sG8mphOI?NV{O?7$v~xlEP*w6RqyBx%o&%k1HrS^kRAR+9mZys_1$RE_2`Prt)>J;=MJLF&`!vCDex_3w)(b>bRH -cb?VAhum|eb8M=$>_=f7(1r_NVP?4bnLqmid?`>_-Ewvv)Pa3(udl%IRPqPC7FO)S`gz{&N_N$B%#p1 -SCY8kMpHA`&8m*C4OTMSr5v`X4k%z-w4xVWM3P;`FzsaHPzRo~1e6(UYG8=sCfOMva(fygQcpujGf?C -YBMD2-iE4pkpobWXF^Yj+u#tn&&$Mn0j)i8idO?oF*A63e_yCk%-oUQc3I|{zj>R=9k-*7PkHHWPL9i?* -$HNvJ=!qD_vYjmtv`epZ@ReiOd8LlKvoZfv0yW_6Ce{vnw}+yYEmrHIxW}P|!%6sS?CiZ?rxs~#_nKM -25Ww#@(!j>$fQfdIBSRv3B;Jy0Bkm2pkiNzIb7%yu2eR0$VB4eqmQV+DCa9|~x25BG~YBgS6^#_ -J7}jP>*Hyau#jUW&z7P&I(T2OKMwLmY%d&+-;*-chX{I+eWK(hg3crD{4%f--e}K1t2>+vhUiu -+HUIDSdO{;e`s3;EV-brv6~TaAn}L{J6W!ho@2C4svHakf4l5|FXTo7f1w{cJjC8?-qax6?&8bMXPag -j1_`99UPg{qaf#)Zx;jNIf}JkW!&HHWYq$y$MOIOYaOFziw%TPEF2XqVc>;P6#Hrz0feV1 -X-?N%7>7f(D*;SXxg4UciP`;#4eH?>O_;_y2wyg1zRrkJt9{g11sK9BgKwTsIW#W -R1D_0ezzo?dM?&x7PN%wm`;WRz+TuJ272-OBp3$r0!!gwEyFa*vV06yFec%u -vnV5m6PTgfG=Kti28l?e=BtsO!-#4gK@e~HCwyEIC5fQYgYM56*avD~;kr;IVH=wmF4na@E)mj -&4v-+-@le+lP1@P+x`Idh~3&YqwJFfg-6?h#x~9nsZOHnDKmDjg(Qy)K&TH3RJfNG|AL6EPulB%>JTd -MH-}^fonDf&@i!F(rc=?tU5_acp)pl$kIL#? -zpUdN<0IT-0em_rsiOyZ2o;X=N)&PXh+!{J>o~TE;rYg>*BY*SBB8NSLExa<7^Qk(qUM9WJsArRAMDhfiUP<@kQ -fJU#@>QdOl*a+r1X`n5O#q%HadMsCNwJFK-||f=`=vqqfLlKM{)>ECiKSV9!tOLEhEX=$oj%QYuy2$( -DLJ>X*(L@Y$n4$S^^@)OGW1h|aS))I4O3li1`t^j!n+K+*}gMn#Cjar@AmJCo=_#|OP5WC$<3=yq#6Z -TgWDik%pg`T!FRL!>haipuqHBrl8OneLY;*j8`B_V{?AfxWHVmO3V3IAL^WH$AESt-8w?E_GETpxD)zA@N98mMm -%RbqcR1Zrf>+>ClPWVWuI4r4Wp~%uDZSKdZj9-mcLag#T~du>x+ZgM&9K_lRj;@~pz>F1^EEdhC!9za5WMrJ~99ckK_zNH7%*}BKjhkgOm~_%)wNY^uX2-^=G?uyGf#%M+(Zt_n0 -%vGmN11*>vRYBzn?6+Y$}JChqJy~M3*uAp=szQm7>DiG)o1H*6zhoNDzyWLVRI*1)wNK`xFm%Op -uWnv?Z2WoA`YX0fQ`yx#}i -j{2V*7rgQ?8IUaGz;mLl{l$KWtB8I0j6s_7Opt7N{yn2TK2gbs(r!lx!;yVAK{*|*3h?n%UjQ5if>v^K2 -YqiDCJ=Fw+YS7ynawxRw@CQkTb5NN<|)5$pS2r -lx)CmsQSln5ZJJX2V`gS6fD?r&tF72SErMse3eSTS<5zR6-`>h^PScW_I|0@;XyjxvC9qBIIb^9(_gu -f$ZeIB3JkfJIM)U+*yhY!1daTCu8I*bal)8xqqf;WE}WCJzE5LXv*CcfJ-wKY*=AP)GO0-9y+7d4^g5 ->_By^=Wyj~Ye)j>XV4|05TbPbZ$%<&xBb(X)^n1qB~*Yp}ZVtITB;PN5k938a$jUtYq#2w?0T)@XxwROR;G -i8kJ+nP%%0uel0L4>Ba9%_7OR$qr9q@G30O4eYBbXpCY)Dg5_i*S1DnG+}~{Az`Uhc-Pj)N2QB~$bXk -RuQI5+y+GS5Wv|B%yQ~VanePpL$@lYC_>QehE%JWigxSq`^Lm6x8KId`O3pEA1U5CvTG~($?6S9arM{ -B(Uqy2@A$d_uo@;?wh_}U(6)+mge56#${zUHk>J|EXj`VPOJYTBJc*Zwny>kz@b)T=2T*Weu?1^SR== -Wu>VY{8PNL~(Cmm&7?mn6>@QrGWGqj&1E9-fhguD$DXb1*XJbifW9^$qBSn>1k`3V|LHbo-zVp)d5oG -Hq2DMkY^2x^O}0&W-iJDY=Qde`sH}|A3pHRA8*$Vkj7I@{Z*+jw4!jDDY4HVanfqKrM%NuE2k!m@!Os -#|E&QVZ<=o4Hac|a>b*FP317^5tWrir27=G~h-;RI7{t(>Z%jBp`%OmV_gw1~X5UUydLh+5oI;p^r0tW%f?R}v@e^o#Z?tf -K)Akc^~VA;$?RDim3g%kXW5j?pHqRsJdEqWPabFke1U?h -pGAfABwi@a@v6I1gn6Q}`;nV+@>}-|5l*%xlrU@{(>y4 -^INI>Q*_-s9;m0tNZPbnj5Gget(E>eV(q}2H8wq%HgaiXRy~Z5cSLe@`3ApT);(bgEltTjR&iY7Tbe~ -BL;12}R_rho|c?L@*Z~db|z2?cyj;T+)uuZXSQ2jQ*m>C|a6ll88dSFXZ$bV24gZ2^J!WNAg$AKh@fvcm^KLM%_54OtHB>x|C>m;V<$jP=yT$S -ihQskqt|;YKOG8u??)RKtz+J+Ol;t`rL-YJ}ug=UFQQqUMvj9?;gjG*)2f&tS?#H;bY)@5DrWbk_QXU ->_&`;>H21bn5tGPq5}ahjgOi70E5;S`U>>eXxe$P8G#0s5{TSK-*9j(m-Z(iphruKs##&f8E6e^!Nym -SoYFu;cW!uUl`dhEOwIe3d4D(?Kf3Zp$Ck5)<705VCT!k6=(mTaZFa)CUsTX1)?IIaPC;c}=f1m|#0H -uw6V*JTO;HSnvxf$`ej*lW+V8p>4pH_H!sFUN^Y(`xJm!2a)#*!WO8n>E<-xCo#m@X?2}-8SO|JM-*A -8SP2z4gLPoRm4tC$9kjjO&Y5xynV&@~ja03-s0j8re*t1m_9kd2wz_rRR~*-L)~QiWA>Jk%sBP_-pU5 -|c4>jM^RX*0a(ODTxM1CaB_F(J-+l$t3N(+R?sZz|zE$HZb6)LtrQ-KMSSY3y_rPI${s}T?>$d*9Wa@ -PUdSy`e`p`D`n4LvwK!{l&H%;tzN_4zW`860|XQR000O8%K%whCl=OM&maH*DS7|^AOHXWaA|NaUukZ -1WpZv|Y%g+UaW8UZabIa}b97;BY%Xwl%{^>cg??$E0lAZY;;TYln*q`D -{D2)!22-pdK{cZdbQE^|jM4hW#|wjsAAhFQ%#9sLwyw$8Ft+i!IKe>Y7GWG+tJ&D>aR$`td4xIf}kqs -}9|9JeK_+RLj4V%_IhXcPz(!Q?J#l*YK5JjHe^aS3Uo}JzdBS^)V~1MZ4+x>{wPGOI%iVDEp71A5`-{ -(${kxdhz$E?!`g0qwe*_JUiEQIEt#?*4A)8swY2+vpo^hx>HrQU0YXWQ~xZAwmgV)J+EOjz^uj~fBgk -Lm;mL#AByW)0H3|7i@A_*97_F`9z)vK>Yknz;RW-*^2gdEJZax|;?9_^j^faSU41Riy)LC&pr^_QFBEZhjDBxh -cm|wO(8#@b^PGh-dXM_Vs!qUZim+^n*7X*Sdg*qg9(M -{)Qo!io0lHli6ao<}kDjq3Z1p1keP^=j(&qiX*OKF6ANN7(W#6o>-?{Ov@GEkvPS{30lY*wiD6lmMK1 -@ld!mvrtATSJ31`rb7&X-WQ1$sO6?GyJhSS%_=#RABFYssHK?;41^E6OFBiwm?fM4#l!;K%@#ujtvLAT3uE%ZrPPswn|wQkq+O_lgLZrYZgV203~l6+LvY -z4y4babJ$G(N4))sBJ4Y38(}Z7HtoFwJ)#n<+ke&v_+UJ*WENGu)}+ij9s!7W3>lp0{Ep7*JTR}HxmJ -?YLk@B(1E$rjwK9)E2^smU^091&^#lN3}wBs`)7I~%5jul&v -rgD!{=&9DS95qJZ}y0)tLUJYFShum^!KFJM*Hdz9^0B-LeCXyQVUw$+uis^*&2Lx)7z%nkH!<|9g>qG -Qc|1%jX&-@$rTFqeToy?;M9NJa9e7{oXN=AF)rkq3c*_dwBcZl>ZBya(9Eq9usXaIrM4rU5p`FddIgJ --=vBI|rb-)={KihFI%ma`_$2^h$SzEh2j2Ny-3r4Z -YUw7_~!1J?^s>q&j|Bs0ZbGpzj0_p^;6oT-uz}^);xYvMJb>9V{FtZ{NIPWZj^D) -OXilhinubahs#i$1fH<+vrFHD0^td;DCP68u46OduB**V%deae!a0#-QZT0|7KvaxHx%~yng;V`TEP> -(sXU(#BN0}a}wtZ31;aMRkondE>d!)lC?P^VLJ;d430+8&AzVo7F;AA5Nm^tH{DH(20EdW90}8Z!()5A&ab2&CfROIHQF -e8X^aeMU??d?j{pPsz3j(&3pS}?Gc`k)PL}-SN(5j*5RD^l>|{5UJ(wnf*99gI?0T?0fB`_=f)|e;!0 -Q~mH_#nhozQ -J7xjEMPSbNxl(NnBXS3y8zhU6qnNJ`>tYAcMBC9}sQ$N=9S7x_~^0rj`&x(r -i=SlCXDgB^D?6{euZXro9=a&{PLo+i@RcA?p3IP`VC0&h6qk;jwUyTCrQt#LLI?I#ai@sTVfOJusl^X -y4Br#2Cqy&VXyT(slP0egr~@Vu%~L*pwkpq+%OJ&N#0js<}F{OPNg0B1LG(75aE0a(9-LW3RO9giI=q -Ll?rWV;MwGy=TpwyLqMln?;PG88;cb=(YCvh)4IV-M+qK{NHyZJ=YbS3y*km?sR&E*vZB1x%I9kSD)c -MXl5up-@Y&Rm@C><_xHvK(8LUq80vEK{huN9=PO)(E-$VG+$E6;NY`(GC+i0ws^kh6yO -<3gcb)M8_i+gJ6#VSEAf`DO8|M-3&{Q2093-mmAY=j_}pSh91@tSaP18@a=WHBXf3+kTGPy>KCX@=%Y -7BCwP3XSoY>R;Q3J{1)amy3y%3X&WC4Fns}$|3I45i!{^8SAsmdYBOqQ45*e^peWZO`UIS4=ztZknj9 -oz;?4k4I-Fttfl6w^w;b>Ctz{b)nXEU$w%pRe8A-iAA9E_a9wnua1`yn}?`m&{<(r9!}gmxd9&F);NY -k^%0yQ>n5_m9dJrE&+uG3P9RqU@`EeJ!M+e8t9NWX&A*0%8=!0((-B1HKwdN11}5)kTKSe5iM!x-Fe< -wrLr0DwEIiWSg>|avO^iFN{%%zSCXm8izBmqmvDxowFDLLi%m!W__0isIu)Z6+LhRM{}+95!ie;{L`& -+CrH1Thv(8DT%89s{q`V$ZBN$XZ0dk}bE{h4h3mJ%p#F-j*qGOp;)Ba@R$)w4GO^eH0Sk0OYF_t@ID) -Vk@ax=qY?QDVzBNtnmIg7s7^l8X27Aw;Dw*AMvctj|Zkv94%e%Xv=zLLRDq?>pJ%+0!^}5^wTpr|3NB -{C$Cm!^x{?4NF(oB*a=VBw)({5&>X_|iDf`Gal*;w0)<|xB4J2o*%b%_-{B1m-Tpbil5uN+wk3fyAXW -k2v(`aLd@ytJ-|Ar_{&W%SLl&0s2>f#C;~tQWZwzK{Z*OUoyE{Nv -#<+0_gzfVg3a#9Ygn;sNm8992EeIhz;LVMiSw7;jW|E^*bnwmH -dhl+A5MMb^xd|if@sBwH4#oMxDMu`$U1)bxff;YjO&`bs@4AzmX;z~1BbfAKmdqmgkCU~^xo73((=m- -#1!u(m_BG~HbQ?y?w0=<=1h%!>tI;1_U*5GbNSFV^zx~s5GEDN6Ag8zxgM#ZY$sKne6$c3CHP+}`e%u -yPY_Yks+PNFe{u(N351d*JhG}kPXXy9|B*Ml_W=hy%Qh>2HtwiAZ$)Ce2&_ev{}?1&zFUO@m_uVQ|V# -4?*Rb9}=C+{meddy40?*oi}V;9+9aQD*yOMxdc12k#EYMy%*;^->1)^888m6;6-A{LXp#oqPvV&+KXZ -B4SzTQ+bSJ9`YIi@=JGFCV8Gbp@3Nn-t(d!k%P?ZrOfmor#`t&bCQBN{Bvg%4|#=!jF}O3l`CH8xWJX -pVT-si5n{le#gztNoK5v8dCjK_Od4p4WE#q?XikX$;X9QhyQV6j=k~@;jUi@Q0G~^p7>;Y6c#}ZIm8L -5v1nC42XRSd=0~jWhS2l!9O@48k(&*_G&*E-L+Jx6yx53nABv@!*$(dmIXl}#o0k9pjdicIUG=6NmoA -z?ob(=&c$)G}X_PK;>Z6oj~F{`c`>uaRk<)xGpml6c#PyJ58b6|nDHlk`Zl*414O~-pHy166@a50M@e -rMTLJe=g~v7d%frfUdu;*EGD)Nl$=3^^qMx&)Jx4lmhAz&I7~FYwIaB=K=cyAfu8i79D@qH@*jV;byQ -i~EC2MryJ5aCr&rHzeY)78^Wys9Tww;lAyF5KZEDs9^+AUftzeQBY*hTQS%Lx|C6Nu+)MJ($Hed5W$iDpQt6vlAK^r -4io2I*AoX`~I>6Z#!w9{rMU}YVMb3ph_*>4i!Zpk6>kamSsyhmb0mSy7AEfgQZ9+IJok#oloQJXr#S0aIuA4OdbZe!4W&2+#?uo4 -bHUeBU0z!+JpeK0b!1>+k7GHQ&*(UQBm%@8=de0OG?QdCfKRN4*!s+L7XGKSvRh -9U*I7gL-=RLCjPbJ5$WZ3Y4Uk2YbcCFa9myCB|R(H@=rsA0?YJFA~-_nMcE|5(kNkx*JsCt6)^*Z=I)7TXh&j{A$Kxdvp3KGv-sid% -U91<$*$?vm_P#NdJ7)NCMOHf?r*7t2ccfKmSJ(3eA2aaEO;pEmU0}LvJ&P{3r3%C8hAp4S!*qu=?-=t -sWc(V*X5&QlUfD-1SMkOFq~h;ma+7z2Oi4;R*!0GXj==FN^*HQ^p!NdGJ7$UtfTgcqBb3Swm85Bw$%n -Kg5>Snm(Nr}4Tl-@BrRy9Hij-afIg_vtdIbu@j{AoOUOY>N|eA;9U1|KBaoa3g=JzM*1Kp)OFf2p$ixo>5{_od|WwsZe?o#LM^5uX1@j1A4gj2K!!7d%~9GXcD7@>4OVGx5n^ZR76-* -Fnz&-O{7KCX1;O_r(5O5@!3g@)t_cuthgIQp4BRhfq7K=rdBwIrh}rC|&#HGt>#07e2glQE -{o&%*;$l@4eC|xjgPn^clj`n*~#}iSbc0)H8)(vklGd!g%hY77_C(&KzBWWK%)%I+{!~xfka*Ej&&UD -|UvPdp{zg($;Id2bkM?8$mbNi6bx1>~?+!7{MVzpN(;<8vOV9osh1luxwzlt{Pj`8PnC< -t4>p9#HKMmzYCZPB&Gp&nGksc^&m(R}r@(LbF^Atm^XR`9XlNZI=|3C0MWaeR~NIdII0t4V4MDzw5iW -LS7OX`X-$@ffrQs*V+FkjH3&Dcoeqt8Rx0v4As*%uz0AZN*8Yz?Tg?M){9$!;*O*1FxVAyrZ}D*ms+$ -vJ*_Gy-c_v?0yEW|BL!la)OQNcYh>yyZkeBp2{7?Rk4qor=v3$suaX?z^NBJB@y_Hp -DGN1(pZo#g0t=xmSWPDL$1y5_$2!$Cq>M&d7U~QYa8cmFXY#TrRzmhJ$l6v-OeCK6o25E-iMfcsYuR* -vqr+5tgfzGQHXyV*R&JW7xYEi)?Qb|M#2WJd0j%8bbqf^k&qy#P~SUA>P4HCxoGgV9Fa=$Z>DqMDX>( -~pd(kj8b@pY#^?QkxU++2rhdDLHb87+{XlwWMmyKbp@ytNrPOgs%d8o~*sBF_}!EtbTC!T30KAw|Xzp -G05QU@HY>)S-$}f7|r0n2p6oKTlaY50n>)LxLe)m6$$-0a0w6o`D9&_D6JQ$>c%0f(#Qa8{oI@iu&?~ -P5ZL|@tnu)7s(y5{~nU8@*rA;UC-CTJQ%S0d&E|;mDa4?MT&xx1_KnD)O;Bb7T4B6nt>)m+B1{7XG@$ -zl-t#zhL>t!G7Mz37ru=ie$xB7GaR|sXVS5x@gMaFoKgcW_r7NioXtf!26NMq-Zx1m+yw_wwu6=AIaf -Zz)oh%rif0@0(8Gf37t%Q8Gg)+TbEPbyRUVz;_tt{{t -Ch*gM6N`*XK|JZ7`bHOj4%8OE50Z!=hjMc<(+Q%J-Nu0Qz!z)I(PvqIuuhiTPAM|Ob_x#K}E^@VGT^? -(jX*`{&~qF6tG%Q-md~S05BWbrO1Tjn#$D-?4cpQ4s2rJxMNH1@Ua6d<6`SY!yQ;}L2V_2 -L8y!LX_hM#oYIrV&qK{rvClv{r-FHz%h|7pGGr#|wwhAS&hlH2h{t~3&(J_gtfgZ}X9#4#wGptUzVlr ->9)xF(HuB|23(K;hS3Qz%F{h)4kcA7|Em -vN~CURl?Kd?qavQ#In}=Q_y~?%97qN*kJ;6ynG)RdtNK@4jAizJUtx+9?UPpAfY~ -30#8LHPij8W9%!+KFlRXtQ+gA#sp?py4A2{5L+1c3e(hHOG{!kiJ7Ub!W{jA*`3a#(kQ{exBVnQ=bh6 -Ynj(KN>5`mNgEDjv32YDS6J(soZHTvZjG9Qu6mT}0?ITqh!#J5bC}J9{Vs<--qi>|MzB<#&c%Eb{AQ( -Xa?I25bJdyTuHcWn?hF*k8*x<&@7SXUv^EcykPI+zZ8kcdyO+xlHoRXW5VKa|ia_@U|15h~~ -)pqLuge#M-J5XsQQ1%JsOH5|Nlj4ABK$xRlyY!5}befpg-k%7)qwZULK42 -TC#F`xIoZ8b*j?(Vz4+9dD~Wb7=>=aHcAV5+z&RsbfS?I~0$|E}xZ8)iD0*M!0gWlk$^G%R -U#z!&(Dqsf1H`Xko&Na2JbN=q6^^jhmf8~<&qD}z;cE0~5iV_4??zTRv^OE;9~`bOxvdo-+pGQNb+m$ -a({I^+iE3T5RQdGVBgY3ks+<7FeRMT3DCck`%g$o!fGRQd(wzV-*t2$Vy3CZsV9{ae0cWiVDyRHt3+T -}wErrXKB+M%SvvFFw?oZcP!Ls51CdByRgT34jA8_yDVYdbRjo{h? -ynT^HmxC2rD -ow(q_+tCcs&!b&)Amy%pm4fHTmo#}-j^z(mqR#upd<@u-Sk00wJ{dNX;dsv=0Ds#tN=aa5XI4~wmT`N -isfHC()`l6ex)FVgbTyf4i&d`biYHA8T|H90>qJW2vG95&KelR(bbVCd -C51Vi$;Fc-+_)2p9HpV2FTmBdCFmZs$BwhUH9Twtqvn34sB)NJ&*A(bOH5ChMZmT5yB8{?|xzWQ~ -|Jg{MJZg6GB{Y?ocJB2q(AcL5pNdOb20Y_zL0B!!9sX}-N`WxPC38#Iq(q-d*)OljG^g%-DyFU+)|E= -uG20+3g@%%ZwN*2Vy3|$vrmp-0nyjkrhhK73W(yprE^amxRX^i!h1ySkR-3)eFsJ_Z?(+671B&SDZnR -8V&Bpqc0k{}9SOqTjsFqZM#d(D~HoT?3PBQPwA$EV-BIfsC426=pZk1OHC?cv-TjmpibpV9AS{Wq -pKyBfi06y`*2yfJWtVjX$;!F_{->jh9{A;Q!Y6zcQT-HA@tjg#kTk&%%vT)Xk~GKY3}c1?VOM$X8O&W -RZ_n88e7l29SVkO8(mo@0rs&GN+9gMsoFGUFO$G&VY?9BzTX1=@0TQ`RRyC7`Avcl1g)!UHJpr>jgv_ -hcoA2`pN>-tqdaW!(ti1B3g35%;VzniMzj0ZoB+*UV5Nx1{?7>p?x5KaV-1q`iQ46E}os)J8zKtod+7 -u%h$!T=PzD;|I>4?HjD07fMLG*^=xK2sY)}wpd8V=oF0ErAh?|Vo}4!q^K_e8-An4!JzXNUBB{dHmDQ -9Pc~dqY7hyQbg22`Pt%ea#3dj02z^N3>9Pqg);E7H4?){mn0m9_76%eO7Q+Jxa-I -y{gUK7mJPT@BB`)vr`Us*Eh~^Wou#!}oqa!-D?Q`e7TH?@F2O+(o8G;zbVJ#g31tA1PdTv@ZwXwb;>u -H55*tZqQKk$5J%RT)|)-(u1y4(^az(PfDlBmHN=klqV*Crl>a&Ho3t|BGhe7@iQb&%pBz{kZn>{75MS -(sE*_DW-vlL?-@FLBd`%zQ@iOZ`e^~Y5dSAm{PQ4~V}AMLFKsLdIc>UjCw~WlGKELFrH)qRA0zo2km5 -EtUY_z7l?8466c|~bCk&47^{-dsa-m&Xj%Cl$b}^hZ3r4z2ewUaNV*FhHUw#Blu^0{ -ysQIa0({B;XZME>jAcT!&Oe&lJ9yr34JAlS>`;Ox^A>Kh0{g3e7STEZrEw4hnPA2c&+ -|9fG$458P!NilxTQ2QAc`5u){s3>v#DHG$mz=(K$w&)l@gfgUT+{KpRUM_lF$TkR~UxhG%yQ<4H6}I6 -m)hjhhx_oVBj1zTo|io6BHjXqbGF~_XAC1MW5+|0gR1w66LfHM{r{o(?}K%fu|ZmK8jZBfww0){E+Bq -Y7|j-?5+rHNTIZ|C0ve`Nw!REz(g_O1(C`+!z5EB80xwvp7p{#!EP>Cxy+ffOUJ$D@+IyE!JJ{}xyK% -CyU~(_BupL^ml1M`j_nxxSbdnWuZf20zeG7N!aK6AdA93hElYOQ((jI -*Hk%?2X#|P$c6%N(wMNOK!KFco2yyi=3{eaS>9ojzGI{#dNjC#8FjU1SHID1p&MxwYPvOK2~lX2V*$E -8GT)yGpZu^EsV;QDe6>^ax9xjc>ut14Wnk>@#p`tCqSAUlIfE7uWqO1EQ1JvVfX&#%3OAI$9*&-X-d& -CCQf=aRb7s!6!LoW34ogKOBc{r>C)nOaRfYbNu$x)(U&>OzW|;d^eX?eNWYk?OdN|4tlt>P4FAV8Uvt -F3|L(W!wPy3A^uFD;*W#jcQYp7jDHpY>IcRHhWi5`CsvUqw4(xfauWOg_WBmRO0@;F=Z3Uc(Fa>=jbB -Nf<^onM+o82lVhi})VYtkoDF?jtYTg+UCB^P2rFk_xkL0i%i(Vdo`u*o_b9-l9mk#@(AFLQ{+tciI1QY-O00;of09jkubWlL>8UO$zUH||h0001RX>c!JX>N37a&BR4FLGsZF -LGsZUv+M2ZgX^DY-}!Yd96HaliN0u-{)6w>2jHJBu#`ZZx>=nq8D->wUlPM3zP6u5G$L%Imu6^S*3qd3L6rt@rh+Z<iDvhJuW9-*Rlry^XTWg{}EbBtlJ%HFu&)${Nh(IoH=26{rySys@EeMHu -)QLLZ2?<|zeJ8{;q+aaWtSCFf{NmSq1?&F=&tH}uh+NZwI5>Y!L%S`zexFyOzSZ!$eGMY8$@Z|TE*pM -7JF95F(Tid4NtPkjSvEU6J1fLG%H*mm+n%1v@N&ujH5dmZ8k@dTJQp#+zFE64cDA?I7TKw0bB9} -Da`=%EB@6`oCjN|z0PUO9a`mF%vtg2)58z5P(OHo7!I%?MZTUiMit!Qu&Esr$)9<)?8m6*e_4+>!!t( -%UfY|2|vN7vT{2;%j1N;BRFun{2V_uv2hed5pZG1^<#o9hvXAu|it+~hc;14sP21x^ZLgBzaT8)K5&OFQZ7*0B -*7b$4QWQ8Ye%Q4X0=6M -Vk7KTY(eO2ynV(URN>EZT3$R~aiCA%K62}^gdP(>HmKb5&)MJ#9bw#Rfs6KQnZAHiU^*+dG+-VZ(c4AWxs9q{Q@NBa@Aa&Sp -a!v-x$cO&8423M&AHb0S+>P*{qP}$9V)qBVafaeqVuSFYrr)gYcn5H7&G>OnGVJ{(gkj--@1SKa2YO1 -`SEx8Pk$w0F#_^7AfQvh}Ua{11K0Vcin(31;d_~5`g4&B)Sg#KMltHM{NVHc1h^3(Jc_+*w4dCufKo$ -(^ubq`|1sCQ7y9V`&M3@pSStyCP%*nH~=NZhRQvkcjUrh%tGx=6bqgwsJf6R -DolsJTVM0P9C14xuD-$d(;&3WF;Gjll5JOXrSqNodrEKFiFd7gq70BT)Z%gDAW`vkwbnv6;T!`AiX#p -3T*&QBpk5iU*nB*JA^u?sDL?js>J(~#JhQ{1)UEr0>l0W8{wD>4~q74eV3QS*!_=k?)t)M)U)eXi7SS -Jov&LQ%cnEg`ySn6m6uYq1hi=DwO;PL4^K*QDj+#!>^@F~FZ=^i^9iReAyleoOVfs9( -M1BjpT|+1V&uF&XLO28%5xG!9_^D`Zj*4~|n<=trA~>!A<&j8$rXB$oOvQP0BaSa1z_Nq~pzW(7lQ$) -WF*mz`z}tskDt@(`MkzRuQxb)?Q -rgux68y9n?ZOF&q3KejNxz2B>5lF#iEj)KOecQmCW}BYG491|%^(PBRgLaUruRe*0SLb7jSFT^_| -tMfdk(4V=~cYb^ujdtqTtmpa6GVxu*(UaNKg)M-nIwNQ}WoGNoVMd@12*+&2v*k~i`zBn7wmC>r1(`* -01W5tDZ$ZZ@lofj#N?BQBC}uRVsxk^>ChUYBBlk;K?9?cZsXCNUNGo{wdGj565e6$#l^Qdi}r$xD5iw^BUrf6m6yEHXYNS2Xg{ -1ap-8fPcaM~MK<*>cDVqJw+_DuTH~W+Ol -Qrd?EEdMitvRaCDgvvx<%osG!?yMWI8LdTBGEHpEq?WS-3hQEfK-sA3AsHb0480u4I|!-R%Plp?I`wcK!qaG -nZqEIb^0JdT-He$Z?z=Jd2I1EKD-j6!bkwGg{v;J?7^Db}vgg{sMg>ZBclvP&FzSZ0-M*$|P?51IBJM -P~0DRnPY?$^;KiN)t~9&+3UGKRFgO)m&i0>K|wW8aZXcQP@z5caZo*;;!#<4OKxP(b<)oEtMr%WK+Ww -9!o}-XAPuLE`%$P8JX&l|KLe8pPm4h>Re@QW{s;Cz&jfC0iKU|Kl&BnXQS*zW4<1$Y2NBda2*Y$ln6H -=;N$7|K}8q^5IkPjZG6!f!I{|XljxlwBSAXOacIfLAdS8%u&RTM*9lG#hEZvfSGA9TC19rhJ_M0OKjt -jt?hyWmCFQ|!HWtkJurzoR$ml^b-}brOio&m;7{~CH?X*>B9W^bMnCKJ%O)7%W{(lSqaRmqLDU1PweH -^MwDsbNonHg0Zm1%%9rdpH3a!UzMqVIAhHbnz`q@oZ)FAF$a>Vo=}TS6Z|5UB`9s*$t$1%d>`A!+78b%L4jPzrco{msZY18o-hRhKnh$O_1(&CkK_@FKN;q5D|@r5A+C>O*{Sr;#2Bid -FK#uK5-aT18DQ*NNf3_H^yqWU4_I*!9yvKW7ct5zivKd&(yW(~r~i9(mQYi>($M8F-0EYMt_Nb0kgQE -WZ<`#QulB||zfCic4pEte!_)D|){K%pu!{AP`J+&|Z~2eJ>GqUYuZGD-w9`r8Ww)n)TYqCAK`PZ7e^X -u=W*b>lo$qV<2`7?3R{h;+Tr2ltMTVek0gAi>TN4}Ks*EDNq=mzXcrN5}-r9uouN`Do24D*sh&plKoM -O~`-ciHLfVuxX{u+hL}jfKFe3|JyKoweQd@c)6WZu^_kVDBN5_nmtl4|F||fY7{Boq&C5+SPAI}IudO -}2aoW-RJ4fXafj#OpxGuHX`=kL0k-cN$XbuOOr{1#ecplnkg1K2qnKGK6?2DR3;t#~AmmV#(fRHvxjFZm1EW$Hb%Qq07Y_&Wv)V)azMR|lF;Wjv3bx0YHVP|dLUa6~e-LeRUaWec&~Z<6?pXsdjMOFxy?s_r7ebkR -!;jewyBdA4dofAt4O?`wD9Xq>2=B~Mrd{o&rA6CgjNAhfI#qN2$ccatQgqv|C7CI@<^l8 -zB^Ir2)kpIz`gap&7rrJ(BZmk>NcV%Q%C17;hG_YN!&(d`h7@!lM!$W<=i^=?%G0QAgHP}S2C85jVZcQC -pV&oy)wYFkNX4y%M7G}IjA(iW25P89MS`${DET`CL16L_B2*iczjk}GP -?s#u+}6KLHa!tFa;x8CkMa1kJw)TL3cMya(6Gou}1G(7YqRo$K+C+=A5pe$>#8r$Q!Tqq25W!^T=TzQ -Xv>h=PS+0f0#Mpbem#ziNyh8;?KusvVpMr#1^!H;0PA#u -+}kz?m7x*8rDtk$s?`{5mkxj$p_V}aLBn_GI;X=3KEygC@Nf^N}V}vIYBd(Cyr8W#As&iT5(j5qY#pv -QIMDRJY}Pd=m#cQHqq-Cz|T0ox2>aJ&}XXJOZRF7ICNK?FVRpiz -^fS;aIuYHu{IlIWi<^(1sj0Z$dcX<|MR(QD{>jHnp+LwHTrQIR>{_f1Ca{@spymjO{SnDgjXbjt=Evi -&$5kH6uxjeTf|eaDuNUte>3pjw7Qjs(I=VbyuAuWeLd3P>iJ$eCN_)ZI8%STPy6kJu7_o!8HL%f>u=D -xamW8*^0O>BD;46s7Teg2U7f*yRUm8g>WGLUzJD!-?E-4n&sRskn~s15h6p{!=yMVQ^hew(_%mvhY-1 -#PjiLmhm9Y44wbJv%=;Ri!Yc!Fgmb6bAYf3euE%)1<8D(wBfu`!c(sVOxN+Y%3gl5RTzqszwNoGnH0(T@KvD2acfai3qk?aCDupQLPopPHvPdnC$6rtrC8ewF -uY={P;@g6A_m5+gVQPis5Ql{#axuSZlBA35?o=roM+!=qS_I9Iq+H(I5>eR4#vA>s7IQD(0ro#qDt7# -t|@`n?}`1Muj=u;4+20O7()6nW5vJ5BwMjiQkc+sdkNcZo{8V4yVYSaY(rJ03;QXB8c5VbEpMjW`H3E -j{^lHrwSzYOLCBoKy3HyL?J=uTJW*j2{Ob;>HU4kI?n7f1j;XRd6UFHvQkIMzs~E~e@6f)iH|Y1zT-d -fY=(yg@W&r2$h}cRNGw*{yxxFd0Cu+}WUq*G4#d!9NHAYcJ|n>LZA9nF8X -gzp>jrjTnrlwkWD1FXzN%ukWOs5>!GD6r`!kRbD+(P$M8>V%;vm&rsaRd_uW-RO0+R>X0{7t^O(*z@0n= -UTYgF>f5qwYAi~U=KOI|oZ4o<1enUopCf3wo~9|N>b}uhCfl -l;*fxntqr+|r9fynJ;EJPo4GfDTywy!DpLYGl9942xdE$wW0X+_1uG|s>;|oh_}6zXnS;{izMj?Yc|*xszbLZr1nHLo6ZnAt2@mK8>>OY5h~-? -Xk7`6#AdysftZT=|8t{)_&fKhs~J-DgImz(FhI43&+#ff9mGqw_DaC^9oBUrb}JxsUt!{=l0~&PH&W -yfp~+sgqoFR~nd}lk-FTCl6);w^Jz6ZVJ-ZSM#n^?o6z<78xHu(A|qaGt0lN{K#o&<*k~F(G;WU%OgcS^i~t^tck9Vhu?QId3}6pnVcN -8U+~e_hK9HWQg)Hcqqghb_;=`YYP!>r_p7bC{MplsYN&*}6svsC9b0rFRbhn%T*_^3b;>fVG8L)iP8O -f0U#vd+Y|$P+PrpEb(XI4;?wsE}aKHTW3!;aMPX^-}JNN1CsIN46Ai>504=4;i*9X&u7xCrMzxk%q7^ -8%Xq9%khLEg-bB|;K3wFR#u#^MyRcXYTr3Fh*%t7%B>SY=0|@At`0t~8GKV_ZzX(jkS#;!}}+fI4()= -wm#O;`BU@J{b|^1ByZ6ETN3y#`w07F78ggge2hIrp=XyikFA@>iOfVAEPh6_*4h~CLRTmQnz+>BnUyS -$qcTqDKPYfc}X>6*>Jo-hw!Tg0|9+iM{AnwkFdr#X}}MK5e+!7#K(O`J-C6hQO>FwdVf=>?R~u^Cxwn -E!RyU@D`}i0PnsqzedC;7)U(@JCDV5K53U+>7NHKf34$JkTg)LdhZkdAxa ->Yzqglqm_^hsIT-3V0CF(;#uP$yckiA?MhT?<_YS%bv-pqc>sVGkuI(9tp9gHrM=>#;Ew;4Tj|`nY!f ->5M?jiHUmJK@1-O{5eubY-mQMUaW~$z9Cm~^S@mBBq;+LhFx8NnG&W*-XxR{nWSuI_`F -p-XWotS)~znGDu%z(O-X^;tOInC3s+;_J<(j1?9$yvbq0NqA!HT!c*;Hf{Y@{lg`-hSVSN>-9lwt|;@ -`mfAI6ME*C`|%2mPi^(SIPn$dQK+<6<%4BjQ2bGj4x6Rf;B9OJ+p#Ya~ -u%ahBubMQ>Q3U)eN1G5kPhvD5c8)ljUOz4Tc3NBxql5A_k_K^D1WbBlC&?6G4g_NyjB1NucP?=dx@_`dM~$mC*m~vCxm8b9sQzfo;!YX;!uvM#Q`VSC^GI^E9kGw#b;N;nT%Ql9;_gpPU3y -PUi?Sw_h*@jnhf;)lW1bL<894r8y&agfIoE$Y^*TK@9@Lx)})B)SsXtOTaIpF*!Hjq5#YIa -zH4-*m$GI!{}sj$$ogq&(Ia(fF+r?xGbr@AiGg)gnVrV~i_CedxaIcYHW{bGmGhm@CNJKwVD>w1kjBI -i!1#xy|T;dnm*{jX4QM3T`JQo3dG|4>T<1QY-O00;of09jiA00002000000000a0001RX>c!JX>N37a -&BR4FLGsbZ)|mRX>V>XUtei%X>?y-E^v7R08mQ<1QY-O00;of09jjk5xxlg0000%0ssIY0001RX>c!J -X>N37a&BR4FLGsbZ)|mRX>V>XVqtS-E^v8`kil-kFbsz8c?usv>Ue-kS+Ao_&_yg-sqj_;wof#oES3$u`*Hqrp@NgJWuX{<;IxOTM>A&f?E{w7v!|>Sz!%4D2 -ra`O3>E{Fp(CKUQ+N*^xnJ#$^(|iLe8NJhgESsAo}Ro5&6E~5+2q!q0?y)TPFywq?;!A{u!$mX5z)tO -n7_Lar?H5UJ%Fa%xH%-r?(597=Fv}KbkL4O9KQH000080LuVbTL1t600IC20000004o3h0B~t=FJEbH -bY*gGVQepQWpi(Ab#!TOZZB+QXJKP`FJE72ZfSI1UoLQY0{~D<0|XQR000O8%K%whjl0M*3K{?a{9OP -5D*ylhaA|NaUukZ1WpZv|Y%g+Ub8l>QbZKvHFKlIJVPknOa%FRGY<6XGE^v9RJZqENMsnZxSB$x(ELp -{cD|g9!-!D6xaR-R;A$+ -kR;cT#oDIMiKxb@l!rMP*<0(<@Q8;#8kRt;9+8`>sFAN)$@$yQb+L)#mC7o{6D5-8S+;HZb$|vF_zjw -gbEr@GCOWwapp+BZWH2QWQ87pa48^>bu`$IjmvcebHBqR7!MvQFiUVsmlTW=tbKN89{+RU$ozb4W0ku?SGmCxQs{sfww$)lZ)~#B -IjzP^u*U+iR##X0S<~I$*X_OeQ=J -vhE`jhsF&t~AS8`vB%}@csqA^*qE6upqgX{~S7!Ru9uc7Oj2~2q?>XzRP=M$`LUi?S|EgFqKh(ne?0A -0F1I~C=};@%mH9Ip4Z?A3+AnM-Xm^PPR>E>yRGCoGQ7m+^X?} -c**MRc{p^fMrwVvN^PeG5&Wd7C-Xi2*_(~M<)^RcVQTVGFK;INy-?EKW^ztf$wp;O-F`IFzK4Hb82X6 -(@R*Yt+=0rom}jS6o{C;3nE#gFBsT+cga3mEL~ftEng_U|L}FAn8L6QZ8pmZWd -BS8^xX)T@?QRH;AA -Bl}2YvPVK%16E4_R8642Z87TuV{8)ppD+qam@(ugUg{FeKWdI8SnmEerir#KO*^6Vav6tenx3oD725N -K?U=#oX&=FaSC^0M(7l>o=Ti4$n>z4jo_T7Z$(6J0;?s9YV`NeFy(?e*jXb01>v=dS9FHxzz85}~HZB -uNbW?ccH(mnxpLc;~`a`IZdKLDHRs_xXeEz$UvAeLtgumBR!6&Lu)_-wFixO3njgd~+{l?dlTa -9H$c>L7@@2cwG2>O6p%`1=S@0SyhS=0zI__)Me;s0h`G*EdJ;+o(|K!Ay0;=?+AeIMRd0l|$E~t`el7 -KW`=igs5$C6H64vEO@n|IUFH~S9K$epC|u}p{;blB9me-PJq|6gj}s+Y-*J|nz2ojZ#fN5Y+&uZbd_W -4{$YoC+pC)M71|;Kd_Z&^0a;Ajn -&V=(lAx0L}4;wMmXLLcbM@Y?1v2lCeA%rzOyT?a@Z_0)eLWJKS|8cVoEc*EkhrZ(o3PL_1|>3d|5r6L -oO=B(ZE)RyNKu!t7NBsxuVxsh~$9DQRzSAkG_y_CYfs;Xq^5OLqY?5p0 -zzAVR_R;tcF9iWFQ?zHZwzZ-3k6b6LImV5Hyypnu^FqQ<;I&v_dV5>%ZxLfz;gg}eRj{_&pKtNy;tN9 -}_Yaa<4|#4GqhzzUOxA%hh*Tmbw!=6J*!c0keNFLrPASvINtVR_k}Si5qG{BY&4u{X&Ki}>;eU -OuAc^Fl!oJk1=cdjxLnc!>8EQptgc_ -WC~a+OVNwMq)I4=y&s1GNTOo+ae!%93CoVu+5@Upui4bbM!(xsN_<{(3h}fn4oL0O^o>84O%qc>v$`{ZI%Oi5{z}l5%P$xT9d1pT-p+&YIsuj!X=brGJ0I#v28jKP?$ -ZoI2S9fFi~qsMX^Wi6r3`O@&gZ2m?9`5EOuw{>#vwE$0e9wf6eTEer!ujt=_(-cu!$bvUCf>Q#Gl0fa -nN|iEE7gAG@)sfI1e3PALk0foMk%0j7h47!ecVJ}l&dyFLM%=^btla{CETvM8OoJPUIGFO#3`0shsdmaVek^g8T>Q(X#lBTvJ_~Gn1ddtS?`OE&H<<0L>ZqmFk=IcHa&dn&?YNRsDw+Hu< -{}lXN(V?a}wwEm7}GYx`fK=#K$v2KNBtoa5jO?J2BCLw6C%BQ4bGo7(eF ->p497bYJM+VZLnT3Z-tl|)YnHwE`$D)nD{o*r&HCy{K83QXYnk>X4>^ -T9bSmDbS5lAoMrDKj3luztxG7w;=F#rb0MGG`?kAU!}YcoEgp=NW6Y%Dbud_9+GDl)b}yA}&J_MU+ki -X0RX?hbxVxJAkb(ayFQYt#9P&LBN|_(;ZK=f>S+0y!RVGN`c~>PBlBft?Ohw3LJtpuk|x -4s-Ym(VF(?Zq(+=bJ~))20t_^QH(4hwauQ3O~I(Glhl=BlSK6kWHsn;;U^q+rG5Q7GxrGL{4 -QU&#+kg#kWy=86U-N#AgS)FCsQg`^C~kig-lg?)1x=V%hEIQ0~6!zpNI4@Hq-ir@?JEQO!e(9%61^#l -`_`fCI;|2E-(=9(B`TKpWT@G!l_*@bvIWyYL6B0&Gy;fblOlX@L{~YXM;#Ca{vwKf|0PiJbkxSk4~x_z~%a^ -$tf#|Z8|o=#h0_T>p~0z;l=?iLo}S;q?Pglet%jizoD3sVoU}qEKp|0O{7|1ZZ-)eUAQGw!3{;YNdPh -SjZ^GOOr6$b=_$-XZm7Ukav(|up#pc_LX4_cMl~)R{MQuMa&W52>_F&t*eR+*H*mmN15KH@mpM0@ufy_{E8K%#GDlb4)`kecTQ2w{_8^5z) -x}*ICC=@_|S1v#Ow~|-)J0CC;!*t9h(J@^sbFW0V#CaZfYnTSDjNNfDO4=YQ*Fkb)4 -%GB{=2D;7q(e2Tf8l}5#ubp6v$}13l%BMazHH0pB1As7Pr~tKEIKRIS;IW&pc5dqL^81mqBKig#o;R! -L8k-AXVBL_Cw2=U*y0t{2;CNoxw^@>lw?f(7bI3D;$ma(7!NZ@XDbdEJH@1a?t~~oqtK7o50uJxXEV) -F!h?>ytgowH!l6PW=iq%(L}okA>Wo$*H^pMkk -jP!NH;@he*f?KaPA%!{?(kFtv8~g$HWEfSdfCVajOU2Koj -zl7j!r2JEw$G%p`N=lLY(f7P}=O#T~Vvz1dNNX)9zd-wK;CmyQ#jO_5jK!hcXi{TN~tYw;}g+n$VcrS -^;p`^_6KLD#yO((m0b~YDDh||P;?Q@PnhiP-V(c`fzjo9{(?=1 -(owJCk&@2Y5-aurdFd73wJRZ)Jo)pRmHdTt>cvW%NMj4|Ybr+?~u%Dy*BBP6A^*-msY8hLWRcL$y2(G -L?V|{~q*wW;OGISc4^a%dy$%0F?;AApWE<6Aj9d|$51tw~SjL*Vk$P=m7R;o4Rqyvp7dUdpKruIw9!Q4_=*a6{{;2tGgA5UF+DP1KX4Y2a>fWpzHd-ZyaP0sZIndbrKp&~sNXMz(Y+ -JQJBn$I{!cG{-s|GeDW2e~$c*rI<+o*-YiEfp_~C0O0BzJq!0C4od}F~`#JbPMrb1=6%gVU)U_l|toB -=iNgvclCh!XcT`R3w7Z5nT{lcc+FLi1SBy2e2Gw7LORNPL`!}P9MBJe@fF9}Q5a(({c83eO*cLcG -VNVrdnZm{tZNz7vcVGU|re?jfA;?Z55j88Z)g9k_nJe<(FStzsd|O4WFC;QnC;bkoKatyg*Q&AC5hO@ -p!IFygQ?OoS`m{T&)mo3DV2)lTSh1gv8$+7uJQu_ruxV5&u*p6aPtEpi)Zd;?r`Xmyo>J7MQC3a)o;I(YJK(*$ez@9Mc<$ -!#xm+?~9ozv6Fg+}%jmHY(K-Q(>i4)#X`(07kwkM#_{6Yp3nI{ud`6R*Y|vy>xPsID90EAi -zTSpv2L+?W3S_&PEveRjTGA;p5oZShU!vxXx2xmW+UxMM3P9%%j=&0^q(|Ew< -ln1bYomzrH+0+o$Vv)uT2q!;rV$+;8M4&RxS@?pI!<7DPcIqHcXOGSh-+bdDTZ58c3N$yp>7Xdx!eI~ -Y@nn7*9R8;Yhd^ByU55tA4i&C(E`G)~2v5Waaedv(0Z7|ObsbD3$A5QyeP6dllTQZ?E(pA23{o1M*%^ -sqG`LkESKWxYl9n>FP>6sFo)@3rxmm_2tGUJy!|-F%s7#1I2srqHBg-)pos27o#}fe4FNB1S?ESCle8hk>eauu<52Wc -(7;q#_-a#=O^D6AL)i$qJ01*k#(H=`Kk5iQUT4h0akbuz|`E5wMzGlhtgH*f}H$6rYp~?(s~tIT4D-+*(D?krIAM -nLH_}2O;F>!!okNsWd}{g(*bOq*I}8i!J+trT$CIkK?;vecniM7ya4KeGS_Z0yD4C* -*M#@j>n&5!TDbN2} -mj|k!2oYfP}6mpZabG@_WW_P>8aT1-jJnM6Gv%KZ-Nm^UmDO>MoE;oay$7Ze-#+GPKX1-+k{)I(@vS? -*_=id<9~v=pnMxQN}G~C1IueS}d>s`Cp>E^I7%9-=9@Ke7=zDGYJaL&dD?|U5)!Hae5X_HG$RKCAV@78C;M5wRfDlJXWkHNDFC|VG{JHy_DQ?XcL4pca(kEYRzQy0d^9f;@Q7kcgxHYmuGNkJhiIZOQiw=7 -Iy=x1G*1&Ol?v)SEr2L7C4^N*uITsV=t|WMlRNqFnp3*Cra+7ST!jHhI3TvCnl6}5$BTte& -=pCoR>uqmbBD=%+3wB??InH=p=i -jB6wlUyHH8qN~KSAB=H8+V91P;F{6(|Ic@0%n#@m+;dJ}fX2^^XdqV!Yc{7Hv+ -Xp#Du9uDf*%LQh?YGrzzi#9k-ZaXcW{z&7|ho)C)JHhggDP6>YEAfOn^kw=8NFkt4&`G)}Dow>T`0rk -{j%jMRGwexR#@17pidNPthG6)(s7hNkPqNs=@!IRe}-z{dOA8yJfn9&`0{LXu~ -6iaO*5JU`moXK-92ayrcVwDWa2K!wYbq20d9g5RCEp)@$;nw%pjjX8EHWwKiw6l;we|rrn>v`m;41pGgS%_IzSRaZ4}h%Sd>gGB-k$^MK_lUOj$-ai_)!liCQqi}X% -@g3BY?3x)lqQvp$-PRjhC3ixGc_TbCFj{^OV9%uRrU&WhI+^c845KVc|+H|&{i+6x| -pA(|`E4e>e^i9oSK~WxRNo`4%+L}iO46oyT@PZyqAISh|naMR!tyX@zfwi1!Z0Hjd&peyPU!uAt>?bu -md=ddIo#Q3IUHZ~WVwSrM-b{)MMqDL5*>7{r1(@bI?k*9T8VbRLRlL3lMg2mq^DUoT6Gbbs_mS+3&cs -z1b52VpFIH)bLaSf>A5cpJ1QY-O00;of09jiA00002000000000l0001RX>c!JX>N37a&BR4FLGsbZ) -|mRX>V>Xa%FRGY<6XAX<{#5UukY>bYEXCaCrj&P)h>@6aWAK2ms3fSz8}v(j6oP006fW001ih003}la -4%nJZggdGZeeUMa%FRGY;|;LZ*DJgWpi(Ac4cg7VlQH0b7d}YdDU2LZ`(Ey{_bBvsK6ito_>IV3fl1l1!{qMUYDT=m}ELj~43li8O@AbKdcT^g!Hmoe0-u4>Gl8H^Hv}L?)mE~5 -b)=VZ1$y7>8sN71yuc|@tqXo^cBzU2kTV6?S446S6E_n&N?L>P$Y<}v=y-L*uTi -!z9aN$bprg%;VdVnsQ<}XMGAf-Zz2Q=Ph5C$d;3>5`bcJCJihuaWG)B>4+8B4SQ8h?YNYUO5;-~n1#% -RTqaNCPJE}^v#DQmc;KU^@DOBo^ckLRSUVe^Qx%dJ>*vmUYj)jdrn?|54a1RhS6y6HQ1!ffA3SPXD4v -M7-)XR!*4L(j|QWHPCtVGW^Y%NC)_JFPw=73O=8&KP{LnwvSpaKA!pz%F(rmQ(igFU(jypD_Fp4MUa< -T5gGcG$0pS7QDD+FH*vnPGd6JEGs@M(dGt|hC27V)9ED9ZnAH|!1u+sME(0R!;IcWR^~p|lS|#hbmEr -SJLEVjAk;pnNWl3n^Jvc(U_L-kORn@O2*e1)k@ns?4;!#+Rl5bk7B-#T=9M%Ko_+60QsY*h29i%_!9R -VAese_Rai>J7603Ymj%SR%%^#rNqanF0u^6Mf1=(;%0qcyljgJhE2&8d6k(X17fcv-t%X6!HW5Wpmd- -8#!*=aaxAyvw(EU%Q3Wtkh0jfJ45;B3Z}T@y`SfMYlW!niKbR**Xzv -xjVqT1g&f%6{D7=_U!NqeM&!YqXms`^kqnu0*BKPJ%{w8pNNf!T~KttkO)qAzI!|Wzce;Mc-mO$`_;X -@+rn7WbqLmGD4Wr;4m{NRhL)WP^)U_fpZqy;RAGe!_I;F2*pnk0ps-7ZrMf{?1|MHYWDm2DNe{CCc|) -g*tIas4;3g<5JqUwO6{tTbN1MYypQZVRjz0-KKRIR|)VDZ3I?#3-Srq1ppMkNr>`u6`A -X%1OVj5Y{Vx*NZ)i926M_+aUal!pWn+>^!0XZ9K*0TGxsMZK9q8n}a -nglHe^J=vyieia-9oKe7S`wV21m=eBURF5;vP5b4!n@pxH0e=|3v|8Po!)D$be>)GR~~hJa+>v{}6j$S}J{EW_*gvkv;pTw#X{u -j2Sp`Pj?D5uHITEdB4~)qN$MWb{QHjs1n@Tlis@E+M?)l{IvLm&Udd_@>||%)PMdGE7W-Ev6Km|KUGJ -i{$DR;#nRgAev60o@FwzM+6Svu#SnxtB$F^0Z>Z=1QY-O00;of09jkf?KIa@6953_PXGWh0001RX>c! -JX>N37a&BR4FLGsbZ)|mRX>V>Xa%FRGY<6XAX<{#9VQyq;WMOn=b1ras?L2F5+enh%^(%U43@jZq+?i -P%aPXZLo1J)ejGIe*8Rr&@AYf>*WpP508a8D|>&t(?di0~3q!cG{4ttB!2iszIS65fpyNX`5%|>O}YS --;ron@-rY@4=Id0jVM-jz*lCKLOtYHn`I`o?`Vdox+#>Sa?^dP#HB{Cep?Ty(n4uPe=yy8RZ`y0NRjU -Y=(^{B-`qzh6B2^ISct_jC2K1_*Pt%uP4pHQREVX1BU7nl|0$%eOghpSJq9UD@i5t~=xRs&DhEEMA16 -V-R*-S^1^Unsx(g|E7y)`9|aZbyJs1dX&}h=oqA1-5S8!!TK-$D$y@%AVZy3seZTATOyL1J)J1{{Fs- -OF0Pu(R&VoG|B%;3S>&Cb)96pRS)aFU)AHAg+H`qU(Z&3-yWY0VQX3OZAVS!Ay0|1%EwKO4T@Il0i+T -xYl%@kZ^+#y=d@>t5!p6>W2l!KN^iu%Sme)JPXmm|e=*py3S-*A5euls23ZIw^FATPfFRZ`#Zmo6YR{ -sURj7-Z(leqH16$=t;k|?RBY&Mt -P($ceEu5-IZ*=!FJ)33(Hp^y{39^kx0Vb2+zasF-ld9EuvG@Gz=Ouhk$+^9@!2oiVeE51YIZ;n;o3a1 -}Rj;ZNG!!VHyRuuWZmrej{_;YxBx{vi?z?qUC+XzH`G3B=cyazqcKO$qs>_0 -z24H7U%a`SFg~vdWFo22R34fuDmEpZh5Tnuno?>il@d0<}`5AqlCM;;8cjD9z41$X0_Xs+FHI5f70Pm -Y$pZtA_SyNeT*q;lf;aHma|t(aLa1J2*pTcH3=L>SEz{jUI>D%Vwhu0S!~bvoNnT4UP$b?umqj`Zx! -zC1JT|Vcr_dT-LC83^V1^2UV&6P&JKtGrfk-VEWrd6Q0P;nf<9 -J_K#x>`^G1dv6jWRBgK8Yj$}Ho-nE3YmF)~XQ#`C9yCYSw`JSZ?1=+9s@A$Ns%v10TIp`N2DI>b={&V -mp)ZICV^ofy#_?5zr`cl62MPz>e6?u+KX4my=SqWH+_Z216~2|M0hpTl-L^q@G2cRaK*)3uJ+i<~;y75sJ1bgdR#IVy4-(OP%%bBP -f9l66+Im+9c%Ir+nX*KCJRqE8PMz?D)TVgK!YBwdNb;nAZTkcXVy^kYm|p4%eBABf7QCt}4{F0V#8&5 -)0<=rL63UX_B$CJS5%=h}3izfvHY!QbuvR-*BEf|_Ts)39j5yKUEoYjhvmOl8+h+<{FQSLMZr~2MZJbW@fv< -zO4a_2eW24G{2U{p$q3V1nu#DrxEK6M;;r;X;J%Dg0v#;9svuFddLt=x{xfaB0D{*R+QdG|IeG5Fo)X -50ZfKd2rI*oCi3yzu1`bX=7WSqe-V2tOg0l99^FgHrrUpGw^p)bN~n(DT5+s{T%(oeey3fe6#!hL*&= ->HMT0R$U+kGNp=KyMW8D0?wDy_Asykm13vm=MET51?Z?572FX#HPZ~;E>2zLuX;p-$B*4dWHG7u5n3| -h&^aMu*Z-!B9Zh`)^i8P&&?UO@dL*P9TN(IA?b1mloh)G&FZ@mDe#Hx!zT?YCLFe}Y@$=TL648kFT$@ -(W6JvYF{cp@$JB~SdQ+!l?(PF?rz^0V*)|7%%paMagMGsx^6W(3 -N17vS`gP16Ns?Wf6eI7M3)=vgVVyC&gH7%*V)9}pOMfC3K6hm#HBXL -K_~Zjr9~sB}2xNh`(BSM0C%grcXIh&E%jiI@WdC!%GX55^J%oeZA`bcSOi$itZ47Kv$B&!0ZW1PDnwSf(@fRW~}Xjk*I?!J1&F -*+Gwhbr6Q+;G}eSViW2bl@OH2dOYkK?lYCGr94!HwI#G3&L1 -roDP}JX2NoF32&s>sB#FFj$bh?nTpFK>g(Bvc6o?G8JuyUFf#jQz; -kS)NTzox6F5V8)e5qQV)j?0a;1=*tl_5QBj7h0dR>i7(T--n{{&+#Wf77frtytHmsl=jE#^mA( -BQjDIw&k5joQSZicsdu(V3_dAE!y6nkJ{Cz*5^iDi0T8KJ}J0#NjMSw~n4SSh88(1MAJB=n=))JMUb^ -bZuT8ubOnfG1qEtoCk17s+Hq8V;4Zhf2?8Qb-{$1n%t3wW{a(4^^9J5@0kJ5z@m2E5Z!|9KBt@T_cr1 -gB74G%(5*hnbp^hy-47QoklqvIMkHL^1?sVOl;gShm_gVm@|q3IBjaXD_%)8#Lo8XGoe1uHzhl~y~3o -VkXn_kA?=D(fFGC}ss)f;jrqo=g^LZ2vmvXOQ-zh15f#XbDw?~xYVyMTj+Iaz@-O(=m%4JM(nh~2<(A -zMvn`>Q3G}AH-|WqjH06QFI~~$=hFl(zls!KoQ>dLQLCT8FLnir$4UDfD+K(#e;=QQEhxm+l0SxbfeJ2t^EA$`E_l}^$yB -xoA+R7PH>f-8&Favpj8+f9A);Bps_#8IzC6LcvG@4OZ=eOumA;9!kBW-m6661mlzbV1Cpj(l*%XN9HX -@n|&0Ct5oatLSiSY!Qe<4harRs5MzaYAa~#rVY!*;@E+JD4|}r9+#BAKYe9MTV@0%0i+QUpyRlF^t=- -@mQ?-77jn;?+&vNR@N5wcNlz(+>~9{`Y6GfH#gAtGTd`FfzA$6F>u*zpo?B$Rlyxy@5-t$i>r30XWl^ ->G?#s4XKSG=ck>^jJpQ&H4kp~keIoMumj -x7}S5;w>&t#&=;=(_B-H*H>MEW!d|Z^Wt0+j5&(5DS&q?vj95(&vMrGseuVP(k}ZQQK}LiqRb4Na!x({JEu -8d8o0%=YdWBLCc#%?1Xa?FpSx5Kti|RQ%j~ti=Q(bT7K9jle=?yz8g= -Y*QOF1?Lo+S;PUT!BmF&SK9AP6W1J9z2%%3wRy)avDu1aA5l(lqOo#5EXajs$7<^u197Izs2X|(GrTs2kI+|e6O`5t7ApnR8tywauRYFC%P?X+TpU>jmkyF^62;x5P?c74j -$PRB@CaRZCkp>IV{4d?i=<)-kscv9^Zcr4pBJ1XC&n9Zw!wx_K|t|rEbdT~xo%mZ%Qe1-%^NiA1Oe?w1wovD(J-okvNp-J=Ylq-Z-76v&2s#n7(xh{uW^cc>4-L@hwHiC^lblh+lbf1a1^6GKoX=k<35>cTkzyM}Vjx$J6v -)LdTo|C?N?Nf~3rQZ_Xe(d|#1j~(opMIOFpIC6Zbt4gT%)tI|vk;&bux!GL4Y{H#Z#6j4oERLV_o}W@!JEF0BuC1_$7ggNHwMBchTy`^6FKe -_OhiJgv{yo$j`^_02i35D4RtI1%%W(JxK+}>c+wNNGm+z%lM6MNr(!AdLhWm8%2FX1j5nV)zfd~AHJD -$0wgE!WvdPb-biFrXk3=jN)spaqmv(@68XmJW1a)^|xG6s1$76!7>X*IW~xL^?B%u)j?tD7b&nXB&xH -S*;C{j2-;uhsqgBxP54%CM4olK$G1?EG7z%y^4OjV}?$#l1%tg3P(4sC~*`KCbt87^WfYxxC=0FX+p6 -1q+^Yu_t}=p{Zj#H`)`zhx7=4dU#0;_mGJ2NO1XRoxzT2!n^X*aU|pW{PZApK#9-ns|A@-C{$jehQTm -xGk`inu~Sa&s2gM&v#bMh@U${P`jDuY5Rr=>DcW{$ltc1`BGCK#7R&xQ%bG8MUd@iaTDz7l5AeM%(Y#2?tDD@QFe|+avdj2O -a0~0qNa!8Sj;>cLK%JnaD`{ILZaO{#0^&e7%V|BdUNB=_yLMRBV^(%b$S*2L1Q+oe@KTN3 -H1#xTj0f5iKPi1aF-HLse*#^r5>(G;$8@?Y|c&Q=JkMuJL_lo}OB0^e*wPth#St8JB+22q*u{q6X2Ks -{r1K6jS=qFF@K3+xv*dkNxtage}64NAkN5{~M%FFYJ6g%#U-9Klg>R-o=gF+m~_Q-1}}{rYD#b-Twek -O9KQH000080LuVbTiBJewG|-%09=Cr04@Lk0B~t=FJEbHbY*gGVQepQWpi(Ab#!TOZZC3Wb8l>RWo&6 -;FJ@t5bZ>HbE^vA6J!^C0#*yFkD-dk?L^>4Kb|rN=Uh#69%{tzSv)QsYaa@WDhRC4=3j}BkC@rJtzo) -zB)$;(!wUd0fI#$^$0y8~5J>5P1n&GX-52CNG8!=OV4P8IW{!neV&~KJ~Eq>UEb}1BuVUJHu$ -H`d+pLgHBeO12v;?>LFzkc((89jPgH4QQ=@I` -ZBgEfcGdO8wp!j*IF+IYjOt!&L_0`3h4#K`>eW~7)gcISE0*=T7QOBNwR!TjIM{PH)Q#=*rt3E_>fgi -)XE+1o+pevb^r&p%(IH6pFgh$mFo8d-r?X7|5ku&O?3#N4Yi_!Jg->?4s;}XJlr<2Z6R-v*>D#I)meu -lBm|idO%d2+i_p|JqTOk^Hkm@GhFU6La#$Wf})^h0UYa01YH+>Duqw79f>7zV&5alI!q*0SfLj*4?S>o)(q3AYfaXhRpzVc8{xYA -Q(WV6tSAgCnascq=6@W8{lf5HWUv(WPxSTkXL{Eb+*gR`Mk6*G5SBe^z>!Asa2w%A{sAfQ?u?a36X^P-fHC#x;p_UQ?RtwFQZD&kuaqpe#9(2o?Y`8e6+l{d^K8IXz -9dyu6vdTB8FPtUEQq8byYXu6Un(UUN0jv^9leyJ_UWr>ae6WJggv -idq%Ri&^yo=glLCKLa{s@gK^qnbbCxAMp}LZo^6XCCIG+-LTwTq=P1S-(j&6+_GIuPCd-l -#`>Xs4q7jR!=5%&;9yJRtDHBY8q+mNm+(AW|&^;+Qv-xSt0qrGHK52KdA!t?_4G~!YK+xNi`#<-W0rb -B35+w2#gSB(@tdX8RTtBd(*s6Q!s{qMF%vwhgxogt8|2|^&9g{qHl!^5s`(-L~-wiy(ovR*5}s3q8bZ -Z%0q5&T((@8@JzCG?|SP;65G4C?1kQ(@!pATUeXJSNm{U4y9!%Q59psa@fih%4e^sCr}%r6kZqK;#`C -PNvg?IBet%YO4Iv8~!P)B~7Nv?64YLP*KAzyxoG50fKId+!4E~!88sX%?S)KWwa)h8;a`K29M_GJY@z -G9}$o_6Jg-SEBF3m_b2!xph;2lr#AP|4KPpuj*=wZ<(=3=pGm!%Af8jEOeRs_V|em2(s`7-|_QQ7xM8MZh8s9BtUB62`a$-_ -7G_!^w;?hz+u{Nbv^ah6Eo(>qg$;;@BQ{nb5!|uAd%YbZ49(=$V}o@T4-T7o;H%YTZ+igrO0GQ0Zw3C -p3-U+|$QtO)$!YW=fGV2{;8c*%+rd1^&UpbN{RH4&bjG(>;pCXAN -1m}{L!nGPd+j5#-=M40i3WLZ1$$#s*AD`**#s9Z1Zffxy?Ve@t{`lKJf)$OSyR1?Io>}p9nsS!GU+9Z -u^?(aAfT2lYQh3au89O~&H`NVzI>Yj|@XSvYT5Zz~g}0>7vTuaN2ox+}Lq>qxOs)lNZW=SdYSjs4KoH -a-0QTFiJ%?8vAl{=2#lT&4Q`K#eVkD%Jhqqa?!pzV|*368o>-Lz_5}CY_O6SZp4(JlcMD-ZCKa(Uh{4 -^dbMu-^U9Bsi>`*6j|z1a=)R9rD~HW}S@kRAm -dopACo&4r0uja4NAmLd)jYf0i>B!w@GJ1J7_Rd1IUgqe0b-GgT0G{Ps=wQ9^Lh5J--)MFX@;Maz*+tg -$u3D*U@uN#&I1luQf882f)ACK6c9UCFWB&ZUQSYC^8Z3%K0i??63>q@AhRbZt`X)q#_R~p)q%Ze$^b5 -7%Yj*sHnBeUGU~2@%5=?jj>`*PuYt`D1bLwCWk3xxOAbjC}N@JlcFGxSAlr~c1F -T^u0j0hFN!*vw1ETzB^$czroJaTrDA|`BeHpseIa_0J&5f4ogA{|ZP$TWg9gpI?%@H;K_IYRY>|ao;^ -bL|vDS`^piRLZf6;Blde=02g7&uWz8A|Oqf9ZHEnr~wJw72jW>ZVDk5Ko~;8bbiz!`ioI*Y`*{KP5KN -=nfW&M3h|ZkD^0SVnAxaTCg|XB(X8HK6qBZT8Dge&xa^sCtyA97GN{=t4z@jyNwE(y^fkCvF&jK>Tr< -iAC2lsdd8n$IZgy`RqRy7Mv?F@!IM{uW6{hPKF-f>s+lEQsG4HiTs_v%heLy51cM=olEjuh-ed1%)9mT%rK@50Ias5Ij_^8pxEws0t%2Q9-H)&=3 -GvvlXdKFk+v&l~y^Jd -`uENb-?iweTI+1|?l-B~(ZXzmgK3lbqmIrKQ%j*mKSKLUc`NeHu^Tg`}8)1u?Acb>LZLw6ucX8xdu(G -r#{nuyn?Axln18Fz%lu@l0<=a+ZCO1M#66f^B^bGh-)un(9u?5})1~F=Ra$i^3`C@U7kudmoaXtk|)j -*cTwFuzIt`Pu3*P6iC*L#4q_mK1X3bWDr{Q|#CkPL;%Ea^A2o8y4UbV#N^9*JpW38=T2G)mi{iJY1FA -bhzcchL4t>LT1mwz{qk`YaY%4kXFbjq#%y9o5?@0DndR55PmETDJ#*nLQcwwS1_D+eu!0oR8?%qaY9A -uiufy;;)BV2s9z%MVUrC%qvJLPE~L -(=EO7=M=6K}*=V4ey6j?G^+R1XX(xAlCmHuu@y)^_nr?GsGr`Z14aIzp#0^Z$fbba&9K=5Hz7~xpw>1 -usW{y(~tNA5(J{iY%4IhVb1KDpEd8PjAaPxC-r=hoPv@28OYF#m`e<&Z!3p>H{u^w*P|2~x6@%PP?hY -Th>xQ?scQrO+V?a$Jg=adaWR~Kw8o!-3w;n8)}EbKl9X0JA>dGKO9ZHa)#i@!2qV~Y7K~MCLt0VZb~}*Ps(Gm -PG5`hG!*=iIe$YXiF&m^8=V)JC?`|kK`+!e;Ty3%oErPI4cf#iG(?$FAuD34ipv+X3x1E_9j_A9qN&y -{9+3BV7ylyAtp=`5y(b4FdCG8yP&uchL-#M?gj)F8tviLt)-q{?OV85e$STCquz6N^bsbE=mjlw0BwT -Kc6G{j&J*oVso*2FPShlM}TC|A6dW3&ZX&*f~vf6dUA0VA}&soScdEfIqWD8{r$k4mG-{kh4|UKvQEb -`XwMN8y`kS|j|J3qfBM2u4x}TZ4cO`CmtH{s -}>NhHEZB;ubqm!S6wW8wfUP-DSZq<1Ls+>>c2%!Dtl_!*Xi{uzwy5$bMc%^qKockfWWVi{$44Rl~QLR -*%BLb(w$%nuZ70_`9*NY;wUQXfl+KZ^U*=8nEW=lc<2bWyHSr|kaxKLLEw7#}q$Sk8R*Nx+rm{h4dKS -~9cFqE9cCn*Eh39Kpbs<>67i+0%)u3ADO#{DVGkB%>s_LCNBw;fNV>M-mF7~kQ?lMR^QWCZhBHkl?As -QI{veWJg%Bt}w)r+7{yqI4MH%d3%SK>49UWVTyD3Zc(EO9|h$i=}*iUJcSP+e+x6ka6RtfJ=Pz$D$GG -{ujiVi91u)hr&21Qu!FT#zNX1$Rk}``M*@ -^;9d6@T5HJtbuLJiiZ-Jgm%)8Ezj~JIf)BqPlS7-v0DvsT_WWxGuK!TFX>M#bP=VkHaEDZMBm06Ni;O -{Ir{H9uo4A^#ghv@7x>fIxIlh`TWa56DcCq$to>A}Z -(n99jYx5bP&{Ue8kiBf^CeiH=Ta!8rKx3(G`E9-kOZr}I8s(&VUx~Cx#q?tK3ni2*{jyqyzI)G8dE$^b#9rMEihlN% -(>5FIfBxq?OP}X|K;9SC2=XSH_t=qr9(joFb;cQn@cFVe=Lo8Yo>0n)q9qeuFbWDduDU>^+-IO#DXn` -t|Ao8Ht)ClB>lMHoOX@%N=7`AG3XYn3SD0mBXK^+ni0$}0{tEliQ4vG@2R3HyiUwb5qPTv7Kp#AlcsF -R>#!7ZTfm?@j+NWfIv)3_LsXm~?P;A-O6;h{^>dY++DvQ&0sMPU80y!$l~h4hLZx1wNyx|3FqV0sfG>ZIiAY50BOsbt@bi!4IV?El -n^i&5zT5Q~JM)*Q-K$Mm(8wzddn{rh8a_d)kxw~!+Adc)AYIbKO;!fJsR6xpTo!mi#vE(fAKKv1+sed-|qSlt6MRGtq_z6)YM->IWw(e`Zu7x=bn_W?)@{DCcc+8^$0B^Th|azi9D$HXhPL``Qy -uB+u8tn>0NoA9$O4&tWTUn|oUrU9#{j+A%;2wNx+oG5}QPO)S|RX$Y-fT^3K(D!KZ78syI%+gGG(lMc -=2c`y0mN;9PxEfKFW4Dwf&X_vp4*a`FA*$JGLOIoCqo9b;2tG86&>YmxuT2fb!9qqDt86@5*Yo~FY=%$U}f&Xs-Fc`?E;=ATi6BVlA`Esx&q!p<$5CZNH`uY$R}nq -`kq_+L;vJhSD)qTOxKHWw3z<1}qJu=GHJ1z*7)tSZ3Cfv%Pmi6AeZN463^T?gW@!Gh^SS6wB{2#uQta -7bRG+cYM+Qxf2akvI}o94(LGRoE&tnY*Nl)#0Ro^QiM^4Y|_{Nz}0$Wdk)!utggY(_~Xs(`$;k<1kxN -4IP8FU`eGyIi2Wj#tWMUkDQ5@C?)Go@x#RYME^FTVA23#2zXw?BCM$~pLl6T`7DlWn$IX-u`>*PaD;& -YCq7YE+-L)V410S7fjC8nEZm|6_d6X~V!!pJ2OF2WPQ-XjUoW5!^u%yE=^hr%dbDyh9gG1(@(K5QPHF0a)DytIP2oD~bp77!#}cX_#%d-6Q4F4&=M4Nx1%uQ9 -s`jVfNU#k9Fj?}x-CTT01WQkipUJ`MpP8KVYoJpDqRY^b;9z#mUyCwycG1^`MDo;I~1%93Ven)p-|Xj -P`A0_Ksbe7AjJWE*V6@;GP}cV2RAtrG(vI(AySFVe;xJvnrd^ssj9m~@wGjg>F3I^>|b=Y$(t{9=a#Aw;~EWtuAb`Bw -!J-3brCp2ZC<*nds@(4DmUCxi`sU~#n2#RwBI!9+y;gJXnIBrvuhPK^wXL;@4m!B2?{s=V;Ad*)o3H^ -0{+68w|x5HdF+2^bMebAOyyi*Q^vn1YLjP*v -U^g;JJr5ylQUq9yzGsUtG8tzs)dN%uja{0=>$iD|aowjxTM9*o$!ZRb8WZ;k-~`pdRUtH3u#g*9x)HC -aXP+^J{zg%C9Y?Lq||4$!?JxG0R{f)eVS2Nx4fVV>H)J-8%%oI^jYasx|NJ=~QcA-!$Dd7yx+FXrT@( -ni@Gi5XJ)-D2(p5K4u^iSUpo?6cp5i1`Apg-7o?_vtrG++-I(VKO`ClAV5p8B8R6#;Z#s#{Ttt+n&f@ -Dd-5nnEItKj=nagkV&h3Hi*Wslx+Nt3SvbtB(dLd7^O_g#iz)d-4sr{i=J5^raj138s*&-g4i!AmnSw -^}1;zne@u2YVmIaW@3U`8Q}Iw@$#!DHylThFj}F8Pp9X2;x}SI+uGqTn45jV^@IoNI3S@L#gk&Z -1p4DxsaAmJ}^HDGZq@uvx>|*C+$>60oyR>)yO&{TE~tYOq1U<;$1t=BEMrl{ov@XqlyaREXxDSNw2j+ -&UiOeg$3<%Cl=-a>|(_HF-O(>d&WdNQ3Fr$aOkH&YSwxk3$oW#%gax$xJzBIlai;il0HP4Jvx}`qIL4 -cf7U;-dSYg%q*&>#zFGTGPw>eE!5%GU0Pi5*D0UcID#t&zmz$iOigimpgthe>AvgEWd}yg@4PT{<1PU -I*TSF*SAdeSBXcnCwU%U@lxS!+X{@Qv}%qBKFeZ9{fHD*uKJQS@HY02eI`Q2VHWvPWKO1@zEl2^AB-O -{h^7AJ_QAAs>ZBn4v35c~n{lC7rYk&*`1dYwPQH=r9?&X0`8k!UW&0%*(Rv48X|S%DtLfYZj1yrc7Q< -@G*sak&RKRB)yzICT;)i=IYGMV3+w>QIGV$M{HvI&fTmKo47xN3#zyRMMTYUxvr)F`KV`eTGT{KKJ|$ -y_yG(KqoL&rznhyJSBH=sCPs@JWqmrbFZrW?&;*dx-pu=nO==i^8kRvY^GP9k`G#h_eXwR(Jha;*+Bk -+BB1(!xgyax!2-lys2t!9Q;T5<%k401<^1{c8`zw7*Z3PV&$s(+^?X}zpVw0Egna(#r~mnjUp}KR@EW -D|;?rMz^4YIG`Q@*E@tN-{sp2e*$77;wtU!xyy5xa5%^h~0*+()_la<33TYTTuZAdvPKgh9iHy&iCB3TKB=qBP{q-8O -&PF+lR!g0|k#1El7|=a348JF+hx{rO&ACGK*J4lOJm6RliKe!=qi6u(U+!lPP!x&}Ii^gTdx}-$4 -}zj>+P%pBQhdu)+cTwXF{y?B(zz@gxnR8NTYw5{G%~!Ep~v16&sQ(qJEhn0%a>(5SlJ3-ePvxMxSBU>@CKs>o(~@C%uAFvRe5DjY(&J?$kfK> -pqj6WqUDpQz^8mN470oCBp>sr5bN6Ek%DrTIOrN!vY`b)kBTZtZO&oN8)yQj|-s&I_0a{z?;qR6z)`a -BfAjh3&yML^X~Fwo5X`@Fpo!Dv4zPq*39I__AxE8>g0l(TUn%sZ>fxoDry+%`ZNM5r3ZLq*rqIOFf6j -{#AY}A;Ge}7A4pu^m~3$U||Gtm2cMmJ=?v!Mh15zs|i-!eU{*nlMv-Kq&`<%o0T -%iONXBo?t2m020NEpFN;6?Vhoa^QEiEA1U;ykg0vDpdFqi4jbe{JK~0=#&!TGi+;sTvhJHDtpQs*F -ZD`IXAL0lFq(CF2XBXCEHnoF_r0M9Y6wa3$*zDzFKshqY%&uq3fisIcPYy{fi=IiNk(tnmJfZ0P2R_M -aAbJW&c&)zeq)4H<8d8U|Q&x1JV}RRYv!1#l*1_@mxfdW>OA{55a-qM~2;r}A7QV2BMXATvwLhh-Gwk -vI3th4?uXZSr@PRG45+ynyDbE?*9JBOTnvA1TMEc2RfFMf)5J@6aWAK2ms3fSzA8nw}0gZ004p)001`t003}la4%nJZggdGZeeUMa%FRGY;|; -LZ*DJgWpi(Ac4cg7VlQTIb#7!|V_|M&X=Gt^WpgfYdF5E&Z`(Eye%D_?xF}{9u7C}DGL`~uumJ(qZos -;|WPw6UruC$}lDOq6SwpK;P2kdoL7i@;ndtr5LQd$QS$f*1XktkEv8MUqX9I+=SVT}w2!pR*O%C#XlS*K7A -hm3$(_}a@dB@V+qQ56we?$|8BdpI^n0b^K&c#ubt>4$O+8^i+%UBXDG>vT7amD@>&w~^dhUDS>sQxav -t;8HuWU>${)Sx~A?kuAi-j>=9a%j1dIKg$m3JhB>Iv0-tvSS+H*$0AAc+SP`VM5ukGjl&^0!bjUiQL| -R)yE~4N-;(O0pwtV*a+|jMt{h*IHf!+#EE(cDTAK=z3^j@9p*VP4Dn^xAW%aCW@kra*-%FkT7Q5QbC{HbV0qIpLVkn -FJJWny}nH(@WBts-Z-3wv?8NNtEz*%C-|ql$)ym?zwg8WZ6T>UqgU?IGo-B~T*%4W29UP-9?n7t -2J|A1f_5vz<&H(7f0&7zSN{}&y&?ZCbF&G*eTMp;w0=*)2<(UH`JKdv!@KLK-7pb2ROjdcWY#ShYy@P0XZ=Ob}-hX`@Tv}QlI$X -@enwfja6sFbBwt-#JD`V@AAK&-AphpS_@18 -Tt@&uAF!|URsU*dnVpWbKm(co65l!$%BT!xD*bKW=0;8=7VxOk5`LBCn`IPga|_Fh>BprYwDiabtexr -3n-ucKt&Ocl#Kxt%Pd;6?Wn-L12u1Eftdl2KSxDH*6)6@F*8a@I6_Hfr~wRi&gM4yR$3|TJrsI{oS@Xaqa9Q3lY-fQY3%yshJPwzd9t2-kzig6D0-i~jhEAC#6EKsFwh&v>6Wvm -OOci7IHa1IN57^r#Ai53@$+A^tMwT;`U2R%8HMUpzQ^EKt$| -YWbJj!8{XJ`J+L@wQb44*L6!QF0frZJbFeQ;?+fK#Q5}6>~tncXNp%}5^+2wdI3+E*2^u!e_Asq@J+@YkN3>yNLXJ;p4OWRv+50}{{e!PEC@ab(N#D|2D)PS9#Sob^MP)P*o -~tl7{q|E;O8n*V-wQPbA{hv8wJwHQ(5_Suk2z0+=B_xHwVvs` -4l{K|<;`{Z4$M!@C%D8%D@OA{az&@jQrn4{UR2{mCcLyhmcJ=BM6nXcNJa2oof>q65qye(w!{|>R7;=>sD#2*5%=Y!VEd8g1^OD`JDHOzU-$ -Ko8c2S;Cs_hvHBH;4N56YPa8vOEUIruZ2c`VUY`0|XQR000O8%K%wh1+IQqp$-55u`B=pF8}}laA|Na -UukZ1WpZv|Y%g+Ub8l>QbZKvHFLGsbZ)|pDY-wUIaB^>UX=G(`E^v9xT5XTpwjF=(Pr*4aA_qF{343# -^yIzt@d+3_pl54i&aFA(>_HieRDkYucqUdMe?@v-9DciYpz~W%u3EK9l} -&$k#aH+4S-Ynb{ext}@VWW%`Z9m@&E>0auik#0+NFIhK5(5*W)pj^mUWgtimFgL)1p!3qbTJvtF?NRh -0p`4i)GVlzG|=CgYdYlKzYSW0QiaiVWDYxSq?1PKi`2Fsl`@Qtvj79c_aE=uXt6+g13;wE651g05@gf -DSK$uZybnwclnPWuHId~&#!;H`R4n#`PJ*ow>MXBt}frXBWqBu^uFK!#{TIZYL8)nc=@p|>NZoO;3ESy*|m(#fW ->6PI4iJk)slouv){^1tsOnRGAx)3F4{1yNQm@1B+AQS3%P?h~Z6$tx%##$)4Cf9?&KYz`YDSmlP>L3bfMh6!B6B4(@Io -6JS43u|Yp*nW^Q!i6h3f0QYNIiGoS1Ch}$?&Z0}%vTK>ZPauT*?-2`YC{Jqqv~tG}$dHZ_x<_6Tx?Vw -mv=!~WDoE}XtGRCFsw)w^*aB#a9JY+`*DMTy3)$oiwAZYqDDvfgK?DLsXbTRUHJrspw0R+F_#NCv5*% -X@-jwWiZ#hE{X$;9G>S0X~50LruhV_9EM^{j?Y_`T`=ma7lR19gPfJXSS2JK+6V9R}IZrVJ)7iB@((~8kJK_DC_YpbMnNGizcb! -wi{2a&1M!}_h?eP^~jW8Yl7`u3b0PhqQEnFz8IthrYmTmzoJdLVhbkP4PUgpFlI77p0-?ZtPOx0nCAd -3W*tZpv&Xbk@%`nK@z=F=Q-)VvQBKRTfD=X$PuNTk_Qng(8aYFq8(u&DcwTXe0u&bCAJ;R0YzTTEdGg -8ZCqa_5dep$g@51%{i-xyi)!Q4P9Sev%mb|&xS6zYw*OZu)6FT2g~ZlGUYtKIlJD&jjNO-2LmIqAwhA -rlr2Jx7Fl412s2f_ME-dY5KRuR4YX^*FBwrXy@d-q9fwB@UL(HGP&e(figpj210yz=^uexW-PZ|8R7l -JQr@l5)_W3-{;S>miYRbCqIoRM^lNw!t329#pZVEjFPoOffBJPI>-8B?;zZLst;EU2A1gk2tP@@+RL= -n^KCyUU^4fH3XtaHx(IoPm85TIQw&dw@zhW*WA5rgXLjpl{OiP%*JCh5FYl{k->$@0ExnD^|Ao`(iE4;dTnJHiHe=;7+<+?Y2 -U8ZV7cMC{W3Xu<05tG+B5v>A||5FjIwqq^}TA}{Wi=(O0Q6g#rLkp%}poeR?k-I;1c6O}{HBd?CKIR# -zFMs?bR@&OD2rY+Yfxlp48t3~OAqzVuyb2Lg)YAM=QOVqNxH}`hj|8^1VK%L&uT^!vTxujuDnk$_T73 -XIo8H6u!ZJqV|#AZ6sN9)HEPR=Ygh4NB}V;*|4638^1akSYV$GV!%DE!(JA^F3wYv_s!M+T0I5Z -vCmye|_J2>fF)pH-xc745QBK0V0Bhfq;Ipa^h^@)DiEo>1tJUNcSA^Z|P!ZAT}E-UA=;`j!c6rmGX%I -Dij?Q_(Q{f)oGAIr5^=Rt2S%qYgLwNM;1M3&eKAl7l%0?37V=CGA1gB${pV~&T?!{Lq!6ekHEa_WyP$ -lY>M-=S8y9JU>*`_dsA2}gIchNAd&`Ii9QkKQY%p&P~AaAFTPMp_K>I>s0UpnGt(fEUyh+sim^^Qh#I -V-UhC!jQ^yO~1%e8k*w%@32BeyrwJ&=7=S`<+-A)%)Rt4Tzr1m~2~lbKJcN*jkFog6qtg!(AF3~fU~=fk0y2{yCnF -qQwbi@HKG3lWSr86~k0n-1`IrYVe0vvj8?_-}REsEiHg+E~LNS#)s2&ZNjGY+60%uF!*zRNm -Z;dKHMYlkuxAVP1k)f`oZd%iNJt|okgQCHJ9SM=-TAGrSO*R5(`p)J-*48m7792PYaZHm}YkZ+V5A@^ -ZGM|59f_kVhmwTd?ZA9uTE1dQ5#%$zTpP($V(v!OTltE-j7$}Vre|UCyR?$eT6P^oDaSC9J+R5V7W)I -PA40!PAd8EZB6Y{7${eQ^9lpp`6S=etlm4azR{(x*mF0hf0BC3weY_gz1vln!bcd2(mdUaix)0G*8z+ -!D)*iuZlhbaph!`^7s)yVomH@NBCJ;Ld<-sjSBVu!ISjrDVX@CCc5_7NYZrU2@x9TH7bCbCVIvRaw4m -!e8VQwei>!G3u6_v~|kL7ONE08C~Az}zg!5=dn}*#)}M`Fy@u1orP17mK0H=J-_>E3S*D@osGq4|>r( -LiBXkL0PesTiG5XCB*EB`eLz2>1%Gke>LARWg%JYr5HyYQVNQFrDcnb*K%EwhqOISX|H+G(;!_-_L6d -U6M&;~NOz^3DZy`0<(ScIaBtNpBo&?6kWN#3lOJhytyFbJ8FLyQ=#79fu( -6rJ8t}V>Tw8i{#4M0^J5-1ub+P!8tiagA@81d%m>8MeBNYd`|v}PM3t3a0#&&V%cprd?`!W? -q|pAuU7Bl)IfD|BEfGG#Lh5j&U_{kAL%-Wy>C@*!|*|QZ>}J0C0MC28oSyyO0%ZyFqPwTv;Cn`yJ`f^ -TVq@rpzO(Ibcgtcmw~(Z`@PdAZuBUbl68-^#W37G+!&cT4ue3dXuBFY=I|;ReKOURDa*3#4gi8;_mI? -Q>}wJX!uICgUE+AU$v1}PBDBJ>?+n8r3F8XdVWVxeeX(qvK0 -5*(8jN^g>PsW1Jn-<}t@4L2F{^!MfGL;d&&-`0|@vfY$J~FNXZvSlkN~ry{N7)e3o?AGk=R)R#S+Xz; -4UG@y&_7RoewuHT^Zu&Ws7-;2vF)1ot`aJqbu}OW7fGlyis)+OJh{gf^OW(xOMOf&(Vw(88jiLpPU`N@=R1+UyR -G=Lm)$!!Td$ZD32)jA#%oVwmX91KoSAn!Hn+Mgu~Wleu^+$b9@%V2X`Y+$GMOze|MCLZQbsZSr`vTlt$NXnTe6U)=f -^}ou48h43qQlk9eDBZSH@@TWkl}^e;z>VxEw|p9*%f*+^F7;5$E986AjUjG7zyj1~{l%UATil3dwWpz -Ml$+0?0J3^fTv6J2AaEV$l^FSO#|eR!iaCNdcy&EnZvc>C$D*iIrHl24f1JQc=@9Su}No0xR!{&hI@j -9xC%FFl%pP?vKMfwR?A0;M$?RW8}sTQ-y`j}vUINxWn*C^wvsfplE2$K#1&CJqu?0%^K9}TP)h>@ -6aWAK2ms3fSzGmN;aS84002b~001ul003}la4%nJZggdGZeeUMa%FRGY;|;LZ*DJgWpi(Ac4cg7VlQ% -KaBp&SWpXZXdEHlCYvV=~efO`pxCA8DRap8`1SzB`g@txowuMsCAjZ;_J&mMMXGX~yH~+n7^f|JX-Rz -dWRD{?X%{}+ruQPJ4RY^&bH@4B7BqYmPX-lH2loeL0%B)to+mouI;Iv&CHf5ZPrm&ez?b)74Y|~-ikD -{(+wd?j2YPZA9@NCES2duE}p4vn0J-n_Cz4uzyQS!i5rgWs)sN#VOxy4tvkM&6DTVeQJzu22bTdWPwZ -$y>JOjzE~4Om(&WMwz>j{j<;=8`MBS`{rHx|+BJNm7Xt=?Pb>RayvR=x13Eegfa=tp~O7-NDyg7_SHy -@`jYlO4=j|3>WzZnH$KMc*xf+R(zUw^f$d#6@!`k2q%RVH2MUY>(qscfF1&m@T+#jY5+&=n{4B8bMep!A;f$(*x -?>DDgnr}jUrcz -S===k`YHw7o16{^>rr3;KtODr2w9g3c-$aBObgvuUC=AKc|NcGlsVz(W(v-ebP$(8L$pN9sCevz6Zga -Oc+#N-(O9q7op_7>yBdXD_RdC}BLS@{qBsk67Kd|L -~=(*t`2v-az6D60|~B9u(>e&EG)E+A!^)aqhW&qIx+7z;&ET21Rn`&t#sg3hDY-LT7*}jpzW8_410F-!HrDKsmvwo^E^$v63g$MM^vIM!Ep)T>0DyR0(O -O8NO&+Ig>*de2*eSxAEJ!6%`O1$>>ckrx9Z+kKl^15+H_6Bduk_UnPMk%)XoXqw}ZoIx`X@gB*dZwY$ -B&?qwpImwA1}o?H6=>CI$n;XtWB~;W<{KAu3saen_09?_+8c_KD~{&V>%wJ&xxL6ZE_*m#g-O!jbi2w -=={G-D7yZ7i;?gdyn`4cFMnJ{@HB%O|0001RX>c! -JX>N37a&BR4FLGsbZ)|mRX>V>Xa%FRGY<6XAX<{#OWpQhBvKWrq;8u3zH>-jNZE4Y#u>N|7o0V!-|?m&dBgO+ovr5Ou -<5mIDq8X{EPqWa7L$w$DR^dvQUOD*5L{^12wEl$`;=rLxo+A_%TWXlL$oYYm{0RY)+zg}8>;3Ya9Ovd -M|fp_{>eVIykQj+x)UH-QN{XI&#B0HPBjQTeEmbpi?)u5vo$6KUQc%yS!5+uioEz-mptRzSr|Isqve9 -YU(O1!8W&urX{cCP7~R;TmgA`rIr$lsD45-af4Ddq){@h4-@-u#-F3hQ0)CDN4Fg{hI2QU~k&Do<`c{ -Y{xHbrBYVMw<@C}ubrZ?fWkNv2*i2@gYP1V}P2h(lgslDwp)Vp|%VnBvq*kxVS|OAWrt8oe1Lhk=hQvyL-rzz>#DjK?rT1YMD#`TjDK`_Mmg~0E~wxmWlbZSTfEqZ+S=^+%Q@5Oure)8K#$;)NWbNv*hJ-s%~h&^ -ZWRz@o~|~xg|zRU-3gv2`Ns|HGO+N$^a2Y$Eq4&$4vK+;7}710XvxI|+|I0sN$?h|i9ad9iP(e{B}29 -de40uT0w@YUCNm%aEx{bw8k7TuC11@Eu-GWpUyMAkM26VA%4P$Y14jSCgkn0}$~PhCADn{#*9%9+#O{ -seqQQzgBM|PTk4fv6xp7TxlR{BW|?QIk(Xc&jnT+=7^)_f^Gb*gs@HKa)mT?2%ql?kK -Kg+TIyF-U78gv-=oiw*6r^*=di&oaMd0C0hdl4!E8Tua-o%D#s@}(`=*_a1$?ss1q?bqSwYbUS-=E2a -zc4UH;SCSI3qtk`6<{zZzOVmdutBRf|R?`NEIVWzV -EFd8U&tmdxOp+vt$xEoibMjl@z!(Cp=mk?IR}Q;@MOR#tN;ZrH>$ai-?nQIDDp?SdE3VhF)ecU>)dkk -`lw@lv&JCYcj0!BlN%%WxuV~A1^9UWCU+>FMfQ@5oHY&-Ez#jO*o0YHwh=O-kW!xWEH)Os -l5LqAMDAYJnCC8^f3~_Qg0U8_ZEgFXj?0z+cTCn55&Jt&2;FC1gk$f(6nw*KrA -IokP*AaGsBtt+_C@>mW@T$DcR;-d*NXm+a(mWbIE=T-ERMF^Y}h(J -DD~@SGs0obqNn7ey)@3C-s`=vZ(Z))2K4Sdm}9KkrUK8`ZhJyu`qnsEWDA)izC!MAJSOUg5(*p3j);a -ix6}RCb_TAMb{PMllKn5y84Mf!12d~$Q)4a^bOk!>Ex$LJqVblKFrLessr3Nj$J-rq)b$8)L&8?jOUg -^UW%NL4$_zcsv*MhUzP%>wW2@uksfG{x&#`px%h^4M{1~+TeUSA-k*1f{{{~P?0|XQR000O8%K%wh!o -psF$`AkmeKY_7F8}}laA|NaUukZ1WpZv|Y%g+Ub8l>QbZKvHFLGsbZ)|pDY-wUIa%FRGY<6XGE^v9ZT -ldKaIWG68+{ ->VKmVG<*zTErBD9+?8y($fBJqYGHIJDH?PuFlD<`CZtA3x*@GlvN!#echQ*mIO{p_k=zplRlm~UbvR2 -jv|2_@x`Eg|y=3kKxy2D-+w)ZPs+>d8%ubH7 -ftUmeoEMgNCr{QWi-jv%Lz_U+tADp3unDRP`=X6)SDPV0@})b#3aYwYtpJJ1=#WZs}(osJw8a-|^toV -pFNwSxZvYs*<(p8FRo9Nsd808*{aMSt_xHAzAC@) -+*_zkt5dT#<#^WC)EQ0kWX6x%YFEo#1+AS{WEhd0PbrxlJP>?r%}O+FRj4~n#*`$seQ47Bi*p^>|=7J+h*8^nEyKU<~ -WQui?azzSg7P!^vY4S&s)Mr;jXVJv}@U+0YQyqu>4H#ZXBTb`FJ2j1HHbnZ^3dQicn_J;B{7AmG3*dt -C8AI)W8Z8^cf!A1t(O?#|u&9vR+VWe+J#nqx_xvaY??I>Q$?Uyq_~?+TfLJ}5;3PqFzX{y`de2&?V9< -7yG%o)4hwIChQ<-w|HNy9|}-QWOad>sm1aUiClEv#ZdTCBYAbqbHb@EpWl~Ss|f3Vv&Kg2Z|Z>ErpW8}2L@Y5YOZCc$GrsKuAoQ%l%V`VLS+h6b_O`48&EHoTbL%8B<4!xWoGGAa(|(~ -K`3qsDY(p7YFnK<(l(7Tl$7!sE;QuFG**#wYo*v0a>xFH#aIP0J3s^wp)I?P3a}OXx=J(|T5ZUj(;PN61@JVUHYV-J -_3}G+fFZSuu5c}W2{I^6Z_WcD+}z+9LBwuwnB?5As0U>CrY-WmZr~~jX~{5%72t7jz;U%W%_(@{i8no -YmO4QpbO27f2OrPgFLgV?-B5ywuQ`TD;Ri1F4Obrd+zv5#IJVBBcW;@A -~b!3*uWeNgR!SOJI^;eA(LU8!*KXaM~CS+K%ar|Ekh*?yOI!3Y@q**eU1yV%M7vAuBVM?jPIR=2MasN -@H9-_}b1VxZ^9Zn1Zh1oDj*xmoz{zHRDydJkLq$-2J>hZZuH --T2C~KquyE)Xn7#i^7nI-wmpsmofbs5(5jk2>+mdlm1zL{IL8s9N7=hO|m)J^L>WHy_9uNqekv98B}6 -xaF`v4>kF{ZX8U(Q#DH8wV#NfCkHyA0MEz?}b=f*LDb%)!a<3(|3XaOB7hH!>lxB4%BVT6sy -u0r?a6%ro&k`Q&1;!etMmxE2WHG6KGWTYIBh06cl02xBoi7#0!???6`;+-QO6At8C)E>RF53JLsBNk%z3b;0*@l9+);W(o4arRz?^Cy<*9>2vLA*%& -xqsC@Z?e*ixaG-r3{ZxlQ2jZ##F+>Oe|p*C2`$hAxv_19LMLYFa~T-J9TKg$5he=)=;Kf4n)aN5#wp6 -Tt$}AC9Qq+(C^^A8hU6PMAPF!l{?HN+Yd3KFt7(~%^5b%N`hu1;|4BWJ1yO%Pf#UIv@Z08#8nDvb$XQ -M#mx~GBj!3$iH{45kxrU@ZQ30o1Rc29-*}@^vd;G5VP_xyFfQm&@01UO6!{^PDS>@f6*%ycdNL7{Jl? -xV62QOoo~2yt@U5oX*xc_Z3u~C1drYM#ToSo4Z<~ncL|#)H+QN++AJZhadZ+S+qmvZm@F~Ap0{dyH{Y -8>wX$4;+Jxc}x+SeK7(rJ>^=IbIS=0esxHeXWlg5+G)YO9b!XP15@ng5P)H%d;SyfHV+jXPvBNe -)XyB3}y^#NkAmogD8_DoVkwl@ZorQnMR7Q0}v@Df0T$|&8MD9ijXwOcI?MAQ38C@&0@SkF~{5XTk@3po5f(37oBC(Eo_mAXRUU# -Cc=QXM*IB#N4$0niQD#|tUtewANAdU?yB_A;P~Jh2vbcA*FB(t@B%HpWIOQ^@%{35(q+waI17&6#HId1*F^Xse&?jM&XjEQuMmbE -)p5(!*A3r>b_3r2uc!JX>N37a&BR4FLiWjY;!MPUukY>bYEXCaCrj&P) -h>@6aWAK2ms3fSzCl@dC}Je000&Z001KZ003}la4%nJZggdGZeeUMb#!TLb1z?NVRB((Z(np}cyumsd -Bs-SQsYJteb-lXnFs8OB2sxIm8}R{EZP{`C5t5#P%xIp*6hj})r<@{U!T(>+t}EcNu9)3@pA -PN#Fes|?z)r7_Wz9U_p?}=te`Iu7iCHP>NmjWws#;JbpQZ6KT`Orx&%yV3@nEFYE0an4=*- -Zoh6?e->VRnXzJV51PO7oMZxl?+WeqJ^WoycGC4VEUmy-PfLLp$`%Iw)-WwWYDSr?0JWeiYC>=d;#R? -ocx_$5>k!=cJ#g$Ni~8=bE*gvz-JsD(m^s5KpNEcNrL=k;~H#uCEvL0DZ;5>Q|3<;;5CM4F|t7O`XUL -K-<+)4~V@JNJPqnVYggn1%G&by2MeA!ispYlN!UdI1;fwFgZ&XhA=3o;k4`5@D^*lmH&(I$JH_nBZK| -T$PA-w<#obrkfp|9+Mz;Z~~dG4yQJKxkR$;_R~@C)#*cmLyh7}2ZQ -Aqn#!WuJ7ZFXz=jnBrL=;bHl8isbgJ?jVaEkAaPuFpJF}_L(8c8@xZ)ki@;pm3`j7I~XqQ53dG@a5o@ -#4$LFpeOLNB!Z|ARfJ^Gwd6UQyRvXaSG6MOiZ|eilZrmT}Da&0-xboJdD#D-#d@f5o4Z@6AEb(CTZNi -8iol?u9C@k8i9TQ*ik$>PrwpgMx!(UD`XUX#0O0;!r_pqyzmOvCu|?}$CH~Ret(hD#dtV~AUum;S$H; -#S}HizABORzPlNC>d>^@8W1u7+tF_kA^+m)Y<_q!PPvh~3-RO@;X@XB5UMA^o>vcShd -P38;*|7GJ;Gzqw@gJ4xaxDAyfFa^$uyRaWVdtF=>g!69eiL>)1|~&t3{I)4tlL*nLnVtN!w*wZhBh=w -vk_MScjfOsd93fT%3aG_>7={w;NHUtpD!aJLuV_yAOVR-0Qj3+xu@T-5$@iKKAK&Cd~2ucEf@2I}A{} -t61s3ZoI49)+6XKz~6L~=_U8euORI${+lDMG3_q*gK*td{P%U=4mCIM+azrAy(^6XSA(hozJ4CI@$W8 -0y>EuU%S_Kz!roF~Zn%#BjOM^}{AqU0& -QEZyg$7oi^)suW}Cdad_y6TvDBei&v{0VlJ@7?)pJN)&dQ8Db{Vfh2$z3eiON9n)>wQM81e+jnaku1l -UJ=RNv9~T58zGpvLiYH%7NZIDfal+;771NvjIv=03u*BLe?=;ZUZk1uIke{Es%^IRx)3yT59=FR1Dch~8?s_@g8^`U?t2kF?Bx2H -e@6aWAK2ms3fSzAhYIX~6`00 -8#`000{R003}la4%nJZggdGZeeUMb#!TLb1z?PZ)YxWd2N!xPUA2ThVOogQTCElsU;3vD(y-O6oiC|1 -X`|X+Dubx9XqleP23;XaZ+@pcLtgVHEHsFp3 -+jz_0(d@LvoO51mp+k4n4QEU!;i^$EFQ6kO*dgp|DY2m)$sp9~1e(MQbh;TNWpu~dg+~d7x9FZ?CYEI -x33t6VZel$c0(7UI_AvhXtxhFrSjX6pvl2k!JIcVnT`ouLE$G4ZgQJ%_bRUQc?$$Sd9tf?0$IbmSdt1 -NUJio5tCtO`K#-a&tF*<_f2{j&z6$4sXau(us35|EvbJj3s_gKq-#X?NVCG14w83Ihe;z>HbC&eU{Ta -0-tjK$*b$8;=U`3p8$VdV77pc{+=FTn{ZA31{DsPrXLkPANX)R|{Ly_mhMo94+AZm9aNW@h6qE7(6 -!O9KQH000080LuVbTT_zpU629*0Fed&03HAU0B~t=FJEbHbY*gGVQepTbZKmJFJW+SWNC79E^v9BR84 -Q&FbuuxR}k){!BQJr4+Dawz%D!Ww(e9E24mB)R@o9nxk)bj?W6pW=4)NIT)VLxvSHrAkr~= -r+iawk=3{;q|ofI=o^!#R0QjHKwB-cr#)I`T?_4T4`WNyH8C=ek@;A9@`^~NDkcs0na^QUp+Bu|uTPHW6TwU!4tj#n) -`?l8Qm4S#k%j($rqY5qIGv0NOKG!2ztL@Srsmr=+Gjg%<;1fWV1J*uQdj#cqSRb;%+Y(E*67Q);Vc)y -0XG;s*pmgwMDV2!!SN!?);iSPN~YLkZs#ygKGV!{C1oE!-`?K!nx3%n`C=Ne^`rcG`Ueq(NiM&`4e0Z -{V*?uTq>e*h_VYU(%Rrp6EM>SJox|kDG5jIzy9RM7@ugm+(c4ewet8^nKkYZL)!Un0ga>L;xVQFMZgt -2g6!CK0kTI!xM`Q2-9+QYlp44wpO9KQH000080LuVbTk(Rx3eN-p0MZEn0384T0B~t=FJEbHbY*gGVQ -epTbZKmJFJo_QaA9;VaCv=JZExE)5dN-TaZo;31{^tQ)^+oi0v(bL9t=B(19(LcD7194iBO_SQg(%|| -9wYNlI<**znE|L_S|!K3d3+|>O2RtWd<^y=cqLBvRz_n;I1)SS;CGkxGXiWGJ{GmlS@?u8P`Vfb#3fb -9EM>K@S>8+fd3PG9V^tbRw?T4O32NImz!=Y`wQLZAXl;g(^T}J)dAB*5F`l`B1zyH9>Wcqq+({9)I1A -E5NeZ=ouqo%XuCDC5VW_;)pUNdSS`b+APB5gRHUFaACsDyJByN}WCbS4ke+3j!Y~SJncRI@+}+NX%b6Ezjh$Y&@Q(3|>|z3xMHehDjd#JO(*ECjE%wzw0|~@l` -;A0RK^w@W5OPncD+DiJ^dW$)F_oT-#~W_8^*T;vF|L}5jVY3GRSPlx_3Hf@@h7ddC03%|UtN5-xVZBA -?*vORBU*9|d0nQKV3wDprm{?}5sd6vO5_5%|1|yE($gvU%h7h^iLJM?78xJ`Xlf!-0~-l!&zd%xiDF! -nj>Xbv0ajb|K?B)mU34H7q^ukmSR>)f2#jN_xML18MnzH+`xeQ3@p&ec2c?c>g{4D(WOs -|?{$??seSv@cu9F-dW#m -RcJl6hPpBdhsQPA1w-|D1`k_&t9d%UBnH2kLp|)7h72o(DKk_oR32CH<)sfp*VNiH9g|9hl`W^jwsXo -JH6_X^I!j> -7eYPTh;VQupBU#PL8#a{Qvb_S)!7olro)5Rm5JxyMB-smmAE&4iY9T~#7_hRU4^d$SC3J-o=cZ<%WYy -a0v*n#pE#2>FuF>af9JGp`bX?G0X-r_^wPAXvP_yx_+8EQEH1B8dL=@NodVz`cYpdRrWg}7Bcl_qpErLVe(9{ -kcphTQS2rXjoin9cYE;kJhCx0?|z<~KJ?3E$CQo+v*{lcw5_J$WWchc^VoJWWwfo}C?uWX(9WFb_z7GS<5R&yr;)mq{dRGOrJ=U+FHlPZ1QY-O00;of09jiI95 -r?I1^@u&82|t&0001RX>c!JX>N37a&BR4FLiWjY;!MTZ*6d4bZKH~Y-x0PUvyz-b1ras%~?%z+cprr> -sO%gq2xrSEhljj%_yC=)9K{WX4*_Iu7-h0NJ2~!90HWAasA)B3xXdYDcMf?5glX-1RmZl-Yym}rSxZ> -@q%eiSQe9=EN;k7=nZ-Q(|hvb;wAY}ra9B%S|md6$p^Mp0i|>@SrsxRVYn)FS@19$#)_!ZA~Ma3yx{O1W-R5y)@_%Ekg7j*%A#(QKbeFq&w&UlqD=^_$O2ptAVZOGub^J=%n -!XksDrh-;We?SR9dDX4ahqUF5 -HuAPS#S!D4v3+a@%JE5Q;EGN(kksOgThLOu)351HlNBiO|T$D9>bwz(JsQ_u<_=4bX2)d)0E|Vq7IW^ -DL{7`eeTJoq^~-qW}PRmDpK%M=S04rFr6s;;qn5Dl}LHi1duaBF?_mR=D1G-;5~*pO)R=sb1Cb;UtLv -AwJSF(c*ZJAjPjRNPyN?7$LywYYPutzXxK&RbgkrhdTMM17xs)KQ2hXN2@YPs@N!KjndgbD;FIK8J3G -M;ai?Sw`*}YI4oym;SH%`d^AYDfwGbEnQpVwS*KYSS;3_&+b1F`Zy)Qcu{Ef&iV2F!L80P#{Zy_nRu< -+LRAhw~Sh;x=cG2BCV@%bC-DEG) --nd0-D1+jJhi1@qe+j0jnh$Q|z*(mc)E6(@g@Uu4Ez8Q8{oy`XUDC+*<=?uK0_rcrT@HK*d*s|Bf`5J -Aq%ss~mvIl65G4kj|M5XDV*5S$vk)TRR#$3}L>9g#teaOk+$b04uIV}`|FHc8)kLM=<3D}_4;G|kp$7X -7^&`RL<(mYEOE`HkKu2{0iMS{&1J&KyNP>q0S3Bo1nTHkXT#|Lrhk9qHA$vE;tVj_v(LyikLM)H6}@w -3yuY*k?y2<-I!Bc`USILJ^2uR?s~e*uVdnJ1sSkoc>3M8?8?PVO+GZF^>2Ee4m{+(_Hq?o^hr>YO464 --HNh;@&SP6J9G6V9_zEK~8O5$_K@hokPX2=+~+~&vu)0m?U9sud{=M>LE^!WvyHcJ8xNIM`OpkF=pP9 -zELyi6%LM4983WlLmPknewZ7jy~KUY%Rc`;L{3$H`unWXCd1wk#2ComU$_ZK=FlS^bvPk&z@af&*q%f -i#E?1-oE&`l^1#(`rw(U(BJSTZ8S8sDv@}0k;3MQv+g93#&@+wV39u^g6lS&3SV-$MrJ9Fo-cZyMgTD -;jVc&v)X*zrN20s+0ds=-m)WG2`wJ8gVyt@(B49us}{gT4-b=;Z_17+r%fVk2DbJ#?yYwlcMt5`cDrw -ezFN3Hr-SY&mIZzdUo|E!dKcmKr`;RJcAMiybT8c!JX>N37a&BR4FLiWjY;!MUVRU75X>DaLaC -u#h!485j5Jd0$6_cJI(S87l9Q_SaVYQ9WlCnzpd&^2ZxTneN+nF`STvNvCIbseoRu^S~B=ny9C&V)bC -yc^KnAjR)h{80dEV|fDA{zK!js~}NVd=)==T5wf)y6J22vJU$QWSsa0;;NUZgER#YzSpE0heE}-Gx=R -s+D+zcX{-W(3Yb>vujc~v%GRVQrqt>eE?8P0|XQR000O8%K%whzrM=1QUw42a1H4;ZWMy!2Wn*DV>WaCxm)U2oeq6n)pPxKJNv4~_!t(U=PaO|}K-(xM5_4Mic)5-o -G3NsXkO8lM0A?jwO%ZqLD@_w9q&moiNd3FOEU3Zd$TQdh@(@MVK%2{w)OQ9P41g_W;EjY6~cx -Fc7vIVx(or9XUhF8o9t!$Cy=rCIuy@9e^b*?kKEXk)f4wzEf1;Z?hYf?82I_rP-zJm+5Ybk!*yHUFtJ -n^sY{#^{winb`XT-DkXEvvpTa#d`(u^6L2)-q%Hc%WY{%V>xBF$+~^3>Mul0=$BUcTab7aNHskEpycx -0TLDz+wdaGZtvcIxO;Q+bpP(5{QdFQ5`X=2{|77}Pj=>jXk?XS*1~(y-i}ax424eO=X9O{{#I>NUI#IKsVhyfHpRR`V649wt7&0Ek1|Mm5`i@pO5x99$e`;2<1kbd2A9ord(sn;{kvEbaM(=1bq!ah?RE*gtvCdM%=uJ&}D1-yJ -$X%NX&+^XG4v%SXx*Kzoye(FXJi2ahLi62|r2w6M~~^I@=yJ{LX3iiW$ler=L_;$YLbK}?F2!TI#WL2-gskAh7g{-5B@{KN`n -9hO~2H)Tj;axx&V;4Vm#?hGWb53KN3y_^)*%NhoAR6d4bopkvNJ(~e*TU3?Blj$fe^3+JZWb$&QSwFH -bPUZ1LUN&1K*(SdGLzg!3%yAwbz`mWl4utg`hT6pO)>8a?%cI5eSns{o5L{fCLI -O@Jh|AXE)HJv=zdOl(zV#A3!^iRVyALpVIIIc8%BP`c4P9HDf=oJj<|-uV&0#)S^Yp7fQLO`u&Ilk{#3$|3`cuPItZsH_zA;#BN7cQ7KDeVuurrgk3w=+WGZ$Biy= -M7MS^UyKC7sHuc)Ove)~=IjEnHhMeB;kQaYxp~hrLjF>C&zAM7l$(G5EPg4W%OS8n5$azGa^bMPuosi -QzN>}~4X|EGrpS>a_3Vp%_{E^&b{R<^_Xw*RCU0wd)WO(r%zoqEQ`QvW!!cz(#bJZ8gw_GlMnr8n3P) -h>@6aWAK2ms3fSz9J3(lhD<004jt001cf003}la4%nJZggdGZeeUMb#!TLb1!6Ra%E$5Uv+Y9Uub1)a -Ak6HE^v9RR^5x-HV}W`zhdw~y5<~{hdeE~B%}lmN?)4hr4*sH$7l6y$(7{Ymh}Gjj^r=NyL+J|hGpgP -eEu}^lWMIxV5}bOpph{UU9YqSE~T>E3MI{MS38|v^f?pKqL#ePJ5^z8@=9ndEgN(@8@|_NBTo1rYn8! -I{_}6T^~+KAxioI^g40&Wr`g@lw7#Y~SV5W;*UXwdw+*w(<#Ns7R4$sd-0gN1*083}xM4l^Z{NOQlTc -%g8fA(5RqGH|BzcDiFjlAV`bYPF*a1CBl6+9|gbu`29ju~PN(nUnHITGJX`PXfk2#rm(=||j;;pE-#e -E8J2ciZcy(k0Q2&%+P-j2hgHx0cOJ~g<91m3~uxz_yAeKYj&_dmSI`7}Zr%i|hz5pGjD7PSp=U%-C(- -~%LIHZC0E&-bB}e$VeN)JoNL!ZJ}kCyU3_udw|}jzO>8&NEJ6(GubAkC`;wG>5pV8Md7FJ<4jEf;X9y -HMkmcV_XBgl>9Uh>Ho+(&e;sd@lEM}*2otrC!t=EP+Q#akC}tFB{H< -R9Jg9&hLy1V*y7%(L6_(TZ=I)Qat@)-MY?t$c`Jh151gghy4!?wF(JO6;})OvfRj>sB}F0;=D28#!03GhV^|(&A%@{Zd!OQ$cp|MyI=bfl!mCS&Rab^P_&w)Z3j8jZL*d}HGAKIy$g>BSe+1wSIs -dmpDd{i1ffB4l~nGDVd60veDSqmzX?CA)qfL9eP1d)fsdjsPrr8{`2i7#=jaEVD?wMNq`S(%ZWy?wQ@ -_pUP#7CPk8Y@E$-!I={u%Z;t*y21C^Q~@}$$Uu<~eHI3t*RvYh3h1b%k0;zXI3mg^tsEXC1?QID$&%o -zl>A||qTJVh+~!Yv;gfcClcVy9kj*2dsXH#4Aktbkf%@VF)tCjNJ@8O^q+GdTsw}{kin;;4ua5geKOg;2t^~x#wnIuQh3wT;6?rBlALxb?6m-+8@@wqwlgy$6w*?_s<%kZJ&xGid`3b~WY -gIktGkDy7hwkO9*|vVE%_G9>Qd=)lZz`n5t*g&Yjp06#1u#tddN2X;7|RT1@UiTEM1WtZu{tr{~B^P> -F5fc^TF`F3J7Pw4gGX5L^9^@XvJ9RA_y675yk^GIiw5MqVQaNS=aM{|sK!v~id^2MI0Z>Z= -1QY-O00;of09jiFSX{=40{{S<2><{e0001RX>c!JX>N37a&BR4FLiWjY;!MVXJ=n*X>MySaCxOw+m72 -d5PjEI43vjRh%INcL4p`)fucx(06`vdL4Yha1T9fA*EA{cvX0aC-#bIytcAB}wtBE7^337z%&oR&4~* -5p2P+r@xo?g2z_m7>dug;wl77v&;mwNS%^h*SCXQCN1(VtftGOzaDLd?%gY^agj+yx!ZKPrXB -@m#uDF%f2d;#^9XPTC@UcN_sT&q_g&Y%V5qGC=rVVS9*>JT;@+4u5D}@*Q1+F466-~pG)I0RbZyCA4@ -HX))UEB*9E{2fnBuOe!0~=wQ?}f`OKg72+IB#=U{t8Mue}x~{>y9@~>T*>df%|?`*cBF?bl5?nJT?;F;2W1*r7D*)gEPyL#aLw$XHa5D4azy7LJ4nAZ!Gkt-Pvl -X^oV<0yht?aT9K3i1<(BJ4RO1B?{75H9bx#zjrNZxQ>wW9it5t*SVN;Z5ziN*yUrn4V>0A(Q_59++%e -U`Ip2ylRNkie -_@!km?zKML1TVLlDODfb0bPWWzB)B9VQJGY(Xc@Wte9$*2sN1okw^)ft<$z-hv|FbAu|qw%}FQl -sS9|x>0P`rQfK2ii@A9|vzGN}B^{}d4d8r!%t_D(#Azas&i99DDE^-a)BBE>(wKP0s<;y~^dB=^e8Wb -8-O(Vz9;Cz25o2FnA8}D&DPMvsh2~b8Uc!qUsB}?ae$xMjYkPa7UQ$b4#ZaM^#itLi-?I;IKE7U%GkV -^gqv5l>+f^tBii^yJmxi{Q*Oa@=mlrQ13?zzAa4aa_%7jxOca(`=y0b -e;6ceBY4EPSQCN6gUoquCkeIQ*9s>rN+6bnB`HGaT3q#WJW$>I%1*R2^x=h0LPCm@X)qIG(#G>|JRPr -bt9tv`9}1^K6JJX{R2=-0|XQR000O8%K%whw$^TKNCf}@{0sm9A^-pYaA|NaUukZ1WpZv|Y%g_mX>4; -ZWo~qGd2nxOZgg`laCx0qO>f*b5WVlO7$^s80YyIL5CaL|q(&2qVOX-Z! -MQAKq=6Hb7umntAhP=FO0cJ0I76qxUH`bHT+IZoW(QY;y-zhJxH>P#d3$uB0&EQqL`E3oSnt -^-O=Y`U6)S}JHR^^>gdJ4kN$Ltu(<)!I*;V0YoZ!drN<^A<+R$P3zc>V6?{MAnv#pU_W7uWQHZfAsl2 -~d(OJl6dD1w8Vlv<_JSyfz5_$=I*oW0N2GpmcP;o2h!@a~7Zq(Y9tKt3p0_soBMW)t4m6D9Po6Ouk05 -f21DhC;2C%blMB+-F)-9HkQIG;=)ZW$a`z5q4eniW_FBCgu9!~W|gd|SXt5c(x$@RyEAH)^S931JNk< -*jh0}=Kj!rGC1U!RVv;1UrS?{|G6I+S;3S!vlyif00ZUlrl)k5+m8fC -qHgbfB|jSP}Rs?L5fJDkCtkj@(e!NR#QV$*WERWQR4~OCRZVW(n>w?Hj{2kxM1KL=ffBJf&lN^lYgrz -)&a4Mrh@_P{ft6nt8;qunb6*vJ&j^K5+Gh!=v)n^fsXPtte#xXk^=Gw34LFYOH1KOdZo5?Bga8nlbj# -1qj5TaGx~H;h-rA9~05s+9staZ*y_eecdrPw$m)h6Qhq(|+BHQ}~=l8J#|R_WXHNY_( -J<8$fphtB0}BG-%}6u0s~=x*wF=yELjATWigtMQ_$-pc0A3v_~N#JKxaB)3Du6g1jW;k5qQ>?4U#Ixq -7kT?Rpb2POX(Fg$~Oy4D5WIV#u#pNaqWVT83R+Ai;u}6~Ifw0AsE<1AN0eHdL#7uB{n-NSHAo!SHEJA -dbVHCZt{}YjlU_e3HdNX<^qzD+MW?95tyk(6 -;p(`Q86fZ{b9IZ(chT#Jsna^|Ax5}rZi<5oqEC?3hfCd&J<`K}nTFtpBTM_{S**=0f+dDe=BWJ*_2WQ -3BG@3*E#kOkT6VLU0)u^&bC#d=@*`o1;#?e0x`fb(PM)Q -ruO^cP*IF>3{E--Q^M&=4P*;3amYBf=9jIhzBaYk?}^@UDzL@gwX?z>poS6tRSly!Zf@e-3IXTE<*8GVbd -g9>{ZaQf`FBrGB&FQvQYSM@w8LsyQFYgUdEVPzk<9PFc8oS~mw3dOT0KMulD1hKI-I^Ncp%xMw@V@Gn -D954bJt#EmUoj5kK`F3r$Zfw}@EXF3U~;&65N8=}RfW40e;_BT*V0|XQR000O8%K%wh&)r~7RR{n8c@ -+QvApigXaA|NaUukZ1WpZv|Y%g_mX>4;ZW@&6?b9r-gWo<5Sd5u_YkK4Er{_bDFIzVIJc@<}`*}E>_q -IYT11}(COydVLBgFs7^%`7ddB(+{0?#K6;q2Bzmo9>6il4pj)d3l~07tZcPnih4aok~-oca?QP$g;FS -25n0}nY8bUawkK!?H;WkJ|&e5{TEJ_x!rY7zVVZS4#})BDr1w0TxZ=$|5T-NGFaD1l|kJHqt~54`(`I -gxlyfDP`j$oMs+f8Pkt@DR_n=;O8U!Og|R-^AKlb3 -GCGdm`QxVy{GQ%cntX<((2z>}4Z@jYdS+Dxr#kVvjH3zpw>fYA-5hGv)n)M9S9MUOhENe62l-!25cq@ -Fl{46W|Bi%C&zagBbU`^zDkB43S*%uKy4U3oms4S#KpDX)FGFxMi<71Zq=FCAd32xtaPQic;)AQz -gkR!dHPlt2yB`z_T235h?gdWqo>xY^k=>R$WMS{x+e{axKh^%%_l|j_-FF8<*=G)87<`jkHq3LM;eWfxj4PsK}hH4Wr6VBRbQ|J;w%kcr(V-ZD3s+2R$8Dp -qm1qZjvDOUINukdwl#a@T4t%JDumne+fBq!P^wNw8%xNVgqxZ|R4<9{rz?lV;dar6F_bJm@=*yi`OG+ -vzuv|kRVB5(&n3RmxM-h!S#r8oZ`{5;j}N@Ec-yb`@87YC}l1e%PlJ9eL<-nK>Gb6sCB2 -2Lh_KLRZ3W|@DL844#2yDBB%8rpdwzWeTWFWtsp6F_ly_n7!)Kr1Tr`0sD=KGW&+UGz)}GI?S2Et+)w -70oFs3!tI$LLEtgQ&~im@QYsYO=B;8faeiPy~v!BLFK|L6uL4EXdtKes`PlyNpr2ZA%ZIDDcAlsc5D% --Xkn|pJ)$N7KVIlS`gwa=yeX;^7_+93q0`xJhv>;ElrxgSsEDo2z1 -05qL}r?kt>7tU2&cl!&~iEXKP~+K}8k%zL#2~E6<7fx*)M>f^or$6k}}kwF#(FVyoFu4# -iO7C!O4AW%5?D|Du4ZF-MY7CMitqx3k0M#%&7Ul-aD4L99wUn*)L;P66#R1+SR0WYwv>K%R1WEHD?O-)UP`S?1lh-L|L`X+9RcVw7v*>H1!vaP)om2v{2GY5wUdWA8YDe%) -MQ=RHoV|q5PSho#DkvC&Mxz(KO1L#oPJNJRQAj;G$bkjN9MM+!M}p-6n&!fm;@6w^Z$5)*omOq5G5RB -_p1%I1F*NP{VaYPLA5bnJM|lDrrmzKNHIHp)EDlBEOC~7hY?4hzvR>a$PiBY*`C -yFX=PRunFt{on5jj7E^#}JE+FN%!%Ivk+hEQEUd$z7Oh;Dkag)CL$ESDVkNElJyH`Jd=pJ5ue)*r#3=wA>VvAQiIi8tyh{=u>epL<&fyc -&kM&SvVmnfH81eCQ_ndEF!lMtGXn!d+6Pcfv+CP%h@p`Hp*+v?30#9fy(4VdPvxeP==tP6}-9N7VA#s -#1qa@gj`wXX3AwBbD=%7A?jUJsPESCN??w*}Bl1^KGIq_K$Ah4e!;`>tVo4G-_Z()rOKK*#KG0mo5)6M%ua;Av9%n{f*zZrX%@pA#nS-lkhTw -*24)cL<^BFT8}gX`(B#cW5(ih6)l48uM; -$<;3Q6*?*lU*@dQsZm{2j4&JLg?9$G{QD(KMG`oS8|j#XVYKn}FwA29=lZ$L?5tPGo|haJp*eZLwduQzj$Y5wKrAe4{Z|FL+UQS5_?;kFrBu6ok&gIe3>plVvl<1o~_h4J*A%_-``T7$TNgtMtU(t_q48YF_et-&tBZLjTdT;m-(q1m3Hm|`zzVim -+YSVVY%Plecq5Tq6rHjZ`kqm^xZpolzjj?-z#=dpN4U=tgm9~m}UC|F|J}vU=@%>q5Hi}-q%`7-x(KB -Tpv°tpolW=BenS>4iAbjQiz$@;^{Z0|XQR000O8%K%whBZAX&P5}S_u>t@9AOHXWaA|NaUukZ1Wp -Zv|Y%g_mX>4;ZW@&6?ba`-Pb1rasjZ(pC0x=N1=PQPsTxtXLD70Y9ibAnPr4>cisM#9GCLzgIyIP!Y!T9!y1gwtHHJV@nPCjACvW{nHrt&%|;0wJG+ -IDSpy*)$$ahHoQy1j_b1MalELb4gFS5OS-;!Uxv|#@G^pEy@CSY+ZC{8>_;)pb?qPZ&eY%Qi;z1FQ7A43MaO}G1hS6gp%{PDJ -2K4017{E!?9Jgcvv!J_QJKIBL;~}nAC9Xx>8aecacJBai1$M$S+V!0|XQR000O8%K%wh_fo6j2nGNEj -t&3-8~^|SaA|NaUukZ1WpZv|Y%g_mX>4;ZXKZO=V=i!ctyk}B<3RMZ2m9EX -vQCQPvcrra!k8)Yirc`OR{w`aCt!rajRd1pho%~FAroZW0*Ye_s@HhD^VWKwv2!$i$kHn-~x0h5wuuP -r7s^6MS-1*iWKWK2B>AJup33EqU||ArL1J_qeJ>ISh#>TNk9#^I36vLGW -1uctRz9HvQ!9TWlMG|YV!Uj?^mmYuI$v1D2=948bhVJm1Q5$tfo{_+Ui?b^W`gL*l!Qv@n*%Ftaj5{-l)@Iet*Fv@SB8rp)7IL?T+HViWw%`l@2h#!+pVtWS@@rpI-M6;vB*zP -&(h+@;&gF(n!U+S-z*nJah9K-%Ee-No{2?zzL>96Iv-3WO@F*voWD9*Enc5FdmsSlmF#`Fyrjua)o06 -z_h-Zf&W>BNZGZoAQ+MleM(A=B^Jwz!X8rN{>w0zd_TBp1htD@xpROmf{b5*tavs6E=fQI>JWlR2hQc -}IGcez!lP10>mdWCLbj@{4_jh>gR$`h`*~q&t?=sm~ihi*GO*<<;w5@Jq`g9W>zKOE)i?fTv{<_*Z;I -k`7@m;7=q@|@QbAv%JX35{VK{b71`~7E4@%blcT2$2`c7yjN^CZUbrwFC -R?b8fc*~T8S(lIg@w|s;b>18J`LriZe$%@e_gF+G2}ESVO;I`7Xhdebt5(?NYR^^CRYU&s+6^;q+=Zp -$9%|Ft#<@f41vf?w*wWzul8uD@21t8R#cN90XxytPQs}g>R}yDsp?zkOs?Ixoo;VgMhdk-2O!LTs)e;iKzE~)B_^?DT?b&}J_Oo2bPL(A|UofQ6r3fzu^t)6|cI?K6DD}l ->{NUox7YVYh+nC<|*TyKp>=xF|*vo+zu{nxf^6Um&6-gm_*&Xo5W~Ls$_1kb>!ndV3ib*vfaQxq+@D5 -QDhzy2mt1#IL!BC4hd^z)!5`|Nd|VUnfj&?nS|m!sj_jYhHJl2rcau}M%u@X;fdMg(8jUdhYCB^(wn| -YHG-q1;V{TM1x7&6_MwkC93rG3P&y>QBlvOWCEU!bC`SKu2=tP(z`X$ZrdB&7%P-Ce=qYAkQ7Q!-wi#>%rVy=RFwUM=9;8Zdk@LP%sq-Jx}OHmmTZ=1u+G&%1Z4!#}Koak2}F~is( -!aJr8KyF-_uznfu_nO7B3TKL#TI8OE2%;}W*M7hwI6o-N~?Vza?#cLwriY@n`H3H>%k*IJp6lZDBaD=b8#rZXfnYp6NYM8nnODr61F`+oE`lOP -)h>@6aWAK2ms3fSzApLULxfN004#)0012T003}la4%nJZggdGZeeUMb#!TLb1!INb7*CAE^v9RSZ#0H -HW2>qUqQGZDg%yybuSn^WLTD^tuZty(yYS}7&0wWHfxDgNh+?<^}p}#NXe3=bRAZINc2KJeD3bK<58K ->1xu2wa+MKD!emiu;}};;JMN@bb~Fmlw%l%Wxf*3SCDnN@QW_BRRa(!RJ7M@L7k-dClu~W#cXxlizDh -1%US0k^zkV@gze>f;VSY1Z*ZNk#@@nK~m9mTzsQ^jwJQg2PQPMg!&P(9&O4@?EbUUr@EG#r-b7dXRb0 -U6aj5b4vm6N%R3uznHM(}l#ZY#AzYqKuX*@(f9aDoTwX$h`98jYT?yR86_R2PL%Yq6$*jOQE8bjBRKE -Hu1vF}vJywGr@0TDIQ`B_4#K0l?b{7!6=}%K<*6vGyk>9ouX56e3W;;}uVLdv4Ycf}#XVS2CCG5RaZ; -yuG=d-(4jyFW$Ul=WJ}Z{Mq+EjEUeiFN8%hwFyYhNVhF4T;_TYq*n)8`s|F@YjQdWVa=9Ha3IZGE&=P -ZG6syv4^tSgEs_;9Q!7}5JC+M>K|+>9A}zq28DS-usOYgL7LVTC-QHZkAih^`fbo4zc}nEJ{qehS_5C --`#tX0`&AGJ{abbN}#^dn?+Z$e%o|aarAbZlSNOu+~7mRPX1kn)KPU*cm+i -1OJ4?M2~h#{=uFV-T1RFO)$B#ErZvuQw|v9(Oyy>({Fz+k#sEZEjzN?8#MfF#RmWZ0XaLkVjcE#_VH8mP-eBmK%h_pP8d@rXoIAtY$pyd -Ooa)tbfU1)q&paYPuMHI1BF9@34SHT@O;k!z(OVgiwMjC*$5`yR~!W$qBQBj1>X0-VS%FA#dvR`2{GF -tBaay&!O5Zlz=*}k(d^o3t=NJ(n$T*{hnWSnQS{ov3(i6pfx@xufVv`@Gp*-zU16xkfkWNA0AXT1s2B -HMnrBQ%1gMLKhhcj44k|4M1g1zM@&gk^2_A}MEMlNQ5oiI~Qlv7IU?iv>tei~oZyi`)nK51cS{~g;gd -7r|8t6SSWDI)hbTmF298556gKzM`URtaiU{v5oSzJQF0BbRrE0`Bp+F_bX?7J#%BpT!+ -k8w_S1x^3ryA3+5}g;oq|xOA(Re_pnnvbtJIFNtDAM37ANmOJCf;Meltx31bB9K)KDU!pSIUuONuy94 -W~X4bDxzGdC{zQVDB-5ZC-!8c$>f+A(k2+3dPuAGC0LYZj4B@#oW!Hj1ERCiI`0%bztiL1N9sczZsdb -dt#XA*ugSCvRpGzxaF7#?cT;G6l@@L*BaV(%0@#Bd*W|=(036Q#Z%!sbLa~kk?$KIlyvAFVo ->GL%iRL)yEPWQMG;X4Hf48*(q9hj4Th}lwQW&0{zK7r>{}_XvX9N8pT?6>$6#_V8XH2{+p!1+-lJ8G# -QS`3SX`A58xaxI$vQ1@ZG4#^z{ko7yo^DwUaqi1^^*VS-vQtFrntWK6Uku_F4~fIcbBs*URGNtZC~1YwB>Cf0j#V6nw4!-ziOwEvJF(;ZqKeDaq3$n2;SG^!?W}D#o~XWCeLv -UwL5#?})6HWLY^XGN-gl<8$A)Lmd2U6!Z$qO)*YE(Buh -BCOiRjqZo}ctOZKCA+B3q&Z-Q&k3Oi>jcK1H?`SuDpj~@=+Y|-&c`D?+l!}=ON9r4%0^ocSqLtEUqJ_tar9xC -J37W!jv8Ocz|_I<$4?1a+L}YBS1Q?%uk9}%gw=d0fdjVQ(}x#_r$ht+b|(I}h6X07&{*Z$)oYE%{31=Ap!6b -b8u^Jr01U3x*Z8%brhmOYsTa|b`j2l}$|MQQXFD5hS7C(LcJr$!v+x2~CwMDxvn2i1bP)h>@6aWAK2m -s3fSzAxdZ1)BZ002xc0015U003}la4%nJZggdGZeeUMb#!TLb1!UfXJ=_{XD)Dgy;^N^+r|<8&R=nbM -iaoM;MnPOQcjqOY|Dq5-z2(PMA+J{0e7NT5q9^++N<^`{m9hdH#+PSawvf^o -utLMm1CX0e+w0oS+GRxF^Uh*ofWu<53X`N<8suWjth8xl|gtRZNfbrAw>ndHZxmj))v@_m`qPDN<-P( -I|T<*r~4O0C%twwsrTCC&bj+eQt;!0%87?h)A!R?xG#NhKr503qZ^EJuU8oK~OAlb}LZJw`o>(SD -T_XBe3tARKZg;z|>MKmIeF|f?3ASHbwDcGd!@cqZ{1mM}3Y13lv|hcv;&8m-=yNa=mLbb%_tYRs~pDR -i;I}sYRjU1bDG?e?GZ*eSUc|W(!__02W*9bk=2ZwN4#F*QaN1&#y0sr{{cK@r*3qu6cX%@#5so@zv@1 -S@P56`vjieo&IB(#7Zb*iR)QPpADLa!Tr(y7~?y%a-)|Je --Qb-mw056<2!T|YdrpPLYMxzmV1bbb{PrSUO%|7V!J*g^Lg(sggPDU9`$gYAQ(Ds!I#d9X+X6Z+<=Cp -#zEZbCy&Eadkb9F1~5XNgeBjHo5RDo;3Zz&~dRF*!lk<2SEU#T{3-V+QXhtT -h<1qwm>F%0jaYJUSXos4l~L5`HQUajs-n)dp-12M*YI$*u($=T?nb$~1F$zMN%g$<$JA*=EgB#ZFGoe -m?%dAObBp3Jz~iKAwI&nHWOLx?Zcvix;!aLdEM{y_97vtHleU)P}$O>t9~JddkFazW(#eZ>>-egstiB -$?z^Mlo7&=2qj9evohnT(J=#8LFpri<~iyorg+Un$wK-TYl+6pJfSdXFF!PqJ~m%hX_=zC;TeLmz_-! --!g8~kz;g{o_==9ih3Ff!)-h}Ty=@!Uu}vq;kVd=7mTW7E0u;YdoJ_h^!h1D9=}1%3ZnMV&>KgPu@sA%SrFC6}4Z$%Byj8)NecgnEQ+fd|3hG$FI*1?*e>pl_9UbOZhd)jZf0`U#1|vHi5QQj3og|?XQ}JRxc2CdJY60Gksnw08c}`mZ>#OyI -yc@P5Gxrug`?T$>C1s$AH9ED2tu1X&E_7Z%A~U{IWM!U3M_s#&M~>2DomLRwI(!~$PJxEN{82iCEkp_ -C2mxeMB(VCLd8&mBGdY(>399z+Gln8_;t3Rtql<6ePKlW~g)P*dx<0r;o7@4rT5>J16)P<;(CpX}5=d%9i -5a`*ESKe1HQT0$nlQR`C>2tE0wF(f7FgDOL^5G9e1gZN+1JF$#l`tWzaHF9L1XpyonFQ~C)bKHcA~9! -;&uUuaX~{58_638hRWJVnWt6m^bf{112>|Kpgk0{Sy(IbY{wDP9yi;82UZLqN=1;=3#z1yexqrua_Zd -12&hGi+|-tv;B<7zegAP3t#V>>7E$k}ZK}_kccZQ+Ogl%s9}wE7>z2)fUumu;eFd|vZ1*ha?gFVSfd8 -HiB)~5FK@k5-iZZ;h%jq;IP16@Z9iC&Y3sKW)g$Tu>+o)@9p7f(1nzGK%T|95{CY_3|$whvtHu1<&O{73qhGwk!B?pQd-rWvS$FQfiuG0ETv;r;N*0DI9) -10g4R#eGg%kRK-v~{PF0DM4`sTOOnnK{t#*_i3 -FtqosDIHLwvu$YMr{85U7<9F*msAe!TY9Jlxms)Ud29%zF{TF`rr -!7fsF*v?f*`uyVU<;1mJZVI958^|KJ>_57k0Nvi)_5jX)ukA6fJ#5snZ9`z1`7 -Io`6pBI$w$lzebuTH>558SP3EB73>(@*>jhZlOrbe!Gr8K8%Fi3d^$Bgw3WI0qQz)0O67$ag@rJvA?; -H(te%IXuAvs=ze4xt;IYXx&+r%EiKo|M!(`ou=gME7&dA=##d=ZdFu+&HZ`TL^GPz(l!77o}9S$QX44 -kYDgq1%5iVUl1V?QugvtbJ=k{u_mo^~tqupdr(EaJ(Pagu)_M_XJQ|%F9CJR|jgNo-tv^Sj>4{p+1yF9qfju -(a822pe7t#$;hL$z5XvWDD;s`fAl1e7_@m<9sev*O@PStZvdZ0TL6^sjfMl$x3A()DbpQ?AaX_GDe;@ -n3#<00K0R)dYmJEmmQ!hppJABvXOFf>nc(D{gA^m^zal`?`fl9Y?gRor-Z#lyezIH(jBbD^;X>Rzj5Z~h}qWID -aZC_B;Lhjl&8@My2_wK$>Ia&2#&W4(GoS7znn9f)ATP4^ls?wRBv<3z4`*#P3!wN)cWbGr~q6(gF6;d -_~`G?tEMsp3c5+7OH4RMQ6PfB|HKW`G;fU$1ST=|9E6n-0Ine -fe=(|a;~pM`UGnF(AN0%Yuy^;dbu)tDGmBl1F2XDzJSF>29BOX(2Trmj{jRqY9cW_Ld%jC -DaDAcAa)eHhuQKdbr|%Ni0RI?IW(({Q6v8f=B*RMa0tm@CY7jnbl_uHo=~RIbA-9*$KqKf}}!ftUeHoS8r2kb2eAy- -%y(UJkA{*O>$Bj)oU@aKR9P$N -47o(zpCOSpLnU%imm_UY))<{?Joa4y-smdv~r6#03ZQ1}`yj=nSKnzuHwh=oUdGWgS?9H@$U0i5WG~S -zxkAV8}n)lm#HbXH;_(kPorqG2{#yz_v1(FFm388Y3tv1SHn7fz_6g`E?rE~Zt88eFD#oDn}pz1CATF-|B -;-^}Uje)Lo^tI|qvrL}6uiDf6F*(z}TYJ@2_0U8UyPThG9Zb^Qp5VPs!QdW2so9pXn2$cZ0^o^cLxrz -Oa$dvusU_}kJf(Z-yuJWx>_Vp`*rVtNM>5tJE#0G}Yz0WLidE=~j>=EmudW{f7c?z~_P`u-Oy-bUVjwcA -iUD4ugIf};I%C4;ZZE163E^v9}JZpE{ww2%YD-e3misMq$!*002db*_l*nAt8yu#Xx`LKC8J1W%T-y`A})%uj_ -a%})Xt86CabC_&CQjpSFuv&b`h&Zp3Tf{US3^g#g(}&ZM#*evfEW$FU$kAXzDDtcQB^C-YDlXDT{e#0 -X9_*qlK%uQgUaGQ%K4@mkA*p#axu3dqAbm;+)@RrFs0mt^oe&nSKrfcY4jU93IW%Tmf3Ogg<}FR#9FSS5np72Z;r9H9g) -U?dV0lS+(eWhxowc&*5Pp&Z{_)vpBiNTiY5CG0puQ0smeWn}9C9&5F3%oSyOBt2obzP;_^;Lb!3xcjx -l9Hti7jU*d{yl33Nj&SXa;ovp&?MiyyV!6GHsF>*s#0g1TQh2Eb+s|A>1KOdz`SnaToyOB -LsJy6Dp6CU5^%HccXpx(CJnmaSX|KBdeyS7r1W^Dy);v+Jl;gendTWXTeFJNtTInxb!{%8IdJ}MZSLn -~wT$aXWq-3UGX)5P^fTwVa0+akL(|T9l`P_=ROS{b4Xkoy?q#7+ojG8D#mXBQ#ZA44kVp}z5c4is$mB -YbcDZ5EXuZw*SEPFYU^>ou(C%_)2X&sTfM{Czg@Dv+dgezFVjD#u9_P?N3y$xddy}1$+vK#Em+-i5R= -K=DvA*~=R`SHK=Vc%|_dWT(-1!UhRR;_#e;~}C%bho}2KGG1SGI|&fyw#O2su83;X1dMmpf+#3=aK_X -1KWM>Kjx_JtP4Ns2A#5RiEstiC{YIe0Tc#BzpCyAKv^qdVl&KC(!@!;OMD%Og}q2JE@$LbOEOusq!ac -?>QKdYD$ZSZ&ue=$O6#2R3UOt_;Z;RzPW){xGbduDf~n+r9;^Ja05D+oxve -PY*hs;AR+Uts@6dL{;w|H=M-+7_%7Pk$-nat0grMpLsEUxuF$Jp_gFfmb`KyLuA;f=WCEnnGFs@q3B% -5!5ET&l~&KFX{P2vJPk{bzb%tqs8Fya;Jpui0nagoi%haSEkM5bV%Vh;xoEjDNMLRRYxEw^6wVMk!~M -GVF~6@C`VkkFrm9nxonKz|`8;v4Z~YRBrm6~8$a{u8k`+H(S8;vWRez%t(k|M^4kkNZz9rcW+AEROcx -_Z-awa?>9}vu>-5eH5E13Mug=Qp5705w2x?EupCr>kPP{sm003Ur*nkhlXBY` -)$VIG#Y`Z)XKW5-utQB!d6~h7oH#JpOHcFhlIXgf3{_Xis=D%mkg8;_*1Ce_7 -ZQ>L&d1AI|L?gWzy8BtnbAhc@v&(Ex}VA?Te$i#P?s;;QG#*T54|!)xEwv1&`yFt{prmns5MmH15vMB -XAMX&5^)b8<6UJE|i35AQj+7h-QV_;E8^uc3zMh`(LKJ70#s|(J -egvg>1u9lXyzWym12-M0{T;jJz^J#$RInZVg|*NzOLLczf$m*EEvq63=oG;PzKXb=>BX->(X_BedsUM -8248||Nx}UfyvWGvAjId%F$%hop2IQyotRN^xjTaY92H9-4v!|T#Gnu?>UyQ7`}=8`00)wW4$JCl|4D -dww7*`&wJPMA?!raA%pY-J*Gi4PZ|QC@(g;p)Aas7F!l6u;-^O9!EVn~p0qarSe;k3xz%qlo(OW^{!( -Jy%^apye9)Yoo@?$&|Z(pAre0A^?H1xM|C54SNM5DkK0LN;G@Fr7a^=uP8_Q(=)(1wxcv>Wefyiv81J -7wfhTU+B{{?c00Dy${kAw-A=P@eG8qj`CtRkA!6^BAMA4$V4RU4iFUrji -vqra#z@h5Gr3J#z(<=zDX_xSQbo%4F)AJJ-exm(C;@CLdrz -NoOI};%;f)I8AFli2f9pIPl0IikUHZd!Qw&B3+JVH_GF6Laz1Y^6BdTA_!bovi)H?rmU5$-=tKljl?* -xMPn1kV)3xJ4p;Bb~jd$#AWt$78_|2HZyJ58^1KxV1_23dmH|3SB^K4ek+)1J2Q -JEHEWk9p`~Y@0f8PWQV(4*3?^crVl7NL!D}XMQAq}(hp`!uE*P+$T1_W^Mr(2 -jyP%Txq@+ToKOyze%Cfl5}^%kC5RWQTuY7ixUDx0c>;NN{w-A`z75SY%BkgH(U%2Hq|ydk)|doScfsc -#yB7+`XWlUSN(yOzo^t=j@M1I!WTEo|OjWRzOX*zOWr-hObs^ys}eu0m02T~_Jmp%Dwy{=b2X-ogs?Z -yzfdw=OXZ^uZ#w=)iauoi=TzXrHG|WU)tzD_%9o|0P}kR~$|-i$xng!ap#+?>W9By9*ivjLPn2H-JA< -;ZIF}pauSx1^h*~clU7jcNz|NBtX;|o9q@P{+-hwY(<^tbJP1^Hxz{5tv;7xl8J+fIyG -J69CcmVxb90)eY@?w1NaegfA%VK*@#;Ch3J&(;t`w{+0#)J?tnuUXsH$X=TAs;{x+b-kkp6bO_9x?hi -)b9E=2ytWVp?`g1QUV3aJ?i^Pu49sGdkdWH?0%?f|EsZmx -H?0v>WQ2`gXY7t|SeMHjbtSkk%_>}Ne$ -Oj-I`6k(FEjw#3SbHH+mpBA+2L2hG;34!Vl72pmSjX&VTxQ{#mR<;;k|hXg3zmE1m@ -gu}apoeQgHijq$@OcHPQ8K?0GnvE~p*BhfO(vaf;2^4$;D~SHjHpMDH0~TfG5#<&@QsoMY4pc#VV5th%4wGtPSA|66de>D8E(CkR+ -l34$8&J*>_>r(ZzYRlOv=7DeNBNG<-z02^8*dnp3&jX|x0+*D2)h8^=Rs{P2{0 -03g-ZBAjtj`iAiQV=d%4U!hzLc4~#prNPhoNaZqYL;}bi5*PhLiY{C3{wX&zM9WPSjU(bC=Fq7dxz2x -nASH%Q(#pmYrbnj_L@mpl9oKf|KU*@Uir9!T^#K&g -xrq6^Y&Y5DxplI7SsUS2Isl3TRtgNs+5~paum)57AmEgQVNX-K00%m=;0O0{+06NuKW=)fB8GnT05Z^W9GP~!vHU9M(H#`p)%`>+)f#N9Q_3GQ)AakURk -tQ9hb9YI3w#6IID$}Tp4|rcb}e_oHtC8Y^c`Z+78Hfna&t2sT)RyIOq9?V8~}kwpuqT|f%6F6&|@I4; -)XLBWbJLn7v|S2bsm$5S!#?I5PyR=SF~!7)0LE!=&Ht1bGfZZ%Kx-s>gO?ginyiV#Hba@u1DY`P{gXP -Xrn0*kGE|)am7(cg&@#ew?t@EsffjgGG6{z#j6#l22+ZtF~_udq^QMIm-R7;Ur^_z1b2mLMNtrzOmS; -J$-yp;bSLaOV|Hi+5&%7YH#t*>4j7b>_O4G|_}S440juP|OajM*`jY(43CNvH6|i&)Bm_N}l$2csBSk -42DjT798V+s~>Rm$QY~!)L!{{Q-lG=wD!fzpfFMw4l7;iXFCzFYVYHTA~a -WoTm6m3k!wcL=~3crC4pfg5Qg+Q+*toMC8`h}+kCPI7Umqwk!LouQGH~7t&**xdDZA%aINK8Y%8B&qSaX;PPH%$yJsmgqWgk7Ic^IYKY-{q)9B^%Y@wui_GO -Nd_C1Iwu^6%ol4n!2U&A>7b~4;LNxmYau&2g-L~9w1Jxlj{`BZXtkES=ycMTZBmaW4Oe6{~srhAO#vN -Oy{8W?oHnz7MMLfAgBQh3k`lo5LXMY4echzw7kbTGd_qZ=;cLi4Fhid?r`IXJTWsG*0Sm(mGJ>@8l|v6 -BbN`+q|@~PZhf9^ -TVRZLw|nS4LK)Qognm%N*|nlY-Ddl_tzA~TDb;VTIfE(aBqMZtcr%uCALFM!H}&C6+aX%xhRyHqOtbq -JeG?be-~cMr$Y^AIV|lcv(7rV6)K>Fz`kt*IxuIBfe!0`nhd5TEwTkBkuC@@&iJIex`v3>cV0|1o)7z -H(aB&J%Vzhn*6M(V{3hi`4V3@E5qdLT@?6OP3I6R2zLgui`6Q3LZ#}&_2)V@ -W`u_|h6enw3bs;l*#Yg3^_m2ep*N-{V@@&kq;(*A&M_a?O>M6)FUK4=58+2*^-8w*bA8{p*VlJ3Ko#CoT*Ls*!gknWJ1Tf*&FAo(m~V(XFrjvjGAUOZ?U*q=cFp -XjW^ylfHSp9)Zkks6*h|3qQ)Tw>d@d*^a -45U`VpT^h}Mm8N;7N;p`{C2Cp-U#km`uTLwGr-Nm*cO~kVo-iWjWlowu!s)cb@h^d7 -HSjc>(0SAk@Z_j3%;=Dj(vFlH(4Ch=8R78{V7XRSzyA7@lY~>A!&v5_y}Hdl77dgs-g(%RHCSxs_M^V -?tRn37k4-bYX3!b|J-wd!JALPqBNJm#spjEdcziEVBX}sk1DAeK6IeVj~aL&O5Db6^IL7sTCWRbjTjH -@UDFflHh$toZyuVbkH|I9wLCM4f2m-fHpK~hl)|)>3b%iuZ^_5ZKuRH$O%{Oz2{+j}>WJ4tYttEcI7% -R1K5=}^Gu8K+?-yArBMM!7oqIEK*yjQbvVF`Q0?Q7^?1oNU$OGvfLgmhsup3tmlP$wLtPX>l3p4zx0* -5x?vz!ENS*F=M#-LS$`ws3IIy$w;C#KL1?s{Y}JKCqRr>f~K9&&2AhR%)N^$+EQbTst-bK8X@FYr{U0dQ%P -9nPtqE6{Vz45O$0)n*m%1N{9=sYa^(!=taiCJCSzJKP -)Y;LhX{H&c;8Pyx(KECNOb(s)e&vv8_AJD>=hDcEsP-&W>nglO9lc7G@#uc&Hey^yG52*ISODl~Yq{k -{Pz&=#7|(Au3m?DpZq$zRV;-o1JG`t?tH-@knG@{cEf`0zpSmpbf0o_X06xCc5u4zHwgH~fafva1;v4 -}Ous9RH!kN(Xtt&)>+S+htr`%fZQU`yG;y4s7B!o~P;*ViXS#j}OV=*(^f0_J%57c#QG!P4V~k`G%w` -865+v-Qz`imbr|a>R5lY9*e?W-P*0C{e90zL&1QlW`&!$qII!vU|w-uZ+vT4Vzr>7^={G047|B(w+!b -{eDjSsI&h*EN+Wd6sB1UdM~Hzwug#@uw8_i4Km7@sP%Yx4XJ7hsh=KrMy+{8gz0e20k9D~y`Ff!DZ5d -Jp%Zd^9yj>z3lcqoga!$F?oQsQ&?fiNUhaT%U$8;K-T6JZ55jLw7%MzV;W`oJh*=|T=RL$mRy0X$Kw# -5BDqNVxB&4oi$D!?4wlt7Rudz=q>Fsr8!x~~pk!iT{1zB+=dCv^1`uAWVTf$=kVVxc}md~sojjJv-Uj1szUF!O-G$y2E^i*ZBK_d=bglq^EhzZ^eFTL1;aQH`VQpmStKwYRArdL#yFYFvCSh& ->0tGQe!$ijOT$`+vz-mdY?-EJBlhrM0*C^p2jWOP6<>F#Q@4bc|hkd|$LUv-V&wl3PUXg!Cm12}T8niZ8~4Jpx`W6+&e(3we~(G7+;I9`BD;;=(mEW -2>;+sKUv1xG;U+t!F)Mz&J7M{o^S$r^(;H8r2p;=B_on{DJ=$~wl#wg-BYo`aY&K_7ZJ(XfZkAjHzEF -*@y$$p-!T9VZYQ3^6|*K62f&F0~c-&-FHH)60J_wf{-v<9YND9;4WR0-#2TdH?^h+;Haqe6B-3Rp+!+ -1kOveUFd%`ZChHIuje^g?SBB(9>o1YPTSIRO!)p)X3e<&J+s~-aQ`Y3x9C4OcW_|9my(f{iidgqQ(EC -4p>QHP`y0*mCnh@8_5UP2^Z8U5pZkbX2|8UybW`VjZ_dZP`(R8(J%IG5Y{mbH)zc9ECl+zsg8VP9BbP -46RS{+L2z=EjR#8=!^+@XJj7pm!n!@H7vtu}|nzW>SFgy<>ekihAlV);i#NC`xz+EUCOW#z{#2kG&QENF!kI= -Fm)W#`H?%moOLn;ANAg!;`Oc`=~Q-O!G;-MJPPAQCKfvzHn*2wZ|W6g*gM&={O%1O_)8#E|qSwn7^{h -WfDnQA&&u#cn1`-7<6EcM{*WUQ}mIB`XknHl_h?E4Qt`^xTabVJlH1O@1ce!S4_0s3Eyv48)3R ->n3M+vIX8LjCA6XQRDdacxhgC7WwEbFjP^0PiaBy#Zv1SOA$SL)<745iv*O8-M+Xx7uaeCby!8f4VPN -rRNMN}@BEC2lgM+x-P#5!SPHQFHd#6uTXpTja?;^|Xa(NJa#ZIawr<%uEg@sZAG|4ifEIsET+L9v=ML -*Aynzy(X@lA%`EwEo{(5t5TLGuaKR?~MIHZG`2A_F6!4Z_e}o1#%j-Oz1%T647wQXG5H&UQnYl@SRSc -%~E<&HXTfXOPuwsNz57>;!X=Sdt1Y)kRIQLIAyq8`ZU{LhnPd%#o!ZU<3=+Yb0x;fCArct_KtPUW(>bcmVu8ifqEF -WYzoAF$arK%NLcMLmv6Yx8Kaf`nOWwAr;-3)QLk1-ZfqlMF%18>lOkFit;EjubDV~l-SIP=yHqLk$p( -@s7)^&3x5K4q-J8Yosih$*EL{j=X1oid>3Aee*m7XH8wqj9N2C|#Qz{Lr^}qqIK;&iO6NVDlO7D0Lo@ -k|c@Wvyco_4StKzD5@MTBK8aio%F=WfrMl#*=@!v$zDdzB3j-{E4u?GU*TjE*?_4*1qChkghIlFj>su -t1PD(j?~UDbO2>C;%|gYSQ5;B-HyjH0?D#s@O@4A)WHT3w)@ye>(d|~yIPVA -#{B=iUj#W&eHES7YV4jr89oorQgH7k?Dc@4!u;|kbHQdE;1@ix@PlFs}^Yl!f{Z=hW4b-u7Vf@>l?8k -N4*>{LZ0w1J>Uz*H+`C*nbmx9lXAc9N0z6T+8TuF$6k|72tJ6TqzJ>8vcxkad_QJmL37Y4UoXl4oC%rgfp4_|&6zWXWT4Llq+ADHFq`$0|5}D|@;#!Nnw!w#*|`Dk|ik!&p_SuF(c?X`5l -(Q>mnth|+mV-jWe^7~QRh?v9ij@yfY6tTIksA_5(f#udn_3Ns(|3^eTPic5>pUvxad(a@@ngbx!=XH- -Bx)o;e>bF>Pl`r|Y&;s$gYg|JLVsMiWRzIMswo4w4~Urc!JX>N37a&BR4FLiWjY;!MiWpr|RE^v8mRAFn|FcAIjUvXq#?BehJ!ke)#_M@ZD`gCzOe;$j|PW -dBQ;4VuHbILOLx^S3&A}H*Sd7K=Wvp=)PW&qPHCt)qHra2QlsM;oE~fJc%Hf?iKjV{WN>@GHvaOme(i -aP*Vpq*b0GG=+wC5y(FUD1;56Xm^uX*kTzrZx07ia!FyJ2~#OfU!hEf^h!S$uNn&BQ6bUN-=OIKw;BUTfm>5)O9`}*~yG(Ew!-|iPizVH#DA`Y$I0uo -~J2CFT+fBwIDMoJ_3Ppc3&g_4%BHwks%g82cf_+ld84FN3?6Ts%_5^_t=Z34`_lc(x&R7lNI*sdUpyBVgp3BrdK1@@rX9PVbyF2`@;X{aC7tTxJ2StIIk*K2*19 -Bsr_BlWt%vaLOFlfHfT8YNni~EF8qVt7J1Q(*A3oJb7$v*@w^vK&MX6``FqZzVVPC;hZvsRwxV#Q -ChVuNHI8dnD`?luoanr-S>97#}_iMCVu#nfB}{faIQ*C29s<|@0|5Zwt+i%CC+@6aWAK2ms3fSzD1m{w8n>000*w001EX003}la4%nJZ -ggdGZeeUMb#!TLb1!psVsLVAV`X!5E^v9RT3c`9HWq&Oui$Dhh~0Ixi+v301?H03256f>CeuE+fj~=i -%tj_%QnDxL_P_5rhonSGN;2JD0~uS=;kjQvQnNLB&x&H(Hmwmw!Q}p+O~bfSy5Wt~%Fbr~Uajx%rMeH -7+J;|tSBo#*MY}l|T?%VwTjX8px)vpI%K4@Y0aUB2h$p*;i^!j}9XVp^$yi{f=&)-tysP6=ZJ`6qXMg*01Uf>3=Bs_)=@IrS07NovbRMZpD{oHhZV{d -#OjK1v~#|e0?!v@X4}_zTX+a -8^L%?!Sv==NQ*on^fC)4GSz#1Mds!}9(q>Y$pRX?% -^rB8}gt-**uN{aImdBauSk;*MBL>V0T%|VwXiCy#>4K0FHX&_ILcF_@u+KP0^D|X%NggHqI*+-^Y@;= -zryFY(g3Y1ePg{X{x+ExjX4>b5{&zmBCGIzeGECH!^6LE&yX?#Brn;-ynZudepmSRy!d#L$i8t7NJCe -EgA%`z-t-~>>;;$;1+i`?^iQ(L`cC&4PM&Vbk905K#A*62u`X2(&IK%z)CWrSzd? -7-V4SFAA-s|}JCYNgbbcrWl@s$xvOkB`kxEA|h}1JpD~@qBjNQsS`It!C&LkW;snu)tJlb%zaGB>qSqDl~T3A$CC<|TM$?F?5ngq@_2I`6GqOQRND{Q`ql%992M)5_MUc%dI6HV}<4n%6L2?xy6rWkMaL(dr@$a5qk1WKXVXl@(3Dr -jy~_GNShr_hK}Wr-)%P7F#ZBQ5EEHa$`q#Yudn+++1Lf9}-0_|7MvCBk)N{Az03)+uxgL4kgjV)_}5E -3z>8EIYd(d`Mtk<`}tpR`@Pumw#M$Joe^O)Oa~WQoAkc -N$%qgF+gvvzuswx7+3CVd#5fLwVb)ib#Lj~v|DGorP)yiq2SAMN1&8Lez6%NsQedEozmJ8?j@(GPulQ -aNUs~Jgb1)KW5N>zlpmS#a5dmA64dc{+m~1K3%-s!Pc3l*DD#2V)Egce|YN-OVB!9Z-dfySfqgc9T;U -8>tv&-4fJE7cy&WwQhzfq!`q{kNvaO10h=V^v{d<0X6lSBNfS)*0b~+pJTPpwtSBoVYuTW$3A|MdoV42 -nA7Vxhj;A3GgQ8OzDX7tAQ=_6fCjJ`KmZ?V#Y6ct}2bcK_I)CR@^p%cNh&;o}Z>ZF0EE^ltGZ@i~)Fg -KX+gzjo|`6Js1Fvnv>@JZv$fJj?i_j7RNyTzlGqxc?`fbG -t=t4OJ!$~6uq?4P8IFMDvoNu2qS=Os?Z!gImC2z9aJN6~i=r!epiCnTV0#4t92`f3nvi3$^IHXA#Cp5 -3%cp+aFx`Pu@EjRBe4alq&Rff{z!ZieKgK0MAB=@biqqaR!9n74-t#^pSrSab&r2sL0TpIQFk(8Ll(CRdSm_QcfIMxe`-qwFlVt$sGv(967Y6^Ofj{Gk1Ic~7AS;1d=6Fj?uzrjr$9Cnp$c==#!Zg;*ZW<8lt2Re#lm*GSXM=ad -yV*-i}@L+1-sK!bIKo>w}8a4c6?j+zff-`I};)^KTW|*_1WuVUZW}MjR(|%)3K$NaV9&!QTTTY->MAE*oS*#?4Onu%uf8j(-y$?WMxUKFgRyN0RkI -y0%B9m#=$fFxr^f2O1r=LpX2877NfM4&{aDqll{>rjnROV08dKY{`5JfWviR6ASDhUmxeoX6g(ASW7G -2rx>I#Jx4^P!zle;u3ZV;D}ThmQkK{EjS6S`~iC<0megh-eFYin<;Uf<94n#`?se8>UW|wwXVQ>Ycur -LSZ8=1hn;`nA-uE>*wN;VZw28$=)9P9Zs!(jzr<>G_a+U2EV^vT3G -m;G@zLXJa(U1NRwj6tAe_FaB07hq$m&XVZo!W6Z_6?%Q;mmPt=G{jR#dcT%haWn$^<-hs40Ptw2Rq1OJanU0f*_>?#^eVIGZPAOVB^< -Su+hz5Tg4O|BlzYVyuJBdF;2nv^K3-!ptKm|f7!`D5`FtUS;kI}<5ugXgB+UOGcC -$_@h1*B>Kev2yHZD9%ZeTG=0z%J^?P#p7xPD!yj1dAME`=P#qL+&Ax7xI#U4Bki8S}$1n}J3;_(?ncd -Y)==%cJ?pV^1aad(H)*jQ|Ih~>HGE@SXr0gP9A?L&i~j#cpDZ7b{QE$k!U`FC1Bpc44z?RMANrF-M(q -=eH|uGpK8S57DtYq=8$_<3`NJBx#$mrF3_pWr*vzCUQCAHCE1EKJ`R^%?Qpiy#X1fALa=0+F7{lB?qX -^AIAppi|^26gv6hD?kc$&6MMB+{(D$vaYDOr%>DyVO9KQH000080LuVbTT;B -q;jIe*09zyg03QGV0B~t=FJEbHbY*gGVQepTbZKmJFLY&Xa9?C;axQRr#aeBX+qMz@u3v%3v<~Rq^*#q`hsS3+%DnAtPXsy(P|T(!pEhlfOTE{v_Lc@#T8HZDdufgu0n~%WPgg-F%*X^>nj&o;|*N{B -*OPi+V2(tJ}R0)hg%tUIfxto4;kmX+1aMyA^SolSyY;dv@^LMekJh1N#DG2YK)h(0+X4bK%YSHEQ%wHv{zy$ZFlf|KL4C{SPH5xo{nrQ_WVX_0~9kNMWbhs0McdI!iuUSQ=?IP0r7E -~*}ZN+q(W4}i3{MFyK6WSV_R2mK;ZA+^(?J$fs&Xe3}f+v5&aFX8eo9iCJqD1o+)6?onR1pZV$%bvBH -!rd%pgue=D55(Dn{sdbIaP#2Q1S7&!t9nt_*i8T03I4ohM+(8{KhO)4-;ycP?9^ -{WzgIX3!RXCeBm19??5z|T-0i|9|>4LRsfpR9aJ<3kghRgX#`bp_}%I`E$&vKv0UeM~L(KoVysO8M&x -(3UIln&>x_8edTa<0JRtg%4oG({vtjZgOwIfF{t#|VqUqP6B}^=QwUoz{jm^y7r+^%?JPkdB)7BxJ=&}c(#-K{T{)Agov*Hcl|U}H~09?`c7oCo=sSAxUWrUqAz``R(}sdcr -NHf$gIl4AgbXk%?b#Ee1S?8OTRgW3oyl*cq!4!Aes*^@4vGjiKFVSB?7q8S$9=Q{TcA(*hf*9^z}AWFy7jdf%mYqB!EeXPf*4iFEBZCFzy|t;kYA3#X5;xN98xJQpt&1e3NNue&`|_gFM80zojebu -sTPH=lxoXu|gv*0^>FNlvj3etlTLLz&p4i1B|VW?TE=%M7ta`*QQ2*lXfZx{}Q{hwK-U;l{f|tK!V0V -0>Jnuy|n#2V9!c+3;k4>wrg?>y}FXQbiGpm=nAW=o6gm`#(bFIr}S+)dK@+dt3=BJ9!$yg0yE`Sf(=P -N4$?FVKzJL?R7|WuD8Svs{!#6-{Ally%=fZ>{}n{1Mu6?)0qw(VEpk~NJEKZP%~}-F;Vr^q)`AhG8!( -YRy$Cb3*krY`CJ**YhYoESk3B^z>(I)5?VH=ErjBrx3XpvCVnvx)1m<(y4oO -~XI?kH&@ZK3GHopTI)2Z4MqLQVx<+Okk)JaqNPe_R{D4#pk3a)OrX!7ao|FG>&tkBm&1uyPJf1I~0FU -GO{?wyRm80)5{**c#}CbE6cy!$9Bn%*Ry`31|a=z#2M -#2?qAeRm!FPoOO17j%7YfjWJ6Tlz8q*ku`o*V#hUh2Ma>BM=p^8N<_Q1|z!k1fQoqZF=UX#iwJ1C3&L -W#?-{JZ?C=KLXu8h2_m;bZ+lFn(`d*SIrxTk#fAzI;s<{W@%{t$l>j&Lim0d-ZAH$Zc{8h_gxRBo6D) -&kp1?u`pt(hMj*K?oWRHhRs<$D&!D~ard`tsAa{{LbK!Fz;JcsW4LpHgQwX<*y)>uIS$!P`mw@UttB= -Gudk2oc|5ym1ws5NkPl2>DCEGgBO%uwvY1S)M+U$ir*^zgR8?+bv0rMz30^IF!c@6j9YR`X>a1%7q|f -wygTi{Ot*YcZ0>SFEer3!*Kc}C)E0C@`bEUDY!sA>cF1;}rfGt8;!P4!`D%$1yZAZ -OnhdJU1&a>vX(o&s4EtQ~AenBtQ5aADKK+d!(|zOh^7jyE4^lOeMM-RJ4uE4t0ktnj!1uM9r{8L+V(s -wL4L4K233qroGeiyxysF(rv^%gggS#cVM6#AE-TbiiCb{iNoRy0Krdj)SU@B+3e -zaTpzz!M*8@b|aty(utQf75u|I{weT@j>42u}1Aba%CK4JxcAE0??Q?O++$*LX&z@$mWB(eRb!DO7@4 -=pRb+%)+EsUW4%OwdWcYk1|mX%ZiNatvih%Ig{`VCUheTbxHov3#RmhMxyuJa9HHdLciC{4~cuGx|gK -C=GR76<}c+#R|i#Ak~PO5B)*vCoMwr8`cB}m#5hWfy`dK-XO`8fB-%mO$Y+B_bW`h>gPxS$$OWtL+x0 -e2JV!x2F?um{Zj*zz@T!3+NV1$Mk>bVG{b-Dh6WLqmjdntfvcpOaDz}bJrmdih3|>oaompV)sI{*i*K -RsC>M=h7SXTeJ(C#US?oE$5tvisCpdwAx@T6RjojhjDUzNpPhnxe9HGJx?F^P6+v3h1@w+S?DglPZ;1 -A$xiIX1!w(!uu7~)d-8&}y>u|xzG!t`cx7}OH}-^MoGn8V*a^F4?50?Gxq9r$rL(!?s&DRVM?A|%MGT -e$Etz&=={;Lsnb?Kzah;5$FuZgjKT<1+wnbh{4wIh=!C+aCi^uf(@hX;`h_?lK$)GzKm{3_6gda4KY3 -0zgwPbXjb@X#oIYB~_{6XH|q(E)RTFgR)qlKOe0gk!(^y8I=V)2*(TV{N(QW=JLt2FFwm2KfQeQ8vF6 -;$8`Ph*B^ZNo8SKK_kZ}~pU%ClNV)(FFBhf;KmOW(?!`tI!RwA!+k@xtfBp5m&)?Cg#viu!f7;t8CE< -albG^(vKN1&oi?fTo0=SG{`wK%qG$%bq&AdZ0@nji1i-Z8;WyL(q97H#jBJ6CvxQ3sdwB$7DGCn2?r1QyFTOqo- -3H}?b}y&jgCH}B@*A#pV(~!&>Zt{uv^+sQNC`#0kOgK3?C4&8r@20C!T)r0hI(hgU(M%K;Ck97rkL>ZkHT^GV_FvZIKWkKAw06EU3@*L4uB0Dvb=7JaIU+KuM-w -L%hcaPu-q;>Ortru22Kmn@7Pi9;T$)9EU&c?2FR>|#yD60UI9$@&_Y9=Hzso}H0>c%@DUXP|l>QS7dW -62Do6SF-y}p|E3^lNyaE&-OZy2kGpDnESZ~n&|AFxlUx0C>C@D9ne3a?W9Os0JwpWf%0pFi{XFAP2Y4 -Nyx11QY-O00;of09jjAqaGw~4gdhIG5`P|0001RX>c!JX>N37a&BR4FLiWjY;!MlZg62^YiVw0E^vA6 -TJLh)wh{k+Pk~b7hOEz&Wv6kfR>`=IoY6F~J+?fPv7(`HJlq{(9?1}-q_g8rzjpc@eX%@AcNYK&ki3( -ew9|elXKWowV6oW0zg;j0g4eaoWG#hamB@H4=8RYAN?r;T2SG5IEUIG7l4Q};O(l|q$#q#&HRE|+)V! -8Ot|k-nEGw2vnJ?{aq3kcUYG9T24(7GLH|?OBR|}a5d;3n7bW7`}MV5(_R*U&8wM?AUqT(|<$jq=W-^ -lVLUle8sr7Yt_<}e}WSuCzoQ4+b9R1R!q8SUt}s)}lC8ty{HxJbPzMLRc~Fu4Bw=0S4&-RbeOS0~S(y -_(Xq)5k9d&%S#zc=YDvh4(DlzW7?IcKae%sFS&@CKHsssMwKJLcA39H}ot_lANzWb~4?o$)v6}hs3H`diL%CV|y$we&iE+K1%rRQBNupEqHc9R7lR6tY$x+i&?V_1Ejt#<_#) -S!6IO8mt1D3>L7BM*<;W1Y{M2s#a?XcRgtrY@x!|=pV_-Lza68VT5icFQQyJnWG)sgZK_J-b<(W4dK* -UU-WLpXs>y+eeJ#NBSHccsG>l!XWV(WJmqiAi$X5Ikbjs?2DHtwchCt4EC9@5=DH5ZJN8om`Ug~~CkK -23vPd};zICD-D43nyaAdqB&&tV<3sIcVq>JWUjB91eu>NB`G*L>qkSPK8d89bI{80-g8`}_Ov1HF=5F -n9qQ(NlOnH{OC5jq~0S!G+A{M2`K*eSkD02CAl6C~Gs5=tYMuS+hHuO8Y`$Q+A*^5*hO3EWC3fjB{ox -F&KP?qo<(Ab0f(WR|7e!!!}fr5J{x7qR3h+cuCY0tE)z!&>$3%RnTlyzo1(Mj?4^oEj4s)A`;42#8W0 -NMFp;a@4+Hri$xC8ni;SH0ko(#_>AW^$Vb+A(qg?X@}#W9LSBPvpTRtOfzCxM8F8sj1bXlkM4@V!GHy -drI~=x!TGO*SDvmIpau9L*f;8px1zSSK{K+pV%uvwWtCtf{reS+OuwaL0N<8r3f4k%PG%J30c84a0N%+c6lm!wvP}Sp6c>v9cAZ9 -i@e*ABSh&x>JYlxff&c=0&Q!r9lleIn+Vt7NWCB2(VEiA!#JidoKYX9a^N+Ggc%Bs>d -OKmox}TcwJZQi!a#0XC?uI`O-j)-j!H38LzZ(@Bz;^tvp~~fL#<}E|86x-j~pqu8xtHu%(W$F@BYtyn -1>3H9d{!F}?q-R=VA7ljDAojS&CC7-@VnD^h}Z=5_&quNH8JOTlK*)3>PoHPR#}r0s?1m0Il=c8!Y&G -c@@rL`0d>O6AbLE5P0`9w0F&iW$8s+N1^r6yiNW0s$gt;wlL9Qp_XB -0vf*5D=6#T7jJ?9(H%m3Ed*%0~kL05fA;5Fr#eBJhyNXPo5-H10}k; -Ci3Q*ctlkx#CjF-+Y^BRV@cA5you;D -sYhA#w48=M^D_u4DTf{;TwNTpVB{l$mq5#Q3X`I(Xi#GlnlfRQ855{;N?_Lxbb)Wzh1XmCu7?=$P^hG -&71uxy<0DM$#kh@~`Lx|x*Pa7@|6%B}E3k!=fPP0Nm_8xsLcV_bu8)$j_cr0YCV!$tj!E}L@E}Dn%MwgivNKOi!sS4DTV|osxayzR1h!U1ys538dKw-& -cF=%45~wT-WBfKDfo@1QA)FS4{k-uYXyx`y1pGxuq$4T?#fPggRJQ-UGXXhE%{k3ctQ3re56p9Fz;u8ksm`a+w# -#s=OGx9Uq-(ry-0)o6z?Ng@6g{V;ZjtqjOW2;jvdK6PrF`*4m^*c3X?N%Y^O76uSADov_F6?hJd5N4pzaisntto5W~d3D?n8S4yxUncOIwoI4r=$modaicmOP$ -u7;YwgJH;atYw~*9Ve3z)yuvQWl=%&w-s>SNMJ1Ps36q8uyt_T5-1?&19FTV0-LURd0Hf0_3Nah_2IT -Wx7b;(M@Mt7LxEE*y08xAiRtkd?M385P56{tt-|tjP_;AYpFl7giqM_gQ#~s{MCmR<4;iW`}i&A7XD^ukqLm+EkTe+jXL%XOaPb-^r0DJ7IwkR!ZLsY?!oazuHWXKX4+(P7 -8!E!-p0f7si<)PNbM{jEz?1&~_KVheQhp2Gu0-Ch0pR8`y -vm%2WK_F2oH#8W4X{%rdhY3?y6-kTWv7qRJ~2k(H?9N&$W8%E`B6G(5tJASd{=1{qMuqhHqnbXOzZv=c~Sbw=6fEY%`swk^Jh{4 -ZKj}O76(S@6wzAIuW&!NnNEv_L7V8bO9h$i8|kB!7~?}=IA=fD4xJ&X??#6SP-ABLV^iH(?wwxI(a -__{zd?a#<^MYHQ0?@|Jlo9c5g?=7+-uroD -DUR#!RO3g+R9;fC;BN#>3FAyOYjj_Q)mog839NY{AJm>NTyV5hx?Ui5#xy4Qb9uQay{o_+jA{^yecv@ -EI`HEllkY(JB;S9gABJ`B9+1o0zI4!0Z(?y=V;U=3YE)DnVM&DSLzF#%p66p(6!fM*7<##|dsFFb+aH -aL4n{MoaLrDLicU-xYJuE$EcP(Py-}tsXADv5^V_H$Qv)I7xlvlPUMCz6H_;UDS&6i -MGVx_Imj=CWx%BGd9F)ELr`kCVg8)3tFag}EUpW6!Hj+$CI-_urmM@4_^VW}YPju6w@7PU-coEuLJU0 -?B^x(XgUVcdqi7Q-`5pC>%+es9c!JX ->N37a&BR4FLiWjY;!Mla%^)haCwbXO>f&U488kTaOx#SQ>W{8=nSO5kZwSMq64~i#jq8D>d3P;MRX;( -K``vUkCf$nbUn<6grdYp^6^P}*`f=8KGVJBgVXlMuwkTD1Xl2vP?O1QngEC6mA)$gpdG4P;ZDHj -UlgGbW^GVB5x`J^J&V~0U4`s36=@Wo!L=$}&S1zUwh#~n1I!9k#L%|Ce_L#P3Y*d?W)#ZT0BBx045i5 -Nd4D1uqJR3}6G3CyX`Im@RLO=K}kF)7*q#nOhGo~l^*-cjkj>W5bNz$C0R!N;w5ECTdL^D)%oD|S#x3 -uw(6Vu5`M0`Q7u)FvdHLO%r6A^o -RHB0@pPPK|;#vb2Dr?_qzG}@O^e5X&qVe+JWUTyO@sl9?P(sb&BbLh6C;jb1J?NRJM@=lY!@Q{*%?^? -cXq;a$m8ttJqGSnpGau|!)z$KoO>Yr(pmM4QwN2&V5cs7L{z$XiL+AeVAqr&6-Adb^h?sO}OcaKX&wa -{3M{-r=nX659TH0UQaq9_vD4Od$^B?ExVto&#u|T4JL^}1zvk7jU;79=r6 -;u`C3&3w!8;deZMum7IeYzMl>@CN`J2k91dPjzYd_^l7b+LDZEMq -G?rKV8}<60E?9bi3+XGZ=SGD<)G^Gt(@)$GJ6;GrDc{EMktpd266HkX95q8j+BgJOEY=Bp%=fEGJDaz -2`*!48c_z8^^_BP8?#so_>@QGD0|XQR000O8%K%whchf_09R&aYe+>WtApigXaA|NaUukZ1WpZv|Y%g -_mX>4;Zc4=~Sbzy8}Zgwtkd5u?XZ`(Ey{_bCKQ$Iu=Y_%IUv?%5dLFN|4x-Mu^Y{NFFv~;q$(xgCAc8 -zuaeRrhwVpmBUAc!R&@9lZ+Ii|8^3wd5OzOk6+Ah)%#9)#A$3ong!lgaU&GV8U}>-N#O_LIf-las#xe -4wLz1}!Yw!`T}BtMD$(b1k+MayrRhzP@_(_NQfj^T*BY@|XPPRa}GYQMe$y&(f#GVUeIs)r8OAxx4aR~szeJCGIEl&yrrytY`$57eeIygNQ_g)cVQB%Hh5NCz7-Q0U-#c@5t^`X0 -CdC585|lh!vv0Y#M7=q+mRvNm9>4+#=IbpO=?sC~kFo6_4EB!g6v*}(FS@)WK`ambIMzob@+LbbJ0IS -&0;bz`+aL1#Udu$O)V^`RCQogvwsmiizVJnX-Qf+`z!VI$n2?m=2fTk9E;_tkhyRNxBtcwwAd|LKEi9Gu`A>aiqgTqJ@EfM -j2<%b5Xc$RBRm-qqkVuzxZn~@1tCg=^v%(fBh0ux($kSfl|WKJjzsxT6a_j5)C8z#4o^WG@Bmny?2|qsU?IBhb>2Rt1#x(0qH#nsso^c+o#zIzySx92lNag2J1ABw5DI1C_)LxME^v? -E{guz83zP)`uf098O_*TpM!IQ_|Jf3#O!N%06)3a)3#D>C@1mq%Vd^LZPOu3C!pY(tX70*qXPBTZt?p -q)!6R|Ur0P|Kz#gx;yj{xvt{7_)+;n$m`wZ`_2k4_numMYrhYh_k61fSO}cy|Xs-a$qDjM@n&+6h5e9 -f^NrMhimhy+me!@UY*Y9zQQ=tO8g~T^mOt;0>(eJny8FD?+K_y>rK`y&EREZJZC(6@+tOo>9u|c@eQc -?%&f%*R1e^f3&_VOUe69C`D|d^+eh1g^f~BL8LlAv}~utQk`jZH>l+MjsXsKqQ3< -t;+9d70um=QCV|x?>xP)-nk{lBWF>hJj(s1vI71*#YDyIp7~s0-wh(?`q;pGl9xa9KqMI`k(8Qmae;Le$|zzeb(#LIe<>G>eB`bP>d{} -~c)dIwJrOW5WN?B}m1S+GsnK_8Dmk1mcCFz7*)<27MzO;e{yxSf!iEEbRidM2Zq60#mVKaWfv*Ml!!^ -qga&%{-ny9hc01@Cm6BZ3;)z8_>>(%p&r4|si-(^0j+wU4V&(X68p_Cnu+#lBBG&gXIBU}x-DMp>!??5V?Y$k1X44^-t?$_D^uA2 -m+d|*q9{drI+np}LYZQ9c^}oON#O)bYUneZ5m}LD$?ffQ7>Y@>Osj%icP!A!9bQJp>GLi}Z+f2$@c -&Rt0|XQR000O8%K%whX2+rgJO%&&bP)gm8~^|SaA|NaUukZ1WpZv|Y%g_mX>4;ZcW7m0Y%XwlomXv-( ->M_R-e2JhI*~|O!}75wR)V|b4jl^!IFN3&s@%kF4T&Au&X!X3zu$}UGdN%ub`R@J8H?J4Jzqxw*?8D3FuP(_`lAh<}aWz9KI1g~Kv+|1TinIq7r6({p01&dAvXv29cFYxtB05O3x)0zXiI%db|m!V)2&R5il{h -|A7yIBB%H6(uiy)pEiiBaNjLmeibELzc{l!r3yG$&%^P(3$)Ej@woP7-$E>A!_CRrlm0U4Be4ZXkJ*Q -cLc6_hnQDMKK*rh`3gK(Rx-00gW%hw*xdDk*vQLO!xP~7vn3z%y8|NYl8r -VTixe0-1a9u6OPJHm?3-!T{B_$(V=ve>@m~>*}!ooA&6TbUSEMdKO~t4xY+RG%c4+q165m!O4wc61r* -YC=d>$CpH1E;-*M>_a7MMd;CB)Ve)j~rtImazmI(>fJXP^P5(L4JX9gM%j~3 -0lO8KX>T_p7v%h2wz40U;zG4m>B@o0f&xIcSS54{w05P)A{p)H$PvVmzz$e<1 -9>8+PQKzPh`I{%fctqp%w(W!qX5KZ)k)kNxU092QGQ2au@XA@dkW;-M-d(#Irl%QmL3F@ekdqv-aQob -%j(yx_d_D<(->+)|diZsp8|w)I| -^+>v1)q~rrs(>D(c@B`l^}+^x={aZk3qvsScP12ol+h_z -IjBY$c-mIfbbn)#g>*q06n#}p1rO}7_7#pZ#fqiquQ-keZMhfVTN0c-Sxw0UqpUhdHD*+=i?e^IuJFl -Gn3yG|nG$xRUT=!P)lwUYN!Bk{Sw@t{(%^^O_CQaJVkQ$vd;Jw@#R3F4U6tVtb~-Q}=j&?L(uf`4RH> -xumXoDN%K6fiKlb28Sg@2rvYX|1Pw-~b~km*3@FI~m;JfU(kTI54>?6f7KOZlw+0Ku;!Y@R9^#V>V8E -HD5um5v$-9jKm7I==cgSi`j>;DnloI#_fvZ=Hbk+4T`oxUNfjVo*FrJ(WIQN$rJh*cOQD5{pQ9F#*L( -s-M-=cV)R!Lz^v9~Nx?7VR -b?zOp)iJQ;e~A#ABy>;2EQug#PxbC>4#Yl8e=V{1n(IP?^*!58HC%@)4@$!&6ffC^23*dH@6aWAK2ms3fSz9ly-o-Bg002}10012T003}la4%nJZggdGZeeUMc4KodUtei% -X>?y-E^v8uQO!;QK@7g{Q#9cqB)HMUlUD2*F3 -M)Iuf&UV#gkGiAWk!m1>vo>PX*;4lP|3|ut(mW*2sKsN+?m!U(q#Wg|8=|&-u`6+{s9g9Ls)R7n$6D& -I4)s0U|k}%Y!UDW0?d^01ZY-BIwl;r>s50KS|aPGGfj#m6TH<#t@DZ3t-&-0OCReLycTY?Q^TA{VimI -{P!1M{aD_|(^*yF4phFXQFAW!Kdy=b~!8#7{)Ij<3C52P26q;xSyO*f@LJv8^0n0Eu__2-AnV`>G%Ll -7AvQPAVLJ_A@*agS}UZtn|V<<&xTo2+R0CFvsBk0zNlI!toz*ON|9KLc3xIij?#hHDFwgsxMGW0|XQR -000O8%K%whcdOf9%me@cmJR>_8vponJ~3vkF;jCxW3yohg -z$Y8`Ie+*%wYzRhF#EuJ3EK?029CY83_9(uPxN;uF^f@=8@@wUXV&RY2hLrm@BERj$@<^@*3t@(;?sQ -%#lqs_f#cVId6RFwNiPFe>mUjOM2%~*?cj@#+`f0R6{Qs`0d&v`Qpq6iU)Wd1^lEA^N}NH| -1vBB{s}7wY6S|Ydy-@R7(Oy-vDlPm2lNP+>2E;y9{GhGd!rD4qU&_o}XiRY-$}jgWh~%Y`ki3;a_W1M -KAm7rG+d^eNKb{i+iyoNaRPz&M`e;`0Cn3h_h-`i)QW(2~#XjlHYYt4V$S&OIa1E=|KIpCG+@=Ms_<4 -Z9aG3)^xv|dbngDZ>33<`8@~54QPo4mg4$27om=GYC&jo|}pejC&=R{ln1NI|hG0?Pqu!?$hPnTE#35 -OsZY;OymegOgz)mZY}JU7HD_#GjO9|(Uc82O(*7#HOV0f>g@Lp&lM@+3cP$*W{yB$w=noC3Hv_l@8-- -1pD9oNV51$=#LB<2{C7QOd*3-#4e57rn)QN50ZBbMyek@ue|?FhvLfJ|G6fE5)&}XO7<>c;Ota=O^Si -7dcvZnk-<19Ex>Q^Y8KXoBJ>twgO}720dQDH8C;84l6hW&`hDu)YuG+Y~L}UH;CMds!Bd?KkJ -4M$SxP4_aDOrsVyE!u?qjN;DkxYnuO6blV{|k%%J16ea;Y-V;3Zi;Tc{TspfEG5JYdmw+U|InH}gqV8 -jwzEU2R?M!Je3hewh!De->73!WTAm4#D-*J?C$xnv_KXzD=HkJf$Dg0xw=JEp9Htvcrc@_{AKNQ&vrpT! -lb0{y>4s$)5q4aN4uLfiLqiuZh?r?=7)+vgG5E>B054%ZA*J^<=ObvfW6NzwZgDy9)eE6Kfva|t8Aj+ -4l`3_^>HsVJCgWT&IZ(Y`y*ppA`yZ&-&g(+e=1RkG0>vU+dNAD9QdpD##isZ;wXORBG_|N#Q}&~&2eh -;5`xln9ufl|2FRPj6a3%JlnTPhcS`YnsJuImCQ>j%S5QYt!ySx3OE|^7BsW24LK*H2`#VY-{!oOEsj1 -GS^vHKlYVsgr;FtCRS!wV-@@IN=U2tjoc+RWog*oEb(rng);zOKpY`3}Fe -**_NFHA#NDl*FN8`c)mXb{Gih-7ppIcyF}$pM@&4ByBzdo%<*G;kH+&D8guOjS4d4u>N|7|x@1XuyjP -zLIHvgh3uI9{>Im1^)q1O9KQH000080LuVbTbuu)p28LY0LV%J02lxO0B~t=FJEbHbY*gGVQepUV{=}ddCgq^bK5qu|J{EDj%miEBT;dB@7+7Sr)youX*|~?9@}l-jO{QG2}x`yQYA<^n!fz+Z+8KZ;5X -UnMXiy&a}dXwb^<7t`|yoj?@9vtXrNw!?Z>C*hnq`%?C%KXIdNj -x_{b78)fc>>R&Z7$o-a`Ag9WO1;-NkqKhVZx;pl2t@D2@i#OU+gw8x_Nby?he%RO}z1gI4wk;^28UPL -b0KFo0dQD=rmiec^Zw_c`6H@B($)zJkRnz5@iu5(qG3iG$dZfQ6$o9@u?WGD3&m(U4ZKyu~i(209G?h -N0#$Vo`phI6I%0+A_&!#`uc4{>2F0YVUA&z7I~JKp6&sI&q-OV&eKIUqFV{!7I7NU5`!WOVE(^}un6)jD*~Kw02`ChPt_V`JUZ|W4uBEBoVZ}K#;*J3IW{ -QMp(gsx!NCDan#iGv7CC<@il6A28w4p|gP_7_2S1))2fv(MUY);t8(g0Kc=o^0BNslfb`vK;F)Mf8FJ -UyzPv$W8gk9uf5r6Xad-pba=F#&v{I@JOk8l0@#YmPkXO -|HrX3_D_R8YX3gvdq&R+vH*lUm~{2WHHV~BDfSxisf3Qg=F^vD5i%63wbUWpJ!#^BdcGZT)pmFwDbOk -j~6H7SI@_Pzx(5_&p&_TKx-KM^n$rpA_OgGVYb0>vmy(##A9*=vbBOK_)PdqpB)0jz89bPdXo^{=_1P -}bDsO~{O|w>?P0d#yr*v#LrN*~kwy8C8o0Oj&%XD*(~Yl0zHPKxm2)J0sCek!yq`?(o_SN<{Cb5P!Y* -S83P`rcOdf}uXo;m6`|M3A3kHf08gvh&f!mr6Pd2dbaxL;WWZ{bEJOmwNGGkmaFo3Mc*c|zlE=tMgiO -~Joub>{;y`+JqCYHn>1v@;QPBgm?S(fAPTljs8zkRdtoA=YZBaeQ40~}L21iS=U{^8x#8L84iBo-k4* -+&^5;b6t(%9SfVg@xoE8~=mNvn-kFu3)@Vdf?V_;jd8$v5$xJ$h5K_^6J?O(wCF;Hf6Z$KO7u)?lkd5$J(%*$x -_`m*<844H(4ZRJ0Y7y)I4umbHu)lCvDElasy7v^KA|jb8EhO2h~2rGKkmfvOoc)O$Id -zKZ1#OrAQVd^pLdOfI0vs^{GC7f)T^MnT93sT7w^YeXn=|D6Z -9!#)Gxsg5x5HqX2LnPh6bzs7K<&i%u -v!%xdqIVp;rTIt*N33eRxez8jrzX&;)bzz9D*M917+)&&Uz{YNrqXOAX7zYI*%GHh#6W9M!h-O{;hiiK_+z!JilVQ+A}&F_NEE0)5w-UG-r=4g|Hg8+o@YrM0=!;;{f-e=qw^Q*O|jFPj -l(LJjOxuR>J}-QRSu93FrqdshB9z_;WLOY8IPhESu&zEj(=yQ#S(v6q_mu(0^@wtbPxoT65au%J%%lk -h2Rm(7VP-Ne<{U=nYFBhCIWs30trdOfK8N=F9kg>B9Z4cZ_ojO%-%RQOAxb+f5-pR`tPdNz?Mykp&D% -VuriW+>+!(Z$3V#-E=bPYUAEYwgxLU&T1DlOIMnQo@=DQC3mXj55Tu# -~`MynFExJg9XYRaZ}2W_Q8IetK$8Wxw7sq*n;dux?a*0NTt5!vp$z50c)A5C&f?ZHGV4Fb)-2kC|Ase -Y}*z0Z*WlgRh!nNo~^ZF9;jXr)B81^W-`i+XV%gUP~4`ZZ5Z5KLP9RO?ifA?S^ZEZi?6SlT^{n);M!Ox8#Wa7G-qM~*b^3c>tAXuRfNL*lH -I$qol3TNJA#8{-{vHK;QiT5rk%WIGR5SRwmvtuA7YyOFT&0e$v8B9vDUwBdqO2l=^KLOu?X(2)KngO3uN)gc#(U8Y$pu -E_@&>~MEa3luNan7<3oUQb`i{hxS^}O7LQ>Nn622>ID(#XH!9s2|S`Ah!UvFxbkj -Jz>ixudiYv8ryREJdJ9@bDFqB -O}?_1BZeyJ4U*S$@;cUDbtZM+2cm!{sB^(W6b3&Hnfzve*N2~Arpk$JmHkF$!EW2OT&cN -WjU_(&jVb&=46|oN=5c*P*>B9$_o_a#;W8*PS_QQt(+^1O6_3v#9@`RnJSt_d9v(c9Ki3tiw14NKzX= -Um9g4`dgi7UIi0G{Y%;(U(AkfXw>jq4v;MXC1K=)GV-9d_4ITW9c>Bmio7=m#P}!ax{YiAp6G4-eeoM -1O`!$Sm243gy(UwJrCQG(chG*BRcS#x?AFOb%a3B_+oTXTdA1RGu{)&7QX(+lNpP5aDN3A3Wtmg -#!HZS^f&}zdZqfXcHY!yBP&if)&NRoO9F^m?wjpg8N(*FaU^SdpXyWh(%(I=?vUx*QG}ux~aPRFyHAL -5XIx-Fk9oEdHWi71vQz8ZD;&nWojsq%E2Np|TD%^!$6EncsZqb?hYyN7EM$nXdi(R -t8Ol2UJG}3usSHLXO46Kxc -0w8T=s@XMjUY%W^z78mCx1)o{DNWm^clg|E+ItecZJUmIHMLSfkHa(&nd;O^Cn1b-&1fu@;aJ97VS>T -05NE3Sa-1=W`EZb_C79Z`oGe)86tGSP$^@ZNnWh43{v3w1P5_{n{Q@W3yYn8K&1eR7qLt)^_O%urDj@ -s`KTG+=IkIav#WZ2{wkR90>y^KZSjA~*cX^;c{9`}&L-gX};1{%Q-ikKj4GKSq>Vsg%Dh9xHVZl}mS_ -75QNV#JRB){7xI>|aFLK0FVCz6ApFA_fFaKpBzx3;+0Ut)5n8b1cI(l{TX+loQ=~Or2Yut~NqMl9d}G -joUTfu|=7NxXY$~iiuMSIDpTEK3R}0EQM>PKlUc|@zb!#owxYWW#_Kvp_gNw^_k6+Y(7ngTOQGF^}j< -&?0wPA_u!|AKK4mSXZ=kFq*rF_OuDzl-(EtcAYGc-4g(37i=o_y-OiPjWA#^)mW1VZC -5EHXeOvgL8X-5JWL-r>E>a+LLt?B!gUnrEdRsGgdfK&b3jw54ML*9x4kH-zq!}@w4)bh5u*xSwGwKHZ -9rR6KqM>g?OGYX$QkP(R=;+{OCkiDou+wx(tKG85dHAu8krs|geOdGT>~y+W8+tQf -5%+JPGoBB;u?8Nrr!?z<{`QKxE$%u`zG$C&%qPL&PR-iw%EuQY;08R=3519#bfVS02~uoDw1Sc;xjxX -m9i^6ej7LJ|>nti0QAc>RgmM${LIm||t>zUDnN{DcBO{p10vT`Oyy{nt_o_#2+}9T+Ogm>n@zcbCz{l -sKAuzb#vA4WjmM-EYA|sV;^OYLl2~fr0N5ogPkWi)3Ag`OdhM(LA5Uq|r-)scV42+ybddQGk>c);NGD -FegFq8VE|7*yI>hsn)iyXy*4E89@1^1AN1$Z}z5Ge!l$4q?{I{F}mO((l9IWXsoc)+Lu@ -zlzlf0>Z1X5Vg;Zu;UKCl@V2tSq?xk(Wo`!_%@Y2vf~7(gkTzJ&Kc*13?>I?S$`-$Hyd$WRFum6z!F^>G0!(b -1`fQ}sg&BcUsza`U{#d%E6zo<`zRPrX}JoqNQp7QF>!m}!JKHA6J1P#{bIqFLo@VG7nR`3)|i(`udz{ -;?s!vqsOl^zYpq-&t&}l6T}+wA$iDns{T5-M(2AEfDd*Est%@Q ->H$z^Jj9KYmJPVB)a6^2h$T;`Fd^B|5s*7NBBWb@<(@VxCC5HHVT%&0y%@nld<9yq<}Ft$6J=j9nOIj -$8gyPD9jZ<;>rMyEA&EpjWfM#dP73NPCbc!{(D~M<8!**X7ZNZku^dm?p8GEKs$NlJ4EFT#Auh+U^ws -khY>jtJ)Ljwey0T$EU!F5P*c$kUZ7O>0wrVmK=dXk@ZxG$eO{UI+fEpC^KU+5NoVSt>058g7(rTn#Pn -Yxv)CndpCu$LLPxV|MW6ho0`)9CY8k={MO~4VV9fGn%J1i!~k+05T(HVrST&^^)TK<@{1n&u{g6OX*C`@D2&A2akwVx4>GH(svpC3-DDJn0&^0_*?>`W6Lti -a1)qQz}2-V%{J`OK$@m)XE_lS43t{vNs-X#!X?ST;q=g4Jvy8m_RhH)0C=}SSCu=8=I!zAi^=V`j*b1 -^nX`H$U}Ky#^A#=gTwfnliyFo#)?tJ$K6&&Bms1R{8fRPV^DMD_phO>bSMp$Ke=psRAKUBO@pA*-WIL -u~*!2-v%I(LC9e$T9{@!8FdQJxt{a*KLK1sF=@DzxM0sxqO&fS9k#-FxsuRh|6MmK7B+yRb8f@R-s>J -lHQKvn5$-xtyw*<0POFG5vp)7d=r@&N#5Z%}s^{>A7?x%qm!&11%cY=uyuszQE+WAtdO(1Bb)_z8iVc -9V>5=fWgWT(-j0oGCknulzPCbM&4_S~PIy%q5s^=O%oo69kQx)l2kNu<9?{QAE{oF@}C^`24Tf*=qaD -6P=YeJ}H)bE?vElmRqWqXIlT3#i?73Vj9@uqpwg~eSkjIIvX@}PvC~{J6=cr9(z;UnoITUAxSm&$M8l -CFAIHWu>n7=3afoW-P}BKS+0sQ`JjA{KF%m>1>?h_~-~9!c%)X*=icO3`Gz-@C}THg(<8M&&#FN) -RllDP5Bp#y&+)SCFmW7NXv=(hUHuH7YD}rPYxi*C&z#N&;NM-+y*?j&JirT0BNFY?yC_1e39=)5L$4$ -v#xITtI(*6f{=hO(nyDEbfLGUOK}=wQ96(yl`9HQah#he8{_2*QjA0&_MU)3G|LRBxF>@+>yUZY`l}N -d9KOJ|C~8F3^UYzgj<jZEmnhN{X)4s!>y2=b7WdRE`_1RcNW2-AwDgcWTo54kWp}xA;N8r4dXR$ZslP1tUBAGO2%d^tLhXp|i)tc2k3WK&zW#?@6xne7RNkx(qQ~vl>7-Qd9@^EyuqCA -psOPkI%s){ek3r%S-CWZNLrE-xAQP1>NmKrZo_J;L_Rlad_<^jBBC>3}D?z)@u2p|_%wfgI+U`8)u!sL5WSg`d`?CJHWl%E^Bs9IF4YPw#10b$+C4$<_U76xm3 -C*<8m8_dcaG26$AFQ*O-s>78yt-Z21UsjjLcsVXMzU0NT5d#~@lwU1$aQPkJysh=4g#5)#*D|73X6yB -!Vn|#zKba@m7+fb4LRM`oa@M8l7cB1+7l!=&5a}xN#1HYTRq&cFVlHp>5omAA5OnJAa!2cZpX8mH}BrKFk@(Hq|AyydzWMSRc862u#LZAWjie7Qnq_F|n% -E`3kC@HD8iCdRkja&J+=gb&)i$)6Y`?{#11&zDjWBZ_>Pb=O)vW&KBW4f6weX^%@eY#_zS=D)Ux!ZLq#wWbi*`Y=ysH;txoaBifE%Ua`P2wxHfmVxL-|56ks_Ba#tlcH%Pnw4gdU++%G1P6_}! -?0`^BIH8d3>5wvllmjofQico6ROP~E!hy;R&*+|XJVMP#0-XUw*bBtU8-BF!dp4~U -$LN*;}L5xm80Awo#KB~SDdO^n3X3h_V`6ud#QClJH9_!!>gm0pe&Cp8$Va9SdOEIxU0<(Q^ec5yXXo -v3|O|R2?W-F*@et?mke0jmn>0_(JO30yTdpxo~J*`e+UsYzO#%T%9by)*Iak2SH1H;yBhp~f#CoVLqA*>oZ?JZ34)9%pjV)%wT-HRQ8KGq4+$R}fdXP -$O^Nwu_RPGfwkOq`N(YS-3q%U5(gMAqX@wt4BN85PnJAKFy>lwP(o+;b1(&Se5O-XP#8DMxk%3Pu`=u -#bgM_akB<_ierK2*v@Te?EPG+(NoAa9pyC=z*Aqoa=HR9!($c|9L|P5pZyKMT)Y()CSq66yOeSpQM79 --pXZZg`p*E)=xVfSn1`qCvh!7H*n7&KUTA73R+N6PzqB_VBwz*K+Tx#TK%w=zJ{a`?y@@%c*VUGQZrrqOpq_0-XNJ`1zxiz88y6TB^12HR9e^Y8hRy9)Q -bGuc@naO4<%8~lLr!oRcf8%hVqhP!r0K6g6oMLcid0%yKDO{VH+u%C}UqYdv?GSXx-yg&*p2m-fB`x@ -v*%^<0%}*dob1o;|i4ULXLA6&H%QFT!zqK;8i5qc4LqJmj~1&$Z&Ng?yihx*)I$`umoc_aIf-HQC632 -cTQv=1A~ek;SzIb(?{+OA5+HjsFQaUh?WWgHv_VBU`1ehxR2ZEX`F+(W*ePOHl4Kb(AJRy>D^T?nxrQ -bA6c?9{66->#;mlfU1@CYUc_HO6H=jjy6@~9pU(*!;v$;%L7@azK_Azr7BN6AYH`bGos&xgXa^;IRd8^)vB(tCI? -Ds_lTdpw@>2&qU7JmXzO9KQH000080LuVbTO86j_-+mW01GYv03QGV0B~t=FJEbHbY*gGVQepUV{J(uleCtTwvey%irl+n#3L_U@+6_{fA2H -nxr1XRwT)UjzE;`%$`NK -7S%g7=0{P^w>Uw}isU|zZnjxemswN_OfN+UL3U{-{6h?$dH%)Uz2(EeI}G&fQtSsa%R>@Attyes@E(g -kCeD%+qN!cR*M5{=;YCWK+_`d$ztVe-;TL}RRq*@~J$aC55@#HzbY5Q?h5KDC1q-Dn(g3J=*u*oYRob -zz87#%Fi0p8@5%E<~SNfI2&#jPn;S-UEQf|DdqW5r9RXg`t+8;#MJgy)Bu}h@>s94^LYEvYd7qMb;Q` -A`!@g#|Gq)Pr>DOTx?A7`+*QrQR%FPN;#fX!a8%c97%beebgY0f^BwJ_UAa?`lC56mLEcDPks>97 -pTO8TTf_PDW{Xjx2MnGi7OaBLf($vU#6G(U+LNJi;!v#D%J>|q&BkYJK50gY3)0?{><7XjD&`1x;Wf{ -$e9K1AN~gloia`;fH8@01T14x+^Ndx>i^UR~UMvSMDU_snEl8%kfE2kDjCO{*Ln9{AWFlb*UqorrCmo -U*?S=2~6LcmCkVu7$RwOs%qQD#eQlz=BrFIs=u+3~Q=*vqo+STF)CMErraxfJ8fW3M9*Y961j2EzLlB ->+Wxl;v`zG!d^ZE6_~4!jei5g8*z6&+0WN`_$4XO>3U*;#9-gI!_f0IrLiM{3svl~WiX?F3OkAk<%_IA#I+`hKpO?$=c=KU -{vxGmCWts?zF2%J}iXmngg@badOoCMfyA;qJ3?ZQ(q_lm?X>vntHmjx3J`REKlcrx6Wx@)b#2KLz!4p -!yS#5yGEQ8BrU7=g-V!LEeZjVmF|hb+4XU+(}mf^ -h6$4dgN+IX-CDjzToE>~mIR+`HXPSx+ZbI)>t!h{a)!eH^e&Q6cSGQSB-46|kL3!+Q$1q&AaGv;r589 -^vf_`P+3N6_&u;l+<5FbUN@PS5pJ#F%Y;`(2D_{-plL70g4+M-aPaWcia$TR6N4UeK}?KLm4G(a8&T?HA&a_<1?3xf9ydu{#8pw=PANrhR8}cKySQRCLU-=aBLa3)fCE%Dv}BbcN^r -L4S(-3ZIKP0LN=mIq7iGGRnw4imaCN-_1k*5PP;lLafuN=8UZQGJ``&^FgEd;R{p8)QIqPmR1~UF44o -PDrr&0mY5ITeUn!IeXqC%Lrh}2s~9)}8GgcK=~;s9rVy2YH_Dl;6+)rDC|cm@mn!F&4Wy;87;bgE9MO -H-lF=IGf2$$Y|6iNeF)I6;$1w+N+~Jx-EP=}2$q^$JzqpBo&GNd^e``F027Am}EKDZe$JxzEd=_g%V|QR|8tXYASu!wz$h(s -OZP{uBQSyp|u@&kCp7`ySB9lk~gc_Q1o4q(=N8wYJ6hhPGT!(^WVCiu>xO2b{qSd=)k0aczw*B2Np{OLMywE$h~$F(gZ})X{8z6s)GAWM!*yBIJO3zX -G%I$6OjUpb>bg^N+h(EG-nBkOKxS{T3FSb>MOt$I4U|xxxmT$sEx9yf&JP -sazXt>4go2{kI=gmF$aWUpQySZ;Qfh~UJ9aJ)_?$OPU#joM-^XfN*E>`%5`u_g8sSCB?|xzw&s_}X9V -sPh^>J%oHi=k7ReZ64Hva3e~_5AJDry-R)T9;{+po@Qbf*%oxfp=lUcv2OYk -|l%6cEO2atkzLTe0KG^}-`vQpA|e$Vb|UzZ}nqa#LJKvPz(-vo{ra`4}uv -o8E&B)Wrhzw4*eMV0PR9_XPH5S`hsLXru{?3jSqw&2J?gSiRMW&;$}`98#T)Tk0Dom8nfyumb&qo5+_ -UTdCT{z42J6jSVA4m5QG(deU{2yKo#~^(UR(ana&n0b(m0uIr7oM{%IrEnBg@glFT)29ubh!q&)JKr< -M$k=JC}xGo9II+Q+9i2U>k{-5GS=)5`Ki5NkJDx6x79!ZI$e7DwOS_2fonV^}RQrE5MP-5&gV%)6;gH -)Pflihhb?CbFR#=5=xI0)=lAXUG(0+?UB5m#K6*UAd)M!I$&O -EH*V8GRpDs?GEsh>LyYAu{FmQb{Tj%w?QXYr%*?E>;odbsW=HZbZzi;r*;=;t{ -Xj`r^;enO_SYSXi)M1${ns(OZDZ~ohrV%%!W?y%x2#_Z@`si-AR-xqCi4yHXSNud~`ij>L}Dp*v`PAg -qQjbC5%<#L?b5!OVWL0;|wqo}*aCEMF~uw$=# -RyI$D}`O1axO9QUOQ#D~;*=RK=j6TS2#HbZ*w5P(ho}tf5|OL$oPuMxpSqdmTGt_N$BDFh%2d(;KXJD -DuUq^QjmLH#00S>i&TDH86QzNo^8R{WbjFY42o7V_3Iu>a$s1sfX`V7n>*+h((=%cAO#v^@6pp!iHz4jp9Tw22# ->efaRVkew^dD09RMWN+%Qg518i5|W+&{KK?+!J+)3{Z7jdrfR{qmX$u+*$l=Riw-L|Z;?~m!y!cGOE* -V!RjMMVYHpZ*1aG<-Q(c_?TBoHO9SH}4ImQ!eVN~i)cDS$h9#Rg?g|dvi4jB=yL-zi=*Y6&^f{Ffk&Y -n)`sznu)18I7$~m4NpAg1LcT%->-3mx;@KkHVd{5DQ1VC5Unv|Uvl79o16s1s0IvS>ORq~ZUz -@>{#lFmwE5-Er{X+Ybm553j-RbE`@LD)$hjdi7*M27ycS?A5 -DeQb@NKFA7To>;5b|lp$H6(8r4YDQ)7x6R2x<@r&QPBzGr+Bm$Ej(W~gtB9jh8MO1DyQjFVbk)4 -MG6%Z@+HBG(F_C+VT8mSab!%Kpj73wT8jEwq8hS=rQ?1Yj+Y&%k@5?q6Y>O;Pe#K -mMWQ1lsocb5UK0-46Hy^J`@foFUS_=sMBGKp;6~wyR|sPUPFVs(AaBf9TZ|3i&SI7yUNpA8Of4^>AI3 -?}mL|)t$PpYPoNU12t4nwspmG;Ei{8TQmb#_q7B>6q3LZ2dCTh9p^}`S&Zed-glzpQigcz*+1`YG>#= -e&s}Tva3i{1(dVUVoENtBu6W1IW8zkxWB;73*ky8C%H-jaI>8$>0XKR`)dMB2$eNdo~P~rtYQ?AnvHaUlxs2?eqf;zHWh3+rHl?O -GdvEFE(|<=@*_~u_p{@nGt8Glho5E^24jVO;5mT-HkuYkAzfTf;WvMoXJtU;XU -hW(Sh7A?*je@=*`@HkFUpd`=@1q3LFI?cl*J5g&Kq{UFfp@co5ndmunLz!oHb-xAuc3>^%6|d;UzT=z -v150^!ld7Mh&zq2VU^XeR -OL8C>3edIaEBAIC61B}OJRX+3#(&z>Nl|59(Dhv*KgkLnKk`9cDXHLo`{tUOXFy|5#7Li9N<`kau$Tg -CfEl%&Wk@AX@{w>;IX4q5vW1|nkOzaI|$77Uag?+eKg2wNYR&zE7l1sgSgU+?|`z#rVV@ZaRtpPJJv{fhpkBFQB4Anq%^-XTDCY7Zw ->E+!P=16ONs)6#{o-s@eUGe}Ryp -ji!zAzq(9&8T{ryYJSntG-KYK|OaweIsgmAu(3nXw2S{@`|kFV>rg_6v|bYZRb^Vvl;ygc>AbWf+*t -oRB%kfs~Vdp?qvq$Mh=mxOaQ*d;+~TA{e--ya8E`+B0DW3p~?5;d6RtQxXV70V3?U@WcAK*8Y}e%uKv -({`dU-q8LW1Yvi@Mlg!#azx51EsMbeAW07ESwzVuKFaWMqZL!68!Vn8|qfKf&II%GULy8(?Xi?v5 -Hlk<#wXVFj>^T -z1BbAlHTC}nL{;{jQi9G+BkX%aKt(^qCQ?xai)nP-?VZ;)-64)uo`jA8I_+)`9|dw6siE_29!g0I0yS -#_Lt-xj6#?+lo-qv77WP-PjH;g4qD3{%lM_gd)<3uA~txxi -i~+F{chOI{>J8DhSp1*i=l$!A9w5Fv=OobTK=Y_AIcx>=cyJ$#;<7O5UrU&7;r`tL$oLak-&5ZGmNVe8`+d>w7-=)wMRtM59S%f%_T<7FM1ubT1ihv(u -LI6NyE?MG(Q0V(ZV_RYrMJi)S91$sacWJsJ(cwB>#IUXSdH`I_;?;q=rti7Kml7n7o$z5r>M=!@ZL^IXVt>k3PUJDA^FPhPY#2dg32vSqf$@YEuauS{-MII+kqbur*Ak&#; -$I;q7+1dsJ2nGv`D$&$;VO&nGe>d(51!cDxBTPc&+-qqF##gm+|iryVwdLxpDuH+IUGx=+m8$y~jxQr -UrL{!dXX94-}GC%LLoPf?{_RFt0h|Us0LTu)u*0*f+jLf9*=dSjQ%XeIYkSak8$a1e?!EjHne=!v7`$>7*+qTN`nQKG!|&6Cj>ydvXG^h0e -~iO+9xvz(pah2h+nFC&oS?p$Z>8V&CbIUhl_&i8Icr^5bhe6ua@9Sm}=vn)g1qi6WYZm3blpu -Md-(PETHJ}x~c$BfdWI3xHSZONoM|2?!4xT!%L8FCa0&!z#M5eav^--^l#j8YzXQB}0h0B&`zGV;o)Q -5k%@$uQb}nGu=0Ehnvpsswo%OYs}61v(JM2KoaX3QtCi*1?Rl+C{(w6+mv$Y|xS`!o7$XLR!|8@iH?p -5}831*!2>lk#v)MG#<^T^JH;^V7S8kyuPB-)YpjRI%cv9CS5+tnyPH1pEg$k=U#aEY!B42r8Y?SQ^qz=3oaM+(g{Vf(;B_x=yV9RsWKxm+wFk`8Lbg*CI6X -$$kV@ROjVLOYy|+Mg@}fC}-lXS)dXD964&xy^YPQ>3LR=Xlv}+jm04s3Qn8|uqH$`Wpb0Zp9T`;Zl6K -_<;K6;L(-B}~X=zf{*Kh`6%wJXm`D)>b+ZPP-Sz2&?tRic};b(d(C -F1o9w92%-+Mb-<5jMyzElQ!Bnl0d!`dLg;aMdrp7dpZE -kW^d|&F3?AVK)iqot3VrCBE-p?ZCn!GF)x8IJXM719!Wjr_Del0QdprJ2CX5Ete&O-2b~8);Ce@>!1^8@JB<+NO;C<#+B}jqcf~;l#R*F@62>!?;3~hYT_9#Bs1dP{5FsuDeE=3kiH8 -G280!Hb!hvXojA8?BjDm+wR=F)OaPcAuHML@c$q+--B?JAkulJtLLyglaMOhX)X^1xBs0~mR8T0CXl6 -7^;Rw@0MFSVTAteI}>d{&~``q|#ji%B5=)&=18`0OpM8Dcu@j7bbL)b0{PgUhR9b~cCgM8e_aADp56o -aKYPKCogfMKkougOZH=Ih9?u`@4MaLSL{;|MMkA*Fn>&L+>i3BBAp9?|%mqR&h`7`omIN2*HcG@wW7- -T!>zg7M_GU_|SJYJ>(WhESzRa{qJ*5RE|5R8m$2!g#xSInt~q}8ar9^VhDSw<#y?b`qhjU|A7+r#*x% -+$4@_g2Tr;g8&)pm6Rfhw-4OLk`#CNr=57AMR6t}^5gJI?5xF;0MA%pWP`5=hLPodoW#7;uin)%+b%D -evXc=ogRD4C8(H70Dt=Pf -syeyc1nz7OMwM_8eSBN& -DRti{)r|Z0Ai62}U0T9Xu_=2$@b0i=o36N-EpQBqK)`Y*PZGPOFsgAR29cliGe!Wa7G;vH+9Yp8hr@m -Z(0}MY;c_Wp!89mAdK&?~lvvTvTZ$e~(6HB6umrz;@)azLUX-Hw`uyzdE9m+4GS-#y3i7sng4LDKa(6 -&>oxzi{?+?%Sdx|jVD(vhg|A=aBbpF&|{UMQY!=5gb%1?n{OZIee`P6M+87?6Ia!T+x2Y;-+O;lPc*cn-#0y>e5idYYGZm#&`Sv;GB_ExpvPM5 -ylbo?zS~3ojUMEd6fYkWiwRVcWgAMIY^-DI>p6jOckVS1vPE&h+@m0g7q=9iiGLW#g!zBiLG}_*h;{$ -sLHya6j7N)e;z0)qX5pmB6R)o!G$cIkC`an&jw(L$wYvVi8Z)5ZF*c&Uyic$MhzQp+6#>w;+{74O6on -l?hhbQ=uo9j{N>05JSB{(hz53g)}T74S$Wy1d5u|WDtDsR2Q{4P>WP_E=NhOr=POT8OzR^Yk0^#Q#M! -zbZ%lXcARTpgIC~zwo^D)PJrm`tFhkr)=eZh}puA-yYv$ZU>UgpUSa@uawP;qF`WFQu@C#;hcCSqzb# -Q>(J={(!x>|xNS|1>;eLo1=GOK&-2!QgTYTDLq{=t;$1|LdrUbaWqQ>cUpIwuiz0GD5WPGEJRfUy2fsBAGzUgc(h18raTp!ez9jFFo-VN9 -Ir{2LufSl#dn|S2cd55TQODS%%S7D-GIH*zj3w3@t|f&ezCj`@RLM;%d -u4-!Pv~PF+nV;xW`V2pJ3N=a0+J;qDv4&~1|d@2;;AvFmHvtP=ASyka()l=47DQCNO`!HJuCF~b{{0kuOw*s -&l_Ip!7&I+KHK02l0I%*HR%}m`x{lX6jdnp?0qB*S(<--6J}$Oxb^@uWFT(-ieCq&C>E+al>8F -3dEIQjZw{7P|=D7kBsMF;)>lp)WhLbaN-iG4!LX$%sccJqzWkSSKCxqG`iFv9;$Z7ZfwjeHO>~r`o7n -QE}}_9sxz!9(qpt2M(~{7#TLuSxE`Zd_9|M$80a5eSPocwurjaWJV;IuCbje{*4%Dc1aqTz^iG~7jJI -bCw{GeL{BQ=@Og;`5T-=(2!#o;IGu=AWKD4G0J$DZEE1e;`?*uazmSCU?K~lkH=2^jg -4d?{S1%a+4m{gaMbg>sus@vDQVv|6J$?Fr|2s?Cdk7lR634 -`P>2noOdd0WLNJ)7|kc*arDrcF|^gfop^_|G$44nee@lzOb4{H@OE~Ae=IbPQ?>Dl@^(v8zBbjNYS0_fT^sAk@LJeQgdoO_xYOJdv_0zBUDaW^bv -~xh|zEKsJ>g7E>twNPjvL6&?Uxxc!uXu5&fmA%YQk&jMV+0>7FY-+XlmdXz^IRb$u#0s1Rls;c36dzc+Se@g3o@(r!I4q!y`O#xC+ -<-~+2SryH5(KZL$LX{a(lY7K)X!ZAmur5Urz`itk--?2l5`rHPMq)%f7Iqnm85SQqBXZj3}5q?|3j#V0`;A`CC=VaIq&)spb$c>UKGuixSkFQKZ&&}K0IGe4tluUcJ)hWo_&6eF%Z{(+ae?6{Wg7ky+UMyEO<@n_Aw66_!b -Wv{1IAzrmV6)&mC5?uv##t*u7M7BT8)e&hT{(uKvr8E^ZGbDU|aq;2}kkGXT!`Rlm8Z{?%$jJv0B^Yh -|>LU6h_A3^2D#}!l>qLS@egsUI+Qp(({J>rOjJ*-dz$d_`KrLhl5^Drx?=aTK*Hl5K1?WonAIiknVgy -|Y2%k!^3j4uvi7}F@i(!GWn?|hh3S4ga4vfUEZWrqFVN|bfSsJTJ+azx|$5LZiqkK}WG3D8d_@E|%g5 -PJ{ANEC7i!gS>;b*vXsA6vaAt3gcz**cMXM|_s -fBONy1mVyZZ}z7YbS*qm>TjJiW*>_R5Iyg-W7jyywFyAH4Jv8MHH+3S8hAw|lsgDgDMz*DrXacE>U4n -d(wW>r1c@u1_v+KI7RHX_S&drm6l?6qkMKK{qnces46Bkom02Jc!Jc4cm4Z*nhEML|SOMJ{r4bgf!la~n6ZeV<=}+80YKir#fL_rbeKmDkq2QI-5rwz8@FB!V*_i -D+gpJD8Eg{Q7%NH!vJZ_9nM#Q+7orfkvbI^yv@ihaXP=EB@))D)heHnNOki<)QfD>55`ubdjq}?Ki#c -y^ChCc>nX4o5jMkVK5tKno#+z#}fBgx^6Hwn%@0u^xQRs!k9RgrHgSp*7d>Mi$F8%U9FN)NJnkCX%Js%#UFq8Bpp%96bsZie#8E@&&n>-Ew+?0va6J|6W{nb>RIU)^NMe?+UY3{7WS+VQ~;dvksD_VUA*OSAR0ixhAJnfI=bKD1BlaGyo{)< -Cf*2He)JRLyZI(fsAD?s%#HTMduHdi{GkZpXG9aJFa{Mi{C+n9bxqsg?!z}0PNQd@%i%2yZ4u?KEe<@iVpNtzVBV^LR&FPIy;p6Q;87d+p-=jh+VIT5bBtWiVZK*hHu( -7!j3TC-rl<>N3GY{r)THT$hCuP_=@JYx4!km?d_^N@SNha7{b=Bn1%6AJ`PYx%}s>t;U`A#-d5tr%5@ -Ou+LEWFFy6v@>veR)*vZA92R!8J?XoEMA<#`w2)?%rbK{$?b`3NYff+5f_QL`DOzT`Fd;r(GdoJTAkW -tv<>4%m1DLmloehE<7e&87r4OIFstM3tU4rbrkmXdnc)wXm3DGlBS8;eO1adq_g4#1-Tk<_~}O7tK7u -A -W3#ct0EGq@2Vpt}NJ*b#-EULqY{nhP;L*k6tfOhlv9>+@8C^N{lovrYe(?9CLl@7Na=^w7gX>?)Ia_B -31W**$f>ok<2>l&EGZ`HUW1Yp{$PaLDALED}Vmbhow>kj}D2KO%u9a7Z7Z3Je&hEVhdT1Ab?l$EEVCY -;9hc{?z2hvRk;LO*^H;e!7?atFYiv_T5e%U+aG(8mNZ*i?aJ(KwlM8ZT)1K0bl9P$7}@D^~1jglc0Bz -d3nBF4w|#bA=Aus?PJUI=UQ@!$!SM7t#VX`PBo+La{#IM*aW$qQgga2fMUKRVG20uqZ+itV~O6WG%1) -fcjkKOE-Yn}r7FB3n@*27@2MNOVvhqKzOv;NBPg*tWu5H2BBOy$@ro4@_KHbIaRf*-}L4KSwAl!kN7( -mj6OX2CjJlbkuG;WH{i`jcrHa7PvUn0N@Ekl)AI!Qd~4taF4H6^!j^y$Is_}<<7B9n;5VPh>vX!=NHi -c_&urQ6p>MkvR)H(2q_X*a``ghf=L?^C@&)uWx&?@Uhcryn|w~58R1TIwVeTxIB8GM@Cjij!9dfrDYB -3{-HR&2PFP=Thmrz+O~#v`UjmNu6aps)O_~wRWOBiad-OsdXN-FJywTsy;Zmel51qwYVH$-XKRd#qrmNdvj*RsULWvVym%TR;iXgUeI%##y4@4)R!w+o4sMPfD -;Z101wo*pT{6OyTs23jlW!N((ggehUovX$ofm608Dh4a$6E@ewSt)g~c>2N-B8+voD>0Hh%I-x0E3zFbq%J+ft$(4`@cum)DNE{>=#ku -RA=r=)ez|0@*184R)!1l&AqG^fP2Fz -fxV?In>uVd?k|$h@5hKeGux@EA4=YLCi?YAniOs!vbn86c3;TZls;J7SsWN0Zrb`AS&esPl^f}In0u?7U9)mWfOVl`9b*~?dsHH2mMC~(-jLT5dY}Yi^8zG9?t*P -wDQcty)s0FV3Bp=rNCB&6*|2MAD-Cmv+U}523SzO65JC!-HZLCH3Z%Tx#3C&29$}wU-@(|k{i}&o@TU -rbr>QWUB0^3!e`-*;PvqAB$(%{xRAVXeT8$G~o+}mF8osR#=B`D5$}(YoWt-!qLr~bAOFW1`x8kK)oq -3V!Oeo(9JwMZ=l(s~2NEc#62?U#=i7nnsuYh2NPn~V3E{LG_n-OdfHKlc&B}m%eS<2M*4(YwpB34Z=` -W2IPIHYU5b<*5M2Mr4q=dG2{JF?Dkua#k2FZ7h%_m}W>}IQ~?rOZMI8y>?Z2BT| -SZ)C&m1#TUhx3*B#le01R5Zad=fE0mh+b!lFkwZOS`g6O0`==+x<;jJB>9k~0z5C*q&T|)nM@=`3Yurf_%J5H5zM+Qbsab6#_f(k*QClw6{#9^NdL!l}?o#`Z#insg&pC~nK-#ucNFDFFhcl!J041>7>lOPMxkW@NMO -XzBt*??!sXU)9){H>v)I~@$UQ%0I_K+h5yJDz7i#6I|s=uI#d+e|{=RRmOM&*BLu^Z;C!HIFPRI)J2L -exsO^|2}P*yQ7v3_nt5@qRwmD4vbwo;K!Diw`$22pBD0r|=9dK41c2k#x7aWrfbcN}5VM`%s*}aj@|; -*3;%5V3064Eojg==aIxJjU*V&3uqYG<2+gQ#0S -5tY#cZEW~}qXGGzCHfT=^e>t3qs}kO!lY=|Ltt=+JZOsBv>7K9QBQdwlphqaZlS%o -#t%Ck3al*^=kEMy@p3_s5Erl`rX(o!{jPL<>JzM+Z9Yyi;ZtWY9mg=EKfLc{*$qpBr=QYDN6T6z*iV_dKeqFGj0JlAwi#ThEE-vsVZMh&AFD~L8I5Abu`JYGL0z|aw79e -GBTgOm5J9tjI%jxh3r!<=@kY}W-)eh?5QpOy)r$3=^=7XK|W+Tb~zL(I0tj<}*b6Bc*!L~w#<~ndJ`A -K2;wZM|tx{*o9e*4IeGVHz5-2*NU!b6mR(xkZup1;FkvEAyGQ(-E^y=$P`7?WVp?2FwYS2m}P@l5hqKvQgle -i?JYGkSco|Hg^FP)2eudlg8)5QW=@UONOF0*FjqFzSd|u2@Re!j}G(qB;mHTjr)<2_gyMT1fR5wN>c| -=WBEd7M$iCa0RlKGy^`SK2YQbkPC5`g45|qIL5$|zgKU&3(e#)xtI_j=%QB6nL4yY~`CBGiAwMR`=V$ -<@V4KQoN?37HMwQp=otNs4sM{aClqM2Kv+~hq8b8j@PPcn_edIkJh`aa0AjMJ%o|CX;6j8Ygs8-_t1^&dzun-zE?+t -d`6t>4B*8#N)? -!mr8izm7bLgdrpZo^*qR~B#}chGeaxP1B_#UJ32V;Ndi%sS2-!sr2YGDJvDi0X9@ -CWU+mFj&)c?r=Dnmw{K|0jn$9euqAuNMz61}pE7Zv6!&W#Sr#1@*C=D#N}$mD05qUBMLbb9h~YvlE?l -1%m>4_|3|AW!ZGWSvIyS9&Mm8$BP*&Q9(r-B+Ic=kc_Sacl&4=fdspmDB#60VT+EDuH$HF8}fKhj-VP -?>O&={?hvDIqUUJKf1(ZKXVQ~_LNnvZ&UlRbT4Q?I1~9jS0}mSgmMk&?#?&KJ}@(^y*ayOjz>-$W;Mo`a -LB$)uq9e_37y{qoc0<@N2mtIu=eOJxxA1Y}a;nXb*M_^&5d`uSL*iypVU`E5#`w}5SV$2Y35L$7q~n$ -y!@A5aW)yME80p@Dl2?a~u-ZIyhdUu>D>Wtkl6C>o-gP637^`ibxJxMQkw8y25ClEsHGq+yvnIiujHC -MsvCx>$Vp$lrvRC~jr`7o^B)H9hrr9N?(?iv0BX{7q!~1)-VzRq=mNO9KQH000080LuVbTLO|0dT9m# -04oy!02crN0B~t=FJE?LZe(wAFJE72ZfSI1UoLQYtyoQu+cprr>sJuYp?2zMlLl>ppgAM<+QPKLnTCALBD(WrTdC6|d95_OsMe{1250-7U2S?@hz*I{1O-clP)*Z&xz -Qvf+zu+H)jQ}VeOE{iC4_TJ)+v~Y(wau}o%5DtlO_usFDzwQ+56s7mceIdtQT_QjP5;U@i|F`^-7z~_ -}jR&lfE7w-OeS+BXMcAZQMX0Muneh(|co`*w!kqMJ;a$xNZ8j&v#KhxPy`l5p*)T&b+iV`a0I(t_Krk=-3hI3NGzPXTMRDctyDUK=>qi0<<+adf4_W -tbtzto*NBJsc&~N{x0Jp1YoU#~LB3*mg|d|6oV!2~DR4kUjx?s*T8lE|HX^X45Hy@2RNkGV-cY65mKg -yPKpCLLDx4uA@2TlLOQAldsNIFLL>%CagD`D?$n{;&DW=JIaxWVJfJ@pa*#`Q&Ch$xmM23BEbF-5OqG>j__pivVP04l@MUX_;tk!X!a9Ri$zQYbIUX$X1UVgoO4 -s14NvFSYlbdvS45DPQ-SG&ju!q%G>%lJfsvSo)_Y=iNouYkl#8EF(7b$RVqz5i6$+ESW*e1c&+g)_RpKK`CESqey4n+(8E>}$1eEfwLrC&Z>1!KC)sq^AZu)e -a^*n`tY;vKpUSn{!VsB$Ez^=L83h}n)@EKxh?9|84FI)_|Y6zH20WS=;y0CdO;aP5pjhg4oPB+;6DEx -owD9{f#OAIb+5L<_sVo_OL7k-^aS3~H}9*dFv#@uzBYR2zMVB8uWrLtHNhUwwznax7d3W33(7746(Q( -uA>LAAG=A#s^0MAU+P4ctt2t(68#Xv?ETz91&8>5r(<&SmwS5n7hXlPbkv}=Gbamii9L|unEJtp(%8t -*jPqSQ^eIRG<(MliEuJIgwT+4vf|2)(QZT24ZYc57mEh)MdCcQ>3)TC9EpRrc%TaRts%Pi_elK#tey% -F?EJ7d-be?o)+7rk-?L7w>$LLAMY>pvXRE_@0me9@c&GUnm%g70CtjxNii1UgajFVtG?BX2bjy -U$H+!W(vl%VMogX_kwQ+Xf*I+oc6x)obvLEohm&NJ0OYgPrnKOK&^uPTsWLQrz>?Dm=|Bt&^b>&Dx+f!D9Fvs$7qMj9`aena~_q< -L#-ie@(G}7K`VE+8uHPb9NT2uawA?(e-mo{2G$fz?Xxb{^;F8TQuDJFjNu+7qLhyL0+H^+8PI`>B4t1 -jjlXzGmfGfd`l=&)dHcx%fFl|N+5l;2ezV&-(#@Ou!7%XyxE|Bsh+v7f`?v*5d@rUnB;&LlK5tC20%> -8u#@>b@w0hEJS#lIvB^FC8;X~Zv -L`5c$%XRE`l{ouu$Mju~Nyv2Ks!#(jhpu?d24^T@31QY-O00;of09jjp|AEHb000120RR9N0001RX>c -!Jc4cm4Z*nhoWo~3|axQdubTyDca>F1DMfaSd7cgV6(@qwwauSLan+78yq!_ -0)qxlx;0Pil#kQ~Yc^>#-xDb4Fi<_PQ1y9u1AQs@?q>^XiA|JfXc#FPgY;lo16f|@g7AR#m|iYa$K@^ZW5v6K}iBm-RWIh@rq#+n!9nUDCfFE{xdgM2})2geM6tC -{cltSJ_;iyDf$@oB%O?ViapN@hM!klrbEIq67)sMZqxz=@&Y-W!%BFRKAMD7zfem91QY-O00;of09jk -c+(2Dr0RR9l0ssIc0001RX>c!Jc4cm4Z*nhVVPj}zV{dMBa&K%eOi4pUPE$oLba-@SkwH(xFbsw7`4v -w)fi$|&Bb>?@qGIZ#b$~c#t=n1Zl8~g7{d?RoBoMhomY?kJJuj>~dT#d->HIpqO6Tb{u=HGErG2nQsT -Egb0@dhEO&)Pi$ega^J9nmdAs^Tpq{$o?qhNAwO)hWM(e^m#BaDsc1IYjmXt*QLXGJc84JvzZ9cxn&2 -9EpFlJkf}y`2ZxUDTr0qIOa>zg5stDahN$=x%1SVHgs9qD-uBXWh3d%vSkgS(M8wAyw+xbQA(Sz4z?J -;O+?Rgrm|sq3iSj>!Dju(u;QF96U!h?G(W_anPPr%^`Z;_3=+YKUCuMw?HD&29vBrUQTeAm3cW;ule? -I{k+9%w%KIGHeZ%lZ?ISw_xUzo7g9YSE8g&w7xz<;#3M`S)A^BI0C|j#>O_@9f0j4)8zwlaxRHx?+O) -lH39bDgZ^Q^3`2&aXctOx=)$xI&j;H#cokV>BP)h>@6aWAK2ms3fSz8Nt7P&wH005-|001Na003}la4 -%nWWo~3|axY_HV`yb#Z*FvQZ)`7LUukY>bYEXCaCwDMO-lnY5WVMD4C$o{HndeQ1;J7k3snR?2qGb5r -`ynM5@u6t|GnAnuI*u=xn%O*e7t!X!28?7=jh2;T=;MxQ*7b~{>B+h$CF!lB%@)asPc{t*q}*X8I7*O -(K1ID&DtQ%U>=m-;AkqwjB(Bv^=si>ByYi#!Bd2#i%4swQ84zRwg@Hys_FpIcPVS~CFr_Vx3zl{H+~n -vdRKr`C01z=R-zuln3QUlc#;zCfS-xwiNaDD+c21zd*dB>y%nB{n*Kqm64WG&#pFg@r;bMj!32?bmwO -WDYJ|#FNK15b&$pa6)P)A^ZToxJi}ia_A@mm?-LTEVUkuXWCwur+hqSJofPWu4sAsD_%J#zv`(~`wXM -Ir^S1{HFqQJOHA<|IF`ZuM_lBxN#a(U`E?9yI$)^=Eyu|H5t0|XQR000O8%K%wh`EO56xdH$H$p!!bA -OHXWaA|NaUv_0~WN&gWV_{=xWn*t{baHQOFJEJAWG--dl~rA9+b|S;_pi{L7aJ0{E9}MKL6>H2Sr*Eg -!p0av*f)-tEoqY6uKo4XXX2!RGQT+1Jv#T?d#;oTy1%>nKKo2Nm~-(gX}i^H_^V5FasK`TT1Z->C6S7 -=WRgrgcu8wu3TC$rFikfUWQi^Xskbn@D9)3OQ%MrmI^MRFwSEhctuU -D&Zn%`~(3=c*c2RnaeijBF5$hj!3p~GB%)i{Ot`m(&{y$sIBUnGc9>nkcJ2iyK -QqDoBPg`i=07ryOYLmxq=5}mpa+42M -_G<$AJK|4@G`u}!eEy02wfIn#ns}|!*%W;Tf!Qfq>4Tl_gVt!Z~S_ekBLar#PzT*4>DTNSb -XL0XWy`bn`b;`KRLF}`$CCjJCwFQGFt*X{?fT^ijQqs{Ekd>`%mm`zLP{|zUsJCX8RB$&=k?* -LRFq`b~B!&uo)cR%_Ts%t$*`>B;ov6~aISUWpmPm7?<+C~ejqi7wkHrKU_X<*iww_RSJarae1np&ytv -DDI-@Xa@N`3<((IzX#z}$io<|=}qMF~UTnr*_Fvg#M}*p1O-S2&N}CqAnqBPJ# -`5^w843GAJ?;G4Cg&$}1*TYaW`7i+DO2RX$kVQf6@PMtgmzKw*3E@VP*dcf|-v?RdC)n|EOX -pi}S)vf|n&X;7J^b{0%ljF}lt^IuX`D8<-L2F9u1>&GPd2T)4`1QY-O00;of09jjg1FwCy2LJ%a82|t -y0001RX>c!Jc4cm4Z*nhVVPj}zV{dMBa&K%eVPs)&bY*fbaCx;@ZExc?68`RAK_n0bSj+sZ#KWQ3!g -H}?KVH50C)+3yu}@rSnawU|msa?bh#=Fj_!vW0h(v%|vg@2j2Uz?VT+I?Ck1Pz6vM3dVArr@3s)F$>l -Laq?%=BzljU{Ck6;h_U9w^Q^xF09rj|!6d=kKZ5TjMAtvWMF6_CGh9aQ$Jk{@2IP?|pWioqe{JX$lcd -=CfaUPU17Gn~OXMpCF5+3Kaae1YPKW$2>1U>DuIn?e^x{j*X~AwA4sug_5cLQR}@EQL(0h*Ocu46jZH -wv$_5C`t#=VHoV=u`*^!qZ^OIo_4do%<_=dwx@)KNitz-oJSkjsV|~rjbkCzlg84W52{A~zpg80^s`|K8Uj-bN0dZ|*x?QJ`5D;}U!LKh;^ZS4>B=r>rgnBV1fA$a_p*YA`vlaW};^xct!erG8thP4*~I}396Gt#wclPkiVnx5V&2934W`q4HU??Jk!aPKbbC0HYM)@b4m4!78Zp^qX4h -rEopYzp@1JU%A@A&r -bI3-F5LSa;>36jNx;5fwilTG)c_J;BGu7s{;ZG)Bv6Tn!Wn>PVYZmesM*@#Xw$w~Q`P}eJORr9Vlrig -vS8FLx8A9e880xWk5ko4vY*I7C%)_?exc6Y0p~y*I^S3QiVB>BnZ~RB&pVdjIfC5wKLIEu)wfJU_;ti -;PHLovW)xu0mRTqFrAQmu+B3gz#5cQWXhv6*?JfSH=aSKu`p7T=m??h)8INP+J06>!D`A+xO6JG}Mc3 -y#`ju!<=D3N1HwdE4VnNBp)36eCk!-m`E)xV5w;OVC%d%wQ*RW=H-=t-9j|s=89skSp|{iRt!$8|F!*D -I-EDa}>bt5rwrO82qzQwqy%%*x> -Bs~LrKxH7#F_DyYAlp(R>9rVR<=*8bCmLWurf0AIJ}7DP**Fi+xloqan|1X4sOK{X -YF4=FujwNQvfjV>bcZfQ0X=#BN0+Bk_e4dgCl~%NJ7W)Uv|s74e)H?uA>6VTz2lJ|hiA!#Z6 -*#?lT~qj=tJ5b6HIrWR-QJ&3J$+F0D0M{KpCcshXXMQvYze4s6v-~x1cg!lz`!-U$5kvo@RKY-^id^= -(S>?Z2BvFQ)B)x&}Y#&zB0T7aDdZIj~^@j5hqLnGCd9A;zUjpI~5bANVtTf)W< -HV(Q^+0=<5eq0_COnk!IEF;p>v7!)!6tJa>#w{eEAd6s2Hzw`AY&VR8iNx^6VX$JH0T<_H^ -TxwmMbSuCmkrZ4>9O259t<<@RHRMU{i+i|t&$Y9&0oFjjM2c<4iGu1>E@)L;$?>1y5I$NrQ<#~=#OW| -unGoq*p6*sYjM1m=oy>fPgO+sB(ikfbn8qqb?6%D#4hpFU7v>a!SqA!o)OyJbqk?kk7^JYy`Qu_teghwK0t5j*-9t_T$mPbfp;^a%;yq2YIfSf-PIHdQLmM+1U^&PfcZZ -T%nxX?o(CJ=UKL(_W<`zgBB7=DM%5l(UwJDsDJH}>pRrU|AAE&A!lZ!H=k$6ETq}}ZQP)h>@6aWAK2m -s3fSzC@Wmhh!jhM<7tMa6}dni!%4-{6^@UiBgPI64%lpkkh(2Ui(i; -pIJucZt~7|El32LJav35_1TH$moY&HVFu=jCH1doYo+F5ITwtMe1DQyzP$kz*@xb$BY5Dl#yG%_03R!%(ZnROrD7fbP!~S*bLY1nVY{Sktw^(sa98Xo&bk=!gA~$ -{9ds@4(F$QfscwwX5TXW@#taJaNWVnNQ+GiHTJ9K-y=ph~SZm=BH{fmz`imJ$ao>*7p* -!QmKaivWDkWG2^Eusf;tCIj^LZ2`ovpY-4y5REDc^%z)Z7nJs%L3YQcxVTP9;B;BhwnnAoV82AIy@L^ -ze0TC?2GA2e+aSl55#)X<}BYmKiJ%q~aNIY5R~3JP1x*EPGwFN^hGv9KSTO+g$s2-`TRco9x=9E6nMCb^HFJor!Pp`v*`<0|XQR000O8%K%wh!`*UJdK3Tvok{=zCIA2caA|NaUv_0 -~WN&gWV_{=xWn*t{baHQOFJo_RbaHQOY-MsTaCy}{YjYb%a^LYQCd*P0uoo+e(lHKFv?Ebej&+o*BI& -!N7#cO$9RO=Cc9DIM#5w-&>F$}?nb{X1`O2xFN+z(F>FMeB3qD}ipWpw-^h2D8vuwM|BmWOLGJr0gWBhT|jH*3MxWw}k_#WCZ_D$C%S#clWtOw55bP12vRwRkC5o~LXy<5|XB#%3FRQZ3#eB}<4Q^sQO|z1hah4XNk$M*LM5GbV_2(p8 -t>Sd0KjuQeE8|W5{!#GzTrA}|V#DJksLD7g^g&y=u84Rk=H=hzqH?zdsQUYx|NMMDKl}6f*K+m*@%lJXOkHfM9K#24Wv$V{!M1Iiwal; -ut@Y%Hh-`9VYbPQ^;{gA90Ppkp)yKD=&OhDEug>3JT%DiY%&%|W --u!fZehuf0Umv~l+3Tb4VC5(hODTi%Dv!tTZ^l4qCT#jWE30iHZsC*9+W)ze%mE$xjJpfQ8Q!8QkWD4 -fA{Q(1#8x0g{ftPO}1ugWzl@6HFL%C*mcV>x_S=6tmQ{ht -8(4z?4G1^wOdGF-zFz$BpeW(hAq^h&iNE*Pjc`;?`EzFg&5wJl&HeI3Y;W4SRGpg$=a%ZInG?tHcU_O -*HY#=Jc-Z@;@k3Wo`t%}5^3q_iYp!!+J+kvbsFOCAbVW~>k(MoVp_Sh3EkBw{395d%90cn4&-a6d9{K -9AG5oX^Mf1Ydyx|8TnScLAAr#2hDs1z*nE+QxqU|jV1EP -<)4((6k|yyOn05XXiY*Gl*!wX}c8o8IEU8Mozy(;59B@yP>` -_Dk-X2bVQF!bHqf-qZb+C>$gz*g_2XBHoT5ThfO)KGC+)#pl7#pJF-pX|AH3j~ -EKbL_7U%9J$O(KPzEGeDjSILb0VyO!m6N_dc$Vpm?S{ -uX~$&U7${&Qaxg3^2&&!G9i0<_&5)14+9+dlXHLFPv&VEE#W|Awfpqs*hyMo@P8L9(5-ed6#G+gVS-y -I?E;q@`d>Q`p@#{ASh2+}wPrQ~IPj+XBKa&qMYX90kj1p-e*>Ct3?ur&n?R1hdn^$+0$BKAQA2xG~US2a)CH&Ud$NF(voLrjHH|DE*9p?dEse0d+^f|3frx)M^}9i -Dj1Yp`VyUcI(|S?|^ViYN?ReZO0I7bzK{-WwTJfxEo`ABu((!inL_?Qt>SbD-1C?p2eP&zJIIvot@_5*Q|0|~8?Qo -}Y0WPr>s#)=`nMK0e!a|o&+4H5H^=P8JiiXId#A=Ol;%KT#qq!p2`zKzkHL0mRzYc(Ecv|4K1Tp|1wI -;~1_w-iv_b!+#PJiDdeQlG)d##FUSjh=w!!~6UZ636BZ=%8+$S2Ay@!+_LF)v<~On7LS64e1=2KD^Ab -dkPe`Xrny`?L(h+2v!*CgSzsXhU!o3#Pc06m1R}gG)Vj$zf0VO?D%+*G1%?oS -0x)59Jwn*LRuv2r1CywgYsgOLMa@H`stsg`5amtRq6fM|2C!p}*=&MkJWPUBed-k{%C}3#4L!L^TEh( -%-)3w`n@>*;)J6kSM@uyWL!A+euy@lPy5>Z<72VvA@EAkCbL9t<%boRd9584xNHt-b8VX_Or=UK3%TS -^6h0Jl^kF|#VSt_a?h)J*3nZ1FNAkc9;WQji?wE#yhEV_sbZBPP&d_E{rn)bhZBa4jtQ9l*cmbL>}jj -(rlmv9fF%>{${cO^|(P^5BZOklKK3lQGs}2M@P`m!>*xdI1s)HHVZWLAZqrH96pdRZ&5dFpdS&Dlu_< -I8bMd1Px>jaDw7KP=$AeU{kEHEZ+*qJZUfym0l4rZ1mdu$FWXlA}R`k@?jAaki*oT9LB`mHtE+yz_L@ -b5CJ2)bz!EX(S*4yrW+k!N#fPE&!Q~qjb;Q0sz>=dJp~eX#z|;)@S4M`B-j52Tt~x7^a1d=&o}yR`6! -gHR6Kx|Kp;?=eVGr^Gx2iT&>;{qIrS>cs=sm_s -fKO)7u9MvF%<5*k13)N({k?%}>|D#q2`W3MUFCFp!|$s -cQ{wquVNVUm7bT<($R^$)=(7qA8(R!)cW#>1=Lo -V6tNLvEC=1V!TTP!He_;SS>Pb?h;gpYW(D)ng3ySNeID1GMF=S?^bK& -B2veLf!!icC*`xTYJ&n7g>le9~T_97H{z`1vJCo^+L+3DN&~()_QpOs*!&0VPYCI^Sg;OX;y&M -Ij2pUk6U?^9dL9@`2p4+^tq&&3n6Jo8i+-BR4vr83Ph3$S!kUE-%vo~Rz(d|tZYSbB!@--g-TBQg-M3 -kfE42bG=QjrSb&xHmDN~9^Q%^#V&Gq-tf;~@VduAdsWOZ2PIe3-(LD~Ra6hH(pz8z1;rhn}GxnbFkmn -yCog5Z*#RXG3{BD&|?IdfJPU@G8*+~9bPK{$t6W2TMGgCS+f!YkSh)i;Uj&Hj+4ureyT<%%|#EtAl4V -b238i)El#|?NE#lyB0`r06|Qvh_-I5J2$DS<38$7%a`2V8%dv$?Nb@5I-BHBWqfVcoMW`U=2NtHpHqf -OSOh+;UK#ld1OI{t&j@I6IRQ64OhOu5i*63z~Kv&iNW|B4=FmYy_GjS&K45?sZf}cv!$JjA$ZX$fy7u -AXs8FSl~yc$&tBllnJVP0h}bExXc1p+{at&3`s53^W&w{zH#NG8BDlb!>L;}E;3E%wwmd#e7@Fn|gBr&9(W#%?sz@;I)Ng(3F+zo<8DwhnJ|5w_Iq4{6c=q~cA*bM^BtyRnO%d$( -Edl<+V&NBMG_z^bCBLS7Ilsx-hdnyrmwCDvra*_@uk4P=sTiJck9?P||kJ8=MYpe9krDvjQ~gnc*}Iu -m*S_2+hu40Ih^1En5W--yvr;YIh!XiUU#Ro1|e3t5W&!LIZ1wO6mqLUdSXzHC7JP4LR@e7|LX<$$xI8 -l?G2G?jho4EXbn$9VF{cLmMz6nHiHvm^@w=E6tI6X24&E9JQqw~gkf<_EK%jMYUu~tWQG<2z -XBY@gw><&kgNS_FxMn2vl-!m&-lSfW-S~{$-D1d)Nv+nBR -PFjI}x7lI>tGQK+UUC&qw}(h;-9c-MX9n9s@fqHC6ytn2=aO=Dnh!H_0U4yM*qHHruOM6;;Xx_ -Naj*M|Gqi#*Dq~2iAI-G!=Va{#;{{1rfq6k-O4U0?KFD>MLm4w&7zt6WwTChzsOoVYiipo8rD&F{>`v -@-R1YuA#Vgv3zT~Mcf{~T14vz$#@UcfqqfqX4j7hte0BIV%NGs$nBV2o)Y*IcwI;#8a_~TsshM;S9O6 -!q+pj%;LSpMjwrmV`1cm@9`_;I^Ku-M)|K7&Nw0B`Q<-4EmrD^zjM%(Z;sBv<^kk;X=U-R%atbO>hYV -AGU2~*P^?6?{)=oZ&ab=!XGRJnh|O3mw|H>lY(NCx&8o}}UtXG~Baw3nnn0awXFY7Io9E+-vi7$p-B% -#nRr4oC6QGsqyW&62sYC{$L~`2o%b1E@#oYBAU@XuKh1*h59*UBg&qWhwMEc+pea{R5`@8{B)o&Fv$kYpoOoIxc20&QO=j;6rW9T)gmsY+c#5 -dm|%K*Vkfl*ss$e#5R@q9Y8s7xc2&qs667Ij2q%frhTq%an>Bzdi*fQ7a{+r-l|c~5)~jk}_^RxO`Dx -_@fp{MNn2~-u0Wwn52g{ix# -}9mUt7lKe#RbJ4>9?<_t}H(mTS0iDK^EpGxNLG|1$(#z;9l~+4NgirTn0_E1Vr49g!>2_CnEsJ;dDbv -X>63M{$Z^Z1_zh@6aWAK2ms3fSz9cDx5@4X006@f001Wd003}la4%nW -Wo~3|axY_HV`yb#Z*FvQZ)`7SX>4V8a$#_AWpXZXdA(OnZ{s);z57=X$ssnzRiYn-chH~; -vjy7wkmpR^CojxI-+$@r7W$p)HzEXYPv8uD{>*XVphvv=9esPwK@%VpWd*3)vSOx-vimA2MJ9gVYLH0 -qw!wQiDqq`E0VVsHIf*HkyrKmRmUcXl^b3td%u1^baHhX(VJ! -be~>F*aFV891=V`uXYU;R}TZ{IGVES2_d%Tj~6n3H<7H)4*3jxwyEH@5dSzYMWju?OkgNrB&G|xB8Rv -sJ0uEZ`9VnDiSGrA-{_o_Q8`bl(Eq%El{hrEXhw1r~<3Vs<_cnrJw|rS1u6jW5S|W%G*fTD(@)bdE8( -P30UrIgmXc;GT|ijY$R{*fS_aaW(AA@+N2uqUJW(?9bhQ9LB85CPhpU}>6Sljgb)X)4Om}Oy&@kH1FZ -_zBK`>?FusNwhN6kSA5$GI5>>BrCNr -YQSv*K{UjE_jaXd}AF;C@h*vulCVLHr%A4yrI7{yO|N{3ATiR+{5K4z8wXYM7XK6}qEC0AM=fW1wR8{ -dx*7%Z%9%R|E=_zke*cqeW4g4{@ -R9=qeosQ^guX2bhXtx2%aLu@Z%1r%&Bx>BME|9%e8~gxAEphN9u!s=GU8VtOK`Y{kvyIXdAIVV1RAal -Hho4WD{+BFn4>yd+#AhsZ&U@!ck!T(W?nyC4wpQgnML#+IxTf^Ly!t(Sc`*nu;m^J%Zu9Eryp -`Mx{HP;GQb$_En;&(Wx7&YpA45A*Bm!sQ`rP=uY$oL^tRyGB6*I)r^quabM*q_Rz{t4keSJu`KSrphnQQR?3O?<6kOucZEL1Ivw%TuK(J4=S-)KkI&Z{|iQvqEK9|q4o8e1; -!G}vh4H>KIke?FYdzg81G@(L7_e=6GO(w_oz%66kd%d>h@OM49ybH`?`s#;k)9jWk7&p>%fogdZdLGlERF3k_>?%|^p2cF!W^Y#Z4kZvhk8h^xX7cwzPf$#I{Kkks>~&9sIK*uzLLUJpHcZ13D+T&S~;IZI(=S89)%Hqks_E{E0;g_JJf#HRIa=B$-|4QL~>dkN- -Xo08A@hgnT~rvDCRI;F44Bp&DhEnnR=uc*?Qpb3iO}TAe8U&s^1(~pk1Hi*C$OEZHvb)mj<}R!}4Tt?>b{H6z&`Y&Iok*-s$45>e(%odtl4L@qCrYEF3Yea#U$8#P6-3*Z8Y<@iBd<)FzQY?@EVoprdL4r= -6x~pwYuw;)Y~2vmu}}Sjhpb}(;aV|I4+nKzzyLGpmp(Bb=vCxc>m3h@kVYkk%GZF^)fH%{PNzEVCjq` -|xEJ!LFmf1N?=ihnw$79;S#DdpdV$_6rnB2$f_gzXmpuG@>oH!8neXIG?H+X2$@~n+Ah`VLngt{FdPU -GBEziXLzln4YgyYS@z%3UuB|?Ym%~u`}X4uVGD!%odfi<_5cOoyT`M2CI{_ua;i_`G7zN*Y>e{BpYmE -;D>wb+G}DT^Zgg^QY%geKb#iHQbZKLAE^v9RSZ#0HHW2>qU%@#rYJ0Z)5~sz4 -Wk8lDO|Ycx8V6kw1OhElF&9eINGgug^}p{(k$Q1ni`8E&N_WTSUU<(*Tj=cL+rNf~l#zX&tp&YWsyVq -;KAMgvJLo{r5S?HldD_})Z56^NG$g4c!{dmgipG=(AMIs0T$16`8@FP?6BGn-u5v+w0MR7lLLr={Twz -6dDqAgkFT{+b5f*N(XhI782rHsjPTQ>qMkXRm;o|a+r&fg|m@+R{lu4%*bCF=Bbq6Q|FUZDapzaD?wa -#dI<<#b1J{|=7M+f_V9KU}zKzr$WfKG9yMF4u-eD1dp{EU^1W?pau9&zF2f-$<-b|#PJ^N*j1%)oonY -R!*N4}!zfIod_-4mi?@P-lTUUr^`B@0|LbGxY9s-fp#85s4AG%_syC$BKxq@!YQmLHprX@%k6Nhlqe- -{ntao-^xb?P`!eu=8uc=p_lSi*ToZp81{^`Adtkgkp3B&9W5%S7$_br363;2RpBr@OC|ienJOJ!mw8V -MUnQVxu2h~e((UznuuI6Wl<0RX$q~r~m5PSe1=xOtMGEQYO2Xm+*e@dMH{$Ft*q9W``qyr^%j-`CG>m -RAqY;|i-a6Z;a7#bP32Z7uK^OYIS`w&SDEHC`j{eT2O5iw1uT|cigPcQ#s6?6IH#8yd#A}B#W5%I6zs=z}3Ve9APS-(+5v&c&1(gF=Y8}c@i<;? -b~K+G|j>NS*X)!TXr6>_*tu}!WEp2b=zWUo -9NQ?UMQQvw%k(EYj;!pqOEVYFPlkqj_)?9?p_{C((6KiY|Z_Vv|;|N0U}g*%$Ye7C`H -1BA9T^FyfMpmRYkXd02*K|*DC*)ZM2TRP*Sb7={6gAeWFN0d$b7=}Dmf-_^Ho#hMY)c3l8RVi!kps>4 -VqV#;x_4~jz{XSUxCnr)#O~b%g*3L_pD&RdkLz(nBTJCVDbQ3t -%q8xSP%ylbu=)%4#{mQl0Wh&*{;V)pBeIO4w%Vs?3oiW&$o?5RL&RTHx>+@zfX}I_t(tF5QTp=Rbyy{S=x5e7mN=1g}N3=U&>8^H@kipk8f{vm*k`7Pj-97P;}Yl>G(jm8XDN)H22=OqvcNFs7W -7>B!F!tW?!*Z?Bw@ce8P@lcLFlf_8qbfnLcCWynmKkKfr+or0$9ty$$CmBhYtU@wMiCPhHJ%Wa>c*xm -EtB+E71b;5YM-X2x%(AWRU*4E4&Qk;2n$)8mqX|`oClHnCY`7E~XV#RD4oy|^Vpi9gjdyB`$+*6#iS0N8-(JujzMsx~_s&+yKmfEYuk5H!rNMQyY` -L4WPGcx@dd+T8(*{zH9`E-*RSo|o+ZfzMA{`lA5j@EP`JWX}LbAQHszK0zcwoeb#<5D!F{yEPc+cGiu1YJ0I$5Sf%7D?h_##m5lF+pHet<>@4yAta>!a=0<8iO -y@dGA)-L2z(*87T^?T3T=CZ8f_6E--mu?d~~R#JTK@vahc*Rgf)NltztA6pEj%KigTO9KQH000080Lu -VbTL1t600IC20000003iSX0B~t=FJE?LZe(wAFJob2Xk}w>Zgg^QY%g$mE_8WtWn=>YP)h>@6aWAK2m -s3fSzE2?V88DM001i#001Qb003}la4%nWWo~3|axY_HV`yb#Z*FvQZ)`7fWpZg@Y-xIBE^v9xSZ#0HH -W2>qUvX1D#0nIlufQO{1G+Tbf^}`s2JHqTaA}E(*~pYeQg*#A|9y9)UMS1yioF2K7u)3V?s)gyJ;xlu -+so&_oW2kRKP#)95;rSz#Xp$@9zA&Y6wZ}M;U!bLl#`>$krRF?QZ6+=y~wyUA{SgG@T_9#il07;A53y -pu3@prYf~$}Sb$hpr82;zEDbZFlzK90)k2Q#@kaU3=VSu-cdrkvl1s%arg-+otxqPCv|w7pTdo8v#P1ket4MQ|aa5u-1`uWtJP9UtBjY(_^ -}5ni`_su5dDQ)CEGGM=dVVHRaLPoIV$3=lwSO#Xrdvoxf@2O#hE}^b -6X81r9w!*AXzvNrUv6f)PWL9tmriA6b5{<|MffbP=4wyzoWqq@vl`Z~~9J6eli3jJb>$BcDl9*;ZEfE -)KDGk=*>;?MA!^5##f-g36o%vvRv+eG+{jY#IYtG)tx?J`Z#Rt&$+w+p)W(0(a;mBEQbpmDxFD8D;xJpJQ#%*$knMkj7*byRo_kEMyDMPjABn=yMZ7j -?LmWf(<5rgsoT@zSh(5E*fjp*uQF0*d1~!2Sl4n^<|r-1Hza}ZDq+nKf$;f9*`eq-M2B^86Lfx^xwL# -M{tG*Z9jyvRr(<*)5k -iTO)NVyBNym=5NKpheVqq!dg8I?yN?{Z{+eOK;f536hRcCxaYzu1F3C2B?fC_M6@dv8cyb_!XCHsKWy -OX>z@fB5Pn_;jq{|nKr3+V2t{ddeEsHC>CwQ`THnJX~@lzb`tGT|6gD#2!eN*vwSyd%zz0?aoms@q+R -!BG&2;?45^)#fr)bYk*6m1P52P5c#06AI(5^{@yRi~BfG9m;c$Y`UUPbi?mm_mTuY9S6u$=BIMhfx{+SiC)#l>c=GDbw -950t|gn_5130Jq$(__@+>t09O8E5D)Uhgl755Ac>sezsbmpgn#Nq7{)v+lc{swg_(-VY(Z(k!V#EuF3md)odBuRZ>lPQkR6!giKM1 -wIy60&CG~N?F}>O14CbM0fX;lm_>YVBZ4_Q$BvWYkSl&Xd7B^(R -nE0|XQR000O8%K%wh*gSt_sX*jo^)N>u^&IhvgMC>Tl`|_-gA!by*l=cZ -XR!*R$sY*Sd|}|?+V<)JBIY~;?rlk);y*6OdBQ1jLaPPo~IxUtnM<9$a4-lq}L@&3s_wS7bMqePm?6C -u+oqul1+vo`j1~g}a|MD$t+JGj6WRMT4n%9a3j0ydaf{oqUyIt0~T0HCqL4VT+C4olutA|Md+0M$mDk -W@otE(pKE3#JXikZ^!-G{@rC(ra_O;W)Q^;^rr%V^|3rl^f6|H^BeImxYCWv>E~Jz$|S%kRF17lX&F1 -zlYo%{J`BEYSOUzM$6FjRTF-0b#&}j*GY4;oe$)>Kp43#}wxPU(Is3b(b3T)*Ya`*Qi@|X@8lKU-iUC -w-=2AqdNK%kB%Y^yzpKQ_NtONKPwQ+@cjI2*|t0lF(XV3(qMm)vc{4>P)h>@6aWAK2ms3fSz9I0PW-a -~001um001ih003}la4%nWWo~3|axY_HV`yb#Z*FvQZ)`7PVPj}zb1z?CX>MtBUtcb8d5wb}e}FnYyAa9cQNVjI)tl(=Q>bo{8!;1GV09gja`77 -{DIAW*u?v%KKaXkIPt6_y+GfsYeKtSq?z{jO>ihw>*mnP)h>@6aWAK2ms3fSzEyRfRL#L -000#b001oj003}la4%nWWo~3|axY_HV`yb#Z*FvQZ)`7PVPj}zb1!CTY-L|#VPj}zE^v9RSKn{kHVl5 -hze2mb*xSX)vb6&a$gnj_yI^TrG#R!c2n?M~pS9SQA<4-d&_8~pB>%iinuq$umMBu>N9q%Oz^*_3@$c -bVQNWq3w@R#Ya|0W*WGBZzzGCM}B{&*gr{ -ds=&*TvO4x2WcZ_Gwpr#No_KL>}-b2b0sT4J$gBYrKB2fh=&W0C2G3E5$u?!qz%VcA8yWQ)U;$}O3-6L%wxsqf)vq_{xO -2gn2tjBqw8$*8MDLaF^8FY(og3;fqa?5mOF{&QywFZW+#@hHXp!1?)`rlXgXU~-vu#|x?i%>)~JdTep -xyT?-jvTW57X{2+0a`JPqiM(1EMmcO!u?B5|aByg|DzL|tRlI~JY#;aWFTK=szhOBz>7I*!R}p0Jk+5kJ~+Ld)QW61KP7mk01Hgx -=VngO87LXJR%@WB7Qce1)HnYJS23pe3x3h^I0A#;At?Zbk7rXhOeY~E1N3a!xmnj{OyC&U?wEqMY%xo -FB(jrjdX##)QFOHwELyMqDF-J6rVnei<^B^!hi?D?mVnrCv*-{?!1TS1osZ^*>p%D*F1w_%d8#Cc%(` -UNZ)ILSGcts?NvTt@A*A2jm?eI?N|jiSay1NHTg)}8uK#&4f=>tTB1c1>cFp9jM!x1EXVTx^hPy9TC+ -7XVHRs03x00r=HWQ;2kajkwQLmQiD`IYnQXKs$Wq4(A^P5SVo`ao>6=_U^=^)7c&DO$a4ubBlVBd}WT -L`5V832cNKJZU4HD#tN-)5w<>H-{3hII`#MUYfUc1GO4m)6i2>h!y?7CAr_`;&4uL3gc&9911 -eU@;9;UQL&&$p=J+-}2eACz6<-w%=Tl7tB*tLT8eWQ3Ps0YN_=ya$*OSjbPXsmT=7c0CHO3RX3x?me& -_^JTgIu=~5T~(#>u_`1_b=2+BrC@Ic_JHot_N#Fnvg17LVjQ;Q&&0d8bDzZi-iCAtn9M|A8}&Q@9368 -u?IBtqt)@+aza!RG4gB6?gsl}{!yW+eK8eZhJ_s&8+*&fbd(p*WPdu6GnQEuT+AD8z$$X=`V_zT(i`v -ty{K80fF9ni^->Q(Fkvd1H)@YkX6|s>qB|9cI32?&4*>$eVzGlWP6(;_ZgO08ZTJL&a%S&v96=^#s68 -`<^GIGJ$-Yb>ZUCvHI@D`EFFg=r@6aWAK2ms3fS -zEmI_qRp@003$Q001rk003}la4%nWWo~3|axY_HV`yb#Z*FvQZ)`7PVPj}zb1!mbWNC9>V_{=xWiD`e -eUwpe(=ZUn-}5Ps_9c;8YfR!L(vYAOHZe^T3<-o#>lv^IlSj?Z` -#qummBIWi+;Ck{6P^$%LL^+Yf@GkrkUG$B+~?rQwepq(f1Y6+a<|QaCW4xE*@5SIXO#8;p236ETg)ST -2DodKL!W~(#HogJidord_-a{4D&p4iNe_%+e^jmT(O;K28=_FNUU_9aNmF9lVP=DgY)mxvvDi|{p85d -6TWNoR1pk+f1dQ*Pf}WJAEoXaH!T;>;mvGo<$*!XEBFM({3N=X$H^%;z7}c50| -*|X_g&nlE)Vs+Y^W0b@JgVa2k#}cDqhoPSGjGH>%VqT}wHlolGW|s2g-pTNgva5KFCrcqEy`LuNpRbC -dTxO#;4oghxf|fl!_U8-8~re*sWS0|XQR000O8%K%wh6ND==J^}y$-2(ss8vpU62+TEpr??z~^IeMIyHj>XiJ#aEPa#w+br#UGBVspLt(IUa%fzDmC4VC#_KFn -W+t5KS?7HRY&{0mkC%{+JX>YjK9Fr$Pe2fpw=HC`~uw10`yVV?K%OB|tbdkM*R(sh~mm^#=>*%P^X1| -#tkn8Fb`c3DfxH*g;?)RQrTAoF~q1Z||mtzTwf(>b0^Lo5+zl0qlt=TNO4diXs(wlHcE<^a`+Rh&?i= -z>)W_|?G2<(Ht3Amu*R!Y7xoZnsjm*3aZlQS4Rq^sBJKc2~6bWIG$Xwdc!Jc4cm4Z*nhVWpZ?BW@#^9 -UukY>bYEXCaCtS#!3uyN3;@u(UomzTF~MVB5-}nc3F=}>{eB^Ie(z8z;C4}lR4&mDyVTCLHw(|n%El0 -FITOJQ3;u`}a4@c)->fZ39S4IC(M{g@h~@=QO9KQH000080LuVbTQ4B3;hq2h0RI3003QGV0B~t=FJE -?LZe(wAFJonLbZKU3FJE76VQFq(UoLQYO^(40!!QU%@1DZ)DV7osP^pJ*& -)eoBZ{0AQK%HU|EcxnK3c)C}f%8Tatn`D?+D>8R8l3m598Si?SglH^2hLsRy@4mgzt&#gg%40m0|XQR -000O8%K%whCe=C#5~CCJ}?gwAZEcE^C&<<0>Q6u(9X=t%-VUX&bjwme`UG%Rs}&2pupGi-N9D$ZhA0n?CmJ+4 -@orU&=P_?tOO;A8GO>n!nuBEWM{~obLZu{QrG{K@k7)(@?<}Y|G&x|&AxREMdH}&g3>}c)*hW|Vwfou~ -X@B@k*(DmxkA;u30b_mxY&;u47z;rvB5(jwwM;?gvddVThwnhlw&Q0;{AA}yD@Bl{#I6g%2fmrV-J)k -l0f42V^UX-;jEN|MfAhWd1zgq!SmigLdSsJYJQ9Dh>jTB -IR|X9|@CTvvSw740y&<}!3=bB_B=Bn65QW03jg>Cfm(w`t_#_+YD2f+RE6M6Ed|7YzhXSFA}`igN@t4 -=?MFhgA`xOzTgPh;A*53MNZR6I|UnSdRrZpQ#M^o&oz(J>V#+z0Apz%K5qJVPM?4OTtdMwRS#L{6l4j -cW3&g)_)vltO%w+nb#U1Zp}WpHkE%d`v^vJf3bV+N$YK0_}6IVI27h629H{iDi`Nsyzeyv}Z&Z4tE_A^8%EJ#TZ{3Fx}RTB~P5;E5NY}kXlQapK3+r8=!3r#(d92s`9Y^Fq{ -c*kRg0_drsV>=UOFxH#7?`o$$OsTKm+ai7@EjVT(bbF@tPOGO1Wjswd~le*D@~4e@|#NrbNiU>hxLLf*4sg5C|Z-sp3`}hy{^4b4?u^G%keR`kZh -^`-apT?&(tn6Py*?oy#`K{mON*U!n3=nR2hR|oMcUpY%MTVmvAw!QE-=hTqG)qYt<%4rYre5M^8Z{>4b!8AcaH}D}npK$&(bm=nY -CUgbBF}cKCxQzR$yCZy4xP!efu)TLT$PYl_HSW)F0{ld3jA!SD$w~k|(3}1OucYC7p -Uz7d2!_y36wz$En<5x~*|LLG=`qW~Fsh{8j;R5FB?57HK*B?|qVy$Qx2M -<(oh~I}$Y$FBP$~y1^L2S%~L=Gf%$kL7MtTZ7;4^(;}@xLMh!ed$OHMaXk1X#wX-^Tq1+b*oOeKPHzN -q|v^E%Qq6h3j!2DK@n`2#lQN8TB!2Qcf?NlcDklvfT2bB8Cgx~LLVxQuin!ZF7l$1)q;d-N(1HXk1QwB8ZZT^wFqkEn_$?2BX{+T -k9<6u|43tGe^+W{mgGX1GgI2ThNl-QONpM}9nxB7G;QB%jFC9%T@&Gi|@q29qbPI6)!?xZh(o>;(Mgc -pREvu1_~&8ZJFsl2FOd=@J;JTh?P%kwM5ly$OY=G<=D7?tB$+(C&d9Z7$@q-CdMYr7rpy66S76%$P18 -s-B8-6F3i}%Pvr)8wQaK{9a0tEaG=^y$3HB2&f0$g#4QfD%^mGNLt2969?gmGtN|dq9WloTk!6_OE2> -S$M=XZjL~8_cg)-q*M46DPCVhx^pde0r{>6nndWrmOoU@5UMq)OG)`S|jI+nkc)`d(O7dza11o -&FT!8aK?DofVC`Ft(Sek=?CLE;@1FPaJ-ve9HNJvWNBAL4v#v%}wUZPp&_L$FISXj@2{h}8f(G7xmxi -!Su>h9i->}F3YZfD7xYu`?>qE(mmOs_39D!-X(o*oM@>e~zENI*BoCQ0UHmKb^5Rzto^ZM0t+*Q&;*qrYX;BQ@As!qI(4-NV577x -1=J|~!1PjyTpLG7-7Z;SCkKYCc_|?Mvs|E)6)xvx?Fx_uW+e%wk)5oDlp@uVnX`bD^6_V5BjKR*gP5; -bm>tdT&U*#fK3IqgwNGHX-Ih@EQZh0xF@ -B73Y-lb8Lup{7k=VKKr9u&soDEQg3vHT80wjXS57ixeInc-VBUdSqpK%kwdg{vXy5g8>c~6PLmz*?2v* -h?fR6KLD{CaKl;Sx$-$=<864f={eg_=^=p*DYP9zu)m^ZV!Qg;jO_hE%VPik1u -a{Cp=EC>ru|IIY#P=$^S_;zss92k<5a)}WzSWX>s343c1rAsIL(ZuIR??W6S^FPCn%^d_PM_2ym16sY -Z33G44a{r=UmY-09y2fT_g$G-HZlu!m#SVE8FjDOoDwfGpC#E|(RDuPOuH4+fdqVJ1$GHO% -hT*PWR!7_IdRjPhxbf6>WHj#dGk+}R8&q5{$P#C@T|BBA;0D&CYOcu%WN7hIn{YIG8! -khauu=2lerT-bsWq_Z%deX%=D0)Y~UJ9IB|W1A)#xJmDsiFaYcq<6z*aga)TJ*7}f5plu)O>$I41^)R -EW=ner56J1bHkdSrnhqJauuTOGBOnlx?K)C=cQ(7RKmbWeZJ5wI0BEW8HsNo3AB(p2w6U9#C3=hL($z1}A8H -17m!?%6VnjeJGMiy7*(`q0PUQB3xRTo(lDj#;IL=-mJF^^zBPsE~;h2OBQ`9nyUN5VPs0eJtSDwsP6+QZ&b?%&NZ7z( -U8pFhXm^G1~6c(^;K2VwXL&Ppr!&#^6zNf3^w%9zvP2A=9NM-78+O_p#AK;rlYgqKM+B)G?eWKUXI=e -m_+Q#Ah#4q@CRz?aA+}o^z0P&1_c+jH9o#o(!GgiV$kTBco5R^M}T3+|QESL9o#F-fjraG*|S1oldj2 -dQXrB@(n>rU0S^pi@2yJDT36A#3d4&A^mEpkXE`5xtMwv}4exl6UT!^ -EceOJn;6Cbg?3O3=6@LMo&+t=7*z%HOPt+xknEb>gyeZemj7DS|%}{b{pVB_RC)R$}DJ!a@mSLmnJg5 -Oi05vJqvsZ`bn7P}f@*r*x-;N{0i1|BSlU4x{cQYvlD#%$7=X}@i#H)ENJ1H|&1s{M#x;*o}H~Ce2j) -bhVwWG;S7<6Qw;kXAwV=o<({XkT(h=x|~8M&FpY!}N(c}63!0=yXl4doR0uUd;Ks2@c|8g&Ld5~q6>V -T?#zXmjuid3a>@z<2cn^`8&tc03Jc`ASz@P@23ErfPw@wt>*o>fYJf3nfB3y59nRBq#MG0J3 -Zi`TY)#NG85qWhm7>>C6ACjiZ&h^l5g1&{*)|(*qdRK+ix)0Xm|2vu2FP|SB@KPp-wb9B-tUt-Xl?)wz-m86! -{HVM$!WXzWZSZaGgndAhS&X#p7*mQG7_lO}d1jZRjnexzgcqlUosNyYR~K*FGypUN4DL2Mpi3A;}gIM -c@PWm%&W#tzq`xHB5ukvm7%s-yHEX=5eIBr>B*seiGbV-R)oOwCgOqt1{|0Gp(9p(}vCljK({>_yeUq -t&|o4-od-AllSiQh8)#0COQAcX&-#(CA|T=^J>1C7}22k>ee#f@ISKcNh)GZ<+gYWKdBu3{_(c9h5xy -&VwxXk4g5RP{&Cj8zccM`v(9tA1b#!ynGagdZ?t@i&pf!|qOXjOzMMpfB$9l0;jKO-VnFFF`8L|CwWP -75P~_aXq~TlbHnvinPu}Y&TgUP-XKX%m0&+Xo$N7jMjeX%n{y6$);L6ExGAt=#&t3PVZo65UovW0!Gk -1}8#}}NTexS*g2kkMB$)S2>9_L4IB(kW<3$-r0E}J#;fou693nX@vJ$>k5ki-8+|( -Z|EMNop5C>1va0g5EJ6dOG0Dr&VuzV@CE{&n;m@X_bl=X^L=AR#vT_vd@3?gO3rkyG5}4BmA|ZRAdP< -LbRT*izr7blMV~*Se(Fbr(5AVKRQ{;x0SsdbIDQ-{<#fs`{^9t~wx7j>gb|rr{uKH9Q(xeR&Yx40buw -L!lgRG$Aa!kyk-lzDdnCR_zWOv-An}gRVmGcB+NSPZ&lVMRj;C5{U?%44myc;Gm4l7oDKQk9x6a)4-m -AUd`A5;_zMTQtRFjP!qDndCqYW=dxkWB}vPAoomv&7Bw8= -mF1v0=Er)}8-n11x$pR7M`P2p1b5QrO!G-1Wahu<6xelOq=n9$?Mv)W-gD8j)D1|{J0#i7$LMM -u$2#S9Ongn%7(L+eRGj`sAHe|&ke1p9h+_G;{_??CmZ%72EpMvIwgXAUyt*7Ge2-`=HEqbG8f*Qfdnyu2(0>L%hN|z5(b8(q!^R()ZB_-E^&Ws_V43NM*8hOV-ZVr -bVrNko5-9?Ji&vfhIjk-_a-e5q+eu{@nK* -6~M56kfv<(iTG5n!T75P8q=m4teO#C-#YI^;6wJk~KhtR`d^B)OJ_14FK_rWS4g2C77v8UfuGyz; -aXSjL$v@)Lz5v&x^A2~9uVI=9T1DhY;MiQ;Yg1?;7+tIc=D7`tyOEVqrl1R_G`KXbarUQXH|4st -iBEG0i88xw?C+0ke_G&>KNKk6Hp|NRM7uKQt7v)$Ayz5hQ3_=pO)zu>0i2?)S6=7s+O&7+<|Lkf>eE? -v6=lncOI)TXuXc`|{{q2m>!pyW&>&YGxB>z=o(HY4F@yQnP(bC3Z!0fw}47Fz%K}bpj@ny!o1{_x((a -cbf%I&W%VI_0g%2zDI#P@rcUJ_YZUIEUPSIt$Q&Vf(%78+jOrpc=fO&=h89Ze#-X>2W2G=Tdx)Ttmnd=sJ`1F+jyI`6P7o_s*a71wM~SZIgM%8s5wNOFr%%5a0 -5YHJt72p4(7*spW1q{5?P1ffeu{&*UQQN1KtL>hu)hIBo6r%C5i|U8?xV7K<~>*Fab63s>GIG9G)q-@ -|7Yqmvu}aed#qgF+^rjzVX!x}Nbfo<$Ea%KU5OfgZ+GnY?0Ik68U~?Iqv6n4#qz_&Can0G!*#smst`D -W81fvXRsR%x9wXuIKvV;e*M<=m~5?)Nyvldks4zi92+JoSqf$g3SQ<7Vkrb^+>sC&^zYd@1CPz+C40{ -%+oFgv=@eQr0VA-Zrp{V#o4ORWa$wwAs>-Z1MK=W5iCrx1SO0Ri`}yG{YvxJJuc-}pMw_^Jf794hXT) -y67=5yN~z3?563QSYI3`0;51PL64z||3@Ac}xd0s#pGMFBi*oLSpKCWIV8GZ_+Shi)iq$|R$-W{?*FGg)h7A9Nr&5D9lxaGC3qp -m(Aw1O?PL0S^8Q1`o3+7)D>4URK(OY@DP%|%A>%_Sv>4Kr7Fc^xGMH#kP$`8i0m(Sa3vS4T4ew+)nvExz4n{rGd|2Q!+#f$4%6{lL}gZfgf@Bo -+{wJK#E3il>wrLDF@jtE08@WtzrMbCoT{*8zpAfa+NOy6C1(H547O?eduA~IBPKRafB8WJeW}I@s|aw -qzB^T0r_)s%Bb&dg>~|5g{OK8NpTB#fA1X9py8d@!VEmAv`j`2sYNTS9^A*@p7J*sdL%l^A=JWeG1{* -61UdQ3=Hk{dZt7A6^RNvR^>vc-!ieml#*f5@bw-LqmfV1`h^`m)Qc(BRGn>(^L -iomx;PhSzY%vk^)kRqX`xfpunMmNFScOZY&>~g?PjgiX%VHxc}88sASD!)33jJ1a=h99_zJVhtEO^J{li!H_A!I+V`F`atZiiPnQ --7z2RYBom!)BUof3xQPN)$KSXC^>Y;8&}@htIdvZkTkx{be5AA5Ewf2kOTbiVxBO1a_kbd$ZrX%TD7v -TiOWASNj?B|%O(-UZ)qprvQB_+X(vC)@+W>b!8Nx#Q4*b4zEX5D{iLdZ9XXZx7(kv3Uu8lX_@|veNgM -sh8l__~4=;$8#z4Ewa*+G7J4=!5Wa#ml-o{rxCpB%jqa0Vpg=SI=RtntnCwSs9rAG -81Gd;xx5W`?OrdlTVk*$(Y28a713``gqlN_kxku*iJCuFC}Fj%{|8YcXqHYp@H<3exi9HMQ6#kYWFf* -V(X7=$9faRLoEG*_ZA+rK~jrs3XJ@4VXYlFzc4OTYe^v`>vGO+a@2HuKauTem+oaIzQSjX;M1F8C2hx -L3oR>@tED}#)9WXXYY5Iz$G?rT)>ICIRp#v%v1^hI0RPVyaZ3aJ;{YxGf}(%-oNp-kdHd&okiI6K$ZitA0TM@|mZ$ -qDT_%{k}X`a~I|*9mdIP=jl*~H!ulKAl -wX|OAQwlsbk*%MbjaYX9lsq(rWHSa144y9J!pwPLzj9tefIi&k};skzqn;~0yj$xX)8dW -E9s7ZsB~C8u0}(`i?%p$(*jYsdWrC>erF4?U~Vq*ABrj`Ln@aqDyKppis~+@mdKY7z%eZTze`?0{=`OA6wzO$cMH`H9AJZ~=j5Nd -;{dnKOlMN%jV%g;A5a8uUcu#2KXMmamT2fQZ%0$=R{L#~L@*-b}Q{Lb$@nypMS7JJmFj>MB{%E#V~=h -NiI8Fy70jshvS>lgTj?SLq{F_MI_6-z^~qiMzY#>0?2kQ3S_)8jtA_%UwF?DWWYSdk-M+0VPAFKgxK& -c!YU@JJ<(Sg| -NY3x08fhFkIz5i7i^p@q+}u?7A>XrO^&D?2dfqy{Pe{SJC+=%j5Mx!k2P{sFN)${aT#^U^e%bj{Ifz?l^s+1%V?*d-Jm -r2YPeZTcy1Mji>1qpq7=O@eD)R7-oY!n8etVFR?y1I_d^-BeEnL<*S4W-8pChI)q?!E4cu#f6uDc%&s -QTM!$od)xDUB9^Cz6|{Lf|NVQGQjO$D@ggtv-W0}0GgZGI2Rn%vooEFL5g$tmekQddDdI!&-wyeXEXn -rCBc!?*`^65Z}Q?|FrLA0D$Dxa(F2p`Th_$tb;m4i`gtgsV9a$Ra6*(O7O5q?nCqIapYg~0zijXSo2F -Xi@6$=y6bDeuB*~DVks`ZdN##2C6)P$scJyYuxcS;$QYz1fCu+3NpzS+6Vd9x}FoE@jkXHq2A{ii2cs9K2I -r{k7YN06`q5v$B&SM!p9Z+I1>ER8(qIEzOY7*SVU{T*3>LBzT>mfjd+XjAhS9s^IJOeZ7SuDYp;pZpNA3hlOjPfr#KSMO!iuaC#=aSUe|D=VpuUo_;rK#>#pnATI3;+KYOx4kZ>hClh+fjRoJ+9j|Sq+f)h}U<~99v6x?j|5md8U1~$4z -cx(%b>u*^?##CEgfb{^TaWtg4-DLezvAoF=e@OM6EkJ*|zzD=GGsf4CJT%peC*d2nn=m`@-obC@YaQ{ -YDg*Fanc?_cA=Go1>;eli!P?~ALEF;xlox{g%K5%nF!IwFip}%J8OPPDwPf=yP -ACJu2dTU5V`rrp6BM4I4)r9qh%2r2}BDCiISaa9$j{R4m@(yG&)Meuw?WqAibw;ObdK9Lw)bbIhH~`6 -OR*F^Do)UgWR?CXaj>*Ozs^RwwdRg9!1IL?NV=!8Q<;84(K4r=IsZ(J(F0BBe2Ech_?o++J`mi4{O_p -%Y3y(x0T3D|bx#JVJ37jotNmBI$huLiK*NWTeSjF^cF5g;SRkC|}#K0iQWz0c1UD4)T)NB^0=8*t1IP -X{(6}#*{{P(B!JAjk0JbxL4_U|Dz1HwOQs`(Khfs3(}VnftJ_rALVIYr49R+1i;_u{J^htWTnlpFn0V -t!x9U*bNeh&4F(DgoSZcJJwV2=;Iv{w-?3>9D865PV-Y77ycZtVTXFM*`kU5>$B=OB2Lj*$asx3cMtQeMC*?q -z7)&mDTxn4>+1^eK)DFA7G{Vv_t2jvkI!yPX8+DK+D*o>W2R_j0Th#hrM#8^&;?G$4$4Nh;Aw_JlaC} -Q6Lm&`?Fbste7=kF2gis8p2oj?{74=ul%Fs<$*w6;ucF9=b5yv*+0Hro~+@K~rMA=)J_&hp}gd4eLOgn@4TVsgW`6$Ybd+g4+R)U6u!T~dL8n -|1KC9^n^<2%YDL3d#T@_D83Dge>F>;F%LMzT^8kO52{vj -L;orrV8ydk2Gqr9t3TIfI3Ugn*e6Cc0(m3~H+NWltDu}Q^-m^kXt7Gv+D4@AsY?Y8Qx5>ATQa -gchg0%aq{d+*U}>ZJA(OB3K@P6Oaeprlx&|7VTxQ>y=Z6W;mMXf2r{NyPEuw;u%$8vbyA85XI={G8EG -Nx(wG^<&5_L(qwF#uuQ-)q}z`5L>c~`tiJ)?Gyb~bzHUD9k*E)8(z=kIKB5fjHeG8FS5{>(c@?|s->n -Fh{=JaTRAOGxS!kxt1R=0=drA{{0vyxlh*v?*x;}I2bSmduWJam#YOE}A6XLq#l*6m~|(x~J4DAYR&Z`R&B$u#;`tgEBM6mmWjQ3gd}(%uulOK$G4dEim5AxW$ -z&M`XRzMFK^of3g2SkQFy?^t!!}xxwTKUx~18H_#I%bC&2NRh_=FY_7TR;4k5B-Wv`H)zCBwA+UgNEU -oqOk0R;8lGU6Su-|8Ky?JP3hARELsw-Z{&iQdJRf4PqYFKr&t^u1uwuseQz`bvY+!%C@8h_0jN{3&3v -43Pg)#b|=^4+VDDf5SJ^04dD|&GS1 -r@Iw`(ve3=+t_tdjlq|vu>}A%l6}7n%#t|EHb~KuVhqzRQa%sYMF -@A09&b3Tr%1gln=$#OAy^zHR+GgsbNU`l-R!ax<-gnt7bQe-!_V7t!gL%@8yvp{Z;z@YSSvonZ|c43OrHAFy8`^UncCWUjy_Y3z -Gj1;`!c=m{RmHD?17X|=_31pmFN=2>nqpWE4R_>9x3H!tY8IM>vE|uP#mt-h;E+p6L06qDqEu=teQR6*ez7n+?WUPVx;W!XOHJmC&Z#WNB80wUYZAj)a3M`6i)q7^;+SgbXGoBc)Cl{8$x(% -o^YH4niby&E#Wab3!EeLlQ9pAC@9Wxgvz>;LONcfIL^{asDn93aUm#w-?RBJmj7Lw=Lw>pH>n6!G8d^ -c8%pqhk~o-p6H%Ryhgb|5hRg>>+*$jP(pe;&L4^+A5Y?|(ADj~I?Z7(v1$LgEBS;y4B42)W`S1w$A~5 -gh_Z9CMZ+LrWXHJ0AAsC?Lom?@E?3{?TEeiLs&sNOtz#pINKuh$a|ZE>^=HQIJWDB?=1- -4MP2JWU}&2Mgl%bc>zr0{v7Ju*6@;_p2I2W1WmFQFj<>5!#{%Qnw3uFDJvdhMk;(J*Rj%~&#>f|zDSq -9#K^~<7J*;yF{xI!D38B>T{6Su&OJ8_TF$kSUVT0j8O9{OvVwmhl -GHK*;bGnVzJ<<>d}0QSp{3mgHM;uDfDEIf6xe}EGc9?N+V~DhkH^vBe13q-mtP4tPyyo1JW_07JkGc* -iReCJ$&q!NsCS2xinZmmz{<|+27)-*%=22kQRkN$Bq$E!z+$ds8z48s$+AW@$ZwNy5ebK(xUV>BPrd_546hmgewYkecau+E07tLD0V3&9*bqPqN7jG$M -=gVZ5Y0BbtO(tsYZk-;)s#xs-PHL?Y`nbgd%uHHjGZSQ2tiCMT}*T?ftyVcgyULW -S{TQ60IgNJwV7-bc$k|ixHxmhXRFL4lxdHrgw`}eI!t!n@YD1_TCTXH`a2sc&h>20(8+fbn>o@T^%zL --R$IQLbvQH_<+9iha+O!4rVo{t5HRxO)5m+I{-u3w(BqS&YD{9E5o+T}`!4_EG;gqe-};#Z>pd<})I4uaW*9)v}R*~bfSepP^YdYa^~=B16UKZ|?YZ2`X*_c%E`OA7P|7f -o`mc|i=16LGuf{FN{T$xp>HV*(T~SgH0EES$x2@j8iKJ(7Op)fH+zN{T{Ct=yRm(a6h03~A&K*kYbyI -&ZpVmc&;(1Gy2Ob%hSVm$WPH_6)jcD)o_;i;J!=qWu_;Ie5Ifd{z$7b1eUeN?BOg6aC -zVGsz&n855!CoW6S5;xGUc@cLTl^)X({-ly#*8#5I2JGT;waIQqh$rc1^^-8yfX9CNxbc=igb9t&bc; -r9yaW-a)u}-;SN`tCPeafWZmbnLVf=gH|GW8WhR0FWyqU*?j9d{Qk~AaDm2>eavn@EuNd=s%;bCZnmV -XQ&K>U2QTAQji>4`$>IwKe`847a*%|iLc(7(L=F;puT!Csj3}N^Q2g`uW(E5XKXUv#jT;`=kr? -T-zl-CXPu?n2I|vMZX|@_hsivH&6%Z=CB{=f`S1KKc<~wy4udnwMcGU&qk5ustNn6k>CQRKIA^pU%*_ -9P@=Cu0k6FSsY><+o+wbmpB~bfp6Q*^Yln+cd7{Ij*N8$)y?zFSkS8{A*tbv$)|2HZ?OLGd;*eJFg@O -1a6!K@lq^yY$6@lPNh8h>Nk_;Jymi)3j7u{Jfou(VMn06+as!{6$-*b#0UFndjp#e%Ol>u~TG<86*^H -%ee?FbEJ$QyCgOOa`(g{nYi(6=Prwnv=Cq_T+jS28(iR1$w9BXMzG5C^XyGs+Zp2zNQPcXVXcXn|4+X -bd8%`cv9UQFQ^J8jaqyMAH2fBf$+>mzw@8lRqz@;}|*{YP~BuNdqb4*q4|9|003DGUcm5+`8-!61^tF -$_dN45mPAZGZ_JM2Sy>YS0^#wq54K4IROEH42ivw|7t9xQA_sx215b9Q|_#11C0N4otS_6_IW(#QJ-M -VDi21>pDD6Z5b&U_#4QK!41XN)3yj7vSH_jg~@jQ3XN&DZ7G4Tz6bfUyC|b*>epGtpDb==X46ZH9wA@)ry|p0k -$f6mz)0DrV8jU<)-FEwbTiajjtw>>A6sHp8;GTzQs6!_iDEI$Cie2DyQ6*%=%REHA#DIl%k%JHVZQ0KxPFAU2$mW1 -#&2o?>|SSozUeHj95rX7MTxW5dqAnMqL?rNc6hQU7SS|@HR>Ci-#TVYtME)aZw>PuA}DeJ=B>j;|?1< -028w!0+g0@)7_43iF}_1ZZzQ1i%y(=F>eur@&^7!esAs$kIey8v`S-7{|YFw1CQ$W0+_4Elb@Z0de#( -JUvsMY3)gS~yo*SbdEtF=DY{EXIxf7NfVPt?RZ;%8DmmLOyw<8)16ytpbGr#-{8il -2Bm+IrBU-~gk6v4M~0n222Uflp@WkJv^FRUp<~b%yr^y1H3q!8F?WJ_F(~iti;7HfZg#?iy*H9lFxP} -Eie2e4*X}sLy#yrJ3#<>jvtPtPI(6M*-mUzF7O#2Uk0Vutr=%5$2WtC>9_%jlZHg&`2gSnGdfB2=cG`($-F -{mdAV4Hv>ng_}$7fck+OkXTM|2wI3ebLxR=NzbJ&b_NVGm!zFV*|^q4k<7LLgFpZ1Z<#%bI|UO*TFI* -Q|xZ9o2RGDsWYj@7hme_(q5x>>TUt>U>*RR?P -Z%SvK;9~F|O9$b9}l*^-YB-3r_SCF|KWdug!>k -Mb48bb|-#w?t%_dK$B%QL!HP0zwgZm={ARQR6Ip6G++yX1M=MV -BqH!hGg29{D?w$koZr!eiGKJ0zF_HYOsZ3M+5)soh=Ft~h`_nGdLOKG -9}f5a5~jUQluw~ZRDeX-AVDz81&O&fDTp>*?I9V?YQ2iN5nIzQYSo2+>G@1mf3U7Tb2(&h1R+Ij8Q;$k?{EC0N?c8gtAw<`}2_At5KZLha}MU1le-QT!`{zb^j1KI~2 -B{Z1xQmn>N^MEV*#H_L%`l}+Vwb=T7wdnCmK^QXBmFPU{K5%ne|C1G^3A{+Ll-jg_a38*VkosRcAwum -PfTOj$4Ui%IcsJoNl{>`2!JSA>;Wu|9^K2-1FNI6P*(dE|?GP|N3hXz*S(-3O>j$PRv6tD^2-_!#h*- -?zI^v3jp_$g75Kof_avF_qqJK)#WZ4M$Y?A4IK-^H!9gB><6HY`G=Wy7j^#+?f3+avN2irS(&h1iGFZ -&!R^F4<9K3zTHv{(d?f+G^krBXr$8eX+srre81yLC%dABbkmTmzs5IZsSm^{@a;=ltUmeqC5s($HS4jv -kFh)e60E}{ADw`3YDnvgtSLAk(oNi$PDQMx6>pk#o#_VH0ng&;(C~!3*0}T`c`*$4wZP(%eO0es_voYI4h{Ug#m@k@XoJs$!Y8kWz8Avyz|PLV<%7#` -i**(={0a|b3Z2pmvA6*?42s>VlKyx)Z_c063WdYjkjJ$aEGd#xi%wij-4e;Wu%fuvIT;$G)Xjzt91T~ -<95+b(u=8LN6jwylVj0`zoLal2v?Nh%J+gCrPCwj<_e2av)ufd=KBaPcL0YHU#M7#4uG>!%ji3D&u|F -90v%MQIOW@oJ%)z^=ko&wGD_v=HMf0598GlU6fH2>%TQFBL|BDXBpZmPYABLUK2J{EPY}lc8;^UmFaW -LF;MPT{9v)bEVO#8w}`PLw#Iy10xPRW+7ugBBdHE#pDXa|Nx!pzIqsv6_l#M+KyyU}dR$&=9#N-haR* -uS&n0-3GNI5Kf;k(bBLw0k4Mo3Eu+Qj5Ny>?l<;h3@_mRnj_0wU2cQPsRySDh%^7;wgj(qL^Xn78UC^ -v8F;~#>9QV%;&1t(A{YegmesJTz8{9ne8-E4+g9ZJrZPj^B@gMg|rkcwYQm!?q1==2$Ams+bO5nQ|s} -?yOl^pPbM1Rnc}oXw^Sq(9h#UAvZ8lly2Xr$JNqw!2T$eSa|XCJs^jURPyyrme;83O{ -72EG_w~YLI>u`XVRR7Q{ZODY%i^NjhE5h%WwiKikSiqf2i8;`*q4|Lgw|>S!3Yq1^xRFX{h05tKy7sQ -o{49u5Cz&eHL+Ej^8!IjXcMS^MZG}%TTl34=*s>plvk+eBNvDwLmCCT|9g-~oTpLF->Mzj -p8kxA75IW~=DU^S~}6bGXjN<}$&}FC)sg;xq&&DE-r2za5ay=4P?wO$Sk(KNQB`F))hr<*`cOq5JaVL -4zZC-?e@dQ`l>0C`(g#wvg{0qI8X_*01%a@FoF{vBAR(Va>omRjD-q;-lpfd)pLY~`+g*hSMVEf}xa5NDR98Io`x8OyGWRUs=kD=f^ -0_(5kZPn;zATwzJp+$U$P~6CVn~8giO#_rrsY#0!Ez(9VbeZx6}eLlb}x=hOk7xBnK6jh#vbo`KP=EF-ze=YPieEA*>!{x66mJU$&)AJy8f> -7BU`X;7|B|@~bX&{BkE}&-|b<3noENq -l#aBKiKM7smL>|#)uX!%&qQ;LFY;)^w*fNt&*RFD4P!# ->lWM3Gz>T|xErP(O_zu4GU4m845un}+3-aFgV7uFn)Ka}CI8C)m$O`pw)-Ou4nf!zoxeQ+oD=*6`TzL -EUlIDRU-)C*34#(diqSNMvLwwA#OHG2oB?w-IRVfzP5>DN0VOyP{sIeSb{lqKfiK63%JDCTQCWc6E99 --pL$IQ-kg4bT}rsswIdL5x -#N;TT+8r}nI>ci*rU0GZnU_t!YA5I6l+>@AR*Lw6i2U04$JF&#vvAp8d`uAlwCQuT$5q+e)yxZN38GA -pJK*5ZxB;D7s@$NFBnqi&@<-UCTvDYJ*nbodqK1;Xe(lg;fxym_>Br0~05uM7Hz8(&4xhqcMAd|KZt+owD(L|4>M|JJLn@BnK@=(&Dn%NK@>lCkny6 -_iaYD5PA6L37@ZB@Xt@*k&oPq56!V_h)kT@x1)JXo~R=nO1XVBdOskM__H_kAJ2t8d2;*#+WWmEa!zs -L?{O2p?DZ!v`SOcpG$s0{Q8Ms*{z2kJ4YA407|-T_Pw;y;vAYTB>d@T3ja{^y8^R{V>2A<3=E;;$V=N -vjAu{_U1zO~Rk=gzIl3&+Xw-a#daeQ`9TOsm(Ej6=jcnV^ER(A)prHORNQ -qtMFejyT&MQ*;hK=2gcKqvEb*H=~@`0vFXngNa%leh}TcO=+Y9=oZuRO3s9?bpmfC!iLl -JPGqKXmz9-$_yldu(qhicdIfM9~R^9UDx>v`JVH?ClwEb4;mA7CYF2AK!YbHS7`xd-ohK*H?~LZbuo{=CIA)JYZIHKI6(mi#=SRLGMne -vFnvUwX0U6SnfLMjC)iSNxsPBrxgnDs$d1-07<^gK+!en=DxXfr#%uY5wrZ^>bhN>1{HXJ@M7<`O5_C -_89c-!HPG(GnY2@bNXd)b^dVk`VQs{hj-Q&yFhsaeG%vCKfHzNJGl!=GOGR;uOYbq5g_lmTh2D_tg&MkfEGop*De8>U7q2 -XN^0^EyR4AD#93%~C?pI@GGx#2MIB{!h_nPx#0>6`?32b!1dr_sF%b}RhGL1eBq&OAtThl=LF}xpkvG -EqJCo;tJi=W`VVtcsa7KbPKHr`YMvjY9vc`SEWK?A-DIz_tostXcgNg>-Nio+nROi#HFH`re3XhHe8@ -w6Q$lB@b4ghd+bdAfs7(gA*|`Z@zay}tP1@yhicf?i}UmzesjxNGdFOhU&6Qas9=ilf^@2(F8D14VMh@|0-l3{8$y0JTkVbuu^`6 -<^Ouri@!l`m)bZlc;%C(3N3dN1>dC-SX~zn>gOi>&D7hXD6mhY$pD!%Lml2{r@V=W!}~2~W?dSyL(${t$1J?33r}-nq{9^BfMA-v(!u_A&qOnAmNB_2 -P$07TZ^M(iWbHizKFKS(sNvo^?*HZd2lq1({EMf3g_=K}@`K+3N|O|hP$a=3IKyHDO0p!01KS}Mr1-3 -0^J#)lL;(Z=CwrT$YUN|(HYx%P{BdBZ%p%|bafKd={x|)(G7RXOw_tGK{Im6JSH4AT!2}@TeDgk7x!T -IoKpG{viOq6w2#T`502<&jiGaaXCIhFYpb<=h+{_jDQg(BnjDc1f;9$JFhfEBB$4>~!oO99F@P(u%^x1hhJXiUsGjLv`Q8NvI=4;!CuP_maUw_jtPGR9W$&51`c)ZbIR0I7Y -hF?%bKP@(z!X!s)f5y4b@8 -uZuG^+=;l88vg+ei5j0=ZLZT6-b*jDhqF4f<>NIu>P+p=ES|wB=`P8`9!}OZyWI!x@K9P!<8+~p4$zz -yO?^;8=dd4-W2Yx0RyC({R7ASMD&tdYGcC~ZfSTTK`+Q&G29LA0m(I{LI_Qmg91e>s)t+?E+p};KMClUWPbZ2iZr -Y)|?O%)AzNM4zM2PJ*b}XRo30Yo)EkbX>XQs0lA9l88_;%~)(~C%);W$*r?KZmyH>qgLvyr!EZ*>?PC4NGfA4GesVT5yONY+*ndF>Tjxkoeo}XC%dwI8UhTdB~O?Ru;p?g1nB8GUn^d8C|u5B~tN*z^IF7!11~*<{h@fTKFoG=~ -1*y3uf~ZdDh^SXq5O4eMTakJ$Wy#nWzZpeRh++&9f?)sVa-9%F|aGSVYBCL;){i^%DvkmW$5!I&MLTy -Y-4AhrhYPN}wgh+W_#iJFS8Lh^Kg@}P~I1z><{W5twXs+){6n(^g4d?61d={sc&Jhmb|XLu3)V<*PBD8zz3U -cWV1q?mzH1ADV>r>lF+6)=IYavMIUF52N4zsOh5JW6-R*}6#u` -5ZdS)Ipl{M5J@@6falGXph{>zbfQPeJPZFDiwEak^x5&yNp$3cr{2vmdnk)3fYMkh61f8$Pmv#zg&>f -L}*ylI~ul#YyPxO&Jwjo3ETW!k_&acU1x(CaH=sb_ucOJ&Zt{%JOmf=->(UzG^X3LVjN|G#1Do;2voGR~m6S^78PEvN0Z;(+QVv8f23~BmM%k9%(k?mQtFI031rigW9{BpQh6D6 -J}K9wJ@!!dqy#v?ys0ybzd|RRm%EeUvYI`|b1S?r@K!fjFTs{q -odeyyVOdvRhNC_eZ~~<^$(;;r+zQZkyZZEa26vtnAhYO6%(OY1rAw^^L*fq8+kRjykuo<6x+bOvY7=b -!&AB)SL*7p?&^NT^*~vw+tmDV&pc}E3dCg6~u{&-S;?iM8lvzgm1YXAp3pct-vb$%Ubuf=gaffhyHQ* -#5?g~q}zDD#S&{dZ+q;tI7@#rO$Huu~b!- -Yi$#jLl1s86&LeQkyhWig{b=}<=2Z-F&G#;nbZEGK31Upba&UKBdKEDF2P*?WF?}+$f*qF&9~t -dJJS|AVgLbv@Nv)pt=QU~FsAtOb@+6AB6NNjLk7^Sm|uD$KgS*YLVm~2m!g{C`i8LcOOy2@UhL}&<>( -@`r(#LnoBV-rXR`CP_MJc2&W_!OfMB?AnHQBGY!BuU`pF*eM+3c0>bhgde%KQor$D=-8Kf7E^vGOx98V&GX`;d_7Du1ey3?DS)~G(>_0VUcHOI7;@ -xqJNJ@AA}BMOaOCEpcAoQ^t;VFx*T{+~e3Cdm<~>s^9HHg@~uNf7J_4l?(d&u9aid=hyW|?KiOQkLS8 -sQnp>)71wO|6hZ$RFZ<2$@^>%!w{Qxau~=}Rolx8STDChmG?O_hyMpq~ -y724Wy12WQ(1=+{J>E__A1d^?9<@0MX`qs&8?6Z=JBI6}S3t5V7NR8a&~Xl -v0@<{bam@9j>pidD5$}Q~#=9=soy-Dijgyux>fL`~B&{x~`ntbyAuVf|#)8fr;<6#H&XSbU%TJGEe`W -&he&U_07y7JGqL6Y3cckPzon7fCbt1@laZdcUSFVMUu2Ah5K4Dt-Yi2W;IFgiS606c2DD)Na4DG-I=!X&=rY0omx;@cUf8!$aJ^{{=LzM9 -lyJq7>kSQoUeSFkeCxJKOc_2oSM)*bx*4brWzal@Je_i0x4naX``+3QfoIHaUe&UncykE9rwcTsNMe@ -L-}(DK$%Rlgj>wzce8?a9j13(LiId4v;c0Q&+(Q!1$kGyll7{#8{L$!P`a&`7J+wu?zNg(>n4)+USt -1UUV8+uPws%(tehxvQZT2sv$!)Z1&4GbJ8zQ6?l$vc!5{MiV%nYeDEggUA~+qobtv5&RY?eBQQ?Csp> -=ztA{CzB43W>-LbVEg%zI_#-`HC9>~!dE?Cg5-_1B@h=$vl9{1YKK(u}YYtE9JJ(ChM;aylm%%#6PbA -vj6GGIB+pvbkuWP-EUNRf|06J0B*POIqzI(;@{83-6gPH~-=Ew5L7~WzZu33Ap{L`ToC!yZ?RH|1$vp -?zR6GiBWoslp?_KG`(dNtT46`NdTKH(4DXJFxy1lIS63*+`gU+ -~&TNd60cRx?aC`X$5|b3T)>;rCTlB(8X&D;u)`<9bi~~F-H&g3uo0eZcf`I3bLl*?x48=iVu_7_X0>B -43&fA<8o^9sYe~-j8IBFlWzeVEalW2qg?~wSgaO77cUfIYGNc^4X`k$jR^jEC+8!CszFHsr#E7tovs0 -@7%%b%n20A__hPLumstEX{w+h<{HlV0q?Yuph8TU)h)lYZN|QrNgW&J7Xc4Vviwr=r9z=+ -+FYpzYrJphF$b)e6A<1Fw|FY-eKkLR@+3}MyJtf?qZo_$BSX3hjX19I;>~@v{&?XMb9Cr-Wc#XbU!p>u@>sZ-j2qZ -$UziheVqp@$^L1<+M?;gF`!mm=@>FWM93HymK0e~}(-J63v=4;IvI<2@edV0PhM{=G^JViMP9;kS;;k -0TI2;msyg^hW@x9r4z6C1=lU#>uY_@3aMcR&eUAxyTc}eQ=g+zFQ?>0t0c|WNI{*#PY(&n7A;0MH|$) -qn=wd>2#xi9a1DE}#b@_V&_^@#Q<)~1KK$-}p)Lh(eF?;P9X((Yn)iiPnbZ-i`tCrK2okQ8@uc@>Vwm -CK3?tRlnYq{_SXU%N7UTW216Wt0;y8p7OV!Jsk0ZXcuGCq?dIOjdi&RGuHNJbly-Ii%X|s2J{hw&ki! -`Sf{KTUK+smHY1;V;{k@I}b(m(E+5Dp6@%Ii^#bz4==TUH9TK_b%y -frWGQgcPjgm;STlP$4wxSkgnGIiWyFi?bY!1J_{{`JTP28^CVR*(d8ZFyWonOhk1 -hWKp!!*2x&EY8|@v_%NN&;t>__*k1LUSTzIDzENF19)NeyHm*>~M`I^1ZG+Cdl0Jg&9cc1tPb$|0EUx -D=Vr~fdzMIj7Mk}OTL6oKLlLw`EI#cUnbI0C}2Y?EIzTRvp8$*fl%3VfUiFqxc!!K=^bnm6;dnA*CyB -!Emd0aOzhS<{>F8v!0M84G&0E4~5;H4LAe!h6!7TOtqycKaAZM{v>@1=8?=YQPzK58T)gQ@l)8y@A^F0czhK+g9@D4CQ~g66o -tv{$8bve?sP`^t6?aU{3dWq;%_I-Fh)%~xxq3V@W*>T&bO7BN;$QEq_OzakB~A`vK -(cP(ZR^`gz3`*maOybZNx)zsq**WkNPp`t;FBm?Vj=|v&vvcQ$sVZAIznP;s}wLWA6(L9FaM1-_?OWQ -C=0`cM2BG-SYs4HEb4Bdg-nS?7?&`(V4|bv*M*kZv7%=YOghq*p;$pFL3bl=2LurU>oe__bqx*<>THu -UnqVuG;4<|Cz$S-^CeB@^Sh=T6nZOn67%$x^RmfqAR3pln!*iUC*$rZ>cIuO27%NHn|78k2v1z_eZaWi8 -rK_iBL~XO3Ple+*x|im%FrhIVvCMQ^60RRY9+KAubmoTIu`2ZpG{4jL6y~)?X1KOUC<${=-tejwqx9N+p7R6q`$v`E1Avu{Q`g_t4fMAm<}Y6I8N~eY^dBIGC25?&7>ve%a*d -|RPuCI&M>pu4(RY_25j!<#cjH* -<~@5UYeiU0zx?2Fzc(d*0B*RM=0#{qVbDUj2#AhZC0+=@Hv&6o`+G;61;UmN -Cx5!fTackPA<{E7eoIHG5Ma^A5M$cILK!yUH|E|nBr|oKSRv+*ndC7{J4^T0x_fN -n&zB5Bl0vvnB;VaXSLvs=NPjUsM+Jjh#l|Nx{qE><)K -!Jb$axUPBr%a83$+Mbv-q{b21=%@60lrms5A<@RHjdGmKcv=UGLM$qNh4G24>1B3pTk`KU?3okX-MOS -AWpYs7-PL)0?MjEEMmg|(4!YHVN5Bys(c>el2_MO?zj2)D0i#r4%Ah%`AFc{oM)R-+t=HkDV5!ZB9Ud -FzG8)=_%rFVmEI43YiM4F&p0(G-=^(=}$(Ni_-}mar234ti2WtlH)D+J<=Z4nIELkzh7_ -nYj$CZ7eEHS$KLeL9&;1@*XokgUfGG%#6TqAa{j~Q7vfom$_X4>v1;ig11H>Xk2K*-D&7cWqz{n`G -@{rHTOUi6q2M5F?L%`VV$`WD}2&z}o1N^0PFakquCi*LmVH?d^?}Pn{ywKagD!#rf2Lr2A4A=*Z?c#i+gzuBI_GcI&nM6bdQ;z!CB97Fsh1n!N&J -i~CO{V3{l~~MO>*&dKezsbEOyfPV|4b%KcR}tEg!_i{sUD$9{aWm==V|O^Ho4!Q00$R{KCHSGxy4wH* -)CV?_fK{Q#OQkViygBI}=U6To{F=I?;UKDpCX#F~)G_Vs1D^O?9klxhnm!<#o=6G?`6x)!=J1zc|9p9 -iK9M^YFa1XIIv`Ow+YvwG*zT{W4gg-XV|Eg}#Q?J9=~(*@qBZRoO{Od@{e9Z>v^`0tvJk$=>Qg(&CJI -)IPs(=GA#)r)q_?Ys5A)-0pPv#`IS2Km&6uBkirhO2_#ljrgHI!1Vow2fBezU{4&cix5Q({d`oRxu=dv3o3*%7n?No9x>WYagBT8JCrTFav9;4s58f0z1${3FUUL)yIdOO?Dy%x6Am{@?~U$x9=#g= -Qx|3meZ2vwI;pmY$TPD;u=Z~M>j3)K;#o4sOQ8h)vWrr+xiWda89FCao`ssU>oYJDW*u+q&a?zh7Oa -2)HkEx4(c<4dx>uK;Ysgc>hb7cYr%%L(Z}Y;`J^9ZH5$HtXD_BY-Wt{m`ukdV?AI#j6wTG|%d^`br_W -)Aft^UFI!sKcdd7sqQNR1zKD5Gv7Y7|R$A%3(iAKJ`*Z6HS)E$2iUCtAySI=1QU48A`1>X^SERw -oTU~WXw`1~07@8|wjKSzeWn-cVd(S>sGQ|h*7^6kdhp3H3m&Jf-vU#LGhT75QSC!ueT3rAii1>73zFfpi39j3x7ij8D4t6ui70?J38)jY2A*G`hA1DG3g*b9&o41_NM$99iY=Fr;^nx5^PI#x9l~;iOKtqRgNwp@B5{al)~mN%Q3_{N|?7Z% -h|6d1o+89bOKOTbuiM1TSZSaZbSd=w~(ex>w{Q!q&e5#@p95(-F`jVO!kqy8gA~qP -L);oY|((vMr8&i!o#q5Dvl+o0}yDiYV-F0_A^5lbN=LWnMrR-8~rO{3nWmn?y?v_hEh3ts! -jM}-cPL%nS83||j&G`x&k)%1KRX8K}V%z=f50U5|!ewCkL^%Rk(fmeH%eIX)s>RicxN2Hg!WhvC>%Cw -gEiMWro{{AD<=R4~NC+godHSv4Ya>A3ziDL93IW(e!qxj(w_SM3PcR3N@*wjcN -X%4a+Ua~XVx!J&noAgKR{6z7qgHeq*%J>!Ddyo(!Sk@%UUOPR^Bl4$oU6l-CL96c|EDP+36g -Mm_^yiV9wVufhuDMoyriu;ia0{9so?-%kiWCV7A$BSaclWu1JC~z(7$92cp_Kz(6+OPXa0{7bMl8xOD -%8YF)6?h|jdEsy5Cj4K$%lE-TpWfkzAO)JGiBEkO)}1cfWSJn3XNw{I&}0%6pm)r=kC0%;L2{7CM*qT -%klqXeG4K&XfbwyDDX=AE!B&WD;Y1O(c@;+B!RSMiN&TDsl`IpCCAuZCfjx1vJ0c>mOENgPO}Y*B0s{#e2sMErAp$&?{;t*d1Y(fK;p^Tf@_YBw++$PuIcb15A=6PO^_ -!x<_T#3e!t?31^33{zmCqQPUm;clnTBiJ_D|DECe8lpj6jMI&@H3^@^=Ah^GeZ*uVN6{+q52Fm-@ozW -dEiDw>>3&CC%_%WO=b}r0N@J9IeH~zebjSB$j_X1_c!l|1{NWyW;+8t`++iPlnJB$a~iKNc{ejT%~}# -H_02-ANi*O$Ug;dM*&;<--IWCAqDlT{Bo|vuV0Uk9r;v;-!(vik;P{XQ0P~Isl;w%J6xxxm#9fHDam=#E8)aX@Z;^muB>xN+ -QY!&6?t^uS(5iW~>Yu@|5bg_NA_$RfzmxkNGYb&hQYl^nOI*-h^(RhP%R9yNTrs+leMj5Lu1hU^L(lW -Z`wRSCf0>NjcW=g3pF~NP1Xd3nY03h`rtg;*#Tlm-|sdk^Cu%r*qk~QA#N0yV5=FOx{t~@YT>hMJ?(4 -*Z|!?1^N@J{1~FDrUP#9JLbj?XN!)X_UoYUEhI$QFAY%WvkvH|4NyJmddE0$F4lBA_KNJU=@lj4(4dr -if}N6nYFQPjsF&04pT4&rPeN5Wc_t8Ag&gQi6jRdnhwIP_k?p^mNYJu|#`mvW9^)_M}vCXPwF!UuNf~7|qR0oJ3b_(K<4V>_jzXqs -X{W=ednLjf|>r#LNDr@d5L=Kz?T}$-$B3QJw9$d%doR8?@oeeU^v9CU(JOUOQfg!@4**!ywLv{m`#iA -i*5z8?y=__@!-EipM`@znDNyB=KaGov<^w@Cp;K4og1vk -_E+gXE9|+P8gSPC0ubo^?=Rj#KcVJ-;WA%g@oz8jgH_W@M2qaT{EBX`A)xiWTHQQMs-vLjmxi+i_r&0i7 -d*fMLY7`d>*4NIfM%hjayZ;30tm0u!@fsbH2N1L_iT(`^DDBzmJJ!29rTkud|v*c|?lNq!y|lz);Ik7Wya4{cN?+ua^6BzTdCn{Z?QTro7dOJ9qHN$GU;J*OOC9TuG%&5ZGQud#AmH!!P#CJ<<4D -I%mW>t4CN`EA~vZTTnPI7lAjmK10eb2+8=ldN4)4K)g&zN&5v{+9~v?C9Zjh%{fY*6NF#a^AYY` -G_A`1n$HX#a-2<8>)4y6Bl*0bknn+YjNv?1Xt^6ftonPVx8+v#zThNem>B -Y0`Ytae2CR8@i)G|w+}3iN3fx8>usi>|zV>#50hF;Z(BQMm3p+Z4{akO`2v@3q^GeUQ|J -%ZXZO#>AcyVG8#oR}z{O`Xtvi7yE`C=@0_by>1v?|FxdJL`wr*pVcywD)(=9hNNEy~g8W3lIO|B5@ -95C>iAXUaj?~$Sia&zFH_6?#d;xDA+ozKRhs@s%~l3b}%Dib|CHxZtY~jTXNd!cc^3uUDCcUC2}p6_j -%55A#_g)w`V-rT^{b)!~0Td*NVy@Tq?}@0CXfBa9w6 -*p$)WXW^GAfO_AwqJn-U=7IA`5|cKzkIE4V*r2o#J`pNQ_0p6NP2?=~#{aHI!TOPk%_%KM_NDYm}B6#;YVZ?P-|SXRS-Bl+jkpTI%)#C;8!g87 -vSB_867F7lP2*oBAJfPalE!xP;%F}Tuc|1%qcJj}1s^M@MU#@I1F`_J#dVMCRbw0$=>=e-Xuhnga{9Q -+PM(5gQ_IGulL#5pzh!=Qm8SrSOPUb*NS0*yFTvPi(G;+3`82=iPSsCW9oo_27R+Nv+OzmYQ#krxLjx -{}%o`pVum!94)L-yC`S(9?51bnx}r>kriIH^MBTPG8-{@WLDOP7hwzfYof&gzCFHzr;!?7aS)O)}d(Z -oK5BE$vw2hHJsjI7`iT&L-A)hWeKIGe3aOB1cYzHG%ZR -JqORw7U*DY?J$IVopZgM8j@#WN0W$x>@NNJ@pI6z(1=Cx5y%DdQIcQbcx4&qTYAM(BK7nTg8>czx!d{ -%a-zCVnIqP$SOkeJ?19#O}cplGt}$w3%tME91EL}@!nGDe3xYucMIsO&k{ -kc8(Qxx%cf_n4tu=c%COGA>m`A`_b&B((;m(X=R2*yJ@W4-R&gDmS9WlrxBlyurGR=u;xa69Vz>bbkz -T4p%!x_7e-hJ!J0Yo{m-_ip)ELY5W&vCw^L@mylr@ti>2t;l5;KHT8f443(H@#u=@H6MMRZ+Jey{CT- -5J}uwr_u0h?ip_3M;da1!G)*o|$w1={tFE0>^s)}?Q=?c%{=vo1Sobe$)xhD)+Zq^;dW`K!-Z$?PUAa -E3Jx(Ozz&BBPL#THX*utn;e%I$#6ZlcPqt%{lynJW*o$x4J<61wJ#A!~i$U(9KW1E+MRwqtTPst8)s` -kWtH{(uU@FMclO7E6)x9@U2Wi@!<`VPT={{SW#qxw2*bSG=a@A2ZJ`S|q-&VSthgZqyivi@o@B!;jQ{ -qwQRf9Xo!4r~7MD&ON5h0!SRYNs&*TR$Qs!hDKo?B?dWMROq=m0G#a3OxA-_u1yr;_b*~rEk&aKi*#{i!?Eh_NS_&sF%*@{^Fo@z@ -q9A-sp)V_ABF_|)&xQ~3)HhS=%_cz-1V$s1Q@g3QH44=W&<;tjKwKj4m=rpyj39@V>A$jkuzKjfvV5?5=$}Yu?o7PDJI&2F>}uHzCQr9skmo*gq -3*TmRuICVCLLeyZgHSIJ#L4e&`z1IBD!uLS6w73G&;uNO}O2;0QDI)R@*Z~L~lkl^_ftu)3pJPB`box -cic&#D)H%@7QcGp$CskqdTI?Rj2 -2%I;I?Yx%V;D&#oF!ig`V~*6(V}blQ*?QK~4D+jmRj@OL;w=$%X5Si9w0xbBkdD4(%zF4nKw#fC`Kh9 -BY`)Alg0Cg)0=4B|$f6xDdaC`Y@I(WbNV*cVUB^_4co$i-obEH?UxIX*E7iZiC(=lWT`X(fw1&XoF`_ -gqee7x}#c_K{tdoqT-BjQ=o#Oci?N=^%T;bJ1F2i#daI37#mCl@|mn#ehNp6kf0Z3$}Mw6m(0%5Ig!Ql!%rp8iuis-;qe-3=tyWmF_E_7Pl!kMhy>H -uoE%|;7~&Sn*V4>m0?v}SE;w?Ir|i$B(KnfVU|2&RoA-0&f_t)(C`xz>9^O91=W$p09ft3-8x2)QeV% -+Qy>p|fv&~yR255_Xt1T*XErx%nEhOV|6??VC%Z6TQHk2*pRGT+}t77|?cE2C$*A=XbN$NPy8?WkNxa -OH0kI*%@=)?a@-g`AYif!A1@BE7Mt~f2c*>89wyoWbDNJv5$VSass(o|*c%w1LYp4$=K9Z^-EgfI&gT -4T()<{U#}z15OgeWZV_k^k$)zMW^%Do`L7X%^YLRy*(^}>*kzb4Fj5R@@`Qda5T^s(g%I5mqVbt`?b8SrT8TaKGLRwR;f0cM99Ggm!Z6%Hj>YQCFawTb8w5lD<6HYUefTz|gQQ5RTE|^uW3JS} -{sngLB4g@*Q^D62!`Mc%kQnMb|Y1LJ-|!Z5KZ`>lK!wWa8n~;7{qPhU7@P-O;y6y>d3eCDwGpKEP7&! -i~d7qVO5tDH%{^Z>bHXC;ZaMB~#H{k$Dl_gMU_WEj1a=w*}tLlcxIt5nxV(-f9f~qQE0~a); -I`}G272glR0!nGUwM99NWY&^2qtU8z8mmc_Oq2E$w4DlCMy~PwycI2&jvf%Ro1zfDdV65)nbJlyZH>2 -dx>=5BSz&QE?DOj}86_yjZ>CsZ`2#;EOH$#)PP-Mw1Jre}mDNKEezB9D+cu%j1yguAOn9EF0v;mJyta -#F_Zsz{#mFU;*ypp8F74Oy6Bke)<|}S$bhAP4jkBh@;;8y0RzdDHJoU*Fyyn~`tlR1ztYI_L+8qvcUw -rD|KasL;pU}>KQzQc62m=31?Edk=-?-dw4*X%g`IlWS_-9EL{y4~C2kY6n&rZ5_9)x{lm#OqfCZOo?0 -{(SE9i@&sW#S`Wkbg4k`&*Rw&_42LZQR*j_NgE)(j#lI(<}%+7JaGTDKAsW$1VUo3TbwxM`TAveCKOe -ei+;p-VgfM< -eLKyDG=YYlx-%`|`_3*Jw6t;t+cq@n*oVM(f^rl_2Mds+@o9?jyDo_zJ9}7C*n-`t-Cl2$y2HB-o{ZL -uns#$}+)gb?{Z3f)yj}*+kk&EPhF&H@te+2%It69D_8g8{|ZnSBej2uNkz@Mss*gS0K!wF9Q>If%u^7 -m%`!#$R6{ynza=r53z~k -gW1@z2GyVBso6l)$N(E7s<%{Dm^=}J9F$4ani8e=7l8Q!ev6VIWA2fP*DReC-XtYcpsS$myBU2m;0Nf -Btw95-B?;njqCPV%TLeXWOzvFwWQ!{n0yjh3Id=DEIQfb_mCx~dL{BGl%7xgUGyKcqQ00cyj!-wJQc+ -*Csngar)8oLErLAcdI -GFiUI4MH{qkKo^3dcW7`Rohxed6Izv!k((_RoJ_O;tsuX3J%zes%uY72*nY;O#Z^qD^G4_WyfP; -tB7*|vzy0VlAcJoW3XDbd|&a)lO1HCZg7{fo_(YKcFQ -h`BM>S1sMngN8h6+X2k8+ia#%-zbTS;dki0b`ZFXw1qRF?&~Nim}un5Qmmz6ijA5;w#`Ql3&Skp?zJdo0=C*WM>`1Y;x!=N2N-j5wK1EMcsGs&*3e1Wx79WO+*s8zjC7 -hwm{u<51E`UgtMaV|ZUFm!{78{EZL3e7mVUc^>=A!+oyLTU-*rC5wrWg|QV*;1RD=z7>VhYERLlp3kr -L6{%6qJ#N}L>tBk@tTUFK$w$#!t9)5b6AW;&!>TgVThUXVjj2tjQ_K(31@iIzE0rpxs$fS8af9KZ^BQ -n!so-y_r><|jT|sXEB+`*R97DMAp5#Mds0@6a#&FPH1pRt#2pH}U8D&Z6?TW}xeC`zLwLS55xO3ohRV -_e1Ob7iEuk0^sI}n4h|z#JT`d7qR?SDcOJ3|>bui_s!!e0yM3=NkOjN%R05(!&+KB8NRG6DOQ -6R?6#=op^FmYvemcsVId$yi>hbxpjo#9YycQ5mWU9@v|rfL8&g~Sc!RR+zezg#Ec_He6q#!{JKZgk5@|C3 -_Q_nGWs&SIkq$?WCXF$G$OO4# -AEo!2?=kHpOly-3MmoXQfuUTJ~@Zp8$Mv34z@xCy~_^x3y1VB!CK`k_JrnsfoN0V&)WLWe=&{1?Y<~V -*v4z56}Q0q5uC^~)FsOrj%HB;r(1rr9Gy;8O1qq}D(2Gt|FUJbA1zn^vypRidl9uy~dT5;;+ -lG);V}G=37Mc=Z(NZkEB+*+FFj7m8cpdfVs8>%|F_an+e?G(r^zHLoi!l1$>?p97iZ{e_QJK{I@n*3^ -E!@&h}l3G&a$5?`QOd2Z7afezFB2Hc9kl$140@y-2HvP84=6|h{7ziA2`-MKe<^)+Za4d~Ueri<{--_ -JmJ*|NTF-Az+h)2uEJs$sBnOyR~GcJTFq%&}H#s8q6FUMIyEhN@&k?3`{|b9wdHD?JGcugwdDvcyDQq -i2QP{L`zUvP-Agp##WbKb%>axo9VGlIh}=&35N*tI+GxquiW|mH3rnq>LBq^{K@yyNqNDArJZK@stGv -lA_L-6(Ptaz2B65^w83aU?N(eQkoSnd61pQA%ZjyqrSIxBCRbauq9htC#aEz1TZiMkplKH#@pi(v|-R -qZgJCQaW=zfsOj$2ld)l6AWU%Sq^cSb&qC1I$ra1!#_|S01W6p>+>Z)Ns>v;5@i~JMZ+ -ALqIOig7xOU@#TPoW+!=Lbx;Y6F{ZWyhdA)fNjHj_@0`c#R24NN`Fw%dtIUWu~LBt9Hwyrqv73IT}0Y -hgN7UhJx6o%XzQy54S~>`GGn*6y*_=AsI5=G*lXtwD`YO9s96SG9>-m55O5e*t|HD -=OvJx~)KDiw5$W);pVQC6J772F>Qx?aPAY2};1u*fc2mQ4i%b`z=9}bW^4a&ZBr)M7x7vw1SB+!2m;E -{f!@~^9e5Oml^@4PIBe#6De{8+u*IbM!@bg0o|DYS$RdWC$H$-vJAj@B{!=p+L_RuMV%>Fh6#9yFpj< -XH9#6CdrCLv1WSbgto@-~A;QkK9`0pDJdA^kKKzgISYMZ)M@= -6*wdu{hjFKW3+>zw9)z3_O3w_emwZQ2G_Ab7l-nnieEHM60Fe|{UP@C7Km2nn+N4LGWU%}{i#plFo33 -4!}Z`Vm806%CbVOz@KZYa>-+mVw+H;E`};e$2mGh|`%ky`t<)6wt6EdqVyQef=crNM%7v(kM>3Qt5-W -Ouj^|AWNM5aQq>=p=Pw%ajyViQ4M$^np+Zpr5Wei=c9|Gev@A=X!K262)e4;VKK-Q3(&w#D2t{I5EJA -nH8O;-w38MZ|7;1>QNyl@D2*puMlq50fS%K=IakUt9wMQy{u7$z -J$|uhF|o+x*LMVLSuq)i3F0goECBJHbs~)ATH>^*Xjy{qtHA_n$PLm&64r%K@s`&-bh#i8y6f}6&;jf -92*V4dN@=77F-{;qXzH~T1$DWdM#2!F%KB*7Dqpo_&CUm0rsAt+{b08nsk09P=yMJmu*fF$9;AI1sdo8yx(3*LS;~mB;ezrkXY8Cav~v>9dln>c{DOpo&^ensWt -~I~{Yp*6IC<6i%_PViARvHO_cvi05H<%zSK}8Cl7bs5X}My;Q~tUumJb$wcd|H~b=Jt%M4~o9NwQ}eg -38bZD9Ach?oK6wS2CPP%Pi79r6gItmFlLEj25f0XlTv5x)tITONPnrIl;zDPd6s{-U0)kdmqcTd8U8Z^Zviwm5lV9*Y5JCo}K%M1QVqcr#5uI)MQfBk@ZEd5{ -bf59L~?fAFv@IkZ#>3%}ae|qdUEd0;M{aA~G;2`o#NX$RYejfwnLobpz+6Sml!yxjJ13R(~pI(E*CG$ -6sNPc7|)1!GA!H*Ra1o?O@eU3g1FVh2%VCq1)@*_2kl}E#03I5I;i8}lq67*=X*vC_X2O#6{XPM~JDu -|(^62pg#-=fb^5kptp~=GbdMvYlGo_TBrBFM@5Y{8=MB| -Zhf>)uRm`Cel@IrGAiI#!}|58eqARD{JK!oSe=2w?GpvvR}(jybKvoTqPxAAQ125wGtZYnbG*Gjsvg~ -xuGc+VF4#Ou^ddZ~{MWhg!s~FkBAdh4mRa?s-3F=xBOK3z`ps83ST3v3Lt$Waq)WK9G$T;^)VE%jX6h -`4i1nu%IHuu9S6h@>m+1u2aO(c;aa0%D?b*&uiH;~6SRM13y>Rd?L`u1+s-%kVOH5x@Kh!wN0L2NCbi -cb0@vgsrA>c0~X%k5V%_%h< -t=~4r8s*wYtFEA?$p5BooMpf-g&rj%OR6SIvL=i4Q^KoKj`38q>x_H8RoL#byMF}^W5>e5Zff}DX;bf -5o3j+>UZLX&bbMd;6jz?VJY?RgePuVtHf{CR{ZZDL&zt6`=aP@gvBBM-?~m{CnGHjY!!DfZ*@6lGD3y -Nb55Ot>wZQAXbQxmJ7S1qF?j)K)4dxP!t1E@%NH)bR1b!iA*c6A>J3^EJP+@D#zcj3t*)dFkx_*8O5V -?i>iNj4PKfeVByJq))FV6$McQ#6=5Z*W$!JVvh1iX?sLtvDyVH`i=Ev8r1PUBChrzv+>!gWpcXRUZNK -$4YPrVo`6r50DRr=Av|TEk8c5I&DEO*~BVA{9!J=*{&Jm#rmoj^0QoP<2p&>OkO$oIQkKBl@+x6L=UD -$6lX7I4OSLVpsJ+j;G!!`|T#)uJG)}7_>xZG4xu3(9SG+0XEiq>6VK|3h(DG#GF^H6tjB3QN?90>CI& -o{Ti+K){Yw?+MXvO_CpAXB-7v-T}i-{gL^VR?_9yUcafdSA-c6_SBIEr#GS0vTdFsQH_1D;A%XLDET* -cns$Hbft?x(#D9?I^PA_|Nacmt}D{Xr*+zB`YZKvl+{+|G9-v8gi>favz*Kqyy(0@bhUpD&hn3my34F -miU1A+5T^obsA8X0&XA^Is}0O8|B`s;$(J&??(j|24{OMVOoA@H!P-m$aXXgU7~{` -Us&Qzgti!4%LAjxAF8i5&6-)Ch-HJ_apbQfz-d?PiqPMDZSVuzp*17m6)b?iJ*!ZiV{oRoPzZ=@`NA^! -F6O_kNbM`znCq%oj#ygz(jli0>@}R1&bc(8b0PVF>sG2caN7?St4lEAxV9Ghp)(dj&D*DaMRDThM6BT -2MH;tsW+~8+*x}=rK<{Ahft_ek>RB!g9BkJ08g=Ru)BH)C<55@|gG%Jc3;;IoBs0nj}Q_pcL4sHvR&h -w1Nz{_$qp;)SfOGj3k5Gp;nL=Y_%x`~L0iLK6H>$n9A3VIfuhtsJ;+c+G8T`1Kj9!}`vi9i@*&0!aGu_;n7PWf1E-a{D`SJ3De -qPg&&mf0_D04Y4T9fYFfeSZvog(7_Z@F!lZo{;Vmz`=AQyhWW>X)@s);YiI`$AM0vBhG_dn`UBw>$RD -rQ4@&#d=e5Ssn97rs&3;n}I$;0vI1kp2?11f3*B=}|LZ!{`)K`fKY)B5-AwH?>3&Yko=9%xo<1y-d$M -9pczz@TP^)+Rw?*t@7jy7Jqj6O?%dZ|JCiG6YnVIK~+55Jgz1NVRLGeLtyI40)B@fw3n_|TJ7!aSMgv -+$g>ug_5jAQQJhT6ANJ2eUz5I&xgR)%}TS#k-JLBS9$Yd!?U>w^NmCjOa#vL!isTXZcnPAO<{!%XpRe -nryqir%2LmgZJ>XTS2~KIXS$vlrWz#Vy0l@X}9=|A)@xYB)sTm5?Sd2z&ewdhg%uKZAk0m1YhD+zKzx -%jz|7@LPV3|EQp96+vF2M8Cem|s>L5zSA48%bgp%4TiVd$3`W*9%(LiX7A08rwiFF<@Wz -p*1^Jp@^k1G>Ra!zB3Y{>(kF#i3(uzW6jk=EV`)?l_93N07dsV8_=4JAzsCv*v#QKXK%PiQi?I3;2lM -(c~lXntzt}!2_R5?1TQ$N67#@;Qr{tqz+AleCG~(;9h>B>yO<8jvOiPoqdqWk=;IwDafN^2mDI~G9fv -@2Kv6*uh7b6pM9sf8HDSvL0bNfguoX$9$o#(hKSAX0=Q2Fd$N6|AlrsP``dyQDYZOIv#vj^3L^rJc@AkX8TNWCj<*eA(%tUG;}=n`NCPKpOjvJ7)-60g(n96-9k -Bj0|ISMl@;sZQH&ASWkg1YRB2?^4yJXIi)}r8pomu)yoR&$Xo^wkK8ycWB&0SHfoE>}K^9Dnw*LC->W -Y&eiE=kh(&h$~H7X)CrT6NejSoD-eIaNwBDRLF7j(#D@jRjTI8S6NEb=Gw{LvDz2G}kBd>=)y%#SW#E -$w`KZf1pkKpNeM4_*8)*d(7qor3U+|2H1B{Gnn6LHoYu=;!&UyJ2sk^+jY?3vtcE{P)wKo7u#QnZJi| -qxPTGTPh;8ZkET6Y3yLm!hP4dynNX0KNM8@Au3jzs*b+6l%KkClR8K%(e4pQ?dL&EUbgQ9CYmjqw7;T -e3z<$z;_nbir4>Li*|2;G7Ypk7?555so)1Q2?uq_}lZ@T>_{#>g|z-Ae&jlHX8-CE|2^XuT5^K2#E1aAOxX=hqn|u#09ym%E?=BWpR$0~+_K4tc2I1?$&BK`~9tn(iI;G_*xV8u{xEi{kT+z -^8|PEs13+tSWm%uM!ZSo!gBK0lU6G*pl$FjH8(gs}-#Y9|S|yiqo8)LriQZ{kF4D?Q9R=-7zSw61S}B-aBXR -+-rzi4D(Qn`JM;cibG#dGfT1^UJ$9BuqE$$@2F1sryLj>jR-WHk9lWa+VNd05;8N@qoNsyRn^HRu~hb -@)lL|qr4_japL*7etT9uX)={-)Jfq&T%t4MzJbj;kqe-r>%jNO+hVWeB`C_F*3Gzad2e}iQgic(Q#=k -5ImTyXOPLX}UuNsnW!LGM`u97CS=P5{IfjlF-~IKm%x+%fk?z{*kP!g -wb68yDyGo8Xh&rkwcdE7Xo}Bo2wS4M1Vff&AI{|_|! --l_zFMj%ibLi+=KFTTgq5LI1lE95+ne=<^LJkK5>2CVDGw5IXp;*^HG>rgyo)yc}IlomH$A2zAX&PA$ -*fJKV2bu-W?DtH9{dL=>OV-{&2M4NH0`Q)G|)`wDEdw_F8nxhgZp=zO)9pRnh-pz170w$?TLzF@aP15m>Xc!+aE1&&x+5D7nVJFIk-)~(&Q -a_31eT#5)DLp1QuqX2o(^{7k@AX-@@PFltZ7PCQp^`;4We`!z479736O=|Eg5%^7BzjJ&=e{H%bR>#g ->$`G#MKEpLyYpY&aM7c1Toz3)Pm1e3sqfGbY=@o-Vj>!X&L(mgxt;prSAL2R1=JYwb0OgCm6*ELzh5bZ^(|wuI)Zgrz5Cmbl?$#@6Ydq+vd0 -YyVZfhW4CbcVl|j#M_cv^g`I6(>`Cgx;jn}TgTsd-*5lR^E6=1`1BA6-VZT!GIit2;VIT57^u%KB$4K@A>SJVy#cjul_Y(=n2;yU;`+oZ^eEA^ -x44tP{`pbzNixEG!V9_AWMYKlK3(fJgZqYE!d-TPkKeu??f<+73apv~x6`seLbDEiH?n$dN{idai&Sw -UpS@i37q|te`pWyQ)ZuHDlRnJrhY-RiZf#{YuV4N4 -p`HX4}|*3$ycLn>&Yu2ocGSbzyF_^NGagw}&pSOz^GgfD -5R7pZ^f~-7 -1S=lb;-LpEz(CfbF(f%+xp6*3d-)w4k$2^KJ+IO5h`QFkG`xkzXA~%`v8oT69cYxN;{3X=(AOIId7te -|Mei5l)CL)x2uszM?6%|h!$)$+j3R{D|GDw^)nz~ZSwdgqD?Vl7Wc`(CH!_gZ^vtzF)(EGt-Ooo}>_H -6wTF=)kgPK(kGkKCe_MO2WBRhN(UJ^pI8L_=xp_*QJ@fhcoA!-JQUNch*8o(h2p$dNz8Ki`<()6is;v3*YfXV%TV)h87wgR^NoeO3w+7g+O}En`06B7s6zsW%PJgWpLb!|3yKU(GS9IkRs+dnD5tk -WaJy22=gxY?rCw1Pc5t$kjdcqRWnSNj`NvxXA&F?L65_x3Jl&E5+s>LNTh>(siQ=0=vU+1G&gGR(ZlZ -*XQko%{Os=q4``*#ogK@#@IgZ>Z>w-L*p!U~?O$2xtPL$ZEpS=7Rj{Qz)h&uFQzg)fih?^h2ULTi53_J>+_MnC; -k38*R5s4ka6P6w$M6n}aB0tNd_($0YhmIzCiu@P;IqEQG{7W-N;nQKX-oFJOET{gPRn|vg$ool<&KI% -;;6D}$UlW_Yy$2vU<2Ddy{h9tRn)F9I6Q60!^NSkK9%TT&kRSQjuNAKIpZcD@1|nA))~?yCUu9I=^gj -n8i|%?Dpu2mpA@Dzmx7+)CCC~gf?e@9#_MjxRJan``emQ3EZ<2Tb1AWzX-9L~xI{BW9h3s(OR@w0Kj_ -$(nUH=v2e+W*1zlcs|CV$H~y7B4rSraIxKP3JlLCQ=7F#AwS0f4~8u)T+B@fV1{+Jl%WV8;n%ZCyg*C -`D^c=OM4gNYog-zWwBVG2+3UE;BFPMrQ*+S}i>@_Io%o))ON|EZJ+<8{}P|&+1$g!VpE1s0Q{L*xQ?q -`5Qb&=g>6~zr2)e4QR~#dMeFbNRjDm-(b?X*cN)m*DN9_i4XLGXTXe&s*>;=S~nBVQc -Sdy>zx0ui*mm*)r!mIo=o8woNFwT)5MO#;?`vG8VdK%WZ^#^)fVYsF_`9C-n1nb}_W6CcMc7?PdDfBe -9|3dXZTbFydvkSy-pKxAHs&XEa&@ -H-{e((DY?82wZ5J;af9}Iyj>Ql#t;?7E3A_Z<~kaNd63X%% -h`lTW8W%ENZdZ)}8G!JOb#CWsFTTqHfKQ~+8lstKj|b=4WXw$hZ{xtF-GVu+WCKUCIuYQVAQOC?vYDf -ir;^QjAN{M~vv-D`LNYTfBodg398E_$y#oxNACow#+%Y=LF@3w&?843YKs)jhTHG|meOLHLCuP9D18f -dEMO+tiKud@)gt^b64)TAx}r55D*m^XT4CPFe^m7;+FS>Oju)6Of -Qty*(I?=nf=AZfG|T0&s^s%WE;SwRgzn3Xh~y2uZ!>0iMeW|Ynq*(*JvJ)zszTQwH%hJq8^fR%nVX{x -%AduVR=AQnwQ0N|T_<@aZ)Eya;IowG`)W!#%k{2mN=xzb}A%654l=J*^>Mb|Vk!IR#^@h{}wpl$#>2D -Pu2rcC-FRYZErv~Aew_7>_bRu-o{R3;Re)NVu_ID-q@?wW}2Al?+le03M>3d{*a)wD+PQX`7UiQ&hxHH?2{Fd2Mwfq!kD!v2N+x=$TN4$h -Mv?CD5}9xd+pQU9GmN4pjNahpDLeLhnB2>PixCyrDv^1IaRp%nb_zb=a-NnDagEjU3OXM~X>XPkUggO -9iHzhLxWQ1FL-9kRm3!S@p4xQoN12KiK}!^b-M;evzwi=_UN9T`VTVP(IuBb!!&lknmY^_DKhc4xEqO -WopMx>?MCzVV*M$lsEk>e14Z1cIBwdz^ab|I{C(X+mqaso2&YOzn}RQXg~!_{uk`{QAxEC0lm(gC6!S -M|$>4E0g@=wkwT%v&jO!`eb==uym2Hztg+FMlN4|&EFh~h`~>z`h$Ww3=C+$=|BHv`?+>G;cGd|iMb_ -C?3W!P0KT2rp^3cZlyWQSIf9+2MVmAetUJ$m=wy%CN=pkFPvO0k9Z53jzO8|sN -bcRYl?kd=vUJl>4Xw#iONyiqWKWI3dmC?mJ@KkhBe_c#LH!MG1rvqy?%r#Nw!PvPHWpNSyyC(B(15gSAY&tu90}T --GO|ZpUy&8lCGbHM>Z0{iKf6ivH#yD!r&@FS`7);lM+1Nhya$JYf4TN-cGIyM=wz-0HY?HlYnxlCMlA -GvFQH_UvF?v#!XcwF -2mjAOlnILzUyeMYTnN3lmT+a?P1Roq5_uF8qZi(p1h^TG|K)|rXnqJ?_lZxUf=L2*v-nrRE9w@U1^jq -DRK5LUfI6LulW0T^1|X(6o(MtA0GE_3G#!FTdI8m)8U=_Hq@WFUD=qd<{xdl%DAQ|(z)n0pV0#Ki)VU -SqrJBEJb;Dw{rr%I;ZkzW%L!%iL({ccpj7!U=G+Z~eM;%bl;#4W(5Rx?-<@)5HGsm|nuEa```XZUH~I -1ni9HJ=xg~VK-`@UCw9D<;3gl2HqMIv>qujj>uELZ?CJyl^b-g(iX@U$Hp;;DX{&FMtDB65)+o(O@ay -}PwsU%%l9)?N)$*G?)i#Hd`cJl=~Y&SNanr(pg2BMm3^Z7ZRuMG*TRyaIK>tw{#Y%@!Yi21-?2@gy1P -93GuOt7`lN8E+U^SEUIEwYs_d@nH>(V~^eo?l_qZ^R`wKksbN*W^l*!4ebN;07^cr<2m%>!TxagA4I# -c981Xo3J+Bi2#P=2G|p%>}<_v(vy7ju1^U9DK|Lq`h9lGOHyl0T`EY~cmUpgNi{1Yn_owM1RC+`UU7l -w@KT+jY@{91zy3k;DYo3t5+3P(4+qDF -cJNd^ -Rx=Sglzn!JN06VcypMN7kshQeO^&z3#~Gr4j_O4+`FOM^M`{X$j{Fn#SzSM*89%B(Ao?MXpB_CJ`xgf -P|F`vA{<-!1@HYkjKht_H|D)FPS4ch6@KkTNB1}TlX|P+_l5p)uw3h?)%h~SGW0Rh*w|+rLE>mNNc}z -)kR%COoUsVDczZ-v0n`*jF5KsJY!`S(GLuX=8!Z7<{h7_QeQI# -d!Or<0sADh^vYY8)|11Zo8bBgVFf}T-AObny%3|3e~zK_Kh=7E1_FnMY3aUGd8NPV;Tq)^RMjTd`46q -`IQ*(Kukq~SoOgv7n{$}6M9`8uOL$>i{=v}#!@0-(gr>brf@I;Gp(3ms}Oz) -+ZPtkVY7(ChHrocg7M(rpr)bxV(55?DzsMX82{$!%&-gHgAVAp3;4#HcC{TYz?=RuHP=Ov)Av=581PV -;aUil&vy8~fC39oamZ7zIHp8q`L?%tkus4$5!e7gHRlpzu{{a@W}JE+tF@tpa{^$WtNvO**vY9M;!dx -E^N6+TDHhoi-<9*d|vy$Y93O0qXITd@^gQ*&9t|Xq77%4S36y+EsqRj>)}@^MjdkO@AjKbDLh*Z8>cQ -olmab+pP&OPGRu4FwOr&JeVGz!#CU54<^AN^eN%mxBpz>^`CeBUgGtu-F_(d!cl}I2$Uj72tgnm1%J8 -Xk`y23AoM_`J?c)f!vg`!k5s~rcstzf*p#Nn>dXEb`}K;;ppZ?p`&B$5S$~9%mR -uZS2%XSkJh*SQt)r}R*)mRkrtm^0D2ssd`jL^H-cjTwl4*4v@9BGJ;#>#=!*sq6ds0wncyKy!Vq^ksNZS8Z}EfZ2uN#e^y>hjxRK3{$B3oTj5c{@pU2>{o9v_{phUWm!HeLa*^8pE -TRYSv$7BBdzlw<

zyL$!_h^Xkg}uZMjWSGakRu5|XjH(c-pFh#UKOkBS%jA+*b=zc2iI!5}lx5Ex@ -f7>K)M-P9Gvcq(J66*DBc5kf%^!>*vUgpzle4uVZpTt{%y_LQs--Kq~%4gA~tRkZjb-8J0S*6lUSfi$ -(@cX;R&|U+44!1n*L_qM>h*V6=Luc9+6)ciJ`s0Mqp4<+v -XBP&$Tb>+c~(kv=L7sD~v@f<4?#s>yKg+Nr@Xl705bxpWY*>b;&2UNBy}+xzNK&M%J3H)y3YBRJbpgp -DPP0r_!vJy!s$jzpbm16s!g)p7VG3C|!p52@xKhl`QHdQX+$}+fzID{Y;3#mdfsLjq`pg>A4gw)AK`D -wG}anAs({Rn;YU(A8L)MzBHH1v}NzObv*a#r}T*yXVud9{n`HoZi(W5a4Vwk`R1lA--wr+n5!Hr-b9Y -SUS`!-t8xn8F0(yurz=JbSI8h@@(R4;a!a7?O}rtx8&CBO=P&hZd&@e5(J6#d%Og^D-nNhGKGi`ouVn -`Fq`5Km^^Dz2fa`=lx6@iKw@4XsX~a`Akd&qD``V3t$sq6VG!enMSQV}gHugT$b^@n4$lQA#L0I;AM*Xo2A(%||2>kURg{XPCOY3Wq|59P`F&x%?_ZrV4Q9-SDjn{md&{r*qi -Q#b*3!9Vg#dML_2T$QAJLz;aT{1*O?#y{Ab<m|5=8V;|c*G#%#5#vb(T;BEFY#-Sh<4~GlA~XX!jCfHBi^GviX_ -S5mIo%sBL2RS__!El;E}1`<3Rj(Iq@S7!~cbsN11U7AJqdXb_9Uf7grQa4ygXq1l`kX;Jqdmts_2T&73Zgz@2(!ceh -zQmV{8Jm$)8z#LN};%p4`UyG>C&JD`_hH|=K|nlSWYvZ%`z;ZXo -c26iJL>wql}R;b`x!w2@UZDJhP^|Q6aB@kLqz@^ -=El?mc>Q_jD{IhOwB6Ay%xJJNkpGtVF!5i=!~(9WtFBxJ-lN_Jar@3&Xs;gN{t6Su5q7}NQvjR!s$mX -p!{40*4w42#W+2IB%RSa?%di@XyE&UrsHHzPxoOZiK@#_Ijw^t!Uw&^~7;-$pbSDXF -c%)MEcqh_-;c+ao!W%r`b5%b_X5FiEtVjk`o1qcvg5W}zkKxO;2%T*rF>F!#&vNDl)ct}YRJ0f=M0MN -n5ALrd^^X`I%moOGz`OQXxTp*Em5k8*+F`d^N<=w%T)CS`k@rIb$wt3c#Qa)V(Wr}?|d~F_!y;bQUU0 -|ym!+S+C1uDvJ8tj;(443@*c$v*a6|cv`J1Zp}SodZ1tq~~lI9%e(70Mh#eZYCSGQ^F48GAe -V5wXCC(v1)a$`Eu6VzB`?K%oVSm3 -%Ho^(Lb3VMF!~5==jz=#`R2@jy@Tu5q)OoH4z4GjdlXERyngA;kfH>{pvb#0x9T?spK-va^2Anzhx*j -5?L{BW#dSR0d`h(zo=ss+c}M?6XzTiDy!11m+L5$nEx9n3XGsIQU=X{(!^U>|YuX -yA4WjJWHxTEEYJigdwawJJm|qf7^o5%uK*Fi^X%cQ-siBwKGph8G4wLqFJK3^8Z~a1U412~#v!uIlP~ -+-E4!7~9)K?lgmyQO!{NXGrt)auzb(VPI=Gn!YNJhJdwMDykODP4A>J7%9Atjc=d}Vddj7S0K6&OADpTbGh(v>t5zJ5F)Nojs|ibZv+?ewv+5RXNq&B00?#2yhOwc(yw8q5%JHnv5W!<1WnwQf9&@9%bYJ11Q7p}WsOWEGnx24f@jg1 -#~h-7o`G?u-WRsrE%Vcv#A=unoB4Aa;*PjX`LDWO!rlFXG1hUV$ncHs2AFQ2l3pj(DibJT;=$SXSF1! -Dv$E_;NLkcC5JLfk-tbK+{h3t*}q2yexywXXc$7^pKulWtdIN*x=jEH!~J0&PKElxns{X}C3n^q`bTg -WLXYk>R6mbhl7tWi#hVmXa;deeR#5D&(J`s_&onT~vj9g{n{*f^h;OX$+@md`DFzh7q3>0KLPk{rC>I -AcUY0j70G7G=W7QMbiVqWC!jza@gsS7}zn?9^-$RUK~fq4+pUNc2)qQ5Ar7QAwN>_@s%b=rW(hOn4H9 -q<|_#KNh~gdk92g7|CmreqQv)5RsNBXiw`_@=b8vN&e~YSqKK -Q?ks-(Y-s*1k!R>rIn7rUR#>5cc`wUYNzyD;V!x>d57=DTInaQ!15*X6M{)R6tkiM1a$I+!5G7 -Wgx`DQP4FA$rAFts2Hx*_2R_wu-e1K>3gLwA$BWwNq;P2pk`(lYFbx=Z(kJaM@2>MUB=qfavigdEgJr -RXB=2*aZF-&abA@*&uy&!Fu0)9Xg1ysn8)GJgDnN>ku9qEbUY<5--YUMF81#}ib$r#tQl6A|!NRwzA~ -2a&){eC_Ej&r#^9pO(peZjvN-hCjaa9QWVuKLFB`VzR;as4zN+94^@95FwbB463ZlpPzPcr+&2tDE}S -QSX_22hz{43;;+p+fQ$uVP#qjk^^eqhSxCr9#J47~jM^D}`y|Bq=wc^mi=xOBSa#l??I_`>Y!(y&p|lb_#2YW?=_Z|=jk3-1?{Y84Ul2E)!q(aPY;;ZKxy$9_%X0lWG+Q -*#s)!S*M7rzo_A|cV6YylRxb`bs+zS2bG-n`8Do2UcH#diR7Ga#JFQ=*Kwo5!it3acSE0Sc3kUrHs*= -V0&e7c2$Cvq?_XAemF!!>)IT!}5B3-X**q~$jn)biF(WAumYPU4psy;ygOw@~exI#WOL-MckJ?gl5-C -v8~-CW6hoRo3C*!J9U*uyL9LbaIFv$%+_o2WD?Xr6)oSyl#I;Q5PGI#<hb!yvGRUOWc`;(v8*zmf -*D4}418af$4L2G$7CV$%gnAL#+LRfw3;3Q2DS5 -OY9%0Dv3>f4uTQ~ST*S!1n4t)XUNgB^s_EC_625biR_6YgpK!ImDxwP{L$u_Vnw(D#VkS!w*SK1%Zq` -qC-G5FennLenv_s|XHo0sCawPfApJs1>+T3Vk3Z${5c}6A(mWsn~F@IUX4SrPOKm;XitH&Tyu|+51!n -*Gh9XO$+$Vw23<;&oiDM_~DRZw4@e6$i+^fjiGgUUv2m?H;N6ZZh%QN~c -vP(uvjuiF6psA0PfC-`@l7L|n+TqlZnj%no$5%K -4<0Oh4BeY?ZgeV$8X`K2d!Qg;B=}~Tp(xX^Ue5p*p2OOe4%C`{uBTXLSLoxJYN`9NaNgb#wPLCwg4zD -8WQ@f#$no36gh(aF~-%lDV`}Dpf`G!}&)L@FPcb^!lS8ogw-t`=r#7Gb8AO!ioL)9PO`B)%ESgOj))F+MnP^@5mnFo+@QU#qSN_{|l&U{ -T-@0L{5>fQPnR8|Cdpf`nOTlw$7LlV&qGZvsNviid%)Jh~LBkp$KWQOyIQCr!IL7WUnEu-*@FQ_0cBA -;C-}hG2=j3 -T?1DF7P1T(4l=$VRwPN(NR@==>?M0o0Nt*SAogoSJ1IGrWb<^`>+=G#5d7{bk~dA=1O$<}R127cQr -Zm$`JS;yNPy2I$QL4(`UJkK*S3rZencIYYkK*mZba*`NQr73v`5DbExv^?i6fj7nC@(ccc~bh9m3$9a -=+GYMSA6x;ye|}qkOR1kuiA!%1OK~aeCfnAX}$pe*<(g#Fkfkl%jO5ZhX~ebB(_|G>ddG`7!fFDt~D@NuTo9%hP~xXsuv4-=*A?gCl|`K2wng38vE23`J2ftpmGbLQ%!P<8c1v -hxZ6Er+(~n@N<=A6O;;7Ne3zoTyGg8QMmqc?FwQv#IZ)CE7+)@fG~oU%C1v#j832R%8kK{p(EIFA5mJ -0%PY)h+=~FQF29L9|O{4)!zoFV7vy;4tb3$M&h>)>?nl}(wQ$2x6#ee+`wV^*mRZEc@Z>4%MkK1~gqt -Le|L%Oy{%Hj`D6=xQR8^QcDb?p?XG+Ve#gdryVk_X=P+2zAk+kxxY-v}jyPt4nUtWjx?Xtw~$rMV@bU -_G?neOG-wqT$PZG6f2w4Pn1B+$ND{(~8RdHDCBVkwlhuUMlGM8#@iH1PBgWtnJlHl+7AXOoB^6N2%mH -y4;M>IxB*AiSlO@jB0yM-*0uQ7;jUwVrdN<_ug;dx%e;UgxMuxKOjo92+1a`ls9a@ZU~~MX$e}ymT~% -Y%fdz5Es9&buX*_J)8!HfcVIfP!EPm!^C`PYwbRc0z{d6za;cFvh}K?dZsL-zdCTUld!ge=$E#jQPNC -)BhG%pGAf`G^V&KFhSk4varldkOQc~*^?W!7= -9mfOixj4?psdL)y|$7doGoZ=H`;*8tx(L;+yjC?0J7jta96M#VxJ@N)=3gV6Nibs5rr@?K3SePH&D4e -&@V56=9PgUo5mVh+cNm_@>gJ6{8|y^cUovf&f<}5(Je#Bwj3vq3~O^+awVUs8X3Gq=i({YAnGwxg$fZ -ugG~t+QRm=liNxyEo#86es^zQCT#7v0 -659+leU;9So0{NL$*EZL9!ck*n$7F+(ei~S=wuKdR0h3WKI0yd -qcn)opD+AqDULpaA&)6Z`;Fj#%5;3Ii6p}m|>9 -N@xczI^3YcW7ez>X2SB8o>(SOJ-9{7^R3r_0EtVwquyLv1t4a6UQt`fVh-h) -${y!s`^lYmlXYe#ll&=VTM`{AiB=C``RFdz|V1+q9b!TZl=+)RkSxBftvFVw%i_jqGZ9QcQ0co(l~GgZ+q?0kI>?x^_yOnMh$iqo|?^-% -y8MGVY#6VN^bv$E07@I5%vkjJ*9Vo++y(sxM^I%4Lnwt`Er5wD4(InEk$TWc -W)zF#^UHBZw0#U0mikm%Hh1g>=nDTB~UlBZe9Yzl)#bA-dGl`%>}%pnhe{kcX>T)*Z2W>RK;%hmL4>qZ{%}&IW>-mIB%lU3q$0E2V^dcl-QV;W_ZlFzdNyI>ZS##gzX4s -N-)}_+6^m(KsD0vB8AH3&a!rC8UE|Jx#5D7?hgP{dqu`4@XC#DXSTGKLF3(RuDzq(DsfG@o9sak51AW -*1p#PIj@h>j+=T7mrlk{*5KU5WI3WgvG$59f6X@o*4jK+7|iDL*se$+C4=M>2g-*fkZDEz7Td`U*6M{ -x(uex&JxL9>VbQF>H@@^8~;a{Op0h(5y%4jpCc5d2F%es(_yK}U1v?$pxclMtW|hr2r|6grwazvC3YR -B%2;BP=PJnA_;j)2N5v|JKm8lXT%$8}Rf``jU#Z(R3Stj(CZM{4V-S9YP4kM{Cb15Rhs+)X?1?2%;-lLlZ~%+2!y1Kwjt>WpUETO -Jg<`O8iM_?DO0ox~NoA{`dq1t$_TjZnBOAvDqIAH`R1D1e!@c86IHref+&5(K6hEDLoZo4QvR%B`t?& -2p)oH(XeQ5YKp74&A=tmW69|y$%4m3TZeqprU%5RdGg~6EowUT|yFH8ky5$ -m{<7i6ycD(0ip}5sBwiy=ddI5Pps*$C4l}_c=1mb`kfX1?IORnpcq1yDDCdp?ATz*6g5%wJex|_~@#r^xiyoV0sp5aGe0;5N8v!9=$AMF72xMZ3g7BNUt6QEm -5szBr~4)0yD`{@v0pK}vHU{5|6)*`UyiAlua%8l`!~~S_<8yD)#SKQf0dpdY+n56>^+X>eAMsGcm>1= -v`8G2{gR%B%q~P~?ijiu@#zpFauJ_@P7EjOA)zS!zLA;#31(6s@x}99`9RKT%Qa0~$cSpzNIs8N>z~a -2S)A{sMw_*>=92^3_rPDf14Mc>}8j -=3yf<=0v5LJvOmz7*im%d^ -z+cT<7GI?;?NW9mwsZk}(kA?VBEbG(kw)NG+vA&g%g+d)3)?EP`Q{!)MQmb9G)VyL-jKej-qmVP-FqB -*fd*$E)lj99Y&{TK<7rgX!)>)}mENcFHe7gBQ>;ni*WGEqF+j@%1WCch=-;Brl$U%9AzA37n1LM6_Gs -$SrekrEhZz=?3JI^|m`tSaNctu8VeJA4>G*6;+@@*{sAh^*$YfQHeNXSNXk&$_%rLmC)F`C|8CywSAje0Vc$yluH6M&!GKp_O8<%G?Ue;EduO$#5D{XV$8XPdwtvsjqr*G;^t&A+UeKd>dN9VKqvpA982V_NAJGo|;~pI$pH6@*KS*Z#%ITl7#Ll;)cP#OK@q -Mi8dheT@b({mwgmHtTQAn!a1rNplR`BrG{oYaM1pbZRJKP@dZ~Wfj_JF_P_cm1w4h=!^u!ma(9kOQvp -kO@DoFy)}GdvpQ;e@@g^J%!|t7uVI0!&Wo=9@1{)B9}cCJK^$gi<%-Om>9_3l|{ou(1URJ(y>oWYU>V -K>qqT@38uQGCYXvQ~e5eWUqPOY2W|3i}sU^B=ABJG1%KD!;Lt7!KnQgyR%Vk|+t0BuS&Y-8`l%KrlhT7=nH~=|X+_e -3D~Q)1g6{AL;rSJCgiI*U=YY2lPYQIOa8eJ7!1^dwYy!Opo0EZkqN}_c-CmxO@a5j!Xyj6GwLc_2}OH -Qfl~K%z%H&2jqvQn}0?uMEPMWiR7@v=Um<2Pb5|{2r$K#0>ewMKt{ -*Mg1ucY7htdYmLeXq?&n96;3PbeP-||KwgT}M@;CBt#=8~lpjSFYFE0j9GohQ^L -k{NO1X>G?8BT(6f$y&>jWu;%XKdyR(anaaa0z+o;8Lh;z7x~e920Eh}K$TXGMi>+8BN81*ATJB^3--> -2ef5#q+?0ke8GkNp&7%5Z;8GXH!tp_Y(3N -5oq!k?cj*Jvo<41lcQTn+A_+o6D)>hERkl{04^;7JSxMJxYMxzY(FfC6WyNF0R{!Q#Muj_%Zgm|n4_zlLm@;zTci+Kp-`2;xn?e%HLf^=jfPKvD$dIi)FEbWQ;{7d~a&9dCRO(R3-H -(;0rad24ah(VJqN*Reo+q>HbZzsP(%6PJ>a&?>y#Z|*>&eaqpeuv412l`~;C4Z`Vx7e?pa2L!NOm=D17*vdf=7DL -6^Fiv=I57#yZRU2zS=Sg%q*mc1V;_Q7*P`v(4)8!g8}|4M@)V -h2#6=o?cb8vv|(9%k{F{W*zS((U_r$!8P_QNT$P{^}!RMCXP39 -v8EL>PfjvA-g0+Ed=+^b=%jnFfxGbW~imgs5%9@rk0GnW?Bo^Bi-O -?q6CchjQI0be3mfTXbuh#QCI%L@SbNVKDl?q2}R4y3Bj%DULV2P9-9KSD>vwtM6P9%F^er|xf!g*q^Q -`}73snxuQLPK$lQ18}dWEWQCBSgFzxtYe>jY2~3uTS{YRh`p+FfHbX11ptznM7mxY=2;;v3U07xbC3q -rzf46?~rx>@gl#>1pU(memf-;rXhL{SLOl@d2~qPeEyiJ?KX -v<&f1P4ys4=Wx{uM)lsm|cY&)v!E2(y(;?X5n%`$dJQ=~I+QDZ0#T*`(V7jJw-0&tSKQy9mAj6i+_v=4Zni`3hJm#lh`$@ZnnAMv -|Bg$RO?~qjqjHyukDp*3b3i&AIPlz8I1f97!+zEc`?FgF{_GzA>=uDPyT?DfMc`Za_&0`onE)YJ7Ig2 -tA(}N=Lxl}8+f;t^8llk=$}ma1*eGQaN}=2=lrr2N$rDjQ1OFsjGvx!_hE@Y(7UY{UYbu9|3RA!ePr& -fjXks4S(=MEatahn5c~10~waMxf;R4A86zf-%W=C4I~B`ZVTCyN9ea$8cVmjr~=-jTL@R1Z -4#AOq)(x`NU=EfjqQ6=nEG(Wj*`8tVnrM^(yQi4YG7iUSJl-3UGuLoV^IaZC>k|Z{L@kTT23o!0xh8$ -^Cidb0AD{>3n@X>{q+qp#UxP_jB7nEe4Y6N%@NXE397DQt{6+IV$_BoE>7TH}33&dTbLR0CtX#wDaad -JXN{kcqYT7Uzh4K(*DYU6yc(LL(oE#DV0}R(DbUkTSC8IZ(8=QHwI(XTKn0PDJ&LIQ+2OIBz0!jeKj(xYBt^@I+0qjWltc_8hxdMq -^+^4!&DA!d4A!^P<$S2_a|o#)P_bIl&SKi8tj_qWYVD6i}J@-ivx*)pM!_Tc$)&LB7m!BqkL`S! -8l!6)yVu&wBbyVkz~7qo0Gm;6ZH76n&E$edccA8CV~c;3MINXPZbg|l2sXy+GABmRSGTQOr#w<>4+e$s=-8Y^YjFe_Ho3T&OChrUxRmeoZqlToDeIey -3FX1w-mcdqj|`|D4|)O}dCJ(E6-u=5l$lLk!et5w(v+siUWPgr=~n@hITnH99b3)!o8&PX&Y|@RSWsA -bQ`N3dCH^(~XG7wv@QE84 -Frx#al>J+4qLPbxD++yh5QQ>5su_J$TYEeNDMu(bvt^^AxyM%bNI!U#T`DMiU|&pSog4?KaimpYk`UZ -9F>g@xT0aL;SIC8TtXLNi}Q7SLQ>U>{xD5(s(OVbfm*hK{vA{!wWCV_0Vw*$>pp@! -+*Qh&oJ>Xt@GQCFAU$|AxU5)1w%MRQY22p6bbLg2?8gn{m1X_CtpLTS&1H+g+Fdm8?a`q(hfnE7W -t`$GRumn -4fb1hignrcdsgKO<{siiSO9||#xRBV7baIqusLzoYK8_IS0Vfmm;07e|4ru=rC!z->fch0BAcsfwwpM -F>d0xon!Bq}-{T@y%ziRPvUG-Pv-0$}Jf!FEEDxoiLq5oI}?ofdIY1BK^9V}g~&McOTd=vjwqm{xht` -l9HyX$I#Q=8WQza#PcVq*GEnqj$C%K -f2Gqf)|0G03*XUj<4NG4w=ghQWl@{<8iZc+g4!Xj)>)}WVN#eKhEi6eN8soxV;csLO9ipFK%tkDh&racT{nBRC7{ux^@5A++YbFjp*a+#P(WCmf9B?F0xN6ujMwu0>HB4#gVMLlO3()1Y^5f3p&Vy3Ocy&FmZTX2S -p$;WQdgCmpmA^m>gEz|~L*UnIhgaJB5yr0FeG;!q`DC8%#PGP(q;sVA)c2<>Lm?lf+J&XySoblcQHvj -LKxUZ>X7qa^LC%h^T;TjaD;p}t7K-=F=uyVXq=68b5LZ*jK9aJ5h%{-{qXb2 -e2nE49Y%WW4@UB0HF}tfn%ySGwf?43oqB`!wUCBYYiPTF}y_(0tsUvYHZy9?7ET_|C&Q( -gmt*p@+_o;3Wq`xdzN_59lW$g6QZPcJZ9HC@jHOe@8&jKbORy|q!#hB*>N7ix1ejfODqQ?xbu%6i?{o -!Ebvcb;?pox_SA*yN)~xmTt}8^l+H#Vx8>KmKe -1bQ-{cQ|4=)OvlQZazo(0+n>zWg90^oF~E2x4Y+fp&|P~8^Cc7u_5!cDMl=S-~6)hKRHRwHEOBxuGgS -@F`0`t+O&`%sp0J;kJuhl^ZS38Spn`Gme-dVO`R=S8}}ZgzfzWKGRJKRkGM57p#u^2=vW?yqHwo=||X -OGl919pAKMptO^y3|6I_Bmb2JQC*MaMoA!5nPx-04^2n?U -C5%8I|%B{p^q$eqjP4K14q9$JGp=*V5Nh{-yN-E_gCQuq{!b8idX&?prV*YBy6pQPj;%3YD{OIi_*0_ -gwC(zGqFEE`wqU@*HA9uX@2(}Ecax&HYODR~HBdT}><1bnepf-_+m3a)DFb6AtKTp2v2w=iC({k?5mAk=({t(Y)3noHnprs3O*yiuzFH}na3S!g~P` --dycqS5xF8Oi1XdUcA7Rn?nLWaM28{ApuBjJb|xSWu&7g#L@>bDwNX7I6Jl^9;pBJtdv{cJj&sDikjs -aZcozJg0*Pl5$s1Q7`kryzec)aiSJe7}CHd6tTdSP|hz}t4xvOtDG8D@-3d`iIWJ7V&5dvL4Q57lc3D%q -03y;*moF`Os|6KeL(QgS~>K2j0w~gUETTnj9o~a=#Ar9mI&`2g9C#jq8*+dJ^YH1HbcMv&M`lg{ -AQ`qq)E;70z9Vq?*-c`6HR+w?LD;qjCaM>ixha~D#~yT_Dap3&>4a**Uj- -yVYlomGSDltvIMK~^jX3zA#x0j3sCbf6QlThg*tdzRk67e&6@gW{z>V)GS_+UkiBnB?7%!XJhpoE@?9(QQN?(J)FK5%bZYgnXv0e7p8@@O-eNI)g$-T_?g0#$lEpqAOo!8b=?anWl -%t${t-Gl*IUWD2l#gq2W)xDE?HT_WhvgkspBaPkkvnnuJm8(74A@Y%G7lnT~h2M!V(?{KwYHeny`d&!=esew{ci -~9P%CS)|)8;bI;){hPI)IXrB_>ygd^ywIlOwgbtQdj|?dEx7j`#tH3Rh<6eT*M0x8rwjHJg`wW$z0-n -ER-hM-zxMe|S$?q3SJqwe6HfTd;(DP=r@ZAOQ*E-tUS-UJS|SE1IxTp9U^p8o -{*T;UqS1=g&l+bWjiRPS^LvOs%O(=eNz5vR{Al#(q_Be&NPJ{P4>g7Cqbijg=ybPw$ae>^rw5aKzb+jgN}%Znd{Xfta#nFUD0Fgxod#bzv4oFgv@^ekj8-(5!JpgY9(!7@S`yUzd2a3% -HQ1)@b#Ad%8vdH!?1POHoM^u8s@YM< -DQvY6!Jb0q{Ox`=df@8`9vf^7VhgI+RTf(ty<>FmQg8yJ3MG`rh+^e$l1J|?cgNaGlmi-Sa40B2PJvcjL^)iO_J>j0v}Ub -x@!VWs-_2_0;J;RdO?1^L73_bSLl|f*8MHMp2TjRZ6i`{+8*9KZE~uz6aGkjvcD26xlS&7pE=P663~C -YU)-q9BlIqr`*KMKin()H?J+qN>h&v!#ELBF@h)Sw_E3Y_ZTTt48p*~T1AC3SEuT?u1b&Fqk{>2_jMPC5vupK#6bsE)UlpLfOwWf6 -O2Z!W#_v_7iZ&a_s0H_7)X|?GAFyTd3)8%ge-4tsA5|Hfy1>TXBAwiX&!;A`Bq8g6%9mpzZb)(Q#yZh -Go>zwB#kRuRpI+t(hUg19m;{5S%4u1I#=Y=1?PfBftHD%bCd+WtGHy+rgDgn;Hw><+Y^F>nnWhppf>M -BW`kqfq8QxlhzhaL)q#e=}TX%nv@20O!8oO6b)3`yk)o#EbFl?Mprwi^Uq$4 -4xzzj?U@DTr98e1wT$$eeQT8<43G{pHiULi&;pz6iHED5iX<;G*Bx! -iuDwF|Nh>bOn?-y24x}-fySwIQ1KXKb#CHCE6d@rg~&kh(lL3Ncy4B-S2-hU=q0CKZ_3DhwghKJgCQA -r2nHZeuf^x?m(Cm&ES<5dy(O;&G%lm7v^r@fmrW(#rVp|uQCtY|9{{V{6xU_4Bb#J=_J0#w{k3y{53j -y+-k&2Zm?numq@*yEM$tXM{LV0Uh&Y0;7&>xkd%#71WH*kKTYND7;Li}7Jy1;d=X|?)ZhtyLAMI_#XW -}|deOl$PqseWLD|6(d>VY2s2>HynIjXklp%ayVPYiSiLzbv%jHiIR(M2|5b;ffwLu~E2MA01n+WSUD?h7kuEK8- -mLs}le5tErY-fO)@1*=Al$F5hrmhFd0uGq- -`h(iuXHn-LKdb7d6;II7fOx&gbjRz|RKE;$SkWwFSiKAg)gk_D*Mxf==Q!@})kRTu -Ct&MMwj0va}2S>f1Y{QupVk@AmoJB@+c}=;W9Ja0=|JR-ReULZTRfCF>$8wnp1fL|bp(%3mwp-_h$k4 -Y-J{b_AAGQLx99TeLK4f|>wY5hBmU#T1zYiLk~nyp^tw?-~L28mr=RcE`iX3bNPJYUCvhp8?qCmwC6> -7Y=>T0PhUwlQ+kVnx2&?LU8Pv+OW++67*=z`$L$j4u>qYTT`qLjZ*3`?^$S{HBNT*wgU8eT^W+5mpu9LFS#_Rg2NwUENDKI0N&-X0dpnk`F}86M3Vki#`kAvwC!=LBlJWit#aT> -_DrlqDON19L*t?<&a3(kA&XuV^l4QarWy;oe#fVKaCY|v!Er;@$ESX5Y(|{ -duigEF+1l{10ekiql3T~@5UPk^Uvu#+!s!yWD^l~dxqHiJ(!mI4l>ELTKBvKcko*P~X*h0*^?1bvWQ` -hB*T^1pUGu5XoR2RmEIJJaK!x_K!W;I1yPYD+R)&zU@fTpMoR(`EbL)RrN=QrWbD`%ex8-B_64{;hlq -L(CkNs(=fp)Q&!V3~5g6ZTonF)jx@jiRP9^ROoa#hD&Jh|%_Z0V-meETvn>c`0xgpIA8Q-Gz4nzum9l -5_ro~!PKI~0Whv{ls#$0{_*8TC`KUvkxj`N>)iBoX=i5-TWjCN&oJ)d)jg?|I#h8|N_-Q~;foB6dq6p -HoD@|R7k6Z#E%A%>WYt?_k&H5EfD3Pd+AUw?=h+q(>VjhSBPZdfKP_&-0GG%}yHkBNwmgZ1?zVyNb12 -%`$oe5In9il75*SYxS3c{AbxxgiJiV|}NmOUb;qo^?IT@V$^T6uMvyDyJWH~5#3`wZx6y8h*>eY0VLy -Y_K(Eb{YBlQHC3{3tq(&F3k3pwtCa=ueF{jV?5J3l@5|2g_-|NF?(TUqwQ(M+3tdEmd>Y&_N3J*xj}% -=%_H{a?xd{R09)^bfy1^S^$`cU%>HS$@mXAEqJu;n~*fS4i&vZ+rV0DgNbs{kBYu?TC;@@o$df#o~jI -m(U;aP~s3~gOUS4LD>h9p@-~y1bw81vL9pQo2ZcboPm%Z^`Vb$8X6sljUYay#$ErJ -|BX?Byy&e+`J!IkVPpxBzt4C%qKTgI!19_nISmKzD_$dJA@MmCfcEDH+ITDEbhOq1apF0fP!Rn3~4+) -3pQ;s7YEY+IbvS`0)!AqAm%&fK@0*h2>tp4RF~~;SHEt*@$IuuL|0j+7#~TQD>GN -F1T@q7HralgYQJwp_HE_g7woI`0{+c{eYIY|zge)a)(iMI3-;A|0sm>ier47%@NL*JnVZ~i5?VtI;5u6}lc4?g&skS{0{BAnpsQ*z}f2mvevjINXJ -o-h8R7x2ZDP8ivzEfGH9iL1~3<7E(33=a0?r7q_xjF|b#^(6innxVbIKp8lbcYkTPSUml$wg(kx=^N! -qF92iH>eY7{cN8oOCkbKeJ^eAC ->C>!m9rQl@X5BL(HVLTp{?dPxQ -0GB&MJMea9k3d!1n1;--&0kV#v7pdcaGZh{&2d$fAYG&;&g%kHg^2e+8wy*MmQuZU_7) -ulrM{YpJhKUIy?nlu$s_(t2xEyKYpm&uuG8*|yDDNPi(W<;^0yaS(Ri1tNycM;h(?AvZOuw`@LTpOPt -z={r^yuQ=sg|FO7ZSg+vXjJ_%#6`yC -iW^laU!k|kygkN1st^7>@VDfd6F{wI}W6SxW-|KXwH`d?<0g+ERPCmYQ~un(%qp^Y+=W -ZjnI?V!LQJLjr-CSax3@wqu>$grKcIszxaJm{{;Z#SQD)}@$F6`4u+p8ZU-cyo1|7yLs0pJ$BX@difS -ds5pFJ^hD4cX`_g@O*B5+!xyVy#1(n50gP`?8HgS9pk!GKp3oN_7OdH>Jvo#;dEd{--v{zp~zs`o8Z4 -;5bhZ&v8k-`nXK)@}qJNj|;UG>7=r)#Pc7|zAof}MI%;~bY7K20=?KgU@&ylR~KKT{ONea^64d -6uzr!_pG0U?vkuBs(9<;`an?phOF&gCNQ --Wxx64RSL-=J>K>9VEmhs;HE9OYwT_z^*%;p4M~w`r&TAL -`{wIqHe6V-_(3?FR*w?|hZ&CoZBa5f2iF^uE$(t}0t{&7aFnsOtV0nK5YPs6t&yD5ZZ}Fe*5csEC{HH -qvwp;vl!GVM8I+G?G=#;Y?o3mgHc+Hd{fG5u2kB7WAb!#E-ra$2mtagc -U$IEYcOuzLYy$-pp7C3UKui9?r?_iZkzL!IZ%>(gdh@q@G0Nj2i*i;w2?gkD);^K!8l1>CLv@Y!Qc)Ob7dZ&404Yn@8|JFO3JAR9;3&RNmBp -lwH5!(s=?Vl4h8)rxzs56*U*NiO4P2)g9kHc3>4rSRh9(@=zh|>_$Y@Qw5AgD8le-4qop;v`K%me{=2 -&E;N?~sygRQpR*&0QOy0ApUi3ev$w3LG`R1p!ZF&Bnh-I-RYZ>KD%wG1#DPpSoy$w)YMsY6SF0pC`?ZMw}?Z8CzhYmz^FQ}=q265*qbUy2iD0rvxEwaQ8e} -p3}J`1&l|*b?HZ0>sq_2SRjarkr6X|&U*C*U=P8FAy0b9NLC1NE2`+zYHv6N{0kFI)jrG*$^ps>}?Oa -n%c%Y*Y&DW&44cD>Ypy^mFZz7o`;{8Uj`^Z=Y3QteBxP&HYT@pQ6c`55-xxyi@7UsskYK -LI(>FY(;1NoJvBg%C+k`0h`1~^>Gi#T5RFtb9FIOFf!Mm61aaU&t#G`z;`&ydAMq2^Ex?sH -X73CR={FnnAziq}WKN4Enu&Ci8qjavFX2x2mM&D!Oio?y&>z%FMh;boKVpOdUh-2D48~ku{8 -s6jb=LDO#FaKFh5;RxM3Zuw~(JTfT)-FEa%Jhz;oO~D9EZF(EL8iPeogS{t=e~I$E?;c8oF;N~Ar^-q -kndn;^vc`H=w7L^zk=D8CUhksL!3wAQ_ck(epVx%N*CL7qKsexfHZmJ^3pW+pP-Yrg0jqbxM&11#IJ{ -7bUY$`@HYQyMhY5x0JIGQx&B#wYg!xrUy1gE*fps-~QXxI!^&Tl3Mwz{v-QCY)gYLS4v;crq5RMChp{ -yz8xW8}oi<;|M;!LGMXo&+(Vx*G0aW!f%dzOO#~!UKW|#_#K!3{rR^wX8+h(Y{w@4fKoyexBc^PDpvn -vC;fzH|8TaIaE^7TwKINi|H8qIGCp!7x(UXd6X>@8#T7GMYTb|xm-ao0YbHrPQ|@-q?&xi_1^^j4@K_l -P{Dwje&*TO>$gBgJfYNdAr@TLvQ8^ALEj?UU|^a>aj}kSVe!29TY5LhVCo*Ny);k@hfbc<%A%_?NgBG -OID?+>50rl$FvH^7EJFhc%4)5k1`wf66%Vxq7tW%Ji+FTEpl<<=M|h1K(V?+k%w=M$I5$ZG84kiR(ufOLaXxLXWX5B9`MHrZy$=zz+WYu^@2O-=EpPx?XYFbg_pK9n0<_|eP7JaZ7armi-9V^FW~#}wVJF -N-vYWl%#PigtyMSFCe)ie%j{=(U0v7hmT}`GAjIs^mp%j8?KQJRsD1xSJicC&`8?o{6MiZ@1Amf;V}7 -Hz{5am=KTS74xda1_LVH(+z=P3v -fn*KuWn(K+#-bj_4kT6>BX}4!7qU5eMDXm#HH{9U?gA1~3s*ge&d^CSPxPk`FF`h=>kRj|Sy0{*lfgo -S~o)8L+1_jH9~37VxO(mUmIyqdwlQ^x}!BImHu3fkN+UcVMJqzN{M*IeKyWK$}ra6w8Shgt@ij$+q;P -dSJ#LCLz4ln+R%dYWVCuCgb#@$rD$AVJt<=P{zNjSXyX@WcBm;WUETtUFr_j9{TNAaHNQXm5KP1T+CX -TT&7I29^<$$J)ok7cPEj@gwr8~uNe4~fPo0J`DV}G2pAva3-9kVSKur1g{DKYR%;2~U7w&i)M7*!)_5 -`M65d*2PRi_Y94qnK%i@3!S_k{{kp}{&f_Spjo}{M_Onjc%`t9YE@>aC4&vGEL3pWjLOjKx)fRX@jFf -{TDpHm)(_*j%UT;Y;-*89YgV$C0<27)kuK09rT_ZPsQtgB9=je^9K2m4)D5I6K{4!DC3iqGMzs2`6h5M0H|jEEJ8GNvEY6@8Xm?#S9kcJUBRf$Z}wbB@R@WFF3`K3k|Zo>Yg9X#UK@s2l-Wgs6d -GThfE~u<@C=^aFd*~B6R%Em`3f@ilXywV3{I|skX?=V -=s*1IaAchVHwrrZBw|JM)B^S?R%(@*|z>~{@+1O;!CL2+uuTr|lbIEqjxMPWEj&=f|K6pmpOMG!Q~kS -OzIJsv@J5+<@yA~)fjUH@j2V?;LDL~0*fuzI%DJ2Kn2w4WtBHj&1CC>fS+E-4PTGm-GtAiX*`^kXc+? -i5!aw>qzIA9@h~Ov@+O2Ny8$_Io`T34Vw7Z1(C1iG5hV?g$H?|;`-nO4Wy9s!9>5Fg&d2^7zwgNS_bJQ!$ad=oQAA6a|B~Hq)gwG -Cp7WSuRLE}bqwsF4KfoT(z|ly+cZf%b#knl1qkh9*TL|?nx2hfxEzm9hWEI@agc)mh9^a<<>^{!C&w% -#U-|G8gtik8T&uyys<5&Yu{fpaD4jgj^qFU_sI}j>zjcGP-OK!}}q_)S^=V9yZn&-PPszx@8*=N#&F` -c_8-E27ntUGhHx86qc*J{t-O3&DIAoNa{G(bPQ7cWVirN8B9HW$xpha{NC#`t`GV`!T{=h{m2?CM!6Q -k5kE>78`K0yw~HY8*@eyWVsp*ObGFrDwglGE9`SUf#xG}nh -ml?n^Ht7z=aU}X4%AVF_1_GMF*#tepg}ghm0u$Z^pW-lVD&kFCCj>MdMmvL`d~kXaz*2z;3%+VO1vXu -tcn|S%!E5t@^TLsjPtm3jYY3Cm9@xI3xh>_S7nvo3H2Y8QWciSZK^<{q05!^$q -rpBI*ncy2z|H7e#W&eQgbv4kF>13$@&j8sJJTyBub`tcvOT+Vii?00s7}_II^seH6P&kf|-La^=gjk52n~ID-SbyFWR5&4-A -p?h@s7;9rgx`hv@h;3;RgA5*2HbE3XTK;zzh3QDNM!O>^DAIup--rU)L -97w9Y&RJzM*Hb|oGOk@>J-SZBWDFek5LgpjYW@ne_XY1GD8I9i)@S|Ebw|w7&X$+m0qjZ=dL1TkSW2WDDOmpJX+~I~m&F%p}U2wraB?dG -6k53+cWc`ir^y|AW{0BfbCgYkZgOUtuZ3z%Yz#G9?&|!OYgjK#}N(A!7#RY`SqMJyhV@jKdT| -bn`fl{@Qk_~hx&>js{?&sG-&DET+ztCsocRIDtFo8?_P0@da}=&uw<|9#eGz!&{#Bw) -z8$s)y_iTucjf@@%zxEUayHgM&9;4l1Fo&{dTs_J`#6@}sgVDT3Niz_B~AKSAE;n%kE5huQo -5NRJPY3{MHz@)_^ucGorW{8$nh;lN=wt=^eK%^4Mu11_h@0N)~1f?SG+0(8J}=jI$-u7y1?Xlp@vb}( -}=d>jsqo;a$SHAcfx^q9W!uu1^Z`%s?bS~{0Hq*$+8v2@LRQ4B?{iCLV8W85?+5`*eYap%a_?~f;cKB -JG(GiP8Db!z5lBpB?l$Dg?_wmc?PY-c25)+HS15f5XD>N$?Hj3q7QtZYL -duH=yT+~d)yCr01b>{KdC!gk;@xdXd)a}E7ANFDC1#P#w7{y90J{r}7QpjG>0H&v%;S2sF~?`pPLk>v -e^Q_Wm1Z<0@y#BcH{se80!s!Lzv3Xd*az!EK=Q=_F1}5LG8dRHZFvnbqqq-YtdzSl*qo8wGpJCqp?c0 -`xSS4}>qgh&XiZo^kuT$%C2zo-D4mZVGewl7J^BMm+3t;bS7+X`mlHlF0)?aN8<-bw#fH>?h>BbB>iD -YZ50l8=fFRB;&OJC^#%B!{*VEAR+;o@O$F*D^fgNAWNn%Fl5~tYmJ~w)fOe&Xi*@&yifDjFU-P31nm9 -K{)*ieV6UW*g9MH6CQzdQnihJToj0sNg@cSBoSC)npDyW*rWU3&}UmKrwO3uOMiB?H)dGQN)hv_Cp3j -#71aQgSIFs{oTtuv>7c$0-93B5)k2>-AVRMyS0nxA9}<*~>N}p*>-W%ME1YPr$+aT_LC5=nG6hqy^E*_y0vmNI?>xpRe~lNWee(#&Bpi6a -l@D4H0BanHKZdgCj{d#)M0tX{i#1nbl#MJ%bu09mQS}V^%v6 -(^rO4Hl*kVWf)`FXLOh(acYY7+ch4=JlwFH2r^>D()kKn(LRWye;E;eGN0e!YJUwXr16?Gx5@$lmqdc -VzD#-Zd{d#G{Az<^)eNS^-U#U0iJs;1o01mxFOvUZ`%W(3TR8`Ux(Qj-N#7Vur2%7jeBY7r4ALYNv&2 -uP*aYJy=J+r@IGZ-OO2>5S~+<@Ny`YF`>zjY6vF!aB?{$C*~{1y7#@)Y}|j1^pN1IL(6{BCQWU^e}uO -$Qm@`5WOr)iL?HaIs4^!Q_^g3CNwe$ZYJv^~B-c+Ypf(Vhr~klD*Y-J^eQPjQm+%W|zWa(5-bAM%U|Z -B$IHT3cYpZqWh5Lk1Iu6=fuXIOgDWMVsDSw8gr8^m4gPGRtdU~{~-445tjar{bS)m;y1-S=<|uNDIm8{1#Zwibb{^GQyO9 -Z3RP|V)*^V}&w90c&Ww0^t8Vz1y9k -N$Rnh#ATApQp1RMYxKd9l9rK -+!2lc6HeY9#rm-gg5l{Fbl4U>RP{gz2;8)RX>I1ox2$wJkShe-ZXO{)^ohnPEouU?F%e`FXgi@MDCqK -0p!9rXO<2Js-9h<$?041qSb1%wHzosFUW- -!V3;coiFGmte&Www*8?hJVwj%iT<@z$bI=3&Qc3@M;7qory<6bw#KtKo`+G!Pl6`;LmYA4Vv<9HJ3f~2Ovu`cOhidjinMjS -|RQ|8?t%GdWwjW9tsa+h2n3dzRl_ao6CF3phymVxqk -!fuXt^%XQ7A)mHIBfY`(B2~VZ@Xw&nIzY<>A*pyqZExV2c`ZmB#o|4^8>p9Ep_}sxzzgek)d73l86Ea -=rXCB;C396EXX)Mo=Tz1LQR;Bg#$Me`KXd|CyoJ)Bg%f`nMg}WF1V7U`&-*u{MumIu0%6-0ySns)B?7 -n^>H=m?&zZkaa)6Lx+gd}~B7{{@rSO(m|elBvt3_KyxaJ##NGA-r}@SizNOS{jXEmBU>Nl9R#+2+Ts@5`(A>_(XbtMovGWZAof33%~2 -F_cYGZOJQPm*LNFnv%9OEYeIU-gC*^o+8?RC0#O?96#B^iNX1j?hnIlQpjYIWn0C8M>JFDF~U^kbVVJ=pzr*k -yRG|*eMR$XOpomq<}$YFY0AS@v-!|X_p20~WXdm_qXHyUA)&o;txY@@X80x~l+{6phv<4daXF-2!6pt -WiN3%;?x#rW?d#3S(c=ngUEfh)V5*t!0a;IdgqJ?_@}?*0|07WLQ-kyWHk|#RzK{O~(*AG{{|IYoWS< -FwZ@Qw}G+=bgfF{U>Y~gt8om~O#CVz)+KruCHfArrFyO{GP&$kh(iC}B(U2!-o4@NXfUap3q*1b0~x9Pr;lTEo%~+pJz7 -#Eg%O>((Rl(KLgohiHYi8)+wcse8|iz9n0?&YaI}a~8CFEfe^Y%oE>KeW!kIdC;RU`&=xoAhS{ba@*~ -G+kd@l;NRT#$GG9|ldphk>9hWseSHtI@4TK{9oJhp&O^n!?;zXjJgm2#82jxz;^Lv%Hu(6xo)`w=J*7 -*$w>}ts`Ctd>u&{UKX#b%vW^ao*DH)(A`8ivW0zPFg!LKD>z)zpPa>VjJ|ZH`ik*3&D -w%N`pZS+&`Lk8Zmh%GX1WjNqnXY^KAF7rAT{yjFmyb-7B*-pnP3mTky^#B9(16O>qqu}504aQbQ%xre -KB7s!9nk@wM;(yVm{hp^a9{<#Xqsar8!Uc!-PU05oLGc-u=)Q!~fYAlU~GH#@p^y7(wy?s+-A5MG~py -v<5*A*(E8?37VNF#7&OLn7KKJ>x1xa3*`DQ^&;!?%q1ei#zafq+7wvhh!f9Z&l`!{#ieZGh!A~O&PDo -J=sKwiM26RuG1PQlq?eX=c*Xd2v9FsUeNY_TU7M>JlDF~jjex{^uOzwgDx*CC!o-`zGs^5Irhj6`8FDlH$)58-cOBKOR)+2KoIV7^g -00Gi?wy^%&bQnuajTVwH`BlRAtcy(aaKd0BwL*>+8Hcyx|RN-e@st^!Dh#iKhhHhE6e(uI^6V&<{Q#A -8_^42%wMv-X25*^U8X=W+2+$`ZRUCJcdTvrLm|ul=r;Tog##ZpJYGV~p8wj7BFwLog`AF`W6_;I1iq| -UD_K}Ig;_hS816$r;roqF$z8t?_(^yy(X8wRhAVg%OEGqX^;jNRQ+d}rtk>@>wL`w~so{PZ3Np)# -#|^hxIQ0Cgj#7V?h^#C;1xuI#2yo3;ENk|!aYgID0_#!x^>R(i*F2H%L -(alCp&sf?p^6(0Vzosr5LZXos@ziM;9+Pd_)uuctj)49Aau#nU<-~|{3utj -?f@pOTm)+X-}BT$l#QLI+mgV2W^!_MIPx01)ERE8cZ$3ER9-{ESyRdbcv|BE1M8L}z!4_wVaMlrJWUQ -!aJW}R+Oc}ZBTe0<6G05)y)Nkb6|*|7QweJD7jYUfhAgocXz}X_imR3tHUY|5mBDE2Q`DhRUPphVS*S -X3alv7_11p2ik-^Nx4@)^PWr~FG?fAAPK;hyi{3yAOt?n^O95t%H^kS6+Cj^Ry;PldcPB97pI>0q^Gf -zSg2QBr%~#IJ~r8KaS58HHa^Hz-6jde8!WzqFrkKeeRfS*2A_{4|;HPxKs$|KAeH>xJVZ`Xm5NcW5*zf -dpZp|i;+C2zg)FD_RRW&-nI~whInrC2!Ga32$ru7$0IOa26~ICQSV8<2Dq^t3YsJnQDN=+D;f}A$M)Z -AKoEo2)Cm<<1S`I?^-HsG504q^F#JX%2L7f2kqNAAFkTlfOD7)m!};h*^*uW+ce$5D#FZZ-D4(m~2H8 -yIpMf#V--6hfvLAy4{Ca`Z`Q|Aw(Xmq}E~1mCqJF5F{ylO&8TKbJ0j0(Rh3p#MNng?dj7!!)Y!GQFu; -J!4&0|YZFX*~3!FYnC$5Smv$VI?qgjspR32ry(cbsr$Y+X902hOO%lJvXohEA9*_$H9-}@dvGu9X*Soe`Gq%U -K8v;RlE>w}ARF?%}(o7#y+31v{NZumU0Qsj%`m4KX!>+AASg2cF}^s5Dx -eV;(5At57}Ob{$$BezJm0C`R*eKfa%@<3m(zh3asS#BM60F@%9=VeI73DCR@ZXHRcC0@|g(I2>zA^P& -^WB?S>aN}QL3gu4AC{4rgZv0Aa4RlpEZO*uH1e1666opL#$-Kzs|BkP_)e(>zcJTfy?y5Vw83N4OlS4 -C&Wf(B}V#S~(e0+s|E_XE5a(X+JgLhYFm=HxEu*?a!eIjh*5`{eUiNr_0 -NN~FRZ*lc$|Bu0|8jBVYn2X>kEq -bl2??Vzr)PNNnOL#7~Am$vPoVcdqHOfAX|V*Z7sUn_@S>`QdWpV?CmQXdk5Wu&UN=S7$wnN?tLG^xnd -n+7pP*GO|x{3S=Zm_&t@TQ!T3(DU6I|2vf!Nryy7^L+M?QRk|w&vvm0v(-3N2-bsrMhC{Wvc4r*(x*= -8s0C)wms_ZXkq)wj}roP{K28+1a)Pv~@Y{fnkb5xTHjPfBd%(*5{jcAuXuRy{m_?C02Y6NHaIWaG24x -n$lE>ujG6rYZO;_YJtrv#wieX5GaT;1t|ncG}sr^JNP(%ljI9Sv$LDBh#vfZN{bWF=6At1>4ns2tfb+ -)whcSf9~`z5YYMY%8m~;uL%fPIBl3!pEWYI?YgIkX1o-Ha(OIv>=?&TD@e> -TqO{Yq6yUh%?kFvG%gdB78_R|;cfKtK>mS!IqJEggl8)i@S-uHar@E&mBiiNYH)Vg2!#Dc|oTC^GSGx -}fowLrFnbt3*g&RZWnsHLuha@=-Hc=C%ce(Ub -Q-~h7p;3LVCZk{;pg6d55oAJ5RT6X~Xxe2+0ug -K_09#P>0j~eLF}s+)9_wZXsSFsmf6mnCfWiSVHH(uLqk48&Ts`@OZ2ruQYAU;Zj$TCsM$CCyrI;qH*h -#iVQt6iR3KO9$YGy!RSnAoj2Z65U-w!%2AiFSl<=$^^C9rnej(*15`yi;nd6S_|&<9sNQ=M=`FT@4l) -FbI#;FM)jLXJIC)QQMksgIt347WP}3SX>VX3OZF$-_!!T@OA-KR-6lPn|RG?SU^EzRk002)!OYGN+h693LvX~8F*xE3LXHLfI -edK9D=fh8809PVICG&6B@SVyGP|keyG}|?k= -L_sleIRU>mg(Bs|A5E>I80D^F;2DnSwJ?pMvX1(~u%4;sBqm-v_~x=p(Q^|+At3vHwTMal0f7rcS4R# -Z{uEonwzUJ@`hoAtXQwR3Ok=|rk++YQSo@o0UxoGGm+@N=Y~8gP1?$Kzyi%G~?NBMHNiZzW?UW$&pT@ -BF}mY3w%?GTvB?1q%oCP9kz>vXm!z -ZN6S3CtN}>_9*#PJ^-?%4ATlU+MMpevm?Rfg4Iyxv$ra6#CE`;#tL=3?jv#3&hrAYvk2BsiN2p$wwxa -TGQ1WQLSnk$nYf%wv?#IlJJ36hX=5U@q{V@Ldo^;GV$Ew>$n4;FH8x-}>8&_09KhU*^=YE3<_`W{XIywTu_>9bQv_= -i8AExG`-aXuw9crqJK}wuO&v~UA&Xo68jt2O7`UcF0%!vTk@YwHu3<(H!U%GhgK_`T47qS>x)L%cA0e}wkxwBge4*QEkn=O}ZkS_a8-?^?k*D@;8r5BYG(^ffkocBZcNEjP7TcvG{Z`1VzR -xsb9zp;ks8iLKb9A>wJn3|vUYS{K!Ci#S8w49FrYg~wjc&Poji(9?L -;SL=`+F-19E>=GXb=K)%Iqn^UCI_61;*QFJ)b6y9z;RGaLU&eAKj>JuSLvWJKXBQsg^|O!Q*_|OM!Es -0`!XuY1Pm*%Z4`3g~-Lr$~2Yo*j$QfWTF>>n$A|DpNybGnTSUhAV>g>qjD~b~PB4W6d$_HXO%8Uo;JO -{0xJs}fj@UrMNP-d&CIQ4Ru!TjpxSvL44r}c|v17K^z{SE`jy7!hXm^#mDb<#&(tfLnt>t@!J)kkZ87jP0WCD9`pF(t+obj3dNgvn?Bo0OcH{%pwectx7yWG@VHNvxc8X3l`*< -&^FSewtldc62>Qe?6)zHEGptz@^4XW79O6#tWII7~-Ja9JMhUn|09b(ka=X2BQapEQtdkF831?W51NW -)fdTK_5G9pRYQHSWFX(5Od^3}Txv -5~W)b01sEU@Uz*N;XYgi-HX=I=BRLZuV0|sS=Io4^>^uB{azi|>NrSr^LlGYPh;DIMEF)h-{dHn&F^e -_6niiq;2Tpj{O3-f-JvYsI~4fgA^C$=4xl3?*L?k^kBas;%sl_kJCx<;4rMDFf4A*_+M$&Hre=XU&4N -a^`%3{bw=(KvQ5JN_EaO6(4lF@uRxV@M7o`4p%4P&rxk;ASaP%yNxqdx3%%G5JiQkW*4>&MXnwLz{?w -Gn_O6G84`}?4;aW?W;8MgN#%HEbhL=)K2| -Os_GEXx|0;N+~3d`%Dj@mTQg>m*yZ$pAWT!mQX?+Pd16wt4W*U^vsh-m7=jbjl@s-3LdjGD1E~y;@ko -21b9PyVY>)@@*dNfGQ~Hpj|v29mcczigbpWe0AB%4)vWNkPxRS_Zo#v!Tr!xZgBuzEeWBoY8#K5?JqD -@HF|ZU!TM+m)a7MW9!02u?Enwyo_JWQVCtUBRWJ -3mhA$E1BMVaxdt5|IGtT1yKPDRqHcZ>Y2gI;)uxXHosPi(qL3Rc(P()9m^E;FQlDRYJsFZXiluY%Jt`OmjSRpg#8^IMUE9LqQ**M8Q0d$V0bk|{pvIS~HVNFk` -o|4{vsc_RM>%p{_iKxspLKn=XSchbHRM_t&9i|LPNLfydkYG>ELjbNLtQOqFYAVuF$k}U0l9=mYfDv( -6A!sV9Y@bf=E-*8D3onem7ajS&?1vAj~Y5wz}OqKc;+rkD5SAS<4kqzYN&m-`O)fba3kHFS5KWTk&dy -B5ITV#o-KMCHaT!_O9X+`ny=~cL?l(Cx%c+cpT4R-CNb90r;dn -TMBF*!VKW8IHwNHV#KHy^yPUZ#de0Ah -ufa9UtU)%t{+VE~FG(NUgj{^XFe|1=sST*a^v>2JHXO8t+It-${TRRf^~~Ij<%e%g(JN%ZqHkm1)_(- -OwTY1UcIZcs3z}?a-Fj89tt3MJOeZ6GS7HwH(t$Ng*Qj3HaM^QHtA?sG -$ejoOD0`-pO4t83sNrHy#IaPW^fh#NNlZ;#*MAjG}7TJ1rXc?{5`NaF0e+{zzGeJt_SM7LkYKpO-C{( -^zt{}u!NO`rSUVxSEI0XqhIbL&p(cLcRh!qytf9a&grIlguuw9ln9tJPq}Nk}Nsm9ko$0$qf$yq$-XT -q4)&YB-JX_1jG3{l%T%kI25{?P|BtyfEyTF~lr>zSGK|%ibTz68JZg7GHR!9vFtn -MoI8A-2#JA}S?JU}u*d_0&Euh+%vwJI8ysHtf``|l$H~PBlnUb4k0*!BF&zRhzoAoUIt~G4Y4oyCSZxdu7WI6b+R;^ho}~i5tW^D&rP8j?*ZWb0&v%YCL10f~MC1NogAN -D$UG+Uh~m%!`!B4KX>qd2D!(Z~J -^pr6f4kkzzbWl~vm%?E%RP7&zx_X}UFyerFASJte=nBgH?F9P`k7R}P04R&a^F;8%HsYT+Xv5~=t-vlxevf{Zv%+^_{Rr -Ej*oP2)ymwIGyKvl!w^s9x2fGwaypP-6^ICMc<ezNqXB=@r0>@g -5Svjqo$abN8SFDPhI~7b1;ek6)C+tPZ_l=2>-Wmit>=3C;t=0B(_DWnc-|hT|KY@s?J+IPUVu$dM*Sw -Ys=kGnT3YkvAS0S-Zv+`w-8we@vQ3qbx4*h6VD(d$RSKTUeF(jBFZF7mc9GQk^PWRg%l&n+R7;+9_M0 -T@yRz0b=>xZ_09iOaLa7=e)PBlGzmi3x3ea6As0*BDrn{$n>tcy&BwwE&i5h!-Ja$w1*&i8LpZRq*+5 -S^Q==8ZrTU&=z)qyjtC)lNFHGjRT?>-(MZ}Z1G0p>RU_@I9}wZRr`{$?B8ZQ(xO6>a@7O@4cv=l3T@? --{T@r@6n!TOsUU2*382HLGMt0ReOI&zy|qo6abaurNwB5HaYLN$QpD^T?~BPwh>+Pya7(@73idwss5O -^DFi}-*-d~J;r-M5+D&Il7KsM5J@0J`1J>5m)mxix4YlH$LZm=T?j2Ll~vE0^9ggdtYozptiEDTY@rC -b#EWGEyvpMOYoNiDc~&K!221Pf5T489T!q+EN6yAs??Yjwq?nILq@Y*Oe5^0lp$Qp+WvvG^bc&`I0W; -w2uKfZ%BlpQLw6z^98ZA(2b{NH#dV!@t(D|Wmisk4wQlBzNa#Bu+3wS|v9gnwbI$_V#nnNg<(#Rb%F$ -i+xOp!62r+5N6!&GJ4>O}HXyB%lN)pkO@to#E|nDVNC!IA5c>cI{O`n*Jn2$A!w(>GHc@bPx1=}aM$$ -Mq_hZ556d9CM7O_+r9fDZrdG5lhGpOy{nO5732>#gUsvWthfW+OLcpKigPCnapIux>eMEaOAiZYYFto}^jxGlz(@&bsu}!{#S~YvVR0#a&`yd -5qmMDR6}L2LT2i|SgzBR|gWOR+-(>V^E@e7}mlTjM7aRlC4ZSZ>rRBOsY5kE*(P%85eyPJJNAIyqu0D -K}rZR-+`^J9^eh>OO**``MsMd)W`;MQx05@D28;B&YO2kM|l93%v%|UeGVAHYviz7d5t$2wRU*b03^W -I{U5kQr~;*8xLOdfD!`1Q3>)|A#~)xJ=jjJebeDj&E}gJ+cFr+hpW@ZmT`r=eX*Tjf%qYSGK{1u>1&+ -I3HH#>Hi)jw;-Td})(^(BB~P8|Qh;$uYJi{mZ~VtY<`<(Zql8`u5v*OZ;oUOE+Z8`xmGGpZ`y>_@7l@ -|0@fApo7mB`yK|+5VX-Vn!q6vgJ1-LKXMI~S<;(9BMm&>bhB@AM1x9-K_*{S^peUvGQO-COln2CQMaXL>d;u`#gSZU=qOPsTetq=KD_LGM -Hn@=-9O)17i{>~F&}yB7|!XN -W~E5VjkSP9FlVXFGPTHZIA6ak1`Ec7+_%5h~HNHFg&@zdDX>3beT$9R7BuX&0rKRY$mPWbLbA)Gn#Y` -jW9rTzpljdq0U|P0jow{`!@(L4axrJ3)h0OPw|6D`~Ujb-TVD!*^JB+8?(3^NxXky5*mD4E*ga%QE-b -dG|#rGQw@EF~eis+vCepp!1JC(}rtpSaZV>q}d3qz9SUWx(-ZrG^V+sZ;mM -}>9HT6W!j>W=vt-NKB7@@5+ciQMI+PC;Z&Wum(mw5TLMCgJ^2m`6Oii&Pj_jDD`c(h<={Wxbx>Zf&adQ6seJ<;5P-)iqMuu0hDIMUc%4XC!1U -JkD#J5u4Z!2+WUQD0O^Y`uFj7&ozDIkA(N8i{~GKGgO_Mow99on40YMHm76Q8W+)lO|DQlwKYcY@Rw= -ATNq;;39#{{Bxq}#jK`AGf>EzE;`v{hg(5l(z9wWz?u!JL8R5CFX%K%LM>qPYVFwANYyj$qvuY}rDyH -L1=|EFrM=?AO`57lJt}I1ig!RkAv05AOBMXYgiW7E(o+|ei>!(qd4dd-k~wg0NhpD`dZa7=Bwu;6FVK -dcMKR%kQk0TWSq%kWNuFkU(s&|ca>bpr#99Nv9xcOIveFCnmY%4-`Mg)7C=`{+drJPjz!SA>xD1rY -Bp;*$SdP}&;a~4f%!>e%Pw=XJ#3gxngJa(w>PC_`DO4|51BmsHV6q^+G2xACOY>so{2gs*h0M$+c(LWSvIN50l5P<(!F2v#b(H4?n(1^3JLIjAelgEpvBFW#D{8ge -^EVi(<}Y8eB#yeLu(UZ&->CL*qkdd)x)YeQ76GI5(Bd}8AGP1@b_tN={^Z -GeDPBaI`>EV2po_>Imv3Sq}5gc9_u(1nbe1r<^u8H@!%>_s%Ssk3MvYn(AWEKKy#%xYtnwy=WL!PQ%B0&_y<4BfsAO3Ofas?} -3fz?%;L)K^1>el>ofMleLgI`Z9yoYIboA);DQXzu9NZJ!I^GVJajoF6e_R3Az=U;8Q|Xb{gSXbQ*uM@ -!uE(dG>zC@K)V&0vsNt7FvC1Bk)&b883iPr}bDH}HU{M_y-618Cm)caux;3m_m`|t;b9`i@6$DIovLD -s5yyBwvq+^pg4l0D?#gX!ho@^&mOu#O6?-P?cq^KIUmzF;l==C;9Xb4ex7rmf=;-71p%KwSl`rkXj8s -_duvu9Aaw(r{h{rJzK{Qt>9-_-g4<|5w%HHso}0)r5G?`}n52*N(i;84V#xPg(q&1kpjgZFGm@ctFOR -WHJ5Pekv{qwofdJ_TwaxqA<8Xles>@!NKG16L%u>tVpy9~AWNJ`HyG7sh`CZykN?Z66qhziMN}cR;tn -GMw0JXbJR3nCu8Ld5a`R@Lm$!fEV;$bi--^wP$L!>%070yhA!XeGAP6J09G7kJ3HmzU>1IxA*&7poZ> -UN7OefLOEHM5Sjx%HXi}d#r_h(?&xZiUmSA< -?7kzT&)@Xo5mygn5A&u{9Fmmvau>rnd8CI&utKV?lD`k2Y~CUv7ksuiuerQ%-%-Gpz~fcM3cr9R31>n -bySNj@>t(GeZYbJbbkzhtLbCQ?}_e*rrv+qe8@HmluD1@>$OAEKve`~QnGPbthQM}5h1Y^S6RYpPOfF -CaAkrVH-zWL-a*W&(cM9Cf*0&_M0e=yg56wt$G;YJ9=Mu}O36aC>$7N|+U50P``Hf;mS6G+Yk;_N?oH -J9e1j58ACdVQz-Gx-%Y@<3;<-nn$sz -+&f%Z3}B6BO30#OtRSym0M2 -Q?*^wNf4p(G&|}+acXgmm5*DEz#OHtsakrK{5l43cyKQm~(q%L;ngcI)BQx&q-4ehm^AA`Yd~jaAJl- -_0qXmv05Iwct;77bBf4U{pmUpV4GNUMyVhw%7#%~BUVEfKY>qn16uZK!v{XSPX|9f_E;X#gvq^%%tIb -j)tw%fS3v>3a|_Xztreyx95_4Z3a4AB({pLo;^u&EufJZ(P*Cvv^e@~?rSe9;BSYiO&E=rqYXDTcxkf -?jP#R4x5I%8O)`YpQkpE%PsWYAV?!tq^C1EwE_@V#3L?@+bCkDk;pK32)HmywL*8FtpgdhYl?|Aq25skmQqXQk+m&kGJ9pG&rs4zN->1icpU21Z&um%0H -d4v#20gfjJatec<@n^Kq60*oHlNOLdxr~MeNf?mGW_<<>|^jK5c1hg3@@(@151-^qNDs*QBq-0L{zOk6gx#$w#f8VAbc^{p0*x{a?eXOV8TWG6G>6mn%Hc(nk5OG -xSm=M~S|ZuoIQ8t9baV#+<^j^!xJDi9%jUE1_+zFdQeld- -4=Byu|v*5unbGL)HnfE*zcTe;+ErLDWfbu5d!C-VMZhK9Rh*?$EZ{%WCr1P!-Eet?D;jgtsU!6-tL2u -xterxQ*a@Y=l}u)QgPO!wYxD&41(Hi);k9i)4?3?_HeCN%hi`i1%qx}z^L+Sfqd4$?UG9`6qK9ip*>9g01J0i2>d6oa7Tr}e*z1 -4R0wB?Ho -|5;&i{zGg0QK$|-sgnI=)*Yn)}DQdJeb;=3|H`k1UmSwmp8YQtskwLM53hfA;?O%D>tB0>jJ2{bWQA( -mdqIr9(G{pOJB^YGv<8Nekq3%)c7BD7N7zJ))|8YxEZsuRB20)%1Q{q?FkO%mj!)IRbzz|U15=}XVZ# -^<*WOv~8WQ#?j1-)L9MDzmc*g10*qU2{T(-XApJugRWYiJu)B0=^ra$X~pWyCooW1Wn2U?(zD>)A<_A -N73ljNWQYgNRm7bmKuvUIDdV{QArAqOr9iiAvjQ;PM0SYdaT2BcM!-JXvJ?qIS@M8#W$s#?n76>AogM -;PRyTBGk0NiNby@&{o~*)IKOAy2J*K<^nbA2&w={yE%kl4CJ>B55E7#?6i2q-Q2f)r+2lP~gTi~0D;( -^NX38yv6bq+O^W@a4(af$gUcaVtee|>n-u#oVyYGjmB>)ZGjzm5B -bo)3fIU-yDafeC~Ri{^%h0j0(kOesEF9(Hu)AAA>h5zvh93VxlDGhj-)&1kKdJ@o!cjSOYn9C(C}W3! -K2?Yd1CLypudTw$xiQet7THj9*I{>(By6mxQ3siGTqbmJ -}-Bz?YYZ(xg0IspTKs7k(+pZ2^gyMeSD+=$xL7Oak#$VSFaO?OQxYL)Oo@U?ZEDQr*tW}do7Xa4*7^F -I%bab#V;n1CC(*ldIk{q=_3!LOO-v9Mc{K=X~$xCp*-FeI0^>Ze9@>oWCe?4OzKonkE2L$VZTwkT)Rd -L5Ym3mvC?DgON0ou47U-SJW;kkRPIfCX*tqhMeCUilpOqALwYHuHO@R{6zV>@P -dvHW)nPam~a!x+ZoXqCV}zd$uzgTR&}7uZjKqYZGjcmP+lYCY&c;bMUY1L&US_7lngYh`ZP|NEv2EEj -hEMGbDqSq?$aCro4*%FE>wo{xyDCFV$qt_gKQEYt77HHSmkPU%|P6y%s_6A3MwE?_bsQW -#+lQ#5*{iJT#{x;SDfn&xR#L8aDjNH?mNH1oJP;#Ua~vynfuOl;U%dbmGr6(H${*x-#zkjX31W&G5cM -4%*xjJ`aQI-$COsfok_ic^;JU<{pjIWv)^`NH# -vjV|ldTq4xKIVDBA7Ky>V9HwcB@Z0>Tc*~{4pJ9BFymX2w8{p-X}&s*`yyOS*S>wmn;08eU+yOgI*$y#=UR_rbBcT2maQhMq#osdavuNXR -Jx(v5+HEo}^w@hUN~Ppuu^idP&xfoG~E{F45WcGYFWppxc}qFO#YsshVe)T-R*mT!84rbVjWR#$mzjr -pl8iXK}-`5!|pHTsoglFb9!9eRgIDzFNqQUupYUTZeuo3&t!%b8yJx(Idnmel#5TPbMAyE=>C;1pj}# -)<1=gKhb;?-e4oP{d*H6j^Q|tQ21xK5vO;Qh`e1JH-xu0xxhPgjG&zqZ(oM=TcteOC39)$Q{0F}d$TM -K@ANu`_X#KD4IN=@2Xqm!tG;btBIsLW7yct6_paP+Z6f{^Zrs2jvR8EW5wLjg=iS?dqWwxj?mGv|CPW3JMFH|3RUo6?f|BDI`r1J6y)$hNr03Ds;Z&ramq5HsJCOpoqY2chE9$v2rV -x(+gb;{znHVX}^JBwbbAo8s|GpZ#Kca;GgIgAd*A!U@D^{O?BSZ_gd(1+Fa$WoasA;OJ?1v+&m$p%$e -m)CUl55AJY=VyK&fm_1{?1+HWS)7(9b2+MB_rFeqT)8uT=S~{<>R{vI1IyFsPZxc~$ebm#lUChm+IF&~cdC6)^^)n#H)w(8SSnB&5y{?9&lElecYO&q8z(F;6WF394KjlP;vmjg4?H -rl^q*O0fZ5UK$ku`hM<9A=!-+%QnZubBG%OHO4p#0w~@U3I=^E-YVYJ*9VAPAVGNfg5{dRGJ4f-;0cF -agszN`9&@mF^*RBgF}}Q|>)JzcaBL$A{h`QM*7$ut(dsMAPT`Qg5sJ?RikL2gNP&1@PNz0^TDc{Co!4)HdJDGU+sR@N?$93P(YLrdP3>iP9C-_kzNb94d%h^ -9@53E$CC?y!CkN8MQ<(orz*9GpfMhUroiL{|_3&Ug!q2o`-D$nW3)zOgu@A#r*Fff%S04u(shaT5D8s -&I5K3t_4Gvn^@5F)1>2#;FlZ!zu92i0#6etqm~biK}M`61xROkX6=uQTw&#gbK3lbNLK>$t@nOQ -A>?-fgymTl{T$Z697!OxDy&!P?$}@G<#u{^2cb4}Ltx%s)T&XL%g>t#-)CmI&!`_Fv6#R6Fu=i8SsaL -dJC%vfhw^d3=VKGr^C-7CdJ29-@ctJgnj%=lvW8CnyS*!))ocam`7gk43l_;WiyDt995Bz5{+Xnwi=O -XrtTKOK%F7z9gb5FYM_W+NWB95Iiw+&CVZZvE@+F6ZCX2&o9wXWn<=nK#S;$+#iQJbnd4Is$Rq6Ydk& -GEU2YdjT6M992F4EYVEp3J+FulcOGqE;vuK!=V=0z=GDATIvF~b!ju+OU?S{*1WSw@K}>(Bm-7SaF5z -0)t;r%NT`K3(N^hsWo60xIoB)_p2^xlo<)MVlwMkt2`n+F6c6AaHYzWF`j8fC+{yrCAum+2sn8S$=P) -Y|8iPc=-PKL?wDvLaf*S62#$K7496NI!k+Adzl#i_BeP#xoXzdDspl(SXbY6#bF*;Aq>4*^U!nqHEC; -z2mb*Bn0Rets-z6pm}7!mj5~yKpN^3bWg*IdNbElI>ypC}-osDO&Dl)nmANguneaT^RTf -HNjbi0>O(n2=Zv_#*UlvA+lF4K`{PyN}GuFu?D!=Gom}M2N>LPL{=|QdPr4P!t;n)&C7Ox-LAGGb)H3 -Q3MVwj>UIsh3CAz#Z7tQ#1A2ev47VEjHuBx0Is0L7C{M3Q0Pk8zW+N)beQ>%ELX5%n&3iawdy!*`X?r8=qgw>}>2x$b*fqf^LJPqcpMX4PG@2Y2; -}k%ZLox|Q_zCOt0_?j3ab9)OK~aHS`tzl)R~sCzS>7h8a9v|u?Fksb@oBJ>nLqoru4?)qCwwWTy~nN^ -;%va_+*g{+;ZvaXm3>%vSJPwaMGL%O)YMZcftT$uVhwR);lyPi%ytahMm}o@Uf;;Mjo9N-9}^nGqy}{ -30eswN=is-_<95iq*DD1gp6@0Q=Zq2M?$hlAHgK7&hil{=DMl~6KEp9hRV$F7&dtL1s5MySd{(i$5SU -W80<9Ex^F~%vOp&D*?ukp*Kg8wf)+ocqC-Mo7%%QN+Lv{{>^US%?Wu&7uqK+&Kl^a-Ybf083dSvSwL9 -_N`9&3p}2iudn`=hRS5|<=PpFH%ozU&22>*SbFjSQ}^CBJUz0CWtk8tK6rEv96lwZtS?Clb>*fBEm1989RQb~NQ*?i^ -IE^pD*j%?C4IxV#(g3JCCEXlQQEbg5AWyAy3HI+X9@CC}o5AOaPZh)I^j0wBMJuK{vYni15cpo5Ch>* -=c;3z7gKH^IX$=b>;7oD3I^kKTKXokg_u-f>o0jmen`|cMWck+@c=QDtiQJBk^#A#IY{OSyZa%hS6Mr -GILKCj!M~8F?#%(r -^5eQhProLGL(>eyfn|=3(T!}7E}-3Wq9aKtCPLt6NkI73GsG8gWt6uWLG%Y0VPH3MI7usN}HlP%f|72EC#0kpyJ;$ ->tHXSC*TK!yUJ#F}hRnGeWPwB@KzAb37Uy<4LP0c{ME-ZijI-{zs46O3^Nw=;W7B#-isFo -ks?CqKss3P>Uh`T0#mk>A$Q|y|KR+)&i#Rz$lnGal8UAF7CVX|U~IwRnVAKcy<{fz)eh;vuXY&v}ygU -$5E6TU4d3tv$x_pN{)^E=*EC2tZKPx_kA*wwF(XG{fP=n(x1m5n`kG3L(2jfa^~v!Y8d -gb`wBS|<}%4GiepDjF?(o$T^Q^#kBTsQVYm@*ncwrX1P5e&L&r&Gct_E)voLdT)Vj@%`RU=gDJPM&z{ -FX1^@V{FLKS`G8?2|0++XA?`LF>i{^ZGz+t+rFw}IZwfuyJ=vY1TibfkOf1>A(mmqvRi=rn7~K$Mt-(;MG9MCXD`XmXI*%Z;HGd#&le9gXcRID -U<8BJw4`>HlLl#e2r@ry%>gdp~05U+?>Y?IIXS(geK=J&_o^MMxT^VQdSMB!xq}9ux{A-)QRl%m;e0K}N(HEnyWY0cg8ZooyC>$KN5 -^W;WHf&sgOzSw{YA$Saw>*!<0-;vMpMVR?_{fAg(?es;~!^q+hyoUpF!sBzYq7ebZ2f8+s8$K -J^J7PWw{VM|iKvh;sa4xSUQ2SRjxoie*=G2xEX2Zc#%zkb}8sW?lXEgCxY;sbUNXM5?+)Ui1(a(o6+{ -Or|M+Y~f#ug9#NkRlx()q?A~#OqG3x1q=#;u*N@?X6%hZ3lGa$1CI2tC$4oU -wbllyB~QBK2D&_NIlohL5Za_)HKHJs5z~W-qaPXgWD^Lc&VhyEk*uDt!#HT*TS6 -5U#fQO!ElzS0jKd<)z9cGwsuYTx0T>OdX_qfC!@aPogcz>AZX}%pizUMA%a=u}q#4InM#Pw4jk93P;` -O@BfQUl>>?)&3w;|KTk7Css9I4{>DGSuViNhl;`NL$2LT$L{gz5=v>4Y2zRORkjbR&FH%uV(ngGMf9w4kwird}==U+fYi(EC?udm>c -%$J?o;cn0K0#;&Hdxd1wGg^Do7g@TsmBBX?^1N6qT*vbVtRhW~%;Ba?(uU16 -N(58ZXz&fUoY{N)HS@y67DS>k0Z-&EkT7@yl&rX0ws#Rr`+c}PeOZ3!r7|J0t3?Nf~+ct=()TwInlTG -}YDo(~MWSax{nC-RwD*Vm;w(f$j~BO7@RBL!~XZU9bdG5i39V5XyG%X)kf9qiX=w?misQFF_+2d=r&dK%bE3RuIy4I7Mi_b32-Uuvij7839qWvVY1gw$r6sUXRO5RXd?aIc;d%e+jQCLr+bI -KWHAw7QogW10=BI;weK1m6=BI*?OZDl}VnuNwAc!6&mcHmNr}h#bPIIm{4_FoZvrOfgUxnlm9L!3nA3 -9;)?rq>5T-*VltnA;z`H=yk+4;ntyf#oBW#od>NPDXS^Uk4roW?v`?v*9pKrRLyvf|BiNI&~x;lu^-3ULeH+@jQ$ -f(RIb5R%04jaCvkLX+sOo{3ZB78daB3y6fEPv!0ccqd?Z`o|YH3iK -_aw9~nC*SsX(?xMToLcCMM1o;#(1?XNC2%(+pA#bxS|pxC2xb*UEGd-w`F5{wPU*olY9MxM)uA{Bo#Xz@dwt3re$@i;;Uu@P8p>B%wJ&Hp@Ms++}Rli_&AI~h`Ps&F{6olYg%;E#f7S}IG&b2(+4U*0kSa!UYJ#0=oZU{glF^<~R_l?}u|tuO8G(IDt=x6&0kB5?#X_RgBag -~$U*rmNOI?|-39E@4hr;Tqn6*a6)B9aGm~~@G&7F19uu|?f8-qn)m=VWw;#*1_repQYZz6@*te_z0sj -mz#;B!h|Ef4Yyw(y@f#+|R3yFh9yW#>Ti1^@=D*O@+@S^q*cHQoCMP^N663hd6*CigMtfd&nx4|#I0l -Pw&lv0*#qx-e;sD_0&!|BdDn;$8`SV@Z4(nW%&ZtWPD{sYfy6`Yo{l{MKUJ`?Qgme;3pEKm#gj=8m^( -r{P1bXd&6>PN#@NAim=C`MOKiv3YEGKpq299SfX5RURHVl+g(=jTqYif{v=JN$4W|Qt`UiSY4apgu|>$On4q<^9nd@ql{6=vdL{(PYIuO>f;O1@=FI+Xw(J#%Q%}LJeC#C>bE00h?i1%&ScEJq@V;4(ZeD#Y%l1FtT)y7tMPFL#?+?Ox(w!48j4kL5p^xN20oDHHHpm@`P>erk;<@Ip(dPWQP1PlMfMerGsIE6PrB1MH^4|Pyt -`{3`F&_TbHxIA@Zfy?mlxei@02E#qXc&345}&eAQq%%3V1;fv!52(x8EJdGiNFazhnJ%X3g{98RmwB- -;3Ce6LQu3(~&HElVWyah@X$9L#?kNxd7S}>5#3y@Z-Up -!NJX4Sse{oSojzf2S~VJ-aW+VV9;4}3{X3EPv^MJ?kmMyo$vQ;+SB!x4AiR1JAg67#^9gjFR3V)aaJG -!$$-Ulkrz;cX@+XBxjx%6S{M+wj~Dr$*$_PcuF2z5DE2)P1O9|!ye9HhC06xJ -)Ylu|F1vmwmC=`@ZiW%BQ#S`gUF5u!*9PR45!G(=paahZeHz;dJ3Q-kp*`1bQPzHavW_ho?sJ+oDz!mJXSv2CQhJm7AV?|rmv&LA|eb$+KojU@T;G=zPIYd}4pciU+{4jo0DHk22ZSfTcW(?o -wH|+91l1EMAKI3}cCiyOyYtSwwo -l=c)tr-@9b9{bb7*GiU6P;Uj~HYum>EQYUYA|Hg5<90Yp4_ZwxHZw5>UJ+LUx|uDZG|M=@H7@n13yng<-pN_PaIk1{JV7G;5%@-jkSEM!{7 --+)@^i?yGTFg6u+~vT&^Lm#O@IRiAQcbEI2@f5guE-^wk33iFqs9!ouV1Ptl#|BlodF#h2)Iv7=zedvt~WQ+Y}z9B5|?)FY*ST+ -zWiWS-5^UxY8;=O#1bkgX?<)^#=#nLVG*7Mr->56m&OnjYjX!s5APF{PJ7}de1P4&Z?~J^Gn+aCHN(K -S2-|K0;bQ4ss%p0U_JCSe6fQ45)+i#UlioeucJMbiQ0BuTf%Ku?Y@ZT^*|sQP{g;d?dB!q@cVa>={;@ -rZx(rTU*I$OhDs?C>%f=QH7jrmZm($+Ik7`Ay8P7JDh~MUWt*~Mwx|LS;WNJU&*%~g?O`-5$GBWtp9z -^9PW`)gQkOF4P1D`j?NtK1yMqU$8~4pD&lgKHL_l&D34@#D`V_;5MASu$TqN~AWg|p3D#{Q>lLgO=i> -;)m*K2z|Nejz#7cJ4UQzj7r#EL3t93SL^q9=Ic$1t@f)*?P6?Rb}`?Ae^wl3JqB}#go4B1T4+qUfG_iDMUvEQ9y~hR6B9SJZqzEMZ -p0@KmS~S1R&dp1(9<)%$lw)|{je8+^aaWf7h)R#Fko}liU(I2;!$5YK?^oQav3Y*I9*};nb^lCNV4nI -jv7J0sN+u2$U(t7=xaJf -U}>~=ccsnQI*XEma$N}zx?rS$1pd%0;HZl=oDBN8`3d3VcFhhbR@Y55R_!ZCmvz37f&bg&TRz_k$?>3 -*=}z=ko+z{R(Z_9{tb)MUg>ohRjfRN)g7vg>d_#XH}=S25B_B(QNGKr -XGJpnODXE&nZ!E`{rarSS^f#X{uh?}0JEPg^#h(oaTdgh3O3rZ>Si;J{<+7qrE$LZyF)@bY1l -Jj0nhmR+oPF+lIuRH{272yVBR^y~DvirLjvVqi1v(J^q#vQ@f}SDZ!5s`ozRQYeLiG+h)VvTqF> -3fLwmDIkiOk-h`o$U2k(aDcvn!x@g1LTuR*^x;5U>VeX+)U-weGm@3sO#>@GUgyT3T1_FwV4-IySM>r -ML}>1dV@0DD$b_wHr$aMRHn@_9P2`N+QkUlVXA>$^rz|d^^~YiWXMgfck#A-Bn|n#BWU}+Mf&lY54J -!h3K35IPiDm$M4Wc61U(?c5nz=rUkJ>GJ#1liq{oL5e*9zOH-O74C6yK-=XfM_^l`RsO)z)&FR;;-$0 -h%AMUWPbk1FDeHA*#&`pyZ9iUnfnsHz6jk{ytd+;K$aHJxyGch=IsZx(#R~`e*A5!-bT@KH67RXA)bu -|wy8DbtlOzN)-BMhOA_=^1Ut-AJxS8}sdO%!@NO;N-0u)CHRL4&TK#yq-lQlNhz;cbwq9^1nhTIj0 -;S^l*JNXLZR(X}(+~QV6MEf|VBDvp|`+~oK&r)F^NNIUTOc;%HV=8Q?6a^kGEP9^Xzs!#S|3rTLoBH^ -%IZ*8i=O{b_mq1TN_+=ok(rUy3dHG}KC9mXI%+SRr>p1S+@sUv1JZ&z%)Z;E1U%GW(@$(Q0r3y@ida -yAE{8$cXwImNOzBtqaRQuv!)u=%>efp#XnK%#)<4FLKKFd4CGSA2#xKe1>Y)MyAEj5p2%iAhw=*TJqK;DgLxIJ)#zM0EkNI4@K2ZN58LBCA}kI2V^=KqOF?Lr&+H==Q>1 -?xL&h%rj7E1hKLpliu5ZeEjVbhuk?Fg0@@>F`zfX;X+RI!awVQts@ANe@;+3fcl!{!i(78z(sr)m)LU$l+70 -2*bQd}b!|k~IRSzbyAAHHXqY0w+BX5`2h8G;KcuzrR-gxczSXAMPI+mGRtOmpIxtvoM{Ea -T+8mdV2Yu@=ZiiMPqbk2Gt-n+YT8J;m^K95Tx%F7FTyEy>A_-Dz$5F(k^K-J3n48WUOgYrcv<6!P}nC -Y$|Nnq9(K7*76i|W38b%}@+bv%F>JI5!4t1?uVxRG(x -`J{5fa;1;2lPz<_U`uu1DSuvPD^k*}AQuakV0Zl=SMnHRRjku^b{W0UKWgJ(<6G!cEweSBj9&Q4!Cv3 --W6IjTeHX-uiPFVcjab2wLLLPwu?Fq84 -#K=3w1$&1sd_~DWLU>PAZ;WIMfm=9*eib2i$zWn%B?$I -R{1yPWfEyrt&t&*+>!jXhdD}w7J3?+dC(^g+=UcmY3-#NhIJUnY_HG^60(bJ}3k0-Ln%|OzXdg7Jz6} -~G%$d9}N-wOdf?dhK_;A0o+)nw4Vcnmzf`i4?&hzWQ;njrh^O0xFZ$pNNqs2F)jcOqS%o|?*J#?+kU* -&%=m6tdCzx}!H>GCh0dh7QWcTYX%XUnNy>=5|f7Js)x;D;^#1N)&B95)0i`T!fP-kCd|DO81Q&$tPfT -V^eji5*(2F4F6IsBp0{QpvsJhvN&2WHP%~3W{?s?#LG(km8GcYwd$OBnS&);G_~kPt3P#(c&kUzr9F@ -)E>w{HDaI}H%(Ej -#>ybL2yL& -_o+doXKd-tK46J&DI8FCKsu+W#f{A@y{2Uacd$L)(r`C8aM+iLQ8cKdSl?4YN3TM~@f#Grh%rs^&ly8 -FY#S4tL!(t+OxWPKeGzcOae?sJz&O0f*+OzZ?>7OCIjj16Q`3R_ch+$anJzyOQD*9R}PV+C?Y@L>qUF -7hY%)=5ivCY>*$Y)^uhYLJiF^jW_06#a%6&X$8VsdS%FY&hACPyIPfL86ZSiGJQQTkFh44Be{CZ^J2l -gedsH){=d|{*OsHm)-`(1Q|x=4?-$;DWBeC*BaE;T+~I{02!Vuo`U6yDbyarN?&{wAT%6I>IWkcYBGR -mwGuB#j`Xe>6I7t_8XX3f4$m2 -v0suz;*8eK?udbpLSTSh!{h52W4cpbWXP%EwnWU{BMG6JJ`B!K*S#Yv2_pnxqAJ6CWY?qs?_c@v`$GR -+ga6S>l*C>BiZ(E%dTQu_x*MhwQeN%uWW|m?R;!&zK^yWAaYN{l3U7lePC#JrOtL;5@HiN-RuR|U2-J -^_>S&FY}-kZ{V;7Eyxs+-+qMSoc7)Mi+6=L$ZSRW2Z|FWnL^!WQlyBy^`!l)^Kco9i$CoFb0qzgb*P-Uw9@|6Et_&6E{DMHhZDAsb!ZNctjC&orD? -Pz_7@=)vQM4tjk?cA33D};K-3}wp5RdIzE$l1hoXo2%KRccXV0WrMwtryL@uZzft@vwM|~F|NraU_QM -Y@_dmSyDd_$FT_1t&7kB*7Sc*dgxYhfTBt<|7iD4uOlh9VQg2NO+At*r-2tq(G34`QM;~xZVaGTnVH3 -(z_${@T+A7{yS2(gt0lN(%)cb({<$m@WJNXC1kEWG$YqljzE-dpUpn)rcx8+}2|C$)|6R4p`kUzGc7RXoyJ(F05%2YVdGLNy?uyjbxAnux$M!n#!vOZXbwiGImYpueEg#WCs3q-18DpP+Ddsq15ObM{UNC?py+e<~vBsTCf -N5ilfkRe0+6sYn2M$95a!{vyRoM-^4=xLt^Q}10UQzG}dwHRE8Sq*rKB!=SHU!N>3j!qW9znh(YG|i~ -E$&!tE0oJm^UJ-9JrAzb}-n2lOU^)Wd-DO-jDnFjooD5BYy*1clj4N?|dAN1IKE@$1&6Q&f2WJ2xJv}NpZ?!l7Vc -&a@d2l^Xo!|rffWdMZNZlWhl=p~}5vlW{oQroAI`3m#TqeB@B49$OX92m9-MhK(z^@tNZY3n}5j#q{+3@K~nxM{)^wBDuU^-j+A&>{T7bh21DAq&$c -vAT=%Jo`#Zn#okIhV9QBeoe&OWRdN~WLB~g`TRs^+rpJ6@9`rYs1#9E_XLklZ%Z_~6P=C7neo`1QuT2 -c5KtcJ~3E`ZpZs#xNFw*BmSZDr$zX=$kOYQq&*uw}yi`%Ms*3%q0UHB3BWqGm?c_~cXaS -aMRMvr(tVXBP>_F&_x2oQ`yCb}nsb>-e;?)gfjC&lOE1&13CF@i2R*8|1BF1F<(V*XO^svo|rk)6TZW -!VbQ9#@N2)<2>nf@p6=dG-O0^64S!3AIIJT)ouEOqQ+XH145N8Qz_cYoHS0oANE=`ULU-f49ngcy|>O -^gSv$OJ+_-(7E8?>z0T|388MNfQ31@aL4w2U~DW`^~4FgE1ESGAZk|>atci;7m6&(f~X%#!7f?qz((_ -+$+s(G-*IDt(%Y*UIfb -5+{b+Q}p=#Dh+o{SOFqZNN2j)mMC;q*a%F#zLp2tfVXwva12t@QMyh}I2Pwzw*ZgWdgWIzL6XzM;_`} -J=DzsJxO^BgToEsKuvS4A|V?hr2wr%m~i>P|a8F)?X#=RBVXn~@VYXF<9_Z{iJ)0VQ*mvxlbOZNd5(h -S%5}Pkn!Lo*Z&_WB(TO&sTkUF|Vh|5Ng~}EL%5M1FX5jJ0!@D?fgEZ0VAvXAj(o^Y3- -VDPr7US(h*oB~82t+#hf7x3Qj9#Lu>Gv*A0w`As@A{Z2R&1|ed_a(|5fwWjBPYt9dx__yZ!p=%7mNrV -Ddj*KD%1>!J)g9J|DI0B+~1&$=WVdrntU^MRN==TevPomwI|0ROC%?vMJ{>SB(cAw9h4^9ualL1XS*^yg~n?vLu1oV>DTC;pl -ftEN5wRcfPNn$VY=m^|vser;>q2QTH9$Nv7Qfq(MY-#<0*Pab=F>W@?y@bg@l+CYA#+Dw(*gxB|Z=YnvfIl3zez540XRMxMLDmjNc<6KMda|Eq^TrLjv -#q^)GC98@a~S~yliz2iKmVqUdP{RmO_z*MR!GO4>+Jf%@>AU7|tr?t=#RF9GG%bqpa@AhZ`bZ?M -grYx88xsqG>EsAp`TEM+1KbKh?)|vfx?O2ZXr=n#mZHl5#Vav`UY>^Kj=z|;U_gQlA@d@i$>)$J8pp$ -T2l*S8=YNE_-P_{79ESh1IsY7ne`CHMNIXap7)lT;g@vb?&qv_B6I*{$O@Jkp+qTSpa+4M!WiTBXA*=9i7iNH#~*I>9Z_%PiaGJ -j=}z9}>BqRZQxMqAcz4azG`U-yR1$^+Vt1=ayDym4_1**>ldNu#|UBKvrnJubt0b>PbBNo+frSf4co| -B}0RTah02%_7~D2IITgGn}J#o^WYI4wh4sUfSTR}csiPK&HYtFxMT@}2uMiPQ$Q8w(b9HW&FNyUPTr`+5t-JZ#-aCxf=Q -1W*|5C(ldQ;p+}cHJS+u_@Xb&8|le&2xRe*|SPO@_l&qm{fy_m!gy~i*b8X+!zQs?oEJAu3v`=!-}!i -ANtMd4_<7xPbgcO+c~m|G?O|dwPd4u#7~bWboER+<@gFkzz;23Q)^u2sW-RWoO#hSetw;yM@OCdeCVx --vLWeYQG%=>N^(2AJu405Ab{WQgT;XhF8hHO+YMY`5aT^-&3ed>64VadaPUOK^f*Q5DY!ZhJ(Rwb<_KLQEgP -?H6+c|lT+xMrS889ym(?YM$kb)BB|hw!Z{UOuP(!2~W?j!kqfj^T9VN|1hhE -rI-*q}aD!-tQ__fL}?9vGEXOXgJ3*8IDU+j(YQW6NBCt+7F6i!1t13KhY)rLc>q<`&=U|AB0NTottV1 -uR@eNZ5ObBy0Tay_X34%b6o9EoO$+6SX{jht4%^Y)5>5k-c@&(dhn}R<-CDZEBlyF?3siN7RWXIZ1l; -CcaC>-3a#6D7;X{?CMQ^VkT?cWX?7EBdxZkGr&&81g}m6;cvk4vqqG2bue^Ka9}iwwbO%9nb7xM2<%KQ*4zNQan}Igo-sX&>7x*c#!V_B<$8J8aF^uQ)08OKHB -v_}42&{e->K@WKy=WA(%L0~0Vfo*g9xTla>j}C8?@FRPZS@KOX?O1B0q0${rXzCLE(`&^FpE}&ZsWkl -`-XJ`Ys{B<2{~V(|BcZjDfE+rQl~uEuU6}*;71$Re@8Hy#6H)aI*4T7oWQDq8%}(>ThCii54(D>SIYZ -m!RZx(_$snTrOisf_Lpv<;ULfD2XAt2`^7q<`}rQR+5+sPA2}%?kqIAoofYAo0)_JLD`?qpxINmh!Qglz0P+yyTL3~GYyKemY*jwS -%UirFFa%wAh#dctr?lSV5d*)xNue0U`&eCtT=o;-=GBHb34OKd2ny1gM0-D5sGE=K=-gg3U5=YE?VwLRt2aff7KXwAK;s#;Z=fb3+}LDDahbV&~V~OD(U#6&sMm&gNr=2_P6(jf` -D&{l# -_7cp8s5NEqz6l3rylZGz=3g!1co=EdHHZLPc<%qj3-N7VLl_sX@_TOJlMN0P_6d7h-Pw-fXh4)Q-+c0 -(0x_g{zW0s&x(dtRt$Wnwk`3ND}&*aN$Gidl7u)1-21)I>v#4~YTMt&0zVo2BUR3IWu*fmUGVhaWeJ2 -nJa&n0x5~jGy>5XrDq|V#PF!NXv-GhLDy2uKoRv*ZeMF{xc%=}oQHScdEV@sSXlqY3}MF01j -yuYrma!HvaGSsx6E9zgBK(xc@g^ndSbaWf;yDtVH8D)* -b(+r-EBi3bgfhZJl4cc;?FID0Dl2Tj?OaTl$d17H+8Z<1bVXb{$L#*+OOdMXZ3_4(Fm>X8_%s&{AqIi -b?jGon))cNjGz#Uxd8%eK+T!bjv8OH*apfE7c{po-lY99nO&5qY~co$4Izc&*I%8Xs3pY -lFs*cM<7tAfSRk4MZbhBFyp>Y -C`Sa&OPEijR@16RH-h@LYmgMay$)Ufc;FviD2Y!(*94{X^+*K12%aZ0;gG7!^2c?8}y*A4mMQRw+gy; -L`h;t`R@s48(US9pkMO~%X-gm2ynb7A)^xoTaKqyQd>m}8y+@RA2qzGqb!G413mRU|w_j6|fv>&O%Qi -wZltKgMJ0rU!dq@6T?zgpT3xtmfaB^X&mN2RkO&#<`vwQ7A5{OltY)famjaqx&C1; -4Jg6f=^U4xff%7KtwMFG-wUWwR`Fa2} -p83-3T#0oLiAYh4A4r}-|cR|?`&`1?QX1WL-%P&bCkJ0muggigtYR -ewqEqlgPF(hMB0jir)39J*Q2OGSSy-_Xn-2CjLC@6c^dc^3%l{faCBK2#a} -NAqa+0Du_stVfg6(f<^Dun7Km^Hg-YqD#Pw;`(o>h*GaI}|yq5-5rCG?+Umn_WEiV=YX&`xGXYl@rIw -!mGmS)yKSFkk9)*f|Wm?Q1qk3qyJ&D2yTva*2Z!ox4Ks_X2AP$q0%VgJH7x|=f#x;l{fjitEqbuEAx42!X7FE_ -?u=)}Xr?|gpGj@@3ca{y{F+LQ$9}W^;7JkdYJnz|e+V_#}s;kJMgN*91a>QO9HTHZ8J6N5XgacjSIJo -wdl>mrTRCF1X+0G5MvCQ|l1XkIaxBlLo5BHpqj?c$6@x!Zh&&L)a+dv7Q7{)3bOy7bEIKQr;BcYnEHQ -1F1my;kikNMTvS*HhfZJ==PJTSXdJz+@ciJz%s%8Ey$h3zMZU3dUeYul3)qAsr5TWm^OM0HS0S0dhZ2 -ga}wi^zP8ggVQI33*2Iqe&J1Q6HEjg$V5mm==>LYb?Kci6ZCm?`H#cmCP}Grym;Pzj8A)z8fmFb+V#@ -{{PvEi$8P4|IUmbaO!tw`ym%gqASKC5tx7p4B6Zk*Uw=Jh9MmO8Pv+iO{Qs6CEtx9qA#sAUrK=EttWr -$ypO&bC4xWq_FC^FvkmmD)H2;ozgJj>?oHlP+_+v)vEIX9CoYGZ^eI@`+VewBb-@5gqOx5BfnTsq -m16tX)flH~5{g>QHe#kOJSZn^m7OAy?3J)!**d{gvXKd!5Q!vBS%T@@ht4a-FV-k=5fO}0~Gve>u{4} -}DMNPTwmPvMp<%0ElM{vndN&SqbxKD=P>@J~L4db=Hm2QWDSvh#&6 -VKybJiJo4T^r`B@Kn1gZ~`%6A^tU}zq4BS=EwL`J4WYw2C7Yf;b@@dDBn>~ZTh(OvCeWmr_;OM!K0CeeLtYT%T>EGA!+NvUGk%x4CKNQ~4Pu -Itpu;W(X7;Jg)4gnR@BgibOz33oMvn$vmL1W1td64M$IHoH9T}hHIH3226tV!G#g`kW1SW#{tj}~!O& -d*zizOpvIGO5`PXF%z$2cwts==ny4me9G=9TedS<_*D5UH{-|INb@cqR&-yw~RBVl^2y+2qw+t+JS1o -8kW~`X$r?bNGBcz*@ZX@`kg?cz`S=50x``6zIF1t#6a|oyD9gzz<>m&*jegQz^=AO8rQf={mP@oGBKjgD}l+HyHMP9*q --1-+}YXW4LF=FIs;739dCLz4I!uy(~z-z7$lXj&Mv+Ra^R7l20Cu$Pe8_jANbP8!u3KZ@xz1^-2i}jZ -T!P+rbd2^Z0GiU;}cs!t`E#*{n#;hKS5}89ITJKJ5XB8={cd<9*#oer}>+q;UG#{`^Pp*8;L!y6UfRM -eq`jqlQ)-&-idq^$cWhK9(O#~_M5vRG~&@x%Y@h*1^bdR1TXvS09wf2k4u>>Gjqnh+r_wN9GHc{J%Dlj8PY0xl -N>ZWT3knCTpT~1&#uOD3zV(Gg3T1@UKsse?5U9{1IP!>`EOxTlv+S%0%tWlH4ZYXM!FI@?!7 -vr>afD^4(z=>~X3EQYB!o4(`>xu~=(iN{TIFiQ4lriXt0qqEDRoX;>NJS4+80dva2jnNqKRZ1g{yga1pKR}FIrQCUgFaLe$_@ -7PvOAPr})Bc#L*iPs`5QR{RKrs+UNRT2RXvLF1_wSB2*q82yX;`u$LUQ*5OUbQc48*qn?G+J*#FjF^( -4Y6fp<96n32&ee-CNu?m$5Ipr_fe@mmpiBAlwbN*SX`}k$Wo>%zgzX(XXPZ>!et^fl&9&ydW5L0ngoNlhlk-^8y?=De;1kx$^VW1QUjj#cU1AM2Snp8kBsHW_R3*aolwc -2QWdgjeHMHA5n(afuYjcU8DIe)U1z@mByo+)Kbj||4}b-HLRep#=QtVCi{As4zz3>a0VJe3%x6%!U;B -Q|pO*~$-8Fw+GVpiT{CUa1@2~miIS}B7B#1;1k>QspqRp)K$g$dpUyZGcDuY(SiAJ1~wr_~lX|wZCV$ -slA9S|<6Smt~e1aOBewxzZbgK>tf#SX$HlD7f_ECH -h45$pV|ywQzNUgHq$=NP?O#D5r-5fc6XDk}fussDXc{x{S9pgBZM{5D -CAH3nl_ZW$-{q;Ilkd4JiNhi!2lA-4@7-Z}j`V;XI(b;LePyw?cA*d{2nh2CsC3D|N%(bi(P-nb6llG -^ZAqP_J)r(5bc!S-(Fzm(u(o1F{V-=Z?B!a@ztfr8s~$j8LzM$`XaL}d-tf0OEeAC+O9(?3V${o4O8q -Vn&o`R6^;z_+Gp@f~tk7h*XY%+Mdu&N!%U5 -1MxXwa!nVB5w-vDG=eJd=(P@R~ec3b-&Q^ae;BIJ+spyhM06I8bElWAHXYKH`Dm)$#&{Z^!e_}jXQOr -Z0OcfXjcj4DbA$&VQRp4fxP_Gx+aAJ -W4L2o$~PTGajbc--`0MO>yB+DG-M3C`Ixt76lT{?U3fk -OFCTL9uJun>;{8M@4-iaMLpE|K2c~VNc`Z&UrT#WH5&zAW$NP@LtdHVVkjAnY42sn_Uf2o(IUz#J2+S -fX1j;#&T3k0|L!`Jsind1A92K?J2TPGh4&P8=@Ke1iDUYoVW#dN0*HL?0^emU5h?XRTztNA4ZeLMy0d -Aj0Mp5xAp>F|PURg~@63GHreD`ks}XKWBaFC7QbE={Ahot5@*UV5TssgOQ)#_|jv| -G=!G|0YiL3rZpKBn#XQ}O-WAJjgAMxVO@Pp-4LAU^w_e_otW;vbv(RyA_ES*X5RaJsP1M6d`QOxaxCN -y=aE^rvekHdU63O%yx_M&Y{8ym`BWHPF2V7 -V1Z&5e;nzyjfCX5%Jzn{;@G4Y*8pjPi3w>(f*rwPw%&LDdku<-?8}0gn=+^w_Di-I7Xr=N*T}u$l^(n -;Vv!ncqj}(+hsYz1!OiSsDFiIN^#eieLh_2d=>OD*6#A7K->@(7MI=1!A&d^N%@Jki}%71h6Yb~tv|9 -Z;!X<>o(M2`&@)*aj(od3sl*j8Z6(4>qSEq&&n*C^{h>KeiP&nd6ZwKj&`yjlzWPB}#ECxEH%}kHX(4 -WPJbC@W%Mhj_a% -nDc^%S$?y@=m<}t-$b6Ga43QJg~adIB6$aYlqKl`O&r5RE>p4(4&$Y1Ssu8HsYaV?0_T(&{p?M3s0i> -e;NwZW0|(Weu(t)xQSIEXo+Q_;E5r#Z5@CXy3o0Vfbn~6GOs~Ru&YTE_<`$M1&go6>4(QONGU9Yq6!c)FkM7lBkHewA0+^g4al+mpB=#U -)y^?Qjx_fYhSix{Uw4!{xiEXj|;EqvpfBD^w5Pk2=7Zf}Q=;34l$qf$6*c3}AM1-@Ec;X!?%f)+$u#- -pWKs})Nl?s@fCbX^Xu>inqO0683AuNt7P*R;rz`L{73&Hl@x<=|g}t~vi>vD&|Ee%*vP!{Y7rQ_ --~ydqrvAUNzj!+bQioA-;b;%Ll&ujTwGylZG%9UfKGJ=nxEnHccG@CO2zxf`mW}gVz7RpEu}H+hHmd? -B51F#+|_{BCy9vtVz!;!Lc5SY7;a#WOtw$Ly%BGPbQrl6Q{uapX0!_kM|B#&qvv+Fx_Jz+XTUhE7x%Q)mMpe^*X+Li)Q -5`Anf&v){R?t>lNb(?hIL4na_H;@+-cbRF3XG3a4dI*!qwe+iUZW$yfTi(kalhy{hI{e7H`)IF_4uZb -doY)W{rp16-K4#ES-f6(jq44fuL-Ou0{vR+}1`^My3-j#hkNIG#DUm9A?hI&xoNp9*lax1r2BUU*+){ -Es^D^HvAGPBOTTjB2GnYMgb_>~TEhAE;GlHYs7ibn8(5ZgyGF1-A|qi>K^$f^B({=P76R7pi4{Fz|DT -d(c^)yLYBn4_i*{E7tQka{g`CJF(hFpM%V9OGA%gOG0CYgq4os0_mx$HxKZF=O|Ji@4G{1xkxGl0dRRe=ALL|pG|dbCjHh -D*-ZACRs$~_eiT>98yz<|d_B1seyh0p$>` -D>y08YjQI+Pwd#4?hLl-@fT1w*KUXA7Uwn!8k-^?FU#1AWMOmP4;qi&*B~64bM@+9sQ&lJo3C?K -X(NqaJ!S0BaJ`-Mjlal#XGN%h6FRJuc0w54Bu@y%=t@?R5W+z`RQ@9t;P)B$fWs>=UxC+(GS&sfxC{| -p#<`5o@XO@XVm$@rB>yS5zTB;Fe4PQ9&bf|)}bG^j}c97B0=7 -AYP;6dz`Ae-2x=I`>%sb|Z<@)W9toH>hk`x@lsxs=}JiDZ+5<;>zyxP -O8E=K@HgdSHV4U_R;u$@bDCsr%Hk;79@E#~n5BYo&Fp-nA4NX`Gn#-nvEHpK8!O;~-&+rs_g~nT2 -QWAEQZVaHLRaMp??P?&0c?Q}1hcYs>-e&9~HK$oYQC#qphYEiQ7c8rN~>vH<#{Li1ntCx32F20rL64b -9IO6{Sv_$H#DS(K5W)87bUXJn>5cD@`asoPAbRhbVi<&y||R?;%gxN?x3-c|X{eu*52T>b1~6yy -=y4L7M&5!0V{g?PgZC`Px-Q#tiQsSd#Na`TRl0tXi-rL_7sj<`rwFWxs$Jtf@vBHb;|o*9$SK?&(PB| -sAh7u8n`v1)EtK+x)X;auS8l*4vY-BMOH&F5m7oM^TY$}i-jA+qb3>v}%>S|{gMlUWA)Jlzt -5a=bWkwQ2+wwfvLq*|U0k1L}e9Ny*O?N8S$s=<+)h%pZ6NE%}lTyaRi1?__)A4h`Y71XDsf`cgnI`LG -Zq{0B^n3?$bQK#|#_-rH6KD`GDO%;+VG7tCh1m;Ln0et+-M`b88{&5R;3-rTX -_xsa|pHKq-fV%O|Xp$l6zfu8f_D-6c-IQe!EVLAiLv2w&_ntJHsS@sTs46SOL?fdc9N0&88;Zgkdmn1 -4an6)%Q@@_Eadf(dC5LQz7v7t}(tn(8`6^=U@L(JIsQ1H8E@>tba*K*{w$}5PyVFrT&7{<*taO!1-o! -H*VaYvdb%a0{#>Dw4o>9SMbSl{u(}gu{Qn>;M0blfZu>m!Umsy?KGWyez#5X-Ho< -<1ME-;g=)~Iri@TDmhn8ie`0b94vg6gdft!rr-eT=;Vwepw$KDsT@l6R2t)Nl1oF*PuCJZ)~ZHV^|O8152liF0UjwI3iLTy$Xc>eJ)7r!WvKF*E8M1$~%32XP&;jM!$yk -StMcduO$5=`5AA?Ps80utV6A^SU#W4{TJu`c3StVIe*B$Qfm|h5p?r+-y$IT(>g~O-+~B=Zmer1Ses` -5-W43dTd;`m?Qm@)hUu1VrG8#fgGIaJ`feH-$6NMuQ~%%f7@*zVa{VOS^{umbBbV^*UxR1A*2~JGEt# -6e`*9w$HQH>YKHzrqrV;^dYFy|SGcR(h)PeR29%S!iUSlx0li)3mCtHn8nr)?#yBNbh;a}57VvBv$w> -n3eOU~t6T(FS%xf6Tto8PpaH+dU}bz@jtWK>_=N&Cr}ma@?GM+A~8!LSO3Nv_tG53Kvw=U<;db3}Ka)d?;-6yM$4-6p$x^Y`B9U-h&`!=XXV*#fDCZhBgP -wp8Qpdiim&WSJZHog}xVDc0gITA3rd!;haTYlKbqu+doeLu@Pk@~B(Rmzu+ibqiRg3hW;e!{si#^|hN -C_+qCyFiz#~vJCEBc%&s?tZ@CKnkLvM74+Iiym_6>Sto;otNfi)&fjQ~q4oDOoZt6Au)X}jmO$K@0Uf -|Giwao`-th;%EDVnkb^h-8l!+d<#Skf3&!_SEQqUx;+}Wo|3BM+O_Mkdo76oG{BF2%CqhpX@f1_$xtx -+n3A92*Y9Vky82?okadLu434SC2{o0y+R_!2hnK>}benaS~fdo}^j!+5N^CiXrzl7Ub -?_0t8Q8PwQ}UWmR7;#qw|)j!K!PI92*_uYI?MU(LxJ9E(n;j)nNsb8+$#z+R3P^pY{Ss{$J0Z*08 -eeim@iO^G|yE#o~oG}HNcq2;3&2HfQKDSoFrbh?#bE3F(LJ!ekR7v=(7hyrRsM|nyVa%RAikWY`tOO= -^wWRKkx--Ak3C-$^H{1NgKkn?KA%Azs$M(mV?N5C!&mGVf&r%qBl7mz(Zp~LCLsuH#+j-0Ttb`nsYJ` -w?ZeWhUUawj=@d$}OhwxBeCsw=m`jCGF0*o0*G7@K(a#6>*a4z)Ei&x09bmYI@Hz|f}r&aOs-RI26}c -ZKsJ1ZXhHrOBY5R}u>|OKvK+BzWBYTttWCC<#kHh!Q@sQ+2*SyAyxbJVPp>VYVy{oiPS@kuCOKeDZX= -ccz~hj~G+6rb#NkK7KUG?4lu`9om&C1I4KXTIk@+55s3quy(`U{;KiFr?%?&N-@2U>sdv1&YdoXQci1 -CSQvew4d8E3=ASmyZ04r4bHlNs5A!=my+*j9!+KKOZc$GVgc_RUd9s~VM^+LYun*ei$+;@IoEhlm98` -dknIui!Ye0Zd{!-_&j8T^*QZFq{pEA`r&C68ckAVsFKJJqXPEn>jm$hZl6fUME@EkFZ!)V2SQR*dBVf -QG}dTPE$+on>RGGtaJj#iuZC(}?RZb-NT%44qR#$|h(qZ9BDhE6C%8@cLvbEAEvoEiL;`y(Mz8Z5UK}Ys!#sqvyzQj1%!5UV7DNOi)n8`Kinb=)!~xT~I6yfaB6 -a=gxEOS_y2Ji-lAFFc#*k!84%Gk>H&~h7eu5K7x>eLrWBkiG$ZhPn7cf(nW8pp-Fs}SLlYzrI3c}-PYo0HUg3dm%#ncdjWZv{-(V9zoWk -=pyy|ctQCkR1&t^u`f;ug&z0-In!D<_>PR6|IwNi!5@cp&VFYia?N&IX@7GxRoGtasLcviod^bEh)*A -Zd{&cIeJ&+XBNtc$$}A^7`GqrW&FfEu5=7lmo+LTim&w63Uo0ziRJ!8IHdXAUEvxy8nx3|7S0L-tk@E -r*Zo)E*!#t*>gXn_x@k@|DpfS*9`17{1;_(pFre)d+rY)`FY+SNHq>_PEi=X8N(6~Mj;r55}UNwN~I} -;!axc}@t>L>eua(53K+MfC`D|UfG^gp=q@ph>^e0oeB1O{p`U`vO@U^w#6x!L!DLs%4&e=8u3zGNSM` -buF??gaDYhZPb>voRm;6ffCfahrD>EmtEklbT8+Z)EjSeT!CZxJP%oo6zgjnR@ayPrGW|?Y-l9?W`1a=L1aFz -aCc?j@vL($_{}o*yKBIu{IZsRx)k7JE&p^Wz&ASoi>3T~`&fquPBQ)!*L~o9$0{8@S4%oSjd!NCQEuc -_bG;S!tPPwQg&FiN3n)_|GoiGG2eVv5ICnsp$*_k&_JUUBS(7cs?wW$pd=g@LhyX#AO%*RYuAw~39yh -X`&Wt+}H90`WfRm>Z_$;6D%U)O+6obaBlJdC}SxQ1rNXqO33Gf1?{=z8Z7gV0+JR81ax026{qXM`h!6 -@S)zTwmhvU*umVfz$I?^+R1uxw+`uM5oT73uk?!lM8Wy183gnXMfh@pl6>@0V52NrI?Mq)C!-E%PuOs -v0{y5kF8a2jdy){YeS}{}mVRJEM*qPyUV8#;;Lr?@URP@A>wBIUfI)-{DhK{xtiK`q(4^A|wc{{2oRj -3M0t%D`MS9Q5=F`1jjH4{b~8l%AMC38N-{HSC(!OZ4+Y5w$>GJC;n@c#n6qMZ>d-6r=dCC4@fuu4elW -{-3}GwUFd9$jS;bNcX%)AiT5)}6x$Cl!!=C*itnde=gYcVV8q@>gKa71HT92O*iR~O)l$j1L}T9B({m3H;e5un;HHm_f7 -ZwT+u<4br!h_XNKoyjfYrB{=EjY;y+ie9uD^w89@ZRrJC#^I!Zh&1#gf7Q?vj9?Z8?8uNxd^84W}WPkPS%mMRh^ -<~x|bPE|r=2F22JyPOUB*sWZ5Lix~PX4#9C{OX()2TL2J-X9P%S`;tK@B&aTlD#G{q -g^0?#;U0#Ih#9cYZ~^tGk(}TB{#G%#)a>H)aU|5@Hm;zCljAY^PnBRo~aWI#;Ii8VPjnz0xM)?AQ^kj -K#yXtmnQZN;q8=#terW|Hvy0(f8jm?1>4`_BYH(ANplhc=3IfY07Kh4HdBs=g_D;t(O<FTKCcq>ss+pXniebXBeRUTN)XN>5=TI8Wl_1cYeGgr+miWq3#I -4dsG=Kr-WWL{j0*7XV9!ne2AgCP)txEz3W@2CkZu~BCPk%d=T;a@MLJ{WW@FanLwYEMJ0|&UloX$O_U -E_LOFcsgEKVV-`MC -I92zm~rfl0iO*q{&s^cM$<+HoY~DEeOr;5=Zn7V_=aatOkcc$?37a-z`0$xN##GU@n*Pdg8{+XvJ|49 -779;m(3`_bK=_Rs7_Bp!M#h~0;@u&9a#&`?CO5?9gPIL{AY(_dYZzojmdAsFp!2raO_X(I3 -mwWs#VX -%8#0(fDvjL<$#MODr6eckExp$!8DOO~dD-L2(%}^7k?#&Edq>l5dfNk#m(fO7Jc7V}D -HJ#>lvFEPAkqYW((lMp1hsuBZdRaQxS62__*uTTYP-sMnT#9B0VYY!DCgWy@8P-yh3SB9O^gRk6D(ahfs6vxVR;4=*gZ#Myl>DFPJa`8lZ-A%782zyA -<v0iY)lHs-@-M8&4Fv92|H*t0P_b08V?gOyC2X9r;WyL>?4>C)d;)jmR2d -(kVjsRbe?$%Gm&*Rnq3)lNLQ~lGmexT*heh0xI{E!45K##;w3V~<>MhFbw56}cg!TT3S!YPE1zij$Yp -N>gNA7jfqYC?&_4PqzlJJ3q0qZ3q8hsYWIK&kZC<`!9bc+Qd^lQO70p1w2a{R{e&@qA3Wm4|c>OAljM -1o}+sVTbPftET%;AP^}IObJ7Wbn&n^%RY<2KG9kDVeNJ#H`AXkRd$SU6U6ZXJ9bS!aV+~Q5y`>ocYFv -Xhaz(4>WAhLIr{Jz{FfkbIGR5AH-7H%i{@i@R=?j9QE2mg)FyIhhZD%^zYaK!ej2OudF&W!gTIQGwY$ -*80ba)rdx=8(nIiZ#6}?Zd2lxdt5>nBxb7==WZL{XV_kC+!=Y<^@HDBhBzxUz(*3>b$gL_x7v3bZBpl -AJftlro7Q%H*E);y|n-dFkfWCeMXwniiK_?K0+!muy@=ibd{P8pDSf7nqWl(qDfgaRD9e5|wC*AY9*m -;C9=1hn<9%p~BqSJLl=1A;4svOO`CT*?W-Z-nQ -Qjg@^`1)Gm04rq5&|+BC)uCNd0=f@ozvcjmg>q*ztnD;!RO3+i0K6$-T4 -Ft(^@uD8=oJR{5n80DpPPp7|s!`gtOSDdAcL0jEBfBL7XYB54#=sgOHn$O_V7)7&3R-R-&Slx!&+=`k -QaLc3AQ@ZB5D8Pu_prA}xWcN=vF>F6=oYageJXMj1ASG7c`*}?=$vA>@gM&d34(A+@H%2bV)UGD5Sib -Q%Qk~gTm-ke5cvlYK{rb(hrlGC0lGE3ibFaH+!j+9=T&I4XlLZOi`vGiC)Gdb`gnIwz^RqOOLdnJ|W{ -JxH0_UO>aA^S7bNLc)$M0>Jv^%fyO5mBZ4yywG%(z)vfzI~JdW5#;Ec$A_bLJ -r;XK#X`N`uB=s&`u)KVu)xCertNxSN=4PkpE6k8?fFuguNHb^{F?9{ -`}1_U<3M&f?d~l=+XrByd&^DOASKIwDcmI^*LQrnIc;g+M)@t_`AgUU2$8_i(~CP2IoPCEOX{TVwH2v -He(%MFy(a;92E?TR7{H#77?m;i;X)z0EA)m`NdcR?Lz=!b4#8W9b=Yl4zC!7Y=?5$7#q7M7#na<7fT8TH4kH@@?CIVkqcI$v#Xd-fXD;qYHd8{WpMt -23g95UH -wYD!DWjbyV}U4#R1-mMUG-QZnM{GsN4#ExD;!ao{2_9jd}J2I3RZ1SuY#{Aly>f<-1;)lV)0!HkF_&y -xj-U)Q{g)`bAxO5&fynv;Z` -9*I2C{V2RZf3PP&+XxSI1_BHa;dg1|@We9<&Tcgz04WJzAM`xi-<|HCeZB6(%kf5_c1-p1(wZo_`o&t -dleyO;i!c>nv$evb-Cf`UovGmA&TC{Cj!L|{8iM2H4hzbd30JN&#McnH{j09B6Y4?DEm8JND`1&nJf`s49+6yV4HRs&>+f6ZFsA@j -|Yq4qPE{13v)-Undf@A_H(3aV0v7aYesL_oIgy;;YDj_tUw=*l?~4CM8TV)EGElh=#OFNW)FmoJJ*`1 -APo<(L6KmXzD!><@Y+SpEsdUYk^0cU;)iGB4CoU^i*7e=Q1Ad}f_EG#VfZl{z`IO4=Dv658+wJa+nlP -*;UxZeL-n2whq4-W(|io-63z!u7t7m3iy`iH=8}&p1IhK)d)|Gu#bD>nyHyscn+);`z;&Xl}ok9mBn@ -1sUx3ES$V6RvX5YAT7hm7V7wBx)l##zIX-a`U;JAIw++`PNaQuw30YobLeY=Ybh(H1Q^@d2A{y$PjhGHR(@X -j4@AX)8!1)Ybd-Ys~euTNY -WVM62Bj};tVm0<(etpTnR8C;%r9~938pWj-T_kqwWTp0$M=Y)N_W#b}G^}(2JSXRF%11Wagf7$^8P#% -e^!(o&s5O3tp5B*)r*K@)Kk4uPg$X$+MwIO9hM5Be!JL?ww%ynHb<53_sMNb0zHbHgC+r%9*MwN_Y1? -gxlYp@9}c!S%49BZ*q+SwdVA}dy-h|%85sy=c2Yr>7D~CkmJY?UiVqU42&urD0yExQX-72Ge$kY6ENz -skEq3YGcYRA*Msn+K0sOS+iMb|be~evMVr#6wW$xox-zfE#Zt`UeTSdVc3~>Cv~#^T| -C0k-vfTx;A5V$OvKewkyZWFy8-2g#LYKmTbq7{eZU`L-{0as^U3Z%4g4lJep@V?U(1uy<>lyB<=h#J2 -ZS(g&A~0WtCq(n&qr$A;fu#8$ZRPZ=QIV}@{ZGnxn4=q{)VLnEc4E3-PJJTZ6FGic#QD6FFtdoa$pWJ -+##X`zqjmBV%v(Emb)B{P3e1T{}L|3OYsAvNnSV;=xb6Z!2bJu&c1H0CYZe^<3!!x^7WpJ;;b`tbhaP -k%2`h*l}*mxQthP)=GjB9s=GLz7565fY#NJu_jm@S_wtw?=lZ^?noUP_G`Pi5>STP%5%}bp=aL&UyQZ -jUE%Xk&r866&Ca`U#07-LlA(`OHEX18oP69ocW`)HC(@DLhyyg3h8fZe~nIkapvxZXvW-}F^V`l_vc4 -y1DLDNIH^_^?7a^STL2VrxkDq;X)qB_6zkQj!x@b7g}lt_?YgVO`UgH_WZ0Mu}a-b-e0ZoV#GALKAi@ -4)2q7cpGe`WiNy`Kli`ubAx{X>RJKOTJtZqb9or`E&tRl-aJ7823b8FQcQ$EJrX+*^(5mjW{cI?EP6A ->-^xK&m?xxeN`1mEZ|m2r0kG1keIiW>O)g^ubGg%+GgQ)SlTzwt?h}(22DYHT*lT5us7%~*A>sFozb| -?bH2Nd@8`?@Xv9F`FiulAj!_VV!xV{;7(`+ifj}^Z<2Z?;G(r3d_$9@$L55|Ac! -nU3T;%){3%^4_>f;SdqyHqa!=n~U|CvI+cDT(Bad`Hb+}&~Aj=yN+53kvyjr)nEqVhjd{tnG|aqI>{@ -INu?cX{>Y0oAhX0HxT+?J@gAE0@K`50N}z9`V_!+@G*t8vejn5Cg9mWphW&Jkxq;{c^#I~1$m|N?1Dj|g4@qhiH#k87U}%QpzpwZz<&w)zE~!HQ$7D0^c~IJRsRw -69mkf7{{{3Nm=E}$fW8Cs0sjr?`##$q_!sE=wb8;ip9QXx!IxMyqe8q?TDgoIgXua6sOoFYTDW2OLcW -^sL}{nqA)f=w^m2GD-+ok5y2e0QZAbpv^WN#cAW8RbI+yeI`vKs%lHT%o#YZ^|p{jT1Pbq))`d+HW`N -^_@!uVNVT*u$|EC9a>w|(3D-PG1jtli_8iiD+7dwB@dyP5)#=hFvhE$5hk -<#%NO-RdBwaJjim{4C4oaL=ON@xqqWG8c27!{48)`Tcy0aC^1M^Yti;;nN`b$5|Hs!4L?s6W5Qh@P|2 -?|G{;>&CUF`*Z862(&S!LNfeyMd?Gdn5Rd@#-r -3=HfA{o6rZ;D2(7@Lk=L7wfAZBG-Hy>G?X*1N>>E=f;XjsEkp_zEdRL?*d$4rn>%ZR>Ni&0n -qS&n?uP8VXXNh*;2+gq<_D!YpyC!jJKc9z%S;}Vb8ccnD|Ji&+6i6pl9 -Qv7U~_3A(o{BSu)UGh?y)+*Q~!iXWX6_r#&y4n|LY5n%XoTBPs7$(2NSZ_aw+&rIPLi4jPJ51T?WZpR2~66MSyhyl_)M6`0>=iwA}w+?=&iP_y=J00TFA=KhF=ij -zSQ{fF!PM4A@%v^+%VY)V-DUgvXluHbuy=gQ@vng( -ycNz#w*3#_`biOoh&v*@_2sB41yXnHR->ZlHRXu@Ssd)8PPlbKOD#R|DS ->5fwW?;jEFJm!XjTDU?7IeK2P{upknLe#FSy@<9U(AR1CLM_kHyM!|{yq8&9WS6x2UFDI7^<rbUO;8k+BU?pZJ@Ve90 -AwQeyI>-${&Btz=W8Tmke1UI`jNvBWml$$6G5Sb7dY$L1QRa--ydrUTdAwTkS{e_IP^=2SsWrr&tk7X -y;T(rZ-0>45M(t+LCjzg+l!0ZN+V^=|*+Y4KVC9F -OG8e(=oQkyZRhCM*i~L{>#QMQfCl9WBU#^FoX90?)}K@t5?>E^Hp#f)E_r6HfwfK5Q0Wn8sDPOEeEiP -LnOE}Ju|2L>@{Ebgr?wHYwEBM=a?Xc%<_Fq*`M+<+vGeuTCI2`-{;&VzkD&YS0p{O7>noi6(;45xL;{ -5l^J@~uAcV#V2qF-S#&;}<;qWmxa|~tED21RDzCVe^euW`-sCGDV!N;%(jvOFwM_d^CnV=~Sl{NkefP -p_un8JTONAodTJnSYuJ)J|FgM8FK#Bs?Z_3?3iLk4Yc;0@&M6b`TKuol(}q<&oy2HMWrJGi~|GLa0gOS13MXt5`G9zh+3n`13@ -W_qoGBFhywS7{>l)cgSU%?lS0(Eg?hr!|rfD27GsL`EOPOe7~H(T#XbhJQGQO{0xrUf&i!(&ogJ4iz< -eXeO7%vR=m0Iz=^532ca2Y(*#yn(#q35Pb+Gijd2^A5vHiq&N?*+n8E8+dH4hT@YAg0955emqd~hj5? -O8Gdb6rdI(2l)U&@nZq70(zFMZ0F=c{Ptx_B4y=$|0_a|{wRGKVRDU~Sd! -?uCHYisz^@1INdduD9O3*h5u?$45hom$NC*;k8nTzcBL(h2WA?zxsSI?y(67TZ$Fkd5ql}qAS&+VKP+kgc{%& -fO4%>U7)p`&k(YJUESf?wbT4A$Ue{<6qo_|nD@MWuMnfNyLh6@t!<0Hk>rnDD#P_im+-n2%;g|d8zW8 -_pe(@?ieDSkiyZB>=|0+QrrDv~(JMBDjQ1L@kckCKzn547Kc --*yYD9ee`-$w6^MxNfm=tte2+EI^b^?F4GC!^p34OH1_v1Ss-+42QeT;%WWo&;{Vy}pQnc5=whehE=d -@Ezt$yj;%wgtAGj})PUZb+x(BGjJuYq_EIKPY8@P&3}w;=z3_9={kDF*beuGQr2dU&lp`W5BnAHOu|f -@aEk5yWIP$?k4_6r^qE@GetW=Pv6)8xRQO>YCh0e>#L)Kk$(h@`Y+XiKR14XKXraJ*e^xNomhyEs-Sw -cZ#2_vmbedGcxAx5@W^i4w8W2baMUOR -&c>Or7&K@soGIG*Tgim1Ra(xX#P`g;QhNcHCgLnLwGCv=?ntH3nh5TAhZ)G(erp0zAbh9jd$Yibs%>y -}Shvoi+@2M#)6xs?VJ7GC?0^4|S$&cp;PxCSmfN>bV4zP48Wu*tn3dmwUM_=S*pqCz7Az@iNAl@G?E| -VP*&$lF(gKUaRKf@#XattC!3r^#~|b=i%f9?=ksyKWDT$U8(!!M7bb(JKvlRZN2+bK>KCC6G3QhofwF -3aSLHWQ)Qc53+z=kJ5{#G!-4wtY-tjBT)dGFAnfRLfLjQWvF)l}EWqpgjd}Xj$y)WDUr(@|87 -D~y$NgTQ7I+gTpiS0Gc!r10T2SBm`O+@Bf1yh!&s(!)h -Un%{qHn_NT@&19oEb{Z=#MxM_}2OTk@9*`$JSIB@D_Z#Kz39OSu%*P4Y~a!4g|ib1OG$8pPKfLlN2LB -nZmN1oRCsZPbK$XUG1LmomrNBcc8{?ut|d%niYrP>~EYLJv}<)jO;xs75Ol@g9?dvO`@J(fAIIew&(6(bb-%A$(sj -f>7E#bXm(RqeSUl?1ntG7nD6#QqJ{c|KUBVew(m?PiY{f;nHh-IGWGlL)yZN2{k*(sM?h?Z)Amq4JnL -jbT)dv2-nTTo=Lv{+~`)t9jr81z?-M1)~P}FF6+KC=o9;><8B~(Bv|xH*x$Jr{kT)Qx&^m&Y&yR7mMx -4X;dP0-J8S~(4I&NRytWl)JjEoOMmclt%b-3ct@7wq?7KXaM)ItC4Cm$yhOVS9Q1jK_LB(8jz`(AdKL -~F^MQ$$)+MM)J-o~xXrM0@YC#gY)zu)+k9b5bYYeHlPL2@@1udW!sxMv*^NZ(qqJ2Rc0KVhi<{OqrI> -iaXlhf@@Z(T6OLrYVYqC>n;L!#N+?u@y}0e-J1E!@uT&=*R6IJH&W9yUnYlym%Oic+6s((Ds$x}S=FB8zsiNbd*9k0;p26+ii1An6XsNaWKpE{<+SLLJS79fa+ -abSHv1`5Dh3kH5m7;QJl&B_9{*qs3evT@mUtH@wrt{aZjk!T0D#r?;<$g#MBXiU$`=zL^wSR<7nysP$ -xDFpIMegmaR-KeubmU!~hW^>g+WvK;AcP#h{#99oBUR-K2R`P3bm_*)zZ?+B?Xz($oh51CbP8TT8Q;v -3hyXI~*qq}QLG!WZj%&iZ0j$bE?p1wIGeY(0$jkApf-{n%4zd{^H;J<3<>#P`u$wVSNPoBiT0@WtJ$f -ePneE!gDDN?A+88OQrag6ALLN_Tw2vA=Rdp?V+O5cruKe$NfJc=kK&%$=u2ab`VmYCPM}VnBB^!r72C -j!OWX;9e7q`IZG7KivbLkJ<}Sr6v7-1d{6DYv#Q%#;l}=99`r&wDU?$TC!!Nl4Kfz2IR3))8{kIX>d8 -+rQwpV@Vl)fjs^SQTJ~E1@*e_z<=Oj{NS3@ByhdDexS6=mx1!eR@kTUn1NXjhL*Qp}SfI_tZqEF2cO? -5d3YkeqZB9a9;yQj3Ra)o5XCP9HZgK%+=@N -N9uaxi$abbmZ|%L@j;9%GVAp*z+IN#QXSO0E0*;+C7K%HNpY|M?H`5T2Z`Q&(M=1@#Hy{3NU$gZ|~w -<~*`7i(|j&mkA-sp+qULqaqcbG|l2;9B`Bu^5BXHeN?BTqnaPbzr=&=b&2?$qkRb-1uiQ-$`)tP@-UHB4Q?cZjMcet|kstDrCLfv=HO{Uu#Y4!Q_`bF{a#92v -G&4I5S92#quH_E|*SFBQq=&x&Nv-}qWT^Iyd&$xadBw9Owm!dJfdnc|w$KZu&VQ%2yk-?TCDx9PL{_3 -81W@*QQYG>4V+KeF6*RP@$sdAWQnS?vV)66VIg57CL?{t@Q(r)p=1x}z_$^)<`w$t99M#mE%}cHZd4! -P-O!Js$CM+HI-+ppdD~-oHv?97k*{2fA0^Nuc-VXY1#7A^!;q^M62LiurinR_X^9(<}~po&M$|^FG@K -^q{Q;=#BH_&^H7{eJP%83g7O$)IJy4P?G&l(D|j)5X%1GO}tWKL2PH?G2d->#!ws^ml^Q03N_c$Ue4b -eF6G&SPft1cNFH3>lkxM7-X^*cGGdxm9EWMzYAw9D?cea*U0~-#08L!kagT&X9>Yw;RWH!tqi8XthNi -A~Tv!SCx}IKGSn|`2@lNhpC6O!DqrM`>T(<$%K*nA5-Ug`%V`XD#Ri7l>lLoddbz@*eCDtF2prMvMdW -24cOfkT^hgQ>kZ>%Z?R9SPZx~~zz9Ezx^9Fdc=r(&VZ#jLi|+iLOzL0(TjF7vB%qjdiL!YFmMcVZ`<& -MF{OMF8D+)1$bt4@u9R-s!n`&sM)I@C&qgKCGVCy@A1k6Crwn_57tJsXW -fOE4Y_X|wgZK&9w@x3J^EhMx_!V0H_oF#rZ=20u4ul#pgHAkc45*Hf9^`Fl;X9p(6B@{}PuWhk@CoJT -!^#WJ`P)oN(as+A5TU~=KxqNQF+gfp_jZ|~={o*}4a1*+IkhLy^=tTPod{CQ2nl>vQC6#2eWMR1A;pd -4hdEJ#XQtrPwTb5A^>PO;y>7oK5~e;fYtqoY05W#OqI9^f3&S?)ikp%FQ*j$S15i~P1jlk}6@#(DW7K -VU3&zuB@dDlt3;p87nu*T;T`s)^fk2eKDP2ZU=EFoMer33|)(1}ep-y}k{ -(F+B`!Ti#&*VK|wGkFW%bxQnVp2hdnXMn@w;n|1gm+3U4l5K$z>{z_hL$DQ2jV9FP%@!^L!P%wd4GM` --%2wZFjIFDI5Viz}2N!ypY#_S&GiZaoyyT+OPyDiC-NUBr`}5xAv2M;>L^g0?D$Srz_+r8$wy?CD0pS -)u;=1LUTh`u&n!l7pA_t(KJdk!r<&O#iJ9U8cF;W@&% -w!WgGoTlMjF8f2o=qI;#)Ry?}L7x@#Plx%vvZ9R)Edx1JLlqd`*ffDNDey#(a@WWTGl4#AVJ4o8>uz8 -XY%Ak-?W)C`t>$U)pVmPU3T`wU+=G1%)Sfet*x2kH?3Cwyn0TE_ -=n}!;ohWbA@&fL6KTU6bb@fsMO8O;l6F0|`Q$=Ffc(X4`HS1*f48)0h219&*c`L#2S6(9&o1%{C; -XqE^`j&HfxZrd02rZYjG{0Er3sY8>HX9AOMbb7l^ql9?C;=_@F#(%JSM<)1V)yJ;%rAYJ63|qc|C1HOHNjNs%D{1zYI -sw0{m`)J>iV_Xa(kENFR@zBmnj}89(FoGsNJx}T*!b6c`O~~OU_*bZlMUVaE=vywE!Bl2b&K`INk>>1 -(@8{3lvSzIv#V_(V594+G(Z~L$2AzC{j!2-D-x&Mr#{K#J#cdw=!*2e&2la0c{_)r!8};(4yZrXm4FA -}u*SBf`^e-;?hoB7&zFXSgT^aDZrTv|i1v@JSezep3`h(J0b}`Nj<6vvoxizvh+>@Fv_B~?=XKE)8XP -0~MZZ9h^MJaY`O4g_F^QpUBX~aK+E6FmNGGz-1u5YOkRcA*Nc5cK -zGyILomZX?%E$%Ds~$sJ?OoT?&A}G%ZlM2ns)xVCjIT -Rz7?T=o$=?QgCY$e~ZD7<`bmV=s$1I%tOn=O;Hirw*Y)ek6bWvgi -~as}MN*GM{?rQ&H-;z!z2qe}cDndbii0Bj`Ilw5>V)PYV8h807x&F@1o3j7ajsDesGPm;NXykk3`{&! -vwk)Q_d;zRtr|k36bU@i9dC;%~Z_QS!r9kF7BBsCilW*PX#*Q)@f@P;~AHwGOPI+{Ec7{K;i`IcT2rt -4*zMhUw?49@_{#$H>~q!dKsr!(iHt?(HZQfEfX}Vl~Qn5!S?M{^OK^Sw2)Xv!vj`v -9s-2t}8Opu#u*iDKf)JffIMtianN^jrFM6+^G5K9;Pfg4sdJfx=+Nk -5GoYO{~WA1=2PsP!tN9sDJgIAaiI_Ib$H?Lj*H`C6=rSwJr?0%hdU;v%j8o%wlWJdW -^g4LQ3H2VGc2k@l+E0ID+4D}Hk~1%(k1kWyoB6=v$BP0Y@XnsXnl9iK -J#~q6nznMVaJ=ygXm-*Wt|XgSBXz~OJFPGxQN3Q)s+!Y>RnREZFwdb?n=K>?q?q?E5H>%yKm|7 -8Gvz1htY;BROiFW!``9O(GAF)11sWaSH5(w9!!GYmci7Gchj>AvJw*GI-$35_Ll!2cw;?6{i&$L$PPc -dFU(i3(NS)uEmEw#}4$cQY6JeQ=p_>aAOv0%v?Wfh!T*Z`kS4`fBpL5%Mt7i~EYI>LIRk(x3x-xgs(C -DBCtr-u)>J+vrkOgvJ+ub?3M9meJdzS%j>I6}2`KGIXr77;SGE+O}DcA8xf;gjx!SjC0<7dk!bDv5It -@-oTY8DGKaa-?ne(rhUqlx0(+vAjdH@v}J2@Jrqw>x8IRS_ZB5Z^SqCvI2fQrKTI71NJ(4xFv#DV?bJ -n#Bx@<6Sn}E_rlO*LTiML@IXzCaiZvj*_~=0BiVLby$Eft1Qh>NBWpmvq$`}RRpl>%_ggG4LI_Omdej -}|>b~C2r{5(PGRLIgphRNdopt4-MF*{QM^s3ColYQPn8Dj?`Gg?CxI1M(JjsHeW!Q?k+oWUL1Q)_~^g -{BKHPjy55T6uHifj~@*Vn=YG)n`NrZ(5kLi{uOto!iLS+Fk{}#%!r?)?i!(n@MvBy~vNvyOxSjrnw6c -6U_uqOAXI#>E78qx+w?>f|n>obQ!H3W6`R5ByHX;6#4~@j+NhDPi?T+eioQ61|XhzKjz{@PGkhqOQPi -jW^~eRM>}Y94sAy=(uW&%1KxCb#>bc2D7m9?u4ia{vvCU`qpX|8wADj<A6E-B3trb5RIU0ZP0w(BvJ!NEC#EK$&!(~%)|R#Q9Y&2(wp-ITI&t#B39VW*`!BE(@?T-4zdGyBSP -A(vR>CkEgASth*$k&4Y$t03LjTfK9WIVd=M4Hx{1BfuAbl8aK%dN3{F9e`IJxBq*2sTN(CFlt;=$+xW -9>93PY?amj%*GDmL1ZnofsVk8RW+}qd45vA^M<7^zV{-sQgf29h>N%5v3hm9Z(KBhLebokvdi!M(|m7 -P&Vx2DOVsLBX#J5l0FU096kI)cJ#DEwe({ayu&%N*zw?hg_T}IV6bN|=~uq#|Fq-#{|#1>|9^#*J9U0G?uG~nSoYPUNp-A=CpLWu4{ftY!FqbLgvxH@2v&^ -<=wPJ?Vfvr{-T6QG}oGidhI$Y!c@3FMXv(u5|K_sila;|1NdYZIG!=Bns^gNlnbFR-O1Wr)4xuY$Y-b -~q`1S*Q}CI9FEPmisxxut#a*Vw(#u=cEa>%AT=dX)(Xcna3@OiYz^$?{G4wpaM$TzH+djncPW7CSl2q -`kbf|zF6DyQ7GU9dsJSfQ*tNu`=eUixToD^l-q{`x2ov_C)8wpM9NhmhIxXJ>M7Gb>u%-IRi9-PFSgB --?^UUe+bbbQe#nL`a~1JLz$z7O4W>)w@o$n<$aetF+!HEI84Z)-?#{nY%X>^hDplrp}w7`*!YuDkF1oTFSQWwL)XXgrTX(Ye*xM&3AWMzxaOs^s~2=QuiZ+pX+hp&6VA -J*rf;7`ibL6tZUpUz%G531a!Y-^9s{6qSVZT1qD)!v@?-6Js~FS~I1&KUAGsnUZ_-;oKGH)IFJ!k5K! -kh~9Y0+o -C8Ue?LHOg=y!8iVo1wnf8B&@UwfXmRY?> -|g|^j&TwTsCcrFPa+dIgPu|D1wDSfz*Q6K$P{u!SksAH}L`UnXU -=zwJy@w>qp?AXjoKFJ1$UYt5MqSDWB4wN5~D*`{9a_P_B)sCBXxU>Vkea$<{!|4MY?F&9)6VYQehg3c -Nq`M{`<@(=9bUoD65%B|7nq=)In^mJA5FP(nDD@4P0RI!5bO00J*T_WwTR7>1mHsC<=~xZ$*EmW4w-G -P^!&s%eT_pY8s*lJm-c)uCcu^smBv+wHSg)|IkSnM(!aF^wCET0p{(7GMm_`yMM4-OB*YXB}eJJLlYz -r(e_~~u#mF`5|8L4*GmcRuku&s85gSTy~2+w0gQZop{aT)?(=iCF$>wSlOr3;b^3w@KV-0;W -Tc#J{?YIc2i@A!1Rgi53)J`CIOc*<-|-cdqe*fv_m{$cHzRX6cIO}0iX$@mGp&Ohd+RXr*= -MJZj@nUZ@lSBeC*2E1kAJgIRotr=RUY;1$m;q?P7CT-ADkXfCjZig@s|~A5$icx`z(91)-R-;|F2gp> -;G33%lfN|!11u>b`z*n1 -y6SOUwUf?#4xx;9@3wqAYd3`P=-PVbBeENIl(HMuNf~w1cP!Ve>TCm2)c;Q_mczJ4`pQjOF?!QS%r|T -feorY@UjLhlMerP0(HIBN%4w3qzSCaq?1BlmL;dk;@i~+?{SGV7FDftm2G3x7jpSRU(c2l(Na3U6q(q -(qo}g1zlYWIGE6=bb(-KfMlzHvQdu>cR|6(qArCkz#sIS;-@QUnZ$hqT$UJc6RRCExn_^Rd13<< -S0tY!1_~%e($La^UZs@-0>1r>FnmAB#}b4qYJ}p-B=VU8G1l9?}Gg{tP5zN6|t)nN^2J&*wRNG2B-J!-tj5p?p9;%7OA2Pu#DTQvXE$TB(^ -IngRHm$BoGZd8yFpshG_BOT(0435uV$Vh$aZ&RZ++qoYi$==Ng~nXh_C;gCQ+BZn1}03SZqVLXi|^}l -@ffsfI?=nqN2f2V*a6(tM9Asda&WzA07I~3G9z!z;W3-h -*)UqyQ)xSZFy6~fQ9L%Y(zUJ!4Vg7&mDaIc`IhKFFb1`81nv?(*VUI-IGpV~QKyQ+E3!m8fs -XXpfA58}6^r-8=KJw&?ll(nKltW_@JH9pV`e2q5Oa4#Itp~2O -=<=2ZT#AbwBm#G5_UJwhp7rvRN3y`9@wIQk}wkj^ya$#ZuylR -{64MTcw)5}vZn@0-r$kGM)qX|^1u1kZgL9APzEpYL?*pclWYc33__w8E|>C$+$!xYa=Y4-cZZY^y`$Xe_g=}2Y^ -66Nt*b_kDLOdG=`Ts-Qn{By?uFInLJjJ}%K1cM;yrY#sqVKt*l|V=!f#{z8@MEXTcBNO@bFMwk7_M?! -DG-^d4B~4dLL+7UTw$*11Mc|6Hjzl&t36~nI9pG|gGj^(scXZRW<`pI^Z_|2 -t5I=XFMkt1_(DW38<v|Zm!U$vCPL)M+h46y!l6FK)Yi^`O~~Rpu9=^BmkVDmu@b3+)CGd!n -`;%N4j*l_LrOP{LiY0pS|dQsQ_$*K5EDH4qn9~{J;l2_WgR_Sg&Gw(3q$|^y{Qt51II)9s?C6-Q(aSY -Q+r^@$8CvH$O)aEt(@q$W{(NM-)Hw6+r@HUxhQm+2+`CG))h$!%5X8h|JIZq17+g?Hx9ax#NPAt1 -FD^dDeIwWj?>k0}|_`8k-hn9#uOp0sQ7}gh6}8OdC<|@3ZDLrl4@CGVSJ(la)hm(Gd$eoX -1lJ3m`N`0cRd(H!95>tH~2@jhnyJ`j{*gE%kCwaICXFjr&g5$@g7Z%(Kl|-x}qae%o%;WM?umkG!Q21sVzrVIA{6CK6+Rn -c?>l^h!fQ>4=vZeKNTXA=|C&!<-)gK)8!LNRK$d9SY1Q7nABty{{gJKj(;xtJyG)>STU>U`6kni-VjR -FC97M^T%R0hmj^R2Q%Hx@>rpb=DHz(Zl>DIY1wKe4Tq%B(DBo@Qk>AUZh%^;!ahAtJRo> -Zh9vAk2UOWb#)fVom^I4j_OC$a)5iFck2wVtXZ&07fxuRl1&Wy*8KxFN|;gFl!mQUd?(`wna{XO-}|~ -0kI7q*RKd(pNaotiHL#$;GO@biz(E*al)Md+cn1Q`!NjnTK)GjQT}-VxSsnjx!hWqsTPlXNuA`mW8y2 -aYRjYi)!R>;YVqt@HspDlD~pQ#y$(kG;=ek6^`7TscNG&D74m7xoQVCexEn<~7vt`j9=EX;cmlf -@pG(uQyC)8fO?Xy%UkVgSYuyZ%HORUM`~YTN%ek-4aj39`X_`F4$X*Bqx$Taj)mZ^q(u?d2ae5Ybe%= -dZ)?=eyZwU%un-H(_E0S12TRsn^mRHz=k{oK# -@^@##UG{^jsh0&56tZOPX!x^>O^zK!_t}?fYZznq6i5kjyMsO-*dsl%sm9oBZ%v!G3JMH$P-2M%V>%Z -9ve7yZg`N=TcFYgOzy-e`r?X7wiRc3ty?ZY23#k@HRqTwU0D>>9l_Fef5{aioS3fqhO*EQTujuf -f&5MKgX|F1y3+7@J(=n3w8w#;M6@Bp(`2uaL4Tu{%+)F*Msr=(#GVzI)w^H|8dnTSMAzer@(dm6R-!u -6KOleCY~b;*K7a`C1|=etH~_2C8{;JAT-x`~l93-hYJizj4^lIRCdpe!zK@TuW+W#kST8d|+X)3_|9gZ{j0hssPS}Y@-7PwB49(nRdn -9wWJ1Xt0bs<(=DjX2Vl}4j||z65yt?R_Zl0U}zU{p8+dmamM6dH2&t76#0KevPamvP?n{~71SFADF!;k_)l@3Hd!mgApL1Jge2gtQFQTgLig|c -FS6|aS8?7$|L@|w<%R!0;{04KhtnhAgNIk@PRNGR^E^g+v}ljSbUNxv+yb2mREKU#hU8gj-rwQQ>U}0 -S+k5U|fs@>dbi6IJX_fudYAj_b1A8)Msk=#-TWEyelDc%_bhB3y=u)U~d@&w!bv6AR967M+|31!_ml9 -Zae~U&iFdF{?#0LU^_Set -yq|@IfIUxSouG_5j)D!Z8WpJO7YI02$Ucll~RGrvS^&a*&U(-a6n-E(J0eaxes1={E3dB*7#Zg+Y@Q$ -fa^{C&_c3vrDAU -%YNn%WwNQlCo16F#ISYb{Yy1cBrpRA03qaYQVceo&fzaK0%elCZ(+!;bW}SdCe>|Sna|`cAg3>MVw{G>Z -4*qR4P+DeIawS4`qFhUfqL^&l>X1I)zd%4&zb>{)*gHai)%NdNK~8X_htJnin1C&n*NmUbD1)^}uc9? -jm6$Q;4VM-uzXm`~%dhl4W^Z^jxUNS$dH&FWV4@^K!vS&p#R=eV3j9eTh%FuhXWxRJs%V9IXpoZ+02)!C}yd)OJiN)^VyI -mRAM0^U!@{eX7}1B`A-l%j -BgW+)g;jc|CyIRax)f?gRfL44ZI03ps>1PvH6Zv27e0B>-F?LmYrVRoAy?A=f}Ao@?ko3*rR -3j)|d`3gz`H@8s0AJpwxMPakyB*k22oiVsN?at>{4b%ja~DpT^t`mtcq#sPlk9dGc2R@QRZa~TsZe`B9nzsms%q(#)n$xG8hfFu< -^zi3mmpsBp&OSYjF4&T$%0`?y_zh1PSZ69ezBkCmGK~aVPI?dsJZ#P3)1io@6_G*M1NnRQbj8 -?{eWp_QJ1h<269UfDL)#n3cS3$X_tjpPqu38QSjb|+Vbu5pA&L+NJGlf8DS!;a@R+B9!oB%S%}TS6dwBh&zA6XIm;^%!DP@2|7)`~ia-Bx#5XgB$-;ev}`s=HNfpZ5gp_M|Gt&iLVIT!p;1T3(0WskQOdBdTK$;NsUF}l>)bZ(qVST2rnC -!g*_=+~8*xqS4mQshp=Qv|wY=agK|20cc7oZrTv1EAI4T@omYqUVxs-VuC;EBxFD5hYO80HP?y!J!bJ -ZxJ^2X4rpBodh#Qwl7Rz*1ng444@@Kzl=b5NvK>b{OFylI)(k -BSu3z3opU*%lP(|?5wI9Jb5xZc<)h^@vvoeqGkDtQ}GUJ$$p@Sec`Ay_dY03dBGmMi?Ro-8(Y=LIE;p -er+M#&h$GoshR=NXtTJQ2f+JqmBc!_NKGRrN8k%c)IeX>h?q)MJ`*wLQB;Mxl5w6P`GM&Bn$sG?wddQ -+WFkIRm5X_gLrz<<&d%R}&!SOH4v+iqUeu^^Prc@Gj;(@os>M?S3Kh~&39_}xZK=;0--d|Sx;7?F96n -QK9lJdOdl?1mweLbT}`(zPfq}N}xMC_Y0-MUP(#Ogo^F|O`)HQN0)$pU02E+6&$ofiJvV?U9qe|qi*i -Tc~~eqc^CO=B2_kO;7RLAt3oJ1Vw)9rEf{|OXhJXEM -zS0@=bLs?xT^o?bx3U%*aH3-4TPsCc=@+>9)mp~F8v|O -o9EluY0ea6IDaq9#BHH1cF7$vR=LZ^MxF&~F(^AoZE~N8zscp-zcpcf+IKVwu% -sUzt61hU@@9hy?ki1FzC=fTd6DmSan_1Z+L#d(`t>_Geseyv|0p$CV^Ij<_D?JxZg7&^ck`LWELDpiT -`5j;U>vShRGm_m^0ihz2%J=T}+UbO9I?(X68b@pRn2X@ybejYAVp(c3-#nNoILr~LW1NQ|(HE;Xar12n)!oU*8{%K8m>R@(>&I)rZ55&Dy8kI! -g7yr8N6Bt5xs!E3&I=wrupM6yy##%_!GMy!O7N33?ossj{g>7N?_?zGcQzCY4tX5Xat7nIRUf3nB&L+ -AN$0HXACLN7l=%KHS1ir=4enT_qEj_Cf1uzt1I?$2vJO{j1gT~!ivGVxJv;OhAZ`Ex;-+#!D -m2G{Rg#2dZ6LgRrfkLlhxQfMW%c8(!DZa*A)HKhekCqFPX1N(MiRW`ASJ=PZ+)GN13beMs -Av-I9UWAARd-cR!^j|J5CUzVp8R#|=}Cb1u({uB(Vy3lotfdGZZq>)XY+rz-QArQ+AsnYWh@Z|)p>xN -gL3w;WB)@e12Zb@K%o2}_XsrxFe;#$PYuc@|-;kQ!R^eZQyp%gNxXG5RXOg72Xob!ba~hLg}-w21A81>A#|{Ccd15ge{zN0ql; -13%|q=aR>Oln(KQ9tsr%fMPX)u}Nx;r~fiD97O8BhA=7J({^(zP^*{9kO_Q&FKRJb>Jjm4NV@9ewvDr -k0Iof0woNAkQ`lb={eVTx~quS)Suw%O79~VcaQY}=?-^apTyypG5XXhE3`Ej`s@;Jey8ab*NaRsArxetdJuiu*& -Iq7H1kgqh{#66Z3$PSg;M_P_jadyTywsekLMHp`+cXM=!^IVl<05I{RJfwf1pHbQAHCNg2N<*QV7PZ%m~E^j3O -x*1zy<%Owgb9l7Po*4rK6Y1~?HA!^413tZb;*)KoDB_FH0V6VEfB2PJKN_Lca6&JO|xlR%kH0D=Yh3I -ez@px+0wdmN%r}f7DG-k2QPY+a%Pbn8(ZA&?}>Eo#eMfg0(jv2aKXJS2z*>iF{_&iHGYe~o5P{` -E6_B!6^Tkt{$POUJn);WnXW>!uQlEJB270Y@pR=I_Krd)=8wbuEWX&F6V!Wp#?s=PhP9r~7$3l3fLL6 -$G+uY!;!_I%L&QV_apT{;;#It=rM*Mf_Vm|ff~cj(=dLY%g3N8Y49ifsumc=-J?bD9cGucu!Zd9-@!I -U{IH31>dC;?WVfR5gs)dFTg->npu9Em>O%9CI}55%5Ef1R-k^%!*0+^;KHG4N2de_xa|2e?Q7J2HEM~ -tPpt?;EpZwy85Bpn!7%~cmLw`|M&l#9pUHp`2WWFKA`QN&-J53497{5!f*zKDG&lj63A!JM&+Q40-lP -S)->PR((5B@dBxCRxnji*Y+C}P0IYnjv7muBBDyu%laJJ2ECKr?c>y$FDYKP$6aXABD9pg32ac?Nezo -;S0WZcR1MPHT+eidqPB;)ufaQ>63k!|`-~w?64d$5i=FUjrUpwe)%?U&p1pvXE0N00ap>ipRgZjs@pb -l!69nxz2R6#JR)?6kQV5HHfQ5GABM8WI4+E=BQ2e_Xb_*oV|C6% -JzO_EH{satRaE>rM<>?8xke41XhL^n&RNoU9Te&qP_t1^jLaEb@JImsZe>{q;hBA)n#Phi!^v95d?tF -LcDHt#{y9%LCq2)9@eDiGuSwEX&<+nC|p~YL)2L_`!Q+tBk^Jk$s_Ti`(yOorm%CXv=xFb0{wfpYCMd -#(u!1ZTkS;MY`ldV`UaAjtHPTml+yeI8l=1gZDPU3BWrK@ndtL-A9Q-6GxZxhB_50Os;cJ3Zp*5DU;8 -b@AT&-a{R9ZGUIZ2I0I?I7{Y>f)+PVI4&O^6ZuFeB!91e$3C7+TGm#xxyIanpT=9lh!{bl6kvE1g3}nYq9Fs*D-)y<9K$FYLunH -JRDya0TKO4_6FK2;g5!6@X(ho6?QiVmpxnJQ -;w03L0xGsIJ!<=YJ|eGd-WFamtBqhbG}OsWUpRlRpm&aDQ$@fi$eF -p78h7yTA4}8ujQZ5@frr;@r~p?4Iq3^}%*E&J*2>>3Oa#GPdnmk!r!?Vf>oa&RM*D(>PiDaNgy=%h3F -jKAIuj`7;(lljU7!+!W2&Yc(v*#)PAGorc*8>!f$bdLL&~i{LO1tEwJF+QN^s@H~Y=gzvQ-8$n$w%+_ -_bJEy%a!xOVwT)6h>lDsj@3MW)!h+eqJ?-$ihKQX?b@u5{jZvSo1#`2MeRt^6(!_|j+JlmvpA8d%eXGjf_4{5{zIw*) -2L|?DxbuV_YH?yr>Pgpdd&QYHUGBScX$_Wr=%&nS6*CrIK`UQZgcQsePmZ;}KWPuOIU*hzoFETYTGAV -DVZJX_?y6hL)l0A!e`79D-xKS1y|J8IeMINdWiM~nE*C=VIfE3WU2=Txp|;v3x&!mlQw=*b@BD@CE_4 -6TLn(a+q>K1Y_Gp;*>wTqt$_e6ICLJR6tWa@2BC95CZg05aB`snR+$lwc78_MByN4!^!mc; -tQdy~xFAm0>4yrCT`pnLCeI9~!5U++rh?cR0FaNB!%N9q#Ymt*$Gj~9a6!}>v!PITzdG! -#%nl)WrxQ1WHpw5nL*FPbTIJU;KSpdWY3`|8A2#SSoD6Yic1_)E+%<;B&{y>!zcyP4dF>Qk3!uN84H@ -UxrirH~wrvA7>np6{*2;*7)%SI99nA7{B;#wL4RwdP}{H;{o|reOV@pJr0^QR6~tV-CNk>v1^eM7~Gw -L3BA=WIQpv;~F2C)>c3U%d}DRvr|7Ar -NPY?z`wv8qS2>SF~$eS4S#my1N>5mpm(vHXju`i|!<y0dQ__?5rF-tqLw7OSS5pw5%_+#pz&E2Ag%?seHuGIHjvUoPJKw4?)pul46% -#%(eQ(GUs}AZFG0u5 -UU)(@9$I~?lGeL5N2Xieb4#KetZ~N#xMDJrY?rWsh{pYgCrq*5=M6a0f54KqC70b#?^DpC_)MdIED&? -<#vG4wxKl+bm*_WXJPJ*u6*XN(hd)a^4cmJ_*_g|jvySCl$pXrBb1HVFW_qD881j6wr8)6w -DxNIZHC1?{BOif^Ko;3A$~lreLaokejCx@Q(v49Kbgh2Fs`L=c})vVvRRU1IY+rJs-`1%4#i$DFW<@N8j!@ -e=>;gda+d6FVQ+fUEYN0>5+|RBS`rXU@RbLMRjXjpN&|Q0PlTzq+S%%?bN{#c8@I&`cfgMij^N766a;&4V6utoJJN9HgZyWtxJv__THDC-P@dS< -c1wYP7mEA_0I|sOYU3MFQjuYhxoI_eCpk4LiN7+u*Xjxp -%^H;qy7r}lXmx^?9kF!}|{`^1)ryltR?AW;I0IuZn6EQ9?j9XZan-Wvl(#yJkAyep8dkW6lHN^jx>Kt -R#WAAO^$1p%Ik{VKg`<(kRHj8|HW#$eelPQNB=pCqGEq#hGrqglq5lOqd-1w|AD7@Fa50zxQPjhpWA?cjH=MDaOyW;$_6k6eN!3y -ho}nu=}qA0tA+mLazArehkl_xaE9JR-3J(g>X)OK@pevB{VTohQ`mtOY9Yk9U~U$M7(Rg64mOdeJqeIaMxWU(Z^>h+x@sZO=EEor|DVw>Z` -cENufWXK#GN!;#E!D$`w=$l -HV#$6_-g>Cux5_6R-!*xqCWd=u7pkHS470^S!?#+3oJqc7fkwUm(nQ2}hE}bk>HvkiY9_4C`Uq&xT;a -meUd}q6tHkdDvOfr|IamScw#GE(}4dnKw1*jUkhTVA@55Y-P`mb}9GzEuEkFM0+uo!?o(??nR1Aku4v -ju2}8;b~{uwT;!G+h?iN&d%WnP;tlPh3_qP5*;3AJVTV#%U*|h^b5KqyUY(1FmDO}jI95Yxh7mp|hI` -aHv~&(sS5`?6(eF4~%e-;KX8IvK*j|@jE5EO0q=>{usO^G7gxyXfr#bQX5|*8XF4)(e!unTM?9R|JW+ -N^h$Q}RY%=ZOf??=>gfEnB{P8duE_5A4R>FSfQ@4wENa8t?|H|`~JijF2W_0aM^s1Lk{7Dh@y+^)}2E -E_R=+^;6*%d&g>R@Xp(sy_HZe84j)bYH(Ft|rx*Cq}YgF(E`@S~_|I`&uG}J_dcjrRepRKTm}4u81IX -`=^je&$5E+i)TXNsaA@1U*4Ut71#TB+#;SVEa81G3hL_>n=|B;eL=E>(j&39sy>wt&2sNE#Tbu&%jooo+5c?9` -iVw&lSBi>(qii-WRL;k+t{`ao#6bbmWPjsts6{b@3UTg5?2sHc3j%R5tD#B1|-8EKjDj>Wo1+-HrQ;$ -O;@d&I_Z>K+eIrt#wx-LtnNpIPeF29R*AhI9ME#PuE>?`~2XAJ=D>v?I{=@fI|0$5+E7Kl0?Es?H)XR -Z~hA{W6Ely_Mc25_#x$$;Q?oNJM|&j=f}Ri8=B!0q^eu#3k%1IB%t$fvMlLndNizRnkUQ -2J`wlQU2!!Vez4{Gi&IS8gvckPO7H4H81cu_D^DSr5#4&V5FA7#>@7C}u9dUrVzI%-dVASTQ`&~ -hJC7;>B*zfjD*q!i?raE|&u%#-FZPt&oHEnpvZ^_qelYDDE|>#XVT{~+=Hd4p)@}Rz|A -L(LuWHM0#v&+;f3{Lfs|DZGl|B7T-Tw68Z%plfIPeEz20W<Yf27zvJTnsTACSYjO*VmBC;jg%{#B{FK84=$V<;d%07dFblt74R0X9cgLIh)g -USS9jMyzycJsFezJN;FuOK-_DE5k##8aBZLwHwzXF}%MWHNVbcYz(`p)`vKC6Ew6!?pnT}H!2G!?9Oqz`UL?_a -i2zFT(@Ac3147areRl=@$K%I_o7fx*Ss-XP#!msta=D~ma+b+cl~mg;_)^p6LBp=>`0tpCv)6cEjE){ -`fgGD7SBt}Blkl1i*qUj48OO~Iv~?_2i#rfhd>jr+lJ1-N0Xq!#*pT?2pP17)ppFLvnQh)hs#=HVF78 -S6tM#c9D}n;X9i7<`h^lMpJ2q;{O7N|N<{KeT;}HNPN)`@A)a@KUCdd@V8p$=@}& -BqaeT_j$z9(_XNIAOYuKHOc~2QSynEpx)AkCY<5JS!V47GaianfRTQ~?Lj|^O<+sED-houb;Hw?1Sc! -`aRcQSYaCQWaVTH{?SV>2ldmd2zr3_NV)jaz1Y^0*_FF;SH&`i@Sr1x1Ze_98fUl9yNA;O27(goWU$A -UvwQ|0;_}kC^*pKHMr949$Fy%QXpoO$|wxM0f-)OxWb-crulXJKoY2#s_XE-(tw9y}NTANz$ -rCcVT%)m*)NyNt3ztp!buei{LGLm$)W0x1Sa%Xk>dD%Trw-csoH%o)PV>@vBCB6jAcr**es#({ko|jr -;G2%_|d~<-<-mQ+D;gi{rOfQa{vF>r-4;RD@Z-Rwhu<&g8GyHP7jcw+3U0=Ii!jciQOQ+K#HO$L -4YF;0V$ZaVdp@1yB -oL>%gQSkh1hhy<{voI%;x#ZyR^)jsX@^p9o;1JH`6k|7A-S;QoNqv(gDQxf+NHC-h-e*>f;0TiqtnzN -`6einS#NQ=a>&{E%HrUBE=c&Y9B2_DtagK0gk!Qiqb00GBT?{Y7P0Y$O%PeSr=e)xy=G?-#J~{76?$| -TUnG^E5tCsU@M6DX%cJX59o}@-^hU!-!Rz$Q+`qfW~nT|cF8g%Ay6xBGu(&#u<_>p(poqLZ?No*VONz -f?qa*NAE)$1peoXzTny5e&>F2WUY6tQ-WvABP@r7h+Pvwl|*ghI=;IKm$mUYz#$^*uFC{}P+XKu3tZ2 -**U`%fX6FnPujSw^wUpN8R(7ys5);mxk*2q@+WNFOA=)Z;YTdDGgo)yMJ}q6M7Mk?krimO2yw|9m&-C -t7^b)9%|<_6Jok`wQ3%jq!|^TkMP@BKO8RMvU^Txx9h4BQm+lU$6a|2^@6^&vM5=YwnNK7!>$MaRYX@(0iv_3WgDbA0zdbQ{EB@TT4u=M++-h0h6quRn>clHmOY~N9&?*r -HWn3(<6!JiScpAY<=n4u&I69~e91VIMI8H6HH0;UL*q!0#&ag4-K^wYgL;K@#X?b#EX3?WT7tDYQqwI -j^7P)q??D6xL(^M)ByfRS&u1+o)3n4J~`X!ipx=_W0uw*2Ik72%r`2fghF=jgvN=2s&K44BCk+2(skf -?nKuVHpBCdBBwv1B-s3$;WIP&+ckydj{6@_YuUWZbN^>$p9sTe&S?b+=l*CsSdtEB7ay3{W!YcR1N!Tf*OSvVU? -L65Un{403#3s^w}i&$Fa8QioDy)48;E;T?Zg4&fb&Q+}zlD7)R*Na*5P!~~d{6ze2mkL8RJh8%DUVVk -o(=(n2^`hLOGd}>DWWN_WKlE{<4VmuboQOXKlMJ!kRF=d<<)SOo9{Irdh<~cEtm5{V}Iei@QazQ6jXAZv=BnbH(#XlT96Ham -C#!d -+Wi*9h_zH*!Ou{6AqA*S&D1k8uz1FpBJqr~0p9Zdy+YB`WII@<9z`ut9aFb_XJ7ldHS6~8eY@35+{CT -I6Q$@<8z(k<4{6){juU^!T|SwFDncKm@iZTjx)*0BpL -!_$O_z0c9HroZ>uQYT@;QzZ05O@~As9s#e;U -l#BH-NOjV$4=S-Lj?=kfyN^8?#g*U$Q!iiqH?2DQkn;Z(ULipMls$2BWZ?W+iId>4=MOyP#O(w>MQ2# -j_u@0?Gxb^BMBzzDPZ7gku+oeU`SHe?X;JU6^^g<6bk^jmUbvLl`LniPMn-Nx!n1hyCJfP*d`9KmGYn?(@u8#V9N0K-nz42ySl$2mbOsn#0P$nbNBnBQM -!9WRh1EFy=FAk=fRsli_f61>a)=&ZY@oWQ{U9DM2nehC(#5V?cTymYnG@ejFiq6>gJx%hG3LP;A34C` -ZvV$FTO=yGVq?f=c-DOhVWuE<$Q~I8I7{pWL|fL;n%(zCY|oDFw2!KA2$$3SN03LSi(AQaFR*2#T%PN0B5&Q=j793PKqMXjei5VwG;H18Pf -@USSN20X}_6>%owJXP>)@AT*edVq2i$ibWU&=A$cSu19ADSbACi`9u*BvVm?A{F -ju%Wdeh^-W`75y*TR}j^(p#KcAth5CFYD!Mrd1uD>u30-*OFVBVxUU&OW0_TU%Hn^d9xW6XnA(A%1We -~fwSy+W>74f^pLBYy21(Cn7)b~$ZbguY?*>3C0Fg(S~PWu`bHQwe`Db$^XVQO7 -h_FqBjhI_MCR`AMM5If%@<0BKv-lpUgwT@NExIE_pk&4?3}4QGdIV^eYM<(?OJ$Jag*@^thdg{l -jk?65D5&_F?LV#NgQR8nw~mev0EaB3+5=qwBqsq2;hXvf3!q1?~#s>por~OsscB_na#qMF&D18gJZH^ -sB$fM2Kg7u;4DLpOvX)@9M3&$)#9L=0Od+`mOqJLxA8Ze?F&sb>@Z}k|t&Jcm*^4FB(Vh@|<3Xn~)m%km=Y#sFKH!FC3JTx$A -dtd%dIHSydVvAMZ}N6neWUE|mJE#I~t70xT5OO*2|@&(fIwHB^rs2v&g5FJe>p-ucK@@wtDvR -_?He@CGWW#T=8iyfTbuTqw3yIxz=rCRSym&-3og`f5b1oq250c*0AiKw@*p8CB*@eAUziR$^~`KDves -{|5#yDd%^`SCt=NWIfKA7}42gmcv~S4xWBO}sn2+g3gYPF9}wx~1>qw>b${L~ZP1PwC@nlozdI&D(*_ -9;f{^j*Jd!V0Vy|J>u|fZDO*@&<0+IYY~ULm#{}%mw`KsE4|p=2^O|O406T9%UqI(JRNU~0zFkqqG-c -<<=l%MSy*{luI%^$n@z=G5AOc~dG@bZ&S>qTsd`McT6S~%@#@FQdmf>w@F_TG;qks`Q{v^!ZMpk1%Au -5w>201h -lrxZhc&&6K6Q{E*=PRIj#gMNOi-NE2W#`^qyIA@KO!gZ^%-5>P#OSCl`LSFZQhM*a!c(U|8bYlm!hSpFe2XtuXFHRDl9F|{d>`OuWStA47ZH*ZF5)SPC%L9jp(a!JJYYOG1=?2xtnP{lSCp{6R)DiKC)zsPbx@4MMdjBtH55gnxq>K{Y3usS{&Ts< -Oa+dp?iF`Iu-TW+j}?aKHk;ljTTAWr(P&pLT}eKd7*Df8}A3*7(eMBl;_e|?f4gcQ_DxJV4a2^hvvn! -qT8WEhfwHxUJmVE9^(BINo2O01uvKHX*8N<}33ce1UStpo_{p>3OHD`Ub7uuhw8q1IWxa-GlfA28rvl -8`{+u-1WS0@`@%3FwVSAt|6@E5}=(0MdgHP&A@AAb9C7bADjJ4p9{7R`UdqC-54ZvB_Gx77Q4HuDp+c -L1mn7Vq9d4*G{(l#jg%WBpD2dHW1^yCCv9hs$tUIL^)SdV@GZQy5LOlwyYy9k}UVlp0^Lr|j>vS9 -czC{87i#NaeM7;2qrd~aMJ-C_G!VS|3$VG~$2RIazT?H|dw6WasYGn+G}Iy7)O5hEEKWS==N3xXkDFQ4`l2i$ITA7>wQfY)w0(>H@cVh289$BF%a5CNu);Jfo%u`|uJ)b-2d8@E#-Xmoq&aA4ml{O6j%FNgE57Ou4Exa?Dbq=eThWt{jgPx2h*yS(oPeR&X2|^ch#NoSz}Su4! -R!q)BfBq|$)hGvct*rX$u`v+noS7W9xf+z3lYj6<{gSM>9`FL~>1;JbbXt_R90e;b#Xf@uUD -otnEon`x=EsJK%yAs_Mn8#{`pBdOKsE#F;J*Bs&Tq90OAM60#JPbGP9?MVnXtu+Pf3G+#VT=|zPADh7 -{2d=Z$_vF;&FM%VIKrmJ_EooSSh~c$bD)BeZpMP9shHGF<59kBJ}}j0je8#7yF6A+e5CX+FqI&(a+s^ -1&;9?$+F8sCfg;!*XzzAH88LI}CoKP-(f_}N*c1BETaarRJfksQX?^_i;5UwKdbvf=GQx|yJ#4 -Q)B8HWbW49gMc?(DwZj85r=?sFg9*?n$&T2cBFWO;J6>Ca)Iwcot9Ut=Jyq!)Q@RV6i!h@#_%_UX*t4 -nnmalAP{uHgF69c{7_x6(#!BN(n05h-tK&7o*2>9lh*0f;T1(wkg4!pY!9n_MdEmZZ5WW12V|mUY(E` -7rqRY|E(ax-)kx%TSAxSvM=#S^)31TdOSKD%;@W{~;SqvL^d(UOl2gH{)HKm3;fH%^Y>05j>Cgw4v#r -*Q+Q+V_fhHCZ+lOJC>O_p&W9u{adZ9Hi^8inDYouQ{IVabIr^0~&VYMmuk;4fh(*-Pn-MZt7zuXf#pO -@T$3g%$)yq7HFVH3o_OY}XqXF=!!-O>2mb)q7Ib)z=-4iYBF?|i8v*4{^@`5p;(hd?vyoR`5uH%0)r_ -l*g{X~fpIeUXt?)5Rk3xTeTuh?KA*5O={1gfALz>L0Bl=X9K%tz}HLNR+Qu>}0724nKcMRb@xmn|pvBT; -5&ARogV9BOT9CU -aHQ*Uw{P733$1%#77%FhY9Jrwc3!0xdB7Q6ck2l_p`!@gyA7!H&B${GwoBn*=fOzaj;G=$LTM)Wo%ia -{ts@8K`>UUK*_<^_MpZ0#%j@0IwJ+86S(cXq)>v7mS-WXY#A#*IG3AF?}~d>bv{|0dv_Oi}M(gN)co7 -P`-;?hz~mwHu(I@V@br!w9{P4g9LA@U~ytOTBLjx4(e)DYT93;cqwK9NTNDdpdWz58tB1UO?Cmy3tYW -i@V)Enn3P@4*9#EZl9r`_vs1jx0%^TZ)bPu*Ht#xS@dd_h~R>gCl?dhv?5jIZhG+ala<*M_$SP6Cw9P -p%ItLR)5;t09kVOtpPAiWZTJPV+ld|UPng}l5Af&A&iqp^Cx&rez(_$De)Kpz#lmAevsom0!OJU|lsx -t{vUb<$ZevGo!0+n7=COjuN-oGAb{(#U7e4{@b&(ENpK8cS#7l7Q&w^GOfulbvEQqFOQzbgq3M@NM@* -!qA)GtOkzfRzBLMf)lg89XI&*Ww0$ -)P&cCt3zakOAdY=wYeo`jC@t8n+smZl>iub9`Ve1ml)AL!>;2^h%XXFYY)+raL4!mPHxQZLuV&#mwg$ -<`eb4ErJ>}1ptniqmpA=r==pbd{_61g+a14!RFvAZF^r)QMxrE6q69)xC{B?uf*~XdL)501i49qaea! -Y3!MJF&cLwjSO3AJ#sdvEP21XnDAo*T;h>*R5Bz|{_z#m)WL2m9Va`ir_pXYu$zHqyQkjXo07lViQm$9yUy+C#Jgg$KLFDG1Ss~l%S+xnL -c8_Awik--7qmSgHWf|%J}^;ydk~zzLaK=#k10j#f#bZE9$VBkM`@9UTN(QqQa|b(`Nj@iJGyWCIOl@U -&(Prq?O1(j7`d4UJpr+&moduL-wb|z@z(pMuLXL3=L}jFFMZWuozo-Wt2OA?jeEfs`yN7na0ca=`t8H -#BKiYJPVTq%cbZgvbICSqroKEz{+G&!`E6gzwEK|)e-4?Js?W{r!9<(rxmlvtEv57<>xX*k)mLUfM;gb^o=UOO0BsBtdXJ -=TN21Xk8tM1gM*ems$lpyc8$xXdgZDMVuOqcTq2S)m7w^t?)H@6?j(6Bg-Ua6PlTi$@yBY6Eg~U#Y;r -FUwM7`r}lXnX4dzq0y_Bt+%?Umr&-vHkGaj@w3D9C+}f{KBJd!lGp@WgVpNbvyT5p&~$0Q$d)g5{4WX -hOPU{fL75M*V*i1*PAjAf`!TdwG#HfeID}CL&9tpP(638GAL0`_aRkIt7H*~;DK4(sL#AzXywUI`Zpl%s|Bag5P1%6%|^W17`2WhI>RA&nlw$M!?g;!1Yta&sn#c6W3_bkre~XmkgA-cztgGr3q -1NV_4OZd!C&wE8x;Iz$M3ixzUgO(L{Ws?ppV+H1f@4gjnEX0Y~&B8Hd&1m2!z7eo|yBY9~G3n0|fEC? -QWw@1p2Nvq&xU+K$2qn=-5WNHg&xJ@#ChwP3~?)aU*G)n%xeDyuH`A7t|-K6Y}3Qxp$y!Z1a<)d;ZP| -%#nY;f%(G}Rb+33&gpmZYP6HTjbqaB+b|`6z}CPC<+i57_e`p~rZdL5 -+)6LkHZ|=9O_RzkVRGa(iuoQoWv;RqPk$x6Ng^N#7;D`iB*IdMI5aV{diby?u?#G -z`yYb0p-Mr&BM8gj}>7Xpnz0E{xYM^A0J0^gI$&G2sL0(C+(K_z9zTI3Oeq ->In&)fP^EJ_xd$S+6U}nL&7d^FB)oN$TMO`lH)n!_lLwBGfFdCqfFHU$s@~M&yrQQlV0g~#1pC~;Flj -lL2ENN2L4gcWpvihwo7}nOtev&>N_K#ogd=Nj&x-==PA7DHgwGSKq%tj-LWDe@R5o&O*3!{EJ}$4ewO -tqt&m5u9-CAHG2vp{ouGPoBHkD|Nq2I}=E#edJ6BP15^LcxcpmsRfq%NcW9kK2%dOX``HMc4jgYp>=3 -;K+gblBNOVqCIT5sDVd`B*i~KEP_0>$()FXfXb8tZaSJ3G@_m~!Uc}(Qy#pyMut#ss-eHUuFn1VoS)y%_y5w`6Zko -8{rfH#9Y&hwZZ6MJ#b3n5f*&wX5KU9+Vcca -}ADIRxDev|0StQV5&ccz5xz@=?qIF(31Ds$@KDa?S*B8T317^K(V_h7yDi_+eL9^tZ -AI7?CdwM=X2ru_9pGP7f;_1RW2Z(h&pKr=SNv>oMNvPZcBz~sFM1&WPR%D1%pNtrTzP@<(NmSGU^J(J -NazII~q`{k2F-vS`PT^GAx>k%fmyP%517=}}p_1ydNwKw`3-HBu+Dzk1OP`oS3-TBcw3zdZx7!L!z-s -KU-dY}z%Ea=OI3(v6K|ALQQLmu#9C@cQ(hG{6I|#=Z+&#b>)C0A}o63;@TShv7Fdjlr_gs6lV`cPkr9 -El2qNZ+*D@tl7isD^_$VX9{(n@CZq!#bMC9mA2KX4^^c#LKj2I(t-54JyeQz54PT$2)xm@%9djB_|$H -7RD`m;Um`@y?oFn{2%7sKuyVM60e(Cf|Y)GTp3%?7E(RoM(*~t(UCa2qmwi@s|3!Bj-eUC$(_!PRaWG -@6oUSK!pB!=f6z9z7wGs1wq87F=-0Lw%>7tB)9)zf+T1dB1sJ0v+Z$;+}l|wjQ(&bZsRcs{U$od+i8> -7$GzycuV((HU+G&|M%mli{1@Z?(XKJK{qT1e3`Oq>6iePMCvWqk^erDZ?u8)x_KhY{(XLpdUxnD``99 -jVdqI)!W4+_0v3y^g+ep=(MW5^;guBUSx_8%XtnpK95QXe%KzqL8+i#T4_lwvW+uI2sdwZPbf1iGNYy -_TB46lQe4n-}~T`szU(FE&a|F`MaPePPOfPFpgdm+jre`h)Fzd^sGzd^sa>!BdYh+*u7N4?>w@_=sg< -jZqvKM2b%0EcH{^^FQGrIj%RywMt8qGekPb<62;tht{MiGpkHuFEc`4JQbw7K!c-%8SfG(zy|0^Muya -9J0O)?)mH@jiStQNOq_u6^ -hlNflHGpo$hnHOXnY++~CWqe!JnGJOq$ed9749M6iY>rde><3p#>)f%kX^_3q7Bk!>BoQ%L7yd3TbSC -CNIRlJ+Ao`9WhPg@n9?Z6;6Y141zGich2!M608W+6RCDDc#?M#+a5eVmvW& -P{30Losi=<2KNd33ymql~;(N@8`y~2o>79)9jC35h%1nr;DFDdLs7kUsOZAhKoma}b9Oq;C>wb}>i|T~?+$umfTe-M>v-&RHQ}(^%|Q++;jMgIwht7U@x*S$Cvp}~u*p|!0G1g{36!WV|4-4l4z=k{?I^a!zBx*+9felFC}sgV?@t9oHnWk|iJ5eZ`&CUUNA5{LsC^ -7`ct^`H41eXK-1Q`zBhJn|ZUejqLJ_^=$m6EMhbk4K7h2 -%Q*XWf-UEkg6?M~c){cu7z#ALu4Qe&54w5XubL|miPDzQQ@mR4s@EDu-f!sr#z;5%M7GO$^$p{Uk5h_ -f{fqqT-6mO`Rc-UHuMu6}4b~$N0&NV6%jRiJuGznSdiSY`#UI;F{wqiP+JExfqkT)%NMv6egQ-0ZkwQ ->}L~xvhP>RNo?T;`)peTiZIO3JS`>5!?tRL+YsmP~#Y5MjqdebQsdDlJhcQrNrke+RSpHn-h+elDEzh -}MReLV_+_lA=++DEQ-hDGg1-Dns6_Hut`?EKoToq9KoY@8~6pM5ut&-R(Fz3{qGwdmg)Dco3P67Mgz3 -#PGseU5EDNmneVCp=Ie-;UiN;e)yY_wI>ZR+EAxU6Mq0}sr~bEN -^Y@L5eVO{ttdj?)qFzQPckkvHLx0-G=)cVk6M@*5r7y{1pXS75H}qe1^mYRHBeLsfSAh47*c13{3Gn= -63BaKG_Ph4U5&-_RO*s?)$)FUWNxi&h<c$+YQ!tVvEK~R6i -TJ?Q{j4wRiSQaXd*OqAb#zhf4`*Uci5&Q=w(cLhx+e^voy?;qdiIUXDxN>g_VuWatreTI*eK=Yi1h4y -a;W-2ccIpj)dA;{3USUuY_S%?-GG(IRk4tsI&{*)iDn}@;uv0}Rd_&M)w)CW(k5xjE!}f_QlB`RL5(# -}-68_V{9xSyguI#L1l9ixR=*m_ZLo#@S77yTAMsCN^}C~er+=|c>LLh@qq|W63B%|ng(;Y%ahRka3dT -u<#OaSHN4!1M>{msHEPwm%CHq_Bct_Ly1+kZRGit97N622=*c9z{O8lLg7L&W;Mn6q#Ks%UI@ -$L?=*O1Bg2rac2KPckg*zetk$31=)|HCOj3SX|8l?W*c=%FIpC&~u(`r+hk`J>6*--gyb{8jv$(E5!M -W|m!ZuAiZG)i`ZAbNnWYFOdEz90S|8>Fckd71*)$pFk_{MZy0Ww0`T5_lDN>kFYY}x9)gn2yQPYhN31 -V@*c+neBy>%Ok+N|7M4r{iOYdk)H^D+I2E2WqA!&_UF_X0Mo{#?vym4MS8L4Tq2dLDqPG&S_qQ56=7^ -n(9*}f|b+8RT5oeMPA1o;w#wqLSyuChK(V#Y^0DFykO^_C9IN~@fi|2wS+a41!=`{v$39cIZnL7)~Du -;IdsEV{XpUSzzlhw3dZUDoX@>u<7;IY&3P-410i*95(!SpaR0=OKSPT+$8uT?9a)a-FK;N$veFwLe-# -B{om+|Hbix6W^C>MC*|oAJ`)E?VnhbmjtJ9ld4U)A6}!PbMkvC4>xw8X&7b@l%$kx5L`YvB^kvs*mbv -(GuQYA0RDAHfn+W0#E^$A=otP^{vWP&4Lmg-DAGxjL~1IvL50ameM1d1cHbm%*9!3Bz013j`Z6=qB4- -Q*ke`a89A7qd#ywbocd{j)#`XVCLngXt-g*B2_Fq)|8ybb{JJ2OqdunUC|9mg093CG?2&A!>UkDQvly -hDRRdBEtL>cZsiMH_Dof3Pu!Gck%c}*mP4*S+;({`%ITSDm{0b#E4oaF_Of6AAR)>XT&WBc06``z^^a -S#hK*RVodV(_ZiXV~dP`Hl_E+{6-&;nMu5*9&NZWq$_{WDmoN3B8)R -ga>uw&iPR>IHCE4vr1>(OqK5@7?i!_N)6*5l?*<07VHy*YJTF^1*334^KTsi{0e6en6k?fDoX=v!fB% -LIWn$9Y05PP{VHfnyil??6W&wuf49_fW!BzX9C^nLG~gmqy@B!>p83N9JJi0vdUkw(|*v!(Cku0FC@YE%cb@bj%&H)V07Bse$bt$+MIGv&^5{zX4p3 -bnJx1Qd`=VnA=uwloI4ryr096Y5*Q=3~c4{O5SnFIZ9Ru@akh68ZytN7G--H07#VM9L~jutg<>gD_-F -TF8EG<$|!M()_d1z!72qE-d2vh?arYox*x@KPj)Ym#u5;A7rAwWc8o`G|U!^+;Zf|@dQ9i(X6t(>&w;mBs<{P;j&EtAcB~@A$QrpI%$>!JljM?BaT+Z>mA48O -i6hOsZC>|#Q$!&0^5~mT1sHc%9_@Rr3LDeS5KiM>QGsyUIU@(TP(UQjC3rHPQ{ZN-cfRN}h;F0H5PCZ -bSCJkN5Le2T~)z4nWi(ajl8YyjuI8I9x@vQX%yYq4zZVbg>CUBq5t;;;o;s*AZ)TfyJ#6i-Bwwt-4p< -9y>3ph?V%1Rm3Qb?@ZCVJ?O4vz`PS64K^hAxe19mg5;U5t`m3W44wIM#IAw@3JIxZ??ur7?ze7I<4a> -x}w#COMy-|6%^+^Xx{|gxka2_3zQ-GH+gRH6g2?u+XM*0zW0vgQbP%#n4R3U9H(WvYGv(wi&Of)9WV> -$PzL$+@Zc2U=&cu73WDlYJOPzJuu9}NKyz;a#xsuUAHt!T{Ub1Tlhg$6Xq?+d+RkPr(e4?@K7i#BN8y -v*G@Jt%I34;w`5PS~>b4=$CPd(sbJ`JxDT+7oY)443$CkE2E%l1@Hp7IyE$Kog3lXcT$Z3hYk7pe`X5 -_Tg#7T8FW;lSo(+H7FC((_Z;#HOA$HKuXZQ*$)rQ`kiE)K=~FbzAxW+i&4Ft$Lz+Xh%M>(AP5b6dmS3 -Mr(ZKt;9{c+~$jBUhWyMEojGJrHgR}N(RYjH(ZaO4pSUEm?QNS&VmY2QJ8%rjzF?J#IPHcl0PQ7G$2% -27vn^#q*_+-p^t+&bt<-4yu>Xg-2TbW7WlInkLn|GzU=6lzss4QJ0)OztMBKHK96lx-{4$}!w;O -aG<*gMVkulQ;iW)JtKvBzy4UJ)+d4ZfZYR%7p%e2%%A#uA6T$?9qhHbaFpmY#sXj?C^72y@W8Mw_({^ -KT%q<8Od>ys*6Wb!P4Se;h{g8U26u*Y5xMy!l6={WtdeDyM(i=R2c493lw{qbUMINgSmxj6w(s!B7ms -AQ+}GibfFR!=Al(@1NT+X~QZU*`3{YeM|2?pqp+@-$^?W`gVERWbDUrXq!^r^L^fFXPYiYseS2@eDA% --_E5WW-A=M;+zi?qj(2OTw~svWs|-mr-vK^;$A@oly}|8Q2^_!e4Yot%Z*@nCUCw|q8#l+UiW^7g*s_Bjnl_O=g=2M -K4@)mhE@>{Dr)D*sI1rc}T?e9;=8r_VsdzqW%wdrG4onw4jNGhg-vI$zq`m-GVboP-0PvGdbL$Ndu2W -B%1E1AcXBfAz|M{nGwM(fD-G+6*zPBEQ0;GZF@XLCI>#tGM`0f37EtymNN-DEJYUZP8B`MG^BGjCjY5 -*RL6hDhas>GWSOMArA1nbsG5v2KfyihM*J8}`x39&a{Qc?H -Qa7;h|#bU*V4Ia55iddogmdmRvmpeV(O`DJ_aGgUlP+3eo29#}+3S~jka2lgUT4Y_vlyN&iYlAPh2xQ|xBjGF -y2m{eW7)URyY8K`&cTa6TP(psVao0pao}ta48ghKX5g2gndITzw3+(HS%8S!;Jdd7OnWr_f41!s$=fa -qnXVv#1v>>)aB1n#W@d6P;f}SF5PV>eG@I(k{+zJ%i*U!~5IAO;^0@W~IipHP(41P9WG(V-_0e?Z^rd -OYLb4|F#c-rDTg_`Hngl+OsUzrs6R!`qA-c=A+_R<`J=xiA}#9fDs(Y-SHrJn)7oNQyeD{kE2ORHLn* -Mhj(*H}-O8SOAv9nlEkh#Wjhqbetd7M65a8E*bfjY)KYCSW?%j^ua&+}3(R(p`xj%S*CE$cu=uBZtN6 -97`RXUb$1GN<$j;VxXMv3V08_tP1Y{LA|&e8{HT>!UeDh;}cXeckmU^jH1{Y2}P|!MpYo!rs5<@?c}8 -Z#y56d#xYa$JK$`$W%Drd;4xA3<*ui4dgV{8s9-__wFHVUoF$#Gs+Y`r>t$nKdl!{n!{yPntV~XTb#( -BQXGm#14+Y#_cIBM -hcQqew6vp9V_5Hnqr|hzfTu)K$f9^BFYz!yVQjOL)Z+U|u1&-$)Ep9ZxM!l}i^2Uc;Q$IQ5w5!q%tY5 -ErgI4|dhh>cdwFT885biRv#OGFO1=x`i0n%xH+ttsi4@Lf|V#p+UK5ca5wl4bwtM)YeDD^bn~zLJU8B -A*@LA&`muC=r1K#O0~R^#dXu#LujrKX(A-WnKT$24%0ypAKmj{v$BdEM0=+0SXazUjyKk+1%3 -~hPh=Uw0=&B7-P%X7q3)b+#|7;uF{=@Oj`8U9@Z39MQFT7X|o&6up~o#(iw1Vhyrh-6ey$lfr%J~!jklzqyW;FdcwT(0FcE47)FZ2 -dmo;=5~d66W3JUcEgFN3<%nJG2`g!X?xX53relj+7>N`HZ3SVvUo{?1+ecju3ljUq+eaINjrG3dpdAjzoKT)>T6aV-C$>ou3cRBfVY%=D}c -y_t=<~;%CMot`B2Dl5}8uXXRFra4`>nu*z6><_|R=9WD(KqgB+7R(!wuBy{9ifsc-v$J;zjZ{NtnGcE -c#YVu=#8om!bFd4XRDC|eLF|?S~h>!IMEB)mb^-_8`%Z`rMuTU;?_W-+DuCIcup_1U33*ueYmxmn`;7 -XsSm&I@2+;RG{xcB*?#>~xx+%*3DA -jt7aTouv`INXe~}P0etgm@fkjd#vw16;9;C$QsW>)_gFeZqd3VYi`wHxS&n@nkG7MBYa@n*_&@{)B_xPB!q{ZDCjR$-mQ&!%xz5cf;A6p~&56;*%-$9-4x^#`niy9PM@a2;QgvkhifTeevG;<6*Wk$X8cI=kvqt^S1 -x(hne|z79IrX-e2Li2LYiJ%|_>m+IU6t3A;Ug744qworF0=9Zp%qs?YdnmFHZ1qA%`*4yi5ap6x?Qcz ->J=;89C?K)vp9l+3Q8vNeVv2RVR8bx7{n@vzV!dNyQ{-hiX(-2PJ6+rw0O=~-Hj&1nIMPJTt1*-`v~E -liUddoPAHzxDl;^sJZMqkXy|%<5#IeoR!}d#UlJU(3v_#519`04NPcNwFtceq!?B$la~Ec%X&HpXZZP -bPn>v#Z`qPNp6xf@6QQP46_rDcD9+-c}^eIbViRv_Y7HFP5$xF^c5<1&f;d^b;r_{>+ -bV+9z#o-8(HX7AZ0j0r$2OU0CTv19ugcSDqi0lez%aUhE^4;tf98Lxv~w53^d2c}FT>ZYnk`&D|9_y} -$>uuxq)_D;Ipe)vU;LMSv|8rzBxvHe8m;Y?jjc|9|3e5b&CQjXVXO?oRlUn90yz$gJ=>b90MdTQ`1%y -~f@9DMKgRH~uUI)yz|_*Hr`?gU3KytxO}qYEw%(LF+OQq!4PjsWEMNMQF%=WoD{1yCx)9G#`Gdn?8+r -yGk1T<}3;o=VqwA^MthsP%5Sx$HWhC58oTEf!P5ISrc5*j9K=kISPxQ7{>d7e*L2n)EWcq_?#f_mZo6 -LQ%KUG3f>$Za!#vD!_D?1B)tN-jB3%B!lI_<(N=_s!j1tc`F(7&%4!uFYW5TtIf0Bxjfcb8#!5jvCx4 -4pHsXT6+jWpugwwVh)Ig-ly|FEJxzZm7@`d2^Yq%d(p`A*8pH!baHuVhF~bU1%t+!8OE9uB!|j85Dp$ -El?ODd88BrSxS-D=Wq+vmRXwhkxxuXpf>|V>{#bED&_mPj$vs2cJ_S%RW -4Y5YM+_pWNDYSodfK11thO7)!`g=TtZ#R>kR?9QCt&HhiPsVPU(~%gXfS5QB$ZEOV(~iIBPY{*HN?q? -r4Sjwa-=gWMHo{u4Pvt!&A?7_oS|k@^Udw!qArUw5IuW0|-b#5U&+bxy#}7fB{IA(nwjx19$htW*U`i -rd<-`6$cXgq_l!L$?XO@UfTB?K_dG57~_K#X_IlB1_Itb -*A%QU)u+^-!0q%3(Q$5(VNoOFD(e9v0tpr#_Q4SLd~@g{2C68!RB4|DPy#Py!p)o(O_bmo9*pKHN -y}ZX*txi?dZwyUxjwVxkh3BQ8f)+N4L$K8rxY_vCxC3}(i9x*f5N};f5^XnyZ@i?FZ{cJ5rlv!6o)aK -#>kC>5i~~A1cAdaMiK-}VtdylM4%K5lN1hpSWt)2U2bpuCBgP(goxM)%cjsbrjxvf{GzvEVT$ZD2O~d -jLw{2u>TM>o?Xl5U!47)ONj}6{Q8npANJlg~7P;wt=Oy1 -4qpKAXwxi47Y=w1|x-u4XiJ7;b?&7L{}?|wIX=+fJMd*gu=^^ULmeL-E?Tlb=`ZRoCW4a6sXl{*seIH -{M&oP8cr1U1?ZJP`j0_u8o!@VB^^^F8-EbJ_yvPWT=7+Bfdo{tEZnsTc51xYr*wE64p!9o$r1(bl&QX -V}DsePM2x)Z(Y~_`-naapF}U&c&peZCEOb6~TlP>|Bv!wola~&x>1UoTwOwbo)H_RD4$B3^r8gA={m8 -0ihTim~3T{^F-9*)j#oi9$}(N4{J0r3!;RY6kQI#6xYBnJI24atQpMYHiNpKxo!zRE)VFs939hsH~{_ -QoWeK8U5HJ{d5xbtmKPu1{m9b9pp{ogslx4^Nr|-o&>t|3e@nH5ejLhWch}_L`C -an7r&A1XR3ySpY=muh?N*Jfk^;ybts4ebSP>7)K|Ph+(&Wn|4d!e0iQKevHPPl@>gZ8Z<5o@INU?n= -CDe-~$Ydymz8)B5c`&klp>&Sll!t=}59Zv(z=6xyfF{XAA9IE?cJeL3EozB+7uYVp1S+L8H!XW@RiZt -XeVkv+Pf@oa}PiIaY6>;A%sfnT#s>wUZSIuLyVQ0B;b_b_PMr_w_RkRpiEz!`$6Zl%9w>sz8fa)8{VN4mxCY*)c22q`N}MsM -Q;>e5;+?x7xS1Bi_TwRn>qtIB%7Mo=D@qEeU?iry1-fAoA;ouyI$uNocDw458Oh-ZnJ(aCd^F7AQPBn -~EXY}VT<@`v*(=+JM+dOF>wUrWz7?3VTOUZ_+zT>4;twQvA=*S74)n@U7(Z;~qWi~SDAr1PIgBuxpaDl{jdi2gwRA?OTog -U$`71gk*>A{r8XN(aSXJp9sNCz==r|pQmxa0XKJzjM2NUwqcfc`5X4F4=+DR^?*tA#UZc7YNfEDgI2D -l*pNGA8-(cumd3ng{M0Ig1XZN@YL6*al{6yZ4`&g>WN!ZnP&miGoyK+RIJbR939j#OfuOUUfr_NUCR3;ktCidgwXXy6!1<*o4tAq!K0Y}X;0Z6zUx=B6lfGB2M|Y-l@9IQSG!|)xao{n -+?j?0#mbh{BISn9%H(ZxvUa#<`4GEw&eC78}FF&W1A8>no4u*EkZ(^H7Lgz`bHx!UY$<-}vs`OS~Ork -F?0!RqT5pvZ#fU(f!(A$DOXD4jaf)Cx1;qt{%9xTUqf^ConzWIuA6VwX~V)v(Q2+J}aXHv+)$HW0wQ$ -Iaw^t9ag2BeeY{cb6!Q8+?;V!bt1LvOW2M?eWX;GU$k=ydREn0pobXXdDkC!lIYmk5qVw;mSq5@|)#! -V51)#U$9L$~A<@WhmJqLpu%5u3_XN*gPD6%|JeMbc_J%5$w_#<8Vsm-15~IS;_j_$Dj8QCjuWL-k9Kw -R0)wn^=+dNi)S&*I6M4vm<~u}1AH>TPh+?gu*>DTAxKpSC#MOwnM+@EJ&4bv&)sB!iPu!`OsIF<<~C) -)oMBQjP!Levk6fXQSYOvI>K?0;d|W8n<(6BlkaQqo677_a?S>@E=?*qidSHeNbjY3u44u(J2Cg!J$B( -gVLar_Jb1wg#;=wnT`#+?Y|A)T*fBP^$yZZldi0@1XVQ6D?5WKOu_W(5op#(-@5CowZvj5r_iO6j;wo -yI&F5U;E_ZZ$B-B&6u=H6a!>NuhI=1FSft-l|u)^>guXJ7TKes``;uyHS{8Mk}8ZaLh}Z?qX{A9?z7f9mP){AEvnsbq@ -Jo-@wL`+lbn;>K3x;JFAe5wif7=fde$_5W~aX6trbthGp*hY|M_5{uxDx -i{461Zoi_z_x(`v0LL*#}^qeVNT})G_x4I;46BQeF)e%*D3ec2y3pZ2v)Dos?6l+_Rm!cuUQNA6MPiQqi*1v2of97i% -Ez{?#jfMhrX-om1-|uZ9dHa7OM3@qq9~|!#VI8;}ojRS}v}GPp)=9@+bo26V7Mxyp-%3LC`d+Q_N**s -=<*{A=9Fi1qpK0xw#xK?Yc(bp-vQ17(v2e{ETv9;H6tt%W)*0Lus<96y-v*zK$`2GtfrxjCz$1C?10{ -p{U{3y%K1ZLi#eSh(wKA763OoHt)1?FmzJDz$a0lM_#fZ`PqxqycE~a(~GQQr@FkxASeb|Jxvf*KW(= -sL}xGoS8o@Iy;rvt9AN9x=<7saEr-pg1y9xBRpq6a9_uaXMM;Bh*>+#cc_Am-8OmE6!XPggJ5(j0=*hdwyT%O -+wF^n`TFd1U+$whQZJi|%Z!V%<+lDAd>JWl(@kEp}HNDmSLizRZP1tKnUJSPu04i-Ra?Gf -~Q9YA7}R4wyj@!k$h)nq84Hm)Cv2rPGB&FOWihAlnM%Xb%0+~ZhtxFEQM3g+H87aqdVU_C}6FizLfJ^ -j>qFVy(Bp(g80l8iRSqAbPwr)ChqK+baCGXse_Ffh9IwA&~x+clsz!u)@wZ4H0q}BsYm8NAXbX;OFbm -`^pH(it+t<_sVURkb%iYStYf;D_8_fU(5fPCOaRH%3TJ#Sl`Of6Y{@-e@%FMd=BamLo3=LDgmby?1g$fp!x@s!fwwUOErvNN(IYmC7XG#b9>7jMp_K0yOY^BqSnZksqt!n`mKN}p)3+vW!X{4D-&ZD<8QQ>|uu0t|BoWKby2&h+VVa -$6-)1!8dI7V7J@T6-ecOGs{Yiz=>2$R%?*O(IVjc!e@d1vb6VQl5~MY#eff)>r|)=n09eFfJAn3TsBW4ah&Q>GY=dg5gkhvq=2|aG_up2`UtzHHn#%0-u{~;d{H~#V8*$L -wQR`RDh?%%Z{VkcCxj?r9|ZA&RHJeZDezw>$czi&nO*T2w;z;*Y$>G1chh_Au=|F#t&`2Xuxgi)9s)| -^>V3B1?Zd(eVT0bL4jYwqE(ydpI`DwPTdl&HOpsl)W;J&oGB`Dp(FAL)P$d>v+iO*K9u%e==S3~Pavb -u-`acwNEpp&os%LY8*uA7YTUEY^Dl_j@DwKH~`{X!l4lJ2p#|!@^wC4FBfLgL#QLI#FJ}OlUyay- -Pk!X^f9{p{L^!h4_h}9GcHcUftx{^(4SRT|}=2}oU`3Bgvca5Ksre+OU`!RUHr4;GuB3@UC5V)a^@0} -)E@mobh5yn{RsRkx=uO9^0SA`Fh?o`T&GhOxYJa!Hw6V5d+uNb0`;lSBSBbxp@dA%GyE2shsrPulWJj -|V~HYv470+h|8oL=1K8JiV_^k$Ztlv$+uC=_e%@WMuJ^%iCYVR)9PfHfm2nq4jzs&d&#h~7JS3c6OWn -pF4cYs#zFg?@oO4)5iRD3(2ioFRVG5!Al=#2B8ip|fRcFbFq(aM+}rfU_M%F_UvT$vLQYjxSp<_35*> ->vhUIfjN6Uq)cnZ3?R$6X_W?#KDJPSzRxEcY%3tYaru1{)4rwGj%<02CH2NnO!@K_&XHXch^M?8yZmz#gxPj823D;A7=a@FlUBrYUTvP4B%t!C>vW?n2N&*VW_|13|8pxM3Zj3~ieTN9@Uq61#^w -QM8gq5u<>|!BFa}&^W~%s9DxapKckohz(XG4@%2p*zoR1G#u5YU>Uv^5@9_AeH;`s=-s~vYr{Ke2}ns{vHUUFI$*SR)U_tLf;lF$n;tpl}3Qvp8~j18wY}{IM$TeTnYzwrC -7NH*`%5(MEs)n&1j -aqT98;KU|lXHFbpXS1y#}{DKq(CID|DYyOz?1RD7;_FkXD1F*gn}@J#WSqxV6K3*lPXt?jE30NxSx9& -Li#)h0_HtmrKlRgg`KQx%Kn|uBWEeb@k`3)oWJ(9zHurDLH_k$|2lGue!J)_{-~q-2lIc`(f#S%?_)4 -2h@^HVP9pnqOX2X5gGmwy3Q;IbK-6JGh*1Pepa{N?;6GK>5y$z1J9e-!k~jj_j=-Y$*iWNB`&2s@g7H -HUMwMi#9`WO{KZ?PmKO#QtpoaU4J9OAO9C -8)x;BWAU>>Y>mkWbd|(OCEdT~Z&5{x7)| -e^eeWG>`X@ER6k{D7<|QejEF}vd)lo+gteM7iHb@c7YVBw*)r&2*LJ6D*1O0epjx)yY5f+9%ZOp8yRDt -_mFp+@9x>|jG1lw&R8AN^`y>6{1cC`K@QZqmtoC#Q01qUp*w+bJn3 -7abUPJD+*PvY4q_4Ra)M#^VQMJ}nFdT%3;}8A&-MiEisRS0^c5=yEZf=F|0V^ik-T`nD;**_XdzB8Gs -!?a6$BX5H8_dKk7glHD$<=uLwGBMvIg0!Ozy@hgays0DAfI#g{p|Ad)pvu9axV?!Y^GROr7YPpN<2+> -Avv5@@0g~>2&l>>_68u0yQul<6df3f(ObefOrH|Xm@8gCskX$VGb3zg%xv0ZON?Fyk(gUFlNSeboeLa -LV;w2Ncw#BW3du^m9!kbh=*MN)QWTroaNQKVbea&V`}av4iO5U1lFEFvb1_{9pvbEE$@Y+vTnBf!Tjg -7z*oVaoZI0k6*?8AGni2l>a+y;Xbce2q^5^2c)Pg^ac?y{CS6ESePz!;~o=o;0IvKUFykU}wo?b0#?y -b67#AjQQXyX(z1!dNF)n_jCxZxE8Ngd3qFu}>3W9XJVT;XmfxUvmQj$SbN5#P6dU~3V-yv!4qZbcNUl -8EdKoD5&Hs{)G!OO%RfSC$i22117ZzE@p(qp&U7x}(KhzOz&KTc;X+uv4(-vQf1PG8b$wS$fK%smloi -xVMM(?)Sv;hT%nl8{B9LXZo3OV8&b1net85`za#*l9#lMswdE=dV|H@3+Hw1i)U5&QUFgpJFLW`@|-c -W5VPM|)ph3+CM4Of-3mw)o>dxf3%- -@^2$x-bTmE$eyMM67+-c~n4l+0vd4MA;D*bh0(m2eYYAHjH|#2qax{!B$FD)yiJt$-j6t>D$0{jBuBHp8G -RWtYVZ;YL`>W>JI-ZpY~DEe(!#nrdvWoT|^;F-IG`dQqcMZt1OC@+SehDRA!SeQi?T6PI>%%?*-UJ@naL#>vuar;kP?22Xou3k^n9|NXCH`=!OiNwtu#TW%d9@RPMO*Zjn -VTs)E`PA2`96FRr$fFW9(?QTEuov*G=3Fb`vYhDdo#bWwLhBmoj){5!8C;)CmV1ACI}eCAn5S4BB3L) -7e!$T$A791!H}aZnST8E;`?FDx6_@S1AP%Sh&l++vBQNQdvhuF^Jw1N#|`@SiO2nPLqv^(-g=N;~coa4}bKp&OKCw!8~j_&6EW&1aw&_Vi;+wpyee-(UaKgf?efAX;rq`! -Kl@S!7yZig0lXO=tZ+&>DV|5P7heDopD-~6G$mV?XdT&}scSIHjH5<&ZW13~Ra_dU+x_(*4xZ|-{+7p -h&K_a1gb!?(uHQ?Z_JC7Xdhio;hXidw*m(DHMl<(i -?jWnBZ}@h_-6gg^R(i@wjlG-oL=!N4D24bA=q1XzZQ<8>nMJ^NHFT8jq2pr%{QOBtc;|)f99IcB=ad( -sevgscD8xAjo5auI*Cm2r(Yl^F|f@-u%6~r&d9rc(eFgW?^RqXGv>8L3G85aJhncuMRUw<_1yA}kD?wE(eI~XE664DS3As9_SA5=U{^N-^Y -*iTPGAn1XH^!=#(Na#2sb)%2NRD1EaL(E=a?f|r-7fv5FCh>DCnT$L-1beB9e6(@YC(-vPFR^2gA`gr -rKXK(0ah%OR0?f0c;X(XSol1Xjl=wuT#Q5QpdQA9~syk|<>G2vNK1i1w&Z#*1!CK@fOvz8~N%C>(Kb& -Z8hg$BClh7Zr-F{nt?w0>J>Qv-6w|qYiq6oiq?+MaRMV4L{VdrB%wfob4Y(e~OL_88a!aqX9kKp4ch^ -TML+o-<~5jY_6K*Z0W|Mw9QI3V#aBjQ&TE$QyxF}qiv6Rl3{2F|sy%q<+nugc6dHGTV`Y3abMnUof(I -IbE56^c;XD(i)&z~?eHFB2!lrX^K_Sp&X*#ZkM2Z%l^<3Ec4?f=u!T419qT!`Ule59#SRA$qTr`S~bB -xWXalE}_h3oUCahPK|pvA!{pBN@>%nOCO``Z52>dpu}(Q=}miXO{yKT+WiB*xp6YF{4k{=s9}bshU)q ->88JgJ?4C<8JZ)tJaXSzLCNO*z&v#49pSUxpmBF@2-@z9wL^S8*%OZkJ59SQw3U! -NI$_8F$AH>lKI-8TAC{;x}VpZVEMXC#eIP_UTf0PDV(o#T&1`Z?}j3_BX1B_f;zI)Gd_MHPi1#{>&F# -N-gKvF-96mB!s-32{>&-rBOnSku`ALBoe>FGS4&rZfWo9}yV1h<<}zg-qWfbmhBi> -3>EJVK`AQ#`z1-|+B#{dyO(rG9w^KK$6CUa(|3x^>GrI7`BezE7@=7(HSp}zcfr{Nwv9kPlX8P~V?*B -m7@q=?NK#eWKEpU0(=Fz_3szhGj<90K+4*FgMVrj~9p^R{0Ns{+WCqKdCO-F$3LKr_Qbo2ip0x`rCAI -m|+T@(dmRK?l*$r+p^>v&8Z=O7lAwa8QDRMiqKT!SDGUzIQss-1NDWs7(ttwFG@2+1r(0O#N#r7t;eS -_ri==JQ4<)O$^y>50<0E`VHxgVd**@6aWAK2ms3fSzDd_1J{uQ0043g000{R003}la4% -nWWo~3|axY_La&&2CX)j}Ma%C=XdF@qAYve`{z3W#L+QVvKELd_gc)^&3>?JJ0kU$8cN2<}Z)6*@wOJ -ggC{Pt9-C5@z=4fdR}u)++os{7T)tE!i#X;Pvmz1MluCI3uMk~|)CfTOEggW!=`Zv)T)LEAve7)9DyF -s_U*ZV#mrq`bi+I1icvl+F@b3VD*UAAM|`CkR~_UlrktrX;Cw5As$-h@zyIYy&qR;TLD|HUWN$hx><5 -zdjZZkH6m4gy8(Cy)HQeuy0*^aAFbu!ZWNLKGf5J?7w`EAec;c9wtEL3?;!hW3m-mgZj;{ -A-n%2j0FB<|9)oLri2=r-L$TL}J<4xL*3rRTAIjqMm;1+`$M3yC5`I3Q$CzdyAG4I_9rTswzhHEIAkz -V%_j)HjcP+s)!n3jjgm85Om6C`oX(6nV{0>|GwzE8VoafzGV`MCL8z3f;#5n+=>6e)047)b=x>47aHjP%jdo}0Xo-mfmU!wWmVm03Z+41+eWHF -;v_a)=*v+#3{vHZia4CJf<(AR(P1kW#^F_Q9$#qil4PP*t^_Zfjvw4#m?i5GTfkS6|XW -$%63!8(@FmSBxL3A*{;y~6y5RYVrxp8))835N}5=4!%O)TH%H7|4`DvFfmT>U_G5#|fb;G1Vv;z;I4z -BhP^Q@7jcf8H*b4eL1ZBAB2T6qc`QS{&-NH(AkD@`3w&!8naAvxvqwVmi7BAOB>zf#R+ZYd_G7+y2+oMw$c|_dS3JCW -^)mCSa-bDvenv_<`D{(_o58&pwybV^Vv!YHpz5)AbT;i(?T89=E6=L?p=`~N2RJl7|zg^PWtmMRiV^; -B}vc(e5Gs%w6MV}7jZNveJW$O>q -#C~wq|3rKTCd4qve -&<&l@zhEgtExRs(=?NqVl5<*W6F38i9%Vdt7hr9Y7 -9v>dRd>SY1^x&cNdhGN7(qakUoSEpy!;j;sGX0{qr|+{_p{J8^&e#ErgHBKS@B+uF8oX{IXeX_~worG -~bkq@e3)S!fqn;cd*n_XkV1^@9P@@dm^c})D+xkm2Xj*B!clPW7-TBGXgc0FV-JcuM@H+0z$=n#|JY?et^z6_9>5SlH8`ILa4z|Ygp*OCJpf -Ln}ID2e5YKD*_^6}YD;B?;;fzlf;Lo6c&>TeMR-b-gsADv)C?RgrIQiq$yVq}5wx2`>PXuY+4Z -!cu6wnJmck^4HEDI8S?`$q>a*QXz>N2yl{^Z?`AtKLSgt5Ewv-vUEkjJJAcXCiv%@!<_-KiVnq{cm%D -{A&(nzEV3TIfxPpnLmq!CopFv*7FE4h|5Zp{mXJ#jn<5n)dm@X|Z6dN -KSmIrl~gky8bQ5jpN{7QwU;*^O^1PR-)CmybHM3G_brf#?9p@ -n=5Kd};dMmFL+QxeMvT}t)ubSFw40h6MnKvptGc#0D*@&JN&8NZ@X)bQa&aE0n^$=|djoTZ3W(((J{# -q2LoO9KQH000080LuVbTY4cwo_YWP0Qvv`04V?f0B~t=FJE?LZe(wAFJxtKWo~3;Zew|0XL4_KaC0wT -UukY>bYEXCaCvo&K?=e!5JmSn#gLUk@djO~h+8i(lyro^WG0NM(A!(v6gTnj{(pSl(=@>eW6Dg4qzom -hz*5x;92_h;&w*pOP$n5k!hAJT>sIs|Z>CeeOWImCe(5&G)$O`+e8zODR2cj= -C+)_=VpP)h>@6aWAK2ms3fSzApcg9o4j006B9001cf003}la4%nWWo~3|axY|MaAj^}Wo~16UuSY}b# -QYpUteuuX>MO%E^v93m0xeuFc84s^C?b#N=FlE>;+8Vs|X@%cugexT3`;N -;*1g{n_Emm)`HC%%4AxR%L80c(S7h}dTsg)zobuJFOoqxq{~9FPhGyUn=YhjJxC7d6}r -;OHYBf@1++RJMGor_y#v!xX?CfuV9;2C#%CopW)&*gNRL&{X5@<&tezyEWm(&|);wbAvrvJFj2cCB$`@>}k%u}?G!AsfDnl)Ec -t08lWnonG}K}w`!A>J*r}3k>Qakl=by`z*1Qfb}S~_wKKY`4pvFV?s--00vX}HbzCEy|6B{2KI79bqd?PQR5N@O@ -rd#;SA(z=aS4@`Vy;#NEdjD3YS=u>twD%7mM|3n+KN(Q8mr1=uLIO`z)fjtE_8FloX)jMn2_|f8FPUd -!Ld=J8~!Q%3%2yC`m{;^Q5Vg#kpJ~U_n?7L?9qoI8nM=d=Yz!ENIfQwl&Vl&3w8kXVox*$&(n+Fv!Mq -pHKvXW=jAI(tJpDA8qGZl?n!8%`OZ%*6b#6!%{&;B!2-=O9KQH000080 -LuVbThD+9Wo>0{bYXO9 -Z*DGddF@$SZ`(E$e)q57yoXwwqHPZYh5#8>ti@1ZD~c`HOJLZHMAvL(Q6r_KS(pF5b9fUe(oWNL*vkw -gwnd&RzjLE^QXI%;vm1v|vdxC@Loeh&XxE8>4qS9_)fta@6%rLSSFlI)L>OLk!0pd8(=)3e25(XbuabV43z%l~AXjvm+w|D6*x$by`GkM -HLT0x!wHj1Jx^HD0ge^6G3;t-oBBWO)Yvkf^9yam(bYNEyGDV(c}0wbI-g)`PaJtMAO3I(HUbfy5HUl -E0Ei>vg%%T7A%7qqmS8oIzV7{hTzs9?YheEB$G`CKu%Wz25CrZmeErNw4jMB$FW}xuNaI2=7v3Dn77- -S`|hE$Jg=H;S*v8V_n7=Fb)NC<(Scn?tiLlC!>|Gd!u$ulqNC*1#{tJ|?*Mo(_O7*0CafSV!v@va74Ct -_p6WR7*}a=Fy+PdnoPP_vgzL0Cf&kYSIe4km99u_Kh!z?W5?)4tZF*lbpcwY!pF&R}YAR%+c!5 -`Io~*CF{*bZj~VWLC|10gh%*;_YrzW+{>hTcYoa5M{HSmRpQA$*2OiTo3@7UcPr^faL)NF8zMiDt4SY -l6H@7unbH0K|hPPP0ZEUW4^5X8AO7X1_~)=Ejx0%=Pd($lxT18)Rz%Cs5f)>J>cAEALvkc-T@x!iL_4C_7ZnijlU(|Qk@n#AElN*BY@AasIl36-)T0-!84yPisgR)$rr0wG -K?Lg7zz?Hw6z1?k6YblZUqd(kMZCVSVnkfK~(Xusch`moh8%IlA)TSg^ -}OomF;;sARgXP7`Ab?1rdg{EWcq-JW^z}aX{sqlb2Qo^I&MxrT0v*Pkt=T8`Qc?(}ZYYAZNoi|bxDjgg4 -wWw&WTE7am}!U<4saFB1lJJ;H3B{j)nNhh10;1V@_t=IuTCZMN)c=O-VwOpHiw9omM_LY0VO3_pIhS4 -rRxxqxWTxIn8ETBb@{vk-cIwya$53<+_nX)xDxU$G=scmp<`Y9oGtvbCVw8YerVQj>qW{}A6s{%o5`JA06iR1=ZE*iju1~Y=bae4#?7 -RL$E}}-@-m&O$D!qUnA?yC@O4s0k1}Kw>d9IY9p~o4?m(Rc0%Ud=`KqGGHrKJE -aL4s*I;v&@eO8K3`xDKddu$Aswl{37i6`xJ&l=LmSu)`LQHh`Vy(+eu^K3Gk_mJ)U5|VsjSGno|u$HaY2xe5|L)-}F4b)3|g&LH4Ze?X7?w0m1< -SYqaSekGC0@540K8oGvlo(o@~;dA%pMFqG>V3i1YWbeEKRvXW;zT8Hhd@vLT9Wlc>BNIdc!FCfvrq=4 -9&NOi&8v -{w?R65Jw$Yu$Q*CJTjpespgMwUft+etgFB953f@|u0|7;ltW7G^zJN)?{r2J$xIChj03wUcC&%fDDL$ -4)YcA-qsJl8#@B;RRk`Af#Dy;;0@-)`ae$)R;i>i92KasY{S~CXfAcy2YO5ibXgd+U)+3mw@D}l6>Jv -upJMQzP}-gFMS^xCpGbWp44}uane72lN#K;)B;Wpf3dW6;K;qHxmKLAAN~xw-TG_SY7mEZyA8#LvXVl -xQbjV-{_}D6Q;xFJ`1Yf*1P18}l}?!*7Yiuv1K1LHHPTYt;2il$IV6@L -RzO9KQH000080LuVbTQ8A0QPlzf0ImiA05|{u0B~t=FJE?LZe(wAFJxtKWo~3;Zew|0XL4_KaC0wTY- -w(EUu0!)Wo~3;Zew|0XL4_KaC0tjd6iXNZ__Xoeb2AB@>2q>Md7VdkPu80(u5`$yiAkjB{!{E5*yoD7 -nS~ZuJh5P-H;$JNqs+j&bh~KVRZ#U6pe2z3IVDz+Io<6t-bV0*Di|2HM=p=IvjtymSRPy=Yp0>tg!BD -#x9@GV4Y)~wU<^m<|{f~KB66qysk=->B>kyR)nQe-A7TBV*!;^b^K}(c79w#shq!c-rm7)_@Zlsw0kA -6eBXVnBcMkWV-rwrwjLg1Rqf;B&+xGCPHqs;T5Z>m9|12n0kMORSD(S+uSQvPU1J5>!gcGnMDrbAtzN -UmNy9>wn36Qy`JB2=!Q}X5rE&=0TIaF4cv3!on?zl(lh+cjQEL9dZkizLbr+eS=jbx4Ob|OgBXmU^XU -fB6PoyB&kFr#`@+}8;T641{t#+y`0Z9#8woq!BlM8alwwOJDq>wJ4)s21Lo3k^L$&3w`c`oEYV}i<-# -Y{N1T$Eun&(rkiac{o4VZmEMt9(qO~zL*#a~z8eO+Q+^AYWDBK^A7P4sZ~@HTvjij!$;`| -;CJ}_e(L78gE=&(^GOOs#Dy-ZEyhjl>=wtTG(`*6z!vL)Fy_=t{6mIXj5J@$GvBS}&@Lvvf{34tlDL! -RrJj6d_+2BRkasfB3=|4N-vqRcL;?jh6al!VUGGk1*5Vfr6jykMij|VoiCdH@PVDfJBkQt^t;^xoLtx -rzC3E6YfeE5iut&@}7k=~Wo4~-f4{S(*!=RDqV!{E-YHr6i#@zhQ2(P9p}EqRPE%1YJ2`;!tdD8Z5w0 -V~%Xknw)QFqj_0UAB5@E(1FK@KVJaViggkkt?JrFd@R_F9hqGbpHfN^aoH&0|XQR000O8 -%K%wh)d5NWBm)2dNe2J`EdT%jaA|NaUv_0~WN&gWWMyz=Ze(R{V|ia^a&L8Tb1z?TX>eb6a$#_AWpXZ -Xd6iY)Yuhjoe$QWV=*jLHL$|j;2BQsQP`c7}gHei5?Q^2lwlqm@8@m4QyOZtM$$qf;AyTB%-FM%2Cza -0BAZ6Kx)}fT3YinEp)ifrkpiSdvvtiAxt#TfR|GvGm&gAI*Ze>f>C40;yEmc@!(|H;Eo}tC2z$V|mcc -!&h=uP#2Za5T7UCG?kR)ygNFID<(Hk%b#g4E4}xkqWWO$xMr0hRXQ#s_zs!u%~TF;nm`2?;-9Kv>4%0DvwB?4bQ*oj>bz -6O!Fu{={dm)(GgFaSJGRJk8fkl!w#v58&{7s~!62Sc7qJxeXg)_f2+k@=6@2Bt -f_6VQ0K486i(jj?G2}gXav!lEQ&9irI-LjgOafi6~xzm=VU5XDKYZGxHs_Qe6PTXaxD5UEBjLD0$)#d -8qY;|$^RldLa`1z99Z^+Cq1*HTVtn5-;Z*am1^qd3(y<(Y|jj8=T+`OC+CFXOxr8@bBd6+X}lRXk%wN -+S(okhmHPb@}k`zDe5jQEHSBF^Bpx>*xk2xOkO&JijoNX_jvk>BSL2hPMNTz_0x-ROauw -5vd@NJti&cWzD=x4n>$I&8oYZ+z5eJD8FYQo*}Bk%ID2+DUN^Mg*a+;H)yLNJ`tm8-}{L%x9bv|uhI)8H;w_JBA6u;dh9!I3TB^T-+E|%oCSi3avDMuS1`J$>8SB&QvuD7c!Jc4cm4Z*nhWWpHI~WMyt+d0%I8Z*_2UFJE+TZERm-Z*6d4bS`jtWsc1X -!Y~ko?|q7;UKCm%px{x#gM9&!5SwiYB-yY@ihX?v9xKqDBgUtrl2Kt>KyoRdDi{AXlh^78sbsLGc%$UhGGPLFmYt -OqID&?H;g48+;z$j#>=9MyVgtyX^ar#>oH$3d2!0VP)h>@6aWAK2ms3fSz7=A000620000000 -1Tc003}la4%nWWo~3|axY|MaAj^}Wo~16UuSY}b#QYpaCt6td2nT90{~D<0|XQR000O8%K%whtq;_lt -PlVI!#Dr{9{>OVaA|NaUv_0~WN&gWWNCABY-wUIOi4pUPE$oLba-^#TU&GD$dZ1qUs1tJ+i_&t#&*wm -PHb$bLADjg23Eq|r}s@j+32*8I4WU=`Str{mTt&kpl4?$c6q?oC9hSL`DIpNx8=R-j|Sd1Ik=Uhdq?_ -%+tF9sbHsn_{0K2zXR{=ht3b<1s)9^SWf;qBrlbneXe}>pZqBtFujSn`oF?)$h^3cky`0M+lYSOXW?2 -}&h{pgV^B@cLWG1^R&QiQ_(*5dMQY$5AS+@A)Q*%S|_PYo^ -gpMFU#OX$w?4JP|p6`ZOdGl6cWQ&OJl3PhFV&^suZ&Dt=Sy74oV4 -~JZGyQRmP;pU@bf6=jVl4@HwlV8!}12bqY>VvY2B8Wu}6;oF}gkkR{@?8Z1=Yp1@C8LMG0c*xb%}JOp -FSWEdg{989ugtkPT?se`!^YdFKIi7iP=NwOEdgc+>q(tKodPU2)G|2tVIQFHSf_x*+!NWW7f9j0E?B} -qT`hPX|Z@ze~}${l!%k~Y>_6sRo-|B^9I&KEWG2s8Vpu2XsppWUY!Z`0??cSO)EH%9taN5cEFKvr85UV&#^3u!>c2L6U -{s^5HeV=ToWtpYr75)M#|^q9=$JC)WO#D|a;& -Vtwm?Zw{;e*0-RN43uD5V@V*w4Hslmkrv^!L*%*y_FV+&~ED7IYQ@7L+4Ii=gaqL>y&S3J7+nA9kdO5 -D_)1tZrXdfVedZL%LeT+I$t$tFDu&e(JtES2JKZvd;LCbUFsWO?U8%c@bgujcAwmapRekE-b;Jgpso4 -&(>~g(2JPj+v~@4Pv3DPB!&ldJ+Mo8(Zu;uF?yH~o(Ox!aI|tKlCY6TH`)C{XemaVF)80>YdvEs9UN& -euN6=n1XfKbTy>8G}Yw+UyJ=)7g20q4DmyJAF=e1Ag%SQUG^E!aGA-B3ZgtoJjcAu{{^Y5XwK=w%nM5 -^>JLIAEcP^X+GO2;2Epy+_APn2#e^?|A$fLl`w0%GS-8YO~CSsTKmU<43cL)uX|0fJWA!i!P(T@`OxG -E$jbC(DYdVlDwd^Au$P&}v?$$yG?jDbxwRlVA{~axz1~3-m+)UFHZ~%F)a)Y1mUJK^zp7xC@l#{3N5Q -bv_Q`vic!m%!8|3S(Jvx4WX`zNjAJN#{hLx7{h&V9*{VYowO&Y7pJ5)M0p|9PN&jG*%qO?wsO!2wp1T -%YHhI=V>Jnunv~Xs87LX3+L_D}TvO2)1uHp;=r!_vO(1hHCG;#=QSob)vOj2|Tf+ommnG4Z->h!4!K@ -#%fsQRwGK?z~TkNfJ;Ds1rRK628zQ_Fo44e+9s8o_@iE57gvzjHQ?4cGx=}9iOKWH%xQxtqO?qI_M23 -}fKHCr)wkS9L3<;d|KeQEc5vODPa+>txz`|@_+NxR#1eBbTgNn8v@_m20~^&R=dw}#(F_k;dBEuVuvL -Kr?}4+(;w`aPSJkPmLx=|kWb$MfmBBhHTa2yp@1sF70evF|*`d35?C8jG>YR#Y!;soP~$w~Lb&)_&{M -Dw`N?JPlvNDb>?SDhKtt{G|0@CYWlS%7D8|G>t$hcw(^p7){SsG*62~mQmkqktWmS1WPf9gJ}I5-bBoVpoK~ct3JX|gs -?>vtmR@k200b}3(yAKCqm)P%`Bem+7eZ`GlNohZ8?-d1ZByE@Fy*z8k;;QS^~OZAgR?-u5DhjaKeWz- -eFz}IX`$L -M049OeO1#Iw_;M21&GDmTWz>P>~uxp>?yThivh=cx@=Q>i(9m>6I{&1{bX&NO!w1*tfx+i*c!xXk&OdG3bu9V$X#{Q~guh4Eu(CGgAmodl> -vtb)mlMFZhUA)H=KSb*-FsZvpWO$SZh7+79rdA$;B)~7L)#m<-KPiJlfx%(I6&0JlhM6B;+=vgy1vbG -hy3b3JdlRM?caLHDGsw3-Me0o89^WS?tbLM$JfTk{=IET|qOe6=%GS|X -RlHgIaG>wz-p=2JGd>l0?M;YUg%NK_B(y`4_l1!HkVZ0u#| -6zjOj6tmgGh{16=2RIicwzKmMS3Yr?%{4jnTB`jg+lo^$enrQL$8UgL~_yM@=hBa4(wJ&B6umwY;c!L -LnY&KgQz4ZHwZ;zmgEVDAFubRpw_JcFKKaT#~ztC(z9>kW&|>n0csmj)OQi__rcUv3#+t0kR;HuVfs| -O&E~nQ>lv_rs8RmYR)ByFmqsjVUk+L94Pog6V9rIidGTb;4XkwvU3er{@j+I*PJonO$!1wAxZKq?Ljp -Ff{7tk$x~E;3w|C`E}&?AXp$GDn;7&tgQ3%JcL$GKun{eje1Ue_=f^;yw7lg}zgAK?M$@4R1Wdhxw&ZQq9-y|=Av@xTW7!x0pB9Pq(_bu7A<<0f^^+k#6#1G-R=CFxPT= -yQBlbc{pY=Ak4q>(Lhf9v#kbhoiz_I4Ur#S3{}+zR*3D!eC-!+UZ&-V?{;J=wr}ViWJx-SoRq-)g60DrqtQQWHF?eASyHmH{i^N05%FkdFKzq$4&H9GiSB*b#5RPR?x?-q8)fBi;c#@=( -Ad>VQW|o5r7fuNLMU1&vQ@BgnR7l)NOm7VCmcyg4C(Cz=xfM}E0I%jOZL-I~T>No8hG?-5dbuH2!GgS -ZwgI}0I -ng3HU2~MvK~vhN_z?R-; -uUrE2(4XYuJuv-h(hJrC1SG6~8cOv|0TqOw|OElm^4IQ%yiYzy+N&PV_zQ=HMBRw_@v$8^q(cW`DWS3aS-tvQe}N)NEu1kb69>qIIfX62|nrzm6+9H2~e4(8t7Cku> -tXiNaObXptBVoYU7AQ|4H7dn0+*4L{<*<$=&eblmo5U?V{>bzL^hZGV -G`&5D}qBU;?z(OFpL^axya-C28RpuA`}p9l|!?)5T27D@$IRH1gAKWc#*U-h%7i-56+=%i_3=0S;PQi -sdb8?A;k)UK^#d_@S+VdZjFm-x|QI;GCDQ!xkOD7>-(W{+)^{^LvaC^PIsH^BPz6t2kswA&-%@8l+ag -ilggSS+8G<6H-QPs%O9MDQC?A_tCvKc)Gi%BkD(;@F=R>%`JC27%$KA-_?{q9FkjqIAqU(FVkr-x5Q} -kH$wR#tki6>r+JwGBZtk&WAOl3=G4{NteYwZG){WtC1<1C|6Jg>CE|@XZxv^@o)veX>lGx1zHDPx&Ff -%{kWh=+c_D=uOutt!jWc>2DxH_zL52b6<40iemXP0p=-P5=@)UZxK!jumZ%0>*a543g9{ -nketXaHZpbN3SSc;r1j${M-@V5(3gK-cIV0OrmH$Qlg<(3W58)&dkgJoShxk>x0}PQC>U5}3@3Nk|h{(`#sllx}$ed8G|TuDZT=2`U42}a)?G!Z#~*UB<#k`YZ%$HM>dfTy#e%w(qHc -#28>YN?hCtXAn+{-9eKF4WZ$`aaPORc*C~-5ya&WPc!Gz!24mZvYYw74jRdF%%n&^Kj^W6zFTpV6J05 -I)w3#bmyxLH-#s2_MO9KQH000080LuVbTX4hMon-+40C55U03QGV0B~t=FJE?LZe(wAFJx(RbZlv2FJ -E72ZfSI1UoLQYl~KV?!axwc?^jIH1Few;2?s;sMXZs8fT8HMS=wp4+U_>fLE-P6mSTuE=d|1*nm`?7dxRrMl%DkO -!y9+rJ2B62hb@Sq`G)Orrq$%bM`_07&6uH^pEIUI5^(G$DKFo&{-{T8eI76?;E+2HdMXTR4PtPTH9kr -Hagl}kT+-5-7jg_RWLoNH%MuG6`e-+QWtU~I7Y_pU15ir?1QY-O00;of0 -9jjIN#!pyD*yodp#T6K0001RX>c!Jc4cm4Z*nhWX>)XJX<{#9Z*6d4bS`jt?S1Wb+c=Wwe?J9Qo~$Uh -%*4)2e|Sf|lXjeG@1)~w>~#08+oPsQ%4VA))g)y{lkIc%HTL!Hlid0MKmY_KB|US`%suNLi6jDrLRF! -@Pylzu{@%Wb%Oopqj>NK=?SGEH?CkFBisN#zsYbgYS{>~O_)|aD`A+&SP8W?hrC&~}s;rKL*cFQ^x|v5uq9|c9_i07lpuL!gS7njH$TNf)!lN2y -KFx|*dHLi@{9Zim?@zNj2g(#Tr)LTY^_@qcFVkkVNNX7CI;vCtX-tos^j$*_;n}jtfURi@m^(?RFS0z -nz=lyRX7HJMnRA~`ua{ZgWCc(c+Q=`p+2pt^X4%c_sH%~vHF6ccFX($a?6S(~C)G=RALFNG(SH7^%nJ -HYFY>IBkLj}oEYGeLr~tlJDa=2;2M|9m%ZBDNrB48$P#>!Na1k}PMKn*DP`*Fa;t-mTadHg^xN%JKl}H;wK=0f6 -ewY7teaY{d-pYEeFnMOKJ7!nN0kB2AMTq^e$4$Or5KPwR4?iic<=flZe%-$hm4C_TnMqnMNEck@8!_~MI^cU12%+jFp;a_NbK)_DeMNru^5_dLI5Vg-sI(Vl(+cEpFu=g6k)!Iqr#-s@ -ZGsoX&9ID1t_{;_~#(p8~p+Q`OR>os5p&qYusnG&w!{u8I~5EpJ^-?^Er)svMS6XsPpHgFbh3alewFKcDuuw62Rw#d^- -fhSi;A8cc1tg(Xmbwngx&#EmdMQe_&wf1T!ZDb!J*X4@34g4t49jMjGe%%ZD@Lr}WV1_jQemL0U9k!j -2<-E9E!DsxANqi$y5+(LQ-tPPt-eX4n_8J#v4Wo{{oXp3)F0hM*^@NL6jcOCq&No1p!4IXNt -v|Ux}O9+&Oy*RFoJKRe90dT2Lx4c>%&ZwqDBm+to4$D^%f&FkKRomK;+`6O6tjh0R;5*H`oD5lG+f^qctF_AIM(qjo?iKYy!zPX|X=EkI2qXsEh_-B}g -@dd~n2Shh{SwCQUjzAH}a8XT70gU`PB{K=7s?#nCzjujb9MU=4Y)9N1LFLAvBG-t~i=|3-}4Q!z_w^1 -R$9O-cdT7(%&5k_7@JQLs-+@~vq3Cl9N&a?V9O&lbAJ93#W<(njpDl+f~xWXB0+FBKn -SkEWpMyhX;ha5Ku<+{TLR*t&8H{>%hD0xj+z{R+yL2%l|e^zb=r>N`Kxn*;(H_p;u9EmQGryCMlG{;! -K{EwB$f+ULUWiAdrA+F%ND3BS3O*k;stV)!}P0QBqkHRe$;Q~$UVygzAmppyUfKwXtch7pjv9e^RLg| -zI-lTon45ifrbkrG&?U5Xoa+fR>*(^)!8gVO<0JB+bq5{0AiG`1tGqTswhT%g&tA`3;>lV^5{CvMF}k -{79Q+_GQE#N&5uKqJWAei>LBc{7=L%07Wg|lZIKo{6x*z+(|k5YnW1^m@CB@UoZU(!Ws|3M?SRvgYyF -P;AVg=UNP^+zpTnz9j768`29_a+I4s|Xtho(_?~46>@K3?=X`Xc-yA#^T!l$9r<5#iKf%O1Fd;`-x-S5L -x#R0iiKu+3!S(&3kP0$Snzvwmc5Y|gAAETBAPXv4>#?c(4CAw>2aL -JJENx(>oS(N2VAhI}v_;?AxYLEs*c(yEPsXHXEaa8nyLX&>3?LZ1RW`nF<>zxGyR(TY>o5pPH@J+_@k -r-I;I)4K+(KYb_f#NsJyh2n+dO@Ta>Gw@gj!)iPOwV3loSwaU_7VZn4{uKX{q5&qFVl$Axk0#%&?PZHGPWLQCNGZZ3`ygEmk2v9)P$^nF=uqbCeN4mjdx3GKl_T*8_Qy=@FOvwud?;zv)m|0u4jza -H+Zzs*tr5oDlRjliux@U8ZaZ#&S7`q9Rf$bW;ROzUDO-6ge%!FoTc2jbB-5P~+gL(9lUW*#j9bd$$=s --w}U<35>F1lIgNn&YGD!)K=OgkWv0E7FHEEJt85@kA?wr~$W|guPw`_;ER??T3`i*s@4Kb-K9$ -Ym?@j!Yk}sFTj;SV~hD0a5mD9%M7T_K}m?&248=ECT=jb1nW3f9Y|M@x|%QZCIhTkY1CwjWU!|k0Sjn -r95Br=1uv_aP6KxN_xA1{qUxr0_eJ1gIGrY03{Au3A_4jbvSo*0yp5+Q#*xm4?Ah&NGb4s&sB#bUsZ2 -Z|?gToB=iqtF!2_C_9yXQ0zQ2YpT?9goM6`1ChQ-xdmvmyA_U$I_aBp^6N%T)F22^Du5{s0p& -~3NaKof5cwM#HXuTo02n@Xz+Ij_GW0X%s0(FvUT)>rSVPPiNaEN|HC_mynBj7KM(3xK-enh0~R_s03@ -KwVCO;qBx)_qo$1sS3N;x!l%W-)jJAS+1!DuWQEU{?gF0~P`KnMRd?Cyr>~X-U+sI7UOtE@~vnffG}M -MC7~4D2Px6`-Q1gBAOXuKTDE~Lmp+G8S0M+?nfpdj=`DOvI3(Sp)nWVf#Y93)CMTk1XGKcWTSC`AdCz -H5A9XKkEZk}m=!D_BX5o&07yiW^j%grHMAa)An5_x{`5Pdt-542;D$Y{80Zv4I~`U^oFXZC%oIIeN_U -M#UY2(lsV^7u5s<5jP|B?ILWDul3^~lc0tvZ6m=N5e3{y67zD!a{6l`0iIjq9_REm}XG{$W!2%QZMP+ -S;3uH!0OwBnb|^n{@;hdJ|P)bSy0s_S*_@e(BIp^1+pyxngACi+=>Q>HM3E9_kzWBeJU0o!z(OZ?F*4 -tkHy08dlrgr+1!oi0WeiddNd(Q$+71kL#1AuFC94t&0Ahwh?@?ohPhYYNLZPK{*{xl@X;50WI4x)0Po6r+V#-f_T -m$Q-yNp~5lWBoLd@b_qF4dBT^htq1(6p0-iyD@IdO4eA?_>r7m=iBCfez%ZO -Ysv1I%Ee*QXybB_A`gHHZj`3IZZKgu0c@k02$YlcP$k#D||T7DiD9cymmmV$>FDJF4rcG*QGrE8b(PH -$kwASZ%WJ&r$ma37e(!lwI~K?k)yZ+K#o=SE^LXZt*qs?lwVt`KdB)LGX}L(_Rg;O&usC6u&!lM8p;# -+XweNP&JV-SIL{<0bQpfR!?kCB$re0c_B$0Ia0i10*)QJ -Uji&BCvb|>!|2SNFZ86UL226iJL^Ey8s%h0y*;jx|mapZ!O1R(0hEa@U)TTCRR$PRo>tAgcp_36FzFxgv4N2>^Ria*z;FgYi%ZhX& -hCfC3(a_)T-v|$@r!khb#dQh}5u@#HbSYp)3s;*CkJW=6qneg0hzz#H;NsB6??iu8{J{*zhA^F&9+!E -}n{5#7T*tZ&Re{rNR0Ck1!`)Gud3JZ}Jguf~ZIQO-<%m!==reW~8_Fzxd(^eFL -h*1Hv^aI3@wbHah)+I}@X^?1oGP=atMXH7C?S`ht<(kl@9R0pJ7g7J#@`X_Rx;Mnfva4CW5&731BK%f -CbMf)Y_6$&#^X2Ek2=YZ;z#O!v0P-r)$Fu{x4j0wTW5HP>Q*KVr4r|M~C#p>P7?MGPk-JS*4SW*N}KF -7jy}6*s5_gX6R3C+8;@BRA_Iq4=88{!W^NydI~LwK&pbEp1@{S){uomy^{O7UTRnu8qZrk5LBDwHuJk|v`vfO~i%Beyu(#8Ih!%v?g#b?l;650rcC!l&&Y -P_bUI)?m7R@2cmppiIVfg|>gOGSvVxZ6ebA`IM`h6Ic`g~H$?I6fIyY1JvRnpCv~8eqM`BTOV*nLTxY -Mq7q0j_ix<*Vzpf)v3)llg^moY4OSo`cTG&SuS^Tkk-<@X{V@R`c{y(0_sm03sT7}b6_?w=$lSM+dO( -*@Pu$J6)75dO_;gg8*^50IW`HI+&G$ -kE-`&t|PC%5A{z#*RJM@cl286Mg?r)i1-43*`cMG}e&1c}`XbhBCko&i04eXqdF#Ply5oNrHtq;BZ`uVa3qQM&my^%fm*T1^)uZKAud=)C60A6I>eXg++ixX1r8(3&Umg)pm?~V%<)tW6JY9BfAdz|EUIa>EDD*KrCsn&E2 -_-n1J1r}HT6f8#3D<$&6sjJv~LsiZaHHYI^5Bsv2n9m)68!H3{bSA4nz2d_c~H7n3{F0MM0jKw-aa{t -ZQ2b(LfC1UOej3kDQKwsOrz-rX5FEQ32SoOwkcfrgaCLWG)1f -^DC72WJpx4qWHCz&vEU1gHQx3jFqF -fJ3NA6A4aWN{eK67>nCYX7hjQAYRd6Zu1E#*>Y!PR)LTny2oZSL{bg{h7vzRhQv60Qug{6Qk51a@-dr -RkKn>?BXn>{95vwYhAZlal(}M3=iqH*ssIn%l6 -`5*I{3U6gvRR{Av2ua=3SCT04upm#F$gWT+IGn)?nWwfjnOHz$>?MuAg+bsdW8y}yb?)0yT(~g-%cEG -e#_z~`0w|i*ORf93?Yw^W+b^-E+|vG;f{-%^_>bT5EaTt)J;IsL55$x|K*ByDD#JC1;c{0q)|-;c8H8 -i6q#<&bV0r9+qL;qy5g{a6t4pW762Y9T>}J-v?K-Ba!K^NTEEx0=vz$)%4M(Cfx(FYvDWiLX)I;eA(s -zJUld8-4RhGbQIqx~M0XT;FqFkbEoD2skv_C5y2Qv9T>mU!`KLL3o37Sr|@&6@wfAGtw -KR3Yh_nojZ=x5?d1I)qJm5GHR1`|N3I*4;TmyBs;Dlgsh)Lspbm1We*Ro3!LeTd6Zn$M)&a#27-Z@K} --4X|IRF>#S_xD}U+m5-joSLs8G7@t{WKSCl|z;o=3z9A9_@66*ps#(nu%YrW5@S-X8g5*;x3jib}jKK -46Wf#F1Xrt&ZR^)}@2YWt%)-=A&av4BSz({4n&wq-Fn{vo!tFR1uv8->Mpu(QCF;QZLI?#+ZFrwnS6q -9K+P1Vm1XP)nWF{fn;6c+M9u)nhAvWT#tzur6(D+|({2+58j)L-8FLZY^tx=Q8Jd5?u6U95 -W&IOU^czR5W028uv%?2Uye=pnztc5=8PC++cLxLQkL|{-b6tB9kKzJaEji=_9mT|_o=zZ8<94jgA8Pq -!9Z9a;*EtddP!>i(R~u&L^(WOw>f4lJ%yq-<-auvhp$THU%iG};jl{_QQxId@th{TJyLc5P==(@wG`&6YOr85hu~`>M--bXB@ -@1=@c7GWswD*65_pwFDK$@_}dRJ0`9=X3pj4k3%p+5Oy$YN)O7~YyR7)AR(mjJ7%o=!#ZNX6|DK1_W< -r__f136~ax^5JNWnaO+>*Mx71A~q{9tQPV}sTa5YVyqr3lMCve}tf3GeC!`-J4Bm{glj -*#PW{utc1Y?0QFfA`s^pMLT5u?VU!)K!aLMe!Y9@jEjS2u3nJLEe6~q!3r=a@OCD;h*2(Kc9}Nx)qoD -VzpR>66|Y^6}SU50JBIZsp@6cutLp>xU3eK8NqWuV7F~CZmcEvN?RlH3xp^>l)MPWMG2JjF$;BoKIa(c)-L2ZWCV-=B&tuTgDCD6(R*nC -XqByS`dO6>MQEbP!7g)c#avA@eHVTmQcyJK7MOkL#Kj#};$p#C9g$|n6QW`S%0!J!8ttj-O%m*=7ZJ= -lyC5wR*I_gi=zlqubp-a+|)Y%d+sbIwXRvUyq#VpI1KNjgb=845P6^ZrqWsd9q-A4E;3xn>ned3iFB% -e#>P!_gKMoyw-z5L?IP9>QKw#)%za=zX{)iq+P?h0Aqtwq%C@~|uYasYaKaOKXhLfc-a$+}q@ZRAYVF -TZNGPyW`keR8#JObpYML9N*Tl8gi1cPMi+=y}tTKXS&Q=@yBDnG}AVFK3BG31RQotz&NSWJ7j&)ScTF -n(!mdoet^E$6K1_&K3}1W((7aa6NtM}#lahB%8Nvo`OO(4KkxpM@43(~vdT(u;)@u -4C&C#CbbAMH*iQGZYzy)6fL&$mB$iBV0-Gajyb|%24ghn`Y>mzg5H?W?wCo^Fox009MQ}V4`ni1jtKB -x3(U%?Ddlag<$Oc&&k71G;Mu_MT4Oo7V&zP+eu!>32F(KA{>D&V3(6Ba<3c((*-FWs|prQ+%tMyWM(A -homEJJe;nDJ!qKR0`8k-K2o~hUSOX$Q^~S77K2zaT#q!rdmgF{hUzZggzJaVMdL3x`739KexjI!+xw8 -TavwR?0&A$wEI1Aw!BOx4@oxhh#T~xWwDH#ek_V;DkPc?0lndy&-BQKC)ntP=@q=C2Y{DM?|~k8thQR -k37MrAj-fc6!2`O$hB8~~$nlu&f;#d|?&Ju){5#ZLCrIlKv|GxcsI|@N!-dyElkL_3JtHf3B99^I`IK -_ywV(+ou*t#F*8Qi2y7LszCyn{VM^yf#H&TyU6eYA5Y8{(UKP2^WxJBwivUMb{NK*-rLoB_J*9!{&DF -q9WHqR40O1J`QG!&2UfThG|ES#(eoj0c|3gPibVahd3ZLUTHPt?%%P*l}2p@CfMzAL`Lf?|AqQl*R39 -#Z%XtDWoRe8C$Ad^CDMg>d$XE3S>32*+8pcrlsyn4x+SUXEe6&*D={4w(BdMJDZl4gHREYgs1T%(F$D -8<#)HF|kgnb}O`fqpCIa@q>FWQR`gzyyXG0aHS$RkWq|!ek4~0ig_ -CS0G*JlQSE;@&*>C(}9Q5JL(gtw7hr7K -G~^~NmVW1i|o5vJ8>Ug2l%)yvb!8&#U&clb7>Wo5bMR;P*bVEs?+(r{;JbvP(95=Kbqjp0n6`|GW$tJ -6mQzGVR)veoF+au%%d?{X~pwOL0 -F_C#D>w!ZR}?Oi(-8jLgD3rOanl8(MC+Vi{U)XduQ_m~|Gmo+!%sybq(TlI>S!3LR8zu%*@Z|9Q6pS9K{1MT%5WV3dcZ8E_v;^T6@h -^nl{j7$kcU|06+IlV80$_Z+`Ke%zv-pQP{-COS?>%#yF2yO#QFq{2^uD@(7wf;M@b9u`ndFlAYt2CKH -yqja_HSELQul&JFdEVOU(A#B5k*;Nkx*b+3cRlo0i8QM7z!Nic-%n-~7Jb%9p5bAnzI4t2Gh6DpTHtB -3PxZL$w8}~q-gU*cn74yP&jlR=-<(thk2lK(zxStj6h3CQ4eJ^RZU7I_3ABFDbJ`Ix$ORY1N06yPZBid3AC6;`HQ=`(EP9X#X$MEByD#{u -k4$z2C@peHnL`8G~AH9G?TM7O;H(S8Wv@(k&pqgGa^-sD`iTEeCv6in_r;zk}{bJgU%3G2lv?(lr4)y -lSw>=U5*y$JYj_AP$#}JZY%kRE;_&!PnQx(e#ynOv11%12b+B#EZ(;i0?ABl(VI7RUCmp|~ycL~W!;aRX+woqpRMZ~LtIXm^Wa=)5dO^3YeiHy8 -|_nRf}8CSIOy^}PLM^D#DS2gKjOS=2_z8k$oKC3fzcnwS{!(k{)3_FmE?o6$|Acy|I4nl@W;(+G?9Tr -;VAaWxj=vOf3W35_Kvx^IymkLVgcn*n2kAuq9BW@8=^|JptmDAK(zKEqK%J{Tftxvn9HaPY5Bzq;qbK -kPW2#eaXd{}q6L906F9XLFd}zxC;1{D(7r8&>@-c(9rNThYFk?jD-E>KhRa(6+5n-?X#&S5e$lM*P2u -`yYJW;-9^Cv*g`m|6m={t+sy+W^c=;zjGdJu|x@e&^V*Gi^r8T8-e=2PC&la=TJ@LXvsd5ZyLw+gySO-c^UAp^d-&(zUtj+oZ{E+|eTi -Y`)hDB)$B*$xKK1owxHtNZd8YEavp3Isf#H4IKRkRgg8z*z1OIzFGY0l!ctroHW%m?ASUm^b2igVkSO|IT;m9r~Jh8LLKWQ3(J=GtudVFn79$2?9tc3QU+@ESQx;PDD(SNBH&iSn8J^L -Z@9G%r12G$v|ws^wT?~|$dn9SBRqrLX^O?FW)kd}Zq8ETD;{CutLE)FSe~OUqW8Gff(Nc}g9pDfRTQWARu#HVf>7zsj* -@TAp#VeA$YtZZ{{Uq9x!g?Gx=X!<4MBhs$INI`#$_J5f^Y5+sTJxwXJRO{k3|JQY(S-Myt^!OwTMiEB -eR;=spD@G -V|L|>(S+U`kVe=0V+^Q{I4h3l7xzl2aN37+I4*?Y@;*Lt*bK9f)$CN47(*^&elGa6s+mT=c;n&N{cI% -RmTg$lazDQSPY_8+ff5S(ZpO>2_fjp^6>!SOrzS_KPpAK&Lly~A54{V{Pt?$y+Q@bX#?QFg79>cP*cV -h%1CI>BgQ4O8s;~f(tY>jd{VL(?pgyLMfceD?JZ}X%#rZ`TfdW+Y}*z0ijQPBG{_fmA1q+IX$NeyA_a -7&ih{x9!O8?C$yH_qL9)QSqX9nachXmo3@TWz+|rM~MXfT8jSO~AfP -V&&JP+r@#y6nZC^8`thR -N@4mHRcpC6?AnH}_7RaKgw|6;arg3wNWdo-AV=_+JZ)*H_{!dd|Yqdogq+I<@RCMaREd3KkDMHX(RL# -g`k8>M1Fr^=9{NljT{nYOhz9uhvlo2y9!i(tu2BM?$>HXM?tG`m~-U`2QDBO9KQH000080LuVbTb;et -fUpYy07NPP03ZMW0B~t=FJE?LZe(wAFJx(RbZlv2FLGsbZ*_8GWpgfYd97M|kK4Er|KFd2jd8Gi^{Ts -<>$NbtZ7*q31W1!ZHYv~=hC*Ak&8#d5qP)A#;l6ukh7XZ?c)fQuf~+mh3}=S(_>t6@U3`1N)_6Z^N!`H|i5FS3k*?5$jjT#1y`1*><0{dBB%Mb2)DO? -}TR!Cn`KJms~7XYfT-jKh3+UCGsGnI -H&P!<;eLFTrT#Y3D#ccle@Inla`ozAElM=q45Lg-vgYfZNRzTEN>SCv8QY3Fp@+$eD=}j)@%w0mI)c} -(t+FxDy`^VSlH`03_D@Hn5pe?5yCN41K4Z471g~LT?sp<*xe%#HkuKv(u@6EOhiWZoj)?xB&PT|XiVb -*H=CV$bNQrDSW2nwG$~7eTksErpgWnZ01rBu(ia|Jh7eCCI5-w!M5M -JM3#(^lBn4iQYEV9nJ_M}v -S$p{@ci)X*)#Y3jiLn5D?!|B0}+nTw&01}N%`Bz@yttJ@x2?FvyVVMr$w57wi#b4r@x<$GujLbzp+nJDy3C*}%ftj#nvzcwS{<&*m46|MgTfQ{!P -+#g=mi-1<)eg}^t*6dF=?cEXiE_Bc}JU|PQJb%b#s=G2)^^wlgYv5^ltKphU|V3_>T$TLqOJyA#vHsnR^NTovQhG -81&i65Zb^IHJ~%2Az?Cmx`zDm!ymp3b>z9ccbQ*G)b-dBi1rE7jZj5K&jwnB7;dVHo0F-ri?4L>CTFuO-wU8yOKD?~e7=H#T;fJ# -IYGcUjx07t?`b7Y6v!ad$Qhb1ZtA0P3Auwai#sqbF$tN3JX_t#;(+Gx8h -`81L1DSLHv*f|1A)wHxgRKpChWWzp%Um@qgDHR3kW&eaUm1T!M5$&tXT0(XOdq&UH9r%Pm7L1+|x-CU -EXbHdQ3q|w6<~sc6+psNLBKyI8$gt+68tt==##E6_kVk&Irgsm2Kdi>x>qCM|quOVOg@VWob;kA%umI -f*8FPP5W0(9*#_W07YD8yatwi3{{}26_(p*@`c~=iC@@mWm$ru;-QAE4U}mGUWz7hd0bDRAj98+{*u- -c9YtY_IY)wrHe2;}U!;euQ)%^6kfXnZgIU3K83zCI8sM-dg1M4Jk0lArOej9&Y$0sh-h1RwLZXxjg6l -?ko_;x~x=GazNH$?$r+^BO5wUKtjUcQ;z8JffNvrXU0cyi0f{p)pIj5Zs7BT7+6>KhXu2^|U5R_fWE+}b5b?g%u2<;Z~J_FA{xgCj*3(_pH@ma9 -pXrW1eHNDn;Y9%eWZ59#=O0}A_Vxwu8y}|;V+!56Sm(A=OtUhVfn#?uQlZu;o1_q)JCT12YbV7?}j5h3d*I0R>ZBVKt6FL%u*c{jjh&lJ!884q$;W75W?*P+;OhHmbf$QDH;2V(Z6A|8+yk=+q2$yf}kK`*_>U)F0}il8-37GG|?~> -eqdjb_?yHby8hBOAPHKi?rw`pcWnHB+4ZXZtYLxDAQ-H$VIJIWAln)KwjWwxf(U6uyu;ETH&;gliE;4 -xzp`v0=i`#+R?NZS4-@j&?x{YcT>sJ5CTV2g#r<;)2&gg|3r;L`%VVy>UaWrHM|SC^tYZ}L4vM~*c&8% -RaHe#QE_7FG^q_*^=QbY?~vkDtPa~~LMz8M9NeAG**6vZn`oP6L4=xP;S+A}lY3uOw@;IM0#lx|Lolc -2rz=MLRGxyw77e427n?TXy)*qM!BZdX7>$q%4=7LK#xpUrEaDfaz?=6vEE|m+&yx^tuXlzsH-K`KwF~ -4}E}k!+41=FI^-~gEXZLkGy-VFOXyJqO^U}5U&IX@S3N+S}>@Os~p__KfmD^tqmW6KOuOK*X!yeHtE{ -ihsS7TSqQz+Nw^gz30#;&G|%jJ||o1`NQx!;Qv&&D;hs6==}A<>;zk0ox1?*9)R;x{bzQx>mN!=GBeA -F15~b-R5*lW05!$+Lxx80~(3CSp9;iu65v>!@@_;|)ac@XWjm3rD|Yi&d!1&+Pk#(wiuilCcUX}?=a9aeGBj3z=y>4uD*M*?!q33U;G96`YcRerjJiyPn8*{0^&OGzQJI7L7P3uDJi{KyI?;{0lYWmooYqym-UELe={fFqf -V=no0uJS|^WIW4`H$WSp*dT65hGYiH|fdEs`*T)#RM@hX|(%QHyD1ndJAR4o38Tw5D%OYFh9vDpK`vU -GaY@AAH+9+UiJ;^tf>nof7Z_Sw?EP}KEm#xCAIoRdf4xei3=KfLoNOn{rrfw)7w#rAxr4)#>?RZlunx -Km?vg3VtaiTZf-){mA!e-j}Au6Ep+euO1~wuEw@#-dn7-G;LS4F>E}ls{JWyKRrIo*Vzx)>cZ!<(Wb^ -hP!D>t6wR#t+bfJj3W{HyK0=Xe~Z%CXOoh~xx-GUITb!#}bH+1bw9&<20X&0EdNi2H3ko)g%wBj|0)G -z}72T)4`1QY-O00;of09jih+_>ik7XSdXNB{sH0001RX>c!Jc4cm4Z*nhWX>)XJX<{#PV{&P5baO6nd -96HccH20T|8)vP9zRlvM8!!@Hsft|kCVvT+Hq_@+v(11lGURmNJ31J91^tb?8i}d@3DXPBzu#sDgY!v -QjU{pb-!+j1PX=vu0p|kZ18x%=E*|Dt3y`E<=`bg>Fjm(*y|*{%fxCeng6=So;^KyK6v)@*>m=*h{HQ -}8GaUN0K-nioX0s|Fqts9=Is1Vu9KKuCQG>uGtQ2aB3^`2z%%&5GZw;nvCPD*kO}-+Wt?w#ED_cv=j> -$idVG304&<$5VZ30k&rYu{Ca>OIon2hwLm1atX32($go#a>WRktQlRTfCb@b0DS*>6{=4XSZubHfbnX1pxN66sta4aLMO#%40>;K2AIaIC@^j^Q6e;J -U3I90*RvE7E(m!g^#n&Ql5oCraoJ70D6mW0JhBT`bfna4nF~pwRz9yg$!pASHIIJl*=UB^qGjKH$1-S -bvj5KV8X~yDS&||^vs`5<8Z^LQy4XUGd@0gcXBm-dvrQE9$#L;1kdw2KmL5Xi5M)K13yRIgWzeGnXQh -x@2-vqFS|fo;9D3aF&}mBc;5Z_r_PUIp7YHtx??au&PUxMix2bpns35xMarvR)~z`HV@cffOvsO)h%E-S>*9QQryF2gA2xFz=Ea0l3-z4Wle9+)*WB>o9_9)PZh)CZy1 -W3ZX|H4{stVgepVUtBV&-R_Kgj1mF!?b^@3#N`F#x4RSERgtYEoGV?4u>mXOEE*L57WCe9Ht^2iaZAa86G@)@!|ze|832|Sf^;U;}wGueYWMWL)<*!l0*x9 -3}%IajhS;c<6vjNR>q+FQG~N#7(j~o)~IK|CQAeZwhQr68w*R^ABRiHv&S$mlzOSY7AZ22dteYim17R -mw*vM*Ba$o&bUk5CL(_169j-X;JWjyC#c1imn@~hVCIGCkiW05Q+cih5CEJFE)c_4;AOHu3*hWdXV4! -Cih$U`I#DM#TZxTR5!~iQ?+=TI*2OX5GBoBy2C|I%d82C5ee)buV@yM&e)2ThQr^IiMO^_Gt`N1JuB( -W#iI?NUz+JJq`!dad~g{(HJs05BwPe$jHZBY9J5)sV5xZ%bzgR%o%rI}1q-b^m9PA0FW=SQ#qa`aEAP -)0d|EC;(MeJ=p9J?MeJc4)?L7f0vkXce}lE -77$Va40ycA(T@0CR3km)KRTVxf-k}E{H;hNx4fQTebw{?P#*jP7*wwg%Y{&;t~txWHz2PG99XhQ8GcK -_VE#PJIjSCbQENVKkl%)z6b228JR8nLJcJ>vJ6woN*}xnK4%&rpK9h0lS^!XEIEWZP?m7S-I6nbQv#k -K!w(BI~HVrHM!GwpIRFw3bRgB!34?OELi=x0~8;p_&MiSx7tLSV`MnR7gX+LvN>DHGg6zs*l{hXnv_S -io2>&R6oaBy~qb%b!`?_b!gT0Nx+8HY&1mZlC`>c4QFmWGkkI?OQaWj+c>oNRl2ZKE=>PGE^59)hQY7 -ztyQ&k&%nA}0eQA9bcp(MCoz9>Uu52z)iWB#-kJggNt*8A8+h>vol}95FQzOu-z&dNfS)jE5lZ%30;g -@u}Ffe1@aIoR^!=Q$80<9Snh^u~$%YrjlO_2N&RUrdXGf>FeXT|Ln$N;`m8)T>-_xmxg`7_PoJY& -P#%&5}0e%5|D8#q)9UR;=3c+^*ar4ntR)9rZWOvi7h)2g51@#%m@ku(uU(Cvd}LL_P(#HTrwJYYk-;k -$`+bus)&~%zk8vAaXj4z)8iS>uHos+6h;Zl2%v(h5>29H$l^2$SDWwh$e365r(H9jXI_SMSO=oTBOw~LfCNvC6Qi8a0 -UO_p$p;y?{B1yR(XaLr;HyJJ3Ys?cXirdS&*i05oa4z^dDDrdHj(KX}&__?C6_6%DO~u^dL*!WkGvXZ<~)_mKf4@f8P)Ga*J-|7(u*JhcN1I^2-MH>3`ms8JKxhb#Z%m%rM!FjSBnu!zC~^n7A0Ho&%G+D^%OlZwrvYY_Sn0; -O738gO%nL`HwdbeZN7_%JFwDk-8c{g!;N(+6F4hGw3ur5++z2@DM&Dzpna*e2U}lWAq3Pr1%tE;vO!f -b$OQ$@6khw)rut~z-63QdTgU^{9NvOB1Z;N=(>xRkDm>Us*rL46eoR`3d;j2DJ~Sh0ijK8&v^nE -NHnLZIzwk}Xz{+>wA#?jNUx3@X{j!Yb%=a!7T6!{Knd-3B5a6wxMF@Zo=M!klV5L-giJkq`AIhIUIm-v3uNSBkYuZ&{yvyvl`@Qm=MgIT-G&qzp -^T8(cSr?0K`BasbfMi|cc3rT0D1^X596TroV-h5?B*x-i-JhBf914C;EGhpTz4`?Ymh_YY6I@Si_- -?}dj)5Ff4bhjwLZOf71!Ijr(H##(wb!(2s&==JPSXWMmKV*;ACauQ9mqv+bmTBXInCsjg?fdsvv0xOs -gQ{t46?P2jL!{U=`@lb-wgQoxjEEU%)hwax3J2qbKS&`p!;p(#@e@D2!6RMLlREW_*REW`zE-QGwG6z -v%hdWk%oq{9sz4808oko&Keif-^9kwmBLc^rJE)mSFfY^ZuETRWi+s()I6Sxqwqio>0S0k5=B;Rb?{-SWzl!>=9(h;gf3ET^-Py@O{z1nek+jTWmlLmV*REzw7 -fTk#kch#*cw%m$DX@HiKE8q&AmLGW~k!||LYjJ6^0gdJBE{XS?jLN|E@A|A5;`=9?}ewl)V7Pyn6tH~ -*QeRMv#Iyzw|Q_gn_JScax{kOo5ZZx>Py@mZG*Mch*G)ST;1JE<>WOaLXYXqL)QxROHmraPk>= -?;8vjJ=~Atjhdb=FP?)UtKmO8PLQu^`aaDlu%-FijPN0JO-VE=0bI1-^enqbQgBI-fE7-OpkTtzY_g?>7NouWME+msAa4nX8f#nUg -AD8Z}#pD_@43K)!{<}Hyy!1xM^&%Sh`M^SJGeg5)g&DCV_i4UGL^LB@lToipzQ149oHW&_mqVSN>JdP63o+2A@`TgaX7?1dcF1Kb1-W_)K+O_a7$ -9gZkx!f#-<}r&{^}s}t|cd|Z-vCoJoG(TKQC)dfxZHye(07jEac7r1cSIQ% -NC=dPp~N+Nt5|n368F4s$-x;d$xp>}2 -@L+an4sjV@&ChMgHJ+Z{z04pa_UkR2tAmSqBbTNnINfkh91q8R7L1})YlgClZ2l){QMeVi7Mxue%Xtc -_g+mVCRY&imR_YybWSMR)iDvhIa(GMGgTiq;iz-muWoSx1^3|S04GrRbz{3u^MrvE`|+rHKw(`T}{4r -P>fE(LiVCR5aPPe%vHjOO_!es&-A(hgjpuyqxgJj8v21IJJaB9DlJ=foz+m6*3!3H&UU76Z(T`?!D~R -iYzV;kP1}xv#+NkO2I^_>6Br`pi&O#E|NMO>GfjfZXSqb*-faaZ5^W -r>R>}YAm1!^>=+JRNG$n?u-ee639T`2C2DpHS37m;T6=jd(v%Jz5hQ-0ODEK*vD13u2WwIyE}HWC`^c -)Hz3Ci#fZ@!YQw5skf6&W)sx>E>kTSsDhImmq3L2(c%BDGoj2&CUkWB;6jhi+8B`7PFb&>&nW`(_JP9 -4dkc2%c^t?jrABU;~O15SSb4YR$nQmrXp??J=c05$NzU%z#MN{sDQ>Xd9>z3u72r)EmlIBCOXL^qevm -y@idVaEZRUHQSnk7pC0wEz!s3kFC|Pn`mgVe{7%--FtG_|%AK@J&;KH9Svuzj$CuDN4fUO^tBz#W6eU -b@h6ceIBU?JEf|MM&(&S2}&i<4kRlBR=MwPL})1S!GLI09hJZ~q+Da8S}J#SGS&;Rh%6*n+xt$E%TZR -1kM`##rxo^%9uZLL-=L@ZFeHum+&;gmfAct3~%K24M^s^w`w1De2% -ffF~=4&vt46VUE;eTh&nMyDLmZInO~#>nG7(u>XqqDuV#4A7T6~RM|5_gS`5{dok(3vF&-ZSvm)bkOt -Z-D2H3em{m6rXQ%ei+Ce9t+Hu4~D0P?VKJ})w?+$=d?%tCehY)b`FRlRkKtP0sNyx76H|xkbSx>hMCAFm3Jxx@pB!T -?2>~vS-{UGlvHUKwWQ`ap -8vC+TUQe~kSFK^`Wwjz=ZqJXQejCJ^53A29x5X+2c%A6JpbpO6*i~`yjMH+1)20y9r)NHI~%~}Q2*~L -ypsk(GfLeOQ^Zd3b}A~a<*Y6?Oey)4A42aj#C$t3Ri@ow{_(n(O5Jc2w(Y0-_=5Unu-i6SfYogzuZ%U2gIUb<#mO>= -ep*-Oy6VUDw8h@diq2*u!fyfrSeQ4c$vQ1nUJ$$e$6Gd)!yJAT?lggqYFdb?`Pl+YmNF&#aU73TEvT% -&V2eK2*Esz=DGL+;p4e#d4EZpd-1~nP-yU(t{DLlUi?tIq|hqAX7uxC?r3ueJ=I-OnBRrj#$&3JxBSZ -mzUsl>iWpzRnyMMrbD>q4Jf%Up4Cv#o`-DQjzF^ZxAqahfZLYYsbbRL2!RftIC+%2rf;E`=qYh{tsh> -jK&ZV?sSB*qxY4sbiEiSfxMh8f=1!N`KAIKXhq1a>R#VitYc62^r%Y=H-t4JBMHxnV-SNnG6YZC@Id(`+VB)-FeTAvP10rsK{)lT=wI6G;YD;iavc2BgTpx70N-L;2CUTtn(nW3& -yOgBX&MH;ng$L^AbO~@-HIfflB1oA2B?zw&a7Sv$VzgLlRo5gL{alcm@`rjfAbYIs|JuFqZcv8^VmJrzN+9uME -QiBbd#_sUExl6et=9HiY~^JICk*DHqKL1wphin|k3(xHAsI->`K`6inMne+x4+--^Zow*`SBs=vCrCT -@4fcgYp=EUUTZ3Df1I;$9A||;P2;$Ioc_;3lx5MQ#c9yiA=F{N?l?7!WeOzY&#rV*wrj_ZJzpX+HcCt -U3C*JV%vpUoVX#+m4Q^3EMK5z%JuGy(lwjRc6Wse9^1|5oRkxViXnr-|$KegiQhCT?FQ$31(R<0k$Sn -m;X=hW8EB^A6KLhQCI>$*k<<5t~(8BIFQ -EYrnsk!$MK*4+co)K+LScNhIOGfCe%mfPCX&e}$ndAI{Z3cmi{?7I3U&6r@F59=0t9KiYyOEsyJRacV -`s;|QCZ&2jBfy(H^DE2(dINuO6IOX|atHegBbj5JF6_xx`t>6pzZNv+`h4lP{P?-35BMXp)Ed6uzdqb -WxyB2L(kNouB(m&|G+Ty9d`d{J|FQE`9#9ksQl3Ngs1X88Ppr=iTSyWgP&I-P(-sqa_nyVx9PkfYt%D -m>8~cora*u_>CS1SybLK6rm|630EGxLI2NfTqcZ3zg-k6nDbeT5nP669Ab&pzr9w0V{BP(Bx(R -HJ0PNY8NZfFi(NARUE5!6-6!XL@@1{%5eSrXPyI;>fhs=4#H5e`dAT+7scJJLOC#Q`zkLj;>*ja2p>> -N>SpuC;#)Tq@gY=8?Q%q%y~2}`z^l*t0POoL)80r;80a@h^pNW>Xx#Ysy=XOO$Hf@u -=`@0UN}SiN4-YuFe+w&vY2M5ne1k%OIpzqiBZVP`LBTCF-_Uq2)ORcfIR(v+&M_-(2|s)x*uNRabj)o -17k30DYksR^iV>x2j)`t=1OYk1KVEuK@W+V>;ITIWbHqud!?`?t77ad#n-=U@&{-CdP6)$Mw}+I+!lG ->LcgC{lvGJc8pYD}g5_`2va31v=~b;ImK5i&z$a&>!GNASZNtfEzsQSC{KlBm&=0e3k1QkZuaQ -6Q0-X0EwIsFBpQ-GZdwii-3L~%Ns{5>1>hSk6Ro)Wr*$rtWk~HVpO-DH;Xp7TGw3(*Yx$5?+5@VLbb* ->PHTWC?Ni{qHGwto7|ir&c{V7)5pfX@Lz!CM61^4U<%C&E?{V`MNeG&d8Q@zke5;BQIKnTh#e(X-58& -Sw_AclB>Kyf*@3yDO%~t2Tb$+iekeTb^!eWM%9RofJ&pjBfwgJn*px~XKt}RmgiUw2*A$*Z -24G&M8-(9`}}NAHj0ej!?=O0FCdpj3OsBlzIQ3W_)ldUg5z--S~lPgS<^(jb3kaq`OBFQn#^#Q2QaZ$ -XgUkh(b*+5{T1%m96OBnrcO+35t?uoRcE_^eBNM>_F?d>+5y*YAcgiJ3_&)XYI7B^37eFx%f!nuPdVh -NtaSE?X(d(NQikI=AyD1@&IBa6yQQok7d)$dDBs%;Y?~OG+Ks6RWM*enGCPht-Tjgk;hgRSKES8a71- -Oy#T>a$6EnTqN01HBJC2uB^^117zF!lwfb`l&UM!=eDuIa!N=#rP6tU76D2=C9WepY9AS1m#5UPH^5@ -Kh8h7e$+I9?+fCrlVCpVPos7VQ1jKe@m3~DDX@JsAj -$I}@n_huwZ#eSVr}**EGZ9p&L7l_g7fmJe$@NPq(bU9%6hh!vhF@bEQC|{TPS&ewbarjr(N71)&Q6nr -PD`hSExsv7cS$0`-Ht*EpW{gIWbinALKpkXCrm#Albac&*qe?Y&fAR-P$6?|NxJl<@!>P$gUCA~wMsgC^|L#HgS@>Rf!yo`+ypJit!VrbXAhX1JuK2gU0W+Vr%G -2q7J1%Oukb`9P@avo`Z;{n@j%sD%yQHW6$}c2$Fs53-_1y>JyUAshbK=J^OFc*+TTS8?jrOys2sNG3a -sJ$WDWBPbDSD_RWfk|6r)2+U#f1KQVd#C^-4o*6lV+-8_=;@o)x;eR-Q9ev<3^}X~C&-#9Fz`!)~x9` ->-|iKCo0S@Nv8_yv7_KiN*av_=e(?-3Y1c7zaB^v;c!mrvggxeX7VSyH5a2Ji9f7_s=0wW@hzCd~vjg -w|Bz5R8 -=9?X_c*~VhFN`E-ijRUA9ohv%y^9K1q>;jIlVpLQ?f^uK -bFEJM6BS^AQ?@$$eLKmxqrVgK2Uxr62!y|e0*8L}+BNIaM`7-hLE@N6Ea78g%9A1+8XwlYgJy3m}^7r(%GUz -?hc9;eatyvSznf4(M-wP8PkiH7wM`{%C5-v><+zT~(q;)`&Y(FgUdQ{`BCK?ZvE~XS|Nf5 -C2jqu9$Hg5Ab7}`&BjnpFs!(EJD57PB~Gb;nN -sM217UZUCxDeZKiOtn(X&V!nyw!{kq4IC?!2GGuan_V=)|%dq71WMh}1rsq -TR14>7%9oHc=iuN8Uzr@`yj%Ldt3yxs61Tj_gD?Hf|Nn-ib(=BB-EbJZ@D}~yZ@oM>MB}Xj}C`+wsjx -#$@w15*wso4crf{jtWM#^AMI{_|TEweh)rcZ~DV^qkL`p&an2ja0pYL9D;S<02m*O;LYQ#?!M0kg}BA -3;+;WaV@MoU+4w>MwEIg{HUk2^xmac=nTmRA+n*7SKK*zx4^q{+M$mOI3 -MV|_Ltxh?ype|cZ+`b-$cY(0C3&e64&ghUl0+Z7zIRcZSR$&(x)q;{Rzt-6$M&omkp@qE`P$%YtmbkY -|f6wr+$L141>vJh&gZyF^=1jw!ZNgso-%`}vP~Xp0iy3M~LcVLwg_@NHt>x0&pldmw^bqh9wzLr&LOe -f~?V6HmXvdF&QoRlpHOEH>jX{K*BGU?lJ; -_O4)R6i**MHyVmrIccIVqoASO_C=GFyBj#|1R3n}0sd2ggOL8U^DNOs7GXU4s)@BBV)I1Z~wNX|K)!# ->n -(vOk+Et4%%!ii;YGVdr;6(`1?&EfBx)k-$};GgU%TW-EvEkRv>ZGqnSxayGM0NDkHbqtvWr -SWUmSCYPnIu?>z-Zg1%6DB^XjkS$PV?lpRI -;a#YkRc$C37`#%8LP`*}HrUa~!1!Wrvwm;|!g3)tfcRtD}DR)rvd1$#ZzbaJ};0N(%18cM14#ZWe`eA -YzFW2XCVlHS7W7(+jDd~J)8cHZKY*Z;9WWNnm1jn)7-V%52y325d#pVry5&873x`ovvHV*=<-!1&MKD -W2bX2WHG`Y!cTkyp#|6?b<$qt)GwraY+8O@D%d3?Ok4%(6hvQen&Aak}VpcGQ{GgtNY1_`x3_!LMD9m -addw|GkXkuH0lb%`?>}iarI&er0{XTHh7JMGj|GcikOweK#jg*^e`eoV2mUa4Dcj%TP*e25ku -ijSa)$Jw#otVW&g{`D(?=m{VDPmsKoU=4VZR^!7c2v?chiK31x-Tm=GA6sz~uH`}Mtul0vI2Afab0@Z -#pm{fIpmeyqLE#NuK@)Xu>{Xn6>TQ~ceS+P?v-_42a`xZ(NNdoNxY!R7YN@S2Z7l-P-7DE+O?rwwehQ -Sh$5DxkhWo)XnR8XCG>mQ9$wX*i@x5Eu?bBUZ;7mq>8l8s -CYuD=l$%pb6CTw{+`7)m-Kxicrh;;jVIRiJO-|gKvk)-lV^-1(nJTn&hdg!DQTPT@asys%_8|cJm(G)Y+(O#g@x(?FauVY?n -8;4-w&@z}JpR)%_19q9zPYydF;tRS-5;{SNo$#uJPvG^*4!mxEAFsQQ;`P8=Xvc4RGC{ZSDY|WMqg&`O-JXlk?KiK&? -WaGaWjEDr4Odu|Fk1huRGtZFri2)NyQyCZpTvg>n-a!)sTFoep}L&v#}&0LxMT;)nU>c?>u^|Yn6(>p -a@s*>vl)$k{dQbaujsFF99hYpKAjTCK5aIVP0M>6*A68fEs|~UNW`Q?cUAKRO$WzXxri@lW)^9l?9{k5z^?_vYj_S7KW(nJro9gHHG+b_IkAh-lqs6`o9>n -$}l_%l7d=T(g1Q>gaL%n{aA64XQguU}1TP;ZZQEF?g^U%Bb;7w5vnC+ -wXAh9>6B6ZT$*nOgw%5Wt#^%-`lD%#4|&_v5s#XM%TV0!)JL{-WmIjtP;=-TgJysd>-lg;@xcCTAZ=h -M$Zcw11GGAQ|ojp7!}!CiZt1$g4z2jkVsQ)L4TiFPNj1B0atrnLb?%Dc{^Ywc=R!?&{})505_AJ(0PJSgC$32HQk8y-! -^Hhj(gJ}Ozy*bm?W0?H&by-cNBfol9^%e!OGkzJQir9@G)V9Z;n)M6<0p`aOOgaq -5OLf;5;_e%xpVo*SMOItj%VNRB0(cHWXblEe#i*0yPA@0H)Gz~VcFy^R-($IFFLskQbmi2b^7F$m+3sP~8#b#XowjHOydLzG=NlG3ZXIQa#JjpvA`uO7e<-q{(wW8qMb<7O5e=3+5NR$+XjK3T$#ZiwX|`O7_KFF07HP7~IMH4% -^5xQKZ?HSgdxK}=7TiC|!(AmcLz$Qk7FB!RP?53ze(h8`Uka;ZRs%b;(4sqo~p$=xJ{i;l~_5{vKau*|Gk}R$ev -DtlOP8Q@j8X-#P>9=^S`h*|z)TKRPqhOY5ng1nrEINWe!xBw^&df>pK;OeL~sNf>(tF{9)-nSc!FRAd)=uo -S1I)!vgQ)PP80&O4j|f4kI=Lqs;fwys+CgGIG*-6<)?5Ymsw5HN_i8{Npqlv8xfJ?_(GrrrA@e$6lqG -;iMR!bfun#;$BwH%!AWDo)i$r*Zp0~0@ahebTU>Ejt?y8!PQbtL+eiV*@I?!N$vIk)UeD%%bRcGDnxz -X=J-Pm*COxz`HUke>xPk@jmg~{nQrw*|o!k|+Q4Wx_#3%*o$i_p|rbBDTwz#aE%xf*Kv@H2f9`}ABQ~ -cxRFVmFzlR787qbpEGf3+av_!-X6NL4G6`r`5YNDg!&kU3N`xT1He!0UG$4}$eq8LmI6RCS(tv7{;&7 -`rt)B{?RZ(?$BLw4i0j*~{4{P6Wd;pFKMWb~d4dJpnJR?o&5XZ!zewMyP}~e_bYUD=inLhpqEL+-TCo -u-lkXJ=$`fR$z0TXGEVfy75M>;^c-b)!n%zUxSv91W1pU-GJ;(s^nDsok(s%AaD@vjCJ>G%ZK=AK=yk7(f~>*@@_w)&?V%pYULx6~Q7{*y$qI9J2Ce8xcJ|9u=9r2Yer! -XNdZOMOXWzQ3)Hw^izQD#UE{^?(l#G0o7&oy808pMYRjOjE;7OSizfH~e8-RfU+~mRCJY~tI0fZkX*M -qJxmd!D8eT!X#x$@(aVdj|RXs=ar5X{cT+LGo7Tir{I8I)E~sf$d@vwDrH1{ZqNGTZYq;CW|6vRyb?+ -$VQ_4K*++Wi~0t$2cI9$!(MGwR4p|mN^^vq0Cm<6nTi3JNp49KIVCoIO=&`%y|LlE-9n6EFb4JyShLd$S36F2-4n($HJ{P -SDaq3kmrQNzM;KGvqkm0JZE|VBVqEuBXRn#LfJ2QxzTOWka81eV03?X_v4XoeZs#e`3E8*r#I@>>E -9gyqaDs&)rZejr0SAhdouDv{VJT#j2(Qs$*!iL&ty^-C -&M)16fueSsJoHL}sv`qXa=avR^mO2GHv>GQ9u4;GH$D7u#DNmc2fL25Ftp9PQ%5iPt^~Lj^Z{6u87iRm}t30`Wx=)N$=x -A(kRS7l~!)zuMQR=zoS7v=CU(p6Er21{$HzEVItOL3`k7q(^%O?UKlS -1JVTcXoju`;jVs4S&CsLL&rg1nq=rpH3rzC6c~F3IwXEfjY1WYD&85r5^9EU5UL?}n#2|_YwIz{r)KE -tpUulYFx1<)!2X_|DtO5JXft^AZYVDbmG!cM-XT(<3kJMdOk+2KHfNPo( -1-PswZ-$PsCT59~~^D}MoW+*ffsE*zxcabPhG2j}%W=E8I1K1{*m-O^Y4N6_NJmV8K&PRRR2xaX{U&( -U78i@_tvNu>hj@G&HB_~1l_%cYg1QrMNrc%p$7VY4gIM@>2BT07vrHxerG -Pe(>h9Png>8}MPmnk+-Zfk*0coBG1JOLTGRTkSwJW3X`M!2blI_m9#5o}H+GgoCLpR6SKjgT2d;DkUO -kP6lyIFxE=2&af&r!z*D3>ud)V^C;M((!0_OtGamf4IwA)0s|;fC -tfZm){A4=DN9NiE%=%i8t#HQ`VS^Zj%PcL%N6~mc;+86QTkQ%oAp%&NBaVQt@Du3^dTgNAKLs19&8V_ -3r!PExJD6?6JFtm|AAX4n`KrYH1*&ZVpn1q|nn$!k`(q8H#wR#QOyMnvKOnQ)#M@<-Bf8 -6h4p6B)nWYP%aRAU8|r+^S_9qX<61hkXdG`S0G(S)>%>*IihiZshKqa~tjZi^1FvkDsFd4jD*~p44wRI?!=p*Z=>Q1 -u`H`!2s}=Q{PA6@P3gI_!AmNr|*t)AEUtEd1QmCSqzkpBAH8@wad -rW{TXLCdRMKxdR#Eq+dMvYgp$kRKT;D=f%rYaIcyCIHZ_~JTdglSqvBFB8IxJ{-9z1Uq%>(X6u`dt1Q -6+av=ev5C!_$;eAKhf-328mHE1x6X}piNK9w4)Qo}1#l$we1E%wwbGy*xgm3_?NBt&HSxM--I?Pzyek -30_2L>#?9K6iE9>2Ud9<#Qv~=LEA?fR2s7#%4|>9$de?S&r~kQMOGU+E&rEO3B8@${O5K`%BfbUAGYJxK(@ -NnYI(dotB-ND!X94E6mP$}5`-jEDo%kmBG!a&0?t~gE_pOJyy#J|Ji -Ac2O@K<7tA50=C@9nvifnLgUaYzkY6qHGZ#UFyoLH!)53fSLQ3W4W;oFR-Ds=0ui{geEok{|-HkRD8_ -IsVum2&AEv^B(n0N$Cr3WpNzp9$)6ko>o5d6+IA=3y8?3VXsto987fY~rYZPL%i5HujXBtX3p dUY -NB%UtR&{@(!zXGv4-LEua=wCnQ{uG=O}D^sUh5RaIn!y#t>Ixp9XE?^GYi#nsfanAqv*hywqSg8Lb=QPvtqb4rd%HRUYAio&<)pU)CTzt*p51 -yc$L+^H#+Oy_65qAzhp6@9E+`N)3TUmp3#_1Jar+QXbU=ALG|17$p7^5%+!M&x*4VT)cHHBgXY8*M3^ -YSA@55dE=yE{At5<3dPShUE*B+eC@(jR3R&Ke^?^VCTLynGPjvj}my}OcZkgwF!YBLPU->yRA6u!2@c -TM{O52X0~LGISTQ;S^OGPddG97lHda&>hff9o<$EX-xQ2Mqi?mr~h%?CZavjnS~mIJmfxz4S4FMGBl< -WO@p02fe%C{Z2>3rOQ@hw^dr|^@AQf5wtFOKKBQlfTRL1d+~CVqZu%+M~BODgJz|SSJGD~W%){(HJ+` -`S2pp=Jf%8cskSyQwV6lK4zX(`>g>toRkB{ch57^(g?PnrK2`vI+gNMU{)T2xJ||zn$+bLJnal%n)g` -9O8zAR+bQNJHYSez(JaaZnX>yKM71h#Hds!Q-mz>Rq0v*TwYCWz|gu~DG(TQ`s9_YmD_E+hepzBe(cG -I -L*?)q%xF$nZ&yOkLLN}6eoz3pw8Bv3GIW2D!`L*a|I^e8YFmif9HQ2ShzcQWZS2cT)V`WQ`6>UEEmBJgwAGA?#q67dzo5qBGiM-pP9LM)9?c3*@N-(z3#dg -Bu}iGDc|5Xa*BRKG_U=?8m9+f|yF^L8x9YO{tYPn -P*A)a*Nf--jK?^3bP~}I?PNR<~fAH6G#S(i(pm>O*hcP{X)|<^zeP5X&OC9ph(lh2BB#*J**I#vhd+K -U$J`Y$MnevClFLI;;|Ycr9r1LQ@1c!%-ujfJZYn$!vjW2d1cmKfWj`ggynSDmNF -j5{n@sowN0pyJYGJF1#IG7inKg+1^DwheJUK|2w3#QWu~kj-7(nUSi5(q*$np8>AsB^Ml`{3?p5o(JS -|9QW~Cu*$iyZLr?`nxB-h4ntH8x)bUHCtZ99dm_s`?C -jl!cr}%kX;l3Z8qXV##*X8$ZU$sN3-t%0km^_j7ETX{Yb4SuRwQc%)XmXTmaN|H_XmYi=octLZrl=T> -%-l3EY+jZttEsTJe;=c*Ql}3caPx-%4Xh`TvZa3-8i`88MO+1(;@UWD!Z|ovFok44VoGy(2#GdY1WLgUk^-jfXUzd!L5fmw0?&kc1vq0ET~M>?ndH)(AgDYe{J8M4K;hudwYPrQat -!7QJ6zXv!BrlsnB#!S~6RKXscC2$jP{<$Lo -(udMk-Ks(H6H$wDr@h%D3!D519*$DYviOMQdbZAnPMtXCBn6DG0rx_Fn;kCv6ajUR*asM;%(LhT~v^)dT<*`Aa387<~$8H1#dF{ -yF4*3|@`=`Oy86)9w=$Z3shh%nW2KDM~L-=64OhUO=-)A4j6KZrstCW5j&933yeRyJl(lwfbLXpks2q -9jrQsbYw)#LZF)3|WoT&!&0M*N|km(P`KO1FfO2wQTN#t(!;h^VNg%%~npQ4&B3i7l}Pm@GqWX_!+i- -loy;Rf*$tDb>Oi$4g1J#> -82T`El(!vt|s0;8n%dv0dCSL&V~p>5Q&n+mnvzaY|5^{6;D9m=vCK!D27@y)y9*T`WiI6h4uzIux^>c -(WdDB~kH&<3Qg5N>oB4HB<18gvQCUhBS?Wby$NKVac&yYYcN`ZmsXlvb(g8s=VLTGB#+f38)Ze3mLib -7ouYyXYgC)0I+m?>P=b_Q1cSTfQdQ$gs-;Gb;XiRbIp=F~Y2n5HHeXL+4&+ggFG{Ti#LpPjFY3csxYEB|z;M>3Ux);rL --ezY&h4?U$1hRjU46c6ZgxJOZws~;ISMq5H+dJ+|wHWhX#EWOo)31F67{R^0*#Hli~ieqJ#E!xUOiz5 -=dHjl^eF14Flq>`;umFwyT}jo(7d!rnHuS=nSo!)cSP`7`^7=I|!_i!Z*opDSEIeq -Uu1i{5?>Gj7c_ZLzyq*m(|~a$QBs@{X+o)w8M))D+s6%>9);{#q>0r#GHm -D>67eRx1}N4_{g4SvzUiT1(*bgwropBg2gko{FENBt*^yz)yW9+4WRiiJPy}>B|WQ8;x*iqgP$!&!?z -d6edE?=H7v=QWDhSZh-b^^$E~;FYe9Hf=N5b1(!y~IobS}NBu@(l+mXkEwmvvojKJO<^teoFq=&rhr3<4&9pwCNGF}M$-C(KP>(p?r{(R_SyyNrKrRt8e -C*&GxK4;b%Hn`Z(2xy4Uk1=&`+JY6(n%=b^RYvYXxP2bGkC`QG@wx?O-HMLU_hGOnud=ks@#vBz{~T=$A&}GLXwU_%VJM^Jb{-1j&S!)Io_;1;fsq+Gw|tt)f5$jvkDhFd*f -qQT9qd?IYmQJ~Ye`DcM&HZ4{b*geALA%8}wyjLTE^WuwL}pz!$diu$%mT^ZU( -$Zg94LofFv5B;e``@9`+gf4<5S_x0@ceDr;>e%bD;fj+ut~zX8hkQ0hxxZ_R0H7n)j7UD;CxBc(m|0y -hlJPcW=O#YsT(-S%Qd>I}c+()%V}w6)9xgX~IS&pa?C;y|*_tqn8aBfrQw39y+KI}_`^QtQ#9wP7A79 -fgk5+?{d@FV}Zk0YhC(ZH2nEF67BlyB|e=y0Xv~UT6<1vjEZUwd!@XO0nFGd)eQSPOE3wN2oHrjIIMq -ER(D*XY;xqM865=ar7MvFPyx802|5kuA8KD>Ow0UiSzP8o8sl;7G*7O$Shq{J&d9%{>zft*v~@fHvMR -3O4#4kLIWW;kUt6r!j!MzYelEVwx#1>Q0h-G+`v+60?_B|K|B-e=A8?zLX#i6q5uXx`!+g{qql|7R0f -IraR&p2gxYuJIr(yOgl8|4TQD7Y5<=KPs&qnI7f6kgk!o3>R39fE_Ealw)m8_*YAKEk?W3NuHD)376) -K=A1%j)-~?SotVnwrrXHE&TR{Mx!=W?a?%vQx*TX1uH%Y)4bY8$q{YFp;Zi(9gQ1JovJ2=*!DPUm}Aq;Tlh?n@3B&@!>=v&{)i*FevyG{S -12V~7kNuxd!G|Ieyby^_KqCFvyb>Z7Z4dSNT@E!mu^=d-KW2eg}2cjmi{_);~ -vb#+VQ}7rTXa0!vOwrI0Z1D0JJ@bj~y9IKXe!efcj`e&$tLOy3|Jx!`sz3ls#+*0H}}RX6S2Z{1vr&* -khD*)Oc$kWSE~xhshaIeH0s7IvvR}q2y_-6V}G<18KYGeOlwBQp1~4MBJuCd;iv(^eld;p?zIMxR!v?us_crxrVV#p!-!4hz!Z!ya*hkNpjo4d!u`I0-&}l|BisB~Pl ->YC0mKpZ7w!^UJ@$NXaJ;d3^#dTLYTgCYXRHG7~1RS#8Y1g}goszw#HHg>CGqk%kD+W~1w`+_KuoDm7 -i)oE$a=4+1ov!mR`$9Yw7bT>_$7UK`xVx7Vl5UQ>j_`z?xc00mCaW|z-Qp8i+6QUFKswWT_`V>K3;(8 -Mud$9qzc+EFTgw6{KR~KHWly#r<>^pWVjjv!8BG=MDt{J0%vDG86Co7Q;zU|W?wb@IC&9LO5h -ArR*1&6hb=~AYS|3^twevDrN0&FZ*%pxN&1^he-re#Mf#gbe=FDD&P3?#j#BIz#4if*IxqkLS6cy -T#68!Iva<@P8O%+z1 -)agItL$=2sg1|JI->yqv$)vy83JF`QiyP -_&{)+MKEc`lWWGsZgEN_5;&H7Ohy#AEjf%2;dk2K;AD7t*9IbhhEwRhzft5dSool5OfmKM_n`nUk -z*^O>m8o*FsP#;XHWagVdP2G4J`gj!{wAn#31K_;P=MOx$9a7My-FF9kJ9sBtD%FIuWv!rQ02(uh@m8 -yIpzgF>It$CGsTL;$@QHQkVFE;3kR$(}JWQo%9R-2sQ5b#n2`KtJJC|jM1K@;`v@NKBQ{c6oP4KwYZT -~s?-plni(954CHlL2iMQC9eG@%Zq>S*Y=7|>=H;#rmtQ6GO_sXBq45no&3+1XcjGn;u%(@HYLzrxEr? -FP)ora+RCp}nW$!=xmNlByF@T8Xh64;06L#A06GT&Itf6RM+Lb1Q*Y;_D_g16DAj-*xTS~8!VaKj0N+JCa@dD6dX&XJhIz+uj@F%Qe>*V -LHyTng4%_b6?cd{-$^k2s>Fg7lR^!mLu+w*ttE)+9%Fq)!@6gw*vJD3$q3Lg^jSckE68 -iOMN%d>MqPZ4O{~n)-q0l}%W*pEGLa2>WprM0@;FHf=yYr1%3yg?Sf9NyQ5uPc5w*Z0H$6iS06KzS4M -0+caw|Ftfxu8~a*IngfpTJAl>P88swiUe{w7h#Ra2OC2{n6`T_pUUKbV&ySuR`KMtd`HuI&L=(sY%ai -yRpfTb~u&xqkz4@#0Azs20vS%d-lbvz4$No5S~Wr*797WMLExiWf-Gyfg`fjws_07G=2bsMNM2R@~%= -mb5qB>;Sd(!35W2TXG>KdlW=e+6MR}8(?h1}M(j`bk`ZfR>sIQd7xUod{*^k}rzHKp#srY!Y}x=uX~m -}rq<#!K+wPl_a<CeHtViU0=$PG$9=k;>yzt-W53q{-!dT5YRK$4fJNq_1<~XjNK@>zbJPW0XBe3lBhNa -eq_=7EF*2d6rDJ|HuBAyjqsG&!z969STSTlj~G;ajmYo0$2i*_AxEiHAEgcMmz-xj4S#Vjm*MM`(tc*!AYrTCLacq*B`&KC?KEH=Hv|t`~6qPb^W(c)vQY2$b -$kGyEPO>w1k`p~luE-Dm{J*A#h3kUKA+CshUpl#t)lY9OOrYlY!ZtwvW4ao7T52<#<1hgTWD*2gw#qz -j7VO*{n;Z@Hk@Tp#Jw=^tw3lX()|GEMA8vu-jDJBo|TIjN(#bHr+uB -ZB&%8M*sVj5?0ps4VYK(o!wA^)$=kwI$1WX{3jJM>B{%8HgSqS5nVQxws1e{c)}FNwQ?RLAU8=xGh~4 -Z*qV#+&OrRcM=s6O?JR2QZaZFbb@G6-6z-!ALu=SdwrGofR10f#p>vJtqpido_qGyBJ%$YThH!88+Z< -lV69kg%})(s8%$1x8{Y#dUsD~G3keYSmrkpiSDl2+;X8xXjuD_R-Ka;rFp?7GsdS<0X@DTTIGlqcC*rWO -heyaw1zzVGw~K!n3HtC6h&~+8X~}FrWP=8j7IS_DKTkeQ$8@E$#_su@#)*Dm+;#B3k%UzOiHd%=SApC -=aDF!C;GzQ0B$K{5<{A5c4xR7_N99yC92QsSR@#59=+H=h6DqK1#A|)e}jwZ!8gH{IWXvAx|oeXO>oXssgjxCNtq -M(kc%k+KA{7)ESjfrLmf;%dRNagjU9m}B|knf6nSVuOFE~!m+pHP>q73O*$4hd_tJC2+)IH{p~$8Y7; -KS^=V+CbTH#4$(DD+4alFv^)3$CDDCiiELd1DVCg#8+@L!-l@A&p!ffKYkP&M{Z##bb+_7g?@hCC1$h8(PX>hfA%_#KW`GpDAm^I -?YjN0F6HIU0VM|nHXuV^*)}sg0|X4L7wDw!4ED#DjB~xBjU9O@d8RiTqeMM2M^{rbeC&+?3g|K$&K7> -5se#t9cy(U5+@$2FOY@5oHz`Z4&c02P{1jm7lNHGb8u1FbtpG|zW`9Vnu*v5z(#|=Iw9_niPEp@&KA% -Pbbe6B4fV-RVp -%Ubxz+()Dg3~~pCZQ~~LW|s1>8N;2NF-AjDxF&+2tufwqd_66qlx9BvNRa|DtFZTbC -4DK>?R?8K-Cd;z%-cJkW?o0yeK-s#JA5-_uV~!VE2hE?6-@giV~iby0DlIdU|y#XTS#jFbnhmEyQ^DM -vljdlZL+{P}>g+^$yH@W&ATO;44^v8RquNgI!(7~x%@bNqVs(7Q~jusfUUEQ%WgsIk-N?%M&=Q5Ois$ -?Q8vlJgF`l^%J67!O!+1uVHr@U2zQSV38oGacpjZ8#0N%1O%-J*u#ENzOC`b`$XYExanY^c)Ck)|S7Zgba0Oc -i&GM)j+N^Xfuivm*Hdt`CnT6VgE#E^mx56&`U=^lim@2x%i_)zPp#_0$S+Z)3Z|-VXNMra%aN1IR@2<^Lep>IWp1`;+hSR`0B9Wh5lztD<<7RX`QZg7j7o!{e& -AD!B`Yog{o~)$-1Z>tw_^k=ZFsGduE6jF)L}p!bgfOdmW#O!^;(v=zr~WRW^Ap{wk%0ci%)Y_b%CvfA -D)GN()0*poA0JmSCni8_=Qj*j4b*{86K<3Q#16oyRebwc-mGIR9=l1xx26fppio8D5sx_!Htlcw~lCP --;lmA_9!VcE@4l5T#ysi&7-04x}2MZJtJHV`#Wdg@*vwfTG;c3a8SY5P`i}PzJ3Gw34;;F|80S@O|n` -#;x1L_S#%vVY1+aqw_TS8_TI+Kt)o?8%WXN+7NW_Qe7j;{QAd1KZ^tu1)2-y`@SJ{_7c`)vs;+k#ib} -60bp^R0L9;QWTtu3cjEwsPo&}LBPN6RjOl!D>@L$iPp5{IU(n$!ZWZ8GeWc{v=b+kL9!jsLG2q03Hj4 -6({08fmlmH#9ulwB1`;TFndL?674jdxZ -89)Lrx|0a3r)TO&p#Dh)8jlP#4#u>YW9>;t2FfwVGiy~&2ZK%lkrw-RR8^^d!L({;3(WYjwv^;N-G<( -6u^_4}iSG%a+wtMvQ(Wwh0oobCi46?TV${g{05;{#zOPLe*W25Mit9bxMc(W`GQidi7WwAy{}pd`J;b -5gJ|W_Q+4MUGnrquqrv>(US;yXs{M`*D%*+1$Tr&1Wd@MWPnkw6@NHvxv5$ceiB7ttL5U -l6us(=1$(yj5j;n-jF}HG{?+h`k5C2ah4AE6~aLQHBq=)jSy}<3vDi++NEZI+9PJm4@2w09I%6p;;SQf`Ebo40OG2&p!ZUhvzJ%p}SH+IfNN2`N}E^)MKPB~gt -DxqzV!`Ym2wED%PE=~?3*?_iNh{8m--9*pbM9JMm;oV5dSPb2)4j#J5&8ocIaI<=Y{FM-%j~z@sTVua -VJ)2_xO>S2BjNBXhaq5#frowZ>`J|f_MMy;d7p#kOKnSlM#Mny#7-P4SPZU0)Pt?ITbe|~tG(Qs^ldx -s6Sq9oGhNEQ$+B`x_G>ByqLUs{C!tysBvD_psZ#W+i7d4z;E#^0zUoV6{;_e|zlN)N?`X33%RWnR~r7x#3|GI#)%_M&5>0-ZcO-*Fg|7b%^H>LPOUo9bzd$tP+~a@P5D0G -!O6J7n-ieyQDi>&D|h0U4i!%Lepei?B0$1a2fS&ID{k2BkVckVXfDYIkq)~d{-$>Rl_*$!o+WEd<9l_ -pzOMnm<2nAUMow4?iQ_YH`GuA86@!}G4#sbzUEXQDY66|{!7T+k2R>wF* -ckzW-ySNE&>88?jLKjemfma+yvsnn}r21;rB4eu%)(2i{%b<$BS!lx6Vg2LiqeNK0p&+ps9Xs?h!p0AaghM#QJ!i>tPo7b$jh;czs5|8lezmG}Q`hvf%~$D~?cI1>nlR|@QOHq- -dlbpV^g$ZM=B1@5PN-daX}XgXCM2DtI)^$*(XWsU%@w8}X)zr9C>U($S(2~FKIlAycRV~64@x5&&fh2 -9qbE^6`Ag^-eQ5CBNOFbzs!CNtko2n>^Cd`k;%Jb=`z|ECtEjx)Sf1`(mCQcOyQ<-W;a$}hcym}gbP0jy`ru__60iDMP@r|wvV?eNB$smqzsP-bN}4fngrtLp8Bl48-Xkpxib+F~kp#pmBt3+d!8)h?YWA{Kl%5uZN9y>w9 -e#WQYdm;=hCi+*uOXICmoK7iv82#z5L^x=jyoDKbe8$FqNUtJv^kK3KbTT$GGJ02tTw*tKf>AV-s+C_ -x8$k5=PC5$#{nwp4D8al!3D!7F`Q3m8WfNvVP41KB*Y1IR$;i=mRm5_wODI*mM -%(@P#WkfoR85t9p%52WStR^K9$K*;iZQ@@b@%)lc)8g6*IVY>SI57!PphEvjwjVu`S~jen_KzUXPZ?oi?_uarqo=Thx1T6U)P{dj%0dh6$ -GjZ3|=U;(`mQ4OI`?wU@s##IEh!uUne0?;mQ{K5?;KyWar&SqhAu`Nh8$rETg9|Hx|X85m-wNhk4?9QNz-jMm~O&ewB0;vl+8QvHhhHX_Px2byj(&Se=+}LNK^*je4&))Mhr1xhJUwIrGXn>)`TLTgEXaN<-j}l9LOYQS -BT0|GSNjI-x4&j#AUfxO_>Qo5tU~UFx+k+jeuc_z3b64cx((JkKNOmNLD4KKE$5U2=ulWfIfQK+bPFp -P%;J!+r(suS%EnR&*LNeC!k(t$qmUlto0+v*ccI4L8O5kFaw0?cLAGvnO?=`T!dvVbLWdJWC__ObIw` -yb*8Tf8+qw -8a4mZ83kfG}3ZzFtkPNzlQ|1_{or<7LW0IP=mzv(zsfAvw4;@3Y3Pypcb+38R0DG>rDM?Dx5{^mSiA{ -vSc8O(qtfuS;;^aMae)GQ?fbQ)@q}^aFt!{ -hd&;SkkDrs$>Pw^NEXD^Dv{U-ArL+=7|7zlU?7W7yP!NaJe0*J7lpED=LRWvfcxf97UXM=VmR -$JGPa_rP!`nn>dJoI`(9`|hnCjye-g?<6CN9x3T2T&p)62*=%FmGD!3$+#Q`If#fhYO_9=$40MqPAtT -)RLnz}KTMRPKi1v2E&SQbZzNTuPiEMC@KsPXpCVp-Hq9ums}NAmv|%VKa${7bPclKf%BvN+oJjj=3ln -=~wz#b3zmK5|Jci!U$eu`C`yyRGozSQel`V!3w@iDThQRTg@YeK1kC)HaZ%c5xsJODd3srImylRl9d+ -APdBwmEt1ZLVE_iSU*u0CRA^K>;v5N90k(jKFL2MbivKewoSYds-Ug$oF2R){E}{7?foPA#3E`iub%J -whTs*6VZkdbhItiD2CrC{3SL2-o(x`b$Ov3P%^w!H!avl!+MWzt@#my@l?q%z==>K2uE2_-eRX^Q -*x{;7oa&c(|Wfyy>S&`|54-8%C&Mfh(|N-Ll$vQQ!&$9TK=AjMGrekE -T^9DNltEz7LLVaiad<(&>X{)kH^s6TB8V@JWf9B&2eB-udU`C2JAU_nCzgf -YF3Pyau&^$TWkC(Pgk)!4x-^yrN+U9{@PWOe;&ex8l`5&3Dvb)zMjSm#NY_PMfQ^pE9z2UWDC!HKx>H -Du3?XDX`;F@EMFA|r3R+#^o4DVrF7Ee=i~9@wxwg8vZSeOr{Cxs{AHcViZ@pS&2j0wZL}KYCI#3)r$KkXnLYLehcUwOr+Wb6mxB=fKTbkm2wOj?0nSg?rUPJT$=3KK&?{6u7^4eK*2Comjt^+1JdY1N;8IC{FMArMF -{)2RSeFlB~TjO>)GJszOe+jhu;-3N)LU@e8pv`YyOts3KYIzgwB;&7uetHlSN+kQbT=vKy=tob{R$&c -TNB>)XCUnN|p9!U3*B$)DWHy_RH7Vf7$4YP*g~hVu?>l|r4O$obnf(}@lBrZC(T@!rm*}wXy*jS=Y#%d0J`5UM8Qx$Lr -v}Y__VK$^dGwrK;PZP})3eFwzc;~b7M`^Ue>qBeT_Y514Af7ahL48JZ>^5}au -oQLoR>o|&R=BY!dchbqL$&-3DVZ5_|Y1{#?3w95$4o%FZ&v9pJDPgJ34CQe36?|`wZHv3ELAgUs^@P -0g%>TIq1y0Mv_n%7$OpL3=e(!+<2k*GO~!|8mQWW!%~aYS6Playf6eZ*iOdDZGWm8Exrf-Ilo!PJG+U -}2@Y=?-~S2@jLCt??Y<-}qORQ#Y+DeT?HGQfBM_#ooIBL{)A7<9lF$QPIKFLem@-6~h$67mnm0i~@n8 -C}=(qWe^a7h8Z8-6m*~%M@(1Htr -!&e^%eu;QfZEtvQSFy^8k$ddH3tz9chO$3L3sT$$+(0D)8RGI!oAKm6zO -?6UF_1Mxxn%erFh3<~MTqGb}R|dwG1p5zcjs20Kbfy%cvM*!*O&iqHWm8;9K`9ZHsm=h`Pp;{Kd$1P= -cYE2|VUf+O(>lA5I4QZMuN{}vm>XZ|tG};u0=9CFTr+_(pzKONx7{r6JA1a|g2Coja-p65aY{+Mdpu`YX(KZAr8Nwb%drNVc5FMl{{;8hN$z -ij$3aE96M|+L!62@Q_&f$0%_l^c+%nsC9@q}R2EyLr-ps&?37V48eeGITt-B*M;u0A|1krG)hB|?3;z -i~@v&R!b`|M{x3)P@)y1-FKDBL=>oo!;-=e8*Q_R3@m+)o;J(EI6Bzw>$EqzrZ-{-R7y(4iaM!1ZY40|P)ii$I<6=I8>-IpqVcYVs_%U&w0dKeb!4) -n-!Cr0GoG~#Qlk4_N=roo`!1>p#BIVo3QlpSLPVw8SKejqg}N$Qk3^g^ektEjlNA}Eh}LjLw1(Sj(_B -$OQ1+gh8djZpsKJz+lA7u?*U8=5ZrkW&spNfeDMll`-I$t+UX9!9C=_}%bf}F|`_=M2qbZ6=Chlyf*w -uSwdFtz?plG%oKWVKVpB0L)QeTflsJJVKQf@TBTB^g+Aec5KZO5eVOmDqKB840G*gBR~+uGG1+=MF?+ -?G_7A@61YZuZ6qdjs#*6S&pN$FN}ri+6D)Z?~?3hPnuIZA7iH-X5@hGNtSeDm!ZrbnWigc_L#~$Adc5 -I)%#d`cIrAkc-!T{!_z=an6~8N>W|5&N;`*=Uj5mY1m`lFCUILXlv)3^YLyo52DGo`D5`Fe-5uOA4`^ -gh>=y(mwoBa*z1_}mzjg?*VLPXm??naNbWJ4wj_y{xXeM0$ugIGaIAdJ-FH9i6 -xo{-nD1>aGe5sA|3xTlDJO*As3eV2~`MWKYRj$4zcu=kz+4xod9F%&;okD1sXadJzOUB$NHA$mw;W*` -9LWu*ZyxA(*=xS%lV%K(dIrnXhvW^`t>>Qo^j?&Nr~@IZY@kGgcKx*g~msgi!Td7lnDDZCVtgWb8NXG -0VC)rjbP}mNCd8G8iAl{3zh?C`KhhQFqz~NsBhB6i(WxQn-1od!~<=5ywtJ?2_0Ih5_O -UF+&NgIY}ep?*ohrTdlzD3?Xh6!Dr>Nl-hbU6oXaYYrSqr?k%srPjXfR@81W*xJedI(LUHLAC(-B}j@ -VLGE~;J28&yBOQ7|e0p%+AieqoMzg^f%|1(37ps{k%NbWx$QhH^lU%z;R<>oAD$M(-Nu8org#juxyyK -KwA;|`u>8YUedsS|D;$3v?l}vY#B~TgQu+e?=|EZ%=Jm3^glnL9BVXQuy>b?gwv|g812}S$@2`& -E^4^gxp{c6($wRvhVJ@8a6?&{X)Gkas5%=k64n$wx|eo)BTl<;8PmPWgvY{yWe_-w;= -wtO&B}UFKu`_SpT}B?Y;C+nIxi3H$9aR@q{sz+ARLG=e0fZxk-v5N|&6o0f^urIy)m*loU~aR*Zu^*U -}_5nRxo+a4|=JuC&u;Q4>-4xZg!T4427bGf(uP?jGvm# -a#vPO4v(xsBGi^Fwog}}c3psJdn%C7@TzA0}1Udf7!tXyaBBc%EDh}FvNn%WYbp_wFEDJ01!M_UwY&&8 -VBcg%`U{3<$aGR$4hY`TClMmchK#zcagFmQKkI$yTro4Op^UH7+#?hA#7hpuIOo_2G_NJF&1KvTwiIj -z2{uf=zYu#Jm}gj&2CUGK0ofQ%*aJ@a}N#vs7G1zh)`=p`A4oeVGE61aa=B0&)wn*JxvrFT%Iq#$hSO}|Y8{j43cj6QZ96#VV%lZOIi2XSzE! -J6i(A1?M|`B*3llPo@!3M>t!RABujRj(luT1(h%`(vgbAzr$wSdhbup`&a_HFIj##+fp%Lu?*DakWt&Y3xDr< -U=9&2BW-OCLUBExdXY{t0c%RX8`?#vtm_kB|$NK0?C9KR!kCiB={V1UgUI~@8NXXtsLK=PP<07GH -UJ1o{B{a50LW3_$=$uJgVw)pmQ(dkf+DWoBW0< -4s0gF#Qj?{Gcf}Gsp|N+OaYgN&2eYV-+y=)4?WJWMU7fJP>C`y{Uf#kv(!oSNjceAb~ -f6YHxf?h`Pm3;+P=gTpSah%tz>>rsFbf+QWqsTebKkb}K6+9(1w;eCk%16A)Dk(HI!lWRhh+;L2o~t4 -0TY?wLOkbKMq?}k@uel9<)WZAJb70C|o-1KE*@Db-tMLSZ{D)>L-Rq&?PC-j -TeO53d1MnFA9QmxC0Lv;FW+UWk6j5b^O-&gHPQ75z<;daZW{vj3OvO}dklqv#9Q(Ly -XINF7apww+$!QLzIeg@t5PJu)PV6#io#qhNh;Xss8#yO5C51YEss*r<_Fafos*ey^8&@q&L$H!Nm7G_ -`$ZF2(X;5n&RS`Q*7lk?U(k{|%RaEV}i1B*vVf*(}+wYt#jVOX@ziL~%J?x7wh`lOxPIDdR*maR}>TE -BhiJ2e98yI?o$cpcdqjimF5ML(B+t@K1CVcuxGV9VYU0L&eGbnYI>(lTpHg*;)Hfw1a*n~SO-XV4b -Gt#D{BKB00cG8=Mnx2)Nd~)^bk<4N*~MDe%nHj1^w~*`5g$lzE|%q(%FsG6^mhf&g%C>Z7AYy|HVx8} -rITXYcSvjOJ!Ohlp9Dq`T+q{-3wkzjfdO35Rd7L26BqOZ7Z3|z4X$|Q*j2n12dIa8*Ztfbm)Cvd?Uxn -bk+RQiZ&ql>CFimfsr^1N13~$K5^z9IQH-6TFCj$?;O`V?N*hw7ABxn%I^`*>io8||>y3OAR){F97RB -Ti0;6YZ#gwY3)Lh_;C4H@QelhgM>!dTn43uhw86{Sb -9#F;nI`B94czm&#Fv<)rNSvzDiCIzv{0CdQo1mcr8HqqlO_vuhBQ%_v!!vuo -F|PEW~MY$nAy@mVdhD_g;^+FCrq297v?f4K$v$*JWod9H7qzuXXuT;h7T#ycfty-*V5<0x)#=h!ivu` -(!0Wny|dDr!ulYre;3wAVcjgOPr&++ux^HRov?0!)gi1xlkhTOt%P;4u%fe)<_jxix|AlY&;ltXswiV -=U%EwjV7@3t2rDMzQmC*VgSD5iLdr(xh4j!O1!d@dkX*te2-Xr|?Fy@1Sc75B6;^Bul+uOO0P8ei4S_XYSVLjGMO -X*J8X>IVu!aija9Dc@D=z$&t`XL;um%XL3D(A3SYu&5Bdl?-ewVADoQSt0!Xp{h4}^6ZtUHBu2CUnJb -vCRo23?HF3mI2mqNR`YeJyWFII^cB&QT)^ovtmF69W9u13E^#pRUj5^VHKR$PwCEUUz*-nKc*g;fjk_>f$8BmH+N{dXKyKJId_@=L>boW{Mb6*^ay>yYnm{C4CFSV~@8^D -|Jbku0AfsW(8zI%B{nU>snNCkVt#MYBpO0Vm?(vOKClc3+BHU8xRYbI#QCu-JA@S$YOC;B*c`#ScB#P -@p=r#kh76`?ea-9C;zJ_z-+C)9d53{^T1+@pbSxpx}hlkqXZeu7;DZxg&q@B+bR&=9-sgfsS@6J$zcN -@SiQ^Awr2WY&^dM`j(F4P-Wu*+^z%g%+RInQyIcgIXFKMC%B;5`?~i?^(LJ>HDhJ14D1SOuga)p7t&g -Dh@Slg>6`4#hsCjN*#nW)j`LLXP;`MeAm8-eCt1>k{#>jKH7$;&z&nL0AHWNMq&ja){L((h)1R|)pM=~Xuqdx) -ATp>%rTtS>3hJcXf60%Eg7C3RDHvnTiQqmQ=aA>=&^pJNz!1fIjski6zrA8e3i~bjTlc({#quF3^{xX=R{sxAoMb -eSyA7y1B-xB0qThV6BOd7;9AZu0oyiLEu_m=Fv}hL>y7-P<*Sco2xq42Ol*1%I5%VWVgpiGdJQB5e3B -YO-!ZH0dEJ28}Z|w&X>-19hT;Z-BpD;m%n?8tJ)#yZ8tk4e_Nk8+cna`QqPXH{)IHfeq%du+K39 -RH>@58B7q@tkyRrHoc0icI(A0Ac{h2SsgVLTsP7bs@8Z`zZ!d#T$jQwR7})58njk>igf=#aydfpGmS@ -z;$9cZZ;urDxT<}tjn!6tGWI7-eqPb%eQiZd5V_tWTq6}%+t4cRlUya&H2gvlk5lKu?dOrPw`$V-luuqv&H)i^?lPkvf+q9cCwrz+-9RG -;K!9D$<=MFvQ3Qah_JSC$*6Q#T@BDjUdC;MY04Wa_D^=NNIEtp7fV!Aay7~PFs^kwXKv8g+pS6pV$8t -Y5gUPv(Ja{pCCdv6{Gii(3UN2iP<-^Vj}D*B;mJn`df*l$Cb}o^WelF;+?wb9qdA_rvaFBMqkFNNw6{ -bZzTJ=~??NDs#w~ML^uf9OFI=BwImR%XKaG+m+v|v`@52t^yN^@bo(4(y{6hn~A0sL$xt$b2T2j(9w0 -lk+Gl~O)vLeWj>_L&!g(67tUB-JXY4jkoVUZY1ej)Ez&T{&=4|-wc5wE**6@8eVd&lh^@OkapdDL#^R4W^;I|}__bA-fnC#AW -=#dU$Hnf#kScB0yQO -Zt^i0xJEy!c+E`8=-%6`77E2U6)cSddequeI4KTuKM18P2%VfpM3|;~*5y|h9;2Yi21^6qTTYkE=I=VC`|4K_V4O29Q4OS_6MA*s@ulk&wCr4K+~>&C3zA1BICMNlo=?X{=TYYmoFb4;5x&*4ab(QeNcNok3t*I*aAlj)bypvK0@Idc;PWoBqB2Pvt@ -%(DCQvjUH$xUZpCyGiWLy4v{GNTkYwbV4z8^sB5=w1_Sg+yEE~;i-PR-zFP~mlp7gV0-e -?BxXLZT$TlL$klBEwMna#$yDv)Q``@Ts*G>S6r)!+4$*icbX7JCvJDWoZqs93Rg;LlpdVz@xWeSNMbM -GllK-Q*Ppnhr76%k`^Bsw3f8QGqa1lKEDUrNow(Q048&QBaM@Dg?F3sld@y2BKVyicVIVj}mW>Q+Oh> -s{-8OTjeYR??i4zME%-xI4$=M=*Jm4=9MP4)ifo#+g#VNS;8-ghzP0CuYD0~hGj^fi_~kdQ)jHkrXA~ -J@1x%q5wXv4iAuT(s^|nVKF+P6&sl+-m4H*ZaFgEZO${f@C<;km7LR;HqCN_qwn4caXCym?zqGDd;GP -1H(E<-(Lv!#*k)M9``HO^>4%!zxt?43%DW9ev?&{=>)sL<@P23c*53D&Fthte1OWpy`FlU9FG*;(iPK -!(;-ns}*em~+|m|%p8vN-xRCIuA@SkomY+W8nNt>l;XB;o`fyIDumgz_NMgjJV#7n&mf!&NIs39ChUh -st#*Tw1Tj7-T2HX$0!V-TJjgxK)Pw{I -2>_RICUKPT;$U3i(Wpw>TF#EP^G=l4WKLlx=WTM(6XA225Yh}c_)YRsl^k}s>I^b*IIK -l-np?Q>VMF>88v?tkaa+x}wiWN*O;x#%r -8v|wOr-_F+(h#|fn&+&9^w;8h_qD1?}_-uOJ!vadkF!TKqO2widpgvt#hTlDdJ9#pzNToG_t;Hsh4t! -7&)#Ky7Y?5DS9fwG)-8i&c8b46lGa$DW}M9@KR3k%o0z^e_zt3ati%YRXN26NLDD41Whk5@zB&on!Uu -EmW|D6xkid=%IVJ;M8nZGY48^`e3&9tS-fcQCmM#ONU?CEN$Th`VmD|9rRs?ot;2GHiHTr$>02yTCY^ -BVq}_H8R;>^=a*0Y<9n!5n7aeG#vG8&ovxV?@?XnK8&gDMT6TDXG4%BbQ`Sm|-R<<+PecSn0bN{Q>fc -V12mKqSZT)dJ7M4>TJ)__<^j3iyO2@M(-6%86c+C1Zwn0Yu-_`0gd<<*N#8Z -DD)<(CA!zEjB7n*T=RXCZp0=X9dU7rj*fjoM+buc!I}|ip5pw?nh_mO3FTik5pAnRM7mwQR3qxzR*m@ -2HdG`2qq%Ct-EFHzeBMhn;^SVb5%0YU)rbYZsv0p#QH^L;R3k8QlnB=7#@h>fvBYuuU^>2A8N8Bc~G)(^$<%nBU<%r-!FXf0S|2@hPFI -uIl8?P`~JK{2E*|xi_KLdBB!fYjVRhF7kW)ESB?1K#owwLQ6zUI)re| -5sX)iAs716XY7u9tYCE=jsYOJ%Hq;`H5NZ(-pl!8?oxRi|UibUeBDSWdqiwy2O6mPZ#fW8Btr)TR2oV -A$9WrBhlysmdeqAx*f+od?GcHq%Xo?rQHx$i?KmC5qh-ioZM$L#1Uc7>4#5;Zl2CZmD#K^BT&4|I9+t -Q4Pe41-U?A0RiD``fAmP4T#Q7A)HH6uc4snCoV@8_i%5w6gTSl$xTCe4VqZ$j2c>vGMAk6vu488K#4T -bdCCQJ0lkY+Z{|a|jiRLTv{ul5ZR)jl%76&4_xr!A6{0bFe1!XOC`7XzhR{Kwnl-*Xj_vg|5U;&5G8J -+YXpm+6Oc(Avj4e>_ZcaB5)9FCBMt>|Ivo+y}q*kV86YtUHw@t4QuyW+q%)40dpSvr)hPI_E>T%WoV#75%ADdK#r(71>VmwW78)Y}eh?UJSHY|M -WTw5@89GMi3>?!%pH$(a4Unk^qN#67IXm5uj3rR;2Z&YIMOL$&L$JRfY@h2KLmDdGBi!bx|~JjZ=VoM -@>SD`pw4T4i2Hvx3B7TF2Kkw@sFwO=~f??b+J+)I7vEZG>(9Dbi@bh_C$+obKITjMOzWwmAuFa0^Aj?b7W|lQ!}SJ -OMkj@N(wGA!gavKuAPo=imRjm+bCtrd6vFdLl7N=9|L~E@ngbIGJfXar?A3_E-ub~uIWa2&e^>pUYlTVa2)rw`ImQ(jY%X05tUP2haK -O*PZOPB|E0=}IFyq1R(r#3HPfkda@+H6V~JxnVegBoKcZ;|J56tcjA!kgjTe@5l=r6lID;c9-G7o_d> -uNzierGfq>E(C@_hf9BEJ8ma(gf~b{>yaH@NG|u-`DtW$yY0u4BqFRxKH6JZsZa@*UDrQl*#vD%LKfT -8p~vTsmpNa*e$E+)MEb#TsC|u+$*EYr$t8N>gb7D?mJ~Q8 -B`q~qQ<2nlN{Vw%pxDD6Bu=2&x<<|k)R&15NJ%>5`?hzS8uuG=!*Vp-p)bRUA>&znX%QD!?B0dVcp&A -w#{?-gv(!&dEiNCsS8#FHrd@7b%g(LrhblH5l^e-|rE=j`cFxuVTOr_->UA%fiEwpa>-s -QkuiY=#XNJ2&L}l4N{X>-%u}ssnkxuAEafhUdJ7IRduAVv#a{24GX_rB(%MJJj*Ij6bA&*a7yTkktr> -eoH{s_`(dUw~#vMSBhKr{FQW!#n+8DMMYsM4Aok<4D7YRc6nOMYvGIsqKeFFR+~<+hXj??wy5!TN4tm -y)Hum}THAw@W(;U&;UJ7=SbU*j1u(A};8|J&$$9-Bb>rEby$0giV$|524MZxPsH1a~h3b+AO(IDN=N_ -{Nc)Y$u`~im2FByWme=;A8~}dVIwFo`q^+%VS8yLRnI2jvfI9FBh+O?+5>G7#@)7|(x)`_+9Wj2oOvD -No#km#hOpQsNs~NyK>Y!3Ux~CgP~*pO^WFCAbG6bgoJZW0Et0k!b%sZZ;5os_c^B$p=Q94J_t>oY3_TM -h2R-7CJ;$`t6;P#=B+9=d|^Xq -&hB5oUb<2G^yA`sZlA|Iw9@bP2$k`Orp+QS927!Owvwtb|Um6{q1Lk)EK9COo9y3rYSmm3f8D-l#Wm< -rY4MRl&qaFU*S{#RmDUf;~xFmGDRpV5ABPwgcE-8(+X?ObdCUG_;(3MoTW_s7sX_{X48^k3s*EBn=($5^EXgKzbs5G^DY(*;$uH- -W5Vcz^0OmIpnu9ehaE5FVuE(MZ)sK+IMk~r7woPG*Rb~AAw9-8on^-#6g?>*d!cb?m#vBz;mP+0Btzv -#<%6F$W{5=?htQ}$cv!cuZcQp^(@D!LVlMb$>!qA5A%d9E>AHV?v92-s}QK#@x~4bB=-!ztr;wlCHi& -sw#vYPXMTce3CjopkYQ^?RUg?4gEL7sREv$z+m)Z9JJcm_Gof53}6>Q_Jj8CGpx4SNn>@VIgZemwdCTnmp!IqcHljO2+(zpl -{+bqPxQW%vb!{KuJ@gk;^1ISlt#Ig^2WKITPeTZqOo0v3kGo~CAKBDxRVlB4?>knu8-sR=~*4BJ(nvE -mQ3Tg=VHp?!tS}4vgTd+rq+N|X&5eQZMvA!!8{jJ`Y0DuhQ05cL~{rAX3E&_l+PrM?n%nUj!~kM{|z@ -&c1yvIfhp2^bxkZn(_zYYjxNg??CyoH@QujMbwN1IX_WS#RP)6Zl^s0A<_>UG3pD}qRbD>Qv0@BS(_F -j7pi*Ck(^IZPj<1WWeh9Ex-2N4h$Ro=Fe}YLgvMflYk!8VLj4T5PCS$OfhQVe!ITd8lV6z~f2Ac&-Xs -}tZvXDF$VgOoz0q9Z;K#L12#MzOG9r#VKZz|hoizo7bz|~gQdFNLgP_ZPaqBtnQe%kT1-WKAj!c|orU -DrCVt5~9~DAvZ?&*IyIudN++^-h==Z?B&*H+@d^n^+zRI~exf6xU8Qz_sPqi2zknvn1l|XJ^*nlcBWc -6Vd7Fa<7%vl6B@ezbB|urs;UDCc7JBrd(fXg;|s;*=}E{>bDY`b-GELXL;EB`~+n)PqU3Vvzsn;o#Kd -kk&?b-_mgnx=SI9haL7kusahE3Sr{nlrjVS86(6Z=-{8orBHdFWK98KE6JHafQ_~;;z|nYu+NqRQL`d&Tl6cpy=K6^+5zoPtO>9UY+ -q^X=h&@r?9q(*-s(T*+oe7}-;LQrwraBN9&_B@0e6fYir6>ixUC&ll)2TX{xetAER3|o1`c2P&nB8p* -@N4oRG@%i6X>s2OTiY1Z$$ygvhs$eUBBY!d16RVhrKwxZEuIG$?Ct}8N^cRzgCW2W`=b2m2F@99-c*B -y#E$TCfkdFaPQ!W-3?>*$kW4U-ppc-1U>(8J1ltKF^whB -71R(_N34(*kP7qEIM=+1TO0bUL1%gU~g9OJ3>Iih*DJ;Qwf*Ayv1XhBT1ospCjo@{HT?C&Hd_z!6aEU -*)FpXdyft6ql!Ji49B={RaCBadGI)e7sYnXvx3_$|HJc2t2$_XAMc$(lfg1rRC2pR~c;s -g~vwM6q#g2wAKmzA@#uZev%!^Bdyo7iJVO)ULElgF>}I};nz&vaQF_3`QW<9uNfBoQ3fQyEb|+(4|KJ -T#gupuZw=pU$SS3G5~|+RHwSjbw?;EYg|ICb4)H!zQy>;SxuGW7uN)8$DMaUCb0sxq=nmL3S}dA)g^_@LUQpAMUxY@kf4a%@g$HvPstb -ddKgW@&0$oQ%_s(7jRs5q#3tNF_LEcfKC=BwtZ=7V%HC_O -v*WhyZc2G};ePa|v*qR7iFjnc^!6gIg|qfp-ONrZF0SDX}qoA-N$pheD4%|}gJO;1fnjgJp-V-_zSQz -V!plb*qbBzg?6ET@MhB@NBY9J*X&l1$YeO%Q8YU|BrhA}V)M!4i*s@&a1{tSOd4o9Y;CUuZ9~p+nCTX -(ZV5MMXvzT607lW>^=9`m@Zp+B2+jn2da3rF@V+(nXp`3+W&YinknTAU@(E4ok_f*{zuwB8>^w9ATT5 -ZMPOFZfTaxe2elj%`V%L3i5?Rid``?Y>10^2#;_G1AlfawW_JaJ!Wcg5$t9*gg7dZ{zQIxMH8$R3(OR -YEj1(0F6lx^}zvy5R0TuJ76FhTeugefx##(EsVUQ@r>4!AdE2bnbLOVc%b33)(~ -`9?J12Kh-s1d%!rQGyHv5vL%a-5qr#n}yEM_qUR`aqPTe7uafd#d{n0Rdg@rISS#6py1QVT%+SOuThh -^H2lyFs*8s|aThbaZ6x6~0cS_LM?@-eF|l6p~7l*u++R&?YAnPezLnt@%$S`41U)w9qA<~fjVwDL -LkvQ)7=+}o2$_c563GT?GKE`~NOQ3Wn@yAyku5_=L|Fns&Oip*Mcc_I8DOB_3_>_>mL*5pMT;W$pe|er5a*9L9Ctgq~r@EM2i^MxZE~0d0+(8 -$xvx=M`y%45=C^v}o@4T240!+EA8|sgM;07D_UI0lit-AW?eYfCV0$yvsAA)so5=ID&Izxo -;J?SewO+6*xhX<_J1l*7wbx`p%~qHo+VDqO=wgtp>`Wkm4>->Q_!Zk4%|bwWegB#a<=aGM{0;!dI8qx -5cZxVriJcjN~y&HSDDjqv1GC~6AnE}`&NF#bIKIuouQZO!lt_o -|<(PZy(IjLp>)i!S5ea4&hCCF($KJ#vpzNbfl!ra|;hg~G)u-c>B*IGN4L0uSay^DddNDU0dI#A@G*7 -fm8WHknw`2_myI+#5~o8n~0$1@2@*73~ujOsqSZ$z=8*^Y||&c0HMeWMa3qfy^7=e%{1-lNn4VMmH}u -m{=b&E6PrMZ-TZ&B)|c@Abzi|8ZLPHvXk7(sKWm@$DX2Z=O7^X++gWcA -0thRT%r`ZIr@Ag$SK_QNwGoYFx+!|3E00XjA+JUlHqfC|-sDr35N+4JaOY+2zldUOTV$ztUFznul2B^{2VYKkxXfz5aZz@=uLx_@6HR{{_$Gb(;O(%>kOGzpH(ka -e$`j&*v)tnmFKJ{rffg@z<69yx+Y2u9{?p?&j}TOK%h1EnUP77_4N~YDa0=U3Zr|*SOYJth;CZhCknX --(T*3;K7F;e&o@|{`&Y68=u^?`KhO$+4Agj&%f|rfBXB3FTMQAt6N{&_WB!dzV-Gy+jmq}?X2GQ?(RK -%_r3T2{trGpaPZJahd=(uC!Zeq?C9smzWDO1<6nRC?TM4$eJ|Dg;6C-wA5Yhw`RQ!k&*$nJ&i`_u@!} -=^FE32Az3^~V3lsm-^#4!i|8HLyYTN&RMfpRQaR)^r_9DAhvET1y$4xxyxM7`_{T?s-dN2D1FZ-Xp?D -u-vaXXJ{-|A(DBA<40dSaKtM2os+mX?#BS+F$4mSM{&$cOv%B8zqM{9MZdTYP?2fpxL)NR#6vW)#_?7 -ua%^Sdt40?S;ZkewUMBNt9y7Ld*DqWoc|WOA#xLE!PVx*lZkzF9(wd@NEG*2PY^qoIKW -i-A`Vl$Ri(lew!_PF@Ee%<6L>nIw8ZBk!P@2tp!#?PQD>8!=Aq&+hQGK=wBoPSjEaY!sX;yG7Yu@LqV -Y>-(Xp0Szxzi%+JFj@$`IH;Fm>(WXQ-jWE4`MObU-xV=BBMUSX`3j7&q8wO}!Y7SUP1KKN!4P1M9udC -nk{F{4tB2?g^F3n&We$dDzF5NnsFi=2llGgLh0WrfsGvaWn1QFkYip>zJ+)wuc*~DmS*SJEJcMG3oM3=EE`cUaHze=I&^+c{!mN)l7VW(7JSvjbI6 -GT=2#2z7hCdeo{XC1Ww9>K${P1M<+G{ -JLdR4zyC`RIVV)&jmB+I)bBb)UZ4oyQDa;fC`Cv3-wK3F}Y-v4jY-)%3v0|$43vV!8(Lba2mHk5pUD< -!ijaT;19C~H{f?>`5_qOmK9%1s(Sa`Fk&2+|%ymCG-kGisd-RS222V3}08`Io>znA~>Mibj%ymI`{Oj -pkT<7sWD5A*36SB`(*r;!Y!UAD1y`omk$Yu&I_Am#qHGqmU9amG| -4O|qzp{V(@+;#p`02LG?ek|$?8j%?_D_Dx#HPK~+`nZxMZI(7dV6#GmGOV(y({~dynki-)_-PVTe@mI -WxXGJWNL}uyEVB525VU28P@9apM6UIzI{sFHQt~4@7B=VBS!AJ5~=4(W4*K~77VD9Gi=#u>~?zdp25( -lSDw|dEi48jW{WxjB1ws`abmdW6@#%~v}qYRwh0B+6zatDEOLDFGIlMCE-K1dn6IQTt)MA!Hdix2B%s -WMDD|&^DNv80A%|zDLRTMhO)AK==UFDu*k)43VuXFlcM3-9t>V4VH_4u7%Nf7iW|>xymXm3T$k>wjzaOgl4BHSHhACc9A7BKA$F>7q}_NEl^opEYGiJrnkqtVc=7JLUgb6|8}sRSn)R7|lr@j&%uK5& --~N7SRDy!(SV5BnkPDCBzR`3MXtn$Wv-Lu0@e5nUhhN*l}wsuLY}=S8;Om#XJ -ygIp88DU3k8z`>jVpSzgg6FlgWsQ3p1=1PdTTdY`w}WHN%>NDNPznk$o~JXUVmY;M26q1uSVaA-4f)J -B)?MwXa~8)_QmTpXYzcGY{8@E2c){pwSbS^OT<+Y=-{%j+?LW`|{~z!ZE~tK~`;J)k9mio*Uh-8; -tnFj32ChuQ2{|Z^!Mb|@lh@sKirG#v+bez@y!@_J;={oo*dKC@0&wr6xQ|fYqm}m<7`_>?6mX603r%zY@Mg!FiQpu2%dVihrq+&ISVPy!fl~zEQzvvy$I4iu*PNzYi -7vPZj^*0J$vGGW+*FL2W*=zoE_N-;UaU%hUG9OxeHfk8S_6c>j^l|8*Wvp8wZbA%w;UK3d&o63FvbFJ -Ik9aen$M+1&V)Y_59zxVX#Kw6a8)TPS?hrlXq-Z6v$ul5Cn3bL&-xk5X)(a&+!d14}=ed-UeZ!yi8KBaw^#mmZc?6jRvk9gVBof3Dj3pROFqj~OpeI3Bf@#Vkrgba%QL*F&*a$IrMnEGtdM6+=JX! -c@2Gz+D;>u-!^TzOvbmF*?%*}wE?TA07py)FOW8on+6|2+Tt^K1Led|Eud%A>|N(a1HN0Q~qI@twi%- -tOM%4?khLDo~#x3WO7{hZ=t!ij`+Lg&%Be8_;$%O4{Ch)N($leoyc`xBB`2AqD+^{Lw1EV4QH8j&ve( -8JQTQttS(6uPyR4#VdS(sH;kHM?D=O6Y*=wguhP1N-$qDkcmnfO(y1K$(okwXvB_?>{i^f=(nU5cjWt -EEADt+^kN`y_sTZh54Pccybbr-Hr&0L)!RL&4R=Eu?%}Ptqfa-Hd-8k@lgK=wY<3%M(JT)bw8Y9EI05y@&o?BhF8&@9XIuH$JHE55BBnw|V6s%H(;jNnow+$Z -P!>GGqvwIB_CNOG{%pIXTR3x62h;QbPYF?7;^gWY0YF4142^H&|t5CHwT#PuZnQmzes=Q3R$JT%Jl3+ -I#kw1Ty`erArTp*8}`?V@Y7+(xrFai2$eh{>B4*e?58cp?4jVz%M)|!M;rAcN$~v$|d2f{caeqW3-YNY?^2bAAiskg>_q=)$e*gXhyJd% -Zk^Wc6Uw)_jId>hvM^(51zs3XSzoNHu@E7>6Zd|j!asN4_f%i*OXU9{3#ym=&-u>R**LBdN`wtKaStu -Xfk^frN<+MB=!8NnGi)A-=PTYThLaaq`=^CGXc-Uw;fOv2C;K1g-FYFg#0pxxvdMt0C=wdB_@c(erwE -Yc@2TCeM_ye3@oW7%>QFtRfm2b_X4P1DO@<$Sq?2Yh3U&j&ITcj^=3sm^HTzO<~*4VqU__%M$IAVd`U -f@fhDPgjpY6CP%IPZAl+JIsLUcxj4z61dTdV<~@d>*9m`mgc%@ZrN*Qc@C2Nl9UM+;K-U9$&wHJ$vrC -=h&;SzAAY9lTSWjCr_UA@OdGYZC<2y`{K3$)Wn|(|HNO*+QEu9+Z&En(ZMcGZ6GKHn4rm -~qcXR=I_1A(gYHDg&-RU3Lx92<*v -GC_4R^R@c0a+{vw$%e_|RPb=vKTZ#qz8??esQz^`FZd>EU>r?BPxcJ?6Oz_#+2*vI^kDo(mo9#nmrZ8%@BiRAs{8|>zc<0*~A6Cftl)D(e{ZYm@yv+DZdl^6Ul~; -TN#UDxW6Dj_DiocBFucP>zDgG-Ie>=tBL-8rk4aX?{Ns3?Vi9eW9xRp}Kp%m_>6rQ3Kc2NpnQwp`6Is -0iCXLVCJJ9j&0^&2=l{}N{x4slQX>nZ+C6n`AWpGxr;Q2ZqnzntPfMDd@d_^(p@ofQ87#Xm;z-Cpr^N -(pqM_}5eXJ`{fd#UD)ZZ=v`zDE{pfe=Ws-n&R)E_(#3sH-FOQ$J;3;W_de|TYm^jg(_$MUAm=nfP^5&T6apR-Y2O~i -L{(bu>{_#ns=!EE)*cgfryoL-83%ez~3CV<{1ofG4eeiYZLx#vHg!JuW=r((*9Ha#_Dn<$@J|vCrQf4jfOHfC{A0{X@d*h@=A_=u0+9c}8%6r#lVTIhN -wG;2Zs^fX2>|~=p*JNmIsF)sg<5LqV-Y|pKau`0V-gi{k~tws5Fv6<%b)O1j~!+L3JH`(j|mgHbm>9? -JpSn;{nJfD&6L3;^1ori1c5?NkAIBbPd~a-xH%y2pmzQmP3Jhm5GW=kj3Y*B@{b-hB0Qph#*RJ#CiusUBX&(-P5vk!VlY#O&YdQw$0Ut260b4loxc3U1$ -B;0il%am_wq*ivBI0&M@7UKp27TzqO`IFSboAb)z27Sr6<1~<{{ -A0#*#jQZAW4RJGuElun%HtGuJU4afR52b|X_@F5dpxu37FM2PW)E6p*sH6}?8S#yvQc|!OhIxk?o-ZE -zmdlb%gDU{{`=V@k37O2d+af`apOj|dGls5R(R!=SHyVm%{Si^V}*ZIKEocN@y3fZR@k*`7yIzT55?H -v#5Z5DZ@&44oj7rVx!rDd>eMOr)0xv^Z1D5XKeLM$FR~5woV`Tvhc0-@*`YZ2GKxAnYM1L5P)E0%I=X -wPqkEcl=C870d^?-M_p;mhVYY!EV=wWO)PL7{#itH2Ihf*8XS;qd#UDxWO%y+Y;?JV^w^RJPDgI*=|7 -D85kK#A=bN?Bq{AZl<|LQoUhav-d^ynczfpr(FsXc=G4;?xbnx=agdh`hH(dWkAL7h6?Kz!D{Z)pGi1 -N#r{9HbxAk=}ds={Im7h3y>FduZ6uFxIp0jiEz(wIAA<0(9!w@rIs#`iBnf&_47A#lc_*y>X~c7Z^%$ -23^ysV~^nLZXDV{r)%HN&o6M0Uf)CO-{Zz!9dzyCFf^!lw*XB*-=TPVwev&z*9QA^p|=kG+P7;LM(wL -l|Lc8wk}pN+Memef;U;K*ojE%)2p05IJE&6M`!XP -|0T=bULhluUjJQ2ijMLByFkGeI)-wXVdT>m?s-}A`>_+l$Yx8A*bqfh}jPC1}myLKd5gic~ADWQUdB! -$?{p_cd`_4q3323}=&lq_)~sNER0>ZZe22stAQJ~8HHVU-E|R8r} -MU(Z@w927aSWK8;W#)EBwfs*2VPlO=_%&}2M+MhKKo4IN%bRehyD2RfBfSgZ+!dhw_j0S=Mj$RyqDVB2*P0<;{I0rX*`9z -JOKPv01YajPE-K?Djlj_g_g9aKsbAV>f8g~?ttHee|2^BTB^ss#1o~UVdu`B0)Ozw(W6I29en-u*Ww* -GqD@d67Uto@hXtT*pgv%t-q3as=l%EJ7x<$Nsr_@*!=+1?xc5Wv7s$O4{?vwV{w?^QJ9my#HzfFf^yt -w8z{ddK|G@_z@VDN2i&LF*8Z(PFjPgf)?Ay0b0C3;CcduwGM~)m3cF+JB0DeO|K^+0N*M8*Oa+LFNA9 -6l?ALm2g#qGqL52@sQ>U*3&carm;&z(C=ZKdaL!GHVq?HfWuLPpRSejUk6)Q6w}_@c~FuAm2X0o-9n- -GBMzmjYl%nWMacFZcko0B@8(>iEf1oX38`x#=U$jR!a%wV(5dy+p$<&Tk|d2E4<0-#0m5cZ~DK#>Pi~ -3;rbQJplYwQ0qd4-_uUe4!|#IUvfV22qO;qlq>22dsik0#C}n~fJwa_0GECib97g9fEN4}nY~ncAOZQXitB_v^B3rDs)Z`1kGG*Te -tR$D^OWtUl0w(I2RFf%brRl?LEXa*tO&kjbB4K9)bdG>Sh-G(1K$;K?BQIh$yhL3m92My5e+&!9nV&y -Y#O_E5Q3b3O<(Q286)p!UC4B=ccw_>;UB{7-Fd9mYdwb1EGwEuaDYL0h~9{2?c|u8QV=`;(DBM>IS|G -;FlV_!ph#e4d+VkT_2}DYs|Hq=e6#+B0ZS+cRiT+cRWR=vyM0`j+^kouPlBah3-_g9sR#2Cw#bMMk=x;G*!1HEFB!6vXB!7u$_%EVi3(@fQLj}Cn;N -2XG1bXvwLL3ijDg!*#~<=S<^M5b#tcARsDSngyn!2VS80LlhrGwU267X975X+j+gFd~Zx9VzSB~K?-f -8475DhKc^Iu3NZCD)5*De~*%X4D5V_^(0&Wz;UB{7?9-1Kz+3CU64YX -b-?!trN(2@9{DEU(g8tuP(oZze6;XXfNTFdpXviJU@6IL~vSRtlOtYW?4`k99 -uQA4mcU$Lw$PDyh7)JoMY}vw}fBt!naghq%b#kS4a`ECte(ncRs -J^`jRSxK@GCDb_t4Oijw{h}@nR$Yy6R#6?%Ei>!!eq_rL^alz1s6;g@#8I8vaatv~E!}cjm}6;1OdC> -Yoz{9{XPOzwfilpYEu0>VG#(nKC6IKR^FH;EMK$Hn4K#NmDLN_1SQPVhh`y}I%$+B4D6ikxZd>5g(*zkYoc$xc3 -S;6R?9p3a|o>M4$Xxu~d!+iW(zcI{fxuD|%=3n6zf7HW+a;ElckvI=7%j7Kpaz=QUD;AuN={DtNlo0f -RmvpU9jNNLaOsV)2({%^niww}g1(^UL#y6L8oSy@^6nVFexlB1#?o_z92K6&zFo|2NnD=I4Z{rBI`@4 -N3lfe+^Qz!x;2FF-$`(t$Y)Xn;HgAEGZr9=lwz{G(@p|2h8N<|R#Ij3x3I1ML|y>EUHbK>t_gUt`CPb -tWYxtpfiLuWz7v|A?(yx4uAan3Er0wrm-vHo@_J>#etnF~Ym=zAN|vxC2+zk;+$UeSkjj9_Be1*P$-Z -9?+&md!~BuZqLuQXwUn%J&tkUBUi&bk&%)2J@n8+VjTo|D{YyVl$79}M&7SqKR#&CAWmazfeY|J+wca -^ph8Pp0N^P#ZOGOS9-GhKE;CX8ERQkJKR>@BhHtxfA^-gS*Z8@ge?Cn0ebws^4?g%HFD)%SyJ5oyzIy -fQpKrVEHlCN4C(51Xm4cT*1KI`X1a6?CB`sKkKtBw=M<0#0g7pZ@BjJWI66y}LwDJ)7=f6Z5{nqs#M5 -ea;L}6j!SAY7`pDqEe`M<2k-^{>Y6X+1~Abc&_>W6TDEue+1_&x% -%9nBTz^B`0{;^)ZCtf#)yHV32?+`O&O7fE^#B@x|M>Ca`HUGeggfd0G^hX?)Oj!FVi;e8?|~;w@DOP8 -<}Hk=5T0bopF=`IM!x#$tM_0I@q6$GZvwzeSZ{_K@WPt2-EJ4{Hzp>An@lD?WXKRv5Aa8vmUOhnTWZ@ --P4E%q0r(2?7~ZyLRo8nEcHr;A-(lzTV>>wOj$1F9 -IK`TTWvO>?#)Rf$EzRzX={dAB8ejNws7~|cXN_iqM!DDfSU@4133d(ga`bCIS~N -3gRaLPe;jL!Px$!w{8rpO_^ac8D(_OzpvuM8=sUcVy&3-A{0|w?vX5;^3rwuB&z?P-Kk&c#FM-$bjy?dmtKUaxT>IF+fV;rod;Uv(&p -PlW`XJDuuJ=QJK-Qr@@}>i2``T--@rNINSn#vjh5?{kU7N?a3Gd(;_4@(I%x8Rkeg75QJ@|X~i{4`x@ -3GC5>ICq_6HkcvpbOza8}J2hL3V-`6?Yn+KK(oS{WAQ43$-u7|I~k%s`IZa(W0(-s^9NWH*iBesNdvJ -XOKnsmVj?ipar;79=}Jpzp-^1@aGkfc^-5C(0OePLw_BL0#WPJ>VTOAF>v5jOG;Ly9UwxN8>K=2k%~q&)ed6Z&? -I?b=?Bx3>v}f#7CECj37~6pFHv{s(zK1_V1rRTx&Sl5arHw3rUmCY`>2b!!)ygK>{$ -SHrF*SvqOPASWByA7XG!KBCb{`062> -=Ed~d{eE_`3XnmE>C|53;J!Jn?`TW@d3>D8-OXWVz+eg9v3R~{clk?$J;Ib>ISin?!c$wO3Fu0E!Z>a -Ol48ii=YEP)slgd`*+kc&*<0TMt#33u>99`}ua5I~My0l|mBD!N>C8_S&gSr7kWiC#&&;Xg_}LL-jcipb6~P@&DT&SvVqHLj6Zx}^ab|k^(loS2Znqwa`H -j@IG#8b#XW(y34Z(_inRgb0Y3tx0@ewfuDr$zzeHPG8&y@MC&4e)MCQs -TSkMYB-%}pW>vx$T)Td;u~6d<0nG@SLRcG(d^h(ZFuw`5&@1c<|sN(*Jz|d;oC)Vmr_Tx(Bp4d-NZXx1 -UcHI05q0Rk8f~yf~Qv&NMcLlMRdY0Y&g5wcCx317BY(|3Q2PIfTz2kF&&Mcc9aiFrz(hE*BUJu)2}dh -oLd7F|PbX|DpY{W5;IBm@y+A*bw}2^dpF?gL#AK2WSC(gx^qr-!2<`K_)K6k8!+qN&OEWd&rO>MWaTI -Dg+%SPMj$89`r73AM`(X0rUVbqV0Ih>BQl^k>iK=kxW$M$AZN3SlQ3D51WEK2XeZgG2$zd#q+uJ>R4D -XHW-VyCZ$GB9oZW>`SHFB{8)B@AEyrOismnA|Ed1#g7(wX(?9ZkU*uhZP3Pz5N5+g9BjP*sDbNLhuEh -9&HfN6RkGwa#Z{*J-62)AAfFI`!dLXj$#gfQpM-D_Hm+^l*&(qqqYc~kbF?8tA;;gK!ZL%y!`u6QB^1 -u@&Oc1$?Aoc;xaR&W^xs9Mlb(@TWI_bLI#9Ke>|Fv%2x(ms@hy!8kfy)3}B0nqg_~Va@=e}I7x%$`yc -08C{N!4TC^UuHFPxMFIE1@JYp%jfX|&^_S8V2(&|CR_C#(ep{FYbM3#$ED~i8&@h#f*e2 -UfNuwSBbN&vL@ph;2IM0iN5B|BFW^4(1+;+PhR#Gk-+1GViPF_!6hi*TjT@Iad-m+~Am+w6qYu4%_l~ -q|*)r0xV@J`qAnzC8L%0rn6X*-fOI*2fC6?A~bQNuj_R%Kv&xjEt`k+4Kizj_WUJ|_WtR9adpbOYw926;t{7if*=hECqIXU~FP*tTdN?U1e!`u|e>2 -z=lbD^^6NPoFMw#n6q&IR%*FuWVbik9Gq43tqzW20l1sFTf?y2k;m$4&)8 -K=)ystE#HNf5(Xyqorv69@<4b4`Z(hiY8(?k$chy|Wwt?(X^fr07w7{gq$v-fqH -W`$_h&$>n|{e!Oqy%$cvvnKS1r(orYo&!7L?f&~jM(mP9+EIvo^l^>Z@ -Wr0~>~MCYjzBe?RcF&6_tP&-y*paftZuEaVWeG5B>E#UN?_jLw7d$-ka@SX8w(xF6Sor{Ti|{!AcK&} -BXT9gUgbpYA5NCbaD$|5$r9q5 -I-V*MXZ2WMW$G33iS(ED`Il=9sG}WUm2P#e4H~BgH1gftpl+*VlTvgi2D$CS`>G}hk;LpenW4eAcldz -xa!4zqK^3T+4s?$4)O?y*+4&x6Kqg02le%5?-%|S{5|-3@HOGuhOq>{RQJVqZmtr3e|h$KpBoRm8uT -N(AoLb6Im{mjY$?i$pxKa3O!F(L(ZTMUm59rhQzTh)#c|8B)`e6KnIZ3Dsb`tHw9wAOZ-O$Tq`-jEf4?T^3q -aM%#SP`BHHXHdQ+>1CHwtLH#Ef?L3GPKUp#D;W8HDibLGe>Yuo!D_BzONzGkrEObiq)R50$UBNY%Ko5 -`fJxoSg#l>2VnKJni5tw#$Q-_t(GLqQSF44*J{VEOc3j?T_3wLL9Dt~N2*&B?q0f-CiS5IvCC(glq~k -$>_O*_(gXAvWkao#KStNVt05G4aQ0)*TZ{fxPq^laUZ+WAQfX|b#|P+)9U_zIuUx5^uDMrSBTE(?jE< -0$B@(9(l4-0V-AC6HN+nVzT|bC=6Wv`NB~9!iUlQAIGNsxch6z#^DJ(vrJR$v_>I&LU5<5_4i$0Zz-R -O(y&a0{OezB8eHho)mey>XuRjLz*Y^9Q7Igk4^s+?X9*NbHAM0E2d_|Y-V>|3xEv1TmK6_F9*>uI# -j~N1=dOx|pIyEmv%Ot^%=~Q3oB{&R8c(@y#=~wVkSnPNiKkoHvRn_8|;c!l2R>7czfrSM{VYg}j(jq& -YotIOX8E##epIurUF3v4$on2gLXNC*g3{aXT6lNCX=jMdV(sN3~`Nc(UQ>BgE)cZ~Sgalqz+CMDz=ef -$-wCH_wXE-Oje`$W%pxEE^O=(WQ{?tlNuhi1~0r>?vy>r6ne~kb6ezEB!btENcKu$qI0seQJW`>iC1{ -C+rDQ%k2KR+=WTd25Ab2AIVIZZv@_A)nLXuR#^Tjy;r>ML)1{-Nn_-Zr48mztQGn3R;Z?w1B@7W*QKidD{Bs+ypx$~y;rQ`69Jd>C5=lNRxA@A -#sb{Dv7T+KV|mHQ+77k#*e!|54sly8$;$)=p4)KF#BQ14eqs$MC`Ux{;1xdwbc@}N3 ->pAk@kYNP1~m()M|#*P$PYz)roCo@3NEZEB2H<#Xe}Cws$%IbdEZ=^2Yog-iklWJ5x`3^KaY&uYo_*p -XAT+xA{<@cq!{TxslvjepsF*ZSpz?Y8`EEXh&$Lo@!(m -`9^50r{8?Vfr|@~aim&Ex@wfR-zK?&-&+zNqH@wqcL%*&6oWI)tyMNdhN=5 -9Ge7oFS?j&c+CGvmEqvR>_4EYQBw~C@9DgBfQ%3)=_x`inEwc1hZuRW!W3!Ml_`UZW0vD(;a9A~mKm~ -ZCC`Hk)oPZFDHG$3QuRq3u|DmhBNQb@J*S4I+T_9+LHPn0Lsb?V<#Ka?686Pg^F9a@F=K9Nu6)A=iWIbYAW@*PB>Z}@Lq%?-PQ+_~;6?sE4n_d~a -)SMJU778B{%)e~C;mCmx2IS)mw2 -VV41R4aHSA@Zx2+@6hhi9@aV&Klj&WY4?V@gnEaTg|>!vhCU0O34I%?qdR&By}Ld^U!s4l-({p2U5y^bXk$5z)q&CIFmJUUurjP?tyiq&)<3Mn)(Pt_)`rxtjq!O|`QdzKYbxnJ!@l6j}LDD;G$r#CiQ8gr; -`Bh3kB8>_&YVQsY}d%NA&8SN~0ZsdO;9`EQw;IW5plKg@4d$m33v6Rp}(n(8=D&sX{ow457WNbBlFzT -CsG?UB$<_xpK3bV0nIa|lixu!qZU*K=YcvMQ#5IU_;8ix`>%|gvXtwJ(6@m45X=SD|kgYlhF&rC5NH8 -V-Wm77z{_f6=rKeO#jvg_Lq5e0hM{p=w&C!XrUAE#%}<-7R@{48(m=DB6=2)BaxXPLX#{o1YNHS})x6 -1-Mk7q7b~&5G_HQ(qnnYNs -wOJF^oaM7|YysOwe6HG0+N~Vbv78Q0n)3|F_;0;Oy;I&MA8U-(BEGZb?($diN5mJk_1p9o`XhR#9@d} -HC+n~1uj}vX2ldl>O{0-#X^`S<|X(HLx05%<4h1HPWiE##xot3)ailG;5Z1%sOSAwZ5^=Sy` --*_+uh_fxXP8u?_4~_B%VlZe};PTiLQ5vMrn2&)VbcO8W)UYMRr>DRKHcBb*9nyz`thha~gwf`eL;RQ -BQh_yCg2r%7|2=HGFNXgJwjNK{*7h6vxB7ScEq(4M`45szKIZrG`}k#kx -nJRr_h0g7`AbOtu#k0za0dTVo}tWFmXMvOqu#7OsAj3B)l}__c5|qv-ioMk#(3A9ZoOw6A>L17AF-Ql -$6ihLAY$u|@4W34^09mcZ{)Ui(_E@=XLP-rzH*jwMm4l7ZG@Jq?>6_59sHMh*!;viMtZ8)8e+X@t+Gy -9w~ms2s?1c5D{I -uZ$x5f`|D(^+m(kem*Q<;*#_PsrW4CduiAB00A9+fFQlgY814-LfDC?C?$`)^jx7R!19q~T*&U)v(T7 -Con7QeCI%x~p~1O-!sR_o;#h+w0Jn1sF8|G<|T))4f|lPly!WaEw!&o);=N-DKbsZ=RDl%2{cr6JkgR -JBAMqP|SFcawTZm9z$0bFG7xr&Vb4v_;w$@-L2RbxGDbg!1rU2{k0Kb*x-hZcbi7D<-oLvlwUXi25B^ -3QJ|3SqAIIda_IR>lUBJ{rzO63>lel`*cH#pba^Y$>Z^ud#J(J=?^#keu!yYkG(sVMp2L>=Zj -omh~LTZY{g6-N0^W-(ufxH~tBywj=Ge*xINOYAbb3cUDvtYY3` -&r8BRB+r_;;HbIP1U?h*H>`?)K5wY<7s1Csh%XiOWEy$r9Lm**7_FPC`(y& ->LkZ=_e@RgZ74m22f%xmK=~Yvq@e{|8V@0|XQR000O8%K%whoCtaWPCfwu0LlUY9{>OVaA|NaUv_0~W -N&gWWNCABY-wUIbT%|CVRCIQWq4)my$yU+SD8P4?wy(B&P)P%nY{4UnFMGj1JnXZNNQ_tLSGW7?UJO` -t-DJCRVRSj6j0JyGXY)ZztcU&uXP1u{Jl~jp8uJ@uWnfDrE=oy@ -^?2gc26jY9iREe8vWgAme1_+v^nfqRMai;cO`}RFHMk$zoC1Xu_XGB#jea{5=6zHiPg2}ne@x#=Lxs0 -@*iY&1KsSES#B+^yKzl5BuF!{o03q>3Ypbjf=51lFOdu{UX)m`JazhOH+_7|CR`W(L=p&5-OUrzxEQO -u_P%@SH`O!toks;s_6=O0lP087@$XuZjFrsCfA^x`6kMz2_;kJ3-YXLFI-;&@8m|8|DV=)leH-svi+f -HJ(1zWG>xM+>KEB~INJumyS}_IJ1C!EKOihXZH~y}Z*o$d03#7?v#L3b|${DNeHLZSm?17&L$4o4se_ -xF2;(U>AF@C3Sm4a0YYiI03FA<#RHg?$bUEh(lOCEpggoywidk@hC!S&qCFfKeViTk~&S1p99R -Q^(iDr2b|mj~ukI+Jh47&SSpFY8N|$I)?J0dTUr(vC9=rt#j>5UBi@O7sEZHdsdv!)KXkajqu|CDtzp -W$--*sy4K8U^IX121(VJkG_l&U6!k5rL~S#!brP#}xsKg755GLs7s;#hMe1Cgx7~wZ1(VO@N29fCE1N=M%@9=Xmykjm=& -U>p%*1YE|nW`M3>j{m%MCZKIyGDt=Ci6u4_JY27OhS3;ZFHSS*V8*k^G>&ouElemi`_=odvGOsJqCJB -2d#nsVWKh78nir2bSL`jd5OkEYoa;PUZ?eApf#1VRx6X>KL(oT`A|pvT)XjZH@$Ne=c#yypX*+{yO-Y -O7U#L}&c)w>7cAv3rCIc#ba?Cm334w({!P#WGxWg%y+{h?F*P!1Vv)0u^>ZdEblxO~-Zq&+$MG{qqdN -yonV~_G1J7BZ6Q=A?4Rq~J=-Nl1Ykv=2`xtcXAE9dux<)WAUexkTS0UrRBWJbecQ@XGGH*uNHvxy`z@ --BETOI^%fy>E{Y*x3dNnwSr0lsQx%l2rx^4LOk$Lij)Gyeg3+Pz7=P3mg=bUwTKs(gtRKB2O}6Q@4E! -|5{pdiSPdNxkk|v~%aIbhL%InjGyhi -*vF3H);zNE0~Ctx=SPb&8bv!MAnk5J#o9BN5LG&;_sVcu1tEM^l5Cj83ONqpdVZZPt}f -bUk~12YLes5&2RYBSfMyZf7*Sop~4_fhfPHyY!?ALDosUCs{kwi|8s -nYD}i&m^;mkw5h9bh6f_uf49{DW{LXS*=Aru6v{-1E#bPJSPXGCtERwZ*Pr7cao0A>*p#P;B5TB$myk -zbP%lNnm>{Dt4%u=C#i?aP)?bgHnIn}()*L3yDLSyU7Ef7b{Rhteyd@te~`p#CH9H$_~E@-8`hL4R%< -JhoG*K_GiJ}oN>81ach#(1M_Nk3)p37u_CamzOi}aprK))>mC74`mRu8_%O%lPiD%ESbCO1*G -lk3k!-E3k==%Ow -zm7R${*ilsmwVp+E~jJ~k!G{n6{)?yE5J{ctRn--oio`y-`xb;^hI^ljUS~O0RC%JdVi-NT^;B#A>B> -tvu^yx(D -W|tHtMtgSC4dm{*{iOZ`$w=E53aEyM3p0yCShi*nY>jzje0eXTkG!7CAm$4!)XW3jWDSG1WdE6P|<3K -1=@KL9_$#17T$&Cdl1!%^!Rf{@@WOYae@Hf7f}t8aZ3@i>?^lpL4i6tn=KrGy$GDbWDTwC%~})-m8Fh -qS<#u_W6zs0_GwYYgd>r!^vP}pYqL&!9UP70dI;0Gg%}a@3%ES_ALC^^C-l0N4bO -WXQ>`(3Cg^XsYXPs_=~PTmtNnSf?nRgfL?8vq1V@)m#5d*1J4DIPnUqR44h58jI(IlqbAqk0f`m8y3G -|h0o(`hK7jI6=Br+u0=V!SBq<|s9Pc;deIH||l@F^C7rry8>}BK+409jhU%Nzq!sK#(%gagA%1Ik9=b -%+kjmnvAsE&L!~D{n*$l_a3&)<<#+~6wvg;)?Zl^c#Q0nJd -q;?6x&ta=l!ua7pi8#(Tczw)qGf(MJKQF>b)12(Mez%liShn}gA?({B;GyG@u0fMEt!ER_*sDayp{qk -lLPRla!}?6QD1`3>%1p3Fp$$Z^J$8+>^xqomhk&DscR$tFouIxeQFa!}+jGDX0A0@v?ChNRPbkZQcP&e@0u@@ie -~0|kF;;W|_gA654WK2#r+Vh0epFss0{BY-pJ0#I1@Ld7y2#nBnHF|98L+4fDvRXgNt6TMt#D%!JKQX1 -k5Hct`PCQjjmnGI!CTm&u?G$sY|Pa1wBO>Zj!tJ0;)Me^{?@|Ui&3}T;49aYbTvYD;a}w?0h7;HEi+* -k$VVXE={2*$LgXVJbJ;A<3Jh*z0g}h7^T0RYfVz1Iw`cKOV`4>gpNX{fTeAYk!LMx4MU`3iZ&2P)h8l -sMhKywz%R_rxM0qGLy-P;kx{J{sqCug?(;uUpPMJ0KAwNs+>_|ilVYh^CI~CUGBY)VzRyi%mhxXGgvD -0jOMWhaNpnDm1suJ>T%J5a^Nxo`pQ$_}-k5!FSrLi#KM)rjCc--zzcaG|1@U(YwT7)RA4{<7$%lnR!) -VCy*$R$$03>=D8FchJelMwfjOI!|c`CT!P^ -0obTNa+wh}=xmJv{{(grG_O7Y7$+d3Br}B9&+$y{E|o<#_osONJ)ZX^v8SjkqU}D}G4kDvns0PIsHS$Hg8?X}%#Jgs^yCBlZqqtv!`+a -gQ%KYd%gG_^8u0~)JkLYFO;(ni;gESwRltwQzho{L#n!6^Y(Q(MeeKX1$FnzP58Su`Tc6N}yDrR>0tm -zva&Q$i4Tj%+Wu*G^^;_&f9&na!CAuaJNKD~~cH(~nBN$GVw6SX^J8_w29cqa>r)pvHoCtbkQ@TLdVHvcq7t2G_|D16Bi@KN2p3tFv^K^J5&6Yp;U-9L*oo7RYSv=ek#1N<^gz9arG -tD_3+oHLz09Ftj4sbVtoKWB7*#^1h*rAo0l!|%-m^F$+SaTb?N!shTiE! -lxE@(eu=TlOFe4)CiKbUrBKXTr~npC#yA%pyF$B;;-AlWN#$$+@q^JhiT!PoQh(47zq!YnOjeb)1`n$Z`z{7+ya5#T7Ie7L~7Ez?!B}1QNVXDW_$5@b;w@Qs9gZ@eQS$I3+ -t0uo!v_m=L?SfBcMLd0xL6;Wuw#38pq8)j4xy;)W!PBkPEJNQA -!)o;|M67oT~jm-!gLOQoajpS**q^R4>z@0h_=wEM=0jE+t1Ewr|u^3(dUc9Haa~#f|nSteFF@JRcGQ( -*;p&d80FMnSiAs;mdc?T_dL11Gqm5o!>31O}&tLD(g)v+H2gC?2G#0?FVfT@4Sjn_D+_*A(?-A~Io&s=%js3z -7VU%2#qos>55dm)k=MDMHMW2@R2F@sXCWt7aew-cNT7g%4e+aceL+%eUHj4LUo%MS0=D?Xvd?bIk0mqCNndHY${CPNn^Jx4fX@~E6lJt_?1 -iBXTooT|G;_(@9aQUzy{uBD}WJ)+vhag`ZV0gW;zg~@UrG}-?y5_MVj6cy{(^Dk5YNaS9<~9On|);^) -9fo!-@_5>HP&;Jczq0)7_;5kEaEG`kuza{dp5&fGv2xF2fvXmTek5aGh5SH?J)8=8w$3Z#dT=uM4EZx_(}Y>Qf#+pZov6fz=O<{1D*xA>Uri&m8X7Q!1*PSr!X#0T3nt -%v~?6ml3gCr?Q;42alN2=%ms{5_|!sYt`0a&;{6UHKweMdOFRppV#lB}(RfH`o`!Qr0AwM -*`88m8C}^>i=l-79Z%KuS8619V>71SXx`lA -zv#k54Ucze6`gF+xUSb4R`692{L)~T~4RHron3h{|$K5NAw!-p#GYz$k&q3cwD#kUC}=N9`*^k)lPbw -!N+05_zV1*2>I4^>_ac<<0|7mxkPQE==L3QSSFEg;j6w;5@}x7(*AZX@k;RTS19ArDD}ThDzBpg#27lgD8t}~*X{fAJluz#!q$|L+0N-th=WPWn=)uosWf3qf!sS=KzI43*K#T8pzB}v -P*rO+?A%_oBLTMre?YRWmn^Jy;f-q8vwvZ$J>;*hln^sW_Vwk687dbe%_5}{u~@C0S^Eo+#$BK`DU!9 --R*&{O755h3k%a}yt`PKawUjbKK4(>`8g4Ex7ZPS=nV4o6G0F(qn<4n2wifFP!^l%9Szk~j`wQ0R)g5KP%`dT!D$@22Km68VlRHGTwpz$e9~_rY8h2j -u*s(?5TEN&+?Mgh>byv{9&#yvXuIIMbEtvxX7pHjNa(y7azwfR{bmw-n?+AQe_%&AQFN>V#F&*kBH9r -ko7ctVA%R&~(z3{6H?_XqlV&h%>xh3{YZ7NBkT67VWzjX~0-2)@z8; -ZenEUWd$e-)zklVBJSa167IKpulN;jmq8nJv>lRkjHw`}1sn737lX`DVDQ5v&3d=}^9Z&11yFJ}jcl& -}5<=!5WbC(r-NV#1}nN!m>zpuk1DJLmbs!H$u9r+esT&bV*+x!}EM?IP-Hd{AxkrTY08lUu9P3e>4mn*LCgH0>eC(jlqrRMWWVb8h=+XeXEC5~Bq0L+^rLM1&Iy=O8%g-jxDV`e}ce!9Pg@! -(%21jU2_bSn+JpwywlVdk?pxc<1hu(dBq4XNn(QHJG -roH+y0gwQx{TVWPodpc5LXmTW4UA-_LFURx;#R%Rkt-Y@R`+U6UmRkJKcYMQ|o^gV;Z?3@?WW38s|8R -_QA!=$_)8X4!u8)G$s>`;bn$i%U2_>W~-4rr13^YJM|#aQN2cxm+IV}#L}tWTS13kh`DRrC#SKbAK-f -?Xxb^Ktlyf1RwNA$n$;to~&Nb+J$`06`& ->R-enc{`M4iWSHoTD!e;J)ybY(|IaNGUpTl|!OK*lxbz&@9I}ALUrIhu9kYTDvALXS!AHe!oMZKTwza -4e(qi)2Hp%pBA2(XAJWZ(0b7qr%Z4pcvaMKmBdow9ZPNt4xcVl-MSe3OG+@6J^t6X&OkxdCKmD|ob`no^G(U6iasN`Ksp%zdXL3$_A{ynOs -j7*Nb-cNQsC&(SAt*-3Q~(`_0srdwP*xf6x1MlGszE26S8~osRcz1Kut^zQ7z77VB$B6dz;)ULElQ@CU@W -(_Tfg#L6dlGX=c(_`u+IoMzIE0{426myZjaYpHgLvF<_}(o{Cw!QsG1POoa19*)N^(Rh|oW{$4ABYYG -#naXpZJi;}r#S+LtY=(GYRq*1mzPjq1U0imM7Uiruv~GHMDPm_G|HS>uq|ZMeA9sER?blMo<&Pkq>63 -Yk97dc-cxF&+rt22*{7#b?bB4aQ4?UnhP5SXm(6Pmk*MWTG$C4aV+S_>^hv#+8pN=2*K8tv<(U3o1l) -d9Nz@Yp|Ggv!~s}a0Pw6Dz>jJH7&mqUl=IP(5C;C%98^e5hIHB9lL(=ziun#)Z2272B7rDT)8pth|e9 -~p7MLMb_5AG@V3HA~&$Z(2V(&FVW6%$v5$3tui>_rcGvU2PVAG^MAsp7UyHh$+Q` -rer==52Ps`bc}x=>@9r=_}0s6@J(*xrni0)bRIBG4`q=rromUNat(ZzqU#Ja=mo9x@8Em!apcj!&#ds%h>jkjA?WP|9qyE@tzB$%*c7E)EpN)^_Us!7b56zr+Huw+IUA`mpaE)&KK>t~7j@u2nLQh7ikBajnJ48M -|;=bTaD<69$Kh9GCzIeu}T9szO@1@{(7WnNTe&4@&^dn>CZQJu_?#i2?M(m&&y^ElXhw!cn@5)Tpk@6 -(XcJ8+X)GU?vBbfOdN*0jW*6P8zqR)bXM}i00Au9U<;>SAk=ud{=Cr05jrNDO>M%snt?$(pI+NsTwMf -*DqcKNFoc9{B~Dj*Y;5@fE3@%S13I``cwEqGVJ8uwEll!XPz&sV^MLx+Y+R=~f2?-pp5az^Oee)w-gD -5DbfXYkLpSbIk_Sq(gYHo7yi{wC0Qc`!ETK=7e*Hp{e*nbJLgdkS<|4gR`WO1}b}U~>Z==wGKKjl8k# -p$@_yao~|R;1~Gkt6QGLcfxOdR%^#TY1#;_1*wZPABE51pghCd{2k;EP`}wQ<;m(TS)t`&4qpeKKdh~ -nc>1ESpO0wQTDMr2k!Ze;1b#EccVwko^fzyW{878^BKoLXk8F~a)@Io;;-fyy11#_n$yc2VTKxC`jX` -9G^!DsV4{LvRt8>&weePd2k9Jl6L&-jQ_LiUI&o24N68tJ=mni9TGPQLz_pJaaj9r4mTqfTLXRcYVAO=W*2e*zYjJbmuy~s?k* -4X9Xjxo1ljq~3~o~!CFlZth(hWYr@o|S8*8LKFPd*cI!SX;_mrG;91pvOL+4qX4XI^J?%)-1)AEZO-0@Xz=`_zy@RZj$~$=t@jT0uNVZ7ke*>962c7Jm&JZW& -jO+%F{KxqiPVqGEBRND|lo#`3={smqJpQdb_N^xqu{U&t)+_qu?0zfgqof|{$eNL9Cp=Strww@8LBCX -dt8CgZ+K`t@ZGW{JIP1^U-?a$5`cJ@te5OAGI{gIqiTqiTw;IbtSvz2mwIcA$fpW7^X4n0jMsrZ^3;8 -*_s?fG>H?iVM$Q<>zPFK>ZuQ=Q`A%VzPhMU_W7ieU}(t-J!=6XKS9D;P)A0AMf!yxUZ(DVjRYq${HJpu -TS}BN|Nu0_xSWMga7vY)vP_2`o%9KN62QsMq_3dZly8L9xt^GTG~2%dm7*LbidcqdBqqItv}%Njxt*q -_0yhSMm7cMXrARJz#+e+l*bFAy{SW6wA(FU1TWsw7Rpw499`!g9lF0^v?_1rt_7fj*9ux&*%`=xZ4+X -(9@xg0^I2fy8Y$3@xTt$9IzBEv(XSHg1PC96@cF!Y)Mh8!VB00=HnH7Vh_>&V0v_W&LUC87hBIQb7yI -76HBz_EJ31d}=)M8>A+$lR{VbqDwrJiJ;qubOXi+b|KMJ}WdpkLjs2w8sy6&=G@|&Eo_vMq{$B&PXuQ -4&no?7tds{B|S@t-2zJQ6>~d-h_qo93Q$U5pmeT%In-M4BPrS?KZZ=_$&8hvG!I6_4$k -zQyz-J>ax`c^#RQRf19(#oSY*Cq79!Mk0p+~pElsOmQ9Yk$77MjDK!M2#)`ry<89EhQKRiaF^`+$dmn -%2@Vq&J-lAOV4XkD;+F4CwK1*>Ade)Xn=0@JnPI_^3dBob(u*D^1+GxDt1Y}_N8rDSdc62Ul8eA#nW1 -R4ac_Tw})h33t#<5(T4}6}{SaKw>zZDy|IuHK84gOy+nLMK=bEsmWSXYqwU*!?<`?ww$`lDYp<%Uimo -;!{)SO`T-~V1%!iDLL7suckI>(Y3}{6Z(8&YU>E59WYaI -?E=hofVlxM6@}(R+E!W@_R-kP3tgzY59K80zs%QibwdVROpQdf{Sd4sF`vhWxS0vwdW^DgIXJItaDFD -o;Jk4$F&tW#WsENg-p132DGz2Gep}XEA{jN=YCwf%cGL+LfGuq5bht{&vVVWyU{&o!uLjMCpACV(8r)@ZD4o26raP(l(Xmok`RqD;ih-8mU6<@=b)IpSke{E359ycT!xOF;AMVolpeEo$`uoNQ&_DUoyg2gL^P*`IFYb45Azu7 -q-j(s9|INRE7fIm7pmWQg#``BHuDev&=^akQ>SU|PPTm6A`!(N-Y~Bn$-cu1-I@)}c{GMhx?aX$h^UD -1~P;jw%Z{1 -h4WsvF@vZ;GiwDvoE~GVU2ZG6C4hY;sS}Cq4^(E!L?pj5dj@~taVZVe(@3Nr&ktMM3k}eqckTf#ljAI -XS2+dd5mm+^{^==q?CwxC9YT0`>AhOv9MK22@6-ikD`7)k8)sU)tBU1gC?8Dy`QDq=djG?*I46MsZTm -9)oz``(ih;JGLN@SrB+-EP>yvT>0UHvyJTHy{W})6;eDkAyqy=@yJG^rE+f7bN8J{sHN|>qdjAzi@t$z?;vR8J6X9>Q3cUZD;z`NM=h_nP+P8pJ!~Jq}PSxm~iuBVs-l?HG;G-LGf2Y@g`y -E=J>H_6II~u+;eHV?fx`CGs_}_+Wr$%po)-;a)LXF-jar7Pn-od|u-h^YSbu#WoyqW1Tc!RDSZ(KVyy -6boo-EY(A?oU8>vRhr3pnD>}3;vepm@L-vE(6}=Ly(St7H*vL+e_&yYX&jyU2%6>W^$W(tyqn4ErVVlHte9P!3fE5o*9d&sE5`R -jtma$Svi7pMuJwNjzL8mKBtDj$YWxoEwTITSzOCVWUdI`>isr_PwddKPqEyySb0M4)%5CyJT>V39t+OWVtl2gJlscp3p6ilaC*R6pavU4ESO1g1Y-Jn#KaW8DV -4AvNi1T$NsTnvSmdW3nlG$I>MK~FANMq7(=mj*rCbevf8>wq6ct1@H(w~% -s8yivvx>g?PWK3PvcTqvxycgAmRNX_)hD-R~{r^iPA2V2q!hNQldGAG=`-{XnbNgDhH!0$saPqAF_;| -q_u|!SyOZ)-W`=nqh;`imdlZ$Tjj{eGp1m)%^W_>gg-Q}ydv@_tzTaY|7E%6znCWaFJg=;%X8>j(T)Q -Dz&BBk^Rx~-ZvA6?eZM?KeYYg2?^pkt`ffu#|NYW+?w4Y9{^%9f`C&tyKbD}*ZkpfuAoYLsmJ)uqNxh -|zvBF`K;u$hoJ!Mh~pQoeD?k%Brv_{5mhF^+CQNP>K!9KVHid+) -o6w*vr<#-9(-Hbu<6hsC@q0&-0P2sbx{8ZyW+w7 -bs1`8$YkqBySu|9*;-fTvB3Ha7TM6U`q>S|Ec_pc^Sn#c@JUn3%5`|QN>?puSi-_Iwm|Jj^Ytt|Q-e= -BT7MjR4F6^-xL&}eHrtDQibC^_QWZPNOM#?Jk{8#zQ7$iDTpQx#Mbh@LaA-2q7l{_uqwF=0q5Xa=J)l7TuNP%Zl9NJ>&L4xEd -?10G9K9IbS@2iN$xKa7X8Z+mB4mV(%g8BBM%Ei-Rp+0H{AwM? -A$9BliZ4C`vep1G5O$wKt+aNm|0Jjseqe5mnAv=Ctt864YKbT-^iuHAA??ekPCnv8nAt%joa%I -Y||SdsFV$;Ecag>F}jvfKf@UJu!lEZgO6$vsoZ$f>daf67RSK}MEZv~`znO -_q^EXALr9uqD!&$+kpFATvMEY>8stU+QcDZDdwU>&_WnpAF^%6lmsqb8b7HxF@I|%hf3iGd{7G#S>BI-)=DhRyl+cqpv{Ui@0Q8>v0fNw>dfY1uVtp1 -~kLUwv@Q6MDKL6K~c>ZsE98B_42i@^PR(TmylzT`k_kP5-H1_7q6a5cYT6TQ7R7cVes>2AuN#b(FW4D -#Rt=r<2(lc})pwFb?aT%H%^jw}MTz(Och_;YLhC%ynGh?=>kdK=QANxw^SFhEhEMYzBC)9)1E$jZrkb -<~20WNh0T+%Ori;ja8dQ&z-;`0e<4WU0JciAxcO$%9=*2Oc(7wmzftW@IjN+^xjp1aH{i8MDE$~!+Q% -Bv-OZ&1T-#A#(gsqs`2tM%bmHXm^zjkPY8cprAzDyeZdV3zHX8fhJnHHG%H>n(NSnbx4VsSh%RMQA3GHPSkmKBW6K?-TY_AH{ -W8TN6B9*QX}db-TtdnvY_=A*KiVx(+FAt4lJ>x49Jm=F6Ao-%B%M{QJ`X^6!71e`rUH{OdCC?<*Sr>f -`w5H}LNs1OKu#{@oeJKP8#f?!FBF2FLk_c1`D>HO9XWY5cpvz`yG?{w*8lpV}1TAK*;MKYTOtFFnS;` -GouV&gx3t57T&92_7o*4^86XAFq&yFKIkHGKq)4<*AAFy%Z1s9R8iZ0{)%Ljq&e^3;)Z%|0({(&mZG4 -2HIkqMOwA`55PH-)o#y(@Ba6pXRR`>as2|_h1f|0n%l(J_&`3+;hyr*z8_XRd-1&?Ke=F3^S5UvX>nL8{2IBuv@e -AfZ2;V%<+K)@HPM`WrORC;D+{Hpl_Ym-AEUjj93BZc^8Oh3I`KXp^?4h7ydLh;Nm+LC>8Q^4 -RIq@6ljWhlRf4k+eii3Os)^@!Tjd%W&lxW-NtAaAwpN#gUx4-m+bxYQzoNS%N&dOdKudNKn4X@Ge8ke>uJLr^AkAOs#h>n__zR89A -nReoIX6o|iAJcZD6ovugm6W67%P3TgSSJNeLoGz_5=#tl{OD5JtYY{hqHpB0BR{NnxwD-gw -KCegMyhxKL8ef7QuA+2S$!hyBWHv}TWD+{$Zt`sD)O4sbUWe=~K;Ku4>yVA>Q1fMUXxoLqvh6Yno%{D -I)VWVyi0Rx8%a!Te-4`b6+|#d39-}7t-loa@wm7+`=PUialOX+c?Rw!7u#Dx$>)kd@Z{qV9^)9}Ali} -SYz0>xW_P#b*@4T03!#*a*^lpg7^e)+;ckR-p^)BUo)Vo%bwpXIohPmzkckA8Tae7xIWgGM^Yh0$^d^ -@^x6RtE*B&JW3;p(2#TO#&Eolkqg^;7?ALv<`poUlgduf-FxPk3h -h&GC$_hoC)FxlMu$(zz1ru8-M}71C(2H>sTal;kUtKY56JN7w{Ux_{=n>q*}k3_mL(zufuHLg#4hvvZ -Eb`vW#14UJ<}ZnmUL52|f-Yox%YbJ3lAZ&J#)gpbwr=<{2>*NXYA*3FhxG)B~b^lnQ<1h(ajoAypXeW -#-7y)CQXs|UQ(nQRPY_m~YZ&43A8(j((tJz%a~V+~leZ^I@fG}m0&nvXP}%vbqbuXA=5{y6INiR-y-i -}cThzLK887LlIPyb0K@qI&SJ?rJr12zokBt2u&Jw6;;tv)z#Pn@GFvYMMjVTiV>i8ufIvj(r{KXO(7h -`qOvse64@-QN%m6*ODyyCqc7i&85$w@cu? -dXUtAqFLPP>fXtpGJA66$K)NfHk5gy=FOd&hOg@%hMn1|(K6F}LMn2O1R`T&>*7Ux}N8GrS(Z(J=OM8 ->65pK8mp0;!r)I>h4ZCSmA_|RgYx6TI{G`@%EP$32UPS);{HQOgkrL=b2$nS;VcOv-SmSZJ;iIobQ8I -uvJmr+h8jkhJ}Z=n5*H^Y{W%T!!hM!UTU^?Drjr}B85PkY>=+;fP}Db`(!dPQ$#rG$fvaA0agw)(RX -Z_6Wlzl*=~-`)G&_<31Vwac)ZqYvc62Ex{pe34v%4z&%YiE_2h;|o6|qy3|{0)ekcL%t-#%HyS2c|_GWBE9a_)^`eeTE#^N -H$KeG)qO?^xFGH8!zU+$hD?SZ#{UcirUOK~MU_k00-2XA@b+CX$C)o(m4{CS=hS4u}|T!Yb_@HssGD~ -v;~7=N^5@%X=aM&RFdh4_C1_@4r9*;CYk=ELoiS>rGHIWgcB;M(E8_|Y~#47rQVXIa0cSxUdpo7C2k3 -BRXP%39xgFn){rT+cgIo&{Q-e?cDad9`$iv0{IEOy){YPo7_+<86ZH(h!_VE3M&TfhsGD@Mq9cLfc?x;v7(~B)&IDvGscY1LZeODBn -7<{P_Oq?+noT94+psE-!Ukgbkzle!n!c+Rd}!eX -&jwuhd=14y}7!gt`;W2j;3xmz&2|?_%xFxs&G?lAeB!=J09 -t?>SK1N;Htb`z-C-`^EQH$M+DPyr0#0Tw+)UWZ-vEB7RRrqnBpIXhiEA2SFcwjv(!)5ntCU<#&kzf8& -~*<&0!4-q}HGG++Hgbm -t4QurXFCtF;O?g+E)hXRG!MnEtV75uJBQ{SUl8@Ev#`G|CCeioJoaX}a?DqwiNe9($+x=+nk}SpJsvN -W9Ng#AFUFHPN~tbUf24x^~41x -)=lRNzU>rqX7|9ibD6|iw0+aLtrlz1f~;{r&EZL~Ho4ERHkn`zB>Sq5I}@*^Cf)7OaLjvGoP9>mlh?2 -%-tQ{$KCpsTwBJsqNt_oM0j(|=X!Wp0t4!2~_J;keMyqAXN^6-ZXQWr7*SoM|wAR{njk;y3HP}@qu?D --;8Jh!5bYmLb8c`ODquUj&=?2~A0cLZ8HQhR$2v;cq-R=Us&uVm=Y7bV?oxt?Bs0W#~9!pkqfI9lv3qW7}ou_*rKE?*5=C0l=)#Yd8(742@MAw!&P?3Nr -Wy%}uc-g*=EADcWbNOjivq3@+fW5(~M#`r=!p>q$`RU@7nTkS7snw?QKrue(Gv)+vScswfNbT9np&Lq -a?xYGN5!_l4O6K1ijR(gIQOYCt#d&2WNXtbv>Yw}sg%i4U^gtBOFJDpDjeBU;!&1uob7vlE*roA$~`7 -EsSPRH|>X<|QdZSSq|^ONG^_&o6X4!+NqwofQyXXfRq@6h_)XAJbY2XaUI9XiM_h5jR56U|c?27OmB) -?eAsu*E%}ZK1gXd1iLFB^vF4Zt-+`QiXmIoNB;P&6CfP0**Z@ot?$hCR$J9!Sxh)MrW$n@H~R+u*vRe -hTM=0ZC|8r88KNs4YWrB&1+bwZt-8FbNke$lfbDGv1ow?gZ5!40E|>TQ=TCXCzZqTb2wsO1{y;mI9|X -Z{G9XH7ALdO85{DE?k5Eioj<$~c2E^-;p^MS(Lxyt*FR?#f_^DxP7wGI<_zB -<(>7K5Kq=Te?q=yC2LE6{-A;yZY*Y@$CIOo>X*uJZ@pO_E6#PQ*&_SX|wCj4N)`LJ78&4Bzb --S4F5`G;d&6#(#on+}*$5rK6%s2Rs-gJTR6uR{5PbR*@Bxma?H}C80?&&xuN~91a&rvj_KoOv^59hcY -@Vug!6R?ciNjL3-}Mqtefg~(qxT0YuDPpFUimM=BF`4Wy|I*& -hx^4i+ztAv_}^|gSFvn>gXGE&V?-EM3J2=~(BztY4VEix_QZ3K -Chxb{^ZV$|(YXCc&(T?=2DtB?6!v+87?kLLnli+3tt*Edd-yMc@=HAUQePK) -%Dsr{9OA&I&=2M? -o^z$|sZ9G4lU*%?PkDT;@U01-&%mw@Judo{PueH$Yh*l+ir`b9W;ELAABrCHQy(INHGXcx>onIuJ1dp -qx6pib%Yu38mP(7X$_W{%f_!*!e_PA;nGLw7cV684BxNPq)gs7YVFmnvii^=6W4p=QjF$4S7q)_QgZ8 -V@Wq9RD(v7_tdK-EyD#qEsgCk}Ze-5*;pXm8<8%g{1P`gUDahm~lp$3~|>@L9G$YSx!AlhJ}1>d!&~w1W&1h{D+nclh>cH-+Rrm(xMmozeWwv^B$i@QJrVU%R)11rgYQ)j7p@jPI9 -PkrKE(XD$gR@yX!FSxiYI-k|->T)e8uCA*<)?j;eX^C)c@yZ2`0H@nyV0IY@jSkd>Qd(jMl?G!3Oc?E -+MGoUI0`xbJ$_->jYXjA%b+X8Q-2&5e$S=OgnNe0EMvYMiuIL_W|K0qU5nFpYj#)6)0`1{{CClvOUI% -U(ssQUofOBwKJ)WJNtdl`Yqh<3)~=JtKb*L4|*dUu9X=$HHWA|0szL!I`Lo=;^u+F*wU -K3u+&0-xo#TI%@@d?R`tp2$_fyo3 -zO_S!?oMWh=?o=RVx>f1qNh!=wVs>~AL4wpXTUTo^kv##Jpu0uL5J_#VmAGL+M_~`Z=~5v^!O&p7-LC -umKft6I^S18JI=uS4O)ErD&+43-v!An#o{|r$IbX<%B%-|;Zr`2yMy -DN1iYJUF+1TIm*)eMbcU}hfg>y#YpeY6G>?-5&3fAo=|6xN+(Q6oLL?#pIps4v`U5qpWz+E#ZtJA6U#> -x_1;RK~|>JOwiJ*(4TmSP*CDv&I(4d)0mxSWqtWxnX61QNY_^`^X+D3-L&mjpcqeL!7DD2f3yFj}KYJ -d7VDQ1KYKG+WWRjyQjU*tlIrKQ?@71g1V&nstfU>c_e+fzLX~XRWGia@S}L664zS%NKgE@wx@~xfe$4 -a%1bhomt-g}DOMis>D-m>tL{!0`}FtW`ck@>qtFGI-D#rFrxNdK@uPkP|M>Z$;=CU4*Tdi=?3T-3{%i -(~$09BSed5zmTTW>{2fBVwJ1;cR*g$}>-ml{KHT=Gg-y`_#z^@IzZkhG=;79U;c5*MpfqHuwD1aZ6w({f84pN#(nLL{sh6a(Vq5-ueu8G8 -YJsVW#OR?MY6ph%F*|IO+4>ul=nB9Lhy^|ocSYeT+bQW_0PfvABEkcGc|H(Z9;Z$3GX|Ed`D&0cp;~A -=ACBVHaQfN7$3I+MjM@9_13Wx*vwKoujo}>Mf;rr0;~YS;T^k=Rp@bcXq{tQjG_A0mdruxBGKi{fu=df8 -U34T_&1O#fqFh)@aVH;(JN2lvyFww+{Iz4du-<(2vTTiiXBADBU4Cn@7Z`lfLi8x4ZD|80g~G-V;5hn -hQ1Y{Lc4w(Ake`WIk@Uqe9!$Sg{`HD8qX?+os_2@~)xjt`N<`D8RETm9^$U-mXp+wwT(0J5t5j8pgOU -c8-T0pO5>Vp3Qbv*p(#q{T>GWPbjf=rTxD}Y+gh21f7V(p$jdDN4C1e9_6nfuJGe|nC9>V`JOR*mxfh -3ucVvi?o@&gZTSA`Hiqz03Q|9c9y=3`)c(BvmGWS|7PmjAwMP#4Yx|W6d#xXtfqGH=PJ1P8otE87Fi- -IK!NJ?F8OVz^gYPLYG+o$-O;0g?h7Ro$eG<==e=mh;O_aX(CE0_Ii9IqsXQ`h9K4>(B$2cMS{$2|2*L -RR_YNfLW$?l+i^Hcr4s*VK^x -7)&cD4wshuAxCzfMnZ-H2XxF+Y#Oh&`(|0=vbHxxKO1L%T>WD9n`8(u`2K79K2v`O*=lp -e&efx{U{W-{?*rmY;2!;a7~VE~i$#22V|*?HmA4~B$owNImpUiLQ2vfs`NZF!Pf_;WiOPNyW$%cUy-_ -Q>F+tf=_)j6SFDD@rBvaI7OUg0u9a4_)b2h=>N(I{mzW~nOM=UWg)L*%gd1Noa|^F()y9kp)AB_h$WOziN=5UzCBduL -XF=`glzNvun(jd&LWKSBNFwWH65e=v#IDrec_G}t$Xj%=(OI_+Cls8PwYoa;eCJ})cawJt<^8_bC@{o ->hSK5cC9bYB4|f&hd?{!MwV&2k<}#Xmm_$~kax94$8wF1{r1Up{9`1#GoFsGBVBt^PU~w4=t%u7Mmiq -13px(iV|4t3{m-N0HOD6RLvlM3P$QYNpDx)D$WAiv@An;<%4TeXE@r`IY=+GcvXd41nqBn4!uRN2!)- ->EM~Cw&XkfG*8=Xem;RxmBi@F5?!;3mAupNwv^_}$nal3v#Q=L`qR_s|;h{T09+9Gn_vTM -@PAbLbbUrG{#Ww(ZGhoAK?e1F>i=!xpqJAw(N3h4oM-uLTt?$(>&L4V48%N8r3%?_o@xA-QrJUyR=Tu -D|XE=}8r7QpD(xD_)M0+z)EF;c!@AOmu1o;Svm%>4r74uR&=l0lm-Ku!})Zam|lA -m+}aTC=|voXA_qu7byv+^Xq_KnVEru7B<-JYbtLo(z>EAJ5KL2#tOl)!deUG*t}hogLaYOj=+D&9${f -v?cJq*xxBFW!}JDfQb_zSophVx{r6;VKi^48{N}m&W$!EE;;hcH^f8VtV-8_&J%!zakf`F$sK>a&hK8 -)6T4uJBO%EfAvbCJUL=c@dY8g`b7b44m6eEt(3CtFJ -AxUc2wB1?yI;1kzCXW?%S&ou=rQ6JhjYxr?f!HJD5aQt!jukd3BZe@Wc(%~GwCa{qD(sesPFl8p{8@H -~`^DCD)w{w7eUQmQ(b9 -OYX3)1vQHpgljR?N>nG@7Mb0bUio)IQ~7cxPLw3e$oeud94fkJ3NU0AJWQck;U2P1lOyr@^vl(SGPM~6p!r)6m ->#c_`;l9B^-?(8%=|FvnuXtLFJu~bm8h&fkV^u`&8<$W95x4{Cw~kOXz -%B(#arTyf8m`SLJK0G?mX)72imvsZXGC2Fsx5B*Qm7&uZlP4BuDw5kCI9p5OcVm{&7w$ac27+Krz}^3 -OEVh0DO6m<({fhnbM6{cPg<=S3f)alD1UueSsyJh#N2C&$C4P*0|?$In<<`##cP+V8pL8t9uNgiNRDJ -WhW;wIjk;IJDKq&;6j7ocfpdAL(dcx}opO1^>zCX&ji;pFsBE?&+*u=~^CMfwTpDZL9OXDrx)7<|MY5c)n2krmMu>v66F%(q-avifMlF3!q8m8EbgJWc9@RFPz?ADi6o%qs`Ybi@t_FLs?B)S&F7Jw)C -Vx<=AS!I%f4B()!xrn|LPi7xpYs`x-w_rOYb)=~-<29LI2-RUhZS;`1QGi!?Y*1~|7*31{*;+Y7LNLo -%IpsYVdfb-x1orI^%6kgsb>c_C;~BdxGScP%3OgJ-K6xm~;eFs{d>)R8?FR@)_I_h*hRZ-X43v8ui!? -v}O=tCaaKHnZ=@oznJ>eRAf=T<}#!3?#;UvqLZC2|u3NzF#P;ozAx1oI!mkh~XvCKhSST^BmGRl3DlH -re(IK@cG}4&{wBrwNgCxrfFKJOPV(BhY?;w@WWtRyVoObI`t(s+AJx&FP`!ap^e{;df8xWSHR9P-nOI -8i(cqhiMsVmEJEq%JEf~anv>MOaqXcFJHGeju>jg){(XU#meLU{D_~IC1xQz^rLkK?8b8uhN(zS+T-4 -X0r?FdEMC-Sh^8E8l&GV*l8ya(`b4%B2XEW2eoxcM=Z$r5a8b2Gr&#wW8I~V;*@H2@WE=y%k)#AA_iP -c6B%U%HeM?u?Nuxsu4>@dwqrExYtw<|M7;49Z!k@s7$C*HN?k)_D1x0}>v(%J$lbAcqvqBT@Dq^JJz4 -l{$-L$_(pZUZZ^^C=z}|8{NV9E2G{+i#6ZuJ%nvI*yZCs2mbsD|-zMTr6f5+`FwcFIT)45B -Eiv<-$;AWhsGG1=796Ww-V!8b6&TC?R8kIeAk;)eRwH$VfEx>)}GiKi2#LW}nb$ms{W85eHbyV!nGjY -C*NdM4#6TcblH9`hU{`-izk(ws)#{nBdo)yeaa1j_~IjWu}^Lydh>GwXsrnG@%*ikO4zDoyI7l)q>UHp -aLy!vW311%RMwuEq7t6n@i7zWB+Y-x*ZlTG^Hk|vV!iGA>ENV!s-+X>sScO8vsj6D@;uej3vz7yx@$1 -;Jk{~@*YkLPW;d1f!90Er`)Q@1JY30?;yU)Bm+FwG2(r}6(w4a-R@{sC3fo&b>&RY^P8P$zZ9@KRl=H@Pw3 -o>&P-yBc{S@rrDcG()=*RY}SdHzs4O?Ddxy3&~yVEW8mc9;~x&dj3hMhlVjYG0?$%PF6OuDDOLmJ!a{ -Go_l7a~s~@=*FAd?Va81D|~rtg#;U`IJcs`75+Dg64~PX8r?g75TvaFUaD(Hvi23Ptqzr?szfw-v39o -s!CsrMg6prIG5?GD6iXtDbmwoR++{RxsEO#7VD3-@e2oSs`n6!Ycp -I>o&zZ{h&>5SR+FbY{6Dz#Npg*5uY#C%NgZM-6cS=g}mBHW3$NH}@Z3gXA^AYIQ+otT5Q_X*h&j;U(j -~gW%N1Q_csPBq=2s-a`v!;I|5^Mjp%*S2A-v%u|BQeJ}$R;Mxzq{a*e+&6(%zCLwZF1WgpKC!dGv}*Y -$mjKg2ee1v2LVe(d(j15gS~Ei7Bc7kAzKxswP})dU5i*t#`960WB1RxkML-5Aoe)G8V^px^(1RtL+b? -~hxbz3e!R1~S+=e}o`P>%XEZ*Zifd?wNV9dus!vk$cDw;>q+%dm#F1`vr|B{ttWa9v@Y8u8*%hlgnI^K<*)^L*SsvAbNlV5_ -`!Gs13x{5~9`id`pmHog{kBiJ&DaCO|I$cz -yaQUgJm=O4g{MWp)7Ol1?#JVtb1lXh(LCmyW9*+C_xFdx=iD74e9o=HIhP%A&i!bdb8c+=jF#>a_0wqocOBO18pXeC)l` -lL+xwQq)Y&`mdq4KO?ro7d>|4-|CslZ-)OdEs(xZLg5Sy_!yY@+cqBv=fZ`&Yqrn%EK;s0|;`sBC+k> -wsY$~6fu%f!l;^w)^;!e|##qc@swKNMNcXje3a%SE66=Uk(FJKa0eh2QYOa5p-X-xTfC>A&bo -6FJe&bXm@QZCqgji38$D(!dg*EeMkU%P2P-WO|$-btDI#>w>Sy7!wy1=A)=o$`Q1%q2W#{#E)DZXERu -`1WE%KRrg@3DPDx%k)unXBfUdd|vifWsgdIV7F2JHxX^z@pq{&+(@0XFb9Iv^qCE1}UzHy -g?N1+yXqRnXqGZ7hw!KRCeB7ET?>J~{B=1br-PVRkpQ;sT;{;l$zUF%&ojpA6X-=7uSjF!2USzf@#hzpUKpN#|#@cSAmj&@-YURcF -An+1wPlh=)dW3{Lo@)r5?Y-k~Me=bEgjc7Mwv{s9Rbi>U(gGbS4OIX_1g|J8LYW!%aJYK?-wh7eJ@Ggh;TD0RPoN52WdommCH;lP)UN(DEC2?)3W~=2L -#=wG}lBESox7{h_4N;)%_9Oir|9M`y=92dzgm1YuT&LQSQ6cen9KG3pcfFHF);lR;y>Z5T+%P -{>_0}7>-o4-_Ag4#JcLUe$no;W=?uYB$Y^-;Ac)c4+F0kJ7qt?4RUadD{JehOjydu{-FpcpwA%CJ@+F -5s@-@s*}fqk(~Ki1EUIt%+14)kB&Q!)d;Yu+(&B54cPKL3O$N#k6`!HGNrC#a8PpGi?Tkrc*>h7}Hl6 -Vpyd;zZJga6;DcSd{J2lm^$53*pAc>8I)2HUu1?UGiCLobQCPy^ua}%Z2i71Ma+31l-90?quES6kOUH<<5;sQ++)!=`*^9^ggqU`bm@{ZP;RKeK7Q>D#Pw)CrkR%6?vlkYCizK^%7_``*Ml&IqEejx>oWnub -fYu*LlCT#Fff>^`^?gS$G-K -YF=v)s2Ce!^vvU(!n$cvpM&|shQD^YdA{D!UdGtWX2ak!-yuqR<-oUQeT -Wc9#7Uc(V#+0JYI`Giao`g7tF)svO<>TBe$GPdzvgJ8hh;uSm$-HvBJAtFj?-vbISB4tT!ykqB_)iTu -Gj||=XBhSeu5J~3KEv=|x$h%I+gO(OZKs8@^>!HdEUeS|k=mb|BHF+DFD9+}-N&5c(fV5zr%KL2Q+6) -Y$I7++h2GFfp6E(Fh%>S~gK;;+w<2nKWp!%+#|qKlisua>}? -h=Y0o6!MbmfmA<8ba4RP7Tlv!a(l@&ECT(e1yL4z_rU --VTJlBW&Y8U>FU~kOu_&BkeV~?PATiS&{hj^ -*?4c>wQZ%o@gk9SSC$C6>XG5e`6NPS3t!`(xvCOv_jfJO+2w66sn4O{~KkmO1r**m44Da?Ji9lTnCzAyHDFvfVN*jKU-6-k#X(k$k_M5{k -%kp`}Aq&9fOnyEMgbGW#C)jw2a?DyG!4;NjTH+%r;oQ8Mw+iT%(@zp>FE(9yu32<6_0ch=kuUtoKxl= -$1N{aO#f)Oje)#aUaVt>K4jDj(OB -pIWD-+>bq#+ISVtIq=3^7Mp*^c?I`40|zI{`{uE>$C3`9AI86NppPlo8?_ks!n}BIK5bz&*SK*k@!_S -z==-oH@6F|SIp@qZ@!o8-g)uGe0<5bmv$t^m`%o6+Tf86Rn-dBZnRpQ^-sHIBDvKW1_ -Ft*IK-jz|%3-2>FsB$Ud}Rcp3Yu)uIP$m2cmp2+f!>&~lkH+ZNPJ(D0C+3@qu1a$z^`_%w~Y0nT{vM5F -U0pl>+d>b^7|7#-b!dPC$QotHHHRtm)jCvmzuOZr=g)p?tN|yh%vSev1zhI*9HbK_DUFmCmb2MW<5q8 -Mb@oh!_wBaB;L9=oq=^FE#OgH?UBk6_gZ=0R}H)wn9UiQVkxY+1R##D*&M>hM2wr#o4dC1r&koTnS#G -*M&9|gDUH_rk2#_C=$1-t^P)^CGlnM1*IjYF*;=0zKokIpNY -rfXwiYWjG~yBis|B{9N=e0)D3+23BHzdi|93CAYR^ZbZkYD9k-Mt|vU`j#Y^{q-e|vY{W(Z$_=3Q}vh -U9O)0ZSNXe7)8DoW^*8$bRWNmq=)o0ionH3agMJ@EzvOd@M)f-qQ$Wh<^mR!z``(*q_!6jZX8#`-7mW -813f|7q_}iz(qC>u2XN={B7-La)1_~cXjc3`&c-$skitUH<3HR5sq|xJBlQgownk-}Y*Nx}@_xD%2xx -dQ(H}+QyeAuhzcxq$~k`z8LCYynWg)#7uc|9>+hMn2Mzwg-fQkN#;(9h{F1nsY{7+Q$4m-1-W1QE>Sg%sdVS7Nce*b5LFT;Y_#Pcrdb8*l>mO -s!9^7>#)K*JBJ;TnoAWn1_m@%R;ZsrWgj~8i_X($ty#^>J2T$y&uO>;dCS*N6*x=_ZZ(<0(a$7{@Q#r -Woo4av9)GJX$r0_T}$L}W8dGLtxYZxO>$n|;3@YgWr;!?Kx*A7dWcSAa4sQt9j? -fZfM?i5?=PRP{s``gO9OtjsJweium9C&Sym}|KdD>ma>9d${u;wTw>ob<0Bi3w}BNSoYn4BOFhWkoExGAh5^CL`tx=R3imMT}p{*p4Ip%y -i0JOVWNCWA1ewQ&u@Ub02>S|ia^^RsDL6PCk@IT2U -vv?Mqxe4APdFws<6v@cZ$Mkb8PFBRhYre^T%>WS!tcryl$n{6N9~_Uie7bnq-?|G9x~dOb}#pdE?bW9 -+rBKku8FAMqYb$-P(Q-&D2AN6!Dj5K0_>^GWy<$7KUo}n0{e^iT#Q}66SQ{;&%Oxw9cEw9nC%i7X9-)?iQ<<$8v2bN4Y?ALLQ`_3FPRlpxxyEb1Fv`ufSoFEv}NcnMHO+S77T -_fWnFg|dK`)%t4+Mg7}$>N0m<2^|+`P{GY1T^N!qAqRG2U*Fvp+nL||l$s&qY -ST35)JJ1!DNwcM0m?ihlZ3*-h%bh(Nzc1sR0{(v^K@?p|yU+yNyjq)>!|`s46LWZ%{{z8xCeOaom`ev -}4|7zz7Hjp4-{xw;e~#~ani`j1lRm7jgPPPsyZW^HKH893vhU;>xx$h)o9*-`cI5JGZA?hJQq^h>EJ95c4t`OC8?S>9zg(X9NBaBN?uXAEYd$N|?+W -%pwsM8MKP>A|&t+}0d>0IyTEA&+^W4(Vte_OmSw~4~EGvPa$=YVT{pu4mSiMH=h{8qB}7AePP_#R0Uo -2Q%)6}99FFKxo4oy5#=AB=yIHyo-ew9EW?W5yuc(^9&auOYtgyuMFQ8 -M#@Z>ZMQTXvYTbrq!{{#4BzKbraBlxnTfOaqh0_oYObx3{nF=Vy~W;I9x+baP^jqCa#8Z^h~0|0kHWD -)dyRWb`oLcbTHQBI_ygv`gN1k=QJ-pP$D#%THYIA9NXK^N9 -)m@VL#x0U@K{wb=Tpgs}4&%*r8 -h%WsACGOK?k@+-UjY6!E)qfVB$Ga-|LA$Wb#}y$KS(`Lp2ipudhjPkJcIa8+1}x^VK_|zoSHBn+egYb -$@Wp$HwkkqlJhB3en~p!^L^B1pMkTZ`y9UaX+_8B^94K$xP-j>2@v=24*!Gnt?iON`f&vkZfh=+c|HQ -*AsN%G|1$c&H&r_Gh5tBx#;z3f{eQ_`Hr>m(Q^$^j|JtN{%D~G88K=@5UqzSI-;dw<*Qob3)Ri@84Xk -JXwpIQI4GjJV_QUZV)In?BzMZPAZPEvBte)TCf4UQu9^ZHUSrxZ^w4ap0N21N%9hR<6 -5>b8=_QVPq8ztTM`_x$RWJb%U9_hC}I^H@M)c3b;obRU$#4~R7IVD>&P6^|CGOq~a%|o3GpNe*`(a~FmByU@6W`M0-E -O+?i^iOcb`C8(r}$enr*xm)n9~nXXQ#1Nk?U0Y!PxjnpJbiVKfyX(8nsUIW31D~#++`8Sf`w*b;^iZr -~O(v745Xkn34 -?@A{^WMG?lJ9wu!QU5*ZmuRi4kg+Fa?y?hafG1~cVa7ts1FRP(iOsA-yv|3xJ -M3Ciw<&nn%NU{)7;h_FpKugt(`<)v=(e5>6_tKC+BWbY^8sM~uD&?K?|dx(#@OrIq+hBD -|KE)G!yTU{8X)tX;2QkQp`YcN&^ExxJ0at<%#(6~jQ=h3=LPQ7f6Y7qj9D@H!!U3E!Xf3iJ43EnoUa$ ->QOhgp{3^hjaX$1-*k-LTkH2*WDb -H8Kg-q6c8!W-4SN4vsO{9rFr0M!4;(lAEz+aos{d{jyjEficga|iqJ*-$%8M-g)T1!Y8*RGY!WnGFdg -D6L7M^`G|3I%W?6E%z6`AuP411Lg&M5lo3TX9v&xXdG$CojW=Qt0;Pi5sH<RJnR~Fomd8X8`uK=+?wk -SSw)b#gg2}i(x&yeHOB`z#;;<6r^}oHzfv(T3pt*jqOEO7>K@|c49KSGqI=&ZA{~5frMqVGl8Q1d)hR -@YD@l}WWWQqQiOzUs-_9}6256r}ROVhR(igfWFPVF>NF8cyI@EyoOAD_1U~&RzC+2Z4%2S8BAKzRr>K -_!RcC=HL&reWoeG}zr8@-abjxc^1ixYdywpz&t`C3m&zE!nhe^KAKFhO)%*tXVth`P!<%-n~Xwtfb_d -!|3zPG7MZdqrw3{x5w)`FxdYSx&}-ET1SA(?-zpIgu`71%Gt)VcH089}dmE#X2W$GT*W+B8?|hSLmlXE-+)M^tkTS!)__lqrfk7ICyo|F#!2;@a20d -xFUmcV;not=I(+xpP}_?6g{=Xc8;{1*7FyJo*G*nhm-gd!JJy5k5cap*g_k&B*K!WnthNKCnmFO!*4) -nj9@Oi6ygAnsFG`kfn{Z+a>I_!VpiHBdG|2fZ^=cI~s8Qzp2Qn}yGV810!hHRg%```E)^Y{t(aeQBk@2*{DXJCkL$UDLGqr -Aw*dO#K4p!UG=#ZmZvP-mFD^c)=!NvPi;*X=&zdNCkXH_e+U8nho=`vM-jBJvwT#wABS&yK@SQRLJ@j -KVV$kMmjsuo>`b#-{I{sUFz<;9tAZYwk#Mxcyex)6-X_YY*q!wM&Gsbcqw%&1Pg+ljO-puq(nLwAQ^?$SHqle!+N-}uHiM#rBD^Ps)bmn~lU12X1W8}Q?6SZ5z#k^?&PP4H>ufQ?Jbs$;CfQER`~-C7 -PDUs^M#I0rP!{KoN{w4XGS@Fi`!+$#CiA`|}PFSrJ}^0BRKW6o{-twbGpFJvJ;rAOgY`T>3?PW2jjH1 -7b6R*^=d&dO)=Tj|Hm-0#h@enwRN3zbwD(Rj>N{Kh;z_)nOd*_ZP3$ngCOWKBs6@%_>fS~v-`FvlHP?sw;v95tT3u*s -tpV2=5z!aIO@H458cFW|CWNcpm)Nej!c{iFkmH!gu3^SJ9GQB3?>9w!bqqYd&iGaFn>zJ$5`JLpNjcB -S`WdrIr&4Ru3|w=!VFn-0g=PCCH;0^bGQ#eSZFvH{eg+-5EYyTPwiOnfh9bYtFn!u)muPepuRFV&2OP*w8NW1@NtQ@QREf#`jGIkJ&uD;_!eq-wk;;ty -*)3{GXA!));jYRo!>6mnffZ -1>InbGtvnc=)^IUmHJcEdjb0#a;9J90sxJ`dr2QIQ}ls)uPLDU7A@1+4EnGJ^x@yL!~6|>Y2UzD;<)G -6nR{O8C6=Op=O>)!hT%|~;_*h?u#CztNO}mI`N~UqY&F4LMre7Rmrw&>Yy`F~Kha4ID#$=fjGkFWooBbBMH~$Ts<-bSD)G5dNrs;^%}^!2k(+E*X?vJTDV9T@NS`lfN+g!i++3tHb6@(R%7x1NcJAx18Nbau#|r`LAEk`8`53FO9BmKtrkOK2K@_s=wMLtg*P#zW*=1hp+1P -)n@lA_2zAd}V>!TgliR;66lcrt7k1$@%R=drr>|`pw+*BFs&Qj4=>Sos-dh~3lt;x(Ua8likaZEPGS- -(?wF-93PQRST${+g-E&yw$%Z;s`I5jmd5v!~&FNry+-wM@55yB4VjIBnRqe0dzZ7WB*e)mZT^)E!Hif -1i;@vHZO-Z4bVvavk#inRyhcb6d(Y_I#*n{qb;4MQcP(#aQ>sr;MCSW*#V>lgv9hGor7*8+}cqAEwb4 -6aRKH$284115~M!#Qi1Gh~xpGVgK3ypN=c&7>8HndB -wY-ETs^1TFXRES!qD_oDpLOVs=izWYC)f2{bqN^fSBa<7p)(@mbiw8eb~^I8U)%JJU&41My_>(b*yYW -mhqxy)H%=)ewz?P$~8Mm}ISbvnTh%`<3hOjuKoYy&;)2i*)z6Afm5RMJt#IU)>ffPF9Zw&01@N}p5lm -~H>1Ugt&nVS7W}D$upLGH+lD@T;b-m~@ylQ1S!xtN4i?Y|%vbbgbP!Duj1L=Q@Qlh28g5Nb#5KgK+J= -Om7H0XTWvBSbcBIc%RDuVb^UD>y-1N>8}{34^g&UVVb~m_IxO8Ul*?}Azh#?Uo5-70mfpI@5QM1(Rnh -L8T9((#VQXypGnhR?+dlvGFA7}*6=Fugf^^cHRv4kR&{EU=dWiBFWN-TNNvs?&klq(Apb?(w9&6C-gW -1iIMruQpJA74QNMrFr_SzPP5s{8r+i`gysOX1oedgapc(nIadsCN`LoNXCv73k>SC;X&F;1J=_Ly%a( -s4&%Wjdfe|U_ON5f<)WoKJG?|LF_3klni-+8tr9;pvqTj>GKWe(3;nfF;NCjE!}?$tGU=M#Lu(3dwE| -6M{GyhYssN}DPE?rl?N>uCNwESH=#;s+J-yxFev&Z@B5p$F%DfOBs<4pLE24-ty&lWx -@|PLqab_MG8dCRjybr8)(AKddm+MHLlm7!a?>#76lOlrD>CoO7JWd17486KZ?-9d*aTy;_6u?XlPgYp~vtV?2LCjdAPQ#>0OaXN+Bzv124J_OA}+$FaMk$JL5)F+M-~*tO -8`2m3Q)kulz7S~!OCn7m-1^JgZE&iQdttjD#+np}2)HTkhIR?_I&Pq8L?#G07nj2_>$G1lg_7;AF|v~`rf=!N4<_^6K^LYpb?g -lX)a14@Ri=f1Pbob!kFVc&)tEnS&BpCI3qTSfPh_@>HJiRy0oUC25y1b}(oo&**iL -IrVJJ1-4n{UY0g?Hs7&z(tcOQ*H{@RiuQe9AMy-`cAfmn{H`ub`hoo61)DNJEAlU<-+j8zKD^*?qn0` -N&bpTNNoX?(e~;oUs{OV;G@h>4rcG*n0BuadUkBsUgO6fPzdq|EpDAw<9#0s#)TWS!LI3>!=ah_A<{o -0q^E1|5UoY1EFBlhd)0GSMXNl_1ascb%U1-fd5Wv1TjydL`f9B@ee6h@rS41A>n{|alo2OztXg6ESZe -24Rs>-s4^R~sG;F!{VotCU{411J4wQhZTw(>{r?)i>BbPi|DnFR11_{JQz(*OT`)7N5rzy4S{pU?l!w -4aHcM^wd>=bW>AK|>cpo4C>Q&yxAw0@gxb2z;C!XN0Niilr|LSTb63q~0{g$9!H}-dBEA;EB7H4r3B{ -pB#DjcQy0GCx{}=nsMMb;CT$=*oyj}$NMRGU$b5B;yoy3oN9}-$=k&oYfWhP+warnPRS>~#533U(OT`(ixpZGU`09CT?`8O^05_M&ko&=o@}>PM1_kq}(#{O~x9T|OKC*u^e9Qy%2 -=6{4+U}%&6>X!Bxdn#oYL{^*G7f|q6T4`u9BQ<6Q6~WU5M*1S<(86QVsrbg;OPR1${xX%ta2gC9GS}v -e$hIbU!00^*D=+s4>1PG)DwOJG7GNFNWb$O8D(n@aqHoxW{Eqt23dZ=cnpjcF^V-fKABal(4{dRNH~&^HmwFY4>|-KVpw~_dXSB3r-{bS6 -BzH^}}#bemU;W4t@Gp%EJjDZ+D#_1^gG}0x)Xc^!eT4z5f$CRhWM3`0Gryt%vg?xhMHRB0IkNbrIDc4 -mND8JG}Czn_xtvI>F+7$w=3r*8XzV^a@k4}`n|39P?(uP;_-W!rKa}aa%Uhx#~MudM(b|JbiDcXZo7X4(1D!x$S`uc;C2sO(*l-?8Unc11bi{{QY6uc;-wQFUN*iW{M_l#y3RT -w7>qY8Wa6z88esdc&?!zr3khdxmYUjyTT$a^5EuP96 -Qx?W4EC!Mw{|LVttmh4S1pZ6xoSn%?BO_DdPmmRrhcK0efu|vND&r3PDRCD+lr;hRXcn=KT(T~0GQhu -IY>QS*@XA$Q{eS_>AX`d_0rHNVFBHAFomkyjAPq$>|T#lvZCGFXCi;U&#BYqGce(9I?vKyR$zX!O~F< -%UIq-ntdeD8Qx3^}y)107kAv1?Wo1J;W@S)!Qh*z%1yuLro|)a(ZwDC?=IVczq_%-um>w`jeq+}}>cY -t;C;zb`sSy-an2?$5{mVs6782Vk`jqAwd^7Xh-Nx^DzbvgR_{NS -nZ^$bAarjN21M|e0`=H;iqyp?wWiz}M?0s~2aNt7{U}s~b@rYD9 -XjcZ_8T(kTtx2536`$87l|cOw+p(r#E924oxZ*4%-bs5wri;O(f#15>rC5j`c8>!nhte*sQCSpIMYvo -ezA__8sItVT=Zn9jk?`E7WzVExHlSdVn>3QwFYyn?hnhYl-syRo_r@P`$WcVkcU-efP%+sJ%Porcd2^EE~Li`0XEy76p?HmZ -`g`q0)w%@TZZlzs4?QF`x=ZSY@oW~KLhDaTR{Y{>gqx>-*Yfke}@5lZgOUD^#! -rB8FiO&UP7klMcv`JpirK0PBlcBa_>K-j^`Fyq4d9Tg&Xim#7iH_}M_BH>cq?3F!KuI&{No{cz{~&Mo$uyLXpr8FjkDi0htd#B~QS=lHf^`{BK-w4{>n2Y~ci3XfM{zlO7#w_Lk*MYGxm%}I;Yn66J(#K+JyePa&o)4KL=fk -C#3-3}_8g0|&K#K#fG+8g{pA33zVH^RyBmlZQL(@yZ57d>**vQU`TFk|Xzs^Oq#dY*$(=M6axya3PBD -Ie4J!o(~XfWlx0$Hx6xOCWB&zNzYiwcUD0|q=FQZa6psYd^%JO~-szn}4EW!~3Cep{lv>r7+3-fzk|W -RJ|@)RXFrODxOtk`{g+@BV=Qf5Q5A&Q91em~3mU#n@|k4~V&NPS|^O)U|&RWM -M?rQL?YMCxp5l&vi?0czPqovoK;j1>6%RpF@7m92>{8Do5TKa&FXb1k%4t+vWN<)4$VR -1NeRc^f1Ky1y-9+joUsl?pSf`WSKQ#UQ)&uN)X+k-3^1-+vJh@txliqb0S^p+h+ExQu;Qhj5A96gt~Y -~t4Grt_VM0J%RFGW3-3I9J8sQ(N6VF_UjuRDm!OSVko}HWNW-&yPko^L!x&?|8|MY>Zp`}hv0!_V^?$ -1F`wG*Ic&Qmu7n1IK8M3JPoBB#Q$8B~)UeEMJ`iRB0D;|CRMZ<_=TnTxV{Z2FFZ`$yFdRS#hSQX(L?^ -N#jK>7TwFxn60QE-?0t;oCSrZgY@91^9Rjy1{hl -}{0yr4Gs}PQ<0(c(yuSR6Pxvy7C8lFgLNMzV|;+dexio%=%RmfVUf^?6=D-Hyz)$ZdB#IglB% -YrKtX1PFDF1j7bT{2|*U&$s-+RD!G~*e(UUy4MW-HFYAmhL?w~PEuxiu8p70yeW;kzO(oO9NNHZ -ss2Z4rnYKfI@ChQvtn1rZ -=5&tqQ7roYn_=H9#YboQ-az}5U-*1%K*krAI=FRjj^(-kJYe$N3(s)mC9dxv%4!zDbq)E}r@QL+p*0Q% -L`_4om`FYLbcD*rn-y{u$W1_|Kaf!8 -OZg|G<6OXWe;8}H8*-{;fzN7b8WZbet11bk>clr19A;z-XCv&s{|M2|Ves$hg#ffg-i|q%F{g8LB(ze -__%C`Il&3<5Of}%l>p)Bz&S -O3NOFg2y3g6}8U2gbYCEs0wcbA0UeUa~`;@#B -nyK=s}6z?t#zgxg}bMbC&_}%CE?kc>yD*Ud5?>>)rpAWw)Fa_&D(1}&$oL2hG*`R> -SyCy6idOS;*d+3$>^-Z0SN9<)n#=e=drlWl;Xpu`yJVBY<_F3k%5QEpzhW6-vzn -(CB_2CCO&J|Z4z`A-N3v@UC_{Z{GDt%AMI@I5FKD|Zy)~}c^yk@NKlW3Q+j~Vav2WUt6VWj)aapuWy8 -Sb;^RQwBhh8ppyj?rJkI@$2wp5!*i9ozx -YJZkwk2sh_RBN7|Nm+l20Kradp@(rVPblexO?9{ufgtejmnv&sr7Bl1Sa%xS!A8SI>W7UDWv<|3glk4L+M=%?9l5-X7;&2|Dx&b-xzEPx-7jSw@d%*ITq>(}O$lzj^J~h -xU;Mu3b9RWRwAp7CnM*N2Utz=Cv{x$l?cwRV>Jd-cqzAS|_nplRAv|ZEI%BdjZDZar{TT*SJeajl4_f -wc6Y&Zs%iZwu1*1k2GwtdVv7JuA0dJ{nh!+T=!1@^?^3+#!-7uXY -vFR&*T|2Ox<;!m(A7GG#j{2F{%G~ZM5j^YVT -KM-EYN?>Gb^3vGJ&QzrZ{Nq}LgO@?`p;X;YpmX|9kq<%hBMD?=d}=V~*~KKeiL?z7p514cQNI#Pa52& -tG&CSN=%zIv3Mw2^03#w}O5PF`teZp3FXh6j#~jz^SX_+_w<|Cg!eYn}_c?$6qye4R6WwV6t0W89Q~O -S|DL?{2q)ujZN2WJzze{95tB(L8&EUT=e~FRa(&97S?GZIVv(CU)eS_cna*!h7<$W8*1|o+Irl8QY6G -%&(yRyXm(||B_o6BRosS2=B3!9-%)G+AZXnN4{t;V9Nirv1dGP>V{WL6$d?_9mKc0(9SC(?WlO~)+Oy${|F}McpMEh|?)06S+*ohPLpS%&{z6<@bzkTEM$1(PukBAu_0bdEn)cD|dapxu)IJ`vBlMS -)*6i}}_MU2EC9(B(aNzeF=dZcE?8pOVxktx#FqyO^MYl3dg_WY8W2ijTY;Wy2=ztMheX57edjKzCmwR -f>U(H6k|XM6uBgwdJgcrMA^mV&H>uymar#NgJ+&?0EAoWb5&3(WkoF`Jj -(z#koY3OcnK) -^ZuJLmmHWAW45FT8Kbr%SLI4$Jll>dgp8xx@p+XGwfRPwAH2R(>95>RU~Ffw{&^_7?Cp`<3cHk#&2!` -CFYJ=G&*1UwSc_C!EBVO$c-mxT)DgY~k}q`(E(_7FwCT`(@Zop$h4mko_JrE#lRXdcrfuII%qa)u=$l -=Q-ygjJo|p0z)_#&EW4aaQg9mg12cEasp1%=me>6dysK)+20-2Tb43O`gx~9@L{pDQn7;|?+?%EByuq -QzT2`~Cp|8_9cHhH4%e>z+EpT=`@D#kNW_~my^-q${6PrgRpmpfAJtrhrP1N<$=vu}b3G9Li%ftJRL; -3CZN9Ol4r1uQcDq8$54jGbdM?HqoIGHjpsbUz(byf*2Nx!<_gmVvf2CU%dJDSU+D>{##kCcTzYkauMkc%nE5eEJGGXPgh`H}SBkzqi*fht;nP=Re$y~nWQMtQmXS1bO+ -3M|xfHBIaH8DQ_Gs@02I)1H5LwbSl=9+Jcs82f)VU1dI%O_j&Mc899?!VPqK$;i8UX%WqkjeU1V9$cK -5GTGF(O>=^(|^zOR~%oie(_hdZMPhx?8}^{Ef|lP(=;37F>_y@!+E)C!PM4sb40Lzs(s5)w#;i03|uL -K9=vB<+=bNp=Y?S{ZMCNffA4Y4`}@7h{yUVYd@rbP%j7vhE|puABB5 -)>GaZ%pT6#)oJGNAF)_tG33sGNlimDk6o}pywA23eO-kGiaQzvY32YX6zjuq1WU{9u~uK0_Bfr$rT-Y=MJ`J(C*dZ^=P&OEH{AXL2I)6@}TfG*9=Izjq*%U{ -4n@n+5j;2%uMWM`p>-0dYL`N3$YDc#6xY!%4{%}Mv;(ti&enlTC(yRRyEG_DzKFKiXK07D -i#bk@riyyncNC{K=UN)|!^e*^zKkaQ{;x`qKK+kZXM#RsZ}M)c#-f*SK3V(wi}x{S;NJYfx`~YSF6u* -nitBQtE&7Mwl-i=&o2GebGtBlbHQMV<126XM+*bNb%e=0D>)@r1Ew`W2OPaMy(%9BI+I-+$E!Jc{eSF -qgymr9c;{r|5Y>t8NT1q^pL`nZTO9}bgS^c3lmMv#YwATWwy0oFWw8gqRL3odX$F?MkD%p%5S6vWF<+x3Hac6@6eUexl^;;eLKST^ -SyRa7?=ZIGMzq^whHD`tepCQZu!6*(O|Y6*{;l8zgJ0%hAqOE5BN88Np -$Tl;9C#SG42p7t!+kvBjUtHTYaGCDcan{h5utO6k7=P1Y;Lv{JSIx=dKQeyY-h$Ra;8HE{==s(9rlnf -a;2>ypvzFlupndWm$wnJ@01q8uJWM5i&(IvLY>V}MMt#PIpv=Q^q)Yu%L6`7-5q`H|-2G^G(CysfIi+ -&RbIlWD(AHk3@Dq-K1@o0%Olg#^DNzqHmk?-?_px7ISl1NGu51hHT$vAf>6c3XM%qdrZE$Xu_`_J&6KmHV8dyH7mEW1$o;lEoQ>8`@Tc;*U^5R4){n6~82QKhr% -*mLkWmWW{-!V90#zR`JY_SL*=JgAFS1;%XjBJ|S8Aue6k2C8QXi7fXaV}I93$A@m#vr5(n;zu+G56N$ -U&gk7;yGo)jWGL?0W;740A_OmFZ$4$usS#fR=Uh*9|5a|JR4wT2dt)xRVJ+ZQJ=7~8TUc4V7C1p*AfM -@te7w>jnAwj+)S7){}eDw2MogS`uKA*^k}Hb0Kch;v|$(T=U~ -2PF~6rZ<`heJ4=!g+tI6JgHP^@T^P}KN-gh$K841hh$dhT#16IIrgeLmdRQ_YyWG~~+nz`#+PfIz(y~ -`pBc^lUI+#Fh6uTaA0Sg)OdtsEfMYQ4n&By1pzYt#D47i4@n&jtB8o4!I$r6Q(@$iqoss7&nTUbW?iP7IL=|`v -du9rRoSmYY~pXc6thH?OtGvN0Irsiy^dZVYL`FuFu5%yNq4a14^nCrb$hf@Q4`DqroH&EQzwdF2ditJhrJRT|jPcTNi$>>eWDy^447QSXYSK3fEtBPZ~j@amb;?~QS}4w2?4SSs3j{ID6P%gEo#c-_(SnIrtneL=g -F0pM7HUH8-9Xd&9-o(?4Hevg&0wse0J{TMSuJ$*JS{}#sUYyL*Pu6q7gz^nkjZ(;n65uDJ$umAI0##V -IIDBa3@lxeoAIz?5$HFxt)r05j*0ki-;F -<3{cy7{c6-|K0?ci7D$BBdU&xdwJ>Qf~>iHKvA6|t}W>*?|182vT$$>JEDGUCKj7Cwdc&;K>t{;!Po8 -IQ3lDt3_>?J>A{@d_>iZM^x(}J%#m>vr4`gHMmM(4*mX^r#koFJr5 -|6s(J*FLRc9z^Fqf*hE3mqqxuo)2~3N{#<9?ZG82Yd2OYrGL($>Kh4)yUV&drsy;`?<22ba$g -%C;3a@-)81Wwlw(-9l=Y{Is!M&;f27r{7J4f=BTHB^9%k^TNV1g%kY7DRr-a*wwF7D{`&9NN7|;If_* -VQX5_hlTCJU1$?`oV18GBZ@5&H4jh4(>(ipA_<>2>^V85nGpBTSg%d9Oka$40`G99ICG+ -s$HPzw-~a&icgyDZc}Gqef9HY_nfqD%B4(4{~5x9`eojix{^Roz>~YQNp-Yy=lm#}hR>R8xBsuwzO36 -4@vi%?Qbv57bH4Jg%3qf;{k!gWE%_ULZa*i|_J=y;eK>VCmqy?!?fLqTs&8B?wp(Vj`%|Oc>fw-tC3( -!qet&3w|0qmG}@yz_9n>59jodTS;JzLm%3-~d%{q(RQ_X -a~%86v|;JzP>q-GO`>`?jC#1U&Ra@!r#WM67t(C#^Og_Z8zhHYbQXnHQFP)~kSz)b+=!bBgmIt$y_{p -|)(mt>r0|2gEM#v~uQl0KclG&-rdSKO1n%BL6>D+2Y7aqhC&`Hc{%8OB*sJUYBZd5=Nz(t)940swLGk -cLMbm)Fm)K5%a;+81%eU%aFSHQY}Z0rBu5}{{OWxrhmnlJiu$ap{wjN?pX>9Uzt7gF&E(Kp0z4J-kxV -we!M*v6<>3Y#UlM-e`>V%Wn+#>(%U2x^;@@_%aSvE=5$wyOuMXF{3sMkF00c5@nsu;Oo@y_}QG(W--P19VfrXi -dP;B=fqAGmzsDNz9$Rw^EDAPmGpE(rwf`&8fvpQRnF*>epI!~z`s|-6`TC#M*m;b{W4>elJ_Y13)cv6 -OSZ_hz49+=|7R&ZQe|kGXizjL-F-LCZJA#KbV%60dJp}Q&0N;kr>MA@0_Rb9kEgqT@)xzo_2KZ|3*QI -F@;}&Z-n%hQt320(IM;h++_`x0UvX~oY8Qhy7{LEirC6uwdKYEDMRuL{E3-V)MOUp|1UbJGe^xk1o~R -jT)|o%6`F%v6#3yBpNST|5wCl~NvTyUZ!B>4C`PZh(fiLL(5M+pb_|9*Ao31SgR4^ad3b$PoX@Ld0zy -E8LZ6Fg)(KThG9cMyGy$Mb|5#{=~7G{&*pB4%?eL(4@;9@Im}t`J#-ykctEJ -mC>j3M8Mlh?nXr=$Yo>BP61H^z?ja(rt!1-O&b;8MMF&y!2>}Ma`!~iOXuQCGWFB%b82qu$^eUm -;rdy^eI1W58$y6e2rc5FgX9gUu=q(`1q2|ST7SFYRUn9ed57~@H%51aXnN5^US+S5EMRSbcB#K94 -N#}%ulXsKzY+;t)OV7xrWZ2z`o(9?L9Gjk+yf}Dl87p1Ld|J -$xGuI}%KeT4~9*h^=4OpXg#uq^QJUcy*b9j#+&*}vAoA(Yf=ZXWerzYyBTHD*16GzI^81q%h#=Qv65z -vNic^+$%>$=DrB#09WvE~l&?fAc-QO@-pvl)$)~rn{ -0-+qyFSFaG+VL`oMh~*;j&G3!~O?6dINLD@csR-kII3yQR;}sS$p~d-bKGD`p102d2hG;aAe**uRlss -^VY$~_A<{j){Q=*n-kR7`M=-DQAd8-l#9Wq>{=>#Z`*X}3ferNFUI2MJ$`BRB$q;+n -_VaQSbtEHcvcqh*|!ZV+~rNgfJe!zGU%Uq;lwv^^O@1&d*7~X^PM5|+Nv08q=q^a_4tHYO{f%61>QcK -O0;=KD3mOKzI4&DyA)}v|Ey_GN@3CBoYjsGtj1KaWKzm5EDZtB){$eg)=-z(tjpX|uJlrbd|#Ae1DG~ -YiW?*v6v4d8VHZKAavDSy%5j5$AyoV`V!50(~#r%uZ>Y?*kU$2j|hg=}lDVT1Ancy4^Bt%F^biIcK~S -%!R~Sw=VPBVzkWj`jPXEX#QN$6pJ_!P)$zlH0H5T56&j@T#X=P~Ov#%b3=sJq3dpf7KDxSrVuCUdC@a -6w$`IC)M5g9DK{do@h$YgLk8iyAou6LK#Pmejy@G*+kn=rZYvMji*xbPP1 -@$3|@Ka=^6dzG)aIYH$MZUHY$UcMG%xlkUPWeF-i%$FYu+gy -x^(+PMozJauTy+jmV#+v#kNuDB*dL3Xk -EU$F8GSf;O>!&i(YDSrTz%NNRTM~{7RNj~zzm4GSnzKaJ_cUikhn8G7U`=ul$~Y@wYwF9nA@E%?_H~kKYX;i#Oh1yFsb#uJcYls?Oivz~_Z; -BykVteM6-jlS%#CrMf?*uS&M`9A!Lu0K!Z-!jeHhz-RRq~)GTQt{qVT^g{iOAvZ~Bq;M=^$@;P3MDUT -(i$v-zyd<7kY*X^bIROP038@1f2DV~yUmKi59csQ-H3GuACfHOHVGu7aoWO^G7Vk`wcy#;j_oQt@a$UKWTv0c@wcWPr^4%dVv2GgeJInrQkY^+BoZkWr(C(p5)b*iV)@N* -DJKC7B%ClSAX`o(y*^E|)HlfvlcCJTXykp`R-v+JShkEI2(p!HHc>e)^x1J2m;}GPJPGdaXq@I*IQAXwe#XCtIr9+xRW3g$?|l-*@M(D$WA -&YW)!N0_;OD@rJmD0Z1Cyzf!nujUvqZlVaBa%{GbuWP#y1!HO%mqz*%<{PK9tFLl-nf)`P{8+k{;2Yx$h}1SuBbdH&4zEK -=X^L1ibr$JelBBMa2-aC2#WyhBEwNJ}5 -!Y(DQ5BlTvXp57$$!B>7@MZL57M%o`RzlqRp`>A7i<>7wtT-50-Ezm#cw~CS^OYb4V-u9xtvETYYJLH -^J&ha|>G-%DaiwZT{qY6ZpI&Z-!e0{4 -PaPCUf}tOi(BxQygGi#WxxxrcY>B!fj$rA=l!8Q9=O|&Hav_ecC%+UZF=5@%>EDSbhjOJusm~NMT=&y -aDWyrWy}m~%U8QJr?d1$tv+duNVO#Z$3V|`elTy*TzotFN39|5YnV8!&&`1&=wKd{{Hf5EQs^&NbF7wzr`u67#u+GiV$uPuNb_mvIsBRr#FXb0>{QFh?6j@%m!n -6+RI=6Dzrct2oQN?wKXGjoeYz-^HMACAX_Ct>0-;K})yH#DcKcLlrYpTf3GIK6GaiE!BW9q^;z30B@8 -zhv$G_bY$9ofEt}e-61fO{B?qOL_Zpn?YL!@cs45#6BttR{I0^X>Sq+3l^mJ)`w5z_ovHfuFEyW^AqZr2kM+w!GW7_#dMDgKte4qUxLA6 -YcU0ii8rKAKT8 -ya-?-`Db|*jvISSrsFYvtoWR`ay<-MD)*d0LovoIg>itkuc7`LeM)_7642XK^kDXc^KhB~gBXYLM&6e -ne^e$>0}rrUP+;r)!8?%4fSoJb9c_`2rR4{foRUD?|FxrE7X@PKygiGj%?Sa8$K-S5rQ{p1e=PhaELh -C@|XT2w#1NvfZ15&d+^a`FBF{4T%wj@@i?A>Lm}z7_S^&XwmwMY2BVAMGZOYH@YjhC)@`>*hWmvxaEz -(Du=5_|VTnZIdF_@cEtLH6$JBr+ueWXZ+IlORTnv;G3bMQ;CoZv2W~mU9!ZT{8i_}366>bR+;;9gN<> -@v2PsM+l~ZLNEwlP|5ZyQUOPS!UKeO#yxwjdjn_OM0+{EKsS)jhPJB^oI@B!NO_uGhA89u$+-@G)1^t -lw#ieMGY!~ljyPzjR!`?vqke##7S~Dxk8#}k;pnf*y{mlq^5o#yBX!}(3V&Ts&OfPD-$EFu1UH!R1FB -r4t#%Owhxx|KRc>XnM5%U+L=bweLq^GUC$Gx6+O{!1sDU2Zle779aY0wLnn`6+6W14|$Xv?G*^P+Igi -Fx1uav0aH1g@RFZDGNZ2wHI@l2(k&C7M=*=klVY6HPlpcPkwR!kT}D^44<;(CKtFn%QaiA5_ -0@w*1Jg7Pk5Rf}(wJx?3-VuC?0rr5&tVh!j8*YMfxcNS@L -Aa|g|Huw0XO(=4d%)D?+{{h-*TI`cbew??72|ehoE<^d{Ms*cwdl!ae@BSKsKYz`-(*2eIai2`SvByA -T{Qcs4*vnX^?gNGuxzX6Qjq7cXEsBF)Id}SZ@TBjxb0NZm-7bgm^j`@3=+taml36s7vA?BSdq28GE<78q-Ze -X3RatJIJ-o``~&p5R1a_?i?FKo{j%eJ837itsFw||CunHOJp;f2_7Tcka#%%x)5$jQ8on%H&O;jpgyw -Z~06wb43r#ye|DTXGgSUS;OPdvPmZ3p#4X_}_-HKgf87Mtkwn&MJLyE#Q-KSn0>v$DYTPKK!kSdlQFE -Y)-VrvIqR_h);X3%mI|{Yx@J_m66(ckg-kvKc=ZZg8DQ-_5ug=lP8A9A?47Z__K{`&k%QyjmjGvsdVMHRJ2qD7!p|kngk -1_xQnYXg$Q9$r)megzuS*a}r%+_*p?Hv-8`}_ucjn((cVm$C0C^iRe_WtnDMc_FuS3JD%XiW+GMgh7-l06)4l@Utycd+VbE4Cr!7mJ2jy0Z``xxM?nr^Y=&@#O@IAUVImR~JZez|qFP4h$PQ -5*-kE4CxW4a34Wd5yY5I@^|&P3=3<6wM_>m1ME8q+Z2(DB1?y_gM};AI@@2Xv3lI5bYd{lfJ*5uGQ@_ -X&V-eT-3hD0|!+;1QJVZ}RHt;m&JwDm)JOSqc$FhhHjR3F^K8qd=Ws!(NOLZUhKNsT#k79A6Lwhj(B^)_^+;%ozd*aY4DIe`Xm`h9T>2e9(>aKr=?r~Dgg65*ID((!2+w!i2kA!kUyXgUr -rnooHl2z7Gv61bq#ED%@qBYUAKh^T_jDBd^sYa`ZMU!R{$HWJt&|YAz8%lLMLQVb*Vg?ET{;sT`!9Yz -F7V=gz7F-`zK`LJ`mcL8VE@$$_aEf@I$Z(X{SNO9yFjMkgU;^;_Por)t-R+i%4zPkhOM{4Z={}G!Zq% -2PbgogV}F2tgDq{J{H3dY=A --{vZsnYlf*ZhaxZm*$H>pozS;5q>UD>8fjm5(18^SN=sAzziVc9#p8p|T@Lsfmf%nlvODyN2N-Ysl@o -gCHTd*zIqwQWyv-D4lXS}in?cA@%91dLA_DPwh17vZg8=|I*->hGNv~b_}nI1trt=UT8Ua?0) -q5BE!JcYYJZ`wh}=fVATkdPbVG>9^zQS6*uvgl8jhnR%QLekSSl@`~^^5FXR<@d(^|ewuM}oi1cAzK8 -6EI5R!k@oy{VL3_Y`G46ur@mmkDdTbBp`M4qYJj$wreEtqP8wxU`8^$SeiSxQk9OQ{-UDreUy>U+@h> -yD1)`cB9KaFjM@Q40@cgkL0H@x{4j2Vices^t<`vi+V3*P&;!8^^i7_oO#GUgZUB6qQIuo`??4{C^8HZZ%JTVB@2kjuP#vTs$VF4CPZd66*s*pIvC!%LZV^p@KcT`X4i+w13FFrTPR$| -zm5N9@o4Tm^*o)g-E`am2E&uwn*1>R3P&l4(rR_ya9=6!JAO?+=r_wV9+x}n|S)&owxn_m)VNQ6T8l! -k__d@cs$KaJmBz;iHcOT(HAfyOJPKE@q5&WGpDAuQJ%pjbaP=={#3-B-4P`&%^U=fT87^~u)x?e$TNk -IvTl>@`BT8`?vw=Os1$d^x{foQovK$79^a$NXE72yQ<_72-PXjk6nf -L@0RXBhE#Mu@}z2#k5`$BF1QFU(+rE_x;B)V-cNWb1%*n{0n2p&(=3ScD(1y8;>22_-V(EZPc+?G{{! -z>;C@YzK1W4!`ernozSu)Vk|brhYcF4OByl+=zF@hIs5c1b#xESvIrWXPHzR`YH>oPvY`mc^r?`?XFI^Y!;f8J}{p&h5zd9WRBW7TKa%B~IK{ -ND0L&gT{HvQ3XyWxrTaE8g|7-H5q2?Y^JeRQqW>e;IB48t(Dg%+DfDVORKfzR&1};(5ea*BkC>#^)m& -hOsLRXOyAWUKZbGUw7te6Mc$2bM@+};=uIUmZ;s= -n@ffkyTlT@CzE-@}hKtigRFS8*Ivk$H%J108hFV)kmB2S1)<-_R#k$TbE&@2X?lF -kBl57hct{2Ipas_vv$DzdnqIamugZ-ryo0j=}I7d+d+S?{nXSR%^>H^djbxac&(tDm=qx(AO#ATtKYT -TfhzX)2r!fEXIBe=l$HzvAlj>>$2ECs2b}6_DvJ^i@h+xA!@UD>Z=s-yA3-+UZwtR)?D(C)MHHW1EOwoB8_jMZMyoen)1w)-ZUXOL{(Zezfa1iaGnl{)KjuMWoS=r*UzgY -y(Zr--$A+dQyx1l$)e(r_q?adh9xf5W<2P!ZrVGCn1^vQ$w -!r1A67w$V9b@L>6)|^+^j(e~#_1>^WK -(1GhRR#x*c5n-C7;NVK=?rTF=N@6kL0?U)l<8;7>M{Ue&|#r|gNC`MbT)9-w*7Votr!pEW -$Z?qfySi)*SM!8_@9o}dTM5RgkZ2fsD03St3%?foJen9q3+7(b%zJic#~D?KMnfczx*uwvvN-+~oa;< -t_rp7DzKcFBVEDQ5JXEbhtA@dxHx+|o)LKl_`2$8rh*a_@f(s^CZt`aa(87q=P(>_mKwPe-5sLXCz|5pIr%W%?9A!3`fx)#>M9pk!R;#)9(CRvaF -|luKieD)idEye3!**SkMnG>fh*&-$+3l#&tAZLql~OK36ujUa8cwem8-e$)hBB7= -%4_#rPBEZw!RL58Puh{I!N^+?(Qus|{~f`LXHv?Z+OVm|hh;JN*5vp{@PXLLGi=M#rdv&;p?A;kOfhk -M^1qTIvBZE=+rLRR>$>FlE-$D~H_`S{z{yo$tFK^sq0R5#?7L+GbK}*jBiIBYApQ8^B+TnjThvt}rYF -;wSqp3@x5;N7${i?hH$VAD)w;*VR;?g7GohkLc&uRQvm|86a0hYvBKgcS&;pTM9z^>aw2JkquDQe(T` -@dN#vP2^rAZKR2{$Ti)S?&De}wMelbv&{ax$=$?B@s_Q^!;e8Tb_Xj?Et~d_&Qd&*0Tx~TOzhjDmFfA -12P#o{0l$C|Pw$Om$yg%dU?82a13PY2;fY;Q?VN-z4f!`A$_D~1V4ii$tZiamQI%-N-8;Emw?6k07J= -qM0@4V18S+|D$7Eu@)GIMI!bj)kU9bw6s*9muq@jb6`{{pzCUTj^7baB0oU2M3cYEa4Wk>93Hfot2y! -H>-ezj -`TN)Wg$3&FirG0a)C?O5mw{jN`hkJ2GvHqZ%*cSM@fG6Llf2l;>S&_|q7!fZZ}&L>K!nl658vRCP(ra -Ot!-LG^6w?}vy<0*}*g*vx1_9-mD*t&UyY>%FvZ3Fb4!K4>Y(uZLE^-@HL8uC$rxB;=Tq&i+=i0fx?`unjJ)u$NR7zo@GXoTWb;a9V&I!^KXL_fO2XPC!^w2{2O^#(a -xft)Qt&K4kN80%m|9s7B(>3YR)sW0o8;^SXZJ%RIEvJPcKnL&qKwm4S8)I8_L -v^(obt>Xu$O0*;RqxUE(;h3dVzBOgWFX1^6weKKpD`X+y(c?Q_Y_*ow4kqpyVN*{gC-aJH-UN*~-Wfx -fku@A;YVLp!H*K0Ykx%xc*>rA!Yd?%Oa_lPbZq8mU=W}7sZ~F0E*z3lpF6SDEusS|wp9g$uK -t_y5``=WEIfu3tV$U>j4k?biR(bOG6}RvHabNshxCi{TreqoU`9GU)DIdyF13D1z2G`4@j`JEh?^VCR -GsPF$Q=HOiKp1Oho1m12WGkU?-_|x8&n;8J;JyvHGq?{T_ZD#9n!D>gefhpqjrBM9%*yc!=6l8DD8YR -HiX5dK#MRqiOf(=3;`8rYLC(5B+>cg28=tHA{n%2aV+zW(0s00h*9Ik2m1~32NtJ7Z($R*z5?Is_### -)J`NO);pp63WNA*mdT$ooo9SVk$Rb?((i)Fe5j)^nf7YXSM1;b8%a!K7i3;{pEXM%j#4 -6C4Zyw6mZony3j<#?wFxUdNa)R%^yW)?^BI6TJLRRfodsPXzdFUo* -9dayp+A89@%~_$%em`e9(s?)?BDQx>-7K@fbU?RvEU(FnQ;hyqyH?7YwYto2H^MRpD}-Mj~ikH_oul# -_9^Jjz&&Y53AnH2?qLDm++7du(}xr?!w?4j#7gcL8n6v8?kmJ}CDjjcw+;a~pDE{0Ty2-G-u1Toyl7W -;rLG6NiZVa9R_sq+tY_0}{MglSV_IlEomYx;JPdR|d&NhHd7@?!>5yK5Zr0Pd2ckG9tZl?aZ+7(^&`} -MH^WO7iT-P{<#Ltfy27Ftyfpz!7__r%tSqo!8arQFq;fb~Y%F3QFhE0A?%md*W=t1ziwA+}6GF*H{mx -ueCUhcql)I(o#DU4l>!EtK_Tbc46KeKr-?kRT@x`i39qTUd9L5%yN0*}FX?`^z_`yUrTp6hxTuU_ej^ -HQ$!Y+`Q*wxN=jubxeO4ZnSq^}ZSJCtS{>;_uz`4iEJJ=ODuHyK&Gxo|*QWhi3)OtKpn7?hAPq+%Jc+ -;E^rZj`yMbaL#Bo;5aYb>xm7B_d<6)cfZ_8x`&ZF#xbSB(7k(BFw7@i1MJNEgE*Zyw$x!UM0X{d=;jw+TCMcqY?cr?Fx$b}?5T#O-@HPapg~+Asw5>T_k -@qwu~?Zo%X>*Tv1B+zMUX^yKDH-AbVz2sgWM6J=q;z4P$f6uw`b599mFL3`upqVVuuvtRQSI;ch*?lt ->$d>6$rYCPY^2Itlx+%MUj@8dAuN}6nZcZ-g_i?&abdm694s)N20_hi$|@jyRb`3U^rUA}T9{FcLS1^ -m?WN<78hgnGR=zorhqW%yXE86xkKa34U}H2m&Nxq7ALeoJ6+#pc%Bfv3k>o# -u$A%9w~RFQJdZZpybN%kaKLByvsQdpz{=Rl@f+F7XT(`j)lHp4I;f#xONoxH#QmJpXq^n7jiumb(&_i -2-|+_C-}CihEHiyKG0>bx#HXL4+Q$HJY2;=p=5sI?R*qI$Vwz^@BJiiL#!Mk5|XKx(0Rt_y~4f<-%0{kv4cy -zWd_~rtoa)iElrGb9f1H6+GI>l5K!;2h3r#J8?IhGUhu=-@;jxbLYbt>bl3H-|$Rr -L~A#2IJ?T<`)+}xCs>3{k-%Kd?3>aVLLRvW>!+2fP5Z7>}6t0T-O!%RB+_HdMIaK;NL>=J=YuR5azu?lXqyJt#k2gMbdkfOTFmzXG3Zqv`>Y$oYtar8SZ7@0`&4K=p`KV5)bJ -{f=>KFCz#L3=1R-v(6{zq9@!W5g!%OajAwc5zS5Aone>JIYgvn>%WxlZ%tsFBtblap5Ap|uzM420bcg -ZdKzEog%oFHit4||khvjC0a_a?po2^t1ow=xUPMWe|9oCbX26hSUquxio2N|Qnl+Jt)1Kd-4aYMs4J< -yqqU2+InTKY5NRnVa-Kc+EKQTVe7_2ApQbeRv+Et`qoE%fJW4ngA$z$1oT@`FAms(Y&GI4ENGLy$dF9P;KVDG|r2-9x-;U5X6IX<8PwXGIC2;`g0rMX2 -!L}h!{FVpP>I!*6osEY)w?Q7Hc-91?_e#@GB*qZvp+ -|IPQW+i;9bIk86@EWP0F#0ovpWaQEP0d@B6?42gP8dFkEDYgGxPiRCZbg=^SwO;`(rH9$U5hVk$$$?4 -2i;#s5FZJgza_DD-{k2Rn(6?Un{qja6~`6Sl$!etdciw~BGw9UZ3o?X*!J+OZ -7t?(q}ygs7Od|q!s3`s4eR2L^-{W9Je=Fd6o|8iUuW_MPF^1%9dyU>rOCnWlAG1JC`NK55#&ycItOa2xU|+AqT4{oot$EcfG -Od8PMXhuQJt#-_uhDJ97bd;XfouQv#zr+RgD=MDG-iZHb|;8PZ8-0{H;WX1CXkK9PsFpIndvHJ -%8c-HOPH}J=C-|IJ$0op(*olCg+M&Tf@aq!Mo-p;t=1uTEuH(wF^=4aD)ZTL-Fb9LYR=I;CELq3adpe -xyqEBKo3c3Wd@nf*U_#m&@ldK*z2(T9kZ6Mci|0ircTFA{B@CTJI;1Bk{Ey^ZK>qKk>HAX-LrJJEeaj -}g5<^mn4m&4S)dw18+P(Nv-{OoGlO`Vi4gMBgD=P4psB&vZcph;|{`m*{Y!xnkc5Hy5nB+*esQ;B91olkTr(Z`8yBKir@&xzI&{fX$GL|bMGY9Ja(bQIA=}0t^=vboi3Z?SsOLVl!-M^?iGb*llfIF5~d4J)aM*c@930n1x!s*2^N~7_TJ`& -+dUKRfSlN;wRFaP;E?!Sd-xY0d)!7;%{Pq>7zH{nu^|60O6G9QF}32)cnD#FbPAJxRKCESegMU6i@?v -DMByUQ1R+&#T;!hVFKH1S6rch~D^!k&baH2kL%mT_hfmZhGi!I_%y*&6>`4KC2&qT}xRzE9&{s^R~T# -(#wduhxWLtHDoca2a7*U}w*2{L3}?MGfAfNq?)xf4j#29gTmb#{UD2f0YLB)8K=e_(wGOs0N?V;2I6C -)!;e}zM#?9MGd~J316?le`+v0p{eg0Y&hXwKf_PB*Ow^5L4=bu{%M2*2p19VNO&#bPK38>{Euq<>oxw -tUkTimu#s?xOrNlUa4F$1!sUc-B79I2|03a5QeP+C_3cU6OX}~Wdw9c1clk!2bnkzXPP*$mjc`lC1t; -C*S#i=m|62%mm--^yP0D-9-9MagFT&GLx$`eN<*u(X!d(bg5e_F@PdGyIuMxNd;kX+2{&+g!)-r!J?) -|5>zbw&&U#sE2y~e$M9jtNJ&qc!Ro8YhQ-?aUcwtZ{+18x7Lt^fX~1#Uxl=FD`H-JB`a2RI|soRJH(z -+}%b`sT7GqX%~hbx|I&T!=0qzmxPm0;#fvnk!8@||bNa?I?{-0=Vv7dUJ~1_8C1 -?GBqI!y(-Zt)^VdEUP(_W9f?-F<8y>q}+x~bFSH8W+UJ}IvU1_!x`%g_anfeGyHD9LxjsOFxyNZFW&c -w7WJnWRLVYm;KxM$yoviouzpO`-{JTJ9%!r;4YW7V@SZRR26`v_7GNBRv;b}*z`qy#`oOOr{JLX2h%_ -7++ywUv;CCnd(&lD4OzFAi9wdL-+;j~_CH7!xj^YAyewK7*X}S5}VX!+)4oik%R-wt3$^Ft$E!J7iaC -5P|K`Y>1hy~O`y7S@^>VgQ`5NVm_EK^~w!;q10l^%wS+b6rVMkQ+3GOlnF?e+&Dn;(5JelVQITByZzxh)Mt(tYQ%b{h3NsxBo7qulv+`m} -8xzOUl2cRT-OE+274Eok##sMly{7svAxq28vsw&ThwK)t6=_Z-iI7_PwEnl>;T9q-&tx^tGH3pkQnlf -lEN9p9(P++%-}U;xL1?VM#a2@uG!2W@;;=x^G2d)6WI-^QPw8)Rn2Rh9SNU8*xz^(wCU;$Zy>aNq^!O -M2-P*e&+h#Upx-@&WKQnacVsOJ5`7=%POcv@!Rrlc|qD6g4g?_~_+HB9TSqfYPG>pQ0#O9jpb{AjJmL -dE){}`LiR7?@cDb51P*fO$s;Jj=T2PPT2uXjz6t6Y-}TbAN4+Mjt9;tLxnkG;LAG;k+TN^j{kYc08}$=fxo*_{UHc2yu=+ -;ujjkK@8~R2lYlM0;_x?ftSQD6O;{Q+k|Ic0F%*qN`F8uWr*jkvI+tcZf>%D$1;X48Lbit2n@WX&z%V -W6)Kcc}aG?EW=I$ma2Y3Aq_?!HYhmE_=a*H4N -x8oOsO9sgCujL2h+qr8NxLvRHzltwhY>VKAbu8k4bH|kAv$|>hp=$xSu?YuKNpD|E~AD{x1$n??!+3HS*^kVv)c<{mv-XK~Pd>H&=?!HYH$C(0bI< ->yd~?MMFTV8hD_dTD?e(qO-gtBSj-9g$jwnhZy7dxMD$3bDLo_8oHZ-kGCLSJyX)r8XBje2^yND@lVtE7ij!TH2(34W0}R8#S&u@nbl-P8lD3G$tf{#3%A365~F{7;% -Ifo4E-(EOpeXHA|u^L&md2ONGFqMHql(71w^eE92L~DrF6 -0IY8f#^k|mxrQ3KH^qDe&4h!znoAzDgw1<|!c%ZQc}-9mIb(MqCKL=O@@O0mSS?CU#6*;p+Bd1K{z`C2x} -F0Ay2Mrb^=zp>_YHqm5*_YfP+G%A9xkoX+a-&I?Q?LW*f-~{PPh(8m`Wmo{D28f}SS|_lXnxWoGs(m -f2*PaP>A!gv)?8#l_?m7k5iu-Yq!8OzHCXQ;5&slko%NLWV+4ga0!*&IT9I6`vEp@c33ORzUqMfg=`k*YvyyYllZ& -_^sA7Mxu|D86C(K+a<0W{IyM&<(m%&I|0FJvzD1oc5b2v;{+)HzK3w;7#gQh2T)1=I>HADdl0T6>`6F@>X#Sci{#%-b -|^vAUl8^utS4+B>?8e&pC92!@^4PqNVtVepKwdUX@px5E+FhrxP)+PnLec#Ao&vxB)o#c2N5nK+=lQL -!fmDg2nS335pFN_M>s_4i*N_2uQt>_NPQ6QB;_R>D)mD+Ov+E#AoWAIv(yjaE>b^)yGs2K?k4k3xVy| -h;hSXs3HOlsC)`u!pKyfCKjB_7|Ac$X{I{k4fp9S4euTpb4!ZC#F3C9xFx1;_>mN(%svb+h8mE}!%oGfp`@v^)LkC)|5c! -Df%!U?jx2`9?(COk=&H{m2%-h`87c@v&0%bV~tS>A+iljR*u{gbTkgmYy63FpfE6J8|qPgtP=ND^UB! -jaUTy$Pq0zYpO8!oGw{2sbCZg0Mf~GQxp`w-62@TuHbs;e&)j2-gtqM7W-CZ^HWaRKExt2uI8O5RN5m -B%DKdI$?zdRM~_*2^SIeCR|F`hwxg$zJ$vOHz&NEa0uZlDKFs~DG%X`QXV-#3n6(3`xEviY#{7IIFhh -0VI$$@gp&w|5KbdJjBo*Ag$8&fGJV1;Wcq~5Wcq}+$n*(U%Jc~zl<5<$k?9kzm+8}hQQv|16AmWqOE{ -cxbHY)CLkPza9!5BgutEdM0?D6niR4dsh2%%LO!6bVUGg&sTqXGtJ}UVUu9NWz*UR`cz|?o7_=JNA`w -|W(JdAJ@VTHcjij(mPPnYotXUp(3z$}vC375+7gxAXOGyzyH!xP>v!xOHO;RjLvWq87MGCbjW8J-5H` -cC9eIGC_9oboUI2}enPBjsQE6P_+H4M?*kj-&iboIv@Pm=qH{TQ(2RtG!W9y=uuRIonZnEbh;(u(elFoGl5;NQFPr8HOrS9lPso!;a+n~7$d`%Y=Tk -fjg|m}?4$)$gV;;#lpX7E>{so^|l#YY2gXGVla>!HlVIpj$c)1iVpQng+Ef^AQQ06jGANoL<$1p~EK| -7>T>U&QHpT8KR&SzwBdyH81JtUL+k5=D%GP$*f)7C&(gN#$>H?p{u$5{0}DvPH-R`t*3@#9&7iVFe9u -_=TtJpFi89t*b;iN_5B`DNwlC93hQkR!zMy{mxxC#n7gJpL#(eH)*TNmj$#c>2j|`q^CGI5j-hKh)m@ -70&@2qw>f4i2hO^4w~PIQ?ZR>$*+*(B$Xf5SIpmN)jx;NFFDK0%;mpTO)rqYCi2;Z_aul?2IO=dd%VSPEhN44$t=lwO!b`b=d?p{X(vXDc -9-2!P9ltpLCvYr(6Z3Z;5B|d^@p2%Wu3I^Ze#FT^^ -e0&3gyuxQf*eH+n?srzMVc1YbIx+ij7>e)o{P9`k+lUTLAi1}5gd}b=Q>4ZgpGDw7S(dGd<6o&nKIkyy;Y4G1ry9kR!!S4-;MTCEDc}w{((J|11hG=OgWEKC -e&5bX|T^)%s)B^rIr3(Q5u(`MQ^@;Oj2`B(+}`e8;6vu@%~{!bDPCoEU;wrkdPjN~s3l;t{})NvB|%kf_t;e}E^gx{3;)2s`YlE18%M>Xq$Ysp{s*X -4vCl={%D8*V3mS&!>9>xfn4ze%P?c%=+aSk}{e!mkt7_Z9dBnVx1HCz$+~5x%Hdw+ttLxh_+$S;veb| -7Qrt5#CRDI^h=yXA|B-xQOthgi8s(CG|=8L#a=~TV;L-e=hY!_@LAmVfkFIBU~x-L-;UZx$Y|0W%d08 -J|y);_$#R&!e2;z5dK8!gYY(~55l{oJ_yUrEQ<&~EA>J6C8-a>r=&gzuaf#8TqE^C_-&~V!Uv>22%nb -vAbecrKT_b6GXI3Xl=&z8kxY;DC)ZJ<$X~Tsp!z4*LF33@+Bei_)>WsIzqE&t>)>*oJe&OGb3?9k_Lc -f1e`!%5*Uja+dMWv@CcKvLbA-zYzf5>LVfh@dBK!v7qlDifEY}_7I%^&ImrD6H>$dgeFV`LAy1rcZ*Y -_7#KBwe5u3Q%mCV#oVhg?^e>+s>^{|ezK!g5?7?FOV>K^*zZ=cZhzmK#P*C;!=m<+{3DhtKXW=gBC2j -h#Xf`O9@%X_p}F7D~zgF~VyJ%l8>+w;-p@%gJ9pr{y}hTu0wd{?fulu5-(8;;P8Mh_GCDzJ>5n@?T6? -+DS+|3c0Q<*U{@J{1U?TgynmKTvwN02ADy#F2j -<;pv3$B`nv`M-a{?{|5-mb@*X~i^%^z!qTol+L4x$|9Z)vu-ve#oNyuG?S$p~gS11Db}Cim|G3l_VL5 -KCBfN&Nv|EvOE%oI8Fkxw@Anhvj0|mZY>PKS-6HNYc94YNyq+Lun`O9%{6ydcpKZGBW{0VO$oK3h~Vv -U_m5&6sab!k`Aj&LdY%W<)^lb~sIwwCpwx)DK~~AD6VF87TEb{_`b&jU7%s`7b3b?PjE%kAASgC4{9NjkNm-CjUh;KN>r$aPq&Ca1>!HVQE( --?Xcp=eFRjML|?$pcf -)4&`5>@Gecu##wwhmobJYG<;9RvlBrbNtc^dih)#r!6b2R?;>##dNhg!b`E_TW9e4c;&Kf`6d#nnC6W -qwD@->0km#rjLSst++=Y*Et_^HJ%1?#7jmSf??o<3_PglcDNa%V1TJ -vNpTK4fwrOy&8@8*@U4aWV{&QXUIk@~zTp$|u%oob6JqTRQFD#rl)0zm3|hE6!Ha7weqaYWiZG&7{^}u?}ca$01@J)S}8G);XN{6 -WF5G2VsXkTkW^S`q?a1Ut+y0M{OTDntn&F`$!jIx1OWQFV@F$)P5_Mc7$-KcrM4Xeu#A+XL}UugL!Iy -B-XK<>5JLoJQr-!_>1=Dj`P&|mq)t^NO{D%uvM);VqM5te+0Iwm#oIvo-!!7m4J0f)mf-^*>ilPqrWT=c=o_>+{e~&+URn!wYx!@N+f#nXA#qJYK(@xY%uE;LzyXLH -&UfXY=Pvwu;@KCk`4{*i`IrD^IcRU#Nx`b^--z`H0$4c-^?e)!j9}VqJf(DxX+?cgiEy3$Q=2Yp}x&T -iwEEtMUk(r`9il?P~uppY$aGs+i06&**dK$2l|)Z9MnAW9-4FdT-$KOr~FUyzt($Gj?C< -nIXpcncJ=o^7^3q!K_EGv>BF5bLakFt}w6eJu3YRWlWcK?QZHZ3Ex?by2reCEt%S_<3E1?a8_QG{Y2h -N=WRH^WxEb!_Sls1N$|5jcE9V;{(>J*<%C6_TR&3_v}>)`cJH6v{GF@;VMF_R=*HaEy}o+R)ba1{KIR -PxJ<_*aANyLxf{@P3jR%(g?04XU`+E4DU$A?1_m|>vBET5()v?qK``$C%dg6r_W_;XywAU}Q+P-rr`I -R^S==ty?X{*I#W9Q~e-){Hp@>iyA-Y~A&kk=;MQ8V*IiEqWVi^eaPmme$%$>|DRLuORMfX6?ccJF7S# -@Cvj_xFe1ZklAZMBEkJ@#`edUW2~s_d}1Qj3?&odA{x0 -@yo;e>f7a8<3p_NQQO7=HyK~@OtOVu`Rvf@o^S8JJ?rtl!!{jX_`CVVl#}zG++^62a^<~`Ug*Ah!o)M -4Y*C-xe_TG;B{?a))tf*66gVv~Xw->?cMSdfiT1bFcZ_3f|pmo@+S&!A(!Dubb7z>eFU@ub&3Y37j0a^T4KW#w6c^Z^Qm?7p?wHSLx{R5WC~zmc -oyYZw}h=<=3Zdt;`7%!soTE`8DtTm9djQ-Sb$FDKC8ed5>?)pL+gPJ_}?VfAGTPi=|!f`7qCSdcNm-v -Agf+Qy-^q{X@rF%YF~PczOTG=lZrUoAdcQ5017hRIFG0*G~;8E`0V8#!kPsV|Zcd;#F-&?$5bp>yncX -KCK@+W@2{p=Z{}qxvJZ#zzf-tJw|VR;)j?;Z|wd=H~rG&>@PQUpS$qkIf+n%GXMDY-rIg^_0qJ}t+$o -1tZh9FpH_eP#lpOi&vl&m=Khlf>yM3yY5Qj4CkN-O4XpiQ#Vy83Uypp@O8&fwH+}nxFW)An|CASes;4 -)9e(z1DsSn=UarUmu4;Id?8{PZUab+=2d+GN?H+!qBuIkh8rd0lZ`nA!@k6Z3GbhCH29a?l{d+p0!ZM -==?!{$%A)Biw}LdPuMs^)Ju;aI`I3_$6=IGwr-&=X|#rFEl%SRGt -HweJhSAFR}M}2zT7bW$@*z)Uhuv4vU1n+y~kS~doQ$ -S#}7WPsQ&Ee+>eJGYCh%Isuk-$URS<%Ligag7hCw0mjCAP54+{h)^Xpwv^?Wh|L^v$sk3}>Q?#vi>jM -=lZtpU5;8TU(^Czr4JFd^XE7o=28U5ORQ2%Ii|BEw!YMr~g`>17C_N-s&XMbfugV+6iVz*cdd-dsk-? -Zobe+)awjPpPKCVhL|?9!wD->ltV_2HBy{!`w#_27}6e@yScc)LgYF$o?+_e^a6BKt7x>&k@UjU{`A9 -qRMy+4?SdcdlC!b=T4M#RqJ2KPuX=5{LUMHuY -9RBI_hOu4RESOe)YvFe{rOcYwF5}H_&(HL|dd4!eZs6cDukr_X^#3)sRaocav9}$4{Fe?dY-&F;8`tY -HuPxl!&vVa&E#Log`dH$|N9@OaYM=UH(ySI6+HC%O+qPd@J|1OX`F{NQeQSEnd~9S?t6yJVX$ZQx=+L -QAhjyLp65VFnsaKZH{N|T=%Rc$;ncI{5d5oC!-6*~;;s4d;WS=p2oF29H<-UVnedRlAHEhpIuTQPY^;^IGF}5r2_7O{Wj=wzQ%jAi5$LD<5x#Ja|MPtU#K3eg`f`rrhr=EMVk -d+Rv6?1yVN00bl^4w|53~s(_$b)~pP!!bjmC1h&U30#Bo5)@Xel0yV4jXqH{<*qDS6qf`5XzHJC0%;y#K*l~8WDB##EL*mOF%?3S+nUu?GDKl-gN13vW#ez4%lTTUd`CBODf;ven2`#nE#=jktA@SXI1`a^ -eJ*m0n&|Bv-@a*IKb!yHdq%I{xjpdx^27lNdtMs&^gC}|dvWf7 -&Ix@7><^6Iy7F4VkS}MvegB%EQCCCujjjJG>f7oEy)Eqy=|Zku+7|ut3p;k_|Df-&%u?} -lf-UjLfu>k}Cf?|9y{`pM4&JB6jq4&1(JSoIS7!b@qVzWC&$UZ*V0Zhmlkx9^*+$ -~te2Z=Zks;Z=o;9d`~{XZ^7Mn=6KVI4ZnqdGX8LPV$AB%$Qy4Nf?S -JpS9OUDg$RUi0IppLl)o?8EQovdlBDpGbVpsxRO8^q5B+RjJGWoc57p-;v$hM_DtzoVM@wZG+EV3(Pf -5{vhP|TaRs9pLl6w(sx%*%(*ASyVr((C)bOKVgDuX-Z^~g&GXJgJ+pVjq^GMF^?v%dXB{73*}dkwjftv{UCr>=SM!gl$qEq9*oy*x8+=*Gud*R&cmeR#XAtACB_@$ -R&n7W6%~tgviF$*{mTj`vv+_i5hUp0(Yckc&;?(m_#rz&Oa$hbf*dnBo<{lxD$9*DRFjyt^}*8~g{O{jz -Q2t=&hf)!a{5TXDIR|ME1Z@kcdy~@(Kb$NY&$tE`Pt_*=eIkR)3R@$ -=F~W|PPj+*|C;l^vzXJGvz440zVJT7sj>c9PTA4#I5qT%_=fvm$>G# -!e2LRC-4C4BjOzQXT#i()<3ATKC%&h9=jdei**JKkA(d-RFtZhgG#p#120%Hg=H-qz?}FF!Hv)PS7m`i?(cuI-r{eQSqPaa -9laM@PoY-8%JhLiE~!Z~WP>EF*fZPfqFc-&Ubs`H`O- -uW&sIjHoM+7SI~w;o$--W(qN@!0ThJYUR@-Z|=nmxG?XGy0}Ei5_)# -=Ga%qeVZ5E>4D$w>hO; -{qSMxFc_4dsbadvATh0}-{OBjw{;@c~dth|+&P%&v&!BoF4RvV))oVogVLLY1Nphq^J{p@mj^v)Uktb1l9m!wBW{bFm^X~}KTE4PL1^q-Uwz3P11!>6ChiyrmC=JpX&GolmpalgFub!zndJykQlc_@kW5E=bm5 -8d0*Tk@jUZh30U?Qfc*yPVqbLficK==6>;PyGD4IeJRUueV0$OpgwD@$|XgammrciwFNW^ml7?ZvL$P -*ngk}A=*OXK8Lu5JeI%H;yi(Y;eHP`zALDKzj+(@8*d@)kBYvxfLnKX3srY5U}H@VQ?9{mv*D2?Rzt3 --(3+8Lw%uarZs&n)G{ZdBWXUyW8XWnC`~tJpU@kIe6!If&h|d%&Ki>xYvnDrlb3sL+!U<{ki*M#Ld_4xdBLk2{_T_6ovTo~2eV$I68yy~PSyw}_K^D1g&HztGWG0J@yvK$kZNHw8+=1UkiCcny3<4+Ed|HZ!pYtU|E=`6olW`}_vnUj&9X*T3%84Bdy<9PLDxbvuU -zmk4DeH+kpE=4~+^FLShA5&Cd&Tt6MqU3>ZULHtcfYL_^K(9UeLVf9hclK;^x(OU}Emkwsv0QtTeCB0 -a9OC#GgDJ}aIqK1`&~EFOZn5?=Tj%zWku~+sSuZg)2xzh8Tk$j`8Qz(W+*K6MzRE9z7=7jx<~z*%_#k -&SjR`&cy0$fwuld+mw7OP}SE@@gg8w{=BOB|I-C?r1IRCw5UCNQ?m%=onKbb$#u3dRur|(QNuL`bV42 -5(&6s`~7xd-GiH#dJCGRD}AiFJLvrghRp{j<#~!~(Y)!p(D`r?O-j% -y|Wl;+}t^TT;O -v{9J>els+FXxsHHI2C|jL6rOXy-^DL_gn>Hom+cVRyV%?zk2<<mFuKTUWIC9v7RMfW~urc<<}$Fc;x$v_2Gm&nW_T%i~kh{<{^v*M@K_!VTV7B~+chmP~l!cj -M+BP?p>-xNl@+Mi%rTRT9T0r*@kj8-rmEYTVHAsJk1l(lu^w@jPyw#tj`WYTVH7ydKpV7@NmGg^pcjMubz-*S{Q2`U({Wxu -c#=-UwK+1niR?0io7-)iU%0DVst6Z3Nhk8b-9$JehJ(i6w@9QLH$%tmIH?S|31R`b1v}liSyCyuxs)| -Gy}@(pr#4gFgt=S+qzR+EzLYKh`M|#ep3zRDfimEY=@(FdEjpHjd~J4HMz=_Af|>N1j#W)o-K1IM-@e -+gv@WL?ojG&n;NpndMfl&DGo4sj1iBsatU^aYAV>(?(*88i2Wu-!lm~0i6uhIb?F|hLF78HG_n|D(|8 -&)TRl56yTNSvy48`=dZdBzGnePFgHdZ?J2zaI$GwQvVk=536n`8ayduv%AX7mpSx6_M`C@4oZO -Y)zA_}+tyyMaUS2+_A!4nTW3(e{uUK=YA`Qk@rEVL_t`jelB4D|s$3GPVMKnZOh2CsWta_bp3f4dtJUG0t-h9*g)y=vGkSd -OJN^^>o(E7Ry=z@=|H>bEdXk(0#B(UoK%#L(dop7b$nWt<^F2%V=&K&6@*L$|>ju!)Co=f-p3*(dund -=oGAxHO41>J-as4~<8Z{d73BTMAMDMQW`4jq^3gy*}hZm(R%d6W&&^i1bqBzgO{q#g92A%r1Xe3_h1k -mdQXSg!B-|m9@sqvikgWCiq1Aq9vNZ}xi-j`(T)JUGHQW!gu;>=SosE5H^ze1i;h(C~RpgYI>GS -R-y|14vppJUAPL*X9{dxcft4 -a`g*-SyqH0c9Y+N4<>hzpA+HXz%Gc;WlFKCGCIcSlf;-|-(bPW;2ER -!3dr;t5NH>~?b=S-4O33prNEhVK-i0>&?s_qTe;ockRQ2Y6`u-|69tS^Utm8*eZWKmao@F0HegD|WL+ -f7$ewSVRTQGk$A8PsRhs+%i_eh&#zYLXUEGq~;eQ&k*A+Dovm1pzgzz=>~iDzOH;XKbk9?v+#1z`R(X -|Fv8kG*rwwmlfb-0QA`nqL?z`W4zL@#(M0<96_i`^{OWharz;?mX@T|3USj3*r&X({Yz&#_#ZK|J|9+ -At>k7F7+My0{gHcZaTTclW3~y3}U#@P|Aw^@o99375 -s$zB0j0JWCI$X&i$-4!&c^U_~OwTqU(*S6TeHv+tmj^2>bqMJEBcU%CU- -QHLpV!GLscn0>9}LW*z9eyXg1a+A8+FFQzJN#uSAWQP>M`&!#$Ykf#1^1-=K9 -oqFh_@t;0TVMpL6`~0T*k`pEEK=_pJ`uv>dfVv}J}#~MjM~22V|{U{&q -v>e_i6BB9}9}(b3N&|8I7eCcxeE4|7Q69pj^OL2nO#G{_v`(z&l6ns9N#tLiY@!mxiF&3B8bGuQ(Y{276CF=9m1r)}`-pBL`WDfBM2`}!C0b9 -kd4`}NL?elgBAQAxo9KL^ONl;CbQ94}h<;AAmgrAJ|0LQnQ&0oZNTQ>NCK632Y9o3d(d9(f5q+NMJ48 -PxdXcCvl}|^akwnLuopl$lP%4kUL`R$4{SWHKnH?!OnOs;9Z_Ub|#?E@Cm>seCR(pP~Io4szjT0*ytg -jci$2*sF#^@cV1PqSp2W3x>*wpkg=#h3_8 -O|vkb3F9T0?2a+GI0rS&YnsX8fKVwG>#SU}NOVjQTf$;<^X+C5a02GcMP47ov4u99+3FZ=u|br4Tk$m -femxdf)6A;)F?hy=>ru3I6oN3snS>Hy&QgmD%xnZAj1ZfjX;#m#oiHXjaZJL%0ev%bbMYyFnFfWXb7i -teI8HX_<xYEMJWcXf_SP^Fm_nuh%wv$V;wlIm`Yf1;tOWq}%WyjCop3jVS_ -My{w~y{bDjRZ9LXdO`D5hT~ULWa#Fqp6@do_UK62sZvad|!dYBeastkyJ)rmsgrej4Jy#~XvuW+bSTj -=!ELL8t*wbFAYB`nRy;@Ei5vk?00TE~Yu17+?DU^%RR=W_0rXfEPT)7bqU@ph2JlF8d0^c8F?>Lw;OiLo`vebShTc)y!#ED8HJcFjw6G^zP&@ -57>a3wExN3bC-IzsO6nF|Qyj8yAnMCYuEp&8D%fI_Hs%sC~vZLX9r8XQSAo3bV4zHai+`Im>ZUp`)PC -fwCT;zDDG5qU>iPdC+W3Sr-Ds@rsXli`hbZ{&BAz2JB3&SM5P^VsZT*P90{I(+nG%@xM_ --z{R+QNP`M>3jAq0j)$jO~)3rfwlxRtu145K=tilYy-3gs3*=D2ZI -PeBY_%#CIO8ES^_i*s38R6qm59guqP@A#l4KqCvd --*L-j%TyirHg(Hw<6Hg1aL%DJ9*}wmDrn3iOJSr90Szv8N#Z_HY<$VYrk_w`&qFG^SVm6Rj;*@BaR1Q?s`u~1wpN&VtsoVGbpYMC_s^!bxd%bJD?|O&zuC?}ATfd+1AkQw -U5Xpe&aNnn!em@LwJo>Ywn@)$hyAW1h1Ns0a_29cFQCevap8Xl6dGzF&n^9VLPrbf`o;(wBtjXcSJ;w -^+)r)6}R)`Y7xxf$V%`-WUYei1TeSj2GTm$~1rXNph^y3+CzynM~`AR>XE*{iL=mXdUa^%ZrcWK8p>j -0h+7o}AW(CM)Ru)iuqVgS!(h&ELN&OrHuK(wcg)hY(UFKegV?1P1P0{2UUc}8c9DLF)^hX-UH!u?9X6 -S&U<4%6A*BSM@(eL+Kc=4gc|2W+WTnmmT_xA4ccV!%6xqn+VAlP5|mAI`J-#%k6P&@bGt96>sdXx2xe -2e`J51gj$;N5DG)d2rex+)u>hx*E`PG|yfA3Wf9~6J`?^b|IFbUau6|>sZYvmGvLftZ6(;4EB^x -RfT>4u1C4Z44zS3p?PFdf8YnQ`0kNOz2CBd@^$ETQVrPTQIqhl_`S35 -e&rP`p_#Ch?>_7Ys+AH+?4f+%G(?*`{5vADyS8B&iKE*uS6!x~6_6T`zrhLb09&ZW}i+X&v2$2A2+e& -*lriHzw`yGc9@?g49+W9u>Kl*>0-d@;twu64$&gV`=8{!Iz1JDVQ9zmL#KQJXW+V;*KQ}>5RW!Gj-l!UVv%D4+BN^>v&n@{kbU!~cYRh4xLS+^leIM?iPjQS=weB_4x*H#pV6r&5St#e|! -DPU(EvPNQGR_i54vdcUDwL4Ur{+ckenc^=nFzSaFf_!;m4ziK;6zcyB@I?J;UqqLwZ+Dn|9-}fvxRxA -3R@*As}f6!sl548Vdn&*$~_v4!5N4>oL9PQzl7Ji=fpUP7q^YQy`aM>Q{f -Xt_4}ap>K~b9D&q9QwTq$5f`0uh`&_B5Ll(g$(rgBMnMQZLs6Aso3=+1GO@4evKqG_TM&U=;5x3nc6W^Qe#bc -p?&Kun$nvRC-i8diD_==-=-S;)d^F$hbG3OT+;0t{9ue*(4CsNUz^|-dzU7{ao_S6O(Xz%-K~kPrtzl -mdo-~G*Rl6%N-iaU=RtSzeVUS|=l#52A*_I3qF-YFs^~X`wa{RHF{TQ@x4^I64E|70X)8@EMSq61LBH -X;s;ws0vpp~L8(?tT(11byy3HQ|Z-$-Oj;cgIYLeKgTli#8SY)s!Ex_0xo%&k^mf2P_Q$J*+QK6Xk%8fg1D~{XGc%4 -u9%5ScBe;)ye>eahwQZJyDuZFyxN%RS)An>bHezD(-L}s)+`pK^GwXS=CU=^SG82hJFTpWq>HMAc?|uE`&Eyz{t7KA9R2t>^z%v71DO0bP51$ -pMrh(uz@RAbiSm(fomGGp(Yic+Vl?`FQx%|<uKevf -EjfTbu`J%f4&{$>(l9Gt~|Ij&XD23=#I_s?kJTa-)u2jmWDS*|Jn7X)~h20!$o4il0{&k@bC0{o)?i~ -gxY%S#%@#~4#J-~*`7mQ4Acbn{%PsW_t;FbMa{)<91Hi(b>j48Yj68vN2&lcyd1ifeP4hWG{kF-_N(O -2But<2My$=yoVF+0JpT6tExdK1&n%@aM%2^f%kh(Nr8Faw#9gC4kYOHyQ9rK+koW@-L2cw4Z2G@*A2s -#rSW%rs50B22B*Ro{iwY!PlmuH#HS!mTrOm;XdC>>4j^)Oa1`jO}-}{&Upr49w6TxO-B5;%nA99abLt -2eCMzm#&N#qM96nl`8>#ZJ;@1^oN&GK`NvLp&Uv5jrSaOw3B#R`?_ZG~zIT)i`SJaj>M{T6Zo&W7)Ig -m>Y7xT5V;K?vxd@je)O3tz0^2Sx-$<>&m*75otQ?T|82W5H?*y*h#&_7^I&EBF@B!p+0^lm@)T60iY{ -u4epXW;3>bbApZLQ}%Q_=4h#Pu5`SGZqxh5Hp(xLO#1ItDqWuYY -=BL;&=a{4zTRa8&V4(4MZ-(i)?Iks_R_WIXC#0E9z0k)_Sj=0E-p^YoHaKp3k0J<^C;@<-5-7+56n9i60YFqhBvz`91q$V)nh~yspuoPs05gUZegRJZD0g1$zDZ^`wic{G -L7g-qmkhRP`Tb`}#H7m-2;udq^7Yk-nOJKOM%^1wL2wFRFQdPtBeStb^CT4SymS1!^Xteq1-)y8F%!t -M}{!3o+~;_wxLOcbwJN0IPZZ>)#)2He}B}lvu{@a<9pLwX2U_KwVew-Iv}fZ;vX=jQijE_m$N^SNo!% -|HHIL_f*&Hvz4gw`=tDJu8=inIafU2E93>p(k3)yH{?dhO_5tj>bZwg*RQ4L-MV!Xp`oE-#E22%>8GE*Ovj -Uwl0;5Uj#$5bz0&bdKmAmkIB}v@&#n9^73QASqD_q`iD#_O$c-`E#e(f;x5rGJDi%OQG25kFu>G3|Xu -`w=-<^-xh(8v*c4CZ_zr@VjePZOuk=tX2bjt;dm>4o`sFWKbC`U!u3e-RuUtom-) -uwi1yDV#Nya$}6vkl`B_@)vH$vyWK7_GBT8Wi;9ZGh7B8(FWb6xt2 -pp>ky!M-6dS6fc;}sW#O~d@#ooPp#s2;K#m66iEDjz#D8BsiOL6u@xj68H6i1F6QF?Le)G2Z9>}hfAC -n>6`s>H>M7nQCsnY=NAhY4{Ll*S!{_RY|nTVU+B!GO*3cF|9E6=UQuF-J}l$#R9*AlHeH<$lATiaUYd -3iuY__XGYA;Ex9W1mGvgX5x9^uLgbr@JoRI5%7<@;I{+5AMhUs{tVzR2Yxp2cLIMu@XLY!74VM%{}k} -g0{?;wzIQu}Z~oBrF&JlNU>sjA#J+5dw{OD;KWZjqc~>F7946%PX+oY}A!OA$AuqV#-wFIyz()X+)F1 -evz(-wGaloGg{O5tc2Ka@*-vRuOfPcaT-%bC1x)s4nQe)uX41D;XDg;B)e7lhSf&T>Xp9cObz~2b`k6 -rM^Z6ZwE4V!I=n5h%YwujgU&9B0YLNwQbX?RlC8pPyGjk1P2EOh6MQf`UggI_G;h0ZJWV^TN^ -b52IDdQ55^yXff1cLu)yHKX6N&e(13vy`^ZA{3-2bqXv>_eY0`Mbz{X-uK2^km?(VP|B<9YYpe`(UB -$$dlE0-uMfg!-FbRv?1iz~=!2LxV#?LI;MnxU2x4x9_Ox9}pTCGB7kSbkJX$-{ma8=N-H{4-tC(0jde -3*CP6|fV2Np{ayQnqKiWZhJ-36R0~G`gZ_xXu6`sT1a&kYH0X{y?m&Us=Mg<_itzIuhz1_P^S=%nq)5 -20_IW_lhE08LvkVLg3`H*wYN438-}o?sZ*PD92i=-BwD?AZMuY|rbW&-2cS!^XbQ=)h-9Wqf5$6N;Rd -Aqx?emCU-X4u_(zIqlppYbnhWJ8}EI`V>kEPgYht60Cc%qw}604&m$rR_z&%KTcd_O@Mk~-Y -Yz^*)P6(&iwx~^tErJ6p9lCtU5(GA8T|tV^Sh%?< -ck5i0zZus!wDju01J-9aZ$V%uh65|GnV_i%{SjvvBD=MSz;C9jUvPfyLRmoAAb0uiVcn*JuHqMJt~eLKQ6xc<{Rc$v_XBO)k`J==PFd-K@u=H}+jJ)5`e*uvws+x`kYyQh^`+qUi7dN=cE+Tm7QH*eXxeS4H`=F!5tlXoX -^U#pH@-uK_^-3$e8yY<$;-q*6Nm-j6;H?V=?+}o()#AcjXS})w0sy%#i1>(sYzUvDQ -*p3Zvsxl{JTjr1R7&4Ky$cJ12nbD!+1+pdmMn>@#i8ROdn34h=F*=L`aOU%C|pMwX9arVo~%F0vV^~gsbee~Ln9Xl2b7%+h66j4t6j2Vx~Zsw$ga=qD^`S-2PXZ(B~KW -%XL(;HCw`tR+c?j7CE?&HNsMTr>T)upH7~*Qu#!n)5y8G_CA6mL}>EjPR_#pev -GcYjFi}l_p{XhQrqeLvTqIK)m-O-23PoF+5OG-*=`+48Kee(0qKUZ`@K8kk!T~SdXzxd({bxod(jF?A -3=-j#QzyH1(diM+Jto-@spXCof{P5!^pM0|U*s)`W(bi8a2R{D=d+QDwmJ|0z=|?=pwrVr!H!^uJGG$ -_9(r@r#{B7jAyciiY*JjAPHoM+u{v-6i{r20-AjekFi6!J==gysqe(J};g9nuijvP6nu1Og&w4)v~lm5MX_ -sT6>wn)faB4$=L%>Ji5cJJP;GHEX@EmgKsUS6*LCJ*!h)Hm7*WklLGekW!0K`DJdl(O4yDZSs3veOPJ -yOc;d{Cz2NPDpwF!i58{mHVzqxeNWZ_56wuf9^5f3wf -o-=vp%Kwd~U`=2sS{Z`7rPo?zxNXkC@r0lsz%7;q9!!9X1f`@i*N!jX6DVHCTvZkhH^|eXAL_d6eZ6^ -IjHe_MsKVm1e1M16TUr9N*T*`n?(EsECGU)Mvln;W3DW6Gsp+?G>{Zd-C)ox8(7yTD6UZgHBd+4EuT0 -s|Je$Ij}P`c`O5Fr -3lHp+(rf#T(+^*~qHWu@-64nNv@`ma`uL!(UrE2%|9sAIfq3*Yu4&Y9*eH4UQp#ZPFc3Txosrj*zDIu -bO7#Eq(@zrP%w?H87}?;XKAF_V$1lJ9BKL1zBQxfP$XBQN%9p45DIy;KTFQ~fq#RZuWhi<0+-c9ynZA -ZS@7C>^KB?U{-JUyd)omIcch8mRhn$yn>(;Fm{P_~nPd%y22lbwMU6&7o7wS6o=)5QLOt`q=og-->OxsTK&fu&vE{D@}T{4J -TPQId*HRf1L=p~lO?Z=m$`HL%FNmQWGZ-A10I;7;6I-L58f0AN}z6O8;SN%NY;R<_tazUdRK-gDdG4>8GFEu+U!?Jkv+!fQJn5kTOH3KmM$gk>7xaN+}; -Zq1!WkQpgvV?3p|m_DmiOd!|qF+M=qtSeJg<8OImIS+$uw7`ZMV_1TEQ3)7|rz2xh2d&`2S`^fAB@Gz -^NJYRKMe)hA@Lj-sja}qqb*mH=}p8XH#KFMdVK0X^hsXcgTy`^5ir07R{zpQ)r?yV4`EFt~$i-gn}^1 -ztK;Dz@%)*5jcpK-ip%)s=ft(V+5zn5GG9`eD113YY9IaQuJd+NHJ%_;CX88Y!P?AaM(wBJ@g{qz@x{ -`c@Sy193T_p -0_5Te0dm3kK>5PV$K_X_l_?%vV~j4&7~@Lx!{00Y2R+M4H|ZiIO{AOlK)MZ?(8s&R#~gpjBlZ97#QyR -v@K6jM*3SbEum##Pc}R8gu+qsx5_x!PfP5|%Jj4Xb`QryF9+>Epdb`9J-QTHS|LHS0hB1y{b~qd|H#b -)@E;6#KOs-}pzyA8GEPE|V?pWMYZe0i-U<+@6heD@4(@U|$?=O?BzVf9he)7dh{_+Lz@GN*(1R -myr2d26)#@`Oq?|(=?bV2Dq^m++-dG^_7WnNyM;^%t!FnA$u<^DJ1&Lur%33%879+W+U2l^z#p6&2St -EcWU*KUmcz7N>{2h9}(Z#F_u}nbg7 -&*YgR1}SL5SqGGU@mT0j3f>=``N<7cjv?qNSAB_-{E@09J^x0ewS5i%nqLvmb>kB^r#X3UVwmMv3u{p -FWmD!;>6s6JjuH^&D0D#k*LM>!v0qCM}+oGEL5#(X1fR;@i7F~&-#Jtx5yu0{XWty`NS)_K&RzjNo#J -z`>Fri>py{u}sFC5P11R5^6$P&s162>FkH{6oI{^2_q2mtIozaDGpE$pgm%juQqSoWqa@`cvv5$3nKT ->xDr1Q5NaHAm2})bt%S}rNsRgCb?JMSocAnl};GBWhakO%pr?s1;OxQ?=*JoEq$=iGS#a? -`&i$S*$FC@-8pe*p5n?)8V{F1(k1dhyC9#WjeOMQg=-KTh -pG1*qiHK#kKjC#_ZTBlcI2g=1onUWihXpW>pw(>?b@wY>)~gfdFD6L`PgHRsZ5(Tc<9)%qntf^wvq$K -3}bD_;Dhv%eq+s-a|1%!18s!%P}kl$X1mTkIDZy5xc)}lqW(jdQWh>;_%ZD?BqT)6ojX^_fjp4@0Rsl -e@bGYTpE4j1MkWu&yq9w^#@Ez)(n&}iB5$s`#h8la;Y({S;k#mSYLO*qrnYzUFX5J%RTyv -f|bEdN2fPet$=jSK8bm^kxz~{uN%SU~>W!Q!RsYmn&)GPX9%FbBd5bDU?>)yXm^n>qZJ$m$Lg+5$Ddw -%P!w-gWTAFdn1J`@j8QBiWpkReL9($muw{p@S@IsYczv;+Dd-e+t|{b8(4JD?5FCRi_hFlhe`^(^u9^ -z?(=Q+Ms!RVnm6pGMb9KihSUe~f-*=6sR#K(-RG1>Y)G_5gWH=$q03j#2Dm1IJiTq#tFU(*M8y`s=cw -pg@g*g@uI*kA01Ps)TQQ5@Y0C(7qG*o)h}M$Mx>tsQ8WX5B)#vaXGJPZ-k_U@}|u~PgN{2dGciCTVd; -Je50?Sej9p7JD^=sCe#zkfIitZ9^&7umolO}IG%%Fo;*Rn+3;FK*QMW8S6y}3RkukS`JxWd7M^?VISH -So#%b3?+Kfyb`WgBnCh8C8M9id}e63x(mTQc5lgV_WwAa#a#Q*5;CFH^Ii}mrr{-zJY*sbOX)G3a&^{ -nYD`Y9VDhTy&q^`5%KKBo;O^JV@U@Au2-ch!ITh`M8JU0w*e#{R?;PsmqZc}3~BYrh+r<)};4bzXA}A -nnHf2;$l`{{iiae%JXg#-8QWOO8S0!&vX9|Ddnqc;w0l`*!2Tjq=r3Usd{S*f2BsHrD1DH}RS}W9%Qm -XJ)y%x&0@!*V6AgF1p4ruCdM4WWsE>+Z8_fVtMjLdZ}CVo#e%!9r0=AAJq5X(N9`nUrPTmelIcRUsvP -BSo1XY?^FyMeqlY2h4}{O>YuJ44j4?MqD^RP -atozGqDQ6v(GU!u%kBw`=+*5e)Z7HAlP|8n#P_$f)2J(7e!-freVs1ALHeF4BhM4m_{Ou|DrElSv4kA -8JMQr;NVqpG;f!`l%g?QuagyZ@?8P`I&&)3JfmizIKQtta+O5a_o{qL^Q{{8UbX=to%pNsn|TodP7>?h}>Ec@ZQ_Ig{;PcK@uDC(t`UK&Lj7&~$P!10lOonsMwH -Dg)EDSXDinZ{IJ+AriD5BEE{rp^5wuE}$c?ON?~4uCo66wDW%GUoT3=dxYK(d-jr&cnFLHF3Y}@ss*~ -ANR`ESJk(4aalTBM;}!CuJ1n7edmFF!tv|OiGwn0@hG*9x!k!1IQEQ`5vTQi(-BAYeP84K0atzG{t)* -W`vQJ)xlXp=n1cCh6yn@buJ2(`{)~Z13&)==D<`WpFs=u4P2Omq`ERH7eNXPUDSa&0_u9C}Lw)2P756 -&1PdNGmg<5K~&v6{{{VDV2&3npqzuSms2^m`#%nwj&iCoX+I-+a)+oxZbPya0CbkNdozrNQ-o8W$@Yx|5V=FFKhWx;|4Pr2^Bk_N7uurHXnzRta- -eCK{2_nEjJN`2%W6ZbY2JNK97j??qYi{cawN*~KE?S0p6zwTae-92E+f5&saYQKwnMO@S5+PI;QTo>p -53il?|9+$KIdE*DEwg$WCW4}F@%bzjX(xpqIRzcD|f9Wv&3-Dh##yJJ6HYtQJw&O)`v!o9jBpkcmquhAv_&i}Ff7 -hZTF>b2Ki8^ygL#^a>Jm{%L`4U!J_1^p49F>(L4;OR>?QJ+4#=vux0pE35_xpSvH_uO-n*$*$i_@eTA -^t&AU=>Mq;><{W9+jiY^`tIn*^32gsVH0)qagu-SUfI>!=a|BE4zB64k2$}BE!M83*WC*S^W41G+UGf2MvZ>#*~_#r8;5dZY=-~S;$KfgKTWB&a4Q^$@S`<}&OkrO9QRO`UYmoHaq7s -lL&eNM>!H{NYzf7Bh5nOG;veFWXQYW&x^bLWw;do>T_SkHYK?kyqCl4HhUa7BV~@np0b>o>&%tw_67YmeJ1JLzJ2?P;`;U|l>IMTwk#?oC1sQ`=cb%V!-NSFWT#G@Ay~KCa1g%%{)c-olelKB@MYX%{TZ#6B?Aco -{EqPla(icvyV{ZP&J6cMZ#RO@iam^y$-;Z{WMC#=0Wq#l9xr^vMSf9LT)EwpIIV2fjx6|N7`KV&HY_* -2y(%)~K~&`bMrf8GDLJs@iVsP4?u`fv)sAJsY;QG_1O`DAO1pjE;wftYt8Yuk?eIMmq -R8&O$JA=MhD)jY}KbrVk6OTWbX`HFO6JZy2 -LXVAxhwQJX|Pfbm|03Y>TMn=XzGcz+U^E-KYd1q0_)SR3gbsuFKq@|?=+wJyk`T6-up2iqPIm4zub;a -j?+KwGNxX$_u>ZpMJoueIcZcKeGz#L@67y5fpzKE|^9aU9rGVocJI?Wi)h-VC&qA!a$rO%ldpMHd}rr -Ud$#m9nUFx#T78FM1nSTO30J_p|P!|#US^T-?Nn`lR*)%bl!+17NqVyYe+j&shPIDh2)m2(BoRVP2c>Y -ujz0dfpa$YALYa`$an{J%d-Pjyv2Bru^wYh#=4wufoFs2y{?>qeqWc``q+l?1ze7DRQH&pPEl -Hc4yqqc?0K8^P&fPy{Nylh92iu>n{tBtwI><0E9#s%fKxF^T=2aK -_lY2Ez(@^r-k)t{V0{XI^#f6kCUb&tNDdP-RiA3j|5Dg88M`PI%;`TC;4>iy3(2@lKn-^?QZ1_OWFwr -wg_Gv3Fc-)783dC;G_(!#$vmb>ad%Nz1H-brFz94Fa6$0N=YSU3GL#{LDa`1I4HoAt0SxK~7;IA(J_i -MX6+bL`%=Yu9DCO!FJ%`FkJ>H)j@C$wrw*&fJvuHPJ|z%^vEsC!WBj@yiwJis!GH1iweYlLL7Anw#L) -D|p4T*BS_mMfXlTd99(dq*bi7NwT}2m_Eedv1_bh@Z;j$lQsPa=#<0KajXN!kWZx{TXpst@1txkuN`p1ES>z5fPK8=5e -U$-~w>&rl;kE2lE^m;Umi|`CL9>r7YQ9PbUT_*R&4T>7{^-Y*O$vi81dVJi}DgD}a?P6(fj-E1Z>iD= -R6Z*9u`RJg|KJCr%Goq%9kD4@fN_4;WbE4zh`}V)}=Dtz!@zImVPMTvz1ykbtwVyeCO7Hk_vC)&G;yX -`{8#jGw{M48koySd`+&e0Ma+g_M+nXmxO^J(%j-N5gwKY^_Hus$|eP+D+Sf{hLR@bQQp;ihHnDNo$W= -@ZrF~@n0JJX}5%|t8F2_Mz+JfeAq6=lzC6&x)R8p2UCqwU3Gqo-%9d#OUem%`@Zt$ -MGw5{o2PwO^T0hZ}z^_i1(FS>Fa$NO?|yHlsj!w80fq9nB}wKBCTwK`R#xu==a -ywWUbVQJxMk!jX6TUt_@J*_CMB&{s1BCRS-q0#-S>DF{xdQ!SQy(qmTy)3;Vy((Q~cx0F -}EE#?oVHuGb)(l%lQieUFD5E5!ETbZ$Dnn#?WSTQAnSPmJnUR^+Oj~ABraiMLvm~=Dvm&!9Q)GE$nX@ -ceepz8zky+L(TUJt*J*z0IB&#f|BC9G31E<66usHl2VU9?L)nRiaIqZ%iM~S1%QQ@d^h-{B+bG9YhFF -PzdGTWMM%TCI+XBTCcWS3=EWLIU29FH7xjwQ!0CoCs2N2r5d#{=-nnv;-Y%Sp^h%1O?#=QwhTb4qebb -INkcb1HHwbERgfMp68M0nP<-P%CqG8vAgmy~AhIB~z*>+{U@J&0NGeD!uop -NAiVBJgN(xE~$_mO0DhetKstT$LM4@}3N1sgoJ3Jj;4j)I5Bis?|NN^-Nk`X)?J4zkpj!H+h!#&$G+bi2AJ19FmJ2pEZJ25*s+mT(IU7B5 -0@PQVD=f>tHVSywmR1xQuTsFgYv -@jV)GL667!NF+hWMI95SuWbB8p&Aju#|F%}X`g!CMcTq&ehnO~jn4taS&RzZ+cEM$}j`8Xh(QplweGI -57IydaAp$RV~cp)j#9xzJHqTv%FIURYUJP1$&uL?VE@-P7)6_pt}r!|k#51bd=A+3v6x+e_``_DXxT- -95!K#Vf@pB`764B{n4?B{3yA#gS41{i#s8;{m;~Kxe|BFIMPE67-}9I#LGxsDf^IKrbxNiLk=RLTjO| -Fsaa9SX5Y2SXNk3SVcO_phJDE(8F%FTkL-JFngrkYPZ>wFnunvm)Ohf74|B-NbyK9r&v<_Qo>RqQ>-b -rl%y1UN>NHlN?A%pN>z$T^++|RT2lQ|!%`ztt*N%uq*QxqQ7U~`1$>uC^ML2Fr1`;f{onUnR`{tT_^2 -Xys4{q`svME)k!yx;^2-g&jm)*?+H#X}?YTv{CGa>E@HRX^nrkKHC|i_^KddGZ1dxbctW0%>&kstAg% -uaUic4U@B0UVY5ee(C!ai)UkR)XzMX-_**hv{Ir6T+P%c%9=^e3hN2T)4`1QY-O00;of09jjHWdj+L% -m4rYrU3vO0001RX>c!Jc4cm4Z*nhWX>)XJX<{#QHZ(3}cxB|hd3Y36);L_fB~2EpSsKD7tpo%j8YV7@ -4Vog|QY{@p0c8{v4MtSdOsHlV5fdvtljhnE&Ww(;>bSh4vv02iP{;yVSOY2sP>iBbO#uePECf>DIp@=#^Vup04*U>$zXg|-Ch@0lT46r(%j1>eUzfgL@F;wFtworE53gB-Gqd4) -twqRB?F+G0SONfK+bqJ!|GF}L|5=1;>v!i~t313yf$u9%ngHS0?H88u2*Sc^OBX$;JSYghT;8T|Gkl% -7uv|X;zm^vhe3HOtY%q3Wxwu^QwI#eDjSXp!1o=OEBEk{r=Z{*L}r -1nh4A_%<)(f2O#J`zU$`msM38+Fh|G2sMhY?tBQrd5`El7P2r_g0A;%)fIa}mNMP@dn1lhI?IIn?dL5 -43n&5)zP%xpR0%k*gbCklclyP-_f2We4xp&WI$rdb3{93TkmaUBfz+OGv+N8OGc-_fU?`poImr~NCG}zGIM2CC$ss`zq-ZqKwWP?8FFM*rf04+N17|$E!`utI-hx9N4Un{AFM}~-ER -a9B$OvSemu=m#AH!5`1oB{K+qFLpnRskZr#%UPY!=VY_bAgd41Ahn#~ -G?CCzI`0U3`4RLYiR=7AvDYyWl0%L{_LW|Ytwn!9^S2kq@gcOVeK!8$7C#>rZ*gNvU(hE*qjyB-XBma -wPfLf3>l{^(T`c9Y28-H4%3qTISX-herA4vZ3(dZvIA@>JU%T4*kS1b>3(T}G=Bl0D#x -L@cO#-^=fEVKir#ZAV$(wH^>mn*=%nP?1cNe3T911JK$h9QoF*(YUXCou^ce2zkHg-s7Aq=XIyd81(4 -PUsy8?la&dFSi+i0G4-2_}4d}Z*;TC|g}gTn_ychS3^$(s+dOPl{l$L2%Bb8s(Y;>P=+A8fpnu<-z}P -`&P;PN^O3HG~@W$gGaf@18h<};)WX{@SF~? -m%_KSel4`Rp&xDS?*j}~)Qh@hJhqiT8@IO1_8ac^2|i_v+G!0r+UsngVZ0>U?qk)g88_>FvN3jCpC&x -_VH>qTE@=MmG24=(>6&2}+Mp3sU1puy^TV+56=|$do8ZCJ7TqqZ+hL`4YsU%QFCYm!WCrAH9#DnFE;& -*Hgkgn;lLhQT8|cCsjoOb4+>4!>72l(Bwicp<7VVTs2@ADFl%UC~VJHJ-Z?{-69blKeonpl;_@H)E`e -$Orb(G#BR*a+cPD*bOD~3{fomkNy9@e)*dxram$Kdf>cnUJja4*5PGxXMKxbHuPu(fGViX6scurag;M -H9eJ#v|^v2fXJMN5K1O$d$9#Z1YB_SIABVyyr0I6UzCPaz;|l5!u_LIArf3$~i+h)?-MPptTc)q7SMb -hc76163A|QVfchHGN7I}lp(Mi4R?D?5E7j=!AoKIODu4X3hV&<C501DZ{Ldh~JwLzX9XRAl1jVvyX#-kg)0A)(ELn^PphJ5uMqoTdCO<(5 -&_c-qJNtr#$_xAc~6Ah&SGd;U2Hy!nq%X;;@#>phC`GBm@hAu`bqrT8z2yX1HfU!4Cg#gd{--J!8)DY -|TUyzz4(a*Gx14j5XXhX(o+t(N>&8TJ9dp`xr<-?baURsn_w;a~+tvT?_J5iKiZcRB^)-zkz)c`nw!? -rb>=108Z$5Xodx_tO6e -OY(NTf_WB}0q-w!T&w7-94fn*OIB!%8fZO4)0KRU?Y0~z#Bh>B>Zbx|_eG7a69v%^dV)dEV?WKy$=Ds -b5_bQLeMK=SNc^#(Gu*=?>(tCk|{7*NIR2gd5b_~I0@6QNEGkaje!RA;xS!UhH=2WksWWXnGRu%R~8q -Azr(`D0M8HeE4Ic2rM2^C5|LXMOXjJJA8V=Tnm$m@Tc(u9>mOYIpeH!45W0v6_$m?hE`q@w)h&-hK^D7{oartLSri+B2dtjg3 -%_-?IsC0z`C$GzJkJXTLd?{^m5>+M%#3!zs;VcHKpH-}OU8Rhb*gkFjAX>v=Q1=Tvf-u8rxV5A4I-wn -M_@I&+upN?FM{Ku1kTpR;sVa>c-mMk&i*xQhkRGlMvIdl^R^Y~sF3%9x^~Wv;U}wQWrq~D#+jyO?pbp -f~0$5zFx0=P%u=pqrn5Eo^KGZRBT*kJt#1ETb2$j%S%_pvb+=}MK)?=Z)536-3I#r>W -7g6Znkl5aQL)gCe)9{Eg?Evv@DBu_s1qA_9U3T(D63}4A -otHI&V%Jp0L5VuKh&Od(RB1fTq(}GDWm~zjiC{|X5+fXBB+fEPP2EbE56_ju;9tG;85dAs_{Fs#`pTx -NWV-Cq#5~0$AB2*YoB$SbM~^$#PwE^|4*T~u_`m4?JA`C_=7ngA3(-NLgi?BsulSdXcZr7YaO(~RsnZ -A;YXqnT?iAi0cS^BI9iA&814pG+UV>-Adk<zSU|K4kgYIk_ZjYI4y@3PSpwAI8@dM --+V*Fbc>Ga0$z1}jl7-D;v!XTJDSJy3{UZ7@4PP;-MYhuOhsy2XYx33k^9FtI@t%7c+bpbzJz@I8MgF -jA7MwWoo>?zr5);$m@h1?m5?H@#*IC_@9}ni^MeM~z{Gtl!khHBGMA53Vw0;a*Zc7gZ7&d-Jnau08v6 -PDpGk26qKwlfNt4M~{4fx0EKo_biqfW_0Kzeh;hFr@IeJ5)JpazR;5qiI3Z>=&wR%>7%Sa2VF4G -AH;%xda#s_VEuFbif?b_OH)8NrC`nSHmZZ*+l}6Q4`Bwqbl?E|&rvYNMggzyjt(eg-l{<}9rh@Nq4!{ -Lc`ImLS_j8sQwI$-A|2pmj1(nm>lQxNV|Y(hpk*=rt-R_mkd(+SW}p7eQ7741dl|7LboYPCi6?ZK!q~ -UnIkQ7o-3Q{`hK`9$P|h;`0bHRz8_y^0n?85=?qT9waDk?fBKC9V)5X7w`njmmhJWu5@7=unP531{B; -IdR7Y;AjKORvOY{j?wCQrk|xg-2FzZX=-Bi(y{!K|A7ld=6})y+O7=&vDBQFvH7X4wN67;~(2Q{xoy^q_)%egINa -(#C2(dK<=Yab{Od^=#BK4`K~jtnyRXS}9K9K^us -;%;D{m(rn6Cdx>&e6+Bc;;5Nj!p(qz}}(1CV$qk%*fW%>`smjN1&Njt7TZggU+_nB(X7S%i&)xonzTg -$8cj0g8OEPA+QPfI@V{@dGHLjW0x`cKBxy7YSuXX90e1kBpjwnW;Q;yqTF?H5@7IGEySH83u`w!4_Gq -%?8Tll#T5mD=RC| -o3t=B8HCSOijDS&%tp|*z(EHFHZ7JeYkIWK;vZ?cOU`-T<)>TNe{TwPTob!B8XVx#GVX$Hu;K9x(d;@pBRg&g`gA|=x_BQ4)U -LtL3{q=c+L1dis&RXw#-FeTu9JIx27<7y^acQDDuL6ue1f>)2T&_(TOXJ@Viy4v`|&K?hc(&I+#+ZWy -G<82lxB#~ckD6D#14(I(|KRGpVD#%D6lp)umwVA0v#=6P^+oikw4coDCNu1C1@dE0Dqxe(4+H%k%xg1 -I1Y?JA$zMF@uShE5NXwElj9B1?!2m(d`{Xl^E1tt%MeK$t};Rj@p4R{j -dgWWYAg%gjqYc>c-MXi8aw`x6CLX#7Dlc%pl_85c75m0$R;+0OsD;bGbG4c`8m2yI@ATN9?pe1NWzAV?0KwWZq7i#s(C(BVu1kD>%;Hox*{=xdSwrSM>ot -)K*vT-&D-pYiA -7S#3HCqc3X;&s4eAI;iEXa%{ihs%cA(es8;wkk7MxEvt|$_<1DHg@dKWbXTf&SBK!xq)cG_&}tfEua= -@g7!C&tfRT+8H#7_-S;H)IFc4(L3Rr5CyN2I3t3ko3WlfU$ehd)#Mj}4fG+OBaC_)#?fB>f6L>2rPWD -!izG#}i(I_ml_cBS;f)(ly>kL9qpyN!?+qi53+3#FIR-dyMO(3nax!DN{zR{;(=kk{V?2~}_x6eAJXQ -YuLrYlK4HZxOClE}z|HlKxZ_@G!Z53(|QU%dXM?V>8GK2#OSARRntKg^_$%V*glZOyz_gF{b=lc9H?@Z7F%X;>)V^ -NdGb603njg)rAnF*VEzH49yAh~5!}OzaVeq9iP&evLSe$w6BGUvAhtH{3S33njW<9Mb!LV~&iWMQAC$ -9>GqLeJ&7~g;u!h+19@IrHv;zbq3q4$u2(VK`ht3As=ZGu|5n1j-tiaU?T_qzQNeQyroW0sNccSbtDu -F>ou+{bJaC3d$Wf*G-H$n0(1U&~|$_=mGE`T#42-P(d9@P^f)c?z5q>*vfE -h#SMZ2C?F=cMot~DrB$~}xn;^A?XF%}QxVo(25i`g2uc1Zq+Eq;hpAG`!^Ah{_o@`%j51Y<(@4Tlr5K -O<7=`6xm`f+|)GBl>5!W3k_BMu4Ve(fY$%MD1AlW;FOi=gNHHqu&vObOsa6zDG`6#Yt`&fuSVx+sD+zg>ssIQry;zNQ|Sv%Jn51lSY~cL~94&OA#aMldK7I@bnKS -Q~T$3t_|H}8nGE#O3?m}J20CL=(2#tnnedqO#cjW62WpVfeDQ$~o=F0Dc$!J>1@{|hJVBa(3*<4+A -n-&-Ti#0)a18#OgN3X)2<(qEWyi1t5^jQmz+hiA6}>MJOK%7;Cw@=TKLVAn*q=8{$?&6{z~`+7B?8nN -pSaFi?1#3FgGe+S4QdHP!cN%2de+7DESIYnqI&Q&EVNrkoAWSy2QCtbqy$7?9p4oH@odr*4>PpyWh92 -ABYg*2b}tHu&z*ypr=2MCJ!4mkoC53!w!;^lX3@46A&m%Euh`3GXLQsej)qdQmO5E|Y&4$Zdx1}!;srXoo>G>Z60#awWzZwMw$f8Tqeif;F<5A!V(*Gm+Vex>P{ -!_siEpp9h!uZBmIx4rl5gr(r7{FJxzIPdfuOJq}6zYV-_dZ!3KK#5}3p=YZZ;#YbV)954!Re -P|4s4|dD8#yAiS1Vl~gsg8mSbdb^u&;eo})YJAnMBKP?AIubKY9NseXfMccDydi)b`|SXZ0zYw&=(eX -EyC^4JN0Z=pxvyFsPBgVb{e@OZkG%89#mgoI4U2KSg@PV(}L;4wW`y|C}D#q17w^tPW=X+$Tntbo|oB -Qs1b#mRy^WO{~74HOCOMny5#7jH*UuL+@T&FEyFCzj~l=leBXkoI3?K#dDW>8i59^U8Vd}4h|#2^g3d -RL4h;1)Iz`kS=DA|(4kQ&=Gcqz6Si=x)s~cg*Gf?dSJgOd~d!--iR -=$Z>6WO+@>;2vgC~GGgzTvxYAxZG8&%{x1oz-%!LEEbR!C9p*CJW!nJd8~WHd(AJNVm&{R`){#bhNAB -fD49&P)zyk8OAt#C3b5PgHNYE<4h+T${9r5R!h~A$Dm|Y-oTM5vccah#~*M{LJatzpYM{lwSP&-pK^2 -I2#pw8NHI~&LOPSOkcAi{@Fgcf0W!~H-LZyHEFz25`xx{M>jLbujKmK@0ef_W_fJ|hbNt$Kr9yFq)MT`PC! -LG}eKqn&7{c5Beq${K1~9X?wts-zx$FzD9vuq_z)$r~*~Y&0OU&wE4+BU>^WVo1?9%!B9X?TwmtGe2U -=Mt-?eF8UaHt%6#Ch`m{Lya-RAX{I&He2$95E -3+{oyMp=uI7L|RJh7&nlL6}!0fB35igyNtJa*(~iU7s|W0Szyb~1O=F7pxm?~2vQa-{CIpR!Gjo*qwY -e1p|dAcq8-O0jNGfl@bft7heWi3#8B1O_L56cx@`eU$~%3`W42d!nfB0>2-3`w2YRs8;i#g*YOdvNhp -yLtB%*@v(IlDOe;N@;RKK_-e9Sg8wGyN@-HHQ>&B32pMlm$pXz-^Dn4b)a4$LSiil@P$WEHzfeq+RkuEVp8~ -^lA?B76o=Yv4SlLDJO*-wkJX`^y<0np!bJ-{wI1U}t@)SfyN^v?oqG56WThE(($}bkPOd6mj|O5x7A@$*f -Y%97+LDe8XHW)%(1tV8;2vcsRgH1PCPWmV}#av4KZZnB#C->*<4YDSj+Wx59^g}8#3ijIo&E996@0}( -bAWc1x)hLPU1R!H}OF*r*7NbTl8{0xLJnVB7`=M@Qs%h%gYeBSX=%Zh%yA4wx8LYjYnn23?po7!~{kC -=~Lsf3p7T-u4f^}-8i}}_#1nMhH!fXrSj02j{fgsO7i-D4&*$?A3E*dYWMHvYp<-$GEN5+QAeKyBnZF -nyn=&Jy9dK|!!uOz7tW(azlcH2=?;=aL%OWfNbN#EB;;@;6$hc4=Zee)qX8v)_}zRVon>ny^0Tyxa`N -|9etg{PmP6mM5(RD1qHi!xMu3yQDj-mNX#D_GZCr(6N;Jde+;(Qr3ZbB`p${Zcg~{ -H#1O#Kmd%ow_l5UFl}2Q$b!@AeZ!mrY$_C}o=oAiE3qpQ`a%Xyk!c7+xl--!0g8M@3YSdV`^64rKiETUv+_+Hd=D078iZ+O=(ch&{@l^B=gM+c1a}Wmgs6i_SofBcQk|_MC%yu3Y4BZ$_7%^u3TK+i -J>lF?1+&czvoBi^>_HmYjoX6`EfAFYA~03I;`qFV#Vl`P!>PXx^_m*7|2s(#)?@sRHX|#3y -7=PN@Ns5rylB8Ix(174M2_Z36fMT0u_ZuSG<_1`bISS=bXGYLMa~~@CK{O!D5LFuoSu&lfamRxnPu^y -PGN7Q`jF{CFifdZeEifaauUH<4YgWuS#*(|L -3nIR_;}bT5Cjii@nUqClzyMUoqQnG~jkj#M(2X3-~sv+txjRPY^v9gTUG6q4jMc5Wk_&mVTAYDl -!Rxh+`4<88u(kY9RcMQ&^ojZYO9Q8(sKTk-3(ae-APn96yD`es-a{!SOsi@xAHzzQDT$S=KSvd@eEEY -u?Ah+jCHkufXX$xnJ62@6!zYX)~_uTzFBZ?sJh9g_qlQCuUTafZR@=^iE25rY8%Y -|{Dw4E&>8f$iO-?QkrkZ7XZ1Iv9N7i2;-~oJwJ5`op9V8QLj5RUYcQJ8HWGon)k^s@LebJET060x^r_sJkr=@#r7oD%w0&uiaO-Vpju!gp}4M5?VA1UO64&rYqwaj2fK=;D$|W7G!!RH{W^1Nsi`3X0Wu_YIuS|h1Ct}}2)vhElJRESOqd@@H8M+aA>96PcVYRdTtWn(1-e1qcw^8(VvcEpc -ESUx+MO_p(98Fv5kk~%m4henSXT`7`n)Rc_VH1iw_THI>J+O5n=)YAq!u=1lU;ltCis2VY=cr5STew* -DP8Gw;c}}7M_w&?%L*O#d{hy$cO{$!3uZDk6qccxU~gnCk -*ie0e!ym`8I3@u+hLs3fpc(IvkO&Bd>l@Nw*iB8Edsg$)hlFcQj$_(%F8A!jfWf{wPrmka*Jo7(>pRQ -?n?D9*kdew%tkcFw`*42zbNO;#Q>C3vUapKB4s=y8OS{ -u1C9yHi+V^8Y!a9@keYblGcgx=UnvuFzrRP>qr8)LT2edrry%JwBi=(YdQ9%W_AmDCN%IU&+WvzhI`C -9m9PgJD*YV%r6Z>21|@@+%Va)!0q`e(MN2<_h%##{2PEj_B1cZkxpX(S -7na#`6bLzUlk4U+W&;Tjm#uGN4_2F&xRD``NSnD9s*>&T=6x14vTDnsD0tPyVmv$6L_9H>bJ~Biul>4 -LMV&;8?@KQ323o9hm)2eM+-d -ULm9DEp&}6}$cz*?^)b`DJlP@$o~iC=PL29!!IpOIl5Uv3x`Q>)OJQ;4baR-+Dh!88}$lNkHS_mpib0Is0J_NG>I}{&lg|GP}BqO?z#D@Z`HsB+}&_02-L5r3n -{wg$fT%&D=1e23i9LBtZzT5?GE2%Q`zox!MzB_29M9nJ5{AK7evE47O8!xfbK0wKQKRYY2y}`)bzZZj -WAy$0R520RAMN~*ykH#Wj^!tsddj})_vT(atiLqudom-2gth!=v&2l%!J8t1=pn^GD)z3xUCb4aDQLS -9mBuCNZz)=q+#7%>(HD~S6?-I5RdH4umB}S82F{U5?-etZ&ZGIP<-^0!C&LsK{PcT7PzX+i8)`&lG0- -Rz+CORT?N@7JlIYShP6*4{e#fpE^!*sFYPxLTbtf(ZpWS&^@Bp&;B-V2<&^FguVWy&uSD~7=X9TWywQ -b4VDqN@w~nk!zJ&2u8b{NCnCI7#A$FH@Ihh?@mqTId`k{4%R{Pp6M{#&C=J#Lc-l)AM2x;tQiDVITug -oNu4Rc&5EU&`6(A>g^PhbaHdc@GGb36oEg5+5=FSPXc&2%uDMsw9=gzt&edC-c!@Q5jO6ARxOLNL<*gef+oy;mZEdMhVdkdvV0yEJ -qhKOVn{}Hq1bSzY?E|{Wg95BRJdYj(jS@PK6CIau=+fv^fa}1aSIdFVR{)EIvOJIq4&W9jmvmf0Uz1g -^d>XNwJ~^64pOurv8pFp+vb~?OZBWtO8MGeaOt(hvioAmv9~AM5Vv!eprC`z;P9dEaa!4ry09Jo^dHo -pug6z5o(jO>ST@yl~5H!8*qw`izFW5ZkI?~m)^&1=CAm2@#WaY`T3izHIRvBVu -_;nc76K!DwnKo0H6$IJ;8D~@&pA&B1V!KH8%M{d%sk(=U^99v`KlPJ-#5^oLt{;QDn`@*eCze0}hdLE -)fT4@`;U{k}WBg7UZxlr7ETcn)=rqtS@>W}I^ -_76|+w~90b{q|33y2r64n`fUgFpP0SaI$g%A{2w^a+YH*CtlzNU)9VuuUp$u|x0@8c^)>#zK;bhrh>b -0K8q&<~L{oXLezelGj*$b$P=Agt2J1R)j%B5&z%u=75S4@*`_u9X{ZIK2+~uLnnchR&F5QF_%mdH=UP -@+M!mp;r{F|sJg=rK~El5Sm|TCO*fwEAV$f3fcWh(0PoM0A1y=BNB9B7hCrh}17*SAQ3G?)bZ9ga=F= -F4n}vo_p_P3L5go|S^@g0ZIH+}^+z)t!vj@0@i2EVn60D>}5v#JWO~A*6u`hX|RT)&}i6=%^e2$bt?* -~JcDcJ|j%nt(FT;b2^@MU#CjkM4}^3!6Cg={MZ4z^80ni#4wy|)14Ap?(X&^6fBN!>V=5%6gbxv9|lY -|^`__CKc~N6mw+7y|BPtbR~t-4g3S%EPHmFmsVfePs4A@cjoL;WW2M3sE>;TGFCn6&WDuhUty&BZvvTs18WYjZDCIdRFVr8IH)f>8+AX -yCU9l4uq#=i66xoa$jnzk7yGrCG{rnG~&t=U=Z(ihuB@M444NSqM%X%HkcSggq3I3u!mv|vi?HMC{HLo}_7 -+Vw%Zn63(Z)g*77a*NtIM)}_QEWT=E)=%OOg9)9zHGHM+g+`tF9Bf(-z^6qoz(q*br@)IEDecCW-*Qz(53{S!H*{C8I?)mSKIkS)lNGcNy_cTbElC*R<(4U(u7JliDv>vZnLa&&S?_78pNA -H$JL_$(3rZ>Muc88nVkG6)o|xx|}9Fanx>nDA!zALn0qKvy$Zz;&~-OUQ*857mWV4g -d%6}7!}N_m)dO>rC(V;*BW2SVZA@RS`}AWKaaG==R90lj+3P)*OzmXyipUt*m+}ns8PQaar2H*Fb}Uv -(R`2EIc&KTnu$*H=xGwZ^@mmfc`s_isKtr?pcGWiXo=}B`_2~bFI&d_WitY7DQtqL#|PNq1$BwDCXe< -G#p`y)qY_V_1Y{m0t5wx#H^A7dSMaD$U$PzzE!utD&-zs!{D?E;9kuxc6r6@=KBCP?1<@etSVdQvpaZ -eDQE_Nv>v8uxUOs`9eoS?lqw<95|UozWQ{ZJsM(8%ms0l07CLL$J^fd9N0P#?`Tt&m -+(3?g@=;a9nj>;M3y1{w+b^p(nHZhMwqtwflT%V1uLcT)dE@Elg&50GRrF*kN8SzcEL$ -?FiV8qbWx1a+Jo(4edN?q^NL|8!>wB5YS+~+i=f*lkdMH0(bBBfMn|)XC`i?#5+c!qc~C)ZPJy1-mC4 -A`_p`Ri!wdpkDeF!G1a)-`I+OCNxAbZ%6ReJT=8Am;_X`Q&e{!a-ZTmTDoe2kY##=)c7(n~jU+cC^no -0`?TYn1&_J2f8L(}kJa&h@JoYBgel(Zcl@V$zT@j5Y#SSXX@G3h7O$@3Uh+k<{b|*#=RtA5MecXX(bX=Xbn8(O12RWZ3wfhvbi;tdRi^jvM;8xljtA#-&J*x4Dq9D{uld6MJ~D@yQQt&YfHg|F8 -Or3`?^~2>#CHdX?@AVL^YUS^!M`_3T4R%JA4*OOrWV?Ec$=3UM_1+CVWHh1oj40^p9Hdwhc?2zm4A&x -Lr9e|>eqJA^1n`3QjOiA;T!qTHu9luKXw4LH-ln-w6=A=_T?Z^`(>Zj}Y4sM^#flJ2*?JG4q1H -*)skP)7Bse}q+~;M!PaCBwD)LK^$s%679F6VJoaeC>jm+>=_ysH+jE+^m5`;FRUub}t6!7lA{a=etA_ -PL4VQc>p1viV~v+&jKcotZTg&*{7LJ9zr@NEX3(iYewj;FftOwSo{JVp;`5y!*$fc}7uuNnsW(5q)1& -$#8HZQ?rL3|FxfsgTRQZvm31f5AS+{PMGi8qW-ZB5QoZMoz1&u#`-v0;pYhK!3vQA)s-ssThVl6HAoI -P@+LU1e>{4L#UlNr1TH9_>k_Z!6A*Rqf8sXEDeRV&Qd##&^VZfr@zF4pC}mis*QuPS|}E`s_Bj#n7ps -?E`tB$-$H)6Q}>H#VI904&nmQjfcFiKI*}jdQt;~dx;DcivvQY(8=q}NP0KwOzGhtbn%(y`D$SFsl4v -KnJMjvoyjg81$hpv;21;0Ebwwx9zFE%Rq-qbwegYDfqX6Azl;JfMb>yk{3le3Uj5eR8UmwWq@b2sbTK -qlZqWA0z-ups#ZydTb6FMxC!of?AgJ-63__oF2^I0nC9t~D=X)Jd6c*@z2ovxIl=?}D#OR4gN%+xsW5 -30FBXuixChrUS}qUH&q0on}o%3usz(2^g;W?06BB_{MKv67sD7&fX6=#%cCXEi?Z>Dbvc$jrrr@)#0V -?ek4k9<7g?i5g_>)^_r=Sr-1fiN9hq+CTX-4j%aNHF#GMu;StBdYlG?>wFQxf(5iN7LGnWZxf$r$0_7 -rdDGjJfie@xLkY*?&=z1h_1R985ZFv-pPT~9Y4D)zxrr{QyTD^DHf9su0g{Mq#+8z={*)ja5cKX=QnrU-hQVej2c_eIvxp#Zn%^LDjQX(3zUvdlOb2Hs_5$};`6_b -LAseATnolb_m}!6?u~Sv3(mH_Ex%-TcZsUf|I|edLkGWG;`yk90ujG{!#l8y*t``<101i{TNJp`;t`k -0D)6Z-E4If{*=TU~lY@iF#)#l-%*eyC@2kqtd57=n@(v3{e&u%5fO@R9}vYZb5x{2Yx|?k*!n#2`ZPG ->IRWqrg9Q%CjJ#V_6AGi(ohra?MoZh2h5Q*V -F^lO@S)Q9A#Y}wF={9YCMWe-7NS8DEL=YAy>Ed9auMGg+AABe@(h{3t>JI;_}eP}wwS*?$=^!&+d}@fg1 -_BGZ;W;#B}hEsHNJy>{cpbE9NGj(-<-~w@2G*aN{ma-C>}%Ak;pHHLY?V{`UuTQTH6L#45*~J;;VRQ0 -9B8#qLPI=X9S~4y;C299j*+ARxG%8y`4&b^;RL&4@;!>uIg`G{0#l)8w{hbP;JgD?hkE=NIowMffD|M -mgnkUp=;29qudQ#rsd^Fw^C|ijVH78uaB5(yziIP!W9^eyIo5|*5y9%%fuS<6TM_QXG-i~)6bEDHRCS -y6t7wAZ4@j1NV8hlsy#q}FKAVt9xQ~e7T3)M-Dq<$YDRP|N7*w#o6rc+W=i_vx>~(NTvzDYz(YN^Ege -hRqIP5}IT|F1>-;W#DwcrVgTz&Q*Y4D>VKWAU*65LFoOC%~5MwGy!+eTvR>O*l9X>DZ6lWIiB$>6K_d -XfAGUD6cp{jB86;9^yI=K&iM>3Y--exA+(TsZ0T{LXJwhlcvw;S#cUM3nsw)@yF5nuRoZGU_=*YRPxN -T3ltG5aU+75~vze%s+O(;xMK`Tb-ueN&GysTl8Cf=QZm36WxS&5#vgt8>cU&q_~Ra^El&N7SFlP`(JJV>Qsi^;ibdr&tcnJ{nVL=PRLDCbe7d{KL+U+xSICy_rSx{j0=3({H&E7@Ub -rL;&^R;jL4bV|7#R*D^Ln;aGX8KNxH9c?AIspnp-jaAyMFT&4Esa6JI8&Js$ibR=mM;YbiI4ayw18!F -HMMykHtvG}W)^c;6#?drC{R2k~yZyyTEo$7@bY9{^5nxOcon8&>mSW+d7809r1gmjB6f;C&vwzy1<(+ -iyaSpjT7g3$kx;*yCmXot@~@ghl@qEXmAY(2`{T0%Og|{l*>SNmDWDO-Nen4}YnMe0E%E7=83*Xk3FM -c7jg|&*(@++v#qt?i^l92z=ns&k!xy{9X^7{5~EyvlZz`hq7CBpK-C!aCig>W_NFNP>G~L#+Bu~3R1UhC$TrCFRXY>^(`x%~ -Jnvcx+M?GxgcisxgWUD|7C7-;7tRIE6Ow(C1=+K|VP)LOWLh6xVxoDHR`Q-8XTaA`i-0M4NfSy4qI-l -X-%0qpk%+vvL0`EOx|JMqH3kD8&2Ce+jh(~0S%S}aKS4i-mlc>g4MF#`ShDUNzo=27TpBD8SqO;oe+x -`m%N4`x`hOmj@3_cWOU;7>MCpw3HV%>&9j;P89kk@}v` ->tnDywyryn64QHJu@!+btQsZ&dUoLE3%MC -8zxz@Fni+Yg;LucE^H2mTi{oVwn-j9B`cZOEAZ`Xc8+E;`FY|Lb1on@jy>dgC6;kFVv6jKlgRJo7ld# -~NWhuCUSV%pez#w1!rj$E{@4*Fde|fsJfuBz+UUd7GE|VRQA9)!RFa;)ZF4km8G!+TiIrA*FuWWPF?M -t1K;m&GZgdP;2MO(c3%X@sW+CTUol&gXffaO6Egi1&`=K&jH+DXjBUpc{U38D+1BQ_CniNc^P6b!@FI -K{+LEE^}Zexdc>uNtb9HeGxuohEm}$4M3QYEMcnBGaLrOlj?Ur$mpyb{*#$snr}P0h`BGq;n7iRTnYx -|c4^~y`No>@*^GN%8!UtEc3aC>u4P`NY^(YTY?S8?c+>2k%Q2zi+%QW2g|AcmsUoQF@i0ms+m7Mrwwp -IPg3|&1jPqtlC6JYjCym3id<3};dRq1yc?v|f$>)@j(B~eP5NCNuOBKc?k7!bBz6Fvl`aqz0=LLjBqis`&vnZ3C^T!tmX0P}z?1O8( -cr$u3Hkq%dsGYWu1D`?b4|?MPtdmVJYtK`_Xqui*#pMROw}&o3sGBDlTjcU?jO`bnmr+nkFFA)*U-$` -5RcQWm|9I*CX_>{abAGu#BISgy5+F>>b94YPUv?ET$$z|n0KCcp81)$vdD -aa;oRhrgLe>8Ti0_7>oy55qpMZ}MfU2T|ws#$BAppe>I25}KcNRhMI}?iU6BOS76lWyw88i(1=c)Kkz -dUph2c~fh4Vys>4?w3n3_mk21?EdV4$SYOeSz7b$AS4)@_rlI*!Nxo)B7~xg?F(}uMwST(H_7HE~5|F -@Ce8|Dw1Xqo_Sa<`UYsxf^&lD&0Epo9rdODI#UyqV3PLO|gb -aq6S{}S_0;DUvKV!T0Q5CtI_J5p$_N#lD%jII~Y4&|AmeUSPaEC?v#7gmu5+1$ND%eudwHmeXV9{Zi+ -z-wQ!tLsmRzX>OI~&~EJ`Gmle(3cNf2R;Q$GKp=UqG|X(5z7xD>2O2H_c1l5a%tTuZOSIES$J`i0%tb&vB$WR5+Hx?{>L(+;pK -?bZkZo-Y%VUFu=~K&pIukadJ^Tr-}7)G_4F&{Uo&DMAI7O2#EIm)(~MGXVl@PacXzwswVTtCN8+}z}# -ziwJ^x7YUUO^^*(sHv*^v4#1coX{TT4BLuOfW_=GY-ecEZkg#^NP(;Chur~eVjFgn=QxzPUiAyg{*81 -rV3dwHtrwS6E1@9P!_S4?F?Rt73GXK1wCu!r~A=FFP_R%>bRMOSUEy{w6VlelEI5HSD2HQcPEuW#y_S+s(be&O -o0Nt4aGxS;ShJ?BL&5tjctN-(z1Q`x)8+2>iniCGd>N*~5y)1Dd!RPUd2G*l9Y0&#QU6AmxIK2j)FfY -2TAmVRHTttu)u3m8^UM}F|7ZD^b7%;587Yy{zThKDCa{f5u|8T#+ckrz2OKumy$cu|^7uZh=9Q%r!@5 -G8Gde|#gR15refkJJ|Zj{IAc7eP{;w#{( -1OEytS6W^6=YHOSPO3>CA%-I%TO`Kg=-ZbwCFJuzHu?;UEj$klHA2*(hV{8INAkwz5v+R^V+_X9sl?& -K|H=$XD(G`$*zZR=Ae^dic!$0=1MidEt7UmT3LKaGP1BO@PDFZ=uz#C9dsl$7!YD-Uk6A#|#6R)8jcL -gJZGx35oXJO94AZT|P!K~&&aoPtD2ieH6=!0A6{b1>kVyXNKp{+3bYXJz6RByx7P+O2eU4~;)x406x4 -0ya4@VLYik>j?vZ7}KD>;C|jIU_?ZYHP1~4yfG3ow&RdYm96fR4UYS+H3Nabglo+1Ri{mW3<;;lwpyl -o$WO)v4T5`7r4~Vd*p~MG9v?cKIEybU#pE);QH@SFvoRT^4p035T(D?^@rZzk5mwWYrI7D=wS1S5&?r -iNb;_TqK!KXC5x=bVQWvbKqssq(`bwoIKwAa5A(6DQ2oY-@XSEJM{V)|1hHFTrSW#hF=|hbas~Z*OFj -PH3FxDEOAF~RO`yZ1)#}Nxf8K0{aY{~)GBjr|+r{t3dP<<*77{DIremQQt~I{Q>`IG1*yuJ^Lva -Q<22eE7Kc`>{dN{a8Ig^k||v7{FtxLk9|db65<+UYEvExx-qu-;<+3y6Q)1IhWNSI8av)z6caq+nK!9 -C69E#zNK@P+M;e)WniPdHf|AUTn&M;^N}9Yw@s4t#RTV#And=IHhucd7Ex~oj5vQz(VjpP~L?gB_C3;eZ0efZMguMPe>LoNEx&G;a(ZUFp24fxO -FBypi0-D>#nZ3&|C1R!K=5L#-bjJH*>0ukWtd@+nhytl2u@Ebi5DIAC#2@{lDKK7ukwk7Mw2oa@Hmy#bu_>0n-vlU>bN15xeLr~&O*QGCJcIRGX0Qqx5|G{O{P;_N?USldPHXpN1HM=M8XngKv-?RmH;NT62 -xMcgz$;Cq%Kq^>lDf^1YzG3-yDk3c!VX`hFT0<=;)L>0yZ|(i2Ms89h7T%*l_`mOwga+J3spL?N}f~( -`opz->fDTVMnRony>)nnE~tuiZjLWYFk1MD4;9a}$z-)R+{S8d>8L~xzM{FGfH+dQ3h(KwlfBL2kAGk -%tLmOrfY7!tp!H5{1_xaxR@{INwPok6V#V`TjAd5RrLtR6(6o~6wCj$tMcK~oK+%=}JA^-oW~-MYevJ -ax;gl5XCjq?0pgz}OnwK@`xh$wp()t1TTGi25yf|6mfvyXmWgOpD=eg^_i}M+ -TzZJRF)CCvR*u$2ekj{0@*8SJ6{(-ZP&l55l;*o;HlWgOZiuWzWn@QAYbJ$HZG+Tf -WDlGU$H}O2Q6|o-g;a0aBu@JOsRC{Zo&7xY_)}k-Ca5}uuRVCoHde*vXz#3b8PB7IsIZr)n3tf_WGq# -k~q7Q=k={R7{P|;qEdcE&?|WF2Hc!Vda;x@oK+@wp4pd}MIS3gZC+23%VS;<4h -UxosJpj^1YGl`WKBdM20SDZJuMajpXs#@c3iT8!5y2ii1&ud=1ZsnZpoAn@Ku`5wOPa#KbSU^OBLhrYVGXV104}S -&A+ub9HzE4=4{(Zw8TY-Go0rdJ`wQ)|QHPR!-^~WT<*t7GAkez*||1aRjPRNQ~a~_ZO_2Js9wZvmq^j -4kx4p&IjE%SOK0PZvI@a5(Qy6dR#_3jd~w_|cqmuetyyuB7ZL?Rj5m5F9!czZ$4LFA3Ug(vUSCyyX+e -34M#XpnUT=z5>2OOK#^=1Y(TzMMCM**joE20e7a&!enPD(|EK$(Q-HN{4GudDGsjNjUD^17gPHzkoVV -Roi_Yr$rgy!=GpU2y*cLKGveUwKX+~(Y{J;zSP{E|E;SCq{7906?LQ6~B{O0L%#(9xU7QhR -JLBl-@>Mg}+-_1p|0nlRn%B!Y7P;XOEyc{TdKPXM7Uu)W@HQx*)+0L&s&&0c#KsVnid%KsWX?v=nKIs -mdD3<;bL5Ays?b9;yLI1)lfqzIr(T1RpHEwsVNSB$l5=f#|e78n-ikq6WG2f857P9xV=~g50Lp|ZWtA -m%bIR#i342&JfUlZe(qUEpIX>8ZRs3qKLDE9EdXOI5(J6Qi8V!ChPG2PcoYx;zA_orVA*@_X}12Lj|0 -MGfyt=bth#I!%V=l>ASec5>K>>diPXy)g_^;PD%aJbRZC!V_mrA7>GE#m>*7z*oS8b)p>V!82}fcfw@ -NAv%}Ff`ZZcr5p8JeJ$oln~mh-ipU^ze2IxYx~4%vD;`b$jAqA>j?$)qwre-cM%!$2A$=PuY)q -)24y@-++%`8)lkr<;IB|LN^0w9yu7RM@#=Ynwu&YsJUZT(4@It2Fybyh_u0?{tbGO69!7V`_q|{>VEc -Iuk*jIBWy&?0+zYnlJHVD$L(GraR^Og?`h?}+mRe0XnI?y6q8g>p@KTi=r_u)b33whA_oyk0y{D)>x`Bmok=Vtm2^Lvf?t(ZaQW -3XrFQ0-hK{$wes(L8iq5-ZU8lLF2^63l?}aTYV+d=x#*Rul?0-)HF?Y@Tj0gU$bpWD$d`&0uppgS##s -XFg8cj6r%|r7j)#%LYInlngM$Jl1aGfpT+V$1!w^joPz?W5F-8p}CJR(zjzP#cVsE{)yB8yF&6BtYJM -5Ew90zpoGhbW9df@tq0>BU|{na-dyYp;gXhmYbsa!#dUXd0;8r-BDs*&X~Q0h$Da>__U3bOOImYVC%m -EGmQd-3!V9Wnxd0=it-v&?3^t;=6T`c!B@Aj8BUe}{+lQ{MvH9`XZX1R9KeBqS+W9CKfhnf@@AU8#_i -4s>?jnlkPDXmv8RC%@YUc_uTn7zRU5zl!G8&70dHC}^R6W>Ne5 -HHql$Kcgp2`&Lii&~2K*SP6x0?=(ZV*__Ov8MLj1k`X`BmL_7iucZaTRtCdoB4B$c57>55TAzSzdhZ* -neZq;YrckSw#%h1nA$hwcab`Er&qX{|8}oW&wS)f;vD(-|B32u3zVL^2{`1rL6O?7%Y_WpE7eG{p%z9 -gPtQ*e}TvFM&M?wXNB?Mlc2^uz4ZlJ2OdR%Xq_IsbBFBQey!fIlOwvKlg_}-v1XAIFkRKr8GVW*$t(b --joTPh)!#xfuNEF}hL2l1-edNf(WqfcfZ;O{Izi7JfHK5s^7UqP|SK8nA7wFVCqy3qa=om)me8w7YfK -6^Rk@9&#m=AA^L*(c(m*>%?>;$G>Nue>>#^Tja6&C>c!#?^8zO-7WP1k2RE)uOoYB`bc9Lr@<8saA!G -cyx_Hl8@pTpxF{%)*@&$>nIe7$7#bDCHCPzJX$CH{<8nIM4a|4$co#*ucJ8atxhEaKJ6D)&DP%UponI -fCS#8BNZ43iB(raP^iy*$Rg*(h;oCfJz}StyF{d3^VD5&{b!_h2pdgQ@zVS|96M!hyqu&3)!>?f;|E! -OfSO*5m=i`NZCbS$-4A#cd>IK2tS0cd5h~Ad@Xj2emKG9}?wjyG1cFIqvGIA*lvn5uB=4;NSfPa3z1P -ZM2APWmm@1i(cVuXcj`^MR#;DAm<|9YOs3we-umv2Duo;735oh0n#x1o&KnW`c9eL$4A6)?bmza -+icsfHmTKnqCQ%!KFV3pB}nKN?OCo(v%xRZmDdA<@!V%_z5vK4+>1l;ThjcWPB@ecuSG!!oapRxkBC4K`t6H)a+){0mwiSOB$y)j}mvrI^B`U`jovt^S=DfQw$Z>87e=!!@Uhrr?}=qLq;E -;pvQa?@^P&l?dg;VJY#_S{MhJ%$iEEkT=n#l$bLM(cM{5pqq6hY`--!5y36gb`n|DmRh(D#eGr%(nx( -J!>@ZRu1>Pgzs@5I7ruzv+d@V|tx&Q(zS|&)Rt -jOi52`a5d`97=U47&yUw2R9UVohi%qDxUJw>Nq1LWT`UD@feIOcJ$*`E734a9)$ety+Q-Z~WVSLuoP8 -mC`rkm5C=`nCC#eixrl>389L>Y=NhPhA;@s^>QHci>Rj84eYjy9j*ll8^HkxpD)@Wa?`jaHtYra|`LE -N4v+NTG$zf>O_)sAxS!njVewB7!1`AdeQmz>e8c1)y=nlWI1dCOn}|6NYMVjW0jS?8b1)*O%Y!uVOfj^>br5wQ -vllh3>w;DZ;NamenM6juGpHWz`4f8|JFQ4*|+;lsMdwEDQ|`?mIhs^C;6x~WLfX=qB+ki`iw -7|&Mmscaqasj?|P)eS>be5xRHo9_XqN|G)&t!?X!OhqvVRCp?>8M~*2?a--&rr)bGnD{miI#sP3ohne -nXEHGD@}N^a!)G&@bT;#{H#!wAmb%fYe(^ytl|Q-!$WoSa%+=bR7zNQeh5a%_Ry?3~lc@V1RP}}C9zqqXJ0IH4*c29AN{V8CI(F>$10A@M?FcQzFGo4NCK& -q&Rrnh6;PkR1)kg9qhQdL5mes&we8`--8s%kg+F?2ck@inQlpARS)kO!(tP@$^K<$nSBr%+X54|9Bsz -RcGRQ)RA*Y5L`&8>i|fj#Kprlwm?EgQ?QbZbjcZpj5s6p?mtHFU$2tsS>I$ujPG&x -qr?t{qkO!@5q1P-Z?UnA(s<>-yt3n^==?Fyl79YLV?+ -WVp+`T~2h>qgZ_SN{beT{gKK8so%Q{$RpmX5+iVy`BYKI<7xUu}^ew9gfzyy&x^jfqXBTl?utQeqxOq}Y;Ts?kt5|-FaEBx}ZshsF(45=lhc)H| -$eXrT=VS+VPWC<@i&E#6?Dy1UbC~D@(M@K0#H>cYD^#E_$DDBfEP(m(Rzaz2fB%m?2}+TrU|9PhUI%A -4P@Sbm6wr<)?4+exIs$z--bhm-=HG^0YAaY;B}m8U6Fc%3N3eUvXCn}W!YVIu&1z#IE#H%%-Y^i5z6!`E}I{&0Y%eo>eCYA7L*dSNa`&o!zf`3w%BC3ca;841t>7mmkG8t -TbV2m0cr_5wgzEyuh>O$Li_DRdhuk=8`_bF5~PY^j!2Dyh$Zw9TchtCZb=?9pEv=rvxvbz8h*p`ay|45|Nbbr+oqE5Iz9FqCiKh=@#FVaz)H;VtM#EHu7Z)R@4nfj$>QE#BQ;G30jniNLZU -iCq9E^tFQAF$jdV@0N7B@~!*Kd0MRXQ&}^ES4XhtzAZ8|i>$(-vCfhbK)ELxrk)s!>&16N>ldeo)m3b -wye*_qEU$pjc;Vl}x&13V*4=A!n$oxZ#UxV^D7 -^imp$(tV-UN~=$+P;vo-UGrVZr27b|mfpneA*a5!lx`Cde(|(q(k3uewM$U@KC1lyY6mlIJXY$ut7Ot -|z4Y)-ohzqgQk9q9J*YRgWYYa!dU%%2bz{k-2fg(0eF@jdl1Xp)c=YhhimOk_q#d9KZ8k!OU@`b*d4Zda7&hNY%iSO#vM&hwQvPJ?TG3XEATX6!tW%a(S3rB^giO}#SVq2ZNi74_k5i5q$K -5CZQMEITrkDR-})07uk15vyZS2l+mU!+Aog4Fs=JJ?+l!CKIc4>G@2&RU1Ryo%>sn0KGPirGkLKX`>V -ju8?zBUw6Yky=fqHk9J8pyrxtH5%T>2Ush8@OEb0kHrZ?ondj+;Pote42WKe6g_Fs+7RV-j+1CW~+93KBC-i -Xofm7F6s*htsycwdiR8EP)=$RCMe$U>8)|8s!goR=YM)~@(!_GnFEn-ihDp6A$Kh#NIs0hNWD{p~xNX -^HaTw%%~4B!7?z(R;3Mj3j07m9~CyS&Azn_>`WOz9yeLWbO+e+T9Y0==^Wsi+ovZomV3G;QdRC({)~x -8;=ms7Lca+23>v76%K5aO+Ls3IY#(Bl=lNqoTm3|F1oMOxH)*HGd}hV-7IUPuzjM-_%FnQ197e)Aq?h -zYRMHPDSI>=)U?8+z79ndTJY8u)>@rPG3PfO<#+mb0>BwU7cgSjZ^C^tbMrEj=#Y|!r7vQLY5T|uAf( -m<_J~cXN8304grCbhQG&Z5F|g|TBU9F_$zqL+?G%vH%w^d)WO@ZF?c_&+FUtE54_9Q^CG3ly2M3SxOZ -bzBGOXn(0~Q2FQYX{Itg8|K@8^)vF%P23@TH8z@dx=KXPRBRMr(k -jlu3O?(AyLU8EAEoe~MBl9o-pk&tn-d8&|8M*2ty!iibew;}>Wf -o)V>ryv(*0*vSxICP)Q)SUIlsDhMqe}i~-Pu -OiDT_`X&WlmX0#~48)77HVr#uLZ(=}O*tf*;ER`ykKysk(jZ0d^O)!8-_iB##nUIEaApM)RXXqXny8w -kW>-IB%R!x}gCGe6`J)LBP#RsnNm8?voMEnXR-4|Y?>#n!ilgzrVfbG^{7+wc-smh#N@$io(Qn@83Nu -N^UBM>lBUgY)`q1SgeIgy)L+~O;+JRv;5M{i7Jl&K~El!$>tFGe6sRWThubGN>E5uP#@f@ibJ(c()Mj -On#1~o+?B7yX+tGDSbJY(t#GQEZO4ls6K?!m-VCfA$Fo5emG{1(4JNK3=eq};8=Ri+ZXxekmVWk#`&e -s5EMlYZ}+(9*+NhzjUqs`PvR5?thO3gY&l+XH(8T!FxciUGh61Os_s8RV-kZeI?OKYv8cIJW(9Wz=M-zlk~BfwPg9TvC&Uzq!*UTB^-j$qLmH~t4XMFCejn -y#C2x*d(HrTceClD$h2%yh;u~BE_{{mD{F{d{tMC-J>j_iBdrp5-@iC*T2SPk(fr -WW#TB8&shZkd~h9RBWtA@mnsY7^4fvp%E25YWZql0ld}`Sflsg_4o3>IzyzSlf(8owLxcHK0(b;==(t26Q|-RQ@jq -QQ$U7-%7<15Bak7W(h@q=(h}z&b460RtM49s(~e6_pt)jsx=V}A4XerevTgM6PA$mf+Um-+b^7wBP(! -Mvk~XYnDoPkTCODP0CU)HE!9_`Wx50r-!nZ3zO{JH#D@Mcre)@7;>({@NoNkuB!|llk4cnfa4zxC@PH -BS?Zcn@ohB*y+f%rezvev7abLe{Yj@|=m7T2+N<>apzNP@l$-;M)G5eJ}*WR|S0`{~Vhi8I7pWo*M_Q -~Y23uB5uDZO}P6`$HMbg2}ji&1QGICtYm%5L1QmaSo_=frbew7Gj&ZxWU5t= -a{ax%aKjN&um4ROjV()*A4`~Bnyx^AAx66Ui1Qo;}GPr0;MPU_W+AoB)WImN>=Vl%JFJ$%ZaClYga%gp9fX*BiZ;W5t;55 -)izJJ#h$EwpAc=J<>!lW;M2>Xe$W{4pjxnYg&Wufhv+7`K+A!wA&4aT+zpVjlF>&2fa5gTXn{FtajY! -_%N^Y(*^V$B+0k@nN4eIWvP4-X?StvEYU*9Hao&KjtE4#cd(lpUriJFlkfs8Fd5lxmwspQnoiZuq7(F -nTqu!U>l;#xl||5{~26MA>nfn?8>mDPMfR(>1=n3cE&nAf#09^@_6iY3Z7%RRpF<=W5nN0b@r?U~B>= -@6v0;c-Y!tmJY+H0>{rd+LBMZ9%@UN^2@6bG70*3wTr`e^k&u3u2A42<H4~!|!w@QBl!j@ra8cUC{43t1mC7>oAAqH$|K`jjrBIk)#jpU@h!Ky -xX}&WEVZ7QSn=mUG(HPJj=v(H$0Pl2Y&AIwMu_BZAgrdT0g~ywRPz5R*|*Q)XQXVFOBWJ9Cu|eHTtr> -gn1djy&Q*%3q{OJd{;>Rv8bJwUw~4Q5tnJ%kLecQQJ=RY;k`Y2C%Y;&DvF|Knf1Glvgb18krU~CmTED -OqF~aOz9ubEQh_Yv5ROtBS1+5hnc -&)S+QfkQ+;3UC=RJRXn2>h^X@&(U5vC**VqZGzQI0uX?ccO3!Q7#hMb%u8F}VHAn!|_=l=-_XMj8#ah7 -K6D49%B6A4!x0jGrLicMEnjP1g3x`1FuFtPjL|rH5Ee>)9%d$`BXOB)d6|k`I#V9z_^7>)9=#WG8`u|?x -zww?=7&066!m@-LUB&_NmkTeBs=u!i#vuMAYYO+7Jr0q3ndXXlsD;ZHe# -{}<^0=js2g`;BQ{*Vsq`xb8LI -44B2F3+L`mjbkNB*UWp^`M4bYcYue*ain`1fm3v}>4j@h%StfNDPlJhu;TRBpwGUnP6m<%t}wIImr;*e -1!iDCVVLR2uL#>f9x>fD5@E4;<+}Md}?uA+bZ55CEIa_0P9l1S^fQ5c6BrtyLPvG@QW`S<(u#(I<462 -mZa -pRU%FqaUBudOp%i8cuJBk;2fn0GEhBmjfvb9&`eg%YA%_CVLJWf`!le=o$b -+B2O4dUP<0W(jaB46BWy>gy5sa1zeU2+!txIEtt@IcmAWUQstH?mN- -={wBZ9|*;da&%#^TfUxZhx9>ypu++(O;eN*~j8O2mSN<{+0h2N}FMKzty(MY=82A{9kiq_64wFy3{{h -;=NsLmiu6qSi4YIdkKbB&jJG&k$cIZ*lrVb&>4abGO?e`=N#mI6);4Dhk$-9aFK6H1axf7^1LliM#OE -TN4^rrbE4ma@UJ--HFJXd)y)ti^cvYw|E@Sbc=`HOFZb(Fj%ZemJZNW;bh-x+4d%`lN?d&jC7Q)nsSN -JO=bA1R$R}4T8S@r98)4Hr{c-%c7kv(lr)|WhH(RwyM5u`Q|=k=ayOdrR%)yK)*U!_dV-cGl;fkk+>K -c7u}m<_C?F%$b}JSzTZv9;qa~cqOE|^rSb!V$H>3SBqo@71LYESATz}7#Y=ckd$ZV+z5U)r0o1hKaKp -6ef2e4Xde4uxbK%-Rl;V-YW)M&*JrlqAiT7+wzHaf18Ao3y4bM -c!Qoa(?@Cp^^xX4qfkY;Dp>Nld -kjQT8MN*(GpTPbSHOO_HDL$H*zQh=0Fj3h&n+0#a&+adk%ZL;{*bu1u3Ii!{mEg+$^$;YdZXaRwq)?G -+y0i}2fsI*Ckc}g=*<0+s3RzSonh^s6gb6_i`xEXxL0)ipuoKVz-j080fZHG -U;4A1I4r?_y%*Q{z=WRvsSaZ!6|+GwR8xhlOB~QLhp8PNH4|#EX^?7T=su)Tc3H1%$;McBS&!i_=Ez5 -5Iw_y9Wo@qLc5sm4PV9cQPhY&Si{2Uc{Jac|K!epEDj?54nNy*uKaIZgD@pAn#^8Hb3$<#$z`lZ(%(4NAhOIW1l2 -fFdkbQc@5+7iDuc(csvUyXEVN$&T|=$XR~B0_pU^$xch>;`@WITNUCY15`T#T$`JU% -WXx27xZhhctTJblrnk@0vfSpJalI9!(N7>~yT<+m7*55maX8IKRt$>cm! -}w<8Pcpt0`Fh6Vwt`&Eczm{7-of}#Pam|}#Q13B%NUQZPRgqoACG)K;}N1;&S1QNd@AD+ -s#_KqpN{+##%CZO%lKU6qZp4*1j><&Uxj=ySGC-$_>D=fJjkY -Z){a$TLBv>5CEJiS=2GHrnbbOgkH0bD<&e=59$be4Uy;SEk(}@S2pP9}Preg%1@0reFS{W2;3!e7f^TaPI6`5hjmZtHN}=7gy~%+!bD3%yFx9 -7`xa38#L24g!h=d&p98)C*air;am+)lbwIr=D~afc2)qd2UMfSo4}Hfcmt$INB|qR36crK9J+C*m7>rA8| -s`2Y$URg&M`P3!!h0w?AO)t#qU)21Johl8607_N>Sj+2dHR?Aa6{ -dSn1O^1Z2*LYmttZe(%Gv7{2!r7vq1)_+NVI_rBu1((kq6n9}e4#XibmaiaX=YWKt)_KZ3%J5<1!)S+r*PEty5gq+7K3l4?yTu8Gt%Y6ahYcz -RK8r!4YAYd&#@tW}x1WNB2C3$dTe=-Ut>Yk{bj?J|^zp0O0r%7UoXIRCsvCFw_Jps6a^S0=wxl=^Z-) -Wsc=cf6n|n@TV0#IVwfL1J*}MXeZMm0&vS8fBHqNsmvDUTNFnPmMQXm6CcQMfPDhk|6mc~q9W1#%kUxD;Q7o0(C_H)wG3)=lqP&Z -z{1Q$(L!^i4S;S)|%t*O%3p&R4k(7u(jiC~{^hru?dy+1aI{m5=BVGQ8YTfbe3$gv%7xe7jS$WyR4CHwjec -{PA1`DbY)In68HijOAvLqbP-}5^vBdwO0OV+AFsIQlG`g;&77}4zibS;`STN2(YN~3RKL%kOC(;5Wd1 -|gxr5r%f=Y(elVl)V0Q&8TWVbfX_TENC_n(}in9LXF}pR1IvUTl&XF)1)CVmIX0z-yO|X;Hzfol!(`u -VHP#~0)3QUeh71Ce}~5@y^p@JcEE*2fcTf^PAYY*b{prVdELy$wU+M--J3pwNrL$dLNCb&Hqk^(8uY> -aG&l6+>?zlfX9di!WSU>ibi4Jp6wM*xE`c9VczQJ7hrLU~b$%K}Z3xuHQAo_X`_YqCqq+GpW*+J`&!O -h&s`($8dBY6V{AOw%ubSUK3QO}O7$n`#HnxX?^-5-4i3nBgEg40#G7>DgV6Nz=zn>lS^7FqRJs1>lKR -dUz*a}j9J9RQ;6s_k!(zi02m55QJnBbJRKp8?nHz*$?8Any!aDOm-$RpbAG}1S(Oj*h+~U6% -BXSr>-YW`;euidDM3FIj)@bLh>v3ke?Gl|uPhA)3%T4vZzbXlZb3({$FXOi@QAY#@jGr3YNpFf7CcM~ -R_Hk?2@f8|)yp@{q=TIl_xYmRczvT6c=2Sm*b_%SJ#geRD92nGCg+Ntp^xD<&(Mb==k%bPmN4Ksivu4 -2oTSPy*AWYhi1fQi&a_Gx6VbaDd{oqgmO{~z+)Gr)MO$H9^osi&Ji?wgAXHjUf&F*FX#fjjF=l-!=rb -Dn{I;xt1P_t?9tBz%^57b}iWpGj?&Yd)Q33a&JHLH;io^Q`;s)&eWatlYeZOovN8sGuWH#=hE;j1Z-DwG{@5H8)%XcW%<&5}@Eq?ga$=>_gT)uR97NfZ -H{JZT7Rg#_sXJSX}|o7%aHeGnHfOo%xxR0RiMg7?d3|A3|kyT+z4bihk!_*PMN#}0r}IQ%SI@Rj^dF6 -Av^xuHLvAWWj*hC|jwnT4!Rllao|w+wdIIat0O;yff?zWz>kf&R3By?$NekB!AetMs){<@fkQ_}cX)K -U3*NR}d($`3 -K0J+us4qKA-?sOg_z=E)p=|`~0l0hpI?ge3H`2DeDm7{$j0)K@zhBTpIQ|L8m{_D>y?*n4Xh);j9?n` -ssQparVJvIWm*JFHiO2^d$boc=^k=+En^X=a$v$`#u0>y726DF{D6YEre4u`FH3COK055H(k8m`aO4s -yFA)>*w!Q9tcVU5aUsGb$(s0>qSHW*>VGF%n}BV6}270C$}D888e7GD@Kw`%nFy`teO)XXVJfvo+SpB -Z`$&6NGFf&7FHNC?`Lzy69_ctFDr=f?9Ry3i;t#G}^iQZ3tcjXQy-H*i}YuNt!*rpY5Xw5q=BW)0g>P -hl4dAaO_NU^BQ5k87vX1Yjm6Ne7_6zLsSoZf{8!O)-~!5!q!@c2v54qaR-OUlU~|xV@di-+fp8Tn&C6 -gm{He>1gLMOPdlgpZk4~P=C@j`i-xlfjP>o)Vl6MnW2*7?7Dic{&Qr-683oFdO+WZ>T^Ez&We>7TPksQt% -1tmAYT|T&I`fY>}H|r4FdvjAHv;q7M&5M4caJ+a$riJGpM2u!;e58Ak+K2I46e|`4PV&WPlbAVo -7_bJp|MGF5sjHNKSt@{*?)O1c^0;Vub!_#G_JH8=&zsVkV!jq!0?NbyC(xy?5M&ZZzf`pO_TOnM>LGK -!kwb>q1hxOTx^*9YekZLGQ@?$ak;!YsgMV%UBR=IBF@Ey2@~dVxuP$}prM^zS -bEXVIZOU!sr#f>;4;LYZQ6YvXXrhscdkQNDgv~@1{#GVI@VJn5{I&qdHoVcT>H%R9P|r#oW5z*d3JaM -OL+iIgtmtE3`Q}?RJ@k-*O!e4EArhmncN6;eL{!B9+%P$S-KO` -U_3p5rtX|kpvCMy2x3mlFr`zu9g#DC#U-#HgLRu&6Ckw@X>fOq*pPtEcXX&myFYKqCX&n3MX{IZ$_E; -RB(jE2_Ju|1KkD+eSRE5fu-OlOFdIUMfyMhT*E2|(yYjI~-Iww$oJ1k40Mcp~IK=TflZow;tXFPg>^sWgqf<=?vvui)Pr5&t%~BmV6Y5B}}V-Q(YWM)7Zty76xZwZp%)Dir@VAMtMww#UC6^y~PygS_x>4_+<)ttZ-= -ZlVoix27nDe@o%-aJ*Y<-Y~Qs{%sOOLB+rQrXBvR3YCZWw{LdFzumb}w|$^D{w)W+suq0frt5c>^=JT -7Q;n>}LV%H0t4?1wot`wSOTzXdiv!(m|3PwgbeS-5wBE7t|#n?u`F!fVlR2^6v)3y$Q2Ye(EEJ>WRJbKL?0Q#R -ma#YdZqs&R~GJGrRzCXHZ1H6ZT! -8`=TlGH6#E8oLp3k3}d7baWIEcOu>j!Id@Wf&!q`Y5g^T+y^JJLFSbJa;H*&+%aANxtkb3Ze{W0d}j! -BofIFpX(hcVbR@K?fH8(98!#JuRFK^8;=2$d7bh4DBX>H($VL6}Zv9{MO=CDj?zJ2u_Zkk7+lNEs_T& -(`{v09~?gApveMZ+$PrCZbCGKG -4hZoY21vY!wEYK|`*Fbq4)0gLatm+dnBd>sutN3lSq2PW>A7+yi%CNsl483s-O1t&$A--1J}Bl@cerQi4mxK10Ids+%_Oop0aL9tww=S`i^{zR -j~?V4X&1N7;guX2c67@(Xe7c8RIr%C!&g}s2#|l^JMTT79VoGPS${nszpN~LT{Q-gx9v}%WNI)kicA3sqm-wU21)? -55pkzWf9bx`zA?~ACB;B(ZnRpf-Fs<_zLlCfy7#s6t6CHGFU1R7eUf{G2oHXwb78`IC|UuGx6z4U)S% -Ih~I|dIPorh`2eclDecmIkoIY}^34xOzy%{qyuR$Yw)QdGOs~NG5YJyO9j)Go>;8S%EPOpj`ra&kiO+ -}j6!2C>o!lx>vBec9dPyjwyqZ5PdQARY!V7z%0n#%Nk1*4t%Lk>8Oj13+ps+6t1g>w)NqVcbN@&vyHG -6#M0u0Onu7EJbahCCYK<_y+K05xRbipc}mHy1$_L+E#yFEH{6U;m=;9j6KBpEO5B0S69seBVz8;hbQ( ->+o=p>9oRWGKKnlX86oX<~XVE>2xG={MI32@~gn%2|jAC1<1>0Wnrwx5D^9r&J4*dOCkms@;%0sU?^{ -t?IFdDxrtG=Pp_lbcb+&5c3oE-{m^2XaN*wN79Z(yK>x*LgiRU^eUlp;-^cn^0cVaYd=Kngrc{gA7=S -Y#7?R7JCF&FgzJFLS?831g -j!DJ&vR1D!6&zpSU0CUZ+qGD|V2l+K`pZWJW{sm*fyGCq)IPtN)gttP#LS)YtJT@6?OGyVkK@F -57iAKsvwHfchxzU*nt|7nwgbJ=AOmOF@15 -c$*?9PlV&j&)N0FGL&!NkpQc)PdS6N)vVT3?F&Ka(4%%F!$UeVKq -YGwYbOGO7-;uTBTvLWUo*A?kE~)GE=5nfK&&MR|ud)>Xm&(#P@_>zLF*4Aanj9hfKMi_^!1s&Ie#oP{ -I5;pmcS&VaUjq~QB8MA)KF`bdc=3}(-|LDwVyl%_VPSgor@v&;-apNl~OQIhsIJ#*Rnsxf^FPyRRox_ --2F3$xbr_+@o0;NE*Q;>IJr3jzJ^dFXoO?W5TeX*Aw%GnK9J9poGgq7*^zym{Yu8+MQboED^ylJv>7p#3u2V}opXl*MG-0M}=7awSX{WcZYz(R3qU1)Ty`&z%$oD8jkZ0#1|^#W`HwJ5LQrJ~oeUM4y163JxI@g$>Waa| -`wS&%daipQG4mO-QJmSaA_;(AkyPR?sCw3tuvvDFwejZG%}8KV83bp+@IZZ7TU1TQ;-)od(nK3ndp2K -0-_f3c4cREI`1BXrw+wG$7Ryr!cM9s?inCl(g?pE6yhIC7>qp8dq!w{d=gsk1G;X8r^zVsA?XJ=9>88 -r1H<40d&>Fr8k{BHK2MpsV2k^BY_2~t><<3zV! -&Klr&aL-2>NY-$8Qlf$PHth(gm8rOOAdUCyo_xZ -a=E>4EF0pH+eSfsT91xpsuRiy0$S&DT)Sz&iPfL@Ee=f>i5@?1c4q5Hg|O4Gif0mOKO|uz@ -j!d1W{!P_)8OY^x6;a)8my_Xz{&iLvL?=_p1}@i#*rN@oej7eLOpiT+B2R7a;6-X2`Um -BZF_5gqXcXyjS>NL44r=@RW+YNIxHmm_(+W4ThZ; -dC3PJ2~CQ>4%(t#c4CA7dQ>Po>2p*qd2{h)4819!Kt0o3Qjk1x`oqioW8^9NlyRDsgCFGHJrwCDsXxS -r>i)Dyu06);RxzRNO&fr87CO6e}U1CKix&9Uwm!X=H5R -|$eG1RaVV1vU*k{>Y3#$xXP$<--;~N|7W@?fO(UsfIvGvkxgV-}9J!uYNs`;|G;%YskVG<@n5nJ+f0M -~__?ry=6uIqX^S@EV0C8Y{2Di>?{zs_(SgJpk>kyJcV-0gkPVr|ZHi)SW!dVReJKXv=lR~nh3;o$(TL -d#zcOFnX=ybGcB2Kadz9KNg-%Q~1c|25g0oW&Q`*v`fNAkHknPh@Gi{x_kbg;V^bevS*+n-GE$D_;Aq -sCv2uNqG^UTS>Qc(CwRyW^pTuZF9JgZ^ZKe@-yV<}TFuVeCD6sStCtRn)=W0`M=Jro6pLD%kecONRIt -c-fdo%yhUjY-VNwRYJ -HJ7W$8Vqo3#>`UUP+qhII`x<@x;UZ&`DWM@);raSUD2lVZC~XNA&3%*{}bAfrG9YY#1_hSk&-qM?_yYGG^51F=J!Lj -gOl!@%kI$CrwV6GWEurjHYRc=A`K}1k21>*5sRKr_7l-rEW;y5^XpMD -Hdj^`05E97c(F%2E9{hRRQB6^vEKtX0bYqJ3a5l)EiQ4MQ&Rl5!Vo4nQq0VTqV}e+*b_ClK18rptv<0 -!zwhlwI0*9%}rg}Nl&*ju^E~KakxJ=qwvepRO8*Ce$w2l-&y*I%BnKbn)s1F{$&^(WtWP-aqau>IUwI -`6)F6Mt8t%Z5uGLy!Zg=v7$oYa3b$9A}ohKgo(_{(>@U_Ook6AMix^L07s7Ev|YTn7HKxMR97b`Ba6w -peJV5K?ZS{^e8qEM=KAWHp~g)3jk;Vw;L>f`LRsUyubV2f2>s9!f%%J5JvDnb~Dd70@`M1hGEALF3|R -=h95$gso>DO=rjQ9`7#i0&pQx*(jj7wFJ^?fDr89Hj9@p^F1FZORZW`%w)M&j(W?UcX;D|MZP+hyUi= -#6TzQUUOo=D3Crn1ntNEc{vGnsG2dd{=}_aO=3iGS$nGXT#=|&7^~Xs|crnk_E9C~}1;*3cPu0I0{=x -EGNmF{|de>EW*Z%|Iq10yogW*kdhi9iHhW@Sq`wlt^KtH;wNow8n_^A3f;U9?qBI>J+tfDa=%O#7cu; -kQI>>U4bWYQJOyAVRN@ffrEUC7O`Ra`-(($-0qy!~%4PZ%pZ9f66w%~55iyMCghe>eP!f6I9*m`7`(g -ZiHBZX=d;8;)jha!i=L@fmnmEw%GJ$1nL3QCH*s|@t~PRYZ?0|(Vd_Y(wsUnqt~PLWf3Dsh%+v!^|G0Xf8h#K{U&G -Z~xq7e~ULaE&)bO}^2v-;Dn0l3(A6%`DXN_Eaml{8=zQom$T&-|*YXDOx)4K1V;}RTiU|GzkeVp3rC6 -XBUGeAkVA5mI?!lAyeo)s}SS>Z7%2I4dw&-w~R`_O8Eaq>aJrbN$y%cg!-S-&bP_#{ -EDWtL#$=jF?E46KOb0s{HH-V@bxGDKB3yAq$EH5O^gM<(4rweB|Sh#ZitOtkP-kHXP_dc%gD`uA7r?F -75sEyg9H$rPKP3>BLQwX{6GFPboz^S|8+L&{3k%kT+-5UP8v;Tj -EIxlQ?}7ihX)fVrv#n*Z+f|96M~Z(ktm-u!<={^NddkXL;e1Zu69`r!_0e3H_8 -EOB24_5B^xn>wf;=%9YEgZj4})c9njxB86^YCKYbW9n-7a3_9-O0J?0%r2A?yR*rpY(VlLCjJQ(xMr7K8dh#5wiR^t -8hFD|~fxTF{cbiVf%-+NDSxSt=4=*r8>A>(6Xi&Kg#V00EeB0gn7F^tz>P?wT|sDg&*;)-H$MPOK$o1 -3fC1+22$SK95J#&ftcb2EIne-lM;-JQ(>aU|@$hRW7=^1+0y3FJ~!9FGo%v9o6!JUvs)%s1E^jzWhau -fUL>=`6^~wK*mmh8Iyw4!Y}r_VV&=*#@!DP-wRm7;LL-Sxzx?alX29l2(8UG|z#IGh`MRGVNe78{FWk -I%MdW4)z?j%xpuBqi{L6VBQQHf_p*slj*KnuP=2)Fqutc@ -f(Y(keieAQKdFC=y7Kz!A`n&>2dLB)dYQbY(=oGuMz<(Zlke64mm)6DVM;4#ei&T%zN3!UeT@hg`&+1 -Hd1e9Jrs7LUU%y;j|Wfc~Pj>3FHZe~GtzRl4oou2YrViTeG@0=BaUq_4qc^QFRQLb%qCTQm871%(=@{79lXJu}lXe+X3X -4wpxIU=NJ#8_vMW9;I*g0Z%O6(dx~9pugMQ)(AoZw5s=3G{KA#!F}m(5TAkf>pC{eOpJRy=3%tk>ZCe2?RbG -z4w%jhRzV6q9GBBnD*oV*bGr=o@Pp^#QDggy*As-lHOdCTqjwnb_?yf8bjNL(mhKVghLdoj6+ANn2ETSt0l2b10tf -=KVV`i79|Ap7TuY^~}U10nChXS~Sl9|rylFLhXdE$KhLk3AwIJQm%ce}|x5{u_L^ayy!Sfux@pLi#}% -{ZfNSzX_mrL{#@19Wrt70X-QY_9T6?Lm>QMqK^x7_%~`Bd~h*i4fKKNK2S>n!3~($H9;B@1S!*m2a@m -!dJ;Z9w7F+va6@3V&K{8ApAv>)9e^-IpxHejEFI}LH=;p}Tiqx^zJyOypDF`o2l0)0r}c>mabV|tK5*$2p=3H!+)%{!!Lb+A2%$GOo@jk7=6wTJ1oL_Z;b=*P2sMSsu65p -oeeMfF$jPxPr^2i%}NmJW!&2=UJjfOPnie&Zt=Rre#WCuA0U_o?MT^SCG2g?xs5g#2yn-4Ir-mZRkcL -hSIp$n7_S+XtYk9#X>}5D4WEN(N*@c}(a@20R$v(5pH$MH3lFBB9(Oq1+;&3?k=JeS28AKMBv)K^X*K -8MxDBnLLH!_xt=(GED+LSA>;Q6$&01^si9y -&g_TGPXa^ce3Ks|nY!JO?=m2Q$p~eq(!(Ys5FmHQ77JH4|PZ*J!qrP%ejdki#&V)&MGpY8+m@jSxTRb -mppYz%p|9#P%DkgD_zZ?(ozn^e0}GuPfB6@C+DZXL$S!LqEs+xZ8yFFs672$4{!cSN%|BU^=w1DezTr -J2v3){^nNic#a2I>aB&gB7pQBuXnrsAqUECiN`Ho$6CC^%WW?$=`|sQ^m;HTMMGtY*Rz3vtPG)S2Rc0 -Er^frUrGzvsBV>R3IK=iF9Rzh{4jGV|0`+A!8K9X-m`j%T2n4Sg7(@n&ksu#^!T&yFpr#jLwNh;pAs? -YU`=z?;cXfAi^O8u&HBOJ*u=)=5Je#+tb0eC)>w69ixn6?rJ8pk8@cZHG%j+S8zR#|Lb`!qws(Ttfgq;mE)h!Rznhx=&?Q1Wv-&Y51d=Jt$6)2Ql-|@Za42sJ8z#e4aGZ6pAu -m*i~h&?zXC^iseI1Jk)XnO`gzcgTc-)5F~dUu{|hndh<@afukoM>AGAH*jT!jBx!>msHL{DAZ{vfj

Y)s&9a=ARv|jYUyqo6fixzArWCMIDml@SW1(K)<{Ycbb`!`26_Gt*O?qv -_l2u%s|>Mz~t+zs^;J_j$qenWJmSE?TBL{HMoL3^(ZcYlld&imy7(4P!|^gw@*qLvr*+ew4PUSJjmWf -@9(C-!_WuvypW-=M8l`)z^nJ`L>X!H)h!ke&ATypQP%;r5Q}>FDS#*6mIY?s^jaYe614>FGmyjt`7g) -g5Wa{CfKsAvMPsk2I`rAjlw$HJ}~nJGXbUDh;arwBv+)4d0uWz3Z4)dpF+Gr=tCXC!l@$fRInsJiP{F -{=uNe9t{E2es*n!Z;DTB`@Hb<@BOY1BEu)#={sDL?0v*qM()7@f9*g<7?H)+aC?C>aQ^{mUtE#~gs&&|p`; -%Y@Gk7jbrDkw287zj8!P%jZ4?W3XPn)08Mf=f+?MF9pZN#|tiwhz{MVn@5cD82dge=Wa%^e!%8_UneK -#0=~A>>-|61g_}BDr?L1#+$CXW|K!+1nJr(**Ks80%Sr-Wtb71=Zs`dNSy3@oBeA?f(7ed^`{5Kj-6F -JpPxRj~}e@J`P%?nPC$X*_7G}GwKB-Mmx`j3xp||V(tR64piQ&wKEGBvlU#jl_cWopG{rtC1fqy73gx -8mka1M{1#;9iPH-m^YRLosaV~SBvEX=mKPqX&-0o~ --nY-hf0dS1Tm=FH`2_XXcMxa!ztyz_jMo%v$kwAG?5wQxaRwk;d8p_&-uVN?R@+~2oF-N@2FJKxlo~f(tHj;p*iG>A4h55Eb(UH$=Imm5P-|2AJ3 -Pf|B1A;1atX_a89TM{ki)`GpSxIwmPO`?0ACsM*Pl!Y?nsaT-3s=yLm}j?TEg(snd7=aUS-|yVY6FUT -F)d~?Pp~d9rC6x*PkaPhrhU2t=K~*5-h|6799dIi?&rFIRh!ewY8MOJWLsRAJ30_sgG*a^+At}9i(-Hm4 -~Y7^05>^;xqH}?&_Egrh@Ew_Pm0`LMOy{V~6@>BEvCOEb>qXKBKwazK+4X0`qac%~3>EA5l9@`kZE5qh{A(Z%efWaNV(=jKDqvT03$+Cy&DOwTVuFEjI_#Ih7&m-DvGx`;)1C4&6iU=V}fUw@ofx5TtIwMJlBB;5qPdc$7v5v1G!m{msua~7O9^58_M< -Y9Ca`Lq5crAKas0VT%E|(DcsFmu0M~{1>FBzc-Z+|e>uUEg@oSa#52?m9#?;EznI6lgwuPtc_}w9W9Ei#fB*KV{g?c@KU~MmyFdK(-y83L5qqZQ*;CINy79k+uPYlle -d8+an>gJG_K!XD%`^SUx6jDW+}6GS>z=0E9OzKyoVw$bfh;<;^wjvy_CGlJMdNuw#;s#6+Z07f^fDLt -y%RI1Moy=28qetjPGdQZ=G4GxB&Rw~Tm2dRfzw7#Kj-u$ryp|Kz-c|Fb(|jHw3^emINi6{9j#&a6W=_pPOoa#7j_2c)TwC%^|&0JsRw2{-3oHlS;$7waE+d19B=_X -EBahkzt3a9a$#&R0Xse#i-PD44>aY{IC)iT=5smy63rzbhB<8(WxTR7dMZSQ|ESKB$w;8fr=o>K#-I! --V7GWs*8jhyc0bPK0{&m8EV`IXIVKX3b=X8vn=(LA?n_Zi{5yd!xz_2=ct!%Io#_tX?N#?9{2-XD&Js -uXv=TKX^jb&UJ3mHWTi{@?5V-|he3>;6B#|DC^IyFcX9;pfYLs{7VnydCFMt?!F@J+_}G|57@Je`Zcry;0Z5+IYK0X8HO5OA1o`!K}dnpJ9xKh(P}WUnuZ0(AhBI-91V}Qh+| -t8^&UwzZvM-aOls0F9v!DzR!WL2O1Iq{cE&_gacg+-_|G%$p#wT7sfSU2kGqo(3S(A0(2ov+!q3m^yF -X|I{}aMtHI1qVjyHNd?~{;FsmWtGJL7P<6^?RAuKErsNYcV6ZCb6^9f%ugo~J;l|z}GjX*~aV|*;o=} -{0C*cX5{MzOGvK0KVIcgt{!14iv5ec@VWXDiU45llZA=v2>zy`zDw6?NHx%<6QOKCKObo0M4neb%ddwz4Eo! -F?v96X+;9!0CniDsf&UQbkOZiQz()a{mcZg*2HKLq{BH&7m;z}5J0j4rQ(6AU0v&%N@Sr~d=uu67mi3@jwgV` -x$sU(73@7hbda<2o0qp7eYM+eiYER;7bJ_>9|Eu -?|`2Gv{xpK-GL7WIwh0&pTC%pZ?jlg$Utw(W@$76eFMI&VE+Knrfg>a2cS>eUb_Gu_&}R5_*MXm?w-RIpc%%WVSUDPi`mcsE0(~7&>uQ!Rq~q^m^=ksqHFrZ=zz)*K?`HMu3 -82c|Tqf5LGH4A;qXFnz__l!k?Lgb$+YUTgOGwo^h$rw{fNoe1G7S7ipkEZT_%{Oet$?{P=xc!vu7r6s -@CKkSR5CkTfu5;k;Wh&`ZGgTV^27}E(+wcA&@hu(&lK|0X^a0X$Iux0wG1pbxzb^C8gR0`%LrS$&s*j(vym#&-~^8PX1RkiJ#}^ -Elvl0G+uH@(p+^(3kc@dV${tbmRd@Kk!DNVlA|{z$0D!KI9MZ*+3iKXZ;k?Q}s;$bD&d?uzXGdiVNpD -DCKGKAfOb(|2-N*R@K%*M*ev5`u-!GxOfY$>348BN^;m?7-`~#EcZ9q -*wva&D(9dQQA0`#MS-hGCJy9TJS8QKfbN1EEq$`|Q{W+tnbfbRMkred|2aNBRrrTY=8O7i@q}1$yWLYZK~$23&+P0(~9Okr$bLyr+ZrjxK@xA -Qd@}6z|Gxhqegmi(DV+0nX!Hm~S|bRN?lK;=K^mM|uP2kWk|hjycH=te)b3w>O_#XB%5&p_J^^cxNm5dQ~gV>BLh#`91wbRM;I$d$6zdsH1<8pv1Y+%tWYGCHX3Q8eW5x`!V -8H^CmzPJJPA4nS;$rw$OdfsoQS!nIFOWTZ_K@o8YVye^pOCh;Ho}+wim7E#G@~k*++SZDO!T!YS01Cv -W6DpLii0n$T)AcqTKGw+zjREgKL^IOpw@xnCvXeul`By#AFHpgr)oOQ*1>;2DVK7n`N}oTx*x!?a*(T -69@Dj=`<$FhOuZiM*B{HtIabTnmza4z=wCuL_O={KaJrtHH5Dy(y@ -As8t7sCE*(368dNQ4PUC;ir499$>RZqcRJYAtXaNhC^1**l`@MB2{JJOVk3kZ0Fn@F>Un%MR6I)YgMf -K(K@nP2bV_;(=W|!{Li;s_sXBMcs^~ka3hHk5;cJTxKA52q}R&Y-Dz9IfcpG~cAy>zU&n%Y06D3{aTZ -M{T|(LUsF(-W-3m*WclW5(pt;Og~_zQs3z2n_c6eBsj+6V_05^J%y} -Z+AC!^R0(ZDJB~DeBld#PY>T<1?4=H%lfa$`Gg4*NOE#AnKy49x$CaG+R6B)O`FI|FTF%|?ASqN{Ns; -5Cf|Phty|9Rc+3FCi1zKf0@ZrNrOiT=kjg2K!r%oj%lZjX?7Mgc+=FB1U=g%j%+;R)K{ -r20*y6ov>$&w{x>C&ZS`SRt&;c(FQW9{0t36|(S%LJ~hz$gEQenR8kpMewD4rI6dcQ%F{ -$LKgp^kdl%TQdwC^9(dpZ^2j6que~dQuj$(QJBAwC8mfIFrHPnAN@5No(@l^|2wqdAeQ6LxlZeC+Eit -~LH00GYtK*QbK9br-mr0noAp06SLN^zV+Yt?CazTGlHLtbzI_G4bSg ->G$SiE?#NKH)@%a<>g^3Bf97VFoqm%eQ4)~zD502DGDy+iQH=>&YnFh_2S~ii{eJXCGpemBnk@)#qHa-rLLfr@!)t7%4-5vnu;8>E3@7_&&FPb4cLWV -5*?|z=tn)p7#c3-Q;JwmtHlvIr2MIPnc-a--jm@wGJH3Ne~aMkKy|={0N3mX82VMzmwq)F?=4wpJMo*82%!|7cl%S9lS?Fj&J_fJq)a+vy*wP!|Sg?UWd@(>Voo`3p& -|T(D~tlE~N-6TrKF94*q3^cV&1^U}pO;dP|T87`o@ZT^z&zW+9;V&?Jp$-0Z9-%9b5XB=*3=d90{jE~1Lfy#P23vPcXe -&p*>=`Hz!)4H7-%&3`uYckH+5^&s6qYCo$IL)0)zQ9{~ydB0t3U{8{>k`on5r^#t=i0(U*r0GWh!j_= -mra3mP8iRd};dN@%a -^&-Tftz+g!*GW8^#`!Q82)OF8nvwFC64u7T^Pw=+xX$Hyj-J3*N*Z6cX>fw7r+|=an>K_U#?N7tA<*) -CYXWY!+rfjx*Lr_#&9PLP|LAq&A+|y!V9&#VFsKJk&VZ9Dsn+M*nsnfppamrF(fF&xyS{4zEKl7en3c -|F(@Q3q|<9ouV@$Gd}Ft!-Gp^~gPcNfE!-OyX#A7oH}4R_EDi}WhDajh42u7Ze|TVXKcHabF`PPes!^ -i`Utl{Q-sZV*KaU`upgW&`ty3q7LT%f5L)G$Cy{dQy83RL@%blDh5p_Cg>fzho-@irKs^vX>!$ZPDf` -c?HZGRyn*w8A#;8D(@a(C?ncoiJzZ#y5}&ZAnT=NufWcVP^HVu;a~73tx5|F*3?TUIJt@ugrsX$W9lZ -q>?Qcz8ZMJixz4hbk4yx8aY5aEu-t_;CJk11{;&;l(l){BYjj%j#+r56^=?tYCi6SFh3|+z{g1fptv? -{rJ|_rCRlNA^yzAVBKkqA1F`r?`>NeIs_LzZPgzhKadqMIH+}dyVJVynKi-w?d?trW6>63cbX_?{GF; -3e9v$qk@IZwx$(vHVsp6(!q+aJD>+Y5`CPAFy<|R;6xrF9dwiDYBc??KiTUvcv2Ie3$X=Kv+V0_;g8f -{VVfkvbhB`7l3U99qSp>ohK?;ow-;d#kGyGJBU&8Qf7=ACq -Ka6vKPXlar%U-6qb}s#JN6^{kewTZ0CT8hBK%R<-e -q{N1T;y+)1r+UnJuJ={IqMQzt6ZXR_idsOEOs=WB(Yqje(aPxS+vfFFg4=yflO*|@Ae8G)j8oyNKMJL -CaO+22jSg~@Y^5tJ>T(zoGx#ygk)Oo&QW&Ge#&H0rI4i#KI(CSnwkMaNJSf&Pld%j-fN|oH%zUnr3wQ -OxZ%OG|5JJ0W4`{hRU+&w(p8!`Soe-n52`tI&^-6egsx#8a7x^?AOF}^D9#T3>D3CfB`-{|9@y9U?e0 -Rd<0{t#X`@vgkYr9?klil!cAcuv>LYqS1uw5Qns2EKT7;1y?QXD}5Vuaaz7sZu5OEb>q6;Ao0HHG!j5a0s2R@OQ`t4 -Cq*$VnYs8xs@L^3I(*sKC7b1o+oteji{PY=tq}wr$%G_%U7*|9$)R(dNyYiSeUNn>I1lKmYtQ9 -XxoD4jnoquLBI{2>AUx!z^9BdiA)C8#j*Qlr3oT-EXKWN&6+KzPoLhmMT-{T -onv5Npc}@0Qv9!9zfPRXq|~cduQl^9`O>9Jl#`QVeZ`vSnuxgnE8MF@dsVIcJ2E2-+#aPr=Na0&2#;Z>zK|@*xp(*4#|LhQv5le!d$iv{8a}HszW -BK1Aj$_`mDM=Evhrlw$3uQb^Y(?PlNwgUwt)`<><;fF$Fa2+__WY5B)fP{J50C*|TTmci;${U>lal+} -vE*VH=PKj*uH{2Qc4$`>n(uGGzNF$l<{Qb=R7;@%Ov@`#qdz8*cF=_}{v9i#QBP{rB?nY6v|>2mbr_@ -24$Wwh+slIA@kN4E{qNd-v{@9k}n=vq#!WUS6L33>x4Apl`4f$OyP?_>Cy?IFavnM6LD`dF&=~-$B$Y -hp5*+qKpefH*ej_Wm~EJSj=_!Z`-ylrGEYTtvQEJW`7BJNE(1Icn-dT9>@Z?<1=J`>eMOO@fkb^zkx6 -G0JH#a@EL|H(C|wl*Dr{YPY~U^cW?2N;LpC^)`7q3N*1a=jh( -;_pf7KpBI=w+WcZ%>4;okoZN4RH!8F7kBf52uXwV@d&uzBSM5*wP!=GpYF3 -=<3SR5VTJ9rKmL2sb{@VmTblL+--{JZ{Wr6J%5kq_hAo@r>!dJ~st#Xl;SsQy;%e$G^nJ7I6Um*+V6?8+<0UL*&LGNJ~>Il0)>wT7}JJVof{qJ(E(k5lzxH#IFGmhIQSWT3p>v?3K~ul1v3pnOhfh+Dowk}ec_|<|NZyhiQ`O>4jNQ -fbQC9(;^?@0_bwgUyo8pIHPXW2zBG53pG2bX&qQzhMAY*PQ3z=GL9=Jp8E<9Jd#&~ipVV-h)t;Mfwc0 -d$+~WRZx5S_2JhN4+R<7*NrvQKGi9H?Ad+4=29f}s{I`ruDf7a1wqrEA8WJmfm!AH`7_RdA3w=^0;PZ -M=xoVxs=vuDtt*>f}aB=)Hd*eBIx8k{#-eJi(##o_Pj>T1(}j`4`|f07667x6&J0``FKiU#1%evfkI4 -J7jzZ(1?BBbk_nB}@Za1pCjonT9@$$KTIcX;AhI8k9Z5CpG_u`TiAAW6;3-ciGJLzef)9U2*ucznA*Y -ww8>12sWqaP_%#s#Dhoa7Vw9kTtCsDvOefQ8BD`+rXg*F75{_+qIWMd4f#ZGUa;CTe3J3UhxQB_ls$t -6WzXFqx9#zVoguz(o@MKxL3MjNin9?#3);rqy#Tr8EdxyW?JuuQy^J!?5eqix0G4}YQbzkU1m4IvBFVXwd&xB+)X3 -w%HPJ=Qhwn}}72ZD`vjdC_L3VSQ42${yQ+EKGyFJ^!11Qc8?J&3rF_rbQWO@?Zmv9~ek8M)svs#|}yw -^f^W|Eys8i{_O9i{xd$wz#Dks2%La7>;ZTynZU>E^JBzc&6I_*}DR_@uOvKC~p>hvvum(wta7nicI&Gnj@CnTAPB!#Jh^%|6HYM{e={2l%rtNd0HMo& -s7v{P07vSS*sB($Jx30c`%EwX}0eTgqV?wlEFSo|y*tBxTR3?2{J9`_jBPKl+4eNM;&7VjBL*dNeUIP -$CMSokoosWuu`aI!clW8hp~aq*B;3(@>0`c~q -+fKFyvzdk6bYYSgF^g@=dJ^5x43aXBF&fkuoNK{IF0ly-gcgnSh10W{e2ff -Xa^-XFZ)NFQahXO&|t(Cm3O+rs1U-@0{cRnB$(uJCW#v}v0`g9gP8960bY`%x(elgUIqdi0>)y?fKgA -Ad}9=gy@$bLL2Vu)YVrpaHP}aYE67H4JEgKZPD57GjQFGXm-GXTbj!?OQhLVU97%nq$D8;gc37h5-G? -uD`s!y{CtSgiM6~u&$@@y1(`M_3JHc!^Gz(F)@+YCJ5iVcI_&2gx$M$OML+Dz!fr5dZpw6`k;GQ=OC| -xEMO0?X=%?a2faP7va{#^Y+Q;wa8YTLr(L^ta~3REAooEqw`R*UVZsCo3k#!q_3BaM#*K+{Yl#c+fNk -hIXi(jr7If&88XLZK|C0W+^`CwmpRG9t;iviL^XJo)DO0Yeq@>WKNt -15&?c0~4qoXC?d0i=W2{gbiKqqhm9rmRHuRXAS7EieU -2HS%Evo57goH+3a?9^y9(%7+Mr5r#5@DB(Gpgw*2$loCY(4abKQ0rc-#gJb^?|~_&K*Dd5!xSoB -^I{#;DtTskt0V+`!yI0}O*way*ZYkSPN9Ym!0rU -#~7_w9Q8;ldmZKbVWNc@@ZnQhv%ab+G(fjxiu<(HBM@CW;bY#)+_h=>U4)~%b=t!2xWN&LZU@Eo6kH| -zlZ2frgXh5jJdh8@5LU=tV@KA3TTz~fADbaeD%xtn(F+LbS?HJ=Kl#UJzP^B=`mbgUPF56hM~x4>P+( -jHje#QG+605J+YRxoNm5q=arh5ui(W({R!Wyu)$`RAWYIPjYJl+V8H9gdM>8TXz1+f-q#dz7|*Bl9=p -Kk)x-kIDEBd&3d9K;E!f)>D~F#Kgo%-^#Ww;~RVp^jqm6>;QHNnLtk<1NdZpK7`L07czo85YL%jyh#U -svwSIWx8bkXRlN@DbsM;WF6aaK -s+_+i$;3^XAQy`mN`?>bMTN1YO5>!~o!~>LZ-jF8OcZF7el|e>wIfLoX47K!@7zhyQ@DLp;*c0p4!du -z?mXTqyNf*)TfjR(tcvoA4bvqv{9jGe0X^w(NfgcN_ltxTw!z^tnw*GC@yGO_lJV3)h1-;0xV??*uIh -cg{~&Jf*(>34h?i_9gY7>#x6-eiuA|y+YrCKkR_@aoID(UE;6b|71U5+y4O#h@Hq^;Qycph)0N>;4#*m;63 -D^_IDu%e2346uZ17uHHEBeFwM^zcZolAwXb$Z6Qqv182fB -PTqm;?7cJd@Y@(gY4J+eeSSCo0eb-_fJb2;p03J@Th#d@QRS9kio(Mn?y`@pbb~(2hCJ0{`*ak*r4t? -O0YjR?v=BwWG6k^wo~zwIf-N4mcKlZpTGDMZ^6)6|`Pnl#nZ8zmNM~?zimLi};R;iv51zqZ;ONiqPpP#y)w6Xw*5PE=R33HtYqXrqJRmqPM>zI(kjwQW6fJwRZXP<=gV -wZaCZY9r!cOId8JRy~uv)SN2QCIX^dXZab87U_8UnoVzRM8wG>UTkB-l3q_r;gSMA@&fzm!zx;>wLezLr@5G)q>O0tzM~&@q^J5LbYtUF;FAP=dd#rOYF -Y;*cM6G#{H|ZPdWqp6K*85N^TUS`z)TdJ@JrUZ -Y43B^y05x_K(CLe524QJ&ByPH>ZCJb3a?)yIL{5!--iMDBL@a9h(B8v#K=7`><42{Ud@mGk4x6NC+cl -dAM>oWHq>~akEl_h)`>ddTi;61J!*c$abEAojvF^_sJ`B<@>v{_TW2L1 -t8MuS=(YD~~a_b;vfLCuf+M_((oufGEqv}~32X6bI;In4LcQYBkfAnm*dq>E!&pPK -sEtU_qE356lfc#}N(nHteaQzJNX94fEn2TQj`$2T?rZ((#bB)&`qEy;DCw@`^EI#>9>vKYpmb_6i)ZZ -vtMRVSgR9rcABghdLAXL!pnTF`>3GS*tIN9boNqCq+pdq&^;eSo^l0-@X=XUjv5xcTDw_^)A$ku&0T= -aix#g7e{>swFz0{(&isGu#=ptvrZp7{->z?k&{iIK0RXb;>BUGMXdLcpM!4T0v_OaI6X?%4N39vt%t?j -Qb)7b>-Of$4P*?0{HG^vj~@7vr@+%9VWYJCPfM9#0Tv*>Gg;M3!3M!7pbmo*mD>ZULbNm{K@*Z$=HF# -n7hGa^3y^wTiZhLDc~54EmV_YDFM@B;n_=g?5U&3gZ#O%$h(I$bN)|0Bm9J9cdB)TvWrz=v70W=X#Xz -l+!h{|{XNKcI`4TVHef?c5Q%a_%VGgq=P{``c<|CC!hRf_)C`>4L{tU$HIP_R{TZVM;cT#qN|)Dm;IL -3Vu5JP#;G;)W=0hZ(;gk<3EpIM$I1<7IxOx*H`YlqBb2B6-6I?^pRZO0Vm}Pl&^$*z?NwKj6>&$K -0HMTtY1o`^Le+!E*l?T=E_MK-EZ;4yZk3r4Jk@?nOsO4`y9X;2a4Lc4N)|xHXkCc*VS*J!Hs`u -TkT{9{n3{yg`r!>T4XAZS?~1UFo5^M?@VtR{g>J{D8+=%Ioth!urU@_cBifDGUl00&&a3;1pLOyIeF`}SF)bTtZT|1)RKj7UpM3sY-u$Qd{c9z2-b-QB5s_wEumweP3uL%0r -lqx=Q#OI)*NO`5Q-6-v{j|AYUTJbCg^jE6mW#C-T1)E3@(=N(yNhF#!VH1I&}@giSFO$B*7)3Epn=C; -jmzlWvYlR!L*kB^tW0r#q^eMQI%yawIy$+@|?E1qy}IX~uMUnBj0ad@a4c=hVlv}DN=xmOI|h&?A&Q+ -&#~<@}gO#b4+W=v6s5Y){oCfdh05H4f}QZQQs~-6#08bKB@I%^oQH416EtoSmHw{ky`vm@cgQCr{h(t -Vy+}?69xL`L{>5d(IPl&o~r$_7?qd&+O`--L=69Yqx*@$`d@d>+(oEIUYJc>KyB%Vd>JP>r5uoE%s5r -EnmL;Un^FuDB^c47E1w-F)Sk^L;lX!luJ)f4^B-@-Ike|DdnkR805@0eN+#RdfJX1JFw4sm&Z86`ga3 -%h_x~FHH+6Ey??ac2j$E8^`difsErC9*FvX}!>N2m*%W+P_(kiQ3Hj+^PHS3yQzSnYh{2c(wx-rZ`dl -#LiggXV@!Frl@J;R;;G1Aaz*Rlpad7K0N*QL&4F_mzC#)Z_e#Ke=YZXskD=lEWP;14S9JoXOG4HCRV4 -34w=QY@(8`d$f7RTBPYd@^}uLP(TK9CYkAO8B_y;*52B~{cw -|p2R^DX3i$n}tGBG<)w!?Jj~90O~PE{q4{_tsl)$vQWD82E5zR~l`Y&{3`@k-H;r$GQP)r*V;;RN)|L2m#uC(uq&j7z(!Ef8PP3CIqejNC1p!kn@>lmHVJyT+%#V14bppnQU*_09UJnmH4ZJZ9c!63GXhO`!eiC4@&PMFswQE-qShS=H7XND8MO -m$j61uvkg4V0zcLz~HxVTi4Z+pTU*c|Yec;q*{f31vsLp9zUfVZ!e75HmB_=b0{l@p$xR_}y2ua(!Xb -dm2}drrI3MZR^df~Z&)>0X%VEyDTVAkm9I1&bb{JAV)6e>?GKv@gpQJg~k8&p|qHb@bKfEyw@aZ@3PU -SoanqM7)R+v0^Zf7A2xZB##^`28m&OO@O?{Q+RQ2$Gw&4FE>t&92y;jfv<@XaUz1RAH�@V7`gc5fa -DXE2(J=*DM8@-+j*5dI#|zq}z5hbm6`14?SB{&nuJOLGh8N -(4m6!V)4C({Pen2%4wFE5E(4{e+%4L3z!>?3jX6VT64h8rdZ^7sSziqgvr3!h4Vg16^pal8za`WVLJ# -TTPt*n~&A@pnE8Jxbv75k=%)sa$qB?}&ti$e8}oV_XtrqGJ;}x{i#GZJ#h;NMuYzLerS20rA5Uh7B6g -bilBf_7Mp&%|N7v@fJY9WWtnA|AJtBT&g8ZAir5vrL6^y0ULGib3yG4$QjCP5}za3p85`tq#4SO##-qmGfl>Y$yfz^(#gCe -36B3*sFJstu4aB`1F&+hF}1Xphl8`1pT+e7iwC)7XGKQJ)3d*?oVLh&vo2l-HTp6QG!-&AO-m|iVCG` -&x{sOBK0xhFL*H9z%EYQ;3iG`BRbv@U6V(uSlZrcFzmpSCh>Q`(-iytMqZJ82b7jwbxYg(DDPYPHk}t -7d1+&q&Q!nUS5bDI+IiPsYKFyo@s$`5A>7cQS;zqPd#c(d=S&Gkco7%zoxB=1_AV^SkCD<~Vbrd4hSG -dA51JIn}(d6{=I9kaZ$60>G!&Cg2BTA7 -udm6Ijp%SyZPA09lUc)cS(0k_e}Rn?;`p5Zu*e)xb -(#I3F-C`>96)z`@e7h1yD-^1QY-O00;of09jkAOc!Jc4cm4Z*nhWX>)XJX -<{#RbZKlZaCyajZF}1`vgr5x3f4Y3q*9r7lBVr?t9HGv+xRqzeQl@h9@$lCiMF|>NQI>AxG($L@4Nv7 -Ku~tt-RJ0OV~GR?gTY`hGZ+lkAFT(w<>Ic&F0Yy(+8qSXww^xQc((QId2n)-28VaeRapebH8+n&h?Y}hO1uY<+-Yp(T}7K>Si{fAF6m(s&D_S%R+t6%gf8GxKuyOT76sGO{ -V31k>whLN`I@%>2=yXYO1^KM*)1)<2rwozMrOxCfKJZdsS6d+d;4%EUM&ko@@t22_3mfE2;+NgHiCNE -Lyuo(^{iZ-|4SSQe7D6H0gX{|9;=xR>|T~2e0O;N|PD(@>?P3gAqdyG2?LsSEUd7}QtGCd-FGmE2CUVzF$7!7-KHKNtqnWO|iO5a6U -Pm(>)?FRQ#HXXPLp5tEO210z>Ld=ktD^SBojW4_=;}9KHlj&C_Za{5`4D-LkmIE|*o(lB{Pp(e7147T{BO4S28N}l-HRGn9@ribE#585hEXRdgGY~$8`BDyLfA6COq -*BqESgM;WS&kYP--1!+7Zl+tV-vARe-f|Xb>)r&jLUk?i~g%wmyflkM_Ubdvmh?&HmmI0It$l*gC4hp -Q7>U?P$+dv2lV^jLN -77-ZL*ZzsDi-@f^-Q*3-%BQbvcJsLduczye?!_BWg`q$5Y`>)48{rBj%FTOk-pPnI^0L-IDv-E-nDxJ -)e>KX==|4euW40yt>t*z1DLo6610hbB53`x@}ZI)F5uVxtu%cQy^(FjaQMO2ma{h|UfLUNfvqVP8A+- -7;MDhAi-T^+2Ii?v~}mS5Ft)Y4jo-!94u#6q#rq}4ntk{pC{LGT)Y*ESMdR>>Ti9tNdCw~m7yg}F48& -Y*TcY!7fikd(oB%EAVOdj-H6C;}Xu1bGGzl3V}`38O@T10LuDnGm>}q{@=>JOu%>oL(WsvZnbnh&7$* -3yz8m1k{OOLNOs_Jn*v~q#1O!3P=QG7g@@^v@z7h=fEZH^oi$5Grcn4>T(%0ylu${AmuV|p!z(%tjgu -0jt1kcGpC5aAy$;{<9VDGvuLp0mQ8a2k5tlyoTict`0Mx%Ncf(_RunD^7)@A!fuH~_C@%sMyWzk9TBk -OccNQ8Bu7JTi!U4u7JPp^v?T$qBppg|MZw=Pk4j$KG@Hp_$5I-d%LGa@$UF69$jouA|FdRI9a8I{wlw -p20HIx{?v!CBy!G!OlC4MxIXcnDbEsJa7lky$?d57FCv#3cs5P(`HC4 -$Y)^~qplKB>)nRDvsAOtOnRz%=m4;K^v<0PB5772&^!Ebj_od@0kVC4oh_iL8su?ak`h9q_)!ammMd; ->*aGTEU!3n=@oTcRtbjI?hf(O0qeH6)vgZ0bGH~UV)x<~8#Nc7V%NiY^DXbS)X`R*gFyZ~AULk{Fz4m -q0Se>BscJ+Mq%d2v+a-r)di+_$$`46Xr5D?1#8kUQ28_8in_dg^=7PoJ;QyqP0CX4=44(3SIC2g -3>+x=n}fD{v4g -QEm`U@6o6{40O$YPqgTpJISgtrgs;&3lh%z(P$cJh -2IfSyoTV{sD4Z`!+<*A&>5#uZf3fu;-~ -4pFah{svuFZiTmh&govd0O6p9h#bMuQV`sIA{C%pdhZ@^t<94X_*o%xgldYHAK2{8hY*QIM -XhKN8g4g(6`=nlRfQrp7!V6KoD~b6`SF{?4@0$g6)`{S!Ep(4Xpu~>lS`O~SIG_P*0A(tIs4;rsRhQJ -UIn)#0SP4(%f)5A3T!GOK3ISH$>$$Ltv^GOKQiSVT>dyOgBejH<%;G5*tw+u@o9hK_emzP8&t8O-MB5 -?eXc9lJ7%_N8;;hx(9=D?FVm;M4{QbAK+&Vf>(2h$d}{bvq29OD{U9}8eb$2=GDvX^OL8!;)VfET+_L -vN1T`ctZ=sxeTP)}2VyxQGlK4EHqQF4XJ*wNMF2l7-%MP7@jza^Ezr9}c_0a%AHFHjKZSc*dDsf%7)h(;F2-^7-A*fLclTF+a{}~!=RFB#T%rqStbO-x -x1UARe9Nke>D<2-t_A2FYn|d+>(+w91^}8xd^ms2p>2Ljd-@rl;U`dqT+K|$f0*qd&t*v-&B%dsj225 -ariC?YV_yVjj;KE-}Lh67B3K<}-TBNg$&5b$*){!NnxoQUjKq5E7AH7>#)-jfep$ZJfrm=C1kBJH4?f -s?ESBY_B^tSZ^Vq4~%ps~Y3<2)_+Z95|fpisjg+bH;zU9m{N^Q6Z3sO7~)_I^m^kV%O2EtwWo3a_D}! -k2i=9WMmc;)cMJ^Lm0$5hO*=MGVPV5U>~p+BprV>Iy$wUa1C%uDSsQ%V#JHz7ymWS#etSa5|$PdzUyS -Q%x!sg}ecOmp3U#u@p^pM~X=k+`^)Zz6`INHOC}gl#9s1Z%DvDWEoUlaeLNMvA<8MK6PwFsP*$`Tp8? -jU(d!?ug*w#OFJJUH27CxYKkKhgouo+0R5o2iDYDf!dFR+{`siBqkV=n0XW28p*6*CvZ`4oFh6br1Yq -v4$+jC>+a5u?PP*48TQB7;L2axlXZGlgZ&fzTbQE{p -8io+c&!}_l_nLSX_!}O4`|i6}|ipV=nWuz^K0>O=sv1NooNzy-EtS-Y{?kAlA+DJjjz}F~vxed79wPM -nWN?af^laeU&Kj1(FE?7L|kJ;757G<3gRylLkFpWnpsjKjqJu0r}SeYDntg2!5RVF5UJf -_9s6qS83`m6H7ce1Ig+A>(-}zZa=vJW$_$tQD^GW65u-y+jWh2} -;JW6iYd8hz&FTAH&3BB@XO!mQ<}%~R64w_ve^E>(`U4aLcTvO4_0jS7=0-6{*x-KMA%o))6OO>I&RrO -Vd1Ol3$&51TfV^SjA)dKbY2$8gJ5CK|keFNT2`6Kya91!<1xPh2nZtce_UGjs$f6i8gBrS!}5nJ-{w3 -@=gI&*)1phhgY9YK$c)I2gX2CiAv0$nm0M0&859sEB_#3Z7f~8}O%uexgeX$P+M_mbvIBpu*<^FbEo3 -O;KfUE;c?72dZHuZ(;pzP>v=PJ^DY#2XBVK1+W5Y^ykU+8iih-UMAB!BoEk`tYd+$k32=NVr)W4su55 -j=Mfg=cRIKpKmQaEc%mT2<}Jjm(V?Q_5W1Tdq=ScN<8VQn7_pEOpXjfaT1E!xkTjJH-Rf#g;;c#Mb>y -f{!~tLflJU<~`~|76Vw#MUae_5HS_~;kBU$ng#|b_1nd_nni(XaJ?DzV%1iA4w@*Nd-6=g$O%h?5epk -UkK7f^T7>u75jYzx^<&$CKh03tjdRKPI+rPY%R<* -npxRluWZ&tfvPiNi2A;Vf6x@J?67CFKOQJNe`D^7_pZcVpMopog;}s8NpDEmi61Z{;9DF-TOop(5sf? -KT&4|RH8}ozXqUyu;F`?hm?8)Gh!M7E -3PYYrB>oE6!QW{hHJ<|9=9BxS{M!zeeK&~HreE?c;3Id!;V>Q`d!trYLdl>L!27JdH6kXd!tqe^O+{9 -eZm4m1Q(D1O;strZ2u;h|#8rWmlALskH-WGNK6NBfz!kOfsrEdtx-t8W -i$ehS|SXV;9rVwleF{-`aC`v3#8;MFRE;aUvcrPpL{h(g-0s(xmSJND%86U5kp>DeX$#xf&=l_I>HP~*K)d|RxteH$A6l9=kflX> -mDSQ~?Gqhabk1?ak1q6#N&PKE=`Ai!!?qo7AXUkmwnw3-dYKa4eEhalPj9-V^jK{XrGcsoeCdMeo6~f -S?VKB{W)9R5~2e16|Uj(rwx_QMJwLL~!wIjiVizgGo8rr9pb@5nHptAz4V`X@aN46%DakmDp68AK(wh -pnBtQ^29up%_24f1V?SRVkhOvp( -0hbNAlQtomNzAg3Z0v|#6Zm@oghH4(`AMJ5gxt_BX&v<;o{XTt2hYWkhdgIJ7I_bqDjgz`8uF0i=duX -8E8l<07ZNw^>P6y@$T5$0t9~D+@S*xaigtrUJr#>xIMuBlf0ZJInQ5wRp4R+1%av4;1Bz@=r~A*3$*$bQ&5)SpnrRy5T3({X&)@9VV5KgrbVkypF{}i7$<4sDEUQ^$N?G$cmUJ8CR{)eya#jqS -Arn+Uz(n@93#6(iQ_xKs12e8>xPOH9Z(Eh9YBZa2(rv*2Ik& -L3eqy;SaW_)1_IVbaVmm}0L1NH=#lJbH5FPjF!#9x%>4IYJ+K=Mz|2wVKp)tR42vg4(-B+^D&$r}I)T -^&N|jNelpahjaImocc?r{q>e^K?m=n8#>jdaP*xFsWM6b4nMg3Ln$1DMZLkds=p^tpcLeRt#F_K7LXw -pFJP1(qTqytHn=>-z|4y5oDkMF13EOrCMGHPv5!i$F-^(S80Soo`+tU*`SVsk)E9>T`dTxD`bjJIQ4u -!M5cAw+AmrxiY~+-{9WI}y%K`yC#n_F1gZj)_MA!8+0?`0!&gN0l)6QojY}lR_m|q0esS7?)gF!S6ZM|1_8z_*FvtUV`!kUQ&jDG+O88tz*Tu}nPqY#e --34>&uTOme6I0_L3Z1XYMzOYzy(C^UN6q-& -US*gnsY%8N%LN8+19Fp~ZKlSekztOq~|f5_6xprh~YFY`PL>dhz-5tSf#Qi-R3(KY)kH-@rZVcYgp_? -G3_gE0Ly3czf;)kKY?&LJ)tF3n(;btiITwExWjf9Y8mVEPA*uTl{qF;|`h~C+UNkiG6hE8D`rsr6dmZ -kbJGKZnT0WNdNe-(R3!iqXX%VYQ41^W=}eil+6Q}(P;L;{Kkq#mC9cwDam%| -0mh6Kjr$e57Aouxf!%exLTe)1;z4JDO4<@%w56zC%IY8FlEdlGeF+JhRxq@F;_;-?AP`h -h59z@SD{$k-g9iL*$*(9Ut?3=`n@)sM%J_zeP-PG9Y_zA&dHQIC8+50?3a`c=OJsO>IeJoC3@m2snJy -`fw2_*)Umc_GTfc4Wg@L;Af?5gHUIB`=>}u!XssgfI;?Z=`oJ1CW^2MyY|)y)yyAjf8PtFO3B*Nt!T$>%NS3tZ^4*xwhyIeo`Dp1Aa4%0=>TCNJiMFk8wFp3IB=!BK`_fQolt?(KfJ_x0rE&hg6*RGUcS&;(5qFGQEKr(Fw@fq|M?23HOBpbg0&9R^RH* -zHnQ88!|m!sb3*!r^=sj7ZUmt~_Q4vTt?$tJXP8M}6sg{vK?gn4y=+6f>iWJb~XoOb)*560I -BJ$Fa7Cies5oE?%lMuPP4-A=HbI-m+rMt-uXMPds{=HLi2qbZzZY3GrrwvL|7{!+>3O7s}C1XX&{WU0 -2HbeP*)88Z6Zv#4cY9yR+C9%FbrMJ^ioBJjcC1mjd@r8g0$ssq?6dr+)!Nu~A|!P?6z;V~Nm -6~vqbGQ;v2ful=exWiIT&dUzs8N8i-I^j{&-uEUd$0s;6ok;5kPio -NP6n3;L;Z5bN$YqpFOn>0wktMhL|3x4ihixj9JoYc6*zikRs#RcCWZtOII_=P{joi{=EC}M1ng7NO%1LYx70axnS6o*DR?j*^ORM@*i@Is%aa-zQIrNoSl1h)N>+QBdw{ -L480%?Yp3h*O+~D;8S{5Z@W-49(`PNS7*IRMP)Je|~QRdiaGw(md4@qB21BF67B=a- -HtYr?7rHXvFp#Lk&>Z2;ZkAWu=<0X6U}TK#h*98QsI^eUEx&4!x=*gz1i}UL7Y -BG_EFiv|0xa@e|dR>S)9Lu^^!Q0{E{xOq$q4?(8dF%$P_g?mI=% -gP5RJD_<1_sBWJTTRbS$;0M_-OlZ#T|9jz(DTt7r)C;G6seK=}J?@Kp~yz#NyS`1`%1zaJd$4XkrjX2 ->l&IHzQ93P<1;f}%=X!%B%~$E|WoAi6o!6lYj23fy-dha{>^Am||p&$`baw#Tc|`q&`5DQ!HTH4I_Fy -mXVx)V%uvBZpv}gVB+giDMPr2*iUo|9(=z8nrzHKS<+1?jza3icB~@nRW^%9vtpFL?r0!1E -7UD;v;A!7(;lGu>_tAJ|ht!&AQbmb+AS!(|C`uXOAJG#2ig-WDoY)#M>zmRh -n4$ivM)LiaYt>#{hL5UMqIM3}E8w}?la5| -)8h)0Vv9flGIZ)*uNVWt9A;3h*;58>l;YE?tzt0aumbLjmIV?iX8qstiv?`TjGEOSr -!)*>k;fF0JdS@x8vcV0UbI`*mDvNemIg%EHwcg3&=Rg0<9P;zeop{c^qY7G1wRov2=8k32@P1`z5<9e -y>Ux&U)|^J--Xv8Wh5toiv0$y@wrl8Hp)g$za!)`TqMQ{U3@{n;`&!XM_=sWM5(5Mwd0<0|dPozxu^? -nCA9M;877S%?lL9Ao8Q5z3dgbhG+pT3yd$*=uDES>L7K;bZIX9lByck=xlYm0w@wjvI-`%)6Hs<ZftCXO5gyT9jIGQIUs*Z+w4F8%)#Uc%v;~A4i93&I^^Pz}3o2fH6YleQWc9epTq3Fkf>Wn(v%{xmJavvFoy2+J^`Jp5g^r -j~jbj|rR%vNeEVKz)7x^f0)MPwxtMTsc7J*LRM<&bJKqiy&%lI2{>aPNrX#Lv;!L0{rZ6tKT_vMvL;R -0M)=)fXwZsGI$5d&BKloy0 -xIu|uBoP?;dD_gi=!ekq>O=s`a0g8a}-D<24Qw#(V0|NjQ7w@244mo|Irp2P!>ns%$&zrrPq$rdYUZK -PK_!&Jr9S0-Is2E@KaEKC<&l{;6TN#TkM!mt$hx8F%V -}`c8=~VXd9=0XrH4rDKCeAh|VWq^{=KbsDUl+`Jm*dmhK;K8WdYZ#PMdF -DB#J?Q>bXpIv;^4oW`}XL(BruFhRv9gv8M+f-j6~Ml1@jap<&1SH(8ch;j%GJ!rpjd$JTt$+XK8#H!; -3?XULoDi4rK;xoio06RRhe7~5w5HAT?rZsJ0CO2TW -QHmvJIcyJ!>C*DF{SC5lEZc?}judKt8--Bqc@lD#6@g(YZ}LcO7>7@~8)LtX+LrYCT=hW2xlcCvhW8> -Xid?MBTq9V)x@cs>jXqlzX!Q~)s?L>ds9CA^qS|R>d!v=KIkg1keVQkgd*Orv*@>snZRxBveNi5YqFq -p4k8$;H#J=)LaFg;KnFvyG57%dO+kx0fD0>Q?!KNEFQJUc0tHjb1;c3(^^9dG=mR0U@vcx1>hWom@*E -X2beay;xf#D%byePQDcO333I@e1@|y*9#71ZQo9@(^3ReAmGUs810yxiCx^}KJ?G$WvDd1*w;Nn~<8ggYPg*E{uuFt(Ux!qr#V-0f+87J0*(yz3B%6a4f;$ExFtCP`=a>DeDYI2hZ -l^XUf@C>hZbc2Mz9R7q+DMg9>M~G)aPNd#!~Coi#LH@!(MCymTeZ;3I1q2D&JmJIcYep4@e-Va2@$bH -MCJA>O>=qfxg*#KDxt)={^-4ZD3^i!$E(PSK$+Uc*=5lJ5d>>@_GS-a!r@**ONmxGmu)EXkdN8!pYn? -{fQS3W@#7ppW*<2dW@XF?eoSYyorD;m>Z3vQMxPG)Z#rr-z#ZZ(uc-Gp&{;`UzKKPZ-2eeYH}NrrYtS -fk&*YdHKj{HgH3D4NfrUa>R+> -$EjbgkDH?K)wsDKgpqor0Vh;6Nn7GQK}7_!;Cerv-d#19*Iu56Kay^!D4w1<8rlGFV8WF&bboQ_X$Tg -sBWSGRC4l0*F*K1Cro-nTh49Kq3jc^nl?K#x)NU3;p9>U1M3Q0dTiL!<%Ln$Dcb_&3PvNc|2;#!KSG- -vBWQ5<1GmdzI;?V9P-7^Kgt%bhbeiUVZwEhIab_fx0Q3mi73ME}Z8|t5MPI`t`{}@VTY-XcRj`3W0xQ -b`R8c|U=0KX_nmeUVDMA_N03tjYQyEGwi>h`sZb6-`Qnu@N6jrGLxE7$$s?+oOotheLLk~G^qU+`9MP -t>~JS$S5hi!W}^@x+r|6ehy5*PoeA@1aI2;YtnC4!C -6JW}7R*5s?Jb{VhDyGyV5q!I14DjTl81unC5q!8rQ-RcV++$OT)d*bJbMcsklhX_ngd4g}?`W6rkG0a -%jybmCKCfu9M -BscE$+LkS{?FC?H7>tRc4ayAGFm;zkR4%uJ~x-Pn7R%T>F6)#L%7j5WR_E_&tqt;PqU=f}l<*epAY-Z|`RIrZ-;3n0!lq0IAjxpFrfepCir7YSk2JI^4)(1MyF4MYmK%FePS*n?r; -W|sP(Vz(H!pg6J%LGF29@<(kU^%nJ_dAel8!}PwgCKLc+I79jdX0yIvb63r -t1=h22hpTclES$EDELj!vH3JNd24)q_K-HYyMMk-f`e^(6m!G}6`ZM=-&LK}M>ow!BYNwbB1W3o2+It -a{|B6kp)&pZW`~Sg(t2Bq5b06IJ{Z-G{8ba%915&qkA|A%Ch@SW5dqD@_`CFuqvq~)hjbOF$Go|bP-| -&jqp3=F>Cz3IbQc_vVs&5i1ghM9878&Xx=c|2G3;3h6jmUV>|@4YT`p8oMflncn2PXmE2Vh2Pz+d)d3 --vLNp!hdC6rM5B;|tYIghH4bIbU9|F*a;%G*L|yo)nl^Oxyf6R6cEu1st)BG#GJBK -A7j)P%V{w$Hmvl%;G}mdrflOUAUvtLr-avn^~hZF1$XdF6m}148E2lt#Tu!r#@NHa$c3f!g=TiK^(^v -2a97Oy1b`4QH8L@Ym_8qv2vecQhMIka!=`UWeor7Du;?hAB`BgUz*w*n8d!T$|>u#Q?AD@O(hShAMkx -Z@#B$gv{EvyBF||t;u#7-VrlI=yvF0CBipFJ-#P~y>!vD;%L&lGP**%Siq(36`$F5Bo6yhO>^fVdvn+ -|r>T(l*ZJs36Ka=9|v(G3(4q)YX3L0MZQ_dHlje^2I(c!S|G%jZ8`w1OlqHR0#O3%8%9Msml%gbAI4? -0>L0kI#M(^x;etdhl*y)?8hTS^G(QPEBBvW`<;4xlV3s{^*o3{PI{S*kDdnxo?htU8Qmrj1-EYR_(bG -R8XSda)?s-fXJ+{jk#R%i!Zt@YInzKw8C-hH~2Wa~Iek!>v@kM%Z)DpX5)M#WTZ5ZASP5Z6OQHxT|cn -%NF@NRAH}*)s_|ih$5w5P=HCID_kYKFL9S*5^NBBHyN19ssB`|%{sbOznJL^=R7MS`{A%Hq=9+P){|@ -7rUwpfU+37xqD!~g*lDU!LM4hKS9=&nn)fa+HP-_@*8@G*gXnm7_fK@5=Z4vHhd=i(f}gQ02d(qXp=> -U|I_mW4GA$_J7QIP;mb=8d2#w5{*`=gMvN>`_eS9~wnV)tfMlvH&2k=wP(`xHPUCC?~gFxr7TYh}-?y^yMIL;*%_EPLmAW**xGc*2*vq2<&2r)>3z(!zC)SAG-aElP;=Pk{**x*xT`Ff~CpApF(uA{&S -D0^uUqjagHJv(|@gK2mE_jb&CgJ`rkS;aD0Mi9*dF<>KJ?v;D;b{?? -XLbhS3VK^$ka9*ToUAx4LhqMsK{?#vpUYq1pL6)<_A191uAQUZm;2xEP4<2~*?WUgmB*M>AdI1OeEAD -qx`SWOe|e_9(_etdzg3xY{tHlbfa+3kN+bC@Y2(WALRH+f9mAotc?2SBTByeCW% -qnjC#ERO6E}=ccYxswP@V^nQIqjD3>Zs8?uWrO#Xtd@8+{O_gPznFXrn7Wx~9KR+(Q5<3eR@#e=#=fC -^&!Pza;R?H4jc&Pf{{oNm}Q+D*=ww&#JOmOY3ax)=B%%ux_xsn#;j?*nSGB>e4%i*K!0lI%3Bw#_#WY -f*Fl`N}5Gdz(Z64{ehtSbr<<(%*{NLY@8*w}u$@sx=CSBOY_HWqz?V*p^d9GOTtFTITxSY=O5Scq9V! -p<}-HWSHlxa7GF&o% -gc+!3!bOu=VRzP!NOyD!tKT-E{k+ -=m=02ZEmm!>CP{5M!|*P3kPaqrorml6uY`P^l~ebbWGc)CA_)4pfZntenD1#aG0lFXZ_q2Swg(s(E$Q -%=7gY%MEzgm`*I?{$b==ejPEnJBb|egLi-}3q5b)uT&b6_l;#D=Ps9U>Q+FSTcz9^3f#<)VbO@=I2pM -PM=U?Z$Y=^Hes?=T!|3w{h{D%J*6yLtIiTu%L*@DA84QhnKlL=@E(Do_YWiwRp&fmR>o37r0>konRH9^hW -0qslBN#PuidWEMFPJVHwjS4aEWsSqe3MSzXS;0Ci!)z3uZ+n?Ryg4Uo7+74I$*VY{B0$78b`6W9->>H -%lQ&aTj8m_W}Gx-H(8nuxj6btVH&#!8W{v=XG$NRA`30eY1bUmIo?Z-{C*6@t<$;pA-D&Fl-~Y52dyp ->bDdP29~wAZ@znT@WY$1H^zBRIi5#bcr(I#{PuKftGlzt@09wB8oWIDawHI1cPmSmSu45`A3`ecUJyrVB -JjcxjG6?NOBHz*rG`M4nY5I%mn{rL6v<6{v0^yZT8XVaY=c=qd7xgPgUb{@`HIwNZMHN3MeZkWa{A;q{D&X>e%1nby?^|A=LDKO-aB!t|1|# -T*Rv<5zyEgnE4ADvSja8`cXB9C@LOwK=%dm2_}p~4s~qPGoDX&IB6_=hOYeSY^|ziGB#!YL+zah@y99(k8S4?V6aID$XK2@&I5Hf_ad!z*)!qJJ}e1VpCS(?uX485|26(y5+(k9VT -sCMzt_%^ei5$QIXsiU*qhE1AP$t;^Dx$Nr@NJSFB&C2;DIJY^m!?4?UyBf!xrC7J+uiBy#IvHO2md@K -$i3_@acm+iH7U!H~L|@_ku+pcePI1K?{cif~sQ;Fi_AI0fyX9421O!(f_u+?`;i7 -;EjVo*|QEGHXngluSP_YIQy*xQNJbuY}9MB2#_oPnEXK3$(_?cXlb;CZ^U6`+&|Cfb~EcxsFZ$OoIb# -Ry5;Cho*n8>sarfuzdlR(46Svg04w7{W1?6+`$Md%C7v-8+_ZqQ+xOr|N{T0jcCj-kr*!4H*NksOaym -K=>yyhx23L3uxbF%x?afsyl)Bdjjk#>{kQhIU`Tknx4afc)VhZpIH_ -ff6>VUS4Pt`H^eRzjk#&6;h3k|rB(=M>;{G{?BHY#AIL)d%bu}gSad+?N1azbS_jdni@9Wh_OJ$V^wT -`9<;yAL}?#d^VS~M6hcaA5=Z}wNPP21UV2&+9bti=i4BBvSHW*C#(6y~>8vY7Bxi11I}2`HC_t#xnCY -mW)%+gZSmEd?Wu8*W&EX;WbfRvXDgX%Ax{R@Z)hftC1U8lGr@~jvp6rY#-R_@|3YgsIQ(vw){7LI(VZ -jo?rqCB|rKPXyV7+7*pF~vLn_eOiqGU%D1UQKp<5Z>w6d6mCMj}r_k|6s~V33Zo!rl+;-=UJ1}`MEj? -z4Po~Y1piu`qvG55);Joa#UCCR0)Qnsf5eGXpFg7N9sZl*UPp=ZNsi-s_5^z@mP_UM`btAF(Sa?Wkol -b3X@Ez@u;*KfiiIi*1)RkBkJgD~sZMG-6^NjCp8s@pCB-r;wom`}@azce|c%-fn?w~i+wi*C*BpIgzR -*AZlj6UPrJKbAgrdMT&304Ry>LbP>R2h^p>}0MQ5fsxOW_l_`yj- -o7&WUnKPA@oEAT;HhIMl*#d}-S;)X_V}d2;+oMJ8lM#e`jT*_Y9SZE#m -coB7ZMEU$#7LL59al`VZjPQXR^>e6?8OGfWamxi!bs{08YKIPNmm#C08!5kBWToUuafFxk_PI(iT*fE -IG_X%!K)|kZdby95X7aJ#6>4pG&oy}&vclrKKVnPTK#vY{wg|3hl^2sO9=^6k^k%UJa=1>5J5_N4NQ> -~4Q*RNFg3TK#-lOpf$ehDxPmyZmA;An-3il2ZYgQ36y(Fx*n^!Sn$?q#xmnE{D{~&Om;Zea|ljG%{&>EzKMnU8olM#f~i;EIJW~C|C -%^v(6Yy*bzGx=gGpkd;Gn=;Nb!IcdzwNsFhenJd -lMH}1qGn5RW`cH3i|v_iAfkGGT4B?0okew5Q4l!l2!BqE6oC`qziShQ!f}jqXl*CP!E;bq^!r^qXWc_ -?30ukB=vQZ;VjKDkU6cwZL(x=18k_o3_Jkyxl*}N+svRe$asq8rj4)*>2e4R8dsp!1Qh(x)-f>C&f$I -#VO%=vx2V*&KyX-?@|cuMK{_o$5;Vd}JX?M@C2xucjl21HyrE!F`n@q*o~VnbTp3+s;^w&#FRQ70#0i -~+&MVdckN8K6k|C4%=v$)swU&KjRlbwn&Nm&t9JwbBm@#q8YJE8$x!oCTcb!$u_h+!?A|9!w>pHM|A` -Er$cKWM6e9hx186I}vU{@xADmv6)SK#kh11g2;_LYE!qpl -&5ozoB`Q6>cBhC_Z@Lyfa8pg?PG?{zGHjdJBP8SbmuUBF=xRK73HqGe-@yu>QjU~9Dv2cTX)ncJX(!9 -|9I%AbLsL{nY87t6L9p9;i~+Z0MzgSQ15_!@L&-@87Eo^dh)*$PU@H|mUfg|3Kp(>G>kPpkP@OEP?F( -`>~bmD5Igr2$gmrQ$bD}X-6nkYr(t!TF8pxI$8N!axB58DiYYIK@I}4ok4v9Y6P&|&i*=<^oO$ogH)W}&LV@&rUDgkr%MqmbX?0+dJ&Zl8`vyU4udGC`(f$KyKi^D4gN%xyc -|Sh1v=wfbTf)ba56VgWTV{>4v}BzgwcR6{#+$qwCmTULis^1NawN(_d}V^R^ZaY7+;_9CLxm-4#htyNmCMNjKZ?J%VkCBtL13dPahP!INCg6y%j4lTRThBNvC5FEV -YJ3*j34Z7Vk*kzQMb5o2ZYe|U$$7oY+8{EWfkUAgT@Up*aJx7dgLW+WAvQ~RPYCbm+PLajJ1j_!5M)o -ntZDxa2wxAj@FX8BByS$QR>|^wUJBmeR`Hod+ozk=fz&LaKwKbfEu^l~JD*x%OrbFpv2F-ArmMn&<}6 ->+D|6I$fsPB5!Z%mW*?Tf)s%{zLSh~e&J=^Bj*4EVZs#>7yRJ%}pdvg;h-;~obkIjv)&gB8gc^$dX@u -TvM_N9M!5{;5cDvB*>8}y#k%%L{PMry@Hf5wHQ2XyaKx#%U7aw9FzeA{wLWt;(qPPNi&_^S21+fHFpb_s%n9D) -fp)gKHcfXrl>RA$1SybaWJs(1(xk;gK~xeph_Z{;*+r=%NbhNn&7rD?!Ri!qT#^=8H_(Bee9cDJEmG( -LE$K{3ww%|Ki>KDs5}6W6jL7TF^*D07(dI1Zq=S9lJhMA^$EO;j2P}{(XDD~hzX%(Yj4zT{|A15xXu6OdT{&3YE-sI-hwh6-;N+@@0HTf`bY}lfa}S^X2~G9z6$Cy3e~SQ} -Yy8)hir&ye{yBzTp{mSTRP1fyv2@M-gU5nUOa!Lp4b_8H2> -OmY=tkSD?6-eK@!>$#F4>w$T-m2ql+mfmbG(nXE`Y-~Y|&2`=QY=iPsJTefh(iqqZ)V7UDeD*P4yeh| -v)t!lSC8xev4lmBCrkd5E7waV0)`M?VDnT$Y@Y0bwkV0%Ypid3YrNeqd%jV(T;l8$VI6^Q6DEcN!^6) -W$Shqo9T{i%#MwO~aW4u$c%FdSv$ybuKLN1GKAMNfkbnG2)wOJoJy{;%^m5pY`nP+-IC(Z4g-e2#1v- -9>9hDjbC9PghT9Q`9Ot|9}K7nb=L=|}^{mzm1N+-=pJSFf=a7V9@oFA7Q1w5)P>0mif@tULqA)t#z)9rfmP1PGHowt&UW7hoKWM -j4Ywd7urLr;-lo24f;d+?JS;R>FEBnm^pHW6OXpGk$Tcyj6=<&904Q+eO}i*ZNi#v%EQT_?pWYBA5SV --N}DNfTTE#0t8H_Leh?_QAq72as{v!oU)8m$oD7cRjIzRRQJwf8cYCJL)HN~JP1wzJfJd8jbLb=UFX| -Zj-l#K6Lm^CqX(dWcAMxD<$iRN**oDlm)pYlff$MeA?^i=?N-0?P`5kb2NJukoU(gr7FQct94%N-_;h -8<`TpvbU7jQ!8&Chzk;q6Ht(B&7s7g@0Bo!$*K#i4^y6t=TdMICo%BuE%6nB?Sk37wsjL25qSQH_0fv6{7Q#dHGG|6=EtGp^wBSc -x>tU{;RJo4{)U+`eR>!J%7Lh(Mq%}qgm`qwqEoh+2*n8TuiaiM7u3wWdkd`;{@)DHV`<9e4pGi^!kCT -ci??Cj4?EX1qcCvRv$ymjG(#iz!LEc|iJ9xlgwxspKY$fORJ%ZIQxmx+;udiSC)i-FdZS=6Gw6k*b!b -JJI?UBEo@@7VIuM^KobEBv+3{=0YBh4nyTkZ+TcIOeEsN@vUw-Yyl5!R$wEu$ETL_vVOw%Wmhbdx -?JNtDWkHdlgmwzGiJJOAZl+d6 -iPq*S^E8xpfhfM7nW8&r+7}Ef^xlNPnTE`oC;A)zx+u_i2#LiYc9ji7|YAZdRsvfteb|_i~{S*0}x;} -TtH!LpkQY%$sL!6USj{4S#)-rMS`pY#Mp0a#vY#K&jN@sSmVwx{!g5<%mxFZ0Z%EOrO?T`e9Qo=RRDh -6do!XkOb91?XBZHO=~#oZ5H$W8q?a-Ne3cbT2`sw?K1g$oS -B7Vg2n9bmjcjBuKL6BylH`l4B#6G!ld>+Mk*wZ$7Js@? -FN;r>+{O>UnV6oDZX{7bSD5*yovpW7v60MYpMDZVpMG+lHM$P;WastQ_&m8zg7=?)I{EaIVX$8`X^vm -^UVr+@$uLNpc6$#%J%0wpp)HCM!*rW!$Z<%Ab@E5eJoQjq6sZ8MMzy|O{I;UP~Iok;^!DDHP|2TPh@MiMz;P_;6_-g0mn}eg*YmyltEpD>PymUfwn -0__x0U!zfjln^a92JW9l==hLM0F~=hs-|@NkznI(`TVPuSu&1>z0NK}?!@kta7GzG-T|Ei%b-20h{Bf@8g_G*5Y7kB1kN5nw8 -mbifWWkyxrZxjEaChIq$3^2C(j@Qpk{xEUH5Z~1T|h2R4IsQ{}N3Em_hAq%V>A_$}g23^-zYdoJBSwe -*}$O`kpvI&F$H3^!5aS_EsFb_uRTU#u^Vw2FVxS`B{;sI=7y#0bMj0j1@h2~K2P|jXHi9>RClSY-C=h -PhCdrs*Oc#i_QwXtDpd(S`r)EU_h12KOD$A>#d%mrV$T(I`I-gsQEVR9oilxoPhFLF7y)f-bRbZuAJ)VtV3m|Ue=%p&}_P*(6HJ -*N(w7g!CRQWK70DfXP_ -CWz{fc4=gIpz`^C6-XFa&U6|T9eScxbxvtuz!s+o+?hR}CIL9XyrL5&FwwSr-B%5&zFBs005bixTdo2 -ZO1VzS^PEsGouHQ$M0Vzu-sw=})Ggyq6$tYs~EWdP16Rocn2bvol68+b~IuvPl9SCjXE@YXMwI-E**Ewcu@64#WFwSga=(2LAcAwsJA_CiY-d9lx1})uN(P&~Ao5lgY -h%_jA#a}0|pwyUnp%1Faoi>%-2K^VBBFa`?C?riT9Lh7$DYi|uN1lV -Wp`ViyGko;BVUxJ#h_l#RnwWIJW6Dqw(o!bOel&+Byz9#grRW;Gx$yn*?(457(7VaWwylOeJ2n??YPk0 -&4Y%K2_~YB}x#wQK@Lx6-HVF3=-f>T1#m%0=yYJnw@yhJ%%px7fakgpE6L)-lck(aovMqN%4(~?>Tz9 -`j-<$6KHGNmy{TaNU8rOLDFDUFccb}p0P92`t-<#?CmOFmzM|}UN?rs;uEH_w~tn9DB%?qf`Jn8jzS|OJDVvW!lvG-7yb9t0Mvn^G%y36AE@9n2Ij5|s -LFAMDfsU?q5q=3f_gR?{xSHK{KjoOynl39uMoD?d?8FB!`yad!-m`W+Zkrf6?)sRgumA>goAROS88C) -y6Z5;pyB-c_u#@;Hf(PAF?`duLYtXY@HQI0kFR0lUH1au6i%Z7#pnGV+>(FX!vFu@|C4{K -A&0E#!WaR$c@!5-+U|5(O9OSR{`ys`v<}HV$Jz4@k#eXo1VFHZb2HousV^)C;Flwig0gyI;w1?|zqu) -oI7>juO7{rMmLj2f&gSffR%reaG7UN>Z{%6UR5F1%Wm>+W4+^S#=1rtN6>Q_Ok`H{2T<{i -3}sYZiSXleHo+V3^CA~hzY#(wf!D8}C5 -h;ffc{M`Vtl9WI*14f@JB!~lAkMDbDFwB#Zi&zr0rbW`~KFJk##2QUfV;msk2$Z7%`^_-o0tOGccM8L -J}9N6k7wpPmY#L%r+$AJh{5ZRNWn5=2jC`fd2{cBZ -e&2N;-m#@nLIrPo{KgpBg?~s5qISF0Y%U0seY9>)=9&0IJkOA6TT?er1)clj?q>Jse=%SoQuQijd-34 -Y=euR5vy?Oc5PAPstjpn4u~;^|I|{#8~eb$ER00qMN@~Op>b~k=$_^WOVh{tgq|P`qkC80{Xn1$Z)H9RyuB-Iu93kefB{w{8RHb90v*D^@dzM0L4QfpsJ+=>Zj)UwWPp#g!KB?Km+`n6e=(_(I -h$U^=TRJI*yIY+XgS4pwKWiZ;eK(DPr6xqz$@QlfVTG$Z?RC4{5bgS|@~GS&yx0xEH|h#PzU6+!B)+Zs$V>TpKj@^id3|K -M4aLxlFgPLWI2;P$LO0oi;eH85*9eDEPB(dsj$bRVps1?6%`w8YWWuO$#JLungvAkz9O7Ee!j$0X=UT -pkH#WxxLZzh}Bip!^E_lnO6+k{uU{F07I4gI-a{x%8o`i8u9^*1N8P&Wwi<3CN?HiJ`+fde#Jw;^-v1OS5A11zihd{U5{1tR7-k)Ki4 -J`wRZ0Z;Bh_FqYkoKw}BnozJ0fZTyihmrrifef`2lhwQujURKwLkO@*umJlhAXxm50}3~3hTP5BIDD8 -A?Ek=~F?n$W`Cp=NH2@c=2nC-4BtAE}>Q*|^wJ;2Z1pqk|Y=;1o#xPtknCh?#`YYItEjj6tf<5p-+7q -f8GZ(!_OBjj+@9aYWFZm-R_y=foH$gR|F%>auLZxfS>+l2|AgJAbivh7N1i-3t6MS#Y)>G8H+mK)e{}s8YysSg&KTex_Y6z(^ -f0zy>wX~1QD&#%isOT!7)uddIjG9CHk79Uo2)RVv*GE|i^@L+6fq=#2@--mBUhM)f4@#*IlNgnn1E55?YpB|(Q7h^`qf#5j^9!OT@i+b^TMjqlPa|hBeVP!G_Dr -BrLOB24(6BZ;3VC~_0DGgG6)nt+>Zpyk@?MaN9xQph4Mn(5&l5Bp!Bw -W@?v4G&bU%_v5VP{9D_kAH=&elsJfW6h`+ADBLq&rc!%9V7p1L}1aH1(m}#U5YjQ{t-rpF*TajPes0uy=%>#*thvS(mhFYAD%_gQU -wI4xU1i_$f*LJ^*YQ60sQjV2%8yqib{Cu5nG@W%Jgi?=t!H#g6udZxo~aEo2+SY)v%VtC+3E7YYvGIB -H8{iqCjG0oLDp9*5d|Y6aNCAi~dJp5N0v4-IJD*CJphkf8hUU0?PzzEZqou_alC^3rtS3)+pd{9S+qc -)?Asr)39>>$T2$kJB?`nud^gFn17|5N}}2KhFZa%jlB-mRku1$)j>OsfD>=nw|vSjcUxmsv4vtPgewXN!*>kIVfst#zmhQfx -Hk1f)cc$E9E=73-&eFja1sDYsIL`84H}hSAh+3SrKGF~ncZddaiE-=mE6ETz|zf5&O$Jvy13rcbN`m^ -I&^R_7s4WFa+5JmfxzK1Ti%NPf(bjl)SWo{9GU222I6`E?W)w+V-JIvb9@Yiz$B7O{sam~i -~$vxA;ZwA_OS8*KnPZ9R)GET5fH3wk@kIpaI?@{VIgNW*+z4Mmc=L?U|7|cIfO#;;}p8UNXXUi_G4wp -gO}1?Q+9(IZG(Dqg&4L;uAZ0+Z3dNOp!jw5kpF`GQ4r|l5O6_FR8~>zlLZ#!^#qaziGhiAQVDTGRpmC -N$lt`ECTIh5lZ@AOwlrW%GqfX}oP^rS3@CsU4d^EI@@lKX);r$d$3s_e2UzUi{`xW-4{}=v^YDfGl3zF4YG?%qetysPW%UNH?Pxa#$vFIgPL3K2u -OvP9vVW808jqLX9Oz0|>)F25^`-Hi%%wpqUilS=q9xHtAa@l^v!dq;#g7WnY_GIaDz$2GFhWf}!xl5| -;1U=QSmR3PDM8u`W_oUj>Qe&6+jTycj!wol)V{;j%ES>)f0*ln^FdKOpqNH^7T@4>bB~g^so)$n??&5 -P0Pi?2<>Y`ZX+D_&9n5*Ymph0tE&kOL{JPsIAW^_Whdd$k_F<19X6rlT{4$wUc&`kj}b}mGFpwY$%b7 -QpCxF84QxDN(l4dre$ej%M_A$BdsM#=t5fc0s&0#VVtA%CPR57ydLEE;6w$z&{8XaGngEsQ~kAV1ZEG -!OaeYHBHwYf(@3t2oU62o*f0u}>q*@KFQt2!iwj&-h#11+)8HHT2YLgKPF({cb3z(f|OazPYsF~ -3S&}RdE%~$5KaV^I*1rWsZ6!P?plu-N}^uU8av7NooBS51ysPZ3sRUtQ2-A|W|z|s=8kBCZ`}&hM=QA;<1#`A@2jG8Oaipm+!PN2Ta&=t1ebf2%Ydexcq7b^4@e23D)Qc@`#)eRlI*GhEJ=M}g -hTc-;$8xn51kR_$-$ILWE^1LHw5XFiVEA5(BD6(NuM=3WPCiQOb+kC5ObF;Q8aYLz{Ao*ak_R>L;qI; -XdVOL!@*pGRlk2X6;WC{xUjed<@@(ikDz;AXW;v>uK$y5kS@u_`D8+-SaB09=S7s@KNIu&`LbIue5`G -60!CTXSFG@%oeWCiP!Rek>lvo09HlA4TZpL~Whs_L -CZHuZI-}6`IrDH3zTc%M8R`HVGvj4zw%V` -b`kFgp00WDY0f5dOY4im-o^Gwp94QAb90n9hmcs`huVA3tFXhGxE7zA*vNv#wG2$&p?gd|KhtsI;#SE -m7A-LrX^1|2?l%}vU^h0t{iJ0femrW=Qsn3+qo%8WoT&{!}QKGP(kk{ -j_Z;C94@vYU9-mC2P(>p?92rveXiUF<^VqxR3Gsij^A@Qej|nT64iFg@`bOF!Jm(NW -OE`q*p2i-)fjyv82v9Wy$j+)DHXx=vkm8{`-mvAqKWhQe5_dq_UxOzXff9Y8J&X> -X-LGHWaZnH7Oa}k(`-z+4ibBcES$1@=;6)xue3b!b&Syn<@(Mg9NmCeR87>h%415V1Ah*Hs&&XF6gF{ -$oR=A0ln6YLWl{_^KMc>?{E!(O%=S9& -egfSui^aDXjOUa6HU*WR7sJO-%dhWgXYTgK8jE)tc7x{7^L@*VC!wsVr)tum{iQSpyP%op}j8&J!OK8 -zp6~gEw~vp2^T0(<{Z$=511D#(@>)e7(y=AjT>{%YRaJL5?o -TlziR8ONXHnlCFGHYiK%A2iYw#oDwOK9jmrN$`7kNz!REd>rVG}4VDk4NDOkg>fG$xtuM?k#j^Xmi!%xKNE7V6{Pq))o@OAonE -=*s4cnw}pJxUifWtTZrZI(i~s1l>)8GvR|kYL&k2`SWrAJtYVg!PiEZ4g82(wD%6$`^1!6m$|b@Aq_h -6RQoA)n1L9Hc(ZpdZQ;*;Fd*oV!i#XwPZs614i= -in$i#R^=qBfTs?@>w$2-4P{H^y)vDbWGa5As>|rYC#-Z0wVD -M{HBDxzz`A`D@INVF?;{MVv*UcI8D7a9SOnSXLCTGnWA%=sW7q5Y>%~P3L_{EB94?`}UoVz1Qgwn~2* -RE_umH?=%$dg>xEeFJ0qjwLHR_pv#0Z!fGYRi0bh*?3VQ2yfepjM6GSoFSJj<0>Uzd{gRu4#CP!t@)F -ra=rAyxHnkf7~q??&Khua9J4fA_-}jhC%AH=3mN=77QDCGRhxV;eZx5#=5p#=+?WeS?Ke0$aSy+&de- -I&cXV#H$^A_~$`lMsmpD!~oP|6n)P$* -Wv=zTuaA$^=S<%}?abgZC`rlow%jNqmKrUqgRhB`tY_-WZTL -&X7&kjX!uIEkivV2-FmTKk^B?yPPxMeh8awY1}Tmivah=TPe<54~W7Ve2K6*AeW1{3dj+A+!;~x-8Wx`8kx5KAch_Z2g-KcWwv7hq#9B*!4>&T{ -c@W`kV%iq#EXa+#K-b)7lT5(~C%9rR>5vCQt$=hz=*&F^rEktV?j65ZeylYF<;uL&J`Lq&i$13zA!1+ -gPw1(u_6gU;+Kp|G6P{S1RZ#qoVD9R2{)j+?bw6jvOiQO|I-B4#1>im{$W^1uio!F_ru%Y?P;#RLk9@ -6;UG@rIJ=IX%(68ovLI#wtRWRjVAxY|&g!0=@MGaSqSl~#~yVSPf9YJ1d5Hrdf7sCw4+@Y5*nq|HPzh -bk6Wb&CX9!2@-O+D}0>ppc#ksTD?`1E{x9^;A2c5upl$HDzK1^8L4AD~{ntZFP@|OR1*=S!DkN&B|m^ -Slo_YAX#KXF4i7hO}a$Y$%Py!1m+(?w825H1q>oa61rtGjj=E+-9i?Bi6gnvY!KFpJ6J|o05zP7i0r4 -90wrkN)bEqRBGsN&Oh61A1!N9gb~?=nlR|Y?kJfI)$xroY4H4U%iYPaB$wD{a{}gwYqBO(jZ2%_bsKr -`6BO_@*%x;URAJiMMu}2l|U)DPhon)MWLUQm5G%+MsH%uqDhpaRQNLtD$h4Q%OqqI#&H&1GFni-K*+n -kn6@irE>)*+Mf*T@$Zsj)}ngm=VGW<)+o_E#G+j*sB{j8ruvsV^SSkK{ln0-1}(r}|IqEA;s6rvhN{) -Px#)q}uK?&zIHa0%O0Most{{=QNT28ePY+;q2uY<_o}>$|t`c1v?wpz#fM{%+DDZ9E`^Hpg1IDm9}$QfKcr5fu@Oi!oj$;6Y -XL<^{^iVXx|ZBM;av15dLIxG@n-RQy&SP38;`>mFT8Cjq;$1%L+a(Tg0kJrD72j2g==CXwn;|B6mv}6{{6l%#P_>vYVtK@v24(K1jy) -)b-m6q=qLHNmHBj_5FH2Vi~T}FO{x$>3iFmp_^chpNc2xcR*tVtD|+%-fSGb*j1QWu$1XSEvDj;?RWm -Dar?;5n6*Y!^-zKM}jXff^X3N~@6LB@B?s;IfJvZA^`qG|U5jsI=C!gzsm??gYRn6ZabU@q1Z5=XW@F -3F)!Qe4N*`g1uxIcf#D{gCF4*G47+rRqlyMXt7#H#a$~3ial8u&0#lWjq3k{+8lPhjGe=0iX#TzB90h -^zhI->U+a^aS8wHvq%+F*rZzE4pV}V%IyJS4lPOh6>x>iI%yJr_I) -RodbPR~mZAN7P$g&s7@^B`I$aJz8_Ycu_O4!37jU6Mn!PEw7=B&)*s>a>u4I%U>ymTEXf1v9br4k^hV -56Y?`^ddnxSJt&IX>dw_mG>R5}{3ofYD3V^lB7Kx2ujAu-#i3$z=m_Gy|6_fR@TskJn8cjHdL&giElL -7?uLEDM{!FI-#dPwW2=!M?~m+XaV{Vj#*dpjSi|jQkRiz#9jcF+bAKZ9&Nz|nUCd@#htkMgbRUmkC8X -sEPRbyrNVo-crqRH&!npr%4?%Vi0`P-kGVu;Q=C)<__Jb4#bBjXL!CsR>k7Q&Qv7ynUM5iEOl|uDnAd1mz2s+u*c+n@teO&1(*_zA17?NQN8Z3)M{Jf+W8Z+yPmmt( -r_;>=ad@e|EwPJzOLRX92rSV6+C8swCcQiPde@+)v_K^I5b8o6yY|MU?Ha_X*)ti6O)VUC7&Syp|NZH?BA -iW1ki_EztSC~?1N>cb1nl(lH~=96D-Vjh&)(*D=r~e(7Bwq;4r;*_d@t?K8P`RSU~vlPI(;E4ziD&Ap8=9{}*`Zy#!M|c?-x -EC4Q)KFYeuqSsB?=VgpSodmcXX=HPQ$5k5-`@mXE~&-#b+s^}|wA$?8Dr!Pw`ea*Jf*A-THJ@u%?2`| -~j?3qvB<~dHUuhU9-^^V0O3C&(o=;jy>A|2* -`S)9cX}?!MN(-H=n?5bX@Zg%Mk!9Vt~06Vf={uf~AVlwb$JXrY>geCp@U0q0{K<0cmwyT8&rPDJ7UnzoCl`x5Iq>ssW;vZ2Kv(I$=;UUd=#BSe-CxD-QOK#lCvSQLg1A -M2GI*`3fGC2zGES1qNKN2#awK_vpXk*1=AZ6ml(tINnGWe@hj=R3!QQ6(Y&xFA<5E3+^_c(HrN2$Eea -6M~aDW+Ipk#&ia^YoP$v(7yJW_charIn5!SbdY`0;MlSdm9lr_jvk}lj#5avuF=&5p6IJQsjKHVx7Ap -6MWEFp_NOE{r70>}Hl_!!kE;tzghN5^zMQ~;PMOlXHu7aa^V?sTG95{gam8)s&?2OfkWXudb&$#9{aW -6;V4$|e|F5YbKP>fk!sTCS7c$%)B_A^`y?$-wj*K7S!wup}hOkTegI-M+8scQpVuoWxZmNp6mJyN7fF!?@29+cKdVpm!;Y1FXN0CeI(6Y{O(Dqs$DX`{&c-yj1d3u`LIa4a^vRv8!0vcoH+D -n8fxKjNk%)6Wuhige_-!*w3#4tVw5>p@veBh@m?Gw&)cg~ -Af1RQ(Ho)h%rD+x{>NwpFx--ay-oSx`yK*7(3uXEITHKXEb+zhuNOoaysPeoRw{0CCs>E$E3$-k{mdB -I1)cG%X$z~K-H8d=nUl?@4yl-`T$s07fQu#T0GxR7w4lE#V%|N{^o9;N8|H~2OS5WgYnN{0B)67U3Cr -p3>$Y&n|B5j+p;00xB3J^_6E$hM;paf|>8c!3Oe?oWvr1fTLJuA(S))$C3+fp-97IJiq{FT88iIQcLd -92NC?{JvkdJLoqKVh+W?uv$QLsegi=f-W#31qJUjmiAh<3?OMwp5_unNx!Y5oSfiFk2fXIT>`8NIMue -EoPz+7bU~29+l9dbQcIO`BbI*I4dQV{ERA${p2-xOC>PD`%5aE%m%fT!Kzji*`&LQY$5n7?)G?nROns -pO4$GGmNpVqh{N#I+a|fB&S_ih5a9DRq)a({D}M(!>$3Xby3_fL -c!4iWNj|{@#ed`g}GvR;bpk{TNU1KQri>rzD#yFr?*b47`~h6U((_Q~=lt2i>kI{?Rg@w*Y;?D+eZD?{_Qgx0-ile3`4;sa(zIOa -ECE0(h47u5{&GWIbBo`rz&!iGmt1IArIAZ8TMnuivfL0T{e5stS*c`P)@-cGLRQh|H?`RXRCy2*xnoQ -w4RMG6DWkc%4 -OK|(fOJ*;9DOppz5sKJf3On0o)6vbUd>@6!*z9^UM(|wSRvW%?a8I>dM@zA*o&zjA2c`{@^ -%fOekfQlToM2zJpWesnw6i14-ziSUW*VjeGA_|Cphu_MoN1oORK;FX`=aUzKQa(Z40s3#YPkq-%mYsr -9F9ij?jcpz&v@B+}pQe!O(cWEYZuB_Fj7MEX@$$C^JY1w+#Sd#8>dkF*fA_{WCI{edXp&bKqgtk=qs7igQNr7}siJxZ3eNJ;mRwU~6Vz5lNO83n9!)eBa#Lt4wkMIgi7WR|Kep0R;(WavXwl&?*>alnk~$ -3|NK5m|~Jr_OEkC=RZncPi=7DX7CH%RLLqG?|Wt`j1qhI!E`Qo|&+i8QtD{7D9XW;^X)DcwxfgG0q=y -_5gwYEOVPd419a%pw%R>AekL)4SjcZSWVSzog=c@g2~hYQ!Pat+Ex|161HMpYra$Ia%+#&*hOuyterR -Ud2%#htSOH0GvSYgKlm1wk|Ff`g&D95OzB*-%2vUkI*e7^jSpqDx^dgZFm^v1;osgvk8zBEDkxl4 -?avZSXOh2;vJqNn22d$KBilBE3usaNsy~u -hW;mO#F6G?wy82+)!b!&j;Vm-^C(AbeH?by`zU^@_ZY1@N00T(QdRr=k4D3DFoC6c>STl{Js0R+j;z( -eu@D{&xz)R@3i5tVo0kll?wkMON-xr=;*Yh+`)k30_5i9)eZ4?tgpelj4&$kbs?8sOmxT<%ayv4wJ+y -pCBrDY;hobB%aJl_c`eQM0BdILQCVt3j%z$HveHk8uP*sg94(vp0GDNShUARZ3GbyyW%wk#z^5q&|qp -<8{=*c_%$W2j!sI{cOb}>MNkNggRK;>a-J5tLkrEFSg`8U -GdM#9;g|5~@Ra$747Ft)1-GmY2F#^c-_IJrOl^uTcv4JKGXo8q)Rlp!%)Q#=9=?t#Kg%B^nV_wM4wEH%mqrC>-R(ehid(bzQ+0dmigU -mmV=Ir{YS1v^tbbr$cJivL<&*OW~>)3H8{eBu~*8;cE0PWI~VmU%Eaw!grJva6O)nre0ZE2KhZOIdtWZ<#Oal{}t89&@}4@YUiHhsIExz^SlWuelEF?pFg0yJQaC)UgBb2#oAI68c2)aM#4pQX9lBqOq?fPj+1uQH}1<{Db-gOA8lRxtx;k13dF(_NpK3p_Os5+TtYL1@Z>Wfk -HBe;z1!enx6v0fV2ei*MORJDP9JEI!_!2WZD_v%$4teHl^>KC;q|S`Ah(el9G3lvL0a%0025t40Y91Q -9qP}FZ4T^!}#$1;OT+&R7M;;h)@Ur36u|1&;xU*B*&&m58*dJi0Yh%&)3(bHqZ41NCO(`P>ID0O_AmW -blveA`w4yJu_~W?{HT2NLjTaDVp6rx)Xp7Poa^y&2kuL=pYVk$SoEozDmPl?Dy!pMLk4L5rVky%4Vg> -C@Epg<213bx0Xhv|V8ow)26&a)lDkp6cZ+@Or~w%V12SbTN>shVFKvDCa;mDMp|lF>bp>5N)l%Fv1-( -OmBs+^DWX4dZ*lkEIB#yB;HKUBKN7_#` -?vA+(Pkpfg>tiuYZ{yrM91{kq()m7{@Tzau;rbo5-u^E5XNMjEJDi!jnDKh*_l%#oSU=16>adYo{xkdqfWYG=^~rp^HR;s! -)?e@}>bxs~uu!XoTQ2vK84w7WU5}%nYA=2}U^;%zp9=_zKlHeXvr=CP5)S#T5P2A@<@J#lUHaOV@SM6 -2n+$RPk&64d-&Tm1e4N6QwgsBYqEhFzOYO7cN7UYhh5|rSP? -yVEN0d}mXK2+rzu^gTB8Qie>;;d$@oS^$-d|JwQ%9g8bP*_^bfDvH2T=6XQ06$YCW`2@LvUzJ2Z?cD@ -2k{1y=)v*yiPxM&4~{d_gEV02_ftb0FA^#6xk*U@V&)FC%ModZS=zx$JFJmUY*~@e4%_q3KmRf+0eH5 -svY`0wBbaeluA$ZDyjd6uJva?nkD+la#H|p9lESFDufTG_(C~$V%(bCFeSb9Df)qJJ)+#AkuetR0W|-CLaAde{g$7=XQ(9ylTDqWTMNbE -$y+m?z0H+{;>dYn{8 -$h(fGDTLk%TZF*TJ0T(M08S%$qK|;^J+a&dY5nuaUmO3mkB*RFsnqnj{~@TRKLB963r}NzjcH+}nVL* -~4WAn=lyz2 --DFIqPklpYyOVCD>a@bxPvFJ_d}nP`_Mt!#X6RnxfUOGMIj7&@=M$T$Kw`UOC?0Uj9WM`B;?^U&xy0> -Weembdi0$7e{n(VAXzv~ZCrhnjFT(<<=QxLlV8M8wXE}dG1jua}NeXi0`o2vE%I+{}z|F8`!J`LkM>IcaYZN#Bf0+oG^Ixi@*!5R??n}@Yr9k#%1r7m -*aRxUrkTZQqj+7ZuCh=$#{s51c1noOojvjG*q1Vf{oLSRqI$bW8?IPksokDsJ(JD4UukN`ywD4YyOZE7_u0D|`b-trP%$T~Udb%oWylOZA`VATewf%N}TwLo#j -`@NEM0vqW;++4*oQBnTh@&)MNmXHUQYp -xqX6oDYS(s*#vgv2t_D9C)-qN=MyLEBUiWte4rMoa8UW-6aX0t~Zr~Qj++-Jx>eNHqDKgAY?t7gch?F -11hL&$bvP>=e71jxBGsy|kc0VSODq3o2M~a9$H6o6a)1eYsPzhZBK%lbq34Xl1%sd)x#b_}`X&UpA^B -^W_F6d)C*Xq#DqRscpC5|C}j!#-Z0@M=|%px!O8a=NUF*P^|()A9Q1UJ*8RIxg0lF~KxZ76qsi(U^sU -*VdNHUU>|0?lm)Ouz-?EsWlDmp%)7quMO|au3eJRx+nIxn@?1Yb#>gU0zb7YT~Bnf-!g)pf4eqTjV}u(wuSC6`i{rm -!>Q9&T= -+R=dfk#u2a9q#z~U_RTGftqIZV1yO2+^HMN%Bkx(bsZk{AE%k6G})Ed#k5wMzvm^*wcw~lCuvvQyi9c -^9K+imv7eI9APP!0+-Ow!EG#>%X@$8-xv#C#+#Eb5SOSA*&$@L3E{^wtErBt!$d{^{eKy -)t(jt0^8enlM8DiWdnBkNCUY+=PP#+}|!D4u)p1M>VHgOO8(aD(hxteD?j#t60Ut8A^MVGWeZ9Cg%A4 -J|faujFunxHCNR%V$4li-#apnK-xB_6qTZr -N(#-R%%|E%gX-`u|>XzE$QD|yaW+yRRfH+|0%{px;LZE|0@(nle$5ZJtftm2SN-TC8XL!z_ea!>~@r+ -XGY^Hd5HQploh-H+P3ai!MVv-Cv|s36DquF?)EtsdQWsFog&b4;&+F~cup)+c3-@fcJgBsaRzeNXDK_ -tqq{82?MP`BjE&?ld%J+h^Jq#sz*Zs4ArVTo&+kGD8XKg=V}IL0Jx3zO*`#q@`SC9gK+lRb{JLVVs-XsOm(Y4 -n$Wbmo>6NE?St-A6Mx~f+@pw8cHP~iI_^I#SX%~d(=nf5G^E{?11sO_0`B4pc8nL>^e=Bf28%qr=QRr -6a)3J{-;L?TK}5b7(Cvtoe`ieTdAW2lO1P-JaYRe*netZ%ANFRD@);x3Q8|6z*vr-Rzx`yFkPb+5>^e8V7lx!@Q0Hp0|T8$i!kitTt%M -57&u%vl~3nHH#&l%{UUSGTZUa_%UV#y?5Zs)Xm{xfVP?L-)mghPG{`^=HCLFL~#vF3n23B&vOn@I^i^ -*$)BGV!XN|IQcDvfmSt0r}UcjI+1oc%|$<3_%k6L@8u{2oqq>Ci2@odG~>bpDtNvwM_pm&_8S&UIZ7_ -oSh{Uu*p>l315*M}D|K;HUxB=uih#h#8p+ist?mJam`VExf9f1`aEuc!{J#^by2h~j)W{^ --94pFi>1xQ|tYC>^^1@Gp9rui;9$+&BIi5YY{A(yzd%-5zeEBc$aHQ10*2yx!`#-9` -3X9rMoIhij3jTc&lrKI~=YGm}~gSRYHz@bf^=Dg!t)*wAv=uTJbXCpK7tfWb*k3w4`Gnq!=L&s;OKnA -3aW_YMUd{V3J&T;RE@1+@$UT)u11Q;&k%s50dNJkwu&4}ChwNHUCVQ3Be07CU*F2goKwqzphgX -JcBU|!f?kTCQz&R|5){Bp8rY?o0VY&URzkDfa=JP{P0cINm)y3QaB++TJGj=nQFE%caSv_4xO7AHK&T -?kJjgV0E$GeyB^$XG++_`R@R?1HOu^E`niW%1;jmoWkEeE{QY}nX1CB0EJkC(3#9fAkQ@EBtLSSJwZ* -4U#T?{nt{W(>D2C5wGoAX194S1p$k7+|vn<-Nffd0zes>>Fj%Y%6PfO=uAlgBvV3#dbYI(RH1&0NLjH -Z`i5{L5{oY&m^bwc5Ho&7#b9)b@k=jS0^{KdIV{2;WrGLohVj4Dd^#LU;tWp1$IYG~5zrLC*3wtxw-J!97dnQL8@GkYmo5y{Y{ON+ad@k3 -ot1`N00rPL%(YZ)jv#kKv%35rUOD)a}KG51K~6&PV*e)2M}9`r^KtNYM*5~ej=?CN_FEooODNGC2RQ) -6+USJd9}Brs2-+!GN~7!XOzCzdE}z!O>Z;(yN+)Cm%3jCaidTYQKE=v+2Gu3pN*tE|FC==t7#lye5e$ -Osse)A0gDu`E-_rskfThmS1~Y!yyD7=J_+9iy?hT;I62F@Z-ijX^`i=CpNHUx>5hW31Cwvk=S=NasAbavEq -6^bJ0npPfloHZ7DFq9sZx3MAEKgaoMJZoiLmJT1T_;Eh>}9jiY7XqT -%_WvrB(Efu06Q()qnx{LPrR~1>&M>lwazr!BhRZ+yYgoTOC-=T8UkRz5em(m-zdTU@F -@X6=I@5~KX4}BE^)Sw5WwD7pTK7mRtavWN4xEJmwN+%+K3{$^<3@(JXK@^@(SeF3*^*1^`Kz{u`b|7YjkI*?@aPiemKQTV2C>TG; -tft@KdPx?70E0ftTqGIgW^65p%#kQqz)G=L{~jI=yngoDOOypDjKB+y~~E)sUtv6)&SVSMDeU1(Z9$s -MS`t93{U*rwtct6c?#A0w-R7IdshR5Wj<(t^mKY6WzqXc1uX%okmaldD|;ihT?ASSa)o -ZVT;u51&u>;4}Dp`i#?O7kv)U=V|(Ufj&FY)%T(Q1CQ9-7=8-EFq(?V!M9U7k>M3pZ8{W&6kLN3zJ^z -J-?r&IcJEhs5B8vAgnx%SFiF}+XdU|K^L6^%i_a&z;aTrE$+bL-2Y8J&rgPGvXCSVO=HMRQ2GOA<-Su -L8%~p;(aFpBM)YM7?3}ft|QCCWf%N(5 -7FFX>ZpU3dB1Cp!l#0t$mvKZ&?=3^%=df8!?-fbB -voF9qpySPK{bpzK@Nk8u3TFU -9iy+I@C|Q1IGfW)6yeGbZtOb+?(nB2 -!cbD0ZWih!477Wa)+3)X+pXlFCU^Mf#9|rK_Wq)4^pT?t?T8r3F&DplxsO>MxiTY?S3W)&thVv_*@^Q -%03JPRql~-(4JjsEugZd!l`DMo*l-u@z!AbFy7XM#kyts0<96NsJ`%f2JiH0*Y4leeh=uJD|t$tllI -|AR?_Pwnx;W$jnqU>@3nYxVRTGT54kEc=%jXytjNyX5k?AO9)A3j#Yac^qxV}AvQQ@-Vqf5OWTP~6lh -gPcBE6jm65qr<`ZsiAXn70Ijm9@}4?^OhB -(fblU*UYd}CkB@e`x(`PZclfBfZy?1U5ZgyVFo(~rmO3re)3#H78k-`%UeNp*3Rr%cNQVO>L}eP?>@AXFAfQNF~cLvK{@hO28rr9zz09P%*p`!fy}LTj`KneE_6)lnq4zhDXOXU!Z+BHTv*r+>xaT# -mpU8ns`!~q^ALCdDD|HRi4nG^k{x+@WjqZ+s(o7=5a$^;t6KxFQYBii+l`v>N~khFxpj}qK(!u95F?t -B6&!=xQoW`NYgnaniUh$E}?U2Xx}hifkxxikTwPFOhX9%3Yik*$y?k$4|xq2=^`c4(+?wpF8Mq0T-lB -^QwRyunzKSV6bd4u;O1cIfUP84Q~#rJ3mtQVIKJq>S6NJ@6oYvOp}BNMq17$y -O*u0;jhv@&BCQJEk&6UDA91s(VNNKi?M$y9-@LxxKn4(SnuP+KAR#q|#nYwu@$bk^XLkR!wHRU2hP7x!$1fk!;n -##Uv9Cp)_Ln2`be#$@P11x}$M5GF3BWIx0XLu@6c?O&4tgHtOuX?=0$|s4sx_ehA;`T+nd#2i09o=%! -q$Ui^1-r3&4^$XoH~kCQnBfB9RS%m(<|1b_SC?+=)kbm^08ZNQu9_Hb{Sfo?f5!W6PHno_fW4O+N%m- -y!l-qV8DCj!J(qY&Ft5O{TfkE#KVkGJHJtWlU%0Mq-ZwrzwSgwOv*I|S>*3ubPVSF^<@xdY~FF1oh6T -`|h-^yY+%uI+9!-h`KSci^Sn?fjp)1D)J~NbynSI*%8fz5E9Hn^{X2>pl)b0oi#}y9igS&!~%GOtFdK -5lk7qguL$JTgbWK&zTf$N{^;mwc`)TmbcIqA-sV6&+_{!uRo1c3E=4C!|_G7x*T_a7S7e1X5=sxg94VO~tB?iccmo?Ocwi)=D0i%x`{g_rB9kv#GeQv@WN -Zz_|1I%Wj`8)9TPT|igMM3`UKOl8Belp@LCT6KfdAZfBU0)Yf=|xjSpiuZp*D-pM?)uZ8r@v9)RrBT? -r$2+!;rJF8fa;?4WjQ?kg_SZ^i0kiuPjOpljS;4w4?jwJMPx(XBFr0GzqT&w@%qT^+I`T<=NDcU#!uh -N@>A3;EFm!9o9nGZ!?@^HBc(}|tW;?p!v-?bynr5dzg*mAGYLhqGq=3Z@~@Nmyz)!%l~Xpm -CVRHHk3cYFkvAhR=ELp@s%Crk0rx+@=3X;{O8Fqj#H(s{9lt{yzN<8Q65T$aTiQkRQ_l(jBDk}mF2%k -hIe6jdig%19u8qRr~D-hr$ZQ5e$_1-0<3zXM%uAKNyi4R#6aC`Yg2b=;)$U$>Qo5?|M!zX(Neq-2v0s -8zPZ~87SO`Nt!zt}zcQ%uMGI#N8TJmajvkN5=fD_plhsb`mrTGEQr-p7^h&gugtPQ$W71pGLMi|=jI@ -H(NW+UX>CVPNZJh1F-=7NcX$G48OFGCNJ59NR;P^6l$iJMX;|otkUG -D`+uxwiLSjjT5oK0uYBe`9r2uS%;IBp8)=(_BOMi&-HCQ+#V?~ShT|4#k_$V$1|=COypQo8(id9^i$O -ZIquE1!X@elFH5O#o@tv}mS)_Yk@gFV5s14_D%j??N>l#3;Kft<|_hT5a7mo!cLESCX{rHF8c+({Zs2 -rq^ogt~AwUZ8cOv0Z?4aDgGLIrS0#Uh6{O1({P6HAArf+wnSxc=?|L8we&I!?K$4OFhfYIRL-nWKBEb -p`s4u>bZHf|sRsM0rFcIDA>-ck6+e>kuaR$8(+#0GhG%c7ZnB{&Y8!tM5WW3DE~3u4Qr>AZdPUHwO}4wp>96Hu+w=F{Kqn+HN}H?q7`$ -^bvW%z|r`3-Se%BqZe33EU|0ctVW>B$NI$EV@&b<4u;EZ^V6>GpWNl50SlHQV?FZj$yPJ(`$+(B)n)PMWGP08lt-;(VPTB~NPCr+ajX58Rodf2R&g>X%F5@+$rYCNi -4GcVxxq&R1W>A98oZ`0QYm&lu|5gST5L8@4$Fy%!(~|MzG3l+2fWmOr+bE)i2T8UxrRfrC7=sSDn83(Ak47Ta#DrmugS9){aQE@tEtFb}2+Bp -CQ%0*XB}fV1K*b!!g=gtA<>s-DHNwid^ucH2g7ou$+S~n=pRJWn-M_4F1&7Y2Msatv1Ono7inCj^gdJ -;aA|%#4oMl=Tz~91i0-}(C{v+sAQTWS18DoS3t*gzNm@N1yK)0Iyo%v -E_`4kbroo?0Bu?SEveg50sP!q+V=la-i&666gZl4U;e(oMCaD%CSKBLqn3;8y{+y+c9+`4PcG~$lz|c -kbJIdu~D)Ot!89*ak5p!u?z^mZHOjkDGgBGx)oUs7l{BrzFPsnBOjR#=N&D8^V7#j-loY0xx4~@?H3f({+M>`M<5~B -F&GAYFcMH_g~8Gd*fzmwVlVg@V)MPQ0W!b$L -4a%GB%v=I$`GE;pLLmrgHnHWoD%HnmotZ`ydisD+09^DDnF)U0bFY -E(NZ~-bnyc-g41%snkrJJ%{3_2)uVeE^illt$Jdas+Hv;AGe7Suw$qQEgL987>e}8+f{I>DEvNFTI1~;TY9cBZ0R;j@QP9)~1q4DsxYwanP$0! -Crp;DXrk0JbnM0EbYR)-igW2Ga)*&h>P0`5xuV7fiB40WDPy)M`3KKYeyro!@d2!w)QJvhpAu*2Gm^muIhfu$ -S*annF^-!+_r7P) -K2{w&ZKsq9p>~$Jb5nVyGo}RxU8Nse8}c{mHI#w8x3wyy$W~bvPlOEuk+!yRY0l&)>a+vr?bKW{`v-F ->e(IOhp -mmIbNk@UL~j8^G@eB+Lt%y53$h0)`4sDwg^?nq35o79_AV~$y}~UW;%+kUYDe~peiBJ>5LkNZ-Q%ys) -}k9$5NA7>$%+4oetFu{gj{3ku6(^k(YkVF-bs>KKPc2&FtSQ)Mmd;71}N8Icgzqdnn;Tv&8eV6QHe~W}qFU=2wzv=l6NDT!Dkl%RbW6K -FZU7q=(IoN!t-K@_jFA$;%sS7CX|q+oLGhwdj7&F(aeLj+;1cJayLP;T=l$MxOrgNV{3Fl6BlgH#3EB -wl+(1N+AKTq`4?}Ea&n}wuwtfCLR2a!@wSuJY05AYCqz#E8yyMkAOo4_A;*=X{WCoX&x7w|Gg> -yJ5#{bd*f|xWwZMTacMVSP1d5c%sKzG3w7GFZ$H0AdOyDOaSUg?hmZMXuGyZX&$;Z8-i50lqkQ!)et| -wOCDR+`268(|;7Tu@j!-YnebG>-g~OZ{4Rxky7Y=jI-{q3Onxhp(lopZ -s0$gM};fb)hki1!PzW(g_j?CJ>A!h$ILB4IxX;6oou|hD?P_h0Gtx{E^HGGAqcuK;{K9E6J=Rvx>~BV -pj|i2$u?%N}U@LM6V^-K(Gfc)R&6iyJIOZG}=|_H5bCM21ke;y5C}J@2cY2{#9BYcvt5^lRwU{siJhd -KZ$fNeLItS-&y&k)p7I4A@a1eFAWoIMD<2s%B9;+Z_4yp2blWBh# -FsiDl*NL+a1W{JjOc9*XmqSa*=cSVFOmq<)7HM=l@Ph -gZ)=be*l@L8Zuo8nevYIUwnQwZT4)SUFwKaR7O}`?zWX+j-U_W=^EIt(o5qa68Jkz3+i$WrgHBn!b__dB=x%N@5Sk@SUQH6a>EAa#A)w+c;zDU-(g%VdKx)e>k;6L$#_4PRJK|^nY3 -g@jJe|C}!9rh8$#Hb!vm;8)pfYjsCczC0bw7A>tp=LO~F4UvLA6}&*-p&}1S3le5R0jM$cZdP`en38; -bjHofoc)A(Rlx7ol-=&YJb$-Gd7EhE#f7zzSFb=IkAc!wNl8=~QbBo(iSeIkwz5StB&BA##Z_xC>)TGx#*e=$sIJ5#WY4KJoyV1&7zRJ(L#+>2aI_^ -D%ps*CQGfF|bYAG;73Ru}EYg`(eckBTygHcmZ2_3&&CIIBrgENoA)URY$ht>*i2DIvYD=gvzes -?v1wKO*_^5LW^=Z3yN1FhXxK_yxNgE6mEy%#oMWQ6u@&d3Dpzn8_$JJJDd*UVh0Mwswmt>x7i@hN)`M -()0oEOCT@C9yY+VcM8}l7!L%cZOak9gcY{jt<%3`)+?}##wtvE|SnLb}b8Pgg{3V%SqTbamK3~DH&*! -mT$!`O;iMj6OfO!X9Emq){tyl8r94=1M6mKBTgSqBimel1{fw>m!MZ<>`{no6*img -AXKxVT#;1uWGQZpe_0nW>@5A_ -Ui3(>+2P-F56a0}C1x~VQL0^A}smz!0WRsn9&nv1LI(k{SlwC3__F1wfm+{S7yCsh~k0JjO6%R$wpe} -G$-=JJ8+GBCi+s=2(Xx>y3-W@;`^sV*S_ZnHI)hg6ra0Jl8N#mX+zBLn&tXw*;2RjGIMr(?BgOM2{MJ -keipZiLgmw>z0$bM-F&g&0H%1G$qnzo-F!+T~N}3pHVFrzB^YgmKOeU#;TpQf~>WxnIN;hp}tdix0Fr4{);eB&50wNN}=sZgj? -aYpLQjAjsL)njD)ANN}-vQLiF@yKAE}wvX+kk?mO8w&cjk{1T@fm$vhWOshrCL@Pz}JrPKOzPhv>Qy< -pV$#x9=(1(NMx(o4tNbzqWU9fziVf^O?_-B{QI8U8%zZ4IZ2F`JJt1(8W!XZ8rwVPkDf!IS?KZg&qE{ -}>U_vLf{Vrz847V1aY-s|s%Q~#`%SmoC>>e&Og=3hNIeSfTS8t+QC=c-Q<1XqK9ab51!nmgl -%H1Q3ZwhKRCoLC@_m|)iiy*J{nM+@DfG%QDf?C!PG%1>{VU2VHqQQw|3nXKV^Fc?7s%l;zOmUus?+wQ -?zaf|CdbR?d~BL;U|j<`<%u4!2bN}OTYeeyjUT9vjNt -2*xDaH0V^>)iek^Q}$_a44nw8ER>Tv!}Pzj>FPt!{ajoor -pHvvBDMCctl1zJFHTi*w-F>iw@2ec7!P_r;L@u4bzV*T3S7+>cuDAb6}Y`)2Lpk>xl{-Q{A4yKA1QU@ -^`HYxWXx#**&0*FUU9aUpJDGCfurZFiA<-0i=}mU(<`B+i(*k9-IGV)g9LsF@0)ahzbw0?5J&W(f!OD -=8}M4yJ^9nI7uohLxZedRrg@sYS -<=@v|U|-;c=<>TwTT7#rrmfXE_yP7}r0KKYCCbaM#1gQ_^w1ifk0>q-A@`VQ<&jp^L+_7}UKlJHk@B` -_PBE&dXeGnoCxdTdkB&}rD~`HAo!a7J_!${3H-aoJw5mC)3w#D))37LJYNKO;qrDJB*#)sqJ;xIqZ-j -O29N1nAX>lnnDunF#Z$v=T;)_^+z6JDSRk|?5iCL-YNOy%^Xi_8Mw>ThtI++%)gQ{4F_$9|)hKa6$PF -B}}&bAJw&3=A+^RE)8OR-6lk#R2C2>ObOk+SA-lw>#MpqZGQ9KcL66^0`nj?`Jh(|UVald;i7{9RdD! -*e^vqd(&TY^e72=k!eTFH^su%me$9n=Wdu`dK*I)W_bmD8$si>>Tb--wV;~57AsnuX$S_Goq+ijT%as -Drx>vBwK$#lK1;v^kG3LE)1jJfGDq=E{mE62NkVArsZ8}L{%K)u}d_u3|;7D89MK(Y%WChr}hdph))I -h-A&CycBwlXgOMEjb7Is3c9|9jz-@;QR|o;vH=0glmvwfa?rbXjs3y{3*srn%{2<&_%ptN!_Ff<#Y}( -x1gC2`m%stHi*tr!z%Yaaor7~qcm@KRb!nu@5VNW{;ROQ4Wb#9h7F=$KkkV6?Bn%r5Dk4o-ynLakPia#O1Td-DU -Um7YOXx;xRI8XHEFp?$*E511JaGkkJn2>1E%3%j8a};prHZL(3`M_TeNIBTU|`iSj&1M#x+0bjNm&uS -}U!ASk##!52flc2Uc|umj1X-*lkSM#US4FI?F>?y=A!T%*OU+3!SMY*i=&PY1)b%VF~Z@8ht)lslg7} -;7@nI4%T(x`+^=hAkye%$ -%9jBCTJdEEhXYgCNmVYbwcMtz6{Cj|ZSM%>q`c;t8S&I8n}3F&(Uwe$EJeyjYu^qT2b-`Gy-@wEiO{`) -egk2+O4&B)fOWXbyuy2aaS#zv8(nR7F)QFBZX~&S>0EA@B8Y$S}>Hjw%EYYJq8|EICxz4J@(aZt+lT< -EKJ*1yO;OX!h6|P-mVI(YuS2--Kp#)-d77d#%;=;bbN9yvy(mV${n`u`MX`!%(v%O6>R|~h0;`%ULE7 -56>u!OhquA*`9s|XyH(o;%bx;yV=d?sWqw+`U24XwXz{k-c$o!d9vq`mD@MGR{p5jI!mVTSuB2E)@~) -&?TjgDuVJiI8LnnxL)$W3~fOOlA9STWv_YY1bCG_@35Q89wqxCmx29kx8*(P)W_C}6frGqo+ -E4~`5EZPKJHX~tbq@ii;Z^dBMLqeO$DJawe7V7Gff43VcJXuJz&l>6?BF<)>LpSZ>@!FV23S?;zAU@X -w#mIoe9gZsZEm=`AeSUS_2`8 -|EvseWBuZDo?-;C?j@@P_c?%qVceC*GzRbP)j#;R)gm*4>5WbB?P=r6$CfYgHIuf$_nl=xsQqNj1F1+ -vu?*T*Uk5){#XuDyp?RRteb}^%YTAx{vG8qe!D^9#ls ->LP&-$`%u8y6tCmE&fxHPv6XY2FM*nYgHx}&qUGgj(5W8dc0SE<&S*cp2V?~MJHcgAuFS9zhBZm8~&jo>}9TrhZ#EW`Spg1Cz;Dn(J!`7;yc#Nl&9I6sm5+vjxNTy-!^Xx@WggoY4l*&ZX3YcZ8<}oxS0YSF=acM8pfv>4R -`q+yw%o96$BNKg7NxR+pSJ|ChX=0UfpoZLZfD)y5Y9(Rg4e7gX_p$g`c2=FD(3&L!oKHz*t$ctDMoi* -Y=jHT=XuMnysx=hFmPQV2|xE-eWtgYI%8c;v*cj{k0BHb?vRy$0zuiQO6k9@mXcsHCXt(;W;7L-Hc`8!T}aRVFsQdVGdZ9_ -G?dX%@pxxU()Pq`CD`0*`*6fQn%40Y$LMM6Y_BZ8)!4eMGmlVO@m5&8N!~PiLT*GT`x -a<^_7&`z6QMemTckF>>GcZs+4fYet`zweq*LBnjhZxA-x13D7ivcSZT=RnCDjbF+6qhNPP1=RS+Nvla -B~A$3`=xc1B}ZQ7cRt0{w`;wZ!@iCxApU9S8GqXabaytISiwj!$~79F>ajgrSeYs#f(RJ8#Mrx?~_a$ -`wrFi+TumbzEdO0db>u{HC#J=z82T=x1b{8t+uczpG-lwl--Fy+wSx&LUdAWf9W|wh-(kFp+!E+ZLgox9RG$r)|ll!=3incXk -#Pq{{BeZe%YV_n2K2xq9#D!FN}S@-l4}SsRTl3azwT)=KyGv%!XdRNN6+x~fZ1n -;V?8p#_?_$(#N-tV?G*H0I-HjGwQVS>2)Q2mD}QDS3=IKGJ?4}HqV7O1QKCw^5l@g%k`RbI8VQ^FRgJ ->eeJIP9<=RePBw)XQ|!ddz0cpPR#lcts&9|-C`1(?!emP{g6%rmzW28$^S*cgu`O)3sh%2`HZ>(*#(QsY&)CBUMMkBK{95H)MXiE_n??e$k&f}G$ -MnMg#cc5#VM8G05Xa?&8cQ;rosWJ;NMeMh#NMhf2tPNiH4OUSXHOIG6^U9gnHPcD(ppAV_@3s8q^t@Z -e1%z)T8jw`x!Uxn)8ffvHZJniLIMGl5ffO^!xbIFQjgSczcjpVc&5nk=*y#cfoWyQa+6A6UYU}Qi6F=eKW6;+;>UuYX#CuVpRD2{D7s -;`!-XGP@3L<%Ra~v6au;fL5WYCvE6jE=|Fp^4&c5C6%hDg5tsX_Ki^E+bY?b*ZU91fjwhjujT^T=ihy -B>4FY|YJ`5g5t8D&43e;VJMf+B2JO3M+3VsgFPR(Vg^7~*hRBiV8$u4Ke&Z^zDWKF@G&6P#11EMCJ`i -B#A_+i53Mw^C-Z8*ZM$-OO-`HJnpz#gTqD+YkEevALx=k1cQPS5mmww7jG^L^xHi8sPo)<=%08$lU{5 -^{q-?_68&zGjTe=ZCEkbnok*^51-k0`M4oWX{Ga}V~D5vy}UwgOXKX7ekJxph0+@7?{00FU+QYf_ZPN -Hv6eK~XywJ{dEY2@V)Er44V78(XaTUHxuwt5Xk~vxTu6)rUGKt?(?8{R!Dh|sRlPrY;Wld=C4V7c7x4 -kh7qqdu`1{pKQRHBZy;WS%#HxUAtVwBuVwCn1)alxQtJX0AC#)m=N|OC&IP<>4%9WrXz^#gYZH<(Elq -0LyWtVN{O6-gFw|QFq0(Mz@DGzU>Z+vnA&halSnH8tBW{Y*WvLsi>1N&TY8mGUll0*@E6nEL$rnxE!& -k?1oDmZHE@giS;X1}dDeHzN2)Ksn8ihS*S;sPQ6whuO(R2`8;b1IafSDA=_tER<&U`3#xsQ?@Mi~GDU -`za&t_2p87Et2Y1f|RM_rK0g1LI77N>kL5!cMvDy>Zs_nQ-Wr{XxGF7!g#anL -f{by!hTLf^lE?L1c8$Usy6aFhtwIPL5Q&*yYB3i|39Q|3c9iq=M$&vCwVqIj5dz;4sxLW*&a8pi{Cy` -8HT$DS4HZmV6Uf&`>GrmEQUouXz%(_F!x-fEm*;F?A}%5Ge*LHJ&spyTUTQs#42Tisf{6mRTAu24PgQ -5P1CX;CoR=~mlqd}!!RHLBQ7p9h}Lq+gcSAnY-j9*X-3W;jsmw;ZX@Y><1Diu) -lumXub)%MqcTuV00$BsCNs!^(g!`M6{yW;Txbce6EC`pwq2e&!#PROrjM1pJ;mZ8jyFbAL>Av*$;h2> -oz)Q$YxXiChEjSEo$UDQfjP+OXwCp%tCB7mc}qiHzHo&mYvIS57=wy=0`Q0pZ%ob|L>N#fEc5V|YWXf -=KtAHzxH3AHU1;qh8mc%}%i1U)(~S<7uR7-7W_g1$jG{#F;kDay4x&Q3q9yRL=XI@70JVlN_u=$i!~W -E-;;i^){HU!Zs)`+%?bMs5r8>_o8Nbr|0GEb6VJ}ylYL@`FZ=C%ayzY*WB_Z)nD>D&ImF-9JPvw^W;BHMSZ&3V<+SF -z?AQ(G@P=Xyoh&gGvnS-tJ0I6b$C{pO-uil@64=ema5&SOf(#o7=Xdxj1Rw -_O^0Z~Vj^?_t!)=djPdk@oGnhnpAP!X8SM8c~GV&W|g@T&>#HjN~Jzt*4HSD!Hc|AA<6>f!@RzxWD#vE()wS}`OcZrh6*U~pM4`F_{65 -31yfG7N+K}ocQmWyvP*Yj?+S-K;H2qrLf6glJK&hSu0Y7|=VU*{{n7CMQq{YaXqKTfY(E9RrP-_eSiJ -qoH-bN7R5wPF7rO2uLU9~jkS3Y1W#7)OkeHG+LzC-vSD%ujH%zcSTDI2e6ORY;j9%Y5Ng?<{7dbP&H1 -mo2jt(C(wj8|)L{*?`K(k|C{dxqn3jZ>5{n -zs7sxYTU2U^A}$0hWj;+E0#=EAvyc+>@=}TY41oEH*GX_u&vzPzJs?!S6{UeWNOad$hXaxw@CcTVJVb)OZhekJ=jIsP0RrsE-nJ>6F$@qArh% -YCGj}bUb~MoexkP|NH#E@T6Z9;sJui1WyxeAlOdu3Biv9u9t)`6WmDvJia;LI|P>CJx!3Kiu1cwQZ6Z}fxO!`SH0$+lG1Q7(W1osi76U-%8O7IlHT7nG(y9iDaTp -{oxJi8EB2%-s+3GxV*5QP1q^3;`Zt0bCFSE}!8$$X5UfFPUT9fDnz*XG4{oh_m`+9E6yETY$Ti@2-6; -`ny*Ba3*li{;ut^fYer^CXxUcZ?zks~|WdL~v$y=*(1Wbav+Cq?{ZfrivUfS!9ZAkwkwzMWRTjKOwqN -FENi`GX0S^F-D9QLq!kK-(c@8`iV#p%IS>}!$r6V79&ImyWCB`qsT3keE5kwDdrHeMY7+a={ZBQlagS -@Bw1q<(rs)XX0QjPr)N%v-)vzM8Nw=33DYEDCdj5g;A`f%nutbhj;6Rrh#2}a`cc2e5N5;0usXhwTv4 -bf5yU>~`i^H5(dkQ6__1pb@)txHS;_ZQkwvDNAd`M=#3|qy#VJOT4V+D6w`7iYD*H_#YI4YyK)*R0%O -v9YEOJStII`($hL}Qc8RP?$Pv;n_nvLVfcQd7%$ -*JO77X76=LZ~tJq%Yx2r5fv0a?Rm)_`ILuRdEMhz%H9pvXLwNWD@0OjxU}4tCS^h$~vEVFm2#vD)~tw -oXw(}=tZvS6x;30JxIZ12Tn$QCe+EP48{@CGnsQIn{mmm5i*2vilVToOsAgbdVb&O$nOjaVP*NqU~Ww -zTFsO~7KNRx<*yohI+-f9dQPcdrW!a}JDuKtB3IYucankM!4%JEEnl)7UsMifGS^V9{~dBsn+{(G9Xd -`r-|C_uu~zsysczB3+c-xjX7r+DHcWrsSQ{448^KS-Ic>Qu#S9<;aMtYFa#J`zdUq^aboKuK*I{ -DA0P&SUEu9Bmd#p)RK`1T=65EgmZV{rm^CTljXov1Fo~Ey#qAcfPP-C{o-f;Z7#byQ}!zB3h -IA0-3m+)0fP)WR`qp;Wg=5WZnjMGTW1x@~K7DaLDXHhSyOG$sg&tA6uwqCjFJn2guw&CN{OcP9{c57L -yspWh0y0X|%tfEtczky;cYoKJ;fM&ehbeBFQ$1LW0|~nD#~*AA0g1>Bq!3#5{$4P1V|7H8kS7pSd}Me -g|uLQ!|B7M~UdbaYxsPZ2;38Mt=k74evRueG1AE^{+eY4eFoS;WA766YSoL-Fs;+q?5A-EU0x=%@8id -s@{bXif1@sF};@C2u*j;U1G@0(%NU;Er_Mo#`QR!2(A-mIqax6)U>l$vrA_j!QaUOz4M{;eG=;_)jW^ -j*1Ot8#aG8s$4kdY$3ah9PghN6mLqLFT|G@b9mJEsEk-uiCn^p@8Qbc2XbP --PxE`nu6T`R}-xHV?H9b8YJ#IZdJsv$g7M>26I(0w}3r8Y7-OW+-m|>YkkBExum6+IT7AF!-*&aj?k~ -BGK>ZByj+~Jwi9rh8Et(maKBxPB3#~|AjTaFbYw&@&4ge`+JGAJuMmGdwmdot%=(xhx#LbmEBA%m@y4 -&q0=h!b%k9>hW6W+4uQM_7arF$q>%c47j@F*G}sZKG3c**Th9TvB32lJ+v%rrM%1GuR=` -hd=nicj4sh;_BAG-J@Zn#-2@@n!K9b)VxK@n{R2=x=q`5x8By?+@WJ9@6KJicDwx!pYA<+_VVq0r(d7 -G{qE}TKOkV>puvHbpdrB_p+kp-g-6^S88v)F^vF>$qsPR?jU9K-_z4s5jlVBp(&WUX6I931M=u<-ioT~qvZ}6{L0{*Lie?~w4TF{^I&G_wxN!IAD`F^AHR?}VEg^!#UdGqGy7 -Zg6UU}4cB`{Lpy4=-Kz$fJ)v{=}0{Er0r%XIDJ;{0l2zT(x@5OE0f|<<-|-f8))6t$SHkls|8H-I>;L{gg8y#R-D#}Vh_{m6RkJ@~u;XS856! -;BV1L+PUuv)~GuR(7*dI06aodJ&-)OMo^pb{Ze7ss@Sdu<29+#Stm^mZHnqW;O9g5t? -8cfHyf?t4w3l#cAR0J(xrT;@y5Z^TOMM%7HusMn>+T`$^7W02cBkgX%p`i;mvSFHLd?vXNskBl5bnam -gA5P)bJzLMATRHsym@()fXhR956#bO;pRpu6c!dzhWYyDMdvLf-`%?QkB-Zu@RUe&G!B|FcgtItN1+5 -Ilai9+;o&|rD{DqpR*m>EBQTr*a|T$)Xd_1Y-PtoSos{TmzKl+tnv`j?2G!vfhrWS2%8_29|EP~sM$# -8`j4MU!lMPLMNXtrjWGqaL1%txuNdBSmQ>OMb_{mO6NHiyBXHF&G99(qjfc|C>Q5#0MjwMrw0FBSl)aSVtTl8s0a|q0z; -EBkbWS&e>L&nU>j4ZpQR*Rl_nk^}NR%*tS`o8BRS*?c8Vrs(uc+IgU5=T-qh#jd+vhJb!JMp|O55gEn -Qg(Jde(;Q`V>EBFHfm*MH+2kQGWQg7ggSGgkJNM(ohf7vnHUXF?Lth>%&=x>rkhg|G7{61vag}9I!~u -0St(0Xt>(#@iAmRkOW|9P?N<8e^ljN -o=Y&j`OlM=`=JvAd~fVp#ePMvYgNJ+IO%$8b7H(Td#e~H4bn%}a)^%3c!^cd|CwzKWTPVW8116w7&6yBOnL29%wYg -|mb0*}ZPR&YBici<->j{aeIo1g?sI|ySoMcSHH-zfap+X#fq;4Fri3jEVmo5M6(I&y2E&H%DEc-o9gNem_v_ -yZeOehquRHpWl|m>&L$@N&SxJi6l$C@gz>Wemd)NuAj~=k6b_ekvFdY{na+^gA|d7^)iP -k?=DumeMWAHM~Ei|w-%SF?K&dreExT>(|>lYQ)?&Va-;=L)OQM2o32P|x#GmFt}#hayrUDWDRE*Ny%~ -F8r1h%3GBL>sONt^>CuL&(I>X7;4hflhoKkXNsFeP)<(jln -gC~(V5jriF=)fas*mmg1qbY_kr7bXHf$LmWXBq@jb470 -Re_?Vc`QF?lI0mh9YwP-qLO$;ehaU#?y#+pt4;S|>Z_CcaJiN#RFghj>$MTc|nR@M2cQbXVOVb%wkpP -8CL>Z&;89F~-j#mn5pezvPi+I|+H_#R9>r_{*_=`q9}mP3D+$fShnN!Pfis=1g=(h06pN=IGCWT?HM# -L$_Old|BCXh5Q34R~MwXP5AtkfcerDN~ZNqbbWNgTxPnj}h+@S6$1mmg5pGm1>Uaj^8?mXAr6h>8TIY -Mmi`XF(xZDBRJDW@jq0nLSROR79t0HGn9=F4HS<~K`%d?`k3X$(3IVJ=+16&geeP>_=u@frI~ -6wjwc$FiWJy`f}@4F3wb&sGz#`clqZB`Y9iMhB)B58x~5WDDAv2iq$ee1i5@PI!BNCImIk7~#AOs=#> -t9ggo0JiNRhxbVrX@jQ4L3x;c>35>i8gU9rcrVS*^#Sld?G*GbXF(Qqqrd?UXwCe -hvT9YvqU7Ig2HGR*p4kDu#N3a=2z8n|N5O4d4(bK2ha`s&Ge^L*-~pD;^pgxcOruP$2kp6 -x8qH#au*l3FnuNL6B)v2w3KG?LzV08dO9;>N=yKnxi`5iR(mrkhW$~CGv@A;Eh@+&NJ^VnQV{!3RKUuCW*-g#G -5)1sLhZ_s~`W^N+?hh9AMVjFSxMdiiu^~XPVCEG?)gFN#1y~oWW{&?E)JFoSB^w>8iFALFot{M{KduT -%soUn*J1Um^z2tFWqhhPK2>jbL_o+VgHP(U!7Ad4WKAcY{2;68#01Y-$C6GRh45`+g95~LGMAc!OwNYIU-1%WH#| -JouHf|CTt2o4kMAt)huhhQDST7nk{o+4OEkVlYCa38^F&3_odK!WZBtq42_E__At5*#BqOz;80V+45w -c>Y~&skKqiKl<(@uEiDi5b00!DWdT#<@NU#F_XfcY#by;e{13PIj@oU|MvN_^=;0z9vAJ=!-KD*h>!Y -Xc6+q7G5*^9xt<#MQr<+j{d36Go<8J%blds?dF`adZIpw8GAdtX)%fZ21CHlzw+8UWP3P9nvv`aNqjeeYcT-s%B*_o{v`Sw#L_tKYjl?8JLl^!KIoj(cqM_op6l689L=?!8~#Kpyl)YMdAv)R-P&C8?zJn__1Pl=abep$Tx?z^I-q(prF`RC&5)vH3Q0`k~ -1z2GK5>d-!XAkR~ncF&k`h+hxMb5(hsRWoMHo(&J@aNq#Hqe& -6CGwYnJO6Kn~W>Xke^&Ne)T>568 -k@J;LtAB;S$Gx0_m&ol)jYqLzpCm8}O?-bom6mUBq|Be}2`X162nuA`ZM?9W^1GJXEDq{Pgbj!QK|PK -YQR1k&q1jcpQ6UUh{Km-xkwz9)A?x*(~zFA@Z>p?DD93<>}u2RS*1L`O%@(9bZ4degWkEW6(fZNx^w6 -f$%@LYV?81szZ4t?EjFIe~j5yS;e2>pZHt$Y^CJS%zs3osGs4V&m%dee&+ZYH&0ED7iy3CSyZ(z$vx? -k7f2$|-T*EHPI*GDsOkYH&7Uz`U_FpaAoGM1feV2b|&p!K1oIQKiA?H~bwpmBzcHL%od&lI_oMakQw%-vQVPm -*wd?$5u)Zz4C#1 -+B7;{>R@!v}^`J@z+ew1R~ym_Lis7NeZwoE+v$XP*@>yzqipvt|wF@49vC#Kw&qxi0(QgA -c^f9qYvWU!>SrA;pIuekk_t-77x&=p%9D$Pw|$C!dJp$B&Dzzy4aBJ9|tV{Z)!nr%th4l$Di<3+H|i- -(HrYqM|}vx^#(U1&?ztYA>RNXh1^aL5+4JlAFd<*qc%THp`pDAlX}tmm|e2IZZqzmx+z?E%AvwqSvRQ -1%>ZO;e9FmAPOHz;m1+Tx2UGY76#fAU|0IRqMBzU%gcnUkw77-JYzNYre5ly=6NgChD@3@E&F&Ghcb1SN9};rfvqCOgFXUT$g -gkP>5Z+AT`%(Bv3O|X$&!q56DEw*)|2Bo+O5t}?cuI5GR}}s%g|Bdg?@lq?O);cW3=1fRH59{6is2N+ -P|-|^Uwcb&VWbonr%7>XnG~1blH&Ix(h+X4dH7(@#yYs9klP^QI?>Hpx}^T3J<(`cK7ia9AAxOL{x-f%m^PGg0DgN-lF-fcWbCc --hmV-e0*R~)bNOi(1`dph@f?=TW-18)6=tEBvQclD1IpZHZ?rNgAMo|92ylK5fK#{)xL%Y`rh>pj(

;QQ^}dPE8}{$Nf*=UV(gc+mLA@%IjhA{Ix5Mno|YoPy4O!aqKww*@FfP#kTB4sG -7NIeBn=kMGwY-qI_S5*SY3ZyP$4QE2D*9&B6AM8W9pjTprq~SW=%)L2S)`4kgN$| -j9-1iyqft(1rZ;^4WAKz2!AH`yZnL!!fSq3<%i-AAwdie?K_~>XG8qNn(&|jwLXi8nom@%&r+7;R+{n -xPm&^0^3UjVxphFDx?G+JtfkLgUr*8XxlyA=u|6_4X_!OzczNbvu`o4MJe3_RHp~wd>zQu|lj|xl*iNy_$7}x8Hu7^}+YvdyjR6PfK1F&yc>cj&y~cJ9mnM2M -@AtaOU(0ar*RWapue!QC?mye*E!A@#`<=SU33Xx8KAcfBYeqU6SH0dOz~Jp`Pu9jVpIiLr3Lu>11l?W ->G`;Fg0{9iDq(x=quH?6wSNmQ*y?jKwj(2qHb!($u&B#O3CQWW@*P(N_UX2@dyG?U2o4ei7%fr -L78-=<3rlw8Ww7TVvUX49G8Z~rt^Ss?;YUA3V%^kNk_Gko$US926x;wdd?1ksnhHi-e=2p(l>8)|6Mh -zSKQ2FZ6xvfh(`bt4=rFTl-r(KJ#oqT%r^1&5Vl>Qw)KHfe)9ekL+cJyHGutNuRhAP)2#6S$VL=cccxc>FP37oY~Q{;mE}h5(ubSRq*jD4mj}Poe582y({`uz*Q$J -d6+qO-9|NZxp+Ee-2XP;q!N}f1zLQ?-*;yckGsh#~=QIs;G_0-27fBal&X=(0|Aw&Fxnon5DbBj=y+5 -i{iYasy -;#X8ubkCi4-U;5d3JD46hIs!f{^!r1m!!)q>(r@JU*h4CAAa~jmXwq@_~O4;i}+_^4g&k`v|?UVA=vr_(c@#0Y`EA9RY{#&!bC1q7r)w6 -#Ef2!*p0Q_~(^FoKeQ%+D0ATM#>NIC47l);}8|3L%gLBIV{-bplMd@kk1Dk+nXNa?%PF*R{R_+Pqo39 -`J{&(E(T$wC2WpnSN77T^Lo0*-aj0ltIh$Ro%NH6@$eEG7ZcBTe^1|4)d>dKS4==lBj-{p~aSI9N9BjnT50_78#7DnQp@1z{_t&}5AN*M -(jzR=1u$;?2#JnvP@GwP%+Th;R1;{&x!Q;mC|KKv=47x(GYrz6$p1;8J2QkxFQJ>3o{42rj?-v}J|Sf!;S~0Tp*({ItvvTcokVqNXR4Dr5Do3$QR`NE=GBFN$BrEx@ -=t9%+WBkp1LYU(fu0vA4|vyU0RB|($&x1%2~>YhAR5LJ9-~jIH0b3SH0b3S -byDx$#P=Oi-VPdw|K@k8{O{pt4%US~)%Ps_RMwV2A3~Ya>CkBb4QLPQlP%zndUE5uAo=El0rFL%VGYr -+(yHQ*YccLCCmIwf!nv{=s%*-L& -Z(3q{q^#!=@?zN){Q^v3tj#P3>eS_d7%T!EAR$xz+I;WbwBER^lMOWqOC&PhG*;i{_cgMvdzOE~X9@5IUNC_Z@J4w -6-g=&(jyLMZXn#Q?aS`bz_^ -0wSwaF;A08r~-wxI7aLsFU>a`~uZU&t?b8C$lW_QKDf9(XfbUC?q+Wm&Es;ah()o&@tXFue<&S{<{1J -1O#*e4#02W!i5eRYSVE&TK@Q>N}eixT7I}VSZ>SjFW=Y7^Lj&hUaisaj7GyFBu7i`50XWxDh+s8$DsB -(l3>O6-2Uzt*FLS0=hXf#8#!`hL`Fu&KH!S-h%zvD?p&t9C|`e;CooZ_YWK0~dOqLzrbEX-c`nwo3)Mvn@J8E!x(d1w^ilK&@Sr>&ddViMu26qt)pSRB) -^&{KT6tbdW#P~8|KNiUOr-0K*75Jrqes8w!u}eDOs&V#Ej;6B8qgi;LwGPdp(Xd+agB -2mO2C3mVWCpqR|I_~0z<~pcqN1Y -aL4HWqmr=jJ@5YTAU#BuG>6@H6bEc#+A@P3q-FLH&@ZpCavOEBH;EFubgp@&e@nWtz -(~<%6+2zf!9_AK3gn^uT9sM0ouD{U2Mtd^wMUAZ@KI%e=fi+`}n5b?PK>%jEp|^MAYNo -_l0^dOGu+`jsq8paJCqbOJZfQJWTwL7*Lm+@p;~S;2S&`jK#hj)c4eEpX7IZ53r@G -`3Z*T8@8#Zis7=4JpgFj>w09nF#Gu(g|#++?78<*eU;9zO7SY*$hJvkrn9bszIQI~A#WkWY1N2m`VSE -!GXcl!7S;e_e78=b#k{E6tp_@W}plhQXpbVf)AYRnLg!@&Br=V4$Nfg}N4iG+N9QX5{YBtIc`He`z^Y;{ -50u}Mj`9W*xFEk#W=T$2mzX+rD%Y)4*13H{T?6^nv- -4lY-GRTZ{}aCpK!aW{)QrqI^bF{aP>BX4va0`oeMd&m-ajxzKV7W4mgzSqFtDF3JJeiy)SjWJb``>{Ri)umjghyVD8$ri|bwR0Ob|(4*XFLNFG=JgSa#P#_><8CmiEHpaE?s^cU2BkO -QU`9-sK=;J!SfnK^FNI{;}6-ro}AYw@5Z_a-}P||@EJ5h)=7@8l8&HIU -ZGuW^AF>W2gjP+{BH)Nfpri6$Pqv?j~bV4GiFo?cX>5%#*C9f1Xidf+5TbEMKhf=Q~X1wQ?2+7>E>x> -$v=UAu4YQrbka;0&2-mHlV-Np%s|b|)l8|HPB3fk+cAs3@nC(Ay9RFJI=Ujp`v~?DT&&e!#5-nGYL5r -j_aLoSr2G0(n;(&slr)m`{Sd12u?!o_&Y*L*6b3m(++SBpIsLShVV|jUY#0m1oWh+uq?~Y2%FljfT&{ --$Xl>`_=5`nL?WR$gu0(xCI_Gax-#VJsB$d;!{6?k -DG^JoJl{fjc?G|{LCp`ApW3j*ix)b^jXdh9pqb)*R4P6#`3ckUPXS`B9Ux+y#%y(i;8}mCDlgAv}pQVpJ -0QEsLsK1b=_wUiqMY_$?gF0^-b6TDE)z=Rgc1wCo*Qdihk^Wu4h&q-{=C0@Dvy -C-JQ!p0disESe^BQ=G2h1WcubvZ!yFIf5pz_S>%=_axcwY-kDfl-aq919%$YMM%{bq!>$5PSTfaFsn8 -y+^o{MorWBQnT#r$zxEqSzPb2OOi!yFUj(PxWVKj`U0|1r*$)}DU{E_l{y^AoG4tK+d3&ix?g{ZbPL$ -YX3Rc?{I%O#L)@tTj)^=~F)`Vfpgqu|_(1Duy51p*2s3_stXP=ElSww -#y`Z?$ZF5m&ok5{GgydmbRFs}NVHn#n|Hot%}f%y^473L*{ICy}^SVv>-)+qldTMs<&KnB(SY5M#C`U -U9QfhVYYz>5pY7joaD(L7Ip@#%Hi`1)!?nZP`gU7IJ%)o{SW@<{1UAo%K1&GN7J&rlAb^Bd<`jB|J3) -1S>59o_yfo@2pWT>;@RSDR}zJlE8JNZ)R^Cp`DubFr8kf<6vB^nSIzZV-5Y7pRZ$4G-pT-+Z9DOw=Wh -23f1K{)dh|d-m*%1q&8T1s@)M_+hU1Q17DcL;Vj~06!p$NZUB)^z-RYe)(>G5U&N$Feao{COmhjkx-uD5|VS@kq+D7oBcAJa`agx7}Nw%)5|Mlq6V+@sh?gyf+$Gi;Ym -PpT%NK0gG1kQ=?ez`Dfs>(}FG&6YQiHm8p?QUB!U=cge)jL -Df!49hqsf%YgnJDckUtX0*=6_H=yHRwj2e -Dvtim;PefoIcW_x`yljy716-;J4m-ORiY4g2#$cH)705pHuvsX>2OQ2WR!BO_~c}d^^8N(b0 -#-BEC-mI?^{JUv8=)aLMP}DQ1`;gD;)~$p5{Y1Pd66*NL-+i#x#OrSY%GaOOZ}mP^?ehXpI@Q>F3;44 -=yT-k{j&MSS+UKwQg?+o`dhw+8uzh{sNFIZqfByLmFTVKVMXIBIUbAM+BQL%5QjL1&_19lNM{#7n`s% -CfPJUchty&fS!V52KedCQcI6w6^4EaoD`ZHsA%%_!>mSUXscZ%aA$=?N(L-dUyuWwQxB<4$X9aJFc*U -y~hs5a~2;TJLu9ZuJ0^fHCIEWS+bGeJN7nADm+yKCsj0&OtTLRr)ML`GdO;U~2Zy!qE_n&ICFHlc1pI -RdWw{*J>Rtd`3%Ro!s1)^|ex5&c*670_4lrM}W~!V7b)=#v9?$UoA3Wo|g@IOnMk_RIw}4)n#*_d?$f -{XXp3SVEfcEZVPZkkUspS#vT^J(T&*W87N3)H< ->5hLR17q^H0pZ2ajDyu4sUzAqbOU5*Hlo1hKX&UEoALpLe1w=zbrBGu{X&`mscW>{&;pDya+#lpRc`h_C~_br7v7_e67=sVx#CB1LF{=!_19C;Vj1}gf>Zj8)PszTsTZ)y -#L~K7^KDsV-U0{Fs2w;(0(RplLcpIITq7FQuxkU9>`0XFJp|+M(6?u<~XA}$U5J5vd;{e?dvoJbjyQ2 -g1<@TV4vYJ8?F{#;Ghf_Vzj?V{usY?=pwn;xY<9DUz`3l1Y;H9<529k5BeOWKUp~A9E?_5Z)l$ -ipL6||VPvKNCSfabLBnFh_=Rg>HuynCKdAo~yqK@s$a8Y?&gpjEzTmIwm=I>eHqw0l%?Ths8#MUsV;2 -76)BeAnzZA`1!ZO+S$1xzU=odUJ{1QIA+!zIS!+sRU|vojR6Hh76v`_Ms#tt8iq+R%6%!+ii>8Hq4w~O6j -h=lDo(u)DrxoQD&pdYySBe5tr$H})328-n)ARBJxq;Ayn-||6?2|!)BB_Dtf&7^K@NX5H6&f0xUN|{W -6dN-wFEKkzAfjS(vhqWLSS{m!<>8+?kN@@FW&F={mGKvhhI1Kz9#5T?n3kBFJaoh@85!^)ym;yaf#M2 -uHR_0ZqQ2-Rlz`IEI5YuGMkVMzG#5RJR-#Sl6zYse;bJ@wKZiHq_iz>>ctx_gcJ@;Q$(QmQSHOsz6m}HCl%@qIc15 -)QnnD8{8H5#tNt4G@OZZ@C@95U&aUUXE=hiBb4kT2S^Lar^R#ut)wrR@rV^_-e<=2v`!K#mr*HTZ?Q}my7sa~$@^h=9=V?^C&x(-YSG*1y>vBQM?a=5^b{2^HVe!(=0@|7+1ZM=?zB$WF-{z-W%cY8_7 -*$Hg!>VnB%Tq^i?>9Jh?A-EWw~D-k>AO-s)OpTdMT=e8lo1dC8}C|t?t$VJy}oHGxR-ru6{@_(U0mH{ -fu6t*XdXF8+wQSKp)Vb>SOw(j`X6u&R!p{zlS{AOZC#d@m~0ki^<@7J@IXL3NFVjxHEA`5-B0`$qVG~ -q>1b%pOKT~M{+w2(#5ovzDJMHB(sMVXG!Z0YmN0j;QE!_&53dP19s`YH!XA)JFA>movq+a2b@Tj%&OR -WwuPMpABuE4x?S8J;78cabaULG`;z;zd)z(g8vF`=4gVeQ$xHbH@T4xHudqatND;XrC}xRLQ6`p%$Ha -26N~{x`z_Z>KyF|12Tzo4|ing+kwGrUr-#(Ua()jQ}N_KtaBB -9ibw{AHp53Zi>Z4SEW#La)JG{XG^q0SEA8JQa5((IlP>Cs`zy>?L0TqJ!yJnoH-&@<9<5@m?kd?D1*&FN!M%=-G;|{kwU&vSTKZyQ -9iIHNGC<5FT172^4e~S0ThvE=;RX3R=(`BxFQMOUT)H0O?cwFiA@`yLt3vUWi3bqVE`KSn0ql4%$`T- -^55qKtE3{mG_xGm{S29X?+PaYvpkk#ZP@+CPz5OwKbu;n4Tls15m?4uFpb>{VEFY~n7)9P;xx2mlh9M -Ac{NpNp*NArB%Mo_U{233g~poM-xZ`PIGvha2?hJ^k#nvK?=T__6rSwR}W55Iujqix -*f#r;$BfB7O3N1t9RN9D?Z;}NJOG`=o%D-1nxxol84C-a*)K)c+hbe9ZjdwXDKtaxzwt)p0$oyXDr7~ -wCC81?4@>{{h}RWb6FW%!j`hfS$nsGdy|{)KH;u*H@hFXN8M9yTYfd~!V`E3|F@VX56Ew2M-`_WHB>E -7E0y7I*4z|c?}8`nMFx(+{c$41sRh_3n@BG@kUm8>({1#7I?SAHR+^8QE6h5x9-{PK^Nbk{GwVicfQ2 -m6Vlc-Ntc6yERcTcLf^F<;?HlcQ+q9W2?XmVedxPC*Z?(7C+wCTMr@hB+w);8*9ORe|gIK=C8O3I>Ic -z@6zsK1cwvKHAZ`ls9yqSH?3vur-h9Z<3$embP&9HI{5WbaaweiR2IsmvX^>L)u}D -&kh(&T&`Wh9#J6K%nT>|er9BQ&(ES0jihKwfH$X(*Lf@eu(rB}v$;@Q)s5u0p@j~l4>sfoPjU2~W;%s -)V1*E=oTiq@^UJMt@Wu0tNyVM8jplVUJ4i9peySWTk;7VMDt8pE!$B~eOI*A@4P7DyHkRn;6ic#V&F+ -t>uVli9Hg(y}jszt4+7i%Gk8UAM78Bj^c`g(jv -g3_u_gFqjA^OaUB90Etq-Vm_cz0eDmaA~k?X9iY+xxU2_c8X>}O6Wc|T*eUi1CR1dFoG43VHH<^OY>; -c^dbvp&I#M&p&j~tNr|49jrbp=veV3l7^C4pw>k^3hrH~uvLtd=amwqR`TrQW(<#M@PE|>pT`43P_0| -XQR000O8%K%whEkq02^(_Ga0FnX#9{>OVaA|NaUv_0~WN&gWWNCABY-wUIcQ!OFVRCIQWq4)my$gJl) -tNtj-gjn_nMnfV%AHAqw39)sKmtjt&70sQ0jo=r+OFOH2%y^upfy}tqSYk0?GC2xGLV)QcQ>H5n@n{V -YiOO`2+-XP(u;uF%eFNEZ4+V@1SEro`G3FXypwq|nW)=sfB(<_^WS_J-rG6PdCv1Z=eeBgNAB1mY(fY --e$l89&j|Hri1+ZP2{Fm@y-A|m_0xi9v|D~!Q0H&hlz;EWPk(CT>bvvTuD<7? -L9ZMFG#fBMe5uF1%7UuA+geanCB`(eAioo|E4I?$Tz&#g-GFlB7PNTbwy11xxO`)DRvqEJmyyPzF!xA#hVsgOsm87UR?~NSrYW=!blNf -f1WPt-hnk!rs|?pv%J`=iD7->^w(^@YwKoQH+)`G2(j+=v1#&!sJ~|8ovSym7UIRAf+=?5`mi=8oq>P -XsAQsQ9$%o~bX;fYiRpT-xmP8Wb-dS1=N*^MxMt&~jcai)X#(1a+i@*UmhP?%pGHE`h_n(;;x{f`<;0 -Zu|JUCXO`M&gi{KR9helet3OTa!n+`0@fR!aeC_wTE`CDmy0ljBimbPFy|u~SHFCeBYh;-g-1JvQ*Orf0bbYqIv -a89ztn2=JZ|GX>H6n&62R=6K3}2hjjx`H?=!jd?9nnO0foO@W$`|KQ -$50`f;2JSyNq&Abqdxy&#v0*Vk}q&CcrTU|3Zo3yGAq2qzseSI&gKwxysov2x`Oy7uTRTf9`5m|1iC?AA&lN_ab!+r%m!Rq9k7X+LL$Y=uiIbEZdW7@s0V{TtR+u?GlzjxNC%WN&a5k@8x~|lKgvde=qN+@je~*>HpKZ^A -EVS6`buDCQWU)-v*jK3b+MR#PPNG<>%J}etyQIfM4LmJN)tm-idrw&fBX>*Szg1ov0kr^|VP}(z)R5B -g3Sx%|4dCJ)m!a&`_Rn3$Iu4dUp44!P#BIYw=v4FK*%WPFzW^M?tU2pf&J6P8yTepyhGWo%E08C5=gI -(wwx9(fU!)n&qrD%Ow0qLGywZ)Dge@y?D2m@A8)v7T_QKxHFO7bCG#V|zx8sL&@ -sHwb^|y*y1-FXMd+*WltuZ&xJfbX3Aj<_b=id`F`4FJ7%qe{#F#_J#_4GVQXXFWvJuax0mg**5z; -{S`2-1ZZ&W!aGLTBdFi)l#~!z9$E<1B3`dJ$E_3z?(M{YdFGpp4I+53kzr-c=AzA0bh)N?O(K?p10cn -4Wy2Ci1u<{K2DczQ)_(R?ps(x>m__A8^CEw)BwTdt4qULev4dTgiAG -a%-kMaoheKf}?t&B!Tc{B(-Cr`>qPjkDRsVNS-4Fwawa$_7HB%T|re6ZlVjeHO`#RntChZ}ZTaVITD` -4H1HCSH%jwpi1ivho2qKN=a!hfy8?5BYsqeQz0#^Wb+0Jcuq6$EEFt%|5qw;jwe6B4Xtaznx7rZTi}4 -k9qXWAvmpd$S2z#+en{nN(dfQ4%1UtjNscFBhlS#qlxWvr5kWIePHP9;vIVkYnka!O!?X7oeNWpo(2MuEs+d85x-{+QC{bK7wAz`Nd7LL48kIgeEe=P{!(a7*>uH|-*?5 -`VIc0<(-9V`t@C(QeXr{>5k=%cfmCbV2bVS$V4<*BhpZx@B|36u*B>=~}1rSiSJX{Uh(8PVZHRIz5_g -NSh_}<9z8^GeB -ykygYzZ@-dsx0#~%07j1`{8%Kj575kWnRiQA}Xf)_s;iHuY>PWuiZ1=OTEUzcWPpX&}_k@LhSdJ84=b -?Xyq@KfZm;7u9UvDQx^xGMr_HnT7{YRBk-G%ZUEmqzpO_Fgg7AIf6|E1XXZDw^%@uKvJ3eN57~o`&$i -as@ougrrqM5_E?~MuJZTrE2-Rnki0=TqzZAYZ67b#8xpOSOOsnlP(^~P}Ig0O&WcUsuU!nG=;X88w54 -uiG*1BGvszYXM66gPfFA}xoABW#x^72FZ!PCIG5ASaSKl}L|*IwY_{HPJhM|ms*>4S%)5A)+r)!#85? -o$%(6cg^0QMjKr;eI|T+$q6tS>T?p{o!cegjMF=1gvSAcy4O;K-n3{*5NF1JX=uCruzD9?$AA-5bYO` -t{rsVCf|cT+1}tR#7qyTh+wDgKIYQ$+9->Knd0~^eb=6I@cl?@Df^@S8ArzAkxgFzaFo}2fg>0N{~<$ -X3tEe0+v|gWHx*?D(?vVsOFmaPf_+mvW_*WnUz&{HYBc=b%na)WF8_nZF4Xhbx=ayXtckX(=! -b6Gd*UACb#2>Qvt4(sU^%2I%Pp4W<@kup&&@J*q*rC;hOQkgGdJ{glqu_W$PuJmH|ydY_2vV3?|+}dV -_lXJW?JGBkhtWUxMa6ogw7fGmm7*$7R&Z(&a_RH-eauq{v`FyMA;Wn-o7;{LD2PF-|mhXYf#n&ylXv? -609`SJqG!y7oy}6?&qMs4WK3Avz`U0AItk%68Mo>qMfiu>jL=8SrN6E}>Rhb|g>kzZtX>`6uorI%FO_Bcgj3;pvPvC`u}K -G<_FZNT=*2=qud?{zIJSOxjEWwq24XwZpGSy_@kE+bMsMTChP^$FYYs9l-t8K#b68SHD7=^HB|HsDP< -vMt8yCBJQv-$-M&(-VD1M3CQLW6p|_n<7#-H&jIMEsO7~M9a=-v(`1xEV?A^D^i0J&RrD|>g$4islim -hO2H5GlzQ8y%B&eJ^X>|1A7bBFc0XmwESoYX)2*(Mwu1e`)RX5^D~89`Q_)hTl{}aBqeT0QF9h0L{x;iV5jw=oT2LPziHgD$?D#< -rn!?fF^@pOWZ%(&%@9aT;xrZG(9I8qY7W?X%ws&oAQnzwmr8MLfl_C_@MJoS|=I)n0B%7j+$Y&vzYq? -$F=iU0)h$px!m(-JMx=myI-&9Qj;CW8>1q%A6|GR704exZx^GDx;@372y+U6m;suksp~_aY -XUFO2Z!2aH{TNjYtn2%B<`*MSO^HqBVO;>s{?2Vlym9gky&qjxcs$l0Bw?&nGDTX&bP@GU-oAy>*UQP -P&b|e52zmrhjwaUt`VkOdMflrLvZ-BrAGNC&fpHf@@FZXN*J-cE=SIym~L&zQo2>yd}r_>quDI@H0Ox@UOpSZ1EOw!gaSlZfo)BpJQE2jTZ;R@D|urE9()Dw -Gm$QAs=i0U8ze(c?9ckp)N3>+?=Ozk>a#mnO4(W=IOxyn<$ki{5rc%hE) -UgG*Dqe6Hkpk0psJf*GJ+i)ep&PH5u+R>BQY^S#k7vL(ZCaeZcd7UE6-MFprw7wU;{JIbWJl6|jNf9h -Gk;wXp-+AZdHc6z1TUC2hrHSZf8qtCsn0T+_R -zMew%N^lP+38NqEDeYTv?gVW(dN*S0$pRkcJE&c4@ydB;B6}0`my3y1F*=JcUlm)-AIA7XU_Q#s?zj| -_z`F@T(PrDHtTw`qQHS_e@oW8eQ@QI?)5U}Z%$fiwzNqLp+Md -CU|>ZF74F(tle%llDw0C_#zMPn;y!?O5|&q`i0W&H3>q>I69$M1k!#UWZjR`8TblQPf*zEW}E!wQ~A2 -JcW-kmc?Kym||bdJ^zvK5KoSF~7&^dr}t>)@^KmK@xojdmgpThQ1SVnVGI+( -SD|g6akq_rgg8LF^A@HC-QZoG6EG5R3!1p~nz&JS<{&1MWzCbmevLIwCcRysGfuKRP9^B>%IzsFFA^Ldr$? -YG9}@g?Nxhs|T$O#M8h^z(CC>(tS9!Fmh<#xP% -3>KihOFDkEBEo0Vs}pTv5P8V+ec(?8WNhy?5qw^kdgM8_vlw4FF3a+czNOlnOK&+L;jOHQJeuNa>j%u -fBRLZ8MO-`eo~A*@x~QjjLC@9!%Al>MX#ry1Gq!2+eUTA&-NQ%u||rjL=yM@%}|3T6cV0C?>Cdj)u~zi%7zwH68)$NurMvXj4pew -rjWCnX+_y_P512l1OE`}Wj$82p?FeeQbk;pb!HT%*3aMq4S`W3iW+KHO(B!q)WGIEAo&1D*A6nrQz$V -*D;m#tXi$&k^m}LP(zl#YeT{hn3vVXJk)-Yq>dN=jADDfgqXXT`D7vsr-j38xE!0taVOSP>&PfJZD+FYl!G3i -03%h%F_e~s@|W?q{an|Gru@Z9%Hl-H=}9YQ;fbRx#?F!6#vU%UzLUWaT*xqm>#o%Dk$%y_w8FMshaOM -d9u?td`zQ+{DvBDaC3@V(o!;mck>~fCd~t%$a8OH1k(aSK>GgdfV@4nQK -$-R~GwlHI>|AD)ftMxT*(%M*{fEc&fs%*i8hwqHn(H-{W~rHW?ric(@$Y9S<1112KaS<=e;`l24jqws+vxtT~4`hk|M3cM<=7@I$s2oYCk3tL(@)TaG)&@% -+nCy`$PksJg9&vH^HOuLl1>bFbGNqVDSJjj->~)op2^Kuwi~SVBu%Ux8SnY9Z1ra5f@Ft4kBEV?{{gE -i#&%nybyd8umN`_l}kmR~I=S385XX($a_MgH~r4;pU2>p~qmCt6WCdrMc4h{1Bd(?}Oh78lo;I)`5qt -D}9fuc}CbThL;H0pQHLo?4~~%^D8NTiF%Ihx)`*jul2M^ODAN+Gt$vZxv=USYd)C{vcUXBrp`j!E}`u -HdXAFekL#+>A+f*43WvP$7TqlSgP>QQVn2`%`H4IGnexHEsyV99x?X2rT^v>n#;a%>GXBi&xMRFr{m> -#;Mr2cp__#;Q9;yJ}r2fzxLH2LDMqcXegD;X#|7Effp`I%0w+9En7cag^zXolWzFCpxd{*W?WDgFesl -54??|q0_M$_yYn`RI(%Mk3%0Q^#StLxc8giYc|)YN%RxPTW -K=#mezNG%mlk||z7UJeD|sAiyNU(@c~$M@gHrc;3mid(i$}e0{a`S$zw?dZ=|sIiM>X57AP;<#P$Oaw=1ewz^Mvy9Qu&K}5eL%O??T)5Xs-1Ix~OwrXN1|7-96s6YP9EawCB}Yygjpzvg- -4wZ<%Q9VjGRH?TwA$#?Xg0f8RBV{uOchTlF4e(_N6yBAb{7n-wN5GUl3r`%KYTfcu6*BV476X7--%U+eOr^hN@N$}?E~h$18r&Trn~#{9lNFrP(+q+c~d -yk56-+VN6z?+4F1qEPh;FaLxzy(j_|_hpx*+-fKU&{~jmu1ftRIg@EJbJ^U9J($XeX6^NCt#@V%9^N# -NEYH4STE73Y9zPSNsd&crM>TWN}X8Iz$^FCml+@4X@mR})`SF=oIOHbT{`=>MXThmd89jJo~_|v{+5& -ti5x~B$Zv;12D>jU@=qAsjQHSk%Ex)854&Hb=fbIUrp%{uj(xU2`w8_ahDW<6IyCg;B$ty{pp%?FHKE -g1nhcIP$gv>bJ+(k4G#txb9uGC9f-L-MSLABVJXY9r3Ti -J?blIM|)6}K$iy4Wj^S#4z_rema)EG>)4~lvA#cc?)cpJc-AGGEi~}AM%NJMzRpp7=tI!YZf*LpPF>k -hSCg@`M4R3=xTvD6DpMT4u$($TG~S^F_Bd0G@X_=Z>WX!ToT6^HP3vO54&AxFa*DDkGg}=&*wVA#sVQ -#jO)VlMLxk=R@94`;188GH+YIEiFV_V^{)5z1Emwn@RswXn7HkMORO?dpLF=_9kfs4l|m!(a-I<>KpKBk**_c7pFc+pin=qlbl<;% -9)mWwuh*gIFqd?2sr$pQ$E0nFHgjXW^p{#{Lhx<>4SD_rzb9+vw{;?4b0uhrx_7$NeIf1#GmUU(rqY@ -FAa8@|c+ODI?AuuH5Sh*J)lQE@>jr^GvzETTA2Q5(^fE8|mH_L=PUAl6|1{JgfVz<%1ItDDC}5E%)bE -AMirRj4KDwLrBP`N@a5{9?`ZG3{@APoAPWe1XI^UXWM8>W|P;*wOf30@u*J7;i%rV`^E$^Tw+b<{an@ -4>yGadA0`)@_u=St=W?L&;!Vi(uF;t7xme7`enHA)qF)oUPxd~s|X -l+O2iWj$ -deg|#2uLtR4?`8i{ig=1`{3~{GK&I1Dg0VD+C3i9%`)t`yH&KQ1U%aJ}$uAEzt9r|9j345RNNx%T1K7I;1K4Qt6gM9R3DaTBEtIU(*`=jYkCyuk9M;zK_$zQ0--gygPF#l(=L_5dv2(JqEwK+ ->j8zkp20Pfq!`v<`J)RpLW{dq>1ai_;IVlVJ!Q+Au?Jl0^sDtCyD@bl|d*_9v7^i1o!Xq1J7b4kA~Rn83n(u^~6M!)y>lsyc5S -L;UC>r%%}Zu=DI+-I8{%B3%6!gsp#TKFs_%PlnM0j*-+5o|0ug*>rx5U;nhADZ}ZeBt$xij#d>YTNjE -2g&@HINv#sFJjX}b3orqpzVO1%Ov)H2(+02+PFX)|NZw3yFnYqpz{mdkD$GssK#$AXyi#1jZ5Iukj8! -)yXS<+GZ*|c($Pm6g5G}6;iHBT)CR&$R`K2G&LEaUVXIWM+8g=3D>e45DRI*x -YIA{F1!1sf?^V|mdLN -Q)UPpRct!sv5>Q`*MK2FFT8&Zi4o!f~1)-U3a0{YIHvo3dm=Tr_7}tloneFYfV9esPaWa~;dqa@yFwC -bIJ*_kVwM{=lWRHt^7{`f1@mOwMmP5yCaP=>x~kn{x-B$W-lQn0;T8A9*4A{D}LyX1L^7Fa0=Q5%}U8 -scv&ND}FBqzjMLw9P<0VEyEuhso1r>aKn`@{d#nhesDM2IGko^;OIR{q -+b+N6cf3E5a5e}}%B@gmKEISkM>g~hd9e|%0h0l}@-(e7GudeX6oxwGYHcMCbcar7^e#s$@vtO$cGEt ->L=1PQ&pW&}d->u4lcSWM{5c{?qB1k{q2_76hI#9YC{snxuV6!%Lh~Eyue;Ys`$jo31Mfcq-wunPS3iu7Yn-~^o;^r8KAXxh+g4?Vnx_#+NH@f!Ssz&vB?l -lV^j*5|hEIjBt<;+miONb^bf968K0_)uUE{Q>sJ4Khz|Pw8?qkH~4+Bj<^m>teorMd{BS=JlQewf-jA -{2~qf7S5IvcNM69>`jnA+VyLmG&O?wN=B`(U0{+L|J*RXj}{2I|0+9o{;c -bUg7`eC#5ps`{3qg{O_{LZeNp-VN(S>_6ML`F|Hzf`%7H>H$azw$<(!x7&|ej(af|W&30AuOC^iq*ch -GG#t_AGh!=LkNN*_PKzDP~=fhOY2aSryuvtFd}LFme9XjfD58$5ht_#k}y`GA*aOAYqkw`sTvd?x>QM -x(oX?pr(z|LZ`zreL%dC|mh^=>GxamHh*U9bwq}Ani15sOxjaF4<>>cI!jCEke6>Hmw^@1JCmd(jNI1 -Xwp!T5>)+;o}}}Q**`Al)7yuaN?3<(3^yYdTZ8uKH|@X@Q!h*f{9fpQY1Daz(;o4my`v30r9pN)*;1! -AYG@1aA&S}G&c38(w`gRa7w5OIO>*vwoPUR_^0{==m&=j99O2V1BK+c95n&AT1q=O~W=Y!`Rz68`xp_ -}?Sv?V56#vPC%lV=mK5iBoOgNizKUIXu&t_e3d>Hl20Uy8m>rKN;Y~nxxc-Zo5@DTpA9N##kcvgML5# -I8*Gio7|zILSP1WnzvsmMDYIC0#--!IBoo?Y@h*OyGTDD%I7%wIs8?4B$TCr%yO3myed$+4j&lcbN7q -vE29xF5^!UQH|+{kA&(?LanSZ?qAvOAY9G$6TO~Gvnx<-09hA#4{aux`AgJ=$C6;eq=ahN}x^JTf12@Vh{3mpBLcuWLfA -b^yBZ#X=F>v_=cIBQEM5iH?qUkM*n6`U~RYBtD-rPP)^m8{Cg5x{b9vH^KH*6v`NPKjW@!6X!RPzxU0 -XBlYY2!{tcB`v%;HV1x1xiJ$@5;`}b+^8A%(Ne{mN8|d=Vo2ijxc1X&m*>*(_{U%TR{p#`W6UW`z_kU -ik#rM^LKczF{am3G6ym=yV?D+hZXgBA$bY6)TbIuUTD)CwJoku&~yG#TNP`}5J?;-HBkn5>zEj7dVb` -ZMg6m*uH%MBg#k*k!S@igAk*Vv0ZKIG-~XXd^DFU8_NOFXXTJ!*(;bQuS_EI82b@6pxzAbm6nKxSbd%mVIR;^@V7Zn&HZ_`P}%VD@mTV>`zIFM){60Q_h>A#Fuj)WIaYKMWxNS`Hk!0OqUMcDd^gH ->IlkA&&|8)3x=z#%L_2CY=Cc_0pl4mVW^ZIayKEOXRzzG)4O{cIY&XXnPD2I;-!GaNZ%5~frv4Rbp2x -^sH6LYQuF)ir);dNU;{)M|k<>_Xe=9z2brJl36a2qqv-yT?_R!zVQ|mRde^U{m-zVFFr9b**+qBSW#B ---$6FE;R-VZq5FCY#s&4>(9_l?{ffsW7O*y`d1&+e%mi5^G>>&&p47sPnRT(fW@Hh-D6jxf^z^N+Z80 -_Hlv+yIzPCwQp2SGcb3<=D)#ov3>Y%9&`sOiTt?{|jilwUC8si!AiX`AJF^tZlagcCPn|@@>yuiS9lH -UMt+QLn}mFZqHs(a&UHnei!r9AMlD74rEp9PF-;x&&}8~|wO>7DMes5 -Guk@4+*hGo0i;`I1%?{Mnunm8gaKCt^`J}7Id#p`&9V;4XMm(kZ-u?34xt_NEf-Nny4za9Da~~^5JR# -$D;LTI$Hal-0fTc3gwe! -#Qa)T?1z}Py3n6v>jYv9n7d0Iah)Bi(QnXgB8M~`G3f78^P7ic$aw$QN6A?;k5rJUzfIa%c>2TbiQL0SFKubf{p6Y=aX^yW9By -R>J%;WjQuPik8SrgqGTHjA=>wW4g0YX=w)*NQIK=itD{^a$s+4o2-=(e)x6J(4mUU6wi=wP{V!U3jk5 -%c8d-O|2ezbFPS-Kkev>K5Y;Gr>^Ebeg%3Qe(Dn7abS~OlyI&K`vf>&^=WN*X~*HkaxKd&;QBF}=*~e -rwF!2X^Mq+@{J_ -nsPb7;mbqqfHC>G;gr=HX1Ip-<;y_EOyeE7H5-kA@-m=@>5JtiNFBzy>Ke=a_N{;BWGiwpmAUNnv4#e -Lqb9v**58lXv4r*X;_Ob3rVY`TpdYgpv=3;by1Hh%9FMJxEFNw?$+)ChpK@+H+ -p}I7K^!`ScXG^ivN_)HHTX6hbK9bDlJntOkpD-h<5yDj$Sz%vEH>wghG)g*X1<4Y|6+Iyp0`ef=YfMYwBLd`7m~Keqc;5Tj3bGfBdIeP%ywh4+xiOq(sTn_m%)U!vX2%} -8_27Mb($&Uuv)Nz+7`3)do)8^Lix9`2+d -%dV7zFii4eZY4G-V;|p?!B65BK|Iy!ut>G@6lXx&OLEAzXe?;?(atD43o|oNIyy9oe{bXF|-Bu%RLs{ -Z!^a{<~x7Br{VLHAK@GbFYs~$|66eFFzFr0oh0$U+N5`S0=)--ch~=b-o!D(H6C{>-oo}Sc!RDIZ(KV -}y2tP)-ET4J9!NrW>aEUqpnJ0ZGWwrfV@~_B6nJyb6SncY(U!P|L+uHX7uw_$?VL-)xdrr5*bmpN=d2 -&PqUN@}3i6#IBCZ>ZNP`>x)?ChGHzKPmMQ|nVIp*kv?dcTGrB2PW$j4X!HfY@%xjrD2IIi#Iy -wnHiPxh4g5ZB$J^_0=)9)zz)f3RFjm;Rq~R!`}m>HnqKVVB`I`pvX|oC!TX8#0iqtt{6v)^Sdpe|AqP -(>6ekeMqz@xNy1?xH(0T&uLQvdpc3ayB+L@7@tm_G{NVhdo&n0UhZp~>;x_+IcuoBc@wzIXn`@EseE`~k6 -pw%3sZqT@9vM%s|vCGvej?bU|z{!WRM9|-`?itp>$$2Xo&NTluMz~0Grup!FMZ@N2)&q8P0d(eD(Yof -@qS23QLztkHpeiG)`N5PT|AVPT&AL|BT%( -m0o4(BpLbQK?W29ZGZ-4a(wXW+Eh>waGduM@0Huo`?!`SK5+-)lgL~wnUh-_$G^~{DPBK#BhtNuks_> -3)m#X3B@v^k3!7Kt#&FlKN)%Lhexx(T0cb@eH_KMj -n^jM3Ea1<}O&qm&ufc$v}8jh7kijqx(`gefyl`=K&VD`*2Bjq6|quTP>b&}W_lYCW&#kQ!HWUdweuvl -nrkTHViC#0>Y5Zx^mbx^R=07auebOwij`&_@guSrSqMCI<%q`;UelGour-Rvy4~a#$2(ou#pig&<_&98w&iP<@mv?=*UG<63;)VA# -lO`F{0mt4ce{mue~c*peKdi8&Qwvi_g(ncKgvI}rx^cSasGYSPm{%Bs>t$p*lc>)jA*a?Y8@CCUPs_S -#PFkug0*me=V}yxkzg62C)@_Xivr*V7g|?a+D4MA>}gy&JSSgZQPB@ylw)FYa~wWPGdezQ&~aBw?lj9 -hn}s&V4KNVn`DugTTd^0^A#lN)N8qk^VBlU6V+tJ;3BDPcSbA9yH`M%D9SdaA0u*oqJj+HQe9L=Ai@_Ug##@j!DmVVe -ruQt3?@(y{9qK=7W&P1NE>+!~p!GTxNCd!(5<#X)E=lA-%oFx5Z^LptWU|Gu-Ny-;V@=O!z=#!iz<=Z -UpvdsL6`&Yd&AlI{p{O|)~ZOQ*GZRr!1w#0QPv9^?IX-nxOIz|=+mcN*anA1s&}Q(hj+y}45yw&Y$^JQo^8!A)!)!wxiESuN1o?fr*@oV|u-j?!w-2bcF{!Z;ESGFVO+xS+z -9o_tn?MRi&Ja@e{wjDJLjNe0yGP~83<*f;_%;)#|yPPEbWb=CI9bj3@Pi#wD&9;!3$J&+>%ePtHWr^h -D%%d@bYi!K0rFGTtMNM>f!&mLbbCzF0QHC|S{-wqbAGhcMt(JB(B*cy?Eid -Rtx@ug#vwKR-y;iuU{22Y=hMKtD95W~Ae~Kq$pLkx0{H5`X&!gd9EK(;y2Dt~wx(~*6$Z`$wTuOzsZM -0dmYe#9rpc8zVff?&Q$o4Lv^hF8%1J6IHHpjWnp4po8k2WI>WBOj^TO?!d1 -F`;c|23+=+_lBAlCfR`(t9115$Kk4UaplueJ7&ny{)U>UJZC>v&9I?K47=Nv;!t|$pIbjRs-hRHLjq; -{5EKFhUVI<+6s~8Q-wws#}zN6iSS*h)8Bqj>bA%$`Dm|fr_e=gr+&QW81`!Lul{{TAeW^5{i87@#<1)f}S>oJo)YBQ0$n;PQwe2R_4a1e>mN{mmM7y4 -$S=#ifH;9Z_gd5|db21a4*A3Z@viP_38!;xSqv}Pf;v*>tn!?L_{56;7Sv~K9S$cMH&41=A+!#-Ub1P -vj{~<=3q+%LvZ$+`!oI`MXy4)H!y*{ajqnoqVlMb_e$<=t>_UATP~RfZwgl}|#>M8llGatFi}B37q!a -VsZsu1sL;mqR+L{YE^JJVo&NgoKB*4qlEbx-p((Huvv9w$(o!GCqrM2Uk6xqK&4$d=Q1>W0{-%I}$dA -}ltzf*+vwq57i-rQr-v2(;4_CF0fQ~I=TG^LrwqfGtm=f

D-b)8R>aR=)6<0eCTT++?J2{3SafgGhhJZa*CA%#UVY>BR|oMoJoxHhq94z3(1 -(-Y-Ui$Q?wgBTVt>{q)S(qNutk^aY!;Siff>6%)5x33mw`Rv{$2;Ra}T0JxxoOwEyk62?t2h?2XEzCy -O*Q8S-;V=@aJV(T$xVxqah7^4qxCs#v$(+f7r1^{Qt|uzwuQOx;&a|h=N#(C-pBJ6TTVo6B23?lYmd2a`AsuK`(BOvryAik;8U#MgWuUFas -$_u*Tnas;rd{=SwFu|gxTM-2H%-)HOg9Rjt?y{_ei0SSppm{@(c<~e@5aw1PlHp$?%tc71vg9&NyJk{ -HpIv7wvCC7GHwAaUM?pT%##AM#cRcW4IqUyrk{FgMaDh9xr3(BqW;`q|RN^eVp6cZWqNz(Y}0AM%+Vk -ECKxIzLuQt;L>u^*!R!%)%oD>ZnP_pHgza>;qE=0L$v5rboaBmk_DHR+g1(U%V(GQ>@uGLGcXb@;rRi -Kw`6^wD`lS>%IVUbX$^8t#e1LM^yjL_qnDdc3Tr(af600zU&l9WuJ^hydKQ*F6tOu%mu)GbrKzI(0Q4 -4h*YhcAe@BkxZAuZ{b@=hD2$Wg(ZyBQf72AwZ4%+R@w&`4(&vh`NM^n%~pa)mO#&A93A;{@BGPpObn) -|ZPF6#E;Ifwgn0Pa7TaPku1sB^foL$_VvSu<)bHQ`l2j=7f&$C|!ufy*<_9>z2DZkS`hEr`vVVF#Iq< -8KdTsC5CiCvwSIqTI)3t9=-6nXV@(1bZ?({I#kcr?{l=D&l$V87HB}Iwb)z}#PvX27gSV`CilyDScoM -b6ly*?W5%?M_mKXGs&k7;c -GcmSXrV*Tj;6TG|m?QJ>szQ$$X$6~R*vfv=lH@Q3Q`f}E3Px&LJ;cjG86k4)sx_|BtlTyt{~c(5Odb1 -i4HU-Xm_;W{C{t14JV`@59uy(;ATrb^B=TteM_`m7p@G1wn}R_^6J)HlL4VL86v;5i`oOVUTze2WpeW -@XHNQ}Wy-q;F?>=(sLj?TOVot~%KJsx(?YI`^fjle^@-LsIf;Hutey&o)c9VTsH@0? -&&KzM#Mt+_YW|#cUnKdyKm^A;SBSCa5n9&kUbM95dh#Muo~?CYbGpU<2X9W|o^1|(T%=ZrZqP1r;Yzf -d;}u0}{amQnCb|dl?A3ajICcXkGsrO|ua+a{@h*XHQ(>O5HRtu{?!AD~gR=fQJwB&@m%8S}&LsRc%I3 -LoEb}u+Po5J`d!}V>zQrCVTYq@u4EUbJKdyGDxs8kMqC3aVJ=3^vd(N1B*oiym#^wV5L7nGL^)#Kaxf -1qdcO5&J5|DeVuusUjb<0-IMYV2C%{$Hs{eGs}o4eu5#_((0r$|?JFyFBD(l)3#Q?B*As-JuR<_@iu=O~FL -jvw|?kNuMo>Cwc2hjpWAIoFh3;=G46HSggr(a5zJRjneEaofY!7+ZZ&wdS1jJ7V<&@3P2yEx)Da4>Q# -sOP7Qwc@8*LITTJmO;hWbZ$>`uzjRP{4(Xc0fGO%^I=ay9y`DPNDY0HO_hp=TsEc^15)^Z9dXeKO14VX1!|uID*tQD5~;7jWQyg4BKWm -E!oR$!Z<72Y7#?Kl!>L%7bbjBFDNWvxWZh4Y@EL5}1<{4Y(UuA5qa%d|(kdbQ0hadO1v~Hi#fP234J1s2#pHepl7ZU^ -j5BJTI}!`73I8KC3PWFzALC%G^F0mb7yd3MncfQy%NOb|0qJv24A7jcscS$;LHqSHM^;Y!2JMnz{d4?130 -~;~*%rNNqmU&j}Fy#2Z@C!q4{K75TUj$w0huw2lt@nJVJw2c1o|B?wr_=SzJ$(h>Tu?;+;m)u+vxOAveohRd4wQT8-%_ZtqLkFCC&^WnG?S0eDGfL5RyPx@8ba$$>dov{7Uj)v(+{!1Ne^# -Bz)n2*u82944W7*PUE^X?Hs*0s;4d&Ws=)sd*d)QG~H1rt!<}>#j-|o=eLvG0C$6DKGurIdSTo*OqK0 -3qoTi47>;5G4DW>@n(-W{r_mpIm(yJn1^!9JqUKAj$81 -{*2RI?xq{Mm0ry$Jd#*dKC%*T3Or8&qYcr%B*W1!AoJQSw7MR+9FE4bpdA1?b9DPllZN@z+uFFk^Z`@ -a=zWPOuxsTZ-(M`Eoq1l3~w6r0P0rto_jTO&$w4C)VoMVXhtw_G24)TQ=YZb96!-Mr`8$ZGSRYe`(@{RLZ|ufT!2Olhep#*gCyy|_L)V!xosA&nE$(UvpKXF=C*n)^K`>-z`_(eowzzKq{L;I{+6o%rp-uUi*A2kzli6Lobcj$!D81x10BpW%(y4gRL(Wa>nW9e4 -&29B7mb&ucFg#wUD~D`r%mkBo4k$G-|L*NIvyVK;@VXz2 -z5YBu#Z&=A{V2<2FG1JqiHSR^*jGaFRKYr;8upNPvo-)2;dne9z6$>_A1YNjq`0@v|S|9KNjFsST_h- -%eS?f-JzXj#y+c-WZN<1y1(LSwOt~Fkvi(=Nd9{HGtc{?ohW0@1t&{_u59p&C?Dn=dmeLue4j&Co4E? -)CJ=`qpR*4X`*-`>f+d)Mf)Uvg)qx#rdBI=rVG@3}Ww(dYC>1}5i+I37@h=ll%O#&e~gOINy>HXxF&_ -C~VCeer#mV)6N?@9Eo;CW<>#)VlCN(Em%$xLvvGH5Hp9KF+yo4&rdMg;re)@{BvezI_&A5j+p_tVm6+ -$=IX*TiCH;z;}R=HFLz6`FDT4@`TM1AlG3bIe-&6Gu=l#_!hxmu-{ -s+6eQsj308O{hF@M<&|^hdEN`#?Piydd!TV0@fpr2fd18r83GB%z;w+&LOMvI`tR?z -y|yAr3H}uX2d)TMaOTf@GTFB`*wx1$G$TZ;6}az|zDj`Y~xW``2$7uVm89WJ0kx<5f!O|D)YalC5o!6RkYhD0?>bZFKbZbb`+Ql-55>zTf3KUM>^qW`{Upi -`#mk;;mOV2`*%SCrA?lYikO|5Z=l8Ndr^vB=kL>q}&AY^vdr)&FTbevd{F1kPYwNNnz=Uxd`i{-H8u2icGS-%N?y>Fyi%7BBHaZ&r)|LON4lW)_(v?^44q0-{D -<#*fOY;|n&S6w($qeyy#JNu%nS=-I<0rK?cx3mWBbw4&2tMy@QV(2T -R@Zh=tCCLOmVq2t~^Mt3LDaW~Snr%Y{oB?%pO -B+&7TCLJHP(DCp8n{?bK={VL;az155vV|e{D&ZQxT8(;nEG~B^;L?z$`Ya}DKN9Yf?+*8432iy;O-O%r0&dSI!Rzr^3Z@J8?O9x__Bjx1h;! -5f+1mC!4Z2ar^;hhNEP}tW8S+|g`ZV0f;R1Y_KAX$e3SDQY{*NGkZ*WaktD0-+pJT}V3<8kBcRaV=(o -QIgUXgpuHJkfhFFScq*M{DF=;1w-qtBr32n=_0n`bT!B%Ha#@_BZ<7EQiaM6!LK8LM)?XY&24fV~Bw0arNr~d^EgjA$^vdft}S$FqAa -m5-)>XqoB-A|j;Y1YY>S;sqx{StjPCn;>hJZUH8zFMU-JAk`Z|kwF2!0o@4IYZDtyFRbuQfCTw52<3* -|b=!N+Vxr#Fe0hccb?cZP_b2O)$(SWp^-;c|VHEl7C_$Kz$Ly)w^Et!a&|#WJ5CFxO-8`+a7ARjeKSguYQ -Y9tW@Ic|m9oi~(I&AKT+YJp8a(&Zl&BmILAX6VByhANl=`cpQ;6q{ewtOdrbWTCVm%H22v`Jpacz#aQ -XtGG!xudC#>*W1rckoCV&XFx#;(=i#lklrwxrl@o)O*86GsbD<= -FCkdA`W>EV-_SSgu%hv)klE-bf(1C2FC!cnS_Je3!q)$%k`_T?^LJ)%I -xbJVEkosNO=%d@*JZC+^*pdC9hww}rvmjrdIUC{r$a0Ti=-MBwzNt2G8f}} -h&whW7O$3mR>w^dWMKAif+(&Wzd8KT(cT5)T&dz1w%4e+1W;w2YyBJd;(+JC;4NY!~G4oXcoA -;ceZu$E9U|%WZEt@lp6q2lecsx!|jAp0Umt^_w0w=1-gS?@rOqJ+!xEG1f<{si`>jpDby9fHXO&qWjB -}vfGH`n`i@Hnv~ndxZ!o%q)?|eY1BU=UIXxtz>{vCMe^$B#W2rOW}h+h4#4(qN4?z8fy<%)gtWP+^MY -shtU%q4X(GaO{LXaMNOOj9_%%oOq~UvSfe6Az%J1{dv`j}>F2G>g`AAn~rb%A?o;+`zY};l^F^ -1_sc^H-rx-0aRqr}=WNf)|57PI?{U??^bcJRAIOY%Rcec8NM|mFFIeR3wAB!c%l6Cy0imRTZulhn%tq -V?e9^u#C9{bIzzgbqcrpEUM2Gri?OFT;Ve9gS3Usms59Ern8c8(iqaPTzr>nu?_(Rt!0f$zF2H}FjK0 -Q>c-pm$?uo^yR)J-$5;oC1|o~zEmfbqK8tNv^(mSLOZJdMK9bJ54|Bki(T_WvIkw~ut$*nOl2OTD?G)IWY7>9R|De4O>m1Ih -Q1j+K99jylJYWo39|+`ZpfR1vNc&Luo!x*mCMgDmxkDNFM;v7`s@onn7YIbxqKb5B<3|B3xsZ<*s-Z_ -9H~631_TXxdIKMBj+#HGG!q>`=F3!1tFr&tMYaH3M?(OwL!h?|@sZg2lF;vb&)NUxn`KMf=!3N7TB1)v)zhF>T2cushvaPuU-#Q#T+DY1r`{(Kw)c7G25;%-}uyxHtyW -@vkbTyc&6mk%#FA@Qt`_0X_#SMdNDd=T~jcP@vM>OKzU(hX@=NE9sjC9@N!)bDmV-|7a`a_}G>B`@rw -S%IeshX!a*+Ej8Q&;!7yA4P|p3@_LkQ-IIiI7RPCvNV~xyjvoN-#P5&X^Mw2TaG!3Z*$P^cj#EI#4WN -_U0|hWQoAq((Y7GbTv;2>ma|zt87iO&3mpk*%RWGhX;a)x-EqA?+>r1+O{Sl$AZf(VV4cs?SQ)5;y7U -s_6em#_nrP@>JcABTIda-Jghn~-(Qk$X~JZ`SpNKSX!`LYuaf^IMv95lJ@3r2)Q1gHZib9 -6zC)E|hb7@_alWugcu3e!wP*-*0I@pA}*$WGzeX(+T*yHRqD|27kXClYWnBv$#I?V`y7%+VWORG*2fn -AAB=8zL9jiYl8O2J|+4PZ=(HeG27peCfZ-n<@i8&8)*4yP2_x)Ix$K6`w4vVuOmOlRu|ihCU2UMa{>r -6d!DhCK5qa#;NBz!fMvj5bfT>_qn$kine+dvSlJcrkn?m}h4N>Zk9l6YZ{|kg(VBzUmY~s(zc)KsA<++>rbWQ+qUVAk7eK*ny%7po4)c>jOz~1Y&^~HXJ(ptXRhRYhvq%=8t3x;;d$!)o_ -Q-jA>Y5hacvskzu$cS{*@nI -{AK~`F`Qb>*f2!jj(lizu0`gc;!<0zNGPV3g4HQ?@LxvR+}r!8hP$>^SUzizOroPeEI%EjcXlz{~`1J -hgNd_MDw=m8XvQhKG&J_xo)LL`bXO&4UfQ|ta9<*xqchlE#CWe=L(LEyWUo59%C#%qS6qC6Yn`^#Chc -(nO4mSrLN%oOZ}l5p0`nbTJcDx2fv1x-qZ=4bKy@>a|9c{psl10q3&of^@d!xzfG)^dS)AR)NjQ~wgs -*OqMm#aGBnGSH|`+=U%Sce5<#vl^qc$6sIzlhYM6)Vt6btZAZjledrR$rqs~^Jf09eEzR|#^jUvCS~4!-zklAOq89ACb4su3iG6$qwJPy(|j?#c -egx?PVJe!9na_G9#53hY-#5n5=u^ILQa2a+PSY!VCS}*HX<<2&T-B0G~bs;;&$$C5w~+&Vdth=?A+HT -uyf<#Pi*Isq?LB=*+IbF3hR}qH!+ -gsoerz-KhMRy_(u<&v{l~$-Zf)B8+Y!Id#? -JjH4*-GRIX9+4Or7XWu|M9XK{yTJw>)F^#_&CzEyCtn0Uu&3xs -HhQskxr#51h+v&Ih>~aG$mhFYOiab21CfvkRw=+_VdE@BUIPH8?%j*g2EyobLbZaM9eEvak1y&NCQ6k -JIwJpPggALEe69DW}IQJ4Mc&&5NBO*Ur>WC$}V!;nDZjFJb -AW?^x@5Nsx{fCBhQ!Se7;;Qw~cu?_lWuY<~hrwXDru>AjeoZuZZ)FNoUT};rNx4^9;v4u0sddgN6slM$_8E=R(C6j57_y$E8_mCJ^rf{p*>n_W6Xh@eX<90dzva`=N!?%y)Wk|(45HwnLDO#7tHXkcTuK&AsX_e-Y;}oQCl`Wp9Ue8(aPUm9(sr*hK<#)P;->@<7#_trx?-;M -i_ibt9x6wp?_rae)oNncJ6Zv-I7=FjgA-@AAzpLW>ZYp~Repe>&yWXMr?R6-Ar%QgPkKs4Z=HvN(ESK -|vyHM`H`$ZGWVxB?p&xbtKgGvSlZ|f5%S)9PLyS_E)&Q^fhiO0k7SwY2{Mr$mx-oLZY -2|Q-?Yj^`a8*l%;VS99GUzkPwhL&#=G@Ta}Cz8Jonxfx@@z{^YiNEcvwr#TA|ipQFao~qn#9= --lGGYcU62h=YS$j7xJYc58#yC1l(flq}Z>^{h9ro*Q%w6!``>Y#Z_JV?{f~1c|a0icx0j(NE(HS(VGMasFg$F1A?tLfL3eo4Kd -d`fbFlCs6kDDq_&fwZN{XflG~C;A24dGm1w5g2DDltY7J=Az5PA(m@f+5&Qmgdipf9J8Foe5F1 -h1&T6Ycc!ZxnKD(r4kC6>QRFXKVI4BkefML(5wf@q1eR>&>yoR44BzD}xeB;mfhu{?La+YUMEr>3AS^ -Q$6`a^LZE7I>L{!~gM4D%Y?Jz7=`{{eB_17vgTv`44(tqW9H2J8@g)bQVD0(Qm9vl2Ux9l@u@5Dc0zA -nqwtHVvqfMBOft85;rWfcpc&6kYaZA -DxLpuEs?VmQuYz-Yd^kJ$QMZ@Z|@dIq~t4Lo8f2SQ;T4m?d*_4REa}{Rr(bMU-a)$vKy;`OZTIulnQX?BkK=w|h9R;(d?=egMv6`PlwSt#5}+bcGtAJ+jjV918 -Z`@>J4a+bn9`mfvo)gO6@F$nPLI@9KOhw;4P?t$U080L@c_-=q9=N#XInr_;P>{GK`r-37KG4ikF@Iy -K3Dg&o9yHnNOG{__WWaK8or4%hkXx0oJujaH|rk2$lZZ`-(j`Dke>3$#;uj1Sfok5^7>>y|MsZ|uRIs -5<@U$ua0dy5835gUo~ZUqEff*j!2Sn1$MX)wIU)A4AuoHu)cLjqam1EzGiPqqYT{G@q8ovSU1c+#$%l -(-*i7u|g+}*@N_rdj{Z8;7+JYYD|Varb+_$A(OtdHPL6-gF@GFP+5poXwsh>cWv8o-gps9zBi2p9A+` -cMSb=KAJBY6{CR%f#GhTf0C6Ai`vFRiG-y*h>|jscz|TXn>+c)2PiF$NX$q}n>gV97uqT~VUI;n27@# -(-XTmoE?bH6DwZL}tPLf@JP9MZeM4C1h#AJvTR~%;!Hljb~NL9`1QQRN)QCST&xudJN9n`dAOEj%A>i -uto9VzVX{0jVeyIK(+asy(~JtS?*p|anhdbUrwj^k-A1eE(L4Vsjx)jIn -6FBwJ&k^&45W9{Ttyy?QP+8`6K8o(o(=AC(Xbmr>34|6dzQ>P`I$@NbphUUt47w|6pCj40LpJ$#ofmY -E(dXba5mSAFO)x)?qI!S_W1%`SX>HU{yGx7XJlU{Ol^m1WLDP%P{Zx0q#P7{{d(r3A9dVv?DvR2*tb= -G>Ri3^L{qLc))V^i=sC^4HEzg)2(Quz8ZP$v(c1`Y|jxy4VGd+XU<}>SM&vcUMfQbRQWR!Ft*e`U_x@ -wkH9>2U=GQXBN@3mI&xvE53^NDQtK{;cDnIRjnx6IO)c>gSeCeP4ZsiT&DKR76R-o9SlbDZRQl*fJi; -@C4?ajX)ubzc4IQJFQBmx#I1H=x^V^SI{<+fD2F>1>)KrZK@Y9N&<*e1Q6-tSdc-^gQcGuYg@Ef4(C4 -e2q=`4e@fQT_{H$m&<~68F&^X6Ylm;u{E2CKbptddbqzky&I-yC1cBbD#pw`hE1ItcZ?dh#pUHJr1lQ -YE$#qKRuI22XvfQ?^}y%*3Yq9FdCjS}12J8VG7%pN-_;>@VMxz?{%M -S?lJ&K@dcM!Y6^OKF0S)E(*YIzt_!@4*zGNEnG?K1;w>Yo$E|m2CgxCk;jBh*D4;v29lXzA}C|?7<@* -LeSHg&GJgSxO5m+5tJ?2m}Lj5;E1TQ0O80;7ZUJ$`l~OII1b15PoQt|I+LaW2UuS%Jy~& -ey0C>jci%StZ_kIkg3N4ux`PTc-22_^(m}A1O1u|5WVW50wnkdDFdzNuB~`wvLCxaZEFnq64rIFybSO -09%mrYBT47p?Qsbm>&YCdh)$1>iDg=O3<_EDlvZ4FW9J@J1=OO5@q}hU*uSEhD&qR6k^*wx&IBXuUoH -gkkggZu@U_|Kkh#nQlCw)FU1L;k2s^gZc~_Td$`|PM19iLqQ0!F$LpiHSN_L;rM{*M)fc}13YuO$ehw -~Zd$kAkc2T{*q9QZaDA=#a(MfizC6CZ8s&-W>z4EX_t#g7vA(kYZ>+B<^PyYx@#Odzd|Au~V -2A1RFgMCP1b#8vhq-9P_FIr9{JkQ_pY)B*&hA+r$hWYUEh&!i+!)i?w_Gxfz*qCghDTcA^IP^X8-?Dt -LzgEfYd62sl(?VK<`mldu`lM)9=)3O|0>d9h&gxi{T=6iy3W$D-Z5(UO-+VN8;NHM9b<0_$%-U1{)Y6 -*RrG&1_2D`CJe$sZJ%?`MXK5^k>N#~o!2JHI(Nfxbp^tVM-(Wb~6uTppqSN<9y4=hjSVHcD%oEv@=EJ@*rZ5p?yz{iF32bI><-fhSfn&lAK)+An2+j=5C+74&;5y(77_J)H% -1ZD5|Aisad`i+DV()orvF15*zE1J{h9udhxRH{;Bb<1-q5Ua(bODg3-}MyF|V4(tq~`FeHoyDJ)>bb- -a|pP~9PliYU0o;ectyE@GnJ3)i27(s&p($id*s~tAUS_WH1_+0{XXdXQ;J)*TB9tO^zJ|G>8-qqqPDH -`67F2`{{a07?QWR|&2?s$6G1%#)+pj|PKbF5>NpubafUbIgW7=+{X%(HTgcOI{o+PpHA+rskukWfE>Q -2S!=i8k1`M$|W6Ce}6Nb>si^IF269y?R!cnF=2B<6FZGU&{}0vGFNy5LtG=Y(xIUqgEN-0tNe)?vzEKGVN5k7F9;?~+CxHp(C3cNDo -_x8od2tf{whXicTA5I&*ViERIKw7zi9)fLMspC#T~fxRz;KH0j@ZPVYhmRh2Ri)dk%QWjK?S=?To7DJ -0%&W#FNgmh~u9rOapO8OxDOh}L3PP`lPe+Kobef{!L>VK!vSNu10Wk$U7Lt0DE(zD?FO9`8QJkB81u| -*dCA2eS%jzn}hA>4id{*`71TOEGJ1FZf}O%E6QaF)#R$1YRXz~>n8&JmBK=wXT9KNs!OV?q91s>_A>dvoah8gR;JF9i==m_<4eJv)wPRA!=yd8TmN^v3e34A?uukLz0cY@p9|+<#xKA5o?_k0?`Nf0FC? -id5GV10N1n3A~vuvxr5DGg-$odTiPxsSbDKjzSS*-BBvlt+24lKg%p{m3+|x+<~r=r!U}kVY2<69dYm -#voBac|1ZUz0)2loj^$kfyU;k(q8bxhi1u!YVGD7W|3ijyW-Pcu?@Jr;9>m6TER*VhGv?Sx-)P^Xl4z -Gt;=ZL0J$9yJP^ue*4Vj#N5BA7aa@qovGZf!u=kbZ-lJAkY|61|_$VwkmeM{?Y97pp9_${R_KrBbJse -$T_)${PXL^(;6mELBD+_;KW<(u_0lvT2gzXOenC|~=uw$?-`v3cl(me+13*^sfiPRZShyL61V`*L-SL -m%&7!4H<6mzY>?^6B6mO$W`jLE^ix3yHMvfJ`L~UgNog>+v@4x@l}{=6Nk|qn&wR6UOZ%<^}5jPD9q1 -R+($&aXTioftM#0d3NgM{c=Q{`MKmN3reNNzGPiamWqo!!r#m0{pom`sSC__>{C4AE_AznQ=oB8}3)WQMUo2m -XhoHu0Po{(r4^KXsihxvV>joT$JwzI8_*0Y!U9^T{|ap-&Vv8le`y0~YzsYQ?-zwxA!v`}`-iq|UzxXVV`61#nU#A9mThaEAwpNi$P -GM+^XCU@pEU*7ekWQIIr$*|J>74LQGMx+hCXo)1$NN(({E}4a&kras>hzx -puXF6d7X%-N&lk-zzk~6+A3x?EJ)`dcd}}+n?|e)Sr(4sdo}riN9m$Zu-Ws}8#OAeTGv8tOj9tOtlmE -22IL!lmq(g^E{@NmZ%1D+C08i0qUs;FZ8>0W&*NNxVI4i4{>XFa*4S&oIy*T!DmHTJG -c+r#B+XK{^^Vt=lI^6c%LKvq;xrw${g4wceLyGujnV94b7nbJ4C0p1hnbK5gTCP+tWT7VmA`jlbO)@Xef4_fLU{vm>cd& -N`k6eEiK3ptMFQ;$aSz$|gn#YE-%f^?0{ty-z=+JkRQ56 -*^8upH0LQ|03g*@&(4}l8A9y9A%s?*86l<$T($0jFT;5oZdb!@_c=gaf&uZVg8P|yBeJ3@t)y&@mtIz -^M2|0pB&5a1�jyLH@$t!Jq2NPpE(e)BB+GIjc3o&4qu;sZ1$yB?FrFUjP=xeD&?#_#)B*nQt--1i;%BzWI<|FkIgeJuRG@97g4zV9>ko`HC&qnO4wld-^T-Tn~iUpp -4WRuFW(aQdHp7M?y*?lEwF__>9db-rU3cDQbu1J9Z{_F;V)3t~{;{dcVe`-SuOZROdC(BUl{9|rtOku -LQZIQ5{%#TjfXbRyV&WIof#W!vnZssYkRcdrumpTJ(M&*VH`Ei8$fXp27&>y*-g*B(n3alTOhY%|GDF -|6)U?w6?4mT`egj7M4(^v~(Er78f4IZ5Sv~1|$1nyzx2R_@Ca?|A<0Ej -~lto-G;5gMhj*b68^|!#Q2WG{L6G7Sjx9lT53&+W0#X?aF}Ktqgi3 -HI4g8yvq>*L_Btv>z{0KVC*ZbU&4^zc2im-!SxuFu6za@;#gOam0FnjH2)3Pn(XF`poT1eDtgayzbXi -1QyGLIHJ+Y8~gn7XrV3v~&JI4r*_<-ndsxZvlHTGISk);U%l< -b4i!m4N`5Be9BSG9dti6PE(G=F%Gn1u|99!5U&EG1bPD|95V!pBfA8+{J(;$hQ=Ho)emM`q>ri& -;vkV@5^(#Vl#Gp|P)?o-fb2@v|9m#4D*E4^n!ZneL-^jyKfL3Xq^|8`8)+wWGX`l~hli#AX!rw{Krn&UqHRq*P_=e5xW=S>lz5GMr<5eQ1S%Fhn -GL0>RO`!Z`mV!QHox8RlHi0|GwER1i(M5-eo}kq<()W10SmKMdacuw58I9$4Q(N5d-JTBL?Lm#0<`@Nz9wvXEP9e}dbE~Ri*uX(v`;)0%j4 -qKmDJ9ts7-#n)lB1oazOjL&CG*2V3sm^O^T@(7<$ef?pA!eC|~>YI8xus@&~&IDHq}Ubt$FDgE5~@7JY7l1{|0Igc*SJAO -5ABpe&-(z_x*$t>vUUK(g>|F3)NR8sbkBfn30oQzm#a7QoKv%f43ETc4Hh%!g@PR>+NCU$2cSS@OH$d -V$Q?#em%W+>}GW%>20kTo8focTH9sBIJNecNY-ZHr7$U_d565QbxwM49lR&6+>kBPs4y&@qQ(! -pwMyhee;jZ!?co^}8%xg>%=& -rrf=@%u4gU+$1GfAKh)9;j&Qr9LCeP$AqMxg7h*^kF>5Stk7*vus#)QDsv@w6`G_(t1Fom7Fk7na1-k -)baV^*K%DzYG0bzxtCPoQ!TQ)jyd)Yn=>uw@u|36@ywb1-{=Pz -1|muQkfeCB?V%Swng4k@h`n1f+s-|cKJp*g<1dSO8Z@hIaR+HdfF@J!Gbyy-H9WANk|^oKlwF;In1ZF -vjh_D#8~DG$E~k};oBBIZ*{FWzHL4d`*<{{D*AvIabwA@*-}GyJyShykVm^4lWv+w}Qp^U@mkyh(W>Z -*Bpuc*s24B<6|FEsy5=6;l~7<1NxH54^q8ZC}mrE~B@t>ZROfQGSZ|c`AGTI(3V|ubcmL;j!fLS0b>c -Cffh%_YJVuC;oMc$FPXyHXGnyxC1;)vO>{X)**bQQl0Pd^&ju7NZ0w;${-(uuLZ^oavH{TA+6_w#Jkd -I%zI>WZ?`Pg(g|@VmP@>@@FKD2a$CNlBsTv`J8pj9HSY6luH6LN)-*wnhC_~m4;QI{FHzr&x`bbc&HD -|~Ejcfw_m_jkLQKnG?{&{A*1C`AG4E~XPElcz)T|pM3yw3zQc!0_9O!l-3m1~!@{ -HpmR)G1pGKTGUQyGwF=G8kYzDj-jU&K#_q)g8f<|)mW)z^*!Uu9UwK@MU+C-kB||M!XS(t4gtY5kN3d -YVjmOdkh*s}4@GR)6?sasF^>xfFIb<>vStlJ{#@L(WRuKQDOooyCs_;j586w@XRRl>Y?Jx-qKq<$CF -ANu;O>W{Ap0X`uYK(UlCuoRcc!2T>8Y&XK#$SPiwbH%WzMPeU@~MO3bkvl#E4B()xDm@-XZy-PG?DI& -EMF`8vt1W|9?w8HVSLx{T=_Th%|TWIIW}O|Ft03*2K3{SwjT18Uo0nx{Bt_z#-X!0q=>yS6`p()dYU( -Aa~Yr6sl5_kLHZyBo4+47ID?eh=EX*G6rPr|%4@4my>8F6@93oTo?*hkkV!*lRkEgDud2lIZ>ToYrzB -rQN$#s>8kN-k(Ty!221`t3=5gFY+FxwFG^1JMj%*nt@L^h)*1%wEVo3@-0mx9x^VRI8i#Llg0bqqd-Y1K-&XsVy;B^BascFU}#BQ~lO2*w4*lJnpe4(q^Sy$8h8K%^zMHut@ -4}izLg4fxjDeFTJ+d*Q3V|%MF=3z3+$^U%1ag{4nV7%dv*fYn88m5R4_Jgv1hyc29Luk2PY%KfyXgyu -Ilmbv>=uH5-0idR@Q=MP0^OPT)PU2eHTb*~sqDm~rjY9$+O1e?x`G4dCY>Y<>CtcD5B~DPx}m>qA}KX -2HakT`11iMEjGwr{b=Gaic?a&BMHE-l5W -$ANT0XPR@$7~_c;i^jQ(>1tLnq_yE5rrwB)H$eY^8#1txdAvls0i1&n&sooXxY%aXzI2cEBDLRKU%Q6 -pNj{GkwUOvsU0ZLF9-v6>r*QpcFPom -v3==Q|Bw}gz7^POOqf!!ZspkJ2F3^0@(WF|6`>J+-1TxCYl}1FtAPda~AeIGCmroCC#;Tnny;wFx -ZdWepqX{W0vZJ?anukX)Pui)2d|VL%fo9iSMa(z{G)#(cC1reX@I1+^PNvzASpZj#$S%O);X*#fNn}H -(C6@=CFtnjJtjwSK_enEyM3chxOR3G`VvmJwB^VeJn?h&ss7kVH?(L2akzj_Lv6M!qRDIpV{It%Ure& -wsA%{O=dwhHPzwlG}KlHv>pGSWvX!T^M>{1u35~7*ljgD-X~jzcx@yfdsK;^p#~n&&!ZakZwHe`?ht3 -{+y(*v2U-NLie~?6#B78-o=bDawNu3za4zr7umNz*8ih3HEy=v)BKV`Vkq+6KAmSr+X%rZB2bh-I7P`EQ#=`dy -eFucRv=2L{hu{0 -&E}P$0Y|bNhidEA^v`PYB;w2aS8ASci3|@L0orJ+yAM2Du{@`xEI?B?{}TAbsFuytu15z6AX?c~9QJ3bis;m~RIbC-Fuq8}dvrmXZ -lEV7Bs@o7`r_8q;2HuyM~-gT*g -BocOrXzGv=`Mptd350mSU2oH$OsUOs7v9@SaJIKrE+09xvVIE%{Ju0X?46W9Vd3`%pi=|1!najF=q25 -I(Kgy#qAvf2DRICQ%7PeR4#7axaaGr&LMr_0zgIOnuCv`Vo_5&vYIaDGzeX{k6HHTW3*wsN8fZy}4;j -t4vdZ@p|TehBl>m+vT(%CN!lzxnWaly72Go?E0QMdMTwlO?tRS?-08XynohbhVAm?ezMVYsw-bN?xUl -}j}sWp=yRGkpz9j2F$wQ~8jtJaS8}}?$!w$>RvNlwG=0=B+nO`@xkQE+F@XMbO!z5~OxP)Ogy|&vWbm -_f7twDZj^#;;tyjdjPxEf4{CCjvOnP3uQ|-V#gA^-bP?Li{2XGfm66>|md^598A8VsEZbIx3;<3=HVc+F7albLht$q8BfySv`v={Nd`2GJe5qJI-tjz>lbJZ-?V -NOVE#=C0b123Xpk`IW_1Ry(@v#9Pk5zqV6N%Y)zdZEj|5Ei#OfOSmcanZ{p3%+dKZ`gm~T;tj4>{gDC -LUYsfd$x?@i|!V`X{dufg%4gCG1rql1^qrHQD>)^{0KMOlL#%;O-u|m(UGsa*b#ofyHf+k_Nc4&j_G49YuiGDao#o -1mYedg=fW&cTa_gtcO1$Y=^dx1+-mK;h)|K-fxEWf08qNqyKs6|KX7SN2Uwye -@c}8E7bpmif1bPg66_*ht&)fiyQuw49K!7&883L)d6&thHLyeEb4m~$cR;r?C4g)VLr -p6Zoou8X}?&~K3Fhc(V)x184UUYn(Mn29&fCE94RmCrTUS14IY9yKw&d4-hTi#2y=y70N?cUB?g4}Ge -azrw8BaEP+~D;JA2G^PWeiuFiq#Bah*V=zq0tpIaTGPj<VTR#`I!-U_M;m&Es)>Y9;!+XfCxaW}|J%QXq%k -x4pzhEmBHvTN>%u)vF4K*2`RJtN`P<@tZLo7tIx`WbU<4T32;7;@vI-7CwAOBJD8w`nC#LL*ncEqWw6 -Rsfts5*|fe8!>h}}$Ai|qMezA00#}7uSgwiN{gUiX<(oMESWMk6ns-*>fBxIm!jWC9aOigCfsfA-O6Q -DK3&H1+4)+kvrP+EuOKmmf%buqe_|I#3cT)X0qyAu5f=8i#z(*@fmz%f5u&sCDJw0EMTpajtSBytfm@ -kI@!v03OLj3 -$jT*3SFm)l1y4JE~HH)*GzfH1N -=?dWYT(qEMg^2Csij{e~?2vRK&5Q-rl9GgUms38}J%!sPB{WaDGX{19Nss9DKXp#aqjB~;+C@Ch8fk} -b@VSr}ZBylr{EJuxKbIxGSE1vd%z>|53gQ_ur|CMJ`=~yWscQ{eN%$hL>l*udPYQYeIPK{tiGLxFd>z -em%G3D=tp#Vgdu8|**qo2)dSY7~n~%86j|>O(R_JY5BZof@>OP^o1Q^s0d??yvxh^`7Bl_myJ)`#<;q -Ms2-`=2dd5k;QucZk5hScj=U?@fBg~v*LPFmBzPwb}sHZ=Z8yKV!I_$QI}*L5U6-Eu~Q&6lwCr?N^VI -q*i9z1`lh^Ye|~B|f80*ihkK#JyC+irsR(8n~F_EHsuMpf=+_YK#xM|}y!l7pnc^`$x`|Kn>>rQsz+zS&Mu6^$}#Ht_6csMyIC5%)bo@k?D&my@^O7?hTw -1e_VROeZc}!jj?raH<}vX;O>4^H7|GVS#=e`^s{QTjtlP}h_w6Z?Y_+OI$4O4ragvEIw#lYi)vWWyZm -~Z!Uc)ND7aL8aE9+_a`Dq;UH4yJ^u+#UIte*BCUjy;*hBV<%(?)Iij@ZY7JXYcN_}pfBdo0Vnm+ud$< -NL!Os4sqn>}(wc%=_=ZR`7ql;R -^0?@=(C>){XA=*Go|nVZRTmVEDRsbLYG0aDu##wi{UM3kjWktMzo8G34(!_}>`gbDq@_MnJip^j2L9# -!ybtN)eRy=Q*yG}hvpnF1KcHu?()XWf{M#4AZ5v56HP=wvYv3Kv3z(;u-c3;&``$>Vwm -7G31D(LXJL+9)HqK#dKqrrGN1rHc&dYIUQ$T$>Jp=YhhW-(Wt2`d(#c9NhfKS&6MR9()Pw=jqnXJOp? -QAyRsTR`wtHB&XY6d3AM{Di*`*7Y0p4NdrCvvgV^;(dmwij?E!DhT9d)=4AD -27Z}?Nb58J@Hf2Mq|6>+X_AbzL;4_q;MdAo}(y=F7>ETVUZm2_vMUTN3~V{ZJKcq8JkACSSr)4cyYF8nQkS=>qc1@5(t{ -IFR@d7=4#Chq%kQ*``68$Sz4@v@^@i}4@lm16v2bbUS58|rfr-EJ>@|BIv%+PIqZRn$9M*S}#S_pj5+ -#%YyD?{KH`%j3d_W|dwxWYV4d98lcJ|KLU8N78eIpFxE5NyjsfvJXbLL;S<|x$0?J^OR`!*ZiUP(v8W -_;w+lRWJi2sIb?B+&)t6#GSgnzeJ9^?rpm@De@8sE@(DGtCr+*R;ZE;2YJm0_9>WdodP>KFuY7_9w#K -mCd>{U$8hD<^Kw_OY3~DWz$*dD)7}9?1@k(?M|{FkPy4`3~}|3Do*f -qg}`^`?$3*#f-}zfVqe_x?y3eV*IXD6F&bfqO@9Kf>+a^4#>B -(dM121mid^Mcb*o4B>O$huFG5ejslYw0SR{<(^bhdv^fWo28B%j>oyyxyOn~j$9uv;zy#q^pNuMsk}~ -F>u##sC@*a->j1sKM*Dk@+*OY=_wc!#)@vhrHBvo@`5qV%bK+ichf$6=A2=UxvIyBA>rJ9x49m-or+q -5U)Lg=Sa$`IWh1C@eI@Sy-8(TxyC9wE9KN3C=&;x=pcBcl;5dC)o5ex?g=XQ6@h>62wx8W- -A-*+&leM-M=g%u6Mt%8|jO@i*(5x@1$>7#KGhLU|p5+Ih-Hm5g(z7dr&+fvrYv|cE!Dn~i*0Xq?i32RoAx!B@)1!a~ -O2aZx1RksueM)3Xh;=mBT|Fbyo(eOn7>^pSZGRSzPF>F7{be$Ha)?0u_MBkA6u`e5}vzKC$0Sd;( -Y`c%Lbm%7mWyyYPFr+L(^BbA`V6pdzWEP6^xLUp*Y;n5cCNajm&Vnzis;~e{Bh2wzjaK=IMCbMU)aWd)K)EK9s~38IF$>T$G}a?p>l+M -V~W>kXO{3YbDldV{76eea2^i94pNz+`$bK6>hhXXmj}x^extm6PY2a)*r-)4X8Tvsc$VmA`b(9_b5fg -gIqncQFwciCXf0>w;%@uej_^C3SI5_hMb+(>BJdzA-%$gvQT+hb&J1R%f|4a1kS}!ocP_7%;QQsq;5 -K#7?8KQTafb${mKaJl%5Wi2*?<&gk8NMG=0}Dv*L|fB|cK7X!@vJ63^bXE^WYACef;GzF?KwOI`xP}X -NZ;=D|2f(N9=Lw_XrrEn=4jqu=-q)?%(Hbp$JAc-qcMT0^!AY8C6T%a>8qXME_gOS4}>4P?EW5y(SoS|$6U)B9npk$BHSt@Lzar(HyFL-JpnjJ6mT{H}ePRpA0~MsVz>fXlA#n~R0pDV=?6HbC^< -KRD`A~Fhgwer?xDGg11l}+BwT(f)N*OV1Ns~CwW!#pupT>T*rg6N8t+e-%+~vcahg-+ffL?~P5xkGn1 -onX;4~OBDM%YK{@ya;yv~9wNDAhYNS?DnEKmK;| -&1Z0Dvy=})6O7nmJPRPHJvU-RT?03z=-fJ6aqH;pd+oJK4$?R7#C*YTG3$^=Rsu%V3Oi~}(ICwq;dozT-oQ_GcfywY&rs;FILhzH#(e -Qq8o|DPKV_%QYuMmEQcQ5++8hH!%7jsit@+RO&C0!SAN~YH;=JmEB&HH`>KQ9s&-@soAVimQE@}3SW? -_yD2C6z}qx9=6dF}H$vL=F?H>%x6;2CL&S7p@lZ${dx9`?ol575_2E;UB=))Q)Tq%VHi&=>HYWqsK3Q -iw$Or7`^yS??Q1d_+8GD;9U-4p;YVlJo~gDcExzP66;`+)G=9{3K5TJ+*iU0d&?PA9%2hTyMz_C&0=+ -zNw95Wm2DR46MUhQ8OM5Svx}G`z=dnOoN+w5wrfOOG50Mz&SqV^IL~oDM{PSz+c^)F -8v2dIPFJkos0SwecLsTp!)b}Vc$Y_*_o4wp+UBP%r6lQs73BRPE6;oy0|mHV_7_&O~jdc&q?FZC&d(! -u8mIMhxW(^E3PPkVa0#IaFwvq=WD(j0hAHobNWjs3wmcC3ol_W{yb(N90*-dRoMra5ohNyfmvFN_|AEcC^bca~vby_AF{U+GgBu{u;5;seGKzzkXE6+TcIgA#v?j5N`*zWS5S)9Qht&g?H$uYfKy -Or)eyn`YHOi$BQuuy9D?R7%^3FX2s;uMSPBSU+2U5Hhe+TWv9btGkQT$~1`F0HFKcA*c*?c~%z6i2iiq*Wf#JbZWu=w-WN-6J=-e -daYp?<0Q{?TXN)mjk0p^A;>zfH0m_)|aPILNlzdAL7dBIbyh())=gXIH!Sn55A~4v$iYz0SP$p)E^+2 -i$!Pk53xP-zNHzPL2;Z?^S^h$ADu?v;vI`dVyA;6X<2o$|AG;m_Omp68D?xn6zO9@wkUC5I<(eh4n$Ip|fuULEZKLTNA3E -T_f(2ak9QblM9J5#%rzf}3oXva#hb2#Sudw^n;)TsM&f8L -|Tl0B}gf~Z^VfhTS(g@#d@3pwa~n8xZrY1e?UqsYwOM*X|}=MP4>7XQ!mmD)9(=_(Ro>bdEC{*{Jyp% -l@&Zeaxm`x;L|;i)-vqok04)aS3#+iXyy0SQd+HhUrs?S>~V-Ls2lblX35$-e92VYX?m6iSm9TwEYzv -(QaTVT>|heB!@X`nlG`pfsQrfz!*3h5v@{1|IVSZpRjI_EX>G!b4-Vm$||MLtw|e2d9x)CYz0x?n|gTz$s~)tl-~?rK^`BC=sI$VW!v2+)k3#4GZ|%Vz+EprtD!Mj4B -P$GPaz^DtS5l{DOJ*9 -u>Y74I3nA~FV9nsabh -&~A3X%yRt;NtR53FO{hbamgyBRO^YUw9aWfcOgFcTvpd4ix{o`g{jHFvp1PCfsQ*{jcMM;u`c&F$U_;aZym5QAL!^`vRK@Q7e$;k#j62e@)IxeEPejMXHC)UN%APq71@n_H$5-T-@sep+iCDuj -p~o&vCsi9KUOcE)eL?M8GjM+w4Jf6^3+O!v-t##>#_Cg`-WG}Z^nDXh(#<>%&8(h7MxXLg;_By34SPM -;s*|r$q?HpRZ1%xqw-l&1asAZufwF%Agy+_jpdwO#VNhSw7JVzH&;b2CarDKWJsr?}4 -JxZ07@x3PH29s5C2zO|1pp44SR{7ijhr(IQB%&)+v6r}Q^b-e;c+(r&W7oZC3*^ztkG%$Vc(1B1Je=r -=1KwyW&Z4C?P0>fZ|zVsoT8M^*wWVur`B*u6-MW+lD9b_3N_EvJ_a5zWy)#7OLitRA`c -by|b&H|g|V_T1;6C;w04!rwL^;ux+Nm>3t{kiR11!f)g;`7Rt6{?V7C$A`Tkq@F7WE*KY{$A|5HIT#= -Iay@;1NEvUu^7-*$FElTY@22^vNL-#fu47cclg4u4+nub}XCDaah|lQt{Tlu=LhgF{Dx!s5m;YQ3J}{ -a=G&xQ3-ter9ZIvH(6}rzWN!%BSbX?LWDsLW(Uxb*c?JPSV@vn_|tLq2Q*EHvg!2dU~T$D9LWgvE^(= -?iY$FSA{ENb|-cfilXcu#dW<3)P-1K?S<9=9|zRqWZ^<~fe#R>pF?m%7H+7C-PL$pZC@^O|v=FYe??& -YXd{Iei3iT@mB+le?k(m969Eih1wC%@dd~8Z9Xs -T079R=OV45>o)z%(Y!(34y#Hn9QB%3U3-B)bz;k$eb>4co-@w1n?{G@@W3OHkQ#Kr`BK%K;ML -(w)uUUPs!X?+h>+iYl`*Q1=Vt@r6@_QyAt{YKZB-ytXN&y{WDaH -8-)=w2|<4qPvYyy`Nexl=Kp}h^3Y4>E34gfz`f8xI_dM8fxSy+2H>|3G7PoK&%+T-M8_$=UF$HJa2$L -}sKBk$qAvl1x3g=lZ3xoKu~!0m(%V+NKl>4sUPL*mS@gLuvzx^60PsnVSn5Au0eSM^%)1ILvucIX_i> -jo`w{)ziMM*_`vk{KP+^jg?oqJPkxg1j)@S#*%w19i$F&#pYkb(m++=iUDheq_LI?O5li_&a>TzRZ&M -X>I(z7H0}i{v$Y7VShDrP`txfq1%GW$e@?5cP=J5bRcOTWz_}&p2s#b&d+oM -z#7n4PBF033ej|?57D>nrCy7owUdnC0gX9I2opN5+%K}=ZjoGX?CrZ%78U4Rvk8~V1e`-AY6Wva}EbyMq{LUpKzm4Ws759bQ!~18Vxh3=Or^ -`25GLm60UL;N9=blCNseE1+Nim#8MUtrwbD>B|;IVO`S3}nY4n1(caW*}TzKbLqKRYgxGI(2xq>K1>f -!?OKs7~J7h34ws^&$?+-j_ril)bXRRoW}d+)wLEdU-GEeN5nX;7{{80t^Jm -95^!+qLS4s7AKvI5SF;adZYyiRG&qLO%)(I%T}FfYIxHoZr4I#%!nN0!RZzImP -%(|I17q~>!TRJznRq+~mtQktcP6)fTyy~&JpY2x>3q+6r!Qrv=Ss<(Oy_LgZ2KzAqUlQmP&ma2TLsrpSH40{eKL99}TB6+MK)PMS(|%c^ACL3Cf$*Av_iQbo|Vbcq(|P$= -p~zcaYon*Q_A+kcS>C`iHpx1U41+Mw}Dle%z5L^ReCicd`D{#5qT~Hk;K89+cudN9&d2C=(xI= -C3~he>o#&;q6utmyFSV6kOve&e@n_zfX~Hoz5|NP>o`RF+n^HTge-BfrnL;y -_sJp}r#Wf|bd;rL73cr+U2|ARjhO||zx)3#<{7~$V{cSPEI@L%8eKjZR -MWBKq6s!t;sw};;2-Qbq%bNpq9`>@JsmRPcXiRv4=9kMs+EVCsp(_AN+oI-9?bbI0HT*f(DdtJ^4)@y -TcC!+xqpT=rvk?Knz9*`Z2y(FO)@p+C*EZ@kVU(b?J#&yaJC+axmU_O3!#&M!Vxkh=|12)qBZp;*x$*#fj>bR{bcT{CzO -e53+u%O=WG#_3c$TFT!iY%BVZ6)u!a`6&A$P1jjT4#I!iTO1R@c^B0`yb>`jRbcKG@_D&I;`36RwWK% -)>Ax{Qe>=u~XMHH^Uhdy+avsHgyH?%uBWiPptm8LP8?tr$rlBQWg=N&9c6wL+;l#FHM(;MBB)S$?EOp -|HlX!uZ=A}!rEOnn`g$~7B1KDSll#vhGpq$8gF_-93Jt+JKT||!_l55OdhN1nRO3pWD+uLcF1q@5j=N+78GDaqL(rjk$&7aueHDzB`WXH{~pK4ySNlTkn -r+EVta8ws6zsYz>dcMbF$8=Gkh&opDMt(%*bm+kJw@#Vx1x9tWn&Sn(FKZl6LjdOfhh@jT{lVX+P#<7 -XI?j6Hl*?xoL<>GMVVy;J`2c)zhZ!ft-!%Do`Gl+{{qGku$~bU4R&&_&z2XEOr4yn+uG -p13HYj@1uv^xw~=^`WNvzoyLdBC6T@{0BMsz4BaN!x-+z&~%c5TjVv+Bz{%a7QfprVL{aTEr#5##yb< -hh+y4viBt5(#NGeYgJ{08Nzh!Juxa7AtnDdUqaaksmW-lfr+XpB<>_fZ-5#_>2b9N!223M@w0?3yS|d -m1ZP9K)8OKIk}mp|{GTi<*c>9wwRT5b?CfY0rY~6JoP=&{_g*Vc(K3@%_HQ@gdO+eyr5aYBNH%NXI)@J!m63T9u@=BTVg;V!$J8hHTD9 -K{9RD@dd&-)u-mfy)1YYyD%lu|5t&e5ar>1P9J-R=!DX|&(VEg47tLnFIXF1&Gg8Ezr-GcbEs5fy!y$ -vMWxzkwXqms3(O-igCRuY^e+<#x)g1SrS`*Nxuzo*joL1}6&iW7Vb!iztH>I{#CNApIE2%|Ax}ltevv$DBd-hy|Dd3zpI(KxkySJd6 -W3!QpwtzLu+IT?U$G{h$VWdOf=W$iRqJ+n6+m~R(Z+9wj$>8BAnYMv1G?uap$CM5V1dy?jW -T(P5+O_tAP&MpZ*%h0+`pjE?SNlN!ic1?E|}bY!%<5l)n1R>VDuUHN2^QhO-y{T-Ddky7FP0BtJ9XbC -CEQ@Y{=ke@$|)>rH`w&DRp_k0S7|-MTNF>wFOZ+9f#_yc;LZvv5X(b0Cgwlc?J$Cb7CfSzz19SqnVku -4ab!l>@{(*RwpF2NJL6cs8=Mz%#1H1Cj^ldrlg^yXTMRd!F*CjXdsW`DY5X<&63m?#F>^&7`}2W*J-E -KcvTIU7n+UHl(n^1bLtjv^Tw>J~pI0(n@+xGx~TVZ0;queQ6F$u9QsHpApTbb^X^&6MH)O?(GWftOv%i+tdZ2WqMWVIH(A=M=`u_RL*%`xI<*gt?*pHJ@&z -A4qwdaV2nmHB!cTh>N=88LX>w64uO_Jg%{J^aq8-ml&!?95;zr_PuTv^Ytjb|Cq|%M{~7ZpRa -?a@cFutXoq!WBKm=zku)?D?TRSv@TG0`TXdRjq&^t!0aoQcqFoVW73j~1DG@@qr8<4k9)q5siA$#^`d -?D-P7%*?_P~DvWf^ojqSFa<==mPWqa+ioUKd-je%(6Z5BBU-&o6#WdT%mI<~Sf(J$5(omSK8+n)bZX+ -j93{p2=(1_4`K@-_Cf}Ik3v)S^jVF$4FPM9Hex=e^P77$DIoG*`4&b9rFqEtK>p7iT!^}leJ%rChbI% -7wG@`LV?eA2hpPXJ*~w~d!qCQ*t#sHHMZDSax=^LQP$|9_I2&s?6f!E)}U1QrnB5q;?+I0){3@~T-x- -w9Wlh4IH&{)~DI9AtAZ91)(ok*9`p!UzAIU>uXBhlQz9 -D!bh`H}Y|&S06%ps|_yJ?*_9hT^Y&AEfibfR@`pWe+5>%6AikeG2v`Wzzn<4xO6*YcIt5{4}@ldxYvo -zgEuQ)XM+UZ_xP;{r{Kdx7Z=yf0(?{j(G$A^z(R@`^{Uwy(f-zkb@*Ub<;c_I-cg~fxdU!<$L^8{(R~ -WWW}S(6yRx;-x(;Z{V%!gTYFGuDLuaeax -3LWIai$5@_2sYf3WL2D8Bm29-~`bNU%2wXCF}eOm_ -QPp5w0FZjh<;1^me_(jXV#4oOTYMfuZ6^&oq{zPQ#!A>N#_;7HA6koc*O|)uO?mreHXNK31ladA;)>eY1vek{*jhv8s`(YTFc@sWzC1 -Q34CIcBKXAiL9Oz~gIYMB_yY5}R0_`LSyJ@*eCoWGj<%klZ!w=mY#fc&;~|;8N<^Czf;>o>e9ty`+qB3w(lg#hO8 -Wq{5q)SRzJNHM@R8uQ30*A7MQua7{0=oR3|q`t=Idmmi=Z3g3}N$QZFY-nvHEB8*dP0gik%}kZ^)VNZ -q1cuaJ$?*`rb}+heZ#_Tjm~GM1p6pY&Swlhd!cbC4jG5PoY7j)mPDzsd_-Z+AId3Z -2Wbs|l-_3li2CbV``k8-^o0bfd(&E_qJ-w0Iq^noZLGPhrkf=pCTBA}hm48y<&RjK9WthC>AV8?SnMU -aa*)~42!ij=}>GI(06^ebv1&W*u8j>osz{Toz)ZM~9tnqvgcrnccS&(0?lg2Ho1Fk -+_P}3R#O<+%#qB{X9^8M3cWL1&Ka67-PE -SccY^j}&;O36^BnhHc<$bJUhBM#K1C*W3}>6!F>F6*f0oujMs?b@le0~plb^5-oWm}HZ>nPNHJcB#0^ -@CY7CTl#eAJ=XM~0@e%6{2SYb#^qgklFqT&3{k7PjODwp*rBbO3fOt6$pM+ApW@^uQ$?p)np-GQ4f{A -9h@i$Qix3`yf8Fe`EDk$F$L2P5D8K-Lu&Kk*~21O27Xs?GbYiY0v*oRs(Bje$Unn{4&9_h~MwtPM>1oQ(lh#;y#>Y*3Omm -9CzEeqbK=^|F*v!V*l?8?j>0Nh&w$*`_M4)ksBH7Mf_3R-MXts3S2ZU?Zl6P-?Wj&hx_mw&zX2$O?58 -CoejNTN_972?VQziqnsVw&QfZ2&{|@a#`42F4ks%-)gP2QAN7ei&wmTqcmB{Ju%Do(!2aC4R^3w6#Ev -Z{UISU9h~|u=N%(l?FtxB6_~w*u9?rkVmh@+vS#ovt-38vP9(##ocII597PzV}E65>UY}V7yj>`qToU~55;#6>DNP@jZ0$8m>SHysxmTrh7qo*J -$8Vl|e_7rkaGK97;db-R>-aopb?DbwVI#<6GM`f7oQOjnjJKHT{k4L4H^F#|_XG2q=5TtV>ML2L`kWG -CP(TI|cwCUXijy|lMGQvVtE4-Nhqabbk_Fc2xVjr42RAfq3B}~v`1s@HpNGp;+9CY3PEbEq1BX%;emd -XR?K0oiak*fBxK+pHf{n`CR^YKBHl)pKv`gS{r917^wk}J!d=ic=2HckEPTUV|d0*JB-E3yLh4R1W=h -OFIIj#0}lEbg1H61SFq&Xw=Z_xAK8_)89-`u_~oCcrQf3x2I<4eME$!_?^g>lLH71kNvw$ubqJCe8Ulw${q2(k=|ebuCV)CPV*9S-5=U7EYpC# -pegVgLhFP6jrtZ)c^^jlQ_KirgJAv#xUCxVBA(DF&H2-O&Wm{UClLFW=Kj+pGn3BX`-tYrKS@3%T6sx -Ub@~rUju-VL7|%;P)z!ZDR1ZC^t1J_lyAd -Ya%{N&d(^R3Y%~!E>LMMU3F_II3D(d3ZdAiTch7;Sbvfn{CYLSNa6j3+AWc|ItOyU>CeaX1!fzl81$t!1J*Wl#PxmubwKbU&~?7vAM6hj+gy(F`d-&s8sgMrX3{I*-#2T=7P~HMInN%`{Vr(zXVCj -qFJ<=PIqLWpFqSAE(#pHG*gw_N0?#P1Pq#x)6*j#oeZuw^bez@)EDZR{+Pu}&Hu#Uj-uda*1r`w6i~D -Ze1@&v%b0W@_XwQZ|-Ts`%PEB=YJ;%?RJ~gcf;zMMf*YZy957sBwI4+ -H>EG1&&(5qML|EjY!PbfHfH3*18Pm4(1uj4oTs*61dBUoa9QpDA~~@p51NOD5)>iywFyakiy|+V%W9eq<=~(bhjnQZLS -6UL{@TqZsDtiLq0=2D$i!y5q~DfZD#w2AWPM|_8`{5VJbJ2PRk`=uRYqzZ5!$Lefz_08^Qlde-)IkBJ -unN6vK(>tQ#+QRp-CyX|UGkfgk8LbB69caPHG{( -eMOGRuK6l`i|pm4$Bv5R}BBXB%W6Fk-oB(`_++5KTLaoDzQBHue)RsV+-fZL+Ju<7(PLG7K%9>dM*%I -56VQ@Z)==?=J`qH!+%je3uGs;1@Ps15AR8y8kR(#gV<@|QNFT5(5Yv8z(pY{Kr9Bu>KSTenqx9dU=Sz_Oyw(R_wvWfO=X< -Q|R@jQ|qi@vLLwm=xY2TiiG_7jpYAgH3MFroP`5pTHDSiGxpU39b&2%f&#pt)H2CQ|13=?_B_*D&GF_ISa@Fnus^@#%f*=Me!PWSr_mEiim)UX~_Z$tiZDFM -ZmPsZm6YLW@MIRR^%-*E$oVEg?%XzEzwL-F|n+$tgux6&&)h?cJ~O@`+MK_|9$`e-+OTV>}Q^tXP$Xx -&YW{*X6FR#f$wj}B?fc`emp!epy_*iKsQJq=d-L={S91pLtLS*MdP!Z4uNRp -Qc5Ukb`qz3E#0?u`?lnY}_M!^s3@X1c9UZPF^L?al!#nt9^{U*PeRNqHGCxaF`#cEvw3OWI&_e|^4dA -oD;OMt}LC>LUwxffj_S1|v+Rvox9z!8cJ4xd2a|;{3&hBmL<(Hp*o$=*k)^!?V?-lC-x@OM|m<6~9KF -@Wt_ObzXpOz4CHZ`}E!Co_lWP1omSo -QtigEV8ZG9A7kNRQ%;Nxhj?>laXoZP%gRj@G--*uKiex)J#p+;-gQsMwU@l7HY~d%-c#! -0*Wlr8p=*f)q*(S_xd17N?;%1X23~!Tz7z)In#Y0E-vo^SRVZ&X=xD`C)f2maMa$TnG=cTEo?vHtu=5VEvmMwOK)PF~kA2)oqFVA<;_+MqqeOAZe(#+aoZ$^ZPccLn151pb2H5!u3g*cI&JpO3Ai7m??b0+*6G@Bb; -hO8!sg~t%IlJqkfNlU6R!oR$(y*xT^ZT0TIWT7UxmNT5zmLI@S*b}$?X2Hl->U(-E_ake#3$La60$Ha -NklyZh-E#cB>t4u3DZjX$f>_lwd!0Qmbf=zZ7I?(|*q -uD=ta`#kQadm7$s&6WvY|3}L$?4x$nfe-kz!L9bl_jxV6w?w_vJ>DD6r%0*i@Bq@)GEFM%mLc_m?_De -l=|22Y0DSL^;RL>S!ti$Ry#otZyL+?WskhYM&}&wvhd=vmX{O}Qe(#nkb%nHQ3)~Y8PlEJpJv`Xi(2? -%1q+Up`t8~5TDyc_2wQDo<4Z^O?QZL!A%~DUsbm8txNGkaJ&J=hi-GA6)A%xeX2ql8ASy)YdyjqWu)f)WCgWCHTxco7&fP_!-%bF7RF0*LnCE+ -3rs8onUwe_}-3%S9n%UxN1wsIBL1WI5Z!3{o1U$3Ut?X;!;k(x{lq#Z>*co>IQ<G-2I!&ds=UOe{{nDa)a8udMn>=FJDa(>wL(KKwV -O=CVU?a9UCCv1Kj&Ch>tI{2}*|i%&PPk>7E{Z-}^2>jh^Jj`}7^DaG&>)H(|b}V-IXU)Ulwq0~yiqFx -hwEF0!s3?g9Bbgy^0$^fw~Xx`A*H+vgKLen8*l)d@aV4(PI}4UI3Wvgv+3SG$vaP0*KI2_QGZ=zVJnS -$Eeb?A=47=pGdPX;^^n2K5`|eJH2<{D6*8^t-3-2HnqVHk7$xfbK?X?*k^AD_H%i$@JZHt+Z(OZCF3y -w7)ug?#5?$Xa{r*B7pwR1bk250q}>L`!+frL&qZNJ|~wT{Aw@aKcOAj_c7ET9RpntJko>yHk=HP@}S{ -r7Jjv-2oJz;!om~e@Lm}1%fk2c5aB&B+@FPalEZtjaJa9(_9LB7xDk1@l)hJ_)e7ES#g?iU8b`Bw -<%X{i^F_Y;d3b+nSkYwsMdCyQ6<<*aWi*XQFLuU8BDrg43q&hY|T&=<}1xeW}0zGwv1Q44LP1pNTdzP -wD}%M$b%59^6xzD{AiFwEa6EC9naPGN+FQ5(GBevj_6y8!xMx;}pa^uKhEg$2<6(mi@u8j1ZLYB-&*3 -c}}>sdOD1$O1tY=|M&W!ZRj)wwB}Dd-=gL*S^cZ3n}mH&wMZ_xI6!CF34Ei-e^y60zTDU%K$!`u4~Kg -?q}@H<20&tAH;SbA27L|$2aORJ>?KL3i+E5cjip;0quk6GksvYfDswJd7A`1H$uF~4LKdnSI_gMaj9+ -=U(e%%t{cM!k{kE(bs^R2{cZT(JrR6wK|aRH-%rSNdePrgXhu`NdR@zX3ZB<7%pb!RJB9gTnB6H%jbS -!9tPt7(53}+xUKbX+*9Kk7#`d@HB5WTDIN#VlzbwAT!rk872Q`!jdMtd4K8wF&^m0E7rmXp5-gn!DkJ8ir^#fzw6HTGEuAfe#{Ngc5}(CcTli&$Etx#bw}0oBYCab4)>B?(0>~xug|?89eqX? -&j;SibWcx{$hJXF`b0jm{&Bi}bPbMoMd0eFfx7(YcDnpP$$K^39~+(pMtW<=x+v&dCg=wKK-YM%I-vX -B0YA^KwR{~CDIx2ozC_l&%=bpGY3;o_3C+z#1uC+R?oXD4^DFK6Z-Ce7RqB6H**E-Mb8KFb)|tAmZr` -PcGN7#y`M1`}_vNUGF8}qY)qKpJ%|}4FwiWVnfz5QCIGqcO9Hz^s;|>AA>P6yrpw8ph89mrLTK5ftb2 -zXvvuPpt5boFi6(mWkb-l@vijwAM#eVEG9X$Kh^!B9ptrhnab^w31C4N4amp-*n<$t$Xsu-_s+cUl`D -R`=|U**OQ8#OV*y)`jWT6GLO3w%$>Q -MxD4&F9Xt -eFLR4$`+TDfgTRJ`5LOr#w3>vGBeOv^vaJ;33FT@e73002jL=S>qIx^78EK;WEbti&>1R+KlK9we5{M3GjHqMaRuZou1uJU2eiUl=~wP(NE*qaHTX!smfkv)afMzD$2{W -kcXNq-LyNYAGRT4sX`pMRXF=RR$!8iJBaM`#nSs44QSKwxt8>1GAYWRib&xk3=g0GO -freb6>m@#>bMz88L*`J*`7pUlSyW^*NybY45g -$tG$S7HTdHP-7q$m+-a%?bcB&BK5#ClxRZp@P?rlN>53aXSD7d2b{hASbr!FR@t56vwq-+`F)b(SI9LbeX1lKNdWYR&%R#o#1!=^j~NzF7Bd{$6|LVuH?o)Nq-(lpUU$*G)S&H6h4#~d>JshG1L -A?aPcy{pO|Beh1M#Gm%T(gQ%aqivaP=r!{;_4I44N0z#~#pE2;bp3bx#@89lbtZc5An-5xrXrCa$ -S;Ghiy+*M#qrwa>!XQLV_Gjf|I|+BAI-A>@}&FaiF3s{ZlEHr5#nl~oYaP?&_`tUChO=sQ8GF^>XWWd -SVezp54b=gSL)mfH#+(qKF=3O_+I0m?B{y|i9SY8$T1NQ`Z?A=fqwr9bpJ7@IXD^*&f^Nl7=g5{q$|1 -h=EG#2=H}+g(g@u>l^$dy<$*WPRMK}Zyb1T$13^YH2*NITz`dlz$AIaWJq>4Z?}R?I1Ai75VXwOre|2 -vjmtKqIxK81;U5ap{n&v1Zq4Jf6O@H+=fQNA>~SpNuSP-KTYQe5|^4d1-t0zJk3}&(fnj9{nD08=a*-ye`k}YCTSpR1nwpc -KM|G>j{kcwd+#4(5oh-SnG5jP-~B9o -BjpAN(Lh^2^|5Z55)A*Le6A%-IMLUczw7|iiB;#EZNAsqW6jzLUB%tBm(xC(It;#S1ni0>mFL%fXWJr -w9U2q6P6W#7&50h#w=?Af8729?>rZ{eU%M*Vv!vBqy#O>B_|sI)aBcB38`g*kqR&7TzVbn66|8kJqrA%bVg_ -)~`H#r>a=IGDP2S*K}puxIQ)V3gn*1D;4rufc9@yCW}B=r2Yt(riL5YQIFmw<_Y7D&#v9yiCE%x4ZiLph8}$&|jsHS1b52Mf_ -R?Kc(Px$Z_J7)GOo-3VvC^8x{FqRmhtZ^6LsY+2Ja8-{C4(D|lN4_f@3tq~QJv-b2AP3f@=20~LJm4p -)DLDELT4{7?lSr{KCBiuSJHNjqHIXT}cK_L7IZ6Y?U3yac%)@^a)okk=ybiM&Z6_uuI%58df1Pegtf@ -&e@DME=M%$SaWtAa6k4AGz-?*YqLC+Y5i~a`pE_gmuvr1w9D1sCCKkUUZ$|8dY5bY -8}-TwqO`><-=uewpm7x1hg2eHj6RECc -^C|eU>rZWJqOPNNGVdCPS{UTa#+YGT01cJbc$`p}!eNNN@N)9s+v9XWo50UXIyd(SyD8d$5+bpFz+lg -M;Bic>CS3BAW61-vmqx<1e2t!91v)%HoeW5qDi;wEvZbFMBQRacf?yv>l?fR;dWX -<14KBhF77T!pp21}YDm-Dvuv7_9Fq`eQnGTab_=sDiM9p(ZPKHD=j28k_xS*q{96>ua&CTWT4+3@#by -Ujq#CW}EPaY0+hDS3vh`+BPD6&~Z%k3fYtbYp`E5-~O35+jx8^jOk3H3~KBau#eJ(MQvh^l?x*_%NoT`l1YCSvGkHT{<` -fm0AZDP~*n{U!*!>M638Er=BIhGhKnly+e$|>Y}n<39=bGFYZmUBD4ZFaY`*W1T#$&Y`d-{pL_WmpXQ -RHxHk?$0!R`eYm}Xa2fcR@*EeTWhC5#em51 -{+*sYcHf~%auou8dNoY1`$RJVf8>75TZ1*#Qso$XVW&6x|EDSA4U(K&C=7IYBexYrhTi1f~ODa+t&k8 -^rwdz|f3mP1*etsQNgP~8e|sV<$-lQ_!m>GUD3aH2@xnvKfxE93oL`2QY%e*_W}v6C$L8_2oIo|QGwA -*b^_K2GuBxvj$8iEmKwCxJWH#~KA+tKdZnzD~iPQt+o0e2>bN?^ST|oFdLM@m!WotR#ix5DPKD-w=|@ -d_YJet-X#$NOYE@JM<~G2ZuZ{!yiUHk6 -?Oer|gV*5&27K$5zbeH4J|T?fMes$5B4loj)V_dpg=EJxccR`m96Lyvs2Ou?*4oJq$-IL99d6ypM9k6 -2wNt*bguqu@O=IA*M$RLCizcgU5{UL`Tn`V3G^pO@QJ|3WoDYPi*kb29j(Rl0}BW|NQg!>C+%28NOMV -+%+xvhv~DN@*U1Tj^pprXTi4=h?NDBC9Et)2r;?n8}?tKF9*tLh5xgdw_Gg_`N#UrfHVgK$gQ9!gN5s -%9$fu5{2%J0IxH+S-6eg&pF@ -WUm>uM_6^l5jAkA&Trk3^5v?-v2)`Sf_BK$(F$_CHanU6(VvbIuIdKjQYwp;YOT9^u -l8i0EQZO(EJK*tXbp>rk7CzBDx;5OdW#{RjUG+PW7km{J~_)8mIcqAqC8IFlPv~=GoNCx&DLkx4a#`& -&f%UI9wk4PM_Aw+Jln>@)1WcO1bNfL4dX>eOxK%btDM5=z~>1Z1 -Hm&pWpPtiTkdQ(RA(4zpmf7*-Rvro4ZFF -Pqmk~Puq5PwD;}c*RfOQE?xca?AE@d4-yM3-n6cx=YbWU -R$tkIZwDb()!py8}Q;vC&#cH!J&dtkTGH?EU_b*6lUH-t*We+~|?=JtpJOBT*|HHx~CQh0>B{FL2wCL -$EGh%1P#V5>~JtuMQz1;t;%l}9C|9?aeC^`QA>EY4?N{&C(Km2p_fX?OqC*@WCsaq(sAJZQq^G@(5LM -iievi~0!8O!a;|7m$0j9Q_Yn^wRwLFTHXvR_Bvm+1 -kLZio3DF<12cibCFJd6#V8jr_k%*y)I>boC1jP9Y{i%pqh&IFm#3ICE#8SjPh~K)AeJLmAXXw)Bi16;AvPd3A~qqCFVS8^UqpYzKtvs4?3b?klaQ -Mc3lNJCOA*Tvs}buFFC#XI{10P2BgP`e+f8J8j)^3|^T!-RDoMz&lgSn%iPzhxELERRsGK1^AfA;Hh- ->2CFg(L<5fswP0?=c(Ae##3-(*8FnXb3MbBF~npG<}Pr`oefm_41u8_Z-zijBnNEG82T{E-UM*!7ltn -$E($z%iZ9oDupgV=>a_M$o|?o0ngvyWH+_Lk)GRQXX6$`wu_Iu@Nr3O7tGEaIg -Hz@j3w-M4G$x{vBV;ByQh<(;L}w2Ki|ooRQwwZSNiGrr;$VQ@ekS){EdNJvf+O#k((9dNykHGvGH(vE -Cl}qL#bntFJU8f5nw+*GTErxG&YXzkSDNlbUR3uhv;`3zvcadn3rrqF6LdxY0kx5ks7&}@AOA5<{vf4 -#avz>axu3Pf?Uk!gd!K~cy-9d+*KrUG1eZ7ydCmH)7&;i*;)S$U7o0L@uw -ZL@w5m6(jG0yac&F@>1kt-CG%Qd7U3}v5u|+xmY(>iCnC^tVZ4oc`b6Wj{)yU;_n#jf41b^h>#!7=+tV;|;F4jGUARml86nQXm9rB^bBax3r9*g{L0_0)H3z0`4FG4;Mc`@?I$V-qXyv@;c;mkT)Q|7kMLc(X%!oe+YS12mbp3Npt{p*k2&`Mec@NgWMf?5w?<>lWyq_pP^8TXy$On -k>pTqt^lplGJC_nN+qWs7Qi}E86MxKa#DDn*C!;$A9ABnsW`6%SY$VVeDMSeH(a^#`NE0N!WycYQw

KpkKQQyd;M13QlD(V~gG*REkqeXoqj -}i5ae1@oRp71T^^JVCsBh$RM13Q_SJXFhgJ|!_Ge!B4XNmG7UoOgzT*4cWBINGK&3Hb0A}>L? -7xFUX-pDJEw?$r!+!uKr@{Y(Gk#`a|Ko_t-K<t}a$dizJAvYuUMqYrtE%G -Ad-H?|Edytoj{P6~_LgbIUTI7$sPUMfgQRFXf5HDi>$bFG_L#{zS7I`Rg32zu9k$WLeMDC3|L+D4IC- -fsP6#9{u2>p0NStj%&uMql?R||c}>x4e!O+ufBb9DpiL++2<8+jn|vB*P_OL&7BDbgcP6zP#?i1f(wM -0)(?MxjWLyjY}1UMk|_4Q9EBkGxXEM_w!9;{afTh>yHU#7C~agz-mW`H_1e4@B;bJXFNT8`MZ4N1iC; -hjL%0=2uQ7n{5gLi>ZQ-W^e~}b6LK?hGsd@IxCQlPU_MrSMz) -egh>IE8a74m}@$863l8WUwV0=*?o=+C0&qAJtb}q*9W#CwW9z4d=aeK1S4n3sc<N5=A$DVuy+47bB-PL5V0A&s}CFkCo+SmOfn0-@?XY;^g=imVca_e+IKRQjSmCANAifnJ)q!ChMo -|k;;WXY&gCXDRT?sLZ6-SSXm!!ue5v4sntWbZ+3K=sa{hMahgrAy!N&4+w -4Y>FZiiiF^ta&YtlSQ6Q^;+2emKStO^R}6DY(UjTNV1US^DWRPhsVp0xe3|ljTy6ydQCB;Qa}0aolsf -U!fX#3tNPq2}e)!)pL$|%x3izBl}s8_Rd7k`;!RyeBtA(4*LvPZ;8lxe-bX&3-2Es>3RR)keir3=l;R -Q&Ob*x;iZdJ)GI%)9sQW|c@V*Tl7;aMF6TdQXW_EHo!gCbI^nlTIAX>9GD&WK{CXX!=y$_pySbmjd@&wrmCJc+KKfhAn -;^FzgQ6eh`An3{@2uCgUb$XZ`)A1gI@deZiCfuqV3Mqtw`9>Wc>1|c<>mG4dY(o&*~NPv*YxpnJL2iX -74;M=>*w*O%KCWx1gG|IW$i!CC4RJ%J=|||T;flX>x0Li<|H?<{3BiBN66QC9zRS`-?QZM=&>GL)5j| -0GZpn1?o>W2^UrJ-9_iE$tgMwpDfGL}E79{R0sYVK7sWWS@Jp;>TwmA3`6A?E-boy>D6d7Hg#09OGx9 -f)7a%V|UW9xt@)G2Sk(VL=3V8+c8syc;Uq@bte4nrv`9WduHO@~V^+hh~U4y&|c?j~4gg=m1A`ex}o5 -Z60dE^<$#Y|q4VqT{JGDp_oTh|H=6lkstDPB0h4_PD7Eug*+1Zc9EZA9w!mytC5R&NqK)2l#6*8F;6M)w}SGQkryKW -0(mj=9mq?OA3|P^{3+y>$ln$IME;5JC-Oa_JjlNm{`!maYT+;B;<_G)yh4-*`BCIz-c`)YMxy+P@E7u -Lg@2GA7ydy0x$p<_y}}>J_X~d@7n|9YBi}0gfqbX%2l6w*AIP5;{y<(Q{BfQ0_k=%?e<}Qd{H*W?@^3 -`>k)IajM}9(-ANi*uKWtxO9yJf;VqQ(BmidI&KOF6PNgQ7*0*eV%&Wyd33^AQ$Tj#5&SSl)oVKBNrR?G$6MlZ$d7fAH+Hou}($Z%=rf4FXZCBH4yo;$i=!9v92W)lDPgib#|{DEy;X2a||$aUUtxy@+)&87LR`y?MxsMR|}vA@n2PjJyQs|Vx0sIqmx>cKZ?8o`D@6VkS|8ACj9<(6LPUmN384dNBKPDV%>~bXA_9>ROF$^KSUmh`~dPqzLFMmv0pQK`!>=5$kA -12>+mbiO{cDhZBnORmjD<8L`eM66FQR#X1_X?k5rD%SCw<>#QQcfxj%BT?k^DeIp -m?pi;zbme+hXa^4-Fp$hQf9A{XnT3Xz+U^KscF^7WEG$Li(#C(e`Q`yoC40(OZDH^|oq&W-Z(Cg%&~@ -^YRj_rIKH$@L+4z6;M**q0+;KR92ckXvuzuKH|p`{F#`sk}>A`RV^DPU9`k;aN`OJAC{;S=P_zUy@~i -@bO}!oF5;LN@im>&U*MfjX}O|U2~>{LF^4GL~i@O&3;m9 -M*;+ZFP~PWo)jeh0U^l!wnl8Rh)=ykVMbFQ4y8ll_y1{fxndJKB%n87@4{Nx#LV{ib2RA^gqfK^*q+c -^b!g$>%K{>)!eNle64{=dCl(kn`vBoEdWde4b4&w_iRFXq4|m_&lglwujGiILgPlQEm@>9r{AK-{$jY ->9W80d{?G?eq<{89Wn1CLioD%Oxb=uKb9%?TUodxgiYp)85iw?&-*yeM?OE8E%!%!9?OwGA1%&y;ueL -RpKq=_TW){ZxSN2mhtCU}4|k2fSmB?=3V-CX_U+*LE_Vhtg}-gsA2@ghyS`+|-1T~5! -}|)0%xy09$>;s;a(uo{z%17fZyokq?mL{roy*JT^%u+b@%eX$J$$}^_9s>ax4CeWOZ*Jk9?rAn_QknX -?jM$*KXPSmcgbHq>U|RUShjCU@co}<;yASK`yXwSt2Yha%*L7YzwO)p(bo9~t_)1!_xY)NZ;bRfSo3( -=Q`b6=&BC#B-xEcK^@GEck4uyKZ0y>9zzq7#s#87Vxqs#CzCB+1^ON-KO6#fYoj+LU4KCULW$J(}DWC -gq{kh)*N4_xsd?qtM`~3?G_>FeG>E?ki7PfsqZFs;vL)=u8?(Nr9vuO6zj}M&mgpAe>DOE@8-nO(`?= -`wFSN-bq<-tb=`24W+!1{hWqv$|@uG_aK6E;_Uq@R0g`}X;twVmkkTY8uGkHo$4&R+wcT${9>4>tC0y -Xw8JTi3iX`}NIJ+Kk>k?Y_DNrwY8c-E7pISW{YE&@J;W=o(T(qla(!Y|g`9g-@;5_j&!qc;9g!Sbn{* -&enNuMY~sfUd}oZGpo>r%gZaPL>BB`0HwQpSak-_V4 -`qOUF4KJB6QGcHcc;KX>QcrXHbBk*_x0?*rcJ(DTXk1G6;&>&L7f?e)CcX#U_s!( -z=32m3$&LPL6IlUL^#2K_R8QOB8)AAY&z!lbx|=+m(8dwJ{sP*vEvKSAzmEVF<5&Fh`^oj7;K(%uj~E -ikuB-S62SuZx&@_|P*0W^F(B^?>h6H@*5M8wJu$t-gGzR?E{~y60gk6II*SQ;$=@Rih&lC`qy_4-}_7ZopaWA*jqr-_72n>uKDEnvg`@3^qB -t67pKiHoE#t4<(-(%s}~h_tUq2fMmOW!gzeXIa;NwI?hS9YP0X-a+r4TM+rIj*{`%REKip&C{;QAM7d -K2Ce0WMp*oz+OL)td)mNZlzzBsGm&$GKHNI%054T``4DY>wiAm^_LIQMidn8xm!E9{G)ZJcigE?y?QKWVe_av7S2xZwC@|;=VOENGM*o -vdE%2detX}TN{7emuP!L~>y0C`ek|2YeZFbVv)jFI?vfsOb?{WeJ}p&*#_hB{#r}@p2xNo&Fgc|h)s6SCDYbjni8CQ&9 -w2N&Zo=4rl;EaHZJ(3L)L+Q;j6D5dSRW9^^K*?9*+h`lo{=Vf(Jh`=T+aI1FDH`$!8am%NrIJ*7{y3{ --W}eSu1^Ky*;=3*oS{54qH*~cITvMw|fpvzjFuqB;Z^{bpA^PhsGWWe)CdOpX~cLt_*#k_Rjn-EsHAZAK>0JB8{_*L#T~pr -q?uP~5H_jXHX&5o8#G~}_eZzi_Xdlq~n}~aBH~iLp`<6Q=WYGD#)SJsb9O`~(TG@}kojn=z(pu{`UiF -)PnvvdabLZE;-n;krJ2r$`*L@uIL)Ei`7CbW{wEgdItNbAH#c(z+;rs2?IIl_foekf!YskpAUe0;>iwUMBfqVYQ-9N3mvL?N*sdQKR*h4$tnq8UY^TG?yk -o}SK#;^Kt>ebOF;-)uzv*?rFJ+65zpEPx0?Y84fqtB{0z4E-B6ppLsV|u!$j`?13|Im`^-**4#$N$=% -*J<|4qhYX7Lxx?+HvEyzAzI(keH!~NY8JIy|VMvXM=Z+550T(r`GE9?C;x@K0mYR7q^33&c>CDxy$$X>() -mnzI(#&u$%wm=I6(piff45eIe$ruAW0*o&Mq3^{?NKE*ihM*K{MR+@+IKXrkG7V!@1LaWD2?+jTvXLN_?5eVK6G-`o?-8WhecoeI -?H2qWx3>WZt0KD#vRljNZwX^^6}H(k0{q?G_L<(V^zn6U+sS~m^B#Rjt!%ieEQw~ -?hJIJ`xTmq}5mmRFSN3XmZQpkw=O=ZK5Ak(D+dkGUKatxsc>M9Ri;9O{{K_k3rsdl^OW&A$Z_w9Yo>B -$f{C4t}PxkFNU}w!;@7(#G`@um~ZQl9(hc~C3n&$Pw=c-emteN@Uu!7kK)|4g=9U0!seRZb`AKmPDz1 -gFEeoFd{(O;!4jy=0CnN9}l%hJ1fOEnc|J*vlA{l1wnWb@1CUf8WacPb_5efR$Bpa06N-_Y;;$NTI)o -q7-M8IZKFWBHb`H7l*lt|Xl~{`sea&KTR={djrbAKN^g_Jb+v&YW+aeA>RkcK_&&rcZ{wQ#AUM@W9G7 -`Mdg_W)m~1Vf!~0e)Zhoh!fo|7Hk^}8E=Hk!Y#`}y$a9>=#n`9T&*J^$9Jm{&~ -d(wAPGwANOcu;%)lPidwKEx=V!Q)n%^a)zvI{`4>vw{w0$4o%Jzk6l^Gl>*#MD72^2yXcM|6lc;cKLFzqqhc -l!*Q2yzEt{4qUa^^BqU|+?A>MB=JRbHI!BAE6RxMx0Qe9-iRJbm*g3|DUdjL -GkfJH+^*Uq5I3kF`~(=SznE`wmmA==*Xtlh02-!g$!g8ph8xFw{lAdX$B))*oZxJG*_w@Pi`^N#$c-G -x_m*jx*l$J;Q<#8*3R~Kjj3I-}NiQg3FsuGQEFH{D$%AKQOHLW9=yxeslP@EZxFm40YkQ(+t=8oni44 --(Xl0*RPI+Yb^}xEFUr~_|@wy({~_&VaazdGSn?-;Nc;|&N2P>=QFIkRKZYl-19t39~i^1qV8#iNvab -J>(=(HM?WysHEm@`YA-U>1P5JU@@tt4b-JAlOH@BGtP3CVotTc4k8gg>Us@0_T3eZ}D&B9+(7yO-;pT -at-miUbXjQ^sRiZW|-fzs31IgN_ms*WG=j*kX9^L)s!hX}WTb3-@_x`GR+EcI0-Eh~)9Br<2H1dY0N& -EZN=cb$)o~do>@ypfvfmzzQ-Oof;KIW?p30u5p_SI-@@rbvt4=qX2F80bSeDym$>dDl`&3bv&+Buop> --*kVRC$kHoBE~XdGsrTHoW(d-S1z_)~0reDPCjF){Y+f*`-OwS=y8H8an=dPNV(3?|?mZ?~K!aHaYNu -`;Hv#hv5fzb$b4OZU03vZVmTkYLCv{eRj_NY;E$UuNQiGjMaV{{qCP#de7HRnRl&U)q*r_Y~+(s>W%k -ma~Dj0bINzw+MbX7@j&<2?$ahcTlQGSLajFS=d$nZBuD#v@n0+aJV$72KD=@u;)f*dBlF+7+~$o1+Kd -HnjsNaSjP{j9-TsPrcCvQEoEY7i9$DJ^lTUhYTd&hzK7V6Fbg)S~a_#GvzPg&M{cwJj=}>yA_TJ0mt9 -53RcJ%2L;U9gNuAPwgL$^0qCTl}~vvlpb@?Pz_y#XKk&PdTd{X>_dXE$YQ!ws;_9deLo7uQOd0bo6Si~vh|siT5YM|9 -is_cW8`oCzz7aG)Im!T@NmR4SS+`RHw((C2jA0OeKl-{5DgneHxRz3tVXe8uZE53^Zj-;WB}<0DXoBG -fbsAANXaqTEIC=44818e%V4>soQ`@n#L^lGN;E7LU%HC?0+(pySkGYN} -j`J_CvA;rcu^A-Ynv$Mep4YU$!0r+*m5NKBe=r?PIAz2TBSw@p#jHX|fwN*X28Acm_9fd}pW`hz97;3 -j#h9(_a&EN?*c -_6DXyv(YdL%ey!}aGHZ5gyJS#5fYQ}92_)~Oy@c`;54{uAZn=e@J8Tl}4BU`@d}j>c}Krp-#2930wm{ -S3Hk0DU_ND+O&uiX3i@-+KU*F)J%4mnx>ITQcj`bglcOmG);zx6=x@Y61<5p+_^OX$;wBTmHbm&o3F8 -8a5_9-Gf(8-`0bTnK>ofke$P~j3RyT?CT?+hng0`_z}?kC)U~0M;f{gtiWGI3REf=866j-YhF%)?36- -A*NSb0phl22s$Aa-8Y+?ho`S|a$=fCJEjcK`D|MZB9pBP`UX%VV7(=Kh&ffv3m -+qnwW@T@?39RvXu9}L4bP23Y=%5Knyv0dNTgNbyl*xe(vDpxvAsHYSRUM+N~8&z$`E5J+yV4u%SE8>* -t+5&At5PgbYS~s+B_%Fl_cDE@&S42WURie)&zFVqcoj4rbOb~4c)pbbRBRm?)*)NM=pLwI{RD=wWljDN=2X{g=?; -BQ_NQ&+FuvSQ2YJGS?+~t@zPp!%b_bq}SAVmLH245FV}t?~iKTiiR!d6X5VnLfc7=*m&XmKPU!Cxej- -I3>=f)udk`I|uV?F&)n^~BC&5JL8wg=%mmia;JsBj{cRuA0 -SFh*`0d&DBdyfcsH*XX_-zb*Br^1PQgKoOEAF??=l9;J(mX`JriFZRE50}$Ch)dDODQ^$o~)b&L{+b%FeLgsZMva(+_ptgV+%<5^*5Wg@XNVpSL}<>X5qfu}b%F&w3X^*9V38pFLD~s12>lmA -o!Xp)LcUtUk>DjdC3516m9egfhJcy(AMjLy!P^Pk3guFk?QKmuA4*i*bxjo(2CuR`nW&U~ih+JPZ<-}q_l8PPvc%A6{JF#=62HS2>{Z# -Pn1ZERY8wY%f6YoK{$-?u06v#rT>~YQ`AYUzu>sk-duR;C6rx5Jdl+tzeu0Cp_3t2&)isb_z`ksBX(n -W`62lxa&J1|ZMg^tFZgk0b0(4mxv?xOD{cF5ZiU%4D2PUIf&&mR7rVq5GBClj;(*@tU9k1!tv$~x!oh -wB@Bs!?Z5EAbKz!S(WxBc2~EpCa!D`eUIZj&o-ejq|Xp4?<4C^$tFbsArg>jOuT}9{BjK=kl>s$7)v{ -$3XtbX+lO~9DkONt1Ua~2&t`eE&iWqSSFeHdfn8r7J%lxQ{``271a(#UG1c{u9Jf -z>1`9vpMss-*(iN~g9p_Ohgz;2{d-sRp~$0lutYAyCvSc0=g0M-a6cRZIxx;g@MS;Vce$KzwEvtq9Lf%#a*X@P2uL@=!8kn)bBvJ;SYS~w<`(dQix-6!Zx{hw)er~~S@)!+mqiqY<#li@%6Fzl -vA3*P)_%-PTsLLX!Ysb4#-#Zlby%54CvoJ?{hx6ap*Qp(*yT@wizH+VE_7~B0@Fn~=T69InFYx&TJ`! -(2=RKrY4gGygYw1J5=8kyY4uUpC*VfbTRNesck4dbi>^A8wBt-2TKL+B{xvI0LyqwO?^YZz5Cv6i>ZB -_^2kHAOw3`csOszx8S!Fx&xZk{3Bw++3ImM+sR=UnU -abvsge`%v3V2>EcM2#LI}@L(z|jBN26oaftULS`Z&Y+=N($_%UJ);%UV15&c5Y4~T;i#~@BboQ;@)n1{FuaRc -Hu#CH*^5Whw|hxk*7qunz5(a#$YcLcl2qmhq6?2p(5aXw;}B3-p=ioq672Vw2zC{tR_9CFD$-e8N!F< -Eo63=uX~+xzD7tCN0WpK^vXuiPhUO=8#~jgDZ53O~?sp3eA%Zl|~y)h -B>5$+m*^A)89{qS*^x&lXyp1jw1_Vqj9I_q}sFCyVB@O#^A~3cN&^t$Tst=NrT&TdzQ@@o^LZG|)8T_tZdk@Bli6;IVkKCNxy~_~Qgd=A<;@{`$w -c}V_MChkMBY%v=2)FPxsukZOlQSUiWB*bC$v(a!joBLleLV;8M1R0vnq@?8&c+w8B)B>0{@~QMto%So -Uqs^mV6>_Z?OgoYeJ?JLqv|)^jg}R?j75z~CB2+N6WE*BA&^{REiB(^Gi1{#(l9IY7hwn#`qOtc1PnfL@*>ZH6BgPAs7NJVqbbF@KTw6mMIs!1MOoNP4CpMHc5_m^}Q%LzmIR -TQ23j}ZS{8(=U3BZDMW(@{Gs)=q;J! -bC6NcNpMi`?IO>BCH^@n)lmwHorGM}k}rCG@*o4=+(B*TZJYnCcIldjw|ri>~iugploSPH#d4ew_Dbl -+m*iLUN_a&@fn^L6))alZ_TDD?ih@hK0o#@>tLu@~B&s6)Z8Pvhx^DDYA$9CM+VBkY}m(8Iz{d?+2L2 -Vrl|p&!OQKG+Zkz(D(kF*u_#M{qP{=4PMBQ~n#_~rxLAB_ib-<-<7nLlpjC<*8KebKInzf16+K73m}x -Lnif@!xLR|0_r9FXrU()?Xa?@3s5C`}n`TT_}$Kk2e!Z_H|x$Zhe --Gc=TOCkRLmruT|X+{6?n;dnBd;eklFvPbZ{^2e4Te^`8TN<{^Yc;;^!I#*#nI7;tdH+cbS#WaU$y=x -AmmjS?U3ZO;VGr_@4cKqDz4yNcAHw4@>h-<%{|txV=O$iXZ{qdXPB-?M=Fx`y@9}@<0lnZqymNlngMly2LY1+H5~|f0uEJ52Eu@UN63`rxAH<0NwlUyhJn{%K&TXx*zCp8Ud4f) -4j~nR1JVLRR*`vzJ#=s3~rJAX!<%*3V4B>RTcH8TrV~Cr)!FnrPu*f&GAug9!;% -8C05~;3wd<&?Stc;lt^kQUydkf?^%fjNm*JFp-76Hz3D`)>F^=xFQ7;uE(m9 -HkR7UsGFgglVM;gND-feZnH=K~`5WhN{+t+}ofg8mFs9zmmXQ^IFoJjY6(@WJ8xn2H~sQ>DSc@o`o!J -w)GY=HPllOaDTS*n}@{k3$-tuTs^Y6vfzM)%%KQx!yWdl~>^dqKX_X}v(b#Xu7QPS=?H2>L>1(!E6MN -CjYTsX^5gM}Om9FO|d-QV8*q66jt#dZ{9T?!lKVCDQdERNpLW=UFLnHrz7LP(!Fu@Qu#8v=V* -$m;6a*CnyR7z>Qi-5rGA*^pDa~8O!x3SFBLpO$ZzyKdJOyw`gM=fy>e4kO^?&`2318N$HJ9RpWxr>RU -G}FAY>^m{}XgC-#RJwNm?K0CEXfY@6f&gOR0P<-8=NWlu^XnVF_RzgqJ_X&MTGg(}Xku&wHBgNm)lK0 -mI;YR6j#Vpa#wdy73)g9$c5a;5?%n>D~wS7d}tV-;1i+7wBFXdMWZnLY|ZARdt)`UZC|-MG0-^$&&6R -TF>BLz;U3bZVMsNfPq{2d0l{5x0SY|WGU|zx|d$Eq<)p^HK_7lrTNuMwSXm%&i^$+@&I+EaK2LhIwA7 -`3ty*usGXBaxAAt;u#N8VSjRSwtpIt+4!Y(dNvhsK_3I_`PF~(>z|g?uvh@OIq*7z -sRbFL1YuZk6SPTxxUC&A)=5H+dCwof*^{VkKZK#IM^=+fRzB>|;W%!1-Hp5b{@Da;x~1z8cTyR#FLmf -_SxupuM?YaxeUnkQ*euP2pi)uIg&=C&=ldhDp@kBeWfzmwan@zUCTQZ}pPzQGOnk0``M`sroD0FF?J1 -P5U#wRQxq)64M=uot4u*LCX9@9u{!eob`~y7W0zJ>ps{+5H^_3(g0xpH{GQd --SAs6BL23P?2K49ZTx~FiRs=R^5hkm?)o=-`t+Dp7#k>A7h7UG5e0MGn1{tvW2JTKK==Ito(M}9p>`j -PJC2G_rzc)lTsNq~0gysGYJ=*OuYS9rUjljjs;0T;O$-79{h{l^7W@$YcGhwE~66YYo6+zS7M`jyh%O -8!Xgo&mOKj=T(h>>u5hNN#p}4hfXDtL>k{+A}hg9jUEz71zzMS5i?+= -mqZ=`O!SdRG39qiq!O^Py+rl^Hu*{s5}eoPrSg -6>9bD)7OQZne7Y>AQ(C0r$V%Oo4!4la7;l9BV>wgLXUxM}*I7EVeppIk!?uT|AI8=i61=ru9R4-h204 -pF}NC@1*^n>;oBC&oWWSB(iA)Ky7q1SK!krMRNX{u^K9mFdfMfKN7fukj|6?oGa(9@4z-{F)1Od1b<0 -j$$XB#Fu=NTd)jFhU~g{&3$lQ6kZR4U;7@4lpT7A}63f3!Mt}<$kG6T@1BTuhPw+ewhZ)KVSO?e$SiDGrej!; -EX_1IKy&tqoaGkyA)@0}P>c3cG?LT8N)d%_JO3+^>NoK&Ev>(lf_5tY{mPqWrB;x_aEV&}E*3JLlLxXwHxkxQVr@ezq#XNw<|$VjT^F|gnLJC*-xiQONCJRy+;2&Zd3?g38MDVM-~5? -zB_1xVKoYv8_yuA6KN_ZL-&bp2`>AYB)<39f&1eI^Z0MP7)w1F;(Ajfixe9t|IfNY`r7@I1s~M7nN?+ -C$gtgn&QkniJcAe>(c@&i0WEfm=NKvwFDID4>HdMTFgs(>+}aT;qGPG1v;WvYovHpn#@t1f=^_sNqzh ->xngh^Z_#zkgm^;m6UoU_6xa;7Q*Sizy&RY(|n6t2&YHcEd!!FprV!V>Q=(*S_yA#CH&S&1|II)O1P$ -#@Q@b5>3ei_D*k(c1}Aywqi{|;*^~5`ij+Eyz^K>4Iq2cw=JhJ}%p(r}cY*slc*W5k%hd)Dpoy)Tz{5s{J -|2oL&0dXGU`Z>?~gT=3^I{YaQ_=D*`L-Xex@^5ZleYlF624U2`>xZwM0roqUGy7LuUsZLz>UXMx@|$t -S21|LuOw1-bXGi~R-*w+89o -sS$jl1!gIoy5n-lO;=*xU})2q9U?w+cvUi&mPvszyA7ba^b=Sxt*Ko&syNwz+Co@C#{JG%nvlbo%RuV -=%b%MO3Tb44?z{Bebn6i&__QkfR>Q?&`-ZDcpJVv^x}ni&CP$MEjx5!)~s0{rA3e24!9t5>Y^FV&3hL -#Ke%$hNAqSaAP+Ua{L$W7RNuZeRR2TG%|HI}M;T$Q_yIy530xQvGr!3H-2`4F0&84 -Ewa1OgPv~Mt$B)#(mjL=6}&lB91kaX(yY>%rniz3ZGf$n#sJ2%_OC+nIvCsCd-#ECo5O3BY-}Wd{P72CEA&teg)4Y0A#I@1sNtgR4ei -DUPWC(C1lBbBk@3yL$h_v6B)@qPDQ;dz_BOvsjx`?<{VC}L>H9$X5J*2B(nmx3xsX1+k?sTEfvke`n; -`uzNM8ZzYao5SQ~Lfx`EP;lh4eN^zXsC34(UIJ^hY3lEu=pU>CZ#@OOXC&NdLQ2`l0>d`W6OleI8tA{ --5@)1um*;>mM=`trV}fE1Bhp4^UDtU-)ErDUc-UjVS8^DGDWs-~&q$A2+eQg!gs}GjrYy%nYEJQktUK -jh>>GQdyE|k!DgpqO{hx{(ENQVM0W&-|u_;?yl{bnRC|O>%G?A>u}Coj^p`?vX|G{^>4Bf?(eKn%^-z -7pQO;SISQRvqYy8W>AVX5k64Q_JSQ*(V;Me%;d!k3REA&3@J}=RdWJ7!_+1RYpW%U4;%3jv|dd{G!^Fs;^%1~(1Qw -q&_Nuf2H6xvv&(7q!ocn!l3WB3US|0u&}GkiY7n;8BThTp;PZ!{!-5|Ys|(Xx%Z!@Fz3z=6&Ydi(U$^coL%9zQxdDm -Xgf&Mvom^Lf6&&vilH1cna|0dQD!!W~_@-0nOtaO~})VI-p($4|KJk6pUlJ4RgKFD~e*0eo0E&W7OpA -G_RnuM4#a{V@Z>Cj^JY+!q}k5uMNt1KjT2wdh_IewU!LUJu(G%j%QPmDikR1C8?CL%gU5Fuue{Ac_V!Ut) -ALNt%jZQQsnUApiEjpq}F-ISmWh~NqC3VM9X00&eu^aG&c0coh{M(s({$M1WWOn>;)^M>2*$F(x{g73tFXkl{lE2e -)tI`G+Vz85YXAJY+~%*roFc385hqM|Eu1b{KyQOTg$+;g{x52*V{4N8NfuJ1x$K1+%)!*IY^du!3oC? -%Z)=LRd`jDAqMaQIGFzkzSog#Dp*(qg1Ccez-WzzYiZAHY%#=X{Y}1_~ERGQ4vE&x}8>y&#Z|G8R>Rf -iEi3r+)fi!E@|kL8(hnYMCJ9#=aw|no6Y5Nm0-7guH`&M=5v!LPZs$|PV)H1+@mOatdf@+p%i9@DVvu -?DCIBaD8s8br(i!9d5Dy_qt2Wmn~%>w|Gctp-8yCc`t^!luUAYalgJfbdF2(655E5T>mpb9xUxuD$N5 -G%=L&oF>`^}a@I#Rs96NeMIePS{a_rbK<>bke%6H#=r~LTCX^|V8J9kd`_19mOH9wKEk-zV|pz^Z;Jr -z+{%3!-Jn8tx_AqTo=InY^@&a_z>L_3s8RHe+J1IijYtZby?9KY*T@Ej0PsLSxdA7m(^ShZKn@X-uEm -Eq?w{7Qyj&+wZV{vC$D6zBevr~D^R`M;W{$oKMe>())&(sH|akx)19J^=v%cx-bwO}B2|-Fo)#;nlI@ -U94xf`}p&b+0(N_`?hUw>DQ@Kw;ONj*8k3%Jv-os0Iwdmwex7_6M%MS`?eVWPu@ -3l;cqwh>d?NuKigN&K6kX~&Sx3qPX5mG`*;7NZ!iCV0Dt^l9M9k1-@muNe@}lwUw3Y}*Q#ev=N&vtIw -{I%Y~rFS=%Km>*W(3gJ)HYPc#d15xJ1*UpKV3MP8b&7H>rAC<WEh|;n5^`T#@AX^=fTgew< -!WAmOmb!zZ1Zv6RIfDm>=6vkMoNh3P2V*wEp2c^L&GnM^A)?T#s7BZ_g=O^Jd395@Z -rPz0zbw};Q#Ks@6v0py+({5Ra8_6ypJD0PG5ihHC0zv)4qNC#C3q-909){W0(yaH*Q?Kb?er}oU%nMU -%q@|Mn*<>etv!|=heUsue9~=+O_N8RjXD#bk9BafOp>E;o*K5_iFJ!bLI?jF0-aruU<>Tu(7<|kfyb=-<(FUR#~*(@^YOKOk@ZCk`DP`_<*);+eR#N;+$F7F!&F7yz|aGq67D;sw!bCH -8nNjGiZPhfWE;_AS2+m<$I#!PlI53sFtzar);{CDiwv8H$L --a|Qu&u4!Lc?cSSFL(~Vf*!~MxZ^Wq|HT(yh>p+TIrt5Hp$DJ^c!U3tvEe(S@PkC!{Y0bo5)FT!XmAz -Nu!pEW)6n+~BA?fZ@(&YTym;}YE2Dl1fA;l_9r(*GWg+{u*a_?a`traRMB{6S!aipHg9ettu=j}WVH( -m85uLwCG-Dr8;Eu-9M62-s>8GEd%c}>nJLIHcOf}Jn51cdvvMt=bo5*kH)#J~;cuk)^eTK3e@?mH2EzQvZUB8@uf&Vy< -xBxi#8Fh<0#x@EXJ|~J|8X}m6@>A5B_5=49FNgonKmSY|XPR`-AiJcaIhiy^$At?QXy0q=$-F3D&WCe;$ffQC<8_RKmnTH5nFPJ4z=>bt{f&jYqQZJIsq`!5A_y4eOZ(v?TVQ4>Jw%j -K`mkI%$yh3>u_8!zT@ToB94GQ9sbY{MWq3_FpB2`LH?s+20HOXIslhJ_MVSbVyo21LDEubPM>yPrkY| -gi4g3Onkexk(@;nBz;UNN!zV?5cFCSWgS2PRAnh4G$#0t&<|lXjVP}XhoM$z5&>*` -z9nIN@qy=s3k`c6N(MT$NViXl;GY#{{(7E~(bm$i+4GB!c)Nhyul|4th>^bCs()7zbf^#KC@X+J{34Ff8PI^7N791d5Nl;#hI5Fw$QjUH&mBQqaz@Zbroq89Seb_HYiHBh)8 -Dqr*?h}%rm;+dq&>TG4Bs8i;}3r!^?&5Zk$oWx*zM$2c0(UR%m^vt}6=!-+uf(CVtG0>G`Tn>Ns_d@>}pM2mAyl@0g -z#H}eyroRw!P6mG&~kcD^r+;8R3*OJwnic2A?! -im17Kjt9kv0&p-@A9)WJPTFGX!5%MD0)iPuBB&Zh0rrh!&6MdGNxfM(}3olWBlbn^Zp0;vn~kzXT4qlTAq69DYDz`f}YmUA!z|@-M&|7_l -n_E$uw+Z8iYMF4e&|Qo^|Y#Udjxn7c#W;52hiXX?U7x_$%wt(&TV~D11_gD#v)`Wb^(X_)GmCHEL8}- -~jyc^70yKaHpdsT7LcYB7M2*MS5#>80}mhMB7~Uyh&xxCKnCsTr~Wh^(g=G5L%V$qybIj7#yD`a9{t8 -h`;YCzi+og<{W?5Oqw()Iz2u8UEm6Pgbn26PjBMmLl(UMHi;FC7zw8Ea5hGzWC<=XAwQ$az&F7}<|>+4Gi2?=C2n+b -6_D=UlU&YerESFaX!{rTsg3%`S0s5x4IH(~>P6>=fuqgW51!JhY8=F!Dpc)ekq-)PS=$5`vK=K{8cE8 -)L=`}R(p>-<^bKVZOsVKZjTNS{7^`bqYqLJkIlfhJCzNK>Xvp}+m@Z}j~0&(m|yJty$N`X2a#2E+oy2 -}uXmFrWec6ncnQh&lE=6Hfb!fd6@V*EIi9jxpbvW5Ax_lU~e@0s2>5e~lhJdR0tJ%u?tN>-rjA_YZyb -)mQCo!^Gz(J3E`$CJ5i}z4u;`BfRz2TS6azJ8*@Jq+UsRfIjFR);Y-QAPd+7Y+Bee%Rz0=#cuZe{?-l -11J|`ic}9#F@!Zuy6TwEOW>eY+-_3KBRTMJx(2W&&#L4)k>w4g($Fnl)=^*|KHl9(w2@nl)>d;5)A?g)V^x*a -heWZlJ@R7VJSF4nyw|qhTx9kH9(-zadA0>_AI1jmiJ{3-IV__kRG)wyVp?$T;%klTZEzJRf-A0nuU8l -7{~M`_qC23xpgHGvwZmqyzW@f4S$2wE>Q>2iOSg!QI{wv(;-4te=&u+<${@LH}8o^h=j6{Rnm%9UV=J -7A+ET01d!DG&GdrlXrlrhE -0UVZ(fwhbv&uZ@lq_paJ~Bz9HL(pdl$Ki6%^#Aau)QG70>_Yw#SOfj8^`{s+G!H--Kn*M=Ry24E8y7e -1JA|Bc64;qC3MWw{&n?AcSNIBP!bT8lsCRp&pFujp7W0w0zwac+UTiiJI}yovQq=m26Acr0P$ej@xRc -nbf&Y11YuEiDx>u&k_1z=7Ayr#kj+k8q4!#JKO~-&QKlx<_m4HzI#S{saHd_Lz_Fus0lm3*-%(Wjz(S -L|R&!@U3j?BEG@bK)`kYr}>O*kjb)aKjB(i+dyfGXH0OuK*3wFE&R9_zfS#v0J -PYpi_vo&Ftw5{2?2eL!hn$y@xJ==dhtd+{}OVd~brkTL0lA++(aeEjVJ2{o#inrWam#LFl)d@3P}M=n -`}t-w^|VyR46JUc3I^z+K?4UjK6J$%kGd27wN_-w*!*Ux#?4rUSg)vSkas_~MH~pQR0>gKoJukGu)rp -)<06z&^96O`A6V4(^TktK*_Nhf(J?Ey)C3r_%{|(1q(k8}Nm0!FPfdi96?~mTT1a-{B8j*uI4RbNpT* -*IzBsBKJIH{SLChZ;*qm$w6lDMW`j91_fGxJI`?~ex^nI)%!nkE!C1tkl!Q!!FSZ<(4kv6{_9`=5`Gs -vfW1QBfj{hk_0e=4aToZj_dnTBH17X^2Ec5PVg9OPVgRbko&uk1HQxO!`H%(@tQ)^HJI -k>jk~}fy4#YTUryiEz6j^#z6JOU8lmf~N5657P{*=DT?oa&XzXV8 -KPD9sL%(U_q^-1lK!{{OLL$+qjM%uA_3Djvj8~tLDvh9V@Q~{!3g(avnWg$2P8GJJ+$3>)69}40autx -Q^sJdf?df?{+lGH8kAc)6Uh~DlO!S*ze>14)^nJ>qUG=Ma6wT@Nx}vd2{X?$T2@UIXQU}=lkL8^KlzC -ZaYKHjmOtF>M6$lyn$%`QKHC$&Keu`f>Ber=S`xAKO{Q%qrjyl96)RLwr$%E=e6A&w&@1=GtN2BvA_M -6{nB^rmpj%o9bi_#1vk?CUKq{r@jvZ@<$E~To8Y@txXeqVAv1Fp1hnN{Vyk+bx+jWgg(|dYi+3UKp#< -~Lah^Z!Ux|IpjC2y#BpBlr!QW-_%U_8TjsMkBDXHh2@`vX*w4j2qI!PRUQs`Oz)c^ut{M$$eW)=(AN} -8O`Ug2b@*j1r)V=-=T+qs0^$FvA=YH(j(QGn7Z}~9I$T!UZ7!r9knKht -KNq?6ZS))kEk)Bwz1q*Us^QH*`HsQDsT|`SbeGX?LNPIE!e#V4EgU`87%5us1;#P6MN%QAF(fv`U+|j -qQ>Q#fAREjVy^KjeH`kruNaO-KzB;w)MmlPo%T|e@xZ~ur9#b4m^SH0WZ$heM0a4BUaQ2us>by+Fv)RYyx$rXIy -o%B`zG$gg)}T4|6~KQ&anw>oeFPa(;E4MP0iCpRQChI`x0L$0gkeS^RQynsK#IW*L7OP{!86V2(PO4pk8|H!cyEn1Yma^=c2@Zs5KpA~)&eiyM1{vWyke -n1y7x4P!^{n3x;)X{@%6K?uAE2Obj*3$flDcI-0o-TNd^%dJ<<6gRZElkP=vUqDvEY%k$jQl>{ph2Qz8x4CNRL1MxY!5I&(9Zo7jo?bp5qAq%X=Ha5BH -dihB5iN{j6Ip=6?eQ47i`|UaSKV>rt0MZHem;qeSS)D{>jFbS`WF$k1>$jf#2B6g$`mb9d!-Zhq{gc -8Ne?94!D39@Z0d2z;ox$ozE(*t5FF1U%h&Dl3uTmlWT6s892(3&RvANP_Q?_s -I;BoL1>Gc$#6z`d$+UlH;GuR%9_@__>fELS=`7uMz&gIXq+zym8}3TEBk1*eix_#GaF^DPH5;Vt -&jc<1cgx^vWC@wkPY7zyUgj8VB~Dwr<@j?-RV%xf|(k%^oQH416EtTwY!d{X50HSfx1kPp-A$Z%w?ev -BSQu$Wz8J$eC^Wpc)?av|YP)VW0H^k5S9|cNTVtwK4R -yl-D3rK6Bm&70mhdx}##Ktr8yBLZ^|#$$UoI6nt62x6U;a^3(mC)(m;ONq(##24gPRnp_j9bHSuj&Nc -AXAAc8y&vD-Z-vm1XuJZ3Ys<)eH&1`3GIL)GSF9DVRte;_(pts~wN|Xjfjjgc^A_hsi5%w)uf -f)xb&i3xIM!ZR`(fROb!QN-JCVa6rvl#aTWDCrAYXjtpL4_*>UQ#ib595R2w1a$e~=Smkh}+V+fxxD- -$K5JTo1V>a$T%9>@TenV_?k@$#_714?g&ysB^=Ife*EN^t2^+j95=1cSqiibpzH;i<8HTIi?&TvNB)R -ty?GZLEIyPx*B8&eR0plt9O-)ynpeu#+)0(z6$H96crAY&=o#+mGXz -~!Pi4iA^9J*W#h&DpX;;lrFUO10{p8Ie#ed-B3G053)G506Jj>@lK_i#He&alJ$stKqUE%+hj?*U+PJ#XLY -LRHbM;R6-9u@oXf$5pX-{|pn+M)pAim-GYd0u(9|4{mfTypuQSkNxe8aQXZd3vTo!$vgUTf>RQX`(bc -9ZK$jdJH@jN(!Drkiju(pMktf{QalFgr{@vn3F+@s<;w0o -5Hs`2B*)mf@>2Pp$Z4;05C;L7=Vo)f3>W$s$ -Kz!-qhZ*DN4oK6J_|j{;`2#O}grB6?qH9I|djPJtaEZ^X -Or;SX?N*g$Tkgp~!DLr*Ya@O29^=v$pMl*VD=DaNNl1JBQK35oRu#bRa$gJdP^DGcPq{8s088#&<^2tgK{TO~ -9o|0xmzx=zu0TjSgre4r~jMeB&?f&htO*|8LWA1mi%_VDL0*jDE&IV~|m6j5j74Q;Y@1LZi-DZmcv`8 -*7dAM#bc1(wG8GT2rhk(Uf7zH5Hh2rgBrIsoGR)sy8WSFSEuRXx5rz&57m=bFR6-tTUIJE6vsBT64Ww -v3OZDmOzWv5^G7cWLR=71s0v9+)`<&w$xhcElQDBk)|lHNLv(JlvtEelv`9#q$?^fsw}E5sx7K7;(%$ -@SOcwEYpgZVnqkeg7Fc!Ga%-iv+FEO^w<^V6#hT*4Vr_A3abj^sac*%zv97qhxU#sqxVE^ySSj%;(Ub -(1XiH*C5=$~la!U$IbR~*-VC-~0II*cz&9 -nOPyg@v3B+nhsv#0R<*^EIR<4{KZ7H@4TedCNmS-!l71}Dws>*80>dG3*@DSZ9)+0}yx6V%&q>I$W>r!; -tx;$N>&Z?`>Rq1MUb-D(fr`}udrw`Ic>f`k(`fPokzEE$~SLmzsHTpVzgWl8NZSXS$86pkwh7?1#A0lw!&@<(UdiR#Szk%2 -Z>jGc}kz&E94|bC5aG9B)oBXPfiPg=VX{!dzvpG1r+J%$^o+i=QRP5^0IIq*$^od6q(pm35+u^#R)8$ -(rC-6y(x_Y}SFoB5P4aQB_e*QC(3(ktfU6kL4Q4GEHH5=2;6lWv{SSS!-Cb4OUNJ+jA4?>XrIEst$YUv3Z56gETaB&G)?o9rd)xi -&LH0;{ygkL9ZO^k8+O75qdzHP$UT1Hxdpf)wevTkVq$A#u;>dR7ISL(CM}?!xQRAp{DB_{VqNvfVt(3 -E-RI`@Uvxa!Fb_B9!#IjaoutpTvbhdI^rLEdlYpb^@b}zfe9%$FvW9^Cd412D$-r))DJSbYwVk9R&`Zquf#HsCLvk>K#g{SE;5nuvA+bTbfv!QJPy?P^v2}FRd)CF0C!CFICFC$ -~0wxW!kdXvc$5CvfQ$QGF@4DS!G#uS#4Q;86MK?g#=jVrPJsFby{7lE>V}E%heU|lD%A4sjJr2>gsih --b=612kN!@Sbd^CL!YZJ(ChT&`bvGZzE)qaR}5YTjUmvWHN+Yc4H<@9LxDkOC^u9ZstvVBG!M}b!` -6yP)h>@6aWAK2ms3fSzG?_qApIw0001v0RS5S003}la4%nWWo~3|axY|Qb98KJVlQ_#G%jU$W#qkid= -yo-0NkDKPC6UaEDd1^L}?I-Xqbp5HfRdErCK_I0-8|}G#YV0WkNOJ2%1=lljhnZjylhoaW-dj)Ok9Nk -1QxA1hTM(uoyrg3Su<{61FTrQr|h}R(HbUy!UtZA -{a+jW^&3^y&-6+B_UkIl{_WSzS@hJhw55g5K3Vwa(`osSKJ(18a@u20q!k99Nqg#q%YUBh`nV_Z_hk9_p7ZcL_4v@9z4T1!`8hpr>gk4Ocky*SXDRQ$dwxmrBaHmxjOQ)%y#J -~EMOfdJeV*ktnI3;R))b%o{bRl3xZac)lVCEnK;UB@mvlG$IQUeUO~!kS$z;PHQ}2@^Ve;PDWrnw$;z -*&$EkB%mv;4$L8J?A27#+f~T -QAR(W-=|jx$yBv(-W(&=6Wr&pyli8HnSnqL-s{qg`2( -EIa=}ncm(7N+>$iJ<2+kYI3mBzC7`rffv|Ch1cm``K*$t7%J)#1okewKU1SbluT7>jh6#`5FbzGGQ`$ -Qa8nFa4j!viJ~(I1+6N!ZZjV~uL@FdRq&G%2St+mVTw -Z-?kM`vlK^pI+ljhs>VrOInk}WLK8NY9%($l_4pc(iZe-Cq=?;%oXQ|bHxY52PJj_+RgE^Fm(OYMeMq -B=~SpXl$+(vJn9cVk)>t;-VoWg(rQTC16hPOYSfiIx-K!u6Y{uXGMiP?1e~1QEVbOlU;O!PVrau9$X2 -S0f^^B7Svu>$d&dN*$#%cb`~4^-bklZl4Gylpfcnbj8W2`Rl3+B})g+w`+K9}Q9+HIgP08rpk-8p2j? -hE_zO+-oXu+YUf#qrTD~-{!eyb -sUz4S4*Qe!(Pc)@CYBPa$=m)6(c7oUDo9jm9E$oH%Tn{Z+GGV8|5^iYEmP-Cqv&qhNoD$7HX}Er=9(e -1HfKUE6S#ESXa($Kunmyo7o6+lG&Wus5{+X;gJ2Vg^@_WD;UL+eNdsYMc18a7*-y4<&4+8L&w!33wqJMt^Z? -GOJX-0D3E0|eErFqYbZo1KRkJ2UFb_-mm@WHzhn!Tilyj8w-z|trp|+E`f(MG8fCuFJ9!N!dZs|#Rgiq<#<-veKbmx1JH`-VLt1L -Z=87@(VZMrkaGw?R&wBSx^mq$oS4`cvLs5=+&47};g7QU%+B(}sQAt{h|OlLM_B$G_ay9!B -#8B}t?Uaw03;v^{}%+*iP?}bU<=Lc)MBXGPoP@$x4fo`JItoy?=UKXeY1~t*+D6xx3niaQ6PJv(CtP1RpB$zp=rpu!KaU6EP)-Eaab|_G|7pW+tic*`h%&_)U{oJNXn;Azi8G5BR -OtC;^w*@&Z2WP`SZ{7C9z!~D0(RiN9Wl~xc%uYocn4*?|<4uQ4eLAiT2I -Xoqyj$_2lt-oKO$YDTu!d;v@7I37vx&xd&n2JisyQaVT}qIjpWrcQ)=rRB}|@U^j7I7U0|?l+Umi0Od -+AP`a*{$B1m7`Y$J~=};)!56gyT*mdVC`yyom+mP{1^jQu`tRuYBL|r=%;;C4XL>2E;SyF`i=st^e`B -^>khV#zh6s|}S%KKrH{jo6*G)9~+RBYj8`k%vnU9V2S>`GmXkv$2sJ9*z!i5)8NWRI1H3gw9y-#v)p6 -JQBrAYpXiJ0syJCCK&yPjECQB8%tmi12lX&JwbK)?mjWuBbe#>jlbP*9UGaP~6w^OyRS{+G3Y}gUOMa -l=>L#9^??t;phcp^r1v*GOa*)apd*#(B6t1kPRBaitdAopocgvn)4~Vn?td3O@J7oyg?nuBa^T-q1;j -c1N;^?>}6=Qj$=}eMCUE2jncuT01PWrdinX#yIEa?NBAb8ah;UyVfB;`0vYiOwjdDGyn57gO`#+XgB)R{1AfC7OZKiik7wVpw^Qm}G~^ -&kV`Zdo>@C_7mein0^r1WB)v+zo}_YZllCP`(r~qIQNBL#;op>p5k&BNCjIe$mfrVI-B|Zp2(WQ9`3y -Q6AKtSPbQ1b1kaT-7}+M`z3X@ji-YWLKwh?qhPVPdLqcGSG<^u79E3{D<2^qV&72OxZ)YuKW<4JYt+DZf?^0Z8eg&x?RkNH$QP2ZFop%6|m~!l)9`w@B{f`v#azf&NlWjiaEWM%M! -|XKSUJMo0KHm%$hJqVc&XQc2j7m$h=~-&`XX -d88Gjo~Ed*5-F>e10#m=(AI54b1RXEhBr*W0Du4ZCVc1l)vq$BR^--xMXy`ct%*7kp;ehOc9hv<7D~*}R -K0Bg1|ZQ>HYN^UOAiXAdkE5!O@8FaVz!YyhPxvJRWT)h=BWo}VCvDW9MF*>_*pylVjS?87|xQEQ+Cp# -f#g-VRNapYn;jR3(NGjCTp%$)cm9RSx2T_DVlYkze3D#7HLuxfXc&{1Qut>O_PfspY<{;Yf9hvvaoFP -|{>+HBY4k<=4J^s4+i6Lb!b2@5C6)6P9<>^y{sK{Jy&%O2e0EGITb3RTjIFnsNAgMG3E$m_2xO;PdkG -n(ow{>hBZ4I?!BX7OD`6dSA{roA4^mTv3wjXo!8Q(xP@ra5Op+RpBCM@p^-S`tz!X|SgzQ^Ae-cvQ -2UdkzJ6PjnH`=D5wcTl9m{urqPh>>0M+FL?Wkk8ybjn_bz%x+Y!1ArK1EA7%^|V1)d(U%*lA@I^fm+F -+j)AZ*(48UMGeT^+WrmvR=}>}Nd#gtW_xE4kkRaEc6qqOra4#{WlMeqv&F0N*;Kb3liI02fylC#*$7||9Xj5b!^#QyP{SeuvT!iW^*ipc~!_;rbL#UVbtBJ|#EHU!%AR_vB8B{oP7a5jqwO -6S#bon0mnoMu1UpKV)nBh2s$m;?D|#`s;m<4-ikPbxYZzd!xJG@(3cAjn5FsxBH;O&V~w0}7N26nBB* -Hm3QMjWIsG0sGvDecmE(63RCU<&EhL0JvBs$k*yzBxY3qbukEeB`;?`1 -yz{Gz;K+YcO>sCNiEl&q$1`7>3-)a#FrL2(3xML1szW2sEI^F6mo=r)IWWP-r-YD$9^_CgkykB#VzV_ -(M(`@TkCKNHv`{NeRsFX -R5azw6eLqN^*ImGf_0P3q{R(CdNGQ}xE}KZ;D{yn{=z$9k}~Z0-m>CH$Tn8I{~rAtFQxC2I*v;<7XS% -tk!r$sh5Y>;stlV@s9KyN7)^&}WJDCG!OehtBW)Z|x(2NtsMI*rsaKY`CtBjc&0H@Ka;8#uwiHf+?)) -hZ-y2rkC{+CvN~Le<)2EK`6&(s68IAj}CIhN-Ky}z`vQSl+B0$^WI!wd{^)Y8RQCqm5lCt{Cur@WY1p -?=M9WA6huL1sRuAzv{kwQxVmV)!)FOVe#SLgVHPk<0>yA!*XxmgN&QQ4V`xay3-9S&B#*Y$FT`$2dOK -Z`=gs}eHPb%n7YGG}TBe9UWN(jeG(b3_mlfLE_Vg61Lm3_odV&vc-t`KV@ -Ikczud8@6Ji6cPo=U9%!Vs%X>cAFN+%n|cmS&NpWel;pODN<-=TX8nF6| -kwRJAv~B(I*#{XuIk@i`{`0WUzt^R*=C8hMsUBH!!~s9|{sEAg==k`DjwtLoG;s)3^+!eh17*(JEJ#= -~{>!FgNEb!XJ>D{7XE?g-Wx0p=slM!D($iC0{kEFZQETg{)f6!=}xBpf4|&hKHBzH0m7}0c7X0u3QwN -&Pc&IuB8aC!*T&!oVlg#fg9A;kx?QLutm_`O46%D7@bGv!z6s_BC!Nks2{Ae(xCe~&iUO9%WovTx5Ll -sGMiC#I7s5^TFCh4J{db#yy;_K`h*V;cT7)7B87z}$9mXImli9Mauju?3nk!6OY$Ju$R$=GQIa3==dZ -APOWWmHLK!MxeeUl)!qoF#w!`mk6n=i3&b~pg-Gwu{NOjx~)!n$lF0wj5;O|9Jx+1VVAe03l93e(R(^$5|=;;McHz -498(`Pe?-#h#tjd$5OxB#Qwnr>o)RI@emaOy|*g)Gg%nZV=08LV)=D^g58-X-04lFT$mSJSW44iG;$0 -Cr2mz&j8r!uvA<{+KrPSi!w7MO-la?=AYHV^oPgsIEQxJ$Lhnwkh{D>U*-=kOyg3wkDVqwbivQQMqpW -p!16Fc1+GrudI{-Bf}d^7Y*uf;wtqh&3PG6`*txgh=K30_^IC#UKiiQDGC3a4Zd32==6f1)JV@H$dM$ ->7TnC2>#m&Ak9uRq5{phgz^Ig?uGyg%ian^0T -5aSoT_b)!wiXyLb+Aa#**UI4?8r63==%8Mt7!vjyyUH*d~-if>2%~ly4nf6JauuCUT+uzlTT1I}J!aK -)7aM;t1n#QwkBL0NY!)lt -z{aXP{8W0W6!JQxP_DxtIk1?eJAbzUVfn9iGe0-Zfyu@L@(cgs-2Lswkp@Ez86A4Uv}MrYGFnj3F$L6 -!x+$D%VAiUjq&7lgrFQIZu$0#v3B5~f6NO-iO98{mK*Nf3XZL3~i+FCMea4S`B;DEa8L>zXz}iW8ju; -2af9_taA%ugMGPkK;okT48whL#+#~vQ%W6Lt|xP9J&zPL4waBFg&1Lef~KO+?s$v=PljF_mq9Of$zNO -9dnI)<$pVl{z`CYjZR$oM#4`pk$!eFZEY(DWxiyXLt9DWg~zT0SJ0VXgN25QNX@&Yhp4o^We~rfkQlz -RnMm4BK`N+O}MI@Q3H=&^ih*7C2<0&aUa0E?koaM!NR29ywo9x=OTV(5`C8=yR1V& -T26k$A)-V*qfE9hg1fSrk<(uL^o>3d@PKx?KN_0pX;;RNd{17rcFM%+wJa0!HX0=txw}S{QvTZ0|t21 -+fgtwJeuWV+sQ#v*xj0O_Ym)Tiecob@L-2Ra5EXV=uRQh5^`$mgA8Ce+UO_V*ZM)tVfW49N{PLw&dgr -cm3K#aaw0>0R7DLf5CayZ)@Zh~%=G_O3Zel#S~%kV=m@$DPULdl1S;-QCuxXhE`cOa1dJZz;Cx$Ksm&-##tHE+=AU*j21kwy@`2!2h1U!ECw*TBnw+GSr*i5%Ns>-%h2sMRcTp85|J@q+zfcc4 -FW0EL`n@S0xScxFPW^LG6-^$H0B{gw$W^D)jTG?&KH8LO-99@jijo7z1oc -C+k-?xyJSoywHy)ZH8Uba(hQx{IQ0PdRnC8-|)>(W%F`_x9L-tsX~YRI=0Qf#n&?bE!RpA_X-d>?^A>lT3kILo?&m#Y0eKN#S8YH!>Yne-uz>eFUbWqh)yW=nJ|hGpR -BGTfYl{4j4>f21_~yd51c5=RC;9fy|KE@r(yK35kO=ts{;28`rf)4AnRsKmtbK%t!STC_AJ?aTP$s4& -B2*@ID!OBncpUCqI>q1AMd2OsYKG6KgYd*mciCGf>*4=s7}&nNbGoxQC5sBNq3xiMNu0`%u%J)uQr4*=*m&7w~;CTW-xwVWpdk7It_>0ZaH22@Ys -ObBZdjtDc|RR>8@FcT2w#?fZejCAN|#h=;DRq8wJW~oE-vu|J-t%N((yZn|`R$tR<^H^F@#7ff!fvlK -^b-}=2g7V=}0LUKqF`*R6l3pK1j9!%o@1c7d)VF@bj}}qQEH9JtzkpW#P|6pyHYv6-@CK4oQj=6%!zZ -R@*|54CI~2*N%ZX1=7AC>AsLrDrw0$+-*+})A51c@af!GnLBXX|;#BK&=`{8bsPz6tqhkb{cE71sB?H -Y)T*Cx)|lAHiAvy`-FhF#9(aKP4l^6411l`bSaZkWNn1{ -%+zHMK-h=I#lmfrxGj#TH&wBR}92XziWAXy|)D=A1!WL$m$AuTl_3n@-uG^T3i4xE$VwwGH#dW0rB(c -xh*RPO3d3-mV&=XHGZl+$7|SLkoW+5!^grt)`l`LNS;702-=ZtaU!7~ipO)Mhe*X1KF$?>Ldg!!y$B_nQP1 -FRT0Tp?+kyP<-DcRb&w{YTQjl+2=?5+g7JeKCmf|5C@lp39!_e6iSgJapl#+G5Q2Gi^dW-|5`Y?*FYP -&!u0;5qKf(Eu}Mt!x{s1H?Rz|AaK(1W!OLy-$s^AmyU4&17~C!oNZrj0MHe(pxdcnfArRbLk(Vwjewz -AQx7et0{r{gukvzUPC)14bCbahw}e|Es5$2fPN&WMT>l4nUaSRRRumP_#v)#XnJ6Pi3AALLQ~l9Eew! -Lnl}_h`xZP9+-%X>%%GDl7r7Bh{};mp5VhN+Ay)b)+`T#U%T8-6snFS6vow0b0K?yg}Tnbo@VtV-R_- -6V3Q-fxjE$hA;Q@!1NAZyB5T18cU{IwJlrojRc+@uCo!S)pq!vwG|M*SVocyGb*%<`g@@H5pS@H49+` -_4d~=-ON}74sXuFS&S((yrB555`H4CEu(?W?tAa^j3DP^bDL8Kf|-E&Bcuv^V0wc~z7tzn3YRe -^55hjU0wo#^GFLIgr>KgyEOA^Qj!DI+A;phJOKGOj|AmN1`^yA_YgWTEIN~G<+Y5<6qFxhKgu-CYrj8 -x(~P9LUUCN*Y>GZ-{SdmEGni4ssED{Ri3|>gd0*D29mvx?ZKjct_B}WP5Fa2@E-oEA028tNfXdVxG6Y -8T;dOl{f6Gxo7wDVjX1UaP_~IjERqWQt-m@hsdt>8o*@_@N!^#;APVIsKx0<3PpOV6K#5+`w*rvds`QuDVeoB-{xqOfG;jy7`pUp8yy|cF;yH6u0t&N3#O+-$1EHCJ@N6gedoc$ybTaCZe#+jAD0wCfMjP;e(;CJ9D^v7CP0 -z^{@|Rj!JlPUzBbFZL)*8(NY-L(yC)0Sq(q3SCmBZTr=lf0uOC(i5ec~F}u<=cJ<0m)KdoQ3Xtf#a2lF=#&~r -34OV9#kq&wb*ZYeD+(NCbVIy4ZiLA2zxLmg_h!aWr|lviNmJ$0CQc5ZglD!&d34UutU18p6?&YI?Twr -8?&NgocdHW>#46G>pEbU)~aP=2*eXAhhxHzzDDs%`hd{pYp|+n-UIahSr7JOu924x*JB~CD&B8I(Nvv -nHK0oU=^^CMokY%C@JX}+kfp0mS#D{EoJ%)*$AI)0*K>hPjL$?qRq#Nqeu9sMei#iarx -vdvX};eq9+ --G*JNdcT|7EMET}&A{s);0z=~K$-HZZo&^d-*`7B-1;8+dvI-vO3xv{>NQ?4+D*1vl8ghfvfc>p=-!3S{NX{)V|MTjT*tA -Z+yG?ir)sk%5C45)jCIQN75=wrFZ*H?Z6lpGOJRjTZv()QNs6Pop+6!{=YE%BQr&|NScTL$yC!Ql=#F -g--p6kyIJFEEQZP%UeKs1|ScTi?hR5)rD)%>jW8L^ -U+)0g70Z^?CutTB@Zm6T5NowPh1Q{*!GUce*?gU)+l?$>S8a2^e* -N#g)K-|H~{w-z5*sY_6qV)SDt*FZ`bhu1F_WDfE+C49{hK{iH6}pQIo|mezOeGuWhr1T%Shu&_7u0=M -@&0FsrL?eMaFUb1(-BWmF6mfMAL3!i(X^HDiN=`2J~oE1Q$oYp9zeD65TChF%uH{+p1vdDVG#5T?Tc{ -~(VcEn(PB5UI2#X?CfQf?4G&LJi(Z&12szxbT|6ES_KP_he~3Mb0b^+x$lC7dGPpoA0Tp-PuE@QN2@8 -I%s}SDlIrBlP98*XIbOK_VJC;Wd=2MOy{YF@fdEcRk9nOUlj|q3ZiMZ7Ea~cnm5L6Ll?L -3D1#dD&c8B -c(lVRK_Gwm#=wn(flv<%$FKQ(=AWLNJooU2-@1ljbgMydx}uHf~Q0)aLQ -DH&BXbnpL|93a=4la(2W7wz39(DS?orZ>AhJTkfpzt4p|IV0eU%1RBA#?)!pWgv!_ZhT-QYN>VsCssVp!dbiT)Sm6lG -bf`-#{21{g8oP567G#6Y00sQ!5OP+jqX(hIhxIJhfk(zPjOu|X)dVXA3U`a%j(X9e@3}F^$pCzMqE -Tz7e%Gi;(QtvD(-;?cEHbQ6K95Jj4+7*QW7w0`^W=yPO{A(N}hC#%%9mVPbkOkPLR{}YM-S^>6}n_CM -2FzI{UAPtw=^HQ(WVLVOEgv+Xcew-PN_yTO#w+Sos-{dP^J~g$3e)@J|1Z1A4WC>n|}~=#!ovNgqV%Q -F}ctwl_~)G!IB1&9R6Eou`f&rnLUhEC&Fp8@eP0W1@y?|H(eO_(TiZerHzaqAL=J|9+$)>YcwLN{VSl -4NwG)afm+&&T*xnEvUm+Uz28geW+$P@~rL6%#84HBQ -|~bO{L2aOHX6!kt8NJN`tpSJn36hDZ3Rbx8V-z#b^XzQwc|1jdO#<^!?1vmWzPo}Gas*H%tQ`He8HyR -opP5umt1(#NFz>1y5YTtnX2z&$yq43EJ>rXlJFAb*DLO;tS!5hi&{GfF}iC5!G4O*nZJ7y@M#@JX_j) -oIUFj>9aYQls{lf?5E`Ct_=+<#C;D0sNpW$BPCg7+do2AGyINBERqU0-rm+w`{wX>bo?Px^Irj@b;y1^HB0GI)~B!K)mU~3z1YhFaI$HgUJu^$HV+F?}7odEZK)K`k2T5fQPeJ_z6 -VWi2a0W7iPAzUCQwou1ACy`7!*-#h*OoxJ90=9Jk6hE5~Ey!TJ_?lKS;UoJissmVX0CGY7WAt}e25+X -9d9E5zT7*}GiMa1qCW!utLm5w7Y}y3K(&V0JE%mi1&HK@817(a@_2*Yq;JHNaF+ZwWYTu4pCQj2w7Q^YanQ5nmvsGer2Ou~fn@e8Y>V -v_^}U*nZH0&Xu-p&391mRk7zDoFix9~MgE@rWTvy$pBL{Rs0=q>)L5e;I)Wa7qOlUwh@V*M5;DCw -1bU>3N$Ya0j}8H=#gEn2q!i9Ek&R*2mobJ8mZcu+5UE9zl|$%gMU)S%_DXGa!Nv#=^oyZpVG4OGC)Ee -H}u1G0v0nsKZn*q8Xm4z8G!6@U5;`hR3wJ;tqU3fyqax0E(`aNbpqmCEDx?#UYy&}eEZQ(zPb^GpfRuBZ9q+qc98OAA>iqM(eix$8&m~6)yj{C -BwAk1s(m#$y2cY(^7K|?ji31qu>>PgH?vhuLblpj_?ze&^Gy_|ZoPzJ*%iJ8WGNlGCq*qDY%+X@bm!h -*6Qw8^iae_A!J(vg57A-d -E_uv$p|FK}9q;3b3n=jr)~LR^g+u)1O7qL@HIN?TNvqV+G -Q7H@MQ<=@a4-G|r5>yzb>p6@AM#DQT<;n!_itYAga5bXG7kaON}%aY)Scb`#Yc&yfH~7HM>u`{Nvx(4 -HKPUDcG;2%%zVByS+X=_Ru^ZJE;Ol%>q75K5#Fdi3$jAKIN1wylQsJ-HKI_SGsr8H=f&O=@|#jXxRpn -;`;z?b=E4Lf+ASOXux^5=*8n{p=XGx_{03WKHGTSkbLnR_Ac8di6>W$pNk0y4oYVl8Zg@vaPheN -ztllXj_pM14^^--<%X_h#fqE$Y3qam$*8l7JC0Wj1k+$;y;1Ehh8@n+FNSI`RraNd?*TO}Oqf2-lwuv -e1^mrm$@J~Y0O#De(BDHH_>ZpmFqibXFSv}6>8v@@njvN8FTg? -{Lt9cMJ+%UF~f)GkA$<*IeGbXjQIU+)QOIF@EQfuI)wb*Phn+z>6@ipol-jNoiEh4_)*$b^*XR#DE`4 -a*tI<r_xB<~m-n6%&$@ii*ruf1+n9{;;QCX3bQ}&4~6RGL*x!goGBjVPK*ZW{Qx^vu{So -tdi3TQm}$4%UL5`l#qPcFtgi?WR4`lybZW>g)2AHj$`ga?PyWCl`|DevyX6M#XL2psAXz~Nl3d92%T+ -cE4As|g3vWf&QFW~4hmdV{c)ncqMn->5r(1u>Ijmkk44tKqE%F!Bts60fL3>!zh|Tz(-?FTU{Mq3#8w -k*U4j1n&A}f|}LVM_QUM2_JUx8cKCZ@O0vy(_2@hKzq+ad!bo&KdaZBcfHQ#^$E(9S5dPk4@K%|YbV) -eE2<|xuD4*F+Itv?<$TZlg(u5574!=%nU%N*SKp7&#&>g9;oxUmZqUI#sk`o!ny~|9${Uz$GvgIU@^T(W -YvI9txM7L^5ZmhJhFRAeaxD;lIp;2c6Sn)7~@he#gh(px7EY8@{^8I)bJQykleuHAG68EImZOq*aJ!A -(HZhCMUyDfog?Hg^|0E^Mm#xIU%Kxa*-Mu~jhB}Y$rjQml#6sbG5 -@ig&E>BplKDK`=&k6vd~BYdrQNWyLsIT3lT*KQ_^8e0cYTzBSOG-G#7uZz{~gAH+~jn$kIRg&nF1O-< -91L=5$VS^#*js6*d~6TQ2zUSgr8hTZeG)qLv~al7Xf*tW0A+KBJ@1Lrz%hTN&?2O;BB1oPo-dSYNOFplcJKJ2GW4Q<9YY%vG1(9fwXvzn$jL%3@Z`vGXO`_+`#M|Ydtb_LC&FUwRM;Y^N8_AbFfid;AeQjLS(Vb;$O(yOV -GrO-#jOJA?b_WL3+dhcqd;oH4dz6cp0)x<4CavDK)<|PHR&9OMuLk<7{E;dl86oFkeuixL$#5?nTg@( -Gw~#yW8P%tg7}AOK5^X!oz^#S{#}Mi3-jC*e64QKav-l}Cn4QIMz!ELne3lcEQk|i6*$TtOy4sDj)k1 -VOOz#~?ZOcKt`TMJU{~aUBhH^1P{>+HnOp*7FK!bn8ENbH*0Ph#9!AFuj*^j%@gqL5T93I!$ta+}n@` -S8Bvph~%mnE!|%6|QMR&LZCHiXif&jH1MrblVYQfghFODha3?!8wItZg99>71wGC-rT-nXBd->t{$DW`;cD9X{3|W~Imzvwm!rZxYt^7uJbpchicYu)%+ -nDz34JmOY}~jIp_vZSJP!C(z_PYiMAnPpP)&TE6t9p9rjnc`H7JL&KiRqm@%ud5Cf-o^uMwf?DrL1Y}{}OO|4qp)2y78t4u?dpq1_=ZOPP$=X;E(&#(ACZ4fIc%m0Y5K -5o3A5d;r=H`b!kQ3+-KMFm%9ooYs%L42UU5#@7l!~I`_#y-F23I$91%j9$|lwN>GyG{E7*+)8z-9WPK -4z&_Zve-0h&^G!cp1azEwvqIb6toS+7qshq<2SmF8X)C@ZM9R%-y)QIW;hDOV2QNoeh|VoYTvLgFunK -!g2w8BkY$Z$=!mIhC1%GRl;LYULMhg9m#-cyaOo1`asQ=36pmU?-Th?c1mP#J!Lb_FR)H1_^R;$_w -}|96Zp*nhB(lsnaayaz2iDJ9}1}oG38|+C@Q*7)e>#Nw{y8y -*H{l4SiAp2#e9Fu1zn(Yi78<1Ozd1mu&Nx!5Y- -WS)~0+xje^feR8axNjb1m^|?OrPt&^+8ToWnI&ByH7)7#_$5);VM498+qH)B5K#ZKf9*i3t$l= -ms6J`H)#4Rqn;WtN9w_8I9bqdwQTQX-b0>}UN!dX{gIg0R81&>gM0kT}W*fq&wKt;!vEw-4<1eRz9E6 -YRY^NGDs{M|W!IoNhcpy1WQ4^UK{Px|`16W%{Y{l>Qb5wREn!{Bbj3l@YM4*hZ@Fy-cOsN -;)Rqe6xsEFCX1Hb97~*i%pVX6h5i68?9oiJk0YHNQA*{37wVT+CLBQ#ziKI`X3j+P<&qrMG>;z&3prC -Ap->VztI&NjttofJORa~a~;#aE@-=#ZCOM&qwyZfIsfQ&oRLbqy4tSNd}u{ef=w{3W{7&#)a?ZM=CFy -wAp=9QF<@se^+hh9)WRW~Jgf`t}%yTz1HH+3S0rhCc?rvds6VF6`ho)o&LBjV4{P`H^T=Rrb&CtG$sC -Y0PtI~Wa*kbaFNpNO>eCcIlaw8)xk`8GR6@Ta)93Zb9V2&UY1*62OT=a0qkeYc1?E!bM9jG76?S@s5< -r_T^JSSCuLS-itVkKIyqS*J4+`gAz%>drPY=lf?!q_MkquPW1GS?X`kpoc(D>A@FP`IN~nU0#Hb9{Ew -R-D@(-58=Ze=;vceF5UT1327y~QvNAGPphCPo!&`ov+}LsYH~79vfNbTV^$Y>{EKV62zrh(uU&UuD&b -tcLTCh7G)eX~#5WOrX&RpMpGd<=-(lpCmjlifOA`ZU?yBPSI3-S(nzDdKWq4D~vX+g{w!Qh|O%!y -A>F^WM^80 -SGg$L@XBfRll9RC}tg8@+h#2=?ME^x~Z8K8FP0KTpLE`FNzEu~w(Vjw1u>M^inJX!oL~JOAp=jE -Huo>ZQJ&u|_*{UUZ)itncfGXehbqsG9C}yu!Xl56v{I?yrF)gcev()OV*&0Pb=16H@+Jz(xH}n+)R`s -4J>&@sQqRX**J-KCF9JFd8WIl?w@p#BA!ZZ_M2juj1R!W@&yT^|tGw3L$S=LDs --J2)*xi&JK#q@i;ao8dq^8h-g@mK039Tz%eqo>|(622^;l+O62xC%Nj{hNjhN=pUK`op2)4JEn#V*9+ -&6lU;Q$vA)*13eCQUg)`e>3GOwk6)*>DGY^W&PMxp9-_ZjGE^RparwxWIF|dk5d7>zIbULzT&e2rIpl_hsS -$JDSZFs33u__*N2sx8(v*79BYc&feE)T(dp>arzScfu)6Syz-|3{y%5jIjr2At|I4n|E1MsMWKEX$(~ -UtdQYxhm1U%kTc09MHZHrq`fn*A?DiL2pac(JQmGdgXQKyk+N(UQq`wi>23rt6%o~mZ@d)#_InE|E(X -uF&S5RZ=u!N72aF>Nv;Wh%a1~Zl16%I7D}p3+%xAjH;}2e?MJJDH0{RH>gUGQ#hJL{)` -9)FW`^U|_WjWXe5^WZ%YO||=?pY)x|ZM8AA0?^3_lhe{{cU16qO^LV!O%eXS?KqBFnMo=FhTrwD>K%r -2Gc|=v}`2S^#DTxq?e}$uh&j%dDo4HJcer&-D%YtYK*GIj6 -rG26=A6SqUc7dJtmNveSh|C~tp4*WfWprf>-e16KjmxW1OZA0^H6RMtSD8##qddC%j#ZOlb0))#0>YX4Fuo)>*ep`FZOMxNk@h@opSQQyIE0cQv7*x)%GW(o&=Q=H0(ctIl-5&9LkwW*f8d=pfm_tw|Hk#N& -XC+KAkS>(p@k3%Bm&#=ulaQ1LQ04G;2jV$aZ4p3)T^W;HAr6!+k2`R8e!91dP8uUz4YV@Lp^LW6hDTS -9&~CCDvfE-gYaGW<%1qw94k~DN4b60o7MP~PJ7{OVO_ -JY(roo_^qpI##c9g?LXmo@m5*1LQ);n+Yn55z&hBKI#SC|J7|1yQBOj`xLkVwTf)9kL7CchN0302 -RgBp1I(&Z2?g(3)L#La#ce(`6KN!;|2t31UA5eWqPlM@L;jK^ug8htld$svYsuQ)fW-ijpy4 -?pYlhmwz!>bkE+w<*U)&0w!|wLW4tNsHU_)_5k`-!%&*OCWFYMI_$Khp4qsGc@q_?#Q}^7sMhK{SK|^ -?~KKz6iRO+qv2+GmuJnAiCJO=D!RQfC&i|2QP{BEQ}*;iy-^+naGc0$JDT1#%kL^jdi?B+w$KQ_wM!G --Eut{B1rWxUXOJT$Tf^b%=-rV+Q8kNR=|bLCU`2q-$ZV!NQkFf4;&pOTDVA%Q!AfXIfOa!-A^{)UCsv -?|M!P2i;JpNmVb=IF;>9}}gpyyNk_LwJGRl}`{UT*VWs5Fa4){V3n7yHe9iB2zM!dJ=r2JSU15}U=70 -CCM9+Y#-5~AfS2PAzX6zRkw*5ezu`Q&mgo -tEM0TK?QGfh8$hg?FLUO714%=RdWQb%D7A0NU~m)ZU5J;Gm0yl1X?CQ&D$}Q1VI)`hv*GV$q!mC=O@X -t(vXq@eKP#U;ZW##PQ=JmO7B&RdO^Mj)y>>qD~wI`pLtD67*^p@7N?K(?Kr?^5JcS0lliEH~8J^vLU{ -Z*Nk_(#NvJN+6#WBhH6?u)DuIZfW#ck>E685`rTi$2>qbS_`wmKyhZU{POOQ -u-%0@?OUgOOx^GuA_+&=fAKq4jwY;~3dEgb+xG?<1hx_akalL-@{*4;1Zx-8(1w>7{k)CFVmyPsBtmy -+<(l2Lt|%;EDDqs8M~(C|I%h&TOM(dva}p>$UaUA!xlph20WZ*8{}0T~3A{a+^Id{`5d3)UBhK<^3%K -n8uIu+B0y>r+#;Xg0}5g>{Kjg$*TU43W3<^-T!%cz2(&RSyTCR=wK25$ep}C|O#x`Ik?J8=5KuxK=L2 -tm?nU5}6Z>^-RoGE?5Fr#omE6rMGAUVSYLe7&DaLtfnGQ>pkjcPtOw55A9MtY-=v7Ll!WznQqbd^Y1T -p{}k~7UcTtmohgNMOYc8-0?vi*Zh^|&wQPQ?6tX>Ctn2p}oy`&6toK@u23KWB;lGhzQWb7aIk`a09{e -8xav=fa6TJXQUm{RC0VZ!oC8g-zBs+|o9TD=v{3bpx2t(C92$;hFFk=d`9f5dAwub*{A`|2AS69Pk3x -`c|6o;A&;5=4HAlXlM-t<%MV)H9J5M2EwXKnz7N{X%oQa6pRAbW0$oP(c%SmSMpw1uF$#=E~yFgsGVY -)-}=9$u$B~wJM&`iw=Dp!7YZ*SbE92ZB3_=<`Ev`@gwRqHN0Ry`%AA!zE?tuNZyW1`CW>Rv@u~j9(4y()UU -TiX(M`nWgbM@_&a!WPx<{}0|;knUK;02E6&0y9J*pU7-a-x2U)r!TPNE(G-ao -6=fZ$kl2U9~NGsP-VR8QaT2`lq7po;15z?(f0Rx9x=#yd#CRXioK>Xt&e_R0r%ANOxDDqd!c_J-p)4z -1b>0j!J6L+cnbq7=d5GllkxxH;{5KfQU=C1wO!8v|U?uJ#wwU6CYD+laaj~Ah~xJCexP?_NlX6Fp{nO -3PczBzk__fS#n=omZZLh+5x*jO)%jNbetrkwR`rf`+%rGmP^WE;iYX1ccE5-1qzk} -Eb*ns!yg25w>&^{Nb0c;9<_18}3@E2{1TbE_AK-sDFnrH1%R -*QX*&awZNLPGlgVj5ut*Enk!yFiHUPsF_T%BTc$Jv|ly^s1pL)1hT{qVUw-aa)Y2V`Kp_U>m6Q@|$bC -TMBpEcE4}lcAb`O-a -729)Ca>(frYWJ|wKBlrWeioLKe9B7duDQQ9O&JZ8;e~8KNVz7ig4K7Zsdy1C;A3bz;8fh>1A{8>q8*I -ss*YYg?JsUa_g)L4ysNG!OzZWT=(U&O-_iNNm3E1Y&06^9T_) -jTN#+FKWl`0j^w!XKp-yb{gTo8qc-H(`!6$h@)reETdDEk0U;=5nde-gIa<~r0^4kl5Ob8+&N7sxsP7 -FLP-_9OczT2MK803k{Xf$^MsP;@v73!hvb<`=c7W&+mxO!lw5B{{$z>E^Z5kS|76c`EfGrRgTx}XZ^{ -u$Zn0q(s*c(PrQOQE*g7xbg5(c7$!HCD*g892Q;XkJk+UV1XGbfu45|?eCoUT-d~WiDp560nyrayMzH -wKlhjnVNd4$iiyzICKR>^~&YN>Ibl0l{V;`Mk;*$e_%cmXPWJ2v45cmo!1;*o6GCHdG$x~yLCbVl3um -K+gEX4BF9L0ErD@T69*$J)-8>cTHjZS2V!ER;C#;^}q$fPq=7UN?(cK?@5>uM_dQ#X65$U$rlMyyx;p -VHIeUA}K@)0BRS~NFpuEx0ZQaj{{_`^_0zVJ>kzUN+OHt+_k7Yc^p-{YUaI0YtJQO*(I8_An0w9Li3x -vtUJ6KZAQFhRK3@tRX%CLPm;$Vsoj=?nOGZ)BRCIZK98l7<5eyx{5x9JCxlnBpga|F*St_)`4k73k_1 -}jXM~bX5K;N85&I6o#8UZDrE{53@;;Uf7)i+JRj%ZL((_PxNc>QfU`qMOaaI>sVdF4(JljvwcNTcE9k -LzG%u4k@o0MKF(dvfZ2L=ZHP!I%82;{ph(!C;x3^w4KB6ra> -DCZTzmIZz5|07D$L@@-xZ3D9W>hFCiX|8B9B=5MG+Ym01mm=rU|c<3A%gU2t%Xk|ze$8D6hF&AecK$D -zXBbUJ*BO3d@hG(EgRuBsjv+LSl>`;5%+`~mn(dJcc(T2a1A`16c?MD5ARklySuYDV=|zm=j^_Y7MuV -(R3Av-wOYKax>SEIj(_eo9+;^PcaJc$2mR}oqB~&dcRLY~J6eV}j@|+|#m{!?J4+j1v>}qi3#u6)phl -_Vw4YUnre-Le*)Lc`OC5?^TeaPOLgk@StOJ1rRs^N6PEYAEfd5N$5)|*rN*78jh|l${0QsJ>^ziwMx* -q6X;BkiocZt-u@PEM?x;w?sQ8pE3c4BR!TYuq(;`;g6!Tg>i -6t-i0Hp+D-9%wSx0K%wrK)vj=m|2vSWCW9EVm3;9OE+LYCj%x?e_4GM2qZF@ppKzBoFrjoDGIbucr%1 -{8Yi*wdAlzJAg(o%*b^9eav)2-%Ny;aFVx!N}NCGCrOE!)R5i5q#DGpZCEDIIs#@3$~zQ7} -6?jQ|Uz=I$hof6i$EPQHfm?(RKDe)n4|5$I9R%IKefMDMcvjW&_mgR27Ec%=x9f&`7e;Yq<&87AkvL@OP4mUT!5XGf6P0Z*4x?j?Wll4e -3{uSu217z510=Fc4m{KOM`0vHE?wRJ=jwq;6Sc18!r%^6vJ<9pT)x(I{Iu@I(_mirE`%yt#S!rE;)s6 -J5V}T2&MgFI08xx$HmYGpY4=P3 -+jl>^C6VD0ZUqfOAl5A)V$n_cwJ{1i?H=-&1|-$+p3{4C0`j{nQGIaXG -u{r5F}N@n4;nbA|%YI5jSO(x?@Meg33oBpip{7Xf~p{xA0j|<%^T8!I`cB#{U#Xnh;A+i~GA<%Rw{i0 -BA9d+dwittOESA5=wzVNBRwYJiB9d_K~nEveP-81vEl< -)Y|(_|uyK-mC+z9(Kaj!;XHz$QbtK$gqtoLN<*Hd+2zqMksekbhBV{;G6{MyE-Mk;c_27eq}s)tN>fP -2yfx%9+LrOyH1}J<|oA-ww-P?!+W-x@f*qcosemz#CFmZjOsH7k-XxSVQAgzXFyW7V{ZP1ce+sOw{)5 -)-$CAEna!#~A39k%l3XmQ(ZDbGZF&;UZ}bj&BslnWS-d9(-qlxM2gJKs+xH*7rv)->>WPo|$8^~x6w^ -{s1gk<6p>$5co5hio?#F2g&6^hacI%PQbTc{mO8E`y+Xwg;AcD5XKH~O#pF*wZ`H!EkMF)_r>RX7z6! -(a@=1sXdoz{0^eEJYW$vb{($cKwqKF7HFjJt@CV3XPaXpThkFDxUHYmni=AOT0T^%*`Jm> -bjxc*8Qa^Vx9fNXZyWAfBd7DAN_$ooRR$Qhp3n0HNFk+;r5{0j$W&oe; -lt4%1<@@w{c$4x~9A#GvxddSL{(ky&29EOJ|HIz9z(rMUkK=oI4lf536vD@-s1!a>C{cXO$c)a|OrnB -XVSq3Q$zvL36w3z)bb%8^T`Rj;*UHMe)wQSE6JLO3zOo{%q^{PX(ZU{xWuE_9d+#$dAXMM`yZ8J3{y* -P);LKY4wf5d?ul+oGuN{Ut`P_iJ%kDu@k+*AIxvs$1IJA*4r~1S`0uZHrut6|}iqAq5R*@5V;TL**q1 -GF(al_lz_*rVy)gdS1WeFOsk=|uC8!r>=gcmwk8KJ&`;eY{9dL -4JL+YyW%iI^Vw+E{YGlgKgyj`6%mMbig7j&zLgB9%`uH&`3W_><-0!nrCHma!aWXxO*F^hEUd7ngp$q -Uh}3`;E=P+*AZf>H7d3BnoIN(zY%re(tWS=_IlxO^d^himl0I1_ED;;I{|w=p2l8D6H&P*)KqKo>M~U -PjM1+7x?;-Rgft3opymmOs<{#&DwCy$M=Pc@fTL_(f!23Z>wD2Um|5c|ap&z7Q~vB>hkH1kc@ql!U=F0(*eGczjY1_~H370-EMX#EBExgic_Ss_GvEd9AklmL~@|d{H>gQ -+phj;;uTSv36PvCQmE*@e`&_^d{{tPtLG+Vv?_`de%{W0Kxc9aL -w-u4@*$^aS<;ojTBB|J=V+U{$tjEMoCe!q7y2VI+GY66((&;E#se#v{-^G@HsgShS-hu)*f_i;SaB}L ->fMtK|3z8{;s>ECao)e|{O9Nq@u6xC8n=WXmzyF9;p3#autSo3$Wo7<#ehUD|58HOAaZuq$tkgI1MveJAE>$_k$vD)YT3d1Gv-Mtu5Woy5r2>*t4jSBB~eBV -F`;?RrL(5k*zQcP}{?~V!)A(P8W6-JM<(&Tr0wO!_i;pC`?y>y&~ZNy)=$?_)^3SR0@j|X8U|P{y6ID -nYEMx4UWO+6_hY8D|C&^yce;n8E~U|LLI7*_>uPAAEaZ_nw&w@v_D<0q -WeQTUWnaQM}PDzUqCuvUY8kc1NhwUsr4C($z`0?N(egi+x$X+8>ulfmU^gBlJQO(fC6Zn -4)jBd+P%H}|5h(bL+K~seh?LuJmJZmp@Ft1M?f6->GDs3X|GWE>pX~vrO@4C>&PrgCN9U{_lRTZ! -;f&IBHYx2&h>}2*=d}3YRO9s43mt+vOIPb?)V@V^X}5>k$KlUN{E5OJRpmvEqe`FLA$+XbxW^f$6^=T -CwZex^f30xb=_PzDH0>g|fh&FA$lY3cf0=!@Ho4J0Wjd}{+OL^T@U-sG>A0uLK6W~OKeoqCCw_Qt3g# -?B+GFtah;YU^K>H?sTXqI%-^Be!n7S`^VPfklEcNVFUUN;{RNz`VJV9xf -C{926P-)xH9A4R4<_8H&FmDn{vhroh-U}-&7xNLNXKg8r>JSYR^LW9Dpm_Pe4#}vG>zPyyxWmNcV=j+ -I_y(`hD;PP(K)PFf|wyG;!hq#9a9ZCL8kg^_r6UxD|+mmGzvaV_`US96tO6k?%;63j6-}1euSjT1>~VcZeaYhsehwRO!gh8q15RW%M_^;MFp_DNhbtNO|A5Q6l=e% -uZ;?+YkMeG76aN<8!&%psD{x7`d0dEzliYR8SELXe?8{WwcQUhUfBqOq0E%Le6Du04L0Bh-+S-U9w7O -k?}i_W^b6X2kE?*nY&9Sx}mU5vVT$YxFKWs!;vCBSg@*j@;w$2|Dle3hIDD-vi>FG95=r_N!qk0$>D~ -yl8dM02mG?A3~=oKEyFZSEFCz9JC8{0dt87)YTL%ar$6$!)mgs>N#Ba>ud&@T-#W?wnbI_SF{kCsick -EJsl-XRTTo3(`jpB#jhl|phx#Kq1Hcr@|l&P+R95lE62gFuc{i?$yINgm*|AA@zY*}f_>VX4X|0?B3* -_NT0h(cikt>sG)-Vb&Vw>@UY#KC&AnG*{mH?&IPl2bf}h|9R~t7r+G%E(EZ%3Ph(@n-ch$ -hqpy{`j#Da>VdQ>}f=o5o`nY}K4{E-X -^)4JVc(>`3&#ph=^ucnHw`2wB>MTomH7Gy5{t7wQDdE?UGobs;PeviQNdOOFN$Youp^a_+N?llzCTZ; -vhnhC>SM^?TOiHDiH$TlNOseYpv7e~wfE3b^2iE9iCfc6KZIP=Pf=21WFl>uZ9kDlEYyDV?xMb%1xK@ -hjxSqB$)z3(hDJR+}?4{QxKmu_8Ehb=Ad2XcgG8E*pKSSlSp$-_#*}F7n7v-ECGSvD63eD}~F476MOM -MYuhkNF3oG`fVtfgZL9EacqNH8y>S5HKWd$+l}#>Mww*GM;dRH{5~sIY%hxhAHJz%c*9?)CIGIuy5!= -<*IdLV)^d)jHu_dWZeTbaWz~xPY5Gz#$O=$3s_!7bhBWRQw7jLw&@wU2Ok4?p+sr_Rxk?P>EAELYs0D -DjTjOs-*T?CCPQuf`;FJmZTMS(|ca^cwyoRU~?>$5!6) --9$;Xrf{w4iMYSU^Za3E@y!>xW@mJt3x(1mK#!lbWo1y)&hx5xOBcQF&)*HOU9|TEanwCCUPgjY0*Z* -r0J~6}{Rc&~bEmA0L*`2%n?bqS)gP=N}Ku85u4%E5Z%dW~3z@k$0;#w&>cl+2wn3WH;@xkr#yW4#o^B -Lj|V_PK|c4xMLsf+t6M_dzLV%1#LS_bl3i~#tIQ*FjjQY4eFJcNVb!I*JYaA9{Gp -2!RX!6vw#fJ0lAE50U0XWy*vtu77*C^xC%)wpj3AORkkaUr<8FDcLDjc0wNw!{K4|k1$JSMJAg74P^1 -ZuK7cJSE}TW9RhnwU93Rr+DV2KUmzGb5B&}lQ17%U}Sxl^ac6*eMiN}Zh?Lz)`pq+?zSO`WD?b^_;74 -42dx@ZYu@y!oKdzvyLSzg0f3)9TCa#--H4_1@H+R9}f5+h6Yx3I(ETiSt0Gk-^=v(SNeO2{{BLLf2O||>2KFw&BAn#oqDO -L*S&_wYK;t90r-zw$3)#TxZ~3>Wylja?H9UK)l4lZ0aMH`haf6_l9jqM5E{KgP40qmh65Vbyv&o7xdJ9yUTqjbtkg>$9fc~R&a>e#s)+qzCJn8@4l?AE$*)Tu0Of&%gWPD39|> -#&Q5yQaUL(p^21BAobUgUo3tL^B5!xymeA;Y-u3N)URI1-Yx^14&oPG@JWRlp4Wb7H(a|@rHp(et8V3 -U=rh6o@{Z4lhp=&e=dyJJACC6ZM+zBn-K%7$RGr$PFGbU;R_F3)DSX1256x9x&tMat9-l(h%-LS>9`( -l@On(p)kw`lx6P+d$nN&?BCyxNTUm$mhpRA&RG;_zR2mf9T>KU-OSpiEtP(Q92uf=>c(}qFcpl(3qQfY)XO&S`g+s2Py=)@FymbX6-2 -8tB=273jXRbq+d9N+nm{Py+E96*2A4RWYxX~Qkg_KDcGrSnU^DsbGaTtQhoYYvmO8p{i`OBu^)q^uV6 -TRqA1OY7LNWq&?oL?)b7O8>8rc!lLfOCzY5_7nB?=SdwkeYhkP88y -25L*^pLo@V4SG`$ARm^^V}AH0Snw_WmXyz-WQixsR|ru%Ro2<5YBo_sJ!^K^a|YuR#b;^6x+#Eqky~D#Un -_qie9g!6E?>*61KdUs)KME)UpST_^F9rgatU&JkQpHW5KwEhHDl+qU*zZFG*7QnwWgTHVS0%jRJV4-D -(tW4mAp;kw)PkRM((q9n)YPi|T{9 -ho)*>qr}~Bkw{I)-87x{(4h#+o~l{FJfHvVq#w+613MPBJo)4FS^9z+)9^ts651jF3E$(8ndvEuEv|a -8^u@m;ribDa-9(l(bZ}$F%8q1UR)EdLqe@YEpUgWi1{<{q(C=8c)XNpuP_rV_pIXoO1Wpb%3Y(yQxsi -dYbh?}Z>Hr5<+zELy9UcWo(U$qmkesV2@9Bg0nlHjCA@@}aH_{v5Z6UJS#(Z^_4qW*q1^iVjY{!z8#} -Ble#nqW@gPG$b{G;XzRr+%@kNRRiO(<#jktv&2Ju0Lq>6VlWQkbKkSuWxL-NH^h9rr%GQ=uoGo(yhVs -mMQ7flSuThBxd!!g%lBEyj{Vyw-jG+rFcaAcbp%5ZvrpAW;a9*JGIQ5@@%c$VSFC9#d+coGNwvv4rs%agpJud0G6F;Yo-eVK}OU7WXmSfcQ>^qiSgJIfgGmd<(;~5Z}n~e8g)Qj% -sqnH4HC9+{*BB#B&*5gZL7LHz00e_$I_P4BvuyBEz>K9?S6Uh!18s9wHP&OX!f6Ds011VxGIM2N;q=r -VB6)^A*3LE!Fi(ut{Jxt;{AAY_4WDP0VHq*km%BSC~x}*rYR?$BX6IXxoC)@6onIg2&T!)Z=OvR1G#K -n9Xu#6ALy+m`ysfi3gh|W|P8flECIoW;2o5Xu#$rW;2Z07{F#5vk74~`C#)Jv-zcnH8)`M9J6U>Hf3P -*IJ5b<$khOW<^9ZZPm!x70?Qg^`FxS9Ndn6|nB}G-H{ZGbAmDxoj?tBLWUi}Gh;M9kkId)Y;W;utV1` -*j9Z9KyIRea&`3&J44*2`kgoztA%JYN$F#Lb_x9-1+J8BZ%#U+*KI>oHI@PMg%9E%LnZXmhXaA7x2!y -R8WxTo85>%#-`D!!Cdc@;l*saCxQo%mcK97JW0;~nGc^7P!~4b%P(i7tC3>@imFwKqFehG-bDwn}>w8 -gUV&0cPG~U=lu(PspQjLe6m&;6V)afyudmTKjq6UeZrjg+^c&!t)yH6jOZy9b?&w8rV6$kh`TOO=`c~P4}gX`rUM#@XryC2F2Lag=Y3qqk=K(RmcvK -OOMc(tijuPujjmS+4l}4)45|Y#U}3TMSw5I(2P-6ZAz)sIvEqV^K#48J>DCQUW^?5-`=Uod9rY-v{WN -WrqS9SpE5oq7pxAhE?REwmJh=YH%c=&0U7`C%&34!_{aC`gq79_qc`Hh*zj*hpfM*yge{Nfq0<-ai2V -^DnYzq_p!VjZWDdIq)yjJ2Unc_Pc{IP98&p$}dR>T=9e`#C!7TP=vHsZQ_-J0urp0^aU>7L?NDdJKBU -GopP7eAjmKNJnJ!#|iqM*}+G#(_6_<_R^^L9TG_&eZ15fvL+6#^lCQm8kHD+&SH-`g6~VESr)E -AK?Ckk1xCSH4MaatYL<@Q-=Crf6VY+QTsZ&HuZj1RNdo@;E^+uv@<>$b(`*F175**$x8X8p!zEO@~@| -dk%zB|(HU|+(et788Kx$~F-$CitPW2SXE(Se)F)RmP13vZnmyqclkg8iz3)|zq2}mIzS_#KhoKTN+>4 -QFULh^f*4b4_c%xO8#;a3v;ha?;m-lJM@t`A2I-(vVcHtl?u>W -TM#bqi`2^JxR`{-`D5W$df+AEaw(^=UB~2BC&cb9`|~H|54~Oo>qlR;f1bl_Kt$#E7j$jS?sA0h+2#R -TXxvQp6f?0uC%O3ZLQe5q&kdn7!~5cBiVk3goUXD50_ZVxVecBOaE72(F$)hjaIdBAJqW>QfI=u(ExaOpNdCoG -~$V2c$gTxJC;{Ad*O%D;aoF=a4-*pYQ!BbFAyb0OtXTkJNvcCO{JsTzxh>?>bwp05m!Tgq-My>m)aVg -$^>HhrM6k;pGqf!88N2zi&U~+nt<1@x1aUmwoWF=&PSN8Ju3Jh$(Qk>gp+3B43#+CPDf>M2u!o;7lz; -oE}fzNRW&NiIiMBx@#Kur*7?L3rG3)^9YHk%*_latXT|OsKSV1oN&^V79Wo4urom8Ot)}{LnsjlEcn| -<)mLWNh(30Mb2fK853mq*GEFOcQ{(kf?Ma+q3#WDeA&)%q-k`Xe*br^E+Chg=V)m_23OF`K8I^`x!oN -B{t@ZbIe@B_<0RoyCQ3P-}!V_(^tgy#fs8=)5#utSwsJ5UP+W~glPV&4IGd*#*oQDeo$%b3Hxedn-Z^ -Y7=-GJ)mLe7Wx&UZc1&hj^oCdzgaf%SuaFCrfN+dirbTL3=ep=$SG!(J7NNIp$oHw+JjK%rOs8oa}F8mH-?~Zz0eE{%6gbFi0aK -_d=Y|`y0|7O!aSZfSV7)+geUp(`khXt6wx*oB^r$pTHzAQFth{O^0-=3Rb$l9&Sq5p;_J -9x^sRI01GuUV^+mMryokHf@1^9N2e4EO&<3k%67+PfU5a>b94ncZFR_DY2`8Ofp6r;`ci|LY8-w>hRJnr=43ZqswKS -zo#sOdh%zG>pC&S=);)hLaCnp^M@4Lw)OFNVH#}i@`{BF%0P9x)@%naj~?y#zPmw<94o#;YDUE*1A0o -E$K@a1Kryw=Z|T@pt%Ya=emN^(`2ZW1oy`$PY=Kxt;Llw;evrGBpKD;og<^*^P?+AsA^YS+lx9gmk!E{a}YU?~yPV~bsDbjw_IM -E+9PJ}U+7iWq;GIbM2|A4xQUiD4H=AYcTK__M6p4#8E`1YL13eX4Z!V__CoV#Jl&cd&m#zsN* -ENv=Z<;tR6RPLgYFXFB7N%?BWVLIkX)Tmb?rowc_ZmEOO`P<{bB>3k&Xh=vsXk=aWc*I@Bw?HQPZr?X -dW6qv@RQtRk43F|BTYGt;IUycZvSo3!u$-k=a^SCE-`yo9YEzm&Ono}2fP$b#LA*NQQOA`9(ysuANO1=ENzh-$u%A!;AEu%GIa@P+W_f{em2=)Q#JT}+R~Zj1?hgeoyF0SY?nlp{c -`3qO@na_wJ|D=4``+Q;ZsJLYv6>eK1nUIsHv-pU5aN-m;+UItY;pwiaO=>iij#F834cwXAro&;$GyE6 -JIt{iG+<@>03qESmVXw(k%qRu+hJ<*;atDRuVhnKHAR@qg+^xjc*6ZAw>R|&gh1XWGsP*oHD*g$nT_~ -p!9Wvc4EKxpR}OjD}VLO(_0!h{zz%N>Te_Y6lbI@N6YJ>={v$^d7e)S^oY!*dn8!Pob3??Dq_Xh-cPv -a*SDOxXms15DwBzUtiS44?Th6-bRo1rsn+n(v$^>z6pUTB8qgI+KmUOk|~48tBq0@$e61FPGYjt -t9G)?!FDTEC-FX4C*h$(LUj0}d<2&Un3?!E$I}8rD&g?U^heC;at?x`0uV~={LDFXsQ!pKUDWKUI^vu -XsRmN2Jyb^oX)DiM&?f()Rn7Ya8q{2-I|AMM&>bJphL -5O^5l4?Sp@a-fn$?*eEToB)Dx;LzqH(uA|FY2PT@&IQWX;FkXSd`tvO=A7D?6|0i}srBb-VMivFPL8H -A9lZ^f-}z&zF4mzctcol-=`NF3~%%XPwGrZd76Q)UvH@JlG5I+(8F_fYV$6O~5%%9Tb8O=z3~w5dvnZ -whb#9!}|`m($G^G~u$!h&|N(#}OXx<;mlywup1GwusM9!dSqL2W5^Se^75uhN}7|x_=nTZz7hKveX|% -ZX6U-=YEAf6b*4MKtc%M9qH$k?7g=Ql)uM3xMYZ@fQyHgHE-hA52Ud>7syn8@tFyI*hX&amoN*8ypja^li|2#L~ -E73~c2IsHfp*>|2;32;k#$&RDI`se=)Q#yP^3T9#1x0FaHVjQ~H@on^x@!S-l3WtQ8Eoi0XmRS|X-9; -*9d5ez-=@Rq1nak;M~#k`B(rfcfjKA1L%VMq1EX^Z9wU?r!+UnNBx^%V2&dT9GKrQp -AF1$e=ZEnaRlNyFb8=t+NZ;BzI{CW7NV}uORF3Wv_W>(mQO8ogh1Dc#^Cl;u -)wJsz;pMk05*__;Qbw)Xv4@q1JyI}_CBs>Lf3u^^+5GZ=-TJYWTLgK;gpq2pk^cBn$Lw+r^=;O0;tGZB_dfswO -tFpBcbY*Uy8GjfZuWU0q`5Is_vxs6TPo&KgSTC_R|bew4X!-nC{; -3B3=HB0d?y8ekCrTOp|PfJ?!j>Mqi&f3^(ktwpsS(i36;F(DA`rjuCwu#7cViU0mW)phvD -X1hMjN18t&gax>v`@6OGNdyPN}SH(>VK!C9dG2PZqxpJ=S%mvoJ=@C&MsN@*S4x+R7P^(@BHyvuekAR -dqG3k?)Gjm*SjbTtgwf)i0oqovgVQ`EFXu6nn>_VJ7+mq2e07=KLEm*qyAv0h#k+&4{!(b=>Fr+a=y6I$jC$P++z}Kb%~O^^_pU;{#0rDpuX;_1*A&$HJNv -2LHlvOtapzz0GL*`TlKm|_hrT}1(9Z+tmsAAm!=e!8R)j#2oL$0tu$JDh!T%ED1aK0V!mtIK8V1~I($47J5XrQ#CW(l-5+oN&iTdJ(ef -HOPp!F3HkGXOkU#Hc3~_ozJUYQ$d%~CgLU&949|W*qIldP>Z|mw3QbX>*BS^=hwucIswQ)^-0zJji@; -$fDO#$4NIpYSwI+G^f%;GP+e8Y9k$WBq~!{`TA&t~2vyZ}^qhz59<+As1I3t?vi9dIQi{`cFFMoQ$G% -9gG{Ep5Z-(ZZ<=SX61cQ$hksiqol?I?n>!;|Q76;nPp^55tkFEIa4Qv4r4~#?!)29WH{Cc7%+&4v4b3ThJqLC6d#>r=ucRN^$f8B{+iC3={5AC3zi>>0)Qk -e&|J9Awl8}B*f)B`otQ+x`e9$9ZtipTDwM1u<0e(}AAT9D~$gIPixYC63YpJ$#r^O>hdnWwMOD13{X- -tRI9e&)Ed2H~>t=&z`>+-lA_hhFt>BkRqmf941lK*?*+Ssj4Lqp+2mH6<}$RrMJ1f7X=Xe7510aR>EZ -IGQ1r3XCQ0vSGL6ec_Zo;T#MKYw&oSX-W}MRc)XJZ~;X8CnEO09tY@>JnLukIWf0pP8tvj4HNXdoeu9VP&E~eVtpNP43yfZ_OpAB -&K3bSgj3#aPmcqKQh?y_ITlXs2oSWL-{s!ANMGPw~BiEZ~u>>6GJj$?8Tn`&cPfPF1WAPWqyUXI#Sp@ -p|QkaCp+4MClybR7Bb*H}%Q?`RUt{q82oh{m=|7Yrq -QjZ0br>p%(YMJBAy(`CM#jt_q`VUhkR)Mm`{TcQ9s{KM>BY}s8g&r8-OIeS|K(SC?CRQIpux=rbiXHz -6EDJ0QN)|FF=?^bc4@2Dcn@(515t{g?U`p=HnNE;rS_#gM{li~r;-Oez$YBl?xHtjKr#=s+t__7edsU -YJ378ekWhES9EEURtp1TT6F9<11v@XihGS&H -x*vmcD+ya9PBo+*RcVINI$3F8=+I&YfL_65I)qa%*|lk{5c!{1RFN7l&p((o) -lN(N0KC8N02zLL_Ys*s5~s8F6=|%pJ`H~`eWM&^|2-4m=tj{n4IaWm(sfkKLk&8Kd?+Py=!1`{zm?#L -kz>)-3-x^;le2cPV6PgF-rWZSdK{irI>FvSqVOI$SX-~Itj#$l=wIo@e5QZP&)!DCnj9<9a6s@+X}kG -ZRAVbbA{mdX4UBwNwZb2Whj(hjoONTV9Tacy?sP`^yiA7=`D0i0D`R@Pn+Rk?pTCg#I8o@LmY-#ol~W~a3Z>K6lo;H7ZZU6s^o -V$W4AO&e7DHeEoA~XsVaA8py2yLDey-jkJ;&!vg`t|kc=EGFqj|&i^Hi#fRMS&MuJ#zJn9_9zJ~Jez^ -K%cZ$b#`Pv>=?sCMY+hnghUcbYsrK#MzU*oZHSUc2}`8}2#Sa1Td%UcZ{tGx1uNGeq-*G%x5H+$k)IYXy;W7?)(6!NebO$8P#x$i=VA}h9}jAT(CtrDm&pS#@_M#VZg9lE -0*d$43O#SCb@gLe^jjuSipLA`t`>_RM1AH6NRanJoHDP4@tvi|_%0pHHGXs~+U!l+4v0ck6+ykZMtSY -NxT0){C0gNnn8N9)?v*6=5!+O+Pm(vIFm-u$9MKepGNnY7YhxkD2~-`=N&929-p)s)h*kY?26eltI-1 -&Fk@v~>L*8Fl!S^&m24*c_nO9Y9z1%%z8`=59Bz_~@IQg`(>-$@z{LvEzZ2Xq8P4l397SdbF_H1xOiM -kb@x0JD24SwF}+=KU&jb7opr>reO>fFQcDccjOs-mKRQ`j%*J!K2~x~D8|j^uGq*$`%O-BT8Dp7uiYp -0Z~;B*}G88Fu2H_mp|Dd&(+BuWlEWP5%P2gBO*(tA-@BPwV}nvW0i_eo@&QORu=7O!J-G2%V+eP6^G! -tNaGP<;CZO26{~y-B@D~-)_U`@v^eFafwK^W1p+x_V`xY<8C~5d8XYYhQ9!vJ)>)Y>0*cY_gStkd%U5 -3k%B7tY7x_xQYQHtgl1>t6?(bW&11^(uDml3A~P^SrRpG=P6x>L=~(Rh4&ZEAqhot3z6hbG9Dlxr)+U -3jr!S=a7S_Pv7v_tX?S4|kcR_?*`X`NA109w$Y90OE4!@do-dUn90`bIbynqv!uypT1mrLbaSYe?Wv! -UaOTYLD4XQ8vi%{+9xXK4B*M0~PmAh%PUhZB9=?m_5&Yj5%G!cRHkyjy#T5RwuvhV_;a-1{mV6_nzi{ -@$x=$anLw-JJ;0Ude>DS;Ve{qQmdGSLTFvmyHE|=7bItw)7TMJT%QSlxJcS4pb!h64I#JxLbg~-Xp!f -*Ywc%;(K+%rcaFM&B>5sR^$0*A>u3gl`{NYF+$%XOWkrsNxhrj3n>()Cl~ZmaFuK0i9G>&8f)d+M1sm -40=aSrQ@wBIDwlExqu|rXlsg2984G$TX2GlxQ#@iI|C*-(w-dtiD2LlQe4E2o4!`2Ci^I^7490S(<8T -p&g&c0+@KFw5;joFrk2w5>!=E_}jA3vvhhsUM#^GEJmvMMIhgnxMn9AXG9L8`M%Hic$2EXO7jYHii2A -6P{&!L^eat=3fxShj29DdH>B@RPJ^Y}TO!Qlc9mvOj~!#WOM;&2~_$2k0k!=E`+j$v>hhiVSvIh@Vm0 -uFEC(9Yoo4jVY!%;8fUzQo}^4nN}Xa}K}d@MjK#IUh%JsNrxChlL#8$>9SWKF47bhsQaz@$!E8 -*K0rarCFYh=V$+uu*i&sKIhIlwbr+FRl0!;}jacAsJjo@61l7H3bP5_@8;3dY2i{3KNh7n#IFiKUlFb -vzb;N{<23+gV -1K8rM`OevjEZ=PZHn{ZN`BgEu_j0K8uJIpc>v1_;o;TF`Z@_qX{tNQZBX2qcn@C-+ahXDPY;~TS}=c1 -OBZBn_Qq_1J`1*98AUF19QKUQgG8k_X6sY4dDvN?c84p*arePWcuhY7k+c7yM@Q8hS(gGKDw(R-Vz!s -x>@0`&=rDF91ku=noCBt2y9EK8GWt*x-9LOFHD_{#^eB7^iu*kS5tb0)IZBvHjP>4(>R(pkLXIm+82wOn&*|WSSxyrj3LfENn6jN?3hrF -sG1}lWk;pNMy7XoCtl}n1b6nSveQ}#ULoA^8Ob!E9*JoughbkcbF*jz4*(Q{5m6j`H -$&-F}dBRj7#T=T#To2jltDopeZ}RWZE8?wS0j-HPNBR}0^%BB3z?vV?bzPp|xuS=+ -$#M1UQpY~1I5?EY3XHwF4eU0}hzdxQ{^uc5l%e07^?Y*Wq-lP2fIQ?w+T@F5E%B8)Os7&A8-sjOWVzr -I+SXM9>Pzq|=yIHA?jhf}uurRq-@W_k9@=I(yrsQTdHwSQY<1frChqQ8YC^v_4a~L;o|CyQlb8{9q50 -E3~=16XAxX8>=+^pf|tGM|{7c)o8^tgGj96vV?<>rRKNx>QStIr3 -)MpVNZ!ii-Zv~ZrRq#85hU*!N-c`$oQ~Ighs7-b8%K(2ByNqx%_V|9W${%waaET|0C3X0=1v -OZD0;cNAmM1ipITZI9kdU_XD)63R=eoTfLrYvO-^_J^U@EEkZFJVR^cwfp1ScGi?LiJ{Hn+dYZFozF53+d@_BRo=is$4{~LgT2wT()Fn{4@ -QhWP;ZY;98Xm@f^Mu^o3CxfR;-cjuqqF2#ik4Yi0&zph%5LU)Irb7XF0@$fvV+#K+))aNo$Dy1^yWH> -X~Ao)wgM~zbG9wVRnC@WHb=J2#jv^A#ncG#Vq8ccX(LUfhqNI4YNUm9Fg%7K3$yJGTW&U`Guu|cG2lz -S!&d4t0lzG{#TE~XGzT-8ONy!GLI;Q0cBF&hF%0@cKj=<8ABC5o2}IDSh=cf_3d}8S-WmM`M>_CulS2~{axvH<(~n1* -x&lA>gBJyj|z`J#UK3h48OtS&+8BV$zj$1)9n9W@LXA@`TyM#pm6{F(WjdPDBOQufAG&O0spUmzn6df -3x_{Xlc(Ds&N9z;ck>_3uRGs8ZRiC#uBW_WLuFO{d&T-B_GgVk@E!i#Z)|H)WUFwK?{Jm6p^JtHbKD -XLgJnF$3l7<(63`Wh`VmjGXl_i^CP*@i4nE!5==cbk>2ZAXhVUv7@jMP(S=v;5^;ikNA0o`snp~Uwp2 -Ohxr$JR41h_@mG?`@$rjO{h_;3)9fm<^0VN939+>d9wqo7{zR!%;u3+9_{%);D=)vZyrLW;@V>MB&JE -?^zP=Dcb#*lqbbNeyYIzO#jvbSfy0{#|LnNuGsBKIgTV7KRp$J5ipP#Q(`j=U)tE|>5=;Ifc8{oz7GB -IFw08d1f)dj` -NWj{ULOD%T0hchY4UV+9^doFONpcq(FK(jA>`=@Fc6Clku0>svr?$%99`PAZ*IM1{2o -q@hHh{}$Dkkt%MfuT}S)qYol$nG8-J&8$g&L#;QviG&Galq>3}U7EmSx#sSy)hPnW`RDSlTC@Rrv*WO -Q|(G$D+>8vqO%?OmLLiCM+u`o?t0nIYtiLLq6T|i?M;n0$WLOk)_x!`*+7e7e -P+m||x|8WG(bItU-K92@ep|4S86<rsw0zuZqR{FC#?>b)neEBKv -jp`zgFN9`!Zq@}0L9+%X6`HUGkMxD)bfl|~XhAaIK9|L(TEG$`tgfVn)$+|LJ-#qC<{%p$~Sm34VSj$ -Ri(F*d^mLjWt_0_*0-!c$2E=<{w(41^27Fn1nQ;IA_C44Dm1h0J~T=h`hiya{g7U~x6uSoyOYxEu3=R -V)i>icPJ2@a@JrePSdqkhO3QTPyEBf{<(B?KsW+PX+u253@q$LsW@=L^UbM8a6nP49-=N!HNCIU|VFzz_#!s -{To9mEsWF!EB3Yj2ahm?XZ0+5^e8C2IYo8x`K%+A<*X6=5xd -gI`M;c=xZ^2pbrUzFyLZU1Sv=mZ2JyJtd@&v+3D1SN(gGkpz!!vC!zeHODKBKYOD7Yu9N -t}Wo>joBG=DeGXu2Rhc1TaIKcvf#L`{lllfyqSg^(Ta8hCi>ANUdle2D_S0KeJ>90_Zb%j?ytgnSBb4 -)>n{_^;4;R?6{52LfM1Ni>vGG?Y{HeYA{H6_J4?63QVG${`YX8o7X#U0Apu#6cu -P5l;MIoGJU8tOwq}`@tCo6~SI4SP@1P>R_VI4I}C_n7;#0)V9GLSG7eR8Q2(ZRSXVhGMxr84g4Owz%A -EdSeB4i735X5z@6?q7^2>3CZzQ%k9y|sMf?>HQ+L-Jp(GJd?h#P#5m4?ClfoLAT)_)65)O9A^RO%Zx# -ad?kjYQrJc=C1GP=llXf)-NPP#nT=}>zE%2o -aE3EuLr7>pVh1ad#0M5m>K?9U_7in)Z9B&*k`BsywB02yk3l?=^&g$zx -6nG98IC!xIFTYY7I_`CRl@qC#@$dqhC9{0?P%_}}?n39AhhLTX5yB$J0NxxHHkfs6pn;$K1nqiDT@v{ -5N1yHBpeM^qBzYpmTW!nF~pj5@+ATrn<2{JnfWOg7Ktmsb|7v(k=;sY6rN^`Y0jeW^-Qwg;BR<|s(@_ -}|Im$%moB04=Pf;{*0BK;->#w!B7FyHZt-qL+t5E+$toA)S1o_B9$p^gUG_>HoS+ -E2*!``u;RSAXIJKNFhV{fQz{34J5HP+uZJh9W0Hzd-v-c%dvKp{&4;wj@kmazS2h>Q!D^ynZ{4MIR9I -*asf+H(W{jr>UTB_apsnbQF^1;vv=pK2C~;@jx`l9`xs_td2wdg!DzFx!RUSnr0{`=-WH`w}l>I)B;n -x$)BxHS?!>*>gz>(6BR@epdbNQk0YQz0-6yo^|OX#1*ay&${I;!Gal+0vkOqT?37qGSCm5=mV;tgFd1eKBg1k-p*;GLVeav0YE(iH8D -Y05M&#xyMkMAaMksDkFc&_)3xv2Phmg_sOJsEJFJyG$&t$aXN8+Zy{5AOVb`1KQnqlr`?;6}o+(F2a5 -pKJm!S2ZS(Dr-SC5=*$?cbALpa1N`XYfD!@EQESx(|Pt!gITOy<(2Vp2DWsCYWI_CUHIsEx53lnr+Wt -OxA(PbM<*{$uhPgP4X49Z5E5EV3`efGi)O%xHf2!m$G5*hanAgG0Y=8jHi7UXBXIKm)I5-6fZBdFlzc -TGLUFXOAD44a|&rC?x;wvVm2kf*JOeESD0?hNJ>er#kF_;Dlf1#m*hGMEwc*>EjMHrq2H&y=i{1Wuem -7jHaiOK1+!M$Eomi-3vw+f`PnvdE7Dnr3o8yQ23t(-LYp*8k(IH9+|TptqNF8c1q+qSG7S>z#n=1|+v>%*Q!Aygq}0NRU$0+qgWe>sd>8ZdvyTM&vc+VwSE|Lb!p-6F-n^WQ#a4&iNYgjfE48H5qZBeR -eJd_`>$tX+(ODxpXW5D@Bi~S1LX+5PE_;!wzM&u+d>F#Y5A#igQ?UA?m--_NyRg1y6tTmGS{>E~m1;y3 -}qd!jE3sQrh+aM(*M51zId7pE9n491(JY;tMRSaJDenVQm!f@+A*5WCHvzg_3lX<|Cn5j=Vf5B-{fE; -%3dZSzLUyAn!c*u_)VCfJ?)RNj7uIotMYBJm1Ux1r<|5NCJ}&9_wQAKm25~^I|>C-F9_^lHYxBEkN5I6~fEsy>!?T#~ox(>ghjxa*rYhqwm?&nBSCKs;BEx^8jb2kIH -XeIH5=12_!iZb2Sy19@14Do!_)+u=Eh{`^Jz;oSaOZq{;h3OA?nFblZ-LJk*m`Zw~p3%Pv}!JUVM?lH -$R5;mSzKkmPrr@4Z|4cxtwyH{~K_ri-iN*?9r$9Q_S^7x+S_B(j`4srL7xqGBP`BKD0A<7^Wy((y8o5W|8*JY@%;Z>B -swPZ`GA)X?cTP+03vI{>J7uW`%|YFeC$aEEB@L>j|^(ol~cG4{O^9^#1ra1=vQ23P{ZM?f6%{{!`Hz7 -;m1!ueib?MxcKKc3@o8;4s -t+`{2z4mWYwz+nxCu1+z4c>|YtEI*uT*95pNEj&J-y`BE&|#sz~cd~hBtngf~)~J9wyd>XbdjL%0W6@V$Dp|1r$YXCkznuXm8@C$g$F)o0k#xVO>fCo5!5a5@wz -y}D6IxN>*&G`dx(lv~q)2^ZV!KidR4KOW$sNi5zhfPYH>o<}Iia{#|iV0;h(MkO*og8>Tg#)6-_06v? -@Xg&vU#kI^&A;1xnS^35QTm$bk@Q?87B#3vAf{X?D5xiMo-MzGHZSUsPmVJEyzcmu$588AKtyaQmuQbs2cU>( -Qr0yz2>7B&{(E4MKIAl#Y->4{QMcuO{nn*h%O*a5E&aD;eH#sqjio|nl3o&#Jc71FxVkppo1{a&gxM;Sggd07B7I$&-B?Q=Q6L-6K<|04jG -u7dmnZUuN#8T1{1X8?R*HA~NSfFG1YyN_uD_+16?8t@K)t2Qt{Ke}RQXI -O#=b|G_>H;Jfg~qYMDN=4Hq;;9~(k3vWK)&jI|)4oDN=n*l0!vOWc2|6R;J9H3(tZ^r;?UI*F0z8K)3 -Hz2=QrU1{ty9sa+;2Uo;8F&-mthX4h1Gse`w7uB>0<7E*?G?5m0B>#rz5|{C@bMJ>P|t~wA22=UCxHjo*FIouxMI*#L+l1Dj?a2NNF@B?m-@ -I1#6;<*CE5su?HLIcO~JU~9j5w7Dn?%{u&;|Sm4IKozrBm9Bm2&2DdIPN{4!EuC(IgSwb_+pw7?&9|U -dD!3(R36mE0Y1JFh(-Dc%qp722Zgb-a@bBX`$DUG9onW%L0nAEow|Uc-3U -ZT2{IQrXg6m$GYEMlES3cv%k0*2H%pr>xFD3;A1;pWSumUYFhyQZ2Y11b1^wUq1UAuOX#>PhS@y8#N% -a<<`{)w!ddIm!Wnu5ta2g`$rs(IC_!_;_K`tDMB@TFC&)~rDf-$@5A9hMGW0Ow{fE5Yy`gaz}eRcICu -A3S)Fn(6df2|wRSm-49lsx=+Tvk+Kn;^tL{m0cJ&_AZ$5Y!vXF=TQMNu`G2lD4me7UfU#lmDnGPNXA9o?Ul$W0Q%fuw+fttJCKfHCsa|fwke8K*#cA -C@$!RfbeNdKWH(hhcAI$YjJ{U4U3U(?^|xpwqP#DHT+@D< -kGOl^=;neGeHD~dA#26D#{6KsQP#n+@EK-nSH$L;g!mX0$y);{oz%?J4`}359PA{dvczbm`KcKGg-KB -A-Vnb+q=nlLqh|3_St92E3dplW&ER$J|bt%oN>vy6}K5|hkCnxhd-9Y9o9Rf*Ye&W>)-kAoxBw#WIYs -7-aC@C{vBr~NWzNs-~N>O8X)T*J#(`p{gSu#z?t;)^mp=1iGKr_xnk~Z^CaojOle*1n0IbY&m`-mC*F -BA9qH}58|kl?q;nT8Y$sHT?|}Yz*f1&7f0#tP`AdG|mHqhhSH<{*Jkn4;Df?y;JflX9B5`qXBtAZ#Or -JiTXti2mG#V-I=Fgu`7A;ytZoKhEvUKTEvMzTvS-yNZx%JjtNl{S|vDs|2{aCwpEqTl_ll<-W>Ez{g$ -)s|*p4_z}g={R*l80>?@?vu#1?@DCaA&F$X -FOihv61o0_MCPB8NGZJOUrOZWuO*VxCXr=lB~npQL27Gj$-Vd9OCEgiLGsW;50TBAHEe3F* -$_U+rrtFOLF+p;%dfcwGT?PSAwiM-k&k^TGklLH41koVtzpBz1UlpH^PoSZmuf_(PbXXLvxt>lC6CGy -1=Ur@PdZ*M0*e)kRe$3=;Bbaapl7cNj)!N+R?^cOJC$1|`>=(GbtZi1k)_v^s#0R72yX(G82IidG?MEYz# -kxt)6q;KveQpbx#>huU74&g^Yco@Lo`p|YWga=w3c$4{R2ww-`AA#_Hhw!^0{4ofB#v?rRYpGC`tuR< -=2!QZ`5FXl}4j2qc>i$HU0pT+s{Ou6_K?wgEgg@>Pp7bNBWDwNZ;V@>3gKB#nISi8DL5xHSyNO5>twf -rCCy{R3Or(2XBGQX3L^^uPBfJ{IUkBk$5PlhiFN5%RL-?%_{$&XNHiU16@DS&{pF;RE5Wd3|ek@S99w --z5g+Bp>CxOB~pzsAy=m?X@_Y)=Z<9vy9-X@U?_e$jAixTZmshg<5 -B+LaGSIwS1V88(I;BwEr&M#9lK@Sn!0PaI)&!$Nj>~c?0`Kpro#p~zl%)sT@!)PXk%;w@TbE;N5$y?< -cKQq&uX*TM4DqWg;jVXzvGSWG9fg;PoW4iKngL5=D9OLklgOt$=Ajw1b8d^8^J -MU7RYj9VoHkJJu`Ebc3yHnf8Xl>P02*+Mx8tU%oOx8FFC}^UxV%`GeKPE61O|@2LwzL6xMHEW{P=cGR -PVsp7hyc2o0NJ)&d`m9?nQ#N1efbazaY7vAZ*qAE2)TK{V>Go!Z0MgFdjvsGZuwnau6}n0q)&QseqAR -r$;|7)Uh!YsPcydwF{u%axOvJ;rmtJx-Cwa|;$MpyQEsmN~Am$J1pqNKJvBY_g@0S2pO$_P?wnlUra+ -0qvaOV~H&Kl8qV406y@*1LWa{A104H@(6kCvB${Pty}3>;pLZKrsKiaUw@sB6+Ucynmi2SjqNa2*tc& -VIdtd{9UGkf@)Y^<%m1&vGl7n(N)zylNCS4O*fTvwT00@^q#;pu2qZvMKva}AqG)S?0Kp`LgjFG|Eg~ -48r4>XZ3lLON1&AzyY@#AA1fqZ|h@cTfP!^HxpZ~j6H#~|Y1e%#MbB>pDdG)H^yZ7JMSC!q)o;`b<{r -mSjKm726^Yc#!O>A)V=uzkN>C?{CU&L9Y=UYy?>}*7Jb#%?t(Wzce>FJmv?dHkq=;k@s$r9&gS?$~-8 -=V1C=uDL_oJF!%{r4f){pt|A1UshrHs!B+e^b3bM(^*W_jlL(2k8Cd_5N9U|6;xWJ-z>|pZhIN`7KWQ -{~D**-|Medty*=CH<%yvuNGV@A|isn16nP#TD9P6)$4~;z2=&~sGQvpR<~BII<+FMt9os{tM$BE^%`~ -R=(X2X4T%Vk2zPD_t6w+brmG^Z(+jS-`s%;jSiM%=h~HmT_b>i8LPP7;kEmSv4|VmPdVjp;>T1E)*N^ -yp<;qu8saWw3^{&0P+LgbnR{y5oSH6ljBC3Y`sZvm-un5LYRVuRne+6IhAA0ut8dp`R60Z7Hz1E*A+^ -C=GEjQ`8?jL^Rf7YoH9uX04TG-WV)ejG^86I9e+~9Yk4xY`dUfrBJBd>LwD2jL${Ij(3X^`t0uICE^A --?($zh&^Kxg@kqJ6n!nzsm4C*R1H@^?zerO7R$Y`Rc%*LPA0yDwCgntX-u_6`nNUX$JnR1AhuaErt2| -fu;E68WsHO{-K=LLs#4*;vYJD_U!xB@2os{@Zh&U{`li}>U(w{IB;OE9(Nu&a^#@qrF*{k;)_M^zWeU -PprD}esHmt2KTg5^b@(aY8%i8{EBn_vofN^!1;&Z{I#y^JrPKW{rIJ-FK -q?RJLu~MuAFp?%XMwe+!>04x)bc%c7#9?-kdtw{G3~(zf -`S6svSo|8j(apl;Qc*%&+J8u7L8c8Y}p76*=`*>cI@4$sj0D0I -vC-B<|4tB -y5JAvQgVPD%Bh>M*HbMaK37f<*5%u9j)t+(Eqq({S0@FO3eefF8*!Pj4ZZO(xaolqS%$H -KxwGtmwB!4bZpJKXdB`|lh0;i2lkz=tzu?9*z#q30*{{Zo9dI(*B;z<=!6F;O>U5^4ONNW|MB;p;>i -t`)g=lgKN3MUEajR;api;{_pi!M}R->Zvtr)@-6Ne5Tq<_+fAWUT6+o!3SOdJFnsWS6_W)Caq9^ -D9@^b%IBDe1lY5$4R9~_hiP2U%}MR7>kE^_RYNbfBok*kYa6Xk;cmtTHCmM1l8)F@14VH`LpAI`%CSd -b%NEC~6D#M?6<#1$$cDuO9Y@ZPd-TE{(F;b`uSM=s96G7|-?rU{L+b*O79aX>h*Vwp>w1y8Yc3vswZ&6w -)vDD*`H+d8VOvVV16eOkzMwy!;}^IOJL4X4dsIik;Y*Q@io>mn!@>hnp7E2;7fOTw*I$1X^)u&8aIn+ -jQ4&u|!sF!0ld@&SESWpJle{n>TAu44V}N+@Tao*|5xHlV$erNusb8N}W}>VBr$ -}dosl%r(eFg`=J~zZBsZFh=HmSPe5VG87TXl>s3I4FKuwwaFACI3uk00nS{=o7Aec;^U0Q_qAWbO04< -kewOGH*z0$x$3;DGrQywVw|u4&4-vzwP$nVD%XstUhCtZhlwk{+38Ra8UY(u2B8oXqNe?B>2_d8~ImV -%OoB`=PVu;7jVEIlqOrikDYvdbX)N}(Mn!X9Ofzx*#mv>rymq~bid+IBy#^=pFU%gI(>dtpTWWEGdNg -%#wOKWX_omV5PtLw|DthL@dO7u1Mw(HM=UOkWg}b2;^8gD^LQ(Hd8pzrxV0QTv`@C5@Zs=?;?Vs&#lf -Y|o&5UTw$Nvjnt$l?&(ZfAILucZR?X-yM-F~pE@Sh(;+d>GX>Rq|A7j*6T{8UG3oHLETehqXFYH8Lfg9L>-Qt4n$ -KI3IU^nqq_%_DsG0kO#;_!M#OIbL)l`K#k0`>V{YLliWx0OkcwUeh3+soKK?PX-ISeZEJLHTNXfx*EY -V>I-~7^T6l_TI?9!jlQyz{L@mfE#@PZp#yFygNR||AHg(|JGA&oYbfdr%vhmDWb4r$)=Plo*+u)K(@c4o@l$V-$xGiUT7s#`t?-$@T~MRThl=t6Yx*mn -WZmQWh*&VDKpq9u^nwE!y&$tRL4*)+!Dw6$hiwiUT&u>T{Obq!-ho<@wYYc~)`AR2-gG9R8tlG&&*H0 -E$g&>xwa6+h4N%2YxI6ty;CJ4Gh5d)KgCt;}D2PS-6}&eM-Jw_kz4VslBWj+gw)q^?9*NpXc~-nCZvi -pDIV0kF}Kvi9Q?{CdN?z+*#+W?~MO_-#LHW0MFI`PQB-zdpf10q-+9K^bsA%$jC4_xaI3YJmH8=1?E^ -KjqMve#W4o@{ET0p14fzg5AehL4jnqc2YLDBmrd+NEHh!k1Q|Sda4`;L;ZYV(7}%sG8RgJt#i0Z{Q`) -!zI!&1}Wu4kisZ*zpJo3mRGI#D=!7rz$r^~>B17*^rNk*@~{PIgvW;0S5ZKdE -Ov7bwcxv+`+~AY-5ZWetn*zx^N-*SFKugt;RZkv+y@)(4cAW-n~?Wo~HPEL;8efQncwQE;-= -9y>Yx#ymfY15_|c*yU87aZ^f_z8;#ISe>pPmx1>A$x3?7%N-zfd80mnlt!pj4{|3W1!F2q!)(X3HTSx -zoMd|Cfs@FouiQ-mG!Bb_cwX{_171u4vT&!Lx&C()d}JJjyvu!F~Zw#zis3J*nt%uS-GY0;v^v>7vIm^ui1`*m4HjT$9!adA?kMh&S~ubyaZZD0Wobi+ -Nt!OlQjn8=h}8{7KftRAxJA2I5meK7|9`PC8aW!bZR)TuIN%$ -TDOKKP&{B_$cUYhGz&2^`Q1@B}vS2*iaN1b!H~$48?p)Fa3vd5;(g-hoRAL(%{BD`<4F^&f6l-OWl(P -2Ks#6HlB0&in7b-%NDc;!wYSeHk)jh~Wc1!`5~z9>5Fyw&qK2z!7~wN6?2reaC0Ja}V-o=OXKG=oa#? -vXnh~^yrV#(@vc_$?)OB4IjV(_}jH>C*8VrGw;I#aIg~`Y~D*QMtqIj11Cph2;AJVMNGx@YD@lEvu4e -vOO`B|Ob&4=@FSZ{WQlq+?*SJz=Rt!88U1bFzP-f6#7M)24GkaooO=TCC`q=gZrCGoggrp6u*dMu);A -O;+HK|SUl{lm?@3LYHVso6jzgc{eDh6%1N5P8sQO`Wh>wq#&Ye3O*_tzFj)5OqLvvmOH+q2m;eBFLg3 -L4vcY&{V>3Qe*9ix)2zkH=$tpx5g)_d#o=Q<2)Xht)?8SJ>C$0UU^gq46ZT^G-LmSIVYh4 -p8+ai@=)(B%<3(+j@zd@BY<6-Fb_QF-Kz_)Hn7|HRvuDqy#+Y@*6<1s=?8Wfg_+RNg4jimqEC~b@(GU9?*8_(xvjk3ojUXwmQrNZ(ExuZsHu7v -Hb(JnRyi|RQPvbFNWXk7u_+8JGLo{C(Nv@EOS42aXq*JFS3R01Q!dt#;5ZxQQpskA6QhsjQp$r9%u8f -vT(6APustPH@pWQY)=lJVT-UMAlV~&S;EKq`bnfR=dowGx!Q -0Ec`&-YaF&x%eNu_E<8o$u*97LYG;PDdrM9$4C7uV9URBh}}3N=Qh!N8|liwfQ`T&0}Z8+<$Uha?09Y -9v2zBTcpD_UyqGiFg=A^-V%A}Bav-C8(7MM0bFmaSg~R=&Fu!LP9MjfY0P<4?d|t!mwr&Y^qI!zIU3v -c(-@fVFx;#&OyiA%efIeJWYj|G^R@EVaz8#SviT>G=nZE7AIsSPCbi+Yy616sKixiSV(aJrnE#B#zd}u%TI?rBL<)W`SFbk&J3VI1nD}Yaro{mRu@m_N{t>&5FTz$6%Mz#X8Lt`Li_ -Z27>G9C-q^3=OhnhS+whQe~4xl+`islRbY<^Fk%f7_X(8T6E#7*u&zwE*9eEmLpWlIi~RJu5491M^Lv -%mYJTe}Fv3_0B6)$G8A_jQ00T8~NoipigJ^*E}hH#*7(pZaCN*Ip@fr-$qT9{sJ}OrO&5cuxGgMgh-mg(t -3-p*M?5e?{x1^Trq6eu#}M_NA`2~UV(wS3AA8PU#HiU@9+1~XQCd8Jkn#Lw=ve=UmD)iH=i4mXkak%S -a7!Y9k_pBFF3FV4FA`Sk2d`-dPUSUsf}BCq%KZ>h2DhearyTj(d#y|*X=HOZ2iHx{3j-xFkwRci!Z(y -hc1%u6Q6@Ous{QjTXPdl-;jP4b=6n>we8vd{sKBde}rD)sD#*J8Zg{8n%=Ej{?V<+AAdYW?SDVpA0RI -vw}U3w9%yl-=u_GBY!}lfpgz6OUtgc&(h2%Z6a9U%k$xB$MjmzFhjf1N%ene*^E332nBU!Jarf?^(}j -9Qh3mbh$3m}eoWhXd?=`x{dG?R}Cr+Fg|I$k@#nBrg9tRGaSKH?Xfdg7#kNAv1|JL*PS)C|J9$m6lqW -vev9zJ|{%J}i)lcB@p$&-!U!|vkyu>Z&c^gtHb+ud{eareh^VD~oFi2!*_YFpeZD{Fsz3Uv-@y3m;XN -_DZgmLAv(v%G;9Z%^$ahxU9d2fx{NRvrhQmB*PG53qfS{$JO>!tNgz7x#5^bhN3v(wk09Oq8dee%j=B -z+`QKwUzJ(+8i$WOg_qfOje9-Yn}zL@|ZL1Hd*|yRN24h3z1UdpEhmUG>VCd83rC185#Y1^yu+!WMrf -~_Sj>l4xE{pX=)cX_krddp}&2$5qbprWCm;Mbz4=o%8dUSG-z<2>b=PW@%8j&=q+iSCEdGsH+Yv;YYy -a%_2c&0N?o4kp3DA%ztW$*HU8K~FOj-&$BrFM{S_Dn3>aYOjyZDz#5jjCV@XnOG`7hfoD~1T@ijkYw*S<7Zw)IyU4xG{_LZ+#@PRo;IJ|9qD6~j)~s -2kR*Y?==45+{m$wT{oUQ-`yAg4{QZ@S_}*@4sc@=43@Pn1mB;q8X -U|@elaq5yZPbr*=g$4-ym|A^u{#SEEI6oZ^nc}*SIql*&6T;ixgE2zvR3Ek=No=n9|oUQr?=K}BnPJ#)8NYMH&CYmsSUI2+Gcox+wq^1Ux-B0k-!VNK(A&xwy -6d@y^VYc?lx$Aa+(d^zy4pU-o{M|Cd6HlatrYQNu6uxgG>?eB{Xd-`)H@<;Mlas_ggNX?aIC|vYf$;p -8o`DfpkGdh|W=dk8rGmrSzAQvb1BKITjBkydkc_%RpF%@uQw;1Fw#EY-}Yk*n9J@%jU)pV#Mkh4KQ_= -FF#&!Db+@>UaX5$_S}5o;3bl5Z?{ae`TcoTG!n0skI&-~rR;#)d(MT^q7x>8REwpCooCZYOUbcN&p!y -V;}bPLcUamzgtXns|_BMChx*OXMXm7B5)0(8T>CdKSmrAp1J>Q@>#B7CkwhAF#fZu_&XB?Aujn=t&Os -k3_To5zBvM4_l9%!pnQ_z1PqbI}IJ`mpmKay!DMdrc+(yq5QaoJvk;XjSuTPy7N~5P#{!obmf>SLkQHUk>2c*D5&twG+ujO3vD3p8PCKb<4077*HOWq@6R+0~Q)uG# -EWxbZRafE@Rtt4H>nDTsnx0NyJx|kj?sL-3E5_fQKkHg{?^vDbdS#mVKKMZWzrWsDo;7cEhU%-H`tCs -etcSUdaf?&MwSJtrI^DJI%}zrzBh9g~c`uT`$YAI&&|IIQD<-)x9HMDJW4%7X?CCiDoEmz4XMNqrTnS -}Tl_p@(M!04g_rL}{`zN=IPfu<*_~x+Cm8qzo)Fe1Dx#E|&u$)3`tDH?=fD~jQH->nM@3kAw&~KgOWWAkj(6SOt((5!$2#FxXl~ -uy$lT_+U2?nSZp2s%kP^%H2G|{Xm*;QH-;rOGe>}gkC)iWh)7;}k1~ -~;VE~{!*R^I%)g?Y>K*5+-@E6CfCw=1tGFLXhqt`ei`bkUU_)wNP}wNbj>6kRb(*IcNpuGMu5bmd*T_ -90!}@l^Iy)g3~0k4W7m#?!&m#na96sHd+d)icyH$}`=Q<(cnU=vnSr>)Gfj@a*vH@)UUvd5(J=Z)I;) -Z?HGiTh|-uZSIZncJOxbcJn^!?dwhT4)u=mKINU_o$k%@&i5|#F88kWZuAy-cX)Ssi@YdKkokLr-Li& -eP0w1MwIeGyJ2Ja@c1(7M>@L~eve)Kp%qhw_lyf}C$*r7QH8(idQm=!d-=n#Gb5nDN=8nqE%AKDZ=!t -$Czm5MptAh2;SjH)W-t@+=*f{`R{#p@p>fXxYKXdWLlFvE -EfCn?6Zrz+R1vO#2a0!-l@;T^MC(P)@jp=FXx|E;%~Aar9Fr*KmYuzXBTaTVfXFJm;3#GE_s`|vFn$e --^#uGj21rJ-2Heju9ny0)$Q_n@vyjE-iz0_cjDdsO(pJb-rn6_zk5YrSNzoV;{M@o@%=mch6Q{v7uTx -Qy&j;Pn|~IxnS~S0gxkoj6I&&FG2o< -5!Emb2{2XU`zFXRizw`N>##5cmU_@MEesmHDAhhx2rJyEPSyMxv1%^=LaS5`nJEY!n6$Z^;htN+dW1v -%$d3>!HJY`V4*3;Q)c6zotapw_YzH!S)@iLs{xvi3^Inx#Z7UC1c*93Q>Wch+L^2MyNGhIgwuInO**q8E4`GYw2Hhc3FBI|B+1|bu* -GIE6Vr?0@M?<)s7T_-dpV)2Rz4_PXcz!NvPdpw;Cyk^x7Ur+Rmz0S&QZTYs(h@KwE50qj5CJSrr|u`? -?$16i7nRn?ZDXt7&)m(V2GGQ+PSv5Kd`e{g`_!D2VY7R5AZsy&f(94!3lwfREeNzYq$is>}Z1zk@r*j -`O3nW{V%|jqIUyOb|px@90Sx1B$=t!pBxf;myP1s_6lPjH6>BZFZUx%CK_bm}^7=&Va8c2QJYICw=tj -?MRN#-K}c06vJUx9{atqA5K7Fk1ttc#vzKdoC3O!$4Cj_O&5-}l?{O8QR+^vx|sh3E>(o46o9o1p(GP -7M0A50@Nj}RZz(*i*{pIHsJ5_$qP9WA1rDH|PJ}dQ3* -1Ln;l00?Qz});q*b0w?R~VB^^1h=TvNH4~l+7kH;$7irq!Iu;1u`IDam?q?jP_ -VGE@e@P@H<#HIgi)1K4w1svm)Oj7oD^85*Sg0pr#?T$gLKY`$PUiss-u79XmHRa>c3eyW|c9 -4TZMKE_sA)g6!r{RcMO3eiGw(#x&7j84}8BHPNEi$qn_$Yj(LrN}zUwAt&g2+Bi^KkM=197ri!?3WA} -5-!%#qz4I`_=ID)2}~qPGTt`43co+6d{rdHK&Wr9br)gVjVoYB0FTH!IJ#JKRxhLiZ^Sr$LMx(|x!ir -UT9ibEP?@)?@Pv0iRX)#F64xq?UK -KDhiGsq~WH)f!dURandB-i|WuTGo8S!#yQcuwc)3e5S#8KQ%*P|I7l6NcZ2ZBDE*-@li%IS5_*pE;oA -F*N2Mk0aLYzccyQReD=cACOR{5l;wX5ep9p3;DXkK3b7#Rkf(6ygb#Q8AK3>2jZ(giifDB7QpBwl=W0 -FV{qPi}dqA^yOUqm1q@h`zjqjGAnUE`r;a5^=va!q*;l|3RO}R1`-4z(3x2r^8%1VoDvLgrv_Lih997 ->YxbJBMsNBT?0JV1)5jNJJ$6lvlR3z4IJ{`B!Y2)4_h@PY@{?*F`ajw*c3W719NSSGpCDa^^WC_@T38 -W*y6YqYJ$(b6eW`W)O&}I!-&WH6gvCsjPj)<0&KX<0gZ@}J=&dBs{Xg)|1(vDZKnaGHV~7+93Oe5qU5 -cGQYvDc6bqL$YN5w2JD#U%XX2ur>HH24Au+q*0&PcqsU4F78x08JhFfTY1RHc8o&6K^?ix1dl$#^4KLr8?o64kmEQaMu -Y!1sUlC?W?#aeL9!~hLTr#6T3e;Oms``JC-U_@V%N8^&R)n%f6@_=wkCZ^J>PG@tr|xp&?>2K>JI7e< -e-x4!*M6^9hcF#=c?G#4{kD*rap@LcR$pz6hdepbWd^OyGK`SPY{YtfCQ@| -Z!TJR?hc^Q@!^V%(}p=9D$1~29jg%C6~SVoh&NRQncxr|a=L}H`4Pk!eU2fO{xWm(=ow|Oa*bg^iz-9 -MBy;+VvqLUi{v@ZRdTO-hl0Yl@KQw;i>e=%4Ve#r_hU9x1*xrF -cVvf&;x2c0@=RstnSbKXKs05wJie-J)G^(l5O#GJSAS-WzuK}M&fqjsvC(uXBETS1Jl>9S;rvoUV+l(_*GS|5Rp2|12I<>NlT(6jNh2G$dmD;0BVFG~;&^rqPZS#T-c}TwLX -+DRBnTUmwXhI)0z6&G#IyH0QZzd1El}+P=P{Du!~WqIv{B<2DY;Fx>gQ31+aoII4nCoh&MrmaH(R= -%CT(D4@QxY>46aB`2Tf{E=Cd)IqTfO_I8y?=rT#n@*V=LGM3=U$cn|&wyBtyyBvn@j-@tO`O)XZ20rn -y{FL?@9Wru~_cUjd7(v^k=$1Q64G>tDO&b_3$79f9ibuSk8Gx7b^{o+0v_xrX^Jsv&LEzX%7{}3G`ValZiwtD#KM>li-$Ki6|BB|u -~@#oTP**2^W)9(p%Opd+`ak%T3&s>c(Zu;i9_i1;$eAn?_VFd3Nqea-N9k+-dx>@x9{%W-rnE%YU^(c -JGxnb{yS`{`PPANW_(vT4HtOMnw`}|&-uVDqKh8k_{)e}%uD^W(K&D;I&Oq(?HJjesdZ|TJ*bC!R-VR -|7p=I47IDjg-av4CJAOVtJbF&b2MuMS%R0>)hlv2YR&M2bI>V>c@s -)u5a)8?@x?3-9pOH`?w*D%C2#lZU(uG_Q8WEaiZqS>?)z;scQx#9Onr@Ac8(p^Mm@sDa-{?L#v%uAIV -sP=%JMpL$9lJ9=VAsYpd)<|@!w)FRX19|xB_vKTlH#q@-DrI^l3hFYncQ#@Jo6?9-2zbXmiF))D~G_L -I3$98JctW2(lA{%kD4z@+wl4ZG9VZszT6PR`waMIF0N{-31o$cytMOGF4RR*yx$OYpQh{aW#LD!+ae3 -CH^%rWC{H>km+d@r1#fE=;|ZXSHI#j={2_LJ3>whyz;UOxJ-u3z2V&l=UDREeRTrjE%v)X(>=5(yS1m -XXuFEU&4qo_h_Bf4_SB7LP3c`7ODSrv*UNLwJ3me07h0V^#L)E(w1RpSYk3uTe~= -36WMCWQ|1|KE$a;l{@)XtFCbZ1p>+LL4QTJlnRkK`{mbJ0zRHhglcz)!NjNaki%@Rry#l}qEbJUzQ2D3BoHU|1KkMTLQAUi_w_Q -v#cd;19tJRgtudMJ5Y1G8g!Wop~!8KvJ&$f#f91zRkTgZlG!37sd?8d^&;#k6%ggUbMC?Zh2z>bp% -j9Ha7Xw)iG2DOamD4xi?#ug|7f_j(d-10M!Ek03HAU0B~t=FJE? -LZe(wAFJx(RbaHPmUtei%X>?y-E^v8ulfh2JAP|P{eG0NCO&Z_8gQgyuRGVx)Y0QKQo1hdnK#h-YyDT -uJxR)~DPiLTC_GI)#?wGL7XdWcj`H$LEKxeS2!4X@}Qi4{cNDNjE%}a6d%a2{r#VYkx7$PH -5#Zzz;WeDFHkfXCWD=c^bL#G(ncD)!Gm_se!F#e{8_Ns)ST)DL(4rbu)gMq -x0|XQR000O8%K%wh9aOw&JOBUyKmY&$9smFUaA|NaUv_0~WN&gWWNCABa&IqRUu|J&ZeL$6aCu8B%Fk -8MOUW!QDau#K%q_?-DpANy%*^BB%1l#;kIzfYO^uINu(efCijN10$HyyKaVY=+h@+v&1prV>0|XQR00 -0O8%K%whN;++8E-e567Qp}j8~^|SaA|NaUv_0~WN&gWWNCABa&IqWX>)XPZ!U0o?R{%^+cvi7cmE0ub -MLC;%FLrPZLg+o$4%T;tEru3J8kDC8JUtGn>9u1kW?JCr~m!z2LJ*fD9L%WbIv+itBpwl8yg!N`vqX* --e2yos`CCM&F{muK~Cmt8eb(vmW*W%HGhkrYG^x)A$agvts{#wF!DDgC%%Dj>@vC3z%6!o< -fN6Tai|EMnm@uMuOw8+Jy_<`tQfk=Id`hN!q>tZDq$y($^EmoBT5NRdmX(okyGnLC)q`8 -637J2$T!Op@dqSRB%z!kz0Rd{5SSN?P_>lPbe>M3U-{K4xsu|lcrD8ux+9jdT%;8aWCiHW07{lF(mJW ->O{-J!-d+?%dl&HL@{)R2TwV$w_zi;XS?{VEz*U69(==bb5l=KN0CHa}%i?u9lht0DLy-lwCz46As#S -AXWoA_oX)VeX;8f%=2CM0{NC0CPL0}*lv%LijdYa3FQYJGRM8-snfNFrlYsIZrbI_MjjiZDOmJ*n8A) -&uB_0%|}S?^eWAhY;U++59h0VI=qCEHA0Lt#CU9Dd?~O9{xuNr20Ng8Ef&~;MFCu-% -OqJ^uVZlpRLE)=96<54q|qfrm_QRbO%_t*atfoLlxuo{vx`yO`y~FGm`UJ`X@V@rbIu?E1ELCwPvXy^ -n+srr8R7*^7FQBkD;A@=+5-qPY6E(&>x)3AvAdVz7DpmO!msVts}v^oo@($>{3*3>8qFG$J29^^>k@6}RQxmYjZum0X%U9N}pCV&hUnqi}RO307s{c%|qC0DxF4EtvWE>#hSFs4v*Z#-6$bv%Z5(Zl%v -#Seh!d($j|(N}5*YxGRrJ_BDQzd$=6dP;swbNg|!O0(HGUo9q5z11*BFv3BA>8z=)kZQdc0}s^I&_;| -tXX%Y>3Qvof%wg{8hR4rOMn8j;|K`c)40>5qF=*(t1QvsN*o&;M(LhA^Wj&4hTNpU9RI0drwR~&3o_M;~e=w -7AWVA{_?_|mj);2|{B_|N1pkjc?+;lhd;wj-HPG|M3$f=I9BK?>8FV -sK}xr{J%`5nFP86o1fC(tjJWEC1_9}!U15uq~huYS2&S~wPLCY&74#Ce!eilSd$nPBZ%&I9z?5|?luA=$zw>v`5{K~>G{{rP=d0R<+4mG+5DAM9m9KUiYS -xiuiZc@S;FXU;uG;Vt_|Av8v12w-WKIe#U=n&FRx|BmCwFAekwj4h|k1NO_4B+ru9(ubNU*tiB>wWNe -&V$R2{qq`hEi6@5GVDK!fRBrf3wfM!Du$El^w{G?)phs6FZ5hfwKi6`M>#jVgK4DQ-Rle?A$VU!0vfE -XpGk8jSc2jAs0X#z8!TDGrt-(P>)bIHVoK6unZSb-b5i$8vOkR1F9C1!p4B&!4vDaAf~k3a5 -5XkxgiPYjlNYyv%Kp{-S17+t{h~ZoZTJOv0wf_#uKdOMdhkUaj|D=b_1Zx`UtrLD3kqs -!!`)gHZ1Ux&-X`(S6_<1b0sYMF(j|CU+4^|qMato@f&)_wyHD*EgPYK-B-tL`yQuRv76Am3igT!8Q|( -X+gw0pEr5M&H8hb%&a#M>La@G|9&e67&}IV2cKF;k^o3Z@PEus&18!z;Lf_{FfN(SR`qSK$9XVXZ}k$ -ePp2ub7Bw7FwypNd-GM6Id8^n=}w9wgU(-+E7=zZU$Qsx0*VcU{@*LMv(c-1X&4Qm*tTK$hWauv&U@Hu7Q49fv0w}9ViY_=5PZuo5wcp>gzl+-sNW|h9^9O3DmRc -Hg^heS9lKf=%6Vem9mcu|Npw9@v8JSZe96y<^$`ZxB?aA@)FeIXfF1{tgw%xGEXwpkQ{8ORvGG|y$~IMM6o!~=wwV<@H4&wEmGBIEICYjbW<2VL -MJf_1{DaC{3lAU-x(W586mU+MXaE~8mizm)<~vhQDH -zXtJ7ts`lB&P*_Ah8yIvEZO2MAPAQuMm>?SG@VsT#)=m;9acr6$NRgo(V!nhK6jy=K%*W9U~8OyN@GG -&boEi2E2Mg-8VF07cSXvIXyy5!&u#-5%;OE_AjP=fcfH~d -uSRRdLgV`2Foz(=deDz8_OX1(Y1Xq%bl(-V4B1~dAL!u-zU>^q>hAH}ep2vhnoJf5E0o`m`I4LxdGTU -J&J=g-fNRdB;<#SS-^C>U43N35hAp2e@Z1?^~&{0ef^LK2O0y8t}epDE^B7DdfvzSfVbhgq?j(QKs8k -MWUl)g6^@KzvUM-NG?H+AXmAWKj54<`;!iAe}76OAHa*yk0=D(bqqSXZZO2k3Vn$cn@U^1b9!<84La@ -&dIp1l&64D7r@P2m{I)!o^7Aa0Q!L*T}Ld^pZL+n6SaGm$xMyjGySN(rp|~zOPdidlWJ)(+?jf`y*X3 -QiFWmH!gmDH`ih&TDbN=eFA2oCd9*P%?j1LCt{?g98qYCK@7Qmeg1^M<+8Rui4$ -3yPxDer>)SW*Xud7dS&obA>iZAs(R?STwc*(_d~{ekePeO7I2!$|IBn&8_!b@&d$>UZRp#jFg$4EPr4 -271&1pgZI(aQoFQ{elE2-2=!3(`cqK=+`nwiZgVh`oWbOmFQbt1=25ZpY~5~;q{iUQGN~%*>wR -y~m?6T3?D25uk)T3x7EH_LmJ8Is;-vFyZpL8I=4$igBm -gbhX$|ju-t0=B-IUeGO_=RS_@t$zzwn1rolYKrAJgKHnlD8KNl_PE!+sv-)c^LDM3BZDg=kqPTz+DjF -8oW-6Rh6wRo>b7x%n!xik=AF_75= -D!zlQ%9csg#Lr6RfOX+wBEk{d3CBzM|w-MwaiVt;Y^dKOpF+#kvqz{957g_I1`)~_iAYH$V02PT5fl -hfJcaQ-&Vj~rvogH6X{B!(1h#{+KnbO9YKa&yz!{P?KiT*hdnM!7+^rA2HhLv;%?I6!ZU+B5>WfLSJ1 -gic<6u^En+`Cb1S+YutFCA|QNid8MR`ngFnh`7?R5&G0qTm`s0pQFm@DPIC~0m2s?B#DBh~A%fxUP!ILry -?z-GnYHr`E;6}R@z(>Fj*Z~%)E1AZhhn;=k1IgtWQ1rB?N&*Mj=@|6w%>F< -z+7$ip-)L^i`m*PWMK{feHoMt%wn7!K0!=vbSXRO&UoNb9ibZ!JA4^8z#816Eq9CsXVP|EK -1NUDp17%FBeVjY7_D3TT%IaoJadj#9lW-u~snXa615iXG3TS9Zdkr+$>QrqBdcN4I?vXcXF^svviolT -g2V&&=yUMUJ{RBEL}51GsEIQv(4l@jAN-8T4jxI1HXdNSV4zw3f>rq+Pp&abney7ZxavKDDPW#lM5<|W{Kv#7Sg4cxuOjh29J$@dj -N^rp35_f&(5vws6lYLs_FYP(TkIX1x!b9ci~`yvgV9g5r4(yW&~4bB^1dmS2k`(po{Rd`_ans)pSc&G -2c+2Om(>NVBgx>ii3#(@=MYIS>cEETG5W4jfKQ`lLgnKdk=>E}y-puRd95?%!-XnM2UcBZffUbBQlqb -`=%WW*7GyP0yVUs(jeEKu57^}yRd;Amu9^b^tD#`UT!iXnn!2d8F5eNaiHECyk3Cyq{~L};i}KCBwi~&5 -l^uw`%!V#!}Eqzuuv~XO|zns*&4O64)ku)cPYj`uaa^`L0DeT#zRF23jKE&#VerCg2qY -j6;d0c(db=fzVBzX)-0e{n`08*70rjw&`jBG8nT`cL$mp;yUpSMxY?nBM{4WRhCUSwb5I?x8DV%|}Vq+ -RswTQ-a{-#A{3{FWKb5JLiWuZ;j;|(}fc9xR{IaT=eQ2aBRM%=+X{Pc -jXkp0?wHxIzk}UrNlQ1FQdC-|kEskvN|cf&b`lPo*s_hnxe3SFfwvD>1uPTrU=8oLrWDT%9f|Ukmpz} -hm?Juxj%V#XDC!iwGVL0(8dT*>5B$y)R3I))AYZ`9oxZSz7EJG`{XG~ejda%v8hHk2zFtfHlxr?B(V-K6T;+c`}mfknmUS -5hGyAJ&!o*C9KeWaX&SPi76?vJ)9?$_4qy4*kc2C!6|rj8$3-kDq0ct=#}50<~~z=j0z2OIis*~r%aU -1h`nR4(`vbV!&hJlXic7$bBl_kTI!&ryJ3PGDv6zYOrM4AA`U#tK{Yczf8~JWmvwD+=HgmFj7VN%m&r -Wmzm`S+BPzeEI@2eW@roCSCFX46$gvUIJV2`Fqo2!28Hvg8GN9!T|#YR5h`$Scp^Hc#O%k1{nOCCRtW -ltNg+a9GI9RO36yr79b2WT#mmd(NN+ce8m=7$4pP3_!ZG-JCbO&AHM#Te|T6oB -_Cnau}cJ~a8eUc5NU*RKYc*y|O|k&_~qLx&5nPCTAdgz`|ppLAw$JPsAY0)Eq1vT3=P{4%5@VxqXsF+ -CP$!g^sEhd(L_Gh!aA$uF@c^b4!*EBK}!dr+qDS -4BRCe@r4T+6QHyv~$JEGkuP?eP}-E5D3mmLOwo%f%hdXlHd-c*(#Fa)dLlYD?+YO~~KP_= -mv8>&mfuNVhFheb&jf5gVn#wB6W8W-8;CX~>Jy6FkklC69h`5(c9qLj0`7Qh^cObay-xar*>xOn|S82 -UcsV;BnK!)BIXOajC%#_KiqIjU)i8B&7y}n07Ht1?v@4sVQdE+Q8C~{`bCVK7fw56z(N#Gee%b@q5ru -pmQM(wmDcmXr8Q3RSF8xdsY7D)v2l;+cHHIrk2HZ*wyUtnw>CnwE%Q`T}55Is`xm}1U4KJ8fxUBz^(E -6yacrCmJH3DTpq3ZjGEDyN|q6&=#v2sa>fwHRT!ApJR`j_AANewE+p-bPF?9pxbgV^_Gcy=1;4}nbUiK}Ze%vUWmRW^AinTY -~g~k;Q1DHpTc^1;7MDjVr5?dS`T8$>CDyx8nriGGOFAV|HOPl_JgIaJ0SsT}+%N)Cy -S*6J^H5E@8Z>=sfs@1&51m$_ZCWZVZqW3_NuKBih -gK(>{B$lEqw6fK#s*b@Msbm5+A~*;RuMWI}J(L0sCOXl9)l^^i?9su5cIe^}>Q?waR0&)l1wmKEYt4H -!OVkH!n1T_80}KE44WgbtG5{L?T!lpTPh`(;3qbcRW2*?Gab3U*RN@Ny_duFqsid0U*f^S524nZUPIre)HSsal7HlJ0Y+9)nFCbX$f*n2t|kS*^k8E -eTlJ?~K-N?By{KXw?f91>&3n4=Ktz;79%Yb$B{S%)oxs)wxvI)>mQEDdo;@6$jhKA%9Sxc6^afhLp># -TLg%H}oOs@oW*m+~BNID{K15iX7hhSL!u?ac|re1X;k(y9jz_w!eK9$35Io0}m&f`6Ca(Z#h@*$ncM0 -pb+Dv-6*iNgJTus!zoc@J=kK0)mP;U5OFFm|0{DqOTGkDiUhgp(=FntYR#lFwa`pn>C_rWJO**FovNp -hV!cU%W+{4wAaG7=|g@9)=EVT;6DL%Lag5=(!UpV2IEy7wxw|l_lki^8|GW(KsQCev77hR{d? -I|gr0rcf<3LwlBWKpvT`mhBfxjd_GFQP_({thYxm8W8|G`f4sVapOf41T%q6yC*=O1ovdhS=0th2O%( -7Io{2KJBw|#6nRs{84SD6zBIGHA>Hi9hO{ik@F_spy)DhI)H+SLVP)CO{TJ`4acz!tHE+z7dHdDfQLt -jmc$ir&GPx_gG?-Czrq^(mr9Q2V<7uluwanup9<^3_dhXyy-poAwEi{w=QvN;=Ud6!lRFwl?LI5o8Q5 -CH>n<~(ldsXP&rb=p@ylU!gsG=S?RYO09ijKW~m#T{YI#|`;VW{i+?}L>Co9cA^z^5RHW_y~h*}UoKE -0N6WX!8y#^`n?&4Kef~adn9A0|{Q;)fKTr%JgOxqWdT{VQ$TpjXA;nl{oFtHtDns95>bO4oRegxuctP -p?NreH2zv=zqTcLzYJT+K`}w2sB`(wJE+cLji(PL25v(zDS`vv{K>!rQFcl;M(p30n$|)Qz)Iqls>Q5kJa6Ky-48oRpT -hQ;p$&jR1Isdz?Y4aRPdi?sbY?Y?&YK#;P{RQN41|9x)TZ`Zd*&iDXhk`GzsT^#Mw_q+ZQXj$6f -w8gY{J9Zg7gP8ud3RM;4~juGH*|;r>BPgyCcEg`1Phx=W>1C!IHGn*47T -n)B#4-jMa1D*by)eQF2Lo*Wv5?)kAjZ5vTMJZHS*67Of)G)LJ&t(!~s^d(IC7iO_0RZ$eO0NOK1Jq1e -#**`WC+(ZW*q1}ZQWwq$&4Cf^5DNoFS*!O1_?dYp4@y5XZ98)7nn-e)B4UNAGJuRk) -@8RMIO}f_Kh&hQD{wf0XnLi)i#TAeFrH&YF?Cvspimd0(buhhWbS`TVjy;7x_#IDi7Ul5g5qFwQ>izl -^`4f4?$+csc>*AMNiybb-b6zGt=EvfEV6HXC+lje}NHt@x@eR?Di_wzd?b*g)81@<6c}<<*KvqYWGb% -gx`lk;8T|sN8I3;UrOko$fpEj-PfQgox^=%`|qtxEBvzSsiy0WW9UGoqS>Mwc}oXgtzJVL#yLX7OHpd -_oi%MZ`WThWy9Na_t7gm9(p6YqgtUZZA_&_@Iu1!E@8!+F1On5SRh<9#;UAU?Ky08%ALAtG7N=<|L*; -JVe4Y!-D*1cp%8ffZ!h@{Tt}*!!K}nbw^1!t^v0#0kbxUqm -K$+8FcdklFWgZJjXlpzZ92G_l!Mjg3)`&GmdQ%8Klw$^RBkw3{}`AKJU5PqNnn+-TgL84kg2o$<^9v2HjfTBg&LR%dvC6Y -KA-9W&OZUm15*kDEWjxbnl*oqOl6;c*$Mz%e=Ml^SZo)T=#{IrKZ{EyOQHeGM5XCWT&48gI`#rokp8@Hb%SB;P019`pmby=*gt}Vdod|fdAxbr*4tB|k` -pF5o3Y1$Q^`_OR>O3Lv3;+uodd5kDmM=jg1TFcfSivMu7VTc?CPru6K8;WOF*YHSA>*vy4_M>3$sSIUc -2%NQsHSVH_xBhzAEnBi)gtHkV`WjsI8`l$`|(H9WOYWl41{WHn8exV;baAxx#eWN1}itg<>ippjP>eP -WiI1s*pY;O(p_UB;EmTcV-|2S<9j#FDa}#(kA2L27#Os^hJYDXt>=MA@gz#OemsD!t07YeRf{NT{^RE -6q@;%S_X<$=D)#oRyt+(BnWfEZ_Iwj406(a_{ELh2>6QD2W|PPcLhYe9t@>@GJ@p66IY@!-Sn+|ocF!Yy230XYVGBKq2(W5sp7Nqy)t^)>%Dy -PasOriKL;%6&7&@ASs_vYqTaQ(p1z{qxG#g0ZK`yV*b< -+7zjq@1%A96C7UFp&b=s2Fq?2?H;FL}s!pLBH{VU#w+nfxq%+-ZO<%FL-XtScZ;eezxVo}9(0CKIbU7 -nx_sB3+DMpN5I|IiEIZlH9FY*wHew{t|R$ttf`JCOAp>6TCyzSp>(Q`23~`2n8Yo+af>rS2#4(m7#VF -V1i$#B*w*+7)FVqgSnMkyLP-DfezsvK?%>?hG$~j4_7{ -XBd}1w9ZP!K=|p+n7iBqKuAv*@AA+p^W^x)ljl#LHo)$Q$BI^nKPAmTSz;5@+$kY(0j1$rURdd=bF>2 -+6--GdXV3VFKKtjvTBOh$i*)w3y1)hc9(Po2TOuvmYnh+6>1wl?HZ#?~{T+= -?p@FvsNyPK>5=%KA<^!uM;wW0}0-2hR^6mlB9c}=Eg4$2t~u3q#qkMuUUXxO$aU@B+eU){X!L+sU)w+IW}18Um@;y1-0e)%x0sJ}_I|r*A8gQ;4w!kWGyUINvtcJ`e3x)=NVcuoo15<-- -KaLnwX*t9f=!7hC7u2t>BD60kUJZq99yH!p)t@Y=|WxGw2;4*#F2s=-G$KaqzAf!NB%G{&u^oryUVnL -qAjvCp_SGg-f+@CavO{Kb_^z|rfKRe?$ZS_%g!*&d>P|9Jn|YIZPGpVi2Y5PNw9jqL3Q2D-+HsjMvdp -PtkN*&%Jto@>l9-KwCVZd?XOUs&VotA -%8^;$_0_)faYJ}i!!lnxDx8wmzbW?-Ss=~c&V&I#$vgtsGqgiH7zHCr|FW;d``H3h@EPUq}5JKL9985 -7v|J5*gzr@ytuasxP?(Y0JXU3Rhlz|VV;A@QO&3#_4D_&L9r-roRwfD|XuT}P3mFs3n+vz>qN*~^X?i#HfJZjQ})&n2h2{3`zD!ig -azJ9eMgnN`$rRkwUoorN~S9!2;-2Dh4h$D<>+;TeQ5bhVGO_$LQiD% -u{S_DL+iMDQaL#|$$?I%5n=gD!UzdpmrFKRl75>#74Ges@RTo}8XFy*>}Zq0of8_OrpnFr;cD^{Vyep -7;rKd6zZy6g&u-)CuR`q4O&V9+%<7)iwPeBexXBUS@5I-P2rZvaS%D{$0rUDi)7n;+~!l#A%sKaq;OG -?~|odDJY2oz${U93A_tDr$d96s37ZNsT|m}#H8|VXUYtIj9v89*Cvaf+b12rNY_Xuz5wsIr*}Cp3NdxVkc$^ih8C&;Xh|h+2`!nb0lzE9)gwV*6Psj!$nCY>N -Of*FRlxrmNNX{oxb&SIb+iXSG2HFaqhX)U>#c@~JTEUI36Pw6(SKvPa5ztVpXl2w%b?AY$q8`p+&fX^q89x%^mUa!ir7Q5Q7P@IDF$`g$@cWUZVOQ+n6aS5)<_f^h|(F+$zb<)0jX{rp>}NHUx8d9~6MI__A -kZp*W26{>g$(1-$7@tz|GA#9*wq*fcTa=-}X&s>o^e;}`h%!v-{C%j4R5G7wT@9^Zd$^K^*Wu%|^38I_0oRurOr>a8#4| -$4XQJDR+~P|jux$*Pt;^{q%GXgWA#1MlCFyqDaDkdEMX|HJ|3fd8jn?MorTrje*sWS0|XQR000O8%K% -wh000000ssI2000008vpc!Jc4cm4Z*nhbWNu+EOi4pUPE$oLZDh4l&2rl|5WeFnc5;em7$vcLYEKXeN -tjZk0)SR@q)8~o3`uwh$R59acOkh}>T@qP#NGY=c9*(3uB -#B=wxO>WzQ^JWV;k-4+9A9Lowh@$*9Oqk-2y7>(iByvcC;ji-v5e*M -?dt~F!}eUl)sCClw2Paa=*fZrE5VK9_-ZPS4)S1T1Fo&oa4cJLrjO~21hIqP6{OWV|<}&x!e(iC)?2> -;hdxMn(2h5^D%{Rc40PBgRWO*&y@gt?^Jy+3sJ!G0A?c~uXhI}*Bb+yx-5j!z^ekkRz}HXV``<@C$4e -wQ!4Q40OIFP39X@4Ah)|xA{e48UKuHL2C~Y_rjXS-0|bDoHmsCeX|T=ISt9*72K5TILggF$$(OQ}=8* -VS(p1E?sukm~F?$j3DksWk7FuS^n;lon4alYmcxLP)fcm1u-R;g; -XLpWas^rV{)jk%$VMZT;c=65NL+0Lw3Q`;zxr`ScHOa`5L`_{ul{dB4HE>?j&!N*pg<@6aWAK2ms3fSz7}a@w*=Z003kI000~S003}la4% -nWWo~3|axZCQZecH9UukY>bYEXCaCvo+u};J=42JhUMbrr-#Dv%w;9y|k1rV~*)NRA1c9o_lczdFx=_ -PliQ~dp@WBKbWkK$gjMWM%utrItu@ZbCE$McJB^_#V9?>3Z|bdpB52X1jUnam>mABp)|5V6VGZub -%e%0k=MiTG#~*VbEru>Ee$9h8%=2Q?5ZPmDwZuOKSVquozF}5%i`vTN3b}`DS|M(e<%$7y~8+Xp%WOS --VHgQoz>KD!EM!rE&Fv?Nj=VK)f))irKa`bB-K&&2(5H7-32(Ioi!6cRRKt20K}8{4A0hCTE|K1!e9< -daN+AezIse?G?}Wi%aUw1OML0GOm{karwK%>sfu1xb+yhEEJe>udP1f4UCJfR)t5;j-$!57A5cpJ1QY --O00;of09jign_y2m0{{SC4FCWc0001RX>c!Jc4cm4Z*nhbWNu+EV{c?-V=i!cFFk?wQMOFIxcir%@CYoN7(p+Rzi9K<*CdVMOpLp$ZNY4ZOlAsG)~ -M8w?C=eA*q*XN(_Km^dZ-uYF###f<@-5hl!~w`Lz!1x50#RchJxQIu(DOOWiI3*FSVE>L=9vZSKqwyy -DypYd*9NwuOT@rF~(tK&^b5T0MjTz7IB=BNh=@7A!VzUW`u+Ae@2^{_-6EK60*IsDz0`dw`nM(qQ8m<@|;;0%e(AVs)KS -BpAktSppI84RI8u{v?E3^1m&~&VcKNTk@>6tJ!x8q>~K+TOOiiV$_wllGC~dQ6?U=6ZcNjVNa~}?LeK -+y(u|}#;Eqjd2?;#JLH|jaITqIlJh;p_me?Vb9W~~#`#wiPCO$ed%dDWWA3^F=@1XP4l<`U1_j4(7Ze -1E5hxhJSoiaJ4=8x@7@IgC*j8Yg32mq-Jkn-1$+U05(y^n+2qH1n1!HmgK(O){sPaF6%A;dS^H(q>-0 -QHuJTz(~t$7a`?E>-1D4?q^iW$TUw}WP3rOvyoa2sk=8K`~)TK-1Wv-+wQYBW$)Eaz1t3cHdgg08>;i -^G!jo<8d?#;2l^Vc_kMqJjeyRtoxjGI`a3l#ue76cBaJLu1KC!5W*|SGFPfb!h7M^E3~;|Ay@$a>EY) -BDD9ng;}qurrS%{sJOS(6AwV$ifdQeYuA-&w?#t{<*GmJw!(|J=KFFexMJFNe$bLM`UOx+0|XQR000O -8%K%why<-UJ#{d8TJOKaz8UO$QaA|NaUv_0~WN&gWX=H9;FJo_QaA9;VaCudaJ&VII42F0A3ehQnTq& -K73+l#}{RxUTaqB -pWdarXtt#XBMtw!aHbv#_6qXOaUX3qC1MP6r1+YHL&09KEW|&AXd1l5A*!P2}xMF=ddZw4Nyx11QY-O -00;of09jkUIKChD3jhGOGXMY>0001RX>c!Jc4cm4Z*nhbWNu+EV{dY0E^v9(8flN)Huk%J1*?9s5Zmx -v?q-lFkj$}7cQPrGNl>g~7_>~=%E*#dl;+r=|9#&hDT$;gAK9ea6_9vDk?%f9hU+xRaxx1!3-V4V-!f -MIRm5SC%-Njgw6n;PHOZgSFkTY%`E8nqNlc?2`4Gb&9XV%Eeh%j`t(OkMIL~OjWW4%b4orh2&S?k$ru -Sh+*AxcUvoHo!(;%XplQZ(XGo=xov1poRY!N=f(9C;``+cu7Rb)=*NzSJkgHIV70EjewEgsn -8p*@D<+)D6;8v^t;kf5u;l{~qi`GxR~Vs2i*@SmOI)TDZWOXJjZ>sUf6HUmVjltJc*s@yRH78Tm+~LK -;`s`w-v7$wN$n70qanvkckOpm`E=#_}SI$$iOH$frEW!Za8B@I31g`7T}rbF(CgPW8nNU+X5>9Et2zZ -aR=a_*ni<{1RVU#7{_4mum=~a>fk|d?5S)9Il-2RS*gg6Rs|04^WJ?-iK#jPKU?8|KSb5v(^+QmurRE -r_%Ey7BBNvSG=6cm@^iH^bq32;vBRC3*>B{{7gniwLTQyD;ja86LOms%s>oq(R{XG?h%k~7~xEaL8pR -@l{Jkg*RxG%iVwfli6gDA;BM`M91FT#CvFQ4dU@Rh85kX+)tE7m)-{a_s%ow -3*8bMG&GNl}dG9FIe-Coyb%LODPKq{(4C5ACnhVw_DCq?sSaj|9@#1Tnrj~r;sR)NlPNh1WJ)#@&&Ej -oi>VmjTAP}#zW -#zK}GbhD_T%4^S_U0W4;iuTo|qpL9Ext>Qf<*W<%1Qu@T@483cVYUNG9r^-Q$%G*y1B*q@*NFntp0R8 -SMG`x$ilZ?G2$*)0#NX2)bSna3*(Xk}731QK1rR(=vS>Wk^XDE?0R!Px=vCDl~( -vu_XlCaBZvw~XlO#*dMo^1D2PqPAaCh0uVS)_AFXOOD>p5~7HA)kWbz07^I#bU184d(lrxSmh-30v~c -wMa5Bm6L}sUy(41*b=OwmY -{Ji!5&c+G@rj_T^q>{1%!ubxuq>88ouM!F@5ho{GqKxpf?Qn;dAZ&D>5;dV#aTvR>BKznJZo^PU3lIV -Mw@8wYcpd8zZThtGt^Ui6OBM8v95|*N5i&)}CU<(8*1g9cnbfYNzH}qCFSLfgks*3=7mLPT_7s?%w&5 -eRZg-2Y$8#(vPW=l2KpN(z;}nUE3_Ve65;_-A$XpA(5wC8MgD*2no^$#RYUaAJi+VAq5r -rRm*zev=JUPZ+x*odE-5rvP&8L8l@9y6}8Q5ET+=gid^B8KZX=8dpY@^XzSr+k{3dPFZ4YgkWE9J4(X -L3N9Q7EnKPc5GrnG$O&*SZI%!ydO~0VX*tR-w`n0wsY0?gr+ut+B?zEbBH|kEVjxeh>Mvu2!?IxP~lB -Qo6xXZ)7L>&#UUY#0IN_Oy84z5f_6cc#tu$VbYPG<-St?gZ;75r*RW9s=R5c}&tAkN*je*J~UuIROg( -F`9)Be3qXS`;tN#NhMYB>~f~1Prz1EI|xMSI=8XGZR81ahIfBxk$U(kF=4Cw9!|S<_J}4;|&QgcaTN2 -QrpP*L$i!NO`5e}$e|R^X&h90L|uE8o)XJFLGe47Gr22YQ`{O;t*eZfW(n( -SgI2k(e{wf63M6U@JhWw3Oykaq*4Acl^PHocuN`z$otVr~jE}wFKEAwuec??kZ)rHGu)MEDa{l(__VV -}JUmIxh?5TA(#8s1v9OrOsO@A$iZ>^gf-meMj1p9f!T6R^K(~T4l5ozdNvNj(iJ~dfPB5m-Qr@t>-H&$E&9O8HTfI>y{V8V -35k+;*({?pP%>Rh_bzW%#MS+Z8jn1+ujQq&AiaH|x|VvDe0NldWs&->!XJfX7(ySj))i#C8mA#TF-cLnD -k05b9p`mA2x9{HDwhqxv$OMe3>wM4LrY#B^OjCMtZbmxkP}u7&$+n$HQ#B36+L5AZ$v|OcK3vn2rX?hJV%OeTJt>JpXqvG#O}q1C4GFfqIg`(irsDQPwMT%(ytvU3%gjI{$ -M!+xyPoI0V|{=tFD6xYBFd{u>gEIu0w`DUFDOJl+`c~^;We>5Q`bVyeU@=NRcN+ -o)$U5+YWZ|4Cu}BSu(5F&v;W#CO}m|!y3Oms0iFy@_;AMB^hH42)4TQ+Y54dGLrXLK*QLY0J%4U>JBf%Z8&h(Jr06E2~0FTN3K4v0;;1voY=sc#YIUlXy& -8W+~mthSDSI+RaY}~TWgj**1Ds5U2VK#NT)pHLI74t;@EkVw)E^w*p6A1V_|B+M63u`;M*vGH{MQ!`z ->N=7s*B@G;*7~C@gpk_}Z}}QwLSLsI(C`dtexbgyiJ7?#uS)>Gq)5*7p$c)YGf#NcMc0#$h=D2W%P%b -kGWt`#PTE>Lh}kb(R@Ym>>RmKGY9n!jg(zSr&-TpT+dxShrbv%3(t%jguIAwXgn!#DGi!=Tu5oa(T1%*xGMtDD-O3n72I<0~)zs< -5Hzy#GZ=1QY-O00;of09jiGrvN+*IsgFTF#!M^0001RX>c!Jc4cm4Z*nhbWNu+EX=H9; -WMOn+E^v8uom;OZIdYxf=T|h^FESv@n|D%x9(eT?mLCMeZ-dsTEw?qE5oX#KLC}9U_sKZbwZ4)IC?vQ -tPiCyeNHRzUbN`F){_4N~^vCc1+fTp!@wr{DkX*T4SmFTVTUr9X3j_WJ%;fBelKe){{L|9 -GK){MC2&-}nFK4?q6+ci;c{cYpFxeE)m>-%ws3|3Zm>DsVEmD%=g;3ZDnAFM(?`I4PVBE(%wJhr-j~r -SN{>?gj4Qfw#hseXncJ*RWgu!<2g*il_D9*(;vy`i$$xdGM+Cx5v8Uvf{29z7_X&eZ=+Au8+7r+4ZRc -Z-#f^!|(}w8P5AZyiNgL?YSc;fNm^K)y{aRIA^v7J`W^LG2a_4V-f8u-)o8NtjAE<|n2TzG2Ehw*~D{P@>@`0@Y1V^aI5x -L2A|q0ifT;VpMh`*E10Z^s$#t9Zd{SUW~{kwn8*lxxit>V{tjCJ~w1nUgmh6(P -M;o5cubD|KmZNE+3qv|R85}x87ylGqA?VY;E$P>vbXxRU3|fp@Oj^uZELyBuY+CGE99o=O1f#9RBS -oito$__c*C}78e4X-j%GW7hr+l6Ab;{Q%U#EPX@^#ABDPOOAz4GPU -$1<<^7YEsE8n1egYpf^Hz?nre1q~0$~P$ApnQY!4azqt-=KVh@(s#2DBq}jqw$~P+C -sC=XHjmkGF->7_}@{P(jD&M4hlk!ciJI^3BRO3*T6S7NZuE7PA(M7ONJ8!i@*D=(QNM7`2$Rn6+56ShWZom9NGssVm>Ee24NK%6BN=p?rt(9m;no-= -Tbm@*T=|DBq!chw>fDcPPQBjHWW0N?|I6DK=+~g0eC~FjyH44fO&#hW)TI^aBaq>Ho` -HXPi#kZ0EmtxRj)MC{=lTlH+yvt7>)Elw>iEp9Cyx${Tp@>S7kQTe_@m#=R14qd)3CM_!8cj)qU -QTe_@m#>S;Hz?nre1q~0%J&_*d{umhF1h?tbb8inQTayY8fS}&yC}h;1go}JRo%O)dslVus_tFYy{o!+Rrjvy-c_5g+O;^eIJ -LO6xV5OwSJURJX@xee(54mIw7G6tzfJ46Y5g|k+mvt9`fXajP3yO5{Wh)NruEyke!JFh*ZS>Rzg_FMY -yEbu->&uB)t|fib637y`F7>om2X$RUHJ~>JCyHGzC-yAQ~i0WKTq}Nss23GpQrls)c1Gl`#Y8I)c1EP->L8K()wM>cj^1PlVmi9PX`;@Ic?A9K3YY)4%huv!PR-3mPxz)(6MsDpix3;2NYvHIrANA*>^?Sr)i%xYGCk| -XR$S+PDxVWIjsKunW%33U1tXgbZ?0V;+#i_-m#jUqe`Z_=;-f2Ug2(c%kT2XVIz7qEnqkr#g#HbrzlKEIQR$bgHxHRAMT0dS#+wi=u~IXs -m`KPokgcQi%xYGo$4$))me0^v*=W3(W%a&Q=KJBlUt(JU{sGy>aj^ZHmS!Z_1L5yo77{IdTdgUO;wBf -h9>pcG_iJI^3BROE8nbqv%aBOTcoUfv%aBO`RW*_=oqII<*Q?yQk1W -baY~6=RK7*|7Uf&CMJigqI^-!jriNT^1gGDC>i%tv{ofs@SF<5 -kBu;|2KnM!{ueT|^XRQei0m8tYKf+`w8m8I3a)V-Iw_tNTKTHQ;ld#QUbb?>Eom-1c8cPZbce3$aw%6 -BW@t$erg-O6_>->rPN^4-dJE8nerH8v|6n-z`C%2r3(>PSZ&>1YRZv;#WY0e!@E<(Hy7x9Y5E)mhW3v -!+#NO{*>zSDiJj4iDN|6oOS3i>odcS6wWwx>#Iwb-3#4aMjh}s>{JumxHS=2UlGVuDTptqeQt?XA!H; -B37M6tU8NWbr!M4tamP2ta@?Nvt7>)Elw>iEp9E!OII~!s2Vd=jTx%O3^hrgQZ;6%N%MUZ_e!pr*U$vvJ+R<0-=xbK1v}dnbYavS{Q;WKacE?qx9IJN6RlDP=GmuqhAgj(mR_%_fcE`2A$f3 -ol#ihlqMHwK}s(h>Rt;+Xf2D1E8j9N@ul=ivSrbTuBs?J~4`Kvm=&M{V9JgvHTT6OWX>f&kD#nY;br& -SkEt1g~ao!qNBxmR^^uj=Gp)yci8lY3Ps_o`0rRh`_cI=NT7x=L49>FO$7mDJh1so$9em`|hQE_tL(5Y2Uq~7PTwa*}ST=c~xigs?O$Boz1H{n^$!L7HPm!!sOi*D)2X4R%fU@oc$=>9He -KOuy29IZg}3PnZ_^dtrfFbJ)4*C)Gc@O$hwDcCFB^7235zy -H-lqO6j_S-y`lU6uMSf*G1}Ha8~N5tIoZukt%yv*}KZ#Rrap3cg^kTn%mQ-($_lbT1Q>$sO#I^_3iHZ -c6WWdJL*^8?xEE^w0?)q^bW1=q18R~{T=%L4z19k@6pirXy}VG^u-xkgJZt5C=Z(2d`-=bpUQXYME6w -sQzyEoHbGOpeX6%losORB{8OEOs#a6AnyS@Qt)^-<)%mA-`_xRfIrYvO_Dl$?9y`@zm&(4>&6ch{EUk -s5Zno6Tmb%$eH(M&*QgN1wvs9d=LEzHZY-wz^G&WlrjBQ=?*}AN=by;Tz{B6-<)ne14*u2#Vw>sh0C6 -XQSw?m6li%X00rfkjX+?uz!HE(lk-saZ4&8>NxTk|%z=521x+uZ8jTbtZH^;TNlTVwt`aVw>-G5^t-z -@syPN2~j2{T{8*qkY*aUW-ACQH%PZlC`MMEoyUt8C{D*i&KkBiz-fiX;BGkV}3Nx{Aix}(ZuqjiRGuN -85L$oTX&jTrD^9!)6S2kogYm*Kbm%aH0}Io+WFC`lA}{4N2f}TPL&*;Dmgk;a&)TXXrB4eJoBS@=123 -)kLH;l%`-o`oOEF9FO(dDG0%SlI^b9t`lWS%PvdM@ax6KSr#WKQyyWij%W2d&AoC}ldo<$qMT69D1(pg*l<--uw2zW8?I^+maDpJ!&Obgy5)`X(Q*L)0R92|1NaB<58xl -bKY)J#{{a30`~&z0@DJc0z(0V01pf&B5&R?gNAQo}AHhF@e+2&s{t^5m_($-M;2*(1f`0=41pW#96Zj -|aPvD=xKY@P&{{;RC{1f;m@K4~Mz(0Y12LBBH8T>Q&XYkM9pTR$ae+K^y{u%r;_-F9X;Ge-igMR`40{ -#X33-}lCFW_Imzkq)Mf5a0N@kFio`^vwEeiiyv=vSd%g?<(KRp{5CUxR)P`Zeg+pkIT24f-|kZ{Xj+z -kz=P{|5dI{2TZ;@bBQ?!M}rl2mcQK9sE1^cku7vuVX1!byJCJnkRJ24rPyWL0RLshWN4}-p6XsEo;2e -HDBtME6TWDXEtvc*JFIpHM`}OdB2JBLRsUzj`2j-*~42-D8t?iz8Uss*q>p4hW#1#BR=mKzhL3xmSJC -keFgRv*jHd*fqjMcY5ds-^wcr=EkoY{ei(mrj6eDSKa58@#v>i$k&f|5$9M#5L$?h78StM0{uBHs_)q -Yk;6K5Cfy6U-9=Hf8X)<9eV0K-Yr8v^e>}Bj}ARL^w_Acjr(uhf -AjC5-vNKkwQy52P;YsmyiwNsPt2FifcvBWo0wmj@XrbVoR}||m@k-^FPOTfbjt~4Ja6%N73GF -<@}-v$3&@ZSahUGU!p|6TCkwZR|$yx`9Z{=DGN3;w -*|&kO#%&|a<${u}%^_~U(88~oAUF0{7`?d>|?f50E_+dANX!2f_h-p|GN^ML;W{{#L9{15nRE{uQaeC -sW*ujgZLdE-3(UiVAfGVZ7SmB;#y$NG-P`o=9gls(D;<%lxu%djuQz6|>^?F0V>dl&3ouy?`U1$(vqx -aEv;LAj#bPzL`6{tNsU_%HBZ;J?9tgZ~Eq4gMSaH~4Sx-{8N&e}n%9{|){d{5SY-@ZaFC?T4Q@KY3o; -7r%7;=+Pc~w8tLpu}6FC(H?uW#~$smM|r)-;QyuG59qUsa>r)-;Qy -uG59qUsa>r)-;QyuG59qUsa>r)-;QyuG59qUsa>r)-;QyuG59qUsa>r)-;QyuG59qUsa>r)-;QyuG59 -qUsa>r)-;QyuG59qUsa>r)-;QyuG59qUsa>r)-;QyuG59qUsa>r)-;QyuG59qUsa>r)-;QyuG59qUsa ->r)-;Q~hIos^nkF$TzP@ILF?bj?Fi>jQ1H?=GczYu^*>nLr%wzoQ^Fy9eZ**Hsy5e%IVma)3Gn7V`EO -o&YX^|IlZzy)BdG5v`<(M=~xfB<%V)cd7wN|RzBFZ(y?vjmhrrZ=V1#)!yXFkd%5L}GW5_dA$Dw(>8; -{?<$3k5rt|r?47~^V5AYx0kG&dgfFJy|qrN)otE0X;>Z_x^x{l9!*ZIO*_9(+1oiFTNQ`tNBz+io)V? -Cu~{iI|4q+|W0V|}D!eWYVOqhmdzb8}3```Yooc8xFZ?GwrsWzF{-`c;Ipb$#KM9m*c%fO14Rp`1}JC -|8sl${po_@F!QX?w2Y(O#9{fG{d+_(*@4?@LzXyL0{vQ0b{XR#>Be&d9*7>3L -uP`a#G2{0NSTel;b6&A%dI4s^V%hYCJ#}xu69@znflOc$SOhi!<=nn2T;0SmEfj}f -6-*I>Yg+L?F39PT9r#G;9#R0Hh1RMcRfLs>2=r@DjKqgRNseU==4KxCsz#uRQESkP~#R0Hcx0`zGNPv -#J>A0JgZd$q*0b06IbECr2Q5~=Fydv<5NFWi&yhkBW2{hiL^NK-W@=q48*t~*zH0sf)N2g)E6Bq<&Sf -^nf74@+X@N3up*a@gtcmjbyB#;PX0);>&&Ea~Q4dEw9QAP2! -%+`MJskCL)WcB^M?F0C@YKVj#a*6yXjtBSe|`Z+z!N~DbdZDo -I!HplKqgQKR054aCol+10*k;VfXY6RdPM3GsRvThFF-va>h{D!pb}^VI)Oo85?BN_0lGta5g`3U`ib- -t`lwT=Y2i%^Z(4ZM^36>X0s2^>j}`h@p^p{%SfP&<`dFcl75Z4Ak1-H9m3nZO7xtXKfk*%i${BphpvZ -$F4~je}^2h|J>1e#76QHJpnvTUkp{9eHGNL~dZSt9HwiwYL#_MO&9p2e2UxfUc{bYv1<03tiqtz*qO)|_KCxIuvi1!6@xR+JkQ2uaZwpkjlHO+IDvDU(l` -d&^#uX}R6}HU64{+Zb|;aFMJg8AokV(2WOowTSwwagk)1_k-w+d@!e% -_O8INqnBioP2_9L?Wh-@z6AUWjJUVvVIY4IM&Cv}<4cxDTe@ks@j+2 -mw4Ihk&m>6V#pndx+yK9>0+XAF2n5gWdD*LF)KB_WQsPwT)AFK4SN*}BAu}YV!bSduTbB!%sW5d?iY&GPAUmy|41PTEfwrCjE -a2%Ig9G6=hms>6n2_yoUKp{{GGy>q#S+|{a+gZ1rb=z6DT-3S1ATS9m0;pSj%kct|%s+KUm};>dGx+_2aIF -OC})Cz%%8^~Hv5v0+<$ffnDFjeS8EfE|Zp)5AXdP}9SH?y$)@Y;q2(;ZV~92}4Blc>zb@n?DHzB7sC8 -6DR~KfkvPc7z8GPMPL&+03!8>v^3JvNJ}FvjkGk<(nw1qEseA^(bB}4ORTxXnoF!XL@2pqi`)eq0Z$+ -h(5Ku{(~E#3;0Xi*-`q42P^ZgWpvw3}<^|9gWPB#`4ZyI2$*#eEZm^#l?A8XmwZU#}uv?pKP$nCc3wQ -$G+=JVNF$lOoCQt})i!lZP7w7~Afk}Y-nvvOc0qn#+j(Rxi;i!kB9*%lA>fxw|qaKcWIO^f3hoc^jdN -}IgsE4B-o_cud;i-qG9-ex5>fxz}ryibqcI=?7`u84-P+UaQJ0|!%rI=e%s&#vJGU*&mXXhp -6sG0yXeU-dY(YGfo#J!Y+3t(wI8WRq#luaMCuW#N2DH@lwA&U;}@s|u)~?lJ{O1t5&=H8@UexDE#Ex0 -6Bq<0fkj{w!1i)3VJ$9UUBD9{+e)^TY%AGTvaMuW$+q)~{>>HC10$T};Zhd}1W>mfz1ZSZ*5Xvw;#Ah -+RMz5D*5Xvw;#Ah+RMz5D*5Xvw1qOjhU=i2^;4s}$4@W&5^>Ea~Q4dEw9QAP2!%+`MJskCL)WcB^M?D -<%aMZ(74^KTj_3+fgQx8u)JoWI@!&47WJv{aB)WcH`Pdz;K@YEwvk3c;F^$64>P>(=80`&;w638WxOC -Xm(E|FX!xv)X$Y*6|)4N4>z_EDVJUHF-vH^91$tlLOEBK3&WBT|n@Jreau)FV-kL_HGqNYo=yk3>BZ^ -+?nsQIAAD67@*bBTyk)FV@mOg%F7$kZcK55_S3sNv#Q4HrLaxcFVe#Sa@Uez|b(RhzW -O&c|B)U;E=P6<0D>@=*?^iByoCG3>2Q$ofco$*I!{LvYI^iKMiVcjo;8=Dz-7w`lEfk+?`$OI~ZMxYZ -I1SWw+U=tvHNBWNR9qBvLcckw~-;q9|iwjT>N4Ab^9oagvb!6+w)|0I#IZtw)Ts*mWa`CKN&$`8|(`G -*1Wfxw|qaKcWIO^f3h -oc^jdU)#LsfVW?o_cud;i-qG9-ex5>fxz}ryibqcH%dIahbs7Ih4fqDe$5vWI?9)WrU>Jg|% -pdNvG1nLo}N1z^odPM3GsYj$9k$PZ2v+*;qZ-C7f1~eOA)p!GqKqsKT1@}T=^p1;O1irb#69@z%fkYq -^CP>(=80`&;gBT$b(Jp%Oz)FV)jKs_S$h}0uek4Q -Zt^@!9XQjbVIBK3&WBT|oVoIg^JNIfF;h}0udk3>BZ^+?nsQIAAD67@*bBT^~lsCQ;$qNGWE#RBU6t|Ju>yk)FV@mOg%F7$kd}yk3u~P^(fS%P>(`A3iT+|qfn1RJqqEa~Q4dEw9QAP2!%+`MJskCL -)WcB^M?D<%aMZ(555^yk@rPsl;TV57#vhLHhhzNV7=Jj%ACB>dWBlP5e>lb;j`4?M{NWgXIL052@rPs -l;TV57#vhLHhhzNV7=Jj%ACB>dWBlP5e>lb;j`4?M{NWgXIL052@rPsl;TV57#vhLHhhzNV7=Jj%ACB ->dWBlP5e>lb;j`4?M{NWgXIL052@rPsl;TV57#vhLHhhzNV7=Jj%ACB>dWBlP5e>lb;j`4?M{NWgXIL -052@rPsl;TV57#vhLHhhzNV7=Jj%ACB>dWBlP5e>lb;j`4?M{NWgXIL052@rPsl;TV57#vhLHhhzNV7 -=Jj%ACB>dWBlP5e>lb;j`4?M{NcWdKPvU8)T2_5Nk3l^K^%&G+P>(@92K5-!V^EJlJtp;-)MHYQNj)a@nABrZk4Zfy^_b -LSQjbYJCiR%qV^WVvJr?y?)MHVPMLibvSkz-tk3~He^;pznQIADE7WG)vV^NPqJvQ~&)MHbRO+7aC*w -kZFk4-%`_1M&7Q;$tOHuc!lV^fdMHNJoP)o*_PH$VRU?|=H;&%gf7Z-4mXAAkJgcYpESKYm>J{eS*1{ -XhP@_5GiJ_a`6a=a2n*ef;zLPrpLL1HMB48qMxO%lH~y=Rr^T8a?#`ynpdE>R!%ZuW&FpDx3_?3KxT` -!p-2W@Gy8PybRt7KWv;+yPetX)NW^XJGI-H-A?UxX17zjo!RZwZfABowcDB9PVIJPw^zHp+3nSCZ+3g -N+ne29?e=E3SG&F0?bU8?c6+tk9L=et-I4bL2emtx-9haRW?)bQgBci_!rkDZ@HBWSybV4ANB_VvC># -w=3TK0h!uB>qH87e~QKgC|RaB{>NtINpWKt!SDp?gLt>R<~CRH$*f=R15SrsR(;$(IwwL6*JN$pN%cU -HTtHMs*rh0X4)c4xCYtKHe`&T4lyyR+Jz&F-vrXR|x2-P!CeYIiZai`rex?xJ=Vv%9F>#q2I>cQKFE7 -WR(73|5aV=CMUR_VKm%5KCb*u&RO846JHkHIJ?8vDGZ8Dpj+Wr0Gl2?A}f9-K=(+Rzov7P3d&gyQ|*a -=yavi?U7v{*^N$DI^Fc{YDqVT=*n`KutS9%))WkF3WhZWLz{wO1D8p)RH|iCEp^Rh4zW}!3kq&rCh)ep+*+60>T)ajZFRY|F1NiqTi=~+!ft)%x4rXQ_1> -oUR=c;^eN@uMYwY8rv5d+x8q27Uv>?Z8kmD`L@fzfK3v#YB0$Z!;r?Q++6Xut~+gNU8xwpa6n0^`SM0sQ -|@v%gSPg;TKk}_eX!O(Xloy=wGZ0b2W#zvw)VkV`=G6Tu+~0kYagt&58B!XYwfT13w+GLqXycLHk -_ugJ@SaNMRTjDQE#-UH)_-yE$WRL^+t<&qei{aqTZ-cZ?vd4YSbGo>Wv!pMvHo*M!nIZ-l&co&5@%za -x_Pd>d4We-l$P;H2;q3-?5n8MeR1Xk45b+X1Bf#v6$Ue?XG5bRl6-1j#VwG_Qc(Avqv_S%3|cGF>-7sRZ{~kbdF7>YG$`a)Ulb}P3>-GcUQYD(2g2tM+>y02HMdA?Wlov>}GdYySv% -l)$VR~Yy2H8{*D@d$8L6O-w=n{J=E@Db`Q0CnB7C|9%lDYyRF}eL+u`Bw+8Xig7`Sp?qPOMwR?JlgTm -3^q;NL4C|nJ03U`Bt!qZ@Ft8A2UyJUfv39G$pTqf*NVV4QJR9NeGqV_w{`kko#PPBd}YQGb$--+7qMC -*5=_B+w~ov8gzw0vy8|JJI@`sQpf~ekW?b6RqEg+V4c`ccS(?@tEC5?Y8bHYIhWEO -cHfW5|7z^)NUK9L>;O`>%5{4RiX`5q7GG}4ONm3Rgw)=k`7gp4ONm3Rgw)=k`7gp4ONm3Rgw)=k`7gp -4ONm3Rgw)=k`7gp4ONm3Rgw)=k`7gp4ONm3Rgw)=k`7gp4ONm3Rgw)=k`7gp4ONm3Rgw)=k`7gp4ONm -3Rgw)=k`7gp4ONm3Rgw)=k`7gp4ONm3Rgw)=k`7gp4ONm3Rgw)=k`7gp4ONm3Rg%pnCY?=8Ps8>~Ra& --WGf+uqppx~jN!!?DZEVsuHch))+t_4nY?`XcG)<;zVzVu2OJkbsElx-7{kcezom`#0!8*#w@wq%}n3 -(g^BIhEVK(CI~?{1}+GD46%tH)L&n@_FE*FL(J|^NB*I*zm1rr)v|HucD1ovG$~Bdtwunu -d41-qrp18T+L5v7E5X|SVOs5o53;+s_l7b)xM!x-_UgayV)a~hUG0kc+(iUnO03_SewnTwz?F4YNzS& -ty!4Zbf&V|Ol8x#$2PpdL18mc!S-EUG;{l=F+`igU{SDHhu*Y9x5?XA)84ghvqbx}cFYnTG_|u -0mPdB$V7m4s-FlM#8VWzcb|+GGYu39Lmfb(X_Mle>Z*Wl99;pjZJye6c!qecT@WZm}%tE(mm9C3K-R4 -iaPLy<8_UXFp(~~_i=_Ac!yAH^@jgGo{Y&VbXI^OB#v0Vo?-8{Cd$9D7Bt{&UXW4ngs-NN#&9^1`hyE -fF_Jho50+B~*TwcFb7uCwLcJhrRHcJtV-&11I!YH0H~%r%F)<}lYBuch##;852b=9)uYbC_!mbQMNpwV|!% -umz~0qv&D7<9p^iL;!5WAR8`TYMdWQ|)hB~qhFJD6)d02} -)v_&4~$U`03dg#|w*PO4%z^Se|-6QN&*PP~>Q=M&^vrTojX>+DioozarH#_EMrmoyF -RS^bHNR}ded)B}vW{`7Lo9QM)kN6U>?RrvFPni{I!mz^xErrNw(IGrPbY#!6`fp7Oy&%_HlH=!K( -6gbUc47SU4KT9!sP;p(@-A#!Vi(kH)dbQlpM84<1_{JaH*(H_Hk2FU<>c}?ueCh~e_037o{%{`q*W}JW8e23-XUosTcFK)RSj{79L4|3AH -<-U02H~DJUBf;rQt|M}P{4<=nN2E*+vu9}Hj<<9fKXIRF_nElQwEG}Q$3HE)3)->XvNs$VlFqJA$c22 -~92bzBX9HRn8t~7QN@T*sr*m*wU$t9453c>|E!*`PnLB%9UEaWtU9XF;Hu?JAUWNR+XX@0IpgB+Wj=O -@i$YTCbV1Icg;#?`Jb1B|%+{bh)1Nbu|a>6=RI!Vc7G)NeMvpk-qYDA;vV(6<5~1~1O7Qo&!MrM -v+(Bq!#j%G`!dn~&u9MEOpd-4r2tjOoxR;+x5( -(yVd^rbjFQ-%7Wkb_vU~H!;!l9bddLm+{I^^>;0KJ;LUJkFFw)UJtBSa>1gi}`HRmC(>u{pacA~p1H* -P^KXx!|XZE9A-OlXC9)|7AezZT^nf-Y0wln*&jbU56y|IsBd$S)K8MZh3v6EqYvmaX-wm18+mtlLeAD -bDrH~ZnA+nfE^&almIH~e-7vmYB8b};(`>!R!yP-`D3e?2l%Dg#FR%M?15l*^ey}+uEOv_;5$FADbk0GW)ShVkfg7@xV@IKlVxNWcFjD#7<^E+Toqd -er%Q47I$sLr8}AZh{tv|`>|VMXR{yMC3ZIZv0q|mvmYBKb~gL5V`5wTxv^zpXR{yu__la&W7EVgW|*w#pW4Oj$G(YO%zm`z+uHM8V1F_D3+yjuKl-n2{rg6|Y7-Rq2>Ywqk6jhJn*G>Tv8&l%VShFI5m) -SL_E*?n&3=qib~XDEx9?{5H`w3I{s#M-*^hW;KU3%Ti03wYZpU-GJr}z-cDLtZ`-V-{I84?!o2|J==+ -;fQ4m-Q)hPZKe(+yiYb~pP6>>po;_yAq{t5f1**{_bH2W9qUuOS;{mbmfcxo@Rf5CnmXKjqL_A>i1Uf9d* -U$B3f{TP?+W%hH!%QpMD<7J!u-14%`e(rhMWvRpI>7(WvHR%mbMKz<;5{D2pB06B6g~a4?J`s2u3L77bK^-eya2C;H{jjy0el*+TUBPU= -N3HoiI?sX_CNk`kGjidR=ZEdeX89@;}&mAp7)r67sD&?W_Sla45vNI@D2Ptc!U1U>~FBYnf(p+H?zON -{$}D)Ym)XBy|1$d*>|bX8hW*>@->`q1{TuDrw)PNP_v -W*A&(DMV9J`OkAz!{~vamsOJ@Y^JSb;CYm6z>x@GIl`XPoGf3E*v)^gR-aF4kDbBiQ)Bip=|$Glc}J@cIRsN3^iPl8bR=DRXahR{7O#h>}5@KQZ*xfwoy -FT=5yv)g?Ru0Qr%%`6SJkQwm(kzfl3!A;<{V-UNwEeGPgZv|UX9=u0QA?lVs-Lb>MQWai3zmUIaiJDp -GA-m;|-=hZ}43EH*;Td=_yaI2Aci_Wt+OrJbz|Vv0yxc9D{YYcY$?Q+CKbidr_9wGH!Tx0SC)l6N{sj -Az*`Hv4GW#>^&t`vy{n_l#us@sq8TMzhKg0fP_Gj3i&HfDgv)Nx@=QEDDNAN1ftK#}JY1o{%p&Hy`u*@?{h63os3 -JA>I7U}rEpBlNWPv!MMf)_xYWpT*kGg7&jm`&rO_7HdBX+RtL`XF>aE`)ckH_FMZ|(0&$cKMUH=V(lm -L9B$d{NBdc<{VZrdi?yEx?Psy}v!MMf)_xYWpT*kGg7&jm`&rO_7HdBX+RtL`XF>Z}toMfeimy#3);`ptX>-GrI~IWc6PI~!_IDYcIehkw+{Qe+23J*H -~Ty6?`F?{J;UxlaQ|WVAGrUp`(r;$8K%#GKEw1G&}W!E6ZTBgXF{K8`b_Baj4$pH{xi+bC;WWb{TJ@P -?EVY)U#8E3KFjo3&}W%G3;I0sf%k~(pWThUq93eJLf$>9|6Ag5uevVTEpMtJM1h!s`pS7Hyyb9%mym=sQ0(Lsoo3jW9?zr{V% -;g1K@ka{moA9!s^e&{XOdaZ6DS>5^shp-{I_fhwDFgr+MPlaNNi0WoTYyw+)#0sPD_?ezvi3{WI@zkE -(-hYraS9bbIP+p&h$k{vq7yn*Vdl&Tym@KlwlRn1G)L_qdO@`*_^P+kHIl^TeO`sHx+_J$4SwCp_D-s -xi-SEKlGtPvDW^qw*(DtmlMhM)OBiy7v!t;l#(>#QV5;;yfoJPb7PI;)!Y}mV__x_g1a{`NlUy{=CQ3Lm~ET30Gn#?PG#3%QN{DcdH=fvQN@p*z7@gH6D`-mE}rNCDS@I+ -hNdm$LVK(-TDJ$%{f-aVoQKfktI`Kjdks&=CBM1M{M{zWKuf`4J}N5zM8FYvq{l{~8Ps92?ZRpU|dt< -P6Ar>^=@g*OoG1fKEv(dXuP!rKX?#J!q5ydf9k0{c0EB#M_MrE>%N8%TDd@Wi84cl--ye}M$b3$*71> -*Lv6{yq+#cx3O9=9=tGUVe3B-r2}J{mJ`CtEO4CLGbgV+Joi}bLv57u$AN2z!LY%wo1GPa+6+`i_3kZ -7cg@tz2tNE$zlUw=_21G8H;TyyFH53$orKrfvL#*z>0>W -Ti+CN~hjsybaj+0>pSXE%yoo&Vu_MyGTN#}A7>nNH7&!5}%@pBff3^`8jom{Tc -4P6XoLsoZ3BK+oWrcKq_Pdp32vk6>9%)swTOmuxwqL!Sp61%i88Wph_6#Hh -IZIHwuL&y3GO_Qm^TgwD_!gzJuhripk^nDI>;!mNHnzj(-0r(+w@S!|yujnT2^-;DT6O{%cFw{mml3^ -QAleB?VmZ(*lX*6)@<7!-**Nbb{^IkO8;XOv#XU8M33F%PgMS}+t7s?KAX^OQDmr@m==wb($E9(gPG)JH={e}Q5r8c#g$M~i%h)kkA}Jp1S0NB%NKf8KL^Z0HTx3FK8?Al -L~wgT-*JG_vI%>7X4~HqltU$Jn_`|B99fiofteZ?ZoDZ$Eg?V!%}y+=r5`PcV|pG7e*shE-fWd|}+)tpVGzmNE_4tvHuMxID^V)DeY6aSy>w -D(2O6YdEh?~x~x{YCobM6weZCvtpFU>0ZC9?Iciu1yZ#qh<+T!2V)lL_h5F0QcB%qS*<&=i|v>eIHd4 -`=HD{VoRf?*xe)E+AqHdf<=QjFrE{LURVFv*7ZJSocLI9c^~l`CB50L;G!3B&k5|<`TXUPb+A=^9sfN -1cvaVOwBE_p$IqwG-Te_puk{eZ<2*_uv)0wcDDET`_R~Wj7o?ld+rsc%66S{? -RY{+q6Np{_*8|MkcF$F31lA^cQ%x6pV@=8nG|i`MH_={Y)X^k}ee`Uib=lkJ#E+DF>gwqD3vvp(vT -)q8@?YrWGnzp(hHU;RH&O9KQH000080LuVbTRLGxW+ejv0BQ#S0384T0B~t=FJE?LZe(wAFKJ|MVJ~T -JbaG*CXJvCPaCv=HQIFa%41VvgusCUIwNk*dKLt#XII7 -QC}GVSvS=l^et#l4gdb@VP;R*_0=|m9S8wEH9Fmx~xCks^xuWZ3fS24haE8>BRTjXW0=)>O_ol&yOK_ -y1x{X6(3CksXd4#`=MHIF1?CV)3S6$SJu#U9HCWs5Gbc!k$!5&p1(`FmaN|5PnB~9E4rt@&eTqUIhMpg}=QY!L}=u -dz{pN^~8Gnb%SKh?I|TH?wUy&a{~lLjo&`E80jt8Czh{_g2UDc2bWg47c(qvl=brle74Dr=bCCYgDVX7YIAt-Nbmpts=|Qh!?vmy -6>VJLZT0fX8!@6aWAK2ms3fSz7=A0006200000000>P003}la4%nWWo~3|axZCQZecH -Qc`kH$aAjlz08mQ<1QY-O00;of09jjQV9FlEsQ>`ErUL*S0001RX>c!Jc4cm4Z*nhbWNu+Eb#!wyHe_ -LRVJ>iaw7qGMTt}8A_&t9GD{E#-)m@9f&7XVrRM)%}C}4WNO!Z)Z>OziETdK5{q|B=9MgtUsK~ftjWo -o5Jtt6!sNr{UVE+<$NV*&`3!ck3f}-*)T0_x@tdU3c7n$3wS1a>wm!e)5Yok38_ey$}EO-M8O=>tEh} ->m#@R;7`~5{O3o_CG8?=pXLB-# -zo;M;`je@1N`&uC@N-KmOCY{lf9?Fmnja=V{PVxCZINh;MOz};Qqh))w -ytQ)MO#m_^+j7D+Dg$@iMCp_HKMImeI+|k^c~j-WTd(>Qhgbz&Wu!V -Myfj_)t|BI&sdH~)t|BI&sg)t^a6T$x0wF;{I&w#dbgv_KhDd%Lxhq6^*O8xwNdG!=cZd|%k$(!2@;Y))h*Z~+dqbqYj@%a_&2{Ad5NWR?4}@l>a~ -*jwM7(_dIYhjC9tsgJpNB)l%jc00@$&guh8cshRT`(OG){L)Vt+D$~$arlH%4#$1_( -TnQ>yg36U?$W?#lsy}nppSkMKT=i$J`ZHJknX7@HtAU@ZfuE~^pR0l2Tk6mMH*AblW2_nz)tIWrOf_~ -@W3C!|sD5woap}E}I7GbnL5tcHPkuDSy#H}(E`(Iim#XJW)AMhIn0Gv)5b@m~x@D#6{?hw|( -fp(9(-5sKN`2#f!$jSyOkL$&!%todA@3W0@_Gn)@qF?|2zmK@@@5Em_wbV)A>`%r$<7e+4&o=fLdbiF -pLlsp`9<*`-!H1X@A=8w;p_M@@X4=3$Pa=~-U%VUJosdH2>G$_$(|7M1L2dsA>>EFC+~)kUm|?+UI5h -?2>U_^17Uv%VIUj`Aq<3rA%ubOeh6V8d=Nqy2p@)!nyxh@btr@&sly=zNgW9xNa|<^K~kZ9sUfLjAp} -Vs51}PV)ys>laFx|4E!rq8+9*TSC_~jKL)9ol)hKn@sQz!1ciJcm)hO?@QQEOl#;H->exr<2qf}t4`m -wGt?JKK^=GU4vsL}ss{U+Mf3~VWTh*Vf>d#j7XVmek3i))FFxAgd$E%M|Hxj1$JL-4^^66uQseX?-UU_`_1YxTGqmEY`pFTyHl0($-O5@Wlgh_cskjFEG -NqIz&$FB&J@`xah=LnPXh#-#_2$S-NAdi;_lk$ilkCzFP@`xah?Sx5rM3Bd;gh_eCkVoiPMX{7e40(i -3RTN8k#E?hmP(`tnM+|v{&TSM+dBl)M=Y(4hkr$;d(!L(rk~6va -}|i5K+UzagC?Lk&^lC4G0uBS#5j8u~_&3^GIsWcm?dQl^Rbj6MzBpD0mgB}%+k^l9k&M2RvlQQ|$KPe -ZpSN|c$267RR(`)~9ih%tP&jxc4hBSaWJ3zw)-sw{Ve_`+u!2vZh3LUiG?aFrgV%92NjEqpl_a{Cd&3 -!jl?Nu?%3aN#quEU6se2rYa@mL-(~9D#+;$g*T|d?T#z8CjN0YBB^BJ|oML$x)Ax!e?YzGCArIQ22~2 -OC~iL!U>;|Wyz#GGRT80OD5%!K^|mTGAWNP5fJjk- -hctO;KJjk-hxIxr~JjejZ_(9Z#JjejZI6{;|9^uM4lJSHnhdjbnb0p&mQ4V>KRgm$8D2F^qnaVgrlw& -FVF=1Y9eMS~Th8?2ZTa3@hV&u|dR)N0>PjgzE6`(75G| -B^6i(fi(_pY$~5lna@zf(zOALOw~lvoD3*hmc9H^}Uc!hQwmj_w(;_?ntQ*`(C@9Bwf|_v*~0=+C_ai -oB9wW>6pR`@^i8|GJFyhm`>DhGK>-xm`-GKWH==%d{2M5mM}~w>ghsGrvk1SDS$%ixWf1U=VV<9X*za -Yy|OL5k|$G&)xFgcw-zkjinFgcw-y??QdFlm>-t$#tzb1j{<#w)TfNRdf;J!G-IDJjftPT0Uz1Xumy7*@{~9BN-$~ -vq!D($IodQ*}@=@d)uNnNg+w{J!;XLVVEeABz)AOH%Wy_GCpe2oBYPAH)D|UCqhy_mZW?PQvO6}^kPZ -M#~|fTgkEnfN%J_Q&@oW6Ac%VEa#moJ= -g){upfkiI9_tCEFi^?LQGNePYS>$6)(Ugp(4|3tXxh$Y(}gY7>N`n<7Z`(v>EC&DE~EZP1TZ2 -yUH@e)h6KL*=>A~Y+pWcy>V{hz-=n3MNnw40x{V~}76QNm&CEFi^?LQHkl~}UvVsR@+U&G5=&A(1}T3cBbJ1D -48r{L&|iv^29FB!W8y^i00{GwAz>a%!aN3H{&~2XijkE6B`IXk=3j<%LX5=xlcf4nIRcTEf0CahM<7` -9lcf43YaW9&KN+&-v1H9-u;wR8Lrc~?F8v4$w->}%ww)wlV%ZihK4Pre!esrp1W6J4FEN&GCqqvqF0p -_nf1$i#W`2zITbO_!%jH{Ko20rhv=Y~SfHX+z0${ktjh7b;--F>IKdvi)lWDGf%E9dZ@g2-0SJ~J}j#wzEAvj|NJL{Jg0G*x>W80NO~(|?Iyhppcqi#%uj{+kqJ0+iWnz7ucS_eZbky;{8X4dnSePz6*?P -8D7JpuJCp6k*cG6c5+s3CQ -nLAyJrs{60m}0{ML^%;QY@kQNQzKBN@}rjow*!A^yiF9CHCieSa_d5}oUC5!9;u|72Qzy*T&yO@ljd0iewIfnGNrkH!P4@aN3Jw~FSuHk3$mqoRLOMd -1o_gZ!v%Lz%l*d4l|CIVxRY9%4&+In4j0@>t=#JbJnrdm!JSARH-R^HI^=N^$>Sz)#7>9)TOxVf1Wwr -LaDA6(Ck#C9>2SfFNFFxk9(SIpyY9r7BhzQaJzzjYat -byw=(HE0V~Y&uU)}1$sFtitng|6HMuFj?XU30kPTTJ$qFZ}w>7`zJG>NrOE#+Y+UvKWjY`1+p9y15X~ -)arOh_E2;DXQi$kUX=n*Z?~*0kf*;+b%zk%DSI69%@@ju-WraD|kDa6S`;x6+Om_?b}cr=XqBgh8&f< -Hdd^RGcZu=QAO7lXko*I}@tU6cqHCFyNJT{J=O9D$x{t@tKfqOgnyHoC#HF3fB0mFv6TheqekRI*}>3 - -Qy(A7!7JD&+%mNdrv3th<+?DLt>l}uyIztELT!9kx1UCA`Y{0m*H6ioD)aN(B5n17+WmV%Ey6E4xx81 -pY=3R78)CPhBynb4IOLM*_zXKG`S(@!V`5mBe&eA;Z$?rf -iwZl0ug?C{i#xXGx*UA`~e&owKCSWg -UtXDU@ePq01D?6e*NvNud`~9tg#qC50{vQ>2K(oh5}{N_ijxca{`-DdmCK+gVcRvNA=Au-jQu=rZn>q -AbN(Qs}ZUMT$hrv!u{vqGgJD6lY1H%W4#9rD6mrDp8yxC$&;RsuWcy&XIdssl3M&6)4V;V_GSGAVsF= -IdVxW#Sf$?5OI#2(Msh#rYH_^j@-~nv7IUMP|uMAD)UfNFq-Gc_|$R^fXh5b#;2BZ04(M?GCsAO1K=; -uk@1oIWeWcC92paxcvS#2g&xcNbCNo(xu -$SjOkul{Q_W$_KVJ1^#Gw_z@LvAFKDY6+D%kv>ClF1a=3|!^;kPpdZifjg!@_e`|%4CXc27dBDoLx!i -)D$%ME}GNzdE>sxs&(7#$l}(KDni?hZp=PLt8hR`H}{gi^w#BYUUv}{xV -%ZAUZUS_6sVuSmme3Xtk_+vbjGesn9e(@G;bcsvfHVWVJ>s=^HQb6<(2&1R{pMBr1WB0npwu|dWotq+ -)L%PnJf;SsXpsv@^`!Pcf0D_da|r}Pga=j_0^p2HL^Unq!{|BsdHt=b60kRdby8M{a1E8ukZr+3kolR -Z&P>${B?!bz~4}K1AMo_Tj1{rTuP=d%a-?LNszuw5bnzeQeRj9mXV--p(}7J0neEWVSzpYX}dcTsj)g`&uuVpE7(1BNWy?$jEjRDkty|U{Kh07AOB7aw&;ZtDEbLH&=SKbbLx=7@=%U -XK{);d?-KBRt2`CElN-KR*@>?q{sz#>svqL6j;id4a7CtUNMWqO+s}*MG -uGhk7ZeU_@b)%^%f8E|`fl1|TZJ=K3S(2=-mTtJ&P#c~N|DR)B(ok0X>|%|bqaY5MIo(DA&-|Rq}3^; -)hVRaDWugYq}3@(`E4?=T1cx?NUKv+>X?f{TAiYjzbmayQOW+5>zAUI{UG&!A@zSD^?%Vyxyc|v*->k -=Qre1A+KN)zin62hRw->oDQ!h5ZAB?>*n@1~SGxs*D&lsdVTI=P -e;Gs{eQH>L5Q_q={(4;`FgJm@`dC|tjRlyFz}uGD;`+{jq=q|~J5E2ZWuC1X+clqD^tzg^1Wbfr8nr< -D0nr9AVd?8}iOPrWH+1-G)0vsfk&m4zIUvNBp($oG{e!<16{mZiM5Ql3gF6ERBZ^OrKAqLkXVl-jqH+ -P9SJgHo;!N~wKI$vjjY^$wNPl2u3bStYe(6{#^yA ->d8v#$x3pjRim&H75_1Xq%6sHFC*r1q<%_N%1dS4qFGlD=FeeYr| -%lS=w>m1GbqsZA=WO)9BPDydB>nb1_#a_q>Yp{iEvfJ*v(mGt{6sd=hajt7~oq&bM05w_27ssO6fhmYzv1J(F6Vu2=Wvw@D7XmZ_*6T{Xube}5rSxQ7DKD|E<#>=~DQh|3YFUu7Zlv7g9IE9Us^!r@wTvk>a_!U@F6!`8R}O{o- -A0}*-6X2_8d>+LkYKAm2g}o`1kKM@R^#;|Ft{i>p$`R`6 -={mR5UeGEWMbgNb)ySFENK4j8OV-F)*vQo6Msnee^oyHXS>DEQ;YW5~ITB_%HHHgc!o%E@Mn(!7IWt> -ng<9mKU6~8h*jkyC)*@-*%CU7GFqINedX -Bm>MnVLm|5l@ds*Ejd#($@_So7h$HUZ-R-WM8c9o;h%B^#4Ze+9n%CT^ow(>;f7D4glyTaggi?H~W;| -H!B57ReV1jd(e_Y>k0L^h8fRCwRw`X^-X;0ywNm!vpa)ato%{Fg$0ec4a)Y~EIq5-qCUTsgi?Jwc9P> -|DBX{B?1cBp=$s^#3k(mmI^`D|F@fZgrO&!=(|_(_##l`W!tuh9!;BR`Oc852LM>X^c8j*F_yk=tLcC -N|}G3>P2^C?u=`$aTAqu?#M+>)X^o%JIDYZ%uuQBj;>}L$+$;!gClEV>FGzYboHZHuJ5B*7Ljw!%4WJ -|S;sui^qXY*P2?&+isdptisiZEu37)KjM_x8tO^#zGQJYUGQbjDD+MX8)_qb+S;o?xan1ToYAtCWZD^AjYCC}NPc4y>9>|FlWUgqEXkERiV}SyP9hIwkFM1#|A*g -|Jln2I@rEp$SXshKm_LMN46tD5sYN@pb{Mh|VFlPpYaoxv79ORtraDb<>|D;NE)S)aEhN&Bd)JGI+NC -6Mbk$@S35^(mCOOh|}wnW+}#nOclRxmI4e^z&S^KC3C0Q!me@63BD?CVCL(l2UZd@+ga3Cgeo9WZqn} -oFKV8P}nugt9x3Oy^fS+PoGrVTTZ+Q3#Gbd`8mCgDytgxWd5aV*6$tbt8+PGdp9_AONNi4o<7s2r&UZ -}_h3JghN!RKyB{lZB2e^dX3^>A70v~ua$62W9 -UnTgtHKrxD(vJy?__=V5*_JP>6Ju0{#o$Xqo{LEm6j-sl-NDOA^FvSXmXa+NMC{=|D -!h;z)sHZ|nn!(KTjOIY;Xu}G{X$CVdG73+0hR)clP^M-uLq}Fss3ts^p~DQT{z@z23f#(Ih7M4yP+?{ -;6V~%{XbY4MxvB~y-`5y5@}=Xhs=~-O+$rc#BVQhjg^_Q#ThO6KzH|guRT%k(I|dzUv#7hC2uyYUE3YW>uw;FRiSq(#V%qR#j=_ODn6YH1 -egDRaF}KvTiJmd}(D>l}5fFGHT>YC(~A?k?$c!jeO~#uBtThrBiCFQhF{`X+%foa#f`f9b19Yh>lL^s -x%o=X?m{G{#GT^h$0=Kk7{g4zLhcmp`XyCtFbTnCyYsts>TsJw=?F)X_$r)P64luGJb0fRz8ZU<2k(=^S7Y}x&j -5rRF -wZmx7f61I&SQNh5@xO$Hu(^Z|3>cia?Nno4v{`>b -XcQ7C9}aH8V!xCW#*bOJr9wn8(GfGHNyZO3JaP!R3B-Ll4gTLVNElK>QlxeBPSXbjjU_tnvI-@QETKz -IM;0CL>8iv)y-V9krTO8jVy5HnvI;uF>YjyGuLe7L=JN!%bdApBPUYZjjVL$nvI-DCpS7g+Qddqd>*{ -4cjlUnocKui>5^yD-t#IPpyD2AMgX5TFYBqfW-|i#$az^(%{7}5z^Bg3s%oy;i~v4(UKUn!&1MAf+4H -itnrk*AfRCToG)9w{5kS_uk!J -Wm-;|krP`!SxwC~8#%G%%e0&fMR1rcpHBE{G9xFpd~z#x)O3xU*z)P5%%*GP#FkGUJ>!~krUe|c>s-THgaP7B+sC6%|=dapEL*1bd8+YKFO14T(glA+b4M#P1NK@PHdlKQoU<7a$@@=kEC(UMow -&>7uzl*?_>k&o70y&~55x#K!j`Y6M*xoX8608DClB_C8l3DiIKq}sR;qN(7 -y(Dv^2steQDagqjz5SDY)p28FbA?Ug=+@?$aY7bjT1E{ -VS%=XIaK$O$y7MPc1P|pi5iolaD?rS++i3sCO6>-+a0;p&^4RU%yvf}mE)StXlA=3Ps?%5W;CmbW?xn{`eC|ie0%c+8#juW?!ZREuENuJ9SHI0!I+b4N6k83t^V*4aZ=elMi -C$>-WfF9RuzRam_|fY@cKy@2F{woV2B8Ov+>|$%eM{^X|Y;k7yV2;eAsX#E83k&r}Eza#8%#oFBO)zPWvJqOG-94BiW7+DRJF -VUi)|zO|97QeL$n*|bqt*m%<|uC2My7YjD77YbGe?QvZDe|fY*QP_jTEik%+s1^#2lqQx3T#`(&McOO -3YE}a~qp4Bz4}J*u)&AKDV*?LelE32~o^Z7G@h8`IB;Q^)98>ko|L%QP3u)8_11s4d*{c2?uRrx`A|B -YpDM@N;zl~(+#BBS`z`7`;bx74W!{(z5S|9^)xUP`!QA$NCO-kV6*=_@eI%8_MA2|-^5c}mKKWI%zO;bX}ur5Mb6sb82Ov6-tyHNe)ky9_^$b4p7D -m}J;pP>Yd)N3ykUEf@r>`9@8=nBNZ(^TXwu -NgYR>VEX(<4{zT1^#k4>I&9F-I3sam`*}r)k~R=8~>zb1$gV;i;(()Z^69x^rt!k?Ultxi~jN;qVLR) -+rpjFy>|y4i|B5y~2?kJ2$Ivc#Crz6pr+{xsAZ59O-x!&b4t*?%kwk`#0&?;wC*?-lS)%oAhjblb&sE -(zC7XS&tjKcFlB<3EI-`~5*_xE -q|{l!hbzr4x!S2y|o`X=As+~oV)mA_wa4{r^Her%t*4TgNbg1uBv?NP8SB+%+Y8Lio>cl<;3YY^daUV -50#ddGR`VQBSZ;@%RU!CGc|EX9wSRx_CWRZpzAv7Q)9g=uq -+kg4T?K2w%9@d`8G`*-*)BuCQ%OFDdcg``($Mf!m(QV)E4!2+<8=>Mx&DS1bIY{6$9XwdS>Xx+=^dDD -4Y@FnQp*wwtt?jSn$fKvi{HY0`IuL1hBiOOOXeE=%5Hb9*KM`w&u2w{){VuvKWo#U&x-!k^>o?>{mJv -b(JK){V1LXDzs7VNFZ>LTeT;Smx6MlH2*YN7%qzRb^xr1Ye@$a??!U(O-~UILtUv6UjSiU;Q>68Md-^ -d2dmVCmvx2n_X{QGB^vYlr&QEd=G}=j-9Ybaxqn#6NyV7dXP};}Bqf8u%!N3bd5Zk(M&1hg*w{|Wus> --&+*NQ@3yNR)-R}HH>x#p#pte0+KNG)*Vj%1Zxv-dvd*M_Rfp^9X1%{jg@Bn%y@R{wD)Rw~prt3ZZpR -&1ARZg1RGk~ekD3BK@v`^y)k(-+8VRE)+ocX8_n?k``EPhTLf(fzf|yDop~pZp1fD!$^H&F~BPMu)2S -O591kH2?N*8SHtG-(s0oz4nUYPDSX%oh+x`Rvj*Dbh_phJ=ndR?wL)mJOs7Y4 -K!&Lo#*X)N}%9vk|WsLdp)@96hY|fb1q&>#G9_=$$T<`Ax$e356^P9rpkwg9H{Ts%-3cZCfKZ5_9F~8 -CZ3ogW+-UXC0$G9WQ0LQw)e)ScTmL%TF7MC?D7{yNjj-ZQKQ8`l -pFnSXT131GXJ)wvX4gODw!fx994+s&{T>aAEI8DUII2KL;ZLQYnM38k2jhZ5fU0 -c4WnCeggvrkA0upqrAZS-YK`cc{rsf)5@EmV{oI*yyQ^#VJwkIVLW0CUGv@mvEREqXRp%v+unYBJ(gR -WK72P9uaA#mI>LcP-m3HKsy_yU6x;sodv=O$8o(>Z-9jH4H`5EFdQbR;Y%OE`(At~dNa4qgYuh_ymKo -0XnF3;=u1_JTpmW)e;+EzSI66w}OSZc70u*%?(iK2q@50(mj21x2iA`i7k~odk`5AF4lTclADX)gUC!#5Z -z9*t7g}Ns}r%?7pG^Nn>WLi_`dNP|SR6Y45DHJ`KyA)cUe3cYRp8b>*T0Rmz10ycF#t+d*^o<{$k?0& -heIwC3Fx=cdQ1lV^4>Wzm9Ry7uaSuV!$J|59ivAj-77=fj=AI?EB^Ib>32sRQ>goMkQh|DY|CUUkp5V -WwD^SlZ-jWNn^k`3@o?X19FHrD{QAfZHVAWeKV9{N@4q(w=y$G=Auw}yddsy_?vX=iI7G2geVitYYgP -bGLX*~!!61~=gq+`)*JsDC<`;f$~d{ZBi_=!Lvh}#4TIo!@@0WsVmP)OlUfkFs(2^2E;sX!ruy9Ejf{ -F6W-fO`ar{=ZkCm}&P36f^C9fnugTAW+P-2L%fG{2kE>Ot#Ujz#I -{;NPC-(Lt6^8N1u^@foG5h;@7hgBU2>D5f_mnHIDeF6q#|v)21j?Njzc%u&T6vd=nBWxq+ldKIk!43raEc5&3R9%WucI -hUiu5|-{Zk~@(V{R#Ze3Wo-C=%Y(8?{vUiV)x=0^q*ylH*o&|O-4`SFaYZycgYk%`AM2Ake_#?)SZJY -%r$oo7t#<;OFozA+=1);DGZ)B45?cZwuF9@p3a&*K`0Zu7Xt9(W$twU-}{Yixt(aou?N@wmoLcplft{ -Nr(r&G0;~n;RAlcbJzSkL$+EkH3pOD8jvI*H&mwV#lq7$s9^?JFG1)hS{zTErEtR`W*_gW!IeD{Yna2yKx?hjLA9SZsG4^yEX3 -i<92b7LI}`R)%Z;>1;KLydgHZOaZd@_n9BBj0fMvO|r0 -!!zg{YUE2zKI91n-jqs=ohe~e0>Y70jZkBSW+6(QPaSR -1dI#8r^fQ18b3HGm+c**Lau{o1;{s6DIWc5(jY~;iahp -ZkNo1;qS53rSz)kEV71n&GnwokHpsB4Ct4zhic)k9q~))>RZ_9cJaU%=l#f{DuyjI1Zuhx!vgBnDO?~xg4^Z;m3L|}59)KjCUUki@@>LFSNx8Ufn9 -wIWN1xJ7N5Uu4~aP(K{v?zyqqjHsw=5eTJ39iyn84m3NB|fwTM}PH5SaTva#b>V`A)eh7pB=6j@9Ztj -aF&)9rugh|mOOh?e0I2j_-<2tcDR99OH+JyxIz4~A1lL+{Jr4lha1VwZo$zHH?j1=(GMRd%i4mYA3i~ -DbqkJu_#`QXHiPM?ZXuTwhZ@clcCz&S-4P=MJCpOSWmhbQO;Ji%ntP;bxXT*#F -^Xo=0H+hg&G>U<&gNw~(jMg8d)9OkP+E_J6pI6h;g7fA}VkI0rr2mgV -y3L*#F^9ay?qG|HHjRa<*Xqhwt)e2m3#Kk4GKY|KUFJ(M?(6;l8j)X>7_84>>Mq$`TLvlLuzX5)b#&g -4~oP9)83#6de8VBi2{Q*B*Yvzl_~-!;dH~*y0?;;bDr-w%9f|Jj^;48|Q{}N|Qq&-=h?}ZLxQ5cvQR< -6+#}w~%Xe`v&wW&kUfe6b7xI0VM^0+IK$b5yBQCy2QPLJ$5f|TM&47)Fi$a@W7vkbR@y -n1eXIHe?fjD}Utp@fVjy6%8ti|rbkr?gRdpO!ejGQT(J=(-t1^W(1kMVfMuEWt|?2BU0;pj2ZlU}}~$ -7#{sV!z>NGi556TH2${L`$^@mX3bKo(#61jD8iCr*^29@90^IWVZ;Kj-F+^kD%%3S)L2nhBA7VXBjr4 -jGpHai_q%md5SWcBH^Q#So(;tj$R^a)f5RIZDkvZK5MT5@_tunGq$#w|kAY0x>@tPf#lE#~OmLYIG-o*L}J1fVJ@%%;ndi)rBs;HPWevC&f! -r0@Nh=nnA#>X$Qr9(7({5r2SP-A@jI`;;G+T+(*Mkt^(-pO+xvF-6Lp4}*-G=7Um9YWmWxA~n>M`_Ht -kEW{lnDgpPP4O}3Y?+GUq*oOpUm>yM`(Pqk!6Is;ggL#CJ`Z@Y!uJ+@|`@#V*`Qm$@4s}P%n -J)BKtXrmru5`)j+NA$yT;j2%Aq{W={rn!YA9s7d1dB`_0q{pKNElk0APFJ0%9RrI+vIRi2@UrB7a^P` -@c6J`o-TLh6&(*wU4d?;G43h^$ZE=y3~`R?JhB=$~C_VC&n!S~4?wkg=YA(`j;$-wL7l2_5tux7h785xfc_@)J5# -H#T*}Cv;Y7Y^sV+=+x2JR284lQJk?UD?Xv)Ghir8T)_UxMd7yK1tDzuxJvbh -W7%qG`r=8BsPQy*$8hZ?3nTsb=weB4uEe8{1td|*0Nn|s>FUxgzl-2-(8%dl?Jk%@iDwX%@cB)G6H)S -pQQyS*J8DwL!nbNF1dAO8#MHZ80RhjF5-M<$)t+;?zcCM435YgHcU+pbkbq@RlteKok}!TQi&NFs30g -Y`6SOv$#vhHwRwn4)ci^HrNod_=&0EHrNnas>Bp=8*B)@35VjGlno(^keC{8gAJiAN=%Wr!Fn2$QEKEH`k@Xr@(sBS|+7eS|Zm>S|ofA`OZm>So`blI?D_kGiF^59F>qDm2p^) -$TFlymY$aj5cdmRe-t|z4(fqNdza!-PL9?XXRZ(_>e4Q9jjYhsGv4Q9g?TVg8U4JgK*nEH2v*^u>2O! -d3LY#2FAOzpeDZ0JiSrt;ljHuTvOQ}=E#%UZ=8*+`N9#MHbS%(7N7CpXTrRsr`sm<_pkT*%!XcSV -rtzDW?65Ta~x+`Z*+}(S#NZWd|7XFjeJ>em<>C#tT(zUNhL9L?gs0_fJ9;n-3``<OfIx#0x -Ze(?0PNt+Kabil{4L0(`H78SU3?0zK)VdpN3`1y%sdhKmNG9LZyBn+z-NeLHyc=u`>5|0Myc=u`7e$G -wdN)`v?lp=zToNay?A>5}=$w2DDUZZdsvFF*b7YRaoaNDO4!xY^@obK~+{h}y9C*2r-`N~@Im_eO9CkSyF4P@r6f -=)@bI|1mcAd;&lpDP3H09bSt8fTrV#@gqHn83>M_g`T!)p$>+`xv{9B;XSb)7lfas$snbF}3KHrldSf -1+y%x@MEy$Kwn{GN<4<6v=&u*wF%!JbXE%njDJczC-MWfk?i3i034TOmn0yPhdWq9fk+;1V>69~8mB8JkzS=mjkcn#ygCL&aEn`Rg*B8M9vOU{z+;2Y2s|kpJ)3?GZTS|NH0G3mo$Q{VSq>ak$kjzu)(^z#;$NAAVopkpI8Wi2 -jWr|9^i{jGqYd|M%xq|3m)2e^d1@AHrYUCk9_!o8K3M -H-_XV8-?<~`D(Wex4p`E0{Jh#uv6fW|Ke-!2s}3LC*+?%{-Yho1rGTSpMFZ&b|0j_D)uV3;9P%GNBPMSG`49K*5;){PdRhHGh -bhiV-=3obdSZ6W??;L2cigZSWnr3I4&Zrr|dxAxp7!e*?~AHBCMzEK;N}-SWnr3-UpAvddd#;E_fW)Q -+A;Dz>}~>n*+TAo`e;r9f)HgL=SoQI}ttPz3)WykaxZl?jb{wagZW56d4B^VndN}upl-R8HWSnu3=J$ -JdZ6?#&g0lWwb7fyDL%Lad#z(J2p%ili0Cg%9zBC4O7M>c5Ik3Cb46~lrf1NgDG>KF{56*&MzhuOquh -F8TImYJ~5+SzRn|N)XVpGlwR#nFJI>kGwS8*d|^hte4QuEsF$zvgBkVmbzU%|UcSx;X4K2qdBBW%`8x -lX(F$*M=(~eN9mXKboO}HH?cx`_lz+dKR?xBBTbsln$^^X;gD4a9M%**h2u~vrWrE&_L6kXvmr)ozjt -0a%LyfT708z&K4;>0>*C?J7gZgqlEu+Y;-Vx?sIus7g(-iBAL47%ImQmzTQjr{o!lBv85(M?-{8&aoeZ9pK2-Mf; -*)V(0pqS3<`g)2duGs^9kjDmcC?~J6kwgxq^I93j?zIg>=EcaNT--~+h`2D -+Cef2Iawr8o85d^$vVlbo0KP?4BZnGGX@=36u%Dr^ug7y?Z+C{jb7-qVBPN(4_46A}=###} -{dtDLcN%%1qhuMN($UjxTaDQ_=T+xTK=*{b)%=-}}LmioW+_CF7+5(lH}tpOB9kOBka?!Z@x+!Z@)5lWF{%Kf(-GN@zcRVCeFY7zPqKyvpvcC6`uwHx;BR6wl -9R-qO#f3M$wy%<%FDRg) -HGcxj9(_9AgHMjiWw8Mf=FWWO-OmK(L~7iM_1fNJ&&Gwczfp8diMFPu@)eqn|e&Zud>FvE*ql!IEBp| -qH|fqZA!BSclGg&AJMG?4EM1)<{x@||J7yfyOWwO(uF%N}rR( -Gasb}+eydh@EvXYSvLOcDtpb0`@k6~#Cq -zoz3!*vn;VO;XZbSGUG3C02Ze4)A~Uq!U0{AJi6VQ!l8E_bA~T}go9w*7J4eaXPr0vKD=$l@evRg75@ -pjDxFNUPpDdTC8(cZ_`n9>E26L2$7r%D{8AWDv-SE+Ux@P&gUQE*~omDuy)xComy6-J@_`12}!6~}5o -Vs%Faxal>O(HXf_b*Q?o-dIZzz3J5Ox8b%nyG8tvxt95A~SkFSbFW{PoT@f{?TzEwGsMwj!HyH5dz}T5B%R`8 -K#RN5(zPC^}?wKmMQvI5-r%x|mvqyfuF|5xn*(qlf^z4q?sVn7JN;*^5aAzjQoBEaf1>;aH8Q<%Ao-I -_8OkMfr4QzU1M$z+2Kjnt*y0_$`Gy8)x(tw(6Wiu}>b;7!jZ(!LIGp@FZMMskK>;+49fgI)}>D$LGsR -9;sQcV4hG{H$Zb&VTZ;ih(XjTuxgEnh=Rp(HVbYWsCpMD)2CRNIzcTU=v?)61%>(~f=C(t~&mq|wxmN -I|4=c#U|L8B4ERSCy>f(=@nDJub7?EDc>+B_(Nh>YCkRm7S#d)HMf}Rqk8z<;_5P^}6*Dm-X#s!oNz3 -=?kvw4e^a;IK8fXp0t|2;0CHSt?gwu^ea*`oOay6UrpQa8vdQkjHq4LTz9_4jHq|6DRcQ#@ziCs1WPh -AqINGSL*b`o>C`oI%L+<7%Z#W!$}VT!sSB2-CcDs1oE$L=yIcD3-7vIf;b9)k@ -fpmrTNboOC7Zhe)nqWKZ-Y><30)!Z?;&b{jG%BH3@qp@?M1A&VlCJ%>DsNOm1EDI(c-$fby6=b5wDb? -AL;({l8<%&dC|wZ|GeyD-G5&AvFtzOR>ZRZkX`Z92g74X9qOfjy8ylDZw(KmbfA~}Pi_s5z -;vjW`UAvDCrGV$Fg)PYfnMIXho^ixum_NWGzk(bZoiMBGzk(b?iOF;r93=*aB&kQd?ZM%SbTA>Vrp`; -GvO^0?VMiht-@Kr?$M}x_~Pb8I%30Jfbj2Y?+oX*T%Fr0;_V6YIK)Un8pr(t^uy@XoIw3Jx=(<99Ni; -8KaB1PkI8kQA4c~H&<~@#1c)Q4Zw-%`b*LXk_ej`}qX)yIbRFo&(ftCf&`ScXfl`l0`&5|he2 -d-{4_ku)q!5B_b3zxNh{Qg^)7{axjv%M3?Iu~jO2~`1PBEoKrh*Q1n32OC#x1DZrm@x60a*zoJ%Y~FV -@=x=%sp(09*VlD?21@+@?@3R|Up+sz5PYg%(B5#%&7qVpU)lj}@pF>wWW^o(m67c6e_J7w_$*E3_BVH -ttcV7w>%v#ewn)_40k_Rtm8uNZW9z{;oE-g1-#;8xQjRNZ+_$fJQG16p|L87xaT{!I8glzW}>0!&O^CLnL(VF7w+KFsqSSrh*vdcaHb=Wfi3ck%+gRpO1wg5r$Vq%m1gfTtb(RXAkWJ=o+|srNGZR -SGuwRSGuwRSGuwRSGuwRSGuwRSLHGRSLKHRSNgxD2gTShf<__&X1-@cby+lk?uP`pd#6M$kd2r?;%$s -lHG@FjY#$%@-<@Ff5_N~W&a^(BbNP#tc_UqAM!S0*?-8~h-LpFcO%yQ2XapLpO=5E`wt|I?mz5n(*1| -6O}hWQ{O{rwW7<*00*Z8b!56C2B9%2Hid34}B9(`A72V}P8(rN+XyUtKZwP+W;J@<jhw>pkVCbcx<*c6AjqLw -PF*9XFlOmcEhmta-)&_TCQp-h&FJ#2tit4J@~#GE*DgF|(fC#K88j5derE>BFChv7?y>Ml=Amxu8#hw3g*OqYjgO%Byvo|rBV^NJj* -yWIHM+sR1f+Rx67pS_)oRIdH(-1yns$w=ke&(4jXy`79yuKn!X_}SaZNafnk&W)eFos3kj{p{TM+1tr -T<=W5Aji0@pj8v}u?A-X-bZ&)1wVaHfy`4-+uKn!X_}SaZl;qma&W)eFolHrt{p{TM+1tsKuBxn=f>TCg>|%c_jBX!zrs3NyZbrZ{qJ6 -39W9TvaLt$rudt5p>$zW;E@vHG=q@izm$Qy8be9)qyz+=B^mr|doLEN}T26(L6YJRkjJ0o@1!W@ZGCy6DmE&DkG<`q -KHF%yy(lYBAWwI8STSx(~(1SyiJ(H?NC)q&^5zFc$HX-ywWbha;g9m6Se6>HVw7z!&>EBxMtmXwJI9z -n)T>LME<*1*@D!1bfcERcdxPqsrBeyj>YqPA-WjmhB(a6<}hQ?VSbE+2`3KA@O~dM+LiBzFyW`Hq7ft -fjMYjjhd?iuA+}y3xv5?vyQi1xt8>YF^?Lg1N}Va@nx*^F(}L~u)s_Bj-6Q>4_FunM6`G4ebCN|OL#u_Bdy&fbE;4;}Cf~crq^nyjD`UR8mUgFT<$IU%EIQYGgR -7;gC_AzT%Siurq<=e>eNx6ZxLWG|GS+XR=WuB_31)pu?8EtP4>=MgsEL!bwl1*==exb(e#fHh>bJq>l -P`(LD!Wo=m0cOSFU@H)o5P({4n>XB&Ez_k=Cqm3VYvr~qDJaw^1(}U+RWy#6o^AnBXu*m;-xulW;40M -r8#Y8GmFZcHnTY_BjQlhNZm}1d1+3Y*-Y+mX-=Ej%yKrT&1@z=yEF{KW};F`bK1=2@EjY5;&su9xPtnayP6OLN-HW}a;3w3*G}o?eF<`I7 -TqnloWG6Ju1GGhsH94_}%yVKx(SRGKqkHj^t~nloWGhudcyYUE2EeF;Wluq7;#>QKmc3pw^Bh>5`#^0 -!Ou!x(HK|Gq>FbFd}cgY8hrcMDNJCAf{jmaqtvLm}TSv_>pJaSXPQpI>4t#$XEt6G|{0gDvFlm)MLk* -g`Rd5`@QK3sFoZ0-S>_tdYTc47Sk9rNn-W!4`@zl!cLRxP8o_M!vM@C<`NBwk?H`FL}OYVdP5-wX!hs -rLaR;82LWWsF5$NOUlB?H!OPPP$OSjo|J`=Z+N7oLydfi=PC;$Ut&U)R!Q9M2LoaSCG@!*3JJa$ -dUFoNQSIwO%jr-^@a@pMaj2Kz;`^aBaHyByXeP8X4n>*oEg{A2P_Ku^PlvvqL%n<_bWE*7A>X$`f83# -7zE_WkFM@oZ3VmvaLcV)Kf6bwg?o>r{2=sS9EvjTABURNp_rjBgkF_Hy?n -=pN7{(i*T3{|H_Z-i?IhoX%8PS%sCG&TA#^yVCjGVZkQaj2K? -2Zr|Sw<)THQLPHCbHy4&xO3BL!lX-WGjhex -zU>;)8s9E$n-oVXYAJrb^ -`9SYBFmd6IT|KUdNNpSy*Z?F`={f`dvJOcMWevGwfZssq~GI0NsjjS1Rq3~4Eg?&*#L0=BdR>v_QCy+ --{koV?qBJ4bAZ-SmJzuB;m52u!2K^i#r7B6|Khvc&*1(?`*`ky`yUC80^I-Aqr9Rw2WY*_o{Tv_OL#K -o0Id(%(t-P5T+2Qoxc?ENpwlaYRX9`DJr|tJGijMXa;_c`eT9)TBl;Pd;XU!r&mJiB -&2aJzRI3VWjHtX+)(&w@d(d-RDK#gE#3;6;W?HdYCnx$XKu4HD{HpwNG6?#6#>dW4QJU$Lj9-3P2A6@ -|1^1tX9@~H4cGEJqXN|Meg0nLn=XFD_6QZA#&ZINXD&1YDnTvIGLM8xP>Zjzbw(*DwMs=TsL58I`zQu -A+{79f)u4t?vuBQ4P{aMaPC+H8#TS{eK_RHc?L50tt!eR%FwW^vc;*MibK$u?&EJbsO`{jsQ$?kw(J{ -8aDAY85lY0Yon#OOk21c2tiC}C{q-pZ9cq{1VCt1g$K-1z8w)-g0G<-UY>N^zkcO%bGlx7B$jhT#-7xNV|p{3jO>M&$8CUM0i}G6w_$8xYx^fw1?O1NU4F -pZwzRSZfnjn=WwK>?=mULgg!YGK;9t#>1PjE -hhc5_J<#4wJ!qLOm$)PJ`BC>0S%wF<+H@yZ}g_C97bJa!=*K8+3^5jES5(pzt=;SwtBFwmzXs`ldsLcJMpAigP7p8SO6id=s@mxbjl3Q9JjJ&v!!sZ3S$cxYM>_-@RksZ$hVdOPJT10ItIBQG9eTYxa~;+%LZXv$gcXM~YQn`j|bAlN&4TKqCJ?=Po`z_vlLyGzS|ZE)m}ZsOA3T)ha32Pa$`MNt0Po<^*GNFgldPgJXR5^9c~c!!g@ZzizI|42An;DM=_UyLYyM -sn1m44A!v=xbNk$NOc$E7YLEzyrBF_p0frrOg%OME7NGG2=6n1%*?FE9si|bjRAPB5{7zBYAXNkrrOb -}RjI3@_pL6rhQ;6>Ul}>oECmw;evb9F2?D>wZ)JkOoUB?P2)wwBLgEF2z>8wVh9L0bU -bYAb0*l0x0zqIw*cS)_FCGe;}8TMy~r9ELEzDgL_-!P2rM`T69j&lC5RyKh_`eU2m+6`i(iJ{wVmw -|g21CUnZ!a6c=V3wNyt}tk|tEUU-Ts8`vHG1Lbb{ZMW}W($KF0dwaSx3sCN7m>kWiz$F#Y~q43JK^ZZ -4qcD$2kC_=U4T|DX#s#Sg_Lbc;}*egS*R`D+g)sC5>E)c3!9&EW%5@mrN_+1{c2- -Pa@*Mw^4_eltGz5`!1_lL~F0EWk(3n+N*1M;2~Q3hn`!Q#UJR_l!JDlvsnV9%?snn*3p8m!g2NExw)HEU>T{RMXuS;x{EJ;O~$ -Q^WQ3-~cC(8Iz9!PH1ap6pe7F+Z%`VcIEeP$QQ?<<$LnI%i$e`L&S(Tl&nNT#ktc)-Dw9#kSeaPik)&#C+CsSCs}G?0OkUu$5pY9-@U{3``R%+D$qf__fH5cTs4*0UOMf{QPzySyrOQ%Jy&^;QK#)-NjqZ>&m9WR^p8T0yVPBnv~MXG@_bvtkms)z40*nzV94{iugT< ->hz+>OK@L=j;D8$q0&Yxi3E0eAq394FxJ=8Z5&?kANSW^+}8}bBA*{W~IWiqv;zM(afDFF2i5vit1(>D}bF=d#(p(U@Wv-AylX{M0UH? -%M@m6E<8x73tE`i8<+rqcj;~Q=6wT;H!ln<-rP^fOYf*RuM8~Syf(0Sa}30H)SI^k<~K*WorC$!QU -2x9Zt>>GySuboJR2)w2a9K?1{TkTy!VK1!HW3O9`$UFuKIDv`=I*q+ -Q8z+8w2xj!0PbvNXFMBNQ}A5nLEd0&21-Q5^ie0OVL?o=$-FFz`tjpe$7d4^)S?qHswSgv1QCteN9^~>w -jvoiyWXLk)Op6%s*d4{cbg=M>gxl^%hcd&T#!oZ?aOTd?DAyHx3?qKoVwSmQVHwG5p4SBB@ABJVTgZa -a-jCU}97!v(1uNQBQ<-CLW&9R(!Fuyr+{Vva{H}4u)ym@Y5@#c{Cta|glfyJ8_1{QA)d2di}UKv=td2 -L|v=8*RW_2#XC`OVF;euH>(=k8f6Z -1bmCaW8hCKJOTcUc>k{P|1S|uP$3i1!4y!fx@J)%Mg(#Z9ZV~(3dx9aY0|YzDjw=x&&&Zzw$*aGBDxLB^r8-PS)4&xXC&AjTRp4;{>7t!MlL_2b4r5M!-p{R8pc)<59yM*5*N8OT3$R+c{ugs+47! -;pa}jWALW9n2rrH~xY6u)gsR)Q1^;F1(Ar(dWXu=o@`5yoKlEoUd`xp^=d|+t5>u7 -oL>!exr6!D3aiig)e57}#j6!YpNm&3tUed7X7#yvHKWhf-H^9pgOHXe*&yU4N;U}50Lca+Gf}cZNKKS -kUKd=7tDY4{xV?5=^uvmtNA1|dyRvO&mGlyp -yN_#E}-khh|HO2g-&UxidiNd;AgDN=VE&p`AjvL -Pi~gLFvA)>MWm5c6KRo>P(fa#BvgBFK!Cln7EIrNe~ONC$JbAaE)<52Qv)&I6HE2eZRe8|Fa$cxhnq-IamGcSGLFpF(Wb!Q#7N -DU{RFn8=>E8%1rTUxDbZ^eY;>(i3-E&p_M_OQB4EV?6_LH!OuRfru_USlkUu@xHpdFtE5A@>aT~S(yo -L?EDt+%FJM6J%4^RM4;thYK`Zw?#BFHE6E@fx^WjMHV83ncY&gVQ10f^N5cO%LGF))|Br$<&PI#7dj= -MFL+VG=-Gzb0-S7_%s=G~mPu&gw;C*#>V_JS@%#6Aja_%F|1V(3`vU=+`2B|phP=fbt -4;h~%(2?U@5LOeP5fTWvD(D%#T=_m{9eqlI-W%`p?dS98q7@6{YL@q0DL -O#EKWF%!R6b8ISpFYY$$y&J_nR^jMi_h6G5px(=@B-K}8B&peJG6U4?HJJfw_L|H9HG5;LIe3DZ%m6i -eO=f_ay(Tk2&EB|P`MaByzq?)WyO9|nzZ;nW%5AEV8KC^48kqqq8>2>MfTBcdWCkcoq()|dqD1N>U$N -Jc-ip1J^jGe+V&xBmyan%2BQxOATbC)nvQFfNn>w>G8o_SWnPv3^zg1_JxDyOlof!vRkyv$R$vd&ytP -###zJNOx0o~;bxNEa13QzsQ6iX-cgo&>1RdN10vzb@2qo%kz^$ylrEA^gDIVecm5o#0dTY5 -E~(|wx{q*PbmW(28WVq$(9WdE6{-v&7f5cgXm9@Q`Or!Ek`V1l`FJ|Vofe1Z4{L*wbsoQmP8N0_W^sb -7m}o6qks70FYNn2P18r9?5s*{NTcisz|cFd5qV7fg<}l%|d9=|`A#p8g7`&ky8;E2h@=H(K?q!U53k( -NN4E>UT5zza9ot457NfaKic#t+_xwqMYIh)df?L0Cj#aN74I&YQ5lG}A>)_*mP@ -pMvuE!3dqFK=?cA|fBPq31tP*&!I&Qj3w)k#6UhzsP2^K*f5cdN9s5s=W!JGkV -XS){`&-6xyd+f*YCy^^tfhK&%`zv;pAf6l>Dwle6RH6%HNTeXfHYvG=GW5Z+$Qmz^<&0-hup%L?~p%d -EWeKZ55{`evH!_fA6+HBqX0@7Dt}%xWQ9VR~o2TUafCo>Ma~^Bu$UC-cgh=TGL9G0&f6w(bV=Co|(%yp0*hJb -$nc;!g>~9*F;qF~9%eKgdNywi+i-H`rA{rGy$3cK%Do{P_6a8S~@gKQrdX$KNm}YveRtKV$yee`C;(j -Q@o(KQjJT#{9_mKNzdzd*o&V3#%{Cg_qYV*z8_m4VBBoBbq$LO{}x|8k6u$I}RF?&rC&T8k2BLg-sfh5lkh18u`DAO`j<-nU^|T%D6aE=v-WZYB -*LJ+!C$F#VcsukT81r_Y{GGPrO%OTJEz+#W7iy7X^;e8}Z}6`PLv~e%F+VyZ#{7CFX3Q)3gfYLCNg4A -4K4Z)iGhN2~OvoAYK7EfdKSlbCdHK=$qK&-#XkF4qUVh}0wvk^vk>A-y-h1%=>a^#*8YeurkslefQfg -!GYLa)_A_?nO#{9_m31fa_+{Tz68Mia$N5&nD`H?|Ids`$%-Nl$689!yrj|?i<+ae$8pBVF6^&ZANYe -8jxTcknV$C#hl_cP{=%L9ZV`{_Z(y!`%|F)zP|81wRbm@zNEM;P<+`x#?ien0>JYx}kyIgTv7_x%+E_ -(cmZ7L}P5k(mbkkXe}t_($0AdUx&hW$dwMc4u~B7!q$oT_~zWin>u6N)$;^qsvGwjYgNz0J96M$z4$7 ->4EuWS7qh7#EI{VIH9r0em~aOWWS$iY_i{9YizRL-)L-#Y(<( -3#T40k3OE!~Xe&l#D5lWg$Pbn5;z+$}RDMf<}rLeD}I)+l%SJ4?mDeS8 -#i=h1u$P9`a9>4`IPwRYz)*{Q(I$wb*T%J~J?tt}#%hQUk9k -7^id0MfeLnShU5;=xSWCqm(dZJAO(t3VX* -oKelT2dp;SIinK>mKyGy)2|Tg40p~bLURBIy>m{#LaZ>{Ij3JCmKW}v(L)^87Vex;bm#zjd*_TUez2- -==bR$A2e98e=k#!aRew9@bQuD2eCM1tEY|z&oKqzC0PcK8A2d2(rQgmOZT42!S0~q2*jI790~Yx0=tC -WcR@heuJy`plGl~Nqu&!_Cj85CFu&++rt+207+pVy#PTN?^w{u3PZ7k&5Iiu4yR`KneQ8L;9tzqYk(# -Zy_-`hE(Q+`X&?ijG(Zs)9i!B}m#b5>_nEVbJ?tLsb18#`xp&c!0TowK_7#0tBevpOU5<#hx4Y3!WUB -`ex#?C7I%2XxceIjhq+^p>5oItQVb#?D!t#?eY+=d8{_eC^-R39IPT)CsF7ZEfg;Rdnv}gjIC%>4a5u -^67+Cbn@xARXSl6os&9Y6`hkhVHKT|Ix3|MXjbuC?XnQ6eycqejVgYtE!Cp~fUdjDk;{JiZ&ph!y^ThMyiRZ~X{yce~x -PRdOTqNvO*q&pj7c3?rTiJkH!B -o(nv#8b-Vjcw9G(_*~%ecr@Zp;Bk>K;$Gl!nK0s|z~e$;#4CZvrNW5U0*{M@5pM(@mkT4_3Oue1MtmX -gxHcH^PT+BMFyg(y^*ncaYw~x56|5jY2jJUA>Ry@NVabf?hxZ)adVgIdomObLa{#)^^bi{@Ix8 -iwez;j{$t$1cS;==x0@$7TNh5fhU`R9lW`)|c%%ZLm6Z^d)a5f}E~if5rCF6_S*&qGIC*ncaYiH>P0sUzcLKbHqh_U7Dum$IIyxIHP#{W)ZoU;yKiai^#ncPoqX$MDC?{>ObNlaxcX*q7fI -7dnukyjkt*1OYuxU;0x(lWaM5*&mtrDLV6Y%xfjy2$jH5ro<&CPh4d^kaxbK3k&$~LJ&TOo3+Y*86{9MtQ+m#}IuEeD3i1QhcD@EK~(RtOCB5tneyy{93H&=9Cb)|@# -D>|>bO4?oLRaZ&7>%8hJX?LAhT_x?V^Qx<)-F04dm9)FgtFDrE*Ll@d((XF1x=Pwz=T%opyX(B_Drt9 -}S6!8`dnN6z^Qo((-E}^7m9)Fgr>>HA*ZI`d((XE+x?0*@=TlcpyX$=FYR&+;TH0IZQCCZQ>lu)%`OM -wboGEe*+YLEhXB)J_^)_*To4TJ@y6&)EV)t|A$0zcBNspNi*XN1*9e+QiPu%bMxKbvVnG*TnA7Yw$L? -^HKhnQv_(J2@FtBl%R(->qh}r^R(agf<8Fz3a#vRk -C9=ui>9MWEDv!ImqA0P-;|}}V$#?E*o}$DmkGp!zE3wMst{(A9tn#>{$GZ}%JPuy_l6~`J;@`in34js -{J?`kiu*5=-@Zp+XDbZ!>F8OS%_y~Wk$(IsqKK@n@C?(c>+|k2Ti8UX0^k7wD%}2ZdcmyXa-_dn_i6t -N5p*7=EV$sJP&9{|U^l?YCZ6y|c+|^uKiA5iGbzN0r(Z^kqny~2O_wNzs#Hx?G44tfem;7K*G><;i^- -S4vC87i~b64|9C3u;;x;8Gs%G@RK4^HMT$x2{k{+_&J@G*a{*~$`Z%-?ISvJ_m*YhP+EuM~{QYhRL#* -z(G+N{sZ?jE`Z}jL#se8S9TSUT^sQMdI@l8e3H)K0l#ZRt2wYs)S})B`TH5C7!3|nqSoVj96b>e;2bN+klL?u_Zyz;ACZr_?NTN$tC^!{eX8Hi@aTC9x -Md8OOT`13Sb;`6h_=V$!)Hiw!9xjzvvO~K=_Nof2{!d}gs$5)fkP#c%HKT+8?bX{h+zDV3pmz|8)xj# -1pzrQ`6-;S5HCvZq^_{M~#+JalQO=OQL3&{a(>#A%aIiPJln_ox{Xj{)07m@?oju(!NxU|2X -wJ#(Wv>kV48gXg=c+0|wOZ)3-wm9=M;F{AGhkgcJ*Oubc&wy)ALLB=UaLs9pb3X&FXQ$%e&w%S0s5tp -E;F{AGM}G!fSD@nT&wy)ATO9rwa6Pvcr+)@q*X`o?&w%U7U7Y_JaLvDn13&|=c>!?(XuvfEBaQ$KxSl -16Ge85bCqm*7(17bXm^cMA;F{GC$AAW0&-2AOpaIu2eQ^+I!1Y{ToCF$h&25XLKm)F)^WrSffNO4hA= -UV8&22BF8o#aQf(xm}Z)?r)=sZv|?j$OZMWO;fI$lfP-z7A$E}@BaiHf>UR8U84rpWz?in>p%dN^us# -O~*-D*86@`3Y{ir%gq(*^qcYL*o4$?IkAnmx=qU#QllqKU#24?oZ6f23}!iiB-H=VioKtp7Q%kS_g*f -v|F6b#@L*xW1tN92_z6{5`GH!iz8}>JvEnb^JzZ*(|Z%H_Hmvxo*7f`TcY1nOVl?K3P_BPsn& -AW?4@w_-A5$bGS~qf^eO31>ri6^kT+;Z;??}G+gKYP~_ZRMb7P26p7C-cxEg3diroZ@q -78m7_QTiD=X$aUWz&8;==XB@13VUpZi;4eR;T^ct4y=4A*)7DI6V1LmErGPtWsP;S>JDN78Vezt4HXp -XZK_q~SV${<-J*XP(GEbGq9uTu+Q#^N#zEOMJd-`N->9p5I)@{UM>6XFlPdKH)!}S3a5Nm0`Hf?duc2 -*Yi~Zndg;oxX$k{6Ti32c>XLi?hj?o{io!WVOGxhc$CmGvLlUv+%IOFTh|ri6!1*oig5~fE^x&-1-uZ -rVw?g#7r0`a0`3H^7^i@Hfh)!-tHg|R>$+l`0$vGRF-`%m1+EyUfHwl?N7J|yIFNJv=2m>O;<5t15V+ -#90^SK+aajTH1+KWPfDZy!Tvk@K(z6to6>#ZUipvVP^en|?1zdWT;<5rRJxg&}0hgYoxU7In&r)1gz@ -=v?E-T>DvlN#VaOqi!%L=&kEX8F7T-sl8Spk>!S6o)WrTrC`6>w>P#bpIt+Fx;5S(V772~{s#=apQ|R -j6>Ca`E9h&(t+nt+JY{R^d9Y%<3%hJQ*J;Yp#Wa>%0=Gxyl=^^UADFj0AOJWUUh;Yt46#gzLQWs2g4x -*K`N^cs+4{&ntr_F_JZjnQB6%%bJG!Ta%diGy{MCZQ^;_guf-UpKzUj{(`R($QFFPVYtryZ87Kmwn)6 -6MWS+ABr4+tJ$rDx&SyT0g#RqKmX=gAXu@_g>8r{!~=* -}`=?66J|GdG6-4a>xhn$@#!5pCZAY7jt@oR=Cbrniq4viejF4KXcD32cMWj`i55~zTwZ~8%N*2Pppp# -*QtHOb;{=ZhBi|Q*SWui>xs|j9ExvxnwfpWInZ#OzJH&n9DK`1YTxoq9=FKBiFve)s8KHbnM!9xZ1Jh#nEd4S8M3KID9SOY8AcjowUE|S>Y4vz3*#U+4w}|QCUrz404qWzZ(xe4FL8GHWxg=7sBt`}xQnuBYy&Z!%n`mTfanSH6eq)N|U5SGsMY8&#X&EZT -*qk=Q0Uf^eO33hjbd((Qt)0dWK -2>Ri`wDWOsyMm~XA+3}ROJ!wqw(5hd_8)&PAl^+6Wsl4=RV`_Oz`)wo%=j-PZxg#9u8NsP8j!(N8!Ko;YkGmPcsUnWgyBWc$9F7mX8Ns -O*j!(Q9!6}U5BX34CqDFM1gB_;PrVtzDMaF9Zvu#Jb!yvdCtB6T2j7g~WY75In-QGs86 -SN!f|EVtvu{RlvgcnBaI$B7`ppPV_WTI}Cws={-;Cg7&-ehG5uEHvJRZ&Hh{vN79r1Xyp(7rT9(2Ux( -SVM4Ji5;jk4Nh{;_>J^M?4-)=ZMFnBEcv4o@ONG#zjIuc7bjE=+-PNO5SgyZN)Ea5yl5=%Icj>Hm9q$9C}Bk4#i;Y>ObOE{E{#1c-WBe8^ -I=}0W$Tsjgru&;_=`xiN}M$Bpwg;l6X9*OXBh1Es4j2v?L -x6#*%nE=t|=8;3|p7gQz4P50;X6JSa-y@!%(k$Ag?C9uH=ccsyuH;_=`liN}MGBpweol6X9*NaFF}A& -JL3N+xx4kdMUK!8{UY2kl6l9h@U^b`Xxl*}*mvX9v|toE^QBCI6DrjBhHSq>WH)Bs5;{8IH``LcQiUA -9*-`E#N*N8ka#@$8xoI4b3@|s=xj(l9&HVY$D^kq@pv>eBp#1$hQ#C1%8+Rd#N)ARop?O9tP_vNes$vU*sMNz{P9Bl1dh-fzjpEz5=+ -pO`h?0YA%giY@xmayZU#1giI}$7Cbs}V36m;fzcJ6WLmUDCz%%g(20vDgeB1 -gop?7opcC&#`*V_H(fgdZI2xZ57f07~;^Js|PFx)Q&WVeo**S4>bUG(4jyC7S#XF0??=1elv-tbY;_o -|)zwa#mzO(rI&f@Poi@)zI{=T#Ldz?E*oV~O7`_AI;JBz>XEdIW?`1{`C?|X~C?=AkmxA^P`(E?+$E)xYIP5Nb?O-VbgQW}%mNGC{%D|wN0RdY4{b2F;g -T>zu7Jolj{QY3@_rv5VB$ft?zaK3Aez5ra!Qk)R%;4|c%;4|c%;4|c%;4|c%;4|c%;4`G?@Su58`Qm< -8PvU-8PvU-8PvTy6k*x37{t7r8N|Gs8N|HH3}W781~Kn4gP3=jLCm|%Am&}hiCmW%oV?2nPTpk(C+{+ -YlXscH$vdd4#EUb7lXn@P$+*m*<6UOZ@h&&$c$XV=yvq$b-sJ`z?{b5VcaT4I%67TI$-CU(pyTah)U14zYt}r-xR~VeUD-2HF6$U5o3WJk(g~ -7=?-q0mnpG--dy)fu_R~U4>D-1f`6?A!Q=jH|z@8$**@8$**@8$**@8$**?|9dcaNQu`9ba@At{W`6g -J4N~bZ)TdZf>yXZf>yXZf>yX&KWGaB;P{{B_3ZIJi9pRjd(nyP~!2W!NiLb-iXIT3MC#7DU^78X>j -%8R5#-BkV1*amj+ia4s|0QUmAqHIL?iDd}R>!;vhHT@s&Z?izD2K$5#enFAi@b9$y)Ry*Rdwczk6L_T -s=c;_(nHiN`~*BpwgJl6ZV&Q1{}bHsbM>LEVcp+K9*326Zn^XCoe88`QlxpN)8YZSeQ5Hu!s28~nYi4 -gTKM27m8rgTHqWEQ!b027m8rgTHsR!QZ>u;O|{y@b|7U_9XYu!)#ou?v`g_+|Bz -|X+_?<=KcNU2cC6;)6XRN<>okjC^7R}#TG=Fc={Jll<_ZH3HTQq-f(fqx!{@(Q#*WX)Qe{XU9y~XwS7 -S|6YmUw(`as9oq{@(Q#;on{)0vM4;JA+ScLyz5&nZk_zxE0551Il{9qCO -gGKlc7U4e_gukB|gukB|gukB|gukB|gukB|gukB|gukB|gukB|gukB|gukB|gukB|i|{>mu4<3>GlTW -_GlTW_GlTW_Gh-3H&kX9{X9o4}GlTl~nL+*g%%J{#W-P+@nZf`2%vgl)Gh-3H&x}R*J~QM44{eord}h -c8J~QM4pK1Bvc(r~K$M=P=HG~PD8^VOojphA5H`EEA8|sA54RyljhC1PMV|l;N4GqEPhKAsCLqqVnp& -@ukwyfZWe?+gvWJ3~Y8&d?CNGc*M63=P3MLqqV+&=9;cGz1U -LmUz4~Gz9Mq4Z%A@L-5Yf5WF`u1n&(E!FxkP@ZQi6yf-ui?+p#XdqYF;-p~-dH#7wA4GqD2LqqV;Y>C -HvLqqW1&=9;gGz9Mr4Z)X&hTuy>L-3`cA^6hJ5PWH92);Bl1Ya5&f-emX!Iy@H;7db8@TH+4cxbl7<4 -Z$B@TH+4_|ni2d}(M1zA`ifUl|&Lue646yjnkrqXEO$8tR0v40Xa+hC1OZL!I!Ip-%Y9P$zt4s1v?2b -{BZ4!6cR{LqqVDp&|In&=7oOXb8SGGz4E88iKD44Z+ujhTv;ML-4hsA^6(R5PWTD2);Hn1Ya8(g0Bq? -!9xuuu~Zuxg0Bq?!Pka{;A=xe@QtA%_{Pu>d}C+`zA-ce4>kDjbYDUbIr#soKEna!y8t}Y-h1pbCxhM -79S%M8-g~OEazH1O-BW#)13KC4p6aR`(8*}`R8QrAPFB0ee(7Ykd#ayu=b`%EQ{9vUIvMVs>ZKge$#V -BpC*^=nrn{&5CO{7H*5o2re*F9otC*fgj(kAP--8YSLZDT*v!EYYnkgnt!2&uxt9H4=(WrYy=lVU4AI^U(cTQ -v-VD*+4AI^U(Vm3p@nZiZPJImDYX*OB20uh$X5Zcn{@x7!-VFZU4F16k{=xK|!3=(g!|Yu_9cHOvFjK -=|riQ^x4TG5)1~WAbW@>;)Oh-b<#4OC96SFXbP|U&%N-+tu5>hbp5_ -&NSh7y7?35F7iF$snek}(N}5}Gjyh7zJN35F7?F$snevM~vU61p*Q_Y&GLarYA9Fmd-1>M(Kl67n!{_ -Y(RrarY7eF>&`23NafA%goB|B}8K8I#7w3b3i7hKe(~XkeMdTtPm|TD@4o83ehq%Lo_!-G&e&uH$yZx -Lo_!-G&e&uH$yZxLo_!-G&e&uH$yZx?VFn+np@Rb35AyhaBc>0VFqwv25?~paA5{;VFqv^0+@X?3`~L -nS}zL*h`lTrp!Tw0fZWT10eUYB1_-_^7z)$CkbGG%K=Wn60MVBP15{rY43K?UFhKWZ!2scx1p|~{77U -PnSujBRW$q4rm$^FxUgqvlc$vFH;$`j*jhDGQL|(Z2@kzl+9Ml=Uck&H75;`-xLqcXw2@RRK9YkdIDx -o4X13S|>oEiMi41Q+@zcYj1nZXY+nFYTygWs9K@66zLX7GD6_`MnY-VAO|5Sm#ql%|0pHM3xV -*35zdVlxW{sLd=GAUCsMfZoi40fI9N1}M%f7%GcAE1@|vcZd4R+#T{Sb9d;!%-taXGk1pq%-kIkFmrc -kz;wK-%p43En7IygVCEc^rBRenf~gry6V_&k)@F#-W{B2ih}LF^)@F#-W{B2ih}LF^)@F#-W{5%-W+7 -UeAzGUuTALwSoAzzY5N*sF3hFQm;KmH##th)b4B*BL;KmH##th&_1Tgz(7?=cu5RpkR2oaeCgAkEPFb -ENu1cMNfNiYZznFK>?8d!+PBp8HkTc -eH`G~gsJGtGV7wtH$>a?|NG5LxIx=}fkdet7f{IMu5JY71hM*ynHv|cpydfyayqYOOc4u- -;H-y`kQELqqxvB~GL7nu=%k=CGb8_a{yaX&QcJaMSYp6K8dIL&lq24mm%=c{o&V)ct+p?Co%!wwVmqD -}H}Q&j1eB>HFzthq@ZM|4>n5_tRdP;kr-UpZLAditeiuwF5c@?0r?*2XuM1L=oGT|RcRm4DQ54h -(mtS5(B4;-P(Y`sy+4)q$4l&!I5ahU@01Ykt4TvZr;Ko4)n@^nQo?;zp9OTv3HMd;9?&T%+*ieWK&Py -5UzO(pozlX6Rh|cQ$_o!vZ5PlfF+5OpVLyWi^+46K0i8VOfhyhuI(g0mRlEmu@|*{%cn|30IS*7V8_> -yf9;o6ypp)l3P{n&dC;L8lnxK7-IvM!E6@pF%e((}OCjAvCw0*U}8;SC{)g*SvG7Tyq&S -a>@#fSnn@5RzDkLP%mk2qB3DA%r9rgbDc{U;(q>qj^=6P{(; -}0m%~u@A&Vnv266<^|#(6cqZ`c1kVNj7QqXF-yry0;5P~G1b&O)Uf^#NycD<|ifb$*4f -s0*uLZ7?OpT?a0e_d^t-#+S_(I_C6TB1nR4g2?x=-Q+)bP!{_-0*d)L6_KaGgQxOnR0sH|k7!mM%EzO -nR0sIqFP$mM%K#OnR0sJL*h&mM%Q%OnR0sJ?cz)mJ%caEyd -LG>bUEXalg@pXpemRjrN%y+4dV5&H?>MYhH3x -1YLUHizqzO~27oygtEB_PH64hwJp&Uzw4Zn~|8Ck(is2 -n46K9n~|8Ckyx0KSeTJmn2}f{>{XaCSeP+bm@!zGF<2yQTA0ztkIxF%&A2Q0h}_Rj51%LA*xdBhdE$4 -?OiIAS -DCY|%A9RgW`tH|gjQyRR*C$H)!}U1tjq|l%m}T_2(8Qrt<4Cn%?Pc{2(8U|R&B;HE4 -YBO4EGg@miTIjG@JMXwK02cWA*-_Z>QCsQ -(Ui3@yJyJwvPS(2}9`cWA}X<~y`zXiML1cF2Vu{?%iD@`nVTUh!l9fx*)|e(X;fJiX+{{*1xXTYl`%8 -9cq_$NnROr}zBWe`4_Tq96MU22XGLvH#5A=~X}WUl=^S>&N~pgWC!IZwy{c@b?+qP4Itba6iHSgTccD -|A4?b?EOC(4151C2E*PzWH9XgBL>6Xf5~9j`>z-bd;gfhu=h_G4151IgJJK#Vem8>EFZu8!{y^@WlXq -b8WL11-T+RuglhUaWT037R5j`iC=NqWjZ_0lmun0eh*i~6IDqN)+yP9t=MG@HJ$C@p?YRS(ZqFUSbbI -aqrrUD|Fx{Rzfa&(!0Zg~&4&Y==)lxWs>Gs?KOt8_K;y|>0huXnhg4VhNA5<@oUh_KV+|Pvgr?gz&?7i>W@i0a -gBV)z{Gu@{1ukjeV;ro__*&Ackpz5)R1AW{%0gu>SLAJ@#_2}j!qr2hO)Oq*3)(ULk_im`NS8?C)A?I -aK{uS%TwFoKXONoJJjSVz?2poa>&W+C&~9hPhP$EB7vtbU0r;Jz*B@SPb-=#z*CIwe$M`Siqh3vuMl{ -O)6FflmMp*}{#q^73UGz*Rx8Z{JVopJrE3J9Tzv10VhaO2`S|*ZS_>ZF$;nqws#WL#ot{%ep_bZwL -f$#5sr8eW2n>7QRD4c=CwuSRR#ZZOVejV@qZZ)F-s{iFJ4ao$zDnLXDy#KX_TiJg_b!vShU#kdq@wcz -JlT77Me!>EhP}^{5abf}kfRO?h4}jFyQ~z?Q4fWig17OTa(UC|&+wZHantJk#3#+gO?#(4<2O0^;7?x -SH+lIii`^&`h95r3)|^8{3=V_&sTg5Ri -c)yxaL1p34f@n#OL#!4?~rx6stCIf6M)|PQ0JSarU%HJb%M?stiq;xSy-3LsKR0uZ~6S@ml;O4n`fC# -G7y0#4l|*DbXx=q&Ew?@^@%D?wL&@QktIUvnCNK&A{((6Q0o4d;-uO^6TXO!(%7N{pby#cH0MV%C2g+ -eE_FHOYOD~;51mN-Sz>T1}U}MK7iBUq;}f}a2k}8AHZo4QoHQ~IF%1-w|xLv&IofnYMHz -_I)_2s-|(EBwV0M5T1K#yvz=U*o<_H|r)i9H)8C#n)2;Q??wXUo4GrN4#aN%#n(Ry!N(<|B+fJ -=ZQ@B{=He=i~Y02xE3~mb^E?KB#H-2xCLfwHIRIWW-_eb8#6u!Z7(c_RBFH{UW9SMi?eP7cYMpVVL|} -Oa+ZF_VV2RX*`P`aqtg6Wq%HnpO07Cjqqgh4trmSw>FP3?0q4o;YJ -uCej%P{2N?Xr>Hu3_b(7@g5y|Si5FCjFnmQ(y7{;B%5P!f3lM)+S4faR{#m_9>I3LMtE+L-gAsMde$>@3i_$q2*XYLEU1!{9 -dvjD8~rZ@fZs2euj}G^q3Ou#?N7~JHm#92L~cE -;<7?RQD8g&MkgH{Lis!YIbxxW)b)#{7U(TiAoJ`hfjn81n=2zOV~n^#MssrLg6@oDRkhy!9T!t!-8&%Pzxou@zx?ofHPxi*RuLDN?{>ZX5 -WrL{WXaN>_}K%C#4`7)~~)|-wmt%hGYS>tY7_x)LZCTzxoYfCD62feTuww?21^bHXtJmUGOc!h+tpD` -V^sK(768K%~u%?!B#Ej7-0yuXURy0EfVXqgwwztiS^m37Cb)bIDz9?$1kq!=V}16H$J)2`Wz`?u@7T? -jud*>h_Plg5O!j$&ylo>tr+W{k~t#wVyu5kz8jk{);}f52fH!W=gDLX+cDPXNuh@Q80#nFWhNsGPkM@ -^9_+|iKTZ4)TQb%c*rOqLy!kG%DE3aQpCRK3HczadCK(^QC)O9pV1)zM*UysS1P$-k7m2~q@&4e=57> -)=#h)W-4^8ja7YU1muJ`L_$w-N|_sh3QeSyCB>n--zu=o}kZP56B{XB69biQA2kr@wK-!E^ISp$0CuU -{Y>H=5tCUtoWY-bmLUC*v#*U|+vTMhl$4zJ8UAWH^F-{W2ln&_sXvF5wl>Mt}VZnVqAN{*ui-&`N*#b -K-4iroTEz>IbybU!5a#DH`gpuaa6BE%n#0kf8`o^*5*DMNcC<8GQF5c`azHzh;~UTI;W`lH`Qu`s=Ht -)UhR*3KG_l -edsh&jMJq>zRJfAGe~gu#Ode{lWh?CW84`qNwd`A1}+N2C7rM}*x#tN!&idw7&?A2O@K=pSoJS2WUGf -2;>n(MNOri5@LQ8_o46dJZAFXs$oe1G#9Tx&B0tXQGGZ`jh$_UpXukW4LBbjKExp!Vm`?Xl!|D>K+)s -WW@-g-%wn>F$R`Uu=^6T+b0|DC~RB5 -@SPQUw%$ld0t -s3E}bT_)pLAx!#_u2*ZZi2nu6uGWR*ZgqK5LxJCWiMTTq_=C5;(v?>Y2|nP&s?gvMZakqS;~FCT@;AD -^s3F4dK2CxOBK+Q~WT1ppesJS78S9{x2V7TnHN^6R8)x(ktA<)0aQ68K=Bp&HV>7_sRbm;)<$JFYr-W -R-%f=MQ<$FZgts$51K0yX7$mP4I$e1!0R((!SL~7{edvB1@5_a?=PswNu&3x}y#HP^9_dX* -r0?mB)8IraknIGI_1sXK-gPTwCNl0}9N6(Jm3~#zX#sjGG2REN)IK1iCr22v^e{k~|QVBqp-}{QxhS2 -5re#1U}a;MGX%qDQBo0mx5gD!t?^D^-#2=kjKNGL#;Ke%~?*>u34lPabX^Ro+Ntb{VZdy%{vDD#_BWS -oaGzd57X$QsK0?sH@)g)+Z6L#!s0`FBXgUJ0YWPBK&_MBdlg-wT50SxvRpl_2h)C51D@d8T035a&1NN -%;nGe)AQna3Id_Zjsauaeng@nbJd>-#kU$JGA-DQ=|q#f5FvLq&5@EJyF+cDEAxs*Rc6hWY!E}fAcgc -L4>e>^IbC1LfGFtO(s&%^*7J3AC5r3KmrxI{^nUSWrwc6d4~Np4E`+p;gi8P7l|`N*WYYuNuq|Xzqvq -M8M^+4tW>U{>u*$_g%O6qFOoPAy8cZzOKU`KCrgKF2>S;&w@C_uvcG$o{W(m2iMS*5_su2Zj?mvXmq| -?v{r%t;eRHt+3uFQd0si3Tmt@`#1%CGh;*U_^cV8egNGR|JH@_l39TNP(t<$6ig9g9*0(r5};5RRm`7 -Jc~&CA3qpuulmCjJZ!e)BS6iJ`%7USTk7{tB7gLW191C6g#f@S9h)%w0o*-@HbqhmhbmSIKY;34Ze$d -#MQX*GcKwCT0SV-ZwuZo&f26bBzo|kluG+B)HXl=n}pkj_I_~d1{sYYzVE(5>KBObyRVX&0L1srJLI)PeBb<>j5ZM -8H}8|g1o3_IA<3x_-w$qmM7$X4`{pC!#Zccj+r;05`u;It{h+>Ypf4Stgq*qu1!aHpOH!La+24Fh3U#6EUt^H~WqrEt348?eD)zrhd@&SHC0^M`-)|SIM9P -ZU5jl)3=1UfBQU{b3@$Uy-gen+WzisQn5nX-~SoO%tG3~{T#`?koI@qVLu&4zx_O^Iic)BfajeU2UJsT>4ZJk^jNbLKckk=+8_IFN`Plv?5yRGwIBQ -~Ii5 -ZU*?WRC`$f60Cr;`p7Hh|AE%7)7E{N0AkW!jif_D`@keA}?Cj*rT5v9pn6O^!4G`{bRlLmp}U9-2d61 -{OC`<_piVAqp_TN><~IE!Df&B;o(32<J -hMNgJRkCsKxmPHSjMNgMSkC#Qymq*W+$L}qAzC3!qJbJ!7dcHh*zC3!qJbJ!7dcGohz9M?QA~tOFd`0 -wpMf7|{^n6A1d`0wpMf800==tW+^Ub5@o5!9PJ>NWfzIpU~^XU2J(eurt=X25Xx#;;^^n5OQJ{N~Z^n -5OQJ{LWoi=NL#&*!7(^U?G9==pr~d_H@6aWAK2ms3fSzB1O -C_0D%003qJ000{R003}la4%nWWo~3|axZOjXK-O-YcE4jP+3V%M`e*gZ<|06h420qPd+7*gB_o$Hi}A -ZS4EAmWSm4fu3^BV!tQo=DdFFDmbg-)azV?yee=D?#@Nwua}K!LTw}djeY#n%Zf}w0={7~K%rC~BH@Y -#kDwVFruT7`8O-j?<&t?D~xFqeVLa!@w2xmfSRVlB_WPvAgo{h#jS>Y-Q=9kg@`b)GJO^;5E&?cbwBp -&Qhb4#Gtl57BL$UfV$(j{TwaGv&DT$1SVa`7;yphT|Zf|XPKUI-N?m4%l1v^m -k-%ymlwMmL7uKW>Xy8Rqu)I{340o$tN@kjB$%xlU>sEANM^xA;K6ZV-7Mfu9R}qn+Ena=<5M3#^z{!< -Qok2SV@hM57RZV@zNJN0ETYMCcKoq_JmNV$9Mb%l?F#G<*zEIdcFgvc!Jc4cm4Z*nhfb7yd2V{0#8UukY>bYEXCaCx0mO^=%}5WV|XjL -3yV$lGjtLh7Mv(+^c`quuPGConK%y)icRIGfGCudyMUZl!ixgkX%H-@G?7%)sv3HV18fh#cO(da-1(> -y33_h_Tk14mdv4*gI)7!ML{om^~hOdBV8~uzGNtV~-hZWQXKL*E4pHmOKf(hIE;Mzc1&kD8{cv;lC$K -zT}q)V@$R_KSzBptpMHy^)!P3AcRf5q&rP^uF3Zj!L}G1?4nbU#M>5LFB -nVLa*VF$Rw`McvW1e-_VhDAoo(1?<+?iX*8Q8Nq7%5V5&Q;9h9 -a7p&K|@=@6s?XP}JdOBT@Ah>+Jx5aneQP(o`gnAlhT3w6}x`w@}pNkQ0+ygcMeBa}zY^ZA*DYVhz6+A -X@h)iHMi)zKmShxC!gfdf{Xxm2}5pPRNg?!aFkhq7#;Gg}Ufv -y;I1UGNSPrBXL;KY3Zl6*>}BPhs`0TaZ$0KP)h>@6aWAK2ms3fSz9|hoztxW0021z001EX003}la4%n -WWo~3|axZOjXK-O-YcFMZV`Xr3X>V?GE^v8`Q_D`nFc7@wD^}%_N~Irw#03$gUMLcUYh~g!v2g6lkCy -&DYda59Kq3dn@yyPScQ>tK&cTa5u=b<*Z1Lcw(e<=j-c&%(VzIbm4xvqKVJtY)Bj8ALV3~Bt;nYcm5R -#nK@9QdsKW=x;&2S1mJ7@MB4#szA!FNo93vW4#5h?b<+8SUZ;A{e%nD*cew9L01Fr@&kHFlK`QtT83Q -ag`~ugBX|Rh6^c1tHKr1f}YWOvBX=#G2UK;J4zQdG;UjgRwpHO*B=%G8M4N{(cW7QyYC`N^Y_%tC&KB!#9$AFPH&)~er@bkd?cQi3d$O7KUUevIYHkNbx0iVjrAg|LtA% -8VFV%sBHq0DK7zkOF@6aWAK2ms3fSz9IGNB4OK003SV000^Q003}la4%nWWo~3|axZOjXK-O-Y -cFMZbS`jtrC4om+cpsX?q6|pu|Y~5RZfC*Hj-jM+72DIpv97{KvBrFOvl<}(iN#RN!R`M-BFZGTC(G9 -Lf$O$-k*DTOtxIfLc@${uCpcYWz}b;i!#-{c_EjO$~@<(&ZJOamjzq$nJ!m3@AZ0V&Xj^zclzpn#e?- -y(6{gU1L&u6#>oonhvB#f`1JerIuv|WaK(jI!0$9@dY1`Us+$!{|BP^F&xp+U9Fj!v&q)%bxf%dz9T> -G%DLvU7gw&9!OeoDn%7bcKCbYAQo`PkH1G(_3C}a^-UM@?eVM-q8ez-#`nH|9NUUT)BUfSB|H4<&jF` -S*7fav$|=y{jQ-&kID`WpX@5{}M->#`7_$`vnyFxr;1rg~Rcdkky2YHHKGajT;je7%)w#f3s#v?^AW> --Pm0FfFrO!{Vsve*)->4cqvW!(Ui^aceomS+zfm{T=5{R6w%Mn%tf#R| -40xYufZ=v|h}DK9|Io6s=-DK}uazI^i@_)4Y=2;z8bzTo=m0)AowhA7+Kt!fQn4bO$Svd!4r*|)2f5WoYR`Rg_*71uVY-;BT -Nf=H_=J|h4UW}C*0+)7*ip08=RF7^2r*V(T=1AN)^^lRuHo%~b7AQG`z_NI8q-17xoYBw`cYk12bo!;Hb6RbPqp?ja -81V{6~E(PwZ|eNxXsPeF}!8>xnwirO(Q~+99mQChF35$1IyuXEpMg<6gNIPdaLZHaSL$P4c=jCRk`3O -A?=QeZG@V0fvF>0o%k4a=)eraQDr>nfAce*)6?)n{L$O*^F!RM@!YfWy>Erau+x(cBkwwmtPNPf?;yU -5tLON&eRjBIrCz*%X8(?MErrt8ul9}VF21lz -fQ(dpqd(*{HqCaUi#ZcW84PW4FCY2SK`?BMga1UyYMVGE?D*^J!2-PO@swg*hXoCZS$uijsE_7w!$OJ -3OOg{se-AlyFiXs7~+K|?$Sc+!*|A@?a6s@;64nVjK5D5lceV!RPU}}*g`Pkg?3Fib3#TXhS$NlBp{?E3JUVTqyrtDN -=VlECUh-RQLW&+edrFbr1@t39H5=oEX;|;>VR{!w-0aTrKNEFv+J}qLJV)$nY=RwfjhLYgSBkC<*fr) -;C-&EjhFhCZ;Mp@6u2bT(C~aKt{1!LDft^|Z57(GnwyEYvyt2?mVjVeN8(0+)vc};`QjE7ms~rq7-pq -H47PhCRYfO9EZe`3AAMJ&dgv6@M3H6`R;@zsdu8P(`CuCYzayJ#br-fEY4`$izQqJmt1{MSDs#bCDU0 -M>MS382;R^1JeR)m&7iQ};M4^T@31QY-O00;of09jik9KmHa8vp=?egFU;0001RX>c!Jc4cm4Z*nh -fb7yd2V{0#FVQg$-VPk79aCzlD{d3z!uD|=Q*s9Z4DUnslPE*HI<;J$`tLKt9XFIR$^!3q{MA@umQb$ -s;eVzXA55RtZkdpFo*SV+3#3C1q1+Z8wu!~)_TCJ0KJ|D-ETeesh348lk-Q*cd?-ujqF3GC6O5rh|vA -gnm0Z^@0YikQi^P*y9RV*i!dU`CkW<`F-rg4>2>0QD^fqITh7LDNH5*`4_jNQa#TvbIIs&-jx@wj-5- -X}!~Ek3j;fB>3TDm6V74u(&AV#~M@|jl*t@L@bW3lj?muUvj1#ZmH1eS)=yAU@C?BAQ7gLw#)RNi9 -v`U;+J2yw)nW|B@dHif#XeU?Q%9tij(Dh-et!RRT0DRiM#A1FYe;%xG3_X%g&b7*^Ixxs9=QBYlX3rs -G$1g<09!Y9DN}FyT#Vl@lVmk-%j5G-F^IiboK`Ld3SUPF9!JH{PA~eOA*EM@fKbvSR3kL>aDMP3Ri=l-(>L(&j6WRzgk*fXwKY0EIe7Q#652d^cYbj -ky*hn;N=1VKNRCG1N{&k`cckl-C-!@&9DnWlx^m-p9qYv;=7M@$Wm1JkA6@-pm!Rb=Bi~e -~zS;?CMG>?9~;!zezIogabZdlN=@%umUVW5I)9xc4f_)R~rPJW|JIes48))86zW94A6$aDKO -TWppgW@17k_gqM5xa%#Lh3NozAN19Fj@>nycU!F+xf7dTZeU830Hr@>+ -${XY5jzpOnFLx!%e;6;HNtA;!#QR8g%y=hUE_-JDEL{GbXmq?6F2}wZJSPI4hap}xDG^MV8uGwXBCGh -&QzEPQ$g52=wFVe@^&Nm?4+Bi@y77n8x?n6-WgX=gFYI^ -USmQovV)3#_!OX}i0@aE0YC<7 -!z)-t1vq^G4j%U;pJG?vr_~L6cX_hC$C6CElO%Th&Z(qFtVukA*}ed}Z}NLxOimv5w9t_0z79D^<=M9 -3U7IBXD4EUkxMHk*0aRw7SENV46gz%l!VsF3Yp|BS|p&C1)nmMMw=GyXg$G`!jfIF%N|4X!J -q%Pz-Kjurcy@_Ii%;)`0I=NAZlvkZ*+%M;&78_KqvN2K}v!|JQmENTl*lLWFK&-r_IIWDCWZJ_KdY8n -|+Rl?FyF1&1o+T8)FXR0I3=aBzGPm>}%*p-!>-4arE6K1yN9k8O4&X9?9;a}n1dw!+f|Ba$KZAUP@Je -vGE0ig*nxGC)b*%_`>REziT4MOY%6wT&0+S6&2_zUnD4-=YN)|;j0mFbZ9Js|#99cd=Dl3?Kdah6bkO -8V|P+Sl#AU1WworuC+mS&%a7(l!+F*MoPNf0QUDv`l-|U1Nf#~H^%2kdYTs6X}u -D6=&!_>TjddHi$jqyuE8uVmr;HeMLtQ&vYl^4SD6E*mDFe=z;*#n -baCv~K`yxJ3E04eo4jw6(Ri|B&(ILFqK62u{z~T*HD_rIrq9B}Tn}qQ5F%e%RM9rC&ZoMKWOH`ynkK>sQM2G^^XLn}ft~wj8QFr^YGg>(u^x+l_ -Dv-Nh%EBH0s0(ON0y|9()y|GS2t4Su8sFTR(zB42(m? -3NNu4rwUn3Pt8k*(q$y~aCB~$~Yje#Jn>++=*jm8wBf5%gk(ViVsDS2WQGgEx9wx=+0en$P8xI-{ciELA -j?tS4ykm6Fk^*JW1%Xue2>@bv#hvvE?a4#c?sN@gSS-jEcB=&zsD$ui+G5i9%^$mDpcCQqO(1@XZ;i0 -vOLoeq;jVBHr=TaSoac*0I)8*|_c8~a1?iW#C3`Q7ma!Ye;>savl7nO*7P`x6m86Tc$oU2~Zj%IJ*8* -K|=nf2Nsi{YVxlamMGF9vm<@BxJHw|2;Gc-VC};XN5^a4F+{uea;TQiB;SI55ha0T}h&fiZj~fHI!KaN2Xy(E6QIM|wK|D!&HQ*ER -~`DWvB1pyy8kICn^^=K!odXvI?i&K{5A$yYfDv7&r7D-++!>b?+DCVEA*d~Ufm3M{B6rbnY7)v|lQmw -|@O4sLAP09@I)0lQxFhTg10ga%)i6kX#Z|H^3^voJRfQr+Bb>`&ptDpykyf3LT=+E~|VuP;IOV+mum?m?}ZU1FISn=0DlpuKC9tLZKE7ZDeU*g6Y8r1}J3_R6 -jpPjyW`|9B6I64@O*wZ06(@B|qjXvQ#1xz1-un+M1wNXLsMcK)u%&8n9Z`Rk)w&ipx@4*aJVp&tm*k&V55I+aO{X= -Pz|M>-m6SW!@(tGM#6`mbjf3VqQoQM90vwWWNS3F}Ni>)E}<4f7>iGhx7@K -&P80^>`OCt0|9y__Z#Qf6_2LPvf6t&bd43~Y0I~A!nlXlWR#Rv8YymzRY?2)*c6`wBYKIyP8+t(=h|L -;8qk^#n;K;SU#_=R#g%21N7~46C%5f{Y+VaQ$5Au@qP@AHvvS6hIV|+S{Q!kwom~Ops2C~Ksya`l4kb -#WBc-67kv-I^r@{~=>0kNqH)0WQI+Ic(CU$2U?zE*WQ|3)R^3-LwliU%>Ik@$Adoq;1YBiagD>#>3pTYwZLEb0zWVWKH9Aj_`hQK2x33i05WCqaYIN*Ze&`Y( -cVX@stQ2Nb>Huq?{VWNq$LRX!HVK{9{8058wLW=5ALFs!o}u10lzQ>eE9D+@9nJH`2^hm8_fHk#RJ51 -4`!E+K696Tq8{5^;Qv~m6Ay!!VmlA(D9*bb<>hUwL-BO?B`mk0zhbfiN^h`szKRERIkSZWQ~wvHbb$pmwtNhe$Gq6{ -&ft)NPuUv8ucz&4@?x^xviU$V=y(V0yCzE5xr-KAL!^x>93A&)}-#pHknPQ9Ew9`R(iPdMf!rMUZ1#M -yPy&H^c928FcphiO2ROlf59;4T|8He!QFWNraznxM$P-axF?IxU9pqfsnioLpS2on#6-{9F-OgfuTN}J3&Gq<$qT$#+PxMKP3UqZ@+O)GL$LI#=+ekRMN!ciYjrT-1e -gUps6p*HJ71{6$5Zi8)afl76_SG66au|fH|9s`DAlCUFF*dF?IoH|bN?|%gQGFsi2JcX#9&R``aUMCj -`RhuqKz!W4BGqjJXNpkTzOat-Z8QiZ**;k-W`Mq%z4Y1=>2zcA7wZOe9f>L8K9RRynt$_|9kH@k{<2Y -+u*;E$AN%ad9{<@j?T+LjqlcFtGFTKJ{0xe|csL&30}WH`1!EtU8AGk(oW60}I^_DWmaA -kwZ%in%lBU!(6L0NHCg|_$q%7m>1dBvDZ{{7oHID?U1A%YdJlD^#e;=(r$EWBXkxhXuDP%;5Rpf?L3%Sf#z!Y`uf#ogSNQ%KU=&`uK7QLLP^*n8_SML@zRRvass|EypqT6Q5D!uuTSE0Zq+!gc;6)%Dn*IT@1B-*;4Xlg`#prCk2Q -^}2BcVWao-0Ov-=dnzp==R;t}B(GddwA?y*r)r0?M+4O}Bzh@aOL$9Iud6B(VWO_*_03rbhCW0>_MuK -3ZfWhV4E;Q}X1XtdA3d8b=i*#d@rXHKRjx1F#9K9%1 -ug?APOZ5V`(z<5UP0~yS|Yut`c6=ngV0qJ)j#p3O-5G_P;Y<^<=Qjl=fD-NH+dz@0j^}|y4d}T1y?&N -QJ$@;inkr$5nLPt?=FuVGOqMd*WD|5vxLZym1p;CZa2T;LR|&3))g0{Gxp}}lD)e)W|u#lUVQO|ed`5 -$yTz#<5rg@Blm>T#R{&iHE@i+(I(sMWlmf_jE)5cv_S{DFof_EEk#jNY2b@|qPA3cQ-2=ald3FyaE4% -6(JvQx=Q)F5P7oF4t_kH8JV<}AIwdV|)jqwKLb(ps=|ib~vdy87M< -T&_@pWZ#E&oI@ueMR3D-HNL}@Or1>+e%>g22%Zr8=KzF@JZ}BZ!=&${Ba!i&U2!l37U_7KjVlCo?d`V -w<%Shib{a{<(*QvV>0~5|2IT#)7Z7=K(_+h$MoR0jMZ0q0rK0Wr!;ab5N3}F=Y6M6()&m!TMN7OB?*X(cM -wMO%}020|(`jYWp6lL*U5=Aw?7C?>^d7Wa3F0gx8uOJ=|wmiZcWO|Q=;w-*Lv%TIP&F2x|1zjR&3$am -NMSTA`;F0bF}n4L`4h6|hgi$`9qR*Y -9}NIO@}90nMi!%2y40ops2*ZAA8LGi1*jk?n1U?7K!}gXf!J+pEXc=;E&)V8XfouFGBp#;M)LhqoeJ> -%F>G>PE=g4(y9C=5+!o953rt4UdCxMw4M=^qRn?u8J`uSo@2CIbqFgFpUvg{3ZRhYc4u$yD9!_@ke(} -uUJzc9&neEGPK%~V>0XfXI0uApA@%IXB`e5UxNwH@~lCvjaNua$Lu#D54!U-kMsVELl>19x2txoJTQ$1SJ3-8Id6T;}C6Mjo~PWgq-_X$1C2aI* -$Nu>JxurWL`!n_!l(x%8~)_ZK4r8=(a4-5}9ZznRr=iafMfEzo(}_0ACFLZa{9YJFg|azI((Psf*+eq -rBE(%`wfN*fYIo$>v%aPOpld{W*G+&DFLLa?A78p#I{px1S<``GyKzjRjrYE4M+oBxO+CL}0vk0N2>biutZLnoqtD8vjB%+GS`H>Z -Vpu0qsErWTi41BZf^{N++6Z7B&%fT|U&qsL^6=~SLB4C@1t9LWgy|bM`}B2q_dLAxJ3}^N*Nv}3_QFS -Y>L6QZvqg@X=4!^hDQa7Dt@(l^)OXc={k@?6!;3!N0M+RP@CF9bX8RPQqizH0k3i2Os?o?s?+^0>x>D -L%({jYTdSd!`A@6-oAs>HUA)l-g@(6!YXSIw?mo5`dT6g8Xe7)c2Pxu&L-geo03^56Ptv7w!98wyP?` -tAH&6}|Z+@wE))kZ9!I0vrc7>V*$+6ZM~tZjya3oN4-UB9VXEnQnZ51O%l2xg&~hc4#uF<4Vq&tq?i+ -v!eYSv%aFc@>aC?^ETaY{&v$o+;nsI`Q}}Dm}a$Isc(R^Z`V8*>AFrLok+1Q -F%gq{o#N+jjXO3&92wz)YMLkLhMLACFsJ37UQeJjKK?s{M&hUbaV(uhM=3--6L-+HumTlD9+1q7>!m~ -cK>5t0)ocj*KlFpvK_xJLt(fQ&n-Gn?9Tv8rytDOxP)h>@6aWAK2ms3fSzF)k(!75F006}R0012T003 -}la4%nWWo~3|axZXUV{2h&X>MmPOi4pUPE$o~PRk0yFc7@wE9Tro|Do1`f{0LmAh8>=kY*vf)!#S0=% -u%LF*DD^3YBOdUkMGbej)g`=j=l!Y-OPeO9Z33DkP#KZoGCo-i^#JDoX02v^2Qjb{(htJoUq47%vl~{ -*=>v?lDTS%3(8>qA+PK2PfRobP!A3uIYcbcVAFT0|XQR000O8%K%whu(zWg=?wq?z9#?xBme*aaA|Na -Uv_0~WN&gWaA9L>VP|P>XD>`iLq$$gMJ_>5K|@GI-CA37+cp+{&#%B}W@qdfMoGKv?zVkWW2aR&l{1p -PnLa`!C}BeqEI?Y;{`x%^+$hOT_H{j()}o1ngLC=LcMkMQeCcP=%W|*82VJVpsaG$&`7dRiHl6r({rt^~gi{hY%JD8V_T5naYtcB(D*P1N;8MkOnrQv9jvMM-OcV#n@DIf_&X9m&eR^#sAM`!BcfZ^q6LvTt3)(^Qyg1i` -hp$Pe3>)mMS@8Fh3W$gfgU@Kjsc10nPxACkM{a3r_mT=;T -SOi*mV9_t*ih@9;I&Nwh8&4;ZlKUiUC+kA4)RfLO?e#1H2sJjV%Q)Yci`G2COY)p`fi7aRmoZPqOV{X -IKc(4q?s<@y$iwE7tux+W%VOi@lgphh?|{80Pb2=C@Ro4hcra>_xZko6^OW*SNgSf+QngI&& -jevZchr|L=i7vj`>-g+3#-ghM4R}HILsbhYsURsJ2v+cZCeT~29@zWfLDVyyV4d@dt)Y7O23Y++ZVz}~5BuFxB8(IZ_e@^)#ONs$e-@ -w*QfNeK!0XsZ+1l~c>MVga(J{4#q!tpa&5xh+0?!{`66BR^d+8*P5tH78wnBsZz@^#f-%;K45UKZE;2 -+OcYHIRVT02mkOg{^ETjt&;rbwobK&Rbt#7UV`($xvvg!KGv&`@7=BQWUW(-vj<7od6MagNqxWlprlD -S&<64qZ?G~atMc-J1Q(8CDlm(2D=|CAG4LnwQ{2u8H`M#AhR~w4LEK&I;!%tV@r7HY2+tDyi}UhQ!v8 -xtRWuZD6ZG#XhsLg+e-Txq0c6E6eyQ+de;&5Lnxw;$gme95y5J<0((o5=jwcRp28>mIvzV@;1?-RZa~ -Dx_GDAB2W7Sj;t43DA|wZ;^L!h&iio-Yr%>TsC^?L(Bl)(nMbjv-JaP|$HhbS-*5Yvw(m>mY0|9~H5@pb+)Zrmm6aK@m_5z%Tz -75XIPm9TRcD6J^f|31-5{&;;zGu&eC*2HQ>3fz!##h+ST_vXL6>&F^N111pk{+`17rJIG$=q}7z2wFU -GE9gWhV_T`Cx<*uO(Hl6Yn5#SFPNo4i3>~ujm5O=t~+JgE_@}|o9Ee`04Q0wMmS)>1BFXRBJFs7qZ2E -AhyCMF1=oluw-C1QlNbII0P(xbv-cl -2az<9K*4KKU!j{Acm>fG!g@43eBj`l+b9T9P=eFuqv(&n4;*t5G&FJ!dAh<5PZ?PAPxJ(ilN}Uxmc@w -GBl^HTX{^#c8WrPgF0}V2_C(`^2jAf8)35%K#t-F84dj3#ELv^o4JHPgJeZ&g~%W&v|6Ru%e85sC*t} -Zu<;ng?4XkB!)Yh+6P3}h=Qi7u2RDgIV;_b6I}dd%h7h -z5y2Ci(?$Kr2BgDm!Mw!^U@5mU6Pj)jlz0%5UI$e -W6cOi+jpDc$_pcm#1lpHoQ1pTyi4J%bEVt};Yup~~Pf(VU!dc1VSbPjX(Wr$(!;5wxQJedC9&o~`cgm -+x+8kbF-A*&EZ&?}3y%+b7Si3<54>%9&-!Uzw=TdP?x2tfO>5oB88FWmxP8OO(m3vW#pK>st1X-Tf(*cywb;21enY&a0&*d9 -ioLgB2u9r=@d~a~UO7%&m5P^UWkevkrZS@)td5pg4a6ShOGZG6Wp$Wa;a~FpAb#CU&-$Wj`m5xI`=pj -<`iVS6{j?XPZtm9-hAIpZUD*WAIzmrE~CMBruE%Y^+P_XqKI3hn2^;04TAt}>3~G_uwS%e}zh(`fI)l -@W`{?utx(L7O6_O-;?>@tIa6pU%ZpB6+xVHqA>4oZVJUs&Hi8I7=ofjd3&?%fwFQq!XAp(hxy!hN!2UV6^90=Eo%NYjxQ(9y~eo~ -K>lp85zEa?{JhvKH^I1{miNEjJ=}{=i}iZ3x?kRI#N9f-mcM&17OOwRKbNZ;pgWpxfS+(W>e5Lx;YyW -_NmB;VoKyy!Izp36rlN+(ox*`|bMOub}1P=jDgx{T~!U@0a -(h+YNtxVBuwaT&%%i4<8n5@$q5(@osa=)s}Ay8@yQn{d?F{)2##D%y3sY4Hxj7HNDkH&*{K5L>In7@s -|)co0sym(K+Bmc-#Qj+7YrlQ|j0zGpPG}R+`4<7p>>bEXUqI=I}X~6c>Haa;qD?Z=`TG`6obcTZ*ZFI?+W5k$|pXkmzGsD$EpQ&p7vEp$fS|F#_|+V+6XI4f8jS3j2~OvuZ?tqxb{i5o -x4#ylA+CRnK$>8b^d+W~Zc!4Vr%6#327V@tJ(I2(pE2P%-^n(0v`wXIXjgazFyrdN=v>;CkSLJrXvLJ -yqqXU_%ukL1lc6ieW|fDi#rfQLd-P-@9q5P_NH&dR{*XmuMjN9p9Kg4$8?39Bz6ST|zfem91QY-O00; -of09jj)Pzy)90ssI&1pojc0001RX>c!Jc4cm4Z*nhiVPk7yXK8L{FHA{8MNU&iE<#g8ol;M4n=ll=`% -^sa)JiDRcHB;mae^g+nQfBhNP#q#2(G}i>bKvslc?!dY0@|(5Bt5}zXx;AZ=-!Wdbn%uA$3FDx1jxeI -C!Xs7VOa47u(Kt4~-lAXdh?q#%V!l-$NXnVC<%DeCgT+A*yTbbndclhzRr4Aqq@x9-A(nJldfi-{9cJ -)0Dug^+#}H{CBe_C)e78ZR!A!Ks|Qw+>Iyeeb>VC=uiw6{#bkb=@6vv-K!m*g1pv-p(zfXI{$+Zz~|c -nKFXTF9jc+)c%lE&vAm> -dWS1stYol)N>a;`Eis{Tt3oSj~wTGz~{Zt(SWYhQ684jV=|)xH~T-B0h~$BhN#TO(-DWI}_ME3>1D1J -)7;G~#iKLe?@;(M~1skqdhOcxfa$)vS$V!+ZaIKJxLP?botjei^vQnkg9MBd? -g=RSw8=fwZ9T|add0{|jG|xYE8MJ@!xa263E^|&I6S_==;>-j)?-j%3P+y*aE;$pS;13+lP_<7kn415 -pun8|oN;hUk(g<1Hqw^icq6ninn#Ocx5v^iEQtbn1$RLKFz%Ju>u0o(ODH4rRZ3Ou+Ky%{}3ck+8ni -q@*H8P;=gy#3SE1^S!IMePZa#qoD*b%BCvAp?WlekF`E1>CC1lgBn+`Sf;?<7XXGC0S*Ain`nO9KQH0 -00080LuVbTYvn?TrmLv0PX<*03iSX0B~t=FJE?LZe(wAFK}UFYhh<;Zf7rFUukY>bYEXCaCuEpOH0Hs -5We?U4BhK)yB56&g1Wm0VG*I?wPc%VJD4URnJg6Xzne`~^fctl_n5~D-Y4=fkV1%62BAPvkJ_Vzpd~u -s34zhg9?Xy)PgeA?95ZbVALu+8U96y~wrdbtCfakm|K(KBA|~+ovOB!Chijltm@|g6ckIBSHJ;eGCHf -7E}FbmNa+OHlZR!UzfqX(LzaRISd1*ZIG69rnKAMXm)f8aln+0bDSybWYRtQq -^rmz#w)(viJc|O9KQH000080LuVbTMqBz -0ptV#0M-lu03iSX0B~t=FJE?LZe(wAFK}UFYhh<;Zf7rFWo%|4UKaCxm(+iu%95PjEIu#S -(#OMawvqfk(2ac`CwDh%;Dki%nV)Eb&k(Y2;(u3j7kMEaGc -u(S0t0^4<3Ocl!P%=irn9$5|SYSeDexzfR$>qR~b!Zc~eOQoVTYO_ISuy;UVKRW1z` -@_K=9Snwj3M}k0HVdiU>;jhqxdD^FR0g36VlMIrmgXX&nx`U2<-83sYtQM7_dm45ML&STIauf6GOZ#Ni6_IKSqa3Gv2O>{a^Hi;7O3OUd1&&^3T+CM(<4ebJir^#fvG_q_9(jM!IEP~ -?Q|XT!f+ItNjp6vS`S;9P!C2^H{4ec3_qybF(kbS{KF+%pw=Y!FZC+<;w)Q%=^NQCw4HsO%HotLie~8 -9C`*71c*fpMA`}9dB_|V_A4}b3uEtg^W?UM}cb7#L--MtY_0c@gf{=Y&dL7_%$lu62(N(|Nxb`k-@lj -4Hw>C{trvMs_!80EX!QiGcKWap4vq@rBM3}X~b8PZ|S`lXi3jfEMQCPJf4VQtjxxt-yW>#wcgA%tvo) -B%{oBUBNn<4DRCRFB%tg-s~T9Gh&#N}Tb~?%DN;RSm3B5<2H0GO-)-to7ZfKiLrBth9NDM(($VJ%-D< -Z_r>aFJ^`Bh&u*ed0e)gKTXL-JU&H^b|+NXH)$&;59MIyU3<^wez;_v5%6_~7lLQH%6_zh$WQx; -q$sa&~hRY{lzZyu&));hT8pTk*P8e%II66@J%tlz3=2DDj5{dZ{juhJ>7XV@wVy`^e4~r_ -)l0K|7s35m@~|331K=H{3o`xb;fv)H4;xg!38z~f2?tO^4%Knx)QuB{U(!XtMSS%k`z_l#th5xf@>(! -%}BU;6kA|kLY~4jTVMiM=u4Xjrf|QOfv>@_n9h~Ox{_&#xu41%)5=YSB=+O4V_9lgXcV4`g}lLuAy=k -*6KUgyUC+8)j*a0zZJ9=k_LQdGO4pK -Rg|iLsR)lRS_Qn6PXnQg2!4RO<(cPx+pXNNN+_cf=RZ(O0|XQR000O8%K%whm|;M{Hw^#)d?f$?BLDy -ZaA|NaUv_0~WN&gWaA9L>VP|P>XD?rEVQzVBX>N6RE^v9>TIqA!HWvS`zXGGov{F~5CHY9=soT1)oqA -TWJ#pG@$LpaX5|U8!Pzf-$HkpT;_&t$w#GtGFCq?&7)Ca -TwSYe5=CBCu=aQliC^j^w;pS>#Ee(=;lNq~%j4iYK)x7aw_+$YiN{3nD~s*7#8n@k%5j<2ubo>_Qh$m -pm5X+k72~5xWt8<{}A1&q8~?h!zVOi2}jd#c9Nq8sXt5t~vhB&d()5p1tP5N+y6jJNHlCy!KCiym)yC2E?v -ld7@>+t~Yv>CT!1pIbtehO3UDZMDZx~irp)n-2)d_Os|9mTcjDgUjy;I!+@Q{>-|uv#B+}^Pxw+~y-+ -Nezvh_|B&Pwv9^s;F$acP~WPR&o#Qw`J(*&ZbzVD8D3?H3s>;ruaed$^x2H%MFvw43=gf6Nx^w~%eg@ -ybFi=o*RPk~r#Hn_Su%Q8~#TX5zcf8|kbpNADDi-l-`Yz;qB`H@WWC!c5WYB3N`BFHtLN1|eF4s*09p -0O3r!ktBwa?PMP7Vw141q+3SVT+JiBZP=SC|e00ij1WTvy$Zr8LGpx&eA|ARbbV6t&Vng6}e2{g=wHX -&;td;yK*IA4EwvwUCtr)y{Wej7rjz+HHB2Pw*{C9cxS=}4T&z8RK}i+IO;n*0ViI-*F2DV!;X&&NAOP -frX45YLTO0GW=@i)cr!KcI2c=}G3T5Ec_&%iadHvyrD9*Pws#$~mQQ^VY{S8hyEjTi!!6U0hc6C4V;T -P+y13YTakwpq6@Mpum?9ihf+3&JWm_5}#?F7i^aQ^jO2p#Vb$-RM|~y}2ho7xjJ1B`h4zJ%{Ty*Ts__xq!2h103#(s~_qEJUKdr1oBppXv)ynkLI_T$xnCyK?D6uf_#WH -Uk1Md^Sj?$X)Kx%FrZ=iv?REY6ncbL(~aYJXs3Gq*g5Ee*grl3b}X09tBGTf|X9$Ly|tS#}%R(g(#Sn -2BKId)o~M<9=M5RhLoppx^`UoBgMs@!mC$S -C885Zj+51pCvun^B_g#Nxw`z%bk|S>SPCTnY~^yIPd}e&77P&SVT$sd|Y-ED~FQk!redvRu(O$_t0wY -an}8+o9<6XD%UB&gRg|0)_$`)37lXP&J{vpJp=;Ro2%~NCaz%NaGPomT2R!!P=n5d6qs1kd%ROrHm@y -k`{3Zbhg~kO}T<`!nR3O$cRaZoC{p@jMQk6ts|~s4>60=Fpq@S%A-qYI-i^Lxi)ptZs7jTqjbKz2o{r -YJl>lx_7~&HcpU6PW4i@TZ@qaon>;_*nN42oZ5J%h*v%@<;RAD$h7%=m%KFy*rrzXv-z(2Ig??019_* -?T@t}WtB~{%{V4l!y2P3v_eaB@)ahMG#g0VRdtq%@8&ua(qk|p=sG4g3AF{b!wA2~;?nDgM_uQ`b`@&=}cR -TTYlP`an&4weB#a%sbHyk2V>uZ$jOz{n3 -VINt6I+}!2L{*G{N2b)bMX4bK?gmc|BXh!p*_9}vXuw#z6N`nB@J(7^D`f+_^$R0UNBqZ^~$%GK~5}S#X1P|JzSf{F?$zV)a_leLC5H?7L7u -s^ZMM$7+o86)fw5#bjn09f{@CX|vN{s$8$DPB2D-v{lP)BM3O2ltTQiN3$!HQ?goDAOvT2KS;#a15ZC -Q+2(Q9+oV#^?;S1ZrDjc~V&-AP8Y(Ojd4Wt_zZ=RRW3-2_FZIm{8LgPDaQeG9ufc+o^6S1r`gQz{Cps -BXt-v5n)k<%mGSVk_BDw)={nz<209a-w2&s -88L(r!oUp>v-~KOfCIZ=48ic~IDj1sYMpG{t~p#0DIzxFXXOea&YoLty%JUM#4$ZEEA_psf_cjHbG^D -uhdpt_eYtAJ2*mu^LeY){=>(cv>^k3c?f2)24UxqJb9op -Q@i$nA`;^ZTYQRE2eGqbZlyl7SpF$@pUFeZ8iS%2{T^|z)Y`0jSR^YZQ=!+(2bi5M%E`lGrD4(n -zg*9K{B5bDuGYr4=sds^E8RnrDdM>I8V1rtn)R?#!pqi;*~5qn>|`$MgFrOz9h9LC}|dv$XU-WWVY=t -ZOBX|*$K2cw;d19F{hSy=9i>>p|bp(;p5;tz{w*%dqGl%89WTWw{@4=-D|b0!q`!KW@~`$nX*WE -+k(+Of^;M>@_Y1ma$>Ofj{GTwvRuje0zO%%?>7eBlB=L9+wlTY7S$j18(}_kQ_Ce&({1!ZIr#+aJVg; -g1ze8wdnlt-k)uyn0`)YNL~8_on*Na6_$Wj-W(?D4(z>eV>gbnN&Tq3^6y&bBLK%<^Dl8~_%jcT -F`olUNg&v*+QJ(a=U<&ozj_*n--%jrAysI*1kl`@aIf+fh#H7}&F0zEJ0?EGdp#1BK~`9n={4tN< -n-kwBZKEr~}i_2gBg{ -SLsC<)&}RKB`l&)MTHtHuj#VB*8emlkEgSc~%hsEOpTfn@5CJfGB%OxsAtZ6pVEBnJ-1WUs<;&_FQ({ -0CQe6ayAt^Q9VmbYNV(sxvEP$>H8F308#_Lc=QYQW;@dr98{1yD8^6wlX7~dZ>lWXv~R9O|7VCp4)CSh*CwvOn12(E!Odve#xB%4KymPx|Cv$1zu`(&HcTNL-+R-QWp|NY08u -Pdpn`Nzi-mEb>6hpW{hfv1e}&Lu#7TUi=xHfl2pBUnAaq|Q%R^0rfUn1*a;H0Q&>xuO+_`UAXmP0zFI -)~@v64my;&HxQ$W$GMj=Vx8l)XDs(ej}1DvAjd&~UgOeIn)+!J+PpVqU6JFkN{~WZ^B -uCU*!%LWfqOCWq29w?B@ElUv|;BL>8A*GR2>W}^?shPeM^^9W|iR^=f3Gm_!iJE#{9`QgZyK5+dy?!O -<|}RcfGdxwoN(2wTL -1211UfMOZvakHdVG$b$$gPl+nqwHL-bL}Q6Nt)0QsPW#yA~BT^B#rg7Icxa{mbMokhsSfG#nU -oIB?_n-ln65xk^hkAal2D=gBcCe`u2wb62lQe(s*YZn?=%xsviq3MYn{G>lZzb0WciTXJ*NNrtNUkBh -ikwP_{{v7<0|XQR000O8%K%wh>DV)-!2|#Rh6(@xBLDyZaA|NaUv_0~WN&gWaA9L>VP|P>XD?rEb#rW -NX>N6RE^v8`R&8(FHW2=jZuinbYwBJh$hK41qV^7y*~tpLff_wl9%KW4k5Kj2(z-f;?X -{MA2R8@5HyWf%l+^cFgV+hForkt}Y6e2r@VFOo2oGqT47;IsSQkl}rb*IR(?VG`z3(lSC5J(=k*PLbc --rrhWSDDuU3p`a8z(G|Z?4~V)>P=GHFt2)w?A|o>M;G*qZQuR8N3%Q*0()fdIlfx?;dr9LpVi=s%c~d -g{`jNy?H~!BuOeIoP+WYy3ogXpX}%;UnD@kR{Uk`0;KkFDD^l@k{TN*J7FD`3`#kGl9}FjyTy;{ -6l!#7JvIm9%c%Rh^hBprf@H={T#7On8rKC|f`gs`p~6@PD9BlzkmFF^OS5iXo7k241!}lASchW -y9)O~$?%9TAXQ1wOpcThHConfAQ(uJa->0^ -g&2bhu9z@-SHRom)RxOZkS@;+wvVP)HEbzR)1u$=X-eMXVv$TK3|CijTS4~p(^Yx(Dxa1{6fcU);&Pg -OJzVa_nw=FbybH&V5QnF?ha*QGrQd_;@HAaBNbNSwK{#lRJA+1QBIcmktjP-^2-=Q#H5`FY6?Pu7KR; -Ys%eCP4(Bbgh+9rLLk`X$=c1^n`=r2KK -bRX?J++S$XCh;+ZEAGK2)1d3D2x4z1X=-*lO<7T9f%yvxlJfiPSTN103tpA1Dry0)n4r5Qw~e&aP8emk7fLY=!)eMmq=n&dL8d~T`3dqEc2f-G -qnR9!G4fqNG@eC6+cukU!*a8sEMX&`W`G4rWV4B+-fYfe`(i&cg4#phawnu$p0;#l=_sn8!LK1jmzdm -cPmrD;5gNganI|BBqZ&UX>=~Bli7L({MLjXlh4}Yn9i+wLD6XkkPc{SF7XaUmeV^$*lqt&10lxAPP&^ -Yt7=~&@xniwEW=y+yDO7b{vqJAM8zU&f{0!z7{GXpo{veSQ*`(9;%TGr;ZXTb@Pr;s^eB@{P^hoG>Wl -7=lcgK?2U0Jpid0D36lARU1RxIe2gzv*pI*wl#mC=pd#l&zbS=8ff!RU_p%_5dW474%pq)=j{@6aWAK2ms3fSzFA4=io63008YL001BW003}la4%nWWo~3 -|axZXUV{2h&X>MmPUvOb^b7gWaaCyyITW{Mq7Jm1y;AkE!rK*8_HR^Vq)R`LV+F9Fa7YG8CmFYx?ENL -hyX=d8}?R$8UNJ^A1!CY2AVu|EAJlBthlpM$TfRoJ=jf#RKD8n=-6s23VAsI3VnYUe1R9BvJ&h00NH&}l9u*}eV2ea!pc68y9-`bjZ+_?0LBO>L0=P@SOhNF&?en -f5eebQK_(DuXI0ntT7~?1>pkOhrd`LDEN>gR1PZ&6Sjfu-F<_(YMER`*&7@#o(&0olE+joSPPQT{2$o ->G)K4SVnDCi14<)L*`>Fo=FZ{?6_L3S{PN8wFP<%1MHsvoPt<09SF3>Cc!=Nm^2bZ#rMS(Fz*Ru2;Nj -_P!(xfq?#6am`G9#<@#H%lOftYsc`RU%0vv5?{6&7Vpds|WI}pg13*7^b;ljCpGfBx>M}3ot~Q$lrOU -fFM7k*MG_djF%)VX&yZSmXSOJ?JDGBfM$cfKTjAIY63f{X6!sIX{dom%N@c6G6sV-gBq(OD<9D^Ci%# -KG>-)>2aYy_8VrXD_x6dx;97`OcR=)gUcE!p_mla0dO4je`J;#peyfcQHUw6SI5&9JG7Mlb$;%yaozE --B>(${WARFdUBgr@LS>e`c>Zs`<;6*RljGPRI5BQF7HV7ApEkSvK0rfjpKOvG=&keIG5e5aQby#Mp`= -~OwcviOd4zqwf^OaD(%pk?mV$*K^(pb)9{@ySSDBo!YDTfTbHcl|3_(Dvfo<-d5xT|G0CDnq^@9{7!I -CesQaOCpa2y)vQm(?6#wDB5UFK5X8yfOv|Dc^bT>M&|v-<484l-)MDq~>*u0e>*B6bh`Q9%@et{9{eg -Ru6`I5*6L*dZ%l}IQ8_k-r8>TYH3G=%x2yq@^E5 -&w5jip7uzzVANXm&DpAxA7+;MXK&8;EFw4)FDS&FwmdcU!Itq;+dqF&dreK{9Vaek!Zr90Sd&N?4mG@!P2e(ir@Lq -56{^XDAlds$h{VTn`?nIzBKqLg@Sahpfp|CEIXPdUR}1EeyOV$cQ(0P2j|Q2#lI$N^pt03vM*qi??lU -T`pbuUkhT`?bCu4wBjcr7H@8B|r>MBsSAisq0iu#SQI_Re9cc1a?S`CM2c~J6uv&;~COfPVWE#KDQfs -ZpEp#$4N~Hx0n=cz|PjV=y1v#S)JB{I3?s5haDj477b7u&F%L;b24$Vlek6@}0)C&-v;#cD(NdIB7nyz?dE>}f>(Bzpnh=kXM -H5SDR#BIA##?>%MSMsHx)M9V?)IP;Qw;JY$6=!udxe#XvF7po05ALV -B!#is%GB&*{|8c-*R|*v`J@eGky}uG}Pxorn%F)9cqfe9NYPy&QSIf!e^gqFRIiAj5ibn0yb8v?8a=V -Sbp%5o1P2y*OC{%*rBBsBLMqgNo^W`V@n-PkOl;(}P#wG6L$A<3OoIL3Tp4DzxERlZ+4 -y`i`=hNhOG6yjWps?clKAhpQ_55O@P_!6Zd=pJHRWw&3n}-Wh>EG4Qt52 -hij_e?ps5|M6InRYA29ZHf1!R{P3w^@8-+~iI5g7U$AP+RfU$Xar6v3YvFFObz?TW!P(0qo56*h#fiV -Q4pZCqb_G@!So-x{BMTz-#;mpF--m&=HKbjQS(lEZ%R4U2o)VAu?<@1zRh4?}9MR4Sy4E#2_A9pxHMb -qg=3}^jDgIoczWy%e=ZW+dLWYSHwFss5N|mQI7yKIs)%5R+{?hoR#XVaeh(uCE04TXT9=XBrdRefbGft1BP*yY)}w<#c>Ln*c=o`+BjSybo6E<#hf_^NguoL -rOx+$wNld5Yi5cut!rsz1)HOEXE<2WdP9O1ddl_g>5!Nwmnm3&&7SpO2dv>ZJBp`WxInZ@EsYcqX}H) -?W6oUfmO137g-`min10iDchPSGBntCHmSlvSxJ6S -dl5`mhxxte$=|^3fX~{RBTF_+tQM^ZB_lm7=$O9KQH000080LuVbTQH;Yl3)P<0G9;-0 -3-ka0B~t=FJE?LZe(wAFK}UFYhh<;Zf7rFb98cbV{~k96Yk4ECk4kPC_6=B(xtq)1U&%%D~o8lVz^&>Su3+`qh -3JayBiEsNn3jTKr2Gk8oMV&E$CWv{#UOHBehM8CLMzpb{r)p)f*?pGw-(ka7fR4+f2wd9fEyX+z?efA#v+R{pTpuAESdImGYnE~6oi*>7<7^)w`Dge6pfNiP1 -J1Bmd(3)N#}G7+I41@pbuY|l7}6%mLu8+mlk^nycvPtQxU=nWr?d4BcF(&i6@Nti)O6XWHGtPy{UT#E -GG9jBq%J-h+a0_V!=uLU{{TBL>pE)qb`r|)YB@O|Jim8J=wa&ooF|_&*|0;33n&m=(m_|-3alIc4H+v -kd91QY-O00;of09jjzt2n+j2LJ%R6aWAt0001RX>c!Jc4cm4Z*nhiVPk7yXK8L -{FJE+TYh`X}dS!AhaCx0rZFAa25dO}uxM(^Pfw3n2SX$CxA}35SxPaSEfhcEm*s8@zb0=aGzp6v4<>a(cT+s%RFkkHNH$K1e9&ln9U)qHpaHp>?S# -V3#sL-V_EcY3F)epHY4`{c==OposPSkJJK-RZOVk(^vVdE)u_Q;AfX235WE_HHdNd(0V*B%*D~(S-z> -=qasEc4YpCBZ$*CpOM5#|ssZS^v5of{G#(94w5qIa+;M&z#+PMQuaJcc_Kkr=)XTjuZ=ubN&j8@qkQ; -i;zJ*f6BX$FwxNL?<$=|>HZp?^LLE_)MyH2dgJ2UDj*gtKrRy$jA{@ZsYTezV}K$ -1J3?X)|r2DmMzDYHE&?8UD-76H-BNeu6@DwU=%r=bcMsQ5YF}XuAn;*Cs-9 -EQ@L@QKRLFLbfjF6>_hm5aw!Hr0^skAY+xc+Y_n_l|;!TG?SJhSO#=kbCh)eS2U?1nvHVT1Bh;pf2^_ -A%?alng$M#uLBat9#SSarv%*z0*J6oc^#hk6v)PY*paxPo{(M -D7c*X=Yv0j*`znn$`{_ERTmnL`@LZ>?49|;s^kf3M=K}p&z<>cu<({8rCBQZT^G5z<+DF}uc+Sv;i>G -us(D9WnJAlWUr!jPl5uRV`G+e~vuk^(z)kpMQX*)HLsn -a(s;bO6zsP?*uSFpu%9od3LP83z&Nj)I4(NzIOxD4kzM!|GI{rE5;MVb4j**j~mEm@Nndf6N=73zb)0>>v@=|jX;}9FACkzheb)=p -ntOY(>jt&fH9VAKY~w1EBq}i1^7l$lB3@qTb6p(qVS`x-QlW1suIXZIoGct=QMvPS;Uyoo7{ -~pQ+L(T@nL8l)Q)SxoHbpV6qkE4r==*;==^~?~5^924UQyB2!m;yX(}%V03&|DDN`*w+Ji!6SeUNKNV -ni!5z7QA%!3BFZpHLJv@G8YOw{C5frDJuo$6M(%avl0Af3`9H?e-1&zl;vK=ih&9-q>r@GaeW%I#m7r -&}z)#E$!!c-ByiW403G?-)!6a;2)Ud$M(yr4L^6Tm0TM#{?}q)uUfI_ -QpW%@GC}TYO(g<*^!~3;eNcV&ptG}h{a)}>Z2|HLKCo7)rwjrTSw=d7Y=3dsgya@D6k|yR`fpUo^0}n -eR&PJCy#nto}Bi9QUmm^6{juPtW~Ex$NvCOO9KQH000080LuVbTZ(}M@B$A201+<$03ZMW0B~t=FJE? -LZe(wAFK}UFYhh<;Zf7rTVRCC_a&so~z{<)loWX9ob_kb!TpFD2T --A81V>(03}^i&Ga4mhJBLmE&vkzI9cta_d_DV#bUAFyTBehUuTM~vRp9upVmBQd6tMmiImkvDk`Sef@ -x806q74@IVyPq|K1z08T*^4R3?kb9(x%b%os0H0{he14-O_`XF|~6`;&Lar)S4@!GgvrDYd~36jY*=D -$}ys0=|{3SRyxNF0^2&&^*gkG?}a_xnXgsaCx>UO2i9Qp0#*T{ZjJ@M|&M$M=`1=g_!t)tB-Yf&9(xtZDL_nCs3A4*^)&uio;uQFU>909$>8=A_vVzEEqrf7LBitM*`)1Bi=-?_VMEVZ&0NA2rG5B;`++>w3Hlom{u)g~aUwNJDShSzv({eKTV3h*sbHZFKK+INp= -Ft;84lI|;8M~A+Uru_P0W4M#7QqH|H07EHGZy4BxsFlK@ud>K)>$QphUMg^qfdW5{&W@}y?_7V^U3L- -w?adg?BnrA_T9njsKMiw91+8l7B=<=Or^EC)?!m?a4j&v8S)X`0@o~+LKXX9g>^|vmFY -)l=*jyLLKx~}Z)w=(rNI5C?a~Z92NQQ-hd{Z?Qek}^Ve#~5NR&rQ*th{pMuwgXzYqt=Hs%a;$ssuh=h -ubV2de_1YioBRx{BDv#ePz5^%}ljTtu|?2=W`_ggf}@|L7tXx?SN_^v$d2fWL$#$m_oQNGmcoK4a3L1 -=v4CIO-rjZa;3}{eaz=agRm#Ak}c+4P8b7fYtMIJ4d^jUq}@p{6$0nBt9Eh<@8*{MCM*cK5{1Tl1*O(qnbQ1c* -;K{Q1@-bGO*a>3OOo|(<>W27XgmYSCl#9tH(NCjR`&;JjknbjYo?8S#O5?f2x`>d!@05g2~oE53ORV| -r+;WzyMKldiYy%ovd?lBl4EWQ> -m>_t&2RnHffhd*qRw|AhPH|E|hthZQN#g&Uk*yx8RD3B(GCrSBRxuj)fUGP7_}hdenpIEImo-QdNDVkd5(si`g6G(cctBy=KqMJ2(VKO6 -_uS@Y2_?a@fceDit{6coYmg!OVpvi0A`w;&Xzr}OE8yB)B1(;_O-4z%`q9bXtp;Sq*X8n7RFJ> -|UDlR?J%q3#j|k(JTi{bopPnl)<=afJY3?ek6lJK|xowZb-gFyFEQ$(HnEm^XTh)bB-?g4%RS9va(%0 -llfPqw)WEkJqqObxbycq8<>n!htYtU8}gG4tbQKd{PCY@JdB@`<1925L0d;S*xcQc7oH<8gqV=OA_z< -@i*loNr&xj2BlQP-fnHYvbgz6M!H$oH$Yg1oN1sFO7qlHD!t6EG*u>i(wA96>5d2Y)+gMY@toXSDcnHf%)PK`P5u`>w;257uuw}P$t4HzZ -@)Cm^CjJ*9~5k=8*2{w@>YqV&90FQ*#Hs(x-+ad$2sZ4b60niO-C2|m~h5#vV3z`!_(c#<{#LHVusv+ -@G*I8-NQ6(=_Tgk8LBB9J*Ni!2Br4p0;E1``U6++gEg;R7$F5BK?veuY_N6x6i)+&bFLgHFclVym)ex -0Jyebf0kz(DrtvwUi>dDdz*q^A?jbSs^%DRjtdQDsFTuXC$ri0FWgZ&Y^5#KKu4eOe#p -)g=U^oz`jke2c76q3o&31PsV^{3&;)OfVMfxBJx{aJtg#U=vn&d40Y!ETn*$&f0G!ZtA`lh}1`z$>Rv -QkkOMRqklX_s02VMF2L*Hl4?Dxc4j*=g#jkewtIlx -Ypx<_eubSb`D6IJ4e31ap2kby6I5$p4mUC(sRdxhCO}|}1-UhwUd@MHM=FzjX{MXgCt0I_t0GFHAPZ& -7i%F>rEQL&tQ+GcW$eix~Wy#)XFW&K1NB9}{!O7aX`-B{;i)%Kp)y)3_~9?*l`1A+L6e&)WS=P0oB*$v@Bp;Nu-AS*GRu9fp6@}043u{B93iJDB(7 -|HmA<=nbg?D2g6`aAaTfBuWLdqTHu;x7hVMK&u-^xr%O1H`=cvL<)lbfz+)EXuXkrJBEZQHm1sQA)%j -SzY1#^8;Axg@t>|K4Yqum{lQ9_ErS10@|wBPd}fXw@Cqo6+_XbfEf)Ug9Dq?%2f(VY#=6OiNXkylW@S -MO-t()ZHN##p#%R$lagU%Zp>w7p>3_SsTCD`+@}3O-Mk*mXRs9! -Xba$bdMX;-6{XBjKKOazshl;-3NKB%;_Sd#aohMvfj1|8o@ -cU9({5_2KhShK4#v*<#rZ-n`g`JkLX{lZ^K=6;om7xp;0E?pB&@cUP=X~HJp$AkxCJm61n`L6D7J`*3LZz*P`i$hdny};M!Chf?YVY7p`5N3Y -EY|liN4fQaBiWQiJ$!%n{wZhqX=f=(HY^Gjq~^PF2&XOo0v9H=CKds|8Z&W$GQq3p`v9BVVebcrFAqkSQpt{FnyFXP$}X!GD^!n4U678{=zpd_X_vB -gzQh>&e)p|Ww`&=Jerp4KF|j5o{_c4_JQR6u@7S_KXvuP!>2A!B*H>>N9jNbzl&N%!ko^8!_=b5qzPtf6RoM!dSJX&izIwje6JRRaOfT_nMue@@t*ns=ALfbg~D -%-1*rbOs=qySJv;?V9U85ACH56OuTDHtHTVLQ63fow-AjVg+^%YGY8jX>C;w(=j^32C(5oW*V`h1PvT -Lh&>D>xK>UqG`AOF(2f1EL1{DRY44gCy7|skF&1W#hz#4PFQPJCn=6h}PqxUTo+vzanp=$k0ui#p&-U -!QtR(DNMZN~Ld^E6rZcit3N{1b53clDVfqjh@3Jsdh>4uXZM!A+XT?-zeEvn13V~mZu(v7-9;K6Tr6I -ElwdT_VS|J=dQYkc^`HG=*kZ429C#8!~UVZhgu89K3UF8PJs`aPPFJnHD@v$Uy^T8XusgeLOw`7-36=d(ZVejpPv`LEykJYF@6!QdsX4HtggWq;Wf|Q#hPUd-KN7s<+btuywO8!n}7T -50b!&o*cqPpQWISk&o#fg0)4TL1n*j;y~_|ITX+70wgMbjj8>S-rVBpuH3GX;SU!cpv%rh%yRn&O%E+ -e&@Gd+L{Eh)&nfwP(O9KQH000080LuVbTcMRVy;mgw0JDbx03iSX0B~t=FJE?LZe(wAFK}UFYhh<;Zf -7rTWprU=VRT_GaCyx=X>;2~mf!U&Fw_T;E((ckPb$kTO|8e4Gp_NqvYdR-sz{J%N{B#!i$gNU>;3Kfu -08=u^p%aO5|d!}>({Sezq=dDs$50W>9TFwN=>Ixwpy1}6D37aHc69}MZLEt&(tc(@_4CKx=yNERsQoU -m89Xbm8$FHQuz=2Iq-DZ_qsiJ{z^;#{mWu_|l+%{QW2mN -l~7a(Dj)cWIeRibS<0vbZcy}h&OSmqof*1smzaBwyNT=%ji(kRwLWW -}%`T&|KyRFu)8goZTRlMf5E#x?$>D%&-H_aE%3yjK0anOJ#J*Fm45`8t|VB -m5f-22Y|@R%9?aoN7+=BkqdIWi(MAU`2`~hYyR%%m+A&h*WPk%@wqM6dxVMdj{}}dpl=!l;jn#<|azj -GAmRX%{Lla*3>Frjt#ixDzR$@J=+?ZXGPsW|Ja?4#XvZG64gy*9x9DQJptCKn{$KJh3yIGz`g^bG}Gy -j@bG|*?}9Rl4*v$sQYcGf{6i>u&%tWy-PdViAww(sCB&^#>q>B=0Rbz2yh?%4D5B-yBY~Yxi6f>{_*5 -h-kS3poA6=aOQhgdn9|IKN;Rp!2$K3RyxPc9jzEVw+CQUN@HOX6bQdMOo{KUK@KS!|l@)URyxCzz?){ -1&Z_*-n&i5l4zg6C&fDnbmA`x^MnU}*X8fz`8W7QDZLJ&QzB=fRG-pZ%#_t7Y!I;`x>X=}3nT$ -bK?ggO?=ZKdNLdOH8az@g4C7E`XoWfSYLfi;c>)s5Wt{HPJhXuGNP52k7I#9p2)UsBKmgxOG+D03p*6(-iwMw_M4s?D1W#B|hTz=ly-QIo4OcYeCTok`0H%m53OE -4$gEtK>!riEgvz3O=s$7L+dSx;T2LR5YRWo6Pj;&7#|@5{c0?RW@%MnJg?dEJnk0t9uA8YxcA|L1Ekb -W{1?Sn-U3%oK4VB-r}@z7})b`Riwy^uvKoaGQbDb${N-c_&RA6vZR$GOmN14fN_`ms3(*Qupvxjibe? -SM@A4F#Xm-e(c6=^(Zk0_OfdOrDvHbsKBT?<*72&Kv((<;*%Ost)l`{_UAzN`_2eI}#&t -*q0RU1Wx-7v#Xxha!2^!RO3OE}!JnmQY -01)WB&Dcz+fumZGB85Q68Wz5=D*-GH3OgMT0c#k+rJx}Sw*2;`$3H-9!vJI0nnk;s!%!%a$}+D(=%s* -4f;4dr(2Gr-&+vIWKpQ}5={lr64+E=G}+}9%!XGD7F(tPEC75}0VMc;xo(nqmS@eTk5rE}UXT9Gl -{`QP?R|Bm(sY^S!8YHUzB@a8@%oqP^A{&CpS=SM?<0$C7H@;G`;9lXe&(!A&m(WS%$$S`Pit$hpn~5`6lp(G%u -@Y=H2)-^<%2;gmZhaJzN{sijOucL`aZy?ELsiz{!r-5&WR6j_;SMLfi1vYv?HRAUamTjwbMumu@6OBB -AV4rmJ)Y -jM4au1Ij900XHEW;hz=;0mB>NlOO^IX)oDi^~Th+DSu^iM&d3TzDlOr|sqRNB03YW<)U9yoE8z+J-S< -O>k87_{NAPi8pGLmSRpboC1`lnCJ)Ng9v@QJVyw$PNZ>(pxWzF5&7Zdzf{Trl-0#SBPMK{CskA-VPN7 -l2xub9n5ck+$zX^jNRo*BYW5o~B;b40rA0K%*e3_|m!tTuP5a_EOT*fQWg%dEmfVp)~>Ag0= -x*wNPw7WHqA5W;aZoT(~)UUJ;re8!vpbPbinL;KJ72OS}%tj6mZh&I}}rc2FD`^PZ8%?C-+=HV5J)8I -R!LKsyD-FJnL$aJ#KCa5EJ8r6u&P*l|b}ix&D5X$ZINT*JNACF&9Mn{t>=&66Sq$(jS}`y+9fr>@+^t -}$Z5d=a=%#F5rY8#6~2z)vNHPzJ;m1CPMOAk^#@$A=NqlHricRs*v=fsNKKt^k@k0NzR2+bxUSB{S~? -jv$P!ch&%BG)`;>R=h1phmh{cDHxpV|bZgCis1p=A8>R_Ieq= -Y1^IU^{478ZLi^_j!8iP_puanFxw38`bgHA9=i#U#1`r&pX{uN*yNOLyO4$;C;)eRw1Fx11WFmQQ;rl -l5R9XPjyY6+eocc*MVB6#TG)L~Zj^CTaxNkz5%s* -vb2d8|aX-`T)E_?2n97E9M*MO13=N?0%Gw0c9RD8VQMk3bJ*PGjY*wuS(ovR}uz;F#rI -(Fk5BmOPmw7Vxi1GFah%sttuzhXX@Ao*Yd_)H2OwFf!fo>p8W$0P4ZFJy-kVuUflbss=Rt1})U#s=l1 -~38DBE|BGa5s|ENV^9@*M(f=50#l)IHQ~x6JbzY_$9kJDbv?AZfTmUTs0i*y8hh$0O;xD35Q24Ks0y= -}2RxHYttOVS^Z9NuPvLI=Ro2pRknWms?SYm040>flA#^@5zGeaJkRD+ZI5F%hWzP^lIvRJZ$k*li4g( -Ynmm}w_;mLIpzn}|8qvDE_%jNycNuPHNYJx!dT_NDr!NogR1#C-dVQU!q3@mbZXyCyd`vm-=pQ?QwO6 -AZdT%mWMuZL|CtgzEss@1k)Sl+y~)`y#`J&!{=EIVu255cTW?n_?Ia^xROr7A6It2{st9K?HzIRQ@KZ -va+osHfNu|IeqoyELyh3g2V4@8Nz?)hEixj438SceTFgHqy|t%^|C6>Dgp>&l1|GpCW!VE6NJ_dL%J3 -0QIy9R_XZ$NEIi)ZM+CuW&r*h&)u>28bb(~L#C#};a&4}Haz&J~rXlqYYFAtWGy(QSxL6B8XQxCOH)f -jBV8i~pj$}~q6YUtBDrtC1SJ(tgl-yq5CRHj{EC+LV4!_a#NtP0GQ`{VgcY|V_oO`gDbrg0*NPN!C@|>{4;-@3J?Re@ShpBk^f}2+JYjNQc;QgO3x^b5nCu>UIKli8a}dp#Yzu50I)RPJUEvPh9-{_sOD~Gx#6%`{M@Z~y-4``E)IpEQJ&}}&A5pOpjmcf -%jYEnyOzsPKV2i%s2daF%s6mWDH{tlmWP7Z{pm1mKF-UvL+rc!zl{5i+o6BO%xlE{zhAO^{S)GEM7Fjx*jU+yMMz*u4QBiv+{hQnJ5 -GTcmQeqU)t2wi}%qp;+DQA~qs!)v6MR{3ZO^ue+CaB>jMI}MFpvXiVJ@2FtIY`bvA(~FWuSK0xXo=!u -TvY;-$8*6k*g)z28UVgIN-{+ERm4Np*K#6W6X -b#6=GeP)eCtrsg3D5iz7P8FF5ktXSl2DlsMSIKdO?&^eA?G;D6@*3?qR`pHvA96>yTps}+7>qv+!QAd -xQjH>B}kp=kQg2%u{m4Y7#xDssXP-Et0m@rZNXCKoDs57ibbpLrKYn2t$q*kH$J6!Xe52&BL$- -loR%@$HIWN*^`ai`y?FvZSLLcCy%QxQ?P%g} -u`goTb{2b5k#k~g>Lh@oX(7KmS~OGrpbfXt6xBVG($$4eYo94z8IVTOe;u#vON0>yjQz=kA|Mt%WDoJ -<&A2g#feiB?d@n*a1Dx#y#xsb7JIUGDFSm%b68k#pJ4n=*+%OYL;REsKy5cq7S|EZeTXquN?xGPZ&!BY -E*4NRFUxE4a%!b?R1swx##2zh8eAeqM%D>MRVV)dPjH9V_<{Dd@O(L^sFPFnWF1}^56fSBqufEX7}1r -w9IPtx1`yPi*#}A)J0TGZnq&UQxY%}#=Xr7|L4RV;sx3rG6&nty*KXtuAqzX9&}m%~)WDxWXd^waEia -71B0Jjo2U;N13c=;Wwb<=h0wm+m5)TlHh)%JtOz7L5T*vEqN(p+7jt-=&ejB!+6$bD -0V45L0wrdi{tYTU$Rj=pOei=(*w%qdr6maLf+(%NXjD{${0Ohq=14mBHdLW#i(b9Zi2^cDD5?`@r*{J -QAe~$ELXBkyfQL3$l-y3mnC}k_!OYpDbx|?B0R8 -Uy5@+c^CJ>zxAWK%Ipc5c3qUWTnGSQS1KuLk*@ylifbPl3NR%_M~ -!o8#Z+%&;`&Y{_l^ogIQ4qaL|jc0H3~VUV6z_G$AWy6v1{aj0d=EX3!VX*l5O#cP*q=E(WC$tAt%+|6 -vX0|+7m(o9fZ2`R%c7t6BaL?9TEi -5VIY}E|Gyr=T}r%7l2L!g%L9j_|z7ZmLGhOf`SLyJ)sCzM%4`tEUIv#%K9#AIJPBN^&Aen&I&n&f!kd -lT0v2Od_!d>82MK)#f3{{SwCnirWCCSl^q*D)&xBq(dX8pIKbShl)7n`%ea5@7Ew@(iX&lp39ZK{IzE -n$!c^$(x$bY6<^0EsdpW}qm^~cg)HKxS@dQgCMNwR*#k~9gqi_$%t7MQuCmhQp)MdL1UV4?XI{|+|!P$M^TY}vGQQ>1d{35zWp3&LB*Ov*#G__lRIS-p+!z<@Tr_xaozR=R3# -9c^8mx02m@fBG_5Ah*Y4*A6#Ka*^e}#gR)3VI(fD -xDR;$35S)mUWWG%qm>i_D;H0lUaFnoe8%sGQiq%DxtX4MW9SWNotf#$r>AIexNdwhQ**>JzHd;0TVK1H#I@d5N{A2}baJa~-tQy5-+BG&l4n+9@gs4JrpDz6SUX(|00C9?fuZF+< -ncvToIjgcOhIK8PX3IH6CxF4z77jJn~E9T!cifO~@o=Ur6Bn!w!%P#~G|z6qPe;&wsh^|{bYC8LI8^m -*baN_`-dLg7-yt`I@Wa+MZzobicjTR*N{qQqX%!Rc97wP9k8uJRytO8GK3JBMkmmP88%>o9bp^~GtS=KM*xQ@Fego#{oAd>EvxhH|OP`Pq -d%;V-``hnaN1g_>Zc#Z*E<(+A*4U>>jp>zs-y>ot}{5G(c`q(5U@5IX1>H`NrU9NjmE4Zw56q8tuQvD ->8V6LX?P{9fh^2@z#U?1=a8DQyGpZwgu{Q41t;z^J2$mlDEX!Sb1Tj1y@asy%3nB#L -X^4#&AOvoC~*SLa5D~GuT~HX_#7anbbKMt&}(_Yd?NP-W-;NG!^-egb5x8HwLsWIjh8kCfcJ3Z4%B7K --0K}^}L~D5^Px6;{XT~Ck4tXb3F+(*$mrbKqZj(Wo3ziq3&*6xz{x_7){J&Ugzgeip|A1Iz!U}77yvA -FX0)OW%_8a`84Qt+n#7LZ*kt%plEhYJ)n)K5!Ife&zI*0ovt6)nRrU4j~pI#dxsYKDytXP9=y{hr779 -mfMblDjUyLr5<8+Grdt06(m0|6J!BP=Vwl)Fz$9A<6E*Bl!YI%&6Blb}9Nzlm;qHk}8a(EB$8?pfhaU -&Dn+6k#uMEA$qfd7Iy)zSm9x-4AaOQ)DLfgFw -(%=V9aZ{dvJqoG4=WmxA4VWodsPEJ`%{kRG;*=OvEZNMvu0>I_HmwZ^B9Tu#u)3E=513yc@-ACahJTF -_pnbVt`K^4mGlrxxx(y8r!hZw|_|rAytQJxm51)+0MqK+xgrzPP#Iieh%eUd%NorTF? -%g9?!BW;;wNBRUSs0&ii9{ea2`81FXdW_ucc6QD&8B{wm&OQ9xv#NJ)0spqcxU&(y?^{@ -dlO+q9%G&8Y`sZ1`?z+gSgdH{ax}+u)(k!M(P|l!OBkWkpYX&yGQp(E8C|WxTwurxAPSGU>0}$$G;Y&X)Y>N6exOJ#T!j4I*5NAg?V -4R5}=WxhNle3*JH|dhDWCA)}t*4!f4b8S#2DwghBWofy)OtefjF&(i2=RM8o4o=blq@mnK^5(2Vz;HW -_{D!IE^kee8q;$315%KlKHrKW8bJY9GDfT?<)GU6I)+jl}lYB`WXFO55^%wOQYL9{S+*u2T^H5a#Py! -9J_sKiIh~pcitJ6SkvgY&zR}vC(>Vn9Lk}Nv9m#_ljgYcA$F{r^kHhgQznh+Z+IPN`3qPk8%KT?C);> -hXmfosC7DLS3l^(y-AK$5FvVFA#fh`@k817DMp||Pur#9QYL(E^gTin~}<5T;^*)h;$mt6wQ?${& -=J^1wR8(VG130^a8v|C*~UypG;(91r+nujc9->170U7iId4Is6;ot=|P6aL0RJ%4&n>Qt-TbS#c`BF; -eccNN{Ek{`iUU5@v9#mhhXYJ^;Clg7b>FD@exhM87@X?j -OY_zd_ri+s1prT-8}{`*^-k5sOb)y8SlZClhf8gtr!KnQP>tnW -L@cAHwpTwuGll5kTS+mLPfB&AKnvuMXO^D2#r0w0w{wZB_;E;tM{|8ZpGo?m;qeRE{!v)@r$5)SdWiS -t{4LynGWtq##R+8o2v=N3b@rFmm*K~2hjP*v}dik%6d7wvnz&Ax=06(GJs4No^c&-daDPPgyyeY*pAK -(Acl+LZDrvD`!JeS)-WHsotv_Ud>U5)>e8MCSBJuN^vM;O5~$`8U1#J7uby%|!}<5er~ZHmgpq6HtkR -a|ey$$ni~A_YVaWjM{mOW=F2+h0V}J(Sp+%MJ5poQaRa$J_gHxOH3NE5`GgABG{g-sEqvS8zSTe$$Ti -<5NM)p}u}yZPPP%%8dpQ-B^*UwOX3xMqrW#IsRbqQ$=N&F5Z0Gu+)9dl`43SVSG9q{}3aBof5H6KJnVX9%> -HXoYC8mr}&f(jLBUXREID^R^rNvD!aVIxV_F)putgR0o!YmW#bl7>u1=(h`Xe^Y|&}jep -2?%Q@unwq*o`c&Co~eM2~DDp@EZm&`%fGDIJi5scSE{=<^znRVK7EM*A?SfC=gvAJ3Jfl8P9b)UuOXvCE -Zv6gOrqkji!y~Y;DqT?96qec5qfEbKZo84lm(w|BqH{ytmzar -sKb^I}DYW39n!V39^p1f2$!tT=$F_Q0-&y3`@HhI__8nadAmdF%>ux_@kS#lncg%R;pG~yGrgq_=qlq -V_+-q36l41LW4ldwceG?4U1G}@MLec1gFfRrS?|}HCO{XDm+icq+gs9VvBSws4(iZwdCoCAeB9w*9^b -)`H1qcIFI}Jki3y!Y@2bPVYV -H~Nta)*Z6YJ%nFC2|{}N>^Y1v!OHQ;?0?tE;+oES##>TnT>1s6sM4T(lbk-Ai{JROGdVKIebUbOq$hT^4$09aYV_MRS&c+;=7Jgo%E?2fo8?VsD_L^NzbQkR -5W?gGTW7w_R{DoV@F@Y`b3ErJNy=eRJfuW={>-kimHw<+tMsbECPXw)djR&!F(`jPG8;ZpPIw*8FyJV -~D{3`i+#|dKkV(-3;cw7W1=c_;5!W+5vjT{G3UC#mK>&%}Y2q*u{C{?UY)>;%U8rr~|L{~_huUY%=R%fg=gGsnF^NDmp11cZ+mQIZWre$nx1-N8=lxe+xGUqnsQEi`hxcXc!J -c4cm4Z*nhiVPk7yXK8L{FK~G-ba`-PWCH+DO9KQH000080LuVbTk4G}yDbC&0D}tv03`qb0B~t=FJE? -LZe(wAFK}UFYhh<;Zf7rYWpQ@`Mw6izwKZYeP=97{6i~zcxbS -1|_i?U`M;1uvK4Bif@Dbjb~f#rRC2TT9clD+;IJcqSV3u^vfKWRC>c}Vqo}eHD^z}ZqT$sW{WFPiILp -0+gV~~Lt`l>-T8`3CB-T*O1{C_Y__VoF>on~T9mhg59w#8%Nv=^XFyMqB>sISDKj1p<4Ei%6qTZy$e} -ULD6>{g?J$+A5(PIGH}K1wUyGisQ$Et#Yx0o-KGsmo{9$+UD4tM878t92J|!s=*0^k0lY?niYnLbIFs --^6<(}=+?NumAA7E3!Ef)r-0s=~s)(2kJ``1)>8;sodWWj)ugYM%5K7Kg=Fo*YC!p^mdp;)bvDqDk58 -~X4@8>(a~3n7VjC){T2{b@@Up3914VXbyhD#uE{cFpZ6ocpCjVxcH_W&Dv?>Fs*SXPr?uMQpx>y%`jy -9oDsbyUnN2x~K3Lkpq8FO`vJ(^>oU?zGgkr=^Ntfm!}~vC05j>LJG@RYOt>IvAZn9a_(Z9!RvQ$tt5u -(@D}U+uyBv>!4eh|-IyNJX1y#vuW&ojm@WbstI)18raFZm&3oqOL|cM=x_r@f*MSf5VDerd4Yn{qN6wBD6 -ItuV1-BMU)+S~FNuCscDpzK&ipmWnCl2PzFB?*{OQQYqq~0WbxbhSTNDob+7D*sZmYTbi{# -yTP-;{nlu6gs!nj+8%N3>`o{#VK>|m6J5L_#J>HfVr1eUYI2*vUEFNa{Um^k-l1VF^(uA@=EIpF{e>u -GG?F$28cnwD_0yT{UE8r4m`4j9&>jQk5wO9KQH000080LuVbTO)hyab+R^0FRsi03!eZ0B~t=FJE?LZ -e(wAFK}UFYhh<;Zf7rZaAjj@W@%+|b1ras?LBLE+t!ia^($byXGPi&B00|KmTM|?61VF;shv~1$=Rc* -BuHFQs6l`QK+%ll{Pvxh`+k5A+1{pY^ud&&8o$+U@eO>)~0#UBuBko?u_=c5PRV -T2{H|abt=Sj0eS=DXyN9l|icn&WZzX9>$==EmFzE9zozm&Kf4j@^z -KtTy$mX=kS&j8z>*jS(;#&sZd^p|p(*1#I#K{VpvFT~$*)U@@bI|`No1I2>+J^#~Q98D`=wvqGE;7H> -5a>?-mIP%ejJt=~yip!&;ql{fe<5|V(nZl|s{vSr}4AIHC#BNAConJ+;IUoyp*|cdnWAa_a=L0?ve{K -{RzPBeD(2yFqvo7T*SlX;eo5tg?{`YCIV$bTjss(Ih9{m5$X^wKZzGh{Vwk=yMTV#N?ie@!SK_=2@NH -B*J1#J?o^Y$7(X)F!~8qurQDIy1=M{7v9!GBBUg$77=2IUlDoC1rl{QSTZ~>}-o{vCmlBZvN{U>-J7V>+1luiNPbk@Sa>nfw? -#1>&UC$Ux9N-gQ6$ltK2n*NQ=+Jtnf0@Ykst0If0%mQ- -5q&ihFK=5&4iM60TaJo*LYZIxwY_}toM)*o}gKS4Sb_*_tE{6SGEG3RAzz+751EQ6oi0)b%j$vu3iG^ -$n5`e%WSSv~=9ZBoTS+WAgX9Qq1i`d`cNE2+1KSPlJbRpOF8UR%sGQC5hby=u)e11mbjL7yX+!I_9?u$;h6Ca>!in -@kYyWb%i!Xz&mG{OuAB`n)YR5i3!l;)pD7@K*)bzyLqaTW6BV1n~fXecTejIb9X7)$qxxQGdhWMGE^k -Yr@o2QzQVj75XYfznHV^kZ4H6^r-lXAP`dzL1Wg#i2-Ed^ML?}e89y4mJX{pknLv`$TYBvSS1|biy^9 -w*hFDyZao|s_}>nx_=ch3m6t&MoGXhtD2|dQKrJrj#UlcCo6ISAs`TlBSb$9i)AhCDlE{#8F;r&8g@1 -xy?!=U!61vhx34sZd;dS~3g`iYxLDh*gXvo;)>n2^jh?;B^j1y5xhRNP=ra%EMH|JfOw7{6*7E^xd=XqQ{RirR7E`cG$tR#M -RO5Ar}^qLZ5Dyc{8h4HEeel05)9#cbBtbmBDh5c%Vu)bnMyG^0DfF`DBD!i%?#o%yR;9t0u4&9)KU42 -i~)bzy8?#*WmJj+|ab1<{-WGMp0ZGgMV8!cy+5*c&XcSKpA$7=WbLAe9%e5n4#xEJ6 -Y8;yX|=0~?sbgM=tQrcj{xpGSK!;mwzUNrxAwC%?OVcyI8fIVoFK#tA2X_)vJ~$AtenctE4YG3~G7KA -jXo_R)Y#oxMm+vlNbexDV>TQlRxs6p*m{d=HiEwAGxOy0BC8x>_yy$t|v}w=~lNj(q;OTH=i>4Jfn_j -}bAGA4zh5P-|~CaIU;Xt(o|A#-H2ZQf&VlV{+;5sG2G*me=X+6i&ihfOnI#HU57K -k8WqT^=xow$h7rgtOwz@_C=d}1GeK0uucT?-6Hq$#rE;tTR`)+9*9SLnPD`$*C}2G#9|i^xow# -!AP+Cnlb@uxa8T;W0Kwhw&_mAB?xqsa#hJD_8MgAs6s8@hjyA%zw{+pn@_F7iLLCR72|PiO%@=U+h1J -&2w$6$=j?TcU{$Yyip@m)4|3N7DG}^xNd{kA*KP=-%C9AkjBLQ-gUSk#PsIYP#_C~kKVhLJ?})fxwQ^z1bUF -R+Y!CP&lx6Ecu1lL;>i*_5Z3oe$mRqzUcFtKHc!~cFz(#+VX;FS6AJG8{dN-oF*Yx!^4o9;32#O@BF5 -xFMI&``)7at>gB6vV8y~8&;R)1<)5E@{q)r{K4#!}Ga6XUJH67MN4;Ux)5op>xi_NENI#j6@6lsWMPu -)yC-hOlnq~|)n9?6`P6Or?NUA%=Jh5O{Z@Vwst0ma#+AKbMf24ek<8fZ*tqK*Ob|Q}{P^8z}&u!}Yex -C!xoT8ba^^K^C2;}+BUz*k8pw~RygUZ&A_!uEI;RGXJiuR+x#l?aKdVWFc!RQaDT!xpb#9f7ntFxFK& -JI=-A`GQ5Lhq;a=%q9A9%v}~+{v#jc~C#u{t6AH8ak-@L -e)8RzTdpvJ+;i~^0UaYuZz>$Y7s!_Pj0%gOa>it6>V<)*!^%Fh;?<>s^Hswh4?`~2~5zqpUT;?;)1>g -nV=IXnIB?|%EAkH0wkf`F;$8p3TiOE!%1IS3NFr~vgk&)B6>BAxNIp+N}FVl}0=kTaEsy{ym{nujtD8 -sqR1Oq^p70rc+~u%<;`W?%!NZ^U*B0bBXBf*FzOwxS|d7%{DWpsfDT1;n9Oh;8r$6w5k4*Y< -g_bzR^@$2$vWT*=CR0_{AC}=osa*C}a4`QJvvkG=%%^Wz=e6866_)4m0)Z>`8U=22x4pZaGiHZW>gq2 -x?zN>yOdC(j9R}@Vb@!Q4lG`hrEhBS8?7JOg=!FAkLCHB}5*mLV+#kGvx>m)NZ8?wL(e4>f&|-0wz}92S8q5I -W$XLR_Qvz4LCw$KFm}1D|G5l7h3pO6V@4o{U;rs&2cn1@{ADs70mfDe<^|aeE4=|R9VAYscc!HMF(n^ -@Y?ZWkhSV!-WdbZDdA^z9~Dj;pWckkc5?jM4=S~^xT%2yHHSn9sa8K{IP*LH)ak5 -^Rs;30mhVfH0yMRG4TM&sbfC2^~SftILAmUTPoT< -sWapb8>)7?tLK3F-0t&Q1^CRwxvUatlR&c7%)~7;abn)be-JSbBXI|_mhe4HgBLlOrG16N5jt36 -uAWqN)U2+q^0zWvYuwOyFbG&sd7l4P<<6O04Kf>`A9jup~DB2((TXZ -OYfLQ$?_(_J;ifW$EkT#Ursi-Eyc9xF_UfJp}xDSe&N8j<9UtPJt%+Q9;$gzfUKMC -$L;yZ8Sk}QXHul@!PCjEN9z5ZIm>DkL5ggES%&pULBo_mbG|w}CJ;$;gnMtKi;>WcED#%|v+GKC3B$~ -6WKD91}h(O;UhvPhzhf5gZ$RMO)&AR?gw3bR(aH~A+N;8!}&Q#Sv+I-1o#x?qTL;rYt_S_{UIi}H~7M -VHa&vxL@Fe5xwRUsjwgOaPlbGVxyk~m@Pn+PWKvVU)|_fOKmKRX7&UGw%zPZ&75RiyjQ$G4c4k -mHvQ%PQ~d21bILgy?FULTB0WPT?R;CfV*tNPkJzfr -@dL_b;?a+cV!Mtl9AK%$CE}L7j-CX?!6KuddhjbH1C%`9O-aR&I|+zR)@>I?M+KUiPbE^sqFR85NToH -Z>W9o*PRmOdJ+y@tW8#G?vo0ZDlP8So@z^#+2S*a(FJofP0jCND64bzXd(8ErIO*ujh-P^`Tjkskqly -kQrUk>-CQ3TnG^=>^23VKC3wR~Kf-)e36&B{joO;<$0%<3H42@6rT7`H-<{lzTW|ugzW0+VVZ=KRqNN -lZ|BBYHUNmN%=X#mL7MwJvGhx#N{d4y^&=ssSRTE6s}`Yf3z(UZ@UGm&>-z*89+5qz9v?B8P>efX1ptxO#s1NHl@HA*tStLhu~ -^kWA-YvvcY+aF-ZmuE9>i~v5NMB}K#904pZB$VU4>&ozqrjdatv#r!7o(LWRjQIWkUv!Q0%7+uicGy# -H89!IK<`)!-+&)bnPjzrG^>Soxs&)z9P@rGHn_!xZC3~M?}sDMV@=1NG~obMWQEznFmAYQ9i@&sr_Io -ya@_6J^K{zZcDhaQ0Pi_QoTHo1q?O~_V{pGSlqGS$EV<8ORohv+U(>oW!grPelKKV3ySOuA&{LInf1; -JI#b%+3<{zDtz=vh1Upli%qHG)#sU{^KSL^cg`h%t5>yi&YHP6e8t~};JHe(qPF{fGUd=+iokFzzeOE -~D3`of8o*52d09@=21Q8I&tw0aQi$gd_?2W_o8`IV8ZaH_O%SJG+7u$IO_|-R1?8H?!9nW0I3mLyqf| -}~Xw{?T*(w|FLhuMm<*9~C3M8rpj1yNuE5g<91}w8Rk=jR( -3-?Q82?WyVk^F0Sj!NI5m`M9k{SMVR`Ah2}^#@vyXayzAW7S1Vw9O>TeIeTMupMVB$HQWWr$%ckm3=D -q)HzmpfVJ}|b&~Yu!L3%Vrjey*oZO76c&a)R?I9g393&1np80*ylWHdT1Pv0g5qCJFB@USST@?8otXv -cwSazmMm*gTk^{QBGAMbUbrTOwgCqF*Q;X2P^FAA+DlbJ)xwV)C&@IWxY4crR>Z>EOA+$_>sUGB)0Gn -ojf(E3;(VsUhlB+2DqGEo4hY?fk;5te?)G$($Is8_J}REjrTvgL5{`0?q4vz)2xh-VbESO5L;FW-JcM -;W%e1GnHjOZE{wizj5RkgIAb!?^Nwv7*YE --Ypi_k83TGlTgiB~m2?2{Dg;F1y>;;N?zof|MG6Vz_=cB3an}{5PWxI^lvrUeGb_1}IuDUN#2iVd{T1 -H&4VO#!RN&{5ksr{}mDmZ4%bE;jwx2mTvX^c?=XO3+1zc7zFQsuZ^P=9gEMx(DvE5Vv^ -3`>lLh&)@{)Uo0jibqrh(D65k<+U3%98>NMIw~VW0UaYeaYZaPX{dcYJ`Be2*O_PSWL)m6_e+*~ao&E -{>60F|EA?*xpQ;&Sb<-S)euBeW9vdlUf#G!&~#bEWKJq{PA;Jq_Je(H*PxdR|bM67m(J6I2LU -#-&W>#KyLf%Lj{j4QMEETqISQqwLv6G~XG$`i4(bpIqK`2Uix7-9;MHuQ -P>TbJDOGol|j1G|WXw1{6Lj5S}-e3EosYbv5tJl+;en!a($`>MiT -zMe^33;D+`6H;or-YJ8Mb|D`S`+s1l;4zbr#1@NGJ9nc5bK>_@z{XQ1I^3RC3*lPN%jF8MTnlL#JM2N -zP0-!Rc$*@=4|&}fj2g*LMAhOQYfWu4CkqP?bUJmoT~7-ln=V4??th4gTfyjQ(mq8KF-tq;J#V6JP(~S*??G6%-0?gn|ckUFP${uz~IZ7b%^nR|2o_VAYUxAkkD+y~dIUjwVoK7aF#JrZ>h^f2?Y53}`_#X -UwM(>i-#cVf0sRK^iZmZFg?GV&|scB@uS2+k3nJ#%(DD}<`Ct<9;gnRagf{oeHCQ2;PbI$G6ED9rOTW -ReRqqD)(uBKA>^QUyf)4!+8?V3L^z9`1yd?7}~Z#Vy=0!za!CUU1byc7u>x{BY1fGs>wt{qPRW3% -h1+I&#-Xk^O{`b6hUy?2BUBN34D12{@9hndd5)su&n!mc_uMMKBqNH-$X3M#GEqAZ*uy=FC%BGBnJxI -o7enXCVW@veO^-~fQVR?Xrlu=i08y-s(wrNGw-Tp66_9#dLuhpksU(58e?mOl^kQH!d0ZkdVO^Mg#9S -H5i}daQH3F1F>@?b0uK2Jxq-RmJ<7}Bpr9HS;OhC?@G{uiOYMe>JxsS+->_lSw)~pM)QX&CrcQ9yx!- -z>sr~+7bhsOrYu;giclgA^L^z?jlgEi*8Nk*o&_v7&RqfSG%x`hA|F`IfBrWY%5t12h`xPy!Ol*>NZW -*xf6qS{3aC4f=Mx}-%e&m*4&qXj4Tcax9hbiL1FJg{B_MBnq8SY$c0CT#S=|X(iJefT!AiXK8wF*fZz -A2L!=GPGgBH_B8hB&aJ2S8~BOJCyUX^lY!WC=SsC@rSgvPVE@=%&PTxISFTJQ*riwjb+cQ=RbCln^eFfX-ur>dZDB+cHJRADjZr$< -m$4dG@i&=Kg0qgToXXhsxC)IUlBtR7 -QHnV#ku{Qt~M20Iqd;@$14f}nPEuIhvy>R=W3xZA41V%&_n(lmpx3bk^85A5k=Dq#))&PKPFlXdO5RX -#HFaVQVdxfrTE_SEf^(uxL9u*=P1O`Zr2BTm_-`ut6g{v$emB2)lZ1!u?Is)FyE7#BY|Ks^R$w!yp0* -rKxbUdJ7JB*}J2)||W%l3{6gHZL+?JFIn)RwmBm1Aj=+FgFbLTiNNz+jcy|s~l~${r?> -oM*xl|dTcE%Oh)2uEwXe|f6qg;8O^>zo*ajar)RPJEEIC{f2BP!;VFJG9wcEv!jJK~Ps!Iwdl1vwGjP -I`8fvW32O{?}r*eTvx*)#h94+71QQ<0B4%v2VrNrv&rqYv3%WmG(M@i{eZd#O}Bq!QtOw%d#Eey^kV7 -CN2LRis=x|Uw%DZSU&DlTDJ?-U8?N%Sl%vyJ_ObQwRih!ggchu?Y-DLDIvQp0Y;~!-1<#Gy!eA*%JI; -pJZ+t(0leN#?)t7ew_{ih##pEZ#JaheC2Yqc+ABsH;81bN@0N)uVQ{u8(pnVmL|9*CWmjX!4Zo)lb84 -MRf6Fa}KE^kV_||Ak(sY7xprGzk)i3@KXdTGOGOt8YlBy-rk)-J+Dyr%YA7DW9LqUc)$qS+cZVk`OM} -<}*sdezj0GCp`PNUw={WcYzK5X^C$WefPhYR3idlhb%8tAbQ#h%?^5=qy^RW+PK(pGQrFh -Lp`e{D*GkTuuTA@LMk<83S;8)Q_(ZY6%|VvmsG^H*w!N{T4j9-U@L7!w#xOIbii%7 -veOL3glV_+D>vYrD(Mef3V1}$S*5vb&>f(aU<8U>NZL`_R*DPDx7R~%uE8jHk$Q`ELueF@N)^dJuEhm -~pNs|1_RCFp6@1qxW?ge>!Ip?8OoI9uVu(%tvPf^5a&GwP%IMpYh=X!Kp^_*}g3i6eybuD4!VfC5pP4 -D)G)Y+RV0KS{ox+!@$ciFJ+Ql-nmeB`p(gE7Aw57~FAUvDz{XaAdwkwfU_-?es6?r2@7OLM>%JE1 -DYU?I8hKsN=`*T7AGV(1!tm9B^A-7MfCK=H&ME%z{W7M-RRYTEuaFVf#f@~_u#Q(^7(a&3J2bWTQa6L -vrL9W+pcdOp>{Q_=_;211|E&YeQPVJ7L=p{-vGF@s9B -nAIMSGvX;reI!Js{ilEB=^L?|})o|L0@H{DY(a15ir?1QY-O00;of09jkTk -tBWy8UO&c!Jc4cm4Z*nhiVPk7yXK8L{FLYsNb1ras?L2F9+eVV#`77q2x)PX>uqb*R -M`w9gwv|M+j;)H6*XdLsNDQg5K>)!3vKXE2w_kV91A_-8Ih(z!t2&h};W5+G)6?(i!45mW5|S-N!WsO -Ht0-ZKSnyQxm{n=a3szonR`OykS+=B`qdZ!`pSP;ni2cM1DYA61!}f!{5sT88n!P%EW7-7ljB|SU_T= -^PyR&0GU`o%5EGrdkz@XrHCPkSQ8-TaW3bx4BdBRK1VqQifk-=cFEV4BV!(~-g1rI|e)_GQxFmRfcQ3 -rqrL<2r?&Ts&mtv(5QBM;w57s~zk@8|=FgpI_9 -A%qy -p^Wz^*e>jHU;n{cZe>e}{9esa%cEDn>C@+A_Bc$~ubiJi1csUK{9pIHWVr++f%Yleu5vcczy1^p%My3 --q#xC}WD6+I}DeUtP_SpvjKe!tV2H|u+e06djI;a9ViUWd0lq{t9o%`OYH)My0>?!nEB$1TtJX#$npq -@v6kC0spE_tf>&PYfZtyozIpbW8a#6Dedo&e|4(x$SEfSi&=1qb#{r2sJlaTDuxRU!*$5vo*+t4Lnq( -|{fS2I48o4FmC8d_=ZJcpx-UA~WFG92cL*fyI$7!cZnzDM3yS*@YKIa{*NLLIg_}uxz}5S-@|)H!|W~ -s{S!w0t}IgG7NpmlO?UR>2knmNdc<9GTrHR$i^?(yDa63Zuo-`0(VeA%ef1Z>=Ur#u+<*mLp%7r<-Hj -d8kx&Z%N@{l9HxRZ2Wr4t@oN-T36un)^MtQqAyF$u!1Lt=EcYe*B$9;9;q7&lR1r{@agcIWiZyotmsb -(!AHbyqkxUZFL@C)Uh=u%+!Q}kmjIBBFVLSp>5YT@eZ4fFp29Q^vX@LV!!@59W42i6hGPwZdhAJasd9 ->n)JBaj0N$dsc_#-@orlPz;-je{nL{+pR#2k=jS^+yP_!w1dRjOv3;RJ!(fE5|&zPLgTm132OrGU{v? -d6=uRZe3%Ami*4Fm%DAHSuz_2GtO>RtG5#D}#T0zrz5ddW}b|%6U<4EJr&12_kVC;88Og -ysL_|%WvBt!tjrQ1kJWVFjzkZjI~6LjPc*Bgk`WIAY0k|17DN`YMy0D4NEK;7(0-Cp|y!mJw~XbhIKP -%*mqfZf|e9{oX3MUR(_+8@2=aS!baT=iH)v6(To?OSX829M`uT;iV=CS1bC4IJ_s7{P@(y4Nwsk8YH# -a9(r=r?-PV>3y<&SfPwSpTm-I>})C&hawk#aBbbG=QyB+&heCgeGu)7=I+PT?lLx?-i9RY33!$Z6v7< -cS%>{i68@!d;+Ga$~43duuI4L43v`vh#CrDGQ*xy8^0e2U8Tve8F_raRRZl_4C)7<(1@9oVlK$ok#03s6^)(pd_Z@B<&_#MfE6^* -CpCD8&$R_`XfMfL$=X(7OXv{D97goLv8TEqPrD2SmcfN$ng^aEab)j-UD1lb7`<&wJ2QNUtt~EFyO?P -OQ5^d~T)&}1Nqbs|Lp1}#9cEc@`Q=bEMF`$ss65&5Ix2j00JRbS6cywCs9SOUa#uqv`D_a_Yw&@(da%(B4!G>jk$5?uOfS;A -`-4;`>Z>@?DJBTm6pdhBv(zSU5TXD<0d)dCOkmgqS>)nU={gJ1Kmy$mXQc!B;$UAhp`9Q!C+&sN$!3jXF5t3S$fxi}6BeEK -(6nl+X76+pzBzvN!C7EF -Z9{lgjMwdhu!wF!|2{{TH1OzQsPD`*XNRl372LR~_#PQ -LST;xxLlogjxrqk!s9lD0M$jsx(SJVAxUp@PJ-%c5ZM6!i(kZ)oV(VjeuRss%HVKTmR>)IQ?K}l$+hZ -#B+*es0~TXp7enC9-uRIzv8`h9!yHhgn(cJ%7)@f)*@HF%d-K=Jep@q$Kp?)0JAI1JSp_2-lG@50kxz -dw5W_WkPy>$v2x5MbiX2bxArH^dpxTYE0F4zkmh2UDeDk;Py{!IEy_@*m%wyng@YI6OP~_u~%6dYmSh -L-v%`w&N=`nz_pj0s6s*pV&*AF!t07 -n-~jlU*WDtCAvNer_T%fbG0oI2fcBhGOzQ6Bb%J@~Y3wh-K;G&lxx2M2cBgd#H3bhHf>?iV#J07*r5W -DcnsM)LkqiciHV6ny>v5yT(bx2~D!#``*6t$y8ClE@Y)X`rdS5f^Ge4s37wS*fc8msOJxs>6ClvZsa9 -m0qgGimzd#x?3o4j&7CTvK_Jye5B1#k?sdC?@s%kc}CEc`ZedAYqMTH{AGq`<7NAut-w1DNXp*vN+O6 -HNa%eGy6d9`qm4vCs+4WeXb1(qx1AqJ>xr%yWY0{lPJ@C8o40V7sx6EWtsQD4W`Df+vr2*v04QxtlB; -bc*XycM<7b6baoL-BL3wIYHP0A_5u^er+tE4K69P(b+M0Ckm#7lhFnQ;%OWe9e>XGN>CQA|e7qQLkLB3)+aFAUltDzqbU?06Y2F!D4^u0Sl6tJ2Zg+VAf1)|&cFJ34JI%;O5h#f&=?b=z>X+Y*b9Zuv-`~=^_T9);hc9z6-{g%* -}%5CHo2EzT=|Eik8^}O+0979W%B1fg9A5`Wx1n1Sc!dOc(U00h*%|zp{sd{z1p>+!{nGyg;5Pp23#)# -5GPJW5X-cwZ>E7NIOy8^n_Wd8rgg1t_Hau9*^!o5!8EnL)Hn>s^8J8?)dzc0ozI#8wcY^niO_B_JeTt?2hcFUW3Q*Fm -AGmO9}AxJ79!YG!unrC&#nTa64Gx~&k5nz8la-OSPVT7Tv+Ezgr1y3j|p;L*74=NfZ`}j!1dZ3@($I{ -Kmj5bjZ;AS__k~E%9_Qp5QpM}q!j_K`ql~#ewcG=-eN?PpM&g*D#Kx@$K97hFAp1RaOozBpt?u^~wK^ -U#$XHPK%a=YH&xzo7)qdI@Z{0eiwO;a!&T9CN=dSF22o9Xj$o_{snN(t2_#5NYhs)(Wlq3j0watvV7s -1(=S0Mf*XH;I_<>YktX1A}YmAt0cjD_sB%cqpsoQrz?fA~oyj807H0sCXl_12Ja* -0x?72ImxInIWN_MGW%jnwW?JKu_j51rAR?6I45qlQBW%cgFmqFvZ&@D+&&j|x! -!1*66HRCEmjtP=)xh@oBnCVTX`H3|ipsoh(`0ko*Wi=GESLwJmFr*F6Vhc1$!=#k$M1vkjSFl>=~>`M -d3=I+f96rJ@A(uCg+y90FZUZ7Nu;DFT#j=cakfc*|MOPnq6G}nF>$Y}?pJIx%J4J~LABsH -jz|sY~s?qjtl4^mHjkJFzAs}H=!HWZj0MiKlMpTw(?62*TmC&d*O^bYQe6P_x+)a2~iquNe`#?#~pvb -ZwE~DQYbREymvR(qC=4XS$AJ~X<*&lB{9zHmV8T--MhY=X7_g756H$`E?}zYtYt-P`Tb~F>M}uepn%_ -Q;NF!QSQFtQXNZ@$=)B5UerQ{RB*GYA4)@tM%KvVC}n9)6$>M^utg{Uozip7fmD0EJ@0Rf5=6V&{>GcKqx<97%Ok3;2n+Ldw;;FFB33IT -xquvmj?2Xvd{^B(4Q#zvsv_lg67ap!_f-AF(N`S5|ru*Kd&gzi6+rWbe@lV7823HsN-~h|hAjJ+lfB) -wF0ei<)ieZ6jEt+Rl$)@{b;0ktn`dWz^_QZr+R;{ksWG~qNYu{8}s;T^iqRl@%KjXY8+TJ&teu8Q&o_ -#(+d+F$o3u1liMa5dJx#q3AgRN7jQh``S$-cJ@ng%k`EHX07q%P<=E%ReWF@u;AE7eYmA4;FqlFA78N -?*%ZGAoD87lTLoY;NAlja2}h_h)EPO1^f?<4of;!_XYuajJ2{?V9paHWtuUl}%_{yD^2r3Yy)_$p{{~ -4h4^rrW@XxHckZH7<-Z?D5 -|u(Nq^CKf2=s5@z+r)YcX_dttOv+t_Yu|0P=c`gj;Gg$x8B*WX -XnS?hd&+vKt(FA-+zC4bdE<-_Po1pA6dp=77iWCot~T?ciWYqOM8YL1`0smqWo#R*o?Pp==#caW_rSS -B~rC_W#p_kmRF4%J`(Z$-jKbp15ya-TEPi6aLw;&Ay5Nk?E7fJkjJ3~9p%_y+H-^He;!GSJ#>qCwa$5 -LcMKF5k0LJ5j+IvHIYh!rp*FXlv=R6YLOvlmszjay(cC#0w7`p&Z<1d%!tDSjYU3WEf#2SNQVV5?neRme-`)*7O;zsI~vLlM8(ujE -D&(4S)pxbn}*W7(WufoLf6#RN20KU?o|L7SIkrGYI215P&&DIc>hnKH;G?m0ELM -UQK50hf@96RdmhuX$jT2>A0EQ9x4YyWYtkK1gzjR3wL?-`G1h$4iLi}^WFd%h13Ok@E#=-1?&`U -G|FB>G)S;TIz^hO0o(nLol{&7GBE-nrM&U@msFR8vGyC*(lV~dVopDV@6Jv=BfL)YFORiIPTL1@REsR -JJtz>MPmP~;G&@J!Ay*DW{S69($aFhLDFQDIs!Ss -4!x@rmWgp+jq_XO*!h>>Ys%UbrdH}otgdipue -AQ|75F;{iLFpS0$$I->v|eIr$B7-m*6=bA}bL|)Lk-LV{unmL>eIYC$S)>SWQez!!;!+N?1j6ItB9#S -P)tjrP?P1(UCB72^xhcY%o*w;5t|_)z?BBjXz2x_Y>^8T;T)?B`JCKI+ -i#?SyfNV}rlGpidW{wz2G)+1K}OD?ffh^`z+cXUs>>i99h7+9DHi+Z&5=5hlO_i8(pkOoZ$2iTrT~>; -UcA!R)`zqg7yrp4Cpuk-{oa5{sWes?tTNWmKfs*+;5gC8$_|rrXzhpDF~K&Fnm8Gty7+WG?W!SXF4Pq -%JO1huoWfA5Pq%gGwRYfJd`tv!l}!M$au#UzqcAF+Rp`PXwB_bd;{lvP3>HLF5yQ=*w9cqRc3DQ9DT% -Zp$%L4(hJ0y@17#y0doSW>&V$+CtOWtl>uD+R-9X^50MooyP~Jk;WgNtSV3t@(I??AojKAq`p0%y~69 -~5e&N3!L{?iLEDq4Jlu%LqtsHQt;arB8#Q1UD2LJOlboPjlP%#;0oS7{A`J)5Cabkt3o -wpJarDyGJ`#@W-5u)t-B9QC8X!z!Knh$hU|H0nYMFnJq(VD@cX&dYbz*>*xM*JlC0*?@nlN2rgmkVs7 -3FXb?I^F(8KkX2}wK**9%U$tgMeG{h@O%&f@pKHvD#J#W)nqt^NIf&DPd&@I>;PwXif%rf-WQ -5vH;d1gFA(7vP~tbVxPBkb`7HArM!+Y{Ffp=_rVMJ&xEK?oIgy(u|E3Ef6K!7(hUL?(m~PJf^QUt?yD>I1Of0lP4lSi%EWUnSn<9j*zL?u^(4m9 -2CR7S&(M3u>PTE2n6Pp62FbdfS^*u1A*~(Bm%KsyW}9n6i1+rsmjf+^E2O4lnAj05xmF9`L13Rb|_@b -QU>HbOa(vqWYBoy#^1N;td_@@Y-Qat7O8dWbX$9VO$qRGwJsr3}b7$1z8=J^+8x~aY?(7I$Ef2i_^62 -;Q*7(>{Mqr3e1|vu+)lCUANh+uhiG$y0z+O7??LRD^hP9VED_6Ys^+)QZnHsmTg4yIqqDj&mF3)6uLw -NjNBsaV$oiLQ;E^!4qbHKlh&~xi71agvb)et3tqypaNbvss_g1;^_Cq=5D@QtGuXRwfw@piT>&M&QJ2KV%nEIpY8H26iV8BGFo$ -ZTpPm!>@EUhYG{h>{13w*XSJBA`Qfztrh`Zf~f8SvXrU#3)+0`gG?BIdDC}>~y;bDy -KtEL1!@++8P)h>@6aWAK2ms3fSzEf@a+;C`006oZ0015U003}la4%nWWo~3|axZXUV{2h&X>MmPb#!T -Lb1rastypby(?$^fu3xbz%tQ*Z#z~WQ>P}KZ2%XF@fgu5AYE*HS&bC0-sduLsr|{pq_ae!16fg`OitW -?x-o8Bh?1~+@o@WB4Sxx~z(~9JfX9+C?O`$4M%0bR4NXi!i*i`>GElGkO&y83BZzvZTE1VrT2!{cXBG -qC~uAbI1AzV?a51(J0onKy^+Y1YQ#u<~wHhREm$wVd@U!lDzH3*dEuRsu+fkoLMzl9{N*ugNTc -l5jy|JhlQiCJnXBMR7y&EVapLvJ^mk&9lXVrmtvDiJ*@2q{{LXO~SFccdVTldwyj78E1tIKu8`%&L!g -uw0ld_OR}J-VqN##?h{ul_>SWwIT50EM?Lc1ki4SjoHOo^9Kiov*VUh==)IAHG(+`lQmJ%8fm(k+B`^ -dKb5`Xk=5ea9!E^-?5Fg1*f%mLWKUdt7>X{t^xke! -JI$D;3;#c_5U=cX+T(Qg+~saS98BbxI@wQPM;s{=F*FdneXc6G>xEXY>=6n$u8q1A5_UJKX)p8=nsT9 -u9twKG$FDXBZEnz52|5h;Tsdi@z>kygENSy*hV%Cr(JgiY(EX>AQow<3Suok6e^38gTVXq9bH*YaFp( ->XSj_{rO-p@cq5|3A*EB__jQQ -W74lO$g-(#V-s-QBYK{Q*0;rQbf(gp@|L|2Q~^7E{II#`Wg!;T`Q6a+Ud -fntFeTfXHhnS|J2C?{sp~PIy;7(@sA`QX#tvNRMw7$zZfLlDD0xsJ+%x)HSyd57G8T2=4?s1XAE@}~bqrFC{ZmYjC&Xe|1)rku=l}LRZVAB(QK{3N2q|{DMPWs*4-4BQ1{;tWk -{b5vIBD_@ze?2$~YRdu^WCe3_gKSXcGDFJWa6|;y0aGLwyjz=LMZSW`3hGEfMJl#(YuLE8yOnfrQjbl -WwDApXd=}wtXaqUVU_);dB*$HXrr-2*${0)t$Nz6@f7g#X)w9q;H@%JF>Zvnp)Av>P_QV1dL#Z8B$ba -1DQiy2AAwanXXY4f~+!lIWMduJiJFb6UXTLCfVXO)lw;7!ys*M^0?%zVR8#a|4KBy;7O9&=o`_WezLCvS_{NEB1S(nQ12ZLEmsI!69<6 -}5#{V|(c2LK1r##3D}wa#p;dA5|T3BOUq>(+Jel<|8~(M`wtMv9@gmw!=U>raL0jZRRy`nYGcB!-EV- -kQ_ljI*lr_I*EqyT=E2do`Y5_B-5o0M{!>-vo>~(?G?lQN2;c1z>L)zb_?3S4O -NE!F~aCW=S%StN7*}CY^BMk7{ntWu_aF5yBtlZwF7LI!I+QCZqMe&}t*?2eX9T6skT}4d0>TaAD?uNO -qGwlw%rFND2x60HSh`Xrf%)?<@&gDGA9VP|P>XD@bTa&u{KZZ2?n-5XtV+qm&vzXG8!UF;VYyTEdXUCd*}W^p1Id`2st -up|yeszk(AX(S}e=Yr)zE)>gV^yBT4hw$-1gZb=}kSfm7-VQqmhCbtIL|{Lhzqc>}I~Rf;emHq|e0qL -tCQRv>%(7h51`ecHW-87zxdwW(OtLUrEEACn7Kxn4i3)n2=k^|=>2z}T_Tu9BwU;Vo*+FSLwsKn -+m|s40z6;&EC>Pse79JhNpY5`{Jpp{2pV)V_xh1ZbqnpZ;FqQa@;)y(ZF -?5z6loAAc9?LU@`70V@*{o`%PV4MjZjaUnjkVNCN5)&`fc%F=&`e4a(Xm-V;HOUmL2T5G`ekL<$M;=t-YLokXpH(>0P*>C{x7#(!@I668q^ma$VFaXI$y8{||ZBd|Mt0wOPtutY`AH-Yz5$#T0@t>o360z$femIRi_=Cc{;3rpG^UDD4mRkR8+&Wqy|`-!+r -&7CI?h{0Hz`#}_8IVr|`x!m|!m+LLDx{^wlHCzHMs$&5S++_7ZR=ow)MB0zEjjlRRqXqm8EA(mvw2xf -^xo`YJiV-=?==V>Ue8SI_mt-&!npo$-8pVMnuvAK|fHeUm_N)rL{tg{ttoUpxecF))7So|&nRq!Z6Gi -9KXx!@6-t`m>RK!m$ -X;t+=ttESiTEQjd)UFP_;;P#Z2F4`^WTZj!!AB={)k#V>g4v1=&(l}%%%<1{;n?Y!c{8d)9*zRPlbe- -l&p0~Ps4d8Awoecp+J;#%6TIO8MMv!^)BEe%!1!S+MWX`h;=a(2biJ1lo9T5`5IzaBoGXumlC~6Ez5- -LMA|Lu?hSqP-;^T4_x_u+*OYx6T@LsEabE)S( -5*%h87E8?*p^-_Uamcey|L99+F|o*DX+bqqr;qUPE*AIr`4~RRaIiJ0)M~99u6!jZ*o -C*3Xos>h)+Q0?b>9-4q$4}A%j;tu#rfSOO_f*B4t_j>N*?9l>(SGLrb_T{(4`>A?Y_$50N<*gE;9i<$ -OQjyXH^96w(7txq0f5=Oquf0CEx$my1ph`Jb@KZETST3^Bh_txfxVU|Kko>G#Fb-cJ) -C0sJ`0N$u5fm|saTNXSI=Pd~&rprs7Lp_}qS*8t1JZB=t_=T}ZA1eR|BxLXthB5Mp#Z>6r2NEfm`UWz -la(2b~1oXfcRA3Vm2wWz(=?W-F#uWuH40v20O{7p14FXK(do5zAdA3R-HWie$0X$^Rz$$6s`dW$r)b_ -(q=f~IA#5O%T`SJ9pkH_!c;?v$N;Wr9mP7XR4FJR%bc^rxTJIu_-y=o3?Q|p#}0l?_+Mg#IGScQiU`LA%NdtxV$nmW>H>d~L{HGrJj~qRV`g@p~=>NfPqZO3leeX+*SCCNa-lDeDVJC*rU%o -UNfSP04I>2IYXNlnr+yaIvCxfZB?GZz=~GJ*bUV$41SSbfBjF(=NsPc?==qK8D(GDs;t6{sl -onE{K81Av+bYxCKU-pk$-HxOVVPW@ee1iD8XKnb5HZ0vZrqV8l8*K4Y(54Kc9rbgk}!p(QYXgNpoxkJuDBNTs2}QcLEMHFhA)*hWH1UK(>#j?$2jvWtQ!{%EI3Lk_m%Za<2Y$`yzcSkb -sKf|3g~EABoJUTPjkRWO@`izR$>_3F4}K(6F(05DFn2?XrCaTv_7X&jWeX_&OsP?3)37*ne!RuF2P0R -FZm>f$oBu{@tcq6G=fJ){tzSK6pk$P8fz&qXD~SI_zA9CWx_I@Rm?$dMEM>~~Y5&b$)-3o`y2Ab3@wK -#0vUb#5rtph}{dlm*-^jrIX+xy0D)hy}#@^)#fE&EyuX0%6vu@J8jT>b1vCRSJ=dtVwQ$MfuQ;$wVXI -z*ibUEdEjg|MtQSz8^bkRfhsZu;7<0PWvd;gN}O-825;-api$S_pGIEwOoUKZD#J~2F-~#%?o5|bHa+ -~o)qiO?Pi^M1pTGJt8G~C#hpm9CF-kDsoCkUiracN0RCdJYIA+4mQ)QZ+1YgAwe5U+UrHorIA8iT|EfKm)He=ONy4pQeLsCq$LG -5W?q%1`mV2!q=9!9}I@Dc>8J63<>vjd;K9Bfu!0}XJ7@UH{SD3ZKzYGob4SeWTp29asWeDW|$rzCkz6+2B*E!K-ME^BlzhTB}}gJA6_tv>`beT9;yywq5aAx*< -wiYiEy7LyJ@~!HnFK$zb!p^kvJ -205pH}Nv;k}?y^wuuy7o|hDo5y-EXq=Q`>-9Yk! -1>JQ}6J*h@sOSK{NNt{SLT9o9v^GZpWNH8&l^r&q6hu-WIrBJ7BHeLr-q2bvKz59RHh -aTFIogF={#8)>bX@rVjkAG*vIeH>d=Z`YjrXnsxdXiYX`EUN5Epv|vBSf4Y>i;J;*YbJ0DErJ{b~1xD -L3&tY4-HauOUUtV7iudgv8u2OynB^Hd#bIyxAI=bM{VADHiW3(g2OYo-z~RaSO@Zdpc9z40C~AdCsn}hb#llMojrSg-FBoi*o@hMH}t!z^aQW-Rv>=AiI!$5u?rymIE}=EzQ4k6K6X)H -ej=dEq}!1`VABoRUA5l84%>;L17sHbp>L`BsA|4S^25qmld&|=x*WU6du()N8m>27TIpxwN-5}@LfAf -=u5+Q_v-0+XZRdXStcM(qCaC0hKjUFus$%n+ilpYVDYh!$OB0aG#yQhhxXuAlJTuUhtK1#a;;eO=F!0 -?hjH!E`$8@Pe?>6F(DAxS%2A+2c@oek*FWfM=ahN^vlOs&( -LKd`z3se}zS#6P=v1HV8SQY&meGkiUpN+sFG!X2t4$;`0nzi5k7!By2g+{$=7rpK6pl7j);+ -r_{n=S}cDn57b`f=Z&<~H!yPQ9Abl-h~apv5$#=IWh#`L>X6z}sIJ(xV&R2QmN~UI_9{meBdikp}2^W -ysf?jP8iidf(Bn-P{))ZD4Zi=0-_=HT+qqSZ5e?VXjX}3^_N~09uRePDBAa&w$J>K(CaZ1F7pEn#2k; -6hR(Dq|trFonfZfUE-P@hvvBW%?K#LWx)Ot^9p;d -6;tqn&Z92VySyDAUTd6`$yje%j!l-ZBFR(*HJN(4z@UZUAhJmP$QumAZC<7G#evWH8;fra+*~6GUV3Z -&FB~Eb^V=m{5H*auj6}r*Xg9RJjdU|^@eDUNalVC3D>smrcG;eVYZ4Vh<@DGE67}_bybUd7{VcnMz5l -;wAj#)(4NtPz-N#P>~(@?9V8YaOoM`!vBz!D3vEe&YRIm8w^l9-t-7|mHI-KG$QXmXX{28|6d)b%Me* -OWsTn^B#PG|iMO2ey3^HrINI~iDCcgy8>u>YPNpS^lD3@Tvz>@DWMy6lKUOI$h9)(5x7PsBqUW;c>A=P|wuxOOm4TX77f2K4zHd%R`QMe -cBzWa&+H?B2f3iWLEN2UJMeURLqIF*?<3*W5GkuOCBJb7gZcUtei%X>?y-E^v9pSZ#0HMiTz6UopoXP^poqw%6it!r3^*qGRG@Ns#2YIm0KfsFlT --;u5>06(z`j?=wqMFP7{ihbzhiu(>Ylq;f`ct(Gmk(;`h1(`q*F!*r2N-9R -my$vP3ug-C-%a~i2>^Vh%q`m_Aym*!Li4>F!LjSlkdD2(r$Q5eYBNYhM&HquSjR?nqBrf$At-_KOC*5 -UmkrD{;4pMU)6?~O!DBjc3%bNN7B`AGj%#{5g+&7H}|Q+sS6^)fUjR568y7E;UUn(noT!J!TkQj(feu -n_uQ)+tpa;x#2w8yry6RD>~1BtZexvt+sm4PcF$rz@c)#AhUoQ9&r!D>D^jOV}h*);JF%X{eemBsrtv -gj2I>W>Q2H#>9K&7I0T^qRLVNvs8xxqtqc5MA?k(D7T|<85TMi!1~27Y-YfMZLQN%&BFP*4$!5vYPl!P9-n7AhEh4Rq!k=zb-t -HD$}22BS5c%^pa;}I#j}uknU)RT82e(X9;GFf$3IpnEY2-KG;&Mw5Eol!AwXd&3(rA0JS{e?fXg(qsY -#IsAp(&oZIwRd>*hvWcr@ytkFVQ9kNgo0hW)pG$Lo;O9^u`o)3rao=wFQq62o?H{EqtP)b72bfBU^oo -xI-$LvJ*q{*e63LD%;>92LKJ*1hWZy*Kn4CwqOQyN{#??6_}L7GS$4F~;^2NOB~?)km*A@q2cUT^F|FV@I=ix(PQw7XratbGLrLu=z%fADVTzquF -#^rGMGcvybz!MgVAu9vF=wX<&9zpPWIec68F*>ip146RsU;q`@Qm!PqY|7T;r-vbpc=6-KH#6lfhhvU -uRYk%a`sXg>Z?3VLk9}3v5px?JJIM?%X9CrC$WB>~xUbzDGjXfQ&-38i+kMD}x(BjNBN{YlR&pM=n;A -9kWVEcm5(=gRSuj#JnMMF7n+)+3~jfe9P$zSKBqxYFf0k$hOu$9&5s3n@nolV!hZLJa=$`s&t)|Nl0X -LFyX$z-0TnU<3Yg-cF$m; -(7+9%EVa!DCirrQ!<+_&hiH$Y7!wIPK#g(96SDpXfv>X2sjgLPz{^2%mfB36Yt5jalj?;yUsxKQ!=(> -Qa1Q9L_dX8`Pt%x$|X{~gvmGg3(p$HA^TrDK}7!?#6TWSn#LuSzFEN;afJv}|8FL{=!7k_5?}z3&TCzAt&hTQfQTA_Xh1#p=jji(ON%)TnVvF -GMFN70zujKZCU4y~KjJmk9jad5O-a<3NAWpPdbs=Ey#u427y%9!@lQG0P_uTuw1=D)<^#DsEEDRpG~c -B7nE}?i0Zxl#j?73_O?sWV9xeL}8k8qNun4TrI*>8bna$u{<=FtK%pSSIZ>A!Kzb5aywKdr?y?I*|i! -KX;mEOs>Att_yMraPCIn=v_rx`;M3Mka_(}i>T0dEe~w=9=1nBy>Vb|LwQe^MbKLn@Rp!?lt&NRd(xlqID(rYS!JGRb>#j=`4(bWLH?P$2I##v~Yo79UBTcR8m@D`&gK)YxJ;D~E>06^Xu}BF<%wfSu4V!Y`LXQiN7bgm}V>z$Yv98YW>RS&T0bdQf;M6rrbzDNc3w;U%gbn?o2KL$G1nXHu8Qn_5zgUyXU^+K}W|L -AeCEc?P(JfM*9&C64C!Xpm)Ua*juB)D|4_#sxS_*k>c%P!vB&i`ZSeQVhrDBQ~sH^a-OtCtVHPtohjz -!d5aE$0$CShZ!uRmYz$PfGV`|JV-A`*KgCeYNjZ+Bg-#wtxp!^(~e4b?(~QdSf?kZyZdMw=WfDN!P{s -{8n{dIkiKoHkl?xzsir?wBW`AL6gmknc~$x-9z2a;JyIqRGu2YbAER;aJI8S;~1S{zC7{?;x75mVbvY -nBggr9q0|9~sP8}=jXgi#+F8rz*%a8F+irF&xW#mq`?2;FQQ3EFSqkMh@6aWAK2ms3fSzF+NmX -FmQ000iX001Wd003}la4%nWWo~3|axZXUV{2h&X>MmPY-wX2j}|$VmSm3fWue&JdI!f -`RBeY${hds=W7yYX|a|Iaz`e}8- -Rn*B$@-(J%qul^n4EXRRgU?$t8X(Zw&&jHSOo>wxw`aIcqv(@tTKVJVdc+>y+&CO4*Ud?Bt$J=GGhk^!L|4UGiblsUK1NpTF0DL5OI?XmZqJSzVBWG4 -d{QD~0bz0&qB-1L`|kyhNCam{w#5hJKLd7=>vdVnct_N&$Gf!>E6A^NWN&3mw2Ap;_d=pe&^Y4RVqd8 -r8T_Su+){R&gLcL@(J<w{gg -f~ke?xex6Mlphtec^F@S5iQ`RWWQN$E-BTbtIQw{W|Fx8v -(6oZbZ7`Q*h!15FW_os;bFeHgc&`9jCI7A6PMPw%;7vok4fwL&{W#oFk9zrl_$S$A}8%NB*(8x~?GVC -k`;F&KO>&1HL0g$AKa+Ha0qTzBo-=H<{Hfn)r??=(j4;~E+mo$Lt1*Mb-1YEN6W&kYngCkZ4^!IZsP@XKfnbeN?WI!#p|=hP_VuE|& -5&nD5}@Au_wPllAO+wBp#@i!^A^R<1Sp8!b&ym^?nD{Rqb{a3PjjQ{n;#0URC?O8sn>xNeuUWk8W&Wp -An#$yW?r?%@S^bpv^|g(=yW?DKgOkKxNaK>06WT4z}s2?fbv{|A;PYhNx$t^2SW;N?>|GM&S5Z0s#vB{3K0RMIJP+Q{CzIB&P+6NFpuFB~DT03 -gR4N*(o8BPFb?aNg*y2B)$X}7P=)au42oWLQ-_Gt?N|xhkYXW)LDO@t3;5@)Lj{kdnauPkRN0u@H=?g -)GO$UA}70&JGHtvhk}YbaS&lY@1h)4iSA_k2tRjmRcY>2$E=I!!}~uh^BtGDVA4wQ)gSmr+tyP41A-veEqi$OiRqdT-tHs?~oeLpJ%z*;4k!A+v -{8)jgv$!wHHr(#5{w>5Db6(3L1@+1Z`E(xDfgLU0JN&SkzyyBCX0>dw#HN4DXg?^O%X~`z5Xon9ZNqF_` -Nu;)zXVh_}4$`oLaKuw!nipE&m;~zd!aK{|DbQ(fO7q7~D)caTsL}yS)t__g;?CKXnmgzA-o>4>`%bc -cwTAgwz=~4qwC}9ZyUTjuY?LoI6Ck$q;V$6?Hn+ks3|jc<8 -(X31$<@i{?=yae?u-Ud9*LJ88I#45oLG~BLSdQ=dg-HX~bBXs+RL+ez~z5xpQvs -&oS2B_YL%~0f*CNvt}yATWv#Rhk|&QSKIFrm?K>r9{*z9MifJd#Z5gf^w3AZM=Q|iJ;U7$bl3oQ1JDdu!=$ug7d;c&DXS -r`T~c0C>gzPcc|(fam{L$?k?ZGAo-yDZRaU{41^%T7avPWeSXKZ_9)O39OYMu|}j`DqL>^(wl)l@1=< -oQ|>;JNJwkislJPb|M^*NUNa9Xf}7LqZZsVfuVM(!F^NC@bkA|L(bsA(+0fLp|?)JKnFsD1?LG$k9t* -YmT2|Csj|)leKP@Y$m&dxX*J$ZtX*ooA)#F)-W4RLE|Lcal07LXk*wx)w)JL^u{`475d}}8FW@4Bq4A ->%i~*l{5Va?=#Ii>1c`OJ_GRbXG{kB8m62*QXo^2W}?$LsaoPa1v5t=w -mH~;h7P>z7CG~QRN_wGru5;v-HA-_?hL=MafQ{63tAmW-dlDjpH&Q1$#+nSaSxQ6TLPOAWOnDSIx*p2 -!OMf2=5D&;O7F}ypd;8rqnCxdR98Hba{fAb&XO=`S~k0*~ACwDeLrkoBnYTwE)g8H< -)eW+=!bv82y_(lDJ2X#|7x5^?s1`sVN>6DX%D6@bKwoIV1ce_wN?VSjI>Boa>?Ea+QoX9~hiK(gUFA+ -1nfxaJkdDSyKkWcEmnu@x|J_Qgun_IF-L^5By6m@wY!n;L6pbe%$1_DC+3#at+NnK@?7)aENAztHHrE -|boOViTg6}v_jTB!eo0S8J^zIn1fQb=P-=*mzqU>cFi`!!3G`QsnUY+1X@+_UnZyCNo2lKMXE^0oF03p}6)Qdt+e19TurrpVd!0;E816J{B>J#0tZKS`CwtLp5+%UxR_&3kRt>_ -r#+{WZRF%d##jHmr`mM;wCN^;_v|{L^S63t`gWmk#9J-KM|yMcv7&JqS)kazAc`09dc1d#6e&^@qK8de$1)*MY# -q1D9;V@FEWTNh7hs_C6734O9Np4%mr#Om7!}diaQZUwoHL$Dhz}2)s}BI6st>P3?E0LL7E^CJ1YQ>vgr6X3Pg9x?q*Y8^YE -TAcBs1LuFXjpoX(UfE$tfLMq}GDqxwj`CK9OvQ}gVo}!Zgs(xAlhal(uNVL>dQd`-sDotN!&|gJ{Uet#hLRWDrkXsX&)!Z_0=|x7ZAELKB!hg|=`km#zqNKL6T~(Ta&McV`<2~Q0~j#wIVw(>PxQoyLo>Ntcwm2k@kb{s^0 -*ljGkpoB>+IDB8#N5J#5#&ROOJc1wtzBdbbigOk2meL`-Q=f74-qbHQm{$2rc5CCtioQC1GxV)d?Lv8 -tB8uRnL+<+=Oj=k7Z{cfb3&``11&!1YFf~coLIZJ^r%E_mZe~L?TMU8y{)VQ&S}&PZL -KysrpCXc=BYy&|)<4yWAL>MyjPKd9CVVW9S`{toq)$A_fe7cpk*&e}9S;QXf|@0nO4_C7kO3*o{@%K< -v#qn6x1vjMk>f{|W?4&Nc(#kPNbjfXMhp{hd!RfMdZsvk(x0tTl^^h|>X=cE4{QjTncFl+s*Z-0(ypT -dnpRT1x}7Zz&al*89#BYu2G(u+??=w&zlfuPgpyg|&IWotX9Pc2SG$z(7 -2@G8+LmHVF=@i|^IJw=}`rX@kp1+!LNnM%4}Y>eySOGP$1vAu!1|(In440%n-0*O4V;P$RQL`Q{e;pG -d(dt1YiZErsBPwDC?Bl>h)mWY?Fns|0pXvQN=>BJQ%yxajK)*k~i{<&kp>srVOWm3-OA0SfsI=kRi?A -XN||6^Di+XkN>~`Oc^6A^4djw^K-QS7%VnRie3_W-hzu%$$f253mhP9X(+ZMZR=`j^WVfR%|8s=8Nl! -k3O)d_~z4$wXxjpAAP!Rgmy5i5_LQHBdZhwZwG(N66yHy$MtdsRLwUkP3t~-tECgLpZ=x|J;h7B~Bz8Rsw3jEv`+ic!QTwtik{e+0nH{QoMdvPV(a7KzojW;3qYlsL+}$-AS$w1O>ztzvRZnNIZ;!V)!}Ir?o?qM8ZP -D(}Z`|0u(6~SUKyQD@Z5SE+hI1#!VbtL{oV&XYBa82Fex37>Q$;EsI*r1tStbP+o!0^fkTTW;Ap~YKZ -IYDg*(D_bA}PpzFQY371S<;?q0oOK%pR2o*)h^4kE`TmoeU60w+o}}=Uoq0S#fhy|^c>F$ThEWs -puzfS&x?hy0;)G_tS^;zWD69OGrQ!~pug4cLK%mZh=PvW!2G-Yc(5xmiWayToc22bcLPH_b&Z#y)x1t -Vw9XWt{S_2IXP{8#Z(Nwa5{HGmI1|;;+juf$H)4Z4p;@gL0pmuDUFQM3)OMI+@@wS4NV~t?B|<9sKVCUHuz?kSNQgo2F+G0DQzG3 -8nD33nrUYG60$vBw>9hXU{~W8Ox$hx;-b^iGu8iG5_RLi0LsEkDM-owVY%p!iCQtTdajvWCxNYlRqr7 -Xe0nEPpJgTSAcgeCQ?V40bv}txAH~?AYkBdh!7oA0gEXhO%Rs~e6#?ZdmM(7#8+^Y2n6o&QNeME1B@2 -rN&4Uz$4*-@Nbq>f(ILPTag`@;RB*n8XxP)9r8TDZ5(Grh> -#(g8aV+X366ak%mn~QodqpiCPA?_2U@sHfnv>zvT(Tw)xh|n%<~-BbgrZ!>(GO;kTysh_S<5LBIz*V! -b5ZDFd_q~7zaqMKsFGhkWLFMwsQzjMt-7yH)+AePTx*C6ZFjla2uCVA>YSaG!g(N*)qt0B|f1A{|uw( -&Rh;>-nJG3;3}w<46OoMfk3_h5@EH#4%%r$b9w7BHj`jS`UToCSM)Mz|~GCR=-7r6v$M0Vgo8r{n~XmBSaDAY -)4%h_n_5GF0*U0H5t-jBg{u3@Gx-!*v058qWpY*nb9GOt*?)a9so5~XU$zU6%75U#HR_H`iKO=Xk41P -ykQK1yFaT~@yVi1ew`Lb9*iLb9ilLvJ+%JmzT-6!QFFH5zTf8A2LF>gD;>XzmatISqGe -H(HH%4*e(%4H3;(;}N=jaRBf4M10Y&rqda)#7(sz1}m2!1`Ui)2=jfSqj|z+3T&IUBxVLjcu#`@L#}~ -!6}@*w(R&>XF#MnfZCc$q1n*7{VT1~wocu(CkLPH6Ndagg6!?iBs)0aPY)Ue^B!DF4qgIS+PUBPpZAP -?moDenOfe8#+&Gb-6?~v8q7J7-KQRu6iMXULE=&)7PLi+|NuzqZz0|OM86}Qm01}N}7q#CNUO9h3SrG -*+Qs+KyzO<+f;Udw&AT4rj-eBE<4=k1!i-o|~mHtrkRxbND=KO3Mxz}y!4ivbE;C1|0)8lbJcGKXc>sjhj?+{;XX5J~d3 -u&FR?&T5e!tWPT`i-}P_|^xe$Q`F(<4ln>pmMX?LPuV#W@9SHQW(QBu$(b+Lms%ZUT?NCKEtf%9ECdt -UNwAxu$zwMC61HW4+fLe(i7U5^fTf43Wc2E-2f)rAG*o({#O{156n#XL;o$#2ZI1fl2UQ@1KrZ(kjOM -{wQ1~m40c;=GwkKfYn9K1^JR@MWupv57_^e?IZ*Fx`iXcS0m(lxXm&FEe+1#52#n*s*Qr5jL<008l0R -a8RN5<9F8gR7MCB6I{xhHEW~T?qJuK`gi+wZi{8mXd$}_v*JDqP_WkcX{(Qu;zkV|K{ec&d}UZs}ikb -xgS1-0>1ujhlKLM_D%KM4mnnl66H78=0g=L5qyc|ntyEL;vcJFKWi{xo|@}BE*q!amYn)17OQgO4bcK -{{N^-}hVjN5K?6B2cjDR84gU!OP*$j*N@N96|5^5Bd&4iFE4v0MgSZC5gP4JNyaN&0NbIlxi|$}=Gyy -g)s`##0yT)Bbcf~HZxG7FTK!M>@L;HqY;Z|i(zC -<`g3YISs?tt!257}D@wgxO3Jw{0WN~rf?j!a(&HNkp14yqWU@1T1mcG*LxL|(ahGjQEjK@_)81%ur}a -#m125ysg|Abps|JIn0QI7+^Wu*H1&5rW9G8E*m3XJ{gA0eQFNZ;A^T9 -Fl4_*ve3Tt}>u?Qp$mKCS~)Ng*Bk*h0rXiXcO$+|M$cQwg`ei0uS8k8|vYk5Y_^Et77T^b-`EqB!Ws{ -JJwnNhVuM#U0sqJd>`~3u-Nfq)@ge`a!p~xf$*>(*%$O^qa -oocg?hQ)RZjff{k6WS?@E+Y>=TRjCpb-h!)C5b=R%VB4qu42Cf0-93myF^4W{Wm+;0usXMC@Xp3D -CoyQd;Dx1X+38%?gO4|D-urKD+o*2gBj8jpD~3BZ!5LKL%b`iZ)A#n?u#I`!*QO -9>&!3JZ#!g9MA!Zo{u2POY1O-5l!?!m=|yUXlLQxH-82F_h)pYZDO<=Oz6nJR=sQW$CnCjj(saJII(F -(PmCoAMzwHoKVDZuUPdh+b6Y*bmb^0jzT)udZkiX^mdny^?`nMJGqP<%FaCx=7XrA~#Ts-mLrSb3;8V -BDomTvlgt$qU#?JMD5Tag_|;$5y@AozL_x`E%Cz<{4c2{iK+3^T(4mLQMDWILhOnqQq-Q}cFyb@6q6* -Un6Q1dXYyZ6Q`NV42oN}saDETNA^ -o}r%TM#dNY{PV|Ip7OxG5w+9X)a@B+ae7!D*YfiTyjzI+Du|UkUUW -a`dYLNQ?ixN5awLzwxbRWfR@m#n{qz$)>qN-wI0 -PqnW6UVoaylewe}R9V>N{2K)6>C?Ne$t5M!pX;TqGG=eg2W8}D7UP;J0pQ3*JYd}f%^B0C!?Q$ek2)( -rOk=@C3t(hdbXX%iT&3h%l11M}!_L!_Csoak+Kt0lkEXsOgREpwpwovL6)fqMO?`f9w -x4WI)L1>whHEA#C&(K~zRb*TIZ@jI9pIDt?48RC!3dy=s>I~1&>k0kIzR{v~~mXnU5_KA!-bga~oOCO -v6U?=0@_VW51&2i8l7`YoAKJoY<>p;{+nwm%QTuTF}cr-K0H|?WHwf1v8(M}f0H@D|=@alFHJ;mndn1 -5b4KLig+WhCJ-sa%6s5c7vzWVw~1y5@}j5~l>pJ>g2>cU8gWLrnuej!@N8ZM}3g;%Zf2y|SNH*JZ`>$ -Twg|TP4v>#(MHM{^?yxalDIOx`~y6f?@Q`ef6ez@3JeEpcQ>$9-#%sXFKL$c1yJKYt{#|R^pu_rUmRZ3u9_}8hs3;(XAS*h| -gCgu+=sVR1vJgm&{K9UxP%Jh7*8CP1y5%a@8%&ws5h=1fMDZojZ=PHuFiD;{(*|&iachq!t6=O1RW+F7#_WkYH*~w8R((+p=@<`Ot1+Ub6&!?)9FMz*^o;m2SuR{{c`-0|XQR000O8%K%wh9#ch%P -XYh{{{sL3ApigXaA|NaUv_0~WN&gWaBF8@a%FRGb#h~6b1zIuLq$$gMO{%#kDD+MzWY~ -XR%5!GR~$s5-`CP!4Wo%_TP7gWM7q#&^*4^xU|#7h2Z=ME7Yi& -J%8-ZqTBe=QP`#Hj7L!4Yg90@Kp!`zaGxAEXE?xBEiHo0>j138}qCLo^)Tzu;~`r9Sebv_L)^a-?%s8 -4gmW}s$cM?ztaAAVxN4g)bb^u1*LaJTG`%9RcpA17LLZ@s^KP2vLU=4nJpyI@+6CtDQ%*d|hk);m88{ -zlOnKemn%g`bQW9MvcP*DhI>crf~iR__+*0j?jf`WS`2M(-#4I2h`>k=38#Oe7y7KD@)*U;VV`{U7rQ -^)msL+L{B`0%QfHRxdIYDXe&f^(ujbG@05Pzzxj{a|E>oUL4_FJz^Pzwp-3UOzlQ^S`R!4<0am3SwFRPY?RDu&?C% -TMedP)h>@6aWAK2ms3fSzCEian4qP001NA0RSZc003}la4%nWWo~3|axZXeXJ2w?y-E^vA5z3p}zN0KP`pHEQ@jxPWSEK0V!$ITwvBTKZKw{3|>lI`};9H0mk$!Y>sa1}rbtNdH}`?I7e7R&i2SvPrt-->CmN}iRQ{qM?aC{rf4SH<)yxvkdoS& -~7Ubv|2{v#ea{1_9_|v)mjcv!cn)=lNb<*HyhYt!DXQ?Oon7Vttf|)ZG*6zL9 -PjK!K07#vSeOe7Sn8=Tom&h23D@JqJ$yd7OShGg!fAZJxXAZw)9T3KA#me%t2LelGRmS=ha1; -ByX>BQ^D&?lD}I*;l@2Ivqj$Qz`R~%H#tHmYN)xof`4i#*W~aQ$44DWCX@ZiWRyIDK|g{Sgcmn?jkTa -d*d3}Vb)cU`1ATc>!J@2aA*v>&64f%V-HI-TB+NS=9@f;AYdT@`b72jCRue@{-?-VEt#wy2N(?!M%W?9IujT28%Io1TL-2yjT=8L) -nHJ+DZRjEV&NBqQqbclcSU5lP4d;bY-)=hT+2;Ce3=etm+jFZZf&dSL0c>$_B&9WM`4DuBsUg7 -sj=yX6t#*BV59&O5go;Bg;7;e!dFK0uxq=c$>aW-{6d-$5%GHaVgG9xOM@;$Xs)0}C*L -VZwA}&<(`#or@X)n9{(@aZz4W$-gCoPe#dS!-FK*g@yS!J4n7 -edGg60;Meo3e0rT+0Q#&;nHd<7)NXbX_|#be*z|f`u8Ku|gj;gZJK4ZYmB}aRXShQUn-Kp1HGW>htju -~t+o`HC--^n7LqO(xZNDM?X?VQw6>(8s45%AGPDb?Q8(9B3oAZws#dO8rj`1n%1OD}GwnPx}W4%Ip@& -^!FKmYU5^S>Ow{F)z~)YYm2(#XG_XY)BWrMhH)y -@m08pGp5Jm|{fB(az9}AMhB_SpDR0=D}@NmoOwdpAD?P`F^o3fhtvV`*L}Gxn31>y9!I59QzuU3>JN1 -ek}!m)3(cJD;{ -6F0v_*?!4YG2Fdsh3gI^Auh4l@8K3OHp)<{vxVexb0HVWgKHne}Od8mq=hZuybB0+p2Hcu3ICuzi3ceS&#sy7FO4FROZHI=G!B+Mh&ne(E&-l38|ygq(=G=B5l$;qqNZ>Utzo|ysy9lOb#)Fn31@W)G%@-5*-;ekS3q`ip+3uEnF1k!sVlR2iqqPp< -G;bFGp)wB$o@Xv6#0BMh8`oO#|N{zT8^jHVv(1!RsU8M1R&_yJ?aW=4Qk8NNPXFM?!!@>?Zf2n7frLyoBgLBfBfkuFsL!@G$!AJ-uNvXZ{rt09qZzJj -WGS!-kaViIj>-=9rpBN0I*9nfV(L&=;_Y*We%tQ45l4m(TQ^k#YgrCIgLH*xmvw5`0R|2fp>r-D6bJX -cFP<{r>DKw4##_Eqa^HTSbJ+_?u|dD5gegqj=ea!zJ&4uM0MPUvNJw9K>%L45Tt?y6}}bL$**$!TF?j -1;+q1fn(4~U7~jlq^4U=dwDttZ&I+mlCxNu_IWU*!c;ne5*^1lDk;g>Fka<6@kC0)8^K9Yj;_WHfPmW -7K8(gP;EmkGx%~$3KXd1;!VeOQ2>Su(X^N$X&u!O0En^Qz0*kyMVxCh8 -Z1Hgv78e8@NNjz8^Dh3ydLb4qy|Z -BBSrWn+{|GfEP@Wsi$HQ7b}`QVF#(JbV1R;TQY`p8S;Z*WsgK54o@OcxXGwuz%naW+0a#`6+$G> -l)n`@cs;Yqht{Q2^>n^GrWB=%LBlkMV`30DDs-_?L4ka-oD*=y)G(sR|J6jXXM;<4JknHs^2zuJt`R%>JF -v87S4lZ|0NZ4_84PXV#uUEJHzPQ>Mhr~qRX{!FIS{y^`MF-s5No)Q*Cd}h@kaf{_8OOJUCpYgoa`J;sLF9=I*dLIK#yu#L@YILQ{-yJE3k(1r!>7U`?mdVEov -0ioxs68@WFI}W-?$KVjJb8Na{rUNzv!ZddQ8tVl#HGTDlWmmW6fHr~_P;e7;B}u- --)0{v7F+%w(5vnOlfspE$;S_K`-5})x_$a}}{t+xox%__RBoTX5uS+C|k5DLD0dG5pn(GFIhEU$f|EL -N917?Rr#WW8zAfSz9ehb7fY-faREY{76<$+$Grt2E6$}3$5r}NbtS&HKeUOu>S&I-DTR8y$a(7r?VuR -=m?S2rZmuqlfbi^6OX`8Y!_$M0&xWI9kzUr-dO7%HIJD428u2|+btQ~2pIz!IEvnxEIA%40zz*S_9s?QydwZCiicI6 -ntSP2W3JnvSH!S;|DGqNxN0vWX63-7WrYLda{m>U=>yPB;x*Rvtx>&Bv6WXo9lj&!=0rQ;rA8Kq&jp_ -3Ic#$pJ4^c2{o}xfggJ^8eWRPRVx7cI2mb4z@EdbBm)(aHr&5FA9Jbwo-+O3kbN}lc}N99dXS0!HFvN -T1q44DUdS1lZ@cWxVSkTLPrmJj!OXF|fJC4G9IPOQ?V{uP=Z(_U74k3&}Zbt2(2-Hb$|Ud4;D@x -n4Rsg12b-MYf)=SUO@yb5a+J3=jm>RdgX`83l2BIU{0Ujo1MEfd}-)jmO+izRgOYZgw0mBB-LSzvhv< -n7Vnkq6*i*oZ!%IGBt%PJg>?NAfl^SyDzKNS15ci??fASgp+7HU}scA-%Wln!W1gyvA=E(fdL_bj?h2 -eg^a5tBLiw{=8}0J!duwZ83*hbZk)hrrF%p(+k&o!#=CVGLB9!l5p8EQ3pxFvxVDpv-3zy5ms;` -q&51Mkb@m(O1R=skM#>buv^kGw}DFnsmhTkqkEqwkNtd3B1FRl{fLU1F}uUE3RlY(hF;IEihz-(Ofg=>9o0a+}q -&fMcx6xjAFuP+eq&-K)=x9O6Yl+t2JBJ0NreO9&k|{_MOVa+RECO-}sVEMFp{RZcg=5^(*5gvAZrma! -(^*nror@w$-DFVxM-T$bP|Tq~$2<1u{02?f66F|z12RZ8*p;6wre0UiY152OOmP?L=k!p&YGVpgkX#h -ptq8sSBHp5nPusu3~d(eYQu+({^#0&UkjLN(@U$T4eY4CU^e?U&tHG!f*ehF|tc -nV+!^Y5^umN{Jf?J6J?UIDxw>psR;1gHG;$ypaTXwN}Y!`TAt>Di8|j@x1@Mwp36@g&JsgGZ0Z_ySnq -p@W0dPYNwKGeihxW@3E4IRH?rQ*S-~i{vk=O@k)*Ilj=fnYl0bh(wJ{=Q4fU!4~-8G!S8UNh5M2x25$*VRQutL4agQ6%`9 -@Dh$Gq;ut0Kclq-nR6jFV6;qKgITeeU*|L)!6qz~70w~NfoedupNc~i)tJ|Y-m)E2z|*b&+okJwGVo+ -(?49^OBAfo8Qu^3(jO&qq4%K;e|xex#-VIKaiuA@@AVy}8z((t>S>)aV2X)zPcHHd -%{qEU9&}mUrK`wx9PE1@aYaK~` -vQ23J_#gfDCk9QP54!@u7c+5CFd9tH9ad@!zROFa329bB4dT@|LMi$HhoK~^1?r4c9HAQ($-8St*ggmLSv!RK>FexJP~v{Rf9xi2|EdE_wU9K%u1t8p#eAYss?&RiWj0Zt!}X%?Y>|X?Q1h&1 -nJIh0F4Wwv^oW4+N_-uro#lctMKHqjqJ?myWyESi_KI6H7puBgyG(+J2>*X)}lB{8kOP=OjjMMwgyOn -^}8V+2MEX$n@S%KN!-dmnp0JctB5%CFZm`AhK}V{1_V;K^+529Vk}5TGX{YLMBeR`K5MBxikV!-xP4Em3P92rBtV{Pgk+iOIZ>+paMhxfTcSYTK -4fHkpgNv)7R|9&`KOwcvO8VGGy@Lm&-G-i2Rq!+fb*2~ITqjBLh -0=Yic;i*e>gm-$tSNn!-TVGYb(_~G&-({Sf4Q6@CqYlfuvBO;-dveCA5h_Cpo|g&PAT0&oyT1YH<2nL -Vwk6UqC1~@ykx@%%wPZd*W@`XhUOcFCZ|jT1O``aGTEx5L(_Y_;1;Z_D)W5SkX;V&8PjV%*uE*s^*cI -%NT1;!2xR64dlxVF#ih1Hwbr*gj44(T6sUm2mOG}&p+=D<#Fr^hLf9yXpaY}8=_~AQDSMRZfZnQJ8Cf -$Z1AAkQR2g>ar)S8q%`+goy`+amWIex@pi0lEk!I;fv%Qa+9%_>3Ucp57Z!&$%71N -2fj-LSr>M8hiw>oQRz3Ab;+@Pwy6VJWVt{%(u;oXA!HYjssu>&vS|yeLEm5eP@UIvx72whAz`YmuoQ{ -4{$!{3)eBdxP}R5LSkJytOtL@oEP^c4;pK0>2%Nu2D6n?Dd#DejumK%rF#qbU2ZTe(LA~n(Lt}kYyAM -&o`Smy(}zRQn3l&#DH_Jzt|D67xQ_mdY$8<=Y_BTEs!=~<46LeG23@ibOYpUo;pJFl8A#Z3MHL-NG4d -*ye^A(o8;9Sw0D)!nf-3Ygq4iEDH4)BNTC9 -yfllk^X1@(lKPP%pb!kuMe=Qlw%=LJ)vKCoJM<^IMr}c9qJUv`A0a3nnjzNo3cyrc75BwBh<~E@9@VJ -{xw=cPFB~I&!I0*pQNAkqQy}ipkRLX -Aef)FVGi4qi66a_^mEvCmvGLLEP~e?eWUWrY`T#(jUAQZY2kom?258+%*rDBTjfk*WB_zHa6fomPEmzxwmW)hLcO=(~v75UtWI=MtUfed`{0p%(_pHB2*3ubVz -}nYS<@##)9XvO3-$5r`=HGs;#-t0sc5heA{}qkMtgOUE5=@oC*9LclJo-KI$;+K5144SbFsmy@CyG8#$ZURl^(XT5Hx{PZKeX2Z(h -bb-VuBRyMxF;x~KGQKoqBbD(aI%~NR|ns)wKFpwJi%IH(#>ee1vy{hdWq-d5jhCtGd9IlgCSNW@Sbqf -D1*TV!kC42O9tDVQ^qKvttVE$)&Q8Xx$RjI;?D$YOcHH}Rb3<#g2}Qpx`0iH#i3y;2R4I4((fpC57X8 -d2srn37~QMZB8!it?lHe)!%;F74NQ66H#hjRv#uTOTHHFpK^qFPzsK(hPS|G|LF9IwdkX{bctv=1EZd -8w{8A8#+;36*UJ^ik@ptBL_s6(c_mxYJB1tlU$BcY54#UeowD;7-0HIhi-LhAS`9%N~85>t`(Fyxo2p -v>B1@Mb1=hIYp&3@xkUWrqnrKAG?3(wL4hp@F}`0F+Nwgpg`S0XR&=z6pfHmQm>WxTQj?IeKu5sk67H -A`~6>VmG{=ku(*rf=d%noyJO-2W!eW^`LeLkFNHbMm3MDpqV1gM11OV!4FYNHCKF2Sq^GK#9Q_XcB&5 -cQV0_&_*a~=#A^j^a}l7@eY9OTD2%v7#+ZyAv-%{MtMIJIBp`V3kzV4!@fw80l9=tCPrk*;D?jRFy6c -veMUQg@7hcAjm*6500)Q=)N>&)R8LGJt!_BuTIm4)7%P7gYq}J5q3{)|0Q=}&K1EkOH&G1I^bgJX=S4 -nqa0x8B&D9%L{mC{Cbr6Hr#vUcX&PIK><$*{SBNDn4#bx*9*WD@j-iE4agY@7Vsj4SZNjE}L?{qi3_Q1l7=O$K_79 -@nwqA!3VopERSvaF9$Lz-SL91mRZ>505Sb- -0rS#}IY&~>e=Dw0K6`ir>1gBmI*kmBG@I%wTp*BJzqY&EzJH6=san>%h8DHm{#yJAom*?~*^1>Qb&&I -22Jk8)l$Iwn)aHgV=Y`6nja$&|JLRnupVO}2ERd98Y8Zk=^2w{fm)n;>{nOtoK^sljPBDCN@MQ#CE^1 -jH>EMqX8HywtBCp89(;**mL>otp81>WKyrgNX3kv -d$J{-9uMq!X|Xr@`82f+D6UaXbBaY*J~-21OS1DFpu4%x^0Yc)Ig7=tx)of`=U~+DTYkpRl>>g{2&78 -p(x9YX4}fQVM2o>njNEfs{*g` -0?qu1R-{Yz_i}A0A&gHGCwXmov6MnV1Gpb&%WPnX3ZD%na?}!AzaF!xD5VxQBn|`(kYM -^AqHlV^d{B^f6DV`a93{=vls7J`S^J%qE;zL0)5X~Rpzv@tWaBq9_K&biY8eEnxYhmFlID{VnsgMcs0 -@UrOjfQe<33pte(P;A*!@kertM6Hf+}2V7ql+%QP99#y}**C97hjXvAHhj{wxAt2+fbRJ+g?UE}oKSY -*U(HxUfF!To?Dp9=Rq{aWO?xWya2S9m62)C}w6PV@kK$v*BQT#ftoQF=yuq;L!YF&-Rz6FXy#3MRTIB -iIQ61He>=B-*nyI%j3|&^a?QOvivWGIhEQjZLDV2(bxZ6EbPW1XF8Er6b$=J9%(N^KNTi6wKnPWv@q2B9FKy)`&+hjR207 -L8paY;Pv1iF4`Dlu=Ed&tNgxRU6cwF~;KGQ5`QbS=*Kp}7YZW=fF)HH+|IC&nO7WN@*1_K5oT0xU>Vz -BHD+}hlLK&3|iWZo&Q-2v_EmlQmkb#&?8FzHf^_qw+5^>Xh+(oy<$isS#z4X**Ini$jAt)c -oN=t*m(HH;{D#km4^}W4J|xYOV`{9HB(j$BJ(%iY?ZnG{!!v;n_vYsfmI#5wLm&OLL*cxGfjxKs9(eO -fSX~&Bm3ZVL|cJzVV~%pD8F_^ca4mOJY?$71~e5SROTKHJs|1ab1V)(s<`O?Xa49P+AFRN{W-JFSgHXF{v7Q|!oDyGiIGR}>5@VpsWf`$8p*r$pekgf?rMYF -em>;(TNANxr?sU(ek*cSFmG(-G6PaTwA25(>z2f;K`Ct+BY5nbz3N%qOVkB8HQ`F6XERn#ckk -;<8RWMvz*hE{*7x>!ggvjZ3?9!&Pg2j^yn;NAYl?L_qi<3*_^hfOIi~WN1aW=4}8v!1%?u5`RFMYkNy -FqmaMDmJD>$EdfRDjoL6+V@rs*^ze~D1TJ5JOvH2rO^M_T(!=(NB{rq{fgP^oU*P8|LFPhxn(+H`5CwmVJO4mCZBJHET@|2+18c--g9LR&`4ek;^-!7QIN@F2?9&IZzgK -?FRY=>NX0!Q-)XgY6V7LXhOU^>#oeYUoPh-pFq`cWyD-KkM^Cv?7=yfq+}lH*Oka~=6c%eHBX0~XeyS -*Ms%yiCk+ql>K$7Xtl(Mj3Vc^M7g{8#q?O;E;bAQ4S%dFXmE|ZDI1f2ZZEuIJf6^r|w&u40+Wu9}yf+ -;lM3uB_#aQJFHUDpF0D~Vhwq0bE^U|%?B|+^^zgSCxHhQw3WUz@TQj70mgcPZYlIaY{*HWWZ3h@S1UXI+h}lD9gd+^ -jV_17Q+PAH$TkUvMdNVqlf?%0n1Aw#4f^503;N!J#nm~roJ5T<7YHoeiSXJ-=9|lPozLBNF)`$I7jSc -O-l35V4J#J&xan|DHh53qRi|-BsXKeG)N2>gHMo-4ALM3*A?4}fv;xjI4)Bl`VxXT5PY(|Eo}LAuD)N -3o8$2=sir5Ia*P&Eomt+ZKfvq_7^PFVM$|wYwYlA|8MF^^z1xpUmUhQX!V1hD0)re#zV%s1^Oe!+~pX -m9Khlr|cZ{?lz=Y8cr((l{)m&xe4v1h|kay)n`DmxsT!D?MNoo1!6;~) -j1Ns)zu!E%?4McpOZs>%Pgzq;7LnXiUrvWgx4%4bUNmn*>p37p|Y!_$t$P3b59wR3xFjnnM!V%G5Y2r -29AUfL>$B7AhmX4EKCxqE!!S(1RgPN6k|j1K7KQDJdYP(Y^3qh101MeV71oGm32z=_E}{zzWI9*KTvi -Y>`A)08CfhTd}L;VttlgOi_L+N2O-E-M{SguXlI21IwvaA5+nw|5skw3q6#N7pQl0TGm;Wod^eK -q0(5oV9sqr-hh41=KYB3pKD#h5}=mks)Xlw}ZT2qmm(Yqlb!;ns+{T5m{%b`u}&z2-!J5YnM%$D{>%a -nW}Ucbaa9oN2^bV&&>iMyVE=v&N$f(eb7#W?MiZS9!6n$^9}C?@~?S79e`fA21PG4j6uDb)3~pw5&qh -Q$ZNq(a=ze&&KSX9c_a4)@>8)pEk!HdDem#^{u`eIc$hGcktrYHR!K{1GTD>W?S!O8!wP*TF`!VMH95 -Buf5T`HSVvfWy7%dcm&a(R|Fb`ET*Balx0j3r(CTgE&SH?>N(c=RsA$VQy(<=~)w39>Lo5}O&x(5P9Vs+*!6AYip#Xh6FV35&u`_r$P+GaJB+e^Pmdfck>Me1o`qj`#^xZFvzT>_!$)(%^6rmKW -pQxlVI&xv%sUZ-r8U|SSZJ-5+q;}WQAO5R6+y6aU1j2P-YVJ5@`0oCV@xr15stWQ~$PC!^ji@QYRYAA -NZY^Mxvk!Ku46fL;xjdJ2rjHqWwZRE%eN{P|yIFL^_TJY$LKhfbD7;KETWnWk$0Tk(x1oC`i<5hFxMh -WAL1kCD5)k$(O&n|7gNy^|z{P=AyE9qpwD~#~F;y4%tHmT+_mzYVyvkDrqu`Q);YF1wUDq4RlCKph%R -PO;ywa)8c1F$9^?-2n!NCp#c&L(I`g4>9^^C!a*C!utc;FoR-azl$Q6l#Z5%>a|sR`l{MF~?v>emd6k{#D-s3*wsyfz8wo -n?4WFj|VEe)8i-Wy6-(FK#1rsycK!BZNC|#Ew&HZonSoZOuVt(BCTK9a0YukRTYdKnbmCdSKJQ1xB(` -3l-?#U=uWIbiIuAOIOsQ`Ga%mB|@_l}QSGS5_Na=~$`th|i_qI&BXxFIUbT{oIj4I^Sl$k#(Q(;>fSE -3_L7&vshHTW*354()Ar`gB86US^>nZPqdb0qmU@m66cpazb&Z_6;`1&j{P+j -GWXDAEkRo4#Hc$f%qHAf_6jEY@4kLJ&kZ7~suxBAEKJ;O%XCn$P7=6PjFa2XfXl@7R8>`YHrN)gZQg) -;eQG47=&~1`MKk{uwT!wSs==I2W^=7N!Xw3YXT*jSwow_JECk)wVU|YQZ+K4J#nng!vZ0pMw -ECYU^5`_n147QdzLp341(AGmXSfr1}$7IZqOs0-QM}!N_Ujl<|0nSz)HlTtEd_1M2xMaT_DNSxRXc7) -9HlKIe3EOLiv5B?D$uY-vZ;D(|BlI;7T2{Drezw63J#j+UG={C0l{8(>c+@#@s#!;<_Z?(s8#-}uJiE -!Djf27$ogOGDjrUUQnK>78q>QMoI^S<~%j7~MR(#P`lp1!2`)rcMt+7%f)aT|2fUSx&xN|Vf-WPzp(&{V|OI8mg2jj?v)rL@8iDg(K})Fl2Kyzg`7&;zdd!RHX -kgOEmQXpdFH4TeL^pC${3I%kp@D;Be*Y!Z7F=|jT#tOu&cE@+{;LRutxlVjw3k|H!^wrsAdh!ql^BhX -VUNrr#%*g^hbseWNS%3!)RppX{bb7Eaxk|i9zJe?(jY~)U(kwI7B1bcY6Y-F@R)NC!27D~tz=i*`A^r_84urQoE -r(^TT}A+sqxojRce$XMki<5N432~rUdG|%;lX%d0G21u8?vZVch(bv0h=$qOkwK!pQL~2PbNcWakr%} -eu7T?{8+8Yq*EohM;VHm^8wJOee}YW&JQ_M2pjn#QQK5&3l0d?6{pr(CXe70c@(NtkPQ3{rx#!~tr_+ -NeI~7i&2|8tkipw=s6FBzXLOyzUbG`+say~(^xy*A$f4|0|0Rm^25J*1?gQQ+OV=HdBa8&+*#(IoAg6 -!i=zx{P?N>YT3tuPWK_IlHOQnbVbigdViBrT`nX0)^6}K38Cv2KhABqZt!qsw|af^0$#d%n6#=P9Evq -{goh!(x~!d*iVlUMt`#SRZO@b1&e^$jO0nhAtK%g>RIN?u07MRk+k69-;XVYBpB5u_%)6S1AQUMfuno -WCF{1v0(o(H0y!D7g0NeM>WhYE(qkqKz(hA3^_QCk{FVw-l3T!~+Y=!IrcL#cr>Ce{12~7pO(U9JeO# -amJQ^HX>8kl~xTkE@RgPx+3;DpLw_HEym8=RwL1guztigFeaR^H%zQ!uUuM;9R#5C9kcwcAigE6n%i4 -ddnm4LR=z>07vf-|TT(jy@6eS5vm~HTOI|hwMX>Qp^Cqv026Z36b*^cRGZs5Jnq~FSA=(?&m@7ei2jB -}Jf(Wv7tdTjUAHzOLoaTa}2&j`d42feJldGWad`iFuEh^>_2pf_~17b&GyLR_Kt(5CN;OQ+YtIcso!O -KicBD&h(A#dbt=NQ_7u)c-sbn9dogj3Gb8N<5Dei{U*i0`dKwjFEeSZ#-@s|Kcg1Syh)1|P?50sSH*P -KJgKF+DB@=}0lFAvHxW2g*S7zs0AuGh@g<(=dOf -A8KcaGSAxUe|Bj9Fc;GM$7l1HQ8Ik@G#nhgM;HcVc-5dMnynQhmqv_YD0Up$uUcaf8xBxIVBaJXv$(| -a&AMn?hv|5R9NjE6aSZV)L?*V5{fLL{l?SI4sBQ4J92C?~=v4uQ+ZB_CX -LO$zNceqD);P4yQ^U>X>}a8f-(_gul`N8Sx@w0ep+n~O`$4c9tbeC$)*@UA -mbnTttr*AcJBf$iuNj*RkvjCvoGrrWl`em)^##U>28{8(6fWKKJ@T=;EDDdcF$?+ym$j*j??9@@1I2f -jq}&pqdSkv|E@W+6~fa|z3}$c!%oqA9gVNp^}8#0sn+UGM5=dFWOgGdHLwq<5>kzT@ -{qhP(hD -OfL3oUuQ@54(EZ)7uER^jYDb{gi5Ra4P2J8xE*Cdl7e0#1POHE8MHJ+@#T$7k061FiuF%sEIOk~lJhc -QpxqdSBDHSvr@*r|0{{JhQs`$AzuI?R@Jm&0Io_CiI5Kf~6+K*vV=$iNTBro%{0dPo>++sFlIDO!pC6 -at182YV6#Yc?xZ!d?Z-MnNv4!ZJS_A2r>Xx(+{9yg^0P`969w(+UvPQEE#j`5~ -nU1KXFCwR=m!LG>g<XZlSB)seiZbe3x_jngk3V$l0L2icXD?H3N1`Pv7jW^-Z;1Mru -Q{*H`uZg^hL?fcjSHJsh;H$;Pb=5hN6%}NtwtNVp>5=jL2t59ZZPPq0|glXV@tEusZ^-HG4Pp0;p;0u -Mu3)A972#IrX{I#^h45EvAsKbTMGa+184x%<%WM_5)1sCQ)Jo-SOvvLn9-~1ay^^mlA|8=cYf*<4h8Kli-dOeEF3|(oRGergNknOQlkUl|>dlVAm1DQoM0wtHGm37`_?v%?-V3so| -1?P&)X^!|@20jAeXw6Uqq!IFdkDJb3G;`E=lfc%k|qc5W+yt+D}mv#L*^0*bX@7|~PioE*H+vy@|1<3 -);Bk5urw5taR-@iaOZhmj1>lr!>$68Mvu7fy-|5iXw697G*4&s_$Ua+&xpG_3P`;fzF9=GI|&32E0&c -sZR{ocIiO(ST)Bs09)~;LxDZmH&IOM4-PLYOy5xoltl?Ao9HwK=u*}BPP)#WopPZU8uS5nyyY{XWK$E -L!E$GM8NA3PJtfx=kl`O=*xszrVhttX^UL&Z76n*P9~?2Ic1c)jy8RfNS4I(wi;zM4*|cBq#+Ia-e9E -1(gP&bN0LPee_S(veVLai_%eBs`Jky}H)BGMa=lF8=AYMF%%G>MUviWf5UZ@F6PF>FM67=U0yEIi{XL -dEPWpR5A^M}_5%qjaBFgdk294Oqobz#P_r*Z9+)^?fxx2cC6%LS(bagJW#@VzW -)1ruFmWH&kIZ?iWsY_}OEx%0W9~KFu^F%Nd|6@aK@^2lYWyZ(zv&eQQz(pWE2)gLHB3T{ei_i+4T|Q| -Wn!fOi#!5?kgqC6*KqJ=xlGin)q?Wo!@G?m{K9MliT0U+z@KYT&2rxNh`Nl8DOz{Hb3JcV?tftnj8*z(eEh?!w| -`D(&i%(=iR~E!FLDEP`i? -;P_cd^h;r`5G$Fn0v_sB{UzRFPv}T6lNY}6=^3!tVw}B+_#>$HBnt1uG#Hx76b(lfNxsdd$rZ3gIZ9P -c3S7%t%&Ic~lWY@f)QnI2ZqsqTm~ykavThaGoX>9oG^;%CrWVT}#indW>DZ1}FTeRQ$u2NTs -pQU5mF-yk48&$lL@w9&gyh{pG8-+yav6(w0%`KNHu)@QGq><78Lby~Iwe$#-SEB?kDWsi56@{SZjolj>jL<3WIm(<>-wYmNx0wn%U9cq?WR33p(DF1aj;gXjAJVY*#$JH76_ -T~)+gv#+#T=3{bH0R*O_!y%~Aa*F~Ox1;MLtR+q$9qS|CFc^Ciy4iUC@MM%8;V~y@J2m(LFB%$YaEvx -;{bUl-yU8R<@kUHc@n^MG;lo=5#l*@$vv_0`ScAi3N9*QlO8RW3(uQxa@C^SjY;S%(j`jSbLS7%&HsS -+L84q*4)MjiYQ)z4&V`h|QhfF5Nrj-++Ff@6_XNed)gpB8$+x%wJy36JW!~C<68R7sD5yF7UM6MI|=x -fNFSuu69e4ejz=cj6!x`q!`q^nGwY{t~fbyjC^-^qO!(XPg9xFR{S*_}f?;=L9XIuBk}D>z{`G8Kf=P -P!5@#hf#YC)X+u?ba7@1*TtCTZ;|@vfj46u%;c@ECo!bX}oKM9a)#14Y<2s#B|x;<970t*)q~#lp|Rd -CmfJ&_re=Z$aWkmpGk7$YK?-)6ij27=FYh`w2U_vB9j2`Ed=cD~sE(fU2|b> -0do$Y5pL;E4aWj@FXnAc{m}f)HM{F|EUnk^cqf3?e4rr0<3=LcIr_F;tP04w_@d#$dq}zX -D>^O2`IeE)1U2+*R7iJ3Fwmo|UUM`f%(F_-+)6k>;YPckB7LG=Ui0B7a)FtN11%`r-|I(bqs~(}K?nW -Q|9rOhAKBjDpX~iHJ$u|AVFi}pdV@56GV(VVD*e6xc#R=Dzx053BCK5(V1kgT?N`crzg0a@sRE7L?)Q -F|u1NXg5?HEYCe>Uote2;C2$rcH^b?$~wBNF7+CI6aLo&Z(z-Kr#Rab16?F4^i0<;i?n@seYOi -V~nS}E9E)%uc|zRy4ZJmJ7W25_XU{*(0MPuAz_arYhZr*ar*d+$Yj9tU@9Is30L`iPAO^Uv%CA8Qt4Dmf9`FQ`{E?~X&)xzlcSSQKYcR(>e)Bn -eEID8Us874w1`H9QPhOxm1z$m)FD4T?cuN9nZkAQ7!VMra77(piD4q)Xndi|HHNq~ZJ7<+C -S-j4~}oBWc35Mr(2hhEC6v|id9YO;3Gdyl0*@LawpcprLZr+I+n@BIjNQ8Ed`QAkCz7V?u~>TlcjroE -0zd#YAPC7-&~{h!iSomADaQ?%FfXgP5KER6rfN0!8Nx#wwh-?M7mXEj$Qc}>Johdi9fxZ3)+&hCAv?e -WBn`x3KfX#HAwN>-IK0orzoTXWmW1$QL5$f_bmoB^_zH`8p%4A;U<4XkE33r=4xMXAg*0eGG`O5=e&@Xt-2?^F-Sp1o -H=A;my|Yn16#>apD=!Y*l=5_5lyK~Y&!zLZo3a)SpKwBJWj+Iqfi{iFpiHM0 -buaIy)@j0h)71j;m#yr{r+(1+l#dAEM)*4_R)73x_L1OE+j|0>TgiJt23Ut -=CA8YQKh7OIes-T;mcaKf{Q4?&Z&wUPClfYR5fNT>CGQrgYP|)s!LyBU(_h}c5L>j7VA^ -NJ21p-Mtif>Loq-VJQq;J5qtcIkF5tY>JlsGjCrb!L?Q$6)91COD~JH4>$m#+3;L>~Cq1N}tbtIa2fj -T?%N#{(E`ULwRx{vZ)~k!X&-yJZl<&#{-@l+zjvsh84V6#F9EO)&;z`ygB0$fK6cq+Y_=}58P=8fd>+ -^ZuTvZik_oGy9z=u)RP?P~%$8hV7vDz$iU$qbKfHlh+c=`A+>0>9}e-QEMN;FUJKlF{Nm^D}abkPivU -v_k*$H?R~Q5@}X;gjhmmpN5C_3$U$W>LR}`|oa7>!JJGj#D?hZ7)i@_|V_vVvp0+oq!fZ2k+Zw-+%bJ --Y=d2zwt^0*3tQ7J^r>S3&s@s}qFzeFOBz=&zhvkv%c`qGOxBvs#& -A%4Hm1abM$O!+?^U0&u>sn7kYPMQ8-=|ap(rDt_J;vE&!+9eEj?s@6|k;pcHEy~8|BoD90#=ZUskJQbHB@H -k?k+a-K7{t(e0W;YZ;q_b!n7y`AlwUbl@vPe$_AE{qLWtXQ7Xg^eu9N{10{T!bFZla}9k>zT?CNIzqN -O(b$EdXL_6VPq6X?rS$-tRIp_-fzFT`g(J?U$x%TzY+~e_duC=BF#jr_UJJ`ujzXdCEXtj9MZesg -y{;;Vx@s~V_35wVlqGaJXDQO*!m%WVGre*0+0@>|WPJ373@a@t84wLutGXvoQ@o{2P>?5nG((nMdaR? -FsKf4_mwEmyFIU}&3VmKN21UhbnUezV$NUSB$siS%lyXHkqPiu1Ob$90(bO+G7cHQ1+(A*uPrZ)V0Q8wL$-)s`k&bI$gb61yG5l5Ik`YAqXZ)-T$NRvBQR21=Mijqh&5}9XSVh$6^2tUp`>40*OR~PV_%q5KDl)3t5{+>&I+9^z^OzC -W-x6cp$Pw4NL^Q^pP>4Aiu;3Dl2^?$*-1L5BjRhMc4dJ>s2UuOCX=jfSo>(fZOIE$pTnxH`zFbrwWzaqTImw@ub1&U8YJZ+8TuCA859)kPEe~TVA!XD0Ec -QyLTE9D_w{B^;0gnxm>fSRJkd2kjNr)NCzfEwX%S$u3fj3e+DD%XbyJyKwQLon7TzO~2^S0bey|!MW0 -YPdjaq*DZCpLBrdjlBGf!+w-iy)c~U@(RW&rOI_5k#T+^*M%FZ8(`6x{&bx&l{#{VSYk}VAUa|@w5Z4IdL?jC_2&Py>eRYQ-))2D$2;R|!_WtsASmpY>|FBbujB=XPT;5KvP$e`oTmRn)X~GM(A>@PUyitP_vd+f#A?MeBLnHbMu!W5W`lHx;bfSos@x#`cT*A6Ze -@yabN(sM=`kw`x^Wr>RWEf;LM-Ld5_|aiRn?BHP#SSe8e3VQ%L07P^Qa4HJ6j&@*oBN&v?1v=bs)0=- -&qI@05`)_SS#_2jK!UAX;*kST0v}d#Ugs0*j98Rcd0lW!8WjENuHa^87YASrF(^+#J_2Omp^%h -XRlg%9BW+FuSlUZ62omiK?{1WqV{!yMffwhqEYSp1-FWVa`bJFD?>}(+Yn@S7;5c-ahV{e%sv8@22*a -UxOY4^WKZ@nISy^1*VsUmrQZ4nq%@-9~it6_m9)-T4S79&A6ZrRHnV15}`MQ{!+~^7iCOEK(__)B)JC -P_Rur9~1PHCRbCgSo1*mYA>B%mTX{>_^cj$0D#8e;sUniQ0QgZ4bYt3R&%(L1E* -kGihfw1l%wl*msFbOeh+!aqiNe$7(eTCkN(D`FLG?^-E|1i9zWiiKMPj-@!yC+;jCD)X2i5UFVBShYv -hpFyzvqUE!b6tyHkg22uN=c!!Nb`h$hn?_ud>n+B#mqfXMN3?#=Nn$`_eY5gsI5yO?4pifz8Z`^#m7# -oU|T(?X)%MqUzC5^FB}q836wrvQBs{ojvK@>Ozw#&Vz*EX9|!tJm{3#7o9VT?rU~+9GI+h|1~1VshZ1 -%w>@_8{ZdffH;+eZwEih>5L2>~@Svhio18ebxNV~cr9<<1(5$vDAi`M33cd>A9>UqHr9;=jJQ|O4;d9 -Zhv7ytT=c?Bo -!a6c2hNoAkiq;IaT}5R=&CzXRulg1JhqJyAQ>@fF98zp0~^17HTj{6o(CkLU>=~>gj{8~$w-balUhMMWa1O -20hqHr_Zk+ZwZ#q%Od~4l`w2$z -(`$Eb$nM=X%Q5X~Y>5-RB6f6X+tG8>1eHUlvg^S8d4w)Ua9-TzoMaD!Khee4LNEwt?V)V{Ru#^Em(HH -g7te;49_e5OdIo`=WEm8f=M8AL9+E_JWO+l?%wp_4ei)(&x@HZwtuea1r{2k+;s$_s0kYlQYDyro?0g -Jo^0!gU4DUuUtW -ajRxJIe}GIB1a{~~DJYYPzoc@INGwhgw-jlK4pl!CVnxFb;&sG5_YDn_T}s{X4rQA8fqC>f|tWWNT$% -u4Cdcc#7@i9dHO#O3Lx(ce{9WL<*3Z7|A4xo9Y$wu2PUne)pr#C&#LsGakCiX_#;RgVrUpNhvrht7uxd`pl2MGjj=h)=NKZ|MFau>^(I=2k{JcTKD$THgiKBImM(WK@h={v^0%R$Z@~^ -cY3QbzQ9PQxi@M0m8PGC6bqszCQg4fv8z&EX0;UU#s5UN)hA|j`-<3nZUf@$1UQdE+r*$vli%rz%+Fp82&;0Hx9&ykwNhUC!7 -v~LG8KP@|-!>Z`MyYb!0Vv`z;`b+|py;!<)0;TMMc7F_(nA3gy$#7LwP}g@4d$-P1CqB9oY+(=BtL;+ -pwEw#GfuA*PP50c%T8=h4u^I>)RuywfWv%@Jx?7=UoNmXf(b(~mK0E)?8@R)nS{|)ub|Ke-RAU6Y*XiO~uQ?eWk&)NuLR}PB8-`EP{V}~5henVuWZNrO}yulN&O10QFk1c -${;_vu1uEQXtvuB~)snx4(E%>_PmV{Vflak8=kviY>Sd*h<;5lykQjCXsZMJ0DzV{nS>$W+jhmxV+u= -(wtqj4+1s0LxA)f7&pj;hjc4MFwv?oPR4-X~XZ++#4|3qDZGSvE(-$;n^XYKpCwcwW9kr|Sci3&t9<9 -r|td_Hk_ct7_kXF>y5Mx@fHe>YQ}XWP8D$nL1*JG2t*26nkjkF{jKaH&CV`IX}pmob#!TocXKd;s{eH -i61vXlGN*zRYT>$kjd)Q)Uc{jYzRU#Hrez^fM_#D4j@1{Lo`Iql$`%@9OeyrE{(t`O|I1T5waO^BW`A&!&#L#{PSXcJvMVbn(0ZPNF_4G!cW -=Mi`^=ZVv*mtA?jo;bDS*;+* -`z9B$QGh#LV-un9u{!3jV7*RHuegoKL*+`MtQxF8X%*jT?Ye_hIX~YEuu|t%h1-=AD%ur^I$-AovxPJ -q$VYn8ntm~TXV>M==P49#qUbAoXuz9vRQFF2M$7lfU$jx`Lph& -jF*F~(X^p2;zy>{s*xy{JzyatGdt#z6h&a6=+}nq*7uV?ByP7y}D0;G$|AP%N*?Z|zKs*)9wJJsrP28 -zm7?2d!Q~Np1ck!L%olx)U}pDg7|ZWO9#LBKeAhc5{q06R%%3TL=Ci$8Xvh6mi=GhU*<0#7hPqM1m(w -^H5TGL_5glX29BVaQ>@oH6r6!WK-RqMUXZ!rnMw8$*D(t=rze?qG=8W2e#R`S-&D797&pG`zi?=CKKP -i8hbdF0QMM=oFi}N?=nNq`Swcp|RK777jq6K?e(f -iWFQ;HFO>|!5Q@DLCYixXo_agz;~-5>roqlAYz*#SCjA7T`OX3;`}$Zc3=L=%yojtp -#1h9p#~gUE!d<u>49cU^#i1pTZis^26T{9`5?(P-$rTw4OndaheV4K`9VUY=a -Bx7(XY3k6Ux_ipFy-2PkY`|MOsE>n^K9=LNa^kV#p=GR3*FP9AC#c2p`&FW{Zey^*2s74a{{7#=Q6f1 -V#a1HKU)>d5g#E~riy9XAntwODdtyv4RkH*X~Wr=sG?=ph#2C9kDi)n=!Aew^M%^b8+~*;8fi*Xn_a~ -nfgJ`CL|j%oA@`VW+n$hA-A@dCtvcdS4DU8>_wSfCixt9|GRU`Z3*Hqh!kl`NOjK*5nJSLZJ2V4yAy| -0q&Tn!mgk}$&kfL*sHTnU>m(`2Hc_TQfmOS^lI~xnd*5Yc)F%XN8I>r_m^i8rxBi|4$Z*aXwz-=eLc0 -61zjn^G346qxZRjz^DnhMHcsA -461bto_jCeH+1?2H6TZLy4+8`r+{1@V;lM^d7AHbeN_M+2Tld~Rh|KG}{9?Q1$|Gwaa=9Ev5{BLg96p -Cd?l#2$hp=3V5N1~2D$DPbZI^OBAjbVJ6s%R&3h -90mRH1SAwIL_s0-KFqy}2eiDt1VD3OKaUqnBXR578-jl;kiw}dI`Xaq!}d)g63-yToybDRR`s!sQaTn -9Um!+(`R333`~gzbHcTT!Rj(v^NaI{+?Onz?H9n*%1Fu6|+v<~l%WH{mTjdYJxqxCjG2v~iCW&rk6b` -NlpYq=$u#I%e7uTXiJ4vau#*fHhYvKZtov`yP(0`dpE@R(&6F=6+Ayf5mYS4Ab-JASrd1nPMR7oKpE>jNv211BM`Fmo+(3p}e@juy&<@6nZ5|8tp8OjF8&q^TtN?c6UB<(ghwiaD_C!vEyZoC -ytTcNuEsSJ8K}Nz{GmOX@xH`0chpU>_}V6JUK2Vw#bd$tIEX -;cm)EtFhN+r@$Y?zF&sN5v8~q~+FmZb1qGuMNxG=@HE*Son`9CBP75`{b&eRcf2SwXlxCJx@A3lq!nItmh^Orz?2lMB>Hb$~7DY;{8_**AN(jwG3MR>qMF4Xguo*U~kkxUEGYeA{% -!dN>9C}Z&<5ez~E3b^an46ubklnR*>v}7Ea$%q{3w5w+riku)!s~RZ&vMN>l>7sVBTPo|d39pQr{WK! -HW$t*X70D{-5vQ6I2x8UB5}~-~#t&6}oz?6B2Vi%9|M3%|sk^^_`bj)WNT+eRG@9y0-Bpi?s0z>sM0c!x3TTZtvAnh*skD_;ojFSZKI4E0s{J~QL3yMg6`*eKE6d@p(?S}mX+MQ8>} -5U_*o9==yi{;vKooTNXM4kLRux&dX1U2vZF$b}^tV9-DglEr3}#6EreWSA&uWcG}g?V7_E+||2Ds%xV -a$Gzg{+xvX8uc`yi-NIg#!20rQ&$hK^hSF2l|CTxDC>9bJO~$9!b2NlJ{CH?$)-9H4HU(}T283Z61`! -UCTt@8llN2a&8IB3)JeqNke7mMtOmEOw9LCMje-7K!hItUU!gDNI)p%Ot_N5I;)FTM8gw_0x#E`wV`F -vZIjmY)R0n+3NPd^?F%|fUvfP14+D`gCI4SG9T&iO -ZHqBXdI)6j -|KbNK??V_PyIqM-y_%i32p;`4d_uY8JKD@Ge%^HEUGmvpl{TRSOKQJL;ZXbN+HW7e=46+D%@Slq=ja) -kcI?AVtvPGR$4n%7g8$Xb%>vw?&y95_S)jw2-|Gq0`PNfUyO{-6H$2Kdh%(&*u-vcqM0#BE9!A5(tg~&D#>6S;Hjd`sF^nb3f<*sjP>n8T#u~_Zo!Qtwm}=ynb)9D2g$JhkceNN*%D*%fWMA*27YjL ->*Z{C~fk{-C>l&^;>_VVsONYe7(9BkOg>e*E&=3aA98Wiww*~2Krml;_p|vT97if=x^yYPTxyTM+;8U -DFPh9Z4Dmh~~@a!kA-W>lm`S{6`&(K^ac}^{iP!K!4oh5@$hDM^@h@@eGYC@U-z=v|xq{(-flAYd~A! -X;}{5lvWz?5I(!B66X701W;&`&T -wp#NIrITE)a1x>zz~76I!Bi<3`1>F*L19w>A0PII_i+^mJmhcZph2fH%2#-Jd&u@+aFJk#?SwJ^!zW! -FTXwzNwm4;To$XV^*P+>7yHZ2>Z&UD7n|i~pHiPR`%nMy@gJVvS%c0fpw6b&A3pi?U#XJYCHRZbrA5* -2*k}(2)4AvF*$0ib(iNaS3JxDE+_Q^yIpsUzyk0Z7!6JKykuwfm;)d<5w(WL0D1+T|+E6N@On-qH^%k -V`zYrTXI_!`U&uW8z9Bnd+de{KE{TLHUO577QuGC%fYWT*dzO1!nPfP|{$~G(FG3!fZv49Cv%q<6nbi -dl2{)N+J0CmpW*3rdAWe2o1p~gsg`Ub=6miiyaIqM1V0`83&H}+IiK^u&qD+pg5x}oe#>mN0>)ubdj|V@c@bB=^zx0MF2aSQL8NYn?ZBUq(!mP>Qi<5uDMn6CO -&%Lw9LxqHU_zFl$`whKaZty+1ImSy^7)XIXPag*X#=qDJzRDly*hsJ?{r>3nm#^L&0bzk3$6vpE_4?@ -fvo}Y;-^d898zBmyPhhfF7HMCxrn%0=_yy%=fSuV4W6KsO6=rsf6zt6Z#C-|X!p$$-b+`0yBKQ$8ZsZ -LvIM7A%j)C@!SZJV(n<aeSTe$iVvYjsjr0fu?{wK8be -?)IR28dT8n=tl455+_gSijrSC}D(Cdd}ypFtV%h{Zp8C1JkWR-*_CgX$id5W038qgK!T08qBS!G%0|4 -7>gSvy9ul@5;78$cjf`B&Q!WXMN~Xf284rLpGN#JygW^P7B$d+b5XP4|v;;>LHH3k!)xtx@wD+pyl}s!>g(dD%T7RML%_M3BNQ -Bz)>g4tbtI%@!X=YH2^yj2_*cf@?HXS@&^m>BCnkY%!?Bt%sRBTqSv5@xS1Xi%rlwyM0<{e&H64<~Bl -&z2W-7$b$@Qg)j@xX@TxeHbsV%x9YErxnMhUOPZt_Deu{6-UJDI>fo3LG(-ANO%(VlNYaV?srU+4J}P -sa%%mDDxJ!VbA;a05SQjh{fpCtXOQq>BRrHL@tz%5vDEVO*IeUk7(GkBB%u5E*$TDQGWQvn{bOD+W*- -EGHX|mI!nl5v^g0QiX)4!-#;QQuFY%Kc+|HaUZxjUeYI9;^7)yxV<3N`D{JS2a*A6qz)V3CIUa*nv-| -a*!j*PhEC?4V5l7((7+TI_)y!~w+aOP;Xr%ImVaEsyYA431U4;`>t*nGy?aAel -4G*oQVjX66}G_jpOXm?2;1E3s3gx(o9Du(VGjAK8T=p(!Yk6m>LH(cN}^>Igp;DeiA_K&GsIw8N<)Y} -9!L+xFJgz+K!_9yAFjiP9W;3th!IYgiE`UAWCD>sXq+LaFX;f%(WTJZBE;XO#pN*CB8Uah0j`l2Aawi -A<5{29f0Dy2jYe6*^3I0y>8!EQzlYezCWhf!$c -zl3J!8eBG;jkq0Sd;zo)j-4`c1EUf2e`+4})lYjHOCNtR4Bj0^$}&ax{Pd^W6fR(4fdhDerI37>GQ%$ --V0dt!|8$x@`ss{H{7c{5pz%*MhZr-EltZkkr=n=P>xSZ6@thzCG&qF`oJYvCtQO`d$g)KA`V7Gs!6p|Ea9^$7inUTYGj80E!3a?n9y#C7hd$BuyK}w>ThrJ( -+l0dotm6nlpI80tYf56HJdEKG5gsoP-{TMSSTUz-+Xl-Ub;WCBwBlBa_9tp+H+jne-_ll1u~NyG+;m3 -5V6{MM$Ne|CttDUq#3*BPHJo$`huKDY>}eP{xEd9GAbr1Q(}${@zy+H_c^`UgMAcnUcpaD`ej=rEkZl -MWRl{aJYCH(+?gt{**(NTW8GOnnly=IH5g%Crje?@uglyk9@DgKf_>OJJKkZv;I50OX^kvZZO4%pxi>tlHbT~NrH|RvZw0jws7K<31)5Xv=a>>G!f^#f%8NulByjY_>7>2J>Q)PxUG7tmX9e -{l*Z*W{HK9yDf7I)(a9)WQTVRuE=3;05<|5DQ$d4Kkspbh$np($t!c^Zv9WA3GGpJYp(zf-J(XBzvIp -(96y!NCQqKNh!3ii<)Ff~JqHAY7(6(LCTRrW~)x5A+D*^AI#Z)d;Q%Imaw2gY3n+_2jZ%uys<|VvQ}(mBR6!y_Jn-zU4Ov%va7`?Y3?4U>m3M%sHAS&|zZsIiOwo|}Y{K(R)o!C(pb$QYz0mAM -8)V$K$g?KJj)`WuV3T;^8Cqq3*eNeI!p%j9Er9so>@ZugD*vz^+LzkEHW=u$dh?ftb8# -``*A;@0-LZab0t<#goCtpLqpZY#6gjN`>lejI=M?0>&{t%{1OemgwtCVqhV05SN{*I#Q-@i6YO8Vr(D -TxRw`)-PbS`XWu}dt8?EJ?=7N8!^-l2DXHOx%zsz(ljf{CqR=WS%6Ufy)WWJ6Y -CAIYXVmlOz!9a~e9c3SAHL!N;fLDzhqTi41%PZz>ZPuVUMYjH@ -o-lDQ!mc2H!!c54D8Sh(1&3MeyHZaFCz1CXukNb5r0*rEX*Z|dJap8{A2>&tL_}2UD-pgvgT^KSa#I( -r&5v)RP|+n@=3l%&*EbY$X4X#b-sx_jW1V7YmvtHV?_+61qJk!VS0_LPQ5C=9GL7eI5S?=xr<0ljFo? -(>4>r~fRcaXI_9?rH~%$o54!G0=UI-LI;0%n7bwp2!4)~KY(&^IRfJ|4HiA{<-0XRo -v5D#DHq5>qt)d99>)Y7~B=r`XUAw+yXp%~@+`)*Oy?CKjA=r+<3|?7c+TU!mKo(I((sEBpI~hmx&O4W -QzkT1{!W5kQtdSOZr(-(Qz=3}{7`;mq{THxeQsO_D*8=IMyTU9}_90I+OBT}GZu{A3L$W;U$Rz+gA<= -rw$CXI5jN4i8I@1Qpyh{O-(1wh|ml{BMiJ^kUrwZDMO3J@=_%BtMpRN@sNF1*EhW;BM48OL#u7BAc}} -xz~EyVqMi7iE9Xwb19aJNX=X>NeLMd9&i}Hi=HvUJ4oKQrtrhCt$A1O1o=A{dp4s`oOw8C*U{7z8BpE -;L^0=X?>#G2haZyn*xrYpC{|ra1%(K)Rp`Qu&ZF}|m>#tCTde23P)YSZid{ -kKOk6Lr%lfJ8vEh>_!g>y*d)Jzi2M?g(M2robktR{kldv -5bk`7F(wDEq<;}*O*n;xXRUv)Bn=4>7AN{Ejq_y++iJyM&+k$WM{qKz|GP%kr)(zbC_hCryKX{P>;fw --uZW0bfPYs -rH>na2;aGc3zga&jbN;13Ke!rZtF;o-cQscA8W<;g~E%g;D&jG3lkOB@h%g6A0{+!fPMDeQxvExH2cD -dq0jS;wN7=#J1P$)?QmG#sK1-NWHlT#%nSN~WQ}g_|K|kG06^YYN6z&P1M`kRemK@xa)+WI%v27PEQ+ -Aii`I>zXaDB&(vx?0LAeMd+xE!8xcxAlId50TPf!&ul>+7W94?2St*7?xukhlkM9+XT`i3+w$@S=cm1 -&3+8;J^nm}64uGzCs2ae$#Qtj!N{h(p<|E1|9(9n{)(+Lkl8a#V4W9 -||2(C|0(H$p~VYg;$I7YD?%LgaAfe85lvH#gIpjpmx?A*6}$8#>dXE)|ZkWhXiE(hQxSuB^!?eZf+Ot_@;Th7e&KV7kz_^8z7)Q1MYEormGQtZy+9?u1${msv@c5Dx> -qZ#4vY^3XGr?m4DYw4n_q<37BLEijKwgT+sT+v>jz>`ll|_rrkGPxk?4m%2 -KIwln0G>To{a&K0=v0m~Z1N+hyqB+RcFKBVu^j2W_HRCLq|>oG!Yax#W<l}WgzG_xQFJVUg2;Xx{wEE!n-^MO|w>YA0use5Dz -|ThU&oWz_+%Io4ITGo`+lfHCrd3lQS40A$P&p81+Yq3obyN*dCYC!LzY#{l{v_dZ%k{WznLfVF -W%&y+amFL5!@LrqtL?yUZYc!O83cNzpMsu8j~TJsT*W5`!6a)w)_;%3BOpP=9mdn*{D-ugCB-lu-oRP -oe9gcd1*!L_&?ORrL{z8@z`PXgP$!CJVx!jBI=z>tBBnV*MT^MM&IDNkBBUpOw^TVGGX;5i8yd$gog* -lb4TVk*c?h%9MPeSVpRIQ}>ggu}ybiV$gIYocBS^r2A0C2hMNu(yGod9oKF+r=>%>k5HVP5? -pf^t;?N)#ZrNA51(uZMZy72=)YNZOeI4uTySCeaO9V7f}Ic3~}4GQ609;Z5K&xoeZ~1q8XIlmgd@PVA -K{Q`kw0Bp=|TO7-h6bkYt_7_L#i4) -)HTPgGC~a9-$JPr6Qpom}$S21&47De76O^lFtPzq!<2|_f&_#PutJOIdxCQ%pD}LjG)u|4;|uYEsXY8 -gwTE4hU{7o -bT{3w3iHi&6?okM-tbT^W^Z6g2xVYNBIdpJ&^ej;i{X;f-k&OA3$aS)HA*=N59Y4nx_L@h)?Bz1R(LF -4wA(_yMg$b=xZf-HkB5Zrgl!o&8GyrkbIv4~no8?gSDmvyV06NDeQaRmh_uD}Uc*>K%p;tzR3Nja19i -gPHMyV?pC)bF&@^WT!h@`RE!H^dBoRTV-l4L-__{+;C6z>7LUKN)p!&=rwRij>-gjs6*gIK+P!YOW7f -HS(f;Ydas&fmCaM=41OwyjICGD`yt9G|oP;#hYC{J=V}Z8^@VhxQ>&hUYON=RlPVtIZq2bGH!Z7^$(H -hOhJrl_ntlV9EI~V{z$GvZ12WtH|{Dq30tTTQx7&3(Q&$Ed_}NWQd$x_mD;rm3|l+5*2y4SgZQeX6Ve -vq~Bgtv*KdIXg{E+iFWm8##yPs%S;CpVpV90IFbr=F=LD%8fYf$;>?J8(Xh|nh}mTM+k~}3=qia^587KHj$YUkf -HD{UY{lqkaT)&1)|fcyqNnA)8@l5@Q6kX(5Fmq1M?|^LS?j+kArR8yVn5j-}!A@LihJcWKi*q!7)BLM -0%@ANNskLrAg7&EMcYj#JJ66OJG=KYCp#uxYIu+Xn{(5D_1+qLc6^}Tafly71fHB$ZeW6!Au>v9wX?_ -LyDJfZI4@k -8-xU9EbywTmwsC~N`&S^;WCSWCV%d{6bw_O#iPLE6SV|;MCPQ;T@J(h!5ez^|QCI8yx3cBc -N2h``<9UibF)o_&^W@l^&YsH-0;qv%b%bGVn~ia6iKYqCo~4T4w4TMgQV)Pr`Z8n5UUizaR&Afol+JM -_VIzDkJ#C}cQCG1M3fEl^@K7;}0fssDNCgJ)6%E(iQGOmkiZ9eFv3mzEkIdYDAK@i!xnzhDyy-5CVSo -_WSxLxL$^98e%GqTEg!)?KOGY`F#3QKyStdXA|mRItZpOZ#-k?aAK#g>1hX{4Kyu#_ZkEmu2*rnIiyG -h0D#|mn9qaz9a}Kp+R};3s98V-H;3w(svzbTHQQf=NfZs9&Z)*#SGfJj%th9E#3T1X}w<};Dq#fT30O -)mv}9~-wtBw{k>g2cTE)jf$tqeH<#y^Z&b=NFB9h(FEWJC5{r^BHIMoK2gfg8|Kas3?DEl`{>kWqmTGF|I0kM-vBI^)yo;|sfAV -5E3&jVI3H_X+PbBuB)vPy9~^aP=r+^J}xYN;)dyGM)AVRAG7-iDu1+p8h>vmL5%J?zxX`&AB57ng`eS -v(09F&^on?~Gj>2cbr5_q-w846pAp6k~A}!5!~zedzQq#qHb%(=>?0rj5NnnX-8d|L=Ae#WrFj_9^V| -j{Wxb_F_59tz9J!cvkJ<`)8-gRTbuh8L=ir#t!9TRE84HB#5(sv(xH{4v6FV>d8nHP@Bdd5zj>We~$M -S#xt4&xEii9_5jkE2c?b>!HD5?V3I`l84^TIXqCDwAZ{O6i=$AWe32ma>O{xM@p1ApnM@yEe)a10t3O -RMbZK&#&FfTNw-a%d^4ygeWse*3s9NW2P#2#n6X#}@$nCy8F{DQ1Xg$p -%Wz)n_annTGUxCEEU5Ky_?hX)0R`X+F -E2a#m3Z-+8?qG372sqz;n(azY>L%$Df!A@9+)p1E06}HZr6S3aF|I<9<)VRsvzM2LWi5zduo=WdCqnY -ZPl-?Je_xNh!3~_(Qb_bIu#!nGq5Vx9k$OH#|gT2=vbQex>8tu{QgXou*u)^B5jd=twojmF!k}_U4v) -Wege@=kC{Eo0g;^Knp(xkPRt?%mrWCIrYmT7Tj-bybXcaJ7=`zDw2qKNB4SEg5d6#=%scg;ah8sC*AoQ`As@+ -azsnnJn4*luFE&EEFw&NIzRJyDgcrJK`id)KxjysXJm9A;CC|>0tKSztF?vg6et{B58URyp3Mr6TV)YPgIq3dlb<_yO8V3G-J! -MJ5{8q0Wym6C8*tj(FyX!B=8g+a{Nj#Fo2lS#Yo%yTe(EYJkZ%ea}R75e(tZ#F;X@oe#t)Z|FHqXReU -=a5C%%d_`C^xoa{&LhwuzrQ^HGfnm;&Zhl^w-Y;@o7AHD{hFBGEFGbh>Lp% -mjjilgE@_>cs)-(ArUU=QPRawyMbxBpPltacsJvzC90@n@r$UivC<)4>+lR(3F097j#xU77hH}y*a2y -9tRDx|H(B2Rw}>aWQn^40MYw8OZkNS6iKY91-NeMI0er(``x}o4&q#B9X@#GJAx`aKaomn_>}o>U -zU}2V40+r%QrHno892D;85d*wjQXN(j;B`dEmD4uI`|M)>?TXElDCJ=j-7&NSIq>?*9T7+v>SVwAJp# -3*|cqHuPNjpqvC2#htVQPzUpT<9PsYQPcG^4!p9`Dj571-gaXMw3GtT3KJ5xLFZ$4KzkZzX-aU^=m3~T~|SAbe;j+cvfhKjL -|N|;=0ajj%hZgsvl#ZmOXY&_l+HCKGARk@j;=uk<&nnn?VWPh3_GpY9vnbVZnTVmz}9fSo<8RWq=A45 -UkX*>1|U#n;U!btUJ{!kv(lown~yI+GnRDeqY@zQv062$qw$SEWvr9{D*FD6~GibPVHuQ9;Xm4Z}H=2 -QWgFg*wTNu!H;ber;8+kq-(noY75IudiBQqVyoImqV -?+*t-15LeKse%cUA>gM+t!zb7v*5vwh6)_>Qn -ou`Ep3)i7JD6xkH~oqE*v8K`yY5l3eWH$q6hpiTZv)&G=_KDSiop3+QC9aJP{oZq%cur{Q3g{m^xEW{ -bJ{}(ZzxLXoI9tE&SRqRA~bD!No*A`_gZ#!H}n%}8l;>kUF!ox-ytbJA6YiRB$|GmqpaWI^}NucFzvm -mh_fBxy>rk7s7zq-2o$$)XV>o2dl;}2U{>pLi=#k(VWIqJxPB-bk+4SVQXW;~|&dZ+3MypE6#6UKq5` -@x)R@LK!G2)(WSqDi-#^r#J0x4+E)_{}#BKOYR!+@o;~*!1Z5$KzA?I`@unC|pQtL(Ky0;0|J{rZ^l( -EYF?HB97wi9m4SHFGslkm~ZiLha(MWmBDo4{iKIy!QoKSz(JA7gUAN5f&nvggK#dKOo>e5R$rbBggqM -vW6DL!*R)7w{cuIIlP7ed80kifYTIQc7t)@lGau>$Zf3srH^JIx@ZFhX!YwDb4}zzn6?HXIxtRmPpaW33Cj -;AesAfojC4sZBZ0h<=+Eh!u_mn^jLuaGGVGv@^y193Mb@dIaKlDpRUBn41WsVj=QN6_<7429JV)rV)< -e2ILB_S+d-(J=7%8c~qAuZBJ;4S^=Z293gW%N5jY~?8cw>P-7$m|FAhKAmteP -ELumz6A-{*F|H|g0F+m@Q#uESV40UvD#;8!lFF+&6yjvSMG9S3i5;_WGOR)fDosZ$LKA96nZEFNO#yi -eB@{?jr9KiN7OQGr+J@uuW%u -++;VD}&AC{Ir66>t&G-M%`tD8Xd2oTI%tB{V{^?VC=iw{n2Ou;Ch -unSP^Hv@rUYRHOx^biZx2aWrN@tsFxYW*YHcw2k*%o%1RFmlY0XmdI}Jvp2|9GR7J2imNyZKe%WAF&e -r$8{M$pxFK^AaZh~0UihN;+`(R=fLkv4r^Yxzx}o9rqsgv`D#agXK8WrTp*b)^kB+KXn@2~iCbIoN=2 -U474rV5QU86iy)>y|2ND)L`&MUGD@K8=&ySSCoa$Fg#acUKa>_~7Pu=>V6c_Aq^1OZ!238J^6<-}+hy -XFLN0a6MoK~e*@H6{d1Onwi%{Hq^Mxd}E -eWF_uG|q*{8>0L|8|iTMB(~gS4O_#|#Fv@heuS;r1<{0LAd5`=z5m*5wI*>z?ww917>N;<9jXz3HShT -!3I|b2*_6CZPP_*Z(UZCSj!2m*vCt%Ad?u2s7rC{)r+jHFseL33ihGjtzh*@?&J!0gG&)XRv(&|-eVX -Bti)Y@I6-8j>;%AM>tkx)lXq>;9a=2W)iOOd1(UZx-?znlG-XR4A3 -E3Ycel)7LlmN4O?wh)U;n^QM8e+o4^v5VI3W5q`kUV+G}Db2e#~b=wa-Xman$wk+lpew!=$8ZMAoAlj -v7k!JrG(@$h!q5)!Jg -|eL+L4-K8XC$xCqzMHr{h&K%~fp!2fDw~nIur0I&IZs$Ik7|f)wR8w6ScB-a2UWu8rp$rH>*d`f<9D+ -q46&Y276>wmzuB=*72u6iL=`QF$%o+e%D;RW*aq-ZiY1WLjkRqe&(~?ZV%o>X=Z=Ldbk#y^`xwJ`vB) ->WF_JS37jmLr=DatL4erapNii(n)!hu5=P8X=S_s!Fcn_RWCjJzIXNyP}JY(41$qhb5zMDARN=t6OE} -AlyZ|);;HI^$2oOkC*83b5DN+&Y^h&RDij!sK<>L;5p2vwNzf=ZsU`9=lBM#Qe3Zt|W^ALoays%QFSE -dfcW329jf}V+p3SI+4fhP^!t?EvRAN?J^QAX!E2((k266`yvJnnd;?2syfP&c@$zBA+dnLBx_W!Mz4o -l*H15ir?1QY-O00;of09jkC)*D@e0ssI(1ONaa0001RX>c!Jc4cm4Z*nhiY+-a}Z*py9X>xNfOi4pUP -E$o)QcG{!FbuxyR}eY1fY)_B?xxsEBg`+zPO}^xTaGn2Rv;b^_yJ&`$4P;ozDn=q`fZX$`i -Ac4*u|M-VJD(<=}Z0d^l6GnwfC4YVlEa54-?<82=XLF2t`O@n;U+4g+YQ)mL|^=8x_?gH9dR5! -%peirF;Gcq#)y}dob#e~70132x2Gc9pMV5W9FcjV#ic{E3JQy}=f5zmnH>@flPjNoWH)6-Yau}xa8}hOD1xq2ab7Po2wPkDrqm&-$~QP*TsB&f>~u#b3G&) -S>a-j-$O$wAN0Yp+x>WV|31DZ-47K%k|)s4QTahfevu`xa#f656w8jUf&jmy13rvpE!=>eZa$;W`vV< -v)=zPA5ecSpek1NmMa0Wf>ISPG7$+Zc!l*Mf~~CA#ij3H*;wpuq;*`rQf`n)MSo&>@qRo>JgL{~c^^5BFvp02385fF9MKpjIJp5ZO(PsYj!E*Ir6|A?9Dop0*7v -**2!U5q&QmrV3~n%=X4n0I#Dg?~T*!$fn|1P9i-$O3045>G7`#D3R8JV;O(=hHe*FcK1-#+ihT9=#fg9P?&Ro?_h-K=(S4X=-F~6WIu>S)-yA81TSS@AqnUhP5qy6U1 -JBY(+6!^uxCd!}HV_Jtuy -NNC@jgTCMl$sN;DKdDn`_$6Xl#61+XX$T<|U6)c3!+?|u{)EB#2U|Qt%Sg!*{9OzADE}KWJTF8O;4{R -1-NzT5{2+pi?d^l-uchOX^~Pt%m*?+~{p0b~xVN8=br3=hdwGPmFkYUYUuj -0+CXQm|m8R?u?@mrn1+NHugL-j!^7j2-z5S|}{81$0KB-8lw@K$<5@ZrqhuKxg0`h=#-}71NUlN7#Re}9fBW@iF3AigFlveCI>X%X1kZ-xD5Df5JemO*aD0n^ -=JdvCsw7wfWv4eV_;v?r$bhg`tR47TF5EM4Cx#%`YB?JU92rB@mvJmL@8$Lhh6Ki9Be7~_|=E^4WpE0Pi@GRe&lToV$BmR!<76&Y{eTpb?1I4F*Vn^D5#`j5V2#@VtRbI1 -!DiiwcvL>`AJq%f|=Ei=|vbp~G9A)MlVZMEnt&evc2HEG&QT3&I!oWz2+CbM*7E>UNh#qtzqO2WWV*!Z66ebJtWu{&XvfoM1EvLdVK7A --MJY*flD8|UUD3=z^nZqE>GOL6RQ1~cJY|?}WA;=Y8L8t6*;511N}djtJC90r=1}>_sI&oS=1RsP-z+ -NKy{5a@Tnx)RPPY-3Yu{>?2sB|CkUL8RM~+GSP`V#h{&}sUlsVQO-vGGF-kbgQP~Jv<6*UC{D -EhF -uE3bSh0ZEH%`)pkx;O$83&av(*&9&9LMgqZntw)2X5;Zp$A3+Rt2?(5z~5 -YI$(BTpT7 -X+c3sjwP>pIGlSMI(yV2Y^PJ#DEnFRxvL45!$z4Na=hCDshubbZK1d~89gS6KT8_zP(6 -$xB&Dlh{S0!B25M7$Kgdw^ZZgn$EWvzv93#}+i;Z`-mQq@uhx6lGr2)C+!Lrt>?ZlUs`3~p83m6Em+x -Wy5mir`kBe?w7Q0i4TEyPcmp@O(v=**vivWg@Vx{LLZLA+t-!+?MV92djNk!1R3>MZPbB`ExO*aq*40 -UeT#%4yuhN{HV#sM=_hE;)|D>Zom21LY3CpB-(H#1($dLz(_~Idqmz17r$K8lmM>7WP#etV+a~P|jV>*7+sIBnjHJur|bDx5t1BOG|IthHZ}pp<41rU};oI&JsfYy(2qIomJ_f?b7;W- -snBe(>#fDekY9s?NT^^N8{3(bIw>F(1H;Cn8u)m0AoFd1kyQ7G8uI7yv7^97JKa!hP5AL>M>YlLVRLu -&)7S{R;0MsArf5(#&-(6%WkI&G*W*hI$+6H*o~H)$DdRl-GEY -+6>qwHB4uH3l*SUaV>DfXhS4QsFS?8-cm2%$1eWe+___NDy;QClV@mg6X~tCz%^UcDPH*P*oZpWQ(oR -q1A_94h^jB(fagDPhVW&3di)*Fmd4YknHU%x>SyA-javUo?{-ShB{ymq@iWP#u`aN6y5s}%}9L!K8VDiOr<5S_(bD1;f~3M#FmGv|6uD{V9hlNu> -|Gq0an<9IXjX>VHfQlAgM$eU#mV41SrD>NQ)v6C4j$}BSl -Go6gR=qQ}qyWt{gO(xLVNFuiSn8O@jk>KZcOwD`Mi}hBPA5-54LsGK?OEjqjcBO9>salMD& -O&-sZ)N_Ro11E_2kcX>+621PYcCcgtRT45mWP9~yzCl8n?!4MkP1|@Fx^A8SfC!;LarWNojaSFt7X?h -+G_%&nd*?DMz$8I@0Kni{>`}pRKCI1BB+Rl2+j|PL(K;|fXe| -Q5fshI1g4z8I73wR@&!@z8nM7n*UY!nix{gZda)GEOa#m{!`9A0$)H@HOGiIBUpQqQ{RHfVMcZA3cEH -I$@Bp1(yzcj(EyAN^>uW6clNn)3HZ3%nd<`afASDb#&|ET0A{~Es&ppiamu~8On-*qe8+MHsw=Ci~Xz -qoT-`ESQzs`%ytVUkM)Yk{#rcPwy8~Si=)$3izZEEDf4GjoyYJV!cDFUP86L$o=PvCk7tVxZ`pCyv|4 -Xrn_8ybJ2b?ePeO=(7cBCi#Xn^HfO-Zbo~_EOv1H460yeUUIlXKTvW{mNj~9=bZHS^S{Y!OZt>R|i;h -yXs(0KaGH?!DGz{s5E{3*`g!}7x0()wZ1dMcar`|Ui7;yZS=ikp0yW!MW>vRK@pJA^Syp+(Sb`01V~l -kM;IE}dpJUf1P8jJ)n2l6#*kt62d=gE&^u&ZEa9IS5wZOWCI;o4c}lI;me409mlE6ZF(Y^sExpE=qq|d) -x1RHH>15L{a;!YUAWCCa;G5B>6t%krkM9kIiqwOBAL(VfbpF$3HxE)P?Euf}Y0a!EwL=@*e6jB(#EZ-b9K}DRKZ4c6B3vy=im7J&5pfvF<77Qc@AQC~I;o -@4F^aeW>P+*WHq6o4yBr%xQ)VH2V3pn7hu8X#;Bcgn#T)Py3B>e(B(y1lZu@mjIJa+W=S5qWE1wVJ)j -sss=%PXe>m|!di4rW$YxvL48F5}Fw@~5V_b*nO!K$|isVOK*E$kzL~_xYZ}r6BmaKo#o#U>oFgk_SGT -#5mXxFQjlfybyFfrcqM7OcmdUnxoc&F=NUlcQIJve#E=saQA3$_xbMUzaBsR`+ta$JvO8R&CUt;@0pb -o&hne@$`e9%2G~KbpfqO3APQGBQ`~K$H6oQfy;)M+Fd-6lLj`BrmF22%oE#T2CCy5&AU)w!QtJ3l(*Q -?#YAr4093H6^3R;e)WHWYKE2SfNF>lysZ1F5bT#KLAt8_N*dG)<3KAq^#MDC -JPO!l_at4vLBTqMQd@3TrxU7*`6GQB`b4C(VVQZqfl+ejcONgB{kAsP)h>@6aWAK2ms3fSzBUzjPzy)002QG0018V003}la4%n -WWo~3|axZXfVRUA1a&2U3a&s?XaA_`Zd97M&Z`(E${qA2uq+jg8Rg$gf!)S{oX}1O|x?l}Z6h$C25*@ -RVM2)2C8rYBDy?jWNpYkI}VNp8dJv_X;Jdt~Ha&k!(BNsQ<(aFh4uQ!n*CviNfbR}6F6P}ksYC?-bXs -Wp=RIlgm31!jrKJl=N%c~)@m6kNoInz^-T1-jxG~@RkcLPoPK+j8FJox*!UvJ*WS0CPA{dN86PY5(=- -x>7?{Fr`;$TF4~LQz2AS~E$tkPWaM;3G#s*6a0>j4DNLGO8y+=I^*vK)w(E1|veAkV5=RN95+?KWDGb -o|(}=o(nfKa*vwX5ui6I-6$mG5=aTjnX+K$4#?BDX2zp>EL=B#eI*&yj8Iar+43sSqh5_1l~mHaS*_2Ek+ -ExyqZKf}U><1O%g}m}4A8ZfVf8v0m}@P=f^{*#y1fjoSHb{YYZy@Fx@ibfmnz*j;n*vM8?<3EDF;6m>+2IFp0&1C`1+g+`MUKFr~NS$U -LXX-Fv8OvFrF&=PYMT%EA2YedDH8lCdRFfIC@CwSF5e6#r9G;r{Hiv+;3!+eFcq)y~(CZ!)$4SAZtO~ --5lwnk$kj*1M6%V`{lQ03x}omC!F!IHj0uCdWJ+*{ -4*~X1fy*@C&dPGZkjTl!69~7MKTciOB;)NriCgJ3s}|Ga(;=o32U&@^MNjCOl(mXlIUni3vR8S%wo=^ -CDp>36?n%54<3K#aMl746|V5^;GLpjZRNBOQsMq%6TG%5)&OkW}dR*X{Ao#AeL0lPm|=u^XJcAzUqJZ -9G#yHBC5)tZ^!Xqld5&W0A1_rt?6e<1%oq*%%(h^8V57;S6GZL$Rwi=m>`Zdu9-_0hl81<^suB|}NtiPG6G~GmKw&U$#Kc=8TOmT*LSQr(cu -}jy%-E`dA2C+67%0VV^Y@K=A-|wQGRDH3u<+R&uqGK$3|#9z0>7P)?gsU9;2xe0O~EEV4m${#nmiXOGL`YW;DzOTb_Y|mBMbsZcl+V|{AD`wd~`lO*EULCTO7brv6rI -dz^cz+h)eg&_1PGVxSkdOrrb|#_qd>dSk7Z=@N@a(;NiE%vOtWtHo5sI6%xNaa|J}xteUl5CbkVY@A7 -p-4i??nxX^s(b8%*yT9M;@UVZgaoz;J4W_6$>e_c}t5`XJRb5;QqfOn-&Hi^N*d>#CZd(g~?s$E$jK^ -5TaJ`He%R2#f)yPH~k%gm1b5yG2(V6J_t{Y6>|3HFs`LvK@n)sVAbC@Q04!E%?aOKa4>mf# -L+KD0Peu$Ro#2MN*;9R$cd?#~+=?tjGjSyYq1PBl7?&k -1v(F((v#jv-rVNCc1L4;3G50E5(zJ1Pu1r`M}hp7G>9Lu;?RmfWHGZ&%0Izhdy#?Jd{JX^4RtQwKH1b -gqF@@<7`&V)b|KoiZM|w)3MU7w9UcpbXH7yEt{4JS<5DL3tjoR(|S*V8e_0oq60{+aSxlN>cA`T6rE| -AQ+;2lynWcB)V_>%PN(gJ0^iYp2%&5Y2AbmfCyWKTBE@NDMNIFrr=p!O#Fm!;4SZR6)1OLu -$t^Tao2fZ-)0LRBH+#}*oYZDg*6(!XWF60*lxl0WCBK{l>YQE91MQnzjp -5qN%R7>;yu2sblUHDXHkbE(dTR(*9PF5L*Rgi@<`&PRc^gd*s89Nx?bEG(dHXu%z=y-UH&%C -LFBcpL(Ix_HZLGo@Wy?yUqYJ^Tg1(OQ!sBz3rlTjKMmS1UhOQJm&2>l`546OKP|fOZ8p(vY#V=iu+Fu -2?WAwBk#-{-R@_4=ht;?X>DF%AM*DwIO9KQH000080LuVbTd|cKBGCi@0J0bW03iSX0B~t=FJE?LZe( -wAFK}#ObY^dIZDeV3b1!XSV{daVaCzNWU2oeq6n)pPAhZCp0asdHjV@T1v`w(4HIO`P2m*?cu57N7sF -Ac2W7vNCE=ftYoW!P9EZgcATjIGN=aBbaOJNw6Ed6v9g<%*3xfCU#G+*ka1WE}nYaumZRV6gjTvRFun -tMV;HPegsQmvJ<*6W&A^VahEm)To-dj9tGr;Cs88~+ijdAmL6RzNQ -ovDc)1ZlyPBA154Xvf>w85?HiyImz2myl{9IVgfRh+Y7(A`nr7M4p&G^3hO6fwg=mibC8g3!E=`#;O! --PIfhUQ`ZHMOr*;aDyI{9z0kT=(>Tsx1%>g8v=J7iqKuJ+WO5xdnJ)MW;+|>zju&vrDmDjmFFq9*t=3 -Cc*p(}r*i4GKWMzq|zQ=?hwgUWowqf@8PPEWvagxx;W_`&%BmJcUB63SC!Tc�(o6GLZME|Y*w_RYt -$f5l8F;w1yWhk2Q3?9X9<-uCk-CB(#pfWhd)W0{jq1){P>}kMrysbR~h7nprjeojH1txf+8Q2S0_Yi= -?+I@K9zt+zM2vP?GDJAXbu_C!U|+uO$bEuh$P9M@rPy;&F=Hm?fTWSHfJ+FY7~l(8p27^oPP^SC+oJ4BDB#LymwgRb -6p`ijrT_i#-a&iyfN~5_cMn-09y+HFjUZF=UR#P1|W^yh{xRY?dsW@{LsRf?&iQTs@)vW;cyeP;W`mZ -u@4WRQ=OUU^{N8@PGEKs))TfQCyl5$ySZnn?kTzA-$gm6WmVs`~(6YRWMqXe(I6enAOOzIL;jjb7sB_ -r0=D;pR0()Kc_8MZYiSU3$Gw=F0BET=Z~@?@1VoTF15B;n48J-W$0zI*Xn3&}svU6lS7FJ -vZZQ_sjRd{j9InMY{Cy{jKb%V8_~}qLw~s?bF2d+=xbQ%kZ^X=9Bk7LNbZck3 -Uek0EJPOgTf>)z1qw#K6>Q+#~!@a|xH_c<#3;(FExQ&n5J_Vju-*B)xWs4QK8Z>k16Y(k$J6ZJwv*##RqhdmT{JcxItI5>%)7vldid{<1ymiAXafcG$+dT -i*|J$E~C{^{A#9h)QEKkV?%{^AgXONelNQGHcKSc(f)1X{>k# -!k2^t`%$hDiT|Ne&+IMMGkF6YlDokVtp!k@wWxh7!J$3tz=4NUfwLG$K`txJOW)zqp#Rp -h#Wk3h(s<(n2j9y+YV5=w`sDGY8a;aG=ug-!_xp$J7BC0`N?B1*YW^X*H@x<3b~p)cgTDY!O9KQH000 -080LuVbTL1t600IC20000003iSX0B~t=FJE?LZe(wAFK}#ObY^dIZDeV3b1!gtE_8WtWn=>YP)h>@6a -WAK2ms3fSzAWTQblvP+<5 -&y2g0>gMF(u_?yy?$ujOQw$PB%X^ck0oE*@p&j%1SQlI!2*yUle_!dZx;Zc)`N1Crc-^vUo%W+V*bxt>cFhJ?>DAvK|CDm2wxq^i?#zX_FpQcubtNk*k&GG{$9zg?K$ELBG -JG~u^7!3BJ^f9R)-r<45m;lD0U!lO?oM?amOf7~VK;*x<vUM6SVeE4a#iixola+*P^FBxZo>k-143SrRQ!h?k -c-ck`+wd4vyp@@#`nI}BKXYH`Qt8mU?c{!hzFvljEw$y{E@t;pe>hVkg+lOajJEu4uU|zHy-n`jRZfI -j73b{X_5ItPJ&y@w1;H1FMDQH(2*(aPxUN$Hwt}g%^|FaB!r>@iimqAC66!;fEzp|Bxi~|a6Kg~z2{P --GnQ&{PbH_f2~)li!5-->gKq?k{*V$XnZW|SqVoi)lQ~nnW<{EabfS1s71$$JTD4ra#`A9Eyl!d@e)xMPMaHClDpCuKD#M6s8iLHOVo7YWOlesrSRTjtI*>N|2=Wv$^F%a{8eyhX6JM-byRUyP6v0z{iXDK -R3f#W9;CQyG+hhD0<`qD}#Jp|F)&oq@buu|KF$2-z!gvgyPnS(#HVrCDg)R|I#od@Deg?$qG^&2G^#7 -)uj8?chFM5G;+ -Jl1vrxhOC3|SZEGN2DbrgWaY|FYal{#>&CVpq(RPy^3=fCfi@H+UH#wCcux}oF##M<}v%?hr0m_x$Q8 -*l+ij3+ZFv?Bh4#%dbq>~7%Ig;mXL{&qrFn%I;{`~;qY6`$GlFEwUsRFUD-h9=2VI -F`(`EY={9fop6(wZcX -#6y}08np-H=-UEvY6vZ3`zSZ-?f>QEvw3;`C^Eb$L>8PA+E+mU|vC(H4m -%ZinM%L%SJ@pHGiZ`V|aT#u@~(x#HuN01RzA{m+neGxg7|2B$}*_C~_G)Ub2jwuxb>Tc(92-Y_Xn20s -n^7nPi_Qr5tbcF?}*M()_!;0b^>z+Zx$E|19Gyw`R+zq~pho}Qf`q(v-lbZTB7ps)WR8GAoC4*IZH;` -=$%@CGftCJ*M~!I{uCPVLmMk~ZP&MUZYuhGiK^)n)qodxmHGnNS+fC(h0v~u0^Z ->HBI883~(caU2;2zOyx;ZEEhv_MsDf&4hxGp5kvE-#F(1~(-j*jAbRHJf;Cl|(Fm`gD0pi!PpHI}cEX -}PXh1DZ4E5l3CDV$GrPy<0lnH8;t|;=tTDG`)Ew -%XnOQ|U;jn_WYy*m9C6Wa{Gk*rMXi`wv~))e5~h8=X()7U3{97#F7Fv!hx3cSx|kmC0&wVo}8OXVC6B -6l16|W-ZwxEvxMq<0x!pUZN@sSk(ADF$0DaX^eKIQP$`d1j(I!}hD{8}`b0pj^H|Vjc!rGLWq!Sa3J -{68)Puk}K8D*Ic7?d1uz?nCp=RT|QLe5K0{m(V)4;tr2&|Q1c|R5orFYcM=6R6unX=E)>QB_zJ1EisS -IR;DGQt7ip!4R)sZo1}DYr2YKyp?h6Hby(~ARHV0-4a6X@zbt(a0$a`kWD(pM1z$fok6$2)5hozX?$293NhBwNjLVc!NNDcq+jg9=1*+iD|&+s!0RRy&1D@5x%Tjcd{fuc1ro+ -Xr(zak{2JRe$h&dj{@Qz8M*J)C3F0@w3oz()_&vmfnAyUwng!YzPBMTwW?gT`jhh#aj-d@&8TwY^P8d -SYgdu2qZ3D_YV7J@3>HGmuO9KQH000080LuVbTMO*&ls*9f0KftO03!eZ0B~t=FJE?LZe(wAFK}#ObY -^dIZDeV3b1!yfa&u{KZZ2?nby7`FgFqC#zhCjv9;(SodT1{sJyleq{fJT%(~u2?ZCzbM>=FrE?=z+uZJ<&VCs>#_T{Y0tpTi8Mxjj7Sj8U?_4 -q7>&i!8IT~Jt|$#dzr%$mg-A|M8c-j)HMJ_OUpaDGow?h;i9kXvjKe!07 -~CP#PJOT9yx?~QdB9sT|zSDGS0$-V~9FX3X&qt)^$Ae(Y;T?*DxJBEd`F+YFMDNGa;0I7k0|XQR000O8%K%wht(cJU%?SVijVAyABLDyZaA|NaUv_0~WN&gWa -BN|8W^ZzBWNC79FL!BfWN&wKE^v9h8f{bJNcua!qK~;PVGAQ7x~sJ_bq?T;H7cCwOijsDXhQ=zLpqyo -RBmqm`#n#02NFmijLuqC5z^25dv{YT7C*cGfIkVVSS%I_!;nviZ4ak%8Zz4^?j+!$B-Ho0q|)WSC=|> -)E}|zf6B=XDkXz}Jtr)4)jv>G4Jul_WwEgHDdddH)3$iqp~4aMpE1~4 -8baZU;Z{uG|WJ&!p`K#Sfx5s3!%JbLf3LZRSzR0v{}EiKV5Yf@+e{On(Ln^%PUoOgX=brtpa@UI}`0So0!eH*YLx+1jE-8RrH3FZyUWcM9}La3u -99x?cL5;Djd|3rrf;6Om~a7Lud1{g!=Q>0lT%o-G -LQcaeTZSNHp*1UX5JEz89CdjrL54W3OiqL-?n1L-BnTT*Ov&zl$aX{XaXy*VusNK7N7t&}GYi>4(&HRBK&`^pJVwVeZ=Oq^rzM^=Zp#Li5* -9tN5WIs}G%G7J}KXS!=24dmIsod{UI`#OIYscuZT0A5D+wS-~WA%sso4z*fF3-%xouQR5E*>wI-gH;tpSCl?&k_{>E~VRVBojhEON&XR1aBT_hlr8z;S9C*& -j0N2hRl=El=BXKtAUZL}{h+gI}p-8R`W>Lw)a&>iJ8g$7<*G6gn5Do8(3l@n^m#Z5r1_1Mb~WGiNdx{jdDUdeREi?Tk&bbo&t*w$n!kcBkGHiyX+?B8A+6p+XB76yR)5fQNuVtb9J -pBkR}{R=)4kbRJHu1xW2iY5_6)LT=#bau}yBFAEO%f!E*&^EOW+?h_8FGb$o8NsisYbKSJYEm;o?WYA -u&SdGduDk84(+fQDrowe{Dx+Yx*X0t?UZ%?BZTGyfF_D;fkUYqnMjCqK9y -}$BH0FzPsMY;=)9y6SuUn0q_Vt%ds)bWux|6NPI)nC_=L|P&5@hW!wwlQj-6hw62JHvAgRL_A&kycK) -QFbL#%oNwGNvW_d5rw}sUrqhC=^hc2_)ZBGmmtORSpo?NhqMp?!QvEiik9$YTCC=Yn?&52^R-0y(s=;$5WXy|31y5*7>JmTG< -9r?8Att89p_OFmRhhUZ|8@V^O0>OUK7rYw5u+FlYpp1h_cYG%sVXc_95CTki*$VQM2u-n?yAk^n_m%#ZTqF`=8ZpKhn;2*LPwsrT^ -c_5NGW8eL23T;l3dnTks&UYp%7x5kiEe3k{qXapX#qa`(%4kYM>Px2Xv=orTL+q!J!QzSO?W#(8&(mp -d|EdyS#BOHrQ(c-x>gVYrdaB9OM;d%G;XcX4SZyOXC}qPhY7E3a%n7(jUWJO~J}G`X_^sY-)sCyX`?Z -6!-TL9t;qLMNQGNH|xKTf@9k%L6zrFvyN+h1K5K=QT&$Sd#lT8Xv8K7HmiCHv-b%&Jd)v~&v(mMlHa6ndyBa` -8{GjESO#Px+_bP!Q|g^T<`JR-DS84;n46y2q+DP^^z%{A=caAVb~{l2bHt>Pz2%fa!Qda_?%Q%Cj>)> -giHb#-}p=L|%SHWZ<_F@IY5QOoz!ArLbWYtB8VzKb=)hP5H=ZTG9Sx|R&X1yWM&wWx$7(s&hBvS20N{ -@+)t$M8jWQv;91iav68fDIWW3OMXz-AAU=W~H{pcv>np4{BET17w@72_2Y8K9|romAJtwM8B$5P&uRK -8-i&v3h6-q@MVngtTy_(#1Dxp$ruLlu{#>?-oqCMHpjw>976qBiHC!dQzq|7?eI_)V+YT%k;h3qC$4M -NYJF~cSgz$au;FHO1{3%Ul?Ii&C^2e$1@e9_0j8yHCX|x33Q;tYhvM3`xcoX(dyMDincNVVec;?a;LLRoU5d$wFWDiZNV?eWeA!1@7qcf`Yhc4vbXA&Bc*ibL5A5jIM2q0e5ZS&Y3(p;e -J%VSP)h>@6aWAK2ms3fSz8MLAIrf4001)u000~S003}la4%nWWo~3|axZXsXKiI}baO9ENkc_WQ$?*( -&5qhI5Wf2?0uXip|F#8QY;+ktWniCHW`j1t3Y-@ap))v|4+UV6&JmyBF{W1A`gQruFt}pt<;i8%B5-jR9KgTu23bdo+_Np~oXclrc=0;cB`mA<}eq^n=}9ymd -V^eFvA(fbC&)m!UQBWM}(kc!9kePGbVk)*rwP@#`+0oLp!3wrv7H0?lCHY=)Ed-gIyt+|_ni_(S8P+P -ym--LvhVpmlv`!_XLp$jSJxgnWeeibinzS1GOQ44f{bM+3e=i2<+Kxoc?MA|e32^R_h!cI^m9#0n7iW -PSIW5+rPojXjxR`Zop0beF)Z0<_j$P}Tnz0l%yvO6S_k$@IQ?UCw9B-C=KlQ{&BGo8$PFaa>FQy_b)U -bj}qhvCtbT8DL#WvE~`eV7>*NGe||Xm3*0N$VHJcd^5Dn@U+yD&ufj3h3{E3WaH7D*)ia}2}+R!pWj{?a)wxZXy-1=nTZS_nxfR8(r7)&- -SN)v^+bL9mr^l@^q*ST;p=WCXTmr3RJLqImC;h)v04kbZ9w<_ro>=LHKy(PYLYOLcJbR;3sm$}1A0Dw -c9AFbjrWsN5zuOvS#}cnO=3(G^{ywTFKsf-6aD$yPx!#-QrC(p=Y?!BU7U=B$`p^OUJ?PzV*nsufE}M -m3GZ2!Tmq8|!(kc+8BKnn_t#nu~IR>)l`~sFC7m7Q+=~u!?Vl3AqiRA%mEG0-K!SxeR&6q-cmhVPdJi -zlMAmpeAn~p=8T~FIkzg&=mn?!xfw0s<;Xlc%;SUBWE2ghZUhJ63e>>o5V!|UjWV4JjlLX#?>o+dnPf -mG{40?B|iaBO9KQH000080LuVbTaA*H9PM -tBUtcb8dA(NKZrer>eb-k^od-z;O{WH0paT4mxK3@rxgb`H0!1ON$f2|qxn%ZYS$*iY_v|i7UE-uKMF -j|;w7X}{%$zyPv$L}i;qzg)s##eRe*E2iMn_8|4GA(kWyKS%ST9mK8~bLw4P$G -$P#Dcy`5#wah=mU`=JCmE2d`-gmL#_szv0R&kwsM0o0BY!|1GF*Wx=F{LY82YH+9`8s_@$(acouv)k1G -v>q*rjR?E1yZM5a0gE1}V5(x@lCC|%mYK07>3bsO4NJFYzvoOA}pPZO8{RnDd(AcHEL5zC5wH$^Bbwx3s0F-p=@;c!^0}^Lg#Dw$QKbJ{Y$nlc(uCfq0bRB2At(hyVE|=8C~~}bouh-59gOJFTeK^f$I0 -eYFX3?YR>fPm-+4cj}K{t!AP(&JfqhSKgIR?yX%{G4>v(Yoe0y>8GNDCO--^vql3yGtcE7<-Y4{WXW8 -7nAB`4^N7e>HUM%Q}#+T`LKc<&s{9QD8ltf&BoUC@ab(#yyPj>8&qPSu8_i@EfP23nfm&VJ{>Yg43`ZOeES6w1Ba&% -?7nf>Dp->#vuW+w-|-V$iY8KIYvrXY!~swWSXT%{H)#$vgl)dcF6OOjQA<%*3x{}@{QJw-k*Izx9qeu -7A^|UfV+7o?ER?5g4l36Y!4l5Wt*Z;s%UiD>n;t^`_6vk6+eaA3@2rY+34#u3A>Rz2uwn5{wvL%M=exr7fD=SwFmTysYaeG{Ny6%jjw%%0WGbg#!DI!+DGSr*_X(B_-rzr8;D`? -JKEjxs_)fh*~~F1RB-;*NmZOCNXnC+9YB=^jIP5oMbB``H83jLW&0CV>}b(^Y;Wyl%O_!=Wx|lX#VGe -FtT{br3uXm^)`S!4|%vcSy;%s^fnm4i;6n04KmOpaObw(5NwPjHW;U*g|MlNV+p1$OxX*`_g231U-PH -Co(vQIXxj02J>Luk0;9O_yC(LbCSdtBOOKgABK7_v2L&jO?_dm4+{P4Y#*!-Q2#dGT@9X{@B@PX-INc -P#o-0fHOeG+`3z#>DrD^9rvL)$u!S=~V7t*;4;Di`FdR4JKHB=BpMbEXMQ$j5~t&XO2R4GVA -Jeda*r6d^-9EP)h>@6aWAK2ms3fSzB(M#1QHL003bD001BW003}la4%nWWo~3|axZXsXKiI}baO9XUu -|J&ZeL$6aCu!%yKciU4BYhWzk_y1G>rpA|)8Y;dycsesLY3wf43Eb}bCheH7B -5k4+DAgt9Lctm*o{(64@Ect9R$Og5?0@mfzynoqkwr}NjdZ9(C<={C>Vk#z;HIBds&yLQ-r=;cLgt|4 -3U*k5a9i&FMkjLzd#bCvBFKRiSB`rpLF;2DXPHD>U+@K}cV>p=FwL4Ad+>fOj*(zm!P)h>@6aWAK2ms -3fSzE_rd?u9w006cI0018V003}la4%nWWo~3|axZXsXKiI}baO9bZ*FsMY-KKRd8Jg_Zqq;zefL+4R; -uhYX|U5aQeE`{N)eS1APP?*N;dW+Sz_--_L{~;@a^oycI*&TB=*H~IcH{eX4Zrd2h`b?s}#A+hmqj9N -YEI69URA4--s-w5+ad^3}dJWAPTuH1M0ok- -^V<)f9)>nd!Xo53hVbg)we`QdzrB0^a0g3~L5z~JBp*6N-NL~X(RbHjXtW^%P`8csXO$sxEtHrf;!#VBq(Cw-{catS(v=zw)rhF1+>&WYrcR=V*- -2u?+oZB;mC!0pTC28_SJY!A)ZSJl)=<~;xMouIz;Sk}jC0>HJE0md08=6`4o__BwR%!{Dy+`DmGL%Kep}V#f+yL1V -dPbOSupFqzw3ElE4P>h7)gbjb8kF*!EzK+UsFYbJBB_2D)n8yf*S`529+_MNwe_d~@eB3IMT(#!utgy -;4@0LJvBlZLb*RLe8o;ShPmQb_bl|ofni;uyONXGoQTsNAeou_AT9EoG>heV*^(Ie^&X%U=Nf@Xh-B| -f72#wGB9A@WO4lTEr@aQH`A~%k!n5EXlAJ!9M0MmpbIhL0(X{pXM926VQW3juvrK=bxn ->pJhH3WKb%);sYF;TlDO?O-6lx<0T1TQ@B2SygO05nx+}BsU9+?zfM6_?5MX -=2BTAKa$@^vh(?T2J~wV+X`!=lR}gRmAm^p{mtCv51F?mg4uBcqH0QXZuE_iW5L_ffQLxDeVWPUpu~1 -(Qn>7naDI07`Skqk{1a>r&i7Y{4+eP^gGPjlhx>0of3Fh{Q;H00jVXLxU-|d%x0ly<>uV^aqR@tmQYW -)H)iaw>CcvAfJTRY&jDaOo{Hs0_zb|9RbE18jRtImn~WryvNjTV0LVsxR}v -$QeB|KnngvzSj{VBL|Jqbio5X*u#DlkkFj8^a#P0GB84rrwlPFpJb75gWBXiY&e+Cq=h17&qkZ<*FH2 -G;E4kL-;~z^hnPXVD8*l8!!XJsMHgxObScS*XoV}PmSh55Bo(hApeGSMdTe~#MrVnfh8Gl34`BDbq-%EhR9fA8n4>+5cV(9unOc -}U>?KSBpG{hn!!thf;wFKH)vbPYO4L`PapBb|!i)&p8I0)?9^qyjGK2k`v*`<0|XQR000O8%K%whWi@i;;|2f#CJ_JtApigXaA|NaUv_0~WN&gWaCv -8KWo~qHFJ^CYZDDkDWpXZXd6idNZ`(E$e)q2++y^;@EfgI#48}4b%^I{o*A7dHVJHHNmMDu0MQS9K)O -+Y}-#LdAB_;0A`N6O$p4)f6bJ23STrl`_$3rW$cX@4W%e@ztj=wzlEa?0dcY>v>tW?~&JUrT#raM}BX -#AF44nNt;7cYL;ynONUd$vC^-w5_wG|0T*?V33u*u}@2Hy?h#&Vw2yOQD_EvRBu?gy-+xUcP>R{Tl7E -N_Z|6*jz#Y3)yzYdRCfBly0%8t!Y^&yL@*qbY*OwqU*g>EZ#l{zuOz5cBb?HA>#Z|2{#p@limKf)4Ua -n#cs!y+U?jm`;=V`kYu-u#i9~5+bLu2`Ys5q==*F-5z3lLC$;ljm+=Y2>1}+BA1i!s^;4r!&th_iPQW -Imf+atKu4T;frP24o0_dE1b0<*B0{rj?((u{G3?^a6Tw{7wv4-CZ21i!zdk*a7R=CdS3I(?0=(hNhfL7wOu=zXuE`p9^Fglj!jQ8H8^X-;V3L;utZ8rB@^ -s(1-uKOPnwNwI)d{yR8)3cIm)*)4$*J{Vg85{EuAeZT>7Zu`;$%Sgk8Qe!QFY*~uV -<5j4!J$JGk)JBU%LCG$P#1p_h5CCoT4tRtf?ICBFVT2n>hO=wB;`%*dpQKm7Q{F46*Jb>*Jj>a8<55+ -G1uODHjt-3mg>9wgDp4hmp4GilQ^$l-j-no$1-NK>-}N3mTH^gBt4v84t(nY4o}@Wo)pkmFflRA6fE- -l42A>82q?~l{S2&${;^`=@!ST8Vf>OtDrWI3|$Ea(roH!dI@Y2t)eOsrnpk5hd4&cRz=py{$ekc6Vq9i>me?Q1!UrA1$^8-Wj -sK;d`Db{`;cD=TPLQ)sd25~Q)C(Dc9`&$jFea!jBSU1pqJBQ8zfa=n3mS@FFJ4}%LNb52e -rWfx3Ih|Sc<4GYBXwfwsm3_Q|a))E;=)QN=}GsnWT7_`g_hh`Kf#p_U#y`xrI_qyB`IgW{RJB^C7B6t -~FAC9Tp6K>WTZ{kO`^tz`H@qWD&qdXie3ZB5?s{C+BsO5Q%TXdf85D*d_9TF3zVJjt$jwzoN7Q~v$+mzv!G(SnXQtm89oK -+D{9ws!O)A(*~}arqPgtR?bJda2@K~TLmg&lbx9h)o+|)4ud+iA*rJ`X;T9ItY?6B-7L`TAyAD`~6-2 -cg_a6C=V4Dw#iRYwjCjQvvtoQZi$L!lenl#6Z;+fr^>ll`oq+(?GxPihy7r73MJkPSP5S`{yD|}}XvYS^HLq3QipwiW*WxmeGvg|H=UaXLC -Q>g1p^C^+33V%D>4t%%+~OgRTZ(wJ^L( -1p@{bi>EK&@M!TbP)h>@6aWAK2ms3fSzA{MhFpmy000$=0012T003}la4%nWWo~3|axZXsXKiI}baO9 -kWq4(BE^vA6J!x+nN0#6HD=HQYHXVwVoC5>yCigOr -?Pcd^(8;!y0WSFc{ZdUqA?-@kuP!L!|6ZggH((@fv!lD_)S;JHWNeIHj^Q4_Vw;;PcslwZtN#dcSwms -j;nJvATITp1r8b|cU1dJj=|yUliU{;ry+ -AL+xv-j(ToscY#xU+(643_}n1t&0yjH$9(~Wl>JlTYXbc)N7y+{PFjzwAR%&Ug?RNzs1&XciVjsn>x* -WRF*}N&5JES#J|$Kc7J4PC4apqdT-(ndLHL-wyWYAG5N8nn-T`TY;;x46TK>ueM6V}Qr`ePW!D1!#sT -i#o6qAco6psmIv>42ju}nV=x2Oy|9W;)13aCC-~OO8eHqtjekorOrt&bTr$t@^$*^^rud*i5{1NDv(z -naqWmz=a3O_d$fWi;;Ra&*6C3+n<8FjfX%7ofyX|Cs`9$nnKCm&%BrJk;e%{I++G@3sA%lmvZhAHYC= -mXH_n|O;8jh3VLH+ud4X1Rp_)--y(KK|2a91P13HC{vGwSIr{{=_u?>hWM}Xhp4|mutP$hoGCbr{C^r -89+V#7NG(G=0YIh5u`*_!`iGiy1pusD2q4CBvv;wbuv8xirgcqplFd^!e)(XVBPVI2E8w$<}ECPg8nh -LjhnjI#Fyy`*kz|yMgCEjARrbC8FG#*y;vw75%yeO7X*h9vEN@`rK>9yXV>wrg5?4>&eb|A;#z%-vjz -o~L4rQO#MFGA>+AVE+N@yH=3gVk^$1zCLyVqAgfxPTU`}Rmhghq*M*9c1Xi#Bp7hpdHb~T_%x;erY~l()&tVLm^mIFj>?Ym`>)Z{?xxR -@v+pKT->8pP)8nudZwZ&}&iV7fC$j!)vAbhI2Y*w&ary!;@vL5Ug)2ak*c$ugFlnH@J10qXUU&VO>OH -r&qZNf~%+(Fh=3!3IaiMT5bCW{3OmTvQGv9N;R4SoY4W7H}Hvo68n!9PGMpgO7)X=8+*=G&%Lk%;J-d -RsPnJoRA%ejVV9bI$POTByhlJPsSk9*~MN%;^3jn;jib2EF02)yRu)O>DTMH*kmt|b;yeXuxJ%N37V#o_C-6imAAW8{3sv@ZV8 -XzR1eajX&0)}pY#z=fstf?RI*92Gis@cT(F%m9b!bE`sH`I$-pQ%?xt}U1+3N37e0m8NfGrP&te>AW> -2?#Y%zK&&qB5p}lRumr$jArW);GGx~RA9a7u -6P!pnPEdISB`xu8w8q1`0BV0t>C&8PSV1B9pTyFc7|b^ysZx|hq70Y^aJqS6#{|%zmp`V)4l(u; -z&^2{-~If;ZhT7~v6o|9#z!C+(>!H)_O9YJ@*^4s64=lUei3uiH@Zd!KoP&$!k|tjlNrbfi0imgq|_KtV0X1aK4a5}c8O@3pEdAA`cWGb2-Kmj>gthLQP{EqMo3cRzoOht4J_ -It;r0mUt}b=1OZ0pYHmuk+QGwwy7-Opic(P-Lv<4@r(%E{V9zGNk%=5~Ohl3B{;u&wCoIjwdDwk=ED# -TiFK&uaWhn_yXylgmzVZl73%~Lf?w$w;HuhSC#6~hkH5ZH{@xr2v>0Pb9LUM(D=qcQp%R=~Ef)imldV -<=X}N=#R_w}G9dU_n_E(KBYASSpg*ke0ZIW`V_16-~LK01kQyJMfZo_x|>TmIoV92&6Y+>Or`dX{PaZ3l410>iUtr2>%(HaTvA~b&l$b92;UpR_w+A^d$dtYfYPQEUFJ2jWm% -gV=Q4$F;GgdXPjsxe`2vXnVzuowOE|u$0_-xX)Q1+(`4c-4GV(YQKUTzWu1Xt0C>lh3*&)yTaxeuoo> -80fop-txD$M5{96kw-E4G%QNU~`u>jA~a3DPM+t-h8t@_1)=%N#9t|iobUB=sr`OIhl(3zeVM;ZclMQ -L)05M_V7K>F@d7kJtE`hs{7mcN0)`%ifIH?dN@#sT0e?iL$OZ0r*4+89w<$s`iz$jM=B -{}#9YC%%p;gi-QUZx6NS6H&N3$Y@LQkFm5dQOyWQIwGhBi;{%A7uA;x&H~K*Ocpo0=#O_RIe1Jn2eu8 -U?zai1`Kxi?)%p-#hfoU;CcMhkAqDuO8w%Xrkoncm6jkd^46)6Gox%eMvr13BsIK*7;o^PvCewonn*_ -vj5}zZ?v?)%s(R_aS6b2mx1vOJ$e#51dM1&dx{ct;7tflxcf0hfjqtfcJbt_|CD -hGa^^Yprfg+oCS@>TRM`svM7s7Ln5!$jt$iT4^16TiJp;Zi6u6r>S2MIFvXPRYoqM5mpAi_&7?zb_k{ -^09A*b8JB5`_E*W1~nM*?KJ);fWmUxDT-%F9P!kWee#fq4Yu>eUC5`N-+yYjj#M%UNHp{Te$ji_OH51 -mu2@nCvGrZhhclup;BKuFz(P08H2J@4*B6oo71d4i{6|6lP3V3fSpcywW|aV4fxL?}r|WfJ7xy6iuGM -?&auUJ2L3a!C(cvh~()>I;2KH_G1ajhPv)kD -zN4TrhKTfEl;Z7ReDh~@hn9S_RwA86TWXQaPzxxyLvP^d*T_9zqd%@20%AJ{l!->H*f0GzwPrjg&N3( -dNQtbJ?go+oCZ^7bOpFkhY#S=@@4#5TNi<^vkwR0UIc76XRL*{s&d8GEaHf<0n@rCtC93j=hy#`F~^x -!A;VpVyvs?OBf$q`|ayhLuWw#FqL{04hakhx!_wEVd`cu!|bJn64`*$M7b0q%^PHKk&h=X=Cw< -8mQ)0VSF`10xMiL0?mq^CD5h&7r)@ryeZvm$fWgalpzVUp&5kU%!fpQ(bOAV+z~>)PLlf8m3bhPq^A -NoWfb&h+Yd-TvZCk{Sf~|ms=^E6&Mf~+7@Ot^LPPGa*?c3JDY=)eI-VH78?f#ADZ#d7PjEI#cs&D_*4 -eSm~3SkV;F`dW{QbE3c*cKIq`b|!NYxa<9gN!c;#lSr`DP7sO3G`aJkTN(eB9Cw*f|HBIZ@#N%H!&LJ -7}yVk=$rT^-Jp;M&4;;FU}>OIKl`#7{i>*+8<;vFbK2Grwd(95`nRuEi{lX-Q^&dy{u|M=F?97rl?8B -{yvo@$iCr>9r16hrSW%V=X?aPTdKKre3&r&EQ!oTX60z|b7X2D5IF>kz1yxH@5Cd6`_A{rH92`W+!H| -`Ddu8)Qw#)&u9-JKmfINq7y!wEMb9Ne?YFql+TS*oKx92$WTri)T>DVt&x@Y=K>OK2{WVHPp5ZI<3$Y ->}^6%K@*!fysUGD6saKiRjp3fKz174M#gb5W!5c-!v*ICvH%A=@}m7DH3}%qfYqPKvn$5V?;B -{Gk^A53Qs{0xKhr)ba2706L^W)ZE>Ozvd{Abv&>p -5892(=PS2~d$xQKA42m5@2e+FBu+ZAED9xoJdNw$m?v6%1cpOZ3R?SwI1TGwCUrIyP-WFuvyJx_jS)8 -KrK^mE9fT~=gofeT6|7_x#nHq3BM_WTcFZWdmuQITN9)ixRx>F_0ra?<;IGoaM!8x)ZxEX95U7;tCVD -SJeOE0eMFY-^7Zj$09);@#_{X}i)mDEK@kl1FV4FzKHjt{FfvXeUEoGoO}&1}7)UR*d{dR`#?DS&|(X -BfWe64nFZx}p%eLr3+cLW$+@^z{}V6Bq^tauL35Wy?4~i6*4PCtnpD&L&`^uAyuPZjD3o179<$BI%;hQLw#vE6yC!spEPg6xG05G&;y9L`xyJXqcmM8NnFQNv5(VZmPmnXPJ5$c*tlQ&vxm -k|n7tW_nec@A-a=>J86OjiAC7@SLs*aD@GiWZg4<(9_3`iRL`w-~Gp-^}PnLhbW_OP-xS~Fy;l#$oMt -j22lwQ=E;Rg8d!eBdpXLsn)rUA+sp1kZZ!{|=~}_Gc5MG%ybRRYfn6+tcz?_9|oo5p -agHKcT!BXSdbHA5^yAw$q6~*5u>r}bZQpO3*$+AxIVb!!^jY;2C2tRQZlFVy5lf)s;0=_f;pi_#-S>& -Xjbw)3sR{I&T$i{6pw@C35Mb-joDF3@k$}Kw3a#_ZC&%)t0Biy0jYb#q%n*4H0P+)N@8-Rq?#lPh6zl -rF&d#s$w$R7hd7cN8565Fwao-YD)nrF9u^4z(p1sG2Z2~@O?-fSU$Mrh^D9c){YZ7CF~zJfjXY5cUaA -EL3M7e+ncA*txd`R#Wj*J`1s(uFFDyKwLFq&izxC8IH01Aqr?-)`3`~A6TM7nC#6`wzWVLpj9sL~PtU -k~ICl_VwO0nJ!mj1>t0vwni9EnLi^zWuu@Hs+o{fx@h{ASP54i0D{kt`z&Ml?~}SCcTXDYWg?Si@2}obcyx&~29N*2_hdX8!h8=>HtmhjgPkN^-y*E5wgArd~va>_;H -8##A|t;b|6yC{9h~s{oHi*l-C`7q- -q0J)WNQxWa8usn%^Di3j8pw9dPwbH@>%yB-e{%T(596#UM(o2~e%%Q&y_WC|6gbHO?Cqm$X*Q3 -(W^?~DQ?XuoB`1Ym+SKI`=63~66AYyQ|K17m*1UBXw~9p3SW7WoJu(vLm_vBv?A$R0vS?5X&yRj`_k5#;l6d9CZbKmpkdceg4X8XB6;HX>{R7-kxPmTgXXM%>XZ)_^!Vx~K}D117zBNU -|E;t5vg&dr^#92wwxkm5vdLoHW>n9O`w?rqK~&~7r&T!>;+=+Q6=ixhg^V7swp$=%$20yv)lGEVX79* --Fpu$+7`SKVgd5MCKJJ|=PvkDM!we1{w?oI0;fH=DiZYafXmd&fbI*c#)*~JPVc2Am9H^^XASPrVexe -Rvl)sMJ^r06$RKOS%tu#}@{)riOt1>3;uF!WF|7Lb^8=Fr!m&z(8z -=TcyA;)tpTzQSt_+()QtVT%oFQW;gcGgl}*uw5u#G6}T*l<>H)aM)fE5q&-2sc -zraLDxZ{~OoB#*(?_ZaN5!ZYe`jf%B1B -9j3$pvnu(a{KdOUC9v|8%YF>J12OSpdi6%;2_-6$b4Vp?qND0_#L!f6{0Bip~e)5`;V_U^+2tzX%8KLRdp62f|%~N28Y(-Juts8wPbl%=*_!*-yH6ytO^uvF-YjYpCt)!2&5{vY5KD*FL*G$~lH7 -}c4D1*o43I=^uPC&lIDq|({zq1CrF7X-^TD~mzhRjf=K|OFo1Do*k!N;uKFOGXI+d!h#bq`q+R~o@QT -RAhpJ13@@0%k7W!ai38QiUDvDZUGoL;ozOIX(-FUM$JKHkqYof8KA=Hx9I?fmh@=U6j?jM*iH&y}X~u -b&HjpQ&;>HJ~AUAubc$x@Eqr+M&X>GR2fEMoUT#r?i}+Dp0CUN(t%kl03%h9EZJd;e9(x!~s+5_A@$( -ncO5O-2z=@H6{j)BHOPeGI|8RZQ5f@v`J|j^b~<12qfh|#pg?1UTUM4iPK$`iKbf)UjFEI`%t -K2qF^T0s;FN^G#XI=d-qeAZrfZB74l|d^(3m0SolguPupUc -g}<|#e9to0GZNX?Rsw7K~2dm&x|;tP-Iu1&{Wp0{od0`N9xsdIR%m$^7HXmpnZBNd}hr&30$sHFiEoA -0Juc!t-4Be8JDZ8RC7wq2_jgql1~oc5mXMpuLsbIhLX5)ZU`d1j6G+7^O`n~sZ=P&|TE=dQZXro(+A_tZ2k&phxGF#)^LuBTC{_=Eqq4yl6rJ -ZDU5Et&l`XX=fMvO!?Oj=TI032R&Jf>rb#d?j0QS7~R*LWhPs`<$7QfI|=KkK4x0@Jz+&)t&WwtZ;#+ -i&^(aZp2zYMO7#d|a*Y&r~hU50%Z$@G1gP2Vt<47&a`%xv405}V>2E6t7FPy5)NO -4R@r#mT}!@pc6b4MGGURVw%ir8VV10w8>4%}Nx^cigE(&8WIYL|W4`azzZC8bc5*dA@Cv!aP_Zhcl(A -^|++2~1DwKP_yzD6#`O0yk_Q>txU%VEOVnTd{89H_LWXVux=<@>`dQ?a9^4*KK&tE)$^-ONJlMoa5(D -seHJ&mtA!kf$GvZ7QXL#q!C_o;f08-foEibVzQubv3X|ky$_l1Q3jHU}%b8@RHFbRTElRqIpr7!^mQW5ncwIkDk3UBt*r -SSbR`&Ly{fO5y6zTETUF1?c*@0a@agf0)@p3m%3lXc>R`EDk?%HmiB{E9Q%^gDjw>Ki*M* -&4$>Fdd5oD45u_=-5v4!bygC($LVPkFnCE@4May^ -0ag1Sn((1h`2V`E9YAyh(;!If58i!bdfx_Z>e@?D=YxSwZ;FyXcSNU#Yuc&_A{6b`ouEVmyDH<;=QVw -1_7k`IL$&2c#=_a<4=aug8qV>+_DBIMh%_+8h6*ND?{Ozr51BE7On}udbx;d>!1491I;FoAA<{nG(YB -Kxp9&D2kfS^UYj12wYjJq-pNVh6?i`)k -#u`CTgb{f7(O06jO+jVX>B>^N*GhhRCv!u9UVo|}H4QY_~c%JodqTY%dT<*1_s_tEe6soM4v1QGY7WP -1j!4i#w(^rI6xmZQh&yEX^!dt66yc86ElJbdZK!2*s0V@dEi+=#as{1B@h7;7IF53W0i!UL;*1U$Rtt -l0+)8O3-G08I>=xbLdTVOSCHXaakM?q4w%&f|68IT+1+&l?P9ewE+=FuuGTylU}=`)i~92gS{@as&X5 -I%_l8;YCPv<}Jec>8yJm*MDxevMr!)e+>h_L(!4{d_J!po=K~iA*cob{WJ^f7`j>L0LMG1H~;D0BGr% -gk@`Dd{Yq{X*;>dSS!U!Z4^O{0vD8!+2=X<-qs2Y2#4ra)*RkBIZ;n{zUQz9)8vz=ORLNyX=b$vW4D5 -ZrXK9acYd(AZ0+S;cg));+_t7>vKx$)Z7(?dH3v;NSr+)pPvZ9NDhWeng?38i&&)!I?H=_KYmL=w(YV -IA)TvTVzcJ{9Uvt968Jo1ZF?E|=Ycd#yl`x&naJa8$OOqhE{@TlA2^S9RSGS)C>-CFF#UcbadouD -nBQ}d9KinsCLn8sy@}6YeKDR}#7h3szZ=O|+!R=rznRQP>g>zd8(-rDGf((8un0S6=K%2-u!4?6y|1L -*Lj@Ix8K=*&MS7mX%y=R4wcY0`#Lxj(7@Wc41@FA-ZiF5~-26Qc7pVba5{pDcS&tOG2>eDCh-adc*YG -%(>@o@oE_d?7s{@LHrHLcVt0nlMRvxCv;)iCmbqNY!Ic -zp-zVRMAlVymFk}qsYwk?&?)~8DGX6HrT`wf_)xRO5 -s5)@Fv<_ghAfa#r8Q83tu@HbgQ`QV!!tc$(D1&{wPlkW|^_+YOA9GR;JhJGvGzElECL>s`>8EF5&s=( -QAwu@d0&4Nc!65!OwdN#$Hahbt6FA$v84+g$)s=;PV3WBd)GkLL((#uZw}Dh&k%%4TXgbfJOZN5tWlorY!#U$XhYt$MCpz#lUbwwp^ -$Ied~I+{g_x9z>+u`w&K^U5JxV*%BrsU!PXI-L5gu{@vqX8m1>>q4S=kuZh)TUU*e!I~W+lAaS|#`HU -v?XJLzQr0!Y@TU_)qqtFeQ>zJULy@;6dK6D#9;LVnF;2#)~))U30eLMtg)qm=;xYr)~%zgrvaXSDRXz -O-pFR%1Wa{ugHSf;(&w_oYQtGsW853R0=zWr1dYbuj!U79dd9Tr}i(9;{^Wf}f1P-U*LNH_c|7wcmBT -Q5N9y{~|7iRfHcU@t46>m>Nvg77{G{!T8G*lTplWfR=N+%{p?spsdp?BMo$If9jNF`2t3!QGM&UQ15# -ra4}76nRl(FeKssXVtEnZsHsIDS$RhtPAgu&w7r5YBc4iutm%rET*^^Mjubq`N>3`Oix-~NqvH*Yr4@ ->{HNtkm-I`YW=hjmk%2+81@LqaKtvl)n|4ozDn~HOe*x -Sd#%u4P~i%n$~{~jjaVTY~8T-^pkB+Uxz@G3NM5t&w|!uvrwJ5}QvdePG@)TbSH -HU`}qeP}T=-pWsZ?>cR21UjzFS6JuJXrcH*s48OulE=(;R{caL4>8Lx#ac2yiXkrKz%Ix;=-@w1$B@6aWAK2ms3fSzG0}oo(;|008X+001BW003}la4%nWWo~3|axZXsXKiI}baO9lZ -)9a`X>MgMaCv=FVQbqk5dE%SabzE2cZo^Dx`C+|+AgrcXiL5AOPU!aJ|_Z8_9Ug$F!tMfmTE6)*9k1i -x_9qAoo+A~FrZO4TZzh=NLD$PLZLhT^L#LOd=5s)xu_oc9OI;jqe0pyQ!0vM0s$L$RO979bsKm6d=95c4)MFtw2UwUG3{&`nsO4Qb5ZHB!6)FCOHi~*Wjs@ -(hXPcDLvdhOXtM^S9zz{*Mt02)I`D%cCYg2h|Z@aEn8&jQdoBo_?xuct^tMez~y)vk9%+J&vbQkRbqO -rkZrT}Q1V3Rzo`e6qNNuXt+M87i$SkCg7LKd!uXI)z!{XNoQw%y5dYtoS&CJ7l=3fs&pfh&|0^mmdjw -ER0|XQR000O8%K%whe-4dud;$OfV+Q~L9smFUaA|NaUv_0~WN&gWaCv8KWo~qHFK}#iXK8LOaCxm$QE -%EX5Pr|EII#z)#4<*;X^X@I>!^mrpaLeXs;bB&2Uv5Q$Tow#^tbO~ry)Q~*J*w_-<`kv?z?lN(eQvym -ah_&M*ErAthltkhr{Dp=S`UiE)lpisFo~3-^B&)P?bR)1kPEUyQ+ju(f3s(PdCav>sdQV{dlY1yz_BqUdXk{eO=^0#3X3Mc6Ety$t41zpwv8-y66DR8uy*p_gik -(z(_Y1bGZU7m|mA!>RcMOJKwK1xxk}dt*_*6ZOeOBSf!R+?1g+rN1Jk_p?lV=U7RA;^c+kOrCF}{YHb -3zdNiTi>3n?K?Q}oDBDW$`dt{I$?Ls=}|CoQ9OlQ7>5j-N{Fn}+!uP#0sUJb^xfz>fa!?@7C*PsPnsx -$$fWT`TOFZ_gJF*B@m0hLjCn5DF;+Ve((pMz;Q7|*A-;mu??o>M=~qup$J*OTe@{(OEWbH8X!Z#KUj4 -W7+(BO=V<=q#XBk@7r`b_Y0J!aDWVN~N1Dlq|uEz-nY2nm`WHE^Lf1Va+rVRPLQ_LsO%`$TZ8>#d!l` --0uG{Su_~4P9oR`3=+3A^WVhGE^)dn##aNVbPp@t7cBZ!IQ=S3q1h-|680Sr$E#QC%BG{Bq1W& -yBBuHk#WsyTQy4LW@l$juL~5Mn*M`N$z!s^s0>d1@e6lE){^#e&uh7&c(T2PGi*jQ0-{a8zsnG3FtCY -IfSUV#-qN3jEbWirOHm01Th-(PGNLk!^s_uR(UsLsJiC2TE&Ouoq|#0-VKm>!JbDG+1v>vTMjTZL`%ENO}_oL4SK^NJ`d9a+#ym=gQ -Fc^4%zw&&@c&=3_`GJ2aN^MsEYyUj6{E$PE9ttV+F#~-}G#q%&WumzPbQPKS%6DKQ;mW>{kR^J16!`( -tg8mUoqvhTNONB<)6qJ@|phBA>+7{(mio;Py@K}Aq-DL9f_-=Cd6Ku|yn8POr8#GU{5h%{#;jic4|9o -16HcTqEJXbt|UlzaG{SWUS=N}hy18c{1B2vsbKpH%em4(!Z+uO?`!~empaJ5ZJw_2#MO!TqZ+7?+>ZYIl3rPRhb-QaVj@w=Ux&|nA&Zkb#~wyeLU}Rv7Rp;A3n`thRH(=U5b?ZlEuq-`S)_U94}WZ7 -I?qCe_A}R=O5;O%s)P(?Ingk2|Ul+@jVj$FJEtuiBw$sso7LPWH^yrpJWa!W-x?0XQ)maKTWdDE`clp -wp`M778>+^FapNHe+rS;!2;Z7E9UZm&1#E^S-Y|q5)sd;K=EW%g->~25fs2|2Gjl|MV_8v>ij!d8kG*6LVc5x8q*-9S(DlCPiE0N=6P1?^Y -=4eX?Q%1R;U&KK4JkZ>j?Bix!kjUY($8|-liivrCO9JM>hDF1YPlYZ#0e~aw@NV(9miWxsO+kB1|_LD ->MNVsjSoLW}n)SI%&eBL}GSm`O1c;ICeT-grdj6C97iw@D6iQ{ksA_b2uh19I^sO5+0VEayQDf@bBB` -DODFhv*%MoXtJdsLu1c+0ROo6;RkLEnD&uAr -6yL{mkWl1xKI%l_nhH$H{B~@^eTawlnUGFJ$8Q92uz2(rm)Sc7T?ozB`0vxS6!oPlE+dSQXP&AY!)um -2*Ol)=MBCREVp+lzPC>l?DGFfBec!g~WQ8R8WFN~`T<8~|PrO~+|jrPl;`3{*s;?CYdK9)5m2;IVhs} -#B}HG%Xg@CI{kPpDMXx#I=J?&h&&h%2vMXqqT`un;`v^_MnmAi~3GFY$JLbh-NGs*e$#Im_KMOf?%Ds -@uJ4E*HIfJE#cgs-vN4COUoxO;TQ0-g-P-qS=aPtZpHG_BZ=1QY-O00;of09jj$!VtEA1ONaO3;+Nh0001RX>c!Jc4c -m4Z*nhid1q~9Zgg`mb7Ns{Ze?;VaCy~MOK;pZ5WerPm>Mv47uYI>f%dSd130mZ6bO>Q>zs;0;!3j;V~ -P|=D)ypie|u+0z4noyJys`aUi{{p2X%3AF#~#M+1y3_``_N)f7sl^&RC+-3#nW(yP!5;uB -;~>X0utYgmZAOi;ufUVt4V-YA&NX+4>o6vYYX_2#A3Rk1AOhKLGJzDKTd!4DQBM`cDu+SzpRc`7=D#A4y -xgVAzme_bZNK>h$Pt%1%wOfT`SFlr}bT?NWF;keK2GrJ~4yNlkKNv}Rt|TSyD$}Y0wtX_iv15q7j^~S -3`E&UMo3p1g&ZH1tz)d*yRFNvwrLRa&FAKI=9vV3>tWRl>$mHr!bygU_koXW(P*FiWX+9W@hl*l|bV{5e`|;9&L*mw-EhEMZp$B%8fM*TEAb==Za-Xni -eKc)8&Yej?ACBM(G;z8!d9%G-?l7XOs=Qt+ke>8wif%*$A@~aSpC*7s)f*FPPIJT-+ISdQO|jv?M0yW -dffa#x+A@Klt*>5*3OfIg5%&3=B|qBkVv`82NEKET=AA55P2{4Jt|q;; -JX^HWK(VV9Lm5{?bthHw==}GJ6l)`fyuMePQ1K}l(z73lT>>X%InVkurPx6|EZ(ZjYM7A$Klk{*YBJu -V0eVZ`0>ti9q%2@3Q5BRFFS8EuSD{*Di~_dCC=92Gap6c^$ef&AKb` -%d;S?-lumZwQDY?iX; -@m+L>(R^@JvxV+$#z?4L7*++UZ08Z^v!A!^gEfsOw(*-$rqgP0h4%e;+XrJM34X+|lhYy_@RzwK&4tE -2#pqaqrbQ`uY^5w?}XvP7m<5f7K9(zE;+e~!?4bi`EvS4VX?8@AH^hShhYB8 -rGv}j)n)N|4wtYvQ_(d1C#5O9X;EXEu1<}Jgfr6PyN#da>>p4|0|XQR000O8%K%wh&Pxrm%?JPhEgt{ -?ApigXaA|NaUv_0~WN&gWaCv8KWo~qHFLQ8cX>NFBcyumsdEHrUZ{s!+{_bBv7+6qBz1lSQg2Ff-u(x -q~3naV6CcPfu7+OZ6V|8UoC(?O2!~N|y!xxF75={nMC@+<)l`IP+a -`drUnzJE9Wa6L!!B;}H2IoKRS7)IHqDr-r^vlxw5b-7XQOPD8SrYKV`x>J#W&G8LOARgSx{UhU#mzj{ -BXvJ!QJFJDYlIK~?6PZ15Evkbbp?IW>HI})MJdKrx_%V~~SY|Q@XGWu|kLN#L&d)!-BQruN9Yqn@l<7 -9-(#){H6u)8v|YiWSa*{3+95h*=Q3Ad -Vtj!1`#wu+SHtyCIdfn|wU3py%!eo$! -}{38^%!bd9C$9arqEqR<;y}!78;PnGg#L*FHR&>k&t991bEEAl}pAxPPHK1AX!1635)(?fdv1NL&F>X -OgTg^y3;E3X%C@Fz}wnge#e#l|gQd|IDL(RxD8PMB;hQP7qwodqrx*89#mjak(&q6O%fX4H*=0%$a2am!_4#?u~n*vBAy -W8fhCRQ42U&|b(#^H4XMGUBAyD5M?YOac!5>kXXQTI)i6~GZin)6hbzf5S+Mp=O8`=KoOIMgfQ_?h6E -mKe#i;tNt)YeO+pWC=*c9LrFZRZ;u^5i9GcX$*|(@rG3u#MV|ORa35yUT>EHJ*7INv2vnOptE4HY(oU -G%(*v)C!>=(#$5$ESlc=eyWoISgu(8ZbLK1hd525)dmWG`b4mM`(a)+hAaYhv!;%08t(W -Js0!QsNW1-o9vC@!JXPSlM@6UPJ|!KA;qbrV$338B%?WmW|3a0sbdT3M$SD<8+!Ea>sng{?+BFBe?;>Fzw&BtV8{0Ct(6O0xwmE6tnOIXxUjrP(y<96D8}<|| -;6A9!$Az9=h6-vhfbi030DyJ`JTP=G1yd!5%Uz$nM0%_NtTCt!P7D=!umpm*P{*#B4cq`>mo@Ne64&= -ZZ*qbjpdzmFEJ26kqO-}Kr{Q5K)I>iRc+V}o0}=_zVBh3}YM}G*8 -;;+*Qd=>rrtRzXcvj$3g7k+gCf9j(*8y6W2lUgZ(9aj$LD-^@HFj#<+22-SPnTZqX=7l*(7ZnhGjXi< -6UTv;o)D}u_Wz5<8;^3kovG?AVSN;(1yZ89e6EB%M%@A5p+Rr<%5U1( -|X#ps1S_A=E8^BZvF7smcsfCNm#Dd;fK%G1L7S&7kXgXDCj;fTR7Oa08KHs&^ -f-EHz6A^o(WYU1|Il5a$JqCG9fSl+f+n9cUsfltCDNrYF*9_neHS$U?sx&DhTmyKmMSR7euE?Y+vEOB -2D5LlYqh3>X{@k1EGf?^&P`BNIJUx-8;6LZvd~U`-jUE{^aPgiemZm=Ic+&I46 -1-2^wuRFp2UVr8;Xz{OoKtDYv9E>#|eE`lajDHoO4_wD_{I68n>^|(h5dK;sm^c5&XMy=fWK%$W#rhi -lHS`fKfPAyW;(g#fco%=-Swx0k7y8`vT^uhD;PQV1Z7!6WZ+(Cz7kd*EnxhYJgr^4$%C-tXw{a7GVWi -^bJ4isC?;`*z=SyKY`VUY`0|XQR000O8%K%wh?Pqs``UU_16d3>j9RL6TaA|NaUv_0~WN&gWaCv8KWo -~qHFLQKxY-KKRd6igOZ`(E$e)q57;s@KCW4K9^6bAZ`F5OmO>xw3O7>q!XCCcVPiyB46F@o&3?;Ktv> -f%e(pf-KJ^PLM1k4T=Job(7h*><_6smdq0+H$IY`mgr%^lzWSoJB-qh*y}SN;bN&4$xcmFt=i6SdPi|-u7F>~am}M+o5|N -NF&DrzwGf{$k|Me4z1Q!y_1^7F_$Vc*`{p^!Cl#lxB0A&jvMvsFT8BFyv7*`@mV-#E^O~F!9ni?Gx`l -2o6Tx^WmRYGgB1uv+xbd~s$b-0A+LuainjJ9EEtS(u~aElUpf&Z+ePf{V*A@}U+DJ=uN&`-@V4Y1)Rt -#FOl3pz=<09EKo(v9#FkR)xnmj)#0f>g;sBffmIVv0KJTJ#E);#EjYGfAUR(U|mIwU{!&qDkq-j7S=; -Y2FJexUM0K>?)&vt&Xa}*b^h#CfJo$Q(PvvDls3jkYZDcZLXUXF2QBB`GBmD885{y*FFN4Hg|HJXkW- -p5txcAtFNo(yof5aUxYrbmz1gueLyWOekuB(s`gs-Sj(ZGpQ-9pz(PuVb%H~?<9-JBsdL<4+(z!^miqufiXFm#O4t_;TRp#23B -!y{7HMC28KApV-<6yl%Xe$FLW(#reX7?>IaB!Lq0X0I3%Qk`M0iar9${Q5thbZRLd-5YWBR}cEJAZJd -2fY@*&Jsr05uhc8)OV`E9EDQl8>Uvn!BZ#)BZAe?ltdwq2BYJ152Z}uW&TUZb11sxWFrMMWYcTSgp(I -k`a!;)R5cNpf`JLa>ZwLAuf2M)Vs-B;z*e3Cqgx0b?_ts@z`jP_9Z8sqqba9HQZg073E&Xr+1tE)1L?7T7N8Jp!-U~}{xsd|cV(E|Y#ZkiCkFa~HbeY|;={I -HPQK=rdBk9J@^+z}RE=J#9{n>5L?y8#>7RxHE_cg;#a%5l?(T>n8V3&>Ey!Gj6hv->$L6)|fwewQb9B -34rNovW0Kb-e*9co?_l~E!15cbFSqXy2k>N@VMh4cEPhdXcKY$%S<1-r{hY+I#qg$=43X`4)fmy>Li%3&)THJ~@FfUi||;4D{m(PL-uT7snvn -Li8VA@nJ)_rbnyIaRmU@733dzwf_)0l7VO*>w+kReE7vRMH|PC1PfUP$7dy2gx2E?~rRdy|&0L;;InK -?#)KDHj_mqYzKdsen|35nkD>wi&SHrrxv-y|cw{u?-8lF<$TBAcVo`hrDO@7Q@ORO!iY9>fj5j3atE=ZGD{Vt3M^}9!gN`LOM0p -&E@6?&hB=Js5ippg99RPZeU5k$WAyrVq4z&fO9KQH000080LuVbTU8^;g@FeE08$tL0384T0B~t=FJE -?LZe(wAFK~HhZDnqBb1!snYh`XOaCxm*TW{mI6@J&RATkehTCXrW!1lqw43OJw2F)Zv>~>KM1`aLJGH -ZzxL@J3?Y4g6wF?jGvY;TM^z4~-~xW#(=v5pN|Aw3 -z95&&Yxn%_+ttne@`h|xLvn7IDB;dCuqH{el4(uWD^z3!O3qAMm;8L;3}p@6L_U*yCAm{F=UbxdQka? -M<<{fyvuU>g{PDBjkSgb}gRp+@btg*ho(K)!#!5a4@q$uT*Iec^1fB;Ey-@s0opwV7(uTkAM)TS6P{U -2zNPKwIf4vq`nDyG4FeiUvO`!*td;sx8a*p9487oVo+nP6NcdQ5E@l>}P7hE*ez|DheG9|&;QukNAuy -Syb*9{+`JK!OK0L=n}QT7avA;Mk=VligxHPR`-0_?}4h!mO>U -KH%jUlafX-VoY#Yk=A`14`hQ=SSSN(B15-bcWz91Id>ItHWX0dm@fMllBKcz{4A(WwFpY61J?UE%DdD+fzx`hCBJUdS)y_fQF-d})7kX{uwe}TYX7KOaDdm}e=DYT -jWAUnKIhMEQXN!G)#=^>o7l8UcKilKa&%Q77H=&L5(}rdHS7_WC+*H*Nf{lPyct}YLY2_s%&%6Ne$ZH -xX(bdTS=#A0VUNFenWbM%-Bd(N$4rfqd=O!V6?`*)@#Ukkw_sf?rW2rOh+?rmic2jbCER_LK(*X6p!0 -@1p&!k)*zIO$V4?BbGPwi43*~ZF-v;5#^D7)pn3#AM)WXEK|^Z(%tXXX1o2>dUQb6x$M>G_}vMjCr}1 -QSD_JYMzFqn~Py8^|?KqC$%3)XiJaTn1P8BfinvNoj76oDX@1+X-gvcUF(4PM`h29=%h*C<$c?>}jzR -z^zxyC2_MMJU}LZkZ)UZP%VKb3yi}Cj`z(8w0s8i$KKIG5I|znN>_uoIPSp)6eTdJ!WsaXBQ+!xm)QZ -z1kd(vEQQHIWXy2v@h!CSWVyPyzrJ|7Ua$V~c(Vjs|8=Cnf=qoVoqM`Kf0%|@9DX0WPZUgFJtV3dhdPhOtG#+VN&pXJs&KBQirl5511$rFRddtwU{dAv=fY%&i+}KvK2!Qzrh)pfxCAya*;G^bku1 -++i`+5X?f=Q%#;bWudz8L{Aw<4C)z3JS8%DorrPDjM07u{TXs-t8>;AJaHBVtseauWeYOsAWA2)x34i -qlMH0Wcmv3sgfO#826FA2B8$0&k>&J{7y0N)Z6xGG34Q;DLMy{EAaZ3`thK;^6w{eV<=YUrV#Ng0s5C?1-&rFVw9c4&UShygcF>nB3jDoP8#c6`z7$d -5H{YigN1%uTPb2eEb;;$M;biUU9Nsm}YZgf@#O0kt*H))XsrC~62y(IAino|V%|hTM3};4U`!Sa4O$f -Z&v1qVqU&FiUhnJ#4yT(N-FGoO?#o43tF|du3LVpT@Q~ic&9s+N9!}G}XRPRFIHwlK+>Lij>ix4=K3{ -<~`z$aetcx0LoxU}z-F=Kum0T0#YR6_*nzS!FVa+*DTNwK)}GFrhAw+Vq?OT}UX1Fg0p@HT2Q0(T)09 -bQCKrIm;R(>oQW1r=ooe8}P>4Xsq?p6>l`s!Db3qr;_tL_{6EiW($K7q1HX!pz5v{gkP{rI*{@@u`#;I)-95$ -q+xDb?x@zxnSf;o+nWO-hkAOE-y^B%2;A3EXhNVR_0^Y(aSfrTKVmAUpFM{5FI_?!>1bk>s-JWuT7HR -~9aiOYXzeevC*3KaNGN8kj!tzXl^l7C0>YJNYkAO9KQH000080LuVbTM -rfW*}zNy0O)XPX>MmOaCz;$YqK85k>~fGpW=*U2 -f$r{uBxuS(U$G8cEURm8ZBcuig1XK>AI5GApr&ejv&k0&+e*Uc4a*WJd{X^qGo0V#L2Epc6HsKy8JWq -M?d<}J9q8h7r*`F^YZ-ltM|Tq{_NH3m(QMmLhyh4%D;CA{}7f}cQ5K)`0TU0Up{}9UgUC@J`FEJdR<; -h%v~xkUx#PU?^1YOK6&x-x9{=tFJF~>S6{rm`#ii(pSE=F8q1Z$lkl7HvfOcVE1G@n -1`NeZBvFdhx|?Uq1Wf)7S6c{cJezoH5%IXWZJ8?Qef@Rm|P%5FBKYl2MdH=I#s -XTvG-oN`lKKxG-|MS29+28%*!@r}5xwH!S?A3ek{HRs*J3mWzKfQbO^o=o3p1!g5>6^Va$5|h>s47t* -KmF2douNP0PHpNm_lVupMvvm2erXR=o_^W#8=vr~ow|U>5wjUQn}=@VX>7IBPHR1%@Q6=;X^gSf#5`Q --a>W!{BEz|eE8Bf}qRxP;bl-5f>z(G@Dcf_(_N7_&;|Y(m(}O}T#{!MeUdSc1;+b_Zov!;69!IA~#B= -ltSx@2#%5ma}D^ -y<0y%gu+q4^1q{{8c`cuwL&uZlBT=t2akoH2-}*UOnG%b7OlDV_kWcrr2h5D)KUK2|(Ig)}S7!5FBAM -hWW7!IHCJd(MLEISY%0!n^EHG8UEy?Nt#x84rGgp5h5m>LRRsadqN3puD+Cj<%k?u+!GzfSHO5)L5XZ -0#y~Lst}2lIttWLpmGAW5|VtDUJKiPTDz`iRs -=^fXkYCZg0tq$Z-&L|j+sAs*gY`N~GdA~g|ZoWz6AP!m!5eq@BilZbO8=p_3mX`Ms|Mpk4T#UNP4(I1i&DA6AhB}tSdNl8+MEAQeV5qd?EUXdtIss>7ZBubU(6`8VS%9bfer -XYDGZ+Yz#S(aR9a-C(0%d}&r9rMZ8^U0ghK@*fbW@e(Cgp1tg6vfl~NzNpkdp+bL$&>9|#Z&o~On!wv -T^8OiOMkz}999@)g}JvJ!hCGd6DqGT)0Tsrl@n%qTM4`#Je2nWU82w&CGcJhPMO#>iY-l#)zzY|86A@m3~_jG?U-j-)O&H8?^+s`40ujFRo$zvC*{95!*c2sVC{w7aukq)Kuc5ZPvmQiIL8mC`ghDXDYI -}_GGH`rAl9_wuq<4S1&Q180q7c9M=B3+bHkT?`r?$v2_6nDfOaVOlAjv2bq?k>Lf8dr8aXXsS2# -T4girXaQ8sni73x)^Ud<>T$q%CK1W2JPWH?Th4vT3aj~s6FuTi`E|9<1DnN=ri+{C&Q;S=gGH%XOY?- -c=!QokM7wRq+wg|)B_Li?EK<}y|#J{hp@3?4hKK`frWE8SkDBegm5rHwny$X3tO+Ja56KsN9mLS6&U5 -=OauutBf1UF9ub(?+5->Wy*==(Fo(Xz3W3#42sw0Gd4~mGqnGD^Ry>v%D>1$DLUcP6H+!k}z*G2UEO; -2HfnJ#R&`Z-Ec&N|z2%XZznHjVOrndILLkTPMOO>ou7M_hB;z^zMm3X5W3a^K$uS#l64%WT_X=AuIae -H*Ge2S;9oZ5gU#axKtx1~MUA2OY{2cEvHsb$-ti3r(6_CeVCY$m9dzL)G(+BTHE7F$2I&_g_Dr`coJm1O$JZM2A|%Eq((rQvE#bGeqNt}$36rN3>ugvOQipnJf1G+}dU^gaW34V<6 -UsnyzIhO6&|hjT=`_M>{amQh_+=is)}+tqcdYZTm_>e5NBIrZfdHBd|IW@&YN;oOps(W-X=uOxSH!(xyeeRqP+btZU-P%f81GPke3VsjXj;Zv{tg -VNuCoz89(D~6Q-Ox&pOU|E~f_iO>UXOC61Gii87^H!Y& -9!qaT++%Y`fJFcxW9SK?hw>&e+=Z$#9MqjB|A!b=1VPjm$?K3^DCl19uHPP7ORXP&-X?#^DO1S0v3+H -rg6le_gY6qgOQNFMw1C{gC{bO@#pw6DnnTKl|87^rz!xi -;IS<@_I@l%R3>RQyLuN1eb>lm&@8J8Ir)RpvQZ`HNTKuyu>+g{)HX}DI+*s9-KXU0}dX4Pa?O=g|Rl; -PSm!A%p~W`dh0xW(bxm2uaP>{(q}oriw-oTU%?y`Z%nlrc`mqUq?YSG31fqbrr>)WFkg;HiP3YG5?Om -DH8hr4v-u394FM1~;b$Zu)VZU8@a~vs}HrIz6pv&cW%lGc>0L>O#0WuYs%6z*Pe`&8JR!b<3m8no|Rd -W@#%LsO$c!tHH`m%W!I7)-34|t%uRfA6lYA*D}qifjS!Np_xC9$$4mCC>p3AI_HN@{m`lPb?#Q9cc-p ->2A=AQ!==-BJ%c6`HJfMDz%)6>BBT9IkC~+E8bwY^#W{MXYm1dDA1;hKc5MrPHBontz{8d&>bwF|30J -ds=M`9DxeDyGHQE+#yx!q1MqN$=uQZ$Mm>8Hk0&N|En)xaxo5M*UyKxD`354c_Kxq2s`a$SuxN6L(rp -}q=B@N{L+~0pC@8^T}b6=4>ZO_zCY#Z%Y9a%p0^UpeKc=G=~eG}KGZ*qqy)vf&<-O;aPt+CJrx{-@)(Tcklx!unv9)Kj<6_9HZ@hq40Ap&>jagKohiIg4=$r(x`)a -=mzBy+j6aMJ>T(s$Mc<)?{Lp=0diPi3wDJcG=~M(VnHz24gL*&00DM`e}jL6e}jL6e}{jEe}^BuLkC% -PN&&d7YYS0I!2P{FU_B@+_h{`ws^DhqN%RRA#R3ztC$$jxTp*1AQy`7NU3}XBLP;s|IZzfhQdLpXjVF8o0%@ve{vSc4@h6S5p3%M-h5~PDIR7&9kg-jOsxiILX49W|Y7b ->h3)(R{A*wK$2{n*iu9sStRj~)Hk(T^Sd*wK$2{n*iu9sStRj~)Hk(T}-tYwIxh>Bo+K?C8gie(dPSj -(+Uu$BusN7LX@xg9uGfO6KUcj&AEbDbZ6MJ=M`uoflTX8tw$y(T}D&>dr!e+tIJM0O$`;ci`z3uG7sz -i!E6YdYKpVDjwd=2hLdMFd4`i`IC%=}DX?cad5Z0UANn1B%HDU_jRMp0AQ3$Sh`UhZ0UCK~)H9g;NjhM7JR` -~jyz;;RJ>$wVt~}$)1HSTLV?4vkGpsxfsaknz=R>s}RxLX-_E8}iu+^vkem2tN+?pDU#3R2dreFkm6 -TR8` -v@LTu+d=v}6g&#xIpu`^grBLjJ%|PwR07d~=s1boKVT|XZQi_6yRqi!=K>?4OKvOl>&c(zrbJMFYp)m3;Y1T3aFn_;RpLuz@?Q6KVYZA -;3xo%+Yhn8lLRwF2d+*$^vk-4XA!KDc<5Jk6%XB|ZsMUY)LlF@Yn4G*X~lYq=fo4lLxa>Pp1yo(Q~** -@Vn9Fvm}n2;1;1CRY5^>?#{!iYDzz4gp~h;J1S$y#%B}Xgan=MVAR|`8j&-nO<=U1K9(qCr0@Vkcv%t -6aQs+BWC`4bYBoN;^#bO0z#mWZ%RSJ3WZ+qoUCjx@MI{6suTVb{u -6!>Sm}iSg#Uyev{VZ42lxa0;HN0i%k@_ptU!h_PZP3SqlV1Rr`>fk2*sng*hON(;jV{m?xG3xSfjIskohr+o`ynirc -9;kgs8<;&v)-r{Z=hZl~gQDh_mPqwoUxvPpb1JewYyWhJGHx0yZZr8mtzP!HM*Z*(RQkIr%HFKbe -AeU`Q}NYffoXa2KWQ~0sbHmJi;U4k?_8B36Y5+^t+ahRTB^8jJtTEfXp~gh-L#_wg|`!{1pJ)TE~H=@ -e@Pf6{z9!3>=a~CM5+xq&n`)4h^)5aM_9;7jJr$nY|rnMA?sX-t6_3I{Q^G1r)uJ`(EI39ZmzsB}w*i -a>i3#GdgQ@(ogF7c-9-|?2BfuuV62i)y%EqHZ-q#oS4^foSWA}^RbRY)4ax+X4W+Ik}YrJz&~$J)7-` -_XWnKlY~z45Z-=J2g>h}MrTKae>e4HSt>_I#otG5@-qT#$3>;tIml^2GaIJc8vFo+ZzR$qpP*<3NBlC -Nhfyt@M&%je%_Q9YOGP9T`qtL&EF%sU(}875+6DB@(k2TqKKCE;UNfDZva#S-& -+w~mcEMN+tooFn`y{{Z86-21V+WSF?>H8-^ika&;)&pjGeP}OXfa?@+oMldrmxw87sM|+xUiW%pKZJR -LI^HcW-NirB5=^=M8mAdTq&>beU%u1=7+OaUn_KH3ZQwpR_zbZoRwti0t~u?DLn=MIfl#@6(-S~*O1x -jInI%{$;nD>IBoA2FD_Bb@bt}fe8tHS;Pm#8latE1x7HjRh@6CyXCUSurGi7^pwhcV5**J -pOc8Ou@y|MM)io~;X=60Yz5bgoyhc^i0qvrsCIJla)RzRCr6K`2o0S`2%U#WA##Lto+8G`S>hQkF-Bq#c -RJ7NXS?+1gtpAj)asxDZOMr8B_cS$P}vxvMO2gPBL1Qx^w -kc}L?+!`YiJd`P;V{M%^Fo#aWD1a_0GR@0DM02jB -9rIrk?31RVsPD{v_>Z9?Pm}Vr;}L`>OnuK2V#G6E@vEPd7NPxIax3qM;T;9+N1yC8cv@Ko9Xu8ms-TmO#c -b$m~lKDBALFRw<3#UT^CVoG>E4s=HxT3H)^BA%f}oab8@EPI6ibDeC9ZMa&nB}det=Q#Ir&V@$^r-M7 -*3_I8KM02#vYSH?tg8_P*EAqvBbi5`!R^2;sL!uNzNCv0JamMmfXbm)>rN#Bhwc$=S+Q@KlNEXZLoO` -q&K%93v8h9AfTr?!zWfAdX!E5s%cw&PR8taF!2tm%Qy9nC-})9g@u9QGno)Pdiu~9m0I{e%mPWjz=J- -Pd4_DGoXWWt5fpXLGz7DQ=J?vo^rf&_KDt}lIF=t-YE&39KHofAlP2jK{6G(AhVVE+A{ZA5eILNoPN`TZPEMGN-tSUw*)pjYTM2U&Sh%g^>j1yrZhm5kzSzNq6}pQjKrgX5*C|6I9Q{l?9DO6(9uZF! -4@n>Wu(i=FK?uhg5su?5=9JC<*>@e`l+FLiX~lVnj9p)fwRO0}^zBchyLeckizM4g($d=lE2SCHgwk6r*-BZV&O(f12IJ~D -3w&M+E1#nW>xx*OVq7$@f*PCSQrU`rCy^{NCJfSFU5yhri$PTu#LczWemnL}zv$~L--2U)ggk;}7)ud -zIn>E+e5Zla@y1Pa#c>xE2raCPRA<;;*AFJ@J^PPOpJ&wb~#}=`B7j>cdB5X0=$eFu)gNf -GbA%Ge9e_+U{;sPwivI0>-^BFfyWF~fhjiw-8u5WnYJ -Detj7L*jC;dI1Zkg>zILh(l*mn{H$4LK3A24T$dR*V1}YoNSwQhuT5bCqsKI6-)7)8h2J!=&tb!Ls7p -nOjzy!V5Z<8zSI3gnHRa&+NDU;E)_J=lsw)BuAR7%) -rqa!Hfp6Xg>po+iQDb%Kw6lQ_L0KRt}I2x!OHOy1hpaPj8%{0pp!;7=PQ9m?KOwdAex#rj7! -1cAw-V7YV879p%(*VtyJOlH9?)D}62&_Za#Z>K`b@t9W@YH=|kHcZ3;w0vnfht{PR6xb&1TtkY-O^KTh5E?E`vg3R%*D*PdO))iuS_P7OR -~pvug+I8)CmLT0AOl~IW;qp$bA}9>*NZ8q-~#xL6z=8<$5i2p6*8CZr;)tsNV==4N;)5B{OI^Lp+25s@s6YcqW9es1Fqo?e%7{^| -(5q8WRA7NBYoSbc;T!u?;Xv!-^HUfET8<(O@AXnqeB4JxN*J1mUbP`Z-;CAJa%lq3>8t@x(NtyuR@x-U -;eO~O=Zjr(qSI46z`<$|4@Wu`HBL2&(xlW={<5e1w@-fc^i7|Y-?VbQ=|4elI8nA7YqlK-6mYZw9Bs$ -|w;Yy3Ws6cm%bBu8C!w|QTlg*f+(KFuhxUQpf%5P>__>|z>!eQMckny-9sC>?x8Hv#{9Kd(Rce$9;Sk -$$1!g@cZ4c<0<7CU(tmRJ3qH7Zvld)VZ*_D*h{bxD(wp>_Q&hIT(FP5v9_L~fae}jL6e}kW!Ez8lb-Q -Y(>TwmsO3jYqj=v1_Te=Xo&3;5T{xwqxq8<16_n*>pWLy+#^g00*fSgw1lT=!V+3aw~kwCJWo1x$Yk3 -}0~R=*+VyWwhLe*dVWPK5jt=TQuicxf!v*gWHUZhvRVzKG;SPZ-{&=3dWuk(X3|y0$ViWSrEXM+iuG> -q~#jYf&jK4fGr4MD;=?r%R(->7Ps7Ww>cV0%3PFNF3K%;a<;-*r99xcE%)J;%R1YSk@~oX0$-<(JNme -zk30IfqmMiKxTB9d`naQyJNmezk30IfqmMiKxTB9d`naQyJNmezk30IfqmMiKxTB9d`nVGmMk -8)=-EzGoI0-YIqz?j1nJt&3upicJza01)1lI@ZTxY|&BVer%ECR=BE5;zce99I*ah_XBDZD>wVgJWtIVdYk)T|5`mP6XpZQ55>clZ<$gr)(!Fq9)EVlb8}xt*1V2Ih#ZTdGtXTLh{M_dh3%`Zm!p}uf7 -gO#Oeh0sU-@(r%V{!01_!syW_!syW_!s!m2U!;Q7x+E=+-(&PzlWde#V()RDf}M(75)`|E}shZSh#L1 --0u|>Jr)!_cH!er;osorma}Z|gET5!!<7wwZsGcpu~YbW__=T^JN!HR=+H!`tmXdLa(`@5DQmetwp<@ -uG|Dy#{|WyI{|Wy|I5%V^5FQ8*q{DSy;W`fuvyH;fRcVRvgYhmA{s=!ErLZ0=T%8uKPD_Fx%(vjON`g -PZ&&sVN_%r+&{tQ1?tR=&r;pcj?WcUmGC}}SR{sMo2pKH^?yg=E7#uDi-*hUiu^@I{-UlD0@Ef;o2+2PgII!S@oNt}askvsbx`2p+*>(tB#jawSo{OT{Q=_s0C9f+!8`cq3I7TI2|stO>4YC -Oph+NmC;S2a0Dpj=>sWv(jlv(`=VmrV_#^xgelBTKgg?R`;pe_KCHNEk34X3_Q-VLipWx>fH)Z%U{26 -{ObW?^u!=K^L@E7l(2c#~_S@)fYL%^OSM1}5D=#EIieh8i&9x8FC`gW>sm+CwD%t>ki&H!g1wE%yBKZwHF04tZX?Ex2 -e$a_RvsFTY~y-b{kcx^?idMBrvK0i>`W*%Req;~;_UedK)phDN@ai5;g(?(}xbp<~nDfCKP5uQ5Epyna!f=Ym8=dHLB>wOhq?FHR9+kZeA4}zD{uXYE -`q&qNQ9H{m?RX@C*fAdDZG{PEFIR-X1?8^5dj4uX?hT*F$qWRSK<&^k4HV+c<{Jo1RtY%_;DGnP_-Lu -5S~$zKxTwyoLGtP;ipBG7}l%Q%`@z^j+O~TzFpI6gB;4;vF$RaB})}iur*Px#E*!gJ1+fD^@nV*y|8v>t%R@Ahtt -6UJpqCtb!ahf=tEm1+PcX7b83CC>B~GS&pb!i1EUT0x3;mivl_5D0u!zUZcRGqkN3Kif)Vc0MRK(?S# -&UB;`l)&F~v3f-C(&ywn1}kBtg2l+9fdS$KXAO_nlcK9HqOGKI=uH`?QX3S^T)WeSxg1DOnDvE-6DsS -(hWL4CN%Vx}Y`UtQ$`tm=&DY)7|Ke*k#|VQ1K4m_;JkW0)1OM1x4BVJT#>4b-OjwI30mPfctN>9e2FG@>;* -L{y3++iCGvMb;2RY=3)UY2@wvb+df0In+@jMwfCLK_nK8s4+XMNGA_~w8I?;>Zt-xdTx{}{Xqc@z!jG -s|+sB7L;n>Yja8K^h=Lx*+O3aS_KeqP<{eSN&S#8)M*w9OSMY?NdPKd~dv*m}%*BuO -+5##W*IaQ%k2>n-UgBC+%a>iF22O3L>kzDKuJQJ!xR2&^>WVWksY~Ap8m7ybu3mDu;^+svqIsIrs%u0 -av&j!_a>XRspooQ_+6--2HTqNqrwR&4n#UAW5vo>2-;aqaA2&DFs*F_Sl$ujH&ym|QH5T`i09;$utHi -C28@X&%cU-nZcU*Jn>a0~kUnG=kk>KSB+N19vuq!lK>~pYrGezuoFv>R(6Wt^UHQ^?*MvG`z;m4vq(5 -@lp*DC=2l^^O=eyCS|sMqAhgF+4tOa)ro9tcj!uRdbU?I#IoJ0$MgLoDF&1G#Z5x&c;gj{_=FRH!5YZ -*F!`9K$Zj>~hbtbK1L?%gEaT?XMhGWT8+ea#*qGhW3!dt>W3Cr+9crgoF;1J{{mP&4E0~QAGZUL)H^< -k8_d5ml%FyIdFQBvXUT{1Lr2|F%;aKhnR%pMXAPzrz9W9`A+oDYDoblbj7lV8etZt)XUX9N -Osce1$OhB)dU|8EjJQzrmtFge2K&sMBavL~Qt26)9FHs&P2vHjAt$(d1#4 -Q#>MdbFmc`pJJ1X^G73gw#+8NHHeXzU6!2=ddb`HHDqH9+E+l3^+#WiilC(3Iltc}8=S7#eH9PtXZBq -@=!)5+7`j5LW{_FSgDKa1@mjfsEi`s)VFz5fL)Cv2uwzl -16i7V?}f;hPkQa{KXw#BW1I_N5^-o$cOmcYAsDswrX2Rd&>q~Eg1vKT3}m_FzaR=4a2uoz>bC@YH8OB -8t*%nr7duq<^^Y$I}e?Qufo9aP&yj<%zm0!K8*#3B(@mdY%JMVdJw6_X@0iO!Nl!wScYjmjdL19Ta4? -QuX+y2R;tl2hh{+8L2Tu~|G-B&5Y8l}Mo&jwlUEJPgv?&Qv+6#3~a)GY%O$CkFLg7TZf4I=4qgKIO!7 -$~hvUAlRa655%Goc?_6i3tN?6&MJZ*$aNr_a3D}3KPv4(P`VcR$c2b(!yzwX;Hd4LSjwYHXZ9}j?k+VQt1nJcv(yIZhn~jRerHC|0l&w -N!tB{cCD9E#Pv|HajAZ$@&4}p!PBGM6Ev_NEYD%!4UapF<)6t)hOyDN&e->3bVU9Ei>HI5UEBCyRwV5 -{YxpjuA5%@=icbscXNac-@e{cvs~lGQAQ)9*Od5z(A#T0@u6 -s$#rtt|o4}sDX-E$i+;X7jvVeI|miZXuBi*;<9PVv@4+-Hm_10%{c+dqb>ZZd!pG?UN^0 -(jJ&y~B@5kv`e`^tm5sEGk_N82e_d5Wvt89X^E%U9HO)0o&OA7`MKiZ4qp-_bO+bWM$Z)fxyMa}Ywh9Bl*?wVXLT -cS)+8nZBfH1`{2S~#aYjdPm8n7Ha8U?avKhcl~dn!%X3>Y$sg5ESKF>n2($)9t*tFCy2n*vhpi4f8!@SgrI^mS -@tu|ohLw)C+tqsoJr`=OxkWyP&Kf)Q%UCrb;B?G4=QyRGTBvgv`OLF-39JEMkfz+Z%n@6@lQI_p!!06HzENjU@61F5--$ZbAL$%r=g33o`jVF@1r@4V>s?DUB?_BUr}dd;(lql7%rkJ_66aBM+Z69>&Jh^(^(~Iq^R8yh@-) -$LDra+^eGGmk2XR*P2z6nQj(OPx69YJf8LvMMB<|(jK?Li24_a7<_6K_zIm=gr1zk7#XlpyxNM;!)$V -SbEb+ypC6wlYx=jQ#l^IPfpy_7BWpRYIVhwDx2>u>t1<_#Bar>M|6Mc0jJ*@XY^5bO`wqQtPbIz*y>3fJ0ciP4S -=i>@_(Hr^1@w7P^b`U}{$RMViS>g`*mOzfiKoQE2HGiq4E94aYd4z*@@gemWaTM9?U}YIbuh$P_NZl_rm9TTvPZ>Un5 -bn>YW%~f88b6g9vT&`F;$n)trL$;<~;Im&anwGs#UP-su?q=wx+dXiPa`HT}SR}5$#a*g;u4eMYG7|+ -|>Bis&=;G5Y*hJ2+&r>ayZY+vFwJ$(zf)}z(oz$9BCRC8m(E9+7>BkY(APR&p_pjlXKj#)VT$2&Ms;5 -0hox7`#?r-zQc;PGb5V2$(iX`e?{0O>2BQ4hG7V;pA$z=(D{he~b-TDWa -qnG(>P?TXm$w&*&+PPggMruJzoIdE84J}{`fU_H4@Abr4P&AU&Z{`&V!Wf9s>zc#>v{pg_2S6tozG+g -?xf1~}QO9z}ebeKAHm^yTrI&_$}S;|5eXwN+do^J0qxsDZ1qr3%lZ^t6}3E1%j?05opJOMkR`>oNQ2U -Nc`>Y(rgc02()vLtD=r+oIw(V?W=p`_fQq}&zxZ8Ad9OK3QGGaSbCTb+d(Z8Zc$N9je<~xjT;MgJ+jcLP_F)M(Jg=y5&d{MbRMK9SC=a>UamL-GNna>6xBh|JnV~-#q# -8|M=w5fB4CJPyVO(pZw4N+mrwOY5(^o*!H)#{BHecxI2(9;m386yF644wFkF&2Ws=VfLRfGSh*ot(b? -D@T*#qzr9HUHL#0i7^y`QX{rsC1wTCF^-!{3h!&t+{J&34|Y7g%4u+VP!wKnq0hb85H-^Qw8lT`yoAW -p6d=~f1HGbU8g_{RMTxnqR4J%FenRc^lF2WmxYU^qnuFy=nnMz -X==#wLootlmuB(5@b;lWKYp`b)p>|GerHt(Z2{s1R(N3&gp|3 -&5ORM$UQJ;cMfu*BOdUUZ?#6b&xCNCA -orE78A>sRM_p1Bc1D{t(?&vWD)XdFd)?x{9iPp6|(WU?JF+`#^6tBUc0{&juyCiTY==4(eUJQIrxHQS{1{K((S%$nw4@ -RH}}NcM%On7D_JaHkZb83o{PI+tjIQd{oQNy6GxaZGWGt?&q9W08{mjqC!m2wckA8GGX8~y24Fe5Nbm -4oPny>rK&|iQ*|?GCOAkn9d*m%8WFJ>yrIZ)zxQbe!M~(-V^UCS`uE(Ql(ryc*KW&pb}TG_c5HNo?oif~n5H<%!-;8%6VntYoA49U6ep%B&Md1?R;^W -5NvxreG(w8cA;srd47NF5j!sZRXjs4-egdp4={*#^g%(Z=CwkxV!eu5Z^f9VaF{)Gfk*RDADn@l`5l( -czVN|DLPN#BISy`pjB8yFfR0HoKwpFX>#@;$M>qGWC|3h2T^)nV!uC7QFqrb4NqKYi0ID!={@mU2o6j -?2S%VDi0c7Api>LQNm%ys;5piz#WfQ`j)Uw>9HgVG*M#6>3caZB1ZX?-pF;Xlik>i_Ik1P5poTU!MH -0|KERWm6KY3fy7cybKHAm>-?5)SdOF+&xEkg4+s=83v^)D|=h=@lWxJ7%Yq(XhmAP5V$)^CcFW3|r# -*Tqo9t9Yx#*r`l3)S37v_p8F)+*dw*v#w#yaarL=>DT6L)m+M#(=oG$)2J9H=uVat97kgBl$JZ)0gAi -wY!UTVIz2NRrhE&C2eu%I>hsyfu1C?;jUJTT+>ltfm^OnoUrf)BB>X@f(?Uwd1g>yvZKsQL*Y&(WOUo -_KoC+L%IFd{?7JST7MGhdO(RJ)6-$Xw(Yt^<5BdExI4qO|!!krFyB1<0GYem`kMTE>F5lM7Zv(3t -=*pTI(VY%1#S3+0Waser`a#qWRu%cq5qGF_Cw6lu9M`d29f-^*&N~2c{FN!2aMHJVZf+#~B%ObZf%PB -WChI3!yPxe4@v$8t^62*GlMmnu%b#_=eb!e9B4@Lu$w+=g}4z(+dilHWUM^<(kST1!~E_GNgbyzNSs8 -(@S4vEmA;?R%MsF;58V){uAoQ3O<5i2M9SXdI>BJBYi8gm_5MH=GPW@ywIC4K#yg7u@O2LJkZrj0j^{_#%WUK+wU>fmRnV>S(G?NaRrCyV -tBAR -ZwJnw@MJ@~q4O3*Y!$6v=^3jTcsVZ|8a?!1JW%#JN1vctQvqMYPUzygZ0I6N#}~KDRfEJwVk8;I&!4pavh>RHg0J&H;({k10i%=Q5DmHqs* -n%z9u7=89hh_qyWqt7Y|wsjS;rJvGq9TB=eIqWL90xyiukh2&FzL>P0dROJ7z_(csrcqINd5>0gJbnY -ZyZ}^X?G^HP;q!pSKr!cM^0l01Wq6t^@@bD`)y4Pt0WcFgV8(hr&&_F8Rh7-9lACrY^J=aMQc#q^ -)K!_S!VG2|WUD_Cb_#ylp(ry?WtG{DJH6%_nVU$fybN4So_gA{-nQhh%#FpsmgF!c55le(2!-rrRa}_ -EL}i3Smy1K%qt2SEX5fUXS!UolmQyCoz@gi*{amZHa-&AZO`3sfWZdK#IGU(h)#zI7Fx*t9TD)pCpSD -cn@Z+SRy7{^rJGScl3>>#p)vYsdwZpZ|z;UEeXT#KGh?UlP*;@M{>g6gka4bdYR%hV2SFclXW!(o3u< -pkU%p;$(2h?Mza<0cL<}r{$^_V3(##u(4I(_SLDCao?2Lz{1`PSxCJM!~G=h)KMbJk}+03NPsM}BB`Z -LXlMY2!Yq5&Jk38sblN-a0m?a^@L0#zCF;`&|`yTY&xQx}b0U;ppBk)Z4G`f5lGwuC2@Zf&^6rqM))r -*kto7Ww;G)5$%;X&+^CV#q-zCo_|>`{LSA!`PIi5Xvc)ndL812w>8TRtL@*?^wlNdb6LRJ#KTtPp1tU -y+Df)xUsY;#Pmk4Q;F)FIu(Nz4JLC_!ZxD05H%fsn<6E~DdcTXH(H}76bWa57vg#zDrpu7iu-zn{dy+ -`E!bkT_IQ5`Q_~Y{>mk82j!bxO{h6yKRi@uo&rw2ulM3U$p1WSZJE}8G|%Z8KKZMwvh9u -ghr(gtzthmAaun+qLwW8~>A!6rnC{)~2?L}qG8wSOa-sRv9&UAVUU0#tnYd$t=zMZP77bx-2{0mD)xa -|^Z$qV!Egw4OXw@%-=E@|4(|x@5f`lJ>vkzwn6A|ATg_NctW`@G@E@pZ-hOuDX;H`~iIglz^rXtpxXB -jWYPZt<@_$qRHSJ%Dp{kBMTw8d%#MO3A^+n{ONYHE;%|;sUY9EuTnt_Yu%_+c+1KbY2Kne-!i!+2r$v -V@JH=$U9x!(X=V^`*=3Cj1r5U*7fIiPnBF(db*snWJKO2HqyvW@h$$U-UmL_Nxb>cD2(|Mi;)A4oi4W -ex!2TudeBF~+{Hld7u~T*{#CQ*>`;Y~&dqR#kNZyAe9xt644_FJkWFhZ~KYp#m<7G`uAi|d&F(eYdR& -$}qpjiwWSuw5oZ`xYet>mTREBi6~Jv(Ce^ffLbt+!wp{ySM18$j1y_=hcwA?`UWjNus;#*pCrdJALsq -(U!qW0%zD4owmShT|R%{GL4MEhv1LD7$Bs?2nls8`7iqRHNXzAm-o3IGG~R(V#Lcl_6z1ER|jIrFe!I -DW0za*2OmMg94^v*6bdj{ULEu(SLH8H4~xFdrdFFL`c0F};Gv&38q6(#0Us3-{__gbs4ae? -_&XyezL1pZ@^u=h1J%DJ=USol)TD5u_kL^&fM8Y>_g)yIg(6XQAj9zb<;2;S4tDAzJ+y%aX$DX1L;N# -)R^pNt-Q->bNzFf% -zMj9UdJ7+!%7~y5-%%xVi-?s#62V@O_N-r(f3V8bGI;TMju#w8}{tQTApY~x^Ey)JR&@OD;T4a)L*fd -cS%xT2J%Fm(tQJY;z4)P$zp7X)Lho`?%Bk<kZH7&}VFnM+eWP}wcS+bTjY}eJEuzzZvTZv -lXx*a##1%RRTi?H6TGCk%ytae{bga9uzgYk^29}*7|Pr1N>3DeJZ|mYo -bpB|1=YQ4~azG(>Vi8!a7E_r`-#v3^2yr;@{4^Mw=4 -J9Qw==xM3&AuS#EWNyEMCBuhQcl=;B4OJb;gF?dhbcCob#=o%j!y8qO-$m);G?}RCA8(B(>c_vis*@W -ti_@awgW}aw!O4xm#RqIFJ}8HMYjE*Km&sHQd84-E#tNgV8@aKQxJ&x#ywUNx|2TePwj*&mrwBv(%Nv4&@heX{UH2wHBdP{W5xzR@Q-H -b$j7riB_t>i)B{O@LCa%$|Huit1uxl!Y(uRov#;J#tXt88CmnDU+g{c!zY)+!&;R`8$+%ZHV^+}O2Lr -7m*?(VWY0J)i;MTUfxn)xAH&aK6?orUnja>-&0U^z|}1%hkmHR%~g0osygz?XU_9yQdviUEjavoLf@} -s|HS`_kV6%2Wov`=pt?|ExPL6YA#Lv`?GprH6MQts|Tt9piNb3q2NJd1Zi@P^|}F56hUzepdv3KjA?dJ(sYYj9C*ckLD(ok*n2#HD -k*{y&o@Ibo}!T91hR21q8=T{p}>~T4Pf8^FcR~90T}t%g;IP7pD2(-a(;_X{Bwd4A3zq0daOV{{=UJ4 -4`2mJOcZl{&rm?Y^F0La1Ic&c`ThbxzATF61&a19;doEK4q%6V`9oBa{K;TC!OGo($C3VhMBE<(!2uE -DR_txq8+fw;djqw0AL;g8!EFjoGf=Q+1ENO!UjmgzUYHNQ9+dU~^i1;o5S&bYHmcZ5y5M3?)Zn8t<^q -jj)yN3K%K$4e7&kJa=G|bH)d*4vHFf=-1`n#)QTz1Y59sVK3N7pPt!D>Ank?Q!lRf=9d> -EcTh7ywo`&L*mg@2(PXu!X)ohA0Jm?|u@v=!ONf$I`OU=cwTnI&!4Y}eJ5UUsoCb#>vo93Za?%7ucYB -%SHo;kuBaZjS^N#YAF3pswQS1sNc(VfMLu{#}B+z7Em#Zv?SbM7@)k-Tn6PEfKE$LE&3pgV;L14PxtG -KS&F})kZIP8?jYxKw2k=tspUc=UXDSj2OTWGj|4%mbqe3Z-s1mI53|37O;z?53E`of`@tl%XI+7I*pj -VbL$PT3sbpo0lO}9c}{>`H;69mU5WfrKs$9ftE71k>xD^Q#d-naLM>Z+#0RimP*K1Yz;#8;V?;5yfw| -fc#V$t#1SaPC{s1~G+@iz$4}=bLe|8KQ6mZnPG9=iyMs?i-ay#7{1V^xOh76_%*j -&6WZHzr^;dg>FrK}UTT^poO3emKli9{7X5G?M8dP|^()N$tzt0E`sq<|pLo26c1;E7CsU^~&!-tkEs3 -Xz2R5k04UhMK=f{HQ9QDAi4ny1%&`TgbTVw`%HsHH)tQ#gn19`qplk(L{435xkxJ$t*w{yK?Kn8eIbC -<`t1$u=OKj8cf{?CG_!&iza#KQQ&(WkJq*oNng1#=G!MaLZjmu=>I|BJ6B+a80l^HHLYDrWAYc9pGZE -)sVJ6=r%;aGJ#yzaXL{MnpUttx0g;o5Zu!=|TypzJ~u2JfvM?XuCescHdXBGM!`U3i@)1Q9^{Txd8&t -F5o?DQXa_a6$BcYl$gy!8wCKY-^0!asoX1HSdaGpO+Jtq!~Xc_{a+K>lW0sM31Pi -qfA3}a{KdbN$1U>RZ -iNmwYIG*^P~SG_#p{s`}>4>9czGS$Y0Ay?gi5pWfLSe#P~4+Nb;TqbK&)AK71bzVzwK$1h&y_SO^o?) -`VJk9X#lfyd)L^Xs47rFZXfmd{?52j6`!eDOtjo*zF>{hj&JS6}!VW}4+Ecd-|_vD_#wMf8ayHQ6K#-{Qk-B;_ustdZ(rJ@yE}@=jG#%dye`tefhd -RIUc=x7hc`fFJ^IiLgjgfQq?b7{|x!@yOP?c&0Svh9xfaoy?;60tIt{mx2faqr;pwpJ?h2ld%X8pe{? -$_fAnq}EVCQ7g7;>58J>Sq9-B8vSDBkW)RmO??$sG*)Av#F)7y%B@2a!kKECNytx2R;_69L;a$859~-G97GH{ES!yj4xB6_Wh+|h0kug^OvvsoZZ^}^YA)-+G^z{wP(3?m}m7_dHJdr`-9)U{`AH3yXC#x=1lKBdo}qUzg7 -DAp^t6`&95KGdrRu^qd*0``>m4HVRxJA*ABi%zj^cxMqnFvzj^uWb$Ki!?n;lC`2+B(%@id+p80tzqy -G|6AbI_e-2Z);W#~UE@7|ww_3LdmYtg^&Q%+0lZ^G}NK7acB(M@iyAH1*TSLpPjXyC(SE?Idvpe%vPDSI=HN{{sm!pS}Ce{|``00|XQR000O8%K%wh+ljQdZw&wd -FDC#18~^|SaA|NaUv_0~WN&gWaCv8KWo~qHFLiWjY%Xwly&7qA8b|g!zoH|jc4lN4kaBF6SW?+KV$~{ -=%a)Sem1Jm!1|}M&ad(e^wY|T6-+SG20aEP56_PSTAMd_LTU%QV2A|n-JQb3*OBIhNrppfVN4DGPeBavb?0&}vOJ*j5y>d915q#QYS_t<1>hk5q>o=hVlf;oowd -k@JZ;tH#+3Df&`I}=ll8Oz5;c)^uw}9qGJe^5pm=cYVl2aB-=2kAn!kk|Gs`zXsRHM<*GoZOQ<7U!Rg -6+YLFp|?*oCsfe*X^4dcIyR=p^VIOCVGjKciC*iNzWr1pM7<7ad`RfSI0N{sZT5Hnz#P@y!D^nt$XOS -{&Bl)xBP=HgM;+Uo-jE>squuxDSLVO^VxTg0DxG`q!`OER^7Z-1igNEVb4SYTG!>0#GpTB -VkHsTH%i)qpuiYVED^G0t!yE%XZ#KjFFBftTuD@wT5>=6Gwm*zxf>Cpd$Cz&`_N~)ksRDlJ&=CKw-7L -S-r1(PGPFsr@NKG`rHrk-J=I2|(1rrbo6I2|)W5|SM`gcqb5!r>n;B0)b(Z1jCOi8T}XCY#4e!UlpVO -9>b`hXt!+Hd8VZS|hTOns`KTkWw@@21b34tmWi$HO|mO7~n~zVUbiyL?$e#Xbi=OjfLq&6A8Z317yBM -sGEp6~AW9002-DFsXR!rx?^`cpph0d1j<06q=tLL>o8Ql -r%AK+5C+jaC)8s<_l-v5D_RvTTA%Kt1q(|4}J|eg+waP9z~RV7o>VtQSCm0Hu*phG6d^U(zgsbu@M{g -L`13aEoilL=%N7)leva#CU`rvJS8`Cx`@74j!4rX=(I2+-!ugwh79j%?+<&kvX${hA^Fdf=|mR=1Ar| -HWP4uM~e+>g?;ubO79Z|F}WWfAhss!JE0*RPIDM_QVlt5=+Ym`2{??0xtS5WnvAF4zf_s%18oon`~96 -h2*2N73T+>Z+f0YngZeVX4df71or55UEnsx))l}N=QN9Vx!~M~y-zQi;MLQrC9i;^nLuz$~68st8i#2 -l3!Rh^6`)IjA#hYpdjw>SBd@)YV`aOi^P6Xv`a@}v7$D#}e&KEWmZ`Tc1vHz-L3QwMD)C<;fj}VzY -p}?E9I$H8d`w){W&a9!l@Sok;j;qZbIrP#UAz5i{B+?BfC&Xr!;m*wb|7Mb&*RQmsPGTizfsFI=>oXA -xl)C|F2UYNs9(}Nmoj@?+?1lm;N=xYE%RQDy0WDDqUP}D=qn5%iS_4M!spd6WyZirpC>VE3)n4S4&a8 -(C$SNF#v?(&wX9q!X(CaO2cYs}NHReHxo~E(Spvk`Lq*y^^P_|H_zUd+f#3_b*215q5OzZ1!w)t_U%K -fd9ZgGUunq%v@nG;mg5qjs=$~U?J@F;AF42#+vKopq>D!U* -FOXq?D==EM&^NAtIXjffXkn(0)Z$fDG62Q{5U<06)YY6#ADfw#A;-jt0fix@hq;0F%0_R*IghC;?&+C -JurV#+2WQ9*3TCskuQ-R5~Pk8||B$ID)Qw0cK4{!RPX>1bl-VhY9M9L~1 -bl-zyNA_6Bs0}R_QXEW$<28Y7Kc -+D!I)d>-lrN-bXD=g_+2qU%i;CQcuKygH_wQP170di^pQ&s8FDh!-#+Hq*vAi!AS-#RB{+ALrm_e}CVG&e~uopRufvTD!H%hWF&FcBWN(5T|XC-UC38K#Co|XC -M9DoA72B9C)QlisI)}fLbWK;kRELW=my*<>OGx((2Q}w!bu3FfC=YMB9R}Nc~POvkKz$HZax=H^wd!Td1(y`mxe59U&*6ys%Qpy9NH9RypRHgS=P8i`N4-@W!OlX08)<(0gWGSIv?{;t3^J -VP=)GVkr!$h&Dk<;>m+w0cO?P{1mMgCU`U-E_3R>>}oN#xb(<;Z*cp4LD5H=z%P`XQPQKMc6}0InDW2 -Vc8kW-%o#X$NxYW@rt?Cz%BK$O0?cjWiXJt43o{4AI>B;iR#eGu(|LMet4b=)++*aS6i596+XE_bxOb -bW&fDZ8&q&a$E3%6EJULrU`p>_2TUGkZrZv?O(q+Y`2dtkJ!I(B+ejTvo|!CYPXNix4fDQOhFfGllok -mWw`;wPf@CMiLPllm&6{MhGQkOnZHw;&{egL0(C>`kR<23+l7Tz)##LitoAl(=mFJG^hCZi6pMEP+lL -AD3?#E0VJ-|uT0F87~ex#H@dXhBsg`ZV99HLQE>YnCd=>+Vjc)48o%NPg#GM -lkN&uksBmu3F51u`xGd#$)A2rNVxqnvkkBgl`8eo}Rih;$5JW9T$+1H3wAJKwWGY;?0~ -TlA)4F5O;UJKFsL0zlVoO^lg!+wJ*$9y+o@smARAu-=X)oo~X~a8$+XcNW9 -HM;&OI{0IsMw$llkv>0{V;@?sqUNvUP29+U~NAWQB0*ckoy{Q_8Amt>Stki6L0Hm=Kg8`>pEAP1M%>b*=Y#3Lp -)o?fFaf;>E@HZ7RhMswF4ALrx^RD^mdH$PCnbtpsd0~2!U@Uy&lmb+s4&xtA08S%s>#5L!A0s9#4oQ$ -c60c@}%{{kG2eRPCeq#oTDaF=vryNxbA!zVn9l_(TXqt2!-8HBB6B=65I$=14_CekPYd6GI4%r&i35j5GWz>@g-^g(RnZBo73XLxm2l^7>$-Ic-n@0ts^vYvaSmjF -qeq0$6!xZRlImWDQd_QmzLQr|ew*y~331+I6#FnJLKsI9*3Tm}Cpbz~E7i6rHO7h$stAR(V+W&^-P>> -XCV7xJsi$Vogx{124b!E65AZBZ+_ZgBPHjfIO~}B{Dh4MwmVz&JRTq^TQ&v)A`lljz7aul0i!nH2x1z -O9KQH000080LuVbTQuD)n71DQ0LYyH04M+e0B~t=FJE?LZe(wAFK~HhZDnqBb1!CTY;Mt -BUtcb8dBt6AciXtJ{+?gKrXM74U0vIG*~EwLIoZ6eyH1?iPB(3mS`YmTQ^U)nt)owJa-r@qftYmcII;m^Yb>k|M7~ny2|PGM -FeUiVw{uS~Pi5r$t^xBA-Wdxe!fOYh+vZ#ME<`6q}neU9RfG=yCToI@sI0|NDczgL~2JCaPC5dew3k) -l#f?qe{x?(d&!vPtVRLrkX5GWM0X`=)3bL=Kjg?XD -(#_=Zlg#HuIoYb9E?)+30htim_%o?Zp -u7*S>!TG7g1gSEyd>Kc5AljBI+xQfTL%}Cl^o8&P~u>wg?~O@Kc!sCYlvg!rnzU?7cw+mkBYZPnOM)}BDy||u3Ormyc!y5Z_Dyp)X -7RQtSL+s+YqzLv8Ia{1_LW0E`Y}aT!SY1(s*aL{Ry|mJJ==k{pdGGnyBp!Mtm&dTU|&)-V-fZLSy+1l -W#TX5Xisp$+vRC#m?mP6~&CqnQ%kwyobioHVlD&E-%E=gy6CH(_7H@&`D*wqiT7DesC_MS}b++NK7x$ -c6~)+I)RG1Xp+bcINzbUJXKYw8m7@y?R=vYiWgl -~tE~vr>6q({N}||^Yy#e&!@81yK+#{cV_d2bb$L0iWVYaOP<>2uOa@DC3+e_rviinxTWB{LbqXI1cvky)I| -BWTpuEzK`m4oIUqCbJk5bAA*`i`+0u_Uc1MyN$z#cYfCdcLI-E92#pLgEPMicNwaHH|)zekgwiZma;C -B4T)R*BfiG1Rl89T&bR+cGw`P=B|G_d7c~y#h`2X_b=6YYB-7m~w;_{aT)2_yIQ)tJk(e8R|8o0J?ggC! -h&x&jyKs+vTaE)3;%9VtF(gb2er4dzBH4uTpPv5;%L}q*D^=Z-?t_(diPFbo0S+*n>d79ulLmF&TJpW4gky` -p0M&3;J^kf#d__haK{GBW#%A(0e5Y{C+HkDRSGsV8UWn00iT01+_wSF@)qD*#uAtkMPa}L8;}+umLA% -G4`8-_EeRW75+{nlfSC^|-DbY!IgKE!UEl!u+mVOT!dMAGa2p -svZva2m92#l?)t&DUd#{l-W#_!xZlAk(NFS0qzmNswh7AOkbsDT;hvTy@a{0cbz -1Yi9@qpuH%f&}>{X(9{7_=n0W^19OTbC)sTB{m?=!X4EFIAFr!F=D0 -O^6Q%q87l&xycCF9tEQv^*noZ3v3my`2M6~ -_J2-5c0BGdnF@^vLOG{TnTYwprp{lBMJ#0(g()p#j1qW~`5P*&pI|?>a7(i!#KDwm90H -W4J1Ui(1hb($KFgSWs3QdTD;1oCpQ=l#V(M1CbeV0*iV30rTQPcS7(i5gYSo%&=pcqVnxG1#QI?bDE& -@}c5OW@8M-$)UXKAKf#w0axxKr$^kAXo;XR!0=GDcm8Q;=Ew)W&`NZ5-9iqcS$=q!X5#TKqji2wYR13 -6165U0QdHBcS%djk4@_BMeh@z{$t?^$Jifef>m7AsVBh_)HF7wOuQ7n6*N>P0lmu~IndaW3n&vU}2S4nwk8+NV?TVBQ`V-P@a!9%$~cBnv=iu#&lJ#DQ&z#-*Ci{n! -9Hcb}Gi)7WQO<#H)K!!r1f0GtR6pp!W(Z%ZE>&?kOn5I{%lxr+t{&}S&Qiv|YJfnS;GA-FZuVhsnVEw -un^rUe7&^SBN$g_`QXIXBPZyych|X(359Gk$9*DGb;q)L(fsb|_7)*hfMiFwLhdf_$U@!$@- -i0F%uqhCaE2_*f^1}e?aRnb7+YUs1+KE6EBu#XiKqj!iPy2_$KT_Ny6q3RP!4!yE(|l3i_+;3Z6cCu3 -IUs#p;Y45np^#*d+&AgqaNz!ZIzemZ1PBHLh_$YyC~HOn1L*i+Bj(ZzAZitv-1uf4_V3X2XT7O!d}AP -*8Rue_d7j4sfIbWcnNvZ>`4~X7w8(_F;qKp`QW?;smTjws2W10OlM}uS(g`Adf@<;9{wqFXURSZs_!dWe -`rlT2x7sZ4sCP0m!CgV=OCF(s{m(dcS!%dIL+ur*EWdt_1^VCIB)dl7<=#IB)=(vXps(0w4z5<$$?dm -i8nq)>3DDKbse{x>vV`-qDIb%d+^xupam4-%6s#7%azuupq>I;`+)(#DYi)QRc=$7(zTVUKCB~mu0a> -A?YWVFc?DOa9pL=0a+FY6e2&#+$+n1*j_AE#PCCiZN$pa0KMgW=hqnH+Fod@L^B`l?$>Z;XROsWa_&5lmNmTsDFk`g!e+WZJx*4B@8KdIA2rHQy|I09Aq|b -~`!w}Me#IM2-+AeusEsXy5cN2PA)No`pbzsxx^6aFE8B5C4#U#`K3F##1Ja}sXM -vE5b7_l+{q<|P=EQ8JGsOV>Mv*Rwc2=$i_?&K0fsKI3JSt{mVypa)}|-UrKj!i6PWqDtB^;AZ$*lb|;q@Lj9$2CzlvP{pFK8xx^6aFIVp55<{r -JT)UG?459vV<4!Iyg!;>8cXEj#)L-<{;^Em*$J*jpEiU%VjyhHs$9j3Op9)p7USRBpP$jqPizx_0sFF -X2WuZ!TBPem55V|HgJ`b8DeI^^f4%!%d%<(E{FzhkMtDwQq6&Uep&|ugz -3$?49Ct9kGk3!U5IGJSKG9K^KUMv^Rj06u7sm97MgiYK@;gdx{XtYf85EYidZx^rHKm?)h>BrWR?jF` -d_)}4|rOar63RN=MiDu-W!XBjsWnpuDCt+ErS)K)DVRL;aVOgkI+AFCY5VpcQ=1aUa3{#OG)jw$?jdr*lW+Bgj4a_F?=Y2%#_K$JT;xm%*?V-?6>q+ofdoAeAio?l -1_#G9|j~yTu>~%VDgA-%eRr3TH3VDlZI!cw>(ggqHCtH^y3hQ4kczYap^ -Lh{(GB=LhYcPaGezP-x$CByR0C7u%qA-Lix!)aN0K&Eh=%`Q>hS1I;^*^$4Y!MLaJp#N+ -W<9c5{q74Fben+Ktbg}HxLKMt_QOrd#P3$wdmz;JyS4lt2zxN}(J2jru!)F%$ABS&VHs?{XTTtYDml% -$<=YVUSnxOq;UkMY4J7dFN#g4ugbjJBk4`#Rd>UcSLlCxs08nx23{#}F2Ii+AqEeuXzT0Z-2ZWeVUGpXl5F2&4z?2l -WaG}46%d5Y^r~ge0UKlfQae?KAyi|t23p0Nkg!+I)E)|fJeeB5*%QG-g>9n{&)padVY8#F&`^}U-cW` -iZ00I=(!mh+P&jwe!4NihM>{GYBp$vov&4HUxbkGSslpaAJe|(O-|nk0=r43*g}4az$ZW3fG~6T8)-T -*{iK$Q*I}JhD1`BZ!^h&zLLVOy6uuT@?!j&i4RH!4HeHOeB=9PtGvcq;U_TqsXgQ&0-spm~7?O`t)zb -cEsNSwU|{aEj~;EiKc$!We&r@zpbwJ?MxiYnz@mh34n?#0-vfNl(iP%7Qd3xibH)(bJ(d%^txf^EJKZ -+Bl9VvKFS5O4Qi7=}m3+;2E(>sh`T))hLy}l7J3&3|KOdy4`cRW@J-yY7j!e{I{{dHn)bANF$_~- -n=!P&U=Z1OkTD?7<_Y;xK*zF^PI6rovvHyiPK)W>4TS6&f -x+HnLKj5fG+An^Ns);qQ{z+_>4(!fRtRTHd&JBCsJe!l*qhe2nppE#huhi=jBix3-AQz7{@N*m`8K7M -WvQjk1e7RLSeW%aaHD>?u!He3m(%y2j}!HOs5uqh~fuxGJ4<1bB*oO?T8~lX%B3>=jLsG_@~Uf|4mD& -FB1uGYsLIB+OHYlvO%7*#)gn#ZTls&3HeC8KaO@lP^WttTWL#hY5QJa~j90V(r@^V{h*cdl@X3SD9mD -4Kzc-7Je2Rr;i~==I`}o1&{kSgxc5&bnK;ahjrqAs~XpuA!L^9`945gwu~XvU;Z{LGVh4*+k1@4%*h# -|GN-NoHyMPusCcuBeKvQYs01KgXYq4)BwD__T#Jp`X{SFq4G?rIp53TDbq(%W=cxVs>zwAO7;8^+?2j -%+=Q*nTNP73@Z7Md%0m~w;$X%8K4KzpJ|QmpR#))e0(|umhwUNPk3 -}T|=Nx`8CDj3=QC3{ZG69S_W3bt5`X7ndq_sYKSe+%f-i@wSX|mFFt)<8-&|+EDpuSq2!n=j4102=mu -3mh=VIp#MKwx{ip`0uyN_*-Q!~THNSCEo*N3^PQC)@K4BQY17T5+PQGpD8&3R^;}++=duq;nZMv=#W7 -My|sQ;f^{V5o%1hl3Yg{YWYUf>DGebn%03i>rnmd6*zR#VOPnoe6Xr+QI9LC`(NOShFd6acCYpo{H)FLz0haw{{VpzC;#MMe>JHh`(*LS;yb!@RH$OrCDFrZ~7zKQ5GY7Ns4C9-&IrP0`*~ -A*mK>XuE`KsEj_lwAHUb`L-od2wgfLXq -Mzk`m9?lcS7w&(=#$Ms>j1B-kPyKO}wwwW_UNEri4F) -bnl}Y)8DX>24##x-|qF$YG=!FPOV_{>*p`6BMnJw>&KP2cwF?)X_d}0`Qs`Dr`d=kG-{~D4B~J*KcT# -A{$HvGQ%LO)pG3aoTvaul{_;;q>c@>ayuGdM#qpCTCr_g%rx!aDTz9k(5G-BQDl0-&o9?%*%^ -DXxQ#Ev>f~nH`ppx`K>^kIE+PpfY?nbkwwsNJGtaid)lbYTy|C&zQ`Z^(_v*YLAU;O^n>G8{p8~NgRUv)yURVJ)+d3N@4(-;WnEF7neP+9iuz -Dx==5*Dp40OHj46QJp&x%mc3(`q)jNAD9!A-Er$A)>4U9VbdLBdo5nx5X}mf -TmSy?dg^HTYWO0(4@LmQ4v*1C~N3R;}@VQPwV?{=smG8lVy>Q~bNUDn$LB~+SYz<UEPsJ0Wk*Me{W?Uq-y$P=l?0J<~X(RpnKDJ5LliH?q< -ecmxv-6Vt7kfET8e?dCyf+|v)Y^P4p?`VWaASRr&(<=~B$grio4?1~ZjYLe&4L4om2NjBbcNC4_zi*; -(N8c?~5?IZ-*%u8wcH!u$Bq#_LjD5+Y%Y}RH>S2E2B+W0IWhY$zQD}mOL+J&|wdZMa)e|?2D#^u3V=hrYXl#)M|1oRc{)hRBH!6e`s^Kd0A{sie?6 -<(`gR*Jxb4t#Uhf>b>Ha9b?yd~+GImlw-z_(Nyc_60`s-dnH-v-P-7X$@ABGYl`v9BujSm#dafa_FcX -t)>Z2$1@=*=Js7gyfq=o_{DVwGaO?TSatxi+id1w(WwMqvwmi@`WHe*qLf1T;S!R}L;(H`o047Y7{Z3 -t2C?b;oWx)Q}YRWM^>GUzu!PTzW~${xN4DxUDs0!AZ440__e>aXk;^KD19y`-x=f;$jfzAvu8w(=X?N -cui%sSfw0f#$c-nOql;TTiE|s%e`*S0+BUPw~LccUwo$mm#B$^eDNSSe2=F`KqnISDP%YwYS?oKGwzF -?!9kYrxUtuK600Z&s2$`b_bB-{avkVjo5K?!EIfsmy)`h+hn@9rFftGRHPjVyU*J`+x-ocbabhnIHsr -1UdVDe6wCCzN=@!pGHGfx_-WfkKkccz;TxLpw_<><$F#TdYGF+7Kkmq1Xa43zt@^T?9WQdp?sZLL-TZ -gv&M-0Fb~d@+9ewe%H;}NzXeS62yY4>HSP91L+SCVmYt^JB3J5^?|6(q~+}*nM?D*v3>Dl@AIH!^7;c -Br=7dN9rUCU#*a3Pk)Tucr(Y9FEN$IH4DdqtrwxP4gFR_1zQ*%s9kpCE&p7S*K(o*=Zv`*D9;^}wuWT -iisAjtltUY@{ut{c)|Fw=G9Ky)D@6aWAK2ms3fSz9&QKakD_000RU001ih -003}la4%nWWo~3|axZXsXKiI}baO9eZ*py6baZ8Mb1z?CX>MtBUtcb8dBs>uZ`?KzzUx=8o5O0;t>7A -H4-5FH~*z1^_0?v))M9sT;#@zL=wY`bUnUJL@P;JY<5La@`1*S~+byh -@yyO6Ectv0-nneslY8|2R8;e|3&WvO-ucD>!ojH;Ck}){SM2=-$lUbgy<_i~dKcK-v60DjU6H+9Wl%_ -ss8Yxe -oX>2?gVFfA;q&d&Q=O^s(iLqBKN7aQvD3HO^ny6Gj7DZ7ZN|u%?C65Ucmuk)KD!pB8NL6tBrLb+I*yn -ozh-wH{rmgErih`kDBprOtya?vcVco5hg~z`vNip-pd^j{jsraf8gau-u3w*Si|T#~x~lZM -ZTOw<0FxSG};tY3~IkL=9Yyz;DhS6Ht01js0Tcoi5r+kkgj) -td*_@Xo1&UYdp{tl4|5`u}g19*E0ZcM%NM0SDv|I0*O?PdHdNuMC_5&;!~Ak62GQ2ir}dbpyN!lAsgd --CO1m#_<7qI8Z6kaO)H$iws8cQ0*=8ioJic0Xs?KU1Sc4*B(ZthM@Q)dMzsPg)~Td!w?OaV+fn%fmf~ -Q)%4gzx|^SGdw5Q6zX@aa5DvtBbURpYFGW=x+R&eJ+pCub4#0Eg^Hr#;licrlHxOTN8k#BQmE16bZkKpP4CrRiVxK#QQn-c>Pdd{(t@qEPci##&xl`u$0)LrOL-9 -zvWeS`7xBCmwE!9bb4kw11EJO@=IPKkNVUcc70m0ATZ47|iR6gg!qSGiysj)%4Yy}rqki%w()$$Cqx5 -Kr8)dr&YV=@}|2t!qKNvQ%FZb`2RSz@%Y!5{e`xuFNZxF%SmaUI8mPE_)FDB_X^UE{$Lt@g?76quQE` -ifZOi7M%B4Lc4&<4w|73GbnVqY!>*=f~&6XkQW`Q2ZuNfN7}^YVx$WlN>CsmK02{KG*63B1p)E#;4AO -i=L&h;6~-X&@#o3ntTk5edfzR+4FHzSdAbJ$s(@u0dmZ#%-RK8d_`ea%_i4gCG4ZYT50Ze3yE$k*n -?>HEa#0V+p5C{mbkhF_N~ZyYy2g!X6&F-fm(`+{@qH*J9r;ZwP}^YUNXxBEXJ!WV_FT8=z53XC%NAl0 -53tkari3`p4b}*#_^xpGto2}u8+jS~7Kz9I -qBgKPo7$tm7fYqTK+|o-3lkC)s)~K`bFb0DNS@9!Dadyjm`0fb7T408h)|CX+mcjnqVin~~ -Jpzy_z7ill-rXCI&r&})a7@aUZOzti*>BrE5LT-26*xH?DZp{D?Y-$<0poGY{pKm7Iua?+8@%FWr_hV2ugScKFrwhtOrON*Ag6o2v+ -g?>yHD^n1iE~*u#P@@gZwJFx5ZhWW&b*`F8U_2x|HLr!eao|f_FPjm%>R$2a$Qjrxll-X3vo#8cT8W0l6Umu8s3N4Iz=Kta_*9v|6(( -I|Dg|0NJ}Gl>c>`?jg6>E2eq&T$4T!b#2GNeW8e8f?|Y{?d|bQv^@*-~j32NEJ<08}LkLa-{Suk~6j< -f9wSw}ozd;n~s&F3(%-$rg!ap;XjV8#b^MS-EB -~6fKkM$NlhpqNIsLo)KKL(CO9KQH000080LuVbTaJqT%pe5-08kJB04o3h0B~t=FJE?LZe(wAFK~HhZ -DnqBb1!CZa&2LBbY*gLFJEn8aByjEXD)Dg%~#uQ<2Dd|&sPj~ADk_+E?RW=p*JAT)r;I<8@)WOD2cM! -NF+;ANra;R-kG5!%Sjq1T_9+HfEUe>=Nt}a=(n(62UCflGEtTdETxuC1ertHfX|$=!GDdDpDSStRZ2g -$&3NZ69A3i356qz-;ubklyErWx9e_Bs#j?T*%yX*zm`Gx$G>Ye6vAQhT^nVFi;3+4(UnR?^S&|vC={mc8Zxp% -?w@<_>dw}$ru*Abawh|iIA~@v`S7{swonsK|06sM30Ox6_%Iz>$Xhld)7w2aTGETILcVD?jK;LYF6du -SCzJnzB(OaQ^P7(-WiMMOgKrRbWmtl{!%n%}O4tOY1XQiPVhIdMkr**P*+{3HkWm&+j6V?t)a&^<1z; -0S2S;;Cuq79iQU~EuEzvFAYb{Kg$n9m_Ah4*}6Q#@4}=x(!j<~@Jw{Ks!P`KGgjX9TRvNFKCh?FPdq4 -i!JQN9X0AHHsm9e`qhhC~@C7oSvUhWA(Dj0o2{RF;^ESuYO}b+2zXz>4e|Me%FU7g)@ -;V>VQs2;f5A1u(j77&l@JG^=ru9yd)9?#nvZn(G>n9n?A|tDK|W;3q1k>vFrsbcJT-T)6AnZii;cxME -|1!oG0cjnRiD_!$KPU;*hbxsk2)AL_W*1%`tc0o(MKu~mPw`a6Y+;SYmnO -0Jsd=#z#{DzP)VS3y`P|6}n@x+azaAs2+;l-#1r#=hMs8}v=O`0u^Bebxtyj}kxBD97mavPlef*>{?l -ZL|szhmsMS8lR8K8`pm#*Cq4G^QHQ#aAdz5kl*9mAYN1JpsD8N@Lwrc`e -F5l$vswKNiwexM=z9hmYIE3v1(;Ool^BCmWv;q&KB(N7Ot~ME5V|)B9C)k3%8`T6{R*BJZP7w-*?CH} -OVzB4_HFI78Aw>p3O&7w6vv?JD$>2=QonOXn1S5rh@a*jQ9Z$ri?~ao423a~QH^^H9j=t2mT!L%)G|U -oIKv?p~pT*AZdiYROyu=a}kO*ih!1kJUh&c0|pW%Y^&=e?!ZMO!#NKw+3}Sy0q$bgMY1?_&9gQIz`VV -l130>=o}f!-#h;RP)h>@6aWAK2ms3fSz7>tTrEcp002`j001Wd003}la4%nWWo~3|axZXsXKiI}baO9 -kWq4(Bb1z?CX>MtBUtcb8dF5JdkK4Er{_bDF?tQSF^QoI0t|(k=i(Hzt2jnh6l0yqD7P4$pYhqxtUcy0M~>;mYhyxU~CPb~i<}%k6qQDH^>K+N5=Ex5AIDawDsJCns*SR&|`tS<@3KkfAjW^bvwP0rK)6Da4Bt8s8SMh87<3JZBfYP4ia_QUaBe(C{(q{+@s9v>Gq| -qRb^QCtjvx1N!u5?tu}LU*-B%wja-Al$>iCqA74LxC(gtN*1`1k-fngE?MvEVx(N5~c0QQ$uF2POm9M -Wxb4SnpqL(tSx`}l2Z5aMInPgdBmRSZa&Zi_NK3=xkXQZ6zTtp9H2`t7i!cuWVYdWXj&;L!`VN17_n3 -g&Zvsnd1X?h{P7Ns&aaXOuaTPKgwmAKhTI7GuRRxn4m9_#lGGu7c$snvh8_hVU#TR(o$aJlD -HOMF4Q&QXHl__&86J~za5o;Xxo>0oOvC2*GgU7asW_ls@dN2tbr7b0^x);<-Oc1#XGa|_=+?1uzBsD| -lCCRyk!fHM1#DZsZ*36lHmc`WP;wN2o|DR92)lDzl0Vu{eMY9sH3D%~5QZLbUX|Mwf66b7){5)FK%PO -{+bF4{em;{b%F9kAcBJSnjT-%(zYg&0=F*#;evdJx&)P=BZUCK@n;Cv^R%fuZrFBQU?aeff&#>~Y|?K -t?D*>XuVNT64hz8S{y9}4lvbMi?$(uk#WPp3Hjn?3FAxJnj7P;v%?;KT$ue -Q#KJhjI1%mo+^)}PRi~&$lBjV(smCznR}Iwqg67rdj9nqCy_5XMlK=Z6;H>*+=?583`WC`X!RpOhNxD -pg{Y0VR7eF-1(S*=HJ2@=hM`Cynqcg)!7j!wKd@%DWS;SrD7C(7>rRIeejo`CoUa_3r^xW{PE`ix&d% -K0p-OIzA%j`>@ef-A-!O3&s<&uPs|G7;#1^>w*sAGh6b9}!9h@MzKAqz55T3n0CX5XvB*7$OxqYgsY+ -(^gUb!T06~LF5Kc~1kA6vK98AD}k4sqm`agOy7;ib9)`Gd5)oI~qYY&2-OBotOlW{A1)B>tulYnmKBF -BbCsda)eflym@4maKQ}z0=z|j>3E+B6$s9rYD}J;u>;Sy2>{jxmhg4lk~}fm4ibk!pOFnp~2=!G3~Y{ -ExS!WxRt+YKDv#kvro7F&oF+sYg+0X*(5WX?UxU2D7Z;L0!RJ$f6QwRk4}Dc^!SkreW_CWv^=nM?|XC -c=#j21?Ccgd%q#B%YsvZb!9M9yW#pHscn1&i`6l>jCEjh}N`PZqlnC;DE!UI`UDZ|UHPY4w2SwQfXG -2Zla#-6mkg>SY%@uMYli8>SOLenH^4zNRmNk{b2SJRP^O;Y4zYmak4p;5A1YMU_c`i0u`Km8E-yg(Yw -_=TeVz5XRd%6ekzo{H?Uf`mJaYZDmd8&~-$aDPTxJ6s7qp0w+VZe@z4>XH?kvI3)rjbQk_OMb_bqk(X -a-FwEIzmov&h24$|7aCl%kR{*RYizF;t^1MPa~H88?qELG!-&KDh8yh6RN$a8@(#!&iK{^Q6*|Dln1A -K_(X5NUpc+YU2UNcwI=wC(CC%M)ETkY`I-wUeuC}RQ&*AqbswuI>nC&3A4XYtx|UUByVGVhosou$p*s -=vq(~c?ZxU2fW5TnDcVw1E$?FJFiDl;!OK0R9tGm^)i+3|a3>$AXa`W4q8G2P4@u4I4(X&!GC061?IP -}M9&wYeqsq7tXecn;Kh|c@8BB@s>P4@#psiLaK`~A2?iL0tnltKcOwH2@4K5rU|NHB+{xi4YC`$%e}r -RaPM@d0bYF+JhI(h=Qw-gEu|Uk_wKGP^fhB!3E%#=bLenyWM<`in^42SYTu8zY~kxDIehfg)FA~$O%b4m|JMy -^JVXuFt>KnUz#B)9Na&WGiF0R71$DeacbAUdw|J18W8?t95LNAWV+UH^9a!~ExkKPE?gECHvqR+YNic -){#!%xBW<(-MAyWxsN)9nc#8oey2cd_t90)l4BC;4W4Z;I11ME~L0PTb<8o6q3rJ2Q1-}Sc^u~l}t(fHIX+DdUxJP=_5XM=t@Z|C_4wL-)7a-yn2M;r`T2NfEQ*u$2kS|hA^zcEA#zPHj<+lMy3! -{X_qX|RuuB=tm3)9(ayaXQF$AyEKv-Q;_d5u=KGYw_trc2{;v8ef+!k|{1CQeEZ^h=z%KOhXwOnHms) -%MKrfy{EEkN0e|j;1q5EVhccq)ASVN-n0NX>qgbq{B|QPbdc?ofx0fUnAf!4$6RT)&Jc=R6saUpwn76 -MGzz$&N(W6CXzeygn+`%1>sNg3BtPfzjFVQDjS~)yAz0}~0a+Yh5$x65b`eZ%A?a2YE -oOlh+zTS#{wTiXH8Dp?xrWZQn318jcTX~38Es0du=nOkBnNRAf{8AkiJ2t5SOT9_O=RFQlvgq~Az3mN}$*-@%ieIp= -@dZ7NPijttGf+6a;&k-YEO*iC`bA7%`uYMF&?KbH(UEb{bmPAOTUK1EF64|CRzwXuN5i`F7=#SoVelgVVIZX4-CgID&Q -pTUobqwegIyvAyJdX7HpW^gIUiub*PY5GopP6|N6EAmrs$@idMsFP-AH2;SpQzS#GCK&O(h$t=OX}Ec -Za>B~R(KV}K!^XI#W9O!g1PkcXjFGzK4Ph7tE&*3Yu_WDMO7JpV#Uzf4r)Tf-+T?4gWrUz2!#j$)?fr -q8yss4>HajB7^;B5Luiu06@m0|*ETw6Vgb(Qo#oi`#}~MzpRTy@x1p+t)zcRc-I=`{dL)KvTmwA5ID* -Ndka|xXejpG-HOWNfCUht9oC;1~S}y4Z*#3pbM01|8AOO5rrXz41+V#2+l6%$ky^1#ursf$i2%u$&Ex -RTmZ%KuQ?}*R2UbY?nbGagXmC`lkx9SoR!3qXeu2Z~{aVK)abh@uu{;dfhjw4YA#Oux2CdN>etN`_?m -O2X^(-a6inf4>o`EaCrJ{ve{zhDF>SjK1WH}(CIXkRFJ+S=mxQA=1N8Xh4+eYa7c#m9?@^Ck<`Bl)@( -$+g;v?&HcrHO76tP+0d_*3pkF2lQ~Jf+Wa-q@6aWAK2ms3fSzCw9=Kmr;001H{0RSfe003}la4%nWWo~3|axZXsXKiI}baO9kWq4(Bb1z? -QVQ_G1Zf7oVd98ihaw9j==KVYcI;l`4QN6KmvG; -qm@3)Vz0P=tYifx1Jzav^qB#`(LC=?1v0Es_NPKUm!lD7Tza -04jS<%4R+FRa(}oIGxsAQdLR!Yu)s^RWvnJ`ORyr@_HqzoC93bmcyW_$zWky(YKK`Vnv -|gS!9vL=98i!?lw)r(L@BZ)Qu!>gI@ -=$u2L$!c-@ilR<=0<-$NFYVY=)U4BJtlslYvm9j8vmbH+lcu=2n>td{Hv-qGa@`n)UbyKKb(4k*HNAqP>G&2t~!JRs3N*u{RsA*e@KL|8T`akA-?G{xxzM4b5P -UP9RW?udMa+HIL1(uG1tSk!CKoTaBU*2}4@U%!^cJ&d713;>CysSMLpXinfzyXGl -&=%$kbBoGGBQdG9?X;h|u=pu`h*tC#8QU-vSn&KfuH>x|8MuI39k}b|JFDFS?HC>(o5Ja?j<%aAy6Ei -$b6xpwtN|4Iafc{n8Q!sn^-FM&o-4D6wPT@FO&IHt%1yiU94wzqnYB2}@j6J165C^D~RY2 -Nk4%5d^ox=D&!OdC}FO>B+T-$0e!oP2{-WjI_*6cBm0x)=W;F8l4Htz4AB{0+Tz5Zj;;bG$=*uG(IYx -TB}U?WRbfw;@r4b|DjRZQ(i8ffG6{hIO8(!}~ic2JlD{>aY&0cIcbS*x4!?f^f=jB@v}>3eh7C-v(u) -c!|1cdfh;eK~7kj(K)DRME6|Bn|qw(ZSKoys1S)brJf>ZEhdvZD`494gwqxKJ6+>%g}=4=C3OsI-<17 -H&Qs3L&d(B6MD{*QlhgodvB?<*{xn@KQ;b1u%?KX67HG1_=3OP-{m>$emLd#G((sMymidenIf6^db-w -IuA9!{S5TXDh{eQZNtfB|oXmonW^T7>yn0C1??jmz7mb$z)&!&0v -++$>Fp~(u?)(Y10rjcUnlpuF|EcvT=?u8$q9)fmdYV^G(^{lH<9C+4Vp-d2xV)CDg3>MpO3`D)J>b$T -}o95kzbzN6G3@=d6F-OqRfy^%S54xC*N)IE?2X!DOeV0fI=vnM`vFiM_J~tUlJ3jj~wf$d!m?_|M=$Y -=gX^$@wa$41H_Q0i#jsYNv>Dh3q2YwVkZJ#rB6`QR=)mBsK!2DsOy=+NUmw-*vi_IXiqVuOc?EiU2*# -K@YbVcM9b76cu2PLvPLoMsi=YGm!yZHq*b=B09tm2!cN01i+n#b+w{&d81va6;;wQWMtPm1O~5euYN* -$d@2NBZq99-rP385%epy|dQ=#i3Kf+6s6wrY%*vMMuT-{-bDM_e_VLLU7a;I7w7fhNq<6zgJMc~ -2Qk<#5ZQ|9I5*mIVRc1UYi~t!)*{wj9`uwZz^h)18!>`X9_tq3dLKj5=D(DN`j-8|If(sCGehRDN9yd -D|6rGUjbHKdV4W3Q7UQ9u%*>2k$w64zmt`?_W -!;}ja!ku#!74?{~dq@4ZyW&vspFh??92_(%>!_uVEXp(e(THU`|99{z;2jAVKx(Dlfk(GZTD98Qsg!v -9O)^cMF#X3D8@zX?W!-n$5@a|vB_TQD)1*My;6N8Xc!H}k?l69`byJgW=uzJp{w%8e#Q`QU}4;yh>5z -yHUGwp|3Cb%k;7VUSvuQ5t6{JZ5g8>&bne4?z4AT>5%BLzyh5GfLH2wSijf4(@4!!8R2PfJSxZmV2>; -sP$~#p36yzrm7NFV4(e7{tt`eqFrZ!Af -MD4Qa9>oq->imybbXv(t|7id6JDvtBnUvDuCf|Q|SXfjgxY?o=* -uyMV~Qy@=do;&oX@3;|k+un(Y~cTFE_U)IT6lU!&1A6j-=PptHRAXY72I~NjY-7jpv%e*@!m(Svhr~3{Oo#icDlGcpS*`v)^u4dg&mG*2bIM(hctuz4Cfd0pfc|{aezt%90tmt%h --xUQ!0o67LtS}gM>#sMqK6fx~TKW4L-t=$KZrai*HX6lgnXXDBcbd4BAQU4kpa~`1XmY;E^FG;p#x95#n1#i& -TOd+qKv=a!zbe2aW*8Lj+aNB@ -Sxj*WYc}uOvnU7-k#f5c!L=DSo64=AGXlcV1K6`+R+YFlv -vwMb`~+nm*;rs(__PD+RR*6u;aqM4cP57_dwt?M9+M(K3K$2?BEg!UB$i%9Genzex&EKo~4b(E@ggEh -J0JV5C_#+acC2PhUomK^&qrx#iASlVE);x2Zpaap{9biV>k^hwLKo^^@N-;D}FJxm7~ha9kWF->d5A! -`3XD@^zQztEQ`xSbLJ#!F~Sj?A?{g0CYClt!#)CUE)(bKf4N^qDgxspaXM24VvwA0(_3`Gjitt@Si6N -I!&_tr^pewSuuNnue+5(O^Erx-}VPA_d^-B-e$GR4{IXaRRgY*M|8sw?Zk#>A%mL^ghb=c&Hp!@3{7o60tJ5>G1vv}oix;RMG#MS_XMqo6 -LquxR8V(UD8BEb)n(ojlwISjhe%?|6=u3cQ>)W6mzIYKBR$sg@rd3(OqwJP#>sv0puqieM?Z-IOt?0oO#X*JS|Jq=?jrFNF(G*)+#mz6H>8+sB -S4#BmUmLr|g(>B<+5c`2C$*S&nKeE|jN6;YPpj9)}*+r8M1|(7gD`OR~)P%%A5zCupjz -_FiWzH&_g&%L5=AE&Cs=kX{t{Z8>@G#&NRpLFpI?t5dP|iqTW(9s27=Kf(q)z*2cMIf|Kn_>>+ -=wwu-sCtZnvEzlNK()E#6fDs@FGp54e7?xugd5ArsCrl4NU*Kxn0Ou6bfm`~3t$f({pSr<2t9Sm~Zt} -?EE_Oe#g|$Kej3l8x)0Yy&Ip$|O9K~e}t%=UQ)Kb_l=h&)Z~G9n7U>?|KlWq{9$UL5clzOb6P4N -?f5`N*071->|De3W^kqinRU8*&JG&&cK?=~6Y1->{YyE4!n?%1!3DDNQTU;ubU1pr!3d{}EGK2OYcY46Dc&xY5Z58F3eUF)#9E -k<{l+-FLA=IOB;RDN8)T1Q|>)hY9aK%<^RlJeU~+`320ukiSQhMKg419{CN}s|n`dX`=Zj3ixiVS%!r -cWgGi3bkkXoLNp|&tvy~ab*UPA&JM%W%Otq7C2KqH`HyD37rDCco7l;oTOUL@Ss#|+8if}Q_I+}kQ{% -Bn^S;X?M|c^h2W3Ras3Ie?o{g-Aglg$k(CJED)`#}~*&1-yuJr5RgJC5nz4+ua9js{0Acnz)*mZHSm}P_j -5s2v)k7R6BQ&;5+6Cyn@#g`{r=29JC6fJhzc-HH=v|kjng+8(|7zwMLAtp*}#kWy2u^}HAUGA3hgCw4 -wL6oQBF}gXO+~cAImIR{EAGqmAk&yyLWNYztxVpR$RZ|vrh-iLLA>@2{vdP-95+J~eygHBF{{;3+o!) -}mpfqjn=|X@nkPe)VbdWV{4w(V@b=y~VN}TlNrjJin4?Pk)l@GnD--S0NSs@LMLWrJxs^O0?Iq9s&E&82H!il*%57Q;M -wEs_k^loVie3QvlB=k-WMPGe|^Y}4YwBuX5paU4Dro5(2~-^h17Qp$0bnlrJ( -p;MXDWY$yC)2oV9s;oex3e|B7dyZz5%7bWQefQ{;0MZtS(iQ@yEkc}D8jfD+#;y)uPh&BFG$CG0%b75 -2uJAI{15*W^A1t^D^@`YrZ<1QwW=L7ZfR|VN-7_R;Qt1|FcAR&s^lvX;%1SxPs`%>m3_!c3Xz70Oj2Y -1z912X+fDdl$L>X8hshAypR5wJ>VF2DzA)f{b`|2#ri+P+*6XTggb^y{RiKgwj#XyBmb_9>j!;pW@?m -x)E>R~tQE;CFN47gupj>Ty<$R(-M*5C`@hBt?DUAFh@GMEsmdJ1p!^)*{x=UdQ -@r}PiGJ5jQd`AHXat}z`(WQNMH1BOqjvtkhMoT_-ub^`?@cRww^phk!vBb@!b1SB7kv`>ZG$Fe`z8`|Wz$+*qBJm)ge-%ENjwKyr;$TQT7|J92nf`4#fzIO+#+S8+9bXAi)^^Ql~?E -1H<2Igl}(E}0A-aBMXvY0wIt|J&8Q&~)Fic-NN7l6tx7hm8xl%h?)-5)rWV+B9Xnj4uQ13UIFi#<&0k -GJ`mgKjVijpZUg?w;1Aa9EGW0L!UxYjDUl(d~!1ssy9V3S*u*z?KA(pXd`M+$Ez2~+eg;Pa$sPR1@?L -r_aP=*8H7+AzOyylSIcB`^+GcT4!SP`WIt}%scz^|LGk6edXnWPpA35^33<=%0Zw6}IKv4rlD`gO6(B -WLz2kRTmMm{rk?6);F6(C77ZROT+$qA|vlaJGK`&U@HaUw?}=@ON|)@tQcVDHT0`%0Q^Ko5%X0S~WY| -4FyuD6EPoGxWBDvTQJ@2jOPtfbAF2s(b_I>iDov)zRqL2O|u;rHjKSyQma8Uee<(mQOqS^*rATxeY4U -jE#Odty@MDl-f*!(vxklGOE>iTa~yf@cm;_8j7WorG@%W_8kkE!uy0Q12TiJl*BUMAji%R0E6sZ*D2; -pl$S=&BSZ~)pD1mV?aSPrWjb&>b@0cV$-o+tZ*gA9F6iEWKh*-0&Zk_p6EmQe@4aYdeuJqt9J7~2Xy> -Dqru;Q&t<%XJ%Irny5Lba1O-aBpt{87o9MeP3VwStY=Mgq_+gKB~UF6&sgUbxcw=5&Ei89cfvb}EO1+ -J#@1M6N2lNuxbItxZx(+lM(z$=lP~IeRiC60V;Z<=DM8Z*ZfGK_PN_50+jt*>Ox0x_8!Kn`Y0q%L({= -E9$FjHuoH@vfCOYw*`H|ssKr5v~`Fc6z(2!TJ`+_D((7J)uTpMaB`mX3GBhuvWL+tzJuOSa -lJ)c8ir|%~3?u+Ox?20VCC2XAPA^L)3Yi0;gr7J$E9QiA_pF@7WxJT91vJH)&xX_4SbBOo1h15%19_YaAX7!t#rOjTtMX9WE -Z-v#;(!G!#X58!v&kQ4reAT7H-yA8Wv3^5^CtDyLgD5z+g8IQa>@g4q?>Cyib%nuj9pu;0dR3>gkvnr -W`S6$*>b|U85q;`tf5?1L~><@m)f#)Q$+;llExFL_xzCylPX!9d?KZD>5NT{8IKpp?=gJ2QtOV$=4h6 -Qbm89O=5g-AP2z2CU6<{ksgbW2SDl6j0H4;VMNyOI`sMx>YG_qZjs!w9_!3yE|hCxH*q)@j#dbj0%}? -lQ6=mM?qA=`{{HVNuC?JJ@Hf~zzHO_0-;bWMu_{AI>e@WUu2W)Tu_C0ZCHYxl3_Q1t_Be3$VQK<@t4D ->$jws*Kh;k!Gln|I8<=gZyA%ovtUmfbe#TvxPgNrqJOBXJNG4|o|Hi@-A(-sXQ0SHMYisk}TAmBKlTG -#k8>D1?#`&}-skQ>2cZA3B?UdqE%BdLkgnV7Bv*;9gJv=0;ybPk?L>~Vn&dis2u;30vH^z^w&!Y9(@- -V?D$PhyeyoXT4!4IZ;Mydtde?JaPW^c5!-%y@X_YkcTL2bkW4F(sWxaY;pZCsPESMQ~(AsN_||(X&}w5n@Vf0o;t#25Al(s5U -w=65qXYk4^`afg$RZYf4lRlFFNKP=xnwxBBhO0=?AjxYH)KhUmD7&6v-dnAwle+#g=pRN}^#UBDI1-t#Vgltb^ME83X^-rn6&9rt^Cv0ZfQqMm!l}7-H^Qat+vjChM1J4BU1%VPdH~5PEJGbI=|~UDXI(C1U9lD5+v^ft -dlka7mt8Gz;DzVa>oSc!E;Q6d3LC~RQ`+{G3YT6K`{l{y`MVq6UVNEF*QodiSt%?I)|Mvdu^Q5=Q%vM -Q*Rkv2#Tr(_FW0yhP7L#Fs*fOBZz~@J8Ll=?OCry~6m|dWsz}MPD#%XYL`(r0S^V!Ax(=665$~|s^e1 -p9y_{JW-@BPL(*f>ZKL_uxqMU!V<@_s^v&JQd;;+H&dY5&q7c#fK{qp7#k!-H%Y7Nt9Wf)wsh={RR)s -yLoxv8P)a6h}(&_D#*KrN=KmyrXYqM=^u5{zDjxF{#EtQCP}?XrpWmWq>MBY+^Vg}C6e^2L3Kb)N@lU -3%5T`puQ!kjvZiN!hIN_3hkt<)=JPH%V_z%^r4_0*|3pveX?x?A|Qr)`A29ngg_W7Jd#y3WAQkFvg(4 -MKSJ*m?HFqByWfL;ri8Lp2RN2Z>!uK@3X7%BnfCDV#~f6wxR3G9~M`Umr`uaNbb)d0SJ~s8M^?L2{;U -t9OGK~hf{Q|yiL~bQ@N#OHsej%FHTbDk+yRh6<86(uoQ#ft}jhywfAy!ROs2xu!SzFie0_H_Yg!-86= -U~v0ZN)KEmiL`WWRgK@4#Xp9pX<9~18h(gy#dCvw7>;b=N1y!(lxR?-cCNWt^GZQxm8>l_PgeZK;q^g -`$WKAD8!M9qw($8o;cr-eo$d+PkSEy<$7+24UcqYZak=K(>)+I`G6(sPOM$y%Z;#ueekg#0~)J$iyH$ -{Q}DZbJtPTP3E>La88C7!gqsF57Wc1aW(^A6G}@+wC2|aWxSb1DqwHP -_Y`?wS&y6ci|0<0e^IkIgWrOeiQj*In11~5WJJK;Qef)Nh_xK;u*!5F6&;9|IZ?|IE*3ndcZ@Kmdwf>ANLVj4hd5uL{uS9`3;JrqcJ$iwV3s1 -lVgldtu_d#~W+KkQy*NgM&8$sQ~q$O^rs+IwSU6&;N@uKPem9r|vg{u^SYLF=`<$<4|N|Rkiu3{ -K0QS(WB!nB(~MSB-DgANOsB00oo-7gNJ`?k3L!C-DBQO<_TU$wN-lB;K}BxD-@-nH9~K`F -$p*VTkbg0_VY3i^!D{ZzxGybmR^lM`P9byK=4)rJ87$d?W95R@`D57j-3s{Ixt8;h}! -ct{I1GmkRd+%b5L(o4+6VufP5M-&0@h|EfDOToJ?+hBew!rTzY7iDkxfB|mBu;TPTIY#3S%5X}icxCS?Zq!ChUP&L`0E51q~0dE+u^Oq8u0n>UpG1OtwZl> -NZ@^}|^REsMxGpwsd4H7ZoNd7lU{Sm6QXNga7-5#2g*MVC$x!~F6^MdBg~Ld0Yu`0dnFAaeRH75EEl; -AZXnq5tBQR;fchl%fpc2{)x8z;H^ -a6^jxR;IDjy@yFX4XDtE>0dfL=bjvrA<-;)O_wa$R(1O%`eK7GSgM%iD>OWYk$zbhD-9jO-eSbQFpQz -(r%)VKnO|vVnoEuCQ*!Kn_5FLIgA%?QkE-jy6hG{-lqDgc^NPMjlgk7i6Fq=>ar7#x+^%Fe1Rm#H_&( -*|qvprUARu!A*26ZGfK)<)IFHtVlMI&?53oQmDp2dJIyLA$Z)^=jPmbAAMg=Rn&RfCE&!u30tfxUAdq -M(W@#WT=6}w(`S)x@+z|G18qPQbx{~#9XtkVz&yskynYpnUD<|jBmmMfcv?^mWnZ*Or(w`RgJ5;J`Sb -_Oq_mbH`$h?c=8;!qS4FEpq4@$8YL$4?TU_3>@qQzC_wGFUEVI|nP!iyp2G2DclAnEe;W{UC@QaIgiv -^tUkJN!Yo>A&Ide#)}F0UWtw<2P^Mt&NefmeSP=T;_;RihTe@ITxe$*0k`q*TTAam%WqpNR$2J -M(+YXU$XWcbyNik{oKrfoa8S&lP$=;aI -NpM== -I*)!qsvNgnT<8395X%y}M=NMoieymyuT5(nuYT2b{$YH~#wk_Q#Uc0J*z`Egt=U2*0Sw#Uuy14zu~7J -9T^43k*Hu+))u{rbj!E~4eS;zkr=r~h9@`zr9pfa4gK-M*I6NGNa{>XahqbZeV@Qm^nw-5PVB%LVsZO -#SyBr-m4YYvNLW!dOrMQpW0xLdSTF(6c>aI)DiOFHYks3}jXHS|5vlCommzszaLYHmi_^Hwymz%s3HZ0j1Z~>Bx@srtOoJ~X6)gq$|Oa|o@RuWBEayh) -OFL;$G?p%Oq0)Ayp2s$iwS5QmQtn-tDNJ6o+{R-Xl0UudJ27tB8#;N&xc<(ALm`(c1;>R$#SjU%&@BF -V+$F$d6r-yu(@oo2sa}*?6*nmO`{H6qinR81P@u`&mAKkI@NNxVmED2_U--{tb*-t*1$aP(meK7POlP -LkPPY}CSz7sGz|`6?{@ZYcMaTC?{QJ9fU}?$)@b$qLLts)xV8&v)=sVzB&`%ymnf0+@;X(Br&;#xz{G -q0iljwy(x_85cwI6tPBhxb9FW2VNgD6&3H1AXpk4s2XpuB}6$Jrm#kRs9*3BB#-A-%}o0i$Vo%VtwS* -aF5Xqo~x+Exr*99S5bSXDsC2+k(PV^Rzf2OND7u?Yw)aJs|`=($=5#fuvLJFQz21@Er8yfE?@z)whV+ -z=!g`=8r<;mNlY(+kGWzLZwkfiw2aR{eTZ4;;bmYeGwJr+H+qv-c^7@T7+-sUVo)PQ(=!NwIoe^WgD< -eEwG5)ER5Y^?O9eNoYpg!c%!$89M7QO~i4w35(iy;-m)q-b^nQ~+9e#NZ3L(-Huz5ir%~5oi6Z_ejaigb!~nl|FT+Q{DHwcR)Vrg0Us=gdZM~{ZB!XLd&t5Cb4>wxf6jj$mZ -c3_cTB{3D8jYEna;_+_Prz7Djl$~-ipWM;7x2;**b)nO+JG-RVsbMquHVPLI@Z; -um4I9a%fgCfTI5~LBL^Vv4pu9K*;nnpoiRGaGP%5sY!N0gT#jbte#PA}mj5e6P#S6>9OXS6As|^sjx_ -s>ub}(@hTlG4k?{xb9qrAD`~-P>d02RMPC8&XMfD&^A!fUI7xWGO;Ka*jkEU&8EqzMY7cW9>El23fVG -(60(wfx%q~1nqta+?Q6t7!wivJL(z4er)5VN;|JB+cK(y7Y3G}?bDtZuzKC-l{ -rXCM$ENSnK8d}!(ygU{6e8K2Sn0(E6W0(6!`Ei$>LJzGs@jp_QQ!mV4Hv};3DQ8tBm0XlgS@! -in3L#*|_&C>#oZJNo(q_MJ~1(KuB`aDT1do%RmV`(R9AzB+0BW7TTF|+_Z&*f9 -o)7$HMz7UNd%z?LLys0t9;-0BBj$Ic0Wj)z^0FR@UU^u+4jPzpI;yt8R1aW|xw(!#>_XBrG)VtHP>VYcxd-(mHQ+!W -E$x#(W5m8|jZh5vyOjS+h0_OS<7s{I8Ns`Ke(^$*)YUS`s;W~jT`BY8N{T(J@sM>zbgkqc6X=1y9dE$ -J76GdR4PWkRBXcZf8iWu#?2mR=fs~!xJH(m0a3tn=uIyTf0HtoN@{mXa%{=_S{k;DIa?AdobF+ImK?` -`Vsz)8db)tDWHy2DDw9YcW` -l?;iQcGL5>%EIDE`qU|qVCO6zQ{I2$t|V`XfYR}g7sy%;yiEa*eSS92_W?|=3hJ0zYB=0yM9>5p8RU{ -bO=TSB!?grAp3KP4hc9W09E}#H?E0mD4Z-ptKLn~FvM>n-6GxY@ed{?JTgt=r^{oPM7~g7(|h;+Hlf$*QM^@;pXlK&jtBgdcgNfl>jV*92RX8Rf-7QmS>acG13%R_?n -U^9hMdG?9bHy?^gQm6hm-eCdKg7#&W>snO&hnd%fWX^64^7fCaHx)LMI1`N=^x^z6Stxqz4$y`;mK(|_J1E3~O%Z;;Z$xS0f`NSjLRhL=GP -BrhWk5@PfLVhm$x!l_)TX$;*}jsSqG~&A8MoXnyjH;h?;G6hJ0D@WP6hR>HV1u=X~? -dg7jMJ$o5`x_COnuoS?{xA^s#f09;P67;EBx!c^DTgQ2~P?hR7a(&M;^Jz_7p5ZQ?9G$TEJBGyiZH#& -GM(gf(nFv~8@5J#J4{2K4}vbO^n5&!6Yd(H^hFNXW=duBfMbZlAGBGb9(7*n4blF3xFUmZc(hNxL;H5 -SzR_!PzC`0;;?m4>VkKs1HPh>X=>6`7%jkPfYL8Yr=V;hA(?}XhMd39qrC(6brOE>@^Ej+P75NKBN0T -Nd4*YEktKc3g(SKvEwun4-Ql?e-$2O0Y1#)pY6{w`3QS~RiAVvUZdSU7p~dim*n}02(;vFF^2LsHevI --FcsQ&131XEA)c#zd9wUy(awH8{}FyKtq)NkPFR~?-rfGVn4g}%LhDKrj54`=`x>5j{Y8QIrTjK*qMz -GB!`9g^v99~2tI>W*Rmp6>MTo{4mPEt_svqj^7o^zLZ~N@I`fZ=bu6}7z2M?;98^k^lzl^_tw(wi0gjFnKq-c{4#_fUV7?=oFrIH8hu*OPGEhb$r`B1=(V8!m(*o!uwkMP>|cMv0Y?(f692+^|N!Fp9K^DN6F*L -MXZ2~KP_0x_?C2QvOblYJ$$j-!J@dKaDS2R#nZVeFal*c69uLE2X&&STs-?RKNzE}6Ob<9KW%9fHzz9 -vI@0#~%EryXHRjIEz>t*MgJYNKVihJNuL$4YMGRnN2Mx_qUNv7J7f!*2En -68Q1!YycvF=1RoDZkPMMp`3)!VHvx~(3zxGi&tmwNQI?XT|k3Q_bs6>1N3F%L&sJx1Ls)B;P6NPe5Ia -MpI5Hn)b;0?3cxclP}QLsNeXq+Xx}}#YWln@@H?EPv<;*q8Wl>P!seW>&4q6VTe7-*jR5OH&s^NvQ5P -<=KNUK!Cp8l(1$g9;QVTawwsk|ejl>KNK_kZjsb=WX*ufK~20)f3x5X(IB@z# -1{cvcxp81Q&P5Ui*o_Uy+V$VkhlMx0~ZmVBPXfxgA-W>`5!UP|4D8^%3eQuBCP7vY?lB$`rBfyDMN5( -``WwyAZ8f1gO-7=ZNaI=p#SBT>%$m+`_K`>cw}~1uPJT8Kzk_<{f!s7K0-3P{d*)J+bnDj;z@lSg3~x -jleI>z_?ChtXK3dcr80*gom6@UGCXyGmm2w*Qu_sWwWVI?%XAZjqhFIDkCFy#eej74qnYs%)Y$}Z-Lpy3LI^PLQw?wmLA+ -M98+Rzd+yV&Wx#7H#`#|KOrs@k46lbtRa$-3nej5wW>IJ*f9sdZlAJ@+;(emXMD_Dx=q(ECJpG<84nB -}WB;Rz52%{%83JwB&~rPv&FjaaYh+a|W|SiYWwp)Rc!yYmzV(s -{Oy&`^EAMsWN->Ni7I;(21cw+(oK8?=KvM;-Uabb-FB#plvd@wx;+%3fXOZ{AdR|D47-&~7-FlV?KwE -$?VCD3bl{_ON&vFI2FMT$L;;s|n#8XBQ( -YMe$QvliT!g3?$P!?&!(bZjE>b!^qrBrYGUU*_N$n!dVh?htG=tqIMEq$gSyKkZi6uk^6}ic)$d4OYG -eHL7kknM=@Nmn73c5i~9K3hH7djP3xOxLA)T)rK-gJ=4j~}((8;Wppr@)~v$oT?$;#9sM>F5RVFKJlB -8WD`RnR#0*v?C}m3as&0nO_1XOTxd_>fg*K9ddbeWU+a*i?qyj;4#ojfCig<2W?FgjBRMy$3fWP#ahF --wQbO~i?W1+33dtsXVkU~*P$1A>=86&l3IKuv;shp2E`+NYYf|etBUe5kMz*46Ba8}81RfN0s_g1_;q -`1yAZEl4`rffn33Z|VVn`N2r9yejGo(MQXjvJb{wTlQkxkP8WAWeC3VLl_af4Ng;-@XLYR9|iOh>CXf -~yHOnp%fU3C^;8rq#`TJLW=Z)7iu3%_naXtL -;!#}N<9qVC1L7je5wy>*a=8=y8exxVZUy=Dt2EcUJo0{f>PqnS*CroV;tUy+lT3bUJ3WEtA-7$7HV{DT2iS0WU|dNmCYCbibx -BwMBBYM&bY=HDjcvj@(+NOj5uALN|jt>zX11XS(ngQdzuHh&@?`JK7dayNSJ}RN1rui>Du9E{+t}nF~u@@&@ybw;ItoztE(b*X78`jU*yXosV9(&9tyx~p5;pbgV`>S7)!wu8~y_7MsR{!X_ -M$(s5J1IpOQgOS!rR@!;t{6N8JMOXS%QSRx}oE5xdHuOj4@^5=IAPS)xp8zPb}*yhnsAhJGl2DC%uw! -}I&ZpRQh5Sc_nmF0M6tqe9G8i}<{zfIA6he3&J -!tTUPphFPp^cjldu4Q9MR7Z!%7;gh7UIR`0^c -fBetEv_xb2jY$dyQBWg@?+vThp}>?CQ -ziA2vLn9OF+i%Hs=E-i(f9^5ZYRRp+!>eYQPkXW>@v!>j`1>S)L7?l!yHQk01%)0cWZn`H9fP_0jim6 -Nm(AMVcyPe+6U;~oli_uMvkFmPfVQt2-4Zj*!cTw~R)-S#6de0T3+ZRq0h7$L~ -J5f(}am_ynvtq`d>>m^C7S{Ko?U8@goz`TpzcG}~f(mUE~3Ji -XF>8MJROhsN$Th#rptAejQyF0yDOuj`BoryykLBwvZK!+(wiMiU~u&WKlMF>Vg7%d;ofo?$W$PH-!+i -pPn&<$wsy8*pRW6KGL4)0Pdr|_}b(|N@1(O*Tz2sW^qbzMsBe+Ugz=yb4&$B3X_V8KeAD^_9|59^zKT -eoezhko2CSR?73e>KSauR6QSlRF$g9WmO{(vW}dhW3u8%FpFGi57|8EtKF>qb8jvrGyFR=689g3)4lm -U$~hRtZ+thVnioZl6DQEK|FD$QlRhMJJpwAc|Ad?UX>=1gId`IgB&RZ^cP-eLZlgSPs2*7;s@*jERWc -G4t(JUCrs~!?vhm#`#^kbkR)&pO?x|e*`%)i@xBmd1uyh!&r8l33SA?2)wvBXclifY&+d0HF9k1f^t6 -pG8Ib4$Z#kF(Lnth?Io4HHT|Sdm;?uSBhOnWFgu9I%w5W@-$QQ?^&g^;`Ndlx@LE~t`4wq%`HeK`(oe -Y8?11LyR_A(eeW{P6uBFoBWbK%Q(hd%7yh~hk^7k6$u+SaVOnaU=84(^A!DM^3X2B1$FhkD;Sbl2s1-yD8`1?2#kXnK|$XTKpqANqph;N$lU_FBIS;OJK7e?IIo0(&kSN1Yv}s -uoois^1Q9?)a{tGWVn-b?@SP5Niu*YiIov&=wF!uFA0hWFa$KQ -PJqyd;8s%_J}yN&iJhN}3bEPh#4OKd;#?#f$gjB{fsZwxmW!&Siq-hkNk59P-3YezwyB -d5Ra)9VsWyo?5qZzq|(%HqXkMt%fs-PHMdB7t?@6`LD2bk&PvOKmtKD7LE{K1o=TA}x00c}|ZmO(G1ATIE%2%~{PLVc1+66asWe#$vgbRM6M}2yFnwio2%uCC9s7M6{Wj8?mya4C -A7}P&}Ehe%ok@$`5{n;Vi10T63F^l-P3~F0508b|i;9_r}c}%*x05ijTyoKgbSV$hzV5FlN{`DYQYEc -*Q4qALb6jcPQTv!B0``Sg6)WnRR%$mm#?-z*HXPBozq)`q^6K4nA&)IWWtdPCW%O)*8@A@&y!;M@Y -nd849VPvSq5gK^lKCY<~G~cdlFYP}+Ondx&Tpsjpf3d?mx%L)bPw(4UYvLoOhO#ql%1!|)doV&02_*E -Pt8a4MO{}=_g4^R5)n8$Izln8?O^EROCnN}GL$?EH%5U(HjPYVORLYIVMhGy)hRR2{0Lc7y_ANZgcSx -+6_p75Q_E#6n1hPE6z0H~L?HdkG3*Uez@5mNXv`_*gKYC|0rF2X#u5w-vE@K -LHE#79p2??9S4sTz?@{UM;PKb^1%bJas{hl)%jEDdA17Pvij9#`&oE^70~b$f1GpMh`7DxkPvO#3d{d -bSUfClcG;+MF?-^KTj5$BrEE*Zytm+gtzw3k7AgSH+muy`CWRlX)QWY8VpH-lR@$WKs*L5?u$O!bncA -1BnQAQFs*dAmRL(}IARUVyT3HQ+{GTqd1{_|IDYd>)0NqzSxtUPI(sT?a`z5SO;8$|wKt(>qhGq%<@2 -WwR>*(paM>IfRIeQ1tUrA7bzIK7Z@#vlbo-;f9gW?Uf64mv!xJb -d`cY^aQW(dc9Vhye_4$<#%Az#sTq!Ve#5!e~Q~grVuTB*I$<=cL|<2e_^fRe7=1uVUmX-e|-H!e6-^n -WpM8%d};vXn&@wN#2!dX}2qLG%|D9gfW<Nn0iRXgvC{|8JD10#Lk$A&5Ya3 -C95ch(3mlPQz?zxEx$#5I|K%I*4cWkLV}$49`ac-n>D;dMyAsAX_4`9aAAy)IJKmv9tS?W9DS+Bd5jlfGC*vT3SU@sP*sJ_Ac5UJ#tTy};7dAhNh18G -!I_gJnRWoU`G75kbYLwHP#XlAkbN5)4_g#@rvRoDGdQ!?sM^dY2bcT`ff_+kp15E$x5C0!fO9KQH000 -080LuVbTd0s@2XZR_0A11m03`qb0B~t=FJE?LZe(wAFK~HhZDnqBb1!UVcx7^PFK~HuXm4&VaCz-LZF -}3ek>B$xxK6i{Y?-m0YkPH*sCjX&xn?(cl5DqkmLgLWBymjF$=wfjh)borl~h& -!j-9yeZIJggXt_ecAn~>b98X<$Ni&&qdz#KD<_#TfF$eUB+Q$49l@CM_=k6Yd-dk6E5U?;&mzH&ohNU -f$?q>;JU#jT?Fm(6%n~mMq0Aj<+nr#(;Bn%dv#TYK$9L{bVm@~k!NMJ0uxQLW+a#u%%LaX)&PUkJ* -JuJ130JF~QtXL+W` -xCl0FBZUni7NmnoW!1 -NTN)RTlz18kvJ>aQoF&j7k~R|HOZP7Ka}x1#pvHT1j4H^5ocHUp-dCifM%B<8kD=y3ya(>#K}(W2Frc -9#^U2QXM~J$Y!FY%?n0y-KFMJ-sJdm|Z+OF$%9K{@A5(74mbC1)OL-_*6)fkEdesaH4fY**&-?sbh{&KJ1>GV$b?;bdNgWdM&! -Tz5Id#lskL8sU5yJo!8YZcMjI-P|t2ffpi{r~!>zdsFlEWGsc+d)fD3+{_3Vex{8ovx)V5nJlEeEigM -W#wzyQ?7ddW2r82tBq~xz+V@it9JkN^nHIY*y}^hccJcB>?B}-KQO$k -ac6+A}9zE>8A0RgOoI&qy?*LFZ-s|+bh)N0(u15RQhzmeeuYyu&*(WRdO8CR+NWhE&GkVwgIpr{u(#S -R_Km^lBIP+8_&WJe+FGg;_BVg)@<2gxC!T}%^@YV0PcE0`Cj|O{C3rUyVN-{{(?+P=p``=j=q`4ip4pW1eY{kG9EmM%rT`(m7T)nGv#Dp -ISIOBGLLGM^Vigp-|v>6X+n-K1VE=9{`2LZVB95Zya7p5zd)2FA@^&l7W~PwGxy&nGgA|qQtwDZF7NJ -!;GA?LaAxY6eB$+Gct40 -lFlPIzCt4MHxlLOB`04X#`Fey$k~Gd&AM=SV<~(T~mq-mUhqG8N3sW;P4R;!hl3X#sv&6e8gKfe2oyZ5_Rw(@ZQ26 -TbzmulL(?yi~eEkF$|=mRo94C8vn_R;=Ht<%~qwfYlfs?8DzK1ufLQb7rAU_vk{^^W^u1yvE9J!Enjp -0d@HNdZo40_!K7rECCcA+bG3%2o#x7hvl_y13+1omxw=PBP0JB>05b3 --gRPc=}h(|vnq1mHc)@iX~&b*B&H!)Kk8~|J{?s^G+X1hTd`u1O!(e{Ry@i0=SizFPIk;{G!8>+HLN4 -0xf;S&jUqN~9Tyv*;?@yi!vjnbJlb!hn$yYifhL&G#LLDcBFmK2@tJtom>yI+Z=sdVjl6{>+It#%d32 -=(lqMcAT8-k>XK@MY+Uk_DQ-pv0;ZNOIgc$ypb+0msr=svW++1!sZ6vc=L4YENi#7YT4E#n -@4omU=zB(dANG`aP{E9>d~Xs@x#^W{x5s?cE5f2;L-2juUuU0>EjH`f^hKh({a1g{mUt)GtMz=qHMGZ -#0qAD6^~aDPdfdP0o(l%q}V38w07B$r5dF{2mcaC|*f9B+&XTw;s8(gp_t_r -j1bX$P3LSFX2sv(whilP9r-&ZjDbuh8U-XH*xiPdvFR=IH((bHC94_wU7Q!eSQr>>dJR0()1uvt%9?+ -dxIM+r^H@I(XtL@{zLX{AB3`2?kgwh@q?Z!36(?-h4FnEJrb|E8K<0BouIPzX00oh!iK2hk(CP8L9fz -xr}R7&(^y*r3yUvR879;5v#(~U|c#^^P;2%)3r!$LX2QBA%jplQqJ4^M%Jj^J2vZtaxzY+$ad%)*md1<-qYjmUmm{y{pzp3bozr&gB5|;B -kn)*V$V-dM+ecOu^~?`@DM{DX9OZ@B^COp`Q3FEUIHK^hnh>rlp~n*v6t^*-`}H-$DNotx;S1R3}fsi9DKMrF`0_ME|#HkJ%^z93#zg^8a+cBxn`&> -iI;2Yw=-Dta-OkQl5cP;Qq@_reQpFeiX9+15$Oy+pC%xeCSn11gtMWXcjWu1^W(-a -MbS@u@vNo3lnxqR!b9f?%Pg*^}^Fx5ro&$oDoQ&GN2jBEv9H>om9-d>kPvl^;EFvY0wqW68brOcbLcp -5xd!D>N^2L)aHeUImkZVg5mit<;bj)k!(rU9F$1p(yG6J6z@UTITNoWD)l)nLN2EdbC4 -p4Hc>--`aR%WDj@}==7-vC@KZ9;%QpqNJTZMY^>Ujp?B)SM<9%0+_cV>ytaqt14*T^wyEBZV|%_5^Fq -=$$r^OH*{m%VtE!GA$~trnL2Y`}Vv&*eD`cz2n_fHT*EgKs15#ba7M#i^9jU -isX{zdG$6A!k7Y{wX6m4nnh!1o2|E;)*HQS$l$$UQZuokOW&js9?a)iWZVGzHsOmG3HC`~)!AyjC6$B ->f{dB~u0+D-#&?QR?^c1m^F~=p)bGn8KSU#BGcS=d8ZQG>x_rj)Z!OkTs*}h_mA@^;-vO%=sNcK}oYB -G`V1OFYe|TY%i!#JlEq9ae2!ZpjFg+tOaYPgy|!L@Eqg_ARUp2! -YdeB&yhOnS%V(KMr~e*$Ik-PzbTAho>bo2#$h}UgfRJv!d;0!yo0gC(^{4VVkc@_3+hMG|08|k`<6il -qwq1VS&g9x?1zujO6zgl`)S-nI4v3rKmnQn%hMtJt}}sV6F`k9ZtR3^BhP{iD}~u1YIbVe$J8^h$0Euh(I3%7g?O7<~o4WtbS<9kOy_Uj3~(}^>7lib<+$M3fnpy82uio4z#X(%54ZjF$ -*Rc%m5lRPs2RQ>p_|6P-w^_IsUYy@6VWjUJEHQ(!&&=$9RUR0HpCO9A|a5f`bJ_med#lK>~|&v+0*P_ -^=lJu=2IT~#+i8n^mEOAbv22LZEho*IqfG_5$@}ZhHY{5kyT-JdBUbgV!tiafy -rr2NyadAB1TBi=8UjJli%Y+vfQ5QF3EH;ZZkQC=FrBzlIX=rR57fUY1a^i^%g0TWJ2uS2(TolK+D2`u -I&((~g^%^)EH~4Z*b!2VU>|N3{^c?s9VHR?~!}+FZE}})_R#QRej*TF#sD)HVn|t%q=M%3lpA!|B2Pp9Y=1JjW~Fntt>|r#~zte`4F@I5@rhr%LYaPaf|-N23xm -*WcJ>W3%#~Wb$AqiClNO7I8|0nl=QRFlV;3{;YZ}{_f(r3Leoo+{ -eYA1ItP%1Z`8C9Xe2SF6FiWCfWh63*^gW4&RGPup*Rh(1^oxiZzBiggadNTKZw*%cXmx3D?~g~H -Ry0?9JnY^l!>wXFt%a_|Yrp+{|D(Ix>wj9MW?}#FY4w8{|4&PaHOMdi2h6%B-f;1;i~R=EuY*=&ZI*CpYV(G9q -iJ$6EWH)=o+Pu(tL)s`f3a{mzZoUUjFv_NZI=YsCGAsJB)A>!W@n%qMkZzq2Xw>j|~mNl_VQO{MCnxw -R=$>*=qtQi`M0VaiUV+hgh`5oM>n?Gbeoc>0u^7JHEwUI{it#{zBqQ#C9uIbWevU0CT!rMzXYeV+`&J -7@^rae(6cg(@i-+L7v@koWZX@ZjJ8-HDh!IGl&3Uy+ipk)zF>jU2g`ro)jfT#U#9XPyAbjz!RVl*Oci -n8nvf=?!`(xs%~-f`=9fSN?qDp@nmS>+~prym%%feT*nwBH@I=IRlvJp96r+yo*enq{__s1yltfX3(I -TOBRNWj!O_3pW(dZ2zMhES4TESBm=3NwR92>9qw=vh_B2pblT|Rg>p7{zo0T~J}i#vd|0%;uosH*___ -s$6C3)QxK)1%U)WtjEu`O!)xW5*db63z?O3yKVV-g=!}sf%wUkV5Hnz7}Yi9qO0llSrzNi8HmoY&?_M -8O%xy2e42K8WdO5<7v;g5{oig9gIakpS=+t|7rHoa|Z+D%&GN*h#w`xUHhWvO4<-d45sOIqBjhJG2F+ -ge+};>W5%=8aj^thb@5+RRkpfpATxIh)(kXxo`8tW29ZGphNjhBUXL+F#J<$EAi;?Z?HAv<2-dMgJpssI~o1L{lwyyY-!Tg=9(-OwRsfNNquExk<6QhDEv|BYY$X_&A32bF|P|d!HkpZk^0WGcVG~&?*mg|% -9T(339-M0GNwF=zEkS{1>-=<=BRdd@4H*c+})NN44Mta*;&?UwHt@XBTKt*YE3nS;}>2WJjTc`e2sC9 -3o!`-B_ZRP*XI@}F9+t$Ge-B+ktuEO1*vaMCHtzBHFl3hvsB~-Nc@Qi(T1^=WM_FT_e9yWG*8>mP_#! -|M8^r}%su8^``pBh!EH`1gwG~w3kQ188Lm6fO6m)v#)1!>KSYE5hV4wK%nuVB1&=I-{+TG{sF6P0pXj -qTI^(v{6yqpS_s`0#lDe}{wqa<_GR`qtKhZd>izUT9JK+6uElhq!l4?qWz0t5bh?>4|~@ZiCizx7FxS -Pd9)%G;70#8oD>8IlW=4B?NoRYx?-B>PK5PV`{4@(2(~o!2H2uR>l=iGBty-%b<8iRj(McCF`+-S66LUZn;h;wSM@1N>Vv+_;0Wnt(WWQp>pRXc!@kPUT6^gg^r4?Hk+ -hWl7Er_E^fCa@@s%U5B195At5*C>X>>lt+vJqm7pjW!G+CsHLwCEq!`Cj|)aPQlv$L*~=nL?&Op>aLs -k3YH%q!vWxs8HZ`_ZFEPN(he!Yth~PPWw@g6?aW8Sqxayn>{fBhQ4T@oaVCBNf$|w^%?mx(;)vZr54P -Sd=$R;(>wOORADEMFNu3x!#TcO9>$41AN0Yw#@Wu6ecGLfJMTF!U&E?{TUJicA=U -f7{18?Y+m#7`uR#}%a0JKuQ& -Fp#?lkps!tN3y0!nsJu2jMoo{@y=3bxCn1P%w_f@)oDZ0{j_Z9A*qp+7}+MQGim->#fTRgh#ZW*UiH( -L{-ms<`@8(tS6XU%PZMYrPu0X5ES{;>EC%F)+y#P#L?lcg@%yq@yp+G}d -3PMHjR>6p&*$X-8MGib?V+3yiB2Jkzc_r_@VT%xu`T`e-I2+dRP5s-D5cQ7~Xxbe&%c6R{TPJ9n6SE2 -WPAx=5B;9ILETV`3usC%TDQhG1sr&AwAyIavtfVP|7~af7l}p2&-p$hQk~7|ifW-icu2Pn1FqEfiVw<@L)GgoTYka#NQ-vv|bvwo|TM6e{Xf#D-7`v5&5d{4#C0ijf9S -Wqsaq3f)QgZ^s>z>IEpi|kO%F@Vf%NU5RAK9y6DgIrxEN7|cN^(1B9vU90dR2;-e#vPoPy^Ved=N$JR -veuzF=+~yHR3mAwNSwOwyH#@t*X&!tBQ2!=HE7nU9kGtQ^Ic%4J9Y!=4P%IujOG(79$rIjk5CkI(=S6 -))ww;(B7&)b3>fylr3-OUVLu9=;wV~IpZ&`!!{^Y8CsRvLw7_Q0)N?MzAW0)TC-v0)yRW-lU8s$#`}^Hy2oX)A{%x-A|qN=`ZgGTM(^bdFOsd^=LG*x( -?<)qAb6c;{8*3IgtQfjM$cUQk|>)8@YVI4IDB@_QL8mrV@188N$_WTjL~S -nda@2bn`}}r82+C2LH;kfdpwv1xHO{!DmYavb=ved7!_%31OgNy0`G2=jFMzx<#y})BgLyPOofeb<(` -9RFZG>KF#IEbyHl-E{QeDEJaTzI(1P*|Ao81B0bj}o6kR*nT&i{O(FbyNL9H+pjqX?L+R=hD@`ws>uZ{MUnr^}+Wl@CZ!24k@+X;VLmM~sL9x;ni&d|s*YQEI= -#gKG4MF*kxGEO4)vmLLZ41LIDH}PSGs0Uqq=>{Lc#qjdr&!1h^rzsXw1KRqdyV@^WvC~Eb#4~uZXOIw1M3ZYWj728M;OdKGDM -tm=hvDlR}58fo4aPXfm=+0@jN?uzIyR|b@Gz_ygiXqxpwOG8m)tPoejg6ciKP?E8ec{yrBxJd}HI5&K -s%#x4FX6=DeW_vDTTxm$}r@>b#)}cgyP?ZO$92K(BMLW6^oT>#LykaO?2>4=-Q-=GWo>v0sO>ABXCjTA*Lj(?i*_LlxZio*qh`9jYL=@$_JN -cBn%7_we-4(3xM%o5Nbm&fXf+&emERO{t~aY@T_mR+p8eU(4pQzx+Z5mz!EsTA|Wrh>ac==?J8N6|8T -@xb3k5p#~(YmA431L4>Cqoq}N>-v1e2Fq<4N_cc~1QLbRp=g~?!4K?#yvK2SyX*aX?@U(q=_u*jmyZS -A!QrkB=39yyby+#r1HhOTQ#cXwbjfw`N)$hlxLRk$>1Dv6~-FhAK5)GOAnieJUR%HkFD_Y^+;1x()TG -+P#Rqi>;cJbEUbks0Twzp9|J3|J88u?O(VP^*XgvC2(359l5XLul(eLZjIKnO#Pong8wqJ8X8Y!X3v= -@amEX21k1gveI5t2j71~1Y3$ACIM6~bnx@`Vj@~$pB2t--NU!W^rdalm?}~=2S`T1sIpl|T_5=OwgZ1 -=2`)cSDkg6CL?leed>BtTEy>Y#w`X8@zj4v9G!F$$=2dYw8&^Z`w;-iLAa_RYI&BBV3UgB&36Vuvd6{ -a@Uc70bq+iY83J|#d*-I%(GN1swwo5P$6?shO!dlawY-l*5v>PV}r=iKVql&nhzcB|92Ca|U2ZhZE&M -O8PdXtkc@ik{(mb@yziNe!VtYNKX_)Un|4YKeS%BMpnwZ+`df?|1k5`|iEt2m62R_usF6`Ae6G)F`f) -t9FAlJ5aG7qm<4_jB}9eu)|OKE$iWh^{7IfEo8?6@Wkv+Yfd|Z7XhP<{CeljR{zZTwz~wDZ@SyOCRc% -4_Si8Xfzft*d5=i`nl_f&y>CDU>n~xsxXu3g=bWkKhi9_8l-F{7vZcu -ad{IF~0r9wL>&(5*}0yp8|I6UmhAc~+A64u_|!cDn6y(yh=*|*H|ldlcDlm|(6mWK+=^sI4Qo}GC$J1 -Va~(Xx@hcJ9B%!+3W7~?@C -rceHjS4H@A!0vLswwjnR*ty8(x(07Rkks_Qrpm-)^2k_wyoq+vvQz?^PN@YjAd0gJ6P?MR@OWDQGnU{ -YCUWfp^b=6Tk1=!l#av-#^frPtR(Y6HNX`a?kxksR;0#It%M#nE25eesxGWZouN=C%XVkER;RRmv{yi -Id0*E^I-q&iR)%%wZHJQQ5GxaHyR%ls5;yONF8-}HQtI|mzQ!GS|9;yiRax-n_B%dv=i&4B06u>qU%w -iM^~~!lxga;qGK(wn$J!uEx@3r%&!_O;RT#*4`bp=k}>h#?Y+( -T{BeV1<7j05y}1S)Oq6AQ(izW&`xuz~*vKR~l|r7xA{cJY^Xt|9to6@$l~_|MTOkH_zS{SLIrS9 -0?w#QlRmCKb@zc5^tc=QpS{miIz3P5FG@Y@*ciJ56JQ}jN&nuI(|yv@sz~lkw3MrsRPCeGKtg*8+zo) -+q(>eN(Ju_SMW-%OT0|-_hR4YA`i_9&|+Vxk|{J(L6GqB&`};#->16T5Gvdugl|w8uxdm@ekV#}T;)xmMu^GEm(GX^mxmO^Eky@;bV(yK7wHUy~^N%MUbsH3IReE-ui~L^6I{sE65-JuNV7x+*|^g1}H?7vgHu0IDS0jWbwIkQmcSyaqIQn<>3wS -nZ_ixL$)}cyWbabxd>G9M)7C8DawTF(vumGL{1n-xw;~7=>wDrpex$u;}~Zs+@_IK}nwgPHr!DmiyfGjB#QQFstN&9(RmQhE|73yP> -Hl|XcuaO79&V=mFEh}wVjqWpFKj;3G(+PGl80E2d}o;a8(NM6{aj1aM<7TY0MW$^eCRvv9eg#BOcR4@+ZKb+cRC28RtMEq`GPMM$e0(amkUaZga(P)w1&sr6U38aIvc?2fwC)Lq -cNWkmG8^~?$jsY&DD_yyqKt`1O@^*RY|ied=7)ju(JF^2~DpIARM4EkV4m8>hP6*z -f(CSjDd0(Oz#=6Di-7sbCMi5{V ->O;HBx$eFR9&h#@Zo@U_&{VnUo<4QijN-i~^iY+Q4w=O?B -~e^gSJCFoCD4u`$@qTF^ytw98zGriNGSs^}$#u -P67Clugm=|%jkvH+in)d6wJ*av-9NNV~ptHg6jz9K0MK^U!>jIOP7E+jf&VW&4I@?(=oJ!m-u!Zg&B2 --w&dV{+M-D4{pulPsjqNE4oKBamkw70?R+0~$!TALQ3SJwo-I50TFb6-(taWro0xl>96xk}19;)zdE7 -7~vMD3i3$O0R(?eSsv1FiFpjw9FYS|BQ1enc$yhp$iZ`=#`Xma6&@=J(C5Jy9A+p{Wd0Xfw_appcFBT -KDk(kNomtLke#zQT0Xb{v{Y8ABZY7HJQh{ZdM##lL&|zccKD6YzA9a4HeO;;Z(seG-NCAE4^ee|=#&) -HbY91 -OB=>KcUhx_c1X_tixRcol+Ex?F{bLLO5&Ekq5K*eUe}`JmFt@6 -aWAK2ms3fSzCf`CZc2m008<2001Wd003}la4%nWWo~3|axZXsXKiI}baO9rba`xLb1z?CX>MtBUtcb8 -d7V^EZ{s!)z57>8wFk>sN5}?f4+3&*w+LUm-+@K8pMpVz+0)~n58oC^C?j}|(qIa|FFuF;FSj@I`^6kerJ+E=An2bF!5@`c?V;v%k{wVMN+ -&}zwP-3r$oI9<7D_IQEPTqkHs4NdEEj#^jbO&yEBl9PWbsZ;$3(Vgb{ZbSOw-&!a@*tenu0au5FHfuxa9@}Hx}q#V68D{7K;R)48r;x -YJj!|5|orEM*cm!n|+0Acs?vYtx`BFFIS1F1-J1y8BdlMpH}b@#%VgH+s@0eo5F!gbd4t5;@-rQAHL& -TX;DiiK;r>5Vdy)wiXL}_sX`m%$Tn@}T3W+iX#^}wo^wJw$^xwW+1)&Qe8}z|K0lqKj2+A=>?ZF68Wm -W=s~^VhGD*GjqHZ9*p>4DT)+yoncJAe5TC;Vsx}=2o?qwo5owdEBgQC6`qi&oeSU6B=!7NLEE`@7q-H -p;;To7xF6K5%Y&yO;kHaB*9o -{DRTS~TIOC}^sO2q%_D=5?5M_$lS}tj3dibY7%J -uo?~s{*2m7ZR{Et-@X(20;xf8XjVQ=7;au#P5625U)BP9yBZ7~w_&i>5PY0s{(UeT5nriDfr3rD>H^V -qYlXoi-4@nnt!a=7VO)Rrk$GrzndY;h^-!vBgf7@ZYpQ$o+4gK(TY&c*q2L}`4%#bur!(maOM9r*swF -ZjqjYKC$aQH^qEgbU}OrG>Z>GF=xt^*Uzsrg%u)PXzphJ;W$`|rUCx!#rk0#Hi>1QY-O00;of09jjm{ -_Gta0{{T<3;+Np0001RX>c!Jc4cm4Z*nhid1q~9Zgg`mb98xZWpgiIZDDY5X>MmOaCwcET~FIE6o&8n -D=h7zAVGJ(X&10EhQvy>qHXM^IEkBBb?nF=g)r^E&mW27QgS*j@avrSdDFx`N#3CgqeYE1%`sywDhVe -NYf_-BMJkuH(dxZDxUO-tp^T_m)6ORE(EJW9?(Wgm{N^5AKA^7`i}^<+4M&e!Jfm}hdjBJ$Txo)f*<` -Z1f4IF`p-$p#`?I2&j0a>-4)*^OP)LS(0Z_r<(I^N0v)F71aMIk@9 -UWB)Ga)wyShBoxgn+1rh&kC{!}NGU5ezP1kO?nwZWspmVDNmD1AkTHT#eJjl^xrQf&3yV;RS22C2p1N -G2kX{I_tXZy)N?2qlfv&W7Du#5r%3oz|{lZsR6#O=2&jW;j&s6*5K%p>dIv2-86}iW|xWQBRRq3bAlW -Vx}{|WY?U{Xm?QS0Zb~7?BdD!9z8e#aZgeM>AcRRRuaf+;=|y7_3@k -zVfO`UZ}9haJ7dN>gB68eb0?ol0OLSs`+O4hD}$Iq=^x<(YVzV^xi3$Y8SfMpZ88)jUa)z%7x|--s{_ -%nAeVFNOWvD+~!E7?$*f6!4m=-sUs<$>Ea5xFd@?l6SbID~+Z7+@R2tN=1)POJ{pQ8XvhxXNlX^&)&E -~clg-V`JlLAL(<{3-D6$?UzysK9p8(DwBIiPXsezJc!Jc4 -cm4Z*nhid2n)XYGq?|UubV{YjZD5Nkc_WQ$<};OK;*Z5Wf3Yj5xJY)O}o5+M7vCu^M9Jn6ey0Vu)8EM -zK@Gzu%caTXm&K(Y(ILj7@7{B@N`RwtcYhfWaeUsXt!_cWfg()Q|A<`;T8>T7RknymaIB%?^yM?Qn7- -INyT{(AvT7uW%fyKH4UM!(c7=1JrFb9Bl&8L)Bm5Y=?jYejh8>yZ#6j)JVm!Gqwm5{2^Yd!D4d*RS3R -z72-kT>+xj!SViJHxXuQ6h^T#5-X~^{6G>yMj=3JF_FjUEi>)6c3^v5U)r66N>+5c8NW*)%b0>Fapz+ -CnU2IQhmM=x--m=W*Y;&?eA0g5Hl2lYx@cIur8o%?}W4*D*wJsC)AIscz~a>1SGKyss!w=7 -KB#>cj6Dy?KIPPT-n6<*Vd6T)PC??=PzWf_I=~XQTWZ6!BVyNZ+7x>J8$nJO1(*-1DrpvdM|~x> -N?ol`@)dd^^8*Qb_Svk5fAUG4$gj<`2M%~g#1eZrCOLR*8*e-Ypq^oCNh}u6609{TWOYRV*ooeFU$_q -0(h~57g=Nph(BvB$`X`ja<$H-z?>}7e3QxI8Rj@wD16BhpD=zo!8mS7hk*dOI3=?#z&By3gmLdVJWl0NJXthchtr -I8I1cGygxS`Oh;mbv^h2l*qJ^&#TJVz`%Px0k^PwXF1O9KQH000080LuVbTbLREok0Nr0J8!B04D$d0 -B~t=FJE?LZe(wAFK~Hqa&Ky7V{~6=Z*OaJFJE72ZfSI1UoLQYb&}mn!!Q)Z-}6^o;)~)e>Z9O`bGRXX -pbilkA#Lw8E=yuQiu>=qZP(#k*M~yRIo$ltC7knDFP%dlAX+G-*5Fpo+b=9)7VT1>&w>Wn64 -TU7W&q!&vz;Wx~>(bCL~u6#qxoA$f&Goof79V=_o?VsZoI;2%n7UCVeM`z2*^PLR7R8nj{2V0zb>nvr -|qgfi%i{GlB0G;Daq^Vk)Us*{OfvJE;%9^Y#|#0(&ZKAbR&m?u@Z*K&rIc1>gX#(RnN+(E;UT7yz9sW -mC{AkPy&GVn)E90$t6{n3``tfQtsE=Mg+_xStwq)&#pHAJdlk{|KM4Urc!Jc4cm4Z*nhid2n)XYGq?|UubV{YjZDOX>D+9E^vA6TWxRKI1>KuU%^v9oL#8 -GVS&RfjP>HuHn+|8+5~NS#Ucn~TA~v+vZxg$yGGIfelsLRk&rcisDYvmH5?A-<(V0p;;TZG75N -}VKH~bjkj*!_wO=c?DwaHBqh%>pBoTSVK2{lD5cAxOtY}WJSgEOmj0##^Hzaf+@z23$9|X?{m88Hgnn -Ko#q7(^}vLSkV{`~aac}S9LmfXY%&9dgZsH>u`^t@_Y@cgQobN4?B*K%}!$R3Z`CND(9%1o7MxRp)GD;A@-j>Y1`lz2SBxR) -V8YfOn6c>MTgL(41KVY}hrc%oqW7->TiY=V%$o<$0*Wfep9Jpqb)~f;+4$0EtZe%MNO| -@uN1hsHa#$hQA2~e`CE^{qMCTKcd;=xo<>|Yet|`yxB4gv@j|o$nHy#9D@t$WHSuj#kE?G*zV -bOsWdM-(5o{I{fYd=MbTv1#LaYKxgwce#Yi|4#@JejSc_mNV1Ce>Sizcn!82ZMBTo^Ly8&3+{d1U5VdBi&j8AA -a3RL!cOHvnr9h1n~#V8u~L4xD5g#R|zfa1o)xJ$EC&^pr#F!w7iB|N=_7m1v#G_mp%Nygw;lq>;pl$h -XFe^*{r1NCHw`zWtb^Np>W*zwKN+n${<@wH^fK}+_#-Qc?%7-<(jEie_p{)Tq?if)~aF5r)vXK5@`E~ -{?z@I2)St(b&S3QsG&@xA-?qhn~lyDhFE2maAItKv2Z#5xJ+fZPS>2Y~z@0NxQEc6|=YG2%hESJxuN$ -iYN}u7axDl)w+7s}(-dsye^s65U0T>9Wp~>D-}cbN`&wIR>Xes5EL;)_cg;ARJT~W*bo=itSjt37F2` -qE$EyM0JsZ9mgKU#N=wb_!v|hq8?}P53zeagMUpX__gNjPid5AAbPiI(13su0nn%7SXWcGHijnu!_1+ -VA2tYT1cWMQx^!&plZ$c{Q@~fnVJt;mCM>Q>#!TT*qIo3XLmQd!zg6RFm`^i61wt`#A`^hT1_}_gjq; -8GX-cLc0TpBUdc*`YEDM(KC5I=F^(hc$084A8o@)odrNBZWTO-H_Aj(9du*=c>70k(r-VnhJ8A0CY;( -_K^g^R4N7+JGY;$*Sxh^FOttp^q|^?Jl5Fc@26r#w|KN&ZK(nyE}P4se1bt6D1jZ?+T|>j9tG2s*RuO -^@K3WMZjX5!(r9N;RVyzy@O6@^)-F(VQNU_u!!vbPGxZ$HoM{5;=JPl7CdG{%i|ym6PwHe|3^Oc`L|< -87{eXf`m%NjIISwu~35gYb9aN3GMOcXx-$R0+2%&VpuIGCK$C=$)Ua@(UQ^~`fr3cJ+K4=|*#zv6G$^O?wGB -00o}htgvwX@+mOv-Cy<1spyc1@3>RMeOm<7Af?p&ZMrKrr|&jxN8x-n}*!moeDVlBl|PpY{f=0zqr!13>9F%;wo_cY!$aTsvyb3zIq^41=BW+A|)u -EAHQ}jGnk;LP1wQQFbk`Cw1pu-m{PDYXr!RYMc#VM|{Aly~n^vZlTh#1OCcG^t~}{o_}{ -v)o={)sb5*@hNAK3K1Qe@Z_d3g&|xk+2TNutV%E}8#IQ|-oRn>efO?_iX;Zu#$7&r!bj`G -<=9@x2$-5QsE|Qc`UT0`WiQG{_IdljT63P6g~7VHYgLbL@xo`J;-!3Y$?p%L$>C@ei -F>MYYw!ZkhuswF#CE2ip$+@t6ktJO#Zf=;%C{cCGNaJ=<;_5JCw|#^M#LD0H{VWr}_MGglA(^Ufmc0*XT^y*+EK^Z(|-p9i9tF -*&G(!uZgBYt~IA>>U{3RDsWa3p)FNm)C53dS;TL!U&p -%iL2>F3palLebf0?G4f>`9ENdEp{=-SwXK@e<2j+Cl1amqg{+!R@5S-ouIn~221Diu|b)ehsLg-WNa> -mW{PsV#8vbUW*_^6rw4&Znz+%6eb@$4f+2?B-(aect*wa@uZ0Mpwv?ZJHB(a?Nfb#jtpW0a^ZD_~jf! ->7UM-vAJ+&cOa?YJuN9{7{?2UqhzzLm@j42hebX@b7VJezH}E2WbR*krktpq=dlqq~9Sop7r~Fg2K3-b?)&zYGO^}uSTMzU_dY~q5ceX)3V0kUk@7V^m==I -m{gbrcpSJ4c0Q~$TN=<8^Ux`6PYosnfe&>Z<-rBcLiXpTH?x>34!bL8>*>KdinBj;bRS5kL9zluglNj -iAwd$)GRk>f$y1|`>T=oxNd2lepb^V$CFq;CQ5O-lEGYt3~ZT;JP1o;ePMaz~|3D0iMbKqwCfq_dBq+ -uCeZ{&f^!Sr!U{ePv4#V_j&yE=-JET*Uv%X{ymq^tz`kF -d$&#YUT-lM>?AAn{lwmE|F(fA@5O+yTFm(D==_Dc{^i|sY3n@IAea}oqH}YZWBa_Z-B`V1Ql-A=O3yq -m>Pl%!s~I$+zkK#+o!4>NX`z0LYg|F14b(`46u~6KTk_DF&AyxXCB!4V8q&UMg0F%(H$eOqaZf`pVCl -{1f&P^XeG{a)T+zR7)1}{W4Px9Q#LtE17q&4{m18Z*Wa2n3)E((%KVI;I7ni$to(2g$*O+;K^(S=HtS -ZWx`<_y3Cv)dp8_M@a@OIV90>F=hx_x&d7`WP^Z+{NHVA1lhZ!thx2==a)izfPZFE;S*)1c;du0O13^ -XAN~oz?^y!)vYkrn;~B%=$pGe-b*-tcMRnw%w^U<#6VuoZWhJU-N0_nrX3h+F*h&4-LF%qd4irgVQep -{YSi}S#UY}KTt~p1QY-O00;of09jiA00002000000000a0001RX>c!Jc4cm4Z*nhid2n)XYGq?|UubV -{YjZDfc`kH$aAjlz08mQ<1QY-O00;of09jkIX1_ld0RR9k0ssIr0001RX>c!Jc4cm4Z*nhid2n)XYGq -?|UubV{YjZDOX>MO|a&Kd0b8|0WUukY>bYEXCaCx0ju};J=4Bhh;R#>hQ>cWCa>; -U$NQ7L~m)MpHQ&sAuoBXR0a&KRb*HPpXMRR(wv0(+0hRm!d;W{Lnr^$5RbWfkx7@e)o9Z<2&g -RMRm(w4$Vg2)HJ90Msx)@cioS3_h5gn&^E^4D0npg7xChjaEBf7NAOz{P5a6;uVa6^`Q|s~3s6e~1QY --O00;of09jiwC3Alr4FCYRF8}~G0001RX>c!Jc4cm4Z*nhid2n)XYGq?|UubV{YjZDOX>MO|a&Kd0b8 -|0WX>MO|a&Kd0b8{|mdEHuVa~ro2{?1>4o+o3{rOtWz(2QDUMzQ7EP5d&lJegEc&nFTTVIL3I1CBDI@ -qh2`;`LrgNw%Bm)RT!w-eR%XefGgF4oQ-nu7qOvU%8R*`2zm3oT=ubs^y$3#k6GeyeODnady#&V!vaxK3sG{G*O%&$1hX<<^C_FlQrEJ`7NQ2(<8mWvS+04h*(R? -=eo=5x!suW2>ep8%FaL4;=rnu!^5svjvmf6)`}uM9V_Ty#OAfA^pqi1 --rAv8bzW7xR&PIdt7w@J -F&7QIc4@^)e26+Y&6)(r;adMSR887E@A*uYUZVPWkwc>XXwvdcnQwAYU@DD|Kg@Zhg)HrK0JxJ}D9mzj28YBh(aR1ynBhhVB9lg-=fm5CfrNCMHmzfj&A -fVPpoX~QG+{6iE`~#muHo*gQz!JC)C)(!$&u2F-+)H+r?th21`Og+88L@0;BQXwZ!ki=HG4g#rMvGC> -e$6N+XPBMb&}@3i^*Gr}kcqLASFM3(1A}5NPR?l*&_`xqoJ|>|fwO>Dzz?++7@=`;2y>7)#3nX1c<~< -W3#x(Wi1I+!+=G$FsD59x=1l9@+i>FUjRR} -$fw*nurtpcoSGd2~a;$>Ju55}d<0gs$am}qm_QJ+*)tM3u4`F`(Q9=Hqn -F8}RuW%31%IkPCV@9F9flYJlthV~E@i#Wbq3RpCm~EOmSJv?Bl65JuvjuU5$nbzO`^02YXt=6TGxSnN -r73b#(3v6N%b7Vu#imoJtgt^@hYekvg8hh1ZZuR5QH;O~;JARTG&m=?1;@^M;he;dAw_^EG -?(UfquFdW6tk}9j^=Y&FKFMC1Y=Yof=$=^3cAYTZLu(}^89{;-hn)QlBd -hE@cv9ogryH&1e>$RfL*R-dIH*BXM#SRM8_C(_i+X-8z0t;s~|91nmn#v%Tt%XpP@d5*}6qlLe8f!BZ -oc)YWAIv66PR@(Hny6}SJNfJ!mq{g|E%*`YU=r}f;!gf+6=|f*Z&c|%qvDe7NE`g|PSs|eX5+#O#e2_ -i#-6A4d)Gzgi#XUR`(N|1v+ECVGZt8|>1B@cdO$V}k9+DGlPz&`tPm;_C=KOR=K9J}FC;!3H%)3n5f# -3Tabsq(%L)RYu8Wr62=9s4k*Pjc`VZiH7WF<-$x4Bw)Dn~rg)zy4wozj5ie2stbl!P -r?(dwVq7c6@d)xd9-zh_O*|F9ZYXp_r;)aFoWRb*JL+VchH;~2iTUX^u4hABxyqYy!L5NAVo>gjFrgGSyW>pa2z!HMTN -=eD^AWvHXi9a-p_8p_1a%EFO{^dFLmcj4d9NyeUuv?Ma2W9)!kx3Fz^?r^9#TTq%wPJGDa@z9~>+T-NWRijs -B#c_)^p%uZ4KfV!upEX;`yaje_>{wb;_peJ=1(W-{s>6J+?nkPl~VUa$L%HTB&sd>82^?Wt1le2$4{O -z~%KOfk?zXc;6CZ^+V_1EhT;zBtcLSH-pl -miV9?-zU^krpY`x^7<2g!iCu`2PqL{&8jLMMh@nlY1D0Ka}YKPjLbPm93GHg@}LdaM6%612Co%ZBG3L -}z9%P_Aa0xozGg7$>C!uZSO@yeBO@q*|wYHQ#{s8!=gWhQB7~p;>0n%~3Y>>CF+U!FmKAvkg -;CMeGM-kRN(Xhn%Z0Qr?3SXADl0uaJ&A?4I6(+ -^x;c_AU$AdA#WHc%?_4P=->GbUj;FrqiUR_))VepC9_5ictanYly#6KP+NW03HLBbDaPr!K%X`n^_zg7FOXD1~k?ZWUbSg`?YbIvWzi^1crIfQpDgYKMm!7%Bv*RH7Uys1sz)WQ-=dA)|Z@y-4~hm}#)on^j0Bso~Dv! -lIK8e7vE#wcmmWjP+(rMyt%$(djHc@W6xGA|wJ_?wv2Q$>-0Obu6l6L$p)zhsieUMEe+$K=o)(!!|L< -z6h_KK9wgsX2?K&+QGs@q-7!$TN4%_`F3c{p$vJHo%8wD5c{$n&Gr=_<)qB5fb8os|RG=t9ZI)Vddj{ -o7To1jJ4Cy*qxtjshvAM9p5z{IoaL>-8Z2X$vl$m#33tT7mcoN;^^plu~Rfrt^~gi{hY%JD8V_T5naYtcB(D*P1 -N;8MkOnrQv9jvMM-OcV#n@DIf_&X9m&eR^#sAM`!BcfZ^q6LvTt3)(^Qyg1i`hp$Pe3>)mMS@8Fh3W$ -gfgU@K -jsc10nPxACkM{a3r_mT=;TSOi*mV9_t*ih@9;I&N -wh8&4;ZlKUiUC+kA4)RfLO?e#1H2sJjV%Q)Yci`G2COY)p`fi7aRmoZPqOV{XIKc(4q?s<@y$iwE7tux+W%VOi@lgphh?|{80Pb2=C@Ro4hcra>_xZko6^OW*SNgSf+QngI&&jevZchr|L=i7vj`>-g -+3#-ghM4R}HILsbhYsURsJ2v -+cZCeT~29@zWfLDVyyV4d@dt)Y7O23Y++ZVz}~5BuFxB8(IZ_e@^)#ONs$e-@w*QfNeK!0XsZ+1l~c> -MVga(J{4#q!tpa&5xh+0?!{`66BR^d+8*P5tH78wnBsZz@^#f-%;K45UKZE;2+OcYHIRVT02mkOg{^E -Tjt&;rbwobK&Rbt#7UV`($xvvg!KGv&`@7=BQWUW(-vj<7od6MagNqxWlprlDS&<64qZ?G~atMc-J1Q -(8CDlm(2D=|CAG4LnwQ{2u8H`M#AhR~w4LEK&I;!%tV@r7HY2+tDyi}UhQ!v8xtRWuZD6ZG#XhsLg+e --Txq0c6E6eyQ+de;&5Lnxw;$gme95y5J<0((o5=jwcRp28>mIvzV@;1?-RZa~Dx_GDAB2W7Sj;t43DA -|wZ;^L!h&iio-Yr%>TsC^?L(Bl)(nMbjv-JaP|$HhbS-*5Yvw(m>mY0|9~H5@pb+)Zrmm6aK@m_5z%Tz75XIPm9TRcD6J^f|31 --5{&;;zGu&eC*2HQ>3fz!##h+ST_vXL6>&F^N111pk{+`17rJIG$=q}7z2wFUGE9gWhV_T`Cx<*uO(H -l6Yn5#SFPNo4i3>~ujm5O=t~+JgE_@}|o9Ee`04Q0wMmS)>1BFXRBJFs7qZ2EAhyCMF1=oluw-C1QlNbII0P(xbv-cl2az<9K*4KKU!j{Acm> -fG!g@43eBj`l+b9T9P=eFuqv(&n4;*t5G&FJ!dAh<5PZ?PAPxJ(ilN}Uxmc@wGBl^HTX{^#c8WrPgF0 -}V2_C(`^2jAf8)35%K#t-F84dj3#ELv^o4JHPgJeZ&g~%W&v|6Ru%e85sC*t}Zu<;ng?4XkB!)Yh+6P3}h=Qi7u2RDgIV;_b6I}dd%h7hz5y2Ci(?$Kr2BgDm!M -w!^U@5mU6Pj)jlz0%5UI$eW6cOi+jpDc$_pcm#1l -pHoQ1pTyi4J%bEVt};Yup~~Pf(VU!dc1VSbPjX(Wr$(!;5wxQJedC9&o~`cgm+x+8kbF-A*&EZ&?}3y -%+b7Si3<54>%9&-!Uzw=TdP?x2tfO>5oB88FWmxP8OO(m3vW#pK>st1X-Tf(*cywb;21enY&a0&*d9ioLgB2u9r=@d~a~UO7 -%&m5P^UWkevkrZS@)td5pg4a6ShOGZG6Wp$Wa;a~FpAb#CU&-$Wj`m5xI`=pj<`iVS6{j?XPZtm9-hA -IpZUD*WAIzmrE~CMBruE%Y^+P_XqKI3hn2^;04TAt}>3~G_uwS%e}zh(`fI)l@W`{?utx(L7O6_O-;? ->@tIa6pU%ZpB6+xVHqA>4oZVJUs&Hi8I7=ofjd3&?%fwFQq!XAp(hxy!hN!2UV6^90=Eo%NYjxQ(9y~eo~K>lp85zEa?{JhvKH^I -1{miNEjJ=}{=i}iZ3x?kRI#N9f-mcM&17OOwRKbNZ;pgWpxfS+(W>e5Lx;YyW_NmB;VoKyy!Izp36rl -N+(ox*`|bMOub}1P=jDgx{T~!U@0a(h+YNtxVBuwaT&%%i4 -<8n5@$q5(@osa=)s}Ay8@yQn{d?F{)2##D%y3sY4Hxj7HNDkH&*{K5L>In7@s|)co0sym(K+Bmc-#Qj -+7YrlQ|j0zGpPG}R+`4<7p>>bEXUqI=I}X~6c>Haa;qD?Z=`TG`6obcTZ*ZF -I?+W5k$|pXkmzGsD$EpQ&p7vEp$fS|F#_|+V+6XI4f8jS3j2~OvuZ?tqxb{i5ox4#ylA+CRnK$>8b^d+ -W~Zc!4Vr%6#327V@tJ(I2(pE2P%-^n(0v`wXIXjgazFyrdN=v>;CkSLJrXvLJyqqXU_%ukL1lc6ieW| -fDi#rfQLd-P-@9q5P_NH&dR{*XmuMjN9p^H=``P)h>@6aWAK2ms3fSz8Jt)Z=Fd006-g001BW003}la -4%nWWo~3|axZdaadl;LbaO9XUukY>bYEXCaCx;@ZExE+68^4VF{^yA45&)%6y1Z-1n9NtwR_#PNOIc` -MWLl7#^zQQwIr3SchURp`;a0f*>U!Iz+I5U(r|{HdFF*PJ|dz#kB*3r^L&bD_eQ1EUL{hp}!WdqYsOiJjtYw -|jtyHvbH(*_IQLl6s{heu1&_6D2sMT!67f}RN&d<+DY#ONyDStvBD?00G*@~NFvNopCi{s<_`}?eM=V -Vf?CevOb(}tN==Y=e}CC*Py!rF4uY3B5&M#+_8n+-Ihiq)$Y3Qer~Dl23SQ;e0nQ7w<6EXxSSe0YC-G -w9r~Z6#R=6-!Z=6waJuyMjGRX4siOrr)AbqBe=j4FAnI -;vbRoD0J)|V&J}s7l96fRSkr=Q!Y-&3-f+q9(VMw>CT2xwnor;^&NBw^c=@EZPkKWk$$4vgjLn9_;P~6lWz40kh78JkO$NQOIVi#AI8w5_IBt9hYy`DoOk!Oz?plLOgA_yiL=-R88NN-fJ86%-5Gs~OMVxBCOV>@ -xDvP9z=S-jCk82nl&mU*h(m8BOy+V3{MQn(BsN^9y^O4@EmFjy3Uu)oPGXv7&$3w;epUhe}jrL!} -dFG6@fFo`wRJ4t*3o4f`uExi5VHcob+?oQjVU!+P2rNVw-wOBs{%TKChAQ*a@on-Trp3^ftob2O~oTh{lF`v@&GfECd+L^q+75~x+5&)9LR&B -eAq1a(9t8a*1dZbNRoPcPxw{7?e|()HG`3VQ3#qQ;eJStjEY9q$&rH&9_?XQ-!*u^_o!N)Pzom~~YN( -!3s36@12q%|;OThI9ib(H_LpR!f$F*IU$C<~@|Wa8V7qb5lBUnU9`b(|a}IPHx)(N;sOJZ9Q!@ha}a9mr|ZTX1$zU$hlU|s^~jjPb;q=6MYeH%Ga79$ihMkQEIDSq?fI)` -K8lwLzGJ{`pO``|2urJanFSO!&^twcy!g@!^NNSUaAL6NtB%G}E*(}A&XLB;^=N3m06hr^#d$`9+%gN -3f&<(qGm=rK&Y>*;8e6c3U~lU~afrF)5d5a><|US_GsJ<}p|4IFJQcqc6@n;<#R`ef;kQGT8 -C^B%6fRQMR>r2y%4W$L30k2}A-~z36sSm-vM1ikwl3qL7KXvd3wK6@o&4G*2D>+%d>D)Rs6eg2Fm0Q9 -n?rte>@m;SmA*V@Ee1-3O1JyYH>e2%b=iul99|3dg^%9Cq&SdDt_vj`#skrhiUYw2Q;lBXp5bLE!O_% -PfvGlVU-AlVOlp%Vr8S1vlw{{o-zwn)b3FdH_~*ymM<>V*ioFSF-MgLQ-9um-zT4?p-Afet@U-6&y~88YB+qNMforE3D@z}$$tY}-smqo(XZPoy5|$>=FKzv9nv)V6HrS71QY-O00;of09jjpxozqM0RRBA0RR9a0001RX>c!J -c4cm4Z*nhkWpQ<7b98erUte}*a&u{KZeL$6aCv2qL2JV>42AFc6(T)sm#CgY=@<+)b}8&qMt3VDZlX4 -(iJjTbSVw;R>^fZ;!#pIt_aWgKV1`!#{+%;SW-#6YjF}gXaDKioed^m#J#ph-nFW;?Ym~P`O_HK;(zR -3+@Cd1)kAbYSskdvgm65%ZTE%dEa*bB7+HBUa*9KNYCzQp&4yKJv<6Lha=JU%UG$d-*P^~MAQU~^vjQ -!ooG3U4l_HK}EeKy$_<2h#wZ;HX6G!cCzrN}k#M$&z!*N{3Pr|TU(9V2A|gxVYe*vg8Ooxb?oi>jfFB -MVWb`;(D}#>MdEGQz)Wl{iitwgX@Cd)U4+#_sm{l0kIR+*q06^-sKrSfJE#tlNWS-GXAlHT<4=dinkW_0qL#c&ofV~Ff1f;zU1) -VDax~zNg>7K#WM+ku25iACaE*}B%2k-dpj9@oX<**lV7o$O1>{(x%t1-Od#kZRLTz#9S1kz&ENj#TLg -NU|4e3@jw<8vi!=kE{wxE%*tyUF8-68b*8z)LDWNNHtEd4t9G@GU$lgTuhr=KV1Njgu`x8IftoRLCH6 -(xpx6RfY+@ycBLw_eHX*RO)n3Ho9=KVbg{nHHBX7+)=});;z2`UUne+t0Y{5p6b|EWgYbJuy5P(HKr! -932zv<2a${Pftqv!x4KXNov0s(6ebQnix(P9k&=N?@@#gh5ccS<+uf_Qpv5jL*ETT-iF<03@9^|)8AR -ZO(sMzid#`?uMdQlz@bBe+I+{%jw$1}T<0Yk1B!H;8vUMin!2CUU#-!gwA{}FAzHO|Dta9uX^GHzGi- -J;gD=2oCUWc-y8@}hy^O?+%yQjCuq*_bGk(MjOcw8EGjOtX%BqqEicIL8Fc=Ndv|q6WVT3f6%MAA^cr -+({^!`!KQW-i-yNRX2&CVtRfy}KO#7HhVBU*0Q=~C0i?2U(b$5@_y%r>debE{*1lRXY?LaW -WS7c;bnQIK@ZiDNTOM_-pqqyz4(CyuIShv$+ynX)F171?!}|yo!z=irDRBiERHEsOHy_`x$JMhe&9)vik)Qkb)yf7L;?+<(e -G}+gTY|0Qh(X1x~UVH%1xuHdhf68&)(koDyzl1Ot-lbi?V2BR@9P{p=Z@ -Te52T4sgXlSL}l&>JBjq(Uv4tSsteZvb=I%hm*|tdoswR_142F0ND~!~;>3e~}aM!@oayYTh-so2{=CRStZ*l&$Ib=iTbcm8yZvSPKS9}tMqkQ=0{akR%3C<^E%Jgs@yj8@ILT -KF$h~wf~=2`lqlBr@a!IS%d8tw`BoRkZRkN+g|t>R_U^PlC1H>#;_y+^Oq&7s{G@A> -PhY|4c}(%$GklYqV-ZXHhC&-(y*18ev9WDHM$_KI5F8`Rm^iJOi2(GOq)w6T_jyoAj#D}qNq^P{oY;C -g1*eFk0n;cS~MHvsfXa0dG8ncOba-8ye%RjbdBl~lO54u9}Nb4DV!&C0x(&1Ib}X834UHyRJgqFe(Ln -l{h`__+e+8VeYLCSV(uSDCW7RUbeoESi~qJZs9?FVX<*qt8x?3;B?r1$+=ZBBFZx<|KG{wpgjP3?A7g -jtcp=M_=Ya$Qg!EEb%8G5>@UID3hap3Fz7`nr-D!a|lR0h6Y8QHQBX#nJpUK7I&+Kl9xw69K1a_pB)~ -|Uc7zv^5kgt`r!OWo{7Qnui0j!(isZNn%`WihBwksqkF81MVZ0|@~ahLQ1{wYY6Ijms}WslmMv$6QmI -P$ZLMzx&pfq=kczXEL8Eu4Srah(*aZK+C-IF~63*0+{G8P_pa6j3dI2LyxJwWO1Xeh08~{ -7#Gqum$<;e(JH`(0yZU)j69;z!6W))+9)WDDt7?yr;P~5(^KRVxz2D}fe64nc=lIt3f^6`@$dBz -0cJw%a}S{wn+;GGd=L;V=3b;LwZtN=~oehI%z^6~~KaHNL>0_bdvZH8gMdms)Li?TxLjK*@gtwdTbK(5^=A#~SzA!F@ -OfWZ-jgZk-a{gnSEZMbdJX6x9CD&78SfDtzAf`y7V@ ->kY^J)t}5V`s}>QYcW^o!tn6M@;D!FY9BpNMk)3+Ukl?GivzZS4*7V$Xt@5bm7lN7PIRbCZiU)bCplo -?nQ$LPCT^1I!jkUhG5+OETkRJ`#UkDWJoOh?CwDS?aOLUF-s+@f1M%5ePjmnFnrU){tgKU_kuWdIZ2s -K+~ygBup4uDsh=zJ`GNAv&rd;M{bGFMJ-eCRoQ@5nS;;*icHz%vmCy_AV6Cr0E|Q2o*GEX1UR$+&K7_{w -XTypj-%xkMgz1^8hSEp6bP~THp`pGuwB^N$S~psV@vND2L`c>lzyoSRmuD@Z!L^lE#*Q{9~+7R2d*a1 -3s6QuoCU`!P&*Pl{Bee3)M8FvBd>v!(yBwOx`7fw5h7mWO_t~ECDdC`Ubnd83t%lk2EuJTS#-uz65s1 -Dz-bHv$W9UKv*A<%eb;)`hE`B3co8HPa>aH%hmPgapvns}Q5Rml1l^TH(AM%po&Bm968y;4+qDPL8;m -58yhc>hda!!{4|>ng`6g=a2B%*OC6Og9Y^P;b`@{-)<&DfD$S}Hc$I -2gfm?V492j+f%L+zRwP#U)VNHkXwVgrdA=iogeH#w#H*v9MUW-6%bhbShdqXqqNKQeIq;bAMsz4NG0~<3l@L?PJO}JfoaW-<7nlScI7l^oGA -84NjnI+p6MGtc5E2Rsfk5}EWE5aF;aznewSB?yPZJ@f_n$SZZRdzG(2Z_A>gu7l%e2-%yIL|2PQ9IJO -@U_Sg)4^b@yEd!W9I6qgE#W2@=`V6iZ&{CF|>E{5m{-1F**D)WOo9;j>)kdk+{I$s9ty-hWkGG1J&U` -=!w8DZ@dTMP!4&%X5SdRfM11q3_=esIU!SQIBJu6j9=>4rod?XV~&Z4blgFThA!HqXTf=7(my6oHyE -k~KNc(E(IekWrm+Gu_85lCWDY=Z%ZzR0#pFEak&}I(*~-*tF;w4Z#eq2os+nS+Xy?N$QFssH^E7+u9NWf-o=cI16Md{3cZ4P?MN_@Xlmp#9Tz1^=j20dWk!WQ(vX@lVfNg~2U0cKmJRA`uit -tGF)KnfYgw>)@iPhy6O?Aq*rZ}U_Za7T}cUsu3_6Wmga9jj5hTS4I;lMa?Fs4{{ -^8<#E?kcbEAga5hv@d(6KJAA53Y$c*>Z#8|K@+3t4PON>naAkg@I}C5<1VHSgxU4;<5RoayB_Y!u`&y9UA`82R+|T8GHFq!q3^imJNo5^kqu;%iAX>jR!x1AHLKx(@@#{aPDC=GNGy3fgr#YmGaM4s(S -2tI@8Seq00ci?8nJ6;Q#-}Dr=IfEe2JSl=I>G4Q4b>vI@Ed8wp+c&?LtNm-|sTcxtGq&hM3k5zkFxH8 -ihM=OECI0`1Q|;{kAc#5#>b!F)zZA)}gb-sOrJQ^Es3s3h+O1=1#Ulp2J;n05!+heD -gMeYmMMj}LlsX&+3r>W(-#&$Zha%gAKy6zdqG0kKZRp|KTAjK~) -l!!MX(U@rWlbWU^G4Ns(1wD2O{<;#yJ}m8-Ni7YoulC9`Wb -KaAp`Osn%ARv7977nJ)@X?EF$r<92{B>US_ZdKt9{7L4%#B3rND>XC{xa81w+kxIBgEYa%|HxqXV;x7 -LcRnCvPj=)z2x``>{GhORUJCMOR~gNCvGYInuVe^>9gXU}q6jYHqeYyx%1_#q!;EaE3!2=dDasD8kez -4c~QHrx##kiQ4FV&H$%Eiq%^a5~a~q0r+S;NP%_LjWA`}#JFD8){6mk6>d9}>F2%(^2hmO -#^FO(mQ3XuCh|(5GESqm(2b>(`+_3M`Q9V|NV6`#SpFpf -TP%Jv;w^kJRAuKnGcCZd}^dPDm9fS(x(U=}m5skCYwFOR5S?AA6t!_9MQZB9YB}h`KhI9ETepzRQB7t -C6!69A?j2#%A2qMqU_x7}&-~gmqtgG24pusBHw%~3B}1p;IaTxO4N8aXqUNhsleC -`?SVpqYOkG;G~l({)r~Fos3O(?llhGVYU1T!eM*e(fx#lx{&5rpP^z#>g-XN-v@HvEkn7T4GX}vGR$O -{qR8q}Lm8B*v1Cu~L01=GpA8FheQ_Lsybxal-J+Pp3^a;t;u#lVsa1ce8*@7bDD_LABCpl`YpiET_Ta -_l_hqCh2gv83QG43u-12$vDXZ#Q}hz%eUkQTUw+6DtEPJsM1^?#pVO!7Eobiz{T#r%yr5 -La+ZA{^%DNyNO-)i^{ZEDB&^U~ib1zg`j`d?l~5a$EIR%?5h*JQUT60z;-sHmFUqLf?X=V$n{h219?f -CL*6GNm^Rrq9;xk0Q-WNS?yNU{#_IM%3(2327KeaYUF4A(6}g)t#&25_&xgqo}^K%p+=0^dw9jppf -)D51qma+8ns?nw!!oLon(RBJgEa)g=6J0#wiM353v9#WFRD120{E8O@5USmawW!}a+9VT=E{$PSNJ;a -~wYBQ9&w4z|b_fIGS2tO+}%o@!sVzHz|%u7p`RvG`a>-?T%ItxSXDfy}>BYoT$W^C7%mcw+kWT96 -l8ibzlNX$|m-8TQ6#&179$Qc-fC0^-=A#T}K$ty6gDjDe4h@WY$!^qjzzG3P;v)ipH+gDcs+3?rq}$h -w9VsTAlms-=Vi(9=(2Zba-%n^l~EJ(&0?K?lt;VAmA=OV>LLNp2}LmWizT4Pj)Q*qe37lk)}*6;raAg -^4+uKnWt9oVMP#W$A*Fhb$^55Z&8e3>jz-^x{`~}@s)QvMRs73_qc9uL2ZUsrqP63?P!uW9qYbU<<}h -)jJc2|!7m{#$CN#2k<102xq+<#VnK)uoaf_x}-AHSh%!d#AKI~?1186v6sJkn*=1 -!Mt%xFpmZh~)U4W=~VY0t}36t8+@rsfCO}*SV)UdzV=K21A{O8jr1bvNXI#boi3_Rix+9*_YFEx5#!G -C6q`}LXTTc@@UwfyfIuB-b$w1z`0+FKj&^VRx^Pq6#9Lj>^g<85P_T%z#Jhn;i(tdg_S->sth&ZoZ2Y -LvyKF=TPeM>0X=>4Omt>fyGkD9S!Iu|n!*oOe@Wi(8Fw7Vu~4H&ZoVQGP_|L@*viW~HU;pytS&{!`IMaD2jlOfnQ&9DcL8LQOHYUnt#3e03gZ#It*1 -T?hWKhniM0{rE$`=z^`27lS7EhIhch|)@yGHb(b=9uB)!OVfAp{AYZa?!TOZW14AolkMqX~-smn9xV! -^(f;mv*U(1HK&Wf25*I49aYV5BDFAJ`Gw){ce<-@}g*oQCxyi=B~~PuN`V@0Gr51VJ6emRxz}MS;ys9 -K#Q37N7-zuFW5pcC-A+f7-w9?^|Ry&s|Gx=Z6U$JR>)Ix?`-#cOLjx9mV8BrPE+Y-`l=et)fge+q9

>UMK<$~+NC-!B5CI+T$ff0y&%-k}U$+zv=E{S-8<2ha_yC}kR -rV`6fOMTz>q^UEJb_r;YM{U0`td?PXC*lH)-uYezwc^CPIfqyZ^*Xzup{NA#Vr3 -9zA`emlc}``?#h-~RELk-TjB)}ooTV!7FQG+LTzuY_j#Gp`6#N&hAkN!viN&^LS-ETWHr;)79NM5g)J -90tI=0ihC3TkB~6Mck#B&3sc8sNJN+D5_!mwa**FcrO!1%4fs2xeeIZHLqVLRvCQUn|kNV4}M7*Zkn8 -Yxc&B(SSr1YN~VfhBq4P+FHXHp)z!pul^du&VN}x&; -e?gY`UVDM*m9$a824NRXe2Wv+H=-;orz+$nnSqXorO|KEEESu@ozyjgct7z67?n54~{xedyp~i{i2OcRxm-XS1)_91yQ`|BjMXo|(u7ZP}OXli$3p_tZH06C67j! -#4bqal$`$(mP7f>JH;{ov)65%iM*>^%E+2j&}ig++mOMcbu)Kk7u8Ot3H*!G0-{748hdAi~MnN{3MCYzR-DT&{&)nnBb7`jC_ZI -$?E{kj0`%Nl1~>3Zf(v#zb*&2cCz;UQAB$(Wn$!Deg>(^6#ey_gf<&o_j0@po0NP^73o6mu4FG@G5nk>A#+Sxd^92Zbj}T} -r$yoP&1Y-n|t1rU#D`%`v~v>NxHdBFA2|MdztkWa5~PF?oM)ZE?8%5yu0ca -I7WB5sU)Iqgf~f4Oohudd$Hj-vQfxR`UI-(T@u6H2k;VtkI8SWe-I8Zzj;gXPkw=WWuTVD><~q_iFvkUV3<{Aa(GUcOctK>cLyu>qaS_FRl3ln=@FonYb%j& -G=2ek;R#yh^jj#8WbfL&wzDll)KZwjA7Xh|cqbX_h*yisAju?4=Iy>n^hYJ;gAp7rV8-+uLchmvsaEwaanl>najAnYjkeO%54ZgwSp2+R?bXvnIKcFZpq -U%%ZGK?7Fk*>O)CnQp+uN%LxB9@fH}nX0x032vu1ER6wyvx#O*9_!?_3rmrI+I++TNfoBK+0Z3qS8Xd --ls-J^Fc5zJV7T0=6!*J??`&!jKTi9WKNTsw1YNFHbZ#W^%z!GG#DjJS(Co265s_$rOmuuJ9O@Tq2Q0 -mAIuI4GT0AFCF{II92Hce-hf&WMZgMc`b}-O4ANk_l!C>xI!*y+@x`M!%8msfa2sZ9yiy=>sfd-^^5+ -1t6z8;H7xfK*Uv}ok$LypzgNUg1Zr%1rv -xb%c{N+Q8S3+;4}4OD=yJ|@&s!UI)GY(*(H>^;rbJB@;|rn^f5I3yRJ&tU@RcCq)&|HXAsK}q3+G=WC -$P+Q4d8ALdX8^BS&puJ5=A83fw?iD9f$5XK_yX_HpFzD~j@0~dEa9)f -68^e_Rn=`$-N&ShuSEJHh*aIhqg&(=Hd%cXefTQ*l|WwtftVVPpiqh#T6}Pb@&A4JUyL#TY8ZWZZjS_ -i0Z>Z=1QY-O00;of09jkI-i@1j3jhGhCjbB(0001RX>c!Jc4cm4Z*nhkWpQ<7b98erVRdw9E^vA68f$ -OcIP$xH1yA`9duuC?wrOG1SY+GmHfWm`Nfy^bQ3$j}+iYY}C8;>sAotsEh7XC7?Kr)~VGoCMfJ7GO#d -&j>D2h6geJ`2PDyF5Lcm5fEI-ScoS7ag5GGk;Sa!q-zh@LZoz%18%LNymTnbSPYm{epcBsupKI}y?4%agOqsit6$PRgbvfnD-Rf@VZrbW4TmTChGysnk-)v^VS!_<#w|nJy*RoswS=i5 -vqGu^Nhvg%NItM5kb*T$2(!M3AD?fH3B>fyj2jcNPn97Xq%LMM1=ra5Oj74p-4zWxm5N@VGIX(`znDN -#ILbF2-aG9%qA2zK?)`ca8@cAI97HGG9pv1m#;+IwBoz}AKuexwARq{|5ZAB)p!GD~haZE -=rz98PeF|br(|p1LW7NLr%z+;t2{Ar&tJlU_XCfb11mdYU{<$ET$Y)HVB&saXC9NzzUP;uXWo0CvG6q -s$_+K*PC1H#X7aR+PKf%y^8Y)V=$gyzimXTBNA+7XnHe-)fM_kpLtlm@_KLv;;q6h)u5rPunyKDa -ILFs5pwzovE($)ay4OAzs)_$lK8y27Y3r=kHc$HDIrL@>C#V!bs?KHdx08$e**0r}DpR9erEW%If_&8 -+kI$Mio^TdxHTv$x~1j#23xrYh}B&Q3WIdg9Y~YkFm|J-6#XW@`I?ij$Z8tO?rW>{&536%)Wq&soG}N -h&)-TQMC(IRjxaZhEe=Y@VvV&arb&Lq844N9>+|YEbe(i8JJ{%V3Ij7XH~ZU#kiAMGYOf3cOnv2F)%P ->abqmV%S?il1PfhN1x#s4CV-P9UB$9#pMXt)FDL -HX%gawjR|wzr!q>IxSZg|gOl|;!dKL64;KAtBGZ4OtF6Z!ZYa=}8y4uLb5{VxKiPv~i6JU*NgzVV&ik -Scr8=7<0%d%!l`O^5dMv@VkF|Zf}jUm}5kzb4YV2Qk)JlDv&O2V(LC-F51;9z$-+qOGQvY&RP@`3zArnQGH&C&6{rE?q)MU)uR|rW^chhcA#o -0EUHcpWzaay_#aTQzhN2_P-z0`Dpc2&z@f?n=BjuQrARCa*sb43j9{>tOIfmQpx_re$&E}!Ip;-`< -o%w{>FL`Pc&O#zp{R#Um^n#)ydN9bt8fS2oS!*Nopf_3^}_S0Z3GbIL)Bt;(GpoSQf1U43I@ZA$qrAT -BraGkNt26pT2Bp-??xt5coc>Hk~yI1&%(vd0wlZW$Gc|>N$p18m!t2VlJbLaZ2Ght5gZ;!Ig~j7 -+VeVuAhiuO9$7PX&9YGTke`zd`{&iygNu`ki)h$sNQ>^bNS_No_w=1sfF3_mxp;hU<(1d4- -~fdTx-<##uRiGexrg<^7J72uQCI=gv61MROW;NtD^x=pG%{p&Uf%nu&DOE2lRT@DVPA8ptvto)^$1^z -#>T`%s|9X~%f+^}O<`Ac^U{C{M}Ufde!9nRToN+7CF(qWC#okn-+et4JOM?Fj&x%f_pCoi5JG;&zYVQ -mg`A~#dt4*C4~vz;r{2O)CHbZv=T!-m>bu1L%b5nCERlW0o6M2lef*?8;T8CgVF`aN@B66j5gL20c6_VCM(VpaU7*rOi+DQVWB0*!O?ow58;Lk((IYA<6q-6ky#gku2qsO|&0^^vO>@;r+`@HIHQHOfgNGCq0wIn8d3ePe#Sz@eQ -^nW3DPD+cRj=?86wie_ZR)pFpb%+;c}{{dshUoK -9;#&>&zsdv*g|St`u4x&N)9FBl_ptqV2W+na%3m1Ddo-7C~|&vV@4aD-xN@9B#dhwVi9Y=57;8iA5=w -o$4CgO|J_&-eHL4ONE-^_(i;xM4`(cDdS>?T~Y}E4Cqk8CP57~GR99+QTBsXf+s>r&JDJzytoQWl#gpG@ -B)%q)P_=9$O(l!)cse|auT(5(QyV40oK;CIaXKDnCr@x(jJU@ATd~wp-#?Q@%R<6;zz$sScxEldsyV< -_Y0HcFm>pZYSUW5PRxgY^1)4B!(j3OK~gK$`?FsB^aT2DffdIywqxli -+x5H*kF0Z!-bCd{YI5WQ8%l_ne6+Zd)+x|NEeesIh<9uSu9kx?JTG<1Z?i$7xb%DbcJwFmyMzTQ^)=I -nWtOe%NNdz;UJl)f6?6Mj^FnVvIT?Y>;zXDDj-*IQKTDMS;+A1;86{WMt^9%QuO9u)Eouvzqy8%XA-- -bTt7Ehtg(19F&2kDC>vK;miwC{}R$Uy@m<{o{9g2U8zGU}#N&x02*8+}1mc9(JE46EC -R$c++M&Sr+V;-Aq`a$+3nyIxaOc7mV%ne`;^GCm8jR=bLvO&~D;wxQXS|mSkTC#I&6Zuv?W3|68 -RxbDNq#OU_?TgZX0Z>Z=1QY-O00;of09jjynjak-0RRBG0RR9U0001RX>c!Jc -4cm4Z*nhkWpQ<7b98erV`Xx5b1rasRgf`G#4r#=`#HrBnnhY=d!#K86%rNN-txxdWJEGHV|yVD;t-sW -li*}S!BqB0|M@@P-W*EO!-BS;20w($o5QRbb(3&g1*icz1_6Y{r$8C|a#r;| -yCvw+L^N>qo9Ckz;EvBra(Cr^HhXny**@_GpaQiHyPH5gTJB%VJD(4_GHEOWti2?%)?tO9KQH000080 -LuVbTO!V%e>?*K03HVb03HAU0B~t=FJE?LZe(wAFLGsZb!BsOb1!3WZE#_9E^v8uRZ(x-HV}U2uQ<35 -wE<69ESIP$+79bdto -s%TI`xg9>o8P1Fd^<4M^VbO#4%8jTI8qqCkM?^c!4iaJmsFsZ1J&WHy7?&x3*+yL(xja>l`1Rjr4nE< -WzP|5o1S^6f3C0R72G|&OT;Dz8QUm1KtACxtPSn(aikVNQ$5bxlp{P8M!tyQ&~`C4y};FK%M6Pm{QVt -|bID9eGlEJ22mV00$-dnBw0&IeT=mdSu^VWJCAp)z}j0h>aQ?`{R~r$RY$et8xR<^Vsb5UUnkz=hDAQU7-cEuqF24 -w_sq1di(G25`riR|kQM(DXTcUR9&__#n3!m}*iN{Y;ec-~_$XaK?YJ3e$FjS1il$6c7t0k)r6!r6phS -ELq7v0{=Ch2mvpZ`8{j2Qdc4QlHd%8iy+08hJaKV7Hus9>Vlm0fXQ9Z9wbeRTJ -&eRNf;X5#R`rPcD7BWA#w~o3&ha&P5h-~FJMTUax$DDZzbbr{K8Xu -_76{Qr-2=R;a=7qJoM1;7SVI%xhkCWsWq8(-K+ImvK`GVI2iM=*F0EW?k5vX>ds`oT4g4 -qKYFtIS7x%J8KRp-w!@FZGT%Lt^Y+3u+kR)4Ximyob7alsSkGFfI+t37(aGqptU4z{c2@N0f}x6o=UX(M=zd-TiTM*Ig*O9KQH00 -0080LuVbTh`v<*u)b60G>zy03QGV0B~t=FJE?LZe(wAFLGsZb!BsOb1!3WZ)<5~b1ras-8^fL+sKjM= -T~%S7$a?DZj$?QnLxgpFCnsIgm93o&n61%&)-mh+M -k|Y;O{-c+=(`nu`_odVq|2z0xTzu0U+N>*Y%c|^7B5w*Qx<=eek=<7GwbxQ?@4BvKAn!Akena2>J5yV -6EW115_POv|_FY+NQ6Abxbz1CHa}ak@$(7IqKF^fg^_5xAv(k$MkiIa#U`8eDB2(^n)1K`2t~|(#3j& -&M%DR)P&ZO#QB?slhSR^T@BLoVA+yrk!URH{+Y#G*eOP8&B}WI9-TagkS<)__)i|Hd%VEC#;(pt4qrOl%#8d -Bzr{bSMvHj-!Aip#2*+2qHlM*LTtgx2(@eQ+{Y=-XTgQFcSH?%W_)Os6r~-DQ}d5In{~a9r9tVWg -l62$LjHn;BqK>-vXUtctJbxI8bs#4ZGpxdU_CF=ODQ2?~6U=7Lu!Qwu(Jc7bzK(!VRC2)2xyA3?t@Wa -J&OE?jiS|xEYE<6ZOBGt62%sPoX|Jd{d6dvA!lAXjc10)EXp@kuCkzYGtRsoEunqz7uZ%BnPu>oB#yU -k_+O1oQGf|o`|@Pp6OjS_1+HW&%;hdCcV#lzjHmGIbV{DPsS@L-wh9B7js7sMYI(5wIj;6$%#4=`SF* -XV9$#q_qd?5ZY@ff$KzfW&j~2#niv8G``V1O7r)!?n-E_naKslNPfkf3I+JT8Q_MT1y6XiSON_b>cQT35%K -9=^vh$x3cDH&BfaCNya-~$-I+m^9rWxl7@RuIj_y>pKY;s*C)~JSfr6HL)7KADmAew7(u{szt3EKUyG -*lXvva7kgB&lyr|8ZH!UZQf7z#Yaz=DUHklV`os$2i1sVBCa(s;9Ov{Nv~FqZ*avZ9E0zn7<#c*v??P -jHtmupv*t=0qj;B_s^sjwzWze^2r403oR8pZLs-ghvpo|A;>1!p6Qjt+w!j{RDLd?w$g#di$TO1_)}_?g~*IDHoG=Y)?}>w<6YMMUVNSD67t>(^coyt1L}s}k=-i7by`K)EDk~-0xr5RO -zF1&BQp+5hRsD7?f(ywv{RX2ft4jEN(=DbS=XtBg|HI -I#^(AP`!HE@u48_t;~67i)6Zwb#_??M{1TjXsf`VRX}4f`p^^?zH{l!@0U8&SZ3~;5^KLEVo4__s6a! -tSXb`A2k<<2=0bwx4kpqCe;%&6s#$d5HLwS18d$T{veV~T8ko3>0>8KuwG}P-Q{e7*PqKCo|inyTK5m -Dq5CXeKHg2u{=h_wYQ)H`74m@`&m4%rH^K+jU+^a2m@Ms^<5L -dYMKnqEQT^s2}f4Qfca1!T=P>QQ^%242_FvO8tJZCN$$J1k!XUY!9*2i -ykD$>*4-0=j0?G3rMXRlg#Uesv6~bDDq}wDDIcDdYNj4|;usssqr5f3cSNg=k$nl$|-c@C8x>zQt+b=2#*Mdt&k@IHBvD==ZBfU$6`GaqS&{WVYRp!GCwt6XacTUf@A` -zuF^A;2X_4BI`&RIG1J -ZxCn^n560FKtDY2?S{jRQvbHL(EodSLvD?`((8+9O3zF+)$Nm2z%h`nx@Bh -D7XkHiXw^#?%a1<{4DmW)QFoniq0T0*y)Gg$}DnJ(yGQ$0vT3WG2sWS;I?=k^oj;8=NY>Sj~)ka6W;sCp4bUBg7*O5$$&|HxC&Y99b*iFlFe -fyk&JUWqThxIZF4=cJ?mUrT3-)z6%32`t=D(3>R1dLi(JU^wNz#C$6@6Hn -}tX^o02M(tAVMec`KTZ3LVM)`Yr8nt91t#p%EH4PYyH#C%W5ZpkXh!^=`dlYDaxtj95s~+(KHm;>aJq -H^{KtvMmLKy_+ADuj4COxTnaq@}NheY7#M8#H+iFw)|rF)WQ0cO-qk?Y8Gz|N36Bz2xlmKpA9THRh6% -UF&K^D#-ftHVqOHggyS2I;uCCq+nI#j99AI3#5bgxN&s#jaQ%mj3URQ*t|{Jz -xa5Jyo*U&pHk3WXZKj}lcR_9ucw?EzR%YQ)WWuov8}@6`3BtZ>)2BT9+O;E7#5>WQ9%7a`cl!ErJj(y -Yod%K@9X5~y2z7-zwlxwk-K$P;_EeY!?@IOwP3r0gXQ!X7USQzAeit>j*53dK*OWZD5MQk#jgra#<1; -(>gGR(Pa(@}~>bhy#||R<#WPme&p*@qC8YdXm>ujI#x7Sm;KKDrEJT8p=v`GB_YH_y0215p(jfoxi2Hz -4T1lZl{|LOKcfV3w~;PfNInxvSqp%H$R}P$C`e!UxXenvQ>MC89JJkNOYH}PNND8+8IGeq`9e|xwuR& -ADImCZ({1O(-Eu~xWH6N2tK)7T>ky<$tB>Ki~^|jV_^?thAp0xeOrK4F&8NX8RT~~&eu+{*JX?&tn{* -))&;G6Jgl*{7D_no`OvE1La*1O0bAmbM$5*LWYoTC+Ea@EFcX&Nnnjj%%niebz;w_J#jWHg~x|&WvKEGNY9 -ObU3j&gb4lM}M>YKHKHjBI)>er%U>)Jvd2x?yLV`HNkGRIESZzrRqQoI!jvf9VmK#Yg-XekMb68btNa -HLkB4ysxmOY-wYnnmeF^NANv2H7Ian*VmLUUYWDET88cV9`A3VM>v`|Ub|DheX -TL(p)+RK7b?=MLMQ_%Q`f*5xM0BOx>F}y9__O;1b0MAyYM{%hPS`a;O9I>{OSAd|BP^yUctN$2xXd4Z6Da#NU5<7Klk8CLlIfXU*`n;5kRC -Ek43`$@+d;2 -srn|F3Av>LI85p>HZ*YU0u_Yzn?d9D-a9Gcbj>{Ps1r$5c){1o1FhQ8%J3-IH5AqT`S|PIv!?1acp!N -uZWJcT`?nk%PyfIC~1jg0ipYx#BvYT93Q6N!k;me=*_Np)s+WU(1ERK}X9Yj0*@tP_SGWBPP^G`W@cb -Coh$~t3WE?U%h9n{^TqZ|m|S+r@{*Y1+oiccC!I11&2Q%Q~Y*Vf{rS8V`{H%XE!@hjh9BaaV{h(p0#L_OgH@8CIJ -8G*-?8~dkxG2;rh59i{Y40!fs_)yEdR|uK`6&MyDWuYdL17|Tya@V8k(mv`3G8e>foFX)%y`qG*$PYrCF!tM2`l@kfeE{Sz^DfW -1yuXbbiL;gUIe1ssQ5xsN3XTuhAs^LW=)@!u*n+j9@KCCkcN;vz|IY7%1q@ -0s&QHu$TPg^5zrd(BPtO&Hd87R4gjsv?s`#bx05(yMdCqtxQIjJkM6*dGX@;?ThC>is#>6J%4re{IBB -Audb7eG0pbnf1-JRF!Vz$+7b^&+pgm6;*b#=T+m-4|FO2ftfRSE4X2=sS(E;a<}UqY_F(IQ%j4wQuC( -gE8WO$czkYzSbh9qSR~_8|CFd5PF@<0dOnDu9Lv)7g$Q9bf*ki{7(Bvj?jO-VOgczS5^MU<*$Oeif7V -SY`;n?JGx3TI5i&YL+)>>WXe}j?Q9Kmvmh2K|x0{3w6e&i8L%QZ$2Z`E(y!K%+^OiLMk?U$wx?hpIoH ->iRguf!kxO_4A8LJ~Lz$c@lxkdyG)@E8caj1_cNzVgoz48DaLzU$RB5e!FN1moq7W0y%kuqNOZ{-{@C -503m15!b@?)2Mj*doU&gT -%wU3fVTU5meOq?kDW2rE4ok&K5{muKg%c-E{6WW7Rd+FZMa-$Ydkj;Ko6oIY;%#SY<5Wr&Y1;IcwgM$ -?l^&ZzrerycGUZ+=+knG#fPt-D!gFnO>@nP;D$3!+C_ZG>IO;7)?5xV9<%otc|;uUkWqh66?SWfHsyo -b`EcM6dAE{d!+2&7Y6pC@&cFnFEhM9`|tFfI5Bip=u1@1{}%Z(NYTk@Yzvnz)iPZ3f};Ic@(xP)h>@6 -aWAK2ms3fSzEw8U^;;X006KM001HY003}la4%nWWo~3|axZdaadl;LbaO9dcw=R7bZKvHb1rasrB_RH -+cprs>sRc!hsYTXd+04QJtR&tNheM+w%Z*X@OqAfRxQHNX5##3Tjhl_fKfG^I7A|(ky -@5Q3Urk*LH^r2->9Q$^ -V5;9GF^Ka;KheMJ$AzR9DYPiu_oJ}R3dtc<5X3k_9S_P2!tKxJg1;UvcwLC%2e%SXsM%~qMTH@#HJQvkWY3`hHSq-f6D8|NKn~<2I~3c2=+n&Zi3>mXN -H-*osMF-w${)!TMR`Y(+>C|OkW)IOHBky`lRrvl;$Wnv09CV72+PDFu0e^;p_I>;oJYyYuYp=`iTrc$ -m9jyP@ww)f!((dd5l@cqD|le -Mk*qQO;tabiPLbT`{KjUY+TSUJVFQt1jPRC!5m5aQ8QRDJ$P`WKFZ7GyRCa2yF5 -p_`rY9&=b)b`tb^LKolA2cVrno$v_h+JrzExXlZ!)E>+N{!(Q#9n%~owm0x4ggIv -G42l?Q>NN5lF4>nVL66k6gX90Gqr5*EO(Ei}(|ks5zbS~MzOl6Y@cKL{$5t`n{G`@i9dFISRam%9r|^ -)555zCr?`6l86mo%S(BMyNlZMAm4gQVMQ?DC2LUB8UcvUp^5%oJu1Ax+Yt|jl53K9a7Hct?MEe0$bDI -i{U`y`3bbPhxklF@Ti^J4tUG#K(;nKBhjyEXPJc%E>_tvz~F6)%A_zV$IxciZ|NtH6Z3wVpPzUH!5?WuT8TY8yd?`pJ -6ZJXW~=vAKHrI>xa?;6SP%7_>^g(W7D}}1SWGgNZjfeb{S>n85#vVNKze5c!Jc4cm4Z*nhkWpQ<7b9 -8erXk~10E^v9ZS4(f>HW0q+S4`Q1ZNL@U<`y8AF49GUu5Bc{y%dE&TXf7sCRLJ3q6Yr&9g3txNw(pls -F2v!a30@0I71RbzR0{P7;L04lnXANU|l_dVx?3DcC09wn(A3f2uTvYZ>Y~&itxKEsM$)j5ARwR?ax{# -$yUidl)Ox{S0-|)z%TM#(36XmDhj^7^3rJ_XUL{BmmZXeBcBv4cXg -n%LNlkun~RxZy4RN^NzS&AA+=z8hN&|cfa%QXz+62rQz-!ya=tO}#?21Qj=`!nJ1L+q22|wWW0)q5hL -w`9JZG?$UQp0IYxwMo=LNYvt<#|_wEU!t9S9Ma~Gq}B*E(^k -el*iq1KS+{+Fi>*xe^jWI;C5&8L%D7RoW1Mxsw~_!pV>PtX^I`ucUW9xT#O)2&bKry6enYMYa&sVFandS9^x`h*^C(Q8Zz~Fov9}td-+mYPyv6 -c=)EBX)YpCjpVH3^po$ng#WEY!cEmS#KIBLOJvsW9Llh1yMt3Oh)C>-?QVMIbiRAAnU-9h^^9iip<)wvi93!cjUlGV -!XY&4PpfcLS^j>D>g8~;<`-NZ-xwBe`o0!!%jH!Pp6&LQLYX@CC{_y&C|F~SVQ?KEi+KXkHSNl>=-3-Np4$j8xh1L4ztE%wIz(tN -Fbo28J53~suaF*F=jTb_Ds>m{n1NV^w;y@+;08mQ<1QY-O00;of09jkDfkzl;0RRBq0ssIV0001RX>c -!Jc4cm4Z*nhkWpQ<7b98erXm4+8b1rasZIQuhgFq05@B0*ky_gVu03nA$4uwL6Hn$LTqGMcQcJ1z(5< ->dy9aoJ>+W}qu*ZJn(-yjHtq0g4Avr(n>!HVyP5@KCR3rAhFH9=$aNfu*yE>@q9vAx`TF6x(C0l|02#cHB(vZ@>2i5Fe!MTszb2 -+-WAU1OgAh!^8UqIGIfL8vWO;Zo4*J*SRj+z5nt*dwXY>MXTW9w$y4_S6x{w|b&)G*P!`LgQ_`}8mQB -jdyazW$cUe@>b)&N!hMC=}Wwvd#`mwIBwwsqtQ#b17Qdg?&vMSG-T$ROngTsbdLCZ}2>Ez7}Ah+A)rO -K%6<@#Lbxz5#_+wQWiQgy1ey6n15d-U+(1@yQ*PnY%j;ig?ZH0XO+v@Lvj_~e^kK26Pp&FtE=$g51Le -VFE7vm^ED>qlSht(tnRirUQNUDa+k_)X`pH@eBXqOSJ1(x%v?^J`t@b(3zJ5+?8@)gPC7gLPU1AlXy! -d7+_8P4k{n!H@NlV7>FMc@@@qOioK>OA+zv&S{}mSQYE$Y>-P*M -zZ0Y%WTXw}JYr6TWZq`6E**J)8SCppt1}8X&CxW$fUQ|FeRaVZinjKtM&nv*lwVt*|xhhjea;gd-L0@Etb!qP@V$`Y_iTr|MRRYv-46j>VALv`sZjte#$l*?3R9L=F56>JLkgrU -$SN<ac9a75P{*gKZq1Wn8)hnVk#mHp*vGgk{*aIM|R0i8{+tq7eG)rI!x7d8s)G+!M* -#nF+FB+ijbnorU7bkCDK0lkEo;^D|QO5?82oz-Uwa#ZKIG%vS=*w62wld|9o$?ywawFwWobpcNDuL9S -n~tT@FP%z&f^}BG-U=%{#reN@`RdubAJ67?yg#1&baDpcee!7U`Rkw0UjBSGfBwU}pZ_pFJ^9m1DF5n -_dI0}D{(A4^?B!ek**9N3-rL(-mRW1b-t7OMijPHgM8o?(%hY*};EvTM#V@wBSb;3>=wUw$r~2ff-)Rb5A7 -od8j~(DJbx=hRI(>7&Q-KfLfiwA!$#m=)wRAdZH0blGL64p4+qlCXcG+QWK( -PDnZe$&Tl*2PN)9z$vTiqW6D69uheDMW?k0^gbC`hlmB7c`J6cAJZ|5DEMU@i%R$FdS2J}@YZ&moaOs -DqecjF+HD-#7hC9QI3m=b;%6y=E7}<6EA$p4AyT-9yx5G_gFxR-=*yqo6rH5JYPKgG{g%0 -!$@KtvrrolFC)08RlRQ&=M~My_Ue!n?i$-n9$wdN;i3=+o&CO<83yVGPYwUbp(M$7jt}PaR@w^*Bl1g -Pov^Zxi}|redbj^Y8OIzom80l(hvOqAhSpds{57`!E>l>`Qpd7jl+=AYVb|R^V6&x_6R$8bNd9ec7 -=5=YTSlR4T~|x!?8!7-LVKAXXn>2KkWS4!L|G8m-~h!CZJu^4w!1AjlQqMXGInDNAb-W?&1?3Dj!Hwe(B{UvC1LzV+K^-5f$HO$Us({;} -vVR>tCg#%vc(Mz?jAt?gBovTw!x>N5Z#>*^RY67HI>zDQg2Z)sT03xv-8-5o|H+yryB$L3in!h}zzqi -K=$i;}P*e2*q*x`H#&BTFz-We<49cRS+UonDsUNB@AXZT5`XLj<;PDl!9>*(|F9x&6xI`lEj#rm~nzT -yS*_Gp%Ohi`BQCc3aI=_0%cbb*6rBJ)E(v~2=Ku2AlNe?YH=weWzQT+qDejJM$^|3}@z%By0fqo$}cx -Osq0|dYTl=xcN8?wMbqcW{PtYWP+w1Wr&bMZ=%6-{oZJEZpNz;SyEjCP}!1xy)OCVKY`WLPCN`9wcHP -Y$MIMi}-wvu@acH2ZPACqN&}3em)e?J5R1Dr;kq`eR2Ri6>qnI`DBgOlOQT7hD^lzb{#x7us2tO{1@i -dfPgS1vsU}vTfcH;j{`>dmIs}+cY};5gP!?Uu7sWSRjvBy1!IKuBCi)bMB_E!3~&%l8WjV2MHT4$XwFMs#jaY0qrSH(Zii2tmR33!vD`VtI -*-two{*Fdg{}n`;bT5&R=E3KSbjN9;)CAI4whRs-AY$oR4VWbQo_hIK@yL{V21j9Z!F!6q!kg9e`r8$ -2@R0HEWusJQ^evLyvlxXyft$fb^YyA4BlU6%n8Wb(#bQzt7idQ8WSUFN`^S}fitpjeZ>3?}G@#R4&Et -cIB)ThK-hZ@~KNX%iT=+LYPSMWNg`^tE!9IFrE6NfSI;rTxo -6X=3JQW(Vo4p-$Runw4Ef?H)qM)H*V_8Lhb$fuh|J*VQ|XAO?n1d9Tck_{~eyGyzx90EE-d*VD>Uhz| -!oiZ^z*csV{wtE!59~~K%^!@kWn=ph4$2w1-2qdfxa~sZa@Vl312QzhG;=1_vyr|N;xp;V?yMw6@A{B -mMRhe-2tM>r#!}o^!^gJCTM^$w5d2eJ8SGT`-^BfK_QYtP$Y{Oc8 -$vsl;K`-iX6EF)*%O;lwNPTTUz6m){^`#Je1TME!pC(iv%-kvPZ7 -an{*~93#O`91MN_@V&E@4G)N;PZ*4)&}sWc3Ij`{Cac;thKn(H53-^WEv@KM*WrdnUI4)%n2qI79%% -@7ai-+E1CY^}MfaZ=hU2ciUxS%Xzxe)BVncTM{b$5IECu+_gjYFa;6Ep}W$D0wX84y&4Z*C4m-Sn)e> -x3mVq>+<5bHxB%UNu5^|~l}ciA(&-gXoK{sJ0deavKlPy!Z8dsHX%#|&61ne-fhMil^^pQGyMiC7@)k --pFsrW%39OCkO2ZY_C+6iSfV%X(Yp#@ou4O;-@GbyjUL>iwO{YaD7@lT#b?mVCR^YFp*HEgHlFLDH)g -CXoHJggXO>1Ux;vL)bf2TX!6%ns&1Xt4&*S23Q5+t)wET_m3Yv+Nn7N{AORw23XZ|qveN-ZR+~m-t)8 ->FphcQ>A+nbfMK~1!tixCf+*Y~Xx_^Y9|Xld7FL;DpbKqVjcGd|S8B95&=*JC2i -f&uSzNIh6m&Spvbk+36N-f3{}FYToM1F#&UyvLS~UG8Ar}>vUli9AW;C{+?;&GpCM3xyO8C!oxDF<{p -h%Xvj?@w?CMHrhTuac9$lsA#xavvo76sBoq>NNXWG5tlm7guM57a_?7IT4}$~O21@`WUmu80!x8KRVv -R8@=Va*q1g=jFa8n-v<~)Xf#fn=+B8(n~%N7)I6t`~IhQr)TP-4j2RWoRTFh{nriVa-mGi@?89!BWR~ -BNdHg2M8%Ax$hfMBxt1ZIyA(jn<{hZcV`krmlRFQ3#_aOxl!ic|C69eO;XyWUJbt%vIck_Ons_YN9lE -ei%SPSN3IFSvq1#!=s|GmOfL>>?9PCJTx5~a8X92v`9nYPorth?`3^L{UXsFc(5~eIC&f|2u+2|&j?o -R2#>`Tl+mt>3_NSJKOwsYngrdb=k3mfim{~{C^B-S!D+CQp!9lx3v%T2^$^xoZkC_SupK>K%5t%&xU$87SQQ8r5!MWRC6wZU3kb8dQ6U;{gjh -FN-QKZ#jT2iKq_&4sf)JbcNP!owe*oBdw>XmgP2QrKpd%o5WLe<((fKc*+56iw&J)P*2j2-mvaE(y#& -k`tb4NCr@4Sb%Gd#)}b)Hbs)IQ?yY(h%M<3*HAyH>-cl?-QV&#At0YaTo%NeJ&f{M{(7i%<9kNblgy!4hvm@~xZS8zjwsc6I^`rQ8H -fljmLqV>as -3+r;BWMBXvy@evY-HshOkd*Y-LCje~a8?A0K@f(1#g6z0Ipv9rWwxd9-gbs!Im{Va6Slmg^PM@Eg&;g -cgNg6S0W5Dz3o0fue94&D3V+qaLdlDtXeZ%EWGibmJA}B`P -%E-^K+0*=FkRA%S8jMDV5N0;*?lLwQTnn1n5EWP{o;Vi~;5X!9G0}&Y4Dj~bl9dX-I -*n({Y51>rk!Wf+jQWfd3KRjvji(@tU%_IZ_rKDCkFp!OnL0CAv6L%iJMJF4xP;$5-b=fW2Mr3l -o(*(W{78mKRymRYJkMu}mT=wf*E$6m6Q=NYrAT&oY^~1J_QW^)&FyP0WdRR?9(bv!a{;3D@*=L2rX+H -)+-z+@h;hY0X`jc(Rdn2EC>5Nn&(?66=AEtrz+Y8c2*)*0oE1bq2zn8LU?5)^g&>Ln0r?3)l5lk+Chb -hWHVV4=Ep-4XxLyD0Kx)yC%tHPk}Xtt@qp+Iz;#W{HJOWYE1mLaDe6DS)^iYmDncpp(Nf|!25v^Rtju -LpTEcrhZM2041}w>D&#(s>qdb$~x4Lvw@?fY6*!8yjT)Z&O3v2O1}YFG8FLlKUnmEo)#HPfz@_ -B{{3KLBuLqctC+_UI)8rcHLu@JvJO%m6aq=_g^^N5}%T#C~f;3=f(c3wndqja)mcBasK2|fv4iUwXXvLOgupC-Sj20&ZGqFKv1J;Mjp;ldoF|Vbat#0QN2)zRN$lKRkT8%i?R!X*^TD2BE(EOXp^ -2>%Ig>^i=O4&Dwny5p!6T0i$VQ -tEz&G>D;kt{)_Q-W*7wAdwt7 -3yHxSGVY&b;sTubDeaM3?~Od#94K^%fc+eOuaP6pV=Xqo0qe7bi6QM(`J2+H>_{0o>X9a9DRBjyEy4; -8_0fPvCDVY=<&_EWI>tpv4;CPlbPgK{?^GzdRK3RA3vPe~lDlRlf${+mz?36KaioMv#3JhN;#}nMNMb -S-Bno*Xli$5OLnp`&FQ2^#PKP0$y}%l%1W7oIsYw~^6{C>9odTMI6$V5+A}fzaBVmbUR&i(|iv`v`>E -Ahd!evtcY3NiXMGF7pbYLdPL2xG1c;374IC&H&%WD<9s?Xc -^R~~q>REZX1O|Xypjh@ujjGGAy^D^L2h>5tbxD=XBp#1vqX5S(x?8IY9mv~8JaWl^<+HEeZSYPoA}rR -lR;3n4AAuhP$+D7e8=BA4ErCcHTM~7GYV>*o`hKu`?k##OpPmkJXwI!*0ymNEd(k#m4sXjK_144zSt; -L&XPk5O&|F?fy3}G}n7pBI3x_*0vG{Gvj(WL2DpfDG6p6GYV98@r-US%Lxz!-(cssGnL+1kyY%Z2Uwhu1wWVC9?SrKOSg!)llkIo$y7b7Zb3$!xA>sqv&y}P0%pEg1WO3Gq --hf81l|Iy0}wCCiFipED1E$!6Gnb7HI0uA&pObRJz|+*cyILZytB7WrmC|*HXUv#pf`^boz3%DK^t!2 -UE0-W201Pz=LRp~eqNcM%P~{Dzn%H{5O!xeI6{Rm`@xRW+#wb_Tqy@f+^v*}bJwQh40TO`%b90YT;COO&Sz>!=Gbw&rr=)@xX04XGO4G3WCw8fhZZ!!3VArL+wf&LN)lzWDNyitd}1UI -=b{1-NidxZfW!RDT2uJ=16An22$dsF9#9jB*yNz8qIu;i}VELJB&m=D3(yCON~pED*8aEnM&48;j6^o -|HwNk5;X3DeH!c;usBT*ICxy-R80+-v3*3thQIAcwsg1tyW+x6QV5Lry5P>tAv^#=DHj;$CT@)OIbXGd49R36mz3??U*YN@1%@=V-meNwUq{YL -wV{cWxLaW@RTqge7ELFUFMv5Prud}{0GpuX3@ejiORVA8|`q-0Y#|2FyFdsmrs5MVt9q=OB0Hmzy8A{ -Ty};th+Xd(=yP)n)F#RESroM71DLy267LQjxqLBO!x;H|Y_Z1dQ{uO<1-cuJSNsCXO;U#7%Y~4M6E|y -`0!w1cufdkJHXeEZP*yE3!C++2^&~s|6|w%C@~ ->aFYmOlbQ1G#|s(&EEDGYvO}omOxUMT&GOF&%;)EIUB;k5c@8i6IHCnxIH2ej0WoJm1^*hgJKq7(6CG -uCw0*{(z4A8VpG9j}o`nUFiJN^3f9V93NgZr3)e!gXvMi8tHO1BGPCk(uRDyRU^hUei|IFzS6;$d7I- -EJ#(#~}woPF5uhE_dz089Vi0Spv_O!zR7MInw&dU8$AzJwPb^^-e|a$6-C`MwcZ;VT~bR%@+M+U5s2HYObX=7ULw0v-XiQ*V_P<{5#Wk;3&S*{NC6?51^2~g@Re*EgkJ3gGy -3j{xNg%LyXS#@BJjo+wiu7k!#y+tx^7hzM&l)CJWCQTJJnLm5f$R*M3_9T3%*&4T9bIfK7}nj`2eBPm -PT)+hFf7Cu7(TtsB~MQqP`KnArdzT0f*5Ut;2KH|Jc-;%|K8k-S$!-Ub6^Xd1ZZO@&d%wai88rSQ*?m -rsTn8&)a_~JC<%yOuP$W2&2y&>5oTyMmGBE -Ccmz&HUM+uB((5y5~DPBA8Qk&i{wrSdl*S2X#7aM13$(p)%K4ZU2@Ag9aG1Uiw_%-#v*;K@P&GKT?|v -W52XuTD8auP8!Kepq?2kyp>T -jnepFXBhH8EDa@fjizMY;zgqDZ8U&V29xx3Ww*8B%4kz2~qwkhk`znY!#Hmj=mzW8OD>yt5=6%OKInZ^zJu!^tA0n -5dYU9~}=Kw3t-og*lAvGUJ033>M^}4Y1KT;F5D<`n-)xylVx*R;BHt_d+XW`8E&RNA?J73>vI~OBsJgmG5Ea23}-7l+cjzs!iGiJHEusUUX9oB% -%#$?`y!?gy)M}VqM~o#Gs8fr@dZk5XXDFEDO4V0k2rkR@8jS0wfntLj8mJ}5Nfb)r&e7miwcn9xrIHg -DR>XkunvW#WwAzp$2K4EixjIF*)BL3RERjDAN^~{iVp6M{`)>Qt*5ECQ<77`dZhJr+fF*D -M5_^nm8`hDcb001W3fujIam68qd+xx?>KF}|ohF4C{js+_Ih&%2VjasFfdJ|m?KQ!R6#m0}L(GXCN4c -xgT+*+!QnWC-{6kYnrK{Vf&gj}Buhv=4cKfA*u_=H*^>LxBPj*l?hCy0X|(KEIc@&I{J%9&#@u~lCxd~M&H*`AkI?$CNdlvn`-d4W(OH8X=R}-l=jL^0@ivcNM>nN25$DkX+fh}cUkpx_Hg^X! -y40wv%j$?m_#HXV4kY--mwJ%@LDYFq1|zS$!9%u}*%t3yCs>>WYM0G*lX!8sm^`~-PtGpd6lUg|OEy= --nmMw9Gv!?{joJvkmcp1!7RX`|yIK|tL_9yqb%$P6TmS6Rv8h5Yv$IFg8PcrjElg~AwRNe%V68cM%)& -EwlmbkFZfXEMG0*1c9BDgAqZv59E4lTwKft-Qz2funxAHlQ!d}(ZtuLIBu;$H -<&-DSLTvKx!>cE9a^ZlNe^kn7DKx8y9q}cIS9RremHx`nEr$o&H;_PkBsFWRjo0&t%lpLFvTs}<4L2* -LlfUXG;ko?U+8~X7oe13F%rE -mRQ97`*&&JUT(5c=Vh$0;xm -WL2`mBo7>{((ep%df$BA1kkMe7S9!Ow&IxcOfG_Osd8`t`|6MQJ~>=Kii0hc5i#GvYVnDgbG^(cu@z -M_oFuLG2sk_z*Vb#$05E!`o*teT{A%o_CpS>d@2J@k0Mx%cIWjMaL2qxp^qEx+ixX|=tXhguBi -`y|DDOVpm~iI1fs6kohjG(YE8tkYXAJl@WYgEPWXHGW0Ay%ET(Pu`+}oi9!XpnBqU$zuBHlTW2A-fDJ0j*}VGm -yo0}K)UefXyL&dH$Xu&?AA3qK?_$E+#5v}bu)*L*YwGZ;#V*cFXQB$o^KQX>rccc_BW>3izgp5(S-ux -RVUr%%86)!|p1C<>2=Ee0@*%z&!SMDJr -c?d;IQFspT7C2AQE>NeNI*mJggBK{J?DHZQTsmeAU>NAB$Swg8lOimrlhFIH={%S_8&FWQ|Ry#paGTyEJL~eheK7HU$ZBVVhPc}Km7QGaCLOdf1+f{tVGz@ -aPHEce$GX@XBj#Qn#7WDt$P)h>@6aWAK2ms3fSzDMNRt32M001Eb001BW003}la4%nWWo~3|axZdaad -l;LbaO9oVPk7yXJvCPaCxnh&u-%&5XSF*iV=IT-LAOoEm9s}_pqnlq6!f+Zcs2#z)7vD@4iD!BBxd_) -pBG0d^6t%xEzGDoH%S++J&?|Y0DStvCyzF7^C;U*iT?uS;^W>1$0h -ml?$lEwRj86YtO};;0zwGI+uFmI%#6l?n>~P8ibh53ekL)YcEa3dkXS|AxZ;uk>1OIHkk=5b{L}!6U4 -DQkQ>66Eonu0oitK)z@p>>MhquZ&;)!PkwEe?Eb?NKKjaCw%e+J;GB}lK7{cGslX1F_24i87kM1u;+w -Gdn+|Of@A;mL2!pbeNs(HOo%V)H-aXYb -If?GOYqlOoVO)YFBIQiE4UC=4k*O8*cG)kv7Wupxev}hahkuO8>Q9yU%~0wn#emPKUk))RK1h9y#)dF -Mt0cDi_!m45U}=t%=`yXO9KQH000080LuVbTf8{G|9Bw)04H|<03ZMW0B~t=FJE?LZe(wAFLGsZb!Bs -Ob1!pcb8~5LZgVbhdF?%GbK5wQ-~B7F$_Gm+nNB9T-K=VqD#!8cp{XLy&f~Q`z -5s{Q?Mpl;h0avu;#V6PW}WK)=x4=!OP^!Cs|*ZnUXQVzes=W9~FKShnYMmD8rB-RNR2yBH%+$*0y3UGAXuqgRW88phxG2l(Oq;PvoBC>ufIq{ZqhxOY^UBtnv -Z_^S>^HM@U+ZkG_f}Q8R*N#vVQ8GUl9%tQrLOeyWwxlf%`&Yu_E28$rOxYgkKZS=tf+NWr1`9=v)loy -%UO}u*^Qo=x`Gu*>vWlJpt)=IAAy-xYjd0DKL+RodkBj2@2FGBbe2#TEU%xA#D;$_nP6Z?R!k=%u;>Rv4>F>(86g$lusXU8IZa+q7CLCPj9U< -ypOzF@e){%LhZH9;l-HIi083nAfNxWE3D_%8$JBbA>}4zbnr -!m4MwU|3sT#oCo_sf$(C8aG&Wy%UQ&K7ksJsqlKJyxQc+XpDh0 -1U7)_0-*nhXd{TQo{_^AOWCtH-NWImElUIKM8b$X5OKo^#X8(alr`bvYs*yPzFbIKfXR$qYz -bG=dHV@um0E96g_d7@seRLV594R8a&b-?Si+K#dLR^4WKt}Zkq1R$2cO(3(XG1zC`u6u5mx<_A-U?Gx -lfVV(z#tE|*h>lyZZRh$&&v*8Z#^L@C)C*!#1XPrDurxK)$z`H$bd{~PBRd?n8amh+tPiBWSm$gXEqx3KuDSm?zf%3(csWm@OG-#<&9R**f2|hnnWf+YXb3~s1 -x3;xXCK3482Ql%516992gxJ!@!W`Jj5@XiYN(fQhb(BvjU*b2-sA8ye457f~vR~A=5^yfv#&a93jDOE -Ju7IDA)~vU>^stMLFyr5vY&Q;M3loN00D*pb2Pyb2yXT&Z6IKes}AD_aOGP7ZOOLE=sUUobiRcBmflw -On7c&>A`=6Ucg&K;)PzMjnSY%TW!H}K(+YGKB)2ym<4)&qCS55LUP>lg}SJ83N6q4p%j<}u+8V�ha -E!?Ni!z7=3Pc#(io7!K^XgAvZ-`eX$6$2GN`Bb5>03CQW -;6C9h_6TH$Ps&R1yan2jg;fZJ1`Bxrh6x2Lk61is-c@tFmqKAgTnG|cCSkonwFg$3(nqfzKkzQ8613q -+MGAm(k$P_WA&$G|uiBFZyvwxGewC24jKAn5o-3bs3laIJEkf>4*B)Ks%6dFLpr>@owJTv;Lz%euz(Z -MM)JJm{rU*x^lcVQ@-C-LA|*7y|vl${ZzLN8FgyqA@iXO>4w5@?00N^vod57lI>AU}X26u}{9`br7)cZ2h>Nl!vJ!)kL9}9c8XX~+t|P&YS)G@PEMVfTvARWmI -N;7i0@pw5pO8ml`!(V2!vVJ+jDilLmtv=OXnT*To -{t<~?Q9su`v(uddcs_~`vYy&F>u3v!vF##(j2USU;bnu-1ZM7t?lI>fvn8CPPf@A6dLl@2@c55#7!3; -<6s|;3ku~nx(9IHQk_vD{g;-JknchR+_2i9JW1)*jg+$fgV0YwSQjeflJZJ?bU=4 -=MP0dmd_jlUOXUQQu;-gMayVxGohQ0ZrXRos7FC!0w --~bsz9e{$K7)u=@qoWf}?>@BkCv+cz^|rq!#m+?JL|Ut|WNT;*Rsn4G_2Meli2M0_>eab%O6qQl?aG|TXorF@Nr`t>eA*{5=BR+g6k|$^IV2 -*&o3vKe2PTGRG^i$ --t(Z^dsKPE(1_fcv>dYGI+Hmve@FBI?Un0d=wo4 -|o(o^C)6cCW_Y9V1V9m#o>|v)JepZ7MSgl4&@s}>_FCG&N#?Dii;+GD%4u_pGa)~Wqtp%Tt(5BDPJ$$ -h#%RK2)evy|T0UeOCr#V{;1FB#vfC7>|Oo15)suN9-$Il);1NW6%Jsrj91?j;!H&`?XVK_KE9P|*5wf -n4MsrEz@<``IR_QMSalR<*VCc}>%JJ*9K8`y{rS3Y$X`?jpsX`cO!B%SsIaXo8vFsLpmzfX~xnowW;VFoZw>j@snuLJw`1X$MF+0iAcJ&@)Vr_KZ3OeV8`V -;u3d*yvt%gqzg1cQ8$8iAkwLQ>zwRcH)>37ve=!i?c4FzIOMv22*&NzVO<{LM%+{jy{|BQ3b0Vq?xQ_ -8IXqmJ>ohBpI$w6%k0QXwU@+$y`2B1w!}w4B`0V%Oz-=^ifES(@hs+jJ80{g!MWJh0VHd%U>xq{w(~t -dQ6k?v3ct-Dn755CWA8%j@ck0Mcw0Z&jB&vJCE&5Sz$4MajaeO9^W!ix2E -H6Ve)Q~Vj6%YQcUVR^MXSBDdOAHzYK_NWX|;7KcD$n{&~Tj=K*b#`dsNlESMucPA0KsPsi7dSus53C1 -KH^gUhCqrz5@O4o}6zt8vj18(!#8Cbzt|%L02JSd|Etfz)UzycM-i@U_%$1ax>)beePCoS3JXm#k8>l -`)L=tYq-AVG=%{lZx8GwB7U;R1n7<0Q*wjyC=RKMx5*CZPRF|uEDCdK7u-zL3mXI1`=FNA4XPri{_6jB?M{UJMC`56yX@W|CWuW5%Sr^Q9Q&S6jYBU^IC+Ow?mH_)j9>b -c+$8g^f;Dk1#~*VV6Vj1J9}`u%08dLWg&D}qce$j;Bi;kvfl@bG5d25N;~dH=!AEtNAD5XJ@$U&V=FE -O{k>kjA4D7ck3=5$pZ+E`3?cvaHvyBFmoyW&iW7@-cFZx_}9eMnGQ`EzLEDM1%=|!qMjmRPW{-xj+8U}s-4FlL08a0a@AA -|Rr&dF}kRK`#z9vG0S$HQMayP!-ZdtQtm8@zDm`?*A{7XTSM0!c-w%CRD*?&W!Q$%j{ba?mOcR7$|S8 -!*{1C>4%=OUYoM%uI&r&W+iQ1voce7{)(;_y#*h+@BsByX6R=_#ikH2{a`i)}s)2PG34MJjL9+G04f? -6%nQ5PoNYYhk3L^RFDCL@q6-ZBT32@S_&u ->4IML}{bgs_j)TErYB-Hj(a3$_ECWDgT{VU>*hqJ(V1Fi$UZ8k;+}%+>X9>|*w*ROrV%A{UC1uuZ353 -k&c|0jZ;{l9N!@(RfoM8J~WtR=7x}bB6`yuR2mF)QroTTlM=YEq`8SaB75mvzTo0t_ooW2HNH<)7ra2 -=et?`H2$-~Ibvm_vv)cFZPG@9^O777Q0|4E(kRx5#GhFg~#xC;5=xg9YQjLlA_8#Ir!3pv=-(8;$^^E -*#-$5uW}iX?e=WA1D#D87MTu2z-xYAM_qel9uq8OVVOI(37eLzidKj9RH}UaOXEJNYtk{M)C0t9f&yd -I7+VEZ5fXvQ5b~^sRt{*TAJ(@?Y^F#fOq1@;2u?YvT*{PEChWNOfQh=K%Z++NBpvNq7D&~==tMw!}^j~0F@ZlR#LIfOq9rpOy#I -^ukGB6sNDG(2j(a8_FsLESdAPfbYO;%R@d%i&L5o3PibCHb+)ny4LOQ!2pQdXCTSM@qStX2z*Sw1lAf -F1lkd73=+Y3%lKKgKnkN{q^AJu4}X0_iK0`rT9wempsPISA?Orabu5$@##a;L8%YEv*SrvV5bK;UfirZaN7{yO*#k8;q#`1Jgg#u)!?ZR95fbot@)GEvM_eZBcwSg*iT -!*=DsQH{)#mCDGn+0o-4K@CgfmX#o#0?>VJH0^pWc7{*hN|M2MPH>{L1S-h2EKp*FO&z0_AHE3vgE=( -94_9YIsNufvXq}8Qn^H}d~RS$bIwt5FMw#Htt-~i8!DA_n=huhoZ8PE&V8|nilJWM8b9nYUXx1|VCG6 -}vh5}>97V7rcy-+(d@_N`3ezZY3S0tGKzjQ)Fjq^v%E_vq25=REqmic{<8*za>KB}!E|IbxRX9I&U`= -RgkV-(Nc$Aps9Nn_*&~na%KgyVW+Lmj|swIuO^=e}r1u|67!<|A6@ej5D;7iE@Itrbmy4mEEblg_X$?;-c-ft1ne8l02Y#A -88#oU^!1JItS%SD;v@a{WGGrCbWIF5$5?4o4<_%Qiw?H3Y&{fXLNm1S~~R+A@>zk9YncBg9Om7&Stw| -V;L@pt=QM$s~XjE!u~l*D`QWw$o(FJFRnpWI;3(Pd{!hI>dtPGJvg%3|^bj0bb5wtN(ZMTKcVwr^lSj -9>&zO_5;ZjJ^!jGnkk@G}s1Uqd^ve!PhX~j5;lOQi`M-i+ha_q^!#tL+Hz3G;m}EOlE8;$(&{ -`!?LZ7G4^ypP6M79MZ@_31@-K3a#>S#gQsCAHUvj9aa`soIJcBIrYU%H7}64~ii~R1+R-otm@3o=+8V -5WwM83*LRW&=n&tk0=^(j#_f8hWq#yHOsZ$?vDUcRz)+U? -mg=7P5L=oH)~Zi>kC~4*;O`+fJXAX1@a^H(scHQJX>d2=;j1F&h(ipS0qcuC?3;-1#w;pCXQm%fP)Dal -os2zY#YT?Qg%BCjAgo48ElvfUULj<|n!xvkZjhj(|bb5!LZi}eZ_l`U=QqpPjd`?-{p)_)J2&feH(&l -ql*PXbI^SR5t&b!A{{+WWB_7!pufB7;B+&NjGm+)n+x7bv$_3Kl=RDL*$m^{_^RoC!y1y%VfX#Lq)iDUH;k7~R=j3qx~4LjK|tPs!DMezY%5MIC(MJlN_o>)hv+Uoo0!Si{rH -`jQANwG}f{7TYYvzqt$9dPIBDm6~rych!&zS@(KyUbWy2r9L=rO{%twUNotoSpkYOEL^BS~^nTK?gM@ -C}l4x5HltNLEmpz({h}RjXFt+u+ST4X&K)-z4i&{q`;0*FUk`1L6vS=-O-7bc_p*%n65`6qoa|#rLUX -Q@Ibwxf=9xUt*n%05p`!R>@XuEiU|0xI+|xViqbU3ozX^Nz$8%gKrM>!#kyN?W(&UAR%lTweukwB{^_ -GTQu36mkJZpN!-ImC)?W$n-G5}`2fe+Z2sDfs$X{@BkP_n|dRMFVRXd?l -t`tHbB?Fy~jVo*FTxEaIf)?@U{4n)3AUvXo8>U0AU=sr9p9U-xU!)q-gEzFid;vSH+tc`>sXQ9jO*|= -G$Ey*`p4^w9`;DBAlRWHti{mPFz8)b?7|kol3UgxSP2q*h|KqoFa3ye+Nt;7z%4Ai0I)XOUdx5U2pi5 -nJRons0<(VSAE2Z4`KO7f2{|jG<+uc9=#4B+l5r{2Yqwo(VpR+F?)#DjJ -5HrN`iJr!?vOxKBxwzl~dx7`gew#ftpR)0`<9&zAIbtF&KiqXW4q; -~z&!96jGU$$h#}slP;{iP`kSChP7$@XXse-}VuoV)>PLh&7C2I85`5?ap^T7vqKn%wJDs+29xeL=PE* -bu@4JFdBV&Ht^U?S3;!4Q9~Hxc3N*9}E_;*M3#=lj@1!fuXu2Ryi$&%><@CP=X)04{m!lxDce%XHZpo -AzvfXbft$p-e32m*50|!f(M?bS`lCGQ-;M!(*&RC-8g5iEFZ!y$?773pH5Y=u|P!v`MI4WZ? -+}V>srdAlFxy){Om0V1qRqOerB{>rP`oJBo7RT4ME1={eJ|e(s2#$Ew-&U0+ojMrI1DxGU^i6 -a!OG7#hJ|_^BT$_P8xPcha8Z_dM+5P0u#1Jh79@_gWr=p}2pV^#5;vW%4ULVI1&f)b`i4UMVV<_tSo9 -K96*r=)IJsH^Ij}PW-r|BO*rIe8EnBYvduffA1T4$j0>vQ69gi#*3YjKhbW_4kMI$mQsu^1dFGo$4Ba-;d?MjjMK*prg-i=Q -FK_(;no0%L&^PvuR+9h$3nj)`f5Y2A^rX{!ms%GYWx)im;(1Uy6y2v1ZgXWA*&|0*B4pE}Z8pxqbPd~Eh+k5fs$2tXfct7mv^f)(YGTR*BIQ9kq`fjs3EwkUEc!~@D+Ja -Z|xV)%=_35He1rE!(rfp&wCyfiokBQNC1tyB6AAtA6mBmsLwDABd7;+%n>R{5pNe`mka4?DY$F}{Kz7 -g$3W_HECHyLHsBDdA^uLMtANpwDM{xAIekI9|&X0fmjp+Z2h>irgAK8~W*JLW+cOQ -Rw^7`cbo6YENU$aHpD??VW$^ytw{!*+=XV-)+*ZjmR59 -ab9BWXg$3;|L&(*vd_S`V7}?Q45g|R@GV(P~$K}FGC%P$GSg~zovx|d+YJW|C(kxK^{rr-eC2`hvzVg -^B^p;1%C{rQC8kUDqbVLnll+W3kV%wu#1yWQ{#`R794;BpXjMub!4tm-OIK#Kg#d^z!b6(bXEpVQ0=( -1&;W;m^Q)~>JZ(=$8=nm9(wh55ElM1ZgVof_b?E<#k$QnnYRrv@eHG?Ja$nzk|mDTn|TiU^c<1F2hK*2+Rhx)sC?U!S}i8fbTW%>-VBt4((w%jW`U{x@!|Z5|Vvv1L#j41IM> -@cwK`+__M2Kwc;Dd=_iybDxAxR8!8Q(fOHJ-v5SX*J1bT?{3{AWdtu4QiJa{6p$vHwOfmv8d<4jduST -;PEyw%WEx>iU42|PTZk0q}sBfd|C!@P3gDXBs3_Lww(bY*@9=a--F);FM@5F3xSTFpu2BM3*o)qua7> -)w`QOep~Jk@{yWiq3CI;M18NB>nT_h7B)5>~zesGr~P6I}tvHPB~uxp`%O^rG*YK;vyiwx*a*Mk8v-8 -D}?|-5RQQ91?gA!JP*Kp*`!qA7AOkHUIvMm!ZzJl$dIn$t8TIT)Q8hG2c>{4?N5FzAvc$#wFc?HA{kq -jO7g-FiB{g#j@7TwiE<}3a2P%oMZ2fWZ0ar&j)mVXHRSV>z2BgRuKJ*8}A-~o}XwM@*V90ermsjTtq) -@GR7Y~DSU|b$58q%gyaUE=}bSU65p2e8jto@?Qd|;f|*V~ekov;Vp0Oh+|vEKQ`IZ~5WrC95G;*^W^^d?Ttuq(%fM -qh$+IA=>cZM8!k*&s!1wRcM{Tu_32a$?fmez>dCcmYn9{UyP^32;e{US*%hZltUTt2+9HRPNoy9kPPW -&bv5*;n_F6S11$W1-{RIZ%eXDlufmm92~afBBw9-p}*m3@V4x$PvJXV_(}j=;x91O-&J1*UND4d#{~q4XM|K#SQUw-U -VSrmSkdeqgE1LM{<0v2F)5!MHD8S$*jz(fMl)miUo7d}=+lHu|BrWTMlrak(GHy6#}fUXgu0Qs2`T20 -8M~R_~{=e5fH>!+GJ4V_DPN(F4sxcoCEjfxzND3%XlDM2|5~01N9UDB1W$sDxkT-cbJaJNVlz&DG+6i -Pa6k^oKX$@MCV_&B6j7%k_?#0O^-Y(=$oxHc!Jc4cm4Z*nhkWpQ<7b98erb97;Jb#q^1Z)9b2E^v8$7t3xNH}Jk+L2 -M9IY8}Ov<-{?NT>3zZ7H!eymIM^FOG&JE$t}sXtp@#y9(w85^-DUO;gXc?R0n%GFV1_0wyx{y?Ow9=x --qSE>oqf*-(=%iE4NBZ#{}!eFetraTWfmOn7%ifY0(32Yb^Wr?c1N3^TLl#y2=~o_flD=$6i|1h>oG$ -6DFEQI;S?BWKsc*WdekIZ`Ee>Aj=@2RgG6h3wx-Pm%Uq5)z^2TA3Aw;RaFdrZ{ECNsvnH?%*x+K>AXw -u39!IB-+cxWfDnXnD_dp3tj|6m@%ib?PcSXK6oc`IU7cNiQuKTCJjz<0o>tYbrk7+X+lw2?`mys0v?f -P^^}1ynUN1gWYxnfsc3L^nKP-)YE)USRyX1 -p?1bPp{alZjAOykFq|QpG5dsfV*4en?2+fAFT1FF`X+&mVpEim^XS?pRigA+Z~<(hj(?NJI5>bvbi1% -V6ucIN*Jx;-Wc8)k`W`xkKmDIN&1@WKp31C{)6{8LFg+(cgD2g`#oes`2X)ee`QH%VPmE6vQ0WI$*NE -c!!aDK&_;7H`n|CV;-YROSK!nj`f*HjcfNE*Qh4yVEPSy%y$BcxrSuue2$0!dpv#5zN^~Vz1ZBdPfe@ -3NQw9a9De3Qn@T?wCs4VUhozj7`+EA+@*GlIsQ7_eV=rI}mg#8k$8eFi5b$yx?3)eJzqab|}%Q;1*H# -f57P>8+IU}1*=MPQ8PvpO+gytZc4Z6O7N<)q{$bTUj9{73{b4xaa>Ra+%XW{d<_5|aMzP#|99ns|Uih -d=CR;pz-cCL~bSD7Y9u!>>-Z7m_j-vI5#SWEgaNaHE$Gkr6UK32~TU*w!B54jpGBp@*V^Olfw$IE^v< -vQwZ2``TJ#DTpfy;zqPQ)@m>bj*~GfisGnaLHtp;cNN=)IP_%Q?<1JkQW7I9OF5^|+Kp;ksS6P)dNJ; --ewI^WMFH$3+}-j(6?-(5X+9?U7n3+DdjT`Bxa?(Yb9VE~(V1&3V;JGg6 -|&&qVktHs@`pn%pZ^XXyW((S@34=<-hLo!=LVC3I}o5~Q^XwciTWO}b}ry8P_C)H0ai57du^-O0cf`% -eK(ysMHoHS&q^vCh6aqq5TuCNQ=a-b3aK?hgYo-N#^0$RtV5i=#q{N&u4;dM7a-9K^$I6~+0Jq(`X#; -Y}A~qR6`eid7tWI8j{2HVG=@7DKII`chmJsEKJw9Z@-I1F|T(L_%{^$5|Y7ivc6Y74QduG)0myjC*IA ->r(Y4MwATVtsJC=ifmARaZ1!Nmd3Uc(h%KuuOURAgPZ_@{zP -&Voilxp}!s`U%E+^08F&!==wbjJ!v@%iLQ0WiTnUIajvF0Pb;01aYZO7UQvm77z;2__B8B$o(QW8>ar -9Gz6oTyHMfq#u!n%f1;FA?h8w^=V3IR6;~R(`h8$Wd@8k}MVW2xuF9p?fO=sLA`S*nVDl -Kft<&3zX83F5_w6JY*hZI}}ig4Tg^K)o0QG{YAqBA|Hp|6az7bXl9R`mxT+5w-UCkF;&jrmG1s!g?RSo*u -dDstjvY>fI=IZuVwb0``)hp=Bca4-SX^?Q~dKwf#Re9$_krwWwfuW_pHTL@dCR<5z|9m{%{;aC$&5<7 ->v+!1WUDL`*Z<2amRTaEn8SEg+&*oRCQFyEd^a+E`)rB{WW<T{uK0lWHDd`J*InL(OEyxs%Y)ylNg&cZ5zGmTu#c%K^nr-X#`l -J4Hy{^$VosEAN{(i_>Lz5V-X7%UnX?w+f&q5WVwP4CI5TfM(IFQ5P_r1Pzd -+Md6|c!w^^GNZNSGrFWNBg&_Iuo!RBbN|v1H5Qp41Z@%8layFYK2A?{#&gRycu5_J2n|y!yOOm(337A -0Ow!%THcWB`6fg_m_W3o27hEgkqrIWg`AnR5e2Y;Fh4OVYu>3UFJ>bfPG9`H%Dt!y?KeCnK7DeRw-&a^1<1&Eal~Qxm{Y%=HQlcLO~MCrxqkGf>SOGY%Z+GM`aVkYkflX`ENrEZ*%O(_3Ve74Tv56b%)O$H{rsjZv32tYLeHsRN|wHxzkIQ}nIA&M?d`k2?ihLs3vdB*!wx>OnSS-rqmVXJAWV<`2rfG-Zo=LpM~mtO9h$FZ+o+MgBe-TG`1v9gg~Cri;XaqLK`Ydwp+dPk -_ZTBM2)Z`v1%hXvLVUuNmZJLbC`#qAPO>wO&bnA@7hW@WC?IqPn}TG?ZFSRkyNALZWWjDQF1@6;Uu=C{2rk){KXEH~Fa|+!J`HKF>xPg%wA?c^Q2acCXcQy!XF8#9S4sSG -ef(EYU&JY47wDFC$5P<%Yry4;j=n&RB(6k_Q3j{v|E3gR~VkXMNCF8WBWx(ON&YCFOqPYbV;8p`SbHw -*Jx?8y|KhOZ_;x)MI=?^bZA21zz|}Uz{wC)9jrdWQk8>FEq);MDfxrNIP5+sXxu&@Rl8Z{_xB)4zwd2 -l|34(KDE!b^ik7V~`+iRhUpBApO1=R6t~}n5{${xL=hH#g@(|`D$IoZAp8e -JbHUB#Df1Ce4`}q}f0skV;2i;(6`g7^FMt7SnU1f$gz13l7S>_U*a{RHwi6y6Z)RK$0T;}96Y&>6tDH -?tst?8chxKyuG_C|~3e^5&U1QY-O00;of09jjH$!AdnDF6V;fdBv<0001RX>c!Jc4cm4Z*nhkWpQ<7b -98erb#!TLb1ras-FKx~;g*in1if$Juz3I=0-Xb!=BJ$!TRDnp$EdTD#MB~+*{s_5hxfO!adMeeYF?zvnO2LkxXw~tDXf~ -+cWS!KCv{fjm8vh3S|z1c%Sxx}CaW*ATpbDM4O^PbssgJ_moqgjN~LdWUFOM5O#sMpzCd8nMgVASWb; -K))=-e@N!2{%HME&!=jKgTm|umhRciCADkfLDwoi4roHS2&wn44ui)l8q&2Ex1&+-df{#(Yy)3TTY4z -rn-VN3a&GS#I{Ut|ECD=o4`6kqE+Ey`$F&H&+kBCu&OPp)))sS{{d4b>tkD;+~UX&sqr6I}xN_+1R}$ -_l!TV^i$B+Su4sbrI)Don7k~Mhf^Vn7U<7Gja*|4b}OwmZtKGCgD7pT)~Lb3Whdc0DyCvUDIur^Bl#* -Ag1e3sZEs^za}H~^3lWH4F!Mkj|b0Q92~_zKYMlf;@QcYqvQC8znvTmR}~(g9C2Bvh+ntIGu#*mQjDm -@g=+tlEEZG=AI_6Xn@03}SL=#@Ug#Pa`8Ml5$2z~x$|9fZyyh|-`-Lv!`fg#WXH_f{8t<+`i0AXNt{3 -qvtAO<@74tiMlzCq>6vXE)KEGQemGtVq>7x1TvZyuxghBI2Vf}?=hkH8AuajAp{zwD%UY#Dj;^))6)Hhk4Cg(H#Ly_Jclx0yic(Keb% -LcXQFtvwyrSmF7-e947eQfl(7RTxCMuFFX=>oXKFMo;(D2_`8CKN00VDqtG3I;sQY;6}L51TO@QSa&-Z-rYl`uqvq7 -tS+T5Cs%P0e1A$WK`Z9rcWn}w%_Wb$5+mra<_479`4qyKXdPe-Ck42Uu)i0KsIyI@?p5M(pAwa$;D5isyc -IVU>H7k7%Zt~?Y6hIMTvQtbB-K-&bxVB#5>VcSmqk?%xF-_z@~(LxF%wOk0ef875@tSiNQJ~HeydUgt -K{k>tTTFJ4RBIvx<1x*jd~tP%^gK=7CQfnz6-tT_$TE*{^j6r@$=K8ql4Ec@#*n_C@>F);JBF9piuR@ -pR!3=RPgjpDE037vMfPte`XrKI}}8e!6fkE4ZNpK^%iM)kfW@Ch%RR-u#q}C_)~oRcLkapf4%tS&C!b -?sOEAGa`qc71GNWc;Q$#UEkYUGUzfVP`#G5{^}%h}rV+5B(|a&Dd-%Q=yEnieC9)03K{A^csQzX*$sI -f?*2rCCY#TcdUxbQjM#!{zBz$wsszfmu{5@F1EtWT?K)c!_5>VE@GJSpt-))k7!nh}wdUB<%5f4QR0_ -nE;QY#}53YY@86rMpKC#X)e17_>#te7OT@3w@3u$)nAdW(|8%~g`8^nOwRmsN`*PhE0vs!5U;c?NB`e -q@W!|dH?#oa<1RZM*{(;QSye`x=JWl$qtYvvcgQY}=HG8e(9Bl%D(hyvmj0EhF^t>bV -D>SwQG{bQ(1h`W%_2;y!>ZvKm|@HQ -EAcBc`s!h0v>oo@CQZr;WNsMsheUa*=bDOTW1p_N%5xi~(GRn)OO6s}{Q_$65?#!#e8DI*0mPYg%?JWo2RIsX+3$VIpIu8qJ -R~I#GLEz0reF@D`A<+8?FR1aNKywgmlCgSdu^>rw>~&I*S(W8go#0rDA!;oLvrQ#+)n0V_6=*D-)VPe -ztaeLq_dDHwq7VoeUQSccHDF9-`Sz{~v#QvB^5pApws!}h@nI1_voV^?GQhQAjRLSOBC^at*a&Vg=+L -l0D;9?y$1W9H6k%d|^_os}stO1&iIRw3!Y4uzS&xzxGkoE9|H -X%lENo86&uL(2lzSz?-0SCcYZ)Id#IM5-DLdYU0*<2D>w)kvVzBHyaj6&l7jEWJkFp+CT@Kx+p!hR7( -t=@Pl6ts!R`J*mP7lgnj(h4p3G>&UsMt6)ed^SDTN#l46AZsRrjur<@5@vp!#rA<$(_x_$Y)iJH&hi| -CS59kww-f3E8n-MHAf&{ -tXd!aEV+70nk(S!cw5CF%}Ddz<)8*1dIO=*%YY>cj4!)(O;#K{saH`q}MoCI;m^xkWO2b+$V!XpCX7& -DmWTC(HUW1X~a`FiBSFSR>linD!=J0~w{u>dbn%0OmmewBRJ_BD)6Z2VlIPpbi>infoJgrzKvj{0MLf -19EH1wj1r)rf#o@aXIc!*+Dy2Vbf$d5DTwb-(nj4W9QTKb3oWnQMozJRbfF-2AWYBuia~Wc#95Zalug -OMyuCF(wOqzhMYQWVdk?8yk=@43^2JCAkSd$l;L#H))-AC*!)x>wCvfAdI&+4PK9%jj4Ri(^Xwc4s*@ -6=hPRs~P6}#W7*C@f?W_^-Ue&~`+5j&zK*D(JkqMhI(T;WuH6r|#C-Ze)PG)7OPXC4t8iNSA69gaw;s -RT@c9ViOH}BJ;ve!m?1&DP4leI&p>`hUoo$6E}#vwgypq@Q^PmHK`S9u0PsVL3g4^7Lx_ubt^hu=icb -hZpIBu6cnu+;#&h_)KlRkjdGAo*YBU;WV`Hq*l1&s;~3u)l-_0RF86o*|CoEMK$TqOFK%LC|9rsSISW5Eg+(r_*4!D?3Vef~QxSH$vIX -X+H@EhP0yri650;yPxj5n8;&?y@4f)B9Eq*&Y;>efjKL>wZsYU8#h-ZM_tOB<(82{&LucHD{0W@eqeiZT3Y&j854Lr;uk -8B|O}LqGuv{JWfM@xfIMC{erk+i;Kzx%lLzk7SCjdK@bC0sBp0GMpuaDjW6#W)!fsOF--zo@qB7bhV#~VJj -4q&CBip+IYeoXEpM4vA{@S0$Qu2z(fw9Cj@pqb)va7Y)#8^3o>xg>L+EC6!dv<~Ws?vc=#ujt?Fnu>_#S5qUO5K#m2sSE2ei@Wz -oOxKSP{B=$87IO3A&MN9$(tmvn+2C!R%(6);{isRaNHww4>_R~7yafqY{#hwz)0K#RtpYuOV;(e0QL#_g_xoAZFEiUqwyz0340b&DIZB01%+T`8h3ywixdNNOd5*O8jFQj8(cte+pIORK -4?tz_K@p~QaW;ET3m++MiU8%)G1qh(gJ0s+X|gRL&HG{^RJUZHdG+K(U5h+Xw(?^-+lL8&uGrVEqKPv -6L_^XaI9ko=9*glP`B99+G^^Z0HbU2F@zMh#gL{VeWHo^B1k_0Kj@eb=$x5gX6}HQuZ8XF;{P1eA3C)g#^y{1T&oE2wQSwSb~U@UOV)x}#U}q0kpJtV;IS>D^#3Bi_V -z6)n@(RmlrjI?8zOeNc?0wBUZVGDp4HXzGP^2vZghF|n_gb1&}Je#d-_2{!qLgaY>Nl^5~wSR`0}E64 -)D(J)X~f5YVYCWC+e89fYqKP6ddO4Nu4ZbNhwLam{L1Wwkr)!2?`Lyq+CPJESPBw=vN>py#8KFiV%i! -vV1aI0);7)Gu@ULbfnZX{}tzgQF}Bx60{e%XQ~U^)ore?-1=;DQb8qKAp$YaQdp{? -IKLO695u<##$1$mpD(zFx{MHPmYY(L+!KC?t4t0qTk!QBYM~hwp#HWggZ8w%K|bckRi3)9oZd8gIh#F -xRh#m#yJ=Nn@WQ6WA0ZHu*|`tF%(`~Bw6_-io-_TKx(gbNn0MIEcuKUcl9MyLIfNJ#oU=LW?R#u7zf1 -S)2zA-&|6lH1B^UkI8?ZmRyUaDFdd*cq{j5ISy))1q704D=I02bhg$#<`CMRb4PfcOZtIFAbsF0lkHHw&ZrAf$Q7t3;{IF8^4a8pQ=eE7#e})LXH -zvdXM)_Zj$Pkmt%wE-t36?Nxr@1VtmgU%(Q7Sybe32@dp}G+6>h|$r5XlOLGFoU3^a*34Q)MQoQ6liN -<^cbBS=^x0jE`AKJ^{huOcrDtY1GTRlrQO^i#q|KK;$tfuz0`yi7PYD(?f5*yVKoxmT_d=QSm{gX?Dw -D7pPYiCJD(+i$;ZtNiSNoiHHhjBrPxaRDwe9<}phFJl$vK!Rl)=GH(T0|+>o6hgK+dZr^94 -m;U0)Xf<`5C5yimWv^U*udj_*KgM)hqS80Qnk?AkoFej>v&X1kCtSc-nCv5FUKsMJ$$)ZTne=p<1$wS$#SJ@nO)r${M5Zk{czFj%DLftnXi8MVF4AI&= -@u2J#7ndIVc+(hj5v4NgZSOXnQo#~0}>T&nKSY;vTj;t(4I1N(Lho^WL}ORwA1&AcVohCOT*PCj;%3uwjM;Yo6mZfPA$)N^Y4XW%kVswW>uO_AB -^1WkUyKyU1(MpfFPCUJ(7ld#IepaxWu%teTsS{@ebLAPQO5hk8Wm)Il)yKuA#iWQ2 -y-_%GPHxaa>kvi>1;c5|@qG8^ETF!!tfSKbp8Ahttp73=Z=qql{7uR}ZV5Z5XMSz+HX;*c=ICkqLD~jU+^)qgG -OH}oxrW2nSLn^nP!qSyQ55AWg*Mq& -C&w60~*VA|3LkU2&rwY3``u>?_%A0XV81V&z#PX`G^HnpRb(I*q{u)L^#2A(`Io2bz{rM`q-_t}811YUG}1v&KCqfQchULREe@jv5!Y5B)X0`|5&x449iGdx_S0EO*3122xQ{CAI=%4A -0PJi*PVBufPP~u>Dn5v9r3SZ6BQo!B!Yzg_TSRCaB?=09h{0-4qfm81uiTQ+Lg!0)I($HqFv-*gY2ov -#xLBVj6?)0At&Bf$V_Wj!x=Qy8*1{uLGly6P%b)~V)%u>AXlOT~{TCvO!W-KG^V>XCqI*2ld@ -%YrNho#AA)RaNjju2BOb`KHj%MeSCdptD30~w%!E}_5sF)3C`Lv1c%3OMgl`!^j)S3=g9b&fh|jS}njc%1lxBt9Jax<3L*&_uvdK}bRa4Gptj$CtJf>n`vki*3iwzh -Tn%Qt|&Ii@wH5X*Mn#9ifspI*Ot``4r6lhdDn`Om-oJvpC%ez>^IK3>h{d9nD -ltm@_U&F$T90jQba;qKo4qpu(T$CGciqxc`ehS`u92Xcd6xGb|qm7M0=x|$@kC%8lZGiDK(SH{Cek=c -aC)!eaV0owf~k5^6sg2Y^FKDKNs4iyNWzDQ=0sG4s-JAQt6$UA)r#%SbubGCGeKhbEZ6POkOEQ1ov@+I-sruLd*>uvtMxSa -{U`JIxlw4{cI(?18%|PcdHRHzQOOfCmzmGSN1AgBoZNq4RvU%KL&Fj?qcaA&@+-nqSu8Km$!i2=hv7} -`CX@y#IoIn+7N7rgz#H*>yErTF@qlYgsXOfHa-ZX@RZZZw=b2lp2RC^wf~XJ? -_xTQ`3#KxC8!0ZYM%_yg>J7olc<6biGv}txRAFW|l%f4kQjDV>-yWHlR`*i^*6aX2TB6Sp>RrN0>$5Z -Iy2LVo??_2+JiLR?>G$XgNtR%0@OtMZ!aAI?)QMm06oY4toq5A9snTX|f!TFJWjR%ocPW&dJ5|ge1X> -aVm9JEa@<(xzz^5{xU=` -j*E0qT1v$SI&clZc$&*1$E%SshTOe_u9bfSl;gxigt1uY93WjoVW0V2p7&cQmP;Fpud4iYN6nG`01qq -&=taVd`uFCmB=xI@761TLXMFe3F241H`?JtBVI8uA{2ZN8RWEu^(CH2pNMO}xm*k%}x2J?8|HiZ -2KvyL0FER~ed18#({AxCKuzU_)TUy+;i%n%Up``s>L2_i(*xl_`06`%?bYf2nX+J$(q|;2HuugQD!yXn?vCY -O55#Ypo+bW%f*5#NqgYtE3Frl%4{rrb8YIfl*37cD|)dJKh8tPJ1JnCH^2yzN@cuzH16<9-witA+hQQ -BgCp&}jtEW6i!?3}I>gF4+YrxcX-whLQ@oJK9FxH7SGcmtXaW=VmB6~uV6Y;pYz|XCi`cmqBxd{Zu*| -DCDbbl^9*x+zNVM*WFePJ$EP;6|CKSa-(#}-`209npdkDInY~r{oIh!h1fKtY}X(C5L_%@}TogM6`>R -8*XMq5K_Z8nVU?@xlg2$HOLx1JnvF%|s%hPl1P+#o2Y&wkW)BlN>ryqn}9y -5ct&m}O}QYck$)!x!{9N~LY`w5Z?mkWiSGI3ruQ{dUz!);3a1vz|*bq<+tqriIN=-2i(TWm+2Q2r4s3*nB5EyM!+r#)E6JLm&=TBw#m%01%Fu40kMm?FaVHn2}ON-Tl5XGtJiUevc<+ -rMaJ!+w^tY`#>!nvw19nuo}SSsr^)AFqr)t12JK~(7FM7b(+WG{>qC@)Y5HzeST$)UtN}UIl?4`aG5I -zD2Pg6w*u(C4LU;v0+#D|!;%c$07T=kLJw^9U6;&W%cg?d#K@L6D>m7V` -NXq1Egcq$^tz0??P>>fVM3gS{J(Ao?axG?ztDm!gq{om~VdFSt%O|`6juOdqDK+w>rjj+ee@2VG`zxs -b?ezl$O2GNt4gM$mSWhj2~wdlK#?b+c$T%hGq|1Y5WL;ES=Pc)Nrl#8J(?^>t&~MLo8KBJ<4%>XA8$+ -LMR_`Sfj*nnMvMu0*-EN>2;)@`B_D(0!1}vbItrDuX;guR2700F3}zxjl7IqdrxS?hJeHX+Se1oh)nq -!qrkbB7AF;LD%**!ngZ%6Oe6b{7l;DE*uJUXhws+eiOevdh?%uB -fE}L))+mn|Ed*H7y6PYEdCGR3L-X5gsjpeqrnaK4`+`%iti;co&OJNCPZ`1h~ReEa$vVrz ->>yic2SVv}Ht%fu}qpbSMJf-D|xi8laO(I}zkkVv1Yak0YjyyXJj8B-`#CTs$oLpc!IGsrf`0(JiNzD -jWt+FK|xT9dHy$s ->Po3S--ThGPpxeUxn7brd|~52+lO52=jQiZ8eHwu9=ygPjMh4}75GKIh}0SONVV+prnC)=oVZP-A$=w -}tq?*euJhs`j~5JGtfp2HDADWs_XlGOOtR;a5;TQcuyDGTMovXa}@9%=^cMzWDB2AnRp?Z_8?VXOi$5Y+`gKmOGq0m&~mx-Vr@D2RGNi==>9T8^m*}qb=m9* -`okbgmLMK?*iIr8~WXIZpv(ln!?tmnR|d%u2SJ(AcwC46XK8$t8Cv8i4!F=Yw!UcPnjuq(Bk6bAK&33 -|EG30N3(sfW4mFvfh_cv(+3;vAu1a{TSXhk?DqKdB~J#$U^@SzgeXb8T=Q-twXDeFm_HdB_Cmf!gC&` -_8LtjaR61^be06Rsu#R6<%JcD0XBP2}F}E -9&@NVE1O#k*$7<9Hhr^o1HBeIcYc(!yuTM9l2Fs!x{>;woQ?MBB&!edNk6nF+f{ba0XS4Y9m6Et0mdQbF&1E*Ug& -ht?r86Q&xm#(_lyW$t{ly -@c3ru-zZ6#A<(U<#@le{n}9_&ZzeH`bXs|P(Dz>CJ5+Y(6TXGs&O`rQHfI85L(n=rE*E5r2a&SKMDEy -DC~h5JrwJ`Bb?;+!^ZTb4xF4JE@5r(!kme>+08~XK|pcIpfIj4#-G<#<+TAMfrS#CF0je4wXV94GmO}RdJ=d~H9d~rbFQQQRM8zy -{97Y*aVkJMgqOdsvIrsSsy;ElC723+0Xnwi>z-kYW`J`aaqgm}T%0f0f7NL~V^dlwv-D$48vgQPO3p* -gcI(C|W0`DzN^vta1C?%Sgh-(p0pKb)UkTaMp2fh~IGd7yT5Q7tz!?QqsQ9$P}loFF$JKBP5{ap1_$v -UX!dpOZAj$R4QkO3$v{&x_d+1b;O5lrqzBL36vNFmA_RW;fuCRGB9I7mCZsTPKa1FMfu0f9cW+ -t;-JI4&1}Xh&E@a8&T*8H~q9yaOfgKl2|$N0PnAy#7QQqa=50VGR706Irjf7T^_a7> -hg+|e -0g8^mF0Un^P(DVqB|e(^iTCngC83PQ{g8Ex8IeY2W;-k+s~~dJkwx~Zp(AN4<1i+0@;w$U3#)-EXKh-!_2{F -KO!rUDfIg2@YXUr4@llg<2LFKGHs392?GIfQ&Bw1;j*W$4?v!b1ZtU;ARQlf4guM)XmQas^rtjA%xgK -zZY5C#bTJ`NR%s--1+!uc&pUj!PcIS&%%hvdUV9=bCz>itS=S6xKE+nK7i$k-Sl%H@0rhKrGZYZ)}@4 -@uhQcqyTq1zheiXZc#8YX$+UCx@Me0iCogxxg5dvAPYjL5|FjiBZW3~k&ZtYki_)cPhv)8369v_8_%S -Rif4L=5|_ouU6xb@pwpaj4nOyD0mBzc87T1ibg7@6aWAK2ms3fSzE!Wt;^d1008d-0015U003}la4%nWWo~3|axZd -ab8l>RWo&6;FHA{8MNU&iRglY$+aMH%_dLbXtXj#9+OE2&suCN=NElMU6V1|zNkWwv%f>TGpS}lAr)f -7}@PGL@c^Gb^Kb<4IJG_U_A3uG`!2X&)48A`>ex7mtI*)BXPd~@+r)%pkt3$YcBV_BwtDhzxdWZq$$z -^cBtPHN9^GDpG*CLAwvnjvW5Jq7Ts@+)}ZQQ)xLfWXEjv=o=zN@mOo13)TZk4#!HRSNlsY0W@zx)uydr3sZ!V74shRAh*&S#C4*zkQ)TSJJQ_4Q8PR3R= -@mLeP+`e~x5VQEL=7YJ8PxgiZ9NRy7D^5->D|VF9%?S#fR%YckkX4AFEVrjpRyawU@*ORc7fK(dYC8z -ys>jA~I^rBT@6R2M@ARCALWRb$g|)DnmplD#G_<4!X`ilv`ve0)up)bjuFugGsuO9KQH000080LuVbT -Q=0RVblNs038AV03rYY0B~t=FJE?LZe(wAFLGsbZ)|pDY-wUIUtei%X>?y-E^v8mkUa~+Fcd}i{EE;? -K?=G%Xm#i&#Z4rn?JES*y!u}B_Z#hZHCyhvgqx$3wpJ;*mp389i4Hk?(uKbezDx-eZRU9fjP^C=N6nB -ygXG67bp<*$?7{g&yV8`}Lk}}}JPkn|7$@i|YYGqLNW=3_Z_u5IkbdK|ZjjmJ7Jk_k@dVJ!z+|dY;li -P(REX|EOoAXF+Qz$h8wnvQ!dk{nXMm-tKmU_UHC>Oo-w(a1lA+U189J-)q@~n$=wE#;`Y*3%Gj+>w6t -^Uuwn_Y*vKDVpO9KQH000080LuVbTX~PlW8?_{0Ma4=03!eZ0B~t=FJE?LZe(wAFLGsbZ)|pDY-wUIa -B^>UX=G(`b1ras?OI)r+qe~d_pc!4CHBB$U|(Id!5~etgCNOb6Le7oflbSlO^igkBsCK+`r~^pNr|Lv -Pc}i&eP{z@5y|A`{XF;ZYHOogwp!Jl>x@{fm~0Q)IL4LIjytK9olI(4c!7 -5U%a_{_j>MMy%tIsS%rr?Ck)>-B71V$gr~RsKmx>CvlTUb*0|MbYD -H7e@u1yam?a6Ev75iKjn>VQEf$N(WD+FaZeS6w9K=1ytuWK*sZV54u_2)@pykgQVCWEMtdx@x8)>#+;!+PE3Oe}T__P7`1O6DGQev}>d+r93ZE!*=4?+j^ -}J%}nCcHjmn(+abLY#hv!daMmu!}uU6x8y-`wqDm#ZC7#ObhchEatEY+Tw1~2X?NEi8oym^UmBxLa`; -ZT)gk3jQl9)jJS8_RDY-Mn5X8<`-?A^1Pj7sld`_zd(?*N>I{Spx*>agGqv;A#>s|r?fPJDH2Czrn+{6)w*Wu6vQ+s9Q72uM*;QOwq)KrI_50*WFaYH6&yT6oEgeiS_>dR%(oO-HB9{mYhg45tH02T}gCozdNag3HaW^XP+Znc_>}3~ -!6_y)8hG+juCs;|O3G0GL3Wd0PievqONJZrHQz<@^Np-&s2%&hS9UF@VWje897I|ff?Fhu|C8uvUC@ -w0x!+lGIkeJ2?AoP9umD(W-vgY5wxzTDvUaLkb&O;nrv(DpspKWEcBw$}ynT8qAER}?a{F`_PpdI(lAIXPqo2F>R;HKKsr7qtXE*ofTH&}6_6`h6fZ0!ZAcS|VO#5?e1MtB>A(rM+Qk -Y~e~fnq|8R<`)B#0t;05Fnmq_*3lj8(WgJtAInei?23kMt(~_M9qeNao*==miei_^`p1}{tH?Sw{8R* -&Y~Hr-DZ*XuzM@NJn-10ST1<-lS -GOv(=a&XCYFaDY;PSB%MN2>O9JHU4t=Dl+Jo9<47rJAqt3kR2uTX@|?al1NOtPT?luY;pc~yW**V)7J -?UTH4^EH~B2|e`8{r+rB&-0icpT(01PRObK}$oE}J$xReNf&{VXNFmPhq&Jcr0E{+Hn}wrP -6fr)N(&beTp^OX8V>gn;iO5n7C*&%T614`8%g=Z#(N!FI8IIi1y%)%yt_q+zC$2 -)A$7)kAsk>j?(wVkho#C?{qJ`gu6~vd#-kXp)i3ee@GFl84{`Dhp`HmuMpm1*hJ3Moy(ed)AWUM(ZDewO1% -nMBX^M=_D2MO?z7659d*%G{w@mSP5_v7g81E9*8TTBRD$)BsrS=%=0RKzY@6aWAK2ms3fSz7=A00062000000018V003}la4%nWWo~3|axZdab8l>R -Wo&6;FK~G-ba`-PWCH+DO9KQH000080LuVbTOD4#64L?z0QCm|03!eZ0B~t=FJE?LZe(wAFLGsbZ)|p -DY-wUIa%FIDa&%>Kb1rasrB%yr+b|Hk>nj#HIe_91(4mLAK^g=|i_|?7L159;Dq%~LDoG`E(2wsfsi& -JXZDZJ$Es`_C*;%S~b_bhH-G*1SQkJTV}l6F8lPn~&k6fRncIGR{pI87X7&Dbb#Z=qmczuP=Y{2r9tql8JLJn1s^%^1~WrLZAd4_{(cR1iDQ#Bj^ -%7dxSD=pd8R>>E5>RAWYyur20dxjL_;a=5+aA7Io&l#*nABd%?Z$;EM5K9jICgR3pAskNeGcQMRNMVF%W=!?mLef{()+=cSyCIY;ig#K*e!iW% -X*MwR1dPS4u{vL&1U23dB3%M7alO$Ss75`D4;_NDpY=!(59DR|3XU+M_8)kF*a(9g&`$tDHK1KTN=VL -gyF<$mzprNK1b5*sDZ_Q-R)X}rmjHkmtat>WM33BSb#s)H9Q6~+iGwm->O9KQH000080LuVbTku8q;a -LX&02CPj03iSX0B~t=FJE?LZe(wAFLGsbZ)|pDY-wUIb98cbV{~&aaCxm*TW{Mo6n@vQ;Jg`!tpu`hT$mEQLIK5J(6-_WclwqhZl*I<L(UIAZu&{PpqediML<+2zI6`>}pCORFl -qt-w6JhZ;ngnTHFM*)8cO#f5qLknV`RdAK6c3leRfui@_o>dZbZ4mn%ygb|JGUBNRU%jl~OH-0qeF$P)RV_50j$DyOfAKGl!{1W40W5IOFrkz&kRkCML2qECJ_ -{ey8_aSZJ@qfPVKa9IY_?BjcstP+q1We=pMACUzd`JcF?R!tQ7hT_T!{Dg}Ee)dsOOI~nA6AUpr$P+( -0d>r$=Pv4pBf*sC`>Iyzs|>&Y6!SCttdE`i*%Lc{P^ob>1rq=!`G((|X?Ks%nEDdm2?hr=5Zbd4+eVb -$Y8DEucjzEy~+Zm!CpK -uHzy>{w=ecjz<KMR+QkKV -?_LK{yiXXuW%pyY@9cDR6ZFVd#>sP^Q5zax;?xozv$6ndjvb(Bo)5StOa@A^JVzjFcwl5KJTk=*Mi@X -5d%~a(J9b)`w3NU;rBwrOh2YqlEMIuU+9pT+40Bv2q%ZP6m4QoKJC&s^HX%_91=cmbhjY-FYf1#fDs9 -FI2gexGS7PMXG=JcLfx%}=kf!?)ktIz?te0xm;pE*sDxYpgg0#5;F{66TM8y7DvrL|gT<0w#v#mTf585XF^ -u1S#)@_3DCWrt-dl0Lv^!8#5jtVV*){XCONm2R8}_<4ds3J8-K-3c8SGSBB75cYC{e9`$RQa@XMJNZ$ -u@{i9J&tYt8r-K1PzOE+=`PV`TCiA%LQdld~c?@b`(b`Z8vlY?CvhVHiJoHKXb#xwRyXs9K}CO0lMbI5IwNZjijvG3I2F+=T0wQy9A)zL#ln -5gS1J9xL~|c$pe8<;=C@|%}@=x`EuqsLDTm7*0Z+5y4B}(yW@Y0`B1`F#r^T<;Mknk%;ql-#2&MVI8m -9(4Yxf7h;=j#bNUC30G$9k(6xc{EA0*^)r^yxa%e=?l{)vz5*%ksKK^*p(c$W1~Bu#4$>xXs -BBrx6$CS-L`k?x=IC{oXEjEYAK`0e~*mZWArajO9KQH000080LuVbTd`2$d+`7O0DuAj04x9i0B~t=F -JE?LZe(wAFLGsbZ)|pDY-wUIa%FRGY<6XGb1z?CX>MtBUtcb8d2LcbZiFxlyyq1Wr&dDygB~in2d=B$ -stRR;ks^hJI;QIL>!mTIKyH~C+v765(tqX*W(0N#EdWb_gQ`YQqq5G~{AX!y-0!RvSwqe$Dq -Ar0Bl^Yc5Z-r!2SV2$0J~-w2Qe@MB?m^k2t6Db1%1mNA!zT{I^2t~UhrLKXf}6T{%0CFq4?8}p;z6*1 -x-YN?|n*XbAiv9zrAK~;aFsx=S+PB$0w{B2IyBA -Ja9qeo;(>j-ao9l4y?Kcd*{kYPda|KQuF`UOx+0|XQR000O8%K%whXZ0Csr~&{02L=EDEC2uiaA|NaU -v_0~WN&gWa%FRGY<6XAX<{#OWpi(Ac4cyNFJWSHbaG*1bS`jtbyU%6<1i3>_g5@_N{i1QVCX@w7jmUg -O3BMnf}_~$M8uYSD=EFeef+H~%W+z2Fj&&=?9A@S4e1WbvT@cCmLtFYJ -jG<)u)+Q*dj@0Vry@N4<-^!U8W;9i|GSfN6aRR)h1N$gt8;8}E7uU+3_5>WEoSaOv$5$Iu^!4to){Qi -{0C9z(S=nu@Kxh_*Az|RQNlRn8BX#%?6F|w%a%XlykCiEe*;2*-^>$o`!tVr?(Yo6G}jbw@6aWAK2ms3fSzEFR#o@yO008I)001oj003}la4%nWWo~3 -|axZdab8l>RWo&6;FLGsbZ)|pDa&s?Za%psBa%pdFE^v8$R7;PWFc7};E2f;1h~)<$Z7*q6?WKoRZmS -h>VAxD!Y-*d$M%(`Pj*SE4p+<@XX6Ad$!;R4`lx5Sq-e6e*wVgH&gi_iGM_Sn=X?%=3b)??T-`847to --*3exO1_RR&KEjo3+yPfnYdo9EVLY|M-C55%HHpD=DvWOfnK54yqT -^0%OFHO^D>JIOs -Hu%Lct=O5oEEx7y`T2s*&U^Y+*r)5s8~$35!53fXSmc*@1T%XED!x>9HsxaGcIz8Tf&eL&c@>J@A@8S -exP!o_Md?-ud0jVYT|EJ0vS$4q42NpAl?1Bj&qaG4gc1e$K -@(@?^qCl(Vmsne!dbJ$o&#!55UzuQxB0?#WwTF%3rUkgPZGC-nC&SM>iMntxDB0|XQR000O8%K%wh?r -F~OP67Y`=mr1)E&u=kaA|NaUv_0~WN&gWa%FRGY<6XAX<{#OWpi(Ac4cyNFJ*XRWpH$9Z*FrgaCx;<+ -iKfD5PjEIOx+g=RDFOUg`k+!6hbI|DWM2mM)8LAt~#>{H6j1rSy?wbaoSQMuq|10&di)MJ2lxIWLe$1 -mN3ge?HWrCa$~H^owA05p!P9t-zalGyx)AkU1jN)Rr+;(a~(r!wZ_ti$8e1XNtHb;N!U?aIv$YeCWde -RgFjnExI^RCrna=}vIT*vkI9z-$aUZGovm7pNzlPpQ&=UbA7^m~KHl0^R~==6u+2@SB}SPeF1R%Nj;?5nLVIBwk9 -AuF>N^;Sgv1f%PDpZwToCuk(8#RqQB6**skJHv5WY}jwCB!ykZykTqEuMR6ey!ymW7PEj$ssgImtYt; -L~Tgu?7!le{tI&g;6rHjQD7Z5#pJN7gLU4R)q{qUqqKi9;YmXGN%R>^M0Z18(zZG^CCISQaD07sbW|p -4_29wu%VEm&UNyx9M_ci-;5^qFpA=lDjM}(8%fHForI)fv8u$N|C4)QS$0R{P4(YWeW009oEk8Mcmsy -}ckt+z?4zn=Y^Q`zALtn>iwv?}faS<@y+7;PakcULI@)>Hdec@!mR@?xv?2S}8dkTI7P)h>@6aWAK2ms3fSz9lsaj@eT005L;001rk003}la4%nWWo~3|axZdab -8l>RWo&6;FLGsbZ)|pDa&s?oWpi(Ab#!TOZZ2?nwLIH)+eVV_`idUu!DJ4FwP$u-c$PWq$c{(rP2%Ht -XU5^tp+KN1A_M^}E}}VJ|NB-|ccZTWDJQ{$O#ywWuBxtk_ogoQY_-~SZCCTvie>w%tXmcrMcKw}Ru;| -0g?yKkdCn8@PQJ=oUbkhL_bHuWkXBc%AdVXZ&Ei0bqWJD}+Oi!2h>{&U;?8x5cKc_q -0m!?Ppjr7eCismZvLR-|CSDoJu(3H2@kXt%CEV{QEs`%KVYnGnk?2@>VuaY8(S*Q0;Hj3)z5wN_a&`w -O4*yB=9b4S^)GWF48RRVKbJ{+?2&8&kz;r8-UT_c(*WG(`0}V_sfT}e8qWXyau>mD>!he7um)p*PPlC -#0DDxGLn8>{p(pziW!rrztG*IsH)4yEai3MqVS#kFk|m(zTq`55^i>^IW{*seit|Vy?l9bvHI!Vhua@ -+fBtZLv-zOKUFc(`2f9MR6*JWbBIl3_O(;JZEvk>auKE$}3)^K#7bu&sY@$Q -QAmM3jA5nm~+!bYwp9&{lSVXO+cUChY7T8;Ud|lU -N?T_KDN3N+FUzEMD%37KJCXq_8(>n9=d!;1xiTuv%qB)~;3>> -kZF08c|BB&OLEzVnVl9!gOv)QNJyTKiti8YvO2C0SBu|oliARc-bI@!{Pu8O7Sl(UqeXiI~ -pp*8{HcQw#PIkX`dy==0F)#sWn8q2t?3=HQ=BpYB1Y?h?2p7`$k)H%tNIWHr2Vo(OKsi=`umnZm^rMd -d3CINNsXg{2r^zQc7!kpnUQ@)AGIFDuIA31B&>}{Pms=3+pGd$UpL1u7IBPh2*|v4I1_4aXCYLnWAm`cWD{t~O~&hrw^uXJRo>+u7rkt0=0fR`p`_T4oNS1?!eNTW` -j6kM?90}ERFHJSc0S^h-FgyqMLoB61cM0siyn~;@_=uCOUi-6#-QOFt+}~@DE#qJ+2%{a0I<)3-+WUJ -6!~H#4KgyO>286+s24HjO6YmXleSoIwFEmTK3JD##9eTvw_M~gbQ#x!jSO#>g2M~qJ))+D{>++kLB47 -ww&a3Z?54+(O*Fnd-;Vo!1wOHPlfJ^TUtwJw6_#Htvkz;${CprU)d}sqRM*@lH0a_Nl`v5GSIpGg8Rg -SnMO>nrW_x8?5*@$uAC=5defLdBreI#unHhK~7`5e3l+O0ayrb%wO4prm~GGs5{u&$+`O!HL9`=m}b#(JW9*0Z>Jc|AldhyC?F}d??rLmnGqyW15TSlKdO&n_{YD*#4S8gGESF -%apXrjqw%AuK4v#$AqdO$IC?InNrph9WD#>@cIqjYK(73{>GV$84%;to?gyW6ovSj~)7MS{Q_4>;dIg%Q7`64~irs8^vXfDoLw>vlYGbHXqcl$14tDgB9m;ywL -?RT_Y|AcBouT6V5Ff#K!ys^1p!rsyOa+rEtCMaILK~$QKeqxMC(w1 -(w#_9`mx?HDVa(s*;AlAh6+S=sa$~fX|9(Xc;Q5Se4a;Y%Q`+2Ul%rS9B58TI^LzS!{>gU!ud&i! -qQ~K++jfc#A-rJJN6>FiVp1I5oxrbj=uMQ2q-`7tZI1_anc3@>WnCL7|T<#gEY -SzvXzImeeA+K^t4GqZ?ex@u6L?tYK_RAz^ku_Onx1-PYzb%gIV2 -Eamv8LY}sVkMOLK_Zc8BdP|z;K)E9^diZD9Hn@eu9Tjs-+ae{&>edk*8A3=c -S%`cQXf5>F>RM-q5HV-QWGmqay0*-BMdmwAq29!XV(Gq;;K%bUmvUwIi<6@UqOAZD{gAeaY -C9!T~EEf_Fz0&ryk!+AJ5qNL{Sf7wbAaHSgh6R&W!a+tjl&IlzXq}BhiNnEX^9+j05ztD0Nnj{dx3!l -(a#TDr$PIPtso}&_zj?gs6A34Yw|wedv9>fLVaqYF=Yg3n$uI`p1?R=qHf<&rs9P$N~WHa2GT%@s5+# -H=o<4*HRi8zp%UD`;P$bN|^_V3LrFev>};=h!wkJ -63=>a;dXlR{zN0~z?I7=6hZd&nL+x~es|vJee>Ib+`udT+ov&DyOsk{QimFJC~@7osivJmB)-CB4A+w -F!96=LnuP8kxKTSNrT?+hXMfFcd~6klHuIG6&5MAVre6PQ7uIDJ&Oiaua(c`}|@tM*sRF5gpgQUjrLu -SQ~r>k%QA?9tqcuR6q^X0Rj&E8T<=QOHO_b;FHy~R50{4ME?#Dc@--kpHaa`3T^=Vr{M>LIB_VlEZhs -)J0G9v`TiP5>~lYS3#IRf05AhDcv+a2iXaP}!fIf&`*c&|Y3MirHUdy -@NMIcLw1tvL<3&bYy+JHxA>2`z3&@eWvza^D|yeIRouOp9=EiA)U-jhUSWhm}OfeL{b%!3;l8{Bu~70 -=xMW)RdFWE#xinqb4mLPhi0uOFG}(22LuFa2PLZ3gN!=OT`&l--nlyI_B@mEjeW*;1)OS}wAd;DSFqk -N^u~9G{fue)WQWn+$?<0m8y->7P8e!tRD?c-(JTRV}L=fnWT=6y9h0k3fDVe -!h6F7J=Y;fRfQCWRD2T*9}#HL@GPux%$N)&P6LBTFL}I`7JPkvtY6AQ+{)uE*Avm<%V?Kj~VuXH|*e*=979k -4%^Biv~x(-0YM2UsGL9?USkGwy?OMb-~U**aHoSZTzQEr|v@@Z$zcMS5>3NAeU;KOt@eYMh0$@d114i7hWmdEkkDvqe8{9eH$HLK^qt74`H#D)Oa>%&H`GkEM7$zic -Dfqsx+9R2uu!kx;)(#j8(VT&0Q6!uY>TEM|XA>edB@31mij_mDG%wWv#Tfy -#E*}es`O8rWPnmsL_S^1m8q8KXmmH$&OJ#yu{UQE0QvQV*kzCXTq~b$E3CD=bfc)6b -yqgR2>=AQwgWN4-?Q#bZSL-2e|#tvem9&^rKXmm|=Cj#0%ZqG$4Rb;_6nX8%%*de8NQ2zT~x)< -8#7e=B@%^_lc`aTQj3$GGBRcc77V8h{IMBqoloMhN<*hrPMto{QNBQcPs)_8K>JCaA&GO4lIc-FlwKM -Z|%isV;1#4SmAOf?l<3@PK#$dAKfTn)TYAj1(m)PwRFq!P)2d>IW#K}pA4LvPj4 -E=X?hK(p64!`%!4Xa-4%vI90AfpjO{n|A7|K&& -e*;MfcKdMFXiFhsFe>U!(1XmlvSfK_APJvd7>%YW27MkCL-Kju0&WPQ;k?!P1i^!o8c(MNu{Z(4Xhl0 -*X4_x9S#&R4IcFt1&MPJq8JLi|JWx;x(*tpybUB&l8F09;% -BA+Qg|tWEhPNSeA;P-}z=8xG#}?kQU1w1Entn6(ZOGp&Vz31d&)#Kyc{?C)83v_V;bRpzkT1l&fdaLgBC4S#aKbF!0RA4jV&5709_3>hj#-~vkyoa$s%w$!K_rEOyM-noZSI+oX -Hns>tUmyNS3|gWL8wQckigIcosf`Da3f=&i(n`epTX>-8}93il+;jehu#%xhOu%3t>T?PVEZG`XM`Tk9{13L#czLo1WOpAbTvJbgY2=O@Uf>gmcCHgJjO!8oxSWf -F1iDV&eXiIm9;a9t(olfrNNJsjy~!RyTUGvMlj}D~bb@Qbp{5a_&k?p)eeqIk-{;hnU7b*l5i5-zF7;?q%`R(qQqz&FChCLmFCDfX-!oSgj!^h8BO1?h(N& -~O28wVXg^9frIpDXrGYAdW#lw7;%^2jd|!>YKluY&Xr_vnQU2MLYEOJ8y*q&8!|lc|#l&fbw1K$?L(` -zAHmyc73g%kYJ2sH|nM6Gp(qLqyAbmETLQV_g?~!dv{U#$qPDSbFBEBWe){-?MqDC -*5K9|S2bA|7>(pP`Zi9!10Z?v{JC*K(a`OW3I|wo7aJF2QkDmNMZg~n2ex2Sf%Q>#t($okcUD<urm+zh_QF!V4R_ax_?EKOS|bgHXV$Ye8YEbQIO4RksGyhKUNx#;A1thTTi&_I}P@(%E -EOluFnbj%_6BXGu<=s5D$`i;eNXrF9s2f>%r`|3MbWsJEPD -{Q3uB6CQImQWNbhsLiU~snJ0w1;Q{)B6sr2R7MPU#HSBREQ7kGE`1iudxo5Guq%Tydq$Mn&MWvyGw=P -lfxn%tzWJ{|i(M3I6+Xqn?9a7`mi~Dk%Jw>~>1Z@c|z&t8=Lm5XWaBl4HnmxqAr0kF5c8xeNt_$vCaR -eN;k~NH;TO2AlmL0i?ZGDvI8gK?tJh4XGC#iYdgivy?)do^B?TzfoKHzNDj?NV=j-H{Ccja~!@+jp*SIu`H7FGCh>1}_wzT{vfR -v|+0IxXbLlQ~RRdekpEyeP$4pBnVd}g;b)5KDn3wic67(E&%od0w*s19X&J~y1=*^#)5@V{}$sAqU4| -fCeW61b^dkBPnp`H!T$qLO9KQH000080LuVbTmP2wN^1fD03ZYa02lxO0B~t=FJE?LZe(wAFLG&PXfI -4jLq$$gMO{%#kDD+MzVjX|(^oGbH<}goHdZ-)kmum$%WLx& -RN&BfLDnJj2fRJ><=LI@LoIm1aC!@2wladg#n(4sUQ8>misnhT~`qxFa-OJ)TSq!9hK|!NrUo0qziLJ -J{g_H8dy`;biEL#=B#<)}z7U7HaQZV{7Dt)-}`F451FBceK6n@DMr!i|US8JkBDmsr$$dK)v@CTy5yw -6ks$yjJ6?;80^sWQ%fG++r2&8y8^-I^gcq;$zuZY8N=DN_DFv-$GS|1-ufy+^8+@{*$b{a<(MsroI9E~ugOVvNl5NHmlv`EGilcmrw1p^E3RQT{3;l$uSOflw52!XQ -%Sb9>TXe6;A0)EeE3w{a*vKs9SX^>+#+DhsNnxlYVF~D4c{(;BlO-@{DTlu&NAex2uu;*ho7Q@6aW -AK2ms3fSzBR)sl@;Y007Dt000~S003}la4%nWWo~3|axZdeV`wj5UukY>bYEXCaCxm*U2o$y7Jc`xAU -qFi2O6U-`edvMWYc6OEt&)~Srm(Opwbd$aYhzZl8VRsu)lrJC8;l4_Uyw3h$WGihvdEIevtF|{DVvjX -2n;_c$Kk@;Ek1eQ?gv?9k*6iTP7=J3jra$lNB#l^Z9%>le=1J%aobrTJ4xU)K1j5-akg{`Cm`B$<6!Q -n_urA-bL)XIz;S-mnGknB4UrVl?q@H`)7rhnMYYAScs?$T3V=5@hp+`!4|3-^s+asQh)I^yZiCG?`N| -l0dkVCn7z!m!X~Mzj4H)EVskCy!Su4;O0c{=_FQK-ITlp%On`YLoZQbeMWgc+mpRj=kRRPfIgP%9K|Ly7V!*g=;`_Jf9W>s@rto -X*zbAfMcH{EB8<#HyzrlPj&PL|?9**n!#*{#+}d!RSae#Hu2*8s$f{T{^Ex|FH3<$-OX@`^#3tC9zZ^ujpNwug;o;v -@vY%gsAy()<6pt@^9@7C?+W(%5Z&7PEzp2ZB!%}!L-Q0#Csd0at_-=M;(M-RJcC-_wy_DW~}2lUpQ&N -{WMR|yD`DZLtwnY30V_7lc(dg#@^oE;f;fwJu%_u)F49a2GEYZse8Pd`7SeCU;Fn)N5v}C -jVq}--i8kZGTjuF_)uYbN~1PiBD~)2{L74iZ$CPRigclmj)e2T+&eB6D^7P#VpK{nWl!r!Bt-+XY~XZ -yz0m9O%Vs7NM;a_DWr|uB1D3q%qnNvO(RSD7J0|1z+>FO>^lV+p1qQfW-J$O5Nu$CaO1uaZ(Q#rr=d) -?fLbD^f}TLVkD}(9Tg`N8->2r -5%6cN7Y$9y#7cM{~7od!a@Wdc)l-bg(${VazebOgBa4%PY+(WZ{^ACU=S{4a-&=js9Pee^c)A -kzNGd#?N=^&WmeDn~hk2`N4}7a$A)6x1AS0>n1>$FS1U)h=V!yP|q|2CrG$SrJU+d -3z2?FQmNCXx`8LHGSI^MybVjh#JqztPAsIJG4qOqZ4_uG~LL#u_eXVU4@-7_bsGj@XRfnqjKk{u2$NituPMQVhY9gdSYeWw)i+QbZy`FnvH -g>IgHSGvruuI9_xe*sWS0|XQR000O8%K%wh(rsf?i3k;@uHKV!b3m;3x -?flL!-D3=znLNCC+~PJRp%EXdm8FXXGLzRqQBC$T&`T~+6tK6NhNIaR~GFo+Uj2W}9>qqxDz5mx>r!j(Hzo5ST3jB6oYw$Y_Cvr6VHuC!**#LD_=ex?5#Bj{yD5qEK6`$nIkP -rmn-65%-p^~vW2zMw}Eo%+JuuOUz42pY0ax^g}6Q<3U>BMMEwENmkhGfELU@iy^CMJu5N64HL=b9O%xRAhhLzke0|)t?Mt*foZyD3!1Kk=zF^}#rD`9bB -*)(!=DOV<*lBj9gb1WegVsl_4$&e=4lm~f9M8S(?KlOSnB0-<@8HEBYHo+=IsGRO7spcT!jvi9S;r^f -l#qek`XTUm}QOcSngPHI|vT2tEpyyI6B25TER=cU^>)Nww9ai%LYd4~ylxL7DX`Jx>sH5311dFgPVXL -F{0(dK8j@Fzb@m#S%7*5MvDIIUsc2s&2)u!|Fz-LA=!}dbzJDq2#8&_uIVX$ir8HGJ(E=hAtEEHy{%r -_@hlKMx%ZmE)93sLRco9~h>x${B9nBP5lqQ|qMyS3Qe_O*HBJvU|D(~nB+(E+SYyTS!3L-p1v;DN_o7 -shNJ%{^kS8|-$^A)*Le=5ZfeKAOcIQNMdOO#%bqo#YYWzLS1fm)z{WyHpD_IGEe|Zw|G(ex)MJT)EoLVIatL*J$YU -FMO_<9Qll&s(@hd44^$x^o45)KROhb-$qi_NcIj}>tCmfyeb3|fkPNW0i -YQ7Qb0_psdY5Wz1#bof -$}nn5;*i`j#srao2zR<-KO%sE5Lz80Dwde=q%nyJm11oNUPvPz2fkjGz1G@Pb6+mDCl@E{Xkpi)Y)n^ -yNsuoKN?u6Hpso8UuURd4RTivH>nO7J(PYBwm3`OtS<+{HhGaV$5YNV(`rE#Xo;O8h*!JTz6pecoBG> -vZ9a!j(Eqe&SUnAuEFF*P2n|JTM_vO9o`~SN0#l0K9{_*bhdmns)hwt9G{=>(g;X(YrdHct^``dmg;0 -lGUAYgp(+4$P6@xcx8aPawf|4WGZ?!$k4|M{QBukQ;< -7;1!4?Y?n+!e$9561hSj`!af@BeeW|HtY(;~-6j)(*l;sDA&WugKv?cky+2=lbEtw+}zPefZXUPS -|NHAN?%(*z0L9(IzumzjIsEHy$%FTAss-}km%q9H)~yF`-8uZ^F1i2b*YAJ$y9e)ocKG(2hi~7i!?Q{ --i|tV_lu=bNc)bdX{z+s@W_8i6#auFN@oukY2M+22@1Af`Jdj~sC3$PH*Olj$A`P>;Un@isLdvS*1y} -enqE*~_9BH#g!%(!gYZQGdS?e(xi7z@xaTH}%4@PsuM2gC&&BY<@qSOnaTi<$Sk8CurH=Em7Rnt(mj1 -Otvpo*o+cxOxVk(^LSL+#eOC5^~ie9VpYiB-7b?`a<@NMhus=$uk8r6_PvZqYxHr&Cz2ny;ByL06fbEK0>fttWpJb5wN6!x*g^d7x^VD33QWqB@P-j<%?sdBQWi$r=HrDhjxyJ$Z}yi{VGLfBlgx4RiL*vyR!8t$0>xtaDi)(& -2G0i)k}heN!&*AWCM~2alAqJwn$r>PDMdYiBQ@vC{y&al~QKn0V-O;FQe9+=wl2~2FGOUNXb72 -F-??nIP9hF!b2ED5OEh)K*(74zsU08-f)`yJvOFG<^C*FV?=ycE)#fnmOPXDpOsrS>;Yz*ID8icCP+Q -MTrnWA6&y3@>qmR7uPIucM0&HCYxHw^L`#;(Cm6gOd_r+vIol28kQ&Gv%MaZMFhNEN{_{+!((QjDlnk -mjMcR2bGKd}a>u$L;1_3f(5ZoT5GhcMO80{P8+rOQJ2Z#SMO@8}yFc~|A7_D~>y0;E{TITwg!sq1gcI -@V~MAm?SL5##A;8gpBOvPTnPnl-YpAfRkv!5|Cj(2u8=g&kfN1WN6nA3X2roNZ7>kG2!LD -1Wsw0e0p9kgy#pGS*w`$DQ~|$!7HM#C-|AD3Sr+NQXBLj5+_x+QIC%n@O9vj>bU(nGud+x%%Iu#`sL$ -NumlH9_@mT#R)XA@A%MJHA8WJ9u8wo;wu>OKxi7q(X7MHE+PHUz1x06v`H+C4z3=FHT@X2;-XMNgXra=X;&Q)Pvxq6SF -X}Y)`*zj%R17*w^HhojMmIVo`>+{2(D>^3@4zzUGO`OCo+saFnl_+GESC -Nh-ZQThgq#A_ip?!%fM=)$mrYI53_MLPkbGN|=8o7?Cs<$tX^d$K);S)IIDamr-=R;)TG}h{U0IweUU ->zuNYtESw**nk`e!8k`nvozqNAN>ZV0o;A7JJ05*)w}Y)SB`@#NU}A5cpJ1QY-O00 -;of09jjww)mB53IG7%CjbB-0001RX>c!Jc4cm4Z*nhkX=7+FUt?u#Y+rY2WOQhAE^v8WnoX;xIT6M8{ -V6WE8KIS=@>0oS7NUZ}>=ayxG%eHUpfk$YBJ=I7bLzQ$b29W)kzexUJ(a5b@6TR-_219`dimR@A3pu~ -@$;u|U;g^j%l9Au@#(|oPd|SD-47rC`Qd-xef#-(1$|No+X0p{;t>e#}(1%vqVozM0DW%)j;U< -(iM?=ZcrOq)7aGJo`38kS~(WTks^Vt_5lm^toPgTT#(uVau -BP>Z&wG)Qy){6crOtj*o-o>tex1iBl8vw{bV3>XsF)?n12znE$gZ=2)l7Ln9IKGh`Dyr^@`Oe>1)_(m`GaVk?DkegxZ%sWm{$^&wV^&A<`k@?imG3t=AM!vA_u3K?c)Vb-C(<7+idO9?r5Fz3v8uCvXxo{;`5MNvMxZY0F^Ryu;Pl!5)uIAusQl1b~H--#zaFT~E+ -=Qu2c|!DaYMm3FKp19$Jk(8(cH@oNTvyW<;g-%XmRw>48V_L8V;IRnc$IWQna|DT&;#4{K%HB=w7J2y -OPzaRL|bfKB;)f~=W~a|owER&Kwgerk4&%20M`(RspwViqeynbsStIVY_LcYR{7kp!Mcg@k0eXVt=)M -ShdgH~E==^Zci6@+u*1UlDYtgt*t8+fI>&~s3#kzEGq!Bnv9T3Ho|8Na7A|BH2tzT89PL@rM9KqVf6g -jhnzSAv&*g|~Gnz>G20JY0>|G(oKJ{$Q2_JP%aZaY=Zk@dT4v6`jz7*oaly7jc^(BWZ?pweu*Tmd7GI -L9vUT_b35yRpIa~{$J;=1=DThVzVI4AH88vGwOlL(>V(h0c)qBE{m(PMJE0?z -3Mrq{rwX}GxLD^xL~}qaS;_-Cp(}(7m-2wP&O!>@Scs5Ipmji#I(H+4JRo*m$^&9!r#zsuq*sNFuqj- -G+9eq3lqcMTI-<>f?z8hX))5h7suROqd6ZyFY!~*#TQOtdR0poOW}&nvHJhwglNBrYs5e|0{6S5{*zu -1#G4|)94$Ke->O7ti=EL+q>c9?nB9v?CU5K~I;NwJK@J5^ThGXMn2F#IH0^q|)wF{v2fwb;3zjpkcQFc)#-)e5{ -oI;JCtfC)S92(dF_1Q52`C~xI`L3%!W%sYSJxo}mwv@BJp-6W3EqjfVoWsDRX4qjR0po?Nvl^+D&{-P -w0c2})<0C&CT+4wo9f@R!qTm1Y3{O6R|nhV;F}z&NbjskJNDIgdK==A4Y -JcxIEcYfomDp6W$LnltD{*V0Q%tUd8ud9U@tS-R?Qd~gPR=)^L-hmxLj`goyqS4G+qH^o@uG4$q)G(3 -b*vJA7)Rz>I{cVeV*d>{2zdeHiZ66ZM5xW1rHqZDbZZ7B6y2jlTMYMD|xY%#H2*b=uY*yvTJI_fsuZ* -$P^I!g0#DpW0sjg{!z)0})7eAIzydLMP*Zp2+4o8~Imd{ENCRpAvu*@YLeY@bNeBM#Lsyc28P4gLm{_ -GDuRLx~F=*ARNL)#;nN!IyxN!S9X-k5EIs>rsgHoi9!j{qs;=F-*e@M@E~>J(x8 --)QM&9OV+Tj`U2d!&!|@E-|K&&VDb6#%K6q(@>2YQu~ah7uaEMVg*_DfKE1Pbt(DxbU4hlq?$ -3J+Xujg_i)Loh{f3o1_w)QBl#-EqihWF@=Euln#f%U#O}5?pQ6Ts}h(7D7)}Td&s75^4EW^TcSH%^;4p#jc0BOv3c$Hl+A0MH#G!3;^lMpK|5ZK&MSCbvCT& -O8?QXPgml!SR+_adgle`1&dD7)}TER{r@hB-+k*b>Wdtipknd*Wr*S?sfKD79*MQ?!5w?O;UfKD8GHohI3#Q9(@3N@u^cJfbcm- --pN1k0FCo-AD+z0kIeL`fMl9AR(l{hgOK@e~p(M2K754m5XK9=~DCu`M*sh3+n}%Ki_?vF2Ln3#?#4R -(XJSc?e8T=6$dS4mvn5g~E*qO2Ks?b@6&ZZfBA9>XhRNH!@>Z>X))T@?^{(Ld&mBHXYjONfVzww5rmg -p6rKN%XOuaUi4Lw|<`n3*AK8j*TsfDAqEjGY-?t_=R553PSDu&tJS45-@p<`2Bo@;iEz#q@H)iQh)YQjE56i7b;(&U-kJ2Qu7rg^_kI}<4CO -;o*CCOOE}B<&eHeQ216id`MClhyX#}#U(EGn>?>nERjn%{AHG8S)hnkh3B9kzJvW@MsZUY9_$MOf>v~ -@D{fr}jK0Y%>%?z~+0iJQt`H_tEnsGHwM~M3VVvi?7kHMld1g#IA0R746cLul5;%+%J7>eTwtm<5?mJ -fxY`<;&lW~pQU!Mc@B -{)FOUgaNGoEhV3{GRq_X?M{#M1b{VXjK^?hi=zv>25jl(EG~Bow&!<-JO3+75X}A0eWV*T^VUI@`sNz -gL4Gku5ozJ?)tdDm`_#LQQMxk!FD|#@b-*c!Jc4cm4Z*nhkX=7+FUuA7?YH43%Z)9b2E^v -9Az1wmeIg%*&zF#qC#%ye6Y}HoXJ=5K@X8X7gyJli1#P8i2MFPn_nzm{0Hmo-*4 -^3q@smF+82-EJqo^;RSG2SYy!(#)6i3y1vA3k?oC9O@&`I{*z>qCrRe&tElY#35Zj@=|{kW8xEy5%uC -E!Js1m4dM_(J|ajSga+(2z%oDSX8fxN`UydULGFA1>~;L4llL&vOO5nn?k7Q>3fTG1x^7MejFU&0<}& -~TCJF99+lZMmdk>(zg8`=+kak&=27IuBG~~iTIL|aro@R+38l1lxKm+tegXUS%@q^382?&+teINHVDi<~5$N9HYyU(mdRQ#xe6|S<+*O={E%BrZY~Wk_vtq7h|U*1}$P4S0Vh(DK>Roe=_o`!} -S)=6K-jQ9r65rX`~Z2&O2eyMcsJ;xLnn_(vJ890?Yf{5W~<{y7&dp;_fgA@lSk*01+1G2LL2|A>!7=) -Np}wR>#Q_CfP6F1jPu(W0qu9J1STlgOYUP)ccni#(z^Z_}7Z4o*4rehw-Qr;NE@&*eFh7EMd@@n6F{p -B~MJpd4HBm4h!OC6^&1Wbm*t}D@RD9XNVrY1vp>2y#&_1XL#0kli1I6_g@(xSc#tDg`}G&{%=!1Mlfx -Ex|1>3-e-nR%`-l%n4z|?-QPZD%gYqaXs&9N{riu$ulK;X!8Hf5VI8t_E~-zCNBw>DwSRN_} -~BEAAb0O{qVz|{;dD@2mQDI(0^lfyg&CfOYYKe`yTMw&RwU7DM`G<7K8h0d@kX+t(AnOd@2PPq?c4~g -@Sl&!1OZ6@+2M$Ta`zFwyMvumi77MH4XKK2KnKPkV8f5Utp!^`$^0Q1U+v3;BtW4-#~kW%)$$-gZ)HU -YXtd>IvaI|LV~a^zr$C& -Fef{sx8SJa-dkso4~UjWFZpRx&py#!cgj(jj7mpVb3`Z+>41%!zyGHn`%exR+>E8LSI^ZE9VudUr)9Q -k2Y;a}mg9mY|Q8)gjqfM5e}M*J)F6A!Phukf51<`K7|69m!%3wq{U+|+=KWxkVv)#452cG(}R&uo1+ -8d=IbzuRTw%g$+(W^rcTQ@kd>1s}+b5IbY7MRHrhRv;E1?O4~XB0=-Fr0Y)sNTEJHS8$%W_{ -s%PfW~X0yA^=_EU|5uWOOlj}wlUa+8pzCc)Pl!PrkZ;=tph(5rC&SV!jG9Pdj%P-b-HX&onthe0REm^ -o?hysu#<{9nmz6|b%%DeERIGgM=TKuTtLw@5KvuVYX6ZYwMW#rXCyVsgdUd|SgM$4Z-%%PawF{RSI=% -W=}vbN>xq{b&4ne`bhcT5qvKI>Vv~-Ue!hGXdXY`Px{&#cSJ4oW`@tDL%t?bta$V79P>)HDP9w3?}oH -Zs^4(^!4W99bVx1Oob4_JM18atm8-8D-fV5fZ}rDUxb!4&W15U)+QzlIw*4Y4%^{StnVeKWz;so$ln3 -Lo&JnPX~qKFtJi>#4RaJfe21-Vm|tOBN%CZXpnCp{xKiY`1?@c^Wsbsn$US|JM`gr4ih-YYwWsnP%U( -DKa{+>!V#p*lcz#E)UNEcuwfA^zI1KZ9huSmD>6-%hY( -j$^Ge**##95}eemDBXt_0l}5$Ub>DB#;kyt2Q*E958+d+{vp$`ODf-(dL+K?ZSX{D2P##&wGg5={9VL -%B)fOB;fo05pla#wQN}G%q}8Q>NNBjFR+sUN`iW*!DlwAzqJfRlSt&+C6Tsm_p>nnqcFF!C;t?*xv~3 -5$gr{=#upd-|aN^51m?xNOTx*?$UA<^!}xj^PR%KB6;aikJy~)I^NfifeCn^0sbg}9n1$jHWAm^905M -VXU6JuqyQWPpc)MOfM=iRUe^x|Gw}f{pg@>sKJqa?V091yg2dj4dYj#rWY2^WBf -)(R&Yf%V!q=35bUw6+0_S=SGeYIyq?uN&hSO`&6K4aR(qKgso1gAOdT2Bn>F`Sk^HvmcC@o&!HZ4}5U`<)0CWy9Z&LN-zscxsF -hN8uyPA}o$q7&^m^M9P0^#AXA0qV)kIg?ZQ5$vwlsa;t*8jTaa?7FxtatW!ANVT_lt_ZV?Dh{W+E$1J -15Xm$8ze$vnKB5YG^W;6M_w?MTjQa%CzH*Nk1>njYQ2>Zb>f;P5XeUT^o70 -<(bzQc@J@fBhPZ}?Y=A(Rfsl -4_gt^%CFUZ9rSM?ty@}!5M~uP|beMwI2JD+V;-4c& -etHjGHuaMccf8|naTDZx1#YQw*QQh6&^X`Sg4RIk&f1yLC=QH-=1CxkY}Cm^IOgAlu~au!`*ECgxM -ItR0>5rF9>z+ASK!)Z)E$<$InH-CpmD$wc?41ttpAE&$;9jXstemc*C8JF6NVn=t4!G1UlCFsq(g4O5 -DU2R8%ZTyeTbS^6#2qMl*~5Jt=KC;%r`mkP42s>fa*2~pL=#)ae&Cb8peq_Lf+lO2;v$vA0-`BUWRL{ -#T}ArpW#X@cWjWX&Jo+Z+>pao6w31^fU;_=?1QuE&X65&2uzKH)Y`44@p -jwbMF;4tJg6IggO(aMUTD_z2q-^g6|}zfNQS5cL}@p|4YH2MY8~oaqDRC6iNk7Ot<`-4LiYbtO+s -{+KhUgGM(_o_OYnEv2p)5>ccxMBhz-f0Q -SgcQwj}te6*4XQpW2ANq^E!GPi^43g9ax84MqdP#3Lr)+59sYt^N1VArUqKIHv^#O|1~m?NyNELfau2t669IGKOL?7 -ysoV7j4***DuwE*THcO`o!zBLUd)x-s3W(_ih>n#3Z11sGEk-FezL+;Z4I(kaYGBo;nFh -k~2h7*bi}gSB~sD7zQRM-`)x|as3+Q@{-dMXbe-jvbl-d2 -uY?dX$&}i6H#~%5XdeYT9D46NNjJ!zBEde|S6wg?CUe}k)%=A})#6y3cOR}7_PoX`L;#fUIZP~Er^Sy -&Vqqd6r`d}v%j}f~;(p7ftg$e41z7*ujM-zgE7|*~sH8 -G-~Q$|JB&Gs7W2hS@A1;_{>Zo+tkan!ps^n-+9JN~Y;(dRqY|LbJMhD$H>@5DT9MNY4qvx -O~3|7kg(;FTjRmaMs8GTgA0^4ityTT~rQ9yT}>^+=AF4*Jf&3oFpCgFtqkwwF$hyZlS*o3|eS#1Bcc+ -+>TEREpElgt;ZD>X{t_47V((Q^#iretb!vO=3=>sr5m-eZ2-VLa~pMRfbg2QV?e<4{r$-k^wwEjs#)E -FwvqIV)?p#hzwr>gr;caf&EV#PYP*;D9To9h!$8%z3ymVcRj#|^Slmb&c->fWUV;e}=RfAk^;=(s<~h -$dp9!v@2jO#SMIUPi`QRxA^Uuk35Io}eO7UbsT!D}z;IYOkW{JdT3Ah~^HYc{l+YSvI6T8E0hlWjw8= -`H8h7F0W!L~!gWCe!pcS1{#abucY>NCz&iE;lO4W+N`nr7J~pl{Tg7TKh5UEej0F-d3t;Hi_4B -thcTQQ9ezP0*7%Kg(3gfVH<4Fs)v1P1%_gv_8Bwg=dnu)b6b*J(AG=$x}BaNrUCbxBH%-araIKxE6mx -VAk*Q&7w36rv^Z6mEWUQ(M!YAgMdwOZC$-2I5n4IJ(?0!z8s{RzR6p4_VTjlMvI#PbHOj{ -zC)g^u&v_8`N^=>dS(0}%eRXjK40>L~4KkS2ySD&VTuPN9h1iHX!rfg&OU*@cJLfC&}n0_gD -04Q@3Kuhcy>`#_K%%9$VdW6<=~H0SZwtLZdAvt8=uYAa^Ch~YrJuS5XO21Sh)|)QBO^P`)_co9Mq+ks -4)^plyR)$ZKzLCtPZPDo{fZK2O-*6Z{09}MZi&ZbfFIMdciC=M(S90>pDV-^OS|+1Xx*><)0CB%x3~$ -21ZI{msb!lQ319CBn9+D6f#Gsz|`tH)Pw^=wf1rVGHy9iarN%9az(bk(3o?{q=-bRLtoGp?#@Q?>iY( -6l)vct8fr{3uP0t8`l%k_jh{?-5LMuJZ1A#U`d;5U^nA#~W5hYuQ};urR}qYRe!~Ao0f=h|n!vUCrFE -ly;+f9(34*4nH}>abrk5uekVUy#XJEXLzMd5#Q;$(MX5r)(wMF;L(Ayb;FVrh^~O#x?w^vx} -E(gg->Fh8JG!Q3%g02!hr80xd!7=u0p`8=%>bfW%aY-VYM8p7YH~`#oZq{hzGv{Bu~1-Qd4!zrW*|S0 -;w00T`(d+#MCQ_Wyp|IEW`Zw6>K)H<95ERWlUY$SVmvSCaK62B$aKa1`z&8FHhy~T!iH&c=(#pCOaa=EYd<d^XG4LA}GN(p@cUP2u8xls%}I0jq3i-B<- -^~_sf4=^}SP2hb5tn4m!rxeVai<23B>F=`xg*IcK4Kqei(Wd3IVao_ImqGh%7&JSv{nMw-)Fd4m)hJ8 -i(3^>36){`hLqPEihZX0|nSST90(OGi8a{4`PFby3&nb*}T*ZKzpN{a_QBOU!*1Yag8Zd=QbY*m`Qfqtd40 -^Uf%4P{U7C|2(xfXL0HJgI|K1mJ$QiWHweTC4+*x|}~+HsHb{1>!od$_Mb+0;YHGj}41bp`iEFj}2>5 -A#v^D9~+j0CbWO@)J;j!oYiQ!W1uX^VilNjlT^Bw=ewH-sBcNaG)#(jD&}r<`>z!+0{x0WX(oI$y#j5 -5y60W)ov7vE0QO4x7|x=n*HE)`7P3TLKR(z8Xj0s-NI@hAix?>8mDG -=bFg=*^;XpPji!u0?1I1MXl?Jh6&O=Vx~pp -yJuWNEkD`5JDnK?n;M2H}i(z9IYVgl&J-0<)PVXWPMn9gn%>R?Vj?wq}e?Hc*gi*ua1cBAJ+h+QM!MO -jS}FJwxK}&R!4RiqJreRmBjv;QyUG5t&LAF@JBKC55Pk|tb&Sp?0Y>G8e9JvFh-`ca0?&_6N9NzJ1s+ -2(IDm9M{$qP^j9AP!fbIi<0}t(AH*F@eB{YI-vq$?ywdI$MkD?N46(WlHwFERReb&vYQif&YUf+DM7) -OLKA(zWtRExrc&Ii|tLOtNZ|UxHXv>P{bitKNN+PMwl(2~svF0K5I7g2p%w&82XE9|6b0CKtI$O5o8628Q}5dkW1eefoi32YcES9^Qtw(HPFH4|M3VH>*My+Rn2} -B7z`W^@o^0P%hlzEof9*x9g~h-l|>4HcVjA`=jgFdI>bwR_r>qTL{_C`m~Ns;+b(38d))P6_;KQ?!gSgM(jYD;$xVeac;~~6b+eYntA8PjH&;-^x1Ie1hzOnV~oSC<0F4bAJjd$_ -T5Pn{%DH=|Co#9L>TWH^!v&4X{w?WMP# -p_@~1e|k-uw8Hj_yHmw1^=1NPnYXtMl+>S%i03~?$|+}!5<>TCPA{m#IejYmNS*IQm2z|v% -h_y}JFi@;na4@dW35r|xvmyAcy`}gZuLf=G;8624H&jeN+#rk%^bhVC9q)J5{+9rA~Bg$ON9j%)C&j> -sb+BEL7@7%0oaVzGJynYgoyeL*a`1U@5SiIy*z1ZmB{>wThaoBmCz#6h-B!aBUOW$87km6{dvcGNb)! -@7zhO|B*45-nuf<`q^jniGDkv43**szs&wEfw|rmaMD8Mj?*+zfO(`%?;^#5@Dc(C@4J0@K^5Ipk^<@ -_nFO0@3h{0(E&~?YY@n1oq%6o!)@Mr8*Oj#&uE+f+VbwOf85c|7jzC%SW7_v;A_n3|w+(D-o+y-12eC -sI9cz5^F2FG8{|wqG{A{)a8qVrJU1}%u>pj79{v?RDMW -H=xm<1Nppx+9`gw@`QgCTRVG5?}^_}m4TPbc4rQY0-sa0bRgdQW3j@p=5f#_W^;#-a*F8506lLIt<~e^5!9eln;1i;0D~mknaL7^%!^A+uBNmOz~l*K0(5PkEe`sM} -S`Fp0ywni&U;Z$T$&a>(&osen>H@2bhOr-OTa6$oB@ftm(V@fKmW3meeGx4s}dDTa~bx?|k=GSEpm?- -)0}41C9^?ilyIC@TBMP90<<0gP#^4hjxG;kGf~&8ysCYcRFBF)A)7%tS$7ouhrEq4JTDxP~YZ(MOXkk -3_*1xHNG9MnYV4&{v+`SDU@Dx)}Q49)cyN$$5f7#g<7V+nQAxv~<;gS4|R_(fA6|*Fzn0FBXoTHx9ax --0K01LRtT+k8a9Be|Te$8>S>s2O~?cX-~{Uj1vTixCh{Oi#vDKbVLUTZ7#w=@^iSYY``o*3oxEJ3%zk -(_V9cU1}+!(-eI}9UjyJ(<`F<6CaN&zvee@))WhOpKsb0rOk)ys9CvbQ`|-e>8IF46lioPeTq#(f#(X-Pq9hB*mmZp1UiYR2er7QR~ -%^vp;W1ZJp-uYf*a5rcbHgS%fMinQBFM*S+-!Mfq$t|*mrS-#x1MrUBZcr7ne*_8NN#bQRG#9dKKa5> -Lt^C)H+vQWl-_QMLjVGG1pM-Lmu2h?J7bK+~ZED4h`3!78_rAbl%AWkZ+toJ<&t>6^Y-=%jmGqGqAjY -pkr-37NAvxr$u2cWm8v75V{S!WkQViQNNlnd_ip21f_ncld@De8c(O7j2$;@rJX|S{7VFlBI3f!S1W= -w$Hz&H1o68{0%c~Fz~FPkfl7S#rVAaNkKU@`ae2vnJgB3$0j*R`JQg7E9PXQ4HyUY2^txe<3Ow2)yl$ -AI0@3AoUN`I!Mz^y+rSM5im@&C{xu_5M4__mIp1I3f0fr*gWumXxENvoyemJ!G%xoeM>_Zb*}lu0t3B3@6CbW@uRk&2 -a9!uX1XNEd*S86{D2A;$KSBtzlQMr494bECG{NF_PLbuJA0hEipQvR9>(l|6&pfB_WY^bdTk^yHJkM%*-?lpiU^1kmPOsk;7cc1 -#L>sM$W3@?d5RpGX)c*5Gx23#Dp_KbJGH!<9sjl)cyL!Lu+;bBldqS1S>1DIq%)XTk$jr3B#2fsHMkK -=Ki`6JcaL+!5(bwK7mE}F`?^wdX|-WY(6I_q8q-Qph0Jl}DYd67klMi0eUb`KDqI(c0EXf$H$Sb-#kc -TmDs7AP2L16Pdfpawc=Z&r+Kpa#Cf(ybWTKPf8v$4(t&Bmvm7I30^htrCXhn|UgyRk@wJk3o8YTzrL_ -(Q}z&+&*UFU7V!&gdwTe?GI!ngJ%e$qO&Wgk*~DHhM^~a%=zM89T*kelL#wF6XCk(&qiZu>INc3nR-? -Ki-Dlhxr2qw2ZPFmFtBZT)GLzr#1O?;2eq%op)^R}4xZjcOXeb0Hhlimm~1q7LEWPqw-&#Z$GlaiAluK@f>QE>$0TzW~Z@Erjqh|Bc=7Dr%j|3(HRhN&q9_OIN -w3t=*J0K$ZmcC$T-hDQ9D>1v5RyF>dE0gi6R1?kP2h-7YZpCGY%pBY$8R)C4KdoUrnIAw;_S3Hg0$3Q -hZ6=gJbGu-Ll@pTxNPNyrx&r%fTR3*ZQY=qjovghY>ecpm9>X4VFuY7S4?CcjONC{yk_ERkmXll-Dy; -&xp-6)TCc~p2I^#-pu#D-am=vfBLVJmGc`!^cNEJp91us6;L3O?RB|xq2uOIH8>Dgnxvw@!0GE<>-T -V5o2Oyx_v&xFB&|v>Rbkrd!I+h8%hzIv{s_|(F^ID&9Kp6Im`veXz6M4if?Yc@5m49xY`#lj5ss~IqI -%6LjQ;!@w9N%xKgrd#{R6!k%*4J5IIU{2g?`6_*SZI%s9mq~P#oDW-!e($C1~$W#H(+eWMlO)fX4MF9 -P3cgxvZrvGNj433i3C8HZk~4ith+3IG+HAKKR_SBH)PhLjx-l2eQVAH&k;j%y%Jzax0=mTO1|y!y;df -yv;ZlsD3Wpz-SY%2$V??LNROx8I?EIp|!6g{QW>hsrUD6aQYG69$w&UB=nQGJ-oozXk5?m_V5C4qqKk -Yw1Jc)Ix~oB?;18S7{6<{RR@SS8nZ?+_w(D&4YV*X5?4#>X|f>h)^5REV`?{g3{4}YtxNZyF&^vC-s) -{==3bzehJn7gu&j+rD06CsDs|v5CM3~s5O(hXL?sE?+`?_bMfQb@5(etP`@sSr7?bDl8UV(~Nm#!L?P -5}h8PN$M6jInlMbhW>uRVmK>!B{kdZ?{o7wm3I3C6Dx@`A4K -tldwRVJ>bSI}OBDR8-v=|=$Oy>Q~ypKiXw7+igP?SO6k3ZPkbX%gq&fTeJ>xc|fkr_I0M%r4a4Gdby)y=^zT9dX&Pp|TReWdbs0~RUP-H(5_q(4W5< -WDLn$;9V~(jnA>o=kaOr?+>2Om<#lK$4Pa&dVFcc_eSr$dVU(U-Xu-0i6n~?R{~|*dV8(T?M^e(gwXf -+*WY0%bVh2?TlUO1RIKhL-uo8p?4PCSoCGQ1-G#0tO{LFI>bEJpwI=qL%?%Q3SBU8h<47d&;_Ygs9kQ -li@VhDa_*!JF6}&?xCCsF(>~paOUwqn%c(qZ37g_#?~GOC1RIipUGQNr>S&=)x0>*1cX7)pa79Xcd0S -3lD_WOZx#bjeg@e5lb|ITANMgHy?J3VLK;lS&b`dU)lxP=W)0iUde}b7KW!lBL(J9pIeAJaQaQxuSzQ -SnX^ASWX+&zM-wWnvNZQpIsZN8d2Ou(j7&U`gDnt<-gZS&RKZ`O$HpE|Wuk -@#mOzD-dKtZG6ahXdCpM^H1ZBWF-Eabpgl)+`)3h1&9UdF6~fTNv59SObD$22g -|>7i(t2_^^ncQ;H1lZ9(28s-DZ$lj&Kp)fg-JnTaehkbZ#vFM1ANKE?a#iCRCl9GNHvFMb#By~N=Sae -F;WMJ=tUCIUrlGZL={3_FliO{34=j*rW4uX6~*n}#e;y>Xld@Ro79Z5yu|0llH$c4>K?2fe%Ac4>T_iEE?3?b3cLD|@%BG -r`7-RNu<3Ds&@k?p0WIAJ4TFw3 -xJLXz!-x}n>>Y9qcbyT*!Zp_7sjwHa*&y>F5AOYsZY=s-<74&Agdn6(6FydtiwH{BiN?q3$&sDj-hrB -Wk_<@F>U7BRE379s!u!W0kPfyV6C#xObWHo0kenoN1=NoTfth6X&sF7Q(2*2|crNF<2q>(!4=RdTAhl -R!X}T7xD^1vXRdtfLSWR8@?p3r-IT)a8IA3g2UE$Gs=e$s+Nok#MUMADDE^TyPB&%|;cfx8ZD=bK2tB -nkUelh*$ilqGNRq#E>)}}&1Z*lIqBwpn}?=$YX1Yc$0+Vb3U$=|}r-X-fW(40sfj?u7AZH$Jm&)ZlvT -&H)EIj+|`=^C%wx1n}izi&ZK5?bXwVxqW}%8Uli+~$6_oqdo5KC^bKt1?UpJF*IjVu9Ty8Ddzb9+asFU95gVr -f!7Tz*VmZ}c9Z=HWteA*}d?lLDe7PMRV-DOW|JX}8Ncb7r2ak6*HYEdk#NG4X3qL|G2;!u4YhiWf+0S -bCh&e{wLVnR#ctj(|>F0O$&Ycnu)UiOYP4o!_2$;K|~I26XKv|-BNzix+3o1@pxK`|N7hT*kyKuiuU1 -NGWD7$hHihwKAEF(O$w1|f%V*UoYS8hu*6>=t`&4|o-%4teKI5wC)QL(q9+$gAMskZ|sZc@?x)(TcKD -%7RJ!du?!OAv^cU208sRottEX-X(754w>R&?~GOE1RIipUGU|Du8vN*vir0JD(%ZZb&4HP(@yhKr_>R --%ZGpJ6ndSBy(@N^*SL`Mc9G{1Ts~X5(SSxrf^IlPEh*`6%MGWPC8;a&bHgcOlYzYpcJUecRZxOA@V3?7>3WCiq|>Z6AGbGiVcjaB^sfKbW+`q|9X~-4*N8HjPL-gU$zSKufz?&Sz{u>~j5_kJ -oA}?A@^HZ!NOx@4-&)*4GMQmiJgx&z!MAAhi)CN!Lb}B -w>4Dy2#r|(?Q=LPLncE&?T*#R`o5>=tb+MbLvY?_H;OD;(^du*=tEK@wX9oH4&UB;HrOAduQTtRm^!yr?1x+drjpk+*q -6<)C+)f+}Tq7{y^cApETLZ#K*}k!L$`o4~W3x_#Wakw=@ba}$$>sIy5PjK}82c>S7H&k-(tne&>I``a%c -G$Fq{@@g}$$&PFADr?wIk+tK2dCIcKK2gTrB5*;S=a}mNa2~pMM9ab4`yDO_`^3f^rCKU%Am^D+Itrc -Jy!4kUbzWaBclZ(=)O7WT;n3Bzqg^bfBdusIZ0^TPneHK9sTU^iDh(DdDa8`AUtE?IAEfN0;;e079mh -qn~e+2=T~=9TzTa&z{a9%-DQ8Q=HnI;);MHE-NzELbPXU$7VvNK|D;|X1)?5+?DsvuQQ5Yqy425z`I% -QFeN|#wdH}HU@=3u~ay$G^Kp`(MC=(w70V?zT3+VaoeE<%8RGVnyDF8y9P52Hpj^q$*Jc2ek?43M{CL -W)HDjDV*au@}A!qN;~Vx%WDi{iwW)v@j`qvm3~;1>Dr7l0)rKajrG^r7N<)!TS(7J>=s^yY|N5WzifF}131K1$8|`n -j@&3DD~UrkHX2$afSo1MOzi4p-=m<*S5=9B__NWBi=*uukhH5y)u51e(Cdk^>nPm%;64eYi{0Hf#BJb -{vM30?(F6{$7i8a^WcJUUy68v>5PuMgB8Eje5LNPN`2m8?Ot_z_Dp)T9APPFF-q)^9Z9acQd(D7*x<=pnTN6+)7Ds)m+{g)n>jdtr>t-u1Dt3%OF- -mz4>t0+GR#n9S4_+I9*?<)%ja`*tr1UVJkky0G@!YBQHsCuExA?$0#@A!mR;MsQsQ?v>GqM+`$$#pyLj -^Bwac;XUA{OqK#z1|jC@5d4Xyo2b3;oN#d=PWPn>k7DJ`Egw@bRjK|{70Aa;g(;Rn?zzZjZlv$u)Mhl -|5-RS)r1BT8-y%SC1nCyGdl;LJ(cHpT52L%{L$|QaBZ%yuI<-@g_-BAWV%!qwS4JCq01>YuxtzzoDiO -5u7=T=V{LHA+w*Moczc8X%J6BL$4Eg&Bdh(%eMvg>?r6E&CRJRq9%2 -Ep($Az5x5Be+;S@~Fp`#Nu&@u7b1q7$J>Wk8d;*DBr#N89~!gj6zu;qd2-GYF~0wRo?uQ2=4$v8Kbou -xQ*V|-!LYfK%=+PH;maP@Lc=$8^$z4*mmZp1UiWb6N{TJRPTGE+FmM>s!cn~t!q51brJ?T&x)c6eDA5 -MGw2E$aW{*{%I8?sg)d@Vqj5=3NO=XklLtlB1v#4mJWz&oO`GSK+mT+9OS8TQxM2^9xNJSb&3f;%_Oa -4TkFN0qR$u)^HTvDn8cOD0kuLb| -0|2s1WjItEyQjKz-_bjaVJz_v5u~=luL2McJo$AxoN>kxYUjY{< -hnH~+JG(Ez0jl5BdzD{HR(5aP(L_agoN8IZb#e0i2wW$3nJkwBZa5rArvBgDV6Zs_|XT?UN7o@718VI -kfgNTN2rvid<>wkaposM{_T;+G0674;Dn;nV=;nObEI5g;4JJ^pi`9^&?)210$ULukyiN~L4%sK -#@0t9}S3>x%R+K>sGF_M0xrH|CDBen`A=2+f0HqbcTGE{}LNZW_??+#^rqr#Eh3sM7u?&=^EnC&|?w? -!h8J!V&*fs$%-$M+`_sW#TLzuuR{G*?(_Eps#Qm$4qq^`p%?r##ASAUD-4an5?As&z?5Xl2m6v6^Rgy -v@!>eQG3tF!2|RiS*}nH+A;u=L_H!re27}PdmD;U$$puFM{YI$cJ&P7Ot?XZ>gdZq8L(o7fTHRj0GOt -%i2J2O)uBBDI_6VH6k5oVSgPLMb%2Wi3El^u0BFt=CD~PagH;$JCQ8YoqoVEC7$g;TX$?2()#7j$HFL -jTiTxpJmJ(Q6g*Isq9^6GuT@`SHgC^cO$ODPIIz`Q?*nFVf^(Me#aY{#2YMZX91Ks^c>TlDv3Z3OxtB -O@U%>sBL&i)EtDN({$eucm(n=OmHqRv0X@VHu`iv&6fbJ6(p%?4wKViLHnR2cgdlFa_OQx_dcaaN@uY -Hhht?De&rRK!n^Wwo4SR=LS)T5X7}pRNh7ZSAZY9WI$g0v(mIXnguT)B-86c6G6wWcJV1h16IDt3p8) -zjB~BYF@dh=q2!k5ZJ{kT(uy!f4VBX*7dBabd@qz7O}EXY)K@t-JCQ?v4ksO()-E@wG({4hSP#@4jy;yqbHuNNARh-pKv=*= -<0j-U~k$2Wc;@X{1k(LW&nRQc=G*wZ}L^ymfU9NyFDQDeSjx@6kgpTa84V13c58o$6GFXsHp^BFNYbiMZJGZX&>RW>nkXH4{+dW#HsL!V*#C*Fe(gArv^LS!bN -DqPI#TyF(>gNtH50o;eC81Gh4S;Y)$SMs7-yrYfq;FTCrHuO1`qUA1 -G-fv>BCz%UUVj=D<73{$y6U{?vT5w!LXS0z^`Cdq2TBXZ|BP#h_9TvU!sIW96+WcnHDR@X{ILWsv -@kUYN4Ty3@sJ{M`9KWh0EjkM2M?>8WtxZNm3IMnRLa8!v#G*f=P!2cGY(K$GNg0Jr!8si>=;Q4_wQX|%%|tX1fOxNaWz`F=<4oY#*KNQNrbVZ{m}k*EZl -n0?g4i0PM2DiXiEgu(uJtnNlHVSu>0uh+*70B}wW-O*n8C^HCb|BNySjDs{&H#s+v2yC|qp}nO|I)MAuT`Iz{zq!bTlZD!YU~tH3L%@M)+c^}U6_k?P -(;;hI>ljfr*9RQ66nk_7ZPj(bY8R*z6?=>5fmpxg300FAk#suy!eI%y_9ELH9)2Kk{k)KnX9FsP$0%x -|Ev5xl%U@0OK1b+{=ygFiR$ot!8Npfy84$KbnD8)$5-Tw1{PN`F;}Jnf+dcBba`hF -i!jlAA%0oZ1tiPWRNkE&-w5FFd?U2+Z~*w#2T=$^9O-q^XLh;W2DU -RF#VLC2Waa6^7EI^N05c2q}zvgfMJcQ({+z;Z~NCrYejE9Vv;b!mz4FiEao(ZB^v5va?76>7$s%sz?B -zJ_%W@N(Bg9$1jUj!PLm?pF4HYkrYTg@;1-v3N$H$eHEJKV`&qg>%>oDpUVe=Z;2w6=;gWILM5!rg0X -t!`W-dLNzd3geMPWIx40gH-q7uUDEYVW#&H<~@bb6$5kpus%R-$AZ92s;l)C9HYQH4f8O~?{DNHFmoc -`g6#Do?2{LwycFP=<#p-9E4_7f|wF@mu?A7Dt4oMj*qy~M(0;1fEYQ;XdEQ9EL><53 -*ZLzhNLZSozqiB*5BwY@j4wx9FOBZ2c>Sik~Nf>O&EXi(_H8^1%>JQJ(^G3_Mn0GoT}ni(hbCrh*kKF -aSj;!`yIy^zuPxBH}3-%(#`G09qagry@~@{26C;piWt=QwyYiu`qd*(K`U7J`*p}Rj;O>E09^H=JDZI -0O_na*P%_Mq41?OJBDVJ6>;ekv_-wSKs_Kocnen#0^xb*X%5 -9j4F2(9Z;l6ffkW6P;WTwIS)qvF!;b_B^4*7iLwsP!8W<>O08=;I6NqWAGn^hXZ!#k)^Oj8}CU77-f` -F-(;L<1K=g2RA?>Mg}Ds5Q!BaXTWA`-r_G1d}e)F8oxk(M7yeWCN6+_!$AqgU61q(x~n&)qq_cmDyWI -z$Amzs7r4RfaG=b0kzKQ$T4Ui{}ChEG&HH`sZ51($)O5P8BBXR^hJp8C8V7@7;KT|b0^qVc``cC -oda4-jE9x|~QX3B|>O`&yWW1=blG;Ce+C)oI!L$mlnDgZq1gXsyFuXIqNEWD*r$TZD{Y#%{Co^Ohf$u=`(6$91RmeJ=zyn*r+PPN4N= -05~6A2cBq7u7b6i$%8|1E~eRr3&5@R7J(%y(|Z##R2JBNN&Q`NpFc_aHc>fsw!c;J)ek)1{z`!)8(-aT-d?a{{n6Oj9K3b#t9EP2feW*K`;Hw=wIHT<9oRYF1&Vw< -3U?XdCNmZN(mSy<(y66a7g3|KovZLWBF5$`Mh(`AD4x62^~Yne<40MT`3&B~IK@mlc142PU=->9?};~jG!Tl{}%z<@ATHDIRDskk -eSeszAD-7^H4QTd1tp|Ov)1P9Bw-77cZOvSsFfNJdbn5Xjno$vgN0E3};*eWR=AwV({%3r=vX;oXNYP -2d_(~#)HSN6KHE_6mhJ1Lv**MHcDg;(T5pI`(ai@wQ2i-vW$v~2QmDXHe0I8_r-Z0Q%$K4$Rsc;Qc4S|(g2v#1m&v7%Z3>CO@Kjx|i%i$Vsj55w}hIqd -!cLYKv3BQX5MfC$%1F4k)pFz09G&iH%-o5`f}wLAmLzgMErAMUxi~kh71r262>pJUFE -Z;SdMX?MLM|=eRn+2jxSj|ZhjCsCaqkuZ4*g`n0EzD>4hXEIs0M(L4hf{vN5NS87^_erlpXys0`^1LQ -7nE#0NWCsM@gXLI=3p#o=9u0msOQC{U*XnRU~T>3}!xN$&jf+D2FEofSbcmFVTmfBx2tp2lOdwEOOk% -)%pl9H|BIsP1xGaiP&^65h?0)Xc)viiTpvVXWxDBGA6}|h{V_PA-jYkQES?}&-||ZeT%Q5ZH9@}QC}i -D_ys_b-|;gaWw>0vhkyf?Xa{)dE&z0zTlg^ZyZKnJ#Nwz(pQ#)pq~ucE?^mqt#%)08^(Nv%kP0=P03> -93@|=UB=ytC-8-jG_k1+^sU%C$9eAG$N+0fcD0iv(f7?1WaHht;Ec -(I4kU6*T&=XwN@{ZprQDiZ&!g0D`&)|P4jS{zG+IC5KU!6iZ%xx1j?5+RDbp^D%VA&A^sI?za0l?8B$ -u>1mAb<*9{H%24Au7e;Wab}=2?|kJwKoT>A4xfnQP+Pd^TWAh+!4UBCvv>q&Be$Rx+@Cen3fuk9&^YE -cPW{d!Xig%(rz|i`YY`iREyRQO7?7zm@ZE=~neQF+#iyIUB7j<>cJHA^bxe5|x1k$H2v?li=G-Yj7l? -L`2rzbXRuQ}?dCx|jkjam3z5}Q@?3qB0`DFwd52cY=xs6-PWY~O_kv3-OOAXrBci2~d24FrKYdr(Qn( -I8li!fHX=xyQ^u7+FzZvZgKw8so`o@3!`#Wqe&V9hx00X)9M=ghdMZKSpZFEG^L`v^hIbL0CMBFZ!pX ->-?JVx(%H)-Z4`w(-yl1(#I9w=#aY$X>5Di*vd`afntfXzywWD*bs>I65){sF$fc@<(}UPT+L_2IA5* -f*sxjnD|jDQn;yI>wJ8vFwk%)d;30uiNjS4dw^io)7;xz07T6%Z$op+mB0KR3K`x2_?(~gDiLsPMJpA -sE6t9I*HuhMh3hJ+D|5{rUKZL{u%;cidCwj2{3h?BoR9K3k -+fA`Cuz{~aF3YbCNLTE+^!fM(>Ovf<^< -^1k5Qkk~Wu08Shnj`4hqH3MO3dbhz)SrwSKoN>fCSRzAdPHJCO%!C(a7Rt5;*3-10#DANoN1tsf&)J& -;vK;ivu$`Qx>4LII&O>6I_FAyMi+imxb~aqdWkVL_O3$hkAZ<;`T+yPqk~dwFV#+>40)(Wg?GG-tX%} -fz=x(kgqa&kG=cdgw6T0YUh5}fZ!6wKLKRceDt+;j{womvvIQ&W7B@EakUhqyBuiaZYe=z|J13Sio{1 -6GCScvxO98|cJ(c^U$N=PdIg8j5xoqH*MyVGi=CaTCKRpEM*!);YyEZjA{BEil2R8jzM-Ftq(gCHK#L -le#;r^k%AvSY0@p9kcb}UecBLEoBfs)M;J%53ugF&+5RO`ZZNkv?O_R!Xfs@9huW(lSg5vc1OEnNzgz -k`aXW&;pyk#Pl0Y13VdH(XDi6Hf~jNT!Va`*?f%6#z&rj7CwG0;+Inm05m9VbLb> -=~&V!GO(9iY0JUBgDR`9T#% -2w5X;3FQ7oL}fXlNq;W;X`G?F@swlvbZo&_y53A<$}mzFF@VuygK0x#{b9kpIsGH{fAY01HLXYj-=_z -M<;xez0gg=-KhC1RQ}9o1r55IYLTw4ipyy0#jG-oBhnYbGSQOW1K#T%{j! -H4!9J~xa8!N}#%d~O&RJ3D*#Tw`SAM{;rwmntCw&g9aEI!+5ek|b;4LXvc?y+;>y3&+tx?)Dk0ggqbC -2U@>D0_jv@Mp{V|pU&iEWR)}tT?0={d1=-I+Eh7N?}u<3LeXt&OPw=C-;w?ZlT5|`QJZwTMM8 -je|SOeun|D+kb1$yVHs*n>|Jr#gDMx2-eCn;s2(akaw5^|;v=`11s%N-J#q_J5V}^wM{dy~a{K2ULIr -w~(jiWw&{`{GN6oc%^p5gt?HIT`zC*XTFIJT8B78^=Zt)lAPlorQSGSc{F6eV^hVLJ+Swq)R_+1ZkVYm9EW$LNBUJTJtEWZ@j -EdVSA!*p5=4Eg3kfezxS`8idu;Jr76S(DqDRL4k(&`NgY=UX&NfrNy -vRYK*ofaFiQu&%;r9v^^8our$t-i`FB(C@+$WYgo)GrOv2LbIsCbT1WZPW@6V<;l&2Yw=QbhmIq1gl6 -0;!9;{z)K%W}V}CjC4OrDeKA_QBSVkL -V3yWwYZEXn~7__i}ZVqnCN70(Kk6nY%SO&QTv7;h#3u@PaLBmC;3)D$&&4eU(=&_|dveSKSUo5#j4@c -eP_Do!E?LGHcT)dj{MR}22oFikM`X=H??*gYgc6KA%)^EDw9;uOazgJK?hG*2U$t&m`V=`*m*%b^N12 -M91=L%AX^hb=Z%Tds+F3g)#8m~nh+)COa2yUfq9RY9R&?5ZZz``B-t`j$OoJ4o@lESCERKcWcnLAXYb -y@TFJm?$94fiiDT3RPH2b#=9Gw0wMO!IkFC8_p-7qj0Et_~jRwNVWfE0NP<=NtLBO-lLyWBJcB`tj>y4)oDlGt@VyWAvngN -3~tPHAHvB(+1#xPp(QxYrJw7I?!x-C#if+=flM!NDc^h8;S^$KD}_{0T-R3)di&PohxL(xAk|!h -9N-ov_t;1VFVBZ*B;?%!vNIT*t_GBe~k~x!6p8(NQaE0+-O9kBUCpU#4c&+5YUYVkxOD%MCwL^xJ?%J -Zn#8k@F1xjVotjOi>OEJn=dVhZK5Dn+BRVjD{=ccNE3@Tfsh6s4Uv#4^)Sf?0|(=lj49H#XWWz(S8WVe09C}^lLejf5JBSALsKW9-h1DiE15MZ_Hv>)CTASVUZE`cX7`XMgMw;b4nIP -4W>IR9{MywXfHgdI)wim3MK^w_BIW&m2N}DI>a?wqPVwZ^Yf_JmQJV;u4rMcN)93-)8!MoXD8cY`UZn -zAC!GolBnT5Gv1aFpB?lxo6g1*}zXhlr_#JdfWR@5$0-)#_eg@?T(E?JvQNOFg;MVjd$@jB@RiPq*>l -7+I(dn5~K`-ezw25nv-IXSq^MV+=gjN&jJ5Wd^`H#=^d$hnKWP2k)`-#%{Mf=ipQc{7`asCkt^P!{fj -&H4H@#|%K&^hL#MZbM=)pf3wva~l(bgX`MjHMc>b_}Dw-Fe(Hil7+*tRPwVnB04g*Hq$zCxHc2J&c^r -L0&kz^wJi^l+9BpSKSbxHpBfSA*vBWgm=!G@X!zt7vm$oIJU+R_Tw!7FhC|FI50cs?X66^qII=wH=u~ -qb+Tzm|{X>J$O%}8v{m>wLlZVT)e`pv2l9Rntu92WvkxX0zB1@w6aCy`FE*#px-2Zzm_d|t@p45Wwo0 -HCERs{9;Hq`cypSB<;3C{?^aS|7~cDg>SuxfQZd}9*K$D@vZ!uP~FLPaI;POOiwf2~8iqVAt6RkOBRN -6-ZU*4FDlQEpciZd&^V!I}I~oji&r9#=N?g*TSY3Ib7{99?&`sV^(PjS88gB$V+yPV;*35gqC<-Z0n7O6VJ=tfTxWZ3OFf4D?}$dukj(F_Bd%(HEj%`nidT3-@NBG?%$t0`K1EHW#INIFv!Eo37-Kl^0=j&5jD>wq1KH_d}$ -&Nx1a0uj}!ijCcaEMy{+fpydPl&+t(?4=UGRkEI-sSi+S$DvEH1@@C`lfq#beQbJHPR?0`3KaAoGgrm;&Q3HjKW^yYnAbgg7sDtp9F9h-8ta| -Fj`Ub@{ywUQ|dH7=N~61Vw3-KiS|#!5@L~M=Qo$G4Ow|!RKXZtsm+LQ9kPMU8yRz_D=|d4kGBuDklw% -by~ry!p|yjGU%bFI0@B8CSijhj{MLIqF&JTvsh)W_Ck?KKkbm3WR3b~=fjB*kfK3R(9Gmb -gX=nPq6u9>JGq0s(pu6sE)5{L~ZUsTW=IZSy9! -p`(w_sqJ(QY%(Esl%-S0y&rwZzA&W)yGK(|N*Fp~8>Qg`}eg{|>aOG2GKWC)9 -{s)$^)wbvZz}yBmP4m@esQ8^vBnSMA-r52%qY45y8BCs5VDci%%>chU`YtbAAhhQEMPFCz}|(4QA?td -$dVzWo${NjB`MDx%A}>{ch_qa?k1?nEY{#ARH*Vn1KeKcpMxowCr@j^-!}$jjzKQ(=2?-r)?jLa)3u1 -Sp9PK&?dmgc9?)*|nm+mN$-NAz{h5EFu@m+6cP(Q%4q!Sb1%BN2ubQe4Ky2b=CZW^6YB0E3_F=H7APZ -M|#Nne5whVY0fK0Cg-yL*M&R(%IrM;!h=ulFPN<5_v&AGV}2n~`;YU-P$OB?7MJ7nv-we3KSOAXsAF8 -RMCELwXi{pQ7Bp2=l`Gmrm0sETMyiiAcW{c-$y4=3J{@!vqUY;6Q@U`8D9objq-sRI -kQX)NJqi~{bf13;U#?vN|6lJ8tp7Gx8vfHXcEKM|O!a2v+EW8g`9qQ5_4166YCDIl`$3`@!~&gw8&hd -}&pmyIQ#pLFddQHyJk_~W7H_h$Vl({J*A(!1Vn%vU}?%fq79O(C?u-PN`|W!{x$W$TvF`upE&^ur23Ht!nE|Dn-; -Hrr+ErqTNE8vBXaE?YN@*8kMl`)0dr-8MU@u^*c4vUSyH{jghfC=5_QHQA|>>D`) -ZylWXXv9*dD5h_bL{y)respI4x=`2}K}_f%;gvZAYShcZqqCxmv1Iy%z24kHtf#tBQI(DWAQPaO4yYo -z_b2N;s;9&hXMPXZMgTY`BK&ne5Po!4$s6yoK}O6kTSPg_svE9IKG2sFyY29?dBpuHaTo<9)Ry=J1IF -Z;6f3&0}%vI6pXm$8*fwfFm<0IVv!u4`9WBx;u)=dS*xd>g^SpqqKamW}4e ->+=O6<699?!&83hXVg38slNdhDL0K`>E;?HKFVhGU>s0L5~e3uFZQD1ty-pYW9|a?c%Vj1k8D2-z!tA -+TO7u66@RSvT^;U5bX2_c>M1a_cofKC7jIpVi^T$6x -9KhrJ}`-dtIE4(j6S8O1zrq$kt{7@k;##257^xR)w^1xy9VN-7*!KG-A1fxgtU#pMb#4dm**xKe@i1( -}5D(}*Gh>?vK4Zk8&6_VA;LmX6~Yj)g$50% -lIlv%D{pJyam+Uafuv}~Pl)47YuPDnMGF2pQTd!wq!VH79roD?wnmyG#$>rj(txH^{+<{Z -&ySMdraVAL1*Cudyv#p(xbWX35+i1U|iqVO9geMMgdb$VuP)R!B+o`cMw%hxtztYxj>+Lz;AP?eLJr~ -lxdpv{78t&-d#-&`m;#OPr?bk4cLvD&ur}f~Z8MpBsrx77{sGzrJgUeOi`v7hw&1ChCbn?4m`KEzRfzzezZ -1wO)gN8oS|LgO6as3PE$R$SE99mSa4NgXxn#Yuv3jtHDV9AcD@6bkn!45Zft61V-`OfM7#+9wJv805b -9&te_!MV|^yDM$DA#djco#V2sl`hA18@Y??eGMaa+U$a$J2erSC3D~)B~e5vEiv!v -q(mksTyk9x-+`XjY;FHFhd<~X9pJi>WJ;0SNC$CUaPXmza+gTwuM{Jm)%iG^IO80R~MGYFJ@FL0hyIK -ezaD8$d_1B`#z8v-L4fjUh|CFl%un59#4IJgHA%m$cEeLi&Jl&c$8;UP=r*&v<;1)bufPDGHUbH0I8* -S}7j&+?s#p?p((J_4ADKd)TE0;N7BI^v22H-_TSf}4uC%Ze+x32%{p%#@Y>*6ssB&N68#zro -NglBPb?8GKB@YLFVhEPtlP7#_^0{qKV}w|@r5urO5Ahf428p}V9qw1DFTN$GLs2JQ*MeSnMyd@rw!YS -^XXFSA6z;;C!g@`bj;`2z@!g2pa1|WugWToP>U6}gMJ(frh8a&u)7teX;%1lGhqvUc6lJox`I4S}#J4 -wNzDWwYs7H07cNw2OT-PoO%^mKBLDo4#gE5+|~M|>c*ouiSeKiYQ`P3*yI0*`91H?av@_0)u)Rl;FE^$=~LVVL -gJzrS^WpSB@nqUYd<6)pzK-?cQ4oRbk4RNfNVS= -JL~0-F=6o?v(2T9CF&j)&B)-6?cb|paZiOUtrmA?vxH%5Qdc8ZoDX -B1x)Z^TWDt)Hr26Wvx4n6!!;S@NT{HWBI73{RdqJu|-%nyjVAhACA6yO)LVf)EL4>T+3sNili3leml) -A|J#pO~l0N{7LB+a_Sa3;n;0x+D4ue3!NpniFkX`8n{S$WER6Fp{$PFK -;%Fm;P}ggYkg<^;;;+>w${^>Xo2qkz0;z5gYO}k(M3*RhNej}{&k<_f$`A61IV^Q6KlB4_JztUi7LlQ -z@KL%wacQSIdtZ@yHH@PkUxzSK{g-g0fj1MYsCt1_Hw42x;tRn9;nYVU=rW`c!bz-eF=9 -MlUz1uBBznK|np9AZ@FPN~+h@K)R(e~XP+Z7G0DWY9tS;yvO?(yNE>P-IA;|m??@#JOe3|cqVSRiOq|BE%2hfe^ -=~Gh<8uUYsn4^SNivwsSNdF^+@Y`C)pw-94*d)^%I`=|AY>gs(sp0q)Po2iOXF-9Bb@qd2!jp^i@hTi -nd3|_M%{XZ0rwFz7Nr>raG&ae4ufo%qhQ@TQW3-a3gcVMJQ*Og`g9-pU2XZ^lUl?vSPwZk@5!MZ@fBj -=r}`>efmRcgi_Z4fig*r&()>Hk$Sb6OsrQFX`2NeZceUgd|ZrA8m-DZs9dEw@n<5#zrBkl=|3 -_Y(4D)L7GIE)|2&qaKIbA(Y>fk;$`MHuykq8@$uKx%r#=De;S8WZ^gIW_`wp8051{Xj}VZ157F`)dfR -ZmXg|-wQ4<>f={n7=5Z+jVRDV@`0QMQJnPSaK?gu5l&eL*O7r5*d@+%J^G@UM@_aT} -K!GC@dnQ$A)1{v~w{84rW7$Q5ApzHWZULDuzyq#9#9CanwKP=cl$lXGUwH6S1833yD})^RSPb|M}XBf -R=l{?Uj1+9y)7iO>u^5k8a3ljLb?pwy=VSHP%;QR=FZ{C0#R44+A<yQd?5LoE)lpW)X0y>oeb3^ -#YG;0@%Y7QUIy2_FCL08Ll%ZwTTVzjmj+_d;0~U?qgEi+)j*u-`&<_=R$bLOR>q|v -&VeI@SYK(=kT0M!hA8UOTOgrB!xPfBrXsW(SJhqM*8SKt9-;HV6H>W~u!(VK&d~%Q-~PrS#b%W2#{PZOv}WWO7&o5IwwvV!Hl1fj!HVmdIlrH*3AXkeCb9>y -l3=KP7-4d)fcn$x2Lg}#iQSj_^%izzrBjO!5IJS`U;Zy!JHP~7t&y7{^h`M3cir?;P830gnS{zkqvz; -voGWi%>wQ(dt&#iw-#T>IgrJJ7$MZ{ViwOZa8gn#M?G+%7ylPS6?6;<*ClNv5q>Pqo~yh0em0;ZYDW^t9^)QK0+j3CtksfWS-vhNBSmdv3Q=px(__U5x(BOdr&%J8_6D=W(~u_$=(gk$$61?TfXB^2JhWpM5kk0~R4m3vd)UY;)>yq)jryZ6JmtAh2 -LdWNGXFVNUwvT0Qh5ATS$X1hqH*YHUc=>!;8(!@CD*WNLP2&SK -TD-%-`zjIeMz=ny#-bUC#!C-r7qYXnrXOOK1P~n7sNK>EH;#yL*RAy5?{Vd%kW1l{$cIoo>!ixpRifv -ZOB+g|K_cz$j-HvTj@+Q2M5npkn^rE;bXx6;4g9kFgm$QBFV8T2Z?r6$1(p(B=api#Hxg*zmFM-n-8= -4hxh%IWZPVfygJ{MhDSRL;)aKI(|<$IiDdvqN0&3oHW()s4SHmO{Y2~kBq3st^KT>ijJH8wl$pXi$R% -vRFO2k*g*8w0)8?udj9+CnFqon@IoOK>4yary03h4^V6w(2u6p} -kxou}l3@iNa*cO#`1YI6RwF~sD{tr=Uz2&+vNvSwU^AUq1U5XY!jIe(?{Q|z0KnNqQX2kW)?Ng4iMl> -oey)M#}06%e+*R3|Sc%4qD2D_6JWTXutu7XGj+k#@yhuFNFMoCO&nMt7scZ3o^#`KADmSj>HzyHYZst -uL~M4^w7b4|Vlk=M1-XxObN!A?(n*P<@3h*wPCxpYGu66w#4roaJN(!=AJy)ayLk!A2>gs$JA_*nFu@ -P}F;bvlM?j4&{-z?T^ZA8%?+Ck`g6ltE`jC+~CltH|>)2{MKi;D@b()d15Rz8?4!fEtU`o)Eh2+#1V+ -$E}F6bSrPTz^=O~b`OQHGkciKx9w*aQS~E{!cT%!Co@0|Gz#sVZ`T|*j4#gbu#uxfJhV`h`+5(B21q^ -MG@VNprk)fix2c9wYW}xw82q6g2UDJ($&d3AxV8r^N7%&&NTzmvyfP;>@~6Y#YvoT;1QF^wn5l6;F3$w+L}_JDP9-U=O$8dPny0Wb`?Uy?La5l<;zt>p*F -MHu6`8XEK}hireHZbr;c=RN%$Th@QHE}kMIEQ`0F{KDyvO=^{o<+jZ5TiH*$%m51!f!67*j``BqfSjo`04o!S6v$kui1ahhq-q*-K)B2f5?ee>mT*|oGV3r@%6g^la3}T<0VY2 -vGyD$3cN}loFvRnfu6B*9uLQ+Z=5cheA>U7fK_n3HJ^Ho$#alelM+3Ou0`7-J(D6sYD>Jx3w`KErX)m -i|96h|$;6G2Oc#rn;pc~O4Aimn=#RdyG+1{{j>>e=w&R3!Mbc;+p_&JujesJ$08?Et2CI0yigISGlmH -42|m=$Ble*M+`I}GqN{+ALTV)`#WkzuCBKPxf*niFqh?8Exu-Fpl%HC`$4!F|T@82_rozkDs6ACwq9W -an^shgaT;D*~RCJtEJV<Bc -&49QJos~Vt*KB4l0J(-g<1>r*MH4tA@rQC4?wazg~_^l2iCiRS%-Rhv%gVY72FVXX3L-#2dgZ{k}B{# -3kKebxyxZS`Mtq3xOVQPPX&jVMz|1I}NAeu*a`+p`8slkF!FoU@Px=Z$R9K(V^|yYvUSPUKZS2k3sTG -&o(BxzeL_7&04`0o~kjmR#W|vYpgu=ZxIjTtrD$6oF;pS0MGORddA?0;Ca+IPCF8XuCL*lxkRHU=1>j -$MI8KgvlCxx+ZXqS{M!~^Xstk({HKhdB{kvm;v~GYn09mLaX9g_=&5iwfMfOkJb363Rdo^G(2Fx>F!) -qG^^KG7wK{|DIgI8s(UkGPx^Bnw$NT%2xt!ClKp&?+Z1cOgz`mb<aSsfwL0Gqxa2 -SSZgVT2&gE)bB_aPr2FyewYcR+aP=^7c^;A(4;v*lXe768W1#TInboZK$ErtO&SR_X&um{SwNHa0F4>~Fwd7- -05k#pZvyk*1lYd`lz$WO{w8q!K0yCp|L{LhO9KQH000080LuVbTjP1}xz+&y04M|i03rYY0B~t=FJE? -LZe(wAFLG&PXfI!7ZEtF6Uvgz|Y+++%E^v8mQp;}JFc7@+D;Dg9Y!#tr0jUp#0yKpiAUWEBK*=jdi6j -aXmD*_g?_E+4+KvewC~{_YcDSX{4cKE>wR@O8tHAB3w&|^T2KYoGdptnn?{xl(h4qXfpUV?t-flMMV$TuGG7GwD-}M{rA -U~3H+z&a+QfH17*Jxk#?t@+^HOd81x`JjT=@Jway6LK##9ilXhb2sn8|-RTpkP4D;=v|PBl5olQ$uaD -cV#No(#;&>7{M+4(k+hANsUY=@X4?r10FouDFz|L~`kxkC=p;Z_Z5OD{#& -qN3Y=V;GT>!^QL3E^{74C(F#@d&lqPJomCc_Ec1w|nVtZDxa0B!_;$EUzot9_?#av19sjq;PDJ9(5-} -ZL3(8N^1x{RGs1%n_m6u$=x<|TWvoP%R@<@Kxl9Mq!kI1bs&e61QRYUvW-iyS0qlc~+av@rC3@UCJw1 -a6wq9Vw}wf>DpNC{s~qY-EPldz+!37w&X5j+&wS@?;^)EI59nhITxIZel53PngQ@)IK2e^5&U1QY-O0 -0;of09jieTGfm<0{{R}2mk;g0001RX>c!Jc4cm4Z*nhkX=7+FUuAf3Z*p{BW^ZzBVRSBVdBs&tZ`(K! -z3W#@xrfAnC3(}LTR5`O)Q!^uaT3JdVha>77>Tl&r9_vaBD;eAdr69tT_;)Wp+$8NHJq=m}N(F|AVgsFa7P$zAz<7w+vC(U_drMXY!vgjCon8ABLK(K$!YI4Xy^!%Dij_U_LZ1 -5g9&da`?{Ys@@ih7VYwyGx_$80y6aO%YXSPvKVHGKs$H>Y*J@wO>4-%u^>5t?Yj)#zG8+M>8O|ggYYA -|Rt>M!`|wRp!_^lJ;DGyH}sbZP%*3GS5rGX?l-uK;HCsv@Ib3$mn9?W)LvZk{bgSBXA!?d1@G`twxB) -BSX5|M2x@XTxYl+{9w3wDgpfiS&R4#cCxumoCoL_YmTOG1oMZ|2TxCl0Cjo4X{7oJj&BD^8{(_xiu2@ -V@5j{np@oU^G1qDoZG$=GP=IOs!`Esg!Vyd#jZt;ZQekTP=fB12)5>J4y#;S(`C~>5j3>zYr2N&+vI+ -my^#&w9pVOLqX-khQ(`qMkU~utea}Y7d@|K_u!XS0u&pY?8k@=G*?fLJU*6oFPe1;n&CKA^6UX@uP)h>@6a -WAK2ms3fSzAfex7(Wl000RA0015U003}la4%nWWo~3|axZdeV`wj5Wq5RDZgXjGZZ2?neNIabf4BH@x! -vzjV6U*IMj8h{fd#eT#FSis91v;g?fVGeBIN&PdT?jxJpt=oxG?G{TbS}XaXUJ_>cI72)c#`wo?iay- -YKYk*B6PUWAU~AYxS#9X6F2oy9O9KQH000080LuVbTkXNa7KZ@<03QPY02=@R0B~t=FJE?LZe(wAFLG -&PXfI!8X>4U~Z!U0ojZ({s8!-^P>nn=ciw9=?0fq!jSi)Y`kmP2}SRS`*ZDjQz^@I%M<0HxAN1PneAo -S|$Dpkt_KA~;RAVZ+Gg+28g1loHREtY*$%H$jwd+QIE``1;2Pue-XbEK5olfi+cHP(^mS)N*FztbWmu -A_uajpvWBUhoTV>}jEpqN=LTB$z-o3Xd-P2DW%GcXILVNC*@MCtx!vhBeSXv@IqyNFsA?KINW5F%KqP(|&$=uDU4lytw+ -eUT>OAwwc)QA$zGs!p+DA)frYbC8sWx>`h8UYS^8(IV6m?O^#5Wcs2ZbM?X921utbvvj0RqVS7ewGtH -j6VQPSmsV$P$Hoas-&^G3FX{^4!guvmU%F^lBoSQ7kP8;Ju=^(8GL8lQJC+)}o>UHx_T#i~UI%oA3o% -Tf#;eCkdKksR&G+he-0JtRp02}}S0B~t=FJE?LZ -e(wAFLG&PXfI!BZgX&DV{|TXd8Ha%kK4HM-M@lR9#*>Ql;ouk22>z-x4q`jd|rLs -*A4RDJ2NCDQj&cQPDpHvoEgqHXULVOx@uY`m9E9MUF@6c5}2vI73G9AQAbrj=-=IQm`vuA+osL71*l0}60QPQl(9 -H;R2lB|T4X0MXfvwN>wQT%Fm+b9<~5`>^>8BiVvb-ZTeoknyQDpIHpvlq|ug@C{K1Dp?N*3mAH4}m`<0 -T_f_f6+fsvG%*x8jl}+%#{2)TV0%E-MaFDa;a2vOTg6m($y-5y^F)3GmWF-U7NszjfqyRfk#{R=K)}# -^{?_0*AdNcv&C$Mzj*ueZHn2n7_y}Obs-zvh{_R-JFh1WEL3F$WaN -H$L^ccPOJ0(6eQb2NqwcQO@gsgycNJ&)OEXTMsn(bv9YJuvs_y^;sW;=6-jJb#H0Lk?hxO4oV9J3XvM2i -siGV2N+^(rW=&b;!zbb^9av04dM}s-fuMleNoujf7{c^;N4H-n5R^F}brVOYXF9U(t34l_>H#4E0SML9M42{6N>IS6qUPIX>om4Tl5H@o)(wD2{6!P5M<~c8OnaboE^{ryM8rvzA;ehzIPlunl%D -+vK8?SyGhwiag%_l!z30vvw)4JZWqFazA;I7_}CZ^h4AvB&B%RjcRl$kY^=lVXXadK+1HScC~n4uY{f -~(__;Jq{fZJSX=%(9v6xvxV?2s;%Hl$p -XBTkw{-0jKUjP6K4egjJwpFJv+xjy9I485!#wsS44Lk})U^gG}5lGXQu%Q@4l*W-pKE#t}J9;n70a9+ -xj%R|hU?b|Zg`4dq;>P(p$yYtfoqW@-`muLQkzw%gNP?Am~#!25+!66)fR_N?u5>%=Z1RZx -q!etw8#AQEAwj3s`9_S8f}>Pgzbb=CrLBkRwOg5`UL||p1$A!tn>4hudFMItHZ|EE_?^TpRt9Op4E^8tCVtQZ(SD6V -4o2Fumm8(Wv5>%HwZV?pSNqaPv#xQ_L6tqjX5VXSkSSVoU^#FmC`~Q_c -jbi$05}9_K$JEoHo}p~%I%4vHc3tqf5UJ8Z_$MgmNeu> -M1wsaM0`A1rl&c|c62PKD3Z%p(%hL#L_w{=BbXOXpP+K+M!#bDM()7&e8Ynr#l35-r|$6(`<2cJ(x~? -xWR)ZO>e-zmDjATfJPbg8rl`-nNt -*I6`5r{Mus7z+I{CJ-xTPOrl5%5O?;pWB9S&Gle^4~Ulr++C}s&=yx)p&^o5+Ure2fv>}f$~~Z4FvQ1 -;TB%!&@j}NRpfV&H(TivnU5|1L3p_&I8 -A{5;opd!yYn{*%eK+;Z64j_1avKpx%+87Id7i3nYWPUnj5{{%7q8sx;QmeOoz1ew*f2XrFcexjr!U@T -xjW!2(m0&_ZQFBN@R(5cOL>@2bZ*e(13xs_ZIP552(WMPtwxSq2M27xE7n?reqONm*-sgJ -!jm$E7UL>V*x7n5&{ -a3O1{H-CU$S1`3aoG3*3ima{Tt`!~0(ZI%< -BLfuHxpa(>gLz7`;;mMXHaY&5FlpeOOa*lzP4tS+5v$B71iFrn*Gx=X-cLz7riYkO -|r;xZ0rSj03Q9#oBaV3`UBbvjd+k>#Kib9iZZrvIyP}acLbPNjYH5X#Oc^;OtYrZ+p>VCUT1WQenD>| -NB1J$#&(Z=O-*y@xxbIK?{tc@_#aS90|XQR000O8%K%wh7OUnPw*vqGnG65`9{>OVaA|NaUv_0~WN&g -Wa%p2|FJEkLXJ2ww{bSpqZk#f+gq^Z*_vcG@8qriUe){{r -ZO7ynl1^@x$kL3Hn~ce1dMM5C*^5bZn_+589D!;XASK?WF+-uf{Cd2Eo(q+snF@!fU+ghKc+5#_G~38 -@?^=z@g2YH{Qv5E7ghWEx7BXqTH)43U?N}PUjB%*bCaW=#xBjb~=$Is}Lc)=9-Ym&X}iVqDqF*QQi|n -mZ4iI1=!iDGn=dncE^MVyrrU@A{wfXD-4F5RMw-ci$Kw8m7uwkJJ^ehrw{1qj~yNe{4&>Rdv?Q+ -2o&{PY;ThQA5iT`f47%Owg-1o?uh<(9pWm<*#`b#E1|D!Kd7s%^tiaf`NXE;{&hpH6gK(u>VbgC>#o=~(@#42{zJIU{x*OUfhmJqun&(L0;TJ8AG -~o=-WAeCi^HlhEZ|6^N0fo;ao;?}%FZwtqjo0Gs701(S!!|QOM$h0CbT6jQs_rox`|nIhO(E-uDJV?H -39c?eaE8zm)F^s5(G1lR`VutR;itIaO-k-Sf>f#E`{T-wcBg7@{*AkdK3Dc>qu}z5a3O1SE6!~9o(Un -Lm;XGsnpc+`D}Fd??BGxw0r?~c)C2$k82|tPaA|NaUv_0~WN& -gWa%p2|FJEkLZ*VSfdCgK?OT#b}eb2AB^r?mE4;aix!Dovvqm<0v)QHn##JrT--1 -EUXX}#S7jZT>k)Ye%~@IdJ0M&km?!SQEw_$0j$C9DoZo|Rc4glMn@ZLO#!Dvvs@{bY9smFUaA|NaUv_0~WN&gWa%p2|FJEqTY;0d4UKaCwzfO^d=X5WV+T44 -xD!tXE<8u%I9my||}RLadWD7!ygRtM0Ggq-sA@R-03ky!U4EW;()}9nfiy4&-icEWy(999KNwF8ygFF -85M@Ujeg43^PK@R|+p*WNZwHbQHs4Pts_vVkr4hlq>EW%nPNaQel6#NFCX1Fpj{lH44a;aoBgL!~ljL -FgF^{mtpNaIjW%2aiUX=T^XZltVJrUhACGLYmP;}E*UW}rpF*P!8?pfzR%>2-s`ow;hEhdDygxGs^m^ -m_sBuRRU#WtZ)*g@OEd43(yUf;h}%lU2;uN76Sjv9>KxmuA*UJV?4b-OZ7T(n(EClbU{D$yR6%#WaOP62ftrw6QK&Xd -2I#>KgG29znO<4KltWfNy4VHWa4|Z8h?Z64^T@31QY-O00;of09jiXp;s-11ONb+8vpc!J -c4cm4Z*nhkX=7+FUvOb;Wps39b1rasota&0BR3dE@BJ$Ty-C0%p0UUB6nc@8&<0AnYz(<5CAhIS88)$ -v?d|^kN~6)|dkjm%N|5tvq@z)u-K_lOu=`L~`_=pAblRxyW4}9`)F=1tXf*msEmvPRTlK{bdpn)>$Dg -mSH`}Z4-PdlvS$C_e-Qo4xes_KI@Ob~cIlip+o9CC^dZR9$y6t-R{iv#mdfB~y``BKNp6(xR|9pCU{& -e%l-Ew)Se(`HvjFkEMtMBdCMOBZLYnS7Mw%VX(tfp0t4V_n}o8@?FN6*JP7klk!;g_?|43&SQ>He1!c -vAUY|T|R$p=Xc#u_8xyT3|mdji!QK%(d(dHi#`5N@6r{KX-AOVe*29*=c-H=#O -*cPP1<2?hMrs_R-z_Z@({}f4yJ+&-1Dmd|o{lBPAYAre4sj1@`lnquV|^uHz+(C;B%8){gTF1#5cw5> -I~osLs7*hguj`O$lm%bwyAt-NIHa6N_6YSMRA}C7D$Qm6$_vQ`3e}6Fki6%N|>)$ASIcvfJuq+6$_+9`HF?&KT8$l%fl9j7BT^A#Dz$}p -ds=VD2PD5A_o=7SL7fg^A$1ZP`;7^2vNS00x5}nMNCSNucSaqkgucwN-|#&gA(N{DS#5?D=CnYylM!T -lptS8fs`O$NrA5}sJ$XRBJSG#$m*?a{`D!>{0fP?aD;7Wq^A!uEB=Z$8DN(* -+fs`m;u|U^YGG76M66PxwKne2|3+Hau&Rpe@66GrvNQv@w?oMwpR`yT!Z=JoC>{~X2A+Du#YKUtoof+ -a#>OoV}E-><)dG73?N`mlfQnGOo)CcB{V2f5 -EV9=c!Jc4cm4Z*nhkX=7+FUvOz-Yc6nkWl%wC!!QuM>lK4facBa$`cTNF$F}rZN>Q3s -B5G^hm0Za0Bdz5o&;=o}GdnX{Cp;isMjuY_!^o2GGA7T#_QseaIgI}6`}4ruuhze|pD1>JjQ7(1Xt8- -jd36T4nx=WpBS>e0^AZyRhfO}DB(OjzO%jyR+e}d1t80s#;=IM%xbia6y=c0}0rv9@$VNpNs -xXu0n5Wa;K#c@$^4Zme?-Q^|mpR)L8KW&8|2MEQ>?_z|Z7`jx*h5Q~~%<88=t&7A$7hq(N_Jsk#kinY -)JWT)VBg0#Hi>1QY-O00;of09jko7`$MQ1^@uf6aWAk0001RX>c!Jc4cm4Z*nhkX=7+FUvgn|X>TrYd -Cgc`Z`(E$e)q39NFF9{N|tPI#!{?U+o3?w4NZYP1cpjWRLoT(JyD92{`#E@MP2Nq>BE2-Xk(E)_wRg% -R@t~3%0}g_an^&WcdlvSO;}izv|X!q+wbJsRI=JW?2NsQ>&mw&yfgIwA3Ks&RSMVopHACcr|@UvT9>< -Ou~_7l@*Z9nw>s%pm#YQf=j7z%rF{Um`CI2Lv|H6ep-W@62W3H{1SG!8x1c<{QxC3dm**(6U}0~*=vB -0RV{E&ID|qj$=Iut|#tLv5S1);~HrCW#orz?>vsj=_Ii5AzyXsCSt!q`qM^bjfuVQF76v%oBXD>KWFqMz>vKbpaJx16Pt$`c^j{{ML0<0 -jemV+mRhHYb?~xbA_}|p|h37^qUrpXQfm?@6D!~_@=4~3!o=Xz#G?~I^QZTsVV71qe0=ncJCxJtaEP= -F{+P;10`2~x&W=Hd>HE|_U;pU0Y6`mBFpxs;q48&v5^AaHd?j1Ay`-m?Ht}(*fCVepLUD7_MC%S(;32 -#3P9;W#rv!;ugp$Ny1kdfBJA#TQ&w)zn?7$ey0+EkcH4$Bylig0l$bt}5%8?I8$74Y*fCs!#Sfg7kM4Rtu|=wk^r&Bp(^=0VC>RMPxbWB3Al?)tS`YKS`GoG6tGx1lC?QJY!lTrjuFFA#*7wTp|@+5+eDX4Ngz7Hn_Z8jQo9Djv6E|AuxFh1tKRkJ; -mY}(YuOJ!I@_y@o`Yh)j_eiEH`Y9M)85ffdGO`@1ZpJ$p0aZiU!!7s -!TzJ>zRS>Dwt{qdn~5L)7a9O02~k;^{#O_=IKEH21m%}frk-hbCAD6Mrh2YLt7E!(rv^kxvO{Ddh&!4 -*|^RY)GoSdTr<;jZE#3Rl7(z?teLab16483?2-R8w`BqnmI^&rNCSOKhAW&+WVJF@A3-5XPJnvxC|f- -T0ev0QiT1+W)vid6V&jM5WQx%L8Y&KeFce89l?RKn81Er1D)mF^D3@}x6ULz%XcFgQ+Vk7=I#+bLmew -6#>3A3KbR+3n-4jlyi-^NckCzrt-}|uQ!QJ65hGpETru8_DJQg0aC)S^Fx7hYKNz_rWh2c^mN?=F;3c -Q@(5u>SIs*iFi_by1qU5R_RRQSrZgOad&RE6w){StHFs)ySSsTm)nJCIi6m3Dfc36bNx7W^S+W$3v~5 -gB>mip|SmWQf#I=jD1pMJ9@_n9Jp$z!+6q*NzVj!2}daz&ERjy7^oVu=QQlj=)4Ah?v?T%@p`7XU^q$ -(YJ31C%7J9`B-2*YL|nnCiDYy+Bo7D^Qm7%IwWt9yOU^?$xp0d|IY*cajbWC*O6!v@%n6RV|s3`^XzJ -T_SjK&29NQVCwPAB!E($YlJ5y;3?`q)8#3Sg=$B~RjlV}9WsNv|{{PGo;(UZT!VkaW#PMIUfhmU)%tM -?~1YOUMP=v7uofPyviz1Zwx@qTKmBPt2y6no{mAdovV~6&_>x-wTVg&mv;O{BcI5<8gkWqsA3-^$yEN --FJi#0ZOrR^lXX)}$__g@caDqvxj)hIe8rf;Dmb)UrRFTI!gW;QxxhN*>_o_I3w$5G~bl8T-|X@Q9jY -KWvnu+6Nhb(X=^6`W)l>B+JaiGknB;Wb(D3pO2J=EWre+ONYiJ%+b&LmIyF9xe9M^j+GB6Xx&JBr1N9 -#xrg5X6^D4o8#tnLi+K#NH@k5o_btMs&@_QPq_PnYOL$OP)h>@6aWAK2ms3fSzE?2tyEtQ007TT0012 -T003}la4%nWWo~3|axZdeV`wj5b8u;HZe?4a#?PLy!G`G>^AI4!7z2hRq$(d^_wByi{PO`eA7Q&&&oezUlAHyyfJ -&Y@55IHzbkB6RGzx`6dlDmRI;~`~lAKYbfs8i%!JapzWJ=waEh#?OBhi1yqP-3XTKR7meo{yQXBpsX>YB9iQeF><4R^&&1O|CEvZ>)NXaODXQaW5 -oX#ozZ8RrEk+cCRIV5KWl{P7-Npkv(G(uCO4q9JMR?}*FP?Du2&4~l4b#92Fvf4oIG;5cf?w90jR!gc -fWlH^8a&%Zp=j0rVIiRMLth6^bq)4Ge$s_az57e*7DNa<=64RH}(rGm}q>bhz+L&BMO)^d!HA<#N`&l -k!bxIvpOHr6RFIbj>Ide2iTf=g0lZLf^b%6alu8NP2cM}=7?q)+Vhr{k+MoE#R -w|RUmM`i+AO&pYsLsGaIRD`7GF9&qz%`%h?Z+$G?S)b74Flo(ekx%CXM_%C7EMV<}#qAQrcJknZ@W?jsL^~p%JvrDS%kW8D=rm{8@EG@@mA*kVst -ThU%jqbnjMVn=&VhT{<_nvjmiV;@C3Bb`zxoS^5SHtA%vw<~s{SE3g2a8K{YQtYS{?)g~y-)K)~n-ux -~e~~eS45Ov7ZltZ<3UTfXe7aa$0G5LE_xb1677!me%v -M(SbE7si^p~52(nqBxvcZAb=DV8T3k0w}gj>cm&1S>D~ig6J2?THjgtn~FgL`~TFiA1EDPiG|D -O<9Rj-Rm6M_k`$uYfs{MltO7s4@-aMW}$v9mkoVTI^z09rjnSpM+{52U|_^3e0OTb)F;5bPM -J~n2Ulm-)TO=d6?t8*~L(_9Gfsr{=UN{m$-IZ5bfPvv}Xb47=JDB_7v-{^V2#0q6drAUJ&-Z=ED|XPI -^b);8i%!4~pfjwYkC_se3_?%c16Tnc%QXym_^)Eq3n)FTTkS#1-N06M|gcZ>cVgm!_e-#Hsyv(B)%2h -4l>9H?VhLUj(@YVt~xnT|RH`52T!jH3q8yLSGXgb0B6t;%z75?fni0E65mYGRecr^Cr)+M%7^sG+F9T -Dl#tT;C!gfAtgd7tG-)G+bJRlrRD`8;hc~ll$w`?giAt#P|64eA;B#rLMbCmx}}$zkg|c^g!BuybOM& -ZmJT(y)S3_x*r*Z^&I<`+LING(E22IoZYk^I+oC?U=0x=+ZYk>_Q`AG2u#d^>)yJl=k4PfqYK? -eIE7&*yC -*8ruRH+64n&E=&fm3mtoDonuR?9dlL2(>`SnxVPA$l1ACSZQCRb^7GN#HT7tC? -PRCwTat0Y=d4O^df_G1=dwqYp~W~U4uoJ-7BzHVPAoL74{nJb=cR~a@;au5m**18`cJ_8?bJ|x&_;WO -<-HFZP**IZ@|6@`xdO*u)ctG2i9F!_h5Yq>pm>HroRpQ3)pvH--Ue-_Ls2lv$ejp3F`r@Em#j>ZNqv5 ->oKe+b%}4negJz5_Cwg)uphyGO!o+|cFOCYuh|mF`WDu6wvS+Ip{KB)v2BI@4c}kTUB-8g^;JquE3Q? -|!Mat=C!u|G`y2Z7Q}%wAy}cFstk&PPy>Ty##`Yn35GrL(k+t3|!D-_eYAYEC8X?WI5M>0`kfRoBkFmCcwWnBn#^0+OMCKi|%lxlDzFDy!NU0;9EjuK>%bq%QYTs{v`>XG__t7KAFdCAF`$jW^i -mM5({#<#0l`X7nS5g_wXs16?GC4KL9mFZUeZb?+uNR=7gMJ?R1?c0@FY@;bTyoBlB#YFgWz;`u>0C>( -W}(kOzYKjE`X%U7&?ou(B`%@z8ArrEIhE41w5urZ)E7l+*J8hUQwllU+}G8CdOzy+{bo#SuQ;YPH5+_ -M>5I@8pwCz1X`|_++wXwa=EfYWasA_w#=LbM`U>=Q=!V-n_!hz{oQFO~d$!`9%cydCFy)H&Cwy-xoyDrb}qEyF$~cey)vq{b$UhcLEaJa -9z*KZ?nT5UvciVLXBH7{((9qDxMFuDAfM(#^qGgfS0ep|%lfb||oV%XhJSr`lKAprEG?IO9z{#?%vsQ -{cyE`d#SGNC$%lB!O-^yxs{OhbDKtLxVbe-wpP?5B)2L*LzN{gAT7B^6Nho@(!;_0>)P`zKHSFdP>tD0flgH{GU6sd`UscmBVYKDqX1@-n&?pu2!WNs#2XFI@T -aOHSf=vJtjt)oBp)+#UtoDZqd_PX+>0(){06}hloj?S@YS8uGEec7f`K#u=WXi|IE2*{p7G%dRQzyER -af6ERafKsU(()xE{7bt}utgUed-}?8wZsb5h@H -_!Hz6m=Tsc41geDQ8^MsdLNo)Ag0}{EDXs>opslVJ-c8xBEf2`|3o=y?QWNKA5Tj4k>g&;$DupfeTkl -?hC8sgSGO(mEz55eWQ4Ds;pO1TJrbacO{Q~Sba@@y1vW;yM%Kh;k-z=AQCQ$gi9jfvPh_lgoa4CA`-5 -OgjXc*QNudPy>NTq)k)KrOIqqqiH4dEJR1>;^CD*)Fl%MaHA%ahnLOMv_LvI;HZeo?FzabB!ea6YWYaX%+(;zhHX!~J4eP2gTMt5 -MubU^Rq$IjmsZOJfD#UM9H)7p{&Mn4Y3*}{ -zvRPUR72Av9-OztI84+U)%jqp&4RBtj38igVhMG305HSrLux?FO3y|Yhj)9s23+02w^1)L1V7Yvtmk*5c!HVmFlFTU -kIe5V1k1X_+$0f+@xti}8M>L=3ZyWoqls*S@taJu_+FgJ4^DJMzQ~`MEO<%nPlK0n_0ABiX20r=;Uws -MSr#F4|#U}Xa8Nf%+Gw|0>`05!TKyUi%N!~xcxtsiX5BTcGc2oc22=eKFQ}WmQj-T=s;H@9?(+7?J^6 ->Y({z~o+z4L*WAvxpo5QxX4X|8BzKmGJn%|CHUr|PXdjOwktjOwgBjcTpDjcTlXjH*_CMinbxqq3F1Q -OPR6sAv^r#H|92m{qXRTW14p-Z~o?jG6@5Fp~fqZsKoKH1V}5nfTe1O?+%BCf+tx6Hl8O6EB-u6Azm@ -lT*I8fc`&>^aVYB%WcXIrxiIjqFPjdqt~&?Sd5YBo142mQQ^Mcw-8`po&7A2umzxnJ4c0`~9I?A@7keoG*gzTbAskbmwLO4x -B-icC1=7M>N@6=p=jWtZ?DPXq~3%SE?PU>6{Abkm4NE9?FVz`OMu4NR;SU;H^!W47jm9`Ws3i}RYl2m``q~zq2^5?r84Sy^Ab*uW{S)jje)#}~aTD?eH)r+-dy_dGA_trAKcMr-+m -foy$YI3(bzD?*495460v+1_;GTLm=XtP11%?6D&8#LM!8u^r0`4>=20|XQR000O8%K%wh=1~*U -+yDRoUjYCB8UO$QaA|NaUv_0~WN&gWa%p2|FJE(XVPk79aCudbO-lnY5Qgvh6)$^Ape^)TQ1s+QL}>6 -*dWh|;8_XBUtg!#yBx{A{Hkl{SJCiX-4_=ZDU$8#K%y_khH_%I>50iz^c(@gE(1Q{}T%0KdM>dyl+31 -r#tAjv2!!SIfc;g((WIf^!Wqi)ou?=t)-Bm(#^eN?-DD%>*D~;foWFa>yY#*Ul;h*${xU~^Slos1Hw| -*v`b3l7~+V0QdUzp}Hb>jw>$seG*`Yi^!v%gmkCUgif)FkjkYZ9bMiNY0if%PIpPoaK64JCCqNQ>I7x -NI&ZegRNR0|XQR000O8%K%whVA>K~$Subb8T*&5p6c7HU0;e@agD--g45z;89SSIQ#EdI9}Ldr7si&E!h2zyNIqD-qr(b%PFwO -V1^Kly*bPwfJ}GGFje*LA{EqM{eKT?I`nt#LKzzUt@1>b8l>AE^v9>JpFUqxQ@U3ui&~f@5ytHYG>QGy-dAx -`D*KVY0`<)&F%Cu8bzXG-O7@WlH+>c$KQScNJ*qbDNfqm?rdh-L?j4+AOM0Oq<+7D7Tqm!wh{hn#Rc< -4x=uotFF3o7lJED~ahhal%-P|GDcghgC}YbsTtiE?NY`=5(j?xpTh5~8Dit}$1{@luo2=KnTF8<4!q1 -JtNH1gX9Hzl~$rBt1Ad5U-We3lnLu=+Pqd=rtI?vr8T|Vc@-a31}k(i$clC9$R-h7<|c?1u=e!t)ANo -H7(Z&y4kp0aJ$n~QYGpfyVF6vyl^+4g#fH;xbRflb(Xn($sU#}V!-_ValvmTWR%{SC^bf6!y_BZmOAq -~e}Mx55|Ov(yi{a8x5m;^gdX(BpB&R20LH@nN> -ja=vT<&+=`|E8gIt*PC6PTztF&(d~`Lz0>oh$vc^c?>9aCvcjI-5<;-vZ!^Uhn -wg;_?LkEN0_+)Y9wqg4oY8cE<1gV5{ULC42fg;44&%ft39U;7WJ}d7P#qj1oWRnNtZu4FF}*`dO3|mV -qM;+wcWv&UVN?4~8sBbLM;=vhBbnELW)bl4yiTau5%!YM!!{;2^79+6UN=i@FW3>Y4@bErkev~m>z%+`9M$qMHH7jU!);5 -gnUkHc`IM|ODBl4A-nhEH9uhermXt|w)Z0=ZltJ7aVPg&AXk6w_g9I$5FeZi%WAie#~nyh!{ -QtSf;3K}NUR!gz2oxFAEvB@T!UxfmtRLvfc?%;FEtk2y_F2L}v&UzDuYxUkg>fR4nSYlts7FrUA83G@y(mlOK*Dq=E5nRiZTpu9` -QKLJg~=4B`K#LZf4Q1yg!uW21BFkcrCJ2d`PnhP0|XqIm~kr-L6qVD!=7k*WofDMXm$Q-XETwd6%chS -2eot{m;>Iy_6#?{M|GJaQyLdcA;jj9C?=jw0W1@oJgL5f!m*=EV_-2-lL0)vjNie-tv5whws1!!I4bV -rsDnrddlPP_g%f)1a=FPEtkNgJ?!1o1_+*EzwScc9iE?@odO?dygIP6pkHL~FhDwNGTGYL(ocd8iwL| -KTCN<7x{kJ#t#}ZDxI)lV=$}>L*`i|JQye~P@Bp7G;GlW4W -K^2(1ixsvf=ClX~s=sxJ5HiLo2<|fKgIjp&d0^q#)2_Q3`UZYawvAlU{L&Gljj$XDW#MYo>Z`*Yi0Sg -hMS`F;c2Ub12IK!Fi%aT2-WCo7GnaROYazvjtwAp^}*nnwJ>eX;hm*38*J94q+y#GfM-QTgEnv2r%NV -gt3_(WCT@^(WaazarP5!MS?>}k)hfvG=<~}G$cB|7Iz!_&ZA@SJD#gQ5&1LP{yFTlnHftc7;jl>xaO^ -#0l{SJAmACQ@ok#Mj@PPsLj&DFj)sUGLobD-l8+3zE%9OhW8&Y+MTbeSm>?SA9g^-34D;4#yR?y_5x< -GzSdFqj{!g?c-V7~K*TBH|;bM9YF}hRV>2 -SO0V!PYNw2yRpru8QLxsQ>U?t7j4SNl*buFTt_OYN~qV5(9(+9w6i%ZtaT -XYq3I-fxZe9P~mBtdheBe1?5Y9ZXWaajc+oF{I9LWQlyyf-Kfm%U$SIi4+i6f#Z+rYNP%#uX;jqBdg0 -RTD^wQSwy~a$RMvL(l&Awy#^39f1~nH!q2t{y2};Aq?_%Qt+V*4^Ry$~b-(&sPF>A==UJ8E1xX=JzVDAL; -hWfjhcAzdSE#zV3^h-;C#(X`31milgoBt^yBu94jT;&G}I};*Su*6@;`2*NznT_9zR)!tZ426tgL}e( -;NW#B3P`c15t24-%ryaN@b^IdF+P=;ZlK$6z -z5o~rsNBu)IBZvCmPx5Hz}3qAzA~F;3{HY;$v&(S97sC1H+2xq4J{^|GA_4oRwkj7607nJ6j-IQ9R9S --n8TE_Gc}l9(HDiBQv;>9tw{Gx{Fd)jJDNqPoU2^P%k!h=FcgT!GvIOO@P@znF9Q?IU1Oc-B?M2>`DO -ENZNnpNH%aFL@%RGPV+Tb>5!!WBsO?-dgI3#H%(TlvplA?ExT?znt|3*9y5~75s7A+%+t>FM;*8e(c_ -Lv#RlJ=1sxPQwKjUY=27Ycci8Mzy1b1wOU5NiFq`d@y8E}Vc{DcQKUnIduCX1n1)}Y}azRy697*Or1Q --o*7)cMsP!|1)sRc?@vz5wJ51Wo=NY=LmV|2^Yg31X0xcms=J+ep|# -=h2Dbf_T+`K%95&{d%V?(;wayeFWf|7zWBBZT}_r*g>MysxmnOXyx;+kYCEX0Sya1eb#o!9x0%FOHwv -plST>4ECIe|4b!%!d4jY(#@K-htD~ZRs_}|VrX6?p7Cy(%d*BOQCe`Xs|0abA#N%%&W$M0qV)>VI;1z -bhB_0qitaoreWzi&PJaC|+xy1Z$E?|f)RXeY=2=`O>nfigQsKw6MtT=K~qI-91+VOpvmm(Rpy;P~D5c -ufR@fxWitii>3_5UuzL&hvO}M^!$)+wO=RWNW`TJv -Xu!l41F}nqmtwXODITM#$}XcTN&nuDV!Uc}*}HTD$1^cRxkWhw6x7O!C!L7&p2!Gh0fa0zQ|xY1@r(7 -qHpk0kXL|7o*7-&)G_`FOQLt#OVUr#QSrY0LDGSYpSld;WOJGr?td^)1iyd7~-@d!@j;2`I_e!=sW78 -yz0+mrbJfBUOtRIvq!=gk?B^*iIpvu`Qg;5|r8<)m-Ie9?qU0rzB)AN&y>zPrLG#Vlkx3M4m!mK}6#$ -CHi)*>@7v7!ZX2MTLJns_>810)VI3nSYECU`?-n*73h8 -@dNq_|4YKZ^zXnv{4V@QhsS^KkpFn3@Q;Z|iT``TA8h?L@muiQCCVrdzUr7A -pQ4RkqDSts_-qH>n(5CI61HVoFuL%EmRL4Kw*9m|fR6QfThcvkAd!hG9?JGO~(`0$gt6|^kU24&*-m% -_`OICHg)VoO9)ORD@6mG*F1PM{@c}~XT`H=kO0qZL#D$7-|F1%56HZcq5RHf+FqPUM{dNk|u~JEftm>@3=r3sbZP*red*hdB~2)yp= -P=KdTdo=?05}DS7D@LW~;Z5Js!RY8ez`D2{0pEPY(2VKk2dzc3NHSJf7yS!hqVwPOoCSEWS8!PqY7(p -fk-c3|-mPD293S6G*Ux(7axZ?kksM2;_+)#cIoT0!S*&3g-ayT*~6NMPBdG_pLp)a4$e|; -7eg3d~a;z#%9`VIM{EH?zVl#1Lb!lf!esvKC&m1(})babtk^leNuqe{&|;&qUxLZ&=bE;8YpGJdk%o7 -sJ3_tjaKpVaAf++?lQfPLe}T62v$dQkQn^cPbTdOPgA*lzMIH)_ku)NiBVr8t0Jm~s811yDr6|Fg>O{IEEkS5QrlS(W5HG)r?6Giivlg7OJPh}Q$6LpsxFrawANgKV?`Bc -lt=SHXe~woA!|8vU*w*CNjYf^O-)+3CNlmRRo3xP0lM8}`>a1yMSK&Jb5#^JX|QRr+g-}g8ni7v#*xb -eu=7|8C9M1oHHK@-#*@%X=a1Jp`5JPiY^83f7zyx76ZCUZOe1djQgy27DH-lw9j&2hzm;mV@%+bBj2g -EcY9UqVw@?gO5c+cKL`!DFi!cSq41*J<>XgAzkKyUm=Oob@sJQq<-o_Cv3NcHe-6gME1V1i7c -Y9SeuR|c?VG1lGJ=Zp0SGc)z4Di9XHyYii0S*m!FhS{aS&s{^bgp9iCWd3|a9}3jGLVuc$HQRaMhW>R -f1Y$rkD6VY$oz>Z{7->q%w&WT~{v1lXpsmLi>OR{6AMMU3JcEt_6d_~4Hvhy2J&AFJzAX)XvhRY?we& -Guiugm|~8$dDM9>gC`VU*3n{JV7Y!1>3$woO-2?b_-iy-co1nc -72o`6Rl2e`&aOL^R@mz=TeU*3LaQ9Aysm+H=~Zg#QU`0FBD|EA9|ohZk9pqGj}p&QA0V^eKPy?beY;D -Q4i;bUk{2d4qUf6x%>^n{I%sXwU^!JK(4Yi#D?P`rK51W&JN@i0S23qgm?5;hrPKeY`Bqa`zkHCoz+M!6qH0aaMsXn%LkM -OL_g=%joaO>KkK9TT%4Y47Ufq*NH+JPY{KkB28}w{^bTods -k8saU4quJ;Pi~q`-!)81Zm$@aVufO-QJJ7tHZrte#p-`tbABQ-KwayN@ -PUpeCzTI*Fi6?T`h^=fxqlsjm_hMZ#mQ#t#(@?#`IOaiGlP1*j_r}Ipo%bZHSo!eVBxetAuYN;V%&Jk2{IDON9FE%p(POt`26K1 -qN=U-Wv#>F~TR@p;e9r>(jq3!6_<552Fk{uP?sqtGbzqeOkEl?ldnoAuZz6PMZ?CoDDGf|4>T<1QY-O -00;of09jj17JZz&0ssJo2LJ#Z0001RX>c!Jc4cm4Z*nhkX=7+FUw3J4WN&wKE^v9RR9kP;Fcf~zudsr -|5@|$(2_7b;qKs}-$_nbrc(mN)WHlslWINp|#DC|+j?;wEib?pPbV7if@E_Gbx1({DsiQ82R=? -Z;!s_bwsy-3K>bBZX#-+QAbOcR?x?uEv6rm8NHh?S6`(k!$8gy6t8!XF -6eyg)j46zTS3EZ=E*ENNKZFLA29#g0h#Z+s_*Nq0N9;SFo?pV=SMQ90n(N?=yv-C$>g9ofNt4=g2kXx -d$^sfJqTVtO=PVPa|jwiwU5_!Gl(q -I|5w&y^2GOY-&lx9^5MC1p|q9Ka@zA^=;-r0fB`%TiGtwvG1V*+P^xW=Jg{=@7HY3SC_lc%tkaVw5cVR>LM1pkAP&mS*!wz$sFxXpH`TQ|JO$OLt7&)%Jn!i3Klm937q`gD&r?r%qi$U}xGtyl?^nr%y -I$q*a$M~mDo3ET0N)A#03`qb0B~t=FJE?LZe(wAFLG&PXfI!PX>Me1cXMBIWo~3;a%FNZaCxOxTa -ThJ6n^JdocM$ZOPtv^6E|@kc89DZ6U6nwH4RG9W`MG79mjvaZ2>Qfw^1Gt$~oWVoE|V{GzGEANwxrz= -9CF=1!mza!4?eh-xbd0@+w&6YGEF7VHmN>$w_Xbcbr9(*^JDWHd>SH>m!=ejMD^HV12v@=bJH+acdr6 -mReBW^eAN!$wGno2}F(;Db7R%)uWkUH0jYuDk*P*O|Et7K*KO197Cchl^jKbm_)?PoUP&*v-Z~Aq)eJ -xAXn0MU1<}(e1ieaaN7WRG#ZVU@C|ba^PJI~5vjcLqU31T-B(u@DM>iYLXIP#8Hgo@(g_Dr+!tbOLj! -l&f)^C(C?0`X)7ji<2O3wNIS2}}=I>ZmNM4Rr6LvD}Ev_;NX5C~PMoOqckqKJe5mR=Fr%}-|Lxw$A9_A|@qRlf@ak;pKo)#huu)`khD|-s -lY=25nb4=Hb)ctt6i$~nrSBNcSBz7J1U2h(Z8EyN5-p=ymjn9>FpGW%3?qK@i`QxttbS0S_rgg}oJNUkLd%r -$GI1A@*cQTfPhk9Or^m3xANKe2j&CXBYt;2^3aI@r4V|K&Ncb^!*gAa -VTe&_V71_1(dVVU`U^)yw++c1SG+QU_Ycxwf|C=bK|B|BxUWql9Gmh`Fu73iRPBP%c6yJ_GxB73>ML% -WZZ7GfI>QpRNUBz{MVTUSlSLMdBM;*UrP-C^+v%xEqMfW -_G43G2gp-Poj=RHBuKx9sU#*U!K^YzIM*hyg<@+j8Q-^_X>eo1MKdBXiRg8N8)-9Y!xOxP7}9FnGc+y -di~3EE7B)5e)NjjRZqD#x3KF5Xd1m+Ht8d=9YClcahYIIOs5;lll`;+FiEILi`G0QGk<{d=WC`6T*@@ -Rz)IcEgrjqD^GxH{F|`xNIGFN9)pP?t6UC8u3Iyw|CzB8Qv({EqXC47Qtluel#K-J(r?nIGxKPB^ATt -|?*v34-Rz3d!fS*Ng-SsJXk*+3cDmJ`nN`V}GZ^Nk$Xe$blL^W_hv(&5Ksb>-xYS8cH$ -wCZq)Os9vGGM*q$*8;CAyuP-Q4-9o9or1efK$(2!^Vz|@)n^Q%!RJR)nQf?H^e1QSXH+U@{F=oR!tOh -N-+oP_@aE1X4vAsEbF1#U;C015VBAX5@5%poAeDRSVOGZ=z&52S}1OQ;=9{zD>3WV!)OZs$YWB*|;AT -URcChy$L2I^9NIuMOJpeJ&ODho@V?)`niVodJpFb7Y}|^$X%1jgTLJqDqLCxCu@{tQnJ?o -dysn?{U5>kv2b=X(RBur*Qh)mBFlkjmN3)8?oHn~sySG^3;5&3!;xp71ReFv!lC|1RzvLqI`%UC`f__ -zvXXi;Sfi6HNpsGSLj@UjZ0`#3`fkz*ep9+kmwoDDlYO|Xt#z$Ekc#33svVV*!VSf>yQ&8tbWE}=!lOyIO4*2W -7rth;y7thu*n*jOd((Iiqtf=F!AbkSImJOHR#HO$GQY_O)7Yt8*5owC?_FV1Zr0?@$pSupSjIROww07 -YoFfj4gNX96Tag+7;-;5^jmY-?e)Z~$f%VzTj1kc)gQp?nN(P~^2tz(t=<$H2Cx}=wfL4F;)eCR8Ei9 -%Pa$2bsHC`3W-JmaCCFwrOYZtKK3&A83%E&_#U}J&$iad+9-150-_R~=Hwxdui=Da#(D&G0;jnp5Lzk -VdJFvPmQ75%Tnz>1PO{YIyj+3Pz^XeL~bfmyN1*C~0%~QNi_xmi2A`CBz!X7n!&W9=Q!G#Mhhx;P;$B -DlkgTH}pt?tJc{m-DjgwkBU9&Y*0+^X(b)qFRD`-W?UTj!sB?s>-dVz@L7NAm7Hm1>iH2xm>PKb{=i- -+t-h%dcFz{OZnYufOr;mA9_GeeL@0JMX^t{*4b9xrQFqqs?s!;ReWq$^EdyfK2;jy62bZ<^@cxjbIm2 -2-oBC@c8P)jf^f2e|hrLH%~tLwmka2oPJ$Se=LuFEvJ8%M}L&lhvoG5CtrQ~_`zT0^viPkGmcNcE2qa -~a9kcfEDs-qi`-6Fd(#p?X6uC)&prC{c!Jc4cm4Z*nhkX=7+FVPa!0aCwbV%TB{E5WMFrR^gIJB_9x}2&lx71LAHu?u -M8qUd!Gh`uD7zhVZI{4^iyd*_|1?4>W=94A{w2V7R;A6-8l29V0whGd!W&W6<4*l|Xe{fUL4C@0_qA2 -W=QS9T8?qPa)GFeDY#bWYQxs0uP$yg8{)ta0Cu<4@`jXEDV9ir=+U35~ -y`qSSN1qJDO#2}Ld>|%6V*?K&5a0FWM{L_!SX$D~M(ZMLXRw-1YVNxEk-^#jvJAOMBQ`#W&XrXcOG@+ -4Z8kI^546t*d+b4}cMpck5xbzS>SwmL>uD7#s{bYUW -Voi5n=DN=$XOq*PghM0KG`P5f{NTEsX8k -XB%9^kY5&#^IV^RBA%b`JT=EhX&N1zo89t1ws22|f_5DSJ7Y!Cn*ybV)$}-|1!fN96}lO9KQH000080 -LuVbTT1)I;(!PM08}Ud02u%P0B~t=FJE?LZe(wAFLG&PXfI)GX=iROaCz+-Ymd`L@Oys6*8OlcG6D5N -KSVVux>8D89s){LQRLf;H^~;iWN(g35&yk2`(E401@2HEY9vVPotd52&TDVgH5K84b)sp?p75$|Izg5 -j%Inz**Tk;n_1Y}F{O#F;{Pc$hr$0S<@-QWj0Fj=TEG17{!JC?vDftBH=gX$ziBt1J1N^ -aVcYtA*e^if;`K5)Z811(mGZ2r}W5sHr0eZH^Cj;<8J(@+aD=qE|e0aHQ*=)vFp)`vbfsbRlaagD7F* -QGwuzgCNv${|kK$_Wp=@#g}XJ>MkvopebQa8d_1$D?STecLeAQyW4?-W&Q7ha|O(CiV;)2dns5o$wlr$h9fS3LHtFLTlthL50M -~(|d{XMfV{&33f<)5$q|~{v`dC)`eTJ;_VC2HzJu|{`Dzid_02ui)bkXp&iej1a&~Qn|luSFOe}FsIc5r|ZO!_2*JU@SlZC8_(%7*d#f#{&6l)QJMqS -kbD_*Q1vN**t}8+w?t-m>gJHd6c-gMiX)IE)HK|GUh!Tt+GJdQM||^L$1p -KEB-Kx&s?ku%#l*Li`Bm~7I_(U;dlB -nV#k_AzXglOmY@|+-P^BrcRgXbiGm-Zj+Ok-pjyJLNL%=ko=UrN5eq)N9TDo3QkTZ2|$l$;jQs$ -ZNhN4=yRR$O4EpJG(3fncltz+Ll=H -RVsXI3ZAm=&Nc)YsIMSkAbvxHH*D(&ERC2^V`Lw+p%u&vva{3yhI6nkK%pu;rBY46xDOcfdRldt=a(Fp(SCAcYOBovf0*95<(?-WSo}>SfB21_r~IOwEhNa;k}uP -3gleZj*rolagSD$6KR#A~C0dCH?J^WPCQqZ0q^fm~)WFvK{Z^JDCpGS4!=Of20N`9m&^#r97x4&1Xu5 -R}u}>ItCvqx`$Q(g4~3*^#Jj(wy)_Vd$chnxmS83)3ZawC(uX7Np-^G9`33V_29;)Se6$~8aIF$MMy7 -x2Y{oFPPb1t*3o&_r%nr%dkNBbAgGIYY3@H8bMZOotDIQ6r}pUC%SkgiI8`&S>03$zRPI!c?t@-lUjh -9}yV~*oasxq7Z2>G;Dzi -4L)0wA(X>nH*rqIo02uu&N3w0h^ObR%;FhMZh5WPv(4wJY6`C -N@b~z53U%#p<4FHYIRhpP6Ld)0q`;>_NZR=^3-(Fw$N9!HHR9O0c##-fuW9NE+~cmiwTvZFg2|}2wAe -_c?DHfo=38#YdgBG=mMezhy>|N(vnJQc{|T9V3($wqb~OTq0>W1DD45f4V;$}&y4%St>+Mxe{io4Ah) -z;r9pd!56VV9(QHUhhX})PW_XRYffdnA(Q}rSf`su;ruwa*9j0Vd@uc3a9V;R;H -_i-Ee~g^W{sK@-0|XQR000O8%K%wh9WLn?vj+eG2^#O}mg8LPt~<@N)5QSWnPL(YXdOdPEZQSPmIR7U>fN@_vd^+l)^m<{wG~g%A4Rb -!5yaxr;o)~KBtP2uwyuO^g7@=~uhD7Q9jK<-X3#jehr9a*^H_e8AqMTb| -wSYJ7fM%e*SnqSp?IR#j06J>)Cws27dg7hJ!%!H#-?5VZUjo_jsPleBDdHV4pPa9igE4|)um==b}*;K -l6jr}%y`8opSDKkM$@GKfBT5z^zG6$gvp%i)iAcf(})^WB2Re6nD}&*1-;!(W$S*aMS;);N#8fA#L=+ -nZN|UfApPRz=!0EV%|Kx77nL1M7Dgm1Mk -$S|_f>rfH2}GqbV)XPvQYNGw^y$^uLHcx2;;oEMp6#lXYyC5D(uj4{SEAV32Q)BqzAY6nzIt0u<)L1a -eH1oW9$l6Vc3x=rQkJ$kQ-E?VbhmP71pOhb#ck*Saa#vwNIc4*YpX&H<{qXKggUZb!P)vj&=tC6o;JM -0VNfSm!`9wX!yo`JG31IWVCHww;sc3K&mL)`>@|J6bC(4kLRe58VT%*!3`FdM;`<|&u&gZ{Q39$PVk6hg^XKGP2@np(Im~-~xW -oD&yFlN}C!ZzOs@GFf#mfp`WRD9^&ytNe^L<&(#YT`}o2YnprQE)c0b1rNq7~dzog3MOkzE+srIB448KQ;_IfX+&CmSI{?H!<=#ID!?pNw6x!9F>2#R -mRl;);z2lXF*WoS0m=V&li;(iIz5Cev2zdvlJk*8;X>EkMA$T;*a_7#X2`^y^z!5DliPI~=7;>0sp4= -KM;7o7$XDjm+jeHesIHq{k-6Q=9eJgm`Mx9-9D9ZQf%O-lDT061WlDv=>7C -{h<#XK>-}p>G*m?*wq3I9vc< -@`{2;F;#-4L|7yTmjX)q5hJvNk2zb5De&yuH48K2dJer|ya1ate?xn`zv7*FK_?>8sz~G2=UuFkf?B6 -=WHdPUHpdOCr22e-@PLuUc(#l{LlQ>|`e3Hh3bIr^zIz^Q-j-xL(XP&5>%$fD5YXJuJC79mr1Kv$yO* -jR6o3s089bu55-OOAwPWz>O<4&X`3U?>sSVgYU!}W-vnTjVONmzRIkMfHm$HM$<6byti_Sp6A`Hgo=L -48S1_|nh7(RL;YR*R8F~(v{^vQsC0#&#tazg4+-ema(G(%=A%+1Z$ZOh|Qf?Xrg{?J8@I>!?tIE~E?> -MjKUra_FJFP4vi$u|5ry1IGAlz0bRmpWCCM(ABLPC+31$92+y;Zp+wO0405H^l-cUVaD|#N+Jh6q^72 -_Zuiqyy7~0*B3u603i{VO&lxp8ZKtd%P;am0*5QC?@JRA4vJLx3B?}(Fx2t$%=k*X&B87{M0q?>Iz11bi|f8>HDP` -@nXmD{Ttw^z%(+r{5U)K!VD#IUTK3$nhgHZp^23v!SudbyXpdHvSC$O@ij!8)5Er=mO6J?hvZ1bUegp -^cR00KXxa*t%}*M^6HRAE0)47%vp6@4@9BzlVM*1B_fMHjj&Vwj2-$g-QQ9Q#O_T8Ro0D+iK`{b1e=8 -#|zc`R_wUvbRyaxJ?0wZhD#u)J067U)*1A7ay`6KG_V#(EWNQBm2<^a(bg7xbsIPZGir@G7hCBL6N-s@&4acQfxr)wv?}cZni>dc)rzC@SLUaisMHVRfc9<2qt7p^Vn?m7hDNOx+}X)+Ybk)Qg}glN -#qNnYNi|3@9OPNlRfG;ZhJxRC~qh>Bv-ZEgI~DqhVT|bk(JrC1Xpg`Ug}*Z*p=Bt(Gk2{6GY7M&q$?D -az1sX5Lhh*2AUNCYxpQ8^!^!D5I-7m{^m1`_{`~Q -?#q4}`F}r+u1yh)h{`vdIQ3?mkpF_)E?Q$K>E}smcfus1tEx)_vH@E!imS5cRvs->btgl|3pF1$bup& -fJRIpqrILl@1yClgXz~5*za;{X61A3U2uw5Ph|VbG)I?QW4Ib)rlB!2I8fT -eb*n|%H_`{uCt0B3*aq~}>K9(PsGXVsi0yX@50a3?1SJws9^+SPF?tj(?50lX-ov(@OKy$^db -fzv*lyZSHAPmr-oK2h#IRBn?ngHl@%U*)@M@~nuX>A{MTW$pc -vkya@=#iFUx_CLrE+c%5gYD%?~&|8B|^Q#%RLy1Sd!y8@|LV5n=APs@s?6hTfedkX_@c10JplgB77)d -YcLPe<{7*czw_y!m8_0#$C-#z4{y0u6s*!KxNCc5B)GbEX#>kmj4A%O9KQH000080LuVbTkfoCU|b3S -0KO>z02crN0B~t=FJE?LZe(wAFLG&PXfI-KcrI{xU)YCr*oam-8Pi=Wj3Pe^}1n2bL%$zKj3ya>Bh)O8Xfx{%bk^l&EeTo|U<$qK3KP(}?3R -lywr2p3>&T{l#+rV?ZorI8}(Z;z^?H#2?tg>_o=xF550_oqBKJh|MBwcXDPg84AJd<1q9)EF|}NKJ+| -{XW-m1lP4k}iL(TJr?@o_hJk)HwT#~qZ|Wq`56p6w{9U4| -G~Pf2LA>tAPkVZaNsu%qK5lpc0B=?Yj+iOQRN81UyhblFsw=hfu}b$L-;UQ(Br)#YnSzLP}0kVr`Eb} -sRmJRow_662STA3MAB;wi3_e7B>KPI)ZF(DQX#MfPS{b{=2&v --3Ms8}KNNE80)OSbR4OBmh{!HH^a~S;atz3x>81{`q!GL!Zr~Tm}$IenCpfEl8?(cPYX!l- -Z@siFu-nwrS@>dS5P69(yPdYxIc*Aq|OyL4|j8sq<%9V)S{>`oJ2k!mB-p>C01NY$Gjf2C%;l0Db{@#Hz;SwD -Fh>CB!P7xJPmmGph!0l8BRe7$_wso*Q7?i`^2y{^Cg40FdV#enqoCd{gsRr?&yx>S8J`|p>FFD4kOOL -~NL`x6~394?G4B$Q}W5Ra+otZfGd>j{)E>AqkMj@yGGuIl~Ll80z6Uh=W6~3q9F&Bf=3J|9-t`qLD7? -Q36yi*C%fkL0gBgypq7=)a|AWM;q1wbU93}eh`Fi^@yn9UnWY9Jrg$5m?no!za$o_njmwbS3Ji3yd0X -jIIRgXK!vcqT^j1kzKsvOf}md|1tRp!+75WDxYlk@3)5knr -P6!5OAY5qDj~l(8FHP|md9eiBNHcVO2$6mCuvj2R5o9o+5u2kPORcuC5G>{vhAM7tQU_u>%(pwXvbb$ -pP?E=q1`x^@V^j?wSE-8;)kc7PFcG>bsL>jQNJ82_wysW{}1n;sV#!> -?fMxo7;+J1I*opW<^q3vBaK$dik2?&v-wM!`QZzk<%H}*kjH%FvvDKYy*UC;5uppMQ>Q`CzH9ssgzIZ -((e<>;+n^=oX5%A?X^%}K9Q?E)0Jzy>i+*$=C#e~ldANoBmcW%Z|;ig>%G27zA#c+u<@VDE;xAHqNRS -9mZv@ZtLndX?fUg=-wssLvmUd@kBf)v*B-Ca(#V#IRwbUcAU##`0U4p$tb=o{0nVB&7qT_Oad-Cb-Pr -BZGW$}yyufVr%=i*o!i8=1oE?wJ(#V?kN_m_*`Ef*g@O&e|x_6 -E1NcB{Y{TQ0t2c~^DTufhFWs&VnfrFhoqh^qg}Qh*sk^zUmA8oYcF*nceN?=0v4UM~Lr)q>5it_+FQV -3`Vo743v*6=9V(4+gjI?AFXR{G)YYW%+M6wO*dJouO#-&y#?1;X~mK6B6~WxvzdQKW -{@&r%?m-Qf)uvyvb!~RQb$9>X-cG+RwJrXbrnbeO{`cv5@#i$PEj~|yH;n3X{zDp)zbxlJ(j>kI65oS -=8tD_{TYP^x|NRyC*6sWuf%)fh{!<;c+uylaTM-rZ&#TzqHN}3EdF{7X;9Iw|!v19$ds8J`*Y6*b>%O -;~fA%yhp~~5HVLqpn{Ojf7LoM7sfVBDn{L@Gu5g$y4Pp-hXZs(s6-)GCk&kfv70h5CNv;rRdVaWYuR_ -T6q1-^AVEAYQ5z~A1w`|8gA{k;p4Z)s`%xN*_G7F#O1ZY`Q^CIZ}7-1Z^_L`U0nfUdojIuo#eiebA0_ -hZ;SC-n9v)DQ5}POVk1JHX^O2}Yh})uK{it6Tb{G{s=nFZHBaGLj}uPb^islAh7u%IsHfMAVljXdmpT -MV!7r@OhGG1}ClNBmCUXI%Q-@&04z76$XRhiwm7qvjI|DG%^>z>v7${n{wBsBP`dQ;wvH7>f#IFDE1~ -(uM4|e*#DpMm?SoR^h4(Z#YaE%xS~Vk1>Dcr@8CnL;C8YGJU;M}{G$@pSpR{M#vt*KB>w^mAM{54Nak -&DuheETdzm;p=EE=^(IiRFW^U|%vyX#%L{BuU$d}qVxFT#;DEbKKjcT;3lqUXkY->nb2-LMU>8IIKr8 -F=qB|2SOzv<+3W*eN9uKCogQp@TU{3@l=JafTgjYKf%rQiSH=@dja%|SexN+2xx -$fi)!j-^yY_f#;KzE3#`&s;oz<#DPa1wRr~tCPNqbyFIiS1Gy%Di_x_^*I35Rp!{%Ey{lEIeZE93TcD -;H9Bl+9X4w_;5#d8oBKX;()DUw>dp6rOHax;f1OUAEwf4?l{;bn+hYrlM&HbM-ztWG>P=v{Jxx<~3`& -ToDAItC)Vi2R`P0yNCT~r=Q|>0G6Pt9y4e0m}P)h>@6aWAK2ms3fSzF!mxc@2#000UU000>P003}la4 -%nWWo~3|axZdeV`wj9Wo&G7E^v9xSX*x!MHGJbuQ+mFc8QbPrlrA(BSM;#NKufQpgvgEcC#LPm#%k~o -!Q2&cqo*&Rze^um5K^+dFTs@NDa3trxNJac8wjA4WnSL(Ep%)C&hnu$V+lM|&lvgZ=`M!1J^`x5_*xcIPwXW= -3Ja}WjQaJ;c?O?zJ@IJ(W=kOi_9`(GyMvv09C4>?0#SQ}-8p~u2&>92BwIkb+EJ6zhu7}121Uy(~BiD -|F^sciS9L5qHew{@EhFs9^!yymg3KOUop_Dc@&02P{MJM%ob9*~Kt -k++@P}k?@R-4+xrHHX>a27Vlk>~rKlgT#E*R^c({Pw1neVvb>7uiP`2m*qCFrN)sAVm!Zj1QTNJO_kp -hfL&VR#P*p6pAtv+31Q@y2RK92quZJ96?Q$9T80eefh{x%2f&7eFP3u;|2@RhM`<)9O;vjnx6D|1eTC -?gt%KQ=pp!U3Rxw+%o-^T?a*-Cs0IjMg-VgV4sLL;HFwgz0fEIRT8gKf-hM8y-97AO`;D=VQ+2Yq&lH11Qnl~7#qgos1*!(^pWc` -d6+<6QmKSIkO}AvyGSPuuocU`RHSExLyZ%j=ZHEp_E;wR#EJ#NZ%^hlhw35sOL!ulP>vMTmjq(DOv-+ -hZUrD;1@w$__Kgp3oBK86_Y8;1y>&sE9{*^o!$71)tfu@R%1L=LnQ6jqk6ra@l? -DqoGd_LTVEhz6pSu}wjNL04l?ne8|Nj#XGFV{^&NrGfu$L>q{4c00)aTG1(ZhCLrwi -1L?CkWSRc)3O2s%6OKyyjTTVlf0M%JyT^yZVbH${?xbE1Ls8k#0>$B9D^2|FC?X;3g(q>M&<$k7~TCQ -kbNo*Gf>3K`uj=hBi{qOkIMUWK6B`MQVne -!`U6}8gXzw)LW%fdT|kvrd#Vd07Syw~B;0w(tH!tlRs8o77)lOFCpB}r&5&rPr9`?}h+1O}Q+}o`r4^ -OpN4EYVQkVX_ImLqTAN` -=?L8V$Li^waK^$WbBTRK7A?2nimIuu{OX=}rj{a77IDX^HNlU;-ss!WXCBv|eYq(KSh`IX5p7p%zMMB -%kBv6e9~X*XJo}77&0_dPd3{NhOPi%R^NmLhtN(p=wg7Y*ew-AENChm?3Z?|Guj9Li8VUwQU}|p*mk?Y(P)Q~$NT2(?f)@-ynNLjX29elHjU<;LE4pD80+{M -0=Avo2p%5TB-I!)>n)ItA4khf`B712nbY-6j^?DW1oW-RG6AtfH2BvU>bj(@{_6nqR!4o}YiI(J@HzQ -HK?RL6JPqMME2R$obN0#ENt(8{vc~L}i%eRFriBVOgqB@l*N3C?U*7f1c`}b>(ebO -bjw*DvDI-$c~B~)1yKo)e7t(eWiqrt7=(8+Y2nK+M-q;M)qK6w;v?_G_?O*avjT3WhAMfP&IrF09g4T*D=6My0 -&%{3~(e>mE**cdO=9DAf^-?!c`44Nyx11QY-O00;of09jjBE8c6z5&!@rM*sjB0001RX>c!Jc4cm4Z* -nhkX=7+FV{dG4axQRrLm8Dy;NaHkvC!q$i^-WNu!6sJY -GTCqS*vq82ghw7&?UO~3rm>kBXZv6gE;f?Ki^DET)_U;zUw^v{p8b6J?4Mu0deITD!mV6h?{~56h}XL -!N|P||2m-WtdN{ns}dplFehL6J -p3ztHraA@G`E3NMqpOb=hld>^Bt!U+nelEl}VwZ*MZFnw{6x9DgW-S4qQ`D~&^csN<_Q#^c`B=1CEu{c4M-z8BBh -p6eE1?VH7Z%;D@(ZVdH+y(h9uQI|B&D}Cxu4NV&!jn3DfbcWh)T|iU2+*3y$sKqm_a#*)20z(XV$PdP8pq9PE6bkZElPfaAZ_ -f4G|dIg5p}ENyAz+pxJ|z8SS@$J&waOixnh2?pXjNFm^P>J7?<144L|NL>ob~wcBTq7ZEmgK`88aVtH -clAxa^GB3LUW&BA0&lV_t+%){^uk!A)GRv=2ga_Nb&J)|*doM)n)Crb~IDYdaC+60SDm=&4avlC86w) -Dm$aOg}c21|{=aovfrMj>Ua_vGZP1|^2HG8bufVhVZ2S*OePp*X%42y(mPVENtWoT6XwYd=@pQ}45``VZ -RCSjn%eB$Zo=D$5H$l(DuutNTx7AU+F4~XBhc99!38PNoRHlgO^%Gxpf(HdbDA)1&LgB?llfv(C91-S2#UTKvW}ZwQEjBO^E8xIdo`g5~jog!w#?8x<3J0 -#JwjWtl^>2}O(nKDXQq39F{Fbq`Xb`3tFsXVE&p)ez8GGhB6+ -1-Vafm>X9+)&(Mn;B@Do#j1CxA00f}4Cz0gMT_RRGjxVT|y;lN`#R&PS?}OLS$jrKxx<_cjMZ=7C|R^ -a_(@88Qd`N+2p(?7KloD{x@)T?v_59?1|Bb;0onVzQncFnPpe4deU_H)j)eeg=*m=VF-4sPT0k)ZEhE -V31~)iXsM?N*IddJc?N^b40}5(C_sKR+dj;158q-Fc{huRTUx+X9O8qI8_L&$1IHO<*FdS8WYG)Ff|Z -h%?M1>R0xQNjPgOyHMXZi`HdYk`Sj6hcR(uU44%Im8EbhA4y03u%i&!1oy0 -3vCQxL0TTlX~(V3BFZw(e^nz#`L*ZQa*E@-QRE9NW4tL7&|o>2G0{%gZcFGyn4ALQ+TH>XmM!>XRiwt5=k`DRdy7cE^9lU^Q@=C( -+{9=TW|k!$VF=2p7fP7zk)z!F+-3vO@Wpp2*iHkPBmT$MOarDhT!a6+NyR>j>6~H{_e$~@4tb;Bj*IBwxmggh`3|s?IB3JsQ3Vk)lgvciU*he4Wn2{)vOf;-GX -Bk*7vUToyAKwJ~4v!F)74z*<4s#pJQ%vZ_)UwZ|R&G~3=Px-9P;7Stml0a{D73Q{icidkHeQ)3{v5_R -Biw|GTOM{7L<*L?S7g_U$c|dYMMZau6wZ{8dy}=w96FzLZ40#Q-L7uaa}u{nz4K!?_!f~PzGt?COBb^ -OzGz{^JY)d(niC~!&aK`^&|@+_zO>n-$Cu{FeBa8ocPrjnlixpx7abA5lohf^H-sZ%!0)C!o)vXd{-v -B$-X%H4M=t55j>jaFW?oKfq}*L{{Y4Drm@XS3b8)>za< -~Q8vkXr3P->|+9TfXiFGa<-xtKn=i7@u%ks3Q0+|+c;m2KSV>iI?%cTy&z_XJKGz|JDx=aAqZ6gP;By -i@!|2^O>q>BfeNhCz6@c*9C)zVbMGd}5|*6&f0Q?$>t!T1Lt7gXmA#G!2`!QpoX?xhkcq72R?3ser*2 -G}XCh$58KnYno(PR^x2c(5~Pim+fVKI$bGj-)5kFGKyz%*%8#M1iyjW1=et)`Xh|aYos^7%SCao$;rx -!pYBF}?6652Lc39{M!(l8a0q3KfN>Dp?Vyx)2ZNwba;ophc)NYd@V -f4W6^FoC_m<<8cqQqzc=U$}x)vD(KHe -^`*M*8sjtKv}=q6^{fph6%11iz|y)FTcDudA*3-v6FG@n+ -xapiq-9Rr&zN|zz#t4JMByc%nRk#f#ps=q+g)-J@Xbg_`y<$eRJgHdzhn>)vsy3H__XbNw8=Y#hhFoK -FP20l0+HNr>d6NXo`?kK9}zmLdh_@UZbX{NDUn_Emx-Nnn==SS=mpV!1cW=6XeDYpBZTEsuh1s8zglQ -yn*RM@YvJHBjR+_NLhS)uecS3g!JpO-k@d>2QB?{&PIs$-e><)pUw(_9Q6s`45depa6RxPuG|m4GfW? -Bb%n*Ss2;s(o~A#{~}wrb#|z|J;|5Yu<+&KaNYaXeV0Pr4m(2x;FhXh1H6%2Iu`BsO<~(@V(g?yJPBv -y_C7nU9lt0E=t;T74jC6X-9Os-S>?>SWJ3C=8JHLUITPbbzlQNDL3?&!Gyu%`m(+^oOw;hVQ)0^I@P$ -=K)kVEU_zaBN^d&E-_uJ#o;uVF15*C5*Ab&$+le$&X4an$I$}0xbORqW?;|Dgy;-j=E-j)F+7A)Dxo7 -kDAo*A8oPqOAk2v2H%4%Cvb$LqT{&FdQl_&QC-S|o04I6gwSffA16Gk(+Y8NaLe{I -HO;Oz)C)7A~cVMtn7KZeKneJw80wbIlP#U2GI7UTE_7yQtWRm4st~`psLn+9WPJc;+(1jy;;Ec4H92$ -p-!nu$_EB%K!)55VOzj<|yx;U?P3*RhqwY7enn4`{t&C_!Yb)*mOh;qT)Tg`6;0Pojtdkq}dh<>%W23 -ow?tLmmoq2`Uvs#MpskG5)|Ga|06Q7#~wT_jV}{Grik6OQ2`JZdfk3Q-H20=DRe?V7T-Z`?e^L=G@!w -lbJ434!B6F^)}Pd{C}?OZg9ct2$P9-420FQX=wt|AxMIB&vF;Upb4%N43T2(9_u} -Y{jHUP|F&m?H@%s<>odN;!joV*50ZN)v5>xOeYxI7{Nw<^x -gvNX0t%qneklnGR-`b<8tXz?`NH*Rj)VV}WJt9I+kfl9ls9B}(}{QvrFPsJZA=?1u7(w$MsQ~%uINwp -{kHE5RRd9Z@&SKTGyugtsS)^XiF#l-}^7*byKv06je7aLig)U!Sm3-IrzulrG4xEB(&j@35Pq)nV?@&)?|~c!=d({}F(r^FhfaGXGl -d81B#VV~I+?%Q}V$PF!N@-;1_e9NP&_#%W+&1xUWg= -dTb@n|kqTlXdCMV4v7%xBeSYO9KQH000080LuVbTWLfMPLcrt03HMY03iSX0B~t=FJE?LZe(wAFLG&P -XfI=LY;SU3baH8MY-MyVaCxOuyKcfj5bXUGBc}n$aU3EVq(lJww-m-|R3S(VpGc&op<4#l79eA*x@B6bA!B7$)Gej -yttP(7W4PmJ*w<5(Abc)!XEmLR_o`YJ<0J-8Xo+%{_#6BPbYxMekm{?`xfKuw(oFON2^DC_AI)jeRgk -C_p;>XhHk__+MDWMcA+IA``(k`waP_d~KITB~Y;1$)T!F7JZKYo6;#k-;rbqwPapbPIFLfv}jjZsMbt -G(nF^|pt79faFDU$VO>)?|yzVrM6zR^W6pLaNO==FmG8Opm!J8 -EW7@E9=03)-M0{m^+QnmwyV@1x)&WQ@lfO+lr+ZO=)R@6aWAK2ms3fSzDe=Qt^}r007 -$?000{R003}la4%nWWo~3|axZdeV`wj9Z)|mKZgVbhd8JufZ`(E$e)q2+JQScB#cf^uLcD#Sspj>sbv? -^M&21IOXaL~UqI%R8o;&XIt(I&lR%3|&37qCPq59bl&9SD>@|D@SVWl|L -Zs#p7QuFTvSLdd%TnCW6DY@f&ZshOx>3|0lX6BvHS|@`HE;F;}ZY$)p6q%f6~54T; -HiMc}huukVCxOZG&%!6e|>n$(;rj0fDgS!w3m?j+4N{Dl0}IdK;ocAsp?ZT&>*zIp&5n%WIqFa_Vcv;FApT14WO`r^aY#DIU2S}U<^Jhb_!sFIsUzSF*81ffZ%0OoRPY;qz0dKpG -)hCWFgiVG^if?>4<07iz$&aFwY7X2aA!M~ACZU@aQ63tlU_~MM&){cuW;H4KQ6a4pr?W~fo~l`L)#Q5 -5^w_@@>s&CJbM+2#M$<`A@QdiCa!`n_8C|}xvgd@Z@#J+%)*5)NgJ^w#ROa0}86xTkUu-?R6i!Mg$Yb -g@Nu~u&$8o=%RwQ6GA?%r7r*UcqCqj-+!A -;0-33?NV+9_!qzDvX~3YU1Im|0@dt`7ta8pc#$?U8`l`7h>7$n`tD+(u5>gf%TACA2acJ_0}Q_DQ_Ol -&{Q~GbXtTlQ|#7WLr)$Yvd78=X_>N9FwB-<UHw23k@i>)2AK4W7|j(3>Q&y2x+w1c9#HyPET9)$RI -)U;e$Zp0$$Q~*~#E}RoU^-&;fd`aSw{*fGCj;0B5er~oR_I -z|>kaJt;TxP+H!3QBBQaIq`JjWiZBrA%DalQ*N{NkZTlzMA-Lk`H8fKqQSDm4`Ps)ltQrvG*JtGium) -6!#o8egu2M)vmc&7&(TS0?i8wVecYcur#^9BlC@;a<8iH~ZSw)4tQqQD!>r5m8m_}NG|_4Wvh9!3!uW -@mzGimI=k@$R#Z^(Aa6IRkOg4--PNC$C2qaujbTOg(&)2`Y}H(Sf1Kx9A|>DRh<%ZC1ag7&i_@aNK(L -u`)P8Eju&fJWeZjjOF8VrUiWy?bHmR-((UPWu!U?cG7W*FJ`XE?*iH1yos!Yb`RGRTnqRj0?w& -4fO{3~PNr&3W!o9JJ-N!r%~X-D`8t?6Jn1S#>z~Mf>S;kV0V2p)>Lbjd!>eRll`1bruD+#x3e -Y^;GD-f3oV%hCEEe+9?gMH(g-g!-=18f?dF&v_-6R$g>b4}ErDSdSl){ESlsHcpEa1iirC;ALN082Qf --|D%VFcvd$WvO-z-b_&~f&8mii=kSxob9r(t8~&l|u1=$D!oBx|(&Cv0A%fr*Egxf#aIbQgh#tnX8V?sS^dYEjA0=aH_ARSSz{fz}bD*LmPrs&WjVz^wgik -xPC?sCgu@~8$|t7>V6Kg@$qgYN5 -DlsuOE9hr%J6MmaR$yrB}X*o0iP|BCGavaL3(QwjrUr_XoEW6>%y`8}(&(q1NQ^~aj8^y|DQj>jUp*H -$HK++@``wJ>QU91n-!tIEG&19BuUJwYgP?mRog1lHO$HU-rhvR3th&@}H2TxPKxWcCS>T=Iq1~$S(V{ -0%dX2H!dz44aIFTI?^7&Q1klWXx*dO0Tl+n{<;M*Qd^IDQC#p-6I`OTS4C_t4Rs)}(Rl-tMO%Hyz0Hx -ruyUS@WhJF`2uyETG4>xV7}#Dd{CJNDp#W0fVGb4k -ZL0G&6eLM6&A$denlTLI)k6#S_!<#s9DAH9EYi -Oo>vf|dA{D_oYHpeRIx4eP!RG9HQ(sE-FBQD*EBx -7sQweyyppD$%%TFPoZzXmDcNgs0@N0;MP2K%kX0JQ#jhY!LQgeqE&KK_Dr0s-$Bi2|ksq3}u!KveWHy -mG{M>AW;VK~_v9XAX2FHlPZ1QY-O00;of09jiH|6Dt1R{#Lc!Jc4cm4Z*nhkX=7+FV -{dMAZ){~QaCz;$Yj+zrvMBmpzoLg(_n35O+E$X8EbTCR6kAC&+m?JKd9vqtyjm@?DRoS-Io(Z58ZGU0 -f5iQT=a(GR13=-?O-gx8&OYHxBA9>Z7!-jn-+_6{otS -T;%sYHm6u6&Hc9y5?l&)n>UZ5#*`&G5n~QQvWu{q^P3Bo$=d~_t9u1ROk-D!1_1Qlh;@^|Mz1d4&e82bNKlTs4q -er`o8~XAho6obec}~Ax7L$g)?W;lA=dX&o(cj?3oPNB{me4BwI8a@ko~)L-&f&5t%7rfarc#5Pl>F-? -ziv$LY!w_GetWx6b?NIQ(`{F)1t2b+T;!7v)VaYH4S!iIN9o6WF)gc6RZK2M>0&jXr?X;i@O<$0)vNT -|{a1TjWn|{Ls>(_cS96hH<~*mPG`}o=EYd1p&K0h8QTM4SdJQkrSy`#s)X3fEmA^mwRmfT29OK -HHqfFH*r0{+mYs=R4#Ofmd5 -RLdPGAX`iUpeae@Zu+cfG6hQJ_f>wb2wcBC!e9KVo|Mbn9czlu%=;?OP1R%krC;UeFi`W3e(IOaja>} -K@MB1SEEAzJQ>Tz75;Yr_g<8%jE)!kk<=(fuZ(p6HCwoV)_YZbor6=F-z1~xR6@l;l=bOW$6aD0b>z_ -b%1M}K+7Ffa%{}=sn+QOPc`g5$dgb!O=#mr@HJIE(^$y^nR9b6e+`UCw`G+USR{M$Th6e-lA!o2us+I -pqFi&u4{O4k-I@M~-9FK^%cO$|KNNO*Dh>h0@;<7AvXfBr@A_SOEu9=?0_Z0m>pgO`Utz%R9PjubO#f -O#b2ane)0{%pIqwIx;oAVd|_%lzFtXEL+W+o+>)e -XZUxb$Rn@dTLn$Qmzi3U3d-Tfyqe9+s}RhroKLZCRW9LgKA#uMx~Sp1;w)vAtLj@@xIrcd<%^4Kah}t -};6A&xb#SOAZ1>>1Jq1=5=qpa%R5w&EJwAE4e}K&T;E<7&rgfvliZmS&W9$!-sAQTZByii|)M~I -hA!2-7FaR@-d48TvZZx0WNzRlItn2viX7lVEJ3}uLeqmrK1g9ZZUl)twa&>7jYjQmR+gRZ-{3x7ob5AoInfgu~BHqPVjLx5p20Y9%e{?nemXdqWH$b7 -hJGs-&;Bp{l7t^3h-h8qANx^>>We{m$73-Xp7BL4cI6)CamAn}U=t2FJqj|V;=Rc@+I#foc=sa5@Zgh&|pc;sppSF0tcCi -X{Od?V3!o#ni5&!J1wFzx --_Ej6zg8jxb(wL**5b?J8j4)A&DPfX)0KQ-R68M&@C0H8X@@YeJl -`8_1SpI1uXae%B#+-tjG-Yyn3XAgebe(wms}ZWhJ~k!9sa0_J+~MJbX_xcENc7=MW3&;e|7ND_VZ0@n -@J%)vpANSE99)#KuIQ+kT)fEYrmt_~tiEek0XX{+t~i^EtWn*&NzU@*5nNe3f;EvDEP3a5Lw2lXN@FJ --q%o3sHKiRHFrYZrJyYM+dHMYgs5_Tcc7B=dsBKuatppnbcQhJxjWTFp{c!8m5;Xg@qj}SJIMHDmWtu -l!U3CiRz@knE|uM|lk@`c20FNwSL?VQ0ckQK6g&yiofDzv1CXtT=%md|;Z72p=M|`2 -H5K|+qpu#&*6qY6D#m8K&)!9t#5dK1RFotz@&{E+Y{%7c8m|XL&cVjoa2sTq{c1)n$ch~CiwfhK8aW_ -8`(p{Wel*<)idwZ?-9^0c!#UN5eXJdUQKjJ5*1n4&jjf5+p?tHPjM6&D)1Ppn2-pPojn4$KshYV~GDT -H^X0Lx!;X2xH*O9@uUgjZSPaE+znJq@r8RYLG)+L@2@JlO%}v2*%gX!N1qYR<V&MItm>SF1 -XnuITwm)0l~}PWE@4O_kR2^o&2rU@^pHB1vOn>y1gM)E>q_2x!i!KH1DAGG$J9RF>T(BIyrWL{T3Q^P -<>yLZ;h}{}O(^U%aE$|O(P`$_la%f%O-F%czxyVubHi`dx^yahgi9Dx-Yx2ZK9hi@Db#x*ZN%_f>8-^ -bJlMF{p_^fz=i-2&Fzv|1-{p-*sz{y$#nHAEGq-g<^Si67%HS@^*_i^RgPz310s;+1_Ax8aE+7GUXzp -)Q3CvYJ>N((|6Rdk^RxMgu@KO~!o8JH}D!d*8`-rT%AycZ+`UR@$kv$15(N(Pgc5?X`7exR&+B~FKT; -x^J(1o6wYPid5-#sR8#Dx1wJ1x9;U9G3l>Cjg#f7z93}uZR`I{SE3nnx#S^R5djUZHlR9?K -p%nKR{e-xE}9J}4h_B5Q(Fr@Bjjh^A!LCM9Eax4wiudN)rLIrg44cLyXJ@J%ZKh6WY{*A#%%oo@Fcc) -lL|{KJ?n~?}$x9#V?gbzxwRIEXdS|QI3!VJy;s!ggFRASw`Afs>?5;zbkeI3D(bD2`5zJ5OlYUgv=1}&ZYHGy?HT42 -5sB9<)$m{j8c?D`wkO8D$Eut6*J13eq!cTY?k^22;VxBl58$>wUR7Fh#Ejbt*q0e7T=_hpaKUHPJz=I -LZr)n!HeY)Cw0#~Pg~wDx+P+-(hW6K&vqlKJH^F|wE)GcNnsk1_rqOa@*GO^xM=r&RR^{#ck=>?FR1g -2!vvgc-PFKN1r$%9{W{4HXBNj(gHnC$<|qJSE(Y%^GkTgKHPXR8nfGk8*W*$kCNSdGD{$$XE -?<#UQyj!^v7)_COJ_h@@JW9+NH|p+ -xm95^Udy6)y974@#H?I#^LFv#`zDGJ=^pko9S+5iI6g_0-&j0SFoyXmV&nu94@=iuqierR$}dRF~@j* -4C?g+%2K!4yXq%Y@3};rIU&(DqU8bax*-Is{f(I;CcAQMUB2;UQ)mc4FvKY{x -@E8>5ulP7GVnQsFu0H12v-A_QvQ#nD%IZyaOF<`vlMuGZG)mm;J{Xpc;KolgE!gT$!KPZLtkm!j -~+NU=tJzU=Uj>q5Ou7cdY&i>mx2~+&Q}nLkk-D;D{bh{}(iF_8Rp~UP24+I2?wyzAIo!sSiY$A80Co882M~l6M3t2FNOdh=eC+pL?w;&^bF}~BKaSI -*w+AQtulLfw?j0TPA0G6aD%>@7ru?6;|KEL(d9w+eevjcfGpzt8=;@(`?N0~*{6`}Xov3WEZD@DKprJ -V0p$%Qdkiz~OL;-`BT;H#H2<|3y+u#3pklfIdDn0(&@yXt6;ZOy)RMV`Qwz@%KsqHWRxW=2RQD}+zQ= -J^W-NUB29DDJ2rP#X>ovu=+r|H4&YjlFVt+A+DHHv8HsF#8Eo*n~@IQ2$~BVE@0!Ci+3)CmS2cvVgm( -)uO}!DtQ-?Bza{O7JP0N@Mfd-i_5NmbsyNI<3LqbVvH0!n-yt9EK_Av%h6JQ{+)rH)D`-2BMyl8>jiT -zT|&qbcKuCNepKRqom{9_nig@B4e&2e&}BuCz%CfUMTnUM_z0dQEnn#ap>3I^7~Dh*XAKxI&6LoJ -FrlC+EouZ0(@%d@;{JB}EbQ46#-TqV0&4};2($oO21hgw>^`ixjNW1nUr^!MUk&pj*nZ}C-UD>T~f+5 -ok0f#>t`Os%n2xO$^#Ef80e=uNv5Y}^C9Ym_u=?R^=q!O_-6eIsm7)y_6)QrAUNNELd)zr=FdWAX>jW -0D)wuH-=RfUWLsjy^WNBdido!GcOP_c5Of%Llc;%|MQg8%-##pEA?#qeHronorPR17b^X2{yAMQ?Z+J -RauWpXN}|q?suoBc0)To9l~$ex=v4Fp{pNumpVNiC9m@ABRILin{3sTP!6P3^_37V6a{59$&K%4()V3 ->#2Co6@9C6X4^5p3N5YsW;I&m+a)VE(MeFU^p}wTfuxxpszG#`nxEi*0gG;XOMPa3x`}9OsQw%r{I>I -F1)q%R4HRN&9)$noNU`g}=O?Zb*lqid;P;E)*G(!%$rX9M}@3Xal0KK~~zgSI6MTxHuiRQ>K93=+bIN -3VbaiHgZ9V{;B$_8X%2rvMQ@bna%{S|cPdk}G`P@+Wlcre6GfB(=O3q5A7+?8r9cT8K^E -A*-=;Ethm8=xN97XGm+a&ss?w2geD9^@HHA+WF&e^G}=B_6Y3R9C018F;`j94C6ae$47NXm?2*tJwyG -eR$A<>u0}t6WYSu&NL87-`wiBgh~{oAS{>j8~kEW6Wnniq^`xWC{tpj>lfvecHue6LJwMsO?H+-NPrN -12S)Hg-9`uF?_2%@6POfKrV&udJ2@|z;niaG@5#`><4<1;~QFSU -!f59^QTr{u)oK8HiVX(tV{&)FtJ*s1c_&Z*vwsp4syVhA@t45bu;W=C^+4O`DzAn8Yn!4&S!<}#3=ak --)1e0z8|1k}^R0=+JZv2J~>eFoaQ%$j0i3&ZiOqyG@{3InBWf?`$pT+;~DQ33`}sCskC8$-RWxg|tv+ -oU<*9I{EY{Zj{=%4>L&@#(-}eG%3N!g^7FFGcZMq3165eb_NH4B9oQb5k>iY`PCzMtz+jV$}Bu!y~;P -wNBSHP{QqU@7@3NpYM}iWU4_}@q!dN{bMMyf*89a`?gb!EBjiGW_IE(u4etX49m`Dd3sh}M;biJ0b7d -A!)xv1g9w_Ii$^Bvm=ZNSdE8OgnW%()8?8xAn||!h;0tFS&U5X5!q@N!CxeSNU43N0^UmH!{wsVv6g> -=FrYN@>17}fQx&4mx`CVTd^**|SV2;?XiFN4dz|(y3v8c+$jt(hEtANd$%X-A`H@DP#ER -FW;Aw?XGy;C_yx5AUdP-xPIq%RvvrP$DFzU*B;^pvZ&95h;@i1@|%D6nB^A$zZ_VPxz!^FubJwO8BlJ -7&WOVr&}_5IqHq{X;ZU522u3*#?B26`lA90feqlDxSqyq_uZn+E*m6w2e#0ju7`= -!n6Q3+OBtcHy8brffgeevY{@bQ1$-N@f!BTHMMyM(?Kn8?S|ngae=^g(u?%Zq4%D*ub{fh3k}&9e9jL -xwQd>kF3>V6$;Sn(I71JiYj)aJbv7@vCQ{GxKvOSZhHfo!o$-m<0m!%6GD|pdz^1vXrG#EI95$&2`O5 -$RMVzkJZJuO`jZF=!@+9MUb5#aqoQpHnjH~Z(43M2yvTGjLAZ7y{8l#2=@ZoB(wpW+dbys&EuA2Hi9j -}q2IQBwoP}PtRNuUZ35#D>%=V@_XH2p!?Dfnn_U@+tIs>aM+eZv4BOiMsYA6stliFR~pf4B~0VNlC%R -1uUmL{V(JHE1%P&k|2q^FCNIP@<7JJsLJ=;IZNLZUH{Qj+x)PFVnVwB#cTtagB+m$KGwg=bdsMjP58a -WGx+u-4~cb_S-boTG{Na%fLp>6Z4tkTwO0Le7Br=FYfBQ7kTHxjxXq9mb5M64HmClxO%}RPhzN*Bj{R -q9g8it$O`Qx>vw+bf^rrUFCv#VtSWYFMMjUrGy>enNKAfSpaSo8i##rFkl*`NJcYd%2*YeDrofsVECA -;|2wE@D5EuTqZ%8vbCP~)AA!&NYEGs*N?A)4kSjRc#yfLFWf?GfzKQ;hG4|#RVcl|Cqv?wU&p8lPD8O -$QMvYORTk4{^y%#G67-xl->pCOqFV)=!g;mhK-YL16dA$^KvWdP4*huvs5gXJD{Md03avdU25?9F7cm>Y}a -%^z(awhI39RtGi`92|yPz%+lZbr>n~|^ENwJZD5fqJy0onio%}(ND7Nics6MO~V -5WSf=s2YMh9@oKb;=faR5=!pII?I*>}lHF6Edr -_{ZSm{m*@pG!+mj@_NOT9*N`j^yQ9&xH>zd@>$f68a4w4n%fg~7$w3HlGz73IRz#bEdBj9a)1#>9?{t -nWpd2G2wCz$GIow#*(MTI@iX!fP)Sy)Yk=vWSaIWH8UbG4Sc)~_n-C@1t5d7o$x_T|c<93~*Kh4C=Yb -{?2;)#=5}@&bHBv4b$*@N!fwv#KdzCH8wyJS~HF){|L0up0zv7&SJJ(u)Kr*n|`K -!SIJ*>)KnO%9y=SsJOP+CL<1hCd`n+jm=@DpF%bk(GlT;-`)?h6pFRa=T;pz0Ns(vC}mwooi1ou#o?@Z(~xccjXfY*y>27oR96$;-vvL#rf;l(jN_G1b36^~iu!jNXFMlfmSH-k?5X-uHh2Z) -0mx4mRlV-PQ}=r&;qP2H%Sw6VE>yJm%=VxmsGwa&^h=c5GfT;=kztc^hz?OFmlgjVu^Q_8my^pSjJz? -wq#F1f&qcLzIs&pp3M=zJ3QKo>DQ!q6SQY(C1COGPn#coJPN;BoF(8o;?v5bZ?asj>tSC)pQJ2=UWG2 -U8b!Y7q+!hP6Fj8R3Nw3jIS=VtxJ5vyeU -cSNeCO-Se%hVYsmCnarnl_OK!3s)xxU}wc9#T -z@Wx%`mu^uv%bUwh_fe+(dslF(GvY)m#k*4 -*Qb!_C3cRj`;(v0=G!;8Hu7vwIEswx0IPoP%%KMZT=^k43qvecSWzFv{#)?NIeH&px0_z)iW*=HAjKc -5OU93|iYz+-JF_hituW8bvF19o@GPK3jU9=;XfA93KtWS(z4aUHIszd!hB-KEF|ujCKu-vM%$>aJh&h -tc=FjIDrMn{MTdOmPq2Wf0RS$R305oBu%=R@VY6HEqIf=rkgT=p6-vC#d{NN81a$2{}uy0P77E)XzNG -gC`qK4t`tqMXFR(ZIFXX?K>#Pt^mA;<>cM1S93EQ9p__9#2ua>e^LnBVTNn|hg@KVaTo27en2mX`6A=E4{$2qAdY -gDLn_+^{^a2*OLG=fZ`Xzm6ixT)fE6O5V5u_gjN$qEAr0Vbgmhsi!*yFy{nWWFhyki0jwE^ySNZ8im;B;T&4=fphsr*lV7sJck=Pqz=nLd|)aUSjI^H0_b -$jd?k;uVk`5r%Sy;H8C&F#C@vU=TCSJIoN>iRdRKq1O#=q%!*6omS^bP)7&UA86;6KiB}sxl*|@2Iiu -7uWQgk|iPJV6?I&*R1LzKWFAlj&kage!^e^8Y9=+HjW&)GBLF3kP(b1^c0&u}PDKD35H>1#S!G-aQ!` -E+iPxilgwMW`FF|E&pnu2MYDFx7G!Ae~C4My&YgXkU9+Vf~@_iD+%IuS -)D&MMTpY1t$>3vF-N_~MSfr|5bcmGOSS=`ce{|30=-=**XUI2@WM01whb%nrp-7P4Tj%~OFqeoscK*+ -;KXe4XAH(l_1p_$dM3?nD16IbVGv&Y78)pcR+MlnQ#31*w)&agz^1TwWeKbk&n}bm%Bw{$sul^1GZgHsNkiHnY(fs&eWh&zeiUz_;#W_mXTJxj1vno19GJ=fj{6^a8s;0 -o9xmS_q>?^e?xJn251IGt#K6Wn!S|U{-80?x^Tmi!9}n&GhOaUvKE9N1X)n>=G{E(&Lkt>fh1ct~v- -EzCC&K_C#D4(@r7F$SoHca;i_Hl&**&dnasve=?&PT4P`exMh%p2Jd~Z1^0%6H?01y+9nGn6w>}N6nn -`qIXuQM1H8a~QIaL|dzyeh-26Lc`fUo_!;1T17TlfI+n*%C9C -G9&jDMpj`WaqHf_wlfsAZk6rsb1hCLH5#IBb6TY%A!14x>=1Wzxd%#d39qA6;n5sU -RAyc*)cMH&v3_ZWU{Ryf4>u8l9V(X7WX6p=Y(#tMQv^y;=i1%} -#wqf3lcXD6U+ej+Jm>!6BhlZGNJ;Zlr&W)lz$oaW0C+{&uBElziuG?-n+9Ba -@;9|DG~Vn_X_o8@Py|qysBjiq$#Ry_7fhw%vV5;y -bFpd*MuGOoKuP0#Vyn;Z@&CRly&}}I#P{=qr&%v?+gIQ6w(hoUCZ&EUFIq%-MsW1IEbW7?)IiaINF(@AQl}PR^s*DyFTen7g*3fXN7FRbH7>Icpk -l9^bdm`lTwU?roP(A`ix;xXEZnBouY6}2BmwPrZlkhpU#!#dl4yukPJc3IVjA -JCOTq^s?ueab}J5id*w^6p@twk#aXA2x -@fZAlu%8wZ=)En?NNQY$Ji_kpvvREu&zfsZ#LeEs9INFR1f8;n43-n4@y3ffw})CnDNlx4=xurlrAe7HdI1%=V3U}pZ%e5Ufug -&a@8SZ -&efyv7B_Ymw3z)d|Bb#(IkB4jlImZ9}(2f2Kjl)MV&uo3r^am}Nl}LO<%pabF%s;f62Z;v~M;BA -cySvvp)QJdB>ahB;sEe~=UNkr61kLbbGcB~&am96%RXbx<-v?2na=hU6o_SuMUIr>Qw8fZw8(?lvry& -e-5TYfiz~DjIfKk*8%@j>+#X?8=6UK}AqNeoV8S3@;sHBAZd?s%d0k7gXz}b((%!k_IQD!_07NLNeX> -*5~`UmAz?>dQr$I}|zkn(V$H$Mkb=2Tw>aW-}1eGS*5HaIq7b#etEf#I0hI8u#suJm#tc95p)74A8&t -tVSSX@oP6Zf#E-(lzh`A7+9QoPY^T$3?S8@*CJnsT!KJuQC&X^Zv6FhqIxHlVf+}48vW~G$UzKG{$Kd -zoY!s?m9$~L7TaM)oAbb39uOT4_O5_h^RuW*bW9VV=ZcZkUs)j9tE$bn0(%Vwxh>-VgSYMDPmX4*FZI -ecu7j>=HvWahcBG;3Kq6qbJ$wSKEFoB&BgEqWKxA7X?p(q3_bf4r@JzF&#f3 -6f#l-xKiaoJ8w?2Z`!F&t)04HQ?{lMQBnR-9{-Us_bP#so|NTsb*C9Cd+3w!sP*rUWrI#dkAD??S0?A -S}^UR$++h$$T|6m5hyEW6W*|2kO+TPL6Qth3HgBSk71H1>|DOo5{#B9Ox6G*$ly@avl<;Cz;^h#JW^P -)mLhx0L-#65?}*k5*X;h5suB|&vLXqRh2?^95I(htbowZ(ah1GKrYVqMfre@0-e5ahF#dyqT*D6ri2F -(axUEKX(!Dz$097Ljt6`lKGQO%tI_jSR%L2u=IX&JL}$H8vqEJabNMiJt-}0-hQY4JMhIUR4gFIiS7> -u9L##iioJJfN|9`lMM*EbfzYWqZ>+qKM$wu6oAeNHHg)Uwc;d?PPg3O~(^n2vXei03d$miUmNb!a7b? -QucT8)Q4rxJLqjL2G~ZtY$||K;ZBHoASXSIs?P*oeX)oi5O@RbZ}$OBI$|J2!i$&3-WXWAUBmp7eugJ -1!FhbH2*<3{YWW>&GHHqzK|5jl4pl66q@)g5u{+KG~5*w0i2Q$GNv70qxZ0xj>U;vR@K`_ZAe`3wEnY -?Y@=SEo>eqd#&tb5umIh&4oz*1f~?Hs1Tq=c@SZ~o?@A78%1J3M75}EB6_BY`i+S-Or>rH~^B -brIlc17M-w+Bx!j~JaqnJ>>Fbiv5oK;!X?fJK6sc8K0)~q4>y$2ExsM}&!>e?oh_LRnxb+0{llH+-T> -b46#Nu&2)QX&2LIk0PbQ9Gj}bi&b7t}^o@)qNhfttKhb=+ZWsaPn31+_y+SC@Bq$&E99OZCN~q0~_{q -SwcLF0+L_muqmNxhJr9e#2>Mi(JfaGUN6rh;il`6(m0_^$64LMU6X<1Q(- -Z)5}C=>mCKEmxeVJ9NnCVKL5;BsQ-ee0%c8U2g>G&|MvMV@toRt-BXce0AWuE9xK@)${Nr2a(C|4}vhLwUGHar -!WSQ|!u*+4>xr%W>Bw80@-rn{4gYu99DVp5D8`QK!04iC3LRio_>nRSvGtp^|3Xh*;|V2{`rxW*HU|?JQ{#{ywN>H7WNY$j>D*>(iiw|CM6lD;7D+&>-cK=QyF9;2kNO9`&n5Jo4w!qENnO*?m5Cv$v0@Ys=nz|#jH_-hD*UJODOB;eRd0#No>2+h26{n -Fi@d15@MnUGZ&DTIoLA_0aS4xV<(Vu+%#gW8rtIL*8HcXoZ%2-!z{YAfdW6kM -MJd2WhWj1r$Ovf$Yx#`U1N<)4j}AFBt{ESCJ?EIFenR=&Nv$L->`tj%~*fA$J7h*lPQTH-HVsD2r(d@ -VampE5#Z4LDEAKNR$Gq<9=H=y93y0Vqsz>l8JsjldTkuoA -O8QzIAH3?CL%CX)=Ch)Yf%CUQVHVGH_UN>f9`K8Ik<4Srl4!NAGU#TNst`^N^3Zx&5x`2h?vV@&mKnUK-+#-}a@795cz=lx45YUrSiRcq~#nh^8XO-HH -y;C0sy)%c7%!Jz_FS7%c)iDI%^jsDGW#O3?I(H}W6~Bh%4|Uzm{~iq5cEg!JLhDh -eh-6l0Whzz`WK!{cpfyA(T`=Z=tsn8_>cfij=i@3_Cj9YN9w)P6IV7ix8d<_Y}R7OK`4x#tJYe&gdJ^ -HS&I{Kdm%Q6VGh|^%v#l0W197a^J*+9tmOhFJVUyBNr&|i5Yrkwx6NG_*nlfW+K{WEjbPWbo#xj|h^4 -LGx0_&X+SVH-*fuxLFxuu};P)nGCL(bdo?~rqCBt+2DLEBIOVS65mIT)Fh+-vkTQz2wMY5dQZ%{6udT -S6WMV7v(`aV1|+n3_PV-=j2?e0f@Wt-nBH7P-QsBkatsu9Bln{9BIha|D$e)tfgT9uFt&ZB`-`A0;n(ayM8@@x?m#1 -_P7flEFd*{@8AP9F8vW87T+!?#CZtnd+^vP`K|Hs8sfP^%GHiI9lL7KCAz%ip>Ntp{%AE9=YD)M#)ss -CGvY&6$9rN(ogd@>URGZUKLWsI&=L0hA??v){<))La>OZbe}U0Y7Ukjz0_98=)_r$PEEX)8lA| -x5f?mk#(gv!mb8AQJg(+!ud1_4n?FYVNdzF~1V{`TByVi7~-PqX}2AGn52rP-e3iOt%rj6q~Hs^jw0~ -MV1fe9qN@XQa-|Go(Mji=KSjOQmE(u27_BqKOK@|U@rR1B%SdNR*H=JVt#!&m|>(ZblEjjZ`i6P@{gk -lId@s}jO@FLAG=Q}R$Xk}c`R0riA<>!?qRkP-mgQ?1H{M-931u%tZfk+*DemeuMMNII_0kP55?@R-tC -eSpg_l3rXv&&QZf3hM13CfUK3w=@=l4T=j=V|Hzv~nZlwNyOy)?F;_J`wM&$hltvJ-RT%}yC%Y -gxEe7jzk?@;gb-#x72_t%~ZuDPdrP2`?T!8C{3@(IUMHwv?+2HZM85*>GRf;)bS`E0v@TKJJ*cQ;UL& -vLQm_gtIRSGnv{7OLs`WEH97b0!>LgGzH -m5XPGEHHJJs@2PZ=jBkmq3NsmaVlM6ccLYn_osKZpVebn}`YwchFL0+D -D`s=8sIt@VYylV+hICHq0QWQyj*ON@&TWy0H*bGNGuG#MA$@6o+Ui5MK^5pB35!s^dj^QKNc{2KlHTI -t?Zio&Ubb9*~@ZSxY*7=uFUG*V9Tx^@sIzDNCHtxt~Qc(qz8c_U&b@pG7hQ92{V*!tQ+9$LolXj>8{> -mie_!}P`Xy9eLxMUYHMn%X=v^l6Cy{t|N-@HOoS&)2Hj^UA$hg%2s1bQ%n<#cHR>gt^X8{t91X^3$nF -kGq7h{%gJO)h1UG4{tyE4n8set=7WuSwE6taDx$EUdT`UO>oiZoqd8zsz|)&;Mmjav5JC+i(4`V24a1 -r!lfjj0XcM$8&~9#6+T}l&x7YJ@x>k5EO286mo$xh48uJLMvON5AUcG*yLQy@$#-Xz!B~d_qDu -djPY8>JkfDgqo{gB^8eXPbxVDVwW3C0%5TmiM8dt4*SbwsnAth_a|It -EP1XZAuBNgszjn93LKx5>}C8K*X~gg2)w9dex<^fGHIR1i(L^(kp$x!fEdal&lN$oAlM7257E|pm^_u -_%!nZ+*lj%PRG=M9K$xWEF)^rngfp-T7pqyh!qs8skSR>Z?_%*+YTv?D&8EwcGjJKH1Wlv8>@E-zC*E -oW|aHQ%`#sbp-ob4tk;zFSda4er0$p)!3KG4rW7 -Gi`u+w6K52_m^l+IZTu5y>fa+H2l{JgA2cx^@NkFIv1o6hl2FV0XpgEwx>99-$pMha^++}Pl0D5}yr9 -Mhi%NV14Ay`IMATkIrdm&REB2#dO8K6ZbG#SPt#&9|@+)8IIWVQU7Vr+sX`wVk`dJ)5-v!ng?Lr_Csg -tjtXmoVW}!>*3_<6d?+dA~dx(fWIch3YHr`Ozrpj4X*4*hr%dUNzw=lYKQA0{@CYA7?zZj~vnMX_yu|^CoHG~#IXN#Nh;=!e)%j0}J3|Kd(uU -uy3*&-%@Ymtu^a$k<>J?~)v8y(?4CWkPG!P@YJ(Q|!?>fvFbgjqeRQ2+C+=TUR+Ul8#F-Ig5IBR#_P( -W3b?*RE))mqd#mAQ@3?gg^#Fpja2*gNBPndbYwUC|8lG3|5*v6w*)a%kMi3VzQBFc>KV=QX-tUMmoAq -Inh}Fwu$^m4SOVZOQ!_f{<|>;&C^OUt$>brE7<|9ws-#AQLK$_$Lm@0V*_8PX1pf6qw{*40qfXSqE!7 -v&=v^j1Ua0jps!pk0l(-D-xwhLDQMQo++^A`+62dW92Zf58_oElPS`xfo9CP4gl59Ae>}5T}>#XUo@M -C#Q5o8V`h_ZB4@R92M=XDoiKhmCM2fiH^GCL2jK#;d6xH5O!1yg2hU-!cWF^)ua6BI -la--jV7j`9CtoUF#?+pvu=#^U=hI=f^2#kZZ*Q!ut)rFo-)V~=_UCCsq@45{6U2^k$37M@~=riwtO5a -HiqjrL4M9)zyl*7+&)&yfnPASkz0fJ&`p9J3Lt}RehvlWFIe!NkczT_Z>uZpbito$ -&>`$7IrLJ5dETE9)aN(oJ)x<_A9iR#|9{BfvX4=Q%2Q+uuEV)H-!@Q0D)TYlOEJ$WHiUMD(L5lF+=D?Gp|1>?9%QF2Cq<^yQp<^<(RK@bM=NzmzT|M -uTGYcC)AI_Z{tLXvSa0kQhVYy#W|~;re9=Bh-qz -d&wcW$MD@lxEofeP-8g@HR`+Rj#pjQ$03&evpZOIN3j%R7GSBjxN8G9G6oKZMv#ybe2cbjei<%mg{TK -ds%fQA;O_>Ii5<1G3KY>sne^=u4V&&w;wWYzeiuslt0AuQG(yrbd0Z|ybK8+je3hfbfloZEfFUgzcce -+OIhf004B!rmGooubO5Vy6OeyGz)9Ri2ZajVUS@p@Y!xhT6X&L+$@V0)?y@-Y;*({Hr-P-M1Td2}e#v -->3(0o79-pXR7$QIeTF(+Wk&Xfs+nSPlIg=l6+E_(X~e?hgpP&)kAs?M^r6>sN)B(7Sz>{FzBVo?$R7 -R$TD{$3ss>}`Q1ddMUf33X=b=Ci22Xx-47otJD?#QNDNHVW8OMz$QIUFVVF7Hh2x654n;)(8 -ePA8qlnY3^}i*LNmb}sV(fXjzp#4Ucn5jdJZ}u1yNA`u8RalPCSjYo|MbHZwJM>e#gCc#d~_%^TY6X^ -b}c3lzQNt*J68kmUs{o0Y6>Y!e`wLI{Em6mT494(Z3qT^_u_ooVLPVlt^DQjsPKHZZzn($) -zsTyj(~dnr#XYQ#A87e}U-Q<7o3`H9tTke~5;JEtC#`ako>k?Ce334TWp1;M2L}N3|DUZEsFx5y$}uN -SmYjZrP~H@5Aaja2Z$SAl^)m;8eOEbAXLYKp+6B=CB9g4sxxi*(Qd|7$8qemJ+2rteB%s*Lk>-4#>3* -SlwjeJe#gQmF0H%e;*q)|>Rf39ihNyw9S7(%wr*V=GpXC?X#{!~Kn5xHlo?JA|vfg?6bXrc-#&M&-EJ -kH@{`9k_oQ3ddU7W*N=AyZrf1+4z{-j(yQO^M3NEy?-$%?iFzc}UpSQUBWJN!)Zj}UhZ?%8X#7INHp_ -~8(q9i~4V9Ui>;o26YSuv%{SO{w}u?N!ZGA)U_brL^#-3%cW*TJ0Wt3Z?z$f^gj1t#-O`vei@RAY~xm -wL*eN2)RTifc<=QKGHXki~%)KP*V^Nh0Vn@lh!q07 -S<^1iaw^=v&T@MV8n*Whl^{{TZ -?2uKZ~}ci58j&%(*2Oow(A1i@&rG9G{J9j1QC!C$>a9|7R;7fC=4=yn470T1AUdRRxmi~`7ZV;4JD~4dR@wPwwgcBk6 -Y$fDON@9`#MSgg5ns1$MQiX&r~gCuv?Zogjxa8MuA@`<#2qep>_6-uygdBjIQ||!>iC_~!cHk`du?+c -#{EJBbnKd2yS*j&5F1sXL3(7Kia_9dyMGZk$F)08UH2JZx0{sLXSMmI$0skXp9~}eIlCC5eQzHze3(y-*V46zDlEPzLE%tVA}%QYPHW -tU*OP;39rJ`~X|u{9wibcvQxX64<$t0h@9PyGAYD*-JDInyWc~Rh*()%=Vut<}N;m=aeV$#yp6>3?sA -vN-q{{5}qfXcy+;kIWwWg0+=sR1TJ3?Pk*Q@-qKYjV9(CaP^;9L!MbvBxmmrn`I(+U4YQ1#O8a9w*PAZGWCTeVTmUO$ENB6Z?R -!CgwH|5GXRVE5V=+EPS@OT|9)y)xpj$Ldm5A@jk8dmmG=jAssCo~`J=Qza1 -Y`&^5TBtO3l(EO%aHfVVs|izCtxUzEuwwDCl-5)UlpGwm2t)b6>fu)-sUpzswJn>5*%J81+1nGO=qWG -6%tKFgqNPP$)_rOUtHX^6*$Vn$VJ)s72RaT%Fm4x#jriLo`I;KARZ&n;4zDYsz4GJ)=kj>IVdl?q^1& -dZU3V7niKkPG&BR##Xs_ZTf9(E@SKBDeJ4{#KnL?Tp3Xa=~8e$<$Hr|Q8TV{0)cB`#M@OWYohC0xf{lkd=oxjFkl6nQq+G!k!+p7$cZvnC5$ -l31r0&Tae5h}#FG%!fEJ|vP!o(3^)_Zu+3{ZQ+K?=_mssfvnOzrn-DqgqjaF;{1ZWCwKqJEz$0sl5Pz -Ua2NO77*h(pR2P&rG0Xed|H?|yBCV>luF8<<(Znx995oS8|o0ZbE;4_h5Dy#at-H!WtD#{%2hqTN#t$Xa1DIU%6=9NvKd|i;gEwco~NhK3Eu#Yh8=;wKQlEz4y%hKr+6V0{9cEM;jH)5kqu3V++4AGeMS! -EuqB}328-o5h$!QuzQuW%ZIt`r${`Gz=YwI=J1H55;lzSHn4D7WB4B4T!L5wqF2t{H)?Kk^Ni3p{=L3 -?KDx!i;-;v2k&0~z7OybnIoF52{DU2}E?X|*GZ0&B+dl@0DwTmKZ*d~30kMZ0j@GR1;83qIaLsk)#vj --xlX{iclhU8{HturOC3uXisF_aMT-h3&zPV8VzsBzCek|i#C5rd@?3Oqj}vqIA`&WrH|z8f=R>Qe#6R -KHR7rqiVU(%en5C8_X`I~8JOjQ?v2pxFDuXo(Z&Es4nOYK=P&QXh6Q_5o(@0 -st|wg1nZ6sJb8UT+!H=Roxt-*K=zQ<#o2tS`h|X6>l}NF1qUSzS4qcN#4MHXPhP+J<@a!QTXitvB`p$ -f*y1+6%gw>1{MTGTVlF~!{BP-bZF~y&5CUi&CHo7|WTE9KynA%3z~tvuxmrx`9)+j$N4Ma_W8NdvdUO -lz!>{@Rjhyre@X;+z`9+>LcaPAxMp6Qtup+UN1B&V944Qpdb^ukUI=I(T@r@zJqQ1;NsDjAgf`QEy(V -%v#N%E{N$k1c5q66>}9K2X!A5r7n?V;a%etDKpr@6M}BG-!Syqg8_$>(Pgu7H`~)c&SqEDd%BW+B3Sv -qL3liuzhXN=>aw&|lTg%e`-R-@ZCYPxg*p?;q^GN>9Gud%Y+0A({$NvI&g#GfQOU|!eS-aAG1K-lPS+yJx47DzH$*yWbg1U -rwUq|{l!}amVX#<%ZHy!0(k!aH^ -f&EsQ`wfOlUbr^BHcNj?hJQB2AgEfHRM`?PMq9EamJw2S|pVj>?bm9pzI!vJ52hOK!qdEF3iD_M?9K2 -4tg>=cN`z2jbWEBkZTdoX3Lzrgosx1|Ff5#Ntw4Stc$rhv(I79Xi#xV{gh%}fA~nAm@sHD_{*g`Qn@J -!v)Hvmn}07<98g>y13@CAYV?7S`R}UBrFG-G;zhlT(7ZBaV8dx}46xQ?OBSfAw0b{cUd -v0WJ3uVFDO8jfkLjqDS;;Tm1!*I_*?GU|HvNn5?2|B&w>(mU;U9iQy+|JOWGk<)lY7P;@Ss$B)qajwE -!AJPecG6FU1zlsV1CW4Hre?Of4emK*g-_+qu9;c98UEPq&N(}MFq7X?9QX9#P#rI+RG>n~Ju^>agfGP -@x8mgk|kL}Sj1mC2L*c=-J;Km%L6cN9$B{$S^wIt-$*{F2{yCyr+XL{`tcbT2%_hiZ!5Cgz@uUG;zXA?(Bu -R{@d(ywwe-mGs(+mXyO9&p`&t0 -2Sg>_F6Yn}>>LIfyvRG7g#J+sQS3E`yPv_H0S`Iij*cC5dTNp+2dhTSrEbXOpur+a`nZS!;-GMcYP1} -lmI^Zo0u&n^ko#gLD-u+Ku;@o0Rg0E)n4)G!x|T-2Eh_MPQ&M)n5h#X!48ko!Z8YH4nW+u?ks=W_San -K9yOU&Epo!dq>H=~YrpDSJV|(IB&$Ell!RagB#1n5 -uFG)-pMTU+lj)KIRQ_m@Ggh4v1x14{%4NTe)V)i9G~<)p54GR8$FGgXr~tmeu3{3C+{u%%_@@xjIRq6 -pZ=AO!E1>)_ySXBULAJTZv9vlWcO4vrgogR2|HPZ|`hb*qt!f(#La|Z)u0(;e=e8d-#I#N~l2|>kv7B -Gd!gsPGRv&j*dt(u`3o%9|P?2u|$-4hu__E@&tV2R~(YSdxZv8F)yo@(^)vdGkp_Hu(V!NpRMUJrn#p -%4-d7d;*x~{&uS~m$>K4llkuF{&)1H_KP`5QLqiKj!Y?)*i!OfO(HNaj1@0v=#SB6kiedA&VG@qbK${ -j38yy2ct&3?|ouBnx*h9trJ|D!XX=F4RN}ikVB)8@m?rSbgAI{48G|~wlVh?!UldD{(a+*)u02WO)FD -4s<VJiIf3eSGgf-y9yD@NJ1WDNHfAj1Bi-rRj^TN=+>J?Pu-_T~U^L97+SG5}sHLG`H6C -(e}3Ok~-a;7mKvP)S%D(#}z{eUYwOpQ(ng2KxneWi3hoallJ1A+xYxh@ItrpW$?x{6Taehz6{&Z$Z|@ -4>b4eCGRTs>{nYD-KUUq69QvmY?))028G-jtU8D-veMRS~Wzp?wUn~fPmaRolHNEbVgB7o(JAZfr#R~ ->9&w&5%#MipLqyCEx=1@fr;&YWNd3uuTykmH3UYym-V4vqzChs&xT)7W1Ad}wz_~iK?pE(O1`UrZCw8 -_rh-EM(_d2zrjE47>}lgE#+;VdWyvXMFp+?b6%W{vgt -gpKHF|F=hT-hihwGekYNy-LGlI6U_^=K+MRJ&9oOz$qf=*Z>PG{gPf5k28`o_ZSnIrR>@PPh4h-{VXn -bz-#p)9EBmK@;Q#P!W5?{}CkawRqTnE0Y;nlWqnu-g4bROfoMU;K;3>e`ne9(%V6~n@O$xpE|iiO#`q -PXXm@36*u-50fqMN69OW0b+bk#Y? -wHDv({Mp)K{1DiZnAaTHoHOEPs1yfNlXF~${|V2gKLgTZSFG!i?wy}EL3`&>xdZXRW0Vhx8NhLXH9jr -I2CBy{j@vCyq;NXwa+HKmTs-qXJ&3#2!Jgp*2kWtIy1Vf8_*xx$>?x|d^v(E2m6pe+Pt}3xi|$I4N<$ -pf9!rdO*QHj*Zi~-G+iRMvFF+ixbda=hZ&DTYa&LN`*B!`fOLiWk0hHVSfI=Q%M3YMgG2j8@=OP1uxP -}b;Gf7J%PyI>BI%n3qHB!pmlUvL3@76T;(+NqO!@DcPXkO9Y?NF7Oh3@K;d9kFpqqE+sn)g4EBSu}x? -akfb0d(R&xrS}jV<)Z-f+F{#ij_m9#+H|fERs^RstRH=_HPF79bmO%+~bNa>S*;93>$48is`r)4w>QK -^}fccw}udz<){gUM0I28@-Vilo$P#(H9K(a1;2W#yY#hx#uWn|n&cc(!jssuh#z=b`c9OLBn~PxsD$w -+x3|gdq^|FhySpCcr>>Q($ghRG6`Zx~H^KLXOOwAP*z)SwSrylQQDhhcsqdfzgH?Cs#k{YPFJ|IXSJy -KLAs9VN{{~GB1FCcNK0@23j`2C_XJTwBwmac&w6XP-ItP^>@^P;^Kg;^tUwkpt|DKIL`@&z7RG0=i?m -hS3UXO*|$$fP*w)!P6KNG@oG3}OM>h@6SmZ`4CUk+nbiRYkHOLLn@sj$_c#e&lS&%>u}6M>V_vPU@k0 -)URN%jjl+g(CoQhQ-Lj5=hS3VscSd|b&6v+(Z00kYug)PM*rEpg -&$x$Q;byzP{m0?<_JH}JDGZsK!SLaF+WFMT$&=)HOBI -G4XLmuh7#Fa5=#80ztz_X|Q-3;t^GqB&yza40+JcZ(a-biVC(gg+TYclnSQ3_GtEcS0C$LyDv6#K5c5Y>7sEK$U -*F%SbNK?$}VeF;#Qp;7pr^$r9a;u>KYgHWc#x_uVMj@#Sr*?1V6p+^ZC3SLbT@j^bc?`dHX1Pw39r)> -p?=CW!2}>1Ftd#)@mC$OseXFiyeHUHqGWny?XcFd!b~rf5K41!+Z^lDf$Xut265r_?W2@?o_$ppSMEb -<`qRK&JOeO1MXRvISev_R22$M5RiO9sg#q0Z+8E^>!z4<2*Eg83t_>7{3`idS$#;pS5_08I)Ya}#KH;p51o-X$bs--r -;rRq7PG4<&q|PMCjE%x2Tce*SEjJXccBbG1mHf3^{Vo=5aN@QEORy0^V~_n&X81OAuy>HjcMJ^yT&e7 -;fNKY!?Sil2APn*KLXO9KQH000080LuVbTZ>H6N!|ef00;#D0384T0B~t=FJE?LZe(wAFLG&PXfI=LZ -gX^UVQFqIaCxniO>dkq5QgvkiV-J}sMTY%(pHMLQQAgHt7tD#R3R@6iz9y^+qBt#UxN)88nu`Bgz-D? -JTr`Gv=U%19g}BZqEp(!_s%jUDKFso^y8*He7QN?eSNr1l4j77m%h7L)RX#~-`=xVCWAsCHND13aqqV -RrA-pntX+2rxaH$2C5_^9+`(^)>A`cE0v=J;sOgF0@zUY6e?-%Bmm#S*HD(O5nlhQigt^a`rs;eRwng -~OYTJTpfP$Xg!78^uvtKAd_TuDey>rne>V85`=lV1{>BAIVg#2d%MzvSgCkvdj$frigHMkgjpl!l&EK1M%$8e!|`v= -kVbZJSd4v-Gd*As@q52YmqNv@7-fE&8oHxL*NKkQ3Zqp#~J1!o!VrUcVR}-F?OU(uq7P0=e1ftCsP)3{0-T@I?BqrWVUFQzqO*E{W$7i2KA|k>_tfj29$yAi!iBuwtD*|Esg%TE1U -rql^8o#!6-1XzqX7v{)?d5KBH67G)=Av{^O%F4Xj&*fc1Q?x^jxjQx^7P)h>@6aWAK2ms3fSz8R!V|T ->`004d!0015U003}la4%nWWo~3|axZdeV`wj9Z*Fv9X>Mh5b1ras-B{m`+cpe-&tIV#7+|kHx9MZh3@ -f@~>(*X};x=Fi0z;R`IkmE7Sn~bou>XA|CEIfBq+Pd%VSYJVB0rKJNr}0l(It~sG*)Y6m@I4ESoTrYx -zg1^7`rG?!amfpI&|Cdg28k3`!6r^7eBsy@$;+8A5vO+1wi~-iQvoX#(lk(#=5t+wUxT!%6(lu)Z#bp -g?_DI$Gz%X(J0OL=^|MyWHC0kA`x;e+v&<^CC1L-^TKVU#|zHB8P=Z -X9aKt#8q9}UK6wuF1d+lWIZshKUH6#hftzp2T1zf|dX7)xf&uRwB3qsk8g5|laq|NhzV}Djo5FjSmcW;N||$82VrAxxVIJ-Q!>&lbQA0WEi>_UvO`WG{h05;qtyts5w>lr(5rZ -vvgKeoFTv27mHa;LF)~<_Wm}R<{VIAimN@4K4lKusBILkVJc!*XlAPqkE4c%xSD=H2w#KK1s7 -JII+f3pwB8o8nBf$Tv`7?<=$uC@{oeSq!4c+>#yGH|T=sS=uds-y$c)x-_)-^R*7z|`t_P4Q&k~=|YA -#XF7XHjSz*bo-3*)1WuRR@`l@TaT9d&Gh>xfIY-6;XQ->Zi?o&Tq>FlYFI|r$I6}_7Q?aJwF6fG^t -~uJ6nrn>iyLoejA_-e$D$qFzM{dmqE#wlAQ3h4aATvaL!lYkDU8kcJe{=PO9Ym8jjqWarGyMZu(CGN9 -NQ$3fbXykPyGt%_GMp;Y~5AMI9KQsaypD&YPhMxhlsaLTkW_NF=Z_6LtA$*DXaN@ebuzVoF~+g8=mfe -;z$Ez%Q{ZJ(9IKi;Cqf7#ZHl3^uOI=x39RYV5oTfm_0oML8WsZ;xsv~?*7Bvd|_PFS@U7#bbzLPh@xs -s9E`+h^qEEr76x@Rg5HEG;N^0vh7Vn7lbt)=>1YNogcU%R1*OwYFHlPZ1QY-O00;of09jiiy)2={2LJ$982|tq0001RX>c!Jc4cm4Z*nhkX=7+FV{dMBa&K%daCx;^?~> -cL5&y2I*fJkXIxOeTX~wNGPO8(g=|fu} -uJr-!Aqq7LHaa7ZPN&Vx5KHE0EhXN|q3r`Pol5KIvci{kMbB8yEUCj_EvRE?h=|bYUQwFympA;rKR*c -;dqAD4}u3;CUHau2>=q#6C}?;BlXaxGNvbwkZy#hOpC#7InVe4&cwA1w6@k%qnPV=oTO2RH6i#3lo-v -OiWm|juCM>n+$G9e>|E^#zXw~(w}yXl<*4fH2g~!;48Rbajb`bG(#Hvnx!;gdhYyOPyWb+i~`&$>D0h%OljCf6eA+;>Vbj8L;lZ!{hDX8ul~j~ajksG^idiR{9NCkz%`R3;%xpy6 -1ZPn}LN;ZOU$n*-dJU9jH2r$u);;NC9lVm>)x!Z{o)SA+xg0!=X -owHW{Z=0B14KDrTJDPYH -Vsd*^&_hZEU3Fn$wtdvYPrc(?FURjj?VB%-Z$`&?*PC3n@9!Vqf6?o|Jq>Vq9N>EVqfcgI^yA^}b^AH -KKL7Nl$u%ausZRzYG8~Njb~)ajl%w3gabdgLzD3^Oeca&RO>R8(XXM>rgoYrq!ECq_LPj!X11W^)6LhOvi3pYPg*Hpo~ -T5YRTy?Y0K|Gm!WJ8k4DB{_g9rgzmXWtA#^wxV<$z)~4yXc|eAfSJkaK^s>g3i*opM(_7pUI`Y -~+Ng;I9~~i;PLHw!AHte=3*D^@%?59^P+7Mk~Sr`1}_ZRQHu}Ug%)Sr1243pSHN%)*`XarAURZszFzy -zyRIU_vqt9EDpgwn6F^&)kh5k-@{oE(KVu?2)d1kbe`E8=Wxy0mvsAINvf^^8s(^%w>QTH+%^ihpCM7 -x%CnzWOywN7z9Bu~>_l=zh}Wy6>h_=G^626!5iHT&6;lP}0#8jHi?=|-Fn3l(D){3mpDkkm^?`7bsTl -+8G)F_h4QFyrHu7=4I5sLr?Ch(x?gPpS2lld$Uu*K~73#Uc&du(1KrQ3X7#@&>@)9P#g2Nx^8D>bV{sjj{fJ8fzVVn)S@TH04MI&!l>F!rSCsoTUh# -@PdZB*qfPhH9btl;`AMjUO?aC|beCk5k#|t4PiZlv%JFDFyfLiwW6NWdglG}Hxh)7 -aLD2=x5FufYM^HW|aQ6Gb{1`>-nBE1M=BXM?Rbp`MZ1t-$A*q!-axdorcvq<)L)1tNk6+ZKfasu7HhAHY1^sJd-{#YwuYSl3N+7_I{9(DVDYs$$a; -UX48T`=5P6w1rjc -38vV0jiihiF%a)c>D&SC!HMABwfb26Pu7*~2gl0xi=r -7m8F#Dv3MnzweOLjUpvGX&>5}=lc$ME<8M>_EMBcA1lfWM9WG@jXuy^FBW?vNR>V^(h6?PkHzA{$4~G -7_^nJu|qd*9U@3s9k8Nb$sDM2d5~gs1BChV{S4W; -_P%>hf*fHJS$~@<^8oq^_Gpy8$*+fbqpJ^}oQbSSbNMrGVuxd6XI;%xKS&3TA17t{8i4(Qa&s45|uCxzNLNZR|TAyd -&GXIN>^vK>1#TB`9KrgFj-)DGz3tHNelhfr}$2fzQk()fD#W337sB^%qn*HSuo2h&3&PYGso*tQrPh5 -~ZLbPV1|%K1>Eml1hdt8+V|^QmAoU$AcLz(5Ziht1Ur=MYx^7YiA -gY<%*DiPDhmX>#5?_Ar?0NM>f@b-4|Xk7ipBqIE(6c(AM(d3vrQguX)xTr}JyLaAmOqFqnV-y8 -K43fvB#q7QFOlYR_Xwd~{Mx6$K7x5veq^`fq6(9kw&x*VFIntpBCjCd+}1$m;WwZJ`P+bhn+5S1B~aT`7U)}{ZpRlV>Ms9PVl{+VrqF}zaL(>o;!_kZzeG>rk2TW%ENsi&46Tr&t|k -ICKD@RVT8hZa6HIfwyz*eK2MSjti|y+2%@4`dyX4~H&D!JaDW~cXdA9-%fRl^wuam0_pHtP5LuqOfD0 -r|`rJ$rLxFLF5DX7^FX^waA^)Pxf4&bR4IxJEXq@8%=OyI!o2{>W6QlXbd7r_o*5t*AP7L%=9K#`=xG -ytV~tH+o_}%d;HsjWpL(W&%Z{$k#F`hD;F-xjyLgRO|70oejmx3Olemz$I!r@`MJbe$a;{W`8|mxl?8kx=^~MjOyb^<*7IH4(L6mhd0i)i72|0diaLyX@PECgj@ei1(ut -AsUtONoP`~SQ5A1`0wmRgNJORS7It@7n*EzY)Pa^AZ$-~wAErnl@?uzrcz6B!!qm#J4;e0|1;$<&$6 -IBH{TjKe_qq1pVzHoml-W0!EJK?OT4tH4@>vEfC_@$0!V=#gE}%9%5NT&~||!tp~6a?{*n( -q>F4D%LHO;|OURZ9o%=*_LJ5|9;V$ -yV~7eK4sp#d1T&9c+*OyL3W5=Zf#966df -Fvbfp4PvLWO=&G7pq%*5*J%jyvi0u^c07;nJ6**MNmYWNounW(&vzD!u%;+ha^sy%P3in4A;77{QKyQ -oSP5fnRyt<>0=knl%c -iW%aSCS}pl}xiD^WHRi4;iMSz)pN-XSUK;tf*2SmfN!4A0Dnrd8U4>CB?issY=zLLNYb>$|`I+|Zjho -11sH+uxV)p|wIuRxjZ%i62Y&Ls+g^y$H71=TqC^MYMh!j_<{TkiCZChlWcVH9m%86xy!lgUB6v0p*SQ -ENG!hZlZki{=?hbxm@$MX2sxPe!EyKL^V$+UjY7si$Y1G#^^WUxQrhmbF+jUmnGLbZcepDmm3%26Vrl -WQma^xJ@KK*5{TswmfWLeQ8Qy;tGq?Bc__GdnFbJqkgWo51^;$nkIaH(4bM<87zfj7ReNhnO|7h<6d~ --1vvZ++Io$^>j!e6ca{%h_II~ad-R*jN_Ytig1JE4EKd0B5tBunRImD~i -^n=2O$Sz^?2pc5IwB9+8NA{ZL&g}e%A=hC*)d`G1(cp9Iowvrg_uUE -xRF0v;19ysY?|-v88V;9%I7l|1n8>(YA!kt9PYmcvDMuioY(i9LT%rMdl%|{P~x84@Q!yIt(BfzYnSf -O2h~XXg4H!I;nDeep`S=Pm|@>Wc*1~G_Ts@WK>f4|r%3PYz-`hyKw6_?Z~2>Vs)pO6DhFcs!y+ka>~+ -@?p2ePI_&?*=uAepmeZ%yZQwv{(JwlNTu3b>)!@Ko?4EQOy1fCr-Aa -8hGE67bFYTo59$b?qhp0vDh?sE8+*Wff}Z-NBDo{+&8Eih%9_AxF7ZQ?RWr?UbD8fS%jl;@S~eN^Pvf -aGzYFp}V!O#PtU|G61yFascpBkd>rv!MmO00&(^>`828Gv4(JfQ&CCl{D8~IgZnIXj89gIBr8Ad1Znyf>NHcCXfBde -&x7J|@5oN1*ooJ4fFIRW^#M7t~0!h}KBHJ1riFk|{~PT?9}<`cKU(6BS7N88kc;{N4*y7 -&)JO9KQH000080LuVbTOvpy)9nBN0D=Ml02%-Q0B~t=FJE?LZe(wAFLG&PXfI`Qa&K~TE^v8`kwH?!F -bqZaIfW*xvf%*iI!t#^X2=N?+XYkCN+qQ|bBsB5n*vo;nL=uG_$NnlQht}4Au1k6#CNZWQeyCq0xjAwz=n8-s?N?}k6y&bml9q -~S`%hXAia4VMb3CAn(cKB$uaPfMmk{dbB8tQvx_z@s{P!`dDc_==8R@cBnAHJ%qK$L_T44Sn7F!{2Y< -L|Q?AJ(U}}O=DS-9T1sr}1xD47Db}s(GO79>g06Vhdx3Syzo@6aWAK2ms3fSzBt}ZxoFJ00 -5H)0015U003}la4%nWWo~3|axZdeV`wjCX>4U*aB^>Wc`k5ybyVAK+b|4$_g4^pvN2B|5TL`ltm}#mZ -65**Lsl57O()tM$&h5E!LWZHCHbbL^^2Y8@bK`ElsI%sTM#N~v^s!2bwchygn#cI3fTO7yeZdTZ`R-L -?r#gYmZt)~3uBWIz_>JG#9!0*XRK6>dN{TCV#m`BeJ6=%V#-wl^aOOJq){!70qgKHNs_AN#=sZR;-gk -Cr%Wj3*zO)aaf5lFNYnI(=A8i!4FMrTwqsyUh8bI{-sZq%4H|2qv9bn^T39q-6$K258F3_@ti}e)Qb= -LTGBemV#fVu!I_OFZpJ+ukZH?7WIlTV>_ex^W;~p2u%5tphY9u)ac)?em?)>PvZyMAq@3~Jyxb4$9%5 -|ZvgY^VhG$PAGMISq@I@ITGp}#p!06^|W)#xSj5Kgk(GVITjpee!%z+CJo>4wG7Y1Q->hMNurL!gt -DQI%^u0No(aOjPNXLP}Zr~OMC6Bu*S=@gdbzcinCL;I*DQ9GyOR!B5&Hut!9bKz@?gt{n7kH<26`7OK -hFV=Bwg%wI>=`XnuGFmYN#e|a-#=8z>ouw)Jt%S^i+#KHd>(CRr -&csv?#-r&GX;1z1hz@8{a%rabqW=`OBB?259V0Z0Z}*YHZ`N-mab&@@RIYn)O9Xd3j&30iP%T0Z>Z=1 -QY-O00;of09jk6yFJ%A1ONcE2><{b0001RX>c!Jc4cm4Z*nhkX=7+FW@&6?b7^{IE^v9ZRc&wEHW2=< -Uvbbr)HXEPNwE$v8Uyjr4gnT4STgJ-2xMA1*@PugBPq9T;QzimQg12xp~EmEp)K*@b8pWb!!Uf`RH>D -!GLS0`8f&Gktagyg0*(BPCWf`rT7AR}a;-|(TU(pue7=)*-)!SlmGisPdq$s(#d2Ps^8Vp62p;wbd2b -`bv=Ex8WCeCF4U{Tt3IwxPO_4#Wj;Lt^A(MHI8Y>$#1|by&4OEC39kmK}SfLgcL4dKE-6h908o~=U1S -+R3_5(Nu;$VzHVi(204NWC21$Zy36xpmMM*1&Bbs}O>WkP3$xOenT3_;K6Peyb`-=&*0SZ(pMG@Y%%|o!u!z$>d#b3mLVsvTm_zN~gX_7k>no(?#UpOkw{o(ULb=wTs58E9zB}2eKttVcSOxC?o&_n?wqR# -&xoBoQhTwfn5F-X#hNbvU2Q0hFXi&p5;3x~+a{*%zOQ!9|I`hXhM!bLQ_i;I&&9#BRt=+aQqwTkhG*o -(VzG$e{-zx{l$_X${-zzzDYI2S@>5r#NO2g;misK~-G`-!#78!dbjT^J%iPB -<+?szI{kxF}Fe}21PLtY58hfGs(jjK5yDC7ZQGQ0fY4cxA0P@<@K9`kvT+QwNG9qkLsQfam+DsZPc0v -E-sZ3q&K^KpBE+3Z%N@Zlc*m~rdyop8KZ#BbPte=!bMifMzt*1LvhyBYLB1q9asG!w3e6Y)zxUwT6E9 -Z5Rf(%q+e-9TN4lu|&($(5qXC;=oZ(G;xNi|N2ebtQwHP+GpW*mSzx@tHr0qCtO=^P{`Q#=In|P6NL> -Wsd^e5a)(tYPm@qR!A-c$>efPu+h2P)h ->@6aWAK2ms3fSzFV)i%mld0043&0018V003}la4%nWWo~3|axZdeV`wjEX=i9`X=iA3WpXZXdA%BIcb -i7`yMM)mbZf8#LJvQ}#;9(#J?Wk{=}EjNdm^+LiIGHtfHEkuYP-LE@15Z>l8|IK+m&quGxzCq&O=foS0Ng#%Ex>*bSk@%}ih?*g67xnzZ?j_hed0w;at$L# -M@K%L5O*BUM?pm0luqfbdcwabvj{NKh;HMzbwuE!D9S4VOFR-S<|CStctUan5=116VTOS)Ar-^M5yyN --vqc&)K?Xlalf0oz0Y#QUHX~jbmbBi!qdriZf|fKj3hw+n<+O-$uNa@=AX0M_4b&Nj9%JNnFr9_)m(f -%On`x!+SGn!T$YYcU;7Od4S%tt1WWy1-ZQL`P;iO5>j+QX(07wZl4!M!=u2K((lv1EYGB8xiVZ -cf(^P(}$ErXw{bSjw!Kipxe#OM#f$r3cc%%#i{op)Zq=mw?hRYHGA(-_SC(}YBEMo=R#E;l*n1Cjv`n -a-xSI3ce46`-(62|=Zl1*6bQk}#mYA={$74<;mF1+t3Vn>@_cS(Wr$F5e~R&7rjtj)rx#5bzK&{4&OI -vQlMgjIN`lXv@la>O`q?A_U`aiy)u1iyaklSrL_klUo1 --@=YAkwSc?=$E?V(-Cktmn?)GPSt;F1*(v8?K{%Uv8S! -JvqT_5U=6~^4WJJk=@mfLrfO#V*JZmb_C -sDhy!WDw(2SOwQ{tj;Tvq_MvkLs;|hxVdpbrX+RUn;RmkfD{##9O!X`Gn9>R!5FDZjiUB{eVKVvK=RV -eeyJ7DkvFIIKzEjgbw4~QhsSZ0K^if8;4GaF(gcls+t!q3Mfw=0yevzDQBKWSy2HD6^VEFME~UvYUbv -vP(&zIHrLXmmn;R02pymOA;o$B;)rK;gjgDo#Zgboq`O>oorv3tXcOM+%@V(4njnJ8`4zBbcJ$;hH_8 -u&&uffgv!{$Ui5xNM@dbZlPkek3j@Xmv--2sh_!TTCO>oI^m&_}!LKCF-J6sdQyL!{HkigXxIh}3zMN -Q*__+XDarKyJ{B2VH*bJ>{|8*}(<#3!~B)p7b?G-GXsOJoe%^bh8vf^cpUjGrE_?CmwVQ_K$e2W+JGc -HWpDBs<(I;1_=w8?HKq|FP&(rF}{;%ntN6Z*g4M9Ew6M&BV)Qw)4H(;#I|M3>(^s)8YkW5`?rWI52gh^=oLnd^T_#XF23Ju&U -Xk{_*j#ank2<@)f?_ix`kd3`{nvf4iIcBSQ;5 -tEMlh{4Z*wmj&#-GtsWAy6kQmgeu+U^eo{bG3Pl+ZG!f+y78fmwOpYbAPG4{x^fVq%hhkW`afg+{Wap -5#(9#Y`b1A#Hz1`?DSm!!Ql2U}P{15(3Mwi#4*Rq@zHtyF8~xdh%(G;Ir!?~XZXuwP~NdJfp%1<`_TO -^#CPeJ*R)Pw(GeoSE%E)!yc~#P>4jb%QxAN|ccw4`D+Oy){iK?5ZuoD=!q=<#K7nDGi_}hAsQN6$X&~ ->8kZ5pvzXi^U$T8hdz$u&yd`oTQANp+P&V3Zs)vr+HH+zW{;WY#w?qM>-)g<`V_Dj<=kJQo=chZ53eY -vn=`LG2pu0hFbDI_h)RvAK{}>+p_><)L9W3nvsE&q5eZm%j3BkZ|G{FjqF$Q!vCuZ@IK;`>Z5Zqf@6A -CQ=Cp65nj&%L)?T*aoDq|RFfY2Mw#g0Q?QOQZxC)e&-5r?iV(@;@?hHAmwjtHq)DN)WaG@k?bb8$_bH -s_LPLC8&@0X&S!Qk2O9z@blg&dVc(1+$yiU~8J9rkhH}aKnK_gs$gM(|_PNz*=X#9e -Fy3obtT0@O8IZF(xx5giA~G#5r(j@Lk+Nw}E=#izq{*dX6so1qcvf*SBmq0^(-&JgbwHGQL3j(sIqeY --jYxinr5{=Z7EB{Ai_19t*dz_}9FHsE%Q8*baKgM8xrK@M#RJNZOl-0sTjoizGAy=l*P0~duU1*7~CuX#s4Prqb82{ObdXgL0%=%QT?n)_Kh`3ELEcP!*Vp-q^_j -`OI7ZLdFnxzvktco?H2>UgU5YgjF>lzu+3~Gn)fM1AOxMQO_!AzYq<?qLssPgTgDWa|R45ko -tes)G4soc`NM`x>FF%v_VDGuNY4Dvv*%YBxxKYZ#?Fh>QOFY%&`fF(cD?-Z*HtVuZIXJ^xo2V2kK->iQEV5BxyxcWKUT=cjRbbJnW_v%ap;q23QOK_Mq%;5bC%G5HfTu`N-gT~qi~NE?Whi$v?g&} -PZ(!fd!|NST0fgl!r$*ZxV9vaKHOizeS=P?-@@OeKs$hSKrjvh#2yD_B&$aQqH+CtX!1000RB1*!)-J -;R%1U}aX3P6J$y}I#Obs!`tup}e45s~3Cc_F^z7W|_Ug#CNsDyqCUl$u-14`Ucv*y$cIVtMjkeJyAP{ -N4=osy@3wZ8f&1+kxKf16kPQ5NHM5ZY&7hZ>2X7}8*&b!?n!4T&3ggAK~;(1pMXBWp+ei-I2<_S~d(W -lEh1rr}KUBQ1xp?D-cju%k|Bkc#R33~ybM%g8QUvu@6aWAK2ms3fSzF*uyoM?T005^G000;O0 -03}la4%nWWo~3|axZdeV`wjGb8l`gaCz-mOK;;g5Wf3YOm&h0SwM4ip`u1M#kL1;&}2`+LZBte=0+w3 -l8)mp_P=+A)Qf&(i%r_o<`B!A;cz(f%{L=gT5X|Wc2kQLh;5^^g=dJEGMws)%9p$GMN --LGsY>-CxCxZ6fHn=S#nKIz( -WB(ZuPKK67PzJ6TqFH5>V~6(a3VQUEy_91;vJGR-n%Ytm-VR4k^6j72DO;Enf|1^wEHLrBoeQy^-!}y -$WU^v~#a_d@yAmMUy}BgE?dB4rSnEQd#no!HOACP-1y#^mkU9kyS3uPV=0ZLWpxdnu0YP5DREo&b-ID*3iGvt_l7*DDCK!PoWtuX2`Xw%b&(uZ&m7KH4`gtaoEE_x$Y&H%?=s#W0UHCLB=e(Q3>{q8hFcc9HDmfyKDY(;+yfRPDaR49MlpFzya*+9tP -((2$hq6Anp?nrw$suY}danSMm*w}pd6ml}xCEous~v*M_H%MK&zZ|R=SI#aa&?bs2Z;Ps -=&C{-8QW0$|7OZtgzFK}|en9wqYgnGS?*w8WChH1DKtQLR2-2MXizUd%zjQckphE7Z7EeE2Y*4}a``g -Z2uVleV$@%rH?poZ5BNWI0sxZP=aL;>zK1(tk&N+fo}VJ)7xJbKrlX{&pt{(|$Cu?*8XZ4uTF@@Zwmi=v -%OJ}0EfX@&+wR!Dq%*y4m^OU8W)8-m>?9Q4PAkR6xZ5JcZk;5?zFKJEDlLm{N3ZGb9?^7UC1quqvv&I -c=ZJo~h=v3s53nc1S9H6$e?MB^@Vd#IGSzD(935$xIodR~6tY$RH)qbo$Z|BB`La3Z#og`V`sQvVp@x -!)agBM;hNf*AhO|w@hAehsxr6@I*jsebV3b>@-3PgC5TZ`*IHnOznj^)ON~Z*d6e-sz3$aB0q%Qshrdbqnql0bxdV}yuT}D#PSve{*s=~|L=trx#KV|AR= -Y-|X_2+^&7H-@J0uGq**BML(w@V!m<0|^-ryhh2nS?L>89(W-zrYu_>1#9^{n3B^ENndcc^X8swMbkO -_ns#m_d2uwd5Tk=s;r}gHm3goP)h>@6aWAK2ms3fSz9>4SjdY6006WM000{R003}la4%nWWo~3|axZd -eV`wjGb#QrfWpXZXdEHiDPvbZcf6u3|a?&o7g$Wp;U+}uAfEr8`OSqO`rGC8uOT@R+mM{`M2F;3gJO?4gyf1>c)o+bY7 -hx{9?ynqu}AQi8RI1{Bl-j)l`2v^Sw)EymCT`aciJ3pQKGT#c~&oK1(u+x({Y>%mIVmi3fwhE$HW*5E -?7?W3X0Cl+3ft}>(zoTX1DXp>ys -71wZBl`Mf6$w%@DRF=!l;d9JtEst_8;BWaC8RW*&j8zW1=E+DdZ? -@0%TP&&0NBr=89S0`-=Ow66kuP<*#mx1WX%c}VC!|_&#E2vtg<+GGE^> -2n8bnA9k5aB9YtAU8kP%9`;F{8)f;+i|3zRJ -7m)Jk}!3fd&JmkZ_UYbqaBZ)_e5*5E6LaPd_+VZ^f9U^t#YX=N-_3e{L=u4iT|!q{jg8cespyF4^Yz -a2WScZ87}P%ne2`=yiOmoon;jY`Ln2h6O2(t7+~fT2>eezru`Eu$| -HZ=Q;#h**8}3z5|@VPx$VZ<*v%fdh{lxNKUgj+OS~#=g^u#|dQ(A{5U)v -H3id@3eV1GWU1gi=y6Lg!6%%FV8Dhp!3KR3k*>l*mySizVlvjwm?T`B7<2L}pzM0bf9PD2jT~qwrko( -?f(VdMmdH>7_jIk!5Mop`2Ulscfyu8`!l@yV?S^NBJkE|lRv=Y>jydHc*M({qS#wiXUC^f#~QUuPm)a -Mdc*-R<9Ga-H<7tHK8(|IIeh^@^ERC9D2j;t^ktW*Njd73O#4dITB$6B`+BD`_~u+l8gA&ok@64$w_X -Q?@6aWAK2ms3fSzCG%*B|Z;006W$ -000^Q003}la4%nWWo~3|axZdeV`wjIVR>(LbS`jt)f?@O+s5&~o`U6{k_kij(tiR}#Bq!yh~pY(r$8Y -H1dm)vj46^ODc>oFdxQ2(`c!?A&dl!Y7b)F2P2E;M?nLhF?9A+Z?`qq(hakTxg6hz<{TMv|`b7fI!`S -D=co5^hEf-sasYcO{ZCeivSq*hn2szkxqR+>+*JGnN?8FFD16l%-@`>6q_W -DzeUIcWx*R4vm(4_PmUw*H`)P4l|iHCkZ!pazPz0XV@!FXUswUnD_85?qU>lsvpUb%LL-#c-@gg=$~z -_JDa0ylR<=uK&)wR)AgJh?+s(-~$izumLAMLQiU_7ds%t;Ke6B)L9E6K}pb8#XiFU1v7}UideX&>+bv_KitNhd+M=7qz>~}1Yg<<#{xN6 -GkLscI0I;aUe=m|cB;RV4-_Lt1Ke1#xO(xx^2QCw{{u($;LG4FyE>ESw*goKI0u~|r(WSao-G3-KhmhB(Ir3^hH>!4 -A3+9p@HJ9(c@&YoxL&b#l5basxZ4I-vi`Oi3B2G4`$c(?(N)oy* -1HV#EZBuS-ty(q@K*&XY==i#wc0!wX9w1!-_w^~W~i_sK%1N`a`CL=as?c3YIRT$J2cpEN0N;SZ`j(r -n^YfA@GK(qPhKmKN+`Y-MJT@suU=`s`J7kh(HvVeY8I?(}#?Q|_tlbKZGe%kGJ5Qy%Xew9fIC{{U3pr -H~5g-9-jTk%WPMT|RfNr_#hlkQGoQ7c7v!RLw);?zRU{KgIvMfB$@8&s>)i?%)2yMa?>Pvg5!QzFd*l(81MWfOTEWhT1FVxKet7T5f5pU8kyVRdMK+h0=Y8CM~zIL|9_F+axPXUy;wz#j^}mw8=Vs@av)>h?I)OoER~RITF -{hD9}+GqnbxR9cUpqVC&8b3uf9iG2+eCL@HZda^tCL1}M4;Vbj#Kypgv#R@ujVh)3fGmVE)9o6mDY!W}5Kq97AA3e6`**IBpCQ?ve#N=4LDapSTD?S -Zm(%dyKWL8wNpy_yaIdvL7Zgka~r-k8&*MGMcZ+&dQxHRQ?GbBV^dgZ7^4cYcO1?rH1(I=q;=T0w@e4 -{ek@=$*GtW}s;Na(I{ohQktrjs$?fgj4sTCwhci6#JRA>)yyq^NX`GS>2)(-j)CJ=HsE(MaTiC@5XU? -dS2SS(b`l^h&yu<^YRrCzPL=wQ?XR$v^+>J{wEKK^4_;LGuR-2I%0g3@qxRa(}SXPJ7HJx}$dOfHEZ$JZ&ME_gd$Yt~f}XRs^tqqekY0Z&*@6r6Oh7v-kqH2+#w)srDhs>T;EI0n0GVplfe8-Bn=hW -yUuF6OZmTYXH|M@wm|`daoJC -m>kmwAIHsUl$aVK4qd|^oCriij3z*-EnZQer -Jkt#l7LrAS^VW6wF#4)P`m7$i9Dvy(Wlnub32$qyakpbrY-ioA%*7OWE0Q$u^d~ELBygNutRxsqW|$H -_Gk898Iy~6p+wuriaKBuQAub1AC{I}#1hz;7w^Gd>0FEIQIQw>poOXZ|F8g6;Im|SY>da4m&vujJ~%; -9pAy@N6_5&ajBZI#ihAm)tfjgVW~hy_+m7zm%v13bO}1MmmxO%2XO7i2cr?51ck{C-#(8m431*sFGZE -=rH_@=&i)}B4eKr)3^{}4mI~1Y1GNF^s%q=))wu8{&VYyC0E}!tKTUe1=Dc09COmc7wYAdvg(Kc(ZE6 -Q}nZDyDH7auK3r*2qVD{Q=N#iN`ATzb`rfopE7!=t9Uv!a>vt$2FCMRX{u( -K9H-L^I%<=)y5(d{(DyhJYsLvKLE&D2wOzo2ig$r&yrB2_5*^ec40)UStww#(g*?FlxCo<@BH$bm|J2 -qnj^!fkv7KWmTXX%}4Ck4ZOYb5`diQI1cirssSyJD@~>#4=XVlRb;92yk}HoyMm3)#&^eIPgk1&bDDP -@h|P?(ogRYkUiGR3F+56-Sxn;3ocX$^!1#PBZox+i#k#YBbJ!xFGm7IGqu~?a$$INAu3S6NJz3rD;5T -kV#I%j4C9JM^vm8h?hvt4ou1>P;1KXXs(C2?}vmzg@Q&O^O;&H=_5SDYsY*I5VBWy{w{V2-Fl~oeF7p -IlErW;*ZvVO4Wpz_s_V|)3A&x}6ehTMf_w|q;v2snC96#rdM@49kDm7c&a -)Y%`;+VRF04ISEf7jZ*7qJ$gwB+rg~X9A*zoy0q93npVmlja%sobYUTeo9zQO)i#9DO3S8)!Rx>kH_a -<9htT3m>x!b!uz%tQn0gDW}A;|cKG>tn{&m;cBlMj;j#E=E_FG5GP;5ipDKw|q1k)xXvFIcDQQLqnKw -JW<`lbuHIu*nfEnsXAod_4aCAaiHxre9$m=6$@l0gQWZue&fZ|&{ox7&xEZg47HTnM3Z17doN47+0VZ -1KotQPyunm=d4I;_u)4xHy5hcx1r76TrX&{KSj&Ur`diC7Lo?%ZTud3sL=M`qd?D5hXRKgYiU{B-9tWn8TuIE)#fS -E8{2L|tG+z%voMY=AEo)rAKK)+ -pfIXlt-j#$)bhj92^`h>XI>1T+mfV*;=YsD!(R@za&XsE475Ll7lAHH3M!1Mmizzs-ri3UBdP?u(!8m -&2qpx5a>&~0vo8BiWf!7fRT*gh98jxH#x~Pp<+wBf>Dt6@`%7bmwZB=|47c-hDtv0gZM&TWGblm17k` -6Hh$?NEnxxx6K0l9>{t&7=_EI-FH5>? -ko;9$eB|56OnUF?-ZN0(IxXoVDcTof7*UgMG_ct%)DuCBHnXw}sf*?B~yJ`u$>Cl7HD$IfxcBh`j4Mo7qj?<-YKt -7C(Skfd=+)6G&!e)Fd9;tgWt#|nL3xV(w9-3zRSmVer9UJa|&4K5h4lH#jB&Uv>IwhQSy9U;$lTNZRE -7Lf0Lt|+`tLV}~h<(k=4?GZ*5n%tg%uoQ-J4+lfCrUKNt6jXr1jHI57CmQIxm$M -I;i7Y|p9}&V{Tp2NDM^jmaj|dp(b$eGV`i`|-Dy1LT8!A#<8+ocW-9vKQgL9KQQmmn_l{++Cq*lGr`g -ph7Q&`)YdN@q)7j(GACl~v6$atGOS36Vh*lkHa3E`B<{W4VlkS%fU8GU5RQ*c&L*; -w#5KBigg{Z}m=_3n)a@2xzE8Hx(E|9wC~xU51+_llUcdyGj?;VhumH*fC{tZoPD+DpCM8~4BDE7HPF9AfA4l08aFdwEqF*S{-+9( -n~Az*)ylvToGtoj-(Tjf%%w3WJq;)G&b4wf>?iOgJ~JTJMeSOq(kc;HrtbOHNtx9O>v6E<)|M^OzjFL -eqKCT4h#`{5OOiF%(8FH~o2PeAeDjY_p!FRp}9_N{lgR_)Zju~>~!SIy9Vf^cq79mijX(IJx$WEwyjB4GJ{S4Vs(5tqMkiOw+3}2QG__&sVc*3VQh#(tG6r0NGfJJzAJn|VF -MaWp7bbLB$D>xjF5^n_drn}$Q&w`L$;NhfGS7MPd`>ABqWHR%)k8leLYvFy+0qJX(uzT(~k2Yi6NN(2 -=NCD%aao7@t$Y|QIcsL__qi&GYY7wwatY04Fql`TbJ1Sb`!LzFKgH+{nqAhHWL8Do`-$KEJ+n|pz%n< -gLAP}U9JRMG)nNoK_nS(QlMCW1CC%z<1Q6IM?*+jDFu_Z~k$1MqPCaxNVJbzq|P9*eyqe#QH*Fj)GK^ -cXxLp9MzC%il%#0Iv}LTm*0HH7^QHY6R}lNkm;OIO#5R8dw*?3l3TUgE`2FD -K6Y&OJ`>DTLc(UO%2Q3!VU#GGDm|sUBNgsb-2ML+1InfxzGO-aWJ~7IwtQsIOmXfs;81aNcg&)D6n+R -Bw&3OeKuiH;%-O!Y$i^OhaDqvf>X;G9R}xC6R<5b?6)}7gDt#a$dY$4QYTFhSSG2G30QG_tCd>%A=tVEeqAd`qU<3{*Z0C~2 -t7+FN`+?6^eny1znqt~zL~(68d|-4U>Jr@2XTR#vR>gG5U}~IOc#JP!1QKd)ZSy?JI}_8B -j3W9ny@}hT<;Nsk|bUYZ<7UQ0g&`2dTHryv^mQk3HW8QXkB{Xc&%Mr(q;mq=Y*R`ZGkBs@;gxi}+DHoDNVYk%b%GcaG5>Q5<0dsFf_? -HK}R5Fg5srUe2K83)>Q{_vQ1k#>2ATT<@vKAJ7AdFs|-F4ogzOWjZ1<>c{yE+k-+EX(9{}UAj<}Rn=5 -%ji{nigjX*a)-kxm=X!}EKrl*W#{>+YLDcyFEZaC`NC_8|F~=*iL22rQ0CprUy6P(uHgwKy?HB>@{QS -9`)TJzM9f{PH$cBN~W$l81%^MhIp7S5;U#Ni=?yT)AG{r5m5~_<^I!yb-VK!dcVx(OBtFJ_pw`7Z{1(&h_FVExyP#L_iU4WvsFIQNqsG?IzZu3gmP_IOn9%-Ry14D -olJ#gJ*;s3T9T~Q!*1vfHcr~osDOjJIU?HLLY6NG0Rla;%jYx(l?6O;n}NkNi33*iKTsWO~a(yfjmUH4Sf3aDSFYmZMOwB`P197?$J-4aw*H%GQ^aZ83`6(+vJWmi)8Hb2sE#&OLCq5}&q!D@<-?pQ-RogQO58w^G=cz+M;;1wY&$T3X-XW@U28Aq1Q -dQ$JVu5!Rt(WQ^+hF=QS9VyUi<@f$Fc5(4a3))kxH{Lh)oX!5!FV< -+YvrP#K%gec7lg(bt^W2?7yms);L38-d@C=n|zYu5EKHBdI_QAp5CG@`BfED15<3EPrE1yR4hs%~QQJ -NSGXM1U#Ni?N^~;6(_~x_s`1z9u{Q9%^_ -{;eh_wo4bo%{Iiy$5{!+Iu`b^CmiZ|K5Fk`_sAmc=7os_wno>{u2Iv=21TJC|~}~1HOFck^1dSkIHy1 -e46U)ApB|@dp9P32%HW%m{izP&+BN3mJPX$F1t&({rZ`I-ryx;$@aTdK)1){r0FV!KT^yT5 -?Zqf;PMt=AnSp -AN?d_azg9bHF4mZb^#aUeVLJhu2THl8-8H6%g-%K2BiV^CMk4(DTj{3ydK4yvBK(SpAK8Y~Hv_%`Y++ -PdE~Y4WMfNyaCcziY1cJ6i?$!G^{DSIQ5y -@n(daHr_}9cH*NLIpeg&+LGTr&X4j}t3Q%BMZpu_ph$HP(jDWHJ2T)4`1QY-O00;of09jiBjB{=30{{TQ4FCWi0001RX>c!Jc4cm4Z*nhkX=7+FY-x -67Uvgz`WMy(LaCzNVOK;;g5WeeI5IRYLRzQ0cpal%Cy9lsN3OVT|2naR=Z->xz7Yw$q(KlgnmfgmvCeVpcT7tmdt${}!6-QL) --DtGVwH;o70}j7>kEa4YX@#!fH+f6mE??is`Roa3OJmgl{2E%wYN0KDX*hHd%e(HJyF6xgN_IF9t3{_ -m!vA`{RMbK*^fU#9$AfYmxiaX;;ai36auBB6GZ6u3lT)c5XloA9g5g(# -Q@>oiADzE4T(7CG>L!AajV#5!@%Ie1xiKfYmGqh>1J_CFC4wXy?S(}1r2tGlqFS3yxRz0I+sO&m=)i3 -f0p5_#Nfu>X?5}Fo3{L$V%(w3miz>YR_1Ms5+#*SuX_z$TM$eS#b8yW-vioA2^~Tx0|m!T^N#c&>e0j}HsA6##uH}LchZeoQ7>+LUB9b-LTt99QTyO>V)f_dH$h!l=YNz`cp5DzayM>S`+6@jO5+is6MRgN -p%B(nK0CvyApb&Ycc(COa9p`-)kSZ{vTOkr7q|I;%GRtrFNEF=&1VMswqwHt^1gG?ig%`EEnRdbuo+= -R$=a8kvs}O`hXi_gO8wgp3t*4h -8=KP)h>@6aWAK2ms3fSzD%nIWB4r006Ww000{R003}la4%nWWo~3|axZdeV`wjIZ)ay|Zf7oVd9_+?b -K5o&{_bCavXd#*Q7FrCdamYpGd{;j+)Lt&t={!U^~kgcNsRfTfTR^qr}x`$7w;fQH#gIB5{Cp9y9;3V -*~Nlcv8z}^@@lim%FV%=RRu|thBjG2NhUhgyXcjqs`G79ra5I}@~+xk(N)D#LAFbpZJymJvXAg8I-18 -FtyS~;Ds6K*hbG-=@jlCwy8dSZ%r6?slDnLOn5=HbO -s)rP)WbdHn3WoZDYi|HBDo3MR&|b+7wW!@a3RXk8~8N`L|3E1xh*(yZR)$s7n{XbB6Bvr?N -=B<#N3p)N3AB>+rwO&+Tj=KesDbe(Nl1}dvHS%Yq1*eqcg5SLY>1-9HuU@R(MHmu4caM_Ez65?R?Gw^VL~ock%;)MoBtEko`W3m7mab; -EAPq>_F!nxV-GvWUE*>!i(_nlzP~X7u55etkZVG)J{);f!0*?Djw0w`6l94g1y+p}bTN&dRz;N}NX}R -szML#slEfkk}TkV(JCd0d`${n??p%1tU1LXVJNCS44V$ND> -I95g0021sdpkc~c*FC&1fr=nilVKaF)|`2r6n`v%>nwNfE0ylqz#g86#X~HS3-5=a_*}6J_8(RoePRU -w_Y`;TQgjJkbw4EJoRbg$twOK`T3i)QE@uOL5hI?GWCLXAK%*tgY8N)~m{Z5MftmqIl-vN(wmP`#YJh -}B1Bn)cGkh9VT+XkPytY`QCoMH<_nDQHfrP{&yyF_yX8JhaSxT-B5=5v@u4m5U~ -OUhR@(%|H5d^xA$G7!0Y+!VA#CY+G32AHDe(7*4HwJrB7dBVDpgdhpe&tQUig}`e}kPvpu%dOeU>1dC^1$g^K=NMIDdr -;q4kU?h0LS@`K1wBj-eB4DENp65d}NwN3#*2?Uof1LmOUzb;}u5WK6mD`JCQFYAp5o@7AUyHxFAu(BK -1%o%|^I&XVKVDut?{CjfUpf6hJKe9&-<{9T?Uu`r^NY(5*LKtV>h#Q^`FJ`%dn+4m`0uwDZ{EIxfAjO -J<=MN_YkLmp#|tZ*^BHuHPMk|h*JQbbyt`R0Lz67gd~JmD+-gQTIW%x|Q)DwhW&p;m7cwixL1@8-^n3 -HDGu0VECZaOzWX60hmTFwk{LuB1d#S)I0iYK%;m+ -%Ljob(Q{LjDZTTtNAvmykzVi>p>m -q>u6B06!19H3Ty##QwzEU2+UO$#3sXj-dp;LF*A5qbVM{%X$1uw0adqe$Zbcb%C39yyqx%Q>;m@&WO! -Vo}(kw@)jvb9TX~ffOE$by?Vjr`Xh&3|MhWg&nbe8v6udUgoif1id}hPcDJAOL1q{T*4ltihJ7JL0lu -@nz8FLodxmEnz4sh?f`iZizWEDo!ui>+yx<*+&d)29vTe}u+I+h0KyCo7HNF!;mxpmI+u@oL1%EFos` -D|AT&H!W!Z5bAPo+)a^tukg$9ROr^C1(>D{4zrX2fdHaJ8keX);i?hu{D#TuJT!hwJx%pD82AW})#H;^B2Fec`=QEuc~1&(U?Eet+ABzY?z?MXbTL0wbUXg+x8!kV`;?+zW8GDv6()^{Hf`U_z0it< -#!T}5Xey**-H{ezSUzgxt$-Vovw8|d;zn65q*bQtl*S$+OM9!R_)ytQX0DuDR2Q9b@3D|C3ntUiCBhs -<$5h(9?}0JJw!H1{Ygqm{3z?C^(}IOPp8UGA`v^JdUMItavw_o6LyDZ%kAz`&!Ir;TLs^*r%M^K%giP -s&|VP+^L;E*gI>X69W^YW@mm -yKpQsZC=2U>mXW|g)?%j1{mQcPxnW=AR}1Bijy{$*Z@e6KC$1_6T)C`vRcLQ&_1+8c(pRtSKghy_*P! -v*JStDY^!oV#>-Mh<%QY={-ccX+grRv7x%GyszRo|NOyVa8#@ -4vP5TyxSOK{Co#2s?3!p{K1KE-q{(y_>atVG+;4C_@qM+v;o4gx1-5$)Q2#AgNkT8`&m|7&8YZpyD@U -TV`T8F+(%SCAsFwr;_#OS3;pN4;6ihtUT2a -grNcqbWuwVni(h71ZK8w`M}f1>Q+iwO%5j5rMkX-P+D7bq1|OEr2>#X;(!mk$*;Cz>lDP_Rk?YX*|2h -EYCO@XLuVYUhDx$!Q(&nqLCR*=^`@cXuT`M+J?xoA9tXiu -@C;@`X+jz&rR^cUR)*=VyM(^8|*teJVNf>a4&`eaFY)iI?Ju`)iJ9#Sh)y6Myo<=k{xa=36gShTc?zHSFEiTc%0JWDOhDpvIicr7cBMAReWj$P=j^lF)W -?nC8IRi1FRWH -}pXr!X3w!_8>68ik==t4s&Q@pqi@MNtjyMo^X;x`p?wgQnnx%Qi!NR=>kY6 -g+@xuPJu~Qd``9;e>ZIgG7q=k)?;?{K^crHNa2G|D3R|Ogb2rOxY!8K(M{DL@P{HvQN3Pv3gyq%yNw5 -pP9n`SqgOr}SFi4Nh+Yj5p_VGReRQzC_!m$jQ;+A(hk?I)cYu>V3j+CUU?!dC0)|Blty*50A;w@zLN^QhL`~CJ_8|efi<);~5zG -=N3Te*YY_;z-R2&@~_XqoS3!Ezd4Q;z;f3Rn84tJX*3O7Hw+32=Y$kH`DdK0c+J|Z`Pi}*9AjyhdY8a -yy$0&VB1m>S496s>SW<5IQvnYZt?_YJ6*RAcn=t{QM=v^{6l7fF^vC0uN7LhzA1yh*TNCkgCjp~%n>W -Es+y|0?;xPbTMGem5`+LV!y=8q@C8^A=`A4k$3l8x$?j%W*I-S5j{O>dUyr+}TdnW!)>U1OiPWF_Ezm -s$gmQPtTfq&vJf8r0CD(Gm$XAlNImny2R!F0)YVgbLyGY0r`0=PiG@yrmvFCddobyc$6YQin>hnFv39 -1RdAX_1u^Fuf@`K011F+-In}q^`w>ObLEj^NFq?;PG6Ak0y22&?vA@Txs@@r8Mj*^vVx)Vlp+*1jKcP -ZHE?$RhR`Y0(b~+$oF`gYR6Z(X92DRmlfDG;LfpA%A6=k=E%xVl`b(~eqHPgPD{c!gMVoQ%TMlcC#uz -|WsIiZ=t;{7yN;{)t?-NuSQ!SFE&mv$phd+V{k@Y&NWWyb!2ByktYF5^ryl9zuFBJ!^ItCJ7a!gLWG* -JWAsCCgW}Id#I3n>glR)Uv!G8fzO9KQH000080LuVbTi#N$M^6d>00SWa02%-Q0B~t=FJE?LZe(wAFL -G&PXfJJHa%**PE^v9h8QXH(Huha#figQ2GNtLJkJ?oeC-EeiHkYy8?gOoIAQ6&TQzT1JmQ|1cea{6X0 -ZEk8ncd}sEdu8pocjgLR#jOyqE@3tU9N=GjmUT@&l*+BJW+Qt=gqQ$g{G`q2-!6CQZ)d^`^~n>iehr%rUTEP2==|+iKm49SOt*TtkReONDYCvh -Ng*eBT-(;Q1qG|W*x~z2z@JZIU>*^fW%)V~6d3VAkzL+2W6b?nFTd+?$Y3h^?fl}+m&@~T6cD`+XP@8wa(>~Gti9j;_< -9o|wtwwMFJ4^-!K;6GfHIorQfqM~mwvC)U`*8Em^0Bxuu383WuZ%uZbec@ur(r>0+7Z+H+73~b)yIeH -zJ{;Wh9=@$~=8eI0#h98uq}hdfeGA1`5;e&|)iOnr6IZ!3(*7f1-gShO8^9D!y#Jr0Vs@M4SGspo7COX`3bsck}&J+Y^b+@s_|T@v>cq~L<|pG+&nGvI -Zp?m-W@SI+)W^M)|ivo_iRqx~7jF0pASS~4AUR@{}hDiyOW`nq1vM#;IJ%L++;Qq4`7?h)+IVtPGQ*b -*y70$odCLMO%SHX5M;713r*voXo=x9g71jIDtqp?O(@kOhZTqC;>#V;~QXVrU#bC9HL8b0|~jL5 -u37j-iR5h0c-)z(g3qkw*ks-cvdfU?`6Ctd>|i;#oH1P3vjxUsc3LlAOW!sD-M+lax#k$@@Mj&*ksL( -(0{IFt;DRB^IO$uqpLJ(h3KP{caj+ZYREh{O8evyW6N0%#s?bCHnwOpI-x7=goQ;uWX(+&E?U;*y3&5 -I8$l4@7Sf~ha((@PG#-r@Mvar!=>EP;p#kD!vNrl7Riqjf=6l!IULa(VilS@|?c)Ak_A)1bFA#k++S@ -Unnm~wZfM0XK -m_(Qq#z+Fh8xTGwR$4|F_40A%5T(OtZTcMI8U&MUVtm;4i%{pabCy1X7st8!fUTMOlo-EWIz*d7iKoY --CO1B*$G!H?l^IBJE%_RDBBqlQ05_pE@zdPWs(uDo40rAIO28~FEnB&w*-f7G~$W~|Q*j9) -b?`pYyEaw9=?`@+CXT`e5Gd=`qG6#XJ+B@m|2@7jOM161K^@ecxL}o&AYM)8kWqe@{;Np#-J(U9t1AM -poFM)FJ`9GiC2lTEdw(8-$Qv+p)qQe!En&UC|=9q0l?vSUzP}|K?|2-D7vdl4L>$S4hlRNIkt)9_t(U -A@*hbt)d*Io9yAij2zNt?z96w6;Di@Ex*v;*VQ^w;@#GxpK -P%b5%*zy-?00G**Ey73AcR`mq10-Q1aJtpnl#i59QY{|AvnWB7xM(9OpgFO_b601=!;tfnQ1#iSi{j+ -e8lEeL4N>^rWAS$YP%OSOpy_8)D!In858v0C~aHHSl4oqwk4SC&UeP$z1X9Ao3#aVufpO{ZZqu=G7%D -DD}#DyRXcGM9Z)EeN+}Wzk~?~!-`v9uh^!iZqSkbz334FQ~GEiC3HK?uY~z?MU<#h1K;|8|RW*C4_{2_4L`C*6R_aq -Y6~bckPpX`DQ%;$l=TvpZFo80+T67GXi&m#X_v_z5>D=ohM3ukfY;vi -k!4MBc}JNg7Gns&_AK0tu7SfNBpd|6f&0RW}ZZD`!o!`m=#IaT_N92&u6iSNxp-|07U=C$aUF!CO$wj -ime=&i~4xgcRU*kUW`-TFjTN!@vwvnaL8FlPOO)x4I}6RVlTvabQ=!#7T=Az9)CBmtermPqSzUyi2{7KO?;=Gr;V_GCCf-B)=whZ$}TC++tZd*^h4?D)kYw5H -zfyIoI{S_rgo(ooE9m&#b@Eodu@y752#Z1;m79!2Hmbfd1!wsoM-g&RszbsHGoFtbjJec+5PUY|N@W$v>2yVZ!6{GQe@5s=TNGnkZ)NgA@NCyq=G-TfK# -Nu3;bB78hQ@c8ax=Wyp{ba(eC*#qgZeU3&JK0sdbImo{Kh2fh_VFK=wx=#utye(75Ll2+78GiUIJnn9 ->CSz@sS~ghRI&j+C++?|O=EdEsO=fv#K6GwCxAZv&^@NAI7oT`1a-i)$pmK}LkP&VE=~fl}RG0Ya -!l&1GJU)3j4aKe6PWq3`zDE?jfLSq`81yLSC5(hb;M3W -~&%b`f(YK#K<65QD4=T?~@$a(E)A$q@ywHMJ$t8Z$Gf?qtY5-?07=ACEzJWh+)`02eQ-Le=xquf0pdb -vt#is_$Ch`^*Zbh7D#qH##X(~NFJ>6_J(I(6D)qJ^L$^t)2O>wr1PZ4a2J>VUrDQK;?;!R#IZF=L4tk -IiK%q<-;35ra`jFI1+eSUv_@g4M=25o9aS4`?vRz(R+<0;g(I>SbYzlK*e%<{uvQNv&Q82AcbnCP<3Q -o1>S72o5cg#+b<((AnO@Ln`cf!m4y1CH$06y4wcu?>4qM4AzxC3U&+>Gs7;)BV>Rz8pcGeS{81{{c`- -0|XQR000O8%K%wh|Ew4UYy|)SxfB2Z8vpui*3~0jiO{dUdeC%er7_ImRPJUl(YN+ZNxqq(IlWBlXQ;FQ4D;1s!_a7U(_s?L -hj9BpPdBVF+Ub<7RWu1~F7a>m=YMBYh2GDP%^gU0Go+l7u#~0LuiyPoq>r&xe*+khNYG_Cd?U3-r!G- -#qR?u%v1+=jaW={J=;@dXafua?K%4)AwK^!jae!)X|=HhYsPsMjigBo*!r}b{u82N7W89T!H;~ju_}q -8L5#Nkk$#P&~X1Nsl)L^oQ;qsD=hrq;2^OqQX%yT9z2b$oEVbMfMFi1cE|5s#Q% -I#P_(*IBA<4y*QWf+AN00Ya&ohn}#syYIIM;10lULjWUzg#)&F_jEDp4k%V2?voFk2A5WYB{lGfh(m# -loA`Lv2PUPwQOYj+ni0kE=>xK2;tJR#8=zN9F5{LLs$;mwn`3smY(S?_<4mmGEsob;SQaIKtrS$kcY{ -OFqhqAN6Zsv|EAKNhu;5j0-0sqGA^E~%h*yU>P7&IfkY!u(_Aq$NttWR!x-^l -;?&%(Mir{3Hf%FPKfJD6>2ME0tdJ!!`e=D+23*uv_sy)hzUWswqsw<1%~k>j#;DF5ptttfMU+aA -YN3I#TEd{3p>u4$pSyZ0Axs1PCyX6le~jS$CQ(io_Zi&To1x?xRiTKbk(S0V7Qoonl<0kDKt$VX<>mO -nqkIQRYuK-c7(N^WM+=_GR

62jP?CSQyYza=jLNAvi+fNESbb+%d+bFlr&%%2H7c3uxIY+K+WG(b -&38%B@a7-Y}J^OIjue>I5S)d)=?ppX0tM72jNdyt*Y{1!FXaaFZ3I!LS@KCG{2c -kM$&J<$%o9=+Fq44f{GYN06>Jwo;IvAlJ(}X}g&RKPi)tr7)dL}E^@me+YP?P@DtEh*%0iLq8L%UQqf -$A&hZQvYc)YQ}+#CsSHht#|?q&Mn}r{-ERcI2`dseMCMfdEwZWt2_ -&AOCyC8<|brUyNVmi?k8(`l?nta&XD)DgtytS`+%^z>*H=u{he!#%N -P|47T_b_xHoZ6vVy8vn8Unpb?Xr<5mEt-!4F7wF7m1`+YqUich9Tb0+|JBUs$|=My6d>u0B^chD!BOL -{b_mf_VncEAK(2z;1}L20>5^Ow}RFLE_Pkb2z(ScP7+?l`-=qdS83G6sug`(vjI0VOZs*zF1-$?oy`i`y`P} -%Op8qn_w6<#D}%jqp`P((30@A&u_I*B5(0XTij>bgw{WG@K3%n2AHmuV~Dy56xg&n?tzx#>NQ9D@T{9 -H_iN&qz(FRzj<`7!JRAF#>dS4cV-pt`NZ~2Nhzdwj&_x#~?d7KeqPZ1Dc37WY^X6_txE)Bqu!+T>zuBuEtPbov^MaDRTSy -jqw;~E8)IL{o%1P@!NzILDkw`XxGGDZbv>&qG84 -T5(l~r62}6X-d;Ct3;MsEm?l1QTo|4#y1?>3{&su+wi9ZU^{=j8zl1u9%QhyO}ZW)*mZ!&DL+bw0x^S -~Q3FvMBf=bgah`IMold5mL8u(XE6OD=#*r@3|eE0eAbQ`>eu#(|@bm};xKCy)Ca$$Z1=LK8EGqvO%*s -doVe_>6vHBR#n26&0BHNCS(FS>YtQ?B7q@FOs_YXY_?k>s!FPU2~0l+bYNbt=670ADG%nG114FTEOcu -?FZWyq4h~XUkx*sMz15$-dTEb8ZEJ;*+4D7r*+ReT>B%yvVIcT -LFGGnrt%m5q|{W2Ck*C|}GA5!4Z5ie3w-ZS`(VlOP|9gB$EBT;}XLgbcAo=eo*!SHvEQ(G! -JXEWSztst!s)Q7M96k%JFJoVRk>MfnUi7?AiB2(R(2GzUmDFYtb;$c0J^I)F&igZA*7o+v^wL7BCmwB -`qBqR`2KB#my4gR4|raIWghN;rbWxnJ_u@1^@9BI=E5y>gDXy|Qe=6qMdRG&Z7aFIm|`WIXT -wOW}|aWt1qTTrfL@@CiUK|S+b=FplY+}`S*HSv~!lSYlw?GOC0>(*2;dpNTO9|cYCC=eexItE;9naNw -;gJ>1J!+t&2LIU}d^G*#5v+ry>znm9(H1_e~(X#k}D{3xSA8^Wnn(0{bB!Fb1^lTpUuZXEd0gfP(dJaNV) -VE?c-VX3x~Afj@o_P}9A4^o#Y=bgN?-IMiw3C}*O*Hl3y;iLXjuI4rx?wPt-0f@UfXxh0O{!{EN&f@i -iPWHZ4O{pHk2v$$Z7@5W^*80AAW2-cJ_X8`qzs54{3wzZ5}=1{a}FVeBoHQES~lr_d@>VdQVWvgVr;a -)W@5~t&()y=M(hVT*jbVJ?c?(_Pgn&%XRZWx6BuL592dza0@mo;s^8=)%x$PYKrQLSEP-Xht>wR#L}a -sL%6|cq2Ihq*D13ZJDp>KBZ>L0RF;A^$O?VSOG`xHqAb$_;A^~LO1f!ifxQVya>Jzm>eTUWQQ|#o-O7 -b#*YK&veDScqozV`~y%+0|XQR000O8%K%wh4#pLNa{&MVJOcm#82|tPaA|NaUv_0~WN -&gWa%p2|FK}UJWpXZXd7YBMZi6rkhVOX_i(Nq4cmSkMYkQw2bvs8X7>Wi7irrPb&7NgXwvZ6$s_n8Ik -Ya!K@86`hrU!K_q3s7_8MeFqy}J`zRdSDfGL@(&hR31R&4TGs2vK&52wqf!Hg$zrA;6s^Np^}5Lbf&* -6buuSyNvM2(~ZIutFQ(-nM+F8)%Q>u&8WEm2WIf_2+u}i5ST>{4vrYKde#sNvb7DZ1=EReI`u4IKVR9 -v5us{n=+uZ+el4=NJdDKH2e1CSx_u)g;1h-k+$hAV>#zc&eVBad>V9%6J^Pnex84Ha4!%AItD0UFppB -zG&@QuC42Q)fKJtDYa`LC2Tg3NZ -G4aOPP3czgqKeEstP)ZoPU($w(|qWveUDiIM3ra{F)%mSb)c!)fN@mzGa4Iq!;i?2tCE_KexOg@v)L= -Je+bX_kp^P)h>@6aWAK2ms3fSz9;&K3}^8002Y{000{R003}la4%nWWo~3|axZdeV`wjMVQgh|bY(7Z -d6ib%ZsRr(efL)k)E6pIl-S+u76ue_H`t~@fp)Q*mm&xV*`_SEx_C)Cv0L=5-_Q^2mvlI!Znl!31`ZR(s0hSD6+_l%v%fZ$-sMAnFyMH4-cU`$^NWb% -$*t}UcV2&7DgITwkK5Eav{{*0)=pPHE>7=tAyFAYOu*p`e)rcoQJS~mi3cAh_C~4Q_#VVdzKIf0eVG2 -n|&9GFZSaV1tWHC2HfHMYvLoJZZzCp;+9#vOp%w>O6-DYxcK?z!{HLyc(903Dnt_aR65vPrW+&cD#?O -B3sJFtAi}?V&<|8{xy}pxGs&IWicvWJ+g2Ny?UO)R!5%=^ImkxFn&op7KR{sYg-VSW5}6O}vHpqE)rz -b7Y3L{W(?&7E4WpwyB{qr?=*XrG9}6IW6PUh@fKnl$9AJwzDJA6I*oKy?^@5!^sPF(RJWO`e1^OGf%O -&R3gjdQi{qzTCJcGVx_^hC~Qom^nKg2Sefp${6i$ui`kp7Ef8#d%e-Ica~LacDjj_3{ -D~do2Xq#sJ_eMvY1?Fs~cDD1c`HqHELPW{} -8UO;Jffh>lEIgfTQQ1BQl8n?(I8gSc^@3-D3;hcMij#BEJbY)Az3ix!h7tTO_Q`Q79QBG8a6*>|& -a1Q(oI{d8z9+IHIZQcS@(FElqb4a_`(B7$jr@r=IhVYYwGUeLESJ4H603^uFq1xP=+5$|Ajt0@*ZP8 -iF$d0Y6maj55jW-U!GvX+JY^ -?0Lj_^_r-7-IhiP)h>@6aWAK2ms3fSzF6kp*Keg005OM000>P003}la4%nWWo~3|axZdeV`wjMVQyt? -E^vA6T3?gfwh@2Vr$E_D}Bj?aWX -?=q+4E-0`)XbD!Hnvv?bcK82mP4an==6n22jz`5Bv&QTKXULqwSQp^(erP_>k+ri`$MH8p&_Q(HC%^t -6xR15Ibt$)q=rgi~};Xj{95vnA;n@5JS#Bjw&Kg`}~m)$678Hi%ne(18CRqI`xg1>A8|5dClxp&RR_F -E6KKfbv>|(t=k^jS=G14X3K;eaD)Ui49kV$a(+`Cx|(a2c)fb{%5q(Hy=Y{rB}l%5Rc>?kx3;ghze}= -AkAK}?4Q{X}MoP8IRxB!QyZ(i5Te+o$aua2^Pw&jvUl06q!dwJ -3vhV&7W*geV3GTDtT{h2Ji3CJ!G(mB8O)Bw?Ht;qy^=Z*3Nl{b;^YQM?$Fh_UZUP9vhTl8aj~4$Ys(V&}?FT7b|7k$e)NP -U1|%*$K=|gz}i<`b-(?*s)^R3!Uh@IrZI@%RF!@58wR*G0#S`(QDmAWmSmx*`BeKy0_JV#Ej -~Fq=tXkP(blk+jgj2#w)|;ez0T66%myXgts^6X^;A1KMeWQ(>1s7&}Iy!Y#Ru7~iN|^`OrW*nlAf_An -PcHudoT;PLR_@jntD5_QH%T!_HKO~a}h%+|op_Nz{xWPjYqw%57dz+nf|E>!4U3^y7zhth*94Ei&`nKb}ymg6)vP_B*NI`{oQUw@x0;Z1P|}lYxoNxGl#*pLvM#yWBs@$9{cy$+&A6|*CfHixAEZnE(Ih!bSTZ| -B)0WqL)tR(pRuHvfrL5LNI?D>ckUq}>UsFElh(hngm@>QW)+tV3`Y~k3eeHn&u;Wg`!06`p!C*W9--` -qycN`3q#Nr{+dLJa{kULNj2sI0gK!#at5W0DI=o3jVHv7^XqbDa5d%lK@!qnLkdsZ4;o2k7e-qzz+-M -k&BblH&El -RYWTJ0iRC18!380sRa;*;mLgLu>!=eEBwOR@Nb?B?nMw|5fweh4h#sa?jZ<1vDZ~)z)di5>@jfO^r&~D#O!*s&(lz5+VMmLQKR6%BJi+*iV=LO_&d@JXuxTh|xfixUSkEcWpKm;3q&bUi@amho -KxGyl{JNW6#F-Dya+PS?E`)OCuz@o~lp$#?z^jyP;XO2>JRVO*spLa8|&cRP&u5IQ`26a<#$20g)ddi -cR-lU*3GSH+86ZLc}v3lFbh_8(KkYFX07own&nAe{4%9MleUbxEatw>b&nQ^H*Q77hk|;#40$?{nmJZ -nNuS2Gc`f>-?H}*0U}=ifx#XIqq`0FrIohXGbTns3B$x2wL6uxUL|%u4QzBxD}(YI>iTn5%i-)9_*$IL1P>%}hq0j8qPM8YTNSYjC}*P~EhL<_SH-wL&WYBoleiY -i`E>|M75BH|K3dne&K3bPA=~J4c?^b4xM|svgHaX4P9w9#w{b3Fm)~4*YYhSE3<$W3Fm!`(Im5~In%$ -d;v?##jCqymk!;(@9l`x==jd#4GZdU_9qHrjKXbQ8dM}D -{CA~GUKc~L8b)LRMcKiC+}Ka&x;u}PeYBhle4F{Zt57}At^QKm?=c4ZyC$7sSW^Kb(Ss5LnWHBOjeBj -o^^e7b*NSlpTU;D@?LSi>qAGq!W$6BG^ECZCLflbefa6kU(a*BT6z|;qG;?YE}+&PttDRl- -ROLQYv+|*`hf!oLx#Bazb*nnr!WVzwXAX@!&ipS(VC}F0VL1184w^exU)D>rK=4V!kb_zO1`ZDSt1Ts -wx(J+1R9I-sg)d@4CVyL4gfG>7R|sXp2o-m2>^G+jXN=+pI<2w)sw|RQe&)S#vt~Nwcl{qMeApbw&GK -x#;EF4`l~$%VL#ptA0tz%i^EgLe|4@6h7J$u_}vd3ICQ|W>RH2)e&8$?06nv(Kk)i+1#pJs0RDp#y0h -1zS)%Z<%_m?zk3ZC1_Tl{Oga+qy#C)8PqQa~e){CE&z^rjp`^$4PQE?sX^N`T&~SCj -Y*aI|G;jOgvB4UTDE$zQLqXuT-(F9q~^yU`{8Rs;BCG#*LRH|1uUy({Wv(@xuRaW%zx)jfF$2`}OI%c -27WN29*oolzxvdi%Z=c_%V_$=1zsTNNV$$gc8^Qadpd@Q$x32b+kboG0T^@qSTkdhv`ZeA>26dnUvkv -1#+mb$%x51_IZ*#P@6tRG>2R`NAEC -dMRf>gGnL1z`|7?>L;G6Ac$AJDKfBsm$FYAf^6|}rA@^0G}h9>3Dby4e -n^~HNb=&#{@G%8mvLCR@V9gzd(k}r{|%$WEIlH@v@&iaLm!e;zs+x6vY_hZptHA|cP6U^wUYOeis*;7 -evG8M>l1e`!*rJVV!nZJQmoWeh6xSq!1?l%G_O?D_rk|$RM&gn|zwUEH#8c1}}7I|MR#dXY*0g@BPhvyT-@}6JWrgO9>aS%-}c2h0A3^)2L11gUNJGTYFg}s# -6@-`K$h33nzsnzD6sAnuwcab1shSrA~Rg|5^>qk{&59f+UL+I(ZEm3eqpKBeGZwTzbZ5XCqXflT;Xwh -*`1l$hL%LCnu7sj_2~Llv6s|YR+11CuiI^b2>Fcx_ZHi(SWYC5P$0NoU`oXTND`=CVx{^<|K0&+dwEH -GycT@}pjFdj4@(qMz+pRqwJ6vpdB3_E!-SSQ`LsYvT24vnr1XdSq3kQW?8_?q5+Bno|CT&Y}T$@e3+Po5(t-N69+;G@JQ~M-V|qks{0n@%l -T#DBdZwmC|hpV>s^NEEqjzdz_ghXFtr(J62K-Cpj5NMkcQO^GnkJcT87jh%^pLtww_N)sbB6TRA`^34 -94?$Qx_rCFaV|XOo2yXO3Eq8ndbDTS0sHW0y#Lnj%|&1gvth4UN0-qhlt3pX;~>Q3w?pUl{GL~pvEj) -!>DGN=oSq~1W`;cr{d+)7e9!*NIZsc^LKeky3mw0W>A2Bfx-nfLd`|pmG6pbH!*7rCQ@^F;Ofn`x7C9 -Fw!tZFo9*S5t$KqhA1x0>O*9e%V-*mkH*$7ysv2ftwT9VlcGYJo#@_{8-cnpsFyWBpcB^5#17Ea#kgop#bjkRRO|Qv!9f*Bg)+vE1*nVN?21M -2WYu-N+Fgi-=Hwle$%mpBXvimO&9qFs)0Pz+0_re#us1(!J>6K)xo-|mI81`>yPADaV@H{wtRn2CxfA -7JcV^*iuI-Ak)ON)y6RoI=KeB^C`jFdl&j5;gQBD-y?`O_cVTO$#Yr?vFU*;KiGw+dZ|3}p7qE+QW?z -Rv<%*oy*I`gvmNWAf2eeCbW?zTuY(a^d-G2~fyF)#80gzA4@Q<4e)RrN2WJG~8l%y^IOyn#{eC6eiz< -zH~ev}oeUQiEKGXde#gUsCm{yPu{Z7+3LvY1Hf%hJ!jToa_UvrsITW-ZJM&18TPgo2&tYoZAKF4OmRWf@h8%%SMSs5^x|R^YCyJfK&KWf7!H)5ZQd3n -MA2sgtfP3JqbjPL9mF+2N`tCN&1YKmh$*KGeL&jBctDpNQW|kosJg03e-(*`M4}ZEFhS4cu!(#NY9rV -a3#3xuVS>1*wrfBGRkjBzV^OkthlV2d0rVVt{q$*<|C -xl$MmFu`l-jjoxBUDah)gDlp6%EKwjlTVGu3XxIS}~Sc_twFWyeXy}j7qzjx2I3#0zrm}m4@4HCi7gFu@FYT0)r7Of( -R4BIhmtWD#wD23|F0$>;9^uSGnsVa&N~ -gu=1idyK?{Nq?pu~=csZA2_esT*{FVJjfv{=gx33D*%{3_6G;Zm$p!k$Bh+-Nb?`Rj1zENNL72+I~AR -`X0M5@1e)jtiU$l#U*Sk>WVtf_C{QojS5YnLNBm%O*2*K*wi6U)E)}a97=imYY$E&@31^0Hj`L3z*I< -OS__4O@zmp;NP=g0%7?f@Kn_wr3CGJkNb3D&w&M}Sjf!bzQ3+d@saD(C0YKc86^KP&CS>n&mBxg2_@h -!YRqUE8tRskhy<%iv$hN*XBKqkQTdoHU%lKcI9oyI2{!s$YM-new_-4p+plx5;#uDrrW8ba>g=pOUubbn|`OPH8_0Dk -98ektyd>+PYceq)kZtGIBWjxqrauM$FshvVnbWGIxC*19{N%<}QO_^pI@zXy2Q}=r2 -|_Vj7d$sU_}AlZl(n#p}!wwxN7v>IRkNGI7Rf#RWTxOcan?@ -giGQY?DLC?^Dx%ynjI#sy)Fa>mHTuZMWsOi_#9Gs~!EoNPEpA^lHT1_5WWE^^};*UYZ7X)0zbvE-NVX -sg5$qSWkgL;xZ6gEVDFP;zp06>=OZV&OA|#MGpSyL?34VvHhZKEZ|!DEq{l@eocKCf=lnaMB`ir`_9U -vZjbP_5R)y8$>c5W)y=rMcJ0+byvCrX@}cwPpz>TUzNDb^`NRGw~#RnofXN8#cf8l=w>DcGUcn%kdsd -EjMJ{)f%Tfmo&;NQ;c!GKiSe?O0lP*W2x*q*&JOy$aF|0Qe3jf#NF1lw@vVvFP{trAPQ`iy7LBBkBs4 -NrFPA083+Gk2@L8K|iD()x2AjW!BV{=a6tmjs2RU6`DDfhW!o}k?sLS6v>Qapa&Lj%b -gHw+`L44iQ(d+yu%no!F;qbflr&y*Hkp?r{ilZ9`^IFwf|&BhEt}Px4&?q{#3_3r${1WGBKn+U9$n7c -AKi~Nrz5lR4SVcTDkWkEh{Gto6XcBGhv81fw*%tC^~eO=!ii)L691)n+4lr=H2s-wSF`J!*y-52fcMBLmkHj&(p*{_poxiv+HyZg^=!8-~+&yVOG?l`EXVMbvt2g8r -jM7D7b3o*GE<6oIDPR8+pvISVLFD61ZG`n#d4n_yd9snI04LW&ZMdQBTRMSnfr}ewC#W>t5U(c7hwItJ|kXh*sVs%S1W0 -DXFIaY+{S$wD_nk{RB4B_~R9D-KocVY;sERK*)GU(9!#+oswUnCx3?Xp(M_X -aSSSJ75z{7pk(0mY39lf6gyBfAk|#lwYJ- -@FCZK{f~xsXonjZsm@eo$pnQv5-YbrK0FVoHRfDffP71B@s>B0oa#sIuTUQrgo?T3u1j|l`gLbdN@kM -hYUCESX-8$0w`m(D5@~aL#MK1ADD%DtJClpWBBjK_Y5Tp*{)qw5lkF>m*Vvz)%^^UAtc`50Km9T^r9oFQ9>23>re3 -cs3}bF?Xe2ksvS{2M;g0Tnrc#a^Nf7m$V--dz9IV%&9_WkL>GvE_LtSmw$z+bleh -wy1~tOe2(3qq>%jh#@+sPgaauaOZ7-?6sTS{ZFyNjLnd&74YzqGUk6OjdG-UIY{9Xg$>|0u)HaPg$S} -Te3S@PIn@A?@`U+2#>6||b?Plq`HKg_qDC=B-U-&Br)2CKtHlmTG7t3fo6dP;WKvrcOUTjEZ92Xo8Zs -N>z>ML>oLZPmZNEwZcgKkn85YZ!dc`CK*AGKbepjU?uDLHo0g$e~?YBS9q)0=foL>6+d*lNZg2MY{Iv -%6nZJaIHhHT{wGi!{OtqZaCvBwwqm6iW%>B?8s-1x2>_YPFu^uxo%;+-gazM70 -O!#>=m!rn2dEmPaOD9y9Xw~FcWo@rsaFP-LjS-rz4scn%10iOkwdukK2X#jW@ZFGX+BPDJk?TE9LtkMuKl8k+!*2)CR>_T1)vf($^k$W2&6?ZD-jnUa%) -Sn7KZRQL0Mdk`*ZL&h7CW1XXrBnkj|H1-X?n%THrI -(&xvi^u;I`LN5{ARZmqq@h25obz!%!N&QQFZr0Gzlds|dHjEJ+h9RQQZ{7ut!C7z0>Id82z+`jQSiI_`+bYdVan?4G}JPH3n*eCU)0_w&m1MBZU==y2>E@oj+}e^<0Sdj*ELi)dbxR -ga4sS|Ft9HZFJp!@HM~kt6SG2UhzVIL9lG@2Yt;S^9!Cb+}iOVBh9|VyE~+7~%L?&&{gXI_B_xtyi8P2gpe3e`{{sow2Gu!4b=8+X^bb4|TJHge0uO3+zS}>D!_d<_bik2rGK+3srSn~nBMc -Rc$3YoX#&Azb`P#fFN#BTv>=Vuo3;*0k7s~c0{UYYB#-AL_vHr(F%Oc -$?2Dcn$bv_c# -Xy1X=y4bcI&d0Lop{ZgS)2|T0)NVS>lIe7sj9oM^%mTNf^=){r0Ye*qwBJXL7B+J^s4vSb8v7eKC1mP -1JUrOcWVZfjV>buElA1Tm7>pe)@AaTrVxb@y8**{s4ifQ-mTq&f6~-KQ!GZ@!Ex6-uI`k95!UZ@ox*i -Vi0)Yvu3@~$3D~uEr94rBK8^DI3Ll{BCi!)&F-8X}YJOc@HRXeckFx9K>dkfars(kp{13mI@=do~t=IMYP5J2e4 -~F7LhgI4*O3deokBxu1{lJk0I)*8S=Uyy0lM^le%;W-Y_}K+>_m -CzeZV}aV^g9mQz2MtVmqg8f+?4Mln(d#K0iGRHHJyw3R2*7g?Z5+}N()f&lv_8NSlt1FGk>r%?SkQvG -n%r}k_wC%4{3zuO|>>*Nn}EX|T{4gsr)o#~hT;N)0Bjc7lZ@w3g6eb|2N+#NrF&cvQFe&f$Re(^!E9$ -clySpO|9>;bNd&j%?r`Z-#<<0IX$YJ#jPZ8RPXvX|Q5SK|I~T#Z{5xge2C;tk|S7SUS=Pz+vADdqUPX -JOzGepq&tpTOS_@b+= -2G~r`RUb=5SB?EW3ib%4%0Ro-yeNCRPPE!Do%ADZ*3=+8k&W{Wk- -gX-6hB@=N5SG_F*>@)fqR{zEy#P{S7M=vJ2k0eDL|d`64DQ>QL<6Vvz?P(9QD9IbqusCdNHH5~h6iyW(~eFpVrPz2%WAFi}mA@Q;LiHlWmZpQOfNVO%t -n}OcmIJ~Fk1TgW>p|B9FUtTeC2H_|@=9B0!A4oJGrMDcVvmB|f9MDzh2*Clmju3E=WVtHIewTg5`7pu -_a%fY_kgBU6AO!-Ek^^bgaa78@xac3?5Vh7qs;pDs+g`egMmiu*=geCpWf68cv74NOL -l`IKm0a;g4evcf>P{6Z!I8p`mFL4^Mr_djT+rr0dJ7lJ^f|rVMYT!acVBC)xsuvwgjAAn6n{T+Q~(Cx8z%nDN=O35vrcy@PDZKkZc8>e*b&#FHl59E -R{_SyuH~r}-D0tHDv0cD~^O`;P{*#zK(-Vnlp(!`X?g3a?!uu1BXgPc@^ZI!AwpJSivHC&4M750p{v8 -5|}Laq`^QaSiHx)H7lvnF1$;^I%k)2D**QhjkhkjCcwYRG4Sn;*f%z;BU^D-dFoxC5hup(Ur=GMytG$cI@ -en|w~I!<)iL5t`seM5sk84U4f6A&KLQxPQ`~+u^52ciJ0I|SjQs>YUE=rR#R^t6Q6ere?@NBC@$mSkI -!^k^pTl_Y6Msj1ZQ?(I2X~+G6Ask1BfjCVt?726LY+T2;zKAO`)v+fK)=Y>A#}`zzT8}fHQ0|J(MGxb -c-wPxPFsJfk1-wb)gyD!xBF#NeAwmOU9;-t#om?$;+ahpU}{? -db^JtstUaMmZ*AKAa-E;yGFxCf>U)8-G4O4R7aqNIB3jg4&N8k&eg+$#`{K9kavOY0( -y6nPKRqB?}A;0JtfceNcZR7EQiMUA?EvWn;++i3J{3aCM{GGH{TWD*Juk$T_2aE1N-Coj>gal@5+a?# -cv1O1fS<7u(<~81;0Tb%xUqt&T@9l&xrs` -lHEqNJL#lWngo}7hC$EC7Yo4kEXU(f!wE&H;&60bHSKL4K6@Vn4YU~)hRpj_%C*mVT#-t_lgxfs-Wf_ -sknI{z2c^99Z?{sy=fZzA|2xrnm6L~i+;xO9KQH00008 -0LuVbTS#u=wG=M^0QlY3pp#F7o!KnsS~=HJ|4bg0AyU^Iujur%tmro6NJO$!)6sG)QJeKA)P5d0Q-UoeQ6*` -Mk|Ixu|r)>+J6I$8WPHA0&S&o7ECO=F=aSc}-p9?2D|u(FG=XF;`y~*;0SSXZ~7I(W=%d^Sr!nZ^k!8 -*{Tet>8}9n_r)z#-N1lyYI{@X*%Zc1XZr4KUbche_4n1}c95J=vdX$%+F>|@fy}S6$?X|@HR%FQdXT{ -H>@Tmr8$bJ(@1Fhf^oJJ%`t-D1^TW4AnbqslANlQBHlJr#^IW}EWt+ciUuI=?o!9*BInIlEEG8{Ky~x -YFE++hP+QL{E7~dC7tDi7o&M!Y?Fg0hZrOx(4eusVFmmiln3p(o=G(opdl(5Cc&J-=9h-IymS9md -@OcpHTW2(8n^l%9A=@&SCVX$z1{Sd>I+MH~D;anG|L6Y7MAbCSMJYkAF)*MCLKHE-xF2x>yb?TGrY1B -0EXS3K0BlULWub4RyX*G0&UgUjY-Y^L9J|2-RH8;!wTmtk3v$ki4j?)pC&hVb!$7Z2gkf5a3B(PI=HR -A`EYe>zg_J52LQZ{EPjhGX1<-u0@7F@axOsU7=RNa9+GM&G;Uk)#qhhUDtWjjIT0N`jz_jZKlCDFhr# -V|At?6V!O6Ie1k8d^)2?sc7MKFl&}!+qs{;WZ`J`k59nWlaB!A2x2MmOkw&Yu$6tB_n9de(L$?$YA)C -UWv?ss!Am59;Jp&@5mjz_ikLTwXKzrs@)?OSO9E?vPi4kIGfQ?Uk1C70AodM7TqFl)sSMM#a?UeuY8A -p4)-cv;rlD0~KwiMO0m_W_zWVJ*pm+0XiLf!)^G|gu~w~Ml9$K$jC6dagLCyD78w5Wxl7#CCcXaQ&yi15~d{8k>G{_;vCFRAW17n=gA~lC$aCteoUJ*^IyXZu0rv4Et`f0v6!1k9FrRD_UI`M`MT -|Z0ZeA4j?Nn+8pap4wto6SXjH>(HwN&=P)gUT3L3l{I1T=tK?|6CO%uF3=K9I|q^+_MHb1 -DNs?!vUO2a;j>Z=FOxos68jZ0o5?My`QT37Afa&*yBts@S3uJOlOgvLz_5@MLi_xxOn(`wrZ=fn%A1QxQ2CcD=w)YSN>w>yTL#X;@L0`r(cpl@#;>`l?#9hB$GLMRh%h -mZJZ590)o6C09pYahJ6^?KLHLbxx6G~y}VSlJ*}@BxsbRiNgD4~|0MZyHeUfpj}_d(RN#isJ<3&ofSkLn9 -vbjsF(oN1rU3{d+(H2STqHt2kt)*ACna)D$XA#_gt>+6vurhQ8%)tFvS+JOOERU2cIr99l%!lOuAsG0 -QPbwj#=uL82Bo1?k3`>sv;;maN*Cr2E;Nx9=5XdW_sGJO%RfmGyBxSrAsOVlVxkx1KqD%`nN>ARRi@L -DY=*3Ho&icIB&$lNV~Dn5h%C_vYkoG*fl+R6a)Eqhf;)rT7g{HBj93mxjRYWJszfMuS3?Wo!8UIK8&- ->=BJOCA-{dYz*tTeS*lo)vPHLW=nLv`)ATye%JW#D3DW_jW2?M!sM~nJ)EGCMX|J2YJr(^ULr-J~agn -=~h8f2ex>S3xOkRe`;_o%}nShyG*x@xO&Y3ufCGE<_X4nvD=9fBTjuK+V683b5O1Uyr33)|&XBMbW_v -}dte(8S{sRrB8TL40oS^1Mt?B%ulv`_=c?<*}euYMq5kUgp`B)ZFOQy#z)1Zag!|3NJlRCeq>l8%U<0`yMRounD4H?FpIW{n^;d}wjOC<3Gs*^2WR03x)|DA-7 -V?7r>9yld1})+d{p%)Jq0Y*4P6n|SM&jAcS<%~fra*> -eFXegj%1qDJ2*NWI-Cx1@dg+|Av$ORxPKYWCyNU4cD-HvD -tO8w&1%{Zo{sg^6YNbo#)zVOa)hqlCjQ8}K5Ds{9Rg*QYRmrk~g0E!N>;ME#fOj~MRkm3$=rDBZxas1 -G_Gpn^3XtmJ@MFxb{4<-e_a+AIRK(J)nwP_$@g=ik&bF@zyfo$JU8DCbmD6q}(k_v59ele^-7~6KeOX -UiLD$wuh&7zv-795Co8=+zDE=Xmz*aH&g7AjL%JV^2P=$I)YMyJX1+daXUA_5G`IXi&yr~;|jStpJME -?Q-DcpO9Gt9doK)yh!OqZEsNBq|fk5SQt*+e~O#rJ4#0o=&sQv4L4#XX`<-$QM<;ep}>s13G?04MOro -!gQy&E~`2}FRQ~1ZuPkp&?sb$mA<|e(Nld8QN-K-WV;6)0JA8-x*c>5dM0p1}x&P9q@xHr -4%e%XG#**Ix-w?@vwr3q@xs@bf`dDKt~$9iO?*bV9Nvhv4`Krs+qzH9LOqXKzO0DyzRf^W9TZ9bpxzD -ZDSk9rwoN3n7p0h1LVzs1ZF(jJxX`2Z#E(X-INLlqa1N!?lr!RdJnu(Ah8UAutu@#)1tn*iX8oF -)!M6!cai^R?J<7VhtVPqUPeN0s$h^>@#XSvrjqUWH=|u__T{B+RCC@5RipmHZqyx(~cT8oe`;892WHN -p4}83ua-U~upQ>U=}ruE>jtWFeZF*8XUj42T09K+!Hv=tVgwlw(xMvhia-$4?wy*@@kZpJguA35{|bjX3j2ExrzH=kB@d@1d!Ckj!tq6L%+N{klB -b3(T5uvOQL&wtBU1dx!&$@M!db)f>aK*vKULk?;Yi`(;J_RlD2CrM;X}*s;XecC7@$kD-&e4RuXY3#4 -OWW;$W?x$%rsGjfD=b-X0^;q$zcjkq35s)AYHH+&;7vNUguW!&F8(k?Ey<+UwfJhzgo>^AS9L7Ry)86 -Ag)$Z`fhkctjysF^SZ9;W>bzbzeBNYQJn0a{1wzC_(nPp1()8@ysV)=xXw)-!{o;j?Qm -Ix-O!RuaP?xHxNI$0vq+652IXY)5&aR{{Ccfw8sND^axz<$latG$8e*-N7MNbxp>77DnpD+lJc>w+b0 -4&xWIO2AwpxfZbku=qtVnm7k5cV4SM@W1Eww1}60gRAYid(Pn(du&m0%~Kk?&@GSn_h!wCJb@nEJMuv -T2_lNr~9%ti`BuSbIPySu{yS8u*sPi9S4z@i6c~+4i7pdr-FhmsGahBZ*d9ixVITPqcBW42aw8=NWM{RS5!D%B3&bVi+7?zqS}&&r>~mqZVrH9@{5I_H*Uo5#c*h-s_fl`WaIi+d>VA=CBbP|0-2{TxTRow~scz6^MPDxNob<`eLPQxVHZAm(#JqU=jF6y;A>Fti&(a|>waT?5lJD38 -w&8X5DKrmbbcesR(dMG$*r~fA=Y7s(A9Dl7H*4H4qz%8U-(St3`9VJiVq85G(k!XM&P(!gZT7*gWjujEU8aE_jKtYx&jmz# -2o1y;j)l@lT4T3bo-IjqV_qeP;YKRhqQHP6Wl(+*w14XON?1W@K;pisH}u_AlA&!9O>@F4S9zd=lQG^ -8JP-24lVH~V&RE}E5hNq1eIF72mS4ki$%v#AhB<5w$3i91P`%kS&n$@+$0_XNaF`LQHBE?$pQHYk}3q -*vKph!7Q4jnyl4>j>|p5)#vimH{~-h)xuS)Y;%L@g>5QQT8@#SUfBFVs*UZQc+0NwLL)wuyP -94m;&g_10GfNyA$zl1NGPsBi>nmCtc|3+|TWP&{jSznmQLqYtE|qr@=}X`6S2{wF)%p>0 -UG*rK969gA$6&gZ< -dxd;wMdLuJOPw1?}DIUFj$H8}Y+{#GrtK#*z4C+3-2x|iGLTEaJ`ldC^mYJu(;~oRgvrV5=ImrQNPzp -6%20$S>F&VwXlSBNkrx^_^St%Gj%Meh+b8McDex9nhT^fghQ0urx5#MmC5V-V|%Jw(aXyz)6w`7G2J5 -fBYsLC;GJ)@PM*OMINi|ahfct|cf(31V0ztdi4z?h>kcgM2_H2d)1WavZNfHAPUq -Mokiz}$*lFukiX5{Rs~x7w~+wab-fmBRv)g``-`B=us1``R_mS4?6$MG~!qSsu$3p*6;LC24Uvy(`KZ -{-p4cM&&ODu1G*w*>X82>%3V1apj1)l2S_QZ`$rCP}`pw?fk0AH$nc00Ce$b7iLtSBt|3OvL8xEVYw) -6ActDezjrC!oy3@F@(yymM*v3MBk8eu7JEjb72W(+AaJTmM&dcTzlU3|mPFNGH<2e~#DCl*F&!iE>}K -lZ3!P0J|9nH6MfA9i2P2)#n}Jr6dT06Zotn3Pj=;!^ue9cn(jah;6K>kg^pW3|v+FaX!H*n^>jQ9ZT3 -l6$Y$7*&D+>_W`z^AG?SBRM16fzIS@Dj<0KQ~mxGW~OV@w=pC2(kz?qV781+ln@=U8{||NPJ6qyPNh` -sx4vkAJt1|6$Vq<3Imb@4_v8d~|dqaX~-f6)tu~(BCK2_$Wg%>%F%HPKL*`5AV4PA7H|tG}{K}qgcT} -S}^tj-ja^anmAKb=@Cv_>eCn`%{WwDyh$LhY_uichKmnjChXN1OG6ioNdFJJ^H^#`W6VM4AarH^gaceXu_N6eH#t5Rqj>W^r% -Gpo=xj09-|xD6K%wp>kqq}~$GXdry+dBeIgQm)>n4KY!^h{b>ByD}OupWM>~2VlciH3V-+ll{DH*n_3 -rJaDL$@h(qD>&sCHQrHk)RRk6n#K$+IHETJb7|mv^T3Og!82RWvc$Unpam(7Fh$!jz2W*bhxyqRg0or -wK=&Uk@^8fYg)WrR8w>WN=9%|z~g5nMk?_zwN3~&<)-pSFI(+?- -_RTJrL6N-31xEcho30$0^(UAOUP@Bi^fp2>QeA|xuhCQq>4BscKz>wjxfSq;BycpCXIm~H3fxT;PrdQ -}m<60&ICFI^TP^o3YMYQLz@WT5W8WL7SZ`DMH%I;kUj3J6)ds)qXc&1vsH!@ACRoOmK;WzT(^)Pva9Q -_ky>K`ed< -lisf8_r95z>BGo^{q?tdGdzqZ&Z3Zactyy9<&|hqMV{RbYgT&+lm4iX=x@A}$)L#74a|fz^HT#sU~9Yy#272o2e5p* -ZRtp}|4l5vmf;0I$$_$X6Haxun9||3Y!unM|j$k5@kJ(znp%@4QD&`TjW5$x> -gAjAaq;?f%3V}k;X63C<(p)ZmG8n!$b$^(B*d$A7y!C{*_)Ze$G0o)?UQ=y0(^OF^Y-TrxNGP~pS%y( -ja=ezaT{ZrAK#31e-kZ`N60Pv;g%G4F%mW*8ie_I5c0~uC<-5_zbe7j-l)K}8b+cMzrMWtQJyvI^jVD -$3{jlK?pgNycD{s}fI;3q!Mf^WUl+D;6hS=-Sa7RS~*80Y|6G!9yA*$kbYW2NK-(-kMsl8}h9K=(S^q -p6hHp(Rsq;fk)#*a)Hu4^D?#=i^=vOD;KjmBXjk(udSQeOOpGN%*h2@ZNH$1y3j2~pfo -?TZtxl|W`ZTXaN^RPnQVzp}VqrTYdglayy{5L(2m+X)0`|INk%IvmkMg*mi)8`1D%}jRwHjhJNzy)z+Z -QJ=?%xlGihehwLHfVWp$`>-79}TZlFCJlthM2W?=!1#&KfGSOuP3IyFUlAM-!4L5}@nTctIzpAXh53b -uudxila8s+v(q9=yvCjtQp>6CfwG5bEOM^K+SWdI2HasoWP_8nwck;Y|Ugq);V5I^< -=|_hC)R^9?}%sg9MO|ujD4#m?&_6}s>5i)!Iu0dnjO4W~7(wro(zvoMExGtM@ -5L8Dw&9L)5*$D!1w|QfgXngK@Rm#0xt86y9}RA7obgvzr|HUsw{GSY+< -3->xvppLR9H92nTf_?RQx6=0PfCZBp6xC3l|2JI*h3A>rmCB)C$=&{W9a|(m!e&4z1p=-H3?drQajrq -u)pUK0;roj7^qXg&j5}SK@>CEFzMi+y|=28!XJCD1Kg)OV-ShtAcqKwXd=+_+`W{csG0v&^`Da{6?57 -GU`-y&nJ@r+-$*Q-F;RGfveBtI8XiS8GpUQ+AT(M){p2x<) -Ey-kMVV6SfNa0%VZ-sr)&;~CC7(t(tgV4Do?;3j?56H(c@QJaM4C1jr+|o^#94wndBoI(_l$0;7k?~6 -tAkx#{QU6vVrXMz9NL5<@!fBL1T{*Ic5egSxG4IG?KVjI));1 -9#2sU8%#iYjsb!~YuPc4WiiDyWrit7?WQ4ffaw_TUMkt(MV^w(9dIE!>R!T6E>O;7IDFmgNKd5 -fq4=Z(WtW2AprE!Tj&FN=3Y%zv?@P-=%H?8vyb=th)+TphQY+7P)?V__JS(m4QuiaWzsQpigcVX$eZ) -vDb&3aGNE$6VUziMBWuehv7Vb+rO&ItQse=U`MAJu10WnbW=q?-~o9=`Auh)fU+rHCNcpO%4{a(!8$ssuc2X0&TT6sn*Z9@rMWlS#5*o_Sfjo7s_uQcBnOD7}`x3vOO -GVpl=B$$WqUDOFV;%3=k)Xux=E2w`gOI%OC5-p+iABQR-cF`o+MfRNDfs`-G*#zhaj#>Kj(1&id7n3} ->@3*n4z6bhOPw;ApN$@OV#^7!FKFtaTS1c60|pOCBBFYB*VtNjO?NpJf4}(s2VHyyg(0QLHO9{t2@M#o4n4MgVlVEr9giH=5~TS{ -T!Get~kje&00|lCO}%Ng{j~whdHun3e@Qd&NW5K7Y%9#s6^6dT=_eakxvesCABe{v;Mi!0sEna#ba|8 -6Q$WOp+kJ%x85+Xbox2L%QFz86gP&WshB=7q>5o02+uV;f+qN&y4wStIR~-H9jHs2VGGX -Z6&)t?jHLYy<%={kOcbl&_wQe6JI1 -1D(1>9u7c3kDT(8%uIQmJcJoKcs)&>S^;-=QV0+aPa4JhjP&=`m+r8#Fv`FunV@I@S%%OS=k0gT`E== -C3Fum~2k9+}N6GxxF#6F6Vl2*-Fi56wR1)G#N;{Z*62}KhMvsVTyFh;cL9h+5H>f-@t=!pxu4-```aP -?`r(*^2eK{I-_dhSu}E3$!W_hso~;8Rrre=umnP#Fh*fMR7#{wh-AN*28E(wTtG=Of#c9184piS06)OIsW=hU=M` -Z7>n%H;2+7pf72C7b*CDg&q{D^EvUUM~?rHo#EmSTxl!d^P$6Yj -ps|)O%|&Kvt^Dt^(t@gXuAUxp-wZbwC@ka@wV6q9jgJa5YX6)z2bcf`g04l(zRY{58yk$>}n9JND}=G -CJWjWnEIcx7t#2J>@GE?^tDVdim5xkFY4ITM;qW=#~q53=EpZbjnE(R5jnIMV#>C^_CRaM|=N35R -B8=ryL;u?KSW>SWom8=JbIYPPd?+z;}Y^BTK@+Pkkic2U0FRNpt%y=Mk@f8y`)eQl7zrJ=O>=~ivp?# -PSDyTVU79MO%AoAw_no9bHV*de%95^o=3JBH}(l4hv(zKW3K>4kqE;i{w%3YjuK;^6;=9?@n^2G!AR< -9>FHPuw&%Q*bk-FxJMDf9qYI7G_p7xd)P11isoN`3a*!l$CU-BrXSr803Cr+ELumJA&3RksjRMEhCW4 -Jb#g(WA$RGZoruk^Fc+}8(4!x2<{ph)4XrA@5G$+lX=zTW4^{zTQRd1!@s#5Ke`uQ>qhyT#I9Fho;a7_cW!zI_0Z|g%#emiR_w49#+F0tBKvHf8jgPKawNed -`bGYtI=JS?lWC{?jXy3Tz)PQ?DJlLZc+4eUxF^s{PSFdejgIZDXO4Wo|X6h_j8a40@0^ -q`X09n;>hzfY2I))_F3TD@l}S-8jd=ZS}HDv-w@NZj$vTUQgi3YCLCchWHT59X(o~f|%L6gOB?#jpWnABr#9{xJ -TApkfW(cGla;Q`|DMSD$<|0X4%;Lm&yWM$|xolveyE{>+KNQzFM{ -Md8bM-ACD{E<$?1!`C5BQf<-k)umF`#UIRiG<+-Hze$DNgDhVInY^t&hG7_rpd?hAG(I{wxMK2+}H=P -i?ZdhqPm7#OPLI6vz#doXStQAeWX?@a0On1ddW=GPms3EYj&j5dz~D3f#HgJ`Gkw*VCb6zFzGX3E)7^;F``D={o7lOT9KOFVzCSr0($UwdJ5N*X8YWjCt2sK;^rmQO-Ctj -8ATG!0W`9QnSLuF=9r%wn|+Cy|uMwBgL^JXairoHX<63Z-QR4;ih`C3msu~TnKVV1J -H&Qx3U?A>w_71>%c{rjH#Xpq@_AD3XshP~h4T;zoh%8ArdQMJqF -QLers1BDa;1I9Zg#Wcl!l`tk`(6QO;F5#v{WK%ELD+CX -rm(Dj^mjb|^#vEAUg>&)pM_LmzCFS{8U<>IUi$DIgyTZZJ5S*|UEU^yK=mtxJn<{)-5XRD~*lXVpIdsLX2ob^H9Gu4yKaT(Ut7xlO5lrY-mi=QGQ(6B1XJ(V@K+gdi840*e43)SmhzB;biw)ge -!U@nU0!c9+$|P!2Xd?8pjQ{2Qgn$OW_5i-Fnlij!+L4*=^2*bKIIy9MT8d7q+->-Cj -gX@-Eg0XEWSY)E%O(|Dx%NiT6_5J4Jmb>8(3hBfO)y%D~@+b)DKR7uZbTsxi(|UENuJ^~y0lieEj-F- -@MThMOQ&(M?X?Zh;iqB{Lz>QO?d~HO5mxbynzSg7pNfk*YVmQgFrTO+8r1GfI%7X~?lw-H{HYs4Xwo)U(w)M$p=_MU&49v7H(io29@oX>M!M#1AjL0%el+f -4=lF9faEGO9dwZxyXtT?9EBk{Ejf-#i_=$QJ4nDe{q9s~Gm(GM4cTXgks!%JBu&0l-9pqkg)tAi4^=F -#L=sD+3*pRCZ?#i90d=yg6(@3L(VCX~l59;DT^^3pa|uQ>ZSu9wXBUw-_7)tN&l4 -1Rl2v_gcGrcL3dM9!qaXESAClA55A`Wt58(H!xgFxkXsA30)nBqNT*$PUUT%R4D``y+CpL=1noW(pb~ -P|t8jg7f3_-ib6v{^w5}%{2KLefT06zoz83(rMIF-RRG1qj4J|1*t{#Hs9NN<#~?Nua(`%DYO{$vT66 -#;5=mrr_V*XoiTr!&*KV}g99|F9GBT5ACJkpbUelwJ>zljgp6J5?0S)%pl2a@{$706ISr%1nq)OOj{^-#ooW*%nEy -s9!M9Qx6QmVOd1vUqoOdNw0oVG>LqD0yEfaP^M9m%bWsIWWjG!R9<3bpRp@GdV^nPGw -Q!%Qb2KAvI6q{!;jCquoWhYLXeq_5LpH}lY#u!6jxE60%q$nG+bP`YyIWRTCRtM_I#xk0n>^93w6K;jGPRdf -F2!i}QRy0Hj*uV9|rwP`c#(5nedeBQXl-uE#64Xo~-3f6|>gJ!7s$%WDTAe%)C%7w+N&A#IrPJ1sfJW -xSufO`EK-M0%t6%g}USXhl-!ISkwuVB?Giq(l<)7X^t^(9IirgEf%f~Dx7Tu-GvF0IHg>|AHKxt+tzZ -44pj1=19=W+GcUu~A{Z!%+N4rH>OEW#3v^_s)U6y3p62;x*~oV4AL`aUguz0RcYbfeHissoqv1dFx90y+=Ksn&Uc=DH%{0o5k3eMx3`O!!G4p;?_q@7Ji7ac(?e`D<}%9{>O#0001RX>c -!Jc4cm4Z*nhkX=7+FaB^>Fa%FRKUt(c$E^v9RS?_M-ND==&Pto{vIy>t)lNI5^2(LJfJ#^^qIBl{*5F -)R~)1EPSY^&Se>~P`(?*Q+DcpRRDs{Y??Pm+^}lu7KauBxv3)xVzRIP`2J$8tQZYSte~J$2RC_bn%lT -x!`=RAqIba(!%$R4*O-t{?lZsnpVGXu7>#{AQ%A+_p3&-!^ASLBN#d;P+n#CSukaxyK;G>N -w%@um#mJhWrK3lCC*K*mj`uCK6yh<}|47k1CGUVeW$!4TTD|oz({KV!_153X$M2d?O{XZ1wB!>r%)gV -5N7`u?obK^S6Wq#2y*Qq%Vz>B^%0G9ffv#4ck?*JD7UcIesSYJ8E0E5~mexIp{9>3|PQM^!*MNuM420 -2vmLp&N@c!G|x5e!@e|k%D@_M~mt*W-WIRyG{x`HP!R{P(lk4?514E7Y|Mz`I@ -h=mKfnIoM1d=&*W&~IL5#0`YzCgGk_GPavyT@iX9!T7DKBGT~*T6^V4vUO@_<)Lj_&~rj5HFva<8+h^ -eD+9XB?t$ScncBbiFn&#Y^Y^g*R&daHeZ*pl7T91tEie9Y@>{Y=6@_X8NlaoJ|j+|Ej9r2WQw;!eRVG -Qt^T@a{nP~}iqK)W7|l&z`~(@+T{8p-5qjMbQ#W114|MlZF{RmYMuuLF0;jxGfJP(m;tT8rFNe1QZ6P -vctkd4T!0Ur2E_1v -vyGa-GftZ9@bJP&lgs_6`>aAmo8+roNRo1fdUW0?(-r)Q!!AKw&{QSED#bZxEE8LD0e;gNSqq5eplnk -JB?gUb3dOM6&Tbk;&N2r6c?$8!frDzfU3)RxTajFUcg4%cJ;966v`#`m)SI)A%oZ;vhEjz|;%+7)s~K -Ckx^w4~)C=sSCwNmRRnX9-EANReB$^3*5MX%E=%ukWE4^-+(cgbq%BA#5z((zl6d`e)Qe(Gxm2Z20vQ -=!J9WK=zax-a-eKH8By?5ffD7O7KVit8(_t%0Ac)9*;Z3qj@02{9Fk$2YhtJc4w@P+PHTBNM-pZT!wC -@<2bjBZI40ymCWOVE$EJ&6HN`3zW7e>%NfMgW0$G~Nw39LH=FQ(|e*KHYx@k|xz{lq@88%2>#M&iHB4 -?IlM=5jY1~Fily1o#rGG#f@&a0+b%45Qx9_~DE{ubJslC-ccS`u}fS*JtYKF92TC#=VPm-@2Emow&2E -TB;6eiT4td&4UE_i|N33me#K*GRl?>7?iDvzPci>ljqq)M4Vd-mPZ%7ErKL(j~D(p!Y-c -?oh*f=*2j9h90q^|W@)vQ$u~aWrJ{MDku`-JnKBF`W8T=;08Sjzd+Wjq?vNF~zPU-E{1fx3_R0VS_aN@SmH~Q_+n3shD1T5*{%`T||cI%ZcobAk9noFfJ_vOlU8^pnme! -^&AWw)meToNDbL(dyA=H+V3l>;>2{C37H$P^3-hUCi0gykY4+V5obCGlqu#65oVU -y>7iuZlJz+#%S2tlZ0zNW8WO~Cy--7K3Dq}6%VkvGSKvH(zFsalPu>Z|N}s956No2~`XMLAzBW!Zn -HWP_HKX`ynrqNFF^zdJ)~F8MiLGGFIjbJG!;^k-AwU{Syk7a|NT3`z{YPKdTvRCBaOs@_I{cf2eC}^w -lCJ-~SIjn=>nj;Dwgo&Pl0|J?NbN9zHPxX|UYgxdEF=(_Hu)yF#s`;w!^t0Y3aVKPC(V|`bX~a6O={YwI_R~LqD~XRh%`P8p2-jSU=@v?E`P_}geqyQ+GCU66pZ}`{07dEuCK1=_fXl9TCHh -!dhW(6C4&}YH#nISrCZXJujMC9yp>@#n8jiANOo0;?#Dvg84+ydvQKX-^^y4dt}M=)l%rOs38~^Kt@{=y)<^xvYQwWQ_%5tQs^#S02FtB!}c&6lk@gh&F -_L27_CML1$$*d6z_4l!WR*Cm47L$tPjd -OfwXL#Uvy}pJIXN1S^3e*JKxi=A#@4y;~jQLXCz{ewl&O46&%$)99e9OaIN(a1XRktroir%q_`m0i0v -icuTO9KQH000080LuVbTb0E6f!_-N0G}`b02%-Q0B~t=FJE?LZe(wAFLG&PXfJSbZ*6dNE^vA68eMbS -HuBxS0_8m%*YJ&m#^V_->~;oq~zZN)be2XJ>B!qLF`-g`pou!_z1^`-yF6Urk_~#{hJ8mRGcGiQ -0J~1Q+qe$DBzSlld8ePZUM=5n0og5w&N-1!?3yB3MN`RuF~=+8OMym@n&FT9w6H);1s=iycZK{2eD{U -6eVXBNXteUF43o@LXb=k+-yjN8NSVlDiVVYGfm4ZJQbRIPnqHiZe-$)&;`fmc_6@cCedn+3ZY$ycFA3 -@rED0Mq+8+W!nj5iB&Qu7yfQ7j>1x~4au^yF1sv?TUKpS130&qOvraH$SYp6fgD2449$hv;-Um+yw)h -?euHelAz0Uo+Cw!m^^ai3SbO$z@SD|Olb7uc=$MhdWkTAzted_gechE6m`2)SFc8>6u$+~*3=DLIrh$ -LBeQ=l*_+bQOC43`^;4(x&4%})Gro4)Mq`RhKZO|Or#G8&tu!oX_VCc+O`7tJXo5U-J^2o -Wz)h&bC5_ZUH9$pk02A=@76Q^6fS6tAebV-2R>2?i0Eay+a80++~GY -4v={V%fu|JXa*ymL}(QfMrf8#BD4W(oH^?-nD8jT00d2|WvN$4Cz(tc*$|@9QWMY_YW1dK==H4bAaqQ -q!2dmqB&M>e*p^mW0tZuUB^=|NDOiIVBAex%4kj5$^p=*D^eLr?R*_21hM39PezPe*BFA&0W!74n+hc -8M-D?Ga=rDn7<|V6i6q-YJLrT%G{B1f8L5Hl4Izu#_1ks$JoncwVMLLd)V-lnXT;8HYXGnwYdJM!;o` -4)m>n8IoYz~6l;*T;I7qPFzG90={frWCQ@MeZY;u~jL-=3}l&cMCp -O3jRNo!$^bLLR7f!C6ncOx|ns?bmi13%J~pPLpV%lW`wepz_-3jhw3-Ji{OgL>Zxicu*`rO3HZ^L#m|MR)%^_r;>8 -QHDJxdf9okR1#|&*AUFf37EMpUDw)utr;0%u`_8mw0`wM>8)xJp4@A&2yEc~ZV{5KZ=<6!tA4Zt{j9^ -C}^L8TlHXFCo;6KB%lr2oPyt7XRun+ntw;DmB6WLVz<*mIb_Gh5|B9*hI-pN5=yp}*6h^DWFDX~>lp` -a3=15exrkdeEK=_eTKC&~$$cU_)uClZ@6zY_J=|+A<6El?+XZV0FQSVL%UBLe+B$^S=9ft`x7Z7RBVB0Z=W8*5H~1#MZgcUmv~6jUy2kAV$nZl-Mn9*} -NF<1^`W8_|hQkDWqlr)ep7YsDe9EL(Nfjq^Z;}!*lGm}ysN8zIGm_ZZPxaFi=ttUBd=a06Hp8RW8w}kUq)rBt4SQ=i_RbNIHRCRG97hO9z)tH>No(w$|g=wK`! -#K2|eIh#qXFyiuEnTN3MpDSFy1>Isf6b7;c$hZ|VCHEs&p7w -^bM`$t!Cb8@*MA3jK|A3iA8Q?uq1ZsJHwuE13lUm(buX(IqjQdH>ph)RZGUr{?3zRm0^M0m&@GOc{?4 -s?2~#1d@4#^fEN3JiH|H^MD1700)g5^Z?!pG@n5lao&X#03rPWD3Qr4}5~F@Fjx7)r8=YK*P^J+&-U+ -a$91vqtK@kjm%3JJ+fCI;J+;#GE=4T9~&`nMSy1k1*;YqOV?0FRm@yE;nldr^Vtbvas;>=)a7?PW(|&8dm(R!B4~WP{wd{6KmO -7TIqVN*h;!Qk^;21Qv4l?C%<8ooVSW6{`HfrMw;a<6%C%--dBv43+3Y>ubl{t1IVjd%!+)Ees*@ft)j -UkZ(}JWH9EP9bVJOwq418cGb@7B|;9;2q!^8*aVLU#3+t-fX<(Tkc&7ORnCNxngKBw<{zc~w>87g6gTC>TCI>7ts -dQ$d`4paQ-19g-<;LO(t2Ym<5|x88eQ9xQ>_2W@5q -iqa?o%sDL6_3JwPohRq$3O%;eypK){H6{>T?5v^=ElcWa<(b@Vf2-GG)h`1k=9m2k(D)UjQ|1O<5H)u -HCvw_t6F$EpK-QZJJ7<1{nY1$5k-QtVtkU=!!5VGkAQ6R6)Xln_`|6J|(uSPzhTGwG3RUxg6qpZ+e7Axesexr^5){mYP;ZPQVdrcCv3XCAQlZDM0K -UM#cG~o3D}T>>=rO5l_Jj&PuF>FcoK+JVmo?N6R|PqB-_FjiB2dovZi&(IsNpyf$Ss%WlE=P~_q(m*! -nuSf`Q?F<@uCdb -WJBUdhq{e>}xs4nxm9oRjK9GxB0Vp6R<_`TJk+`{^_ftf7Lr8I@63{73T!NW1IMD?ZX}P;Me|;l7b#6poc84zUaSeRQ-0>|V+Q2!28JG)NZ{=S -Nqas>ODa*}^d-*k@S@7+)WJe4Y{Ra;9f9f}kA;ic}eEU6#&(Uy~uijm}dUtts`Tk;Tj)Bh~Whnz07YZ -FXP=U>JUR9-J_;njH3B7v(zoL&qmB(6zJp#&$F63ViYM%LQ$?y|eK(cgDgvMk9qg;p3FOYgm%_hv?gQ9E$87IF<@ -rtg<6`ST?xJV8jSF>@V5*sS8#27xHX=ve@CH54s$!VV~NWfCK@GRc& -*s=`iVMUT+%w8D>CV-T>H<@57mwS0etM+n1^i}Hv@Z0@ID@23yd#^zE_AGTKIh0sTZbdx@t%5FD#LHH -mDaxADcX3ok{R${hgD&0`=GjBA8FW(?ls%k4G-rPfgjBp7p5O%zV6rt6@9MT$A-hn^Gvkbbukq|Y1{3 -R>}P9aOgFB3 -K_2|KR8{zJ~nO8qI5~n$UK>~ot;e#*wLv1>H=uxh#^R<9zt3~tFvk%9!PaP{3v~6?7niK$Sh`iW5ofh ->*A8m*ZTJ3A^4K?Mxvgv7k~)H&Vw^&{c!Jc4cm4Z*nhkX=7+FaCt6td2nT90{~D<0|XQR000O8% -K%whUBa%{bpQYWrT_o{8UO$QaA|NaUv_0~WN&gWa%p2|FLGsPX>V>WaCt?GF%H8Z5CnT(v2v9(d4TKG -DWZG;;R8;B1HoB|e;?u$Y&EOh@vUU&QCR{w*U}UZn`wLPHIXsKCEGim6u8J`F4xsyWLcIwc^2xhXynv -N*pqF8e*%h!MnZRuNN(e8PVKXTMNd2?zIoF)Ya|X)|NRS4O9KQH000080LuVbThcic9To)u08bGB02l -xO0B~t=FJE?LZe(wAFLG&PXfJYQaB?nid9_zdZ{s!)zUx=8>LF606;c%F!L1gXrU?)?-E$*6M+>bcTrS+Y(kdcD1WR2kyeZGtQ&y@L*e*a-R?w;KI8CzL2aia`jV-#ItEoGl5hwG6VNnzVtNqu2ADx+yP=-jG~}o#S#DwD -?F$Jn={^8ioAG^?_k*2V>v@bG?Pg9P|>9MqQgYXl8GGXjytv#ugi7v4pvOVleBin9U~&*H_U$Io(uUJ -YXF-Nz7kTgIf}l40C2(bt~E=z7^@K1{c<6frrNiZhs~@(=m?~nMPOp$$fgEJ7fL3QKUjyN)Uj=zZ! -9+&n5{|jxx?fnd+V(v=NG5JLh+CQ~6c(?4kC^x9e}t0*-I)A$vaumC=+0Hlp$%5Sc={UF_#?-PmNdIx -AK%~2?j9em?$gJI`w!!fbjumE)#K6@8KcHByQ>dO4o0E-mQ2%MZa)0+IQ=;~Bh;`N+lP5a!qEwd%Uq} -%mVo2Lz%8DOhZP==3jWTQ5{rDmipI4>T?e`;X67f-j!}eGc6u7fJIX6|G$(|=0f+~H`3jqaX|8P^ga~ze>1=zQ2w(i{s8%%aRxg_(6@%meFA4nK}j*ZSX -Uu+tXnH$OlB;~|Dk9Wx2xQw3=w^DY;`0kKu@?xw5+rsBBrSymBPiAV~`PEvM_|5E3$lPbERRvUkO*i- -lbOyss!TiO35To8&ZssRTY5x3@m#Ny?=MFL9m98*gUYX(YH*|*TpleRGgY(KRK@?t{= -WfGO9KQH000080LuVbTasx5ZAAqD0QwOC02lxO0B~t=FJE?LZe(wAFLG&PXfJYgY-KKRdA(RsZ{#!(e -$THM!b6=S%~5nuD}q!VNC(n^R1SDJPLXfp>>6d0V0+u$tC08+{6c;SV>^!RWZP2pATQm-<8NlZ`DWs@ -5&aI-ap2ts_-^Qhf}c9xcS+Ldm4dc>)47#Pee;^BZC{VeYu2{XE$f@AWnB)Esmk0*{((8uzLR~+obgM -0o4X%B3<}-K9IjYbGjYY_zE$Ja-}b{%G2ym;XK!Eg1MfzQ9iwtDCh*txcE>sgA=T0Q{!!nI3yK|-d;S -K0CrPqysg!WFZy5<)Wvc}6O4IZbTr$E0OW)&9)k7`l4N%zjf`8~cMOzRE0;J^~%ZkKozZ4s3f%K)g(g -ud|w^E7R#PBYM-q1nYY-OHkbn6Rv2Wratl7Stx9L{j~0bffvqQo1Xaa5uBwW7l33g1ag|pvuQHSfrr%OZ~z&K?_WbH=-C1h -3&*TS0?rF}`wOhk;=h^$AIh-hk+UZjJpX;*ad?j`+2B3OgA}tDSf$XzntQN^E4?w6sczmvl>&S`zVN% -UNmj3B@hgAW+=e$Y=p|eb*j=BCoo_vz!YqjSe$NkzLgBi@spXtQ=QE0uY{mO)g0C?Ns&>== -t**!NzAs*4;NVV8E==^bf6na3%Hmh#T^QSb<1@h>J}<{_kzW~a7eS7TW#Mc+#jI1udjJr}I5Vfx(00F5hgQ!Nsg$HHOb3*msVmF{gUo -Ym|)%c@eQeva#Nq$P)CaIT#S1L>TDI(8-{n!+@0!Vn$eOmppjSJh0X0^?)J5BDjswxH|2GwM+Gxm+pU -{y(cB}fJI!yssAT4Cyeah(_l{;{|67JgAaVU=uLY(LM;#T7M5T+x2BOFeT9iD}Ya75OlMnD|}UAEGTM&IDX&DqqLcrPyc}rMRkpO>P4ZGy5m0-c83zj; -H!N+(j4*lCpOn_FW#9Yc387P!MAu^IhgN<#g6n9mY~n!mtM5BJySZTrSnI!@Od!^)D2DF>GDOE`~y%+0|XQR000O8%K%wh4~6wBbOZnZ8w&sc82|tPaA|NaUv_0~WN&gWa%p2|FLPs -WaAhuVd5u6=+FNiTq`6)xcMWNvlw~nmTi(@GmRw0rvXuV!jwIW%cN3@~8(A -}&muF@kxz?(IQVGFI%at@4-Iu_dR%r{LY1?vnSkz9(p4@s6J^p!rSKNNOyZ!Z-FF&VnBTp&();bGRr1 -03af-S -v{Z}1>h6Tm570gCU3**1Ctfg-uz-wu^P1BC`_qZmiXg-xdGCHwl!9JTy`Y&@^8aG%dJ4Q0jup0)eds_ -q9hAi27MQQv6I=rUMYdHN8R-@j-I704jr#plz*hc^a>4qDFl-Tdz|>;(#~BtEsPOEeM@Tz-~9;W0s#` ->B7}rJXixU2Bs@{go!nrAK;gijm<*_73xDE9A51tq;sr?$*Kz|l7?NLOJ4#Erbj!ggNZO!fDl6O)9W& -Wz$Ri4g$6NJ@EMrbVp21J7rk$_`tn&Fq10&dwI51~c?xgqbR9rZ!25wnn=H^^&qPbap-ZmN3sN061J2 -&bJ8m-2nLV{JAS^P|2c+)k|%Q46vL8cAtS*uSjdz(69dgl#cRqCQ6oJ#9WPwz%6;8mF%E-m9(^~ -FzP&cm%?h90((Kys&Zeei=!DT7f7vLBa^S4vAjOzL_0@h84_|#;=9kEO5E=zkY}V1KZRj0v=T~o5duu ->cSlsFWhKXMk5gE4ZLlo%h+{}ad!&12c}ayhWdj!O!NyOaz%-rraaVh&A(6^Djm%i8t$F75n2@3%n7Z -2)>e*q|XXBGVLoOp$C1LbKWfbOA+=+R5}~jMF_!v0u+V2 -QwnVT9I)yQ@+H=O1Mm&BnjRXuanx9*>R)4V(m-w7=yZNnJ&Wipv?%ZyEGr8ZC -tMR6qeRQGE?+>YU5OABsnq3XgOF>W_b+f@qd6t2>6EyHyhJ_hK1{aDMw`^6zgd -#hvMz&gC?amBw?{f9w`u%!!{lkZepMPUXJ-)LKFrkpwJj;YmAzQu+d+2jynVzQm%J{0R-(^?VYlQYHT -V1ceU(O2nA5cpJ1QY-O00;of09jic!Jc4cm4Z*nhkX=7+Fb7OL4Wo|BT -d2LkNj+-zLedjAiJduj>0goOVzEA0|u-*wq-lrreEK|HW-uTA%VG^Idf+0UduDst72j -Rc2-hb_@*qAg76ZKKfd2LhtKziuU{S?vaFwCD?^7f6?Tq(_MfxUUO8P{)5r0uXmB=MN;NyD_l6T=vn* -qMT56vGK0c|H!pMkOIM?(XyrbM2Jbe@GPgHbNKT&hxc55Xk&oJ5_w7e+)xkG9#QZ(?~YD$HtmFM{#=p -`HMNG$Y>bAzw=N%fhbdSF}*5Ya)ar7}SOw$zv^b0~LuF#DzG3a&_r?e=gO8Shph)sp7bE*70jkYRS1H -jfcbr2Rc$S4hMqRPd4Z1jRtY5+0=ph}f9t7EA3RPw9k+j`q+rOfcItaUzEDJ_`9R!kAl}GHoK#{ke4W -3u?zMh7zSXYsnPYoMneuCn-6GeIne1xE;Lz;L@VHZjlsIRr3c)NL^{xj@!j@bI+1qs?}I(A5g5j+NIH -c8ntgIb!~}b)Yh~XcH-9rem4R`aG<7PW#S4QzJaR*{@Cq5(~q;+sq?8l@0HR&q(&+j7Y)2qcCQydZ! -8;A)lB-EY5s%o7-p)FUq2399TBF+DT=?w+2(EUID|hmMo$MAD!fenjQK5vR5t5s`Ot)h -86zZ2Y^HllLc7ik8s7I9Wfh^)B1P4ipTek+<&cq1xqElN?vO;M@lxM?Z?X;gdjP+FbCodBMzAHF}1j= -noS`s>NrwHk~*8*9mluXPMw%r)e$UpT$d&T&+sga$YuRk;fVTx)gs~iF_{GEU>GGQvUTm2l&y>a -EB2a&qZdaX2hbjj%q+R6A`vX_i57H)!dzUmhk&bl%AERgob>L#bjzo^>!7oE73jZde4({(@AX0>RoMjI|Q^k3m#f=D>A~-y|JYl -#o46{SyeSLg-3aISUkE6qji<84|kE4s@tM?aY@cs?G|L*+VG5q)=1DzfpUcj$Eu`+yj{vkSge{p$^@B -Yl-CvT6+E55TQlnT5W*q!A5MZ69b*l|W*tt56_ -cC)L-w$(kIBeyM&Mo9$>A`${v~??B5_%wK`D&v(UlE`j$lymQ)k?BJ*ae{h^S-0ibC-0>1@>qi%#YLZ -b395NE*hw0!;9AAYUWdcc~WjSc#VhMVu$d=pD@wgVHszVG4>0?nUfaRM{{RZHGeLhdO-)a)&V0JkYO1 -ENr~PT4al1u=gcus-X-pL`uNP%|LNTb}a9J<3w`D%kaUOKrhV -TH0Icz+)<@b5D{xeS}+GRnFgw8mf&@e8rcbm;F{17hZ*i@21JoI#wao@;898YOBr$ur1vY-;hhBk`jAu*A+`H-8HL06}7C7zSFMa?(@PEZ9xA2FNt%ZuB~rXw+pq*Pag_*T3VVa1SMkzs$jcTgt{7D*DI>9W;idwAV(Y`wL80#fkhf!)W&U27?gGSc&ggFze1;f{ -*}X3d%*7Q9H3m|LL;;5LO7X85J@%<_ -*Ztt1j%NT_pb__$IDFOfhhamxCaVeU7ABAazHNc~b59dkvHIa9vFDbn -05OYsfJBHrD5KLQ~k($%LW-DqqFIGM2W*^+7@c -A}aO>daa~f3A&J*8-Bg38LxgD%5HA9NmlD1)buG62j`({DCD51M&c%}vQ%-&3TagjzK(05Dlu=xB8D` -ZMnR^!Ll{(j<6OEG4N8bhi%0!)taoc|YU?R2yg`w@f(QZ2I-O-)J*wlS&~QV)^W&vOq({&>a&5afRz3 -$aCiFlO7DaKAL{ZDZw1@8c8Z&2t&)f#?_s_fqq^lKgQuuVJnV;r5` -_*jHpSg>9`Ayi|xGpMiw3;4EmXde5d1S{4obo6=5~EZkv%it?s~5Q#m>4_L13(*>eU%`i30u%jYk4;( -|+UxLlxB>pgkEt9|*5-juD)H5Jb2vE5L9)o`wM9~Xbi!3dncB+k;p9Me#QCcZuZ9^B2e9DQBvyg~9bt -?66!G`t@{L5e-3Bo+-)rj|FxQ;W1^;JVPH<_pU20%C8KfK`=Dr^{8cn^EWQ^@oW$;E4g=>;Jza^&X6t -0hMpz`sD3KP%gp6-C-seDo3^PINrNLZ4|#^Le&PM5N?Xd*B}JbXsnP -#vO6V2$;dYgWF2PTtI+zfhu}4kuMk66z~WW`x(4_=*a$7i{x1y#kFw#n -exP$4&bbkd&6grnUe?S=};3`|ohB)Dn3%!nWqDsPPugyh$+{?{ybC=k8!|2>>9eLZ{w#7!C$XVT4LVq -bY8z(=lYL11b<_@MZ?l$N?N^ezS$35&dI6gnC181=4SL?`a{Ga2{(bh|mMAB<_h1kD>!G|ekNEu!ZBu -~Kv4&cAu?tY{t;$4ro77d#Usj*>O>H6p6SYeN`Hn@%F8Dz|%JKl?hERO^>GkI+#YmG>q5NA1tpC`0Y$ -ZE~!*0Bh6%7AmOyvPW|#psYkMsBv!9yd1U9)`~eQ9A3k<0K^NR0M8iv4%bN36n&I9;b5N?bsWDu5_&p -`YQ53b08i;yrFtM6`=qOlx(=MxTT_$Gs*XHA_E5%t9chdZtg!&z^dC*UxD^Hzl -s5oj#NwZnBgidlQUt7WJ?zqK3O3Mtsh7@>BKE^a$*D^-Ji5%~6tQ<6>B^wsDxjt0BIHO0JO5d0ws<&b -;!1VfddPq*+at>LdHrjy88h=1ut+!~>ktLp?nf%tNCR{*b_HV@mo6sQok)y92~NOs$!e4&qYRhv@*Lu -=PV_mR4L(3-N-!!Iv<9-VrqP6T2d&6VX;uM(}J(o|x%$scBr`+rfEMbQxs;+n=k -@x)Dw}SUh%|X7l+g=+32>sA8|DpCxBF8gK2}TrCgV8mV5NKH~1)g6De*Ysp`R}d$z2Rt^<8@aK=H2jM -J3M`=+d9e7<~O>*$URPH6=%D(St3)O>Tu7r^vbHfi%VgoR7)|n*PGS`xjtd*6&#Hhguv^pz}O0;%$pc -v?2tFKw=aHi7z+(Pda2~Za|z}b+jd*fhNR#F1$SJ1+==w|j{Kb&&o9cB&dc>v>#c@f#-Dk+8Q|dWux& -9E)JV)MP8MrDRKBw2ZGhKpWeN4bw`_9=vcqA6QxB=gS7BU4_Lx>@9w$!51oLe>3t&b -hus;;3g=5&atbV(RH;ZZg|-3mWv&g_9Z>l6qP>p6VjoK~rcW4pj!M0JJr~-8Lt^RQae?41pB}xJh>aE -dzhbnFy72LUy^z61Co=wY{5LlI*rhF5BgiGdc?3<*MUYC0hM+uWs}ks%QUp1HT|Q&}05E!86-)S! -j@Nr~dxnR+mNP!%g=N0-$mR+|8peKb|;J2#MLW1V8iLAgR_EbC0-=ea2HE{vii+R$70tjq9a%_okn9TjKF&VH1Qq&$XyV76 -EI#n$H~S`_QAw>p4rU_O_z>m1Vje@=X?l<)ah*u@_eT3}m@VkTHbr;?}LV^!#twlTEO<{d)tq70km -}QqA^x6Ymln)+Ycxxrt_8AKIuDr)(G5r@FbrXwq=Z8VC@^TLRfBVZwz6qjrpu^|v)G$4lAmC%E*!sgx -jE)@93&BENhX-iS^AT~9J(h3ihjcZj2gs$%@65Nd5R2Y5mSRD*TtP4BWnmXd&leykp280xV#!gn>z<9YK4g}^&*x+%xjim+J;9-#o+6WA8Kzvqs7gQ1+6t}Ub@l)7_}A{z -usSv~2?QiChJBaer~8>?mlk0LKtkk#=e8V|)PxR3cd54vd61b4`yYP^z!PSVJW-Q-6+Q50~3;=QHtiM -1xxeO=Y|FiG@v5j9ZL+zXLg-la&9v7878RO5EhIhiR#firh1WYR( -Kvl2$?&_oEnKN12dQJ^B9>Ghw_9T&yk$sPOTA%60)}@RrdB;lMY18DRt2yp8iE1elypz;ODwo6KkorQ -6HA%NLqaG56oZ>pTpfH!}!{U1T?Ua+8F5)q^887ey1OXWvACT=cP02skK+IkR7*T8g<0UhzqA-cQ}jv> -ZpDfQLmjrdw(KH%hPRP(Of4eb#qY)2}t)RL+*bS9sAV!Or=XV$eG3wGI>Y`arz_p~=(zINlfs0OG5Ut -1=vX2iaCJ3a)1b@M{_0f}rE+hoZ`o>sQm?4Tfs8F$G#)8aOxBK5bMRx8?}1Hecb -n>;jkL)3iE9q-7h8&Rq|%r5Z$?ba2@@@_-H5_Sf&}s4I-N2b>g!kOl+Z3_KZ& -`%InJg7NwBD2ptaaJgZrJdx>T}@q3*aXNp$hBu7t#VGgPi5$5#%^i>_OohfJ`9wP6qM%@Vhdg+Jy$V4 -8`$eMO!@6aWAK2ms3fSz8y+oGB&+0 -04Fn000{R003}la4%nWWo~3|axZdeV`wjPaA|ICWpXZXdCgZ*Z{s!$e$TJqJPfls*O%_mfCm&^aVzfD -9opW*5Cn#r$cb={Y)DFb!La{6O0s3iNrr8A56ep|Q>3Vm97vMn8hRLj3cjP}tbAsL*K(uWHuDp7#~@W&k^=6{b6b1ORPF_I2fb|tT<$>f)i&>CB=;69D -(UnVZ66{udkxYHLAS8xs%z*0s0O`dXm+#J=2)}vN25p3N(-~ih!t&`1IgiMR#OB<8AC}jdg0(nJus06 -_AS(C?--QG;(aSvV6$Uv4Qbu9RLAzLpr$L0CXLxM=C1O25BdY;iUqL0V~1NQTUo$e?P9={_#R;p%tNv -7Osw&sBd1k9gxOhNvl^<33$Cjw9o=V5okf8z5GI*Lkt-mHLzh5_hR018S&VVvt8bBbw38)j8St_gyaF -tEC`2fnjKJkjcIya+mWOfSblrtZU%LzA{?cz^pI5BYBZ`r>G+29CS|_6Q7EKxUwbvZvs2eNd1*EPt--VI4+C$8z`%SHtE?FibTQ;RjjT2X^ -L|e6Wt(YG3W}eUEoz=cuQ_ZTunAs2%fg!1?r{1GtfL?$h?pZh)>ROHOdS*_>bWK7|PR49IaMrAyTaMz -={(yYL9A+sYN8a@f=&N}bbg&|-Z3A`S+kA|yWO^?}sA)VNy!WSO)=Y4N-jZ|I<%A7Xh>|Cg(HSHvCyO -}V5+`i<3lr8>PnG5u_b7WGZu*ovtbix8;sURSF0WOf`CxB0QH~v(bR@K(DRL*y9G+mo=3^dSz&~SQPY -MGIeZpubst1E%{zgxSlsANu+8viu9Re3Y;j%|DC~q~$0iFrARR0a_ -zoH-(m+lO0hwYigZjtN29S~25qeP>p8#&+GCDcD`6nPBZyBiP=t7IEuwb!A;0n^hC{ -0C)LQYz;%?0N>w!>6)QOFrRRaHyRdKQ8{|*zsRxjnnPv?AU=t{n7fYa1_;&Kkv;WPG5AKxN58|;2O46!81dmvbo0Af6rC_i;3 -8y_?4$(5BQf%#<9r>y&l(gWrfN -!CbbhD!cQVF)77beH=Jl9X$SDtDZ>t^RaDc{O=-v#0ZTj?NskIzmO(`5W4E}ha?r5B6nPW@+O+AG)_x -GA8tO+vCR7LtMdla0jsNJvZ-L3UN{GY3n^e$}MIXrFbBhTZgTB=#w*ubdzm!6M}`GdOQyC}hlbr0 -Q2sb4fMse>7z6dXqUqNUeDix}bw&)iD8kVNpkYUY$CHoW@0xeNCJ6RM&DzX -3lj_;x@JLxt7Z7lKbxx43HNHzo8BXR5rM|9t_L7;~LS7d{$zaQZ1sxj-d+ggD?dO=x-+Y~7YEs|BN2U9+S+WSzcquC^P&6;*BM}1LUqENH8>VDKnru))Caek8R5~xV1nT`a&8!wt`D -xDdcTgcBNB+ojbfIh&4{Y5dfluARu%Hzu)cAK&zciQWBbqY9@GeVyx6zi(S3~!|LnA@(fb*X$=P8gw! -Z*T82dks1_Pr$;3)_PC^Pzyw9#Y>Oh)-@LkzUw+RrsXyR_%ZvYUSHuBJ0pnW3JDo -ff?QI@d}h_p>QE3tgmTz>s+IcA< -%!GIZ0~UbD^Qgk4uHsi;wv;ck0z)0 -wa_F)T?twPwWriw(QqEHHl7o>uvnR@Fu{OWJpCN#AdJ+77gFM;Z>=Vpbrw{Lc<*m%;X^I4!Cy+RSiu8U -IogP+HlnclK+~Lw`j~uE}UePAAU;6v_8FTXUp(5(zdL1e};dUY)iq4=nQ{d_CZ+&)GF~MnqS5(H(!5y -J2Q30pHvfN|Zm1UG8^aJV#yxrqm4KDTshXM(p@0E1#?rXdyq+j{9GZ#uNVb==#*-)%*YWAX|D>059_j -n8W!s{FML~CGN5H)Hyj0t(I9s4LjnA=_7$0$9DKlO_q?_M>!zKxPsYF}F&?bG!w) -FQ5rm-Nl{ZD)vjE275@YvD}Pj&NYOuW;8%kK{I?fTPs5OA!&M#NDeM{DqBhyLo?=UR}9=AkkB}%)re? -p@004kpp?X&kIKl8^A#7^;f_I=O@r|UxR|=yC^)sx|lZLWd%x~;nKV;qX%6P9ZB1==V*G^4ZZ%9Oe-t -v2il3)c()x0>0J4=P(U@}V^tB(n-Q+-K*2vyO9KQH000080LuVbTcN@$e7zO`0B&gj02u%P0B~t=FJE -?LZe(wAFLG&PXfJbgd2D4aaCz-LYi}DzlHc_!I$R(jeKb#^EytF_MZAu8*Ep>cNb(&HjzEvdrZh1*!< -`x0=IM05{p!*0>3PV(?FGRS1~lDWQ&rtv-Sz0|ak|;&MJ398d9W(-jaconMV064Ql#9nE_U-pvbd4lo -^NH5RC%GBld3ANWmU<7+qdcB!&>Ur#TUcY&9I=}eqtG~ -U19w_Jw1U@*J-z4SDoCj3QoY5T~2>1==)Q7jW&jC{EN+tQTH^5EcLu};cT% ->R^DC(Qp~@B)yRiz3~k5LLN=?)@w9D-uQ8@GY*=Y634D(3|slxz4L{J_lOw2GE4~u>n|xM&x!rAC_{x -8sm%eEI&bhN8;fh#hW~nOdo#;GXW4lbR@p}PC>iKfF*+~zxw4w_lQa$rSgOX@15=_5z#{V?h01Vr6rf -H@U2NOC*o_vhXuu7A*)@H@heeq@zqQm9t7anLm2Mwtt|4$6NApsYuV(no^YqbYO#Q6hlN(cw8OOUsu{ -kgzwbw5sJcoPAFhjh2lAp@l{8X2R3{?aZLUCu5ryf-Sd_cP4G~RBah0!^W06+LI$a1^EhYjMfdegWl8 -;h60;Uouod{S@dsyU|kSRPC;tJ$4BV@z(?=Om-e2>)g)e6ghPS&NWC`6XS4hJ#@^*B;ct-D@amv;PBU -s#mY{s0nv2{_c`j!t-gz;2mAtK{cuB7Tso1ondxWg3Z03|cib_eX4pxfMMOqZp#pdFWncxJJ3Pi^FBQ -X~uw!T?o`9HsVn2U1*@EvVfgRHe)g%-MzZPJ+BedlG{<^^g=Xk=4l0rr$4qrew3hYS{Tz1U6^2JDNuX_6sK78_UttFmfq5pa-l1;az-6>D9K -fqR;JJ)wa?^cd**G=E!g1!JPgDUA#^%sKZ$AfZy@t>!!<}WTT&R+lN?ZvBeuiFFAJcASdAFE_#ge~gHZHrf%T?(6R!G&mSA;my^gp$a5#WcC2HBdSL8}*~x+EA}Hd~DT^Zfiq5Z|> -Nrp4_&0wcX*petTWm?V8(Yv&LqAa9Z^4i;d>lY1ca{wwq_SJ<@O&aTPp&z`p;$K8D+7&3wBAamDS}zE -a<7`@975c@p>!H6Rf9P8gQkPB1egoscA!{h+3uU`kv&;go=O!m&{H?`=1SB#WI;S|~dKwGeg!2BVBpQ -iBN;Dh5W_+kTPl1DP5HzAh=ZIQ{@IRC*0BS)s%QwMjlmR^CG;{56>;2bs!IXII5#Rw6BRFoIF3Xv|1R -aOU`Vvc*^pH;)bi{uvQ-iTD)ak}C*FjondzLwjo$i>_(JqX0rofyX}s=XW#o`tO%$Eup&xgwSL41;C6 -sxML0gdE3)EnDV$n$j5a^>vV;{PwT+!Q-y#>PZ}VnHPA;9UbVjy!aj*`t9_mt3)2X<%B!gj*Kvej$7c -!QkB=fuvo}&><8T^b+V~tbP7WVO*fwrOZT1eI1)N)%3#HGKJYUdia)`fr4gN?Tsbd&NIjr-FKNLyoO$ -&^YN9sijm=#76z!pfBM3%lTu&ENg7C0?~EO#w{S^$;MwLmKAQxeyw(^D;OEx;^yTGmQ{NI2sFJN^J5%imfbGd9g_{4$^CkIw1 -Y_^mI<%T?=n`UZPZ}wf@%SxMq2Cz(N~25i8pRU?Zrjzao>LRJ{9BTHP>BJ_E^eCkj2WGI~aRD}5VA#d -ZT$MknHh7{UNa4&aW2GH3vnWOhYiX|8Px3*pZMkfTIwl=zE@xD!Z2a-z{`5eXyTqdd;`H*Z~) -?}NTLVL0SOdDMEH*B=_n0l%a3C*m-8i%c|YR`!=W_ONofV>SUY;xChxOQv9hdV`xM=c_P=vqI>$BOzTa)ovCYsJA{68If$Ceov9 -jvp%wFcKQboZlDiqu*^=drCWxt)(MPWD8*?;stSoY!u;xUve~YsFU+NqF(NV{cGJa&8L8$xn^Va#Tl@ -pW?$zL(5xmGuiB^bN8y>$dayuSkh~sf0{!Q*bFC$HOCM~zmDy6I=_s -mP=bLYM!*rk40J|=@g;V3G1^JK?WR6DdH1I6Okia!%DP -U7KfoTT;bsl4A=#&vdMgeqFdelM59^e>acJ*6&7Odt1dozfWcZ3gZGsv7h?%<3MkL*m+(sQPgn!5+!x -DT0WKxY~fn)aadq=S+^y)vZqE@^3LkZxk%b`aBtUWS<7d@V5z(oO7X2QhV8W@xD`-I7uO&9bA#=CGga -JnJjf(}-A4V{+YRfirCC1G?=ydHr=9JKoG8)iQGeA8EcI`*zbkPtrs?`!SKn+ncbEcp -pymgw2o&!J0V9Xj2I>!{9CP_Lf1x-&aE0rH0($o5z2J&?eDmn1`k{eCcorqHniGvgnl&)HVk -sy!U)m0rUo9xxwZzDJ!3=Lz%ZUGOeUlQstQcd(hm@+a8)@LZ6*(=AnJ&p!w8)tq1`k`yQx%Zr-X7|QX -2A!o0HsNjVYt(#hdfjWPg=pfK@lZyqoHYzq-rf5)(t0n9fS%wl3!Z*@8hUFbfX0(<*^sr^h9R(Hwp)9 -wSR+IYy1K^1A3}g#lp9{PL1&T)BXn2CL5i)-!NB3^>PfjO{ZXzIM~iGWqWCGP-Gb$fkXf9rZx=xG%D4 -4`fgKB74*WS*KlvKNNk;lm>^Sz-0B|<8h;oj1d{?yzrKH;$47udUc{h!ACpsY%GqN?AQP6Uff4(;WCt -;t59<-^H14zku25Oz1nU@6HK;uw!hJ@(QXduXqSD>){~>g!L!l*8D*PC0qfEb(|0OfIe4yg9CV}}S~H -Lh@wM;GRh{7``CO*|C1-d!wAkA>Ny6dqN=c~MwQAN4Q;MwIQqEELAfLqHgtJ!V`?nLl0#m4X(jFqPT2;dY;_z?$IWw9?4$WAfDuPfrJ@inM{MQih -1}|!M-#Fk&iByI&ve140U@er9HH%SG+F1LWHD?z)Wb_w^Ho}u70=`Y3XqjfK-FB$;kdskdNceJpb>-y -L?e4hF*;_&9R=KoOaM+13fHJ=Z(EzHs|0DKAW6Zn`gg3O(E(%r^A==BE@3YdYl(^V$d%9DVB{Kd=FW? -2&GF?d%f&vG>m?>X@{K9w1JYrc7w1-R_HGq>!PjlDlwN8BEaM;_0O`zAH_j2K>ZFZ&RcAJfG^8ahrRolc_c+cf!Nlek~-=c@l=_AQiyd -*YO&iHBM~$h)$kCtFqLW;+o_0eA?c?j@lCYGgP}Xcz=d;0zd;xxbh)3J0*kW-+n67HQ)Sew -0P#mr{m(s%_k1bwwH7SkL$vcT$oI|j%+Q5eild(G>^|Zc -$>#gVYY$K^4{sCjIqIo_!OcRSQq(pLB+lX6a>Z{_8J -^~S#h0nz;QE<}I+~0`PJUofiKdV6y|AD_vtlb%gDf;p$wjU^&)s`W~4jYEt4bR38I)3#y8?gz$q@u(K -y8X=<)(#G~o`Z&AZtivY&OBZkJ=QYTeEaqPYdyU^JFTrI=V>hHn@^6i^FH -sS)m}?R-OKZH)v;n1#9{e!QZw}z#v}eoDXa3!=Ik|dW(q+1%+$9Pv?=VT+}kbIb?9o2h`#FUO{70JHl -Kei;z;nuvxmTjG>jur7f2eNNewi98I(O`oJepFB|B(@pzWTMrp*hGa!t{^Tf@#@S4sN&VZ)FkB`Oog3 -0L)pljA1xZZG0tDMIRBsOavWIsse_l>K+>V42*=&v%0ccnqRETms^PT=0iI-uLg*_UVKUXhga(uARX( -FltA+CyyqYe-cADSb>}J@RLlhZ>p0pjndCjYnDQC+-vW -LEs_YOcH6;ooznx`P}_OeBHvsA2FFc>txbOw(Ekiu^mYKht`95miWdkIaDit;-TrDRhK3el$2%?O+*0 -M5x!#%e_L@zHBMCI;iUHR#ifFCfX6QhV*{X7CMy-Q;lcefa#sfXJY83m9D1Z4eFH?1yL0?8tmk+>`2Z -w$wYYu&Vz#sh1Mx3R0yT4!DFOs#@-9tY(!7;;h3A2zca7vsDjnTgULt}_eE<*qii6l1U@&zw?hRM9+D -xrU#5c2rDi8ZRAVeF8qcQDe1xbJB!PoG*_d9jvBK}cZIm&9}VWlMtP+T8uGodW3ic)ykN#_SCF)i2;A -{&Q{oR@TOSuj6)KTI?NhuZuT|FyB&}n?uHNOI)*$RFfPhVwtX1lFt0>oi{|>swJIo1v=LWQJ4cHOxj4Bvd;9 -9;kEcJLIUmoG*U4f(|CDCS{8P!_pwvB7-@}SuB}C@wh^Uq{-^WnJ?$wu|Imd$K#yGqcI2icXdHuxu=m -11M5v5}&Lmn5tX~iD|sX{pm`eLZoV?fNu<8p^*t210gvZ|IL -#4N-UMp&vRgE(-`CR4YmP$qE6l5*ig&s9zL6HW^7B7Y6+aUH9ez=*hVSdRbzt6#j%F`Q0*JSxb`WYu( -Y4!l6qsH-1Jgd_9hYbehXg(wp2czRfM>Y_)E}yIn#Y%LXoBs9}{>D)>`x%4h{~i>uS!@M&^I+pz?W -@EYv+u$^$I1Y;IAMP5(#7oZwpin+}zpsCOPW=IE;=ZT%w~e4@ocK_ufL@o0rt0z3;)jo)hU9e0Adbn6 -iMs<`{(pfMMBSD?#gakJYLdU|L{@_w3dp~|;5u@8teC!SLCyN5?kkG4DYT{KIJfa#on{I|j%wIG|E^Pozy!TcFyr3DKb{< -yli#J|p&9h8&`)qXv?n6$i$>nDc_+2_Yw~hZz{4F0 -0(pti&o?U?6ve!jQOywM_08+ruJ9t!*Z7m+@e2*_b4r-#%MT4$#T7{%D>uCBY8hV0`zc50e5%6Uj9lY<>Rs5bZ|hy -Mp7l{g;`6=B+qhUNZq%V6MYb#|rIL+KA~X#UN=GInrKAS<0ot&BWHm_T1b=X|ZLt=*q8~1lGL^41ZnH -MR{T#l8RhmgYMDUY0uxEJRgniBzPfo6U7!)Z=AWFjfaC7e*(LPJQaEk0dP)h>@6aWAK2ms3fSzF>n+0 -*_Y000w?000^Q003}la4%nWWo~3|axZdeV`wjPd2V!JcrI{x?LF;w+{ShP`4lMhqq`v%lANTc|9ai}<)RbLY;T`!#oF*KM-4^|gdTfQd>-L?x?Zt=wwk@_tYpl1pTB-_9zj%_sbJw@Uvfqg -QrdcrzR#}x-3hp(wKp*=3wydw!``J+h*J-^^=;mp;?D_6_FWTa&68!X{?Bx4b#YU{ocU$%4n=SUHsQB -qj{AnlZrBF|@kO62`m;5q00CmG~f9h3gA^I0W1&=xfpE^d`UB=Q -UuRMAtjfb@Ju9f&a=%;FpOyyMA8@Rpl8y6kYUMex?_PN?RDw8->wC$-a3fYI~5LG#mWJ#Cz7ZP0P1mi -CVPfk{|vc_P0&D8pcZA?KW4URU-x?+pE%NeGBX$#|9ha+wX45UUb_6hjKoF=dLd+h31p0=(<-;|7EkQ -SAkIlByZAXQ};z#8w2$Mp1Kj}ZLwMz-1?dy6Ls^n1S~q9O;kaeS5323;CK&%+;+YEx)fEV+rC4UDucE{Y#pUN%RJ45$zk%?J9-I_q*-BA6WbM{P? -o`P%3;jqUb<&D2SK*h*XyVouDz$%*l_5y6F^{$H58z!I>ywP`oIRbGSb`B33|%DC61t7DP-ErD=L`bo -AZxS5LqB?hHPGh)TCr(Zg)oL|GQaw`Kk7&*Gz}&%QkU_QiRA{`Iq$&j2poG7-q={zn#t9>7qUmJ7w`LeuyQ -MoWTlYRc^BLb|>tw>FZW(+h*xhwA>f9sz*+82jZ&Q2?rv@4op;}L_zwUU)`p-25briin1jAVpmpuSx? -kRU|;U)B~sXdyKaTB0r8&@)u_fK+TpaRKs=aaqJku9g4IuzdiG%{wgJSv4m=ghh6po46Oar|wesr!x1 -udkXF8=?v9fAMY0(!0AB=#V_8^+CcKy@<65x4p?Xp5!Uf=Y&$q_8)9k{)~S*R-`B{9KOQqur=D$xs%| -F}`38w&n;-HFyOVDW>3{aO?&15uaG3hZombz?&w93oZpHHf@zf2aXcum^h>g2K-MF8omnnWv|3{_tsK -j2`V``R-GQ8d={RB#zH!6M`pD?UekH;CB~8f%>PUkp#cHP%`VkuWYg!`xp^5*10iBXp!NL6Ni?%n2953R!=;sUJmyuzyElW4H9l8C>oiuR -rQfO7H2qgUU){NmZ0v;4_7FTQ>GDpz*o>(i%CpTGJF>iq2J=xB*EEIQi*F+PAXxe+sPgXbrNb{xlFfE -gJv448va0ahwPMm>V20B}T*|LVBS_--ZEhMR^yA~|BSPE_k8qWioj(WYhhJX(CtRAK}q5?_fPTPEY1v -A8b5v(c_2S4P;VQf*NJ;^06%-+;l5P+~=+5JeB7ZChM#ij$~rqGj_Q{NRIqxB_)>-J&nX-JwoQU<>4l -hB6Ti`A@-54wLqkjo~uxhhaN2MhX5cdgw_dt_x6lKrDlB_Ap|1^SrFfKF`%SA;bv>7}#&YfqH9s?H3n -BidRi-Py&mma`2bM@K}EnavhajNy2d{X4HBDL*LGan(^w}z5QZnQu!~cZK7P}M+~^~xmAlGm0^ -3|Ke?S4$XbBA8V_xWsuOm-lXViNM-FKXo@9SJo2Vppp~oxZd6p@`cg^xD+~)k~=bY*kpBtE!J7Ufvj| -*DS@7mhuy|)7n7sDX%V1t>(^7$}>*_53oxJ!!xgCNU=v7iAm$0MzGRW);J@R$iN#S=DiZ)18NUwH6xp -aUi3mp~btRlLs{jVyi;N5>K+nBjbbct`PrM|8zWu()DBu0Y`kRj6C4B+Q&6HKHc>x(F}hi*SLmFRF6s -72(%og$BLumeXHCMdtyTFv7;jeH1cP(2Doz<;Q6+Yx=mQv=%}W)B$_{kQbv`+@rUhdd`p)X^6N}9kP4h0=ZRLb4YcfFi0eb&S$;=0g8JUVGxo=8vN!# -Vk-hui~M70nrWQ*&D#r&q}dVH=#0XN-NEHSp|loRQ78qMN$(}3k1UllF?`ExwSqSl9bWW2Ixm8Q6o-< -nHO_+%>qS~&pyi&5KA32rL`i?=Me+ncfyT&FLtHA)&iZg1{Eg7A~TWd|l=?@WRaW&vc>b=8XEo%cm|9 -RI#OSvmTkH1L7}5iKqC}dXOR$-4uSyy256M;h -zgjlAd9-3I>h*Hl!q1VLNZkiO>|-evF@qTQ`QAggx<>+yctXtR3hmi=?|65PUb2ALfn+KQkA<20o@}Q -@dOCF4H2lTUt?5j=g4afoggsQtsJld)*DV;Roa@@XDemcyG>)RDJkB9#7 -eVAxFB&x@o%S1U{pgVN`sk#0OVIF_`g+A3F%M^qlGG2b_SMPIFJ*Us%mOU(5JTchl^uRdfaOk>w`#h% -K94CE-zb#k&%}Mkf{=srfIi2LJoS}j*>t(egIbha02cosn0#HqN*azwGq`mwAdF{ -c~|~9G{pE8sasq*!}*(`IIc$Im^w4gd3EsjAv58Kro -XE%$5~cAsC;sf_p|StTl}np+Jr4<*BHhr7P`t9_J4E4x_rHW~GE4OXl1CGb%?=oUDG(+==H9ikauU`s -LnYl^0X7I#pLF9E*jAva&x${LXT9E$=b01j7kxio{l>YNSYoQdHCkA^a+FDK;{!fVh=Xk~q4{Hrh~a(lvG{#P2x+yF#@=P&S(pV -%WKP$Qe%~l+l>JGh~dVb(=IVCyVn*0R4zBPNp^VSlu&kj!EWHxi>wip=1QKy9pgGlN5HAXnp -No(={b0gBg+#|&dGNx%;_kXDMB9(fyJPkn0B}8*NcCSJ&~;EwYC1%g}dhnM>myxLoDJ{ONR

Tk!R7n@?)#Ws$g)ThTh4~OtSqg!v!;!o(>Ys6jmo@L3;=FopC=NqqI-T^0 -$|7lJ*UiWpxRln?7{t8YxUZ2=8$LnWx&UH&~$eDA*vI;RH_c@_e3?FTZJ+k|@#rQm}cyV-|_Bfad!j7 -$GB;sc%JLhRDidBx=3T8^6XX?hVz|a>4R%NRsAlNgscv55A4%MDedl~*?)Fs_kC@}LTOybv__!2#1(H -#(@8UIN}MPOz=1i(n(WW#MTBJ8r^HklG3ne&{C$cSvXO@b3TQ#VO?vS#`&nVg57wV#a7pUiqnLbD>1e -}|7#ma@qp6e1c-V;PqS*-~Wqz_T49!m5bhi>fR-CF`{&ugsMIy0OEyXrP`g5|YR8eREaTv`Z#tp2>tHJt)VM1+;tf -4oQG486>{xa9nOXw5Xph4neN6Jx0;W~N@}}vaGkTCU${^KPQm1z?eoyQD85y#;>uD=(V3)Ea>eUeMn; -qcL5#YNT^vZx5SUSEnFrfnNjkJFdMu@hS{h_xmav}PMula4xdNaNr%BMK1Dj!Y`Vj -<|MH^vfH$!eKS@(4I%83UWPLGX2$5x1lJ*5xrxT<7nz5k{7n93}w|R{WP|v)KF9(ldZXhuXjfk=D$Zwd%X*^G><`x95U!O%EcNgwV*Z4{35}=8Co00?<-N@i -6F0f$VKC+WW!fA?u@)Z4dZIJ=ah@*44cZy47kC17)c%U1(P?;&1U_NCv6y#IG|E1asvF@CHK!dud11>>7=+V7FI#CZFeJnooWCct9nmss!McA -B@^cl+H#$J(7}ZkjdPAd>ouKLgiW*7(+I70mQsyUF2IQ13OV`irZ?H0nT& -vGE0W?zem>zLAxa%ZQEc|Xw0O<5Eg94ZOM^jOYBB0)(P6TD&qptu*k>?!OI6zT<3%Y|QQAX4yKT{;e` -*AfIPJd@Yr|MHVh=srTi9c}8Hr6M|GJD0EoMXGBOdyMjHnxS(rHFo%YG*3tGisUZjIUuj!qhSsO&jj| -D*?6zNCd$L$dhB8A=R4+ZQR87Tc|;S2Np4_D!yGZO%VWmOhAHHZ3w1xQUb+wbFi8Q&;;)F501!-n6bm -9jNV`6yjEhRfp;6>s<@LEoNA?^4y4mT$5H%o@|8y17Mk&^dWAih8Moch#jEdPQv#XGqx6wy&|O5@jDF -#^k7EvLmkW`#Sb3c!z*`T1D?`D!Te{7sAH;tCn`8*Zpm!(LWSL`00K~@6@S{5E$i@2@lHf9_pkTS4Gi -HvT12;kT~$5eD#tr2dbTjO4G0*&lX#I8qKzMEP-1=zo7URu?3h!+{lsfBS=}2Af=i-W+;Y6lP9O)*37 -7@Pdlvt(j>BG#y7r~0R#qQ|7QpCMO@zmo`v$-{aZzxT$-^JkDtQCr_=rDg>i0N$!aU16nX~UiM&6b9iw)juB72f2aF6OwI49b$5N0v-EG<5Fh}2EiU8@3v*_{ -2$k2XrkhfC$8RG&C-EXEd`NlD_yXz`4({=i#$zas3<$5941F_)enSMXr5?(y!zPB(aOCIs?y{xBI>2o`m=Qm3jc>-EkDNgQdBs{$2QprgqGx7z!^5{uSfNCoEE-7Su -e{->wCoiaJRNNxh?KK7Bl_F+vMWZ8rn2wl^)R#11nTGJGptf0=x{D -!Xk}>~TjCa7X`-om{#WtHA|$VJ?WHrS*70vG_y^cXPXtJV5LbLrgkKG8aQrSBB>KwOVftBqMm~0=l3< -0AAcv8T+)Q@bay({+?J^UtGGmFRQUu@Y5%0!(;wZhiF7NhTFE%;Q0*DX}2vK*HgJw9%NYvml3~q#%f~x5`Nj3vjue^b(`|XaXdR4;z$IwT6T2odX(P -;CTM=5dv|7Jv5Jf5QFkdNR2Nw@G__P604*S^F&Hb}SH;Kd{DP$TP -n%_trc*W<Hh6nAAM^+T7$N7Z@{rSoLfWG3+A -FmQEvci#Xt#%uKd+vMtYoG#qpt7nX#(2z% -?wCRF6rnH5&LFXjoUBysfhFlJmXi|XM`)0Th6g}TaPbo*GI;9}XBjWS|~45Dj`Im6AC1|nBd4DYX+zH -c`CeoCedcpLk`bfM#+USlkb-FCJt@RjdOj!z2HTozhPiMU1dIGAna&1?^d1cTSwgf+1U^gS3#v*o#!9 -H$$000ZL5A{z?5M~U&yh=O6PHE-&?t`fC2CeX=R87I;4YkE1;O&jND@Zzeysd8plmbe=;{UmdLA1LRU -x+|yl>#>DLS?SQdkuW5P;pRQefZcjseh922J&4Xq%*TzEnB)b%8?F=^`(`M7Xch?XyreJFj)A(44_b) -wbm&mgDVxivd^9wIS2)CL@32r~pr%04Ez2^;<=ilxan%mVjbZ)Vs%mC0pDPW3|N7`t9k|xCu9{oS5Cv -C%8+-WP*{Q;d6$Pb^DKCkKHynCL6tvFERR4F|Rnd!?o6~>j~U~Uk>02MzX+JoA -LfTl@-AW6#?IYupCtNVW$T#)ZXE2E4?K`ON0Ov;50H@`3V1Z$5W+n -q!_Mlh0o5Bamjj#%*M7$Y$6IYhhp)k>cA3$yjk<#`i>U~|8C2Lnt5bolYmQbJ6Us!=h4^i -vOz+Q7`I{JYNGggq^#*Kuv#a{>Gn@;^Y7joRh=mMUib??A18qRzlmc-I=?@Z+!RiQNWX7NH9-Eyd&y+ -oDS3h^yX1jd$N0@BgkcHk{vL&{*YuIj8bc>mS@SZG)ePZ?Xt=9!6KIS``{=IU6Jw(PfC ->Eb(3h%T0kjEMqoBfywg06O_4J9S66IRhdO4&PN9>8go -yQ&PGy=-T(Ox_}W^o*t$p%NtbjGk|bw}W>YSusEIgDnok{}dbYEq@e3I~A$pP+}X&GR@ymku^eLOElv -UA*a1_Zew*-FL**70&OtI10|3@h}x$Ce_ImM-SB$T*0Pd_}{O>BKsIWCZ0UHqftkPmnIfg6r9t)7B{o -lWF$ifmq8JDJgbMgE7#)jFd>;)S2y#tMV^|C1Pwl$a|cH&F4vAy(G^dp7_o7xsc!0r#9IB}-=exp&ZA -u!cVsLiK9L2Hy)tRJ+kwQ*)S={35gAUAnaT;Qv;RGqUU5 -vH`(E8JXj3Y`)QI&Ao3+1vF_LhP{gzY-Dk(M`*sbev$4br^Mp>x6Y2@AcE1T?+r3>%+O2Ttq_8e^4(d -M86*bpYD!!U}5WNtXGF7ZrQL{ti?h0jKL3^@_J`%<9ChEY@RvAO6mF+$}N}L*BrC+3y -erQETl-^!;(PjsAt?5S=F|VtZFdNYqMwUT`P(>HT{}rR38vpS>W4pay`HGe{`LcyZIW>NIC2}x9A?vm -x#l2>fpv+c3$Nkn-aouI`_2suO&Y!<{{_2@MbDbY7Ku?mwi`e4u>B2TIZO$dn{|>ooY0gF -6284K9TtUO=gxr3fKIHZ26vtyc3;Qi-m+fC>7BY4zxEA3w%i!+(1S+4zF8cB+yF({SXDFJ7pO3Z9BUz -U%{aVvTBaK(en^TOE4u#wD29^s>7aKpy9)WdoF9;S1%>r~KO9R>qp@Rnr<@2(s?wHkl<6lOa9-f$mv4 -gOnHx5d7rD0PHxIF#qjTof}&yj`CP^P4ACM(2f_j=c)P -uaCyw7v-~mFhPXWTi<+@v25)=3P5VVEK5Stg9UbXa(YwecR~DVe$M> -MnhtWFhCpK+g$z#lK+gF?)14`|yYq6!X>gLZGpbaFa9j1CWTb%0McFj0)eHaewzN0gAqui++dv}26le -xTHG2heKh7I4EOHhM6!mQ}IYLKU4nFTHDId0ap%^a&zH>NI6!SX{v)NO^~p2~*G9VU=Rg@NiXvHS;n~ -8}k<>RPb8%`g5xZ3|)oVh<+E%%u-;S7`~jby&G+{O9S#UNG$?T*qQY%q?u(zGiV2sjm}%6JYaGG#U?; -5bgd8Yyq!yvI39W@(0?3rZxFe+4q-PktSnKr3C{zmPT!uw@7O8_A_;$vhIXUpIsIyRo^QYc&-3^s_)~ -^7to-ZfMSHz1+D@47x_zh4d4(?+YE@7vNQ@o&nXl=~AktUzw?lwwviOblV;1&tYpmVgl*=2L`eihff& --8jvW21U7|UX`^3zDt{nrBQRdSU??1vEYnP^Gw*T& -tm$PyWff>Q={v_q@(0(t03ioXU>t@)B~zco+QGjanuX-+wa|dx=`9(;3bO%FIjk}n3wSica3GGdDI@`>%vVj#H~6O69mr|WY0_ -j^PVK*xYzl85N{pC!5H(XsuvJG2}9`PFr%uG$9Ue0)SZoS*Q|RN?E}bRpRkzQr%AMhJYW{e4TpJdz^b -?p?mZ@rFBY^a8YdE!;fb5qKV_3BVP^eC?A;~i;>sq~5Lg}&Vea;lJ96eXpiNo~%(!Qh-iI_FoN105D- -X6WXT3g<%A3}F){F7!PPlA&m}J5kJm~qW@pz<0Zvf&JWp?I0Mrok~z9ifZh>dYF*uis;6zBGDnW^J+! -F2eCxpe}HMJYhIqijR!0{y{Gbn1+EK3KZkra9QiE3nTo*)~;bUYZgz@AF6+L;NjMuV<_|m2%6l`qa>- -GNqT_o6FFcFpKX<*`vX;FyeD68E=oj{|X~QXSbTaO2BqWrhcx=ivwgMN31d)1J6wypURNNo3Z0axohD -wN%;_0f(V@C>YZ3hfcH7JlnUJ8>{3Z;=KxdUFXt539|+Wb`=h&^%#Kw#qmc%FB>*oy+Bsre{{c`-0|XQR000O8%K%whJO8{DwI2Wgahw1E82|tPaA|NaUv_0~WN&gWa%p2|FL -YsIY-KKRdF?%GliS9T-}NhSbgm3a5NlVClXx#kMUj-)x!6ugI;T>5TUaa!?!p2I1^~6%OC`Vk`Z1552 -SD!HPUVA8r4@F(*7>FMt28Lqo_ldQ_2TsCFjSA9}%c5OG7xXmw}1WftK!9%uU`E3SKs_0qgOBM<-ni58mg|mYAXKrr@B|~zbQA>>g|4~OTOI=b-OJa{`RK&+rHW -^Rha<$pDjRTkLHhxTl`He~z)KWgY*y?m#k@pEz2wv8_AN?= -Zx(wAZdE$Qm3{ZM7-s~zRS=u1?D -t)zA-?9Xn`)~$-<7MC=#oG4*GzwSSp#ejLep~Db?wkD+eTNYd*FP#;$BiPeRYi#G{_tNstbp^#_;n7z -hwO1+q<1cFjOB*?YHnbqrU|3s9w7x=8QU)>n(xL)?(-Jh0hBv$ov;PY1 -TmSuQ({yCcfB(Ab_l+8`FytQn*36Da^cT2j_M0uYU^bh*tk(4w=2Dp~sUW$nhnso}-~ag*X)GQQgl{8{`Cb&w$U?bdV`j_aY^LUj-T!tU;peGd`^Y^$!kC -7^0`4Pz(mt13C`s`c4rx!Zx{>kpudK3vYD{2}oM$DGR{lKAq~^Y8vzz~7ga*S=g=1u<--gnG^u)62T2 -?ngM$S;GyCvQ@kN6#mvdtO<8Xiz9ryYH$0bsjkcAT>_9Gva+F(pFHkB!&cp6;1=5MviP>!3n6MPmIi9 -eVNI1%>z7s2fVg1Hg9fgzZ(zt>wF8V9(C*cGn7icClPo!%i@DG?O}Xo2V?&LF>C*BWQK^bOhH`z&IU81T@=%zM5I=z*Rw&fH$-T&)K2_M4d}PD=27 -HKVTc7CX)k|0E2hab~swgEoh-i@lp55b^RXN4$ROWRe?uw*}y2-!swmS<~%Hmql7@O^x)u%UgKFHTqK -0R1NH$#xMGwD7tQ3l)IF>vTyaWBWQsfhGu1o7LU{mcRr4^{-?M=T>nx%b=#=G{Y{*39F&*+dTY@F*$X -WAzAH-2^L)160tTni_MbyccrVQ5$v@nJW4U81#25>MRRViSnvKh+5@+_ITM(`}j0Y-zmWoN=*HnU!9d -h`w1SSyhOP+4 -f`nkefc}J$pob~BXv#ma$di<7W@a?PC$g?)WPi@q`=*&$S-)4VE$|TWa)}c#(4vg!s@R9*Evq36EMOI -&yQx>Jrc#92wiBdV<5rc&$w>W81d4?aJL*f{GY8B8941!gt`DbYRcH_(iy#GU(&8%$309B_X@KwA75^pQ87Cw}ZK3K -tU~<1{@?%fZJAmI0JWd0OThc@mKiP_M(6%r{s(Rc;~(n#~z|!p_XZ}2A{F*?sBx<=8OS~0PF4!{M%+d -PoDlJnQ`=FWI?h7u7(xG&A>jpcImY3-sXo-ehGmkvTe9=wrow2Y5{??P@JUyV35q -DIKxR)o)1-<7wO)fxwF3g?E=FM}40K&~?EI3?;xA}sVwyF~pSoO3~j3t6chHBp{3ecynt>-rYpao0mp*pvqHSOfd{ED7!B%$GSi27 -uqop=ZkRssQ{%sT{gO{*+0t(;n5S_PQZa(@=)wg4!bDcPu*;(C->U}czcT$nP?Vi>C|*6We_Vi}%9GQ -omfe8w=Sb!;rcmG{Qo<$Y;fx1dRXHoq;6?VvCYWKJbjyk;$j -j5Qtj;N0*z~d^7#0FKC62^;=uQNx_ETS3LW -AfUy0V3svC)X+t>Isv;O2u1l8(>eC}i21L&4#VvBoz}&S8aUW;qf;$FX?%H<|aK|*u9Ze7!rUp)Yiw; -=>=zNc1oS3Q(K%C4(Z4bl~sxet|JO6{3Yt*<-pLNw0x&x{3B0#(tDA!mxW>TnVa_;Dj1s4kK-Dx+pHc ->sRa_|9Ts_u2nE@cohXXY-PX#k8fMO;DW{NLOGuRzs>2$T!KyBajgx2fcR9@KDJn`%lNlBXc##>Q3Ae -!K4Km9^cLO@Ck0ZPIu-Il^vnje@{z@l1koyM^~&CDkdJsgcXadH22VJ|llB0KkR%$9zR7ciJ{$s(T`!o -ZX>@32&Reh=a@U#@!zow)XOrV0_g}c+!V<{6scMROW{nnOC&vo)Xvy`5EYjII|7VLvcSspppG>-Txh%Z#Js^cgf@ARHP;k$Ck5Ajk|)n%)kk21 -KUd*LGh)1*I`=3*2G-V$%qXM!xjRFw1MktD2&~rI$dp3o?#XNeVS#7&1oQpY%0|hTP8T_4h>}w5{|`o --!$nw5U?=0+}TQ%pX8d!1vbDkFY2AM!SbI@(^c7A*kfM8JeasdNQAU --VJ_4Or69Bw?IK=ldizct;F(#|L>zIlJe%Ys5P-t5FzwC4DT&z_$XN$0WbEu6;@+)Whp(C+__e<{boOz&UL*4@3n3E#!3av+H -;41S1Ub|=Xs{$tuDJ13|5&hTgO>3U^a50HlvNWJz0bB1?LOB(~Ts=d!!>9I8UA=v3{M)9=;D*l}h(Ov -iQKzP2CxQqd)YZV#9tO>qTH%j&w(kncKdx`m%{&(QkIC8}bR%jHocS{3sL4Enxl)vk6n-zKxb{dny`u*dg{4tw!r8`(Ut6L3uQ*sXC -7(1U5!7WByUXItUOq~)J#Z{|RbuhjI2-&e~aChUhaooha*=?|c$Q^PTU^B|gLo7+93E-60O37a-g%iHND1?1v2> -7Hps2RZ1!3lc*!$^n6{nmTjgXH5X4%YoKvN$M)Kd2r$a^xm>#O6#^+;r~YId1UPSh!Y!R^t`P48WSzO -a|mwQr0q5x*J^F5|AF2+Do^sqIg&BrtW44UQBGmFeh#II2Xg+sN-Ti4K2h5q`8xJ_B0@R5 -Uen(seiBusAYJG{NuT=vdY~c7KC4;KmOXu{3O^rv97ML?@f7(D@8ioaLYnPnU5ALotCTrv#4Uq^@l8~HFEtS3nzNco<@8o=H^otpE -9}ml!aqEr@vWj#O3OL87(ye3Ae>k?Vmbkch}yICi_^cK_*=UXHXtwzK))vIBc0=EDYmwmb>X~9s!u7H -255g2saeU<&8K+fdn=%RA@s$+D>d3pt{lLXmF>{?Hn32o<1v80w62t&_WUPvPdi#nl$2NU7(`5>wB -w?5T$lWoR5mJQu@bT5r~6lSfw~V+tRodHJ}Lo-1N_2sI24g)=EA1!a6@*!^r6w*^M-jo(Q -YeGyG*K;QIWUHle_GMyY9$yVd8a^Y!txYWyC}m;8Nz87Dbi6vejOR7LTR|1ul)sF-OFIHJ@+V#+LB -%>ts{!b)31suy3w0xHapOlsO(e4cv?uX=5RgCncmwz@TnITcztDc9)ziw&ar!qJmr|Ht3?F?%ylPcIg -7W#IgcVDm)9z<;#OlbO7@OavQ&Wi$|-TtJ6&KBKB`dP`^jyY&VyM;Aq;|6i~nV@;YPcT{+H#DE;Nd-4 -N!PV6TNbT}J2OwNzji-0Y5RiWq#IWh+>n{3Yt8slmg({<>zIeN>CI_tb5UT&dB=yD0=F+|_$R -o}sqogWl)zY3CnnDzGTHfC#5uLP9aG-P2fvy#F^n(=$nNs0e;pU`i3|$CFJ`|QaC9xUdu@j|0R!9LmxbCJoQ&nHv%Sw-HSvR14 -srxT*Qn-VyA*UbbuZdo~C-V<=qmD}rxl^)x`4z{Armb16bJifzhmY#DaY<6>@{wdqN^d7ND{wxN}z)+ -I%mb#$i-V39i5GqwF42BTr&^mQP1h9Len=aaoco0b6GZEOy<@PRRd=YzgU4N)ng%woP06kGO&~emcSR -K1^ja#oq%$ex4|EqVY6uB;oWCdS@Oal*vP6!xdT3P7I%)uVVML*~^4aVM6NL@E!e7wmgl+%&TsimDF4 -IH#tWG6-ZH%P<|m@bYlgi --#!6f0{m0IAxGGT&Y~`%ek1m3Lqky5xH{`}EHjAThpWxGJ7|?#OA#tm)2dJfLPzxAMTmAJ;_7zbxHu% -#CkbBA4U2T&b^|W@c)zw$MR?azbLT6dqeO^L^#!GDWzWv&1G_Ad3l>7+DJB?-NBRFC6 -w>(AolW;Mm;}wP!-a238&%skkV_j!=B^1>4%m<0Gnhmv!urI`|!bB@o`07$YsHCI%6YPbT8L=g6O$lR -WujbX6@IEHU>F-O)u5$Eg~ZZzz!L{uUQAF2KOz;*ZBgWp=_|{ -_7pTTXeT#N6rR)&n=^B0OLLjG_pO?sYETnI75rY6TXo?~tuD(VyG+mBO?j8}H(kAbH@eG=>C^s>GDQQ -a=|g+yn1+>H@|h-$n@Wf)`u<|r)Wf -)&E?n~S>Kc%HZ4I2kL3cxmFt{OgWmj8g3hlaQc$O?xkq4bO)lu6t+z3Qs+U37q|nF*tZ>lr;EunMH@8;cgwQ`T9e5054_kXzIZsCik+>8#Ba2X-7YrY?wPfp#0EaKTQmuWmrlUh2Kg13K(ekf -sx?B1#9G8Mbd)5x`Ld|`pjT9-upGsR=-NcM1C?I=pP(wr!|Pnnx&xE+xi@y7KSHNm2U<^e{2I}cid_{ -`Dg7{Q}k5Uy~T6;K|i+^IwO(b{%kMeAbArsy6Pky}^H*YUgjqBVGwo<~RMg&56loRCLk5e%}fymhVz= -D8!bilW&Y;$D%0H+*T0xplHdoZzUrqrpG27KAzFwLDH6yvr%Pb9rGJo(|(N(1`is)SGa*R5R05!e=fL -p&%%EBji+pOD3&?R!PG)M6=+IiSuD8WK?Y)?x4u)iyS20y&aw;!>q#NBq{41BS{ILxk%=2Y)!0ho;3b -kdM9kWc6L0kd5UC7s)&h%2o$&X3v;O%$&nBw??CbO!_f)PB%(!{U6GKHg{vDg>*o&X2f)i==o}O?oud -gh5SR>YG~0qV -}BG0-ap?O(0xZXHlnDkWy>3dGpkA6&yq}Ku;`?5S$mJbWL+%2L~p)DI49sbS_hoNsQKZj~+4>r^EG*4 -;tzlKe;~kTO9Ztof9LE7oXnuc=3^Yg#Koer0+-m<5@YMB)`YK4%P06`yJtROtFi&pk*gL+bY>sd{wmm8BKlK!}ghuqRL>QYn~Wl;~?rbAwZUh3(Nv5MuU|fh>cKmWP_rj3SCnLiuOBRcTO -`yjB=3}Qd!VOS2D@GcJm$~-)tA415y`QUyk5LqOD`RecQ8{%kZ8T%XyqVgv68G9T?`e#CxsF}~@IUI^ -T~or$B42BTFH=$mg}!OWW^<&cU5PjUr*JM)aHxvn3l9@A$N1U5TT@x31NM+*&9I7 -9Abxi9v^@AISRUg_mtou_p4F~B%+eURC+BmBG(AEV>Z6h~t;iLRjhd8R4q21Z!4~P5+3awBk{AqtKiA -{6of-F6GotH-j(j2=|EkCIf{?H4{++B}?@P}T1=JRL-4i5t699ZG>jfn1JJ}AJ$;FJzzpgk3c4}*;)@ -!Bi1a9`2Vi|);(xDp6n_2OXp8A%!s5;t5pw81z#9~eD#isW)_|EQsL;m1rBpMphmHD^4KEXPOmRx{m% ->8cYkJbc`%+Kkzw#eQn$lU!Rz(jh~;NPCTB#l*9b^uV)uNwK|^dmOyb4-tqRN7j!{-&WRkG;v!$`kb+ -wWfzAv$D|c-{Is-w#Lmy^fed;bmz(D(PQ~s_LjQQ^FBgGuM9fMROn5Wsj%k -H63$=5|FEZMmr1nG(8#*Uh8R(I(fn3b`Oy}GU&Xa0%tJ3i`yqALy{Y7a@^9<~6V!x$IYK>&rB1J9$E|pLHPJ{yu5UVo-kK_inOX5s>Y2ae{iW?XB2uJC%16@f(TQWY_TkMc@}=)w{ggC5**TS58VsSJeTOSNCk*@QOQ06ylkxPEgnDfbg!TEjG=j7D{f{>iN -aqwugo#V0-zz}gE30%hx#;qMFhj~p)JkmF0iPuVc|@}PBpoMUga8yGooN0i=J2v3>EM?~>_;*XH3x*~ --v{i4|5u0t_9Lt|mrmDii{46ou_qAyRlIPs(2EsFOrjoNNOzj>m6i@VsQ@*xXDbm^8+l#|%P3R&UM<8^_KcBrB$}ag!+4afS?fbgwJ-<_(aL;G2@ -u%CX59MapsOv)GkoqiqKBHwx00}RJPua(=J;rDDx`Nr$S2(ep8~A2dgtJ)QmD`!VR$`{U>1Ykz;1UkR -faL5g-GyM#Y|3jGTIFo2%I{#5e!E`RORVQNR3p0Xs%krnw${0Aj+*+iS|-mQXP*=8U*R2}Z`!8C<3M&uycyc=4vT*F*%z -N>r=Oo@pZ)xoQ3=IccFNd)Ioh)1)ozh|cY1otP5cwu@PDjUbe)TI~P)hEwP+m1)-+ -HC;J|g$mneNe!Ur;-K;j{I+z1~;&L+DYlMVs*Rpb6fcFFfK?zS{-wJsG>UjbnSet_FoTkCt_NnDj>55 -XyNRQe%1K70_j3^huDAzjNj(E(yi$n7vXugKt)!`xF#u^%M1GPN~ajf>~sK2gaG^nwI(%9W0xs6yVh< -JfEKgNI9nb`N=F4&pnyfM4ashuU&L0j#nst4lYm5XRCUHbCP}`^dex_0_W$S0_dk%@+pFTdhu`G00y< -2+owU(R9Sv1T-sPG|Lq&;Z1X!K*%v}M9Bo}zXFsYa{vS|F0|XQR000O8%K%whIwRA+{sRC2Dh&VtA^- -pYaA|NaUv_0~WN&gWa%p2|FLY&cZE0>{Y+rO}Wo>0HaCx;>QE%He5PsLMAo65yo*|MFB^hV|qwyM}aa -^?8R=@}ZH=z?@wq!^Q*8cn5k&-Qna<)Fyh?wYj_wl~_?&$2Q@<7!0V_rOw{BSHQMQ-z2P00GF^0L_OC -nSA2?xpAL&~y4w?#n7uRSq`QM!Z8~YHW?)aSMMd*?N35ub!6~UxXxY4D=>1;JkLZ_dM@mzpZPM%Ic6q -2I;dr$eBmrb9Hq!Cprb9&`Mrv`AGhGmIaaD@F1k|P!@Fw@bFv7LQOp#I{X3U7&vV;GYUM)9m%r1$W@k&PN22i?MSQt)7yNpc*M*f80 -PxHDt<<#5ICdy*h2=B)3c)DhT3HBk>p1H;CdnXKR&C+b#JISBV?}cmn0~Czi@pR3G!nuGJYR~LD<-Vu -`Pyjtw>$~-b`SOB(6iMr+6_3#fK$^vskO|n(k9B+y=wuTsDH;MYW3a_-?~ucZAq6|N{oz#VQ*Z>-xXxaemg#1|r$X5Wxh*@TiSRv2}6WblB3&=U-U3&!dD8-|NtvSc)VFw15Jw})S`$Pnm{{V*u2vz48# ->30{V%Pa@<6&~ZC5YTc!tn^HtqZD)da=mvYXeSwi+a$aK#Vg6dS-!CSKhnj`>rd&cU#$UPn&&jXI9t -p|l1c1J96^E+u!`+V#1GH}L*~2b32={k?pk9=TUhk@6aWAK2ms3fSzA+?Q -=FeD006+b000;O003}la4%nWWo~3|axZdeV`wjQWq5QhaCz;0+mhQxlIXj>0_N<7Nr~b}esXk}iqYEg -tY$4~!s?lcjW!R3CehunK!OIqZuM%lapHW#e#3s)e#yyORb^cOvgOOfdd-=z8bBedDl0Q9D{m+_ySnM -Jrnq;})SK*Lzg=~8UA0-s4|aLem3cLV7VG^=T5a;~$~3rKt@&oXE1JBk8}p#-ip^!wbw$I^yW3s4y); -k9_cHjM{_CHgEx-KdXJ7uy^Ot{|(ygc4TmIwAysGl^s^F`y%2mgIJqL*H{x@YS_h06lVtu;bnJ2GyU0 -H8o=ydhx97nR{3)Si`TWB=BH@U}yoGq)m-kCw--^(g*JM(Z?uHIVAz~9Srz?FG;)8zKeb$MN`i)B@A1 -!c3mDlfO@@l(8bQFK?eAaAxRDu;BuSXRY$nzco@q^q|4Hv?%^Z@U}-H@5vZaMh}(O|z^FI7vMR>S=)3m8Ka7A3Pm6aQVgQ~KZJxmjPO}l-9xd*TMx&?Mu7c&Kl?8qRh -Rd^)6;Nq=@&;SI$;xfp<=a&;gVy)%{jul>t%jcfu=(>0fiI?(Rn9E_7Yi6L0nJ|T^6l6$*W`qdP6OHG -&D&y~?RPMhOMsB!uWd2y41_3xM+xwW6aDZeG5x8Dmoq -fhElxLUab+Hxb4oUa}Zsgn97ix44a4?T)_Ob@a7r*^vZs&REW&4=Anj#fUdEqk#;u^qI9>YyqA!)rSw -_|#PJD$j^6Yr(Z>`Mq?M(tfUlW2vo)s)|K3$p=1)Uz(U03P~ZKx?7=9M&4w`}{UUy_@^f=Cbt>C5Rkj -A1B$%{pK8GbbXPnuJR^db+{|QlIH6*{^NC|ze@4w6>Ve1|>bH(+5zycouwrDEV7)F`xM -jSFI#S2p>I*^Sacj|!^ioFVzjtscdXURxO#ma&QdDixzhY(&Mphha4b1Jr_=q`lRac-4{GNLd9QgReZ -8;m$ep$;V0Z9Fd)tPL+r8n2Sdz;%NPDZ1dB1oM1INEzXnE#Zwg|xhYK@CIvo;9p^WTJXy8&ak~yQ+$d -4tBWF3e?D=!M`r{RfYeOa2WZ0;5OoVsYNr3yD>VO>j#AlK}XdM{)MG0cWv1YcCfAa2r$O&-N$g*lIBJ -D$2NEpnxg845C}t}Qx3sLGK$9jr4jpv2RcI!0lvS3-&P^L%g^ah5$`0#S#7pMnE5#k^UWI^%9}R`X%8 -#==8dS^GXv($o1vj2rvVe=VeJe#LleskII)5U*!2!hnZ*U1X4XPH^kN@mtD!8HZB=*eatRE1<`{a3xE -M{{4T7m}2<7QphD`K5Sy_DxP^~Ww8uYy-4c@y0XS&PoN^UK*idb&T4pc;kF_g0&Un -%$p!!`$` -Ll5lO!G4zz`}23cL@B~e{|F$NF)EvBiF(-^ItXsYHn7?e~i+DGzkB;6%=#3CILE -R2^{8e*2M83I?e*L2oI3*0PWyv?4xm_*#aKxgA=@yJ~s)}J0H66=2%1UdIx$qbS|WK=|Fz>6Xh2~NAD -d7I5^5B8pFVt8c$_6!IqOrsD!kr63A(7Rqt-clS -G3VY-Okw61=bh&63-NbAMTP&o^*pMRQQG)>A5e2xnBEe_yP+#1g`kwsn`77ELIrR0aL^u4Oi#bN6zIO -1@eb_n5Es1jIVm-enr`6kuV}Goa^WXK{pQniN$MRzEuQ6w@H3;nz8+sd|=|MThoXJIS^xS9T);E|oir -BrPHc06^Tn%)4=Pjtko3Z%HRnKCXKYL>7~V>^;Dro6#9}vY5tga0TaQ(S6(N_|Y_b94YSlhYDPLI4~F -$8cDqf>%|ryWga23PuMUqZg_F7H+lReSr9RFuZP@dm8OLGB!%c@qP9aP7ePVW)jRxefjbSoG%L4j_@^ -C{w6|+A@rMf|W>yhkJf7mtXcj;~|ZyMi0&!!1F^B)T~g9N8=J4Bpk1Js53KVriga|673J;{ppjS4Nr -`(4k5C}T1fd~O?#0vY_!!5*RFxt=ZKySjtrDf6nj%~nm3cX-nP&oA5b(O+B}+)SJaMZ3y($|70syWPG -MjR#HWs0=RLbJ;^u0+j)qZ6x!>4oVslwNs@+PJ9->KPGGCOi14}=UdpErIA>7l>q92Ljvox*?$xeVrX -7@3nVr0MRNYL7LYrBTini_uaJbT5)8)L2)KG=Z?e{bdM+)5Xd!M1>&6-pR}}!j!?u;ldITq-05Wpe_8 -ZFxQF)jqc=Bhqwd^vJSQx2^aHClE;aRS)yNl7;@846C4~r4mr+8u{$Uv7ovkxZ$uIVf|`@TjCAE7Il-Nhdm|{HcWK%zO8zx7^v4Nf^DJR;`ub+y)QSJ}g}01T?eG2~(Z?DG;R -{s9h!L#Urx#;IVT=4F>IXia>s5)1#)VpUwOzbodjP2!}V>Q)2a}0sWG?A_9jOy;vW=!SWmTlwe{A#?r -#qWQ!3=K4l7+(|@Nv+04DGNgAp#!Xzv$emlEx3A2jdGB%*tL$GRgNY28xRQ{Y&_E%s5{QM{Zn=^^>!P -~gFV~P3nA>QYB+DM{qZvL3YOfzSRZyz{`zUG<2%+$QcTH=ctlQF8h2&(WwHLn_2P-F4n-NjRBrJPGsIuLSUye*>{2 -x0uf8A+Z81$DOVbzYg7jmQK42ix_(G8si=jyh2ALs1q>OvQxm?QuwuU>!1~#4^#=egM((Oqct2vdZJV -#F)SNXp|Ut`za=k5{UG*^Hzn44$(bbFif(;g5+KGB`u0(WoDna_dviXZj`-pB2ih)s;ChBX-RN;^9R> -NMsVc$_zvdo=O3S$ky+!E(M7cvFtL?QnrzH`xO(^ag*iocc`*{7nn|DYQVtRoU2QCNHq&1vXgc@E>mn -BblmEwyhZ!Hnp9aK9Af8FtE|%%e07Bvi-24r%3znua3I?!9U1n=1c*JslU}nhw)JhzX~&i{zwft)en~P%n$(PlWr+{R_@8L+RGVhx>mif)u&c-Fd3lQx8+f>~nn4 -hUX&p=)9Flz!c?aFovlH^f7<0FJZ4pjDz4T~dW`YBswVbmQ&t{4=zO-FrA>xeNZa5$Wqfr -e9Zo}yrG$Al;uK$Bs#^hGP-4eV@qHz? -kJ+O9G3f{Xh0a5_jtbnDBF0ITv+tUgg`jONzze --k4r&SYV&##9~`0F>!_ih^VF`pVMA3KHqobZKmLDn`OlQKX1f1mL-lQXwzti0_kk*C2bvtU?((wtys467rtOgmWrzBUk`H#^!V -)XDNMI8>1*hO4h};7!KrE1e@a@IIU*~2f}APPh=*QcA`7uJe9U>m+Ab`wQ;ZVcp$KdCMH{s+w1=RoB3 -kTgx-gkA+u}0s%Ikvj@hG3Sbx|o3Ut@ig(6~K(dAj#AO5ixE8(({s%#Zk#)NeOUyW9PZ4LLa6&KNUaW -RDWt7d@IAAP*9OToUaCQ}M|{BggiFq1zE~1Nn@7Op{mgxkx2+cZB(G*1Wfq$ptRibo`Lf`17juo497R91Ik6vSc+T$Mzsk4ks%YB&3<^sk#WlwNnisM$IJ2Fjmp!?{msHz&HlB@M-l59Ey9IUEx^>#kGE{mI4S -6^OM#rRWt=hKM^AifjnzF>RnVcOOB{FmYu`dp(~y`?~A*xWA4ccv)A=no^L5|j34-aF3!^cs&wWh-)s -eNAkB&Nlfi?zXEdES1s|ibRu`XYW54OPG<+994;TytcFb4rD_yet$9g!~3_z?T68cTR~g7CB -x!PwSzJo(T=&`H{leO2C(rYW%Kz_Q(603wa)fi4CtQ{U1{sJu{1GCFn{@O9e_hrCF8ajw4RD~irWfc@ -L7J)PnSu*Jn5GpI?zey5WAX%FUxils3VftY`pcT1K~CK;U#=O@`nYQ2o&(g3ghJztq2Sq1?`=dOeT77 -D;j^-M|s^wn3dP -B;ysMB{8#b{)ND6CS4Gn6I}W42|$ZFH#gpr>lkp3XI%W8kep7VB6UxM~8uUN9?_56R}UmDg-@DU5hex -KrdOQ(sd#OeKJocr3@mh2i|aOA5-I;nvX_lcXU{9LCt6?ZX`$45sRy$3uNaDt1XBhO30E+GvJM@{93N -_jp2yAr8`*t$}z^UoM2#m-Y-|G!Hln8bnqLXrv_9L=$qD7N(=7UgU}`}VCLBf{&!!X2K|I9jW~~Kvro -en%X&-&lvyDaGpO5%>5ohN%ymb!*|cF|8~(g3j3|q2kr^xyD!Z}m5RBdloU#aBe}3SorolJ$=;oj#c=_0MLo&Jj#Lh0A@Fg -Ub63&*?1|kQ1J|Z0h|k+X5Gns$)TQ4Xj;7m8hsH562%2@RCN9Ipa6Q%v7->G|^ci9*YThtxdSXK(Nyg -CUvyEwRc!2^34@TQ+2B=IUcP9MnuFnSD@c!`D#_{Iuu{!IllEe4&(?(16>XYr=b? -p&)%?=;o~lTII21m%~WQ&0Sjk|v6xM>m;WU@| -9WYH{9|-wo}0JP&$%7Kyd~Oea+>XKn$F<4h_7Y!w6EAUO?-_hbODAaQFbQsTcv&U(s&8kGiHa?&Op1x -%T)Cw*Ifu*&0QQ2S@9K~I#rhCe14kgAWp+Z2_7n=vJ}oCC6L+l6ojC)ql;OTq1_**_J}6Rup7r+Yy!FZjngPx|5LP+1fevoxQObCbfhu_}?X -91Ei9qRpKK~LM7V)Z~-+|F*xgEWNIvqlTtPic22~FB9sO()sENfqRKDZ`6HG$9_C5h0JP_SeEI6zXJ0 -;j{mg<3Pr#aQo!lHh^4ZqRIl+5Iws?dPiV%D?14}ULWsBRgrrb5V|f}N`|NV9*`6@Q;mcHFG6|YHF?4_^-mt}K -%;IYs?0Qen0fVf@aJZ~WSX64er4HNTjb5^%1g`F0AMT;47K6KqF=Q(423fcf$$auL>q55p@|1w03p<= -V8C+%76ub40!tMT>JP8MSQq2@{ba5O6(U+1ebL*-Vh0AZBn*v -RQ&@gemB^!%WQbqFjK&}wJg>MBTAi0-R*5Q)rHyJ_L)qPvqRJz;XWBhZYCWrvtS~%0jCZv*Lw7?PW4% -Wfs7vD1;VihgGFW{psO_xkIk*&0FXTnT(f7i-&kmATr37PQ#(_xYs?@$IOnTVl@0KLKe!2kID4Z2X|9 -Lu+$JR3+doewBc=iUBoS3A-b#g2SA{byl#dq7$#ARKn{rk2os=4-Izg%HKGE9J+3GTK%3% -Y%}c$@7<)C})~rRb{%P4lKB!B+PkJ!ma(?*j7!!))OxejZs8mgJ2WgJJrRgORhuRSGo23@(IuDxlg&G -oq;-G_lGz9#Bz%^q3|~LhcOLc=&hfZx_`lRn#stQEh7;wqeY+{NVOOjTPzI^RG2ju{MdjW8_V(MyHMF -~m8?+!?NX-Gq!t}ZPH0Eji?zWPObrwpactc*W!Dv3V^1~GG^w2|1t#jfQiK06eM@<@UYj^`u)qt-B+C -0yHDdRk8Pb?oTR#tMyxIvMAEOvY&H!gvXLs6mUsYp$)GoIB4St@YSla+wjQ<_m`(rgvzXA=PX*X0i@j -ba;P6Fe;u9-s(rq*a1_WL0Qd-iD2=eN#@SwZ*|V6VC5@&Ga{$%@!(w?|rz?^dgqcGbSs(gCwj@ -tC+>+0wUzvjluXqSNjGef+3icgA;#V(ww|vvdVDZ<#}p3VTViG3{G$C$W*+q<$6JDZZG+x59o;;6U6W -oiql;C<~oNBhTle#!+UG+aAJIJ`ljVn?+38~4B)!@$fIj*G#At9=Q2A);H7FDl`{Myx_0<;0~gQo -w^__C+j68=9yWd0kT#aRxqlj-rvVv;yn5Wu7nkV$!LyE^9dpqQlRMM_z*1Gdi4gA$%nP^7H_0PK>WW$ -%a;?H{SV!>SZE=V!|Lld=`c=TrLPdj3?KXt%D0F=0Tad@w`><{OV_TfjDiA37RTDp-SaHb3l?JT{ot& -i8O*_tG)Q;)b(0Hu*Zrxzic%-RvMST?Db*m0JQZ -|_ZT-lIV4n~mol_I*-4GvbRq#@)9QP347G*Qh&CDOgG}xG4pOjTXe`7PT{U16YO6XZ5?^o8ct}=hRY> -+;@c|gB_`25@?kquWqPn*18qrWa1wAlk3a&?k}ez;_Y&R6zMFt3fx6M -*^tfX{m(Ie|1r-*qDsIou}!0^Pp6$UCNaQclF`pg`E7{jk6QH`vBv5dO`RocpxB*4W#!qq!&>A^K#2|g?upo7?;3LOxT0bKS})A5G6~e5B=8 -%-=-<~fnteppM#hQ;(zonONK&{(AW%JVzX!Ab01)kD)dBL>F*3E>wV}8iQ>2fT3F -K^4p7?_nh1O|qi5zL^A=gQlMd&PD#)_S*VZ-7=XZLP-T+Qzkh0TXNQ`DxZ!eAE$b4js#yXt2U}j_-Vl -VP{Ld)%(3;EveY?{pVa_&V}nrb_@xjAhyr}2e{fg4EX)iG -W6Kk5_KXe1}k(ZM9u>P0fys%d@|2Q<>j_+il7j_CF^AQZQatjpPnchjkwb5c0O0QPcwet-c{cxRU?@y -8hV&L#^B!3fB)bAip6!mfVpcw(d3Z`Fs#M;;0BOj<**)hA|J~R%!(k4da=X#@cz^6QyXgfVYKLZxmPw ->p`!~S^gHIoA&%h-^%}ig`Sw<9h0OApV0bM?@z5LYWkSHTqp;VIsFJZ02clsU-(*Myy6WKAFVtv~^dg -4LKH%p--@QomzY&es7sd%5#1Pp`2+8`uh=`RgjF1AzTc^$~u<4PdFZsJR;aeF=Zj%{^HI9PPjU!Q=4YnJ?0MKT~o<(71}~!OsOsOay;BkT33f -uq`5aX5(Kn)Ai_A)lHB{yaMF&I9G%&TQp{dD(EJ%-G(=IUg=k@#FXP -@;Cf;J%hPjBA`#!Slk!0Q7MbGH9p95~Kk+l6likkZZ^rEN;Tnp~z-_7c~KC4cQ`5!WS1rJb+WI{#=JR -5CrHE;9FQ3kohwkz{0;Q^S)qyJQG@$P1>vGY0s7Z0TTmL`bClg)+6H(!iSJgRlK@K*e_>UsUC~f!k8Q_>q*J#l)9ifjPR~>X0Kbmd5`s<-r -CIQP#9T^N1#d8JAL6)1T^I4E+*Lf{2*tt?{58N=34Z!Wwn=?;dWbU)am0`2gpc8XaCje0@%7tc8|DUP -d{+hfsG9p$?NZ6%i(x&BDNo1epte9C|K+q~&byNc+&ocdjZSJ&wY6c6pgHJJp#yV>qlhR}uhqdnI10S -PkUuy6pM!;zM6%qX{pCTskIaJ}&}wLiJXKASOc{X*fDw{bz~pI=a;M~NvcaJbHy#X!F;LSAksovtjUe -zQq+?Eee&PzL6x`UigCwF$cDE)llv6g6JIT>5MW8b@0WAx{ -{cO^an1KLpO?&NO@|B3aYBUSyF6WuQT@>idNL~dX1x`x?nbP3vfhnXj3i@*<3T5AFL!{M;IUDUb9h&q -J^+O5LAc1auCft;b?u%ZTxUa{t8gRvlw)%26#(!U`r@-0}jJ+t}Wc2Tz=v!C@iV#A= -HC4my+Q7KBq%ZL(#KBfK;G-Ksq3SoL;;)$C#%BbE_v?j!)4dzJ)}=3R#+NkYf8s^ajK=4hiHBuGp3#+R8RVL`WzuqsL9 -W~yXRi0nZ`MWf?j4AlJ3R)xUaK2`79&mor -(7|t>58N_Q=ey=uIszbLAst^yQ`|mOvhqoYF>VX4Jc>&(?D7Gpt?cdOj>rC@}qHag!ObqQbw=i|(e-q -f^s!`jHuzq%g^4yjG^+M~HxaITPECd}d`_s4<>Bnr8RSwRiAw!4bGP%O>Am7Is#c1mVM)2CykD!b<{@ -3=^9_Wgn%pgq?EPlm6>aepG-p2Oc`lnGjT2h{Or -*P)4ANaSlS}=~}W&h$u+2F5yex3FK&Bab<>v6CEYiaGD|2&x1vq3p1qk4l^Q}pxBhGp+q-Mo;rIFrOFkiGfp&R4q(c -VXgXlQR=<+Tn=`)#uD*cop(O(N@sst#75h5$;q&ao|6)6c9K3-}by1pF2Dc0L>?B{8>g%d(NLGpWT+rwsDjQqkoJm7Gsjt@O?HnZDVe?$;UkWz(Lw03lcWhnW~?dpx9(magSNH=! -#Xh#FSK+7)pMZl@!*=eY)@Jhx|gOUZ5?xUf;ExtYl^xTPp(;KpM49_f>r&a^L%CDNy}%)hZuPVirkz1 -xEj8oZX*{XZI)nFiMqxg~l&npLqVw^Ow&Y%n!<_+T4aSb;t<|w((M|_p4%jxq>swnM~Z2TiO!+`*SrR -lKl7|v(FZ^k)%b*ZZvk5+~rK%|?jJZ!|p_^h)-##h&u9PO)pKg8gk5P`Z7?|YDE0()qzNeqY3*hT~n`msCtn0as**Z?AKC?(I -lkYiYs;L>>ZQ+IrfH(>8~F~nt$YR=9l5u^2M(#Vt0$4QWfj98I4mhf<1$a`xvdrnXmO;!l7r!=#{4Ku -${_`?u?yJF`++p@|yUkwGjR+VYSjDj>_ZX54YHIyDt)qXx*?z#`_{3j%qUQcA9ae1<)951_r2$n|e)E ->bMmsqIyMrXozy#D^Plrx~}j{3u-eAr{hYiAUSc!faP*mwkuSY3t)QGY1Z&UR+Re|7yx>}ZZQ_5T$gL~xb3mR);Y!p7afDfV -93jDj%h(Z?DJXn7mU4u4L4t-p%4Cm>6H0qn*Fe+;Cie^xCYImXjY{Nf8daOv#J>~HY(c^dm@nK4jN|# -ek*+gW2$)q&=qc#2og3=z)!P>k?>qeZQC`)6_rdaNlm)0tNpIS>KNBJ)2wJ=OIVdvWiSL2+5@j%mU)+ -9?a&-6nZU)qnPtyNO)B;B~*50jbG3VPcng7Jeb)Y(#53g1vck?he5{ZHW@7Eug$AS0=s1P}kfL*4TeAYxZz&WX -_+TqS5P9>B=c!Jc4cm4Z*nhkX=7+FbZBL5WiD`e -tykY~+cpq>_g`_)UZhMB&{qQ6CPZ{HnB{j!uG8>$EE2g&2zci(q -Bi@H(TLZReVJT^=lt`}RaYOse!$VU)<{M}6om!hyKTnmH0f3k*#Q@ClYP?A*(zbY&FmhCH>8+)j@X_s -GK{eJQBdY#{`|GK`q3sc>ccE%mOr2e)(G<>mG6csZD);nJF&8O6uS(U8CCJ@H-G<}Eu6^<{b;;#x`NX&MY=Io2Y>EUP$A4@2^UcN;6ZJ}{*u+WOi8dR* -vR1017YafdnLJZmw}~l;H$7Zklp0Y0U>u>3k&gsQq^eJcVQ&pJG;$V?|IfD<>T|ymT(%8A4Ox3=Ryjb -=W&k$ueNEovQ6+NqD;HndI==*R9>0=1&SeoiAhHHE1(&hylQ?v=3|AJO^j{~Ov5AH;5xkPRtQCV04TTyu~+8tu(=$H{*+O=WY3Zjko^! -}5K;Ae>Nmi?`S9K*SdjuhV$E;zdRPtk19sBvD#!{h`2dIb6>&Fx;x;7UNisk!EMr^*flsv|iC)caIaC -JhVS2rB1>D2p04YOf1EU^IPx1$znJ$d#dJ1o5?EWv6IY((r|Uvrwy&JDfw2p=(@ddZ3+?5CAh@A=t^#T2Jvk7;~H+tpmI1`g0}(lF<;fd)8&T4W*2lK;0sG2eI&!VTeZq -9N>II@xWG5VNnB7se((}|!fy+XbSg_{_|1z~d@F}%v!xsrn6C5#-H@EbR+!&V~3WGi110K(}2aHgT(o -4eGpA!B@9f|GL!ZMZ6RXY;oGw^&)K(8dcd?n#sOF)W^3CBKttM(xQW=555v|n2?E0gtAaZU!_`h<|9c -I(cv;-4$6m5#5z72H9TbUoNxhUXt|zdMm|nrqOB3eVe;Fd_hb*oKU`BWAkCps#b&0`xWYL^J)FHyFG1 -t`i#rJ8peO&5iGAtWSzxPuw535AJlz@x;`EIv?tEH>Zsg>1X3l&2T)9mp|gUyBb^CoWHrJ@4Z{i-t2P -gE5yXHc|Fg#i;>=WW6O*BU`+T?+2ZvH#gVbvyyV7`G -qCv$F}-DV&`RGxRj+D2Jn=>zwIAMus!&ez`0|MNj@u(p59a`7`-MpkgLVD1|AQE|WH+=e06+rdz|v@X -U`#TPLtu!JCuvzy!v9D(>i`bmbWJir&EkYq5!*LjkA7|0Vc^AQ93Tn&;$Jl;<`6d&~1^3Gfp%nmyKRN -q>b3^#%WW_MB{wP`_KFg|xAc-&eI{@h?zI0|XQR000O8%K%whQFZoEM*si-W&i*H8UO$QaA|NaUv_0~ -WN&gWa%p2|FLY>SZDn&VaCu8B%Fk8MOG!;jEX^s2FD|LfNi9~$%q_?-Dp7E8b#n~$35gF5iS%&|=1K! -AEy+mDO@*ll0dcsvxS%Q(Y#|~Vs5&*d08mQ<1QY-O00;of09jirWX9x|9smFpi~s-~0001RX>c!Jc4c -m4Z*nhkX=7+FbaG*1Wny7tYc6nk?LA$O+{kh7`zz?tgE+d;oK}~I!{8`emUMSQbdn&+J_qXs#ToYO@^ -*$ClN{~N2?CrrZ#EDlh=at4AM)|wmq4(c0Qmy)UlLosJmoLAs`|Ux&6(Agk@L~M)v!ORtE;Q4tE;Pv# -d1~EZB&%aip|=+g1;_`lFhQ&F_Q^ZBY!oUW^Z0s%P7yZ3l=S_+p1cqVnwqk -W{hXHbv9$uY<40uPqW2I+D=98W~HHjQswIfdjv_N=#i|3uc~ZO9?8P-UIfqeaxiRq%MRx$^F1EK1sO#uM(of7b;C -_Sqd>t)$!8qurDESqYxE#EBc3@l8U~IpjIk35qQG%+ -bi9E7fGuT_2keNQeXP?*kg}?c1W|RBl}wTKKE8;IW>&3KUu!}!r8Jq}ZZs_(;)~4Klx=pVa_DU&{=Hu -{1=2~?CssTcjNkrUyo&dMVe3)s|d^iH%t9BOC+IjUo)+PFaj26mVCt)gs!{mXWMaK-M -K24MCuTo>BqoC<4g3(FB_&r4iR2#GXN3(tX&His}0cdJs!n$se0kBvgelQcK3NH94|1BKpypO=BzS!a -BPOpvJ|>ph-Bv(=~wlF4ZvR+sd&&Q_CX3JW^qeV9S-d1iB5oq~*>L5eS%=cG%Vovb`U40+?7$=)g7(6 -Z$z&!vD>{ukmd!^2Tzy^l2JmNs+<1wrr8qUtv1wn3T*n}|7OWQX -WgB2U05N5-CIiwehxhh45-WsSwNL>=X-EyBSkGEagBCn;P_Dvphv!F8C6a~{x$H) -56`j@~2`U4%);+1BH(I(CgtG!Qo0gy=#7m&$qCrk#0@AV5!tz&_un_x2t9`pLTf?4g_#l01aOFw4e31%h%eBwHRUZ}_l43XNYp0mX -9Cf^m+U^Iz2`nAg);zvlB}|W&&CU)29N4%3xG+X@Jy7Dg?%eNvRXXI0Nmzmb3z5kqC>>763ZEsNi2Bf -sGysBn{e-Sz+LL(Ew5#4i<_=PclFma{6I+1YUQfKFQ9C<$4*3d>Sm6P=Ge<4yFvq05qCTYxV?41U-Yw -HHjyp7A%`O)6^jFu{qR>6&I_!U?oZljM{o#&K!K86(AE5#bMfR(}slh(EMY_6(v#qpBHCR*5uV1M}sP -Y`dT|Jp;Y#_)ttYr20GW%mNU}UG7;t@UMv#DaNeUBKyk@&=8=>0%8y*gNr%z9|={32O34uqQ>eZVYiO))Sb<8Sdj7*lfVWnXz0zL^YIzj%+EFOF$w8cHFc;@; -brZ4^g>OTeXHW8fsfIaWAl@<6^6@Mo4<3KvV9tJ}#Q{v-N6&doDd17C_PZPzzv0qg#;jha4|d^*C*gt -Lj7}gf@+I68RQ*Rpv>ySOcBy^X%)>SOgN|w_T_s&67?_umTsKNhify!4Qv0Hz{0 -&Ghl+kC+?;X?TVJfN~|}BHbU+s@`s(Rb=gG!)J29}mQ|CE^vKx5kH+*VR-05$p>(#Wnsv?M0`?!E;cE -7dq6;98vZG}-!BufqJrPrpnx3X;mjpI{`V0#;EiPh3$<r|=|`wP$<{i!fxOpUCj|m+b56 -_EJ47+JgRFpYq3|=eBQ9u5!}mFfz(Q5x5v=6cmUdA3BoOv`6-I^!HAE}&}AVG*oC1o26hGLiLN6S<*x -0;1iR0b2!Sef{)*G#nk_+KAp3!sDfJ$bOEW_`;<}-q8aSU)RO-0dN);M*(W6d;vBw9r`s~EhFQU^JfR -|`vb=gd+8hUiEz-CUr?ZaC;?F@k>9cK%c)pU>(tye&eb(0-2-O>fA(?yOj#Emw~a@sApVh1Kwb*kt)X -BjA#K-gvt8${3&b2eQc9pMH8xA;|kVzkyC1D)Xiz&ntKdL_y@2`aBiF*9U)XL8dt^++NEt1K70O;mPay~$SCqt5K_h~UH{?A_9Qn$kRn@{4Ra&9lf6oRgh -pr6DFsr`9lxMOrB3oBnn50x~Yn06e|m127_Ez-$fc%Ch*}6r4t{pcg3eM3s0*!wTXFI&oCOdUSwiU6) -}20Fjd)8WjU;zvkbXb|`d=piA!Q$q*^sd1CaU0uHKA3qTx^m&+vpY1uMT5b&uMxq*T>uy9(!$4*AmdA -E$f)dsn@xAzToX1YgzqYqK)aD3%K-~>0|%1B>UWH+SuZZ5q63Sxzl_E@z-;L#`FH1fl>@m-M4hjOXlb -+QJA3(uw2Ga{a`zK|TTPewqfI0+h@r++Cwbh(8L&b!G4rS6o!+hHOsnp{x8vGfdid -PdT>x0xdb^NrDE_Bl%$HBk_{47LL>@<-XV(J&)Q?-I5Tz+JjD{KjLZ#Sa!?QAuiyV5FP7KFm);%X?UaYAatBdrCGB6D*q_CeCJW!5d`s=7MLfhnl~nTdu`A! -H+|Gu*#@8*7q+9d}CLGoS)A_b(FAi=Z!jZ(Y>5%FoHpCJcTcFt~~ZbcYc;8Fb(|RuiI9LNjnI6uT}C{ -RA6@4jwExLuUo*ro -P2{OGbm;z~vc7t55>bkQNCqvOlcbV9!IzrXRW6DJDZ{q6B&|%32tf>AGdz$&0!Mkf&4PAUZLx$gJx6CfQBd!8b))v~IW_# ->7H*|ec|{4=!Mw8nXihQ>1I4jBC_=&lPGG)4yW}MUMI={hM(6~1W}b*@! -w4N?qq~`%`Fakp@%c($`AT20^%XqG8z^=e=(!2%`;D5RGg;;y;b|uoR8F)&wQV~lYe%uiLxa5Ql=mUz -k2`g_Iul#FLt(^#;kgdqU?rqzPjOO2p&B$%)-H_!{-R36{OPI -de!G3HmYw9;T)?2Q7pqBYJplb8!U!it#+WS&@#&wfSM>S1heozZ^wdrY7{>$auxU)Opmtm}lDzE6)7~ -Vd0=3?KzDdhh9#&%F&?j1?`q-Y8dovC?-t;q9`wGP4UEgkJEj3aA&K=j_?18Q0&wd?5k%eIqv)4JwYO -5pFC61an6q?9795OIfyO|_0@SxJss=x0|6rzFtM9xFb6Y->td)^%M@Y!j(;#8K`q+IZ@QE};92;u?E; -`;+_k9z05Ke0o#Y@c7^yu@+5|v0?ikq$w3l)4_;eC_x@MFkEoSE60&wUBP8taG}1kHYRJIM+DkeXFLU -bS`aQ}O*UsZ58{+d3hjXoi5)NMn^k$S6|SB}<3!37!K5jUAzt{)jEfrc7`k#Y;OO -T-7<`~jKnQIZ%`hi)fdk3;L=jA;jA6V%bx2d^`h~1TY++ZUBu{b4tc -m>5Ekdl7&|f>#aWV%s!IQctFl+JR1ziWY1|0R?DB9om`zsqK0|q$^A4ta@jTowXt~dIHI1j?2f{^+AE9Aq~giaQ&K7i^d5?(5$MMy8CV(wGq( -M}G4#uRX;LH#)fZEz4k9G~8m$7!VLY2>Bb;MCK-w;7H?`5-?+(TGB8r>7Z0$ipP3y!4G@jFy -vGJ_Zq0l;&U&AGZtuRfmeTVqxQlO}g2AiUR?ab3_Y`=)!eK)$g6@-ip1yJxRSXc?*vG)3vgb?*rJgaz -%($MMLIg4Xpr*IYj(&kSMca-J?*;d0XfmGd6f06u>^q1WO`+5R%AZ)sy}~XfkSI38IHeh9H%7Kd -Xu|HY^n*Rj4~}ig{NO@e^ITi<{afA8EBLT6LY*kCa+gx-!4!c+%x2|4=Fg@H}dt47BD~$p~6N$YEvxG -~bOTof9k(ca%dbnC_k;C|S4(EK&t8IUFKaO#q@HPXVuC`>TqyZiSeQF5{Og~_$4~j6gFg51u -k|aXR4JODVI+G4&L{1p+vp%Ci$RF2y$eIl&cM-zRqy~#CbjG6`j>jN?v6lJs}N&nJ0gZSEq|BPt2Uxuw$ujQ**PlrZVWzz5VR)_mX!5DMb7i)%nPZi{xo -lEH!Y;w!-%8Wq(LqiCAh8$1%w$4}jfPh9Lu_WS85HKI}b+hR_^#kB;zwlrE$a1 -nxeyTeE|MyZ0X5I&^4)5P%+5-!aW?TaA?pfz-880(z81UvA*PY!4LB=1>h`n!56gJ_?=x_#C^)Ta4Uf -9M3dAHvEej{5N+=l3&fB?`1fMj>YCIY&Q0Ju5i1OwO|_!jtoo!hBZE9o+DkI`Fcb*(Jj8z?Zg(L#&wM -O;Ng0DS&!<4r*g!fZIQGgnN9Wk6$N}=NmBYOZvIb~m~76Yjr9x&O)$FmyUgknplT4s!FS7p;t<1sXH$ -%ssR@YTzA5}{K_NrAO#?q>aEyMr=oN@gva0;#yJdh#UAFRzq?bqk=9(OKDO$O>jn{U2gdf15P=Q7#tJ -!Q_*DUYDUM1ti-s@%vj#Xt_scde0deY+#-<4#jDehwZ)ECXS(q8Y!|j|R%AImd -`j)YZg@9cn9k9S)m9z&d@YFvA2cl}D%S>`EsH$jVrL*eo7fA%BlXC=C)sfYj-GnK(gNpNPy#bL9!#v# -WTXj|Mek3mMSiR`9Chs`0T4nS+0E^;)!_W)j$4fNjbH-?ZmFQ8rDeFdk87mP$~@QUz<@Pd1_QAy&nkB -B5;r?0WRY_@!Y@KGJCi{6HislCK4}GKv~Au`zS;>$#Ye8lM@A@lui1Wr)n-930c+F~2U}O4d`rz{Zp*gGh35>HWk- -ZvHS=o?yU0|%#tmU8YqH|ojf5~{fY6|b1fsk&Fr0&Q^+Db+{((9x -wiA$)CT&1HzN}NRk_$id36fdaa(#TqYX>rx@LIX5>uS#g&pyuWT)grpger&keOa$WDI-uiEo`@$9c(v -i4EY8XlcU)(Rg`Zegz|Eu5i>aEti7}UmFaGC^A$Tu0}>#Se{WbhrQ=THN(p7IP2?clVct{jm{czO5)L -M=o6QUg*f{S$K>p#*^Zq1txrUHTjg=twk?Hx?c{5qVKD1i_0F}Lc@7_Z7CZAae8T-8Y$wZ(iS4kn1XA -U3$O+Nb7|_TWaI`R>^xETik$}cy-yQ7+X9_seYX1d)kfv}o;ke)r)YR*8D{hOcz#__Hkmj-#q?dX&?* -!R_*>|OL$^EQ68pZHAAMi}q3s)(-wqhbq+^NB=+{AWebFrZ?Rm5);01ItAW;CnrF4j@ -#Z7g0%lkaG?H0*W*qmI%g_4?PP*b6y(UX7mC=m6sgm>hr4+3f4wB>1=AyAQ(PwRhufjlfHhOQ^p#lez -w%hJ$F&w`(0M8c!rO-)XSFu@HE=7va%Ly28fmeZbPS8MffkPFMSa4slT(ui{#k&yZEW9NuooYyEjiP%GdSHrrMo2O#b77%6VkND3t`TWw#-eREB8p__A@^JRFDkv88)aPh_FFZ;}b3ujBH{7BJW6%yopYh|*)+^_(Q -QcfRo(LQH>7#1ChOg$kb-)xW0%%{|Q`)hsZ6=P_FA{hNEaq4&&OJ|x`fmV(mQTtHnnP5E3P9~N{06!VA;Xw+$5G5);T -O+;{6F9SlmGqc-$R)fKl{5E|M2r){p&ye`p=$Y?hk(X;+H@EzyJKviy!{QZ~o(_zxmPMLU~C4;`z^C{ -NT@3;}r?geQjf<@4#Z}yHRF*`*mJbaTfaMsrL$tYgYQ4j1{)4^+3m?1^Rm2uk4=It(S42TkXz#aBO9KQH000080LuVbTbBnf7o-XR0P7?G02lxO0B~t=FJE?LZe(wAFLG&PXfJ -efWo0gKd9@nLZrsT6K3~zHb0F8$KqEia8h{2ITkExVZ7(Ek_AnVR)Uao!SA2mFO9SMP*D;3#@ghJFAO -Z3R`6PcqJ|R`r-DEe(k(8`M$U9_rRdv0qx_j4Up0avZh-|||TI6NTUe&ye>b&&X*P^O@_O7T!o<)hz& -YL3PlgS!!reTujg~q}6a4kxNr}rePY6ZB;vno%x27N9+ef=WIS3W!CSqv**B>a3=aQ%JCt0vK;zHf>h -aHmPVmY;9Lwa64i%A=|&+r{3<&y;7HM8!Awr2)?5r=F~L9iwOX;me=1I$B+-cjV55k1d3FGI`mEn4nl -7Rko;V(5CgD!q$1oB31)5;y0@rAT%7uX(jk0_V*dQOGPc@71n -KK86oQ@c=9p05Bx5OM0p5lZ-onIFbh!!*VBF__x>a(QESuCm|iFRzY6-f+Quu0n1*^13Jfq(S~u`k;> -LzS*rqb@z0=oe8^SyVy%GSA1Uf~Y7o&1ZSc_mTo{Rz^4dGGMMlvp#cvrGD=p9MJDC@zk2Z-4vv;n(n|zCP$L&mYD6UHkUWj~{*mf9mTm`fD8Thd;G%|M>XBAK_1Z{Z)S{yrmR7=4%#)A`^ -8Ox}B%;WbNA#Fz@HVcKS2l_V9U^jzBBYZ~Za99RZYn=nwpG&d|u1x4^ahp=-7CWLkgdC1U^{J9x_8<{ -6(`f=^|{2U2c(O2%6t1+}NlfDBAAZ7>}%bE;ws%(6>#o}~@e24K+Yzy|1|D(KYE2hkxCbc*QA9>wt|v -$(Fsi`Jqj6zX=5j$L$jIvJ5O{e3(nM!yP4WJm*cLq5w -O(sJjlNBukM(&6-q#NkY}54{8%r6LaRM2GP-uQ*~pig?EfRU^eeYKB%ydRw-HUGZwN(mi|t8dfCg=1# -C0|i93b&ZoH|s?Sz!M^BaB@Vtm28>hUSuT(5K84X@ieqU4;mV;!L;Nm?+9?kNtr-dMuLiliyHeMj<9; -}}SR<9L|EI?^qzvu~a~KY_i`ySL}B-hK;jU%Wg0%4a8UUjXSc$~N46XowIh**Yn^IW%t?UGk7va0M9T -dd@^v`&R!z_EN}p6hj!k0er~bdKI|`N%(t6(jzI(j9YxbmzA#=+X{OPe3Y3;U1oc@AeNVQSV1HVDM%& -)Hftk(dXYocp}296y3q!`64yJDbBErI+!&4L;?NW);hSi+3$K8~{Hiii!E%gg6?xiNE-s;rVAzt>TB+ -brpcDv(5j(#y{K{Ke5#b65EVtc-y9*R$>3OE-cgP_{k9KX^=eqk{*fTQD9kKg(up7_cD|sYbrSV-fS| -EE0)|pC;VQlKY61*}s)|7JaBYj%<(NZ8c#ffPGJajV9Tgj$U$1}{;65*8Is@+*p -+}aT4p+wqyh>SqFx4{^w#o##IG8O}g&WTz!^IH=@C{dziN&<-ZR>|^1|AUupWC4z&DJ7oY);8(_h}f_ -1LoML)S+CED*apZWVq5$#Y2HtO<7|0yKm%n+U)LCp0$0sjmijqQS!C8?(D5nn7!Gcq%-r_yq|1z2N%j -Sp02pbDLL#86e^FeBu}!rQYp$Gfa)~Z6meAZAt`*WL|ktJ4VgyQA&rc}9;$WtPXB!A{tT( -dSh)gMEgd2SJ^;3JGIf2^SApMN2P(*Pn$@fiFi#45&WnhINjh=Ub;SkPMNei+TgMP4QOmFDlD7!vU#g -YUX=uTPPzLiG^sKuq)#;5Vw2@aK#&NFp^!d@!bDIeX7NQxoi<4V?#s>cumn=*&2KGrP{=;oD*kzk}=? -8hC?ONNN3tQ%l-*SJraq$lqUnctN0zoJLi6>lA$|7gtr?_6oy9H6O%iq+R-&{ -l;OXuUnAX_ETw9LGs=Kn3uJxGTuVd`<8UUKyF(ouj#!lr%hza5AL5Nv(-G|PhD$$VJ)SMy2$IY`_d)8 -x-cMbbe7e1#ap`jF){nM(*yNHxs#X1QJHJ$9B5hLY4?CZ@9*(K2qQJZ*!VRF^ky|)>+e~CXq;se{cbN -*B`pXvme{L_?0ot}dwZGrmvmz)@JA{u~3p*{TH+Q8k@N#`~Cl9l*W$5*8kYH$UDmna&J?yWgi~ -&tmNmO6M^q6d|Nr(oQ2BL%+G?T=umHcRAfAZ_X%>o>WF)8JJ(++CqXE>OvBr-UhOaumF7Nrn`n3_e2G}uv>?7OI}~eP%C6~NJn4ca$I`U_;L;vyd;ea!u`X<`3qen#b2~o|H9F -J#P>X>259T9tQ{J@_*n)weNem+hNS)55?pIa=llVZw#KAX)XuU}*Nf6Ze-WP#v=BmdEZWBVjRLy1a$>is`ITHOn<*-Z@<)AUq9?-kRte6M_&G9+uW+ -p4eAi6Mguld_GX_?jl3LZLzDjIl5iFg**VRlxO*6gv!BwGt8Rwf|r$Y#)-^wDma5SF1rd%^4+_44@yL -of=JUeo;+n=!rm$O*>`!F#6Oulg%OR(!U(!7d0Vm-*`{jwRjaA{D>`R){{8nq{EICv7`f{56c6p`V#h -c3S#(L%-FKOstCJHPM)SJPd<|tYT;~bwS -71y4a7Nvm5rycg>KTt~p1QY-O00;of09jjISapJT0ssIo1ONaT0001RX>c!Jc4cm4Z*nhmZ*6R8FHA{ -8MNU&iT~f(z+b|Hl^D729C4i8&Irc_NbVO7eC_1i>vP4^KWzi6+8u|Oq$WGD}Fd%T2w~ceDA+xO+yg$%uS>aT9nm3=9!0>R!Bls9R=8`k4X^w}?DdF1c -p-Pl=jxW2ldohZ^)6y-sSCnfmkKykTxl6^GOnQ7sH&_5(kBR8$YQOKMdYH;k -I0HO5HI+EI^|hLRE%#>y&`=Om(@kN#;)pMn+hrs -AID#;Wm%AoxGc27XM`@5eqY_nT13E=tSOMSD)Wd@oVZagAh=hED+)yopDuzU^xo9s9S{=1GsLQCnJm5 -yKeE42O9KQH000080LuVbThBUA@azBp06GBx0384T0B~t=FJE?LZe(wAFLZBhY-ulFUukY>bYEXCaCu -#iL2H9R5QXpliiciWNF=qlKo1fPSy(mT2AY$YI7yf2u8eDG|9v&dsh637dEdM@0m^;yxs>`$W}u~L$T -)q`#A{?idBJ17)h%so|I7ONN`Q`ZPMrh_6i4O --7;TRaM=#RfP|{_^qv*!C&A$%v}|kOz3PIN|`PDuao?nc+l3|5kCsw!<7#NNDAbILYfE#wG$X4e6zVvG)1TgjZHKVxYaXV&y6Y`F{PSO64)&;n_Xki*0|XQR000O8%K%whUyVp -UWEubfvSa`N8~^|SaA|NaUv_0~WN&gWbZ>2JX)j-JVRCb2axQRr%{^;Z+s3xv{VS&8gYDKRK$6qe>(a -FxAgynJaDkk*hp-e`9%B_*awR$7a{2Ff?>!IA=wTa@c3rR40@lo){eI18vc)dW4u9J|4g(%%e0wI=*tx1~9nCDFv`6A_>$HLh>Npt4MagzIan8aDDCBMs-nHZig=R8v*PyG2jjBn1; -_@F)nCn^GddFeRnV)9-ZIt*q{GI@Vx7aX_K!7aAHL@3en)=I(`GSQJ$qGce$Z*A!2a -7ZV1YH1>Pm?IQ34ssM62Qbf!-3-@Sm2zD1^y!AY!0$zXG<6yqh6$b2w$EnI7ev;O3}HShQXAD8GHY(+ -XGmjD{Qu$F9S9Ln$jh^n{pmy>@JKV7NmYQ1?2){lF!7NCF_97;AEE#f*?u9h-{t!Bmo7G%2+-PC4>=Y -lkgsp4FRVq{6QMnhvCyapJ)9SFKC*Za(@iZakd*KFL?YS0&U3h7vpd;;VF*;o_WY>Z%zcd(|i_fWw{^ -RZX;@wD7ni3>Ki^ufC~h)o#EO@YZ&(mh64NeI71Era#1)7k(Z;TKspLXDR99BPzTc_AyIq+lW=c%t|8 -E*PWKb@Gf>C-aJHDi5`}1mWJydgA3#O?Gmm;WJOrbQ5LY%sJmnBr+mB88^++!83|N?f6@!+riJu~<

p`qlnYc2-k+QtonCH=M(gtZ$+7 -qAMV)pyjQAx-c!SeJP`qDp`C_zuXsL@s%y~0_*zoS?_cA8MIC8lTxa={X+)TqiZlhV8%>PWYeDV42e) -*TfoWYxe!=tx9|1$jVcgH8E@6Z0{{NnP%zkmJh_y54do!z~cuU`M_`yYOUKejywq6e-wJRM#RkrVd1H -a;bUoi6aH0Fq;E712<(AImpN5XXA$>BM#9>_gf{CSbv@PwO64C=CyiMfD1yI{Bx6xXo0 -hKAtAIT$T*t41mg*SFUMV%)-?N|E%ms16+RFZ$Je&DJd4u|p48E-hKNJ@Fba*8DT6n;?XTT$!tHLY%m -8(seELe=SnIi*<{!X|Tfvg)Oi&*3qJ+{+BestM)Y}YgtHGZG%+IW)p5lpn^Kyk#IM{d))DY-&>;g-27 -4*h8#>Vg;zs#}H!Wfc_G^AvzOJRVD*N-F29M8udlamc$q2m&8_%7;D>x{n5&G)3{@j0gDTk#+nWM=4m -3-C}IBXap34O8nXtfZ_l4*QkCZA^U6O%D(pL`fDvx@r(UXF`s@yw8a1bHTSsVd6Rh`D)y_<65`-BC{I -sBEpLJ~dEcndk#_vqW!pcoaTw%RVD(h#U+dURqeUpBir-iD@&qCqOGbb2Ag6$j<3|7y*Qc_!BNR4Imx -Z*D&XGGNPbHu8`+(1LHc+A42-x@$)jTGCB7d+1?_lSM%@!GkXO_%)EGlk^;tW;+3Ar~eW;^T?4(PNeN -#f?oENX+VOGoYQr$WqKLE~u>i&Q`u8SypiWELOyjo-Tl|1%IkMnUqqoW}$&PX6@!?BK_p-DWw!3~;Zm -P64(kW60(_P^ET|#GiReg|_oC1*62&K^P}^-xxV1fXMk#h<%_abJ;P^A-5YREK6`shJBiFvQ&AETO-* -tAmL{q`*{SZcSkJQW1_?{MBTAv-7BQq-B9im=ja~j#QY7j5ut@Wa>MNh%qvg`T`<)@7k(7`Gwykh?Pf -6x``e-4lLZRr{UV<)a;c$HP-g5{zJrQU}}l3gX00pB -&S{?i)3|ATjmhjE0j}1BsCeWXU28xWMFYibimmE)1`@#dckGaU0IXGzj&SIl@rz307ewXwP7%7fb=S4 -hXU*BB-oa4d1rK_FPFpx`;R^ijV^#D4HBaHf~}ApKT^RIy#PqM~H0sQIO1#N#(C^`SQhQ^vWQKFfBiY -nh`;wWug8?A$W)mW;3(oGska2$YD@1h#B3=n&7{3E~albjd-1d(ndC)u^xyKod^Uviny!iSk{9dWIu` -`4X%Lw*F~HXRx`xVT5^u~83C@hW{eeZ_Q~$-p%l{hjhIf$CRi3Ocm)LLTD3(`fwzy6@JJFK>V%;`U`j -n{JeDKu)&NPB#{^oHE@<}!qcO?3wx+FNcnK;%(vC!ZfO2;d`H;s?u!|A?1% -823@z$xNxE81ysw9sTCo$xcc6zAPH7vIms3uE|@KD>=!h4>yz5%+nN_0sFUuq9H!&nsh+B-r7QEOP9= -%sXUW&dDeC?0ic=fBd<%Wyte(ABxAzt>p5sWaAk{#5HY9SLK8S0X|8F8oPO+hlc2Z79IBuP_rK)a$XI -`!4E=Zt=WDP1gw{{e>O+9HQc!nibSDK|hhb(9Sb66PFV|z0IK7gyML%psj?$%VOzCbcxLcY`+^@4qf3 -;-=4q!&(UeoG~BTP(|&`%Y3gT?D}qnv_vn*DY^w_c7*n|$02Yu~M@f9cQwgQWEyW%chB@bo*1vXGwv2 -QCUf^-`+u8B(V0cNp4h1^9HBB1K;wk+@?_uOonB05hJo-jvNWak*9Oj7FBYAi%&5<5P&4t4LDqqYaen -q=-P`}eOY|VBHbo}vHyCzZ4fSCGXBM$-fsh81&rnRICusWSvHM481Qt^Tf#A7Q%1EBP@hB8ZGX#mosr -|BK9&wQWuL%S=H4j?y5O^z(Ed+kB;RpLhtx)?f%rCMY*B}f5Gzv^%AUYD9c6L?CUqnMynU395YY%GUA -u5Pcd9I+khTwC}+F*!$OT)~<>!q=b;8PzCds#i4c`fQXWQ9*Q3l?--vUT(qy%nQKuIYP$Y0oj>I7lY3QBSl&xVq&@+_Em7RAzoFcBVPJD$^n>M@|v -ua?6XjdF!Rh~EhybZqz5$n>03{;RMb0;El4uLi-+WvA>IaaxWR!Vl!a?G;F34JLMKtx%lm^hYg;ICQq101T -*}uT)t#V&ws3tL#`WU+yIHgBrbGLpy1$KM40=KjwB0wRrHK0V23+$HJ|0V67QAb_mY$HAm0W~_K0<3Y -mWvwG*a5Hx~haeT+#7WAp;$+*8m)l~+B^M`B!7X7y3zVHd8a_9i -4__iN8-^(;-n1DZ)ewmmc+6VLq!iUUdl!EK7w{zvRHlw?%#!4nb^!1mpo=JhJ>Q40y!^i^X)?p$4ypE`()$x|>#?RapL1h3|EwC`#~b -@Ld|_ocWBJ^Cle)UJ<=IWE5$h-G%wI<6Jvp__{SM%`3ce`?`%gTbR7dexxsYc9P9>NTjY2SvF$f8-OF -WE?fQYFVf{bLzn+yWqzuLVtX{_wS7`j+KRevF+4LwQYePT$(`tihB=H3@jy?WKn)-gLOz@4OOX+R0?l -04Eh&D{BGYQR#b6ko-BxKM!*f7DO0>juUYzKjr^z%N(YbT+l0jJ;{w@Qn%lrRdSlAbvSk=a2*(iUyk? -omAR@&9GGthcF>zO;{jYl6!8KyFzo?xPBM{<4K-!;t70NbM)(ThGu`#XZ3C-oftQRWpBV3^zy`_2HPw -m%|ZrC!&Ywm~*cB;dS{10q|Uy?V`4I1&%mCfuM2qz|FSE!ITHL!eYk>9b3h@V&eh6;o<~Z -%0Y0nQ-Zr2p?t$azt>aiiqvxT!F97dTpEO4-f&3@eAO>W!r-|Nb5i{VgW?85Zs2ZA+dSKYW<$e->yR8 -s%>Dm@LGpZn90c%-gTivUdKJp@7LYK0-qbQ!i@m_ZFo4uLF^Zt -Wk?V)d0yt@CVj-VH9g@T;@~(5gwG%b0vduwK8)zUAJuWYLzwGfE~bprW*uyReHBgPoX?m_5(L;pV*)(TLCl%G1j2_8h5ij)vmU -u1PkT)SMalGNmj=6w$pBy4#ayHPrQ+zg+UFJHAkMX{~h{~(F?>>Q8*}+tbG#75Pki^8!LgO@)ku^LVA -#dL^+;v)Mb~63B}K^9&a$5UeURi8iZ-delZ638RV8vwWC&T{@Fs_paoKDOAxG8okDWj&vE{mIJMgXb6 -0f>8)+-7)HwCnS0q@QQvn&sb)&B=)2yeVi|uW8!tX;IL=~lmSe(t|nQATaM30icu -ddzphH;&Wi<1J>ysg{(s_c1{8MZ2H)o|%?tVob8KxgcA`zQYBben-}s^%jP;2L-WivrBQSq4s%-mg8qr@AtUI`_(G -HhVUFmXtJf23s0#v@l)GAT~%T#UOqYVHNB+H7CfE>{(Oa}{MJY#^f?-_lh>T42%sJ$N>w@TU7KEWPkA -=I2o?7Q08rlt=i^64Zzti+0ODutJeplM+7U2b<)H0L5dVG7;w0WSo)j6IK74J!!G0CK9m&MIxzr1`ri -Yn6dNC2VqtHoi*OY&OJe?N*So?Cvrlx?O+P5W|jlw?v>NY|7@B}D7)HGF0Tvy(#aiv$bs=AF^d7BBu+yGhOB0O!O -PpVkzixlN{>S`6?>P%EjzE&I*K8$!=9DuDw14m)GT=hqCzOFrS3qX$Uq6x_7>iz)xHWkQKvc;UIqE{# -;trd$ZB4I*9&r_O(Tv`(l%7dwTtsdL`msV48(#%`LA+F^CgCjqcRg9ed)XgiYJ%!JdFVOq!gII-$RVr -j?IMwDD-nyjFv7}ro>7QH^>o`COGTku(kMcB`Pem3|*TyjXP3g!>p_)#+?xaiQUZc`!7k#d{eviD`{L -uJ|tqc=;lTF%C8W=?(Tk_I)Ghh|#UN)5%bjHW7FYJ1cZYK~+=@DZdg_!9(#%sK|rQ6JE#qM%wu)lEH? -RM!Ji;7TITsu`iQ;VusQduFh46obKKBF(mxVE^luc`Y^a?(Cusrb=$+IF#}!Ee&RkzNreZ{dA9j9%Ij -7rW(m3B-mx^$nfLUeM)48Hz~lIcNOO1zuApp@1>JLbvu>9EzkDsTiiN+^7BNKk#M!m#*nwN(w@xHCLd -D#A=!RMc{zXG_tuPW{nkK*|%)3AqQ?UqEg=5Y({4KS|F}ejBRq!nN|!!O+?46sv4D8T%&I_>wej4Ez1 -^Uc2(z0H{L*a#->4IiUZzf7RJ83V4yj-QGlBI!-iVc2(zk-Zk) -R+4-{7hjAMiZGc!GwR^Ft6hmp^VGkY0iWW8(T%NkjHh>p%(sZM7vd^fx+hECDogj0(z_72F5RHQ$94@ -nao9tZJCEX;)6BSEq~@T-SK3^YQFeH8_Wt~GaC+H?Sc``V@D72H4zn=uv_KAuXU;jr+yIXcQT7M%Itp -(gbWM|Ftbn~i*DU>9rWB&qojt+fSS4B|y_A`VO!LZ3&CPe!Bn8I;P7@8*j$chiN;%+{(qK2!HrQ8G+= -naxFL|jTov%hZPpLRsvz_s(SG!-ue@?49K`Di(NJKHpF(}n(jZwM9nShCh5vT4 -^0bM8#qr{?+|ZeCcDA=dBlq9zHX1Mni(k0;?8&YK_pm^~HFZbjd -z=p+(d?u- -Bup)Q8n2~@=8*Cq$Wb@EN@GsjnWXFjYjPCV?GV@fypNU|>GKqWX{!7=)B5~g^{dWxsctBtXjx+P|gqk -eoCPp}Uk`&JvioV1$vYZ)NA#@dr)QP-1uRa%;c(aPI$>F?yvZNvfpUWLo6I$RK@WX>8}tM=$b{@2uO! -*A%Qe%(v?60yz>`!9TTRF1?yUccV^@io09(|P^<%U7=|19o?JU%`KB`L!3BnHi=y$*unsxZr)hsQgLl -rBOboeYDj{VhIPCo#+(Y;m7*MV|D(1y>xUp17#?Tb{*0>j>EO7ci<|~x4lANJjD8Pv4_M -U!eX4CGaF?Ho0rm&xayB)`*Y_*@6aWAK2ms3fSzCE`)!`2X005&5000;O0 -03}la4%nWWo~3|axZjmZER^TUvgzGaCxOxZExE)5dN-Tad5s!sU*m5mlt7fmZr`cEJ=gJE4F!tqR5oZ -RVEFJDq_d^?>mxub+Qh_48sCKdZtl*ZU^5~hhBMBjpbv%Q0#4_GkR=B4Eh`m#2~m_5vJfPSfG$ -^DNQfEZGM1FHywsV+l1REFm4;fowbX^?cSbT9o5u}jO1Pmc<;G1~utaj6<&~dI6j2f<3u2W5K`d+h`^ -_ji`!qWH<>LBh7{Ym+<*1NCXCt%#KiaKauPB>WK2O=&g>%Z{%xYBp2$e`3_a+uOInQ`3Erf04&wc#!A -PB;@gD?pEWBuvKtZ003HHvOW(cSnqGyyOcJ@(W4;JK6t8Vr0gG#p -4>>HKtXAd4IYZ-9#cv$;Bd}*==Jdv3F)|$NWrYEQwY$tuJs?G;(oR<2GBrZ@U$s8HFiiN~#ZhlNvOVs -MU4y%FE#vL~&?XWAwQ$#5haZ)~NccuM=jHWuJSKYMi$>R$7%A~P&EQHuK{K?{ZKExJ$8Qp&Pd^d7&6M -J;|`Rw#kQR3A$waKE@Kc0?9|EO)6nyRjrw0L^={OovkpfTL`Q8<`3>q&o5a_y~dpaxVQNrF^{{_)X~* -GH4Xnj(|2gsFadYNz&G!DLVV*i4AAquG$#n#$@g)ytQkpR8Ovl5Ngh=L#MY8fL2{*@ -Cb<+?mG{0GJX`%}iop;#(Y_His^Go_?!9rX#@~Hq=~oZ;fYYxlzW2?DxQ%I!!{JMu6=Wm?>fyAJ3sSn -9aoog_2niR0;Gb) -IF?EGH%#ZU)Q<>(=`vZ_x+ty($Op7Xlu3)&<^Qd%zjO$?+^cmo(Z=1QY-O -00;of09jkJ?-FOl0002~0000R0001RX>c!Jc4cm4Z*nhmZ*6R8FJE+daAk8YaCtqA!D_=m3`Fny6@pw -H)8-iHA-I@?HeiSkh2X=cR$P&_v}k3>{(U!X>2==Bn-!e8*N^o(eIh4ZA4X)OE1_=Tc<9~g7q=9!k(i -aHFYEDcTi85po}u@QfndMfZp6r?{wEs21iDiT3y-Map%FHv%VMIin-p;_N40ZTl^HBY;t6Puq6Rz86* -N8se+gJYzr=v9sQVldJhPZL5>Br3ny?Mx4PO2S=W{Y0s)uE8Zbf3WC#j^w=DqvP4^KWzi6+8u|Oq$WGD}Fd%T2w~ceDA+x -O+yg$%uS>aT9nm3=9!0>R!Bls9R=8`k4X^w}?DdF1cp-Pl=jxW2ldohZ^)6y-sSCnfmkKykTxl6^GOnQ7sH&_5(kBR8$YQOKM -dYH;kI0HO5HI+EI^|hLRE%#>y&`=Om(@kN#;)pM -n+hrsAID#;Wm%AoxGc27XM`@5eqY_nT13E=tSOMSD)Wd@oVZagAh=hED+)yopDuzU^xo9s9S{=1GsLQ -CnJm5yKeE42O9KQH000080LuVbTWdJ7;*J0S0I2`~03QGV0B~t=FJE?LZe(wAFLZBhY-wM2FJE72ZfS -I1UoLQYEsn7YgD@C`_dLasE-grC-8%>h4Wv5scX0Y?(Lx|GA=>)-Rq1jZ9N*nBKb>R76-U{Be4J#4-N -=(*{tjufVcc_hQMo!x;DEA0-4Lo$6X>QNFuJaAFY2~1rc*E#;k9<;e;zGqqp90KO+vg=@Q605bztm6Q -giUm@}hW2b{-O}k0H(2eq_5&;s;Pm0|XQR000O8%K%whn^YOTGYJ3yF&h8?9smFUaA|NaUv_0~WN&gW -bZ>2JXU?q?dgO$H&-m(8k!rmPrW -ZzjsEmY|A#;w)sGUW;B}XGb1HP;uUe6Amdp=9fz>hI!-uoqbTMsXK|EvIsx){aTrohtn}Q4rx?%O^_o -R@lE-&B|Ebmfl;HYYfWg^xV`P7fXbmq6}oZD4Xg8JbTx54 -Og^;;5Jnl`4xa{hOZNFbTt)HvAxU|*d3@SF8^jpB8NZ#pJAFU><-^(e#pTCeudZ)?`~8nUKm8ww{o|8 -wUVr=D_domyUydB3V|7MXSEEl3Mgfm9LzZ#~Yqk!l)j1#i>6}e3CeGR9;`HV{(2x85PG@|1{&6(EaVF -Q}k=#~~w^Vc|K84D+S-;=^!5F|_w+q82Ziq{N~o6I2%et639=aUoPWM;NJXOqZX(V<~fQYcvW9uvl?DR_G{@?xKwmR0dc`l3s~cCh=4HrMTy=3+qpy|uDNi -DC3U$1F1cfxXI{2CI-KRt|<61PzpCT9{Gz08LNy>_V9RLeLj|bzhif4O -CV_9TIGdNMf^QLuX99@}b!~B~NdMU!i*fifnVI(#2pfa}yP}~rFVf~7$e1-kat{t%rDhFcQrO7d$fM@ -QR$j7nxRHAGfCiOpnkdwkhFE18(@;o515z_S4I@o(U(bJ%w{_jwx7?TpVbaREs-VX0ER*xb? -g_XN#>51b{L --T^mXX!e0w~mO|5X}_M?CliuUU*%Qg}>E$JyO`M8o;rsB24wDYMBkCUljV({3!rcwn1A*83fdbroc)! -G*p<`*CA7{Oo4d_0F?w(ILX0z!d76CrWyg3rg2D)DVk{z#xA#OW4+as4;tY%>yR|DP0I*a?(t_91@V` -LmYd7-GvdcEJz$BY*;>r$aS*`7D+H&NspTK)1N5d5d2D3Ak$kl=X9yvx&qN2$s7%fW&&qKQCQo-;?O8 -?O4@z%haI0jAph|L%XkU>t_j#*sJTQB~UrD^7WEJ~aSX}1B_Pl}c|{^MzZ7_whEVV-k$aGHG%>Z|M{$8Xqh}!UsucibgK8_{0QKb+@b?Yf -8TEWAW+5GybpOD#zog!MKJy0*E}CjibQoeb^R+U}*Fi5Vq?T#E?y+{ZnUpgxk`H5!I#;&mmkF5hcYEn -ny|+|2ObvjH9Wk^8@Rm*TIa4FZ|BFIuR>I<#{jg-aJ6e`x*a-l$_fxP -gOzlgm4>y4_`r5Y8~5*%3#vw2=SV6c;4Vu!gM>aGUYGxHb>t|Ous#sD~X&LV+pHsvpDFLzcAFC%Hpss -jiH8@{(hyQ<}@dLKt(1G|UipRnUtkD+)uf?avkYr&MN1gOid-PHZO71N3=#1C;Z(c{6BZ5fC-3b)R}O -{m!BE!_mK8T$JYz!bDxn&9&i%9Q%iCy%aqiPn|i9)fXUv@wm~A_5Sg-T~lXz;iK*b?X2xj;kSjPp=l6t$QWdEg~^i53ols(T8Oz3=l2%-8(CURgRMgXQkzqPy!7aS$v33(wusYR>oboYEk7>l%|{n -ugCTAXF!3nss(Y%|q_7Awx`qK;y5-vlHkn>RVR(hsZQLf}WeufTt*BXu=}LdDVOphal$j*&dVGn9mEUm!Sx58cK@r^N^ -I>JwP`@jLXi&AqeBtE<_=vE?dDZDzi|U=ZmelG!$Wm5+ND&FzRe8q=#oB}apIdO!Lnv2gLyCcoTc1+4 -LR61bjYkG8KxazYC#v@hp-@vE`MPVy=xS0MO9&WNu40jIPUrQ4~YQWrXC^}bm?wj(M@qtv_FAK`P(Jy -(H?Q}y_Bxjd@Guv9CsiP%;Txm1Ibmez#~fpeCdKvFbQ0$^aK1ZUq{+rU_={HlY^x-goDp1Ul{H*H~j` -!zNA?KmNGBJtAZ~ahkoNQ@=3u*Dc^?p=DUUnLvIrQMnXV#$z2@Ffx}zI%%ehISY;cko|{US$x$R%R$k -ZjZoc2Wc)8izn)5>%nKhvn`DsY%|FtMkG?LF3wLP7`0Z>Z=1QY-O00;of09jirDFFB!0000Q0000S00 -01RX>c!Jc4cm4Z*nhmZ*6R8Uw1EXc`kH$aAjmuR`5+M%1$j(NXyJgRY=P(QV4JjP%t$!Eu{Mb#$CTw55<|QSF^bKizWvSw+Nzb1(2VE%n~CWy6w*NMS~~^{4_G`hmiqJCXYR)@ND -8~;!L9oNXkintvIqRppd4{Z?X`%7HqClN%?@wQoxWEpVX+%joV>*v3n-%}sHd2Si4}KW@m7=wN>>MQu -zZf(4)a?B~J7>Awc=Qpi-Yn1o&YxGkKZU*-2H4yC!cfk5AT`*&VHKpzQ=q*{k4Evt<08 -d0kVclt5=zc3>Lh`dXd0Znzh;(zzNMuvjepPUhd$xEVBf}pGu3m2Bn!Ss$2@}$uiA1nJiymiMUeX=Pd -C%MpC1I1n*Q*)I?SYovyLuOPNcvOV~=9k~mjN1BZ%hBhyXJHB=j2sal|XhP0(DR~l7BAxiUvs@MbZiU -rhbp64XR_y*l;@&~D^c3Qry4Xjn332ZI}I^)Y++@vs6nsZqsknw`Qh#94jMKc<^xr6OmP>b|9{!$}VN -ikAY8jWRwDYf~4Ze=YJ;9AzSkd;5T6_>fCU}lqc%Jgh`JUK+P)h>@ -6aWAK2ms3fSz8Y^bEsJY001Zj001HY003}la4%nWWo~3|axZjpb#rucbZ>HHFJE72ZfSI1UoLQYjZ?u -++%OQm=PO3$Kq76Uh2>DK+5@FjLTU>VRWArdPBKf396Q*Ph5UWT-pyuN5Na;5?3p)j-kT9Z{6(i$161 -fptyComJ@j6g2QZS<8o>{qumy6xC-7t)dLcxTsJ63?!ZSz@1^@YO6%Y9uombWr%B(DWfpjs0lX!O1 -h=<8uggb{&mWMA-mF8r3hSb3q%vs+=Qmf3`KM7n9KjkrK#7%X5xTzAssd&9$Sk1j37)xQxt2=HQez0X -*hWZGbMVaotYC$aeg2^;i!A2E_Z@2%-kRJ54#rZzPes+m3StJL8@fSk)g(7;4&fUu@p@y4*YOEoD(q- -q@+sv%-Jk~HnDLIGwRNwNoruWVr#j4_BR<~6Jf5|LCa%34Xt>En8C`~7JgfU$?l;>9b25iPNdOc?$p>Px -r85|PfwC0Cg+9h)Hs5?b{88nXWTG3Sy2(9Hh9W}jpRKb8!jfe%d`1ulElIcHj8WVM%J(wJ6R8y6ou5f -DA=4kF>we>b5Ju}9TF4woIfsUz!3AWzh@%(2T)4`1QY-O00;of09jioVaNzW3jhF>EC2u=0001RX>c! -Jc4cm4Z*nhma&>cbb98TVWiMZ0aA_`ZdCgi|Z`(K)e)q2+l83dit}-(Lx_}ohrcKkCVACc@G94^}K%p -ht=0+B^B-O;g{`ftIq$E-o$LR*e<{_4-bK`d|JR-$6g^-#gBG>FmXM8c@c0{P^s>rBb3c0E7l}H|#uJ -2SsW -vR5%LNf1JoR)bC)JaB_+DT}oI=TrMzN~?WV4|9B#uGB1lY>8bo{C3xFf-EJlpgJ1H5r`VlCMfo;gFt^ -_#2aaxs40PWWuCYF%V}XmJMq9NF~S(6>VrDzE6!}fgn+^931f-q7WME-st9^m#6W`@24mKIQ#OOC*D4 -eK?*40;Oa`vhjO`Oax^+5pBx=iYnC13v$skw~u&S1Wpm_qKN~if%OjE7p7 ->K6ipdmY$lICV&NDoOVGw{MOS&J-HMC1f4;b66R%t0?Wj^zm>Vo5@kj#~-x$VkpuihR~i1~Z}BW7a}g -1GU^*bnvgLJiLe?=pdD3|8ycges ->;B+cAg8s<$n*0(S{}TN~Xqv)YDJdk&1jTF6QW+QYm(5MEEu8R1V_CP7X{VH8ZE@o>9+`NuXTbe>iDzFK+R22PMD*j8T -eE&M%G#vYIbzA0tw|JO2p=zP=W -L-JWci6!Hal_!&8y5?hmy;&)mOYZeAGy$82Dd{k7{tG>1s#$>i#n7|v|`#umsb}z7bh3z@y+>ld~$wv -`sHR4GVL4GrdWYkNugSyV+N|#6*FNGmaA5oA!fL)5N>hCMA2-I;?uX#PhPn&c!Qqh5I23w_l3|-+jG( -L{lsKuQqC+ne3b*??S`x5u7caM?fvDMt(c_(_-!{paP8$wo={wrpen;^RCf#tu{MH!g<1nD^>&9c^nY -xPisP|j+0r=QP}dVZBp*47S<~bJ{R>$}s1(phL1I)D0-^>gM)IIMHn69)GSmeHvK{Q^0t?BJ0c^uO&c -`*&Vaiy(a*@9|x!meCOcdJXz#i$r8ITz(-!Op&Pbq -Jt<1hAXt}&Rd4EWfUj+X&}JHm5xc5CC;G72EMwDj?Q6VJZWOtPUfbbFUY-EU{5i2h}*kKSg#Jr4K6Oi -JcH=wSa%mJp%7l=QPLt_Z~?AN1ylK7bc>g7HZ0|KMM`Z?vZi^e*7Sk3R{K=6c)gex>OMZ5xp%Y3{&)|qwF;Vp8CeLCfx=I)nzM$)2PYOeE=M`oP?Ljw9yLG;s~1!aNm -naaz?fL7A<_`3nQNHQVJ@+JT7|96tSTGV_P&qYc)cPs)18gH3i4+Yizb&`!NQo0;usPYOp3iAI3$ZwV -+5xH0zc%%J_*Vd@OtD`Zvd?uT`_3H!USMX1-1G-9<_zim1v#-ek&@mb2g1VD!8{xCDxFpu9@z#c{h(v -pSZpT5l{+nl~Z^bM!n$(>qs{(3}0kHNUCM4^TIc-#Y|dhh^wr!6{-p8-Wc_xv)kBsVLA7%2MN6t4DP! -O1JOJm81K>86X5f{!9HV6(9Yd&Uu%YcL0^M^J@v=o+*gAJ%8%<2NV@sgJ)qb?#z-d)==ZSITB#kMy(r86l;JRfufP*_*Aw%Zz#2e$2{7(%UZ+iSq!@3al3 --ab(EO~f~UUR_KpbLZ}NTH7)8ttP@6a*Liaf4fAF4Jr9IPYt3>JB_SBW9Nj}<)%!rPWHd5zNXY&vzJ-@M6tmWkT%kGkC`_m0( -|ml^9=ur)kyFRJ}0c1mEY|9m0Liz;sh%V{HFCktt8KEimu~dbiVNpWhQ~62-m@?<*TS0KiC}RWI5sHM -<_;7SvUcRvinJ05$Y37NXnB>7@VaU=;8Wb0Nf>_UGaPUaRFc#=Va^(9StVT6Sxi2CT?Y6g63h4!vzWb -~iJRUtq!Wit!D|>1|Zn1?Dxhxt^$!?nPfC`~`9hOuf)Rq4rJRKTzt;rzRS4cc|`7nqWa2c0_l`=7fiX -7*_!ME=^PNjb^3mCU?eY$PC~Z#-T}7bg|w0f}U`*fmD3a40p(IYIjhy=l+~Q`vb8<(sQW5a+`@&j7x5 -a1eBZ|K(yGL^wwy5!hO2xN%mL$L5RM(DAsty{WZ)1E%ntrOgIsfTc62-B0y^$n*ad#85&7VmR6Xu}AIefY{ReLqHUTeM<g#;U3;fP4`I~vmwR%m)jl*&y}O94-TUM!T-$kT -BN%s>sW_GD?DW88eT}*H|8mTc5!DX>T?arcAeicqpMZ`e>G1&;e#Doe150HA8Na%qb`VPPvy(vI#BsT -g4=1_h?Mog_8G>XM-90{&~Rh`vszUTWT@`D5HtE-8L-hIaVLbw(a*QwN-8BLt5{8BOs(faZj43cWRrxBXh8eTlkxGC{Cqr_d1sq_0-TNZAZ -iwFdUoDfYOvcpbBurb<>cby*_Yp9a9g;Wj -nG@kdB^IS@NVQjg!})&YR{CYklY8Comf=gtaVQxj4s^!hop{MbGIjM@Ol_v(K)B^_vJ_3?@G+?S8I;_ -xCFI_-#3^HfTU=Y3h%VoNT4f -7lqosr^ZY^g;J}?5*k~EFp%WPH@aH`|;EPhT!7tg&){xy7pYIX!HnaLa|Md1_0?4BJv(ZM`v*%cLlV< -c^P)h>@6aWAK2ms3fSzF#Ww@#T8000+I001BW003}la4%nWWo~3|axZjpb#rucbZ>HHFJEn8V{daVaC -yBP{Zkve@pt|TJu_46dB&I*nlhf!mJsM;N+8ALy>>F0oAaHFubl5(Ne0|B|9f}!b<%zLplSU>V(YZ4) -$VG)v`4aBD?1lbZiQkKj&!7`8O&Q7olBOKvLo{Fn8+QoFyWM3N`lWB%bDUfe&)TKEZvZsDft^Y9FLAkn8hTcPn -406QNq_N1fubfWVs+;D@y_W-$pgVhhM(+YEQ}tA$jL*N -KV<~b5D*uz`U$cdM1zXe8xVQSC4YBzrZif!L~Kwu3%yh7WS;egq<1jga6y*`R|@yR1q0kqIeXv-X4+_O#_UiJO -wU=)=taKH;!r5#P+7;H(Nor|ZABlzS+{}-&eS_FsWilem3iGcySU6^`YZ>~TEf6oFwo5NEXrfl&=YG` -;o18TH21_>Q#;!qJ7p}~_(LDoBAlhP=7~}kZ|Dw8b9LnMy`n{vi9@$wAx@0j=&P{-5|E~O6pEaw2^P` -L&2%b4vF376ZOV{P#w7;_de7yDI5x1mFF#eI1H*~(V>vh|F{M1oeP%J8fxUSLc8mMmxmlrvFVgUyyW& -E9KBG(|b1+n*#4H0oz*!&jG>JAX6X5)DvHdw;r@{(PQ(>Mq5ov9#Jvu(wXN7leceIw8q@2GPmqPzbgUVSbX)xM5k(sniU<0Hx#n?bm -j23I6~Hic|QB)6;Bo`Ur6(`sWW+EbZx2{qKPkg_h$L9B6C&YAZdxgGcOel^n%PhFMmOhb#WYrp{v(7+!BE -3wXE7*?ojrOU@z&}25zq8<)I93_?3aL$rz$ZR0PLDv{a)dB(KS2Vj9 -ixL@L6uAY`T~QL*zTUfypj8#>w4ecXnQZc!S~!(5RQbI1*!c-n;cXi<#!pRJHO3{$<(vr+x9>#fMK9a -XcO%-Y?$Tb$;@Y-wx?0gJ{skOqS@TOs$^ra#be*ifgu?R*gQ9%)ycm58XjiQ_P?E+|A*VL1yX}jOCXuvQ68~mRg(d9S+3C74TR8C3D`LQM1)wR`A7)@qtK~-rk< -b*a|Da?CnPLe@{8>Y#>uokGURwm>t{pJ@~cS$*!z^F`P;8>%+naVQw(U0`y2)B*v--Gk&E@csBaY}=nbD8|~nM|W5q$nE&6MbDbf+S~1fLGidk#{V`PkYl -M|Gz~VISDgD6=idI{b5rr!jO!u6-CD{~s;lq^ggP!8@Bx1P%xTEuZqxx8fqPnJKy1H5z0$&5OH>QI2D)A>A -!5}Jv0GwRinH_hE+aZd%w*NmV&;3Q5MXsVqIgc@PMny8-sCEsj4pUdl8BUKp~2b$^XY5#Ei)6u{KQ~K -UTzP@m~2)F{EsHl!3S=G^2_3)?rd(s!7+Kg84de@aWK@TM$(KbRRzC|7t -j+|b`3TuQQA}$XHz+;Uyo>D%K}7H`Qwpox3~FAKFR2DxugDWPL^vb&G@~qu2+#!cL -xF%#Ap@1flcIqm47-hkC@Cf$F3X=2@FbHOdoov7xAYEuf$ZMMSCpyyS9^j&>y)g&0xzlr`>nU*L_eIT -qv`ZY$yBd0aPE$`)u4f`(Sk+~c=ucK$JOwRbNIG|gZ-t7zy8wXe~Vt;fqb_mN=5JQhSyfAUyE^PCraV -BF$^}$Dc#=Nwjc~)RQb2gf;hpv8Samt$H(LG_;5Vd2MS09@-N}}5}b{o$n~O-e5QA?|Mqup-~84A_&M -j|I$Y8Z315aHTIf0BpPV_40MDOS38S%^b$od2%(~8%FznN#HzsVHZ9w#a$iWB%u=Oxtbl#U{XhP&O)p -}#9Sd$8GVQv!dC7nSt~-?9IvZuqwbCr{mm}6QV8$A -gpE!W8GEptQrcGUIdG+}soWx8EgLHm)H#$9{OWsy6W4c2Srsi)+cnW}SFr|@iBDOAF30oEPKuGw{y`% -#BA-xPU@YJ`r5ow-q`%Y~MK@zb{&xzARf1hmS-ycRmy&<36R8?$E|lYw-g3UqB57AJ)LY2dA8o;Np$G -DdO)_2H_y| -tL{ANRZiq^EW}Krv-O1P;I$IQ$-jSflF8;R#6-Eb^9YB`G%Lsr*PAM@(bzTVwo+&73M}mh_a2>0I%kP7|xu{@0nwv12jS0hw*k -ads(-4u<{_M13(dmy1Fl=rU&;t+AgnBv_`PiZBZH2A}`X(}4z>DyanMrXU?9N)22)5vwLYuj!Q2Na+r -fHAU~4R}=AZ)W;5XeLvMHty$Wjf^hk$9e%uKBH!RQM-|`d$52I7k -hW5qj?_rgL0nJe6J_D?f4Q?1rR1)geg~#+uHe&@S!C9?ELmFxW2hQA5<%Wn}nuueNJAxcFiuj5#kKk0 -9KE!C}qmjClEKNgW=#Rb6>v~frM)KwU|Yv4fj=dO$K>RqLMx*l5gt~*>TiLzq%fJ3gFJv&+1X)|1 -1Qc|LFq#S3z7xE}cdV@g)Et0MIb<)<}*T?SqVVMdaU#&IM>?;9Q=isN!y45dReKp~Q8!DC79Zf6{d1~ -N@S)Zx7Y*+q`JUYLe-(Ft)Jvh6)3eJ8$Kl=;(dwO~8Q@Be`W5^4hSGbfg+R9`fn5xvPRxlVcxxKOl+( -YA#`l9%xf!VEbiusJIL$8fn%;%*p)ogp{2xtoq>(m+m3O;YwoTA=R@$$&pFN&PlKOoD5_^|ON5jUbkO5x7mz=x4E+h5 -O9Q&TAs@f<{=L%?W2d=aZ^WPtR^*X -QahA(bWl)8jnt9?kebttYRvdNqdWb}yl{b?7bqg80ROdTH?DE$HaMub@ilQ@G*eDa=6MHH7dPVQd8)=Vd$kK!|66e{TFmLoWx8FE6Wmg|ge3!L6zdz7D7E21M@16yD7sA$&aO$9%GIdAB -@l11?KDs;a+~i|E^|_8lAOZW%=@^od6|07nTM@o#L|AwGYK&>eJ61E@JCbsVjtDe4ZHb__m#s;mEGkH ->s<>V>`;;@iUa$9yTn{t_ZTwzYS@y{3$IHTZSCv!gfz$Pc))&5V^KFKURG*)nM5 -mm9_5IwFC&d$O6i0wgjn&(!NENUW{p*#<-zl2wmLpOK1gn+`_uP-;Rmxcp -B*eiuzlO_;bR10RSRbRAYz&`6FY4V&Og8_69EEvICN_SmmFoJjjmo|4q>BZA+ -yQ%-l`jkc%TpY=pYOt`K)*iPKwv`RmN$E6x8c%)X^j4*H{{C?WAL2dq(&5h -!Ps%C`V?5egeOZkhf&n0|v{r*4)s(}v!*mf;3(h%rTb@KAg!T$yM+IHLC+`6fest&hNYQa>(3cy_*>(T$!vHGsC -WIK+2967gKRR3{>#{O{a2Mglx^al2xc6(2Azt+H%J*FYWv|gxVS-)z6*NE61P5NW(@H#ZppJbMHsDn> -^aqcFZ`uN*+v4&lpfUzn9E*=Qm_hR!TRrzQFA2)y(Hwu2BS<^YzaIW+k@|tuz@4b`f(D(!U)+?VmBLQ -;|#)e7Kx7l5I7l9r?>x(gtfGe4`X$MEKs$Ck}Sh!3F$P`b(U*->B|75eEVu(FW#N01QS{z_qY{9z&M> -1g)Z3nD2Y!$*RT_;wQcgJ#Gg(4~_aC{?5Btt%N5 -n^9?t|5v9PX4tXsj(=$ChI)vtUpEPMwg4@8kkP+M%8Jc+Syr{seT#?GB)wi9ul&ryRsSGQ>_aU@gM6+ -Dtt|~l-Ag%RNhUszya#ea|Es~v^1mjUn}Z)K$ZR7-MqhgU~pxK0`sMmg=CWq>Nv+ybCBn9#*<&1dwo`BK3h6Qt8#dG($>b>4k1ceI|gM0)kvib(gd -N+u4gN1%IX-Un30y2gR&yNto@*>#{U6OO9KQH000080LuVbTb)jZkQf920Bi{W03rYY0B~t=FJE?LZe -(wAFLZKsb98fbZ*pZXUvF?_ZgX>NE^v9pR!wi?HW0n*R}7p2v9Xf8>E_S`O;Oi&5CdrzwG;G^rUE66E -Fu!Al2j6-?SJo3q-@9Q-D`c2X_9Z|&6^LI*Ns#LGARtcnTn?YcRNz|D~0_^>#ARxO~b`H2u>hlBIh|X -s3Bzlkep$^r#-`E{9i^?vD?DcN$1CFIt?3XnYY>g_&v4Dj33chX?CjNAo -nYT+@a`9?)J9T>_h<@^=7cUGdM)mV}Cj~xT_L6g@O$b?u*)zPBXi8K-4RNzu*D3q+bY9ad -`tr2Ue_%m9_23~np!J5Ayv>F&aSXv=nqV6bQI)P5#xW^LHHJZdbN?uU$Vv{t?lmxC6vU~%EJ$GrNs+U -F2IRe5Niir*iQqnO)zU58?ciIDr7kFXHg8`(wWF5g)#oPuj`tabDkyn?eJlrCwbHUNH&P~aTdvvDojX -F{vq1c^%mEvr%{9hO@1h72kHns4IrIXXeeN_B_P5eGbdAP~Eg<%pqezE)d -(I`rkhpc?ed?rhdU$m-awTCuRQ<93wtuWN?f@t6s;8GVm#D?5aa!%&wWRh_d3gVbI1YG==2TM;FdDre -owTy9N%dW4QGkZZeFrO~>e-RaBeQ)Mh8I@q*piF6wy2>4kKSMFWvV30dwsSfJmktFzW}Dxkk3K4z2R8 -pG@{%bI+`fAOXJ9>ZU&`NR*uYIFqj@IG*_HXXK%}iX%(>DfNu@jcZe0IW+o;>eN_o*_5{ee^8WYYS7= -%#;F?hl1ii$9)t$hEt;qXPAesN#IFE@&Rt~L|1EsyUd(}E;@4S2A`T|8%U7QL1Q>cdBb?T*lsvk~t*X -kCKWF?7o36}X|x}i7KlRMO?$aYdnZ7j7e3Gfftu}^(!zrvtPie2_I^}gh!w4K^QMzVnvGuCRfqk+ab3 -OmVj*zus6PFr4?hzoC_i-)^vf)~#G97<}L<5ef!a{k-HVmY6MU6ylc`$iy{rQxEG=8OSuToW`*>)=07 -O9KQH000080LuVbTWP>t&&2@%0BQsP04M+e0B~t=FJE?LZe(wAFLZKsb98fbZ*pZXUvqP8Ut@1>b97; -DbaO6nd6iSkZsRZvyz487_Rs`K`~d;7mtOX;FQAKFib9ZWI?=9784{h=e_zUWYqdyrQNstD8ge*8%C? -;k3^6)o=g%R&e)7;+Iw*AwU%jir6R-z9BKTkrmW+eSnHV8@ctWTQK^4$2f@P>YMH?B3-5>?dH^ -kJ6I70l5!^HIxHz=p#q=0W(tyt;Zr-uhvR0kW8trd#c5*5x1KMUi|Em3gaH+A0OX1Psa6_n8y(%Pp2M -zbAkU7=`^B+6P1$Hki{$B`&f`^5%znys=CqVtDORQ{$Eyxer~Y;%FdmT^dFv3p^NwlN?)eU{shU{9+j -%G0Uc4wN2jYf^@PDTgO+Q7tC1o*+QQK+x36B1(}ua%EX< -`rmK%MedVSf)oc-1saMwGdr)Hota%Z+3xZJdvTs&^o-Nwp_8yTWtoaceB343V`p9DTQ7z$sPy>A`y{) -bOg!%efkBsE@|fiXazOe;kr$`Hb~_$?=Ckp1=`Z^9OMr`gjQ0g%U*j;|L`Aq0-}Y&qJ&G?+d9td94hm -#cV*XU12%~VHC2_t&;lqBtMg<$%_)g%@r6C5!`C=v#>?}Xb{c!ZgAH5yJgx)Yab{N=H*Au -GkMtgvAJPgYXFz$@&;>K{cJun~}d*j@}H%Q-693|MRZ6P;4Yv&KAnM#fSNFzL>oWC;shl^r0T=bmBBB -OOIfMqZwHD=ViCDZ10i^7OoSNuF4S5m{!TBd$u}^P=WmntObHDLHTEWrsiH>0^2=QU}yyp!W$HzSOc< -sB>9L5Qe*RSAN|CCOg{d$IsGCR7U@|Y$CeHY#h!L?3z0;l0A*DzwvTC*#AA93e!K3oh>R^USbp};!#0{ -wT7N{m(t6Ow#(=y6%@qbx@AJWb+b8`|n{U!)6!iv&GI=`z`(e2=Z{tVbIZ|FS?&c^u&+&u$7-ZZcGs1 -PL%mn9@kQfh}uejYmQKYoq|sxkO-c-xtksal3^ms*2?pb# -`=d7UN#AuoZj&O+j@hR)knbW_(fMhr1mt9ZL69ltcNe|BO)vk)IlY^TW4YGk( -kJwL1kL)a(+B@8%PqfqVnAe2KPWKSf&z->wi}4{Mu^B95vyI{zdo8#l%Eo2dK)!cJ`sS*f})u!HpvgD_|HB{6MUq|^-{G?&;bAcF`;4$`+_o(BL|9wU~Tvyn1t|RF%0GSJ -)4Y2AHLPLzd8)}c;n1S){wH2jo2RHNgxVP0g3iBqUtgukeRN0WA<&^E$8kDf@Be~%*4KX>n4?$b8Mb9 -K$CP|wv{%@M+3ZdETwB<47M#$HP5?zeRU|k~LrlO%9q~9~-RWrwM*RI|eESWww1G4Za>^L -;(M6tPTQy?<()f20t~vUTFMs1HYS)Om2`Xb>FM)b^qg$>TbkppNHzpz5JWg+r#m -{vDY`}&#e~7Te6>#<(@7;&>*546m;t&+{SJk -Ao$n^{OKkQ6a4yW(#RxS0%Z16ip*y3xl -|hV$=XFdh+$!X5)q%YagmiTq?C0+Y|)hxgOB)7g7qeyAf|G%zk^o(_R=4exaW-uneaTV4ihXs=F~j=$ -9~e`?V%n5Ys{Q}2z{I!rT5AJjPm>nPLxWH%_w5$!0i2}{JOafnW5h1OoUiLzCSLLe9-lHYZq?AWLH(tF7 -BwC{zNq}|>3qT;cnk3H|@Kgqu=K}yzMmSd0ORGa7vZ<=ESSddIbX?paEC@E3t%>s645v?&QycmW^m;D -yuNmOD!l&!p*yx2zg4r-t15miJfjq2c|+oJ;of^;pT?14nI*+Neuq(p<}rhfLAWC*k&G&g92Z6!8P#H8+lSqx(_iDx^M(L_!cuNqLn6vp|<2|r@-jN^uCq)HJrVqkkhj2uf;` -IH9IE&^nv{QN-D!ZNo?)dbs0%rsifwQ<&@9*ao>(xZCfAS~L*Lai~<0C0IN3;FFH%itP`C_-1}wqYl?5j)Q(-akN!)Y#HXvM93dU^oX^=`M8F$dPqfrFn_0QKTNx02Rq3KTy_54Q(-;ln*te;U+2 -{Z<3v{Yx7nd*pV%I6>`)B^Y|}SS?B1T;kS&Bv8_Gcd_)s#8_W)HTN9u?MXnMUA0+f3P_RT`vGH?cuo< -3piUKCzrZM+cJAymYmfig_CyOePx-8;A>Uj}vkz-I#l68w|w(o2&Pe|$J%RDR<3GiVPr+~g(BqAdanx -PZiTnp5kLe}GAn0+L6T@SUBVv32R;4U(vP>?Q@gV?@1uNwPgi;8yve^oH|sIRM;Fhh>c8HH9T2d8=fV -eTM-LlEElDQlg@5M4&Q`bAb6G0=ddPsuF&BHSiWRS&TA|F13N2tV;50xuwvSh%}rU%Qw&Fgfn0X0~L) -UKKqfR)r6dU>c`ba-*3j`ICYCw3EweAXj1Z}uDxT^_s;vw0^)(Cz_>WdnKz^!ULs9HS*1LJHj?b+miB -z;TQDw(e*%{jn^ZZ}ApllV)yIp=PxUP1)0XCR6==8geFM&HczyOPT`m)`X5Ji2QP!=;dnQ0b56lE+3q -gLvEzfK~gcEnbs9IALttopI5P1_)BV5U9KL1ofXin6WYN@(DRev=PeX5@7Yp)`wN<~-oE-Ioe^(Tl^@ -rJt>5t1!)a){$0ekKfSUM}INpeWn?f{q~_A#xd4aC&a|fqw6Xwjtf%wi`Gkc8l!S`>S`+>pI$)8XuHm -$b`N!@7%bd2lrf5gTESus5a!Py)y39^sImb^`eZJK$_W(#*QxloL$E2%7Z=0{b+6lZ|jR78fXOzkyYA -uM4M_&J6V6BVU>5JMvIBuvsEJsBhIiC9oQSb6)sA@V*ZIgy!qPcMTc7AKwNyW6FaX}r9^ACu5OZ|#9o -9k+U{_@nehFRY>KewNBb04duMqBtHa((4m7+X&vB`d>L_G}Zb`eR1V -i#UHQLKz&NeE%&>Yip70Z558{tZN)&sYALeDYU;l?b$i%Jja_M@3g1c5$E($ag_gVw<=ko$~=JUVWoC -*(aKlzGt#qN0Wm7AHoo^E7VxG>@=W{5fIW+0DzccEk#eD#@N -lO%~bLnwRZ11?B<=<3IOMntz{<2_8Pe>BUPRFmT=r!*KMQ%_|j;7@nza?s6#G|w%3l-JhFgVpR0DlOS -460ToQeX(q#3Ly!zBuYw}99cEw9l(-@`aZYN*Dk}EB(D7rOewtUlgP!6)8coRcX7t7*M0?8c1_ps9L%f<>kXZNy*O#SWSRn9yJefhLVc= -0#PSWcSO}H%#R>~utO#?9aE&P7Cm*9tg5^iP{a&!Cdol~2gBkcWZfZhJ4srR;f}T(X(9~tTQA4;PNrr -g=Za?9ur0c1ov^-k2a{?C;HOimZs*Z#SD)lIA?V5+c -0?!+$Q-)d)+qlT1b{{j44&!w-y$g=EsG)EWXJsPmvG7UCT9OJ&|?o2 -d4C3wF^RRV*Ep)Vou3qt&q-xYiR@MM*CpT5Z8^Q#w!#o|Ms^&Ql=f5nMUI+fF`IV_Wa5^ew^Mnqo@ov -JNnw`;vPo4I*;X!Pu{>Dw4m2F?AD7UHnR^>|r&uUS%y|bAl8JvjS(c-$og13DGKgNRhW?{c_rd5=~!^ -P*Jl@Mx6tbA1RXMV*BM{_u~B3>s|vcS=PjKYP;SHM-6v`KI?<(#h_E`>{?^y*trW_Qh}weX3lFsTAHS -feNggw;ROg0!yd>69-{;{doZ7DFS56@VnM^q$!*sIN$~-BPult502wTVm>c+z;0;07@M`PR+a#NOhgi -9r;O|NZ@lg?PUJ|!aoul$T)yDsT4b3bg6iD!c^_jFbd4j@YI%DTI$+KTP|L%SKO_&p|D)nrgbf>K=fvLB -2^o~T+HA1Z*U|eY#p3<OVaA|NaUv_0~WN&gWbaHibbaQlXa%C@Yc`kH$ -aAjlz08mQ<1QY-O00;of09jj_7^gIp0ssJ81ONaZ0001RX>c!Jc4cm4Z*nhna%^mAVlyvHNkc_WQ$;R -xcywJ;TW{Jh6n^)wIOzf`x`&hVBj5ivC&<4NP!&tP;Yxt4pj5@HT!;lc@Cwfl32dZZ6>AMRp?I#hP%MGxTl -gXKMFQeysYF#lp_t6dR0^!ge4efsGJgSloRAmz-eo2=g6aZ@@D3$KMPQjw^A#TXOr}zA6SkB(C(LD`f -J4cZmh*MW6_jgL7L`Ey1;XYsUn*n~naK47S+NG<6%SCYc$yLw<7-r}NFRPLOVaA| -NaUv_0~WN&gWb#iQMX<{=kUtei%X>?y-E^v93R_kxuHW2@=zv7^xh+TL}*A2s9z`(pFZGyJ0k+|E3VG -wBPY;&PRm89%wfBlXxOU`3O{UOmE?|%2-J(#9xvh2)G)^N32EkW?LVyc4!H#;ypg{=HRvkgKa>lzEgN -wASp@OnFe4Hrn1HlW+4k;+V<)u>>jIM$`;Ccx?vDyc9@$<9W}3i5o@npR<+1Fx(Ou(g(=HJE!kNnXG$ -8YpqYT46xQ?Ut;;j<5uQ8wqSLc?s!49+?~2)@2F@rfQPaSrV{=awOxqcO@c+!$m8^mB&lGKo*bizQD% -D9yGIjk4oEI1c5v<@;pfN`LgMXayG|rk0(nynK<6Oz*7SClK>FlF+MR8` -3g3-Z={#cH;tGzArF;*)1hV5#JNccdCiJv6CMHxxS#qqqsYAkrExNbYNky=N7>!+b{-M5#FyGgM(g=* -AMmQ^PBi?Ga?<$)N-}7EiZm@*O3^7R_ungT?37C9FQf&E<0b_QU-B<;^Y37Z>oy<>KNqeUl7HE(J1*r -bK}Tsmi&AwUEW##D7y=@B-*V)}kYEN+^&`YK@;@jkaP+u8rbrD%ukEoWZhNUT3auaSA(Q8a+Kf-%<;; -Ybu25JZK;PbFC>+=Wl-b^>sqwPZOYD2r0 -;N-7W7Fs&058a;Vz4*M~3Ec;8(g2lhQ#lO$M+oS}P*eEQR4byh=feS$y0ELy@6EE#Scnu7EXVbCIuiP -Bd_9nZ}=&G5Y20+p-(=HFwANRM^3i%2MuO#2lG5j{HsU*uV0S)o8{k=?PvSwSZUrbiJ~?;~XH~a4J|cLO8b1efoG7gmBU)CPF+Kw -C<5LXy742?9j+_ofe7ER*__7nQKFhs&YFOd!Q%gP&}e0U~7SStgFa>b|6K8*tC%z&WXr7okPT>iwg5P -BmIxg$x)uytfKsnGYTNg?$Kb5kH(~y&?bRL0gikbM9hoJ-@m*c`$n<5%KpDX(YW2B;A_jNa5@WP`=7z -Bq19Mj^*hPz5Ik}QUo>u7VHF)tiR;!|NgPcV(W+WImAJsz#@N3eo}28%kr%MC-BKZyX$duAsjbO%-$; -4RYi{y9YdU*E8FJ_`KeSF1HH8Ui6gh3KzB>tkI3;yv=@!3j=2{Yfq8U=$QDJ$1e0)^F|8o;gFhF>=Nb5w;|lYe8Kj>iKZhxHK5oE($Mj~2(HogBH_8m05Hv -{Mg!(LD7E@$yr6Nkt2%@hf-3Q8=;NDa$f=OLMV?w#ld@#p!#4B<3_+LoeBv6@{+hRn=&yw2=}(+=J-Rr>X&tu+dYK4?VK4)>2 -jf~l~eIJmCdxu(}gUD&0GmEU-@Gq~x(s3l+~4y@C#l{Ei1u_`UHg@}6wKHm*ZWaQoA -k-`urf!MztAF+n)t`ySzo00b=*&3z4^T@31QY-O00;of09jkX!apXo4FCXaEC2u_0001RX>c!Jc4cm4 -Z*nhna%^mAVlyvaV{dG1Wn*+{Z*FrgaCxm-ZExE~68`RAv5*mnY9%60?@*ip2barB8#HNyBt>!X8T5) -=*;`X2LoRKphx_mM%)XGjBrQ9=(}zSBXJ_YaXP%kmrY^TENj9x&Ymp>OZmY6ZjIWzAZQ-$zHP-k% -;Zd^WToE%RKYN|r?v^L6T0z2{XWi>oDj-zvV&h5h?T>{3xF_EuB9uIsX%=JgPtPzZHh3!dS3x9sD)GQ -FAFjIm?(A?IpS*4tH6iBxW6ddhw-EsuurTP}0j>PYiXbKW$DpV^Q;nb^QmqEMnvlBf~+X3;;SPlZKCL --L(emU_t`TB7!qFyC%@-r}pOc`DXCy_w{8bf1lWsWVZF?3GNFC5nyaKGE9(o5<`(=qPGSRbeZBB|4f$ -?gmd{{N$cWXnP~}%~F5UB15Vq;dzc|Ib-Luk0OOl^8K%k$X=F(;<6C+Y{_One*XNy>`X$=E}%^EbDQT -_Rjz-BI?)0@SyN9ji9+{V`^;uD{r`u$yp^3vxRY -5y)@3oK9JVv)s+y~G!aPiw|I2^U@oWeGZ;=r5h`Ugiz_90X~9C}iOs3TvFRlc&o-y|bUoLU?C_sN=*^ -gSY(_KKB`b#{ctZz^M<9GxLitJ3=#{&XEZ~r!xYDZLbR~K?d*CK~Q)?A0x6?2sweHL@ppwsEo*^w57< -+^s^{33IhF=yoNK(a)=UljP9hm?nMbRg?9MaP_4KstB9ltXa_{+Ah^I!@d=E -{nQ6BkOjB$BPXx>oF(-wN_^r_6wz&MM4h6{fR__GWix*WyYR1@X4=(d!h(zy{2UXrRl2JrB=KXm0u>+ -AI&*53xjYMM&1_>U9mqm$snFiAY4uu7yKKD68? -T2?>2_5NMi~2@Bq{@=x#6>+M85!*05s!`FU*&ONM;k -%1uwsl4)5BpwnBmXFgj2U$A1do`REW6_11JkAt~I%*%nP0NYg$N2XCp1ZGkdw^hbfk -HKEp-+^idu#jBZsIsQ)N`_;`a*bX1rq_i*bWI8cD6t*ThGe0~x#UGEv~M-Hk8fVG@1K45G-mw?tQ>@; -aNj^7Wf*qsC|nR3uph}PcxreJch%OgQvceJI+vGTt%l~!9!l@@k-w|)Yh#Q_J9FjKVd6{6^T{2IUM0a -fvb2NJooVMVQ6*#c24Z@g4KJTN`iP=Yuro`MTND5*E5sf6gpmu8$<8QIrV#92lRnP_OrtuU~7ZCK@HqpX0>pFeje3| -VV%JQqU}(lnZX5^D9bEN`TkFWG!kmS<~T&+j3Imse>=yykzLeQ9^QG4=VJ0v}~zh5B-(uIGet=HJi9b -e~>-b{sGCQzycbb?nEhURv(&EO)cg^6&bC3H)z4Xt=**-+W^dfDioJ3cX-xx~0ySi@4)8!0Q3J_qq^| -4k_O`#v&7vV1TSQct!zBXF0Ou%Yhpg_tXK%Q0)^!a&Bc%MAkQ#A4%b-oJqEOonyUZ -fJ*=!W-fBuw1Li -(I@ulNbBCFIMFF2Q?&M%yS%?CQO_X;{cSi<8@^#zfD46cauLv?5MAO0&KA -%q|Vn)qLTUp%F+>sUmr;EHN>xS*_cNYW2{%*frDXH-M2gUM|6)EV8z)lXY>Dvwzf{g$KR>9Tq#n -#aN$=ui-%O-p&5O{Vpc?$gW<>fPT4&~Aj)?ybIv9ke+#$&Tp6xrv&F^@z(-G&FANz`ORA?L{{3dHuw2Q8dq0d1SO6YFhBg_H4L78vL(!$GWs&m4K^WHLZ2N&m)2Lv`qbe4xSUhV9jqa{dmrkR8u$FR7 -l>?z-FahNDuF@A$yqBQff3K36n&C%%4hz?1z8h8`FuG>0+_;n7fz3(B+cBk9l09c91I(glSf8QK#NLd#BdH+fzDXgMrc*WgYArG}5uL9-Ju;F8X>2R}HkF*}tZhFsQ5%7&@Ms<%;$e+?pvoQxVo^-|iyy&2#^U>4 -0M+2UZx?E8;pmlV#Xhz!Q{6Q&15*hx}^s1> -WjKHCs|*{o(az8ZrI&`o$~8(^LbxD|$df!oQu+P3R|c-Oo<`8vkV)WLk+5CsEQD`18OZGdG%fU8E)LG -+RzTe+S`PekJaGUZ#3(dEawN-1n~4W1->GSMmxL8vgo@uG-&2RKoV&!;9d2S_kNH>$D}NuWU5vGZdpH -mS{}C9J7D$lnyIRk>ViN}TWB^7gR12ROJWFzGdnK20O(BEA3Vkz2=RJoh`V-A43 -2#*mL`{t)w1w?%e}@}!BLG>aK~!lJ&z$JwORk|ayhpO0I2zrF0IT+=Q0;o@Q@Kc*14OgXx=4olfBO_7 -OkqGlYkS0yF4LYUz$LgQB7y|-2Fb+X~^defkKcJXp}BzD2RB^KU=7;G^#b()uj2q!3u@L@c#;V3LMRl -iG*3C4NLTrLQ6xko`^j|sg^KritWY2NB=lffb~)ueb5)B(?9w85XR367aXG%h8N-4@IVmZ43(z6lUQBsMKYX|4^>G)U%)=_!`R8367s -E$J5yct__loa5peig3zNY3gV3_Zn2_7RU}EIL(zR_~gZhx909@N6ia4MBmrgZ9QIMgUl`Udk``EOt;c -qqf?8Oc}uV6cP5bg_am%)9wQ^EdH-NXFV`t^zvl02rK7t?x3F9{00w!ceJqPsWJ%to>}aMv`7br=&!$ -&yFGA!& -w#IZfu2us{3TU9$`6m~lGK`HxETF9eZxZ^uemtuW+OVaA|NaUv_0~WN&gWb#iQMX<{=kUv_13b7^mGE^v8MRtU++ELK -R%%t=)M(gj8NC8^0JsVNF&nTZO{{w}Wm;eM_`!Cdk2WvNBQnfZC~@d~yIN``tyW_m^jN?ZU?O9KQH00 -0080LuVbTUT?1zE2kb09sD~03iSX0B~t=FJE?LZe(wAFLiQkY-wUMFJo_RZe?S1X>V>WaCz-K{de0ol -E3S(K$XjxYGvkg_fqZY?W(rY_|>(uw$i-4Y?h`-NMcQqEJ0e)_PT%j&I|w&04XQw>w9-URPBo?5SSSZ -<{N-kWxi!Wu&S%Nxu{R`(U8jYUu*_h -97_1Pne5aA}vdl{s3U*D+bNt%Z4++TYs_nUu4FlQPFT&_HTyvrOi{yjz#)l?aZl^ACGHCMr2FI-s^dh%|`utl}SFhE -=(rAdr)MTO=uu*)FLz?E3or&jSV@J!6p$==(RJ;P8@I$)?!u>#qY&L^1D-#!%as&8n(Qmay{leMGokr -(w0q%k8l!c$BP?Xu>XX#OV_b01BWl@{CKIQ5p(iVfOG#!-LU;y5XmQo&z~4Y6D#RY6EO~HUT!l3d6t0 -c`;@?epGT6#xckat2T*XRSthS%fc-eun-(%geAbaFk?%u_`qXHVY_mg5l)Nd#RWj428W!d#>AYV-bdz -TO-ez%`0b|lO_|^F>~&IbJvFDLJ2i7&6Kl!rk@w9^3NDCcno3>>Q`}$iFy`fH5>;B>nFkZA@-m6SbPU -D;wIyar#_pbuG_X5f3RDpVD9{Z-pn7mkPlL=emseQ^@<(W#r4@4Jtu#sM)m~J5i= -dk3k)fs<@%49SuY*_f+3EahOkZBlzB!+(mg_N#Tw6gpj+$(f4&3Br6|4XeFN-qCDoKVi0CkzB$?}<2f -B7yEK~jLclw1he0#A8*$AbuAmHa2C8PO_Ujx3j%x}1f#(KAq7h?Yte=S=qdYoh*onrVJ{w!89fX5^8ER;Uj~B*?CpkUB>g# -P1DSr*kGeolB-Q0EOlyu-It1xNC$xf@uj@1{fjrGLsr0{_FDgOkgP#q8|jnjWq$&x -4a!v#a0=hn3uqq%rr(;k$nvfAQcy9{%UimtXn+2yXuOUyqN!4v+siJN}>dLx3}oifVf!&+$%UmXmlZt -sM@BjVp*5=+}gROm2|VzVubT4dI4?2VmfR?vfao4t58GuMAPDMOD@jC^CUuf&&DH9*bj~L0BbAU}Uvt -X}ISlu&vs0SiC$1V*o&6mhl}tN!ufQX?kX0_z-89vRl62MO2%MBvYx;ooGNI@p7-M69QyKJ_4!Mpg?Wn8#132(X@-*V;%{uc$avDK1c=sybaey_p*-Z5Vn@?coz+z*nhImddu4N`s -BYA&QUk8A?7l0!g6VH#QynuP%q=z`V=!fPSPmXOW4d5O=#ztY -M`{vlV@%rH(76-G;f=Fk8RqYC7bq8Gz&gi)Cbq2euBJXjM*fFn52iTVv({;3A2V&wN~XvwR(%%l|w;! -IEq=a*>g!GQSM9RG!y1*vma`^Pf%I$#8@r4i7@UTc8tNY6d|jBvJT`_KMd2=^zh7x16j2-E426H+4h(K^DLugA=Nd(zbYZjmbbr)JfqjFha?Iy7q2pz{x9RRmJ|8cds~W~z(i(}!X-?0|TvCT+}D^&gEtX=Za>s^@)^>(<>4%# -96ShE&D2Y+#6m0%H*;FvlU#%ny;KtQn3lv=GAM8greXssU2!D&qt8ZJtAp!e&Lm<{8*~Ojoo?iF2?7= -x_lPykQCmNr;OV;#o9G$|erYu6FoEWECYZAW&@$gsc`Clr=K1M7g6)ovDrNRd~P_m#3E#_A>dvWDiwB -##vzu6Y~N+MV=YBfc-kXun1{zI4Hgk%iB)Q6iTL*yrwL$3oWdx92A+Onm%s1qCmK7r1V+R(O6X{3br{ -)ylep%b){UP;BckQ*9Z~2W(F1a-2u#B*EP-c@*lT0z%8SXtZJ{vcx6Bi2SeK^QHq -VYLnoYC`ONd)AWPMmNZ$*(4IN21OZUs&K#z-#;(d4fo-YUG&=p*XxCNiJ5)Ru^BXv0DclV6}Kx%K;mL -gmkp)_94)XdesB-W^!fsvMCszhl#>xmiHr&>d}jxWMt#Go^b;YmIY?wgK{Zo(bH@1p++T*@svL#*aos -Dj@FcR(!S$=vpZ47MPQxz)REx`cA=)9*U#h)gB&=+0oGd5u#Ra8x|fA3+|`F`kkPyEw4l4Xfzu4$^$O -oJ*&RV2yg!~w7mxalEKKDuShk{@H)$2X -m#(*%I& -ZUM-#dG&o$U`-=>k1qG(+ZoWsUxgGw5hXO5cOXrw5Ac7rC@C{a13``SF@vhy9HrWs2bohWkeYkR-TFY -Wd5#*VmyoTSeA)nY2r|{IitFFFBP4Ig&f+HCD|I>^1UF3UU~06eM#RNi*lP3K}fqhK||$~4Q11@ZW^| -u38y_?rENos&<^&bETHu%mQ-2m)ko8`uW{U>pEgeo+>a`OL1m96Xnln#>f^(L-r_@kYu6%eUCVnvsUT -@~BTNUQHQ1g9q)VAt0X^-~eC~7>8pK?K02`@9b?5|xAl!;@&o+M~R -Q~m8A-OlKtJRC)s52nt8R|1xf)5^ilNd&$cTXqpJG~(rAg4Md0>ZKk_nv~F5skVy^r)|3*zoV=-?MI< -VRRw1zw(r0i`DtI%vWj=q@9-D5ZA+bR`7k_Gq&jekhK*gYU1A>q7|TXasNuK%dr&*0nh%Wr1q!Rh?v?9KTiSe$)7zkIVWfV8I!UGTI8( -H!z|-eN*UUq%yStKr@)CRuetmW(y+^3S-T+NiiS`%sPMXwKo5mwL22G(1F`rJuVnG&>a>{nb&^mM`)E -WpftO?q)NIVjQc*v>aDjs&qA+jDA~=lGP~o#Q)lVtfDq5QtuMlXu#680gbFhQ}9R4Ugu1mkHvRB^YD1 -fnGoREA3XcRAG`IFyvn0IrI3iUvK4cP)JU@qiTP(raiiy-F5TJ+IL%6h4@hNK!TX`29Xa#{Oo{R8K=i -#ZYBge$*J&(sz$GqJn$n4@2>gP1hRjL_?Of0D4HW(Rr4kJ41{oSe+B7Qxm0uW!z-=BN4y%? -|ueDsls$^hp`o`V%G7@~jxf8dA`3A4|c;>Q(|0MTZRb0VI9g%33Yjx-nH#`!_V`U;t(bM^pE2NBRa4* -1&F(>VoQ=j`W-z@!!m-m{^}fjn6+6@D+0gPro1aLQwA5 -{B}Pa>~aW{bm8XLkJgDF;4wn;Fn~H!6WL`~x3GR+Y1y*mYOADmu1$ns2L`6tf2>Ha`}IT{U3or8xnc+ -nr*j{kCR}YRbmLx=_C;A6q)S$wfZY-bJeq$5h|b!>?XbM%1}L*L(%dhQN%oRMQvZLksnCMF1W_oHxIJ -Cc9#z1ShTK6{-jhxhp#0(CN1JR{9;^+aL$MMjL9V(sX3^<-Po1wDT6D)70w@9upMrd~a!)1uRUSivHD -dqGjI+xoX~tli3Vm_M$A_Lx)ZWB#D38<2W7{x6NQ@YgNE7gxf{Gp0N0ZmIU1`nRB%!P{HNcyy9P`fXFv2v?b7icU5)F?W6-83SZ-fbXSEZC*CV -$*h=D+81Mu`n+maZ2Ezb!kgWCg53Vg2R&Q3ZE`j3xohAxpjAT;!od%dU#{nh|Zn4xLwR3qBk& -bP@2oWDEBCNcsQC}UN^@G)dFKx1&j$M*z4quXXeJh+q}HRU<~9NsCFP2uEqKF-4om_pa=g6W-ZmG%y%qgr@Sb+w5!T48)NmsGCHDm -SaCi@mR&=o4eCLmZv+#>UX5`op)FY4pwkv{gn8D_Lz*fUvDKW1zZiOgb35fLo(V7{pXGyx7kr&pi9F& -;pdV?bwrB5gZn5GkWRYE$IYiRQ8$%AOt`X2V1brs&?y3J6{%bBm%A)NtMb2(%JeL(DVS3FpLE{8ulGwvRU2ZZMyPw(5YxQ+fAYDCF#EJKk6ijTl1>myrhE4GReXcOcKN=Q} -oLEu-QFHo1{@c(VoAJtL$+;h^JVKsSQ4p7O(1!c$I!8l8lf+}WSY0~+?p%Un -kR55C^XRNa#*nv+M5)?h^HCEm4sq(rXyn_y{FJbM21f8B=#sgFTVl2rJkp-wFH+I -JIe{PY+h{M>T^*&1!NIG1ug6qa$n{sEj!xN$J)w+I<3}l -F?s)H4jvyx*Bx@wZ|nd)X308eU}wfxmB)3VgcpUf`3=?FTPw%Ij{XRGFFP)lprULcbO -de@wqKbet!$eX+lEbD%Ztp28u-Njg-=6L$k9-yxU+!t%peEZ{VDP49H+8=&8Oq&|d2VHV+lG^!+3@0& -gB0D!?Hq*VncfH~J{UQ53QMwy6gP|oKu?vN9S*^Ka(sNynOsHhB0m>?T%Iis1*;5uPeq9dr|0kdMcp} -p)MNUxHzopQLk~)#mA*2UGrQhQox%2#c(DQrHL!aL;=s6VC^xca-r1AGgG};l$->w=AC{F1mMaDZkZX -+%ID?gWY$?8H>;Hb9FMF2chgUB6-BQ=t&e^@|rOt;@CegeLy$v;CK*qpuAKOX^Itjiq=hP)bp*9@;`7 -u1C=8OTeL@+vB@{|o9(*ZvA)4$a=7(dMhmf%Uuhs@5U53!2}e@j74(PF!OUOSDn(Z=Qm>R6poZ+7 -X}$ukFQ?Cjj9GWmvi+otDAB0Z>Z=1QY-O00;of09jk$+~6CSD*yo4od5tO0001RX>c!Jc4cm4Z*nhna -%^mAVlyveZ*Fd7V{~b6Zg6jJY%Xwl?S1=m+t!uv@A@mS)ObW@Daf+ZCe^Icacw2htRMO;xt>lo9|;kc -6ewdjhvYV|WeUR_=MsIpu&uZl${4HtD)aXZ~?yqCb|4_T`&FJIFDrhKKTZ&h_sY?O2HRliZQ2d -&g>n(^yRakO+i2Rx>(l#$C(qg2qx -C5Zni8+HQUr}fnE4*{Rw7ms{k`^2#aPLMkD_L*m}EW@WZ{^w&iS{b&H$X4d7dqtqtIp02Yn|pp23jau%}HTeO%P*LZ}slJrzOM86qlNVjr6xY2h(r0>nM$iMC -zlJY3gf@-3^~Ywe@9lkRkTAfo%z;Vy(%v301-LG(A%&%R4PR9=p3QJ~XS2!P-t3iHW_{T~yZZXifW~9 -ExCa9V$|P6Xy+sMAl~^i)H^+Q)azgVMjYfuEC)|9A9HcCr;{k-TF579+_KO?XPjUnJ{7|AKst08sPm+0b?%Oq; -J)-2W(p=+q!RbQ#f~bsv+pr0sQPYv{^pfTmy69r)Kr&3v(3Mi??;9=n4^tmt_mrEWMyRxmtp-P*g=Xn -~hslE~kkgK6HJG048QVaJ)EcY+x_qr^yP~v6x+qM{jFto(SfuJRME;%(K{GO7~C0vpVKs@LW!*A%CCv -PtS;o&IAblNz3{k&@*U9OL>~G>|DXWcI<$CFRqmYKmJVZHwApnDgSk=EBpn2^fdflK+49`1=uZx}$Hh>^l4A*|EA_JUx0{TpzK>3GMgGXTN#2H@hz30uZzr3BY(V{zl)@vllG}mSb<~+FfwPxt1@I)RbJc`d7rsVSbB#QGoWDIvSeA-S%-@Qi=aSAm{KS -EmLRGJTMvD_GuY0nBx2))AO+SzJG7iA}`BxpcBBP|gRAnyg_N3@{LK)Zq%Yy<%F_H7JxIr`2Mz4603& -7fW1992hhNr0N$PjO8Y4ooNCQVg#8ovcvybd7o_&OLmNpr5t}R*4d{Pn4R5=xBdDWI)}-+1QU?9&F+v -`!otFO6ACLlZfWucrinMI$$o@SsXqf2vGMUeb2F~aBY^&%*60M*5>*tQyWO&o6yL~KDMetU4I{MDnuduu@hu`2q8}XVs`uWwb4tVP#s*YYUJE) -p7}sAmTeL4m6@gUuMnDodVQWFg_%oM(VuJn0x|sf2U}DWOckLf(XfC7(BIU)E&}aII4Ay61~r4NbPGK -KEoIA$)Z}6JKK -FQr+MB31hdl%=Zn|^Xf@d7nY;NfPPW@(tT$M#C`y8t|{s6?z#D)&rz}RtORmhX7?eIywx93;uNcIv7B -BxA8?ZIS5a4HNTeRcNzix01_W>@EL&fa~vvJK_%o*vsvM1ewpf+Ioaa*TqjjDLd{XTxyhCLh1xoQik_n)phq7DZ0={Sm+%JKG6xI55TQAhjUKqC5yP}G7eE~P+uw -(tU~I7u@Ne7ke)2ssRAjGg!0b0kTeDn~*T9I->NcJvXh7{8=8IGj_*ZF%iL&Di$vs=B%>4cRFzc0W`i ->Nki94998--eGgDkWKipmlxpd*fiv{ZH7lgTUD^vwoT+BWs4UG?jk%VbZJr*Q+roua#PjNpboP%f(3h -4p8(pPW^sJipCWr8+>-b5%FMJk~S;d;+j$inKbrt3hKyt02fy%O+c`f$bA3O#_f0N%z6ib6eH-S~UWi -Poi&<}*~5=uabBrqR# -(T=+P>qq(4G@YDd(zN(3=pl+8OV9Nv42T?#oGb^t1fwrO^q&|LCw2byPC~?ToZy1ulJ{{u4yZqDlRx(@kAbigmU#(MAqm*EF{7YJR^PLlwaS`nVET!8&K#L@A0HtZ7`tJrd0pq*M4 -^3%w%Z3^INRouAN=kvlr{Tkz3!!RIEu=S^I9!gnpe)_u}wv^CZx8(bWnZJPll#!;T|(GQ_ZkH_mEC$1 -g=%CQuoy0A&Zun(6T`8iS+oPX9r2ezn><@rbSt=QX)2E4{1l29jXVg-!MehlO45V{3M^8Bv0CtCvZSw -WDhzVXyHpT<}#T{fsl;8ed@qPTAfz0t_&di -<^`T3kg#rd2xS?JfYO+sF``&I^L)N%amG8N(HQj9}D7Vfj6(ucmwlBTfuzCDHNSUzf(ykK~K&fmJHF -uBm5FJMj#@<20-d&y#3V}=8gPn8f#zpVb^n`>y@=;jtRY*Fqxxem?D@24EZU0{MO@j!@;R&#G5`$qVb -GY;;YfzVg@|l^8!E78gb3Qv#cNo?H?Phdi$J#~w$ah%Sur3-YepvaHZN9BC#@xYT& -IXvhF0)xWQ(w3LN1OYvpVs&FAsm%DKU@_NOuBJ(5Tv9q}veoXFhaBf-R2$IA;UuzEW_VQlfMblcuo`(y#3@iX9 -OFj-A1RKALcXl*Py0tzZfG9BUc_C-=LY%EgWf%NZYUgTUuB!oVhBv% -M$I`qbRW=oA`aocWp8G8M7BliAi{MnfM*MgaXaHUSaJUTwHVhYy2@} -tUrZU3TG`c*&Wl6@3zJMb9la}Inya@-C#5p9Fiepx?YuE$aE;Oj&FPI)jjd-5--@#WItNA5k;$vSo(tRDgDObgn=%V&8l;ChrMwi|EEWyH(nx9k|F67LwN -F%<&1ysixDYYxUKw%Eg_PeDxKsIgoq}bil9ViQ}rlI9P{G)@WHFCv(Eg!xaQY)^-BN>fZ6bKT>%2T$AfkD%EOabA7-K6kL^{Hzl^o(O5?p<1D*J4@nP_mnt(s(i=%O7u#n{hFYej -h}p4;KFito9-NL~a`91c(mwQONI5@$ccydU2F{a+h)8tT+2Y;h(d(k>KLyQbL -8;Bz<&Ph(UcLMHmbM~X|Jn+jQ*M6o{@vBPt3O?w9i6>=_xA1kv&*w9dbDEmO&^DHgX~ -3OjYvU#CTV1Onyi=-JcbuaAz8kB)zLl;s#t*2A(M7S*9~n;&L}Cc6BvZa*E~s-{xqA>2L+4}lVSl>#` -b%?5#qLk3n7-J(K;HSjDb60K -jv}MOK7>*Xg1=5$s2Hq*!Z-aGW{&Te*7bc_pT=?w$G+) -5U7aTa4GWcW*D@~N9lKO~@kxj5~g1{YSuSh9r?GRV{AGRXG~`5_-h9_qn1lx`X++62z -aQB#?1+pxc!F{X#lB#UUPAwm_|L&WOpDI2k8V5=1LI#7A=$fzE9qTuXB*#f&r8Z_L!W6>s%q8U3Xa^F}Ckvr&)Ae0KM3x7+CZtSDK9r8F9oE-*&?1RnH|J7{LgSU_WZU=lZk(GCc -BpBA&CTGm7(LvN8Ijge3juA5^4^ZIAjy2-kmIGytY-j9-3r(s}J7z{y;ih$ghN*b(wK)_psM%%Yv4VT==vjexXkVRd{61y+)&ag -Js#PMpesIOx1BV{p^O22c<6>_v{LaOd&^Xtp`Em=rP|9gT+FQRJ<44M*=@y*t4i4vEdpN|Qp4?9Pf-I -J)aCm1ZDum6Cuo92ZX+JDGCxTHz|WW2K1}oR$=0%A+#7C8^t)$cI*$Od%6?Z0133vZiR!8pbJI975-3 -8d)@MI*CN;g1lAyI9cJg77N9Fg~{fslmr7GECYB5!eEDkkt^eP?J0<7>=BU9q?zj~1)goxNiYdGna9O -24me2n_>x4EMsa#lKx2hBnFANxj+_MM4xMB-@Y3yylI`f62e*6|gEu#zX_0ae?D7+ptYm9l*hF@mBTO^Kl5m8~u?g0Zj;I8dpO0kQFZ# -XD#y{;?DaO*IWV1q%GO@V=Rl50m3hGSJwFt`{-q19cH`xbl9-%W`PKdonODs|`}C7IfA-m?pI?(XuO? -GQL#w(Kbqem`v}P7ltz_v8j1twF3sM}n#(A8YPLFX<#oG~GVtQCwk7i;QIvb_aCnYzu!W?g(b=>&(|6 -T+@#3aP!ChCthTqXt>yoX>#%I694@G|EO+}P9m^y&V86yFrN?aUy~RncM2>&#c+ei7D|XbSk-v50c+OEqEq-1~J}t|DmB}~PS!b9ugm!&xtddCX!0)Jc;lBrhxxLC@bH^s<=kHu#e -CK?Q%LGys6(vDqA*WZPdr*w^LyyTF;J6<2 -zUt*PlW|xDhc0mJ(;7DsnOdO!VJ3Z5O9r*+SqD8(XKYWbC60|}xfh^g6#uXHR)R6{TyaPw`xmth1#a4)!XWb__Is*J7-(O_etgGK%aWUi446SzYpw%1O)txd7_$AvXAB%q$M$L?+LS6FON> -MuChiO1ZnAA+C0wdyH_c8Da!`Ks(oVD5tSsaRDs)&cT27Tbz+0%%cJ{)C{B;pY@F`y~tSNKOR@ZSm3MsHNZlK~Gk1BX#>AShIsE(eufC;5-cmT^I%+h -f)Vp$tJak4iKhrlYjj?J -o)=OCl%-sC%cWf$IhuwgN0e*D0OAU(O6oA|n7L7kCSzflL}c!3bPs5B!m2>3Gy0di9KQ4qy!UWZZ~Bt -7P_z4N>7jUqc&b>-#K8uCw!3SZNOX4RlPytFZNQI+pKiSnj;cGi~7)FV38#(t4pn$tsR3xhBmd?e)c?Aehi6XO?=vcrKTO=ixv8$7zLF2E;+vP8>t -9%Zfp3vwi?#ZiN`sav9^A2vDD1X!uNtSLyCK#W%CbpSnn>SYOIbQv0n0wWTNtR0ztZkoDZp}mHqQu@j -iw#%@@7FxxY4m7t>!o#9E8&0UX;Z%c`t7~_rLVI$IZ#PU(CVR*A$2fYc=vqwzjh57If*J_>4rEMHpos -Nk|F^$6PJTN6nIt-67#w;A8~ull_mV)bMZG=}B53{+-1Z3IeivC{abN6#wX?UV;Z(EMC9ZRc9C)Dcwe -LW*r8=h0cPZO39U!RYY41?JQE(n?f~hg1aIMAM#55skVE)V%E#JL;d-n1QQwkhlWQ#LBxi>)zh&X -$EAV%1;zR#RM^B1&A0|tmjr+ewhQ@hr|jXy7z|V!BOmfU8^5vwJ(rYT^PwV0SYenI4kuKzX$Gj!f6`c -p6v27a)ChM1P!l&9%j>G#87`|oEMhx=m=d=5XkH76W+45B#ij~{9y=boN7Z{_SWfL+2KQdw7qB8EV(D -t^Qzm01++YY&-Q*@O*i>w{Ajp(OZ+8IPKxpltu*MAeNX$g~aV-*(j1W|XAG66;H+d-DU|$#S+Tj!7Y* -b$j`O+ACT1=-TgDuISeL}918dL}hg7Y5;zW1L4B8G|8hOF3v5L~R^x;l;Qpu<``IC>0z#HH6H3Z;wt# -x6d8p7En=VY|7W9Pl6P+Z(j^_?CVqS}e`R=`;Q=4upPj*Tbnn-cm$Yhr(*`6NT2`Cn&ptTNyVZ4DvQB5u;gFa|S!wlf*0&0| -^fLfa~JjpQRIvNtbmDfEGzLlc -I3#zI~?M}gr`DKGvS?G?ZEKydeA5{stHkiAxZkfV-@^V-C(r_rB*R{`md6R7Fm7shY*PR_X5D3M3=^P=FmO;cYo1$JY!=PqyN4Syl?9ih*u1i8&!ADIr} -3vZpH=sE*P=DYuINCI=tizkB<`?2l*fFVEk-oxT0=<_~A@56}rW08cMHn~i`bbt4tJ}bQqtEWLyqhyr+892|lQE3YVrW=ob`oG!XUw_6zLLiLT71 -FQ6n8Xp;I_@aCfuGTRQ^We{@LhIqQVO$-V(=+Ue+KqhAwAs)od)=SdtiVb;Aye19NU5Gf5MX|UPh1sGL8YEU?a3w6DnwVpU?p0`e8`q`bPyGiO<%8l~@}=qZI8lC%a -&1jIc)NEC*te*NYYt@KkkloId+)`Uz9-aR2hd9v3^k>DJ{AkoyGlkj8RITXcw5_`*Q6HGVWYz@~k!Od -CZ<`Uo}l(-uNbHQBLMPN$#k!R4?6Z=jn@0=)e!a*SSb;WC{EPv!Cv>wv0$>3BTJVApItQDE#=R04s&> -8L_pq5@JE6^Tx%U=~&*iP783@1>2$RFIh!I+Gd9U~!AK5_Q^c+f;|jk0(Dp`}qNlRd?Q&N^QmwIrc3u -wrY$`i^QWMX#zgEVv~*|9cc9_Bxg4NTwRwKar2Y -!v#ov!f_=#WG?)7QT^(lx=^%F>)YO!2Fm^KNl-;77AuRD41Mq-FY(z^W3x)Vtt!vvpzgtw=0HkL8b_Wk{`By1H -)XY=AanisB8$$0cAA|k?$8Z!KE|>`;P3maRcTl|pA8uthsNOqrMo}A(AK2ahQihLm;a$gzn}c*2eAF# -+|pNOrM$O~7>dNGajccNyAi{q)?+UbB)T2Z>LB?ICAjBLpAS(1cCWg%II;yQ2cpis|Ndq2yW``hQ!Gb -J<39dP`ZfHw0~=9W`r+&<7L4|j7cfghgUKe(#M(r5pu>eNA+4&=N`5zx+DQ+d#s*&eSYM0(#^SN|qFA -HtlEh{l@r*q{Ax|C5$9+f`@OZyoe`85uG7F8GGF -K!)ep=q+fqMcAEjP&^WBsJx+AH46vqC(0(k6%BG4+&X4G{kqq?!@zUy6qAHWS7X9iJttr~jg|`!-4N- -W0NB0avFlTKqIm-DE9e;EcKTDU7Es4La)h9gcaSlXyt)ryETI(ktg)pqXPIrW=PHhVR4q^^+H#$giKc -Jcgvu>!dXu`E`m~{~{B1^?nI55uup#KH1=<)Z_SlD8Wl)FEM`?v+5pr_w|(0N(j7uBi@b|Vb2qKVX9O -%oX=&mK=EFwW&?^&HC-@l0fqJ3tlxBl<4&|+|VQ?u${bP^ld7N>AHib|3n^glxlOP7Ft}X(N;IfJ;qq=40cDy97-Tjt42 -MPZ59z6f1#;q{JSR>icb;Py-FdBuzM+76o)w5U_+zF1$@LlKB4)=%n{<-)1nL9CENn`Y&QAa-P;dlQ8 -C07RC*)vGh1t)^EIeI6H-{}xn8{}lX`9s%cgFyT>!cX{iLX9zSw)9uOfx#WLu<7puO%RQA6Z{@`qcKX -l(mw-Of?o(qz794(gs^S!m*31mNDP6nzx-wac0$wO%3yN`UF{Rj#erh>i$_)1zu?wHLg#Z!M1hVk?Mp -Krq{jAWVPS2MKd;N8z2WR7ipfGAU}D4g&BpFfTtVIO7I0|TABnlCJRrk?T+SuT(8vhu63c;xNO=etOD -G-{im+%vDl$~gN8BAPiSGIMJJ3>-I}bT`wj(kXqUlaa)6m2rRNVQLODlIEWf%$Ojg&UWKOMn=3B^Af7 -5&)vGq^Z?rg4x;APNpT@`1#)`vJ=q4X>gH@EuzQCg6D1P=y64kZuZ|>G(kZ{XNCnBrTxZ0+S<7ga8}o -*+H9}Uz{M}-*Geyf=k4=j%6e|>=!!mqQ2umWbeh9v8#CCTyen~LB_3QQ4UL8u%F%IPD#G`ljz;o{MVxVhYBRmpzl&L -|g?@OiQeeLV{f^Y*n8GaeYOiCV^g$eIV{U5VJ!>GHvq|!h)jO<=#yE>$n)UAiykA0t`(|Ih_ojM&3N^ -`&9I~^a#n;yd%zl1C9e3N8I;r(Fj(D(OidRI`}1MwaY*dLhCPKIm9o5%N)@9Sn^^iyFTri@5WQ8XGt7 -`+_Qfrkne`^6{Mv<#H&g7M4AoGQG|eH0lW+D>P5yJ%;3B7*s+OuVcuxm#mGvLlG-L#3e!VR80{b9r!= -a7pO+9z1Ce(p@1yn6`?U#x`muGKIjwsH+uLPS5WOa*Y48V~Rk?In+yqm4j(fBSM6qfdHcOs>OUaC5a> --8P3kA`?!)d3jxj^YHTqUr2myobt3@KT*h`7zQ+Xi|DQem1^01}idfoQKMD+#jA(rE!0mlt-~53pEoL -UOmZYf48ZZM>r(0@+vL+D^jipe(@Up!D?zE8l9zsCGt<1$pb(V!+iceSiBWM!Mgvqi(iHXGWd*eSpK5 -d8{sLS<2ii?s&f_#r+tFdShrq^-xV&x1kO)ikTqpejcKKj5hG2YFG$E*uQ?{gnWiI;$7N|0@Od}p|xs -Z$uvfZ8?96ia1cX$mSi$(8DO8i-I-E!n^@Z(Y-t9b>woLRvhPvrie;!E#H(efQmWZje%lP&(3fRo5$! -*f5dxVB$U8=z`TcU+Lh7vn!0r)eSWcKC&35Q$>>AbGEA%kIy;IgRbW&XMt9vSb2_du;!t -&z6M24d_glnv(7beU_5I}1dFk3zJhsup=r*C`Df=c`++35oWyiBj=!ExrVs4(MNcv^*Jrl4 -XYqvS4{CfF!b~2@7GkZJ%jKWl(glJ6_~)nNlfz>UIMsESQIrt8+)> -@6aWAK2ms3fSzBJrNww<<000>v001EX003}la4%nWWo~3|axZmqY;0*_GcRR$V`Xr3X>V?GE^v9hSzT -`%yAgfYuOJi{kqbo$dS3$JIzXs@k7awQ8F?QieQko&1!Tkb;x#FCce%<#Y>gsqRv?WARy -PKs2_VAhDH=OC>m^&i|QskW2w*^zqiPhTPPm)5Dbw6b~7(v3Du;0()RPAU!(G$_Vd4q|EziP(Twq>?U -*CRU`IDNUHNOjrvZ2paGoj?Qq9UD2FspfYFHku<)rb9Q0J?1D|&QZJj@jYx`EgjEfNZ`fDZMlR#W6p_ -jclWPl+P)#rbf=PM9qKHt6qA8o)XeFi$|6W$;-HLw|!a5IN1?95LR9t!;X$$OXhqWs8UAvUX6vfVQW#kz1*F8qpkZI(=3< -{UTzWh_klJ3#T*gWDI+OMd;c!$5Gh#LMsOZ`R?wRi|)?5$T}TALnuv+R%Zg)1K3K5ryMYbksZ%-p@6e -w^3M3f_1*}Z1G%0u*qb>VK*cjy2d-%;XX^seKq_eqc5j9}qVn%DVAv9bYS^ -{ln?9QV!Jlr0Ga;CoWH2008Gn*lkug1PXgDn}G6t_yn3jN2rt%xRuxzs6Wrn;kmy}5wiZ5%aO_%mlXN!mHOs%b3?6}&B -04rK{i0t070qtaOe{VWXui3*UpRUC#<0BJcb0uiXVk5CHlZ-I*tot!+~9J7Za -sX;pHbBWST;~1OsfX0nItYwoV){`3IbO2V3Oa$C#T?>$dHJ$aqVMWBab3ERG0f>fzBW=;UQtgD1&g%ej1 -$mgX>e?~vxCf%vrvI01mduLI)`J`L`ftRclV&D8CQVsS1=%#{;Il940|)VSD(wz5D{*KRJ?!2VJWY>m -x%hc=bG!U{zgm77NZ89PcgMN~ctO7}^O-3iK@%REhFi@-7U*C~6sDumrZ*wz)y!D#%8WXai9?qL>2tyFp8Ua; -X8H#XRCYF57ikQpLkXE-cYW-{d_)WHxLIm#nIdMxNk6lqY}pB%n|Sf3q8ZkYzkMK5=2?bf2A|55zTM$k4hiZ*|_j}%M=KZ5Fk{G9NT)BV@c?AqMM9bUTpSK8s({Q}GI##m0 -HFiwXSHG5v?_Yc8|A`5?uo1Q+|2|z^HGu*M7OL{{lB03qQEw3+RHQ7Y^h(66-1{B@zkTdip(LtVmFpKk_O;JYHbMM%1@vLoYF%^Qu+J*tGUj!7`iBR$L@lBsi)!<+#e*~mz+-=4as%*=v&Tzfw1F#G^O -&u9e1J4D$vb?4lstf&r1fpB)3iT+GNvjo}vR(oXtLmx@Vj=eHkzKS&FhOf%9(dr8ooIOI65$V>9KnQd -{?c_3y7g4_(H(u_0%pMtX@a`T9vE+JXZ~uJ`rgve2bRi5Pt+4%VJf!C5n?0P1)N)`@1R!q*I1SrehBt -n5|iRKqOiZo<#TwpuZwH*rvn$too!XXrb2DOw^$adw+;dTTZ!t%q`aKDrdi$$AUAIURA_y0wDmp^jbX -6{{&H-OP9`z9-G>KA{7&-p!B5+3Xs==bf%p$94@i!hsMO-ovkr#vLIz>#Mq6K->V;3M{hZv8wQW5~pr --=bmzcRZy}}GPl;e(rE+ktY9rPZC6v0m_Wh=pSPf!V|a&>9J=}5V<9=dD<2X(4+(7fwj6Ch!=56m(^D -3zBd@AT5F2mNx=G>GT{ziZ`*16tdzv^tTdlgnDe%qswgWWePRIt@r#n3W6YV!(gxRE~bLSHa}hc>4tL -Eo6=$@4}GB>i`&Xx1ZnR?UOh-arQS5D|*W|yfWL-DBdscCEZwY-(l?NT#|t2F0|Vv^Ji7f{fJ8mh#pIWhjjHr7^4=N~_O{J`FQ3JR{O0Sr-m70Bd<`fK?_cQ^Ix40v8v<51C7>& -EkT1e66kURw671;cTqy{OS#kQZnxJ5F7+UpFoD8r07w6q(2k^Y`zu`NOXkzK<)hBDI4`+QCfjZ>v3Rm -Y}z(OhCM!)m{EO>4p4ZsGzD}e4N`G^~HH*>ZGqK`6|LiTv@JJ2$fjiJ2#_??JhjWmpS3;y|MX$x36?Q -v0VqbFWnffkN8dxLO=4yI?DaJDZQ9D(Df12><$YeY9dm-ot1BI>Gm?iMEx9j!z0YPQ?Q)kK8f(Yt-p7 -@XiTGPIcVe!5nPah9vufqjJqN`t~&+dW!K#+BMwp}SRP}&Hi%z%alxQm)dk-4(4Aj-l}gty<``NpFA~ -68gT%Je{tHk`0|XQR000O8%K%wh$0$oIAPN8ggCPI_9RL6TaA|NaUv_0~WN&gWb#iQMX<{=kW@%+?WO -FWXdEHriZ`;Tb|6iYC!5|Qslh(F;K?Aia>T~S70C5^5IkbgsP%CmJF{Vf!cNtqv)6d?SeUQtCe955?p -n?I3%bnSI{N}Y@m-&u_;kuGl$-|H&yCN?oVJnfRmE@s*9*xX8-?1d6l}u7$mv+gH%YDJ?g_4g(Yh+6c -7TvK87nZ9?9!8^4%-3X7aUsGe&m_-eh~)k{NqNS0d_rIYt5Uh}SW%=&#AK3ZM|mWm$7HzWEa -qjXv#KDgICZKd=m#8Sta^$g_0c!k|*D%~&lBa4yRHQxXG1CdhKR+DpLm65+ -Ch6PFg+fR~2dnR3b$oif#~MuDmp*bD7vO5)rwneoVsugm{z_HI -3;M|^Xe~aY?(Ptc@c0lqBfBJCs`g!>7;_d6R^NXvCHy_>u{b#kc@@}f*#Ku5*G51Sp_{g)A4P$Z{qy{I_SY{S)gh0|^==h*5&E%Y*@pWl!$eE)9XV^w6>Nj_-vZ+-eil&@ws798X2H@~vWRBzMxn!5WU -`==p!Cf-8X4b^L!-vsIB>YFYl0mv<=P%aW~i^V5{@lVduY(cdjYrKv&z*TgHnWuFp|5h`X>m8sqUnE( -vFD&Yu_)@Vm52?vhqAJgE`_6_%^5#2!SS1X$2x@~}VK^0X&@$rd{06M+P6PPlTUO^!0O -cK62|=s5ENgNdchsAiD(4Pt0JiebKS63MB`D%y<5-w-F@evQ2QlN!lXb-mL)0q_o4O`=x}MZe3dGHk;I! -t+I%4rA&zwz0*1^@rEQ^C;Y8oeYQ6d^DmoJn+oi{Zg7PUz29SZ*oC4KvabuaX4U)&z0hm!#cgo|$;gC -Vw!uCTkiq&@H_6~hpE+dQsaf=yf~D@Z)Tk5Z^~3S&+-nS6sW9W&Z|prAgf6MpHSPFx8_PV6r$X)#v?yIaZQ$p9Rp51xzm7bahK8k|xyY8niPLuiytvaln=*_)SY6+N!R -p`cX#fZ@>i+Pz*v1e2>*R~JN82&2mVmX};bTKx)H*l?vaE6m~if$e~3J|DI4F<$5ScuvMER*ol~H|&n -14Di~=ZR%D^KFINejLkimKsC*)anPYJIZJDL;P24q1<(9uLqOqY><3*As8Q^&_`6~OwX$Ym)@lwQTC! -sse)O^6o1R3_kAXc6!mUpcTdPQylEtak)tu^^}I_xR~z!o -82QjY5m4KbH5|3=-m}!~KyfQ0GRWcc7E$FuRT(IidHyC7v@FsQ|1S4oEnD=g2zwXgk$eztOyxj(XY7a -tnqJraK6bux@pISC@tR5V(??;m6_Q)U}dlCKXCwl{C_yfAHht>(8HfSthZn%iF^<(VDSu#lM;sm1Cd-MJe8T0IewT-RgZ_*Iz3bl* -tTgSR`(yY~i7{q7}b*ChR7Id)GUcdsLwTa5vR)}W6twQ?U|JZSz_*bLh27JIxt1hA-7UVX#(Z5+Q_xS -6q~(RD9GD$kqr7JG@f(}dcO4-tXk<4v4P0hm-^*><{e0001RX>c!Jc4cm4Z*nhna%^mAVlyvhX>4V1Z*z1m -aCwzg!H(ND5WVXw1}YYi3Y$S1AShhyC1?Wpkfca@DVjo}q|usPid0BycYVpXXGl@9EqgZ`0fr@V=FQA -|!(nT52Sw2iZZKFBAi7=~2fVUc4i1a)J4xbtB^0-{5Yu<98?5d0rESuBJfSC?H+zS6`!N~USu0RBcG~ -=eUk9}AonMRwJ+E*14bCQ3JS=t?Els;slH42N(7<)fgh{P$hdxUO*RFq0^M7Ccmn2DpEz~=$Ef$p?RK -v|a%Qp$olcs6-|2P=HuHwdUYlXUz`?aMg62V|Ed5s+^2dy^Hi4L9C(y|34I{}v)QP@qWmq{afhJq -h|JM~C1L0D(0i^eNMT9g7P8wF+~wc~BF6$%0gWKsRe1GAmaKy7rj|tJru;S-DMj@v%Fm8ZDJPvhdG7UKK -OkrY=SfvtNUMekR!?S}_I<_6BPbtl4bfC3~IxK*Yay2)t>86IyufJ5ldqkJ{%q2SkLQ>_ztTEftSX77=}a_&47k7dtKX(GzZSzXerM`6F9&g -LI@6aWAK2ms3fSz9}1+nApg000(F001HY003}la4%nWWo~3|axZmqY;0*_GcRy& -Z)|O0ZeeF-axQRrwLEKY+eWtE^(*G8C@keFY?JK0=o(Og+%#>2CTS3-yI5ocilRo=)D)@ZkhaxD|NA| -Ud6Tl8?yig45@+Vj%z3|#}Q_$~~ -x>%S%;XP9|G~o#nwA3;xFsAKv{ai;Zl*Q)Rah?`0|Lt~&mr<>p0GcZI5DYu1-dUGwq->=HtF;YQxU$b -z0-qz`d+zCPx<+{wK(Fy(D257CKc69GT3o3g;ie<`Y7zHHm3UC{G4#qE39wa4JmySBMKUJ%e<>Z`iBs -Yjmc{(#WshR1$+|2kPQHwQWtn;GTGZtTK($imyRJLrT@M3QxUNNPc<}Hg9EHQ!F*;-p4tYX_JD*H0E?(z9z5Lr>e}4bn1i*UXu+RJ_qQlUk+E08#SwGnG605{iaBR8UJU=ZDZ7QJR@FsQq>FG2hnDjX%7X_(H%sKY(`|d5rwPk5%3I$DP{CruQRw ->TT5&^x*>!`q{{MO6q7f^PfK#%qR`KHC)4wtgEKH8XM3}-M59BV$b;d4-Nwh5^M+% -N2?s+FKb~61tU0>YaVg_acST)F(QE|_W*|f_cpOp7s^%GsHLeZDJAUTUi@c|utMJwoc1SF+0+*7!W7QlK8;%Ckyd%Sz1P)N>G+0|}-5dW~j)z923DkM -p_c`a>lH-~?UAz~ckLS&$xH;w}FfplB=8$y<0R+XIY^Nzy=v*cqrw&OxrFz~>t}D`R6s$N`cTTcabqV -j$7a5Dx03vezBxb(kwy-nGy>QbnpRt9~PCwNVeMhHs!yaCU)OWUp$q@AugR_5DxqbQyXOnLp%gYG%M) -0)+$jiO3u1p;+IFDcmoI%iI+025^TT$#&`Kd_vac1p&D00Dcx8Vq{x5&156DB8Oi^Uv+fQx@q@C1>E1 -xNcs@j3tdMDi8r0n-xsYcIvHqL0Xh!3v9Jq;SxBq|M=e^=#I*K#SvGVxMgN}9p0X_n_4+T2(kf`m6fMk;Fd!R -#$E@oFoMO)qYg-VyWd3gBzrZY_7kwF1SvuqKUQ6;&tO8qrgNyHKn_*l^`)MO;yPS8la37&!;+Q4otV_ -8HgahKBnA0x*cU*~xl%4>WUZ3=M0qkv5(;9K0K4ki1&iGZ3OzD?~MXA(4tXpHA2v4^jjtOo5v#BwXV! -@;u;!$CbJQZN2^#nAo}o+@Y$WlXg)nY=QLPcSq5 -rZ+?c%a&o0sVZ?^s)VB(!tfkC8O8$#E&81eOc=VAlwz-6bDudU`sBwb+b_*jD?G=>tohtxIYB3S_2b9 -^aea`p!um9h3Kpkp!S@_1mCMAJw&SIEI-Rq-av#}l4Ph~fEXh;MBOqK|1AtP?SeydK(plT_8nU;)~JX -qP+Dd)4LVC}JPSo~fiiJNqC?}MoF>bRX^dEsry)5!V~+D4&BXKKQTU9kg8UA%M~|!#osf4E>O$5{`{I -Ba$yh#QFK0W9@*;byqp*hjB&8$8hD*`>(B*G{rR{XhX4D!Ei0AR18>t~i-Vm&pMY$6}R7-kHTSW$@ow -y`8xhA$c+Wpzl=m8u;yR4cu+~6qCL{l86>Wu;$+0)a8xzM`q*(9}%fYE=l-bEj1dhU}nm^J8BQO!;`F -$=4<=L-)&|4|6PPbP>$rli#I0qxnwQaNthk(>?yov>98@I8xj9ErmPo@{f&&Ti4J73d?NfUzVRyaH7j -B?Wh*L=+HR9>_mvi}DKSM_{~!C)STagQpX_Bz3>fY1niB+kU?W?f`}ufI$BUxJX)mYol~o7*`LTREX3V(kvWHt6^Wit1jvRj -c^EHZQCp5@oIUa9o>jh}tT3uwaE^S&&8Rf;b;v@1LT9lFxQWb50I7kny;U4Eo7f;a*{Gr%ABY!^7=ZM -V^^^nZ0S07iupAJ>m|%Pw&AtstB5%ta{e1p(T~y}LI)KAe0=;G3ZAQX-2LF9gX|mR>eB2fa-Rc@GZU? -6V^CEnXK0O$%3{F@?{_`nx -WD|trX@+rBI%#=IJ4FJ>qVT+etM-?>D0JTeQmJ1ra5{D~5(f3ofk(908)eZ*c3dWE9Uix#7)VKcki`pUpzyfAlC#1$xEBsxeW%A?Lo_A3DqpVw -xB|Lhu?{#E@Vch&?C+B=_XRA$L@bmVwqFX}(b~h{0H~-QogOKi*dAau?S7!1G)89S0(wf=8#_GE{$9}~reH7IOE`hcT3WyOOaqjmep;;@r0b$ -pB+YpJ#5Je<2bji0?^8q!R0$7+IiOJHottT-6uAtMorpO(1H)4MZ2Gq0oVGs5Qf(E0o9s3ZfX9m-|9<&QY9xVRj`>LY9y5|1$Inr&JdjALSOKrAz(?ZhA1$ZUrm{HWP)*iU{$Y1<%K4j4$oH4medsGJSEFc@1Brs2K&wadSLCx%kq0jzswsgRlFHeK~kD3wZ -iTk8uz^L|6z&ztsCr(|G+h&*+yMwT5vJOg8Tn7Tj?0HaiLj6yCOR46V)fcW>6-Ok_pm9QXoPWub1m+$ -4DPku20&UdL!vvlsjkswf*~@*=l{*OWz(YkJw>bpKB?c2YCpc~*vKZxAqI3b#%kY3G5Ej$(Xs%*kYRu -_#-rOWP(;Kc$QF1Cd2&Z_Z%?S&Pjic4~i+(hn?A*v?R_kM?=7ga*iJ%~Cen7xH5WeVc4&fgyvOgL1Fz -O|rmRyiCBed}8!7ks0+N-K)(MdVS2^jYoqF+xk|R? -oyOL_aQoP)90z5G5^q;fnlnEHb=`P3kdz+Hl5KqY8ze=MHX=od37cjXP0o}x=0L{EMona;f#YR+Mm})&SM(q$Q>y-DHjZkW7KSY2{ELoiS&KoW5STIlVu -DuXgfoE-k(TutnCpv66d4c=M%%t>^xi79;$x;W@e8id+blw9>)imKdQlPsG2^yg7ci>C&}H#r@kh}7zIC#wg%5{MaQ>?rIDpEqYHRhwEmz?Xut|YYW`WEoI -0h5|qDN|&GL3dJ0VKk@V|8>@cygFH@Z@$M)-dR0nD!*`;^((-U%vR@H7&I_k8>sFlE}Joqx5(JFwIH_ -PjUBq`l@41<#$^IpG^fOM01wZ0Yc>jG*%nHwILqQB+>y$K8iQ9X7~iz{$fCxHe?7+BxjR4kf1|n8qqe -5`0)Ay$%KCb-wo*DK)pM5yQY4M&LC-$-uK62A`V-|D~WmPg1hq`P7u?oYKtHLwR8pK5lT(QVP$_{vAW -IP{p~6Dy8uoG^vX_kC5}a2<8Kbk2Esk1dKI^?al2h0El8%8Qp!Apr@`i==@87>wJNME7#YxK*5UD#0h -%)5#M2l^z;J+jbkPk@aRT4CaNrD6qH_76?WqnJ)%_wb4m)oPZ7T-W=NYyKWM(_>+z>mNR@Jg?w0=zfv -)vnC8&D;YiPVFZ^+tyd%j3RPL4ezm-Z36MrRP?w$nKhYxE=GwY@aSLDiVaaFLx|Epnr-35aiG*5GV)k -GhaA_Jx}PKPYXJo-;FN!EfRWWWj=@ -(TvoDsyn~RnF#wKs*zs6tb^-C+86B=7q`KCD7MrOMS=tZyRtT8jCE`0!+#J6}uZYGwhSWq)@9x_<^|gV?H9C@^GIp~N(8Clk+fA0@P -yX7c?W+CaL4f}RSlTwT>3grS#g|$<{uX{u6IX9cFW-HuFqxB@xdBvX>TDNi@bc&^Xue8%K>rWlHvzi; -iXIxhLt!1SV9kmP;p`xWt&!HPCQzK{-KyHYv$il1CJtDS+SmpZ4tB-S>c#Dn>{q~r<5yNN;BCSfnUig -tCL69F2wc=5ZE6R2IxF8C$eB7dT(xuUKokc-q)u-PIe2guikqoDfZlzO#%Oiyc}r}5cDX-pr%fVuLcv -!SHwfN?BS7SsH`vWAUf6O7B`6VrbT9|2=wT_SWvknVC<)#Ti&4w?#9%Y%>9I*zIDN6M=v5_{8V9llML -tL%K$Ku$(nsM|e5U(@z?tdGr&KbzwR6nnQo|uc)(|-{y{x74T*-u^&6Be+)9iudSdhZEt*WhMtFlsvb -Cdg}Ze+~lQj!EOo}@+uMv6n;&Oi`f)`&7>RhNryQY{1u!a -NT4ng+Zv=WeK-(*2YdX_-^UgNtmzg7&&HOx7Yl;7uV;gmi<7*aWTj%J{%GE%^qHRcuBTVzgB(%x_#qJ -ZW+?!E8c?D?pwk*Cbk6$E{VNboeYp1^PP|R=pKh`ow$+ZO$nUt+{e@cm8Ec-uuQZE4BdtdN!byV403R -rLV9u7-4nD7f4BlPLZDHp_XaY{<%OyB~RwzxyYkjmX=(+8Zr1o#F$UmnxZE|3KYy!8;CP2l{oE;D0hR -qM_Bma9c8Vu>^k!<)*ulMn8({nL>^wp#3@18id?7&+YB&Rce9_dMXGMc5xKTdzNd1A9LTsfMBuuHT4f -NAg9;JqTlHg?{>qrhRS8fZ6X7xVE-#==z+()vO6pme@5Dfbl`@2M@i)O?yN{T^Sm2x@vh;;|6{Ff8l? --YIhr_XwN!77}32jMPy`k%XXIaU;)6sgbv{J0;*?8iWiKjjK+Y+Al)G;OPumMgIGqOHyC9(#~Kpvqn%`6nS|+&OP^%&9o|5l4P~5w -MY^si%MzB_}Zwvwj%M*lSvpi2Q!&uv?k7EE-GbgvqY2|m5QXOb1N&ZZIUTnq}=i_zv7!Gz7?jKZ{*9- -n;1iJp`P)p)_E@1c%zMI7GTWpF_|O@&vR&d$v#IviT~8X*gIi&DmBrU$z-z0xiM_`Vy4!=h>e{uCk&q -`irn{mRX&SSLatzie32!C*37Mkru&_u&2v$R5b|yuh`WTXAt5|JB(BV^srJ3Bo=ZA$K3Z96=@+1TwhOC%0PDdk3Cn`*F- -b66nM$k$NC&({M89Z-yg8lr{k8DQnBv=09`f8gY^A=e@^&5Z$asu-0Ds$JlubR0S`aZKUa_MZ{AS9RXIeJmj{L)X)inrCN0~`JZHI5Ppqy;KaaS -o(yG#uThT?2G~<>cPf=v56pcxIjr3x6uUdE@(%9Dv>lqzXy)C7cJeR);&0ssnDxu9X*x^RlPVm$Xvhc -a_10-6Oo4ig1bhmH|^l!qkV0D=bWU0cYx{*O~VOxomY=w!JDwi8+^F!QbuKUr$T+$eG$-YJ&D$#;Pzz -=;9g6RAH;qmtV{)ow(m8U0L({aZyVYZ>9y8McV@XD -ybla1UTny-A=^R|Ul{^-M#JTZxztSUOSWp+w4u|un>+8k8*mc>)etW^F{HFNbm%si8>A;s`(8jB_NUJ -szz=F4=K3+wJNUZ%8A~^nwbE8AuMMnOa|2>go>*2ZuQJAwk!i>-`aJHC{rKZx^Ik1-v#}IuOsYu@XOo -~@aA?U3gx6A17rd-_-r*xd?%SB%TDIXOiiX>c=&hhqZKSO^)gj~`>Od}-vBWA=T8Q-nsnADNG>>vPg~ -cqTU=dd+WT*)Rp(T<1PLvEK4hap+`GN=*Y=s4Xj#P`q8}A&Kceq8a2{^~QKd+m&hX5Z)ns<=M -jrh$Pk=y?`m5#Hy?}z_$V)qia>j_#}Gy`kmHd~QZS)IG8ecW&568oOCE(<2?W*nWsR8vwQnd}>etE)W -yLeyar9IQF~;HqDE8!ek~z2$vqH3e}H)3dWijkn -=21mW)}qkX{q;f3@F<$n?i$v3|phQ}RfKw&HKpTyg%)u=O|!wkXU7@RL??@p0rR+pRZ*z)GdZxwV*sd -9v~Ti=+9A9BUh@$!mtXsLCelS<6)sZHFYIT=^NpA_R3ncdEf>~`H0o1mj~W$@zVe58|~4)^Zdb=MvN+ -S>2>oTkjWGDv4)*My_If)1$(mv(9pAZ}F2*ZHA?HqS&jYEgralH$8zD+?<3O2zC3#qL>XcK|5(fvq9t -sR?!`NN!8)*%n@J_5D?CcRlq#q3)qOpd(=@?>z5Au5IwCv3L@Pi)WtK)FNoeHtg!-hgKB5B?E1Z4lia -`tbOpk%#5B(NLw%h!h+r1-d#I)&Y7PEU+Ezl+9@QW(Jy9yr`rF*sg+zVCj(XrhdJ=xNy(aKN==t+y5@ -Si7=m}7JMr%eyPNERPRzzV8WIIYr8kzwMU+NJt7Xo5!vgGtS6T4vVJ9B#!bf-@ZQ`H>l~q+y3MsF3vP6}@W`zVjy`)x(7B9$yfLQ%+7}1AetrXcebZh~k=k3`G|FLMelqu)o%HW(g -(s5e*WHT85C3`>3f=G`Tcvd_X69hx!u>R#lY|&HNxsBM;^rnvuMhP%TpM>)9?n;1TyNbO?P*WH=EL~+ -GoL=U88H<45iMOm!bld_1XSQPV5tN}6W4@ZkUhVPU%*I*T!g$t770pATJDZf5IESx)Z4A_Efd}!T$fV -{@M0(On@WB+{iCT@cFO#J0T7dstH~I&W8&ZCX;w!~Z`<5E&D(RoUu=9Y)m$A*>hI`o21Rg8TD8@{kn8 -nYtGdo_R0AUf+gV$g<;6v(4Nu{uRBcQw)rBZ8j3pB;p5L7p5_?dNoo{QIig;&>d?+@6^E&`S&I|xI`7 -cmQ0|XQR000O8%K%whP$0NEV;}$kv3CFf9{>OVaA|NaUv_0~WN&gWb#iQMX<{=ka%FRHZ*FsCE^vA6J -!^N{xRu}iE3oqEu{4o6wlhz6+)m;+9e0yB$M$U6ljca21j(!^Qsu*n`eA?j-Uk2zAV_<*v$OkQb!Hrk -1TKJ!i~GLtvdGtBGFjGDUC7BqBtWk)-;sl*sU$g+kq@A)M2h7)sgC-IaCJVH2!^z7}6zf3;7yu7_}G0;M!z -egU0@FHCp%Os1_GytcK6@9k;^v`6Y4%Q9OsBBIb0eO9SeRDf`d-3+=#Wi$2qK8+vKV4j_hd1!_n0dyB -7jNENeVDv@ar*)uo`i<@rDwMjmB-(RXQB(8+_%K#KyQDAk-et>jiU^Jkm2kq85c#o8;&!hcc?Zi -NKzo@2j!_QbOEbjwE=eWv!P6&^A0GH8p5aFh5PqQRLz35}*&J)Rt;YM}a+T$Ur1gn-S;<0FcT&(9Wc{ -Io&^x!@#$Ky0VBeo#0E9#Q?G^IGU!-Zi1)Ro73c^jS;$jBEQ^a$`Pz-8-NHUSbVhXW>0mM8_WLA|yXx -)2hTEvORh}21MHm=TU;B^y+b*nXpDd_Q$A3?Q)r(eP)gGU<_)um7hUsBSXk&P)X#eC)^y$*63-j$nrg -UU%MW`KHez~DB!0}*EnalqXVgw}Mdj%AX~CDPtbly@;uWNb7J0}| -s6czt9k_{_g6ZAIAbm_S_~=!G%VOA)v3GIlGp|8|{lTb1ax_gkB)ZE7KrH?;K${|;BNS)h%d{@NzmmNXg9&EAEcN`nTd -DH|I`50|oAckMjb1z!#7j^Segp!-KGucn)|u8#o|3t6_r`+|{+#PAWtuYNFGM`cwco -8f3g2V;$MAP`Xr16ISqcrg0Z3>G96XWR8iACEqn?SHbs#wjdt3LqT-hFlC?pdQUB=DX;YmX6N8N9b}^ -4?h)5pmkWQUVdwF4vA8?J>3nzp#U48fEl!U+Dci|sm)=DSs7Tfq7BJqy9$?*9$j9 -csqGyvY-CJSqR}`3I*)jWm%yjDaj_^5=f=7g+XOiKDUESDy#wPrV$mAe15U-Pu86>m**cxhz+j(FVd( -@s&Ec>tXz{!&3pmq<3%SN8+dIkUV>-nxgkHeFHnOO8h&dGUXjnu7#DW6Hi^i8}nG1iu2^_aaDKlyhv7Pi5aI -yGyVrAoy0u>Vp{+wMOIrHR=yDAUWuQ8V*{Ocdn}R)PzQ5;kQuhd^??SH)@p%ze$3e9tX_#tU2O6a_^Y -^tz70TBQMaI73RWQ9b;2Pp5?MB@yQR&ehx)j(e1Qd2W*z@o7b%eUbL}|-(<8dZpVNja@v{X|#zFONGi -m&zmSyG6LuRm$x!LVWO5U+%S{MA@NI8BlK<}F0qwk5Fb&6%da5K>IWC5IBtZSI|I<73)_YkMG!~>+|E -W6?q9gLL`fyW~fR~ma`R?*pJ@r331)w|n^cej&Q7w1>6*q;8#Jezbg;-7<0-3+=}W((3Y1O+UcdtYC? -y1c$Pzn$FNzPSDQ=7L6gesnw*&yP;<$8-GgE&llKQ$UbAvU_a4;b}P=44MawiSi65*E~f#p)SvknkOh -q;oRDffNhoV#P3-o^U7}Kagj4R(U$ILAARp->u7p|ZSp+z9yMNJ@5Q=Yc~2DWdr!mUc}DIdyDw4p1i0 ->9zGwmFmvC~tH*sC%>}R!Bs=YUgXcDkrP_WG9P;JiGbAX5J3e^;-ME%w*sHNb09bDjgP|rcb6oGaC{o -=k+r@x>+lP=bF*>Kq50Bsy0YrzJ2UwuN -vhzRt@j`A<%qtXp@$J%8W&%_-rhOW%8N!GFwfm-9`e@ -(lkavBUqsV|M5|XXBuANKi(TF;B#i3b7z9u)Z|GG5C^`IX`$R5e_y-+BWVlLv(y4?&S93rvf8C2M?l- -k$|dZu)+zqzPEvJf`J0=UKwitr-|niI06=YJs}qP(mf=8;XtWrVG=ou;=38m|4EDvVJUc)&Z(g(KZ6k -q9K+A<`&r#jPMa5)B(iT>z+$@rE4hk&3N}A);*sEt^2w${)M=A(T6g5YcL_S1_0{_(>%Wi5?+G}gG7!&3njQG -sDHulX~t~%hp9Zvyh9V08*c%`J8+XzUe%$f?+pzd7Unhh;=uiY@h;tUVm(seaaeCd5Vay?siQFJ&xex -jLQaiD_CXRH1fm%en=(9(I~=cz(inpA+e*mw)6og1Uhi+j&zE+XBL_O4H^Yeq22SUVWON1Eyb!k&gPs -$J*$C5I;t@rb+c(eA|U5zJ6n6@g!+ipsi=U@;zF|uu}b(^SItm12tUN)@J|ja=_xNj5IVjwNSrE_f(UKxYh< -FOFs1-+Pj%b5D6)71uQGYuNV%(Q@;EA41tlg$RY9B>QS=q9O4`zsJ*}P$?GQnj4Fp}2IMtGy&2Us69RD -DBrIsRoh?5lkU*WkLJ-yA7wS$M7*qj-_s(wM^+oo`(fwUtL`{WU9rc6$#CCrr%V4Qi0KV-aB=;s49Y! --@;g5Z+L1H1jyun`Z!7zOBQS{mM^9bULEc5I*l=rGrrTw(+Rofo9)Wx}L8MXkPRzKT?e*s%4Iw%}S%w -rCIl%cf|Y=T1SsGYCVu)#|H`Z~)RPwP^THekMXqcSs$xWU&C9v;z{Hnd{f*;@gwwM*w3ko=4B4lZd0v -2>0o8Qg|gLKNl*T)hrV>r#0)-E?9aS+phzWUp{q>=+&QKIDg++E(tnXvM4ulo-EPQU1LB2=+E<&l}Lj -Dcrz)qCpg9^=?Z!V9b@Czc3o7lmv_Fgjs8Y -*Zb@dt6--eMvO@-;eXf#Cy*pMkesI?A$WNFhw8)Va>cJBoC#kTv{WIU0!H3D4zWf7(kk3|TEM+T@#IF -uCMEaFmO1a*2V`O8W{&q#KU@gE$JK(gkU0QRnDfFO1DV4)%(6lr*X$vAt4Y4}xUO*Kn#pu`9>oje4?MUIb-#1GoZ2beyA?=*rf_$j#f)5VKd19zSV|3@toNfDT*T<8#LpFu0D!+Jx9*Lc^^!;{ -SFN|HGxk$cC*F04K512I(cRC7qxULPGYpNl(g|q$aJ{3QHc@WBWqTmolk_ooXr*>81>cgxDKwPDeS;lgk_K`oV -p=^vb@6etjzGrBe*>WeL(O7xaeaMtJvo2#;)cUK!^^zaO+q_4lyvBw~L)H7O9jAE)G)mHdjQ(V4x8i(}szyRK(Ae+9;r7sI{RB9xmwmK18w -|pUZ_w?T*Ud?kvfJv|^^abekHH+Vg~=Rs!+y-2V-wKMw_tHU$W&}N4W_B884C;xp)Fjo|Y2_CT;gJse3jnJ?#yeR7a(S6z8}AS -_upv;b(1k$`Rmyu&B-1vNm -{;Q=vbk+G6^q23!8I1rk*maz3M@l7k**rURnBgu)<0*&+A$5jtmzpa&tPR3MlM=4k=#fBE*sIpLL4hS -+;vNJ8F1;2L^tLOt)H^icLcf&Zc7w{|?-qt5+cFAh^=>3PGtNXDsl8=^sLPG*tc%M2t^fkFnR?ys97b -wIU-NgyRxcSXL132%)uA3x;Df=Qd$&&C5%XWC{+(w#d^=h!>tXTf3}nh*@UNXXpccqNh^!AW;Ct4uOU -&_qRyq?*@-2geSK*l&~`)XXcq4VD*h4e+rG3NdnrlXM3jqS)ywSAMS=>;FPbBF=V{3XQpsl<|-tuh?Q -AcJ46K_V{WN^7A|)9AmAtBk7?}<8C#MXv)LDi2;ilUNxy{9=^0-CofsOTqbj!jK`TCZE%=g3<3}pAjB -1iNhxbQ7I!%+6H6aSj87+Tu7OwOc@J8;LW3?oEIJAzLQ^cw&&e)q+{ksMQWBYbVC${Zl7R3jta^%`oM|v) -VhUp~k~y+2_Q9f8e3r^J9*;)nLK$X9Zjtd=leG8c3-LVq4qZv-Dp&dkc<_DnEzEEpBkKX2f%vZ7s0O; -)%F{2uTmk0mSp@RrOQk)Uf0MGTW%=dF*T>KI>33+fMw90F`0JCezyALFqoc@~hYpY?8K&C;LDRNwc2} -?#!1^H)D?G(tOD;Kb``ZgRK4J@ja>^N}i(IlTj(gl}Z&7BOz{OE?^qnFN= ->zcu8r5W_=YeWIzRaD9~y{l_XtqIFq}7Ij$i9dCqVa`PKTjf;me!J&Bgi8*B7_fFW%j}zq-C{wH@k)@ -Be!B{^A`T0+?O7D|JdC47$t#^~2|5aeQRu7>!zX)RG)8Ngln#qj-7c8O{6Rr7RQFQ>L(p3Fl(b4zdLo -1F3&b*7cerbD~WCDaXEm1mOtd%qtAofsEYIQM@Asa%mWY#X)8}QlmW>%8LreUC~zLHJBP?Plfykh$TG -QC*n`ZOQZ_oHLUtbCARe3eXYW*gpakM)s%F?05ci442OMj^7S|3Eews;FNc=%HE43~q#c}UDn9wnX~W -g5_+Gg*akKI3znz6#`WPDKw&4J-{@BBU{)Go&?9M0bBE%ktXHECE7b6&mXbowP)bCwdZdA%d_r!WjJn -=IF2=}cwO#jigNGcg9_NAGCsZ6sWRRa^L>NzF}#`l_->k`Xefo8oE;_h_gg1`~MxFJuvoiEcu8(2w{N -_KfgB#62)*|PxrtgGtZ#Teo$KoB&R*VK~yuGHsTY(L|iAG%2_Wbk(dB$9Ok;vG?_QaHLQJve!Ty=$yW -Y~jQ$776M|ag*KJRPUT(jgm9H>6G~7bm~lSI<>~rN+wAbcZ{%OiZa*)D*4hPeNrlfN(r6ySx_LII -C3sTjUqO;l9BFldJ}9H!tq}EE6u{$7(Vd7y9!7n@&|_zcr8H708i-6F5-cB;{!8FBNk?M!{ocCG0jHU -$dsAM;76jSgTqfP(u*Ra0`R;lP(l*UD{|CwEy>%KE&l2p?+1Ru($G02o%g)7^jH1S(%)!~tve6!Gde{X@;;5LWnQcg71x@dikJ0jwbMl ->6p#-oRQkgnfR;y?l=&=I386iq&ipGoRg(Wya|PUs_^9OxT#7bO4J}+}#~;+KBi#dK$ -*gGL`4O@vl(sR|mLPaxTeEJ^p$uPL7UG#1JO~b{5!QWyG+eyjZbub^OiA_uq|R@D6ZXREit~ETMS_qi -E^^V%1zvzwR6-Ckm-+G9 -%g?!9M1S@s10-c7AsC(P<{ZFtkT5p=9Ia*=vXzwV>vLF1yh&KI5xG9oGlEgg|09#;DKo2-SJFq~cp;ISJTKLhKS@xU_OSe*H$6>l -~vr*d$eW+aP86{PPZ`=wKU$1o(?s8}k(k)|DYQR&H0`~m5_UKN{(p6QA8^907c6s+LO`fdF<8Ok>e23 -otLf2}8&8Z8xmwBNSH&knsT^>PGBra9Rjui)>ec=U#8kAOUI2F6OCxo_8#m*dl2*VHrM;=gQgoYp -C_;l(UwKi%>|#JSvx|6Clpm!0>jZf?#a3pl95qZ#87U!ZWqdWVC6Z}Xw67V54tO1YPL%$RSN4nqv$q%Rj_EvaIwFT86=h|>;>s}8OSSqVZ1y=kTDZX0+U3f9(PrC-jblR -?^FQ(HlAvdMr1=b65`Eix*;S+-IosZ7#ML$UK -8HeO9=k8~)eePry1TE{H4lX<9Y;qIXqlWg#n%xz6lZQ -~;KJ8*7Y$rI;<&54*qyynvw$u9z$Y3&8F*|)_f#yiY7jPK+W7{LrY -=05vS1J1+`8cj7rlxu3~*B?RM-0Kp_d#H0C{JSjtSV@y3@ -hHfr0pwOa+M~USMkxc0GHa3M;}~Is`%0yTIbs*{8?u~z&N^37jYH%-czng&`mq(Ox%Y*$+$Jc?@^xI~HLyT?-Eeg0MX~`M&uwo-yn=Ttw!E+QHfZmG(;()NTNva^J6laDuk3N -3x0&xFJK*2=sPD&Xm&e<+v*djJM-4)9K}-`b;Zqmzd44RPS-Ql_Vf-pra&Go(4v03WV>zKCX69a2+ORFf~Y#`AzHkXbseE{;$`4 -C}Jg)FIjcWGTJO(CgGqjpn(Gogrwz5;QJ#l2;{rZA>1=PTkZ$MeC+Y8;ZPDwu2a9ZS5wq&{?NHUm2wt -=@o1D<}FgAX8rN-ur;JW@Fmj0oI%+EA*b?-_>ZY!iiOOhrY_k>dd|H+I!^`GCAG+$2!_~Ls&NxbvhJP -I81*fSA~_>pu+F?*kYcl75<|twq}Fh14lR>eLQN4t-{xEIM*d%?Ic2huHQVlaB^SQ)#_?{+*ZmyHUVg -`mSh}d6VazOw5`ps*hLfbNHQe6>N2$!8VUy-oJRsP_&bl1z}n&I`f5^36F`cDpppRCiu3ys*G=UEHbzd -F(12-8nmBuEnqIU{Bf=`G&tgAz$dpsX@KzKylWa4HQZS`v0__-^7#On~Y-0k -gg@A+p$ujbV9m+8xfSCeZm(>^JeQ(C3e%+s5GYeDw#XNxUsZDm(u8(AOEjF1=5D4Z9&KMN;p1(&X-(i^SgCar{kKK72EeXQHnl@>@R?|1Kac2OS(gLN&PxkVrV$dq)OP&0u1GeZV=U -#yeWUG;?qh4~}_AgQd-8@N}aKx|{=?=RtVg;RA&(6Ph9ZSwQSP=sS) -cwmxw3hgHk>dn=`lT!w?Ea^YXjRE+|wsYuitR8*H-CSWnSZr`K0kBrKM%*-j_@JK7Q -`p{OJOxleu9PwL?p_v7}v-QNn4!Tt4XC9^0QIq3`PU|8@yFrR)X~OrXk(U60GqYx(i;_*2xnGq=ZK=C -~hQ7hkJO^+9uM|4}DXbo8-tWzeNkbOd!8he}=wTy)=#k^=!}e6tRM2xQyZHoI}C{4EPyg2PN=HW);I& -y#G(a~+9i-1p?y-E^v7R08mQ<1QY-O00;of09jiqgXtBu0RRBK0{{Rq0001RX>c!Jc4cm4Z*nhna%^ -mAVlyveZ*FvQX<{#5VQ_F|Zf9w3WnX1(c4=~NZZ2?n#gjp6+%OP@@BS5oF5SRlLoR_p4hwNY4&4N2+a -5|WiZpAB${Hb!o$kM{WOf^88+xs89*@3xdT%U*D0aPxFpwTCf)6wqjp-ewi@*dL85INf2pjLAcAaqu= -q3}$4d}QmM1mA%@Dvy*7Db_P4<@$Kdz{->7uN-(Cm@f(A;*|FEsw=F4{X@VOR0;N}K|KX6a(@=Cnr@> -i1YqW*xCO?#VlHoEMPS2JK1{aiO+>!y8vyxV=-G__d6@fsIpx@L;})o{NOw>Y6CpSQ6Ri=8>&rvD)Ae -$Hv}>-Z=1QY-O00;of09jiK?vsgR3;+NeD*yl}0001RX>c!J -c4cm4Z*nhna%^mAVlyveZ*FvQX<{#7aByXAXK8L_E^v9ZTI+M$xDo%(zXDYzW9dew^O)%ePjww%;^aJ -+T-;chn@)Q}LnI{OOcFeTl&yB!zrDM75g;kWZd%PGrohEwAHRJ7&}1@+mRl}KA+k1Sq^iY^XG{`GDj{ --G(2{OgO`0ujNGsvRdm%PJcu`g4vfS{Joyh3+%jXa8Sd(ta4XbxNW#muWrm5ul*;$&4Hj6iEQk;t7j8 -+v>^UgAn%Clsl0~;#HjNSec1BhB-N3y(YYk~nb(~41!6Hx*-T;*Mq!b -Nk9wFHZ6d>tVAjMkLWG~v9fB-p9m7Fo@#Ba7WV?@1G#mO>f9EHMXs5ZR*d1(6j?`FgL=wH7%rL*2os%F?&DAZ)y -+MSgp0bCrt9a+mCA&B0VmcHxPi_ZOGRmz&Gv=F^*>E|)jSkIUuthpQjRIXP+SmYqb=>zG`vk-x&9LL0 -g5z(wBW?12O0=+P!zn;B{2Yf%=gY> -3l4jwVoHB9)7iWZgDk(j>77m#(D9+lD3jdlZ>*F6>h-HXDpI^YxzAC47~JV;r-Gl$o$JBvWN+wCH7Bi -+b+9{-WpVfsgdzGi&O-M{`|^Mx-M5hHh>?c;A-1U|d`9FTY-oma&Z2uu=|gE@i`LM# -S1A29w(fD_YQ#WvWO60Oz{&Nl0X?-(W80GRHQ^&5}URmqJqnlL4d13}x`-haA#kXzQ@Q#nPzAzOB(i4 -nIRaUZGkL-aOAqO*s~$a#$||FWy26)@9Z2{GzGKJl#>AV-AWvBAZj2h82~vhfaVn2ecZ=uXD7`x&^Ux -VerzpY%86t1yO%FW!8UZYZV~f|aM)rAC9)&2fJ!Wf?zPv8L%WWku*6|G*7_Yt76GMF_8`)T>a$?fD~0hSU@ -+_Y16#{td}CAwk -MYA<&DM}2pfx?~Z@aclk;jvH!1qvs@2EV(g$X%!cvCR5@R|kai^qTj#b<0doY;pKOzICs;f#*ZK0!fH -H93BJ`R?M&hov7Y01;Lj;XS;<)V3}a!+`q1(-bOg((EgypFkBevW*V-&3}CvOg=uT7m6PSw$IDM&mL1`D-`TvH+}+ -y@(O(n%rrel7Lt;HVHGk($1P_j`N3tKiMMx~EMuL7eyCB6Zc7uOdkQc$521;TcrYKtwI{3xv|6zU#EN -(v)-RFX43-rM7t#;a?@4LQ_%wNRwKmRo(qe^Yqz?2w9*+KzlNY*#cVm!{!9(~dwvkXLnohVy$K69H2g -Nv#8I~&d(EiVW2ML54pFs9>metBzTx`c6-u}swwthj8f?D)G18`vJY`pmCNB^u=DLu{ecO&TFcPU{Vu -U`>K;@s!QZ2kJkq?%}7aeYUA^E6x2H1qEd?&cL)K(lUSoS7BOu`Jz}US)t+SR|jW{;qYw4qQORPdd!$ -_#eIokamI21m@ror~ocopLkm{2Lo1^YI+?H&+h`+&ATj#MMEyVLSe&3&-P1+pYUiJDOn9oy7mzh?!uzKeQVKonFt$b3o8@TPO)Tq&>D;iI?|vhV$~IDP@gF0@ -6?>J34nY)zz|LIqXUBjQj{$7+OsW{XY*%agqzYne7+3K%`jL(%zJ57X%P!gV&)Ko{=|Omfoq|wOh#Y? -+5MSisEW7IyDiclj_Nlta5BGF@#Ya9?N^;g|?Bfr*5H1SdiB*UJvQv2nmKJjn2t6kWUJt2SIm%}?409cOfj!5CR$zFvo1CiO>MqY_*!Bkc4oz+)lsRKn0;p6 -l+fRdO!v%7&i8HopH`B>yBL{Bpo=bB2JthWc^AkT7c<@Q%$7=xkA$vazX$$WLH -LAViH$`tSt2@s`t$;+C$nCnG?8prPHnu7rImSq>0ZlnEI~Xm4}=&CfR6dep?fwafR3YM(eo(49s=tqK -Gkk?9H6}&&ye7&0DS~FzY5I5pqvHU)&uEE9T_~LUvOEnBkVN?{x(@(d{?!|x(kEptY565m(5yzF8==F -5nQ;!>VQ2_N#x2Kufu1^&{oB5Gf_2fPWm`ZNu-Hf(!`9);X2#;lXTi;&W0z)@R;_!PwM@v!4s9IX`45 -3nDz0F0Do;_Pt@jiz1iT{#8AWuUJ%%e615@(hI^Un+f)DjFrgl~C$XO+o9+z+eVbFaC^Qfaw%pJe9XO -$?*$xjGblj?wSfYYiS)@A#u}T~`tNQwlM(6qw=UDFHNpSeG2N`=-t&uW3u^wE8c@A~ds#iPMBkEfW{4 -GGjp+Ow1(KS2PUJL(9WEON0w)3+fP#b%DGLAP<{`axn-&cS!89%WyF7*!u9B4*Qvr@4s@fOy|sm}1AD --O>gu>Y@UO`!qB8a)Dv#{puP2i+x8^o(5O915^*a;5%F#_(4g2-|{!f8Zaf+Eva|^@~4_H%)7r!nP#x -f4T^M9Wsf1ifHVYj$fbVelQ7xxDm}Cv1X4h8nl`RXNF|MiAk|5 -PH&O-w0I(4N04D$d0B~t=FJE?LZe(wAFLiQkY-wUMFJo_RbaH88FK%>fZE$aIY%Xwl%~#)V+cpq>_g` -^S5!CKf#YtCSFklQp>LtO_v}j!PAqfOpI$LZcQ6s6i+0g%eM^Zm6Ck6H}EH9zRB7?h-sd0pb#SUAriTkd{t_UVf6?o`F(^jN~I;sN23d+N`x0jQy&!K3=*x2+2LU%VhlDyY*Gv&k -sTI?rA$-4dUMFl26Y&tW<33k$Md7<$Rh-6DKZGdb*W2i@qAjU!>MS&=1c??^grm#YgXwHwpbMAN!l{0ua4yLr6B2( -x$v*alkhC=eH1*o#HSh9z_XEA}}ZwN)T(%3E`^lT3E!CV_9a&%xM+DjEO{pz&H&&-@1EOu|i8<9J~H#G8$3rVcv#(9ELf|D -1&He6s1flFv;By?s!kYCo>1ybKoyB0%wpYiUADjL*(#4MPUBZomd-+sBvlwsPSnBDoV&F!SyVo)ohGw -%6y)49fo6t>Dsqs`A}?qdnoz#;T4k`MgLyiZ_TKJ+l6fYDsvNi22`OjCGqD^mmnOqda%sz|vWd$+fq4@Wy^5XX$Tn0L0_Oz -#3JH@d-&ZRfPg^60K_L2S*5vQBE`xD8EO+1%*F{RMnPX~=fu+8j+N88=fl-q)HaFv4xI -+oJ3iP%Z*Eh+FwNPzS0KgXhlPW044@_e|Ht)vQcL%`ib`+8hPrZ?8>3ZwB5Uh{fxZls};Ny)h?}Lbt_ -~_X#-`lK*9u!s{5RqGPq)ZeMpq1- -F)`JDk}LqR;lf$fz+L|uXUubB?77H1!p7bnX___$$szd283%pM*rD_Z_bD&>;O?oT@Rr --y+v&GW$`+Jxi5nw+1z?bI0z?UWZd?inPp2|~;xJ(i}c@A~M$bq+vjrb`RgAvPP`oEy6i`n-?+o -gI+pcf?FHd*)3L1b1F@R6IVKPA9#nf3j}`TkjIrC^3(0^Vdzm6N92^02(5e+cyGLeUu7i8$rjLb5o%Z -tFM%WFy11)zNlN1*L$jjj%zC#J&uEQK^&VP^;6S!A(e`)OivpgGwWqJUxT`ty`^HQtHtXZJawG9c{^*xp4<$Ls81SDeSea1qJQ -x%B3(8jFgVQnoa48`HFC-l+0V3DRn6lLNB%zECOyYPG1+g4nCd~vnBIDJyE29TQ`?RQ(dUw7u+JM97VV^yso|1Wj%zY)EG= ->A*%o7~Ju0y~*YRn@|SpXP8poes*kV?j0i%k*eizq>W(RhlL7tE;n${TXbfpa)Fz_$3voU%B4o*)sON -*qy3&>@Hi=Tovp-qC1t@m)JcVwr>X@wDfG}D#0!?iS&azw<*y+IVpvVl()rVf0Z;}8Pe+}Zohj^sFjE -x=QjOzYhq`XPL>?i{7rW%=JXp+&zG|7SQ4Kob*RkmI{N}@mpjj43T*vm+bg4;2i^BUxptp;wqQGtyj` -w-^wVU(UWYTwmF1|ZPQPZtZ|?68uG~?Ye*jQR0|XQR000O8%K%whPFm|b(G&mxxk3N{C;$KeaA|NaUv -_0~WN&gWb#iQMX<{=kV{dMBa%o~OaCvWVWo~nGY%XwlC2ZYoqpX6$5kYEHY_T8}MnR -GZj7+nMZaDn&seBymg;T!OT$+1mT<*WCa}kfLN~_CDN;r=}v2jYfB)yYX}b>h*fw-TPa%iHjAxy_|$c -s?0K(7i=MOcD*fDGG%W9b|I5QZdkDrO0YOpMPAN}Sf+}_3$~Rd+sHDByd=ICtdQ(M#RMZAd -TybwT~Iy^6BAqL2jMPk9rq+m%r%XuCHVI;j*sVrW>2bTkzuf+U5PM5!@NmC&roDPV16vZeIPe7FXo`m -2@9Kj&3Ap(rcl}HwW=iRMhMLZNqu^=6RIm`I`o-avWRq!Mc(a`f=v1zuw%0zm5`+gcSAD9&|Jz4JL`` -|b@?jy}S-)4ne<~&<%k!&5O@w!~Ili+Re-GD7}xn|i`0w1Lk8WBe+cPZZl$A-@3)%U!2awEX9gL*tz4 -nBbVU{RNNz&5LRzH(pVECZ$_zDWez2(N^t1m!^YT)-N{xtJHpmaRmNuNe$vWyV(g0Zo#NgctFH0D^_s -)E3Ic-%A)7fjNJM6`AwY8Vzl_*5+wicP`g!o}x*E0KP=xXqL$N{cy-;T&+Bh!Qa1uj^rmtV68Z-c-ES -`>XLG>p5O-Hl@H`79ZoX<@sh+FF3NO1oc?E-C*W{zgSkwLJe~y=?;uWp0b7J|TFB5)P4$YP_gEK3(Y= -{~KoR7PGqD8!UUQs^y}P@+CgXsYsDP2Hp>bwlX_;aSG0vt_6&GS&szR>gU&VCltz}du5ET%7ykI=tG6 -}}ZSrM-VtLK8ZS(c!7;PQH6#Y%+}Pp4Mm>A=Jz%-vyZZD4KgBHs?_1OCNpN&sN?AbbEXka@7zRE(>7O -)3rtE$R1(PCU*2)gOn&E&%vmb!p=nuh&=mMK(eJY=h)$kg!lYmo+<_&&}=6!Spl%h&kr^{vPu6@- -_3Ey76}8CRgu^BkQ;^L&a8%^!j+^D>W%Z73eIM8-v*7*)$ybJ=V*PMiQ6Gh-c#HLd$Jh$T^Cpw<*iND -k!;ylVaW{+H-AMR5;yBnq`KT$Dwbi!ihv$!AI?Wg$ZS-Sf5mbf}B*a#pCtW+t6mPbbtF -ypI#X@GkD=KCLxdjsvwAlXvp+tIm^6TJ76R#p=z%z6pHq-6AF-~ng~L&)opLq_a#uVXjr4OmY+7C8Wi -5KUY?|KfQsaN~fyzu{X&s8R*qyVKhcpN`mFUJ9i7z%%VWR(pV+*A^%u*vUZ38CW;tIb`cX`f(^7pfuxR}CQy9EMo2S^q<}XXBf@@bc<%+;obc#y69TpXtS*NqcpDH@UhzecwrcJs#h*=x@gV@ -o{o9KHp1E23 -ri>_41e}1F1eAV%A*fCpoCg%)r<=*$IQ;Sb)!AQeCx08m%qQQ!{oy;$lX<)hF|xxti%_7OuMPq5IfFC -}naR?iSV8?5RuT3acxiT>U@40C@KNM`7^a8{LlE18%9?dzaP3dM4>G+M+apL(P%YX7Mm1ZN!j1^qLIUsI-RMHB*->95U5m`q~^jKDgI#2Pj(xx`5K?2^(q>ZNx)fV@%o87 -~v-jLu>~XD+sa58(u2ge>kF@p1FLmVS%MGRC~ZNAU=1iTy7MbI$&{%jVq|+wp9M5vZ$%;?69>9K#mdi -mcaiZQ#6`8qtQulGW|V`}41LN*H -t=L4ESNZUswhaYNOpYs?nnqvUJdZp2!D_sQ$*MPuu_s7)C=*qgy2GwG~43?|7+~r$VX22w(>RGY}YI| -WjBNZEw2Q|G(5HDEORAa}%$#=9R@)>|ls`(pr4JK_%JB!nZ5H2jPU?J6Unjab~oLA^N58qzMuUy&^lQSCOO2d07HEDgSdMIkBvg(avg@of7)X?`+kxAk=_MyaCRqS`Ng{Mw2`$ -R}X3LcuNYcWyWqTZ&qPN0GpHQN7%!w`Y?Hhy`VBz&?qplT6f*NRSh;u56;Zk%CPNmg>d#l|T$iY&vqHXVMn5OlcXPBU%gmYAjwA>pk-k^p{$x)%oT!LRE)_PVy%||bfT49q7xvBd&%dlN${e -(Uc+<Kacc93?oU_3y4_Y9 -29%Jb7V#tQAD_$Ks{qnS*$aAKnuI3rplksI5`-jTHD8bI@6zHQE#HX#`J9N<60_l#PIBWBn#!2W?{PJ -cV<=d@WsW8F@_HSk_Y;}64H3*`E`rJVoB4QjY;h164zE?<85;N@I7wJ*jOcakjrwZ&l!*((^K -bxP1L<%4=szLIQdvlfe-Z*#!@>DY`M_;-L&pVN*X3Ws-5u<Mr6X;m>du6dW`k -|+1Y{*C18h3pKi8tg)K_K7Hgn+Fd5=<*&|Cz#-r*0>rJ`ow97eB<9Jj_82Q#zl+_bQ>A5NKhdYiuXth -`C8?UFS~NCSRBEVI~UT$4;*&io6{AK-H?)%C}Mi_E6Y|T!uZzUNqr|Rn$R?)*dVX(yRv|6dSd6_aG|Q -S?JiqZ@s^mydU5Fd_4|NZ!b><^^ETWIc@(9whBK}^!D_URj@8 -#r|Q0$1c5e@bfyI<%#Vwg$}n4Sg?5Gvl>&v#Fo{>E4gK9S9$hq$6>AV0>_L@;Z-VX}zaLOD%RSN|>I$ -u1%Hl$nc-&{_egUxp{s!Jr%{8#jjz$pctk?oPP({!Vvd>@KiE8{rQ=XUlNK9074=RGL{`a6c+#&h1NJ -Y*Q)bDPa36~A+%#^S6(>|gP;VO4(w+M;?Dnu4&fchx(?G9H6VuFAD=*rk%ZLIW#EJ_gLzr -b*;7F#X;ZpsXc9AcPsMgt{;njwtRisoCD3qDqUZ)mt&M%>mlt78-qPpFMyCorBL7JFbt+-1$veNS=?N -UY{(W=-;T`pfTK8l75Ge>_(V*S0P3`B)mS=8;BOQi38^lPkKAKwjqY`i?1{y#Fd?MEK!hZjpcp_wM9W -m-J*1wUg?8yZuyw;f%$sNqlC~gJHNcuDYufspf+Yv8`o(9!o`7{0}cKRJA-tM;kCaM;A@KcK4MUr_-13tk54nqbi$Wi>;$?!`WRS`+a?O1VVu?Fjy%vOv?SdeFWKiYhMrFNkG|{TnJE -(SNY>^tpS>r*5u5 -xt9uZ;3ML0$_l)%Gi24dR|p6Z93hNynfbbv4UZrq@Jp`!0xMS9^`U^^`HU~M-R3e$U;=mpDKOh++m13 -K1l$xCA)ZYqiviYAMy7!|6dcg;KAI=Y`En_uP{@TEVYki!4RqC~9*mn_^`%K)P0%<7M0Qc2qPcfh72a9{{4TyZ8d>_5`wBudscQ3SsO)yjM>^v#vZ(2z1fwotQ+rE -1TKaJEC0KeFQ{WbYz68nKgIMq1~N(e%Qqg2Lru{$M=d70;L8Z?9h6Z}5{N;7&aSy9meCX|8=h%-U?)l -}Y2MX{*ivAdl@4w`fjd_33mDj0X|m6l?_8{|Ino?_&xjCkO*8Grqd$brfu7eP_MqcG<+V6)`oGoO5>{ -vy7X#I&}J*$mI>q(ouC;jU1y<&9N$rNN!Sy>{0V*>1KR3p8PaE4{zUnygR@8bot+*(joCb!Ty0K%iY3 -bcBUTmhC}YaY?=eD?}W>s5j&uqxtHMgyHhL6`sNYmG+w+sU`F+if{5B`-ZtS`;IaePC=YCW0j6KzCK`G!? -XfcbNpWswa5S0V)oez{lAIWJ>K{~Dr~EORJE;iM4E4h@q{zV(mSXO+ixM1k#Q2@JyydQ6M@53Uol1`) ->L69?%N_bK8zFj#J9Oj-46B%&*D2gY_FHm;r(xRvt3u~@6g~it2iyiQ{q37uIj0}l7_R^P`xEd_j(kv -;7LelU@idF%jCTL1H%S!YcT$qp&GrlU0_FCWboNxJ9OvxmUY~bXttu8wsgmiND@bDo8(%M^#!|_{Pn{ -aeG89%-CJ0~UV2paGz0ItV_Gl;>Lvx%G%D9LG%NOtowGs#>nLmFnbJ6VEuh6Gc2$hqi{=k)d8#3Sbd-w|onPQWnr^g4jvkx?WKRCPW5AeTE9;Zn`BZnQw0VGR3?4R|@n8S!e^<;Yqa -#={I)a_j}J%hU!=xV>kvCB!mZ8(YU82rYNX_Se-#Z~cMo^uKK$*ahf*p``b>^0QT*hyQc_*XGHI(d!> -T2^<&=wdq1$whZGJdc!o&K2kC#`37^vT~zS6gadsiRme|Z(r@SmFV9I<7>vNhLRp?n?CjF({yUMOs7o -q3G`5a;29)i%+17)hUPp~iTw!%Hp&W-FZf*8mif*lg6XuDvzmeUQ8PD5@V8v<#oVVmPS*e13u~W)$#$ -<}+LvEi&M+??TbDAeQ5!t_tzpvaR6A`W`bksWaBg&u{si1$DT-~SnPGD%`0Z4Nh51HcP~@GA28^1PJA -PlwT%Qaz41r87m9Kg16I|cJ3^6SB2*3Bj=jVDpYOp$-`sRe-PtL=(rtO*cX>0Y-^#0zvq@QW{mjSbqp -i^RaHt~9BehO)+0JW@k)~z{n=ZF^9d7Q1lMQtxte`{vnSkZ5w(z)=>`vLy$AV4Rjo^%Tm8X|0{>QB{O -0rw1hS6odU;_;rH$Pt##QjI2b(&@O{oMU<#VfP)sprX|mgq-xHsgxe#_U&6raFhJ>hrKEgE1d8UD|$a -NANI1^!F;uOx|P?WS&f_Ej^v_l=U%y>uw~~-Z6wX -&XPVMr;18s?Wv2{EFf3dWYdDLG1T4_KjE}qj|Y4i0h)5cO;j{A*jAgB{Fc8m@hgE%|F2LCGlqw3>ccW -G*Z0YpQsiFvt8iHH>54j3Kg;Q7v0yo^w=p4J4XjPjNEtlI(h38r)}H9*j_Mhn%_+Jwr0{2Z+ -6I2654`Xvyfx{Ub{cQ0SkJBeddgZp1RLhO!fz;CLtlLBeAB;{#2;(wDx`iD$*}6ugn)7l3MSO$FxYq0 -v@?g@leKr!TlWr2+JOTM>h>!m=p(LX$>aE1Yo?t%{Tz_EM`|j^8c%jSvkBRoqxWaWRIh9Swg)+JfUYN -6931v`h*=QWSzC4BS9O_?o4eFcSIGzZmPY%1y-RdGTEVYxBlW-s@Lt5eE+wh*%*N8w9i?Wro^kv!lCe -+6F#!I}K*Rf;Vz+tv45qFe*j)H;LLN<_FHMU)4^yfCjX;Byy$Tef*BP6cX*4aivyZ+10#Hi>1QY-O00 -;of09ji%{d6_GB>(`9h5!IA0001RX>c!Jc4cm4Z*nhna%^mAVlyveZ*FvQX<{#PWn*=6Wpr|3ZgX&Na -&#_mdF6fkbK^GB;P3h?u$tN$c_MSi<0N0o7w@h{mSm-A -ZNxPb+hi$>wbrNxMW<@P0FfT`GUiA!Sg0?OUhu*Rn6gPc -=vV<+mw7?6%H;nlqf9!B5frcUj_$dMHkt-O9H(2k;7Ph2g#?;Gf5k7VOOzB67HPpt1I=OjQCj9et~gE -Zj)4YiXZp}n&EM-FZCB|=(zJ1TZbmoayKhfnf#O)m2bmNlV6_uOa-m3`my)?T2MQKqyNbe9X9KGSzXN -%I_?avaduku1Ns&c!;GsnsOHaHn5p7zBe1#hWaqMHyMAAe$dN7IH3ezLjLwHLD3S$CX#jEh&ryw+T0K -B@7U|rK?#~pf_dY(t?2BQU;L}Kz{Q#Wqb7dZ?IbLOo=JQ5{;mm!Q=oSk(#0)EsxSUk@MDlO8?Q(Mzhg -L*mfPoeoo1i&EnBZA~S^LQv9h?)8dWbdfc`50_AJKPjNj6WzI2-5|vTp;E_UIf`9%7K8e2XkC(*hxr2 -YP3vCnuYj+>-s)T7Oz2iVE4|6%$g%t(V7;09tZ!R1wjk|?I>2f;EBs -L<4vB)AWxINUjGHHbQv#@MS(ndw173xJR7DD^w&HEnK0jJ5eHi&jzXEh;@}Fv3ytsV92UXz2=MVdPQ! -b>D#LWM%Rm?vq8A>CcW+O=JBI&#U(Z_nh>ZQ$?Hqph5?TD2``rKs0$Z#7j|>9C-NBL+a?uydOp;s*S3 -!0M>;Y&8$xdv5&Pe8Y5kv`6Bp_HJ;J~yeiLflTLB>-E@;nWr0APZM3(GZZu7I}X683j4dbnEMsp`=^> -JvB^Y$1bK?->2M;&+gSz-=MI=|iJ1E@`{y*KxFtcoc>PAc#O(%Jm7~KC*ta#D67CaZ}FYC|~sjuoNU$7hqNM$8qDcWDT;j{EO&pAz5 -86a^(t6PRfc;p}q70VvRdIdBxsi+KU$B%rNiJxafNwbiS91wsyakqIbqtMC+!sbz>pj;b?)^}xM~*!k>fp&&XXD}LtHG_f`FwjbnT`OB3jlpRzP`MLkw%|J*E0`BhG$~*H~2$LKMt<0a5#cD_zW|- -MO?*ja`W@;_`}DU_&B+`7{SB$BfxO*{%XX-!32j_gYl=nxEOpId>B!$2|(g`VpGOceEB%S$2jf){u|E -5lWUyPaB@Ang+Kc+WjhccnT= -iWhzn=YE;oC+Rlv!@v)~uqsr85IY4~7VLZ+(BFW8sNMhxVJX8U%EIK18>x438Z7)UjzE|*1ke&uqFxp -f5d+og-^-nP0lK`%c4tDo5;xk~&o?rRmQi@7jhc7VpXSZ$tf%Osu1HB5cI4MkZVEA`HzS%r=!_vY{m0?s{jO|Xx~LBH|!i%!P}O2)wb=i--civa`Tt_xhB1@N3`tDpbc>AVut>oEvgu< -pSxNVEIA^Kmfs@!`3cWhFpQL5};#cz*qc_)z_6Gyp$i$XiUeL0usEo7O0=9n>63zl4 -Fgte9OFb$=dg+gGiwDpg#9qoT(77RY1<5}GXTJAW%yfQY0IJj^EpBz8X9G35Q16-g-TWv@F5tLuj3ei -Em9D&;FSSC$I-on-)o~uyMOt*yt|8~i;9B)0s~7qEYOHmb^{S0ku>L9xwKIrwWzUU+aFg#i~-VJDRb~ -rU%{_mDid9x)ddzbnU!qmkkmnc5`jhuZo!;K0b6^KCa(+B<`E4Fu#m|4pSG~cQOq!sn&vq>eQZS{eaU -3MOQ9tg(X9PJ;Uh}$a{^>5v_G^-L9r`hAvko9QX4qX6Hs9P!eRmj2S1y1Q^o<=emSgu5=s|Cc1{-T=* -_FwH8_ufDfhJ=n+dTl!6tJtKyz8{*wrN#q#YR*WJFcCPnr) -5V#Hmbf6Rpk79o`tbfWn5Y@}6j8|D9}pZ+-zET!{AZb=gb9!`+SEmB-0S#Z#brKdnH+YZQaIih`1fx? -T*`}T%HY^I9el#*yjEM@6+A!(=#{uaTUIER0pw4VjYkd`D>N;ef5GpRZS3D{_u>p&&2(4vcF)Zp34yXSCy> -%_d%|#39XPQ%qq`F6IWg&DAR{`r>KL$m`qS#lLZ$DG(plO>Qg)z;DDOIlLyqK)nUOUzyg;r1BV9&4{K2A>a&%Tx84umbbK(5i!@AQAKaRClz{W|6;$N^CSs5xHyz^UvMA!`J8`xwlkn`fmSy?vFjK6=x>_V9a#&mArLVi6R9KQEU{JiVtsz(c(&Rz98S%F -I~?*M+p=Wwk?&PdKPHo#`WOnHQ{saTEeJ$2bqFdZ_{)bS}(sG;359;__8I)QDVSIJMfvjb?5{!n>1&O -@O1SGsZ^j{xA<5r-qL|jIbjxWR?L-j1$T_h^zMVAeS{VHgr#g>pC@HA`vv;1%Mtc%kj(GU+XSKw3Zkj -XuyVPs&~D5f)o%pX~&}?Z=_{{P~IA#O}Y|oR~Q90n*Z3q*zTZcs&Q1r#G^F*=+7`>vybYBQ7&TgZ4iN -ENHb#NoPd3RDaJ&A9=yZK -)|RS+zL0vNhNPYqx)=4dXPIiw~eQ3zN5+6Kmhj#9DjTwHouhf{q87x1;H3hI`35qc)D+|ZL4tCYv>3%Ta@~;0trNR2S2C=l=`Kh^}kC067Ux*U@*Ek#S1i -XPJwAY4oGByXF&h|DdG99kU2USOxL|C*xU}#O38SAL|pW&RkeX0{ZsWft=dyGH#Pjx!R249v1yEdt;% -+LEB(V%F|{5%TF|mI`%vpzyGc`o=^u5Jv%*9^%z?!Nml)JGr5fpDNdH?f|rjd4L*+HrD@pip|OgxR-K+j -E?7xTg;e1S3CzuApK+#;>mp+JxDr!rekgpZ{ox7^@2qcoo%W_T$wa9Tu6BkL4o-+MRl|v2=x0ov03E= -0}#yC#N+a{T&I}Ris{y3s}d3?9FScOBbd*6XiBDL@nq3*17ROF*G -#w?67P`hHLMFW89t+wC#V6vhEamM^?sQ(CSbQ`NBs1~=nl%n=WYXdcA?S}k(Madpc4d0ickJDDPCp;X -f1bqZW7sPK`}QVz9J%c|%ifdMGXp-2r-|BzW)+l31h<2pNF%QX#v7UqK?lw}6E -Ehxi6CmF4z0Zfgkl6yFvB$k27*w^?V8p@KyqJ~T}9O7sIQ<-U$a@57N{dm{q-K{Xvn_G1c?vPkI;e7C -rMhg=)Yzg)*yboE8)(F9Lhz!n3)b^LbZtcdG*&weQyj9lRU+dgX&6w*Tfen$>98W672ZNW58iZ&tyx! -8Zt#zuXP(XNcSX?oeoUVph)Ucb0T&jFyk&5x3CKO1kAh~rcV!Sl7g|K~{`m>6V2}H&IoWozia&e+{3) -4&`n{7Z+zzV6lfEs;Dk2hs1AJXy{-h3m+tyc($fGgEq1$5@F2Iv^0kFc#64V;tk9?=~<~>M@ -PLk_5;Hd%#qBt;R1^@N-!mLF;pwr!WAqj`WmHv%dl0i`ZqUuyT=F|IUE$ko1{7n%#{IA{rv=<8eH96* -B!K`Qb@PcR=KxTr#Ab(dQ`Q-WN`|BwB?ZZYLxf;?)#Pir*LDDpeDLCyuCH|Qdl=wMhV7)Y$WVdeQ$Vq -D{=G4o-FG;N7jZpaP88ps3bak0?jC=pf| -MSfOQ<<|vaW<)fVV`oGZ*&n4CuakpBckb@p*G-x8?;(g6CwE+cSq$%&(LY21{@s(c7NIMQ-TA0sw4L&9oUba5gVLqJ-dc(dD=UrZ=C7KQ6lBkD_OS? -it;&ynkW*ac1P*m-& -lwm2O@~3)7`%Ghy5^)T2B|I`vB{QM14c2S9z>-MIR*JzwMz7~wob)k19X+5)iIo;IvzqZE^zHiUHMs_ -F0iQGp7q#7*ZVF^@E>fTxmtHrDx_unr?OqEXe}kSnh&IKXqRH^5(M&t5@W{v7u+eRwhHxOcmXQi;=9G -g#lA+hdJ&i>=I78mAU?YPm3}WYxIW9SKgDF{^O~8V6k3c0emaIq$8NxQ;xCm7?VC@q6isi-<9jffii* -j_IdwU?`4<)puV8d9A`l6f7yA*9)Husv@5|(7?(V4En)pZ?~>pSzFF%@AYU6v+LHY{Wsa@PTM$$6GCY+N#7HG31Oqx$mnQ8`NRh0&T#;tCj_CFJQM=7_(svr!x#TJMh?$&`}qz -uUFftj}=20F(@-J9ORriiRB9gmgv$UkOM5bORyBi$(Tf*z>S~x>##zQ-y?5C$^}Icp$eW2)G(4^RzJ; -4_(?)a(d7CNvqP(4gC*QIQf9M?w+&O0k67lr4Az^7e)LS+27i=tr2Z|cpkX-0WjL0t|s*j+nOLDMx4Nf4Vt*z|RmRD-3=P&~f)PXJf? -&RD5`1Xg>@4x*I+YF@a0@VQ~;AVR?1z`oer+)iJoE#Xy1^5g1`c?Uboky9gu6=e)3I_B%1{p7_ctoHC -tf&MmzEXADyfHBzn5LG2?Fi+DR-%iYaxhH`tg!95GMvwuThb?-hD06fi3c(PL|3m{7R%!wo?y5T@z5N -*NHIO|8elY&6m5z&u*JBa^U&7b&ow^K?JZOsPsfC9IcG(g_w=+QfNcPrAfx@K3zcAkx{TH(hlfE+((k -B`l(x*hM;TQI^!~;@K1D9W8XEaIT>I4Vg0Jp)wX%QqE>VduuuF85%)I7{&K;womZE%ihrWs#B5wvcFe -r}ka6uZBwjXHr6#=3%Aq|bsCg~)_0H@t99ho%zzks6oXV{D6?3NB;stmaWUHoJ!=aQYKDAA{Lx!et0O-Z8jmg|5ocHwby(!77wl$oC4?G0BAt(TrqpQQA1$tq -!H_teii)xuk84j?cwO6A&NnxAfLg*9ti-c2H5$dAZD^e;9aGH5$4^gOBOdAsxT%sa|y_%JZYLCv`BouBOdQT%=m2haOFB+sh|608}C;N`E -GIsv{+86Gp3SVjhQ+M5agXcuR_6uJRQbED862DHm!Jj7!vQ{C61ba-I+C+EaQVz1NZELK}jP5a_GIPu -yR_NeHQt3O}Iw*`TG^HkV`DR$W9?Jkl^UJ%Qn5s*@&_L;{gc%&cm8*TbbEAN!uM>N#H`nQP3YuwWG0@ -Azczu#d0To3t%0f~Ur5H2$!k7?4##ca69fi8*0Q;D3CEYZvZkxpbx{Cmv_(fp2@hy$|6!{IH@Wug08< -|!Fw_aae_^N102g!z0_jiu{LXR%(vaSORJM&($u?E^XjKZ-zBxOrJJur(_X?T3V;sb|m=7P8DMwA<^Y -SC1OFv~+{zO8)n#P*tdjI&;sC-{!9bz0zsQs_g&%B!SI=U{_Od$j;`C8%f|i)_Jd@F;_|+DZ%CdRG$8 -y0$q=955S@``|dn)!oIEcx;r&;1$I11TR-+!DC2mmruu-Sgdowke|vsCzrqVn6POJJjM9NbEd7jybhT -V?|2U8m`{i2-Wa|#si?5t0;zB-)#X5nny?fG;Y_wuJzJT%R$E4Xq&lB$6n(h)m+H3~#+?mgj|Jm5`ca -t&k4nWJb-(|UdWEvPA5$zWa8QF1rZ3bi&ydTSHJ&7h3?XFCw^ia$%0autc?my%;d-V1US^)*p(b8jhB -upWDb1GqJfB%PW1*F2inIt~zpAmSn1j43U_GP*`-Ql}T|3Z!3*o@XYQZ$!&=a=+S3Z#Fl}QD`o!v -cMz{cgsAPnXQunl1wem9gY-MFD(OfHo2IbX)!ZKaUI4Il(a^S1n^%l4Q9+S)FS=MV{XEa;-t+@s51B) -u?r1HVf~DgID4{gw?kD27XA&d{kbSS=J%Ua*i%c;ikJ>zH1RP{V}DYuE+jg`J-cfYkw>Pozaw91>}6{ -(ciq+)-q|;2qyFOYk_-H4@p&ItyL+8cQ$?Za66)M2a1|mvw87HH^*I%mSexT>cDF)_mZVrPSj(wbMw^ -cuK|`UotuTwg3FjW@Wb~e)Ns-sQFF<-Q~)RgH)Td?RT$P0SaMC4lwk -?G*t3R04hxt1K=4e;nztT9n0r0~$!GaH4VrjRoWy6-eM3Bnp*m15g*ZvMfP$;tR)%2(bG;sR;l+0Evj -_(TrW98D(@?S*rvF}aXCE4^rNo!j1&$x78Fv00ku9)=On^izM4A?swM%dJ?!jz??wDjo9uJqEnLD>ls -JMpF5XrY2`LxdncpAl~RB&_((NKh+`DW~e>g$Sk0vPOu%ywaTwFa_y^6m~0(S8gCso86D5nBwv3^NG{ -{x&iu5FVk08~fbK2EFJUCrIiCT~NF^pOlM{p+Uysop)Tw|WCmdsvj4NUsm44(Bi*5*36$`NDKnywoCV -+g9VHjVJz-r?Qe*RTPiy -SqrXc;c*NY&vP`RyXW4R>BsjLu{rKjanvA!)lw!0MEYyidrxuuj!`Z73Q}j)!olL!)jtJK?P7l^fpcJ -|01e5fAW{7&lx5En$KY#$6`lODc?!a(@(wPLBUh{hn3mWDrgFuO0LQcNCtssS!Jmb$_*h`a~V(Dx7$yZn -W@Oc)V*pSLmu6TD!jk^2s5{C%*^s0l?1%Yu*;x4wi2*eL2`Y1m+7|4|p?e9Pyb8Jj^5&U20M@Z?)A_@ -xQav+PtK1EH8;~5O~^&V7sgTdwey#Uex&J$)jIhsnFKIlxMe27gC?=$@Alrr;xx|Bm<2$pnZ<1oO4EY -hbH;h7hMivd*02u`tS<0!8vDhz6r9s9R7K8zs29&Vi8qsR(ET?wzQ6**B_$b&pH~=sGzw5qe~h{>O|@ -u5wN)cVW~bU`0C{3yHj!{(c3zz)p!lXjKL|YHoiyW(w4h6XAZn!U9eGq>E>M2-?mX#akSr|+I}a;SOj -aZ!t_NP3_xF#;Zjn4eXq|h8$69wQWAk8LM;RRp$*u;1z2NPN224pN{!<}%ea%UrpDQ6}@Fo%&KE$}EL -x3w@A?TN`C>X@)h5EG_5wk<(F*Cjt@4RwJY(lX8tA6uM8L`fB0yN>9QKt?`F+*?rWZlu~_+Q$?0Ya4+ ->KJdBE;UfR_{%vJ^KKITAfR}yms-4k_Hh{oTcahS~#6KW~cUoN|u3)18>y)K_Ol=lg1A4iE)-%rAiM! -H$RwP-R_@2v+tx5tghkccIMxX{`=?g?dYO5f0U4;y1S~`_2hb_Ml*oBvvtR9&e)NoxNULd% -F1lwMR=&IqxCqL46QEQP_TR~5N+Z6X{tW@if8%lan!Ah@Rlt*Z)K}uZfF~rp*nI;R(nI`{N-#JDDLeW -AGfM+OQJt1&r{(Zry*yB7?Aj`Ooe>q7nO-hC2oFTDX>f+Tx1g*e~CEYumpahRc|;QWTXvFx91qQno&K -jT|T1kIZ;bUWLI^VKw_>=!ps>A^m8N-P;p~@fwotM#c)pxe5#BA -KB=Gx$nrQ~NLxLFHJafb?vX<~mtpms15{VaoW~ER7pv+IjJ`Cp0w+HaG4!|P;yQq=x{z3BxHc^a` -hjHiC(YanGLSvS-zNkHl6luU_NUu3cx1A~tM2k^lsjqT++zP_hgEetw`@skuK5*zS`AAz3SI#b*r}PE -J6f!CZ^zwUjS}xJy(-)L!sQfd&P`GH^ty96 -#7o}j<))>##&_t7q>dF)XA6d+^@2V6P8`obE(ZQ2Gg+<)Ea=LE%4zXMd63y(L26tD-FoZ1augLgt(8; -@R>;ysQhKQ!$@Jk+j^}v$CuL<*!PH%HWz9Uxr)ztpAUPyE^Rp#~HGGHqMR8?c)XW03Sx_-4t{pwf%Q$ -*B}Foh%x05EI3u-?ytyF0n?WxCAWCU(UaZvuMTFc`M#<~ly@=JXJ%q9w~f>JU6-Xe!$)lDToufB_FBD -V6e4s!6_z)_p!2mi%>b0-~=C4fwiDohfL2enp`N7VgdHMtt|}Y30A8{G@NP*9{eheDO*}Tle)I+Pcl! -h}%r`dreJ>0|X -QR000O8%K%wh%TmZcSqK0Cxf=igBme*aaA|NaUv_0~WN&gWb#iQMX<{=kV{dMBa%o~Ob8lm7b1rastr -=@?+eY#`f5n19P|1+!w&xTFA}U(sfd+?5jg!+{Q3Ujg99cvtF3Vk75sLozo7o4seEFO<{9uv0^Pbsxt -e>#;#hRre%gei%HEO^9C#?iw7zQ_oT(YCcngS4_R6H*wt7>td%UqPaV6v$yQ7g6=HQWCD;$K_Z{gi`b -`&t~?DOZQQ44S$q^4&io=F_vN(%@aK?C+0k!Al086qLk+iM`!et{5+~Kz}?#Qi4IrcStBl8vCY9l@=* -i=M@9w3pGP&6)y^Qecp=nqhz96Lb+sMB_Szs0od)WB)HpK;$QP9Km&l)D?e5R906@Wd^+g~Z{HxrnwN -J#;#zIl9=K|1kP&;M7%!v<`beTE#0iaV894Ab2-ck~Zg0t;=LCsvzS7cN@%ZvGilW=wWeYe7Hv)TRG_ -Bz%6tI_Xw^L}(4s&>gNYUeSXg=0E33bVj7zOAtR&gm$qRym_64hE%lL&$+VkOEG#{pwxDU6UBQgK9EX -OJ-^SR-54Xz9>DNXV{ebKr_h4?r9-+>p*$Gz8H~i2#ZMMxupL1g^9mXl+q6+MmcZ6D63iB^iyCh!!ey -Z-WVVMjlR!90Or3^9+OM041G`5- -2yVZygGJHQJla)f7#728#{e;#by7su4JK}NYxunZsbcM5%r3CrVE} -!MV^*EjL6iZRmUUqicHVq$D1ZNh-d!JrRN+HPJ)2XCOoN)(`0KO`rnF>JS~_Bvj)-R4;8v;flw*LjZq -uYbQ(Ya-&mEKR@F-%yrF5=uyvdHGJqFeK(X2vmL6N67(bMge?Ep;GLPUVBy7qV55mc -wM4Vr4duJiXwx9_L2_lB~)}O%k=!-AQ*Uu9ns4|FVWMz(G!}ua1{f?)g?*_l^oBaHlt0@+(neAOkQVU -pIYWb!D16LSwWib(IZ8H>EGu2h_Tn0D2}2gvY331n(7KY3`24Ur7DTeaHXooqmJTaB9&1tE}^{qCi?c -f=+h;Y!8+eX2X!o-XxpxB$?GUCSMBS_VrFZp|JFR05E#!ADNxSPKpvtX-0EfRuqil|A3-##gS;^dmS_ -~Y14Aj!p!{@mbN!-SUFq*_w=hPmgeazwuW@HYfEI(<2(oZiOs&_7y@gYU@97#v#QVrgen$z|P$2sMHc -Ji)Q+N@f=Nw4IYYBABCQgGaX3qldu?>5RWgrNGw7{ci@=GV%qObqb;L+)%-yd03U>aU4%-q^4%(K2OI -`&(#`<%P%Ws*YFQqnIO?D4GUWv-HBA)(l>*affeB>r4{Jkguwt>wo3DABQ)2y5;QGhk@JY7zD~!)1rs -R8TJlTCOz3Wz-5VWbMgV!YLuGw2+=m&L;G{!V|qKa(O_N5#F~7*-0{ucJ3!VgGm0y~(XPAAJ*Z=Sjj;cc$tmQ+l+?BdQ -R_VG)J1mwMb|ijF#0u$hPyBe+bk#i3p!{4hs=%ri~E?CVlBi%qgUgiZVN*selOQ*gTvaJOZ3MgqG#k6 -&k{oW`Y{DLk8)4U+vBLtR|B_~~se1267%uIeTCe%RJ&YKf*-VusQZ+<92t4`V4lV-4BZW(-cCCB9fQ#M -pXji*uIW}G);UUGx?ZE%hn}x4AhJ|j-iXj#I`Ga51@%kV)4X_(U+% -&~9z;`Q3$_%Xj6yeP -vlLaS${?O^qN@vEBP+CrDJC_4&rkkaD_6AZmk-)l7c8pq?M0`8`2HiS$xY*9e2jL}JfP8k>tHUnIJV@ -T8(KgYC|y-!m6X=H3g*9f6m!>vyKbXq2EiMUOHrpIJE=74gV#O$Q{#o!kJxCLF4G{T_GR*e}LQ7Q0(~ -QBbEjU^*teouIsZJn7+t(xa~{hn)u&-fS@H&259@ZYkXG>Ej!x)tJk080}(6x;E9dy?Kd_o^Rf1Xi2n -)Ad}rWF;`R6xpu8B&6{C7XZuwjQ2LPDDbPPB#)jT}&{1sfAiAZty`tI2gLGX}f4OXq38^SI-+bMz9`N -v>g$J+~7o8b3l52Im%BQEb49(MyZHrdMWv?4|a#AZme0tZ5gGHhl(M=7lmqO{iema+!VU*)*fFr%O2; -;C{3y^sMIONdol5IBZ+i_)Z(ycm?Hin01E!9K@67)xsHlO0`iSYnW)H_--(#4H -ITiY5x^o*8vF_c28I1pb^n@A=d2wI=e&goPbFnk1Zuh3Z^x)Ql(boX1ozCL!gbWC|`73Tu&HzGBS+RC -Nx6@q%jC0Iy!z+&u&d1|H+`9wui^1p?q$g6F1XnKet>%**JtKF{6^W4`GoZo$08mQ<1QY-O00;of09j -iA00002000000000u0001RX>c!Jc4cm4Z*nhna%^mAVlyveZ*FvQX<{#5b7f<7a%FUKVQzD9Z*p`mUt -ei%X>?y-E^v7R08mQ<1QY-O00;of09jiu&n->v5C8z+L;wId0001RX>c!Jc4cm4Z*nhna%^mAVlyveZ -*FvQX<{#5b7f<7a%FUKVQzD9Z*p`mVrgzMn8E^v9p8*6XkNb)4-5Cv;+_) -lO3&A$e1ShVr-uudF-#hs&2R4?FTrsySmY626xxHtE#JC&1SQ)+}R=7dz;vyBo1jvtmw^$|00j9ZFfU -V^0M=wZymeVj(A(Lg&&daxgjy3+kzjF=|&;h2Hw75P%94TGBDlH_kyTPmODxwJ<}mO8cym*>sOtrLF -Am9ut|KJwy#P(%zO(m0CLn9a{M~jRfi{{ZFl~o5$6=R@ya0qZBN9+E^xO`IVFH9ZFLpL$O_{P_%79a4 -4bSXr?L|AXej{<~hYlg{hww|ILEw+(`EP;!xQj^3YLnAm|KcP3bD1eg{`mqw|NXL -t-(8nR|KfOa8tmw~b -zK+)-jP3_Z&>fhO5_R=fvCFmcSu)^@;KTR1pPBf)64JB%DK+feX57%J=+-E7#s7%dw{fz5oU13;D&Z% -`J|?bv&pbK%s01TzrxABdF>ledG>V{h?)%2f2@wPT074gn7W^6NN)#t<9X0C3SjrxO_Hkb-f6iVZp>E -+(X8?#sdj4ICvVwsGJBYwisj4@}57|A$%;wjlYf=RnM%n&3Mdo5fAy0T`?18XSn} -#eKo`A`qL4f`j5`qn&hxI|xX0J#IjXk-=4zR0~iH7M2jc1OuDJYfj~EBFC*xjM-@UFj(khY>@e4_H8` -Whom_$;Jeu&598(S?0!iAVlkL5zmwUB45r`7&*SN^L-b$f3*9itY{AHl@8*-S4(;QqHn|^;r(ekp;7( -^tG8x~Emq2(qV@wlrW8FZ4cltuRh0npwcrspo=i@dSFQ-U*G+U4XnGY7rv35ThEXe$RF`pSah%p4x)A -4k)08aXyK3#T!Gqe%?8+;Jsb}*SBJ4SDC51K4cR-(=3-xuSrw@Y$6n+$bmywO3z!OcYH>_91+3;toP)}H62zuzJ%eq)we`QpoVqq9@-;dt(&_#N86(^V -wCVg)Xi0iQj4gY??6$lW-CG@r=5g*kSWOc@Oh)$MkD(|R%=0W`E<1td%kAziUqwqb8mV|2r}FrxcboB -ZvXG>2vYvEBT__%jSLX6SklctNWo7T+0bP>ir_R&qV-DrrJ?m8n4fGcQjwSVpXCi=lDTMvMu* -mr1aZ~84zza!?!Zs4!(sXhq-hO1BMy)$@N}bf|vwM3F7xJ|AC&)=8xd*ODgE!$&LL;j5uI55w&)%ZLz|p--l8wA}_ -2O;}pmyrXqbOSQ4Ezji6+qD1~FLsY7wHXCziekR;U6aYb>0jq2z5eQiXi(jGi6)mjU -$DCDylO6sETIBIOpJi2bA?PT|P)I{2`?P^&HH0Y&=>>yMqQ>(I9ZW>tzrp0?=2B`jKFg@z0{Er347hf -r%R6TQI)1!)H@;lqUEW)2&0Mo=ee(Sh;FLVRJq -{dtRrxFojmz20>uHsZDorWUd`r(lg{G8xE1u&izPfT0+7twv2~dMkapJwt!@CE#`?9Z)>AuP%c?!

    RqLHYZCxvb8RPlT-Z*2=BT?qV1ehyY0W9=EyW+3|zVwq)vDi$Ql!avgHCNEcZ1U>5Q1>KxGKJuw>*#-Zl9(&D?2>b)=>=w6aE -Foq}Vv55frllAx-5rCNK1_ZxDL@%YPo2ONd4N+WS3zBy@ -)mjF$y>yFQ?w+-+aVB1TPYt^YMYbwJ?af%ZIV+@Wm?TGL9N4W{7HWWn+bLrKHAWLCwu5 -b}i=T$tFcB3!D5Do&{1zn(+A!M7u4BK7bUoKs8n45c!NK?u4{UxQGR_sxi*&(RhYTLv>K)GNvP|+)+f -$n@}Nme5e`toYEkd`*ZBmz)AOQO8F*)YO<}$I62;#*Oc;nU!KUL4B}|#L6=Prnm@~ZA{!i&MJ8NVtkk -8Bc!J`4<$`3S-r)m7Zx3l=3T4)xqPI%3?K?k3!sPY2i+S!GCdSYQynPq~s%E(*cO*5S|M_Ad~C)2 -$P*C`+~%`5NtDXw^`znr=<04$6~NGm2ovxm{UumaKnL4az6Ca!G22r|#m`# -9t%7x)om<%sKY&a)xrz|DY%Qm_UD}72o3pKWI++=tb=GmLt}Qs+_px7R}Rrw>)B3<6Clm@q>mHWDc3G -nrW@%2a6`0nAk2&E&FAAMJ2{l$<-;W86*VMz_p+}km`jnU(yS -CHl4Uo8OkCQhe&M66-Fvgv$hmpm6}fTVmGQ1sDfu11$m9mjsa?MA4ClHke@JsT7n2-@7Of~cW2AhBuu -K7h00cr9Q^XuXD62xq>az2*jJfLa?62RI(I62r$^Uuad1>W~)i-&ORoPKq#-Vi`0Hl4)gWTl* -4?jltVec&5;>Q=2JoH$Z|~0h(~FaNP{@Zx#l2BZW!f`H~&u3a*cR+oNn{e=+l3sAif!vZ4&?gxA>GT6 -6vda<&eRppCyx>bgH68T^>|Av5~%t6qOSkwV`r{FyXH>RBm~tz*0kbdQ2h3R~49a?o(_mAQevI3T-Ni -bRJp}zfY*TbTl#t5A=6|8Ri_KYrCuTd|Rf_A|}hws=-4ocSIlAd-^xll{;z -IMhlAf9djoL$^Oi52T}4z7HQm+AVA*Ft&@*4;W!x=yx&m4p!~??7>nblF9j4nLejo^NisS*eHS;*^Y< -0xX}h)z2Qpa&4Xk4LuXz;YOqU&B7Nk2Umr65}Nhn=Ypel=&oT`?YlhtE?$i9*&vJ&jRNjV;F;g_73=>Jt?KcJVdVYB&-w(B%e_CG1}$CP>unQU{(~>RaiG~5iPnqDj5v*+R#A*%Oc5&h${HF -Cdu8V_bK$J2`#wGvgdAM9a3T7`_$c(`9+RMzL-8vuGU=5urRw8*8lm1$Pv`KIZ9wL8OzrF8Q)_$xs+7 -R%=IKW#*l7JC~~>mN#0VZ`&1fO{S~RmRc4VVMo;CPx?22-6pnK|dQ-19=WV*LF8Ue<3|4lI0tiH5J;S{ -{4ilU-J4{|^WU@{;#I_?FDHq7biC%TMV{cBeV?O^dR}XR2z9RQjf_n{r@3p7OQc{c?0ASxYpxdau2!U -a!~h^)lhbGV3o;L6EK%JilRf1KIv*<{C73rU9OSgBHCER%`^ELUX^+dIqhhymdJ-L~9>f&x@_t{HfTC -l&_KnrJ=roBpVNMC*Gk?9zPI3B>nCgq0iLC`}Gk%yZeC86NJLd40crHgR=H}h4wr1HQO<|CZ9il&Xm_ -28!VmN`^>b5A*QCdi<{btT6}Y0uHPB-zcF#(}ts~QRU|<3Kq6~ZlMY@v%yetEp;dqPN?w`12X%MK}S06t^#7SKb_-6D5X0xZH^GxbwC*_;30 -R+%C7T6aBfT=h2B76JB0s5B(fWaEwlcQhCg8hO!^cwDCWn+#&pJWp3L2Ar(*p;V!|C!{I(9eCLl+J0oy~egyA@|vMHb -@>Q-_iDMd~&tJFMMGd2N2A>(6bT*{?5e2_pV9)htvvy79@&Y6#3zolamd;AoqMclJ>z?UCY~_C=ZNJ! -elNv>0ey*VE8|Vp|j#FR!pW9p3PCdEe{!WrE9}m9se~3D|{)r8tBIA^x|^$Rl8dmw}Z2*b4+-zTJKcL -AuW!gob?X#>7QN+B)}JZZVvG&M;TS~(QXIu_q4Z~&nDmdXT1vrRU3S+;&u%2Dn+bgj9%~as2ZU}D?re -XkCJwyQgNQVW^CJ4SyY$1DE@%R!zAe^w|N*Md0I<1?0t1vI(S=FDf*Gc<(DPB_;xx(*hk0V6JkME= -r+PnDfT}08mQ<1QY-O00;of09jkEDyO_z5dZ*UHUI!Q0001RX>c!Jc4cm4Z*nhna%^mAVlyveZ*FvQX -<{#5b7f<7a%FUKVQzD9Z*p`mY;Sj8Y-M(3Y%Xwl)f;VZ+qn6=e+AFQq4LcY(w7~I0NIeH>0Pk4DdKd+ -&?^L5rej@XQAbK~y}cM)H7ZxuO0&g@H>$zOYkb~lbDIj-b}Nc1FWUv|E;cJ&DlvzxS3)mE=_whVX`^20wpvKUNvr`TAT -H!w<37~_M%EfOS<6b^sI{sZf-Q1g<;nmNMFl$-R;=}+E!7B7ljU05M!;O42`FM+UUCi**20jJlCcGKG2ues%2XW3Aout!p0$UlyU6I^+KhiF$k(yvlOINKR@$td9$sR^>0O^Orv$# -4e0L4DF0+_**!L<%^UyLsCB(8fZWm2=(R|_yoLLI_Ua4-+)PEx-YF|e(QHSf2f#*Pn^8AyDUDE{YlFN -88NetIObT#g|w<)Q?7GkD3)&abX#+10r?0f1?)*EJxQj0F75$B*#m#rZE6Zw1 -b~cy)1hdVNl#zg0>4(_bgQRrGfL_T|ou@f0Dxyu3br{dzPRjYbQ#6j{E^5TgvK&*s}k*@XUY3~*s`Od -F#h@b@ppovHxoe!^7bMv-iRQ7?(I`R}0of|yy5h5-}eZ|l}JOpt;vR|@n7nKa*$BIL3xl}V$GjcQs`4 -MdPK)H6bWIfOrhNw;UOPRml~veD+HTFSNrS%w6ZstPG4I1zUWvT-m^=26quB`}G99bixi%8l}MeuC%r -ysA+DbzYEj04lm2Vyc6kZShY98iBaFjYl-)f%t&_!>Jg%j$oa*Ra>^mJ6X2MlHQS-D`DkY32hdh*EID -+B+bX8PwWk8rrQl?I2bhN@qszOFb`-)TGwE;SHRVuhLlZ`pKwtJq2K76VDK`>nS(0RE&C3|>E=2zqXF -K&9~vMy{lZ2_Rgc);)b3-lP&JE^zX9DW~ATBoew08w+$DMZ#iBl}=o!LY&hLd -JZ5L(Vf;27+Y|p5|G;k|wJ=%)AEhtR3Mo4lKA2MYB5Nf$x+d?h^2q9X+mLsOJ@gA^83N`qgvK&Ooym% -Z|Z>6Z09?WYQS9O-kF`&qM5G`oK-RX&X5&)hR;m#XuX~LkhIuN)}e3L%?{L&StOAjR7mhR7Msh_agdMG2zcq{OYu4zBOt&bP>)j7z1w&lYafAdn%TctuYnH(FO+ -jB;HV3Lo719`4V`T+#|{S3N?K%!^nb;#gBF@ei -VtCPKIU<6M}re8F-_!wkj}DO~~;N?3|S!OijG!S>9h#V$}OnmlLJpJ=$!SXQLu1#Mq&NsVI40djAI|6 -9p#qm4;Xgar}?AAU9p=hHDW@v~tXK5QIZt92LU0%GsKEFyj|IKSA -F6cTM4|?ujRO4LsJ1~GwPy-C6U-rd;D+Ud5TLN+DqR&hn4Ta<6nX*1W}jcWoH=D?@15(K%CWOZ#r$TFECsHh6A+`A66(9G#QkG-b_@9-T_{rhaoa -2(Js(FF!!x|aKnR2KEUR=z`kc`JQ)m9b-7GI*$_ef^G@+FO*G)mp+alT#X0y%8=nciO4h|9kQ=w)p_y -h5#(SZR)&|x}HIXMwW$AJ#uP9q=wXiC-*?`ysk9?pR6B!D+qyx%CN8IC@phoC^pUAGOp{XNcOunDjEl -8cLHds^Hs&I|+wb+_+{XvgE8zz&^ud^wvnvT3bzee7>K=EMTpHL&4?lgQJ$B%JhhFf(Bk`rH)tgPbrI -ShZRw@Gs-C_)rN-A2EZ%<|_dcS;nNDh>yKjqdr -JF|M{6SK%h2Xr8h>mb@F&Jp};|7$dRa_kS?*a(SX+rm`i{^d6`r5%I_Q4#GW|%RtuvsSAnU;vdAge*; -?iBq86TrUbNquC4NhW7-Wavn;A|*7W7zunFn7?zIM}P6b*8hnhhwPJ*@UjqbAYG-tJ5XVt?-F79YwUz -kz}wPDB6KOTI72upyJ|Y<-`Wj7L0L93yeqE4sA@Ir6RCn72i_V4^uc>m+&*m)+F4(;_DB8)%6#z3c6%%|F5Up+ra@^Lz_Kg -d0^bm$^pGXTlrn<`DSOnw>1X+PQCPAXuAk!QaHZJsUz=7D1>%O{Mc+5V&fd50FF`aNnWE3D3Ml{-IS-T^}p7V!9QZX$}v_hrx|1}&MXJJZ3;5=Ss+9zb{rmVzo8>4W5cD<(4T -yeiy~vtuN@S^1-tj%rRU=yVFtYdHJSRRe1+HtPpT&ie{nNz+y!u5vpKcVdSRltbNVh~s`g+3|@*<3O# -Ige_R8=#){e>syPM;VOIu{Z^GSMLv*HDlVXUQgShMur}%;i0ekXBdD%L*)Qu!WU(Wd3ftXoE2l!=^--cKmq`6lm -jR2E#(c*jv}C`%dT<6vLvC$b)bKb*q-#Ev`UFgSNv7=TPIc*+NSGv0;L0^4gak3|heK~BU^qKtjeCl8AtDA!y;)j{jWxT -({XMATQO4;b8pu1PmgOU7nCo?t(88P*^h;;btUmQX;-MY6Zl=x5#Q%oKm_lca|juqKBieHu6HsePpe> -6j%ba$e6TF#jol9w^d0QV$TglTV;bl#h@d^!D+9i|+pz^(h&Sm?lLBDi4#Bt -}89+Xwp~W>|=#oOYP6*5ujo3Xpu;Z$DaZph`5T{*(#|J!sj!*-YC3u0S4 -!8XHXEO+cjOHpguQ7y5Q@*igSbkAeTT@gSv`>J&Vj2UbjlsVx^(r0j0px;N#9|FsP2HvP4VWLN3WYsU -L>sk%LL>Z9Ef8u8!UF>w8s3)a5 -AFwLbg{pl3xD?1kC?FTIX)vL4P0?5sT|)kQ+wR);p(VfI-OKd!BK{D)99pyZ6muI8VYz -7qV;;MfAKZU>Y}Qxw7jW$4?lihWB4y!>TwM(!Ddeoh!ay;!eM+c0GqG3hV~)gDUPg2Bp?py{Q -x->!0gISG2db@rZS?2kTeGy -?O%5gn97pO%N#mTXo9upCjcw&emIT#0TCrf|ziR{?OP+zc`qG9t_z_Gyu8f52`JZE(@Xf!pEY2g -Vg0i*urTJyKt|5EjS6-<)|c#3AxikFT=LWQLo8Cf*??9u%&rX0xsblsBi^34pdDZa7`J -2QSRb*hY7bH0L%^EaaTQUVeQHs57e*a*XPssQCGyfg6|Qb%pU`*t>4b!|X2u -_c919`hG>0TMDUuqv#A&%poLEta68O_=r~vslR|U(gTu1Umz?~xq9FL2R%g=t%w76r8Zd@T=Cr><9)4 -6?qCmctUjxN6c)3Jc$3|enl1qf<`WW8%8jwgy9I^mHaThTY5vhF>u!!5q;KKg$1U%@W~(7?w>eFe -JjjE;~(er(!2 -7w&5lF$bR9vJ>}OgrSy>{m@BU=bzVKQI4In&_H}s!8K7(0#h^%rtGhmkbMvf{|oLNRAK;&h0pctS@!P -g^7``Z@^$w2^Q+m#<=Yvk+rK!;hxm2(WPB`=qp5h}Bus-(aq#I>9Qi>2bobZ|K8^-cKl~&fe&U9o!tl -R+4MxV2HKonk3TuTb%Vj1@td+rC;~$s|%C8r=0JQ;t3YrMq&9Qrr%)|7N;fS--!4*JdvbFqu$`52?dz -tak?fc2KM&T7aQnrKwvxGHApO2oKutW5t -W=fz!w7iuhnYxohDOWmqKzN_Irz9|?-#$)AIJ=rPIq`S_1jLwpYSoPt9kk8KTt~p1QY-O00;of09jiA -00002000000000e0001RX>c!Jc4cm4Z*nhna%^mAVlyvrVPk7yXJvCQUtei%X>?y-E^v7R08mQ<1QY- -O00;of09jjHlo_lhApihrhX4R00001RX>c!Jc4cm4Z*nhna%^mAVlyvrVPk7yXJvCQb7^=kaCyZ&`*Y -hilfUb)z$w$Ilq$T$PJ4Hmx>Lt-8r|pD$xhQ=z6?!?kj0uJRX%L1-+cf3?E?V8r(`*I=gCAQu)A0+b{ -C5WfoI|6aB+Wc -<+zIr}5Q?^N&{oASQ=rSHFw%_u}yEck#FJ*}IWA{_SFNe0eF(CqQg`dT}y7hSc%d(aFbmBOvvDe07F&-=9y!p}07lT#b)Do*YiZ#mC9T`Q{42bO%MXVqCs-IzJNyWBm>`$p==|dM$@rHKSK`C@$-84ne0vPs9lk -v|=E9(AM<<8l(~)?0czXEDF=3qp9jqY&GLPcZhht2{G7sV3(bf3;3@dYVes(p1*Adipa%GS|jW3T!;& -3v)M6tY|oS$MfQLun>PGkV{?3n4G(93NDAo%|A^4O3R?~V^ofZ8P@b44|{)9dy7{f}7?g2sVP%t0yc6 -L=;)^d)G52bP$lT7XwsgXu+LkMwRI5PQ9O0t)ZC^E}HF>AE6Vq8@?1rf~?yROm#v4 -eyWrZgDQJjoapy-ZL+fUN~XP@>v}oxkw9q&qF-});>W2<`gukCI&pIIQy2VxJh_uI4gG5d96{+qoo;-> ->K*dK}4LlD_q0O*^ZfIoCr(Tz(I0Neq*kA{>JfET}!Ns$E!Dr**kS1vN}LFHufA)^fBBTFc48bn^Q)~ -VBVhO)+t)5G5`$Nzn-SYCP$Y4Dfq$uO18yiL7Sa#|OQB8;P(c0ReWog>V;g2y`Fc@&_AHt>!$;)3VI@ -+$N)XxpU&zAnj9>_7(nKfF7yKja$V-!t)_tleIC@2ykpLsGH72c)yqF;ImppbmqJPjrK`h^FM93>df2P6XVF9-VlV(q5*=w`#BNRp2DZ4gs;VR4LqzMO;gg%Xr{0Dac -Hl8t|f1k;R!Vzp=Tvt(Tc1}cN!2E(CAln*nx%EUQn=^Xl|<0OfbGRiZM&T+UQKfK^H(->g3?zv~NmRa^_;b8^)e56I+!lSy1j-97%8zBxuj~U5IW5=*31D=O}T5Cp -`44fe*$}?C;QLptB^tAi;5w3b`Tp)x?=Q7Xp?%l!M4y%7XfsIp@5QYEtyk*imz;@aEwEC;=2E1In2x) -YZgYrrWdhf^3>7q@OYK3-H4#=bm_ft>4LXmXvjmbq<2UOex(+Slp5rA<@A!s3HFJEW8_8paEGh^p%AP -F<%~LXv0Xth1WQ|lY^l}8pN9sN28*gkwP4j$whbZEf}xsrdXEKuJIUKLGT3sJhXF>--HkA6-VYKC{1sSvfn@>YOtNsXG4WnkLx6hc -P-l+~$d%kD;mHVI)&5Xt)kRD8QlbLf}{7ER`S=ol*HI@rCJl?Iei-blResxVV&jGasp0v*LXdEeneEI -_D+rFB}5DlSsVtzvqb%CMrg*my=*_+n8~xE+dn=Y=`uD?$TT8#7U=>331eh*5`Ct|_%_tg5lO>*N>eJ -GZu5YGaGkhBoH98fT_nm1gS=)H^DyBzF9t`w9aF6Yz*d!Sk02+;;$oPwZbXWWzTc+;H(r)KJ&c -rjvfD%8ke{QxFH|Q(_m~MGB+J&=-K7BmqUf0b5v&B=Y*DY-Yw(A7^Q(1Y~XefGJoE*&e5x*Je{Rg%YdV5$;r=$}qQ4d>^T7^dkTGr&SsTmsYdU)YV8lhk -frs6v#-*M#PaXI@Hk)*e(0$u!p08rVFtlBM99oxO@}Ea4Y5vh*QHg9v9-XL3O1YsJfzp*fR<|C@rvdc -Q+}N@Gm0Rd|=Cx+;UNh^1uXLF%r&*{yfe(E{4ZlVWCjgUD%~eBRv -509V9Ha@ZJ4tXF9L+*-_mYwWH^|D7?>nyra4E4*c>TRK{I?lYm0WnKiK1Itjbco+Ziq20r9@DiR_c_; -q-vQgn?Lg)!4;Lj7A%UybZOGJneE^X^<3ToMh^YfnyoIi&92=`Ah|TolbITAP<4=AihjyVOUzhiebwK@Nm_xtZV^fVZEMfqJ@LzctZ~ZtgwVkYMjYsYL8T`kOULSxnNHZPJ+UBty- -ps*NsoXJc?)OCATOZ)o#dLQ{Yd?m6*RjyQKNf@i`7%N-sf-|C^NIs1bi2giu?8*Yn}{GR+|!913L=Fz -Tt9R-va1Huu1n)F_uot6(KrRG5wroYjJDTV=&UTQEMRGIGb5nguW}u6IZlM29!3K-#Z0y;3t?C9d -pxH^I*0Xdc|GaEXuH=@GqO@OuM7Q3p%>t~Z_aM~zu6$qr)ZB`JI1Pp|py50x#|X*E}4wN9i`i`Cl2=O -#FRx^!J*Oa=gZ~V%@z<$__hq%(#r^qlKGJY^B-C;Wg<=7FS2X}!A*P{$ -UnSog#Y!g_N}eA1lad3+gg-ahW(%xR^liYY2->ieo@pi4@BD;p+6{!n_gss#Q}J^I-rb -XpT!Y=bR=?x%7h2T?PxK0SV0Na0~j6h|#(Kxo}oObMw~DXmg)0lQ57s+2-h38;w>w!??3tBXsS+_6H`Q>dX+er=+0^vk%6P)pcE;gW -_kWt^t8O>|}#!R(f!4C`^lPfZAa<}xRk|DuN`kVLK$rSbUK2$ROMTbZQ=@AQ7z&Zu!aqF+4e#3qF7mV -#LXyHbxGV6Go{VN8zvoq1Auj>_vo#~pSBiTChUCu6~MkxFL?_C3B<(@XWPl~j`}!F3WOy(qY2J*o3og -6krf1r+IB@sCEw)K0AImcZ4I#)8)_EC>{6cU;O1S$fqpip0(qj(nz}9!%>-oeBgTgmE%k#If@K+3gXg -9N8(F4-2xJxqLg}@lps5$ds|k)?Dk1YIFBQxZ12sJA`rLqUK(`?!zJiUrm@nFQ+|INm$b! -#>aJE*d$Ze}^*2vK@JrmW1EQYSpviGs8*wTS3WIz5E|L&)xmGKr^gR)jT`SH+ -l}+C=MN878Zl72t4!a`^nx733)Vw9;+;c6u_oI4U|WMX2byzTnnS-xv2mGkaO?ix+Che>ZRC??8yBZ1l}Q|vb=&sXzagXRfqGf(H(% -T$cfXBEInepAY@hkLSk?b70-zB-Tde$}*Ij~5FM%5!}+t?Lha&3sX=W~@K7R_bY8j+f{l%e75zm&rt0 -_0fx)Wdp`(#$Ks_?Pn7y-k;EB9%USxoq|2l0Wp8JJ32G_XZjS}a2I5NfnieO-Ry|`M;t^uqT29}#Z6J -%JL2ZiQ%cUGoy#_pXvV8=kLIqzJbgTF9_?7DLfoy`VyvghWU)(|n>$F2M2XufV%O{@(2)_Yt~|b;)R- -{FN!B~!X2x_Li7<@Zyz>nl(=s(YhoIW0Np+k$`Mb-VG#hDad~c0o)~k4Y~*5je5T -UAE5R$ja~3>WVTA;hoB{3*?>J!*}K&IoP_o+!0kD0u44lUbUg3m+=HPAIgSrrju~l>!&b6bN5fI$OSlr`#AAWZ@vnRPeHa&?KP2JxZH{6;WW(>Z# -I=G^?MiKn)KJG;Mbu^|D(A$ilMONEd0i*9|28_JH4q6!8(Gau-c347LVr}m0*@Ly-S2l?vV|% -NZAQA=+#eS)+mKj?#t%HRvm93z#o4ytJjVi~gvkkx4yt`&xw!~ADM}((3u`Mx8U -#W1}=(k4J?uD_7w%V3}tsWb{1y`S$eD0g+aI<;A^xXKU9aq`Jzw~y|^LE|t<5xnZ1X54rpK0ImVsrTz -5HG@0jh6U1Q?HfGwU0fP -A<1BOC&yW6uuw2-nAAz*;*@Ebf#5YM_d?TpVqan(}{(#-ZFUVZ{TmTPpa_Px+`&J%m73bj-=5Z!LV)& -CMlJnJzL1zg3Vv4ijshsEY^hnD2`mGFe6UI|7R3PdFA&SrX%2f!K^w?Ab$im0A3fS*WH83x2RVXF2)W -LqoUfSq`=ilMh5f}IGAYo46xzx0q9ooY&K?I7GbjSEq|#V(J56#2lAIm%pa<0T{R`5bIhY@oJYQ^PXrpMggOT* -gVEKed5o*RoIFs6W1Z3~c7e4|`E+TFp$;sXg=xJs%wqI$(TW_jDSdrfH5{efH5MBfY;2HfT?<(HuQ+g -I_yj4m6vF3lAckrUYIiND`jK2%Hfe6dSzK001ASHp!s<@2*Q(3r+$n-+YK0MY3ctvVPm)xL0#C`LkEG -T4V{6?e1_V0`)L81hsq%&-YJob6j}C?a5c(j~IIf|MrQB6J23HX=bODwdYechMTNRVNTL&5dRB$Y(*N -&%wD!7F&Y)4fvjrNaD>sGGQ0Uqe^{Pw=d;{hSxAs~Vnma%D<#b$g3)+w9;0zKQp&qTFLhs{m*pJ(B-8LZ&VxfaUW9j;*~u-%NsE`9}nJKV3Yv=a~;$%KUtDP(7; -JiwuhVL_%u#Pb}s2>(EsUef9T-gS(O!PJbXozJ;1fK1|_dN?Mr>0s4pLAE1nGpFbHa-ssZrXW;2MHMg -0vhg2aw~qpY`xs&mwfVHfc6(Zy54lI#fj)9m=q+zT;MuBLH7-SB1O=T~3Ej|bwFwZ)ooCv?)oMBd1Uf -}oYS>enzTumpALWwV9es%Qb9#Z`EnQbs8sK(ZxoTe0>CmikqX-K@v@EVoZ_a_c1=-{9HSnHTw92oao^ -g1V&;x-B*+xH?jHkUEK6h2BgKlK$WS-X974frlaO1|=9*rZaCPwaplt;D@RWU{X;FrR0NY0}FBDL;B~ -JUu<`8^nEbR=OYocpyl(=fyT2fztI>M;}m -{tWq*XC8&5o?_!;XaLUO$!yjq?*r+oYfu9UB4f+{suN1i~e)wPsW0ObkFC^uQvfr7DKg@@1c-5!`_-=&f|Qq{p -l)CD>l0z7#5@MeZf%3bhW(dRALsFV)1UO -`uwbvhV===03d5T&G0j7As+IHTs#O7??SgGBQFMVhTc$r#1-4+cTz5X{$f2^2FbY<3;hX^zh#kBzqhu9+sk0qF73E-I+^HFrj -J)ZrBLns>D;2i|~Bz17dhl2*0Q)OyUE7W3f4t9XlhczHN(zIF^RaP_feC257v|q3J*#;5NLRf0)7V^Z -)W2sT7S{3FRuk(0aL%Yvc?)2Ll0Dttj(W=KtQ7AJ20G1XM>T&NyIE5LgX}tG%(3r_gWmP93rB*+5FLG -szucl;HqX1~pCUpr^__<2*a$}I^uIMb0OiO2g>pyl0IE$>7wvf|ho@RJbEwYL??|ACkjJ|c~W8_zAF6 -p#@TFoV?+*UOQ(63!{BDrPFk<%?|4ux>Fn-&T!azO-+Ht=!w2RAGWK*{-}V7!b8z29J1CX2k=m8^sf$ -WC%Ykw>lcx-bEzOPJdTRR2vA>D}6lwlfL-Jq#l2#*;z+?@_;OX}q(+f)M?#!$b>`NUD(%P-rl=y1KEs -5>|tX*5O&n6$3>-1Pu|4bX&*>S~c-9l6YG0K7s~CCO-qrl;y(7K&43LBw(W}O8HS8$fXu*xnn(0UGgK -kXuE#}x82#l&ci%iRJ&tL19)FAAM98%Vrtyov^xv%&;&>;t!B8Z0>u&UG+-;l9+j3<)nz8bX5q -aor1x6ML0z&zCj-j|Za6KF(tF&DgKuvzYqP3` -0};uYZ$EtH}D43k?i#~SCHNlDd+|_+&dR3YMLG2u;UorbIZaRSPD)Fqe_BbBxp7c3Q2#O99~=;Pu#=H -%keK~cuCr20k61h0RMRL?(phZfmRME*Bk0|X=Tl;jhXSUdPj!lymChd{%vtZ3jWPvXn?5Jt2JE~kd_K -fS}xzPYo_E{LQ~fY9%&1>OPz9${_eQCea(dQA9?GbLs`)LNkHNbUXyd>DpTJ+!NT=A$Wi -~Bhy4qqugA`8~R5NN*-ZbIEruD@|G0wlNLNB7Lr{0hon&P)kBJib${vQIQrFV;^(5Z`xmZ492jR9IH? -_@~VfvPPo>V8Od&n9i1q^a=~FLJu1X)vOMqM4QIjjX -AMCzOUDNc=+I{Dq_N}7#5Dw$LHKE1}FT?Jir^Yqdjp89ylPMiccjDgqdK%w-!g@>$hTSwG~%tJ#D%*9 -+561HP85d4`W!^^@IU08vddAKjHSs$AB)?l%Yjmv0CfISNmO -NAJE+i^4kw+;Suxu5CbODUh$F7xfWkkDH`QFPxIx^E;2mL&ozxemQo{>4 -wGdh-r#u6_Fc>X+n{#wUTi`F -eIYWz<0tSsFlR#bFkUk2K=THJ_ZE*9u#r@2@_is2XHzH1I?$ -(0ob0hozpH{07C>jd=U#*ZLhB=3E+Lnsyfi45$@SwGAj0X7P%Q^zhd5F^K~|XwOPfhJxHW@rd0VY$dMPMXE@qcf3`Z;#0B2 -FO0lfEBL47vO>h}Mp8R}ngihvuf7SZ&7?r0ovZxKc8k{Ay_Q|l`)ePI8a6kujvo=STzO36ksJQzBt8k -*fo41#|}%%g#dc0Mp^M63qM!aoq9x19PO-2?PSUtnaY01J9gA$yRL*CA4L-}+IE$F62lU0;lceVQwh) -D`P$ti~GyU?|43ggAQ~&($?jeB%_lpB3IMm9KP?`UaJVP_S=-!=Nv=uc)eh-={`@yoU_Og?RrUsun)K -{{9tV5HcZ|Z^!V0CBB?W?EmnC_<17f{8;~xCLge7IAN-G9Dty7P>#33L6}(&!F8TvE=zP%=_RuTZ8xC -XAq>-sH}Nx?Au5IctCs}y@xU{oJ{AK@1PcXM-EdhW2K7NiA4kfyU`1OcG2ky#N410(0j&iFp%D!nfDl -bY@MpgK61)T`XZ?~aDAt)}aI?rPf?%$5s1uLqXu~1O2ucA4!DZc(U4wm+x -fWf*5OsC7kh5E?iAB#8S8ErmS*-Tc6v1nTAzdhM5ra$#sQekCmK3>HKbX&f$tbfuq4IiV$3@+XMjjex -LZ=!xUDH)|S`;Pek1+XK%)(BJDYyD3g{3C^9F&KU#flki!Y$MQwCFu=@fhLgoMdHRLP|cSB3Xf_A2_{ -Ifexv*KQJGV4txItP)h>@6aWAK2ms3fSz7=A0006200000001%o003}la4%nWWo~3|axZmqY;0*_GcR -yqV{2h&WpgiLVPk7>Z*p{VFJE72ZfSI1UoLQY0{~D<0|XQR000O8%K%wh-ne74NCE%=i3I=vG5`PoaA -|NaUv_0~WN&gWb#iQMX<{=kaA9L>VP|D?FJfV1YjAIJbaO9lVQXb(X>4UKaCvorV1tSo`RSW7<^*iE9u$E#Lb3X(i8ltAR+&lmiHp5o8+02l -R2?vs}Q^*Vg16tX93IPUk`xAF2_x4z1;yTOVqJSsg9_ -`$pE&UU!xkoW>o&3h?7U!gxjk|`zRRJ!1z -K__OC}TwR5wLgYx{PUF$h1a#=A^fZVRXv{QwK-cKI5~_6K$k7AToyM45Bx_izJKuaal(b-lEeyOv05% -UDi)Q-cQC_VWl15U8>Sg1#k`Sx{vjap#o+j2^a8oT+C(udA|5`|K)CddB0f9Z`QZV@AFeYQ49%jG=i1 -K?eRfqMZjPQ+E_yy4Na?UHbwx=j=rI3m2kE^y8-)tmZvrca5%meI)wtL)NF~amoiTmapPsVP|D?FJfV1Y -jAIJbaO9vWnpV_WoBPyX>MU`X?kTYaCy~OTW{Mo6n@vQARGjf14qEHmmyt-w!;b(D28Du(1##UXo<2} -OQb>&8gh$^q -7;&8QN%A_9Gq)A&&(3YNVn4Bmhrpdh?uQ4~2-qz%Qd1e-usp9?*BX)MOg1Qm@wL@;>qPGVPm- -`MoiI-8+QtRjSuZ2#i2~p?N%EZ6H3+?BcZ*{y7WYY#6g9V&9dD(2{ZSNdd6O{s0m|QqtkQ-B67D3Jr> -n@y*R^1t%M#OA)QVNDDpu@@n~jCX)$V|AOTY+5E6X@504K?ae1@cWiR`eI#g-jpU9+_yT2aEHP)NX}W -gBkRd;?n=whAZ8jJ@BA0K{0woNNRIrk9UwL6US#Jo38bPFRAPBGM6162w*614cQ8ug`6Y-)j^`6zK^YSy|h2H5*bl^K$}p?-sp|t4HHiiGK#AXh#M{yviBN%M7(ocFc)k97I-)#*N{J -m!kJd|-ZQOYQh~Tf5CAbVB|wEh7DiA|J&@NOV6QkB7$pWdSxkw9UM|?1-rf*?R%-QwV_W_x2)K110W| -fxDi1JxudlZ3vSs?bf8PKTI*fAXZc%I>0Y>0!TB3$o#{xFv_Xte7K#Ytz#Wox;Buw3L% -a7JiULq@tc>fTD{8QX`10J=+F2@-)Wed40bNe^;t>Nh@YnZDBtD`-KtO -_z=>~=dHt;ds@KJ-U@B;k;yUh{zS7!bfv7 -qMwfD3lo5Wt{2h#B<>oN57#i -qJG}*Cmrf->)pCyBe%|dI=FN%#O{j@{U7=7yE*42%prR?)4JYO9kwj*5znOgd>3ty>D=Sz`BO>oC -$sHDYf<3;hPY|26N<=G@|!OB;uH2*u^a(+g@yzb@D%OI^1Qsg&ip;tgDXAQel$H0Tj95q-6EoQN+?_} -8nMu;TBO-WgHEG2bC}Q2PaLq?m5RqN2gh-vNh^fikco2mCa9q(wa*wl3IT-e&4YqZ`w0D_#W74Ts -e*JWYpvD0>VNc4t!)vxCUsJLY=e?Iznj{bMzSzA$ynpdxWlQeF%}j1+D51+6aV1s%M2rJKW2BZ8dVUg -^QMvn+;S<{9^Eo5W1-;aX9*{bJBTd+$Ek0v3KludVsd`zO$XpJoQ%VU3wM%&n4ZrPD=;R!mPo3=#SfB -&q3(x?)W;p$B!t=(N_+-xQ-hVev%VUqG%oZzHALXVh8)erE>JCVOGA7rlSI%djrba?}9)71yD-^1QY- -O00;of09jk+ikX_&0RRAl1ONae0001RX>c!Jc4cm4Z*nhna%^mAVlyvwbZKlaUtei%X>?y-E^v8mlfi -D=Fbsz8ehMMGY=F@>=poBIG-!|*c6!)}P;9ypVp|#|C+pi!jzvdl>XZ2W6eazn8`7NsXa+YB0tnR^O- -{&z)$QOArZ`EyiQk&UK~|@Wq}qx~cSbsOP_1$wsW7C^s>ZP03U`!F3>ItQv^bzRBH>fgjE6l{y6>@aO -80!4vT%b?lQstHkWKh^KvxT*z- -(>E7W#H^tIgBnOS^-;oTdK5+je-JTJuQS}bldpzFy><#d4PQnN-Bn?rjWZ{dZM!z2NaZR=<7Ias|2zA -mOmEMjLP_Q_jTZeB8p{bqLGPNvrp;2@a7p?^FtAKSx9>>sl)r#uqpp=8Dmb3FUZARcyR52Nu}h=zlus -B1I2pBTn>9ejY-KFhh -K2*A-pM^HM8k#4_C~SY30q7K#^K?S#=P9pqn1UOSezBCy+Sy{`T%DIv -aOLt}=JThIo%99hiIbu`(66t5g^>-x^eev0OyMSCG7Fu7-~1>^09#o`l_mM<3Iv==7 -1^S&feutEFfz&FFo7VT1NaE2+1KA?%R1w0Oo%?Dv3bifB#bH|)BKDtvF)9QE%UiHME!wCALXr*60pMh -ELQWvi1h`)6#>Ord@sgE7{PZ(%pA|SI(xdCQj*#bpeisd9f7EVd_YtFQNqApHAK$6)~V!Zbofnqn@aM -*H27e-pwKistJ`|3AYv8ePDg5VC4ihu>1y0t4jhEEIWOw;+^<$ -5ZPzc -CSj{9@`5G9+=e;WfSW`k(9*Y|8pLRE3x!G}JQA{Olsqt@Ak5>Qm@SN(@w$|}2RGM3L=@8*N9_}X -hnG~+gJxi3B4kW!E{+SOUnrFDR(}0SYrWmE-=g*Hn-hM8Ay8Cqd@_Am&?|;2}e3?&UK{Tf1MZEuk5IX -HDrH@10n9Nc&MFf+ZSMlI~E%-(X=%Y5k4IXHtRI`)|> -?p1b`Yy@^QSQF)?R+RHIL#@W4an{52Yc+yj&SN17)VjmhZ_!sXUbr!TtsFkMPWuwT3VlN5d`whn{hP>Tt^9p1wRr`C-2-wTCsq8e=fAWp!S64h6i%PNapbfkzL0f_OkC`23q}(uJc -|qaKx^4P7owTxMeEE&NoPuOaE`ds{r6@R!Vby9qhHtzzlWE9Yk`%2b@EV%W6!aC7@^S$6X$-+M9eaNK -_|M{b{@dN{uLtm9t8l=zHTKflXs4%8GKC>W_P{-cF+_dcg*avmOdCY<;84^YKC&~aqEG4Pf{GhAcA_F -Kp^}a@D)|m$6&X8KWx|D8=L0bXUMz+rAG`c;@kl$htVmPHCrT&fU+XeMmy_GMr#LSri=UHYHADLPiX7 -skY{A^GigaG=)&Fl()B|FYBQ8#p)% -!C!E2`Bvv&RhsKjt-z34y7D!m@P!TyYURyhwa?kFwUGvVzIBQ=s+l -<{OG~dFK^=fh}I$?c)Sm_$fb4$W7n146gu{df)X{y;Su}A7vch3*m-dg -Tw9HJ@si{= -tSvtK$n9L4b;!ug_y73r-4Fj`qaBvv&6%#@drKS^-b{uE4$7c=jMGnS9(U=VNkce#D7~Z6Ts -0DZ&y@6oESXN*~W=Z;{0R4Nyx11QY-O00;of09jiK&s+yy0ssI-1^@sd0001RX>c!Jc4cm4Z*nhna%^ -mAVlyvwbZKlaaB^>Wc`k5yeN;`4n=lZ)^D9R5utY*!dRr-{sNGhova8+nj2sw(C1TUDNjCq!W8;AHwO -oMle7yH&#trE`6vlKV$bGNLLPaPnmIHO2+4p>%5wj}&#W|Ip)AQ`Duk|Z@Yz)=b-% -LeB@%ck>fi5Z2T}$$G$6N>DI70xT;MZpJ2CLWg`p0U+ex!$8>-~N8BJug2dqH9kf8YPiAjgXCac14bK~c<`v#dOa3qo_a_kht -lkmm-CYeTPfw--w6kc$2kCpDJtNHM3}FqF}D1{Qf2kjVnU$XC8UmC`gQq_q4$H%8P8cA*`u_rhxXz5NIx!hu*7I52-#F-5}>$(V$#XSC5s2*;X%O~moqT;>37HYh -{f0BUb*>nlxEWHwqYU6?aFyu$CV#JxMdDYKN5_#JBMidE8*_~6IM0+mbIj7i?HJE+PwE8K{o -v!fEOCIk>pX^w?f81-afi>RE_LDpEq_T;ABCu5>iT%?bf!1BSpHuo^1xx90+WzqK{pQ0u`mN)lfHhl& -EMRGeiZrRR4UQ9%E;3R!~#^PQavQENU$|CkTq@5OZ88jZNrHku%iXw_z|17`{;}bb2{{c`-0|XQR000 -O8%K%whP6PIz`~Uy|@&Nzc!Jc4cm4Z*nhna%^mAVlyvwbZKlaa%FLKWpi{caCxOy>u=jO5d -W^fg3wUd#%#62iefO3qE70r!Ga`M(tQ~UnU>Bm8;R6NDvn$9f8QPTq7=K`!w6!Fy!ZX?p^a8Ki{njcO -AT?%L|!Ou8DATfl@{XaIS4jHx8??3y=-XKuU>*oR0{QN6v1Nxg%wJghUSCRuoE)n>kLj+x?gIobVFY7 -TiV-Gwlf$fCd3dvThE$|eyS*@5=%-o9ti6Ix(3VFwj%7MM -Vy}f2N8M9lCo@rn_O(A8xv?}LTBs|OZ5whNHfhn2opE8)TAnS+;pH{0yLXOUslGc+w;p`N#EUPQA|_VWZXg+1o6AY`Ni_>?egLRZ`Jw}BzBCP0g|ws- -$Pu-U>q>~+yfLb&%^_qTq+4ucUajIH}-fpWKQhN@eUC@B8f5RTko*Y>Tw?*yh?>3)Wiu-wvF1!++BrX -=)P~Zn3SZnhHaE)s1G!Ex!3Sl3C1>wi~>hKFSzEp=S(GyKIa=u*sU(Xr7ukl8zo9gj0&Q8?pS7yWmL= -TiyVZalY7P`TO;7uTB!`UoU%+9J7s9Du^LIEmbB5&tme{Co(#cyzb*TughX_%_bg -F)&O5Rjr==o9LohQbl@iANL0}Q|-RSZ*9}3V=XfSv~4|eM$)sSO*Rk9tI=$0vyEsAfM!(wl@O3q51Wn -%sF^B2*pu&1%hycaW-A&UW`ppqj^lIMUd#_W_%>(IHRkBC=k@tIedassipM59(7Q5?+L5PbjD1SeGJ8K#}Jsyhhl6u^vBYmKYpk -Mu!B~1sYL#P4wIN5O2^BmWQcq345p^P9W#%o-~rVJkk0~I?CXAKnsyqyzMma7L~B!z~OTZ{^XuB|4`6 -;1@b69(JJ4n{a?xV4ZMmS{CPiw)D0OHH#{I0-{h&jLz}GNf86%V_F%ai~_wn2h&42-xg?W|Pqv2+;Gx&fb-j>&c2j14ucp6grA3sIL!cVAEOoa5vSpw -Q}A=?C3rd(^0b1iV4c1s5(y;;jNqyk}qdmC-aXp|N(Skg#(7ut6ik69+}pJC{BieqSFRYZNqBGJ%l;8yz$V)bntF7`34eOxh?a*#-ip}kcZ#E`pwF@z -mN>*im45x?z;)ChL@bW3e?efvY=i;m=GIvS*IwK$<7P-%v{f1QY-O00;o -f09ji$DAlxJ1polH4FCWn0001RX>c!Jc4cm4Z*nhna%^mAVlyvwbZKlaa%FRHZ*FsCE^v9RR!wi?HW0 -o0R}kJqY`{^1=CXAW6zOiZ$syaIn=T3zg+ogti;YBjNy=WM=zs4Ul9FX7$f7`Xu_bao_~y+Umew?4u_ -#;LT3RfGYSzYjA(zh7t*1qNAB{@hM~zfN9+jI0kE{&>6F8@`EYh)3m8qSHrsvJNmREpq8d-0ys4^o1x4Iep@VBQGA~?#FE4>E|19>s( -9X&G+BE?Y~<>r8Vc2?PG-1SORh7y;V!pK4892Z7rR9CF6~f0>BAVmMT}W-k<+8oqsEL*-L4}B>smKR2 -ehq>m)^Pw3kYwo~1!ewF$L-UYbh!fx8uXhYV{1+=Vb2c~!D?OiwYPY?X1LanEG@o0iL(m_7x0vJH;YvouCYd>l^`)8UM)UsR!k7VFqTllbb6Xnz&<4tQVJBJC8F -+>lrjy1=!0PR^BWUGD@tWcsuD^;d6Y4tvr$+8X%*v7`{8UAZEb98?0-5Gj`~LTaI3G`wz&Mf(Ba$Zv6 -R*k9bsS#0f$(Z54F5P{CqJn4mFoHk@tx$Eo$qE587h8A25lqP#FG5d<#?vqW&krqwO(sk6pZ_w^2!d3 -yf5&|h4=%qc51VU~p)Rvo&f*`b5WgSYAa{@U?&;Y}d_r2FvtT{g0V{jM$gSg!pvb5YWU_isTz_)Ziop -UDlAQMl^Q-`Ula4DWMbQGXFzA$!yF2%*@+2^cQN)t -%E-xw;FUmZvD;|mOy`Fr}w{eoHFKK>zn`F1@Yr)l1m2CC%J}xroRQA?5(=ty}zO0fwE7eq4yydH^-lb -f0Nh)4cd7kPHRs50fHwms6M?eXj>pV-rQY@RnDv;gJ_X6d@&|Xwd8Rp -0SI#idk6|^%6IRt@8rtyDNCfvkJzsYQx!LlfCvccuU0jWU_U!JI?KxY -<87yX--*H%EQZj)GPg$I;2$y9VmnGXL)rPGQR{+OXEEhW+_M!{EjuEX1%NKuz83UX~42ozk-vE%Hdf4 -&7m{9ujuE=-5O%L>K#nsj<>I(v0>5&in=ayET^c{2-^X(iquX*demyOR6h&tAX2`gM9Cr?~;>Sy=%q_JsNt+X&>k(V -2HZN7+~B5zbB~IgS?Zi(NjM;=XOu!F>yEYUF>{#Bc6b<|L*tW%QoUM#{UVBzIypGdUN*gf0V;55Kl} -x7xL5ct_54t#(Lv(rAMXz1N;3Q5UYbqhLd`K23snOSF1=H4uP=&2GS~^7lAMqB&GWOKD`A?G+QUPbwN -uB1^uV-awF^|76b9OJ1}1n{bj*{JYY*@d@q1X3q%+z^=?JfqqV`#lsQ2I6YYVF9-v+suX*|auZ@|~gTrmA)@<}E;b6 -K63Rof6VS{LnC$Vt8@9i;Eb1uMF -I+yoysJR%ZB>W%YK!3*l&piL0E&X`1gtHUs~xWO%Ikif2nsTg`Hqy?|3zXH^**ir#RX0jT{;o&mA(+4 -c#5lfmhUaUCd&)-kRXyeI|>Cx}KwbV42puxs#CaTrozofKuIJfY?Yx%2o%(voHeuMN)Zd`^^#@+{JaN -)BiePo6|WzE}C4ZR6}gwZIO*$w*0FaKRHq0+K12wYb>^d_CG;*-h*h>>QjKvA61kNnZ*9iPzg4q*cj_ -JLT-+M=%EXVZd$mma`2AnkOY?kZFZxV(L5_*ptJ&K{wo9DowD;^{4lA8lMqEosgO{Jls#Zzbeeg}(09k9CPq^(Qh^TwIRb=e15Z8s2291hgMoXp@ -&zOvXIkp_fnFf$5-zY^zN^kYEKh1^%r2<0ZP(^Z=m`sqQLYD4UpXJo`mE|POf&4T=a>KiCQ$|Q$P)^O -nEb_ZGsb~V@nv;@p06SE}QLt7n#Vj>JVAPAfUBvFkT7Ii{a{&P4YCh2m?+ -Bw7*JY!2Cl!8e>cyKTQ1c%sy(B}{WqGHO#DF&8!2Mb5mO~ks_&cT+vM_qs%9_!BM8WABv71t)>i7CL(Swp^(U<@B}Fiq!_R3Y&o8Xw4S-OHb)e`AB7{(pri1V{(cL8!!Z1y2` -X0tGfH1P<7l4(_osPwYxx7fsad-uumT8OETSl5b4gGNF)7q*sF18UltudM+>WfhKMmmA(ii(< -)*Qzcsv8neeCYKtx59l&Rc?IseV6r4zf*1z9HV+9nu2_mF1fCZwhes`?moJxf0h5Iy>?8()^uW=uG!e -m4!2vfnH*ecgpe|82L=#I+I(E%nfr2#qT?0P!8{*UPF=tBdK)9Csg#r4GF$w3l3C>6|4?Qv=!*QdrB?l~%-nvaF8fh;qj;n&Go-m0-*f7&~^HAiC&(j(Hd;5<5 -T})q3XH$H-dON$k`sD^6XV+)v)BXne8S?&98PrzCH_VkA#`)>5&Lvw0tBi&i -MIl?2>!CXTt-Ols;W5SUvXEl(&NPZKkYNR2i_>bM!KGi@hC);|f^#poUp|d16N_}!V>gmXR6rn{FMP@ -xEPuE5YlFFT&jLaJq&qWJmXt7?rYP^zOu=`qUpZAh{KJT5!lM|I<_dyv>@Bp|5z{J9#>8a0(nzyve#k -#FhMmNyyl5sxpPEKk;Q57m&XoGvGB#gAIbywel_n-|diwks{-&;A4YLJn0a)7rThx=T2$-W&2cRZ)XqFjuqcx{>p`q3x{p@(YzM}86F-* -DC(_{n;ipW5@N}D`^n^zFg0R|gdFqW%4UPNAZSX}V^q} -!Q9|PUo2>M!>7@@6kvR+}hj$;XxGqTAS-W#{O=l@MSec&y-1--_*jv;z)yf<$5jV^A}Hhv*I03Xk8O+ -rNNYV?5|VCxn$1V8ULA=0x524wVXP|9?}`Jl%Rj6IR@eLxlj`0gL~Rs{eV*hnO>cD-qZjbToXD{_>*V -IQBN$FpF2=;PzJN!98wr`ty=d5b&Uc43^Dlh!ucGD2Fp+#TatCN1nbJbd^j?M&M-%0|$8>5fpAjm@w= -Jv)t;N7&AX(`hX|at4Mz+u=vobE29-E_h(K$sIUcz`@9f7zscHzu&|d7Z+_{<6DRH84gCPK{UdEjEhh -)m12P8tScH%Sub#WL-s4y3M`Wqw}31toQCfhMA^l}cYNF@XVh%WUM*2od?dNv{ve66w-+_~8v*Zo~BVs}rOrutQNl6o4$Z92g3kVJ3=X{UG@IYtdVt$UhUm3y4}XU -_b<%%}NxA$yk|H3q($mih7c7WoeIozyA}+M_`6kdA!WGJD}|%Nt5b9SLx)biNIk-I<7(tfhY)xsSW|2 -K=8idqso?4`45cHkHVk8QAB6E2F$Uyk`fss&XzpT7!o568nYn@WgokyN9$+|oCWIpfLIsj*Ihx=MrP5 -lprA=pJzS4U^V@L67xit>KhLu}kU89tjS@{`_Eo`-{}ulBt!=VqYcPVoD{g7sz#EkqWV4SppBF@;(+BT9xt>9v(sl#-uE9$(JzQozDE0Dwr2s-e!4W!f -WNgk8M+G>B}NP8gCb?nB9-ry*5hp9RK3-rgeIzV;+&4@#fejS>Sk7R$yxAPHUNzAX>m;u?zueU|Mf1zJmQnxmoVwCGZK> -z?I|Y&*{!inD7crxZ?KtrrM^@i}mvR(-)&>rL25<5uS$My7D5VQi-7G|JQ$(U%ww6|M1}(c=)G&*W~z --X(*r@3Ol=>?$#95q8m!uNF)U&;x)@Z9|0}iDFgIC~!z6Ga0@Zu|GbeCCyc#-CAA`7ChDczhSrU^mWY4j&WUyl&Jz2ss=6Zw?XH0;2 -~#e;U4*3d0~BwA?5HE^{;MU-|B=tz<8+XgC=VXfRTC7dOz&Y;;wZM4^~EEMDzpS+I{75L_c>KE+%zpl -o23-f>(WJL)%7Am6|%YrfRLFaRetuT+Q^5kAL2Ms?_HbMTMycD1k#E6v2`(dP;Y9C^4vC2!CSPrHE(t ->K04Tm}Ls$-lgLWW*_3ism#7iXB1!+7vlJcF7(fY@u$=zx`9Mz%@1 -+Ta=Oyc8QWQGDXwE@R8gD&o0lX>iP{u6^<~I8z%+Wqh(FC(s5-MH?Khb;?7~kdP0@wjjBSbHcTp2 -+8}Rm=mIg9))wlPa{`MTTK9)N69!sQM-&AZOUZ4L#i_v1Kh!?JmMhL4%`i0cMbsj)P!&Q;5heywsICG -gFDNA@q9lSK=Hv*Y!l&?p-vZ~$3Uc%~lB%6#9asj9d)Dsd?16`~bmwVru(&{CE?&FHpKti`V@b&p8HT -d!BF%maZ)?Ts9oDg|_>N~Q!enjsTUL53(|RSs8AyVz0D|j5iCEf8s*usel!ks83NZ?g(Zz~dSFKT<=>mA=BO-{YGIfKlbyK8nouudzsi2$bFc|nOYPWs~H(H@D2Qd1@)^4au& -8|8u!T(;#QC*}hMM&CpF|Wzu^4>*wt<`@=l^<0RFKq}eXI!h2+4lR!8WXPWY{<%0+dq=!mfEhT3VrGR -2w+s5EA~TL9=Gvy}(7u$0ABqO)s|Tu -LIxdyHze1&fhUvOaQ~N&XSRjd{Lc&Y$5&T9$uTG=iDoD{aASgiu9?sgHu?y?6E>)#Uui{ul3vtmj&W% -`!M!fT0aC)Ynxe2J}U?@fmZQinCD5_MZMRq1@Q>mh|FJ4fBnvW;!!~iq}Hnyy>4hR+4XlY|LxJ5tBnh -zy$VYju{>DGCuIvgW_%(CK|Zff`I>?0V}S>J;TO-^>FnG~3KK_noL;6SZkTAfnf;@>?}o|-tJKl(nhq -G5Z9&>!4#dpM|BXO{NuwYSsVWE$Usz2#oE5VtD;LYn*E&Kd**i@>;?|_t977(O@jn8ig1^!lI6!}h9t4rDcgq0@h*B3tr -Vsk$kYw7b)@9QVq;DPtvZza(Vf6|7}XRA!)lVkCGR!k}1i!YxC2HrrODp=galmQKiBdLDR4uM7>MDa? -NM0rRT$t=V%GL2WuCX<#po-Z~@6aWAK2ms3fSz9?X>2cZb8KHOaCyyK> -vQ9{k^ip00+Fk3q%2L3A3MqU&UdAiHEUHow$Jk1)+CjqAQF-gQ=~#tmX*!sZ@=!wg8)T&X6y3dT&^)6 -K%>#<{`CXk%!_=*qG-OUHid{HCRgjcs2HD>dAg}Yr0zQ%{d~#GWh!U(cEw|JotNgS+}X=5FEW`e%FZ0 -enaPZov6N<%rFc!{d>61yrY}3HVHk?nSgb3VXQgS}>oE8kXQhaN)rXW<^SoI7qy`A+P`q|B35&>&)`;n)m=!_@P!$<#4^u -K`OIU(Au*3CW}$*Y+dBZCI&e{XIbhV$nMR@&v(P<)8GLH`x&oCZv0tk*j~3s`rReBbQ{tatGqZptM;rMnLo>hA3O58Zyh^Gm1GNyMB*F=$8y|E|^?^jC2i&0q -yl$UPpgS^tuT?RLBLcc_}OYMbjxl_(SO*9}jNjDQ)U%QKLXT#2wlU7tXTrpsyr@|y|lTJ6??l~r*v+f -||rsp+Etdjj2kDmH@2IW2(g%BvXXo+WAxSq^WDtt>_8aaxFSlcJ)4G4_D79`CbbWU)8vyP}*|0W0KUS -s~wllj|Px!Ha_Lpa%%R-a$7d3y1pOuZRNsA4NhBjLh5duh@H#cB`a;IlJfp1+YZ^`)=@+73z07Rk6Ed -?1()fJ)9A0{#}hCc37rr)@X}xW1bI3k56E6|MhwNIJyqlw&3fi%;RTKbrg!QF*PJRR{~h`vQ#6 -tQZz;hT;1>zq2oV)6{!$tRo2oqQP=vvLe0v=pxkk}7O|Ym_>w*BswJ4`dHA7~bpr*_)9plQt$hLvyqu -lGl6h?oabjom<|vvb2x))dZx|e#9Ae<>-Qo7^ja77;M#J0j=n)AngLLAK#~xiP4Z0X5+cvxO`j-bOaLw8nQo}6suvkX`H-(;Hd -FfSG-6QnMy9b6PIqKGCk5wucXuk9ACF$D|_0fSD!t^GyBRe3T4xH(=I9K0PcbRx4vyoSqam3rNmyG%T -XD{;aLXqQ66B`*?z@x$>B=5o9hv*VJ+FOOlIW6k^tF^>*oAu>^bYmpENfIE>mFBo85h`A{ATv@r1;5x -udaEPPk_+l-JIq->sf*2}6Q5GWJvEKE~=wvj6fBS(;h#s*V_+WT}@|}sQ(v(odSuDIBJlnCsXmb4H@D -ur>;gZAe*?UFAydC~9JU|RP0X?Qx8P>t0ZYUEmOL`_^z9})(C=!^KIOcIIQjiaXY1jqC1DeQk -tN5_nl_*Y&E{U?Blfwpn9E*xx|c@&ERB43^J>@~=BriYQ{esAbHm#H|| -`&sPXd~XmyzxUQ5&3Qu1Me&k@U8&_TLk3LKZd(tBFLU`8Z~CfbcrD8`SHM;pCOEdgLZ0zv&8wxoNqEK -Srj@Z;+=@oWYXexLs3F)QVj&%$4E_`x5)zvp_qxLYz#h1u3KBMWhBIl-kvDB1Bk#k!U -ZUWGe&~7pMWp}o`K(e0gqI&?mn$ekwQLxu22B-vAJA;AB5B%Q!Ty*I9V4Gd~muWC$-``ru=jZEmSSA6 -zU43G7&4=t#UR8jzab&-$3t-ZPtr|13XZXAro6zuoiL}!7z{3Knd;3Jbe*VcL>?XmPiq$+sd2?+w_D( -z`Fz-Ym&!hDDxAMot%beAHvrYp1fcTlN@hK@a+`$wU9f)h|VmO<6q3ro5d*e)A#P4f!w0Br4w&B4H4Z --4?FgU?Har>EiGQPG}T~Wo9g=mRHH)t)wXp&zNIj -h0V88vXl5by^hrmdQsc_ux^IwGS!N-Nr!GPXS}-eM~4aMTvlYgU^7ecn#hHO^(32{2^En-^ik*_6qbV`T{A}&208Su+ -&i$Cbqe8Z4Y#$E5u&;E^V^8(Zv3%a{ALihFgdDv8fNU7shU5Iz38yebrPyz~D05l ->b%j$>fc~3V55##c=76E$>onW7v{i&ERvVV6_`VJ$3*` -aV*?H^td!kj{o`-&`__P{cMFK%NG12vDTrNR?{tqtms_Lp#7$k>*l1GZhtcxj9r%Y>Mb=BiW&-5rdyi -g~gU;87Fh_6f5f@c}g7q`K04t&#+C-j^6@T~2K%ZwjhY)pTs|kQ?-o6p8>Nz$&tWeBmIJIOMA2ZL8>D -|8TEx4mhB3rkWQ9x(W!!_-qC-j3gscgZ3sl+v#vZB~&{$wTxkoUx9R5`><9aOyX=+ia1ekwO_FbZVuLQ?P7lV;0+H&Nm{dB71pdW03D -^Qz=)?+3)*47~Xtqf8SRrFO<1f}joIMfr>gv>;zQ$-_pir5$5g=}2uD$OywR<%Pa)nHAXs_Cjsh<*qJ -Y)dD{Y$6bVQ0f2+Oai$uBSK43Z#tE!QC|OqS$DXfl*dB~xd -EE3!ReyLXI_!vpKxZ??C!lz+Km|Dc}k<^#3SPMw{G=~TrKfThzZ5uZ -*q%CnS1V35iTpmmnNh`t{dJK9-W3vI265)d`9)h%XnLtqY)QJv>*C8(vTZfOUIKh3Qoh11kJ6X5qTza -{RLa=@Te|W=qG2x(^T(N%-Q`~WS9f%x`Y?QTXU-?_|sC54}!VCi;XN4GfwirD -BaT4Z9UnfcF>m6w8ckHz2UVA)PqlSG14-vTi__dDHZ~Ain_BIHJt{Hp&f;Sm)=89L6ohzuhlM5gIl}Z -Q;Ui`tiLk?K~V~+h`TgcC=h9#KPjYi2o2F^7aKC?{W{7Pf&v{?U>c`IqDkSN{$;aLYMjrJzY9t01U3` -%s9yWKRlu?e?cY!sOMxox0*DHC`&qW1qrmImSWTy9$#`AES(HWs{;7_AlgGsdx?o3 -z`MVHE;#XUU7ze3R0fgkutrZ^}L5f4_3KY8&I6V>Sd@`9fQMfG~iltI~{>f&P9Y;d6(uc4QthgYN5s9 -~(Z7KaChf7#!;s2*}3DPs8`$Idpmw0$st5&fouQ|M2LJ2|Im3;a*O|Gg#Af7qZV8_|=kJq3_4rBjq>4YRn|t?i*MMYW= -oSCRvlM9W&*o@I(Y00Z47^e~PN@gSyJxY(I%~cb1xVqr2=$J+XL3#>gEZ -Q@M#}H5^RhZ#!qm$phw#_m?4)JYJCv&U`Qr^cI{k2d5weGy0r0JfWg>=^;l+-6z~G6*{-BI;Dp6a|oS -MW*NSb=sF)JldrwTdS)DgC)UH^Us?Loy5IVcuWyPvYGY=vIa(jZjUMc$BM-2#=#gdJ`m_d5vXC@Gw~{ -;?)zy`<6zwVR5m7lw}JP~#S@+>p!RiZ;GpLKR~vF6lJCk;&oRO( -HT#b%)`NBWUBE8-YTwu70L^d>5&~TK$S!<@Yts$tt*n0gmWk1;mfnDWc6j7d`Sf_j*pkgY)z -<7sGGOrr%j}=rLDULcV%->C799idX<17-r?$%v<6BlqpU(3EJP>t@LP+5Jd*I#SqJ}j`?a2X>4&(vUe4GM#r_)0c_hzDT=aZ0;BXq6 -9coLm?v+LLD}Kfw**LnfV#4!qB*#>CE$TdbliX~v>E@Kmbh;l{wB3wNFdudU>DQ?gOAH^Yx*r?`e2{t -i-oS>Bkq>ktCNFyk({Ee7}cW!J;LKZB72cVo>77Dg%=X3Xp-+hQhykIxxN4KskzkcH2eEQMTsn&r@(oHB} -edhNn2xcJU~6J(h~eq1b@^{bkJ!O-26(9KKY^!OnI{kTqboXrRI-v09%l5mWl9u6M{V8NCrTGPcN|^q -)N_i`KVDY@LBLO>I{$OqJr0F{b_7r%KjG4O+C61PBEI$ueU2hIkdN#C|Y*Q0QjD%3;#Xe&#Z#shm+W41H#;=7U^4+nJaWS;T`Itb!b9Yt|=KkBMQ?E%`)$@IN>xi(Qx?0& -7|U-QbqqdA%5CKc^+!hMl_dU^@o7S_Y)6I1(4RLC$}f|eRGRojO-BSYhhN|iXiE(MqA+SF^7KdZg7`q -LA=OQB5#0(NAfZOg*1LB>ug1ii6-xtE3?!F>fWHU%3q7&WG^PGS=l@ie8OjUOXPL1>0&BeNWQBSv`+w -T<4%Hm6u1DZ@>F3>r(w5!1vb@wN^_GKHZEu>zCK7GAPQS$`tx1g@Uufhwy5;Z7~RE}ZMX^`56af`v(# -gOKj6r&iu+dxxh3Yi>_*xYPrWcI{;~p7+N0>PshfKcS00s$Fx(!ry5XdP*Wyt~zQgIth6{R3N5p#h#9 -z^j-ac*`g9+GhoKnl&GcIvEB+$6NprPl39{(OV%@GQx(Brxqmep@ql -(k>bS`BIqv0w7n8Y(^L1YX7OIS3@q(sr@?l;+Q<9>9O?N~CUoQ&|LdCd_iW#C|Km^cs!&wp4?s>uSdZ$Q3Zt+hY -*f(Vu(u5uWQZNXti;*tjvgs?lhEuT=pQPYjZ1VKk~@wJ(M_+k`Q!d1cKX>6p*ZaM&i}wL-pekhD9<*v --?^!^GBPv9K@`nD;|m8?f1?GCII|>N6U4haCEcA-m%&(IV|TS^|oo_dmP?sfgO-W!Jg_6gJ7(GO#K%_ -SW4qGh&YSxo{mu@NwT_h3T)x0gsmr28#K*9)qmQl)PLfM)HhF~9{9RhDW}mXV+j8C -5(a)?8c`!z#xK89U!5p(t}B;yths*%y3I;zOYt%Nt4tgmx9_I@hK$brJ$$jiV)Nc>g2~dBwPa4jmp+xDu7Nshw{}J8v__}N)v{hk8x@0$IOx%Hgf;;dfz;F2TtYsjt!2%u0^WPtYVlxm<-2bDj0(1&r&Crc97p$DUj^khEdB<%{m%aaP)h>@6aWAK2ms -3fSzAAGx4P*F004d#001rk003}la4%nWWo~3|axZmqY;0*_GcR>?X>2cZb8KI2VRU0?UubW0bZ%j7Wi -D`etyud{8%YxW-G4=CP{f#+A;4`mXB{h{gzO?^feXoAblIHMp6S7zjJxNi+Zf+pe_wUaqYZ45dk~43e -pG!`_0^-DB+0vlq@{5AVzw~WX|a^)n97yToidur#w_V-;}-aQoZ;Vuq_t8zB`R7kq^6Zs`i8Er-(*RW -j7B?jVVs=OyM?k;s!9_6U+F?NilPB_T0rky0{L9zU-*bD$Qe4yo*aR -q*G|Q&4->L7aQJ!lr|fv-Z?z< -H##A;_(v_Y=%$$R!%=y|*duu=3YE`v>9ozr$Q*cR1+Bhn7z{5X{bgM?wYf)Jl^zG19Aed)%2>9{7-Do -H7rtOJJmTn1aC$!hdrBFUsP9Qk5b){U|>}I>;7C_?>3|mQ-gsn~pcKxZ3&jQL=V%RvGehS!npJJ=Kp5 -fESVCKXl{|EG~WPF87juoAV&uY4k-lg% -U&PV09`iXW{GNp1UMzTLQZX1ov1X+l9_dRLOK3y!EIyD)3c2G{0>h%Pr$Sfe2o -wP;}gHXFh!I4>U%y5IL40j1y`y@~vVtjmW>KB1Y>Q%dM5^wkZBKJKNAJI2^r0_3XZng~-U7tp>wFoWc -ne~*Ntr6E{?Y9Skx^OKW4pSrb9+aIJ0{#?x!svI3=?MZGFd8Ob^y8h+z{hL>Garusf!=c)H)F&?qs!= -F#ialL$rzjO731_SqDqoDca7U5O7oril!zq|tH}Ez}9aUm3t56sNJNOP0b>9=+Z)cq=2L7WM?Ikfc|H -fTvcW;-Ax#$_iTvX1=s+=$(Ja(^EBlp=NF2b(jqBw}^>DaGvO~^!DOI@VN$A4s>p0=4Is|ZyFz9RS5$ -}Q65PVet0sQAqglZV@}0pASKaTcBapTZQ$L33~9AH@l-l*_sFy^LWax(LweejBq`&iih$D81V6u`_O7&92vC>2B=69&bLISW -C#~Ma>>x{}nHCuAZcKwbT6#wCu;?$RUBVw?Y&)WTN-kH962GUxj3?1$!8R`1KJ~VQIeccj3BglY -fyeyO9oExZ%sd)t3@Z@V!pqUKXNz^y2V^r+gRcV8*KA+Jmew4NvW|n9v--iz(u_xrt+d#n>IN#R>w2? -oS?Eef>=0pCly$yh2XXrAAyK$fDpd(5;jSrdrnK9~4`}X-5gRHku;*)H6)&_vW3@A^}F5c*>dF3lexI -3vR2)_H|~d^}c2q{1?AlR&nk4@UD -(OUCb^c2CN-Q-x<$`aW!Xe@^JZizomcmZ$ud=JGI*hSF#%97nfe(bKOO`-`~sMWB%UfM;J1<}(SL_yT -@Jf;%^0)q?Q@c`;3|s>09FRYkt!qW)t{<=4EniXR=AtROU!#v?$lPibH3ye*G-**`QR_nkVRKn}_1Vf -wAu{)*NTX{^+|5e;}-hY6&Y!L4k-Wxej{7W*E-<+1S);xWil!s=VOb<1awUumDX(F@YT87Dq@xZb46c -2!3E>PjM4g)ee5poWE(a%JLIidzE<6PaeKiJYU!3CzJP2kcNTRMiWf4DJQD!} -gWIo0zKV|-IAFZri6fLlBmVC}_AmHwA|@0M(QP;wmj6x$9RT&J2;X`+KOK9-R9swhz%YNHup^~hFIr`?G${X~cNJKHzj^L+z*@}LcM<6C%0wXF(6*%O=46H13~uK1l0M)0MmQ- -1MJdeq}_{PgGF?Tayn4flQ<_ZHM`Y)_hdh&s_9pM84t5%pcJ0p;QH#^9yS-)juc!Jc4cm4Z*nhna%^mAVlyvwbZKla -b8~ETa$#y)#2fCM_u&^ux9OkhR5`hx7iTmDNrg$D|%f|IO;+vvBP%TaHF+qbrtW;<*qzC$+LHoZcJ#e0(gnq7t -8!t#|lPtQz_qFFx~)XsNAy9*rI!Z}ZFByN`c(y!p!&o3ppS{mt*bAB{#u$*pB@x-eW>r2k>W;4>bN>E -FdxB&Y0P_|7t}n0S`fNwo=*!=n`w*70R2?N$&8FtS1`B?=Tj8xi71t(@WJ$ed>PN4M4L4T)NUHp+3NW -O|Lfj|4np62Qw+A1piS8rh01mx{oux^xm)+2Zywy2=98M!2mo8M|?SzvITqqAs~9kF2(0U6)L*2~P>J -0&;?D#UQ+cbX%$LYY-ic;SCmyZds+RLvi7*#t~Nu_gEtsK`BLgPEtt1PFjZnMa>|-1?bw!E?wpCuFk|BeY#f;+FhSgB_k -VtuKF5ziGY!~i(xE1KPBv&1rlk4k<;WT-CR29U%3=$aKx>Zs!A=f=Q1>ao-gRlG$rDjp@&Bg^-*Z_)% -40Bqm(;fkqI)kQz&^cBaSyRAU7nvrwW3_lLMF{v;%h6+UYL9f8K-z9=09sCW|WDbg3S%zY{V)TdTr+m -QwxF|NR*F8K(d7rb*gRL=o+e6$m7~#bsI{UtQQ!?&!j4Jg}ifQ$o?oQZH|{Wcav#Kr)#&-{d6f|pN`1 -X^Z1O>Ycg-T%h{w$gdalf)F1BDg&#;wS)&aJ5kcnovJe1p2`HbsjO$Wksp2m_+z%Sqi0J01sBr0%61oQMU>g!O -{Hf#k=okY+0|@!ca61;N|2KGy?$uYr~ab7yAzOy5t)>zld<5F49Bxt2w(!>xr)=f8u2=u8h%UGRB1dr -!YFOU=^lDjIBgL$0$o|v~=}&`byEyM1}(3bdcNYG-GOcbi#)G{^e94Q6fFy9xoA%HWM!u_P8@w;U^r< -(e{(ZH_f3av4th<1KjAfUOSTYJXw$Ltxzk_B$yHA2>wgiE!-Sgt$-4W73ZRTf@2i4cr0#vDGn2qZ_2o -3_-jUgvw~YU8DEa+IjWiC<5Wmq1!%HyLz8w%Xj#yUybSLHN4IP4vF1^xg7_z>zY^aMf17p~Jm5ETwoi -@zKeBj|Y|y$CK+V5UEe}RI;YIDFY6Egou7aA1_3OMbS9&GpV>51o5XyhZQvuCk75zk|njBW*P)g-H_D -gNKT?e+#?TS+3Oj+YZ+lWJ%=x;`|Xz;gCR93N#{D0hv;`r{K{R82Gr7Z3kxa9aQnOajl3Wc#j -DXRWgWq>Y0@y)(3Y_ypZHLFLOcdX?0tcOv02&kC4fpM!MLPTvZ3rtEx(Zhw`)7zS?^@v|XznR{KWG0U -n~x9pXF2}&mtsjOS}4d}>}$g&epH_^$hTz^R?{*2n$Z=rY1El9Q2bB=*J0=(MHDgK)^KXtQwCcCiX2# -5V%U@fn7ZKss&k5^RJaPxQLFA<{6Oe%nph{`og2Xue?yYR`n{$9bUbI@4luqyFCH1<3wFM8g@8Vf; -vxJethvMZ=$jtH2a|b*_JkH`LSMkQo4Y3%tabq23{6e_ZAx!jb*BkL7BcEAsp&>MKed2b;p}!k=f2!z -JWn*4`02S1BjinIyxs`66T%c==VmM<8{ylKYJSbZCqrYmmmvm-SfeXoH?dNy7e{!H)lkM7$=w#r{$<` -kn2c#9lzmv#c)3$KER{|dx_kxH0AnAt`Y(~y5b%~mAm?5g`x?ZQA-0^ZXkC4s>0|0dQRieOVD}8svB% -;x5L^+lCriRC#(8;@i-_Yzb0rEW`MD>Ox?`QBKXiy=q;yMax45C@&<&J~my;8*6WFNI(JN+p3-S1POQ -jwManWfOB7^FpG?{3mn*syN^1MkDJTbzXy83fR5Tur;M`1y)OylBP~2ZJI3Lhh*e%$UY -k^_Cuj5LqvppG8FHtT9{AVw4a-pf{UQ<50g#r{+qFhF_+Ebvo+AKV8;4$q`Nx5Y(*BV!6fWWKe6XE^M -Sl5R%^ZB>$(=U0VI<*YBF-eV*S9y}}*q_GaGz1Sa~VXS;Nx{{T=+0|XQR000O8%K%wh{T7%r<_iD-xF --MrBLDyZaA|NaUv_0~WN&gWb#iQMX<{=kb#!TLFLY^bWp8zKE^v9(T3d75xD|f)ufW*ju{6V_-A&VNx -$DlRNjA+)c9WSn+eZl^Aqfpda0$?g`q1Cra}F*9L@IXDcBa)sB9XwsxqRmeuRFD4i^aM(eJ2(RCU>ps -4C5=U>fVTj{e5=ko*TInXJ;4eZYx+N*1WF`)2jR^>?_rqjP+VnY;|B;W7=A-m{CkNl`Odt%xt+~jq;1 -xp6lF?>H(YKcm;~on|F(*orI*s;C&q$%C2Ks)JvxQgvZ>^t~)ZQTZz;OGW; -y8Sl4$yg_Nb^Bd@GxzHeF0l9=9Pl%5_(oWXVdJg)KP>HpZo#6c9Nd+6;Rxo6tb`_0#uDRilzIL*p}&T+%&Y3wE9!SQ^Z5zONR1dF)M) -`|Z#InAwPllcd?$?PG@^I8{6#flxqKYojar|9283tL06A`;et{?$`cf;$#p8=6Zg7j*)0z -_K#IJOvXDq_l$1P^aOaNzKJS{J_y<_kAPN}8jpP_(#I7|td)?skKuO$#kJ{LN^c>C0h=KfIeB;su~t-huqboUhroblbFT_>XoXWUDlB8` -T5&@x;j+XQPNLWvg6;Sxx$Ad~?^F*OP;0BfnQTC-u%ps*Jaf;b4;qdP1e`#~X%!c8*w&%y;k)k3Qh*r -j?M_;BC&;$qdjE{Aajbi1TYb=Qj^B~7wo@x`paWat+f5C>0=7SR2Jdi_`Td5F{;N%j*b0duiT%ZaE1$xEl#GKwv4; -mC9_x09X39}cX!Q-St~7h` -%X?g`YINU9L4XDhq1q9L5Lr>xA%+EVs<8&Ns$*~AM;8R4m>Nq%lK7NOa>%Bv`)aW9R&KWJ`MVDc^ahc -Ve(F7MZO{Z4{LO!3djSzswBcZ8=#4C+cFx`k!k ->nJHa)S)BE9`7LZf9(?9L7XWci(f^=_Ze65B#{?z -p6qUzEJpB4TVX54XtBHe~`YvIB~%gF#c;9igD6qSoepX}9ru&kuTt&!`BlAh((Y)iP-s`n-ai$4$Rvq8lA^9_(5c -KIy(1FgXQ_6i2EbkSisU2FhUPEN!9FuL46eMKWf*YTs+ap+2eo^bfg$27mJw|^*Xb7c6F$)N_~6fDpU -V+UPRWY*ZIO%2Dk1v@Fgo~>wNj?%Uo!!+_&@j5ed@T$^G~iHsgQCJFOp{+)1xtEB!w@RU6)!i2WfGSP -h{CoQ5KyolBoz>QKL5c1}n5#j5hMo1e2wHVfMbN05XpxqPu`AR#XnS+XEe=+rAx#rV0EGdy>_%>~E+P -P7LoP8b954ra*E+NzW5ta|39~td -nJC=B{`?ryqShfKeukw9@F6qgDj7qSM+|}gJ>QD*qs@nW=nxDO>1;0_+h_(->MkGR*PHavCAS~u7q#rQC3mmjSF4P(tdp#6Lh1I$})>E~~6e2u+g@CHD?se>JBnD -%WC8mNINV8w@G0k)e46>5U#aYwuL?=t9ff|@Vl{T2=A>LueRtf+A5S^mhi{!`3%;zC2;EbT8kXC4UY_ -;$maAVJe{XRYw$x#k;6OcNaq(TKHoEaWW?3_HoM&p>K)UtsP)0AT7@VByaj$}6~HqvVvFi74oSmF)p2 -zybYB)GQ|du`aRhm3YU-i^zfsOAj{P1g>w(R7EDMcQUX7*^S_ADi8Q^@#B-65{J+v}(W9Gy=Q?&s||i -aMD>(CS;tF4udbcnREX6k+i494QTk`mfaMi!yXofSbAo=Nav?eV1P_fbLB-t2N(-P6wJu%ZluCvKRdP -g$o<&O0~)bf@!yy?PFLPTB6UrT?<8FMD2ZW=)CCD$EFIU&p+vVuCuPELJ`U)iZs -_Bl%S{HBOZQ~Cj0#cwv~Ylhy}`5BeN#r0aQyf=q$OZe)3n~g4m(rx8;)~CD4l(YY^536`#>jTL$6d?7 -`?}XiqmR}laGVcT~wv|LHbZc4f!s45j;;HWrod=5Br&!Mn`#@0wQF3vKcy$9Xn03K)R -0+tBhfLR5%@HPmj~5j772#Y96rHSNWWi2bF+j(ZP{9+u75a8w)aI+D3)Q1WL*`6`uPu)kqv+FTj -*8zP@qLbPnNwI`Ud82Y%7 -8ksNQ!i#Zc)-Xi9pR2~@qu_lf?gwRM+>Ro~f8HbfIY{*29OYd<&DiZ1-}(M?~0*%MmtK!rsgOoYUZt+ -YF9C`8tZ<_Gu=OVb*KKAbH0rg}G5KBBKrl^!ELXCV_=PLK(AZo*9dV9yv|CC=u@hMw3nm<}2t|8P;c` -@#tezVgwSW=x{yEm{t0%Tl;OO-&iTz*JT&lf#x%Zu!!YaC1ICm&r#iNzwJcf&&npd6C^9JB0*#cw&ez -M8o5eFfKEmQ+(MBv%!?rBka=D&7Xml?vgyi?vEiynq-gu8?+zW0LGTYsCFc!Jc4cm4Z*nhna%^mAVlyvwbZKlab#iPjaCxmedtci~lK=Zt^o -nII1tasGjBXf0f|GDZfHg3?8*qedwXrs`>zRW_Wov!S6|&X84U5^zDi9KXabU@nM{JcT*l^#-P< -&VgKMPj>0+n;zb$DViv{8)C+^JoB^YheF?G&^CI^C^77ogu$bUy10uW*gW)y{qA>?yuKFheI?h2=5{{ -Xtal|d8JLGJVWiz{Vde;ANcyj#l@byX0y?FD!_n~(IQ!x#EO%s6c6dni_MLZ0rFem1PA&bW>oA8t~m{ -rCTP~sP!aOU$QxaU42y#W9qwUhB^``4FyZ8nYlJme0sPn~Y}4us5y_9&iqXLFD@TA$2kJn2e=)_Jf0r -gwVj_In^!!Vz#5gxpMw)oX8ke7OF0WBu2`x2^An)e`bIfPOL$9Dx3Mxc+e{@_%)jzJG -wHSN^uGxTW1GeQ@2Zn?Lux{79{&aKGJy?NP+MDZ7G@?P?9KhfL)Pb{N+Bou8qa((+eRX>L=Ip4)9`VP -~+2!HMNv8$NaD4v9j(dK1i4r5`H7Q}Uv;FfSW1u&DxdSgxF1EMSTcGo8!XL -hizmMO7+5OIU%dp3|2{RK0_a&_k!*}~i3ajO&`4Er5Hv^yyhZS%}m~L;ufa3GL<=%Tq;E6K_XNG3F!k -+l-nQXGT#|+0{7DIA`g3_*{Ojhbx-}b(*n)u2}rS6tdG#nta+f31MDl~WNG_Mz^A(erkx$~Tq*1-b(o -hJG(Hg!50{q3dnb)pB$m~A&Q+gieG>xYU0dhEkUXJwC;7TW6?VG;!Kl&Z+!7#c8L? -sm@{dM>xRvG6uU1Q_<<~HPc^4Zcir%bikino%VYdz5XA)qrz4kzJ7Dmd;9MF@qhl;$%oUk^S}2mF0cO -h$3Oq|(Ho9@KE9g-|Nb0Kqj>gbl4kk+mxuY+jm@p?o!z}(UjAykH|zgxp!M0)jE2G1nDJ=X|Bi#mynb -M-Oqc_(8|(o*64Q?Sz(Qa!g&)sOd+za(MrL(u_i@!iV}=A5WCx_d=Nwl9p<_PT%grgzf2xNv`x!1b21zBN!}lx}O7W~1hB2BZo1{n^E(dw$lxl-dVQ -=4$##|Dg(lWARH996hMUcv0f2Or0fe*s2__H}4Po7&oO -n6B7NO4;r2`h-j`}yGOl~;|~=M#$W#Wx3+1$SnK@9J_au!#PR9dvjV0j*dtpc^a!rnF}}B8hVM(fDj| -e;J2OeA_>K_W{E)HK5T!v;)9swML1wxCsaFh~0Z(PDn({8RyZ+xt(Dw?*hkLo4t-9godD0#l&9Zv -k3*Xz^n+ji6}tLahOOA5pU$Bd_9Ojlr+d7#%11U6em841wnQH?Hh(Mt6j{89$YsK<1B^L1-7>(#$W@$ -c*MWBuGla58HH)5_#wpG$FRE`#H?ar2xVUw!dzo<$g3ez2C^`wk`ZlKt`2^JG0`EFMR7(y>@@R|EX7p -AgngnxE4FCeyJ!ROhfYf_K#&qK4FKHs)Gu&?3!AYY9*ZxX|Y3!>yW -NGBw9ykq*ExNARyQ5)hvmDRW>hH5J}+9Hd7vswW-6>zlCi(goN>nm-y*AnilYX!ypW@xya&RH`2DWA? -Iu3)(b|Qg~4YI+s?~ubsF@E0uviqE)yr=FoQIMjUF^ebSM*Mu_!?_3Uje>;VjV1@zB9Qg*A{On**E)4 -Q2(4*{Uofz7zzOE+vv#gMDy-`=CT{iOYwP`7_al9(2 -6vSACsM0sk!4zumlsNTG_}tpSayQ6cJim#fdyX{)0D%dBG4T+kRM6!Fl#AAkrkFxrANr6>!^VLn2|1Y -AB+es$9hPt=Hkh=1wr>0`CwxJ2uypW>O}#lS-H;El3yE^MTS7LqI>fd$blhfLoGYgf>z^Xaf*p0O -ZPVj|H$^8idUoGLaHq|sQK+`hBP6{H0@QQ~03(2G9Py1_^ovWW-Q)&ZfX@jdq~yVOkY!JE#oqTo2K%2 -NLR`|s=uc5fk=*~X#+)*RyaE+mKFoq9poZCiCj8`ZCmE=8{6Dpu_dYv(CX&_*~j@P)8ZeH_G{f)wES< -wmDMz_xAH?&4uXKsqM)po~)6S}ItL(<{ZsdAUIp4~X1KL91~q{-l+Z=0UR+9Le&DDV%%W~l3BHBE=?q -+_I}RXLHo=wqJPfAkLgWZ1$Hzz!`~deIhi%$hyMo<2h}9zI$?Zfg0m8^2!(S+ -rBc5&}vvu2OOdXtTbM6x@$QnPDhI)Dd&)Ww7sj4MPZXJSHU)$qqv5jF1GXnMU$Znq1|Y1p#I8G<%0;M -^{mM9LL?EmmCbeuV_9Fy8D`uqt`gBLFh*CIckumEK<{7w~J~qzP=cN-A^I3iMW@9GWv<~>5wOd;!_(i -VBS9WLW=l9FS+A&sLFCrgUAj12e3BCtr#LueGuvSScVWpw0M?~Q+pglzUfyE8NT|BZB#e;D8>vV=auk -BNdaqy9MK&MWci{>qLdl_*4`-5eZirbx*Dp8rw$DyYqD0~YU$(Gjzk;4_P$mm*4QRMT;WgVEV&3&> -+m1A|N=iA-wSCY*r!TmjtRi0u)w~mMW|1sDKa}DiQXvqhmvAhkd4uKV;>McUh~&QUl#)Y{xB?pL})d6!IA(n@1=nIlk3`Wb>eY{0Zv+YHwLA~x*$49QYpxFj$|lLojZ -QXp*!B@b8e$6|chYq8UO7?mkh=uL;dCoGyHzU(^)IA`Vt#3M6Iz^sm@S;Ub#gOJ`Ol>q&6*@S$i+%K? -eLFq|+2ZkGi7ek!IAayHJf^zqpdTCfUXplom5rl?2+%?#UUbLJjSfWAky# -M6H17nl4C+_zCGaR90(g*O5ybKfHv=Aoss`uO*J4}jZ*03Dgm6rg{5pW}Hozk}sKB@o$N?PuHSuS~N| -_BBLhmDJ!$eUHXfvwGa6$Grwz>7Pz_5Kzi(;WBDkr?mgsax9uG}@D-Db176wQXi#=!=oyr2OrW1C7y> -w3!>sQqHuU&o++bF0{OmQ^uyb<%QDYP(p<#EDG8#X!mRQIG|~AR#){*)gI40vK?k>B%TbB(!q3OPaXSioCwZ+jd(8AtN}L2o -xFYN-k{4)Zuq^&P7Cz-BaTV~IO0`5s1d$_9kd!b(U4ve*z%l&K`;2tMVYB%qvHvAIHx?O%}XB42@E3p -%C*wKOS?aV=m@SnbniAS40u!VGutTgo+yXCzGMY0>@)2*WFtBoa>WZy3eS{-wlN3{5gLnTGKLlEc0NdhiVo!qcJ> -J>O>--B3NFlyF7G>v$Xh#A8_k88?o!W!kMK0V^q{0CTXZWDU!1)y0*IUdXD-Jzd$o5@3aqDC)s3rg5g -M0#P<8Y0EUQQX~z)DZsR&5my)xR}6#5OXhAipYa7NN@-*z_`tjmN3Ck1O(we$Yr3>ir3sE9nBDII*-B -H#Cf6-#SK!AArAkMx=Q)rIQ%?k2g>Rk-1qXgddzWgyT%q3a$y;0pvle5gd3+jLOjk^~uAl=%Mk$(Pkm5 -oaNK?wlipi=3&ZzhR)^^IwhoV_05@vgzD59S%9c8aA9+$$X6OxhXvr)Y9`U%5bTXs>yEmxo$v;~GgO!4(pO+YrT -n1ICi5_@2{PT*BUUzSM=p{2Vj>(|QCU{{Kp8S)@X?jl`H<>8f#_LtZnL9uLU3nay%8Aq1IwKln>4$>3 -FS1Cx*8!Jn%Qr!YTDz@xo!2b)yPgUZ~u@PFhK*I2uL!Ef%*1k%@TkRgK&0d4v*ZFuXs?tKA5Xggj7ZJ -8NJwb#CRmd}O57lOG10MQfLbD{k576yXp9TVSoS|^B)U+ZB2~J=2$H}ND?@vX6Fz~RJBayhm3FEt9Bs -%mw#E?NXz*5>3Ll5UR;of}^Lg*V2I`sm5j3xt-Qr+IlEbaDI9(T#e!2n{4M-wOIkcrS{OgY$6#R(E{% -CJ>dZ}fuc5luMY=-jGa@9onC+@ESxVUtPT5m3D*x(d)PaoQd2DuB2G(3W&M*k`hqaW6?YFE^Tb3YyC} -zTD_&R}P-!)g8z2!O%;dWg}-%n~*4Fi`xuX(AMXs6$Wvu;xcR5S+B@6v;&9CW!#;{Ian%JcC=*IO_D0 -E8mmCDJdyYjI7XK)J``P#ir>)tY;h7~^W_aTaXVQWG-On@X80*H71^>cI%BGCT)joHERAT$IDYI_hK^ -TEsfsE(imq5sm*Rl^64p+z_YvUnHjP(?RYR4)IG37>D%+H3p!K_5T0BN98+4~hE$_Byx~#`=QkG^nAm -&)TNLAw`h`899dK3}FB?5m*=t1ki6~&HKy+-Y68-NYhNT>oMOX~63p>%wM*(yeZtY18%b?Tz -18>gJVs+SI2Jz}J&&2PB=4>?zZCLNpc3PEg;0k*yYgDQCnL4k9^JELRq#z7{V7;LnM0fne^K96P9OjcRx8Ltsah|9C1Gl%tQ`-lqQ7i`XgAj5=vpO;B$>wclSEt`Trv|rCR)YZ -*7;|U`mVame}E=|Z6;9!@D-GU--i%XI#&rVD$Xi<2H%a0AR#cGyJX)#Kxchg$+Ni -2m2`O13&iU#n46tuP%A5g-wLy8=9!hljK^`U=X@WJ~!2Wfp4?p*2+IAASYz|`MU>c4OslP=JdL*4>Ov -P_FtGj#wOu^KLF>Og8x!l`r&@&{pF>q)~=ab1G0($??GGF*7>`PDw_ta7W0OuMGxnJr@kyeugH8QIKM -cO~(e{=we=08mQ<1QY-O00;of09jjB)K=BM2LJ#Y6#xJr0001RX>c!Jc4cm4Z*nhna%^mAVlyvwbZKl -acVTICE^v9(SX*x!#}$6(ub5H~qFjnrYXkvYvw;gZsF74MYz2M@3}U!@mYi~SXEQUa6@mWvd}l7pB}F -;zTT9sDWzPBL+%K!@)>tpd+Qv*&{NhMc;fHT~H>s?tMb$NhH%2>=PFiWbl6o@ncK>=J@CtNZ)s#tU(- -?0WRZQrypw^4`;rDme8v6`O_$=Lsb+TGXtye3t6u(Yy&1bp4#~#8z#Wl5GtK -lE04ntqu~CWV&HiDzME?}DJpG*?l;7UW>N5dSp2a9h*qQuHJkS(rw+!0LXOT!7?$OBTZ8zweFk`<9$A -m3aH%y%T%Wi9$A_b0i>908q_lgz;MVEzV?9?15FAk}S@=^i*{^f@rlPL9NmH5^!+4TZn&jo~>&dy%+RAY$iy9yRyBdaIl)DB!`gr7np&?_@!ws{lIl3VbppK&S -0<;Q?U_b;X8y{H2>Hh3-QTBuSZ-~AR2XcCE!f9?Cfe0g5cKlbFoFRl!2w#8UyMz`JSjbaIvtF?FVt!u -#lzVu#;N+(x^vr^TBbT-XbHQI;ffrFd-zzPGOn{NCz%x1WVL%hlqoroMs^)SS6Ka%V=*l1g-5n1m9L| -w%1KpuMw}Ms1)p3k}Irgy|FIFNNW_7h1sH%LIs~!>JfN+TQDSUPSUcjt)+^NtpU|OtEMzPD11l{;4~( -K&L9i~*Cd@6ANGC=z|xRoC$`o+sFI_vx3X=?0*-}O{tCg?`5?um6L&uyV$Y1O8;q9$LoEBPa%|aWP5C -#`YNW$qwDBe(1>63Xqx`*{1heOOrj@^A9^d*hcM3gKeK0_<*3@e2D#hL>Y#vQK2^@SYh*<{Yt#6 -h*TW`{6>8EWM?6&90;s<=uoaHWI%E7ep+{Cz2Aujb;rRK+gDzhXU!HlD|LvE;2n#DJUa;KX{5IH+W&e -c-}pv0|eAD&TrMtC$z*W;&ms`30H|;8iorXOju*oiD#b(PXnE>7x*C@!wsds{A7x)wc0i|=y -WE@&W0r%UnI#rH=@sj1Eq-ZASHcTsvO-G#WMg|r*@>&=%-L>fvpDk0a96zI5dZTxnSk{ve&eu4L!U?8a+Jvn*mpTWw>?aPD5uE!ob%QOm?H{#|z@G0mCY|cb`7(~sb^YuD8R0zL -RI@iMFuQ)fN(ztf4-hSr_8d&mQcMgn|%r2`;+e~vX5>L1PiX{O&Jut%7iC!_RmmWh|fGll!gnpgoe%a -TH`d;@dk#RBRpIfUOm+)aB9Pi)pbR|4S&5)IJLBaE6f9NpkNu~TGL|Ow@kO$&J&6o!2|#*6B-&XQf^9 -lf|Pn1N_{KQyoBK-=8H;Ip-U1^L_i2SW*LG35?UU#hJ1u-aYR6ap2aB+xv);5gTBo(W5=x={Vu8`>x_ -oBhh7owS2b`?n%-wAv*X)5d1sz%pWl3I>&CQ{>rf3+XO@Napp>h&d>Fz^p<$!6r -GXe908`507c(LpGQ<67&{Q2!~+Tx4?!KCKx& -wp|nbk{1&G$p3xmw4rh45iK`+M&zT$74Ta<+#Ur#RI||cc9$`&WsZD2N6X$9Xrz|wc9qn7`3(oz77WQ -pa4J*Us7Z*yNd^P!(8;cbMAL3wkG~0YN@c4D&tRwHv#Z4s41It^F!4o4qrMV@I8c!$6klvB?XgWy8oP -Nn8-{$`duE!;5xsQt(7hcSE#g-pGXsRjCLdoh}8g$&tm38UqL{FDC*HxaIsSIVl2+szv^{KFyj-F@ab -p=krB2Z)bio05>`DXA=iL%2pJm(b04wsz`SMy%+V+XkZ#H&ZqBrD{6LJBe_86aQyb)(H1UGT(Gxd=TU -()>DM;s$c2OeaqP4GXJ;lV`cFOyh^?S0T?2K?nD~dVLy8w`&^M2SDZhJpVKh -o@}BDJndUQ{PPw|o?#-m3!yD|Whhh>RXKd61?y-E^v8JO928D0~7!N00;of09jja_pjHy0RRA20{{RI000000 -00000001_fh7R|0B~t=FJE76VQFq(UoLQYP)h*<6ay3h000O8%K%whB^=tD)dBzjss#W56#xJL00000 -00000q=60t003}la4%n9aA|NYa&>NQWpZC%E^v8JO928D0~7!N00;of09jkj$G|nf0000U0RR9D0000 -0000000001_fffb;0B~t=FK~G-ba`-PWKc^10u%!j000080LuVbTMcdTY&Zb`0RI6102u%P00000000 -000PTSQ2LJ$YX>c!JX>N37a&BR4FJE72ZfSI1UoLQYP)h*<6ay3h000O8%K%whgjc!iDG&evc{cz68U -O$Q0000000000q=9b;003}la4%nJZggdGZeeUMVs&Y3WM5@&b}n#vP)h*<6ay3h000O8%K%wh#Uk5%2 -MhoJX(#{y761SM0000000000q=C>F003}la4%nJZggdGZeeUMV_{=xWiD`eP)h*<6ay3h000O8%K%wh -@S@g#N)Z46=r{lX9smFU0000000000q=5@1003}la4%nJZggdGZeeUMV{dL|X=inEVRUJ4ZZ2?nP)h* -<6ay3h000O8%K%whP|GebiXi|1m~#LC8vpc!JX>N37a& -BR4FKuCIZZ2?nP)h*<6ay3h000O8%K%whkG}tUb_M_d%Mkzo8UO$Q0000000000q=9f*003}la4%nJZ -ggdGZeeUMaCvZYZ)#;@bS`jtP)h*<6ay3h000O8%K%whyO09K8w&sc7a;%uBme*a0000000000q=6n_ -003}la4%nJZggdGZeeUMb7gF1UvG7EWMOn=WM5-wWn*hDaCuNm0Rj{Q6aWAK2ms3fSzEm6vGX7b001& -00015U0000000000005+cduspyaA|NaUukZ1WpZv|Y%h0cWo2w%Vs&Y3WMy(LaCuNm0Rj{Q6aWAK2ms -3fSzD|mDhX%+004sk0012T0000000000005+c*L45@aA|NaUukZ1WpZv|Y%gPMX)j-2X>MtBUtcb8c~ -DCM0u%!j000080LuVbTgO#3@r?)o02v$r03!eZ00000000000HlF>b^riyX>c!JX>N37a&BR4FJo+JF -JX0bZ)0z5aBO9CX>V>WaCuNm0Rj{Q6aWAK2ms3fSzFom@aG2$000ak001EX0000000000005+cNPhqT -aA|NaUukZ1WpZv|Y%gPMX)j`7b7fy+Z*6U1Ze%WSc~DCM0u%!j000080LuVbTMa&|9YP@h05N$003QG -V00000000000HlGAiU0s`X>c!JX>N37a&BR4FJo+JFJoMd?cwb|0ZEaz0WG--dP)h*<6ay3h000O8%K%wh-O;>)6a@eP{|W#A82|tP0000000000q=9&~ -003}la4%nJZggdGZeeUMV{Bc!JX>N37a&BR4FJo+JFK}UUb7gWaaCuNm0Rj{Q6aWAK2m -s3fSzGJuA!B|8002A^001HY0000000000005+coz4INaA|NaUukZ1WpZv|Y%gPMX)kbcZ)b94b8}x}V -RCaWaCuNm0Rj{Q6aWAK2ms3fSzDNCi2n2q008GT001BW0000000000005+cThssmaA|NaUukZ1WpZv| -Y%gPMX)khRabII^ZEaz0WG--dP)h*<6ay3h000O8%K%whvI1y49ti*d!W;kq9RL6T0000000000q=Aj -%003}la4%nJZggdGZeeUMV{BbYEXCa -CuNm0Rj{Q6aWAK2ms3fSzCJ{NTMhS001T+0018V0000000000005+cbMXKGaA|NaUukZ1WpZv|Y%gPP -ZEaz0WOFZLVPj}zE^v8JO928D0~7!N00;of09jl7duma@0ssKS2mk;d00000000000001_f!O*00B~t -=FJEbHbY*gGVQepBZ*6U1Ze(*WV`yb#Yc6nkP)h*<6ay3h000O8%K%wh2-05in*{&>vJn6PBme*a000 -0000000q=D7_003}la4%nJZggdGZeeUMV{dJ3VQyq|FJo_QaBO9CX>V>WaCuNm0Rj{Q6aWAK2ms3fSz -Bi86XIeE004O>001Wd0000000000005+cvI7ACaA|NaUukZ1WpZv|Y%gPPZEaz0WOFZLZ*FF3XLWL6b -ZKvHE^v8JO928D0~7!N00;of09jikZx9Ik2mk<;8UO$v00000000000001_fmsg$0B~t=FJEbHbY*gG -VQepBZ*6U1Ze(*WWMyJ?XD)DgP)h*<6ay3h000O8%K%whWDMTSumu1B(-QyyA^-pY0000000000q=At -b0RV7ma4%nJZggdGZeeUMV{dJ3VQyq|FJy0bZftL1WG--dP)h*<6ay3h000O8%K%wh{#IW~@B;t<8w> -ydAOHXW0000000000q=A7R0RV7ma4%nJZggdGZeeUMV{dJ3VQyq|FJ^LOWqM^UaCuNm0Rj{Q6aWAK2m -s3fSz9b^5>VO#0055$0015U0000000000005+cuOa~eaA|NaUukZ1WpZv|Y%gPPZEaz0WOFZQVRL9Ma -CuNm0Rj{Q6aWAK2ms3fSzEL~pz+`V002}30015U0000000000005+c#U%j%aA|NaUukZ1WpZv|Y%gPP -ZEaz0WOFZQWo&RRaCuNm0Rj{Q6aWAK2ms3fSzD@O%Ud@F004Uw0018V0000000000005+c;U@tAaA|N -aUukZ1WpZv|Y%gPPZEaz0WOFZRZe(S6E^v8JO928D0~7!N00;of09jk3koovT1ONbO3;+Nj00000000 -000001_fmJR60B~t=FJEbHbY*gGVQepBZ*6U1Ze(*WX>N0HWn*+MaCuNm0Rj{Q6aWAK2ms3fSzA?UNT -#bE000Pg001EX0000000000005+c*fIeCaA|NaUukZ1WpZv|Y%gPPZEaz0WOFZRZgX^DY-}!Yc~DCM0 -u%!j000080LuVbTWK>c!JX>N37a&BR4FJo_QZDDR? -b1!UZb963nc~DCM0u%!j000080LuVbTbR62h2jPP0HqcH03HAU00000000000HlF>UjYDcX>c!JX>N3 -7a&BR4FJo_QZDDR?b1!UfV{0yOc~DCM0u%!j000080LuVbTXpFM(rO0)0G1X203ZMW00000000000Hl -GKW&r?jX>c!JX>N37a&BR4FJo_QZDDR?b1!pcVRB<=E^v8JO928D0~7!N00;of09jj52c?+22><|s9{ ->Oz00000000000001_fk18n0B~t=FJEbHbY*gGVQepBZ*6U1Ze(*Wb7*gOE^v8JO928D0~7!N00;of0 -9jiw&KKa_1ONaW4*&oo00000000000001_fjD>p0B~t=FJEbHbY*gGVQepBZ*6U1Ze(*Wb#7^Hb97;B -Y%XwlP)h*<6ay3h000O8%K%whIZn|;+6DjseHQ=#9{>OV0000000000q=8m_0RV7ma4%nJZggdGZeeU -MV{dJ3VQyq|FL!8VWo#~Rc~DCM0u%!j000080LuVbThnz7E;|7L09pe804D$d00000000000HlF!gaH -6>X>c!JX>N37a&BR4FJx(RbaH88b#!TOZgVeRUukY>bYEXCaCuNm0Rj{Q6aWAK2ms3fSzD&7p8M4T00 -1Tj001KZ0000000000005+c?S%mVaA|NaUukZ1WpZv|Y%gSKb98cPVs&(BZ*FrhVqtS-E^v8JO928D0 -~7!N00;of09jl1K{v9T0RRA@0{{Rh00000000000001_fdh#F0B~t=FJEbHbY*gGVQepCX>)XPX<~JB -X>V?GFKKRbbYX04Wn?aJc~DCM0u%!j000080LuVbTV1H44yOhH0OJ_|03-ka00000000000HlHCiU9y -{X>c!JX>N37a&BR4FJx(RbaH88b#!TOZgVelWNCABE^v8JO928D0~7!N00;of09jjltDV9+0ssJ11po -jf00000000000001_fzFWu0B~t=FJEbHbY*gGVQepCX>)XPX<~JBX>V?GFL!8VWo#~Rc~DCM0u%!j00 -0080LuVbTX2Jdbsqo#0384T03QGV00000000000HlFRlmP&6X>c!JX>N37a&BR4FKKRMWq2=NUukY>b -YEXCaCuNm0Rj{Q6aWAK2ms3fSz8-dbR&5a0022Z001BW0000000000005+crj!8yaA|NaUukZ1WpZv| -Y%ghUWMz0SV{dG1Wn*-2axQRrP)h*<6ay3h000O8%K%whJUcH~wIl!lvzGt>B>(^b0000000000q=8+ -i0RV7ma4%nJZggdGZeeUMX>Md?crS2aV{2h&WnX4#Ze(S0E^v8JO928D0~7!N00;of09jk#z}3a-2LJ -%SApig#00000000000001_fmO``0B~t=FJEbHbY*gGVQepHZe(S6FLQ5oa${w4E^v8JO928D0~7!N00 -;of09jj}_uI$r5C8yaHvj-500000000000001_fqT^f0B~t=FJEbHbY*gGVQepKZ)0I}X>V?GFJE72Z -fSI1UoLQYP)h*<6ay3h000O8%K%wh@MJ@%-3R~xR~G;PB>(^b0000000000q=BmD0RV7ma4%nJZggdG -ZeeUMY;R*>bZKvHb1z?HX>)XSbZKmJE^v8JO928D0~7!N00;of09jjtrH&%w2><{j9smF(000000000 -00001_fy(az0B~t=FJEbHbY*gGVQepKZ)0I}X>V?GFJE(cb7OCAW@%?GaCuNm0Rj{Q6aWAK2ms3fSz8 -W~Ai#J8008z00018V0000000000005+c@c97%aA|NaUukZ1WpZv|Y%gqYV_|e@Z*FrhVqtS-E^v8JO9 -28D0~7!N00;of09jk1nmd{v2LJ%T761Su00000000000001_fu;Td0B~t=FJEbHbY*gGVQepLWprU=V -RT_HUtei%X>?y-E^v8JO928D0~7!N00;of09jk|a&Ce80{{S*3IG5f00000000000001_fdd8t0B~t= -FJEbHbY*gGVQepLWprU=VRT_HUutu2ZZ2?nP)h*<6ay3h000O8%K%whwVq`iB^m$#Ok)559smFU0000 -000000q=7mL0swGna4%nJZggdGZeeUMZDn*}WMOn+FJfVHWiD`eP)h*<6ay3h000O8%K%whnQSPd&kF -zmFev~4CjbBd0000000000q=B0x0swGna4%nJZggdGZeeUMZDn*}WMOn+FK}yTUvg!0Z*_8GWpgfYc~ -DCM0u%!j000080LuVbTX{ifu3-QG0EYko04x9i00000000000HlGhFaiK@X>c!JX>N37a&BR4FKuOXV -Ps)+VJ~TIaBp&SY-wUIUtei%X>?y-E^v8JO928D0~7!N00;of09jjpU#A6o1ONc^3IG5r0000000000 -0001_fm$&F0B~t=FJEbHbY*gGVQepLWprU=VRT_HX>D+Ca&&BIVlQ7~Z*6d4bS`jtP)h*<6ay3h000O -8%K%who`0Lp5DEYQ<{{{d6)f4~#DF6Tf0000000000q=9xo0swGna4 -%nJZggdGZeeUMZDn*}WMOn+FKKOXZ*p{OX<{#5Wo~wJE^v8JO928D0~7!N00;of09jj~Qj4QH0000!0 -000W00000000000001_fxJfo0B~t=FJEbHbY*gGVQepLZ)9a`b1z?CX>MtBUtcb8c~DCM0u%!j00008 -0LuVbTg%^TJ4XQk0Pz9<03iSX00000000000HlF5NCE(GX>c!JX>N37a&BR4FKusRWo&aVV_|M&X=Gt -^WiD`eP)h*<6ay3h000O8%K%whn~;%qlm`F+n;8HAA^-pY0000000000q=C9g0swGna4%nJZggdGZee -UMZEs{{Y;!MUX>w&_bYFFHY%XwlP)h*<6ay3h000O8%K%wh|0GbUQUd@0rwIT6CIA2c0000000000q= -Al60swGna4%nJZggdGZeeUMZEs{{Y;!MWZ*py6bYEj{Zgg^QY%XwlP)h*<6ay3h000O8%K%wh-MySaCuNm0Rj{Q6aWAK2ms3fSzFdSe=5`k002W60015U0000000 -000005+c$8`b#aA|NaUukZ1WpZv|Y%gtZWMyn~FK~HmZ)0mNaCuNm0Rj{Q6aWAK2ms3fSzF)amiI~l0 -02J%0015U0000000000005+c*Lwm0aA|NaUukZ1WpZv|Y%gtZWMyn~FLPsPWo>0HaCuNm0Rj{Q6aWAK -2ms3fSz8L>y3AMx005g2001Na0000000000005+cU3~%oaA|NaUukZ1WpZv|Y%gtZWMyn~FLPyKa${& -;b7OCCWiD`eP)h*<6ay3h000O8%K%wh=H8j>3j+WE;0FKzCjbBd0000000000q=EN?0swGna4%nJZgg -dGZeeUMZEs{{Y;!MjWo%_*bZKvHUvP3|W^*oZc~DCM0u%!j000080LuVbTPB)h_Z|fR0Fw{^03`qb00 -000000000HlFOhXMd_X>c!JX>N37a&BR4FKusRWo&aVbYXI5WprO~d30!RZZ2?nP)h*<6ay3h000O8% -K%whly^$V+5-RpXbS)U9RL6T0000000000q=Bf60swGna4%nJZggdGZeeUMZEs{{Y;!MnXk}$=E^v8J -O928D0~7!N00;of09jjO5aR(c0000n0000X00000000000001_fxVCd0B~t=FJEbHbY*gGVQepMWpsC -Ma%(SNUukY>bYEXCaCuNm0Rj{Q6aWAK2ms3fSzDdJmTb!v007BQ0012T0000000000005+cEs+8MaA| -NaUukZ1WpZv|Y%gwQba!uZYcFASbZ9Pcc~DCM0u%!j000080LuVbTY+Npkb(vP0R9pH03HAU0000000 -0000HlF4rvdc!JX>N37a&BR4FK%UYcW-iQFJob2Xk{*Nc~DCM0u%!j000080LuVbTX&CK@Iem%0 -E#gH03iSX00000000000HlHOtpWgWX>c!JX>N37a&BR4FK%UYcW-iQFJy0bZftL1WG--dP)h*<6ay3h -000O8%K%whQy)_w-3b5y-W>n{BLDyZ0000000000q=9k00swGna4%nJZggdGZeeUMZe?_LZ*prdY+-t -NUw3F_Wo#~Rc~DCM0u%!j000080LuVbTl=Ch;~N+N0Q5=#03ZMW00000000000HlGB#{vLwX>c!JX>N -37a&BR4FK%UYcW-iQFLPycb7^mGE^v8JO928D0~7!N00;of09jjVsPjnB1pom14*&oj000000000000 -01_f#}}?0B~t=FJEbHbY*gGVQepMWpsCMa%(SjbZKmJE^v8JO928D0~7!N00;of09jkpymbr70ssIe2 -LJ#c00000000000001_f%N490B~t=FJEbHbY*gGVQepMWpsCMa%(SmZESLIV=i!cP)h*<6ay3h000O8 -%K%wh000000ssI200000Bme*a0000000000q=EYA0swGna4%nJZggdGZeeUMZ*XODVRUJ4ZgVeRUukY ->bYEXCaCuNm0Rj{Q6aWAK2ms3fSzC(}invJ!000IT001EX0000000000005+cKIsAgaA|NaUukZ1WpZ -v|Y%gzcWpZJ3X>V?GFJowBV{0yOc~DCM0u%!j000080LuVbTjw_%Eolk>0DdL_03rYY00000000000H -lG&@B#pEX>c!JX>N37a&BR4FK=*Va$$67Z*FrhW^!d^dSxzfc~DCM0u%!j000080LuVbTa4r$p_Lr~0 -P=7E03!eZ00000000000HlF$`vL%PX>c!JX>N37a&BR4FK=*Va$$67Z*FrhaB^jEVRB_IaCuNm0Rj{Q -6aWAK2ms3fSz7=A0006200000001fg0000000000005+cMHvGCaA|NaUukZ1WpZv|Y%gzcWpZJ3X>V? -GFJg6RY-BHAUukY>bYEXCaCuNm0Rj{Q6aWAK2ms3fSzD`a?XZpp005&B001ul0000000000005+cj~N -31aA|NaUukZ1WpZv|Y%gzcWpZJ3X>V?GFJg6RY-BHDb!lv5UvzR|V{2t{E^v8JO928D0~7!N00;of09 -jks1zWLQ0ssJw1pojl00000000000001_fodQF0B~t=FJEbHbY*gGVQepNaAk5~bZKvHb1!0bX>4RKZ -Dn*}WMOn+E^v8JO928D0~7!N00;of09jj`2F-DM0ssKz1poju00000000000001_fe#}C0B~t=FJEbH -bY*gGVQepNaAk5~bZKvHb1!0bX>4RKZDn*}WMOn+Uu9%zbYWs_WiD`eP)h*<6ay3h000O8%K%wh@IZ? -F@c{q;a0CDVCjbBd0000000000q=DNd0|0Poa4%nJZggdGZeeUMZ*XODVRUJ4ZgVeUb!lv5FL!8VWo# -~Rc~DCM0u%!j000080LuVbTb_oUv_t{`0LBFX05AXm00000000000HlEtC<6d+X>c!JX>N37a&BR4FK -=*Va$$67Z*FrhVs&Y3WG{DUWo2w%Wn^h|VPb4$E^v8JO928D0~7!N00;of09jkj5U=_)0000o0000i0 -0000000000001_fukw|0B~t=FJEbHbY*gGVQepNaAk5~bZKvHb1!Lbb97;BY%gD5X>MtBUtcb8c~DCM -0u%!j000080LuVbTb8bJ0=ppq05oj?04V?f00000000000HlE-D+2&c!JX>N37a&BR4FK=*Va$$6 -7Z*FrhX>N0LVQg$KcW7m0Y%XwlP)h*<6ay3h000O8%K%wh&&ODO=mP)%;R^r&9RL6T0000000000q=6 -$%0|0Poa4%nJZggdGZeeUMa%FKZUtei%X>?y-E^v8JO928D0~7!N00;of09jjo(zp0=6aWB~NB{sK00 -000000000001_fksgS0B~t=FJEbHbY*gGVQepQWpOWKZ*FsRa&=>LZ*p@kaCuNm0Rj{Q6aWAK2ms3fS -z9jkM;K`X0024)001cf0000000000005+c_GJSAaA|NaUukZ1WpZv|Y%g+UaW8UZabIL*aAj^}Wo~16 -UuSY}b#N|lc~DCM0u%!j000080LuVbTSjL|2AvlG0HRI+0384T00000000000HlGYX#)UoX>c!JX>N3 -7a&BR4FLGsZFLGsZUuJ1+WiD`eP)h*<6ay3h000O8%K%whCl=OM&maH*DS7|^AOHXW0000000000q=9 -~c0|0Poa4%nJZggdGZeeUMa%FKZa%FK}X>N0LVQg$JaCuNm0Rj{Q6aWAK2ms3fSzDA^6;L<>000aN00 -0~S0000000000005+ci=qPnaA|NaUukZ1WpZv|Y%g+UaW8UZabI&~bS`jtP)h*<6ay3h000O8%K%wh* -mO`p@EQOBBVGUiA^-pY0000000000q=EgV0|0Poa4%nJZggdGZeeUMa%FKZa%FK}b#7^Hb97;BY%Xwl -P)h*<6ay3h000O8%K%wh000000ssI200000Bme*a0000000000q=7BL0|0Poa4%nJZggdGZeeUMa%FR -GY;|;LZ*DJNUukY>bYEXCaCuNm0Rj{Q6aWAK2ms3fSzCD#z6ksP002P(001BW0000000000005+calr -!saA|NaUukZ1WpZv|Y%g+Ub8l>QbZKvHFJfVHWiD`eP)h*<6ay3h000O8%K%wh000000ssI200000D* -ylh0000000000q=Bo#0|0Poa4%nJZggdGZeeUMa%FRGY;|;LZ*DJaWoKbyc`sjIX>MtBUtcb8c~DCM0 -u%!j000080LuVbTaCNOGYT320Q_A504o3h00000000000HlHT!UF(sX>c!JX>N37a&BR4FLGsbZ)|mR -X>V>XY-ML*V|g!fWpi(Ac4cxdaCuNm0Rj{Q6aWAK2ms3fSz7=A0006200000001ul0000000000005+ -cN!QbZKvHFLGsbZ)|pDY-wUIUtei%X>?y-E^v8JO928D0~7!N00; -of09jifWzroa1polI6951!00000000000001_ftcL`0B~t=FJEbHbY*gGVQepQWpi(Ab#!TOZZC3Wb8 -l>RWo&6;FJfVHWiD`eP)h*<6ay3h000O8%K%wh$?Y`PQxgCH=1%|sF#rGn0000000000q=5zF0|0Poa -4%nJZggdGZeeUMa%FRGY;|;LZ*DJgWpi(Ac4cg7VlQK1Ze(d>VRU74E^v8JO928D0~7!N00;of09jku -m9w=KApiheg8%?700000000000001_fv5HZ0B~t=FJEbHbY*gGVQepQWpi(Ab#!TOZZC3Wb8l>RWo&6 -;FJ@t5bZ>HbE^v8JO928D0~7!N00;of09ji;=(m661^@tp7XSb^00000000000001_fe07`0B~t=FJE -bHbY*gGVQepQWpi(Ab#!TOZZC3Wb8l>RWo&6;FJ^CbZe(9$VQyq;WMOn=b1rasP)h*<6ay3h000O8%K -%wh1+IQqp$-55u`B=pF8}}l0000000000q=7;o1ORYpa4%nJZggdGZeeUMa%FRGY;|;LZ*DJgWpi(Ac -4cg7VlQxVZ+2;9WpXZXc~DCM0u%!j000080LuVbTlH+=S;PbY07VY~051Rl00000000000HlF1E(8E@ -X>c!JX>N37a&BR4FLGsbZ)|mRX>V>Xa%FRGY<6XAX<{#OWpHnDbY*fbaCuNm0Rj{Q6aWAK2ms3fSzC9 -QbZKvHFLGsbZ)|pDY-w -UIa%FLKX>w(4Wo~qHE^v8JO928D0~7!N00;of09jkY!d`&N5C8yuGyniE00000000000001_f&MuJ0B -~t=FJEbHbY*gGVQepQWpi(Ab#!TOZZC3Wb8l>RWo&6;FLGsbZ)|pDaxQRrP)h*<6ay3h000O8%K%wh0 -00000ssI2000009{>OV0000000000q=6Mn1ORYpa4%nJZggdGZeeUMb#!TLb1z?CX>MtBUtcb8c~DCM -0u%!j000080LuVbTZCzO(bok402U1Z03!eZ00000000000HlFbO9TLLX>c!JX>N37a&BR4FLiWjY;!M -PYGHC=V{cz{Wq5QhaCuNm0Rj{Q6aWAK2ms3fSzAhYIX~6`008#`000{R0000000000005+cY*7RNaA| -NaUukZ1WpZv|Y%g_mX>4;ZUu<{c00000000000001_fk##Z0B~t=FJEbHbY*gGVQepTbZKmJFJo_QaA9;VaCuNm0Rj{Q6aWAK2ms3f -Sz8DkHFfm{008D0001cf0000000000005+cQCkE6aA|NaUukZ1WpZv|Y%g_mX>4;ZV{dJ6VRUI?X>4h -9d0%v4XLBxac~DCM0u%!j000080LuVbTlvktub2P;0Pz3-03QGV00000000000HlG9VgvwiX>c!JX>N -37a&BR4FLiWjY;!MUVRU75X>DaLaCuNm0Rj{Q6aWAK2ms3fSzEup%C}Ml003|f001HY000000000000 -5+cVq*jVaA|NaUukZ1WpZv|Y%g_mX>4;ZWMy!2Wn*DV>WaCuNm0Rj{Q6aWAK2ms3fSz9J3(lhD<0 -04jt001cf0000000000005+c^k@VCaA|NaUukZ1WpZv|Y%g_mX>4;ZWNC6`V{~72a%^8{Wo&R|a&sZ3F;tX>c!JX>N37a&BR4FLiW -jY;!MVXJ=n*X>MySaCuNm0Rj{Q6aWAK2ms3fSzET&Zf!^f008_9001HY0000000000005+c>Tm=AaA| -NaUukZ1WpZv|Y%g_mX>4;ZWo~qGd2nxOZgg`laCuNm0Rj{Q6aWAK2ms3fSzFKDU`|yC004Ov001EX00 -00000000005+ca&`m&aA|NaUukZ1WpZv|Y%g_mX>4;ZW@&6?b9r-gWo<5Sc~DCM0u%!j000080LuVbT -O)$gb4~#O0I>o903ZMW00000000000HlEie*^$c!JX>N37a&BR4FLiWjY;!MWX>4V5d2nTOE^v8J -O928D0~7!N00;of09jl2Qmf$z1^@t#4gdfg00000000000001_fs%j(0B~t=FJEbHbY*gGVQepTbZKm -JFK29NVq-3Fc~DCM0u%!j000080LuVbTTK&QBIO1E0EQC)0384T00000000000HlG`h6Dg`X>c!JX>N -37a&BR4FLiWjY;!MYVRL9@b1rasP)h*<6ay3h000O8%K%whPt9!i1`YrKOfLWc9smFU0000000000q= -EE}1ORYpa4%nJZggdGZeeUMb#!TLb1!UfXJ=_{XD)DgP)h*<6ay3h000O8%K%whU<2TMR2={SPFnx~8 -vpc!JX>N37a&BR4FLiWjY;!MgVPk7yXK8L{E^v8JO928 -D0~7!N00;of09ji`nsm^00ssKD1pojX00000000000001_fyulC0B~t=FJEbHbY*gGVQepTbZKmJFLG -sca(OOrc~DCM0u%!j000080LuVbTaiEhCU6S?02dc!JX>N37a& -BR4FLiWjY;!Mjbz*RGZ)0V1b1rasP)h*<6ay3h000O8%K%whQoPCGtqTAETOOV0000000000q -=6^P1ORYpa4%nJZggdGZeeUMb#!TLb1!sdZE#;?X>u-bc~DCM0u%!j000080LuVbTUDbTByJ7>0Io6s -03ZMW00000000000HlEt*8~7?X>c!JX>N37a&BR4FLiWjY;!MlZg62^YiVw0E^v8JO928D0~7!N00;o -f09jjbkyNwv0ssI(1^@sX00000000000001_fxYDf0B~t=FJEbHbY*gGVQepTbZKmJFLiQkb1rasP)h -*<6ay3h000O8%K%whchf_09R&aYe+>WtApigX0000000000q=D+_1ORYpa4%nJZggdGZeeUMb#!TLb1 -!yja&&cJY-MhCE^v8JO928D0~7!N00;of09jjR$D#y01^@tb5dZ)j00000000000001_fky5G0B~t=F -JEbHbY*gGVQepTbZKmJFL!8VWo#~Rc~DCM0u%!j000080LuVbTQ9BN#V-K>08|110384T0000000000 -0HlGu^aKEKX>c!JX>N37a&BR4FLq;dFJE72ZfSI1UoLQYP)h*<6ay3h000O8%K%whcdOf9%me@cmJR> -_8vp(1ORYpa4%nJZggdGZeeUMc4KodVqtn=VR9~Tc~DCM0u%!j000080LuVbTbu -u)p28LY0LV%J02lxO00000000000HlE{`vd@RX>c!JX>N37a&BR4FLq;dFK20VE^v8JO928D0~7!N00 -;of09jirK51Vp1^@uZ6#xJp00000000000001_fg=(H0B~t=FJEbHbY*gGVQepUV{V?0Z*FvQZ)`4bc~DCM0u%!j000080LuVbTQ5`lfFch70J0c!Jc4cm4Z*nhEML|SOMJ{r4bWlqH0u%!j000080LuVbTLO|0dT9m#04 -oy!02crN00000000000HlHUP6YsPX>c!Jc4cm4Z*nhRUukY>bYEXCaCuNm0Rj{Q6aWAK2ms3fSzCVpf -yUeb00373000&M0000000000005+cl~n}*aA|NaUv_0~WN&gWc4cm4Z*nelcyv%p0Rj{Q6aWAK2ms3f -SzE^3KwV`4001rm001Na0000000000005+crB($1aA|NaUv_0~WN&gWV_{=xWn*t{baHQOFHA{8MNU& -iE_8TwP)h*<6ay3h000O8%K%wh3wIW|Kmh;%r2+r|Bme*a0000000000q=8IW1pshqa4%nWWo~3|axY -_HV`yb#Z*FvQZ)`7LUukY>bYEXCaCuNm0Rj{Q6aWAK2ms3fSzGyUPffW3007Ab001BW000000000000 -5+c&RGQjaA|NaUv_0~WN&gWV_{=xWn*t{baHQOFJEJAWG--dP)h*<6ay3h000O8%K%whbpx+`wg&(J# -~Ac!Jc4cm4Z*nhVVPj}zV{dMBa&K% -eV_{=xWiD`eP)h*<6ay3h000O8%K%wh!`*UJdK3Tvok{=zCIA2c0000000000q=D9F1pshqa4%nWWo~ -3|axY_HV`yb#Z*FvQZ)`7PZ*FvQZ)|L3axQRrP)h*<6ay3h000O8%K%whEP=Pl?gjt=!w>)fCjbBd00 -00000000q=A%t1pshqa4%nWWo~3|axY_HV`yb#Z*FvQZ)`7SX>4V8a$#_AWpXZXc~DCM0u%!j000080 -LuVbTPyCQ&a?&q01*=a044wc00000000000HlG(g9QL^X>c!Jc4cm4Z*nhVVPj}zV{dMBa&K%eXk~SB -X>)XGV{c!Jc4cm -4Z*nhVVPj}zV{dMBa&K%eaCt6td2nT9P)h*<6ay3h000O8%K%wht?6LD?*;$>D--|#B>(^b00000000 -00q=5j71pshqa4%nWWo~3|axY_HV`yb#Z*FvQZ)`7fWpZg@Y-xIBE^v8JO928D0~7!N00;of09jkuJb -z{70RRAr1poje00000000000001_fijW>0B~t=FJE?LZe(wAFJob2Xk}w>Zgg^QY%h0mVQ_F|axQRrP -)h*<6ay3h000O8%K%whCD2a%vj6}9F984mD*ylh0000000000q=8tJ1pshqa4%nWWo~3|axY_HV`yb# -Z*FvQZ)`7PVPj}zb1z?CX>MtBUtcb8c~DCM0u%!j000080LuVbTfq8&kf{X#02L4b04)Fj000000000 -00HlFbl?4EBX>c!Jc4cm4Z*nhVVPj}zV{dMBa&K%eV_{=xWpgiPX>4U*V_{=xWiD`eP)h*<6ay3h000 -O8%K%why!H3DMgjl;Y6SoQE&u=k0000000000q=87A1pshqa4%nWWo~3|axY_HV`yb#Z*FvQZ)`7PVP -j}zb1!mbWNC9>V_{=xWiD`eP)h*<6ay3h000O8%K%wh6ND==J^}y$-2(ss8vpc!Jc4cm4Z*nhVWpZ?BW@#^9UukY>bYEXCaCuNm0Rj{Q6aWAK2ms3fSz9k4t>K; -k009300018V0000000000005+c-=GBmaA|NaUv_0~WN&gWV`Xx5X=Z6JUteuuX>MO%E^v8JO928D0~7 -!N00;of09jk)QUn6dYXSf!as&V#00000000000001_fx4ju0B~t=FJE?LZe(wAFJonLbZKU3FJob2Wp -Z>baAj>!O928D0~7!N00;of09jj|`~%mK0{{SW4FCWe00000000000001_fy5060B~t=FJE?LZe(wAF -JonLbZKU3FJo_VWiD`eP)h*<6ay3h000O8%K%wh000000ssI2000008~^|S0000000000q=AkR2mo+t -a4%nWWo~3|axY_La&&2CX)kbjE_8WtWn@rG0Rj{Q6aWAK2ms3fSzC!J>(+t-002t_001cf000000000 -0005+c$`J?vaA|NaUv_0~WN&gWWMyz=Ze(R{V|ia^a&L8Tb1zIuLq$$gMJ{xBbWlqH0u%!j000080Lu -VbTY4cwo_YWP0Qvv`04V?f00000000000HlGF6bJxtX>c!Jc4cm4Z*nhWWpHI~WMyt+d0%I8Z*_2UFJ -E72ZfSI1UoLQYP)h*<6ay3h000O8%K%whO(cT{paK8@tp)%9DF6Tf0000000000q=8fw2mo+ta4%nWW -o~3|axY|MaAj^}Wo~16UuSY}b#QYpUteuuX>MO%E^v8JO928D0~7!N00;of09jks5C0D02LJ$R9{>O_ -00000000000001_fjbxo0B~t=FJE?LZe(wAFJxtKWo~3;Zew|0XL4_KaC0wTX>D+9Wo>0{bYXO9Z*DG -dc~DCM0u%!j000080LuVbTQ8A0QPlzf0ImiA05|{u00000000000HlFyAP4|(X>c!Jc4cm4Z*nhWWpH -I~WMyt+d0%I8Z*_2UFJEkFZggK{WpHI~WMyt+d0%I8Z*_2UE^v8JO928D0~7!N00;of09jkr0ZIQP0{ -{R?2LJ#q00000000000001_ftDi(0B~t=FJE?LZe(wAFJxtKWo~3;Zew|0XL4_KaC0wTaA|N~cXDBHa -Ak5XaCuNm0Rj{Q6aWAK2ms3fSz8B4&nTz>0012U001oj0000000000005+c1t$mqaA|NaUv_0~WN&gW -WMyz=Ze(R{V|ia^a&L8Tb1z?XZ*6Q}V{dJ6VRSBVc~DCM0u%!j000080LuVbTL1t600IC200000044w -c00000000000HlHVCkOy=X>c!Jc4cm4Z*nhWWpHI~WMyt+d0%I8Z*_2UFK~G-ba`-PWKc^10u%!j000 -080LuVbTdfb&ovaW50K+%{03QGV00000000000HlFFCX>c!Jc4cm4Z*nhWX>)XJX<{!-Nkc_WQ$ -;Rxcyv%p0Rj{Q6aWAK2ms3fSzB?y-E^v8JO928D0~7!N00;of09jjIN#!pyD*yodp#T6K00000000000001_fyp`u -0B~t=FJE?LZe(wAFJx(RbZlv2FJo_QaA9;VaCuNm0Rj{Q6aWAK2ms3fSzDdG)qt=I002ZP001BW0000 -000000005+cH)aR`aA|NaUv_0~WN&gWWNCABY-wUIa%FRGb#h~6b1rasP)h*<6ay3h000O8%K%whA>6 -p<1{VMTv`7E|9smFU0000000000q=6-J2mo+ta4%nWWo~3|axY|Qb98KJVlQ)Ja%pgMb1rasP)h*<6a -y3h000O8%K%whgi3i5>bw8|0Db`g8vpc!Jc4cm4Z*nhW -X>)XJX<{#QHZ(0^a&0bUcx6ya0Rj{Q6aWAK2ms3fSzBCX0~wRd0001{0RS5S0000000000005+c6@>} -_aA|NaUv_0~WN&gWWNCABY-wUIbT%|DWq4&!O928D0~7!N00;of09jkAOLI8vpy3jlCwa4%nWWo~3|axY|Qb98KJVlQ_yGA?C!Wl&220u%!j000080LuVbT -P;Kj+Vw2~005E#03QGV00000000000HlH1P7DBWX>c!Jc4cm4Z*nhWX>)XJX<{#THZ(0^a&0bUcx6ya -0Rj{Q6aWAK2ms3fSzG?_qApIw0001v0RS5S0000000000005+c41Em%aA|NaUv_0~WN&gWWNCABY-wU -IcQ!OGWq4&!O928D0~7!N00;of09jiO<6DGi4*&o?EC2u+00000000000001_fs;ZG0B~t=FJE?LZe( -wAFJx(RbaHPmOi4pUPE$otO928D0~7!N00;of09jl6U~t?c0RRBi0{{RX00000000000001_fiqGL0B -~t=FJE?LZe(wAFJx(RbaHPmUtei%X>?y-E^v8JO928D0~7!N00;of09jidRJ>_C0000$0000U000000 -00000001_fs#`W0B~t=FJE?LZe(wAFJx(RbaHPmUteuuX>MO%E^v8JO928D0~7!N00;of09ji~I&Et% -EdT%(!2kdp00000000000001_feKU(0B~t=FJE?LZe(wAFJx(RbaHPmWNCABa&Inhc~DCM0u%!j0000 -80LuVbTL1t600IC20000002=@R00000000000HlF(ferw0X>c!Jc4cm4Z*nhWX>)XPZ!d6pE_8WtWn@ -rG0Rj{Q6aWAK2ms3fSzE*6oA4?F000FB000{R0000000000005+ct$_{zaA|NaUv_0~WN&gWX=H9;FH -A{8MNU&iE^TB`O928D0~7!N00;of09jiD7xB9v0RRAG0{{RV00000000000001_fe?ib0B~t=FJE?LZ -e(wAFKJ|MVJ}}_X>MtBUtcb8c~DCM0u%!j000080LuVbTOgZYPdWnt0A39M02u%P00000000000HlFw -h7JI5X>c!Jc4cm4Z*nhbWNu+EV{c?-V=i!cP)h*<6ay3h000O8%K%why<-UJ#{d8TJOKaz8UO$Q0000 -000000q=DLr4ghdza4%nWWo~3|axZCQZecHDZ*6d4bS`jtP)h*<6ay3h000O8%K%whzBs-g_6q<2w=) -0$7ytkO0000000000q=DIr4ghdza4%nWWo~3|axZCQZecHDZ*pZWaCuNm0Rj{Q6aWAK2ms3fSz89D06 -Yvj0081K0RS8T0000000000005+c1(yy0aA|NaUv_0~WN&gWX=H9;FKJ|MVPs)+VJ>iaP)h*<6ay3h0 -00O8%K%whI$=a+B?ABeY6k!S9RL6T0000000000q=8G(4ghdza4%nWWo~3|axZCQZecHJZgg^CZf9k4 -E^v8JO928D0~7!N00;of09jlAIQrrj0000L0000W00000000000001_fveOG0B~t=FJE?LZe(wAFKJ| -MVJ~oDV{2h&WnW}rbYU)Vc~DCM0u%!j000080LuVbTL1t600IC20000002u%P00000000000HlEc)eZ -n~X>c!Jc4cm4Z*nhbWNu+EaCt6td2nT9P)h*<6ay3h000O8%K%whWnju4#Hj!PxuydE9RL6T0000000 -000q=7lr4ghdza4%nWWo~3|axZCQZecHVbaON|WMOn+E^v8JO928D0~7!N00;of09jjDwJ17>0RRAI0 -ssIT00000000000001_fjEB;0B~t=FJE?LZe(wAFKu&YaA9L>FGEjISxHVuP)h*<6ay3h000O8%K%wh -DhN0*;sF2vRRjP49{>OV0000000000q=ET>4*+m!a4%nWWo~3|axZOjXK-O-YcF44X>MtBUtcb8c~DC -M0u%!j000080LuVbTRS_Q)2#sj067Ez03iSX00000000000HlE!gAV|3X>c!Jc4cm4Z*nhfb7yd2V{0 -#Ecw=R7bZKvHb1rasP)h*<6ay3h000O8%K%whCE!Q*c?JLgUKRiV8UO$Q0000000000q=5v54*+m!a4 -%nWWo~3|axZOjXK-O-YcFMZbS`jtP)h*<6ay3h000O8%K%whBpktIHX8r{hJFA59{>OV0000000000q -=B`I4*+m!a4%nWWo~3|axZOjXK-O-YcFPDY;0m-V{0yOc~DCM0u%!j000080LuVbTi@=|yng@y0L1_R -0384T00000000000HlE^rw;&dX>c!Jc4cm4Z*nhiVPk7yXK8L{FHA{8MNU&iP)h*<6ay3h000O8%K%w -hu(zWg=?wq?z9#?xBme*a0000000000q=Dk64*+m!a4%nWWo~3|axZXUV{2h&X>MmPOi4pUPE$oLK~O -VP|P>XD>`iLq$$gMJ_^9L{Lis0u%!j000080LuVbTYvn?TrmLv0PX<*03iSX00000000000HlEjx -eowvX>c!Jc4cm4Z*nhiVPk7yXK8L{FJE72ZfSI1UoLQYP)h*<6ay3h000O8%K%wh4)5auMmPUuA4&W@&6?E^v8JO928D0~7!N00;o -f09jj@VL-t*4FCXqB>(^;00000000000001_fttS$0B~t=FJE?LZe(wAFK}UFYhh<;Zf7rFZDDSCY-w -(FcrI{xP)h*<6ay3h000O8%K%wh>DV)-!2|#Rh6(@xBLDyZ0000000000q=6F64*+m!a4%nWWo~3|ax -ZXUV{2h&X>MmPUu|`BY;0+6b$Bjtc~DCM0u%!j000080LuVbTg-yz;4ujR0PQCL03ZMW00000000000 -HlEw(hmS|X>c!Jc4cm4Z*nhiVPk7yXK8L{FJEwBa&u*JE^v8JO928D0~7!N00;of09jiwqw-46h8X>c!Jc4cm4Z*nhiVPk7yXK8L{FJE -+TYh`X}dS!AhaCuNm0Rj{Q6aWAK2ms3fSzC&M1n>e6000p$001BW0000000000005+co8%7waA|NaUv -_0~WN&gWaA9L>VP|P>XD@AGa%*LBb1rasP)h*<6ay3h000O8%K%whp_MniS0w-dvxfixApigX000000 -0000q=DP>4*+m!a4%nWWo~3|axZXUV{2h&X>MmPZDn*}WMOn+E^v8JO928D0~7!N00;of09jiA00002 -000000000U00000000000001_fpHlS0B~t=FJE?LZe(wAFK}UFYhh<;Zf7rWc`kH$aAjmrO928D0~7! -N00;of09jk=jVik>1ONbo3jhEm00000000000001_fv*`50B~t=FJE?LZe(wAFK}UFYhh<;Zf7rYWpQ -~Uox0057i001KZ0000000000005+cA07|@aA|NaUv -_0~WN&gWaA9L>VP|P>XD@SbWn*b(X=QSAE^v8JO928D0~7!N00;of09jkTktBWy8UO&VP|P>XD@YhX>4;YaCuNm0Rj{Q6aWA -K2ms3fSzCV>$`Wf4001yS001BW0000000000005+c)?p9;aA|NaUv_0~WN&gWaA9L>VP|P>XD@bTa&u -{KZZ2?nP)h*<6ay3h000O8%K%wh9o!vi{RaR5UlsrWDgXcg0000000000q=A2P5CCv#a4%nWWo~3|ax -ZXUV{2h&X>MmPY-wXMmPY-wXc!Jc4cm4Z*nhiYiD0_Wpi(Ja${w4FHA{ -8MNU&iP)h*<6ay3h000O8%K%whc~f!DR)PQkBjy1BB>(^b0000000000q=9Um5CCv#a4%nWWo~3|axZ -XeXJ2w?y-E^v8JO928D0~7!N00;of09jkC)*D@e0ssI(1ONaa000000000000 -01_fdwED0B~t=FJE?LZe(wAFK}#ObY^dIZDeV3b1zIuLq$$gMNmrt0u%!j000080LuVbTUKc!Jc4cm4Z*nhiY+-a}Z*py9X>xNfUtei%X>?y-E^v8JO928D -0~7!N00;of09jj;(H^6_0RRBu1pojf00000000000001_fvPDH0B~t=FJE?LZe(wAFK}#ObY^dIZDeV -3b1z?CZDDC{Utcb8c~DCM0u%!j000080LuVbTbgeR6F3I|05u~303!eZ00000000000HlGZD-i&2X>c -!Jc4cm4Z*nhiY+-a}Z*py9X>xNfVQyq{Z)s#MaCuNm0Rj{Q6aWAK2ms3fSzBUzjPzy)002QG0018V00 -00000000005+cATtpFaA|NaUv_0~WN&gWaBN|8W^ZzBWNC79FJW+LE^v8JO928D0~7!N00;of09jkHl -^i0`1ONcC7ytkv00000000000001_fx|ix0B~t=FJE?LZe(wAFK}#ObY^dIZDeV3b1!XSV{daVaCuNm -0Rj{Q6aWAK2ms3fSz7=A0006200000001EX0000000000005+c(?1aaaA|NaUv_0~WN&gWaBN|8W^Zz -BWNC79FK~G-ba`-PWKc^10u%!j000080LuVbTSm-MMR^GT0NN-303ZMW00000000000HlEwKoJ0NX>c -!Jc4cm4Z*nhiY+-a}Z*py9X>xNfb#7^RE^v8JO928D0~7!N00;of09jiL?Cz940RRBN0ssIb0000000 -0000001_fy+q|0B~t=FJE?LZe(wAFK}#ObY^dIZDeV3b1!yfa&u{KZZ2?nP)h*<6ay3h000O8%K%wht -(cJU%?SVijVAyABLDyZ0000000000q=8CH5dd&$a4%nWWo~3|axZXfVRUA1a&2U3a&s?tX>Me1cXKXq -c~DCM0u%!j000080LuVbTMGam%fSKw05b&u02}}S00000000000HlFfRS^JiX>c!Jc4cm4Z*nhid1q~ -9Zgg`mOi4pUPE$otO928D0~7!N00;of09jj&l9e3u1ONc13jhEh00000000000001_fm2u!0B~t=FJE -?LZe(wAFK~HhZDnqBb1z?CX>MtBUtcb8c~DCM0u%!j000080LuVbTW+1i5b6K`0AT?D03ZMW0000000 -0000HlG0T@e6qX>c!Jc4cm4Z*nhid1q~9Zgg`mUteuuX>MO%E^v8JO928D0~7!N00;of09jkdV|*r+0 -ssKE1^@sb00000000000001_fvjE;0B~t=FJE?LZe(wAFK~HhZDnqBb1!3WZgX#JWiD`eP)h*<6ay3h -000O8%K%whcJWzR`~m;~b_W0e9smFU0000000000q=9~65dd&$a4%nWWo~3|axZXsXKiI}baO9eX>4? -5axQRrP)h*<6ay3h000O8%K%whWi@i;;|2f#CJ_JtApigX0000000000q=B|%5dd&$a4%nWWo~3|axZ -XsXKiI}baO9eZ*py6baZ8ME^v8JO928D0~7!N00;of09jjC3Wi*XB>(^wiU0r|00000000000001_f! -J#i0B~t=FJE?LZe(wAFK~HhZDnqBb1!UVcx7@faCuNm0Rj{Q6aWAK2ms3fSzG0}oo(;|008X+001BW0 -000000000005+coRJX#aA|NaUv_0~WN&gWaCv8KWo~qHFKusRWo&6~WiD`eP)h*<6ay3h000O8%K%wh -e-4dud;$OfV+Q~L9smFU0000000000q=Cwl5dd&$a4%nWWo~3|axZXsXKiI}baO9oY;|X8ZZ2?nP)h* -<6ay3h000O8%K%whjivwtO#}b{01N;CAOHXW0000000000q=AB#5dd&$a4%nWWo~3|axZXsXKiI}baO -9qWoKo0Z*X)jaCuNm0Rj{Q6aWAK2ms3fSzC+35Vn8>000yW0018V0000000000005+c4Vw`FaA|NaUv -_0~WN&gWaCv8KWo~qHFLPsIZf<3AE^v8JO928D0~7!N00;of09jkkOAWKl2mk;r9{>O$00000000000 -001_fy$o|0B~t=FJE?LZe(wAFK~HhZDnqBb1!pnXlZVEWq5QhaCuNm0Rj{Q6aWAK2ms3fSzGOAcZ2!{ -000yj0012T0000000000005+c)~FExaA|NaUv_0~WN&gWaCv8KWo~qHFLQKxY-KKRc~DCM0u%!j0000 -80LuVbTU8^;g@FeE08$tL0384T00000000000HlEnun_=oX>c!Jc4cm4Z*nhid1q~9Zgg`mbZ={AZZ2 -?nP)h*<6ay3h000O8%K%wh4;A&2ZVc~DCM0u%!j000080LuVbTQuD)n71DQ0LYyH04M+e00000 -000000HlF$5)uG#X>c!Jc4cm4Z*nhid1q~9Zgg`mW@&76WpZ;bUtei%X>?y-E^v8JO928D0~7!N00;o -f09ji#+CPxa1^@sF6#xJ$00000000000001_fow7o0B~t=FJE?LZe(wAFK~HhZDnqBb1!CZa&2LBbY* -gLFJE72ZfSI1UoLQYP)h*<6ay3h000O8%K%whj*9)vAO!#bP!IqBD*ylh0000000000q=ADu5&&>%a4 -%nWWo~3|axZXsXKiI}baO9eZ*py6baZ8Mb1z?QVQ_G1Zf7oVc~DCM0u%!j000080LuVbTL6MwEk_Li0 -8=jj04D$d00000000000HlHJJ`wc!Jc4cm4Z*nhid1q~9Zgg`mY-M<5a&s?VUukY>bYEXCaCuNm -0Rj{Q6aWAK2ms3fSzCw9=Kmr;001H{0RSfe0000000000005+ccuW!iaA|NaUv_0~WN&gWaCv8KWo~q -HFKlIaWpZ;bUu|J?y-E^v8JO -928D0~7!N00;of09jjm{_Gta0{{T<3;+Np00000000000001_fjzqt0B~t=FJE?LZe(wAFK~HhZDnqB -b1!prd2D5KFJEn8aByjEXD)DgP)h*<6ay3h000O8%K%whM~lQneF6XgIRpRzBLDyZ0000000000q=B8 -j5&&>%a4%nWWo~3|axZXsaB^>IWn*+-Xm4+8b1zIuLq$$gMNmrt0u%!j000080LuVbTbLREok0Nr0J8 -!B04D$d00000000000HlFi!4d#)X>c!Jc4cm4Z*nhid2n)XYGq?|UubV{YjZDOUukY>bYEXCaCuNm0R -j{Q6aWAK2ms3fSz8}vF|INU003Az001Na0000000000005+c;ldICaA|NaUv_0~WN&gWaCvZYZ)#;@b -YEz1Z)5&&>% -a4%nWWo~3|axZXsaB^>IWn*+-Xm4+8b1!gtE_8WtWn@rG0Rj{Q6aWAK2ms3fSzEGZzdsiN001ol001) -p0000000000005+cna>gcaA|NaUv_0~WN&gWaCvZYZ)#;@bYEz1Z)BWpi^cUtei%X>? -y-E^v8JO928D0~7!N00;of09jiwC3Alr4FCYRF8}~G00000000000001_fdSDH0B~t=FJE?LZe(wAFK -~Hqa&Ky7V{~6=Z*OaJFJEbHUvP47V`X!5FJEbHUvP47V`X!5E^v8JO928D0~7!N00;of09jj-=AX;p4 -FCY1CjbB(00000000000001_fpXpw0B~t=FJE?LZe(wAFLGsZb!BsOb1zIuLq$$gMNmrt0u%!j00008 -0LuVbTM8r8<7Wo|0KpRg03ZMW00000000000HlG8>=FQQX>c!Jc4cm4Z*nhkWpQ<7b98erUtei%X>?y --E^v8JO928D0~7!N00;of09jjpxozqM0RRBA0RR9a00000000000001_fid(F0B~t=FJE?LZe(wAFLG -sZb!BsOb1z?Cc4cyNX>V>{UoLQYP)h*<6ay3h000O8%K%wh$SBO|hXMcq*98CoCjbBd0000000000q= -9zz5&&>%a4%nWWo~3|axZdaadl;LbaO9XX>N37a&BR4Uv+e8Y;!Jfc~DCM0u%!j000080LuVbTLRuU^ -j8@G06k{_03ZMW00000000000HlFJ_!0ncX>c!Jc4cm4Z*nhkWpQ<7b98erVPs)&bY*gLE^v8JO928D -0~7!N00;of09jje{x$+p1^@st82|tq00000000000001_f!7fe0B~t=FJE?LZe(wAFLGsZb!BsOb1z| -VX)bViP)h*<6ay3h000O8%K%whvfhoGdJ6ym$|nE-8~^|S0000000000q=8`=698~&a4%nWWo~3|axZ -daadl;LbaO9Zb#!PhaCuNm0Rj{Q6aWAK2ms3fSzCvi9~~P3006lG0012T0000000000005+c6(kb?aA -|NaUv_0~WN&gWa%FLKWpi|MFJonLbaO6nc~DCM0u%!j000080LuVbTO!V%e>?*K03HVb03HAU000000 -00000HlFyB@+N}X>c!Jc4cm4Z*nhkWpQ<7b98erV{dJ6VRSBVc~DCM0u%!j000080LuVbTh`v<*u)b6 -0G>zy03QGV00000000000HlHAC=&p1X>c!Jc4cm4Z*nhkWpQ<7b98erV{dP3X=QURaCuNm0Rj{Q6aWA -K2ms3fSzEw8U^;;X006KM001HY0000000000005+cE7}CjbEViU0r}00000 -000000001_fd@$w0B~t=FJE?LZe(wAFLGsZb!BsOb1!XgWMyn~E^v8JO928D0~7!N00;of09jj@A65m -q0RR9Y1ONaa00000000000001_fmCo40B~t=FJE?LZe(wAFLGsZb!BsOb1!gVV{2h&WpgfYc~DCM0u% -!j000080LuVbTf8{G|9Bw)04H|<03ZMW00000000000HlFSauWb>X>c!Jc4cm4Z*nhkWpQ<7b98erb7 -gaLX>V?GE^v8JO928D0~7!N00;of09jkCNtR`s1^@u!5C8xq00000000000001_fdP~g0B~t=FJE?LZ -e(wAFLGsZb!BsOb1!prVRUtKUt@1%WpgfYc~DCM0u%!j000080LuVbTTv2BxmyGP0ALFM03rYY00000 -000000HlH5niBwUX>c!Jc4cm4Z*nhkWpQ<7b98erb98cbV{~0384T00000000000HlF=pA!IZX>c!Jc4cm4Z*nhkWpQ<7b98erb#!TLb1rasP)h*<6ay -3h000O8%K%wh!Ktmw+W`Oo?*ae-9smFU0000000000q=C4|698~&a4%nWWo~3|axZdab8l>RWo&6;FH -A{8MNU&iP)h*<6ay3h000O8%K%whHq^9X)Bpeg9RdIVA^-pY0000000000q=C=M698~&a4%nWWo~3|a -xZdab8l>RWo&6;FJE72ZfSI1UoLQYP)h*<6ay3h000O8%K%whd5_9tRWo&6;FK}{ic4=f~a&sc!Jc4cm4Z*nhkWpi(Ac4cg7VlQxcE_8WtWn@rG0Rj{Q6aW -AK2ms3fSz8@my%N&`008v|001KZ0000000000005+cN7fSnaA|NaUv_0~WN&gWa%FRGY<6XAX<{#OWp -HnDbY*gLE^v8JO928D0~7!N00;of09jk`Mfc%Z2LJ#R82|tw00000000000001_fm_)V0B~t=FJE?LZ -e(wAFLGsbZ)|pDY-wUIb98cbV{~&aaCuNm0Rj{Q6aWAK2ms3fSzEDC;(PG`004jj001li0000000000 -005+c^WYNzaA|NaUv_0~WN&gWa%FRGY<6XAX<{#OWpi(Ac4cyNFJE72ZfSI1UoLQYP)h*<6ay3h000O -8%K%whXZ0Csr~&{02L=EDEC2ui0000000000q=7Et698~&a4%nWWo~3|axZdab8l>RWo&6;FLGsbZ)| -pDa&s?XVsmtIVPkYIaCuNm0Rj{Q6aWAK2ms3fSzEFR#o@yO008I)001oj0000000000005+cAmtMPaA -|NaUv_0~WN&gWa%FRGY<6XAX<{#OWpi(Ac4cyNFJp3PbY*gBZ*DGdc~DCM0u%!j000080LuVbTkdJj@ -J<2%0O$q)04@Lk00000000000HlE~=o0{NX>c!Jc4cm4Z*nhkWpi(Ac4cg7VlQ%Kb8l>RWpZ;bWq4y{ -aCB*JZgVbhc~DCM0u%!j000080LuVbTQ8?^u;Uj10F+(;04@Lk00000000000HlG)>JtEPX>c!Jc4cm -4Z*nhkWpi(Ac4cg7VlQ%Kb8l>RWpZ;ba%FRGY;|;LZ*DGdc~DCM0u%!j000080LuVbTmP2wN^1fD03Z -Ya02lxO00000000000HlHU0u%snX>c!Jc4cm4Z*nhkX=7+FOi4pUPE$otO928D0~7!N00;of09jjMgQ ->*;2mk=e7XSbp00000000000001_ftUpp0B~t=FJE?LZe(wAFLG&PXfI!1X>MtBUtcb8c~DCM0u%!j0 -00080LuVbTheV~Q;7=z0N5S?02}}S00000000000HlG^4HN)yX>c!Jc4cm4Z*nhkX=7+FUteuuX>MO% -E^v8JO928D0~7!N00;of09jjww)mB53IG7%CjbB-00000000000001_ft48)0B~t=FJE?LZe(wAFLG& -PXfI!5Wo&F;cWGpFXmc)bc~DCM0u%!j000080LuVbTYoO@fGlwU0Lvo+03ZMW00000000000HlFEBoq -K}X>c!Jc4cm4Z*nhkX=7+FUuA7?YH43%Z)9b2E^v8JO928D0~7!N00;of09jk(dGEQ_0RR9f1ONac00 -000000000001_fv1%e0B~t=FJE?LZe(wAFLG&PXfI!7ZEtF6Uvgz|Y+++%E^v8JO928D0~7!N00;of0 -9jieTGfm<0{{R}2mk;g00000000000001_fxVX$0B~t=FJE?LZe(wAFLG&PXfI!7cyMoWbYEs~a&2LB -E^v8JO928D0~7!N00;of09ji})VJH4000090RR9V00000000000001_fi#;G0B~t=FJE?LZe(wAFLG& -PXfI!7cywiMb7^mGE^v8JO928D0~7!N00;of09jk@!NV4Z0RR9W0{{RU00000000000001_feV}z0B~ -t=FJE?LZe(wAFLG&PXfI!8X>4U~Z!U0oP)h*<6ay3h000O8%K%wh-=?)RT?+sJxFrAp8~^|S0000000 -000q=Czw6aa8(a4%nWWo~3|axZdeV`wj5X>N0HWn*+MaCuNm0Rj{Q6aWAK2ms3fSz8vX<{P&I005Z`0 -018V0000000000005+cVyP4WaA|NaUv_0~WN&gWa%p2|FJEkLXJ2w)0B~t=FJE?LZe(wAFLG&PXfI!EZ*OoeaCuNm0Rj{ -Q6aWAK2ms3fSzEm*)3c)i0043Y0015U0000000000005+cvaS>WaA|NaUv_0~WN&gWa%p2|FJEqTY;0 -d4UKaCuNm0Rj{Q6aWAK2ms3fSz8yOS1pAE005U80012T0000000000005+ck+2j1aA|NaUv_0~WN -&gWa%p2|FJEwBY-MzGWpgfYc~DCM0u%!j000080LuVbTj}ubOyK|k0H*-}02u%P00000000000HlFYw -G;qwX>c!Jc4cm4Z*nhkX=7+FUvOz-Yc6nkP)h*<6ay3h000O8%K%wh(ips8j|Kn$%@hCt8UO$Q00000 -00000q=9L+6aa8(a4%nWWo~3|axZdeV`wj5a$$67Z!U0oP)h*<6ay3h000O8%K%wh#xbo_Uk(5O&rJX -T9RL6T0000000000q=7KK6aa8(a4%nWWo~3|axZdeV`wj5b8u;HZe?Db4E -^v8JO928D0~7!N00;of09jiOAN76G6953?SO5Sb00000000000001_f&0xA0B~t=FJE?LZe(wAFLG&P -XfI!PX>KzzUt@1>b8l>AE^v8JO928D0~7!N00;of09jj17JZz&0ssJo2LJ#Z00000000000001_feqp -m0B~t=FJE?LZe(wAFLG&PXfI!PX>Me1cXKXqc~DCM0u%!j000080LuVbTcif2J>3ET0N)A#03`qb000 -00000000HlEfc!Jc4cm4Z*nhkX=7+FUw3J4WN&wKUvgz`WMy(?axQRrP)h*<6ay3h000O8%K -%wh=jik%Lj?c;Obq}482|tP0000000000q=6&o6aa8(a4%nWWo~3|axZdeV`wj5cXDBHE^v8JO928D0 -~7!N00;of09jiOYei9w0RRAc0{{RQ00000000000001_ft~FX0B~t=FJE?LZe(wAFLG&PXfI)6V=i!c -P)h*<6ay3h000O8%K%whO8dp)fCvBpR44!d82|tP0000000000q=8@W6aa8(a4%nWWo~3|axZdeV`wj -7Y-wk1E^v8JO928D0~7!N00;of09jidF6kGu2LJ#G8vp -c!Jc4cm4Z*nhkX=7+FVqtPFaCuNm0Rj{Q6aWAK2ms3fSzGR`YG7On006!z000*N0000000000005+cj -|3F}aA|NaUv_0~WN&gWa%p2|FJf -P0000000000005+cfff}2aA|NaUv_0~WN&gWa%p2|FJo_PZ*nehc~DCM0u%!j000080LuVbTWLfMPLc -rt03HMY03iSX00000000000HlF`DHQ;4X>c!Jc4cm4Z*nhkX=7+FV{dG4a$j_EX>e?1bS`jtP)h*<6a -y3h000O8%K%who=j5llm`F++ZzA?8vpc!Jc4cm4Z*nhk -X=7+FV{dMAZ){~QaCuNm0Rj{Q6aWAK2ms3fSzC)t(n;O{000OD0012T0000000000005+czls$AaA|N -aUv_0~WN&gWa%p2|FJo_Rb98cHX>KlXc~DCM0u%!j000080LuVbTMW`;cf|z&0Dct!03HAU00000000 -000HlG~j1>TIX>c!Jc4cm4Z*nhkX=7+FV{dMBVQFqc!Jc4cm4Z*nhkX=7+FV{dMBa&K%daCuNm0Rj{Q6aWAK2ms3fS -zAOlSyLwm002QC001HY0000000000005+c*P0aoaA|NaUv_0~WN&gWa%p2|FJxtAVRdYDUvqSMY-MvU -aCuNm0Rj{Q6aWAK2ms3fSzAALI+oV~00035000~S0000000000005+cKB5%>aA|NaUv_0~WN&gWa%p2 -|FJx(9XKrtEWiD`eP)h*<6ay3h000O8%K%wh;RV|O%>w`cKM4Q;82|tP0000000000q=8SQ6##H)a4% -nWWo~3|axZdeV`wjBZEtF6E^v8JO928D0~7!N00;of09jiiNFmeh0001j0ssIS00000000000001_fm -5gz0B~t=FJE?LZe(wAFLG&PXfI`Qa&K~TE^v8JO928D0~7!N00;of09jjV-ft9*0ssJ$1^@sa000000 -00000001_fq1DE0B~t=FJE?LZe(wAFLG&PXfI}IY-L|?a&LHfE^v8JO928D0~7!N00;of09jk6yFJ%A -1ONcE2><{b00000000000001_fkCVl0B~t=FJE?LZe(wAFLG&PXfI}IY-Mw4dSxzfc~DCM0u%!j0000 -80LuVbThqIXO+yO+0CFV&03QGV00000000000HlGmuoVDsX>c!Jc4cm4Z*nhkX=7+FXlZ9?Y-wj`bY* -fbaCuNm0Rj{Q6aWAK2ms3fSzF*uyoM?T005^G000;O0000000000005+cG`tl6aA|NaUv_0~WN&gWa% -p2|FKTmdZZ2?nP)h*<6ay3h000O8%K%whIKf!Riv$1wvc!Jc4cm4Z*nhkX=7+FY+-qCb#yLpc~DCM0u%!j000080LuVbTiGDR&Z!Ro08Bdo02lxO0000 -0000000HlF_)D-}5X>c!Jc4cm4Z*nhkX=7+FY-x67E^v8JO928D0~7!N00;of09jiBjB{=30{{TQ4FC -Wi00000000000001_fnDSk0B~t=FJE?LZe(wAFLG&PXfJGOc4c33Wo~3;axQRrP)h*<6ay3h000O8%K -%whrhqvvY7PJZv@iew8vp00SWa02%-Q00000000000HlE@_Z0weX>c!Jc4cm4Z*nhkX=7+FZDDe2 -b#N|lc~DCM0u%!j000080LuVbTmP&W1Z)KW0J#(Z02=@R00000000000HlGg0TuvoX>c!Jc4cm4Z*nh -kX=7+FZDnC|b#i4caCuNm0Rj{Q6aWAK2ms3fSzG^SPo&lb001l#000{R0000000000005+cQwJ6RaA| -NaUv_0~WN&gWa%p2|FK}UGWNB_^E^v8JO928D0~7!N00;of09jiO#ub8d0RR9z0{{RS000000000000 -01_fnp670B~t=FJE?LZe(wAFLG&PXfJSKXJv9OaCuNm0Rj{Q6aWAK2ms3fSz9;&K3}^8002Y{000{R0 -000000000005+c3=b9naA|NaUv_0~WN&gWa%p2|FK}UOWps39E^v8JO928D0~7!N00;of09jkhSfMvZ -3IG6=D*yl)00000000000001_fdCX10B~t=FJE?LZe(wAFLG&PXfJSKZe?sPaCuNm0Rj{Q6aWAK2ms3 -fSz9pk4CHMe001YB000^Q0000000000005+cejXM8aA|NaUv_0~WN&gWa%p2|FK}{YbaZ(xaCuNm0Rj -{Q6aWAK2ms3fSzAbM;c!Jc4cm4Z*nhkX=7+ -Fa&>HFE^v8JO928D0~7!N00;of09jiPh4m|R1ONaV3jhEa00000000000001_f$oqN0B~t=FJE?LZe( -wAFLG&PXfJbPZ*XNUaCuNm0Rj{Q6aWAK2ms3fSzA9f{;pC2001`z000^Q0000000000005+cnUxj*aA -|NaUv_0~WN&gWa%p2|FLPsZWo2$IaCuNm0Rj{Q6aWAK2ms3fSzA^q8>EyJ005_8000{R00000000000 -05+cBbXKdaA|NaUv_0~WN&gWa%p2|FLPyQZDnqBE^v8JO928D0~7!N00;of09jiX&zvbH1pokc5C8xh -00000000000001_f$*#r0B~t=FJE?LZe(wAFLG&PXfJbcX>M+1axQRrP)h*<6ay3h000O8%K%wh0$S! -e_yhm|ND%-48UO$Q0000000000q=8Se765Q*a4%nWWo~3|axZdeV`wjPbYXOLb1rasP)h*<6ay3h000 -O8%K%whp~5SCy%qofZfO7j82|tP0000000000q=A38765Q*a4%nWWo~3|axZdeV`wjPba`xLE^v8JO9 -28D0~7!N00;of09jk{3Z_ui0RRB%1ONaU00000000000001_fpg9l0B~t=FJE?LZe(wAFLG&PXfJbgd -2D55E^v8JO928D0~7!N00;of09jk&MA_5+AOHXpjsO4}00000000000001_fq>8!0B~t=FJE?LZe(wA -FLG&PXfJblZggRIE^v8JO928D0~7!N00;of09ji*|GX8o9{>PxoB#kB00000000000001_fwu7$0B~t -=FJE?LZe(wAFLG&PXfJeOVr*qDaCuNm0Rj{Q6aWAK2ms3fSz9_I)4%=$001ft001HY0000000000005 -+cq7fGWaA|NaUv_0~WN&gWa%p2|FLY&cZE0>{Y+rO}Wo>0HaCuNm0Rj{Q6aWAK2ms3fSzA+?Q=FeD00 -6+b000;O0000000000005+c;1w4DaA|NaUv_0~WN&gWa%p2|FLY&ibS`jtP)h*<6ay3h000O8%K%whq -u(SR$^-xayAA*V82|tP0000000000q=B_Q7XWZ+a4%nWWo~3|axZdeV`wjQXk~3>E^v8JO928D0~7!N -00;of09jj6b@ot40001H0000Q00000000000001_fwn^z0B~t=FJE?LZe(wAFLG&PXfJeVWo>11E^v8 -JO928D0~7!N00;of09jirWX9x|9smFpi~s-~00000000000001_fi*-I0B~t=FJE?LZe(wAFLG&PXfJ -efVPj=tVPk79aCuNm0Rj{Q6aWAK2ms3fSzDI}FBhZ=008SG000;O0000000000005+c2V)liaA|NaUv -_0~WN&gWa%p2|FLZKcWiD`eP)h*<6ay3h000O8%K%whU08L3cLD$aFa!Vq82|tP0000000000q=DgW7 -XWZ+a4%nWWo~3|axZjmZER^TOi4pUPE$otO928D0~7!N00;of09jklI#2NI0000w0RR9U0000000000 -0001_fsb$(0B~t=FJE?LZe(wAFLZBhY-ulFUukY>bYEXCaCuNm0Rj{Q6aWAK2ms3fSzBL?NIqm5006R -N000~S0000000000005+cws98#aA|NaUv_0~WN&gWbZ>2JX)j-JVRCb2axQRrP)h*<6ay3h000O8%K% -whd3V*}4+Q`KqYMB57ytkO0000000000q=8h77XWZ+a4%nWWo~3|axZjmZER^TUvgzGaCuNm0Rj{Q6a -WAK2ms3fSzEL35@*E#008~~000{R0000000000005+cnUNO&aA|NaUv_0~WN&gWbZ>2JX)j-Nd2nTOE -^v8JO928D0~7!N00;of09jirDFFB!0000Q0000Q00000000000001_ftQjO0B~t=FJE?LZe(wAFLZBh -Y-ulWc`kH$aAjmrO928D0~7!N00;of09jjISapJT0ssIo1ONaV00000000000001_f$NeN0B~t=FJE? -LZe(wAFLZBhY-wM2FHA{8MNU&iP)h*<6ay3h000O8%K%whYdEsvjsO4vsQ>@~9{>OV0000000000q=B -227XWZ+a4%nWWo~3|axZjmZER^@cQ0RGX>MtBUtcb8c~DCM0u%!j000080LuVbTbooFzB35`05KZ?03 -HAU00000000000HlFtmKOkUX>c!Jc4cm4Z*nhmZ*6R8Uw1EGcXDZTWpXZXc~DCM0u%!j000080LuVbT -PrC5_#6NL02%-Q02}}S00000000000HlG{pBDgdX>c!Jc4cm4Z*nhmZ*6R8Uw1EXc`kH$aAjmrO928D -0~7!N00;of09jkb4+^k-0ssI$1ONaY00000000000001_fh(XF0B~t=FJE?LZe(wAFLZKsb98fbZ*pZ -XOi4pUPE$otO928D0~7!N00;of09jiPHFKz00ssIg1pojd00000000000001_f#jkW0B~t=FJE?LZe( -wAFLZKsb98fbZ*pZXUtei%X>?y-E^v8JO928D0~7!N00;of09jioVaNzW3jhF>EC2u=000000000000 -01_fqkVH0B~t=FJE?LZe(wAFLZKsb98fbZ*pZXUtw@*E^v8JO928D0~7!N00;of09jk!H@8li6951gP -yhfR00000000000001_f&8!+0B~t=FJE?LZe(wAFLZKsb98fbZ*pZXUu|JyZ*wkic~DCM0u%!j00008 -0LuVbTb)jZkQf920Bi{W03rYY00000000000HlG_#1{Z?X>c!Jc4cm4Z*nhma&>cbb98TVWiMZEaAj_ -Db8Iefc~DCM0u%!j000080LuVbTWP>t&&2@%0BQsP04M+e00000000000HlE{$rk`{X>c!Jc4cm4Z*n -hma&>cbb98TVWiMZIb8KH@Z*FsRVQzGDE^v8JO928D0~7!N00;of09jj6{tboz5C8y4MgRaJ0000000 -0000001_fi}w*0B~t=FJE?LZe(wAFLZKsb98fbZ*pZXUw3J4WN&wKE^v8JO928D0~7!N00;of09jiA0 -0002000000000V00000000000001_fp*&$0B~t=FJE?LZe(wAFLZKsb98fbZ*pZXaCt6td2nT9P)h*< -6ay3h000O8%K%whni!`vlmY+%TLb_A9{>OV0000000000q=B{D7XWZ+a4%nWWo~3|axZmqY;0*_GcQa -@Lq$$gMJ{xBbWlqH0u%!j000080LuVbTTg=^C~5@&00j*I03QGV00000000000HlG2-WLFHX>c!Jc4c -m4Z*nhna%^mAVlyvaUukY>bYEXCaCuNm0Rj{Q6aWAK2ms3fSzE!vKPI#d003+(001KZ000000000000 -5+cE#wyfaA|NaUv_0~WN&gWb#iQMX<{=kUt@1V?GE^v8JO928D0~7!N00;of09jkux;?u<0 -000$0000V00000000000001_fgV?DZ*OcaaCuNm0Rj{Q6aWAK2ms3fSzBJrNww<<000>v001EX0000000 -000005+c2sRi1aA|NaUv_0~WN&gWb#iQMX<{=kWq4y{aCB*JZgVbhc~DCM0u%!j000080LuVbTgNC%E -FcO10D~a_0384T00000000000HlF3Ko|gUX>c!Jc4cm4Z*nhna%^mAVlyvhX=Q9=b1rasP)h*<6ay3h -000O8%K%whhpWG%eggmihzS4y9{>OV0000000000q=Ah}7yxi-a4%nWWo~3|axZmqY;0*_GcRUoY-Mn -7b963nc~DCM0u%!j000080LuVbTRUdkn4cB^02WRF03rYY00000000000HlFPP8a}iX>c!Jc4cm4Z*n -hna%^mAVlyvrZ*OdEVQyh(WpXZXc~DCM0u%!j000080LuVbTL4*}jI#&;03#Xz03HAU00000000000H -lE@Wf%Z(X>c!Jc4cm4Z*nhna%^mAVlyvtWpQ<7b963nc~DCM0u%!j000080LuVbTTmdlI%6OJ0I_!f0 -3QGV00000000000HlEyZWsV?X>c!Jc4cm4Z*nhna%^mAVlyvtWpi+EZgXWWaCuNm0Rj{Q6aWAK2ms3f -Sz7=A0006200000001Wd0000000000005+cwvHG8aA|NaUv_0~WN&gWb#iQMX<{=kV{dMBa%o~OUtei -%X>?y-E^v8JO928D0~7!N00;of09jiqgXtBu0RRBK0{{Rq00000000000001_f&Gpc0B~t=FJE?LZe( -wAFLiQkY-wUMFJo_RbaH88FJEDBaAj_1X>Mg-Wo~w9a&K-faCuNm0Rj{Q6aWAK2ms3fSz8M3lZj*u00 -0~-001Ze0000000000005+c0g)I0aA|NaUv_0~WN&gWb#iQMX<{=kV{dMBa%o~OVQ_F|Zf9w3WiD`eP -)h*<6ay3h000O8%K%wh6=~@=QU(A3un_c!Jc4cm4Z*nhna%^mAVlyveZ*FvQX<{#Md2euKZgX>NE^v8JO928D0~7!N00;of09ji%{d6 -_GB>(`9h5!IA00000000000001_fmyj20B~t=FJE?LZe(wAFLiQkY-wUMFJo_RbaH88FLPyMb#i5Na$ -#c!Jc4cm4Z*nhna%^mAVlyveZ*FvQX<{#5b7f<7a%FUKVQzD9Z*p`mUtei%X>?y --E^v8JO928D0~7!N00;of09jiu&n->v5C8z+L;wId00000000000001_fmP@j0B~t=FJE?LZe(wAFLi -QkY-wUMFJo_RbaH88FJE(IV|8+6baG*Cb8v5RbT49QZe(e0XLBxac~DCM0u%!j000080LuVbTdpdnyj -c+d0Ae-(0672v00000000000HlGL`4|9jX>c!Jc4cm4Z*nhna%^mAVlyveZ*FvQX<{#5b7f<7a%FUKV -QzD9Z*p`mY;Sj8Y-M(3Y%XwlP)h*<6ay3h000O8%K%wh000000ssI200000C;$Ke0000000000q=88b -831r;a4%nWWo~3|axZmqY;0*_GcRyqV{2h&WpgiIUukY>bYEXCaCuNm0Rj{Q6aWAK2ms3fSzBC`8LTE -D003!+001KZ0000000000005+ckqa3BaA|NaUv_0~WN&gWb#iQMX<{=kaA9L>VP|D?FLP;lE^v8JO92 -8D0~7!N00;of09jiA00002000000000o00000000000001_f%q&L0B~t=FJE?LZe(wAFLiQkY-wUMFK -}UFYhh<)b1!0HV{344a&&VqUtei%X>?y-E^v8JO928D0~7!N00;of09jk!xMQD1^@ux6aWA^00000000000001_f#NS20B~t=FJE?LZe(wAFLiQkY-wU -MFK}UFYhh<)b1!0HV{344a&&VqcV%H~a%E;;W@&C=Y-xIBE^v8JO928D0~7!N00;of09jk+ikX_&0RR -Al1ONae00000000000001_fyFi%0B~t=FJE?LZe(wAFLiQkY-wUMFLiWjY%gD5X>MtBUtcb8c~DCM0u -%!j000080LuVbTkNXj51I!603{Ou044wc00000000000HlH6I2izNX>c!Jc4cm4Z*nhna%^mAVlyvwb -ZKlaV{dM5Wn*+{Z*DGdc~DCM0u%!j000080LuVbTMExy2VMdI07V7>03iSX00000000000HlGvKp6mV -X>c!Jc4cm4Z*nhna%^mAVlyvwbZKlaaB^>Wc`k5yP)h*<6ay3h000O8%K%whP6PIz`~Uy|@&Nz?X>2cXb!ByBE^v8JO928D0~7!N00;of09j -i%j&1R)1pok@4*&oo00000000000001_ft5rV0B~t=FJE?LZe(wAFLiQkY-wUMFLiWjY%g+Uadl;LbS -`jtP)h*<6ay3h000O8%K%whHYnA!U?X>2cYWpi+EZgXWWaCuNm0Rj{Q6aWAK2ms3fSzBKfFZT);001IZ001EX0000000000005+c -BTyLtaA|NaUv_0~WN&gWb#iQMX<{=kb#!TLFLGsca(OOrc~DCM0u%!j000080LuVbTR1R}=I#~%05n7 -Z03ZMW00000000000HlFzXBhx+X>c!Jc4cm4Z*nhna%^mAVlyvwbZKlab8~E8E^v8JO928D0~7!N00; -of09jic!Jc4cm4Z*nhna%^mAVlyvwbZKlabZKp6Z*_DoaCuNm0Rj{ -Q6aWAK2ms3fSz9aB=w25S007uG0018V0000000000005+cIh+{)aA|NaUv_0~WN&gWb#iQMX<{=kb#! -TLFLiQkE^v8JO928D0~7!N00;of09jjB)K=BM2LJ#Y6#xJr00000000000001_fsL;j0B~t=FJE?LZe -(wAFLiQkY-wUMFLiWjY%h0VX>=}dc~DCQ1^@s60J;IX0sn^p0E)L60000 -""" - - -if __name__ == "__main__": - main() diff --git a/convert/parakeet-tdt-ctc-110m/output/Decoder.mlmodelc/analytics/coremldata.bin b/convert/parakeet-tdt-ctc-110m/output/Decoder.mlmodelc/analytics/coremldata.bin deleted file mode 100644 index 18de02b6523b282aa52b4e816db388a961557778..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/output/Decoder.mlmodelc/analytics/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:fefed2be21166b9afcc5913192d8b86b4a621ee236b459a2cc754584020d4878 -size 243 diff --git a/convert/parakeet-tdt-ctc-110m/output/Decoder.mlmodelc/coremldata.bin b/convert/parakeet-tdt-ctc-110m/output/Decoder.mlmodelc/coremldata.bin deleted file mode 100644 index f6cebbf78c8aa9626ae27c1a9d8b681c3bf0f441..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/output/Decoder.mlmodelc/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:66953a7662cfdc8247e5f3da0ec38847d1e198a5b5dd94b51fa1c711f7b777ac -size 543 diff --git a/convert/parakeet-tdt-ctc-110m/output/Decoder.mlmodelc/metadata.json b/convert/parakeet-tdt-ctc-110m/output/Decoder.mlmodelc/metadata.json deleted file mode 100644 index 84ebb9b5a894782e6def44d0b3707784e36a1757..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/output/Decoder.mlmodelc/metadata.json +++ /dev/null @@ -1,123 +0,0 @@ -[ - { - "metadataOutputVersion" : "3.0", - "shortDescription" : "Hybrid TDT Decoder (110M)", - "outputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float16", - "formattedType" : "MultiArray (Float16 1 × 640 × 2)", - "shortDescription" : "", - "shape" : "[1, 640, 2]", - "name" : "decoder_output", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float16", - "formattedType" : "MultiArray (Float16 1 × 1 × 640)", - "shortDescription" : "", - "shape" : "[1, 1, 640]", - "name" : "h_out", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float16", - "formattedType" : "MultiArray (Float16 1 × 1 × 640)", - "shortDescription" : "", - "shape" : "[1, 1, 640]", - "name" : "c_out", - "type" : "MultiArray" - } - ], - "storagePrecision" : "Float16", - "modelParameters" : [ - - ], - "author" : "Fluid Inference", - "specificationVersion" : 8, - "mlProgramOperationTypeHistogram" : { - "Select" : 1, - "Ios17.squeeze" : 2, - "Ios17.gather" : 1, - "Ios17.cast" : 5, - "Ios17.lstm" : 1, - "Ios17.transpose" : 2, - "Ios17.add" : 1, - "Ios17.concat" : 1, - "Ios17.greaterEqual" : 1, - "Identity" : 1, - "Ios17.expandDims" : 2 - }, - "computePrecision" : "Mixed (Float16, Int16, Int32)", - "isUpdatable" : "0", - "stateSchema" : [ - - ], - "availability" : { - "macOS" : "14.0", - "tvOS" : "17.0", - "visionOS" : "1.0", - "watchOS" : "10.0", - "iOS" : "17.0", - "macCatalyst" : "17.0" - }, - "modelType" : { - "name" : "MLModelType_mlProgram" - }, - "inputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1 × 1)", - "shortDescription" : "", - "shape" : "[1, 1]", - "name" : "targets", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1)", - "shortDescription" : "", - "shape" : "[1]", - "name" : "target_lengths", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 1 × 640)", - "shortDescription" : "", - "shape" : "[1, 1, 640]", - "name" : "h_in", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 1 × 640)", - "shortDescription" : "", - "shape" : "[1, 1, 640]", - "name" : "c_in", - "type" : "MultiArray" - } - ], - "userDefinedMetadata" : { - "com.github.apple.coremltools.conversion_date" : "2026-01-01", - "com.github.apple.coremltools.source" : "torch==2.8.0", - "com.github.apple.coremltools.version" : "9.0", - "com.github.apple.coremltools.source_dialect" : "TorchScript" - }, - "generatedClassName" : "Decoder", - "method" : "predict" - } -] \ No newline at end of file diff --git a/convert/parakeet-tdt-ctc-110m/output/Decoder.mlmodelc/model.mil b/convert/parakeet-tdt-ctc-110m/output/Decoder.mlmodelc/model.mil deleted file mode 100644 index 3d8abb8f2d6b323d3d0fd6e3d34bc1f8145e5ca4..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/output/Decoder.mlmodelc/model.mil +++ /dev/null @@ -1,51 +0,0 @@ -program(1.0) -[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3510.2.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.8.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "9.0"}})] -{ - func main(tensor c_in, tensor h_in, tensor target_lengths, tensor targets) { - tensor y_batch_dims_0 = const()[name = tensor("y_batch_dims_0"), val = tensor(0)]; - tensor y_validate_indices_0 = const()[name = tensor("y_validate_indices_0"), val = tensor(false)]; - tensor decoder_prediction_embed_weight_to_fp16 = const()[name = tensor("decoder_prediction_embed_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; - tensor targets_to_int16_dtype_0 = const()[name = tensor("targets_to_int16_dtype_0"), val = tensor("int16")]; - tensor cast_4_dtype_0 = const()[name = tensor("cast_4_dtype_0"), val = tensor("int32")]; - tensor greater_equal_0_y_0 = const()[name = tensor("greater_equal_0_y_0"), val = tensor(0)]; - tensor targets_to_int16 = cast(dtype = targets_to_int16_dtype_0, x = targets)[name = tensor("cast_9")]; - tensor cast_4 = cast(dtype = cast_4_dtype_0, x = targets_to_int16)[name = tensor("cast_8")]; - tensor greater_equal_0 = greater_equal(x = cast_4, y = greater_equal_0_y_0)[name = tensor("greater_equal_0")]; - tensor slice_by_index_0 = const()[name = tensor("slice_by_index_0"), val = tensor(1025)]; - tensor add_1 = add(x = cast_4, y = slice_by_index_0)[name = tensor("add_1")]; - tensor select_0 = select(a = cast_4, b = add_1, cond = greater_equal_0)[name = tensor("select_0")]; - tensor y_cast_fp16_cast_uint16_axis_0 = const()[name = tensor("y_cast_fp16_cast_uint16_axis_0"), val = tensor(0)]; - tensor select_0_to_int16_dtype_0 = const()[name = tensor("select_0_to_int16_dtype_0"), val = tensor("int16")]; - tensor select_0_to_int16 = cast(dtype = select_0_to_int16_dtype_0, x = select_0)[name = tensor("cast_7")]; - tensor y_cast_fp16_cast_uint16_cast_uint16 = gather(axis = y_cast_fp16_cast_uint16_axis_0, batch_dims = y_batch_dims_0, indices = select_0_to_int16, validate_indices = y_validate_indices_0, x = decoder_prediction_embed_weight_to_fp16)[name = tensor("y_cast_fp16_cast_uint16_cast_uint16")]; - tensor input_1_perm_0 = const()[name = tensor("input_1_perm_0"), val = tensor([1, 0, 2])]; - tensor input_lstm_h0_squeeze_axes_0 = const()[name = tensor("input_lstm_h0_squeeze_axes_0"), val = tensor([0])]; - tensor h_in_to_fp16_dtype_0 = const()[name = tensor("h_in_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor h_in_to_fp16 = cast(dtype = h_in_to_fp16_dtype_0, x = h_in)[name = tensor("cast_6")]; - tensor input_lstm_h0_squeeze_cast_fp16 = squeeze(axes = input_lstm_h0_squeeze_axes_0, x = h_in_to_fp16)[name = tensor("input_lstm_h0_squeeze_cast_fp16")]; - tensor input_lstm_c0_squeeze_axes_0 = const()[name = tensor("input_lstm_c0_squeeze_axes_0"), val = tensor([0])]; - tensor c_in_to_fp16_dtype_0 = const()[name = tensor("c_in_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor c_in_to_fp16 = cast(dtype = c_in_to_fp16_dtype_0, x = c_in)[name = tensor("cast_5")]; - tensor input_lstm_c0_squeeze_cast_fp16 = squeeze(axes = input_lstm_c0_squeeze_axes_0, x = c_in_to_fp16)[name = tensor("input_lstm_c0_squeeze_cast_fp16")]; - tensor input_direction_0 = const()[name = tensor("input_direction_0"), val = tensor("forward")]; - tensor input_output_sequence_0 = const()[name = tensor("input_output_sequence_0"), val = tensor(true)]; - tensor input_recurrent_activation_0 = const()[name = tensor("input_recurrent_activation_0"), val = tensor("sigmoid")]; - tensor input_cell_activation_0 = const()[name = tensor("input_cell_activation_0"), val = tensor("tanh")]; - tensor input_activation_0 = const()[name = tensor("input_activation_0"), val = tensor("tanh")]; - tensor concat_1_to_fp16 = const()[name = tensor("concat_1_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1312128)))]; - tensor concat_2_to_fp16 = const()[name = tensor("concat_2_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4588992)))]; - tensor concat_0_to_fp16 = const()[name = tensor("concat_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7865856)))]; - tensor input_1_cast_fp16 = transpose(perm = input_1_perm_0, x = y_cast_fp16_cast_uint16_cast_uint16)[name = tensor("transpose_2")]; - tensor input_cast_fp16_0, tensor input_cast_fp16_1, tensor input_cast_fp16_2 = lstm(activation = input_activation_0, bias = concat_0_to_fp16, cell_activation = input_cell_activation_0, direction = input_direction_0, initial_c = input_lstm_c0_squeeze_cast_fp16, initial_h = input_lstm_h0_squeeze_cast_fp16, output_sequence = input_output_sequence_0, recurrent_activation = input_recurrent_activation_0, weight_hh = concat_2_to_fp16, weight_ih = concat_1_to_fp16, x = input_1_cast_fp16)[name = tensor("input_cast_fp16")]; - tensor var_30_axes_0 = const()[name = tensor("op_30_axes_0"), val = tensor([0])]; - tensor h_out = expand_dims(axes = var_30_axes_0, x = input_cast_fp16_1)[name = tensor("op_30_cast_fp16")]; - tensor var_31_axes_0 = const()[name = tensor("op_31_axes_0"), val = tensor([0])]; - tensor c_out = expand_dims(axes = var_31_axes_0, x = input_cast_fp16_2)[name = tensor("op_31_cast_fp16")]; - tensor transpose_0_perm_0 = const()[name = tensor("transpose_0_perm_0"), val = tensor([1, 2, 0])]; - tensor var_48 = const()[name = tensor("op_48"), val = tensor(2)]; - tensor var_49_interleave_0 = const()[name = tensor("op_49_interleave_0"), val = tensor(false)]; - tensor transpose_0_cast_fp16 = transpose(perm = transpose_0_perm_0, x = input_cast_fp16_0)[name = tensor("transpose_1")]; - tensor decoder_output = concat(axis = var_48, interleave = var_49_interleave_0, values = (transpose_0_cast_fp16, transpose_0_cast_fp16))[name = tensor("op_49_cast_fp16")]; - tensor target_lengths_tmp = identity(x = target_lengths)[name = tensor("target_lengths_tmp")]; - } -> (decoder_output, h_out, c_out); -} \ No newline at end of file diff --git a/convert/parakeet-tdt-ctc-110m/output/Decoder.mlmodelc/weights/weight.bin b/convert/parakeet-tdt-ctc-110m/output/Decoder.mlmodelc/weights/weight.bin deleted file mode 100644 index 4543ad4ea89b21ebed0de0d5ec3975d07eace7fb..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/output/Decoder.mlmodelc/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:dd90b58597ee2c172c672dffe13b1110898ba07394c1a15efc96cc8c6b18411b -size 7871040 diff --git a/convert/parakeet-tdt-ctc-110m/output/Decoder.mlpackage/Data/com.apple.CoreML/model.mlmodel b/convert/parakeet-tdt-ctc-110m/output/Decoder.mlpackage/Data/com.apple.CoreML/model.mlmodel deleted file mode 100644 index 70eaf8653921b3a9c2809f32857e034913cc2e86..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/output/Decoder.mlpackage/Data/com.apple.CoreML/model.mlmodel +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:75a807f2b8ec1cde294f77fe8a1faf50229f011a8e159d7d6c6beaab778ce8c1 -size 7501 diff --git a/convert/parakeet-tdt-ctc-110m/output/Decoder.mlpackage/Data/com.apple.CoreML/weights/weight.bin b/convert/parakeet-tdt-ctc-110m/output/Decoder.mlpackage/Data/com.apple.CoreML/weights/weight.bin deleted file mode 100644 index 4543ad4ea89b21ebed0de0d5ec3975d07eace7fb..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/output/Decoder.mlpackage/Data/com.apple.CoreML/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:dd90b58597ee2c172c672dffe13b1110898ba07394c1a15efc96cc8c6b18411b -size 7871040 diff --git a/convert/parakeet-tdt-ctc-110m/output/Decoder.mlpackage/Manifest.json b/convert/parakeet-tdt-ctc-110m/output/Decoder.mlpackage/Manifest.json deleted file mode 100644 index b24fa578e7986b6fc0ac0dba9f4481b1a819b0fe..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/output/Decoder.mlpackage/Manifest.json +++ /dev/null @@ -1,18 +0,0 @@ -{ - "fileFormatVersion": "1.0.0", - "itemInfoEntries": { - "6BDFA15B-7DF3-44B6-A308-B95282CBC63A": { - "author": "com.apple.CoreML", - "description": "CoreML Model Specification", - "name": "model.mlmodel", - "path": "com.apple.CoreML/model.mlmodel" - }, - "A04A1CE0-C74A-460D-B289-34CAF7B6F6C9": { - "author": "com.apple.CoreML", - "description": "CoreML Model Weights", - "name": "weights", - "path": "com.apple.CoreML/weights" - } - }, - "rootModelIdentifier": "6BDFA15B-7DF3-44B6-A308-B95282CBC63A" -} diff --git a/convert/parakeet-tdt-ctc-110m/output/JointDecision.mlmodelc/analytics/coremldata.bin b/convert/parakeet-tdt-ctc-110m/output/JointDecision.mlmodelc/analytics/coremldata.bin deleted file mode 100644 index f7d619494d2f2061ae52418edbb5c5fc2ec1f11e..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/output/JointDecision.mlmodelc/analytics/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:e51c4ee19e8d7e8392896e5d3b5464892f9e16ba271335ff7eb3d2eb4a015f27 -size 243 diff --git a/convert/parakeet-tdt-ctc-110m/output/JointDecision.mlmodelc/coremldata.bin b/convert/parakeet-tdt-ctc-110m/output/JointDecision.mlmodelc/coremldata.bin deleted file mode 100644 index cc76621d9ad55dc03a554246d198a4851607172f..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/output/JointDecision.mlmodelc/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:5c45adaff9e0e6d8c49949666f07a6fcde0affc7a631a8f93d12a0e51c1393d3 -size 509 diff --git a/convert/parakeet-tdt-ctc-110m/output/JointDecision.mlmodelc/metadata.json b/convert/parakeet-tdt-ctc-110m/output/JointDecision.mlmodelc/metadata.json deleted file mode 100644 index 4d8cd2c6a2393dfafad11338ff96f388b2d59295..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/output/JointDecision.mlmodelc/metadata.json +++ /dev/null @@ -1,104 +0,0 @@ -[ - { - "metadataOutputVersion" : "3.0", - "shortDescription" : "Hybrid Joint Decision (110M)", - "outputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1 × 1 × 1)", - "shortDescription" : "", - "shape" : "[1, 1, 1]", - "name" : "token_id", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float16", - "formattedType" : "MultiArray (Float16 1 × 1 × 1)", - "shortDescription" : "", - "shape" : "[1, 1, 1]", - "name" : "token_prob", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1 × 1 × 1)", - "shortDescription" : "", - "shape" : "[1, 1, 1]", - "name" : "duration", - "type" : "MultiArray" - } - ], - "storagePrecision" : "Float16", - "modelParameters" : [ - - ], - "author" : "Fluid Inference", - "specificationVersion" : 8, - "mlProgramOperationTypeHistogram" : { - "Ios17.reduceArgmax" : 2, - "Ios17.linear" : 3, - "Ios17.transpose" : 2, - "Ios17.log" : 1, - "Ios17.add" : 1, - "Ios17.sliceByIndex" : 2, - "Ios16.relu" : 1, - "Ios16.softmax" : 2, - "Ios17.expandDims" : 3, - "Ios17.squeeze" : 2, - "Ios17.cast" : 5, - "Ios17.gatherAlongAxis" : 1 - }, - "computePrecision" : "Mixed (Float16, Float32, Int16, Int32)", - "isUpdatable" : "0", - "stateSchema" : [ - - ], - "availability" : { - "macOS" : "14.0", - "tvOS" : "17.0", - "visionOS" : "1.0", - "watchOS" : "10.0", - "iOS" : "17.0", - "macCatalyst" : "17.0" - }, - "modelType" : { - "name" : "MLModelType_mlProgram" - }, - "inputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 512 × 1)", - "shortDescription" : "", - "shape" : "[1, 512, 1]", - "name" : "encoder_step", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 640 × 1)", - "shortDescription" : "", - "shape" : "[1, 640, 1]", - "name" : "decoder_step", - "type" : "MultiArray" - } - ], - "userDefinedMetadata" : { - "com.github.apple.coremltools.conversion_date" : "2026-01-01", - "com.github.apple.coremltools.source" : "torch==2.8.0", - "com.github.apple.coremltools.version" : "9.0", - "com.github.apple.coremltools.source_dialect" : "TorchScript" - }, - "generatedClassName" : "JointDecision", - "method" : "predict" - } -] \ No newline at end of file diff --git a/convert/parakeet-tdt-ctc-110m/output/JointDecision.mlmodelc/model.mil b/convert/parakeet-tdt-ctc-110m/output/JointDecision.mlmodelc/model.mil deleted file mode 100644 index 7057845c459a15cef5cb1a4baf0695b949c7f703..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/output/JointDecision.mlmodelc/model.mil +++ /dev/null @@ -1,66 +0,0 @@ -program(1.0) -[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3510.2.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.8.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "9.0"}})] -{ - func main(tensor decoder_step, tensor encoder_step) { - tensor input_1_perm_0 = const()[name = tensor("input_1_perm_0"), val = tensor([0, 2, 1])]; - tensor encoder_step_to_fp16_dtype_0 = const()[name = tensor("encoder_step_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor input_3_perm_0 = const()[name = tensor("input_3_perm_0"), val = tensor([0, 2, 1])]; - tensor decoder_step_to_fp16_dtype_0 = const()[name = tensor("decoder_step_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor joint_enc_weight_to_fp16 = const()[name = tensor("joint_enc_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; - tensor joint_enc_bias_to_fp16 = const()[name = tensor("joint_enc_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(655488)))]; - tensor encoder_step_to_fp16 = cast(dtype = encoder_step_to_fp16_dtype_0, x = encoder_step)[name = tensor("cast_5")]; - tensor input_1_cast_fp16 = transpose(perm = input_1_perm_0, x = encoder_step_to_fp16)[name = tensor("transpose_1")]; - tensor linear_0_cast_fp16 = linear(bias = joint_enc_bias_to_fp16, weight = joint_enc_weight_to_fp16, x = input_1_cast_fp16)[name = tensor("linear_0_cast_fp16")]; - tensor f_axes_0 = const()[name = tensor("f_axes_0"), val = tensor([2])]; - tensor f_cast_fp16 = expand_dims(axes = f_axes_0, x = linear_0_cast_fp16)[name = tensor("f_cast_fp16")]; - tensor joint_pred_weight_to_fp16 = const()[name = tensor("joint_pred_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(656832)))]; - tensor joint_pred_bias_to_fp16 = const()[name = tensor("joint_pred_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1476096)))]; - tensor decoder_step_to_fp16 = cast(dtype = decoder_step_to_fp16_dtype_0, x = decoder_step)[name = tensor("cast_4")]; - tensor input_3_cast_fp16 = transpose(perm = input_3_perm_0, x = decoder_step_to_fp16)[name = tensor("transpose_0")]; - tensor linear_1_cast_fp16 = linear(bias = joint_pred_bias_to_fp16, weight = joint_pred_weight_to_fp16, x = input_3_cast_fp16)[name = tensor("linear_1_cast_fp16")]; - tensor g_axes_0 = const()[name = tensor("g_axes_0"), val = tensor([1])]; - tensor g_cast_fp16 = expand_dims(axes = g_axes_0, x = linear_1_cast_fp16)[name = tensor("g_cast_fp16")]; - tensor input_5_cast_fp16 = add(x = f_cast_fp16, y = g_cast_fp16)[name = tensor("input_5_cast_fp16")]; - tensor input_7_cast_fp16 = relu(x = input_5_cast_fp16)[name = tensor("input_7_cast_fp16")]; - tensor joint_joint_net_2_weight_to_fp16 = const()[name = tensor("joint_joint_net_2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1477440)))]; - tensor joint_joint_net_2_bias_to_fp16 = const()[name = tensor("joint_joint_net_2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2795904)))]; - tensor linear_2_cast_fp16 = linear(bias = joint_joint_net_2_bias_to_fp16, weight = joint_joint_net_2_weight_to_fp16, x = input_7_cast_fp16)[name = tensor("linear_2_cast_fp16")]; - tensor var_35 = const()[name = tensor("op_35"), val = tensor(-1)]; - tensor joint_out_softmax_cast_fp16 = softmax(axis = var_35, x = linear_2_cast_fp16)[name = tensor("joint_out_softmax_cast_fp16")]; - tensor joint_out_epsilon_0 = const()[name = tensor("joint_out_epsilon_0"), val = tensor(0x1p-149)]; - tensor joint_out_cast_fp16 = log(epsilon = joint_out_epsilon_0, x = joint_out_softmax_cast_fp16)[name = tensor("joint_out_cast_fp16")]; - tensor input_begin_0 = const()[name = tensor("input_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor input_end_0 = const()[name = tensor("input_end_0"), val = tensor([1, 1, 1, 1025])]; - tensor input_end_mask_0 = const()[name = tensor("input_end_mask_0"), val = tensor([true, true, true, false])]; - tensor input_cast_fp16 = slice_by_index(begin = input_begin_0, end = input_end_0, end_mask = input_end_mask_0, x = joint_out_cast_fp16)[name = tensor("input_cast_fp16")]; - tensor duration_logits_begin_0 = const()[name = tensor("duration_logits_begin_0"), val = tensor([0, 0, 0, 1025])]; - tensor duration_logits_end_0 = const()[name = tensor("duration_logits_end_0"), val = tensor([1, 1, 1, 1030])]; - tensor duration_logits_end_mask_0 = const()[name = tensor("duration_logits_end_mask_0"), val = tensor([true, true, true, true])]; - tensor duration_logits_cast_fp16 = slice_by_index(begin = duration_logits_begin_0, end = duration_logits_end_0, end_mask = duration_logits_end_mask_0, x = joint_out_cast_fp16)[name = tensor("duration_logits_cast_fp16")]; - tensor var_48 = const()[name = tensor("op_48"), val = tensor(-1)]; - tensor probs_cast_fp16 = softmax(axis = var_48, x = input_cast_fp16)[name = tensor("probs_cast_fp16")]; - tensor token_id_1_axis_0 = const()[name = tensor("token_id_1_axis_0"), val = tensor(-1)]; - tensor token_id_1_keep_dims_0 = const()[name = tensor("token_id_1_keep_dims_0"), val = tensor(true)]; - tensor token_id_1_output_dtype_0 = const()[name = tensor("token_id_1_output_dtype_0"), val = tensor("int32")]; - tensor token_id_1_cast_fp16 = reduce_argmax(axis = token_id_1_axis_0, keep_dims = token_id_1_keep_dims_0, output_dtype = token_id_1_output_dtype_0, x = input_cast_fp16)[name = tensor("token_id_1_cast_fp16")]; - tensor token_id_axes_0 = const()[name = tensor("token_id_axes_0"), val = tensor([-1])]; - tensor token_id_1_cast_fp16_to_int16_dtype_0 = const()[name = tensor("token_id_1_cast_fp16_to_int16_dtype_0"), val = tensor("int16")]; - tensor token_id_1_cast_fp16_to_int16 = cast(dtype = token_id_1_cast_fp16_to_int16_dtype_0, x = token_id_1_cast_fp16)[name = tensor("cast_3")]; - tensor token_id_cast_int16 = squeeze(axes = token_id_axes_0, x = token_id_1_cast_fp16_to_int16)[name = tensor("token_id_cast_int16")]; - tensor token_id_cast_int16_to_int32_dtype_0 = const()[name = tensor("token_id_cast_int16_to_int32_dtype_0"), val = tensor("int32")]; - tensor var_57_axes_0 = const()[name = tensor("op_57_axes_0"), val = tensor([-1])]; - tensor token_id = cast(dtype = token_id_cast_int16_to_int32_dtype_0, x = token_id_cast_int16)[name = tensor("cast_2")]; - tensor var_57 = expand_dims(axes = var_57_axes_0, x = token_id)[name = tensor("op_57")]; - tensor var_58 = const()[name = tensor("op_58"), val = tensor(-1)]; - tensor token_prob_validate_indices_0 = const()[name = tensor("token_prob_validate_indices_0"), val = tensor(false)]; - tensor var_57_to_int16_dtype_0 = const()[name = tensor("op_57_to_int16_dtype_0"), val = tensor("int16")]; - tensor var_57_to_int16 = cast(dtype = var_57_to_int16_dtype_0, x = var_57)[name = tensor("cast_1")]; - tensor token_prob_cast_fp16_cast_int16 = gather_along_axis(axis = var_58, indices = var_57_to_int16, validate_indices = token_prob_validate_indices_0, x = probs_cast_fp16)[name = tensor("token_prob_cast_fp16_cast_int16")]; - tensor var_62_axes_0 = const()[name = tensor("op_62_axes_0"), val = tensor([-1])]; - tensor token_prob = squeeze(axes = var_62_axes_0, x = token_prob_cast_fp16_cast_int16)[name = tensor("op_62_cast_fp16")]; - tensor duration_axis_0 = const()[name = tensor("duration_axis_0"), val = tensor(-1)]; - tensor duration_keep_dims_0 = const()[name = tensor("duration_keep_dims_0"), val = tensor(false)]; - tensor duration_output_dtype_0 = const()[name = tensor("duration_output_dtype_0"), val = tensor("int32")]; - tensor duration = reduce_argmax(axis = duration_axis_0, keep_dims = duration_keep_dims_0, output_dtype = duration_output_dtype_0, x = duration_logits_cast_fp16)[name = tensor("duration_cast_fp16")]; - } -> (token_id, token_prob, duration); -} \ No newline at end of file diff --git a/convert/parakeet-tdt-ctc-110m/output/JointDecision.mlmodelc/weights/weight.bin b/convert/parakeet-tdt-ctc-110m/output/JointDecision.mlmodelc/weights/weight.bin deleted file mode 100644 index 747ec9d24e2d3e008bc869d2c404bb3d2653325c..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/output/JointDecision.mlmodelc/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:b3f771cb65b190f1873e39629676ed79b65a8361522f451b37bdba8b1106e6ff -size 2798028 diff --git a/convert/parakeet-tdt-ctc-110m/output/JointDecision.mlpackage/Data/com.apple.CoreML/model.mlmodel b/convert/parakeet-tdt-ctc-110m/output/JointDecision.mlpackage/Data/com.apple.CoreML/model.mlmodel deleted file mode 100644 index 00748a425eda1d051f13fc524fcfc160d1cb7ca1..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/output/JointDecision.mlpackage/Data/com.apple.CoreML/model.mlmodel +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:73d76247715622313f066e93874f5eff0706410e8572e1b3c51f6f7f2b14a78b -size 9790 diff --git a/convert/parakeet-tdt-ctc-110m/output/JointDecision.mlpackage/Data/com.apple.CoreML/weights/weight.bin b/convert/parakeet-tdt-ctc-110m/output/JointDecision.mlpackage/Data/com.apple.CoreML/weights/weight.bin deleted file mode 100644 index 747ec9d24e2d3e008bc869d2c404bb3d2653325c..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/output/JointDecision.mlpackage/Data/com.apple.CoreML/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:b3f771cb65b190f1873e39629676ed79b65a8361522f451b37bdba8b1106e6ff -size 2798028 diff --git a/convert/parakeet-tdt-ctc-110m/output/JointDecision.mlpackage/Manifest.json b/convert/parakeet-tdt-ctc-110m/output/JointDecision.mlpackage/Manifest.json deleted file mode 100644 index 51edbc5622d98d269493af788ab212f926544b0c..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-ctc-110m/output/JointDecision.mlpackage/Manifest.json +++ /dev/null @@ -1,18 +0,0 @@ -{ - "fileFormatVersion": "1.0.0", - "itemInfoEntries": { - "601AC192-A217-4675-8F63-01F426D7BC52": { - "author": "com.apple.CoreML", - "description": "CoreML Model Weights", - "name": "weights", - "path": "com.apple.CoreML/weights" - }, - "76B2BF33-9547-4031-8668-01C0E473C306": { - "author": "com.apple.CoreML", - "description": "CoreML Model Specification", - "name": "model.mlmodel", - "path": "com.apple.CoreML/model.mlmodel" - } - }, - "rootModelIdentifier": "76B2BF33-9547-4031-8668-01C0E473C306" -} diff --git a/convert/parakeet-tdt-v2-0.6b/.DS_Store b/convert/parakeet-tdt-v2-0.6b/.DS_Store deleted file mode 100644 index 4a654c3e785ab1ee5f435f3eda3c58ea35b780d4..0000000000000000000000000000000000000000 Binary files a/convert/parakeet-tdt-v2-0.6b/.DS_Store and /dev/null differ diff --git a/convert/parakeet-tdt-v2-0.6b/coreml/.gitignore b/convert/parakeet-tdt-v2-0.6b/coreml/.gitignore deleted file mode 100644 index e2e12b1c4fd2deeef4b38477dccac90f560a627c..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-v2-0.6b/coreml/.gitignore +++ /dev/null @@ -1,4 +0,0 @@ -mlpackages/ -parakeet_coreml/ -parakeet_coreml_quantized/ -compiled/ diff --git a/convert/parakeet-tdt-v2-0.6b/coreml/README.md b/convert/parakeet-tdt-v2-0.6b/coreml/README.md deleted file mode 100644 index 8ef855d29d09b61830ca71b880c99e953f44f667..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-v2-0.6b/coreml/README.md +++ /dev/null @@ -1,136 +0,0 @@ -# Parakeet‑TDT v2 (0.6B) — CoreML Export, Parity, and Quantization - -Tools to export NVIDIA Parakeet‑TDT v2 (0.6B) RNNT ASR to CoreML, validate numerical parity with the NeMo reference, measure latency, and explore quantization trade‑offs. All CoreML components use a fixed 15‑second audio window for export and validation. - -## Environment - -1. Create or reuse the local environment with `uv venv`. -2. Activate the repo `.venv` and install deps via `uv pip sync`. -3. Run everything through `uv run` to keep resolutions reproducible. - -## Test Environment - -All tests and measurements referenced here were run on an Apple M4 Pro with 48 GB of RAM. - -## Export CoreML packages - -Exports preprocessor, encoder, decoder, joint, and two fused variants (mel+encoder, joint+decision). Shapes and I/O match the fixed 15‑second window contract. - -``` -uv run python convert-parakeet.py convert \ - --nemo-path /path/to/parakeet-tdt-0.6b-v2.nemo \ - --output-dir parakeet_coreml -``` - -Notes -- Minimum deployment target: iOS 17. Export uses CPU_ONLY by default; runtime compute units can be set when loading the model (Python or Swift). -- Audio is 16 kHz, single‑channel. The 15 s window is enforced during export and validation. - -## Validate parity and speed (Torch vs CoreML) - -Runs Torch and CoreML side‑by‑side on the same 15 s input, records diffs and latency, and saves plots under `plots/compare-components/`. The tool updates `parakeet_coreml/metadata.json` with all measurements. - -``` -uv run python compare-components.py compare \ - --output-dir parakeet_coreml \ - --model-id nvidia/parakeet-tdt-0.6b-v2 \ - --runs 10 --warmup 3 -``` - -Output comparison: - -![./plots/compare-components/mel_encoder_time_l2.png](./plots/compare-components/mel_encoder_time_l2.png) -![./plots/compare-components/joint_decision_prob_u0.png](./plots/compare-components/joint_decision_prob_u0.png) -![./plots/compare-components/decoder_steps_l2.png](./plots/compare-components/decoder_steps_l2.png) - -Latency: - -![./plots/quantize/all/all_components_compile.png](./plots/quantize/all/all_components_compile.png) -![./plots/quantize/all/all_components_compression.png](./plots/quantize/all/all_components_compression.png) -![./plots/quantize/all/all_components_quality.png](./plots/quantize/all/all_components_quality.png) -![./plots/quantize/all/all_components_latency.png](./plots/quantize/all/all_components_latency.png) - - -Quants: - - - -### Key results (quality first) - -Numerical parity is strong across components on the fixed window: -- Preprocessor mel: match=true; max_abs≈0.484, max_rel≈2.00 (near‑zero bins inflate relative error). -- Encoder: match=true; max_abs≈0.0054, strong agreement over time (see plot). -- Decoder h/c state: match=true; value deltas within tolerance. -- Joint logits: match=true; max_abs≈0.099, distributions align (see top‑k plot). -- Joint+Decision: Fused CoreML head exactly matches decisions computed on CoreML logits (token_id/prob/duration). PyTorch logits produce slightly different argmax paths (expected from small logit differences). - -### Speed (latency and RTF) - -Component latency on a 15 s clip, Torch CPU vs CoreML (CPU+NE) from `parakeet_coreml/metadata.json`: -- Encoder: Torch 1030.48 ms → CoreML 25.44 ms (≈40.5× faster, RTF 0.00170) -- Preprocessor: 1.99 ms → 1.19 ms (≈1.68×) -- Joint: 28.34 ms → 22.66 ms (≈1.25×) -- Decoder (U=1): 7.51 ms → 4.32 ms (≈1.73×) - -Fused paths: -- Mel+Encoder (Torch separate vs CoreML fused): 1032.48 ms → 27.10 ms (≈38.1× faster) -- Joint+Decision (CoreML joint + CPU post vs fused CoreML head): 50.05 ms → 64.09 ms (fused is slower here; prefer CoreML joint + lightweight CPU decision on host). - -Plots -- Latency bars and speedups: `plots/compare-components/latency_summary.png`, `plots/compare-components/latency_speedup.png` -- Fused vs separate: `plots/compare-components/latency_fused_vs_separate.png`, `plots/compare-components/latency_fused_speedup.png` -- Quality visuals: mel composite (`mel_composite.png`), encoder L2 over time (`encoder_time_l2.png`), decoder step L2 (`decoder_steps_l2.png`), joint top‑k/time L2 (`joint_top50.png`, `joint_time_l2.png`), joint‑decision agreement (`joint_decision_token_agree.png`, `joint_decision_prob_u0.png`). - -## Quantization (size ‱ quality ‱ speed) - -`uv run python quantize_coreml.py` evaluates several variants and writes a roll‑up to `parakeet_coreml_quantized/quantization_summary.json`. Plots are mirrored to `plots/quantize//` (we include `plots/quantize/all/`). Quality here is reported as 1 − normalized L2 error (1.0 = identical). For JointDecision we report token‑id match rate, duration match, and token‑prob MAE. - -Quick highlights (ComputeUnits=ALL): -- int8 linear (per‑channel): ~2.0× smaller across components with minimal quality loss - - MelEncoder quality≈0.963; latency≈31.13 ms (baseline≈29.34 ms) - - JointDecision acc≈0.995; latency≈1.96 ms (baseline≈2.15 ms) -- int8 linear (per‑tensor symmetric): large encoder quality drop (≈0.50) — not recommended - -Quantization plots (ALL) -- Fused: `plots/quantize/all/fused_quality.png`, `fused_latency.png`, `fused_compression.png`, `fused_size.png` -- Component breakdown: `plots/quantize/all/all_components_quality.png`, `all_components_latency.png`, `all_components_compression.png`, `all_components_size.png`, `all_components_compile.png` - -## Reproduce the figures - -1) Export baseline CoreML packages -``` -uv run python convert-parakeet.py convert --model-id nvidia/parakeet-tdt-0.6b-v2 --output-dir parakeet_coreml -``` - -2) Compare Torch vs CoreML and generate parity/latency plots -``` -uv run python compare-components.py compare --output-dir parakeet_coreml --runs 10 --warmup 3 -``` - -3) Run quantization sweeps (mirrors plots into `plots/quantize//`) -``` -uv run python quantize_coreml.py \ - --input-dir parakeet_coreml \ - --output-root parakeet_coreml_quantized \ - --compute-units ALL --runs 10 -``` - -Examples -- Encoder 6‑bit palette only: - `uv run python quantize_coreml.py -c encoder-palettize` -- MelEncoder 6‑bit palette only: - `uv run python quantize_coreml.py -c mel-palettize` - (By default, the script derives the component whitelist from the selected - variants. Use `-m/--component` to explicitly restrict or `-m all` to force all.) - -## Notes & limits - -- Fixed 15‑second window shapes are required for all CoreML exports and validations. -- Latency measurements are host‑side CoreML predictions (CPU+NE or ALL); on‑device results can differ by chip/OS. -- For streaming decode, the exported decoder uses U=1 inputs with explicit LSTM state I/O. -- Minimum deployment target is iOS 17; models are saved as MLProgram and eligible for ANE when loaded with `ComputeUnits=ALL`. - -## Acknowledgements - -- Parakeet‑TDT v2 model from NVIDIA NeMo (`nvidia/parakeet-tdt-0.6b-v2`). -- This directory provides export/validation utilities and plots to help the community reproduce quality and performance on Apple devices. diff --git a/convert/parakeet-tdt-v2-0.6b/coreml/agents.md b/convert/parakeet-tdt-v2-0.6b/coreml/agents.md deleted file mode 100644 index d6765848d59f76846549712d8c31a4fce733a98a..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-v2-0.6b/coreml/agents.md +++ /dev/null @@ -1,7 +0,0 @@ -# Agent Notes - -- Preferred Python workflow uses `uv` (https://github.com/astral-sh/uv). - - Create and manage environments with `uv venv`. - - Install dependencies with `uv pip install` or `uv pip sync` as needed. -- When working in this repo, activate the local `.venv` and run tooling through `uv run` to keep resolutions reproducible. -- Keep CoreML conversions constrained to the fixed 15-second audio window when exporting or validating Parakeet components. diff --git a/convert/parakeet-tdt-v2-0.6b/coreml/audio/yc_first_minute.wav b/convert/parakeet-tdt-v2-0.6b/coreml/audio/yc_first_minute.wav deleted file mode 100644 index 8682a32b66200c635a60706e722c2b03b5701db0..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-v2-0.6b/coreml/audio/yc_first_minute.wav +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:004afd9a8b6265d321b1faeb3b22ca5efcc78b570566fca633327d73f545f3fa -size 10584142 diff --git a/convert/parakeet-tdt-v2-0.6b/coreml/audio/yc_first_minute_16k.wav b/convert/parakeet-tdt-v2-0.6b/coreml/audio/yc_first_minute_16k.wav deleted file mode 100644 index de71da9ab403797061b7051b820337a754336481..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-v2-0.6b/coreml/audio/yc_first_minute_16k.wav +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:555404115420c8b4a77fcfacc450ecfc767211ca7b449a9dcb292b1734256103 -size 1920090 diff --git a/convert/parakeet-tdt-v2-0.6b/coreml/audio/yc_first_minute_16k_15s.wav b/convert/parakeet-tdt-v2-0.6b/coreml/audio/yc_first_minute_16k_15s.wav deleted file mode 100644 index 0b8040f5f124d20a1d7e812c576c831a7573eddf..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-v2-0.6b/coreml/audio/yc_first_minute_16k_15s.wav +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:c79c8bc763b4efccb3e12f199ec0a59aa2edc5e9e4d21ca70fde8f36762d4147 -size 480078 diff --git a/convert/parakeet-tdt-v2-0.6b/coreml/compare-components.py b/convert/parakeet-tdt-v2-0.6b/coreml/compare-components.py deleted file mode 100644 index 114b74c388a2562cc2576f1ac331b8cd61dc518e..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-v2-0.6b/coreml/compare-components.py +++ /dev/null @@ -1,958 +0,0 @@ -#!/usr/bin/env python3 -"""Compare Parakeet TDT v2 Torch vs CoreML components on a fixed 15s window. - -Writes numeric diffs to the specified output directory (metadata.json) and -saves plots under a repo-tracked directory: plots//. -""" -from __future__ import annotations - -import json -import time -from dataclasses import dataclass -from pathlib import Path -from typing import Dict, Optional, Tuple - -import coremltools as ct -import numpy as np -import soundfile as sf -import torch -import typer - -import nemo.collections.asr as nemo_asr - -# Optional plotting -try: - import matplotlib - matplotlib.use("Agg") - import matplotlib.pyplot as plt - HAS_MPL = True -except Exception: - HAS_MPL = False - - -@dataclass -class ValidationSettings: - audio_path: Optional[Path] - seconds: float - seed: Optional[int] - rtol: float - atol: float - - -def _compute_length(seconds: float, sample_rate: int) -> int: - return int(round(seconds * sample_rate)) - - -def _prepare_audio( - validation_audio: Optional[Path], - sample_rate: int, - max_samples: int, - seed: Optional[int], -) -> torch.Tensor: - if validation_audio is None: - if seed is not None: - torch.manual_seed(seed) - return torch.randn(1, max_samples, dtype=torch.float32) - - data, sr = sf.read(str(validation_audio), dtype="float32") - if sr != sample_rate: - raise typer.BadParameter( - f"Validation audio sample rate {sr} does not match model rate {sample_rate}" - ) - if data.ndim > 1: - data = data[:, 0] - if data.size == 0: - raise typer.BadParameter("Validation audio is empty") - if data.size < max_samples: - data = np.pad(data, (0, max_samples - data.size)) - elif data.size > max_samples: - data = data[:max_samples] - return torch.from_numpy(data).unsqueeze(0).to(dtype=torch.float32) - - -def _np(x: torch.Tensor, dtype=None) -> np.ndarray: - arr = x.detach().cpu().numpy() - if dtype is not None: - return arr.astype(dtype, copy=False) - return arr - - -def _to_t(x) -> torch.Tensor: - if isinstance(x, torch.Tensor): - return x.detach().cpu() - elif isinstance(x, np.ndarray): - # Ensure a separate tensor (avoid shared memory weirdness) - return torch.from_numpy(np.array(x, copy=True)) - else: - return torch.tensor(x) - - -def _max_diffs(a, b, rtol: float, atol: float) -> Tuple[float, float, bool]: - # Use NumPy for comparisons to avoid invoking the PyTorch C-API in contexts - # where the GIL may not be held (which can trigger PyEval_SaveThread errors). - na = np.array(a, dtype=np.float32, copy=True) - nb = np.array(b, dtype=np.float32, copy=True) - if na.size == 0: - return 0.0, 0.0, True - diff = np.abs(na - nb) - max_abs = float(diff.max()) - denom = np.maximum(np.abs(na), np.abs(nb)) - with np.errstate(divide="ignore", invalid="ignore"): - rel = np.where(denom == 0.0, 0.0, diff / denom) - max_rel = float(rel.max()) - ok = bool(np.allclose(na, nb, rtol=rtol, atol=atol)) - return max_abs, max_rel, ok - - -def _plot_line(x_ref: np.ndarray, x_ml: np.ndarray, title: str, path: Path, also_delta: bool = False): - if not HAS_MPL: - return None - if also_delta: - fig, axes = plt.subplots(2, 1, figsize=(8, 5), sharex=True) - axes[0].plot(x_ref, label="torch", linewidth=1) - axes[0].plot(x_ml, label="coreml", linewidth=1, alpha=0.8) - axes[0].set_title(title) - axes[0].legend() - delta = np.asarray(x_ref) - np.asarray(x_ml) - axes[1].plot(delta, color="C3", linewidth=1) - axes[1].set_title("Delta (torch - coreml)") - axes[1].set_xlabel("time/step") - plt.tight_layout() - plt.savefig(path) - plt.close(fig) - else: - plt.figure(figsize=(8, 3)) - plt.plot(x_ref, label="torch", linewidth=1) - plt.plot(x_ml, label="coreml", linewidth=1, alpha=0.8) - plt.title(title) - plt.legend() - plt.tight_layout() - plt.savefig(path) - plt.close() - return str(path.name) - - -def _plot_image(img: np.ndarray, title: str, path: Path, vmin=None, vmax=None): - if not HAS_MPL: - return None - plt.figure(figsize=(6, 4)) - plt.imshow(img, aspect='auto', origin='lower', interpolation='nearest', vmin=vmin, vmax=vmax) - plt.title(title) - plt.colorbar(shrink=0.8) - plt.tight_layout() - plt.savefig(path) - plt.close() - return str(path.name) - - -def _plot_mel_composite( - mel_torch: np.ndarray, - mel_coreml: np.ndarray, - path: Path, - vmin=None, - vmax=None, -): - """Create a single PNG with mel torch, mel coreml, abs diff heatmap, and mean-over-time curves with delta.""" - if not HAS_MPL: - return None - mel_torch = np.asarray(mel_torch) - mel_coreml = np.asarray(mel_coreml) - absdiff = np.abs(mel_torch - mel_coreml) - mean_t = mel_torch.mean(axis=0) - mean_c = mel_coreml.mean(axis=0) - delta = mean_t - mean_c - - fig = plt.figure(figsize=(12, 8)) - gs = fig.add_gridspec(2, 2, height_ratios=[1, 1]) - ax1 = fig.add_subplot(gs[0, 0]) - im1 = ax1.imshow(mel_torch, aspect='auto', origin='lower', interpolation='nearest', vmin=vmin, vmax=vmax) - ax1.set_title("Mel (Torch)") - fig.colorbar(im1, ax=ax1, shrink=0.8) - - ax2 = fig.add_subplot(gs[0, 1]) - im2 = ax2.imshow(mel_coreml, aspect='auto', origin='lower', interpolation='nearest', vmin=vmin, vmax=vmax) - ax2.set_title("Mel (CoreML)") - fig.colorbar(im2, ax=ax2, shrink=0.8) - - ax3 = fig.add_subplot(gs[1, 0]) - im3 = ax3.imshow(absdiff, aspect='auto', origin='lower', interpolation='nearest') - ax3.set_title("Mel |diff|") - fig.colorbar(im3, ax=ax3, shrink=0.8) - - ax4 = fig.add_subplot(gs[1, 1]) - ax4.plot(mean_t, label="torch", linewidth=1) - ax4.plot(mean_c, label="coreml", linewidth=1, alpha=0.8) - ax4.plot(delta, label="delta", linewidth=1, color="C3") - ax4.set_title("Mel mean over time + delta") - ax4.legend() - - plt.tight_layout() - plt.savefig(path) - plt.close(fig) - return str(path.name) - - -def _plot_latency_bars( - labels, - torch_means, - torch_stds, - coreml_means, - coreml_stds, - path: Path, -): - if not HAS_MPL: - return None - x = np.arange(len(labels)) - width = 0.35 - fig, ax = plt.subplots(figsize=(8, 4)) - b1 = ax.bar(x - width/2, torch_means, width, yerr=torch_stds, label="torch", color="C0", alpha=0.9) - b2 = ax.bar(x + width/2, coreml_means, width, yerr=coreml_stds, label="coreml", color="C1", alpha=0.9) - ax.set_xticks(x, labels, rotation=15) - ax.set_ylabel("latency (ms)") - ax.set_title("Component latency (15s window inputs)") - ax.legend() - # Add value labels on bars - def _annotate(bars): - for bar in bars: - h = bar.get_height() - if np.isnan(h): - continue - ax.annotate(f"{h:.0f}", - xy=(bar.get_x() + bar.get_width()/2, h), - xytext=(0, 3), textcoords="offset points", - ha='center', va='bottom', fontsize=8) - _annotate(b1) - _annotate(b2) - plt.tight_layout() - plt.savefig(path) - plt.close(fig) - return str(path.name) - - -def _plot_speedup_bars(labels, torch_means, coreml_means, path: Path): - if not HAS_MPL: - return None - speedup = [] - for t, c in zip(torch_means, coreml_means): - if c and c > 0: - speedup.append(float(t) / float(c)) - else: - speedup.append(np.nan) - x = np.arange(len(labels)) - fig, ax = plt.subplots(figsize=(8, 4)) - bars = ax.bar(x, speedup, color="C2") - ax.set_xticks(x, labels, rotation=15) - ax.set_ylabel("torch/coreml speedup") - ax.set_title("CoreML speedup vs Torch (higher is better)") - ax.axhline(1.0, color="gray", linestyle="--", linewidth=1) - # Add value labels - for bar in bars: - h = bar.get_height() - if np.isnan(h): - continue - ax.annotate(f"{h:.2f}", - xy=(bar.get_x() + bar.get_width()/2, h), - xytext=(0, 3), textcoords="offset points", - ha='center', va='bottom', fontsize=8) - plt.tight_layout() - plt.savefig(path) - plt.close(fig) - return str(path.name) - - - -app = typer.Typer(add_completion=False, pretty_exceptions_show_locals=False) - - -@app.command() -def compare( - output_dir: Path = typer.Option(Path("parakeet_coreml"), help="Directory containing mlpackages + metadata.json"), - nemo_path: Optional[Path] = typer.Option(None, "--nemo-path", exists=True, resolve_path=True, help="Path to .nemo checkpoint"), - model_id: str = typer.Option("nvidia/parakeet-tdt-0.6b-v2", "--model-id", help="HF model id if --nemo-path omitted"), - validation_audio: Optional[Path] = typer.Option(None, exists=True, resolve_path=True, help="15s, 16kHz wav for validation (defaults to audio/yc_first_minute_16k_15s.wav if present)"), - seed: Optional[int] = typer.Option(None, help="Random seed for synthetic input when audio is not provided"), - rtol: float = typer.Option(1e-3, help="Relative tolerance for comparisons"), - atol: float = typer.Option(1e-4, help="Absolute tolerance for comparisons"), - runs: int = typer.Option(10, help="Timed runs per model for latency measurement"), - warmup: int = typer.Option(3, help="Warmup runs before timing (compilation, caches)"), - symbol_steps: int = typer.Option( - 32, - help="Number of sequential decoder steps to validate with streaming U=1 inputs", - ), -) -> None: - """Run Torch vs CoreML comparisons and update metadata.json with plots and diffs.""" - output_dir.mkdir(parents=True, exist_ok=True) - if symbol_steps < 1: - raise typer.BadParameter("symbol_steps must be >= 1") - - meta_path = output_dir / "metadata.json" - exported_meta: Dict[str, object] = {} - if meta_path.exists(): - try: - exported_meta = json.loads(meta_path.read_text()) - except Exception: - exported_meta = {} - exported_max_u = int(exported_meta.get("max_symbol_steps", 1)) - if exported_max_u != 1: - typer.echo( - f"Note: CoreML export reports max_symbol_steps={exported_max_u}; " - "comparison still drives decoder step-wise with U=1 inputs." - ) - if nemo_path is not None: - typer.echo(f"Loading NeMo model from {nemo_path}
") - asr_model = nemo_asr.models.EncDecRNNTBPEModel.restore_from(str(nemo_path), map_location="cpu") - else: - typer.echo(f"Downloading NeMo model via {model_id}
") - asr_model = nemo_asr.models.EncDecRNNTBPEModel.from_pretrained(model_id, map_location="cpu") - asr_model.eval() - - sample_rate = int(asr_model.cfg.preprocessor.sample_rate) - max_samples = _compute_length(15.0, sample_rate) - default_audio = (Path(__file__).parent / "audio" / "yc_first_minute_16k_15s.wav").resolve() - chosen_audio = validation_audio if validation_audio is not None else (default_audio if default_audio.exists() else None) - if chosen_audio is not None and validation_audio is None: - typer.echo(f"Using default validation audio: {chosen_audio}") - - audio_tensor = _prepare_audio(chosen_audio, sample_rate, max_samples, seed) - audio_length = torch.tensor([max_samples], dtype=torch.int32) - - asr_model.decoder._rnnt_export = True - # Disable fused loss/WER computation for simpler joint inference - asr_model.joint.set_fuse_loss_wer(False) - # Important: ensure the joint returns raw logits (not log-softmax) - # RNNTJoint applies log_softmax on CPU by default when `log_softmax is None`. - # Our exported CoreML joint emits pre-softmax logits, so make the Torch - # reference do the same to avoid systematic offsets in comparisons/plots. - try: - # Some versions expose this as a plain attribute - asr_model.joint.log_softmax = False - except Exception: - pass - - # Generate reference outputs directly from NeMo model components - with torch.inference_mode(): - # Preprocessor - direct NeMo call - mel_ref, mel_length_ref = asr_model.preprocessor( - input_signal=audio_tensor, - length=audio_length.to(dtype=torch.long) - ) - mel_length_ref = mel_length_ref.to(dtype=torch.int32) - - # Encoder - direct NeMo call - encoder_ref, encoder_length_ref = asr_model.encoder( - audio_signal=mel_ref, - length=mel_length_ref.to(dtype=torch.long) - ) - encoder_length_ref = encoder_length_ref.to(dtype=torch.int32) - - vocab_size = int(asr_model.tokenizer.vocab_size) - num_extra = int(asr_model.joint.num_extra_outputs) - decoder_hidden = int(asr_model.decoder.pred_hidden) - decoder_layers = int(asr_model.decoder.pred_rnn_layers) - blank_id = int(asr_model.decoder.blank_idx) - - blank_targets = torch.tensor([[blank_id]], dtype=torch.int32) - blank_target_lengths = torch.tensor([1], dtype=torch.int32) - blank_targets_long = blank_targets.to(dtype=torch.long) - blank_target_lengths_long = blank_target_lengths.to(dtype=torch.long) - - def _decoder_rollout_torch(num_steps: int) -> Tuple[torch.Tensor, torch.Tensor, torch.Tensor]: - outputs = [] - h_state = torch.zeros(decoder_layers, 1, decoder_hidden, dtype=torch.float32) - c_state = torch.zeros(decoder_layers, 1, decoder_hidden, dtype=torch.float32) - state = [h_state, c_state] - with torch.inference_mode(): - for _ in range(num_steps): - y, _, new_state = asr_model.decoder( - targets=blank_targets_long, - target_length=blank_target_lengths_long, - states=state, - ) - outputs.append(y.detach()) - state = [new_state[0].detach(), new_state[1].detach()] - if outputs: - decoder_seq = torch.cat(outputs, dim=-1) - else: - decoder_seq = torch.zeros(1, decoder_hidden, 0, dtype=torch.float32) - return decoder_seq, state[0], state[1] - - decoder_ref, h_ref, c_ref = _decoder_rollout_torch(symbol_steps) - - with torch.inference_mode(): - logits_ref = asr_model.joint( - encoder_outputs=encoder_ref, - decoder_outputs=decoder_ref, - ) - - # Convert tensors to numpy for CoreML - def _np32(x): - return np.array(x.detach().cpu().numpy(), dtype=np.float32, copy=True) - - # Prepare plot dir (write to repo-tracked plots//) - plots_root = Path(__file__).parent / "plots" - plots_dir = plots_root / Path(__file__).stem - plots_dir.mkdir(parents=True, exist_ok=True) - - encoder_np = _np32(encoder_ref) - decoder_ref_np = _np32(decoder_ref) - - summary: Dict[str, object] = { - "requested": True, - "status": "ok", - "atol": atol, - "rtol": rtol, - "symbol_steps": int(symbol_steps), - "audio_path": None if validation_audio is None else str(validation_audio), - "components": {}, - } - - # Preprocessor - pre = ct.models.MLModel(str(output_dir / "parakeet_preprocessor.mlpackage"), compute_units=ct.ComputeUnit.CPU_AND_NE) - t0 = time.perf_counter() - pre_out = pre.predict({"audio_signal": _np32(audio_tensor), "audio_length": _np32(audio_length).astype(np.int32)}) - t1 = time.perf_counter() - pre_first_ms = (t1 - t0) * 1000.0 - mel_ml = np.array(pre_out["mel"], dtype=np.float32, copy=True) - mel_len_ml = np.array(pre_out["mel_length"], dtype=np.int32, copy=True) - pre_atol, pre_rtol = max(atol, 1.0), max(rtol, 1e-2) - a_mel, r_mel, ok_mel = _max_diffs(_np32(mel_ref), mel_ml, pre_rtol, pre_atol) - ok_len = int(_np32(mel_length_ref).astype(np.int32)[0]) == int(np.array(mel_len_ml).astype(np.int32)[0]) - mel_t = _np32(mel_ref)[0] - mel_c = mel_ml[0] - vmin = float(min(mel_t.min(), mel_c.min())) - vmax = float(max(mel_t.max(), mel_c.max())) - pre_plots = { - "mel_composite.png": _plot_mel_composite(mel_t, mel_c, plots_dir / "mel_composite.png", vmin=vmin, vmax=vmax), - } - # Latency measurements: Torch and CoreML - def _time_coreml(model: ct.models.MLModel, inputs: Dict[str, np.ndarray]) -> Tuple[float, float]: - # Warmup - for _ in range(max(0, warmup)): - _ = model.predict(inputs) - times = [] - for _ in range(max(1, runs)): - t0 = time.perf_counter() - _ = model.predict(inputs) - t1 = time.perf_counter() - times.append((t1 - t0) * 1000.0) - arr = np.array(times, dtype=np.float64) - return float(arr.mean()), float(arr.std(ddof=1) if arr.size > 1 else 0.0) - - def _time_torch(fn, *args, **kwargs) -> Tuple[float, float]: - with torch.inference_mode(): - for _ in range(max(0, warmup)): - _ = fn(*args, **kwargs) - times = [] - for _ in range(max(1, runs)): - t0 = time.perf_counter() - _ = fn(*args, **kwargs) - t1 = time.perf_counter() - times.append((t1 - t0) * 1000.0) - arr = np.array(times, dtype=np.float64) - return float(arr.mean()), float(arr.std(ddof=1) if arr.size > 1 else 0.0) - - pre_torch_ms_mean, pre_torch_ms_std = _time_torch( - asr_model.preprocessor, input_signal=audio_tensor, length=audio_length.to(dtype=torch.long) - ) - pre_coreml_ms_mean, pre_coreml_ms_std = _time_coreml( - pre, - {"audio_signal": _np32(audio_tensor), "audio_length": _np32(audio_length).astype(np.int32)}, - ) - seconds = 15.0 - pre_coreml_rtf = float(pre_coreml_ms_mean / (seconds * 1000.0)) if pre_coreml_ms_mean > 0 else None - pre_torch_rtf = float(pre_torch_ms_mean / (seconds * 1000.0)) if pre_torch_ms_mean > 0 else None - - summary["components"]["preprocessor"] = { - "mel": {"max_abs": a_mel, "max_rel": r_mel, "match": bool(ok_mel)}, - "length_match": bool(ok_len), - "latency": { - "runs": int(runs), - "warmup": int(warmup), - "coreml_first_ms": pre_first_ms, - "torch_ms": {"mean": pre_torch_ms_mean, "std": pre_torch_ms_std}, - "coreml_ms": {"mean": pre_coreml_ms_mean, "std": pre_coreml_ms_std}, - "rtf": {"torch": pre_torch_rtf, "coreml": pre_coreml_rtf}, - }, - "plots": {k: v for k, v in pre_plots.items() if v}, - } - - # Encoder - enc = ct.models.MLModel(str(output_dir / "parakeet_encoder.mlpackage"), compute_units=ct.ComputeUnit.CPU_AND_NE) - t0 = time.perf_counter() - enc_out = enc.predict({"mel": _np32(mel_ref), "mel_length": _np32(mel_length_ref).astype(np.int32)}) - t1 = time.perf_counter() - enc_first_ms = (t1 - t0) * 1000.0 - enc_ml = np.array(enc_out["encoder"], dtype=np.float32, copy=True) - enc_len_ml = np.array(enc_out["encoder_length"], dtype=np.int32, copy=True) - a_enc, r_enc, ok_enc = _max_diffs(_np32(encoder_ref), enc_ml, max(rtol, 5e-3), max(atol, 5e-2)) - ok_enc_len = int(_np32(encoder_length_ref).astype(np.int32)[0]) == int(np.array(enc_len_ml).astype(np.int32)[0]) - enc_t = _np32(encoder_ref)[0] - enc_c = enc_ml[0] - enc_plots = { - "encoder_time_l2.png": _plot_line( - np.linalg.norm(enc_t, axis=0), # L2 norm over features (D) for each time step - np.linalg.norm(enc_c, axis=0), # enc_t shape is (D, T), so axis=0 is features - "Encoder L2 over time", - plots_dir / "encoder_time_l2.png", - also_delta=True, - ), - } - enc_torch_ms_mean, enc_torch_ms_std = _time_torch( - asr_model.encoder, audio_signal=mel_ref, length=mel_length_ref.to(dtype=torch.long) - ) - enc_coreml_ms_mean, enc_coreml_ms_std = _time_coreml( - enc, {"mel": _np32(mel_ref), "mel_length": _np32(mel_length_ref).astype(np.int32)} - ) - enc_coreml_rtf = float(enc_coreml_ms_mean / (seconds * 1000.0)) if enc_coreml_ms_mean > 0 else None - enc_torch_rtf = float(enc_torch_ms_mean / (seconds * 1000.0)) if enc_torch_ms_mean > 0 else None - - summary["components"]["encoder"] = { - "encoder": {"max_abs": a_enc, "max_rel": r_enc, "match": bool(ok_enc)}, - "length_match": bool(ok_enc_len), - "latency": { - "runs": int(runs), - "warmup": int(warmup), - "coreml_first_ms": enc_first_ms, - "torch_ms": {"mean": enc_torch_ms_mean, "std": enc_torch_ms_std}, - "coreml_ms": {"mean": enc_coreml_ms_mean, "std": enc_coreml_ms_std}, - "rtf": {"torch": enc_torch_rtf, "coreml": enc_coreml_rtf}, - }, - "plots": {k: v for k, v in enc_plots.items() if v}, - } - - # Decoder (sequential U=1 rollout) - dec = ct.models.MLModel(str(output_dir / "parakeet_decoder.mlpackage"), compute_units=ct.ComputeUnit.CPU_AND_NE) - - zero_state_np = np.zeros((decoder_layers, 1, decoder_hidden), dtype=np.float32) - blank_targets_np = np.array(blank_targets.detach().cpu().numpy(), dtype=np.int32, copy=True) - blank_target_lengths_np = np.array(blank_target_lengths.detach().cpu().numpy(), dtype=np.int32, copy=True) - - def _decoder_rollout_coreml(num_steps: int) -> Tuple[np.ndarray, np.ndarray, np.ndarray, float]: - outputs = [] - h_np = zero_state_np.copy() - c_np = zero_state_np.copy() - first_ms: Optional[float] = None - for i in range(num_steps): - t0_i = time.perf_counter() if i == 0 else None - res = dec.predict( - { - "targets": blank_targets_np, - "target_length": blank_target_lengths_np, - "h_in": h_np, - "c_in": c_np, - } - ) - if t0_i is not None: - t1_i = time.perf_counter() - first_ms = (t1_i - t0_i) * 1000.0 - outputs.append(np.array(res["decoder"], dtype=np.float32, copy=True)) - h_np = np.array(res["h_out"], dtype=np.float32, copy=True) - c_np = np.array(res["c_out"], dtype=np.float32, copy=True) - if outputs: - decoder_seq = np.concatenate(outputs, axis=-1) - else: - decoder_seq = np.zeros((1, decoder_hidden, 0), dtype=np.float32) - return decoder_seq, h_np, c_np, (0.0 if first_ms is None else float(first_ms)) - - dec_ml, h_ml, c_ml, dec_first_ms = _decoder_rollout_coreml(symbol_steps) - h_ref_np = _np32(h_ref) - c_ref_np = _np32(c_ref) - - a_dec, r_dec, ok_dec = _max_diffs(decoder_ref_np, dec_ml, max(rtol, 1e-2), max(atol, 1e-1)) - a_h, r_h, ok_h = _max_diffs(h_ref_np, h_ml, max(rtol, 1e-2), max(atol, 2.5e-1)) - a_c, r_c, ok_c = _max_diffs(c_ref_np, c_ml, max(rtol, 5e-2), max(atol, 1.5e0)) - - dec_t = decoder_ref_np[0] - dec_c = dec_ml[0] - dec_plots = { - "decoder_steps_l2.png": _plot_line( - np.linalg.norm(dec_t, axis=0), - np.linalg.norm(dec_c, axis=0), - "Decoder L2 over steps", - plots_dir / "decoder_steps_l2.png", - also_delta=True, - ), - } - - def _time_decoder_coreml() -> Tuple[float, float]: - for _ in range(max(0, warmup)): - _decoder_rollout_coreml(symbol_steps) - times = [] - for _ in range(max(1, runs)): - t0 = time.perf_counter() - _decoder_rollout_coreml(symbol_steps) - t1 = time.perf_counter() - times.append((t1 - t0) * 1000.0) - arr = np.array(times, dtype=np.float64) - return float(arr.mean()), float(arr.std(ddof=1) if arr.size > 1 else 0.0) - - dec_torch_ms_mean, dec_torch_ms_std = _time_torch(lambda: _decoder_rollout_torch(symbol_steps)) - dec_coreml_ms_mean, dec_coreml_ms_std = _time_decoder_coreml() - dec_coreml_rtf = float(dec_coreml_ms_mean / (seconds * 1000.0)) if dec_coreml_ms_mean > 0 else None - dec_torch_rtf = float(dec_torch_ms_mean / (seconds * 1000.0)) if dec_torch_ms_mean > 0 else None - - summary["components"]["decoder"] = { - "decoder": {"max_abs": a_dec, "max_rel": r_dec, "match": bool(ok_dec)}, - "h_out": {"max_abs": a_h, "max_rel": r_h, "match": bool(ok_h)}, - "c_out": {"max_abs": a_c, "max_rel": r_c, "match": bool(ok_c)}, - "latency": { - "runs": int(runs), - "warmup": int(warmup), - "coreml_first_ms": dec_first_ms, - "torch_ms": {"mean": dec_torch_ms_mean, "std": dec_torch_ms_std}, - "coreml_ms": {"mean": dec_coreml_ms_mean, "std": dec_coreml_ms_std}, - "rtf": {"torch": dec_torch_rtf, "coreml": dec_coreml_rtf}, - }, - "plots": {k: v for k, v in dec_plots.items() if v}, - } - - # Joint (sequential U=1 rollouts) - j = ct.models.MLModel(str(output_dir / "parakeet_joint.mlpackage"), compute_units=ct.ComputeUnit.CPU_AND_NE) - - def _joint_rollout_coreml(decoder_seq_np: np.ndarray) -> Tuple[np.ndarray, float]: - logits_steps = [] - first_ms: Optional[float] = None - for u in range(decoder_seq_np.shape[2]): - dec_slice = decoder_seq_np[:, :, u : u + 1] - t0_u = time.perf_counter() if u == 0 else None - res = j.predict({"encoder": encoder_np, "decoder": dec_slice}) - if t0_u is not None: - t1_u = time.perf_counter() - first_ms = (t1_u - t0_u) * 1000.0 - logits_steps.append(np.array(res["logits"], dtype=np.float32, copy=True)) - if not logits_steps: - raise RuntimeError("No decoder steps provided for joint rollout") - return np.concatenate(logits_steps, axis=2), (0.0 if first_ms is None else float(first_ms)) - - logits_ml, joint_first_ms = _joint_rollout_coreml(decoder_ref_np) - logits_ref_np = _np32(logits_ref) - a_j, r_j, ok_j = _max_diffs(logits_ref_np, logits_ml, max(rtol, 1e-2), max(atol, 1e-1)) - joint_plots = {} - if HAS_MPL: - lt = logits_ref_np[0, 0, 0, :] - lc = logits_ml[0, 0, 0, :] - top_idx = np.argsort(-np.abs(lt))[:50] - path = plots_dir / "joint_top50.png" - plt.figure(figsize=(8, 3)) - plt.plot(lt[top_idx], label="torch") - plt.plot(lc[top_idx], label="coreml", alpha=0.8) - plt.title("Joint logits (t=0,u=0) top-50 |torch|") - plt.legend(); plt.tight_layout(); plt.savefig(path); plt.close() - joint_plots["joint_top50.png"] = str(path.name) - - # Delta-over-time visualization (fix u=0; summarize over vocab) - jt = logits_ref_np[0, :, 0, :] - jc = logits_ml[0, :, 0, :] - l2_t = np.linalg.norm(jt, axis=1) - l2_c = np.linalg.norm(jc, axis=1) - path2 = plots_dir / "joint_time_l2.png" - _plot_line(l2_t, l2_c, "Joint L2 over time (u=0)", path2, also_delta=True) - joint_plots["joint_time_l2.png"] = str(path2.name) - - def _time_joint_coreml() -> Tuple[float, float]: - for _ in range(max(0, warmup)): - _joint_rollout_coreml(decoder_ref_np) - times = [] - for _ in range(max(1, runs)): - t0 = time.perf_counter() - _joint_rollout_coreml(decoder_ref_np) - t1 = time.perf_counter() - times.append((t1 - t0) * 1000.0) - arr = np.array(times, dtype=np.float64) - return float(arr.mean()), float(arr.std(ddof=1) if arr.size > 1 else 0.0) - - joint_torch_ms_mean, joint_torch_ms_std = _time_torch( - asr_model.joint, encoder_outputs=encoder_ref, decoder_outputs=decoder_ref - ) - joint_coreml_ms_mean, joint_coreml_ms_std = _time_joint_coreml() - joint_coreml_rtf = float(joint_coreml_ms_mean / (seconds * 1000.0)) if joint_coreml_ms_mean > 0 else None - joint_torch_rtf = float(joint_torch_ms_mean / (seconds * 1000.0)) if joint_torch_ms_mean > 0 else None - - summary["components"]["joint"] = { - "logits": {"max_abs": a_j, "max_rel": r_j, "match": bool(ok_j)}, - "latency": { - "runs": int(runs), - "warmup": int(warmup), - "coreml_first_ms": joint_first_ms, - "torch_ms": {"mean": joint_torch_ms_mean, "std": joint_torch_ms_std}, - "coreml_ms": {"mean": joint_coreml_ms_mean, "std": joint_coreml_ms_std}, - "rtf": {"torch": joint_torch_rtf, "coreml": joint_coreml_rtf}, - }, - "plots": joint_plots, - } - - # Fused components - # 1) Mel+Encoder fused vs separate - mel_enc_plots = {} - try: - mel_enc = ct.models.MLModel(str(output_dir / "parakeet_mel_encoder.mlpackage"), compute_units=ct.ComputeUnit.CPU_AND_NE) - t0 = time.perf_counter() - mel_enc_out = mel_enc.predict({ - "audio_signal": _np32(audio_tensor), - "audio_length": _np32(audio_length).astype(np.int32), - }) - t1 = time.perf_counter() - mel_enc_first_ms = (t1 - t0) * 1000.0 - mel_enc_ml = np.array(mel_enc_out["encoder"], dtype=np.float32, copy=True) - mel_enc_len_ml = np.array(mel_enc_out["encoder_length"], dtype=np.int32, copy=True) - # Compare fused output vs Torch reference encoder - a_melenc, r_melenc, ok_melenc = _max_diffs(_np32(encoder_ref), mel_enc_ml, max(rtol, 5e-3), max(atol, 5e-2)) - ok_melenc_len = int(_np32(encoder_length_ref).astype(np.int32)[0]) == int(mel_enc_len_ml.astype(np.int32)[0]) - # Also compare fused vs separate CoreML pipeline (pre -> enc) - a_melenc_vs_sep, r_melenc_vs_sep, ok_melenc_vs_sep = _max_diffs(enc_ml, mel_enc_ml, max(rtol, 5e-3), max(atol, 5e-2)) - - # Plots: L2 over time (fused vs torch) - enc_t_ref = _np32(encoder_ref)[0] - enc_c_fused = mel_enc_ml[0] - mel_enc_plots["mel_encoder_time_l2.png"] = _plot_line( - np.linalg.norm(enc_t_ref, axis=0), - np.linalg.norm(enc_c_fused, axis=0), - "Mel+Encoder (fused) L2 over time", - plots_dir / "mel_encoder_time_l2.png", - also_delta=True, - ) - - # Latency: fused CoreML vs separate (CoreML pre + CoreML enc) - mel_enc_coreml_ms_mean, mel_enc_coreml_ms_std = _time_coreml( - mel_enc, - {"audio_signal": _np32(audio_tensor), "audio_length": _np32(audio_length).astype(np.int32)}, - ) - sep_coreml_ms_mean = float(pre_coreml_ms_mean + enc_coreml_ms_mean) - sep_coreml_ms_std = float((pre_coreml_ms_std ** 2 + enc_coreml_ms_std ** 2) ** 0.5) - # Torch baseline (separate torch pre + enc) - sep_torch_ms_mean = float(pre_torch_ms_mean + enc_torch_ms_mean) - sep_torch_ms_std = float((pre_torch_ms_std ** 2 + enc_torch_ms_std ** 2) ** 0.5) - - mel_enc_coreml_rtf = float(mel_enc_coreml_ms_mean / (seconds * 1000.0)) if mel_enc_coreml_ms_mean > 0 else None - sep_coreml_rtf = float(sep_coreml_ms_mean / (seconds * 1000.0)) if sep_coreml_ms_mean > 0 else None - sep_torch_rtf = float(sep_torch_ms_mean / (seconds * 1000.0)) if sep_torch_ms_mean > 0 else None - - summary["components"]["mel_encoder"] = { - "encoder": {"max_abs": a_melenc, "max_rel": r_melenc, "match": bool(ok_melenc)}, - "length_match": bool(ok_melenc_len), - "vs_separate_coreml": {"max_abs": a_melenc_vs_sep, "max_rel": r_melenc_vs_sep, "match": bool(ok_melenc_vs_sep)}, - "latency": { - "runs": int(runs), - "warmup": int(warmup), - "fused_coreml_first_ms": mel_enc_first_ms, - "fused_coreml_ms": {"mean": mel_enc_coreml_ms_mean, "std": mel_enc_coreml_ms_std}, - "separate_coreml_ms": {"mean": sep_coreml_ms_mean, "std": sep_coreml_ms_std}, - "separate_torch_ms": {"mean": sep_torch_ms_mean, "std": sep_torch_ms_std}, - "rtf": {"fused_coreml": mel_enc_coreml_rtf, "separate_coreml": sep_coreml_rtf, "separate_torch": sep_torch_rtf}, - }, - "plots": {k: v for k, v in mel_enc_plots.items() if v}, - } - except Exception as e: - summary["components"]["mel_encoder_error"] = str(e) - - # 2) JointDecision fused vs CPU PyTorch post-processing - jd_plots = {} - try: - # Fused CoreML joint decision - jd = ct.models.MLModel(str(output_dir / "parakeet_joint_decision.mlpackage"), compute_units=ct.ComputeUnit.CPU_AND_NE) - def _joint_decision_rollout_coreml(decoder_seq_np: np.ndarray) -> Tuple[np.ndarray, np.ndarray, np.ndarray, float]: - token_ids = [] - token_probs = [] - durations = [] - first_ms: Optional[float] = None - for u in range(decoder_seq_np.shape[2]): - dec_slice = decoder_seq_np[:, :, u : u + 1] - t0_u = time.perf_counter() if u == 0 else None - res = jd.predict({"encoder": encoder_np, "decoder": dec_slice}) - if t0_u is not None: - t1_u = time.perf_counter() - first_ms = (t1_u - t0_u) * 1000.0 - token_ids.append(np.array(res["token_id"], dtype=np.int32, copy=True)) - token_probs.append(np.array(res["token_prob"], dtype=np.float32, copy=True)) - durations.append(np.array(res["duration"], dtype=np.int32, copy=True)) - if not token_ids: - raise RuntimeError("No decoder steps provided for joint decision rollout") - return ( - np.concatenate(token_ids, axis=2), - np.concatenate(token_probs, axis=2), - np.concatenate(durations, axis=2), - (0.0 if first_ms is None else float(first_ms)), - ) - - token_id_ml, token_prob_ml, duration_ml, jd_first_ms = _joint_decision_rollout_coreml(decoder_ref_np) - - # CPU PyTorch decision using Torch logits - vocab_with_blank = int(vocab_size) + 1 - with torch.inference_mode(): - logits_t = logits_ref - token_logits_t = logits_t[..., :vocab_with_blank] - duration_logits_t = logits_t[..., -num_extra:] if num_extra > 0 else None - token_ids_t = torch.argmax(token_logits_t, dim=-1).to(dtype=torch.int32) - token_probs_all_t = torch.softmax(token_logits_t, dim=-1) - token_prob_t = torch.gather( - token_probs_all_t, dim=-1, index=token_ids_t.long().unsqueeze(-1) - ).squeeze(-1) - if duration_logits_t is not None and duration_logits_t.numel() > 0: - duration_t = torch.argmax(duration_logits_t, dim=-1).to(dtype=torch.int32) - else: - duration_t = torch.zeros_like(token_ids_t, dtype=torch.int32) - - # Also derive CPU decision from CoreML joint logits for "separate" path - token_logits_c = _to_t(logits_ml)[..., :vocab_with_blank] - duration_logits_c = _to_t(logits_ml)[..., -num_extra:] if num_extra > 0 else None - token_ids_c = torch.argmax(token_logits_c, dim=-1).to(dtype=torch.int32) - token_probs_all_c = torch.softmax(token_logits_c, dim=-1) - token_prob_c = torch.gather( - token_probs_all_c, dim=-1, index=token_ids_c.long().unsqueeze(-1) - ).squeeze(-1) - if duration_logits_c is not None and duration_logits_c.numel() > 0: - duration_c = torch.argmax(duration_logits_c, dim=-1).to(dtype=torch.int32) - else: - duration_c = torch.zeros_like(token_ids_c, dtype=torch.int32) - - # Compare fused outputs to CPU PyTorch decisions - a_tid_t, r_tid_t, ok_tid_t = _max_diffs(_np(token_ids_t), token_id_ml, 0.0, 0.0) - a_tprob_t, r_tprob_t, ok_tprob_t = _max_diffs(_np(token_prob_t), token_prob_ml, max(rtol, 1e-2), max(atol, 1e-1)) - a_dur_t, r_dur_t, ok_dur_t = _max_diffs(_np(duration_t), duration_ml, 0.0, 0.0) - - a_tid_c, r_tid_c, ok_tid_c = _max_diffs(_np(token_ids_c), token_id_ml, 0.0, 0.0) - a_tprob_c, r_tprob_c, ok_tprob_c = _max_diffs(_np(token_prob_c), token_prob_ml, max(rtol, 1e-2), max(atol, 1e-1)) - a_dur_c, r_dur_c, ok_dur_c = _max_diffs(_np(duration_c), duration_ml, 0.0, 0.0) - - # Plots: token_prob over time for u=0 (fused vs torch CPU) - if HAS_MPL: - prob_t = _np(token_prob_t)[0, :, 0] - prob_ml = token_prob_ml[0, :, 0] - jd_plots["joint_decision_prob_u0.png"] = _plot_line( - prob_t, - prob_ml, - "JointDecision token_prob (u=0)", - plots_dir / "joint_decision_prob_u0.png", - also_delta=True, - ) - - # Agreement heatmap for token_id - agree = (_np(token_ids_t)[0] == token_id_ml[0]).astype(np.float32) - jd_plots["joint_decision_token_agree.png"] = _plot_image( - agree, - "token_id agreement (torch CPU vs fused)", - plots_dir / "joint_decision_token_agree.png", - vmin=0.0, - vmax=1.0, - ) - - # Latency: fused CoreML vs separate (CoreML joint + CPU PyTorch decision) - def _time_joint_decision_coreml() -> Tuple[float, float]: - for _ in range(max(0, warmup)): - _joint_decision_rollout_coreml(decoder_ref_np) - times = [] - for _ in range(max(1, runs)): - t0 = time.perf_counter() - _joint_decision_rollout_coreml(decoder_ref_np) - t1 = time.perf_counter() - times.append((t1 - t0) * 1000.0) - arr = np.array(times, dtype=np.float64) - return float(arr.mean()), float(arr.std(ddof=1) if arr.size > 1 else 0.0) - - jd_coreml_ms_mean, jd_coreml_ms_std = _time_joint_decision_coreml() - - # Time CPU post-processing only (Torch) on top of CoreML or Torch logits. Use Torch logits. - def _decision_torch_call(): - with torch.inference_mode(): - tl = logits_ref - tl_token = tl[..., :vocab_with_blank] - tl_ids = torch.argmax(tl_token, dim=-1) - tl_probs = torch.softmax(tl_token, dim=-1) - _ = torch.gather(tl_probs, -1, tl_ids.long().unsqueeze(-1)).squeeze(-1) - if num_extra > 0: - _ = torch.argmax(tl[..., -num_extra:], dim=-1) - return None - - jd_decision_torch_ms_mean, jd_decision_torch_ms_std = _time_torch(lambda: _decision_torch_call()) - sep_joint_plus_cpu_ms_mean = float(joint_coreml_ms_mean + jd_decision_torch_ms_mean) - sep_joint_plus_cpu_ms_std = float((joint_coreml_ms_std ** 2 + jd_decision_torch_ms_std ** 2) ** 0.5) - jd_coreml_rtf = float(jd_coreml_ms_mean / (seconds * 1000.0)) if jd_coreml_ms_mean > 0 else None - sep_joint_cpu_rtf = float(sep_joint_plus_cpu_ms_mean / (seconds * 1000.0)) if sep_joint_plus_cpu_ms_mean > 0 else None - - summary["components"]["joint_decision"] = { - "vs_torch_cpu": { - "token_id": {"max_abs": a_tid_t, "max_rel": r_tid_t, "match": bool(ok_tid_t)}, - "token_prob": {"max_abs": a_tprob_t, "max_rel": r_tprob_t, "match": bool(ok_tprob_t)}, - "duration": {"max_abs": a_dur_t, "max_rel": r_dur_t, "match": bool(ok_dur_t)}, - }, - "vs_coreml_joint_cpu": { - "token_id": {"max_abs": a_tid_c, "max_rel": r_tid_c, "match": bool(ok_tid_c)}, - "token_prob": {"max_abs": a_tprob_c, "max_rel": r_tprob_c, "match": bool(ok_tprob_c)}, - "duration": {"max_abs": a_dur_c, "max_rel": r_dur_c, "match": bool(ok_dur_c)}, - }, - "latency": { - "runs": int(runs), - "warmup": int(warmup), - "fused_coreml_first_ms": jd_first_ms, - "fused_coreml_ms": {"mean": jd_coreml_ms_mean, "std": jd_coreml_ms_std}, - "separate_joint_coreml_plus_cpu_ms": {"mean": sep_joint_plus_cpu_ms_mean, "std": sep_joint_plus_cpu_ms_std}, - "rtf": {"fused_coreml": jd_coreml_rtf, "separate_joint_coreml_plus_cpu": sep_joint_cpu_rtf}, - }, - "plots": {k: v for k, v in jd_plots.items() if v}, - } - except Exception as e: - summary["components"]["joint_decision_error"] = str(e) - - # Latency overview plots (saved alongside component plots) - latency_plots = {} - labels = ["preprocessor", "encoder", "decoder", "joint"] - torch_means = [pre_torch_ms_mean, enc_torch_ms_mean, dec_torch_ms_mean, joint_torch_ms_mean] - torch_stds = [pre_torch_ms_std, enc_torch_ms_std, dec_torch_ms_std, joint_torch_ms_std] - coreml_means = [pre_coreml_ms_mean, enc_coreml_ms_mean, dec_coreml_ms_mean, joint_coreml_ms_mean] - coreml_stds = [pre_coreml_ms_std, enc_coreml_ms_std, dec_coreml_ms_std, joint_coreml_ms_std] - lat_path = plots_dir / "latency_summary.png" - spd_path = plots_dir / "latency_speedup.png" - latency_plots["latency_summary.png"] = _plot_latency_bars( - labels, torch_means, torch_stds, coreml_means, coreml_stds, lat_path - ) - latency_plots["latency_speedup.png"] = _plot_speedup_bars( - labels, torch_means, coreml_means, spd_path - ) - - # Fused vs separate latency summary - fused_labels = ["mel+encoder", "joint_decision"] - fused_baseline_means = [ - float(pre_torch_ms_mean + enc_torch_ms_mean), - float(joint_coreml_ms_mean + jd_decision_torch_ms_mean if 'jd_coreml_ms_mean' in locals() else joint_coreml_ms_mean), - ] - fused_coreml_means = [ - float(mel_enc_coreml_ms_mean if 'mel_enc_coreml_ms_mean' in locals() else np.nan), - float(jd_coreml_ms_mean if 'jd_coreml_ms_mean' in locals() else np.nan), - ] - fused_latency_path = plots_dir / "latency_fused_vs_separate.png" - fused_speedup_path = plots_dir / "latency_fused_speedup.png" - latency_plots["latency_fused_vs_separate.png"] = _plot_latency_bars( - fused_labels, fused_baseline_means, [0, 0], fused_coreml_means, [0, 0], fused_latency_path - ) - latency_plots["latency_fused_speedup.png"] = _plot_speedup_bars( - fused_labels, fused_baseline_means, fused_coreml_means, fused_speedup_path - ) - - all_ok = ( - summary["components"]["preprocessor"]["mel"]["match"] - and summary["components"]["preprocessor"]["length_match"] - and summary["components"]["encoder"]["encoder"]["match"] - and summary["components"]["encoder"]["length_match"] - and summary["components"]["decoder"]["decoder"]["match"] - and summary["components"]["decoder"]["h_out"]["match"] - and summary["components"]["decoder"]["c_out"]["match"] - and summary["components"]["joint"]["logits"]["match"] - ) - summary["status"] = "ok" if all_ok else "mismatch" - - # Update metadata.json - meta_path = output_dir / "metadata.json" - try: - meta = json.loads(meta_path.read_text()) - except Exception: - meta = {} - meta["validation"] = summary - meta_path.write_text(json.dumps(meta, indent=2)) - - typer.echo(f"Validation {'passed' if all_ok else 'mismatched'}. Updated {meta_path}") - if HAS_MPL: - typer.echo(f"Saved plots to {plots_dir}") - - -if __name__ == "__main__": - app() diff --git a/convert/parakeet-tdt-v2-0.6b/coreml/compile_modelc.py b/convert/parakeet-tdt-v2-0.6b/coreml/compile_modelc.py deleted file mode 100644 index 9695e03b41fce9280f2ca2ee0e9585c26b2393cc..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-v2-0.6b/coreml/compile_modelc.py +++ /dev/null @@ -1,91 +0,0 @@ -#!/usr/bin/env python3 -"""Compile Core ML packages into ``.mlmodelc`` bundles via ``xcrun``. - -This script walks through the default Parakeet CoreML directories, finds -all ``*.mlpackage`` bundles, and compiles each of them with -``xcrun coremlcompiler`` into ``./compiled`` while preserving the -relative directory structure. -""" -from __future__ import annotations - -import shutil -import subprocess -import sys -from pathlib import Path - -BASE_DIR = Path(__file__).resolve().parent -OUTPUT_ROOT = BASE_DIR / "compiled" -SOURCE_DIRS = [BASE_DIR / "parakeet_coreml", BASE_DIR / "parakeet_coreml_quantized"] - - -def ensure_coremlcompiler() -> None: - """Ensure ``xcrun coremlcompiler`` is available for the active Xcode.""" - xcrun_path = shutil.which("xcrun") - if xcrun_path is None: - print("Error: 'xcrun' not found on PATH. Install Xcode command line tools.", file=sys.stderr) - sys.exit(1) - - try: - subprocess.run([ - xcrun_path, - "--find", - "coremlcompiler", - ], check=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE) - except subprocess.CalledProcessError: - print("Error: 'coremlcompiler' not found via xcrun. Check your Xcode installation.", file=sys.stderr) - sys.exit(1) - - -def gather_packages() -> list[Path]: - """Return a list of all ``*.mlpackage`` bundles under the source dirs.""" - packages: list[Path] = [] - for source in SOURCE_DIRS: - if not source.exists(): - print(f"Warning: {source.relative_to(BASE_DIR)} does not exist; skipping", file=sys.stderr) - continue - packages.extend(source.rglob("*.mlpackage")) - return packages - - -def compile_package(package: Path) -> None: - """Compile a single ``.mlpackage`` bundle using ``xcrun coremlcompiler``.""" - relative_pkg = package.relative_to(BASE_DIR) - output_dir = OUTPUT_ROOT / relative_pkg.parent - output_dir.mkdir(parents=True, exist_ok=True) - output_path = output_dir / f"{package.stem}.mlmodelc" - - if output_path.exists(): - shutil.rmtree(output_path) - - cmd = [ - "xcrun", - "coremlcompiler", - "compile", - str(package), - str(output_dir), - ] - - print(f"Compiling {relative_pkg} -> {output_path.relative_to(BASE_DIR)}") - subprocess.run(cmd, check=True) - - -def main() -> None: - ensure_coremlcompiler() - packages = gather_packages() - - if not packages: - print("No .mlpackage bundles found to compile.") - return - - for package in packages: - try: - compile_package(package) - except subprocess.CalledProcessError as exc: - print(f"Failed to compile {package}: {exc}", file=sys.stderr) - sys.exit(exc.returncode) - - print(f"Finished compiling {len(packages)} package(s) into {OUTPUT_ROOT.relative_to(BASE_DIR)}.") - - -if __name__ == "__main__": - main() diff --git a/convert/parakeet-tdt-v2-0.6b/coreml/context/Efficient Sequence Transduction by Jointly Predicting Tokens and Durations.pdf b/convert/parakeet-tdt-v2-0.6b/coreml/context/Efficient Sequence Transduction by Jointly Predicting Tokens and Durations.pdf deleted file mode 100644 index b3fdd0192810231ef4b0ac61b608a7e06d9d4f6c..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-v2-0.6b/coreml/context/Efficient Sequence Transduction by Jointly Predicting Tokens and Durations.pdf +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:a29e637f2f977a44bece33863aeb01b45b0fc73a9bb723aea2960986e41c5bba -size 3071390 diff --git a/convert/parakeet-tdt-v2-0.6b/coreml/context/FAST CONFORMER WITH LINEARLY SCALABLE ATTENTION.pdf b/convert/parakeet-tdt-v2-0.6b/coreml/context/FAST CONFORMER WITH LINEARLY SCALABLE ATTENTION.pdf deleted file mode 100644 index d0aef3ea5d69a76e18bbd94161a522246ccfd1fc..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-v2-0.6b/coreml/context/FAST CONFORMER WITH LINEARLY SCALABLE ATTENTION.pdf +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:eb8c7706a567ec5795333139f8cd41ff9c127ba99b6dc85d378eb3fa3213bf8d -size 586484 diff --git a/convert/parakeet-tdt-v2-0.6b/coreml/context/coreml_component_io.md b/convert/parakeet-tdt-v2-0.6b/coreml/context/coreml_component_io.md deleted file mode 100644 index 85a34b5927f4e7e98a4c672ebc7cc249992c3100..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-v2-0.6b/coreml/context/coreml_component_io.md +++ /dev/null @@ -1,96 +0,0 @@ -# CoreML Component I/O (Parakeet‑TDT‑0.6B‑v2) - -This file documents the CoreML I/O contracts for each exported sub‑module. All shapes below target the fixed 15‑second window at 16 kHz (240,000 samples). Batch is fixed to 1 for on-device use, and we do not plan to support other window sizes in CoreML (re-export if the requirement changes). - -## Conventions - -- Dtypes: `float32` for real‑valued tensors, `int32` for lengths and token IDs. -- Batch: `B=1` everywhere. -- Vocab: `V=8192` BPE tokens; RNNT adds blank (+1). TDT adds 5 duration logits appended to the class axis. -- Subsampling: overall encoder time reduction ×8. - -## PreprocessorWrapper - -- Inputs - - `audio_signal`: `[1, 240000]` float32, PCM range ~[-1, 1] - - `length`: `[1]` int32, number of valid samples (<= 240000) -- Outputs - - `mel`: `[1, 128, 1501]` float32 (center=True => frames = floor(240000/160)+1 = 1501) - - `mel_length`: `[1]` int32 (=1501 for full 15 s) - -## EncoderWrapper - -- Inputs - - `features`: `[1, 128, 1501]` float32 (mel) - - `length`: `[1]` int32 (=1501) -- Outputs - - `encoded`: `[1, 188, 1024]` float32 (wrapper has time‑major last: it transposes `[B,D,T] -> [B,T,D]`) - - `encoded_length`: `[1]` int32 (`ceil(1501/8)=188`) - -## DecoderWrapper (stateful LSTM, 2 layers) - -- Inputs - - `targets`: `[1, U]` int32 (token IDs in [0, V)) - - `target_lengths`: `[1]` int32 (=U) - - `h_in`: `[2, 1, 640]` float32 (LSTM hidden, L=2 layers) - - `c_in`: `[2, 1, 640]` float32 (LSTM cell, L=2 layers) -- Outputs - - `decoder_output`: `[1, U, 640]` float32 - - `h_out`: `[2, 1, 640]` float32 - - `c_out`: `[2, 1, 640]` float32 - -Notes - -- For step‑wise greedy decoding, set `U=1` and feed back `h_out`, `c_out`. -- For batched token scoring, `U` can be >1 to evaluate multiple symbols per call. - -## JointWrapper - -- Inputs - - `encoder_outputs`: `[1, 188, 1024]` float32 (time‑major) - - `decoder_outputs`: `[1, U, 640]` float32 -- Output - - `logits`: `[1, 188, U, 8192 + 1 + 5]` float32 - -Notes - -- Split the last dimension into `[token_logits: V+1]` and `[duration_logits: 5]` when post‑processing. -- `log_softmax` is disabled for export; apply on CPU if needed. - -## MelEncoderWrapper (fused) - -- Inputs - - `audio_signal`: `[1, 240000]` float32 - - `audio_length`: `[1]` int32 -- Outputs - - `encoder`: `[1, 188, 1024]` float32 - - `encoder_length`: `[1]` int32 - -Notes - -- Fuses preprocessor + encoder to avoid a mel round‑trip. Shapes match chaining `PreprocessorWrapper` then `EncoderWrapper` on the fixed 15 s window. - -## JointDecisionWrapper (fused post‑processing) - -- Inputs - - `encoder`: `[1, 188, 1024]` float32 - - `decoder`: `[1, U, 640]` float32 -- Outputs - - `token_id`: `[1, 188, U]` int32 (argmax over token logits, includes blank) - - `token_prob`: `[1, 188, U]` float32 (softmax probability of chosen token) - - `duration`: `[1, 188, U]` int32 (argmax over 5 duration logits) - -Notes - -- Embeds the common host‑side steps: split joint logits, softmax over token logits, argmax over both heads, and gather chosen token probability. Replaces separate `runJoint` → `splitLogits` → `softmax/argmax` in Swift. - -## Suggested CoreML tensor names - -- Preprocessor: `audio_signal`, `audio_length` → `mel`, `mel_length` -- Encoder: `mel`, `mel_length` → `encoder`, `encoder_length` -- Decoder: `targets`, `target_length`, `h_in`, `c_in` → `decoder`, `h_out`, `c_out` -- Joint: `encoder`, `decoder` → `logits` -- MelEncoder (fused): `audio_signal`, `audio_length` → `encoder`, `encoder_length` -- JointDecision (fused): `encoder`, `decoder` → `token_id`, `token_prob`, `duration` - -These names align with wrappers in `individual_components.py` and simplify downstream wiring. diff --git a/convert/parakeet-tdt-v2-0.6b/coreml/context/coreml_conversion_plan.md b/convert/parakeet-tdt-v2-0.6b/coreml/context/coreml_conversion_plan.md deleted file mode 100644 index cf3cd4c09dbd02b882cca3f37f0802f7bfa569b1..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-v2-0.6b/coreml/context/coreml_conversion_plan.md +++ /dev/null @@ -1,65 +0,0 @@ -# CoreML Conversion Plan (Parakeet‑TDT‑0.6B‑v2) - -This plan describes how we export Parakeet’s sub‑modules to CoreML, validate numerics, and prepare for on‑device decoding. The pipeline keeps a fixed 15‑second audio window. - -## Goals - -- Export preprocessor, encoder, decoder, and joint as separate mlpackages. -- Also export fused variants: `mel_encoder` (preprocessor+encoder) and `joint_decision` (joint + split/softmax/argmax). -- Preserve streaming decoder state I/O for on‑device greedy decoding. -- Validate component outputs against the NeMo reference on a 15‑second clip. - -## Environment - -- Use `uv venv` and run via `uv run` to ensure reproducible resolutions. -- Python 3.10.12, `torch 2.5.0`, `coremltools 8.3.0`, `nemo-toolkit 2.3.1`. - -## Export settings - -- `convert_to = "mlprogram"` -- `minimum_deployment_target` = `ct.target.iOS17` (we only target iOS 17+ for the CoreML deployment) -- `compute_units` = `.CPU_AND_GPU` (or `.ALL` on iOS) -- `compute_precision` = `ct.precision.FLOAT16` or `None` (start with fp32 for validation, then try fp16) -- Fixed window: 15 seconds at 16 kHz → shapes as in `context/coreml_component_io.md`. No variable-length support is required; CoreML graphs can assume the 15 s window. - -## Steps - -- Load the NeMo checkpoint with `ASRModel.from_pretrained("nvidia/parakeet-tdt-0.6b-v2")`. -- Extract modules: preprocessor, encoder, decoder, joint. -- Wrap modules with `PreprocessorWrapper`, `EncoderWrapper`, `DecoderWrapper`, `JointWrapper`; derive `MelEncoderWrapper` and `JointDecisionWrapper` for fused exports. -- Trace each wrapper with static 15‑second inputs (batch=1). Ensure outputs match I/O contracts in `coreml_component_io.md`. -- Define CoreML inputs with explicit names and fixed shapes (15 s audio → mel 1501 → encoder 188). Keeping the window fixed simplifies CoreML deployment; re‑export if window size changes. -- Convert via `ct.convert(...)` with the settings above. -- Save each mlpackage under the chosen `output_dir`, along with a `metadata.json` capturing shapes, sample rate, vocab size, tokenizer path, and export metadata (author, conversion commit, etc.). -- Emit fused graphs (`parakeet_mel_encoder.mlpackage` and `parakeet_joint_decision.mlpackage`) alongside standalone components to reduce host I/O and post‑processing. -- Provide a helper like `compile_modelc.py` to batch `xcrun coremlcompiler` invocations when packaging for release. - -## CLI layout - -- `convert_parakeet_components.py` → conversion-only entry point (no validation), parameterized by compute units, precision, and optional fused exports. Model inspection (encoder strides, etc.) should auto-populate config, inspired by Silero VAD’s converter. -- `compare_parakeet_models.py` → validation/plotting script that loads reference NeMo modules and CoreML packages, runs diffs per component, and reports max/mean errors plus RTF metrics. Use lightweight wrappers similar to Silero’s `PyTorchJITWrapper` / `CoreMLVADWrapper`. -- Future helpers: `compile_modelc.py` (reuse pattern from VAD project) and optional quantization experiments once parity is established. - -## Validation - -- Run inference on a known 16 kHz clip (trim/pad to 15 s): - - Compare preprocessor mel and lengths (max abs/rel diff thresholds: atol=1e-4, rtol=1e-3). - - Compare encoder outputs and lengths. - - Compare decoder outputs for a fixed target sequence and initial zero states. - - Compare joint logits on the same `(T_enc, U)` grid; split `[V+1]` token logits from the last 5 duration logits when computing metrics. -- Record per‑component diffs in `metadata.json` for auditability. - -## Decoding (device) - -- Implement greedy RNNT in app code calling CoreML: - - Either: use modular `joint` and perform split/softmax/argmax on host. - - Or: call fused `joint_decision` to receive `token_id`, `token_prob`, and `duration` directly. - - Preprocess → Encode once per window (or call fused `mel_encoder`). - - Maintain `(h, c)` across symbol steps. - - Blank handling: `token_id == V` indicates blank. - -## Known caveats - -- RNNT joint logits are large: `[188 × U × ~8200]` per window; consider fp16 and/or tiling over `U` to reduce memory. -- Length maths: mel frames use `center=True`, yielding 1501 frames for exactly 240,000 samples; encoder length computed via ceil divide by 8. -- For accurate timestamping, use the original NeMo decoder on server to validate any device‑side greedy implementation. diff --git a/convert/parakeet-tdt-v2-0.6b/coreml/context/hf_model_card_notes.md b/convert/parakeet-tdt-v2-0.6b/coreml/context/hf_model_card_notes.md deleted file mode 100644 index 082919ab08fdfb6067b8537d0af73b3c5001ff5a..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-v2-0.6b/coreml/context/hf_model_card_notes.md +++ /dev/null @@ -1,24 +0,0 @@ -# Hugging Face Model Card Notes: nvidia/parakeet-tdt-0.6b-v2 - -Key details extracted from the public model card for quick reference. - -## Summary -- Task: Multilingual automatic speech recognition (ASR) -- Architecture: FastConformer encoder + RNNT + TDT -- Params: ~0.6B -- License: CC‑BY‑4.0 - -## Languages (25 EU + Russian/Ukrainian) -bg, hr, cs, da, nl, et, fi, fr, de, el, hu, it, lv, lt, mt, pl, pt, ro, sk, sl, sv, es, en, ru, uk - -## Input / Output -- Input: 16 kHz mono audio (`.wav`, `.flac`) -- Output: text with punctuation and capitalization - -## Notes relevant to conversion -- The model card reports accurate word/segment timestamps; Parakeet TDT uses duration buckets which we expose from the joint head (last 5 logits along the class axis). -- Long‑audio inference is reported for GPU; our CoreML export intentionally uses a fixed 15‑second window to fit on‑device constraints. - -References -- HF repo: https://huggingface.co/nvidia/parakeet-tdt-0.6b-v2 -- NeMo FastConformer: https://docs.nvidia.com/deeplearning/nemo/user-guide/docs/en/main/asr/models.html#fast-conformer diff --git a/convert/parakeet-tdt-v2-0.6b/coreml/context/mel_encoder_ane_behavior.md b/convert/parakeet-tdt-v2-0.6b/coreml/context/mel_encoder_ane_behavior.md deleted file mode 100644 index e55cf015d7b17e89d1ce1dedeb54512a056cf5ae..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-v2-0.6b/coreml/context/mel_encoder_ane_behavior.md +++ /dev/null @@ -1,67 +0,0 @@ -# MelEncoder ANE behavior and iPhone 13 issue - -Summary of the failure seen on iPhone 13 (A15 / E5): - -> ANE model load has failed for on-device compiled macho. Must re-compile the E5 bundle. Invalid layer: Tensor dimensions N1D1C1H1W240000 are not within supported range, N[1-65536] D[1-16384] C[1-65536] H[1-16384] W[1-16384]. - -## Root cause - -- The fused MelEncoder model takes raw waveform input shaped `[1, 240000]` for the fixed 15 s window (16 kHz × 15 s). -- In Core ML’s internal layout this maps to `N1 D1 C1 H1 W240000`. -- Apple Neural Engine (ANE) enforces a per-dimension cap of ≀ 16384 for H and W. `W=240000` violates this, so any ANE partition that “sees” the waveform will fail validation on A15 (and other iPhone ANE generations). -- The failure is caused by the preprocessor stage inside the fused MelEncoder (waveform → STFT/mel). The standalone encoder (mel → encoder) uses inputs around `[1, 128, 1501]` and is ANE-safe. - -## Why earlier models “worked” - -- Prior exports either: - - Avoided fusing the waveform preprocessor with the encoder; or - - Exported a “mel→encoder” model only; or - - Relied on Core ML partitioning that kept the preprocessor on CPU/GPU while the encoder ran on ANE. -- Even with `computeUnits = .cpuAndNeuralEngine`, Core ML will fall back to CPU for non-ANE-eligible subgraphs. With split models, the preprocessor silently ran on CPU and the encoder on ANE. -- With the new fused MelEncoder, the compiler tries to start an ANE partition earlier (due to the encoder being ANE-friendly). That exposes the 240000-wide waveform to the ANE plan and triggers the dimension error. - -## Device differences (M‑series Macs vs iPhones) - -- Inputs are identical across devices; what differs is how Core ML partitions the graph. -- On M‑series Macs, the waveform preprocessor commonly stays on CPU/GPU; the ANE (if present) or GPU is used for the encoder. No ANE error is logged. -- On iPhones (A14/A15/A16/A17), the ANE compiler (E-series) may attempt to include earlier elementwise ops, making the NE subgraph input be the waveform itself, which fails validation with `W>16384`. -- Treat this limitation as universal across iPhones with ANE; the precise logging/behavior can vary, but no iPhone ANE can accept a subgraph that sees `W=240000`. - -## Current repo behavior and fix - -- The exporter now sets CPU+GPU compute units for waveform‑in components to avoid ANE attempts while preserving the 15 s window: - - `parakeet_preprocessor.mlpackage` - - `parakeet_mel_encoder.mlpackage` (fused waveform→mel→encoder) -- See the changes in `convert-parakeet.py`: - - `convert-parakeet.py:225` and `convert-parakeet.py:237` — preprocessor exported with `ct.ComputeUnit.CPU_AND_GPU`. - - `convert-parakeet.py:289` and `convert-parakeet.py:301` — fused mel+encoder exported with `ct.ComputeUnit.CPU_AND_GPU`. - -## Recommended runtime settings (iOS) - -- Preprocessor (waveform→mel): set `MLModelConfiguration.computeUnits = .cpuAndGPU` to skip ANE attempts and use CPU/GPU efficiently. -- Encoder / Decoder / Joint: set `computeUnits = .cpuAndNeuralEngine` (or `.all`) to leverage ANE. -- Using `.cpuAndNeuralEngine` for all models also works with the split setup (preprocessor falls back to CPU), but setting `.cpuAndGPU` on the preprocessor avoids ANE compile warnings and extra load-time overhead. - -## Using the split models (recommended) - -- Pipeline: - 1. `parakeet_preprocessor.mlmodelc` (CPU+GPU): input `audio_signal [1, 240000]`, `audio_length [1]` → output `mel [1, 128, 1501]`, `mel_length [1]`. - 2. `parakeet_encoder.mlmodelc` (ANE): input `mel [1, 128, 1501]`, `mel_length [1]` → output `encoder [1, 1024, 188]`, `encoder_length [1]`. - 3. Decoder / Joint / Decision: keep on ANE. -- Shapes are captured in `parakeet_coreml/metadata.json`. - -## If fused‑on‑ANE is required - -- The current fused API (`[1, 240000]` waveform input) cannot compile to ANE on iPhones due to the hard ≀16384 per‑dimension cap. -- Viable approach: introduce a chunked fused variant with input like `[1, 15, 16000]` and implement overlap‑aware preemphasis + STFT + mel inside the model to preserve exact parity with global STFT (center padding, hop=160, win=400). Ensure no op sees a dimension >16384. This changes the model API and requires careful seam handling. -- Alternative: stream ≀1.024 s windows with state and accumulate outputs, but this changes runtime control flow. - -## FAQ - -- Q: Which model causes the iPhone 13 error? - - A: The preprocessor stage inside the fused MelEncoder (not the standalone encoder). -- Q: Is this only iPhone 13? - - A: No. Treat it as affecting all iPhone ANE generations; some may just silently partition or fall back differently. -- Q: Why does M4 work? - - A: Likely because the preprocessor subgraph runs on CPU/GPU; the ANE never “sees” `W=240000` on macOS. - diff --git a/convert/parakeet-tdt-v2-0.6b/coreml/context/nemo-joint-fuse-mode.md b/convert/parakeet-tdt-v2-0.6b/coreml/context/nemo-joint-fuse-mode.md deleted file mode 100644 index 1c5eeae7a0f89608d7a6abddf40a1412c13c642a..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-v2-0.6b/coreml/context/nemo-joint-fuse-mode.md +++ /dev/null @@ -1,87 +0,0 @@ -# NeMo Joint Network Fuse Mode - -## Overview - -The NeMo RNNT Joint network has two operational modes controlled by the `fuse_loss_wer` flag. Understanding this is crucial for proper validation and inference. - -## Fuse Mode Details - -### Normal Mode (`fuse_loss_wer = False`) - -**Purpose**: Standard inference and model export -**Interface**: `joint(encoder_outputs, decoder_outputs)` -**Returns**: Raw logits `[B, T, U, V + 1]` -**Use case**: -- Model inference -- Model export (CoreML, ONNX, etc.) -- Validation and comparison - -```python -# Simple joint computation -logits = model.joint(encoder_outputs=encoder_out, decoder_outputs=decoder_out) -``` - -### Fused Mode (`fuse_loss_wer = True`) - -**Purpose**: Memory-optimized training with integrated loss/WER computation -**Interface**: Requires additional parameters -**Returns**: Loss and WER metrics -**Use case**: Training with memory constraints - -**Required parameters when fused**: -- `encoder_outputs` (mandatory) -- `decoder_outputs` (optional, required for loss) -- `encoder_lengths` (required) -- `transcripts` (optional, required for WER) -- `transcript_lengths` (required) - -**Memory optimization**: Processes batch in sub-batches to reduce memory usage - -```python -# Fused mode - more complex interface -result = model.joint( - encoder_outputs=encoder_out, - decoder_outputs=decoder_out, - encoder_lengths=enc_lengths, - transcripts=transcripts, # Ground truth tokens - transcript_lengths=trans_lengths -) -``` - -## Why We Disable Fused Mode for Validation - -### 1. Interface Simplicity -- Fused mode requires ground truth transcripts that we don't have during validation -- Normal mode only needs encoder/decoder outputs, matching CoreML model interface - -### 2. Export Compatibility -```python -def _prepare_for_export(self, **kwargs): - self._fuse_loss_wer = False # Automatically disabled for export - self.log_softmax = False -``` -NeMo automatically disables fused mode during model export, indicating this is the correct mode for inference. - -### 3. Validation Purpose -- We want raw logits for comparison, not loss/WER calculations -- Fused mode is training-oriented, not inference-oriented - -### 4. CoreML Conversion -CoreML models are converted from the non-fused mode, so validation should use the same mode to ensure fair comparison. - -## Implementation in compare-components.py - -```python -# Disable fused mode for proper inference comparison -asr_model.joint.set_fuse_loss_wer(False) - -# Now we can use the simple interface that matches CoreML -logits_ref = asr_model.joint( - encoder_outputs=encoder_ref, - decoder_outputs=decoder_ref -) -``` - -## Key Takeaway - -For validation and inference, always use `fuse_loss_wer = False` to get the standard joint network behavior that matches exported models and enables fair comparison with CoreML implementations. \ No newline at end of file diff --git a/convert/parakeet-tdt-v2-0.6b/coreml/context/parakeet_tdt_v2_architecture.md b/convert/parakeet-tdt-v2-0.6b/coreml/context/parakeet_tdt_v2_architecture.md deleted file mode 100644 index ca086b16c6a2b05b0bcf49a3d31cd9c98d7844d9..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-v2-0.6b/coreml/context/parakeet_tdt_v2_architecture.md +++ /dev/null @@ -1,61 +0,0 @@ -# Parakeet‑TDT‑0.6B‑v2 Architecture Overview - -## Model topology - -- Streaming‑capable RNN‑Transducer (RNNT) ASR model trained at 16 kHz. -- Time‑Depth Transducer (TDT) extends RNNT to jointly predict token and duration buckets (commonly [0, 1, 2, 3, 4]). -- Four major learnable blocks packaged in the `.nemo` checkpoint: - 1. `AudioToMelSpectrogramPreprocessor` – waveform → 128×T mel‑spectrogram. - 2. `ConformerEncoder` (FastConformer‑style) – 24 conformer blocks, `d_model=1024`, `n_heads=8`, depthwise CNN subsampling with total reduction ×8. - 3. `RNNTDecoder` – 2‑layer LSTM prediction network, hidden size 640, `blank_as_pad=True` for efficient batching and blank handling. - 4. `RNNTJoint` – linear projections from encoder (1024) and decoder (640) into joint space 640 with ReLU, then to class logits. - -## Audio front‑end - -- `sample_rate = 16000`, Hann window, 25 ms window, 10 ms hop (`n_fft=512`, `win_length=400`, `hop_length=160`, `center=True`). -- 128 mel filters (Slaney‑style), log‑magnitude power 2.0 and per‑feature normalization. -- Outputs `(mel, mel_length)` feed the encoder. For CoreML, both tensors must be surfaced to preserve valid‑lengths through subsampling. - -## Encoder (FastConformer) - -- Implemented as `nemo.collections.asr.modules.ConformerEncoder` configured per FastConformer. -- 24 blocks, `d_model=1024`, FFN expansion ×4, 8‑head relative‑position attention, conv kernel 9, batch‑norm. -- Dropouts: 0.1 general, 0.1 pre‑encoder, 0.0 embedding, 0.1 attention; stochastic depth disabled. -- Depthwise CNN subsampling yields overall time reduction ×8. Returns features `[B, D=1024, T_enc]` and lengths. - -## Decoder (`RNNTDecoder`) - -- Stateful LSTM prediction network (2 layers, hidden=640). See `nemo/collections/asr/modules/rnnt.py`. -- `blank_as_pad=True` adds a learnable pad/blank to the embedding; embedding returns zeros for blank, simplifying blank handling. -- Export toggles `_rnnt_export` and uses `.predict()` without prepending SOS so decoder outputs shape `[B, U, 640]`. -- Streaming state is a tuple `(h, c)`, each `[L=2, B, H=640]`. Utilities exist to batch and select states during beam search. - -## Joint network & TDT - -- `nemo.collections.asr.modules.RNNTJoint` with `joint_hidden=640`, activation=ReLU, dropout=0.2. -- Input shapes: encoder `[B, T_enc, 1024]`, decoder `[B, U, 640]` → output logits `[B, T_enc, U, V+1+E]`. -- Vocabulary `V=8192` (SentencePiece BPE). RNNT blank adds `+1`. TDT uses `num_extra_outputs=E=5`, appended in the last dimension of the joint logits. The TDT loss splits the tail 5 entries as duration logits for buckets `[0,1,2,3,4]`. -- Training often enables `fuse_loss_wer=true` with `fused_batch_size=4`; export/inference use just `joint.joint(...)`. - -## Auxiliary heads - -- No separate CTC decoder is attached in the public v2 checkpoint. -- Loss is TDT: `fastemit_lambda=0.0`, `durations=[0,1,2,3,4]`, `sigma=0.02`, `omega=0.1`. - -## Tokenizer - -- SentencePiece BPE (`tokenizer.model`) with 8192 subword units plus control symbols (e.g., punctuation/timestamp markers). RNNT adds blank internally. - -## Notes for CoreML export - -- Exportable subgraphs map 1:1 to NeMo modules: preprocessor, encoder, decoder, joint, and (optionally) a greedy decoder. In addition, we export two fused variants for on‑device efficiency: (1) `mel_encoder` which combines preprocessor+encoder, and (2) `joint_decision` which applies split/softmax/argmax to joint logits to output `token_id`, `token_prob`, and `duration`. -- Decoder state I/O: expose `h` and `c` separately as `[L=2, B, 640]` multi‑arrays to enable on‑device streaming. -- TDT duration logits are appended to the class dimension; downstream code must split `[V+1]` token logits from the last `5` duration logits. - -## Fixed 15‑second window (CoreML constraint) - -- Audio: `B=1`, `T_audio = 15s × 16kHz = 240,000` samples. -- Mel frames (center=True): `T_mel ≈ floor(T_audio / 160) + 1 = 1,500 + 1 = 1,501` → mel `[1, 128, 1501]`. -- Encoder length (×8 subsampling): `T_enc = ceil(1501 / 8) = 188` → encoder `[1, 1024, 188]` (wrapper transposes to `[1, 188, 1024]`). -- Decoder step budget: choose `U_max` to cap per‑window decoding (e.g., 256–384). Joint outputs then shape `[1, T_enc, U_max, 8192+1+5]`. - diff --git a/convert/parakeet-tdt-v2-0.6b/coreml/context/silero_coreml_reference.md b/convert/parakeet-tdt-v2-0.6b/coreml/context/silero_coreml_reference.md deleted file mode 100644 index 2efa8c5dad24884352d705092051bc1fedce9ca7..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-v2-0.6b/coreml/context/silero_coreml_reference.md +++ /dev/null @@ -1,53 +0,0 @@ -# Silero VAD CoreML Reference (Techniques to Reuse) - -This note distills patterns from `/models/vad/silero-vad/coreml` that we can reuse for Parakeet-TDT and highlights CoreML limits to plan around. - -## Tooling Patterns Worth Copying - -- **Dedicated CLI entry points**: `convert-coreml.py` performs all exports, while `compare-models.py` focuses solely on validation/plotting. Keeping convert vs. compare independent shortens iteration when only one side changes. -- **Module wrappers**: Lightweight `nn.Module` wrappers (`STFTModel`, `EncoderModel`, `DecoderModel`) map state-dict weights into export-friendly graphs. They surface state tensors explicitly (hidden/cell) and clamp activations where CoreML would otherwise overflow. -- **Shape contracts via CoreML types**: Inputs/outputs use `ct.TensorType` to document tensor layouts. Silero leans on `ct.RangeDim` for flexible windows; for Parakeet we can simplify by locking to the fixed 15 s window (240 k samples → 1501 mels → 188 encoder frames). -- **Metadata & publish scripts**: Conversion annotates author/version on each mlpackage and reports disk size. `compile_modelc.py` shows how to mass-compile mlpackages to `.mlmodelc` bundles via `xcrun`—handy when preparing release artifacts. -- **Multiple deployment variants**: The project exports both modular (STFT→Encoder→Decoder) and unified pipelines. The 256 ms unified variant demonstrates how to encode fixed iteration counts inside the traced graph (8 chunks + noisy-OR aggregation) for batch workflows. - -## Techniques Directly Applicable to Parakeet-TDT - -- **Separate convert & compare CLIs** - - `convert_parakeet_components.py` should mirror the Silero convert script: accept output directory, compute-unit/precision knobs, optional fused graph exports (e.g., preprocessor+encoder). - - A new `compare_parakeet_models.py` can follow the VAD wrapper pattern: run NeMo vs. CoreML components, collect per-layer diffs, and optionally chart mel/encoder/token probabilities over time. -- **Wrapper modules per component** - - We already have `PreprocessorWrapper`, `EncoderWrapper`, `DecoderWrapper`, `JointWrapper`; extend them with deterministic input padding/casts just like Silero’s wrappers to avoid CoreML runtime surprises. -- **Fixed-window CoreML shapes** - - Keep CoreML inputs fixed to the 15 s window. If we ever change the window length, re-export the models instead of relying on dynamic shape bounds. -- **Explicit state I/O** - - Follow Silero’s practice of exposing LSTM states as inputs/outputs every call. For Parakeet, that means separate `h`/`c` arrays of shape `[num_layers, batch, 640]`, enabling streaming decode loops outside CoreML. -- **Release automation** - - Borrow `compile_modelc.py` to generate `.mlmodelc`; consider a similar helper for quantization experiments once baseline parity is achieved. - -## Parakeet Components to Export - -1. **Audio front-end (Mel)** – same strategy as Silero STFT: wrap the NeMo preprocessor, trace with fixed 15 s audio (240 k samples), and export with constant shapes. -2. **Encoder (FastConformer)** – export with constant shapes for the 188 encoder frames corresponding to the 15 s window (transpose to time-major for CoreML compatibility). -3. **Decoder (RNNT Prediction Network)** – stateful LSTM; replicate Silero’s explicit state tensors but with `[2, 1, 640]` dims and integer targets. Provide zero-state helper for CoreML consumers. -4. **Joint network + TDT head** – trace head-only module; outputs `[B, T_enc, U, V+1+5]`. Document token vs. duration slices (see `tdt_decoding_notes.md`). -5. **Fused graphs** – export both: - - `mel_encoder` (preprocessor+encoder) to reduce I/O. - - `joint_decision` (joint + split/softmax/argmax) to avoid host-side post-processing on logits and return `token_id`, `token_prob`, and `duration` directly. - -## Likely CoreML Pain Points - -- **Joint tensor size**: At 15 s the joint produces ~188×U×(8192+6) floats. Even at `U=256`, that is >400 MB of fp32 activations. Plan for fp16 exports and/or tiled decoding (call joint per-step rather than full grid). -- **Dynamic loops**: RNNT greedy/beam search loops can’t be represented in a single CoreML graph (no `while` support). Expect to orchestrate decoding in Swift/Python by issuing repeated `decoder` and `joint` calls. -- **Large vocabulary softmax**: Full RNNT softmax over 8193 classes is CoreML-friendly but expensive. Avoid embedding beam-search logic in CoreML; keep it host-side. -- **Duration logits**: Splitting the final 5 logits inside CoreML is trivial, but combining them with token control flow should stay in host code (mirrors Silero’s noisy-OR aggregation done in Python before tracing). -- **Tokenizer & text post-processing**: SentencePiece decoding, punctuation fixes, and timestamp formatting remain in Python/Swift; CoreML returns logits only. -- **Feature normalization stats**: Stick to constant statistics baked into the wrapper as Silero does for STFT weights; do not rely on CoreML `LayerNormalization` with dynamic reductions over time. - -## Python / Host Responsibilities - -- Audio chunking, state caching, and greedy/beam decoding loops (leveraging exported decoder + joint). -- Token → text conversion (`sentencepiece`), duration bucket interpretation, punctuation/TDT alignment heuristics. -- Validation tooling (diff calculations, plotting) analogous to Silero’s `compare-models.py`. -- Optional post-processing (e.g., noise suppression, segmentation) that require dynamic control flow. - -Reuse these patterns to keep Parakeet’s conversion pipeline maintainable while respecting CoreML’s graph restrictions. diff --git a/convert/parakeet-tdt-v2-0.6b/coreml/context/tdt_decoding_notes.md b/convert/parakeet-tdt-v2-0.6b/coreml/context/tdt_decoding_notes.md deleted file mode 100644 index a1c5fb6ada701bc349921fbf288504b95fe34840..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-v2-0.6b/coreml/context/tdt_decoding_notes.md +++ /dev/null @@ -1,36 +0,0 @@ -# TDT Decoding Notes (RNNT + Duration Buckets) - -Parakeet‑TDT extends RNNT by appending 5 duration logits to the joint’s class axis. These describe coarse duration buckets associated with an emitted symbol. This note summarizes how to consume them alongside standard RNNT greedy decoding. - -## Joint output - -- Shape: `logits[B, T_enc, U, V+1+5]` where: - - `V=8192` BPE vocab, `+1` RNNT blank, `+5` TDT duration buckets. -- Split last dimension: - - `token_logits = logits[..., :V+1]` - - `duration_logits = logits[..., V+1:] # shape [..., 5]` - -## Greedy decoding sketch - -- Initialize decoder state `(h, c)` and `u=0`; iterate over encoder time `t=0..T_enc-1`. -- At each `(t,u)`: - - Either use raw joint logits: retrieve `token_logits[t, u]` and pick `k = argmax(token_logits)`. - - Or call the fused `JointDecisionWrapper` CoreML model to get `token_id[t,u]`, `token_prob[t,u]`, and `duration[t,u]` directly. - - If `k` is blank: advance `t += 1` (RNNT time advance), keep `u`. - - Else (emit token): append token `k` to hypothesis, update decoder state via `DecoderWrapper` with that token, increment `u += 1`. - - Duration (optional): `d = argmax(duration_logits[t, u_prev])`. Map `d ∈ {0..4}` to bucket durations per training config. Use to assign sub‑frame timestamp estimates. - -## Using duration buckets - -- Bucket IDs 0..4 correspond to the configured `tdt_durations = [0,1,2,3,4]` used during training. -- A simple heuristic for timestamps per emitted token: - - Keep track of current mel/encoder frame index. - - When emitting a token at `(t,u)`, associate bucket `d = argmax(duration_logits[t,u])`. - - Convert bucket to time by multiplying with the encoder frame stride (`8× hop_length / sr = 8×0.01 s = 80 ms` per encoder step) or use mel frame stride if preferred. -- For better quality, prefer server‑side timestamp decoding identical to NeMo’s implementation and use duration buckets as a hint when post‑processing on device. - -## Notes - -- NeMo trains the joint with `num_extra_outputs=5`. These logits are not part of the token softmax and should not affect argmax over `V+1`. -- The model card claims accurate word/segment timestamps; TDT buckets contribute to that, but exact use may vary per downstream aligner. -- If durations are not needed, you can ignore the last 5 logits entirely. diff --git a/convert/parakeet-tdt-v2-0.6b/coreml/convert-parakeet.py b/convert/parakeet-tdt-v2-0.6b/coreml/convert-parakeet.py deleted file mode 100644 index eaaa19ce95913996b3772611443ad6a7c0be7044..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-v2-0.6b/coreml/convert-parakeet.py +++ /dev/null @@ -1,619 +0,0 @@ -#!/usr/bin/env python3 -"""CLI for exporting Parakeet TDT v2 components to CoreML.""" -from __future__ import annotations - -import json -from dataclasses import asdict -from pathlib import Path -from typing import Dict, Optional, Tuple - -import coremltools as ct -import numpy as np -import soundfile as sf -import torch -import typer - -import nemo.collections.asr as nemo_asr - -from individual_components import ( - DecoderWrapper, - EncoderWrapper, - ExportSettings, - JointWrapper, - JointDecisionWrapper, - JointDecisionSingleStep, - PreprocessorWrapper, - MelEncoderWrapper, - _coreml_convert, -) - -DEFAULT_MODEL_ID = "nvidia/parakeet-tdt-0.6b-v2" -AUTHOR = "Fluid Inference" - - -def _compute_length(seconds: float, sample_rate: int) -> int: - return int(round(seconds * sample_rate)) - - -def _prepare_audio( - validation_audio: Optional[Path], - sample_rate: int, - max_samples: int, - seed: Optional[int], -) -> torch.Tensor: - if validation_audio is None: - if seed is not None: - torch.manual_seed(seed) - audio = torch.randn(1, max_samples, dtype=torch.float32) - return audio - - data, sr = sf.read(str(validation_audio), dtype="float32") - if sr != sample_rate: - raise typer.BadParameter( - f"Validation audio sample rate {sr} does not match model rate {sample_rate}" - ) - - if data.ndim > 1: - data = data[:, 0] - - if data.size == 0: - raise typer.BadParameter("Validation audio is empty") - - if data.size < max_samples: - pad_width = max_samples - data.size - data = np.pad(data, (0, pad_width)) - elif data.size > max_samples: - data = data[:max_samples] - - audio = torch.from_numpy(data).unsqueeze(0).to(dtype=torch.float32) - return audio - - -def _save_mlpackage(model: ct.models.MLModel, path: Path, description: str) -> None: - # Ensure iOS 17+ target for MLProgram ops and ANE readiness - try: - model.minimum_deployment_target = ct.target.iOS17 - except Exception: - pass - model.short_description = description - model.author = AUTHOR - path.parent.mkdir(parents=True, exist_ok=True) - model.save(str(path)) - - -def _tensor_shape(tensor: torch.Tensor) -> Tuple[int, ...]: - return tuple(int(dim) for dim in tensor.shape) - - -def _parse_compute_units(name: str) -> ct.ComputeUnit: - """Parse a human-friendly compute units string into ct.ComputeUnit. - - Accepted (case-insensitive): ALL, CPU_ONLY, CPU_AND_GPU, CPU_AND_NE. - """ - normalized = str(name).strip().upper() - mapping = { - "ALL": ct.ComputeUnit.ALL, - "CPU_ONLY": ct.ComputeUnit.CPU_ONLY, - "CPU_AND_GPU": ct.ComputeUnit.CPU_AND_GPU, - "CPU_AND_NE": ct.ComputeUnit.CPU_AND_NE, - "CPU_AND_NEURALENGINE": ct.ComputeUnit.CPU_AND_NE, - } - if normalized not in mapping: - raise typer.BadParameter( - f"Unknown compute units '{name}'. Choose from: " + ", ".join(mapping.keys()) - ) - return mapping[normalized] - - -def _parse_compute_precision(name: Optional[str]) -> Optional[ct.precision]: - """Parse compute precision string into ct.precision or None. - - Accepted (case-insensitive): FLOAT32, FLOAT16. If None/empty, returns None (tool default). - """ - if name is None: - return None - normalized = str(name).strip().upper() - if normalized == "": - return None - mapping = { - "FLOAT32": ct.precision.FLOAT32, - "FLOAT16": ct.precision.FLOAT16, - } - if normalized not in mapping: - raise typer.BadParameter( - f"Unknown compute precision '{name}'. Choose from: " + ", ".join(mapping.keys()) - ) - return mapping[normalized] - - -# Validation logic removed; use compare-compnents.py for comparisons. - - -# Fixed export choices: CPU_ONLY + FP32, min target iOS17 - - -app = typer.Typer(add_completion=False, pretty_exceptions_show_locals=False) - - -@app.command() -def convert( - nemo_path: Optional[Path] = typer.Option( - None, - "--nemo-path", - exists=True, - resolve_path=True, - help="Path to parakeet-tdt-0.6b-v2 .nemo checkpoint (skip to auto-download)", - ), - model_id: str = typer.Option( - DEFAULT_MODEL_ID, - "--model-id", - help="Model identifier to download when --nemo-path is omitted", - ), - output_dir: Path = typer.Option(Path("parakeet_coreml"), help="Directory where mlpackages and metadata will be written"), - preprocessor_cu: str = typer.Option( - "CPU_ONLY", - "--preprocessor-cu", - help="Compute units for preprocessor (default CPU_ONLY)", - ), - mel_encoder_cu: str = typer.Option( - "CPU_ONLY", - "--mel-encoder-cu", - help="Compute units for fused mel+encoder (default CPU_ONLY)", - ), - compute_precision: Optional[str] = typer.Option( - None, - "--compute-precision", - help="Export precision: FLOAT32 (default) or FLOAT16 to shrink non-quantized weights.", - ), -) -> None: - """Export all Parakeet sub-modules to CoreML with a fixed 15-second window.""" - # Runtime CoreML contract keeps U=1 so the prediction net matches the streaming decoder. - export_settings = ExportSettings( - output_dir=output_dir, - compute_units=ct.ComputeUnit.CPU_ONLY, # Default: CPU-only for all components - deployment_target=ct.target.iOS17, # iOS 17+ features and kernels - compute_precision=_parse_compute_precision(compute_precision), - max_audio_seconds=15.0, - max_symbol_steps=1, - ) - - typer.echo("Export configuration:") - typer.echo(asdict(export_settings)) - - output_dir.mkdir(parents=True, exist_ok=True) - pre_cu = _parse_compute_units(preprocessor_cu) - melenc_cu = _parse_compute_units(mel_encoder_cu) - - if nemo_path is not None: - typer.echo(f"Loading NeMo model from {nemo_path}
") - asr_model = nemo_asr.models.EncDecRNNTBPEModel.restore_from( - str(nemo_path), map_location="cpu" - ) - checkpoint_meta = { - "type": "file", - "path": str(nemo_path), - } - else: - typer.echo(f"Downloading NeMo model via {model_id}
") - asr_model = nemo_asr.models.EncDecRNNTBPEModel.from_pretrained( - model_id, map_location="cpu" - ) - checkpoint_meta = { - "type": "pretrained", - "model_id": model_id, - } - asr_model.eval() - - sample_rate = int(asr_model.cfg.preprocessor.sample_rate) - max_samples = _compute_length(export_settings.max_audio_seconds, sample_rate) - # Prefer a bundled 15s 16kHz audio if available - default_audio = (Path(__file__).parent / "audio" / "yc_first_minute_16k_15s.wav").resolve() - if not default_audio.exists(): - raise typer.BadParameter(f"Expected 15s trace audio at {default_audio}; add the file to proceed.") - typer.echo(f"Using trace audio: {default_audio}") - audio_tensor = _prepare_audio(default_audio, sample_rate, max_samples, seed=None) - audio_length = torch.tensor([max_samples], dtype=torch.int32) - - preprocessor = PreprocessorWrapper(asr_model.preprocessor.eval()) - encoder = EncoderWrapper(asr_model.encoder.eval()) - decoder = DecoderWrapper(asr_model.decoder.eval()) - joint = JointWrapper(asr_model.joint.eval()) - - decoder_export_flag = getattr(asr_model.decoder, "_rnnt_export", False) - asr_model.decoder._rnnt_export = True - - try: - with torch.inference_mode(): - mel_ref, mel_length_ref = preprocessor(audio_tensor, audio_length) - mel_length_ref = mel_length_ref.to(dtype=torch.int32) - encoder_ref, encoder_length_ref = encoder(mel_ref, mel_length_ref) - encoder_length_ref = encoder_length_ref.to(dtype=torch.int32) - - # Clone Tensors to drop the inference tensor flag before tracing - mel_ref = mel_ref.clone() - mel_length_ref = mel_length_ref.clone() - encoder_ref = encoder_ref.clone() - encoder_length_ref = encoder_length_ref.clone() - - vocab_size = int(asr_model.tokenizer.vocab_size) - num_extra = int(asr_model.joint.num_extra_outputs) - decoder_hidden = int(asr_model.decoder.pred_hidden) - decoder_layers = int(asr_model.decoder.pred_rnn_layers) - - targets = torch.full( - (1, export_settings.max_symbol_steps), - fill_value=asr_model.decoder.blank_idx, - dtype=torch.int32, - ) - target_lengths = torch.tensor( - [export_settings.max_symbol_steps], dtype=torch.int32 - ) - zero_state = torch.zeros( - decoder_layers, - 1, - decoder_hidden, - dtype=torch.float32, - ) - - with torch.inference_mode(): - decoder_ref, h_ref, c_ref = decoder(targets, target_lengths, zero_state, zero_state) - joint_ref = joint(encoder_ref, decoder_ref) - - decoder_ref = decoder_ref.clone() - h_ref = h_ref.clone() - c_ref = c_ref.clone() - joint_ref = joint_ref.clone() - - typer.echo("Tracing and converting preprocessor
") - # Ensure tracing happens on CPU explicitly - preprocessor = preprocessor.cpu() - audio_tensor = audio_tensor.cpu() - audio_length = audio_length.cpu() - traced_preprocessor = torch.jit.trace( - preprocessor, (audio_tensor, audio_length), strict=False - ) - traced_preprocessor.eval() - preprocessor_inputs = [ - # Allow variable-length audio up to the fixed 15s window using RangeDim - ct.TensorType( - name="audio_signal", - shape=(1, ct.RangeDim(1, max_samples)), - dtype=np.float32, - ), - ct.TensorType(name="audio_length", shape=(1,), dtype=np.int32), - ] - preprocessor_outputs = [ - ct.TensorType(name="mel", dtype=np.float32), - ct.TensorType(name="mel_length", dtype=np.int32), - ] - # Preprocessor compute units (parametrized; default CPU_ONLY) - preprocessor_model = _coreml_convert( - traced_preprocessor, - preprocessor_inputs, - preprocessor_outputs, - export_settings, - compute_units_override=pre_cu, - ) - preprocessor_path = output_dir / "parakeet_preprocessor.mlpackage" - _save_mlpackage( - preprocessor_model, - preprocessor_path, - "Parakeet preprocessor (15 s window)", - ) - - typer.echo("Tracing and converting encoder
") - traced_encoder = torch.jit.trace( - encoder, (mel_ref, mel_length_ref), strict=False - ) - traced_encoder.eval() - encoder_inputs = [ - ct.TensorType(name="mel", shape=_tensor_shape(mel_ref), dtype=np.float32), - ct.TensorType(name="mel_length", shape=(1,), dtype=np.int32), - ] - encoder_outputs = [ - ct.TensorType(name="encoder", dtype=np.float32), - ct.TensorType(name="encoder_length", dtype=np.int32), - ] - # Encoder: CPU only - encoder_model = _coreml_convert( - traced_encoder, - encoder_inputs, - encoder_outputs, - export_settings, - compute_units_override=ct.ComputeUnit.CPU_ONLY, - ) - encoder_path = output_dir / "parakeet_encoder.mlpackage" - _save_mlpackage( - encoder_model, - encoder_path, - "Parakeet encoder (15 s window)", - ) - - # Optional fused export: Preprocessor + Encoder - typer.echo("Tracing and converting fused mel+encoder
") - mel_encoder = MelEncoderWrapper(preprocessor, encoder) - traced_mel_encoder = torch.jit.trace( - mel_encoder, (audio_tensor, audio_length), strict=False - ) - traced_mel_encoder.eval() - mel_encoder_inputs = [ - # Keep fixed 15s window for fused Mel+Encoder - ct.TensorType(name="audio_signal", shape=(1, max_samples), dtype=np.float32), - ct.TensorType(name="audio_length", shape=(1,), dtype=np.int32), - ] - mel_encoder_outputs = [ - ct.TensorType(name="encoder", dtype=np.float32), - ct.TensorType(name="encoder_length", dtype=np.int32), - ] - # Fused mel+encoder compute units (parametrized; default CPU_ONLY) - mel_encoder_model = _coreml_convert( - traced_mel_encoder, - mel_encoder_inputs, - mel_encoder_outputs, - export_settings, - compute_units_override=melenc_cu, - ) - mel_encoder_path = output_dir / "parakeet_mel_encoder.mlpackage" - _save_mlpackage( - mel_encoder_model, - mel_encoder_path, - "Parakeet fused Mel+Encoder (15 s window)", - ) - - typer.echo("Tracing and converting decoder
") - traced_decoder = torch.jit.trace( - decoder, - (targets, target_lengths, zero_state, zero_state), - strict=False, - ) - traced_decoder.eval() - decoder_inputs = [ - ct.TensorType(name="targets", shape=_tensor_shape(targets), dtype=np.int32), - ct.TensorType(name="target_length", shape=(1,), dtype=np.int32), - ct.TensorType(name="h_in", shape=_tensor_shape(zero_state), dtype=np.float32), - ct.TensorType(name="c_in", shape=_tensor_shape(zero_state), dtype=np.float32), - ] - decoder_outputs = [ - ct.TensorType(name="decoder", dtype=np.float32), - ct.TensorType(name="h_out", dtype=np.float32), - ct.TensorType(name="c_out", dtype=np.float32), - ] - # Decoder: CPU only - decoder_model = _coreml_convert( - traced_decoder, - decoder_inputs, - decoder_outputs, - export_settings, - compute_units_override=ct.ComputeUnit.CPU_ONLY, - ) - decoder_path = output_dir / "parakeet_decoder.mlpackage" - _save_mlpackage( - decoder_model, - decoder_path, - "Parakeet decoder (RNNT prediction network)", - ) - - typer.echo("Tracing and converting joint
") - traced_joint = torch.jit.trace( - joint, - (encoder_ref, decoder_ref), - strict=False, - ) - traced_joint.eval() - joint_inputs = [ - ct.TensorType(name="encoder", shape=_tensor_shape(encoder_ref), dtype=np.float32), - ct.TensorType(name="decoder", shape=_tensor_shape(decoder_ref), dtype=np.float32), - ] - joint_outputs = [ - ct.TensorType(name="logits", dtype=np.float32), - ] - # Joint: CPU only - joint_model = _coreml_convert( - traced_joint, - joint_inputs, - joint_outputs, - export_settings, - compute_units_override=ct.ComputeUnit.CPU_ONLY, - ) - joint_path = output_dir / "parakeet_joint.mlpackage" - _save_mlpackage( - joint_model, - joint_path, - "Parakeet joint network (RNNT)", - ) - - # Joint + decision head (split logits, softmax, argmax) - typer.echo("Tracing and converting joint decision head
") - vocab_size = int(asr_model.tokenizer.vocab_size) - num_extra = int(asr_model.joint.num_extra_outputs) - joint_decision = JointDecisionWrapper(joint, vocab_size=vocab_size, num_extra=num_extra) - traced_joint_decision = torch.jit.trace( - joint_decision, - (encoder_ref, decoder_ref), - strict=False, - ) - traced_joint_decision.eval() - joint_decision_inputs = [ - ct.TensorType(name="encoder", shape=_tensor_shape(encoder_ref), dtype=np.float32), - ct.TensorType(name="decoder", shape=_tensor_shape(decoder_ref), dtype=np.float32), - ] - joint_decision_outputs = [ - ct.TensorType(name="token_id", dtype=np.int32), - ct.TensorType(name="token_prob", dtype=np.float32), - ct.TensorType(name="duration", dtype=np.int32), - ] - # JointDecision: CPU only - joint_decision_model = _coreml_convert( - traced_joint_decision, - joint_decision_inputs, - joint_decision_outputs, - export_settings, - compute_units_override=ct.ComputeUnit.CPU_ONLY, - ) - joint_decision_path = output_dir / "parakeet_joint_decision.mlpackage" - _save_mlpackage( - joint_decision_model, - joint_decision_path, - "Parakeet joint + decision head (split, softmax, argmax)", - ) - - # Single-step JointDecision for [1,1024,1] x [1,640,1] -> [1,1,1] - typer.echo("Tracing and converting single-step joint decision
") - jd_single = JointDecisionSingleStep(joint, vocab_size=vocab_size, num_extra=num_extra) - # Create single-step slices from refs - enc_step = encoder_ref[:, :, :1].contiguous() - dec_step = decoder_ref[:, :, :1].contiguous() - traced_jd_single = torch.jit.trace( - jd_single, - (enc_step, dec_step), - strict=False, - ) - traced_jd_single.eval() - jd_single_inputs = [ - ct.TensorType(name="encoder_step", shape=(1, enc_step.shape[1], 1), dtype=np.float32), - ct.TensorType(name="decoder_step", shape=(1, dec_step.shape[1], 1), dtype=np.float32), - ] - jd_single_outputs = [ - ct.TensorType(name="token_id", dtype=np.int32), - ct.TensorType(name="token_prob", dtype=np.float32), - ct.TensorType(name="duration", dtype=np.int32), - ] - # Single-step JointDecision: CPU only - jd_single_model = _coreml_convert( - traced_jd_single, - jd_single_inputs, - jd_single_outputs, - export_settings, - compute_units_override=ct.ComputeUnit.CPU_ONLY, - ) - jd_single_path = output_dir / "parakeet_joint_decision_single_step.mlpackage" - _save_mlpackage( - jd_single_model, - jd_single_path, - "Parakeet single-step joint decision (current frame)", - ) - - metadata: Dict[str, object] = { - "model_id": model_id, - "sample_rate": sample_rate, - "max_audio_seconds": export_settings.max_audio_seconds, - "max_audio_samples": max_samples, - "max_symbol_steps": export_settings.max_symbol_steps, - "vocab_size": vocab_size, - "joint_extra_outputs": num_extra, - "checkpoint": checkpoint_meta, - "coreml": { - "compute_units": export_settings.compute_units.name, - "compute_precision": ( - export_settings.compute_precision.name - if export_settings.compute_precision is not None - else "FLOAT32" - ), - }, - "components": { - "preprocessor": { - "inputs": { - "audio_signal": list(_tensor_shape(audio_tensor)), - "audio_length": [1], - }, - "outputs": { - "mel": list(_tensor_shape(mel_ref)), - "mel_length": [1], - }, - "path": preprocessor_path.name, - }, - "encoder": { - "inputs": { - "mel": list(_tensor_shape(mel_ref)), - "mel_length": [1], - }, - "outputs": { - "encoder": list(_tensor_shape(encoder_ref)), - "encoder_length": [1], - }, - "path": encoder_path.name, - }, - "mel_encoder": { - "inputs": { - "audio_signal": [1, max_samples], - "audio_length": [1], - }, - "outputs": { - "encoder": list(_tensor_shape(encoder_ref)), - "encoder_length": [1], - }, - "path": mel_encoder_path.name, - }, - "decoder": { - "inputs": { - "targets": list(_tensor_shape(targets)), - "target_length": [1], - "h_in": list(_tensor_shape(zero_state)), - "c_in": list(_tensor_shape(zero_state)), - }, - "outputs": { - "decoder": list(_tensor_shape(decoder_ref)), - "h_out": list(_tensor_shape(h_ref)), - "c_out": list(_tensor_shape(c_ref)), - }, - "path": decoder_path.name, - }, - "joint": { - "inputs": { - "encoder": list(_tensor_shape(encoder_ref)), - "decoder": list(_tensor_shape(decoder_ref)), - }, - "outputs": { - "logits": list(_tensor_shape(joint_ref)), - }, - "path": joint_path.name, - }, - "joint_decision": { - "inputs": { - "encoder": list(_tensor_shape(encoder_ref)), - "decoder": list(_tensor_shape(decoder_ref)), - }, - "outputs": { - "token_id": [ - _tensor_shape(encoder_ref)[0], - _tensor_shape(encoder_ref)[1], - _tensor_shape(decoder_ref)[1], - ], - "token_prob": [ - _tensor_shape(encoder_ref)[0], - _tensor_shape(encoder_ref)[1], - _tensor_shape(decoder_ref)[1], - ], - "duration": [ - _tensor_shape(encoder_ref)[0], - _tensor_shape(encoder_ref)[1], - _tensor_shape(decoder_ref)[1], - ], - }, - "path": joint_decision_path.name, - }, - "joint_decision_single_step": { - "inputs": { - "encoder_step": [1, _tensor_shape(encoder_ref)[2-1], 1], - "decoder_step": [1, _tensor_shape(decoder_ref)[2-1], 1], - }, - "outputs": { - "token_id": [1, 1, 1], - "token_prob": [1, 1, 1], - "duration": [1, 1, 1], - }, - "path": jd_single_path.name, - }, - }, - } - - metadata_path = output_dir / "metadata.json" - metadata_path.write_text(json.dumps(metadata, indent=2)) - typer.echo(f"Export complete. Metadata written to {metadata_path}") - - finally: - asr_model.decoder._rnnt_export = decoder_export_flag - - -if __name__ == "__main__": - app() diff --git a/convert/parakeet-tdt-v2-0.6b/coreml/individual_components.py b/convert/parakeet-tdt-v2-0.6b/coreml/individual_components.py deleted file mode 100644 index 5640330bbaa9e73ef18644458bcf023971ca70cf..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-v2-0.6b/coreml/individual_components.py +++ /dev/null @@ -1,229 +0,0 @@ -#!/usr/bin/env python3 -"""Export Parakeet TDT v2 RNNT components into CoreML and validate outputs.""" -from __future__ import annotations - -from dataclasses import dataclass -from pathlib import Path -from typing import Optional, Tuple - -import coremltools as ct -import torch - - -@dataclass -class ExportSettings: - output_dir: Path - compute_units: ct.ComputeUnit - deployment_target: Optional[ct.target.iOS17] - compute_precision: Optional[ct.precision] - max_audio_seconds: float - max_symbol_steps: int - - -@dataclass -class ValidationSettings: - audio_path: Optional[Path] - seconds: float - seed: Optional[int] - rtol: float - atol: float - skip: bool - - -@dataclass -class ValidationDiff: - name: str - max_abs_diff: float - max_rel_diff: float - - -@dataclass -class ValidationResult: - source: str - audio_num_samples: int - audio_seconds: float - token_length: int - atol: float - rtol: float - diffs: Tuple[ValidationDiff, ...] - - -class PreprocessorWrapper(torch.nn.Module): - def __init__(self, module: torch.nn.Module) -> None: - super().__init__() - self.module = module - - def forward(self, audio_signal: torch.Tensor, length: torch.Tensor) -> Tuple[torch.Tensor, torch.Tensor]: - mel, mel_length = self.module(input_signal=audio_signal, length=length.to(dtype=torch.long)) - return mel, mel_length - - -class EncoderWrapper(torch.nn.Module): - def __init__(self, module: torch.nn.Module) -> None: - super().__init__() - self.module = module - - def forward(self, features: torch.Tensor, length: torch.Tensor) -> Tuple[torch.Tensor, torch.Tensor]: - encoded, encoded_lengths = self.module(audio_signal=features, length=length.to(dtype=torch.long)) - return encoded, encoded_lengths - - -class DecoderWrapper(torch.nn.Module): - def __init__(self, module: torch.nn.Module) -> None: - super().__init__() - self.module = module - - def forward( - self, - targets: torch.Tensor, - target_lengths: torch.Tensor, - h_in: torch.Tensor, - c_in: torch.Tensor, - ) -> Tuple[torch.Tensor, torch.Tensor, torch.Tensor]: - state = [h_in, c_in] - decoder_output, _, new_state = self.module( - targets=targets.to(dtype=torch.long), - target_length=target_lengths.to(dtype=torch.long), - states=state, - ) - return decoder_output, new_state[0], new_state[1] - - -class JointWrapper(torch.nn.Module): - def __init__(self, module: torch.nn.Module) -> None: - super().__init__() - self.module = module - - def forward(self, encoder_outputs: torch.Tensor, decoder_outputs: torch.Tensor) -> torch.Tensor: - # Input: encoder_outputs [B, D, T], decoder_outputs [B, D, U] - # Transpose to match what projection layers expect - encoder_outputs = encoder_outputs.transpose(1, 2) # [B, T, D] - decoder_outputs = decoder_outputs.transpose(1, 2) # [B, U, D] - - # Apply projections - enc_proj = self.module.enc(encoder_outputs) # [B, T, 640] - dec_proj = self.module.pred(decoder_outputs) # [B, U, 640] - - # Explicit broadcasting along T and U to avoid converter ambiguity - x = enc_proj.unsqueeze(2) + dec_proj.unsqueeze(1) # [B, T, U, 640] - x = self.module.joint_net[0](x) # ReLU - x = self.module.joint_net[1](x) # Dropout (no-op in eval) - out = self.module.joint_net[2](x) # Linear -> logits [B, T, U, 8198] - return out - - -class MelEncoderWrapper(torch.nn.Module): - """Fused wrapper: waveform -> mel -> encoder. - - Inputs: - - audio_signal: [B, S] - - audio_length: [B] - - Outputs: - - encoder: [B, D, T_enc] - - encoder_length: [B] - """ - def __init__(self, preprocessor: PreprocessorWrapper, encoder: EncoderWrapper) -> None: - super().__init__() - self.preprocessor = preprocessor - self.encoder = encoder - - def forward(self, audio_signal: torch.Tensor, audio_length: torch.Tensor) -> Tuple[torch.Tensor, torch.Tensor]: - mel, mel_length = self.preprocessor(audio_signal, audio_length) - encoded, enc_len = self.encoder(mel, mel_length.to(dtype=torch.int32)) - return encoded, enc_len - - -class JointDecisionWrapper(torch.nn.Module): - """Joint + decision head: outputs label id, label prob, duration frames. - - Splits joint logits into token logits and duration logits, applies softmax - over tokens, argmax for both heads, and gathers probability of the chosen token. - - Inputs: - - encoder_outputs: [B, D, T] - - decoder_outputs: [B, D, U] - - Returns: - - token_id: [B, T, U] int32 - - token_prob: [B, T, U] float32 - - duration: [B, T, U] int32 (frames; duration buckets as defined by the model) - """ - def __init__(self, joint: JointWrapper, vocab_size: int, num_extra: int) -> None: - super().__init__() - self.joint = joint - self.vocab_with_blank = int(vocab_size) + 1 - self.num_extra = int(num_extra) - - def forward(self, encoder_outputs: torch.Tensor, decoder_outputs: torch.Tensor): - logits = self.joint(encoder_outputs, decoder_outputs) - token_logits = logits[..., : self.vocab_with_blank] - duration_logits = logits[..., -self.num_extra :] - - # Token selection - token_ids = torch.argmax(token_logits, dim=-1).to(dtype=torch.int32) - token_probs_all = torch.softmax(token_logits, dim=-1) - # gather expects int64 (long) indices; cast only for gather - token_prob = torch.gather( - token_probs_all, dim=-1, index=token_ids.long().unsqueeze(-1) - ).squeeze(-1) - - # Duration prediction (duration bucket argmax → frame increments) - duration = torch.argmax(duration_logits, dim=-1).to(dtype=torch.int32) - return token_ids, token_prob, duration - - -class JointDecisionSingleStep(torch.nn.Module): - """Single-step variant for streaming: encoder_step [1, 1024, 1] -> [1,1,1]. - - Inputs: - - encoder_step: [B=1, D=1024, T=1] - - decoder_step: [B=1, D=640, U=1] - - Returns: - - token_id: [1, 1, 1] int32 - - token_prob: [1, 1, 1] float32 - - duration: [1, 1, 1] int32 - """ - def __init__(self, joint: JointWrapper, vocab_size: int, num_extra: int) -> None: - super().__init__() - self.joint = joint - self.vocab_with_blank = int(vocab_size) + 1 - self.num_extra = int(num_extra) - - def forward(self, encoder_step: torch.Tensor, decoder_step: torch.Tensor): - # Reuse JointWrapper which expects [B, D, T] and [B, D, U] - logits = self.joint(encoder_step, decoder_step) # [1, 1, 1, V+extra] - token_logits = logits[..., : self.vocab_with_blank] - duration_logits = logits[..., -self.num_extra :] - - token_ids = torch.argmax(token_logits, dim=-1, keepdim=False).to(dtype=torch.int32) - token_probs_all = torch.softmax(token_logits, dim=-1) - token_prob = torch.gather( - token_probs_all, dim=-1, index=token_ids.long().unsqueeze(-1) - ).squeeze(-1) - duration = torch.argmax(duration_logits, dim=-1, keepdim=False).to(dtype=torch.int32) - return token_ids, token_prob, duration - - -def _coreml_convert( - traced: torch.jit.ScriptModule, - inputs, - outputs, - settings: ExportSettings, - compute_units_override: Optional[ct.ComputeUnit] = None, -) -> ct.models.MLModel: - cu = compute_units_override if compute_units_override is not None else settings.compute_units - kwargs = { - "convert_to": "mlprogram", - "inputs": inputs, - "outputs": outputs, - "compute_units": cu, - } - print("Converting:", traced.__class__.__name__) - print("Conversion kwargs:", kwargs) - if settings.deployment_target is not None: - kwargs["minimum_deployment_target"] = settings.deployment_target - if settings.compute_precision is not None: - kwargs["compute_precision"] = settings.compute_precision - return ct.convert(traced, **kwargs) diff --git a/convert/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/decoder_steps_l2.png b/convert/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/decoder_steps_l2.png deleted file mode 100644 index e9cc0ba8359231e77deff906e60dfa254568533b..0000000000000000000000000000000000000000 Binary files a/convert/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/decoder_steps_l2.png and /dev/null differ diff --git a/convert/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/encoder_time_l2.png b/convert/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/encoder_time_l2.png deleted file mode 100644 index fa3053e9ba9fd5e0646cbc19b3dc7c3ca417ce5b..0000000000000000000000000000000000000000 Binary files a/convert/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/encoder_time_l2.png and /dev/null differ diff --git a/convert/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/joint_decision_prob_u0.png b/convert/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/joint_decision_prob_u0.png deleted file mode 100644 index 1afb0785b0064069044b5e06f2270cf5248d112b..0000000000000000000000000000000000000000 Binary files a/convert/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/joint_decision_prob_u0.png and /dev/null differ diff --git a/convert/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/joint_decision_token_agree.png b/convert/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/joint_decision_token_agree.png deleted file mode 100644 index 253ad19382995a4aced910a330b59e902b076bbd..0000000000000000000000000000000000000000 Binary files a/convert/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/joint_decision_token_agree.png and /dev/null differ diff --git a/convert/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/joint_time_l2.png b/convert/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/joint_time_l2.png deleted file mode 100644 index dc8c869bd6444a83f95e68aec1aa51c4c0e04ef7..0000000000000000000000000000000000000000 Binary files a/convert/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/joint_time_l2.png and /dev/null differ diff --git a/convert/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/joint_top50.png b/convert/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/joint_top50.png deleted file mode 100644 index 4292db8718f5779d18186ced8c43af327922563c..0000000000000000000000000000000000000000 Binary files a/convert/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/joint_top50.png and /dev/null differ diff --git a/convert/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/latency_fused_speedup.png b/convert/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/latency_fused_speedup.png deleted file mode 100644 index a965cbcac59d25eba8a7dfe7837297df75b53dc2..0000000000000000000000000000000000000000 Binary files a/convert/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/latency_fused_speedup.png and /dev/null differ diff --git a/convert/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/latency_fused_vs_separate.png b/convert/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/latency_fused_vs_separate.png deleted file mode 100644 index d637cdb676c2a30c57dcfb2fea7dc796fa5e9f0f..0000000000000000000000000000000000000000 Binary files a/convert/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/latency_fused_vs_separate.png and /dev/null differ diff --git a/convert/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/latency_speedup.png b/convert/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/latency_speedup.png deleted file mode 100644 index 2688f5b6f2643641f97177665920216c340efa41..0000000000000000000000000000000000000000 Binary files a/convert/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/latency_speedup.png and /dev/null differ diff --git a/convert/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/latency_summary.png b/convert/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/latency_summary.png deleted file mode 100644 index 3c51c8fca7ee437d5296a6e82c2a8f8523bbb444..0000000000000000000000000000000000000000 Binary files a/convert/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/latency_summary.png and /dev/null differ diff --git a/convert/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/mel_composite.png b/convert/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/mel_composite.png deleted file mode 100644 index 50870080ad1be000c4b6fa4b3b8f333ebfaabf3d..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/mel_composite.png +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:b46c8f85cf9be45d71847d38173d2a2912e285cda2f1d66c21980f769e39b055 -size 318316 diff --git a/convert/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/mel_encoder_time_l2.png b/convert/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/mel_encoder_time_l2.png deleted file mode 100644 index e713f7a17b7658678531e6a7eadafc6a4be26242..0000000000000000000000000000000000000000 Binary files a/convert/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/mel_encoder_time_l2.png and /dev/null differ diff --git a/convert/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/all_components_compile.png b/convert/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/all_components_compile.png deleted file mode 100644 index 11ce0c66d03ed9a5df57c487d06a3d9ebe6062a4..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/all_components_compile.png +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:d3a9beaa0ed9d9893fa5a123240b7acdaca8e15a734deff20d94def1d4b4fd2d -size 103825 diff --git a/convert/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/all_components_compression.png b/convert/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/all_components_compression.png deleted file mode 100644 index d71982a212f6b7ec1998c6b6b874f05c9f8da353..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/all_components_compression.png +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:dd7ad4cdd309d603728dae0726d779e274ad48798c69b3f64c1cddebec5a9561 -size 101215 diff --git a/convert/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/all_components_latency.png b/convert/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/all_components_latency.png deleted file mode 100644 index 51d3633deb37d6a46b3419a6d8f6831972597ee9..0000000000000000000000000000000000000000 Binary files a/convert/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/all_components_latency.png and /dev/null differ diff --git a/convert/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/all_components_quality.png b/convert/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/all_components_quality.png deleted file mode 100644 index b702e1cae49f41220a043cbab815e943f37b8d9e..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/all_components_quality.png +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:a25157bcfeaad91955b959303a8632bc6b3868dc61b5fdf4a1a1ccb91dc35cb1 -size 113723 diff --git a/convert/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/all_components_size.png b/convert/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/all_components_size.png deleted file mode 100644 index 8c0209963b2e28b05d8ea0dc7b2a075ff0e65b7b..0000000000000000000000000000000000000000 Binary files a/convert/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/all_components_size.png and /dev/null differ diff --git a/convert/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/fused_compression.png b/convert/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/fused_compression.png deleted file mode 100644 index 25522432876d20012049ee5cf830ecd0bc8a44be..0000000000000000000000000000000000000000 Binary files a/convert/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/fused_compression.png and /dev/null differ diff --git a/convert/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/fused_latency.png b/convert/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/fused_latency.png deleted file mode 100644 index b92312755973edc5c407f20fa28d42069bb7f57a..0000000000000000000000000000000000000000 Binary files a/convert/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/fused_latency.png and /dev/null differ diff --git a/convert/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/fused_quality.png b/convert/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/fused_quality.png deleted file mode 100644 index 289f61fe3343cd8265674b9ffabd6e21ead217ae..0000000000000000000000000000000000000000 Binary files a/convert/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/fused_quality.png and /dev/null differ diff --git a/convert/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/fused_size.png b/convert/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/fused_size.png deleted file mode 100644 index 8c794846f1c578688ed41de3c3a8df11c4ed161b..0000000000000000000000000000000000000000 Binary files a/convert/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/fused_size.png and /dev/null differ diff --git a/convert/parakeet-tdt-v2-0.6b/coreml/pyproject.toml b/convert/parakeet-tdt-v2-0.6b/coreml/pyproject.toml deleted file mode 100644 index d73aaacc5f26aa71ce35b13f3dd9fc69b9616bc7..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-v2-0.6b/coreml/pyproject.toml +++ /dev/null @@ -1,251 +0,0 @@ -[project] -name = "parakeet-coreml" -version = "0.1.0" -description = "Add your description here" -readme = "README.md" -requires-python = "==3.10.12" -dependencies = [ - "absl-py==2.3.0", - "accelerate==1.8.1", - "aiohappyeyeballs==2.6.1", - "aiohttp==3.12.13", - "aiosignal==1.3.2", - "alembic==1.16.2", - "annotated-types==0.7.0", - "antlr4-python3-runtime==4.9.3", - "anyio==4.9.0", - "appnope==0.1.4", - "argon2-cffi-bindings==21.2.0", - "argon2-cffi==25.1.0", - "arrow==1.3.0", - "asttokens==3.0.0", - "async-lru==2.0.5", - "async-timeout==5.0.1", - "attrs==25.3.0", - "audioread==3.0.1", - "babel==2.17.0", - "backports-datetime-fromisoformat==2.0.3", - "beautifulsoup4==4.13.4", - "bleach==6.2.0", - "braceexpand==0.1.7", - "cattrs==25.1.1", - "certifi==2025.6.15", - "cffi==1.17.1", - "charset-normalizer==3.4.2", - "click==8.2.1", - "cloudpickle==3.1.1", - "colorlog==6.9.0", - "comm==0.2.2", - "contourpy==1.3.2", - "coremltools==9.0b1", - "cycler==0.12.1", - "cytoolz==1.0.1", - "datasets==3.6.0", - "debugpy==1.8.14", - "decorator==5.2.1", - "defusedxml==0.7.1", - "dill==0.3.8", - "distance==0.1.3", - "docopt==0.6.2", - "editdistance==0.8.1", - "einops==0.8.1", - "exceptiongroup==1.3.0", - "executing==2.2.0", - "fastjsonschema==2.21.1", - "fiddle==0.3.0", - "filelock==3.18.0", - "fonttools==4.58.4", - "fqdn==1.5.1", - "frozenlist==1.7.0", - "fsspec==2024.12.0", - "future==1.0.0", - "g2p-en==2.1.0", - "gitdb==4.0.12", - "gitpython==3.1.44", - "graphviz==0.21", - "grpcio==1.73.1", - "h11==0.16.0", - "hf-xet==1.1.5", - "httpcore==1.0.9", - "httpx==0.28.1", - "huggingface-hub==0.33.1", - "hydra-core==1.3.2", - "idna==3.10", - "inflect==7.5.0", - "intervaltree==3.1.0", - "ipykernel==6.29.5", - "ipython==8.37.0", - "ipywidgets==8.1.7", - "isoduration==20.11.0", - "jedi==0.19.2", - "jinja2==3.1.6", - "jiwer==4.0.0", - "joblib==1.5.1", - "json5==0.12.0", - "jsonpointer==3.0.0", - "jsonschema==4.24.0", - "jsonschema-specifications==2025.4.1", - "jupyter==1.1.1", - "jupyter-console==6.6.3", - "jupyter-events==0.12.0", - "jupyter-lsp==2.2.5", - "jupyter-client==8.6.3", - "jupyter-core==5.8.1", - "jupyter-server==2.16.0", - "jupyter-server-terminals==0.5.3", - "jupyterlab==4.4.4", - "jupyterlab-pygments==0.3.0", - "jupyterlab-server==2.27.3", - "jupyterlab-widgets==3.0.15", - "kaldi-python-io==1.2.2", - "kaldiio==2.18.1", - "kiwisolver==1.4.8", - "lazy-loader==0.4", - "levenshtein==0.27.1", - "lhotse==1.30.3", - "libcst==1.8.2", - "librosa==0.11.0", - "lightning==2.4.0", - "lightning-utilities==0.14.3", - "lilcom==1.8.1", - "llvmlite==0.44.0", - "loguru==0.7.3", - "mako==1.3.10", - "markdown==3.8.2", - "markdown-it-py==3.0.0", - "markupsafe==3.0.2", - "marshmallow==4.0.0", - "matplotlib==3.10.3", - "matplotlib-inline==0.1.7", - "mdurl==0.1.2", - "mediapy==1.1.6", - "mistune==3.1.3", - "more-itertools==10.7.0", - "mpmath==1.3.0", - "msgpack==1.1.1", - "multidict==6.6.2", - "multiprocess==0.70.16", - "nbclient==0.10.2", - "nbconvert==7.16.6", - "nbformat==5.10.4", - "nemo-toolkit==2.3.1", - "nest-asyncio==1.6.0", - "networkx==3.4.2", - "nltk==3.9.1", - "notebook==7.4.3", - "notebook-shim==0.2.4", - "num2words==0.5.14", - "numba==0.61.0", - "numpy==1.26.4", - "omegaconf==2.3.0", - "onnx==1.17.0", - "optuna==4.4.0", - "overrides==7.7.0", - "packaging==24.2", - "pandas==2.3.0", - "pandocfilters==1.5.1", - "parso==0.8.4", - "peft==0.15.2", - "pexpect==4.9.0", - "pillow==11.2.1", - "plac==1.4.5", - "platformdirs==4.3.8", - "pooch==1.8.2", - "prometheus-client==0.22.1", - "prompt-toolkit==3.0.51", - "propcache==0.3.2", - "psutil==7.0.0", - "ptyprocess==0.7.0", - "pure-eval==0.2.3", - "pyaml==25.5.0", - "pyannote-core==5.0.0", - "pyannote-database==5.1.3", - "pyannote-metrics==3.2.1", - "pyarrow==20.0.0", - "pybind11==2.13.6", - "pycparser==2.22", - "pydantic==2.11.7", - "pydantic-core==2.33.2", - "pydub==0.25.1", - "pygments==2.19.2", - "pyloudnorm==0.1.1", - "pyparsing==3.2.3", - "python-dateutil==2.9.0.post0", - "python-json-logger==3.3.0", - "pytorch-lightning==2.5.2", - "pytz==2025.2", - "pyyaml==6.0.2", - "pyzmq==27.0.0", - "rapidfuzz==3.13.0", - "referencing==0.36.2", - "regex==2024.11.6", - "requests==2.32.4", - "resampy==0.4.3", - "rfc3339-validator==0.1.4", - "rfc3986-validator==0.1.1", - "rich==14.0.0", - "rpds-py==0.25.1", - "ruamel-yaml==0.18.14", - "ruamel-yaml-clib==0.2.12", - "sacremoses==0.1.1", - "safetensors==0.5.3", - "scikit-learn==1.5.1", - "scipy==1.15.3", - "send2trash==1.8.3", - "sentencepiece==0.2.0", - "sentry-sdk==2.32.0", - "setproctitle==1.3.6", - "shellingham==1.5.4", - "six==1.17.0", - "smmap==5.0.2", - "sniffio==1.3.1", - "sortedcontainers==2.4.0", - "soundfile==0.13.1", - "soupsieve==2.7", - "sox==1.5.0", - "soxr==0.5.0.post1", - "sqlalchemy==2.0.41", - "stack-data==0.6.3", - "tabulate==0.9.0", - "tensorboard==2.19.0", - "tensorboard-data-server==0.7.2", - "termcolor==3.1.0", - "terminado==0.18.1", - "text-unidecode==1.3", - "texterrors==0.5.1", - "threadpoolctl==3.6.0", - "tinycss2==1.4.0", - "tokenizers==0.21.2", - "tomli==2.2.1", - "toolz==1.0.0", - "torch==2.7.0", - "torchmetrics==1.7.3", - "tornado==6.5.1", - "tqdm==4.67.1", - "traitlets==5.14.3", - "transformers==4.51.3", - "typeguard==4.4.4", - "typer==0.16.0", - "types-python-dateutil==2.9.0.20250516", - "typing-inspection==0.4.1", - "typing-extensions==4.14.0", - "tzdata==2025.2", - "uri-template==1.3.0", - "urllib3==2.5.0", - "wandb==0.20.1", - "wcwidth==0.2.13", - "webcolors==24.11.1", - "webdataset==1.0.2", - "webencodings==0.5.1", - "websocket-client==1.8.0", - "werkzeug==3.1.3", - "wget==3.2", - "widgetsnbextension==4.0.14", - "wrapt==1.17.2", - "xxhash==3.5.0", - "yarl==1.20.1", - "pip>=25.1.1", - "seaborn>=0.13.2", - "pyannote-audio>=3.3.2", - "funasr>=1.2.6", -] diff --git a/convert/parakeet-tdt-v2-0.6b/coreml/quantize_coreml.py b/convert/parakeet-tdt-v2-0.6b/coreml/quantize_coreml.py deleted file mode 100644 index 2067bd179da56ba63fe260b8334c4adc56321970..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-v2-0.6b/coreml/quantize_coreml.py +++ /dev/null @@ -1,1207 +0,0 @@ -#!/usr/bin/env python3 -"""Quantize CoreML mlpackages and compare quality, compression, latency, and compile time. - -This script focuses on the fused models: - - parakeet_mel_encoder.mlpackage (waveform -> encoder) - - parakeet_joint_decision.mlpackage (joint + softmax/argmax) - -It also quantizes the rest of the components in the directory for completeness. - -Variants tried by default (examples): - - int8-linear (per-channel) - - palettize 6-bit (Mel-only) - - jd-only variants (int8 and palette) - - prune + int8 - -Outputs: - - A new directory per variant under // with quantized mlpackages - - quantization_summary.json with aggregate metrics (baseline vs each variant) - - Plots saved under /plots/ and mirrored to /plots/quantize// - - fused_quality.png / fused_latency.png / fused_compression.png / fused_size.png (latency chart also shows compile time) - - all_components_quality.png / all_components_latency.png / all_components_compression.png / all_components_size.png (latency chart also shows compile time) - -Notes: - - Uses CoreMLTools optimize.coreml linear/palettize/prune for MLProgram models. - - Keeps the fixed 15-second window shapes as per context/coreml_component_io.md. - - Run via `uv run` to ensure reproducible environment. - - Sets minimum deployment target to iOS 17 for all outputs and loads models with `compute_units=ALL` by default to enable ANE. - - Tracks offline compile time (mlpackage->mlmodelc) and includes that metric in plots. -""" -from __future__ import annotations - -import json -import platform -import shutil -import subprocess -import time -from dataclasses import dataclass -from pathlib import Path -from typing import Dict, List, Optional, Set, Tuple - -import numpy as np -import soundfile as sf -import typer - -import coremltools as ct -from coremltools.optimize.coreml import ( - OptimizationConfig, - OpLinearQuantizerConfig, - OpPalettizerConfig, - OpThresholdPrunerConfig, - linear_quantize_weights, - palettize_weights, - prune_weights, -) - - -# Optional plotting -try: - import matplotlib - matplotlib.use("Agg") - import matplotlib.pyplot as plt - HAS_MPL = True -except Exception: - HAS_MPL = False - - -BASE_DIR = Path(__file__).resolve().parent -BYTES_IN_MB = 1024 * 1024 - - -DISPLAY_LABEL_OVERRIDES = { - "mel6bit-palettize": "mel6bit", - "enc6bit-palettize": "enc6bit", - "int8-linear": "int8 linear", -} - - -# When no explicit components are provided, we derive the set of -# components to quantize from the selected variants' whitelists. -# If any selected variant has no whitelist (i.e., applies globally), -# we quantize all components. -DEFAULT_TARGET_COMPONENTS: Tuple[str, str] = () - - -# Formatting helpers ------------------------------------------------------- - - -def _format_labels(labels: List[str]) -> List[str]: - formatted: List[str] = [] - for label in labels: - pretty = DISPLAY_LABEL_OVERRIDES.get(label, label) - pretty = pretty.replace("_", " ") - pretty = pretty.replace("-", "\n") - formatted.append(pretty) - return formatted - - -def _friendly_component_name(name: str) -> str: - return name.replace("_", " ").title() - - -def _get_metric_series( - metrics: Dict[str, Dict[str, List[float]]], - component: str, - key: str, - length: int, -) -> List[float]: - values = metrics.get(component, {}).get(key) - if values is None: - return [float("nan")] * length - if len(values) >= length: - return list(values[:length]) - padded = list(values) - padded.extend([float("nan")] * (length - len(values))) - return padded - - -def _plot_bar_rows( - out_path: Path, - display_labels: List[str], - rows: List[Dict[str, object]], - title_suffix: str, -) -> None: - if not HAS_MPL: - return - if not rows or not display_labels: - return - - n_labels = len(display_labels) - fig_width = max(8.0, 1.3 * n_labels) - fig_height = max(2.6 * len(rows), 3.0) - - fig, axes = plt.subplots(len(rows), 1, figsize=(fig_width, fig_height), squeeze=False) - axes = axes.flatten() - x = np.arange(n_labels) - - for ax, row in zip(axes, rows): - values = np.asarray(row.get("values", [float("nan")] * n_labels), dtype=np.float64) - color = row.get("color") - bars = ax.bar(x, values, width=0.6, color=color) - ax.set_title(str(row.get("title", ""))) - ax.set_xticks(x) - ax.set_xticklabels(display_labels, rotation=25, ha="right") - ylim = row.get("ylim") - if isinstance(ylim, tuple) and len(ylim) == 2: - ax.set_ylim(ylim) - ax.grid(axis="y", linestyle="--", alpha=0.3) - # Add value labels for bars - for bar in bars: - h = bar.get_height() - if np.isnan(h): - continue - ax.annotate( - f"{h:.2f}", - xy=(bar.get_x() + bar.get_width() / 2, h), - xytext=(0, 3), - textcoords="offset points", - ha="center", - va="bottom", - fontsize=8, - ) - - if title_suffix: - fig.suptitle(title_suffix, fontsize=12) - fig.tight_layout(rect=(0, 0, 1, 0.95)) - else: - fig.tight_layout() - - out_path.parent.mkdir(parents=True, exist_ok=True) - plt.savefig(out_path) - plt.close(fig) - - -def _plot_fused_category_charts( - plot_dir: Path, - labels: List[str], - mel_quality: List[float], - mel_latency_ms: List[float], - mel_compression: List[float], - mel_size_mb: List[float], - jd_acc: List[float], - jd_latency_ms: List[float], - jd_compression: List[float], - jd_size_mb: List[float], - mel_compile_ms: Optional[List[float]] = None, - jd_compile_ms: Optional[List[float]] = None, - title_suffix: str = "", -) -> List[Path]: - if not HAS_MPL: - return [] - outputs: List[Path] = [] - display = _format_labels(labels) - prefix = f"Fused Components — {title_suffix}" if title_suffix else "Fused Components" - - quality_path = plot_dir / "fused_quality.png" - _plot_bar_rows( - quality_path, - display, - [ - { - "title": "MelEncoder quality (1 - norm err)", - "values": mel_quality, - "color": "C0", - "ylim": (0.0, 1.05), - }, - { - "title": "JointDecision token-id match rate", - "values": jd_acc, - "color": "C0", - "ylim": (0.0, 1.05), - }, - ], - f"{prefix} — Quality", - ) - outputs.append(quality_path) - - compression_path = plot_dir / "fused_compression.png" - _plot_bar_rows( - compression_path, - display, - [ - { - "title": "MelEncoder compression ratio", - "values": mel_compression, - "color": "C2", - }, - { - "title": "JointDecision compression ratio", - "values": jd_compression, - "color": "C2", - }, - ], - f"{prefix} — Compression", - ) - outputs.append(compression_path) - - size_path = plot_dir / "fused_size.png" - _plot_bar_rows( - size_path, - display, - [ - { - "title": "MelEncoder size (MB)", - "values": mel_size_mb, - "color": "C4", - }, - { - "title": "JointDecision size (MB)", - "values": jd_size_mb, - "color": "C4", - }, - ], - f"{prefix} — Size", - ) - outputs.append(size_path) - - latency_rows = [ - { - "title": "MelEncoder latency (ms)", - "values": mel_latency_ms, - "color": "C1", - }, - { - "title": "JointDecision latency (ms)", - "values": jd_latency_ms, - "color": "C1", - }, - ] - if mel_compile_ms is not None and jd_compile_ms is not None: - latency_rows.extend( - [ - { - "title": "MelEncoder compile (ms)", - "values": mel_compile_ms, - "color": "C3", - }, - { - "title": "JointDecision compile (ms)", - "values": jd_compile_ms, - "color": "C3", - }, - ] - ) - - latency_path = plot_dir / "fused_latency.png" - _plot_bar_rows( - latency_path, - display, - latency_rows, - f"{prefix} — Latency", - ) - outputs.append(latency_path) - - return outputs - - -def _plot_all_component_category_charts( - plot_dir: Path, - labels: List[str], - metrics: Dict[str, Dict[str, List[float]]], - title_suffix: str = "", -) -> List[Path]: - if not HAS_MPL: - return [] - outputs: List[Path] = [] - display = _format_labels(labels) - prefix = f"Component Breakdown — {title_suffix}" if title_suffix else "Component Breakdown" - comp_order = [ - "preprocessor", - "encoder", - "mel_encoder", - "decoder", - "joint", - "joint_decision", - ] - n = len(labels) - - quality_rows: List[Dict[str, object]] = [] - for comp in comp_order: - friendly = _friendly_component_name(comp) - if comp == "joint_decision": - values = _get_metric_series(metrics, comp, "acc", n) - title = f"{friendly} token-id match rate" - else: - values = _get_metric_series(metrics, comp, "quality", n) - title = f"{friendly} quality (1 - norm err)" - quality_rows.append({"title": title, "values": values, "color": "C0", "ylim": (0.0, 1.05)}) - - compression_rows: List[Dict[str, object]] = [] - for comp in comp_order: - friendly = _friendly_component_name(comp) - compression_rows.append( - { - "title": f"{friendly} compression ratio", - "values": _get_metric_series(metrics, comp, "compression", n), - "color": "C2", - } - ) - - size_rows: List[Dict[str, object]] = [] - for comp in comp_order: - friendly = _friendly_component_name(comp) - size_rows.append( - { - "title": f"{friendly} size (MB)", - "values": _get_metric_series(metrics, comp, "size_mb", n), - "color": "C4", - } - ) - - latency_rows: List[Dict[str, object]] = [] - for comp in comp_order: - friendly = _friendly_component_name(comp) - latency_rows.append( - { - "title": f"{friendly} latency (ms)", - "values": _get_metric_series(metrics, comp, "latency_ms", n), - "color": "C1", - } - ) - - compile_rows: List[Dict[str, object]] = [] - for comp in comp_order: - friendly = _friendly_component_name(comp) - compile_rows.append( - { - "title": f"{friendly} compile (ms)", - "values": _get_metric_series(metrics, comp, "compile_ms", n), - "color": "C3", - } - ) - - quality_path = plot_dir / "all_components_quality.png" - _plot_bar_rows(quality_path, display, quality_rows, f"{prefix} — Quality") - outputs.append(quality_path) - - compression_path = plot_dir / "all_components_compression.png" - _plot_bar_rows(compression_path, display, compression_rows, f"{prefix} — Compression") - outputs.append(compression_path) - - size_path = plot_dir / "all_components_size.png" - _plot_bar_rows(size_path, display, size_rows, f"{prefix} — Size") - outputs.append(size_path) - - latency_path = plot_dir / "all_components_latency.png" - _plot_bar_rows(latency_path, display, latency_rows, f"{prefix} — Latency") - outputs.append(latency_path) - - compile_path = plot_dir / "all_components_compile.png" - _plot_bar_rows(compile_path, display, compile_rows, f"{prefix} — Compile") - outputs.append(compile_path) - - return outputs - - -app = typer.Typer(add_completion=False, pretty_exceptions_show_locals=False) - - -@dataclass -class VariantConfig: - name: str - # List of (op_kind, opt_config) steps applied in order; op_kind in {'linear','palettize','prune'} - steps: List[Tuple[str, OptimizationConfig]] - category: str - whitelist: Optional[List[str]] = None # component names to apply; others copied - - -def _dir_size_bytes(path: Path) -> int: - total = 0 - for p in path.rglob("*"): - if p.is_file(): - try: - total += p.stat().st_size - except OSError: - pass - return total - - -def _load_metadata(input_dir: Path) -> Dict[str, object]: - meta_path = input_dir / "metadata.json" - if not meta_path.exists(): - raise typer.BadParameter(f"Expected metadata.json in {input_dir}") - return json.loads(meta_path.read_text()) - - -def _prepare_audio( - seconds: float, - sample_rate: int, - audio_path: Optional[Path], -) -> Tuple[np.ndarray, np.ndarray]: - max_samples = int(round(seconds * sample_rate)) - if audio_path is None: - # random but deterministic sample - rng = np.random.default_rng(1234) - audio = rng.standard_normal(size=(1, max_samples), dtype=np.float32).astype(np.float32) - else: - data, sr = sf.read(str(audio_path), dtype="float32") - if sr != sample_rate: - raise typer.BadParameter( - f"Validation audio sample rate {sr} != expected {sample_rate}" - ) - if data.ndim > 1: - data = data[:, 0] - if data.size < max_samples: - data = np.pad(data, (0, max_samples - data.size)) - elif data.size > max_samples: - data = data[:max_samples] - audio = data.reshape(1, -1).astype(np.float32, copy=False) - length = np.array([max_samples], dtype=np.int32) - return audio, length - - -def _predict_latency(model: ct.models.MLModel, inputs: Dict[str, np.ndarray], runs: int = 10, warmup: int = 3) -> Tuple[float, float]: - # Warmup - for _ in range(max(0, warmup)): - _ = model.predict(inputs) - # Timed runs - times: List[float] = [] - for _ in range(max(1, runs)): - t0 = time.perf_counter() - _ = model.predict(inputs) - t1 = time.perf_counter() - times.append((t1 - t0) * 1000.0) # ms - arr = np.array(times, dtype=np.float64) - return float(arr.mean()), float(arr.std(ddof=1) if arr.size > 1 else 0.0) - - -def _max_abs_rel(a: np.ndarray, b: np.ndarray) -> Tuple[float, float]: - a = np.asarray(a, dtype=np.float32) - b = np.asarray(b, dtype=np.float32) - if a.size == 0: - return 0.0, 0.0 - diff = np.abs(a - b) - max_abs = float(diff.max()) - denom = np.maximum(np.abs(a), np.abs(b)) - with np.errstate(divide="ignore", invalid="ignore"): - rel = np.where(denom == 0.0, 0.0, diff / denom) - max_rel = float(rel.max()) - return max_abs, max_rel - - -def _save_mlpackage(model: ct.models.MLModel, path: Path, description: str) -> None: - path.parent.mkdir(parents=True, exist_ok=True) - # Ensure iOS 17 target for proper MLProgram ops (e.g., blockwise shift/scale) - try: - model.minimum_deployment_target = ct.target.iOS17 - except Exception: - pass - model.short_description = description - model.author = "Fluid Inference" - model.save(str(path)) - - -def _offline_compile_time_ms(model_path: Path) -> float: - """Compile mlpackage -> mlmodelc and return wall time in ms (host offline compile). - - Returns NaN on failure. - """ - compiled_dir: Optional[Path] = None - expected_dir = model_path.with_suffix(".mlmodelc") - try: - # Delete any prior compile artifact in-place so we can measure a fresh build - if expected_dir.exists(): - shutil.rmtree(expected_dir, ignore_errors=True) - - t0 = time.perf_counter() - compiled_path = ct.utils.compile_model(str(model_path)) - t1 = time.perf_counter() - compiled_dir = Path(compiled_path) - return (t1 - t0) * 1000.0 - except Exception: - return float("nan") - finally: - if ( - compiled_dir is not None - and compiled_dir.exists() - and compiled_dir == expected_dir - ): - shutil.rmtree(compiled_dir, ignore_errors=True) - -def _chip_spec_string(compute_units: str) -> str: - try: - chip = subprocess.check_output(["sysctl", "-n", "machdep.cpu.brand_string"]).decode().strip() - except Exception: - chip = platform.processor() or platform.machine() - mac_ver = platform.mac_ver()[0] or platform.platform() - return f"Host: {chip} ‱ macOS {mac_ver} ‱ CoreMLTools {ct.__version__} ‱ ComputeUnits={compute_units} ‱ Min Target: iOS17" - - -def _quantize_dir( - input_dir: Path, - output_dir: Path, - variant: VariantConfig, - global_whitelist: Optional[Set[str]] = None, -) -> Dict[str, str]: - """Quantize all mlpackages in input_dir into output_dir using given variant config. - - Returns a map of component name -> saved relative path. - """ - meta = _load_metadata(input_dir) - comps = meta.get("components", {}) - saved: Dict[str, str] = {} - for name, cfg in comps.items(): - src_name = cfg.get("path") - if not src_name: - continue - src_path = input_dir / src_name - if not src_path.exists(): - continue - dst_path = output_dir / src_name - # Use CPU+GPU for preprocessor to avoid NE preprocessor input size issues; others use CPU+NE - cu = ct.ComputeUnit.CPU_AND_GPU if name == "preprocessor" else ct.ComputeUnit.CPU_AND_NE - base_model = ct.models.MLModel(str(src_path), compute_units=cu) - # Target iOS17 when running optimizations so the right ops are chosen - try: - base_model.minimum_deployment_target = ct.target.iOS17 - except Exception: - pass - - if name == "decoder": - typer.echo(f"[{variant.name}] Skipping decoder quantization; copying baseline: {src_name}") - _save_mlpackage(base_model, dst_path, "Baseline copy (decoder quantization disabled) - decoder") - saved[name] = dst_path.name - continue - - skip_reasons: List[str] = [] - if variant.whitelist is not None and name not in variant.whitelist: - skip_reasons.append("not targeted by variant") - if global_whitelist is not None and name not in global_whitelist: - skip_reasons.append("not in requested components") - - if skip_reasons: - reason = "; ".join(skip_reasons) - typer.echo(f"[{variant.name}] Skipping {name} ({reason}); copying baseline: {src_name}") - _save_mlpackage(base_model, dst_path, f"Baseline copy ({reason}) - {name}") - saved[name] = dst_path.name - continue - - typer.echo(f"[{variant.name}] Quantizing {name}: {src_name}") - - try: - q_model = base_model - for step_kind, step_cfg in variant.steps: - if step_kind == 'linear': - q_model = linear_quantize_weights(q_model, step_cfg) - elif step_kind == 'palettize': - q_model = palettize_weights(q_model, step_cfg) - elif step_kind == 'prune': - q_model = prune_weights(q_model, step_cfg) - else: - raise ValueError(f"Unknown variant step: {step_kind}") - except Exception as e: - # If quantization fails (e.g., unsupported op), fall back to copying baseline. - typer.echo(f" ! Failed to quantize {name} with {variant.name}: {e}. Copying baseline.") - _save_mlpackage(base_model, dst_path, f"Baseline copy (failed to quantize) - {name}") - else: - _save_mlpackage(q_model, dst_path, f"{variant.name} quantized - {name}") - saved[name] = dst_path.name - # Persist a variant metadata shim - out_meta = { - "variant": variant.name, - "base_dir": str(input_dir.resolve()), - "components": saved, - } - (output_dir / "quantization_metadata.json").write_text(json.dumps(out_meta, indent=2)) - return saved - - -@app.command() -def quantize( - input_dir: Path = typer.Option(Path("parakeet_coreml"), help="Directory containing baseline mlpackages + metadata.json"), - output_root: Path = typer.Option(Path("parakeet_coreml_quantized"), help="Root output dir for quantized variants"), - validation_audio: Optional[Path] = typer.Option(None, exists=True, resolve_path=True, help="Optional 15s, 16kHz wav for evaluation (defaults to bundled audio if present)"), - compute_units: str = typer.Option("CPU_AND_NE", help="Compute units for evaluation of non-preprocessor models. Preprocessor is forced to CPU_AND_GPU."), - runs: int = typer.Option(10, help="Timed runs per model for latency measurement"), - categories: Optional[List[str]] = typer.Option( - None, - "--category", - "-c", - help="Only run quantization variants in these categories (e.g., linear, mel-palettize). Can be repeated.", - ), - components: Optional[List[str]] = typer.Option( - None, - "--component", - "-m", - help="Component names to quantize. Defaults to mel_encoder and joint_decision. Use 'all' to keep every component enabled.", - ), -) -> None: - """Quantize models, then compare quality, compression, latency, and compile time. - - Variants include int8-linear (per-channel/per-tensor/block), palettization (6-bit), - jd-only probes, and prune+int8. Baseline is the pre-converted models. - """ - meta = _load_metadata(input_dir) - sr = int(meta.get("sample_rate", 16000)) - seconds = float(meta.get("max_audio_seconds", 15.0)) - - components_meta = meta.get("components", {}) - component_lookup = {name.lower(): name for name in components_meta.keys()} - # Defer computing component_filter until after variant/category selection so - # that, by default, we quantize exactly the components targeted by the - # selected variants (instead of a hard-coded subset). - component_filter: Optional[Set[str]] = None - - # Default audio if present - default_audio = (BASE_DIR / "audio" / "yc_first_minute_16k_15s.wav").resolve() - audio_path = validation_audio if validation_audio is not None else (default_audio if default_audio.exists() else None) - if audio_path is not None and validation_audio is None: - typer.echo(f"Using default validation audio: {audio_path}") - audio, audio_len = _prepare_audio(seconds, sr, audio_path) - - # Load baseline models and helpers for inputs - # Baseline models for input preparation - # Force preprocessor to CPU+GPU and all other components to CPU+NE for evaluation - pre_base = ct.models.MLModel(str(input_dir / "parakeet_preprocessor.mlpackage"), compute_units=ct.ComputeUnit.CPU_AND_GPU) - enc_base = ct.models.MLModel(str(input_dir / "parakeet_encoder.mlpackage"), compute_units=ct.ComputeUnit.CPU_AND_NE) - mel_encoder_base = ct.models.MLModel(str(input_dir / "parakeet_mel_encoder.mlpackage"), compute_units=ct.ComputeUnit.CPU_AND_NE) - decoder_base = ct.models.MLModel(str(input_dir / "parakeet_decoder.mlpackage"), compute_units=ct.ComputeUnit.CPU_AND_NE) - joint_decision_base = ct.models.MLModel(str(input_dir / "parakeet_joint_decision.mlpackage"), compute_units=ct.ComputeUnit.CPU_AND_NE) - joint_base = ct.models.MLModel(str(input_dir / "parakeet_joint.mlpackage"), compute_units=ct.ComputeUnit.CPU_AND_NE) - - # Prepare typical inputs once using baseline models - pre_out = pre_base.predict({"audio_signal": audio, "audio_length": audio_len}) - mel_ref = np.array(pre_out["mel"], dtype=np.float32, copy=True) - mel_len = np.array(pre_out["mel_length"], dtype=np.int32, copy=True) - - enc_out = enc_base.predict({"mel": mel_ref, "mel_length": mel_len}) - encoder_ref = np.array(enc_out["encoder"], dtype=np.float32, copy=True) - encoder_len = np.array(enc_out["encoder_length"], dtype=np.int32, copy=True) - - # Decoder inputs from metadata - dec_in = meta["components"]["decoder"]["inputs"] - targets_shape = tuple(int(x) for x in dec_in["targets"]) # e.g., (1, 256) - h_shape = tuple(int(x) for x in dec_in["h_in"]) # e.g., (2, 1, 640) - # Use zeros as targets for reproducibility without needing blank idx - targets = np.zeros(targets_shape, dtype=np.int32) - target_length = np.array([targets_shape[1]], dtype=np.int32) - h0 = np.zeros(h_shape, dtype=np.float32) - c0 = np.zeros(h_shape, dtype=np.float32) - dec_out = decoder_base.predict({ - "targets": targets, - "target_length": target_length, - "h_in": h0, - "c_in": c0, - }) - decoder_ref = np.array(dec_out["decoder"], dtype=np.float32, copy=True) - - # Baseline sizes per component - pre_base_size = _dir_size_bytes(input_dir / "parakeet_preprocessor.mlpackage") - enc_base_size = _dir_size_bytes(input_dir / "parakeet_encoder.mlpackage") - mel_base_size = _dir_size_bytes(input_dir / "parakeet_mel_encoder.mlpackage") - dec_base_size = _dir_size_bytes(input_dir / "parakeet_decoder.mlpackage") - joint_base_size = _dir_size_bytes(input_dir / "parakeet_joint.mlpackage") - jd_base_size = _dir_size_bytes(input_dir / "parakeet_joint_decision.mlpackage") - - # Baseline latencies - pre_base_inputs = {"audio_signal": audio, "audio_length": audio_len} - enc_base_inputs = {"mel": mel_ref, "mel_length": mel_len} - mel_base_inputs = {"audio_signal": audio, "audio_length": audio_len} - dec_base_inputs = {"targets": targets, "target_length": target_length, "h_in": h0, "c_in": c0} - joint_base_inputs = {"encoder": encoder_ref, "decoder": decoder_ref} - jd_base_inputs = {"encoder": encoder_ref, "decoder": decoder_ref} - pre_base_ms, _ = _predict_latency(pre_base, pre_base_inputs, runs=runs) - enc_base_ms, _ = _predict_latency(enc_base, enc_base_inputs, runs=runs) - mel_base_ms, _ = _predict_latency(mel_encoder_base, mel_base_inputs, runs=runs) - dec_base_ms, _ = _predict_latency(decoder_base, dec_base_inputs, runs=runs) - joint_base_ms, _ = _predict_latency(joint_base, joint_base_inputs, runs=runs) - jd_base_ms, _ = _predict_latency(joint_decision_base, jd_base_inputs, runs=runs) - # Cache baseline joint logits for comparisons - logits_base = np.array(joint_base.predict(joint_base_inputs)["logits"], dtype=np.float32, copy=True) - - # Variants - variants: List[VariantConfig] = [ - VariantConfig( - name="int8-linear", - steps=[( - "linear", - OptimizationConfig(global_config=OpLinearQuantizerConfig(mode="linear", granularity="per_channel")), - )], - category="linear", - ), - # 6-bit palettization for MelEncoder only - VariantConfig( - name="mel6bit-palettize", - steps=[( - "palettize", - OptimizationConfig( - global_config=OpPalettizerConfig(mode="kmeans", nbits=6) - ), - )], - category="mel-palettize", - whitelist=["mel_encoder"], - ), - # 6-bit palettization for Encoder only - VariantConfig( - name="enc6bit-palettize", - steps=[( - "palettize", - OptimizationConfig( - global_config=OpPalettizerConfig(mode="kmeans", nbits=6) - ), - )], - category="encoder-palettize", - whitelist=["encoder"], - ), - # (removed) Global palettization variants - ] - - available_categories = {variant.category for variant in variants} - category_lookup = {cat.lower(): cat for cat in available_categories} - selected_categories: Set[str] - if categories: - normalized_categories = [cat.strip().lower() for cat in categories if cat.strip()] - if normalized_categories: - invalid_categories = [cat for cat in normalized_categories if cat not in category_lookup] - if invalid_categories: - available_str = ", ".join(sorted(available_categories)) or "none" - bad = ", ".join(sorted(set(invalid_categories))) - raise typer.BadParameter( - f"Unknown category (--category): {bad}. Available categories: {available_str}." - ) - selected_categories = {category_lookup[cat] for cat in normalized_categories} - variants = [variant for variant in variants if variant.category in selected_categories] - else: - selected_categories = available_categories - else: - selected_categories = available_categories - - if not variants: - typer.echo("No quantization variants matched the requested categories; nothing to do.") - raise typer.Exit(code=0) - - typer.echo("Running variant categories: " + ", ".join(sorted(selected_categories))) - - # Resolve the component whitelist now that variants are known. - if components: - normalized_components = [comp.strip().lower() for comp in components if comp.strip()] - if any(comp == "all" for comp in normalized_components): - component_filter = None - else: - resolved: List[str] = [] - invalid: List[str] = [] - for comp in normalized_components: - match = component_lookup.get(comp) - if match is None: - invalid.append(comp) - else: - resolved.append(match) - if invalid: - available = ", ".join(sorted(component_lookup.values())) or "none" - bad = ", ".join(sorted(set(invalid))) - raise typer.BadParameter( - f"Unknown component(s) for --component: {bad}. Available components: {available}." - ) - component_filter = set(resolved) - else: - # Derive from selected variants' whitelists - derived: Set[str] = set() - has_global = False - for v in variants: - if v.whitelist is None: - has_global = True - break - derived.update(v.whitelist) - component_filter = None if has_global or not derived else derived - - if component_filter is None: - typer.echo("Quantizing components: all components (derived)") - else: - typer.echo("Quantizing components: " + ", ".join(sorted(component_filter)) + " (derived)") - - # Aggregate results (baseline + variants) - summary: Dict[str, Dict[str, object]] = {} - variants_names: List[str] = [] - # Build arrays including baseline as first label - fused_labels: List[str] = ["baseline"] - mel_quality_scores: List[float] = [1.0] - mel_latency_ms: List[float] = [mel_base_ms] - mel_compression: List[float] = [1.0] - mel_size_mb: List[float] = [float(mel_base_size) / BYTES_IN_MB] - # Offline compile time (host) for fused models - mel_compile_ms: List[float] = [_offline_compile_time_ms(input_dir / "parakeet_mel_encoder.mlpackage")] - jd_accuracy: List[float] = [1.0] - jd_latency_ms: List[float] = [jd_base_ms] - jd_compression: List[float] = [1.0] - jd_size_mb: List[float] = [float(jd_base_size) / BYTES_IN_MB] - jd_compile_ms: List[float] = [_offline_compile_time_ms(input_dir / "parakeet_joint_decision.mlpackage")] - - # For the all-components chart, collect per-component metrics similarly - all_metrics: Dict[str, Dict[str, List[float]]] = { - "preprocessor": { - "quality": [1.0], - "compression": [1.0], - "latency_ms": [pre_base_ms], - "compile_ms": [_offline_compile_time_ms(input_dir / "parakeet_preprocessor.mlpackage")], - "size_mb": [float(pre_base_size) / BYTES_IN_MB], - }, - "encoder": { - "quality": [1.0], - "compression": [1.0], - "latency_ms": [enc_base_ms], - "compile_ms": [_offline_compile_time_ms(input_dir / "parakeet_encoder.mlpackage")], - "size_mb": [float(enc_base_size) / BYTES_IN_MB], - }, - "mel_encoder": { - "quality": [1.0], - "compression": [1.0], - "latency_ms": [mel_base_ms], - "compile_ms": [mel_compile_ms[0]], - "size_mb": [float(mel_base_size) / BYTES_IN_MB], - }, - "decoder": { - "quality": [1.0], - "compression": [1.0], - "latency_ms": [dec_base_ms], - "compile_ms": [_offline_compile_time_ms(input_dir / "parakeet_decoder.mlpackage")], - "size_mb": [float(dec_base_size) / BYTES_IN_MB], - }, - "joint": { - "quality": [1.0], - "compression": [1.0], - "latency_ms": [joint_base_ms], - "compile_ms": [_offline_compile_time_ms(input_dir / "parakeet_joint.mlpackage")], - "size_mb": [float(joint_base_size) / BYTES_IN_MB], - }, - "joint_decision": { - "acc": [1.0], - "compression": [1.0], - "latency_ms": [jd_base_ms], - "compile_ms": [jd_compile_ms[0]], - "size_mb": [float(jd_base_size) / BYTES_IN_MB], - }, - } - - # Populate baseline entry - summary["baseline"] = { - "components": { - "preprocessor": { - "quality": 1.0, - "latency_ms": pre_base_ms, - "size_bytes": float(pre_base_size), - "size_mb": float(pre_base_size) / BYTES_IN_MB, - "compression_ratio": 1.0, - "compile_ms": all_metrics["preprocessor"]["compile_ms"][0], - }, - "encoder": { - "quality": 1.0, - "latency_ms": enc_base_ms, - "size_bytes": float(enc_base_size), - "size_mb": float(enc_base_size) / BYTES_IN_MB, - "compression_ratio": 1.0, - "compile_ms": all_metrics["encoder"]["compile_ms"][0], - }, - "mel_encoder": { - "quality": 1.0, - "latency_ms": mel_base_ms, - "size_bytes": float(mel_base_size), - "size_mb": float(mel_base_size) / BYTES_IN_MB, - "compression_ratio": 1.0, - "compile_ms": all_metrics["mel_encoder"]["compile_ms"][0], - }, - "decoder": { - "quality": 1.0, - "latency_ms": dec_base_ms, - "size_bytes": float(dec_base_size), - "size_mb": float(dec_base_size) / BYTES_IN_MB, - "compression_ratio": 1.0, - "compile_ms": all_metrics["decoder"]["compile_ms"][0], - }, - "joint": { - "quality": 1.0, - "latency_ms": joint_base_ms, - "size_bytes": float(joint_base_size), - "size_mb": float(joint_base_size) / BYTES_IN_MB, - "compression_ratio": 1.0, - "compile_ms": all_metrics["joint"]["compile_ms"][0], - }, - "joint_decision": { - "acc": 1.0, - "latency_ms": jd_base_ms, - "size_bytes": float(jd_base_size), - "size_mb": float(jd_base_size) / BYTES_IN_MB, - "compression_ratio": 1.0, - "compile_ms": all_metrics["joint_decision"]["compile_ms"][0], - }, - } - } - - for var in variants: - variants_names.append(var.name) - out_dir = output_root / var.name - out_dir_exists = out_dir.exists() - out_dir.mkdir(parents=True, exist_ok=True) - - expected_components = [] - for comp_cfg in meta.get("components", {}).values(): - rel = comp_cfg.get("path") - if rel: - expected_components.append(out_dir / rel) - - missing = [p for p in expected_components if not p.exists()] - if out_dir_exists and not missing: - typer.echo(f"[{var.name}] Output already present at {out_dir}; skipping quantization step.") - else: - if out_dir_exists and missing: - missing_names = ", ".join(sorted(p.name for p in missing)) or "unknown" - typer.echo(f"[{var.name}] Output directory exists but is incomplete (missing: {missing_names}). Re-quantizing.") - shutil.rmtree(out_dir, ignore_errors=True) - out_dir.mkdir(parents=True, exist_ok=True) - _quantize_dir(input_dir, out_dir, var, component_filter) - - # Load quantized models and pre-compute offline compile time - pre_q_path = out_dir / "parakeet_preprocessor.mlpackage" - enc_q_path = out_dir / "parakeet_encoder.mlpackage" - mel_q_path = out_dir / "parakeet_mel_encoder.mlpackage" - dec_q_path = out_dir / "parakeet_decoder.mlpackage" - joint_q_path = out_dir / "parakeet_joint.mlpackage" - jd_q_path = out_dir / "parakeet_joint_decision.mlpackage" - - pre_compile_ms = _offline_compile_time_ms(pre_q_path) - enc_compile_ms = _offline_compile_time_ms(enc_q_path) - mel_compile_q_ms = _offline_compile_time_ms(mel_q_path) - dec_compile_ms = _offline_compile_time_ms(dec_q_path) - joint_compile_ms = _offline_compile_time_ms(joint_q_path) - jd_compile_q_ms = _offline_compile_time_ms(jd_q_path) - - # Match compute units for quantized artifacts: preprocessor on CPU+GPU; others on CPU+NE - pre_q = ct.models.MLModel(str(pre_q_path), compute_units=ct.ComputeUnit.CPU_AND_GPU) - enc_q = ct.models.MLModel(str(enc_q_path), compute_units=ct.ComputeUnit.CPU_AND_NE) - mel_q = ct.models.MLModel(str(mel_q_path), compute_units=ct.ComputeUnit.CPU_AND_NE) - dec_q = ct.models.MLModel(str(dec_q_path), compute_units=ct.ComputeUnit.CPU_AND_NE) - joint_q = ct.models.MLModel(str(joint_q_path), compute_units=ct.ComputeUnit.CPU_AND_NE) - jd_q = ct.models.MLModel(str(jd_q_path), compute_units=ct.ComputeUnit.CPU_AND_NE) - - # Preprocessor quality vs baseline - pre_q_out = pre_q.predict(pre_base_inputs) - mel_q_in = np.array(pre_q_out["mel"], dtype=np.float32, copy=True) - a_pre, r_pre = _max_abs_rel(mel_ref, mel_q_in) - l2_pre_ref = float(np.linalg.norm(mel_ref)) - l2_pre_err = float(np.linalg.norm(mel_ref - mel_q_in)) - pre_norm_err = (l2_pre_err / (l2_pre_ref + 1e-8)) if l2_pre_ref > 0 else 0.0 - pre_quality = float(max(0.0, 1.0 - pre_norm_err)) - pre_q_ms, _ = _predict_latency(pre_q, pre_base_inputs, runs=runs) - pre_q_size = _dir_size_bytes(out_dir / "parakeet_preprocessor.mlpackage") - all_metrics["preprocessor"]["quality"].append(pre_quality) - all_metrics["preprocessor"]["latency_ms"].append(pre_q_ms) - all_metrics["preprocessor"]["compression"].append(float(pre_base_size) / float(pre_q_size if pre_q_size > 0 else 1)) - all_metrics["preprocessor"].setdefault("compile_ms", []).append(pre_compile_ms) - all_metrics["preprocessor"]["size_mb"].append(float(pre_q_size) / BYTES_IN_MB) - - # Encoder quality vs baseline (feed baseline mel to both) - enc_q_out = enc_q.predict({"mel": mel_ref, "mel_length": mel_len}) - encoder_q = np.array(enc_q_out["encoder"], dtype=np.float32, copy=True) - l2_enc_ref = float(np.linalg.norm(encoder_ref)) - l2_enc_err = float(np.linalg.norm(encoder_ref - encoder_q)) - enc_norm_err = (l2_enc_err / (l2_enc_ref + 1e-8)) if l2_enc_ref > 0 else 0.0 - enc_quality = float(max(0.0, 1.0 - enc_norm_err)) - enc_q_ms, _ = _predict_latency(enc_q, enc_base_inputs, runs=runs) - enc_q_size = _dir_size_bytes(out_dir / "parakeet_encoder.mlpackage") - all_metrics["encoder"]["quality"].append(enc_quality) - all_metrics["encoder"]["latency_ms"].append(enc_q_ms) - all_metrics["encoder"]["compression"].append(float(enc_base_size) / float(enc_q_size if enc_q_size > 0 else 1)) - all_metrics["encoder"].setdefault("compile_ms", []).append(enc_compile_ms) - all_metrics["encoder"]["size_mb"].append(float(enc_q_size) / BYTES_IN_MB) - - # MelEncoder quality - mel_q_out = mel_q.predict(mel_base_inputs) - enc_q_fused = np.array(mel_q_out["encoder"], dtype=np.float32, copy=True) - a_mel, r_mel = _max_abs_rel(encoder_ref, enc_q_fused) - # Normalize error into a [0,1] quality score: 1 / (1 + normalized L2) - # Use relative measure derived from L2 norms as more stable than max. - l2_ref = float(np.linalg.norm(encoder_ref)) - l2_err = float(np.linalg.norm(encoder_ref - enc_q_fused)) - norm_err = (l2_err / (l2_ref + 1e-8)) if l2_ref > 0 else 0.0 - mel_quality = float(max(0.0, 1.0 - norm_err)) - mel_q_ms, _ = _predict_latency(mel_q, mel_base_inputs, runs=runs) - mel_q_size = _dir_size_bytes(out_dir / "parakeet_mel_encoder.mlpackage") - mel_ratio = float(mel_base_size) / float(mel_q_size if mel_q_size > 0 else 1) - mel_size_mb.append(float(mel_q_size) / BYTES_IN_MB) - all_metrics["mel_encoder"]["quality"].append(mel_quality) - all_metrics["mel_encoder"]["latency_ms"].append(mel_q_ms) - all_metrics["mel_encoder"]["compression"].append(float(mel_base_size) / float(mel_q_size if mel_q_size > 0 else 1)) - all_metrics["mel_encoder"].setdefault("compile_ms", []).append(mel_compile_q_ms) - all_metrics["mel_encoder"]["size_mb"].append(float(mel_q_size) / BYTES_IN_MB) - - # JointDecision quality: token-id and duration match rates - jd_base_out = joint_decision_base.predict(jd_base_inputs) - token_id_base = np.array(jd_base_out["token_id"], dtype=np.int32, copy=True) - duration_base = np.array(jd_base_out["duration"], dtype=np.int32, copy=True) - token_prob_base = np.array(jd_base_out["token_prob"], dtype=np.float32, copy=True) - - jd_q_out = jd_q.predict(jd_base_inputs) - token_id_q = np.array(jd_q_out["token_id"], dtype=np.int32, copy=True) - duration_q = np.array(jd_q_out["duration"], dtype=np.int32, copy=True) - token_prob_q = np.array(jd_q_out["token_prob"], dtype=np.float32, copy=True) - - # Accuracy metrics - id_match = float((token_id_q == token_id_base).mean()) - dur_match = float((duration_q == duration_base).mean()) - # Aggregate a single "accuracy" number as token-id match rate (primary) - jd_acc = id_match - jd_q_ms, _ = _predict_latency(jd_q, jd_base_inputs, runs=runs) - jd_q_size = _dir_size_bytes(out_dir / "parakeet_joint_decision.mlpackage") - jd_ratio = float(jd_base_size) / float(jd_q_size if jd_q_size > 0 else 1) - jd_size_mb.append(float(jd_q_size) / BYTES_IN_MB) - all_metrics["joint_decision"].setdefault("acc", []).append(jd_acc) - all_metrics["joint_decision"]["latency_ms"].append(jd_q_ms) - all_metrics["joint_decision"]["compression"].append(float(jd_base_size) / float(jd_q_size if jd_q_size > 0 else 1)) - all_metrics["joint_decision"].setdefault("compile_ms", []).append(jd_compile_q_ms) - all_metrics["joint_decision"]["size_mb"].append(float(jd_q_size) / BYTES_IN_MB) - - # Decoder quality vs baseline - dec_q_out = dec_q.predict(dec_base_inputs) - decoder_q = np.array(dec_q_out["decoder"], dtype=np.float32, copy=True) - l2_dec_ref = float(np.linalg.norm(decoder_ref)) - l2_dec_err = float(np.linalg.norm(decoder_ref - decoder_q)) - dec_norm_err = (l2_dec_err / (l2_dec_ref + 1e-8)) if l2_dec_ref > 0 else 0.0 - dec_quality = float(max(0.0, 1.0 - dec_norm_err)) - dec_q_ms, _ = _predict_latency(dec_q, dec_base_inputs, runs=runs) - dec_q_size = _dir_size_bytes(out_dir / "parakeet_decoder.mlpackage") - all_metrics["decoder"]["quality"].append(dec_quality) - all_metrics["decoder"]["latency_ms"].append(dec_q_ms) - all_metrics["decoder"]["compression"].append(float(dec_base_size) / float(dec_q_size if dec_q_size > 0 else 1)) - all_metrics["decoder"].setdefault("compile_ms", []).append(dec_compile_ms) - all_metrics["decoder"]["size_mb"].append(float(dec_q_size) / BYTES_IN_MB) - - # Joint quality vs baseline (compare logits) - joint_q_out = joint_q.predict(joint_base_inputs) - logits_q = np.array(joint_q_out["logits"], dtype=np.float32, copy=True) - l2_joint_ref = float(np.linalg.norm(logits_base)) - l2_joint_err = float(np.linalg.norm(logits_base - logits_q)) - joint_norm_err = (l2_joint_err / (l2_joint_ref + 1e-8)) if l2_joint_ref > 0 else 0.0 - joint_quality = float(max(0.0, 1.0 - joint_norm_err)) - joint_q_ms, _ = _predict_latency(joint_q, joint_base_inputs, runs=runs) - joint_q_size = _dir_size_bytes(out_dir / "parakeet_joint.mlpackage") - all_metrics["joint"]["quality"].append(joint_quality) - all_metrics["joint"]["latency_ms"].append(joint_q_ms) - all_metrics["joint"]["compression"].append(float(joint_base_size) / float(joint_q_size if joint_q_size > 0 else 1)) - all_metrics["joint"].setdefault("compile_ms", []).append(joint_compile_ms) - all_metrics["joint"]["size_mb"].append(float(joint_q_size) / BYTES_IN_MB) - - # Decoder deltas for JSON - a_dec, r_dec = _max_abs_rel(decoder_ref, decoder_q) - # Joint deltas for JSON - a_joint, r_joint = _max_abs_rel(logits_base, logits_q) - - # Store metrics - summary[var.name] = { - "components": { - "preprocessor": { - "quality": pre_quality, - "latency_ms": pre_q_ms, - "size_bytes": float(pre_q_size), - "size_mb": float(pre_q_size) / BYTES_IN_MB, - "compression_ratio": float(pre_base_size) / float(pre_q_size if pre_q_size > 0 else 1), - "max_abs": a_pre, - "max_rel": r_pre, - "compile_ms": pre_compile_ms, - }, - "encoder": { - "quality": enc_quality, - "latency_ms": enc_q_ms, - "size_bytes": float(enc_q_size), - "size_mb": float(enc_q_size) / BYTES_IN_MB, - "compression_ratio": float(enc_base_size) / float(enc_q_size if enc_q_size > 0 else 1), - "compile_ms": enc_compile_ms, - }, - "mel_encoder": { - "quality": mel_quality, - "latency_ms": mel_q_ms, - "size_bytes": float(mel_q_size), - "size_mb": float(mel_q_size) / BYTES_IN_MB, - "compression_ratio": mel_ratio, - "max_abs": a_mel, - "max_rel": r_mel, - "compile_ms": mel_compile_q_ms, - }, - "decoder": { - "quality": dec_quality, - "latency_ms": dec_q_ms, - "size_bytes": float(dec_q_size), - "size_mb": float(dec_q_size) / BYTES_IN_MB, - "compression_ratio": float(dec_base_size) / float(dec_q_size if dec_q_size > 0 else 1), - "max_abs": a_dec, - "max_rel": r_dec, - "compile_ms": dec_compile_ms, - }, - "joint": { - "quality": joint_quality, - "latency_ms": joint_q_ms, - "size_bytes": float(joint_q_size), - "size_mb": float(joint_q_size) / BYTES_IN_MB, - "compression_ratio": float(joint_base_size) / float(joint_q_size if joint_q_size > 0 else 1), - "max_abs": a_joint, - "max_rel": r_joint, - "compile_ms": joint_compile_ms, - }, - "joint_decision": { - "acc": jd_acc, - "duration_match": dur_match, - "prob_mae": float(np.mean(np.abs(token_prob_q - token_prob_base))), - "latency_ms": jd_q_ms, - "size_bytes": float(jd_q_size), - "size_mb": float(jd_q_size) / BYTES_IN_MB, - "compression_ratio": jd_ratio, - "compile_ms": jd_compile_q_ms, - }, - } - } - - fused_labels.append(var.name) - mel_quality_scores.append(mel_quality) - mel_latency_ms.append(mel_q_ms) - mel_compression.append(mel_ratio) - jd_accuracy.append(jd_acc) - jd_latency_ms.append(jd_q_ms) - jd_compression.append(jd_ratio) - mel_compile_ms.append(mel_compile_q_ms) - jd_compile_ms.append(jd_compile_q_ms) - - # Write summary JSON - out_root = output_root - out_root.mkdir(parents=True, exist_ok=True) - (out_root / "quantization_summary.json").write_text(json.dumps(summary, indent=2)) - - # Plot - plot_dir = out_root / "plots" - title_suffix = _chip_spec_string(compute_units) - fused_paths = _plot_fused_category_charts( - plot_dir, - fused_labels, - mel_quality_scores, - mel_latency_ms, - mel_compression, - mel_size_mb, - jd_accuracy, - jd_latency_ms, - jd_compression, - jd_size_mb, - mel_compile_ms, - jd_compile_ms, - title_suffix, - ) - component_paths = _plot_all_component_category_charts( - plot_dir, - fused_labels, - all_metrics, - title_suffix, - ) - - typer.echo(f"Wrote summary JSON: {out_root / 'quantization_summary.json'}") - if HAS_MPL: - all_plot_paths = fused_paths + component_paths - repo_plot_dir = BASE_DIR / "plots" / "quantize" / compute_units.lower() - repo_plot_dir.mkdir(parents=True, exist_ok=True) - for path in all_plot_paths: - typer.echo(f"Wrote plot: {path}") - if path.exists(): - dest = repo_plot_dir / path.name - shutil.copy2(path, dest) - typer.echo(f"Mirrored plot: {dest}") - else: - typer.echo("matplotlib unavailable; skipped plotting.") - - -if __name__ == "__main__": - app() diff --git a/convert/parakeet-tdt-v2-0.6b/coreml/speech_to_text_streaming_infer_rnnt.py b/convert/parakeet-tdt-v2-0.6b/coreml/speech_to_text_streaming_infer_rnnt.py deleted file mode 100644 index 07d8fcc04df44b41c758ad52f518d36a1ad797ae..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-v2-0.6b/coreml/speech_to_text_streaming_infer_rnnt.py +++ /dev/null @@ -1,341 +0,0 @@ -"""Streaming inference helper that stitches Parakeet CoreML components using the RNNT greedy loop from Nemo.""" -from __future__ import annotations - -import argparse -import json -import logging -from pathlib import Path -from typing import Iterable, List, Optional, Sequence - -import coremltools as ct -import librosa -import numpy as np -import torch - -from parakeet_components import CoreMLModelBundle - -LOGGER = logging.getLogger("parakeet_streaming") - - -class BatchedHyps: - """Minimal port of Nemo's batched hypothesis buffer.""" - - def __init__( - self, - batch_size: int, - init_length: int, - device: torch.device, - float_dtype: torch.dtype, - ) -> None: - if init_length <= 0: - raise ValueError("init_length must be > 0") - self._max_length = init_length - self.current_lengths = torch.zeros(batch_size, device=device, dtype=torch.long) - self.transcript = torch.zeros((batch_size, self._max_length), device=device, dtype=torch.long) - self.timestamps = torch.zeros((batch_size, self._max_length), device=device, dtype=torch.long) - self.scores = torch.zeros(batch_size, device=device, dtype=float_dtype) - self.last_timestamp = torch.full((batch_size,), -1, device=device, dtype=torch.long) - self.last_timestamp_lasts = torch.zeros(batch_size, device=device, dtype=torch.long) - self._batch_indices = torch.arange(batch_size, device=device) - self._ones = torch.ones_like(self._batch_indices) - - def add_results( - self, - active_mask: torch.Tensor, - labels: torch.Tensor, - time_indices: torch.Tensor, - scores: torch.Tensor, - ) -> None: - self.scores = torch.where(active_mask, self.scores + scores, self.scores) - self.transcript[self._batch_indices, self.current_lengths] = labels - self.timestamps[self._batch_indices, self.current_lengths] = time_indices - torch.where( - torch.logical_and(active_mask, self.last_timestamp == time_indices), - self.last_timestamp_lasts + 1, - self.last_timestamp_lasts, - out=self.last_timestamp_lasts, - ) - torch.where( - torch.logical_and(active_mask, self.last_timestamp != time_indices), - self._ones, - self.last_timestamp_lasts, - out=self.last_timestamp_lasts, - ) - torch.where(active_mask, time_indices, self.last_timestamp, out=self.last_timestamp) - self.current_lengths += active_mask - - -class CoreMLStreamingDecoder: - """Use exported decoder and joint CoreML models with Nemo's greedy RNNT loop.""" - - def __init__( - self, - decoder_model: ct.models.MLModel, - joint_model: ct.models.MLModel, - *, - vocab_size: int, - blank_id: int, - num_layers: int, - hidden_size: int, - durations: Sequence[int] = (0, 1, 2, 3, 4), - max_symbols: int = 10, - device: torch.device = torch.device("cpu"), - ) -> None: - self.decoder_model = decoder_model - self.joint_model = joint_model - self.vocab_size = vocab_size - self.blank_id = blank_id - self.num_layers = num_layers - self.hidden_size = hidden_size - self.durations = torch.tensor(durations, dtype=torch.long, device=device) - self.max_symbols = max_symbols - self.device = device - - def _predict_decoder(self, labels: torch.Tensor, h_in: np.ndarray, c_in: np.ndarray) -> tuple[torch.Tensor, np.ndarray, np.ndarray]: - outputs = self.decoder_model.predict( - { - "targets": np.array([labels.cpu().numpy()], dtype=np.int32), - "target_lengths": np.array([labels.numel()], dtype=np.int32), - "h_in": h_in, - "c_in": c_in, - } - ) - decoder_output = torch.from_numpy(outputs["decoder_output"]).to(self.device) - return decoder_output, outputs["h_out"], outputs["c_out"] - - def _predict_joint(self, encoder_frame: torch.Tensor, decoder_output: torch.Tensor) -> torch.Tensor: - outputs = self.joint_model.predict( - { - "encoder_outputs": encoder_frame.unsqueeze(1).cpu().numpy().astype(np.float32), - "decoder_outputs": decoder_output.cpu().numpy().astype(np.float32), - } - ) - logits = torch.from_numpy(outputs["logits"]).to(self.device) - return logits.squeeze(1).squeeze(1) - - def decode(self, encoder_output: torch.Tensor, encoder_lengths: torch.Tensor) -> List[List[int]]: - batch_size, max_time, _ = encoder_output.shape - encoder_output = encoder_output.to(self.device) - encoder_lengths = encoder_lengths.to(self.device) - - float_dtype = encoder_output.dtype - batch_indices = torch.arange(batch_size, device=self.device) - labels = torch.full((batch_size,), fill_value=self.blank_id, device=self.device, dtype=torch.long) - time_indices = torch.zeros_like(labels) - safe_time_indices = torch.zeros_like(labels) - time_indices_current = torch.zeros_like(labels) - last_timesteps = encoder_lengths - 1 - active_mask = encoder_lengths > 0 - advance_mask = torch.empty_like(active_mask) - active_mask_prev = torch.empty_like(active_mask) - became_inactive = torch.empty_like(active_mask) - - hyps = BatchedHyps( - batch_size=batch_size, - init_length=max_time * self.max_symbols if self.max_symbols else max_time, - device=self.device, - float_dtype=float_dtype, - ) - - h_in = np.zeros((self.num_layers, batch_size, self.hidden_size), dtype=np.float32) - c_in = np.zeros((self.num_layers, batch_size, self.hidden_size), dtype=np.float32) - - while active_mask.any(): - active_mask_prev.copy_(active_mask) - decoder_output, h_in, c_in = self._predict_decoder(labels, h_in, c_in) - logits = self._predict_joint(encoder_output[batch_indices, safe_time_indices], decoder_output) - - scores, labels = logits[:, : self.vocab_size].max(dim=-1) - duration_indices = logits[:, self.vocab_size : self.vocab_size + len(self.durations)].argmax(dim=-1) - durations = self.durations[duration_indices] - - blank_mask = labels == self.blank_id - durations.masked_fill_(torch.logical_and(durations == 0, blank_mask), 1) - time_indices_current.copy_(time_indices) - time_indices += durations - torch.minimum(time_indices, last_timesteps, out=safe_time_indices) - torch.less(time_indices, encoder_lengths, out=active_mask) - torch.logical_and(active_mask, blank_mask, out=advance_mask) - - while advance_mask.any(): - torch.where(advance_mask, time_indices, time_indices_current, out=time_indices_current) - logits = self._predict_joint(encoder_output[batch_indices, safe_time_indices], decoder_output) - more_scores, more_labels = logits[:, : self.vocab_size].max(dim=-1) - labels = torch.where(advance_mask, more_labels, labels) - scores = torch.where(advance_mask, more_scores, scores) - duration_indices = logits[:, self.vocab_size : self.vocab_size + len(self.durations)].argmax(dim=-1) - durations = self.durations[duration_indices] - blank_mask = labels == self.blank_id - durations.masked_fill_(torch.logical_and(durations == 0, blank_mask), 1) - torch.where(advance_mask, time_indices + durations, time_indices, out=time_indices) - torch.minimum(time_indices, last_timesteps, out=safe_time_indices) - torch.less(time_indices, encoder_lengths, out=active_mask) - torch.logical_and(active_mask, blank_mask, out=advance_mask) - - torch.ne(active_mask, active_mask_prev, out=became_inactive) - hyps.add_results(active_mask, labels, time_indices_current, scores) - - if self.max_symbols is not None: - force_blank = torch.logical_and( - active_mask, - torch.logical_and( - torch.logical_and(labels != self.blank_id, hyps.last_timestamp_lasts >= self.max_symbols), - hyps.last_timestamp == time_indices, - ), - ) - time_indices += force_blank - torch.minimum(time_indices, last_timesteps, out=safe_time_indices) - torch.less(time_indices, encoder_lengths, out=active_mask) - - results: List[List[int]] = [] - for hyp in hyps.transcript: - tokens = [int(token) for token in hyp.tolist() if 0 < token < self.vocab_size] - results.append(tokens) - return results - - -class StreamingTranscriber: - def __init__( - self, - bundle: CoreMLModelBundle, - *, - blank_id: Optional[int] = None, - num_layers: int = 2, - hidden_size: int = 640, - durations: Sequence[int] = (0, 1, 2, 3, 4), - ) -> None: - self.preprocessor = ct.models.MLModel(str(bundle.preprocessor), compute_units=ct.ComputeUnit.CPU_ONLY) - self.encoder = ct.models.MLModel(str(bundle.encoder), compute_units=ct.ComputeUnit.CPU_ONLY) - self.decoder = ct.models.MLModel(str(bundle.decoder), compute_units=ct.ComputeUnit.CPU_ONLY) - self.joint = ct.models.MLModel(str(bundle.joint), compute_units=ct.ComputeUnit.CPU_ONLY) - self.tokenizer = self._load_tokenizer(bundle.tokenizer) - - vocab_size = max(self.tokenizer.keys()) + 1 - if blank_id is None: - blank_id = vocab_size - 1 - self.decoder_helper = CoreMLStreamingDecoder( - self.decoder, - self.joint, - vocab_size=vocab_size, - blank_id=blank_id, - num_layers=num_layers, - hidden_size=hidden_size, - durations=durations, - ) - self.blank_id = blank_id - - @staticmethod - def _load_tokenizer(tokenizer_path: Optional[Path]) -> dict[int, str]: - if tokenizer_path is None: - raise ValueError("Tokenizer JSON is required") - with Path(tokenizer_path).open() as f: - data = json.load(f) - return {int(k): v for k, v in data.items()} - - def _tokens_to_text(self, tokens: Iterable[int]) -> str: - pieces: List[str] = [] - for token in tokens: - piece = self.tokenizer.get(token) - if piece is None: - continue - if piece.startswith("▁"): - if pieces: - pieces.append(" ") - pieces.append(piece[1:]) - else: - pieces.append(piece) - return "".join(pieces).strip() - - def _preprocess(self, audio: np.ndarray) -> tuple[np.ndarray, int]: - audio_2d = audio.reshape(1, -1).astype(np.float32) - length = np.array([audio_2d.shape[-1]], dtype=np.int32) - outputs = self.preprocessor.predict({ - "audio_signal": audio_2d, - "audio_length": length, - }) - return outputs["melspectrogram"], int(outputs["melspectrogram_length"][0]) - - def _encode(self, mel: np.ndarray, mel_length: int) -> tuple[torch.Tensor, torch.Tensor]: - outputs = self.encoder.predict({ - "melspectrogram": mel.astype(np.float32), - "melspectrogram_length": np.array([mel_length], dtype=np.int32), - }) - encoder_output = outputs["encoder_output"] - if encoder_output.ndim == 3: - encoder_output = np.transpose(encoder_output, (0, 2, 1)) - length = torch.tensor(outputs["encoder_output_length"], dtype=torch.long) - return torch.from_numpy(encoder_output.astype(np.float32)), length - - def transcribe(self, audio_path: Path) -> str: - audio, _ = librosa.load(str(audio_path), sr=16000) - mel, mel_length = self._preprocess(audio) - encoder_output, encoder_length = self._encode(mel, mel_length) - token_ids = self.decoder_helper.decode(encoder_output, encoder_length)[0] - return self._tokens_to_text(token_ids) - - def transcribe_many(self, audio_paths: Sequence[Path]) -> List[str]: - results: List[str] = [] - for path in audio_paths: - LOGGER.info("Transcribing %s", path) - results.append(self.transcribe(path)) - return results - - -def _resolve_bundle(args: argparse.Namespace) -> CoreMLModelBundle: - base = Path(args.model_dir) if args.model_dir else None - if base is None and not all([args.preprocessor, args.encoder, args.decoder, args.joint, args.tokenizer]): - raise ValueError("Either --model-dir or explicit model paths are required") - return CoreMLModelBundle( - preprocessor=Path(args.preprocessor) if args.preprocessor else base / "Melspectrogram.mlpackage", - encoder=Path(args.encoder) if args.encoder else base / "ParakeetEncoder.mlpackage", - decoder=Path(args.decoder) if args.decoder else base / "ParakeetDecoder.mlpackage", - joint=Path(args.joint) if args.joint else base / "RNNTJoint.mlpackage", - tokenizer=Path(args.tokenizer) if args.tokenizer else base / "tokenizer.json", - ) - - -def _build_parser() -> argparse.ArgumentParser: - parser = argparse.ArgumentParser(description="Streaming RNNT inference with CoreML components") - parser.add_argument("--model-dir", type=Path, help="Directory containing exported CoreML models") - parser.add_argument("--preprocessor", type=Path, help="Path to the preprocessor .mlpackage") - parser.add_argument("--encoder", type=Path, help="Path to the encoder .mlpackage") - parser.add_argument("--decoder", type=Path, help="Path to the decoder .mlpackage") - parser.add_argument("--joint", type=Path, help="Path to the joint .mlpackage") - parser.add_argument("--tokenizer", type=Path, help="Path to tokenizer JSON") - parser.add_argument("audio", nargs="+", help="Audio files to transcribe") - parser.add_argument("--blank-id", type=int, help="Blank token id") - parser.add_argument("--num-layers", type=int, default=2, help="Prediction network layer count") - parser.add_argument("--hidden-size", type=int, default=640, help="Prediction network hidden size") - parser.add_argument("--durations", type=int, nargs="+", default=[0, 1, 2, 3, 4], help="RNNT duration bucket values") - parser.add_argument("--verbose", "-v", action="count", default=0, help="Increase log verbosity") - return parser - - -def _configure_logging(verbosity: int) -> None: - level = logging.WARNING - (10 * verbosity) - logging.basicConfig(level=max(logging.DEBUG, level), format="[%(levelname)s] %(message)s") - - -def main(argv: Optional[Sequence[str]] = None) -> None: - parser = _build_parser() - args = parser.parse_args(argv) - _configure_logging(args.verbose) - - try: - bundle = _resolve_bundle(args) - transcriber = StreamingTranscriber( - bundle, - blank_id=args.blank_id, - num_layers=args.num_layers, - hidden_size=args.hidden_size, - durations=tuple(args.durations), - ) - transcripts = transcriber.transcribe_many([Path(p) for p in args.audio]) - for path, text in zip(args.audio, transcripts): - print(f"{path}: {text}") - except ValueError as exc: - parser.error(str(exc)) - - -if __name__ == "__main__": # pragma: no cover - main() diff --git a/convert/parakeet-tdt-v2-0.6b/coreml/uv.lock b/convert/parakeet-tdt-v2-0.6b/coreml/uv.lock deleted file mode 100644 index 627100c36a840640c5f07947fd14ac1ef80534a0..0000000000000000000000000000000000000000 --- a/convert/parakeet-tdt-v2-0.6b/coreml/uv.lock +++ /dev/null @@ -1,4725 +0,0 @@ -version = 1 -requires-python = "==3.10.12" -resolution-markers = [ - "sys_platform != 'linux'", - "sys_platform == 'linux'", -] - -[[package]] -name = "absl-py" -version = "2.3.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/03/15/18693af986560a5c3cc0b84a8046b536ffb2cdb536e03cce897f2759e284/absl_py-2.3.0.tar.gz", hash = "sha256:d96fda5c884f1b22178852f30ffa85766d50b99e00775ea626c23304f582fc4f", size = 116400 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/87/04/9d75e1d3bb4ab8ec67ff10919476ccdee06c098bcfcf3a352da5f985171d/absl_py-2.3.0-py3-none-any.whl", hash = "sha256:9824a48b654a306168f63e0d97714665f8490b8d89ec7bf2efc24bf67cf579b3", size = 135657 }, -] - -[[package]] -name = "accelerate" -version = "1.8.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "huggingface-hub" }, - { name = "numpy" }, - { name = "packaging" }, - { name = "psutil" }, - { name = "pyyaml" }, - { name = "safetensors" }, - { name = "torch" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/bd/c2/b9e33ad13232606dded4c546e654fb06a15f1dbcbd95d81c9f9dd3ccc771/accelerate-1.8.1.tar.gz", hash = "sha256:f60df931671bc4e75077b852990469d4991ce8bd3a58e72375c3c95132034db9", size = 380872 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/91/d9/e044c9d42d8ad9afa96533b46ecc9b7aea893d362b3c52bd78fb9fe4d7b3/accelerate-1.8.1-py3-none-any.whl", hash = "sha256:c47b8994498875a2b1286e945bd4d20e476956056c7941d512334f4eb44ff991", size = 365338 }, -] - -[[package]] -name = "aiohappyeyeballs" -version = "2.6.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/26/30/f84a107a9c4331c14b2b586036f40965c128aa4fee4dda5d3d51cb14ad54/aiohappyeyeballs-2.6.1.tar.gz", hash = "sha256:c3f9d0113123803ccadfdf3f0faa505bc78e6a72d1cc4806cbd719826e943558", size = 22760 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/0f/15/5bf3b99495fb160b63f95972b81750f18f7f4e02ad051373b669d17d44f2/aiohappyeyeballs-2.6.1-py3-none-any.whl", hash = "sha256:f349ba8f4b75cb25c99c5c2d84e997e485204d2902a9597802b0371f09331fb8", size = 15265 }, -] - -[[package]] -name = "aiohttp" -version = "3.12.13" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "aiohappyeyeballs" }, - { name = "aiosignal" }, - { name = "async-timeout", marker = "python_full_version == '3.10.12'" }, - { name = "attrs" }, - { name = "frozenlist" }, - { name = "multidict" }, - { name = "propcache" }, - { name = "yarl" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/42/6e/ab88e7cb2a4058bed2f7870276454f85a7c56cd6da79349eb314fc7bbcaa/aiohttp-3.12.13.tar.gz", hash = "sha256:47e2da578528264a12e4e3dd8dd72a7289e5f812758fe086473fab037a10fcce", size = 7819160 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/8b/2d/27e4347660723738b01daa3f5769d56170f232bf4695dd4613340da135bb/aiohttp-3.12.13-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:5421af8f22a98f640261ee48aae3a37f0c41371e99412d55eaf2f8a46d5dad29", size = 702090 }, - { url = "https://files.pythonhosted.org/packages/10/0b/4a8e0468ee8f2b9aff3c05f2c3a6be1dfc40b03f68a91b31041d798a9510/aiohttp-3.12.13-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:0fcda86f6cb318ba36ed8f1396a6a4a3fd8f856f84d426584392083d10da4de0", size = 478440 }, - { url = "https://files.pythonhosted.org/packages/b9/c8/2086df2f9a842b13feb92d071edf756be89250f404f10966b7bc28317f17/aiohttp-3.12.13-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:4cd71c9fb92aceb5a23c4c39d8ecc80389c178eba9feab77f19274843eb9412d", size = 466215 }, - { url = "https://files.pythonhosted.org/packages/a7/3d/d23e5bd978bc8012a65853959b13bd3b55c6e5afc172d89c26ad6624c52b/aiohttp-3.12.13-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:34ebf1aca12845066c963016655dac897651e1544f22a34c9b461ac3b4b1d3aa", size = 1648271 }, - { url = "https://files.pythonhosted.org/packages/31/31/e00122447bb137591c202786062f26dd383574c9f5157144127077d5733e/aiohttp-3.12.13-cp310-cp310-manylinux_2_17_armv7l.manylinux2014_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:893a4639694c5b7edd4bdd8141be296042b6806e27cc1d794e585c43010cc294", size = 1622329 }, - { url = "https://files.pythonhosted.org/packages/04/01/caef70be3ac38986969045f21f5fb802ce517b3f371f0615206bf8aa6423/aiohttp-3.12.13-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:663d8ee3ffb3494502ebcccb49078faddbb84c1d870f9c1dd5a29e85d1f747ce", size = 1694734 }, - { url = "https://files.pythonhosted.org/packages/3f/15/328b71fedecf69a9fd2306549b11c8966e420648a3938d75d3ed5bcb47f6/aiohttp-3.12.13-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:f0f8f6a85a0006ae2709aa4ce05749ba2cdcb4b43d6c21a16c8517c16593aabe", size = 1737049 }, - { url = "https://files.pythonhosted.org/packages/e6/7a/d85866a642158e1147c7da5f93ad66b07e5452a84ec4258e5f06b9071e92/aiohttp-3.12.13-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1582745eb63df267c92d8b61ca655a0ce62105ef62542c00a74590f306be8cb5", size = 1641715 }, - { url = "https://files.pythonhosted.org/packages/14/57/3588800d5d2f5f3e1cb6e7a72747d1abc1e67ba5048e8b845183259c2e9b/aiohttp-3.12.13-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:d59227776ee2aa64226f7e086638baa645f4b044f2947dbf85c76ab11dcba073", size = 1581836 }, - { url = "https://files.pythonhosted.org/packages/2f/55/c913332899a916d85781aa74572f60fd98127449b156ad9c19e23135b0e4/aiohttp-3.12.13-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:06b07c418bde1c8e737d8fa67741072bd3f5b0fb66cf8c0655172188c17e5fa6", size = 1625685 }, - { url = "https://files.pythonhosted.org/packages/4c/34/26cded195f3bff128d6a6d58d7a0be2ae7d001ea029e0fe9008dcdc6a009/aiohttp-3.12.13-cp310-cp310-musllinux_1_2_armv7l.whl", hash = "sha256:9445c1842680efac0f81d272fd8db7163acfcc2b1436e3f420f4c9a9c5a50795", size = 1636471 }, - { url = "https://files.pythonhosted.org/packages/19/21/70629ca006820fccbcec07f3cd5966cbd966e2d853d6da55339af85555b9/aiohttp-3.12.13-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:09c4767af0b0b98c724f5d47f2bf33395c8986995b0a9dab0575ca81a554a8c0", size = 1611923 }, - { url = "https://files.pythonhosted.org/packages/31/80/7fa3f3bebf533aa6ae6508b51ac0de9965e88f9654fa679cc1a29d335a79/aiohttp-3.12.13-cp310-cp310-musllinux_1_2_ppc64le.whl", hash = "sha256:f3854fbde7a465318ad8d3fc5bef8f059e6d0a87e71a0d3360bb56c0bf87b18a", size = 1691511 }, - { url = "https://files.pythonhosted.org/packages/0f/7a/359974653a3cdd3e9cee8ca10072a662c3c0eb46a359c6a1f667b0296e2f/aiohttp-3.12.13-cp310-cp310-musllinux_1_2_s390x.whl", hash = "sha256:2332b4c361c05ecd381edb99e2a33733f3db906739a83a483974b3df70a51b40", size = 1714751 }, - { url = "https://files.pythonhosted.org/packages/2d/24/0aa03d522171ce19064347afeefadb008be31ace0bbb7d44ceb055700a14/aiohttp-3.12.13-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:1561db63fa1b658cd94325d303933553ea7d89ae09ff21cc3bcd41b8521fbbb6", size = 1643090 }, - { url = "https://files.pythonhosted.org/packages/86/2e/7d4b0026a41e4b467e143221c51b279083b7044a4b104054f5c6464082ff/aiohttp-3.12.13-cp310-cp310-win32.whl", hash = "sha256:a0be857f0b35177ba09d7c472825d1b711d11c6d0e8a2052804e3b93166de1ad", size = 427526 }, - { url = "https://files.pythonhosted.org/packages/17/de/34d998da1e7f0de86382160d039131e9b0af1962eebfe53dda2b61d250e7/aiohttp-3.12.13-cp310-cp310-win_amd64.whl", hash = "sha256:fcc30ad4fb5cb41a33953292d45f54ef4066746d625992aeac33b8c681173178", size = 450734 }, -] - -[[package]] -name = "aiosignal" -version = "1.3.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "frozenlist" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/ba/b5/6d55e80f6d8a08ce22b982eafa278d823b541c925f11ee774b0b9c43473d/aiosignal-1.3.2.tar.gz", hash = "sha256:a8c255c66fafb1e499c9351d0bf32ff2d8a0321595ebac3b93713656d2436f54", size = 19424 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/ec/6a/bc7e17a3e87a2985d3e8f4da4cd0f481060eb78fb08596c42be62c90a4d9/aiosignal-1.3.2-py2.py3-none-any.whl", hash = "sha256:45cde58e409a301715980c2b01d0c28bdde3770d8290b5eb2173759d9acb31a5", size = 7597 }, -] - -[[package]] -name = "alembic" -version = "1.16.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "mako" }, - { name = "sqlalchemy" }, - { name = "tomli", marker = "python_full_version == '3.10.12'" }, - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/9c/35/116797ff14635e496bbda0c168987f5326a6555b09312e9b817e360d1f56/alembic-1.16.2.tar.gz", hash = "sha256:e53c38ff88dadb92eb22f8b150708367db731d58ad7e9d417c9168ab516cbed8", size = 1963563 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/dd/e2/88e425adac5ad887a087c38d04fe2030010572a3e0e627f8a6e8c33eeda8/alembic-1.16.2-py3-none-any.whl", hash = "sha256:5f42e9bd0afdbd1d5e3ad856c01754530367debdebf21ed6894e34af52b3bb03", size = 242717 }, -] - -[[package]] -name = "aliyun-python-sdk-core" -version = "2.16.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "cryptography" }, - { name = "jmespath" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/3e/09/da9f58eb38b4fdb97ba6523274fbf445ef6a06be64b433693da8307b4bec/aliyun-python-sdk-core-2.16.0.tar.gz", hash = "sha256:651caad597eb39d4fad6cf85133dffe92837d53bdf62db9d8f37dab6508bb8f9", size = 449555 } - -[[package]] -name = "aliyun-python-sdk-kms" -version = "2.16.5" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "aliyun-python-sdk-core" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/a8/2c/9877d0e6b18ecf246df671ac65a5d1d9fecbf85bdcb5d43efbde0d4662eb/aliyun-python-sdk-kms-2.16.5.tar.gz", hash = "sha256:f328a8a19d83ecbb965ffce0ec1e9930755216d104638cd95ecd362753b813b3", size = 12018 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/11/5c/0132193d7da2c735669a1ed103b142fd63c9455984d48c5a88a1a516efaa/aliyun_python_sdk_kms-2.16.5-py2.py3-none-any.whl", hash = "sha256:24b6cdc4fd161d2942619479c8d050c63ea9cd22b044fe33b60bbb60153786f0", size = 99495 }, -] - -[[package]] -name = "annotated-types" -version = "0.7.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/ee/67/531ea369ba64dcff5ec9c3402f9f51bf748cec26dde048a2f973a4eea7f5/annotated_types-0.7.0.tar.gz", hash = "sha256:aff07c09a53a08bc8cfccb9c85b05f1aa9a2a6f23728d790723543408344ce89", size = 16081 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/78/b6/6307fbef88d9b5ee7421e68d78a9f162e0da4900bc5f5793f6d3d0e34fb8/annotated_types-0.7.0-py3-none-any.whl", hash = "sha256:1f02e8b43a8fbbc3f3e0d4f0f4bfc8131bcb4eebe8849b8e5c773f3a1c582a53", size = 13643 }, -] - -[[package]] -name = "antlr4-python3-runtime" -version = "4.9.3" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/3e/38/7859ff46355f76f8d19459005ca000b6e7012f2f1ca597746cbcd1fbfe5e/antlr4-python3-runtime-4.9.3.tar.gz", hash = "sha256:f224469b4168294902bb1efa80a8bf7855f24c99aef99cbefc1bcd3cce77881b", size = 117034 } - -[[package]] -name = "anyio" -version = "4.9.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "exceptiongroup", marker = "python_full_version == '3.10.12'" }, - { name = "idna" }, - { name = "sniffio" }, - { name = "typing-extensions", marker = "python_full_version == '3.10.12'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/95/7d/4c1bd541d4dffa1b52bd83fb8527089e097a106fc90b467a7313b105f840/anyio-4.9.0.tar.gz", hash = "sha256:673c0c244e15788651a4ff38710fea9675823028a6f08a5eda409e0c9840a028", size = 190949 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/a1/ee/48ca1a7c89ffec8b6a0c5d02b89c305671d5ffd8d3c94acf8b8c408575bb/anyio-4.9.0-py3-none-any.whl", hash = "sha256:9f76d541cad6e36af7beb62e978876f3b41e3e04f2c1fbf0884604c0a9c4d93c", size = 100916 }, -] - -[[package]] -name = "appnope" -version = "0.1.4" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/35/5d/752690df9ef5b76e169e68d6a129fa6d08a7100ca7f754c89495db3c6019/appnope-0.1.4.tar.gz", hash = "sha256:1de3860566df9caf38f01f86f65e0e13e379af54f9e4bee1e66b48f2efffd1ee", size = 4170 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/81/29/5ecc3a15d5a33e31b26c11426c45c501e439cb865d0bff96315d86443b78/appnope-0.1.4-py2.py3-none-any.whl", hash = "sha256:502575ee11cd7a28c0205f379b525beefebab9d161b7c964670864014ed7213c", size = 4321 }, -] - -[[package]] -name = "argon2-cffi" -version = "25.1.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "argon2-cffi-bindings" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/0e/89/ce5af8a7d472a67cc819d5d998aa8c82c5d860608c4db9f46f1162d7dab9/argon2_cffi-25.1.0.tar.gz", hash = "sha256:694ae5cc8a42f4c4e2bf2ca0e64e51e23a040c6a517a85074683d3959e1346c1", size = 45706 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/4f/d3/a8b22fa575b297cd6e3e3b0155c7e25db170edf1c74783d6a31a2490b8d9/argon2_cffi-25.1.0-py3-none-any.whl", hash = "sha256:fdc8b074db390fccb6eb4a3604ae7231f219aa669a2652e0f20e16ba513d5741", size = 14657 }, -] - -[[package]] -name = "argon2-cffi-bindings" -version = "21.2.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "cffi" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/b9/e9/184b8ccce6683b0aa2fbb7ba5683ea4b9c5763f1356347f1312c32e3c66e/argon2-cffi-bindings-21.2.0.tar.gz", hash = "sha256:bb89ceffa6c791807d1305ceb77dbfacc5aa499891d2c55661c6459651fc39e3", size = 1779911 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/d4/13/838ce2620025e9666aa8f686431f67a29052241692a3dd1ae9d3692a89d3/argon2_cffi_bindings-21.2.0-cp36-abi3-macosx_10_9_x86_64.whl", hash = "sha256:ccb949252cb2ab3a08c02024acb77cfb179492d5701c7cbdbfd776124d4d2367", size = 29658 }, - { url = "https://files.pythonhosted.org/packages/b3/02/f7f7bb6b6af6031edb11037639c697b912e1dea2db94d436e681aea2f495/argon2_cffi_bindings-21.2.0-cp36-abi3-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:9524464572e12979364b7d600abf96181d3541da11e23ddf565a32e70bd4dc0d", size = 80583 }, - { url = "https://files.pythonhosted.org/packages/ec/f7/378254e6dd7ae6f31fe40c8649eea7d4832a42243acaf0f1fff9083b2bed/argon2_cffi_bindings-21.2.0-cp36-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b746dba803a79238e925d9046a63aa26bf86ab2a2fe74ce6b009a1c3f5c8f2ae", size = 86168 }, - { url = "https://files.pythonhosted.org/packages/74/f6/4a34a37a98311ed73bb80efe422fed95f2ac25a4cacc5ae1d7ae6a144505/argon2_cffi_bindings-21.2.0-cp36-abi3-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:58ed19212051f49a523abb1dbe954337dc82d947fb6e5a0da60f7c8471a8476c", size = 82709 }, - { url = "https://files.pythonhosted.org/packages/74/2b/73d767bfdaab25484f7e7901379d5f8793cccbb86c6e0cbc4c1b96f63896/argon2_cffi_bindings-21.2.0-cp36-abi3-musllinux_1_1_aarch64.whl", hash = "sha256:bd46088725ef7f58b5a1ef7ca06647ebaf0eb4baff7d1d0d177c6cc8744abd86", size = 83613 }, - { url = "https://files.pythonhosted.org/packages/4f/fd/37f86deef67ff57c76f137a67181949c2d408077e2e3dd70c6c42912c9bf/argon2_cffi_bindings-21.2.0-cp36-abi3-musllinux_1_1_i686.whl", hash = "sha256:8cd69c07dd875537a824deec19f978e0f2078fdda07fd5c42ac29668dda5f40f", size = 84583 }, - { url = "https://files.pythonhosted.org/packages/6f/52/5a60085a3dae8fded8327a4f564223029f5f54b0cb0455a31131b5363a01/argon2_cffi_bindings-21.2.0-cp36-abi3-musllinux_1_1_x86_64.whl", hash = "sha256:f1152ac548bd5b8bcecfb0b0371f082037e47128653df2e8ba6e914d384f3c3e", size = 88475 }, - { url = "https://files.pythonhosted.org/packages/8b/95/143cd64feb24a15fa4b189a3e1e7efbaeeb00f39a51e99b26fc62fbacabd/argon2_cffi_bindings-21.2.0-cp36-abi3-win32.whl", hash = "sha256:603ca0aba86b1349b147cab91ae970c63118a0f30444d4bc80355937c950c082", size = 27698 }, - { url = "https://files.pythonhosted.org/packages/37/2c/e34e47c7dee97ba6f01a6203e0383e15b60fb85d78ac9a15cd066f6fe28b/argon2_cffi_bindings-21.2.0-cp36-abi3-win_amd64.whl", hash = "sha256:b2ef1c30440dbbcba7a5dc3e319408b59676e2e039e2ae11a8775ecf482b192f", size = 30817 }, - { url = "https://files.pythonhosted.org/packages/5a/e4/bf8034d25edaa495da3c8a3405627d2e35758e44ff6eaa7948092646fdcc/argon2_cffi_bindings-21.2.0-cp38-abi3-macosx_10_9_universal2.whl", hash = "sha256:e415e3f62c8d124ee16018e491a009937f8cf7ebf5eb430ffc5de21b900dad93", size = 53104 }, -] - -[[package]] -name = "arrow" -version = "1.3.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "python-dateutil" }, - { name = "types-python-dateutil" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/2e/00/0f6e8fcdb23ea632c866620cc872729ff43ed91d284c866b515c6342b173/arrow-1.3.0.tar.gz", hash = "sha256:d4540617648cb5f895730f1ad8c82a65f2dad0166f57b75f3ca54759c4d67a85", size = 131960 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/f8/ed/e97229a566617f2ae958a6b13e7cc0f585470eac730a73e9e82c32a3cdd2/arrow-1.3.0-py3-none-any.whl", hash = "sha256:c728b120ebc00eb84e01882a6f5e7927a53960aa990ce7dd2b10f39005a67f80", size = 66419 }, -] - -[[package]] -name = "asteroid-filterbanks" -version = "0.4.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "numpy" }, - { name = "torch" }, - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/90/fa/5c2be1f96dc179f83cdd3bb267edbd1f47d08f756785c016d5c2163901a7/asteroid-filterbanks-0.4.0.tar.gz", hash = "sha256:415f89d1dcf2b13b35f03f7a9370968ac4e6fa6800633c522dac992b283409b9", size = 24599 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/c5/7c/83ff6046176a675e6a1e8aeefed8892cd97fe7c46af93cc540d1b24b8323/asteroid_filterbanks-0.4.0-py3-none-any.whl", hash = "sha256:4932ac8b6acc6e08fb87cbe8ece84215b5a74eee284fe83acf3540a72a02eaf5", size = 29912 }, -] - -[[package]] -name = "asttokens" -version = "3.0.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/4a/e7/82da0a03e7ba5141f05cce0d302e6eed121ae055e0456ca228bf693984bc/asttokens-3.0.0.tar.gz", hash = "sha256:0dcd8baa8d62b0c1d118b399b2ddba3c4aff271d0d7a9e0d4c1681c79035bbc7", size = 61978 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/25/8a/c46dcc25341b5bce5472c718902eb3d38600a903b14fa6aeecef3f21a46f/asttokens-3.0.0-py3-none-any.whl", hash = "sha256:e3078351a059199dd5138cb1c706e6430c05eff2ff136af5eb4790f9d28932e2", size = 26918 }, -] - -[[package]] -name = "async-lru" -version = "2.0.5" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "typing-extensions", marker = "python_full_version == '3.10.12'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/b2/4d/71ec4d3939dc755264f680f6c2b4906423a304c3d18e96853f0a595dfe97/async_lru-2.0.5.tar.gz", hash = "sha256:481d52ccdd27275f42c43a928b4a50c3bfb2d67af4e78b170e3e0bb39c66e5bb", size = 10380 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/03/49/d10027df9fce941cb8184e78a02857af36360d33e1721df81c5ed2179a1a/async_lru-2.0.5-py3-none-any.whl", hash = "sha256:ab95404d8d2605310d345932697371a5f40def0487c03d6d0ad9138de52c9943", size = 6069 }, -] - -[[package]] -name = "async-timeout" -version = "5.0.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/a5/ae/136395dfbfe00dfc94da3f3e136d0b13f394cba8f4841120e34226265780/async_timeout-5.0.1.tar.gz", hash = "sha256:d9321a7a3d5a6a5e187e824d2fa0793ce379a202935782d555d6e9d2735677d3", size = 9274 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/fe/ba/e2081de779ca30d473f21f5b30e0e737c438205440784c7dfc81efc2b029/async_timeout-5.0.1-py3-none-any.whl", hash = "sha256:39e3809566ff85354557ec2398b55e096c8364bacac9405a7a1fa429e77fe76c", size = 6233 }, -] - -[[package]] -name = "attrs" -version = "25.3.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/5a/b0/1367933a8532ee6ff8d63537de4f1177af4bff9f3e829baf7331f595bb24/attrs-25.3.0.tar.gz", hash = "sha256:75d7cefc7fb576747b2c81b4442d4d4a1ce0900973527c011d1030fd3bf4af1b", size = 812032 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/77/06/bb80f5f86020c4551da315d78b3ab75e8228f89f0162f2c3a819e407941a/attrs-25.3.0-py3-none-any.whl", hash = "sha256:427318ce031701fea540783410126f03899a97ffc6f61596ad581ac2e40e3bc3", size = 63815 }, -] - -[[package]] -name = "audioread" -version = "3.0.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/db/d2/87016ca9f083acadffb2d8da59bfa3253e4da7eeb9f71fb8e7708dc97ecd/audioread-3.0.1.tar.gz", hash = "sha256:ac5460a5498c48bdf2e8e767402583a4dcd13f4414d286f42ce4379e8b35066d", size = 116513 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/57/8d/30aa32745af16af0a9a650115fbe81bde7c610ed5c21b381fca0196f3a7f/audioread-3.0.1-py3-none-any.whl", hash = "sha256:4cdce70b8adc0da0a3c9e0d85fb10b3ace30fbdf8d1670fd443929b61d117c33", size = 23492 }, -] - -[[package]] -name = "babel" -version = "2.17.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/7d/6b/d52e42361e1aa00709585ecc30b3f9684b3ab62530771402248b1b1d6240/babel-2.17.0.tar.gz", hash = "sha256:0c54cffb19f690cdcc52a3b50bcbf71e07a808d1c80d549f2459b9d2cf0afb9d", size = 9951852 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/b7/b8/3fe70c75fe32afc4bb507f75563d39bc5642255d1d94f1f23604725780bf/babel-2.17.0-py3-none-any.whl", hash = "sha256:4d0b53093fdfb4b21c92b5213dba5a1b23885afa8383709427046b21c366e5f2", size = 10182537 }, -] - -[[package]] -name = "backports-datetime-fromisoformat" -version = "2.0.3" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/71/81/eff3184acb1d9dc3ce95a98b6f3c81a49b4be296e664db8e1c2eeabef3d9/backports_datetime_fromisoformat-2.0.3.tar.gz", hash = "sha256:b58edc8f517b66b397abc250ecc737969486703a66eb97e01e6d51291b1a139d", size = 23588 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/42/4b/d6b051ca4b3d76f23c2c436a9669f3be616b8cf6461a7e8061c7c4269642/backports_datetime_fromisoformat-2.0.3-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:5f681f638f10588fa3c101ee9ae2b63d3734713202ddfcfb6ec6cea0778a29d4", size = 27561 }, - { url = "https://files.pythonhosted.org/packages/6d/40/e39b0d471e55eb1b5c7c81edab605c02f71c786d59fb875f0a6f23318747/backports_datetime_fromisoformat-2.0.3-cp310-cp310-macosx_11_0_universal2.whl", hash = "sha256:cd681460e9142f1249408e5aee6d178c6d89b49e06d44913c8fdfb6defda8d1c", size = 34448 }, - { url = "https://files.pythonhosted.org/packages/f2/28/7a5c87c5561d14f1c9af979231fdf85d8f9fad7a95ff94e56d2205e2520a/backports_datetime_fromisoformat-2.0.3-cp310-cp310-macosx_11_0_x86_64.whl", hash = "sha256:ee68bc8735ae5058695b76d3bb2aee1d137c052a11c8303f1e966aa23b72b65b", size = 27093 }, - { url = "https://files.pythonhosted.org/packages/80/ba/f00296c5c4536967c7d1136107fdb91c48404fe769a4a6fd5ab045629af8/backports_datetime_fromisoformat-2.0.3-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8273fe7932db65d952a43e238318966eab9e49e8dd546550a41df12175cc2be4", size = 52836 }, - { url = "https://files.pythonhosted.org/packages/e3/92/bb1da57a069ddd601aee352a87262c7ae93467e66721d5762f59df5021a6/backports_datetime_fromisoformat-2.0.3-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:39d57ea50aa5a524bb239688adc1d1d824c31b6094ebd39aa164d6cadb85de22", size = 52798 }, - { url = "https://files.pythonhosted.org/packages/df/ef/b6cfd355982e817ccdb8d8d109f720cab6e06f900784b034b30efa8fa832/backports_datetime_fromisoformat-2.0.3-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:ac6272f87693e78209dc72e84cf9ab58052027733cd0721c55356d3c881791cf", size = 52891 }, - { url = "https://files.pythonhosted.org/packages/37/39/b13e3ae8a7c5d88b68a6e9248ffe7066534b0cfe504bf521963e61b6282d/backports_datetime_fromisoformat-2.0.3-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:44c497a71f80cd2bcfc26faae8857cf8e79388e3d5fbf79d2354b8c360547d58", size = 52955 }, - { url = "https://files.pythonhosted.org/packages/1e/e4/70cffa3ce1eb4f2ff0c0d6f5d56285aacead6bd3879b27a2ba57ab261172/backports_datetime_fromisoformat-2.0.3-cp310-cp310-win_amd64.whl", hash = "sha256:6335a4c9e8af329cb1ded5ab41a666e1448116161905a94e054f205aa6d263bc", size = 29323 }, - { url = "https://files.pythonhosted.org/packages/be/03/7eaa9f9bf290395d57fd30d7f1f2f9dff60c06a31c237dc2beb477e8f899/backports_datetime_fromisoformat-2.0.3-pp310-pypy310_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:90e202e72a3d5aae673fcc8c9a4267d56b2f532beeb9173361293625fe4d2039", size = 28980 }, - { url = "https://files.pythonhosted.org/packages/47/80/a0ecf33446c7349e79f54cc532933780341d20cff0ee12b5bfdcaa47067e/backports_datetime_fromisoformat-2.0.3-pp310-pypy310_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2df98ef1b76f5a58bb493dda552259ba60c3a37557d848e039524203951c9f06", size = 28449 }, -] - -[[package]] -name = "beautifulsoup4" -version = "4.13.4" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "soupsieve" }, - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/d8/e4/0c4c39e18fd76d6a628d4dd8da40543d136ce2d1752bd6eeeab0791f4d6b/beautifulsoup4-4.13.4.tar.gz", hash = "sha256:dbb3c4e1ceae6aefebdaf2423247260cd062430a410e38c66f2baa50a8437195", size = 621067 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/50/cd/30110dc0ffcf3b131156077b90e9f60ed75711223f306da4db08eff8403b/beautifulsoup4-4.13.4-py3-none-any.whl", hash = "sha256:9bbbb14bfde9d79f38b8cd5f8c7c85f4b8f2523190ebed90e950a8dea4cb1c4b", size = 187285 }, -] - -[[package]] -name = "bleach" -version = "6.2.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "webencodings" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/76/9a/0e33f5054c54d349ea62c277191c020c2d6ef1d65ab2cb1993f91ec846d1/bleach-6.2.0.tar.gz", hash = "sha256:123e894118b8a599fd80d3ec1a6d4cc7ce4e5882b1317a7e1ba69b56e95f991f", size = 203083 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/fc/55/96142937f66150805c25c4d0f31ee4132fd33497753400734f9dfdcbdc66/bleach-6.2.0-py3-none-any.whl", hash = "sha256:117d9c6097a7c3d22fd578fcd8d35ff1e125df6736f554da4e432fdd63f31e5e", size = 163406 }, -] - -[package.optional-dependencies] -css = [ - { name = "tinycss2" }, -] - -[[package]] -name = "braceexpand" -version = "0.1.7" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/54/93/badd4f5ccf25209f3fef2573073da9fe4a45a3da99fca2f800f942130c0f/braceexpand-0.1.7.tar.gz", hash = "sha256:e6e539bd20eaea53547472ff94f4fb5c3d3bf9d0a89388c4b56663aba765f705", size = 7777 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/fa/93/e8c04e80e82391a6e51f218ca49720f64236bc824e92152a2633b74cf7ab/braceexpand-0.1.7-py2.py3-none-any.whl", hash = "sha256:91332d53de7828103dcae5773fb43bc34950b0c8160e35e0f44c4427a3b85014", size = 5923 }, -] - -[[package]] -name = "cattrs" -version = "25.1.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "attrs" }, - { name = "exceptiongroup", marker = "python_full_version == '3.10.12'" }, - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/57/2b/561d78f488dcc303da4639e02021311728fb7fda8006dd2835550cddd9ed/cattrs-25.1.1.tar.gz", hash = "sha256:c914b734e0f2d59e5b720d145ee010f1fd9a13ee93900922a2f3f9d593b8382c", size = 435016 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/18/b0/215274ef0d835bbc1056392a367646648b6084e39d489099959aefcca2af/cattrs-25.1.1-py3-none-any.whl", hash = "sha256:1b40b2d3402af7be79a7e7e097a9b4cd16d4c06e6d526644b0b26a063a1cc064", size = 69386 }, -] - -[[package]] -name = "certifi" -version = "2025.6.15" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/73/f7/f14b46d4bcd21092d7d3ccef689615220d8a08fb25e564b65d20738e672e/certifi-2025.6.15.tar.gz", hash = "sha256:d747aa5a8b9bbbb1bb8c22bb13e22bd1f18e9796defa16bab421f7f7a317323b", size = 158753 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/84/ae/320161bd181fc06471eed047ecce67b693fd7515b16d495d8932db763426/certifi-2025.6.15-py3-none-any.whl", hash = "sha256:2e0c7ce7cb5d8f8634ca55d2ba7e6ec2689a2fd6537d8dec1296a477a4910057", size = 157650 }, -] - -[[package]] -name = "cffi" -version = "1.17.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "pycparser" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/fc/97/c783634659c2920c3fc70419e3af40972dbaf758daa229a7d6ea6135c90d/cffi-1.17.1.tar.gz", hash = "sha256:1c39c6016c32bc48dd54561950ebd6836e1670f2ae46128f67cf49e789c52824", size = 516621 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/90/07/f44ca684db4e4f08a3fdc6eeb9a0d15dc6883efc7b8c90357fdbf74e186c/cffi-1.17.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:df8b1c11f177bc2313ec4b2d46baec87a5f3e71fc8b45dab2ee7cae86d9aba14", size = 182191 }, - { url = "https://files.pythonhosted.org/packages/08/fd/cc2fedbd887223f9f5d170c96e57cbf655df9831a6546c1727ae13fa977a/cffi-1.17.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:8f2cdc858323644ab277e9bb925ad72ae0e67f69e804f4898c070998d50b1a67", size = 178592 }, - { url = "https://files.pythonhosted.org/packages/de/cc/4635c320081c78d6ffc2cab0a76025b691a91204f4aa317d568ff9280a2d/cffi-1.17.1-cp310-cp310-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:edae79245293e15384b51f88b00613ba9f7198016a5948b5dddf4917d4d26382", size = 426024 }, - { url = "https://files.pythonhosted.org/packages/b6/7b/3b2b250f3aab91abe5f8a51ada1b717935fdaec53f790ad4100fe2ec64d1/cffi-1.17.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:45398b671ac6d70e67da8e4224a065cec6a93541bb7aebe1b198a61b58c7b702", size = 448188 }, - { url = "https://files.pythonhosted.org/packages/d3/48/1b9283ebbf0ec065148d8de05d647a986c5f22586b18120020452fff8f5d/cffi-1.17.1-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ad9413ccdeda48c5afdae7e4fa2192157e991ff761e7ab8fdd8926f40b160cc3", size = 455571 }, - { url = "https://files.pythonhosted.org/packages/40/87/3b8452525437b40f39ca7ff70276679772ee7e8b394934ff60e63b7b090c/cffi-1.17.1-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:5da5719280082ac6bd9aa7becb3938dc9f9cbd57fac7d2871717b1feb0902ab6", size = 436687 }, - { url = "https://files.pythonhosted.org/packages/8d/fb/4da72871d177d63649ac449aec2e8a29efe0274035880c7af59101ca2232/cffi-1.17.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2bb1a08b8008b281856e5971307cc386a8e9c5b625ac297e853d36da6efe9c17", size = 446211 }, - { url = "https://files.pythonhosted.org/packages/ab/a0/62f00bcb411332106c02b663b26f3545a9ef136f80d5df746c05878f8c4b/cffi-1.17.1-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:045d61c734659cc045141be4bae381a41d89b741f795af1dd018bfb532fd0df8", size = 461325 }, - { url = "https://files.pythonhosted.org/packages/36/83/76127035ed2e7e27b0787604d99da630ac3123bfb02d8e80c633f218a11d/cffi-1.17.1-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:6883e737d7d9e4899a8a695e00ec36bd4e5e4f18fabe0aca0efe0a4b44cdb13e", size = 438784 }, - { url = "https://files.pythonhosted.org/packages/21/81/a6cd025db2f08ac88b901b745c163d884641909641f9b826e8cb87645942/cffi-1.17.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:6b8b4a92e1c65048ff98cfe1f735ef8f1ceb72e3d5f0c25fdb12087a23da22be", size = 461564 }, - { url = "https://files.pythonhosted.org/packages/f8/fe/4d41c2f200c4a457933dbd98d3cf4e911870877bd94d9656cc0fcb390681/cffi-1.17.1-cp310-cp310-win32.whl", hash = "sha256:c9c3d058ebabb74db66e431095118094d06abf53284d9c81f27300d0e0d8bc7c", size = 171804 }, - { url = "https://files.pythonhosted.org/packages/d1/b6/0b0f5ab93b0df4acc49cae758c81fe4e5ef26c3ae2e10cc69249dfd8b3ab/cffi-1.17.1-cp310-cp310-win_amd64.whl", hash = "sha256:0f048dcf80db46f0098ccac01132761580d28e28bc0f78ae0d58048063317e15", size = 181299 }, -] - -[[package]] -name = "charset-normalizer" -version = "3.4.2" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/e4/33/89c2ced2b67d1c2a61c19c6751aa8902d46ce3dacb23600a283619f5a12d/charset_normalizer-3.4.2.tar.gz", hash = "sha256:5baececa9ecba31eff645232d59845c07aa030f0c81ee70184a90d35099a0e63", size = 126367 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/95/28/9901804da60055b406e1a1c5ba7aac1276fb77f1dde635aabfc7fd84b8ab/charset_normalizer-3.4.2-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:7c48ed483eb946e6c04ccbe02c6b4d1d48e51944b6db70f697e089c193404941", size = 201818 }, - { url = "https://files.pythonhosted.org/packages/d9/9b/892a8c8af9110935e5adcbb06d9c6fe741b6bb02608c6513983048ba1a18/charset_normalizer-3.4.2-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:b2d318c11350e10662026ad0eb71bb51c7812fc8590825304ae0bdd4ac283acd", size = 144649 }, - { url = "https://files.pythonhosted.org/packages/7b/a5/4179abd063ff6414223575e008593861d62abfc22455b5d1a44995b7c101/charset_normalizer-3.4.2-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:9cbfacf36cb0ec2897ce0ebc5d08ca44213af24265bd56eca54bee7923c48fd6", size = 155045 }, - { url = "https://files.pythonhosted.org/packages/3b/95/bc08c7dfeddd26b4be8c8287b9bb055716f31077c8b0ea1cd09553794665/charset_normalizer-3.4.2-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:18dd2e350387c87dabe711b86f83c9c78af772c748904d372ade190b5c7c9d4d", size = 147356 }, - { url = "https://files.pythonhosted.org/packages/a8/2d/7a5b635aa65284bf3eab7653e8b4151ab420ecbae918d3e359d1947b4d61/charset_normalizer-3.4.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8075c35cd58273fee266c58c0c9b670947c19df5fb98e7b66710e04ad4e9ff86", size = 149471 }, - { url = "https://files.pythonhosted.org/packages/ae/38/51fc6ac74251fd331a8cfdb7ec57beba8c23fd5493f1050f71c87ef77ed0/charset_normalizer-3.4.2-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:5bf4545e3b962767e5c06fe1738f951f77d27967cb2caa64c28be7c4563e162c", size = 151317 }, - { url = "https://files.pythonhosted.org/packages/b7/17/edee1e32215ee6e9e46c3e482645b46575a44a2d72c7dfd49e49f60ce6bf/charset_normalizer-3.4.2-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:7a6ab32f7210554a96cd9e33abe3ddd86732beeafc7a28e9955cdf22ffadbab0", size = 146368 }, - { url = "https://files.pythonhosted.org/packages/26/2c/ea3e66f2b5f21fd00b2825c94cafb8c326ea6240cd80a91eb09e4a285830/charset_normalizer-3.4.2-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:b33de11b92e9f75a2b545d6e9b6f37e398d86c3e9e9653c4864eb7e89c5773ef", size = 154491 }, - { url = "https://files.pythonhosted.org/packages/52/47/7be7fa972422ad062e909fd62460d45c3ef4c141805b7078dbab15904ff7/charset_normalizer-3.4.2-cp310-cp310-musllinux_1_2_ppc64le.whl", hash = "sha256:8755483f3c00d6c9a77f490c17e6ab0c8729e39e6390328e42521ef175380ae6", size = 157695 }, - { url = "https://files.pythonhosted.org/packages/2f/42/9f02c194da282b2b340f28e5fb60762de1151387a36842a92b533685c61e/charset_normalizer-3.4.2-cp310-cp310-musllinux_1_2_s390x.whl", hash = "sha256:68a328e5f55ec37c57f19ebb1fdc56a248db2e3e9ad769919a58672958e8f366", size = 154849 }, - { url = "https://files.pythonhosted.org/packages/67/44/89cacd6628f31fb0b63201a618049be4be2a7435a31b55b5eb1c3674547a/charset_normalizer-3.4.2-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:21b2899062867b0e1fde9b724f8aecb1af14f2778d69aacd1a5a1853a597a5db", size = 150091 }, - { url = "https://files.pythonhosted.org/packages/1f/79/4b8da9f712bc079c0f16b6d67b099b0b8d808c2292c937f267d816ec5ecc/charset_normalizer-3.4.2-cp310-cp310-win32.whl", hash = "sha256:e8082b26888e2f8b36a042a58307d5b917ef2b1cacab921ad3323ef91901c71a", size = 98445 }, - { url = "https://files.pythonhosted.org/packages/7d/d7/96970afb4fb66497a40761cdf7bd4f6fca0fc7bafde3a84f836c1f57a926/charset_normalizer-3.4.2-cp310-cp310-win_amd64.whl", hash = "sha256:f69a27e45c43520f5487f27627059b64aaf160415589230992cec34c5e18a509", size = 105782 }, - { url = "https://files.pythonhosted.org/packages/20/94/c5790835a017658cbfabd07f3bfb549140c3ac458cfc196323996b10095a/charset_normalizer-3.4.2-py3-none-any.whl", hash = "sha256:7f56930ab0abd1c45cd15be65cc741c28b1c9a34876ce8c17a2fa107810c0af0", size = 52626 }, -] - -[[package]] -name = "click" -version = "8.2.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "colorama", marker = "platform_system == 'Windows'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/60/6c/8ca2efa64cf75a977a0d7fac081354553ebe483345c734fb6b6515d96bbc/click-8.2.1.tar.gz", hash = "sha256:27c491cc05d968d271d5a1db13e3b5a184636d9d930f148c50b038f0d0646202", size = 286342 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/85/32/10bb5764d90a8eee674e9dc6f4db6a0ab47c8c4d0d83c27f7c39ac415a4d/click-8.2.1-py3-none-any.whl", hash = "sha256:61a3265b914e850b85317d0b3109c7f8cd35a670f963866005d6ef1d5175a12b", size = 102215 }, -] - -[[package]] -name = "cloudpickle" -version = "3.1.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/52/39/069100b84d7418bc358d81669d5748efb14b9cceacd2f9c75f550424132f/cloudpickle-3.1.1.tar.gz", hash = "sha256:b216fa8ae4019d5482a8ac3c95d8f6346115d8835911fd4aefd1a445e4242c64", size = 22113 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/7e/e8/64c37fadfc2816a7701fa8a6ed8d87327c7d54eacfbfb6edab14a2f2be75/cloudpickle-3.1.1-py3-none-any.whl", hash = "sha256:c8c5a44295039331ee9dad40ba100a9c7297b6f988e50e87ccdf3765a668350e", size = 20992 }, -] - -[[package]] -name = "colorama" -version = "0.4.6" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/d8/53/6f443c9a4a8358a93a6792e2acffb9d9d5cb0a5cfd8802644b7b1c9a02e4/colorama-0.4.6.tar.gz", hash = "sha256:08695f5cb7ed6e0531a20572697297273c47b8cae5a63ffc6d6ed5c201be6e44", size = 27697 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/d1/d6/3965ed04c63042e047cb6a3e6ed1a63a35087b6a609aa3a15ed8ac56c221/colorama-0.4.6-py2.py3-none-any.whl", hash = "sha256:4f1d9991f5acc0ca119f9d443620b77f9d6b33703e51011c16baf57afb285fc6", size = 25335 }, -] - -[[package]] -name = "colorlog" -version = "6.9.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "colorama", marker = "sys_platform == 'win32'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/d3/7a/359f4d5df2353f26172b3cc39ea32daa39af8de522205f512f458923e677/colorlog-6.9.0.tar.gz", hash = "sha256:bfba54a1b93b94f54e1f4fe48395725a3d92fd2a4af702f6bd70946bdc0c6ac2", size = 16624 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/e3/51/9b208e85196941db2f0654ad0357ca6388ab3ed67efdbfc799f35d1f83aa/colorlog-6.9.0-py3-none-any.whl", hash = "sha256:5906e71acd67cb07a71e779c47c4bcb45fb8c2993eebe9e5adcd6a6f1b283eff", size = 11424 }, -] - -[[package]] -name = "comm" -version = "0.2.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "traitlets" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/e9/a8/fb783cb0abe2b5fded9f55e5703015cdf1c9c85b3669087c538dd15a6a86/comm-0.2.2.tar.gz", hash = "sha256:3fd7a84065306e07bea1773df6eb8282de51ba82f77c72f9c85716ab11fe980e", size = 6210 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/e6/75/49e5bfe642f71f272236b5b2d2691cf915a7283cc0ceda56357b61daa538/comm-0.2.2-py3-none-any.whl", hash = "sha256:e6fb86cb70ff661ee8c9c14e7d36d6de3b4066f1441be4063df9c5009f0a64d3", size = 7180 }, -] - -[[package]] -name = "contourpy" -version = "1.3.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "numpy" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/66/54/eb9bfc647b19f2009dd5c7f5ec51c4e6ca831725f1aea7a993034f483147/contourpy-1.3.2.tar.gz", hash = "sha256:b6945942715a034c671b7fc54f9588126b0b8bf23db2696e3ca8328f3ff0ab54", size = 13466130 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/12/a3/da4153ec8fe25d263aa48c1a4cbde7f49b59af86f0b6f7862788c60da737/contourpy-1.3.2-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:ba38e3f9f330af820c4b27ceb4b9c7feee5fe0493ea53a8720f4792667465934", size = 268551 }, - { url = "https://files.pythonhosted.org/packages/2f/6c/330de89ae1087eb622bfca0177d32a7ece50c3ef07b28002de4757d9d875/contourpy-1.3.2-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:dc41ba0714aa2968d1f8674ec97504a8f7e334f48eeacebcaa6256213acb0989", size = 253399 }, - { url = "https://files.pythonhosted.org/packages/c1/bd/20c6726b1b7f81a8bee5271bed5c165f0a8e1f572578a9d27e2ccb763cb2/contourpy-1.3.2-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:9be002b31c558d1ddf1b9b415b162c603405414bacd6932d031c5b5a8b757f0d", size = 312061 }, - { url = "https://files.pythonhosted.org/packages/22/fc/a9665c88f8a2473f823cf1ec601de9e5375050f1958cbb356cdf06ef1ab6/contourpy-1.3.2-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:8d2e74acbcba3bfdb6d9d8384cdc4f9260cae86ed9beee8bd5f54fee49a430b9", size = 351956 }, - { url = "https://files.pythonhosted.org/packages/25/eb/9f0a0238f305ad8fb7ef42481020d6e20cf15e46be99a1fcf939546a177e/contourpy-1.3.2-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:e259bced5549ac64410162adc973c5e2fb77f04df4a439d00b478e57a0e65512", size = 320872 }, - { url = "https://files.pythonhosted.org/packages/32/5c/1ee32d1c7956923202f00cf8d2a14a62ed7517bdc0ee1e55301227fc273c/contourpy-1.3.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ad687a04bc802cbe8b9c399c07162a3c35e227e2daccf1668eb1f278cb698631", size = 325027 }, - { url = "https://files.pythonhosted.org/packages/83/bf/9baed89785ba743ef329c2b07fd0611d12bfecbedbdd3eeecf929d8d3b52/contourpy-1.3.2-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:cdd22595308f53ef2f891040ab2b93d79192513ffccbd7fe19be7aa773a5e09f", size = 1306641 }, - { url = "https://files.pythonhosted.org/packages/d4/cc/74e5e83d1e35de2d28bd97033426b450bc4fd96e092a1f7a63dc7369b55d/contourpy-1.3.2-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:b4f54d6a2defe9f257327b0f243612dd051cc43825587520b1bf74a31e2f6ef2", size = 1374075 }, - { url = "https://files.pythonhosted.org/packages/0c/42/17f3b798fd5e033b46a16f8d9fcb39f1aba051307f5ebf441bad1ecf78f8/contourpy-1.3.2-cp310-cp310-win32.whl", hash = "sha256:f939a054192ddc596e031e50bb13b657ce318cf13d264f095ce9db7dc6ae81c0", size = 177534 }, - { url = "https://files.pythonhosted.org/packages/54/ec/5162b8582f2c994721018d0c9ece9dc6ff769d298a8ac6b6a652c307e7df/contourpy-1.3.2-cp310-cp310-win_amd64.whl", hash = "sha256:c440093bbc8fc21c637c03bafcbef95ccd963bc6e0514ad887932c18ca2a759a", size = 221188 }, - { url = "https://files.pythonhosted.org/packages/33/05/b26e3c6ecc05f349ee0013f0bb850a761016d89cec528a98193a48c34033/contourpy-1.3.2-pp310-pypy310_pp73-macosx_10_15_x86_64.whl", hash = "sha256:fd93cc7f3139b6dd7aab2f26a90dde0aa9fc264dbf70f6740d498a70b860b82c", size = 265681 }, - { url = "https://files.pythonhosted.org/packages/2b/25/ac07d6ad12affa7d1ffed11b77417d0a6308170f44ff20fa1d5aa6333f03/contourpy-1.3.2-pp310-pypy310_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:107ba8a6a7eec58bb475329e6d3b95deba9440667c4d62b9b6063942b61d7f16", size = 315101 }, - { url = "https://files.pythonhosted.org/packages/8f/4d/5bb3192bbe9d3f27e3061a6a8e7733c9120e203cb8515767d30973f71030/contourpy-1.3.2-pp310-pypy310_pp73-win_amd64.whl", hash = "sha256:ded1706ed0c1049224531b81128efbd5084598f18d8a2d9efae833edbd2b40ad", size = 220599 }, -] - -[[package]] -name = "coremltools" -version = "9.0b1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "attrs" }, - { name = "cattrs" }, - { name = "numpy" }, - { name = "packaging" }, - { name = "protobuf" }, - { name = "pyaml" }, - { name = "sympy" }, - { name = "tqdm" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/7b/4b/10f7409775150955bb7649124c592480d764491ebfacd10835130f964485/coremltools-9.0b1.tar.gz", hash = "sha256:55c0e91b0362865041cb3d8625065b9c2b9ac741b5bc8506ae3ad72012178c53", size = 1610618 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/28/47/8e4dd7b739043be8c19a89e505b751797e0d95a4827ae5a136b414890bf5/coremltools-9.0b1-cp310-none-macosx_10_15_x86_64.whl", hash = "sha256:857213355eadd3aeaa61796adff095a4c1b2d925936d866fcf9d3ac0d8c11fb2", size = 2784755 }, - { url = "https://files.pythonhosted.org/packages/81/bc/12179b5d17f4fa75d321b3ace57689a89bfb3f9728e2aed6cf49698a1c8a/coremltools-9.0b1-cp310-none-macosx_11_0_arm64.whl", hash = "sha256:beb692994682bb7e9d5739580a15a50d500bccdfabeb36d541b1f0db483fd438", size = 2759581 }, - { url = "https://files.pythonhosted.org/packages/12/a9/98c8177de8771fe79d58cc8c06be0a0c5900aa9cd3292db61350e17f35dc/coremltools-9.0b1-cp310-none-manylinux1_x86_64.whl", hash = "sha256:4cb3f710a0edcf82c110ec041b6c57cd15c76f2c5c9c015ea49afcee2975c1be", size = 2303866 }, -] - -[[package]] -name = "crcmod" -version = "1.7" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/6b/b0/e595ce2a2527e169c3bcd6c33d2473c1918e0b7f6826a043ca1245dd4e5b/crcmod-1.7.tar.gz", hash = "sha256:dc7051a0db5f2bd48665a990d3ec1cc305a466a77358ca4492826f41f283601e", size = 89670 } - -[[package]] -name = "cryptography" -version = "45.0.7" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "cffi", marker = "platform_python_implementation != 'PyPy'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/a7/35/c495bffc2056f2dadb32434f1feedd79abde2a7f8363e1974afa9c33c7e2/cryptography-45.0.7.tar.gz", hash = "sha256:4b1654dfc64ea479c242508eb8c724044f1e964a47d1d1cacc5132292d851971", size = 744980 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/0c/91/925c0ac74362172ae4516000fe877912e33b5983df735ff290c653de4913/cryptography-45.0.7-cp311-abi3-macosx_10_9_universal2.whl", hash = "sha256:3be4f21c6245930688bd9e162829480de027f8bf962ede33d4f8ba7d67a00cee", size = 7041105 }, - { url = "https://files.pythonhosted.org/packages/fc/63/43641c5acce3a6105cf8bd5baeceeb1846bb63067d26dae3e5db59f1513a/cryptography-45.0.7-cp311-abi3-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:67285f8a611b0ebc0857ced2081e30302909f571a46bfa7a3cc0ad303fe015c6", size = 4205799 }, - { url = "https://files.pythonhosted.org/packages/bc/29/c238dd9107f10bfde09a4d1c52fd38828b1aa353ced11f358b5dd2507d24/cryptography-45.0.7-cp311-abi3-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:577470e39e60a6cd7780793202e63536026d9b8641de011ed9d8174da9ca5339", size = 4430504 }, - { url = "https://files.pythonhosted.org/packages/62/62/24203e7cbcc9bd7c94739428cd30680b18ae6b18377ae66075c8e4771b1b/cryptography-45.0.7-cp311-abi3-manylinux_2_28_aarch64.whl", hash = "sha256:4bd3e5c4b9682bc112d634f2c6ccc6736ed3635fc3319ac2bb11d768cc5a00d8", size = 4209542 }, - { url = "https://files.pythonhosted.org/packages/cd/e3/e7de4771a08620eef2389b86cd87a2c50326827dea5528feb70595439ce4/cryptography-45.0.7-cp311-abi3-manylinux_2_28_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:465ccac9d70115cd4de7186e60cfe989de73f7bb23e8a7aa45af18f7412e75bf", size = 3889244 }, - { url = "https://files.pythonhosted.org/packages/96/b8/bca71059e79a0bb2f8e4ec61d9c205fbe97876318566cde3b5092529faa9/cryptography-45.0.7-cp311-abi3-manylinux_2_28_x86_64.whl", hash = "sha256:16ede8a4f7929b4b7ff3642eba2bf79aa1d71f24ab6ee443935c0d269b6bc513", size = 4461975 }, - { url = "https://files.pythonhosted.org/packages/58/67/3f5b26937fe1218c40e95ef4ff8d23c8dc05aa950d54200cc7ea5fb58d28/cryptography-45.0.7-cp311-abi3-manylinux_2_34_aarch64.whl", hash = "sha256:8978132287a9d3ad6b54fcd1e08548033cc09dc6aacacb6c004c73c3eb5d3ac3", size = 4209082 }, - { url = "https://files.pythonhosted.org/packages/0e/e4/b3e68a4ac363406a56cf7b741eeb80d05284d8c60ee1a55cdc7587e2a553/cryptography-45.0.7-cp311-abi3-manylinux_2_34_x86_64.whl", hash = "sha256:b6a0e535baec27b528cb07a119f321ac024592388c5681a5ced167ae98e9fff3", size = 4460397 }, - { url = "https://files.pythonhosted.org/packages/22/49/2c93f3cd4e3efc8cb22b02678c1fad691cff9dd71bb889e030d100acbfe0/cryptography-45.0.7-cp311-abi3-musllinux_1_2_aarch64.whl", hash = "sha256:a24ee598d10befaec178efdff6054bc4d7e883f615bfbcd08126a0f4931c83a6", size = 4337244 }, - { url = "https://files.pythonhosted.org/packages/04/19/030f400de0bccccc09aa262706d90f2ec23d56bc4eb4f4e8268d0ddf3fb8/cryptography-45.0.7-cp311-abi3-musllinux_1_2_x86_64.whl", hash = "sha256:fa26fa54c0a9384c27fcdc905a2fb7d60ac6e47d14bc2692145f2b3b1e2cfdbd", size = 4568862 }, - { url = "https://files.pythonhosted.org/packages/29/56/3034a3a353efa65116fa20eb3c990a8c9f0d3db4085429040a7eef9ada5f/cryptography-45.0.7-cp311-abi3-win32.whl", hash = "sha256:bef32a5e327bd8e5af915d3416ffefdbe65ed975b646b3805be81b23580b57b8", size = 2936578 }, - { url = "https://files.pythonhosted.org/packages/b3/61/0ab90f421c6194705a99d0fa9f6ee2045d916e4455fdbb095a9c2c9a520f/cryptography-45.0.7-cp311-abi3-win_amd64.whl", hash = "sha256:3808e6b2e5f0b46d981c24d79648e5c25c35e59902ea4391a0dcb3e667bf7443", size = 3405400 }, - { url = "https://files.pythonhosted.org/packages/63/e8/c436233ddf19c5f15b25ace33979a9dd2e7aa1a59209a0ee8554179f1cc0/cryptography-45.0.7-cp37-abi3-macosx_10_9_universal2.whl", hash = "sha256:bfb4c801f65dd61cedfc61a83732327fafbac55a47282e6f26f073ca7a41c3b2", size = 7021824 }, - { url = "https://files.pythonhosted.org/packages/bc/4c/8f57f2500d0ccd2675c5d0cc462095adf3faa8c52294ba085c036befb901/cryptography-45.0.7-cp37-abi3-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:81823935e2f8d476707e85a78a405953a03ef7b7b4f55f93f7c2d9680e5e0691", size = 4202233 }, - { url = "https://files.pythonhosted.org/packages/eb/ac/59b7790b4ccaed739fc44775ce4645c9b8ce54cbec53edf16c74fd80cb2b/cryptography-45.0.7-cp37-abi3-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:3994c809c17fc570c2af12c9b840d7cea85a9fd3e5c0e0491f4fa3c029216d59", size = 4423075 }, - { url = "https://files.pythonhosted.org/packages/b8/56/d4f07ea21434bf891faa088a6ac15d6d98093a66e75e30ad08e88aa2b9ba/cryptography-45.0.7-cp37-abi3-manylinux_2_28_aarch64.whl", hash = "sha256:dad43797959a74103cb59c5dac71409f9c27d34c8a05921341fb64ea8ccb1dd4", size = 4204517 }, - { url = "https://files.pythonhosted.org/packages/e8/ac/924a723299848b4c741c1059752c7cfe09473b6fd77d2920398fc26bfb53/cryptography-45.0.7-cp37-abi3-manylinux_2_28_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:ce7a453385e4c4693985b4a4a3533e041558851eae061a58a5405363b098fcd3", size = 3882893 }, - { url = "https://files.pythonhosted.org/packages/83/dc/4dab2ff0a871cc2d81d3ae6d780991c0192b259c35e4d83fe1de18b20c70/cryptography-45.0.7-cp37-abi3-manylinux_2_28_x86_64.whl", hash = "sha256:b04f85ac3a90c227b6e5890acb0edbaf3140938dbecf07bff618bf3638578cf1", size = 4450132 }, - { url = "https://files.pythonhosted.org/packages/12/dd/b2882b65db8fc944585d7fb00d67cf84a9cef4e77d9ba8f69082e911d0de/cryptography-45.0.7-cp37-abi3-manylinux_2_34_aarch64.whl", hash = "sha256:48c41a44ef8b8c2e80ca4527ee81daa4c527df3ecbc9423c41a420a9559d0e27", size = 4204086 }, - { url = "https://files.pythonhosted.org/packages/5d/fa/1d5745d878048699b8eb87c984d4ccc5da4f5008dfd3ad7a94040caca23a/cryptography-45.0.7-cp37-abi3-manylinux_2_34_x86_64.whl", hash = "sha256:f3df7b3d0f91b88b2106031fd995802a2e9ae13e02c36c1fc075b43f420f3a17", size = 4449383 }, - { url = "https://files.pythonhosted.org/packages/36/8b/fc61f87931bc030598e1876c45b936867bb72777eac693e905ab89832670/cryptography-45.0.7-cp37-abi3-musllinux_1_2_aarch64.whl", hash = "sha256:dd342f085542f6eb894ca00ef70236ea46070c8a13824c6bde0dfdcd36065b9b", size = 4332186 }, - { url = "https://files.pythonhosted.org/packages/0b/11/09700ddad7443ccb11d674efdbe9a832b4455dc1f16566d9bd3834922ce5/cryptography-45.0.7-cp37-abi3-musllinux_1_2_x86_64.whl", hash = "sha256:1993a1bb7e4eccfb922b6cd414f072e08ff5816702a0bdb8941c247a6b1b287c", size = 4561639 }, - { url = "https://files.pythonhosted.org/packages/71/ed/8f4c1337e9d3b94d8e50ae0b08ad0304a5709d483bfcadfcc77a23dbcb52/cryptography-45.0.7-cp37-abi3-win32.whl", hash = "sha256:18fcf70f243fe07252dcb1b268a687f2358025ce32f9f88028ca5c364b123ef5", size = 2926552 }, - { url = "https://files.pythonhosted.org/packages/bc/ff/026513ecad58dacd45d1d24ebe52b852165a26e287177de1d545325c0c25/cryptography-45.0.7-cp37-abi3-win_amd64.whl", hash = "sha256:7285a89df4900ed3bfaad5679b1e668cb4b38a8de1ccbfc84b05f34512da0a90", size = 3392742 }, - { url = "https://files.pythonhosted.org/packages/13/3e/e42f1528ca1ea82256b835191eab1be014e0f9f934b60d98b0be8a38ed70/cryptography-45.0.7-pp310-pypy310_pp73-macosx_10_9_x86_64.whl", hash = "sha256:de58755d723e86175756f463f2f0bddd45cc36fbd62601228a3f8761c9f58252", size = 3572442 }, - { url = "https://files.pythonhosted.org/packages/59/aa/e947693ab08674a2663ed2534cd8d345cf17bf6a1facf99273e8ec8986dc/cryptography-45.0.7-pp310-pypy310_pp73-manylinux_2_28_aarch64.whl", hash = "sha256:a20e442e917889d1a6b3c570c9e3fa2fdc398c20868abcea268ea33c024c4083", size = 4142233 }, - { url = "https://files.pythonhosted.org/packages/24/06/09b6f6a2fc43474a32b8fe259038eef1500ee3d3c141599b57ac6c57612c/cryptography-45.0.7-pp310-pypy310_pp73-manylinux_2_28_x86_64.whl", hash = "sha256:258e0dff86d1d891169b5af222d362468a9570e2532923088658aa866eb11130", size = 4376202 }, - { url = "https://files.pythonhosted.org/packages/00/f2/c166af87e95ce6ae6d38471a7e039d3a0549c2d55d74e059680162052824/cryptography-45.0.7-pp310-pypy310_pp73-manylinux_2_34_aarch64.whl", hash = "sha256:d97cf502abe2ab9eff8bd5e4aca274da8d06dd3ef08b759a8d6143f4ad65d4b4", size = 4141900 }, - { url = "https://files.pythonhosted.org/packages/16/b9/e96e0b6cb86eae27ea51fa8a3151535a18e66fe7c451fa90f7f89c85f541/cryptography-45.0.7-pp310-pypy310_pp73-manylinux_2_34_x86_64.whl", hash = "sha256:c987dad82e8c65ebc985f5dae5e74a3beda9d0a2a4daf8a1115f3772b59e5141", size = 4375562 }, - { url = "https://files.pythonhosted.org/packages/36/d0/36e8ee39274e9d77baf7d0dafda680cba6e52f3936b846f0d56d64fec915/cryptography-45.0.7-pp310-pypy310_pp73-win_amd64.whl", hash = "sha256:c13b1e3afd29a5b3b2656257f14669ca8fa8d7956d509926f0b130b600b50ab7", size = 3322781 }, -] - -[[package]] -name = "cycler" -version = "0.12.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/a9/95/a3dbbb5028f35eafb79008e7522a75244477d2838f38cbb722248dabc2a8/cycler-0.12.1.tar.gz", hash = "sha256:88bb128f02ba341da8ef447245a9e138fae777f6a23943da4540077d3601eb1c", size = 7615 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/e7/05/c19819d5e3d95294a6f5947fb9b9629efb316b96de511b418c53d245aae6/cycler-0.12.1-py3-none-any.whl", hash = "sha256:85cef7cff222d8644161529808465972e51340599459b8ac3ccbac5a854e0d30", size = 8321 }, -] - -[[package]] -name = "cytoolz" -version = "1.0.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "toolz" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/a7/f9/3243eed3a6545c2a33a21f74f655e3fcb5d2192613cd3db81a93369eb339/cytoolz-1.0.1.tar.gz", hash = "sha256:89cc3161b89e1bb3ed7636f74ed2e55984fd35516904fc878cae216e42b2c7d6", size = 626652 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/1e/d9/f13d66c16cff1fa1cb6c234698029877c456f35f577ef274aba3b86e7c51/cytoolz-1.0.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:cec9af61f71fc3853eb5dca3d42eb07d1f48a4599fa502cbe92adde85f74b042", size = 403515 }, - { url = "https://files.pythonhosted.org/packages/4b/2d/4cdf848a69300c7d44984f2ebbebb3b8576e5449c8dea157298f3bdc4da3/cytoolz-1.0.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:140bbd649dbda01e91add7642149a5987a7c3ccc251f2263de894b89f50b6608", size = 383936 }, - { url = "https://files.pythonhosted.org/packages/72/a4/ccfdd3f0ed9cc818f734b424261f6018fc61e3ec833bf85225a9aca0d994/cytoolz-1.0.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e90124bdc42ff58b88cdea1d24a6bc5f776414a314cc4d94f25c88badb3a16d1", size = 1934569 }, - { url = "https://files.pythonhosted.org/packages/50/fc/38d5344fa595683ad10dc819cfc1d8b9d2b3391ccf3e8cb7bab4899a01f5/cytoolz-1.0.1-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:e74801b751e28f7c5cc3ad264c123954a051f546f2fdfe089f5aa7a12ccfa6da", size = 2015129 }, - { url = "https://files.pythonhosted.org/packages/28/29/75261748dc54a20a927f33641f4e9aac674cfc6d3fbd4f332e10d0b37639/cytoolz-1.0.1-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:582dad4545ddfb5127494ef23f3fa4855f1673a35d50c66f7638e9fb49805089", size = 2000506 }, - { url = "https://files.pythonhosted.org/packages/00/ae/e4ead004cc2698281d153c4a5388638d67cdb5544d6d6cc1e5b3db2bd2a3/cytoolz-1.0.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:dd7bd0618e16efe03bd12f19c2a26a27e6e6b75d7105adb7be1cd2a53fa755d8", size = 1957537 }, - { url = "https://files.pythonhosted.org/packages/4a/ff/4f3aa07f4f47701f7f63df60ce0a5669fa09c256c3d4a33503a9414ea5cc/cytoolz-1.0.1-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:d74cca6acf1c4af58b2e4a89cc565ed61c5e201de2e434748c93e5a0f5c541a5", size = 1863331 }, - { url = "https://files.pythonhosted.org/packages/a2/29/654f57f2a9b8e9765a4ab876765f64f94530b61fc6471a07feea42ece6d4/cytoolz-1.0.1-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:823a3763828d8d457f542b2a45d75d6b4ced5e470b5c7cf2ed66a02f508ed442", size = 1849938 }, - { url = "https://files.pythonhosted.org/packages/bc/7b/11f457db6b291060a98315ab2c7198077d8bddeeebe5f7126d9dad98cc54/cytoolz-1.0.1-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:51633a14e6844c61db1d68c1ffd077cf949f5c99c60ed5f1e265b9e2966f1b52", size = 1852345 }, - { url = "https://files.pythonhosted.org/packages/6b/92/0dccc96ce0323be236d404f5084479b79b747fa0e74e43a270e95868b5f9/cytoolz-1.0.1-cp310-cp310-musllinux_1_2_ppc64le.whl", hash = "sha256:f3ec9b01c45348f1d0d712507d54c2bfd69c62fbd7c9ef555c9d8298693c2432", size = 1989877 }, - { url = "https://files.pythonhosted.org/packages/a3/c8/1c5203a81200bae51aa8f7b5fad613f695bf1afa03f16251ca23ecb2ef9f/cytoolz-1.0.1-cp310-cp310-musllinux_1_2_s390x.whl", hash = "sha256:1855022b712a9c7a5bce354517ab4727a38095f81e2d23d3eabaf1daeb6a3b3c", size = 1994492 }, - { url = "https://files.pythonhosted.org/packages/e2/8a/04bc193c4d7ced8ef6bb62cdcd0bf40b5e5eb26586ed2cfb4433ec7dfd0a/cytoolz-1.0.1-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:9930f7288c4866a1dc1cc87174f0c6ff4cad1671eb1f6306808aa6c445857d78", size = 1896077 }, - { url = "https://files.pythonhosted.org/packages/21/a5/bee63a58f51d2c74856db66e6119a014464ff8cb1c9387fa4bd2d94e49b0/cytoolz-1.0.1-cp310-cp310-win32.whl", hash = "sha256:a9baad795d72fadc3445ccd0f122abfdbdf94269157e6d6d4835636dad318804", size = 322135 }, - { url = "https://files.pythonhosted.org/packages/e8/16/7abfb1685e8b7f2838264551ee33651748994813f566ac4c3d737dfe90e5/cytoolz-1.0.1-cp310-cp310-win_amd64.whl", hash = "sha256:ad95b386a84e18e1f6136f6d343d2509d4c3aae9f5a536f3dc96808fcc56a8cf", size = 363599 }, - { url = "https://files.pythonhosted.org/packages/d9/f7/ef2a10daaec5c0f7d781d50758c6187eee484256e356ae8ef178d6c48497/cytoolz-1.0.1-pp310-pypy310_pp73-macosx_10_15_x86_64.whl", hash = "sha256:83d19d55738ad9c60763b94f3f6d3c6e4de979aeb8d76841c1401081e0e58d96", size = 345702 }, - { url = "https://files.pythonhosted.org/packages/c8/14/53c84adddedb67ff1546abb86fea04d26e24298c3ceab8436d20122ed0b9/cytoolz-1.0.1-pp310-pypy310_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:f112a71fad6ea824578e6393765ce5c054603afe1471a5c753ff6c67fd872d10", size = 385695 }, - { url = "https://files.pythonhosted.org/packages/bd/80/3ae356c5e7b8d7dc7d1adb52f6932fee85cd748ed4e1217c269d2dfd610f/cytoolz-1.0.1-pp310-pypy310_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:5a515df8f8aa6e1eaaf397761a6e4aff2eef73b5f920aedf271416d5471ae5ee", size = 406261 }, - { url = "https://files.pythonhosted.org/packages/0c/31/8e43761ffc82d90bf9cab7e0959712eedcd1e33c211397e143dd42d7af57/cytoolz-1.0.1-pp310-pypy310_pp73-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:92c398e7b7023460bea2edffe5fcd0a76029580f06c3f6938ac3d198b47156f3", size = 397207 }, - { url = "https://files.pythonhosted.org/packages/d1/b9/fe9da37090b6444c65f848a83e390f87d8cb43d6a4df46de1556ad7e5ceb/cytoolz-1.0.1-pp310-pypy310_pp73-win_amd64.whl", hash = "sha256:3237e56211e03b13df47435b2369f5df281e02b04ad80a948ebd199b7bc10a47", size = 343358 }, -] - -[[package]] -name = "datasets" -version = "3.6.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "dill" }, - { name = "filelock" }, - { name = "fsspec", extra = ["http"] }, - { name = "huggingface-hub" }, - { name = "multiprocess" }, - { name = "numpy" }, - { name = "packaging" }, - { name = "pandas" }, - { name = "pyarrow" }, - { name = "pyyaml" }, - { name = "requests" }, - { name = "tqdm" }, - { name = "xxhash" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/1a/89/d3d6fef58a488f8569c82fd293ab7cbd4250244d67f425dcae64c63800ea/datasets-3.6.0.tar.gz", hash = "sha256:1b2bf43b19776e2787e181cfd329cb0ca1a358ea014780c3581e0f276375e041", size = 569336 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/20/34/a08b0ee99715eaba118cbe19a71f7b5e2425c2718ef96007c325944a1152/datasets-3.6.0-py3-none-any.whl", hash = "sha256:25000c4a2c0873a710df127d08a202a06eab7bf42441a6bc278b499c2f72cd1b", size = 491546 }, -] - -[[package]] -name = "debugpy" -version = "1.8.14" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/bd/75/087fe07d40f490a78782ff3b0a30e3968936854105487decdb33446d4b0e/debugpy-1.8.14.tar.gz", hash = "sha256:7cd287184318416850aa8b60ac90105837bb1e59531898c07569d197d2ed5322", size = 1641444 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/fc/df/156df75a41aaebd97cee9d3870fe68f8001b6c1c4ca023e221cfce69bece/debugpy-1.8.14-cp310-cp310-macosx_14_0_x86_64.whl", hash = "sha256:93fee753097e85623cab1c0e6a68c76308cd9f13ffdf44127e6fab4fbf024339", size = 2076510 }, - { url = "https://files.pythonhosted.org/packages/69/cd/4fc391607bca0996db5f3658762106e3d2427beaef9bfd363fd370a3c054/debugpy-1.8.14-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3d937d93ae4fa51cdc94d3e865f535f185d5f9748efb41d0d49e33bf3365bd79", size = 3559614 }, - { url = "https://files.pythonhosted.org/packages/1a/42/4e6d2b9d63e002db79edfd0cb5656f1c403958915e0e73ab3e9220012eec/debugpy-1.8.14-cp310-cp310-win32.whl", hash = "sha256:c442f20577b38cc7a9aafecffe1094f78f07fb8423c3dddb384e6b8f49fd2987", size = 5208588 }, - { url = "https://files.pythonhosted.org/packages/97/b1/cc9e4e5faadc9d00df1a64a3c2d5c5f4b9df28196c39ada06361c5141f89/debugpy-1.8.14-cp310-cp310-win_amd64.whl", hash = "sha256:f117dedda6d969c5c9483e23f573b38f4e39412845c7bc487b6f2648df30fe84", size = 5241043 }, - { url = "https://files.pythonhosted.org/packages/97/1a/481f33c37ee3ac8040d3d51fc4c4e4e7e61cb08b8bc8971d6032acc2279f/debugpy-1.8.14-py2.py3-none-any.whl", hash = "sha256:5cd9a579d553b6cb9759a7908a41988ee6280b961f24f63336835d9418216a20", size = 5256230 }, -] - -[[package]] -name = "decorator" -version = "5.2.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/43/fa/6d96a0978d19e17b68d634497769987b16c8f4cd0a7a05048bec693caa6b/decorator-5.2.1.tar.gz", hash = "sha256:65f266143752f734b0a7cc83c46f4618af75b8c5911b00ccb61d0ac9b6da0360", size = 56711 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/4e/8c/f3147f5c4b73e7550fe5f9352eaa956ae838d5c51eb58e7a25b9f3e2643b/decorator-5.2.1-py3-none-any.whl", hash = "sha256:d316bb415a2d9e2d2b3abcc4084c6502fc09240e292cd76a76afc106a1c8e04a", size = 9190 }, -] - -[[package]] -name = "defusedxml" -version = "0.7.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/0f/d5/c66da9b79e5bdb124974bfe172b4daf3c984ebd9c2a06e2b8a4dc7331c72/defusedxml-0.7.1.tar.gz", hash = "sha256:1bb3032db185915b62d7c6209c5a8792be6a32ab2fedacc84e01b52c51aa3e69", size = 75520 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/07/6c/aa3f2f849e01cb6a001cd8554a88d4c77c5c1a31c95bdf1cf9301e6d9ef4/defusedxml-0.7.1-py2.py3-none-any.whl", hash = "sha256:a352e7e428770286cc899e2542b6cdaedb2b4953ff269a210103ec58f6198a61", size = 25604 }, -] - -[[package]] -name = "dill" -version = "0.3.8" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/17/4d/ac7ffa80c69ea1df30a8aa11b3578692a5118e7cd1aa157e3ef73b092d15/dill-0.3.8.tar.gz", hash = "sha256:3ebe3c479ad625c4553aca177444d89b486b1d84982eeacded644afc0cf797ca", size = 184847 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/c9/7a/cef76fd8438a42f96db64ddaa85280485a9c395e7df3db8158cfec1eee34/dill-0.3.8-py3-none-any.whl", hash = "sha256:c36ca9ffb54365bdd2f8eb3eff7d2a21237f8452b57ace88b1ac615b7e815bd7", size = 116252 }, -] - -[[package]] -name = "distance" -version = "0.1.3" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/5c/1a/883e47df323437aefa0d0a92ccfb38895d9416bd0b56262c2e46a47767b8/Distance-0.1.3.tar.gz", hash = "sha256:60807584f5b6003f5c521aa73f39f51f631de3be5cccc5a1d67166fcbf0d4551", size = 180271 } - -[[package]] -name = "docopt" -version = "0.6.2" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/a2/55/8f8cab2afd404cf578136ef2cc5dfb50baa1761b68c9da1fb1e4eed343c9/docopt-0.6.2.tar.gz", hash = "sha256:49b3a825280bd66b3aa83585ef59c4a8c82f2c8a522dbe754a8bc8d08c85c491", size = 25901 } - -[[package]] -name = "editdistance" -version = "0.8.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/d5/18/9f4f975ca87a390832b1c22478f3702fcdf739f83211e24d054b7551270d/editdistance-0.8.1.tar.gz", hash = "sha256:d1cdf80a5d5014b0c9126a69a42ce55a457b457f6986ff69ca98e4fe4d2d8fed", size = 50006 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/09/c9/302658ce7f4c537a4e85cf578d11bbf7af120a712e1d78fedc6cb8823c65/editdistance-0.8.1-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:adeb705f32b93accc74960d227875abff150ee42d676e428536361fe5f8f5388", size = 106150 }, - { url = "https://files.pythonhosted.org/packages/45/80/0b3c7d2c0e183725986fea5dd2df11f0b4b46320e9a64f6077a121ab1f64/editdistance-0.8.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:3de77951b105d0972deec7684a0b3d1a9dee69c9b5d34f6e2acc0d76cd4a1c52", size = 80551 }, - { url = "https://files.pythonhosted.org/packages/b5/14/681460965c6a4a48321b07f88de2273d097fdca0491ff55db891aacbd291/editdistance-0.8.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:5e88efb052d45e924606c305cb833a80579dca3e8e4ff01309d50ba2c1c0bbd5", size = 79142 }, - { url = "https://files.pythonhosted.org/packages/ed/0d/abdbc8e394a9461cf2ae27c16564fadaa65f52bd242dd1582ae5e7736dc3/editdistance-0.8.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0247e7a1e9c66ea75211a97e725366bff19a52aac2c838ed5f90025630e976dd", size = 396768 }, - { url = "https://files.pythonhosted.org/packages/c2/fb/2940d26ebda12efd280ae939436f17ac482930d862df9e774cb8b771ab03/editdistance-0.8.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:67d143429a49ab552411505f550a0fb4285a1d4336e096804d233ec495ac20fc", size = 401846 }, - { url = "https://files.pythonhosted.org/packages/53/cc/c63d75c7f387d4df0645682c1ab8706c2dfe5c9c0c4999723ce9a3ba0853/editdistance-0.8.1-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ca9d3be2b10e5d44a950a4bd1e84bca9ebbecd364bce0cf5693bf8224c78eaef", size = 397543 }, - { url = "https://files.pythonhosted.org/packages/8e/38/bb0f734a7571e093184606b930734b12da5b6bff2635eba9312fe4536dd9/editdistance-0.8.1-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:5c72aa1df8535f2e2b3d8773a1a7da091bc1a7e52bb396e7e48d375ba687e7b2", size = 898934 }, - { url = "https://files.pythonhosted.org/packages/1c/9f/624fc7a09918f850a057465f02e86f269e139a457f48ff8cabfb12701756/editdistance-0.8.1-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:9a606c34a2a6cc190e4fffc856b36333cdcf1f1fab5b22bd3088e585c22d6ca0", size = 959637 }, - { url = "https://files.pythonhosted.org/packages/5e/5c/7fa6cc277f91c477ee370807d51c1826891dc6dfc307544223ce7f2687de/editdistance-0.8.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:5af173d442ffac33b7c7990132f97f88818a3abf4b21c0c702a7022df37c0c5c", size = 911024 }, - { url = "https://files.pythonhosted.org/packages/ad/97/556215f71184291155aee340a6d34f0676e7238fdfd10615b6b775ce25fe/editdistance-0.8.1-cp310-cp310-win32.whl", hash = "sha256:fd64b58f5a7b59afd9d75982aaeeacd2a98498bf472fa0360c122ffe6ea4c871", size = 80834 }, - { url = "https://files.pythonhosted.org/packages/c8/d1/7ec5f5cbb95838d0eff7f980a660c81acd1363d658f2f5d4ceba38877c5a/editdistance-0.8.1-cp310-cp310-win_amd64.whl", hash = "sha256:6c7c62c3cae45ca1fa01bb2722b297b9de1e3a244ac44cfba88bdcb488fe6aee", size = 79614 }, - { url = "https://files.pythonhosted.org/packages/d4/4c/c9d02eeb47815d35f8d324b52f6704ea7beb032bcb209358cac44047d413/editdistance-0.8.1-pp310-pypy310_pp73-macosx_10_9_x86_64.whl", hash = "sha256:a4a90c6b03094c07358572027a8d0a13cca7450b1aa6caca98a5f1fa4f0b8961", size = 76455 }, - { url = "https://files.pythonhosted.org/packages/af/b0/2818fa6a24595dac069b0bfb9d05658406779a1ded8fd2b0c9066396cf99/editdistance-0.8.1-pp310-pypy310_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:510a4f9ced348a4fd89ae2e102357d4d801a771e29bb2bc2f130a1692193407f", size = 84104 }, - { url = "https://files.pythonhosted.org/packages/1f/d1/3d5e09bcf7fdb7aed705bf74047a8634bd2b8fd92177c25a2547e6dbadfb/editdistance-0.8.1-pp310-pypy310_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:4787fa7228ba6a34b430066d174320f011d605015baa7299c2c4911e6ea6bd46", size = 89058 }, - { url = "https://files.pythonhosted.org/packages/cd/88/fca5d7b1a1edf66ce1e5b6b60bff75842e6814b4f5facbdf4585d88c912d/editdistance-0.8.1-pp310-pypy310_pp73-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ee02601375073afccd6b4d811129ce1cb696d47db734784d8dbd1fddcea75447", size = 84635 }, - { url = "https://files.pythonhosted.org/packages/a9/91/0e6285bbe2358d81fd16313d30306b2d0036387348f7bc11d8c076ca3c72/editdistance-0.8.1-pp310-pypy310_pp73-win_amd64.whl", hash = "sha256:bc7ad9f9a20e6f351523de77c59249f005242e3f317b5de45d02c378d24f6531", size = 77389 }, -] - -[[package]] -name = "einops" -version = "0.8.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/e5/81/df4fbe24dff8ba3934af99044188e20a98ed441ad17a274539b74e82e126/einops-0.8.1.tar.gz", hash = "sha256:de5d960a7a761225532e0f1959e5315ebeafc0cd43394732f103ca44b9837e84", size = 54805 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/87/62/9773de14fe6c45c23649e98b83231fffd7b9892b6cf863251dc2afa73643/einops-0.8.1-py3-none-any.whl", hash = "sha256:919387eb55330f5757c6bea9165c5ff5cfe63a642682ea788a6d472576d81737", size = 64359 }, -] - -[[package]] -name = "exceptiongroup" -version = "1.3.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "typing-extensions", marker = "python_full_version == '3.10.12'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/0b/9f/a65090624ecf468cdca03533906e7c69ed7588582240cfe7cc9e770b50eb/exceptiongroup-1.3.0.tar.gz", hash = "sha256:b241f5885f560bc56a59ee63ca4c6a8bfa46ae4ad651af316d4e81817bb9fd88", size = 29749 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/36/f4/c6e662dade71f56cd2f3735141b265c3c79293c109549c1e6933b0651ffc/exceptiongroup-1.3.0-py3-none-any.whl", hash = "sha256:4d111e6e0c13d0644cad6ddaa7ed0261a0b36971f6d23e7ec9b4b9097da78a10", size = 16674 }, -] - -[[package]] -name = "executing" -version = "2.2.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/91/50/a9d80c47ff289c611ff12e63f7c5d13942c65d68125160cefd768c73e6e4/executing-2.2.0.tar.gz", hash = "sha256:5d108c028108fe2551d1a7b2e8b713341e2cb4fc0aa7dcf966fa4327a5226755", size = 978693 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/7b/8f/c4d9bafc34ad7ad5d8dc16dd1347ee0e507a52c3adb6bfa8887e1c6a26ba/executing-2.2.0-py2.py3-none-any.whl", hash = "sha256:11387150cad388d62750327a53d3339fad4888b39a6fe233c3afbb54ecffd3aa", size = 26702 }, -] - -[[package]] -name = "fastjsonschema" -version = "2.21.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/8b/50/4b769ce1ac4071a1ef6d86b1a3fb56cdc3a37615e8c5519e1af96cdac366/fastjsonschema-2.21.1.tar.gz", hash = "sha256:794d4f0a58f848961ba16af7b9c85a3e88cd360df008c59aac6fc5ae9323b5d4", size = 373939 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/90/2b/0817a2b257fe88725c25589d89aec060581aabf668707a8d03b2e9e0cb2a/fastjsonschema-2.21.1-py3-none-any.whl", hash = "sha256:c9e5b7e908310918cf494a434eeb31384dd84a98b57a30bcb1f535015b554667", size = 23924 }, -] - -[[package]] -name = "fiddle" -version = "0.3.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "absl-py" }, - { name = "graphviz" }, - { name = "libcst" }, - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/73/36/7a4fac76351619b36bbc7937abf59f7b601326dc4efc253b3c16819f782a/fiddle-0.3.0.tar.gz", hash = "sha256:5d083d3299a479868345513385a6c5546141bd92086c15d3dcbf8008a90075d3", size = 277884 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/3b/98/a38e949a91ff9e15874487fd8329ff53c25f3413c0cfc809eb6ff7eb7fa1/fiddle-0.3.0-py3-none-any.whl", hash = "sha256:f4824541c103a94a2f33f6c93eeddf6007c3a7300440087a95907f3e74362e61", size = 419830 }, -] - -[[package]] -name = "filelock" -version = "3.18.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/0a/10/c23352565a6544bdc5353e0b15fc1c563352101f30e24bf500207a54df9a/filelock-3.18.0.tar.gz", hash = "sha256:adbc88eabb99d2fec8c9c1b229b171f18afa655400173ddc653d5d01501fb9f2", size = 18075 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/4d/36/2a115987e2d8c300a974597416d9de88f2444426de9571f4b59b2cca3acc/filelock-3.18.0-py3-none-any.whl", hash = "sha256:c401f4f8377c4464e6db25fff06205fd89bdd83b65eb0488ed1b160f780e21de", size = 16215 }, -] - -[[package]] -name = "fonttools" -version = "4.58.4" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/2e/5a/1124b2c8cb3a8015faf552e92714040bcdbc145dfa29928891b02d147a18/fonttools-4.58.4.tar.gz", hash = "sha256:928a8009b9884ed3aae17724b960987575155ca23c6f0b8146e400cc9e0d44ba", size = 3525026 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/ed/86/d22c24caa574449b56e994ed1a96d23b23af85557fb62a92df96439d3f6c/fonttools-4.58.4-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:834542f13fee7625ad753b2db035edb674b07522fcbdd0ed9e9a9e2a1034467f", size = 2748349 }, - { url = "https://files.pythonhosted.org/packages/f9/b8/384aca93856def00e7de30341f1e27f439694857d82c35d74a809c705ed0/fonttools-4.58.4-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:2e6c61ce330142525296170cd65666e46121fc0d44383cbbcfa39cf8f58383df", size = 2318565 }, - { url = "https://files.pythonhosted.org/packages/1a/f2/273edfdc8d9db89ecfbbf659bd894f7e07b6d53448b19837a4bdba148d17/fonttools-4.58.4-cp310-cp310-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:e9c75f8faa29579c0fbf29b56ae6a3660c6c025f3b671803cb6a9caa7e4e3a98", size = 4838855 }, - { url = "https://files.pythonhosted.org/packages/13/fa/403703548c093c30b52ab37e109b369558afa221130e67f06bef7513f28a/fonttools-4.58.4-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:88dedcedbd5549e35b2ea3db3de02579c27e62e51af56779c021e7b33caadd0e", size = 4767637 }, - { url = "https://files.pythonhosted.org/packages/6e/a8/3380e1e0bff6defb0f81c9abf274a5b4a0f30bc8cab4fd4e346c6f923b4c/fonttools-4.58.4-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:ae80a895adab43586f4da1521d58fd4f4377cef322ee0cc205abcefa3a5effc3", size = 4819397 }, - { url = "https://files.pythonhosted.org/packages/cd/1b/99e47eb17a8ca51d808622a4658584fa8f340857438a4e9d7ac326d4a041/fonttools-4.58.4-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:0d3acc7f0d151da116e87a182aefb569cf0a3c8e0fd4c9cd0a7c1e7d3e7adb26", size = 4926641 }, - { url = "https://files.pythonhosted.org/packages/31/75/415254408f038e35b36c8525fc31feb8561f98445688dd2267c23eafd7a2/fonttools-4.58.4-cp310-cp310-win32.whl", hash = "sha256:1244f69686008e7e8d2581d9f37eef330a73fee3843f1107993eb82c9d306577", size = 2201917 }, - { url = "https://files.pythonhosted.org/packages/c5/69/f019a15ed2946317c5318e1bcc8876f8a54a313848604ad1d4cfc4c07916/fonttools-4.58.4-cp310-cp310-win_amd64.whl", hash = "sha256:2a66c0af8a01eb2b78645af60f3b787de5fe5eb1fd8348163715b80bdbfbde1f", size = 2246327 }, - { url = "https://files.pythonhosted.org/packages/0b/2f/c536b5b9bb3c071e91d536a4d11f969e911dbb6b227939f4c5b0bca090df/fonttools-4.58.4-py3-none-any.whl", hash = "sha256:a10ce13a13f26cbb9f37512a4346bb437ad7e002ff6fa966a7ce7ff5ac3528bd", size = 1114660 }, -] - -[[package]] -name = "fqdn" -version = "1.5.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/30/3e/a80a8c077fd798951169626cde3e239adeba7dab75deb3555716415bd9b0/fqdn-1.5.1.tar.gz", hash = "sha256:105ed3677e767fb5ca086a0c1f4bb66ebc3c100be518f0e0d755d9eae164d89f", size = 6015 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/cf/58/8acf1b3e91c58313ce5cb67df61001fc9dcd21be4fadb76c1a2d540e09ed/fqdn-1.5.1-py3-none-any.whl", hash = "sha256:3a179af3761e4df6eb2e026ff9e1a3033d3587bf980a0b1b2e1e5d08d7358014", size = 9121 }, -] - -[[package]] -name = "frozenlist" -version = "1.7.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/79/b1/b64018016eeb087db503b038296fd782586432b9c077fc5c7839e9cb6ef6/frozenlist-1.7.0.tar.gz", hash = "sha256:2e310d81923c2437ea8670467121cc3e9b0f76d3043cc1d2331d56c7fb7a3a8f", size = 45078 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/af/36/0da0a49409f6b47cc2d060dc8c9040b897b5902a8a4e37d9bc1deb11f680/frozenlist-1.7.0-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:cc4df77d638aa2ed703b878dd093725b72a824c3c546c076e8fdf276f78ee84a", size = 81304 }, - { url = "https://files.pythonhosted.org/packages/77/f0/77c11d13d39513b298e267b22eb6cb559c103d56f155aa9a49097221f0b6/frozenlist-1.7.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:716a9973a2cc963160394f701964fe25012600f3d311f60c790400b00e568b61", size = 47735 }, - { url = "https://files.pythonhosted.org/packages/37/12/9d07fa18971a44150593de56b2f2947c46604819976784bcf6ea0d5db43b/frozenlist-1.7.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:a0fd1bad056a3600047fb9462cff4c5322cebc59ebf5d0a3725e0ee78955001d", size = 46775 }, - { url = "https://files.pythonhosted.org/packages/70/34/f73539227e06288fcd1f8a76853e755b2b48bca6747e99e283111c18bcd4/frozenlist-1.7.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3789ebc19cb811163e70fe2bd354cea097254ce6e707ae42e56f45e31e96cb8e", size = 224644 }, - { url = "https://files.pythonhosted.org/packages/fb/68/c1d9c2f4a6e438e14613bad0f2973567586610cc22dcb1e1241da71de9d3/frozenlist-1.7.0-cp310-cp310-manylinux_2_17_armv7l.manylinux2014_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:af369aa35ee34f132fcfad5be45fbfcde0e3a5f6a1ec0712857f286b7d20cca9", size = 222125 }, - { url = "https://files.pythonhosted.org/packages/b9/d0/98e8f9a515228d708344d7c6986752be3e3192d1795f748c24bcf154ad99/frozenlist-1.7.0-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ac64b6478722eeb7a3313d494f8342ef3478dff539d17002f849101b212ef97c", size = 233455 }, - { url = "https://files.pythonhosted.org/packages/79/df/8a11bcec5600557f40338407d3e5bea80376ed1c01a6c0910fcfdc4b8993/frozenlist-1.7.0-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:f89f65d85774f1797239693cef07ad4c97fdd0639544bad9ac4b869782eb1981", size = 227339 }, - { url = "https://files.pythonhosted.org/packages/50/82/41cb97d9c9a5ff94438c63cc343eb7980dac4187eb625a51bdfdb7707314/frozenlist-1.7.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:1073557c941395fdfcfac13eb2456cb8aad89f9de27bae29fabca8e563b12615", size = 212969 }, - { url = "https://files.pythonhosted.org/packages/13/47/f9179ee5ee4f55629e4f28c660b3fdf2775c8bfde8f9c53f2de2d93f52a9/frozenlist-1.7.0-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1ed8d2fa095aae4bdc7fdd80351009a48d286635edffee66bf865e37a9125c50", size = 222862 }, - { url = "https://files.pythonhosted.org/packages/1a/52/df81e41ec6b953902c8b7e3a83bee48b195cb0e5ec2eabae5d8330c78038/frozenlist-1.7.0-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:24c34bea555fe42d9f928ba0a740c553088500377448febecaa82cc3e88aa1fa", size = 222492 }, - { url = "https://files.pythonhosted.org/packages/84/17/30d6ea87fa95a9408245a948604b82c1a4b8b3e153cea596421a2aef2754/frozenlist-1.7.0-cp310-cp310-musllinux_1_2_armv7l.whl", hash = "sha256:69cac419ac6a6baad202c85aaf467b65ac860ac2e7f2ac1686dc40dbb52f6577", size = 238250 }, - { url = "https://files.pythonhosted.org/packages/8f/00/ecbeb51669e3c3df76cf2ddd66ae3e48345ec213a55e3887d216eb4fbab3/frozenlist-1.7.0-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:960d67d0611f4c87da7e2ae2eacf7ea81a5be967861e0c63cf205215afbfac59", size = 218720 }, - { url = "https://files.pythonhosted.org/packages/1a/c0/c224ce0e0eb31cc57f67742071bb470ba8246623c1823a7530be0e76164c/frozenlist-1.7.0-cp310-cp310-musllinux_1_2_ppc64le.whl", hash = "sha256:41be2964bd4b15bf575e5daee5a5ce7ed3115320fb3c2b71fca05582ffa4dc9e", size = 232585 }, - { url = "https://files.pythonhosted.org/packages/55/3c/34cb694abf532f31f365106deebdeac9e45c19304d83cf7d51ebbb4ca4d1/frozenlist-1.7.0-cp310-cp310-musllinux_1_2_s390x.whl", hash = "sha256:46d84d49e00c9429238a7ce02dc0be8f6d7cd0cd405abd1bebdc991bf27c15bd", size = 234248 }, - { url = "https://files.pythonhosted.org/packages/98/c0/2052d8b6cecda2e70bd81299e3512fa332abb6dcd2969b9c80dfcdddbf75/frozenlist-1.7.0-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:15900082e886edb37480335d9d518cec978afc69ccbc30bd18610b7c1b22a718", size = 221621 }, - { url = "https://files.pythonhosted.org/packages/c5/bf/7dcebae315436903b1d98ffb791a09d674c88480c158aa171958a3ac07f0/frozenlist-1.7.0-cp310-cp310-win32.whl", hash = "sha256:400ddd24ab4e55014bba442d917203c73b2846391dd42ca5e38ff52bb18c3c5e", size = 39578 }, - { url = "https://files.pythonhosted.org/packages/8f/5f/f69818f017fa9a3d24d1ae39763e29b7f60a59e46d5f91b9c6b21622f4cd/frozenlist-1.7.0-cp310-cp310-win_amd64.whl", hash = "sha256:6eb93efb8101ef39d32d50bce242c84bcbddb4f7e9febfa7b524532a239b4464", size = 43830 }, - { url = "https://files.pythonhosted.org/packages/ee/45/b82e3c16be2182bff01179db177fe144d58b5dc787a7d4492c6ed8b9317f/frozenlist-1.7.0-py3-none-any.whl", hash = "sha256:9a5af342e34f7e97caf8c995864c7a396418ae2859cc6fdf1b1073020d516a7e", size = 13106 }, -] - -[[package]] -name = "fsspec" -version = "2024.12.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/ee/11/de70dee31455c546fbc88301971ec03c328f3d1138cfba14263f651e9551/fsspec-2024.12.0.tar.gz", hash = "sha256:670700c977ed2fb51e0d9f9253177ed20cbde4a3e5c0283cc5385b5870c8533f", size = 291600 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/de/86/5486b0188d08aa643e127774a99bac51ffa6cf343e3deb0583956dca5b22/fsspec-2024.12.0-py3-none-any.whl", hash = "sha256:b520aed47ad9804237ff878b504267a3b0b441e97508bd6d2d8774e3db85cee2", size = 183862 }, -] - -[package.optional-dependencies] -http = [ - { name = "aiohttp" }, -] - -[[package]] -name = "funasr" -version = "1.2.7" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "editdistance" }, - { name = "hydra-core" }, - { name = "jaconv" }, - { name = "jamo" }, - { name = "jieba" }, - { name = "kaldiio" }, - { name = "librosa" }, - { name = "modelscope" }, - { name = "oss2" }, - { name = "pytorch-wpe" }, - { name = "pyyaml" }, - { name = "requests" }, - { name = "scipy" }, - { name = "sentencepiece" }, - { name = "soundfile" }, - { name = "tensorboardx" }, - { name = "torch-complex" }, - { name = "tqdm" }, - { name = "umap-learn" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/4d/ca/0f3e0df85bfea75acd5950a3c8da67c857af700b7e781eaa47322e18e593/funasr-1.2.7.tar.gz", hash = "sha256:fe943487bab0137813b213ffab14e819d8762fe1df6f776bb2db0bf89958d5e2", size = 555042 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/f7/6f/491bc744f9d23be35d848479dd21ba6576788e4a2cffbdd410725669fe5c/funasr-1.2.7-py3-none-any.whl", hash = "sha256:b53f748e479e5bf6af172407c50eccaa6818ed91bdf8656abcd7ea6c5e3d2b0d", size = 703175 }, -] - -[[package]] -name = "future" -version = "1.0.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/a7/b2/4140c69c6a66432916b26158687e821ba631a4c9273c474343badf84d3ba/future-1.0.0.tar.gz", hash = "sha256:bd2968309307861edae1458a4f8a4f3598c03be43b97521076aebf5d94c07b05", size = 1228490 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/da/71/ae30dadffc90b9006d77af76b393cb9dfbfc9629f339fc1574a1c52e6806/future-1.0.0-py3-none-any.whl", hash = "sha256:929292d34f5872e70396626ef385ec22355a1fae8ad29e1a734c3e43f9fbc216", size = 491326 }, -] - -[[package]] -name = "g2p-en" -version = "2.1.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "distance" }, - { name = "inflect" }, - { name = "nltk" }, - { name = "numpy" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/5f/22/2c7acbe6164ed6cfd4301e9ad2dbde69c68d22268a0f9b5b0ee6052ed3ab/g2p_en-2.1.0.tar.gz", hash = "sha256:32ecb119827a3b10ea8c1197276f4ea4f44070ae56cbbd01f0f261875f556a58", size = 3116166 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/d7/d9/b77dc634a7a0c0c97716ba97dd0a28cbfa6267c96f359c4f27ed71cbd284/g2p_en-2.1.0-py3-none-any.whl", hash = "sha256:2a7aabf1fc7f270fcc3349881407988c9245173c2413debbe5432f4a4f31319f", size = 3117464 }, -] - -[[package]] -name = "gitdb" -version = "4.0.12" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "smmap" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/72/94/63b0fc47eb32792c7ba1fe1b694daec9a63620db1e313033d18140c2320a/gitdb-4.0.12.tar.gz", hash = "sha256:5ef71f855d191a3326fcfbc0d5da835f26b13fbcba60c32c21091c349ffdb571", size = 394684 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/a0/61/5c78b91c3143ed5c14207f463aecfc8f9dbb5092fb2869baf37c273b2705/gitdb-4.0.12-py3-none-any.whl", hash = "sha256:67073e15955400952c6565cc3e707c554a4eea2e428946f7a4c162fab9bd9bcf", size = 62794 }, -] - -[[package]] -name = "gitpython" -version = "3.1.44" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "gitdb" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/c0/89/37df0b71473153574a5cdef8f242de422a0f5d26d7a9e231e6f169b4ad14/gitpython-3.1.44.tar.gz", hash = "sha256:c87e30b26253bf5418b01b0660f818967f3c503193838337fe5e573331249269", size = 214196 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/1d/9a/4114a9057db2f1462d5c8f8390ab7383925fe1ac012eaa42402ad65c2963/GitPython-3.1.44-py3-none-any.whl", hash = "sha256:9e0e10cda9bed1ee64bc9a6de50e7e38a9c9943241cd7f585f6df3ed28011110", size = 207599 }, -] - -[[package]] -name = "graphviz" -version = "0.21" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/f8/b3/3ac91e9be6b761a4b30d66ff165e54439dcd48b83f4e20d644867215f6ca/graphviz-0.21.tar.gz", hash = "sha256:20743e7183be82aaaa8ad6c93f8893c923bd6658a04c32ee115edb3c8a835f78", size = 200434 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/91/4c/e0ce1ef95d4000ebc1c11801f9b944fa5910ecc15b5e351865763d8657f8/graphviz-0.21-py3-none-any.whl", hash = "sha256:54f33de9f4f911d7e84e4191749cac8cc5653f815b06738c54db9a15ab8b1e42", size = 47300 }, -] - -[[package]] -name = "greenlet" -version = "3.2.4" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/03/b8/704d753a5a45507a7aab61f18db9509302ed3d0a27ac7e0359ec2905b1a6/greenlet-3.2.4.tar.gz", hash = "sha256:0dca0d95ff849f9a364385f36ab49f50065d76964944638be9691e1832e9f86d", size = 188260 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/7d/ed/6bfa4109fcb23a58819600392564fea69cdc6551ffd5e69ccf1d52a40cbc/greenlet-3.2.4-cp310-cp310-macosx_11_0_universal2.whl", hash = "sha256:8c68325b0d0acf8d91dde4e6f930967dd52a5302cd4062932a6b2e7c2969f47c", size = 271061 }, - { url = "https://files.pythonhosted.org/packages/2a/fc/102ec1a2fc015b3a7652abab7acf3541d58c04d3d17a8d3d6a44adae1eb1/greenlet-3.2.4-cp310-cp310-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:94385f101946790ae13da500603491f04a76b6e4c059dab271b3ce2e283b2590", size = 629475 }, - { url = "https://files.pythonhosted.org/packages/c5/26/80383131d55a4ac0fb08d71660fd77e7660b9db6bdb4e8884f46d9f2cc04/greenlet-3.2.4-cp310-cp310-manylinux2014_ppc64le.manylinux_2_17_ppc64le.whl", hash = "sha256:f10fd42b5ee276335863712fa3da6608e93f70629c631bf77145021600abc23c", size = 640802 }, - { url = "https://files.pythonhosted.org/packages/9f/7c/e7833dbcd8f376f3326bd728c845d31dcde4c84268d3921afcae77d90d08/greenlet-3.2.4-cp310-cp310-manylinux2014_s390x.manylinux_2_17_s390x.whl", hash = "sha256:c8c9e331e58180d0d83c5b7999255721b725913ff6bc6cf39fa2a45841a4fd4b", size = 636703 }, - { url = "https://files.pythonhosted.org/packages/e9/49/547b93b7c0428ede7b3f309bc965986874759f7d89e4e04aeddbc9699acb/greenlet-3.2.4-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:58b97143c9cc7b86fc458f215bd0932f1757ce649e05b640fea2e79b54cedb31", size = 635417 }, - { url = "https://files.pythonhosted.org/packages/7f/91/ae2eb6b7979e2f9b035a9f612cf70f1bf54aad4e1d125129bef1eae96f19/greenlet-3.2.4-cp310-cp310-manylinux_2_24_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:c2ca18a03a8cfb5b25bc1cbe20f3d9a4c80d8c3b13ba3df49ac3961af0b1018d", size = 584358 }, - { url = "https://files.pythonhosted.org/packages/f7/85/433de0c9c0252b22b16d413c9407e6cb3b41df7389afc366ca204dbc1393/greenlet-3.2.4-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:9fe0a28a7b952a21e2c062cd5756d34354117796c6d9215a87f55e38d15402c5", size = 1113550 }, - { url = "https://files.pythonhosted.org/packages/a1/8d/88f3ebd2bc96bf7747093696f4335a0a8a4c5acfcf1b757717c0d2474ba3/greenlet-3.2.4-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:8854167e06950ca75b898b104b63cc646573aa5fef1353d4508ecdd1ee76254f", size = 1137126 }, - { url = "https://files.pythonhosted.org/packages/d6/6f/b60b0291d9623c496638c582297ead61f43c4b72eef5e9c926ef4565ec13/greenlet-3.2.4-cp310-cp310-win_amd64.whl", hash = "sha256:73f49b5368b5359d04e18d15828eecc1806033db5233397748f4ca813ff1056c", size = 298654 }, -] - -[[package]] -name = "grpcio" -version = "1.73.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/79/e8/b43b851537da2e2f03fa8be1aef207e5cbfb1a2e014fbb6b40d24c177cd3/grpcio-1.73.1.tar.gz", hash = "sha256:7fce2cd1c0c1116cf3850564ebfc3264fba75d3c74a7414373f1238ea365ef87", size = 12730355 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/8f/51/a5748ab2773d893d099b92653039672f7e26dd35741020972b84d604066f/grpcio-1.73.1-cp310-cp310-linux_armv7l.whl", hash = "sha256:2d70f4ddd0a823436c2624640570ed6097e40935c9194482475fe8e3d9754d55", size = 5365087 }, - { url = "https://files.pythonhosted.org/packages/ae/12/c5ee1a5dfe93dbc2eaa42a219e2bf887250b52e2e2ee5c036c4695f2769c/grpcio-1.73.1-cp310-cp310-macosx_11_0_universal2.whl", hash = "sha256:3841a8a5a66830261ab6a3c2a3dc539ed84e4ab019165f77b3eeb9f0ba621f26", size = 10608921 }, - { url = "https://files.pythonhosted.org/packages/c4/6d/b0c6a8120f02b7d15c5accda6bfc43bc92be70ada3af3ba6d8e077c00374/grpcio-1.73.1-cp310-cp310-manylinux_2_17_aarch64.whl", hash = "sha256:628c30f8e77e0258ab788750ec92059fc3d6628590fb4b7cea8c102503623ed7", size = 5803221 }, - { url = "https://files.pythonhosted.org/packages/a6/7a/3c886d9f1c1e416ae81f7f9c7d1995ae72cd64712d29dab74a6bafacb2d2/grpcio-1.73.1-cp310-cp310-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:67a0468256c9db6d5ecb1fde4bf409d016f42cef649323f0a08a72f352d1358b", size = 6444603 }, - { url = "https://files.pythonhosted.org/packages/42/07/f143a2ff534982c9caa1febcad1c1073cdec732f6ac7545d85555a900a7e/grpcio-1.73.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:68b84d65bbdebd5926eb5c53b0b9ec3b3f83408a30e4c20c373c5337b4219ec5", size = 6040969 }, - { url = "https://files.pythonhosted.org/packages/fb/0f/523131b7c9196d0718e7b2dac0310eb307b4117bdbfef62382e760f7e8bb/grpcio-1.73.1-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:c54796ca22b8349cc594d18b01099e39f2b7ffb586ad83217655781a350ce4da", size = 6132201 }, - { url = "https://files.pythonhosted.org/packages/ad/18/010a055410eef1d3a7a1e477ec9d93b091ac664ad93e9c5f56d6cc04bdee/grpcio-1.73.1-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:75fc8e543962ece2f7ecd32ada2d44c0c8570ae73ec92869f9af8b944863116d", size = 6774718 }, - { url = "https://files.pythonhosted.org/packages/16/11/452bfc1ab39d8ee748837ab8ee56beeae0290861052948785c2c445fb44b/grpcio-1.73.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:6a6037891cd2b1dd1406b388660522e1565ed340b1fea2955b0234bdd941a862", size = 6304362 }, - { url = "https://files.pythonhosted.org/packages/1e/1c/c75ceee626465721e5cb040cf4b271eff817aa97388948660884cb7adffa/grpcio-1.73.1-cp310-cp310-win32.whl", hash = "sha256:cce7265b9617168c2d08ae570fcc2af4eaf72e84f8c710ca657cc546115263af", size = 3679036 }, - { url = "https://files.pythonhosted.org/packages/62/2e/42cb31b6cbd671a7b3dbd97ef33f59088cf60e3cf2141368282e26fafe79/grpcio-1.73.1-cp310-cp310-win_amd64.whl", hash = "sha256:6a2b372e65fad38842050943f42ce8fee00c6f2e8ea4f7754ba7478d26a356ee", size = 4340208 }, -] - -[[package]] -name = "h11" -version = "0.16.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/01/ee/02a2c011bdab74c6fb3c75474d40b3052059d95df7e73351460c8588d963/h11-0.16.0.tar.gz", hash = "sha256:4e35b956cf45792e4caa5885e69fba00bdbc6ffafbfa020300e549b208ee5ff1", size = 101250 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/04/4b/29cac41a4d98d144bf5f6d33995617b185d14b22401f75ca86f384e87ff1/h11-0.16.0-py3-none-any.whl", hash = "sha256:63cf8bbe7522de3bf65932fda1d9c2772064ffb3dae62d55932da54b31cb6c86", size = 37515 }, -] - -[[package]] -name = "hf-xet" -version = "1.1.5" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/ed/d4/7685999e85945ed0d7f0762b686ae7015035390de1161dcea9d5276c134c/hf_xet-1.1.5.tar.gz", hash = "sha256:69ebbcfd9ec44fdc2af73441619eeb06b94ee34511bbcf57cd423820090f5694", size = 495969 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/00/89/a1119eebe2836cb25758e7661d6410d3eae982e2b5e974bcc4d250be9012/hf_xet-1.1.5-cp37-abi3-macosx_10_12_x86_64.whl", hash = "sha256:f52c2fa3635b8c37c7764d8796dfa72706cc4eded19d638331161e82b0792e23", size = 2687929 }, - { url = "https://files.pythonhosted.org/packages/de/5f/2c78e28f309396e71ec8e4e9304a6483dcbc36172b5cea8f291994163425/hf_xet-1.1.5-cp37-abi3-macosx_11_0_arm64.whl", hash = "sha256:9fa6e3ee5d61912c4a113e0708eaaef987047616465ac7aa30f7121a48fc1af8", size = 2556338 }, - { url = "https://files.pythonhosted.org/packages/6d/2f/6cad7b5fe86b7652579346cb7f85156c11761df26435651cbba89376cd2c/hf_xet-1.1.5-cp37-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:fc874b5c843e642f45fd85cda1ce599e123308ad2901ead23d3510a47ff506d1", size = 3102894 }, - { url = "https://files.pythonhosted.org/packages/d0/54/0fcf2b619720a26fbb6cc941e89f2472a522cd963a776c089b189559447f/hf_xet-1.1.5-cp37-abi3-manylinux_2_28_aarch64.whl", hash = "sha256:dbba1660e5d810bd0ea77c511a99e9242d920790d0e63c0e4673ed36c4022d18", size = 3002134 }, - { url = "https://files.pythonhosted.org/packages/f3/92/1d351ac6cef7c4ba8c85744d37ffbfac2d53d0a6c04d2cabeba614640a78/hf_xet-1.1.5-cp37-abi3-musllinux_1_2_aarch64.whl", hash = "sha256:ab34c4c3104133c495785d5d8bba3b1efc99de52c02e759cf711a91fd39d3a14", size = 3171009 }, - { url = "https://files.pythonhosted.org/packages/c9/65/4b2ddb0e3e983f2508528eb4501288ae2f84963586fbdfae596836d5e57a/hf_xet-1.1.5-cp37-abi3-musllinux_1_2_x86_64.whl", hash = "sha256:83088ecea236d5113de478acb2339f92c95b4fb0462acaa30621fac02f5a534a", size = 3279245 }, - { url = "https://files.pythonhosted.org/packages/f0/55/ef77a85ee443ae05a9e9cba1c9f0dd9241eb42da2aeba1dc50f51154c81a/hf_xet-1.1.5-cp37-abi3-win_amd64.whl", hash = "sha256:73e167d9807d166596b4b2f0b585c6d5bd84a26dea32843665a8b58f6edba245", size = 2738931 }, -] - -[[package]] -name = "httpcore" -version = "1.0.9" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "certifi" }, - { name = "h11" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/06/94/82699a10bca87a5556c9c59b5963f2d039dbd239f25bc2a63907a05a14cb/httpcore-1.0.9.tar.gz", hash = "sha256:6e34463af53fd2ab5d807f399a9b45ea31c3dfa2276f15a2c3f00afff6e176e8", size = 85484 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/7e/f5/f66802a942d491edb555dd61e3a9961140fd64c90bce1eafd741609d334d/httpcore-1.0.9-py3-none-any.whl", hash = "sha256:2d400746a40668fc9dec9810239072b40b4484b640a8c38fd654a024c7a1bf55", size = 78784 }, -] - -[[package]] -name = "httpx" -version = "0.28.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "anyio" }, - { name = "certifi" }, - { name = "httpcore" }, - { name = "idna" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/b1/df/48c586a5fe32a0f01324ee087459e112ebb7224f646c0b5023f5e79e9956/httpx-0.28.1.tar.gz", hash = "sha256:75e98c5f16b0f35b567856f597f06ff2270a374470a5c2392242528e3e3e42fc", size = 141406 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/2a/39/e50c7c3a983047577ee07d2a9e53faf5a69493943ec3f6a384bdc792deb2/httpx-0.28.1-py3-none-any.whl", hash = "sha256:d909fcccc110f8c7faf814ca82a9a4d816bc5a6dbfea25d6591d6985b8ba59ad", size = 73517 }, -] - -[[package]] -name = "huggingface-hub" -version = "0.33.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "filelock" }, - { name = "fsspec" }, - { name = "hf-xet", marker = "platform_machine == 'aarch64' or platform_machine == 'amd64' or platform_machine == 'arm64' or platform_machine == 'x86_64'" }, - { name = "packaging" }, - { name = "pyyaml" }, - { name = "requests" }, - { name = "tqdm" }, - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/a4/01/bfe0534a63ce7a2285e90dbb33e8a5b815ff096d8f7743b135c256916589/huggingface_hub-0.33.1.tar.gz", hash = "sha256:589b634f979da3ea4b8bdb3d79f97f547840dc83715918daf0b64209c0844c7b", size = 426728 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/d0/fb/5307bd3612eb0f0e62c3a916ae531d3a31e58fb5c82b58e3ebf7fd6f47a1/huggingface_hub-0.33.1-py3-none-any.whl", hash = "sha256:ec8d7444628210c0ba27e968e3c4c973032d44dcea59ca0d78ef3f612196f095", size = 515377 }, -] - -[[package]] -name = "hydra-core" -version = "1.3.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "antlr4-python3-runtime" }, - { name = "omegaconf" }, - { name = "packaging" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/6d/8e/07e42bc434a847154083b315779b0a81d567154504624e181caf2c71cd98/hydra-core-1.3.2.tar.gz", hash = "sha256:8a878ed67216997c3e9d88a8e72e7b4767e81af37afb4ea3334b269a4390a824", size = 3263494 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/c6/50/e0edd38dcd63fb26a8547f13d28f7a008bc4a3fd4eb4ff030673f22ad41a/hydra_core-1.3.2-py3-none-any.whl", hash = "sha256:fa0238a9e31df3373b35b0bfb672c34cc92718d21f81311d8996a16de1141d8b", size = 154547 }, -] - -[[package]] -name = "hyperpyyaml" -version = "1.2.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "pyyaml" }, - { name = "ruamel-yaml" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/52/e3/3ac46d9a662b037f699a6948b39c8d03bfcff0b592335d5953ba0c55d453/HyperPyYAML-1.2.2.tar.gz", hash = "sha256:bdb734210d18770a262f500fe5755c7a44a5d3b91521b06e24f7a00a36ee0f87", size = 17085 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/33/c9/751b6401887f4b50f9307cc1e53d287b3dc77c375c126aeb6335aff73ccb/HyperPyYAML-1.2.2-py3-none-any.whl", hash = "sha256:3c5864bdc8864b2f0fbd7bc495e7e8fdf2dfd5dd80116f72da27ca96a128bdeb", size = 16118 }, -] - -[[package]] -name = "idna" -version = "3.10" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/f1/70/7703c29685631f5a7590aa73f1f1d3fa9a380e654b86af429e0934a32f7d/idna-3.10.tar.gz", hash = "sha256:12f65c9b470abda6dc35cf8e63cc574b1c52b11df2c86030af0ac09b01b13ea9", size = 190490 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/76/c6/c88e154df9c4e1a2a66ccf0005a88dfb2650c1dffb6f5ce603dfbd452ce3/idna-3.10-py3-none-any.whl", hash = "sha256:946d195a0d259cbba61165e88e65941f16e9b36ea6ddb97f00452bae8b1287d3", size = 70442 }, -] - -[[package]] -name = "inflect" -version = "7.5.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "more-itertools" }, - { name = "typeguard" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/78/c6/943357d44a21fd995723d07ccaddd78023eace03c1846049a2645d4324a3/inflect-7.5.0.tar.gz", hash = "sha256:faf19801c3742ed5a05a8ce388e0d8fe1a07f8d095c82201eb904f5d27ad571f", size = 73751 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/8a/eb/427ed2b20a38a4ee29f24dbe4ae2dafab198674fe9a85e3d6adf9e5f5f41/inflect-7.5.0-py3-none-any.whl", hash = "sha256:2aea70e5e70c35d8350b8097396ec155ffd68def678c7ff97f51aa69c1d92344", size = 35197 }, -] - -[[package]] -name = "intervaltree" -version = "3.1.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "sortedcontainers" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/50/fb/396d568039d21344639db96d940d40eb62befe704ef849b27949ded5c3bb/intervaltree-3.1.0.tar.gz", hash = "sha256:902b1b88936918f9b2a19e0e5eb7ccb430ae45cde4f39ea4b36932920d33952d", size = 32861 } - -[[package]] -name = "ipykernel" -version = "6.29.5" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "appnope", marker = "platform_system == 'Darwin'" }, - { name = "comm" }, - { name = "debugpy" }, - { name = "ipython" }, - { name = "jupyter-client" }, - { name = "jupyter-core" }, - { name = "matplotlib-inline" }, - { name = "nest-asyncio" }, - { name = "packaging" }, - { name = "psutil" }, - { name = "pyzmq" }, - { name = "tornado" }, - { name = "traitlets" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/e9/5c/67594cb0c7055dc50814b21731c22a601101ea3b1b50a9a1b090e11f5d0f/ipykernel-6.29.5.tar.gz", hash = "sha256:f093a22c4a40f8828f8e330a9c297cb93dcab13bd9678ded6de8e5cf81c56215", size = 163367 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/94/5c/368ae6c01c7628438358e6d337c19b05425727fbb221d2a3c4303c372f42/ipykernel-6.29.5-py3-none-any.whl", hash = "sha256:afdb66ba5aa354b09b91379bac28ae4afebbb30e8b39510c9690afb7a10421b5", size = 117173 }, -] - -[[package]] -name = "ipython" -version = "8.37.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "colorama", marker = "sys_platform == 'win32'" }, - { name = "decorator" }, - { name = "exceptiongroup", marker = "python_full_version == '3.10.12'" }, - { name = "jedi" }, - { name = "matplotlib-inline" }, - { name = "pexpect", marker = "sys_platform != 'emscripten' and sys_platform != 'win32'" }, - { name = "prompt-toolkit" }, - { name = "pygments" }, - { name = "stack-data" }, - { name = "traitlets" }, - { name = "typing-extensions", marker = "python_full_version == '3.10.12'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/85/31/10ac88f3357fc276dc8a64e8880c82e80e7459326ae1d0a211b40abf6665/ipython-8.37.0.tar.gz", hash = "sha256:ca815841e1a41a1e6b73a0b08f3038af9b2252564d01fc405356d34033012216", size = 5606088 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/91/d0/274fbf7b0b12643cbbc001ce13e6a5b1607ac4929d1b11c72460152c9fc3/ipython-8.37.0-py3-none-any.whl", hash = "sha256:ed87326596b878932dbcb171e3e698845434d8c61b8d8cd474bf663041a9dcf2", size = 831864 }, -] - -[[package]] -name = "ipywidgets" -version = "8.1.7" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "comm" }, - { name = "ipython" }, - { name = "jupyterlab-widgets" }, - { name = "traitlets" }, - { name = "widgetsnbextension" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/3e/48/d3dbac45c2814cb73812f98dd6b38bbcc957a4e7bb31d6ea9c03bf94ed87/ipywidgets-8.1.7.tar.gz", hash = "sha256:15f1ac050b9ccbefd45dccfbb2ef6bed0029d8278682d569d71b8dd96bee0376", size = 116721 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/58/6a/9166369a2f092bd286d24e6307de555d63616e8ddb373ebad2b5635ca4cd/ipywidgets-8.1.7-py3-none-any.whl", hash = "sha256:764f2602d25471c213919b8a1997df04bef869251db4ca8efba1b76b1bd9f7bb", size = 139806 }, -] - -[[package]] -name = "isoduration" -version = "20.11.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "arrow" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/7c/1a/3c8edc664e06e6bd06cce40c6b22da5f1429aa4224d0c590f3be21c91ead/isoduration-20.11.0.tar.gz", hash = "sha256:ac2f9015137935279eac671f94f89eb00584f940f5dc49462a0c4ee692ba1bd9", size = 11649 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/7b/55/e5326141505c5d5e34c5e0935d2908a74e4561eca44108fbfb9c13d2911a/isoduration-20.11.0-py3-none-any.whl", hash = "sha256:b2904c2a4228c3d44f409c8ae8e2370eb21a26f7ac2ec5446df141dde3452042", size = 11321 }, -] - -[[package]] -name = "jaconv" -version = "0.4.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/d2/e1/670cefc7f00b0e1890e114a37a98ea425f7e06131342aeb9636856ac663c/jaconv-0.4.0.tar.gz", hash = "sha256:32da74b247f276e09a52d6b35c153df2387965cb85a6f034cc8af21d446f8161", size = 17402 } - -[[package]] -name = "jamo" -version = "0.4.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/b1/a2/bda770579809726e929ca6356743f9f50f64a2cbaee578fa9d4824afb00e/jamo-0.4.1.tar.gz", hash = "sha256:ea65cf9d35338d0e0af48d75ff426d8a369b0ebde6f07051c3ac37256f56d025", size = 7386 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/ac/cc/49812faae67f9a24be6ddaf58a2cf7e8c3cbfcf5b762d9414f7103d2ea2c/jamo-0.4.1-py3-none-any.whl", hash = "sha256:d4b94fd23324c606ed2fbc4037c603e2c3a7ae9390c05d3473aea1ccb6b1c3fb", size = 9543 }, -] - -[[package]] -name = "jedi" -version = "0.19.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "parso" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/72/3a/79a912fbd4d8dd6fbb02bf69afd3bb72cf0c729bb3063c6f4498603db17a/jedi-0.19.2.tar.gz", hash = "sha256:4770dc3de41bde3966b02eb84fbcf557fb33cce26ad23da12c742fb50ecb11f0", size = 1231287 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/c0/5a/9cac0c82afec3d09ccd97c8b6502d48f165f9124db81b4bcb90b4af974ee/jedi-0.19.2-py2.py3-none-any.whl", hash = "sha256:a8ef22bde8490f57fe5c7681a3c83cb58874daf72b4784de3cce5b6ef6edb5b9", size = 1572278 }, -] - -[[package]] -name = "jieba" -version = "0.42.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/c6/cb/18eeb235f833b726522d7ebed54f2278ce28ba9438e3135ab0278d9792a2/jieba-0.42.1.tar.gz", hash = "sha256:055ca12f62674fafed09427f176506079bc135638a14e23e25be909131928db2", size = 19214172 } - -[[package]] -name = "jinja2" -version = "3.1.6" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "markupsafe" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/df/bf/f7da0350254c0ed7c72f3e33cef02e048281fec7ecec5f032d4aac52226b/jinja2-3.1.6.tar.gz", hash = "sha256:0137fb05990d35f1275a587e9aee6d56da821fc83491a0fb838183be43f66d6d", size = 245115 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/62/a1/3d680cbfd5f4b8f15abc1d571870c5fc3e594bb582bc3b64ea099db13e56/jinja2-3.1.6-py3-none-any.whl", hash = "sha256:85ece4451f492d0c13c5dd7c13a64681a86afae63a5f347908daf103ce6d2f67", size = 134899 }, -] - -[[package]] -name = "jiwer" -version = "4.0.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "click" }, - { name = "rapidfuzz" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/12/1e/963dfc249d5bbe88079b57a22556e981ddd9208e4b6116e48bdbfc01f26b/jiwer-4.0.0.tar.gz", hash = "sha256:ae9c051469102a61ef0927100baeeb4546f78d180c9b0948281d08eaf44c191e", size = 28074 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/69/c9/172c525330c739a068c01050759a6f855ce16212db10a0359e690a03ac48/jiwer-4.0.0-py3-none-any.whl", hash = "sha256:7efaf0bd336b095d99ddef9dd67e1ee829d75d58aa2a81d9639870b01d6d95ea", size = 23034 }, -] - -[[package]] -name = "jmespath" -version = "0.10.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/3c/56/3f325b1eef9791759784aa5046a8f6a1aff8f7c898a2e34506771d3b99d8/jmespath-0.10.0.tar.gz", hash = "sha256:b85d0567b8666149a93172712e68920734333c0ce7e89b78b3e987f71e5ed4f9", size = 21607 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/07/cb/5f001272b6faeb23c1c9e0acc04d48eaaf5c862c17709d20e3469c6e0139/jmespath-0.10.0-py2.py3-none-any.whl", hash = "sha256:cdf6525904cc597730141d61b36f2e4b8ecc257c420fa2f4549bac2c2d0cb72f", size = 24489 }, -] - -[[package]] -name = "joblib" -version = "1.5.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/dc/fe/0f5a938c54105553436dbff7a61dc4fed4b1b2c98852f8833beaf4d5968f/joblib-1.5.1.tar.gz", hash = "sha256:f4f86e351f39fe3d0d32a9f2c3d8af1ee4cec285aafcb27003dda5205576b444", size = 330475 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/7d/4f/1195bbac8e0c2acc5f740661631d8d750dc38d4a32b23ee5df3cde6f4e0d/joblib-1.5.1-py3-none-any.whl", hash = "sha256:4719a31f054c7d766948dcd83e9613686b27114f190f717cec7eaa2084f8a74a", size = 307746 }, -] - -[[package]] -name = "json5" -version = "0.12.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/12/be/c6c745ec4c4539b25a278b70e29793f10382947df0d9efba2fa09120895d/json5-0.12.0.tar.gz", hash = "sha256:0b4b6ff56801a1c7dc817b0241bca4ce474a0e6a163bfef3fc594d3fd263ff3a", size = 51907 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/41/9f/3500910d5a98549e3098807493851eeef2b89cdd3032227558a104dfe926/json5-0.12.0-py3-none-any.whl", hash = "sha256:6d37aa6c08b0609f16e1ec5ff94697e2cbbfbad5ac112afa05794da9ab7810db", size = 36079 }, -] - -[[package]] -name = "jsonpointer" -version = "3.0.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/6a/0a/eebeb1fa92507ea94016a2a790b93c2ae41a7e18778f85471dc54475ed25/jsonpointer-3.0.0.tar.gz", hash = "sha256:2b2d729f2091522d61c3b31f82e11870f60b68f43fbc705cb76bf4b832af59ef", size = 9114 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/71/92/5e77f98553e9e75130c78900d000368476aed74276eb8ae8796f65f00918/jsonpointer-3.0.0-py2.py3-none-any.whl", hash = "sha256:13e088adc14fca8b6aa8177c044e12701e6ad4b28ff10e65f2267a90109c9942", size = 7595 }, -] - -[[package]] -name = "jsonschema" -version = "4.24.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "attrs" }, - { name = "jsonschema-specifications" }, - { name = "referencing" }, - { name = "rpds-py" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/bf/d3/1cf5326b923a53515d8f3a2cd442e6d7e94fcc444716e879ea70a0ce3177/jsonschema-4.24.0.tar.gz", hash = "sha256:0b4e8069eb12aedfa881333004bccaec24ecef5a8a6a4b6df142b2cc9599d196", size = 353480 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/a2/3d/023389198f69c722d039351050738d6755376c8fd343e91dc493ea485905/jsonschema-4.24.0-py3-none-any.whl", hash = "sha256:a462455f19f5faf404a7902952b6f0e3ce868f3ee09a359b05eca6673bd8412d", size = 88709 }, -] - -[package.optional-dependencies] -format-nongpl = [ - { name = "fqdn" }, - { name = "idna" }, - { name = "isoduration" }, - { name = "jsonpointer" }, - { name = "rfc3339-validator" }, - { name = "rfc3986-validator" }, - { name = "uri-template" }, - { name = "webcolors" }, -] - -[[package]] -name = "jsonschema-specifications" -version = "2025.4.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "referencing" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/bf/ce/46fbd9c8119cfc3581ee5643ea49464d168028cfb5caff5fc0596d0cf914/jsonschema_specifications-2025.4.1.tar.gz", hash = "sha256:630159c9f4dbea161a6a2205c3011cc4f18ff381b189fff48bb39b9bf26ae608", size = 15513 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/01/0e/b27cdbaccf30b890c40ed1da9fd4a3593a5cf94dae54fb34f8a4b74fcd3f/jsonschema_specifications-2025.4.1-py3-none-any.whl", hash = "sha256:4653bffbd6584f7de83a67e0d620ef16900b390ddc7939d56684d6c81e33f1af", size = 18437 }, -] - -[[package]] -name = "julius" -version = "0.2.7" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "torch" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/a1/19/c9e1596b5572c786b93428d0904280e964c930fae7e6c9368ed9e1b63922/julius-0.2.7.tar.gz", hash = "sha256:3c0f5f5306d7d6016fcc95196b274cae6f07e2c9596eed314e4e7641554fbb08", size = 59640 } - -[[package]] -name = "jupyter" -version = "1.1.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "ipykernel" }, - { name = "ipywidgets" }, - { name = "jupyter-console" }, - { name = "jupyterlab" }, - { name = "nbconvert" }, - { name = "notebook" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/58/f3/af28ea964ab8bc1e472dba2e82627d36d470c51f5cd38c37502eeffaa25e/jupyter-1.1.1.tar.gz", hash = "sha256:d55467bceabdea49d7e3624af7e33d59c37fff53ed3a350e1ac957bed731de7a", size = 5714959 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/38/64/285f20a31679bf547b75602702f7800e74dbabae36ef324f716c02804753/jupyter-1.1.1-py2.py3-none-any.whl", hash = "sha256:7a59533c22af65439b24bbe60373a4e95af8f16ac65a6c00820ad378e3f7cc83", size = 2657 }, -] - -[[package]] -name = "jupyter-client" -version = "8.6.3" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "jupyter-core" }, - { name = "python-dateutil" }, - { name = "pyzmq" }, - { name = "tornado" }, - { name = "traitlets" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/71/22/bf9f12fdaeae18019a468b68952a60fe6dbab5d67cd2a103cac7659b41ca/jupyter_client-8.6.3.tar.gz", hash = "sha256:35b3a0947c4a6e9d589eb97d7d4cd5e90f910ee73101611f01283732bd6d9419", size = 342019 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/11/85/b0394e0b6fcccd2c1eeefc230978a6f8cb0c5df1e4cd3e7625735a0d7d1e/jupyter_client-8.6.3-py3-none-any.whl", hash = "sha256:e8a19cc986cc45905ac3362915f410f3af85424b4c0905e94fa5f2cb08e8f23f", size = 106105 }, -] - -[[package]] -name = "jupyter-console" -version = "6.6.3" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "ipykernel" }, - { name = "ipython" }, - { name = "jupyter-client" }, - { name = "jupyter-core" }, - { name = "prompt-toolkit" }, - { name = "pygments" }, - { name = "pyzmq" }, - { name = "traitlets" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/bd/2d/e2fd31e2fc41c14e2bcb6c976ab732597e907523f6b2420305f9fc7fdbdb/jupyter_console-6.6.3.tar.gz", hash = "sha256:566a4bf31c87adbfadf22cdf846e3069b59a71ed5da71d6ba4d8aaad14a53539", size = 34363 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/ca/77/71d78d58f15c22db16328a476426f7ac4a60d3a5a7ba3b9627ee2f7903d4/jupyter_console-6.6.3-py3-none-any.whl", hash = "sha256:309d33409fcc92ffdad25f0bcdf9a4a9daa61b6f341177570fdac03de5352485", size = 24510 }, -] - -[[package]] -name = "jupyter-core" -version = "5.8.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "platformdirs" }, - { name = "pywin32", marker = "platform_python_implementation != 'PyPy' and sys_platform == 'win32'" }, - { name = "traitlets" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/99/1b/72906d554acfeb588332eaaa6f61577705e9ec752ddb486f302dafa292d9/jupyter_core-5.8.1.tar.gz", hash = "sha256:0a5f9706f70e64786b75acba995988915ebd4601c8a52e534a40b51c95f59941", size = 88923 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/2f/57/6bffd4b20b88da3800c5d691e0337761576ee688eb01299eae865689d2df/jupyter_core-5.8.1-py3-none-any.whl", hash = "sha256:c28d268fc90fb53f1338ded2eb410704c5449a358406e8a948b75706e24863d0", size = 28880 }, -] - -[[package]] -name = "jupyter-events" -version = "0.12.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "jsonschema", extra = ["format-nongpl"] }, - { name = "packaging" }, - { name = "python-json-logger" }, - { name = "pyyaml" }, - { name = "referencing" }, - { name = "rfc3339-validator" }, - { name = "rfc3986-validator" }, - { name = "traitlets" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/9d/c3/306d090461e4cf3cd91eceaff84bede12a8e52cd821c2d20c9a4fd728385/jupyter_events-0.12.0.tar.gz", hash = "sha256:fc3fce98865f6784c9cd0a56a20644fc6098f21c8c33834a8d9fe383c17e554b", size = 62196 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/e2/48/577993f1f99c552f18a0428731a755e06171f9902fa118c379eb7c04ea22/jupyter_events-0.12.0-py3-none-any.whl", hash = "sha256:6464b2fa5ad10451c3d35fabc75eab39556ae1e2853ad0c0cc31b656731a97fb", size = 19430 }, -] - -[[package]] -name = "jupyter-lsp" -version = "2.2.5" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "jupyter-server" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/85/b4/3200b0b09c12bc3b72d943d923323c398eff382d1dcc7c0dbc8b74630e40/jupyter-lsp-2.2.5.tar.gz", hash = "sha256:793147a05ad446f809fd53ef1cd19a9f5256fd0a2d6b7ce943a982cb4f545001", size = 48741 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/07/e0/7bd7cff65594fd9936e2f9385701e44574fc7d721331ff676ce440b14100/jupyter_lsp-2.2.5-py3-none-any.whl", hash = "sha256:45fbddbd505f3fbfb0b6cb2f1bc5e15e83ab7c79cd6e89416b248cb3c00c11da", size = 69146 }, -] - -[[package]] -name = "jupyter-server" -version = "2.16.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "anyio" }, - { name = "argon2-cffi" }, - { name = "jinja2" }, - { name = "jupyter-client" }, - { name = "jupyter-core" }, - { name = "jupyter-events" }, - { name = "jupyter-server-terminals" }, - { name = "nbconvert" }, - { name = "nbformat" }, - { name = "overrides" }, - { name = "packaging" }, - { name = "prometheus-client" }, - { name = "pywinpty", marker = "os_name == 'nt'" }, - { name = "pyzmq" }, - { name = "send2trash" }, - { name = "terminado" }, - { name = "tornado" }, - { name = "traitlets" }, - { name = "websocket-client" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/41/c8/ba2bbcd758c47f1124c4ca14061e8ce60d9c6fd537faee9534a95f83521a/jupyter_server-2.16.0.tar.gz", hash = "sha256:65d4b44fdf2dcbbdfe0aa1ace4a842d4aaf746a2b7b168134d5aaed35621b7f6", size = 728177 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/46/1f/5ebbced977171d09a7b0c08a285ff9a20aafb9c51bde07e52349ff1ddd71/jupyter_server-2.16.0-py3-none-any.whl", hash = "sha256:3d8db5be3bc64403b1c65b400a1d7f4647a5ce743f3b20dbdefe8ddb7b55af9e", size = 386904 }, -] - -[[package]] -name = "jupyter-server-terminals" -version = "0.5.3" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "pywinpty", marker = "os_name == 'nt'" }, - { name = "terminado" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/fc/d5/562469734f476159e99a55426d697cbf8e7eb5efe89fb0e0b4f83a3d3459/jupyter_server_terminals-0.5.3.tar.gz", hash = "sha256:5ae0295167220e9ace0edcfdb212afd2b01ee8d179fe6f23c899590e9b8a5269", size = 31430 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/07/2d/2b32cdbe8d2a602f697a649798554e4f072115438e92249624e532e8aca6/jupyter_server_terminals-0.5.3-py3-none-any.whl", hash = "sha256:41ee0d7dc0ebf2809c668e0fc726dfaf258fcd3e769568996ca731b6194ae9aa", size = 13656 }, -] - -[[package]] -name = "jupyterlab" -version = "4.4.4" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "async-lru" }, - { name = "httpx" }, - { name = "ipykernel" }, - { name = "jinja2" }, - { name = "jupyter-core" }, - { name = "jupyter-lsp" }, - { name = "jupyter-server" }, - { name = "jupyterlab-server" }, - { name = "notebook-shim" }, - { name = "packaging" }, - { name = "setuptools" }, - { name = "tomli", marker = "python_full_version == '3.10.12'" }, - { name = "tornado" }, - { name = "traitlets" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/e2/4d/7ca5b46ea56742880d71a768a9e6fb8f8482228427eb89492d55c5d0bb7d/jupyterlab-4.4.4.tar.gz", hash = "sha256:163fee1ef702e0a057f75d2eed3ed1da8a986d59eb002cbeb6f0c2779e6cd153", size = 23044296 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/f8/82/66910ce0995dbfdb33609f41c99fe32ce483b9624a3e7d672af14ff63b9f/jupyterlab-4.4.4-py3-none-any.whl", hash = "sha256:711611e4f59851152eb93316c3547c3ec6291f16bb455f1f4fa380d25637e0dd", size = 12296310 }, -] - -[[package]] -name = "jupyterlab-pygments" -version = "0.3.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/90/51/9187be60d989df97f5f0aba133fa54e7300f17616e065d1ada7d7646b6d6/jupyterlab_pygments-0.3.0.tar.gz", hash = "sha256:721aca4d9029252b11cfa9d185e5b5af4d54772bb8072f9b7036f4170054d35d", size = 512900 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/b1/dd/ead9d8ea85bf202d90cc513b533f9c363121c7792674f78e0d8a854b63b4/jupyterlab_pygments-0.3.0-py3-none-any.whl", hash = "sha256:841a89020971da1d8693f1a99997aefc5dc424bb1b251fd6322462a1b8842780", size = 15884 }, -] - -[[package]] -name = "jupyterlab-server" -version = "2.27.3" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "babel" }, - { name = "jinja2" }, - { name = "json5" }, - { name = "jsonschema" }, - { name = "jupyter-server" }, - { name = "packaging" }, - { name = "requests" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/0a/c9/a883ce65eb27905ce77ace410d83587c82ea64dc85a48d1f7ed52bcfa68d/jupyterlab_server-2.27.3.tar.gz", hash = "sha256:eb36caca59e74471988f0ae25c77945610b887f777255aa21f8065def9e51ed4", size = 76173 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/54/09/2032e7d15c544a0e3cd831c51d77a8ca57f7555b2e1b2922142eddb02a84/jupyterlab_server-2.27.3-py3-none-any.whl", hash = "sha256:e697488f66c3db49df675158a77b3b017520d772c6e1548c7d9bcc5df7944ee4", size = 59700 }, -] - -[[package]] -name = "jupyterlab-widgets" -version = "3.0.15" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/b9/7d/160595ca88ee87ac6ba95d82177d29ec60aaa63821d3077babb22ce031a5/jupyterlab_widgets-3.0.15.tar.gz", hash = "sha256:2920888a0c2922351a9202817957a68c07d99673504d6cd37345299e971bb08b", size = 213149 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/43/6a/ca128561b22b60bd5a0c4ea26649e68c8556b82bc70a0c396eebc977fe86/jupyterlab_widgets-3.0.15-py3-none-any.whl", hash = "sha256:d59023d7d7ef71400d51e6fee9a88867f6e65e10a4201605d2d7f3e8f012a31c", size = 216571 }, -] - -[[package]] -name = "kaldi-python-io" -version = "1.2.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "numpy" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/80/45/e3e542ffa8970ebd782fcece35e2295de9c60e8c396c2c1a403410d1b24e/kaldi-python-io-1.2.2.tar.gz", hash = "sha256:4ebb4029c6c58296cc0abf96edff02832ba341d290ed37624a8d00105f0f7c00", size = 8814 } - -[[package]] -name = "kaldiio" -version = "2.18.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "numpy" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/8d/85/92435e8e62eb3d43eded9f24643fc2a6dbce031cebceed11528147c7873f/kaldiio-2.18.1.tar.gz", hash = "sha256:0283d197fac6ac683f7a9e6af8d18aad9dbd2c4a997f22e45294f2ac1ee3c432", size = 35570 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/ae/e3/6c3b42233225f398f7a72988b524f654ae818cca0d441db847a2761203e9/kaldiio-2.18.1-py3-none-any.whl", hash = "sha256:397a4cd18977acaae7acabfba6807ee0a6978c620064381a266eac15b3c1a0a0", size = 29330 }, -] - -[[package]] -name = "kiwisolver" -version = "1.4.8" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/82/59/7c91426a8ac292e1cdd53a63b6d9439abd573c875c3f92c146767dd33faf/kiwisolver-1.4.8.tar.gz", hash = "sha256:23d5f023bdc8c7e54eb65f03ca5d5bb25b601eac4d7f1a042888a1f45237987e", size = 97538 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/47/5f/4d8e9e852d98ecd26cdf8eaf7ed8bc33174033bba5e07001b289f07308fd/kiwisolver-1.4.8-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:88c6f252f6816a73b1f8c904f7bbe02fd67c09a69f7cb8a0eecdbf5ce78e63db", size = 124623 }, - { url = "https://files.pythonhosted.org/packages/1d/70/7f5af2a18a76fe92ea14675f8bd88ce53ee79e37900fa5f1a1d8e0b42998/kiwisolver-1.4.8-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:c72941acb7b67138f35b879bbe85be0f6c6a70cab78fe3ef6db9c024d9223e5b", size = 66720 }, - { url = "https://files.pythonhosted.org/packages/c6/13/e15f804a142353aefd089fadc8f1d985561a15358c97aca27b0979cb0785/kiwisolver-1.4.8-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:ce2cf1e5688edcb727fdf7cd1bbd0b6416758996826a8be1d958f91880d0809d", size = 65413 }, - { url = "https://files.pythonhosted.org/packages/ce/6d/67d36c4d2054e83fb875c6b59d0809d5c530de8148846b1370475eeeece9/kiwisolver-1.4.8-cp310-cp310-manylinux_2_12_i686.manylinux2010_i686.whl", hash = "sha256:c8bf637892dc6e6aad2bc6d4d69d08764166e5e3f69d469e55427b6ac001b19d", size = 1650826 }, - { url = "https://files.pythonhosted.org/packages/de/c6/7b9bb8044e150d4d1558423a1568e4f227193662a02231064e3824f37e0a/kiwisolver-1.4.8-cp310-cp310-manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:034d2c891f76bd3edbdb3ea11140d8510dca675443da7304205a2eaa45d8334c", size = 1628231 }, - { url = "https://files.pythonhosted.org/packages/b6/38/ad10d437563063eaaedbe2c3540a71101fc7fb07a7e71f855e93ea4de605/kiwisolver-1.4.8-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d47b28d1dfe0793d5e96bce90835e17edf9a499b53969b03c6c47ea5985844c3", size = 1408938 }, - { url = "https://files.pythonhosted.org/packages/52/ce/c0106b3bd7f9e665c5f5bc1e07cc95b5dabd4e08e3dad42dbe2faad467e7/kiwisolver-1.4.8-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:eb158fe28ca0c29f2260cca8c43005329ad58452c36f0edf298204de32a9a3ed", size = 1422799 }, - { url = "https://files.pythonhosted.org/packages/d0/87/efb704b1d75dc9758087ba374c0f23d3254505edaedd09cf9d247f7878b9/kiwisolver-1.4.8-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:d5536185fce131780ebd809f8e623bf4030ce1b161353166c49a3c74c287897f", size = 1354362 }, - { url = "https://files.pythonhosted.org/packages/eb/b3/fd760dc214ec9a8f208b99e42e8f0130ff4b384eca8b29dd0efc62052176/kiwisolver-1.4.8-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:369b75d40abedc1da2c1f4de13f3482cb99e3237b38726710f4a793432b1c5ff", size = 2222695 }, - { url = "https://files.pythonhosted.org/packages/a2/09/a27fb36cca3fc01700687cc45dae7a6a5f8eeb5f657b9f710f788748e10d/kiwisolver-1.4.8-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:641f2ddf9358c80faa22e22eb4c9f54bd3f0e442e038728f500e3b978d00aa7d", size = 2370802 }, - { url = "https://files.pythonhosted.org/packages/3d/c3/ba0a0346db35fe4dc1f2f2cf8b99362fbb922d7562e5f911f7ce7a7b60fa/kiwisolver-1.4.8-cp310-cp310-musllinux_1_2_ppc64le.whl", hash = "sha256:d561d2d8883e0819445cfe58d7ddd673e4015c3c57261d7bdcd3710d0d14005c", size = 2334646 }, - { url = "https://files.pythonhosted.org/packages/41/52/942cf69e562f5ed253ac67d5c92a693745f0bed3c81f49fc0cbebe4d6b00/kiwisolver-1.4.8-cp310-cp310-musllinux_1_2_s390x.whl", hash = "sha256:1732e065704b47c9afca7ffa272f845300a4eb959276bf6970dc07265e73b605", size = 2467260 }, - { url = "https://files.pythonhosted.org/packages/32/26/2d9668f30d8a494b0411d4d7d4ea1345ba12deb6a75274d58dd6ea01e951/kiwisolver-1.4.8-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:bcb1ebc3547619c3b58a39e2448af089ea2ef44b37988caf432447374941574e", size = 2288633 }, - { url = "https://files.pythonhosted.org/packages/98/99/0dd05071654aa44fe5d5e350729961e7bb535372935a45ac89a8924316e6/kiwisolver-1.4.8-cp310-cp310-win_amd64.whl", hash = "sha256:89c107041f7b27844179ea9c85d6da275aa55ecf28413e87624d033cf1f6b751", size = 71885 }, - { url = "https://files.pythonhosted.org/packages/6c/fc/822e532262a97442989335394d441cd1d0448c2e46d26d3e04efca84df22/kiwisolver-1.4.8-cp310-cp310-win_arm64.whl", hash = "sha256:b5773efa2be9eb9fcf5415ea3ab70fc785d598729fd6057bea38d539ead28271", size = 65175 }, - { url = "https://files.pythonhosted.org/packages/1f/f9/ae81c47a43e33b93b0a9819cac6723257f5da2a5a60daf46aa5c7226ea85/kiwisolver-1.4.8-pp310-pypy310_pp73-macosx_10_15_x86_64.whl", hash = "sha256:e7a019419b7b510f0f7c9dceff8c5eae2392037eae483a7f9162625233802b0a", size = 60403 }, - { url = "https://files.pythonhosted.org/packages/58/ca/f92b5cb6f4ce0c1ebfcfe3e2e42b96917e16f7090e45b21102941924f18f/kiwisolver-1.4.8-pp310-pypy310_pp73-macosx_11_0_arm64.whl", hash = "sha256:286b18e86682fd2217a48fc6be6b0f20c1d0ed10958d8dc53453ad58d7be0bf8", size = 58657 }, - { url = "https://files.pythonhosted.org/packages/80/28/ae0240f732f0484d3a4dc885d055653c47144bdf59b670aae0ec3c65a7c8/kiwisolver-1.4.8-pp310-pypy310_pp73-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:4191ee8dfd0be1c3666ccbac178c5a05d5f8d689bbe3fc92f3c4abec817f8fe0", size = 84948 }, - { url = "https://files.pythonhosted.org/packages/5d/eb/78d50346c51db22c7203c1611f9b513075f35c4e0e4877c5dde378d66043/kiwisolver-1.4.8-pp310-pypy310_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7cd2785b9391f2873ad46088ed7599a6a71e762e1ea33e87514b1a441ed1da1c", size = 81186 }, - { url = "https://files.pythonhosted.org/packages/43/f8/7259f18c77adca88d5f64f9a522792e178b2691f3748817a8750c2d216ef/kiwisolver-1.4.8-pp310-pypy310_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c07b29089b7ba090b6f1a669f1411f27221c3662b3a1b7010e67b59bb5a6f10b", size = 80279 }, - { url = "https://files.pythonhosted.org/packages/3a/1d/50ad811d1c5dae091e4cf046beba925bcae0a610e79ae4c538f996f63ed5/kiwisolver-1.4.8-pp310-pypy310_pp73-win_amd64.whl", hash = "sha256:65ea09a5a3faadd59c2ce96dc7bf0f364986a315949dc6374f04396b0d60e09b", size = 71762 }, -] - -[[package]] -name = "lazy-loader" -version = "0.4" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "packaging" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/6f/6b/c875b30a1ba490860c93da4cabf479e03f584eba06fe5963f6f6644653d8/lazy_loader-0.4.tar.gz", hash = "sha256:47c75182589b91a4e1a85a136c074285a5ad4d9f39c63e0d7fb76391c4574cd1", size = 15431 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/83/60/d497a310bde3f01cb805196ac61b7ad6dc5dcf8dce66634dc34364b20b4f/lazy_loader-0.4-py3-none-any.whl", hash = "sha256:342aa8e14d543a154047afb4ba8ef17f5563baad3fc610d7b15b213b0f119efc", size = 12097 }, -] - -[[package]] -name = "levenshtein" -version = "0.27.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "rapidfuzz" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/7e/b3/b5f8011483ba9083a0bc74c4d58705e9cf465fbe55c948a1b1357d0a2aa8/levenshtein-0.27.1.tar.gz", hash = "sha256:3e18b73564cfc846eec94dd13fab6cb006b5d2e0cc56bad1fd7d5585881302e3", size = 382571 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/b3/b1/9906a75b98dd9c008015a72d7658be53851e361a35492631edf1b1f334ab/levenshtein-0.27.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:13d6f617cb6fe63714c4794861cfaacd398db58a292f930edb7f12aad931dace", size = 174542 }, - { url = "https://files.pythonhosted.org/packages/3b/57/e26e0164a93fb045316856603111d95538cac8224a3709e4ac96a6bb74f3/levenshtein-0.27.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:ca9d54d41075e130c390e61360bec80f116b62d6ae973aec502e77e921e95334", size = 156367 }, - { url = "https://files.pythonhosted.org/packages/6d/dd/92fcb71d48c1fe69c46c211156adafb8175037dc63e80e970106aef3f9d5/levenshtein-0.27.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:2de1f822b5c9a20d10411f779dfd7181ce3407261436f8470008a98276a9d07f", size = 152189 }, - { url = "https://files.pythonhosted.org/packages/5e/23/3f331f5fbfa93634126439cfc8c01b31f7ef1fbedb81663581e27a69da4d/levenshtein-0.27.1-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:81270392c2e45d1a7e1b3047c3a272d5e28bb4f1eff0137637980064948929b7", size = 184271 }, - { url = "https://files.pythonhosted.org/packages/5a/76/d6ac541a1a80bdc5c98584a6a2d2301e677af4cb2e4092247207791b56a6/levenshtein-0.27.1-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:2d30c3ea23a94dddd56dbe323e1fa8a29ceb24da18e2daa8d0abf78b269a5ad1", size = 185078 }, - { url = "https://files.pythonhosted.org/packages/2d/ed/d0c5abe8cfcf6a7f2a4197e889e12b7a0c2145a0ef3354b1c000bf367305/levenshtein-0.27.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0f3e0bea76695b9045bbf9ad5f67ad4cc01c11f783368f34760e068f19b6a6bc", size = 161505 }, - { url = "https://files.pythonhosted.org/packages/f3/28/a5b78e1818211bc6407590876bbdcc6d79671e529a0c186780492c1f2136/levenshtein-0.27.1-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:cdd190e468a68c31a5943368a5eaf4e130256a8707886d23ab5906a0cb98a43c", size = 246968 }, - { url = "https://files.pythonhosted.org/packages/77/7f/981b903583956cb67b33bed39d9840ab5e4c7062bceec564b7bf2c3f6f49/levenshtein-0.27.1-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:7c3121314bb4b676c011c33f6a0ebb462cfdcf378ff383e6f9e4cca5618d0ba7", size = 1116000 }, - { url = "https://files.pythonhosted.org/packages/75/1d/c4be47d5f436fd310373c5ebdf05828c1d95be9a44c3e94f29c40937b30c/levenshtein-0.27.1-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:f8ef378c873efcc5e978026b69b45342d841cd7a2f273447324f1c687cc4dc37", size = 1401162 }, - { url = "https://files.pythonhosted.org/packages/91/e4/0b107676efe3ecd5fada1ed3a3bbddd4c829e2ef34e980b76374c116235b/levenshtein-0.27.1-cp310-cp310-musllinux_1_2_ppc64le.whl", hash = "sha256:ff18d78c5c16bea20876425e1bf5af56c25918fb01bc0f2532db1317d4c0e157", size = 1225141 }, - { url = "https://files.pythonhosted.org/packages/29/f0/f3f88d766fdbb1d39fe98dc5527223bae099444e501550ae088c47ddd97b/levenshtein-0.27.1-cp310-cp310-musllinux_1_2_s390x.whl", hash = "sha256:13412ff805afbfe619d070280d1a76eb4198c60c5445cd5478bd4c7055bb3d51", size = 1419707 }, - { url = "https://files.pythonhosted.org/packages/b8/1c/f51ac1db4064a85effa50df240250e413f428164301d836c312baf09381e/levenshtein-0.27.1-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:a2adb9f263557f7fb13e19eb2f34595d86929a44c250b2fca6e9b65971e51e20", size = 1189284 }, - { url = "https://files.pythonhosted.org/packages/e0/67/5ace76bc964b93ed6203a9f8c4dcde1a50e336468f7da3a21dd29febaf46/levenshtein-0.27.1-cp310-cp310-win32.whl", hash = "sha256:6278a33d2e0e909d8829b5a72191419c86dd3bb45b82399c7efc53dabe870c35", size = 88036 }, - { url = "https://files.pythonhosted.org/packages/06/e0/d9737dbbe85842ddb300cb7974fc065edc56ec647652863f95ac1977d378/levenshtein-0.27.1-cp310-cp310-win_amd64.whl", hash = "sha256:5b602b8428ee5dc88432a55c5303a739ee2be7c15175bd67c29476a9d942f48e", size = 99922 }, - { url = "https://files.pythonhosted.org/packages/27/b8/13e22789ab700db0da98f973a508643dbe2d25bd0fb5dc53239e0e2852c1/levenshtein-0.27.1-cp310-cp310-win_arm64.whl", hash = "sha256:48334081fddaa0c259ba01ee898640a2cf8ede62e5f7e25fefece1c64d34837f", size = 87846 }, - { url = "https://files.pythonhosted.org/packages/25/ed/37e2d1f5e690d7376cd7e8bdd19411479ff352a3df9ab5f845dd680ef779/levenshtein-0.27.1-pp310-pypy310_pp73-macosx_10_15_x86_64.whl", hash = "sha256:c92a222ab95b8d903eae6d5e7d51fe6c999be021b647715c18d04d0b0880f463", size = 170482 }, - { url = "https://files.pythonhosted.org/packages/6d/9f/30b1144b9d1da74743e7d7cdf47575b7013c9767e608c7454dbd318aacd2/levenshtein-0.27.1-pp310-pypy310_pp73-macosx_11_0_arm64.whl", hash = "sha256:71afc36b4ee950fa1140aff22ffda9e5e23280285858e1303260dbb2eabf342d", size = 153106 }, - { url = "https://files.pythonhosted.org/packages/b1/c5/18d0bec94a166cebaefa3db4beab9a7e0d75412b52e9626f5dce1ca8d149/levenshtein-0.27.1-pp310-pypy310_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:58b1daeebfc148a571f09cfe18c16911ea1eaaa9e51065c5f7e7acbc4b866afa", size = 150984 }, - { url = "https://files.pythonhosted.org/packages/55/b4/4b80eb0c96caabdb683256cac9cc2cc9a73dee8ea80ab7cc3ee8aebd603f/levenshtein-0.27.1-pp310-pypy310_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:105edcb14797d95c77f69bad23104314715a64cafbf4b0e79d354a33d7b54d8d", size = 158673 }, - { url = "https://files.pythonhosted.org/packages/81/14/a43daefbc6d5e5561176150363cbac73003795b85ae136ffd4d0691af3fb/levenshtein-0.27.1-pp310-pypy310_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:d9c58fb1ef8bdc8773d705fbacf628e12c3bb63ee4d065dda18a76e86042444a", size = 244419 }, - { url = "https://files.pythonhosted.org/packages/d0/55/34f133f4f0998d7335bd96b9d315dc888b118e48e999c3d2c621b84965b9/levenshtein-0.27.1-pp310-pypy310_pp73-win_amd64.whl", hash = "sha256:e52270591854af67217103955a36bd7436b57c801e3354e73ba44d689ed93697", size = 97932 }, -] - -[[package]] -name = "lhotse" -version = "1.30.3" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "audioread" }, - { name = "click" }, - { name = "cytoolz" }, - { name = "intervaltree" }, - { name = "lilcom" }, - { name = "numpy" }, - { name = "packaging" }, - { name = "pyyaml" }, - { name = "soundfile" }, - { name = "tabulate" }, - { name = "torch" }, - { name = "tqdm" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/a4/28/4ad81e95f2abaf101532cc7f599223abbfa0faa19caa1232496e7cb60f19/lhotse-1.30.3.tar.gz", hash = "sha256:f39eaeab795b85d35cf7e6ed315246252c6565ee4d0f767781ba23a8992f2df9", size = 643240 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/f7/35/a0ddacd91f13267653262eba0933d2750fe63cff2fddcb8d73bb584090cb/lhotse-1.30.3-py3-none-any.whl", hash = "sha256:80df401000ac4b016721bcf27a575d72a4f31ace914bb95db2e118e61556a3ac", size = 851420 }, -] - -[[package]] -name = "libcst" -version = "1.8.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "pyyaml", marker = "python_full_version == '3.10.12'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/89/aa/b52d195b167958fe1bd106a260f64cc80ec384f6ac2a9cda874d8803df06/libcst-1.8.2.tar.gz", hash = "sha256:66e82cedba95a6176194a817be4232c720312f8be6d2c8f3847f3317d95a0c7f", size = 881534 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/3c/2e/1d7f67d2ef6f875e9e8798c024f7cb3af3fe861e417bff485c69b655ac96/libcst-1.8.2-cp310-cp310-macosx_10_12_x86_64.whl", hash = "sha256:67d9720d91f507c87b3e5f070627ad640a00bc6cfdf5635f8c6ee9f2964cf71c", size = 2195106 }, - { url = "https://files.pythonhosted.org/packages/82/d0/3d94fee2685f263fd8d85a83e2537fcc78b644eae450738bf2c72604f0df/libcst-1.8.2-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:94b7c032b72566077614a02baab1929739fd0af0cc1d46deaba4408b870faef2", size = 2080577 }, - { url = "https://files.pythonhosted.org/packages/14/87/c9b49bebb9a930fdcb59bf841f1c45719d2a4a39c3eb7efacfd30a2bfb0a/libcst-1.8.2-cp310-cp310-manylinux1_i686.manylinux2014_i686.manylinux_2_17_i686.manylinux_2_5_i686.whl", hash = "sha256:11ea148902e3e1688afa392087c728ac3a843e54a87d334d1464d2097d3debb7", size = 2404076 }, - { url = "https://files.pythonhosted.org/packages/49/fa/9ca145aa9033f9a8362a5663ceb28dfb67082574de8118424b6b8e445e7a/libcst-1.8.2-cp310-cp310-manylinux_2_28_aarch64.whl", hash = "sha256:22c9473a2cc53faabcc95a0ac6ca4e52d127017bf34ba9bc0f8e472e44f7b38e", size = 2219813 }, - { url = "https://files.pythonhosted.org/packages/0c/25/496a025c09e96116437a57fd34abefe84c041d930f832c6e42d84d9e028c/libcst-1.8.2-cp310-cp310-manylinux_2_28_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:b5269b96367e65793a7714608f6d906418eb056d59eaac9bba980486aabddbed", size = 2189782 }, - { url = "https://files.pythonhosted.org/packages/b3/75/826b5772192826d70480efe93bab3e4f0b4a24d31031f45547257ad5f9a8/libcst-1.8.2-cp310-cp310-manylinux_2_28_x86_64.whl", hash = "sha256:d20e932ddd9a389da57b060c26e84a24118c96ff6fc5dcc7b784da24e823b694", size = 2312403 }, - { url = "https://files.pythonhosted.org/packages/93/f4/316fa14ea6c61ea8755672d60e012558f0216300b3819e72bebc7864a507/libcst-1.8.2-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:a553d452004e44b841788f6faa7231a02157527ddecc89dbbe5b689b74822226", size = 2280566 }, - { url = "https://files.pythonhosted.org/packages/fc/52/74b69350db379b1646739288b88ffab2981b2ad48407faf03df3768d7d2f/libcst-1.8.2-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:7fe762c4c390039b79b818cbc725d8663586b25351dc18a2704b0e357d69b924", size = 2388508 }, - { url = "https://files.pythonhosted.org/packages/bc/c6/fa92699b537ed65e93c2869144e23bdf156ec81ae7b84b4f34cbc20d6048/libcst-1.8.2-cp310-cp310-win_amd64.whl", hash = "sha256:5c513e64eff0f7bf2a908e2d987a98653eb33e1062ce2afd3a84af58159a24f9", size = 2093260 }, - { url = "https://files.pythonhosted.org/packages/b0/ac/4ec4ae9da311f72cd97e930c325bb605e9ad0baaafcafadb0588e1dc5c4e/libcst-1.8.2-cp310-cp310-win_arm64.whl", hash = "sha256:41613fe08e647213546c7c59a5a1fc5484666e7d4cab6e80260c612acbb20e8c", size = 1985236 }, -] - -[[package]] -name = "librosa" -version = "0.11.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "audioread" }, - { name = "decorator" }, - { name = "joblib" }, - { name = "lazy-loader" }, - { name = "msgpack" }, - { name = "numba" }, - { name = "numpy" }, - { name = "pooch" }, - { name = "scikit-learn" }, - { name = "scipy" }, - { name = "soundfile" }, - { name = "soxr" }, - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/64/36/360b5aafa0238e29758729e9486c6ed92a6f37fa403b7875e06c115cdf4a/librosa-0.11.0.tar.gz", hash = "sha256:f5ed951ca189b375bbe2e33b2abd7e040ceeee302b9bbaeeffdfddb8d0ace908", size = 327001 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/b5/ba/c63c5786dfee4c3417094c4b00966e61e4a63efecee22cb7b4c0387dda83/librosa-0.11.0-py3-none-any.whl", hash = "sha256:0b6415c4fd68bff4c29288abe67c6d80b587e0e1e2cfb0aad23e4559504a7fa1", size = 260749 }, -] - -[[package]] -name = "lightning" -version = "2.4.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "fsspec", extra = ["http"] }, - { name = "lightning-utilities" }, - { name = "packaging" }, - { name = "pytorch-lightning" }, - { name = "pyyaml" }, - { name = "torch" }, - { name = "torchmetrics" }, - { name = "tqdm" }, - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/56/d0/78ea244ac044cd4df15aa8294a50ff3561fb177e7e5ba788aaa542046cae/lightning-2.4.0.tar.gz", hash = "sha256:9156604cc56e4b2b603f34fa7f0fe5107375c8e6d85e74544b319a15faa9ed0e", size = 620632 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/a3/2c/85eaf42c983b0cd81bcda5876da2c8e2a9fd347908666ea9855724369171/lightning-2.4.0-py3-none-any.whl", hash = "sha256:560163af9711cf59055c448232c473150a299089efce0d2be3cc3288082d8768", size = 810971 }, -] - -[[package]] -name = "lightning-utilities" -version = "0.14.3" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "packaging" }, - { name = "setuptools" }, - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/0f/bb/63a6a8c9e7a96b6ba92647fa5b1595c2dbee29f8178705adb4704d82ecba/lightning_utilities-0.14.3.tar.gz", hash = "sha256:37e2f83f273890052955a44054382c211a303012ee577619efbaa5df9e65e9f5", size = 30346 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/1a/c1/31b3184cba7b257a4a3b5ca5b88b9204ccb7aa02fe3c992280899293ed54/lightning_utilities-0.14.3-py3-none-any.whl", hash = "sha256:4ab9066aa36cd7b93a05713808901909e96cc3f187ea6fd3052b2fd91313b468", size = 28894 }, -] - -[[package]] -name = "lilcom" -version = "1.8.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "numpy" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/8b/b5/97422825e61a2683dd39d78165fc470c1153b4a908dd5cd0711c0e9d262c/lilcom-1.8.1.tar.gz", hash = "sha256:69c62037c92e71e601ac3bb3ae19811f22ceffbdf58b0fdbf81cc6a0ec6fc3c5", size = 45813 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/e1/5c/51ece6a8f6ed39c27eb18bb1c09fe6508613bc67bcc5872caa788cdc9eef/lilcom-1.8.1-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:7e896dd6172011d320c26a4ca648edddf1e6b47bb1de749cf27ac096d73f63c9", size = 122436 }, - { url = "https://files.pythonhosted.org/packages/58/fd/8008d8b82965eeb2e35a46edb118e195d10bc424e3d80f2e3dafe3044e1f/lilcom-1.8.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:6b7c5488267c9c3c6ab49c3e3ec287031029d7de57f4c2307264dd75f5e77211", size = 86593 }, - { url = "https://files.pythonhosted.org/packages/7b/02/30e76ca2bfc3fe19adbcca254dde03ba48a6e623eb55df2752f4c06cfeb0/lilcom-1.8.1-cp310-cp310-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:7460e2dd481b38fc8fa1f0cc48c5aaa1b4206d0c27130c001bf23f1d4c99143b", size = 98968 }, - { url = "https://files.pythonhosted.org/packages/d7/cd/e3e45ef07c2e19f5718d7c9d33b40438b83a3397bdd4c70bd3d1bbf059ad/lilcom-1.8.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9bdeea4aa3d7cad9e38cac99ef012a61f843a87a661b9475f565787f73828ec5", size = 92337 }, - { url = "https://files.pythonhosted.org/packages/84/da/3b8a45be54f4d12a2d118c84d9d2aab81e5b75ae4a7b8469213a28b121dd/lilcom-1.8.1-cp310-cp310-win_amd64.whl", hash = "sha256:c254a74156ff8da2e43c8b891ed47eefcf0c1fee82a0c5984388052d97573a1b", size = 68524 }, -] - -[[package]] -name = "llvmlite" -version = "0.44.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/89/6a/95a3d3610d5c75293d5dbbb2a76480d5d4eeba641557b69fe90af6c5b84e/llvmlite-0.44.0.tar.gz", hash = "sha256:07667d66a5d150abed9157ab6c0b9393c9356f229784a4385c02f99e94fc94d4", size = 171880 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/41/75/d4863ddfd8ab5f6e70f4504cf8cc37f4e986ec6910f4ef8502bb7d3c1c71/llvmlite-0.44.0-cp310-cp310-macosx_10_14_x86_64.whl", hash = "sha256:9fbadbfba8422123bab5535b293da1cf72f9f478a65645ecd73e781f962ca614", size = 28132306 }, - { url = "https://files.pythonhosted.org/packages/37/d9/6e8943e1515d2f1003e8278819ec03e4e653e2eeb71e4d00de6cfe59424e/llvmlite-0.44.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:cccf8eb28f24840f2689fb1a45f9c0f7e582dd24e088dcf96e424834af11f791", size = 26201096 }, - { url = "https://files.pythonhosted.org/packages/aa/46/8ffbc114def88cc698906bf5acab54ca9fdf9214fe04aed0e71731fb3688/llvmlite-0.44.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:7202b678cdf904823c764ee0fe2dfe38a76981f4c1e51715b4cb5abb6cf1d9e8", size = 42361859 }, - { url = "https://files.pythonhosted.org/packages/30/1c/9366b29ab050a726af13ebaae8d0dff00c3c58562261c79c635ad4f5eb71/llvmlite-0.44.0-cp310-cp310-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:40526fb5e313d7b96bda4cbb2c85cd5374e04d80732dd36a282d72a560bb6408", size = 41184199 }, - { url = "https://files.pythonhosted.org/packages/69/07/35e7c594b021ecb1938540f5bce543ddd8713cff97f71d81f021221edc1b/llvmlite-0.44.0-cp310-cp310-win_amd64.whl", hash = "sha256:41e3839150db4330e1b2716c0be3b5c4672525b4c9005e17c7597f835f351ce2", size = 30332381 }, -] - -[[package]] -name = "loguru" -version = "0.7.3" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "colorama", marker = "sys_platform == 'win32'" }, - { name = "win32-setctime", marker = "sys_platform == 'win32'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/3a/05/a1dae3dffd1116099471c643b8924f5aa6524411dc6c63fdae648c4f1aca/loguru-0.7.3.tar.gz", hash = "sha256:19480589e77d47b8d85b2c827ad95d49bf31b0dcde16593892eb51dd18706eb6", size = 63559 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/0c/29/0348de65b8cc732daa3e33e67806420b2ae89bdce2b04af740289c5c6c8c/loguru-0.7.3-py3-none-any.whl", hash = "sha256:31a33c10c8e1e10422bfd431aeb5d351c7cf7fa671e3c4df004162264b28220c", size = 61595 }, -] - -[[package]] -name = "mako" -version = "1.3.10" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "markupsafe" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/9e/38/bd5b78a920a64d708fe6bc8e0a2c075e1389d53bef8413725c63ba041535/mako-1.3.10.tar.gz", hash = "sha256:99579a6f39583fa7e5630a28c3c1f440e4e97a414b80372649c0ce338da2ea28", size = 392474 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/87/fb/99f81ac72ae23375f22b7afdb7642aba97c00a713c217124420147681a2f/mako-1.3.10-py3-none-any.whl", hash = "sha256:baef24a52fc4fc514a0887ac600f9f1cff3d82c61d4d700a1fa84d597b88db59", size = 78509 }, -] - -[[package]] -name = "markdown" -version = "3.8.2" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/d7/c2/4ab49206c17f75cb08d6311171f2d65798988db4360c4d1485bd0eedd67c/markdown-3.8.2.tar.gz", hash = "sha256:247b9a70dd12e27f67431ce62523e675b866d254f900c4fe75ce3dda62237c45", size = 362071 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/96/2b/34cc11786bc00d0f04d0f5fdc3a2b1ae0b6239eef72d3d345805f9ad92a1/markdown-3.8.2-py3-none-any.whl", hash = "sha256:5c83764dbd4e00bdd94d85a19b8d55ccca20fe35b2e678a1422b380324dd5f24", size = 106827 }, -] - -[[package]] -name = "markdown-it-py" -version = "3.0.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "mdurl" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/38/71/3b932df36c1a044d397a1f92d1cf91ee0a503d91e470cbd670aa66b07ed0/markdown-it-py-3.0.0.tar.gz", hash = "sha256:e3f60a94fa066dc52ec76661e37c851cb232d92f9886b15cb560aaada2df8feb", size = 74596 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/42/d7/1ec15b46af6af88f19b8e5ffea08fa375d433c998b8a7639e76935c14f1f/markdown_it_py-3.0.0-py3-none-any.whl", hash = "sha256:355216845c60bd96232cd8d8c40e8f9765cc86f46880e43a8fd22dc1a1a8cab1", size = 87528 }, -] - -[[package]] -name = "markupsafe" -version = "3.0.2" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/b2/97/5d42485e71dfc078108a86d6de8fa46db44a1a9295e89c5d6d4a06e23a62/markupsafe-3.0.2.tar.gz", hash = "sha256:ee55d3edf80167e48ea11a923c7386f4669df67d7994554387f84e7d8b0a2bf0", size = 20537 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/04/90/d08277ce111dd22f77149fd1a5d4653eeb3b3eaacbdfcbae5afb2600eebd/MarkupSafe-3.0.2-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:7e94c425039cde14257288fd61dcfb01963e658efbc0ff54f5306b06054700f8", size = 14357 }, - { url = "https://files.pythonhosted.org/packages/04/e1/6e2194baeae0bca1fae6629dc0cbbb968d4d941469cbab11a3872edff374/MarkupSafe-3.0.2-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:9e2d922824181480953426608b81967de705c3cef4d1af983af849d7bd619158", size = 12393 }, - { url = "https://files.pythonhosted.org/packages/1d/69/35fa85a8ece0a437493dc61ce0bb6d459dcba482c34197e3efc829aa357f/MarkupSafe-3.0.2-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:38a9ef736c01fccdd6600705b09dc574584b89bea478200c5fbf112a6b0d5579", size = 21732 }, - { url = "https://files.pythonhosted.org/packages/22/35/137da042dfb4720b638d2937c38a9c2df83fe32d20e8c8f3185dbfef05f7/MarkupSafe-3.0.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:bbcb445fa71794da8f178f0f6d66789a28d7319071af7a496d4d507ed566270d", size = 20866 }, - { url = "https://files.pythonhosted.org/packages/29/28/6d029a903727a1b62edb51863232152fd335d602def598dade38996887f0/MarkupSafe-3.0.2-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:57cb5a3cf367aeb1d316576250f65edec5bb3be939e9247ae594b4bcbc317dfb", size = 20964 }, - { url = "https://files.pythonhosted.org/packages/cc/cd/07438f95f83e8bc028279909d9c9bd39e24149b0d60053a97b2bc4f8aa51/MarkupSafe-3.0.2-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:3809ede931876f5b2ec92eef964286840ed3540dadf803dd570c3b7e13141a3b", size = 21977 }, - { url = "https://files.pythonhosted.org/packages/29/01/84b57395b4cc062f9c4c55ce0df7d3108ca32397299d9df00fedd9117d3d/MarkupSafe-3.0.2-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:e07c3764494e3776c602c1e78e298937c3315ccc9043ead7e685b7f2b8d47b3c", size = 21366 }, - { url = "https://files.pythonhosted.org/packages/bd/6e/61ebf08d8940553afff20d1fb1ba7294b6f8d279df9fd0c0db911b4bbcfd/MarkupSafe-3.0.2-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:b424c77b206d63d500bcb69fa55ed8d0e6a3774056bdc4839fc9298a7edca171", size = 21091 }, - { url = "https://files.pythonhosted.org/packages/11/23/ffbf53694e8c94ebd1e7e491de185124277964344733c45481f32ede2499/MarkupSafe-3.0.2-cp310-cp310-win32.whl", hash = "sha256:fcabf5ff6eea076f859677f5f0b6b5c1a51e70a376b0579e0eadef8db48c6b50", size = 15065 }, - { url = "https://files.pythonhosted.org/packages/44/06/e7175d06dd6e9172d4a69a72592cb3f7a996a9c396eee29082826449bbc3/MarkupSafe-3.0.2-cp310-cp310-win_amd64.whl", hash = "sha256:6af100e168aa82a50e186c82875a5893c5597a0c1ccdb0d8b40240b1f28b969a", size = 15514 }, -] - -[[package]] -name = "marshmallow" -version = "4.0.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "backports-datetime-fromisoformat", marker = "python_full_version == '3.10.12'" }, - { name = "typing-extensions", marker = "python_full_version == '3.10.12'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/1e/ff/26df5a9f5ac57ccf693a5854916ab47243039d2aa9e0fe5f5a0331e7b74b/marshmallow-4.0.0.tar.gz", hash = "sha256:3b6e80aac299a7935cfb97ed01d1854fb90b5079430969af92118ea1b12a8d55", size = 220507 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/d6/26/6cc45d156f44dbe1d5696d9e54042e4dcaf7b946c0b86df6a97d29706f32/marshmallow-4.0.0-py3-none-any.whl", hash = "sha256:e7b0528337e9990fd64950f8a6b3a1baabed09ad17a0dfb844d701151f92d203", size = 48420 }, -] - -[[package]] -name = "matplotlib" -version = "3.10.3" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "contourpy" }, - { name = "cycler" }, - { name = "fonttools" }, - { name = "kiwisolver" }, - { name = "numpy" }, - { name = "packaging" }, - { name = "pillow" }, - { name = "pyparsing" }, - { name = "python-dateutil" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/26/91/d49359a21893183ed2a5b6c76bec40e0b1dcbf8ca148f864d134897cfc75/matplotlib-3.10.3.tar.gz", hash = "sha256:2f82d2c5bb7ae93aaaa4cd42aca65d76ce6376f83304fa3a630b569aca274df0", size = 34799811 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/d0/ea/2bba25d289d389c7451f331ecd593944b3705f06ddf593fa7be75037d308/matplotlib-3.10.3-cp310-cp310-macosx_10_12_x86_64.whl", hash = "sha256:213fadd6348d106ca7db99e113f1bea1e65e383c3ba76e8556ba4a3054b65ae7", size = 8167862 }, - { url = "https://files.pythonhosted.org/packages/41/81/cc70b5138c926604e8c9ed810ed4c79e8116ba72e02230852f5c12c87ba2/matplotlib-3.10.3-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:d3bec61cb8221f0ca6313889308326e7bb303d0d302c5cc9e523b2f2e6c73deb", size = 8042149 }, - { url = "https://files.pythonhosted.org/packages/4a/9a/0ff45b6bfa42bb16de597e6058edf2361c298ad5ef93b327728145161bbf/matplotlib-3.10.3-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8c21ae75651c0231b3ba014b6d5e08fb969c40cdb5a011e33e99ed0c9ea86ecb", size = 8453719 }, - { url = "https://files.pythonhosted.org/packages/85/c7/1866e972fed6d71ef136efbc980d4d1854ab7ef1ea8152bbd995ca231c81/matplotlib-3.10.3-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a49e39755580b08e30e3620efc659330eac5d6534ab7eae50fa5e31f53ee4e30", size = 8590801 }, - { url = "https://files.pythonhosted.org/packages/5d/b9/748f6626d534ab7e255bdc39dc22634d337cf3ce200f261b5d65742044a1/matplotlib-3.10.3-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:cf4636203e1190871d3a73664dea03d26fb019b66692cbfd642faafdad6208e8", size = 9402111 }, - { url = "https://files.pythonhosted.org/packages/1f/78/8bf07bd8fb67ea5665a6af188e70b57fcb2ab67057daa06b85a08e59160a/matplotlib-3.10.3-cp310-cp310-win_amd64.whl", hash = "sha256:fd5641a9bb9d55f4dd2afe897a53b537c834b9012684c8444cc105895c8c16fd", size = 8057213 }, - { url = "https://files.pythonhosted.org/packages/3d/d1/f54d43e95384b312ffa4a74a4326c722f3b8187aaaa12e9a84cdf3037131/matplotlib-3.10.3-pp310-pypy310_pp73-macosx_10_15_x86_64.whl", hash = "sha256:86ab63d66bbc83fdb6733471d3bff40897c1e9921cba112accd748eee4bce5e4", size = 8162896 }, - { url = "https://files.pythonhosted.org/packages/24/a4/fbfc00c2346177c95b353dcf9b5a004106abe8730a62cb6f27e79df0a698/matplotlib-3.10.3-pp310-pypy310_pp73-macosx_11_0_arm64.whl", hash = "sha256:a48f9c08bf7444b5d2391a83e75edb464ccda3c380384b36532a0962593a1751", size = 8039702 }, - { url = "https://files.pythonhosted.org/packages/6a/b9/59e120d24a2ec5fc2d30646adb2efb4621aab3c6d83d66fb2a7a182db032/matplotlib-3.10.3-pp310-pypy310_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:cb73d8aa75a237457988f9765e4dfe1c0d2453c5ca4eabc897d4309672c8e014", size = 8594298 }, -] - -[[package]] -name = "matplotlib-inline" -version = "0.1.7" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "traitlets" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/99/5b/a36a337438a14116b16480db471ad061c36c3694df7c2084a0da7ba538b7/matplotlib_inline-0.1.7.tar.gz", hash = "sha256:8423b23ec666be3d16e16b60bdd8ac4e86e840ebd1dd11a30b9f117f2fa0ab90", size = 8159 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/8f/8e/9ad090d3553c280a8060fbf6e24dc1c0c29704ee7d1c372f0c174aa59285/matplotlib_inline-0.1.7-py3-none-any.whl", hash = "sha256:df192d39a4ff8f21b1895d72e6a13f5fcc5099f00fa84384e0ea28c2cc0653ca", size = 9899 }, -] - -[[package]] -name = "mdurl" -version = "0.1.2" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/d6/54/cfe61301667036ec958cb99bd3efefba235e65cdeb9c84d24a8293ba1d90/mdurl-0.1.2.tar.gz", hash = "sha256:bb413d29f5eea38f31dd4754dd7377d4465116fb207585f97bf925588687c1ba", size = 8729 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/b3/38/89ba8ad64ae25be8de66a6d463314cf1eb366222074cfda9ee839c56a4b4/mdurl-0.1.2-py3-none-any.whl", hash = "sha256:84008a41e51615a49fc9966191ff91509e3c40b939176e643fd50a5c2196b8f8", size = 9979 }, -] - -[[package]] -name = "mediapy" -version = "1.1.6" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "ipython" }, - { name = "matplotlib" }, - { name = "numpy" }, - { name = "pillow" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/9e/b2/451be65c13d2d69b7601eded7ddd3f150884486715a9b3a705ffb08d0177/mediapy-1.1.6.tar.gz", hash = "sha256:9f44b760400964d8bea5121a213f94dc9a225d026d6a819901283a695e585634", size = 25459 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/e3/a0/0d55c59ea8a5f1b13b3eb931e1c0eab9c2a07322cad79cb51a596d2d2a5c/mediapy-1.1.6-py3-none-any.whl", hash = "sha256:c74370808b445666f95272bfdf0eb5707a43b7e05e5527f2dd0830e6892f976f", size = 24955 }, -] - -[[package]] -name = "mistune" -version = "3.1.3" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "typing-extensions", marker = "python_full_version == '3.10.12'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/c4/79/bda47f7dd7c3c55770478d6d02c9960c430b0cf1773b72366ff89126ea31/mistune-3.1.3.tar.gz", hash = "sha256:a7035c21782b2becb6be62f8f25d3df81ccb4d6fa477a6525b15af06539f02a0", size = 94347 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/01/4d/23c4e4f09da849e127e9f123241946c23c1e30f45a88366879e064211815/mistune-3.1.3-py3-none-any.whl", hash = "sha256:1a32314113cff28aa6432e99e522677c8587fd83e3d51c29b82a52409c842bd9", size = 53410 }, -] - -[[package]] -name = "modelscope" -version = "1.30.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "filelock" }, - { name = "requests" }, - { name = "setuptools" }, - { name = "tqdm" }, - { name = "urllib3" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/cc/cb/2c8a53914c17c2de204ea45b6ece2e54dd5afb6973cc3b3a295a05bdbaf3/modelscope-1.30.0.tar.gz", hash = "sha256:1ebedc419881852a784890330e3b6638ab934fa57f82ffa7e2b81b191fd5f76e", size = 4439412 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/8b/2b/3f7efc0538766ecfc082e47d8358abdfad764397665e8576e0c00edaf0c1/modelscope-1.30.0-py3-none-any.whl", hash = "sha256:a391c63a20f286e50010498ec3cc2aa1a86210a9dd74c55cfa4609fd2f8a999f", size = 5922191 }, -] - -[[package]] -name = "more-itertools" -version = "10.7.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/ce/a0/834b0cebabbfc7e311f30b46c8188790a37f89fc8d756660346fe5abfd09/more_itertools-10.7.0.tar.gz", hash = "sha256:9fddd5403be01a94b204faadcff459ec3568cf110265d3c54323e1e866ad29d3", size = 127671 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/2b/9f/7ba6f94fc1e9ac3d2b853fdff3035fb2fa5afbed898c4a72b8a020610594/more_itertools-10.7.0-py3-none-any.whl", hash = "sha256:d43980384673cb07d2f7d2d918c616b30c659c089ee23953f601d6609c67510e", size = 65278 }, -] - -[[package]] -name = "mpmath" -version = "1.3.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/e0/47/dd32fa426cc72114383ac549964eecb20ecfd886d1e5ccf5340b55b02f57/mpmath-1.3.0.tar.gz", hash = "sha256:7a28eb2a9774d00c7bc92411c19a89209d5da7c4c9a9e227be8330a23a25b91f", size = 508106 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/43/e3/7d92a15f894aa0c9c4b49b8ee9ac9850d6e63b03c9c32c0367a13ae62209/mpmath-1.3.0-py3-none-any.whl", hash = "sha256:a0b2b9fe80bbcd81a6647ff13108738cfb482d481d826cc0e02f5b35e5c88d2c", size = 536198 }, -] - -[[package]] -name = "msgpack" -version = "1.1.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/45/b1/ea4f68038a18c77c9467400d166d74c4ffa536f34761f7983a104357e614/msgpack-1.1.1.tar.gz", hash = "sha256:77b79ce34a2bdab2594f490c8e80dd62a02d650b91a75159a63ec413b8d104cd", size = 173555 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/33/52/f30da112c1dc92cf64f57d08a273ac771e7b29dea10b4b30369b2d7e8546/msgpack-1.1.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:353b6fc0c36fde68b661a12949d7d49f8f51ff5fa019c1e47c87c4ff34b080ed", size = 81799 }, - { url = "https://files.pythonhosted.org/packages/e4/35/7bfc0def2f04ab4145f7f108e3563f9b4abae4ab0ed78a61f350518cc4d2/msgpack-1.1.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:79c408fcf76a958491b4e3b103d1c417044544b68e96d06432a189b43d1215c8", size = 78278 }, - { url = "https://files.pythonhosted.org/packages/e8/c5/df5d6c1c39856bc55f800bf82778fd4c11370667f9b9e9d51b2f5da88f20/msgpack-1.1.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:78426096939c2c7482bf31ef15ca219a9e24460289c00dd0b94411040bb73ad2", size = 402805 }, - { url = "https://files.pythonhosted.org/packages/20/8e/0bb8c977efecfe6ea7116e2ed73a78a8d32a947f94d272586cf02a9757db/msgpack-1.1.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8b17ba27727a36cb73aabacaa44b13090feb88a01d012c0f4be70c00f75048b4", size = 408642 }, - { url = "https://files.pythonhosted.org/packages/59/a1/731d52c1aeec52006be6d1f8027c49fdc2cfc3ab7cbe7c28335b2910d7b6/msgpack-1.1.1-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:7a17ac1ea6ec3c7687d70201cfda3b1e8061466f28f686c24f627cae4ea8efd0", size = 395143 }, - { url = "https://files.pythonhosted.org/packages/2b/92/b42911c52cda2ba67a6418ffa7d08969edf2e760b09015593c8a8a27a97d/msgpack-1.1.1-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:88d1e966c9235c1d4e2afac21ca83933ba59537e2e2727a999bf3f515ca2af26", size = 395986 }, - { url = "https://files.pythonhosted.org/packages/61/dc/8ae165337e70118d4dab651b8b562dd5066dd1e6dd57b038f32ebc3e2f07/msgpack-1.1.1-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:f6d58656842e1b2ddbe07f43f56b10a60f2ba5826164910968f5933e5178af75", size = 402682 }, - { url = "https://files.pythonhosted.org/packages/58/27/555851cb98dcbd6ce041df1eacb25ac30646575e9cd125681aa2f4b1b6f1/msgpack-1.1.1-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:96decdfc4adcbc087f5ea7ebdcfd3dee9a13358cae6e81d54be962efc38f6338", size = 406368 }, - { url = "https://files.pythonhosted.org/packages/d4/64/39a26add4ce16f24e99eabb9005e44c663db00e3fce17d4ae1ae9d61df99/msgpack-1.1.1-cp310-cp310-win32.whl", hash = "sha256:6640fd979ca9a212e4bcdf6eb74051ade2c690b862b679bfcb60ae46e6dc4bfd", size = 65004 }, - { url = "https://files.pythonhosted.org/packages/7d/18/73dfa3e9d5d7450d39debde5b0d848139f7de23bd637a4506e36c9800fd6/msgpack-1.1.1-cp310-cp310-win_amd64.whl", hash = "sha256:8b65b53204fe1bd037c40c4148d00ef918eb2108d24c9aaa20bc31f9810ce0a8", size = 71548 }, -] - -[[package]] -name = "multidict" -version = "6.6.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "typing-extensions", marker = "python_full_version == '3.10.12'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/03/5d/d72502cd6dd64b0c5a5117b1701f05c38e94ffb4a1b4ab65ff0cd9b974e8/multidict-6.6.2.tar.gz", hash = "sha256:c1e8b8b0523c0361a78ce9b99d9850c51cf25e1fa3c5686030ce75df6fdf2918", size = 100939 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/1d/9d/9bcb4da29ff4e5a5dd7dccefaf49de8acae5b027e1a8b53296ac61ba14ab/multidict-6.6.2-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:cfd9c74d337e710d7ee26e72a7dbedbd60e0c58d3df7c5ccbb748857e977783c", size = 76829 }, - { url = "https://files.pythonhosted.org/packages/8d/40/4ca4b3eb34d4b57bb0a7385ca206fc28bc62aeb99daee47e72463efcdfc6/multidict-6.6.2-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:9d2c5867a1bd182041a950e9ec3dd3622926260434655bd5d94a62d889100787", size = 44799 }, - { url = "https://files.pythonhosted.org/packages/f5/d7/30ef84053dcb9f4a3ae9b0057b89da3236683ece29ded9b489793addf660/multidict-6.6.2-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:bc8551dd0000ce3f1d909906415ec18970fedb78e685dcac3a0b331a3422d810", size = 44476 }, - { url = "https://files.pythonhosted.org/packages/dd/3a/4cc34184902534abd2f82d9cfd159a333fd56602aa4de4644aaa441f3e6b/multidict-6.6.2-cp310-cp310-manylinux1_i686.manylinux2014_i686.manylinux_2_17_i686.manylinux_2_5_i686.whl", hash = "sha256:9a23d9360f656c316518c8534685ca7c9f18877f782c11bcfce97ff1012ba256", size = 225204 }, - { url = "https://files.pythonhosted.org/packages/f0/20/fb362a4b56a050c10480a81d4d04ce461e01b2f8d02f1e41d2367849670d/multidict-6.6.2-cp310-cp310-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:37fe64cfc6f73fce34f2ef9e099efb8333650b85b50929ba37789311283f476f", size = 244463 }, - { url = "https://files.pythonhosted.org/packages/4e/a0/84aec746dc7e5db95da3c5777aafd8b78ff91a66b3e7d55bcea783d5b3e3/multidict-6.6.2-cp310-cp310-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:2429b25566ff8c12cdf472ee82084ea96ea085675822d6d85aee85efd1d36cc0", size = 221250 }, - { url = "https://files.pythonhosted.org/packages/d9/9b/549656e890c5134b666928fd56d99b7d7eb1579ab62e821a0a3a07d20738/multidict-6.6.2-cp310-cp310-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:66c596bd9bc833bad98445539ad53165b214c2c87bf386dbb819fabd1acdb462", size = 255154 }, - { url = "https://files.pythonhosted.org/packages/c6/de/8ca2242eda642d264a6e6f43a8c1ad9fee5e5aff15b39db8b00afaba5971/multidict-6.6.2-cp310-cp310-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:eb8c006b565a0e53b298e9d48ef5aafe343f77de65c4fa7adb3d3b752a22d10b", size = 251359 }, - { url = "https://files.pythonhosted.org/packages/e5/d7/34c3743e2dce6777c45dff9f451297b0e9a64e145ba3b59c6d5a8834a245/multidict-6.6.2-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:7d09a7ade505d4556aa00c18f5635c9e7fe5973b98fee4e034162b02e48da7bc", size = 242695 }, - { url = "https://files.pythonhosted.org/packages/33/ab/20d63595785766d1d0aac9850b972b9ff20d533371a9140d499904dc7ace/multidict-6.6.2-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:6c95573274630213748ecee465410d4e5e44532d97ba9b09481968efd3c1fd2c", size = 240935 }, - { url = "https://files.pythonhosted.org/packages/6c/1e/a7c9b9a756ad45f2c9750471750345eb8ed8b7a921f742cec30fa70a4070/multidict-6.6.2-cp310-cp310-musllinux_1_2_armv7l.whl", hash = "sha256:e45ebeeee1ce0f9a68151cee1afe02eef56f3b6977a580873c179175e5108275", size = 233923 }, - { url = "https://files.pythonhosted.org/packages/44/c6/bb6e4cca146748e2b787d9338009e8c845af48808600b0769215b6a1fea7/multidict-6.6.2-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:329ecbdd77402648ebcb077b342ad6e67396dcf377c67418a733e88476ff3a11", size = 241715 }, - { url = "https://files.pythonhosted.org/packages/bf/24/d3c01293132168f6a29b2a5490ce4a07d34f0bdb5d73a4b2a177623b88bb/multidict-6.6.2-cp310-cp310-musllinux_1_2_ppc64le.whl", hash = "sha256:f449699e273579a7eda79e36a8b7a6aae06a601d115c54e1aeebf08e07ea3ea1", size = 251433 }, - { url = "https://files.pythonhosted.org/packages/0f/c1/dd47ff9571905e722ce9d71161d21bb970d9632224fa7bfdfe4ae59073c3/multidict-6.6.2-cp310-cp310-musllinux_1_2_s390x.whl", hash = "sha256:ed4bb96a59976e4da7e1fbe3a7c37bcb4a16f3b20c5bba8af9a0ce459e14039a", size = 243316 }, - { url = "https://files.pythonhosted.org/packages/1b/51/73906c1101792b8c6232ecbbbb2fe943a01d820b502a3e882b3ed986bad6/multidict-6.6.2-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:4d05496c2779af4698ba8a841b226247a9a515210eff3a029f48d5345255b1d3", size = 238648 }, - { url = "https://files.pythonhosted.org/packages/a8/8d/0174c5602a531da3c1c0e6e8497bd98702ad1793ff3a9988628de8d75a41/multidict-6.6.2-cp310-cp310-win32.whl", hash = "sha256:f96af5fbf6bab448d6dab34e8126f32f86de65034539d4a7077193f7b64a08f6", size = 41362 }, - { url = "https://files.pythonhosted.org/packages/b1/58/40b720fd0a9ba2f492497c27c7d047606b20540be64a4315693054bd1d09/multidict-6.6.2-cp310-cp310-win_amd64.whl", hash = "sha256:44468089034383be86735f64f5d7daa6a1297e338b739403871a63750b95866d", size = 45892 }, - { url = "https://files.pythonhosted.org/packages/72/53/1ab0ca0093516836b3e89aa9a9e7247f06109300a24b7d9fa3c983122394/multidict-6.6.2-cp310-cp310-win_arm64.whl", hash = "sha256:4e36b00dfb630a81f8efd4eb8a67b5b45f0918da3f2c8c4c14d16fc12b682d33", size = 42983 }, - { url = "https://files.pythonhosted.org/packages/0c/30/7b7d121f76ea3ea7561814531e5cc19e75e9b6646818491179c2c875b591/multidict-6.6.2-py3-none-any.whl", hash = "sha256:a7d14275ff2f85a8ff3c2a32e30f94b9fc8a2125b59a4ecc32271a347fad6e78", size = 12312 }, -] - -[[package]] -name = "multiprocess" -version = "0.70.16" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "dill" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/b5/ae/04f39c5d0d0def03247c2893d6f2b83c136bf3320a2154d7b8858f2ba72d/multiprocess-0.70.16.tar.gz", hash = "sha256:161af703d4652a0e1410be6abccecde4a7ddffd19341be0a7011b94aeb171ac1", size = 1772603 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/ef/76/6e712a2623d146d314f17598df5de7224c85c0060ef63fd95cc15a25b3fa/multiprocess-0.70.16-pp310-pypy310_pp73-macosx_10_13_x86_64.whl", hash = "sha256:476887be10e2f59ff183c006af746cb6f1fd0eadcfd4ef49e605cbe2659920ee", size = 134980 }, - { url = "https://files.pythonhosted.org/packages/0f/ab/1e6e8009e380e22254ff539ebe117861e5bdb3bff1fc977920972237c6c7/multiprocess-0.70.16-pp310-pypy310_pp73-manylinux_2_28_x86_64.whl", hash = "sha256:d951bed82c8f73929ac82c61f01a7b5ce8f3e5ef40f5b52553b4f547ce2b08ec", size = 134982 }, - { url = "https://files.pythonhosted.org/packages/bc/f7/7ec7fddc92e50714ea3745631f79bd9c96424cb2702632521028e57d3a36/multiprocess-0.70.16-py310-none-any.whl", hash = "sha256:c4a9944c67bd49f823687463660a2d6daae94c289adff97e0f9d696ba6371d02", size = 134824 }, - { url = "https://files.pythonhosted.org/packages/ea/89/38df130f2c799090c978b366cfdf5b96d08de5b29a4a293df7f7429fa50b/multiprocess-0.70.16-py38-none-any.whl", hash = "sha256:a71d82033454891091a226dfc319d0cfa8019a4e888ef9ca910372a446de4435", size = 132628 }, - { url = "https://files.pythonhosted.org/packages/da/d9/f7f9379981e39b8c2511c9e0326d212accacb82f12fbfdc1aa2ce2a7b2b6/multiprocess-0.70.16-py39-none-any.whl", hash = "sha256:a0bafd3ae1b732eac64be2e72038231c1ba97724b60b09400d68f229fcc2fbf3", size = 133351 }, -] - -[[package]] -name = "nbclient" -version = "0.10.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "jupyter-client" }, - { name = "jupyter-core" }, - { name = "nbformat" }, - { name = "traitlets" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/87/66/7ffd18d58eae90d5721f9f39212327695b749e23ad44b3881744eaf4d9e8/nbclient-0.10.2.tar.gz", hash = "sha256:90b7fc6b810630db87a6d0c2250b1f0ab4cf4d3c27a299b0cde78a4ed3fd9193", size = 62424 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/34/6d/e7fa07f03a4a7b221d94b4d586edb754a9b0dc3c9e2c93353e9fa4e0d117/nbclient-0.10.2-py3-none-any.whl", hash = "sha256:4ffee11e788b4a27fabeb7955547e4318a5298f34342a4bfd01f2e1faaeadc3d", size = 25434 }, -] - -[[package]] -name = "nbconvert" -version = "7.16.6" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "beautifulsoup4" }, - { name = "bleach", extra = ["css"] }, - { name = "defusedxml" }, - { name = "jinja2" }, - { name = "jupyter-core" }, - { name = "jupyterlab-pygments" }, - { name = "markupsafe" }, - { name = "mistune" }, - { name = "nbclient" }, - { name = "nbformat" }, - { name = "packaging" }, - { name = "pandocfilters" }, - { name = "pygments" }, - { name = "traitlets" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/a3/59/f28e15fc47ffb73af68a8d9b47367a8630d76e97ae85ad18271b9db96fdf/nbconvert-7.16.6.tar.gz", hash = "sha256:576a7e37c6480da7b8465eefa66c17844243816ce1ccc372633c6b71c3c0f582", size = 857715 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/cc/9a/cd673b2f773a12c992f41309ef81b99da1690426bd2f96957a7ade0d3ed7/nbconvert-7.16.6-py3-none-any.whl", hash = "sha256:1375a7b67e0c2883678c48e506dc320febb57685e5ee67faa51b18a90f3a712b", size = 258525 }, -] - -[[package]] -name = "nbformat" -version = "5.10.4" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "fastjsonschema" }, - { name = "jsonschema" }, - { name = "jupyter-core" }, - { name = "traitlets" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/6d/fd/91545e604bc3dad7dca9ed03284086039b294c6b3d75c0d2fa45f9e9caf3/nbformat-5.10.4.tar.gz", hash = "sha256:322168b14f937a5d11362988ecac2a4952d3d8e3a2cbeb2319584631226d5b3a", size = 142749 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/a9/82/0340caa499416c78e5d8f5f05947ae4bc3cba53c9f038ab6e9ed964e22f1/nbformat-5.10.4-py3-none-any.whl", hash = "sha256:3b48d6c8fbca4b299bf3982ea7db1af21580e4fec269ad087b9e81588891200b", size = 78454 }, -] - -[[package]] -name = "nemo-toolkit" -version = "2.3.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "fsspec" }, - { name = "huggingface-hub" }, - { name = "numba" }, - { name = "numpy" }, - { name = "onnx" }, - { name = "protobuf" }, - { name = "python-dateutil" }, - { name = "ruamel-yaml" }, - { name = "scikit-learn" }, - { name = "setuptools" }, - { name = "tensorboard" }, - { name = "text-unidecode" }, - { name = "torch" }, - { name = "tqdm" }, - { name = "wget" }, - { name = "wrapt" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/b1/9e/534533c7b80578b47e4d59f095f8e2b782afcbf8bb3f090ea882e42201ea/nemo_toolkit-2.3.1.tar.gz", hash = "sha256:8ac1444651fd36a315efad10b2837a91cba9b79f63c00da211454b451e627853", size = 4364579 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/b2/b4/821c0253875f9cd206ff9c5c2f0245b42947e84a201c7c3d9118511b737e/nemo_toolkit-2.3.1-py3-none-any.whl", hash = "sha256:2ecbc805ab368123fe047dd712ee5a6494c2a14eb00bb54481f5c12fcafd077a", size = 5965438 }, -] - -[[package]] -name = "nest-asyncio" -version = "1.6.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/83/f8/51569ac65d696c8ecbee95938f89d4abf00f47d58d48f6fbabfe8f0baefe/nest_asyncio-1.6.0.tar.gz", hash = "sha256:6f172d5449aca15afd6c646851f4e31e02c598d553a667e38cafa997cfec55fe", size = 7418 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/a0/c4/c2971a3ba4c6103a3d10c4b0f24f461ddc027f0f09763220cf35ca1401b3/nest_asyncio-1.6.0-py3-none-any.whl", hash = "sha256:87af6efd6b5e897c81050477ef65c62e2b2f35d51703cae01aff2905b1852e1c", size = 5195 }, -] - -[[package]] -name = "networkx" -version = "3.4.2" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/fd/1d/06475e1cd5264c0b870ea2cc6fdb3e37177c1e565c43f56ff17a10e3937f/networkx-3.4.2.tar.gz", hash = "sha256:307c3669428c5362aab27c8a1260aa8f47c4e91d3891f48be0141738d8d053e1", size = 2151368 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/b9/54/dd730b32ea14ea797530a4479b2ed46a6fb250f682a9cfb997e968bf0261/networkx-3.4.2-py3-none-any.whl", hash = "sha256:df5d4365b724cf81b8c6a7312509d0c22386097011ad1abe274afd5e9d3bbc5f", size = 1723263 }, -] - -[[package]] -name = "nltk" -version = "3.9.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "click" }, - { name = "joblib" }, - { name = "regex" }, - { name = "tqdm" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/3c/87/db8be88ad32c2d042420b6fd9ffd4a149f9a0d7f0e86b3f543be2eeeedd2/nltk-3.9.1.tar.gz", hash = "sha256:87d127bd3de4bd89a4f81265e5fa59cb1b199b27440175370f7417d2bc7ae868", size = 2904691 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/4d/66/7d9e26593edda06e8cb531874633f7c2372279c3b0f46235539fe546df8b/nltk-3.9.1-py3-none-any.whl", hash = "sha256:4fa26829c5b00715afe3061398a8989dc643b92ce7dd93fb4585a70930d168a1", size = 1505442 }, -] - -[[package]] -name = "notebook" -version = "7.4.3" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "jupyter-server" }, - { name = "jupyterlab" }, - { name = "jupyterlab-server" }, - { name = "notebook-shim" }, - { name = "tornado" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/dc/21/4f83b15e483da4f4f63928edd0cb08b6e7d33f8a15c23b116a90c44c6235/notebook-7.4.3.tar.gz", hash = "sha256:a1567481cd3853f2610ee0ecf5dfa12bb508e878ee8f92152c134ef7f0568a76", size = 13881668 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/76/1b/16c809d799e3ddd7a97c8b43734f79624b74ddef9707e7d92275a13777bc/notebook-7.4.3-py3-none-any.whl", hash = "sha256:9cdeee954e04101cadb195d90e2ab62b7c9286c1d4f858bf3bb54e40df16c0c3", size = 14286402 }, -] - -[[package]] -name = "notebook-shim" -version = "0.2.4" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "jupyter-server" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/54/d2/92fa3243712b9a3e8bafaf60aac366da1cada3639ca767ff4b5b3654ec28/notebook_shim-0.2.4.tar.gz", hash = "sha256:b4b2cfa1b65d98307ca24361f5b30fe785b53c3fd07b7a47e89acb5e6ac638cb", size = 13167 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/f9/33/bd5b9137445ea4b680023eb0469b2bb969d61303dedb2aac6560ff3d14a1/notebook_shim-0.2.4-py3-none-any.whl", hash = "sha256:411a5be4e9dc882a074ccbcae671eda64cceb068767e9a3419096986560e1cef", size = 13307 }, -] - -[[package]] -name = "num2words" -version = "0.5.14" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "docopt" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/f6/58/ad645bd38b4b648eb2fc2ba1b909398e54eb0cbb6a7dbd2b4953e38c9621/num2words-0.5.14.tar.gz", hash = "sha256:b066ec18e56b6616a3b38086b5747daafbaa8868b226a36127e0451c0cf379c6", size = 218213 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/d6/5b/545e9267a1cc080c8a1be2746113a063e34bcdd0f5173fd665a5c13cb234/num2words-0.5.14-py3-none-any.whl", hash = "sha256:1c8e5b00142fc2966fd8d685001e36c4a9911e070d1b120e1beb721fa1edb33d", size = 163525 }, -] - -[[package]] -name = "numba" -version = "0.61.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "llvmlite" }, - { name = "numpy" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/3c/88/c13a935f200fda51384411e49840a8e7f70c9cb1ee8d809dd0f2477cf7ef/numba-0.61.0.tar.gz", hash = "sha256:888d2e89b8160899e19591467e8fdd4970e07606e1fbc248f239c89818d5f925", size = 2816484 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/77/97/8568a025b9ab8b4d53491e70d4206d5f3fc71fbe94f3097058e01ad8e7ff/numba-0.61.0-cp310-cp310-macosx_10_14_x86_64.whl", hash = "sha256:9cab9783a700fa428b1a54d65295122bc03b3de1d01fb819a6b9dbbddfdb8c43", size = 2769008 }, - { url = "https://files.pythonhosted.org/packages/8c/ab/a88c20755f66543ee01c85c98b866595b92e1bd0ed80565a4889e22929a8/numba-0.61.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:46c5ae094fb3706f5adf9021bfb7fc11e44818d61afee695cdee4eadfed45e98", size = 2771815 }, - { url = "https://files.pythonhosted.org/packages/ae/f4/b357913089ecec1a9ddc6adc04090396928f36a484a5ab9e71b24ddba4cd/numba-0.61.0-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:6fb74e81aa78a2303e30593d8331327dfc0d2522b5db05ac967556a26db3ef87", size = 3820233 }, - { url = "https://files.pythonhosted.org/packages/ea/60/0e21bcf3baaf10e39d48cd224618e46a6b75d3394f465c37ce57bf98cbfa/numba-0.61.0-cp310-cp310-manylinux_2_28_aarch64.whl", hash = "sha256:0ebbd4827091384ab8c4615ba1b3ca8bc639a3a000157d9c37ba85d34cd0da1b", size = 3514707 }, - { url = "https://files.pythonhosted.org/packages/a0/08/45c136ab59e6b11e61ce15a0d17ef03fd89eaccb0db05ad67912aaf5218a/numba-0.61.0-cp310-cp310-win_amd64.whl", hash = "sha256:43aa4d7d10c542d3c78106b8481e0cbaaec788c39ee8e3d7901682748ffdf0b4", size = 2827753 }, -] - -[[package]] -name = "numpy" -version = "1.26.4" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/65/6e/09db70a523a96d25e115e71cc56a6f9031e7b8cd166c1ac8438307c14058/numpy-1.26.4.tar.gz", hash = "sha256:2a02aba9ed12e4ac4eb3ea9421c420301a0c6460d9830d74a9df87efa4912010", size = 15786129 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/a7/94/ace0fdea5241a27d13543ee117cbc65868e82213fb31a8eb7fe9ff23f313/numpy-1.26.4-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:9ff0f4f29c51e2803569d7a51c2304de5554655a60c5d776e35b4a41413830d0", size = 20631468 }, - { url = "https://files.pythonhosted.org/packages/20/f7/b24208eba89f9d1b58c1668bc6c8c4fd472b20c45573cb767f59d49fb0f6/numpy-1.26.4-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:2e4ee3380d6de9c9ec04745830fd9e2eccb3e6cf790d39d7b98ffd19b0dd754a", size = 13966411 }, - { url = "https://files.pythonhosted.org/packages/fc/a5/4beee6488160798683eed5bdb7eead455892c3b4e1f78d79d8d3f3b084ac/numpy-1.26.4-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d209d8969599b27ad20994c8e41936ee0964e6da07478d6c35016bc386b66ad4", size = 14219016 }, - { url = "https://files.pythonhosted.org/packages/4b/d7/ecf66c1cd12dc28b4040b15ab4d17b773b87fa9d29ca16125de01adb36cd/numpy-1.26.4-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ffa75af20b44f8dba823498024771d5ac50620e6915abac414251bd971b4529f", size = 18240889 }, - { url = "https://files.pythonhosted.org/packages/24/03/6f229fe3187546435c4f6f89f6d26c129d4f5bed40552899fcf1f0bf9e50/numpy-1.26.4-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:62b8e4b1e28009ef2846b4c7852046736bab361f7aeadeb6a5b89ebec3c7055a", size = 13876746 }, - { url = "https://files.pythonhosted.org/packages/39/fe/39ada9b094f01f5a35486577c848fe274e374bbf8d8f472e1423a0bbd26d/numpy-1.26.4-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:a4abb4f9001ad2858e7ac189089c42178fcce737e4169dc61321660f1a96c7d2", size = 18078620 }, - { url = "https://files.pythonhosted.org/packages/d5/ef/6ad11d51197aad206a9ad2286dc1aac6a378059e06e8cf22cd08ed4f20dc/numpy-1.26.4-cp310-cp310-win32.whl", hash = "sha256:bfe25acf8b437eb2a8b2d49d443800a5f18508cd811fea3181723922a8a82b07", size = 5972659 }, - { url = "https://files.pythonhosted.org/packages/19/77/538f202862b9183f54108557bfda67e17603fc560c384559e769321c9d92/numpy-1.26.4-cp310-cp310-win_amd64.whl", hash = "sha256:b97fe8060236edf3662adfc2c633f56a08ae30560c56310562cb4f95500022d5", size = 15808905 }, -] - -[[package]] -name = "nvidia-cublas-cu12" -version = "12.6.4.1" -source = { registry = "https://pypi.org/simple" } -wheels = [ - { url = "https://files.pythonhosted.org/packages/af/eb/ff4b8c503fa1f1796679dce648854d58751982426e4e4b37d6fce49d259c/nvidia_cublas_cu12-12.6.4.1-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:08ed2686e9875d01b58e3cb379c6896df8e76c75e0d4a7f7dace3d7b6d9ef8eb", size = 393138322 }, - { url = "https://files.pythonhosted.org/packages/97/0d/f1f0cadbf69d5b9ef2e4f744c9466cb0a850741d08350736dfdb4aa89569/nvidia_cublas_cu12-12.6.4.1-py3-none-manylinux_2_27_aarch64.whl", hash = "sha256:235f728d6e2a409eddf1df58d5b0921cf80cfa9e72b9f2775ccb7b4a87984668", size = 390794615 }, -] - -[[package]] -name = "nvidia-cuda-cupti-cu12" -version = "12.6.80" -source = { registry = "https://pypi.org/simple" } -wheels = [ - { url = "https://files.pythonhosted.org/packages/e6/8b/2f6230cb715646c3a9425636e513227ce5c93c4d65823a734f4bb86d43c3/nvidia_cuda_cupti_cu12-12.6.80-py3-none-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:166ee35a3ff1587f2490364f90eeeb8da06cd867bd5b701bf7f9a02b78bc63fc", size = 8236764 }, - { url = "https://files.pythonhosted.org/packages/25/0f/acb326ac8fd26e13c799e0b4f3b2751543e1834f04d62e729485872198d4/nvidia_cuda_cupti_cu12-12.6.80-py3-none-manylinux2014_aarch64.whl", hash = "sha256:358b4a1d35370353d52e12f0a7d1769fc01ff74a191689d3870b2123156184c4", size = 8236756 }, - { url = "https://files.pythonhosted.org/packages/49/60/7b6497946d74bcf1de852a21824d63baad12cd417db4195fc1bfe59db953/nvidia_cuda_cupti_cu12-12.6.80-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:6768bad6cab4f19e8292125e5f1ac8aa7d1718704012a0e3272a6f61c4bce132", size = 8917980 }, - { url = "https://files.pythonhosted.org/packages/a5/24/120ee57b218d9952c379d1e026c4479c9ece9997a4fb46303611ee48f038/nvidia_cuda_cupti_cu12-12.6.80-py3-none-manylinux2014_x86_64.whl", hash = "sha256:a3eff6cdfcc6a4c35db968a06fcadb061cbc7d6dde548609a941ff8701b98b73", size = 8917972 }, -] - -[[package]] -name = "nvidia-cuda-nvrtc-cu12" -version = "12.6.77" -source = { registry = "https://pypi.org/simple" } -wheels = [ - { url = "https://files.pythonhosted.org/packages/f4/2f/72df534873235983cc0a5371c3661bebef7c4682760c275590b972c7b0f9/nvidia_cuda_nvrtc_cu12-12.6.77-py3-none-manylinux2014_aarch64.whl", hash = "sha256:5847f1d6e5b757f1d2b3991a01082a44aad6f10ab3c5c0213fa3e25bddc25a13", size = 23162955 }, - { url = "https://files.pythonhosted.org/packages/75/2e/46030320b5a80661e88039f59060d1790298b4718944a65a7f2aeda3d9e9/nvidia_cuda_nvrtc_cu12-12.6.77-py3-none-manylinux2014_x86_64.whl", hash = "sha256:35b0cc6ee3a9636d5409133e79273ce1f3fd087abb0532d2d2e8fff1fe9efc53", size = 23650380 }, -] - -[[package]] -name = "nvidia-cuda-runtime-cu12" -version = "12.6.77" -source = { registry = "https://pypi.org/simple" } -wheels = [ - { url = "https://files.pythonhosted.org/packages/8f/ea/590b2ac00d772a8abd1c387a92b46486d2679ca6622fd25c18ff76265663/nvidia_cuda_runtime_cu12-12.6.77-py3-none-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:6116fad3e049e04791c0256a9778c16237837c08b27ed8c8401e2e45de8d60cd", size = 908052 }, - { url = "https://files.pythonhosted.org/packages/b7/3d/159023799677126e20c8fd580cca09eeb28d5c5a624adc7f793b9aa8bbfa/nvidia_cuda_runtime_cu12-12.6.77-py3-none-manylinux2014_aarch64.whl", hash = "sha256:d461264ecb429c84c8879a7153499ddc7b19b5f8d84c204307491989a365588e", size = 908040 }, - { url = "https://files.pythonhosted.org/packages/e1/23/e717c5ac26d26cf39a27fbc076240fad2e3b817e5889d671b67f4f9f49c5/nvidia_cuda_runtime_cu12-12.6.77-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:ba3b56a4f896141e25e19ab287cd71e52a6a0f4b29d0d31609f60e3b4d5219b7", size = 897690 }, - { url = "https://files.pythonhosted.org/packages/f0/62/65c05e161eeddbafeca24dc461f47de550d9fa8a7e04eb213e32b55cfd99/nvidia_cuda_runtime_cu12-12.6.77-py3-none-manylinux2014_x86_64.whl", hash = "sha256:a84d15d5e1da416dd4774cb42edf5e954a3e60cc945698dc1d5be02321c44dc8", size = 897678 }, -] - -[[package]] -name = "nvidia-cudnn-cu12" -version = "9.5.1.17" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "nvidia-cublas-cu12" }, -] -wheels = [ - { url = "https://files.pythonhosted.org/packages/99/93/a201a12d3ec1caa8c6ac34c1c2f9eeb696b886f0c36ff23c638b46603bd0/nvidia_cudnn_cu12-9.5.1.17-py3-none-manylinux_2_28_aarch64.whl", hash = "sha256:9fd4584468533c61873e5fda8ca41bac3a38bcb2d12350830c69b0a96a7e4def", size = 570523509 }, - { url = "https://files.pythonhosted.org/packages/2a/78/4535c9c7f859a64781e43c969a3a7e84c54634e319a996d43ef32ce46f83/nvidia_cudnn_cu12-9.5.1.17-py3-none-manylinux_2_28_x86_64.whl", hash = "sha256:30ac3869f6db17d170e0e556dd6cc5eee02647abc31ca856634d5a40f82c15b2", size = 570988386 }, -] - -[[package]] -name = "nvidia-cufft-cu12" -version = "11.3.0.4" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "nvidia-nvjitlink-cu12" }, -] -wheels = [ - { url = "https://files.pythonhosted.org/packages/1f/37/c50d2b2f2c07e146776389e3080f4faf70bcc4fa6e19d65bb54ca174ebc3/nvidia_cufft_cu12-11.3.0.4-py3-none-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:d16079550df460376455cba121db6564089176d9bac9e4f360493ca4741b22a6", size = 200164144 }, - { url = "https://files.pythonhosted.org/packages/ce/f5/188566814b7339e893f8d210d3a5332352b1409815908dad6a363dcceac1/nvidia_cufft_cu12-11.3.0.4-py3-none-manylinux2014_aarch64.whl", hash = "sha256:8510990de9f96c803a051822618d42bf6cb8f069ff3f48d93a8486efdacb48fb", size = 200164135 }, - { url = "https://files.pythonhosted.org/packages/8f/16/73727675941ab8e6ffd86ca3a4b7b47065edcca7a997920b831f8147c99d/nvidia_cufft_cu12-11.3.0.4-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:ccba62eb9cef5559abd5e0d54ceed2d9934030f51163df018532142a8ec533e5", size = 200221632 }, - { url = "https://files.pythonhosted.org/packages/60/de/99ec247a07ea40c969d904fc14f3a356b3e2a704121675b75c366b694ee1/nvidia_cufft_cu12-11.3.0.4-py3-none-manylinux2014_x86_64.whl", hash = "sha256:768160ac89f6f7b459bee747e8d175dbf53619cfe74b2a5636264163138013ca", size = 200221622 }, -] - -[[package]] -name = "nvidia-cufile-cu12" -version = "1.11.1.6" -source = { registry = "https://pypi.org/simple" } -wheels = [ - { url = "https://files.pythonhosted.org/packages/b2/66/cc9876340ac68ae71b15c743ddb13f8b30d5244af344ec8322b449e35426/nvidia_cufile_cu12-1.11.1.6-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:cc23469d1c7e52ce6c1d55253273d32c565dd22068647f3aa59b3c6b005bf159", size = 1142103 }, - { url = "https://files.pythonhosted.org/packages/17/bf/cc834147263b929229ce4aadd62869f0b195e98569d4c28b23edc72b85d9/nvidia_cufile_cu12-1.11.1.6-py3-none-manylinux_2_27_aarch64.whl", hash = "sha256:8f57a0051dcf2543f6dc2b98a98cb2719c37d3cee1baba8965d57f3bbc90d4db", size = 1066155 }, -] - -[[package]] -name = "nvidia-curand-cu12" -version = "10.3.7.77" -source = { registry = "https://pypi.org/simple" } -wheels = [ - { url = "https://files.pythonhosted.org/packages/42/ac/36543605358a355632f1a6faa3e2d5dfb91eab1e4bc7d552040e0383c335/nvidia_curand_cu12-10.3.7.77-py3-none-manylinux2014_aarch64.whl", hash = "sha256:6e82df077060ea28e37f48a3ec442a8f47690c7499bff392a5938614b56c98d8", size = 56289881 }, - { url = "https://files.pythonhosted.org/packages/73/1b/44a01c4e70933637c93e6e1a8063d1e998b50213a6b65ac5a9169c47e98e/nvidia_curand_cu12-10.3.7.77-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:a42cd1344297f70b9e39a1e4f467a4e1c10f1da54ff7a85c12197f6c652c8bdf", size = 56279010 }, - { url = "https://files.pythonhosted.org/packages/4a/aa/2c7ff0b5ee02eaef890c0ce7d4f74bc30901871c5e45dee1ae6d0083cd80/nvidia_curand_cu12-10.3.7.77-py3-none-manylinux2014_x86_64.whl", hash = "sha256:99f1a32f1ac2bd134897fc7a203f779303261268a65762a623bf30cc9fe79117", size = 56279000 }, - { url = "https://files.pythonhosted.org/packages/a6/02/5362a9396f23f7de1dd8a64369e87c85ffff8216fc8194ace0fa45ba27a5/nvidia_curand_cu12-10.3.7.77-py3-none-manylinux_2_27_aarch64.whl", hash = "sha256:7b2ed8e95595c3591d984ea3603dd66fe6ce6812b886d59049988a712ed06b6e", size = 56289882 }, -] - -[[package]] -name = "nvidia-cusolver-cu12" -version = "11.7.1.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "nvidia-cublas-cu12" }, - { name = "nvidia-cusparse-cu12" }, - { name = "nvidia-nvjitlink-cu12" }, -] -wheels = [ - { url = "https://files.pythonhosted.org/packages/93/17/dbe1aa865e4fdc7b6d4d0dd308fdd5aaab60f939abfc0ea1954eac4fb113/nvidia_cusolver_cu12-11.7.1.2-py3-none-manylinux2014_aarch64.whl", hash = "sha256:0ce237ef60acde1efc457335a2ddadfd7610b892d94efee7b776c64bb1cac9e0", size = 157833628 }, - { url = "https://files.pythonhosted.org/packages/f0/6e/c2cf12c9ff8b872e92b4a5740701e51ff17689c4d726fca91875b07f655d/nvidia_cusolver_cu12-11.7.1.2-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:e9e49843a7707e42022babb9bcfa33c29857a93b88020c4e4434656a655b698c", size = 158229790 }, - { url = "https://files.pythonhosted.org/packages/9f/81/baba53585da791d043c10084cf9553e074548408e04ae884cfe9193bd484/nvidia_cusolver_cu12-11.7.1.2-py3-none-manylinux2014_x86_64.whl", hash = "sha256:6cf28f17f64107a0c4d7802be5ff5537b2130bfc112f25d5a30df227058ca0e6", size = 158229780 }, - { url = "https://files.pythonhosted.org/packages/7c/5f/07d0ba3b7f19be5a5ec32a8679fc9384cfd9fc6c869825e93be9f28d6690/nvidia_cusolver_cu12-11.7.1.2-py3-none-manylinux_2_27_aarch64.whl", hash = "sha256:dbbe4fc38ec1289c7e5230e16248365e375c3673c9c8bac5796e2e20db07f56e", size = 157833630 }, -] - -[[package]] -name = "nvidia-cusparse-cu12" -version = "12.5.4.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "nvidia-nvjitlink-cu12" }, -] -wheels = [ - { url = "https://files.pythonhosted.org/packages/eb/eb/6681efd0aa7df96b4f8067b3ce7246833dd36830bb4cec8896182773db7d/nvidia_cusparse_cu12-12.5.4.2-py3-none-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:d25b62fb18751758fe3c93a4a08eff08effedfe4edf1c6bb5afd0890fe88f887", size = 216451147 }, - { url = "https://files.pythonhosted.org/packages/d3/56/3af21e43014eb40134dea004e8d0f1ef19d9596a39e4d497d5a7de01669f/nvidia_cusparse_cu12-12.5.4.2-py3-none-manylinux2014_aarch64.whl", hash = "sha256:7aa32fa5470cf754f72d1116c7cbc300b4e638d3ae5304cfa4a638a5b87161b1", size = 216451135 }, - { url = "https://files.pythonhosted.org/packages/06/1e/b8b7c2f4099a37b96af5c9bb158632ea9e5d9d27d7391d7eb8fc45236674/nvidia_cusparse_cu12-12.5.4.2-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:7556d9eca156e18184b94947ade0fba5bb47d69cec46bf8660fd2c71a4b48b73", size = 216561367 }, - { url = "https://files.pythonhosted.org/packages/43/ac/64c4316ba163e8217a99680c7605f779accffc6a4bcd0c778c12948d3707/nvidia_cusparse_cu12-12.5.4.2-py3-none-manylinux2014_x86_64.whl", hash = "sha256:23749a6571191a215cb74d1cdbff4a86e7b19f1200c071b3fcf844a5bea23a2f", size = 216561357 }, -] - -[[package]] -name = "nvidia-cusparselt-cu12" -version = "0.6.3" -source = { registry = "https://pypi.org/simple" } -wheels = [ - { url = "https://files.pythonhosted.org/packages/62/da/4de092c61c6dea1fc9c936e69308a02531d122e12f1f649825934ad651b5/nvidia_cusparselt_cu12-0.6.3-py3-none-manylinux2014_aarch64.whl", hash = "sha256:8371549623ba601a06322af2133c4a44350575f5a3108fb75f3ef20b822ad5f1", size = 156402859 }, - { url = "https://files.pythonhosted.org/packages/3b/9a/72ef35b399b0e183bc2e8f6f558036922d453c4d8237dab26c666a04244b/nvidia_cusparselt_cu12-0.6.3-py3-none-manylinux2014_x86_64.whl", hash = "sha256:e5c8a26c36445dd2e6812f1177978a24e2d37cacce7e090f297a688d1ec44f46", size = 156785796 }, -] - -[[package]] -name = "nvidia-nccl-cu12" -version = "2.26.2" -source = { registry = "https://pypi.org/simple" } -wheels = [ - { url = "https://files.pythonhosted.org/packages/69/5b/ca2f213f637305633814ae8c36b153220e40a07ea001966dcd87391f3acb/nvidia_nccl_cu12-2.26.2-py3-none-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:5c196e95e832ad30fbbb50381eb3cbd1fadd5675e587a548563993609af19522", size = 291671495 }, - { url = "https://files.pythonhosted.org/packages/67/ca/f42388aed0fddd64ade7493dbba36e1f534d4e6fdbdd355c6a90030ae028/nvidia_nccl_cu12-2.26.2-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:694cf3879a206553cc9d7dbda76b13efaf610fdb70a50cba303de1b0d1530ac6", size = 201319755 }, -] - -[[package]] -name = "nvidia-nvjitlink-cu12" -version = "12.6.85" -source = { registry = "https://pypi.org/simple" } -wheels = [ - { url = "https://files.pythonhosted.org/packages/9d/d7/c5383e47c7e9bf1c99d5bd2a8c935af2b6d705ad831a7ec5c97db4d82f4f/nvidia_nvjitlink_cu12-12.6.85-py3-none-manylinux2010_x86_64.manylinux_2_12_x86_64.whl", hash = "sha256:eedc36df9e88b682efe4309aa16b5b4e78c2407eac59e8c10a6a47535164369a", size = 19744971 }, - { url = "https://files.pythonhosted.org/packages/31/db/dc71113d441f208cdfe7ae10d4983884e13f464a6252450693365e166dcf/nvidia_nvjitlink_cu12-12.6.85-py3-none-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:cf4eaa7d4b6b543ffd69d6abfb11efdeb2db48270d94dfd3a452c24150829e41", size = 19270338 }, -] - -[[package]] -name = "nvidia-nvtx-cu12" -version = "12.6.77" -source = { registry = "https://pypi.org/simple" } -wheels = [ - { url = "https://files.pythonhosted.org/packages/b9/93/80f8a520375af9d7ee44571a6544653a176e53c2b8ccce85b97b83c2491b/nvidia_nvtx_cu12-12.6.77-py3-none-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:f44f8d86bb7d5629988d61c8d3ae61dddb2015dee142740536bc7481b022fe4b", size = 90549 }, - { url = "https://files.pythonhosted.org/packages/2b/53/36e2fd6c7068997169b49ffc8c12d5af5e5ff209df6e1a2c4d373b3a638f/nvidia_nvtx_cu12-12.6.77-py3-none-manylinux2014_aarch64.whl", hash = "sha256:adcaabb9d436c9761fca2b13959a2d237c5f9fd406c8e4b723c695409ff88059", size = 90539 }, - { url = "https://files.pythonhosted.org/packages/56/9a/fff8376f8e3d084cd1530e1ef7b879bb7d6d265620c95c1b322725c694f4/nvidia_nvtx_cu12-12.6.77-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:b90bed3df379fa79afbd21be8e04a0314336b8ae16768b58f2d34cb1d04cd7d2", size = 89276 }, - { url = "https://files.pythonhosted.org/packages/9e/4e/0d0c945463719429b7bd21dece907ad0bde437a2ff12b9b12fee94722ab0/nvidia_nvtx_cu12-12.6.77-py3-none-manylinux2014_x86_64.whl", hash = "sha256:6574241a3ec5fdc9334353ab8c479fe75841dbe8f4532a8fc97ce63503330ba1", size = 89265 }, -] - -[[package]] -name = "omegaconf" -version = "2.3.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "antlr4-python3-runtime" }, - { name = "pyyaml" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/09/48/6388f1bb9da707110532cb70ec4d2822858ddfb44f1cdf1233c20a80ea4b/omegaconf-2.3.0.tar.gz", hash = "sha256:d5d4b6d29955cc50ad50c46dc269bcd92c6e00f5f90d23ab5fee7bfca4ba4cc7", size = 3298120 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/e3/94/1843518e420fa3ed6919835845df698c7e27e183cb997394e4a670973a65/omegaconf-2.3.0-py3-none-any.whl", hash = "sha256:7b4df175cdb08ba400f45cae3bdcae7ba8365db4d165fc65fd04b050ab63b46b", size = 79500 }, -] - -[[package]] -name = "onnx" -version = "1.17.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "numpy" }, - { name = "protobuf" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/9a/54/0e385c26bf230d223810a9c7d06628d954008a5e5e4b73ee26ef02327282/onnx-1.17.0.tar.gz", hash = "sha256:48ca1a91ff73c1d5e3ea2eef20ae5d0e709bb8a2355ed798ffc2169753013fd3", size = 12165120 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/2e/29/57053ba7787788ac75efb095cfc1ae290436b6d3a26754693cd7ed1b4fac/onnx-1.17.0-cp310-cp310-macosx_12_0_universal2.whl", hash = "sha256:38b5df0eb22012198cdcee527cc5f917f09cce1f88a69248aaca22bd78a7f023", size = 16645616 }, - { url = "https://files.pythonhosted.org/packages/75/0d/831807a18db2a5e8f7813848c59272b904a4ef3939fe4d1288cbce9ea735/onnx-1.17.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d545335cb49d4d8c47cc803d3a805deb7ad5d9094dc67657d66e568610a36d7d", size = 15908420 }, - { url = "https://files.pythonhosted.org/packages/dd/5b/c4f95dbe652d14aeba9afaceb177e9ffc48ac3c03048dd3f872f26f07e34/onnx-1.17.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3193a3672fc60f1a18c0f4c93ac81b761bc72fd8a6c2035fa79ff5969f07713e", size = 16046244 }, - { url = "https://files.pythonhosted.org/packages/08/a9/c1f218085043dccc6311460239e253fa6957cf12ee4b0a56b82014938d0b/onnx-1.17.0-cp310-cp310-win32.whl", hash = "sha256:0141c2ce806c474b667b7e4499164227ef594584da432fd5613ec17c1855e311", size = 14423516 }, - { url = "https://files.pythonhosted.org/packages/0e/d3/d26ebf590a65686dde6b27fef32493026c5be9e42083340d947395f93405/onnx-1.17.0-cp310-cp310-win_amd64.whl", hash = "sha256:dfd777d95c158437fda6b34758f0877d15b89cbe9ff45affbedc519b35345cf9", size = 14528496 }, -] - -[[package]] -name = "optuna" -version = "4.4.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "alembic" }, - { name = "colorlog" }, - { name = "numpy" }, - { name = "packaging" }, - { name = "pyyaml" }, - { name = "sqlalchemy" }, - { name = "tqdm" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/a5/e0/b303190ae8032d12f320a24c42af04038bacb1f3b17ede354dd1044a5642/optuna-4.4.0.tar.gz", hash = "sha256:a9029f6a92a1d6c8494a94e45abd8057823b535c2570819072dbcdc06f1c1da4", size = 467708 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/5c/5e/068798a8c7087863e7772e9363a880ab13fe55a5a7ede8ec42fab8a1acbb/optuna-4.4.0-py3-none-any.whl", hash = "sha256:fad8d9c5d5af993ae1280d6ce140aecc031c514a44c3b639d8c8658a8b7920ea", size = 395949 }, -] - -[[package]] -name = "oss2" -version = "2.19.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "aliyun-python-sdk-core" }, - { name = "aliyun-python-sdk-kms" }, - { name = "crcmod" }, - { name = "pycryptodome" }, - { name = "requests" }, - { name = "six" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/df/b5/f2cb1950dda46ac2284d6c950489fdacd0e743c2d79a347924d3cc44b86f/oss2-2.19.1.tar.gz", hash = "sha256:a8ab9ee7eb99e88a7e1382edc6ea641d219d585a7e074e3776e9dec9473e59c1", size = 298845 } - -[[package]] -name = "overrides" -version = "7.7.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/36/86/b585f53236dec60aba864e050778b25045f857e17f6e5ea0ae95fe80edd2/overrides-7.7.0.tar.gz", hash = "sha256:55158fa3d93b98cc75299b1e67078ad9003ca27945c76162c1c0766d6f91820a", size = 22812 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/2c/ab/fc8290c6a4c722e5514d80f62b2dc4c4df1a68a41d1364e625c35990fcf3/overrides-7.7.0-py3-none-any.whl", hash = "sha256:c7ed9d062f78b8e4c1a7b70bd8796b35ead4d9f510227ef9c5dc7626c60d7e49", size = 17832 }, -] - -[[package]] -name = "packaging" -version = "24.2" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/d0/63/68dbb6eb2de9cb10ee4c9c14a0148804425e13c4fb20d61cce69f53106da/packaging-24.2.tar.gz", hash = "sha256:c228a6dc5e932d346bc5739379109d49e8853dd8223571c7c5b55260edc0b97f", size = 163950 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/88/ef/eb23f262cca3c0c4eb7ab1933c3b1f03d021f2c48f54763065b6f0e321be/packaging-24.2-py3-none-any.whl", hash = "sha256:09abb1bccd265c01f4a3aa3f7a7db064b36514d2cba19a2f694fe6150451a759", size = 65451 }, -] - -[[package]] -name = "pandas" -version = "2.3.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "numpy", marker = "python_full_version == '3.10.12'" }, - { name = "python-dateutil" }, - { name = "pytz" }, - { name = "tzdata" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/72/51/48f713c4c728d7c55ef7444ba5ea027c26998d96d1a40953b346438602fc/pandas-2.3.0.tar.gz", hash = "sha256:34600ab34ebf1131a7613a260a61dbe8b62c188ec0ea4c296da7c9a06b004133", size = 4484490 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/e2/2d/df6b98c736ba51b8eaa71229e8fcd91233a831ec00ab520e1e23090cc072/pandas-2.3.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:625466edd01d43b75b1883a64d859168e4556261a5035b32f9d743b67ef44634", size = 11527531 }, - { url = "https://files.pythonhosted.org/packages/77/1c/3f8c331d223f86ba1d0ed7d3ed7fcf1501c6f250882489cc820d2567ddbf/pandas-2.3.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:a6872d695c896f00df46b71648eea332279ef4077a409e2fe94220208b6bb675", size = 10774764 }, - { url = "https://files.pythonhosted.org/packages/1b/45/d2599400fad7fe06b849bd40b52c65684bc88fbe5f0a474d0513d057a377/pandas-2.3.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:f4dd97c19bd06bc557ad787a15b6489d2614ddaab5d104a0310eb314c724b2d2", size = 11711963 }, - { url = "https://files.pythonhosted.org/packages/66/f8/5508bc45e994e698dbc93607ee6b9b6eb67df978dc10ee2b09df80103d9e/pandas-2.3.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:034abd6f3db8b9880aaee98f4f5d4dbec7c4829938463ec046517220b2f8574e", size = 12349446 }, - { url = "https://files.pythonhosted.org/packages/f7/fc/17851e1b1ea0c8456ba90a2f514c35134dd56d981cf30ccdc501a0adeac4/pandas-2.3.0-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:23c2b2dc5213810208ca0b80b8666670eb4660bbfd9d45f58592cc4ddcfd62e1", size = 12920002 }, - { url = "https://files.pythonhosted.org/packages/a1/9b/8743be105989c81fa33f8e2a4e9822ac0ad4aaf812c00fee6bb09fc814f9/pandas-2.3.0-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:39ff73ec07be5e90330cc6ff5705c651ace83374189dcdcb46e6ff54b4a72cd6", size = 13651218 }, - { url = "https://files.pythonhosted.org/packages/26/fa/8eeb2353f6d40974a6a9fd4081ad1700e2386cf4264a8f28542fd10b3e38/pandas-2.3.0-cp310-cp310-win_amd64.whl", hash = "sha256:40cecc4ea5abd2921682b57532baea5588cc5f80f0231c624056b146887274d2", size = 11082485 }, -] - -[[package]] -name = "pandocfilters" -version = "1.5.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/70/6f/3dd4940bbe001c06a65f88e36bad298bc7a0de5036115639926b0c5c0458/pandocfilters-1.5.1.tar.gz", hash = "sha256:002b4a555ee4ebc03f8b66307e287fa492e4a77b4ea14d3f934328297bb4939e", size = 8454 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/ef/af/4fbc8cab944db5d21b7e2a5b8e9211a03a79852b1157e2c102fcc61ac440/pandocfilters-1.5.1-py2.py3-none-any.whl", hash = "sha256:93be382804a9cdb0a7267585f157e5d1731bbe5545a85b268d6f5fe6232de2bc", size = 8663 }, -] - -[[package]] -name = "parakeet-coreml" -version = "0.1.0" -source = { virtual = "." } -dependencies = [ - { name = "absl-py" }, - { name = "accelerate" }, - { name = "aiohappyeyeballs" }, - { name = "aiohttp" }, - { name = "aiosignal" }, - { name = "alembic" }, - { name = "annotated-types" }, - { name = "antlr4-python3-runtime" }, - { name = "anyio" }, - { name = "appnope" }, - { name = "argon2-cffi" }, - { name = "argon2-cffi-bindings" }, - { name = "arrow" }, - { name = "asttokens" }, - { name = "async-lru" }, - { name = "async-timeout" }, - { name = "attrs" }, - { name = "audioread" }, - { name = "babel" }, - { name = "backports-datetime-fromisoformat" }, - { name = "beautifulsoup4" }, - { name = "bleach" }, - { name = "braceexpand" }, - { name = "cattrs" }, - { name = "certifi" }, - { name = "cffi" }, - { name = "charset-normalizer" }, - { name = "click" }, - { name = "cloudpickle" }, - { name = "colorlog" }, - { name = "comm" }, - { name = "contourpy" }, - { name = "coremltools" }, - { name = "cycler" }, - { name = "cytoolz" }, - { name = "datasets" }, - { name = "debugpy" }, - { name = "decorator" }, - { name = "defusedxml" }, - { name = "dill" }, - { name = "distance" }, - { name = "docopt" }, - { name = "editdistance" }, - { name = "einops" }, - { name = "exceptiongroup" }, - { name = "executing" }, - { name = "fastjsonschema" }, - { name = "fiddle" }, - { name = "filelock" }, - { name = "fonttools" }, - { name = "fqdn" }, - { name = "frozenlist" }, - { name = "fsspec" }, - { name = "funasr" }, - { name = "future" }, - { name = "g2p-en" }, - { name = "gitdb" }, - { name = "gitpython" }, - { name = "graphviz" }, - { name = "grpcio" }, - { name = "h11" }, - { name = "hf-xet" }, - { name = "httpcore" }, - { name = "httpx" }, - { name = "huggingface-hub" }, - { name = "hydra-core" }, - { name = "idna" }, - { name = "inflect" }, - { name = "intervaltree" }, - { name = "ipykernel" }, - { name = "ipython" }, - { name = "ipywidgets" }, - { name = "isoduration" }, - { name = "jedi" }, - { name = "jinja2" }, - { name = "jiwer" }, - { name = "joblib" }, - { name = "json5" }, - { name = "jsonpointer" }, - { name = "jsonschema" }, - { name = "jsonschema-specifications" }, - { name = "jupyter" }, - { name = "jupyter-client" }, - { name = "jupyter-console" }, - { name = "jupyter-core" }, - { name = "jupyter-events" }, - { name = "jupyter-lsp" }, - { name = "jupyter-server" }, - { name = "jupyter-server-terminals" }, - { name = "jupyterlab" }, - { name = "jupyterlab-pygments" }, - { name = "jupyterlab-server" }, - { name = "jupyterlab-widgets" }, - { name = "kaldi-python-io" }, - { name = "kaldiio" }, - { name = "kiwisolver" }, - { name = "lazy-loader" }, - { name = "levenshtein" }, - { name = "lhotse" }, - { name = "libcst" }, - { name = "librosa" }, - { name = "lightning" }, - { name = "lightning-utilities" }, - { name = "lilcom" }, - { name = "llvmlite" }, - { name = "loguru" }, - { name = "mako" }, - { name = "markdown" }, - { name = "markdown-it-py" }, - { name = "markupsafe" }, - { name = "marshmallow" }, - { name = "matplotlib" }, - { name = "matplotlib-inline" }, - { name = "mdurl" }, - { name = "mediapy" }, - { name = "mistune" }, - { name = "more-itertools" }, - { name = "mpmath" }, - { name = "msgpack" }, - { name = "multidict" }, - { name = "multiprocess" }, - { name = "nbclient" }, - { name = "nbconvert" }, - { name = "nbformat" }, - { name = "nemo-toolkit" }, - { name = "nest-asyncio" }, - { name = "networkx" }, - { name = "nltk" }, - { name = "notebook" }, - { name = "notebook-shim" }, - { name = "num2words" }, - { name = "numba" }, - { name = "numpy" }, - { name = "omegaconf" }, - { name = "onnx" }, - { name = "optuna" }, - { name = "overrides" }, - { name = "packaging" }, - { name = "pandas" }, - { name = "pandocfilters" }, - { name = "parso" }, - { name = "peft" }, - { name = "pexpect" }, - { name = "pillow" }, - { name = "pip" }, - { name = "plac" }, - { name = "platformdirs" }, - { name = "pooch" }, - { name = "prometheus-client" }, - { name = "prompt-toolkit" }, - { name = "propcache" }, - { name = "psutil" }, - { name = "ptyprocess" }, - { name = "pure-eval" }, - { name = "pyaml" }, - { name = "pyannote-audio" }, - { name = "pyannote-core" }, - { name = "pyannote-database" }, - { name = "pyannote-metrics" }, - { name = "pyarrow" }, - { name = "pybind11" }, - { name = "pycparser" }, - { name = "pydantic" }, - { name = "pydantic-core" }, - { name = "pydub" }, - { name = "pygments" }, - { name = "pyloudnorm" }, - { name = "pyparsing" }, - { name = "python-dateutil" }, - { name = "python-json-logger" }, - { name = "pytorch-lightning" }, - { name = "pytz" }, - { name = "pyyaml" }, - { name = "pyzmq" }, - { name = "rapidfuzz" }, - { name = "referencing" }, - { name = "regex" }, - { name = "requests" }, - { name = "resampy" }, - { name = "rfc3339-validator" }, - { name = "rfc3986-validator" }, - { name = "rich" }, - { name = "rpds-py" }, - { name = "ruamel-yaml" }, - { name = "ruamel-yaml-clib" }, - { name = "sacremoses" }, - { name = "safetensors" }, - { name = "scikit-learn" }, - { name = "scipy" }, - { name = "seaborn" }, - { name = "send2trash" }, - { name = "sentencepiece" }, - { name = "sentry-sdk" }, - { name = "setproctitle" }, - { name = "shellingham" }, - { name = "six" }, - { name = "smmap" }, - { name = "sniffio" }, - { name = "sortedcontainers" }, - { name = "soundfile" }, - { name = "soupsieve" }, - { name = "sox" }, - { name = "soxr" }, - { name = "sqlalchemy" }, - { name = "stack-data" }, - { name = "tabulate" }, - { name = "tensorboard" }, - { name = "tensorboard-data-server" }, - { name = "termcolor" }, - { name = "terminado" }, - { name = "text-unidecode" }, - { name = "texterrors" }, - { name = "threadpoolctl" }, - { name = "tinycss2" }, - { name = "tokenizers" }, - { name = "tomli" }, - { name = "toolz" }, - { name = "torch" }, - { name = "torchmetrics" }, - { name = "tornado" }, - { name = "tqdm" }, - { name = "traitlets" }, - { name = "transformers" }, - { name = "typeguard" }, - { name = "typer" }, - { name = "types-python-dateutil" }, - { name = "typing-extensions" }, - { name = "typing-inspection" }, - { name = "tzdata" }, - { name = "uri-template" }, - { name = "urllib3" }, - { name = "wandb" }, - { name = "wcwidth" }, - { name = "webcolors" }, - { name = "webdataset" }, - { name = "webencodings" }, - { name = "websocket-client" }, - { name = "werkzeug" }, - { name = "wget" }, - { name = "widgetsnbextension" }, - { name = "wrapt" }, - { name = "xxhash" }, - { name = "yarl" }, -] - -[package.metadata] -requires-dist = [ - { name = "absl-py", specifier = "==2.3.0" }, - { name = "accelerate", specifier = "==1.8.1" }, - { name = "aiohappyeyeballs", specifier = "==2.6.1" }, - { name = "aiohttp", specifier = "==3.12.13" }, - { name = "aiosignal", specifier = "==1.3.2" }, - { name = "alembic", specifier = "==1.16.2" }, - { name = "annotated-types", specifier = "==0.7.0" }, - { name = "antlr4-python3-runtime", specifier = "==4.9.3" }, - { name = "anyio", specifier = "==4.9.0" }, - { name = "appnope", specifier = "==0.1.4" }, - { name = "argon2-cffi", specifier = "==25.1.0" }, - { name = "argon2-cffi-bindings", specifier = "==21.2.0" }, - { name = "arrow", specifier = "==1.3.0" }, - { name = "asttokens", specifier = "==3.0.0" }, - { name = "async-lru", specifier = "==2.0.5" }, - { name = "async-timeout", specifier = "==5.0.1" }, - { name = "attrs", specifier = "==25.3.0" }, - { name = "audioread", specifier = "==3.0.1" }, - { name = "babel", specifier = "==2.17.0" }, - { name = "backports-datetime-fromisoformat", specifier = "==2.0.3" }, - { name = "beautifulsoup4", specifier = "==4.13.4" }, - { name = "bleach", specifier = "==6.2.0" }, - { name = "braceexpand", specifier = "==0.1.7" }, - { name = "cattrs", specifier = "==25.1.1" }, - { name = "certifi", specifier = "==2025.6.15" }, - { name = "cffi", specifier = "==1.17.1" }, - { name = "charset-normalizer", specifier = "==3.4.2" }, - { name = "click", specifier = "==8.2.1" }, - { name = "cloudpickle", specifier = "==3.1.1" }, - { name = "colorlog", specifier = "==6.9.0" }, - { name = "comm", specifier = "==0.2.2" }, - { name = "contourpy", specifier = "==1.3.2" }, - { name = "coremltools", specifier = "==9.0b1" }, - { name = "cycler", specifier = "==0.12.1" }, - { name = "cytoolz", specifier = "==1.0.1" }, - { name = "datasets", specifier = "==3.6.0" }, - { name = "debugpy", specifier = "==1.8.14" }, - { name = "decorator", specifier = "==5.2.1" }, - { name = "defusedxml", specifier = "==0.7.1" }, - { name = "dill", specifier = "==0.3.8" }, - { name = "distance", specifier = "==0.1.3" }, - { name = "docopt", specifier = "==0.6.2" }, - { name = "editdistance", specifier = "==0.8.1" }, - { name = "einops", specifier = "==0.8.1" }, - { name = "exceptiongroup", specifier = "==1.3.0" }, - { name = "executing", specifier = "==2.2.0" }, - { name = "fastjsonschema", specifier = "==2.21.1" }, - { name = "fiddle", specifier = "==0.3.0" }, - { name = "filelock", specifier = "==3.18.0" }, - { name = "fonttools", specifier = "==4.58.4" }, - { name = "fqdn", specifier = "==1.5.1" }, - { name = "frozenlist", specifier = "==1.7.0" }, - { name = "fsspec", specifier = "==2024.12.0" }, - { name = "funasr", specifier = ">=1.2.6" }, - { name = "future", specifier = "==1.0.0" }, - { name = "g2p-en", specifier = "==2.1.0" }, - { name = "gitdb", specifier = "==4.0.12" }, - { name = "gitpython", specifier = "==3.1.44" }, - { name = "graphviz", specifier = "==0.21" }, - { name = "grpcio", specifier = "==1.73.1" }, - { name = "h11", specifier = "==0.16.0" }, - { name = "hf-xet", specifier = "==1.1.5" }, - { name = "httpcore", specifier = "==1.0.9" }, - { name = "httpx", specifier = "==0.28.1" }, - { name = "huggingface-hub", specifier = "==0.33.1" }, - { name = "hydra-core", specifier = "==1.3.2" }, - { name = "idna", specifier = "==3.10" }, - { name = "inflect", specifier = "==7.5.0" }, - { name = "intervaltree", specifier = "==3.1.0" }, - { name = "ipykernel", specifier = "==6.29.5" }, - { name = "ipython", specifier = "==8.37.0" }, - { name = "ipywidgets", specifier = "==8.1.7" }, - { name = "isoduration", specifier = "==20.11.0" }, - { name = "jedi", specifier = "==0.19.2" }, - { name = "jinja2", specifier = "==3.1.6" }, - { name = "jiwer", specifier = "==4.0.0" }, - { name = "joblib", specifier = "==1.5.1" }, - { name = "json5", specifier = "==0.12.0" }, - { name = "jsonpointer", specifier = "==3.0.0" }, - { name = "jsonschema", specifier = "==4.24.0" }, - { name = "jsonschema-specifications", specifier = "==2025.4.1" }, - { name = "jupyter", specifier = "==1.1.1" }, - { name = "jupyter-client", specifier = "==8.6.3" }, - { name = "jupyter-console", specifier = "==6.6.3" }, - { name = "jupyter-core", specifier = "==5.8.1" }, - { name = "jupyter-events", specifier = "==0.12.0" }, - { name = "jupyter-lsp", specifier = "==2.2.5" }, - { name = "jupyter-server", specifier = "==2.16.0" }, - { name = "jupyter-server-terminals", specifier = "==0.5.3" }, - { name = "jupyterlab", specifier = "==4.4.4" }, - { name = "jupyterlab-pygments", specifier = "==0.3.0" }, - { name = "jupyterlab-server", specifier = "==2.27.3" }, - { name = "jupyterlab-widgets", specifier = "==3.0.15" }, - { name = "kaldi-python-io", specifier = "==1.2.2" }, - { name = "kaldiio", specifier = "==2.18.1" }, - { name = "kiwisolver", specifier = "==1.4.8" }, - { name = "lazy-loader", specifier = "==0.4" }, - { name = "levenshtein", specifier = "==0.27.1" }, - { name = "lhotse", specifier = "==1.30.3" }, - { name = "libcst", specifier = "==1.8.2" }, - { name = "librosa", specifier = "==0.11.0" }, - { name = "lightning", specifier = "==2.4.0" }, - { name = "lightning-utilities", specifier = "==0.14.3" }, - { name = "lilcom", specifier = "==1.8.1" }, - { name = "llvmlite", specifier = "==0.44.0" }, - { name = "loguru", specifier = "==0.7.3" }, - { name = "mako", specifier = "==1.3.10" }, - { name = "markdown", specifier = "==3.8.2" }, - { name = "markdown-it-py", specifier = "==3.0.0" }, - { name = "markupsafe", specifier = "==3.0.2" }, - { name = "marshmallow", specifier = "==4.0.0" }, - { name = "matplotlib", specifier = "==3.10.3" }, - { name = "matplotlib-inline", specifier = "==0.1.7" }, - { name = "mdurl", specifier = "==0.1.2" }, - { name = "mediapy", specifier = "==1.1.6" }, - { name = "mistune", specifier = "==3.1.3" }, - { name = "more-itertools", specifier = "==10.7.0" }, - { name = "mpmath", specifier = "==1.3.0" }, - { name = "msgpack", specifier = "==1.1.1" }, - { name = "multidict", specifier = "==6.6.2" }, - { name = "multiprocess", specifier = "==0.70.16" }, - { name = "nbclient", specifier = "==0.10.2" }, - { name = "nbconvert", specifier = "==7.16.6" }, - { name = "nbformat", specifier = "==5.10.4" }, - { name = "nemo-toolkit", specifier = "==2.3.1" }, - { name = "nest-asyncio", specifier = "==1.6.0" }, - { name = "networkx", specifier = "==3.4.2" }, - { name = "nltk", specifier = "==3.9.1" }, - { name = "notebook", specifier = "==7.4.3" }, - { name = "notebook-shim", specifier = "==0.2.4" }, - { name = "num2words", specifier = "==0.5.14" }, - { name = "numba", specifier = "==0.61.0" }, - { name = "numpy", specifier = "==1.26.4" }, - { name = "omegaconf", specifier = "==2.3.0" }, - { name = "onnx", specifier = "==1.17.0" }, - { name = "optuna", specifier = "==4.4.0" }, - { name = "overrides", specifier = "==7.7.0" }, - { name = "packaging", specifier = "==24.2" }, - { name = "pandas", specifier = "==2.3.0" }, - { name = "pandocfilters", specifier = "==1.5.1" }, - { name = "parso", specifier = "==0.8.4" }, - { name = "peft", specifier = "==0.15.2" }, - { name = "pexpect", specifier = "==4.9.0" }, - { name = "pillow", specifier = "==11.2.1" }, - { name = "pip", specifier = ">=25.1.1" }, - { name = "plac", specifier = "==1.4.5" }, - { name = "platformdirs", specifier = "==4.3.8" }, - { name = "pooch", specifier = "==1.8.2" }, - { name = "prometheus-client", specifier = "==0.22.1" }, - { name = "prompt-toolkit", specifier = "==3.0.51" }, - { name = "propcache", specifier = "==0.3.2" }, - { name = "psutil", specifier = "==7.0.0" }, - { name = "ptyprocess", specifier = "==0.7.0" }, - { name = "pure-eval", specifier = "==0.2.3" }, - { name = "pyaml", specifier = "==25.5.0" }, - { name = "pyannote-audio", specifier = ">=3.3.2" }, - { name = "pyannote-core", specifier = "==5.0.0" }, - { name = "pyannote-database", specifier = "==5.1.3" }, - { name = "pyannote-metrics", specifier = "==3.2.1" }, - { name = "pyarrow", specifier = "==20.0.0" }, - { name = "pybind11", specifier = "==2.13.6" }, - { name = "pycparser", specifier = "==2.22" }, - { name = "pydantic", specifier = "==2.11.7" }, - { name = "pydantic-core", specifier = "==2.33.2" }, - { name = "pydub", specifier = "==0.25.1" }, - { name = "pygments", specifier = "==2.19.2" }, - { name = "pyloudnorm", specifier = "==0.1.1" }, - { name = "pyparsing", specifier = "==3.2.3" }, - { name = "python-dateutil", specifier = "==2.9.0.post0" }, - { name = "python-json-logger", specifier = "==3.3.0" }, - { name = "pytorch-lightning", specifier = "==2.5.2" }, - { name = "pytz", specifier = "==2025.2" }, - { name = "pyyaml", specifier = "==6.0.2" }, - { name = "pyzmq", specifier = "==27.0.0" }, - { name = "rapidfuzz", specifier = "==3.13.0" }, - { name = "referencing", specifier = "==0.36.2" }, - { name = "regex", specifier = "==2024.11.6" }, - { name = "requests", specifier = "==2.32.4" }, - { name = "resampy", specifier = "==0.4.3" }, - { name = "rfc3339-validator", specifier = "==0.1.4" }, - { name = "rfc3986-validator", specifier = "==0.1.1" }, - { name = "rich", specifier = "==14.0.0" }, - { name = "rpds-py", specifier = "==0.25.1" }, - { name = "ruamel-yaml", specifier = "==0.18.14" }, - { name = "ruamel-yaml-clib", specifier = "==0.2.12" }, - { name = "sacremoses", specifier = "==0.1.1" }, - { name = "safetensors", specifier = "==0.5.3" }, - { name = "scikit-learn", specifier = "==1.5.1" }, - { name = "scipy", specifier = "==1.15.3" }, - { name = "seaborn", specifier = ">=0.13.2" }, - { name = "send2trash", specifier = "==1.8.3" }, - { name = "sentencepiece", specifier = "==0.2.0" }, - { name = "sentry-sdk", specifier = "==2.32.0" }, - { name = "setproctitle", specifier = "==1.3.6" }, - { name = "shellingham", specifier = "==1.5.4" }, - { name = "six", specifier = "==1.17.0" }, - { name = "smmap", specifier = "==5.0.2" }, - { name = "sniffio", specifier = "==1.3.1" }, - { name = "sortedcontainers", specifier = "==2.4.0" }, - { name = "soundfile", specifier = "==0.13.1" }, - { name = "soupsieve", specifier = "==2.7" }, - { name = "sox", specifier = "==1.5.0" }, - { name = "soxr", specifier = "==0.5.0.post1" }, - { name = "sqlalchemy", specifier = "==2.0.41" }, - { name = "stack-data", specifier = "==0.6.3" }, - { name = "tabulate", specifier = "==0.9.0" }, - { name = "tensorboard", specifier = "==2.19.0" }, - { name = "tensorboard-data-server", specifier = "==0.7.2" }, - { name = "termcolor", specifier = "==3.1.0" }, - { name = "terminado", specifier = "==0.18.1" }, - { name = "text-unidecode", specifier = "==1.3" }, - { name = "texterrors", specifier = "==0.5.1" }, - { name = "threadpoolctl", specifier = "==3.6.0" }, - { name = "tinycss2", specifier = "==1.4.0" }, - { name = "tokenizers", specifier = "==0.21.2" }, - { name = "tomli", specifier = "==2.2.1" }, - { name = "toolz", specifier = "==1.0.0" }, - { name = "torch", specifier = "==2.7.0" }, - { name = "torchmetrics", specifier = "==1.7.3" }, - { name = "tornado", specifier = "==6.5.1" }, - { name = "tqdm", specifier = "==4.67.1" }, - { name = "traitlets", specifier = "==5.14.3" }, - { name = "transformers", specifier = "==4.51.3" }, - { name = "typeguard", specifier = "==4.4.4" }, - { name = "typer", specifier = "==0.16.0" }, - { name = "types-python-dateutil", specifier = "==2.9.0.20250516" }, - { name = "typing-extensions", specifier = "==4.14.0" }, - { name = "typing-inspection", specifier = "==0.4.1" }, - { name = "tzdata", specifier = "==2025.2" }, - { name = "uri-template", specifier = "==1.3.0" }, - { name = "urllib3", specifier = "==2.5.0" }, - { name = "wandb", specifier = "==0.20.1" }, - { name = "wcwidth", specifier = "==0.2.13" }, - { name = "webcolors", specifier = "==24.11.1" }, - { name = "webdataset", specifier = "==1.0.2" }, - { name = "webencodings", specifier = "==0.5.1" }, - { name = "websocket-client", specifier = "==1.8.0" }, - { name = "werkzeug", specifier = "==3.1.3" }, - { name = "wget", specifier = "==3.2" }, - { name = "widgetsnbextension", specifier = "==4.0.14" }, - { name = "wrapt", specifier = "==1.17.2" }, - { name = "xxhash", specifier = "==3.5.0" }, - { name = "yarl", specifier = "==1.20.1" }, -] - -[[package]] -name = "parso" -version = "0.8.4" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/66/94/68e2e17afaa9169cf6412ab0f28623903be73d1b32e208d9e8e541bb086d/parso-0.8.4.tar.gz", hash = "sha256:eb3a7b58240fb99099a345571deecc0f9540ea5f4dd2fe14c2a99d6b281ab92d", size = 400609 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/c6/ac/dac4a63f978e4dcb3c6d3a78c4d8e0192a113d288502a1216950c41b1027/parso-0.8.4-py2.py3-none-any.whl", hash = "sha256:a418670a20291dacd2dddc80c377c5c3791378ee1e8d12bffc35420643d43f18", size = 103650 }, -] - -[[package]] -name = "peft" -version = "0.15.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "accelerate" }, - { name = "huggingface-hub" }, - { name = "numpy" }, - { name = "packaging" }, - { name = "psutil" }, - { name = "pyyaml" }, - { name = "safetensors" }, - { name = "torch" }, - { name = "tqdm" }, - { name = "transformers" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/33/65/faa18cd8ffbe0f742c3f2559770646cce2574b9cd28a2a05e8d36f64e968/peft-0.15.2.tar.gz", hash = "sha256:7059029f4d42a092ded1aa117dd366a46084aef638bdd593f6ab0195d5427fcd", size = 472952 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/68/85/8e6ea3d1089f2b6de3c1cd34bbbd7560912af9d34b057be3b8b8fefe1da3/peft-0.15.2-py3-none-any.whl", hash = "sha256:0dfc942b03b7af4b7267cd4e30b15e3a4a1d277adc581ce6245fc13f1f93d0a0", size = 411051 }, -] - -[[package]] -name = "pexpect" -version = "4.9.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "ptyprocess" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/42/92/cc564bf6381ff43ce1f4d06852fc19a2f11d180f23dc32d9588bee2f149d/pexpect-4.9.0.tar.gz", hash = "sha256:ee7d41123f3c9911050ea2c2dac107568dc43b2d3b0c7557a33212c398ead30f", size = 166450 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/9e/c3/059298687310d527a58bb01f3b1965787ee3b40dce76752eda8b44e9a2c5/pexpect-4.9.0-py2.py3-none-any.whl", hash = "sha256:7236d1e080e4936be2dc3e326cec0af72acf9212a7e1d060210e70a47e253523", size = 63772 }, -] - -[[package]] -name = "pillow" -version = "11.2.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/af/cb/bb5c01fcd2a69335b86c22142b2bccfc3464087efb7fd382eee5ffc7fdf7/pillow-11.2.1.tar.gz", hash = "sha256:a64dd61998416367b7ef979b73d3a85853ba9bec4c2925f74e588879a58716b6", size = 47026707 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/0d/8b/b158ad57ed44d3cc54db8d68ad7c0a58b8fc0e4c7a3f995f9d62d5b464a1/pillow-11.2.1-cp310-cp310-macosx_10_10_x86_64.whl", hash = "sha256:d57a75d53922fc20c165016a20d9c44f73305e67c351bbc60d1adaf662e74047", size = 3198442 }, - { url = "https://files.pythonhosted.org/packages/b1/f8/bb5d956142f86c2d6cc36704943fa761f2d2e4c48b7436fd0a85c20f1713/pillow-11.2.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:127bf6ac4a5b58b3d32fc8289656f77f80567d65660bc46f72c0d77e6600cc95", size = 3030553 }, - { url = "https://files.pythonhosted.org/packages/22/7f/0e413bb3e2aa797b9ca2c5c38cb2e2e45d88654e5b12da91ad446964cfae/pillow-11.2.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:b4ba4be812c7a40280629e55ae0b14a0aafa150dd6451297562e1764808bbe61", size = 4405503 }, - { url = "https://files.pythonhosted.org/packages/f3/b4/cc647f4d13f3eb837d3065824aa58b9bcf10821f029dc79955ee43f793bd/pillow-11.2.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c8bd62331e5032bc396a93609982a9ab6b411c05078a52f5fe3cc59234a3abd1", size = 4490648 }, - { url = "https://files.pythonhosted.org/packages/c2/6f/240b772a3b35cdd7384166461567aa6713799b4e78d180c555bd284844ea/pillow-11.2.1-cp310-cp310-manylinux_2_28_aarch64.whl", hash = "sha256:562d11134c97a62fe3af29581f083033179f7ff435f78392565a1ad2d1c2c45c", size = 4508937 }, - { url = "https://files.pythonhosted.org/packages/f3/5e/7ca9c815ade5fdca18853db86d812f2f188212792780208bdb37a0a6aef4/pillow-11.2.1-cp310-cp310-manylinux_2_28_x86_64.whl", hash = "sha256:c97209e85b5be259994eb5b69ff50c5d20cca0f458ef9abd835e262d9d88b39d", size = 4599802 }, - { url = "https://files.pythonhosted.org/packages/02/81/c3d9d38ce0c4878a77245d4cf2c46d45a4ad0f93000227910a46caff52f3/pillow-11.2.1-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:0c3e6d0f59171dfa2e25d7116217543310908dfa2770aa64b8f87605f8cacc97", size = 4576717 }, - { url = "https://files.pythonhosted.org/packages/42/49/52b719b89ac7da3185b8d29c94d0e6aec8140059e3d8adcaa46da3751180/pillow-11.2.1-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:cc1c3bc53befb6096b84165956e886b1729634a799e9d6329a0c512ab651e579", size = 4654874 }, - { url = "https://files.pythonhosted.org/packages/5b/0b/ede75063ba6023798267023dc0d0401f13695d228194d2242d5a7ba2f964/pillow-11.2.1-cp310-cp310-win32.whl", hash = "sha256:312c77b7f07ab2139924d2639860e084ec2a13e72af54d4f08ac843a5fc9c79d", size = 2331717 }, - { url = "https://files.pythonhosted.org/packages/ed/3c/9831da3edea527c2ed9a09f31a2c04e77cd705847f13b69ca60269eec370/pillow-11.2.1-cp310-cp310-win_amd64.whl", hash = "sha256:9bc7ae48b8057a611e5fe9f853baa88093b9a76303937449397899385da06fad", size = 2676204 }, - { url = "https://files.pythonhosted.org/packages/01/97/1f66ff8a1503d8cbfc5bae4dc99d54c6ec1e22ad2b946241365320caabc2/pillow-11.2.1-cp310-cp310-win_arm64.whl", hash = "sha256:2728567e249cdd939f6cc3d1f049595c66e4187f3c34078cbc0a7d21c47482d2", size = 2414767 }, - { url = "https://files.pythonhosted.org/packages/33/49/c8c21e4255b4f4a2c0c68ac18125d7f5460b109acc6dfdef1a24f9b960ef/pillow-11.2.1-pp310-pypy310_pp73-macosx_10_15_x86_64.whl", hash = "sha256:9b7b0d4fd2635f54ad82785d56bc0d94f147096493a79985d0ab57aedd563156", size = 3181727 }, - { url = "https://files.pythonhosted.org/packages/6d/f1/f7255c0838f8c1ef6d55b625cfb286835c17e8136ce4351c5577d02c443b/pillow-11.2.1-pp310-pypy310_pp73-macosx_11_0_arm64.whl", hash = "sha256:aa442755e31c64037aa7c1cb186e0b369f8416c567381852c63444dd666fb772", size = 2999833 }, - { url = "https://files.pythonhosted.org/packages/e2/57/9968114457bd131063da98d87790d080366218f64fa2943b65ac6739abb3/pillow-11.2.1-pp310-pypy310_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:f0d3348c95b766f54b76116d53d4cb171b52992a1027e7ca50c81b43b9d9e363", size = 3437472 }, - { url = "https://files.pythonhosted.org/packages/b2/1b/e35d8a158e21372ecc48aac9c453518cfe23907bb82f950d6e1c72811eb0/pillow-11.2.1-pp310-pypy310_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:85d27ea4c889342f7e35f6d56e7e1cb345632ad592e8c51b693d7b7556043ce0", size = 3459976 }, - { url = "https://files.pythonhosted.org/packages/26/da/2c11d03b765efff0ccc473f1c4186dc2770110464f2177efaed9cf6fae01/pillow-11.2.1-pp310-pypy310_pp73-manylinux_2_28_aarch64.whl", hash = "sha256:bf2c33d6791c598142f00c9c4c7d47f6476731c31081331664eb26d6ab583e01", size = 3527133 }, - { url = "https://files.pythonhosted.org/packages/79/1a/4e85bd7cadf78412c2a3069249a09c32ef3323650fd3005c97cca7aa21df/pillow-11.2.1-pp310-pypy310_pp73-manylinux_2_28_x86_64.whl", hash = "sha256:e616e7154c37669fc1dfc14584f11e284e05d1c650e1c0f972f281c4ccc53193", size = 3571555 }, - { url = "https://files.pythonhosted.org/packages/69/03/239939915216de1e95e0ce2334bf17a7870ae185eb390fab6d706aadbfc0/pillow-11.2.1-pp310-pypy310_pp73-win_amd64.whl", hash = "sha256:39ad2e0f424394e3aebc40168845fee52df1394a4673a6ee512d840d14ab3013", size = 2674713 }, -] - -[[package]] -name = "pip" -version = "25.2" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/20/16/650289cd3f43d5a2fadfd98c68bd1e1e7f2550a1a5326768cddfbcedb2c5/pip-25.2.tar.gz", hash = "sha256:578283f006390f85bb6282dffb876454593d637f5d1be494b5202ce4877e71f2", size = 1840021 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/b7/3f/945ef7ab14dc4f9d7f40288d2df998d1837ee0888ec3659c813487572faa/pip-25.2-py3-none-any.whl", hash = "sha256:6d67a2b4e7f14d8b31b8b52648866fa717f45a1eb70e83002f4331d07e953717", size = 1752557 }, -] - -[[package]] -name = "plac" -version = "1.4.5" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/23/09/26ef2d614cabdcc52a7f383d0dc7967bf46be3c9700898c594e37b710c3d/plac-1.4.5.tar.gz", hash = "sha256:5f05bf85235c017fcd76c73c8101d4ff8e96beb3dc58b9a37de49cac7de82d14", size = 38988 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/15/36/38676114a0dbee137ec366daa86603d667a07e9a52667d5ebf5c580100ba/plac-1.4.5-py2.py3-none-any.whl", hash = "sha256:87187786b4e446688b1cf5112e18fed8a23ab3b316c25fe91266a10bd1736b16", size = 22468 }, -] - -[[package]] -name = "platformdirs" -version = "4.3.8" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/fe/8b/3c73abc9c759ecd3f1f7ceff6685840859e8070c4d947c93fae71f6a0bf2/platformdirs-4.3.8.tar.gz", hash = "sha256:3d512d96e16bcb959a814c9f348431070822a6496326a4be0911c40b5a74c2bc", size = 21362 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/fe/39/979e8e21520d4e47a0bbe349e2713c0aac6f3d853d0e5b34d76206c439aa/platformdirs-4.3.8-py3-none-any.whl", hash = "sha256:ff7059bb7eb1179e2685604f4aaf157cfd9535242bd23742eadc3c13542139b4", size = 18567 }, -] - -[[package]] -name = "pooch" -version = "1.8.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "packaging" }, - { name = "platformdirs" }, - { name = "requests" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/c6/77/b3d3e00c696c16cf99af81ef7b1f5fe73bd2a307abca41bd7605429fe6e5/pooch-1.8.2.tar.gz", hash = "sha256:76561f0de68a01da4df6af38e9955c4c9d1a5c90da73f7e40276a5728ec83d10", size = 59353 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/a8/87/77cc11c7a9ea9fd05503def69e3d18605852cd0d4b0d3b8f15bbeb3ef1d1/pooch-1.8.2-py3-none-any.whl", hash = "sha256:3529a57096f7198778a5ceefd5ac3ef0e4d06a6ddaf9fc2d609b806f25302c47", size = 64574 }, -] - -[[package]] -name = "primepy" -version = "1.3" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/35/77/0cfa1b4697cfb5336f3a96e8bc73327f64610be3a64c97275f1801afb395/primePy-1.3.tar.gz", hash = "sha256:25fd7e25344b0789a5984c75d89f054fcf1f180bef20c998e4befbac92de4669", size = 3914 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/74/c1/bb7e334135859c3a92ec399bc89293ea73f28e815e35b43929c8db6af030/primePy-1.3-py3-none-any.whl", hash = "sha256:5ed443718765be9bf7e2ff4c56cdff71b42140a15b39d054f9d99f0009e2317a", size = 4040 }, -] - -[[package]] -name = "prometheus-client" -version = "0.22.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/5e/cf/40dde0a2be27cc1eb41e333d1a674a74ce8b8b0457269cc640fd42b07cf7/prometheus_client-0.22.1.tar.gz", hash = "sha256:190f1331e783cf21eb60bca559354e0a4d4378facecf78f5428c39b675d20d28", size = 69746 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/32/ae/ec06af4fe3ee72d16973474f122541746196aaa16cea6f66d18b963c6177/prometheus_client-0.22.1-py3-none-any.whl", hash = "sha256:cca895342e308174341b2cbf99a56bef291fbc0ef7b9e5412a0f26d653ba7094", size = 58694 }, -] - -[[package]] -name = "prompt-toolkit" -version = "3.0.51" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "wcwidth" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/bb/6e/9d084c929dfe9e3bfe0c6a47e31f78a25c54627d64a66e884a8bf5474f1c/prompt_toolkit-3.0.51.tar.gz", hash = "sha256:931a162e3b27fc90c86f1b48bb1fb2c528c2761475e57c9c06de13311c7b54ed", size = 428940 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/ce/4f/5249960887b1fbe561d9ff265496d170b55a735b76724f10ef19f9e40716/prompt_toolkit-3.0.51-py3-none-any.whl", hash = "sha256:52742911fde84e2d423e2f9a4cf1de7d7ac4e51958f648d9540e0fb8db077b07", size = 387810 }, -] - -[[package]] -name = "propcache" -version = "0.3.2" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/a6/16/43264e4a779dd8588c21a70f0709665ee8f611211bdd2c87d952cfa7c776/propcache-0.3.2.tar.gz", hash = "sha256:20d7d62e4e7ef05f221e0db2856b979540686342e7dd9973b815599c7057e168", size = 44139 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/ab/14/510deed325e262afeb8b360043c5d7c960da7d3ecd6d6f9496c9c56dc7f4/propcache-0.3.2-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:22d9962a358aedbb7a2e36187ff273adeaab9743373a272976d2e348d08c7770", size = 73178 }, - { url = "https://files.pythonhosted.org/packages/cd/4e/ad52a7925ff01c1325653a730c7ec3175a23f948f08626a534133427dcff/propcache-0.3.2-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:0d0fda578d1dc3f77b6b5a5dce3b9ad69a8250a891760a548df850a5e8da87f3", size = 43133 }, - { url = "https://files.pythonhosted.org/packages/63/7c/e9399ba5da7780871db4eac178e9c2e204c23dd3e7d32df202092a1ed400/propcache-0.3.2-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:3def3da3ac3ce41562d85db655d18ebac740cb3fa4367f11a52b3da9d03a5cc3", size = 43039 }, - { url = "https://files.pythonhosted.org/packages/22/e1/58da211eb8fdc6fc854002387d38f415a6ca5f5c67c1315b204a5d3e9d7a/propcache-0.3.2-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:9bec58347a5a6cebf239daba9bda37dffec5b8d2ce004d9fe4edef3d2815137e", size = 201903 }, - { url = "https://files.pythonhosted.org/packages/c4/0a/550ea0f52aac455cb90111c8bab995208443e46d925e51e2f6ebdf869525/propcache-0.3.2-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:55ffda449a507e9fbd4aca1a7d9aa6753b07d6166140e5a18d2ac9bc49eac220", size = 213362 }, - { url = "https://files.pythonhosted.org/packages/5a/af/9893b7d878deda9bb69fcf54600b247fba7317761b7db11fede6e0f28bd0/propcache-0.3.2-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:64a67fb39229a8a8491dd42f864e5e263155e729c2e7ff723d6e25f596b1e8cb", size = 210525 }, - { url = "https://files.pythonhosted.org/packages/7c/bb/38fd08b278ca85cde36d848091ad2b45954bc5f15cce494bb300b9285831/propcache-0.3.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9da1cf97b92b51253d5b68cf5a2b9e0dafca095e36b7f2da335e27dc6172a614", size = 198283 }, - { url = "https://files.pythonhosted.org/packages/78/8c/9fe55bd01d362bafb413dfe508c48753111a1e269737fa143ba85693592c/propcache-0.3.2-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:5f559e127134b07425134b4065be45b166183fdcb433cb6c24c8e4149056ad50", size = 191872 }, - { url = "https://files.pythonhosted.org/packages/54/14/4701c33852937a22584e08abb531d654c8bcf7948a8f87ad0a4822394147/propcache-0.3.2-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:aff2e4e06435d61f11a428360a932138d0ec288b0a31dd9bd78d200bd4a2b339", size = 199452 }, - { url = "https://files.pythonhosted.org/packages/16/44/447f2253d859602095356007657ee535e0093215ea0b3d1d6a41d16e5201/propcache-0.3.2-cp310-cp310-musllinux_1_2_armv7l.whl", hash = "sha256:4927842833830942a5d0a56e6f4839bc484785b8e1ce8d287359794818633ba0", size = 191567 }, - { url = "https://files.pythonhosted.org/packages/f2/b3/e4756258749bb2d3b46defcff606a2f47410bab82be5824a67e84015b267/propcache-0.3.2-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:6107ddd08b02654a30fb8ad7a132021759d750a82578b94cd55ee2772b6ebea2", size = 193015 }, - { url = "https://files.pythonhosted.org/packages/1e/df/e6d3c7574233164b6330b9fd697beeac402afd367280e6dc377bb99b43d9/propcache-0.3.2-cp310-cp310-musllinux_1_2_ppc64le.whl", hash = "sha256:70bd8b9cd6b519e12859c99f3fc9a93f375ebd22a50296c3a295028bea73b9e7", size = 204660 }, - { url = "https://files.pythonhosted.org/packages/b2/53/e4d31dd5170b4a0e2e6b730f2385a96410633b4833dc25fe5dffd1f73294/propcache-0.3.2-cp310-cp310-musllinux_1_2_s390x.whl", hash = "sha256:2183111651d710d3097338dd1893fcf09c9f54e27ff1a8795495a16a469cc90b", size = 206105 }, - { url = "https://files.pythonhosted.org/packages/7f/fe/74d54cf9fbe2a20ff786e5f7afcfde446588f0cf15fb2daacfbc267b866c/propcache-0.3.2-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:fb075ad271405dcad8e2a7ffc9a750a3bf70e533bd86e89f0603e607b93aa64c", size = 196980 }, - { url = "https://files.pythonhosted.org/packages/22/ec/c469c9d59dada8a7679625e0440b544fe72e99311a4679c279562051f6fc/propcache-0.3.2-cp310-cp310-win32.whl", hash = "sha256:404d70768080d3d3bdb41d0771037da19d8340d50b08e104ca0e7f9ce55fce70", size = 37679 }, - { url = "https://files.pythonhosted.org/packages/38/35/07a471371ac89d418f8d0b699c75ea6dca2041fbda360823de21f6a9ce0a/propcache-0.3.2-cp310-cp310-win_amd64.whl", hash = "sha256:7435d766f978b4ede777002e6b3b6641dd229cd1da8d3d3106a45770365f9ad9", size = 41459 }, - { url = "https://files.pythonhosted.org/packages/cc/35/cc0aaecf278bb4575b8555f2b137de5ab821595ddae9da9d3cd1da4072c7/propcache-0.3.2-py3-none-any.whl", hash = "sha256:98f1ec44fb675f5052cccc8e609c46ed23a35a1cfd18545ad4e29002d858a43f", size = 12663 }, -] - -[[package]] -name = "protobuf" -version = "4.24.4" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/52/5c/f2c0778278259089952f94b0884ca27a001a17ffbd992ebe30c841085f4c/protobuf-4.24.4.tar.gz", hash = "sha256:5a70731910cd9104762161719c3d883c960151eea077134458503723b60e3667", size = 383850 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/00/0f/3dc2f86e9c3d6e73c56d915a3563ecc96ebee8144fb39614f0d6c1fb023d/protobuf-4.24.4-cp310-abi3-win32.whl", hash = "sha256:ec9912d5cb6714a5710e28e592ee1093d68c5ebfeda61983b3f40331da0b1ebb", size = 409968 }, - { url = "https://files.pythonhosted.org/packages/c2/59/f89c04923d68595d359f4cd7adbbdf5e5d791257945f8873d88b2fd1f979/protobuf-4.24.4-cp310-abi3-win_amd64.whl", hash = "sha256:1badab72aa8a3a2b812eacfede5020472e16c6b2212d737cefd685884c191085", size = 430493 }, - { url = "https://files.pythonhosted.org/packages/88/12/efb5896c901382548ecb58d0449885a8f9aa62bb559d65e5a8a47f122629/protobuf-4.24.4-cp37-abi3-macosx_10_9_universal2.whl", hash = "sha256:8e61a27f362369c2f33248a0ff6896c20dcd47b5d48239cb9720134bef6082e4", size = 409417 }, - { url = "https://files.pythonhosted.org/packages/db/61/9c7b481771fe4702fb3be1152812fecec9b06f9c36d523ad52b98cb46800/protobuf-4.24.4-cp37-abi3-manylinux2014_aarch64.whl", hash = "sha256:bffa46ad9612e6779d0e51ae586fde768339b791a50610d85eb162daeb23661e", size = 310587 }, - { url = "https://files.pythonhosted.org/packages/c8/2c/03046cac73f46bfe98fc846ef629cf4f84c2f59258216aa2cc0d22bfca8f/protobuf-4.24.4-cp37-abi3-manylinux2014_x86_64.whl", hash = "sha256:b493cb590960ff863743b9ff1452c413c2ee12b782f48beca77c8da3e2ffe9d9", size = 311559 }, - { url = "https://files.pythonhosted.org/packages/e5/a7/bb962b8b981dd890a44a34d0e922b76c32e5db443ff9f9b9ce6149069070/protobuf-4.24.4-py3-none-any.whl", hash = "sha256:80797ce7424f8c8d2f2547e2d42bfbb6c08230ce5832d6c099a37335c9c90a92", size = 175677 }, -] - -[[package]] -name = "psutil" -version = "7.0.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/2a/80/336820c1ad9286a4ded7e845b2eccfcb27851ab8ac6abece774a6ff4d3de/psutil-7.0.0.tar.gz", hash = "sha256:7be9c3eba38beccb6495ea33afd982a44074b78f28c434a1f51cc07fd315c456", size = 497003 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/ed/e6/2d26234410f8b8abdbf891c9da62bee396583f713fb9f3325a4760875d22/psutil-7.0.0-cp36-abi3-macosx_10_9_x86_64.whl", hash = "sha256:101d71dc322e3cffd7cea0650b09b3d08b8e7c4109dd6809fe452dfd00e58b25", size = 238051 }, - { url = "https://files.pythonhosted.org/packages/04/8b/30f930733afe425e3cbfc0e1468a30a18942350c1a8816acfade80c005c4/psutil-7.0.0-cp36-abi3-macosx_11_0_arm64.whl", hash = "sha256:39db632f6bb862eeccf56660871433e111b6ea58f2caea825571951d4b6aa3da", size = 239535 }, - { url = "https://files.pythonhosted.org/packages/2a/ed/d362e84620dd22876b55389248e522338ed1bf134a5edd3b8231d7207f6d/psutil-7.0.0-cp36-abi3-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:1fcee592b4c6f146991ca55919ea3d1f8926497a713ed7faaf8225e174581e91", size = 275004 }, - { url = "https://files.pythonhosted.org/packages/bf/b9/b0eb3f3cbcb734d930fdf839431606844a825b23eaf9a6ab371edac8162c/psutil-7.0.0-cp36-abi3-manylinux_2_12_x86_64.manylinux2010_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4b1388a4f6875d7e2aff5c4ca1cc16c545ed41dd8bb596cefea80111db353a34", size = 277986 }, - { url = "https://files.pythonhosted.org/packages/eb/a2/709e0fe2f093556c17fbafda93ac032257242cabcc7ff3369e2cb76a97aa/psutil-7.0.0-cp36-abi3-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a5f098451abc2828f7dc6b58d44b532b22f2088f4999a937557b603ce72b1993", size = 279544 }, - { url = "https://files.pythonhosted.org/packages/50/e6/eecf58810b9d12e6427369784efe814a1eec0f492084ce8eb8f4d89d6d61/psutil-7.0.0-cp37-abi3-win32.whl", hash = "sha256:ba3fcef7523064a6c9da440fc4d6bd07da93ac726b5733c29027d7dc95b39d99", size = 241053 }, - { url = "https://files.pythonhosted.org/packages/50/1b/6921afe68c74868b4c9fa424dad3be35b095e16687989ebbb50ce4fceb7c/psutil-7.0.0-cp37-abi3-win_amd64.whl", hash = "sha256:4cf3d4eb1aa9b348dec30105c55cd9b7d4629285735a102beb4441e38db90553", size = 244885 }, -] - -[[package]] -name = "ptyprocess" -version = "0.7.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/20/e5/16ff212c1e452235a90aeb09066144d0c5a6a8c0834397e03f5224495c4e/ptyprocess-0.7.0.tar.gz", hash = "sha256:5c5d0a3b48ceee0b48485e0c26037c0acd7d29765ca3fbb5cb3831d347423220", size = 70762 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/22/a6/858897256d0deac81a172289110f31629fc4cee19b6f01283303e18c8db3/ptyprocess-0.7.0-py2.py3-none-any.whl", hash = "sha256:4b41f3967fce3af57cc7e94b888626c18bf37a083e3651ca8feeb66d492fef35", size = 13993 }, -] - -[[package]] -name = "pure-eval" -version = "0.2.3" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/cd/05/0a34433a064256a578f1783a10da6df098ceaa4a57bbeaa96a6c0352786b/pure_eval-0.2.3.tar.gz", hash = "sha256:5f4e983f40564c576c7c8635ae88db5956bb2229d7e9237d03b3c0b0190eaf42", size = 19752 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/8e/37/efad0257dc6e593a18957422533ff0f87ede7c9c6ea010a2177d738fb82f/pure_eval-0.2.3-py3-none-any.whl", hash = "sha256:1db8e35b67b3d218d818ae653e27f06c3aa420901fa7b081ca98cbedc874e0d0", size = 11842 }, -] - -[[package]] -name = "pyaml" -version = "25.5.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "pyyaml" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/c1/40/94f10f32ab952c5cca713d9ac9d8b2fdc37392d90eea403823eeac674c24/pyaml-25.5.0.tar.gz", hash = "sha256:5799560c7b1c9daf35a7a4535f53e2c30323f74cbd7cb4f2e715b16dd681a58a", size = 29812 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/aa/7d/1b5061beff826f902285827261485a058b943332eba8a5532a0164735205/pyaml-25.5.0-py3-none-any.whl", hash = "sha256:b9e0c4e58a5e8003f8f18e802db49fd0563ada587209b13e429bdcbefa87d035", size = 26422 }, -] - -[[package]] -name = "pyannote-audio" -version = "3.4.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "asteroid-filterbanks" }, - { name = "einops" }, - { name = "huggingface-hub" }, - { name = "lightning" }, - { name = "omegaconf" }, - { name = "pyannote-core" }, - { name = "pyannote-database" }, - { name = "pyannote-metrics" }, - { name = "pyannote-pipeline" }, - { name = "pytorch-metric-learning" }, - { name = "rich" }, - { name = "semver" }, - { name = "soundfile" }, - { name = "speechbrain" }, - { name = "tensorboardx" }, - { name = "torch" }, - { name = "torch-audiomentations" }, - { name = "torchaudio" }, - { name = "torchmetrics" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/ec/1e/efe9619c38f1281ddf21640654d8ea9e3f67c459b76f78657b26d8557bbe/pyannote_audio-3.4.0.tar.gz", hash = "sha256:d523d883cb8d37cb6daf99f3ba83f9138bb193646ad71e6eae7deb89d8ddd642", size = 804850 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/79/13/620c6f711b723653092fd063bfee82a6af5ea3a4d3c42efc53ce623a7f4d/pyannote_audio-3.4.0-py2.py3-none-any.whl", hash = "sha256:36e38f058059f46da3478dda581cda53d9d85a21173a3e70bbdbc3ba93b5e1b7", size = 897789 }, -] - -[[package]] -name = "pyannote-core" -version = "5.0.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "numpy" }, - { name = "scipy" }, - { name = "sortedcontainers" }, - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/65/03/feaf7534206f02c75baf151ce4b8c322b402a6f477c2be82f69d9269cbe6/pyannote.core-5.0.0.tar.gz", hash = "sha256:1a55bcc8bd680ba6be5fa53efa3b6f3d2cdd67144c07b6b4d8d66d5cb0d2096f", size = 59247 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/84/c4/370bc8ba66815a5832ece753a1009388bb07ea353d21c83f2d5a1a436f2c/pyannote.core-5.0.0-py3-none-any.whl", hash = "sha256:04920a6754492242ce0dc6017545595ab643870fe69a994f20c1a5f2da0544d0", size = 58475 }, -] - -[[package]] -name = "pyannote-database" -version = "5.1.3" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "pandas" }, - { name = "pyannote-core" }, - { name = "pyyaml" }, - { name = "typer" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/a9/ae/de36413d69a46be87cb612ebbcdc4eacbeebce3bc809124603e44a88fe26/pyannote.database-5.1.3.tar.gz", hash = "sha256:0eaf64c1cc506718de60d2d702f1359b1ae7ff252ee3e4799f1c5e378cd52c31", size = 49957 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/a1/64/92d51a3a05615ba58be8ba62a43f9f9f952d9f3646f7e4fb7826e5a3a24e/pyannote.database-5.1.3-py3-none-any.whl", hash = "sha256:37887844c7dfbcc075cb591eddc00aff45fae1ed905344e1f43e0090e63bd40a", size = 48127 }, -] - -[[package]] -name = "pyannote-metrics" -version = "3.2.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "docopt" }, - { name = "matplotlib" }, - { name = "numpy" }, - { name = "pandas" }, - { name = "pyannote-core" }, - { name = "pyannote-database" }, - { name = "scikit-learn" }, - { name = "scipy" }, - { name = "sympy" }, - { name = "tabulate" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/39/2b/6c5f01d3c49aa1c160765946e23782ca6436ae8b9bc514b56319ff5f16e7/pyannote.metrics-3.2.1.tar.gz", hash = "sha256:08024255a3550e96a8e9da4f5f4af326886548480de891414567c8900920ee5c", size = 49086 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/6c/7d/035b370ab834b30e849fe9cd092b7bd7f321fcc4a2c56b84e96476b7ede5/pyannote.metrics-3.2.1-py3-none-any.whl", hash = "sha256:46be797cdade26c82773e5018659ae610145260069c7c5bf3d3c8a029ade8e22", size = 51386 }, -] - -[[package]] -name = "pyannote-pipeline" -version = "3.0.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "docopt" }, - { name = "filelock" }, - { name = "optuna" }, - { name = "pyannote-core" }, - { name = "pyannote-database" }, - { name = "pyyaml" }, - { name = "scikit-learn" }, - { name = "tqdm" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/35/04/4bcfe0dd588577a188328b806f3a7213d8cead0ce5fe5784d01fd57df93f/pyannote.pipeline-3.0.1.tar.gz", hash = "sha256:021794e26a2cf5d8fb5bb1835951e71f5fac33eb14e23dfb7468e16b1b805151", size = 34486 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/83/42/1bf7cbf061ed05c580bfb63bffdd3f3474cbd5c02bee4fac518eea9e9d9e/pyannote.pipeline-3.0.1-py3-none-any.whl", hash = "sha256:819bde4c4dd514f740f2373dfec794832b9fc8e346a35e43a7681625ee187393", size = 31517 }, -] - -[[package]] -name = "pyarrow" -version = "20.0.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/a2/ee/a7810cb9f3d6e9238e61d312076a9859bf3668fd21c69744de9532383912/pyarrow-20.0.0.tar.gz", hash = "sha256:febc4a913592573c8d5805091a6c2b5064c8bd6e002131f01061797d91c783c1", size = 1125187 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/5b/23/77094eb8ee0dbe88441689cb6afc40ac312a1e15d3a7acc0586999518222/pyarrow-20.0.0-cp310-cp310-macosx_12_0_arm64.whl", hash = "sha256:c7dd06fd7d7b410ca5dc839cc9d485d2bc4ae5240851bcd45d85105cc90a47d7", size = 30832591 }, - { url = "https://files.pythonhosted.org/packages/c3/d5/48cc573aff00d62913701d9fac478518f693b30c25f2c157550b0b2565cb/pyarrow-20.0.0-cp310-cp310-macosx_12_0_x86_64.whl", hash = "sha256:d5382de8dc34c943249b01c19110783d0d64b207167c728461add1ecc2db88e4", size = 32273686 }, - { url = "https://files.pythonhosted.org/packages/37/df/4099b69a432b5cb412dd18adc2629975544d656df3d7fda6d73c5dba935d/pyarrow-20.0.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:6415a0d0174487456ddc9beaead703d0ded5966129fa4fd3114d76b5d1c5ceae", size = 41337051 }, - { url = "https://files.pythonhosted.org/packages/4c/27/99922a9ac1c9226f346e3a1e15e63dee6f623ed757ff2893f9d6994a69d3/pyarrow-20.0.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:15aa1b3b2587e74328a730457068dc6c89e6dcbf438d4369f572af9d320a25ee", size = 42404659 }, - { url = "https://files.pythonhosted.org/packages/21/d1/71d91b2791b829c9e98f1e0d85be66ed93aff399f80abb99678511847eaa/pyarrow-20.0.0-cp310-cp310-manylinux_2_28_aarch64.whl", hash = "sha256:5605919fbe67a7948c1f03b9f3727d82846c053cd2ce9303ace791855923fd20", size = 40695446 }, - { url = "https://files.pythonhosted.org/packages/f1/ca/ae10fba419a6e94329707487835ec721f5a95f3ac9168500bcf7aa3813c7/pyarrow-20.0.0-cp310-cp310-manylinux_2_28_x86_64.whl", hash = "sha256:a5704f29a74b81673d266e5ec1fe376f060627c2e42c5c7651288ed4b0db29e9", size = 42278528 }, - { url = "https://files.pythonhosted.org/packages/7a/a6/aba40a2bf01b5d00cf9cd16d427a5da1fad0fb69b514ce8c8292ab80e968/pyarrow-20.0.0-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:00138f79ee1b5aca81e2bdedb91e3739b987245e11fa3c826f9e57c5d102fb75", size = 42918162 }, - { url = "https://files.pythonhosted.org/packages/93/6b/98b39650cd64f32bf2ec6d627a9bd24fcb3e4e6ea1873c5e1ea8a83b1a18/pyarrow-20.0.0-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:f2d67ac28f57a362f1a2c1e6fa98bfe2f03230f7e15927aecd067433b1e70ce8", size = 44550319 }, - { url = "https://files.pythonhosted.org/packages/ab/32/340238be1eb5037e7b5de7e640ee22334417239bc347eadefaf8c373936d/pyarrow-20.0.0-cp310-cp310-win_amd64.whl", hash = "sha256:4a8b029a07956b8d7bd742ffca25374dd3f634b35e46cc7a7c3fa4c75b297191", size = 25770759 }, -] - -[[package]] -name = "pybind11" -version = "2.13.6" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/d2/c1/72b9622fcb32ff98b054f724e213c7f70d6898baa714f4516288456ceaba/pybind11-2.13.6.tar.gz", hash = "sha256:ba6af10348c12b24e92fa086b39cfba0eff619b61ac77c406167d813b096d39a", size = 218403 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/13/2f/0f24b288e2ce56f51c920137620b4434a38fd80583dbbe24fc2a1656c388/pybind11-2.13.6-py3-none-any.whl", hash = "sha256:237c41e29157b962835d356b370ededd57594a26d5894a795960f0047cb5caf5", size = 243282 }, -] - -[[package]] -name = "pycparser" -version = "2.22" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/1d/b2/31537cf4b1ca988837256c910a668b553fceb8f069bedc4b1c826024b52c/pycparser-2.22.tar.gz", hash = "sha256:491c8be9c040f5390f5bf44a5b07752bd07f56edf992381b05c701439eec10f6", size = 172736 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/13/a3/a812df4e2dd5696d1f351d58b8fe16a405b234ad2886a0dab9183fb78109/pycparser-2.22-py3-none-any.whl", hash = "sha256:c3702b6d3dd8c7abc1afa565d7e63d53a1d0bd86cdc24edd75470f4de499cfcc", size = 117552 }, -] - -[[package]] -name = "pycryptodome" -version = "3.23.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/8e/a6/8452177684d5e906854776276ddd34eca30d1b1e15aa1ee9cefc289a33f5/pycryptodome-3.23.0.tar.gz", hash = "sha256:447700a657182d60338bab09fdb27518f8856aecd80ae4c6bdddb67ff5da44ef", size = 4921276 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/db/6c/a1f71542c969912bb0e106f64f60a56cc1f0fabecf9396f45accbe63fa68/pycryptodome-3.23.0-cp37-abi3-macosx_10_9_universal2.whl", hash = "sha256:187058ab80b3281b1de11c2e6842a357a1f71b42cb1e15bce373f3d238135c27", size = 2495627 }, - { url = "https://files.pythonhosted.org/packages/6e/4e/a066527e079fc5002390c8acdd3aca431e6ea0a50ffd7201551175b47323/pycryptodome-3.23.0-cp37-abi3-macosx_10_9_x86_64.whl", hash = "sha256:cfb5cd445280c5b0a4e6187a7ce8de5a07b5f3f897f235caa11f1f435f182843", size = 1640362 }, - { url = "https://files.pythonhosted.org/packages/50/52/adaf4c8c100a8c49d2bd058e5b551f73dfd8cb89eb4911e25a0c469b6b4e/pycryptodome-3.23.0-cp37-abi3-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:67bd81fcbe34f43ad9422ee8fd4843c8e7198dd88dd3d40e6de42ee65fbe1490", size = 2182625 }, - { url = "https://files.pythonhosted.org/packages/5f/e9/a09476d436d0ff1402ac3867d933c61805ec2326c6ea557aeeac3825604e/pycryptodome-3.23.0-cp37-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c8987bd3307a39bc03df5c8e0e3d8be0c4c3518b7f044b0f4c15d1aa78f52575", size = 2268954 }, - { url = "https://files.pythonhosted.org/packages/f9/c5/ffe6474e0c551d54cab931918127c46d70cab8f114e0c2b5a3c071c2f484/pycryptodome-3.23.0-cp37-abi3-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:aa0698f65e5b570426fc31b8162ed4603b0c2841cbb9088e2b01641e3065915b", size = 2308534 }, - { url = "https://files.pythonhosted.org/packages/18/28/e199677fc15ecf43010f2463fde4c1a53015d1fe95fb03bca2890836603a/pycryptodome-3.23.0-cp37-abi3-musllinux_1_2_aarch64.whl", hash = "sha256:53ecbafc2b55353edcebd64bf5da94a2a2cdf5090a6915bcca6eca6cc452585a", size = 2181853 }, - { url = "https://files.pythonhosted.org/packages/ce/ea/4fdb09f2165ce1365c9eaefef36625583371ee514db58dc9b65d3a255c4c/pycryptodome-3.23.0-cp37-abi3-musllinux_1_2_i686.whl", hash = "sha256:156df9667ad9f2ad26255926524e1c136d6664b741547deb0a86a9acf5ea631f", size = 2342465 }, - { url = "https://files.pythonhosted.org/packages/22/82/6edc3fc42fe9284aead511394bac167693fb2b0e0395b28b8bedaa07ef04/pycryptodome-3.23.0-cp37-abi3-musllinux_1_2_x86_64.whl", hash = "sha256:dea827b4d55ee390dc89b2afe5927d4308a8b538ae91d9c6f7a5090f397af1aa", size = 2267414 }, - { url = "https://files.pythonhosted.org/packages/59/fe/aae679b64363eb78326c7fdc9d06ec3de18bac68be4b612fc1fe8902693c/pycryptodome-3.23.0-cp37-abi3-win32.whl", hash = "sha256:507dbead45474b62b2bbe318eb1c4c8ee641077532067fec9c1aa82c31f84886", size = 1768484 }, - { url = "https://files.pythonhosted.org/packages/54/2f/e97a1b8294db0daaa87012c24a7bb714147c7ade7656973fd6c736b484ff/pycryptodome-3.23.0-cp37-abi3-win_amd64.whl", hash = "sha256:c75b52aacc6c0c260f204cbdd834f76edc9fb0d8e0da9fbf8352ef58202564e2", size = 1799636 }, - { url = "https://files.pythonhosted.org/packages/18/3d/f9441a0d798bf2b1e645adc3265e55706aead1255ccdad3856dbdcffec14/pycryptodome-3.23.0-cp37-abi3-win_arm64.whl", hash = "sha256:11eeeb6917903876f134b56ba11abe95c0b0fd5e3330def218083c7d98bbcb3c", size = 1703675 }, - { url = "https://files.pythonhosted.org/packages/9f/7c/f5b0556590e7b4e710509105e668adb55aa9470a9f0e4dea9c40a4a11ce1/pycryptodome-3.23.0-pp27-pypy_73-manylinux2010_x86_64.whl", hash = "sha256:350ebc1eba1da729b35ab7627a833a1a355ee4e852d8ba0447fafe7b14504d56", size = 1705791 }, - { url = "https://files.pythonhosted.org/packages/33/38/dcc795578d610ea1aaffef4b148b8cafcfcf4d126b1e58231ddc4e475c70/pycryptodome-3.23.0-pp27-pypy_73-win32.whl", hash = "sha256:93837e379a3e5fd2bb00302a47aee9fdf7940d83595be3915752c74033d17ca7", size = 1780265 }, - { url = "https://files.pythonhosted.org/packages/d9/12/e33935a0709c07de084d7d58d330ec3f4daf7910a18e77937affdb728452/pycryptodome-3.23.0-pp310-pypy310_pp73-macosx_10_15_x86_64.whl", hash = "sha256:ddb95b49df036ddd264a0ad246d1be5b672000f12d6961ea2c267083a5e19379", size = 1623886 }, - { url = "https://files.pythonhosted.org/packages/22/0b/aa8f9419f25870889bebf0b26b223c6986652bdf071f000623df11212c90/pycryptodome-3.23.0-pp310-pypy310_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d8e95564beb8782abfd9e431c974e14563a794a4944c29d6d3b7b5ea042110b4", size = 1672151 }, - { url = "https://files.pythonhosted.org/packages/d4/5e/63f5cbde2342b7f70a39e591dbe75d9809d6338ce0b07c10406f1a140cdc/pycryptodome-3.23.0-pp310-pypy310_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:14e15c081e912c4b0d75632acd8382dfce45b258667aa3c67caf7a4d4c13f630", size = 1664461 }, - { url = "https://files.pythonhosted.org/packages/d6/92/608fbdad566ebe499297a86aae5f2a5263818ceeecd16733006f1600403c/pycryptodome-3.23.0-pp310-pypy310_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:a7fc76bf273353dc7e5207d172b83f569540fc9a28d63171061c42e361d22353", size = 1702440 }, - { url = "https://files.pythonhosted.org/packages/d1/92/2eadd1341abd2989cce2e2740b4423608ee2014acb8110438244ee97d7ff/pycryptodome-3.23.0-pp310-pypy310_pp73-win_amd64.whl", hash = "sha256:45c69ad715ca1a94f778215a11e66b7ff989d792a4d63b68dc586a1da1392ff5", size = 1803005 }, -] - -[[package]] -name = "pydantic" -version = "2.11.7" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "annotated-types" }, - { name = "pydantic-core" }, - { name = "typing-extensions" }, - { name = "typing-inspection" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/00/dd/4325abf92c39ba8623b5af936ddb36ffcfe0beae70405d456ab1fb2f5b8c/pydantic-2.11.7.tar.gz", hash = "sha256:d989c3c6cb79469287b1569f7447a17848c998458d49ebe294e975b9baf0f0db", size = 788350 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/6a/c0/ec2b1c8712ca690e5d61979dee872603e92b8a32f94cc1b72d53beab008a/pydantic-2.11.7-py3-none-any.whl", hash = "sha256:dde5df002701f6de26248661f6835bbe296a47bf73990135c7d07ce741b9623b", size = 444782 }, -] - -[[package]] -name = "pydantic-core" -version = "2.33.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/ad/88/5f2260bdfae97aabf98f1778d43f69574390ad787afb646292a638c923d4/pydantic_core-2.33.2.tar.gz", hash = "sha256:7cb8bc3605c29176e1b105350d2e6474142d7c1bd1d9327c4a9bdb46bf827acc", size = 435195 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/e5/92/b31726561b5dae176c2d2c2dc43a9c5bfba5d32f96f8b4c0a600dd492447/pydantic_core-2.33.2-cp310-cp310-macosx_10_12_x86_64.whl", hash = "sha256:2b3d326aaef0c0399d9afffeb6367d5e26ddc24d351dbc9c636840ac355dc5d8", size = 2028817 }, - { url = "https://files.pythonhosted.org/packages/a3/44/3f0b95fafdaca04a483c4e685fe437c6891001bf3ce8b2fded82b9ea3aa1/pydantic_core-2.33.2-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:0e5b2671f05ba48b94cb90ce55d8bdcaaedb8ba00cc5359f6810fc918713983d", size = 1861357 }, - { url = "https://files.pythonhosted.org/packages/30/97/e8f13b55766234caae05372826e8e4b3b96e7b248be3157f53237682e43c/pydantic_core-2.33.2-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0069c9acc3f3981b9ff4cdfaf088e98d83440a4c7ea1bc07460af3d4dc22e72d", size = 1898011 }, - { url = "https://files.pythonhosted.org/packages/9b/a3/99c48cf7bafc991cc3ee66fd544c0aae8dc907b752f1dad2d79b1b5a471f/pydantic_core-2.33.2-cp310-cp310-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:d53b22f2032c42eaaf025f7c40c2e3b94568ae077a606f006d206a463bc69572", size = 1982730 }, - { url = "https://files.pythonhosted.org/packages/de/8e/a5b882ec4307010a840fb8b58bd9bf65d1840c92eae7534c7441709bf54b/pydantic_core-2.33.2-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:0405262705a123b7ce9f0b92f123334d67b70fd1f20a9372b907ce1080c7ba02", size = 2136178 }, - { url = "https://files.pythonhosted.org/packages/e4/bb/71e35fc3ed05af6834e890edb75968e2802fe98778971ab5cba20a162315/pydantic_core-2.33.2-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:4b25d91e288e2c4e0662b8038a28c6a07eaac3e196cfc4ff69de4ea3db992a1b", size = 2736462 }, - { url = "https://files.pythonhosted.org/packages/31/0d/c8f7593e6bc7066289bbc366f2235701dcbebcd1ff0ef8e64f6f239fb47d/pydantic_core-2.33.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6bdfe4b3789761f3bcb4b1ddf33355a71079858958e3a552f16d5af19768fef2", size = 2005652 }, - { url = "https://files.pythonhosted.org/packages/d2/7a/996d8bd75f3eda405e3dd219ff5ff0a283cd8e34add39d8ef9157e722867/pydantic_core-2.33.2-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:efec8db3266b76ef9607c2c4c419bdb06bf335ae433b80816089ea7585816f6a", size = 2113306 }, - { url = "https://files.pythonhosted.org/packages/ff/84/daf2a6fb2db40ffda6578a7e8c5a6e9c8affb251a05c233ae37098118788/pydantic_core-2.33.2-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:031c57d67ca86902726e0fae2214ce6770bbe2f710dc33063187a68744a5ecac", size = 2073720 }, - { url = "https://files.pythonhosted.org/packages/77/fb/2258da019f4825128445ae79456a5499c032b55849dbd5bed78c95ccf163/pydantic_core-2.33.2-cp310-cp310-musllinux_1_1_armv7l.whl", hash = "sha256:f8de619080e944347f5f20de29a975c2d815d9ddd8be9b9b7268e2e3ef68605a", size = 2244915 }, - { url = "https://files.pythonhosted.org/packages/d8/7a/925ff73756031289468326e355b6fa8316960d0d65f8b5d6b3a3e7866de7/pydantic_core-2.33.2-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:73662edf539e72a9440129f231ed3757faab89630d291b784ca99237fb94db2b", size = 2241884 }, - { url = "https://files.pythonhosted.org/packages/0b/b0/249ee6d2646f1cdadcb813805fe76265745c4010cf20a8eba7b0e639d9b2/pydantic_core-2.33.2-cp310-cp310-win32.whl", hash = "sha256:0a39979dcbb70998b0e505fb1556a1d550a0781463ce84ebf915ba293ccb7e22", size = 1910496 }, - { url = "https://files.pythonhosted.org/packages/66/ff/172ba8f12a42d4b552917aa65d1f2328990d3ccfc01d5b7c943ec084299f/pydantic_core-2.33.2-cp310-cp310-win_amd64.whl", hash = "sha256:b0379a2b24882fef529ec3b4987cb5d003b9cda32256024e6fe1586ac45fc640", size = 1955019 }, - { url = "https://files.pythonhosted.org/packages/30/68/373d55e58b7e83ce371691f6eaa7175e3a24b956c44628eb25d7da007917/pydantic_core-2.33.2-pp310-pypy310_pp73-macosx_10_12_x86_64.whl", hash = "sha256:5c4aa4e82353f65e548c476b37e64189783aa5384903bfea4f41580f255fddfa", size = 2023982 }, - { url = "https://files.pythonhosted.org/packages/a4/16/145f54ac08c96a63d8ed6442f9dec17b2773d19920b627b18d4f10a061ea/pydantic_core-2.33.2-pp310-pypy310_pp73-macosx_11_0_arm64.whl", hash = "sha256:d946c8bf0d5c24bf4fe333af284c59a19358aa3ec18cb3dc4370080da1e8ad29", size = 1858412 }, - { url = "https://files.pythonhosted.org/packages/41/b1/c6dc6c3e2de4516c0bb2c46f6a373b91b5660312342a0cf5826e38ad82fa/pydantic_core-2.33.2-pp310-pypy310_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:87b31b6846e361ef83fedb187bb5b4372d0da3f7e28d85415efa92d6125d6e6d", size = 1892749 }, - { url = "https://files.pythonhosted.org/packages/12/73/8cd57e20afba760b21b742106f9dbdfa6697f1570b189c7457a1af4cd8a0/pydantic_core-2.33.2-pp310-pypy310_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:aa9d91b338f2df0508606f7009fde642391425189bba6d8c653afd80fd6bb64e", size = 2067527 }, - { url = "https://files.pythonhosted.org/packages/e3/d5/0bb5d988cc019b3cba4a78f2d4b3854427fc47ee8ec8e9eaabf787da239c/pydantic_core-2.33.2-pp310-pypy310_pp73-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:2058a32994f1fde4ca0480ab9d1e75a0e8c87c22b53a3ae66554f9af78f2fe8c", size = 2108225 }, - { url = "https://files.pythonhosted.org/packages/f1/c5/00c02d1571913d496aabf146106ad8239dc132485ee22efe08085084ff7c/pydantic_core-2.33.2-pp310-pypy310_pp73-musllinux_1_1_aarch64.whl", hash = "sha256:0e03262ab796d986f978f79c943fc5f620381be7287148b8010b4097f79a39ec", size = 2069490 }, - { url = "https://files.pythonhosted.org/packages/22/a8/dccc38768274d3ed3a59b5d06f59ccb845778687652daa71df0cab4040d7/pydantic_core-2.33.2-pp310-pypy310_pp73-musllinux_1_1_armv7l.whl", hash = "sha256:1a8695a8d00c73e50bff9dfda4d540b7dee29ff9b8053e38380426a85ef10052", size = 2237525 }, - { url = "https://files.pythonhosted.org/packages/d4/e7/4f98c0b125dda7cf7ccd14ba936218397b44f50a56dd8c16a3091df116c3/pydantic_core-2.33.2-pp310-pypy310_pp73-musllinux_1_1_x86_64.whl", hash = "sha256:fa754d1850735a0b0e03bcffd9d4b4343eb417e47196e4485d9cca326073a42c", size = 2238446 }, - { url = "https://files.pythonhosted.org/packages/ce/91/2ec36480fdb0b783cd9ef6795753c1dea13882f2e68e73bce76ae8c21e6a/pydantic_core-2.33.2-pp310-pypy310_pp73-win_amd64.whl", hash = "sha256:a11c8d26a50bfab49002947d3d237abe4d9e4b5bdc8846a63537b6488e197808", size = 2066678 }, -] - -[[package]] -name = "pydub" -version = "0.25.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/fe/9a/e6bca0eed82db26562c73b5076539a4a08d3cffd19c3cc5913a3e61145fd/pydub-0.25.1.tar.gz", hash = "sha256:980a33ce9949cab2a569606b65674d748ecbca4f0796887fd6f46173a7b0d30f", size = 38326 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/a6/53/d78dc063216e62fc55f6b2eebb447f6a4b0a59f55c8406376f76bf959b08/pydub-0.25.1-py2.py3-none-any.whl", hash = "sha256:65617e33033874b59d87db603aa1ed450633288aefead953b30bded59cb599a6", size = 32327 }, -] - -[[package]] -name = "pygments" -version = "2.19.2" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/b0/77/a5b8c569bf593b0140bde72ea885a803b82086995367bf2037de0159d924/pygments-2.19.2.tar.gz", hash = "sha256:636cb2477cec7f8952536970bc533bc43743542f70392ae026374600add5b887", size = 4968631 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/c7/21/705964c7812476f378728bdf590ca4b771ec72385c533964653c68e86bdc/pygments-2.19.2-py3-none-any.whl", hash = "sha256:86540386c03d588bb81d44bc3928634ff26449851e99741617ecb9037ee5ec0b", size = 1225217 }, -] - -[[package]] -name = "pyloudnorm" -version = "0.1.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "future" }, - { name = "numpy" }, - { name = "scipy" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/75/b5/39d59c44ecd828fabfdbd796b50a561e6543ca90ef440ab307374f107856/pyloudnorm-0.1.1.tar.gz", hash = "sha256:63cd4e197dea4e7795160ea08ed02d318091bce883e436a6dbc5963326b71e1e", size = 8588 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/58/f5/6724805521ab4e723a12182f92374031032aff28a8a89dc8505c52b79032/pyloudnorm-0.1.1-py3-none-any.whl", hash = "sha256:d7f12ebdd097a464d87ce2878fc4d942f15f8233e26cc03f33fefa226f869a14", size = 9636 }, -] - -[[package]] -name = "pynndescent" -version = "0.5.13" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "joblib" }, - { name = "llvmlite" }, - { name = "numba" }, - { name = "scikit-learn" }, - { name = "scipy" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/7e/58/560a4db5eb3794d922fe55804b10326534ded3d971e1933c1eef91193f5e/pynndescent-0.5.13.tar.gz", hash = "sha256:d74254c0ee0a1eeec84597d5fe89fedcf778593eeabe32c2f97412934a9800fb", size = 2975955 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/d2/53/d23a97e0a2c690d40b165d1062e2c4ccc796be458a1ce59f6ba030434663/pynndescent-0.5.13-py3-none-any.whl", hash = "sha256:69aabb8f394bc631b6ac475a1c7f3994c54adf3f51cd63b2730fefba5771b949", size = 56850 }, -] - -[[package]] -name = "pyparsing" -version = "3.2.3" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/bb/22/f1129e69d94ffff626bdb5c835506b3a5b4f3d070f17ea295e12c2c6f60f/pyparsing-3.2.3.tar.gz", hash = "sha256:b9c13f1ab8b3b542f72e28f634bad4de758ab3ce4546e4301970ad6fa77c38be", size = 1088608 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/05/e7/df2285f3d08fee213f2d041540fa4fc9ca6c2d44cf36d3a035bf2a8d2bcc/pyparsing-3.2.3-py3-none-any.whl", hash = "sha256:a749938e02d6fd0b59b356ca504a24982314bb090c383e3cf201c95ef7e2bfcf", size = 111120 }, -] - -[[package]] -name = "python-dateutil" -version = "2.9.0.post0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "six" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/66/c0/0c8b6ad9f17a802ee498c46e004a0eb49bc148f2fd230864601a86dcf6db/python-dateutil-2.9.0.post0.tar.gz", hash = "sha256:37dd54208da7e1cd875388217d5e00ebd4179249f90fb72437e91a35459a0ad3", size = 342432 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/ec/57/56b9bcc3c9c6a792fcbaf139543cee77261f3651ca9da0c93f5c1221264b/python_dateutil-2.9.0.post0-py2.py3-none-any.whl", hash = "sha256:a8b2bc7bffae282281c8140a97d3aa9c14da0b136dfe83f850eea9a5f7470427", size = 229892 }, -] - -[[package]] -name = "python-json-logger" -version = "3.3.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/9e/de/d3144a0bceede957f961e975f3752760fbe390d57fbe194baf709d8f1f7b/python_json_logger-3.3.0.tar.gz", hash = "sha256:12b7e74b17775e7d565129296105bbe3910842d9d0eb083fc83a6a617aa8df84", size = 16642 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/08/20/0f2523b9e50a8052bc6a8b732dfc8568abbdc42010aef03a2d750bdab3b2/python_json_logger-3.3.0-py3-none-any.whl", hash = "sha256:dd980fae8cffb24c13caf6e158d3d61c0d6d22342f932cb6e9deedab3d35eec7", size = 15163 }, -] - -[[package]] -name = "pytorch-lightning" -version = "2.5.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "fsspec", extra = ["http"] }, - { name = "lightning-utilities" }, - { name = "packaging" }, - { name = "pyyaml" }, - { name = "torch" }, - { name = "torchmetrics" }, - { name = "tqdm" }, - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/01/3e/728fbdc671d07727ad447f9401d98a43570573965beb3cb2060f9a330b4f/pytorch_lightning-2.5.2.tar.gz", hash = "sha256:f817087d611be8d43b777dd4e543d72703e235510936677a13e6c29f7fd790e3", size = 636859 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/e2/42/47c186c8f9e956e559c89e6c764d5d5d0d0af517c04ca0ad39bd0a357d3a/pytorch_lightning-2.5.2-py3-none-any.whl", hash = "sha256:17cfdf89bd98074e389101f097cdf34c486a1f5c6d3fdcefbaf4dea7f97ff0bf", size = 825366 }, -] - -[[package]] -name = "pytorch-metric-learning" -version = "2.9.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "numpy" }, - { name = "scikit-learn" }, - { name = "torch" }, - { name = "tqdm" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/9b/80/6e61b1a91debf4c1b47d441f9a9d7fe2aabcdd9575ed70b2811474eb95c3/pytorch-metric-learning-2.9.0.tar.gz", hash = "sha256:27a626caf5e2876a0fd666605a78cb67ef7597e25d7a68c18053dd503830701f", size = 84530 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/46/7d/73ef5052f57b7720cad00e16598db3592a5ef4826745ffca67a2f085d4dc/pytorch_metric_learning-2.9.0-py3-none-any.whl", hash = "sha256:d51646006dc87168f00cf954785db133a4c5aac81253877248737aa42ef6432a", size = 127801 }, -] - -[[package]] -name = "pytorch-wpe" -version = "0.0.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "numpy" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/db/39/8d94737fd6fab4028687575099566a125100f3ba8c638f861506747d7b7c/pytorch_wpe-0.0.1.tar.gz", hash = "sha256:fc7e706b5411800c4483fe94db7dcd82ecf6c57bc013af529ab4fb675c9cc29c", size = 4457 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/23/de/c47967a11bfe68cb28d2f19e55c7027993c3721eba79813db65d245e4ced/pytorch_wpe-0.0.1-py3-none-any.whl", hash = "sha256:fa0dc9f818fba81b36c1a51a53331cf6ed975f29b33f23e07b0deb4bee82eaad", size = 8080 }, -] - -[[package]] -name = "pytz" -version = "2025.2" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/f8/bf/abbd3cdfb8fbc7fb3d4d38d320f2441b1e7cbe29be4f23797b4a2b5d8aac/pytz-2025.2.tar.gz", hash = "sha256:360b9e3dbb49a209c21ad61809c7fb453643e048b38924c765813546746e81c3", size = 320884 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/81/c4/34e93fe5f5429d7570ec1fa436f1986fb1f00c3e0f43a589fe2bbcd22c3f/pytz-2025.2-py2.py3-none-any.whl", hash = "sha256:5ddf76296dd8c44c26eb8f4b6f35488f3ccbf6fbbd7adee0b7262d43f0ec2f00", size = 509225 }, -] - -[[package]] -name = "pywin32" -version = "311" -source = { registry = "https://pypi.org/simple" } -wheels = [ - { url = "https://files.pythonhosted.org/packages/7b/40/44efbb0dfbd33aca6a6483191dae0716070ed99e2ecb0c53683f400a0b4f/pywin32-311-cp310-cp310-win32.whl", hash = "sha256:d03ff496d2a0cd4a5893504789d4a15399133fe82517455e78bad62efbb7f0a3", size = 8760432 }, - { url = "https://files.pythonhosted.org/packages/5e/bf/360243b1e953bd254a82f12653974be395ba880e7ec23e3731d9f73921cc/pywin32-311-cp310-cp310-win_amd64.whl", hash = "sha256:797c2772017851984b97180b0bebe4b620bb86328e8a884bb626156295a63b3b", size = 9590103 }, - { url = "https://files.pythonhosted.org/packages/57/38/d290720e6f138086fb3d5ffe0b6caa019a791dd57866940c82e4eeaf2012/pywin32-311-cp310-cp310-win_arm64.whl", hash = "sha256:0502d1facf1fed4839a9a51ccbcc63d952cf318f78ffc00a7e78528ac27d7a2b", size = 8778557 }, -] - -[[package]] -name = "pywinpty" -version = "3.0.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/06/df/429cc505dc5f77ab0612c4b60bca2e3dcc81f6c321844ee017d6dc0f4a95/pywinpty-3.0.0.tar.gz", hash = "sha256:68f70e68a9f0766ffdea3fc500351cb7b9b012bcb8239a411f7ff0fc8f86dcb1", size = 28551 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/15/f9/13d62974debb0c74ce3fa3d96b32cee6fce4f2d634789217e67aebf339f6/pywinpty-3.0.0-cp310-cp310-win_amd64.whl", hash = "sha256:327b6034e0dc38352c1c99a7c0b3e54941b4e506a5f21acce63609cd2ab6cce2", size = 2050843 }, -] - -[[package]] -name = "pyyaml" -version = "6.0.2" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/54/ed/79a089b6be93607fa5cdaedf301d7dfb23af5f25c398d5ead2525b063e17/pyyaml-6.0.2.tar.gz", hash = "sha256:d584d9ec91ad65861cc08d42e834324ef890a082e591037abe114850ff7bbc3e", size = 130631 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/9b/95/a3fac87cb7158e231b5a6012e438c647e1a87f09f8e0d123acec8ab8bf71/PyYAML-6.0.2-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:0a9a2848a5b7feac301353437eb7d5957887edbf81d56e903999a75a3d743086", size = 184199 }, - { url = "https://files.pythonhosted.org/packages/c7/7a/68bd47624dab8fd4afbfd3c48e3b79efe09098ae941de5b58abcbadff5cb/PyYAML-6.0.2-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:29717114e51c84ddfba879543fb232a6ed60086602313ca38cce623c1d62cfbf", size = 171758 }, - { url = "https://files.pythonhosted.org/packages/49/ee/14c54df452143b9ee9f0f29074d7ca5516a36edb0b4cc40c3f280131656f/PyYAML-6.0.2-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8824b5a04a04a047e72eea5cec3bc266db09e35de6bdfe34c9436ac5ee27d237", size = 718463 }, - { url = "https://files.pythonhosted.org/packages/4d/61/de363a97476e766574650d742205be468921a7b532aa2499fcd886b62530/PyYAML-6.0.2-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:7c36280e6fb8385e520936c3cb3b8042851904eba0e58d277dca80a5cfed590b", size = 719280 }, - { url = "https://files.pythonhosted.org/packages/6b/4e/1523cb902fd98355e2e9ea5e5eb237cbc5f3ad5f3075fa65087aa0ecb669/PyYAML-6.0.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ec031d5d2feb36d1d1a24380e4db6d43695f3748343d99434e6f5f9156aaa2ed", size = 751239 }, - { url = "https://files.pythonhosted.org/packages/b7/33/5504b3a9a4464893c32f118a9cc045190a91637b119a9c881da1cf6b7a72/PyYAML-6.0.2-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:936d68689298c36b53b29f23c6dbb74de12b4ac12ca6cfe0e047bedceea56180", size = 695802 }, - { url = "https://files.pythonhosted.org/packages/5c/20/8347dcabd41ef3a3cdc4f7b7a2aff3d06598c8779faa189cdbf878b626a4/PyYAML-6.0.2-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:23502f431948090f597378482b4812b0caae32c22213aecf3b55325e049a6c68", size = 720527 }, - { url = "https://files.pythonhosted.org/packages/be/aa/5afe99233fb360d0ff37377145a949ae258aaab831bde4792b32650a4378/PyYAML-6.0.2-cp310-cp310-win32.whl", hash = "sha256:2e99c6826ffa974fe6e27cdb5ed0021786b03fc98e5ee3c5bfe1fd5015f42b99", size = 144052 }, - { url = "https://files.pythonhosted.org/packages/b5/84/0fa4b06f6d6c958d207620fc60005e241ecedceee58931bb20138e1e5776/PyYAML-6.0.2-cp310-cp310-win_amd64.whl", hash = "sha256:a4d3091415f010369ae4ed1fc6b79def9416358877534caf6a0fdd2146c87a3e", size = 161774 }, -] - -[[package]] -name = "pyzmq" -version = "27.0.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "cffi", marker = "implementation_name == 'pypy'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/f1/06/50a4e9648b3e8b992bef8eb632e457307553a89d294103213cfd47b3da69/pyzmq-27.0.0.tar.gz", hash = "sha256:b1f08eeb9ce1510e6939b6e5dcd46a17765e2333daae78ecf4606808442e52cf", size = 280478 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/9c/09/1681d4b047626d352c083770618ac29655ab1f5c20eee31dc94c000b9b7b/pyzmq-27.0.0-cp310-cp310-macosx_10_15_universal2.whl", hash = "sha256:b973ee650e8f442ce482c1d99ca7ab537c69098d53a3d046676a484fd710c87a", size = 1329291 }, - { url = "https://files.pythonhosted.org/packages/9d/b2/9c9385225fdd54db9506ed8accbb9ea63ca813ba59d43d7f282a6a16a30b/pyzmq-27.0.0-cp310-cp310-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:661942bc7cd0223d569d808f2e5696d9cc120acc73bf3e88a1f1be7ab648a7e4", size = 905952 }, - { url = "https://files.pythonhosted.org/packages/41/73/333c72c7ec182cdffe25649e3da1c3b9f3cf1cede63cfdc23d1384d4a601/pyzmq-27.0.0-cp310-cp310-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:50360fb2a056ffd16e5f4177eee67f1dd1017332ea53fb095fe7b5bf29c70246", size = 666165 }, - { url = "https://files.pythonhosted.org/packages/a5/fe/fc7b9c1a50981928e25635a926653cb755364316db59ccd6e79cfb9a0b4f/pyzmq-27.0.0-cp310-cp310-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:cf209a6dc4b420ed32a7093642843cbf8703ed0a7d86c16c0b98af46762ebefb", size = 853755 }, - { url = "https://files.pythonhosted.org/packages/8c/4c/740ed4b6e8fa160cd19dc5abec8db68f440564b2d5b79c1d697d9862a2f7/pyzmq-27.0.0-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:c2dace4a7041cca2fba5357a2d7c97c5effdf52f63a1ef252cfa496875a3762d", size = 1654868 }, - { url = "https://files.pythonhosted.org/packages/97/00/875b2ecfcfc78ab962a59bd384995186818524ea957dc8ad3144611fae12/pyzmq-27.0.0-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:63af72b2955fc77caf0a77444baa2431fcabb4370219da38e1a9f8d12aaebe28", size = 2033443 }, - { url = "https://files.pythonhosted.org/packages/60/55/6dd9c470c42d713297c5f2a56f7903dc1ebdb4ab2edda996445c21651900/pyzmq-27.0.0-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:e8c4adce8e37e75c4215297d7745551b8dcfa5f728f23ce09bf4e678a9399413", size = 1891288 }, - { url = "https://files.pythonhosted.org/packages/28/5d/54b0ef50d40d7c65a627f4a4b4127024ba9820f2af8acd933a4d30ae192e/pyzmq-27.0.0-cp310-cp310-win32.whl", hash = "sha256:5d5ef4718ecab24f785794e0e7536436698b459bfbc19a1650ef55280119d93b", size = 567936 }, - { url = "https://files.pythonhosted.org/packages/18/ea/dedca4321de748ca48d3bcdb72274d4d54e8d84ea49088d3de174bd45d88/pyzmq-27.0.0-cp310-cp310-win_amd64.whl", hash = "sha256:e40609380480b3d12c30f841323f42451c755b8fece84235236f5fe5ffca8c1c", size = 628686 }, - { url = "https://files.pythonhosted.org/packages/d4/a7/fcdeedc306e71e94ac262cba2d02337d885f5cdb7e8efced8e5ffe327808/pyzmq-27.0.0-cp310-cp310-win_arm64.whl", hash = "sha256:6b0397b0be277b46762956f576e04dc06ced265759e8c2ff41a0ee1aa0064198", size = 559039 }, - { url = "https://files.pythonhosted.org/packages/93/a7/9ad68f55b8834ede477842214feba6a4c786d936c022a67625497aacf61d/pyzmq-27.0.0-cp312-abi3-macosx_10_15_universal2.whl", hash = "sha256:cbabc59dcfaac66655c040dfcb8118f133fb5dde185e5fc152628354c1598e52", size = 1305438 }, - { url = "https://files.pythonhosted.org/packages/ba/ee/26aa0f98665a22bc90ebe12dced1de5f3eaca05363b717f6fb229b3421b3/pyzmq-27.0.0-cp312-abi3-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:cb0ac5179cba4b2f94f1aa208fbb77b62c4c9bf24dd446278b8b602cf85fcda3", size = 895095 }, - { url = "https://files.pythonhosted.org/packages/cf/85/c57e7ab216ecd8aa4cc7e3b83b06cc4e9cf45c87b0afc095f10cd5ce87c1/pyzmq-27.0.0-cp312-abi3-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:53a48f0228eab6cbf69fde3aa3c03cbe04e50e623ef92ae395fce47ef8a76152", size = 651826 }, - { url = "https://files.pythonhosted.org/packages/69/9a/9ea7e230feda9400fb0ae0d61d7d6ddda635e718d941c44eeab22a179d34/pyzmq-27.0.0-cp312-abi3-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:111db5f395e09f7e775f759d598f43cb815fc58e0147623c4816486e1a39dc22", size = 839750 }, - { url = "https://files.pythonhosted.org/packages/08/66/4cebfbe71f3dfbd417011daca267539f62ed0fbc68105357b68bbb1a25b7/pyzmq-27.0.0-cp312-abi3-musllinux_1_2_aarch64.whl", hash = "sha256:c8878011653dcdc27cc2c57e04ff96f0471e797f5c19ac3d7813a245bcb24371", size = 1641357 }, - { url = "https://files.pythonhosted.org/packages/ac/f6/b0f62578c08d2471c791287149cb8c2aaea414ae98c6e995c7dbe008adfb/pyzmq-27.0.0-cp312-abi3-musllinux_1_2_i686.whl", hash = "sha256:c0ed2c1f335ba55b5fdc964622254917d6b782311c50e138863eda409fbb3b6d", size = 2020281 }, - { url = "https://files.pythonhosted.org/packages/37/b9/4f670b15c7498495da9159edc374ec09c88a86d9cd5a47d892f69df23450/pyzmq-27.0.0-cp312-abi3-musllinux_1_2_x86_64.whl", hash = "sha256:e918d70862d4cfd4b1c187310015646a14e1f5917922ab45b29f28f345eeb6be", size = 1877110 }, - { url = "https://files.pythonhosted.org/packages/66/31/9dee25c226295b740609f0d46db2fe972b23b6f5cf786360980524a3ba92/pyzmq-27.0.0-cp312-abi3-win32.whl", hash = "sha256:88b4e43cab04c3c0f0d55df3b1eef62df2b629a1a369b5289a58f6fa8b07c4f4", size = 559297 }, - { url = "https://files.pythonhosted.org/packages/9b/12/52da5509800f7ff2d287b2f2b4e636e7ea0f001181cba6964ff6c1537778/pyzmq-27.0.0-cp312-abi3-win_amd64.whl", hash = "sha256:dce4199bf5f648a902ce37e7b3afa286f305cd2ef7a8b6ec907470ccb6c8b371", size = 619203 }, - { url = "https://files.pythonhosted.org/packages/93/6d/7f2e53b19d1edb1eb4f09ec7c3a1f945ca0aac272099eab757d15699202b/pyzmq-27.0.0-cp312-abi3-win_arm64.whl", hash = "sha256:56e46bbb85d52c1072b3f809cc1ce77251d560bc036d3a312b96db1afe76db2e", size = 551927 }, - { url = "https://files.pythonhosted.org/packages/09/6f/be6523a7f3821c0b5370912ef02822c028611360e0d206dd945bdbf9eaef/pyzmq-27.0.0-pp310-pypy310_pp73-macosx_10_15_x86_64.whl", hash = "sha256:656c1866505a5735d0660b7da6d7147174bbf59d4975fc2b7f09f43c9bc25745", size = 835950 }, - { url = "https://files.pythonhosted.org/packages/c6/1e/a50fdd5c15018de07ab82a61bc460841be967ee7bbe7abee3b714d66f7ac/pyzmq-27.0.0-pp310-pypy310_pp73-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:74175b9e12779382432dd1d1f5960ebe7465d36649b98a06c6b26be24d173fab", size = 799876 }, - { url = "https://files.pythonhosted.org/packages/88/a1/89eb5b71f5a504f8f887aceb8e1eb3626e00c00aa8085381cdff475440dc/pyzmq-27.0.0-pp310-pypy310_pp73-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:d8c6de908465697a8708e4d6843a1e884f567962fc61eb1706856545141d0cbb", size = 567400 }, - { url = "https://files.pythonhosted.org/packages/56/aa/4571dbcff56cfb034bac73fde8294e123c975ce3eea89aff31bf6dc6382b/pyzmq-27.0.0-pp310-pypy310_pp73-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:c644aaacc01d0df5c7072826df45e67301f191c55f68d7b2916d83a9ddc1b551", size = 747031 }, - { url = "https://files.pythonhosted.org/packages/46/e0/d25f30fe0991293c5b2f5ef3b070d35fa6d57c0c7428898c3ab4913d0297/pyzmq-27.0.0-pp310-pypy310_pp73-win_amd64.whl", hash = "sha256:10f70c1d9a446a85013a36871a296007f6fe4232b530aa254baf9da3f8328bc0", size = 544726 }, -] - -[[package]] -name = "rapidfuzz" -version = "3.13.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/ed/f6/6895abc3a3d056b9698da3199b04c0e56226d530ae44a470edabf8b664f0/rapidfuzz-3.13.0.tar.gz", hash = "sha256:d2eaf3839e52cbcc0accbe9817a67b4b0fcf70aaeb229cfddc1c28061f9ce5d8", size = 57904226 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/de/27/ca10b3166024ae19a7e7c21f73c58dfd4b7fef7420e5497ee64ce6b73453/rapidfuzz-3.13.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:aafc42a1dc5e1beeba52cd83baa41372228d6d8266f6d803c16dbabbcc156255", size = 1998899 }, - { url = "https://files.pythonhosted.org/packages/f0/38/c4c404b13af0315483a6909b3a29636e18e1359307fb74a333fdccb3730d/rapidfuzz-3.13.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:85c9a131a44a95f9cac2eb6e65531db014e09d89c4f18c7b1fa54979cb9ff1f3", size = 1449949 }, - { url = "https://files.pythonhosted.org/packages/12/ae/15c71d68a6df6b8e24595421fdf5bcb305888318e870b7be8d935a9187ee/rapidfuzz-3.13.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7d7cec4242d30dd521ef91c0df872e14449d1dffc2a6990ede33943b0dae56c3", size = 1424199 }, - { url = "https://files.pythonhosted.org/packages/dc/9a/765beb9e14d7b30d12e2d6019e8b93747a0bedbc1d0cce13184fa3825426/rapidfuzz-3.13.0-cp310-cp310-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e297c09972698c95649e89121e3550cee761ca3640cd005e24aaa2619175464e", size = 5352400 }, - { url = "https://files.pythonhosted.org/packages/e2/b8/49479fe6f06b06cd54d6345ed16de3d1ac659b57730bdbe897df1e059471/rapidfuzz-3.13.0-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ef0f5f03f61b0e5a57b1df7beafd83df993fd5811a09871bad6038d08e526d0d", size = 1652465 }, - { url = "https://files.pythonhosted.org/packages/6f/d8/08823d496b7dd142a7b5d2da04337df6673a14677cfdb72f2604c64ead69/rapidfuzz-3.13.0-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:d8cf5f7cd6e4d5eb272baf6a54e182b2c237548d048e2882258336533f3f02b7", size = 1616590 }, - { url = "https://files.pythonhosted.org/packages/38/d4/5cfbc9a997e544f07f301c54d42aac9e0d28d457d543169e4ec859b8ce0d/rapidfuzz-3.13.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9256218ac8f1a957806ec2fb9a6ddfc6c32ea937c0429e88cf16362a20ed8602", size = 3086956 }, - { url = "https://files.pythonhosted.org/packages/25/1e/06d8932a72fa9576095234a15785136407acf8f9a7dbc8136389a3429da1/rapidfuzz-3.13.0-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:e1bdd2e6d0c5f9706ef7595773a81ca2b40f3b33fd7f9840b726fb00c6c4eb2e", size = 2494220 }, - { url = "https://files.pythonhosted.org/packages/03/16/5acf15df63119d5ca3d9a54b82807866ff403461811d077201ca351a40c3/rapidfuzz-3.13.0-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:5280be8fd7e2bee5822e254fe0a5763aa0ad57054b85a32a3d9970e9b09bbcbf", size = 7585481 }, - { url = "https://files.pythonhosted.org/packages/e1/cf/ebade4009431ea8e715e59e882477a970834ddaacd1a670095705b86bd0d/rapidfuzz-3.13.0-cp310-cp310-musllinux_1_2_ppc64le.whl", hash = "sha256:fd742c03885db1fce798a1cd87a20f47f144ccf26d75d52feb6f2bae3d57af05", size = 2894842 }, - { url = "https://files.pythonhosted.org/packages/a7/bd/0732632bd3f906bf613229ee1b7cbfba77515db714a0e307becfa8a970ae/rapidfuzz-3.13.0-cp310-cp310-musllinux_1_2_s390x.whl", hash = "sha256:5435fcac94c9ecf0504bf88a8a60c55482c32e18e108d6079a0089c47f3f8cf6", size = 3438517 }, - { url = "https://files.pythonhosted.org/packages/83/89/d3bd47ec9f4b0890f62aea143a1e35f78f3d8329b93d9495b4fa8a3cbfc3/rapidfuzz-3.13.0-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:93a755266856599be4ab6346273f192acde3102d7aa0735e2f48b456397a041f", size = 4412773 }, - { url = "https://files.pythonhosted.org/packages/b3/57/1a152a07883e672fc117c7f553f5b933f6e43c431ac3fd0e8dae5008f481/rapidfuzz-3.13.0-cp310-cp310-win32.whl", hash = "sha256:3abe6a4e8eb4cfc4cda04dd650a2dc6d2934cbdeda5def7e6fd1c20f6e7d2a0b", size = 1842334 }, - { url = "https://files.pythonhosted.org/packages/a7/68/7248addf95b6ca51fc9d955161072285da3059dd1472b0de773cff910963/rapidfuzz-3.13.0-cp310-cp310-win_amd64.whl", hash = "sha256:e8ddb58961401da7d6f55f185512c0d6bd24f529a637078d41dd8ffa5a49c107", size = 1624392 }, - { url = "https://files.pythonhosted.org/packages/68/23/f41c749f2c61ed1ed5575eaf9e73ef9406bfedbf20a3ffa438d15b5bf87e/rapidfuzz-3.13.0-cp310-cp310-win_arm64.whl", hash = "sha256:c523620d14ebd03a8d473c89e05fa1ae152821920c3ff78b839218ff69e19ca3", size = 865584 }, - { url = "https://files.pythonhosted.org/packages/d5/e1/f5d85ae3c53df6f817ca70dbdd37c83f31e64caced5bb867bec6b43d1fdf/rapidfuzz-3.13.0-pp310-pypy310_pp73-macosx_10_15_x86_64.whl", hash = "sha256:fe5790a36d33a5d0a6a1f802aa42ecae282bf29ac6f7506d8e12510847b82a45", size = 1904437 }, - { url = "https://files.pythonhosted.org/packages/db/d7/ded50603dddc5eb182b7ce547a523ab67b3bf42b89736f93a230a398a445/rapidfuzz-3.13.0-pp310-pypy310_pp73-macosx_11_0_arm64.whl", hash = "sha256:cdb33ee9f8a8e4742c6b268fa6bd739024f34651a06b26913381b1413ebe7590", size = 1383126 }, - { url = "https://files.pythonhosted.org/packages/c4/48/6f795e793babb0120b63a165496d64f989b9438efbeed3357d9a226ce575/rapidfuzz-3.13.0-pp310-pypy310_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8c99b76b93f7b495eee7dcb0d6a38fb3ce91e72e99d9f78faa5664a881cb2b7d", size = 1365565 }, - { url = "https://files.pythonhosted.org/packages/f0/50/0062a959a2d72ed17815824e40e2eefdb26f6c51d627389514510a7875f3/rapidfuzz-3.13.0-pp310-pypy310_pp73-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:6af42f2ede8b596a6aaf6d49fdee3066ca578f4856b85ab5c1e2145de367a12d", size = 5251719 }, - { url = "https://files.pythonhosted.org/packages/e7/02/bd8b70cd98b7a88e1621264778ac830c9daa7745cd63e838bd773b1aeebd/rapidfuzz-3.13.0-pp310-pypy310_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6c0efa73afbc5b265aca0d8a467ae2a3f40d6854cbe1481cb442a62b7bf23c99", size = 2991095 }, - { url = "https://files.pythonhosted.org/packages/9f/8d/632d895cdae8356826184864d74a5f487d40cb79f50a9137510524a1ba86/rapidfuzz-3.13.0-pp310-pypy310_pp73-win_amd64.whl", hash = "sha256:7ac21489de962a4e2fc1e8f0b0da4aa1adc6ab9512fd845563fecb4b4c52093a", size = 1553888 }, -] - -[[package]] -name = "referencing" -version = "0.36.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "attrs" }, - { name = "rpds-py" }, - { name = "typing-extensions", marker = "python_full_version == '3.10.12'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/2f/db/98b5c277be99dd18bfd91dd04e1b759cad18d1a338188c936e92f921c7e2/referencing-0.36.2.tar.gz", hash = "sha256:df2e89862cd09deabbdba16944cc3f10feb6b3e6f18e902f7cc25609a34775aa", size = 74744 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/c1/b1/3baf80dc6d2b7bc27a95a67752d0208e410351e3feb4eb78de5f77454d8d/referencing-0.36.2-py3-none-any.whl", hash = "sha256:e8699adbbf8b5c7de96d8ffa0eb5c158b3beafce084968e2ea8bb08c6794dcd0", size = 26775 }, -] - -[[package]] -name = "regex" -version = "2024.11.6" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/8e/5f/bd69653fbfb76cf8604468d3b4ec4c403197144c7bfe0e6a5fc9e02a07cb/regex-2024.11.6.tar.gz", hash = "sha256:7ab159b063c52a0333c884e4679f8d7a85112ee3078fe3d9004b2dd875585519", size = 399494 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/95/3c/4651f6b130c6842a8f3df82461a8950f923925db8b6961063e82744bddcc/regex-2024.11.6-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:ff590880083d60acc0433f9c3f713c51f7ac6ebb9adf889c79a261ecf541aa91", size = 482674 }, - { url = "https://files.pythonhosted.org/packages/15/51/9f35d12da8434b489c7b7bffc205c474a0a9432a889457026e9bc06a297a/regex-2024.11.6-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:658f90550f38270639e83ce492f27d2c8d2cd63805c65a13a14d36ca126753f0", size = 287684 }, - { url = "https://files.pythonhosted.org/packages/bd/18/b731f5510d1b8fb63c6b6d3484bfa9a59b84cc578ac8b5172970e05ae07c/regex-2024.11.6-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:164d8b7b3b4bcb2068b97428060b2a53be050085ef94eca7f240e7947f1b080e", size = 284589 }, - { url = "https://files.pythonhosted.org/packages/78/a2/6dd36e16341ab95e4c6073426561b9bfdeb1a9c9b63ab1b579c2e96cb105/regex-2024.11.6-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d3660c82f209655a06b587d55e723f0b813d3a7db2e32e5e7dc64ac2a9e86fde", size = 782511 }, - { url = "https://files.pythonhosted.org/packages/1b/2b/323e72d5d2fd8de0d9baa443e1ed70363ed7e7b2fb526f5950c5cb99c364/regex-2024.11.6-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:d22326fcdef5e08c154280b71163ced384b428343ae16a5ab2b3354aed12436e", size = 821149 }, - { url = "https://files.pythonhosted.org/packages/90/30/63373b9ea468fbef8a907fd273e5c329b8c9535fee36fc8dba5fecac475d/regex-2024.11.6-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:f1ac758ef6aebfc8943560194e9fd0fa18bcb34d89fd8bd2af18183afd8da3a2", size = 809707 }, - { url = "https://files.pythonhosted.org/packages/f2/98/26d3830875b53071f1f0ae6d547f1d98e964dd29ad35cbf94439120bb67a/regex-2024.11.6-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:997d6a487ff00807ba810e0f8332c18b4eb8d29463cfb7c820dc4b6e7562d0cf", size = 781702 }, - { url = "https://files.pythonhosted.org/packages/87/55/eb2a068334274db86208ab9d5599ffa63631b9f0f67ed70ea7c82a69bbc8/regex-2024.11.6-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:02a02d2bb04fec86ad61f3ea7f49c015a0681bf76abb9857f945d26159d2968c", size = 771976 }, - { url = "https://files.pythonhosted.org/packages/74/c0/be707bcfe98254d8f9d2cff55d216e946f4ea48ad2fd8cf1428f8c5332ba/regex-2024.11.6-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:f02f93b92358ee3f78660e43b4b0091229260c5d5c408d17d60bf26b6c900e86", size = 697397 }, - { url = "https://files.pythonhosted.org/packages/49/dc/bb45572ceb49e0f6509f7596e4ba7031f6819ecb26bc7610979af5a77f45/regex-2024.11.6-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:06eb1be98df10e81ebaded73fcd51989dcf534e3c753466e4b60c4697a003b67", size = 768726 }, - { url = "https://files.pythonhosted.org/packages/5a/db/f43fd75dc4c0c2d96d0881967897926942e935d700863666f3c844a72ce6/regex-2024.11.6-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:040df6fe1a5504eb0f04f048e6d09cd7c7110fef851d7c567a6b6e09942feb7d", size = 775098 }, - { url = "https://files.pythonhosted.org/packages/99/d7/f94154db29ab5a89d69ff893159b19ada89e76b915c1293e98603d39838c/regex-2024.11.6-cp310-cp310-musllinux_1_2_ppc64le.whl", hash = "sha256:fdabbfc59f2c6edba2a6622c647b716e34e8e3867e0ab975412c5c2f79b82da2", size = 839325 }, - { url = "https://files.pythonhosted.org/packages/f7/17/3cbfab1f23356fbbf07708220ab438a7efa1e0f34195bf857433f79f1788/regex-2024.11.6-cp310-cp310-musllinux_1_2_s390x.whl", hash = "sha256:8447d2d39b5abe381419319f942de20b7ecd60ce86f16a23b0698f22e1b70008", size = 843277 }, - { url = "https://files.pythonhosted.org/packages/7e/f2/48b393b51900456155de3ad001900f94298965e1cad1c772b87f9cfea011/regex-2024.11.6-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:da8f5fc57d1933de22a9e23eec290a0d8a5927a5370d24bda9a6abe50683fe62", size = 773197 }, - { url = "https://files.pythonhosted.org/packages/45/3f/ef9589aba93e084cd3f8471fded352826dcae8489b650d0b9b27bc5bba8a/regex-2024.11.6-cp310-cp310-win32.whl", hash = "sha256:b489578720afb782f6ccf2840920f3a32e31ba28a4b162e13900c3e6bd3f930e", size = 261714 }, - { url = "https://files.pythonhosted.org/packages/42/7e/5f1b92c8468290c465fd50c5318da64319133231415a8aa6ea5ab995a815/regex-2024.11.6-cp310-cp310-win_amd64.whl", hash = "sha256:5071b2093e793357c9d8b2929dfc13ac5f0a6c650559503bb81189d0a3814519", size = 274042 }, -] - -[[package]] -name = "requests" -version = "2.32.4" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "certifi" }, - { name = "charset-normalizer" }, - { name = "idna" }, - { name = "urllib3" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/e1/0a/929373653770d8a0d7ea76c37de6e41f11eb07559b103b1c02cafb3f7cf8/requests-2.32.4.tar.gz", hash = "sha256:27d0316682c8a29834d3264820024b62a36942083d52caf2f14c0591336d3422", size = 135258 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/7c/e4/56027c4a6b4ae70ca9de302488c5ca95ad4a39e190093d6c1a8ace08341b/requests-2.32.4-py3-none-any.whl", hash = "sha256:27babd3cda2a6d50b30443204ee89830707d396671944c998b5975b031ac2b2c", size = 64847 }, -] - -[[package]] -name = "resampy" -version = "0.4.3" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "numba" }, - { name = "numpy" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/29/f1/34be702a69a5d272e844c98cee82351f880985cfbca0cc86378011078497/resampy-0.4.3.tar.gz", hash = "sha256:a0d1c28398f0e55994b739650afef4e3974115edbe96cd4bb81968425e916e47", size = 3080604 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/4d/b9/3b00ac340a1aab3389ebcc52c779914a44aadf7b0cb7a3bf053195735607/resampy-0.4.3-py3-none-any.whl", hash = "sha256:ad2ed64516b140a122d96704e32bc0f92b23f45419e8b8f478e5a05f83edcebd", size = 3076529 }, -] - -[[package]] -name = "rfc3339-validator" -version = "0.1.4" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "six" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/28/ea/a9387748e2d111c3c2b275ba970b735e04e15cdb1eb30693b6b5708c4dbd/rfc3339_validator-0.1.4.tar.gz", hash = "sha256:138a2abdf93304ad60530167e51d2dfb9549521a836871b88d7f4695d0022f6b", size = 5513 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/7b/44/4e421b96b67b2daff264473f7465db72fbdf36a07e05494f50300cc7b0c6/rfc3339_validator-0.1.4-py2.py3-none-any.whl", hash = "sha256:24f6ec1eda14ef823da9e36ec7113124b39c04d50a4d3d3a3c2859577e7791fa", size = 3490 }, -] - -[[package]] -name = "rfc3986-validator" -version = "0.1.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/da/88/f270de456dd7d11dcc808abfa291ecdd3f45ff44e3b549ffa01b126464d0/rfc3986_validator-0.1.1.tar.gz", hash = "sha256:3d44bde7921b3b9ec3ae4e3adca370438eccebc676456449b145d533b240d055", size = 6760 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/9e/51/17023c0f8f1869d8806b979a2bffa3f861f26a3f1a66b094288323fba52f/rfc3986_validator-0.1.1-py2.py3-none-any.whl", hash = "sha256:2f235c432ef459970b4306369336b9d5dbdda31b510ca1e327636e01f528bfa9", size = 4242 }, -] - -[[package]] -name = "rich" -version = "14.0.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "markdown-it-py" }, - { name = "pygments" }, - { name = "typing-extensions", marker = "python_full_version == '3.10.12'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/a1/53/830aa4c3066a8ab0ae9a9955976fb770fe9c6102117c8ec4ab3ea62d89e8/rich-14.0.0.tar.gz", hash = "sha256:82f1bc23a6a21ebca4ae0c45af9bdbc492ed20231dcb63f297d6d1021a9d5725", size = 224078 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/0d/9b/63f4c7ebc259242c89b3acafdb37b41d1185c07ff0011164674e9076b491/rich-14.0.0-py3-none-any.whl", hash = "sha256:1c9491e1951aac09caffd42f448ee3d04e58923ffe14993f6e83068dc395d7e0", size = 243229 }, -] - -[[package]] -name = "rpds-py" -version = "0.25.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/8c/a6/60184b7fc00dd3ca80ac635dd5b8577d444c57e8e8742cecabfacb829921/rpds_py-0.25.1.tar.gz", hash = "sha256:8960b6dac09b62dac26e75d7e2c4a22efb835d827a7278c34f72b2b84fa160e3", size = 27304 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/cb/09/e1158988e50905b7f8306487a576b52d32aa9a87f79f7ab24ee8db8b6c05/rpds_py-0.25.1-cp310-cp310-macosx_10_12_x86_64.whl", hash = "sha256:f4ad628b5174d5315761b67f212774a32f5bad5e61396d38108bd801c0a8f5d9", size = 373140 }, - { url = "https://files.pythonhosted.org/packages/e0/4b/a284321fb3c45c02fc74187171504702b2934bfe16abab89713eedfe672e/rpds_py-0.25.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:8c742af695f7525e559c16f1562cf2323db0e3f0fbdcabdf6865b095256b2d40", size = 358860 }, - { url = "https://files.pythonhosted.org/packages/4e/46/8ac9811150c75edeae9fc6fa0e70376c19bc80f8e1f7716981433905912b/rpds_py-0.25.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:605ffe7769e24b1800b4d024d24034405d9404f0bc2f55b6db3362cd34145a6f", size = 386179 }, - { url = "https://files.pythonhosted.org/packages/f3/ec/87eb42d83e859bce91dcf763eb9f2ab117142a49c9c3d17285440edb5b69/rpds_py-0.25.1-cp310-cp310-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:ccc6f3ddef93243538be76f8e47045b4aad7a66a212cd3a0f23e34469473d36b", size = 400282 }, - { url = "https://files.pythonhosted.org/packages/68/c8/2a38e0707d7919c8c78e1d582ab15cf1255b380bcb086ca265b73ed6db23/rpds_py-0.25.1-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:f70316f760174ca04492b5ab01be631a8ae30cadab1d1081035136ba12738cfa", size = 521824 }, - { url = "https://files.pythonhosted.org/packages/5e/2c/6a92790243569784dde84d144bfd12bd45102f4a1c897d76375076d730ab/rpds_py-0.25.1-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:e1dafef8df605fdb46edcc0bf1573dea0d6d7b01ba87f85cd04dc855b2b4479e", size = 411644 }, - { url = "https://files.pythonhosted.org/packages/eb/76/66b523ffc84cf47db56efe13ae7cf368dee2bacdec9d89b9baca5e2e6301/rpds_py-0.25.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0701942049095741a8aeb298a31b203e735d1c61f4423511d2b1a41dcd8a16da", size = 386955 }, - { url = "https://files.pythonhosted.org/packages/b6/b9/a362d7522feaa24dc2b79847c6175daa1c642817f4a19dcd5c91d3e2c316/rpds_py-0.25.1-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:e87798852ae0b37c88babb7f7bbbb3e3fecc562a1c340195b44c7e24d403e380", size = 421039 }, - { url = "https://files.pythonhosted.org/packages/0f/c4/b5b6f70b4d719b6584716889fd3413102acf9729540ee76708d56a76fa97/rpds_py-0.25.1-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:3bcce0edc1488906c2d4c75c94c70a0417e83920dd4c88fec1078c94843a6ce9", size = 563290 }, - { url = "https://files.pythonhosted.org/packages/87/a3/2e6e816615c12a8f8662c9d8583a12eb54c52557521ef218cbe3095a8afa/rpds_py-0.25.1-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:e2f6a2347d3440ae789505693a02836383426249d5293541cd712e07e7aecf54", size = 592089 }, - { url = "https://files.pythonhosted.org/packages/c0/08/9b8e1050e36ce266135994e2c7ec06e1841f1c64da739daeb8afe9cb77a4/rpds_py-0.25.1-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:4fd52d3455a0aa997734f3835cbc4c9f32571345143960e7d7ebfe7b5fbfa3b2", size = 558400 }, - { url = "https://files.pythonhosted.org/packages/f2/df/b40b8215560b8584baccd839ff5c1056f3c57120d79ac41bd26df196da7e/rpds_py-0.25.1-cp310-cp310-win32.whl", hash = "sha256:3f0b1798cae2bbbc9b9db44ee068c556d4737911ad53a4e5093d09d04b3bbc24", size = 219741 }, - { url = "https://files.pythonhosted.org/packages/10/99/e4c58be18cf5d8b40b8acb4122bc895486230b08f978831b16a3916bd24d/rpds_py-0.25.1-cp310-cp310-win_amd64.whl", hash = "sha256:3ebd879ab996537fc510a2be58c59915b5dd63bccb06d1ef514fee787e05984a", size = 231553 }, - { url = "https://files.pythonhosted.org/packages/78/ff/566ce53529b12b4f10c0a348d316bd766970b7060b4fd50f888be3b3b281/rpds_py-0.25.1-pp310-pypy310_pp73-macosx_10_12_x86_64.whl", hash = "sha256:b24bf3cd93d5b6ecfbedec73b15f143596c88ee249fa98cefa9a9dc9d92c6f28", size = 373931 }, - { url = "https://files.pythonhosted.org/packages/83/5d/deba18503f7c7878e26aa696e97f051175788e19d5336b3b0e76d3ef9256/rpds_py-0.25.1-pp310-pypy310_pp73-macosx_11_0_arm64.whl", hash = "sha256:0eb90e94f43e5085623932b68840b6f379f26db7b5c2e6bcef3179bd83c9330f", size = 359074 }, - { url = "https://files.pythonhosted.org/packages/0d/74/313415c5627644eb114df49c56a27edba4d40cfd7c92bd90212b3604ca84/rpds_py-0.25.1-pp310-pypy310_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d50e4864498a9ab639d6d8854b25e80642bd362ff104312d9770b05d66e5fb13", size = 387255 }, - { url = "https://files.pythonhosted.org/packages/8c/c8/c723298ed6338963d94e05c0f12793acc9b91d04ed7c4ba7508e534b7385/rpds_py-0.25.1-pp310-pypy310_pp73-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:7c9409b47ba0650544b0bb3c188243b83654dfe55dcc173a86832314e1a6a35d", size = 400714 }, - { url = "https://files.pythonhosted.org/packages/33/8a/51f1f6aa653c2e110ed482ef2ae94140d56c910378752a1b483af11019ee/rpds_py-0.25.1-pp310-pypy310_pp73-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:796ad874c89127c91970652a4ee8b00d56368b7e00d3477f4415fe78164c8000", size = 523105 }, - { url = "https://files.pythonhosted.org/packages/c7/a4/7873d15c088ad3bff36910b29ceb0f178e4b3232c2adbe9198de68a41e63/rpds_py-0.25.1-pp310-pypy310_pp73-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:85608eb70a659bf4c1142b2781083d4b7c0c4e2c90eff11856a9754e965b2540", size = 411499 }, - { url = "https://files.pythonhosted.org/packages/90/f3/0ce1437befe1410766d11d08239333ac1b2d940f8a64234ce48a7714669c/rpds_py-0.25.1-pp310-pypy310_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c4feb9211d15d9160bc85fa72fed46432cdc143eb9cf6d5ca377335a921ac37b", size = 387918 }, - { url = "https://files.pythonhosted.org/packages/94/d4/5551247988b2a3566afb8a9dba3f1d4a3eea47793fd83000276c1a6c726e/rpds_py-0.25.1-pp310-pypy310_pp73-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:ccfa689b9246c48947d31dd9d8b16d89a0ecc8e0e26ea5253068efb6c542b76e", size = 421705 }, - { url = "https://files.pythonhosted.org/packages/b0/25/5960f28f847bf736cc7ee3c545a7e1d2f3b5edaf82c96fb616c2f5ed52d0/rpds_py-0.25.1-pp310-pypy310_pp73-musllinux_1_2_aarch64.whl", hash = "sha256:3c5b317ecbd8226887994852e85de562f7177add602514d4ac40f87de3ae45a8", size = 564489 }, - { url = "https://files.pythonhosted.org/packages/02/66/1c99884a0d44e8c2904d3c4ec302f995292d5dde892c3bf7685ac1930146/rpds_py-0.25.1-pp310-pypy310_pp73-musllinux_1_2_i686.whl", hash = "sha256:454601988aab2c6e8fd49e7634c65476b2b919647626208e376afcd22019eeb8", size = 592557 }, - { url = "https://files.pythonhosted.org/packages/55/ae/4aeac84ebeffeac14abb05b3bb1d2f728d00adb55d3fb7b51c9fa772e760/rpds_py-0.25.1-pp310-pypy310_pp73-musllinux_1_2_x86_64.whl", hash = "sha256:1c0c434a53714358532d13539272db75a5ed9df75a4a090a753ac7173ec14e11", size = 558691 }, - { url = "https://files.pythonhosted.org/packages/41/b3/728a08ff6f5e06fe3bb9af2e770e9d5fd20141af45cff8dfc62da4b2d0b3/rpds_py-0.25.1-pp310-pypy310_pp73-win_amd64.whl", hash = "sha256:f73ce1512e04fbe2bc97836e89830d6b4314c171587a99688082d090f934d20a", size = 231651 }, -] - -[[package]] -name = "ruamel-yaml" -version = "0.18.14" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "ruamel-yaml-clib", marker = "platform_python_implementation == 'CPython' and python_full_version == '3.10.12'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/39/87/6da0df742a4684263261c253f00edd5829e6aca970fff69e75028cccc547/ruamel.yaml-0.18.14.tar.gz", hash = "sha256:7227b76aaec364df15936730efbf7d72b30c0b79b1d578bbb8e3dcb2d81f52b7", size = 145511 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/af/6d/6fe4805235e193aad4aaf979160dd1f3c487c57d48b810c816e6e842171b/ruamel.yaml-0.18.14-py3-none-any.whl", hash = "sha256:710ff198bb53da66718c7db27eec4fbcc9aa6ca7204e4c1df2f282b6fe5eb6b2", size = 118570 }, -] - -[[package]] -name = "ruamel-yaml-clib" -version = "0.2.12" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/20/84/80203abff8ea4993a87d823a5f632e4d92831ef75d404c9fc78d0176d2b5/ruamel.yaml.clib-0.2.12.tar.gz", hash = "sha256:6c8fbb13ec503f99a91901ab46e0b07ae7941cd527393187039aec586fdfd36f", size = 225315 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/70/57/40a958e863e299f0c74ef32a3bde9f2d1ea8d69669368c0c502a0997f57f/ruamel.yaml.clib-0.2.12-cp310-cp310-macosx_13_0_arm64.whl", hash = "sha256:11f891336688faf5156a36293a9c362bdc7c88f03a8a027c2c1d8e0bcde998e5", size = 131301 }, - { url = "https://files.pythonhosted.org/packages/98/a8/29a3eb437b12b95f50a6bcc3d7d7214301c6c529d8fdc227247fa84162b5/ruamel.yaml.clib-0.2.12-cp310-cp310-manylinux2014_aarch64.whl", hash = "sha256:a606ef75a60ecf3d924613892cc603b154178ee25abb3055db5062da811fd969", size = 633728 }, - { url = "https://files.pythonhosted.org/packages/35/6d/ae05a87a3ad540259c3ad88d71275cbd1c0f2d30ae04c65dcbfb6dcd4b9f/ruamel.yaml.clib-0.2.12-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:fd5415dded15c3822597455bc02bcd66e81ef8b7a48cb71a33628fc9fdde39df", size = 722230 }, - { url = "https://files.pythonhosted.org/packages/7f/b7/20c6f3c0b656fe609675d69bc135c03aac9e3865912444be6339207b6648/ruamel.yaml.clib-0.2.12-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:f66efbc1caa63c088dead1c4170d148eabc9b80d95fb75b6c92ac0aad2437d76", size = 686712 }, - { url = "https://files.pythonhosted.org/packages/cd/11/d12dbf683471f888d354dac59593873c2b45feb193c5e3e0f2ebf85e68b9/ruamel.yaml.clib-0.2.12-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:22353049ba4181685023b25b5b51a574bce33e7f51c759371a7422dcae5402a6", size = 663936 }, - { url = "https://files.pythonhosted.org/packages/72/14/4c268f5077db5c83f743ee1daeb236269fa8577133a5cfa49f8b382baf13/ruamel.yaml.clib-0.2.12-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:932205970b9f9991b34f55136be327501903f7c66830e9760a8ffb15b07f05cd", size = 696580 }, - { url = "https://files.pythonhosted.org/packages/30/fc/8cd12f189c6405a4c1cf37bd633aa740a9538c8e40497c231072d0fef5cf/ruamel.yaml.clib-0.2.12-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:a52d48f4e7bf9005e8f0a89209bf9a73f7190ddf0489eee5eb51377385f59f2a", size = 663393 }, - { url = "https://files.pythonhosted.org/packages/80/29/c0a017b704aaf3cbf704989785cd9c5d5b8ccec2dae6ac0c53833c84e677/ruamel.yaml.clib-0.2.12-cp310-cp310-win32.whl", hash = "sha256:3eac5a91891ceb88138c113f9db04f3cebdae277f5d44eaa3651a4f573e6a5da", size = 100326 }, - { url = "https://files.pythonhosted.org/packages/3a/65/fa39d74db4e2d0cd252355732d966a460a41cd01c6353b820a0952432839/ruamel.yaml.clib-0.2.12-cp310-cp310-win_amd64.whl", hash = "sha256:ab007f2f5a87bd08ab1499bdf96f3d5c6ad4dcfa364884cb4549aa0154b13a28", size = 118079 }, -] - -[[package]] -name = "sacremoses" -version = "0.1.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "click" }, - { name = "joblib" }, - { name = "regex" }, - { name = "tqdm" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/1d/51/fbdc4af4f6e85d26169e28be3763fe50ddfd0d4bf8b871422b0788dcc4d2/sacremoses-0.1.1.tar.gz", hash = "sha256:b6fd5d3a766b02154ed80b962ddca91e1fd25629c0978c7efba21ebccf663934", size = 883188 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/0b/f0/89ee2bc9da434bd78464f288fdb346bc2932f2ee80a90b2a4bbbac262c74/sacremoses-0.1.1-py3-none-any.whl", hash = "sha256:31e04c98b169bfd902144824d191825cd69220cdb4ae4bcf1ec58a7db5587b1a", size = 897476 }, -] - -[[package]] -name = "safetensors" -version = "0.5.3" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/71/7e/2d5d6ee7b40c0682315367ec7475693d110f512922d582fef1bd4a63adc3/safetensors-0.5.3.tar.gz", hash = "sha256:b6b0d6ecacec39a4fdd99cc19f4576f5219ce858e6fd8dbe7609df0b8dc56965", size = 67210 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/18/ae/88f6c49dbd0cc4da0e08610019a3c78a7d390879a919411a410a1876d03a/safetensors-0.5.3-cp38-abi3-macosx_10_12_x86_64.whl", hash = "sha256:bd20eb133db8ed15b40110b7c00c6df51655a2998132193de2f75f72d99c7073", size = 436917 }, - { url = "https://files.pythonhosted.org/packages/b8/3b/11f1b4a2f5d2ab7da34ecc062b0bc301f2be024d110a6466726bec8c055c/safetensors-0.5.3-cp38-abi3-macosx_11_0_arm64.whl", hash = "sha256:21d01c14ff6c415c485616b8b0bf961c46b3b343ca59110d38d744e577f9cce7", size = 418419 }, - { url = "https://files.pythonhosted.org/packages/5d/9a/add3e6fef267658075c5a41573c26d42d80c935cdc992384dfae435feaef/safetensors-0.5.3-cp38-abi3-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:11bce6164887cd491ca75c2326a113ba934be596e22b28b1742ce27b1d076467", size = 459493 }, - { url = "https://files.pythonhosted.org/packages/df/5c/bf2cae92222513cc23b3ff85c4a1bb2811a2c3583ac0f8e8d502751de934/safetensors-0.5.3-cp38-abi3-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:4a243be3590bc3301c821da7a18d87224ef35cbd3e5f5727e4e0728b8172411e", size = 472400 }, - { url = "https://files.pythonhosted.org/packages/58/11/7456afb740bd45782d0f4c8e8e1bb9e572f1bf82899fb6ace58af47b4282/safetensors-0.5.3-cp38-abi3-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:8bd84b12b1670a6f8e50f01e28156422a2bc07fb16fc4e98bded13039d688a0d", size = 522891 }, - { url = "https://files.pythonhosted.org/packages/57/3d/fe73a9d2ace487e7285f6e157afee2383bd1ddb911b7cb44a55cf812eae3/safetensors-0.5.3-cp38-abi3-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:391ac8cab7c829452175f871fcaf414aa1e292b5448bd02620f675a7f3e7abb9", size = 537694 }, - { url = "https://files.pythonhosted.org/packages/a6/f8/dae3421624fcc87a89d42e1898a798bc7ff72c61f38973a65d60df8f124c/safetensors-0.5.3-cp38-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:cead1fa41fc54b1e61089fa57452e8834f798cb1dc7a09ba3524f1eb08e0317a", size = 471642 }, - { url = "https://files.pythonhosted.org/packages/ce/20/1fbe16f9b815f6c5a672f5b760951e20e17e43f67f231428f871909a37f6/safetensors-0.5.3-cp38-abi3-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:1077f3e94182d72618357b04b5ced540ceb71c8a813d3319f1aba448e68a770d", size = 502241 }, - { url = "https://files.pythonhosted.org/packages/5f/18/8e108846b506487aa4629fe4116b27db65c3dde922de2c8e0cc1133f3f29/safetensors-0.5.3-cp38-abi3-musllinux_1_2_aarch64.whl", hash = "sha256:799021e78287bac619c7b3f3606730a22da4cda27759ddf55d37c8db7511c74b", size = 638001 }, - { url = "https://files.pythonhosted.org/packages/82/5a/c116111d8291af6c8c8a8b40628fe833b9db97d8141c2a82359d14d9e078/safetensors-0.5.3-cp38-abi3-musllinux_1_2_armv7l.whl", hash = "sha256:df26da01aaac504334644e1b7642fa000bfec820e7cef83aeac4e355e03195ff", size = 734013 }, - { url = "https://files.pythonhosted.org/packages/7d/ff/41fcc4d3b7de837963622e8610d998710705bbde9a8a17221d85e5d0baad/safetensors-0.5.3-cp38-abi3-musllinux_1_2_i686.whl", hash = "sha256:32c3ef2d7af8b9f52ff685ed0bc43913cdcde135089ae322ee576de93eae5135", size = 670687 }, - { url = "https://files.pythonhosted.org/packages/40/ad/2b113098e69c985a3d8fbda4b902778eae4a35b7d5188859b4a63d30c161/safetensors-0.5.3-cp38-abi3-musllinux_1_2_x86_64.whl", hash = "sha256:37f1521be045e56fc2b54c606d4455573e717b2d887c579ee1dbba5f868ece04", size = 643147 }, - { url = "https://files.pythonhosted.org/packages/0a/0c/95aeb51d4246bd9a3242d3d8349c1112b4ee7611a4b40f0c5c93b05f001d/safetensors-0.5.3-cp38-abi3-win32.whl", hash = "sha256:cfc0ec0846dcf6763b0ed3d1846ff36008c6e7290683b61616c4b040f6a54ace", size = 296677 }, - { url = "https://files.pythonhosted.org/packages/69/e2/b011c38e5394c4c18fb5500778a55ec43ad6106126e74723ffaee246f56e/safetensors-0.5.3-cp38-abi3-win_amd64.whl", hash = "sha256:836cbbc320b47e80acd40e44c8682db0e8ad7123209f69b093def21ec7cafd11", size = 308878 }, -] - -[[package]] -name = "scikit-learn" -version = "1.5.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "joblib" }, - { name = "numpy" }, - { name = "scipy" }, - { name = "threadpoolctl" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/92/72/2961b9874a9ddf2b0f95f329d4e67f67c3301c1d88ba5e239ff25661bb85/scikit_learn-1.5.1.tar.gz", hash = "sha256:0ea5d40c0e3951df445721927448755d3fe1d80833b0b7308ebff5d2a45e6414", size = 6958368 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/f8/59/d8ea8c05e61d2afa988dfcfe47526595b531e94d23babf58d2e00a35f646/scikit_learn-1.5.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:781586c414f8cc58e71da4f3d7af311e0505a683e112f2f62919e3019abd3745", size = 12102257 }, - { url = "https://files.pythonhosted.org/packages/1f/c6/ba8e5691acca616adc8f0d6f8f5e79d55b927530aa404ee712b077acf0cf/scikit_learn-1.5.1-cp310-cp310-macosx_12_0_arm64.whl", hash = "sha256:f5b213bc29cc30a89a3130393b0e39c847a15d769d6e59539cd86b75d276b1a7", size = 10975310 }, - { url = "https://files.pythonhosted.org/packages/5c/c6/e362563cc7dfe37e4699cbf2b2d22c2854be227c254976de1c4854fc6e84/scikit_learn-1.5.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:1ff4ba34c2abff5ec59c803ed1d97d61b036f659a17f55be102679e88f926fac", size = 12496508 }, - { url = "https://files.pythonhosted.org/packages/f2/60/6c589c91e474721efdcec82ea9cc5c743359e52637e46c364ee5236666ef/scikit_learn-1.5.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:161808750c267b77b4a9603cf9c93579c7a74ba8486b1336034c2f1579546d21", size = 13352348 }, - { url = "https://files.pythonhosted.org/packages/f1/13/de29b945fb28fc0c24159d3a83f1250c5232c1c9abac12434c7c3447e9cc/scikit_learn-1.5.1-cp310-cp310-win_amd64.whl", hash = "sha256:10e49170691514a94bb2e03787aa921b82dbc507a4ea1f20fd95557862c98dc1", size = 10966250 }, -] - -[[package]] -name = "scipy" -version = "1.15.3" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "numpy" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/0f/37/6964b830433e654ec7485e45a00fc9a27cf868d622838f6b6d9c5ec0d532/scipy-1.15.3.tar.gz", hash = "sha256:eae3cf522bc7df64b42cad3925c876e1b0b6c35c1337c93e12c0f366f55b0eaf", size = 59419214 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/78/2f/4966032c5f8cc7e6a60f1b2e0ad686293b9474b65246b0c642e3ef3badd0/scipy-1.15.3-cp310-cp310-macosx_10_13_x86_64.whl", hash = "sha256:a345928c86d535060c9c2b25e71e87c39ab2f22fc96e9636bd74d1dbf9de448c", size = 38702770 }, - { url = "https://files.pythonhosted.org/packages/a0/6e/0c3bf90fae0e910c274db43304ebe25a6b391327f3f10b5dcc638c090795/scipy-1.15.3-cp310-cp310-macosx_12_0_arm64.whl", hash = "sha256:ad3432cb0f9ed87477a8d97f03b763fd1d57709f1bbde3c9369b1dff5503b253", size = 30094511 }, - { url = "https://files.pythonhosted.org/packages/ea/b1/4deb37252311c1acff7f101f6453f0440794f51b6eacb1aad4459a134081/scipy-1.15.3-cp310-cp310-macosx_14_0_arm64.whl", hash = "sha256:aef683a9ae6eb00728a542b796f52a5477b78252edede72b8327a886ab63293f", size = 22368151 }, - { url = "https://files.pythonhosted.org/packages/38/7d/f457626e3cd3c29b3a49ca115a304cebb8cc6f31b04678f03b216899d3c6/scipy-1.15.3-cp310-cp310-macosx_14_0_x86_64.whl", hash = "sha256:1c832e1bd78dea67d5c16f786681b28dd695a8cb1fb90af2e27580d3d0967e92", size = 25121732 }, - { url = "https://files.pythonhosted.org/packages/db/0a/92b1de4a7adc7a15dcf5bddc6e191f6f29ee663b30511ce20467ef9b82e4/scipy-1.15.3-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:263961f658ce2165bbd7b99fa5135195c3a12d9bef045345016b8b50c315cb82", size = 35547617 }, - { url = "https://files.pythonhosted.org/packages/8e/6d/41991e503e51fc1134502694c5fa7a1671501a17ffa12716a4a9151af3df/scipy-1.15.3-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9e2abc762b0811e09a0d3258abee2d98e0c703eee49464ce0069590846f31d40", size = 37662964 }, - { url = "https://files.pythonhosted.org/packages/25/e1/3df8f83cb15f3500478c889be8fb18700813b95e9e087328230b98d547ff/scipy-1.15.3-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:ed7284b21a7a0c8f1b6e5977ac05396c0d008b89e05498c8b7e8f4a1423bba0e", size = 37238749 }, - { url = "https://files.pythonhosted.org/packages/93/3e/b3257cf446f2a3533ed7809757039016b74cd6f38271de91682aa844cfc5/scipy-1.15.3-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:5380741e53df2c566f4d234b100a484b420af85deb39ea35a1cc1be84ff53a5c", size = 40022383 }, - { url = "https://files.pythonhosted.org/packages/d1/84/55bc4881973d3f79b479a5a2e2df61c8c9a04fcb986a213ac9c02cfb659b/scipy-1.15.3-cp310-cp310-win_amd64.whl", hash = "sha256:9d61e97b186a57350f6d6fd72640f9e99d5a4a2b8fbf4b9ee9a841eab327dc13", size = 41259201 }, -] - -[[package]] -name = "seaborn" -version = "0.13.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "matplotlib" }, - { name = "numpy" }, - { name = "pandas" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/86/59/a451d7420a77ab0b98f7affa3a1d78a313d2f7281a57afb1a34bae8ab412/seaborn-0.13.2.tar.gz", hash = "sha256:93e60a40988f4d65e9f4885df477e2fdaff6b73a9ded434c1ab356dd57eefff7", size = 1457696 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/83/11/00d3c3dfc25ad54e731d91449895a79e4bf2384dc3ac01809010ba88f6d5/seaborn-0.13.2-py3-none-any.whl", hash = "sha256:636f8336facf092165e27924f223d3c62ca560b1f2bb5dff7ab7fad265361987", size = 294914 }, -] - -[[package]] -name = "semver" -version = "3.0.4" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/72/d1/d3159231aec234a59dd7d601e9dd9fe96f3afff15efd33c1070019b26132/semver-3.0.4.tar.gz", hash = "sha256:afc7d8c584a5ed0a11033af086e8af226a9c0b206f313e0301f8dd7b6b589602", size = 269730 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/a6/24/4d91e05817e92e3a61c8a21e08fd0f390f5301f1c448b137c57c4bc6e543/semver-3.0.4-py3-none-any.whl", hash = "sha256:9c824d87ba7f7ab4a1890799cec8596f15c1241cb473404ea1cb0c55e4b04746", size = 17912 }, -] - -[[package]] -name = "send2trash" -version = "1.8.3" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/fd/3a/aec9b02217bb79b87bbc1a21bc6abc51e3d5dcf65c30487ac96c0908c722/Send2Trash-1.8.3.tar.gz", hash = "sha256:b18e7a3966d99871aefeb00cfbcfdced55ce4871194810fc71f4aa484b953abf", size = 17394 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/40/b0/4562db6223154aa4e22f939003cb92514c79f3d4dccca3444253fd17f902/Send2Trash-1.8.3-py3-none-any.whl", hash = "sha256:0c31227e0bd08961c7665474a3d1ef7193929fedda4233843689baa056be46c9", size = 18072 }, -] - -[[package]] -name = "sentencepiece" -version = "0.2.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/c9/d2/b9c7ca067c26d8ff085d252c89b5f69609ca93fb85a00ede95f4857865d4/sentencepiece-0.2.0.tar.gz", hash = "sha256:a52c19171daaf2e697dc6cbe67684e0fa341b1248966f6aebb541de654d15843", size = 2632106 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/f6/71/98648c3b64b23edb5403f74bcc906ad21766872a6e1ada26ea3f1eb941ab/sentencepiece-0.2.0-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:188779e1298a1c8b8253c7d3ad729cb0a9891e5cef5e5d07ce4592c54869e227", size = 2408979 }, - { url = "https://files.pythonhosted.org/packages/77/9f/7efbaa6d4c0c718a9affbecc536b03ca62f99f421bdffb531c16030e2d2b/sentencepiece-0.2.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:bed9cf85b296fa2b76fc2547b9cbb691a523864cebaee86304c43a7b4cb1b452", size = 1238845 }, - { url = "https://files.pythonhosted.org/packages/1c/e4/c2541027a43ec6962ba9b601805d17ba3f86b38bdeae0e8ac65a2981e248/sentencepiece-0.2.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:d7b67e724bead13f18db6e1d10b6bbdc454af574d70efbb36f27d90387be1ca3", size = 1181472 }, - { url = "https://files.pythonhosted.org/packages/fd/46/316c1ba6c52b97de76aff7b9da678f7afbb52136afb2987c474d95630e65/sentencepiece-0.2.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:2fde4b08cfe237be4484c6c7c2e2c75fb862cfeab6bd5449ce4caeafd97b767a", size = 1259151 }, - { url = "https://files.pythonhosted.org/packages/aa/5a/3c48738a0835d76dd06c62b6ac48d39c923cde78dd0f587353bdcbb99851/sentencepiece-0.2.0-cp310-cp310-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:4c378492056202d1c48a4979650981635fd97875a00eabb1f00c6a236b013b5e", size = 1355931 }, - { url = "https://files.pythonhosted.org/packages/a6/27/33019685023221ca8ed98e8ceb7ae5e166032686fa3662c68f1f1edf334e/sentencepiece-0.2.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1380ce6540a368de2ef6d7e6ba14ba8f3258df650d39ba7d833b79ee68a52040", size = 1301537 }, - { url = "https://files.pythonhosted.org/packages/ca/e4/55f97cef14293171fef5f96e96999919ab5b4d1ce95b53547ad653d7e3bf/sentencepiece-0.2.0-cp310-cp310-win32.whl", hash = "sha256:a1151d6a6dd4b43e552394aed0edfe9292820272f0194bd56c7c1660a0c06c3d", size = 936747 }, - { url = "https://files.pythonhosted.org/packages/85/f4/4ef1a6e0e9dbd8a60780a91df8b7452ada14cfaa0e17b3b8dfa42cecae18/sentencepiece-0.2.0-cp310-cp310-win_amd64.whl", hash = "sha256:d490142b0521ef22bc1085f061d922a2a6666175bb6b42e588ff95c0db6819b2", size = 991525 }, -] - -[[package]] -name = "sentry-sdk" -version = "2.32.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "certifi" }, - { name = "urllib3" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/10/59/eb90c45cb836cf8bec973bba10230ddad1c55e2b2e9ffa9d7d7368948358/sentry_sdk-2.32.0.tar.gz", hash = "sha256:9016c75d9316b0f6921ac14c8cd4fb938f26002430ac5be9945ab280f78bec6b", size = 334932 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/01/a1/fc4856bd02d2097324fb7ce05b3021fb850f864b83ca765f6e37e92ff8ca/sentry_sdk-2.32.0-py2.py3-none-any.whl", hash = "sha256:6cf51521b099562d7ce3606da928c473643abe99b00ce4cb5626ea735f4ec345", size = 356122 }, -] - -[[package]] -name = "setproctitle" -version = "1.3.6" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/9e/af/56efe21c53ac81ac87e000b15e60b3d8104224b4313b6eacac3597bd183d/setproctitle-1.3.6.tar.gz", hash = "sha256:c9f32b96c700bb384f33f7cf07954bb609d35dd82752cef57fb2ee0968409169", size = 26889 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/7d/db/8214810cae49e2e474ea741aaa7d6111486f27377e864f0eb6d297c9be56/setproctitle-1.3.6-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:ebcf34b69df4ca0eabaaaf4a3d890f637f355fed00ba806f7ebdd2d040658c26", size = 17412 }, - { url = "https://files.pythonhosted.org/packages/a4/45/909b0dcd68b16d2e58de0e861c0c0b67748ccc87ff9b59136e9710b528b1/setproctitle-1.3.6-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:1aa1935aa2195b76f377e5cb018290376b7bf085f0b53f5a95c0c21011b74367", size = 11966 }, - { url = "https://files.pythonhosted.org/packages/8a/f4/f1cd54fedae1cdacf1d1db833dc096bfb1f029451f60e68563e4c26ed2f7/setproctitle-1.3.6-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:13624d9925bb481bc0ccfbc7f533da38bfbfe6e80652314f789abc78c2e513bd", size = 31350 }, - { url = "https://files.pythonhosted.org/packages/5a/5f/f159b22d286a349633d4090090b8e6632fb84575a64f189b68e70a613c65/setproctitle-1.3.6-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:97a138fa875c6f281df7720dac742259e85518135cd0e3551aba1c628103d853", size = 32704 }, - { url = "https://files.pythonhosted.org/packages/9c/25/e5ea2673d951dafc04b6544d7b33dd9283733d715c8f40e93d39ae35d6a0/setproctitle-1.3.6-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:c86e9e82bfab579327dbe9b82c71475165fbc8b2134d24f9a3b2edaf200a5c3d", size = 29833 }, - { url = "https://files.pythonhosted.org/packages/67/2b/c3cbd4a4462c1143465d8c151f1d51bbfb418e60a96a754329d28d416575/setproctitle-1.3.6-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6af330ddc2ec05a99c3933ab3cba9365357c0b8470a7f2fa054ee4b0984f57d1", size = 30884 }, - { url = "https://files.pythonhosted.org/packages/27/04/b43a622a9fbf0f216a50b523067d3b07739ede2106a7226223e33abf6659/setproctitle-1.3.6-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:109fc07b1cd6cef9c245b2028e3e98e038283342b220def311d0239179810dbe", size = 30798 }, - { url = "https://files.pythonhosted.org/packages/41/60/8eb197b56b0a3110eef2a1d2ddb61b3f5809dbab9d975aa40c86e5d4b312/setproctitle-1.3.6-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:7df5fcc48588f82b6cc8073db069609ddd48a49b1e9734a20d0efb32464753c4", size = 29758 }, - { url = "https://files.pythonhosted.org/packages/db/1d/c394322a5425c12f4ada0696eb6d194768752d4e3acaca0c9d593025feb4/setproctitle-1.3.6-cp310-cp310-musllinux_1_2_ppc64le.whl", hash = "sha256:2407955dc359d735a20ac6e797ad160feb33d529a2ac50695c11a1ec680eafab", size = 32157 }, - { url = "https://files.pythonhosted.org/packages/e7/24/ce080682b144f057814efbe95daac09149e90f7689e2515897817a941686/setproctitle-1.3.6-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:38ca045626af693da042ac35d7332e7b9dbd52e6351d6973b310612e3acee6d6", size = 30291 }, - { url = "https://files.pythonhosted.org/packages/a7/4f/4db265493567865428dcec376f8142a9c65d27c10c3ac915d173b4053afb/setproctitle-1.3.6-cp310-cp310-win32.whl", hash = "sha256:9483aa336687463f5497dd37a070094f3dff55e2c888994f8440fcf426a1a844", size = 11492 }, - { url = "https://files.pythonhosted.org/packages/38/b0/64c3948f7957db44b4c5edfe9c197de493144dc915ddf95cf36aeca0dc52/setproctitle-1.3.6-cp310-cp310-win_amd64.whl", hash = "sha256:4efc91b437f6ff2578e89e3f17d010c0a0ff01736606473d082913ecaf7859ba", size = 12204 }, - { url = "https://files.pythonhosted.org/packages/d0/2b/f19977b646b64c1440dade2c385c89c39f74ce5254defa102dfd9c163e0b/setproctitle-1.3.6-pp310-pypy310_pp73-macosx_11_0_arm64.whl", hash = "sha256:3cde5b83ec4915cd5e6ae271937fd60d14113c8f7769b4a20d51769fe70d8717", size = 11471 }, - { url = "https://files.pythonhosted.org/packages/78/46/db58cf700f1408cf0f63d3f939f7b077bd450da8e037310f70e74c41262f/setproctitle-1.3.6-pp310-pypy310_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:797f2846b546a8741413c57d9fb930ad5aa939d925c9c0fa6186d77580035af7", size = 13520 }, - { url = "https://files.pythonhosted.org/packages/5c/46/0b89e7ebe77543e721c67077ad93fc8c7c3c31a8db3b12e00d02950f6adc/setproctitle-1.3.6-pp310-pypy310_pp73-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4ac3eb04bcf0119aadc6235a2c162bae5ed5f740e3d42273a7228b915722de20", size = 13094 }, - { url = "https://files.pythonhosted.org/packages/f7/78/03f2e42eb83bce6f853d7751ae95f8a3ae7408145a0b6cdd717be01497d7/setproctitle-1.3.6-pp310-pypy310_pp73-win_amd64.whl", hash = "sha256:0e6b5633c94c5111f7137f875e8f1ff48f53b991d5d5b90932f27dc8c1fa9ae4", size = 12241 }, -] - -[[package]] -name = "setuptools" -version = "80.9.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/18/5d/3bf57dcd21979b887f014ea83c24ae194cfcd12b9e0fda66b957c69d1fca/setuptools-80.9.0.tar.gz", hash = "sha256:f36b47402ecde768dbfafc46e8e4207b4360c654f1f3bb84475f0a28628fb19c", size = 1319958 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/a3/dc/17031897dae0efacfea57dfd3a82fdd2a2aeb58e0ff71b77b87e44edc772/setuptools-80.9.0-py3-none-any.whl", hash = "sha256:062d34222ad13e0cc312a4c02d73f059e86a4acbfbdea8f8f76b28c99f306922", size = 1201486 }, -] - -[[package]] -name = "shellingham" -version = "1.5.4" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/58/15/8b3609fd3830ef7b27b655beb4b4e9c62313a4e8da8c676e142cc210d58e/shellingham-1.5.4.tar.gz", hash = "sha256:8dbca0739d487e5bd35ab3ca4b36e11c4078f3a234bfce294b0a0291363404de", size = 10310 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/e0/f9/0595336914c5619e5f28a1fb793285925a8cd4b432c9da0a987836c7f822/shellingham-1.5.4-py2.py3-none-any.whl", hash = "sha256:7ecfff8f2fd72616f7481040475a65b2bf8af90a56c89140852d1120324e8686", size = 9755 }, -] - -[[package]] -name = "six" -version = "1.17.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/94/e7/b2c673351809dca68a0e064b6af791aa332cf192da575fd474ed7d6f16a2/six-1.17.0.tar.gz", hash = "sha256:ff70335d468e7eb6ec65b95b99d3a2836546063f63acc5171de367e834932a81", size = 34031 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/b7/ce/149a00dd41f10bc29e5921b496af8b574d8413afcd5e30dfa0ed46c2cc5e/six-1.17.0-py2.py3-none-any.whl", hash = "sha256:4721f391ed90541fddacab5acf947aa0d3dc7d27b2e1e8eda2be8970586c3274", size = 11050 }, -] - -[[package]] -name = "smmap" -version = "5.0.2" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/44/cd/a040c4b3119bbe532e5b0732286f805445375489fceaec1f48306068ee3b/smmap-5.0.2.tar.gz", hash = "sha256:26ea65a03958fa0c8a1c7e8c7a58fdc77221b8910f6be2131affade476898ad5", size = 22329 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/04/be/d09147ad1ec7934636ad912901c5fd7667e1c858e19d355237db0d0cd5e4/smmap-5.0.2-py3-none-any.whl", hash = "sha256:b30115f0def7d7531d22a0fb6502488d879e75b260a9db4d0819cfb25403af5e", size = 24303 }, -] - -[[package]] -name = "sniffio" -version = "1.3.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/a2/87/a6771e1546d97e7e041b6ae58d80074f81b7d5121207425c964ddf5cfdbd/sniffio-1.3.1.tar.gz", hash = "sha256:f4324edc670a0f49750a81b895f35c3adb843cca46f0530f79fc1babb23789dc", size = 20372 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/e9/44/75a9c9421471a6c4805dbf2356f7c181a29c1879239abab1ea2cc8f38b40/sniffio-1.3.1-py3-none-any.whl", hash = "sha256:2f6da418d1f1e0fddd844478f41680e794e6051915791a034ff65e5f100525a2", size = 10235 }, -] - -[[package]] -name = "sortedcontainers" -version = "2.4.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/e8/c4/ba2f8066cceb6f23394729afe52f3bf7adec04bf9ed2c820b39e19299111/sortedcontainers-2.4.0.tar.gz", hash = "sha256:25caa5a06cc30b6b83d11423433f65d1f9d76c4c6a0c90e3379eaa43b9bfdb88", size = 30594 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/32/46/9cb0e58b2deb7f82b84065f37f3bffeb12413f947f9388e4cac22c4621ce/sortedcontainers-2.4.0-py2.py3-none-any.whl", hash = "sha256:a163dcaede0f1c021485e957a39245190e74249897e2ae4b2aa38595db237ee0", size = 29575 }, -] - -[[package]] -name = "soundfile" -version = "0.13.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "cffi" }, - { name = "numpy" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/e1/41/9b873a8c055582859b239be17902a85339bec6a30ad162f98c9b0288a2cc/soundfile-0.13.1.tar.gz", hash = "sha256:b2c68dab1e30297317080a5b43df57e302584c49e2942defdde0acccc53f0e5b", size = 46156 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/64/28/e2a36573ccbcf3d57c00626a21fe51989380636e821b341d36ccca0c1c3a/soundfile-0.13.1-py2.py3-none-any.whl", hash = "sha256:a23c717560da2cf4c7b5ae1142514e0fd82d6bbd9dfc93a50423447142f2c445", size = 25751 }, - { url = "https://files.pythonhosted.org/packages/ea/ab/73e97a5b3cc46bba7ff8650a1504348fa1863a6f9d57d7001c6b67c5f20e/soundfile-0.13.1-py2.py3-none-macosx_10_9_x86_64.whl", hash = "sha256:82dc664d19831933fe59adad199bf3945ad06d84bc111a5b4c0d3089a5b9ec33", size = 1142250 }, - { url = "https://files.pythonhosted.org/packages/a0/e5/58fd1a8d7b26fc113af244f966ee3aecf03cb9293cb935daaddc1e455e18/soundfile-0.13.1-py2.py3-none-macosx_11_0_arm64.whl", hash = "sha256:743f12c12c4054921e15736c6be09ac26b3b3d603aef6fd69f9dde68748f2593", size = 1101406 }, - { url = "https://files.pythonhosted.org/packages/58/ae/c0e4a53d77cf6e9a04179535766b3321b0b9ced5f70522e4caf9329f0046/soundfile-0.13.1-py2.py3-none-manylinux_2_28_aarch64.whl", hash = "sha256:9c9e855f5a4d06ce4213f31918653ab7de0c5a8d8107cd2427e44b42df547deb", size = 1235729 }, - { url = "https://files.pythonhosted.org/packages/57/5e/70bdd9579b35003a489fc850b5047beeda26328053ebadc1fb60f320f7db/soundfile-0.13.1-py2.py3-none-manylinux_2_28_x86_64.whl", hash = "sha256:03267c4e493315294834a0870f31dbb3b28a95561b80b134f0bd3cf2d5f0e618", size = 1313646 }, - { url = "https://files.pythonhosted.org/packages/fe/df/8c11dc4dfceda14e3003bb81a0d0edcaaf0796dd7b4f826ea3e532146bba/soundfile-0.13.1-py2.py3-none-win32.whl", hash = "sha256:c734564fab7c5ddf8e9be5bf70bab68042cd17e9c214c06e365e20d64f9a69d5", size = 899881 }, - { url = "https://files.pythonhosted.org/packages/14/e9/6b761de83277f2f02ded7e7ea6f07828ec78e4b229b80e4ca55dd205b9dc/soundfile-0.13.1-py2.py3-none-win_amd64.whl", hash = "sha256:1e70a05a0626524a69e9f0f4dd2ec174b4e9567f4d8b6c11d38b5c289be36ee9", size = 1019162 }, -] - -[[package]] -name = "soupsieve" -version = "2.7" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/3f/f4/4a80cd6ef364b2e8b65b15816a843c0980f7a5a2b4dc701fc574952aa19f/soupsieve-2.7.tar.gz", hash = "sha256:ad282f9b6926286d2ead4750552c8a6142bc4c783fd66b0293547c8fe6ae126a", size = 103418 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/e7/9c/0e6afc12c269578be5c0c1c9f4b49a8d32770a080260c333ac04cc1c832d/soupsieve-2.7-py3-none-any.whl", hash = "sha256:6e60cc5c1ffaf1cebcc12e8188320b72071e922c2e897f737cadce79ad5d30c4", size = 36677 }, -] - -[[package]] -name = "sox" -version = "1.5.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "numpy" }, - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/3d/a2/d8e0d8fd7abf509ead4a2cb0fb24e5758b5330166bf9223d5cb9f98a7e8d/sox-1.5.0.tar.gz", hash = "sha256:12c7be5bb1f548d891fe11e82c08cf5f1a1d74e225298f60082e5aeb2469ada0", size = 63905 } - -[[package]] -name = "soxr" -version = "0.5.0.post1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "numpy" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/02/c0/4429bf9b3be10e749149e286aa5c53775399ec62891c6b970456c6dca325/soxr-0.5.0.post1.tar.gz", hash = "sha256:7092b9f3e8a416044e1fa138c8172520757179763b85dc53aa9504f4813cff73", size = 170853 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/7d/96/bee1eb69d66fc28c3b219ba9b8674b49d3dcc6cd2f9b3e5114ff28cf88b5/soxr-0.5.0.post1-cp310-cp310-macosx_10_14_x86_64.whl", hash = "sha256:7406d782d85f8cf64e66b65e6b7721973de8a1dc50b9e88bc2288c343a987484", size = 203841 }, - { url = "https://files.pythonhosted.org/packages/1f/5d/56ad3d181d30d103128f65cc44f4c4e24c199e6d5723e562704e47c89f78/soxr-0.5.0.post1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:fa0a382fb8d8e2afed2c1642723b2d2d1b9a6728ff89f77f3524034c8885b8c9", size = 160192 }, - { url = "https://files.pythonhosted.org/packages/7f/09/e43c39390e26b4c1b8d46f8a1c252a5077fa9f81cc2326b03c3d2b85744e/soxr-0.5.0.post1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8b01d3efb95a2851f78414bcd00738b0253eec3f5a1e5482838e965ffef84969", size = 221176 }, - { url = "https://files.pythonhosted.org/packages/ba/e6/059070b4cdb7fdd8ffbb67c5087c1da9716577127fb0540cd11dbf77923b/soxr-0.5.0.post1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:fcc049b0a151a65aa75b92f0ac64bb2dba785d16b78c31c2b94e68c141751d6d", size = 252779 }, - { url = "https://files.pythonhosted.org/packages/ad/64/86082b6372e5ff807dfa79b857da9f50e94e155706000daa43fdc3b59851/soxr-0.5.0.post1-cp310-cp310-win_amd64.whl", hash = "sha256:97f269bc26937c267a2ace43a77167d0c5c8bba5a2b45863bb6042b5b50c474e", size = 166881 }, - { url = "https://files.pythonhosted.org/packages/5d/e3/d422d279e51e6932e7b64f1170a4f61a7ee768e0f84c9233a5b62cd2c832/soxr-0.5.0.post1-cp312-abi3-macosx_10_14_x86_64.whl", hash = "sha256:fef509466c9c25f65eae0ce1e4b9ac9705d22c6038c914160ddaf459589c6e31", size = 199993 }, - { url = "https://files.pythonhosted.org/packages/20/f1/88adaca3c52e03bcb66b63d295df2e2d35bf355d19598c6ce84b20be7fca/soxr-0.5.0.post1-cp312-abi3-macosx_11_0_arm64.whl", hash = "sha256:4704ba6b13a3f1e41d12acf192878384c1c31f71ce606829c64abdf64a8d7d32", size = 156373 }, - { url = "https://files.pythonhosted.org/packages/b8/38/bad15a9e615215c8219652ca554b601663ac3b7ac82a284aca53ec2ff48c/soxr-0.5.0.post1-cp312-abi3-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:bd052a66471a7335b22a6208601a9d0df7b46b8d087dce4ff6e13eed6a33a2a1", size = 216564 }, - { url = "https://files.pythonhosted.org/packages/e1/1a/569ea0420a0c4801c2c8dd40d8d544989522f6014d51def689125f3f2935/soxr-0.5.0.post1-cp312-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a3f16810dd649ab1f433991d2a9661e9e6a116c2b4101039b53b3c3e90a094fc", size = 248455 }, - { url = "https://files.pythonhosted.org/packages/bc/10/440f1ba3d4955e0dc740bbe4ce8968c254a3d644d013eb75eea729becdb8/soxr-0.5.0.post1-cp312-abi3-win_amd64.whl", hash = "sha256:b1be9fee90afb38546bdbd7bde714d1d9a8c5a45137f97478a83b65e7f3146f6", size = 164937 }, -] - -[[package]] -name = "speechbrain" -version = "1.0.3" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "huggingface-hub" }, - { name = "hyperpyyaml" }, - { name = "joblib" }, - { name = "numpy" }, - { name = "packaging" }, - { name = "scipy" }, - { name = "sentencepiece" }, - { name = "torch" }, - { name = "torchaudio" }, - { name = "tqdm" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/ab/10/87e666544a4e0cec7cbdc09f26948994831ae0f8bbc58de3bf53b68285ff/speechbrain-1.0.3.tar.gz", hash = "sha256:fcab3c6e90012cecb1eed40ea235733b550137e73da6bfa2340ba191ec714052", size = 747735 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/58/13/e61f1085aebee17d5fc2df19fcc5177c10379be52578afbecdd615a831c9/speechbrain-1.0.3-py3-none-any.whl", hash = "sha256:9859d4c1b1fb3af3b85523c0c89f52e45a04f305622ed55f31aa32dd2fba19e9", size = 864091 }, -] - -[[package]] -name = "sqlalchemy" -version = "2.0.41" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "greenlet", marker = "(platform_machine == 'AMD64' and python_full_version == '3.10.12') or (platform_machine == 'WIN32' and python_full_version == '3.10.12') or (platform_machine == 'aarch64' and python_full_version == '3.10.12') or (platform_machine == 'amd64' and python_full_version == '3.10.12') or (platform_machine == 'ppc64le' and python_full_version == '3.10.12') or (platform_machine == 'win32' and python_full_version == '3.10.12') or (platform_machine == 'x86_64' and python_full_version == '3.10.12')" }, - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/63/66/45b165c595ec89aa7dcc2c1cd222ab269bc753f1fc7a1e68f8481bd957bf/sqlalchemy-2.0.41.tar.gz", hash = "sha256:edba70118c4be3c2b1f90754d308d0b79c6fe2c0fdc52d8ddf603916f83f4db9", size = 9689424 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/e9/12/d7c445b1940276a828efce7331cb0cb09d6e5f049651db22f4ebb0922b77/sqlalchemy-2.0.41-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:b1f09b6821406ea1f94053f346f28f8215e293344209129a9c0fcc3578598d7b", size = 2117967 }, - { url = "https://files.pythonhosted.org/packages/6f/b8/cb90f23157e28946b27eb01ef401af80a1fab7553762e87df51507eaed61/sqlalchemy-2.0.41-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:1936af879e3db023601196a1684d28e12f19ccf93af01bf3280a3262c4b6b4e5", size = 2107583 }, - { url = "https://files.pythonhosted.org/packages/9e/c2/eef84283a1c8164a207d898e063edf193d36a24fb6a5bb3ce0634b92a1e8/sqlalchemy-2.0.41-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:b2ac41acfc8d965fb0c464eb8f44995770239668956dc4cdf502d1b1ffe0d747", size = 3186025 }, - { url = "https://files.pythonhosted.org/packages/bd/72/49d52bd3c5e63a1d458fd6d289a1523a8015adedbddf2c07408ff556e772/sqlalchemy-2.0.41-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:81c24e0c0fde47a9723c81d5806569cddef103aebbf79dbc9fcbb617153dea30", size = 3186259 }, - { url = "https://files.pythonhosted.org/packages/4f/9e/e3ffc37d29a3679a50b6bbbba94b115f90e565a2b4545abb17924b94c52d/sqlalchemy-2.0.41-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:23a8825495d8b195c4aa9ff1c430c28f2c821e8c5e2d98089228af887e5d7e29", size = 3126803 }, - { url = "https://files.pythonhosted.org/packages/8a/76/56b21e363f6039978ae0b72690237b38383e4657281285a09456f313dd77/sqlalchemy-2.0.41-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:60c578c45c949f909a4026b7807044e7e564adf793537fc762b2489d522f3d11", size = 3148566 }, - { url = "https://files.pythonhosted.org/packages/3b/92/11b8e1b69bf191bc69e300a99badbbb5f2f1102f2b08b39d9eee2e21f565/sqlalchemy-2.0.41-cp310-cp310-win32.whl", hash = "sha256:118c16cd3f1b00c76d69343e38602006c9cfb9998fa4f798606d28d63f23beda", size = 2086696 }, - { url = "https://files.pythonhosted.org/packages/5c/88/2d706c9cc4502654860f4576cd54f7db70487b66c3b619ba98e0be1a4642/sqlalchemy-2.0.41-cp310-cp310-win_amd64.whl", hash = "sha256:7492967c3386df69f80cf67efd665c0f667cee67032090fe01d7d74b0e19bb08", size = 2110200 }, - { url = "https://files.pythonhosted.org/packages/1c/fc/9ba22f01b5cdacc8f5ed0d22304718d2c758fce3fd49a5372b886a86f37c/sqlalchemy-2.0.41-py3-none-any.whl", hash = "sha256:57df5dc6fdb5ed1a88a1ed2195fd31927e705cad62dedd86b46972752a80f576", size = 1911224 }, -] - -[[package]] -name = "stack-data" -version = "0.6.3" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "asttokens" }, - { name = "executing" }, - { name = "pure-eval" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/28/e3/55dcc2cfbc3ca9c29519eb6884dd1415ecb53b0e934862d3559ddcb7e20b/stack_data-0.6.3.tar.gz", hash = "sha256:836a778de4fec4dcd1dcd89ed8abff8a221f58308462e1c4aa2a3cf30148f0b9", size = 44707 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/f1/7b/ce1eafaf1a76852e2ec9b22edecf1daa58175c090266e9f6c64afcd81d91/stack_data-0.6.3-py3-none-any.whl", hash = "sha256:d5558e0c25a4cb0853cddad3d77da9891a08cb85dd9f9f91b9f8cd66e511e695", size = 24521 }, -] - -[[package]] -name = "sympy" -version = "1.14.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "mpmath" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/83/d3/803453b36afefb7c2bb238361cd4ae6125a569b4db67cd9e79846ba2d68c/sympy-1.14.0.tar.gz", hash = "sha256:d3d3fe8df1e5a0b42f0e7bdf50541697dbe7d23746e894990c030e2b05e72517", size = 7793921 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/a2/09/77d55d46fd61b4a135c444fc97158ef34a095e5681d0a6c10b75bf356191/sympy-1.14.0-py3-none-any.whl", hash = "sha256:e091cc3e99d2141a0ba2847328f5479b05d94a6635cb96148ccb3f34671bd8f5", size = 6299353 }, -] - -[[package]] -name = "tabulate" -version = "0.9.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/ec/fe/802052aecb21e3797b8f7902564ab6ea0d60ff8ca23952079064155d1ae1/tabulate-0.9.0.tar.gz", hash = "sha256:0095b12bf5966de529c0feb1fa08671671b3368eec77d7ef7ab114be2c068b3c", size = 81090 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/40/44/4a5f08c96eb108af5cb50b41f76142f0afa346dfa99d5296fe7202a11854/tabulate-0.9.0-py3-none-any.whl", hash = "sha256:024ca478df22e9340661486f85298cff5f6dcdba14f3813e8830015b9ed1948f", size = 35252 }, -] - -[[package]] -name = "tensorboard" -version = "2.19.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "absl-py" }, - { name = "grpcio" }, - { name = "markdown" }, - { name = "numpy" }, - { name = "packaging" }, - { name = "protobuf" }, - { name = "setuptools" }, - { name = "six" }, - { name = "tensorboard-data-server" }, - { name = "werkzeug" }, -] -wheels = [ - { url = "https://files.pythonhosted.org/packages/5d/12/4f70e8e2ba0dbe72ea978429d8530b0333f0ed2140cc571a48802878ef99/tensorboard-2.19.0-py3-none-any.whl", hash = "sha256:5e71b98663a641a7ce8a6e70b0be8e1a4c0c45d48760b076383ac4755c35b9a0", size = 5503412 }, -] - -[[package]] -name = "tensorboard-data-server" -version = "0.7.2" -source = { registry = "https://pypi.org/simple" } -wheels = [ - { url = "https://files.pythonhosted.org/packages/7a/13/e503968fefabd4c6b2650af21e110aa8466fe21432cd7c43a84577a89438/tensorboard_data_server-0.7.2-py3-none-any.whl", hash = "sha256:7e0610d205889588983836ec05dc098e80f97b7e7bbff7e994ebb78f578d0ddb", size = 2356 }, - { url = "https://files.pythonhosted.org/packages/b7/85/dabeaf902892922777492e1d253bb7e1264cadce3cea932f7ff599e53fea/tensorboard_data_server-0.7.2-py3-none-macosx_10_9_x86_64.whl", hash = "sha256:9fe5d24221b29625dbc7328b0436ca7fc1c23de4acf4d272f1180856e32f9f60", size = 4823598 }, - { url = "https://files.pythonhosted.org/packages/73/c6/825dab04195756cf8ff2e12698f22513b3db2f64925bdd41671bfb33aaa5/tensorboard_data_server-0.7.2-py3-none-manylinux_2_31_x86_64.whl", hash = "sha256:ef687163c24185ae9754ed5650eb5bc4d84ff257aabdc33f0cc6f74d8ba54530", size = 6590363 }, -] - -[[package]] -name = "tensorboardx" -version = "2.6.4" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "numpy" }, - { name = "packaging" }, - { name = "protobuf" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/2b/c5/d4cc6e293fb837aaf9f76dd7745476aeba8ef7ef5146c3b3f9ee375fe7a5/tensorboardx-2.6.4.tar.gz", hash = "sha256:b163ccb7798b31100b9f5fa4d6bc22dad362d7065c2f24b51e50731adde86828", size = 4769801 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/e0/1d/b5d63f1a6b824282b57f7b581810d20b7a28ca951f2d5b59f1eb0782c12b/tensorboardx-2.6.4-py3-none-any.whl", hash = "sha256:5970cf3a1f0a6a6e8b180ccf46f3fe832b8a25a70b86e5a237048a7c0beb18e2", size = 87201 }, -] - -[[package]] -name = "termcolor" -version = "3.1.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/ca/6c/3d75c196ac07ac8749600b60b03f4f6094d54e132c4d94ebac6ee0e0add0/termcolor-3.1.0.tar.gz", hash = "sha256:6a6dd7fbee581909eeec6a756cff1d7f7c376063b14e4a298dc4980309e55970", size = 14324 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/4f/bd/de8d508070629b6d84a30d01d57e4a65c69aa7f5abe7560b8fad3b50ea59/termcolor-3.1.0-py3-none-any.whl", hash = "sha256:591dd26b5c2ce03b9e43f391264626557873ce1d379019786f99b0c2bee140aa", size = 7684 }, -] - -[[package]] -name = "terminado" -version = "0.18.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "ptyprocess", marker = "os_name != 'nt'" }, - { name = "pywinpty", marker = "os_name == 'nt'" }, - { name = "tornado" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/8a/11/965c6fd8e5cc254f1fe142d547387da17a8ebfd75a3455f637c663fb38a0/terminado-0.18.1.tar.gz", hash = "sha256:de09f2c4b85de4765f7714688fff57d3e75bad1f909b589fde880460c753fd2e", size = 32701 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/6a/9e/2064975477fdc887e47ad42157e214526dcad8f317a948dee17e1659a62f/terminado-0.18.1-py3-none-any.whl", hash = "sha256:a4468e1b37bb318f8a86514f65814e1afc977cf29b3992a4500d9dd305dcceb0", size = 14154 }, -] - -[[package]] -name = "text-unidecode" -version = "1.3" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/ab/e2/e9a00f0ccb71718418230718b3d900e71a5d16e701a3dae079a21e9cd8f8/text-unidecode-1.3.tar.gz", hash = "sha256:bad6603bb14d279193107714b288be206cac565dfa49aa5b105294dd5c4aab93", size = 76885 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/a6/a5/c0b6468d3824fe3fde30dbb5e1f687b291608f9473681bbf7dabbf5a87d7/text_unidecode-1.3-py2.py3-none-any.whl", hash = "sha256:1311f10e8b895935241623731c2ba64f4c455287888b18189350b67134a822e8", size = 78154 }, -] - -[[package]] -name = "texterrors" -version = "0.5.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "levenshtein" }, - { name = "loguru" }, - { name = "numpy" }, - { name = "plac" }, - { name = "pybind11" }, - { name = "regex" }, - { name = "termcolor" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/9b/47/9a391643961698df3c804172f005e8b56c9693c14c4170abd9d3c961e971/texterrors-0.5.1.tar.gz", hash = "sha256:7fa24b2ca6ed5e05681b5cfdbb6c1fd0e4ae6518f8939e9782294f620d4eb3b1", size = 23813 } - -[[package]] -name = "threadpoolctl" -version = "3.6.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/b7/4d/08c89e34946fce2aec4fbb45c9016efd5f4d7f24af8e5d93296e935631d8/threadpoolctl-3.6.0.tar.gz", hash = "sha256:8ab8b4aa3491d812b623328249fab5302a68d2d71745c8a4c719a2fcaba9f44e", size = 21274 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/32/d5/f9a850d79b0851d1d4ef6456097579a9005b31fea68726a4ae5f2d82ddd9/threadpoolctl-3.6.0-py3-none-any.whl", hash = "sha256:43a0b8fd5a2928500110039e43a5eed8480b918967083ea48dc3ab9f13c4a7fb", size = 18638 }, -] - -[[package]] -name = "tinycss2" -version = "1.4.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "webencodings" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/7a/fd/7a5ee21fd08ff70d3d33a5781c255cbe779659bd03278feb98b19ee550f4/tinycss2-1.4.0.tar.gz", hash = "sha256:10c0972f6fc0fbee87c3edb76549357415e94548c1ae10ebccdea16fb404a9b7", size = 87085 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/e6/34/ebdc18bae6aa14fbee1a08b63c015c72b64868ff7dae68808ab500c492e2/tinycss2-1.4.0-py3-none-any.whl", hash = "sha256:3a49cf47b7675da0b15d0c6e1df8df4ebd96e9394bb905a5775adb0d884c5289", size = 26610 }, -] - -[[package]] -name = "tokenizers" -version = "0.21.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "huggingface-hub" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/ab/2d/b0fce2b8201635f60e8c95990080f58461cc9ca3d5026de2e900f38a7f21/tokenizers-0.21.2.tar.gz", hash = "sha256:fdc7cffde3e2113ba0e6cc7318c40e3438a4d74bbc62bf04bcc63bdfb082ac77", size = 351545 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/1d/cc/2936e2d45ceb130a21d929743f1e9897514691bec123203e10837972296f/tokenizers-0.21.2-cp39-abi3-macosx_10_12_x86_64.whl", hash = "sha256:342b5dfb75009f2255ab8dec0041287260fed5ce00c323eb6bab639066fef8ec", size = 2875206 }, - { url = "https://files.pythonhosted.org/packages/6c/e6/33f41f2cc7861faeba8988e7a77601407bf1d9d28fc79c5903f8f77df587/tokenizers-0.21.2-cp39-abi3-macosx_11_0_arm64.whl", hash = "sha256:126df3205d6f3a93fea80c7a8a266a78c1bd8dd2fe043386bafdd7736a23e45f", size = 2732655 }, - { url = "https://files.pythonhosted.org/packages/33/2b/1791eb329c07122a75b01035b1a3aa22ad139f3ce0ece1b059b506d9d9de/tokenizers-0.21.2-cp39-abi3-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:4a32cd81be21168bd0d6a0f0962d60177c447a1aa1b1e48fa6ec9fc728ee0b12", size = 3019202 }, - { url = "https://files.pythonhosted.org/packages/05/15/fd2d8104faa9f86ac68748e6f7ece0b5eb7983c7efc3a2c197cb98c99030/tokenizers-0.21.2-cp39-abi3-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:8bd8999538c405133c2ab999b83b17c08b7fc1b48c1ada2469964605a709ef91", size = 2934539 }, - { url = "https://files.pythonhosted.org/packages/a5/2e/53e8fd053e1f3ffbe579ca5f9546f35ac67cf0039ed357ad7ec57f5f5af0/tokenizers-0.21.2-cp39-abi3-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:5e9944e61239b083a41cf8fc42802f855e1dca0f499196df37a8ce219abac6eb", size = 3248665 }, - { url = "https://files.pythonhosted.org/packages/00/15/79713359f4037aa8f4d1f06ffca35312ac83629da062670e8830917e2153/tokenizers-0.21.2-cp39-abi3-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:514cd43045c5d546f01142ff9c79a96ea69e4b5cda09e3027708cb2e6d5762ab", size = 3451305 }, - { url = "https://files.pythonhosted.org/packages/38/5f/959f3a8756fc9396aeb704292777b84f02a5c6f25c3fc3ba7530db5feb2c/tokenizers-0.21.2-cp39-abi3-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:b1b9405822527ec1e0f7d8d2fdb287a5730c3a6518189c968254a8441b21faae", size = 3214757 }, - { url = "https://files.pythonhosted.org/packages/c5/74/f41a432a0733f61f3d21b288de6dfa78f7acff309c6f0f323b2833e9189f/tokenizers-0.21.2-cp39-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:fed9a4d51c395103ad24f8e7eb976811c57fbec2af9f133df471afcd922e5020", size = 3121887 }, - { url = "https://files.pythonhosted.org/packages/3c/6a/bc220a11a17e5d07b0dfb3b5c628621d4dcc084bccd27cfaead659963016/tokenizers-0.21.2-cp39-abi3-musllinux_1_2_aarch64.whl", hash = "sha256:2c41862df3d873665ec78b6be36fcc30a26e3d4902e9dd8608ed61d49a48bc19", size = 9091965 }, - { url = "https://files.pythonhosted.org/packages/6c/bd/ac386d79c4ef20dc6f39c4706640c24823dca7ebb6f703bfe6b5f0292d88/tokenizers-0.21.2-cp39-abi3-musllinux_1_2_armv7l.whl", hash = "sha256:ed21dc7e624e4220e21758b2e62893be7101453525e3d23264081c9ef9a6d00d", size = 9053372 }, - { url = "https://files.pythonhosted.org/packages/63/7b/5440bf203b2a5358f074408f7f9c42884849cd9972879e10ee6b7a8c3b3d/tokenizers-0.21.2-cp39-abi3-musllinux_1_2_i686.whl", hash = "sha256:0e73770507e65a0e0e2a1affd6b03c36e3bc4377bd10c9ccf51a82c77c0fe365", size = 9298632 }, - { url = "https://files.pythonhosted.org/packages/a4/d2/faa1acac3f96a7427866e94ed4289949b2524f0c1878512516567d80563c/tokenizers-0.21.2-cp39-abi3-musllinux_1_2_x86_64.whl", hash = "sha256:106746e8aa9014a12109e58d540ad5465b4c183768ea96c03cbc24c44d329958", size = 9470074 }, - { url = "https://files.pythonhosted.org/packages/d8/a5/896e1ef0707212745ae9f37e84c7d50269411aef2e9ccd0de63623feecdf/tokenizers-0.21.2-cp39-abi3-win32.whl", hash = "sha256:cabda5a6d15d620b6dfe711e1af52205266d05b379ea85a8a301b3593c60e962", size = 2330115 }, - { url = "https://files.pythonhosted.org/packages/13/c3/cc2755ee10be859c4338c962a35b9a663788c0c0b50c0bdd8078fb6870cf/tokenizers-0.21.2-cp39-abi3-win_amd64.whl", hash = "sha256:58747bb898acdb1007f37a7bbe614346e98dc28708ffb66a3fd50ce169ac6c98", size = 2509918 }, -] - -[[package]] -name = "tomli" -version = "2.2.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/18/87/302344fed471e44a87289cf4967697d07e532f2421fdaf868a303cbae4ff/tomli-2.2.1.tar.gz", hash = "sha256:cd45e1dc79c835ce60f7404ec8119f2eb06d38b1deba146f07ced3bbc44505ff", size = 17175 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/6e/c2/61d3e0f47e2b74ef40a68b9e6ad5984f6241a942f7cd3bbfbdbd03861ea9/tomli-2.2.1-py3-none-any.whl", hash = "sha256:cb55c73c5f4408779d0cf3eef9f762b9c9f147a77de7b258bef0a5628adc85cc", size = 14257 }, -] - -[[package]] -name = "toolz" -version = "1.0.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/8a/0b/d80dfa675bf592f636d1ea0b835eab4ec8df6e9415d8cfd766df54456123/toolz-1.0.0.tar.gz", hash = "sha256:2c86e3d9a04798ac556793bced838816296a2f085017664e4995cb40a1047a02", size = 66790 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/03/98/eb27cc78ad3af8e302c9d8ff4977f5026676e130d28dd7578132a457170c/toolz-1.0.0-py3-none-any.whl", hash = "sha256:292c8f1c4e7516bf9086f8850935c799a874039c8bcf959d47b600e4c44a6236", size = 56383 }, -] - -[[package]] -name = "torch" -version = "2.7.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "filelock" }, - { name = "fsspec" }, - { name = "jinja2" }, - { name = "networkx" }, - { name = "nvidia-cublas-cu12", marker = "platform_machine == 'x86_64' and platform_system == 'Linux'" }, - { name = "nvidia-cuda-cupti-cu12", marker = "platform_machine == 'x86_64' and platform_system == 'Linux'" }, - { name = "nvidia-cuda-nvrtc-cu12", marker = "platform_machine == 'x86_64' and platform_system == 'Linux'" }, - { name = "nvidia-cuda-runtime-cu12", marker = "platform_machine == 'x86_64' and platform_system == 'Linux'" }, - { name = "nvidia-cudnn-cu12", marker = "platform_machine == 'x86_64' and platform_system == 'Linux'" }, - { name = "nvidia-cufft-cu12", marker = "platform_machine == 'x86_64' and platform_system == 'Linux'" }, - { name = "nvidia-cufile-cu12", marker = "platform_machine == 'x86_64' and platform_system == 'Linux'" }, - { name = "nvidia-curand-cu12", marker = "platform_machine == 'x86_64' and platform_system == 'Linux'" }, - { name = "nvidia-cusolver-cu12", marker = "platform_machine == 'x86_64' and platform_system == 'Linux'" }, - { name = "nvidia-cusparse-cu12", marker = "platform_machine == 'x86_64' and platform_system == 'Linux'" }, - { name = "nvidia-cusparselt-cu12", marker = "platform_machine == 'x86_64' and platform_system == 'Linux'" }, - { name = "nvidia-nccl-cu12", marker = "platform_machine == 'x86_64' and platform_system == 'Linux'" }, - { name = "nvidia-nvjitlink-cu12", marker = "platform_machine == 'x86_64' and platform_system == 'Linux'" }, - { name = "nvidia-nvtx-cu12", marker = "platform_machine == 'x86_64' and platform_system == 'Linux'" }, - { name = "sympy" }, - { name = "triton", marker = "platform_machine == 'x86_64' and platform_system == 'Linux'" }, - { name = "typing-extensions" }, -] -wheels = [ - { url = "https://files.pythonhosted.org/packages/46/c2/3fb87940fa160d956ee94d644d37b99a24b9c05a4222bf34f94c71880e28/torch-2.7.0-cp310-cp310-manylinux_2_28_aarch64.whl", hash = "sha256:c9afea41b11e1a1ab1b258a5c31afbd646d6319042bfe4f231b408034b51128b", size = 99158447 }, - { url = "https://files.pythonhosted.org/packages/cc/2c/91d1de65573fce563f5284e69d9c56b57289625cffbbb6d533d5d56c36a5/torch-2.7.0-cp310-cp310-manylinux_2_28_x86_64.whl", hash = "sha256:0b9960183b6e5b71239a3e6c883d8852c304e691c0b2955f7045e8a6d05b9183", size = 865164221 }, - { url = "https://files.pythonhosted.org/packages/7f/7e/1b1cc4e0e7cc2666cceb3d250eef47a205f0821c330392cf45eb08156ce5/torch-2.7.0-cp310-cp310-win_amd64.whl", hash = "sha256:2ad79d0d8c2a20a37c5df6052ec67c2078a2c4e9a96dd3a8b55daaff6d28ea29", size = 212521189 }, - { url = "https://files.pythonhosted.org/packages/dc/0b/b2b83f30b8e84a51bf4f96aa3f5f65fdf7c31c591cc519310942339977e2/torch-2.7.0-cp310-none-macosx_11_0_arm64.whl", hash = "sha256:34e0168ed6de99121612d72224e59b2a58a83dae64999990eada7260c5dd582d", size = 68559462 }, -] - -[[package]] -name = "torch-audiomentations" -version = "0.12.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "julius" }, - { name = "torch" }, - { name = "torch-pitch-shift" }, - { name = "torchaudio" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/31/8d/2f8fd7e34c75f5ee8de4310c3bd3f22270acd44d1f809e2fe7c12fbf35f8/torch_audiomentations-0.12.0.tar.gz", hash = "sha256:b02d4c5eb86376986a53eb405cca5e34f370ea9284411237508e720c529f7888", size = 52094 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/21/9d/1ee04f49c15d2d632f6f7102061d7c07652858e6d91b58a091531034e84f/torch_audiomentations-0.12.0-py3-none-any.whl", hash = "sha256:1b80b91d2016ccf83979622cac8f702072a79b7dcc4c2bee40f00b26433a786b", size = 48506 }, -] - -[[package]] -name = "torch-complex" -version = "0.4.4" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "numpy" }, - { name = "packaging" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/bf/2b/17cb15a383cf2135330371e034d13b9043dc6d8bd07c871b5aa3064fbed1/torch_complex-0.4.4.tar.gz", hash = "sha256:4153fd6b24a0bad689e6f193bfbd00f38283b1890d808bef684ddc6d1f63fd3f", size = 10025 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/f4/c5/9b4d756a7ada951e9b17dcc636f98ed1073c737ae809b150ef408afb6298/torch_complex-0.4.4-py3-none-any.whl", hash = "sha256:6ab4ecd4f3a16e3adb70a7f7cd2e769a9dfd07d7a8e27d04ff9c621ebbe34b13", size = 9125 }, -] - -[[package]] -name = "torch-pitch-shift" -version = "1.2.5" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "packaging" }, - { name = "primepy" }, - { name = "torch" }, - { name = "torchaudio" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/79/a6/722a832bca75d5079f6731e005b3d0c2eec7c6c6863d030620952d143d57/torch_pitch_shift-1.2.5.tar.gz", hash = "sha256:6e1c7531f08d0f407a4c55e5ff8385a41355c5c5d27ab7fa08632e51defbd0ed", size = 4725 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/27/4c/96ac2a09efb56cc3c41fb3ce9b6f4d8c0604499f7481d4a13a7b03e21382/torch_pitch_shift-1.2.5-py3-none-any.whl", hash = "sha256:6f8500cbc13f1c98b11cde1805ce5084f82cdd195c285f34287541f168a7c6a7", size = 5005 }, -] - -[[package]] -name = "torchaudio" -version = "2.7.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "torch" }, -] -wheels = [ - { url = "https://files.pythonhosted.org/packages/34/26/abc66c79092ad2eaaade546dc93e23d99ddf2513988261b943d274f5c01a/torchaudio-2.7.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:1c4a646c9e9347836c09e965eebc58dd028ec6ef34c46d3e7891bffd8dc645ea", size = 1842304 }, - { url = "https://files.pythonhosted.org/packages/ee/f7/17b8fbce19280424e612f254e1b89faf3c7640c022667a480307f2f3ca76/torchaudio-2.7.0-cp310-cp310-manylinux_2_28_aarch64.whl", hash = "sha256:9e4073992f4f8e7113e4b505d95095361ceb2f21dd7b9310776160a24266f8f6", size = 1680682 }, - { url = "https://files.pythonhosted.org/packages/f2/df/ee0097fc41f718152026541c4c6cdeea830bc09903cc36a53037942a6d3d/torchaudio-2.7.0-cp310-cp310-manylinux_2_28_x86_64.whl", hash = "sha256:f7c99f7c062d6a56a3e281e3c2b779099e64cad1ce78891df61c4d19ce40742e", size = 3444849 }, - { url = "https://files.pythonhosted.org/packages/65/a6/e1903c1b3787f0408d30624536d2ae30da9f749720f3cf272a4fb7abc490/torchaudio-2.7.0-cp310-cp310-win_amd64.whl", hash = "sha256:a5443422640cbe532aaacd83ad2ee6911b0451f7f50e6b3755015e92df579d37", size = 2492239 }, -] - -[[package]] -name = "torchmetrics" -version = "1.7.3" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "lightning-utilities" }, - { name = "numpy" }, - { name = "packaging" }, - { name = "torch" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/48/22/8b16c4ec34d93ee15024924cbbe84fbd235bb3e1df2cc8f48c865c1528e7/torchmetrics-1.7.3.tar.gz", hash = "sha256:08450a19cdb67ba1608aac0b213e5dc73033e11b60ad4719696ebcede591621e", size = 566545 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/6f/f2/bed7da46003c26ed44fc7fa3ecc98a84216f0d4758e5e6a3693754d490d9/torchmetrics-1.7.3-py3-none-any.whl", hash = "sha256:7b6fd43e92f0a1071c8bcb029637f252b0630699140a93ed8817ce7afe9db34e", size = 962639 }, -] - -[[package]] -name = "tornado" -version = "6.5.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/51/89/c72771c81d25d53fe33e3dca61c233b665b2780f21820ba6fd2c6793c12b/tornado-6.5.1.tar.gz", hash = "sha256:84ceece391e8eb9b2b95578db65e920d2a61070260594819589609ba9bc6308c", size = 509934 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/77/89/f4532dee6843c9e0ebc4e28d4be04c67f54f60813e4bf73d595fe7567452/tornado-6.5.1-cp39-abi3-macosx_10_9_universal2.whl", hash = "sha256:d50065ba7fd11d3bd41bcad0825227cc9a95154bad83239357094c36708001f7", size = 441948 }, - { url = "https://files.pythonhosted.org/packages/15/9a/557406b62cffa395d18772e0cdcf03bed2fff03b374677348eef9f6a3792/tornado-6.5.1-cp39-abi3-macosx_10_9_x86_64.whl", hash = "sha256:9e9ca370f717997cb85606d074b0e5b247282cf5e2e1611568b8821afe0342d6", size = 440112 }, - { url = "https://files.pythonhosted.org/packages/55/82/7721b7319013a3cf881f4dffa4f60ceff07b31b394e459984e7a36dc99ec/tornado-6.5.1-cp39-abi3-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:b77e9dfa7ed69754a54c89d82ef746398be82f749df69c4d3abe75c4d1ff4888", size = 443672 }, - { url = "https://files.pythonhosted.org/packages/7d/42/d11c4376e7d101171b94e03cef0cbce43e823ed6567ceda571f54cf6e3ce/tornado-6.5.1-cp39-abi3-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:253b76040ee3bab8bcf7ba9feb136436a3787208717a1fb9f2c16b744fba7331", size = 443019 }, - { url = "https://files.pythonhosted.org/packages/7d/f7/0c48ba992d875521ac761e6e04b0a1750f8150ae42ea26df1852d6a98942/tornado-6.5.1-cp39-abi3-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:308473f4cc5a76227157cdf904de33ac268af770b2c5f05ca6c1161d82fdd95e", size = 443252 }, - { url = "https://files.pythonhosted.org/packages/89/46/d8d7413d11987e316df4ad42e16023cd62666a3c0dfa1518ffa30b8df06c/tornado-6.5.1-cp39-abi3-musllinux_1_2_aarch64.whl", hash = "sha256:caec6314ce8a81cf69bd89909f4b633b9f523834dc1a352021775d45e51d9401", size = 443930 }, - { url = "https://files.pythonhosted.org/packages/78/b2/f8049221c96a06df89bed68260e8ca94beca5ea532ffc63b1175ad31f9cc/tornado-6.5.1-cp39-abi3-musllinux_1_2_i686.whl", hash = "sha256:13ce6e3396c24e2808774741331638ee6c2f50b114b97a55c5b442df65fd9692", size = 443351 }, - { url = "https://files.pythonhosted.org/packages/76/ff/6a0079e65b326cc222a54720a748e04a4db246870c4da54ece4577bfa702/tornado-6.5.1-cp39-abi3-musllinux_1_2_x86_64.whl", hash = "sha256:5cae6145f4cdf5ab24744526cc0f55a17d76f02c98f4cff9daa08ae9a217448a", size = 443328 }, - { url = "https://files.pythonhosted.org/packages/49/18/e3f902a1d21f14035b5bc6246a8c0f51e0eef562ace3a2cea403c1fb7021/tornado-6.5.1-cp39-abi3-win32.whl", hash = "sha256:e0a36e1bc684dca10b1aa75a31df8bdfed656831489bc1e6a6ebed05dc1ec365", size = 444396 }, - { url = "https://files.pythonhosted.org/packages/7b/09/6526e32bf1049ee7de3bebba81572673b19a2a8541f795d887e92af1a8bc/tornado-6.5.1-cp39-abi3-win_amd64.whl", hash = "sha256:908e7d64567cecd4c2b458075589a775063453aeb1d2a1853eedb806922f568b", size = 444840 }, - { url = "https://files.pythonhosted.org/packages/55/a7/535c44c7bea4578e48281d83c615219f3ab19e6abc67625ef637c73987be/tornado-6.5.1-cp39-abi3-win_arm64.whl", hash = "sha256:02420a0eb7bf617257b9935e2b754d1b63897525d8a289c9d65690d580b4dcf7", size = 443596 }, -] - -[[package]] -name = "tqdm" -version = "4.67.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "colorama", marker = "platform_system == 'Windows'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/a8/4b/29b4ef32e036bb34e4ab51796dd745cdba7ed47ad142a9f4a1eb8e0c744d/tqdm-4.67.1.tar.gz", hash = "sha256:f8aef9c52c08c13a65f30ea34f4e5aac3fd1a34959879d7e59e63027286627f2", size = 169737 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/d0/30/dc54f88dd4a2b5dc8a0279bdd7270e735851848b762aeb1c1184ed1f6b14/tqdm-4.67.1-py3-none-any.whl", hash = "sha256:26445eca388f82e72884e0d580d5464cd801a3ea01e63e5601bdff9ba6a48de2", size = 78540 }, -] - -[[package]] -name = "traitlets" -version = "5.14.3" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/eb/79/72064e6a701c2183016abbbfedaba506d81e30e232a68c9f0d6f6fcd1574/traitlets-5.14.3.tar.gz", hash = "sha256:9ed0579d3502c94b4b3732ac120375cda96f923114522847de4b3bb98b96b6b7", size = 161621 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/00/c0/8f5d070730d7836adc9c9b6408dec68c6ced86b304a9b26a14df072a6e8c/traitlets-5.14.3-py3-none-any.whl", hash = "sha256:b74e89e397b1ed28cc831db7aea759ba6640cb3de13090ca145426688ff1ac4f", size = 85359 }, -] - -[[package]] -name = "transformers" -version = "4.51.3" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "filelock" }, - { name = "huggingface-hub" }, - { name = "numpy" }, - { name = "packaging" }, - { name = "pyyaml" }, - { name = "regex" }, - { name = "requests" }, - { name = "safetensors" }, - { name = "tokenizers" }, - { name = "tqdm" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/f1/11/7414d5bc07690002ce4d7553602107bf969af85144bbd02830f9fb471236/transformers-4.51.3.tar.gz", hash = "sha256:e292fcab3990c6defe6328f0f7d2004283ca81a7a07b2de9a46d67fd81ea1409", size = 8941266 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/a9/b6/5257d04ae327b44db31f15cce39e6020cc986333c715660b1315a9724d82/transformers-4.51.3-py3-none-any.whl", hash = "sha256:fd3279633ceb2b777013234bbf0b4f5c2d23c4626b05497691f00cfda55e8a83", size = 10383940 }, -] - -[[package]] -name = "triton" -version = "3.3.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "setuptools" }, -] -wheels = [ - { url = "https://files.pythonhosted.org/packages/76/04/d54d3a6d077c646624dc9461b0059e23fd5d30e0dbe67471e3654aec81f9/triton-3.3.0-cp310-cp310-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:fad99beafc860501d7fcc1fb7045d9496cbe2c882b1674640304949165a916e7", size = 156441993 }, -] - -[[package]] -name = "typeguard" -version = "4.4.4" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/c7/68/71c1a15b5f65f40e91b65da23b8224dad41349894535a97f63a52e462196/typeguard-4.4.4.tar.gz", hash = "sha256:3a7fd2dffb705d4d0efaed4306a704c89b9dee850b688f060a8b1615a79e5f74", size = 75203 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/1b/a9/e3aee762739c1d7528da1c3e06d518503f8b6c439c35549b53735ba52ead/typeguard-4.4.4-py3-none-any.whl", hash = "sha256:b5f562281b6bfa1f5492470464730ef001646128b180769880468bd84b68b09e", size = 34874 }, -] - -[[package]] -name = "typer" -version = "0.16.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "click" }, - { name = "rich" }, - { name = "shellingham" }, - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/c5/8c/7d682431efca5fd290017663ea4588bf6f2c6aad085c7f108c5dbc316e70/typer-0.16.0.tar.gz", hash = "sha256:af377ffaee1dbe37ae9440cb4e8f11686ea5ce4e9bae01b84ae7c63b87f1dd3b", size = 102625 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/76/42/3efaf858001d2c2913de7f354563e3a3a2f0decae3efe98427125a8f441e/typer-0.16.0-py3-none-any.whl", hash = "sha256:1f79bed11d4d02d4310e3c1b7ba594183bcedb0ac73b27a9e5f28f6fb5b98855", size = 46317 }, -] - -[[package]] -name = "types-python-dateutil" -version = "2.9.0.20250516" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/ef/88/d65ed807393285204ab6e2801e5d11fbbea811adcaa979a2ed3b67a5ef41/types_python_dateutil-2.9.0.20250516.tar.gz", hash = "sha256:13e80d6c9c47df23ad773d54b2826bd52dbbb41be87c3f339381c1700ad21ee5", size = 13943 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/c5/3f/b0e8db149896005adc938a1e7f371d6d7e9eca4053a29b108978ed15e0c2/types_python_dateutil-2.9.0.20250516-py3-none-any.whl", hash = "sha256:2b2b3f57f9c6a61fba26a9c0ffb9ea5681c9b83e69cd897c6b5f668d9c0cab93", size = 14356 }, -] - -[[package]] -name = "typing-extensions" -version = "4.14.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/d1/bc/51647cd02527e87d05cb083ccc402f93e441606ff1f01739a62c8ad09ba5/typing_extensions-4.14.0.tar.gz", hash = "sha256:8676b788e32f02ab42d9e7c61324048ae4c6d844a399eebace3d4979d75ceef4", size = 107423 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/69/e0/552843e0d356fbb5256d21449fa957fa4eff3bbc135a74a691ee70c7c5da/typing_extensions-4.14.0-py3-none-any.whl", hash = "sha256:a1514509136dd0b477638fc68d6a91497af5076466ad0fa6c338e44e359944af", size = 43839 }, -] - -[[package]] -name = "typing-inspection" -version = "0.4.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/f8/b1/0c11f5058406b3af7609f121aaa6b609744687f1d158b3c3a5bf4cc94238/typing_inspection-0.4.1.tar.gz", hash = "sha256:6ae134cc0203c33377d43188d4064e9b357dba58cff3185f22924610e70a9d28", size = 75726 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/17/69/cd203477f944c353c31bade965f880aa1061fd6bf05ded0726ca845b6ff7/typing_inspection-0.4.1-py3-none-any.whl", hash = "sha256:389055682238f53b04f7badcb49b989835495a96700ced5dab2d8feae4b26f51", size = 14552 }, -] - -[[package]] -name = "tzdata" -version = "2025.2" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/95/32/1a225d6164441be760d75c2c42e2780dc0873fe382da3e98a2e1e48361e5/tzdata-2025.2.tar.gz", hash = "sha256:b60a638fcc0daffadf82fe0f57e53d06bdec2f36c4df66280ae79bce6bd6f2b9", size = 196380 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/5c/23/c7abc0ca0a1526a0774eca151daeb8de62ec457e77262b66b359c3c7679e/tzdata-2025.2-py2.py3-none-any.whl", hash = "sha256:1a403fada01ff9221ca8044d701868fa132215d84beb92242d9acd2147f667a8", size = 347839 }, -] - -[[package]] -name = "umap-learn" -version = "0.5.7" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "numba" }, - { name = "numpy" }, - { name = "pynndescent" }, - { name = "scikit-learn" }, - { name = "scipy" }, - { name = "tqdm" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/6f/d4/9ed627905f7993349671283b3c5bf2d9f543ef79229fa1c7e01324eb900c/umap-learn-0.5.7.tar.gz", hash = "sha256:b2a97973e4c6ffcebf241100a8de589a4c84126a832ab40f296c6d9fcc5eb19e", size = 92680 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/3c/8f/671c0e1f2572ba625cbcc1faeba9435e00330c3d6962858711445cf1e817/umap_learn-0.5.7-py3-none-any.whl", hash = "sha256:6a7e0be2facfa365a5ed6588447102bdbef32a0ef449535c25c97ea7e680073c", size = 88815 }, -] - -[[package]] -name = "uri-template" -version = "1.3.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/31/c7/0336f2bd0bcbada6ccef7aaa25e443c118a704f828a0620c6fa0207c1b64/uri-template-1.3.0.tar.gz", hash = "sha256:0e00f8eb65e18c7de20d595a14336e9f337ead580c70934141624b6d1ffdacc7", size = 21678 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/e7/00/3fca040d7cf8a32776d3d81a00c8ee7457e00f80c649f1e4a863c8321ae9/uri_template-1.3.0-py3-none-any.whl", hash = "sha256:a44a133ea12d44a0c0f06d7d42a52d71282e77e2f937d8abd5655b8d56fc1363", size = 11140 }, -] - -[[package]] -name = "urllib3" -version = "2.5.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/15/22/9ee70a2574a4f4599c47dd506532914ce044817c7752a79b6a51286319bc/urllib3-2.5.0.tar.gz", hash = "sha256:3fc47733c7e419d4bc3f6b3dc2b4f890bb743906a30d56ba4a5bfa4bbff92760", size = 393185 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/a7/c2/fe1e52489ae3122415c51f387e221dd0773709bad6c6cdaa599e8a2c5185/urllib3-2.5.0-py3-none-any.whl", hash = "sha256:e6b01673c0fa6a13e374b50871808eb3bf7046c4b125b216f6bf1cc604cff0dc", size = 129795 }, -] - -[[package]] -name = "wandb" -version = "0.20.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "click" }, - { name = "gitpython" }, - { name = "packaging" }, - { name = "platformdirs" }, - { name = "protobuf", marker = "python_full_version == '3.10.12'" }, - { name = "psutil" }, - { name = "pydantic" }, - { name = "pyyaml" }, - { name = "requests" }, - { name = "sentry-sdk" }, - { name = "setproctitle" }, - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/62/1f/92be0ca87fb49eb48c16dcf0845a3579a57c4734fec2b95862cf5a0494a0/wandb-0.20.1.tar.gz", hash = "sha256:dbd3fc60dfe7bf83c4de24b206b99b44949fef323f817a783883db72fc5f3bfe", size = 40320062 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/c9/18/afcc37d0b93dd6f6d0f0c5683b9cfff9416ae1539931f58932a2938c0070/wandb-0.20.1-py3-none-any.whl", hash = "sha256:e6395cabf074247042be1cf0dc6ab0b06aa4c9538c2e1fdc5b507a690ce0cf17", size = 6458872 }, - { url = "https://files.pythonhosted.org/packages/e6/b5/70f9e2a3d1380b729ae5853763d938edc50072df357f79bbd19b9aae8e3f/wandb-0.20.1-py3-none-macosx_10_14_x86_64.whl", hash = "sha256:2475a48c693adf677d40da9e1c8ceeaf86d745ffc3b7e3535731279d02f9e845", size = 22517483 }, - { url = "https://files.pythonhosted.org/packages/cc/7e/4eb9aeb2fd974d410a8f2eb11b0219536503913a050d46a03206151705c8/wandb-0.20.1-py3-none-macosx_11_0_arm64.whl", hash = "sha256:99cce804c31ec1e0d1e691650a7d51773ed7329c41745d56384fa3655a0e9b2c", size = 22034511 }, - { url = "https://files.pythonhosted.org/packages/34/38/1df22c2273e6f7ab0aae4fd032085d6d92ab112f5b261646e7dc5e675cfe/wandb-0.20.1-py3-none-macosx_11_0_x86_64.whl", hash = "sha256:ce3ee412677a1679e04b21e03a91e1e02eb90faf658d682bee86c33cf5f32e09", size = 22720771 }, - { url = "https://files.pythonhosted.org/packages/38/96/78fc7a7ea7158d136c84f481423f8736c9346a2387287ec8a6d92019975c/wandb-0.20.1-py3-none-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5e58ca32c7147161158f09b0fb5f5896876f8569d0d10ae7b64d0510c868ce33", size = 21537453 }, - { url = "https://files.pythonhosted.org/packages/88/c9/41b8bdb493e5eda32b502bc1cc49d539335a92cacaf0ef304d7dae0240aa/wandb-0.20.1-py3-none-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:591506ecbdd396648cc323ba270f3ab4aed3158e1dbfa7636c09f9f7f0253e1c", size = 23161349 }, - { url = "https://files.pythonhosted.org/packages/7d/f2/79e783cc50a47d373dfbda862eb5396de8139167e8c6443a16ef0166106f/wandb-0.20.1-py3-none-musllinux_1_2_aarch64.whl", hash = "sha256:382508532db09893f81cc926b1d333caa4c8a7db057878899fadf929bbdb3b56", size = 21550624 }, - { url = "https://files.pythonhosted.org/packages/26/32/23890a726302e7be28bda9fff47ce9b491af64e339aba4d32b3b8d1a7aaf/wandb-0.20.1-py3-none-musllinux_1_2_x86_64.whl", hash = "sha256:29ea495e49393db860f17437fe37e48018da90436ce10949b471780f09293bd7", size = 23237996 }, - { url = "https://files.pythonhosted.org/packages/af/94/296e520b086b2a4f10e99bcea3cd5856421b9c004824663501e3789a713b/wandb-0.20.1-py3-none-win32.whl", hash = "sha256:455ee0a652e59ab1e4b546fa1dc833dd3063aa7e64eb8abf95d22f0e9f08c574", size = 22518456 }, - { url = "https://files.pythonhosted.org/packages/52/5f/c44ad7b2a062ca5f4da99ae475cea274c38f6ec37bdaca1b1c653ee87274/wandb-0.20.1-py3-none-win_amd64.whl", hash = "sha256:6d2431652f096b7e394c29a99135a6441c02ed3198b963f0b351a5b5e56aeca0", size = 22518459 }, -] - -[[package]] -name = "wcwidth" -version = "0.2.13" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/6c/63/53559446a878410fc5a5974feb13d31d78d752eb18aeba59c7fef1af7598/wcwidth-0.2.13.tar.gz", hash = "sha256:72ea0c06399eb286d978fdedb6923a9eb47e1c486ce63e9b4e64fc18303972b5", size = 101301 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/fd/84/fd2ba7aafacbad3c4201d395674fc6348826569da3c0937e75505ead3528/wcwidth-0.2.13-py2.py3-none-any.whl", hash = "sha256:3da69048e4540d84af32131829ff948f1e022c1c6bdb8d6102117aac784f6859", size = 34166 }, -] - -[[package]] -name = "webcolors" -version = "24.11.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/7b/29/061ec845fb58521848f3739e466efd8250b4b7b98c1b6c5bf4d40b419b7e/webcolors-24.11.1.tar.gz", hash = "sha256:ecb3d768f32202af770477b8b65f318fa4f566c22948673a977b00d589dd80f6", size = 45064 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/60/e8/c0e05e4684d13459f93d312077a9a2efbe04d59c393bc2b8802248c908d4/webcolors-24.11.1-py3-none-any.whl", hash = "sha256:515291393b4cdf0eb19c155749a096f779f7d909f7cceea072791cb9095b92e9", size = 14934 }, -] - -[[package]] -name = "webdataset" -version = "1.0.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "braceexpand" }, - { name = "numpy" }, - { name = "pyyaml" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/5a/3a/68800d92e065cf4750ebecf973b13979c0c929b439e1293012938862038d/webdataset-1.0.2.tar.gz", hash = "sha256:7f0498be827cfa46cc5430a58768a24e2c6a410676a61be1838f53d61afdaab4", size = 80090 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/d9/00/aca6beb3658dab4ed3dbb41a78e6e7f31342e0b41d28088f205525751601/webdataset-1.0.2-py3-none-any.whl", hash = "sha256:3dbfced32b25c0d199c6b9787937b6f85742bc3c84f652c846893075c1c082d9", size = 74956 }, -] - -[[package]] -name = "webencodings" -version = "0.5.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/0b/02/ae6ceac1baeda530866a85075641cec12989bd8d31af6d5ab4a3e8c92f47/webencodings-0.5.1.tar.gz", hash = "sha256:b36a1c245f2d304965eb4e0a82848379241dc04b865afcc4aab16748587e1923", size = 9721 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/f4/24/2a3e3df732393fed8b3ebf2ec078f05546de641fe1b667ee316ec1dcf3b7/webencodings-0.5.1-py2.py3-none-any.whl", hash = "sha256:a0af1213f3c2226497a97e2b3aa01a7e4bee4f403f95be16fc9acd2947514a78", size = 11774 }, -] - -[[package]] -name = "websocket-client" -version = "1.8.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/e6/30/fba0d96b4b5fbf5948ed3f4681f7da2f9f64512e1d303f94b4cc174c24a5/websocket_client-1.8.0.tar.gz", hash = "sha256:3239df9f44da632f96012472805d40a23281a991027ce11d2f45a6f24ac4c3da", size = 54648 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/5a/84/44687a29792a70e111c5c477230a72c4b957d88d16141199bf9acb7537a3/websocket_client-1.8.0-py3-none-any.whl", hash = "sha256:17b44cc997f5c498e809b22cdf2d9c7a9e71c02c8cc2b6c56e7c2d1239bfa526", size = 58826 }, -] - -[[package]] -name = "werkzeug" -version = "3.1.3" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "markupsafe" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/9f/69/83029f1f6300c5fb2471d621ab06f6ec6b3324685a2ce0f9777fd4a8b71e/werkzeug-3.1.3.tar.gz", hash = "sha256:60723ce945c19328679790e3282cc758aa4a6040e4bb330f53d30fa546d44746", size = 806925 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/52/24/ab44c871b0f07f491e5d2ad12c9bd7358e527510618cb1b803a88e986db1/werkzeug-3.1.3-py3-none-any.whl", hash = "sha256:54b78bf3716d19a65be4fceccc0d1d7b89e608834989dfae50ea87564639213e", size = 224498 }, -] - -[[package]] -name = "wget" -version = "3.2" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/47/6a/62e288da7bcda82b935ff0c6cfe542970f04e29c756b0e147251b2fb251f/wget-3.2.zip", hash = "sha256:35e630eca2aa50ce998b9b1a127bb26b30dfee573702782aa982f875e3f16061", size = 10857 } - -[[package]] -name = "widgetsnbextension" -version = "4.0.14" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/41/53/2e0253c5efd69c9656b1843892052a31c36d37ad42812b5da45c62191f7e/widgetsnbextension-4.0.14.tar.gz", hash = "sha256:a3629b04e3edb893212df862038c7232f62973373869db5084aed739b437b5af", size = 1097428 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/ca/51/5447876806d1088a0f8f71e16542bf350918128d0a69437df26047c8e46f/widgetsnbextension-4.0.14-py3-none-any.whl", hash = "sha256:4875a9eaf72fbf5079dc372a51a9f268fc38d46f767cbf85c43a36da5cb9b575", size = 2196503 }, -] - -[[package]] -name = "win32-setctime" -version = "1.2.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/b3/8f/705086c9d734d3b663af0e9bb3d4de6578d08f46b1b101c2442fd9aecaa2/win32_setctime-1.2.0.tar.gz", hash = "sha256:ae1fdf948f5640aae05c511ade119313fb6a30d7eabe25fef9764dca5873c4c0", size = 4867 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/e1/07/c6fe3ad3e685340704d314d765b7912993bcb8dc198f0e7a89382d37974b/win32_setctime-1.2.0-py3-none-any.whl", hash = "sha256:95d644c4e708aba81dc3704a116d8cbc974d70b3bdb8be1d150e36be6e9d1390", size = 4083 }, -] - -[[package]] -name = "wrapt" -version = "1.17.2" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/c3/fc/e91cc220803d7bc4db93fb02facd8461c37364151b8494762cc88b0fbcef/wrapt-1.17.2.tar.gz", hash = "sha256:41388e9d4d1522446fe79d3213196bd9e3b301a336965b9e27ca2788ebd122f3", size = 55531 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/5a/d1/1daec934997e8b160040c78d7b31789f19b122110a75eca3d4e8da0049e1/wrapt-1.17.2-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:3d57c572081fed831ad2d26fd430d565b76aa277ed1d30ff4d40670b1c0dd984", size = 53307 }, - { url = "https://files.pythonhosted.org/packages/1b/7b/13369d42651b809389c1a7153baa01d9700430576c81a2f5c5e460df0ed9/wrapt-1.17.2-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:b5e251054542ae57ac7f3fba5d10bfff615b6c2fb09abeb37d2f1463f841ae22", size = 38486 }, - { url = "https://files.pythonhosted.org/packages/62/bf/e0105016f907c30b4bd9e377867c48c34dc9c6c0c104556c9c9126bd89ed/wrapt-1.17.2-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:80dd7db6a7cb57ffbc279c4394246414ec99537ae81ffd702443335a61dbf3a7", size = 38777 }, - { url = "https://files.pythonhosted.org/packages/27/70/0f6e0679845cbf8b165e027d43402a55494779295c4b08414097b258ac87/wrapt-1.17.2-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0a6e821770cf99cc586d33833b2ff32faebdbe886bd6322395606cf55153246c", size = 83314 }, - { url = "https://files.pythonhosted.org/packages/0f/77/0576d841bf84af8579124a93d216f55d6f74374e4445264cb378a6ed33eb/wrapt-1.17.2-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:b60fb58b90c6d63779cb0c0c54eeb38941bae3ecf7a73c764c52c88c2dcb9d72", size = 74947 }, - { url = "https://files.pythonhosted.org/packages/90/ec/00759565518f268ed707dcc40f7eeec38637d46b098a1f5143bff488fe97/wrapt-1.17.2-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b870b5df5b71d8c3359d21be8f0d6c485fa0ebdb6477dda51a1ea54a9b558061", size = 82778 }, - { url = "https://files.pythonhosted.org/packages/f8/5a/7cffd26b1c607b0b0c8a9ca9d75757ad7620c9c0a9b4a25d3f8a1480fafc/wrapt-1.17.2-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:4011d137b9955791f9084749cba9a367c68d50ab8d11d64c50ba1688c9b457f2", size = 81716 }, - { url = "https://files.pythonhosted.org/packages/7e/09/dccf68fa98e862df7e6a60a61d43d644b7d095a5fc36dbb591bbd4a1c7b2/wrapt-1.17.2-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:1473400e5b2733e58b396a04eb7f35f541e1fb976d0c0724d0223dd607e0f74c", size = 74548 }, - { url = "https://files.pythonhosted.org/packages/b7/8e/067021fa3c8814952c5e228d916963c1115b983e21393289de15128e867e/wrapt-1.17.2-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:3cedbfa9c940fdad3e6e941db7138e26ce8aad38ab5fe9dcfadfed9db7a54e62", size = 81334 }, - { url = "https://files.pythonhosted.org/packages/4b/0d/9d4b5219ae4393f718699ca1c05f5ebc0c40d076f7e65fd48f5f693294fb/wrapt-1.17.2-cp310-cp310-win32.whl", hash = "sha256:582530701bff1dec6779efa00c516496968edd851fba224fbd86e46cc6b73563", size = 36427 }, - { url = "https://files.pythonhosted.org/packages/72/6a/c5a83e8f61aec1e1aeef939807602fb880e5872371e95df2137142f5c58e/wrapt-1.17.2-cp310-cp310-win_amd64.whl", hash = "sha256:58705da316756681ad3c9c73fd15499aa4d8c69f9fd38dc8a35e06c12468582f", size = 38774 }, - { url = "https://files.pythonhosted.org/packages/2d/82/f56956041adef78f849db6b289b282e72b55ab8045a75abad81898c28d19/wrapt-1.17.2-py3-none-any.whl", hash = "sha256:b18f2d1533a71f069c7f82d524a52599053d4c7166e9dd374ae2136b7f40f7c8", size = 23594 }, -] - -[[package]] -name = "xxhash" -version = "3.5.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/00/5e/d6e5258d69df8b4ed8c83b6664f2b47d30d2dec551a29ad72a6c69eafd31/xxhash-3.5.0.tar.gz", hash = "sha256:84f2caddf951c9cbf8dc2e22a89d4ccf5d86391ac6418fe81e3c67d0cf60b45f", size = 84241 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/bb/8a/0e9feca390d512d293afd844d31670e25608c4a901e10202aa98785eab09/xxhash-3.5.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:ece616532c499ee9afbb83078b1b952beffef121d989841f7f4b3dc5ac0fd212", size = 31970 }, - { url = "https://files.pythonhosted.org/packages/16/e6/be5aa49580cd064a18200ab78e29b88b1127e1a8c7955eb8ecf81f2626eb/xxhash-3.5.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:3171f693dbc2cef6477054a665dc255d996646b4023fe56cb4db80e26f4cc520", size = 30801 }, - { url = "https://files.pythonhosted.org/packages/20/ee/b8a99ebbc6d1113b3a3f09e747fa318c3cde5b04bd9c197688fadf0eeae8/xxhash-3.5.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7c5d3e570ef46adaf93fc81b44aca6002b5a4d8ca11bd0580c07eac537f36680", size = 220927 }, - { url = "https://files.pythonhosted.org/packages/58/62/15d10582ef159283a5c2b47f6d799fc3303fe3911d5bb0bcc820e1ef7ff4/xxhash-3.5.0-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:7cb29a034301e2982df8b1fe6328a84f4b676106a13e9135a0d7e0c3e9f806da", size = 200360 }, - { url = "https://files.pythonhosted.org/packages/23/41/61202663ea9b1bd8e53673b8ec9e2619989353dba8cfb68e59a9cbd9ffe3/xxhash-3.5.0-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:5d0d307d27099bb0cbeea7260eb39ed4fdb99c5542e21e94bb6fd29e49c57a23", size = 428528 }, - { url = "https://files.pythonhosted.org/packages/f2/07/d9a3059f702dec5b3b703737afb6dda32f304f6e9da181a229dafd052c29/xxhash-3.5.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c0342aafd421795d740e514bc9858ebddfc705a75a8c5046ac56d85fe97bf196", size = 194149 }, - { url = "https://files.pythonhosted.org/packages/eb/58/27caadf78226ecf1d62dbd0c01d152ed381c14c1ee4ad01f0d460fc40eac/xxhash-3.5.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3dbbd9892c5ebffeca1ed620cf0ade13eb55a0d8c84e0751a6653adc6ac40d0c", size = 207703 }, - { url = "https://files.pythonhosted.org/packages/b1/08/32d558ce23e1e068453c39aed7b3c1cdc690c177873ec0ca3a90d5808765/xxhash-3.5.0-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:4cc2d67fdb4d057730c75a64c5923abfa17775ae234a71b0200346bfb0a7f482", size = 216255 }, - { url = "https://files.pythonhosted.org/packages/3f/d4/2b971e2d2b0a61045f842b622ef11e94096cf1f12cd448b6fd426e80e0e2/xxhash-3.5.0-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:ec28adb204b759306a3d64358a5e5c07d7b1dd0ccbce04aa76cb9377b7b70296", size = 202744 }, - { url = "https://files.pythonhosted.org/packages/19/ae/6a6438864a8c4c39915d7b65effd85392ebe22710412902487e51769146d/xxhash-3.5.0-cp310-cp310-musllinux_1_2_ppc64le.whl", hash = "sha256:1328f6d8cca2b86acb14104e381225a3d7b42c92c4b86ceae814e5c400dbb415", size = 210115 }, - { url = "https://files.pythonhosted.org/packages/48/7d/b3c27c27d1fc868094d02fe4498ccce8cec9fcc591825c01d6bcb0b4fc49/xxhash-3.5.0-cp310-cp310-musllinux_1_2_s390x.whl", hash = "sha256:8d47ebd9f5d9607fd039c1fbf4994e3b071ea23eff42f4ecef246ab2b7334198", size = 414247 }, - { url = "https://files.pythonhosted.org/packages/a1/05/918f9e7d2fbbd334b829997045d341d6239b563c44e683b9a7ef8fe50f5d/xxhash-3.5.0-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:b96d559e0fcddd3343c510a0fe2b127fbff16bf346dd76280b82292567523442", size = 191419 }, - { url = "https://files.pythonhosted.org/packages/08/29/dfe393805b2f86bfc47c290b275f0b7c189dc2f4e136fd4754f32eb18a8d/xxhash-3.5.0-cp310-cp310-win32.whl", hash = "sha256:61c722ed8d49ac9bc26c7071eeaa1f6ff24053d553146d5df031802deffd03da", size = 30114 }, - { url = "https://files.pythonhosted.org/packages/7b/d7/aa0b22c4ebb7c3ccb993d4c565132abc641cd11164f8952d89eb6a501909/xxhash-3.5.0-cp310-cp310-win_amd64.whl", hash = "sha256:9bed5144c6923cc902cd14bb8963f2d5e034def4486ab0bbe1f58f03f042f9a9", size = 30003 }, - { url = "https://files.pythonhosted.org/packages/69/12/f969b81541ee91b55f1ce469d7ab55079593c80d04fd01691b550e535000/xxhash-3.5.0-cp310-cp310-win_arm64.whl", hash = "sha256:893074d651cf25c1cc14e3bea4fceefd67f2921b1bb8e40fcfeba56820de80c6", size = 26773 }, - { url = "https://files.pythonhosted.org/packages/ab/9a/233606bada5bd6f50b2b72c45de3d9868ad551e83893d2ac86dc7bb8553a/xxhash-3.5.0-pp310-pypy310_pp73-macosx_10_15_x86_64.whl", hash = "sha256:2014c5b3ff15e64feecb6b713af12093f75b7926049e26a580e94dcad3c73d8c", size = 29732 }, - { url = "https://files.pythonhosted.org/packages/0c/67/f75276ca39e2c6604e3bee6c84e9db8a56a4973fde9bf35989787cf6e8aa/xxhash-3.5.0-pp310-pypy310_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:fab81ef75003eda96239a23eda4e4543cedc22e34c373edcaf744e721a163986", size = 36214 }, - { url = "https://files.pythonhosted.org/packages/0f/f8/f6c61fd794229cc3848d144f73754a0c107854372d7261419dcbbd286299/xxhash-3.5.0-pp310-pypy310_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4e2febf914ace002132aa09169cc572e0d8959d0f305f93d5828c4836f9bc5a6", size = 32020 }, - { url = "https://files.pythonhosted.org/packages/79/d3/c029c99801526f859e6b38d34ab87c08993bf3dcea34b11275775001638a/xxhash-3.5.0-pp310-pypy310_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:5d3a10609c51da2a1c0ea0293fc3968ca0a18bd73838455b5bca3069d7f8e32b", size = 40515 }, - { url = "https://files.pythonhosted.org/packages/62/e3/bef7b82c1997579c94de9ac5ea7626d01ae5858aa22bf4fcb38bf220cb3e/xxhash-3.5.0-pp310-pypy310_pp73-win_amd64.whl", hash = "sha256:5a74f23335b9689b66eb6dbe2a931a88fcd7a4c2cc4b1cb0edba8ce381c7a1da", size = 30064 }, -] - -[[package]] -name = "yarl" -version = "1.20.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "idna" }, - { name = "multidict" }, - { name = "propcache" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/3c/fb/efaa23fa4e45537b827620f04cf8f3cd658b76642205162e072703a5b963/yarl-1.20.1.tar.gz", hash = "sha256:d017a4997ee50c91fd5466cef416231bb82177b93b029906cefc542ce14c35ac", size = 186428 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/cb/65/7fed0d774abf47487c64be14e9223749468922817b5e8792b8a64792a1bb/yarl-1.20.1-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:6032e6da6abd41e4acda34d75a816012717000fa6839f37124a47fcefc49bec4", size = 132910 }, - { url = "https://files.pythonhosted.org/packages/8a/7b/988f55a52da99df9e56dc733b8e4e5a6ae2090081dc2754fc8fd34e60aa0/yarl-1.20.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:2c7b34d804b8cf9b214f05015c4fee2ebe7ed05cf581e7192c06555c71f4446a", size = 90644 }, - { url = "https://files.pythonhosted.org/packages/f7/de/30d98f03e95d30c7e3cc093759982d038c8833ec2451001d45ef4854edc1/yarl-1.20.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:0c869f2651cc77465f6cd01d938d91a11d9ea5d798738c1dc077f3de0b5e5fed", size = 89322 }, - { url = "https://files.pythonhosted.org/packages/e0/7a/f2f314f5ebfe9200724b0b748de2186b927acb334cf964fd312eb86fc286/yarl-1.20.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:62915e6688eb4d180d93840cda4110995ad50c459bf931b8b3775b37c264af1e", size = 323786 }, - { url = "https://files.pythonhosted.org/packages/15/3f/718d26f189db96d993d14b984ce91de52e76309d0fd1d4296f34039856aa/yarl-1.20.1-cp310-cp310-manylinux_2_17_armv7l.manylinux2014_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:41ebd28167bc6af8abb97fec1a399f412eec5fd61a3ccbe2305a18b84fb4ca73", size = 319627 }, - { url = "https://files.pythonhosted.org/packages/a5/76/8fcfbf5fa2369157b9898962a4a7d96764b287b085b5b3d9ffae69cdefd1/yarl-1.20.1-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:21242b4288a6d56f04ea193adde174b7e347ac46ce6bc84989ff7c1b1ecea84e", size = 339149 }, - { url = "https://files.pythonhosted.org/packages/3c/95/d7fc301cc4661785967acc04f54a4a42d5124905e27db27bb578aac49b5c/yarl-1.20.1-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:bea21cdae6c7eb02ba02a475f37463abfe0a01f5d7200121b03e605d6a0439f8", size = 333327 }, - { url = "https://files.pythonhosted.org/packages/65/94/e21269718349582eee81efc5c1c08ee71c816bfc1585b77d0ec3f58089eb/yarl-1.20.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1f8a891e4a22a89f5dde7862994485e19db246b70bb288d3ce73a34422e55b23", size = 326054 }, - { url = "https://files.pythonhosted.org/packages/32/ae/8616d1f07853704523519f6131d21f092e567c5af93de7e3e94b38d7f065/yarl-1.20.1-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:dd803820d44c8853a109a34e3660e5a61beae12970da479cf44aa2954019bf70", size = 315035 }, - { url = "https://files.pythonhosted.org/packages/48/aa/0ace06280861ef055855333707db5e49c6e3a08840a7ce62682259d0a6c0/yarl-1.20.1-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:b982fa7f74c80d5c0c7b5b38f908971e513380a10fecea528091405f519b9ebb", size = 338962 }, - { url = "https://files.pythonhosted.org/packages/20/52/1e9d0e6916f45a8fb50e6844f01cb34692455f1acd548606cbda8134cd1e/yarl-1.20.1-cp310-cp310-musllinux_1_2_armv7l.whl", hash = "sha256:33f29ecfe0330c570d997bcf1afd304377f2e48f61447f37e846a6058a4d33b2", size = 335399 }, - { url = "https://files.pythonhosted.org/packages/f2/65/60452df742952c630e82f394cd409de10610481d9043aa14c61bf846b7b1/yarl-1.20.1-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:835ab2cfc74d5eb4a6a528c57f05688099da41cf4957cf08cad38647e4a83b30", size = 338649 }, - { url = "https://files.pythonhosted.org/packages/7b/f5/6cd4ff38dcde57a70f23719a838665ee17079640c77087404c3d34da6727/yarl-1.20.1-cp310-cp310-musllinux_1_2_ppc64le.whl", hash = "sha256:46b5e0ccf1943a9a6e766b2c2b8c732c55b34e28be57d8daa2b3c1d1d4009309", size = 358563 }, - { url = "https://files.pythonhosted.org/packages/d1/90/c42eefd79d0d8222cb3227bdd51b640c0c1d0aa33fe4cc86c36eccba77d3/yarl-1.20.1-cp310-cp310-musllinux_1_2_s390x.whl", hash = "sha256:df47c55f7d74127d1b11251fe6397d84afdde0d53b90bedb46a23c0e534f9d24", size = 357609 }, - { url = "https://files.pythonhosted.org/packages/03/c8/cea6b232cb4617514232e0f8a718153a95b5d82b5290711b201545825532/yarl-1.20.1-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:76d12524d05841276b0e22573f28d5fbcb67589836772ae9244d90dd7d66aa13", size = 350224 }, - { url = "https://files.pythonhosted.org/packages/ce/a3/eaa0ab9712f1f3d01faf43cf6f1f7210ce4ea4a7e9b28b489a2261ca8db9/yarl-1.20.1-cp310-cp310-win32.whl", hash = "sha256:6c4fbf6b02d70e512d7ade4b1f998f237137f1417ab07ec06358ea04f69134f8", size = 81753 }, - { url = "https://files.pythonhosted.org/packages/8f/34/e4abde70a9256465fe31c88ed02c3f8502b7b5dead693a4f350a06413f28/yarl-1.20.1-cp310-cp310-win_amd64.whl", hash = "sha256:aef6c4d69554d44b7f9d923245f8ad9a707d971e6209d51279196d8e8fe1ae16", size = 86817 }, - { url = "https://files.pythonhosted.org/packages/b4/2d/2345fce04cfd4bee161bf1e7d9cdc702e3e16109021035dbb24db654a622/yarl-1.20.1-py3-none-any.whl", hash = "sha256:83b8eb083fe4683c6115795d9fc1cfaf2cbbefb19b3a1cb68f6527460f483a77", size = 46542 }, -] diff --git a/models/.DS_Store b/models/.DS_Store deleted file mode 100644 index a6bf794785e6093a33b848512a9844f6e47e495d..0000000000000000000000000000000000000000 Binary files a/models/.DS_Store and /dev/null differ diff --git a/models/parakeet-ctc-110m-coreml/.DS_Store b/models/parakeet-ctc-110m-coreml/.DS_Store deleted file mode 100644 index 287c48c58388c597f097f3f208a17a86407f7b8c..0000000000000000000000000000000000000000 Binary files a/models/parakeet-ctc-110m-coreml/.DS_Store and /dev/null differ diff --git a/models/parakeet-ctc-110m-coreml/CTCHead.mlmodelc/analytics/coremldata.bin b/models/parakeet-ctc-110m-coreml/CTCHead.mlmodelc/analytics/coremldata.bin deleted file mode 100644 index 4093e2df453ec508a533d293a574198b2372845b..0000000000000000000000000000000000000000 --- a/models/parakeet-ctc-110m-coreml/CTCHead.mlmodelc/analytics/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:fc681823d92eca3dbece3a30c975afa7251eedae0e718b07ffbf1a8b4313b87e -size 243 diff --git a/models/parakeet-ctc-110m-coreml/CTCHead.mlmodelc/coremldata.bin b/models/parakeet-ctc-110m-coreml/CTCHead.mlmodelc/coremldata.bin deleted file mode 100644 index db4ae8de920867977480465fffece41bc606c0a2..0000000000000000000000000000000000000000 --- a/models/parakeet-ctc-110m-coreml/CTCHead.mlmodelc/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:2ebec8fc38c063de4b2159e21b1f981309fa5947c24d7e4883aca20f7c15fbb9 -size 377 diff --git a/models/parakeet-ctc-110m-coreml/CTCHead.mlmodelc/metadata.json b/models/parakeet-ctc-110m-coreml/CTCHead.mlmodelc/metadata.json deleted file mode 100644 index 4abe30bf0b35b72d2c9c013b61cc78b4f26a85e1..0000000000000000000000000000000000000000 --- a/models/parakeet-ctc-110m-coreml/CTCHead.mlmodelc/metadata.json +++ /dev/null @@ -1,66 +0,0 @@ -[ - { - "metadataOutputVersion" : "3.0", - "shortDescription" : "Parakeet 110M CTC decoder head", - "outputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 188 × 1025)", - "shortDescription" : "", - "shape" : "[1, 188, 1025]", - "name" : "ctc_logits", - "type" : "MultiArray" - } - ], - "storagePrecision" : "Float16", - "modelParameters" : [ - - ], - "author" : "Fluid Inference", - "specificationVersion" : 8, - "mlProgramOperationTypeHistogram" : { - "Ios17.cast" : 2, - "Ios17.conv" : 1, - "Ios17.transpose" : 1, - "Ios16.softmax" : 1, - "Ios17.log" : 1 - }, - "computePrecision" : "Mixed (Float16, Float32, Int32)", - "isUpdatable" : "0", - "stateSchema" : [ - - ], - "availability" : { - "macOS" : "14.0", - "tvOS" : "17.0", - "visionOS" : "1.0", - "watchOS" : "10.0", - "iOS" : "17.0", - "macCatalyst" : "17.0" - }, - "modelType" : { - "name" : "MLModelType_mlProgram" - }, - "inputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 512 × 188)", - "shortDescription" : "", - "shape" : "[1, 512, 188]", - "name" : "encoder_output", - "type" : "MultiArray" - } - ], - "userDefinedMetadata" : { - "com.github.apple.coremltools.source_dialect" : "TorchScript", - "com.github.apple.coremltools.source" : "torch==2.9.0", - "com.github.apple.coremltools.version" : "8.3.0" - }, - "generatedClassName" : "parakeet_ctc_head", - "method" : "predict" - } -] \ No newline at end of file diff --git a/models/parakeet-ctc-110m-coreml/CTCHead.mlmodelc/model.mil b/models/parakeet-ctc-110m-coreml/CTCHead.mlmodelc/model.mil deleted file mode 100644 index 67b3b5f87ce33caed71f9233828f90775bb8ac9d..0000000000000000000000000000000000000000 --- a/models/parakeet-ctc-110m-coreml/CTCHead.mlmodelc/model.mil +++ /dev/null @@ -1,24 +0,0 @@ -program(1.0) -[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3510.2.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.9.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "8.3.0"}})] -{ - func main(tensor encoder_output) { - tensor var_4 = const()[name = tensor("op_4"), val = tensor(-1)]; - tensor var_18_pad_type_0 = const()[name = tensor("op_18_pad_type_0"), val = tensor("valid")]; - tensor var_18_strides_0 = const()[name = tensor("op_18_strides_0"), val = tensor([1])]; - tensor var_18_pad_0 = const()[name = tensor("op_18_pad_0"), val = tensor([0, 0])]; - tensor var_18_dilations_0 = const()[name = tensor("op_18_dilations_0"), val = tensor([1])]; - tensor var_18_groups_0 = const()[name = tensor("op_18_groups_0"), val = tensor(1)]; - tensor encoder_output_to_fp16_dtype_0 = const()[name = tensor("encoder_output_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor module_decoder_layers_0_weight_to_fp16 = const()[name = tensor("module_decoder_layers_0_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; - tensor module_decoder_layers_0_bias_to_fp16 = const()[name = tensor("module_decoder_layers_0_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1049728)))]; - tensor encoder_output_to_fp16 = cast(dtype = encoder_output_to_fp16_dtype_0, x = encoder_output)[name = tensor("cast_1")]; - tensor var_18_cast_fp16 = conv(bias = module_decoder_layers_0_bias_to_fp16, dilations = var_18_dilations_0, groups = var_18_groups_0, pad = var_18_pad_0, pad_type = var_18_pad_type_0, strides = var_18_strides_0, weight = module_decoder_layers_0_weight_to_fp16, x = encoder_output_to_fp16)[name = tensor("op_18_cast_fp16")]; - tensor input_perm_0 = const()[name = tensor("input_perm_0"), val = tensor([0, 2, 1])]; - tensor input_cast_fp16 = transpose(perm = input_perm_0, x = var_18_cast_fp16)[name = tensor("transpose_0")]; - tensor out_objects_softmax_cast_fp16 = softmax(axis = var_4, x = input_cast_fp16)[name = tensor("out_objects_softmax_cast_fp16")]; - tensor out_objects_epsilon_0 = const()[name = tensor("out_objects_epsilon_0"), val = tensor(0x1p-149)]; - tensor out_objects_cast_fp16 = log(epsilon = out_objects_epsilon_0, x = out_objects_softmax_cast_fp16)[name = tensor("out_objects_cast_fp16")]; - tensor out_objects_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("out_objects_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; - tensor ctc_logits = cast(dtype = out_objects_cast_fp16_to_fp32_dtype_0, x = out_objects_cast_fp16)[name = tensor("cast_0")]; - } -> (ctc_logits); -} \ No newline at end of file diff --git a/models/parakeet-ctc-110m-coreml/CTCHead.mlmodelc/weights/weight.bin b/models/parakeet-ctc-110m-coreml/CTCHead.mlmodelc/weights/weight.bin deleted file mode 100644 index 23cedd570125191064c6111997f08c48898245f7..0000000000000000000000000000000000000000 --- a/models/parakeet-ctc-110m-coreml/CTCHead.mlmodelc/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:fb9bead064427ffcb7529c0e3f378e421b4dde8e6d81447b6d1ca3352ca850e1 -size 1051842 diff --git a/models/parakeet-ctc-110m-coreml/Decoder.mlmodelc/analytics/coremldata.bin b/models/parakeet-ctc-110m-coreml/Decoder.mlmodelc/analytics/coremldata.bin deleted file mode 100644 index dd015548a03cc453e6477a24374e8fee955e75ef..0000000000000000000000000000000000000000 --- a/models/parakeet-ctc-110m-coreml/Decoder.mlmodelc/analytics/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:990455f6431342750254f66edf27bfb41be62a7ba17a18e1dd6afd4f5f56e9eb -size 243 diff --git a/models/parakeet-ctc-110m-coreml/Decoder.mlmodelc/coremldata.bin b/models/parakeet-ctc-110m-coreml/Decoder.mlmodelc/coremldata.bin deleted file mode 100644 index cbea0affef0c3bb63be750e134dc0c9ed70aadb1..0000000000000000000000000000000000000000 --- a/models/parakeet-ctc-110m-coreml/Decoder.mlmodelc/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:29009727821ad8551ab5fe9271e93c597d92a9714f64b94aa533a9ceb6e22b93 -size 498 diff --git a/models/parakeet-ctc-110m-coreml/Decoder.mlmodelc/metadata.json b/models/parakeet-ctc-110m-coreml/Decoder.mlmodelc/metadata.json deleted file mode 100644 index 786f4bb2823b1aa7ebb55fe335c9385ee0065b7c..0000000000000000000000000000000000000000 --- a/models/parakeet-ctc-110m-coreml/Decoder.mlmodelc/metadata.json +++ /dev/null @@ -1,118 +0,0 @@ -[ - { - "metadataOutputVersion" : "3.0", - "shortDescription" : "Parakeet 110M decoder (RNNT prediction network)", - "outputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 640 × 1)", - "shortDescription" : "", - "shape" : "[1, 640, 1]", - "name" : "decoder", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 1 × 640)", - "shortDescription" : "", - "shape" : "[1, 1, 640]", - "name" : "h_out", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 1 × 640)", - "shortDescription" : "", - "shape" : "[1, 1, 640]", - "name" : "c_out", - "type" : "MultiArray" - } - ], - "storagePrecision" : "Float16", - "modelParameters" : [ - - ], - "author" : "Fluid Inference", - "specificationVersion" : 8, - "mlProgramOperationTypeHistogram" : { - "Ios17.squeeze" : 2, - "Ios17.gather" : 1, - "Ios17.cast" : 6, - "Ios17.lstm" : 1, - "Ios17.transpose" : 2, - "Identity" : 1, - "Ios17.expandDims" : 2 - }, - "computePrecision" : "Mixed (Float16, Float32, Int16, Int32)", - "isUpdatable" : "0", - "stateSchema" : [ - - ], - "availability" : { - "macOS" : "14.0", - "tvOS" : "17.0", - "visionOS" : "1.0", - "watchOS" : "10.0", - "iOS" : "17.0", - "macCatalyst" : "17.0" - }, - "modelType" : { - "name" : "MLModelType_mlProgram" - }, - "inputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1 × 1)", - "shortDescription" : "", - "shape" : "[1, 1]", - "name" : "targets", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1)", - "shortDescription" : "", - "shape" : "[1]", - "name" : "target_length", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 1 × 640)", - "shortDescription" : "", - "shape" : "[1, 1, 640]", - "name" : "h_in", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 1 × 640)", - "shortDescription" : "", - "shape" : "[1, 1, 640]", - "name" : "c_in", - "type" : "MultiArray" - } - ], - "userDefinedMetadata" : { - "com.github.apple.coremltools.version" : "8.3.0", - "com.github.apple.coremltools.source_dialect" : "TorchScript", - "com.github.apple.coremltools.source" : "torch==2.9.0" - }, - "generatedClassName" : "parakeet_decoder", - "method" : "predict" - } -] \ No newline at end of file diff --git a/models/parakeet-ctc-110m-coreml/Decoder.mlmodelc/model.mil b/models/parakeet-ctc-110m-coreml/Decoder.mlmodelc/model.mil deleted file mode 100644 index f69c7247525bc9d95f9341acc4f11eca709c7d00..0000000000000000000000000000000000000000 --- a/models/parakeet-ctc-110m-coreml/Decoder.mlmodelc/model.mil +++ /dev/null @@ -1,45 +0,0 @@ -program(1.0) -[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3510.2.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.9.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "8.3.0"}})] -{ - func main(tensor c_in, tensor h_in, tensor target_length, tensor targets) { - tensor y_axis_0 = const()[name = tensor("y_axis_0"), val = tensor(0)]; - tensor y_batch_dims_0 = const()[name = tensor("y_batch_dims_0"), val = tensor(0)]; - tensor y_validate_indices_0 = const()[name = tensor("y_validate_indices_0"), val = tensor(false)]; - tensor module_prediction_embed_weight_to_fp16 = const()[name = tensor("module_prediction_embed_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; - tensor targets_to_int16_dtype_0 = const()[name = tensor("targets_to_int16_dtype_0"), val = tensor("int16")]; - tensor targets_to_int16 = cast(dtype = targets_to_int16_dtype_0, x = targets)[name = tensor("cast_8")]; - tensor y_cast_fp16_cast_uint16 = gather(axis = y_axis_0, batch_dims = y_batch_dims_0, indices = targets_to_int16, validate_indices = y_validate_indices_0, x = module_prediction_embed_weight_to_fp16)[name = tensor("y_cast_fp16_cast_uint16")]; - tensor input_3_perm_0 = const()[name = tensor("input_3_perm_0"), val = tensor([1, 0, 2])]; - tensor input_lstm_h0_squeeze_axes_0 = const()[name = tensor("input_lstm_h0_squeeze_axes_0"), val = tensor([0])]; - tensor h_in_to_fp16_dtype_0 = const()[name = tensor("h_in_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor h_in_to_fp16 = cast(dtype = h_in_to_fp16_dtype_0, x = h_in)[name = tensor("cast_7")]; - tensor input_lstm_h0_squeeze_cast_fp16 = squeeze(axes = input_lstm_h0_squeeze_axes_0, x = h_in_to_fp16)[name = tensor("input_lstm_h0_squeeze_cast_fp16")]; - tensor input_lstm_c0_squeeze_axes_0 = const()[name = tensor("input_lstm_c0_squeeze_axes_0"), val = tensor([0])]; - tensor c_in_to_fp16_dtype_0 = const()[name = tensor("c_in_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor c_in_to_fp16 = cast(dtype = c_in_to_fp16_dtype_0, x = c_in)[name = tensor("cast_6")]; - tensor input_lstm_c0_squeeze_cast_fp16 = squeeze(axes = input_lstm_c0_squeeze_axes_0, x = c_in_to_fp16)[name = tensor("input_lstm_c0_squeeze_cast_fp16")]; - tensor input_direction_0 = const()[name = tensor("input_direction_0"), val = tensor("forward")]; - tensor input_output_sequence_0 = const()[name = tensor("input_output_sequence_0"), val = tensor(true)]; - tensor input_recurrent_activation_0 = const()[name = tensor("input_recurrent_activation_0"), val = tensor("sigmoid")]; - tensor input_cell_activation_0 = const()[name = tensor("input_cell_activation_0"), val = tensor("tanh")]; - tensor input_activation_0 = const()[name = tensor("input_activation_0"), val = tensor("tanh")]; - tensor concat_1_to_fp16 = const()[name = tensor("concat_1_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1312128)))]; - tensor concat_2_to_fp16 = const()[name = tensor("concat_2_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4588992)))]; - tensor concat_0_to_fp16 = const()[name = tensor("concat_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7865856)))]; - tensor input_3_cast_fp16 = transpose(perm = input_3_perm_0, x = y_cast_fp16_cast_uint16)[name = tensor("transpose_2")]; - tensor input_cast_fp16_0, tensor input_cast_fp16_1, tensor input_cast_fp16_2 = lstm(activation = input_activation_0, bias = concat_0_to_fp16, cell_activation = input_cell_activation_0, direction = input_direction_0, initial_c = input_lstm_c0_squeeze_cast_fp16, initial_h = input_lstm_h0_squeeze_cast_fp16, output_sequence = input_output_sequence_0, recurrent_activation = input_recurrent_activation_0, weight_hh = concat_2_to_fp16, weight_ih = concat_1_to_fp16, x = input_3_cast_fp16)[name = tensor("input_cast_fp16")]; - tensor obj_3_axes_0 = const()[name = tensor("obj_3_axes_0"), val = tensor([0])]; - tensor obj_3_cast_fp16 = expand_dims(axes = obj_3_axes_0, x = input_cast_fp16_1)[name = tensor("obj_3_cast_fp16")]; - tensor obj_3_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("obj_3_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; - tensor obj_axes_0 = const()[name = tensor("obj_axes_0"), val = tensor([0])]; - tensor obj_cast_fp16 = expand_dims(axes = obj_axes_0, x = input_cast_fp16_2)[name = tensor("obj_cast_fp16")]; - tensor obj_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("obj_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; - tensor transpose_0_perm_0 = const()[name = tensor("transpose_0_perm_0"), val = tensor([1, 2, 0])]; - tensor transpose_0_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("transpose_0_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; - tensor transpose_0_cast_fp16 = transpose(perm = transpose_0_perm_0, x = input_cast_fp16_0)[name = tensor("transpose_1")]; - tensor decoder = cast(dtype = transpose_0_cast_fp16_to_fp32_dtype_0, x = transpose_0_cast_fp16)[name = tensor("cast_3")]; - tensor c_out = cast(dtype = obj_cast_fp16_to_fp32_dtype_0, x = obj_cast_fp16)[name = tensor("cast_4")]; - tensor h_out = cast(dtype = obj_3_cast_fp16_to_fp32_dtype_0, x = obj_3_cast_fp16)[name = tensor("cast_5")]; - tensor target_length_tmp = identity(x = target_length)[name = tensor("target_length_tmp")]; - } -> (decoder, h_out, c_out); -} \ No newline at end of file diff --git a/models/parakeet-ctc-110m-coreml/Decoder.mlmodelc/weights/weight.bin b/models/parakeet-ctc-110m-coreml/Decoder.mlmodelc/weights/weight.bin deleted file mode 100644 index 4543ad4ea89b21ebed0de0d5ec3975d07eace7fb..0000000000000000000000000000000000000000 --- a/models/parakeet-ctc-110m-coreml/Decoder.mlmodelc/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:dd90b58597ee2c172c672dffe13b1110898ba07394c1a15efc96cc8c6b18411b -size 7871040 diff --git a/models/parakeet-ctc-110m-coreml/Encoder.mlmodelc/analytics/coremldata.bin b/models/parakeet-ctc-110m-coreml/Encoder.mlmodelc/analytics/coremldata.bin deleted file mode 100644 index d2787d4e5376d7f894d2b0a5a09f497a9b486c73..0000000000000000000000000000000000000000 --- a/models/parakeet-ctc-110m-coreml/Encoder.mlmodelc/analytics/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:b7ae65e2af616df46066b7efca2d7c19941666ac0685f4ed005666890a052b0d -size 243 diff --git a/models/parakeet-ctc-110m-coreml/Encoder.mlmodelc/coremldata.bin b/models/parakeet-ctc-110m-coreml/Encoder.mlmodelc/coremldata.bin deleted file mode 100644 index 9d9cdbfd392e8f7391b47c1c96f0fcbaab440389..0000000000000000000000000000000000000000 --- a/models/parakeet-ctc-110m-coreml/Encoder.mlmodelc/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:0713c2d6ac5f8f6fb9582be250351ebd8efc925f71f4261191165f1406f2ee5d -size 437 diff --git a/models/parakeet-ctc-110m-coreml/Encoder.mlmodelc/metadata.json b/models/parakeet-ctc-110m-coreml/Encoder.mlmodelc/metadata.json deleted file mode 100644 index 8e3f6cc9ddca206205987b2c828d2e595e19bfc2..0000000000000000000000000000000000000000 --- a/models/parakeet-ctc-110m-coreml/Encoder.mlmodelc/metadata.json +++ /dev/null @@ -1,105 +0,0 @@ -[ - { - "metadataOutputVersion" : "3.0", - "shortDescription" : "Parakeet 110M encoder (15 s window)", - "outputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 512 × 188)", - "shortDescription" : "", - "shape" : "[1, 512, 188]", - "name" : "encoder_output", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1)", - "shortDescription" : "", - "shape" : "[1]", - "name" : "encoder_length", - "type" : "MultiArray" - } - ], - "storagePrecision" : "Float16", - "modelParameters" : [ - - ], - "author" : "Fluid Inference", - "specificationVersion" : 8, - "mlProgramOperationTypeHistogram" : { - "Ios17.logicalAnd" : 2, - "Ios17.reshape" : 103, - "Ios16.softmax" : 17, - "Ios17.matmul" : 51, - "Ios17.transpose" : 123, - "Split" : 17, - "Ios17.expandDims" : 17, - "Select" : 51, - "Ios17.add" : 128, - "Tile" : 8, - "Ios17.sliceByIndex" : 34, - "Ios16.sigmoid" : 17, - "Pad" : 34, - "Ios17.logicalNot" : 2, - "Ios17.layerNorm" : 85, - "Ios16.silu" : 51, - "Ios17.less" : 5, - "Ios17.sub" : 3, - "Ios17.conv" : 56, - "Ios16.relu" : 3, - "Ios17.linear" : 137, - "Ios17.cast" : 11, - "Ios17.floorDiv" : 3, - "Ios17.mul" : 77 - }, - "computePrecision" : "Mixed (Float16, Float32, Int32)", - "isUpdatable" : "0", - "stateSchema" : [ - - ], - "availability" : { - "macOS" : "14.0", - "tvOS" : "17.0", - "visionOS" : "1.0", - "watchOS" : "10.0", - "iOS" : "17.0", - "macCatalyst" : "17.0" - }, - "modelType" : { - "name" : "MLModelType_mlProgram" - }, - "inputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 80 × 1501)", - "shortDescription" : "", - "shape" : "[1, 80, 1501]", - "name" : "mel_features", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1)", - "shortDescription" : "", - "shape" : "[1]", - "name" : "mel_length", - "type" : "MultiArray" - } - ], - "userDefinedMetadata" : { - "com.github.apple.coremltools.source_dialect" : "TorchScript", - "com.github.apple.coremltools.source" : "torch==2.9.0", - "com.github.apple.coremltools.version" : "8.3.0" - }, - "generatedClassName" : "parakeet_encoder", - "method" : "predict" - } -] \ No newline at end of file diff --git a/models/parakeet-ctc-110m-coreml/Encoder.mlmodelc/model.mil b/models/parakeet-ctc-110m-coreml/Encoder.mlmodelc/model.mil deleted file mode 100644 index 344bd1abf313c03c9ed9287c00893739f5d361bd..0000000000000000000000000000000000000000 --- a/models/parakeet-ctc-110m-coreml/Encoder.mlmodelc/model.mil +++ /dev/null @@ -1,2690 +0,0 @@ -program(1.0) -[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3510.2.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.9.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "8.3.0"}})] -{ - func main(tensor mel_features, tensor mel_length) { - tensor var_23 = const()[name = tensor("op_23"), val = tensor(-1)]; - tensor x_1_perm_0 = const()[name = tensor("x_1_perm_0"), val = tensor([0, 2, 1])]; - tensor mel_features_to_fp16_dtype_0 = const()[name = tensor("mel_features_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor tensor_1_axes_0 = const()[name = tensor("tensor_1_axes_0"), val = tensor([1])]; - tensor mel_features_to_fp16 = cast(dtype = mel_features_to_fp16_dtype_0, x = mel_features)[name = tensor("cast_182")]; - tensor x_1_cast_fp16 = transpose(perm = x_1_perm_0, x = mel_features_to_fp16)[name = tensor("transpose_207")]; - tensor tensor_1_cast_fp16 = expand_dims(axes = tensor_1_axes_0, x = x_1_cast_fp16)[name = tensor("tensor_1_cast_fp16")]; - tensor expand_dims_0 = const()[name = tensor("expand_dims_0"), val = tensor([[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390, 391, 392, 393, 394, 395, 396, 397, 398, 399, 400, 401, 402, 403, 404, 405, 406, 407, 408, 409, 410, 411, 412, 413, 414, 415, 416, 417, 418, 419, 420, 421, 422, 423, 424, 425, 426, 427, 428, 429, 430, 431, 432, 433, 434, 435, 436, 437, 438, 439, 440, 441, 442, 443, 444, 445, 446, 447, 448, 449, 450, 451, 452, 453, 454, 455, 456, 457, 458, 459, 460, 461, 462, 463, 464, 465, 466, 467, 468, 469, 470, 471, 472, 473, 474, 475, 476, 477, 478, 479, 480, 481, 482, 483, 484, 485, 486, 487, 488, 489, 490, 491, 492, 493, 494, 495, 496, 497, 498, 499, 500, 501, 502, 503, 504, 505, 506, 507, 508, 509, 510, 511, 512, 513, 514, 515, 516, 517, 518, 519, 520, 521, 522, 523, 524, 525, 526, 527, 528, 529, 530, 531, 532, 533, 534, 535, 536, 537, 538, 539, 540, 541, 542, 543, 544, 545, 546, 547, 548, 549, 550, 551, 552, 553, 554, 555, 556, 557, 558, 559, 560, 561, 562, 563, 564, 565, 566, 567, 568, 569, 570, 571, 572, 573, 574, 575, 576, 577, 578, 579, 580, 581, 582, 583, 584, 585, 586, 587, 588, 589, 590, 591, 592, 593, 594, 595, 596, 597, 598, 599, 600, 601, 602, 603, 604, 605, 606, 607, 608, 609, 610, 611, 612, 613, 614, 615, 616, 617, 618, 619, 620, 621, 622, 623, 624, 625, 626, 627, 628, 629, 630, 631, 632, 633, 634, 635, 636, 637, 638, 639, 640, 641, 642, 643, 644, 645, 646, 647, 648, 649, 650, 651, 652, 653, 654, 655, 656, 657, 658, 659, 660, 661, 662, 663, 664, 665, 666, 667, 668, 669, 670, 671, 672, 673, 674, 675, 676, 677, 678, 679, 680, 681, 682, 683, 684, 685, 686, 687, 688, 689, 690, 691, 692, 693, 694, 695, 696, 697, 698, 699, 700, 701, 702, 703, 704, 705, 706, 707, 708, 709, 710, 711, 712, 713, 714, 715, 716, 717, 718, 719, 720, 721, 722, 723, 724, 725, 726, 727, 728, 729, 730, 731, 732, 733, 734, 735, 736, 737, 738, 739, 740, 741, 742, 743, 744, 745, 746, 747, 748, 749, 750, 751, 752, 753, 754, 755, 756, 757, 758, 759, 760, 761, 762, 763, 764, 765, 766, 767, 768, 769, 770, 771, 772, 773, 774, 775, 776, 777, 778, 779, 780, 781, 782, 783, 784, 785, 786, 787, 788, 789, 790, 791, 792, 793, 794, 795, 796, 797, 798, 799, 800, 801, 802, 803, 804, 805, 806, 807, 808, 809, 810, 811, 812, 813, 814, 815, 816, 817, 818, 819, 820, 821, 822, 823, 824, 825, 826, 827, 828, 829, 830, 831, 832, 833, 834, 835, 836, 837, 838, 839, 840, 841, 842, 843, 844, 845, 846, 847, 848, 849, 850, 851, 852, 853, 854, 855, 856, 857, 858, 859, 860, 861, 862, 863, 864, 865, 866, 867, 868, 869, 870, 871, 872, 873, 874, 875, 876, 877, 878, 879, 880, 881, 882, 883, 884, 885, 886, 887, 888, 889, 890, 891, 892, 893, 894, 895, 896, 897, 898, 899, 900, 901, 902, 903, 904, 905, 906, 907, 908, 909, 910, 911, 912, 913, 914, 915, 916, 917, 918, 919, 920, 921, 922, 923, 924, 925, 926, 927, 928, 929, 930, 931, 932, 933, 934, 935, 936, 937, 938, 939, 940, 941, 942, 943, 944, 945, 946, 947, 948, 949, 950, 951, 952, 953, 954, 955, 956, 957, 958, 959, 960, 961, 962, 963, 964, 965, 966, 967, 968, 969, 970, 971, 972, 973, 974, 975, 976, 977, 978, 979, 980, 981, 982, 983, 984, 985, 986, 987, 988, 989, 990, 991, 992, 993, 994, 995, 996, 997, 998, 999, 1000, 1001, 1002, 1003, 1004, 1005, 1006, 1007, 1008, 1009, 1010, 1011, 1012, 1013, 1014, 1015, 1016, 1017, 1018, 1019, 1020, 1021, 1022, 1023, 1024, 1025, 1026, 1027, 1028, 1029, 1030, 1031, 1032, 1033, 1034, 1035, 1036, 1037, 1038, 1039, 1040, 1041, 1042, 1043, 1044, 1045, 1046, 1047, 1048, 1049, 1050, 1051, 1052, 1053, 1054, 1055, 1056, 1057, 1058, 1059, 1060, 1061, 1062, 1063, 1064, 1065, 1066, 1067, 1068, 1069, 1070, 1071, 1072, 1073, 1074, 1075, 1076, 1077, 1078, 1079, 1080, 1081, 1082, 1083, 1084, 1085, 1086, 1087, 1088, 1089, 1090, 1091, 1092, 1093, 1094, 1095, 1096, 1097, 1098, 1099, 1100, 1101, 1102, 1103, 1104, 1105, 1106, 1107, 1108, 1109, 1110, 1111, 1112, 1113, 1114, 1115, 1116, 1117, 1118, 1119, 1120, 1121, 1122, 1123, 1124, 1125, 1126, 1127, 1128, 1129, 1130, 1131, 1132, 1133, 1134, 1135, 1136, 1137, 1138, 1139, 1140, 1141, 1142, 1143, 1144, 1145, 1146, 1147, 1148, 1149, 1150, 1151, 1152, 1153, 1154, 1155, 1156, 1157, 1158, 1159, 1160, 1161, 1162, 1163, 1164, 1165, 1166, 1167, 1168, 1169, 1170, 1171, 1172, 1173, 1174, 1175, 1176, 1177, 1178, 1179, 1180, 1181, 1182, 1183, 1184, 1185, 1186, 1187, 1188, 1189, 1190, 1191, 1192, 1193, 1194, 1195, 1196, 1197, 1198, 1199, 1200, 1201, 1202, 1203, 1204, 1205, 1206, 1207, 1208, 1209, 1210, 1211, 1212, 1213, 1214, 1215, 1216, 1217, 1218, 1219, 1220, 1221, 1222, 1223, 1224, 1225, 1226, 1227, 1228, 1229, 1230, 1231, 1232, 1233, 1234, 1235, 1236, 1237, 1238, 1239, 1240, 1241, 1242, 1243, 1244, 1245, 1246, 1247, 1248, 1249, 1250, 1251, 1252, 1253, 1254, 1255, 1256, 1257, 1258, 1259, 1260, 1261, 1262, 1263, 1264, 1265, 1266, 1267, 1268, 1269, 1270, 1271, 1272, 1273, 1274, 1275, 1276, 1277, 1278, 1279, 1280, 1281, 1282, 1283, 1284, 1285, 1286, 1287, 1288, 1289, 1290, 1291, 1292, 1293, 1294, 1295, 1296, 1297, 1298, 1299, 1300, 1301, 1302, 1303, 1304, 1305, 1306, 1307, 1308, 1309, 1310, 1311, 1312, 1313, 1314, 1315, 1316, 1317, 1318, 1319, 1320, 1321, 1322, 1323, 1324, 1325, 1326, 1327, 1328, 1329, 1330, 1331, 1332, 1333, 1334, 1335, 1336, 1337, 1338, 1339, 1340, 1341, 1342, 1343, 1344, 1345, 1346, 1347, 1348, 1349, 1350, 1351, 1352, 1353, 1354, 1355, 1356, 1357, 1358, 1359, 1360, 1361, 1362, 1363, 1364, 1365, 1366, 1367, 1368, 1369, 1370, 1371, 1372, 1373, 1374, 1375, 1376, 1377, 1378, 1379, 1380, 1381, 1382, 1383, 1384, 1385, 1386, 1387, 1388, 1389, 1390, 1391, 1392, 1393, 1394, 1395, 1396, 1397, 1398, 1399, 1400, 1401, 1402, 1403, 1404, 1405, 1406, 1407, 1408, 1409, 1410, 1411, 1412, 1413, 1414, 1415, 1416, 1417, 1418, 1419, 1420, 1421, 1422, 1423, 1424, 1425, 1426, 1427, 1428, 1429, 1430, 1431, 1432, 1433, 1434, 1435, 1436, 1437, 1438, 1439, 1440, 1441, 1442, 1443, 1444, 1445, 1446, 1447, 1448, 1449, 1450, 1451, 1452, 1453, 1454, 1455, 1456, 1457, 1458, 1459, 1460, 1461, 1462, 1463, 1464, 1465, 1466, 1467, 1468, 1469, 1470, 1471, 1472, 1473, 1474, 1475, 1476, 1477, 1478, 1479, 1480, 1481, 1482, 1483, 1484, 1485, 1486, 1487, 1488, 1489, 1490, 1491, 1492, 1493, 1494, 1495, 1496, 1497, 1498, 1499, 1500]])]; - tensor var_87_axes_0 = const()[name = tensor("op_87_axes_0"), val = tensor([1])]; - tensor var_87 = expand_dims(axes = var_87_axes_0, x = mel_length)[name = tensor("op_87")]; - tensor time_mask_1 = less(x = expand_dims_0, y = var_87)[name = tensor("time_mask_1")]; - tensor var_89_axes_0 = const()[name = tensor("op_89_axes_0"), val = tensor([-1])]; - tensor var_89 = expand_dims(axes = var_89_axes_0, x = time_mask_1)[name = tensor("op_89")]; - tensor var_91_reps_0 = const()[name = tensor("op_91_reps_0"), val = tensor([1, 1, 80])]; - tensor var_91 = tile(reps = var_91_reps_0, x = var_89)[name = tensor("op_91")]; - tensor var_97_axes_0 = const()[name = tensor("op_97_axes_0"), val = tensor([1])]; - tensor cast_3_to_fp16_dtype_0 = const()[name = tensor("cast_3_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor var_91_to_fp16 = cast(dtype = cast_3_to_fp16_dtype_0, x = var_91)[name = tensor("cast_181")]; - tensor var_97_cast_fp16 = expand_dims(axes = var_97_axes_0, x = var_91_to_fp16)[name = tensor("op_97_cast_fp16")]; - tensor input_1_cast_fp16 = mul(x = tensor_1_cast_fp16, y = var_97_cast_fp16)[name = tensor("input_1_cast_fp16")]; - tensor tensor_3_pad_type_0 = const()[name = tensor("tensor_3_pad_type_0"), val = tensor("custom")]; - tensor tensor_3_pad_0 = const()[name = tensor("tensor_3_pad_0"), val = tensor([1, 1, 1, 1])]; - tensor tensor_3_strides_0 = const()[name = tensor("tensor_3_strides_0"), val = tensor([2, 2])]; - tensor tensor_3_dilations_0 = const()[name = tensor("tensor_3_dilations_0"), val = tensor([1, 1])]; - tensor tensor_3_groups_0 = const()[name = tensor("tensor_3_groups_0"), val = tensor(1)]; - tensor module_pre_encode_conv_0_weight_to_fp16 = const()[name = tensor("module_pre_encode_conv_0_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; - tensor module_pre_encode_conv_0_bias_to_fp16 = const()[name = tensor("module_pre_encode_conv_0_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4736)))]; - tensor tensor_3_cast_fp16 = conv(bias = module_pre_encode_conv_0_bias_to_fp16, dilations = tensor_3_dilations_0, groups = tensor_3_groups_0, pad = tensor_3_pad_0, pad_type = tensor_3_pad_type_0, strides = tensor_3_strides_0, weight = module_pre_encode_conv_0_weight_to_fp16, x = input_1_cast_fp16)[name = tensor("tensor_3_cast_fp16")]; - tensor cast_1_to_fp16_dtype_0 = const()[name = tensor("cast_1_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor var_108_promoted_to_fp16 = const()[name = tensor("op_108_promoted_to_fp16"), val = tensor(0x1p+0)]; - tensor mel_length_to_fp16 = cast(dtype = cast_1_to_fp16_dtype_0, x = mel_length)[name = tensor("cast_180")]; - tensor var_109_cast_fp16 = add(x = mel_length_to_fp16, y = var_108_promoted_to_fp16)[name = tensor("op_109_cast_fp16")]; - tensor var_110_promoted_to_fp16 = const()[name = tensor("op_110_promoted_to_fp16"), val = tensor(0x1p+0)]; - tensor var_111_cast_fp16 = add(x = var_109_cast_fp16, y = var_110_promoted_to_fp16)[name = tensor("op_111_cast_fp16")]; - tensor var_112_promoted_to_fp16 = const()[name = tensor("op_112_promoted_to_fp16"), val = tensor(0x1.8p+1)]; - tensor var_113_cast_fp16 = sub(x = var_111_cast_fp16, y = var_112_promoted_to_fp16)[name = tensor("op_113_cast_fp16")]; - tensor var_21_promoted_to_fp16 = const()[name = tensor("op_21_promoted_to_fp16"), val = tensor(0x1p+1)]; - tensor floor_div_0_cast_fp16 = floor_div(x = var_113_cast_fp16, y = var_21_promoted_to_fp16)[name = tensor("floor_div_0_cast_fp16")]; - tensor var_115_promoted_to_fp16 = const()[name = tensor("op_115_promoted_to_fp16"), val = tensor(0x1p+0)]; - tensor current_lengths_3_cast_fp16 = add(x = floor_div_0_cast_fp16, y = var_115_promoted_to_fp16)[name = tensor("current_lengths_3_cast_fp16")]; - tensor cast_4_dtype_0 = const()[name = tensor("cast_4_dtype_0"), val = tensor("int32")]; - tensor expand_dims_1 = const()[name = tensor("expand_dims_1"), val = tensor([[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390, 391, 392, 393, 394, 395, 396, 397, 398, 399, 400, 401, 402, 403, 404, 405, 406, 407, 408, 409, 410, 411, 412, 413, 414, 415, 416, 417, 418, 419, 420, 421, 422, 423, 424, 425, 426, 427, 428, 429, 430, 431, 432, 433, 434, 435, 436, 437, 438, 439, 440, 441, 442, 443, 444, 445, 446, 447, 448, 449, 450, 451, 452, 453, 454, 455, 456, 457, 458, 459, 460, 461, 462, 463, 464, 465, 466, 467, 468, 469, 470, 471, 472, 473, 474, 475, 476, 477, 478, 479, 480, 481, 482, 483, 484, 485, 486, 487, 488, 489, 490, 491, 492, 493, 494, 495, 496, 497, 498, 499, 500, 501, 502, 503, 504, 505, 506, 507, 508, 509, 510, 511, 512, 513, 514, 515, 516, 517, 518, 519, 520, 521, 522, 523, 524, 525, 526, 527, 528, 529, 530, 531, 532, 533, 534, 535, 536, 537, 538, 539, 540, 541, 542, 543, 544, 545, 546, 547, 548, 549, 550, 551, 552, 553, 554, 555, 556, 557, 558, 559, 560, 561, 562, 563, 564, 565, 566, 567, 568, 569, 570, 571, 572, 573, 574, 575, 576, 577, 578, 579, 580, 581, 582, 583, 584, 585, 586, 587, 588, 589, 590, 591, 592, 593, 594, 595, 596, 597, 598, 599, 600, 601, 602, 603, 604, 605, 606, 607, 608, 609, 610, 611, 612, 613, 614, 615, 616, 617, 618, 619, 620, 621, 622, 623, 624, 625, 626, 627, 628, 629, 630, 631, 632, 633, 634, 635, 636, 637, 638, 639, 640, 641, 642, 643, 644, 645, 646, 647, 648, 649, 650, 651, 652, 653, 654, 655, 656, 657, 658, 659, 660, 661, 662, 663, 664, 665, 666, 667, 668, 669, 670, 671, 672, 673, 674, 675, 676, 677, 678, 679, 680, 681, 682, 683, 684, 685, 686, 687, 688, 689, 690, 691, 692, 693, 694, 695, 696, 697, 698, 699, 700, 701, 702, 703, 704, 705, 706, 707, 708, 709, 710, 711, 712, 713, 714, 715, 716, 717, 718, 719, 720, 721, 722, 723, 724, 725, 726, 727, 728, 729, 730, 731, 732, 733, 734, 735, 736, 737, 738, 739, 740, 741, 742, 743, 744, 745, 746, 747, 748, 749, 750]])]; - tensor var_124_axes_0 = const()[name = tensor("op_124_axes_0"), val = tensor([1])]; - tensor current_lengths_3_cast_fp16_to_int32 = cast(dtype = cast_4_dtype_0, x = current_lengths_3_cast_fp16)[name = tensor("cast_179")]; - tensor var_124 = expand_dims(axes = var_124_axes_0, x = current_lengths_3_cast_fp16_to_int32)[name = tensor("op_124")]; - tensor time_mask_3 = less(x = expand_dims_1, y = var_124)[name = tensor("time_mask_3")]; - tensor var_126_axes_0 = const()[name = tensor("op_126_axes_0"), val = tensor([-1])]; - tensor var_126 = expand_dims(axes = var_126_axes_0, x = time_mask_3)[name = tensor("op_126")]; - tensor var_128_reps_0 = const()[name = tensor("op_128_reps_0"), val = tensor([1, 1, 40])]; - tensor var_128 = tile(reps = var_128_reps_0, x = var_126)[name = tensor("op_128")]; - tensor var_134_axes_0 = const()[name = tensor("op_134_axes_0"), val = tensor([1])]; - tensor cast_5_to_fp16_dtype_0 = const()[name = tensor("cast_5_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor var_128_to_fp16 = cast(dtype = cast_5_to_fp16_dtype_0, x = var_128)[name = tensor("cast_178")]; - tensor var_134_cast_fp16 = expand_dims(axes = var_134_axes_0, x = var_128_to_fp16)[name = tensor("op_134_cast_fp16")]; - tensor expanded_mask_3_reps_0 = const()[name = tensor("expanded_mask_3_reps_0"), val = tensor([1, 256, 1, 1])]; - tensor expanded_mask_3_cast_fp16 = tile(reps = expanded_mask_3_reps_0, x = var_134_cast_fp16)[name = tensor("expanded_mask_3_cast_fp16")]; - tensor input_3_cast_fp16 = mul(x = tensor_3_cast_fp16, y = expanded_mask_3_cast_fp16)[name = tensor("input_3_cast_fp16")]; - tensor tensor_5_cast_fp16 = relu(x = input_3_cast_fp16)[name = tensor("tensor_5_cast_fp16")]; - tensor input_5_cast_fp16 = mul(x = tensor_5_cast_fp16, y = expanded_mask_3_cast_fp16)[name = tensor("input_5_cast_fp16")]; - tensor tensor_7_pad_type_0 = const()[name = tensor("tensor_7_pad_type_0"), val = tensor("custom")]; - tensor tensor_7_pad_0 = const()[name = tensor("tensor_7_pad_0"), val = tensor([1, 1, 1, 1])]; - tensor tensor_7_strides_0 = const()[name = tensor("tensor_7_strides_0"), val = tensor([2, 2])]; - tensor tensor_7_groups_0 = const()[name = tensor("tensor_7_groups_0"), val = tensor(256)]; - tensor tensor_7_dilations_0 = const()[name = tensor("tensor_7_dilations_0"), val = tensor([1, 1])]; - tensor module_pre_encode_conv_2_weight_to_fp16 = const()[name = tensor("module_pre_encode_conv_2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(5312)))]; - tensor module_pre_encode_conv_2_bias_to_fp16 = const()[name = tensor("module_pre_encode_conv_2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9984)))]; - tensor tensor_7_cast_fp16 = conv(bias = module_pre_encode_conv_2_bias_to_fp16, dilations = tensor_7_dilations_0, groups = tensor_7_groups_0, pad = tensor_7_pad_0, pad_type = tensor_7_pad_type_0, strides = tensor_7_strides_0, weight = module_pre_encode_conv_2_weight_to_fp16, x = input_5_cast_fp16)[name = tensor("tensor_7_cast_fp16")]; - tensor var_154_promoted_to_fp16 = const()[name = tensor("op_154_promoted_to_fp16"), val = tensor(0x1p+0)]; - tensor var_155_cast_fp16 = add(x = current_lengths_3_cast_fp16, y = var_154_promoted_to_fp16)[name = tensor("op_155_cast_fp16")]; - tensor var_156_promoted_to_fp16 = const()[name = tensor("op_156_promoted_to_fp16"), val = tensor(0x1p+0)]; - tensor var_157_cast_fp16 = add(x = var_155_cast_fp16, y = var_156_promoted_to_fp16)[name = tensor("op_157_cast_fp16")]; - tensor var_158_promoted_to_fp16 = const()[name = tensor("op_158_promoted_to_fp16"), val = tensor(0x1.8p+1)]; - tensor var_159_cast_fp16 = sub(x = var_157_cast_fp16, y = var_158_promoted_to_fp16)[name = tensor("op_159_cast_fp16")]; - tensor var_21_promoted_1_to_fp16 = const()[name = tensor("op_21_promoted_1_to_fp16"), val = tensor(0x1p+1)]; - tensor floor_div_1_cast_fp16 = floor_div(x = var_159_cast_fp16, y = var_21_promoted_1_to_fp16)[name = tensor("floor_div_1_cast_fp16")]; - tensor var_161_promoted_to_fp16 = const()[name = tensor("op_161_promoted_to_fp16"), val = tensor(0x1p+0)]; - tensor current_lengths_5_cast_fp16 = add(x = floor_div_1_cast_fp16, y = var_161_promoted_to_fp16)[name = tensor("current_lengths_5_cast_fp16")]; - tensor cast_6_dtype_0 = const()[name = tensor("cast_6_dtype_0"), val = tensor("int32")]; - tensor expand_dims_2 = const()[name = tensor("expand_dims_2"), val = tensor([[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375]])]; - tensor var_170_axes_0 = const()[name = tensor("op_170_axes_0"), val = tensor([1])]; - tensor current_lengths_5_cast_fp16_to_int32 = cast(dtype = cast_6_dtype_0, x = current_lengths_5_cast_fp16)[name = tensor("cast_177")]; - tensor var_170 = expand_dims(axes = var_170_axes_0, x = current_lengths_5_cast_fp16_to_int32)[name = tensor("op_170")]; - tensor time_mask_5 = less(x = expand_dims_2, y = var_170)[name = tensor("time_mask_5")]; - tensor var_172_axes_0 = const()[name = tensor("op_172_axes_0"), val = tensor([-1])]; - tensor var_172 = expand_dims(axes = var_172_axes_0, x = time_mask_5)[name = tensor("op_172")]; - tensor var_174_reps_0 = const()[name = tensor("op_174_reps_0"), val = tensor([1, 1, 20])]; - tensor var_174 = tile(reps = var_174_reps_0, x = var_172)[name = tensor("op_174")]; - tensor var_180_axes_0 = const()[name = tensor("op_180_axes_0"), val = tensor([1])]; - tensor cast_7_to_fp16_dtype_0 = const()[name = tensor("cast_7_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor var_174_to_fp16 = cast(dtype = cast_7_to_fp16_dtype_0, x = var_174)[name = tensor("cast_176")]; - tensor var_180_cast_fp16 = expand_dims(axes = var_180_axes_0, x = var_174_to_fp16)[name = tensor("op_180_cast_fp16")]; - tensor expanded_mask_7_reps_0 = const()[name = tensor("expanded_mask_7_reps_0"), val = tensor([1, 256, 1, 1])]; - tensor expanded_mask_7_cast_fp16 = tile(reps = expanded_mask_7_reps_0, x = var_180_cast_fp16)[name = tensor("expanded_mask_7_cast_fp16")]; - tensor input_7_cast_fp16 = mul(x = tensor_7_cast_fp16, y = expanded_mask_7_cast_fp16)[name = tensor("input_7_cast_fp16")]; - tensor tensor_9_pad_type_0 = const()[name = tensor("tensor_9_pad_type_0"), val = tensor("valid")]; - tensor tensor_9_strides_0 = const()[name = tensor("tensor_9_strides_0"), val = tensor([1, 1])]; - tensor tensor_9_pad_0 = const()[name = tensor("tensor_9_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor tensor_9_dilations_0 = const()[name = tensor("tensor_9_dilations_0"), val = tensor([1, 1])]; - tensor tensor_9_groups_0 = const()[name = tensor("tensor_9_groups_0"), val = tensor(1)]; - tensor module_pre_encode_conv_3_weight_to_fp16 = const()[name = tensor("module_pre_encode_conv_3_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10560)))]; - tensor module_pre_encode_conv_3_bias_to_fp16 = const()[name = tensor("module_pre_encode_conv_3_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141696)))]; - tensor tensor_9_cast_fp16 = conv(bias = module_pre_encode_conv_3_bias_to_fp16, dilations = tensor_9_dilations_0, groups = tensor_9_groups_0, pad = tensor_9_pad_0, pad_type = tensor_9_pad_type_0, strides = tensor_9_strides_0, weight = module_pre_encode_conv_3_weight_to_fp16, x = input_7_cast_fp16)[name = tensor("tensor_9_cast_fp16")]; - tensor input_9_cast_fp16 = mul(x = tensor_9_cast_fp16, y = expanded_mask_7_cast_fp16)[name = tensor("input_9_cast_fp16")]; - tensor tensor_11_cast_fp16 = relu(x = input_9_cast_fp16)[name = tensor("tensor_11_cast_fp16")]; - tensor input_11_cast_fp16 = mul(x = tensor_11_cast_fp16, y = expanded_mask_7_cast_fp16)[name = tensor("input_11_cast_fp16")]; - tensor tensor_13_pad_type_0 = const()[name = tensor("tensor_13_pad_type_0"), val = tensor("custom")]; - tensor tensor_13_pad_0 = const()[name = tensor("tensor_13_pad_0"), val = tensor([1, 1, 1, 1])]; - tensor tensor_13_strides_0 = const()[name = tensor("tensor_13_strides_0"), val = tensor([2, 2])]; - tensor tensor_13_groups_0 = const()[name = tensor("tensor_13_groups_0"), val = tensor(256)]; - tensor tensor_13_dilations_0 = const()[name = tensor("tensor_13_dilations_0"), val = tensor([1, 1])]; - tensor module_pre_encode_conv_5_weight_to_fp16 = const()[name = tensor("module_pre_encode_conv_5_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(142272)))]; - tensor module_pre_encode_conv_5_bias_to_fp16 = const()[name = tensor("module_pre_encode_conv_5_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146944)))]; - tensor tensor_13_cast_fp16 = conv(bias = module_pre_encode_conv_5_bias_to_fp16, dilations = tensor_13_dilations_0, groups = tensor_13_groups_0, pad = tensor_13_pad_0, pad_type = tensor_13_pad_type_0, strides = tensor_13_strides_0, weight = module_pre_encode_conv_5_weight_to_fp16, x = input_11_cast_fp16)[name = tensor("tensor_13_cast_fp16")]; - tensor var_215_promoted_to_fp16 = const()[name = tensor("op_215_promoted_to_fp16"), val = tensor(0x1p+0)]; - tensor var_216_cast_fp16 = add(x = current_lengths_5_cast_fp16, y = var_215_promoted_to_fp16)[name = tensor("op_216_cast_fp16")]; - tensor var_217_promoted_to_fp16 = const()[name = tensor("op_217_promoted_to_fp16"), val = tensor(0x1p+0)]; - tensor var_218_cast_fp16 = add(x = var_216_cast_fp16, y = var_217_promoted_to_fp16)[name = tensor("op_218_cast_fp16")]; - tensor var_219_promoted_to_fp16 = const()[name = tensor("op_219_promoted_to_fp16"), val = tensor(0x1.8p+1)]; - tensor var_220_cast_fp16 = sub(x = var_218_cast_fp16, y = var_219_promoted_to_fp16)[name = tensor("op_220_cast_fp16")]; - tensor var_21_promoted_2_to_fp16 = const()[name = tensor("op_21_promoted_2_to_fp16"), val = tensor(0x1p+1)]; - tensor floor_div_2_cast_fp16 = floor_div(x = var_220_cast_fp16, y = var_21_promoted_2_to_fp16)[name = tensor("floor_div_2_cast_fp16")]; - tensor var_222_promoted_to_fp16 = const()[name = tensor("op_222_promoted_to_fp16"), val = tensor(0x1p+0)]; - tensor current_lengths_cast_fp16 = add(x = floor_div_2_cast_fp16, y = var_222_promoted_to_fp16)[name = tensor("current_lengths_cast_fp16")]; - tensor cast_8_dtype_0 = const()[name = tensor("cast_8_dtype_0"), val = tensor("int32")]; - tensor expand_dims_3 = const()[name = tensor("expand_dims_3"), val = tensor([[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187]])]; - tensor var_231_axes_0 = const()[name = tensor("op_231_axes_0"), val = tensor([1])]; - tensor current_lengths_cast_fp16_to_int32 = cast(dtype = cast_8_dtype_0, x = current_lengths_cast_fp16)[name = tensor("cast_175")]; - tensor var_231 = expand_dims(axes = var_231_axes_0, x = current_lengths_cast_fp16_to_int32)[name = tensor("op_231")]; - tensor time_mask = less(x = expand_dims_3, y = var_231)[name = tensor("time_mask")]; - tensor var_233_axes_0 = const()[name = tensor("op_233_axes_0"), val = tensor([-1])]; - tensor var_233 = expand_dims(axes = var_233_axes_0, x = time_mask)[name = tensor("op_233")]; - tensor var_235_reps_0 = const()[name = tensor("op_235_reps_0"), val = tensor([1, 1, 10])]; - tensor var_235 = tile(reps = var_235_reps_0, x = var_233)[name = tensor("op_235")]; - tensor var_241_axes_0 = const()[name = tensor("op_241_axes_0"), val = tensor([1])]; - tensor cast_9_to_fp16_dtype_0 = const()[name = tensor("cast_9_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor var_235_to_fp16 = cast(dtype = cast_9_to_fp16_dtype_0, x = var_235)[name = tensor("cast_174")]; - tensor var_241_cast_fp16 = expand_dims(axes = var_241_axes_0, x = var_235_to_fp16)[name = tensor("op_241_cast_fp16")]; - tensor expanded_mask_13_reps_0 = const()[name = tensor("expanded_mask_13_reps_0"), val = tensor([1, 256, 1, 1])]; - tensor expanded_mask_13_cast_fp16 = tile(reps = expanded_mask_13_reps_0, x = var_241_cast_fp16)[name = tensor("expanded_mask_13_cast_fp16")]; - tensor input_13_cast_fp16 = mul(x = tensor_13_cast_fp16, y = expanded_mask_13_cast_fp16)[name = tensor("input_13_cast_fp16")]; - tensor tensor_15_pad_type_0 = const()[name = tensor("tensor_15_pad_type_0"), val = tensor("valid")]; - tensor tensor_15_strides_0 = const()[name = tensor("tensor_15_strides_0"), val = tensor([1, 1])]; - tensor tensor_15_pad_0 = const()[name = tensor("tensor_15_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor tensor_15_dilations_0 = const()[name = tensor("tensor_15_dilations_0"), val = tensor([1, 1])]; - tensor tensor_15_groups_0 = const()[name = tensor("tensor_15_groups_0"), val = tensor(1)]; - tensor module_pre_encode_conv_6_weight_to_fp16 = const()[name = tensor("module_pre_encode_conv_6_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147520)))]; - tensor module_pre_encode_conv_6_bias_to_fp16 = const()[name = tensor("module_pre_encode_conv_6_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(278656)))]; - tensor tensor_15_cast_fp16 = conv(bias = module_pre_encode_conv_6_bias_to_fp16, dilations = tensor_15_dilations_0, groups = tensor_15_groups_0, pad = tensor_15_pad_0, pad_type = tensor_15_pad_type_0, strides = tensor_15_strides_0, weight = module_pre_encode_conv_6_weight_to_fp16, x = input_13_cast_fp16)[name = tensor("tensor_15_cast_fp16")]; - tensor input_15_cast_fp16 = mul(x = tensor_15_cast_fp16, y = expanded_mask_13_cast_fp16)[name = tensor("input_15_cast_fp16")]; - tensor tensor_cast_fp16 = relu(x = input_15_cast_fp16)[name = tensor("tensor_cast_fp16")]; - tensor x_3_cast_fp16 = mul(x = tensor_cast_fp16, y = expanded_mask_13_cast_fp16)[name = tensor("x_3_cast_fp16")]; - tensor var_275_perm_0 = const()[name = tensor("op_275_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_276 = const()[name = tensor("op_276"), val = tensor([1, 188, -1])]; - tensor var_275_cast_fp16 = transpose(perm = var_275_perm_0, x = x_3_cast_fp16)[name = tensor("transpose_206")]; - tensor input_17_cast_fp16 = reshape(shape = var_276, x = var_275_cast_fp16)[name = tensor("input_17_cast_fp16")]; - tensor module_pre_encode_out_weight_to_fp16 = const()[name = tensor("module_pre_encode_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(279232)))]; - tensor module_pre_encode_out_bias_to_fp16 = const()[name = tensor("module_pre_encode_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2900736)))]; - tensor linear_0_cast_fp16 = linear(bias = module_pre_encode_out_bias_to_fp16, weight = module_pre_encode_out_weight_to_fp16, x = input_17_cast_fp16)[name = tensor("linear_0_cast_fp16")]; - tensor cast_12_dtype_0 = const()[name = tensor("cast_12_dtype_0"), val = tensor("int32")]; - tensor var_314_axes_0 = const()[name = tensor("op_314_axes_0"), val = tensor([-1])]; - tensor encoder_length = cast(dtype = cast_12_dtype_0, x = current_lengths_cast_fp16)[name = tensor("cast_173")]; - tensor var_314 = expand_dims(axes = var_314_axes_0, x = encoder_length)[name = tensor("op_314")]; - tensor pad_mask_1 = less(x = expand_dims_3, y = var_314)[name = tensor("pad_mask_1")]; - tensor var_316_axes_0 = const()[name = tensor("op_316_axes_0"), val = tensor([1])]; - tensor var_316 = expand_dims(axes = var_316_axes_0, x = pad_mask_1)[name = tensor("op_316")]; - tensor var_317 = const()[name = tensor("op_317"), val = tensor([1, 188, 1])]; - tensor pad_mask_for_att_mask_1 = tile(reps = var_317, x = var_316)[name = tensor("pad_mask_for_att_mask_1")]; - tensor var_319_perm_0 = const()[name = tensor("op_319_perm_0"), val = tensor([0, 2, 1])]; - tensor var_319 = transpose(perm = var_319_perm_0, x = pad_mask_for_att_mask_1)[name = tensor("transpose_205")]; - tensor pad_mask_for_att_mask = logical_and(x = pad_mask_for_att_mask_1, y = var_319)[name = tensor("pad_mask_for_att_mask")]; - tensor const_63 = const()[name = tensor("const_63"), val = tensor([[[true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true]]])]; - tensor att_mask = logical_and(x = pad_mask_for_att_mask, y = const_63)[name = tensor("att_mask")]; - tensor mask_9 = logical_not(x = att_mask)[name = tensor("mask_9")]; - tensor pad_mask = logical_not(x = pad_mask_1)[name = tensor("pad_mask")]; - tensor input_21_axes_0 = const()[name = tensor("input_21_axes_0"), val = tensor([-1])]; - tensor module_layers_0_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_0_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2901824)))]; - tensor module_layers_0_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_0_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2902912)))]; - tensor var_9_to_fp16 = const()[name = tensor("op_9_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_21_cast_fp16 = layer_norm(axes = input_21_axes_0, beta = module_layers_0_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_0_norm_feed_forward1_weight_to_fp16, x = linear_0_cast_fp16)[name = tensor("input_21_cast_fp16")]; - tensor module_layers_0_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_0_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2904000)))]; - tensor module_layers_0_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_0_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(5001216)))]; - tensor linear_1_cast_fp16 = linear(bias = module_layers_0_feed_forward1_linear1_bias_to_fp16, weight = module_layers_0_feed_forward1_linear1_weight_to_fp16, x = input_21_cast_fp16)[name = tensor("linear_1_cast_fp16")]; - tensor input_25_cast_fp16 = silu(x = linear_1_cast_fp16)[name = tensor("input_25_cast_fp16")]; - tensor module_layers_0_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_0_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(5005376)))]; - tensor module_layers_0_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_0_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7102592)))]; - tensor linear_2_cast_fp16 = linear(bias = module_layers_0_feed_forward1_linear2_bias_to_fp16, weight = module_layers_0_feed_forward1_linear2_weight_to_fp16, x = input_25_cast_fp16)[name = tensor("linear_2_cast_fp16")]; - tensor var_352_to_fp16 = const()[name = tensor("op_352_to_fp16"), val = tensor(0x1p-1)]; - tensor var_353_cast_fp16 = mul(x = linear_2_cast_fp16, y = var_352_to_fp16)[name = tensor("op_353_cast_fp16")]; - tensor input_31_cast_fp16 = add(x = linear_0_cast_fp16, y = var_353_cast_fp16)[name = tensor("input_31_cast_fp16")]; - tensor query_1_axes_0 = const()[name = tensor("query_1_axes_0"), val = tensor([-1])]; - tensor module_layers_0_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_0_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7103680)))]; - tensor module_layers_0_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_0_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7104768)))]; - tensor query_1_cast_fp16 = layer_norm(axes = query_1_axes_0, beta = module_layers_0_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_0_norm_self_att_weight_to_fp16, x = input_31_cast_fp16)[name = tensor("query_1_cast_fp16")]; - tensor module_layers_0_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_0_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7105856)))]; - tensor module_layers_0_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_0_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7630208)))]; - tensor linear_3_cast_fp16 = linear(bias = module_layers_0_self_attn_linear_q_bias_to_fp16, weight = module_layers_0_self_attn_linear_q_weight_to_fp16, x = query_1_cast_fp16)[name = tensor("linear_3_cast_fp16")]; - tensor var_370 = const()[name = tensor("op_370"), val = tensor([1, -1, 8, 64])]; - tensor q_1_cast_fp16 = reshape(shape = var_370, x = linear_3_cast_fp16)[name = tensor("q_1_cast_fp16")]; - tensor module_layers_0_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_0_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7631296)))]; - tensor module_layers_0_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_0_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8155648)))]; - tensor linear_4_cast_fp16 = linear(bias = module_layers_0_self_attn_linear_k_bias_to_fp16, weight = module_layers_0_self_attn_linear_k_weight_to_fp16, x = query_1_cast_fp16)[name = tensor("linear_4_cast_fp16")]; - tensor var_375 = const()[name = tensor("op_375"), val = tensor([1, -1, 8, 64])]; - tensor k_1_cast_fp16 = reshape(shape = var_375, x = linear_4_cast_fp16)[name = tensor("k_1_cast_fp16")]; - tensor module_layers_0_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_0_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8156736)))]; - tensor module_layers_0_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_0_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8681088)))]; - tensor linear_5_cast_fp16 = linear(bias = module_layers_0_self_attn_linear_v_bias_to_fp16, weight = module_layers_0_self_attn_linear_v_weight_to_fp16, x = query_1_cast_fp16)[name = tensor("linear_5_cast_fp16")]; - tensor var_380 = const()[name = tensor("op_380"), val = tensor([1, -1, 8, 64])]; - tensor v_1_cast_fp16 = reshape(shape = var_380, x = linear_5_cast_fp16)[name = tensor("v_1_cast_fp16")]; - tensor value_3_perm_0 = const()[name = tensor("value_3_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_0_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_0_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8682176)))]; - tensor var_392_cast_fp16 = add(x = q_1_cast_fp16, y = module_layers_0_self_attn_pos_bias_u_to_fp16)[name = tensor("op_392_cast_fp16")]; - tensor module_layers_0_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_0_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8683264)))]; - tensor var_394_cast_fp16 = add(x = q_1_cast_fp16, y = module_layers_0_self_attn_pos_bias_v_to_fp16)[name = tensor("op_394_cast_fp16")]; - tensor q_with_bias_v_1_perm_0 = const()[name = tensor("q_with_bias_v_1_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_7_transpose_x_0 = const()[name = tensor("x_7_transpose_x_0"), val = tensor(false)]; - tensor x_7_transpose_y_0 = const()[name = tensor("x_7_transpose_y_0"), val = tensor(false)]; - tensor var_396_to_fp16 = const()[name = tensor("op_396_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8684352)))]; - tensor q_with_bias_v_1_cast_fp16 = transpose(perm = q_with_bias_v_1_perm_0, x = var_394_cast_fp16)[name = tensor("transpose_203")]; - tensor x_7_cast_fp16 = matmul(transpose_x = x_7_transpose_x_0, transpose_y = x_7_transpose_y_0, x = q_with_bias_v_1_cast_fp16, y = var_396_to_fp16)[name = tensor("x_7_cast_fp16")]; - tensor x_9_pad_0 = const()[name = tensor("x_9_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_9_mode_0 = const()[name = tensor("x_9_mode_0"), val = tensor("constant")]; - tensor const_70_to_fp16 = const()[name = tensor("const_70_to_fp16"), val = tensor(0x0p+0)]; - tensor x_9_cast_fp16 = pad(constant_val = const_70_to_fp16, mode = x_9_mode_0, pad = x_9_pad_0, x = x_7_cast_fp16)[name = tensor("x_9_cast_fp16")]; - tensor var_404 = const()[name = tensor("op_404"), val = tensor([1, 8, -1, 188])]; - tensor x_11_cast_fp16 = reshape(shape = var_404, x = x_9_cast_fp16)[name = tensor("x_11_cast_fp16")]; - tensor var_408_begin_0 = const()[name = tensor("op_408_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_408_end_0 = const()[name = tensor("op_408_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_408_end_mask_0 = const()[name = tensor("op_408_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_408_cast_fp16 = slice_by_index(begin = var_408_begin_0, end = var_408_end_0, end_mask = var_408_end_mask_0, x = x_11_cast_fp16)[name = tensor("op_408_cast_fp16")]; - tensor var_409 = const()[name = tensor("op_409"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_1_cast_fp16 = reshape(shape = var_409, x = var_408_cast_fp16)[name = tensor("matrix_bd_1_cast_fp16")]; - tensor matrix_ac_1_transpose_x_0 = const()[name = tensor("matrix_ac_1_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_1_transpose_y_0 = const()[name = tensor("matrix_ac_1_transpose_y_0"), val = tensor(false)]; - tensor transpose_51_perm_0 = const()[name = tensor("transpose_51_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_52_perm_0 = const()[name = tensor("transpose_52_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_52 = transpose(perm = transpose_52_perm_0, x = k_1_cast_fp16)[name = tensor("transpose_201")]; - tensor transpose_51 = transpose(perm = transpose_51_perm_0, x = var_392_cast_fp16)[name = tensor("transpose_202")]; - tensor matrix_ac_1_cast_fp16 = matmul(transpose_x = matrix_ac_1_transpose_x_0, transpose_y = matrix_ac_1_transpose_y_0, x = transpose_51, y = transpose_52)[name = tensor("matrix_ac_1_cast_fp16")]; - tensor matrix_bd_3_begin_0 = const()[name = tensor("matrix_bd_3_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_3_end_0 = const()[name = tensor("matrix_bd_3_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_3_end_mask_0 = const()[name = tensor("matrix_bd_3_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_3_cast_fp16 = slice_by_index(begin = matrix_bd_3_begin_0, end = matrix_bd_3_end_0, end_mask = matrix_bd_3_end_mask_0, x = matrix_bd_1_cast_fp16)[name = tensor("matrix_bd_3_cast_fp16")]; - tensor var_418_cast_fp16 = add(x = matrix_ac_1_cast_fp16, y = matrix_bd_3_cast_fp16)[name = tensor("op_418_cast_fp16")]; - tensor _inversed_scores_1_y_0_to_fp16 = const()[name = tensor("_inversed_scores_1_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_1_cast_fp16 = mul(x = var_418_cast_fp16, y = _inversed_scores_1_y_0_to_fp16)[name = tensor("_inversed_scores_1_cast_fp16")]; - tensor mask_11_axes_0 = const()[name = tensor("mask_11_axes_0"), val = tensor([1])]; - tensor mask_11 = expand_dims(axes = mask_11_axes_0, x = mask_9)[name = tensor("mask_11")]; - tensor var_12_to_fp16 = const()[name = tensor("op_12_to_fp16"), val = tensor(-0x1.388p+13)]; - tensor scores_3_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_1_cast_fp16, cond = mask_11)[name = tensor("scores_3_cast_fp16")]; - tensor var_424_cast_fp16 = softmax(axis = var_23, x = scores_3_cast_fp16)[name = tensor("op_424_cast_fp16")]; - tensor var_11_to_fp16 = const()[name = tensor("op_11_to_fp16"), val = tensor(0x0p+0)]; - tensor input_33_cast_fp16 = select(a = var_11_to_fp16, b = var_424_cast_fp16, cond = mask_11)[name = tensor("input_33_cast_fp16")]; - tensor x_13_transpose_x_0 = const()[name = tensor("x_13_transpose_x_0"), val = tensor(false)]; - tensor x_13_transpose_y_0 = const()[name = tensor("x_13_transpose_y_0"), val = tensor(false)]; - tensor value_3_cast_fp16 = transpose(perm = value_3_perm_0, x = v_1_cast_fp16)[name = tensor("transpose_204")]; - tensor x_13_cast_fp16 = matmul(transpose_x = x_13_transpose_x_0, transpose_y = x_13_transpose_y_0, x = input_33_cast_fp16, y = value_3_cast_fp16)[name = tensor("x_13_cast_fp16")]; - tensor var_428_perm_0 = const()[name = tensor("op_428_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_429 = const()[name = tensor("op_429"), val = tensor([1, -1, 512])]; - tensor var_428_cast_fp16 = transpose(perm = var_428_perm_0, x = x_13_cast_fp16)[name = tensor("transpose_200")]; - tensor input_35_cast_fp16 = reshape(shape = var_429, x = var_428_cast_fp16)[name = tensor("input_35_cast_fp16")]; - tensor module_layers_0_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_0_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9068416)))]; - tensor module_layers_0_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_0_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9592768)))]; - tensor linear_7_cast_fp16 = linear(bias = module_layers_0_self_attn_linear_out_bias_to_fp16, weight = module_layers_0_self_attn_linear_out_weight_to_fp16, x = input_35_cast_fp16)[name = tensor("linear_7_cast_fp16")]; - tensor input_39_cast_fp16 = add(x = input_31_cast_fp16, y = linear_7_cast_fp16)[name = tensor("input_39_cast_fp16")]; - tensor x_17_axes_0 = const()[name = tensor("x_17_axes_0"), val = tensor([-1])]; - tensor module_layers_0_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_0_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9593856)))]; - tensor module_layers_0_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_0_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9594944)))]; - tensor x_17_cast_fp16 = layer_norm(axes = x_17_axes_0, beta = module_layers_0_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_0_norm_conv_weight_to_fp16, x = input_39_cast_fp16)[name = tensor("x_17_cast_fp16")]; - tensor input_41_perm_0 = const()[name = tensor("input_41_perm_0"), val = tensor([0, 2, 1])]; - tensor input_43_pad_type_0 = const()[name = tensor("input_43_pad_type_0"), val = tensor("valid")]; - tensor input_43_strides_0 = const()[name = tensor("input_43_strides_0"), val = tensor([1])]; - tensor input_43_pad_0 = const()[name = tensor("input_43_pad_0"), val = tensor([0, 0])]; - tensor input_43_dilations_0 = const()[name = tensor("input_43_dilations_0"), val = tensor([1])]; - tensor input_43_groups_0 = const()[name = tensor("input_43_groups_0"), val = tensor(1)]; - tensor module_layers_0_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_0_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9596032)))]; - tensor module_layers_0_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_0_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10644672)))]; - tensor input_41_cast_fp16 = transpose(perm = input_41_perm_0, x = x_17_cast_fp16)[name = tensor("transpose_199")]; - tensor input_43_cast_fp16 = conv(bias = module_layers_0_conv_pointwise_conv1_bias_to_fp16, dilations = input_43_dilations_0, groups = input_43_groups_0, pad = input_43_pad_0, pad_type = input_43_pad_type_0, strides = input_43_strides_0, weight = module_layers_0_conv_pointwise_conv1_weight_to_fp16, x = input_41_cast_fp16)[name = tensor("input_43_cast_fp16")]; - tensor x_19_split_num_splits_0 = const()[name = tensor("x_19_split_num_splits_0"), val = tensor(2)]; - tensor x_19_split_axis_0 = const()[name = tensor("x_19_split_axis_0"), val = tensor(1)]; - tensor x_19_split_cast_fp16_0, tensor x_19_split_cast_fp16_1 = split(axis = x_19_split_axis_0, num_splits = x_19_split_num_splits_0, x = input_43_cast_fp16)[name = tensor("x_19_split_cast_fp16")]; - tensor x_19_split_1_sigmoid_cast_fp16 = sigmoid(x = x_19_split_cast_fp16_1)[name = tensor("x_19_split_1_sigmoid_cast_fp16")]; - tensor x_19_cast_fp16 = mul(x = x_19_split_cast_fp16_0, y = x_19_split_1_sigmoid_cast_fp16)[name = tensor("x_19_cast_fp16")]; - tensor var_453_axes_0 = const()[name = tensor("op_453_axes_0"), val = tensor([1])]; - tensor var_453 = expand_dims(axes = var_453_axes_0, x = pad_mask)[name = tensor("op_453")]; - tensor input_45_cast_fp16 = select(a = var_11_to_fp16, b = x_19_cast_fp16, cond = var_453)[name = tensor("input_45_cast_fp16")]; - tensor input_47_pad_0 = const()[name = tensor("input_47_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_47_mode_0 = const()[name = tensor("input_47_mode_0"), val = tensor("constant")]; - tensor const_73_to_fp16 = const()[name = tensor("const_73_to_fp16"), val = tensor(0x0p+0)]; - tensor input_47_cast_fp16 = pad(constant_val = const_73_to_fp16, mode = input_47_mode_0, pad = input_47_pad_0, x = input_45_cast_fp16)[name = tensor("input_47_cast_fp16")]; - tensor input_49_pad_type_0 = const()[name = tensor("input_49_pad_type_0"), val = tensor("valid")]; - tensor input_49_groups_0 = const()[name = tensor("input_49_groups_0"), val = tensor(512)]; - tensor input_49_strides_0 = const()[name = tensor("input_49_strides_0"), val = tensor([1])]; - tensor input_49_pad_0 = const()[name = tensor("input_49_pad_0"), val = tensor([0, 0])]; - tensor input_49_dilations_0 = const()[name = tensor("input_49_dilations_0"), val = tensor([1])]; - tensor const_234_to_fp16 = const()[name = tensor("const_234_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10646784)))]; - tensor const_235_to_fp16 = const()[name = tensor("const_235_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10656064)))]; - tensor input_51_cast_fp16 = conv(bias = const_235_to_fp16, dilations = input_49_dilations_0, groups = input_49_groups_0, pad = input_49_pad_0, pad_type = input_49_pad_type_0, strides = input_49_strides_0, weight = const_234_to_fp16, x = input_47_cast_fp16)[name = tensor("input_51_cast_fp16")]; - tensor input_53_cast_fp16 = silu(x = input_51_cast_fp16)[name = tensor("input_53_cast_fp16")]; - tensor x_21_pad_type_0 = const()[name = tensor("x_21_pad_type_0"), val = tensor("valid")]; - tensor x_21_strides_0 = const()[name = tensor("x_21_strides_0"), val = tensor([1])]; - tensor x_21_pad_0 = const()[name = tensor("x_21_pad_0"), val = tensor([0, 0])]; - tensor x_21_dilations_0 = const()[name = tensor("x_21_dilations_0"), val = tensor([1])]; - tensor x_21_groups_0 = const()[name = tensor("x_21_groups_0"), val = tensor(1)]; - tensor module_layers_0_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_0_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10657152)))]; - tensor module_layers_0_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_0_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11181504)))]; - tensor x_21_cast_fp16 = conv(bias = module_layers_0_conv_pointwise_conv2_bias_to_fp16, dilations = x_21_dilations_0, groups = x_21_groups_0, pad = x_21_pad_0, pad_type = x_21_pad_type_0, strides = x_21_strides_0, weight = module_layers_0_conv_pointwise_conv2_weight_to_fp16, x = input_53_cast_fp16)[name = tensor("x_21_cast_fp16")]; - tensor input_55_perm_0 = const()[name = tensor("input_55_perm_0"), val = tensor([0, 2, 1])]; - tensor input_55_cast_fp16 = transpose(perm = input_55_perm_0, x = x_21_cast_fp16)[name = tensor("transpose_198")]; - tensor input_57_cast_fp16 = add(x = input_39_cast_fp16, y = input_55_cast_fp16)[name = tensor("input_57_cast_fp16")]; - tensor input_59_axes_0 = const()[name = tensor("input_59_axes_0"), val = tensor([-1])]; - tensor module_layers_0_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_0_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11182592)))]; - tensor module_layers_0_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_0_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11183680)))]; - tensor input_59_cast_fp16 = layer_norm(axes = input_59_axes_0, beta = module_layers_0_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_0_norm_feed_forward2_weight_to_fp16, x = input_57_cast_fp16)[name = tensor("input_59_cast_fp16")]; - tensor module_layers_0_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_0_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11184768)))]; - tensor module_layers_0_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_0_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(13281984)))]; - tensor linear_8_cast_fp16 = linear(bias = module_layers_0_feed_forward2_linear1_bias_to_fp16, weight = module_layers_0_feed_forward2_linear1_weight_to_fp16, x = input_59_cast_fp16)[name = tensor("linear_8_cast_fp16")]; - tensor input_63_cast_fp16 = silu(x = linear_8_cast_fp16)[name = tensor("input_63_cast_fp16")]; - tensor module_layers_0_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_0_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(13286144)))]; - tensor module_layers_0_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_0_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15383360)))]; - tensor linear_9_cast_fp16 = linear(bias = module_layers_0_feed_forward2_linear2_bias_to_fp16, weight = module_layers_0_feed_forward2_linear2_weight_to_fp16, x = input_63_cast_fp16)[name = tensor("linear_9_cast_fp16")]; - tensor var_495_to_fp16 = const()[name = tensor("op_495_to_fp16"), val = tensor(0x1p-1)]; - tensor var_496_cast_fp16 = mul(x = linear_9_cast_fp16, y = var_495_to_fp16)[name = tensor("op_496_cast_fp16")]; - tensor input_69_cast_fp16 = add(x = input_57_cast_fp16, y = var_496_cast_fp16)[name = tensor("input_69_cast_fp16")]; - tensor input_71_axes_0 = const()[name = tensor("input_71_axes_0"), val = tensor([-1])]; - tensor module_layers_0_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_0_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15384448)))]; - tensor module_layers_0_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_0_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15385536)))]; - tensor input_71_cast_fp16 = layer_norm(axes = input_71_axes_0, beta = module_layers_0_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_0_norm_out_weight_to_fp16, x = input_69_cast_fp16)[name = tensor("input_71_cast_fp16")]; - tensor input_73_axes_0 = const()[name = tensor("input_73_axes_0"), val = tensor([-1])]; - tensor module_layers_1_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_1_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15386624)))]; - tensor module_layers_1_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_1_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15387712)))]; - tensor input_73_cast_fp16 = layer_norm(axes = input_73_axes_0, beta = module_layers_1_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_1_norm_feed_forward1_weight_to_fp16, x = input_71_cast_fp16)[name = tensor("input_73_cast_fp16")]; - tensor module_layers_1_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_1_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15388800)))]; - tensor module_layers_1_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_1_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(17486016)))]; - tensor linear_10_cast_fp16 = linear(bias = module_layers_1_feed_forward1_linear1_bias_to_fp16, weight = module_layers_1_feed_forward1_linear1_weight_to_fp16, x = input_73_cast_fp16)[name = tensor("linear_10_cast_fp16")]; - tensor input_77_cast_fp16 = silu(x = linear_10_cast_fp16)[name = tensor("input_77_cast_fp16")]; - tensor module_layers_1_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_1_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(17490176)))]; - tensor module_layers_1_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_1_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19587392)))]; - tensor linear_11_cast_fp16 = linear(bias = module_layers_1_feed_forward1_linear2_bias_to_fp16, weight = module_layers_1_feed_forward1_linear2_weight_to_fp16, x = input_77_cast_fp16)[name = tensor("linear_11_cast_fp16")]; - tensor var_526_to_fp16 = const()[name = tensor("op_526_to_fp16"), val = tensor(0x1p-1)]; - tensor var_527_cast_fp16 = mul(x = linear_11_cast_fp16, y = var_526_to_fp16)[name = tensor("op_527_cast_fp16")]; - tensor input_83_cast_fp16 = add(x = input_71_cast_fp16, y = var_527_cast_fp16)[name = tensor("input_83_cast_fp16")]; - tensor query_3_axes_0 = const()[name = tensor("query_3_axes_0"), val = tensor([-1])]; - tensor module_layers_1_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_1_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19588480)))]; - tensor module_layers_1_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_1_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19589568)))]; - tensor query_3_cast_fp16 = layer_norm(axes = query_3_axes_0, beta = module_layers_1_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_1_norm_self_att_weight_to_fp16, x = input_83_cast_fp16)[name = tensor("query_3_cast_fp16")]; - tensor module_layers_1_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_1_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19590656)))]; - tensor module_layers_1_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_1_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(20115008)))]; - tensor linear_12_cast_fp16 = linear(bias = module_layers_1_self_attn_linear_q_bias_to_fp16, weight = module_layers_1_self_attn_linear_q_weight_to_fp16, x = query_3_cast_fp16)[name = tensor("linear_12_cast_fp16")]; - tensor var_544 = const()[name = tensor("op_544"), val = tensor([1, -1, 8, 64])]; - tensor q_7_cast_fp16 = reshape(shape = var_544, x = linear_12_cast_fp16)[name = tensor("q_7_cast_fp16")]; - tensor module_layers_1_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_1_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(20116096)))]; - tensor module_layers_1_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_1_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(20640448)))]; - tensor linear_13_cast_fp16 = linear(bias = module_layers_1_self_attn_linear_k_bias_to_fp16, weight = module_layers_1_self_attn_linear_k_weight_to_fp16, x = query_3_cast_fp16)[name = tensor("linear_13_cast_fp16")]; - tensor var_549 = const()[name = tensor("op_549"), val = tensor([1, -1, 8, 64])]; - tensor k_5_cast_fp16 = reshape(shape = var_549, x = linear_13_cast_fp16)[name = tensor("k_5_cast_fp16")]; - tensor module_layers_1_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_1_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(20641536)))]; - tensor module_layers_1_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_1_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21165888)))]; - tensor linear_14_cast_fp16 = linear(bias = module_layers_1_self_attn_linear_v_bias_to_fp16, weight = module_layers_1_self_attn_linear_v_weight_to_fp16, x = query_3_cast_fp16)[name = tensor("linear_14_cast_fp16")]; - tensor var_554 = const()[name = tensor("op_554"), val = tensor([1, -1, 8, 64])]; - tensor v_3_cast_fp16 = reshape(shape = var_554, x = linear_14_cast_fp16)[name = tensor("v_3_cast_fp16")]; - tensor value_5_perm_0 = const()[name = tensor("value_5_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_1_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_1_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21166976)))]; - tensor var_566_cast_fp16 = add(x = q_7_cast_fp16, y = module_layers_1_self_attn_pos_bias_u_to_fp16)[name = tensor("op_566_cast_fp16")]; - tensor module_layers_1_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_1_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21168064)))]; - tensor var_568_cast_fp16 = add(x = q_7_cast_fp16, y = module_layers_1_self_attn_pos_bias_v_to_fp16)[name = tensor("op_568_cast_fp16")]; - tensor q_with_bias_v_3_perm_0 = const()[name = tensor("q_with_bias_v_3_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_29_transpose_x_0 = const()[name = tensor("x_29_transpose_x_0"), val = tensor(false)]; - tensor x_29_transpose_y_0 = const()[name = tensor("x_29_transpose_y_0"), val = tensor(false)]; - tensor var_570_to_fp16 = const()[name = tensor("op_570_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21169152)))]; - tensor q_with_bias_v_3_cast_fp16 = transpose(perm = q_with_bias_v_3_perm_0, x = var_568_cast_fp16)[name = tensor("transpose_196")]; - tensor x_29_cast_fp16 = matmul(transpose_x = x_29_transpose_x_0, transpose_y = x_29_transpose_y_0, x = q_with_bias_v_3_cast_fp16, y = var_570_to_fp16)[name = tensor("x_29_cast_fp16")]; - tensor x_31_pad_0 = const()[name = tensor("x_31_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_31_mode_0 = const()[name = tensor("x_31_mode_0"), val = tensor("constant")]; - tensor const_80_to_fp16 = const()[name = tensor("const_80_to_fp16"), val = tensor(0x0p+0)]; - tensor x_31_cast_fp16 = pad(constant_val = const_80_to_fp16, mode = x_31_mode_0, pad = x_31_pad_0, x = x_29_cast_fp16)[name = tensor("x_31_cast_fp16")]; - tensor var_578 = const()[name = tensor("op_578"), val = tensor([1, 8, -1, 188])]; - tensor x_33_cast_fp16 = reshape(shape = var_578, x = x_31_cast_fp16)[name = tensor("x_33_cast_fp16")]; - tensor var_582_begin_0 = const()[name = tensor("op_582_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_582_end_0 = const()[name = tensor("op_582_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_582_end_mask_0 = const()[name = tensor("op_582_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_582_cast_fp16 = slice_by_index(begin = var_582_begin_0, end = var_582_end_0, end_mask = var_582_end_mask_0, x = x_33_cast_fp16)[name = tensor("op_582_cast_fp16")]; - tensor var_583 = const()[name = tensor("op_583"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_5_cast_fp16 = reshape(shape = var_583, x = var_582_cast_fp16)[name = tensor("matrix_bd_5_cast_fp16")]; - tensor matrix_ac_3_transpose_x_0 = const()[name = tensor("matrix_ac_3_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_3_transpose_y_0 = const()[name = tensor("matrix_ac_3_transpose_y_0"), val = tensor(false)]; - tensor transpose_53_perm_0 = const()[name = tensor("transpose_53_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_54_perm_0 = const()[name = tensor("transpose_54_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_54 = transpose(perm = transpose_54_perm_0, x = k_5_cast_fp16)[name = tensor("transpose_194")]; - tensor transpose_53 = transpose(perm = transpose_53_perm_0, x = var_566_cast_fp16)[name = tensor("transpose_195")]; - tensor matrix_ac_3_cast_fp16 = matmul(transpose_x = matrix_ac_3_transpose_x_0, transpose_y = matrix_ac_3_transpose_y_0, x = transpose_53, y = transpose_54)[name = tensor("matrix_ac_3_cast_fp16")]; - tensor matrix_bd_7_begin_0 = const()[name = tensor("matrix_bd_7_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_7_end_0 = const()[name = tensor("matrix_bd_7_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_7_end_mask_0 = const()[name = tensor("matrix_bd_7_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_7_cast_fp16 = slice_by_index(begin = matrix_bd_7_begin_0, end = matrix_bd_7_end_0, end_mask = matrix_bd_7_end_mask_0, x = matrix_bd_5_cast_fp16)[name = tensor("matrix_bd_7_cast_fp16")]; - tensor var_592_cast_fp16 = add(x = matrix_ac_3_cast_fp16, y = matrix_bd_7_cast_fp16)[name = tensor("op_592_cast_fp16")]; - tensor _inversed_scores_5_y_0_to_fp16 = const()[name = tensor("_inversed_scores_5_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_5_cast_fp16 = mul(x = var_592_cast_fp16, y = _inversed_scores_5_y_0_to_fp16)[name = tensor("_inversed_scores_5_cast_fp16")]; - tensor scores_7_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_5_cast_fp16, cond = mask_11)[name = tensor("scores_7_cast_fp16")]; - tensor var_598_cast_fp16 = softmax(axis = var_23, x = scores_7_cast_fp16)[name = tensor("op_598_cast_fp16")]; - tensor input_85_cast_fp16 = select(a = var_11_to_fp16, b = var_598_cast_fp16, cond = mask_11)[name = tensor("input_85_cast_fp16")]; - tensor x_35_transpose_x_0 = const()[name = tensor("x_35_transpose_x_0"), val = tensor(false)]; - tensor x_35_transpose_y_0 = const()[name = tensor("x_35_transpose_y_0"), val = tensor(false)]; - tensor value_5_cast_fp16 = transpose(perm = value_5_perm_0, x = v_3_cast_fp16)[name = tensor("transpose_197")]; - tensor x_35_cast_fp16 = matmul(transpose_x = x_35_transpose_x_0, transpose_y = x_35_transpose_y_0, x = input_85_cast_fp16, y = value_5_cast_fp16)[name = tensor("x_35_cast_fp16")]; - tensor var_602_perm_0 = const()[name = tensor("op_602_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_603 = const()[name = tensor("op_603"), val = tensor([1, -1, 512])]; - tensor var_602_cast_fp16 = transpose(perm = var_602_perm_0, x = x_35_cast_fp16)[name = tensor("transpose_193")]; - tensor input_87_cast_fp16 = reshape(shape = var_603, x = var_602_cast_fp16)[name = tensor("input_87_cast_fp16")]; - tensor module_layers_1_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_1_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21553216)))]; - tensor module_layers_1_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_1_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22077568)))]; - tensor linear_16_cast_fp16 = linear(bias = module_layers_1_self_attn_linear_out_bias_to_fp16, weight = module_layers_1_self_attn_linear_out_weight_to_fp16, x = input_87_cast_fp16)[name = tensor("linear_16_cast_fp16")]; - tensor input_91_cast_fp16 = add(x = input_83_cast_fp16, y = linear_16_cast_fp16)[name = tensor("input_91_cast_fp16")]; - tensor x_39_axes_0 = const()[name = tensor("x_39_axes_0"), val = tensor([-1])]; - tensor module_layers_1_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_1_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22078656)))]; - tensor module_layers_1_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_1_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22079744)))]; - tensor x_39_cast_fp16 = layer_norm(axes = x_39_axes_0, beta = module_layers_1_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_1_norm_conv_weight_to_fp16, x = input_91_cast_fp16)[name = tensor("x_39_cast_fp16")]; - tensor input_93_perm_0 = const()[name = tensor("input_93_perm_0"), val = tensor([0, 2, 1])]; - tensor input_95_pad_type_0 = const()[name = tensor("input_95_pad_type_0"), val = tensor("valid")]; - tensor input_95_strides_0 = const()[name = tensor("input_95_strides_0"), val = tensor([1])]; - tensor input_95_pad_0 = const()[name = tensor("input_95_pad_0"), val = tensor([0, 0])]; - tensor input_95_dilations_0 = const()[name = tensor("input_95_dilations_0"), val = tensor([1])]; - tensor input_95_groups_0 = const()[name = tensor("input_95_groups_0"), val = tensor(1)]; - tensor module_layers_1_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_1_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22080832)))]; - tensor module_layers_1_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_1_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23129472)))]; - tensor input_93_cast_fp16 = transpose(perm = input_93_perm_0, x = x_39_cast_fp16)[name = tensor("transpose_192")]; - tensor input_95_cast_fp16 = conv(bias = module_layers_1_conv_pointwise_conv1_bias_to_fp16, dilations = input_95_dilations_0, groups = input_95_groups_0, pad = input_95_pad_0, pad_type = input_95_pad_type_0, strides = input_95_strides_0, weight = module_layers_1_conv_pointwise_conv1_weight_to_fp16, x = input_93_cast_fp16)[name = tensor("input_95_cast_fp16")]; - tensor x_41_split_num_splits_0 = const()[name = tensor("x_41_split_num_splits_0"), val = tensor(2)]; - tensor x_41_split_axis_0 = const()[name = tensor("x_41_split_axis_0"), val = tensor(1)]; - tensor x_41_split_cast_fp16_0, tensor x_41_split_cast_fp16_1 = split(axis = x_41_split_axis_0, num_splits = x_41_split_num_splits_0, x = input_95_cast_fp16)[name = tensor("x_41_split_cast_fp16")]; - tensor x_41_split_1_sigmoid_cast_fp16 = sigmoid(x = x_41_split_cast_fp16_1)[name = tensor("x_41_split_1_sigmoid_cast_fp16")]; - tensor x_41_cast_fp16 = mul(x = x_41_split_cast_fp16_0, y = x_41_split_1_sigmoid_cast_fp16)[name = tensor("x_41_cast_fp16")]; - tensor input_97_cast_fp16 = select(a = var_11_to_fp16, b = x_41_cast_fp16, cond = var_453)[name = tensor("input_97_cast_fp16")]; - tensor input_99_pad_0 = const()[name = tensor("input_99_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_99_mode_0 = const()[name = tensor("input_99_mode_0"), val = tensor("constant")]; - tensor const_83_to_fp16 = const()[name = tensor("const_83_to_fp16"), val = tensor(0x0p+0)]; - tensor input_99_cast_fp16 = pad(constant_val = const_83_to_fp16, mode = input_99_mode_0, pad = input_99_pad_0, x = input_97_cast_fp16)[name = tensor("input_99_cast_fp16")]; - tensor input_101_pad_type_0 = const()[name = tensor("input_101_pad_type_0"), val = tensor("valid")]; - tensor input_101_groups_0 = const()[name = tensor("input_101_groups_0"), val = tensor(512)]; - tensor input_101_strides_0 = const()[name = tensor("input_101_strides_0"), val = tensor([1])]; - tensor input_101_pad_0 = const()[name = tensor("input_101_pad_0"), val = tensor([0, 0])]; - tensor input_101_dilations_0 = const()[name = tensor("input_101_dilations_0"), val = tensor([1])]; - tensor const_236_to_fp16 = const()[name = tensor("const_236_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23131584)))]; - tensor const_237_to_fp16 = const()[name = tensor("const_237_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23140864)))]; - tensor input_103_cast_fp16 = conv(bias = const_237_to_fp16, dilations = input_101_dilations_0, groups = input_101_groups_0, pad = input_101_pad_0, pad_type = input_101_pad_type_0, strides = input_101_strides_0, weight = const_236_to_fp16, x = input_99_cast_fp16)[name = tensor("input_103_cast_fp16")]; - tensor input_105_cast_fp16 = silu(x = input_103_cast_fp16)[name = tensor("input_105_cast_fp16")]; - tensor x_43_pad_type_0 = const()[name = tensor("x_43_pad_type_0"), val = tensor("valid")]; - tensor x_43_strides_0 = const()[name = tensor("x_43_strides_0"), val = tensor([1])]; - tensor x_43_pad_0 = const()[name = tensor("x_43_pad_0"), val = tensor([0, 0])]; - tensor x_43_dilations_0 = const()[name = tensor("x_43_dilations_0"), val = tensor([1])]; - tensor x_43_groups_0 = const()[name = tensor("x_43_groups_0"), val = tensor(1)]; - tensor module_layers_1_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_1_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23141952)))]; - tensor module_layers_1_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_1_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23666304)))]; - tensor x_43_cast_fp16 = conv(bias = module_layers_1_conv_pointwise_conv2_bias_to_fp16, dilations = x_43_dilations_0, groups = x_43_groups_0, pad = x_43_pad_0, pad_type = x_43_pad_type_0, strides = x_43_strides_0, weight = module_layers_1_conv_pointwise_conv2_weight_to_fp16, x = input_105_cast_fp16)[name = tensor("x_43_cast_fp16")]; - tensor input_107_perm_0 = const()[name = tensor("input_107_perm_0"), val = tensor([0, 2, 1])]; - tensor input_107_cast_fp16 = transpose(perm = input_107_perm_0, x = x_43_cast_fp16)[name = tensor("transpose_191")]; - tensor input_109_cast_fp16 = add(x = input_91_cast_fp16, y = input_107_cast_fp16)[name = tensor("input_109_cast_fp16")]; - tensor input_111_axes_0 = const()[name = tensor("input_111_axes_0"), val = tensor([-1])]; - tensor module_layers_1_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_1_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23667392)))]; - tensor module_layers_1_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_1_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23668480)))]; - tensor input_111_cast_fp16 = layer_norm(axes = input_111_axes_0, beta = module_layers_1_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_1_norm_feed_forward2_weight_to_fp16, x = input_109_cast_fp16)[name = tensor("input_111_cast_fp16")]; - tensor module_layers_1_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_1_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23669568)))]; - tensor module_layers_1_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_1_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(25766784)))]; - tensor linear_17_cast_fp16 = linear(bias = module_layers_1_feed_forward2_linear1_bias_to_fp16, weight = module_layers_1_feed_forward2_linear1_weight_to_fp16, x = input_111_cast_fp16)[name = tensor("linear_17_cast_fp16")]; - tensor input_115_cast_fp16 = silu(x = linear_17_cast_fp16)[name = tensor("input_115_cast_fp16")]; - tensor module_layers_1_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_1_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(25770944)))]; - tensor module_layers_1_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_1_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27868160)))]; - tensor linear_18_cast_fp16 = linear(bias = module_layers_1_feed_forward2_linear2_bias_to_fp16, weight = module_layers_1_feed_forward2_linear2_weight_to_fp16, x = input_115_cast_fp16)[name = tensor("linear_18_cast_fp16")]; - tensor var_669_to_fp16 = const()[name = tensor("op_669_to_fp16"), val = tensor(0x1p-1)]; - tensor var_670_cast_fp16 = mul(x = linear_18_cast_fp16, y = var_669_to_fp16)[name = tensor("op_670_cast_fp16")]; - tensor input_121_cast_fp16 = add(x = input_109_cast_fp16, y = var_670_cast_fp16)[name = tensor("input_121_cast_fp16")]; - tensor input_123_axes_0 = const()[name = tensor("input_123_axes_0"), val = tensor([-1])]; - tensor module_layers_1_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_1_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27869248)))]; - tensor module_layers_1_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_1_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27870336)))]; - tensor input_123_cast_fp16 = layer_norm(axes = input_123_axes_0, beta = module_layers_1_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_1_norm_out_weight_to_fp16, x = input_121_cast_fp16)[name = tensor("input_123_cast_fp16")]; - tensor input_125_axes_0 = const()[name = tensor("input_125_axes_0"), val = tensor([-1])]; - tensor module_layers_2_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_2_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27871424)))]; - tensor module_layers_2_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_2_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27872512)))]; - tensor input_125_cast_fp16 = layer_norm(axes = input_125_axes_0, beta = module_layers_2_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_2_norm_feed_forward1_weight_to_fp16, x = input_123_cast_fp16)[name = tensor("input_125_cast_fp16")]; - tensor module_layers_2_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_2_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27873600)))]; - tensor module_layers_2_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_2_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(29970816)))]; - tensor linear_19_cast_fp16 = linear(bias = module_layers_2_feed_forward1_linear1_bias_to_fp16, weight = module_layers_2_feed_forward1_linear1_weight_to_fp16, x = input_125_cast_fp16)[name = tensor("linear_19_cast_fp16")]; - tensor input_129_cast_fp16 = silu(x = linear_19_cast_fp16)[name = tensor("input_129_cast_fp16")]; - tensor module_layers_2_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_2_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(29974976)))]; - tensor module_layers_2_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_2_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32072192)))]; - tensor linear_20_cast_fp16 = linear(bias = module_layers_2_feed_forward1_linear2_bias_to_fp16, weight = module_layers_2_feed_forward1_linear2_weight_to_fp16, x = input_129_cast_fp16)[name = tensor("linear_20_cast_fp16")]; - tensor var_700_to_fp16 = const()[name = tensor("op_700_to_fp16"), val = tensor(0x1p-1)]; - tensor var_701_cast_fp16 = mul(x = linear_20_cast_fp16, y = var_700_to_fp16)[name = tensor("op_701_cast_fp16")]; - tensor input_135_cast_fp16 = add(x = input_123_cast_fp16, y = var_701_cast_fp16)[name = tensor("input_135_cast_fp16")]; - tensor query_5_axes_0 = const()[name = tensor("query_5_axes_0"), val = tensor([-1])]; - tensor module_layers_2_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_2_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32073280)))]; - tensor module_layers_2_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_2_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32074368)))]; - tensor query_5_cast_fp16 = layer_norm(axes = query_5_axes_0, beta = module_layers_2_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_2_norm_self_att_weight_to_fp16, x = input_135_cast_fp16)[name = tensor("query_5_cast_fp16")]; - tensor module_layers_2_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_2_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32075456)))]; - tensor module_layers_2_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_2_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32599808)))]; - tensor linear_21_cast_fp16 = linear(bias = module_layers_2_self_attn_linear_q_bias_to_fp16, weight = module_layers_2_self_attn_linear_q_weight_to_fp16, x = query_5_cast_fp16)[name = tensor("linear_21_cast_fp16")]; - tensor var_718 = const()[name = tensor("op_718"), val = tensor([1, -1, 8, 64])]; - tensor q_13_cast_fp16 = reshape(shape = var_718, x = linear_21_cast_fp16)[name = tensor("q_13_cast_fp16")]; - tensor module_layers_2_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_2_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32600896)))]; - tensor module_layers_2_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_2_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33125248)))]; - tensor linear_22_cast_fp16 = linear(bias = module_layers_2_self_attn_linear_k_bias_to_fp16, weight = module_layers_2_self_attn_linear_k_weight_to_fp16, x = query_5_cast_fp16)[name = tensor("linear_22_cast_fp16")]; - tensor var_723 = const()[name = tensor("op_723"), val = tensor([1, -1, 8, 64])]; - tensor k_9_cast_fp16 = reshape(shape = var_723, x = linear_22_cast_fp16)[name = tensor("k_9_cast_fp16")]; - tensor module_layers_2_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_2_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33126336)))]; - tensor module_layers_2_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_2_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33650688)))]; - tensor linear_23_cast_fp16 = linear(bias = module_layers_2_self_attn_linear_v_bias_to_fp16, weight = module_layers_2_self_attn_linear_v_weight_to_fp16, x = query_5_cast_fp16)[name = tensor("linear_23_cast_fp16")]; - tensor var_728 = const()[name = tensor("op_728"), val = tensor([1, -1, 8, 64])]; - tensor v_5_cast_fp16 = reshape(shape = var_728, x = linear_23_cast_fp16)[name = tensor("v_5_cast_fp16")]; - tensor value_7_perm_0 = const()[name = tensor("value_7_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_2_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_2_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33651776)))]; - tensor var_740_cast_fp16 = add(x = q_13_cast_fp16, y = module_layers_2_self_attn_pos_bias_u_to_fp16)[name = tensor("op_740_cast_fp16")]; - tensor module_layers_2_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_2_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33652864)))]; - tensor var_742_cast_fp16 = add(x = q_13_cast_fp16, y = module_layers_2_self_attn_pos_bias_v_to_fp16)[name = tensor("op_742_cast_fp16")]; - tensor q_with_bias_v_5_perm_0 = const()[name = tensor("q_with_bias_v_5_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_51_transpose_x_0 = const()[name = tensor("x_51_transpose_x_0"), val = tensor(false)]; - tensor x_51_transpose_y_0 = const()[name = tensor("x_51_transpose_y_0"), val = tensor(false)]; - tensor var_744_to_fp16 = const()[name = tensor("op_744_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33653952)))]; - tensor q_with_bias_v_5_cast_fp16 = transpose(perm = q_with_bias_v_5_perm_0, x = var_742_cast_fp16)[name = tensor("transpose_189")]; - tensor x_51_cast_fp16 = matmul(transpose_x = x_51_transpose_x_0, transpose_y = x_51_transpose_y_0, x = q_with_bias_v_5_cast_fp16, y = var_744_to_fp16)[name = tensor("x_51_cast_fp16")]; - tensor x_53_pad_0 = const()[name = tensor("x_53_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_53_mode_0 = const()[name = tensor("x_53_mode_0"), val = tensor("constant")]; - tensor const_90_to_fp16 = const()[name = tensor("const_90_to_fp16"), val = tensor(0x0p+0)]; - tensor x_53_cast_fp16 = pad(constant_val = const_90_to_fp16, mode = x_53_mode_0, pad = x_53_pad_0, x = x_51_cast_fp16)[name = tensor("x_53_cast_fp16")]; - tensor var_752 = const()[name = tensor("op_752"), val = tensor([1, 8, -1, 188])]; - tensor x_55_cast_fp16 = reshape(shape = var_752, x = x_53_cast_fp16)[name = tensor("x_55_cast_fp16")]; - tensor var_756_begin_0 = const()[name = tensor("op_756_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_756_end_0 = const()[name = tensor("op_756_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_756_end_mask_0 = const()[name = tensor("op_756_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_756_cast_fp16 = slice_by_index(begin = var_756_begin_0, end = var_756_end_0, end_mask = var_756_end_mask_0, x = x_55_cast_fp16)[name = tensor("op_756_cast_fp16")]; - tensor var_757 = const()[name = tensor("op_757"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_9_cast_fp16 = reshape(shape = var_757, x = var_756_cast_fp16)[name = tensor("matrix_bd_9_cast_fp16")]; - tensor matrix_ac_5_transpose_x_0 = const()[name = tensor("matrix_ac_5_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_5_transpose_y_0 = const()[name = tensor("matrix_ac_5_transpose_y_0"), val = tensor(false)]; - tensor transpose_55_perm_0 = const()[name = tensor("transpose_55_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_56_perm_0 = const()[name = tensor("transpose_56_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_56 = transpose(perm = transpose_56_perm_0, x = k_9_cast_fp16)[name = tensor("transpose_187")]; - tensor transpose_55 = transpose(perm = transpose_55_perm_0, x = var_740_cast_fp16)[name = tensor("transpose_188")]; - tensor matrix_ac_5_cast_fp16 = matmul(transpose_x = matrix_ac_5_transpose_x_0, transpose_y = matrix_ac_5_transpose_y_0, x = transpose_55, y = transpose_56)[name = tensor("matrix_ac_5_cast_fp16")]; - tensor matrix_bd_11_begin_0 = const()[name = tensor("matrix_bd_11_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_11_end_0 = const()[name = tensor("matrix_bd_11_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_11_end_mask_0 = const()[name = tensor("matrix_bd_11_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_11_cast_fp16 = slice_by_index(begin = matrix_bd_11_begin_0, end = matrix_bd_11_end_0, end_mask = matrix_bd_11_end_mask_0, x = matrix_bd_9_cast_fp16)[name = tensor("matrix_bd_11_cast_fp16")]; - tensor var_766_cast_fp16 = add(x = matrix_ac_5_cast_fp16, y = matrix_bd_11_cast_fp16)[name = tensor("op_766_cast_fp16")]; - tensor _inversed_scores_9_y_0_to_fp16 = const()[name = tensor("_inversed_scores_9_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_9_cast_fp16 = mul(x = var_766_cast_fp16, y = _inversed_scores_9_y_0_to_fp16)[name = tensor("_inversed_scores_9_cast_fp16")]; - tensor scores_11_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_9_cast_fp16, cond = mask_11)[name = tensor("scores_11_cast_fp16")]; - tensor var_772_cast_fp16 = softmax(axis = var_23, x = scores_11_cast_fp16)[name = tensor("op_772_cast_fp16")]; - tensor input_137_cast_fp16 = select(a = var_11_to_fp16, b = var_772_cast_fp16, cond = mask_11)[name = tensor("input_137_cast_fp16")]; - tensor x_57_transpose_x_0 = const()[name = tensor("x_57_transpose_x_0"), val = tensor(false)]; - tensor x_57_transpose_y_0 = const()[name = tensor("x_57_transpose_y_0"), val = tensor(false)]; - tensor value_7_cast_fp16 = transpose(perm = value_7_perm_0, x = v_5_cast_fp16)[name = tensor("transpose_190")]; - tensor x_57_cast_fp16 = matmul(transpose_x = x_57_transpose_x_0, transpose_y = x_57_transpose_y_0, x = input_137_cast_fp16, y = value_7_cast_fp16)[name = tensor("x_57_cast_fp16")]; - tensor var_776_perm_0 = const()[name = tensor("op_776_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_777 = const()[name = tensor("op_777"), val = tensor([1, -1, 512])]; - tensor var_776_cast_fp16 = transpose(perm = var_776_perm_0, x = x_57_cast_fp16)[name = tensor("transpose_186")]; - tensor input_139_cast_fp16 = reshape(shape = var_777, x = var_776_cast_fp16)[name = tensor("input_139_cast_fp16")]; - tensor module_layers_2_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_2_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34038016)))]; - tensor module_layers_2_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_2_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34562368)))]; - tensor linear_25_cast_fp16 = linear(bias = module_layers_2_self_attn_linear_out_bias_to_fp16, weight = module_layers_2_self_attn_linear_out_weight_to_fp16, x = input_139_cast_fp16)[name = tensor("linear_25_cast_fp16")]; - tensor input_143_cast_fp16 = add(x = input_135_cast_fp16, y = linear_25_cast_fp16)[name = tensor("input_143_cast_fp16")]; - tensor x_61_axes_0 = const()[name = tensor("x_61_axes_0"), val = tensor([-1])]; - tensor module_layers_2_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_2_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34563456)))]; - tensor module_layers_2_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_2_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34564544)))]; - tensor x_61_cast_fp16 = layer_norm(axes = x_61_axes_0, beta = module_layers_2_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_2_norm_conv_weight_to_fp16, x = input_143_cast_fp16)[name = tensor("x_61_cast_fp16")]; - tensor input_145_perm_0 = const()[name = tensor("input_145_perm_0"), val = tensor([0, 2, 1])]; - tensor input_147_pad_type_0 = const()[name = tensor("input_147_pad_type_0"), val = tensor("valid")]; - tensor input_147_strides_0 = const()[name = tensor("input_147_strides_0"), val = tensor([1])]; - tensor input_147_pad_0 = const()[name = tensor("input_147_pad_0"), val = tensor([0, 0])]; - tensor input_147_dilations_0 = const()[name = tensor("input_147_dilations_0"), val = tensor([1])]; - tensor input_147_groups_0 = const()[name = tensor("input_147_groups_0"), val = tensor(1)]; - tensor module_layers_2_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_2_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34565632)))]; - tensor module_layers_2_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_2_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35614272)))]; - tensor input_145_cast_fp16 = transpose(perm = input_145_perm_0, x = x_61_cast_fp16)[name = tensor("transpose_185")]; - tensor input_147_cast_fp16 = conv(bias = module_layers_2_conv_pointwise_conv1_bias_to_fp16, dilations = input_147_dilations_0, groups = input_147_groups_0, pad = input_147_pad_0, pad_type = input_147_pad_type_0, strides = input_147_strides_0, weight = module_layers_2_conv_pointwise_conv1_weight_to_fp16, x = input_145_cast_fp16)[name = tensor("input_147_cast_fp16")]; - tensor x_63_split_num_splits_0 = const()[name = tensor("x_63_split_num_splits_0"), val = tensor(2)]; - tensor x_63_split_axis_0 = const()[name = tensor("x_63_split_axis_0"), val = tensor(1)]; - tensor x_63_split_cast_fp16_0, tensor x_63_split_cast_fp16_1 = split(axis = x_63_split_axis_0, num_splits = x_63_split_num_splits_0, x = input_147_cast_fp16)[name = tensor("x_63_split_cast_fp16")]; - tensor x_63_split_1_sigmoid_cast_fp16 = sigmoid(x = x_63_split_cast_fp16_1)[name = tensor("x_63_split_1_sigmoid_cast_fp16")]; - tensor x_63_cast_fp16 = mul(x = x_63_split_cast_fp16_0, y = x_63_split_1_sigmoid_cast_fp16)[name = tensor("x_63_cast_fp16")]; - tensor input_149_cast_fp16 = select(a = var_11_to_fp16, b = x_63_cast_fp16, cond = var_453)[name = tensor("input_149_cast_fp16")]; - tensor input_151_pad_0 = const()[name = tensor("input_151_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_151_mode_0 = const()[name = tensor("input_151_mode_0"), val = tensor("constant")]; - tensor const_93_to_fp16 = const()[name = tensor("const_93_to_fp16"), val = tensor(0x0p+0)]; - tensor input_151_cast_fp16 = pad(constant_val = const_93_to_fp16, mode = input_151_mode_0, pad = input_151_pad_0, x = input_149_cast_fp16)[name = tensor("input_151_cast_fp16")]; - tensor input_153_pad_type_0 = const()[name = tensor("input_153_pad_type_0"), val = tensor("valid")]; - tensor input_153_groups_0 = const()[name = tensor("input_153_groups_0"), val = tensor(512)]; - tensor input_153_strides_0 = const()[name = tensor("input_153_strides_0"), val = tensor([1])]; - tensor input_153_pad_0 = const()[name = tensor("input_153_pad_0"), val = tensor([0, 0])]; - tensor input_153_dilations_0 = const()[name = tensor("input_153_dilations_0"), val = tensor([1])]; - tensor const_238_to_fp16 = const()[name = tensor("const_238_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35616384)))]; - tensor const_239_to_fp16 = const()[name = tensor("const_239_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35625664)))]; - tensor input_155_cast_fp16 = conv(bias = const_239_to_fp16, dilations = input_153_dilations_0, groups = input_153_groups_0, pad = input_153_pad_0, pad_type = input_153_pad_type_0, strides = input_153_strides_0, weight = const_238_to_fp16, x = input_151_cast_fp16)[name = tensor("input_155_cast_fp16")]; - tensor input_157_cast_fp16 = silu(x = input_155_cast_fp16)[name = tensor("input_157_cast_fp16")]; - tensor x_65_pad_type_0 = const()[name = tensor("x_65_pad_type_0"), val = tensor("valid")]; - tensor x_65_strides_0 = const()[name = tensor("x_65_strides_0"), val = tensor([1])]; - tensor x_65_pad_0 = const()[name = tensor("x_65_pad_0"), val = tensor([0, 0])]; - tensor x_65_dilations_0 = const()[name = tensor("x_65_dilations_0"), val = tensor([1])]; - tensor x_65_groups_0 = const()[name = tensor("x_65_groups_0"), val = tensor(1)]; - tensor module_layers_2_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_2_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35626752)))]; - tensor module_layers_2_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_2_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36151104)))]; - tensor x_65_cast_fp16 = conv(bias = module_layers_2_conv_pointwise_conv2_bias_to_fp16, dilations = x_65_dilations_0, groups = x_65_groups_0, pad = x_65_pad_0, pad_type = x_65_pad_type_0, strides = x_65_strides_0, weight = module_layers_2_conv_pointwise_conv2_weight_to_fp16, x = input_157_cast_fp16)[name = tensor("x_65_cast_fp16")]; - tensor input_159_perm_0 = const()[name = tensor("input_159_perm_0"), val = tensor([0, 2, 1])]; - tensor input_159_cast_fp16 = transpose(perm = input_159_perm_0, x = x_65_cast_fp16)[name = tensor("transpose_184")]; - tensor input_161_cast_fp16 = add(x = input_143_cast_fp16, y = input_159_cast_fp16)[name = tensor("input_161_cast_fp16")]; - tensor input_163_axes_0 = const()[name = tensor("input_163_axes_0"), val = tensor([-1])]; - tensor module_layers_2_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_2_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36152192)))]; - tensor module_layers_2_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_2_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36153280)))]; - tensor input_163_cast_fp16 = layer_norm(axes = input_163_axes_0, beta = module_layers_2_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_2_norm_feed_forward2_weight_to_fp16, x = input_161_cast_fp16)[name = tensor("input_163_cast_fp16")]; - tensor module_layers_2_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_2_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36154368)))]; - tensor module_layers_2_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_2_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(38251584)))]; - tensor linear_26_cast_fp16 = linear(bias = module_layers_2_feed_forward2_linear1_bias_to_fp16, weight = module_layers_2_feed_forward2_linear1_weight_to_fp16, x = input_163_cast_fp16)[name = tensor("linear_26_cast_fp16")]; - tensor input_167_cast_fp16 = silu(x = linear_26_cast_fp16)[name = tensor("input_167_cast_fp16")]; - tensor module_layers_2_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_2_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(38255744)))]; - tensor module_layers_2_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_2_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40352960)))]; - tensor linear_27_cast_fp16 = linear(bias = module_layers_2_feed_forward2_linear2_bias_to_fp16, weight = module_layers_2_feed_forward2_linear2_weight_to_fp16, x = input_167_cast_fp16)[name = tensor("linear_27_cast_fp16")]; - tensor var_843_to_fp16 = const()[name = tensor("op_843_to_fp16"), val = tensor(0x1p-1)]; - tensor var_844_cast_fp16 = mul(x = linear_27_cast_fp16, y = var_843_to_fp16)[name = tensor("op_844_cast_fp16")]; - tensor input_173_cast_fp16 = add(x = input_161_cast_fp16, y = var_844_cast_fp16)[name = tensor("input_173_cast_fp16")]; - tensor input_175_axes_0 = const()[name = tensor("input_175_axes_0"), val = tensor([-1])]; - tensor module_layers_2_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_2_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40354048)))]; - tensor module_layers_2_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_2_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40355136)))]; - tensor input_175_cast_fp16 = layer_norm(axes = input_175_axes_0, beta = module_layers_2_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_2_norm_out_weight_to_fp16, x = input_173_cast_fp16)[name = tensor("input_175_cast_fp16")]; - tensor input_177_axes_0 = const()[name = tensor("input_177_axes_0"), val = tensor([-1])]; - tensor module_layers_3_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_3_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40356224)))]; - tensor module_layers_3_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_3_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40357312)))]; - tensor input_177_cast_fp16 = layer_norm(axes = input_177_axes_0, beta = module_layers_3_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_3_norm_feed_forward1_weight_to_fp16, x = input_175_cast_fp16)[name = tensor("input_177_cast_fp16")]; - tensor module_layers_3_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_3_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40358400)))]; - tensor module_layers_3_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_3_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(42455616)))]; - tensor linear_28_cast_fp16 = linear(bias = module_layers_3_feed_forward1_linear1_bias_to_fp16, weight = module_layers_3_feed_forward1_linear1_weight_to_fp16, x = input_177_cast_fp16)[name = tensor("linear_28_cast_fp16")]; - tensor input_181_cast_fp16 = silu(x = linear_28_cast_fp16)[name = tensor("input_181_cast_fp16")]; - tensor module_layers_3_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_3_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(42459776)))]; - tensor module_layers_3_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_3_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44556992)))]; - tensor linear_29_cast_fp16 = linear(bias = module_layers_3_feed_forward1_linear2_bias_to_fp16, weight = module_layers_3_feed_forward1_linear2_weight_to_fp16, x = input_181_cast_fp16)[name = tensor("linear_29_cast_fp16")]; - tensor var_874_to_fp16 = const()[name = tensor("op_874_to_fp16"), val = tensor(0x1p-1)]; - tensor var_875_cast_fp16 = mul(x = linear_29_cast_fp16, y = var_874_to_fp16)[name = tensor("op_875_cast_fp16")]; - tensor input_187_cast_fp16 = add(x = input_175_cast_fp16, y = var_875_cast_fp16)[name = tensor("input_187_cast_fp16")]; - tensor query_7_axes_0 = const()[name = tensor("query_7_axes_0"), val = tensor([-1])]; - tensor module_layers_3_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_3_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44558080)))]; - tensor module_layers_3_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_3_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44559168)))]; - tensor query_7_cast_fp16 = layer_norm(axes = query_7_axes_0, beta = module_layers_3_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_3_norm_self_att_weight_to_fp16, x = input_187_cast_fp16)[name = tensor("query_7_cast_fp16")]; - tensor module_layers_3_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_3_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44560256)))]; - tensor module_layers_3_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_3_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45084608)))]; - tensor linear_30_cast_fp16 = linear(bias = module_layers_3_self_attn_linear_q_bias_to_fp16, weight = module_layers_3_self_attn_linear_q_weight_to_fp16, x = query_7_cast_fp16)[name = tensor("linear_30_cast_fp16")]; - tensor var_892 = const()[name = tensor("op_892"), val = tensor([1, -1, 8, 64])]; - tensor q_19_cast_fp16 = reshape(shape = var_892, x = linear_30_cast_fp16)[name = tensor("q_19_cast_fp16")]; - tensor module_layers_3_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_3_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45085696)))]; - tensor module_layers_3_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_3_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45610048)))]; - tensor linear_31_cast_fp16 = linear(bias = module_layers_3_self_attn_linear_k_bias_to_fp16, weight = module_layers_3_self_attn_linear_k_weight_to_fp16, x = query_7_cast_fp16)[name = tensor("linear_31_cast_fp16")]; - tensor var_897 = const()[name = tensor("op_897"), val = tensor([1, -1, 8, 64])]; - tensor k_13_cast_fp16 = reshape(shape = var_897, x = linear_31_cast_fp16)[name = tensor("k_13_cast_fp16")]; - tensor module_layers_3_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_3_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45611136)))]; - tensor module_layers_3_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_3_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46135488)))]; - tensor linear_32_cast_fp16 = linear(bias = module_layers_3_self_attn_linear_v_bias_to_fp16, weight = module_layers_3_self_attn_linear_v_weight_to_fp16, x = query_7_cast_fp16)[name = tensor("linear_32_cast_fp16")]; - tensor var_902 = const()[name = tensor("op_902"), val = tensor([1, -1, 8, 64])]; - tensor v_7_cast_fp16 = reshape(shape = var_902, x = linear_32_cast_fp16)[name = tensor("v_7_cast_fp16")]; - tensor value_9_perm_0 = const()[name = tensor("value_9_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_3_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_3_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46136576)))]; - tensor var_914_cast_fp16 = add(x = q_19_cast_fp16, y = module_layers_3_self_attn_pos_bias_u_to_fp16)[name = tensor("op_914_cast_fp16")]; - tensor module_layers_3_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_3_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46137664)))]; - tensor var_916_cast_fp16 = add(x = q_19_cast_fp16, y = module_layers_3_self_attn_pos_bias_v_to_fp16)[name = tensor("op_916_cast_fp16")]; - tensor q_with_bias_v_7_perm_0 = const()[name = tensor("q_with_bias_v_7_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_73_transpose_x_0 = const()[name = tensor("x_73_transpose_x_0"), val = tensor(false)]; - tensor x_73_transpose_y_0 = const()[name = tensor("x_73_transpose_y_0"), val = tensor(false)]; - tensor var_918_to_fp16 = const()[name = tensor("op_918_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46138752)))]; - tensor q_with_bias_v_7_cast_fp16 = transpose(perm = q_with_bias_v_7_perm_0, x = var_916_cast_fp16)[name = tensor("transpose_182")]; - tensor x_73_cast_fp16 = matmul(transpose_x = x_73_transpose_x_0, transpose_y = x_73_transpose_y_0, x = q_with_bias_v_7_cast_fp16, y = var_918_to_fp16)[name = tensor("x_73_cast_fp16")]; - tensor x_75_pad_0 = const()[name = tensor("x_75_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_75_mode_0 = const()[name = tensor("x_75_mode_0"), val = tensor("constant")]; - tensor const_100_to_fp16 = const()[name = tensor("const_100_to_fp16"), val = tensor(0x0p+0)]; - tensor x_75_cast_fp16 = pad(constant_val = const_100_to_fp16, mode = x_75_mode_0, pad = x_75_pad_0, x = x_73_cast_fp16)[name = tensor("x_75_cast_fp16")]; - tensor var_926 = const()[name = tensor("op_926"), val = tensor([1, 8, -1, 188])]; - tensor x_77_cast_fp16 = reshape(shape = var_926, x = x_75_cast_fp16)[name = tensor("x_77_cast_fp16")]; - tensor var_930_begin_0 = const()[name = tensor("op_930_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_930_end_0 = const()[name = tensor("op_930_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_930_end_mask_0 = const()[name = tensor("op_930_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_930_cast_fp16 = slice_by_index(begin = var_930_begin_0, end = var_930_end_0, end_mask = var_930_end_mask_0, x = x_77_cast_fp16)[name = tensor("op_930_cast_fp16")]; - tensor var_931 = const()[name = tensor("op_931"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_13_cast_fp16 = reshape(shape = var_931, x = var_930_cast_fp16)[name = tensor("matrix_bd_13_cast_fp16")]; - tensor matrix_ac_7_transpose_x_0 = const()[name = tensor("matrix_ac_7_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_7_transpose_y_0 = const()[name = tensor("matrix_ac_7_transpose_y_0"), val = tensor(false)]; - tensor transpose_57_perm_0 = const()[name = tensor("transpose_57_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_58_perm_0 = const()[name = tensor("transpose_58_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_58 = transpose(perm = transpose_58_perm_0, x = k_13_cast_fp16)[name = tensor("transpose_180")]; - tensor transpose_57 = transpose(perm = transpose_57_perm_0, x = var_914_cast_fp16)[name = tensor("transpose_181")]; - tensor matrix_ac_7_cast_fp16 = matmul(transpose_x = matrix_ac_7_transpose_x_0, transpose_y = matrix_ac_7_transpose_y_0, x = transpose_57, y = transpose_58)[name = tensor("matrix_ac_7_cast_fp16")]; - tensor matrix_bd_15_begin_0 = const()[name = tensor("matrix_bd_15_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_15_end_0 = const()[name = tensor("matrix_bd_15_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_15_end_mask_0 = const()[name = tensor("matrix_bd_15_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_15_cast_fp16 = slice_by_index(begin = matrix_bd_15_begin_0, end = matrix_bd_15_end_0, end_mask = matrix_bd_15_end_mask_0, x = matrix_bd_13_cast_fp16)[name = tensor("matrix_bd_15_cast_fp16")]; - tensor var_940_cast_fp16 = add(x = matrix_ac_7_cast_fp16, y = matrix_bd_15_cast_fp16)[name = tensor("op_940_cast_fp16")]; - tensor _inversed_scores_13_y_0_to_fp16 = const()[name = tensor("_inversed_scores_13_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_13_cast_fp16 = mul(x = var_940_cast_fp16, y = _inversed_scores_13_y_0_to_fp16)[name = tensor("_inversed_scores_13_cast_fp16")]; - tensor scores_15_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_13_cast_fp16, cond = mask_11)[name = tensor("scores_15_cast_fp16")]; - tensor var_946_cast_fp16 = softmax(axis = var_23, x = scores_15_cast_fp16)[name = tensor("op_946_cast_fp16")]; - tensor input_189_cast_fp16 = select(a = var_11_to_fp16, b = var_946_cast_fp16, cond = mask_11)[name = tensor("input_189_cast_fp16")]; - tensor x_79_transpose_x_0 = const()[name = tensor("x_79_transpose_x_0"), val = tensor(false)]; - tensor x_79_transpose_y_0 = const()[name = tensor("x_79_transpose_y_0"), val = tensor(false)]; - tensor value_9_cast_fp16 = transpose(perm = value_9_perm_0, x = v_7_cast_fp16)[name = tensor("transpose_183")]; - tensor x_79_cast_fp16 = matmul(transpose_x = x_79_transpose_x_0, transpose_y = x_79_transpose_y_0, x = input_189_cast_fp16, y = value_9_cast_fp16)[name = tensor("x_79_cast_fp16")]; - tensor var_950_perm_0 = const()[name = tensor("op_950_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_951 = const()[name = tensor("op_951"), val = tensor([1, -1, 512])]; - tensor var_950_cast_fp16 = transpose(perm = var_950_perm_0, x = x_79_cast_fp16)[name = tensor("transpose_179")]; - tensor input_191_cast_fp16 = reshape(shape = var_951, x = var_950_cast_fp16)[name = tensor("input_191_cast_fp16")]; - tensor module_layers_3_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_3_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46522816)))]; - tensor module_layers_3_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_3_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47047168)))]; - tensor linear_34_cast_fp16 = linear(bias = module_layers_3_self_attn_linear_out_bias_to_fp16, weight = module_layers_3_self_attn_linear_out_weight_to_fp16, x = input_191_cast_fp16)[name = tensor("linear_34_cast_fp16")]; - tensor input_195_cast_fp16 = add(x = input_187_cast_fp16, y = linear_34_cast_fp16)[name = tensor("input_195_cast_fp16")]; - tensor x_83_axes_0 = const()[name = tensor("x_83_axes_0"), val = tensor([-1])]; - tensor module_layers_3_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_3_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47048256)))]; - tensor module_layers_3_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_3_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47049344)))]; - tensor x_83_cast_fp16 = layer_norm(axes = x_83_axes_0, beta = module_layers_3_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_3_norm_conv_weight_to_fp16, x = input_195_cast_fp16)[name = tensor("x_83_cast_fp16")]; - tensor input_197_perm_0 = const()[name = tensor("input_197_perm_0"), val = tensor([0, 2, 1])]; - tensor input_199_pad_type_0 = const()[name = tensor("input_199_pad_type_0"), val = tensor("valid")]; - tensor input_199_strides_0 = const()[name = tensor("input_199_strides_0"), val = tensor([1])]; - tensor input_199_pad_0 = const()[name = tensor("input_199_pad_0"), val = tensor([0, 0])]; - tensor input_199_dilations_0 = const()[name = tensor("input_199_dilations_0"), val = tensor([1])]; - tensor input_199_groups_0 = const()[name = tensor("input_199_groups_0"), val = tensor(1)]; - tensor module_layers_3_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_3_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47050432)))]; - tensor module_layers_3_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_3_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48099072)))]; - tensor input_197_cast_fp16 = transpose(perm = input_197_perm_0, x = x_83_cast_fp16)[name = tensor("transpose_178")]; - tensor input_199_cast_fp16 = conv(bias = module_layers_3_conv_pointwise_conv1_bias_to_fp16, dilations = input_199_dilations_0, groups = input_199_groups_0, pad = input_199_pad_0, pad_type = input_199_pad_type_0, strides = input_199_strides_0, weight = module_layers_3_conv_pointwise_conv1_weight_to_fp16, x = input_197_cast_fp16)[name = tensor("input_199_cast_fp16")]; - tensor x_85_split_num_splits_0 = const()[name = tensor("x_85_split_num_splits_0"), val = tensor(2)]; - tensor x_85_split_axis_0 = const()[name = tensor("x_85_split_axis_0"), val = tensor(1)]; - tensor x_85_split_cast_fp16_0, tensor x_85_split_cast_fp16_1 = split(axis = x_85_split_axis_0, num_splits = x_85_split_num_splits_0, x = input_199_cast_fp16)[name = tensor("x_85_split_cast_fp16")]; - tensor x_85_split_1_sigmoid_cast_fp16 = sigmoid(x = x_85_split_cast_fp16_1)[name = tensor("x_85_split_1_sigmoid_cast_fp16")]; - tensor x_85_cast_fp16 = mul(x = x_85_split_cast_fp16_0, y = x_85_split_1_sigmoid_cast_fp16)[name = tensor("x_85_cast_fp16")]; - tensor input_201_cast_fp16 = select(a = var_11_to_fp16, b = x_85_cast_fp16, cond = var_453)[name = tensor("input_201_cast_fp16")]; - tensor input_203_pad_0 = const()[name = tensor("input_203_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_203_mode_0 = const()[name = tensor("input_203_mode_0"), val = tensor("constant")]; - tensor const_103_to_fp16 = const()[name = tensor("const_103_to_fp16"), val = tensor(0x0p+0)]; - tensor input_203_cast_fp16 = pad(constant_val = const_103_to_fp16, mode = input_203_mode_0, pad = input_203_pad_0, x = input_201_cast_fp16)[name = tensor("input_203_cast_fp16")]; - tensor input_205_pad_type_0 = const()[name = tensor("input_205_pad_type_0"), val = tensor("valid")]; - tensor input_205_groups_0 = const()[name = tensor("input_205_groups_0"), val = tensor(512)]; - tensor input_205_strides_0 = const()[name = tensor("input_205_strides_0"), val = tensor([1])]; - tensor input_205_pad_0 = const()[name = tensor("input_205_pad_0"), val = tensor([0, 0])]; - tensor input_205_dilations_0 = const()[name = tensor("input_205_dilations_0"), val = tensor([1])]; - tensor const_240_to_fp16 = const()[name = tensor("const_240_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48101184)))]; - tensor const_241_to_fp16 = const()[name = tensor("const_241_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48110464)))]; - tensor input_207_cast_fp16 = conv(bias = const_241_to_fp16, dilations = input_205_dilations_0, groups = input_205_groups_0, pad = input_205_pad_0, pad_type = input_205_pad_type_0, strides = input_205_strides_0, weight = const_240_to_fp16, x = input_203_cast_fp16)[name = tensor("input_207_cast_fp16")]; - tensor input_209_cast_fp16 = silu(x = input_207_cast_fp16)[name = tensor("input_209_cast_fp16")]; - tensor x_87_pad_type_0 = const()[name = tensor("x_87_pad_type_0"), val = tensor("valid")]; - tensor x_87_strides_0 = const()[name = tensor("x_87_strides_0"), val = tensor([1])]; - tensor x_87_pad_0 = const()[name = tensor("x_87_pad_0"), val = tensor([0, 0])]; - tensor x_87_dilations_0 = const()[name = tensor("x_87_dilations_0"), val = tensor([1])]; - tensor x_87_groups_0 = const()[name = tensor("x_87_groups_0"), val = tensor(1)]; - tensor module_layers_3_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_3_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48111552)))]; - tensor module_layers_3_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_3_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48635904)))]; - tensor x_87_cast_fp16 = conv(bias = module_layers_3_conv_pointwise_conv2_bias_to_fp16, dilations = x_87_dilations_0, groups = x_87_groups_0, pad = x_87_pad_0, pad_type = x_87_pad_type_0, strides = x_87_strides_0, weight = module_layers_3_conv_pointwise_conv2_weight_to_fp16, x = input_209_cast_fp16)[name = tensor("x_87_cast_fp16")]; - tensor input_211_perm_0 = const()[name = tensor("input_211_perm_0"), val = tensor([0, 2, 1])]; - tensor input_211_cast_fp16 = transpose(perm = input_211_perm_0, x = x_87_cast_fp16)[name = tensor("transpose_177")]; - tensor input_213_cast_fp16 = add(x = input_195_cast_fp16, y = input_211_cast_fp16)[name = tensor("input_213_cast_fp16")]; - tensor input_215_axes_0 = const()[name = tensor("input_215_axes_0"), val = tensor([-1])]; - tensor module_layers_3_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_3_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48636992)))]; - tensor module_layers_3_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_3_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48638080)))]; - tensor input_215_cast_fp16 = layer_norm(axes = input_215_axes_0, beta = module_layers_3_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_3_norm_feed_forward2_weight_to_fp16, x = input_213_cast_fp16)[name = tensor("input_215_cast_fp16")]; - tensor module_layers_3_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_3_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48639168)))]; - tensor module_layers_3_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_3_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(50736384)))]; - tensor linear_35_cast_fp16 = linear(bias = module_layers_3_feed_forward2_linear1_bias_to_fp16, weight = module_layers_3_feed_forward2_linear1_weight_to_fp16, x = input_215_cast_fp16)[name = tensor("linear_35_cast_fp16")]; - tensor input_219_cast_fp16 = silu(x = linear_35_cast_fp16)[name = tensor("input_219_cast_fp16")]; - tensor module_layers_3_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_3_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(50740544)))]; - tensor module_layers_3_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_3_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52837760)))]; - tensor linear_36_cast_fp16 = linear(bias = module_layers_3_feed_forward2_linear2_bias_to_fp16, weight = module_layers_3_feed_forward2_linear2_weight_to_fp16, x = input_219_cast_fp16)[name = tensor("linear_36_cast_fp16")]; - tensor var_1017_to_fp16 = const()[name = tensor("op_1017_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1018_cast_fp16 = mul(x = linear_36_cast_fp16, y = var_1017_to_fp16)[name = tensor("op_1018_cast_fp16")]; - tensor input_225_cast_fp16 = add(x = input_213_cast_fp16, y = var_1018_cast_fp16)[name = tensor("input_225_cast_fp16")]; - tensor input_227_axes_0 = const()[name = tensor("input_227_axes_0"), val = tensor([-1])]; - tensor module_layers_3_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_3_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52838848)))]; - tensor module_layers_3_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_3_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52839936)))]; - tensor input_227_cast_fp16 = layer_norm(axes = input_227_axes_0, beta = module_layers_3_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_3_norm_out_weight_to_fp16, x = input_225_cast_fp16)[name = tensor("input_227_cast_fp16")]; - tensor input_229_axes_0 = const()[name = tensor("input_229_axes_0"), val = tensor([-1])]; - tensor module_layers_4_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_4_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52841024)))]; - tensor module_layers_4_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_4_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52842112)))]; - tensor input_229_cast_fp16 = layer_norm(axes = input_229_axes_0, beta = module_layers_4_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_4_norm_feed_forward1_weight_to_fp16, x = input_227_cast_fp16)[name = tensor("input_229_cast_fp16")]; - tensor module_layers_4_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_4_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52843200)))]; - tensor module_layers_4_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_4_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(54940416)))]; - tensor linear_37_cast_fp16 = linear(bias = module_layers_4_feed_forward1_linear1_bias_to_fp16, weight = module_layers_4_feed_forward1_linear1_weight_to_fp16, x = input_229_cast_fp16)[name = tensor("linear_37_cast_fp16")]; - tensor input_233_cast_fp16 = silu(x = linear_37_cast_fp16)[name = tensor("input_233_cast_fp16")]; - tensor module_layers_4_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_4_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(54944576)))]; - tensor module_layers_4_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_4_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57041792)))]; - tensor linear_38_cast_fp16 = linear(bias = module_layers_4_feed_forward1_linear2_bias_to_fp16, weight = module_layers_4_feed_forward1_linear2_weight_to_fp16, x = input_233_cast_fp16)[name = tensor("linear_38_cast_fp16")]; - tensor var_1048_to_fp16 = const()[name = tensor("op_1048_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1049_cast_fp16 = mul(x = linear_38_cast_fp16, y = var_1048_to_fp16)[name = tensor("op_1049_cast_fp16")]; - tensor input_239_cast_fp16 = add(x = input_227_cast_fp16, y = var_1049_cast_fp16)[name = tensor("input_239_cast_fp16")]; - tensor query_9_axes_0 = const()[name = tensor("query_9_axes_0"), val = tensor([-1])]; - tensor module_layers_4_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_4_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57042880)))]; - tensor module_layers_4_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_4_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57043968)))]; - tensor query_9_cast_fp16 = layer_norm(axes = query_9_axes_0, beta = module_layers_4_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_4_norm_self_att_weight_to_fp16, x = input_239_cast_fp16)[name = tensor("query_9_cast_fp16")]; - tensor module_layers_4_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_4_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57045056)))]; - tensor module_layers_4_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_4_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57569408)))]; - tensor linear_39_cast_fp16 = linear(bias = module_layers_4_self_attn_linear_q_bias_to_fp16, weight = module_layers_4_self_attn_linear_q_weight_to_fp16, x = query_9_cast_fp16)[name = tensor("linear_39_cast_fp16")]; - tensor var_1066 = const()[name = tensor("op_1066"), val = tensor([1, -1, 8, 64])]; - tensor q_25_cast_fp16 = reshape(shape = var_1066, x = linear_39_cast_fp16)[name = tensor("q_25_cast_fp16")]; - tensor module_layers_4_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_4_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57570496)))]; - tensor module_layers_4_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_4_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58094848)))]; - tensor linear_40_cast_fp16 = linear(bias = module_layers_4_self_attn_linear_k_bias_to_fp16, weight = module_layers_4_self_attn_linear_k_weight_to_fp16, x = query_9_cast_fp16)[name = tensor("linear_40_cast_fp16")]; - tensor var_1071 = const()[name = tensor("op_1071"), val = tensor([1, -1, 8, 64])]; - tensor k_17_cast_fp16 = reshape(shape = var_1071, x = linear_40_cast_fp16)[name = tensor("k_17_cast_fp16")]; - tensor module_layers_4_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_4_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58095936)))]; - tensor module_layers_4_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_4_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58620288)))]; - tensor linear_41_cast_fp16 = linear(bias = module_layers_4_self_attn_linear_v_bias_to_fp16, weight = module_layers_4_self_attn_linear_v_weight_to_fp16, x = query_9_cast_fp16)[name = tensor("linear_41_cast_fp16")]; - tensor var_1076 = const()[name = tensor("op_1076"), val = tensor([1, -1, 8, 64])]; - tensor v_9_cast_fp16 = reshape(shape = var_1076, x = linear_41_cast_fp16)[name = tensor("v_9_cast_fp16")]; - tensor value_11_perm_0 = const()[name = tensor("value_11_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_4_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_4_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58621376)))]; - tensor var_1088_cast_fp16 = add(x = q_25_cast_fp16, y = module_layers_4_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1088_cast_fp16")]; - tensor module_layers_4_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_4_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58622464)))]; - tensor var_1090_cast_fp16 = add(x = q_25_cast_fp16, y = module_layers_4_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1090_cast_fp16")]; - tensor q_with_bias_v_9_perm_0 = const()[name = tensor("q_with_bias_v_9_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_95_transpose_x_0 = const()[name = tensor("x_95_transpose_x_0"), val = tensor(false)]; - tensor x_95_transpose_y_0 = const()[name = tensor("x_95_transpose_y_0"), val = tensor(false)]; - tensor var_1092_to_fp16 = const()[name = tensor("op_1092_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58623552)))]; - tensor q_with_bias_v_9_cast_fp16 = transpose(perm = q_with_bias_v_9_perm_0, x = var_1090_cast_fp16)[name = tensor("transpose_175")]; - tensor x_95_cast_fp16 = matmul(transpose_x = x_95_transpose_x_0, transpose_y = x_95_transpose_y_0, x = q_with_bias_v_9_cast_fp16, y = var_1092_to_fp16)[name = tensor("x_95_cast_fp16")]; - tensor x_97_pad_0 = const()[name = tensor("x_97_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_97_mode_0 = const()[name = tensor("x_97_mode_0"), val = tensor("constant")]; - tensor const_110_to_fp16 = const()[name = tensor("const_110_to_fp16"), val = tensor(0x0p+0)]; - tensor x_97_cast_fp16 = pad(constant_val = const_110_to_fp16, mode = x_97_mode_0, pad = x_97_pad_0, x = x_95_cast_fp16)[name = tensor("x_97_cast_fp16")]; - tensor var_1100 = const()[name = tensor("op_1100"), val = tensor([1, 8, -1, 188])]; - tensor x_99_cast_fp16 = reshape(shape = var_1100, x = x_97_cast_fp16)[name = tensor("x_99_cast_fp16")]; - tensor var_1104_begin_0 = const()[name = tensor("op_1104_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_1104_end_0 = const()[name = tensor("op_1104_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_1104_end_mask_0 = const()[name = tensor("op_1104_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_1104_cast_fp16 = slice_by_index(begin = var_1104_begin_0, end = var_1104_end_0, end_mask = var_1104_end_mask_0, x = x_99_cast_fp16)[name = tensor("op_1104_cast_fp16")]; - tensor var_1105 = const()[name = tensor("op_1105"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_17_cast_fp16 = reshape(shape = var_1105, x = var_1104_cast_fp16)[name = tensor("matrix_bd_17_cast_fp16")]; - tensor matrix_ac_9_transpose_x_0 = const()[name = tensor("matrix_ac_9_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_9_transpose_y_0 = const()[name = tensor("matrix_ac_9_transpose_y_0"), val = tensor(false)]; - tensor transpose_59_perm_0 = const()[name = tensor("transpose_59_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_60_perm_0 = const()[name = tensor("transpose_60_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_60 = transpose(perm = transpose_60_perm_0, x = k_17_cast_fp16)[name = tensor("transpose_173")]; - tensor transpose_59 = transpose(perm = transpose_59_perm_0, x = var_1088_cast_fp16)[name = tensor("transpose_174")]; - tensor matrix_ac_9_cast_fp16 = matmul(transpose_x = matrix_ac_9_transpose_x_0, transpose_y = matrix_ac_9_transpose_y_0, x = transpose_59, y = transpose_60)[name = tensor("matrix_ac_9_cast_fp16")]; - tensor matrix_bd_19_begin_0 = const()[name = tensor("matrix_bd_19_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_19_end_0 = const()[name = tensor("matrix_bd_19_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_19_end_mask_0 = const()[name = tensor("matrix_bd_19_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_19_cast_fp16 = slice_by_index(begin = matrix_bd_19_begin_0, end = matrix_bd_19_end_0, end_mask = matrix_bd_19_end_mask_0, x = matrix_bd_17_cast_fp16)[name = tensor("matrix_bd_19_cast_fp16")]; - tensor var_1114_cast_fp16 = add(x = matrix_ac_9_cast_fp16, y = matrix_bd_19_cast_fp16)[name = tensor("op_1114_cast_fp16")]; - tensor _inversed_scores_17_y_0_to_fp16 = const()[name = tensor("_inversed_scores_17_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_17_cast_fp16 = mul(x = var_1114_cast_fp16, y = _inversed_scores_17_y_0_to_fp16)[name = tensor("_inversed_scores_17_cast_fp16")]; - tensor scores_19_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_17_cast_fp16, cond = mask_11)[name = tensor("scores_19_cast_fp16")]; - tensor var_1120_cast_fp16 = softmax(axis = var_23, x = scores_19_cast_fp16)[name = tensor("op_1120_cast_fp16")]; - tensor input_241_cast_fp16 = select(a = var_11_to_fp16, b = var_1120_cast_fp16, cond = mask_11)[name = tensor("input_241_cast_fp16")]; - tensor x_101_transpose_x_0 = const()[name = tensor("x_101_transpose_x_0"), val = tensor(false)]; - tensor x_101_transpose_y_0 = const()[name = tensor("x_101_transpose_y_0"), val = tensor(false)]; - tensor value_11_cast_fp16 = transpose(perm = value_11_perm_0, x = v_9_cast_fp16)[name = tensor("transpose_176")]; - tensor x_101_cast_fp16 = matmul(transpose_x = x_101_transpose_x_0, transpose_y = x_101_transpose_y_0, x = input_241_cast_fp16, y = value_11_cast_fp16)[name = tensor("x_101_cast_fp16")]; - tensor var_1124_perm_0 = const()[name = tensor("op_1124_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_1125 = const()[name = tensor("op_1125"), val = tensor([1, -1, 512])]; - tensor var_1124_cast_fp16 = transpose(perm = var_1124_perm_0, x = x_101_cast_fp16)[name = tensor("transpose_172")]; - tensor input_243_cast_fp16 = reshape(shape = var_1125, x = var_1124_cast_fp16)[name = tensor("input_243_cast_fp16")]; - tensor module_layers_4_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_4_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59007616)))]; - tensor module_layers_4_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_4_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59531968)))]; - tensor linear_43_cast_fp16 = linear(bias = module_layers_4_self_attn_linear_out_bias_to_fp16, weight = module_layers_4_self_attn_linear_out_weight_to_fp16, x = input_243_cast_fp16)[name = tensor("linear_43_cast_fp16")]; - tensor input_247_cast_fp16 = add(x = input_239_cast_fp16, y = linear_43_cast_fp16)[name = tensor("input_247_cast_fp16")]; - tensor x_105_axes_0 = const()[name = tensor("x_105_axes_0"), val = tensor([-1])]; - tensor module_layers_4_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_4_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59533056)))]; - tensor module_layers_4_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_4_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59534144)))]; - tensor x_105_cast_fp16 = layer_norm(axes = x_105_axes_0, beta = module_layers_4_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_4_norm_conv_weight_to_fp16, x = input_247_cast_fp16)[name = tensor("x_105_cast_fp16")]; - tensor input_249_perm_0 = const()[name = tensor("input_249_perm_0"), val = tensor([0, 2, 1])]; - tensor input_251_pad_type_0 = const()[name = tensor("input_251_pad_type_0"), val = tensor("valid")]; - tensor input_251_strides_0 = const()[name = tensor("input_251_strides_0"), val = tensor([1])]; - tensor input_251_pad_0 = const()[name = tensor("input_251_pad_0"), val = tensor([0, 0])]; - tensor input_251_dilations_0 = const()[name = tensor("input_251_dilations_0"), val = tensor([1])]; - tensor input_251_groups_0 = const()[name = tensor("input_251_groups_0"), val = tensor(1)]; - tensor module_layers_4_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_4_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59535232)))]; - tensor module_layers_4_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_4_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60583872)))]; - tensor input_249_cast_fp16 = transpose(perm = input_249_perm_0, x = x_105_cast_fp16)[name = tensor("transpose_171")]; - tensor input_251_cast_fp16 = conv(bias = module_layers_4_conv_pointwise_conv1_bias_to_fp16, dilations = input_251_dilations_0, groups = input_251_groups_0, pad = input_251_pad_0, pad_type = input_251_pad_type_0, strides = input_251_strides_0, weight = module_layers_4_conv_pointwise_conv1_weight_to_fp16, x = input_249_cast_fp16)[name = tensor("input_251_cast_fp16")]; - tensor x_107_split_num_splits_0 = const()[name = tensor("x_107_split_num_splits_0"), val = tensor(2)]; - tensor x_107_split_axis_0 = const()[name = tensor("x_107_split_axis_0"), val = tensor(1)]; - tensor x_107_split_cast_fp16_0, tensor x_107_split_cast_fp16_1 = split(axis = x_107_split_axis_0, num_splits = x_107_split_num_splits_0, x = input_251_cast_fp16)[name = tensor("x_107_split_cast_fp16")]; - tensor x_107_split_1_sigmoid_cast_fp16 = sigmoid(x = x_107_split_cast_fp16_1)[name = tensor("x_107_split_1_sigmoid_cast_fp16")]; - tensor x_107_cast_fp16 = mul(x = x_107_split_cast_fp16_0, y = x_107_split_1_sigmoid_cast_fp16)[name = tensor("x_107_cast_fp16")]; - tensor input_253_cast_fp16 = select(a = var_11_to_fp16, b = x_107_cast_fp16, cond = var_453)[name = tensor("input_253_cast_fp16")]; - tensor input_255_pad_0 = const()[name = tensor("input_255_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_255_mode_0 = const()[name = tensor("input_255_mode_0"), val = tensor("constant")]; - tensor const_113_to_fp16 = const()[name = tensor("const_113_to_fp16"), val = tensor(0x0p+0)]; - tensor input_255_cast_fp16 = pad(constant_val = const_113_to_fp16, mode = input_255_mode_0, pad = input_255_pad_0, x = input_253_cast_fp16)[name = tensor("input_255_cast_fp16")]; - tensor input_257_pad_type_0 = const()[name = tensor("input_257_pad_type_0"), val = tensor("valid")]; - tensor input_257_groups_0 = const()[name = tensor("input_257_groups_0"), val = tensor(512)]; - tensor input_257_strides_0 = const()[name = tensor("input_257_strides_0"), val = tensor([1])]; - tensor input_257_pad_0 = const()[name = tensor("input_257_pad_0"), val = tensor([0, 0])]; - tensor input_257_dilations_0 = const()[name = tensor("input_257_dilations_0"), val = tensor([1])]; - tensor const_242_to_fp16 = const()[name = tensor("const_242_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60585984)))]; - tensor const_243_to_fp16 = const()[name = tensor("const_243_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60595264)))]; - tensor input_259_cast_fp16 = conv(bias = const_243_to_fp16, dilations = input_257_dilations_0, groups = input_257_groups_0, pad = input_257_pad_0, pad_type = input_257_pad_type_0, strides = input_257_strides_0, weight = const_242_to_fp16, x = input_255_cast_fp16)[name = tensor("input_259_cast_fp16")]; - tensor input_261_cast_fp16 = silu(x = input_259_cast_fp16)[name = tensor("input_261_cast_fp16")]; - tensor x_109_pad_type_0 = const()[name = tensor("x_109_pad_type_0"), val = tensor("valid")]; - tensor x_109_strides_0 = const()[name = tensor("x_109_strides_0"), val = tensor([1])]; - tensor x_109_pad_0 = const()[name = tensor("x_109_pad_0"), val = tensor([0, 0])]; - tensor x_109_dilations_0 = const()[name = tensor("x_109_dilations_0"), val = tensor([1])]; - tensor x_109_groups_0 = const()[name = tensor("x_109_groups_0"), val = tensor(1)]; - tensor module_layers_4_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_4_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60596352)))]; - tensor module_layers_4_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_4_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61120704)))]; - tensor x_109_cast_fp16 = conv(bias = module_layers_4_conv_pointwise_conv2_bias_to_fp16, dilations = x_109_dilations_0, groups = x_109_groups_0, pad = x_109_pad_0, pad_type = x_109_pad_type_0, strides = x_109_strides_0, weight = module_layers_4_conv_pointwise_conv2_weight_to_fp16, x = input_261_cast_fp16)[name = tensor("x_109_cast_fp16")]; - tensor input_263_perm_0 = const()[name = tensor("input_263_perm_0"), val = tensor([0, 2, 1])]; - tensor input_263_cast_fp16 = transpose(perm = input_263_perm_0, x = x_109_cast_fp16)[name = tensor("transpose_170")]; - tensor input_265_cast_fp16 = add(x = input_247_cast_fp16, y = input_263_cast_fp16)[name = tensor("input_265_cast_fp16")]; - tensor input_267_axes_0 = const()[name = tensor("input_267_axes_0"), val = tensor([-1])]; - tensor module_layers_4_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_4_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61121792)))]; - tensor module_layers_4_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_4_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61122880)))]; - tensor input_267_cast_fp16 = layer_norm(axes = input_267_axes_0, beta = module_layers_4_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_4_norm_feed_forward2_weight_to_fp16, x = input_265_cast_fp16)[name = tensor("input_267_cast_fp16")]; - tensor module_layers_4_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_4_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61123968)))]; - tensor module_layers_4_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_4_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(63221184)))]; - tensor linear_44_cast_fp16 = linear(bias = module_layers_4_feed_forward2_linear1_bias_to_fp16, weight = module_layers_4_feed_forward2_linear1_weight_to_fp16, x = input_267_cast_fp16)[name = tensor("linear_44_cast_fp16")]; - tensor input_271_cast_fp16 = silu(x = linear_44_cast_fp16)[name = tensor("input_271_cast_fp16")]; - tensor module_layers_4_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_4_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(63225344)))]; - tensor module_layers_4_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_4_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65322560)))]; - tensor linear_45_cast_fp16 = linear(bias = module_layers_4_feed_forward2_linear2_bias_to_fp16, weight = module_layers_4_feed_forward2_linear2_weight_to_fp16, x = input_271_cast_fp16)[name = tensor("linear_45_cast_fp16")]; - tensor var_1191_to_fp16 = const()[name = tensor("op_1191_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1192_cast_fp16 = mul(x = linear_45_cast_fp16, y = var_1191_to_fp16)[name = tensor("op_1192_cast_fp16")]; - tensor input_277_cast_fp16 = add(x = input_265_cast_fp16, y = var_1192_cast_fp16)[name = tensor("input_277_cast_fp16")]; - tensor input_279_axes_0 = const()[name = tensor("input_279_axes_0"), val = tensor([-1])]; - tensor module_layers_4_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_4_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65323648)))]; - tensor module_layers_4_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_4_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65324736)))]; - tensor input_279_cast_fp16 = layer_norm(axes = input_279_axes_0, beta = module_layers_4_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_4_norm_out_weight_to_fp16, x = input_277_cast_fp16)[name = tensor("input_279_cast_fp16")]; - tensor input_281_axes_0 = const()[name = tensor("input_281_axes_0"), val = tensor([-1])]; - tensor module_layers_5_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_5_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65325824)))]; - tensor module_layers_5_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_5_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65326912)))]; - tensor input_281_cast_fp16 = layer_norm(axes = input_281_axes_0, beta = module_layers_5_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_5_norm_feed_forward1_weight_to_fp16, x = input_279_cast_fp16)[name = tensor("input_281_cast_fp16")]; - tensor module_layers_5_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_5_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65328000)))]; - tensor module_layers_5_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_5_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67425216)))]; - tensor linear_46_cast_fp16 = linear(bias = module_layers_5_feed_forward1_linear1_bias_to_fp16, weight = module_layers_5_feed_forward1_linear1_weight_to_fp16, x = input_281_cast_fp16)[name = tensor("linear_46_cast_fp16")]; - tensor input_285_cast_fp16 = silu(x = linear_46_cast_fp16)[name = tensor("input_285_cast_fp16")]; - tensor module_layers_5_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_5_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67429376)))]; - tensor module_layers_5_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_5_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(69526592)))]; - tensor linear_47_cast_fp16 = linear(bias = module_layers_5_feed_forward1_linear2_bias_to_fp16, weight = module_layers_5_feed_forward1_linear2_weight_to_fp16, x = input_285_cast_fp16)[name = tensor("linear_47_cast_fp16")]; - tensor var_1222_to_fp16 = const()[name = tensor("op_1222_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1223_cast_fp16 = mul(x = linear_47_cast_fp16, y = var_1222_to_fp16)[name = tensor("op_1223_cast_fp16")]; - tensor input_291_cast_fp16 = add(x = input_279_cast_fp16, y = var_1223_cast_fp16)[name = tensor("input_291_cast_fp16")]; - tensor query_11_axes_0 = const()[name = tensor("query_11_axes_0"), val = tensor([-1])]; - tensor module_layers_5_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_5_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(69527680)))]; - tensor module_layers_5_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_5_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(69528768)))]; - tensor query_11_cast_fp16 = layer_norm(axes = query_11_axes_0, beta = module_layers_5_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_5_norm_self_att_weight_to_fp16, x = input_291_cast_fp16)[name = tensor("query_11_cast_fp16")]; - tensor module_layers_5_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_5_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(69529856)))]; - tensor module_layers_5_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_5_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70054208)))]; - tensor linear_48_cast_fp16 = linear(bias = module_layers_5_self_attn_linear_q_bias_to_fp16, weight = module_layers_5_self_attn_linear_q_weight_to_fp16, x = query_11_cast_fp16)[name = tensor("linear_48_cast_fp16")]; - tensor var_1240 = const()[name = tensor("op_1240"), val = tensor([1, -1, 8, 64])]; - tensor q_31_cast_fp16 = reshape(shape = var_1240, x = linear_48_cast_fp16)[name = tensor("q_31_cast_fp16")]; - tensor module_layers_5_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_5_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70055296)))]; - tensor module_layers_5_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_5_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70579648)))]; - tensor linear_49_cast_fp16 = linear(bias = module_layers_5_self_attn_linear_k_bias_to_fp16, weight = module_layers_5_self_attn_linear_k_weight_to_fp16, x = query_11_cast_fp16)[name = tensor("linear_49_cast_fp16")]; - tensor var_1245 = const()[name = tensor("op_1245"), val = tensor([1, -1, 8, 64])]; - tensor k_21_cast_fp16 = reshape(shape = var_1245, x = linear_49_cast_fp16)[name = tensor("k_21_cast_fp16")]; - tensor module_layers_5_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_5_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70580736)))]; - tensor module_layers_5_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_5_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71105088)))]; - tensor linear_50_cast_fp16 = linear(bias = module_layers_5_self_attn_linear_v_bias_to_fp16, weight = module_layers_5_self_attn_linear_v_weight_to_fp16, x = query_11_cast_fp16)[name = tensor("linear_50_cast_fp16")]; - tensor var_1250 = const()[name = tensor("op_1250"), val = tensor([1, -1, 8, 64])]; - tensor v_11_cast_fp16 = reshape(shape = var_1250, x = linear_50_cast_fp16)[name = tensor("v_11_cast_fp16")]; - tensor value_13_perm_0 = const()[name = tensor("value_13_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_5_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_5_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71106176)))]; - tensor var_1262_cast_fp16 = add(x = q_31_cast_fp16, y = module_layers_5_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1262_cast_fp16")]; - tensor module_layers_5_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_5_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71107264)))]; - tensor var_1264_cast_fp16 = add(x = q_31_cast_fp16, y = module_layers_5_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1264_cast_fp16")]; - tensor q_with_bias_v_11_perm_0 = const()[name = tensor("q_with_bias_v_11_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_117_transpose_x_0 = const()[name = tensor("x_117_transpose_x_0"), val = tensor(false)]; - tensor x_117_transpose_y_0 = const()[name = tensor("x_117_transpose_y_0"), val = tensor(false)]; - tensor var_1266_to_fp16 = const()[name = tensor("op_1266_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71108352)))]; - tensor q_with_bias_v_11_cast_fp16 = transpose(perm = q_with_bias_v_11_perm_0, x = var_1264_cast_fp16)[name = tensor("transpose_168")]; - tensor x_117_cast_fp16 = matmul(transpose_x = x_117_transpose_x_0, transpose_y = x_117_transpose_y_0, x = q_with_bias_v_11_cast_fp16, y = var_1266_to_fp16)[name = tensor("x_117_cast_fp16")]; - tensor x_119_pad_0 = const()[name = tensor("x_119_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_119_mode_0 = const()[name = tensor("x_119_mode_0"), val = tensor("constant")]; - tensor const_120_to_fp16 = const()[name = tensor("const_120_to_fp16"), val = tensor(0x0p+0)]; - tensor x_119_cast_fp16 = pad(constant_val = const_120_to_fp16, mode = x_119_mode_0, pad = x_119_pad_0, x = x_117_cast_fp16)[name = tensor("x_119_cast_fp16")]; - tensor var_1274 = const()[name = tensor("op_1274"), val = tensor([1, 8, -1, 188])]; - tensor x_121_cast_fp16 = reshape(shape = var_1274, x = x_119_cast_fp16)[name = tensor("x_121_cast_fp16")]; - tensor var_1278_begin_0 = const()[name = tensor("op_1278_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_1278_end_0 = const()[name = tensor("op_1278_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_1278_end_mask_0 = const()[name = tensor("op_1278_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_1278_cast_fp16 = slice_by_index(begin = var_1278_begin_0, end = var_1278_end_0, end_mask = var_1278_end_mask_0, x = x_121_cast_fp16)[name = tensor("op_1278_cast_fp16")]; - tensor var_1279 = const()[name = tensor("op_1279"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_21_cast_fp16 = reshape(shape = var_1279, x = var_1278_cast_fp16)[name = tensor("matrix_bd_21_cast_fp16")]; - tensor matrix_ac_11_transpose_x_0 = const()[name = tensor("matrix_ac_11_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_11_transpose_y_0 = const()[name = tensor("matrix_ac_11_transpose_y_0"), val = tensor(false)]; - tensor transpose_61_perm_0 = const()[name = tensor("transpose_61_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_62_perm_0 = const()[name = tensor("transpose_62_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_62 = transpose(perm = transpose_62_perm_0, x = k_21_cast_fp16)[name = tensor("transpose_166")]; - tensor transpose_61 = transpose(perm = transpose_61_perm_0, x = var_1262_cast_fp16)[name = tensor("transpose_167")]; - tensor matrix_ac_11_cast_fp16 = matmul(transpose_x = matrix_ac_11_transpose_x_0, transpose_y = matrix_ac_11_transpose_y_0, x = transpose_61, y = transpose_62)[name = tensor("matrix_ac_11_cast_fp16")]; - tensor matrix_bd_23_begin_0 = const()[name = tensor("matrix_bd_23_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_23_end_0 = const()[name = tensor("matrix_bd_23_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_23_end_mask_0 = const()[name = tensor("matrix_bd_23_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_23_cast_fp16 = slice_by_index(begin = matrix_bd_23_begin_0, end = matrix_bd_23_end_0, end_mask = matrix_bd_23_end_mask_0, x = matrix_bd_21_cast_fp16)[name = tensor("matrix_bd_23_cast_fp16")]; - tensor var_1288_cast_fp16 = add(x = matrix_ac_11_cast_fp16, y = matrix_bd_23_cast_fp16)[name = tensor("op_1288_cast_fp16")]; - tensor _inversed_scores_21_y_0_to_fp16 = const()[name = tensor("_inversed_scores_21_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_21_cast_fp16 = mul(x = var_1288_cast_fp16, y = _inversed_scores_21_y_0_to_fp16)[name = tensor("_inversed_scores_21_cast_fp16")]; - tensor scores_23_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_21_cast_fp16, cond = mask_11)[name = tensor("scores_23_cast_fp16")]; - tensor var_1294_cast_fp16 = softmax(axis = var_23, x = scores_23_cast_fp16)[name = tensor("op_1294_cast_fp16")]; - tensor input_293_cast_fp16 = select(a = var_11_to_fp16, b = var_1294_cast_fp16, cond = mask_11)[name = tensor("input_293_cast_fp16")]; - tensor x_123_transpose_x_0 = const()[name = tensor("x_123_transpose_x_0"), val = tensor(false)]; - tensor x_123_transpose_y_0 = const()[name = tensor("x_123_transpose_y_0"), val = tensor(false)]; - tensor value_13_cast_fp16 = transpose(perm = value_13_perm_0, x = v_11_cast_fp16)[name = tensor("transpose_169")]; - tensor x_123_cast_fp16 = matmul(transpose_x = x_123_transpose_x_0, transpose_y = x_123_transpose_y_0, x = input_293_cast_fp16, y = value_13_cast_fp16)[name = tensor("x_123_cast_fp16")]; - tensor var_1298_perm_0 = const()[name = tensor("op_1298_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_1299 = const()[name = tensor("op_1299"), val = tensor([1, -1, 512])]; - tensor var_1298_cast_fp16 = transpose(perm = var_1298_perm_0, x = x_123_cast_fp16)[name = tensor("transpose_165")]; - tensor input_295_cast_fp16 = reshape(shape = var_1299, x = var_1298_cast_fp16)[name = tensor("input_295_cast_fp16")]; - tensor module_layers_5_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_5_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71492416)))]; - tensor module_layers_5_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_5_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(72016768)))]; - tensor linear_52_cast_fp16 = linear(bias = module_layers_5_self_attn_linear_out_bias_to_fp16, weight = module_layers_5_self_attn_linear_out_weight_to_fp16, x = input_295_cast_fp16)[name = tensor("linear_52_cast_fp16")]; - tensor input_299_cast_fp16 = add(x = input_291_cast_fp16, y = linear_52_cast_fp16)[name = tensor("input_299_cast_fp16")]; - tensor x_127_axes_0 = const()[name = tensor("x_127_axes_0"), val = tensor([-1])]; - tensor module_layers_5_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_5_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(72017856)))]; - tensor module_layers_5_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_5_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(72018944)))]; - tensor x_127_cast_fp16 = layer_norm(axes = x_127_axes_0, beta = module_layers_5_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_5_norm_conv_weight_to_fp16, x = input_299_cast_fp16)[name = tensor("x_127_cast_fp16")]; - tensor input_301_perm_0 = const()[name = tensor("input_301_perm_0"), val = tensor([0, 2, 1])]; - tensor input_303_pad_type_0 = const()[name = tensor("input_303_pad_type_0"), val = tensor("valid")]; - tensor input_303_strides_0 = const()[name = tensor("input_303_strides_0"), val = tensor([1])]; - tensor input_303_pad_0 = const()[name = tensor("input_303_pad_0"), val = tensor([0, 0])]; - tensor input_303_dilations_0 = const()[name = tensor("input_303_dilations_0"), val = tensor([1])]; - tensor input_303_groups_0 = const()[name = tensor("input_303_groups_0"), val = tensor(1)]; - tensor module_layers_5_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_5_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(72020032)))]; - tensor module_layers_5_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_5_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73068672)))]; - tensor input_301_cast_fp16 = transpose(perm = input_301_perm_0, x = x_127_cast_fp16)[name = tensor("transpose_164")]; - tensor input_303_cast_fp16 = conv(bias = module_layers_5_conv_pointwise_conv1_bias_to_fp16, dilations = input_303_dilations_0, groups = input_303_groups_0, pad = input_303_pad_0, pad_type = input_303_pad_type_0, strides = input_303_strides_0, weight = module_layers_5_conv_pointwise_conv1_weight_to_fp16, x = input_301_cast_fp16)[name = tensor("input_303_cast_fp16")]; - tensor x_129_split_num_splits_0 = const()[name = tensor("x_129_split_num_splits_0"), val = tensor(2)]; - tensor x_129_split_axis_0 = const()[name = tensor("x_129_split_axis_0"), val = tensor(1)]; - tensor x_129_split_cast_fp16_0, tensor x_129_split_cast_fp16_1 = split(axis = x_129_split_axis_0, num_splits = x_129_split_num_splits_0, x = input_303_cast_fp16)[name = tensor("x_129_split_cast_fp16")]; - tensor x_129_split_1_sigmoid_cast_fp16 = sigmoid(x = x_129_split_cast_fp16_1)[name = tensor("x_129_split_1_sigmoid_cast_fp16")]; - tensor x_129_cast_fp16 = mul(x = x_129_split_cast_fp16_0, y = x_129_split_1_sigmoid_cast_fp16)[name = tensor("x_129_cast_fp16")]; - tensor input_305_cast_fp16 = select(a = var_11_to_fp16, b = x_129_cast_fp16, cond = var_453)[name = tensor("input_305_cast_fp16")]; - tensor input_307_pad_0 = const()[name = tensor("input_307_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_307_mode_0 = const()[name = tensor("input_307_mode_0"), val = tensor("constant")]; - tensor const_123_to_fp16 = const()[name = tensor("const_123_to_fp16"), val = tensor(0x0p+0)]; - tensor input_307_cast_fp16 = pad(constant_val = const_123_to_fp16, mode = input_307_mode_0, pad = input_307_pad_0, x = input_305_cast_fp16)[name = tensor("input_307_cast_fp16")]; - tensor input_309_pad_type_0 = const()[name = tensor("input_309_pad_type_0"), val = tensor("valid")]; - tensor input_309_groups_0 = const()[name = tensor("input_309_groups_0"), val = tensor(512)]; - tensor input_309_strides_0 = const()[name = tensor("input_309_strides_0"), val = tensor([1])]; - tensor input_309_pad_0 = const()[name = tensor("input_309_pad_0"), val = tensor([0, 0])]; - tensor input_309_dilations_0 = const()[name = tensor("input_309_dilations_0"), val = tensor([1])]; - tensor const_244_to_fp16 = const()[name = tensor("const_244_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73070784)))]; - tensor const_245_to_fp16 = const()[name = tensor("const_245_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73080064)))]; - tensor input_311_cast_fp16 = conv(bias = const_245_to_fp16, dilations = input_309_dilations_0, groups = input_309_groups_0, pad = input_309_pad_0, pad_type = input_309_pad_type_0, strides = input_309_strides_0, weight = const_244_to_fp16, x = input_307_cast_fp16)[name = tensor("input_311_cast_fp16")]; - tensor input_313_cast_fp16 = silu(x = input_311_cast_fp16)[name = tensor("input_313_cast_fp16")]; - tensor x_131_pad_type_0 = const()[name = tensor("x_131_pad_type_0"), val = tensor("valid")]; - tensor x_131_strides_0 = const()[name = tensor("x_131_strides_0"), val = tensor([1])]; - tensor x_131_pad_0 = const()[name = tensor("x_131_pad_0"), val = tensor([0, 0])]; - tensor x_131_dilations_0 = const()[name = tensor("x_131_dilations_0"), val = tensor([1])]; - tensor x_131_groups_0 = const()[name = tensor("x_131_groups_0"), val = tensor(1)]; - tensor module_layers_5_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_5_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73081152)))]; - tensor module_layers_5_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_5_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73605504)))]; - tensor x_131_cast_fp16 = conv(bias = module_layers_5_conv_pointwise_conv2_bias_to_fp16, dilations = x_131_dilations_0, groups = x_131_groups_0, pad = x_131_pad_0, pad_type = x_131_pad_type_0, strides = x_131_strides_0, weight = module_layers_5_conv_pointwise_conv2_weight_to_fp16, x = input_313_cast_fp16)[name = tensor("x_131_cast_fp16")]; - tensor input_315_perm_0 = const()[name = tensor("input_315_perm_0"), val = tensor([0, 2, 1])]; - tensor input_315_cast_fp16 = transpose(perm = input_315_perm_0, x = x_131_cast_fp16)[name = tensor("transpose_163")]; - tensor input_317_cast_fp16 = add(x = input_299_cast_fp16, y = input_315_cast_fp16)[name = tensor("input_317_cast_fp16")]; - tensor input_319_axes_0 = const()[name = tensor("input_319_axes_0"), val = tensor([-1])]; - tensor module_layers_5_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_5_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73606592)))]; - tensor module_layers_5_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_5_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73607680)))]; - tensor input_319_cast_fp16 = layer_norm(axes = input_319_axes_0, beta = module_layers_5_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_5_norm_feed_forward2_weight_to_fp16, x = input_317_cast_fp16)[name = tensor("input_319_cast_fp16")]; - tensor module_layers_5_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_5_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73608768)))]; - tensor module_layers_5_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_5_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(75705984)))]; - tensor linear_53_cast_fp16 = linear(bias = module_layers_5_feed_forward2_linear1_bias_to_fp16, weight = module_layers_5_feed_forward2_linear1_weight_to_fp16, x = input_319_cast_fp16)[name = tensor("linear_53_cast_fp16")]; - tensor input_323_cast_fp16 = silu(x = linear_53_cast_fp16)[name = tensor("input_323_cast_fp16")]; - tensor module_layers_5_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_5_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(75710144)))]; - tensor module_layers_5_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_5_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(77807360)))]; - tensor linear_54_cast_fp16 = linear(bias = module_layers_5_feed_forward2_linear2_bias_to_fp16, weight = module_layers_5_feed_forward2_linear2_weight_to_fp16, x = input_323_cast_fp16)[name = tensor("linear_54_cast_fp16")]; - tensor var_1365_to_fp16 = const()[name = tensor("op_1365_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1366_cast_fp16 = mul(x = linear_54_cast_fp16, y = var_1365_to_fp16)[name = tensor("op_1366_cast_fp16")]; - tensor input_329_cast_fp16 = add(x = input_317_cast_fp16, y = var_1366_cast_fp16)[name = tensor("input_329_cast_fp16")]; - tensor input_331_axes_0 = const()[name = tensor("input_331_axes_0"), val = tensor([-1])]; - tensor module_layers_5_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_5_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(77808448)))]; - tensor module_layers_5_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_5_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(77809536)))]; - tensor input_331_cast_fp16 = layer_norm(axes = input_331_axes_0, beta = module_layers_5_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_5_norm_out_weight_to_fp16, x = input_329_cast_fp16)[name = tensor("input_331_cast_fp16")]; - tensor input_333_axes_0 = const()[name = tensor("input_333_axes_0"), val = tensor([-1])]; - tensor module_layers_6_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_6_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(77810624)))]; - tensor module_layers_6_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_6_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(77811712)))]; - tensor input_333_cast_fp16 = layer_norm(axes = input_333_axes_0, beta = module_layers_6_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_6_norm_feed_forward1_weight_to_fp16, x = input_331_cast_fp16)[name = tensor("input_333_cast_fp16")]; - tensor module_layers_6_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_6_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(77812800)))]; - tensor module_layers_6_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_6_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(79910016)))]; - tensor linear_55_cast_fp16 = linear(bias = module_layers_6_feed_forward1_linear1_bias_to_fp16, weight = module_layers_6_feed_forward1_linear1_weight_to_fp16, x = input_333_cast_fp16)[name = tensor("linear_55_cast_fp16")]; - tensor input_337_cast_fp16 = silu(x = linear_55_cast_fp16)[name = tensor("input_337_cast_fp16")]; - tensor module_layers_6_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_6_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(79914176)))]; - tensor module_layers_6_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_6_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82011392)))]; - tensor linear_56_cast_fp16 = linear(bias = module_layers_6_feed_forward1_linear2_bias_to_fp16, weight = module_layers_6_feed_forward1_linear2_weight_to_fp16, x = input_337_cast_fp16)[name = tensor("linear_56_cast_fp16")]; - tensor var_1396_to_fp16 = const()[name = tensor("op_1396_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1397_cast_fp16 = mul(x = linear_56_cast_fp16, y = var_1396_to_fp16)[name = tensor("op_1397_cast_fp16")]; - tensor input_343_cast_fp16 = add(x = input_331_cast_fp16, y = var_1397_cast_fp16)[name = tensor("input_343_cast_fp16")]; - tensor query_13_axes_0 = const()[name = tensor("query_13_axes_0"), val = tensor([-1])]; - tensor module_layers_6_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_6_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82012480)))]; - tensor module_layers_6_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_6_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82013568)))]; - tensor query_13_cast_fp16 = layer_norm(axes = query_13_axes_0, beta = module_layers_6_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_6_norm_self_att_weight_to_fp16, x = input_343_cast_fp16)[name = tensor("query_13_cast_fp16")]; - tensor module_layers_6_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_6_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82014656)))]; - tensor module_layers_6_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_6_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82539008)))]; - tensor linear_57_cast_fp16 = linear(bias = module_layers_6_self_attn_linear_q_bias_to_fp16, weight = module_layers_6_self_attn_linear_q_weight_to_fp16, x = query_13_cast_fp16)[name = tensor("linear_57_cast_fp16")]; - tensor var_1414 = const()[name = tensor("op_1414"), val = tensor([1, -1, 8, 64])]; - tensor q_37_cast_fp16 = reshape(shape = var_1414, x = linear_57_cast_fp16)[name = tensor("q_37_cast_fp16")]; - tensor module_layers_6_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_6_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82540096)))]; - tensor module_layers_6_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_6_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83064448)))]; - tensor linear_58_cast_fp16 = linear(bias = module_layers_6_self_attn_linear_k_bias_to_fp16, weight = module_layers_6_self_attn_linear_k_weight_to_fp16, x = query_13_cast_fp16)[name = tensor("linear_58_cast_fp16")]; - tensor var_1419 = const()[name = tensor("op_1419"), val = tensor([1, -1, 8, 64])]; - tensor k_25_cast_fp16 = reshape(shape = var_1419, x = linear_58_cast_fp16)[name = tensor("k_25_cast_fp16")]; - tensor module_layers_6_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_6_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83065536)))]; - tensor module_layers_6_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_6_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83589888)))]; - tensor linear_59_cast_fp16 = linear(bias = module_layers_6_self_attn_linear_v_bias_to_fp16, weight = module_layers_6_self_attn_linear_v_weight_to_fp16, x = query_13_cast_fp16)[name = tensor("linear_59_cast_fp16")]; - tensor var_1424 = const()[name = tensor("op_1424"), val = tensor([1, -1, 8, 64])]; - tensor v_13_cast_fp16 = reshape(shape = var_1424, x = linear_59_cast_fp16)[name = tensor("v_13_cast_fp16")]; - tensor value_15_perm_0 = const()[name = tensor("value_15_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_6_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_6_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83590976)))]; - tensor var_1436_cast_fp16 = add(x = q_37_cast_fp16, y = module_layers_6_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1436_cast_fp16")]; - tensor module_layers_6_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_6_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83592064)))]; - tensor var_1438_cast_fp16 = add(x = q_37_cast_fp16, y = module_layers_6_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1438_cast_fp16")]; - tensor q_with_bias_v_13_perm_0 = const()[name = tensor("q_with_bias_v_13_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_139_transpose_x_0 = const()[name = tensor("x_139_transpose_x_0"), val = tensor(false)]; - tensor x_139_transpose_y_0 = const()[name = tensor("x_139_transpose_y_0"), val = tensor(false)]; - tensor var_1440_to_fp16 = const()[name = tensor("op_1440_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83593152)))]; - tensor q_with_bias_v_13_cast_fp16 = transpose(perm = q_with_bias_v_13_perm_0, x = var_1438_cast_fp16)[name = tensor("transpose_161")]; - tensor x_139_cast_fp16 = matmul(transpose_x = x_139_transpose_x_0, transpose_y = x_139_transpose_y_0, x = q_with_bias_v_13_cast_fp16, y = var_1440_to_fp16)[name = tensor("x_139_cast_fp16")]; - tensor x_141_pad_0 = const()[name = tensor("x_141_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_141_mode_0 = const()[name = tensor("x_141_mode_0"), val = tensor("constant")]; - tensor const_130_to_fp16 = const()[name = tensor("const_130_to_fp16"), val = tensor(0x0p+0)]; - tensor x_141_cast_fp16 = pad(constant_val = const_130_to_fp16, mode = x_141_mode_0, pad = x_141_pad_0, x = x_139_cast_fp16)[name = tensor("x_141_cast_fp16")]; - tensor var_1448 = const()[name = tensor("op_1448"), val = tensor([1, 8, -1, 188])]; - tensor x_143_cast_fp16 = reshape(shape = var_1448, x = x_141_cast_fp16)[name = tensor("x_143_cast_fp16")]; - tensor var_1452_begin_0 = const()[name = tensor("op_1452_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_1452_end_0 = const()[name = tensor("op_1452_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_1452_end_mask_0 = const()[name = tensor("op_1452_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_1452_cast_fp16 = slice_by_index(begin = var_1452_begin_0, end = var_1452_end_0, end_mask = var_1452_end_mask_0, x = x_143_cast_fp16)[name = tensor("op_1452_cast_fp16")]; - tensor var_1453 = const()[name = tensor("op_1453"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_25_cast_fp16 = reshape(shape = var_1453, x = var_1452_cast_fp16)[name = tensor("matrix_bd_25_cast_fp16")]; - tensor matrix_ac_13_transpose_x_0 = const()[name = tensor("matrix_ac_13_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_13_transpose_y_0 = const()[name = tensor("matrix_ac_13_transpose_y_0"), val = tensor(false)]; - tensor transpose_63_perm_0 = const()[name = tensor("transpose_63_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_64_perm_0 = const()[name = tensor("transpose_64_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_64 = transpose(perm = transpose_64_perm_0, x = k_25_cast_fp16)[name = tensor("transpose_159")]; - tensor transpose_63 = transpose(perm = transpose_63_perm_0, x = var_1436_cast_fp16)[name = tensor("transpose_160")]; - tensor matrix_ac_13_cast_fp16 = matmul(transpose_x = matrix_ac_13_transpose_x_0, transpose_y = matrix_ac_13_transpose_y_0, x = transpose_63, y = transpose_64)[name = tensor("matrix_ac_13_cast_fp16")]; - tensor matrix_bd_27_begin_0 = const()[name = tensor("matrix_bd_27_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_27_end_0 = const()[name = tensor("matrix_bd_27_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_27_end_mask_0 = const()[name = tensor("matrix_bd_27_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_27_cast_fp16 = slice_by_index(begin = matrix_bd_27_begin_0, end = matrix_bd_27_end_0, end_mask = matrix_bd_27_end_mask_0, x = matrix_bd_25_cast_fp16)[name = tensor("matrix_bd_27_cast_fp16")]; - tensor var_1462_cast_fp16 = add(x = matrix_ac_13_cast_fp16, y = matrix_bd_27_cast_fp16)[name = tensor("op_1462_cast_fp16")]; - tensor _inversed_scores_25_y_0_to_fp16 = const()[name = tensor("_inversed_scores_25_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_25_cast_fp16 = mul(x = var_1462_cast_fp16, y = _inversed_scores_25_y_0_to_fp16)[name = tensor("_inversed_scores_25_cast_fp16")]; - tensor scores_27_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_25_cast_fp16, cond = mask_11)[name = tensor("scores_27_cast_fp16")]; - tensor var_1468_cast_fp16 = softmax(axis = var_23, x = scores_27_cast_fp16)[name = tensor("op_1468_cast_fp16")]; - tensor input_345_cast_fp16 = select(a = var_11_to_fp16, b = var_1468_cast_fp16, cond = mask_11)[name = tensor("input_345_cast_fp16")]; - tensor x_145_transpose_x_0 = const()[name = tensor("x_145_transpose_x_0"), val = tensor(false)]; - tensor x_145_transpose_y_0 = const()[name = tensor("x_145_transpose_y_0"), val = tensor(false)]; - tensor value_15_cast_fp16 = transpose(perm = value_15_perm_0, x = v_13_cast_fp16)[name = tensor("transpose_162")]; - tensor x_145_cast_fp16 = matmul(transpose_x = x_145_transpose_x_0, transpose_y = x_145_transpose_y_0, x = input_345_cast_fp16, y = value_15_cast_fp16)[name = tensor("x_145_cast_fp16")]; - tensor var_1472_perm_0 = const()[name = tensor("op_1472_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_1473 = const()[name = tensor("op_1473"), val = tensor([1, -1, 512])]; - tensor var_1472_cast_fp16 = transpose(perm = var_1472_perm_0, x = x_145_cast_fp16)[name = tensor("transpose_158")]; - tensor input_347_cast_fp16 = reshape(shape = var_1473, x = var_1472_cast_fp16)[name = tensor("input_347_cast_fp16")]; - tensor module_layers_6_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_6_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83977216)))]; - tensor module_layers_6_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_6_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84501568)))]; - tensor linear_61_cast_fp16 = linear(bias = module_layers_6_self_attn_linear_out_bias_to_fp16, weight = module_layers_6_self_attn_linear_out_weight_to_fp16, x = input_347_cast_fp16)[name = tensor("linear_61_cast_fp16")]; - tensor input_351_cast_fp16 = add(x = input_343_cast_fp16, y = linear_61_cast_fp16)[name = tensor("input_351_cast_fp16")]; - tensor x_149_axes_0 = const()[name = tensor("x_149_axes_0"), val = tensor([-1])]; - tensor module_layers_6_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_6_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84502656)))]; - tensor module_layers_6_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_6_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84503744)))]; - tensor x_149_cast_fp16 = layer_norm(axes = x_149_axes_0, beta = module_layers_6_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_6_norm_conv_weight_to_fp16, x = input_351_cast_fp16)[name = tensor("x_149_cast_fp16")]; - tensor input_353_perm_0 = const()[name = tensor("input_353_perm_0"), val = tensor([0, 2, 1])]; - tensor input_355_pad_type_0 = const()[name = tensor("input_355_pad_type_0"), val = tensor("valid")]; - tensor input_355_strides_0 = const()[name = tensor("input_355_strides_0"), val = tensor([1])]; - tensor input_355_pad_0 = const()[name = tensor("input_355_pad_0"), val = tensor([0, 0])]; - tensor input_355_dilations_0 = const()[name = tensor("input_355_dilations_0"), val = tensor([1])]; - tensor input_355_groups_0 = const()[name = tensor("input_355_groups_0"), val = tensor(1)]; - tensor module_layers_6_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_6_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84504832)))]; - tensor module_layers_6_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_6_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85553472)))]; - tensor input_353_cast_fp16 = transpose(perm = input_353_perm_0, x = x_149_cast_fp16)[name = tensor("transpose_157")]; - tensor input_355_cast_fp16 = conv(bias = module_layers_6_conv_pointwise_conv1_bias_to_fp16, dilations = input_355_dilations_0, groups = input_355_groups_0, pad = input_355_pad_0, pad_type = input_355_pad_type_0, strides = input_355_strides_0, weight = module_layers_6_conv_pointwise_conv1_weight_to_fp16, x = input_353_cast_fp16)[name = tensor("input_355_cast_fp16")]; - tensor x_151_split_num_splits_0 = const()[name = tensor("x_151_split_num_splits_0"), val = tensor(2)]; - tensor x_151_split_axis_0 = const()[name = tensor("x_151_split_axis_0"), val = tensor(1)]; - tensor x_151_split_cast_fp16_0, tensor x_151_split_cast_fp16_1 = split(axis = x_151_split_axis_0, num_splits = x_151_split_num_splits_0, x = input_355_cast_fp16)[name = tensor("x_151_split_cast_fp16")]; - tensor x_151_split_1_sigmoid_cast_fp16 = sigmoid(x = x_151_split_cast_fp16_1)[name = tensor("x_151_split_1_sigmoid_cast_fp16")]; - tensor x_151_cast_fp16 = mul(x = x_151_split_cast_fp16_0, y = x_151_split_1_sigmoid_cast_fp16)[name = tensor("x_151_cast_fp16")]; - tensor input_357_cast_fp16 = select(a = var_11_to_fp16, b = x_151_cast_fp16, cond = var_453)[name = tensor("input_357_cast_fp16")]; - tensor input_359_pad_0 = const()[name = tensor("input_359_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_359_mode_0 = const()[name = tensor("input_359_mode_0"), val = tensor("constant")]; - tensor const_133_to_fp16 = const()[name = tensor("const_133_to_fp16"), val = tensor(0x0p+0)]; - tensor input_359_cast_fp16 = pad(constant_val = const_133_to_fp16, mode = input_359_mode_0, pad = input_359_pad_0, x = input_357_cast_fp16)[name = tensor("input_359_cast_fp16")]; - tensor input_361_pad_type_0 = const()[name = tensor("input_361_pad_type_0"), val = tensor("valid")]; - tensor input_361_groups_0 = const()[name = tensor("input_361_groups_0"), val = tensor(512)]; - tensor input_361_strides_0 = const()[name = tensor("input_361_strides_0"), val = tensor([1])]; - tensor input_361_pad_0 = const()[name = tensor("input_361_pad_0"), val = tensor([0, 0])]; - tensor input_361_dilations_0 = const()[name = tensor("input_361_dilations_0"), val = tensor([1])]; - tensor const_246_to_fp16 = const()[name = tensor("const_246_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85555584)))]; - tensor const_247_to_fp16 = const()[name = tensor("const_247_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85564864)))]; - tensor input_363_cast_fp16 = conv(bias = const_247_to_fp16, dilations = input_361_dilations_0, groups = input_361_groups_0, pad = input_361_pad_0, pad_type = input_361_pad_type_0, strides = input_361_strides_0, weight = const_246_to_fp16, x = input_359_cast_fp16)[name = tensor("input_363_cast_fp16")]; - tensor input_365_cast_fp16 = silu(x = input_363_cast_fp16)[name = tensor("input_365_cast_fp16")]; - tensor x_153_pad_type_0 = const()[name = tensor("x_153_pad_type_0"), val = tensor("valid")]; - tensor x_153_strides_0 = const()[name = tensor("x_153_strides_0"), val = tensor([1])]; - tensor x_153_pad_0 = const()[name = tensor("x_153_pad_0"), val = tensor([0, 0])]; - tensor x_153_dilations_0 = const()[name = tensor("x_153_dilations_0"), val = tensor([1])]; - tensor x_153_groups_0 = const()[name = tensor("x_153_groups_0"), val = tensor(1)]; - tensor module_layers_6_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_6_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85565952)))]; - tensor module_layers_6_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_6_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86090304)))]; - tensor x_153_cast_fp16 = conv(bias = module_layers_6_conv_pointwise_conv2_bias_to_fp16, dilations = x_153_dilations_0, groups = x_153_groups_0, pad = x_153_pad_0, pad_type = x_153_pad_type_0, strides = x_153_strides_0, weight = module_layers_6_conv_pointwise_conv2_weight_to_fp16, x = input_365_cast_fp16)[name = tensor("x_153_cast_fp16")]; - tensor input_367_perm_0 = const()[name = tensor("input_367_perm_0"), val = tensor([0, 2, 1])]; - tensor input_367_cast_fp16 = transpose(perm = input_367_perm_0, x = x_153_cast_fp16)[name = tensor("transpose_156")]; - tensor input_369_cast_fp16 = add(x = input_351_cast_fp16, y = input_367_cast_fp16)[name = tensor("input_369_cast_fp16")]; - tensor input_371_axes_0 = const()[name = tensor("input_371_axes_0"), val = tensor([-1])]; - tensor module_layers_6_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_6_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86091392)))]; - tensor module_layers_6_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_6_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86092480)))]; - tensor input_371_cast_fp16 = layer_norm(axes = input_371_axes_0, beta = module_layers_6_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_6_norm_feed_forward2_weight_to_fp16, x = input_369_cast_fp16)[name = tensor("input_371_cast_fp16")]; - tensor module_layers_6_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_6_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86093568)))]; - tensor module_layers_6_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_6_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(88190784)))]; - tensor linear_62_cast_fp16 = linear(bias = module_layers_6_feed_forward2_linear1_bias_to_fp16, weight = module_layers_6_feed_forward2_linear1_weight_to_fp16, x = input_371_cast_fp16)[name = tensor("linear_62_cast_fp16")]; - tensor input_375_cast_fp16 = silu(x = linear_62_cast_fp16)[name = tensor("input_375_cast_fp16")]; - tensor module_layers_6_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_6_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(88194944)))]; - tensor module_layers_6_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_6_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90292160)))]; - tensor linear_63_cast_fp16 = linear(bias = module_layers_6_feed_forward2_linear2_bias_to_fp16, weight = module_layers_6_feed_forward2_linear2_weight_to_fp16, x = input_375_cast_fp16)[name = tensor("linear_63_cast_fp16")]; - tensor var_1539_to_fp16 = const()[name = tensor("op_1539_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1540_cast_fp16 = mul(x = linear_63_cast_fp16, y = var_1539_to_fp16)[name = tensor("op_1540_cast_fp16")]; - tensor input_381_cast_fp16 = add(x = input_369_cast_fp16, y = var_1540_cast_fp16)[name = tensor("input_381_cast_fp16")]; - tensor input_383_axes_0 = const()[name = tensor("input_383_axes_0"), val = tensor([-1])]; - tensor module_layers_6_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_6_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90293248)))]; - tensor module_layers_6_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_6_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90294336)))]; - tensor input_383_cast_fp16 = layer_norm(axes = input_383_axes_0, beta = module_layers_6_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_6_norm_out_weight_to_fp16, x = input_381_cast_fp16)[name = tensor("input_383_cast_fp16")]; - tensor input_385_axes_0 = const()[name = tensor("input_385_axes_0"), val = tensor([-1])]; - tensor module_layers_7_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_7_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90295424)))]; - tensor module_layers_7_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_7_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90296512)))]; - tensor input_385_cast_fp16 = layer_norm(axes = input_385_axes_0, beta = module_layers_7_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_7_norm_feed_forward1_weight_to_fp16, x = input_383_cast_fp16)[name = tensor("input_385_cast_fp16")]; - tensor module_layers_7_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_7_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90297600)))]; - tensor module_layers_7_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_7_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92394816)))]; - tensor linear_64_cast_fp16 = linear(bias = module_layers_7_feed_forward1_linear1_bias_to_fp16, weight = module_layers_7_feed_forward1_linear1_weight_to_fp16, x = input_385_cast_fp16)[name = tensor("linear_64_cast_fp16")]; - tensor input_389_cast_fp16 = silu(x = linear_64_cast_fp16)[name = tensor("input_389_cast_fp16")]; - tensor module_layers_7_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_7_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92398976)))]; - tensor module_layers_7_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_7_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(94496192)))]; - tensor linear_65_cast_fp16 = linear(bias = module_layers_7_feed_forward1_linear2_bias_to_fp16, weight = module_layers_7_feed_forward1_linear2_weight_to_fp16, x = input_389_cast_fp16)[name = tensor("linear_65_cast_fp16")]; - tensor var_1570_to_fp16 = const()[name = tensor("op_1570_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1571_cast_fp16 = mul(x = linear_65_cast_fp16, y = var_1570_to_fp16)[name = tensor("op_1571_cast_fp16")]; - tensor input_395_cast_fp16 = add(x = input_383_cast_fp16, y = var_1571_cast_fp16)[name = tensor("input_395_cast_fp16")]; - tensor query_15_axes_0 = const()[name = tensor("query_15_axes_0"), val = tensor([-1])]; - tensor module_layers_7_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_7_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(94497280)))]; - tensor module_layers_7_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_7_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(94498368)))]; - tensor query_15_cast_fp16 = layer_norm(axes = query_15_axes_0, beta = module_layers_7_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_7_norm_self_att_weight_to_fp16, x = input_395_cast_fp16)[name = tensor("query_15_cast_fp16")]; - tensor module_layers_7_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_7_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(94499456)))]; - tensor module_layers_7_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_7_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95023808)))]; - tensor linear_66_cast_fp16 = linear(bias = module_layers_7_self_attn_linear_q_bias_to_fp16, weight = module_layers_7_self_attn_linear_q_weight_to_fp16, x = query_15_cast_fp16)[name = tensor("linear_66_cast_fp16")]; - tensor var_1588 = const()[name = tensor("op_1588"), val = tensor([1, -1, 8, 64])]; - tensor q_43_cast_fp16 = reshape(shape = var_1588, x = linear_66_cast_fp16)[name = tensor("q_43_cast_fp16")]; - tensor module_layers_7_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_7_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95024896)))]; - tensor module_layers_7_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_7_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95549248)))]; - tensor linear_67_cast_fp16 = linear(bias = module_layers_7_self_attn_linear_k_bias_to_fp16, weight = module_layers_7_self_attn_linear_k_weight_to_fp16, x = query_15_cast_fp16)[name = tensor("linear_67_cast_fp16")]; - tensor var_1593 = const()[name = tensor("op_1593"), val = tensor([1, -1, 8, 64])]; - tensor k_29_cast_fp16 = reshape(shape = var_1593, x = linear_67_cast_fp16)[name = tensor("k_29_cast_fp16")]; - tensor module_layers_7_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_7_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95550336)))]; - tensor module_layers_7_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_7_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96074688)))]; - tensor linear_68_cast_fp16 = linear(bias = module_layers_7_self_attn_linear_v_bias_to_fp16, weight = module_layers_7_self_attn_linear_v_weight_to_fp16, x = query_15_cast_fp16)[name = tensor("linear_68_cast_fp16")]; - tensor var_1598 = const()[name = tensor("op_1598"), val = tensor([1, -1, 8, 64])]; - tensor v_15_cast_fp16 = reshape(shape = var_1598, x = linear_68_cast_fp16)[name = tensor("v_15_cast_fp16")]; - tensor value_17_perm_0 = const()[name = tensor("value_17_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_7_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_7_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96075776)))]; - tensor var_1610_cast_fp16 = add(x = q_43_cast_fp16, y = module_layers_7_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1610_cast_fp16")]; - tensor module_layers_7_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_7_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96076864)))]; - tensor var_1612_cast_fp16 = add(x = q_43_cast_fp16, y = module_layers_7_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1612_cast_fp16")]; - tensor q_with_bias_v_15_perm_0 = const()[name = tensor("q_with_bias_v_15_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_161_transpose_x_0 = const()[name = tensor("x_161_transpose_x_0"), val = tensor(false)]; - tensor x_161_transpose_y_0 = const()[name = tensor("x_161_transpose_y_0"), val = tensor(false)]; - tensor var_1614_to_fp16 = const()[name = tensor("op_1614_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96077952)))]; - tensor q_with_bias_v_15_cast_fp16 = transpose(perm = q_with_bias_v_15_perm_0, x = var_1612_cast_fp16)[name = tensor("transpose_154")]; - tensor x_161_cast_fp16 = matmul(transpose_x = x_161_transpose_x_0, transpose_y = x_161_transpose_y_0, x = q_with_bias_v_15_cast_fp16, y = var_1614_to_fp16)[name = tensor("x_161_cast_fp16")]; - tensor x_163_pad_0 = const()[name = tensor("x_163_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_163_mode_0 = const()[name = tensor("x_163_mode_0"), val = tensor("constant")]; - tensor const_140_to_fp16 = const()[name = tensor("const_140_to_fp16"), val = tensor(0x0p+0)]; - tensor x_163_cast_fp16 = pad(constant_val = const_140_to_fp16, mode = x_163_mode_0, pad = x_163_pad_0, x = x_161_cast_fp16)[name = tensor("x_163_cast_fp16")]; - tensor var_1622 = const()[name = tensor("op_1622"), val = tensor([1, 8, -1, 188])]; - tensor x_165_cast_fp16 = reshape(shape = var_1622, x = x_163_cast_fp16)[name = tensor("x_165_cast_fp16")]; - tensor var_1626_begin_0 = const()[name = tensor("op_1626_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_1626_end_0 = const()[name = tensor("op_1626_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_1626_end_mask_0 = const()[name = tensor("op_1626_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_1626_cast_fp16 = slice_by_index(begin = var_1626_begin_0, end = var_1626_end_0, end_mask = var_1626_end_mask_0, x = x_165_cast_fp16)[name = tensor("op_1626_cast_fp16")]; - tensor var_1627 = const()[name = tensor("op_1627"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_29_cast_fp16 = reshape(shape = var_1627, x = var_1626_cast_fp16)[name = tensor("matrix_bd_29_cast_fp16")]; - tensor matrix_ac_15_transpose_x_0 = const()[name = tensor("matrix_ac_15_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_15_transpose_y_0 = const()[name = tensor("matrix_ac_15_transpose_y_0"), val = tensor(false)]; - tensor transpose_65_perm_0 = const()[name = tensor("transpose_65_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_66_perm_0 = const()[name = tensor("transpose_66_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_66 = transpose(perm = transpose_66_perm_0, x = k_29_cast_fp16)[name = tensor("transpose_152")]; - tensor transpose_65 = transpose(perm = transpose_65_perm_0, x = var_1610_cast_fp16)[name = tensor("transpose_153")]; - tensor matrix_ac_15_cast_fp16 = matmul(transpose_x = matrix_ac_15_transpose_x_0, transpose_y = matrix_ac_15_transpose_y_0, x = transpose_65, y = transpose_66)[name = tensor("matrix_ac_15_cast_fp16")]; - tensor matrix_bd_31_begin_0 = const()[name = tensor("matrix_bd_31_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_31_end_0 = const()[name = tensor("matrix_bd_31_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_31_end_mask_0 = const()[name = tensor("matrix_bd_31_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_31_cast_fp16 = slice_by_index(begin = matrix_bd_31_begin_0, end = matrix_bd_31_end_0, end_mask = matrix_bd_31_end_mask_0, x = matrix_bd_29_cast_fp16)[name = tensor("matrix_bd_31_cast_fp16")]; - tensor var_1636_cast_fp16 = add(x = matrix_ac_15_cast_fp16, y = matrix_bd_31_cast_fp16)[name = tensor("op_1636_cast_fp16")]; - tensor _inversed_scores_29_y_0_to_fp16 = const()[name = tensor("_inversed_scores_29_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_29_cast_fp16 = mul(x = var_1636_cast_fp16, y = _inversed_scores_29_y_0_to_fp16)[name = tensor("_inversed_scores_29_cast_fp16")]; - tensor scores_31_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_29_cast_fp16, cond = mask_11)[name = tensor("scores_31_cast_fp16")]; - tensor var_1642_cast_fp16 = softmax(axis = var_23, x = scores_31_cast_fp16)[name = tensor("op_1642_cast_fp16")]; - tensor input_397_cast_fp16 = select(a = var_11_to_fp16, b = var_1642_cast_fp16, cond = mask_11)[name = tensor("input_397_cast_fp16")]; - tensor x_167_transpose_x_0 = const()[name = tensor("x_167_transpose_x_0"), val = tensor(false)]; - tensor x_167_transpose_y_0 = const()[name = tensor("x_167_transpose_y_0"), val = tensor(false)]; - tensor value_17_cast_fp16 = transpose(perm = value_17_perm_0, x = v_15_cast_fp16)[name = tensor("transpose_155")]; - tensor x_167_cast_fp16 = matmul(transpose_x = x_167_transpose_x_0, transpose_y = x_167_transpose_y_0, x = input_397_cast_fp16, y = value_17_cast_fp16)[name = tensor("x_167_cast_fp16")]; - tensor var_1646_perm_0 = const()[name = tensor("op_1646_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_1647 = const()[name = tensor("op_1647"), val = tensor([1, -1, 512])]; - tensor var_1646_cast_fp16 = transpose(perm = var_1646_perm_0, x = x_167_cast_fp16)[name = tensor("transpose_151")]; - tensor input_399_cast_fp16 = reshape(shape = var_1647, x = var_1646_cast_fp16)[name = tensor("input_399_cast_fp16")]; - tensor module_layers_7_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_7_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96462016)))]; - tensor module_layers_7_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_7_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96986368)))]; - tensor linear_70_cast_fp16 = linear(bias = module_layers_7_self_attn_linear_out_bias_to_fp16, weight = module_layers_7_self_attn_linear_out_weight_to_fp16, x = input_399_cast_fp16)[name = tensor("linear_70_cast_fp16")]; - tensor input_403_cast_fp16 = add(x = input_395_cast_fp16, y = linear_70_cast_fp16)[name = tensor("input_403_cast_fp16")]; - tensor x_171_axes_0 = const()[name = tensor("x_171_axes_0"), val = tensor([-1])]; - tensor module_layers_7_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_7_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96987456)))]; - tensor module_layers_7_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_7_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96988544)))]; - tensor x_171_cast_fp16 = layer_norm(axes = x_171_axes_0, beta = module_layers_7_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_7_norm_conv_weight_to_fp16, x = input_403_cast_fp16)[name = tensor("x_171_cast_fp16")]; - tensor input_405_perm_0 = const()[name = tensor("input_405_perm_0"), val = tensor([0, 2, 1])]; - tensor input_407_pad_type_0 = const()[name = tensor("input_407_pad_type_0"), val = tensor("valid")]; - tensor input_407_strides_0 = const()[name = tensor("input_407_strides_0"), val = tensor([1])]; - tensor input_407_pad_0 = const()[name = tensor("input_407_pad_0"), val = tensor([0, 0])]; - tensor input_407_dilations_0 = const()[name = tensor("input_407_dilations_0"), val = tensor([1])]; - tensor input_407_groups_0 = const()[name = tensor("input_407_groups_0"), val = tensor(1)]; - tensor module_layers_7_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_7_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96989632)))]; - tensor module_layers_7_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_7_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98038272)))]; - tensor input_405_cast_fp16 = transpose(perm = input_405_perm_0, x = x_171_cast_fp16)[name = tensor("transpose_150")]; - tensor input_407_cast_fp16 = conv(bias = module_layers_7_conv_pointwise_conv1_bias_to_fp16, dilations = input_407_dilations_0, groups = input_407_groups_0, pad = input_407_pad_0, pad_type = input_407_pad_type_0, strides = input_407_strides_0, weight = module_layers_7_conv_pointwise_conv1_weight_to_fp16, x = input_405_cast_fp16)[name = tensor("input_407_cast_fp16")]; - tensor x_173_split_num_splits_0 = const()[name = tensor("x_173_split_num_splits_0"), val = tensor(2)]; - tensor x_173_split_axis_0 = const()[name = tensor("x_173_split_axis_0"), val = tensor(1)]; - tensor x_173_split_cast_fp16_0, tensor x_173_split_cast_fp16_1 = split(axis = x_173_split_axis_0, num_splits = x_173_split_num_splits_0, x = input_407_cast_fp16)[name = tensor("x_173_split_cast_fp16")]; - tensor x_173_split_1_sigmoid_cast_fp16 = sigmoid(x = x_173_split_cast_fp16_1)[name = tensor("x_173_split_1_sigmoid_cast_fp16")]; - tensor x_173_cast_fp16 = mul(x = x_173_split_cast_fp16_0, y = x_173_split_1_sigmoid_cast_fp16)[name = tensor("x_173_cast_fp16")]; - tensor input_409_cast_fp16 = select(a = var_11_to_fp16, b = x_173_cast_fp16, cond = var_453)[name = tensor("input_409_cast_fp16")]; - tensor input_411_pad_0 = const()[name = tensor("input_411_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_411_mode_0 = const()[name = tensor("input_411_mode_0"), val = tensor("constant")]; - tensor const_143_to_fp16 = const()[name = tensor("const_143_to_fp16"), val = tensor(0x0p+0)]; - tensor input_411_cast_fp16 = pad(constant_val = const_143_to_fp16, mode = input_411_mode_0, pad = input_411_pad_0, x = input_409_cast_fp16)[name = tensor("input_411_cast_fp16")]; - tensor input_413_pad_type_0 = const()[name = tensor("input_413_pad_type_0"), val = tensor("valid")]; - tensor input_413_groups_0 = const()[name = tensor("input_413_groups_0"), val = tensor(512)]; - tensor input_413_strides_0 = const()[name = tensor("input_413_strides_0"), val = tensor([1])]; - tensor input_413_pad_0 = const()[name = tensor("input_413_pad_0"), val = tensor([0, 0])]; - tensor input_413_dilations_0 = const()[name = tensor("input_413_dilations_0"), val = tensor([1])]; - tensor const_248_to_fp16 = const()[name = tensor("const_248_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98040384)))]; - tensor const_249_to_fp16 = const()[name = tensor("const_249_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98049664)))]; - tensor input_415_cast_fp16 = conv(bias = const_249_to_fp16, dilations = input_413_dilations_0, groups = input_413_groups_0, pad = input_413_pad_0, pad_type = input_413_pad_type_0, strides = input_413_strides_0, weight = const_248_to_fp16, x = input_411_cast_fp16)[name = tensor("input_415_cast_fp16")]; - tensor input_417_cast_fp16 = silu(x = input_415_cast_fp16)[name = tensor("input_417_cast_fp16")]; - tensor x_175_pad_type_0 = const()[name = tensor("x_175_pad_type_0"), val = tensor("valid")]; - tensor x_175_strides_0 = const()[name = tensor("x_175_strides_0"), val = tensor([1])]; - tensor x_175_pad_0 = const()[name = tensor("x_175_pad_0"), val = tensor([0, 0])]; - tensor x_175_dilations_0 = const()[name = tensor("x_175_dilations_0"), val = tensor([1])]; - tensor x_175_groups_0 = const()[name = tensor("x_175_groups_0"), val = tensor(1)]; - tensor module_layers_7_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_7_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98050752)))]; - tensor module_layers_7_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_7_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98575104)))]; - tensor x_175_cast_fp16 = conv(bias = module_layers_7_conv_pointwise_conv2_bias_to_fp16, dilations = x_175_dilations_0, groups = x_175_groups_0, pad = x_175_pad_0, pad_type = x_175_pad_type_0, strides = x_175_strides_0, weight = module_layers_7_conv_pointwise_conv2_weight_to_fp16, x = input_417_cast_fp16)[name = tensor("x_175_cast_fp16")]; - tensor input_419_perm_0 = const()[name = tensor("input_419_perm_0"), val = tensor([0, 2, 1])]; - tensor input_419_cast_fp16 = transpose(perm = input_419_perm_0, x = x_175_cast_fp16)[name = tensor("transpose_149")]; - tensor input_421_cast_fp16 = add(x = input_403_cast_fp16, y = input_419_cast_fp16)[name = tensor("input_421_cast_fp16")]; - tensor input_423_axes_0 = const()[name = tensor("input_423_axes_0"), val = tensor([-1])]; - tensor module_layers_7_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_7_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98576192)))]; - tensor module_layers_7_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_7_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98577280)))]; - tensor input_423_cast_fp16 = layer_norm(axes = input_423_axes_0, beta = module_layers_7_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_7_norm_feed_forward2_weight_to_fp16, x = input_421_cast_fp16)[name = tensor("input_423_cast_fp16")]; - tensor module_layers_7_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_7_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98578368)))]; - tensor module_layers_7_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_7_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(100675584)))]; - tensor linear_71_cast_fp16 = linear(bias = module_layers_7_feed_forward2_linear1_bias_to_fp16, weight = module_layers_7_feed_forward2_linear1_weight_to_fp16, x = input_423_cast_fp16)[name = tensor("linear_71_cast_fp16")]; - tensor input_427_cast_fp16 = silu(x = linear_71_cast_fp16)[name = tensor("input_427_cast_fp16")]; - tensor module_layers_7_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_7_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(100679744)))]; - tensor module_layers_7_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_7_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102776960)))]; - tensor linear_72_cast_fp16 = linear(bias = module_layers_7_feed_forward2_linear2_bias_to_fp16, weight = module_layers_7_feed_forward2_linear2_weight_to_fp16, x = input_427_cast_fp16)[name = tensor("linear_72_cast_fp16")]; - tensor var_1713_to_fp16 = const()[name = tensor("op_1713_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1714_cast_fp16 = mul(x = linear_72_cast_fp16, y = var_1713_to_fp16)[name = tensor("op_1714_cast_fp16")]; - tensor input_433_cast_fp16 = add(x = input_421_cast_fp16, y = var_1714_cast_fp16)[name = tensor("input_433_cast_fp16")]; - tensor input_435_axes_0 = const()[name = tensor("input_435_axes_0"), val = tensor([-1])]; - tensor module_layers_7_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_7_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102778048)))]; - tensor module_layers_7_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_7_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102779136)))]; - tensor input_435_cast_fp16 = layer_norm(axes = input_435_axes_0, beta = module_layers_7_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_7_norm_out_weight_to_fp16, x = input_433_cast_fp16)[name = tensor("input_435_cast_fp16")]; - tensor input_437_axes_0 = const()[name = tensor("input_437_axes_0"), val = tensor([-1])]; - tensor module_layers_8_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_8_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102780224)))]; - tensor module_layers_8_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_8_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102781312)))]; - tensor input_437_cast_fp16 = layer_norm(axes = input_437_axes_0, beta = module_layers_8_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_8_norm_feed_forward1_weight_to_fp16, x = input_435_cast_fp16)[name = tensor("input_437_cast_fp16")]; - tensor module_layers_8_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_8_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102782400)))]; - tensor module_layers_8_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_8_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(104879616)))]; - tensor linear_73_cast_fp16 = linear(bias = module_layers_8_feed_forward1_linear1_bias_to_fp16, weight = module_layers_8_feed_forward1_linear1_weight_to_fp16, x = input_437_cast_fp16)[name = tensor("linear_73_cast_fp16")]; - tensor input_441_cast_fp16 = silu(x = linear_73_cast_fp16)[name = tensor("input_441_cast_fp16")]; - tensor module_layers_8_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_8_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(104883776)))]; - tensor module_layers_8_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_8_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(106980992)))]; - tensor linear_74_cast_fp16 = linear(bias = module_layers_8_feed_forward1_linear2_bias_to_fp16, weight = module_layers_8_feed_forward1_linear2_weight_to_fp16, x = input_441_cast_fp16)[name = tensor("linear_74_cast_fp16")]; - tensor var_1744_to_fp16 = const()[name = tensor("op_1744_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1745_cast_fp16 = mul(x = linear_74_cast_fp16, y = var_1744_to_fp16)[name = tensor("op_1745_cast_fp16")]; - tensor input_447_cast_fp16 = add(x = input_435_cast_fp16, y = var_1745_cast_fp16)[name = tensor("input_447_cast_fp16")]; - tensor query_17_axes_0 = const()[name = tensor("query_17_axes_0"), val = tensor([-1])]; - tensor module_layers_8_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_8_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(106982080)))]; - tensor module_layers_8_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_8_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(106983168)))]; - tensor query_17_cast_fp16 = layer_norm(axes = query_17_axes_0, beta = module_layers_8_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_8_norm_self_att_weight_to_fp16, x = input_447_cast_fp16)[name = tensor("query_17_cast_fp16")]; - tensor module_layers_8_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_8_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(106984256)))]; - tensor module_layers_8_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_8_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(107508608)))]; - tensor linear_75_cast_fp16 = linear(bias = module_layers_8_self_attn_linear_q_bias_to_fp16, weight = module_layers_8_self_attn_linear_q_weight_to_fp16, x = query_17_cast_fp16)[name = tensor("linear_75_cast_fp16")]; - tensor var_1762 = const()[name = tensor("op_1762"), val = tensor([1, -1, 8, 64])]; - tensor q_49_cast_fp16 = reshape(shape = var_1762, x = linear_75_cast_fp16)[name = tensor("q_49_cast_fp16")]; - tensor module_layers_8_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_8_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(107509696)))]; - tensor module_layers_8_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_8_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108034048)))]; - tensor linear_76_cast_fp16 = linear(bias = module_layers_8_self_attn_linear_k_bias_to_fp16, weight = module_layers_8_self_attn_linear_k_weight_to_fp16, x = query_17_cast_fp16)[name = tensor("linear_76_cast_fp16")]; - tensor var_1767 = const()[name = tensor("op_1767"), val = tensor([1, -1, 8, 64])]; - tensor k_33_cast_fp16 = reshape(shape = var_1767, x = linear_76_cast_fp16)[name = tensor("k_33_cast_fp16")]; - tensor module_layers_8_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_8_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108035136)))]; - tensor module_layers_8_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_8_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108559488)))]; - tensor linear_77_cast_fp16 = linear(bias = module_layers_8_self_attn_linear_v_bias_to_fp16, weight = module_layers_8_self_attn_linear_v_weight_to_fp16, x = query_17_cast_fp16)[name = tensor("linear_77_cast_fp16")]; - tensor var_1772 = const()[name = tensor("op_1772"), val = tensor([1, -1, 8, 64])]; - tensor v_17_cast_fp16 = reshape(shape = var_1772, x = linear_77_cast_fp16)[name = tensor("v_17_cast_fp16")]; - tensor value_19_perm_0 = const()[name = tensor("value_19_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_8_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_8_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108560576)))]; - tensor var_1784_cast_fp16 = add(x = q_49_cast_fp16, y = module_layers_8_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1784_cast_fp16")]; - tensor module_layers_8_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_8_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108561664)))]; - tensor var_1786_cast_fp16 = add(x = q_49_cast_fp16, y = module_layers_8_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1786_cast_fp16")]; - tensor q_with_bias_v_17_perm_0 = const()[name = tensor("q_with_bias_v_17_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_183_transpose_x_0 = const()[name = tensor("x_183_transpose_x_0"), val = tensor(false)]; - tensor x_183_transpose_y_0 = const()[name = tensor("x_183_transpose_y_0"), val = tensor(false)]; - tensor var_1788_to_fp16 = const()[name = tensor("op_1788_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108562752)))]; - tensor q_with_bias_v_17_cast_fp16 = transpose(perm = q_with_bias_v_17_perm_0, x = var_1786_cast_fp16)[name = tensor("transpose_147")]; - tensor x_183_cast_fp16 = matmul(transpose_x = x_183_transpose_x_0, transpose_y = x_183_transpose_y_0, x = q_with_bias_v_17_cast_fp16, y = var_1788_to_fp16)[name = tensor("x_183_cast_fp16")]; - tensor x_185_pad_0 = const()[name = tensor("x_185_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_185_mode_0 = const()[name = tensor("x_185_mode_0"), val = tensor("constant")]; - tensor const_150_to_fp16 = const()[name = tensor("const_150_to_fp16"), val = tensor(0x0p+0)]; - tensor x_185_cast_fp16 = pad(constant_val = const_150_to_fp16, mode = x_185_mode_0, pad = x_185_pad_0, x = x_183_cast_fp16)[name = tensor("x_185_cast_fp16")]; - tensor var_1796 = const()[name = tensor("op_1796"), val = tensor([1, 8, -1, 188])]; - tensor x_187_cast_fp16 = reshape(shape = var_1796, x = x_185_cast_fp16)[name = tensor("x_187_cast_fp16")]; - tensor var_1800_begin_0 = const()[name = tensor("op_1800_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_1800_end_0 = const()[name = tensor("op_1800_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_1800_end_mask_0 = const()[name = tensor("op_1800_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_1800_cast_fp16 = slice_by_index(begin = var_1800_begin_0, end = var_1800_end_0, end_mask = var_1800_end_mask_0, x = x_187_cast_fp16)[name = tensor("op_1800_cast_fp16")]; - tensor var_1801 = const()[name = tensor("op_1801"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_33_cast_fp16 = reshape(shape = var_1801, x = var_1800_cast_fp16)[name = tensor("matrix_bd_33_cast_fp16")]; - tensor matrix_ac_17_transpose_x_0 = const()[name = tensor("matrix_ac_17_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_17_transpose_y_0 = const()[name = tensor("matrix_ac_17_transpose_y_0"), val = tensor(false)]; - tensor transpose_67_perm_0 = const()[name = tensor("transpose_67_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_68_perm_0 = const()[name = tensor("transpose_68_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_68 = transpose(perm = transpose_68_perm_0, x = k_33_cast_fp16)[name = tensor("transpose_145")]; - tensor transpose_67 = transpose(perm = transpose_67_perm_0, x = var_1784_cast_fp16)[name = tensor("transpose_146")]; - tensor matrix_ac_17_cast_fp16 = matmul(transpose_x = matrix_ac_17_transpose_x_0, transpose_y = matrix_ac_17_transpose_y_0, x = transpose_67, y = transpose_68)[name = tensor("matrix_ac_17_cast_fp16")]; - tensor matrix_bd_35_begin_0 = const()[name = tensor("matrix_bd_35_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_35_end_0 = const()[name = tensor("matrix_bd_35_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_35_end_mask_0 = const()[name = tensor("matrix_bd_35_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_35_cast_fp16 = slice_by_index(begin = matrix_bd_35_begin_0, end = matrix_bd_35_end_0, end_mask = matrix_bd_35_end_mask_0, x = matrix_bd_33_cast_fp16)[name = tensor("matrix_bd_35_cast_fp16")]; - tensor var_1810_cast_fp16 = add(x = matrix_ac_17_cast_fp16, y = matrix_bd_35_cast_fp16)[name = tensor("op_1810_cast_fp16")]; - tensor _inversed_scores_33_y_0_to_fp16 = const()[name = tensor("_inversed_scores_33_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_33_cast_fp16 = mul(x = var_1810_cast_fp16, y = _inversed_scores_33_y_0_to_fp16)[name = tensor("_inversed_scores_33_cast_fp16")]; - tensor scores_35_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_33_cast_fp16, cond = mask_11)[name = tensor("scores_35_cast_fp16")]; - tensor var_1816_cast_fp16 = softmax(axis = var_23, x = scores_35_cast_fp16)[name = tensor("op_1816_cast_fp16")]; - tensor input_449_cast_fp16 = select(a = var_11_to_fp16, b = var_1816_cast_fp16, cond = mask_11)[name = tensor("input_449_cast_fp16")]; - tensor x_189_transpose_x_0 = const()[name = tensor("x_189_transpose_x_0"), val = tensor(false)]; - tensor x_189_transpose_y_0 = const()[name = tensor("x_189_transpose_y_0"), val = tensor(false)]; - tensor value_19_cast_fp16 = transpose(perm = value_19_perm_0, x = v_17_cast_fp16)[name = tensor("transpose_148")]; - tensor x_189_cast_fp16 = matmul(transpose_x = x_189_transpose_x_0, transpose_y = x_189_transpose_y_0, x = input_449_cast_fp16, y = value_19_cast_fp16)[name = tensor("x_189_cast_fp16")]; - tensor var_1820_perm_0 = const()[name = tensor("op_1820_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_1821 = const()[name = tensor("op_1821"), val = tensor([1, -1, 512])]; - tensor var_1820_cast_fp16 = transpose(perm = var_1820_perm_0, x = x_189_cast_fp16)[name = tensor("transpose_144")]; - tensor input_451_cast_fp16 = reshape(shape = var_1821, x = var_1820_cast_fp16)[name = tensor("input_451_cast_fp16")]; - tensor module_layers_8_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_8_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108946816)))]; - tensor module_layers_8_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_8_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109471168)))]; - tensor linear_79_cast_fp16 = linear(bias = module_layers_8_self_attn_linear_out_bias_to_fp16, weight = module_layers_8_self_attn_linear_out_weight_to_fp16, x = input_451_cast_fp16)[name = tensor("linear_79_cast_fp16")]; - tensor input_455_cast_fp16 = add(x = input_447_cast_fp16, y = linear_79_cast_fp16)[name = tensor("input_455_cast_fp16")]; - tensor x_193_axes_0 = const()[name = tensor("x_193_axes_0"), val = tensor([-1])]; - tensor module_layers_8_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_8_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109472256)))]; - tensor module_layers_8_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_8_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109473344)))]; - tensor x_193_cast_fp16 = layer_norm(axes = x_193_axes_0, beta = module_layers_8_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_8_norm_conv_weight_to_fp16, x = input_455_cast_fp16)[name = tensor("x_193_cast_fp16")]; - tensor input_457_perm_0 = const()[name = tensor("input_457_perm_0"), val = tensor([0, 2, 1])]; - tensor input_459_pad_type_0 = const()[name = tensor("input_459_pad_type_0"), val = tensor("valid")]; - tensor input_459_strides_0 = const()[name = tensor("input_459_strides_0"), val = tensor([1])]; - tensor input_459_pad_0 = const()[name = tensor("input_459_pad_0"), val = tensor([0, 0])]; - tensor input_459_dilations_0 = const()[name = tensor("input_459_dilations_0"), val = tensor([1])]; - tensor input_459_groups_0 = const()[name = tensor("input_459_groups_0"), val = tensor(1)]; - tensor module_layers_8_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_8_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109474432)))]; - tensor module_layers_8_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_8_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110523072)))]; - tensor input_457_cast_fp16 = transpose(perm = input_457_perm_0, x = x_193_cast_fp16)[name = tensor("transpose_143")]; - tensor input_459_cast_fp16 = conv(bias = module_layers_8_conv_pointwise_conv1_bias_to_fp16, dilations = input_459_dilations_0, groups = input_459_groups_0, pad = input_459_pad_0, pad_type = input_459_pad_type_0, strides = input_459_strides_0, weight = module_layers_8_conv_pointwise_conv1_weight_to_fp16, x = input_457_cast_fp16)[name = tensor("input_459_cast_fp16")]; - tensor x_195_split_num_splits_0 = const()[name = tensor("x_195_split_num_splits_0"), val = tensor(2)]; - tensor x_195_split_axis_0 = const()[name = tensor("x_195_split_axis_0"), val = tensor(1)]; - tensor x_195_split_cast_fp16_0, tensor x_195_split_cast_fp16_1 = split(axis = x_195_split_axis_0, num_splits = x_195_split_num_splits_0, x = input_459_cast_fp16)[name = tensor("x_195_split_cast_fp16")]; - tensor x_195_split_1_sigmoid_cast_fp16 = sigmoid(x = x_195_split_cast_fp16_1)[name = tensor("x_195_split_1_sigmoid_cast_fp16")]; - tensor x_195_cast_fp16 = mul(x = x_195_split_cast_fp16_0, y = x_195_split_1_sigmoid_cast_fp16)[name = tensor("x_195_cast_fp16")]; - tensor input_461_cast_fp16 = select(a = var_11_to_fp16, b = x_195_cast_fp16, cond = var_453)[name = tensor("input_461_cast_fp16")]; - tensor input_463_pad_0 = const()[name = tensor("input_463_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_463_mode_0 = const()[name = tensor("input_463_mode_0"), val = tensor("constant")]; - tensor const_153_to_fp16 = const()[name = tensor("const_153_to_fp16"), val = tensor(0x0p+0)]; - tensor input_463_cast_fp16 = pad(constant_val = const_153_to_fp16, mode = input_463_mode_0, pad = input_463_pad_0, x = input_461_cast_fp16)[name = tensor("input_463_cast_fp16")]; - tensor input_465_pad_type_0 = const()[name = tensor("input_465_pad_type_0"), val = tensor("valid")]; - tensor input_465_groups_0 = const()[name = tensor("input_465_groups_0"), val = tensor(512)]; - tensor input_465_strides_0 = const()[name = tensor("input_465_strides_0"), val = tensor([1])]; - tensor input_465_pad_0 = const()[name = tensor("input_465_pad_0"), val = tensor([0, 0])]; - tensor input_465_dilations_0 = const()[name = tensor("input_465_dilations_0"), val = tensor([1])]; - tensor const_250_to_fp16 = const()[name = tensor("const_250_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110525184)))]; - tensor const_251_to_fp16 = const()[name = tensor("const_251_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110534464)))]; - tensor input_467_cast_fp16 = conv(bias = const_251_to_fp16, dilations = input_465_dilations_0, groups = input_465_groups_0, pad = input_465_pad_0, pad_type = input_465_pad_type_0, strides = input_465_strides_0, weight = const_250_to_fp16, x = input_463_cast_fp16)[name = tensor("input_467_cast_fp16")]; - tensor input_469_cast_fp16 = silu(x = input_467_cast_fp16)[name = tensor("input_469_cast_fp16")]; - tensor x_197_pad_type_0 = const()[name = tensor("x_197_pad_type_0"), val = tensor("valid")]; - tensor x_197_strides_0 = const()[name = tensor("x_197_strides_0"), val = tensor([1])]; - tensor x_197_pad_0 = const()[name = tensor("x_197_pad_0"), val = tensor([0, 0])]; - tensor x_197_dilations_0 = const()[name = tensor("x_197_dilations_0"), val = tensor([1])]; - tensor x_197_groups_0 = const()[name = tensor("x_197_groups_0"), val = tensor(1)]; - tensor module_layers_8_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_8_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110535552)))]; - tensor module_layers_8_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_8_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(111059904)))]; - tensor x_197_cast_fp16 = conv(bias = module_layers_8_conv_pointwise_conv2_bias_to_fp16, dilations = x_197_dilations_0, groups = x_197_groups_0, pad = x_197_pad_0, pad_type = x_197_pad_type_0, strides = x_197_strides_0, weight = module_layers_8_conv_pointwise_conv2_weight_to_fp16, x = input_469_cast_fp16)[name = tensor("x_197_cast_fp16")]; - tensor input_471_perm_0 = const()[name = tensor("input_471_perm_0"), val = tensor([0, 2, 1])]; - tensor input_471_cast_fp16 = transpose(perm = input_471_perm_0, x = x_197_cast_fp16)[name = tensor("transpose_142")]; - tensor input_473_cast_fp16 = add(x = input_455_cast_fp16, y = input_471_cast_fp16)[name = tensor("input_473_cast_fp16")]; - tensor input_475_axes_0 = const()[name = tensor("input_475_axes_0"), val = tensor([-1])]; - tensor module_layers_8_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_8_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(111060992)))]; - tensor module_layers_8_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_8_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(111062080)))]; - tensor input_475_cast_fp16 = layer_norm(axes = input_475_axes_0, beta = module_layers_8_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_8_norm_feed_forward2_weight_to_fp16, x = input_473_cast_fp16)[name = tensor("input_475_cast_fp16")]; - tensor module_layers_8_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_8_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(111063168)))]; - tensor module_layers_8_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_8_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(113160384)))]; - tensor linear_80_cast_fp16 = linear(bias = module_layers_8_feed_forward2_linear1_bias_to_fp16, weight = module_layers_8_feed_forward2_linear1_weight_to_fp16, x = input_475_cast_fp16)[name = tensor("linear_80_cast_fp16")]; - tensor input_479_cast_fp16 = silu(x = linear_80_cast_fp16)[name = tensor("input_479_cast_fp16")]; - tensor module_layers_8_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_8_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(113164544)))]; - tensor module_layers_8_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_8_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(115261760)))]; - tensor linear_81_cast_fp16 = linear(bias = module_layers_8_feed_forward2_linear2_bias_to_fp16, weight = module_layers_8_feed_forward2_linear2_weight_to_fp16, x = input_479_cast_fp16)[name = tensor("linear_81_cast_fp16")]; - tensor var_1887_to_fp16 = const()[name = tensor("op_1887_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1888_cast_fp16 = mul(x = linear_81_cast_fp16, y = var_1887_to_fp16)[name = tensor("op_1888_cast_fp16")]; - tensor input_485_cast_fp16 = add(x = input_473_cast_fp16, y = var_1888_cast_fp16)[name = tensor("input_485_cast_fp16")]; - tensor input_487_axes_0 = const()[name = tensor("input_487_axes_0"), val = tensor([-1])]; - tensor module_layers_8_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_8_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(115262848)))]; - tensor module_layers_8_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_8_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(115263936)))]; - tensor input_487_cast_fp16 = layer_norm(axes = input_487_axes_0, beta = module_layers_8_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_8_norm_out_weight_to_fp16, x = input_485_cast_fp16)[name = tensor("input_487_cast_fp16")]; - tensor input_489_axes_0 = const()[name = tensor("input_489_axes_0"), val = tensor([-1])]; - tensor module_layers_9_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_9_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(115265024)))]; - tensor module_layers_9_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_9_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(115266112)))]; - tensor input_489_cast_fp16 = layer_norm(axes = input_489_axes_0, beta = module_layers_9_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_9_norm_feed_forward1_weight_to_fp16, x = input_487_cast_fp16)[name = tensor("input_489_cast_fp16")]; - tensor module_layers_9_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_9_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(115267200)))]; - tensor module_layers_9_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_9_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(117364416)))]; - tensor linear_82_cast_fp16 = linear(bias = module_layers_9_feed_forward1_linear1_bias_to_fp16, weight = module_layers_9_feed_forward1_linear1_weight_to_fp16, x = input_489_cast_fp16)[name = tensor("linear_82_cast_fp16")]; - tensor input_493_cast_fp16 = silu(x = linear_82_cast_fp16)[name = tensor("input_493_cast_fp16")]; - tensor module_layers_9_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_9_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(117368576)))]; - tensor module_layers_9_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_9_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(119465792)))]; - tensor linear_83_cast_fp16 = linear(bias = module_layers_9_feed_forward1_linear2_bias_to_fp16, weight = module_layers_9_feed_forward1_linear2_weight_to_fp16, x = input_493_cast_fp16)[name = tensor("linear_83_cast_fp16")]; - tensor var_1918_to_fp16 = const()[name = tensor("op_1918_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1919_cast_fp16 = mul(x = linear_83_cast_fp16, y = var_1918_to_fp16)[name = tensor("op_1919_cast_fp16")]; - tensor input_499_cast_fp16 = add(x = input_487_cast_fp16, y = var_1919_cast_fp16)[name = tensor("input_499_cast_fp16")]; - tensor query_19_axes_0 = const()[name = tensor("query_19_axes_0"), val = tensor([-1])]; - tensor module_layers_9_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_9_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(119466880)))]; - tensor module_layers_9_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_9_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(119467968)))]; - tensor query_19_cast_fp16 = layer_norm(axes = query_19_axes_0, beta = module_layers_9_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_9_norm_self_att_weight_to_fp16, x = input_499_cast_fp16)[name = tensor("query_19_cast_fp16")]; - tensor module_layers_9_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_9_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(119469056)))]; - tensor module_layers_9_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_9_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(119993408)))]; - tensor linear_84_cast_fp16 = linear(bias = module_layers_9_self_attn_linear_q_bias_to_fp16, weight = module_layers_9_self_attn_linear_q_weight_to_fp16, x = query_19_cast_fp16)[name = tensor("linear_84_cast_fp16")]; - tensor var_1936 = const()[name = tensor("op_1936"), val = tensor([1, -1, 8, 64])]; - tensor q_55_cast_fp16 = reshape(shape = var_1936, x = linear_84_cast_fp16)[name = tensor("q_55_cast_fp16")]; - tensor module_layers_9_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_9_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(119994496)))]; - tensor module_layers_9_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_9_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(120518848)))]; - tensor linear_85_cast_fp16 = linear(bias = module_layers_9_self_attn_linear_k_bias_to_fp16, weight = module_layers_9_self_attn_linear_k_weight_to_fp16, x = query_19_cast_fp16)[name = tensor("linear_85_cast_fp16")]; - tensor var_1941 = const()[name = tensor("op_1941"), val = tensor([1, -1, 8, 64])]; - tensor k_37_cast_fp16 = reshape(shape = var_1941, x = linear_85_cast_fp16)[name = tensor("k_37_cast_fp16")]; - tensor module_layers_9_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_9_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(120519936)))]; - tensor module_layers_9_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_9_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121044288)))]; - tensor linear_86_cast_fp16 = linear(bias = module_layers_9_self_attn_linear_v_bias_to_fp16, weight = module_layers_9_self_attn_linear_v_weight_to_fp16, x = query_19_cast_fp16)[name = tensor("linear_86_cast_fp16")]; - tensor var_1946 = const()[name = tensor("op_1946"), val = tensor([1, -1, 8, 64])]; - tensor v_19_cast_fp16 = reshape(shape = var_1946, x = linear_86_cast_fp16)[name = tensor("v_19_cast_fp16")]; - tensor value_21_perm_0 = const()[name = tensor("value_21_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_9_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_9_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121045376)))]; - tensor var_1958_cast_fp16 = add(x = q_55_cast_fp16, y = module_layers_9_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1958_cast_fp16")]; - tensor module_layers_9_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_9_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121046464)))]; - tensor var_1960_cast_fp16 = add(x = q_55_cast_fp16, y = module_layers_9_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1960_cast_fp16")]; - tensor q_with_bias_v_19_perm_0 = const()[name = tensor("q_with_bias_v_19_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_205_transpose_x_0 = const()[name = tensor("x_205_transpose_x_0"), val = tensor(false)]; - tensor x_205_transpose_y_0 = const()[name = tensor("x_205_transpose_y_0"), val = tensor(false)]; - tensor var_1962_to_fp16 = const()[name = tensor("op_1962_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121047552)))]; - tensor q_with_bias_v_19_cast_fp16 = transpose(perm = q_with_bias_v_19_perm_0, x = var_1960_cast_fp16)[name = tensor("transpose_140")]; - tensor x_205_cast_fp16 = matmul(transpose_x = x_205_transpose_x_0, transpose_y = x_205_transpose_y_0, x = q_with_bias_v_19_cast_fp16, y = var_1962_to_fp16)[name = tensor("x_205_cast_fp16")]; - tensor x_207_pad_0 = const()[name = tensor("x_207_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_207_mode_0 = const()[name = tensor("x_207_mode_0"), val = tensor("constant")]; - tensor const_160_to_fp16 = const()[name = tensor("const_160_to_fp16"), val = tensor(0x0p+0)]; - tensor x_207_cast_fp16 = pad(constant_val = const_160_to_fp16, mode = x_207_mode_0, pad = x_207_pad_0, x = x_205_cast_fp16)[name = tensor("x_207_cast_fp16")]; - tensor var_1970 = const()[name = tensor("op_1970"), val = tensor([1, 8, -1, 188])]; - tensor x_209_cast_fp16 = reshape(shape = var_1970, x = x_207_cast_fp16)[name = tensor("x_209_cast_fp16")]; - tensor var_1974_begin_0 = const()[name = tensor("op_1974_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_1974_end_0 = const()[name = tensor("op_1974_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_1974_end_mask_0 = const()[name = tensor("op_1974_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_1974_cast_fp16 = slice_by_index(begin = var_1974_begin_0, end = var_1974_end_0, end_mask = var_1974_end_mask_0, x = x_209_cast_fp16)[name = tensor("op_1974_cast_fp16")]; - tensor var_1975 = const()[name = tensor("op_1975"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_37_cast_fp16 = reshape(shape = var_1975, x = var_1974_cast_fp16)[name = tensor("matrix_bd_37_cast_fp16")]; - tensor matrix_ac_19_transpose_x_0 = const()[name = tensor("matrix_ac_19_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_19_transpose_y_0 = const()[name = tensor("matrix_ac_19_transpose_y_0"), val = tensor(false)]; - tensor transpose_69_perm_0 = const()[name = tensor("transpose_69_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_70_perm_0 = const()[name = tensor("transpose_70_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_70 = transpose(perm = transpose_70_perm_0, x = k_37_cast_fp16)[name = tensor("transpose_138")]; - tensor transpose_69 = transpose(perm = transpose_69_perm_0, x = var_1958_cast_fp16)[name = tensor("transpose_139")]; - tensor matrix_ac_19_cast_fp16 = matmul(transpose_x = matrix_ac_19_transpose_x_0, transpose_y = matrix_ac_19_transpose_y_0, x = transpose_69, y = transpose_70)[name = tensor("matrix_ac_19_cast_fp16")]; - tensor matrix_bd_39_begin_0 = const()[name = tensor("matrix_bd_39_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_39_end_0 = const()[name = tensor("matrix_bd_39_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_39_end_mask_0 = const()[name = tensor("matrix_bd_39_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_39_cast_fp16 = slice_by_index(begin = matrix_bd_39_begin_0, end = matrix_bd_39_end_0, end_mask = matrix_bd_39_end_mask_0, x = matrix_bd_37_cast_fp16)[name = tensor("matrix_bd_39_cast_fp16")]; - tensor var_1984_cast_fp16 = add(x = matrix_ac_19_cast_fp16, y = matrix_bd_39_cast_fp16)[name = tensor("op_1984_cast_fp16")]; - tensor _inversed_scores_37_y_0_to_fp16 = const()[name = tensor("_inversed_scores_37_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_37_cast_fp16 = mul(x = var_1984_cast_fp16, y = _inversed_scores_37_y_0_to_fp16)[name = tensor("_inversed_scores_37_cast_fp16")]; - tensor scores_39_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_37_cast_fp16, cond = mask_11)[name = tensor("scores_39_cast_fp16")]; - tensor var_1990_cast_fp16 = softmax(axis = var_23, x = scores_39_cast_fp16)[name = tensor("op_1990_cast_fp16")]; - tensor input_501_cast_fp16 = select(a = var_11_to_fp16, b = var_1990_cast_fp16, cond = mask_11)[name = tensor("input_501_cast_fp16")]; - tensor x_211_transpose_x_0 = const()[name = tensor("x_211_transpose_x_0"), val = tensor(false)]; - tensor x_211_transpose_y_0 = const()[name = tensor("x_211_transpose_y_0"), val = tensor(false)]; - tensor value_21_cast_fp16 = transpose(perm = value_21_perm_0, x = v_19_cast_fp16)[name = tensor("transpose_141")]; - tensor x_211_cast_fp16 = matmul(transpose_x = x_211_transpose_x_0, transpose_y = x_211_transpose_y_0, x = input_501_cast_fp16, y = value_21_cast_fp16)[name = tensor("x_211_cast_fp16")]; - tensor var_1994_perm_0 = const()[name = tensor("op_1994_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_1995 = const()[name = tensor("op_1995"), val = tensor([1, -1, 512])]; - tensor var_1994_cast_fp16 = transpose(perm = var_1994_perm_0, x = x_211_cast_fp16)[name = tensor("transpose_137")]; - tensor input_503_cast_fp16 = reshape(shape = var_1995, x = var_1994_cast_fp16)[name = tensor("input_503_cast_fp16")]; - tensor module_layers_9_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_9_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121431616)))]; - tensor module_layers_9_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_9_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121955968)))]; - tensor linear_88_cast_fp16 = linear(bias = module_layers_9_self_attn_linear_out_bias_to_fp16, weight = module_layers_9_self_attn_linear_out_weight_to_fp16, x = input_503_cast_fp16)[name = tensor("linear_88_cast_fp16")]; - tensor input_507_cast_fp16 = add(x = input_499_cast_fp16, y = linear_88_cast_fp16)[name = tensor("input_507_cast_fp16")]; - tensor x_215_axes_0 = const()[name = tensor("x_215_axes_0"), val = tensor([-1])]; - tensor module_layers_9_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_9_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121957056)))]; - tensor module_layers_9_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_9_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121958144)))]; - tensor x_215_cast_fp16 = layer_norm(axes = x_215_axes_0, beta = module_layers_9_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_9_norm_conv_weight_to_fp16, x = input_507_cast_fp16)[name = tensor("x_215_cast_fp16")]; - tensor input_509_perm_0 = const()[name = tensor("input_509_perm_0"), val = tensor([0, 2, 1])]; - tensor input_511_pad_type_0 = const()[name = tensor("input_511_pad_type_0"), val = tensor("valid")]; - tensor input_511_strides_0 = const()[name = tensor("input_511_strides_0"), val = tensor([1])]; - tensor input_511_pad_0 = const()[name = tensor("input_511_pad_0"), val = tensor([0, 0])]; - tensor input_511_dilations_0 = const()[name = tensor("input_511_dilations_0"), val = tensor([1])]; - tensor input_511_groups_0 = const()[name = tensor("input_511_groups_0"), val = tensor(1)]; - tensor module_layers_9_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_9_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121959232)))]; - tensor module_layers_9_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_9_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123007872)))]; - tensor input_509_cast_fp16 = transpose(perm = input_509_perm_0, x = x_215_cast_fp16)[name = tensor("transpose_136")]; - tensor input_511_cast_fp16 = conv(bias = module_layers_9_conv_pointwise_conv1_bias_to_fp16, dilations = input_511_dilations_0, groups = input_511_groups_0, pad = input_511_pad_0, pad_type = input_511_pad_type_0, strides = input_511_strides_0, weight = module_layers_9_conv_pointwise_conv1_weight_to_fp16, x = input_509_cast_fp16)[name = tensor("input_511_cast_fp16")]; - tensor x_217_split_num_splits_0 = const()[name = tensor("x_217_split_num_splits_0"), val = tensor(2)]; - tensor x_217_split_axis_0 = const()[name = tensor("x_217_split_axis_0"), val = tensor(1)]; - tensor x_217_split_cast_fp16_0, tensor x_217_split_cast_fp16_1 = split(axis = x_217_split_axis_0, num_splits = x_217_split_num_splits_0, x = input_511_cast_fp16)[name = tensor("x_217_split_cast_fp16")]; - tensor x_217_split_1_sigmoid_cast_fp16 = sigmoid(x = x_217_split_cast_fp16_1)[name = tensor("x_217_split_1_sigmoid_cast_fp16")]; - tensor x_217_cast_fp16 = mul(x = x_217_split_cast_fp16_0, y = x_217_split_1_sigmoid_cast_fp16)[name = tensor("x_217_cast_fp16")]; - tensor input_513_cast_fp16 = select(a = var_11_to_fp16, b = x_217_cast_fp16, cond = var_453)[name = tensor("input_513_cast_fp16")]; - tensor input_515_pad_0 = const()[name = tensor("input_515_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_515_mode_0 = const()[name = tensor("input_515_mode_0"), val = tensor("constant")]; - tensor const_163_to_fp16 = const()[name = tensor("const_163_to_fp16"), val = tensor(0x0p+0)]; - tensor input_515_cast_fp16 = pad(constant_val = const_163_to_fp16, mode = input_515_mode_0, pad = input_515_pad_0, x = input_513_cast_fp16)[name = tensor("input_515_cast_fp16")]; - tensor input_517_pad_type_0 = const()[name = tensor("input_517_pad_type_0"), val = tensor("valid")]; - tensor input_517_groups_0 = const()[name = tensor("input_517_groups_0"), val = tensor(512)]; - tensor input_517_strides_0 = const()[name = tensor("input_517_strides_0"), val = tensor([1])]; - tensor input_517_pad_0 = const()[name = tensor("input_517_pad_0"), val = tensor([0, 0])]; - tensor input_517_dilations_0 = const()[name = tensor("input_517_dilations_0"), val = tensor([1])]; - tensor const_252_to_fp16 = const()[name = tensor("const_252_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123009984)))]; - tensor const_253_to_fp16 = const()[name = tensor("const_253_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123019264)))]; - tensor input_519_cast_fp16 = conv(bias = const_253_to_fp16, dilations = input_517_dilations_0, groups = input_517_groups_0, pad = input_517_pad_0, pad_type = input_517_pad_type_0, strides = input_517_strides_0, weight = const_252_to_fp16, x = input_515_cast_fp16)[name = tensor("input_519_cast_fp16")]; - tensor input_521_cast_fp16 = silu(x = input_519_cast_fp16)[name = tensor("input_521_cast_fp16")]; - tensor x_219_pad_type_0 = const()[name = tensor("x_219_pad_type_0"), val = tensor("valid")]; - tensor x_219_strides_0 = const()[name = tensor("x_219_strides_0"), val = tensor([1])]; - tensor x_219_pad_0 = const()[name = tensor("x_219_pad_0"), val = tensor([0, 0])]; - tensor x_219_dilations_0 = const()[name = tensor("x_219_dilations_0"), val = tensor([1])]; - tensor x_219_groups_0 = const()[name = tensor("x_219_groups_0"), val = tensor(1)]; - tensor module_layers_9_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_9_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123020352)))]; - tensor module_layers_9_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_9_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123544704)))]; - tensor x_219_cast_fp16 = conv(bias = module_layers_9_conv_pointwise_conv2_bias_to_fp16, dilations = x_219_dilations_0, groups = x_219_groups_0, pad = x_219_pad_0, pad_type = x_219_pad_type_0, strides = x_219_strides_0, weight = module_layers_9_conv_pointwise_conv2_weight_to_fp16, x = input_521_cast_fp16)[name = tensor("x_219_cast_fp16")]; - tensor input_523_perm_0 = const()[name = tensor("input_523_perm_0"), val = tensor([0, 2, 1])]; - tensor input_523_cast_fp16 = transpose(perm = input_523_perm_0, x = x_219_cast_fp16)[name = tensor("transpose_135")]; - tensor input_525_cast_fp16 = add(x = input_507_cast_fp16, y = input_523_cast_fp16)[name = tensor("input_525_cast_fp16")]; - tensor input_527_axes_0 = const()[name = tensor("input_527_axes_0"), val = tensor([-1])]; - tensor module_layers_9_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_9_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123545792)))]; - tensor module_layers_9_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_9_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123546880)))]; - tensor input_527_cast_fp16 = layer_norm(axes = input_527_axes_0, beta = module_layers_9_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_9_norm_feed_forward2_weight_to_fp16, x = input_525_cast_fp16)[name = tensor("input_527_cast_fp16")]; - tensor module_layers_9_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_9_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123547968)))]; - tensor module_layers_9_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_9_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(125645184)))]; - tensor linear_89_cast_fp16 = linear(bias = module_layers_9_feed_forward2_linear1_bias_to_fp16, weight = module_layers_9_feed_forward2_linear1_weight_to_fp16, x = input_527_cast_fp16)[name = tensor("linear_89_cast_fp16")]; - tensor input_531_cast_fp16 = silu(x = linear_89_cast_fp16)[name = tensor("input_531_cast_fp16")]; - tensor module_layers_9_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_9_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(125649344)))]; - tensor module_layers_9_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_9_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127746560)))]; - tensor linear_90_cast_fp16 = linear(bias = module_layers_9_feed_forward2_linear2_bias_to_fp16, weight = module_layers_9_feed_forward2_linear2_weight_to_fp16, x = input_531_cast_fp16)[name = tensor("linear_90_cast_fp16")]; - tensor var_2061_to_fp16 = const()[name = tensor("op_2061_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2062_cast_fp16 = mul(x = linear_90_cast_fp16, y = var_2061_to_fp16)[name = tensor("op_2062_cast_fp16")]; - tensor input_537_cast_fp16 = add(x = input_525_cast_fp16, y = var_2062_cast_fp16)[name = tensor("input_537_cast_fp16")]; - tensor input_539_axes_0 = const()[name = tensor("input_539_axes_0"), val = tensor([-1])]; - tensor module_layers_9_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_9_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127747648)))]; - tensor module_layers_9_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_9_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127748736)))]; - tensor input_539_cast_fp16 = layer_norm(axes = input_539_axes_0, beta = module_layers_9_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_9_norm_out_weight_to_fp16, x = input_537_cast_fp16)[name = tensor("input_539_cast_fp16")]; - tensor input_541_axes_0 = const()[name = tensor("input_541_axes_0"), val = tensor([-1])]; - tensor module_layers_10_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_10_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127749824)))]; - tensor module_layers_10_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_10_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127750912)))]; - tensor input_541_cast_fp16 = layer_norm(axes = input_541_axes_0, beta = module_layers_10_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_10_norm_feed_forward1_weight_to_fp16, x = input_539_cast_fp16)[name = tensor("input_541_cast_fp16")]; - tensor module_layers_10_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_10_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127752000)))]; - tensor module_layers_10_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_10_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(129849216)))]; - tensor linear_91_cast_fp16 = linear(bias = module_layers_10_feed_forward1_linear1_bias_to_fp16, weight = module_layers_10_feed_forward1_linear1_weight_to_fp16, x = input_541_cast_fp16)[name = tensor("linear_91_cast_fp16")]; - tensor input_545_cast_fp16 = silu(x = linear_91_cast_fp16)[name = tensor("input_545_cast_fp16")]; - tensor module_layers_10_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_10_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(129853376)))]; - tensor module_layers_10_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_10_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(131950592)))]; - tensor linear_92_cast_fp16 = linear(bias = module_layers_10_feed_forward1_linear2_bias_to_fp16, weight = module_layers_10_feed_forward1_linear2_weight_to_fp16, x = input_545_cast_fp16)[name = tensor("linear_92_cast_fp16")]; - tensor var_2092_to_fp16 = const()[name = tensor("op_2092_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2093_cast_fp16 = mul(x = linear_92_cast_fp16, y = var_2092_to_fp16)[name = tensor("op_2093_cast_fp16")]; - tensor input_551_cast_fp16 = add(x = input_539_cast_fp16, y = var_2093_cast_fp16)[name = tensor("input_551_cast_fp16")]; - tensor query_21_axes_0 = const()[name = tensor("query_21_axes_0"), val = tensor([-1])]; - tensor module_layers_10_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_10_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(131951680)))]; - tensor module_layers_10_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_10_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(131952768)))]; - tensor query_21_cast_fp16 = layer_norm(axes = query_21_axes_0, beta = module_layers_10_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_10_norm_self_att_weight_to_fp16, x = input_551_cast_fp16)[name = tensor("query_21_cast_fp16")]; - tensor module_layers_10_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_10_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(131953856)))]; - tensor module_layers_10_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_10_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132478208)))]; - tensor linear_93_cast_fp16 = linear(bias = module_layers_10_self_attn_linear_q_bias_to_fp16, weight = module_layers_10_self_attn_linear_q_weight_to_fp16, x = query_21_cast_fp16)[name = tensor("linear_93_cast_fp16")]; - tensor var_2110 = const()[name = tensor("op_2110"), val = tensor([1, -1, 8, 64])]; - tensor q_61_cast_fp16 = reshape(shape = var_2110, x = linear_93_cast_fp16)[name = tensor("q_61_cast_fp16")]; - tensor module_layers_10_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_10_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132479296)))]; - tensor module_layers_10_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_10_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133003648)))]; - tensor linear_94_cast_fp16 = linear(bias = module_layers_10_self_attn_linear_k_bias_to_fp16, weight = module_layers_10_self_attn_linear_k_weight_to_fp16, x = query_21_cast_fp16)[name = tensor("linear_94_cast_fp16")]; - tensor var_2115 = const()[name = tensor("op_2115"), val = tensor([1, -1, 8, 64])]; - tensor k_41_cast_fp16 = reshape(shape = var_2115, x = linear_94_cast_fp16)[name = tensor("k_41_cast_fp16")]; - tensor module_layers_10_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_10_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133004736)))]; - tensor module_layers_10_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_10_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133529088)))]; - tensor linear_95_cast_fp16 = linear(bias = module_layers_10_self_attn_linear_v_bias_to_fp16, weight = module_layers_10_self_attn_linear_v_weight_to_fp16, x = query_21_cast_fp16)[name = tensor("linear_95_cast_fp16")]; - tensor var_2120 = const()[name = tensor("op_2120"), val = tensor([1, -1, 8, 64])]; - tensor v_21_cast_fp16 = reshape(shape = var_2120, x = linear_95_cast_fp16)[name = tensor("v_21_cast_fp16")]; - tensor value_23_perm_0 = const()[name = tensor("value_23_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_10_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_10_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133530176)))]; - tensor var_2132_cast_fp16 = add(x = q_61_cast_fp16, y = module_layers_10_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2132_cast_fp16")]; - tensor module_layers_10_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_10_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133531264)))]; - tensor var_2134_cast_fp16 = add(x = q_61_cast_fp16, y = module_layers_10_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2134_cast_fp16")]; - tensor q_with_bias_v_21_perm_0 = const()[name = tensor("q_with_bias_v_21_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_227_transpose_x_0 = const()[name = tensor("x_227_transpose_x_0"), val = tensor(false)]; - tensor x_227_transpose_y_0 = const()[name = tensor("x_227_transpose_y_0"), val = tensor(false)]; - tensor var_2136_to_fp16 = const()[name = tensor("op_2136_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133532352)))]; - tensor q_with_bias_v_21_cast_fp16 = transpose(perm = q_with_bias_v_21_perm_0, x = var_2134_cast_fp16)[name = tensor("transpose_133")]; - tensor x_227_cast_fp16 = matmul(transpose_x = x_227_transpose_x_0, transpose_y = x_227_transpose_y_0, x = q_with_bias_v_21_cast_fp16, y = var_2136_to_fp16)[name = tensor("x_227_cast_fp16")]; - tensor x_229_pad_0 = const()[name = tensor("x_229_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_229_mode_0 = const()[name = tensor("x_229_mode_0"), val = tensor("constant")]; - tensor const_170_to_fp16 = const()[name = tensor("const_170_to_fp16"), val = tensor(0x0p+0)]; - tensor x_229_cast_fp16 = pad(constant_val = const_170_to_fp16, mode = x_229_mode_0, pad = x_229_pad_0, x = x_227_cast_fp16)[name = tensor("x_229_cast_fp16")]; - tensor var_2144 = const()[name = tensor("op_2144"), val = tensor([1, 8, -1, 188])]; - tensor x_231_cast_fp16 = reshape(shape = var_2144, x = x_229_cast_fp16)[name = tensor("x_231_cast_fp16")]; - tensor var_2148_begin_0 = const()[name = tensor("op_2148_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_2148_end_0 = const()[name = tensor("op_2148_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_2148_end_mask_0 = const()[name = tensor("op_2148_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_2148_cast_fp16 = slice_by_index(begin = var_2148_begin_0, end = var_2148_end_0, end_mask = var_2148_end_mask_0, x = x_231_cast_fp16)[name = tensor("op_2148_cast_fp16")]; - tensor var_2149 = const()[name = tensor("op_2149"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_41_cast_fp16 = reshape(shape = var_2149, x = var_2148_cast_fp16)[name = tensor("matrix_bd_41_cast_fp16")]; - tensor matrix_ac_21_transpose_x_0 = const()[name = tensor("matrix_ac_21_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_21_transpose_y_0 = const()[name = tensor("matrix_ac_21_transpose_y_0"), val = tensor(false)]; - tensor transpose_71_perm_0 = const()[name = tensor("transpose_71_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_72_perm_0 = const()[name = tensor("transpose_72_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_72 = transpose(perm = transpose_72_perm_0, x = k_41_cast_fp16)[name = tensor("transpose_131")]; - tensor transpose_71 = transpose(perm = transpose_71_perm_0, x = var_2132_cast_fp16)[name = tensor("transpose_132")]; - tensor matrix_ac_21_cast_fp16 = matmul(transpose_x = matrix_ac_21_transpose_x_0, transpose_y = matrix_ac_21_transpose_y_0, x = transpose_71, y = transpose_72)[name = tensor("matrix_ac_21_cast_fp16")]; - tensor matrix_bd_43_begin_0 = const()[name = tensor("matrix_bd_43_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_43_end_0 = const()[name = tensor("matrix_bd_43_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_43_end_mask_0 = const()[name = tensor("matrix_bd_43_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_43_cast_fp16 = slice_by_index(begin = matrix_bd_43_begin_0, end = matrix_bd_43_end_0, end_mask = matrix_bd_43_end_mask_0, x = matrix_bd_41_cast_fp16)[name = tensor("matrix_bd_43_cast_fp16")]; - tensor var_2158_cast_fp16 = add(x = matrix_ac_21_cast_fp16, y = matrix_bd_43_cast_fp16)[name = tensor("op_2158_cast_fp16")]; - tensor _inversed_scores_41_y_0_to_fp16 = const()[name = tensor("_inversed_scores_41_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_41_cast_fp16 = mul(x = var_2158_cast_fp16, y = _inversed_scores_41_y_0_to_fp16)[name = tensor("_inversed_scores_41_cast_fp16")]; - tensor scores_43_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_41_cast_fp16, cond = mask_11)[name = tensor("scores_43_cast_fp16")]; - tensor var_2164_cast_fp16 = softmax(axis = var_23, x = scores_43_cast_fp16)[name = tensor("op_2164_cast_fp16")]; - tensor input_553_cast_fp16 = select(a = var_11_to_fp16, b = var_2164_cast_fp16, cond = mask_11)[name = tensor("input_553_cast_fp16")]; - tensor x_233_transpose_x_0 = const()[name = tensor("x_233_transpose_x_0"), val = tensor(false)]; - tensor x_233_transpose_y_0 = const()[name = tensor("x_233_transpose_y_0"), val = tensor(false)]; - tensor value_23_cast_fp16 = transpose(perm = value_23_perm_0, x = v_21_cast_fp16)[name = tensor("transpose_134")]; - tensor x_233_cast_fp16 = matmul(transpose_x = x_233_transpose_x_0, transpose_y = x_233_transpose_y_0, x = input_553_cast_fp16, y = value_23_cast_fp16)[name = tensor("x_233_cast_fp16")]; - tensor var_2168_perm_0 = const()[name = tensor("op_2168_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_2169 = const()[name = tensor("op_2169"), val = tensor([1, -1, 512])]; - tensor var_2168_cast_fp16 = transpose(perm = var_2168_perm_0, x = x_233_cast_fp16)[name = tensor("transpose_130")]; - tensor input_555_cast_fp16 = reshape(shape = var_2169, x = var_2168_cast_fp16)[name = tensor("input_555_cast_fp16")]; - tensor module_layers_10_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_10_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133916416)))]; - tensor module_layers_10_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_10_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134440768)))]; - tensor linear_97_cast_fp16 = linear(bias = module_layers_10_self_attn_linear_out_bias_to_fp16, weight = module_layers_10_self_attn_linear_out_weight_to_fp16, x = input_555_cast_fp16)[name = tensor("linear_97_cast_fp16")]; - tensor input_559_cast_fp16 = add(x = input_551_cast_fp16, y = linear_97_cast_fp16)[name = tensor("input_559_cast_fp16")]; - tensor x_237_axes_0 = const()[name = tensor("x_237_axes_0"), val = tensor([-1])]; - tensor module_layers_10_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_10_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134441856)))]; - tensor module_layers_10_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_10_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134442944)))]; - tensor x_237_cast_fp16 = layer_norm(axes = x_237_axes_0, beta = module_layers_10_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_10_norm_conv_weight_to_fp16, x = input_559_cast_fp16)[name = tensor("x_237_cast_fp16")]; - tensor input_561_perm_0 = const()[name = tensor("input_561_perm_0"), val = tensor([0, 2, 1])]; - tensor input_563_pad_type_0 = const()[name = tensor("input_563_pad_type_0"), val = tensor("valid")]; - tensor input_563_strides_0 = const()[name = tensor("input_563_strides_0"), val = tensor([1])]; - tensor input_563_pad_0 = const()[name = tensor("input_563_pad_0"), val = tensor([0, 0])]; - tensor input_563_dilations_0 = const()[name = tensor("input_563_dilations_0"), val = tensor([1])]; - tensor input_563_groups_0 = const()[name = tensor("input_563_groups_0"), val = tensor(1)]; - tensor module_layers_10_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_10_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134444032)))]; - tensor module_layers_10_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_10_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(135492672)))]; - tensor input_561_cast_fp16 = transpose(perm = input_561_perm_0, x = x_237_cast_fp16)[name = tensor("transpose_129")]; - tensor input_563_cast_fp16 = conv(bias = module_layers_10_conv_pointwise_conv1_bias_to_fp16, dilations = input_563_dilations_0, groups = input_563_groups_0, pad = input_563_pad_0, pad_type = input_563_pad_type_0, strides = input_563_strides_0, weight = module_layers_10_conv_pointwise_conv1_weight_to_fp16, x = input_561_cast_fp16)[name = tensor("input_563_cast_fp16")]; - tensor x_239_split_num_splits_0 = const()[name = tensor("x_239_split_num_splits_0"), val = tensor(2)]; - tensor x_239_split_axis_0 = const()[name = tensor("x_239_split_axis_0"), val = tensor(1)]; - tensor x_239_split_cast_fp16_0, tensor x_239_split_cast_fp16_1 = split(axis = x_239_split_axis_0, num_splits = x_239_split_num_splits_0, x = input_563_cast_fp16)[name = tensor("x_239_split_cast_fp16")]; - tensor x_239_split_1_sigmoid_cast_fp16 = sigmoid(x = x_239_split_cast_fp16_1)[name = tensor("x_239_split_1_sigmoid_cast_fp16")]; - tensor x_239_cast_fp16 = mul(x = x_239_split_cast_fp16_0, y = x_239_split_1_sigmoid_cast_fp16)[name = tensor("x_239_cast_fp16")]; - tensor input_565_cast_fp16 = select(a = var_11_to_fp16, b = x_239_cast_fp16, cond = var_453)[name = tensor("input_565_cast_fp16")]; - tensor input_567_pad_0 = const()[name = tensor("input_567_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_567_mode_0 = const()[name = tensor("input_567_mode_0"), val = tensor("constant")]; - tensor const_173_to_fp16 = const()[name = tensor("const_173_to_fp16"), val = tensor(0x0p+0)]; - tensor input_567_cast_fp16 = pad(constant_val = const_173_to_fp16, mode = input_567_mode_0, pad = input_567_pad_0, x = input_565_cast_fp16)[name = tensor("input_567_cast_fp16")]; - tensor input_569_pad_type_0 = const()[name = tensor("input_569_pad_type_0"), val = tensor("valid")]; - tensor input_569_groups_0 = const()[name = tensor("input_569_groups_0"), val = tensor(512)]; - tensor input_569_strides_0 = const()[name = tensor("input_569_strides_0"), val = tensor([1])]; - tensor input_569_pad_0 = const()[name = tensor("input_569_pad_0"), val = tensor([0, 0])]; - tensor input_569_dilations_0 = const()[name = tensor("input_569_dilations_0"), val = tensor([1])]; - tensor const_254_to_fp16 = const()[name = tensor("const_254_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(135494784)))]; - tensor const_255_to_fp16 = const()[name = tensor("const_255_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(135504064)))]; - tensor input_571_cast_fp16 = conv(bias = const_255_to_fp16, dilations = input_569_dilations_0, groups = input_569_groups_0, pad = input_569_pad_0, pad_type = input_569_pad_type_0, strides = input_569_strides_0, weight = const_254_to_fp16, x = input_567_cast_fp16)[name = tensor("input_571_cast_fp16")]; - tensor input_573_cast_fp16 = silu(x = input_571_cast_fp16)[name = tensor("input_573_cast_fp16")]; - tensor x_241_pad_type_0 = const()[name = tensor("x_241_pad_type_0"), val = tensor("valid")]; - tensor x_241_strides_0 = const()[name = tensor("x_241_strides_0"), val = tensor([1])]; - tensor x_241_pad_0 = const()[name = tensor("x_241_pad_0"), val = tensor([0, 0])]; - tensor x_241_dilations_0 = const()[name = tensor("x_241_dilations_0"), val = tensor([1])]; - tensor x_241_groups_0 = const()[name = tensor("x_241_groups_0"), val = tensor(1)]; - tensor module_layers_10_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_10_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(135505152)))]; - tensor module_layers_10_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_10_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136029504)))]; - tensor x_241_cast_fp16 = conv(bias = module_layers_10_conv_pointwise_conv2_bias_to_fp16, dilations = x_241_dilations_0, groups = x_241_groups_0, pad = x_241_pad_0, pad_type = x_241_pad_type_0, strides = x_241_strides_0, weight = module_layers_10_conv_pointwise_conv2_weight_to_fp16, x = input_573_cast_fp16)[name = tensor("x_241_cast_fp16")]; - tensor input_575_perm_0 = const()[name = tensor("input_575_perm_0"), val = tensor([0, 2, 1])]; - tensor input_575_cast_fp16 = transpose(perm = input_575_perm_0, x = x_241_cast_fp16)[name = tensor("transpose_128")]; - tensor input_577_cast_fp16 = add(x = input_559_cast_fp16, y = input_575_cast_fp16)[name = tensor("input_577_cast_fp16")]; - tensor input_579_axes_0 = const()[name = tensor("input_579_axes_0"), val = tensor([-1])]; - tensor module_layers_10_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_10_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136030592)))]; - tensor module_layers_10_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_10_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136031680)))]; - tensor input_579_cast_fp16 = layer_norm(axes = input_579_axes_0, beta = module_layers_10_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_10_norm_feed_forward2_weight_to_fp16, x = input_577_cast_fp16)[name = tensor("input_579_cast_fp16")]; - tensor module_layers_10_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_10_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136032768)))]; - tensor module_layers_10_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_10_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(138129984)))]; - tensor linear_98_cast_fp16 = linear(bias = module_layers_10_feed_forward2_linear1_bias_to_fp16, weight = module_layers_10_feed_forward2_linear1_weight_to_fp16, x = input_579_cast_fp16)[name = tensor("linear_98_cast_fp16")]; - tensor input_583_cast_fp16 = silu(x = linear_98_cast_fp16)[name = tensor("input_583_cast_fp16")]; - tensor module_layers_10_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_10_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(138134144)))]; - tensor module_layers_10_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_10_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140231360)))]; - tensor linear_99_cast_fp16 = linear(bias = module_layers_10_feed_forward2_linear2_bias_to_fp16, weight = module_layers_10_feed_forward2_linear2_weight_to_fp16, x = input_583_cast_fp16)[name = tensor("linear_99_cast_fp16")]; - tensor var_2235_to_fp16 = const()[name = tensor("op_2235_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2236_cast_fp16 = mul(x = linear_99_cast_fp16, y = var_2235_to_fp16)[name = tensor("op_2236_cast_fp16")]; - tensor input_589_cast_fp16 = add(x = input_577_cast_fp16, y = var_2236_cast_fp16)[name = tensor("input_589_cast_fp16")]; - tensor input_591_axes_0 = const()[name = tensor("input_591_axes_0"), val = tensor([-1])]; - tensor module_layers_10_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_10_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140232448)))]; - tensor module_layers_10_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_10_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140233536)))]; - tensor input_591_cast_fp16 = layer_norm(axes = input_591_axes_0, beta = module_layers_10_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_10_norm_out_weight_to_fp16, x = input_589_cast_fp16)[name = tensor("input_591_cast_fp16")]; - tensor input_593_axes_0 = const()[name = tensor("input_593_axes_0"), val = tensor([-1])]; - tensor module_layers_11_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_11_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140234624)))]; - tensor module_layers_11_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_11_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140235712)))]; - tensor input_593_cast_fp16 = layer_norm(axes = input_593_axes_0, beta = module_layers_11_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_11_norm_feed_forward1_weight_to_fp16, x = input_591_cast_fp16)[name = tensor("input_593_cast_fp16")]; - tensor module_layers_11_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_11_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140236800)))]; - tensor module_layers_11_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_11_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(142334016)))]; - tensor linear_100_cast_fp16 = linear(bias = module_layers_11_feed_forward1_linear1_bias_to_fp16, weight = module_layers_11_feed_forward1_linear1_weight_to_fp16, x = input_593_cast_fp16)[name = tensor("linear_100_cast_fp16")]; - tensor input_597_cast_fp16 = silu(x = linear_100_cast_fp16)[name = tensor("input_597_cast_fp16")]; - tensor module_layers_11_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_11_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(142338176)))]; - tensor module_layers_11_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_11_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(144435392)))]; - tensor linear_101_cast_fp16 = linear(bias = module_layers_11_feed_forward1_linear2_bias_to_fp16, weight = module_layers_11_feed_forward1_linear2_weight_to_fp16, x = input_597_cast_fp16)[name = tensor("linear_101_cast_fp16")]; - tensor var_2266_to_fp16 = const()[name = tensor("op_2266_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2267_cast_fp16 = mul(x = linear_101_cast_fp16, y = var_2266_to_fp16)[name = tensor("op_2267_cast_fp16")]; - tensor input_603_cast_fp16 = add(x = input_591_cast_fp16, y = var_2267_cast_fp16)[name = tensor("input_603_cast_fp16")]; - tensor query_23_axes_0 = const()[name = tensor("query_23_axes_0"), val = tensor([-1])]; - tensor module_layers_11_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_11_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(144436480)))]; - tensor module_layers_11_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_11_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(144437568)))]; - tensor query_23_cast_fp16 = layer_norm(axes = query_23_axes_0, beta = module_layers_11_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_11_norm_self_att_weight_to_fp16, x = input_603_cast_fp16)[name = tensor("query_23_cast_fp16")]; - tensor module_layers_11_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_11_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(144438656)))]; - tensor module_layers_11_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_11_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(144963008)))]; - tensor linear_102_cast_fp16 = linear(bias = module_layers_11_self_attn_linear_q_bias_to_fp16, weight = module_layers_11_self_attn_linear_q_weight_to_fp16, x = query_23_cast_fp16)[name = tensor("linear_102_cast_fp16")]; - tensor var_2284 = const()[name = tensor("op_2284"), val = tensor([1, -1, 8, 64])]; - tensor q_67_cast_fp16 = reshape(shape = var_2284, x = linear_102_cast_fp16)[name = tensor("q_67_cast_fp16")]; - tensor module_layers_11_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_11_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(144964096)))]; - tensor module_layers_11_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_11_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145488448)))]; - tensor linear_103_cast_fp16 = linear(bias = module_layers_11_self_attn_linear_k_bias_to_fp16, weight = module_layers_11_self_attn_linear_k_weight_to_fp16, x = query_23_cast_fp16)[name = tensor("linear_103_cast_fp16")]; - tensor var_2289 = const()[name = tensor("op_2289"), val = tensor([1, -1, 8, 64])]; - tensor k_45_cast_fp16 = reshape(shape = var_2289, x = linear_103_cast_fp16)[name = tensor("k_45_cast_fp16")]; - tensor module_layers_11_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_11_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145489536)))]; - tensor module_layers_11_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_11_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146013888)))]; - tensor linear_104_cast_fp16 = linear(bias = module_layers_11_self_attn_linear_v_bias_to_fp16, weight = module_layers_11_self_attn_linear_v_weight_to_fp16, x = query_23_cast_fp16)[name = tensor("linear_104_cast_fp16")]; - tensor var_2294 = const()[name = tensor("op_2294"), val = tensor([1, -1, 8, 64])]; - tensor v_23_cast_fp16 = reshape(shape = var_2294, x = linear_104_cast_fp16)[name = tensor("v_23_cast_fp16")]; - tensor value_25_perm_0 = const()[name = tensor("value_25_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_11_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_11_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146014976)))]; - tensor var_2306_cast_fp16 = add(x = q_67_cast_fp16, y = module_layers_11_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2306_cast_fp16")]; - tensor module_layers_11_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_11_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146016064)))]; - tensor var_2308_cast_fp16 = add(x = q_67_cast_fp16, y = module_layers_11_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2308_cast_fp16")]; - tensor q_with_bias_v_23_perm_0 = const()[name = tensor("q_with_bias_v_23_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_249_transpose_x_0 = const()[name = tensor("x_249_transpose_x_0"), val = tensor(false)]; - tensor x_249_transpose_y_0 = const()[name = tensor("x_249_transpose_y_0"), val = tensor(false)]; - tensor var_2310_to_fp16 = const()[name = tensor("op_2310_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146017152)))]; - tensor q_with_bias_v_23_cast_fp16 = transpose(perm = q_with_bias_v_23_perm_0, x = var_2308_cast_fp16)[name = tensor("transpose_126")]; - tensor x_249_cast_fp16 = matmul(transpose_x = x_249_transpose_x_0, transpose_y = x_249_transpose_y_0, x = q_with_bias_v_23_cast_fp16, y = var_2310_to_fp16)[name = tensor("x_249_cast_fp16")]; - tensor x_251_pad_0 = const()[name = tensor("x_251_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_251_mode_0 = const()[name = tensor("x_251_mode_0"), val = tensor("constant")]; - tensor const_180_to_fp16 = const()[name = tensor("const_180_to_fp16"), val = tensor(0x0p+0)]; - tensor x_251_cast_fp16 = pad(constant_val = const_180_to_fp16, mode = x_251_mode_0, pad = x_251_pad_0, x = x_249_cast_fp16)[name = tensor("x_251_cast_fp16")]; - tensor var_2318 = const()[name = tensor("op_2318"), val = tensor([1, 8, -1, 188])]; - tensor x_253_cast_fp16 = reshape(shape = var_2318, x = x_251_cast_fp16)[name = tensor("x_253_cast_fp16")]; - tensor var_2322_begin_0 = const()[name = tensor("op_2322_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_2322_end_0 = const()[name = tensor("op_2322_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_2322_end_mask_0 = const()[name = tensor("op_2322_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_2322_cast_fp16 = slice_by_index(begin = var_2322_begin_0, end = var_2322_end_0, end_mask = var_2322_end_mask_0, x = x_253_cast_fp16)[name = tensor("op_2322_cast_fp16")]; - tensor var_2323 = const()[name = tensor("op_2323"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_45_cast_fp16 = reshape(shape = var_2323, x = var_2322_cast_fp16)[name = tensor("matrix_bd_45_cast_fp16")]; - tensor matrix_ac_23_transpose_x_0 = const()[name = tensor("matrix_ac_23_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_23_transpose_y_0 = const()[name = tensor("matrix_ac_23_transpose_y_0"), val = tensor(false)]; - tensor transpose_73_perm_0 = const()[name = tensor("transpose_73_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_74_perm_0 = const()[name = tensor("transpose_74_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_74 = transpose(perm = transpose_74_perm_0, x = k_45_cast_fp16)[name = tensor("transpose_124")]; - tensor transpose_73 = transpose(perm = transpose_73_perm_0, x = var_2306_cast_fp16)[name = tensor("transpose_125")]; - tensor matrix_ac_23_cast_fp16 = matmul(transpose_x = matrix_ac_23_transpose_x_0, transpose_y = matrix_ac_23_transpose_y_0, x = transpose_73, y = transpose_74)[name = tensor("matrix_ac_23_cast_fp16")]; - tensor matrix_bd_47_begin_0 = const()[name = tensor("matrix_bd_47_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_47_end_0 = const()[name = tensor("matrix_bd_47_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_47_end_mask_0 = const()[name = tensor("matrix_bd_47_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_47_cast_fp16 = slice_by_index(begin = matrix_bd_47_begin_0, end = matrix_bd_47_end_0, end_mask = matrix_bd_47_end_mask_0, x = matrix_bd_45_cast_fp16)[name = tensor("matrix_bd_47_cast_fp16")]; - tensor var_2332_cast_fp16 = add(x = matrix_ac_23_cast_fp16, y = matrix_bd_47_cast_fp16)[name = tensor("op_2332_cast_fp16")]; - tensor _inversed_scores_45_y_0_to_fp16 = const()[name = tensor("_inversed_scores_45_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_45_cast_fp16 = mul(x = var_2332_cast_fp16, y = _inversed_scores_45_y_0_to_fp16)[name = tensor("_inversed_scores_45_cast_fp16")]; - tensor scores_47_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_45_cast_fp16, cond = mask_11)[name = tensor("scores_47_cast_fp16")]; - tensor var_2338_cast_fp16 = softmax(axis = var_23, x = scores_47_cast_fp16)[name = tensor("op_2338_cast_fp16")]; - tensor input_605_cast_fp16 = select(a = var_11_to_fp16, b = var_2338_cast_fp16, cond = mask_11)[name = tensor("input_605_cast_fp16")]; - tensor x_255_transpose_x_0 = const()[name = tensor("x_255_transpose_x_0"), val = tensor(false)]; - tensor x_255_transpose_y_0 = const()[name = tensor("x_255_transpose_y_0"), val = tensor(false)]; - tensor value_25_cast_fp16 = transpose(perm = value_25_perm_0, x = v_23_cast_fp16)[name = tensor("transpose_127")]; - tensor x_255_cast_fp16 = matmul(transpose_x = x_255_transpose_x_0, transpose_y = x_255_transpose_y_0, x = input_605_cast_fp16, y = value_25_cast_fp16)[name = tensor("x_255_cast_fp16")]; - tensor var_2342_perm_0 = const()[name = tensor("op_2342_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_2343 = const()[name = tensor("op_2343"), val = tensor([1, -1, 512])]; - tensor var_2342_cast_fp16 = transpose(perm = var_2342_perm_0, x = x_255_cast_fp16)[name = tensor("transpose_123")]; - tensor input_607_cast_fp16 = reshape(shape = var_2343, x = var_2342_cast_fp16)[name = tensor("input_607_cast_fp16")]; - tensor module_layers_11_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_11_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146401216)))]; - tensor module_layers_11_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_11_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146925568)))]; - tensor linear_106_cast_fp16 = linear(bias = module_layers_11_self_attn_linear_out_bias_to_fp16, weight = module_layers_11_self_attn_linear_out_weight_to_fp16, x = input_607_cast_fp16)[name = tensor("linear_106_cast_fp16")]; - tensor input_611_cast_fp16 = add(x = input_603_cast_fp16, y = linear_106_cast_fp16)[name = tensor("input_611_cast_fp16")]; - tensor x_259_axes_0 = const()[name = tensor("x_259_axes_0"), val = tensor([-1])]; - tensor module_layers_11_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_11_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146926656)))]; - tensor module_layers_11_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_11_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146927744)))]; - tensor x_259_cast_fp16 = layer_norm(axes = x_259_axes_0, beta = module_layers_11_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_11_norm_conv_weight_to_fp16, x = input_611_cast_fp16)[name = tensor("x_259_cast_fp16")]; - tensor input_613_perm_0 = const()[name = tensor("input_613_perm_0"), val = tensor([0, 2, 1])]; - tensor input_615_pad_type_0 = const()[name = tensor("input_615_pad_type_0"), val = tensor("valid")]; - tensor input_615_strides_0 = const()[name = tensor("input_615_strides_0"), val = tensor([1])]; - tensor input_615_pad_0 = const()[name = tensor("input_615_pad_0"), val = tensor([0, 0])]; - tensor input_615_dilations_0 = const()[name = tensor("input_615_dilations_0"), val = tensor([1])]; - tensor input_615_groups_0 = const()[name = tensor("input_615_groups_0"), val = tensor(1)]; - tensor module_layers_11_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_11_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146928832)))]; - tensor module_layers_11_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_11_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147977472)))]; - tensor input_613_cast_fp16 = transpose(perm = input_613_perm_0, x = x_259_cast_fp16)[name = tensor("transpose_122")]; - tensor input_615_cast_fp16 = conv(bias = module_layers_11_conv_pointwise_conv1_bias_to_fp16, dilations = input_615_dilations_0, groups = input_615_groups_0, pad = input_615_pad_0, pad_type = input_615_pad_type_0, strides = input_615_strides_0, weight = module_layers_11_conv_pointwise_conv1_weight_to_fp16, x = input_613_cast_fp16)[name = tensor("input_615_cast_fp16")]; - tensor x_261_split_num_splits_0 = const()[name = tensor("x_261_split_num_splits_0"), val = tensor(2)]; - tensor x_261_split_axis_0 = const()[name = tensor("x_261_split_axis_0"), val = tensor(1)]; - tensor x_261_split_cast_fp16_0, tensor x_261_split_cast_fp16_1 = split(axis = x_261_split_axis_0, num_splits = x_261_split_num_splits_0, x = input_615_cast_fp16)[name = tensor("x_261_split_cast_fp16")]; - tensor x_261_split_1_sigmoid_cast_fp16 = sigmoid(x = x_261_split_cast_fp16_1)[name = tensor("x_261_split_1_sigmoid_cast_fp16")]; - tensor x_261_cast_fp16 = mul(x = x_261_split_cast_fp16_0, y = x_261_split_1_sigmoid_cast_fp16)[name = tensor("x_261_cast_fp16")]; - tensor input_617_cast_fp16 = select(a = var_11_to_fp16, b = x_261_cast_fp16, cond = var_453)[name = tensor("input_617_cast_fp16")]; - tensor input_619_pad_0 = const()[name = tensor("input_619_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_619_mode_0 = const()[name = tensor("input_619_mode_0"), val = tensor("constant")]; - tensor const_183_to_fp16 = const()[name = tensor("const_183_to_fp16"), val = tensor(0x0p+0)]; - tensor input_619_cast_fp16 = pad(constant_val = const_183_to_fp16, mode = input_619_mode_0, pad = input_619_pad_0, x = input_617_cast_fp16)[name = tensor("input_619_cast_fp16")]; - tensor input_621_pad_type_0 = const()[name = tensor("input_621_pad_type_0"), val = tensor("valid")]; - tensor input_621_groups_0 = const()[name = tensor("input_621_groups_0"), val = tensor(512)]; - tensor input_621_strides_0 = const()[name = tensor("input_621_strides_0"), val = tensor([1])]; - tensor input_621_pad_0 = const()[name = tensor("input_621_pad_0"), val = tensor([0, 0])]; - tensor input_621_dilations_0 = const()[name = tensor("input_621_dilations_0"), val = tensor([1])]; - tensor const_256_to_fp16 = const()[name = tensor("const_256_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147979584)))]; - tensor const_257_to_fp16 = const()[name = tensor("const_257_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147988864)))]; - tensor input_623_cast_fp16 = conv(bias = const_257_to_fp16, dilations = input_621_dilations_0, groups = input_621_groups_0, pad = input_621_pad_0, pad_type = input_621_pad_type_0, strides = input_621_strides_0, weight = const_256_to_fp16, x = input_619_cast_fp16)[name = tensor("input_623_cast_fp16")]; - tensor input_625_cast_fp16 = silu(x = input_623_cast_fp16)[name = tensor("input_625_cast_fp16")]; - tensor x_263_pad_type_0 = const()[name = tensor("x_263_pad_type_0"), val = tensor("valid")]; - tensor x_263_strides_0 = const()[name = tensor("x_263_strides_0"), val = tensor([1])]; - tensor x_263_pad_0 = const()[name = tensor("x_263_pad_0"), val = tensor([0, 0])]; - tensor x_263_dilations_0 = const()[name = tensor("x_263_dilations_0"), val = tensor([1])]; - tensor x_263_groups_0 = const()[name = tensor("x_263_groups_0"), val = tensor(1)]; - tensor module_layers_11_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_11_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147989952)))]; - tensor module_layers_11_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_11_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(148514304)))]; - tensor x_263_cast_fp16 = conv(bias = module_layers_11_conv_pointwise_conv2_bias_to_fp16, dilations = x_263_dilations_0, groups = x_263_groups_0, pad = x_263_pad_0, pad_type = x_263_pad_type_0, strides = x_263_strides_0, weight = module_layers_11_conv_pointwise_conv2_weight_to_fp16, x = input_625_cast_fp16)[name = tensor("x_263_cast_fp16")]; - tensor input_627_perm_0 = const()[name = tensor("input_627_perm_0"), val = tensor([0, 2, 1])]; - tensor input_627_cast_fp16 = transpose(perm = input_627_perm_0, x = x_263_cast_fp16)[name = tensor("transpose_121")]; - tensor input_629_cast_fp16 = add(x = input_611_cast_fp16, y = input_627_cast_fp16)[name = tensor("input_629_cast_fp16")]; - tensor input_631_axes_0 = const()[name = tensor("input_631_axes_0"), val = tensor([-1])]; - tensor module_layers_11_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_11_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(148515392)))]; - tensor module_layers_11_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_11_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(148516480)))]; - tensor input_631_cast_fp16 = layer_norm(axes = input_631_axes_0, beta = module_layers_11_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_11_norm_feed_forward2_weight_to_fp16, x = input_629_cast_fp16)[name = tensor("input_631_cast_fp16")]; - tensor module_layers_11_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_11_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(148517568)))]; - tensor module_layers_11_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_11_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(150614784)))]; - tensor linear_107_cast_fp16 = linear(bias = module_layers_11_feed_forward2_linear1_bias_to_fp16, weight = module_layers_11_feed_forward2_linear1_weight_to_fp16, x = input_631_cast_fp16)[name = tensor("linear_107_cast_fp16")]; - tensor input_635_cast_fp16 = silu(x = linear_107_cast_fp16)[name = tensor("input_635_cast_fp16")]; - tensor module_layers_11_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_11_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(150618944)))]; - tensor module_layers_11_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_11_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(152716160)))]; - tensor linear_108_cast_fp16 = linear(bias = module_layers_11_feed_forward2_linear2_bias_to_fp16, weight = module_layers_11_feed_forward2_linear2_weight_to_fp16, x = input_635_cast_fp16)[name = tensor("linear_108_cast_fp16")]; - tensor var_2409_to_fp16 = const()[name = tensor("op_2409_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2410_cast_fp16 = mul(x = linear_108_cast_fp16, y = var_2409_to_fp16)[name = tensor("op_2410_cast_fp16")]; - tensor input_641_cast_fp16 = add(x = input_629_cast_fp16, y = var_2410_cast_fp16)[name = tensor("input_641_cast_fp16")]; - tensor input_643_axes_0 = const()[name = tensor("input_643_axes_0"), val = tensor([-1])]; - tensor module_layers_11_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_11_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(152717248)))]; - tensor module_layers_11_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_11_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(152718336)))]; - tensor input_643_cast_fp16 = layer_norm(axes = input_643_axes_0, beta = module_layers_11_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_11_norm_out_weight_to_fp16, x = input_641_cast_fp16)[name = tensor("input_643_cast_fp16")]; - tensor input_645_axes_0 = const()[name = tensor("input_645_axes_0"), val = tensor([-1])]; - tensor module_layers_12_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_12_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(152719424)))]; - tensor module_layers_12_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_12_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(152720512)))]; - tensor input_645_cast_fp16 = layer_norm(axes = input_645_axes_0, beta = module_layers_12_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_12_norm_feed_forward1_weight_to_fp16, x = input_643_cast_fp16)[name = tensor("input_645_cast_fp16")]; - tensor module_layers_12_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_12_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(152721600)))]; - tensor module_layers_12_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_12_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(154818816)))]; - tensor linear_109_cast_fp16 = linear(bias = module_layers_12_feed_forward1_linear1_bias_to_fp16, weight = module_layers_12_feed_forward1_linear1_weight_to_fp16, x = input_645_cast_fp16)[name = tensor("linear_109_cast_fp16")]; - tensor input_649_cast_fp16 = silu(x = linear_109_cast_fp16)[name = tensor("input_649_cast_fp16")]; - tensor module_layers_12_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_12_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(154822976)))]; - tensor module_layers_12_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_12_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(156920192)))]; - tensor linear_110_cast_fp16 = linear(bias = module_layers_12_feed_forward1_linear2_bias_to_fp16, weight = module_layers_12_feed_forward1_linear2_weight_to_fp16, x = input_649_cast_fp16)[name = tensor("linear_110_cast_fp16")]; - tensor var_2440_to_fp16 = const()[name = tensor("op_2440_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2441_cast_fp16 = mul(x = linear_110_cast_fp16, y = var_2440_to_fp16)[name = tensor("op_2441_cast_fp16")]; - tensor input_655_cast_fp16 = add(x = input_643_cast_fp16, y = var_2441_cast_fp16)[name = tensor("input_655_cast_fp16")]; - tensor query_25_axes_0 = const()[name = tensor("query_25_axes_0"), val = tensor([-1])]; - tensor module_layers_12_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_12_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(156921280)))]; - tensor module_layers_12_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_12_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(156922368)))]; - tensor query_25_cast_fp16 = layer_norm(axes = query_25_axes_0, beta = module_layers_12_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_12_norm_self_att_weight_to_fp16, x = input_655_cast_fp16)[name = tensor("query_25_cast_fp16")]; - tensor module_layers_12_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_12_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(156923456)))]; - tensor module_layers_12_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_12_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157447808)))]; - tensor linear_111_cast_fp16 = linear(bias = module_layers_12_self_attn_linear_q_bias_to_fp16, weight = module_layers_12_self_attn_linear_q_weight_to_fp16, x = query_25_cast_fp16)[name = tensor("linear_111_cast_fp16")]; - tensor var_2458 = const()[name = tensor("op_2458"), val = tensor([1, -1, 8, 64])]; - tensor q_73_cast_fp16 = reshape(shape = var_2458, x = linear_111_cast_fp16)[name = tensor("q_73_cast_fp16")]; - tensor module_layers_12_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_12_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157448896)))]; - tensor module_layers_12_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_12_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157973248)))]; - tensor linear_112_cast_fp16 = linear(bias = module_layers_12_self_attn_linear_k_bias_to_fp16, weight = module_layers_12_self_attn_linear_k_weight_to_fp16, x = query_25_cast_fp16)[name = tensor("linear_112_cast_fp16")]; - tensor var_2463 = const()[name = tensor("op_2463"), val = tensor([1, -1, 8, 64])]; - tensor k_49_cast_fp16 = reshape(shape = var_2463, x = linear_112_cast_fp16)[name = tensor("k_49_cast_fp16")]; - tensor module_layers_12_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_12_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157974336)))]; - tensor module_layers_12_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_12_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158498688)))]; - tensor linear_113_cast_fp16 = linear(bias = module_layers_12_self_attn_linear_v_bias_to_fp16, weight = module_layers_12_self_attn_linear_v_weight_to_fp16, x = query_25_cast_fp16)[name = tensor("linear_113_cast_fp16")]; - tensor var_2468 = const()[name = tensor("op_2468"), val = tensor([1, -1, 8, 64])]; - tensor v_25_cast_fp16 = reshape(shape = var_2468, x = linear_113_cast_fp16)[name = tensor("v_25_cast_fp16")]; - tensor value_27_perm_0 = const()[name = tensor("value_27_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_12_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_12_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158499776)))]; - tensor var_2480_cast_fp16 = add(x = q_73_cast_fp16, y = module_layers_12_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2480_cast_fp16")]; - tensor module_layers_12_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_12_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158500864)))]; - tensor var_2482_cast_fp16 = add(x = q_73_cast_fp16, y = module_layers_12_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2482_cast_fp16")]; - tensor q_with_bias_v_25_perm_0 = const()[name = tensor("q_with_bias_v_25_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_271_transpose_x_0 = const()[name = tensor("x_271_transpose_x_0"), val = tensor(false)]; - tensor x_271_transpose_y_0 = const()[name = tensor("x_271_transpose_y_0"), val = tensor(false)]; - tensor var_2484_to_fp16 = const()[name = tensor("op_2484_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158501952)))]; - tensor q_with_bias_v_25_cast_fp16 = transpose(perm = q_with_bias_v_25_perm_0, x = var_2482_cast_fp16)[name = tensor("transpose_119")]; - tensor x_271_cast_fp16 = matmul(transpose_x = x_271_transpose_x_0, transpose_y = x_271_transpose_y_0, x = q_with_bias_v_25_cast_fp16, y = var_2484_to_fp16)[name = tensor("x_271_cast_fp16")]; - tensor x_273_pad_0 = const()[name = tensor("x_273_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_273_mode_0 = const()[name = tensor("x_273_mode_0"), val = tensor("constant")]; - tensor const_190_to_fp16 = const()[name = tensor("const_190_to_fp16"), val = tensor(0x0p+0)]; - tensor x_273_cast_fp16 = pad(constant_val = const_190_to_fp16, mode = x_273_mode_0, pad = x_273_pad_0, x = x_271_cast_fp16)[name = tensor("x_273_cast_fp16")]; - tensor var_2492 = const()[name = tensor("op_2492"), val = tensor([1, 8, -1, 188])]; - tensor x_275_cast_fp16 = reshape(shape = var_2492, x = x_273_cast_fp16)[name = tensor("x_275_cast_fp16")]; - tensor var_2496_begin_0 = const()[name = tensor("op_2496_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_2496_end_0 = const()[name = tensor("op_2496_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_2496_end_mask_0 = const()[name = tensor("op_2496_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_2496_cast_fp16 = slice_by_index(begin = var_2496_begin_0, end = var_2496_end_0, end_mask = var_2496_end_mask_0, x = x_275_cast_fp16)[name = tensor("op_2496_cast_fp16")]; - tensor var_2497 = const()[name = tensor("op_2497"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_49_cast_fp16 = reshape(shape = var_2497, x = var_2496_cast_fp16)[name = tensor("matrix_bd_49_cast_fp16")]; - tensor matrix_ac_25_transpose_x_0 = const()[name = tensor("matrix_ac_25_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_25_transpose_y_0 = const()[name = tensor("matrix_ac_25_transpose_y_0"), val = tensor(false)]; - tensor transpose_75_perm_0 = const()[name = tensor("transpose_75_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_76_perm_0 = const()[name = tensor("transpose_76_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_76 = transpose(perm = transpose_76_perm_0, x = k_49_cast_fp16)[name = tensor("transpose_117")]; - tensor transpose_75 = transpose(perm = transpose_75_perm_0, x = var_2480_cast_fp16)[name = tensor("transpose_118")]; - tensor matrix_ac_25_cast_fp16 = matmul(transpose_x = matrix_ac_25_transpose_x_0, transpose_y = matrix_ac_25_transpose_y_0, x = transpose_75, y = transpose_76)[name = tensor("matrix_ac_25_cast_fp16")]; - tensor matrix_bd_51_begin_0 = const()[name = tensor("matrix_bd_51_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_51_end_0 = const()[name = tensor("matrix_bd_51_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_51_end_mask_0 = const()[name = tensor("matrix_bd_51_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_51_cast_fp16 = slice_by_index(begin = matrix_bd_51_begin_0, end = matrix_bd_51_end_0, end_mask = matrix_bd_51_end_mask_0, x = matrix_bd_49_cast_fp16)[name = tensor("matrix_bd_51_cast_fp16")]; - tensor var_2506_cast_fp16 = add(x = matrix_ac_25_cast_fp16, y = matrix_bd_51_cast_fp16)[name = tensor("op_2506_cast_fp16")]; - tensor _inversed_scores_49_y_0_to_fp16 = const()[name = tensor("_inversed_scores_49_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_49_cast_fp16 = mul(x = var_2506_cast_fp16, y = _inversed_scores_49_y_0_to_fp16)[name = tensor("_inversed_scores_49_cast_fp16")]; - tensor scores_51_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_49_cast_fp16, cond = mask_11)[name = tensor("scores_51_cast_fp16")]; - tensor var_2512_cast_fp16 = softmax(axis = var_23, x = scores_51_cast_fp16)[name = tensor("op_2512_cast_fp16")]; - tensor input_657_cast_fp16 = select(a = var_11_to_fp16, b = var_2512_cast_fp16, cond = mask_11)[name = tensor("input_657_cast_fp16")]; - tensor x_277_transpose_x_0 = const()[name = tensor("x_277_transpose_x_0"), val = tensor(false)]; - tensor x_277_transpose_y_0 = const()[name = tensor("x_277_transpose_y_0"), val = tensor(false)]; - tensor value_27_cast_fp16 = transpose(perm = value_27_perm_0, x = v_25_cast_fp16)[name = tensor("transpose_120")]; - tensor x_277_cast_fp16 = matmul(transpose_x = x_277_transpose_x_0, transpose_y = x_277_transpose_y_0, x = input_657_cast_fp16, y = value_27_cast_fp16)[name = tensor("x_277_cast_fp16")]; - tensor var_2516_perm_0 = const()[name = tensor("op_2516_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_2517 = const()[name = tensor("op_2517"), val = tensor([1, -1, 512])]; - tensor var_2516_cast_fp16 = transpose(perm = var_2516_perm_0, x = x_277_cast_fp16)[name = tensor("transpose_116")]; - tensor input_659_cast_fp16 = reshape(shape = var_2517, x = var_2516_cast_fp16)[name = tensor("input_659_cast_fp16")]; - tensor module_layers_12_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_12_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158886016)))]; - tensor module_layers_12_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_12_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159410368)))]; - tensor linear_115_cast_fp16 = linear(bias = module_layers_12_self_attn_linear_out_bias_to_fp16, weight = module_layers_12_self_attn_linear_out_weight_to_fp16, x = input_659_cast_fp16)[name = tensor("linear_115_cast_fp16")]; - tensor input_663_cast_fp16 = add(x = input_655_cast_fp16, y = linear_115_cast_fp16)[name = tensor("input_663_cast_fp16")]; - tensor x_281_axes_0 = const()[name = tensor("x_281_axes_0"), val = tensor([-1])]; - tensor module_layers_12_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_12_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159411456)))]; - tensor module_layers_12_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_12_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159412544)))]; - tensor x_281_cast_fp16 = layer_norm(axes = x_281_axes_0, beta = module_layers_12_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_12_norm_conv_weight_to_fp16, x = input_663_cast_fp16)[name = tensor("x_281_cast_fp16")]; - tensor input_665_perm_0 = const()[name = tensor("input_665_perm_0"), val = tensor([0, 2, 1])]; - tensor input_667_pad_type_0 = const()[name = tensor("input_667_pad_type_0"), val = tensor("valid")]; - tensor input_667_strides_0 = const()[name = tensor("input_667_strides_0"), val = tensor([1])]; - tensor input_667_pad_0 = const()[name = tensor("input_667_pad_0"), val = tensor([0, 0])]; - tensor input_667_dilations_0 = const()[name = tensor("input_667_dilations_0"), val = tensor([1])]; - tensor input_667_groups_0 = const()[name = tensor("input_667_groups_0"), val = tensor(1)]; - tensor module_layers_12_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_12_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159413632)))]; - tensor module_layers_12_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_12_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160462272)))]; - tensor input_665_cast_fp16 = transpose(perm = input_665_perm_0, x = x_281_cast_fp16)[name = tensor("transpose_115")]; - tensor input_667_cast_fp16 = conv(bias = module_layers_12_conv_pointwise_conv1_bias_to_fp16, dilations = input_667_dilations_0, groups = input_667_groups_0, pad = input_667_pad_0, pad_type = input_667_pad_type_0, strides = input_667_strides_0, weight = module_layers_12_conv_pointwise_conv1_weight_to_fp16, x = input_665_cast_fp16)[name = tensor("input_667_cast_fp16")]; - tensor x_283_split_num_splits_0 = const()[name = tensor("x_283_split_num_splits_0"), val = tensor(2)]; - tensor x_283_split_axis_0 = const()[name = tensor("x_283_split_axis_0"), val = tensor(1)]; - tensor x_283_split_cast_fp16_0, tensor x_283_split_cast_fp16_1 = split(axis = x_283_split_axis_0, num_splits = x_283_split_num_splits_0, x = input_667_cast_fp16)[name = tensor("x_283_split_cast_fp16")]; - tensor x_283_split_1_sigmoid_cast_fp16 = sigmoid(x = x_283_split_cast_fp16_1)[name = tensor("x_283_split_1_sigmoid_cast_fp16")]; - tensor x_283_cast_fp16 = mul(x = x_283_split_cast_fp16_0, y = x_283_split_1_sigmoid_cast_fp16)[name = tensor("x_283_cast_fp16")]; - tensor input_669_cast_fp16 = select(a = var_11_to_fp16, b = x_283_cast_fp16, cond = var_453)[name = tensor("input_669_cast_fp16")]; - tensor input_671_pad_0 = const()[name = tensor("input_671_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_671_mode_0 = const()[name = tensor("input_671_mode_0"), val = tensor("constant")]; - tensor const_193_to_fp16 = const()[name = tensor("const_193_to_fp16"), val = tensor(0x0p+0)]; - tensor input_671_cast_fp16 = pad(constant_val = const_193_to_fp16, mode = input_671_mode_0, pad = input_671_pad_0, x = input_669_cast_fp16)[name = tensor("input_671_cast_fp16")]; - tensor input_673_pad_type_0 = const()[name = tensor("input_673_pad_type_0"), val = tensor("valid")]; - tensor input_673_groups_0 = const()[name = tensor("input_673_groups_0"), val = tensor(512)]; - tensor input_673_strides_0 = const()[name = tensor("input_673_strides_0"), val = tensor([1])]; - tensor input_673_pad_0 = const()[name = tensor("input_673_pad_0"), val = tensor([0, 0])]; - tensor input_673_dilations_0 = const()[name = tensor("input_673_dilations_0"), val = tensor([1])]; - tensor const_258_to_fp16 = const()[name = tensor("const_258_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160464384)))]; - tensor const_259_to_fp16 = const()[name = tensor("const_259_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160473664)))]; - tensor input_675_cast_fp16 = conv(bias = const_259_to_fp16, dilations = input_673_dilations_0, groups = input_673_groups_0, pad = input_673_pad_0, pad_type = input_673_pad_type_0, strides = input_673_strides_0, weight = const_258_to_fp16, x = input_671_cast_fp16)[name = tensor("input_675_cast_fp16")]; - tensor input_677_cast_fp16 = silu(x = input_675_cast_fp16)[name = tensor("input_677_cast_fp16")]; - tensor x_285_pad_type_0 = const()[name = tensor("x_285_pad_type_0"), val = tensor("valid")]; - tensor x_285_strides_0 = const()[name = tensor("x_285_strides_0"), val = tensor([1])]; - tensor x_285_pad_0 = const()[name = tensor("x_285_pad_0"), val = tensor([0, 0])]; - tensor x_285_dilations_0 = const()[name = tensor("x_285_dilations_0"), val = tensor([1])]; - tensor x_285_groups_0 = const()[name = tensor("x_285_groups_0"), val = tensor(1)]; - tensor module_layers_12_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_12_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160474752)))]; - tensor module_layers_12_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_12_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160999104)))]; - tensor x_285_cast_fp16 = conv(bias = module_layers_12_conv_pointwise_conv2_bias_to_fp16, dilations = x_285_dilations_0, groups = x_285_groups_0, pad = x_285_pad_0, pad_type = x_285_pad_type_0, strides = x_285_strides_0, weight = module_layers_12_conv_pointwise_conv2_weight_to_fp16, x = input_677_cast_fp16)[name = tensor("x_285_cast_fp16")]; - tensor input_679_perm_0 = const()[name = tensor("input_679_perm_0"), val = tensor([0, 2, 1])]; - tensor input_679_cast_fp16 = transpose(perm = input_679_perm_0, x = x_285_cast_fp16)[name = tensor("transpose_114")]; - tensor input_681_cast_fp16 = add(x = input_663_cast_fp16, y = input_679_cast_fp16)[name = tensor("input_681_cast_fp16")]; - tensor input_683_axes_0 = const()[name = tensor("input_683_axes_0"), val = tensor([-1])]; - tensor module_layers_12_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_12_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161000192)))]; - tensor module_layers_12_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_12_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161001280)))]; - tensor input_683_cast_fp16 = layer_norm(axes = input_683_axes_0, beta = module_layers_12_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_12_norm_feed_forward2_weight_to_fp16, x = input_681_cast_fp16)[name = tensor("input_683_cast_fp16")]; - tensor module_layers_12_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_12_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161002368)))]; - tensor module_layers_12_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_12_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(163099584)))]; - tensor linear_116_cast_fp16 = linear(bias = module_layers_12_feed_forward2_linear1_bias_to_fp16, weight = module_layers_12_feed_forward2_linear1_weight_to_fp16, x = input_683_cast_fp16)[name = tensor("linear_116_cast_fp16")]; - tensor input_687_cast_fp16 = silu(x = linear_116_cast_fp16)[name = tensor("input_687_cast_fp16")]; - tensor module_layers_12_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_12_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(163103744)))]; - tensor module_layers_12_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_12_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165200960)))]; - tensor linear_117_cast_fp16 = linear(bias = module_layers_12_feed_forward2_linear2_bias_to_fp16, weight = module_layers_12_feed_forward2_linear2_weight_to_fp16, x = input_687_cast_fp16)[name = tensor("linear_117_cast_fp16")]; - tensor var_2583_to_fp16 = const()[name = tensor("op_2583_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2584_cast_fp16 = mul(x = linear_117_cast_fp16, y = var_2583_to_fp16)[name = tensor("op_2584_cast_fp16")]; - tensor input_693_cast_fp16 = add(x = input_681_cast_fp16, y = var_2584_cast_fp16)[name = tensor("input_693_cast_fp16")]; - tensor input_695_axes_0 = const()[name = tensor("input_695_axes_0"), val = tensor([-1])]; - tensor module_layers_12_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_12_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165202048)))]; - tensor module_layers_12_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_12_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165203136)))]; - tensor input_695_cast_fp16 = layer_norm(axes = input_695_axes_0, beta = module_layers_12_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_12_norm_out_weight_to_fp16, x = input_693_cast_fp16)[name = tensor("input_695_cast_fp16")]; - tensor input_697_axes_0 = const()[name = tensor("input_697_axes_0"), val = tensor([-1])]; - tensor module_layers_13_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_13_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165204224)))]; - tensor module_layers_13_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_13_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165205312)))]; - tensor input_697_cast_fp16 = layer_norm(axes = input_697_axes_0, beta = module_layers_13_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_13_norm_feed_forward1_weight_to_fp16, x = input_695_cast_fp16)[name = tensor("input_697_cast_fp16")]; - tensor module_layers_13_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_13_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165206400)))]; - tensor module_layers_13_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_13_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(167303616)))]; - tensor linear_118_cast_fp16 = linear(bias = module_layers_13_feed_forward1_linear1_bias_to_fp16, weight = module_layers_13_feed_forward1_linear1_weight_to_fp16, x = input_697_cast_fp16)[name = tensor("linear_118_cast_fp16")]; - tensor input_701_cast_fp16 = silu(x = linear_118_cast_fp16)[name = tensor("input_701_cast_fp16")]; - tensor module_layers_13_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_13_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(167307776)))]; - tensor module_layers_13_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_13_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169404992)))]; - tensor linear_119_cast_fp16 = linear(bias = module_layers_13_feed_forward1_linear2_bias_to_fp16, weight = module_layers_13_feed_forward1_linear2_weight_to_fp16, x = input_701_cast_fp16)[name = tensor("linear_119_cast_fp16")]; - tensor var_2614_to_fp16 = const()[name = tensor("op_2614_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2615_cast_fp16 = mul(x = linear_119_cast_fp16, y = var_2614_to_fp16)[name = tensor("op_2615_cast_fp16")]; - tensor input_707_cast_fp16 = add(x = input_695_cast_fp16, y = var_2615_cast_fp16)[name = tensor("input_707_cast_fp16")]; - tensor query_27_axes_0 = const()[name = tensor("query_27_axes_0"), val = tensor([-1])]; - tensor module_layers_13_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_13_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169406080)))]; - tensor module_layers_13_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_13_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169407168)))]; - tensor query_27_cast_fp16 = layer_norm(axes = query_27_axes_0, beta = module_layers_13_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_13_norm_self_att_weight_to_fp16, x = input_707_cast_fp16)[name = tensor("query_27_cast_fp16")]; - tensor module_layers_13_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_13_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169408256)))]; - tensor module_layers_13_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_13_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169932608)))]; - tensor linear_120_cast_fp16 = linear(bias = module_layers_13_self_attn_linear_q_bias_to_fp16, weight = module_layers_13_self_attn_linear_q_weight_to_fp16, x = query_27_cast_fp16)[name = tensor("linear_120_cast_fp16")]; - tensor var_2632 = const()[name = tensor("op_2632"), val = tensor([1, -1, 8, 64])]; - tensor q_79_cast_fp16 = reshape(shape = var_2632, x = linear_120_cast_fp16)[name = tensor("q_79_cast_fp16")]; - tensor module_layers_13_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_13_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169933696)))]; - tensor module_layers_13_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_13_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170458048)))]; - tensor linear_121_cast_fp16 = linear(bias = module_layers_13_self_attn_linear_k_bias_to_fp16, weight = module_layers_13_self_attn_linear_k_weight_to_fp16, x = query_27_cast_fp16)[name = tensor("linear_121_cast_fp16")]; - tensor var_2637 = const()[name = tensor("op_2637"), val = tensor([1, -1, 8, 64])]; - tensor k_53_cast_fp16 = reshape(shape = var_2637, x = linear_121_cast_fp16)[name = tensor("k_53_cast_fp16")]; - tensor module_layers_13_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_13_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170459136)))]; - tensor module_layers_13_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_13_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170983488)))]; - tensor linear_122_cast_fp16 = linear(bias = module_layers_13_self_attn_linear_v_bias_to_fp16, weight = module_layers_13_self_attn_linear_v_weight_to_fp16, x = query_27_cast_fp16)[name = tensor("linear_122_cast_fp16")]; - tensor var_2642 = const()[name = tensor("op_2642"), val = tensor([1, -1, 8, 64])]; - tensor v_27_cast_fp16 = reshape(shape = var_2642, x = linear_122_cast_fp16)[name = tensor("v_27_cast_fp16")]; - tensor value_29_perm_0 = const()[name = tensor("value_29_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_13_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_13_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170984576)))]; - tensor var_2654_cast_fp16 = add(x = q_79_cast_fp16, y = module_layers_13_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2654_cast_fp16")]; - tensor module_layers_13_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_13_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170985664)))]; - tensor var_2656_cast_fp16 = add(x = q_79_cast_fp16, y = module_layers_13_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2656_cast_fp16")]; - tensor q_with_bias_v_27_perm_0 = const()[name = tensor("q_with_bias_v_27_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_293_transpose_x_0 = const()[name = tensor("x_293_transpose_x_0"), val = tensor(false)]; - tensor x_293_transpose_y_0 = const()[name = tensor("x_293_transpose_y_0"), val = tensor(false)]; - tensor var_2658_to_fp16 = const()[name = tensor("op_2658_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170986752)))]; - tensor q_with_bias_v_27_cast_fp16 = transpose(perm = q_with_bias_v_27_perm_0, x = var_2656_cast_fp16)[name = tensor("transpose_112")]; - tensor x_293_cast_fp16 = matmul(transpose_x = x_293_transpose_x_0, transpose_y = x_293_transpose_y_0, x = q_with_bias_v_27_cast_fp16, y = var_2658_to_fp16)[name = tensor("x_293_cast_fp16")]; - tensor x_295_pad_0 = const()[name = tensor("x_295_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_295_mode_0 = const()[name = tensor("x_295_mode_0"), val = tensor("constant")]; - tensor const_200_to_fp16 = const()[name = tensor("const_200_to_fp16"), val = tensor(0x0p+0)]; - tensor x_295_cast_fp16 = pad(constant_val = const_200_to_fp16, mode = x_295_mode_0, pad = x_295_pad_0, x = x_293_cast_fp16)[name = tensor("x_295_cast_fp16")]; - tensor var_2666 = const()[name = tensor("op_2666"), val = tensor([1, 8, -1, 188])]; - tensor x_297_cast_fp16 = reshape(shape = var_2666, x = x_295_cast_fp16)[name = tensor("x_297_cast_fp16")]; - tensor var_2670_begin_0 = const()[name = tensor("op_2670_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_2670_end_0 = const()[name = tensor("op_2670_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_2670_end_mask_0 = const()[name = tensor("op_2670_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_2670_cast_fp16 = slice_by_index(begin = var_2670_begin_0, end = var_2670_end_0, end_mask = var_2670_end_mask_0, x = x_297_cast_fp16)[name = tensor("op_2670_cast_fp16")]; - tensor var_2671 = const()[name = tensor("op_2671"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_53_cast_fp16 = reshape(shape = var_2671, x = var_2670_cast_fp16)[name = tensor("matrix_bd_53_cast_fp16")]; - tensor matrix_ac_27_transpose_x_0 = const()[name = tensor("matrix_ac_27_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_27_transpose_y_0 = const()[name = tensor("matrix_ac_27_transpose_y_0"), val = tensor(false)]; - tensor transpose_77_perm_0 = const()[name = tensor("transpose_77_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_78_perm_0 = const()[name = tensor("transpose_78_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_78 = transpose(perm = transpose_78_perm_0, x = k_53_cast_fp16)[name = tensor("transpose_110")]; - tensor transpose_77 = transpose(perm = transpose_77_perm_0, x = var_2654_cast_fp16)[name = tensor("transpose_111")]; - tensor matrix_ac_27_cast_fp16 = matmul(transpose_x = matrix_ac_27_transpose_x_0, transpose_y = matrix_ac_27_transpose_y_0, x = transpose_77, y = transpose_78)[name = tensor("matrix_ac_27_cast_fp16")]; - tensor matrix_bd_55_begin_0 = const()[name = tensor("matrix_bd_55_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_55_end_0 = const()[name = tensor("matrix_bd_55_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_55_end_mask_0 = const()[name = tensor("matrix_bd_55_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_55_cast_fp16 = slice_by_index(begin = matrix_bd_55_begin_0, end = matrix_bd_55_end_0, end_mask = matrix_bd_55_end_mask_0, x = matrix_bd_53_cast_fp16)[name = tensor("matrix_bd_55_cast_fp16")]; - tensor var_2680_cast_fp16 = add(x = matrix_ac_27_cast_fp16, y = matrix_bd_55_cast_fp16)[name = tensor("op_2680_cast_fp16")]; - tensor _inversed_scores_53_y_0_to_fp16 = const()[name = tensor("_inversed_scores_53_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_53_cast_fp16 = mul(x = var_2680_cast_fp16, y = _inversed_scores_53_y_0_to_fp16)[name = tensor("_inversed_scores_53_cast_fp16")]; - tensor scores_55_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_53_cast_fp16, cond = mask_11)[name = tensor("scores_55_cast_fp16")]; - tensor var_2686_cast_fp16 = softmax(axis = var_23, x = scores_55_cast_fp16)[name = tensor("op_2686_cast_fp16")]; - tensor input_709_cast_fp16 = select(a = var_11_to_fp16, b = var_2686_cast_fp16, cond = mask_11)[name = tensor("input_709_cast_fp16")]; - tensor x_299_transpose_x_0 = const()[name = tensor("x_299_transpose_x_0"), val = tensor(false)]; - tensor x_299_transpose_y_0 = const()[name = tensor("x_299_transpose_y_0"), val = tensor(false)]; - tensor value_29_cast_fp16 = transpose(perm = value_29_perm_0, x = v_27_cast_fp16)[name = tensor("transpose_113")]; - tensor x_299_cast_fp16 = matmul(transpose_x = x_299_transpose_x_0, transpose_y = x_299_transpose_y_0, x = input_709_cast_fp16, y = value_29_cast_fp16)[name = tensor("x_299_cast_fp16")]; - tensor var_2690_perm_0 = const()[name = tensor("op_2690_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_2691 = const()[name = tensor("op_2691"), val = tensor([1, -1, 512])]; - tensor var_2690_cast_fp16 = transpose(perm = var_2690_perm_0, x = x_299_cast_fp16)[name = tensor("transpose_109")]; - tensor input_711_cast_fp16 = reshape(shape = var_2691, x = var_2690_cast_fp16)[name = tensor("input_711_cast_fp16")]; - tensor module_layers_13_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_13_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171370816)))]; - tensor module_layers_13_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_13_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171895168)))]; - tensor linear_124_cast_fp16 = linear(bias = module_layers_13_self_attn_linear_out_bias_to_fp16, weight = module_layers_13_self_attn_linear_out_weight_to_fp16, x = input_711_cast_fp16)[name = tensor("linear_124_cast_fp16")]; - tensor input_715_cast_fp16 = add(x = input_707_cast_fp16, y = linear_124_cast_fp16)[name = tensor("input_715_cast_fp16")]; - tensor x_303_axes_0 = const()[name = tensor("x_303_axes_0"), val = tensor([-1])]; - tensor module_layers_13_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_13_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171896256)))]; - tensor module_layers_13_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_13_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171897344)))]; - tensor x_303_cast_fp16 = layer_norm(axes = x_303_axes_0, beta = module_layers_13_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_13_norm_conv_weight_to_fp16, x = input_715_cast_fp16)[name = tensor("x_303_cast_fp16")]; - tensor input_717_perm_0 = const()[name = tensor("input_717_perm_0"), val = tensor([0, 2, 1])]; - tensor input_719_pad_type_0 = const()[name = tensor("input_719_pad_type_0"), val = tensor("valid")]; - tensor input_719_strides_0 = const()[name = tensor("input_719_strides_0"), val = tensor([1])]; - tensor input_719_pad_0 = const()[name = tensor("input_719_pad_0"), val = tensor([0, 0])]; - tensor input_719_dilations_0 = const()[name = tensor("input_719_dilations_0"), val = tensor([1])]; - tensor input_719_groups_0 = const()[name = tensor("input_719_groups_0"), val = tensor(1)]; - tensor module_layers_13_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_13_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171898432)))]; - tensor module_layers_13_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_13_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(172947072)))]; - tensor input_717_cast_fp16 = transpose(perm = input_717_perm_0, x = x_303_cast_fp16)[name = tensor("transpose_108")]; - tensor input_719_cast_fp16 = conv(bias = module_layers_13_conv_pointwise_conv1_bias_to_fp16, dilations = input_719_dilations_0, groups = input_719_groups_0, pad = input_719_pad_0, pad_type = input_719_pad_type_0, strides = input_719_strides_0, weight = module_layers_13_conv_pointwise_conv1_weight_to_fp16, x = input_717_cast_fp16)[name = tensor("input_719_cast_fp16")]; - tensor x_305_split_num_splits_0 = const()[name = tensor("x_305_split_num_splits_0"), val = tensor(2)]; - tensor x_305_split_axis_0 = const()[name = tensor("x_305_split_axis_0"), val = tensor(1)]; - tensor x_305_split_cast_fp16_0, tensor x_305_split_cast_fp16_1 = split(axis = x_305_split_axis_0, num_splits = x_305_split_num_splits_0, x = input_719_cast_fp16)[name = tensor("x_305_split_cast_fp16")]; - tensor x_305_split_1_sigmoid_cast_fp16 = sigmoid(x = x_305_split_cast_fp16_1)[name = tensor("x_305_split_1_sigmoid_cast_fp16")]; - tensor x_305_cast_fp16 = mul(x = x_305_split_cast_fp16_0, y = x_305_split_1_sigmoid_cast_fp16)[name = tensor("x_305_cast_fp16")]; - tensor input_721_cast_fp16 = select(a = var_11_to_fp16, b = x_305_cast_fp16, cond = var_453)[name = tensor("input_721_cast_fp16")]; - tensor input_723_pad_0 = const()[name = tensor("input_723_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_723_mode_0 = const()[name = tensor("input_723_mode_0"), val = tensor("constant")]; - tensor const_203_to_fp16 = const()[name = tensor("const_203_to_fp16"), val = tensor(0x0p+0)]; - tensor input_723_cast_fp16 = pad(constant_val = const_203_to_fp16, mode = input_723_mode_0, pad = input_723_pad_0, x = input_721_cast_fp16)[name = tensor("input_723_cast_fp16")]; - tensor input_725_pad_type_0 = const()[name = tensor("input_725_pad_type_0"), val = tensor("valid")]; - tensor input_725_groups_0 = const()[name = tensor("input_725_groups_0"), val = tensor(512)]; - tensor input_725_strides_0 = const()[name = tensor("input_725_strides_0"), val = tensor([1])]; - tensor input_725_pad_0 = const()[name = tensor("input_725_pad_0"), val = tensor([0, 0])]; - tensor input_725_dilations_0 = const()[name = tensor("input_725_dilations_0"), val = tensor([1])]; - tensor const_260_to_fp16 = const()[name = tensor("const_260_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(172949184)))]; - tensor const_261_to_fp16 = const()[name = tensor("const_261_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(172958464)))]; - tensor input_727_cast_fp16 = conv(bias = const_261_to_fp16, dilations = input_725_dilations_0, groups = input_725_groups_0, pad = input_725_pad_0, pad_type = input_725_pad_type_0, strides = input_725_strides_0, weight = const_260_to_fp16, x = input_723_cast_fp16)[name = tensor("input_727_cast_fp16")]; - tensor input_729_cast_fp16 = silu(x = input_727_cast_fp16)[name = tensor("input_729_cast_fp16")]; - tensor x_307_pad_type_0 = const()[name = tensor("x_307_pad_type_0"), val = tensor("valid")]; - tensor x_307_strides_0 = const()[name = tensor("x_307_strides_0"), val = tensor([1])]; - tensor x_307_pad_0 = const()[name = tensor("x_307_pad_0"), val = tensor([0, 0])]; - tensor x_307_dilations_0 = const()[name = tensor("x_307_dilations_0"), val = tensor([1])]; - tensor x_307_groups_0 = const()[name = tensor("x_307_groups_0"), val = tensor(1)]; - tensor module_layers_13_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_13_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(172959552)))]; - tensor module_layers_13_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_13_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(173483904)))]; - tensor x_307_cast_fp16 = conv(bias = module_layers_13_conv_pointwise_conv2_bias_to_fp16, dilations = x_307_dilations_0, groups = x_307_groups_0, pad = x_307_pad_0, pad_type = x_307_pad_type_0, strides = x_307_strides_0, weight = module_layers_13_conv_pointwise_conv2_weight_to_fp16, x = input_729_cast_fp16)[name = tensor("x_307_cast_fp16")]; - tensor input_731_perm_0 = const()[name = tensor("input_731_perm_0"), val = tensor([0, 2, 1])]; - tensor input_731_cast_fp16 = transpose(perm = input_731_perm_0, x = x_307_cast_fp16)[name = tensor("transpose_107")]; - tensor input_733_cast_fp16 = add(x = input_715_cast_fp16, y = input_731_cast_fp16)[name = tensor("input_733_cast_fp16")]; - tensor input_735_axes_0 = const()[name = tensor("input_735_axes_0"), val = tensor([-1])]; - tensor module_layers_13_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_13_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(173484992)))]; - tensor module_layers_13_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_13_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(173486080)))]; - tensor input_735_cast_fp16 = layer_norm(axes = input_735_axes_0, beta = module_layers_13_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_13_norm_feed_forward2_weight_to_fp16, x = input_733_cast_fp16)[name = tensor("input_735_cast_fp16")]; - tensor module_layers_13_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_13_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(173487168)))]; - tensor module_layers_13_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_13_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(175584384)))]; - tensor linear_125_cast_fp16 = linear(bias = module_layers_13_feed_forward2_linear1_bias_to_fp16, weight = module_layers_13_feed_forward2_linear1_weight_to_fp16, x = input_735_cast_fp16)[name = tensor("linear_125_cast_fp16")]; - tensor input_739_cast_fp16 = silu(x = linear_125_cast_fp16)[name = tensor("input_739_cast_fp16")]; - tensor module_layers_13_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_13_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(175588544)))]; - tensor module_layers_13_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_13_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(177685760)))]; - tensor linear_126_cast_fp16 = linear(bias = module_layers_13_feed_forward2_linear2_bias_to_fp16, weight = module_layers_13_feed_forward2_linear2_weight_to_fp16, x = input_739_cast_fp16)[name = tensor("linear_126_cast_fp16")]; - tensor var_2757_to_fp16 = const()[name = tensor("op_2757_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2758_cast_fp16 = mul(x = linear_126_cast_fp16, y = var_2757_to_fp16)[name = tensor("op_2758_cast_fp16")]; - tensor input_745_cast_fp16 = add(x = input_733_cast_fp16, y = var_2758_cast_fp16)[name = tensor("input_745_cast_fp16")]; - tensor input_747_axes_0 = const()[name = tensor("input_747_axes_0"), val = tensor([-1])]; - tensor module_layers_13_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_13_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(177686848)))]; - tensor module_layers_13_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_13_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(177687936)))]; - tensor input_747_cast_fp16 = layer_norm(axes = input_747_axes_0, beta = module_layers_13_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_13_norm_out_weight_to_fp16, x = input_745_cast_fp16)[name = tensor("input_747_cast_fp16")]; - tensor input_749_axes_0 = const()[name = tensor("input_749_axes_0"), val = tensor([-1])]; - tensor module_layers_14_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_14_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(177689024)))]; - tensor module_layers_14_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_14_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(177690112)))]; - tensor input_749_cast_fp16 = layer_norm(axes = input_749_axes_0, beta = module_layers_14_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_14_norm_feed_forward1_weight_to_fp16, x = input_747_cast_fp16)[name = tensor("input_749_cast_fp16")]; - tensor module_layers_14_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_14_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(177691200)))]; - tensor module_layers_14_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_14_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(179788416)))]; - tensor linear_127_cast_fp16 = linear(bias = module_layers_14_feed_forward1_linear1_bias_to_fp16, weight = module_layers_14_feed_forward1_linear1_weight_to_fp16, x = input_749_cast_fp16)[name = tensor("linear_127_cast_fp16")]; - tensor input_753_cast_fp16 = silu(x = linear_127_cast_fp16)[name = tensor("input_753_cast_fp16")]; - tensor module_layers_14_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_14_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(179792576)))]; - tensor module_layers_14_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_14_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(181889792)))]; - tensor linear_128_cast_fp16 = linear(bias = module_layers_14_feed_forward1_linear2_bias_to_fp16, weight = module_layers_14_feed_forward1_linear2_weight_to_fp16, x = input_753_cast_fp16)[name = tensor("linear_128_cast_fp16")]; - tensor var_2788_to_fp16 = const()[name = tensor("op_2788_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2789_cast_fp16 = mul(x = linear_128_cast_fp16, y = var_2788_to_fp16)[name = tensor("op_2789_cast_fp16")]; - tensor input_759_cast_fp16 = add(x = input_747_cast_fp16, y = var_2789_cast_fp16)[name = tensor("input_759_cast_fp16")]; - tensor query_29_axes_0 = const()[name = tensor("query_29_axes_0"), val = tensor([-1])]; - tensor module_layers_14_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_14_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(181890880)))]; - tensor module_layers_14_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_14_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(181891968)))]; - tensor query_29_cast_fp16 = layer_norm(axes = query_29_axes_0, beta = module_layers_14_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_14_norm_self_att_weight_to_fp16, x = input_759_cast_fp16)[name = tensor("query_29_cast_fp16")]; - tensor module_layers_14_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_14_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(181893056)))]; - tensor module_layers_14_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_14_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182417408)))]; - tensor linear_129_cast_fp16 = linear(bias = module_layers_14_self_attn_linear_q_bias_to_fp16, weight = module_layers_14_self_attn_linear_q_weight_to_fp16, x = query_29_cast_fp16)[name = tensor("linear_129_cast_fp16")]; - tensor var_2806 = const()[name = tensor("op_2806"), val = tensor([1, -1, 8, 64])]; - tensor q_85_cast_fp16 = reshape(shape = var_2806, x = linear_129_cast_fp16)[name = tensor("q_85_cast_fp16")]; - tensor module_layers_14_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_14_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182418496)))]; - tensor module_layers_14_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_14_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182942848)))]; - tensor linear_130_cast_fp16 = linear(bias = module_layers_14_self_attn_linear_k_bias_to_fp16, weight = module_layers_14_self_attn_linear_k_weight_to_fp16, x = query_29_cast_fp16)[name = tensor("linear_130_cast_fp16")]; - tensor var_2811 = const()[name = tensor("op_2811"), val = tensor([1, -1, 8, 64])]; - tensor k_57_cast_fp16 = reshape(shape = var_2811, x = linear_130_cast_fp16)[name = tensor("k_57_cast_fp16")]; - tensor module_layers_14_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_14_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182943936)))]; - tensor module_layers_14_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_14_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183468288)))]; - tensor linear_131_cast_fp16 = linear(bias = module_layers_14_self_attn_linear_v_bias_to_fp16, weight = module_layers_14_self_attn_linear_v_weight_to_fp16, x = query_29_cast_fp16)[name = tensor("linear_131_cast_fp16")]; - tensor var_2816 = const()[name = tensor("op_2816"), val = tensor([1, -1, 8, 64])]; - tensor v_29_cast_fp16 = reshape(shape = var_2816, x = linear_131_cast_fp16)[name = tensor("v_29_cast_fp16")]; - tensor value_31_perm_0 = const()[name = tensor("value_31_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_14_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_14_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183469376)))]; - tensor var_2828_cast_fp16 = add(x = q_85_cast_fp16, y = module_layers_14_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2828_cast_fp16")]; - tensor module_layers_14_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_14_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183470464)))]; - tensor var_2830_cast_fp16 = add(x = q_85_cast_fp16, y = module_layers_14_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2830_cast_fp16")]; - tensor q_with_bias_v_29_perm_0 = const()[name = tensor("q_with_bias_v_29_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_315_transpose_x_0 = const()[name = tensor("x_315_transpose_x_0"), val = tensor(false)]; - tensor x_315_transpose_y_0 = const()[name = tensor("x_315_transpose_y_0"), val = tensor(false)]; - tensor var_2832_to_fp16 = const()[name = tensor("op_2832_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183471552)))]; - tensor q_with_bias_v_29_cast_fp16 = transpose(perm = q_with_bias_v_29_perm_0, x = var_2830_cast_fp16)[name = tensor("transpose_105")]; - tensor x_315_cast_fp16 = matmul(transpose_x = x_315_transpose_x_0, transpose_y = x_315_transpose_y_0, x = q_with_bias_v_29_cast_fp16, y = var_2832_to_fp16)[name = tensor("x_315_cast_fp16")]; - tensor x_317_pad_0 = const()[name = tensor("x_317_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_317_mode_0 = const()[name = tensor("x_317_mode_0"), val = tensor("constant")]; - tensor const_210_to_fp16 = const()[name = tensor("const_210_to_fp16"), val = tensor(0x0p+0)]; - tensor x_317_cast_fp16 = pad(constant_val = const_210_to_fp16, mode = x_317_mode_0, pad = x_317_pad_0, x = x_315_cast_fp16)[name = tensor("x_317_cast_fp16")]; - tensor var_2840 = const()[name = tensor("op_2840"), val = tensor([1, 8, -1, 188])]; - tensor x_319_cast_fp16 = reshape(shape = var_2840, x = x_317_cast_fp16)[name = tensor("x_319_cast_fp16")]; - tensor var_2844_begin_0 = const()[name = tensor("op_2844_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_2844_end_0 = const()[name = tensor("op_2844_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_2844_end_mask_0 = const()[name = tensor("op_2844_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_2844_cast_fp16 = slice_by_index(begin = var_2844_begin_0, end = var_2844_end_0, end_mask = var_2844_end_mask_0, x = x_319_cast_fp16)[name = tensor("op_2844_cast_fp16")]; - tensor var_2845 = const()[name = tensor("op_2845"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_57_cast_fp16 = reshape(shape = var_2845, x = var_2844_cast_fp16)[name = tensor("matrix_bd_57_cast_fp16")]; - tensor matrix_ac_29_transpose_x_0 = const()[name = tensor("matrix_ac_29_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_29_transpose_y_0 = const()[name = tensor("matrix_ac_29_transpose_y_0"), val = tensor(false)]; - tensor transpose_79_perm_0 = const()[name = tensor("transpose_79_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_80_perm_0 = const()[name = tensor("transpose_80_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_80 = transpose(perm = transpose_80_perm_0, x = k_57_cast_fp16)[name = tensor("transpose_103")]; - tensor transpose_79 = transpose(perm = transpose_79_perm_0, x = var_2828_cast_fp16)[name = tensor("transpose_104")]; - tensor matrix_ac_29_cast_fp16 = matmul(transpose_x = matrix_ac_29_transpose_x_0, transpose_y = matrix_ac_29_transpose_y_0, x = transpose_79, y = transpose_80)[name = tensor("matrix_ac_29_cast_fp16")]; - tensor matrix_bd_59_begin_0 = const()[name = tensor("matrix_bd_59_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_59_end_0 = const()[name = tensor("matrix_bd_59_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_59_end_mask_0 = const()[name = tensor("matrix_bd_59_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_59_cast_fp16 = slice_by_index(begin = matrix_bd_59_begin_0, end = matrix_bd_59_end_0, end_mask = matrix_bd_59_end_mask_0, x = matrix_bd_57_cast_fp16)[name = tensor("matrix_bd_59_cast_fp16")]; - tensor var_2854_cast_fp16 = add(x = matrix_ac_29_cast_fp16, y = matrix_bd_59_cast_fp16)[name = tensor("op_2854_cast_fp16")]; - tensor _inversed_scores_57_y_0_to_fp16 = const()[name = tensor("_inversed_scores_57_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_57_cast_fp16 = mul(x = var_2854_cast_fp16, y = _inversed_scores_57_y_0_to_fp16)[name = tensor("_inversed_scores_57_cast_fp16")]; - tensor scores_59_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_57_cast_fp16, cond = mask_11)[name = tensor("scores_59_cast_fp16")]; - tensor var_2860_cast_fp16 = softmax(axis = var_23, x = scores_59_cast_fp16)[name = tensor("op_2860_cast_fp16")]; - tensor input_761_cast_fp16 = select(a = var_11_to_fp16, b = var_2860_cast_fp16, cond = mask_11)[name = tensor("input_761_cast_fp16")]; - tensor x_321_transpose_x_0 = const()[name = tensor("x_321_transpose_x_0"), val = tensor(false)]; - tensor x_321_transpose_y_0 = const()[name = tensor("x_321_transpose_y_0"), val = tensor(false)]; - tensor value_31_cast_fp16 = transpose(perm = value_31_perm_0, x = v_29_cast_fp16)[name = tensor("transpose_106")]; - tensor x_321_cast_fp16 = matmul(transpose_x = x_321_transpose_x_0, transpose_y = x_321_transpose_y_0, x = input_761_cast_fp16, y = value_31_cast_fp16)[name = tensor("x_321_cast_fp16")]; - tensor var_2864_perm_0 = const()[name = tensor("op_2864_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_2865 = const()[name = tensor("op_2865"), val = tensor([1, -1, 512])]; - tensor var_2864_cast_fp16 = transpose(perm = var_2864_perm_0, x = x_321_cast_fp16)[name = tensor("transpose_102")]; - tensor input_763_cast_fp16 = reshape(shape = var_2865, x = var_2864_cast_fp16)[name = tensor("input_763_cast_fp16")]; - tensor module_layers_14_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_14_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183855616)))]; - tensor module_layers_14_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_14_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184379968)))]; - tensor linear_133_cast_fp16 = linear(bias = module_layers_14_self_attn_linear_out_bias_to_fp16, weight = module_layers_14_self_attn_linear_out_weight_to_fp16, x = input_763_cast_fp16)[name = tensor("linear_133_cast_fp16")]; - tensor input_767_cast_fp16 = add(x = input_759_cast_fp16, y = linear_133_cast_fp16)[name = tensor("input_767_cast_fp16")]; - tensor x_325_axes_0 = const()[name = tensor("x_325_axes_0"), val = tensor([-1])]; - tensor module_layers_14_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_14_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184381056)))]; - tensor module_layers_14_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_14_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184382144)))]; - tensor x_325_cast_fp16 = layer_norm(axes = x_325_axes_0, beta = module_layers_14_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_14_norm_conv_weight_to_fp16, x = input_767_cast_fp16)[name = tensor("x_325_cast_fp16")]; - tensor input_769_perm_0 = const()[name = tensor("input_769_perm_0"), val = tensor([0, 2, 1])]; - tensor input_771_pad_type_0 = const()[name = tensor("input_771_pad_type_0"), val = tensor("valid")]; - tensor input_771_strides_0 = const()[name = tensor("input_771_strides_0"), val = tensor([1])]; - tensor input_771_pad_0 = const()[name = tensor("input_771_pad_0"), val = tensor([0, 0])]; - tensor input_771_dilations_0 = const()[name = tensor("input_771_dilations_0"), val = tensor([1])]; - tensor input_771_groups_0 = const()[name = tensor("input_771_groups_0"), val = tensor(1)]; - tensor module_layers_14_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_14_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184383232)))]; - tensor module_layers_14_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_14_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185431872)))]; - tensor input_769_cast_fp16 = transpose(perm = input_769_perm_0, x = x_325_cast_fp16)[name = tensor("transpose_101")]; - tensor input_771_cast_fp16 = conv(bias = module_layers_14_conv_pointwise_conv1_bias_to_fp16, dilations = input_771_dilations_0, groups = input_771_groups_0, pad = input_771_pad_0, pad_type = input_771_pad_type_0, strides = input_771_strides_0, weight = module_layers_14_conv_pointwise_conv1_weight_to_fp16, x = input_769_cast_fp16)[name = tensor("input_771_cast_fp16")]; - tensor x_327_split_num_splits_0 = const()[name = tensor("x_327_split_num_splits_0"), val = tensor(2)]; - tensor x_327_split_axis_0 = const()[name = tensor("x_327_split_axis_0"), val = tensor(1)]; - tensor x_327_split_cast_fp16_0, tensor x_327_split_cast_fp16_1 = split(axis = x_327_split_axis_0, num_splits = x_327_split_num_splits_0, x = input_771_cast_fp16)[name = tensor("x_327_split_cast_fp16")]; - tensor x_327_split_1_sigmoid_cast_fp16 = sigmoid(x = x_327_split_cast_fp16_1)[name = tensor("x_327_split_1_sigmoid_cast_fp16")]; - tensor x_327_cast_fp16 = mul(x = x_327_split_cast_fp16_0, y = x_327_split_1_sigmoid_cast_fp16)[name = tensor("x_327_cast_fp16")]; - tensor input_773_cast_fp16 = select(a = var_11_to_fp16, b = x_327_cast_fp16, cond = var_453)[name = tensor("input_773_cast_fp16")]; - tensor input_775_pad_0 = const()[name = tensor("input_775_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_775_mode_0 = const()[name = tensor("input_775_mode_0"), val = tensor("constant")]; - tensor const_213_to_fp16 = const()[name = tensor("const_213_to_fp16"), val = tensor(0x0p+0)]; - tensor input_775_cast_fp16 = pad(constant_val = const_213_to_fp16, mode = input_775_mode_0, pad = input_775_pad_0, x = input_773_cast_fp16)[name = tensor("input_775_cast_fp16")]; - tensor input_777_pad_type_0 = const()[name = tensor("input_777_pad_type_0"), val = tensor("valid")]; - tensor input_777_groups_0 = const()[name = tensor("input_777_groups_0"), val = tensor(512)]; - tensor input_777_strides_0 = const()[name = tensor("input_777_strides_0"), val = tensor([1])]; - tensor input_777_pad_0 = const()[name = tensor("input_777_pad_0"), val = tensor([0, 0])]; - tensor input_777_dilations_0 = const()[name = tensor("input_777_dilations_0"), val = tensor([1])]; - tensor const_262_to_fp16 = const()[name = tensor("const_262_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185433984)))]; - tensor const_263_to_fp16 = const()[name = tensor("const_263_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185443264)))]; - tensor input_779_cast_fp16 = conv(bias = const_263_to_fp16, dilations = input_777_dilations_0, groups = input_777_groups_0, pad = input_777_pad_0, pad_type = input_777_pad_type_0, strides = input_777_strides_0, weight = const_262_to_fp16, x = input_775_cast_fp16)[name = tensor("input_779_cast_fp16")]; - tensor input_781_cast_fp16 = silu(x = input_779_cast_fp16)[name = tensor("input_781_cast_fp16")]; - tensor x_329_pad_type_0 = const()[name = tensor("x_329_pad_type_0"), val = tensor("valid")]; - tensor x_329_strides_0 = const()[name = tensor("x_329_strides_0"), val = tensor([1])]; - tensor x_329_pad_0 = const()[name = tensor("x_329_pad_0"), val = tensor([0, 0])]; - tensor x_329_dilations_0 = const()[name = tensor("x_329_dilations_0"), val = tensor([1])]; - tensor x_329_groups_0 = const()[name = tensor("x_329_groups_0"), val = tensor(1)]; - tensor module_layers_14_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_14_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185444352)))]; - tensor module_layers_14_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_14_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185968704)))]; - tensor x_329_cast_fp16 = conv(bias = module_layers_14_conv_pointwise_conv2_bias_to_fp16, dilations = x_329_dilations_0, groups = x_329_groups_0, pad = x_329_pad_0, pad_type = x_329_pad_type_0, strides = x_329_strides_0, weight = module_layers_14_conv_pointwise_conv2_weight_to_fp16, x = input_781_cast_fp16)[name = tensor("x_329_cast_fp16")]; - tensor input_783_perm_0 = const()[name = tensor("input_783_perm_0"), val = tensor([0, 2, 1])]; - tensor input_783_cast_fp16 = transpose(perm = input_783_perm_0, x = x_329_cast_fp16)[name = tensor("transpose_100")]; - tensor input_785_cast_fp16 = add(x = input_767_cast_fp16, y = input_783_cast_fp16)[name = tensor("input_785_cast_fp16")]; - tensor input_787_axes_0 = const()[name = tensor("input_787_axes_0"), val = tensor([-1])]; - tensor module_layers_14_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_14_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185969792)))]; - tensor module_layers_14_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_14_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185970880)))]; - tensor input_787_cast_fp16 = layer_norm(axes = input_787_axes_0, beta = module_layers_14_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_14_norm_feed_forward2_weight_to_fp16, x = input_785_cast_fp16)[name = tensor("input_787_cast_fp16")]; - tensor module_layers_14_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_14_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185971968)))]; - tensor module_layers_14_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_14_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188069184)))]; - tensor linear_134_cast_fp16 = linear(bias = module_layers_14_feed_forward2_linear1_bias_to_fp16, weight = module_layers_14_feed_forward2_linear1_weight_to_fp16, x = input_787_cast_fp16)[name = tensor("linear_134_cast_fp16")]; - tensor input_791_cast_fp16 = silu(x = linear_134_cast_fp16)[name = tensor("input_791_cast_fp16")]; - tensor module_layers_14_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_14_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188073344)))]; - tensor module_layers_14_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_14_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190170560)))]; - tensor linear_135_cast_fp16 = linear(bias = module_layers_14_feed_forward2_linear2_bias_to_fp16, weight = module_layers_14_feed_forward2_linear2_weight_to_fp16, x = input_791_cast_fp16)[name = tensor("linear_135_cast_fp16")]; - tensor var_2931_to_fp16 = const()[name = tensor("op_2931_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2932_cast_fp16 = mul(x = linear_135_cast_fp16, y = var_2931_to_fp16)[name = tensor("op_2932_cast_fp16")]; - tensor input_797_cast_fp16 = add(x = input_785_cast_fp16, y = var_2932_cast_fp16)[name = tensor("input_797_cast_fp16")]; - tensor input_799_axes_0 = const()[name = tensor("input_799_axes_0"), val = tensor([-1])]; - tensor module_layers_14_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_14_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190171648)))]; - tensor module_layers_14_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_14_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190172736)))]; - tensor input_799_cast_fp16 = layer_norm(axes = input_799_axes_0, beta = module_layers_14_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_14_norm_out_weight_to_fp16, x = input_797_cast_fp16)[name = tensor("input_799_cast_fp16")]; - tensor input_801_axes_0 = const()[name = tensor("input_801_axes_0"), val = tensor([-1])]; - tensor module_layers_15_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_15_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190173824)))]; - tensor module_layers_15_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_15_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190174912)))]; - tensor input_801_cast_fp16 = layer_norm(axes = input_801_axes_0, beta = module_layers_15_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_15_norm_feed_forward1_weight_to_fp16, x = input_799_cast_fp16)[name = tensor("input_801_cast_fp16")]; - tensor module_layers_15_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_15_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190176000)))]; - tensor module_layers_15_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_15_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(192273216)))]; - tensor linear_136_cast_fp16 = linear(bias = module_layers_15_feed_forward1_linear1_bias_to_fp16, weight = module_layers_15_feed_forward1_linear1_weight_to_fp16, x = input_801_cast_fp16)[name = tensor("linear_136_cast_fp16")]; - tensor input_805_cast_fp16 = silu(x = linear_136_cast_fp16)[name = tensor("input_805_cast_fp16")]; - tensor module_layers_15_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_15_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(192277376)))]; - tensor module_layers_15_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_15_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194374592)))]; - tensor linear_137_cast_fp16 = linear(bias = module_layers_15_feed_forward1_linear2_bias_to_fp16, weight = module_layers_15_feed_forward1_linear2_weight_to_fp16, x = input_805_cast_fp16)[name = tensor("linear_137_cast_fp16")]; - tensor var_2962_to_fp16 = const()[name = tensor("op_2962_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2963_cast_fp16 = mul(x = linear_137_cast_fp16, y = var_2962_to_fp16)[name = tensor("op_2963_cast_fp16")]; - tensor input_811_cast_fp16 = add(x = input_799_cast_fp16, y = var_2963_cast_fp16)[name = tensor("input_811_cast_fp16")]; - tensor query_31_axes_0 = const()[name = tensor("query_31_axes_0"), val = tensor([-1])]; - tensor module_layers_15_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_15_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194375680)))]; - tensor module_layers_15_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_15_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194376768)))]; - tensor query_31_cast_fp16 = layer_norm(axes = query_31_axes_0, beta = module_layers_15_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_15_norm_self_att_weight_to_fp16, x = input_811_cast_fp16)[name = tensor("query_31_cast_fp16")]; - tensor module_layers_15_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_15_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194377856)))]; - tensor module_layers_15_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_15_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194902208)))]; - tensor linear_138_cast_fp16 = linear(bias = module_layers_15_self_attn_linear_q_bias_to_fp16, weight = module_layers_15_self_attn_linear_q_weight_to_fp16, x = query_31_cast_fp16)[name = tensor("linear_138_cast_fp16")]; - tensor var_2980 = const()[name = tensor("op_2980"), val = tensor([1, -1, 8, 64])]; - tensor q_91_cast_fp16 = reshape(shape = var_2980, x = linear_138_cast_fp16)[name = tensor("q_91_cast_fp16")]; - tensor module_layers_15_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_15_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194903296)))]; - tensor module_layers_15_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_15_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195427648)))]; - tensor linear_139_cast_fp16 = linear(bias = module_layers_15_self_attn_linear_k_bias_to_fp16, weight = module_layers_15_self_attn_linear_k_weight_to_fp16, x = query_31_cast_fp16)[name = tensor("linear_139_cast_fp16")]; - tensor var_2985 = const()[name = tensor("op_2985"), val = tensor([1, -1, 8, 64])]; - tensor k_61_cast_fp16 = reshape(shape = var_2985, x = linear_139_cast_fp16)[name = tensor("k_61_cast_fp16")]; - tensor module_layers_15_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_15_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195428736)))]; - tensor module_layers_15_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_15_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195953088)))]; - tensor linear_140_cast_fp16 = linear(bias = module_layers_15_self_attn_linear_v_bias_to_fp16, weight = module_layers_15_self_attn_linear_v_weight_to_fp16, x = query_31_cast_fp16)[name = tensor("linear_140_cast_fp16")]; - tensor var_2990 = const()[name = tensor("op_2990"), val = tensor([1, -1, 8, 64])]; - tensor v_31_cast_fp16 = reshape(shape = var_2990, x = linear_140_cast_fp16)[name = tensor("v_31_cast_fp16")]; - tensor value_33_perm_0 = const()[name = tensor("value_33_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_15_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_15_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195954176)))]; - tensor var_3002_cast_fp16 = add(x = q_91_cast_fp16, y = module_layers_15_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3002_cast_fp16")]; - tensor module_layers_15_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_15_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195955264)))]; - tensor var_3004_cast_fp16 = add(x = q_91_cast_fp16, y = module_layers_15_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3004_cast_fp16")]; - tensor q_with_bias_v_31_perm_0 = const()[name = tensor("q_with_bias_v_31_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_337_transpose_x_0 = const()[name = tensor("x_337_transpose_x_0"), val = tensor(false)]; - tensor x_337_transpose_y_0 = const()[name = tensor("x_337_transpose_y_0"), val = tensor(false)]; - tensor var_3006_to_fp16 = const()[name = tensor("op_3006_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195956352)))]; - tensor q_with_bias_v_31_cast_fp16 = transpose(perm = q_with_bias_v_31_perm_0, x = var_3004_cast_fp16)[name = tensor("transpose_98")]; - tensor x_337_cast_fp16 = matmul(transpose_x = x_337_transpose_x_0, transpose_y = x_337_transpose_y_0, x = q_with_bias_v_31_cast_fp16, y = var_3006_to_fp16)[name = tensor("x_337_cast_fp16")]; - tensor x_339_pad_0 = const()[name = tensor("x_339_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_339_mode_0 = const()[name = tensor("x_339_mode_0"), val = tensor("constant")]; - tensor const_220_to_fp16 = const()[name = tensor("const_220_to_fp16"), val = tensor(0x0p+0)]; - tensor x_339_cast_fp16 = pad(constant_val = const_220_to_fp16, mode = x_339_mode_0, pad = x_339_pad_0, x = x_337_cast_fp16)[name = tensor("x_339_cast_fp16")]; - tensor var_3014 = const()[name = tensor("op_3014"), val = tensor([1, 8, -1, 188])]; - tensor x_341_cast_fp16 = reshape(shape = var_3014, x = x_339_cast_fp16)[name = tensor("x_341_cast_fp16")]; - tensor var_3018_begin_0 = const()[name = tensor("op_3018_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_3018_end_0 = const()[name = tensor("op_3018_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_3018_end_mask_0 = const()[name = tensor("op_3018_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_3018_cast_fp16 = slice_by_index(begin = var_3018_begin_0, end = var_3018_end_0, end_mask = var_3018_end_mask_0, x = x_341_cast_fp16)[name = tensor("op_3018_cast_fp16")]; - tensor var_3019 = const()[name = tensor("op_3019"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_61_cast_fp16 = reshape(shape = var_3019, x = var_3018_cast_fp16)[name = tensor("matrix_bd_61_cast_fp16")]; - tensor matrix_ac_31_transpose_x_0 = const()[name = tensor("matrix_ac_31_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_31_transpose_y_0 = const()[name = tensor("matrix_ac_31_transpose_y_0"), val = tensor(false)]; - tensor transpose_81_perm_0 = const()[name = tensor("transpose_81_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_82_perm_0 = const()[name = tensor("transpose_82_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_82 = transpose(perm = transpose_82_perm_0, x = k_61_cast_fp16)[name = tensor("transpose_96")]; - tensor transpose_81 = transpose(perm = transpose_81_perm_0, x = var_3002_cast_fp16)[name = tensor("transpose_97")]; - tensor matrix_ac_31_cast_fp16 = matmul(transpose_x = matrix_ac_31_transpose_x_0, transpose_y = matrix_ac_31_transpose_y_0, x = transpose_81, y = transpose_82)[name = tensor("matrix_ac_31_cast_fp16")]; - tensor matrix_bd_63_begin_0 = const()[name = tensor("matrix_bd_63_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_63_end_0 = const()[name = tensor("matrix_bd_63_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_63_end_mask_0 = const()[name = tensor("matrix_bd_63_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_63_cast_fp16 = slice_by_index(begin = matrix_bd_63_begin_0, end = matrix_bd_63_end_0, end_mask = matrix_bd_63_end_mask_0, x = matrix_bd_61_cast_fp16)[name = tensor("matrix_bd_63_cast_fp16")]; - tensor var_3028_cast_fp16 = add(x = matrix_ac_31_cast_fp16, y = matrix_bd_63_cast_fp16)[name = tensor("op_3028_cast_fp16")]; - tensor _inversed_scores_61_y_0_to_fp16 = const()[name = tensor("_inversed_scores_61_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_61_cast_fp16 = mul(x = var_3028_cast_fp16, y = _inversed_scores_61_y_0_to_fp16)[name = tensor("_inversed_scores_61_cast_fp16")]; - tensor scores_63_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_61_cast_fp16, cond = mask_11)[name = tensor("scores_63_cast_fp16")]; - tensor var_3034_cast_fp16 = softmax(axis = var_23, x = scores_63_cast_fp16)[name = tensor("op_3034_cast_fp16")]; - tensor input_813_cast_fp16 = select(a = var_11_to_fp16, b = var_3034_cast_fp16, cond = mask_11)[name = tensor("input_813_cast_fp16")]; - tensor x_343_transpose_x_0 = const()[name = tensor("x_343_transpose_x_0"), val = tensor(false)]; - tensor x_343_transpose_y_0 = const()[name = tensor("x_343_transpose_y_0"), val = tensor(false)]; - tensor value_33_cast_fp16 = transpose(perm = value_33_perm_0, x = v_31_cast_fp16)[name = tensor("transpose_99")]; - tensor x_343_cast_fp16 = matmul(transpose_x = x_343_transpose_x_0, transpose_y = x_343_transpose_y_0, x = input_813_cast_fp16, y = value_33_cast_fp16)[name = tensor("x_343_cast_fp16")]; - tensor var_3038_perm_0 = const()[name = tensor("op_3038_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_3039 = const()[name = tensor("op_3039"), val = tensor([1, -1, 512])]; - tensor var_3038_cast_fp16 = transpose(perm = var_3038_perm_0, x = x_343_cast_fp16)[name = tensor("transpose_95")]; - tensor input_815_cast_fp16 = reshape(shape = var_3039, x = var_3038_cast_fp16)[name = tensor("input_815_cast_fp16")]; - tensor module_layers_15_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_15_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196340416)))]; - tensor module_layers_15_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_15_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196864768)))]; - tensor linear_142_cast_fp16 = linear(bias = module_layers_15_self_attn_linear_out_bias_to_fp16, weight = module_layers_15_self_attn_linear_out_weight_to_fp16, x = input_815_cast_fp16)[name = tensor("linear_142_cast_fp16")]; - tensor input_819_cast_fp16 = add(x = input_811_cast_fp16, y = linear_142_cast_fp16)[name = tensor("input_819_cast_fp16")]; - tensor x_347_axes_0 = const()[name = tensor("x_347_axes_0"), val = tensor([-1])]; - tensor module_layers_15_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_15_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196865856)))]; - tensor module_layers_15_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_15_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196866944)))]; - tensor x_347_cast_fp16 = layer_norm(axes = x_347_axes_0, beta = module_layers_15_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_15_norm_conv_weight_to_fp16, x = input_819_cast_fp16)[name = tensor("x_347_cast_fp16")]; - tensor input_821_perm_0 = const()[name = tensor("input_821_perm_0"), val = tensor([0, 2, 1])]; - tensor input_823_pad_type_0 = const()[name = tensor("input_823_pad_type_0"), val = tensor("valid")]; - tensor input_823_strides_0 = const()[name = tensor("input_823_strides_0"), val = tensor([1])]; - tensor input_823_pad_0 = const()[name = tensor("input_823_pad_0"), val = tensor([0, 0])]; - tensor input_823_dilations_0 = const()[name = tensor("input_823_dilations_0"), val = tensor([1])]; - tensor input_823_groups_0 = const()[name = tensor("input_823_groups_0"), val = tensor(1)]; - tensor module_layers_15_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_15_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196868032)))]; - tensor module_layers_15_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_15_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(197916672)))]; - tensor input_821_cast_fp16 = transpose(perm = input_821_perm_0, x = x_347_cast_fp16)[name = tensor("transpose_94")]; - tensor input_823_cast_fp16 = conv(bias = module_layers_15_conv_pointwise_conv1_bias_to_fp16, dilations = input_823_dilations_0, groups = input_823_groups_0, pad = input_823_pad_0, pad_type = input_823_pad_type_0, strides = input_823_strides_0, weight = module_layers_15_conv_pointwise_conv1_weight_to_fp16, x = input_821_cast_fp16)[name = tensor("input_823_cast_fp16")]; - tensor x_349_split_num_splits_0 = const()[name = tensor("x_349_split_num_splits_0"), val = tensor(2)]; - tensor x_349_split_axis_0 = const()[name = tensor("x_349_split_axis_0"), val = tensor(1)]; - tensor x_349_split_cast_fp16_0, tensor x_349_split_cast_fp16_1 = split(axis = x_349_split_axis_0, num_splits = x_349_split_num_splits_0, x = input_823_cast_fp16)[name = tensor("x_349_split_cast_fp16")]; - tensor x_349_split_1_sigmoid_cast_fp16 = sigmoid(x = x_349_split_cast_fp16_1)[name = tensor("x_349_split_1_sigmoid_cast_fp16")]; - tensor x_349_cast_fp16 = mul(x = x_349_split_cast_fp16_0, y = x_349_split_1_sigmoid_cast_fp16)[name = tensor("x_349_cast_fp16")]; - tensor input_825_cast_fp16 = select(a = var_11_to_fp16, b = x_349_cast_fp16, cond = var_453)[name = tensor("input_825_cast_fp16")]; - tensor input_827_pad_0 = const()[name = tensor("input_827_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_827_mode_0 = const()[name = tensor("input_827_mode_0"), val = tensor("constant")]; - tensor const_223_to_fp16 = const()[name = tensor("const_223_to_fp16"), val = tensor(0x0p+0)]; - tensor input_827_cast_fp16 = pad(constant_val = const_223_to_fp16, mode = input_827_mode_0, pad = input_827_pad_0, x = input_825_cast_fp16)[name = tensor("input_827_cast_fp16")]; - tensor input_829_pad_type_0 = const()[name = tensor("input_829_pad_type_0"), val = tensor("valid")]; - tensor input_829_groups_0 = const()[name = tensor("input_829_groups_0"), val = tensor(512)]; - tensor input_829_strides_0 = const()[name = tensor("input_829_strides_0"), val = tensor([1])]; - tensor input_829_pad_0 = const()[name = tensor("input_829_pad_0"), val = tensor([0, 0])]; - tensor input_829_dilations_0 = const()[name = tensor("input_829_dilations_0"), val = tensor([1])]; - tensor const_264_to_fp16 = const()[name = tensor("const_264_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(197918784)))]; - tensor const_265_to_fp16 = const()[name = tensor("const_265_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(197928064)))]; - tensor input_831_cast_fp16 = conv(bias = const_265_to_fp16, dilations = input_829_dilations_0, groups = input_829_groups_0, pad = input_829_pad_0, pad_type = input_829_pad_type_0, strides = input_829_strides_0, weight = const_264_to_fp16, x = input_827_cast_fp16)[name = tensor("input_831_cast_fp16")]; - tensor input_833_cast_fp16 = silu(x = input_831_cast_fp16)[name = tensor("input_833_cast_fp16")]; - tensor x_351_pad_type_0 = const()[name = tensor("x_351_pad_type_0"), val = tensor("valid")]; - tensor x_351_strides_0 = const()[name = tensor("x_351_strides_0"), val = tensor([1])]; - tensor x_351_pad_0 = const()[name = tensor("x_351_pad_0"), val = tensor([0, 0])]; - tensor x_351_dilations_0 = const()[name = tensor("x_351_dilations_0"), val = tensor([1])]; - tensor x_351_groups_0 = const()[name = tensor("x_351_groups_0"), val = tensor(1)]; - tensor module_layers_15_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_15_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(197929152)))]; - tensor module_layers_15_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_15_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198453504)))]; - tensor x_351_cast_fp16 = conv(bias = module_layers_15_conv_pointwise_conv2_bias_to_fp16, dilations = x_351_dilations_0, groups = x_351_groups_0, pad = x_351_pad_0, pad_type = x_351_pad_type_0, strides = x_351_strides_0, weight = module_layers_15_conv_pointwise_conv2_weight_to_fp16, x = input_833_cast_fp16)[name = tensor("x_351_cast_fp16")]; - tensor input_835_perm_0 = const()[name = tensor("input_835_perm_0"), val = tensor([0, 2, 1])]; - tensor input_835_cast_fp16 = transpose(perm = input_835_perm_0, x = x_351_cast_fp16)[name = tensor("transpose_93")]; - tensor input_837_cast_fp16 = add(x = input_819_cast_fp16, y = input_835_cast_fp16)[name = tensor("input_837_cast_fp16")]; - tensor input_839_axes_0 = const()[name = tensor("input_839_axes_0"), val = tensor([-1])]; - tensor module_layers_15_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_15_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198454592)))]; - tensor module_layers_15_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_15_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198455680)))]; - tensor input_839_cast_fp16 = layer_norm(axes = input_839_axes_0, beta = module_layers_15_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_15_norm_feed_forward2_weight_to_fp16, x = input_837_cast_fp16)[name = tensor("input_839_cast_fp16")]; - tensor module_layers_15_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_15_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198456768)))]; - tensor module_layers_15_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_15_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(200553984)))]; - tensor linear_143_cast_fp16 = linear(bias = module_layers_15_feed_forward2_linear1_bias_to_fp16, weight = module_layers_15_feed_forward2_linear1_weight_to_fp16, x = input_839_cast_fp16)[name = tensor("linear_143_cast_fp16")]; - tensor input_843_cast_fp16 = silu(x = linear_143_cast_fp16)[name = tensor("input_843_cast_fp16")]; - tensor module_layers_15_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_15_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(200558144)))]; - tensor module_layers_15_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_15_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202655360)))]; - tensor linear_144_cast_fp16 = linear(bias = module_layers_15_feed_forward2_linear2_bias_to_fp16, weight = module_layers_15_feed_forward2_linear2_weight_to_fp16, x = input_843_cast_fp16)[name = tensor("linear_144_cast_fp16")]; - tensor var_3105_to_fp16 = const()[name = tensor("op_3105_to_fp16"), val = tensor(0x1p-1)]; - tensor var_3106_cast_fp16 = mul(x = linear_144_cast_fp16, y = var_3105_to_fp16)[name = tensor("op_3106_cast_fp16")]; - tensor input_849_cast_fp16 = add(x = input_837_cast_fp16, y = var_3106_cast_fp16)[name = tensor("input_849_cast_fp16")]; - tensor input_851_axes_0 = const()[name = tensor("input_851_axes_0"), val = tensor([-1])]; - tensor module_layers_15_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_15_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202656448)))]; - tensor module_layers_15_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_15_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202657536)))]; - tensor input_851_cast_fp16 = layer_norm(axes = input_851_axes_0, beta = module_layers_15_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_15_norm_out_weight_to_fp16, x = input_849_cast_fp16)[name = tensor("input_851_cast_fp16")]; - tensor input_853_axes_0 = const()[name = tensor("input_853_axes_0"), val = tensor([-1])]; - tensor module_layers_16_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_16_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202658624)))]; - tensor module_layers_16_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_16_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202659712)))]; - tensor input_853_cast_fp16 = layer_norm(axes = input_853_axes_0, beta = module_layers_16_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_16_norm_feed_forward1_weight_to_fp16, x = input_851_cast_fp16)[name = tensor("input_853_cast_fp16")]; - tensor module_layers_16_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_16_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202660800)))]; - tensor module_layers_16_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_16_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(204758016)))]; - tensor linear_145_cast_fp16 = linear(bias = module_layers_16_feed_forward1_linear1_bias_to_fp16, weight = module_layers_16_feed_forward1_linear1_weight_to_fp16, x = input_853_cast_fp16)[name = tensor("linear_145_cast_fp16")]; - tensor input_857_cast_fp16 = silu(x = linear_145_cast_fp16)[name = tensor("input_857_cast_fp16")]; - tensor module_layers_16_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_16_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(204762176)))]; - tensor module_layers_16_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_16_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(206859392)))]; - tensor linear_146_cast_fp16 = linear(bias = module_layers_16_feed_forward1_linear2_bias_to_fp16, weight = module_layers_16_feed_forward1_linear2_weight_to_fp16, x = input_857_cast_fp16)[name = tensor("linear_146_cast_fp16")]; - tensor var_3136_to_fp16 = const()[name = tensor("op_3136_to_fp16"), val = tensor(0x1p-1)]; - tensor var_3137_cast_fp16 = mul(x = linear_146_cast_fp16, y = var_3136_to_fp16)[name = tensor("op_3137_cast_fp16")]; - tensor input_863_cast_fp16 = add(x = input_851_cast_fp16, y = var_3137_cast_fp16)[name = tensor("input_863_cast_fp16")]; - tensor query_axes_0 = const()[name = tensor("query_axes_0"), val = tensor([-1])]; - tensor module_layers_16_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_16_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(206860480)))]; - tensor module_layers_16_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_16_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(206861568)))]; - tensor query_cast_fp16 = layer_norm(axes = query_axes_0, beta = module_layers_16_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_16_norm_self_att_weight_to_fp16, x = input_863_cast_fp16)[name = tensor("query_cast_fp16")]; - tensor module_layers_16_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_16_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(206862656)))]; - tensor module_layers_16_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_16_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(207387008)))]; - tensor linear_147_cast_fp16 = linear(bias = module_layers_16_self_attn_linear_q_bias_to_fp16, weight = module_layers_16_self_attn_linear_q_weight_to_fp16, x = query_cast_fp16)[name = tensor("linear_147_cast_fp16")]; - tensor var_3154 = const()[name = tensor("op_3154"), val = tensor([1, -1, 8, 64])]; - tensor q_97_cast_fp16 = reshape(shape = var_3154, x = linear_147_cast_fp16)[name = tensor("q_97_cast_fp16")]; - tensor module_layers_16_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_16_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(207388096)))]; - tensor module_layers_16_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_16_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(207912448)))]; - tensor linear_148_cast_fp16 = linear(bias = module_layers_16_self_attn_linear_k_bias_to_fp16, weight = module_layers_16_self_attn_linear_k_weight_to_fp16, x = query_cast_fp16)[name = tensor("linear_148_cast_fp16")]; - tensor var_3159 = const()[name = tensor("op_3159"), val = tensor([1, -1, 8, 64])]; - tensor k_65_cast_fp16 = reshape(shape = var_3159, x = linear_148_cast_fp16)[name = tensor("k_65_cast_fp16")]; - tensor module_layers_16_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_16_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(207913536)))]; - tensor module_layers_16_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_16_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208437888)))]; - tensor linear_149_cast_fp16 = linear(bias = module_layers_16_self_attn_linear_v_bias_to_fp16, weight = module_layers_16_self_attn_linear_v_weight_to_fp16, x = query_cast_fp16)[name = tensor("linear_149_cast_fp16")]; - tensor var_3164 = const()[name = tensor("op_3164"), val = tensor([1, -1, 8, 64])]; - tensor v_cast_fp16 = reshape(shape = var_3164, x = linear_149_cast_fp16)[name = tensor("v_cast_fp16")]; - tensor value_perm_0 = const()[name = tensor("value_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_16_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_16_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208438976)))]; - tensor var_3176_cast_fp16 = add(x = q_97_cast_fp16, y = module_layers_16_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3176_cast_fp16")]; - tensor module_layers_16_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_16_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208440064)))]; - tensor var_3178_cast_fp16 = add(x = q_97_cast_fp16, y = module_layers_16_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3178_cast_fp16")]; - tensor q_with_bias_v_perm_0 = const()[name = tensor("q_with_bias_v_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_359_transpose_x_0 = const()[name = tensor("x_359_transpose_x_0"), val = tensor(false)]; - tensor x_359_transpose_y_0 = const()[name = tensor("x_359_transpose_y_0"), val = tensor(false)]; - tensor var_3180_to_fp16 = const()[name = tensor("op_3180_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208441152)))]; - tensor q_with_bias_v_cast_fp16 = transpose(perm = q_with_bias_v_perm_0, x = var_3178_cast_fp16)[name = tensor("transpose_91")]; - tensor x_359_cast_fp16 = matmul(transpose_x = x_359_transpose_x_0, transpose_y = x_359_transpose_y_0, x = q_with_bias_v_cast_fp16, y = var_3180_to_fp16)[name = tensor("x_359_cast_fp16")]; - tensor x_361_pad_0 = const()[name = tensor("x_361_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_361_mode_0 = const()[name = tensor("x_361_mode_0"), val = tensor("constant")]; - tensor const_230_to_fp16 = const()[name = tensor("const_230_to_fp16"), val = tensor(0x0p+0)]; - tensor x_361_cast_fp16 = pad(constant_val = const_230_to_fp16, mode = x_361_mode_0, pad = x_361_pad_0, x = x_359_cast_fp16)[name = tensor("x_361_cast_fp16")]; - tensor var_3188 = const()[name = tensor("op_3188"), val = tensor([1, 8, -1, 188])]; - tensor x_363_cast_fp16 = reshape(shape = var_3188, x = x_361_cast_fp16)[name = tensor("x_363_cast_fp16")]; - tensor var_3192_begin_0 = const()[name = tensor("op_3192_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_3192_end_0 = const()[name = tensor("op_3192_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_3192_end_mask_0 = const()[name = tensor("op_3192_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_3192_cast_fp16 = slice_by_index(begin = var_3192_begin_0, end = var_3192_end_0, end_mask = var_3192_end_mask_0, x = x_363_cast_fp16)[name = tensor("op_3192_cast_fp16")]; - tensor var_3193 = const()[name = tensor("op_3193"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_65_cast_fp16 = reshape(shape = var_3193, x = var_3192_cast_fp16)[name = tensor("matrix_bd_65_cast_fp16")]; - tensor matrix_ac_transpose_x_0 = const()[name = tensor("matrix_ac_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_transpose_y_0 = const()[name = tensor("matrix_ac_transpose_y_0"), val = tensor(false)]; - tensor transpose_83_perm_0 = const()[name = tensor("transpose_83_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_84_perm_0 = const()[name = tensor("transpose_84_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_84 = transpose(perm = transpose_84_perm_0, x = k_65_cast_fp16)[name = tensor("transpose_89")]; - tensor transpose_83 = transpose(perm = transpose_83_perm_0, x = var_3176_cast_fp16)[name = tensor("transpose_90")]; - tensor matrix_ac_cast_fp16 = matmul(transpose_x = matrix_ac_transpose_x_0, transpose_y = matrix_ac_transpose_y_0, x = transpose_83, y = transpose_84)[name = tensor("matrix_ac_cast_fp16")]; - tensor matrix_bd_begin_0 = const()[name = tensor("matrix_bd_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_end_0 = const()[name = tensor("matrix_bd_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_end_mask_0 = const()[name = tensor("matrix_bd_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_cast_fp16 = slice_by_index(begin = matrix_bd_begin_0, end = matrix_bd_end_0, end_mask = matrix_bd_end_mask_0, x = matrix_bd_65_cast_fp16)[name = tensor("matrix_bd_cast_fp16")]; - tensor var_3202_cast_fp16 = add(x = matrix_ac_cast_fp16, y = matrix_bd_cast_fp16)[name = tensor("op_3202_cast_fp16")]; - tensor _inversed_scores_65_y_0_to_fp16 = const()[name = tensor("_inversed_scores_65_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_65_cast_fp16 = mul(x = var_3202_cast_fp16, y = _inversed_scores_65_y_0_to_fp16)[name = tensor("_inversed_scores_65_cast_fp16")]; - tensor scores_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_65_cast_fp16, cond = mask_11)[name = tensor("scores_cast_fp16")]; - tensor var_3208_cast_fp16 = softmax(axis = var_23, x = scores_cast_fp16)[name = tensor("op_3208_cast_fp16")]; - tensor input_865_cast_fp16 = select(a = var_11_to_fp16, b = var_3208_cast_fp16, cond = mask_11)[name = tensor("input_865_cast_fp16")]; - tensor x_365_transpose_x_0 = const()[name = tensor("x_365_transpose_x_0"), val = tensor(false)]; - tensor x_365_transpose_y_0 = const()[name = tensor("x_365_transpose_y_0"), val = tensor(false)]; - tensor value_cast_fp16 = transpose(perm = value_perm_0, x = v_cast_fp16)[name = tensor("transpose_92")]; - tensor x_365_cast_fp16 = matmul(transpose_x = x_365_transpose_x_0, transpose_y = x_365_transpose_y_0, x = input_865_cast_fp16, y = value_cast_fp16)[name = tensor("x_365_cast_fp16")]; - tensor var_3212_perm_0 = const()[name = tensor("op_3212_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_3213 = const()[name = tensor("op_3213"), val = tensor([1, -1, 512])]; - tensor var_3212_cast_fp16 = transpose(perm = var_3212_perm_0, x = x_365_cast_fp16)[name = tensor("transpose_88")]; - tensor input_867_cast_fp16 = reshape(shape = var_3213, x = var_3212_cast_fp16)[name = tensor("input_867_cast_fp16")]; - tensor module_layers_16_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_16_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208825216)))]; - tensor module_layers_16_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_16_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(209349568)))]; - tensor linear_151_cast_fp16 = linear(bias = module_layers_16_self_attn_linear_out_bias_to_fp16, weight = module_layers_16_self_attn_linear_out_weight_to_fp16, x = input_867_cast_fp16)[name = tensor("linear_151_cast_fp16")]; - tensor input_871_cast_fp16 = add(x = input_863_cast_fp16, y = linear_151_cast_fp16)[name = tensor("input_871_cast_fp16")]; - tensor x_369_axes_0 = const()[name = tensor("x_369_axes_0"), val = tensor([-1])]; - tensor module_layers_16_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_16_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(209350656)))]; - tensor module_layers_16_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_16_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(209351744)))]; - tensor x_369_cast_fp16 = layer_norm(axes = x_369_axes_0, beta = module_layers_16_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_16_norm_conv_weight_to_fp16, x = input_871_cast_fp16)[name = tensor("x_369_cast_fp16")]; - tensor input_873_perm_0 = const()[name = tensor("input_873_perm_0"), val = tensor([0, 2, 1])]; - tensor input_875_pad_type_0 = const()[name = tensor("input_875_pad_type_0"), val = tensor("valid")]; - tensor input_875_strides_0 = const()[name = tensor("input_875_strides_0"), val = tensor([1])]; - tensor input_875_pad_0 = const()[name = tensor("input_875_pad_0"), val = tensor([0, 0])]; - tensor input_875_dilations_0 = const()[name = tensor("input_875_dilations_0"), val = tensor([1])]; - tensor input_875_groups_0 = const()[name = tensor("input_875_groups_0"), val = tensor(1)]; - tensor module_layers_16_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_16_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(209352832)))]; - tensor module_layers_16_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_16_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210401472)))]; - tensor input_873_cast_fp16 = transpose(perm = input_873_perm_0, x = x_369_cast_fp16)[name = tensor("transpose_87")]; - tensor input_875_cast_fp16 = conv(bias = module_layers_16_conv_pointwise_conv1_bias_to_fp16, dilations = input_875_dilations_0, groups = input_875_groups_0, pad = input_875_pad_0, pad_type = input_875_pad_type_0, strides = input_875_strides_0, weight = module_layers_16_conv_pointwise_conv1_weight_to_fp16, x = input_873_cast_fp16)[name = tensor("input_875_cast_fp16")]; - tensor x_371_split_num_splits_0 = const()[name = tensor("x_371_split_num_splits_0"), val = tensor(2)]; - tensor x_371_split_axis_0 = const()[name = tensor("x_371_split_axis_0"), val = tensor(1)]; - tensor x_371_split_cast_fp16_0, tensor x_371_split_cast_fp16_1 = split(axis = x_371_split_axis_0, num_splits = x_371_split_num_splits_0, x = input_875_cast_fp16)[name = tensor("x_371_split_cast_fp16")]; - tensor x_371_split_1_sigmoid_cast_fp16 = sigmoid(x = x_371_split_cast_fp16_1)[name = tensor("x_371_split_1_sigmoid_cast_fp16")]; - tensor x_371_cast_fp16 = mul(x = x_371_split_cast_fp16_0, y = x_371_split_1_sigmoid_cast_fp16)[name = tensor("x_371_cast_fp16")]; - tensor input_877_cast_fp16 = select(a = var_11_to_fp16, b = x_371_cast_fp16, cond = var_453)[name = tensor("input_877_cast_fp16")]; - tensor input_879_pad_0 = const()[name = tensor("input_879_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_879_mode_0 = const()[name = tensor("input_879_mode_0"), val = tensor("constant")]; - tensor const_233_to_fp16 = const()[name = tensor("const_233_to_fp16"), val = tensor(0x0p+0)]; - tensor input_879_cast_fp16 = pad(constant_val = const_233_to_fp16, mode = input_879_mode_0, pad = input_879_pad_0, x = input_877_cast_fp16)[name = tensor("input_879_cast_fp16")]; - tensor input_881_pad_type_0 = const()[name = tensor("input_881_pad_type_0"), val = tensor("valid")]; - tensor input_881_groups_0 = const()[name = tensor("input_881_groups_0"), val = tensor(512)]; - tensor input_881_strides_0 = const()[name = tensor("input_881_strides_0"), val = tensor([1])]; - tensor input_881_pad_0 = const()[name = tensor("input_881_pad_0"), val = tensor([0, 0])]; - tensor input_881_dilations_0 = const()[name = tensor("input_881_dilations_0"), val = tensor([1])]; - tensor const_266_to_fp16 = const()[name = tensor("const_266_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210403584)))]; - tensor const_267_to_fp16 = const()[name = tensor("const_267_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210412864)))]; - tensor input_883_cast_fp16 = conv(bias = const_267_to_fp16, dilations = input_881_dilations_0, groups = input_881_groups_0, pad = input_881_pad_0, pad_type = input_881_pad_type_0, strides = input_881_strides_0, weight = const_266_to_fp16, x = input_879_cast_fp16)[name = tensor("input_883_cast_fp16")]; - tensor input_885_cast_fp16 = silu(x = input_883_cast_fp16)[name = tensor("input_885_cast_fp16")]; - tensor x_373_pad_type_0 = const()[name = tensor("x_373_pad_type_0"), val = tensor("valid")]; - tensor x_373_strides_0 = const()[name = tensor("x_373_strides_0"), val = tensor([1])]; - tensor x_373_pad_0 = const()[name = tensor("x_373_pad_0"), val = tensor([0, 0])]; - tensor x_373_dilations_0 = const()[name = tensor("x_373_dilations_0"), val = tensor([1])]; - tensor x_373_groups_0 = const()[name = tensor("x_373_groups_0"), val = tensor(1)]; - tensor module_layers_16_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_16_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210413952)))]; - tensor module_layers_16_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_16_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210938304)))]; - tensor x_373_cast_fp16 = conv(bias = module_layers_16_conv_pointwise_conv2_bias_to_fp16, dilations = x_373_dilations_0, groups = x_373_groups_0, pad = x_373_pad_0, pad_type = x_373_pad_type_0, strides = x_373_strides_0, weight = module_layers_16_conv_pointwise_conv2_weight_to_fp16, x = input_885_cast_fp16)[name = tensor("x_373_cast_fp16")]; - tensor input_887_perm_0 = const()[name = tensor("input_887_perm_0"), val = tensor([0, 2, 1])]; - tensor input_887_cast_fp16 = transpose(perm = input_887_perm_0, x = x_373_cast_fp16)[name = tensor("transpose_86")]; - tensor input_889_cast_fp16 = add(x = input_871_cast_fp16, y = input_887_cast_fp16)[name = tensor("input_889_cast_fp16")]; - tensor input_891_axes_0 = const()[name = tensor("input_891_axes_0"), val = tensor([-1])]; - tensor module_layers_16_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_16_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210939392)))]; - tensor module_layers_16_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_16_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210940480)))]; - tensor input_891_cast_fp16 = layer_norm(axes = input_891_axes_0, beta = module_layers_16_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_16_norm_feed_forward2_weight_to_fp16, x = input_889_cast_fp16)[name = tensor("input_891_cast_fp16")]; - tensor module_layers_16_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_16_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210941568)))]; - tensor module_layers_16_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_16_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(213038784)))]; - tensor linear_152_cast_fp16 = linear(bias = module_layers_16_feed_forward2_linear1_bias_to_fp16, weight = module_layers_16_feed_forward2_linear1_weight_to_fp16, x = input_891_cast_fp16)[name = tensor("linear_152_cast_fp16")]; - tensor input_895_cast_fp16 = silu(x = linear_152_cast_fp16)[name = tensor("input_895_cast_fp16")]; - tensor module_layers_16_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_16_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(213042944)))]; - tensor module_layers_16_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_16_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(215140160)))]; - tensor linear_153_cast_fp16 = linear(bias = module_layers_16_feed_forward2_linear2_bias_to_fp16, weight = module_layers_16_feed_forward2_linear2_weight_to_fp16, x = input_895_cast_fp16)[name = tensor("linear_153_cast_fp16")]; - tensor var_3279_to_fp16 = const()[name = tensor("op_3279_to_fp16"), val = tensor(0x1p-1)]; - tensor var_3280_cast_fp16 = mul(x = linear_153_cast_fp16, y = var_3279_to_fp16)[name = tensor("op_3280_cast_fp16")]; - tensor input_cast_fp16 = add(x = input_889_cast_fp16, y = var_3280_cast_fp16)[name = tensor("input_cast_fp16")]; - tensor audio_signal_axes_0 = const()[name = tensor("audio_signal_axes_0"), val = tensor([-1])]; - tensor module_layers_16_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_16_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(215141248)))]; - tensor module_layers_16_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_16_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(215142336)))]; - tensor audio_signal_cast_fp16 = layer_norm(axes = audio_signal_axes_0, beta = module_layers_16_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_16_norm_out_weight_to_fp16, x = input_cast_fp16)[name = tensor("audio_signal_cast_fp16")]; - tensor obj_1_perm_0 = const()[name = tensor("obj_1_perm_0"), val = tensor([0, 2, 1])]; - tensor obj_1_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("obj_1_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; - tensor obj_1_cast_fp16 = transpose(perm = obj_1_perm_0, x = audio_signal_cast_fp16)[name = tensor("transpose_85")]; - tensor encoder_output = cast(dtype = obj_1_cast_fp16_to_fp32_dtype_0, x = obj_1_cast_fp16)[name = tensor("cast_172")]; - } -> (encoder_output, encoder_length); -} \ No newline at end of file diff --git a/models/parakeet-ctc-110m-coreml/Encoder.mlmodelc/weights/weight.bin b/models/parakeet-ctc-110m-coreml/Encoder.mlmodelc/weights/weight.bin deleted file mode 100644 index c5bffa9667270bb21c093e55898f4fb2e299e426..0000000000000000000000000000000000000000 --- a/models/parakeet-ctc-110m-coreml/Encoder.mlmodelc/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:cecf7994b2758397d992802a4f6e5d656e3a1aeb7bbedc2aa430b1316d62474c -size 215143424 diff --git a/models/parakeet-ctc-110m-coreml/JointDecision.mlmodelc/analytics/coremldata.bin b/models/parakeet-ctc-110m-coreml/JointDecision.mlmodelc/analytics/coremldata.bin deleted file mode 100644 index 4b4b92e5e8e2b45c67e95237edd74d11f323a65a..0000000000000000000000000000000000000000 --- a/models/parakeet-ctc-110m-coreml/JointDecision.mlmodelc/analytics/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:983ba26dd9276b8d2d4f75f3475aefb1817c542df87dbd0fdac95bd63647494f -size 243 diff --git a/models/parakeet-ctc-110m-coreml/JointDecision.mlmodelc/coremldata.bin b/models/parakeet-ctc-110m-coreml/JointDecision.mlmodelc/coremldata.bin deleted file mode 100644 index 74e1d92e032200815d325fa2c7c84ee8562d6aad..0000000000000000000000000000000000000000 --- a/models/parakeet-ctc-110m-coreml/JointDecision.mlmodelc/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:0800e3bdf4ecb1bd46fd27e1826d33125cd574f9ae1e15dd9ff70ea42944ca2d -size 476 diff --git a/models/parakeet-ctc-110m-coreml/JointDecision.mlmodelc/metadata.json b/models/parakeet-ctc-110m-coreml/JointDecision.mlmodelc/metadata.json deleted file mode 100644 index fe49b70e7e98f7ec9524130b0027886d862f6f8a..0000000000000000000000000000000000000000 --- a/models/parakeet-ctc-110m-coreml/JointDecision.mlmodelc/metadata.json +++ /dev/null @@ -1,102 +0,0 @@ -[ - { - "metadataOutputVersion" : "3.0", - "shortDescription" : "Parakeet 110M joint + decision head (split, softmax, argmax)", - "outputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1 × 188 × 1)", - "shortDescription" : "", - "shape" : "[1, 188, 1]", - "name" : "token_id", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 188 × 1)", - "shortDescription" : "", - "shape" : "[1, 188, 1]", - "name" : "token_prob", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1 × 188 × 1)", - "shortDescription" : "", - "shape" : "[1, 188, 1]", - "name" : "duration", - "type" : "MultiArray" - } - ], - "storagePrecision" : "Float16", - "modelParameters" : [ - - ], - "author" : "Fluid Inference", - "specificationVersion" : 8, - "mlProgramOperationTypeHistogram" : { - "Ios17.reduceArgmax" : 2, - "Ios17.squeeze" : 1, - "Ios17.cast" : 4, - "Ios17.linear" : 3, - "Ios17.transpose" : 2, - "Ios17.sliceByIndex" : 2, - "Ios17.add" : 1, - "Ios16.relu" : 1, - "Ios16.softmax" : 1, - "Ios17.gatherAlongAxis" : 1, - "Ios17.expandDims" : 3 - }, - "computePrecision" : "Mixed (Float16, Float32, Int16, Int32)", - "isUpdatable" : "0", - "stateSchema" : [ - - ], - "availability" : { - "macOS" : "14.0", - "tvOS" : "17.0", - "visionOS" : "1.0", - "watchOS" : "10.0", - "iOS" : "17.0", - "macCatalyst" : "17.0" - }, - "modelType" : { - "name" : "MLModelType_mlProgram" - }, - "inputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 512 × 188)", - "shortDescription" : "", - "shape" : "[1, 512, 188]", - "name" : "encoder", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 640 × 1)", - "shortDescription" : "", - "shape" : "[1, 640, 1]", - "name" : "decoder", - "type" : "MultiArray" - } - ], - "userDefinedMetadata" : { - "com.github.apple.coremltools.version" : "8.3.0", - "com.github.apple.coremltools.source_dialect" : "TorchScript", - "com.github.apple.coremltools.source" : "torch==2.9.0" - }, - "generatedClassName" : "parakeet_joint_decision", - "method" : "predict" - } -] \ No newline at end of file diff --git a/models/parakeet-ctc-110m-coreml/JointDecision.mlmodelc/model.mil b/models/parakeet-ctc-110m-coreml/JointDecision.mlmodelc/model.mil deleted file mode 100644 index edfc5050f47ffbe1f5344799cef5dc0b21c392e0..0000000000000000000000000000000000000000 --- a/models/parakeet-ctc-110m-coreml/JointDecision.mlmodelc/model.mil +++ /dev/null @@ -1,58 +0,0 @@ -program(1.0) -[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3510.2.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.9.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "8.3.0"}})] -{ - func main(tensor decoder, tensor encoder) { - tensor input_1_perm_0 = const()[name = tensor("input_1_perm_0"), val = tensor([0, 2, 1])]; - tensor encoder_to_fp16_dtype_0 = const()[name = tensor("encoder_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor input_3_perm_0 = const()[name = tensor("input_3_perm_0"), val = tensor([0, 2, 1])]; - tensor decoder_to_fp16_dtype_0 = const()[name = tensor("decoder_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor joint_module_enc_weight_to_fp16 = const()[name = tensor("joint_module_enc_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; - tensor joint_module_enc_bias_to_fp16 = const()[name = tensor("joint_module_enc_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(655488)))]; - tensor encoder_to_fp16 = cast(dtype = encoder_to_fp16_dtype_0, x = encoder)[name = tensor("cast_6")]; - tensor input_1_cast_fp16 = transpose(perm = input_1_perm_0, x = encoder_to_fp16)[name = tensor("transpose_1")]; - tensor linear_0_cast_fp16 = linear(bias = joint_module_enc_bias_to_fp16, weight = joint_module_enc_weight_to_fp16, x = input_1_cast_fp16)[name = tensor("linear_0_cast_fp16")]; - tensor joint_module_pred_weight_to_fp16 = const()[name = tensor("joint_module_pred_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(656832)))]; - tensor joint_module_pred_bias_to_fp16 = const()[name = tensor("joint_module_pred_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1476096)))]; - tensor decoder_to_fp16 = cast(dtype = decoder_to_fp16_dtype_0, x = decoder)[name = tensor("cast_5")]; - tensor input_3_cast_fp16 = transpose(perm = input_3_perm_0, x = decoder_to_fp16)[name = tensor("transpose_0")]; - tensor linear_1_cast_fp16 = linear(bias = joint_module_pred_bias_to_fp16, weight = joint_module_pred_weight_to_fp16, x = input_3_cast_fp16)[name = tensor("linear_1_cast_fp16")]; - tensor var_23_axes_0 = const()[name = tensor("op_23_axes_0"), val = tensor([2])]; - tensor var_23_cast_fp16 = expand_dims(axes = var_23_axes_0, x = linear_0_cast_fp16)[name = tensor("op_23_cast_fp16")]; - tensor var_24_axes_0 = const()[name = tensor("op_24_axes_0"), val = tensor([1])]; - tensor var_24_cast_fp16 = expand_dims(axes = var_24_axes_0, x = linear_1_cast_fp16)[name = tensor("op_24_cast_fp16")]; - tensor input_5_cast_fp16 = add(x = var_23_cast_fp16, y = var_24_cast_fp16)[name = tensor("input_5_cast_fp16")]; - tensor input_7_cast_fp16 = relu(x = input_5_cast_fp16)[name = tensor("input_7_cast_fp16")]; - tensor joint_module_joint_net_2_weight_to_fp16 = const()[name = tensor("joint_module_joint_net_2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1477440)))]; - tensor joint_module_joint_net_2_bias_to_fp16 = const()[name = tensor("joint_module_joint_net_2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2795904)))]; - tensor linear_2_cast_fp16 = linear(bias = joint_module_joint_net_2_bias_to_fp16, weight = joint_module_joint_net_2_weight_to_fp16, x = input_7_cast_fp16)[name = tensor("linear_2_cast_fp16")]; - tensor token_logits_begin_0 = const()[name = tensor("token_logits_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor token_logits_end_0 = const()[name = tensor("token_logits_end_0"), val = tensor([1, 188, 1, 1025])]; - tensor token_logits_end_mask_0 = const()[name = tensor("token_logits_end_mask_0"), val = tensor([true, true, true, false])]; - tensor token_logits_cast_fp16 = slice_by_index(begin = token_logits_begin_0, end = token_logits_end_0, end_mask = token_logits_end_mask_0, x = linear_2_cast_fp16)[name = tensor("token_logits_cast_fp16")]; - tensor duration_logits_begin_0 = const()[name = tensor("duration_logits_begin_0"), val = tensor([0, 0, 0, 1025])]; - tensor duration_logits_end_0 = const()[name = tensor("duration_logits_end_0"), val = tensor([1, 188, 1, 1030])]; - tensor duration_logits_end_mask_0 = const()[name = tensor("duration_logits_end_mask_0"), val = tensor([true, true, true, true])]; - tensor duration_logits_cast_fp16 = slice_by_index(begin = duration_logits_begin_0, end = duration_logits_end_0, end_mask = duration_logits_end_mask_0, x = linear_2_cast_fp16)[name = tensor("duration_logits_cast_fp16")]; - tensor var_43_axis_0 = const()[name = tensor("op_43_axis_0"), val = tensor(-1)]; - tensor var_43_keep_dims_0 = const()[name = tensor("op_43_keep_dims_0"), val = tensor(false)]; - tensor var_43_output_dtype_0 = const()[name = tensor("op_43_output_dtype_0"), val = tensor("int32")]; - tensor token_id = reduce_argmax(axis = var_43_axis_0, keep_dims = var_43_keep_dims_0, output_dtype = var_43_output_dtype_0, x = token_logits_cast_fp16)[name = tensor("op_43_cast_fp16")]; - tensor var_49 = const()[name = tensor("op_49"), val = tensor(-1)]; - tensor token_probs_all_cast_fp16 = softmax(axis = var_49, x = token_logits_cast_fp16)[name = tensor("token_probs_all_cast_fp16")]; - tensor var_58_axes_0 = const()[name = tensor("op_58_axes_0"), val = tensor([-1])]; - tensor var_58 = expand_dims(axes = var_58_axes_0, x = token_id)[name = tensor("op_58")]; - tensor var_59 = const()[name = tensor("op_59"), val = tensor(-1)]; - tensor var_61_validate_indices_0 = const()[name = tensor("op_61_validate_indices_0"), val = tensor(false)]; - tensor var_58_to_int16_dtype_0 = const()[name = tensor("op_58_to_int16_dtype_0"), val = tensor("int16")]; - tensor var_58_to_int16 = cast(dtype = var_58_to_int16_dtype_0, x = var_58)[name = tensor("cast_4")]; - tensor var_61_cast_fp16_cast_int16 = gather_along_axis(axis = var_59, indices = var_58_to_int16, validate_indices = var_61_validate_indices_0, x = token_probs_all_cast_fp16)[name = tensor("op_61_cast_fp16_cast_int16")]; - tensor var_63_axes_0 = const()[name = tensor("op_63_axes_0"), val = tensor([-1])]; - tensor var_63_cast_fp16 = squeeze(axes = var_63_axes_0, x = var_61_cast_fp16_cast_int16)[name = tensor("op_63_cast_fp16")]; - tensor var_63_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("op_63_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; - tensor var_66_axis_0 = const()[name = tensor("op_66_axis_0"), val = tensor(-1)]; - tensor var_66_keep_dims_0 = const()[name = tensor("op_66_keep_dims_0"), val = tensor(false)]; - tensor var_66_output_dtype_0 = const()[name = tensor("op_66_output_dtype_0"), val = tensor("int32")]; - tensor duration = reduce_argmax(axis = var_66_axis_0, keep_dims = var_66_keep_dims_0, output_dtype = var_66_output_dtype_0, x = duration_logits_cast_fp16)[name = tensor("op_66_cast_fp16")]; - tensor token_prob = cast(dtype = var_63_cast_fp16_to_fp32_dtype_0, x = var_63_cast_fp16)[name = tensor("cast_3")]; - } -> (token_id, token_prob, duration); -} \ No newline at end of file diff --git a/models/parakeet-ctc-110m-coreml/JointDecision.mlmodelc/weights/weight.bin b/models/parakeet-ctc-110m-coreml/JointDecision.mlmodelc/weights/weight.bin deleted file mode 100644 index 747ec9d24e2d3e008bc869d2c404bb3d2653325c..0000000000000000000000000000000000000000 --- a/models/parakeet-ctc-110m-coreml/JointDecision.mlmodelc/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:b3f771cb65b190f1873e39629676ed79b65a8361522f451b37bdba8b1106e6ff -size 2798028 diff --git a/models/parakeet-ctc-110m-coreml/JointDecisionSingleStep.mlmodelc/analytics/coremldata.bin b/models/parakeet-ctc-110m-coreml/JointDecisionSingleStep.mlmodelc/analytics/coremldata.bin deleted file mode 100644 index ab3e67b973cf846c0af5e1f075f1b4b7ffb77a00..0000000000000000000000000000000000000000 --- a/models/parakeet-ctc-110m-coreml/JointDecisionSingleStep.mlmodelc/analytics/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:c7c11c6bb985fab7f835ba687a575f1eb04f4c93b0783155d634adbc49f0e797 -size 243 diff --git a/models/parakeet-ctc-110m-coreml/JointDecisionSingleStep.mlmodelc/coremldata.bin b/models/parakeet-ctc-110m-coreml/JointDecisionSingleStep.mlmodelc/coremldata.bin deleted file mode 100644 index bc71cc6aa73de6959f4c6718500197ebbf364633..0000000000000000000000000000000000000000 --- a/models/parakeet-ctc-110m-coreml/JointDecisionSingleStep.mlmodelc/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:1af2cb9bcc13eec83ce006e4f1c2cf158393745cd9187428333fbcb6917da244 -size 535 diff --git a/models/parakeet-ctc-110m-coreml/JointDecisionSingleStep.mlmodelc/metadata.json b/models/parakeet-ctc-110m-coreml/JointDecisionSingleStep.mlmodelc/metadata.json deleted file mode 100644 index b96db3615aa1d5e110c61c1f4694def827cb0a1f..0000000000000000000000000000000000000000 --- a/models/parakeet-ctc-110m-coreml/JointDecisionSingleStep.mlmodelc/metadata.json +++ /dev/null @@ -1,123 +0,0 @@ -[ - { - "metadataOutputVersion" : "3.0", - "shortDescription" : "Parakeet 110M single-step joint decision (current frame)", - "outputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1 × 1 × 1)", - "shortDescription" : "", - "shape" : "[1, 1, 1]", - "name" : "token_id", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 1 × 1)", - "shortDescription" : "", - "shape" : "[1, 1, 1]", - "name" : "token_prob", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1 × 1 × 1)", - "shortDescription" : "", - "shape" : "[1, 1, 1]", - "name" : "duration", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1 × 1 × 1 × 64)", - "shortDescription" : "", - "shape" : "[1, 1, 1, 64]", - "name" : "top_k_ids", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 1 × 1 × 64)", - "shortDescription" : "", - "shape" : "[1, 1, 1, 64]", - "name" : "top_k_logits", - "type" : "MultiArray" - } - ], - "storagePrecision" : "Float16", - "modelParameters" : [ - - ], - "author" : "Fluid Inference", - "specificationVersion" : 8, - "mlProgramOperationTypeHistogram" : { - "Ios17.reduceArgmax" : 2, - "Ios17.linear" : 3, - "Ios17.transpose" : 2, - "Ios17.sliceByIndex" : 2, - "Ios17.add" : 1, - "Ios17.topk" : 1, - "Ios16.relu" : 1, - "Ios16.softmax" : 1, - "Ios17.expandDims" : 3, - "Ios17.squeeze" : 1, - "Ios17.cast" : 6, - "Ios17.gatherAlongAxis" : 1 - }, - "computePrecision" : "Mixed (Float16, Float32, Int16, Int32, UInt16)", - "isUpdatable" : "0", - "stateSchema" : [ - - ], - "availability" : { - "macOS" : "14.0", - "tvOS" : "17.0", - "visionOS" : "1.0", - "watchOS" : "10.0", - "iOS" : "17.0", - "macCatalyst" : "17.0" - }, - "modelType" : { - "name" : "MLModelType_mlProgram" - }, - "inputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 512 × 1)", - "shortDescription" : "", - "shape" : "[1, 512, 1]", - "name" : "encoder_step", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 640 × 1)", - "shortDescription" : "", - "shape" : "[1, 640, 1]", - "name" : "decoder_step", - "type" : "MultiArray" - } - ], - "userDefinedMetadata" : { - "com.github.apple.coremltools.source_dialect" : "TorchScript", - "com.github.apple.coremltools.source" : "torch==2.9.0", - "com.github.apple.coremltools.version" : "8.3.0" - }, - "generatedClassName" : "parakeet_joint_decision_single_step", - "method" : "predict" - } -] \ No newline at end of file diff --git a/models/parakeet-ctc-110m-coreml/JointDecisionSingleStep.mlmodelc/model.mil b/models/parakeet-ctc-110m-coreml/JointDecisionSingleStep.mlmodelc/model.mil deleted file mode 100644 index 42bd4ac5def601f37ece871ff09bd6242de025aa..0000000000000000000000000000000000000000 --- a/models/parakeet-ctc-110m-coreml/JointDecisionSingleStep.mlmodelc/model.mil +++ /dev/null @@ -1,69 +0,0 @@ -program(1.0) -[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3510.2.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.9.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "8.3.0"}})] -{ - func main(tensor decoder_step, tensor encoder_step) { - tensor input_1_perm_0 = const()[name = tensor("input_1_perm_0"), val = tensor([0, 2, 1])]; - tensor encoder_step_to_fp16_dtype_0 = const()[name = tensor("encoder_step_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor input_3_perm_0 = const()[name = tensor("input_3_perm_0"), val = tensor([0, 2, 1])]; - tensor decoder_step_to_fp16_dtype_0 = const()[name = tensor("decoder_step_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor joint_module_enc_weight_to_fp16 = const()[name = tensor("joint_module_enc_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; - tensor joint_module_enc_bias_to_fp16 = const()[name = tensor("joint_module_enc_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(655488)))]; - tensor encoder_step_to_fp16 = cast(dtype = encoder_step_to_fp16_dtype_0, x = encoder_step)[name = tensor("cast_9")]; - tensor input_1_cast_fp16 = transpose(perm = input_1_perm_0, x = encoder_step_to_fp16)[name = tensor("transpose_1")]; - tensor linear_0_cast_fp16 = linear(bias = joint_module_enc_bias_to_fp16, weight = joint_module_enc_weight_to_fp16, x = input_1_cast_fp16)[name = tensor("linear_0_cast_fp16")]; - tensor joint_module_pred_weight_to_fp16 = const()[name = tensor("joint_module_pred_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(656832)))]; - tensor joint_module_pred_bias_to_fp16 = const()[name = tensor("joint_module_pred_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1476096)))]; - tensor decoder_step_to_fp16 = cast(dtype = decoder_step_to_fp16_dtype_0, x = decoder_step)[name = tensor("cast_8")]; - tensor input_3_cast_fp16 = transpose(perm = input_3_perm_0, x = decoder_step_to_fp16)[name = tensor("transpose_0")]; - tensor linear_1_cast_fp16 = linear(bias = joint_module_pred_bias_to_fp16, weight = joint_module_pred_weight_to_fp16, x = input_3_cast_fp16)[name = tensor("linear_1_cast_fp16")]; - tensor var_23_axes_0 = const()[name = tensor("op_23_axes_0"), val = tensor([2])]; - tensor var_23_cast_fp16 = expand_dims(axes = var_23_axes_0, x = linear_0_cast_fp16)[name = tensor("op_23_cast_fp16")]; - tensor var_24_axes_0 = const()[name = tensor("op_24_axes_0"), val = tensor([1])]; - tensor var_24_cast_fp16 = expand_dims(axes = var_24_axes_0, x = linear_1_cast_fp16)[name = tensor("op_24_cast_fp16")]; - tensor input_5_cast_fp16 = add(x = var_23_cast_fp16, y = var_24_cast_fp16)[name = tensor("input_5_cast_fp16")]; - tensor input_7_cast_fp16 = relu(x = input_5_cast_fp16)[name = tensor("input_7_cast_fp16")]; - tensor joint_module_joint_net_2_weight_to_fp16 = const()[name = tensor("joint_module_joint_net_2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1477440)))]; - tensor joint_module_joint_net_2_bias_to_fp16 = const()[name = tensor("joint_module_joint_net_2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2795904)))]; - tensor linear_2_cast_fp16 = linear(bias = joint_module_joint_net_2_bias_to_fp16, weight = joint_module_joint_net_2_weight_to_fp16, x = input_7_cast_fp16)[name = tensor("linear_2_cast_fp16")]; - tensor token_logits_begin_0 = const()[name = tensor("token_logits_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor token_logits_end_0 = const()[name = tensor("token_logits_end_0"), val = tensor([1, 1, 1, 1025])]; - tensor token_logits_end_mask_0 = const()[name = tensor("token_logits_end_mask_0"), val = tensor([true, true, true, false])]; - tensor token_logits_cast_fp16 = slice_by_index(begin = token_logits_begin_0, end = token_logits_end_0, end_mask = token_logits_end_mask_0, x = linear_2_cast_fp16)[name = tensor("token_logits_cast_fp16")]; - tensor duration_logits_begin_0 = const()[name = tensor("duration_logits_begin_0"), val = tensor([0, 0, 0, 1025])]; - tensor duration_logits_end_0 = const()[name = tensor("duration_logits_end_0"), val = tensor([1, 1, 1, 1030])]; - tensor duration_logits_end_mask_0 = const()[name = tensor("duration_logits_end_mask_0"), val = tensor([true, true, true, true])]; - tensor duration_logits_cast_fp16 = slice_by_index(begin = duration_logits_begin_0, end = duration_logits_end_0, end_mask = duration_logits_end_mask_0, x = linear_2_cast_fp16)[name = tensor("duration_logits_cast_fp16")]; - tensor var_43_axis_0 = const()[name = tensor("op_43_axis_0"), val = tensor(-1)]; - tensor var_43_keep_dims_0 = const()[name = tensor("op_43_keep_dims_0"), val = tensor(false)]; - tensor var_43_output_dtype_0 = const()[name = tensor("op_43_output_dtype_0"), val = tensor("int32")]; - tensor token_id = reduce_argmax(axis = var_43_axis_0, keep_dims = var_43_keep_dims_0, output_dtype = var_43_output_dtype_0, x = token_logits_cast_fp16)[name = tensor("op_43_cast_fp16")]; - tensor var_49 = const()[name = tensor("op_49"), val = tensor(-1)]; - tensor token_probs_all_cast_fp16 = softmax(axis = var_49, x = token_logits_cast_fp16)[name = tensor("token_probs_all_cast_fp16")]; - tensor var_58_axes_0 = const()[name = tensor("op_58_axes_0"), val = tensor([-1])]; - tensor var_58 = expand_dims(axes = var_58_axes_0, x = token_id)[name = tensor("op_58")]; - tensor var_59 = const()[name = tensor("op_59"), val = tensor(-1)]; - tensor var_61_validate_indices_0 = const()[name = tensor("op_61_validate_indices_0"), val = tensor(false)]; - tensor var_58_to_int16_dtype_0 = const()[name = tensor("op_58_to_int16_dtype_0"), val = tensor("int16")]; - tensor var_58_to_int16 = cast(dtype = var_58_to_int16_dtype_0, x = var_58)[name = tensor("cast_7")]; - tensor var_61_cast_fp16_cast_int16 = gather_along_axis(axis = var_59, indices = var_58_to_int16, validate_indices = var_61_validate_indices_0, x = token_probs_all_cast_fp16)[name = tensor("op_61_cast_fp16_cast_int16")]; - tensor var_63_axes_0 = const()[name = tensor("op_63_axes_0"), val = tensor([-1])]; - tensor var_63_cast_fp16 = squeeze(axes = var_63_axes_0, x = var_61_cast_fp16_cast_int16)[name = tensor("op_63_cast_fp16")]; - tensor var_63_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("op_63_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; - tensor var_66_axis_0 = const()[name = tensor("op_66_axis_0"), val = tensor(-1)]; - tensor var_66_keep_dims_0 = const()[name = tensor("op_66_keep_dims_0"), val = tensor(false)]; - tensor var_66_output_dtype_0 = const()[name = tensor("op_66_output_dtype_0"), val = tensor("int32")]; - tensor duration = reduce_argmax(axis = var_66_axis_0, keep_dims = var_66_keep_dims_0, output_dtype = var_66_output_dtype_0, x = duration_logits_cast_fp16)[name = tensor("op_66_cast_fp16")]; - tensor var_72 = const()[name = tensor("op_72"), val = tensor(64)]; - tensor var_76_axis_0 = const()[name = tensor("op_76_axis_0"), val = tensor(-1)]; - tensor var_76_ascending_0 = const()[name = tensor("op_76_ascending_0"), val = tensor(false)]; - tensor var_76_sort_0 = const()[name = tensor("op_76_sort_0"), val = tensor(true)]; - tensor var_76_return_indices_0 = const()[name = tensor("op_76_return_indices_0"), val = tensor(true)]; - tensor var_76_cast_fp16_cast_int16_output_indices_dtype_0 = const()[name = tensor("op_76_cast_fp16_cast_int16_output_indices_dtype_0"), val = tensor("uint16")]; - tensor var_76_cast_fp16_cast_int16_0, tensor var_76_cast_fp16_cast_int16_1 = topk(ascending = var_76_ascending_0, axis = var_76_axis_0, k = var_72, output_indices_dtype = var_76_cast_fp16_cast_int16_output_indices_dtype_0, return_indices = var_76_return_indices_0, sort = var_76_sort_0, x = token_logits_cast_fp16)[name = tensor("op_76_cast_fp16_cast_int16")]; - tensor var_76_cast_fp16_cast_int16_1_to_int32_dtype_0 = const()[name = tensor("op_76_cast_fp16_cast_int16_1_to_int32_dtype_0"), val = tensor("int32")]; - tensor var_76_cast_fp16_0_to_fp32_dtype_0 = const()[name = tensor("op_76_cast_fp16_0_to_fp32_dtype_0"), val = tensor("fp32")]; - tensor top_k_logits = cast(dtype = var_76_cast_fp16_0_to_fp32_dtype_0, x = var_76_cast_fp16_cast_int16_0)[name = tensor("cast_4")]; - tensor top_k_ids = cast(dtype = var_76_cast_fp16_cast_int16_1_to_int32_dtype_0, x = var_76_cast_fp16_cast_int16_1)[name = tensor("cast_5")]; - tensor token_prob = cast(dtype = var_63_cast_fp16_to_fp32_dtype_0, x = var_63_cast_fp16)[name = tensor("cast_6")]; - } -> (token_id, token_prob, duration, top_k_ids, top_k_logits); -} \ No newline at end of file diff --git a/models/parakeet-ctc-110m-coreml/JointDecisionSingleStep.mlmodelc/weights/weight.bin b/models/parakeet-ctc-110m-coreml/JointDecisionSingleStep.mlmodelc/weights/weight.bin deleted file mode 100644 index 747ec9d24e2d3e008bc869d2c404bb3d2653325c..0000000000000000000000000000000000000000 --- a/models/parakeet-ctc-110m-coreml/JointDecisionSingleStep.mlmodelc/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:b3f771cb65b190f1873e39629676ed79b65a8361522f451b37bdba8b1106e6ff -size 2798028 diff --git a/models/parakeet-ctc-110m-coreml/Preprocessor.mlmodelc/analytics/coremldata.bin b/models/parakeet-ctc-110m-coreml/Preprocessor.mlmodelc/analytics/coremldata.bin deleted file mode 100644 index d34457003aea6648633a14e7cd6134edc9c30f17..0000000000000000000000000000000000000000 --- a/models/parakeet-ctc-110m-coreml/Preprocessor.mlmodelc/analytics/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:a1ac15543fbb9301fba5f018b147e44d767479dec352aaa91dfe7bcf65949693 -size 243 diff --git a/models/parakeet-ctc-110m-coreml/Preprocessor.mlmodelc/coremldata.bin b/models/parakeet-ctc-110m-coreml/Preprocessor.mlmodelc/coremldata.bin deleted file mode 100644 index cd81f47cc6672a441cbe9556757a30ffc0ff0bc7..0000000000000000000000000000000000000000 --- a/models/parakeet-ctc-110m-coreml/Preprocessor.mlmodelc/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:4940877938cc1b6d8830bbdd68ac8a49377cc57d75b61308883da5235b6a1914 -size 439 diff --git a/models/parakeet-ctc-110m-coreml/Preprocessor.mlmodelc/metadata.json b/models/parakeet-ctc-110m-coreml/Preprocessor.mlmodelc/metadata.json deleted file mode 100644 index 6087d5e76f7dd060cf4f135f1f3ae0bbbeb2f30a..0000000000000000000000000000000000000000 --- a/models/parakeet-ctc-110m-coreml/Preprocessor.mlmodelc/metadata.json +++ /dev/null @@ -1,112 +0,0 @@ -[ - { - "metadataOutputVersion" : "3.0", - "shortDescription" : "Parakeet 110M preprocessor (15 s window)", - "outputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32)", - "shortDescription" : "", - "shape" : "[]", - "name" : "mel_features", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1)", - "shortDescription" : "", - "shape" : "[1]", - "name" : "mel_length", - "type" : "MultiArray" - } - ], - "storagePrecision" : "Float16", - "modelParameters" : [ - - ], - "author" : "Fluid Inference", - "specificationVersion" : 8, - "mlProgramOperationTypeHistogram" : { - "Range1d" : 3, - "Ios17.equal" : 1, - "Ios17.notEqual" : 1, - "Ios17.reshape" : 2, - "Identity" : 1, - "Ios17.matmul" : 1, - "Select" : 6, - "Ios17.expandDims" : 12, - "Ios17.add" : 3, - "Tile" : 2, - "Ios17.sliceByIndex" : 3, - "Ios16.reduceSum" : 4, - "Shape" : 4, - "Ios17.gather" : 4, - "Ios17.logicalNot" : 1, - "Pad" : 1, - "Ios17.log" : 1, - "Ios17.less" : 2, - "Ios17.sub" : 4, - "Ios17.conv" : 2, - "Ios17.pow" : 2, - "Ios17.cast" : 10, - "Ios17.concat" : 3, - "Stack" : 1, - "Ios17.floorDiv" : 1, - "Ios17.realDiv" : 4, - "Ios17.sqrt" : 1, - "Ios17.greaterEqual" : 1, - "Ios17.mul" : 1 - }, - "computePrecision" : "Mixed (Float16, Float32, Int16, Int32, UInt16)", - "isUpdatable" : "0", - "stateSchema" : [ - - ], - "availability" : { - "macOS" : "14.0", - "tvOS" : "17.0", - "visionOS" : "1.0", - "watchOS" : "10.0", - "iOS" : "17.0", - "macCatalyst" : "17.0" - }, - "modelType" : { - "name" : "MLModelType_mlProgram" - }, - "inputSchema" : [ - { - "dataType" : "Float32", - "hasShapeFlexibility" : "1", - "isOptional" : "0", - "shapeFlexibility" : "1 × 1...240000", - "shapeRange" : "[[1, 1], [1, 240000]]", - "formattedType" : "MultiArray (Float32 1 × 1)", - "type" : "MultiArray", - "shape" : "[1, 1]", - "name" : "audio", - "shortDescription" : "" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1)", - "shortDescription" : "", - "shape" : "[1]", - "name" : "audio_length", - "type" : "MultiArray" - } - ], - "userDefinedMetadata" : { - "com.github.apple.coremltools.source_dialect" : "TorchScript", - "com.github.apple.coremltools.source" : "torch==2.9.0", - "com.github.apple.coremltools.version" : "8.3.0" - }, - "generatedClassName" : "parakeet_preprocessor", - "method" : "predict" - } -] \ No newline at end of file diff --git a/models/parakeet-ctc-110m-coreml/Preprocessor.mlmodelc/model.mil b/models/parakeet-ctc-110m-coreml/Preprocessor.mlmodelc/model.mil deleted file mode 100644 index ae325b6071a8ee5dad74a2e516e60889d89771d8..0000000000000000000000000000000000000000 --- a/models/parakeet-ctc-110m-coreml/Preprocessor.mlmodelc/model.mil +++ /dev/null @@ -1,191 +0,0 @@ -program(1.0) -[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3510.2.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.9.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "8.3.0"}})] -{ - func main(tensor audio, tensor audio_length) [FlexibleShapeInformation = tuple, dict, tensor>>, tuple, dict, list, ?>>>>((("DefaultShapes", {{"audio", [1, 1]}}), ("RangeDims", {{"audio", [[1, 1], [1, 240000]]}})))] { - tensor var_9 = const()[name = tensor("op_9"), val = tensor(1)]; - tensor var_10 = const()[name = tensor("op_10"), val = tensor(160)]; - tensor var_12 = const()[name = tensor("op_12"), val = tensor(0)]; - tensor var_34 = const()[name = tensor("op_34"), val = tensor(512)]; - tensor var_35 = add(x = audio_length, y = var_34)[name = tensor("op_35")]; - tensor var_36 = const()[name = tensor("op_36"), val = tensor(512)]; - tensor var_37 = sub(x = var_35, y = var_36)[name = tensor("op_37")]; - tensor floor_div_0 = floor_div(x = var_37, y = var_10)[name = tensor("floor_div_0")]; - tensor var_40 = equal(x = audio_length, y = var_12)[name = tensor("op_40")]; - tensor var_41 = const()[name = tensor("op_41"), val = tensor([0])]; - tensor mel_length = select(a = var_41, b = floor_div_0, cond = var_40)[name = tensor("seq_len")]; - tensor audio_to_fp16_dtype_0 = const()[name = tensor("audio_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor audio_to_fp16 = cast(dtype = audio_to_fp16_dtype_0, x = audio)[name = tensor("cast_27")]; - tensor var_43_shape_cast_fp16 = shape(x = audio_to_fp16)[name = tensor("op_43_shape_cast_fp16")]; - tensor gather_0_axis_0 = const()[name = tensor("gather_0_axis_0"), val = tensor(0)]; - tensor gather_0_batch_dims_0 = const()[name = tensor("gather_0_batch_dims_0"), val = tensor(0)]; - tensor gather_0_validate_indices_0 = const()[name = tensor("gather_0_validate_indices_0"), val = tensor(false)]; - tensor var_43_shape_cast_fp16_to_int16_dtype_0 = const()[name = tensor("op_43_shape_cast_fp16_to_int16_dtype_0"), val = tensor("int16")]; - tensor select_0_to_uint16 = const()[name = tensor("select_0_to_uint16"), val = tensor(1)]; - tensor var_43_shape_cast_fp16_to_int16 = cast(dtype = var_43_shape_cast_fp16_to_int16_dtype_0, x = var_43_shape_cast_fp16)[name = tensor("cast_26")]; - tensor gather_0_cast_uint16 = gather(axis = gather_0_axis_0, batch_dims = gather_0_batch_dims_0, indices = select_0_to_uint16, validate_indices = gather_0_validate_indices_0, x = var_43_shape_cast_fp16_to_int16)[name = tensor("gather_0_cast_uint16")]; - tensor gather_0_cast_uint16_to_int32_dtype_0 = const()[name = tensor("gather_0_cast_uint16_to_int32_dtype_0"), val = tensor("int32")]; - tensor const_0 = const()[name = tensor("const_0"), val = tensor(0)]; - tensor const_1 = const()[name = tensor("const_1"), val = tensor(1)]; - tensor gather_0_cast_uint16_to_int32 = cast(dtype = gather_0_cast_uint16_to_int32_dtype_0, x = gather_0_cast_uint16)[name = tensor("cast_25")]; - tensor var_44 = range_1d(end = gather_0_cast_uint16_to_int32, start = const_0, step = const_1)[name = tensor("op_44")]; - tensor var_45_axes_0 = const()[name = tensor("op_45_axes_0"), val = tensor([0])]; - tensor var_45 = expand_dims(axes = var_45_axes_0, x = var_44)[name = tensor("op_45")]; - tensor var_46_axes_0 = const()[name = tensor("op_46_axes_0"), val = tensor([1])]; - tensor var_46 = expand_dims(axes = var_46_axes_0, x = audio_length)[name = tensor("op_46")]; - tensor timemask = less(x = var_45, y = var_46)[name = tensor("timemask")]; - tensor var_49_begin_0 = const()[name = tensor("op_49_begin_0"), val = tensor([0, 0])]; - tensor var_49_end_0 = const()[name = tensor("op_49_end_0"), val = tensor([1, 1])]; - tensor var_49_end_mask_0 = const()[name = tensor("op_49_end_mask_0"), val = tensor([true, false])]; - tensor var_49_squeeze_mask_0 = const()[name = tensor("op_49_squeeze_mask_0"), val = tensor([false, true])]; - tensor var_49_cast_fp16 = slice_by_index(begin = var_49_begin_0, end = var_49_end_0, end_mask = var_49_end_mask_0, squeeze_mask = var_49_squeeze_mask_0, x = audio_to_fp16)[name = tensor("op_49_cast_fp16")]; - tensor var_50_axes_0 = const()[name = tensor("op_50_axes_0"), val = tensor([1])]; - tensor var_50_cast_fp16 = expand_dims(axes = var_50_axes_0, x = var_49_cast_fp16)[name = tensor("op_50_cast_fp16")]; - tensor var_52_begin_0 = const()[name = tensor("op_52_begin_0"), val = tensor([0, 1])]; - tensor var_52_end_0 = const()[name = tensor("op_52_end_0"), val = tensor([1, 0])]; - tensor var_52_end_mask_0 = const()[name = tensor("op_52_end_mask_0"), val = tensor([true, true])]; - tensor var_52_cast_fp16 = slice_by_index(begin = var_52_begin_0, end = var_52_end_0, end_mask = var_52_end_mask_0, x = audio_to_fp16)[name = tensor("op_52_cast_fp16")]; - tensor var_54_begin_0 = const()[name = tensor("op_54_begin_0"), val = tensor([0, 0])]; - tensor var_54_end_0 = const()[name = tensor("op_54_end_0"), val = tensor([1, -1])]; - tensor var_54_end_mask_0 = const()[name = tensor("op_54_end_mask_0"), val = tensor([true, false])]; - tensor var_54_cast_fp16 = slice_by_index(begin = var_54_begin_0, end = var_54_end_0, end_mask = var_54_end_mask_0, x = audio_to_fp16)[name = tensor("op_54_cast_fp16")]; - tensor var_55_to_fp16 = const()[name = tensor("op_55_to_fp16"), val = tensor(0x1.f0cp-1)]; - tensor var_56_cast_fp16 = mul(x = var_54_cast_fp16, y = var_55_to_fp16)[name = tensor("op_56_cast_fp16")]; - tensor var_57_cast_fp16 = sub(x = var_52_cast_fp16, y = var_56_cast_fp16)[name = tensor("op_57_cast_fp16")]; - tensor x_3_interleave_0 = const()[name = tensor("x_3_interleave_0"), val = tensor(false)]; - tensor x_3_cast_fp16 = concat(axis = var_9, interleave = x_3_interleave_0, values = (var_50_cast_fp16, var_57_cast_fp16))[name = tensor("x_3_cast_fp16")]; - tensor var_60 = logical_not(x = timemask)[name = tensor("op_60")]; - tensor var_16_to_fp16 = const()[name = tensor("op_16_to_fp16"), val = tensor(0x0p+0)]; - tensor input_1_cast_fp16 = select(a = var_16_to_fp16, b = x_3_cast_fp16, cond = var_60)[name = tensor("input_1_cast_fp16")]; - tensor concat_1x = const()[name = tensor("concat_1x"), val = tensor([1, 1, -1])]; - tensor input_3_cast_fp16 = reshape(shape = concat_1x, x = input_1_cast_fp16)[name = tensor("input_3_cast_fp16")]; - tensor input_5_pad_0 = const()[name = tensor("input_5_pad_0"), val = tensor([0, 0, 0, 0, 256, 256])]; - tensor input_5_mode_0 = const()[name = tensor("input_5_mode_0"), val = tensor("constant")]; - tensor const_3_to_fp16 = const()[name = tensor("const_3_to_fp16"), val = tensor(0x0p+0)]; - tensor input_5_cast_fp16 = pad(constant_val = const_3_to_fp16, mode = input_5_mode_0, pad = input_5_pad_0, x = input_3_cast_fp16)[name = tensor("input_5_cast_fp16")]; - tensor concat_2x = const()[name = tensor("concat_2x"), val = tensor([1, -1])]; - tensor input_cast_fp16 = reshape(shape = concat_2x, x = input_5_cast_fp16)[name = tensor("input_cast_fp16")]; - tensor expand_dims_3 = const()[name = tensor("expand_dims_3"), val = tensor([160])]; - tensor expand_dims_4_axes_0 = const()[name = tensor("expand_dims_4_axes_0"), val = tensor([1])]; - tensor expand_dims_4_cast_fp16 = expand_dims(axes = expand_dims_4_axes_0, x = input_cast_fp16)[name = tensor("expand_dims_4_cast_fp16")]; - tensor conv_0_pad_type_0 = const()[name = tensor("conv_0_pad_type_0"), val = tensor("valid")]; - tensor conv_0_pad_0 = const()[name = tensor("conv_0_pad_0"), val = tensor([0, 0])]; - tensor conv_0_dilations_0 = const()[name = tensor("conv_0_dilations_0"), val = tensor([1])]; - tensor conv_0_groups_0 = const()[name = tensor("conv_0_groups_0"), val = tensor(1)]; - tensor expand_dims_1_to_fp16 = const()[name = tensor("expand_dims_1_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; - tensor conv_0_cast_fp16 = conv(dilations = conv_0_dilations_0, groups = conv_0_groups_0, pad = conv_0_pad_0, pad_type = conv_0_pad_type_0, strides = expand_dims_3, weight = expand_dims_1_to_fp16, x = expand_dims_4_cast_fp16)[name = tensor("conv_0_cast_fp16")]; - tensor conv_1_pad_type_0 = const()[name = tensor("conv_1_pad_type_0"), val = tensor("valid")]; - tensor conv_1_pad_0 = const()[name = tensor("conv_1_pad_0"), val = tensor([0, 0])]; - tensor conv_1_dilations_0 = const()[name = tensor("conv_1_dilations_0"), val = tensor([1])]; - tensor conv_1_groups_0 = const()[name = tensor("conv_1_groups_0"), val = tensor(1)]; - tensor expand_dims_2_to_fp16 = const()[name = tensor("expand_dims_2_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(263296)))]; - tensor conv_1_cast_fp16 = conv(dilations = conv_1_dilations_0, groups = conv_1_groups_0, pad = conv_1_pad_0, pad_type = conv_1_pad_type_0, strides = expand_dims_3, weight = expand_dims_2_to_fp16, x = expand_dims_4_cast_fp16)[name = tensor("conv_1_cast_fp16")]; - tensor stack_0_axis_0 = const()[name = tensor("stack_0_axis_0"), val = tensor(-1)]; - tensor stack_0_cast_fp16 = stack(axis = stack_0_axis_0, values = (conv_0_cast_fp16, conv_1_cast_fp16))[name = tensor("stack_0_cast_fp16")]; - tensor var_19_promoted_to_fp16 = const()[name = tensor("op_19_promoted_to_fp16"), val = tensor(0x1p+1)]; - tensor var_75_cast_fp16 = pow(x = stack_0_cast_fp16, y = var_19_promoted_to_fp16)[name = tensor("op_75_cast_fp16")]; - tensor var_77_axes_0 = const()[name = tensor("op_77_axes_0"), val = tensor([-1])]; - tensor var_77_keep_dims_0 = const()[name = tensor("op_77_keep_dims_0"), val = tensor(false)]; - tensor var_77_cast_fp16 = reduce_sum(axes = var_77_axes_0, keep_dims = var_77_keep_dims_0, x = var_75_cast_fp16)[name = tensor("op_77_cast_fp16")]; - tensor x_11_cast_fp16 = identity(x = var_77_cast_fp16)[name = tensor("x_11_cast_fp16")]; - tensor x_13_transpose_x_0 = const()[name = tensor("x_13_transpose_x_0"), val = tensor(false)]; - tensor x_13_transpose_y_0 = const()[name = tensor("x_13_transpose_y_0"), val = tensor(false)]; - tensor const_4_to_fp16 = const()[name = tensor("const_4_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(526528)))]; - tensor x_13_cast_fp16 = matmul(transpose_x = x_13_transpose_x_0, transpose_y = x_13_transpose_y_0, x = const_4_to_fp16, y = x_11_cast_fp16)[name = tensor("x_13_cast_fp16")]; - tensor var_84_to_fp16 = const()[name = tensor("op_84_to_fp16"), val = tensor(0x1p-24)]; - tensor var_85_cast_fp16 = add(x = x_13_cast_fp16, y = var_84_to_fp16)[name = tensor("op_85_cast_fp16")]; - tensor x_15_epsilon_0 = const()[name = tensor("x_15_epsilon_0"), val = tensor(0x1p-149)]; - tensor x_15_cast_fp16 = log(epsilon = x_15_epsilon_0, x = var_85_cast_fp16)[name = tensor("x_15_cast_fp16")]; - tensor var_87_shape_cast_fp16 = shape(x = x_15_cast_fp16)[name = tensor("op_87_shape_cast_fp16")]; - tensor gather_5 = const()[name = tensor("gather_5"), val = tensor(1)]; - tensor gather_6_axis_0 = const()[name = tensor("gather_6_axis_0"), val = tensor(0)]; - tensor gather_6_batch_dims_0 = const()[name = tensor("gather_6_batch_dims_0"), val = tensor(0)]; - tensor gather_6_validate_indices_0 = const()[name = tensor("gather_6_validate_indices_0"), val = tensor(false)]; - tensor var_87_shape_cast_fp16_to_uint16_dtype_0 = const()[name = tensor("op_87_shape_cast_fp16_to_uint16_dtype_0"), val = tensor("uint16")]; - tensor select_6_to_uint16 = const()[name = tensor("select_6_to_uint16"), val = tensor(2)]; - tensor var_87_shape_cast_fp16_to_uint16 = cast(dtype = var_87_shape_cast_fp16_to_uint16_dtype_0, x = var_87_shape_cast_fp16)[name = tensor("cast_24")]; - tensor gather_6_cast_uint16 = gather(axis = gather_6_axis_0, batch_dims = gather_6_batch_dims_0, indices = select_6_to_uint16, validate_indices = gather_6_validate_indices_0, x = var_87_shape_cast_fp16_to_uint16)[name = tensor("gather_6_cast_uint16")]; - tensor gather_6_cast_uint16_to_int32_dtype_0 = const()[name = tensor("gather_6_cast_uint16_to_int32_dtype_0"), val = tensor("int32")]; - tensor const_5 = const()[name = tensor("const_5"), val = tensor(0)]; - tensor const_6 = const()[name = tensor("const_6"), val = tensor(1)]; - tensor gather_6_cast_uint16_to_int32 = cast(dtype = gather_6_cast_uint16_to_int32_dtype_0, x = gather_6_cast_uint16)[name = tensor("cast_23")]; - tensor var_89 = range_1d(end = gather_6_cast_uint16_to_int32, start = const_5, step = const_6)[name = tensor("op_89")]; - tensor var_90_axes_0 = const()[name = tensor("op_90_axes_0"), val = tensor([0])]; - tensor var_90 = expand_dims(axes = var_90_axes_0, x = var_89)[name = tensor("op_90")]; - tensor concat_3_axis_0 = const()[name = tensor("concat_3_axis_0"), val = tensor(0)]; - tensor concat_3_interleave_0 = const()[name = tensor("concat_3_interleave_0"), val = tensor(false)]; - tensor concat_3 = concat(axis = concat_3_axis_0, interleave = concat_3_interleave_0, values = (gather_5, gather_6_cast_uint16_to_int32))[name = tensor("concat_3")]; - tensor shape_8 = shape(x = var_90)[name = tensor("shape_8")]; - tensor real_div_0 = real_div(x = concat_3, y = shape_8)[name = tensor("real_div_0")]; - tensor time_steps = tile(reps = real_div_0, x = var_90)[name = tensor("time_steps")]; - tensor var_93_axes_0 = const()[name = tensor("op_93_axes_0"), val = tensor([1])]; - tensor var_93 = expand_dims(axes = var_93_axes_0, x = mel_length)[name = tensor("op_93")]; - tensor valid_mask = less(x = time_steps, y = var_93)[name = tensor("valid_mask")]; - tensor var_95_axes_0 = const()[name = tensor("op_95_axes_0"), val = tensor([1])]; - tensor var_95 = expand_dims(axes = var_95_axes_0, x = valid_mask)[name = tensor("op_95")]; - tensor var_96_cast_fp16 = select(a = x_15_cast_fp16, b = var_16_to_fp16, cond = var_95)[name = tensor("op_96_cast_fp16")]; - tensor x_mean_numerator_axes_0 = const()[name = tensor("x_mean_numerator_axes_0"), val = tensor([2])]; - tensor x_mean_numerator_keep_dims_0 = const()[name = tensor("x_mean_numerator_keep_dims_0"), val = tensor(false)]; - tensor x_mean_numerator_cast_fp16 = reduce_sum(axes = x_mean_numerator_axes_0, keep_dims = x_mean_numerator_keep_dims_0, x = var_96_cast_fp16)[name = tensor("x_mean_numerator_cast_fp16")]; - tensor x_mean_denominator_axes_0 = const()[name = tensor("x_mean_denominator_axes_0"), val = tensor([1])]; - tensor x_mean_denominator_keep_dims_0 = const()[name = tensor("x_mean_denominator_keep_dims_0"), val = tensor(false)]; - tensor cast_6_to_fp16_dtype_0 = const()[name = tensor("cast_6_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor valid_mask_to_fp16 = cast(dtype = cast_6_to_fp16_dtype_0, x = valid_mask)[name = tensor("cast_22")]; - tensor x_mean_denominator_cast_fp16 = reduce_sum(axes = x_mean_denominator_axes_0, keep_dims = x_mean_denominator_keep_dims_0, x = valid_mask_to_fp16)[name = tensor("x_mean_denominator_cast_fp16")]; - tensor var_101_axes_0 = const()[name = tensor("op_101_axes_0"), val = tensor([1])]; - tensor var_101_cast_fp16 = expand_dims(axes = var_101_axes_0, x = x_mean_denominator_cast_fp16)[name = tensor("op_101_cast_fp16")]; - tensor x_mean_cast_fp16 = real_div(x = x_mean_numerator_cast_fp16, y = var_101_cast_fp16)[name = tensor("x_mean_cast_fp16")]; - tensor var_104_axes_0 = const()[name = tensor("op_104_axes_0"), val = tensor([2])]; - tensor var_104_cast_fp16 = expand_dims(axes = var_104_axes_0, x = x_mean_cast_fp16)[name = tensor("op_104_cast_fp16")]; - tensor var_105_cast_fp16 = sub(x = x_15_cast_fp16, y = var_104_cast_fp16)[name = tensor("op_105_cast_fp16")]; - tensor var_106_cast_fp16 = select(a = var_105_cast_fp16, b = var_16_to_fp16, cond = var_95)[name = tensor("op_106_cast_fp16")]; - tensor var_19_promoted_1_to_fp16 = const()[name = tensor("op_19_promoted_1_to_fp16"), val = tensor(0x1p+1)]; - tensor var_107_cast_fp16 = pow(x = var_106_cast_fp16, y = var_19_promoted_1_to_fp16)[name = tensor("op_107_cast_fp16")]; - tensor var_109_axes_0 = const()[name = tensor("op_109_axes_0"), val = tensor([2])]; - tensor var_109_keep_dims_0 = const()[name = tensor("op_109_keep_dims_0"), val = tensor(false)]; - tensor var_109_cast_fp16 = reduce_sum(axes = var_109_axes_0, keep_dims = var_109_keep_dims_0, x = var_107_cast_fp16)[name = tensor("op_109_cast_fp16")]; - tensor var_111_to_fp16 = const()[name = tensor("op_111_to_fp16"), val = tensor(0x1p+0)]; - tensor var_112_cast_fp16 = sub(x = var_101_cast_fp16, y = var_111_to_fp16)[name = tensor("op_112_cast_fp16")]; - tensor var_113_cast_fp16 = real_div(x = var_109_cast_fp16, y = var_112_cast_fp16)[name = tensor("op_113_cast_fp16")]; - tensor x_std_1_cast_fp16 = sqrt(x = var_113_cast_fp16)[name = tensor("x_std_1_cast_fp16")]; - tensor var_115_cast_fp16 = not_equal(x = x_std_1_cast_fp16, y = x_std_1_cast_fp16)[name = tensor("op_115_cast_fp16")]; - tensor x_std_3_cast_fp16 = select(a = var_16_to_fp16, b = x_std_1_cast_fp16, cond = var_115_cast_fp16)[name = tensor("x_std_3_cast_fp16")]; - tensor var_25_to_fp16 = const()[name = tensor("op_25_to_fp16"), val = tensor(0x1.5p-17)]; - tensor x_std_cast_fp16 = add(x = x_std_3_cast_fp16, y = var_25_to_fp16)[name = tensor("x_std_cast_fp16")]; - tensor var_120_axes_0 = const()[name = tensor("op_120_axes_0"), val = tensor([2])]; - tensor var_120_cast_fp16 = expand_dims(axes = var_120_axes_0, x = x_std_cast_fp16)[name = tensor("op_120_cast_fp16")]; - tensor x_cast_fp16 = real_div(x = var_105_cast_fp16, y = var_120_cast_fp16)[name = tensor("x_cast_fp16")]; - tensor var_122_shape_cast_fp16 = shape(x = x_cast_fp16)[name = tensor("op_122_shape_cast_fp16")]; - tensor gather_7_axis_0 = const()[name = tensor("gather_7_axis_0"), val = tensor(0)]; - tensor gather_7_batch_dims_0 = const()[name = tensor("gather_7_batch_dims_0"), val = tensor(0)]; - tensor gather_7_validate_indices_0 = const()[name = tensor("gather_7_validate_indices_0"), val = tensor(false)]; - tensor var_122_shape_cast_fp16_to_uint16_dtype_0 = const()[name = tensor("op_122_shape_cast_fp16_to_uint16_dtype_0"), val = tensor("uint16")]; - tensor select_7_to_uint16 = const()[name = tensor("select_7_to_uint16"), val = tensor(2)]; - tensor var_122_shape_cast_fp16_to_uint16 = cast(dtype = var_122_shape_cast_fp16_to_uint16_dtype_0, x = var_122_shape_cast_fp16)[name = tensor("cast_21")]; - tensor gather_7_cast_uint16 = gather(axis = gather_7_axis_0, batch_dims = gather_7_batch_dims_0, indices = select_7_to_uint16, validate_indices = gather_7_validate_indices_0, x = var_122_shape_cast_fp16_to_uint16)[name = tensor("gather_7_cast_uint16")]; - tensor gather_7_cast_uint16_to_int32_dtype_0 = const()[name = tensor("gather_7_cast_uint16_to_int32_dtype_0"), val = tensor("int32")]; - tensor const_7 = const()[name = tensor("const_7"), val = tensor(0)]; - tensor const_8 = const()[name = tensor("const_8"), val = tensor(1)]; - tensor gather_7_cast_uint16_to_int32 = cast(dtype = gather_7_cast_uint16_to_int32_dtype_0, x = gather_7_cast_uint16)[name = tensor("cast_20")]; - tensor mask_1 = range_1d(end = gather_7_cast_uint16_to_int32, start = const_7, step = const_8)[name = tensor("mask_1")]; - tensor gather_8_axis_0 = const()[name = tensor("gather_8_axis_0"), val = tensor(0)]; - tensor gather_8_batch_dims_0 = const()[name = tensor("gather_8_batch_dims_0"), val = tensor(0)]; - tensor gather_8_validate_indices_0 = const()[name = tensor("gather_8_validate_indices_0"), val = tensor(false)]; - tensor select_8_to_uint16 = const()[name = tensor("select_8_to_uint16"), val = tensor(0)]; - tensor gather_8_cast_uint16 = gather(axis = gather_8_axis_0, batch_dims = gather_8_batch_dims_0, indices = select_8_to_uint16, validate_indices = gather_8_validate_indices_0, x = var_122_shape_cast_fp16_to_uint16)[name = tensor("gather_8_cast_uint16")]; - tensor gather_8_cast_uint16_to_int32_dtype_0 = const()[name = tensor("gather_8_cast_uint16_to_int32_dtype_0"), val = tensor("int32")]; - tensor concat_4_axis_0 = const()[name = tensor("concat_4_axis_0"), val = tensor(0)]; - tensor concat_4_interleave_0 = const()[name = tensor("concat_4_interleave_0"), val = tensor(false)]; - tensor gather_8_cast_uint16_to_int32 = cast(dtype = gather_8_cast_uint16_to_int32_dtype_0, x = gather_8_cast_uint16)[name = tensor("cast_19")]; - tensor concat_4 = concat(axis = concat_4_axis_0, interleave = concat_4_interleave_0, values = (gather_8_cast_uint16_to_int32, var_9))[name = tensor("concat_4")]; - tensor expand_dims_0_axes_0 = const()[name = tensor("expand_dims_0_axes_0"), val = tensor([0])]; - tensor expand_dims_0 = expand_dims(axes = expand_dims_0_axes_0, x = mask_1)[name = tensor("expand_dims_0")]; - tensor var_126 = tile(reps = concat_4, x = expand_dims_0)[name = tensor("op_126")]; - tensor mask = greater_equal(x = var_126, y = var_93)[name = tensor("mask")]; - tensor var_129_axes_0 = const()[name = tensor("op_129_axes_0"), val = tensor([1])]; - tensor var_129 = expand_dims(axes = var_129_axes_0, x = mask)[name = tensor("op_129")]; - tensor cast_15_to_fp16 = const()[name = tensor("cast_15_to_fp16"), val = tensor(0x0p+0)]; - tensor processed_signal_cast_fp16 = select(a = cast_15_to_fp16, b = x_cast_fp16, cond = var_129)[name = tensor("processed_signal_cast_fp16")]; - tensor processed_signal_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("processed_signal_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; - tensor mel_features = cast(dtype = processed_signal_cast_fp16_to_fp32_dtype_0, x = processed_signal_cast_fp16)[name = tensor("cast_18")]; - } -> (mel_features, mel_length); -} \ No newline at end of file diff --git a/models/parakeet-ctc-110m-coreml/Preprocessor.mlmodelc/weights/weight.bin b/models/parakeet-ctc-110m-coreml/Preprocessor.mlmodelc/weights/weight.bin deleted file mode 100644 index d0c6ac87e5f78315662149af9ab4ea3658497dbe..0000000000000000000000000000000000000000 --- a/models/parakeet-ctc-110m-coreml/Preprocessor.mlmodelc/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:c062338de852a26607ce4101f74e6895de3a4134a57b07232bd72bfc6f1d7f1a -size 567712 diff --git a/models/parakeet-ctc-110m-coreml/config.json b/models/parakeet-ctc-110m-coreml/config.json deleted file mode 100644 index aaa7174ca014092519980c51451ac7ca3832a352..0000000000000000000000000000000000000000 --- a/models/parakeet-ctc-110m-coreml/config.json +++ /dev/null @@ -1,7 +0,0 @@ -{ - "bos_token_id": 1, - "eos_token_id": 2, - "nemo_model_type": "parakeet", - "pad_token_id": 0, - "vocab_size": 1024, -} diff --git a/models/parakeet-ctc-110m-coreml/metadata.json b/models/parakeet-ctc-110m-coreml/metadata.json deleted file mode 100644 index 6db872ae8a7d463c95dc180210ace0fce019f4a6..0000000000000000000000000000000000000000 --- a/models/parakeet-ctc-110m-coreml/metadata.json +++ /dev/null @@ -1,247 +0,0 @@ -{ - "model_id": "nvidia/parakeet-tdt_ctc-110m", - "model_type": "hybrid_rnnt_ctc", - "sample_rate": 16000, - "max_audio_seconds": 15.0, - "max_audio_samples": 240000, - "max_symbol_steps": 1, - "vocab_size": 1024, - "joint_extra_outputs": 5, - "encoder_dim": 512, - "decoder_dim": 640, - "decoder_hidden": 640, - "decoder_layers": 1, - "blank_id": 1024, - "checkpoint": { - "type": "pretrained", - "model_id": "nvidia/parakeet-tdt_ctc-110m" - }, - "coreml": { - "compute_units": "CPU_ONLY", - "compute_precision": "FLOAT32" - }, - "components": { - "preprocessor": { - "inputs": { - "audio_signal": [ - 1, - 240000 - ], - "audio_length": [ - 1 - ] - }, - "outputs": { - "mel": [ - 1, - 80, - 1501 - ], - "mel_length": [ - 1 - ] - }, - "path": "parakeet_preprocessor.mlpackage" - }, - "encoder": { - "inputs": { - "mel": [ - 1, - 80, - 1501 - ], - "mel_length": [ - 1 - ] - }, - "outputs": { - "encoder": [ - 1, - 512, - 188 - ], - "encoder_length": [ - 1 - ] - }, - "path": "parakeet_encoder.mlpackage" - }, - "ctc_head": { - "inputs": { - "encoder": [ - 1, - 512, - 188 - ] - }, - "outputs": { - "log_probs": [ - 1, - 188, - 1025 - ] - }, - "path": "parakeet_ctc_head.mlpackage" - }, - "mel_encoder": { - "inputs": { - "audio_signal": [ - 1, - 240000 - ], - "audio_length": [ - 1 - ] - }, - "outputs": { - "encoder": [ - 1, - 512, - 188 - ], - "encoder_length": [ - 1 - ] - }, - "path": "parakeet_mel_encoder.mlpackage" - }, - "decoder": { - "inputs": { - "targets": [ - 1, - 1 - ], - "target_length": [ - 1 - ], - "h_in": [ - 1, - 1, - 640 - ], - "c_in": [ - 1, - 1, - 640 - ] - }, - "outputs": { - "decoder": [ - 1, - 640, - 1 - ], - "h_out": [ - 1, - 1, - 640 - ], - "c_out": [ - 1, - 1, - 640 - ] - }, - "path": "parakeet_decoder.mlpackage" - }, - "joint": { - "inputs": { - "encoder": [ - 1, - 512, - 188 - ], - "decoder": [ - 1, - 640, - 1 - ] - }, - "outputs": { - "logits": [ - 1, - 188, - 1, - 1030 - ] - }, - "path": "parakeet_joint.mlpackage" - }, - "joint_decision": { - "inputs": { - "encoder": [ - 1, - 512, - 188 - ], - "decoder": [ - 1, - 640, - 1 - ] - }, - "outputs": { - "token_id": [ - 1, - 188, - 1 - ], - "token_prob": [ - 1, - 188, - 1 - ], - "duration": [ - 1, - 188, - 1 - ] - }, - "path": "parakeet_joint_decision.mlpackage" - }, - "joint_decision_single_step": { - "inputs": { - "encoder_step": [ - 1, - 512, - 1 - ], - "decoder_step": [ - 1, - 640, - 1 - ] - }, - "outputs": { - "token_id": [ - 1, - 1, - 1 - ], - "token_prob": [ - 1, - 1, - 1 - ], - "duration": [ - 1, - 1, - 1 - ], - "top_k_ids": [ - 1, - 1, - 1, - 64 - ], - "top_k_logits": [ - 1, - 1, - 1, - 64 - ] - }, - "path": "parakeet_joint_decision_single_step.mlpackage" - } - } -} \ No newline at end of file diff --git a/models/parakeet-ctc-110m-coreml/preprocessor_config.json b/models/parakeet-ctc-110m-coreml/preprocessor_config.json deleted file mode 100644 index 54683c2afb1e841dfe59d03ad92c060a6cf50c0d..0000000000000000000000000000000000000000 --- a/models/parakeet-ctc-110m-coreml/preprocessor_config.json +++ /dev/null @@ -1,13 +0,0 @@ -{ - "feature_extractor_type": "ParakeetFeatureExtractor", - "feature_size": 80, - "hop_length": 160, - "n_fft": 512, - "padding_side": "right", - "padding_value": 0.0, - "preemphasis": 0.97, - "processor_class": "ParakeetProcessor", - "return_attention_mask": true, - "sampling_rate": 16000, - "win_length": 400 -} diff --git a/models/parakeet-ctc-110m-coreml/special_tokens_map.json b/models/parakeet-ctc-110m-coreml/special_tokens_map.json deleted file mode 100644 index 815fbdc56abc60a0f29e51ff2373c998d7779805..0000000000000000000000000000000000000000 --- a/models/parakeet-ctc-110m-coreml/special_tokens_map.json +++ /dev/null @@ -1,16 +0,0 @@ -{ - "pad_token": { - "content": "", - "lstrip": false, - "normalized": false, - "rstrip": false, - "single_word": false - }, - "unk_token": { - "content": "", - "lstrip": false, - "normalized": false, - "rstrip": false, - "single_word": false - } -} diff --git a/models/parakeet-ctc-110m-coreml/tokenizer.json b/models/parakeet-ctc-110m-coreml/tokenizer.json deleted file mode 100644 index 21e87f185d71820f5089456aa9c79c97c0a0ce8b..0000000000000000000000000000000000000000 --- a/models/parakeet-ctc-110m-coreml/tokenizer.json +++ /dev/null @@ -1,2456 +0,0 @@ -{ - "version": "1.0", - "truncation": null, - "padding": null, - "added_tokens": [ - { - "id": 0, - "content": "", - "single_word": false, - "lstrip": false, - "rstrip": false, - "normalized": false, - "special": true - }, - { - "id": 1024, - "content": "", - "single_word": false, - "lstrip": false, - "rstrip": false, - "normalized": false, - "special": true - } - ], - "normalizer": { - "type": "Sequence", - "normalizers": [ - { - "type": "Precompiled", - "precompiled_charsmap": "ALQCAACEAAAAAACAAQAAgMz8AgC4BQAAjSIAgMzkAgC4PQAAgSIAgMzsAgC4BQAAkSIAgMw8AADNvAAAngkAgKEJAICkCQCAgx0AAIAZAACBGQAAQx0AgDsdAIBTHQCASx0AgIAxAACBMQAApwkAgIkxAAA9WAMAPEgDAEMKAIA+aAMAAYUAAIQBAQADjQAAAokAAAWVAAAEkQAAB50AAAaZAAAJqQAACKEAAAutAAAKpQAADbkAAAy9AAAPvQAADrkAABHFAAAQwQAAE80AABLJAAAV1QAAFNEAABfdAAAW2QAAGeUAABjhAAAb7QAAGukAAB31AAAc8QAAH/0AAB75AABhOAkAax0AgGNADgBi8AgAZSgPAGSADgBn2A8AZvAPAGlwDABoMAwAa/AMAGrYDABtSA0AbBwNAG8QEgBubA0ASQoAgHAMEwBzqBMAcuwTAHUoEAB0TBAAd9ARAHYUEAB50BYAePQQAGMdAIB69BYAex0AgHMdAIB/fQEAiQwAgEGAAgDhCwCAQxgAAELAAABFSAAARGAAAEeQBgBGhAEASSgGAEhsAQBLOAcASvAHAE1wBwBMRAcAT/AEAE7MBACqCQCAUCwFAFOgCgBSEAUAVQAKAFRQCgBX0AgAVhALAFlICABYuAgAhBEAAFo8CACA9QAAgZ0AANsLAIAzHQCAg2kCAIJFAgCBNQIAgDUCAIdtAwCGVQMAgTkAAIRlAgAaDACAigEEAInVAwCI7QMAjwkAAKsLAIAsDACAjAkAADIMAICJMQMAkQkAAMzYAABbHQCAgx0AgMMaAIBPCgCAgGUDAIENAwCGPQAAgx0DAMwQAgDNhAEAgikAAMx0AwCjgQYAyxoAgICxAgCBsQIA0xoAgIEpAAClwQAA2xoAgMzoAwDNYAIAVQoAgKjxAABbCgCAYQoAgGcKAIDjGgCAgWkAAMzcBACCEQEA6xoAgG0KAIDzGgCAAxsAgAsbAID7GgCAtgkAgMygBADN3AQAzAgBALkJAICrHQCAhhEBAOEAKwDgfCcA44hIAuIMOAKjHQCAh5EBALsdAICzHQCAgNkBAIE1AADMxAIA6kRkApsdAIATGwCA7/hnAoERBwCC8QEA8BiGAolVAACB5QEAGxsAgIfhAQCAbQAAgQ0AAIN5AAB5CgCAgXkAAICVAQDMOAEAzRQBAIzBAQB/CgCAvwkAgKMVAQDDlBcAwpwUAMWEFwDEUBcAx+wXAMaAEgCTHQCAiwoAgMvQFgDK4BYAzRQWADgMAIDPvCAAzpwZANHMJADQ2CUA0+gkALFRAQA+DACAp90HAMMdAIDWvCQA2cgnANjUIgDb+CcAMxsAgIftBwCFCgCAzPgEACMbAIArHQCAh8kGALMJAICR3QcAvAkAgCsbAIBzCgCAOxsAgIsdAICPDACAjPkGAA4MAICA1QYAgcEGAMzEAgDNBAUAglEAAIN1BwCArQYAgbkGAIY1BwCHKQcAhEEAAJEKAICn7QAAQxsAgIjpBwCJzQcAlwoAgI/BBwCM3QcAnQoAgO0LAICnXQYAsJ0AAKMKAICpCgCAo0EGAEsbAIBbGwCAgAwAgFMbAIBjGwCArXEGAGsbAIDCCQCAzPgDAM0sAwDFCQCAo+UAAMgJAICMTQAAtQoAgKfxAAC7CgCAsT0GAIedAACGlQAAqB0HAISJAADBCgCAgqkAAIHVAACtAQcAzQoAgJE9AACCmQEAywkAgM0MBQDMCAUAgT0AAIeFAQCIvQEAexsAgMsdAICxCwCAjJEBAEQMAIBKDACA0x0AgID1AQCBhQEAgoEBAIOdAQCEiQEAxwoAgIapAQCHXQAAiG0AAIlNAABzGwCAzBACAIxdAACCDQAA0woAgI9JAACw6QAAgxsAgPMLAICjKQEAgCUBAIFVAQCLGwCApzUBAMykAQDNEAIA2QoAgJMbAICBNQAA3woAgK4JAQDrCgCAzOgBAM0oAgCbGwCAo/EAAIQFAACjGwCA5QoAgLMbAICotQAAqxsAgIFdAAC7GwCAzPwBAM3AAQDDGwCAyxsAgIGFAwAUDACAgeUDAPEKAICH6QMAzgkAgIylAwDTGwCA/QoAgK0JAIDbGwCAgZkDAIHdAwCMvQMAzSQBAMwgAQDMEAIAzTACAIH5AACHUQAAgFUAAIFZAAD3CgCAg0kAAIxBAADrGwCA4xsAgNEJAICBfQAAgHEAAMwgAwDNsAMAo30DANQJAICjEQMA8x0AgIEtAQCx/QAApzEDAK1BAwDrHQCAo20DAAMeAID7HQCA8xsAgKdtAwCANQAAgR0AALFtAwCILQAAmwwAgKeVAACBcQAAgFkAAINxAACj9QAAgVEAAK2BAAD7GwCAsQkDAIldAACEPQAAzDgBAISdAQCBGQAAgAkAAIRlAAADHACAzNAHAMzwBwALHACAkYkAAMxMBgDNBAYAzHAGAM10BgDMQAcAmy0PAMyoBwDNrAcAhg0AAIdVDwCEQQ8ADAsAgIIBDACDVQ8AgDUBAIHZAQCnDACAj+kAAIztAACVDACA4x0AgIv1AACIbQ8AiQ0AABILAIC3CwCAgiUAAFAMAICBQQAAVgwAgBseAIATHgCAKx4AgCMeAIAzHgCACx4AgIApAACBKQAA/wsAgBMcAICEeQAAGxwAgIFNAQCAoQEAGwsAgKP9DwDMOAIAzUgDACMcAICBWQAAzXwCAMykDQAnCwCAXAwAgKjJDwCHOQAA2gkAgImhDwAGCwCAkREAAKEMAIDdCQCAnAsAgGIMAICAuQ8AgbkPANsdAICDjQ8A+QsAgCscAICEBQAAMxwAgCELAIA7HACALQsAgIGdDwCHIQAAh7UPAMyoAgDN6AIAzLQMAM3cDACmzQAAp8UAAFMcAICPgQ8AjIkPAKPlAAAzCwCAQxwAgD8LAICxyQAAhwUAAFscAIBLHACAhz0AAGMcAIB0DACAOQsAgKMFDwCB+QAAzKgDAGscAIBLCwCAjEkAAKPxAABzHACAegwAgEULAICnlQAAgxwAgHscAIDMrAMAzcgAAOAJAICHaQAA4wkAgIG9AACCeQAA5gkAgIe5AQBRCwCAkaUAAIEdAACjHACAVwsAgIgFAACrHACAm5EAAF0LAIDpCQCAjJEBANULAIDJCwCAwwsAgM8LAICDRQAAgrkBAIG5AQCApQEAQx4AgIZxAABjCwCAhEkAAIsVAACKPQAAiTkAAIhFAACP+QAAaQsAgL0LAICMBQAAp1EBAKZJAQBoDACAsHkAAKNZAQCMqQAAgKkAAIGpAACBlQAAgJUAAK1xAQBuDACApQsAgISNAABTHgCASx4AgKMhAABjHgCAWx4AgGseAICBbQAAgG0AALEFAQCkOQAAOx4AgIscAIBvCwCAqAUAAJscAICTHACArQkAAMywAQCBvQMAgL0DAIPNAwCzHACAuxwAgMMcAIDMvAEAzYQBAInpAwDMHAEAgdkCAIDFAgDNOAEAzDwBAMxoAgDNRAIAg00AAMscAICH2QAAhy0AAIBFAACBEQAAggUAAHULAIDbHACA0xwAgOMcAIDMOAIAiBUAAIjhAACAbQAAgTkAAMyEAgDNUAEAo0UDAIQ5AQDrHACA8xwAgMzcAwDNSAIAcx4AgOwJAIB7CwCAix4AgK0MAICBbQAA+xwAgIELAICj0QAAgx4AgHseAIDMiAQAgXUAAIB1AACECwCAo7UAAMwABADNVAIAAx0AgIoLAICETQEAkAsAgAsdAIATHQCAzNAOAMwsAQDMAAUAzVwFAO8JAIDyCQCAzJgOAIHBAADMzA8AzDwOAMwIAQDNnA4AzNQPAM14DwDMPA4AzTgOAIHlAQCA5QEAg+UBAILlAQDXCQCAhOUBAIfhAQBHHQCAiaUBAIjZAQCByQcAPx0AgFcdAIBPHQCAzDQBAPgJAICA3QAAgekAAEYKAICD/QAAgM0AAIH5AACBEQcAbx0AgGcdAICJ0QAAzCgBAH8dAIB3HQCA5AsAgMw0AQDeCwCAgF0AAIFlAACjAQEAg2EAAIFxAACASQAANx0AgB0MAICuCwCAiVUAAC8MAIA1DACAXx0AgIcdAIDHGgCAUgoAgIIdAACDeQcAgBkHAIEZBwCGIQAAhykAAISRBwD1CQCAimkAALHZBgCIaQAAifUHAEwKAICP3QcAjNkHAIwMAID7CQCALx0AgP4JAICRoQcAgEEHAIFBBwCHBQAAzxoAgIKRBwDXGgCA3xoAgKOVBgCGhQcAp+0AAMyQAgDN4AUAsekAAKPBAABYCgCAXgoAgGQKAIBqCgCAAQoAgKVlBwDnGgCAzLgDAKhVBwDvGgCAcAoAgPcaAIAHGwCADxsAgP8aAIAECgCAo60AAAcKAICMJQYACgoAgIxNAACvHQCAgm0AAIE9BgCCAQYAgWUAAKcdAICHZQAAvx0AgIcRBgCHrQEAtx0AgMxQAgDNxAIAgeEBAIDJAQCD4QEAkYkAAID9AQCB1QEAnx0AgIydAQCJNQAAdgoAgIB1AACBXQAAhi0AAIc1AACEfQAAFxsAgIKFAQCDfQAAgJ0BAIGRAQAfGwCAj+kAAIzhAAB8CgCAggoAgA0KAICIDQAAifkAAKc5AQCXHQCAjgoAgDsMAICjJQEAQQwAgLBZAACPHQCAggUAAMcdAICtFQEAkgwAgDcbAICGBQAAiAoAgCcbAIAvGwCAp2kAAIANAQCBAQEAhzEAAKNJAACxGQEAzBACAD8bAIARDACAlAoAgK1RAADM1AEAzfgBAKhBAABHGwCAzTgBAMw8AQCB7QMAmgoAgKAKAICMDQAA8AsAgKYKAICBxQMAzGgCAKwKAICCxQMATxsAgITJAwCHKQAAhjEAAF8bAICCbQAAgwwAgFcbAICHYQAAZxsAgG8bAIAbHQCAzKgDAM2sAgCB+QAAiC0AABAKAIATCgCAFgoAgIw1AAC4CgCAvgoAgLHVAADECgCAfxsAgM8dAIC0CwCAzDABAEcMAIBNDACA1x0AgMwEAQDKCgCAdxsAgKelAADWCgCAo40AAMwUAgCAuQAAgbkAAKeFAAALDACAgmUAAIcbAICMNQAA9gsAgMzsHADN/AMAjxsAgK6tAADcCgCAlxsAgMzABgDN0AYAsL0BAMyQBwDiCgCAgckBAMwYHQDNIAIAhBEAAO4KAIDNuAYAzKwGAKcbAIDoCgCAgSkAALcbAICvGwCAo+0BAMxAHQDNEAIAvxsAgMcbAICBCQAAzxsAgMxAHQDN0AIAqNkBABcMAIDMkAcAzBwBAMxgBgDNZAYA9AoAgB8KAIDXGwCAkSkBAAALAICBzR8A3xsAgPoKAIDvGwCA5xsAgMzEBgDNwAYAgTEAAIDZAAAiCgCAJQoAgIK5AQCDRQEAgLkBAIG5AQCGXQEA9x0AgIRdAQDvHQCAzcAAAMzwAACIARwAiXkBAAceAICPVQEAjGEBAP8dAICB3R4AgRUfAJ8bAICBXR8AjIEfAIdBHwDMGAMAzWgDAIBNHwCBpR8AKAoAgIOpHwCMFR8AjNEeACsKAICHtR8AgJUfAIGZHwCBEQAAg70fAICFHwCBiR8A9xsAgIQ9AACeDACAiZkfAP8bAICIBQAACQsAgAccAICADQAAgf0AAA8cAICj2R8Ao3keAKOFAAAPCwCArTUfAKdhHgCnqR8ApAwAgIQNAACqDACAozUfAC4KAICtiR8AhHEAAKchHwCxPR4AsYUfAJgMAIDnHQCAFQsAgLoLAIDMtBwAzbAcAFMMAICxQR8AWQwAgJ8LAIAfHgCAFx4AgC8eAIAnHgCAgLkeAIG5HgCCIQEAgzUBAIRhAQA3HgCAhokBAIe9AQCIkQEAiekBAN8dAICL/QEAjOUBAIINAAAPHgCAj90BAIO5AQCRrQEAgb0BAIC9AQCAoQEAgaEBAPwLAIACDACAhD0AABccAICJlQEAm4EBAIHNHgCAzR4AzPwCAM3wAgCB5QAAHxwAgIHtAACjpQAAzJABAM1cAgCHHQAAHgsAgKj5AAAnHACAKgsAgF8MAIBlDACALxwAgIQFAAA3HACAo9UAACQLAIA/HACAgVEAAMz0AQDN0AEAMAsAgIc9AABXHACANgsAgEccAIBCCwCAhwUAAF8cAIBPHACAh/EDAIHZAwCBmQMAgZEAAGccAIB3DACAjPkDAMwkAQCHuQMAgfkDADwLAIDMZAIAgskDAIyZAwBvHACAh9EDAI+RAwCB3QYAkfUDAMwABADN7AMAh2UAAB8dAIBOCwCAdxwAgH0MAIBICwCAzBgBAIg5AACHHACAfxwAgMxcAwCMJQAAMQoAgMwsAQCx/QAAozkDADQKAIA3CgCApxwAgKdZAwDMdAMAiAkAAKNRAwCvHACAYAsAgINtDQCnnQAApq0AAKOdAACxDQMAzCgBANgLAICntQAAprUAAMwLAIDMMAEAgdUHAMYLAIDMKAEA0gsAgEceAIBmCwCArYkAAGwLAICAzQEAgd0BAMxEAQDNnB4AhPUBAMALAIDMWAEAzUwBAIDtAQCB/QEAg7UAAGsMAICM3QEAcQwAgMwIHgCM8QYAzDgBAM08AQBXHgCAiREAAIEFBgBPHgCAZx4AgF8eAIBvHgCAgz0AAIAhAACBOQAAgDkAAIEhAAA/HgCAjxwAgMwoAQCB2QYAcgsAgIH9BgDMJAEAnxwAgJccAIC3HACAgCEBAIE1AQCjBQAAvxwAgMccAIDPHACAzIwFAM1AAgC3HAMAeAsAgIfNBwDfHACA1xwAgCMdAIDNiAAAzJAAAIzdBQCjhQAAGQoAgMzgAgDnHACAiNUHAIFNAACATQAAVAsAgO8cAIBaCwCAkTkHADoKAICIxQcAqAsAgIrJBwD3HACAmz0AAIflBwB3HgCAgYUHAICFBwA9CgCAgvkHAILVBgCDRQAAgMkGAIHdBgCG4QYAfgsAgIRRAACPHgCAipUGAIuZBgCIeQAAiZ0GALAMAICPWQcAjG0HAP8cAIDMgAMAzSQCALARBwBACgCAhx4AgCcdAIB/HgCAhwsAgICNAACBnQAAzOwDAM3oBAAHHQCAjQsAgKNJBwCTCwCADx0AgKO9BwAXHQCAGwAAgOoHAIALAACApKUHAOsEAICKBQCAAwAAgKhhBwDfDQCAZQAAgMgDAIAeCQCArWkHAIAtAQCBPQEAgl0BAINRAQCEYQEAuAQAgKwEAICHYQEAiK0BAIm1AQCKvQEAjykVALwFAIAgDACAzHgCAM3YBQCB3QEAgXEAAOcLAICC/QEAhBkAACYMAICH7QEAIwwAgMw0BADNMAQA6gsAgJ9pFQApDACAjMkBAM34BADM8AIAsUkBACEHAICB1QAAoxUBAKCZFQB2CACARgcAgIT1AADMKAQAzSwEAMYIAICveQEAqH0BADcNAICqaQEAVQkAgLQlAQC1KQEAowkBAAUMAIDqBgCA7gYAgLIFAQCzPQEAvPUAAL39AAC+2QAAOwgAgLgBAQC5AQEAugEBADwHAIBDBwCAhgwAALOdAwCyiQMAtggAgIC9AwBsBwCAbwcAgBUJAIDkBgCA5wYAgDgIAICJhQMAzOQHAL+hAwAIDACA2gwAgIxlAADN5AwAzCQMAIlBAACIVQAAi0UAAIpFAACFtQMAhLUDAIeVAwCGgQMABA0AgAcNAIAKDQCAmCwAABMAAICmyAAAzYwGAMyoBgCFaQAAFwAAgDEAAIBpAACAzPADAAcAAIA1AACA1AwAgLGVAAArDQCAs5UAALKVAAA7DQCAPg0AgEYNAIBBDQCANA0AgHUAAICmBgCAJQAAgJsJAIAjIQCAv1UDAEkNAIAfIQCAGyEAgGcgAIC4bAAAlGUNAJIAAgCcrQEAnaUBAJqJAQCbiQEAmJkBAJmJAQDMIAYAzQQGAMxABgDNXAYAzDwHAM04BwDMvAcAhXUAAIABDwCBDQ8AbyAAgLqZAQCFBQAAdyAAgF8gAIC+hQEAgSkPAIAlDwBrIACAgiEPAIUpAAC0pQEAhREAAHMgAICziQ8AsoUPALHJAQCwAQwAt4EPALbtAQC17QEAtO0BAIFlAQCAZQEAg2EBALi1DwDMPAsAhHkBAIDhDwCB3Q8AeyAAgGMgAIDMyAQAzbgEAIWtAACFFQAAJyEAgD8hAIDM6BkAzbQZAKRdAQBMDQCAok0CAKPxDwCgVQEAod0PAIIIAIBxCQCAPgkAgPMeAIBvCQCA+x4AgHoJAID3HgCAtAgAgJMNAACzHgCA/x4AgITVDACF6Q4AlGkAAIfdDgC7HgCAmbQCAMMeAIDLHgCAtx4AgEMhAIC/HgCAn3QBAMceAICRGA0AgI0OAIGBDgCGhQ4AlYwDAISJDgCXRAIAghEAAKm4AACA0QAAge0AAM8eAIBPDQCA6x4AgIVZDwCDiQAAoTQNAIFFDgCASQ4A7x4AgKU0AQCFYQ8AzPAUACMfAIC5xAUAzMgDAM3cAwCA3QAAgcEAACsfAIC/kAUAhREAALHsBwCA9QAAgcEAAKcgAIC1jAYAMx8AgLdABgCA3Q4AgekOAMwoAgDNtAIAgM0OAIH5DgCFKQAAg4UBAIB1AQCBsQEAgPEBAIHVAQCvIACAOx8AgIUFAAC3IACAgJkBAIG9AQCCfQAAk9UBAJThAQCFDQAAnyAAgCcfAICACQAAgRkAAC8fAICTrQEAlC0AAKsgAICFDQAANx8AgIUFAACzIACAPx8AgIUpAACCGQAAhTUAAIDxAACB4QAAuyAAgKMgAIBHIQCAhQUAAGchAICDdQEAgO0BAIEpAQDM8AEAzbABAFINAIBjIQCAXyEAgKkNAIBjHwCAax8AgIA9AACBDQAAcx8AgHsfAICALQAAgR0AAIIVAABnHwCAzSwBAG8fAIB3HwCAfx8AgIjFAwCrIQCAzJACAM28AgCE7QMAVQ0AgIb5AwCjHwCAgIEDAIH9AwCAPQAAgTUAAIFJAACAQQAAzdwBAIJBAACrHwCApx8AgK8fAIDNMAEAlJ0DAJMhAIDN8AEAzAwBAIG5AwCAxQMAg6EDAJOlAwCArQAAgdUAAICdAACBqQAAjyEAgFgNAICBwQAAgMkAAIC1AACBgQAAiyEAgINpBADMcAMAzbQDAIchAIDNPAEArA0AgJMBBADNjAIAzPQCAIANAACBNQAAlNkGANcfAIDbHwCA3x8AgMwIAQDNHAEAgREAAIApAACvIQCAghkAAICRAQCBkQEAzWgFAMyUAgDMEAkAzSgWAMxYDgDNeA4AzBQNAM3YCgDMKAwAzYwNAMzgFwDM4AoAzDgLAM30CACFEQAAWw0AgIBRBwCBUQcA5yAAgM2QDgCFBQAA7yAAgMzYDgDN7AEA9yAAgM0ADgCFGQAAzfAPAM08DgDNVA4AzGgBAM1sAQDfIACAZAgAgJSZBwDMwDsAgGEBAIHZAACFKQAAzWQOAMx4AQDNfAEAga0HAICtBwCFZQAAgp0HAIBRAQCBUQEAlOEHAM3AAACEeQEAk8UHAIZhAQDrIACAiCEBAIUNAADzIACAzRgBAMzYAADNtAAAgN0HAIHNBwCfHwCAhQkAANMfAID7IACAAyAAgOMgAIALIACAEyAAgBsgAIAPIACAByAAgLMhAIAXIACAHyAAgMy4AgDNHAMAgGUAAIF1AACCfQAAIyAAgIUJAACFQQAAByEAgK8NAICAmQYAgSEHAIUZAACDfQAADyEAgIVZAAADIQCA/yAAgIDNAACB2QAAkx4AgIURAACE6QAAmx4AgIblAABHIACAgDUAAIENAACjHgCAhR0AAE8gAICrHgCAhQUAAFcgAICAVQAAgW0AAIJ9AACTRQAAlA0AAIUNAAA/IACAlx4AgIAJAACBEQAAnx4AgIUdAABLIACApx4AgIUFAABTIACAgOkBAIHxAQCCBQAArx4AgIUJAACFCQAAWyAAgEMgAICAbQEAgXkBAIIZAACDpQEAEyEAgIV1AACFBQAAFyEAgAshAIAnIACAzMgCAM3cAgCyDQCA0x4AgIA5AACBOQAA2x4AgOMeAIDXHgCA3x4AgIAdAACBDQAA5x4AgCsgAICAxQAAgdUAAM3AAADMJAIAgNUAAIHFAACFOQAAg8kAACshAIC1DQCAgNUAAIEJAACFBQAAMyEAgAMfAICHIACAgAkAAIERAAALHwCAk5kAAJS5AAATHwCAhWUAAIU9AACPIACAk10AABsfAICFEQAAzXAFAMx0BQCUATwAlyAAgH8gAIDNKAEAiyAAgJMgAICFGQAAmyAAgIMgAIA7IQCALyEAgC8gAICFJQAAhTkAAMz4AgDNxAMAzTwBALgNAICBlQMAgI0DAM3EAQCCpQMAhVEAAIVJAADMKAEAzSwBAM04AQDMPAEAgGk+AIFpPgBPIQCASyEAgM04PADMVDwAgdE8AJOdPgDMSAEAzcgCAM00AQBTIQCAlLk+AF4NAICAoT4AgaE+AIKhPgCIjTwAWyEAgIWtAACALQAAgSEAAIXVPwCbHwCAgO0AAIHxAACGpQAASx8AgISpAADNJAEAzSgBAFMfAICI+T4AhfE/AFsfAIBPHwCAhcU/AM0wAQDNEAEAzfQGAIDdAQCB6QEAzbwGAM1wBgDM4AYAzVwBAMxoBgDNkAYAzWQGAM14BgDMrAcAzagHAMzoBwDNyAcAgk0/AIP9AgCANQIAgekCAFcfAIBfHwCAgAU9AIV9AQBXIQCAMyAAgM0UAQAvDgCAge0BAIDhAQDNPAEAgs0BAM0sAQCCdQEAgW0BAIBZAQCAZQEAgcUAAIsfAIDNJAEAzTgBAILxAACB+QAAgFkBAIApAACBcQAAzBgBAM18AQDNLAEAkx8AgIEdAACAHQAAjx8AgJcfAIB3IQCAzSQBAMzkPQDNXA8AzegAAMwMAQCA1QEAgckBAIKZAACD5T8ADx8AgBcfAIAfHwCANyEAgCkOAIB7IQCAQx8AgDcgAIBHHwCAMg4AgIBNPwCBQT8Agx8AgG8hAICHHwCAayEAgIAlPwCBKT8Ak5E/AIN9AAAsDgCAlEEAAMzYAgDNrAIAcyEAgJNVAACACQAAgR0AALsNAICDIQCAlEEAALMfAICAnQAAgaEAAIAdAACBEQAAhKUAALsfAICGpQAAwx8AgIjxAACC0QAAgdkAAIDNAACAJQAAgSkAAIIFAADLHwCAtx8AgL8fAIDHHwCAk7EAAJQRAADPHwCAgB0AAIEVAACAJQAAgS0AAII9AAB/IQCAgO0AAIHRAACCFQAAg4EAAIHQPQA7IACAzCACAM3cAQCFeAIAlyEAgDUOAICfIQCAiRgDAOMfAICALQAAgTUAAIAJAACBbQAA6x8AgMcgAICRsQAAkKkAAJPdOwCSAQQAlaUAAJSVOwDzHwCAlqEAAIUJAACTQQAAzyAAgPsfAICFBQAA1yAAgJT1AAC/IACAgLkAAIHdAACC5QAA5x8AgO8fAICF6QAAgAkAAIE1AACFBQAAyyAAgPcfAICFHQAA0yAAgP8fAICFBQAA2yAAgLHBBQCwxQMAwyAAgLLFAwC12QUAtM0DAKMhAICFOQAAuf0DAKchAICbIQCAwQ0AgNMNAIAdDgCABx8AgAsOAIDZDQCAzIgCABEOAIDN4D4AzZABAMwkAQB2DQCAlA0AgEcOAICDDgCAgLEAAM3UPgDN5D4AiQ4AgMy8PgDNuD4AgNEDAIHtAwCC/QMAhmkAAEQOAICFnQMAzTwBAD4OAIDM6AIAzTw/AIjlAADNGAEAjw4AgIhBAABBDgCAfQ4AgM0sAQCbDgCAgNUAAKEOAICG4QAAhukAAE0OAIDNJAEApw4AgM0QAQCI0QAAiCkAAMz4AgBTDgCAzfgCAMwkAQCtDgCAhS0DAMygPgDNbD4AgNUDAIHNAwCCAQMAg/kDAMxkAwDNzAIASg4AgM0kAQDMDAIAzQgCAIERAADMnAMAzLA+AM20PgDMxD4AzcA+AMyAPgDNuD4Asw4AgMyEAgDMmD8AzVA+AMwgPgDNoD4AzQw/AM0wPwDNeD8AzQQ/AIhZAADFDgCAzfgBAMzEAQBQDgCAyw4AgNEOAIDMFAIAzAgBAM3IAQCIBQAA1w4AgN0OAIDMKAIAvw4AgIgNAACG0QAAgB0BAITNAACI9QAAzDwCAIQ1AQDMRAIAhikBAIYOAICIZQEAjA4AgKdEBQBoDgCAi+0AAIjtAACBDQAAiCUAAIZlAADMcAIAzXQCAMwwAgDN2AUAYg4AgJIOAICAOQAAZQ4AgMzgBQCADgCAzCgBAM0UAQCGJQAAiFUAAA4OAICGhDAAyg0AgIDVBwCG/QcAng4AgMwkAgCIPQAApA4AgHEOAICIPQAAqg4AgMxIAgDNeAIAVg4AgLAOAICXwAUAlnAFAJUYBQCAaQAAk1gFAIE5AACIZQAAkPg8AIZZAACeqAUAhEUAAG4OAIDM1AIAmrQFAIBdAACYrAUAp+wEAIgRAADM2AIAzdwCAKO8BAC2DgCAzGACAMgOAIB0DgCAzg4AgK0IBADUDgCAq/QEAMwsAgCIBQAA2g4AgLfoAwC2HAQAtSgEAMwAAgCzKAQAi3kAAIh9AACwdAQAhkEAAL6kAwCEdQAAiB0AAOAOAIC6TAMAzNwDALj8AwCDqAIAiA0AAMIOAICIFQAAh5QCAMw4AgBrDgCAzAQCAIvcAgCPDQAAdw4AgI8ZAADMIAIAeg4AgI3wAgCIdQAAmCADAJksAwCVDgCAmg0AgMxMAgCWcAMAzCQCAIg9AACYDgCAzCwCAIgFAAC5DgCAzCQCAIgNAAC8DgCAh/UAAKjUAwCpxAMA4w4AgNlgAgDYDwCA2w8AgOEPAICUNQAAkzEAANloAgDeDwCA2UwCAJQFAADkDwCAlSEAAJQpAABWEACAehYAgEkXAIDYFgCA2WACAD0XAIC12AMAtPADAJQ1AADZWAIAYBcAgJQFAADZVAIAlA0AADcXAIDgdAEAisgAALwVAACIyAAA4IACAI0XAICBoAAApOwCAKTIAgCoXAAAvA0AAJ8XAIDghAIAvAUAAKMXAICk+AIA4PQCALDMAwCV0AAAYxcAgLPgAwCmyAIAp2ACAJLYAABqFwCAvsEAAHEXAICXwQAAeBcAgH8XAICGFwCAzXg/AMy8PwC+gA0AkRcAgLx4DAC9gA0AuvQMALtUDAC49AwAmBcAgLwXAIC3uAwAwBcAgLWMDACyoAMAs6AMAKcXAICxQAMArnACAK9kAwC4BQMArUgDAK4XAIC1FwCAqEQDAKnYAwDgFwCAp9gDAKRoAgCliAMAtjUDALc9AwCSyAIAtT0DAJldAQCYTQEAm2UBAJppAQCdZQEAnGUBAJ+FAQCemQEAh5wCAL6tAACWpQAAl70AAMw0BQDNjDcAzLg4AM2sOACflQEAth0AAJ2ZAQCc9QEAs7EBAK54AgDnFwCAxBcAgJk9AADLFwCAmxkAAJoJAADSFwCA2RcAgOBIAgCeCQAArFwCAK30AgAAGACA/BcAgAQYAIDuFwCAh2ADAPUXAICvVAIAvhEAAJcFAAAIGACA4KwCAAwYAICG+AMAh+wDAOC0AgAUGACAr0gCAK6QAgDgPAIAvg0AABAYAICXGQAA4NgCAIaEAwCWEQAAvwAMAJ1tAACcYQAAGBgAgLFMAgCzUAIAlQ0AABwYAICGnAMA4MgCALMEAgCCBQAAKBgAgLNQAgCVDQAALBgAgCAYAIAkGACA4LQCAIaMAwCH3AMAvg0AAJVpAACWeQAAMBgAgLToAgC1UAIAlwUAADgYAIDg1AIAtPQCAL4ZAADgoAIANBgAgODUAgCZjAMAt9QCAIoFAAA8GACAQBgAgIoVAAC3NAIAjx0AAEQYAIBIGACAswUAAEwYAICzBQAAYRgAgJwJAACdCQAAUxgAgFoYAICMBQAAaBgAgHMYAIB6GACAgRgAgJ9JAACIGACAjxgAgGwYAICWGACAnRgAgN8YAIDVGACA8BgAgOYYAICkGACAg8kBAIH5AQCyGACAuRgAgMAYAIDHGACAzhgAgKsYAICAtAIApYgDAOEIAgCuHQAA9xgAgLwJAACN9QEA+xgAgOEAAgCSlQEA45QQAJNFAACXiQEAhRQAAId4AQCGAAQAVzoAgFs6AIBfOgCAYzoAgGc6AICdeQAA74xoAJyhAQBrOgCAbzoAgKKZAABzOgCAdzoAgHs6AIB/OgCAp4kAAIM6AICHOgCAqUkBAIs6AICsqQAAjzoAgJM6AICXOgCAsyUBAJs6AICfOgCAozoAgLchAQC2OQEAtTEBAKc6AICrOgCAufkAALkRAQC4GQEArzoAgLM6AIC3OgCAuzoAgICwAQCEiAIAvzoAgIPIAQCEVAMAhFwEAMM6AICEXAUAgN0DAIEtAACCMQAAvjwCAMs6AIDPOgCAh4gDAIacBACzLQMA0zoAgNc6AIC+AAQAvhwFALbRAwC12QMA2zoAgLv5AwC68QMAmljTAYTgBwC/xQMAvtkDAL3dAwC83QMAvgAYAKUFAwCmDQMA3zoAgIQcGADjOgCA5zoAgKPxAwCsAQMArQEDAK4FAwCvGQMArKQbAq3cGgKqLQMAqyUDAL5MGQC+SBoA6zoAgL6AGwC04BoCtdQdArYwHgLvCAIA7zoAgOGgAQC6OBoC4/gCALoAAAC9ZBwCvvQcAr8AEAKRBNMBkOT2AeBEAQCSCD4C8zoAgPc6AID7OgCA/zoAgL6sHAADOwCABzsAgAs7AIAPOwCAEzsAgBc7AIAbOwCAgbBtAICAAQCDHFIAgth3AIUgmgCEkL4AhwjPAIaM5gCJbDcBiOAsAYsYfgGK2BMBjeClAYzwWgGP/OsBjliPAbDVFwCxAWgAso1rALOdawC0SWsAtZVvAB87AIDgcAEAIzsAgCc7AIArOwCALzsAgIAZAACBGQAAggUAADM7AIA7OwCAoaUCAKJJBwCjQQcApEEGAKXVGwCm3RsAp8EaAKgBHACp4R8AqkkfAKsBEACs9RMAra0TAK4BFACv+RcAqDEGAKkxBgCqTQYAq0UGAKxNBgCtmQYAro0GAK+FBgCGgAMAhxgDAD87AIBDOwCARzsAgEs7AIBPOwCAUzsAgLhtBwC5dQcAun0HALt1BwC8bQcAvc0HAL75BwC/+QcAsKkGALGFBgCyeQcAs3kHALRpBwC1aQcAtl0HALdVBwDHOgCAs8EGAFc7AIA3OwCAth0GAFs7AIBfOwCAtcEGALppBgC7RQYAYzsAgGc7AIC+qQcAv6kHALypBwC9qQcAo4UGAGs7AIBvOwCAczsAgHc7AICmWQYApYUGAHs7AICrAQYAqi0GAH87AICDOwCAr+0HAK7tBwCt7QcArO0HAKjBBgCpLQEAqiUBAKs9AQCsJQEArS0BAK4lAQCvlQEAhzsAgIs7AICPOwCAkzsAgJc7AICCvQAAgb0AAIC9AAC4nQEAua0BALqlAQC7bQAAvHUAAL19AAC+dQAAv20AALD1AQCx/QEAssEBALPBAQC0tQEAtb0BALa1AQC3rQEAmzsAgJ87AICjOwCAs6EBAKc7AIC1oQEAtqEBAKs7AICGgAEAh8QBALo9AQC7NQEAvBkBAL0ZAQC+fQEAv3UBAKPtAQCvOwCAszsAgLc7AIC7OwCApu0BAKXtAQC/OwCAq3kBAKpxAQDDOwCAxzsAgK85AQCuMQEArVUBAKxVAQDLOwCAzzsAgNM7AIDXOwCA2zsAgOGsAQDfOwCA42AGAOM7AIDnOwCA6zsAgO9UBgDvOwCA8zsAgL60GgD3OwCA+zsAgP87AICGaBwAh4wDAAM8AIAHPACACzwAgA88AICAOQAAgTkAAIIFAAATPACAGzwAgB88AIAjPACAJzwAgKgdAwCpQQMAqkEDAKtBAwCsQQMArUkDAK5xAwCvcQMAhCAdACs8AIAvPACAMzwAgDc8AIA7PACAPzwAgEM8AIC46QAAufUAALr9AAC78QAAvJEAAL2RAAC+iQAAv4kAALDhAACx4QAAsuEAALPhAAC04QAAte0AALbZAAC32QAA4wwHAOEgBwDhMAEA4wgHAEc8AIBLPACATzwAgFM8AIBXPACAWzwAgF88AIBjPACA75gHAGc8AIBrPACA74gHALOJAgBvPACAczwAgL6AGgB3PACAtokCALWJAgB7PACAu2UBALplAQB/PACAgzwAgL9pAQC+ZQEAvXUBALx1AQC3PQYAtj0GALU9BgC0IQYAszUGALI1BgCxAQYAsAkGAL9ZBgC+UQYAvVkGALxNBgC7bQYAunkGALlxBgC4eQYAgJ0AAIGtAACCpQAAizwAgI88AICTPACAlzwAgJs8AICvcQYArmkGAK1tBgCsbQYAq4EGAKqZBgCpkQYAqJkGABc8AICHPACAnzwAgKPFHQCjPACApcUdAKbFHQCnPACAhgADAIdkAwCqKR4AqykeAKw5HgCtOR4ArikeAK8lHgCzOR4AqzwAgK88AICzPACAtzwAgLb9HgC1/R4AuzwAgLvZHgC60R4AvzwAgMM8AIC/aR8AvmEfAL1pHwC8wR4AqPEeAKnxHgCq8R4Aq/EeAKw1HgCtPR4ArjUeAK8tHgDHPACAyzwAgM88AIDTPACA1zwAgNs8AIDfPACA4zwAgLjlHwC57R8AuuUfALv5HwC86R8AvZEfAL6RHwC/jR8AsFUeALFdHgCyVR4As/0fALTlHwC17R8AtuUfALfdHwCjeR8A5zwAgOs8AIDvPACA8zwAgKa9HwClvR8A9zwAgKuZHwCqkR8AhogAAIdMAQCvKR4AriEeAK0pHgCsgR8AgEkAAIFJAACCWQAAs5keAPs8AIC1iR4AtlEBAP88AIADPQCABz0AgLotAQC7JQEAvD0BAL0lAQC+JQEAvxUBAKhNHgCpVR4Aql0eAKtVHgCsTR4ArZ0BAK6JAQCvgQEAhKwBAAs9AIAPPQCAEz0AgBc9AIAbPQCAHz0AgCM9AIC4ZQEAuW0BALplAQC7fQEAvGUBAL1tAQC+ZQEAv9kAALClAQCxrQEAsqUBALO9AQC0rQEAtZ0BALaVAQC3XQEAo9UdACc9AIArPQCALz0AgDM9AICmHQIApcUdADc9AICraQIAqmECADs9AIA/PQCAr1kCAK5pAgCtaQIArHECAEM9AIBHPQCASz0AgE89AIBTPQCAVz0AgFs9AIBfPQCAgDkAAIE5AACCBQAAYz0AgGs9AIBvPQCAh0ADAIZcBACETAQAcz0AgHc9AICEBAUA4yABAHs9AIDhqAEAfz0AgO+UGgCDPQCAhz0AgIs9AICPPQCAkz0AgJc9AICbPQCAs6EDAJ89AICjPQCApz0AgKs9AIC2fQMAtX0DAK89AIC7WQMAulEDALM9AIC3PQCAv/0AAL79AAC9/QAAvEEDAKhRAgCpWQIAqmkCAKtpAgCstQIArb0CAK61AgCvrQIAhKgHALs9AIC/PQCAwz0AgIKpAADHPQCAgKkAAIGpAAC4aQEAuWkBALoJAQC7CQEAvBkBAL0ZAQC+CQEAvwkBALDVAgCx3QIAstUCALNpAQC0eQEAtXkBALZpAQC3YQEA4bgBAOHUHwDjOB8A4wwbAMs9AIDPPQCA0z0AgNs9AIDfPQCA4z0AgOc9AIDrPQCAvjwJAO89AIDvhBsA74QbAKOhAgDzPQCAhugEAIe8BQD3PQCApn0CAKV9AgD7PQCAq1kCAKpRAgD/PQCAAz4AgK/9AQCu/QEArf0BAKxBAgCzhQYA1z0AgAc+AIALPgCADz4AgLaJBgC1jQYAEz4AgLuRBgC6iQYAFz4AgBs+AIC/9QYAvokGAL2BBgC8iQYAHz4AgCM+AIAnPgCAKz4AgC8+AIAzPgCANz4AgO+EHQA7PgCA4QAEAD8+AIDj/AQAgBEAAIEdAACCBQAAQz4AgKjxBgCp8QYAqg0GAKsFBgCsBQYArQkGAK49BgCvNQYARz4AgEs+AICGiAAAhxADAE8+AIBTPgCAVz4AgFs+AIC4EQYAuRkGALohBgC7IQYAvPUHAL39BwC+9QcAv+kHALBNBgCxVQYAsl0GALNVBgC0TQYAtTEGALYxBgC3MQYAo4UHAF8+AIBjPgCAZz4AgGs+AICmiQcApY0HAG8+AICrkQcAqokHAHM+AIB3PgCAr/UHAK6JBwCtgQcArIkHAHs+AICz4QYAfz4AgIM+AIC25QYAhz4AgIs+AIC18QYAur0GALuNBgCPPgCAkz4AgL59AQC/ZQEAvJUGAL11AQCoHQYAqSUGAKotBgCrJQYArD0GAK0hBgCuXQYAr00GAJc+AICbPgCAnz4AgKM+AICnPgCAgrkDAIGxAwCAuQMAuO0BALmFAQC6jQEAu4UBALydAQC9hQEAvo0BAL+FAQCwPQYAsQ0GALIFBgCz5QEAtP0BALXlAQC25QEAt9UBAKOlBQCrPgCArz4AgLM+AIC7PgCApqEFAKW1BQC/PgCAq8kFAKr5BQCGCAwAhxwDAK8hAgCuOQIArTECAKzRBQDDPgCAs/ECAMc+AIDLPgCAtlUDAM8+AIDTPgCAteECALpxAwC7eQMA1z4AgNs+AIC+MQMAvz0DALxRAwC9UQMAqCUCAKk1AgCqPQIAqzUCAKwtAgCtkQMArpEDAK+RAwDfPgCA4z4AgOc+AIDrPgCArAAAAO8+AIDzPgCA9z4AgLiZAwC5rQMAuqUDALttAwC8dQMAvX0DAL51AwC/bQMAsPEDALH5AwCywQMAs8EDALSxAwC1vQMAtrUDALepAwD7PgCA/z4AgAM/AIAHPwCACz8AgA8/AIATPwCA76gaAL5oDADhlAEAFz8AgOMcBgCADQAAgXEAAIJxAAAbPwCAo/UDAB8/AIAjPwCAhEwCACs/AICmUQIApeUDAC8/AICrfQIAqnUCAIbIDACHLA0ArzkCAK41AgCtVQIArFUCAOFQBgAzPwCA4xQHAITADAA3PwCAOz8AgD8/AIBDPwCARz8AgEs/AIBPPwCAUz8AgFc/AIBbPwCA73gbAL74DwBfPwCAYz8AgGc/AICzjQEAaz8AgLWZAQC2jQEAbz8AgGc9AIBzPwCAuoUBALtNAQC8VQEAvV0BAL5VAQC/SQEAo0EOACc/AIB3PwCAez8AgH8/AICmQQ4ApVUOAIM/AICrgQ4AqkkOAIc/AICLPwCAr4UOAK6ZDgCtkQ4ArJkOAIBtAACBCQAAgh0AAI8/AIDvGAkAkz8AgJc/AICbPwCA4zwNAJ8/AIDhWAwAoz8AgIbQAACHvAMApz8AgKs/AICokQ4AqZkOAKrJDgCrxQ4ArN0OAK3BDgCuwQ4Ar/UOAIToAACvPwCAsz8AgLc/AIC7PwCAvz8AgMM/AIDHPwCAuMEPALnBDwC6wQ8Au8EPALzBDwC9wQ8AvsEPAL/1DwCwjQ4AsUUOALJNDgCzRQ4AtF0OALVBDgC2QQ4At0EOAKhRDgCpWQ4Aqo0OAKudDgCshQ4ArY0OAK6FDgCvvQ4Ayz8AgM8/AIDTPwCA1z8AgNs/AIDfPwCA4z8AgOc/AIC4kQ4AuZkOALqtDgC7RQEAvF0BAL1FAQC+RQEAv3UBALDFDgCxzQ4AssUOALPdDgC0xQ4AtbUOALa9DgC3tQ4AswUOAOs/AIDvPwCA8z8AgPc/AIC2DQ4AtQ0OAPs/AIC7CQ4AugEOAP8/AIADQACAv3EOAL4BDgC9CQ4AvBEOAIJtAACjQQ4AgFUAAIFlAACmSQ4AC0AAgA9AAIClSQ4AqkUOAKtNDgCGSAAAh3gAAK5FDgCvNQ4ArFUOAK1NDgCoXQIAqWECAKplAgCrdQIArG0CAK2xAgCusQIAr7ECAITsBAATQACAF0AAgBtAAIAfQACAI0AAgCdAAIArQACAuHEDALlxAwC6cQMAu3EDALzVAwC93QMAvtUDAL/NAwCw0QIAsdECALLRAgCz0QIAtFEDALVRAwC2UQMAt1EDAC9AAICz6QIAM0AAgL6ABAC2NQIAN0AAgDtAAIC14QIAuhECALsRAgA/QACAQ0AAgL6RAwC/kQMAvAECAL0BAgBHQACAS0AAgKOlAgBPQACApa0CAFNAAIBXQACApnkCAFtAAIBfQACAq10CAKpdAgCtTQIArE0CAK/dAwCu3QMAqNUCAKndAgCqLQEAqyUBAKw9AQCtJQEAri0BAK8lAQBjQACAZ0AAgGtAAIBvQACAc0AAgHtAAIB/QACAg0AAgLiFAQC5iQEAup0BALuVAQC8sQEAvbEBAL55AAC/eQAAsF0BALHlAQCy4QEAs/kBALTpAQC13QEAttUBALe9AQDh8A4Ah0AAgOMUDgCLQACAgb0AAIC9AACPQACAgq0AAIYABACH7AUAk0AAgJdAAICbQACAn0AAgO9gDgCjQACAp0AAgKtAAICFXH0Ar0AAgLNAAIDjZAEAt0AAgOG0AQC7QACA76AOAL9AAIC3PgCAhPgFAMNAAIDHQACAy0AAgLMlBgB3QACAz0AAgNNAAIDXQACAtiUGALU1BgDbQACAu6EGALoZBgDfQACA40AAgL+ZBgC+rQYAva0GALy1BgCCbQAA7zAEAIBVAACBZQAAvlwDAOdAAICG+AAAh2wDAOtAAIDvQACA80AAgPdAAID7QACA40QEAP9AAIDhjAcAo6UGAANBAIAHQQCAC0EAgA9BAICmpQYApbUGABNBAICrIQYAqpkGABdBAIAbQQCArxkGAK4tBgCtLQYArDUGAB9BAICz+QcAI0EAgCdBAIC2SQcAK0EAgC9BAIC1UQcAulEHALtRBwAzQQCAN0EAgL41BwC/OQcAvEUHAL09BwCoNQYAqT0GAKo1BgCriQYArJ0GAK2NBgCusQYAr7EGADtBAIA/QQCAQ0EAgEdBAICADQAAgbEAAIKxAABLQQCAuKEGALmtBgC6vQYAu7UGALytBgC9XQEAvlUBAL9NAQCw0QYAsdEGALLVBgCzrQYAtLUGALW5BgC2qQYAt6UGAKO9BgBPQQCAU0EAgISEAgC+kAEApg0GAKUVBgBbQQCAqxUGAKoVBgCGCAAAh3wBAK99BgCucQYArXkGAKwBBgBfQQCAs60BAGNBAIBnQQCAtqkBAGtBAIBvQQCAta0BALptAQC7dQEAc0EAgHdBAIC+XQEAvzUBALxlAQC9VQEAqGECAKlhAgCqYQIAq2ECAKxhAgCtbQIArp0CAK+VAgB7QQCAf0EAgINBAICHQQCAi0EAgI9BAICTQQCAl0EAgLiVAgC5nQIAuqECALuhAgC8cQMAvXEDAL5xAwC/cQMAsO0CALH1AgCy9QIAs8UCALTdAgC1tQIAtrECALexAgCbQQCAn0EAgKNBAICj5QIAp0EAgKXlAgCm4QIAq0EAgK9BAICzQQCAqiUCAKs9AgCsLQIArR0CAK4VAgCvfQIAt0EAgLtBAIC/QQCAhEB8AIAVAACBHQAAggUAAMNBAIC+7HwAy0EAgIZIfQCHCAMAz0EAgNNBAIDXQQCA20EAgKidAgCpxQIAqsECAKvBAgCsxQIArc0CAK7xAgCv8QIA30EAgONBAIDnQQCA60EAgMkAAADvQQCA80EAgPdBAIC4wQEAucEBALrBAQC73QEAvM0BAL31AQC+/QEAv50BALBBAQCxQQEAskEBALNBAQC0QQEAtUEBALZBAQC3QQEA4TgGAPtBAIDjaAYA/0EAgANCAIAHQgCAC0IAgISUfQC+rHwAD0IAgBNCAIAXQgCAvrh/ABtCAIDvEAEAH0IAgCNCAIAnQgCAK0IAgC9CAIDhkAEAM0IAgONEAAA7QgCAgS0AAIAtAADvgAAAgjkAAD9CAIBDQgCAB0AAgEdCAIDhsH8Ax0EAgOPUfABLQgCAN0IAgE9CAICGuAAAh9QCAFNCAIBXQgCAW0IAgF9CAIBjQgCAZ0IAgO8gfABrQgCAs4l9AG9CAIBzQgCAd0IAgHtCAIC2jX0AtY19AH9CAIC7RX4AukV+AINCAICHQgCAv0V+AL5FfgC9VX4AvFV+AKNJfQCLQgCAj0IAgJNCAICXQgCApk19AKVNfQCbQgCAq4V+AKqFfgCfQgCAo0IAgK+FfgCuhX4ArZV+AKyVfgCCbQAAszF+AIBVAACBZQAAtvF/AITcAwCnQgCAtSF+ALrNfwC70X8AhgAEAIfUAAC+dX8Av3l/ALzBfwC9wX8AqOV/AKn1fwCq/X8Aq/V/AKztfwCtNX4Arj1+AK81fgCrQgCAr0IAgLNCAIC3QgCAu0IAgL9CAIDDQgCAx0IAgLjZfgC54X4AuuF+ALvhfgC85X4Avel+AL6ZfgC/mX4AsE1+ALFRfgCyUX4As1F+ALT1fgC1+X4Atul+ALfpfgCjdX8Ay0IAgM9CAIDTQgCA10IAgKa1fgClZX8A20IAgKuVfgCqiX4A30IAgONCAICvPX4ArjF+AK2FfgCshX4A50IAgLMxfgDrQgCA70IAgLbFAQDzQgCA90IAgLXRAQC6yQEAu8kBAPtCAID/QgCAvs0BAL+xAQC8yQEAvckBAKjdfQCp9X0Aqv19AKvxfQCsHQIArQECAK45AgCvOQIAA0MAgAdDAIALQwCAD0MAgIIFAAATQwCAgBEAAIERAAC4EQIAuRkCALohAgC7IQIAvNUCAL3dAgC+1QIAv80CALBJAgCxSQIAslkCALNZAgC0TQIAtTECALYxAgC3MQIAvgADAKNxfQCEiAIAvoAEAKaFAgAbQwCAH0MAgKWRAgCqiQIAq4kCAIYoBACHDAMAro0CAK/xAgCsiQIArYkCACNDAICEyAMAhcwFALPlAwAnQwCAteUDALbtAwArQwCAL0MAgDNDAIC6bQMAu2UDALx9AwC9ZQMAvmUDAL9VAwA3QwCAO0MAgL8ABACjJQIAP0MAgKUlAgCmLQIAQ0MAgEdDAIBLQwCAqq0CAKulAgCsvQIAraUCAK6lAgCvlQIAT0MAgFNDAIBXQwCAW0MAgF9DAIDjzAMAY0MAgOGsAQBnQwCA7xwDAGtDAIBvQwCAc0MAgHdDAIB7QwCAf0MAgOFwfwBXQQCA4wR+AINDAICLQwCA4ZQBAI9DAIDjWAEAgNkAAIHZAACCJQAA7+R+AJNDAICXQwCA7+B+AJtDAICzAQEAn0MAgIboBwCHLAQAo0MAgLY1AQC1BQEAp0MAgLvxAAC64QAAq0MAgK9DAIC/sQAAvtEAAL3ZAAC84QAAF0MAgIdDAICzQwCAt0MAgKEBBACgEQQAoxkAAKLFBACotQYAqb0GAKrpBgCr/QYArO0GAK3VBgCu3QYArz0HALBFBwCxVQcAslUHALNtBwC0dQcAtRUHALYdBwC3FQcAuC0HALk1BwC6MQcAuw0HALwZBwC9GQcAvgkHAL8JBwCjQQYAu0MAgL9DAIDDQwCAx0MAgKZ1BgClRQYAy0MAgKuxBwCqoQcAj8ltAM9DAICv8QcArpEHAK2ZBwCsoQcAld11AJTBdACXzXAAli1zAJFdaACQVWgAk9l0AJJNaQCd5XgAnB17AJ9tBwCeuXgAmR1/AJhVcACboXwAmvl8AIJhbACDhWkA00MAgNdDAICGEXUAhxF1AISVaQCFjWgAij10AIvFcgDbQwCA30MAgI7dfgCPMX0AjD1xAI2dcQCSGX0Ak716AONDAIDvkAkAltUGAJdRBQCUXXkAlQl5AJpxBQCbvQUA50MAgOtDAIDvQwCA4agFAJx5AQDjuAgAoYUBAPNDAICjqQ0AogEMAKUBCACkOQ0Ap6kJAKa9CQCppRUAqAEUAKsBFACq/RUArbkRAKyxEQCvARwArqEQALH9HACw5R0As+kZALIBGAC1ASQAtH0ZAIQUAAC+FAAAgI0AAIGVAACCbQAA+0MAgIZQDwCHZAAA/0MAgANEAIC61QcAu90HALjBBwC5wQcAvjEEAL8xBAC88QcAvfEHALKtBwCztQcAsK0HALGlBwC2nQcAt/UHALSlBwC1lQcAqmkHAKtpBwCoaQcAqWkHAK5pBwCvaQcArGkHAK1pBwAHRACAC0QAgA9EAIATRACAF0QAgBtEAIAfRACAI0QAgKgRBQCpHQUAqjkFAKs5BQCsLQUArVEFAK5JBQCvQQUAJ0QAgCtEAIAvRACAM0QAgDdEAIA7RACAP0QAgENEAIC4XQIAuWkCALrBAwC7wQMAvPkDAL35AwC+kQMAv7UDALAJBQCxCQUAsuECALPhAgC0dQIAtX0CALZ1AgC3bQIAs7EEAIQAAgC+BA0AR0QAgEtEAIC20QQAtaUEAE9EAIC7zQQAus0EAFNEAIBXRACAv7kDAL6xAwC9NQMAvDUDAFtEAICj9QQAX0QAgGNEAICmlQQAa0QAgG9EAICl4QQAqokEAKuJBACHqA0AhswMAK71AwCv/QMArHEDAK1xAwDhUAYA4TQHAONAAADjWAcAgNEAAIHdAACC1QAAc0QAgHdEAIB7RACAf0QAgINEAICHRACAi0QAgO+cAADvyAcAj0QAgJNEAICzNQIAl0QAgLW1AQCbRACAn0QAgLa1AQC+7AwAo0QAgLuRAQC6mQEAvVEBALyJAQC/UQEAvlkBAKjtDQCp/Q0AqvUNAKttDgCsdQ4ArX0OAK51DgCvbQ4AZ0QAgKdEAICrRACAr0QAgLNEAIC3RACAu0QAgL9EAIC49Q4Auf0OALr1DgC7QQ8AvEEPAL1JDwC+cQ8Av3EPALAVDgCxHQ4AshUOALPNDgC01Q4Atd0OALbVDgC3zQ4Ao30NAMNEAIDHRACAy0QAgM9EAICm/Q4Apf0OANNEAICr2Q4AqtEOAISoAgDXRACArxkOAK4RDgCtGQ4ArMEOAIBNAACBVQAAglUAALNRDwDbRACAtXEPALZxDwDfRACAhuAAAIcEAwC6XQ8Auy0PALw1DwC9OQ8Avi0PAL8lDwCoVQ4AqV0OAKqVDgCrrQ4ArLUOAK29DgCutQ4Ar60OAONEAIDnRACA60QAgO9EAIDzRACA90QAgPtEAID/RACAuGkBALlpAQC6eQEAu3kBALxpAQC9aQEAvt0BAL/VAQCw1Q4AsaUOALKtDgCzoQ4AtKUOALWtDgC2nQ4At1kBAKMdDgADRQCAB0UAgPdDAIALRQCApj0OAKU9DgAPRQCAq2EOAKoRDgATRQCAF0UAgK9pDgCuYQ4ArXUOAKx5DgAbRQCAH0UAgCNFAIAnRQCAK0UAgC9FAIAzRQCAN0UAgIANAACBFQAAgh0AADtFAIA/RQCAQ0UAgIR4AQC+FAAA4xQPAEtFAIDh4A0AhAADAIawBACHFAMAT0UAgFNFAIBXRQCAW0UAgF9FAIBjRQCA78APAGdFAIBrRQCAb0UAgHNFAIB3RQCAe0UAgLNtAwB/RQCAtX0DALZ1AwCDRQCAh0UAgItFAIC6UQMAu1EDALz1AwC9/QMAvukDAL/hAwCPRQCAk0UAgJdFAICbRQCAn0UAgKNFAICnRQCAq0UAgKhxAgCpeQIAqokDAKuJAwCsmQMArZkDAK6JAwCviQMAsPkDALH5AwCyTQMAs0UDALRBAwC1SQMAtnEDALdxAwC4IQMAuSEDALohAwC7IQMAvCEDAL0hAwC+IQMAvyEDAICdAQCBEQAAghEAAIQEBQDvFAAAr0UAgLNFAIC+EAUA48gAALtFAIDh0AEAv0UAgMNFAIDHRQCAy0UAgM9FAICqeQIAq3kCAIboBACHYAUArsECAK/JAgCs3QIArdUCANNFAICjRQIA10UAgNtFAICmXQIA30UAgONFAIClVQIA50UAgOtFAIDvRQCA80UAgPdFAID7RQCA/0UAgO+EDgC+rAQA4dAOAANGAIDjFAEAB0YAgAtGAIAPRgCAE0YAgLPdAQAXRgCAG0YAgB9GAIAjRgCAtv0BALX9AQArRgCAu90BALrdAQCE4AQAL0YAgL+hAQC+vQEAvb0BALy9AQCoBQYAqR0GAKoVBgCrLQYArDUGAK09BgCuNQYArykGALdFAICC9QcAgeUHAIDlBwAnRgCAM0YAgIYcAACHsAMAuCUGALnFBgC6zQYAu8UGALzdBgC9xQYAvs0GAL/FBgCwWQYAsVkGALIpBgCzKQYAtDkGALUlBgC2JQYAtx0GAKOdBgA3RgCAO0YAgD9GAIBDRgCApr0GAKW9BgBHRgCAq50GAKqdBgBLRgCAT0YAgK/hBgCu/QYArf0GAKz9BgBTRgCAs/UHAFdGAIBbRgCAtu0HAF9GAIBjRgCAteUHALqNBwC7kQcAZ0YAgGtGAIC+dQcAv30HALyBBwC9fQcAqCUGAKkpBgCqOQYAqzkGAKwpBgCtKQYArnkGAK91BgBvRgCAc0YAgHdGAIB7RgCAf0YAgINGAICHRgCAi0YAgLjVBgC53QYAuuEGALv9BgC85QYAve0GAL7lBgC/mQYAsA0GALERBgCyEQYAs+0GALT1BgC1/QYAtvUGALftBgCjsQYAgi0AAIEVAACAsQAAR0UAgKapBgCloQYAj0YAgKvVBgCqyQYAk0YAgL5oAQCvOQYArjEGAK05BgCsxQYAm0YAgLPxAQCGaAAAh3wBALZdAQCfRgCAo0YAgLVVAQC6SQEAu0kBAKdGAICrRgCAvj0BAL8hAQC8OQEAvTUBAK9GAICzRgCAhAQDAL6AHAC3RgCA4RwGALtGAIDjAAYAvwguAL9GAIDDRgCA78gHAMdGAIDLRgCAz0YAgNNGAIDXRgCA20YAgKN9AgDfRgCApdkCAONGAIDnRgCAptECAOtGAIDvRgCAq8UCAKrFAgCtuQIArLUCAK+tAgCusQIAqW0FAKhZBQCrDQIAqrkCAK0dAgCsHQIArwUCAK4NAgC+aB0A80YAgPdGAID7RgCAgB0AAIEJAACCmQEA/0YAgLnhAwC4KQIAu+EDALrpAwC94QMAvPkDAL/hAwC+6QMAsU0CALBNAgCzIQIAsi0CALUlAgC0OQIAtxECALYlAgCowQIAqdECAKrRAgCr5QIArP0CAK0VAQCuHQEArw0BAANHAIALRwCAD0cAgBNHAIAXRwCAG0cAgB9HAIAjRwCAuAUBALkJAQC6HQEAuxUBALwxAQC9MQEAvv0BAL/1AQCweQEAsUEBALJBAQCzXQEAtEUBALVNAQC2RQEAtz0BAIagHQCHxB0AJ0cAgO/YAAArRwCAL0cAgDNHAIDvxAYAhGwcAOH0BgA3RwCA47AGADtHAIDhlAEAP0cAgONEBgCzGQIAQ0cAgEdHAIBLRwCAhewsALbVAQC1NQIAT0cAgLvFAQC6/QEAU0cAgFdHAIC/yQEAvsEBAL3JAQC81QEAo9kdAAdHAIBbRwCAX0cAgGNHAICmFR4ApfUdAGdHAICrBR4Aqj0eAGtHAIBvRwCArwkeAK4BHgCtCR4ArBUeAIBpAACBaQAAggUAAHNHAIB3RwCAe0cAgIcQAwCGfAMAf0cAgINHAICHRwCAi0cAgI9HAICTRwCAl0cAgJtHAICopR8Aqa0fAKqlHwCrvR8ArKUfAK2tHwCupR8ArxUfAJ9HAICjRwCAp0cAgKtHAICvRwCAs0cAgLdHAIC7RwCAuA0fALkZHwC6IR8AuyEfALzZAAC92QAAvskAAL/BAACwcR8AsXEfALJxHwCzRR8AtEEfALVNHwC2PR8AtzUfALMtHgC/RwCAw0cAgMdHAIDLRwCAti0eALUtHgDPRwCAu7UeALq1HgDTRwCA10cAgL+JHgC+hR4AvZEeALylHgCCKQAAo2keAIAdAACBFQAApmkeANtHAIDfRwCApWkeAKrxHgCr8R4A40cAgITgAQCuwR4Ar80eAKzhHgCt1R4AqNUBAKnlAQCq7QEAq+UBAKz9AQCt5QEAru0BAK/lAQC+oAEAl0YAgOdHAIDrRwCAhhAAAId0AQDvRwCA80cAgLh9AQC5wQAAusEAALvBAAC8wQAAvckAAL7xAAC/8QAAsJ0BALFFAQCyTQEAs0UBALRdAQC1RQEAtk0BALdFAQD3RwCA+0cAgP9HAIADSACAB0gAgO80AgDv7B4AC0gAgOHwHQDj4AIA4zAeAOGEAQAPSACAE0gAgBdIAIAbSACAsyUCAJQAAAAfSACAI0gAgCdIAIC2JQIAtTUCACtIAIC7wQIAuhkCAC9IAIAzSACAv8ECAL7ZAgC90QIAvNkCADdIAIA7SACAP0gAgKPpAgBDSACApfkCAKbpAgBHSACAS0gAgE9IAICq1QIAqw0CAKwVAgCtHQIArhUCAK8NAgCAYQAAgWEAAIIFAABTSACAW0gAgIQABAC+FAQAX0gAgIbABACHUAMAY0gAgGdIAIBrSACAb0gAgHNIAIB3SACAqK0CAKm9AgCqtQIAqw0BAKwVAQCtHQEArhUBAK8NAQCE7AQAe0gAgH9IAICDSACAh0gAgItIAICPSACAk0gAgLgdAQC5LQEAuiUBALvNAQC81QEAvd0BAL7JAQC/wQEAsH0BALFVAQCyXQEAs1UBALRNAQC1PQEAtjUBALctAQDhGB4Al0gAgOM4HgCbSACAn0gAgKNIAICnSACAq0gAgK9IAICzSACAvmAEALdIAICBdQAAgHUAAO/gHwCCbQAAu0gAgL9IAICG6AQAh3wFAMNIAIDhkAEAy0gAgOOgAADPSACA00gAgNdIAIDvtAAA20gAgN9IAIDjSACA50gAgLUFBgBXSACAx0gAgLYFBgDrSACA70gAgLOlBQDzSACAvRkGALwRBgC/YQYAvhEGAPdIAID7SACAuwkGALohBgCj/QUA/0gAgANJAIAHSQCAC0kAgKZdBgClXQYAD0kAgKtRBgCqeQYAE0kAgBdJAICvOQYArkkGAK1BBgCsSQYAqFEGAKlZBgCqYQYAq2EGAKxhBgCtYQYArmEGAK9hBgAbSQCAH0kAgCNJAIAnSQCAgA0AAIGxAQCCsQEAK0kAgLhNBwC5VQcAul0HALtVBwC8TQcAvXUHAL59BwC/cQcAsMUHALHNBwCyxQcAs90HALTFBwC1zQcAtsUHALd5BwCz6QcAL0kAgDNJAICEwAEAvtgBALbhBwC16QcAN0kAgLsJBgC6AQYAhogAAIesAQC/CQYAvgEGAL0JBgC8EQYAO0kAgKOtBwA/SQCAQ0kAgKalBwBHSQCAS0kAgKWtBwCqRQYAq00GAE9JAIBTSQCArkUGAK9NBgCsVQYArU0GAKhZBgCpZQYAqm0GAKtlBgCsYQYArWEGAK5hBgCvYQYAhKwBAFdJAIBbSQCAX0kAgGNJAIBnSQCAa0kAgG9JAIC4kQEAuZkBALqhAQC7oQEAvHEBAL1xAQC+cQEAv3EBALDxAQCx8QEAsvUBALPdAQC0xQEAtbEBALaxAQC3sQEAs+UFAHNJAIB3SQCAe0kAgH9JAIC24QUAtekFAINJAIC7NQIAujUCAIdJAICLSQCAv3UCAL4BAgC9CQIAvCECAI9JAICjoQUAk0kAgJdJAICmpQUAm0kAgJ9JAIClrQUAqnECAKtxAgCjSQCAvigDAK5FAgCvMQIArGUCAK1NAgCA1QAAgd0AAILhAACrSQCA4yABAK9JAIDhqAEAs0kAgO80AgC3SQCAhggMAIdoAwCsAAAAu0kAgL9JAIDDSQCAs40DAMdJAIDLSQCAhIAMAM9JAIC2vQMAtYEDANNJAIC7TQMAuk0DANdJAIDbSQCAv00DAL5NAwC9TQMAvE0DAKhBAgCpTQIAqkUCAKtZAgCsSQIArX0CAK51AgCvuQIAvmgNAN9JAIDjSQCA50kAgIRsDADrSQCA70kAgPNJAIC4TQEAuVUBALpVAQC7ZQEAvH0BAL0VAQC+EQEAvxEBALDJAgCxyQIAstkCALPZAgC0yQIAtckCALZ9AQC3dQEA4XgHAOOYAADjuAYA4VwGAPdJAID7SQCA/0kAgANKAIAHSgCAC0oAgA9KAIATSgCA7AAAAO9cAADv6AYAG0oAgIFpAACAYQAAo4UCAIJhAACliQIAH0oAgCNKAICmtQIAhkAMAIfEDACrRQIAqkUCAK1FAgCsRQIAr0UCAK5FAgCojQ4AqZEOAKqVDgCrqQ4ArKUOAK2tDgCupQ4Ar9kOABdKAIAnSgCAK0oAgC9KAIAzSgCAN0oAgDtKAIA/SgCAuHUPALl9DwC6dQ8Au90PALzFDwC9zQ8AvsUPAL/9DwCwqQ4AsbUOALK1DgCzhQ4AtJ0OALVRDwC2UQ8At1EPALMdDgBDSgCAR0oAgEtKAIBPSgCAti0OALUtDgBTSgCAu3EOALptDgBXSgCAW0oAgL+VDwC+WQ4AvVEOALxhDgBfSgCAo1kOAGNKAIBnSgCApmkOAGtKAIBvSgCApWkOAKopDgCrNQ4Ac0oAgHdKAICuHQ4Ar9EPAKwlDgCtFQ4AqL0OAKnRDgCq0Q4AqykBAKw5AQCtOQEArikBAK8pAQCADQAAgRUAAIIdAAB7SgCAf0oAgINKAIC+dAIAh0oAgLjtAQC5hQEAuoEBALuBAQC8hQEAvY0BAL6xAQC/sQEAsFkBALFZAQCy7QEAs+UBALT9AQC15QEAtuUBALfVAQCLSgCAtqkBALWhAQCPSgCAs0kOAJNKAICGOAAAh9wBAL8xAQC+KQEAvSEBALwpAQC7jQEAuo0BAKdJAICXSgCAoxkOAJtKAICfSgCAo0oAgKdKAICm+QEApfEBAKtKAICr3QEAqt0BAK9KAICzSgCAr2EBAK55AQCtcQEArHkBALdKAIDv3A8Au0oAgL9KAIDDSgCAx0oAgMtKAIDPSgCA00oAgNdKAIDbSgCA30oAgONKAIDj6A4A50oAgOGMDgCAEQAAgREAAIIRAACEQAIA60oAgO9KAIDzSgCAvhADAIbABACHRAMA+0oAgP9KAIADSwCAB0sAgAtLAIAPSwCA7yQCABNLAIAXSwCAG0sAgB9LAIAjSwCAJ0sAgCtLAICE7AQAL0sAgDNLAIA3SwCA4+wCADtLAIDhOAEAP0sAgLNVAwBDSwCAR0sAgEtLAIBPSwCAth0DALUdAwBTSwCAuwkDALo5AwBXSwCAW0sAgL/9AAC+/QAAvfkAALwRAwCogQIAqYkCAKqdAgCrsQIArNUCAK3dAgCu1QIAr80CAIDNAQCBCQAAghkAAF9LAIBjSwCAa0sAgL5wBQBvSwCAuFkBALlZAQC6aQEAu2kBALx5AQC9eQEAvmkBAL9lAQCwvQIAsY0CALKFAgCzbQEAtHkBALV5AQC2aQEAt2kBAIYgBACHCAUAc0sAgHdLAIB7SwCAf0sAgINLAIDvXAAAhOwEAOFcDgCHSwCA44wOAItLAICPSwCAk0sAgJdLAICjVQIAm0sAgJ9LAICjSwCAp0sAgKYdAgClHQIAq0sAgKsJAgCqOQIAr0sAgLNLAICv/QEArv0BAK35AQCsEQIAqGkGAKlpBgCqeQYAq3kGAKxpBgCtaQYArp0GAK+VBgBnSwCAt0sAgLtLAIC/SwCAw0sAgMdLAIDLSwCAz0sAgLj1BgC5+QYAuo0GALuFBgC8nQYAvYUGAL6FBgC/tQYAsO0GALH1BgCy/QYAs/UGALTtBgC10QYAttEGALfRBgCz8QYAghUAAIG1AACAtQAA00sAgLbpBgC14QYAvtQDALsxBgC6KQYA10sAgNtLAIC/FQYAvikGAL0hBgC8KQYA30sAgKO1BgCGyAAAh8gAAKatBgDjSwCA50sAgKWlBgCqbQYAq3UGAOtLAIDvSwCArm0GAK9RBgCsbQYArWUGAKg1BgCpOQYAqoEGAKuBBgCsgQYArYEGAK6BBgCvtQYA80sAgPdLAID7SwCA/0sAgANMAIAHTACAC0wAgA9MAIC4nQYAua0GALqlBgC7aQEAvHkBAL15AQC+aQEAv2kBALDRBgCx0QYAstEGALPRBgC0tQYAtb0GALa1BgC3rQYAswkGABNMAIAXTACAG0wAgB9MAIC2AQYAtQkGACNMAIC7FQYAuhUGACdMAIArTACAv3kGAL5xBgC9BQYAvAUGAC9MAICjTQYAM0wAgPdKAICmRQYAN0wAgDtMAIClTQYAqlEGAKtRBgA/TACAQ0wAgK41BgCvPQYArEEGAK1BBgCB6QMAgN0DAISIAwCC4QMAhrA8AIeIAgC+VAMAS0wAgE9MAIBTTACAV0wAgFtMAIBfTACAY0wAgGdMAIBrTACA4/AGAG9MAIDhMAYAhAA8AHNMAIB3TACAe0wAgH9MAICDTACAhTQ9AIdMAICLTACA77AHAI9MAICTTACAl0wAgJtMAICfTACAo0wAgL7EPACnTACAgp0BAIGdAQCAnQEAqA0CAKllAgCqfQIAq3UCAKxZAgCtWQIArpkDAK+ZAwCw6QMAsekDALL5AwCz+QMAtOkDALXpAwC2XQMAt1UDALhtAwC5dQMAunUDALtFAwC8XQMAvTUDAL4xAwC/KQMAq0wAgK9MAICzTACAu0wAgOFgAwDv9AMA40QCAL9MAIDDTACA4zwDAO/0NwDh/AEAx0wAgMtMAIDPTACA00wAgIZkPwCHaD0AhTQhALOZAwDXTACAtb0DALa1AwDbTACA30wAgONMAIC6QQIAu0ECALxBAgC9QQIAvkECAL9BAgDnTACA60wAgO9MAIDzTACA90wAgPtMAID/TACA7/gBAIRoPADhPAYAA00AgOMcBgAHTQCAC00AgA9NAIATTQCAoxUDABdNAIAbTQCAH00AgCNNAICmOQMApTEDACtNAICrzQIAqs0CAL5kPgAvTQCAr80CAK7NAgCtzQIArM0CAKgdPgCpJT4Aqi0+AKslPgCsPT4ArSU+AK4tPgCvJT4At0wAgIL1PwCB5T8AgOU/ACdNAIAzTQCAhgAEAIecAwC4LT4AuTE+ALoxPgC7MT4AvNE+AL3RPgC+0T4Av80+ALBdPgCxIT4Asjk+ALM5PgC0KT4AtSk+ALYZPgC3FT4As6U+ADdNAIA7TQCAP00AgENNAIC2pT4AtbU+AEdNAIC75T4Aupk+AEtNAIBPTQCAv+0+AL7tPgC97T4AvO0+AFNNAICj4T4AV00AgFtNAICm4T4AX00AgGNNAICl8T4Aqt0+AKuhPgBnTQCAa00AgK6pPgCvqT4ArKk+AK2pPgCPBSUAsyU+AG9NAIBzTQCAtik+AHdNAIB7TQCAtSk+ALp9PgC7RT4Af00AgINNAIC+tT4Av70+ALxdPgC9vT4An304AJ5lOQCd8TgAnFE0AJtZNQCaUTUAmfEwAJgNMQCXZTEAlsEwAJVZLQCUTS0Ak+EsAJLZKQCRWSkAkPEoALSlGQC13RgAh00AgIQIAACwkRUAsQEVALIBGACzvRkAgA0AAIGtAwCCpQMAi00AgKNhAACiHT0AoZk9AKBxPACkxQUApUEEAKYBCACn4QkAR0wAgKH1AQCi6QEAo90FAKwBEACtxREArtkRAK85EACoZQgAqQEMAKrZDQCrCQ0AijEuAIuhMwCPTQCAk00AgI65MwCPETYAjB0yAI1NMgCCJSYAg6krAL5kAwCEYAQAhqEvAIcVLgCEGSoAhZEqAJphPgCb7T4AhsgEAIfcAwCbTQCA4Vw+AJyJAwDjAD4Akmk2AJN5NwCfTQCA7xg+AJZNOwCXuT8AlME7AJVdOgCpnT0AqIk9AKu5PQCqrT0Arak9AKyhPQCvyT0ArqE9AL7oBACjTQCAp00AgKtNAICvTQCAs00AgLdNAIC7TQCAuVk9ALhRPQC7eT0AumU9AL1pPQC8YT0Avx09AL5hPQCxgT0AsLk9ALNpPQCyiT0AtXk9ALRxPQC3aT0AtnE9AKMhPAC/TQCAw00AgMdNAIDLTQCApi08AKUtPADPTQCAq0E8AKp5PADTTQCA100AgK+5PACusTwArbk8AKxZPADbTQCA300AgLN9AwDjTQCAtdkDAOdNAIDrTQCAttEDAO9NAIDzTQCAu8UDALrFAwC9uQMAvLUDAL+tAwC+sQMA900AgPtNAID/TQCA71wDAIAVAACBHQAAgjEAAO+MPgCE7AQA4fw+AANOAIDjHD4AC04AgOGUAQAPTgCA4yAAAKP1AwATTgCAh+gEAIZsBAAXTgCAplkDAKVRAwAbTgCAq00DAKpNAwAfTgCAI04AgK8lAwCuOQMArTEDAKw9AwCXTQCAB04AgCdOAIArTgCAL04AgDNOAIA3TgCAO04AgKhxBgCpTQYAqo0GAKuFBgCsnQYArYUGAK6NBgCvhQYAsP0GALFBBwCyQQcAs0EHALRBBwC1SQcAtnEHALdxBwC4IQcAuSEHALolBwC7OQcAvCkHAL0VBwC+HQcAv/0HALMlBgA/TgCAQ04AgEdOAIBLTgCAtiUGALU1BgBPTgCAu6UHALoZBgBTTgCAV04AgL+tBwC+pQcAvbUHALy1BwBbTgCAo2EGAF9OAIBjTgCApmEGAGdOAIBrTgCApXEGAKpdBgCr4QcAb04AgHNOAICu4QcAr+kHAKzxBwCt8QcAqLEGAKm9BgCqzQYAq90GAKzNBgCt/QYArvUGAK8VAQCA+QEAgc0BAILFAQC+ZAIAhpAAAIcAAQB7TgCAf04AgLjRAQC52QEAuuEBALvhAQC8kQEAvZ0BAL6VAQC/iQEAsG0BALF1AQCyfQEAs3UBALRtAQC18QEAtvEBALfxAQCzRQYAd04AgINOAICHTgCAi04AgLZ9BgC1RQYAj04AgLuxAQC6qQEAk04AgJdOAIC/NQEAvqkBAL2hAQC8qQEAm04AgKMBBgCfTgCAo04AgKY5BgCnTgCAq04AgKUBBgCq7QEAq/UBAK9OAICzTgCAru0BAK9xAQCs7QEAreUBAOEoAQC3TgCA41ACALtOAIC/TgCAw04AgMdOAIDLTgCAz04AgNNOAIDXTgCA204AgIFxAACAGQAA75wCAIJ5AADfTgCA404AgITIAgCzxQMA604AgLXFAwC2xQMAvhADAIbADACHRAwAuqkDALulAwC8vQMAvaEDAL6hAwC/lQMArhEGAK8ZBgCsAQYArQEGAKqlBgCrEQYAqEU5AKlxOQDvTgCA804AgPdOAID7TgCA/04AgANPAIAHTwCAC08AgL7tBwC/TQcAvNEHAL3lBwC63QcAu8EHALg1BgC51QcAtjkGALcNBgC0JQYAtTkGALIxBgCzPQYAsFEGALFRBgCoOQIAqTkCAKqBAgCrgQIArIECAK2JAgCusQIAr7ECAIRsDQAPTwCAvmANABNPAIAXTwCAG08AgB9PAIAjTwCAuE0BALlVAQC6XQEAu1UBALxNAQC9dQEAvn0BAL91AQCwoQIAsa0CALKlAgCzuQIAtKkCALWdAgC2lQIAt3kBAOFUBgDh1AcA4zgGAOOwBwAnTwCAK08AgC9PAIAzTwCAhOQMADdPAIA7TwCAP08AgENPAIBHTwCA72wAAO/kBwCjSQIAS08AgE9PAIBTTwCAW08AgKZJAgClSQIAX08AgKspAgCqJQIAhkgMAIfcDACvGQIAri0CAK0tAgCsMQIAqFEOAKmlDgCqrQ4Aq6UOAKy9DgCtpQ4Arq0OAK+lDgCA5Q8Age0PAILlDwBXTwCAY08AgGdPAIBrTwCAb08AgLjVDwC53Q8AutUPALvpDwC8+Q8AvfkPAL7pDwC/6Q8AsN0OALFBDwCyRQ8As10PALRFDwC1TQ8AtkUPALftDwCzJQ4Ac08AgHdPAIB7TwCAf08AgLYlDgC1NQ4Ag08AgLuFDwC6GQ4Ah08AgItPAIC/iQ8AvoEPAL2JDwC8kQ8Aj08AgKNhDgCTTwCAl08AgKZhDgCbTwCAn08AgKVxDgCqXQ4Aq8EPAKNPAICnTwCArsUPAK/NDwCs1Q8Arc0PAKjRDgCp2Q4AqjkBAKs5AQCsKQEArSkBAK6dAQCvlQEAq08AgK9PAICzTwCAt08AgIANAACBtQAAgr0AALtPAIC4lQEAuZ0BALqhAQC7oQEAvHEAAL1xAAC+cQAAv3EAALDtAQCx9QEAsvUBALPFAQC03QEAtbUBALaxAQC3sQEAv08AgMNPAICzuQEAvsACALWpAQDHTwCAy08AgLahAQCGgAEAh8QBALs5AQC6IQEAvRkBALwpAQC/eQEAvhEBAKPxAQDPTwCA504AgNNPAIDXTwCApukBAKXhAQDbTwCAq3EBAKppAQDfTwCA408AgK8xAQCuWQEArVEBAKxhAQDnTwCA608AgO9PAIDzTwCA4agBAPdPAIDjQAIA+08AgL8oFQD/TwCA73QCAANQAIAHUACAC1AAgA9QAIATUACAF1AAgON0DwCEiAMA4TQOABtQAIAfUACAI1AAgCdQAICADQAAgRUAAIIRAAArUACAL1AAgO+kDwAzUACAO1AAgKgZAwCpQQMAqkUDAKtdAwCsTQMArX0DAK51AwCvnQAAhaQVAL58AwCGCAQAhxwDAD9QAIBDUACAR1AAgEtQAIC49QAAuf0AALr1AAC7jQAAvIEAAL2BAAC+gQAAv4EAALDlAACx7QAAsuUAALP5AAC07QAAtdEAALbVAAC3zQAAT1AAgFNQAIBXUACAs8ECAFtQAIC1yQIAtvECAF9QAIBjUACAZ1AAgLotAQC7JQEAvD0BAL0hAQC+JQEAvxkBAKapAgCESAIAa1AAgKWRAgBvUACAo5kCAHNQAIB3UACArn0BAK9BAQCsZQEArXkBAKp1AQCrfQEAe1AAgH9QAICDUACAh1AAgItQAICPUACA7+QAAJNQAICXUACAm1AAgOMQDgCfUACA4VgOAKNQAICALQAAgREAAIIVAAC+sAUAs3UBAKtQAICHFAUAhmwEAK9QAIC21QAAtWUBALNQAIC7/QAAuvUAALdQAIC7UACAv6EAAL69AAC93QAAvN0AAKh9BgCptQYAqr0GAKu1BgCsrQYArRUHAK4dBwCvFQcAp1AAgL9QAIDDUACAx1AAgMtQAIDPUACA01AAgNdQAIC4OQcAuTkHALrJBwC7yQcAvNkHAL3ZBwC+zQcAv8UHALBxBwCxeQcAskkHALNJBwC0OQcAtSUHALYhBwC3IQcAozUGANtQAIDfUACA41AAgOdQAICmlQcApSUGAOtQAICrvQcAqrUHAO9QAIDzUACAr+EHAK79BwCtnQcArJ0HAPdQAID7UACA/1AAgANRAIAHUQCAgj0AAIE9AACAPQAAC1EAgA9RAIATUQCAhKADAL6kAwAXUQCAhvgAAIfgAACoxQYAqdUGAKrVBgCr5QYArP0GAK0xAQCuMQEArzEBABtRAIAfUQCAI1EAgCdRAIArUQCAL1EAgDNRAIA3UQCAuN0BALntAQC65QEAu40BALyVAQC9nQEAvpUBAL+NAQCwUQEAsVEBALJRAQCzUQEAtPUBALX9AQC29QEAt+0BALNdBgA7UQCAP1EAgENRAIBHUQCAtrEBALV1BgBLUQCAu5UBALqVAQBPUQCAU1EAgL85AQC+MQEAvYUBALyFAQClLQYAV1EAgFtRAICm6QEAX1EAgGNRAICjBQYAZ1EAgK3dAQCs3QEAr2EBAK5pAQBrUQCAN1AAgKvNAQCqzQEAb1EAgHNRAICExAMAvwD0AHdRAICCPQAAgT0AAIA9AAB7UQCAf1EAgINRAIC+YAMAi1EAgI9RAICTUQCAl1EAgIbgHACHAAMA7wwHAJtRAICfUQCAo1EAgKdRAICrUQCAr1EAgLNRAIC3UQCAu1EAgOHABgC/UQCA4ywHAMNRAIDHUQCAy1EAgM9RAIDTUQCA11EAgNtRAIDfUQCA41EAgKiBAwCpgQMAqoEDAKuBAwCsgQMArYEDAK6BAwCvgQMAsEUDALFNAwCyRQMAs10DALRNAwC1fQMAtnUDALcZAwC4KQMAuTUDALo9AwC7MQMAvAEDAL31AAC+/QAAv+0AALMpAgDnUQCA61EAgO9RAIDzUQCAtiECALUpAgCEUB0Au6kCALqhAgD7UQCA/1EAgL+ZAgC+qQIAvakCALyxAgCBTQAAgE0AAO+cAwCCXQAAhvAcAId4HQC+EB0AA1IAgAdSAIALUgCAD1IAgBNSAIDhkAEAF1IAgONgAwAbUgCAH1IAgCNSAIAnUgCAK1IAgC9SAIAzUgCAN1IAgO+UAQCE7BwA4XAGADtSAIDjUAEAP1IAgENSAIBHUgCAS1IAgKPpAgBPUgCAU1IAgFdSAIBbUgCApuECAKXpAgBfUgCAq2kCAKphAgBjUgCAvqgcAK9ZAgCuaQIArWkCAKxxAgCoMR4AqTEeAKoxHgCrMR4ArF0eAK1FHgCuTR4Ar0UeAPdRAICCzR8AgfUfAID9HwBnUgCAa1IAgIYcAACH+AMAuMUeALnNHgC6xR4Au90eALzFHgC9zR4AvsUeAL9ZHwCwPR4AsQUeALINHgCzBR4AtB0eALUBHgC2BR4At/0eALO5HgBvUgCAc1IAgHdSAIB7UgCAtsUeALXVHgB/UgCAu8EeALr5HgCDUgCAh1IAgL/FHgC+2R4AvdEeALzZHgCLUgCAo/0eAI9SAICTUgCApoEeAJdSAICbUgCApZEeAKq9HgCrhR4An1IAgKNSAICunR4Ar4EeAKydHgCtlR4AqCkeAKkpHgCqVR4Aq20eAKx1HgCtfR4ArnUeAK9pHgCnUgCAq1IAgK9SAICzUgCAt1IAgLtSAIC/UgCAw1IAgLjpHgC59R4Auv0eALv1HgC87R4AvZEeAL6RHgC/kR4AsB0eALHlHgCy7R4As+UeALT9HgC15R4Atu0eALflHgCz3R4Ax1IAgMtSAIDPUgCA01IAgLb9HgC1/R4AhFgBALshHgC62R4AvigAANtSAIC/IR4AvjkeAL0xHgC8OR4AgU0AAIBNAACjlR4Agl0AAKW1HgDXUgCA31IAgKa1HgCHUQCA41IAgKtpHgCqkR4ArXkeAKxxHgCvaR4ArnEeAIYABACHRAMAs4ECAOdSAIC1gQIA61IAgO9SAIC2gQIAiAAAAPNSAIC74QIAuu0CAL3lAgC8+QIAv9ECAL7lAgD3UgCA+1IAgIREAwC+jAMA4UgCAP9SAIDjAAIA7/wfAANTAIDhPB4A79wCAONgHwAHUwCAC1MAgA9TAIATUwCAqQUCAKixAgCrBQIAqgUCAK0NAgCsBQIArzUCAK41AgCEbAUAF1MAgBtTAIAfUwCAI1MAgCdTAIArUwCAL1MAgLnpAwC44QMAu/kDALrhAwC96QMAvOEDAL9dAwC+4QMAsSkCALAlAgCzPQIAsiECALUZAgC0LQIAt9kDALYRAgAzUwCAN1MAgDtTAICjhQMAP1MAgKWFAwCmhQMAQ1MAgEtTAIBPUwCAqukDAKvlAwCs/QMAreEDAK7hAwCv1QMAgEkAAIFVAACCVQAAo6kCAL6YBAClQQEApkEBAFNTAICG4AUAh+AFAKotAQCrOQEArBEBAK0FAQCuDQEArwUBAFdTAIBbUwCAX1MAgO/cAABjUwCAZ1MAgGtTAIDviB4AhCwHAOHsHgBvUwCA4xweAHNTAIDhlAEAd1MAgOMwAACzJQIAhWDmAHtTAIB/UwCAg1MAgLbNAQC1zQEAh1MAgLu1AQC6oQEAi1MAgI9TAIC/iQEAvoEBAL2JAQC8nQEAR1MAgJNTAICXUwCAm1MAgJ9TAICjUwCAp1MAgKtTAICoAQcAqQEHAKp1BwCrrQcArLUHAK29BwCuqQcAr6kHALDZBwCx7QcAsvkHALP1BwC0mQcAtZkHALaJBwC3gQcAuIkHALmJBwC6bQAAu2UAALx9AAC9ZQAAvm0AAL9lAACBCQAAgJkAAK9TAICCHQAAs1MAgLdTAIC7UwCAv1MAgKgNBQCpfQUAqk0FAKuhBgCspQYAra0GAK6dBgCv/QYAsIUGALGRBgCyqQYAs70GALSlBgC1rQYAtqUGALd5BgC4SQYAuUkGALpZBgC7WQYAvEkGAL1JBgC++QcAv/kHALNdBgDDUwCAhigCAIcsAQDHUwCAtp0GALWdBgDLUwCAu4kGALq9BgDPUwCA01MAgL/9BgC+/QYAvYEGALyNBgDXUwCAoxkGANtTAIDfUwCAptkGAONTAIDnUwCApdkGAKr5BgCrzQYA61MAgO9TAICuuQYAr7kGAKzJBgCtxQYAqBkBAKkZAQCqjQAAq50AAKyNAACtvQAArrUAAK/dAADzUwCA91MAgPtTAID/UwCAA1QAgAdUAIALVACAD1QAgLhpAAC5aQAAunkAALt5AAC8aQAAvWkAAL7dAwC/1QMAsKkAALGpAACyvQAAs7UAALSZAAC1mQAAtlkAALdZAAC+LAIAE1QAgBdUAIAbVACAH1QAgCNUAIArVACAL1QAgIAtAACBNQAAgj0AADNUAICGkAwAh+gCADdUAIA7VACAs0UDAD9UAIBDVACAR1QAgEtUAIC2fQMAtUUDAE9UAIC7LQMAui0DAFNUAIBXVACAvx0DAL4dAwC9IQMAvCkDAKvNAwCqzQMAW1QAgF9UAICv/QMArv0DAK3BAwCsyQMAo6UDAGNUAIBnVACAa1QAgG9UAICmnQMApaUDAHNUAIB3VACAe1QAgH9UAICDVACAh1QAgII9AACBPQAAgD0AAItUAICPVACAk1QAgIRgAwCG0AwAhzADAJtUAICfVACAvkQCAKNUAICnVACAq1QAgOEAAACvVACA46gGALNUAICE7AwAt1QAgO/QAwC7VACAv1QAgMNUAIDHVACAy1QAgLNtAQDPVACA01QAgNdUAIDbVACAthEBALVlAQDfVACAuz0BALo1AQDjVACA51QAgL/9AQC+/QEAvRUBALwVAQDrVACA4fwGAO9UAIDjPAcA81QAgPdUAID7VACA/1QAgANVAIC+bAwAC1UAgA9VAIATVQCAF1UAgBtVAIDvFAYAgV0AAIBdAACj5QEAgm0AAKXtAQAfVQCAI1UAgKaZAQCHqAwAhuQMAKu1AQCqvQEArZ0BAKydAQCvdQEArnUBAKgZDgCpGQ4AqiUOAKs1DgCsLQ4ArVEOAK5RDgCvUQ4Al1QAgAdVAIAnVQCAK1UAgC9VAIAzVQCAN1UAgDtVAIC47Q4AufUOALr1DgC7jQ4AvJUOAL2dDgC+lQ4Av40OALAxDgCxOQ4AsgEOALMBDgC0+Q4AtfkOALbdDgC31Q4AqHkOAKl5DgCqjQ8Aq4UPAKydDwCtgQ8AroUPAK+5DwA/VQCAQ1UAgEdVAIBLVQCAT1UAgFNVAIBXVQCAW1UAgLiRDwC5mQ8AuqEPALuhDwC8UQ8AvV0PAL5JDwC/SQ8AsM0PALHVDwCy3Q8As9UPALTNDwC1sQ8AtrEPALexDwCzBQ4AX1UAgGNVAIBnVQCAa1UAgLYBDgC1FQ4Ab1UAgLsRDgC6CQ4Ac1UAgISgAQC/dQ4AvgkOAL0BDgC8CQ4AgmkAAKNBDgCAWQAAgVEAAKZFDgC+WAEAd1UAgKVRDgCqTQ4Aq1UOAIbIAACHrAEArk0OAK8xDgCsTQ4ArUUOAHtVAIB/VQCAg1UAgIdVAICLVQCAj1UAgCdUAICTVQCAqAkOAKkJDgCqGQ4AqxkOAKwJDgCtYQ4ArmEOAK+VAQCw7QEAsfUBALL9AQCz9QEAtO0BALV1AQC2fQEAt3UBALhNAQC5VQEAul0BALtVAQC8TQEAvfEAAL7xAAC/8QAAl1UAgJtVAICfVQCAo1UAgKdVAIDj6A4Aq1UAgOE0DgC+AAQA79wPAK9VAICzVQCAt1UAgLtVAIC/VQCAw1UAgLPxDQDHVQCAy1UAgM9VAIDTVQCAtoENALXhDQDXVQCAu1ECALpJAgDbVQCA31UAgL/RAgC+SQIAvUECALxJAgCjMQ0A41UAgISIAwDrVQCA71UAgKZBDQClIQ0A81UAgKuRAgCqiQIA91UAgPtVAICvEQIArokCAK2BAgCsiQIAgKkAAIGpAACCTQAA/1UAgOFkEgDjTAIA4wgLAOGsAQADVgCA7zwCAO8YFgAHVgCAhlAGAIdIAwALVgCAD1YAgKiBAgCpgQIAqoECAKuBAgCsgQIArYECAK6FAgCvHQEAE1YAgBdWAIAbVgCAH1YAgCNWAIAnVgCAK1YAgIS4BQC4dQEAuX0BALp1AQC7CQEAvBkBAL0ZAQC+CQEAvwEBALBlAQCxbQEAsmUBALN9AQC0aQEAtV0BALZVAQC3TQEAL1YAgDNWAIA3VgCAO1YAgD9WAIBDVgCA7zQAAO/ADgDhXA4A4UwPAOOUAADjnA4AR1YAgIJlAACBfQAAgH0AAEtWAIBPVgCAvsQHALNFAgBTVgCAtUUCALZNAgBbVgCAhkAGAIeQBAC67QEAu+UBALz9AQC95QEAvuEBAL/VAQCflQgAngUIAJ3dDQCcPQwAmzEMAJr1DQCZ7RAAmD0QAJfVEQCWsRUAlQUUAJTlFQCTtRkAkjEYAJE5GACQDRwAj2EcAOdVAICz1QYAX1YAgLX9BgBXVgCAY1YAgLaRBgBnVgCAa1YAgLuVBgC6lQYAvVUHALxVBwC/VQcAvlUHAG9WAIBzVgCAqo0GAKuFBgCsnQYArYUGAK6BBgCvtQYAhKgAAHdWAIB7VgCAoyUFAH9WAIClJQUApi0FAINWAICHVgCAi1YAgI9WAICTVgCAl1YAgJtWAICfVgCAo1YAgKdWAICrVgCAr1YAgLNWAICjqQUAotEEAKHZBACgZQUAgiEdAIM1HQC3VgCAu1YAgIaVGACH3RQAhBkZAIUZGQCKDRUAi7EUAL9WAIDDVgCAjsURAI/VDACMzRAAjR0RAJJhDQCTdQ0AvkwAAMtWAICWxQkAl80EAJSNDACVXQkAmkEFAJtBBQCGyP8Ah0wAAIFZAACAeQAAnCEEAIJRAAChxQEAz1YAgKMB/ACi2QEApRX9AKS1/QCnufkApgH4AKkJ+AColfkAqwX1AKqt9QCtsfEArAHwAK8d8ACurfEAseHtALAB7ACzAegAsv3sALVd6QC09ekA01YAgNdWAIDbVgCA31YAgONWAIDnVgCA61YAgO9WAIDzVgCA91YAgKiNBACplQQAqpUEAKulBACsvQQArdkEAK75BACv8QQAhGz8APtWAID/VgCAA1cAgAdXAIALVwCAD1cAgBNXAIC4eQUAucUFALrNBQC7xQUAvN0FAL3FBQC+zQUAv+0FALCZBACxmQQAskkFALNJBQC0WQUAtVkFALZJBQC3SQUAox0EAL7M/AAXVwCAG1cAgB9XAICmWQQApTUEACNXAICrXQQAql0EACdXAIArVwCAr50FAK6dBQCtnQUArJ0FAC9XAICznQIAM1cAgDtXAIC2UQIAP1cAgENXAIC1uQIAukkCALtVAgCGSP0Ah8D8AL41AgC/PQIAvEUCAL09AgCo3QQAqUkDAKpRAwCrbQMArHUDAK2VAwCunQMAr7kDAICNAQCB5QEAguEBAEdXAIBLVwCAT1cAgFNXAIBXVwCAuJUDALmdAwC6lQMAu60DALy1AwC9vQMAvrUDAL9VAgCwyQMAsdUDALLVAwCzrQMAtLUDALW9AwC2tQMAt60DAFtXAIBfVwCAo9EDAGNXAICl9QMAZ1cAgGtXAICmHQMAb1cAgHNXAICrGQMAqgUDAK1xAwCsCQMAr3EDAK55AwDhKAcAd1cAgOPkBgB7VwCA4SgGAH9XAIDjaAEAg1cAgIdXAICLVwCA71gAAI9XAICTVwCAl1cAgO/IBgCbVwCAqE39AKmB/QCq0f0Aq9H9AKzx/QCt8f0ArvH9AK/x/QA3VwCAghEAAIEZAACA0f8An1cAgKNXAICEdAMAvnQDALh1/gC5ff4AunX+ALvF/gC83f4AvcX+AL7F/gC/9f4AsJH9ALGR/QCykf0As5H9ALRV/gC1Xf4AtlX+ALdN/gCzWf0Ap1cAgIasAACHRAMAq1cAgLZx/QC1ef0Ar1cAgLtV/QC6Vf0As1cAgLdXAIC/mf4AvpH+AL1F/QC8Rf0Au1cAgKMd/QC/VwCAw1cAgKY1/QDHVwCAy1cAgKU9/QCqEf0AqxH9AM9XAIDTVwCArtX+AK/d/gCsAf0ArQH9AKjN/wCp0f8AqtH/AKsh/gCsIf4ArSH+AK4h/gCvIf4A11cAgNtXAIDfVwCA41cAgOdXAIDrVwCA71cAgPNXAIC4jf4AuZH+ALqV/gC7rf4AvLX+AL25/gC+qf4Av6n+ALDh/gCx4f4AsuX+ALP5/gC06f4AtdX+ALbd/gC3uf4As1n/APdXAIDHVgCA+1cAgP9XAIC2of4Atan+AANYAIC7Jf4AuiX+AAdYAIALWACAvxH+AL4t/gC9Lf4AvDH+AIIZAACjHf8AgGUAAIEZAACm5f4AD1gAgBNYAICl7f4AqmH+AKth/gCEZAEAviAAAK5p/gCvVf4ArHX+AK1p/gAbWACA4zT+AB9YAIDhfP0AhrAEAIcIAwAjWACAJ1gAgCtYAIAvWACAhCQDAIQkBAAzWACA70j+ADdYAIA7WACAs+kCAD9YAIC+RAQAvkAFAENYAIC2nQIAtZkCAEdYAIC7iQIAur0CAEtYAIBPWACAv1kDAL5RAwC9WQMAvJECAKkdAgCoFQIAqyUCAKolAgCtWQIArFUCAK9NAgCuUQIAvmQGAFNYAIBXWACAW1gAgF9YAIBjWACAZ1gAgGtYAIC5+QMAuPEDALtNAwC68QMAvUEDALxZAwC/cQMAvkEDALEJAgCwPQIAs8kDALIBAgC12QMAtNEDALfJAwC20QMA4ZABAG9YAIDj8AAAc1gAgHdYAICCPQAAgT0AAIA9AAB7WACAf1gAgINYAICLWACAj1gAgJNYAIDvLAAAl1gAgKPpAwCbWACAhugEAIdgBQCfWACApp0DAKWZAwCjWACAq4kDAKq9AwCnWACAq1gAgK9ZAgCuUQIArVkCAKyRAwCvWACAs1gAgLdYAIC7WACAv1gAgMNYAIDHWACA71gBAISgBADhVP8Ay1gAgOOEAQDPWACA01gAgNdYAIDbWACAs9kBAN9YAICFzBkA41gAgOdYAIC28QEAtfkBAOtYAIC7pQEAutkBAO9YAIDzWACAv50BAL6dAQC9pQEAvK0BAKgBBgCpDQYAqhEGAKsRBgCsMQYArTEGAK4pBgCvJQYAh1gAgILJBwCBwQcAgPEHAPdYAID7WACAhhwAAIf8AwC47QYAufUGALr9BgC79QYAvO0GAL1RBwC+VQcAv00HALBdBgCxIQYAsjkGALMxBgC0GQYAtRkGALbdBgC31QYAo5kGAP9YAIADWQCAB1kAgAtZAICmsQYApbkGAA9ZAICr5QYAqpkGABNZAIAXWQCAr90GAK7dBgCt5QYArO0GABtZAICz8QcAH1kAgCNZAIC2gQcAJ1kAgCtZAIC1mQcAuo0HALtlBwAvWQCAM1kAgL59BwC/ZQcAvH0HAL11BwCoLQYAqTUGAKo9BgCrMQYArFUGAK1FBgCuRQYAr3UGADdZAIA7WQCAP1kAgENZAIBHWQCAS1kAgE9ZAIBTWQCAuOkGALn1BgC6/QYAu/UGALztBgC9kQYAvpUGAL+NBgCwDQYAseUGALLtBgCz5QYAtP0GALXlBgC27QYAt+UGAKO1BgBXWQCAW1kAgF9ZAIBjWQCApsUGAKXdBgAXWACAqyEGAKrJBgBnWQCAa1kAgK8hBgCuOQYArTEGAKw5BgCASQAAgUkAAIJZAACzRQEAb1kAgLVFAQC2RQEAc1kAgIZAAACHZAAAuikBALslAQC8PQEAvSEBAL4hAQC/FQEAd1kAgHtZAICEBAMAvgAMAOMoBgDv4AIA4RAGAH9ZAIDvkAYA4zwCAINZAIDh1AEAh1kAgItZAICPWQCAk1kAgJdZAICbWQCAo8ECAJ9ZAIClwQIAo1kAgKdZAICmwQIAq1kAgK9ZAICroQIAqq0CAK2lAgCsuQIAr5ECAK6lAgCpBQIAqLECAKsFAgCqBQIArQ0CAKwFAgCvNQIArjUCAISoDACzWQCAt1kAgLtZAIC/WQCAw1kAgMdZAIDLWQCAuekDALjhAwC7+QMAuuEDAL3pAwC84QMAv10DAL7hAwCxKQIAsCUCALM9AgCyIQIAtRkCALQtAgC32QMAthECAKitAgCp1QIAqtUCAKsNAQCsFQEArQkBAK4xAQCvLQEAz1kAgNNZAIDbWQCA31kAgONZAIDnWQCA61kAgO9ZAIC4IQEAuSEBALrtAQC75QEAvP0BAL3lAQC+7QEAv+UBALBVAQCxXQEAslUBALMtAQC0NQEAtTkBALYtAQC3JQEAgD0BAIGlAACCrQAA79QHAPNZAID3WQCA+1kAgO8oBwC+LAwA4fQGAP9ZAIDjkAcAA1oAgOGUAQAHWgCA4wwGALMdAgALWgCAh0QNAIZMDQAPWgCAtskBALXdAQATWgCAu9kBALrRAQAXWgCAG1oAgL+9AQC+sQEAvbkBALzBAQDXWQCAH1oAgCNaAIAnWgCAK1oAgC9aAIAzWgCAN1oAgKgJDwCpCQ8AqhkPAKsZDwCsCQ8ArQkPAK6pDwCvqQ8AsNkPALHtDwCy+Q8As/UPALSVDwC1hQ8AtoUPALe1DwC4jQ8AuWEAALphAAC7YQAAvGEAAL1hAAC+YQAAv2EAAKNdDQCCLQAAgRUAAIAdAAA7WgCApokOAKWdDgA/WgCAq5kOAKqRDgBDWgCAR1oAgK/9DgCu8Q4ArfkOAKyBDgBLWgCAs/UPAIboAwCHvAMAtu0PAE9aAIBTWgCAteUPALp5DwC7TQ8AV1oAgFtaAIC+NQ8AvyUPALxJDwC9RQ8AozEOAF9aAIBjWgCAZ1oAgGtaAICmKQ4ApSEOAG9aAICriQ4Aqr0OAHNaAIB3WgCAr+EOAK7xDgCtgQ4ArI0OAHtaAIB/WgCAg1oAgIdaAICLWgCAj1oAgJNaAICXWgCAm1oAgJ9aAICjWgCAp1oAgIANAACB1QAAgt0AAKtaAICoQQEAqVEBAKpRAQCrZQEArH0BAK2RAACukQAAr5EAAK9aAICzWgCAhGQBAL5kAQCGkAEAh4QAALtaAIC/WgCAuJEAALmRAAC6kQAAu5EAALyxAAC9sQAAvrEAAL+xAACw8QAAsfkAALLBAACzwQAAtLEAALWxAAC2sQAAt7EAALPZAgDDWgCAvnADAL5EBADHWgCAthEDALX1AgDLWgCAuz0DALo1AwDPWgCA01oAgL91AwC+dQMAvRUDALwVAwDXWgCAo50CANtaAIDfWgCAplUDAONaAIDnWgCApbECAKpxAwCreQMA61oAgO9aAICuMQMArzEDAKxRAwCtUQMAqDkDAKk5AwCqjQAAq50AAKyNAACtvQAArrUAAK/dAADzWgCA91oAgPtaAID/WgCAA1sAgAdbAIALWwCAD1sAgLhpAAC5aQAAunkAALt5AAC8aQAAvWkAAL7ZAQC/2QEAsKkAALGpAACyvQAAs7UAALSZAAC1mQAAtlkAALdZAAATWwCAF1sAgBtbAIAfWwCA70QAACNbAICGmAUAh+QCAOOYAACEqAIA4fgBACtbAICAOQAAgTkAAIItAAAvWwCAs0UBADNbAIA3WwCAO1sAgD9bAIC2fQEAtUUBAENbAIC7LQEAui0BAEdbAIBLWwCAvx0BAL4dAQC9IQEAvCkBAE9bAIDhUA4AU1sAgOM8DwBXWwCAW1sAgF9bAIBjWwCAZ1sAgGtbAIDjAAAAb1sAgHNbAIB3WwCAhPQFAO/kDgCuqQEAr6kBAKydAQCtlQEAqpkBAKuZAQB7WwCAf1sAgKbJAQCDWwCAh1sAgKXxAQCC/QcAo/EBAID9BwCB9QcAJ1sAgItbAICPWwCAk1sAgJdbAICbWwCAhrgDAIeQAwCoDQcAqRkHAKptBwCrZQcArH0HAK1lBwCuZQcAr1UHALAtBwCxxQcAssEHALPdBwC0xQcAtc0HALbFBwC3/QcAuMUHALnJBwC62QcAu9kHALypBwC9qQcAvp0HAL+VBwCzxQcAn1sAgKNbAICnWwCAq1sAgLbFBwC11QcAr1sAgLshBwC6yQcAs1sAgLdbAIC/KQcAviEHAL0pBwC8NQcAu1sAgKOBBwC/WwCAw1sAgKaBBwDHWwCAy1sAgKWRBwCqjQcAq2UHAM9bAIDTWwCArmUHAK9tBwCscQcArW0HAKgVAQCpgQEAqoEBAKuBAQCsgQEArYkBAK6xAQCvsQEA11sAgNtbAIDfWwCA41sAgOdbAIDrWwCA71sAgPNbAIC4ZQAAuW0AALplAAC7fQAAvGUAAL1tAAC+ZQAAv90AALChAQCxrQEAsqUBALO5AQC0qQEAtZ0BALaVAQC3XQAA91sAgIIdAACBHQAAgB0AAPtbAID/WwCAA1wAgL5YAQCErAIAB1wAgIcIAQCGjAEAC1wAgLdaAIAPXACAE1wAgLNJAQAXXACAG1wAgB9cAIAjXACAtkkBALVJAQAnXACAuykBALolAQArXACAL1wAgL8ZAQC+LQEAvS0BALwxAQC+2AMAM1wAgO/4BgA3XACAO1wAgD9cAIDv4AIAQ1wAgOGUAQBHXACA43QCAEtcAIDhmAUAT1wAgOMMBwBTXACAV1wAgFtcAICjwQIAhIwDAKXBAgBfXACAY1wAgKbBAgBnXACAa1wAgKuhAgCqrQIAraUCAKy5AgCvkQIArqUCAKgxAwCpPQMAqjUDAKtJAwCsWQMArVkDAK5JAwCvQQMAgMUAAIEJAACCGQAAb1wAgHNcAIB7XACAh2wDAIYcHAC47QAAufEAALr1AAC7jQAAvJUAAL2BAAC+gQAAv70AALAJAwCxCQMAsu0AALPhAAC04QAAteEAALblAAC32QAAf1wAgINcAICHXACAs7ECAItcAIC13QIAttUCAI9cAICTXACAl1wAgLrBAgC7wQIAvDUBAL05AQC+KQEAvykBAKaNAgCbXACAn1wAgKWFAgCjXACAo+kCAKdcAICrXACArnEBAK9xAQCsbQEArWEBAKqZAgCrmQIAr1wAgLNcAIC3XACA4YQGALtcAIDjJAYAv1wAgOGUAQDDXACA4ywAAL7oHQDHXACAy1wAgO/IAACE/B0AvvAcAM9cAIDvSAcA01wAgNdcAIDbXACA31wAgIEdAACAHQAA41wAgIIFAACGQBwAh8QcAOtcAIDvXACA81wAgPdcAID7XACA/1wAgKi1HgCpBR8Aqg0fAKsFHwCsAR8ArQkfAK45HwCvOR8A51wAgANdAIAHXQCAC10AgA9dAIATXQCAF10AgBtdAIC4yR8AudUfALrRHwC76R8AvPkfAL3tHwC+mR8Av5kfALAlHwCxLR8AsjkfALM1HwC0LR8AtQ0fALYFHwC3/R8As4UfAB9dAIAjXQCAJ10AgCtdAIC2iR8AtYkfAC9dAIC76R8AuuEfADNdAIA3XQCAv8kfAL7pHwC94R8AvO0fADtdAICjwR8AP10AgENdAICmzR8AR10AgEtdAIClzR8AqqUfAKutHwBPXQCAU10AgK6tHwCvjR8ArKkfAK2lHwCo6R4AqekeAKr5HgCr+R4ArOkeAK3pHgCuPQEArzUBAID5AQCBzQEAgsUBAIRgAgBXXQCAW10AgIdoAQCGnAAAuNEBALnZAQC64QEAu+EBALyRAQC9nQEAvpUBAL+JAQCwTQEAsVUBALJdAQCzVQEAtE0BALXxAQC28QEAt/EBALNxHgBfXQCAY10AgGddAIBrXQCAtmkeALVhHgBvXQCAu5EBALqJAQBzXQCAd10AgL81AQC+iQEAvYEBALyJAQB7XQCAd1wAgKM5HgB/XQCApSkeAINdAICHXQCApiEeAItdAICPXQCAq9kBAKrBAQCtyQEArMEBAK99AQCuwQEAk10AgJddAICbXQCAn10AgKNdAICnXQCAq10AgK9dAICzXQCAt10AgLtdAIC/XQCAw10AgMtdAIDPXQCAvnADAOHkHgCESAIA4+gfAIQABACAeQAAgXkAAIJpAADTXQCAhsAEAIdEAwDXXQCA210AgN9dAIDjXQCA7yAfAOddAIDrXQCA710AgPNdAIDvSAIA910AgPtdAID/XQCAA14AgL7oBAAHXgCAC14AgA9eAIATXgCA4ZABABdeAIDj6AIAs0kDABteAIAfXgCAI14AgCdeAIC2SQMAtUkDACteAIC7LQMAuiUDAC9eAIAzXgCAvxUDAL4VAwC9IQMAvCkDAKg1AgCpgQIAqoECAKuBAgCsgQIArYkCAK6xAgCvsQIAgP0BAIHNAQCCxQEAO14AgIaQBACHBAUAP14AgIRwBAC4SQEAuUkBALpZAQC7WQEAvEkBAL1JAQC+eQEAv3kBALChAgCxqQIAsr0CALO1AgC0kQIAtZECALZ5AQC3eQEAQ14AgEdeAIBLXgCAT14AgFNeAIBXXgCAW14AgO/QHgC+6AQA4VweAF9eAIDjkAAAY14AgGdeAIBrXgCAb14AgKNJAgBzXgCAd14AgHteAIB/XgCApkkCAKVJAgCDXgCAqy0CAKolAgCHXgCAi14AgK8VAgCuFQIArSECAKwpAgCoNQYAqT0GAKpVBgCrZQYArH0GAK1lBgCubQYAr2EGADdeAICPXgCAk14AgJdeAICADQAAgbEAAIKxAACbXgCAuOkGALnpBgC6+QYAu/UGALyVBgC9nQYAvpUGAL+NBgCw4QYAseEGALLhBgCz/QYAtOUGALXtBgC25QYAt9kGALPdBgCfXgCAo14AgKdeAICrXgCAtuUGALX1BgCvXgCAuyUGALolBgCGmAAAh6wAAL8pBgC+IQYAvSkGALw1BgCzXgCAo5kGALdeAIC7XgCApqEGAL9eAIDDXgCApbEGAKphBgCrYQYAx14AgMteAICuZQYAr20GAKxxBgCtbQYAqC0GAKk9BgCqiQYAq4kGAKyZBgCtmQYArokGAK+JBgDPXgCA014AgNdeAIDbXgCA314AgONeAIDnXgCA614AgLiNBgC5lQYAupUGALulBgC8vQYAvXEBAL5xAQC/cQEAsPkGALHNBgCy2QYAs9kGALTJBgC1yQYAtr0GALe1BgCzAQYA714AgPNeAID3XgCA+14AgLYZBgC1EQYA/14AgLsJBgC6PQYAA18AgAdfAIC/DQYAvg0GAL0NBgC8DQYAC18AgKNFBgDHXQCAD18AgKZdBgATXwCAhFgAAKVVBgCqeQYAq00GAL5oAQAXXwCArkkGAK9JBgCsSQYArUkGAIDBAwCByQMAgt0DAKPNAgAbXwCApdkCAKbNAgAfXwCAhoANAIeUAwCqxQIAqw0DAKwVAwCtHQMArhUDAK8NAwDhnBcA4xgGAOMUAwDhNAYA7xgCACNfAIAnXwCAK18AgOPQAgAvXwCA4VACADNfAIA3XwCA7ywGAO/kJQA7XwCArE0CAK1RAgCuUQIAr2UCAKgBAgCpCQIAqlkCAKtVAgCE7A0AP18AgENfAIBHXwCAvvgNAEtfAIBPXwCAU18AgLxRAwC9WQMAvmEDAL9hAwC47QMAuVEDALpRAwC7UQMAtM0DALXVAwC23QMAt9UDALAdAgCx1QMAst0DALPVAwDjyAAAV18AgOG4AQBbXwCAhFQPAF9fAIBjXwCAZ18AgKHpAgCgFQYAo6UDAKINAwDvIAAAa18AgG9fAIBzXwCAd18AgHtfAICFNCYAs40DAH9fAIC1mQMAto0DAINfAICGwA8Ah5QNALqFAwC7TQIAvFUCAL1dAgC+VQIAv00CAItfAICPXwCAk18AgJdfAICbXwCAn18AgI/d6wDvxAYAvuAPAOGMBgCjXwCA44AGAID1AACB5QAAguUAAKdfAICZbR8AmMUfAJvJGwCaeRoAnXUaAJzFGwCf+QcAnhkGAJFpFgCQsesAk20XAJLNFwCV0RMAlGkSAJdREgCWzRMAg1XkAIJB5ACHXwCAq18AgIeNHQCGkRgAhTkYAISVGQCLERwAigUcAK9fAICzXwCAj4UVAI6ZEACNORAAjJUdAJNRFACSRRQAt18AgLtfAICXYQkAlnUIAJWdCQCU+RUAm0EMAJqtDQC/XwCAw18AgMdfAIDLXwCAz18AgJzxDAChbQ0A018AgKMBBACihQAApZkEAKSRBACnGTgApsUFAKkJOACoKTgAq4k8AKoBPACtATAArB08AK8pMACunTAAseE0ALABNACzASgAsv00ALXZKAC00SgA118AgNtfAIDfXwCA418AgOdfAIDrXwCAgB0AAIEJAACC2QEA718AgKgRDwCpGQ8Aql0PAKtVDwCsTQ8ArXEPAK51DwCvbQ8A818AgPtfAICGiAAAhxABAP9fAIADYACAB2AAgAtgAIC4TQ4AuVEOALpRDgC7UQ4AvGUOAL1tDgC+ZQ4Avx0OALAdDwCxwQ8AssEPALPBDwC0xQ8Atc0PALbFDwC3eQ4As9UPAA9gAIATYACAF2AAgBtgAIC28Q8AtcUPAB9gAIC7BQ8AutkPACNgAIAnYACAvwkPAL4BDwC9FQ8AvBUPACtgAICjkQ8AL2AAgDNgAICmtQ8AN2AAgDtgAIClgQ8Aqp0PAKtBDwA/YACAQ2AAgK5FDwCvTQ8ArFEPAK1RDwCogQ0AqYENAKqBDQCrgQ0ArIENAK2BDQCusQ0Ar6ENAEdgAIBLYACAT2AAgFNgAIBXYACAgrkAAIG9AACAvQAAuDUCALk9AgC6zQIAu5UCALyNAgC9tQIAvr0CAL+1AgCwbQIAsU0CALJFAgCzJQIAtD0CALUdAgC2FQIAtw0CAFtgAIBfYACAswENAGNgAIC1AQ0Aa2AAgISUAwC2CQ0AviwEAG9gAIC7gQIAuqECAL35AgC8mQIAv9ECAL7xAgBzYACAd2AAgHtgAICjRQ0Af2AAgKVFDQCmTQ0Ag2AAgIbgBACHpAQAquUCAKvFAgCs3QIArb0CAK61AgCvlQIAqCUCAKk1AgCqPQIAqzUCAKwtAgCtkQIArpECAK+RAgCHYACAi2AAgI9gAICTYACAzAAAAJdgAICbYACAn2AAgLiZAgC5rQIAuqUCALttAQC8dQEAvX0BAL51AQC/bQEAsPECALH5AgCywQIAs8ECALSxAgC1vQIAtrUCALepAgCjYACA44QOAKdgAIDh9A4Aq2AAgK9gAICzYACAt2AAgIQgBQC7YACAv2AAgMNgAIDHYACA7+wOAMtgAIDPYACAs/UCANNgAICG6AQAh4wEAL5cBAC2UQIAteUCANtgAIC7fQIAunUCAN9gAIDjYACAvzkCAL41AgC9VQIAvFUCAKM1BQBnYACA12AAgOdgAIDrYACAppEFAKUlBQDvYACAq70FAKq1BQDzYACA92AAgK/5BQCu9QUArZUFAKyVBQCA+QcAgfkHAIKNBwCzjQYA+2AAgLWdBgC2iQYA/2AAgANhAIAHYQCAuk0HALtFBwC8XQcAvUEHAL5BBwC/QQcAC2EAgA9hAID3XwCAE2EAgBdhAIAbYQCAH2EAgCNhAICoNQYAqQEGAKppBgCraQYArHkGAK1lBgCuZQYAr50HALDlBwCx7QcAsuUHALP5BwC06QcAtekHALZZBwC3VQcAuHEHALlxBwC6cQcAu3EHALxVBwC9XQcAvlUHAL9NBwCjwQcAJ2EAgCthAIAvYQCAM2EAgKbFBwCl0QcAN2EAgKsJBgCqAQYAO2EAgD9hAICvDQYArg0GAK0NBgCsEQYAgGkAAIFpAACCBQAAQ2EAgL6YAQCEmAEAR2EAgEthAICGADwAh8QBAE9hAIBTYQCAV2EAgFthAIBfYQCAY2EAgKhdBgCpbQYAqmUGAKuBAQCsgQEArYkBAK6xAQCvsQEAZ2EAgGthAIBvYQCAc2EAgHdhAIB7YQCAf2EAgINhAIC4VQEAuV0BALpVAQC7yQAAvNkAAL3ZAAC+yQAAv8EAALCxAQCxuQEAsokBALOJAQC0cQEAtXEBALZ1AQC3bQEAs+0FAIdhAICLYQCAj2EAgJNhAIC2CQIAtQkCAJdhAIC7fQIAunUCAJthAICfYQCAv7UCAL61AgC9XQIAvF0CAL5gAgCjqQUAo2EAgKdhAICmTQIAq2EAgK9hAIClTQIAqjECAKs5AgCzYQCAhOADAK7xAgCv8QIArBkCAK0ZAgC+iDwAu2EAgKotAwCrJQMArD0DAK0lAwCuLQMAryUDAID1AACB/QAAgsEAAKPBAwC/YQCApcEDAKbBAwDDYQCAhmA8AIdUAwDHYQCAy2EAgM9hAIDjqAIA02EAgOGkAQDXYQCA71wCANthAIDfYQCA42EAgOdhAIDrYQCA72EAgPNhAIDjjAcA92EAgOE8BAD7YQCA/2EAgANiAIAHYgCAhCACAAtiAIAPYgCAE2IAgBdiAIDvbAcAG2IAgB9iAICzLQIAhEQ9ACNiAIArYgCAL2IAgLYtAgC1LQIAM2IAgLvJAgC6wQIAN2IAgDtiAIC/yQIAvsECAL3JAgC80QIA4XgHAOPAAADjOAYA4VwGAICpAACBqQAAgtEAAD9iAIBDYgCAR2IAgL6kPABLYgCAT2IAgO8cAADvkAYAU2IAgIZgPACHBD0AV2IAgLNxAQBbYgCAtRkBALYJAQBfYgCAY2IAgGdiAIC6AQEAuwEBALwBAQC9AQEAvgEBAL8BAQCohT4AqbU+AKq1PgCrxT4ArN0+AK3FPgCuwT4Ar/0+AGtiAIBvYgCAc2IAgHdiAIB7YgCAf2IAgINiAICHYgCAuFE/ALlRPwC6UT8Au1E/ALx1PwC9fT8AvnU/AL9tPwCwiT4AsYk+ALKZPgCzmT4AtIk+ALWJPgC2eT8At3U/ALdhAICjOT4Ai2IAgCdiAICmQT4Aj2IAgJNiAIClUT4Aqkk+AKtJPgCXYgCAm2IAgK5JPgCvST4ArEk+AK1JPgCASQAAgVEAAIJRAACzkT8An2IAgLW5PwC2RT8Ao2IAgIZAAACHBAMAukU/ALtdPwC8TT8AvT0/AL4pPwC/IT8AqE0+AKlVPgCqVT4Aq2U+AKx9PgCtiT4Arrk+AK+5PgCnYgCAq2IAgK9iAICzYgCAt2IAgLtiAIC/YgCAw2IAgLhhAQC5YQEAumEBALthAQC8YQEAvWEBAL5hAQC/YQEAsM0+ALHVPgCy1T4As6U+ALShPgC1qT4Atpk+ALeZPgCj3T4Ax2IAgMtiAIDPYgCA02IAgKYJPgCl9T4A12IAgKsRPgCqCT4A22IAgN9iAICvbT4ArmU+AK1xPgCsAT4A42IAgOdiAIDrYgCA72IAgPNiAID3YgCA+2IAgP9iAICAOQAAgTkAAIIFAAADYwCAvrgBAIS4AQALYwCAD2MAgKitAgCp1QIAqtUCAKstAwCsNQMArT0DAK41AwCvLQMAE2MAgBdjAIAbYwCAH2MAgCNjAIAnYwCAK2MAgC9jAIC46QMAuekDALqJAwC7iQMAvJkDAL2ZAwC+iQMAv4kDALBVAwCxXQMAslUDALPpAwC0+QMAtfkDALbpAwC34QMAs10CADNjAICGKAQAh8wDADdjAIC2vQMAtb0DADtjAIC7mQMAupEDAD9jAIBDYwCAvz0DAL49AwC9PQMAvIEDAIUAFACjGQIAR2MAgEtjAICm+QMAT2MAgFNjAICl+QMAqtUDAKvdAwBXYwCAW2MAgK55AwCveQMArMUDAK15AwDjVD4A4dw/AOHQPgDjPD4AX2MAgO8cAABjYwCAZ2MAgGtjAIDjwAAAb2MAgOHUAQDvYD4Ac2MAgHtjAIDvRD8AgGEAAIFtAACCfQAAhAAFAIbwBACHnAUAvhAFAH9jAICDYwCAh2MAgItjAICPYwCAk2MAgJdjAICbYwCAn2MAgLiJPQC5iT0Aupk9ALuRPQC8uT0Avbk9AL7RPQC/0T0AsAU+ALENPgCyBT4Asx0+ALQFPgC1DT4AtgU+ALe5PQConT4Aqa0+AKqlPgCrvT4ArKU+AK2tPgCupT4Ar30+AISsBAC+rAQAo2MAgKdjAICrYwCAr2MAgLNjAIC3YwCAqPkFAKn5BQCqKQYAqykGAKw5BgCtOQYArikGAK8pBgB3YwCAu2MAgL9jAIDDYwCAx2MAgMtjAIDPYwCA02MAgLiNBgC5kQYAupEGALulBgC8vQYAvUUHAL5BBwC/QQcAsFkGALFZBgCy7QYAs/0GALTtBgC13QYAttUGALe1BgCzoQYA12MAgNtjAIDfYwCA42MAgLa5BgC1sQYA62MAgLudBgC6nQYA52MAgAdjAIC/GQYAvikGAL0pBgC8OQYAglEAAKPlBgCAQQAAgUEAAKb9BgDvYwCA82MAgKX1BgCq2QYAq9kGAIZIAACHbAAArm0GAK9dBgCsfQYArW0GAKg5BgCpWQYAqmkGAKtpBgCseQYArXkGAK5pBgCvaQYA92MAgPtjAID/YwCAA2QAgAdkAIALZACAD2QAgBNkAIC4ZQEAuW0BALplAQC7fQEAvGUBAL1tAQC+ZQEAv9kBALAZBgCxGQYAsoEGALOBBgC0gQYAtYEGALaBBgC3gQYAs+EGABdkAIAbZACAH2QAgCNkAIC2+QYAtfEGACdkAIC73QYAut0GACtkAIAvZACAv0UGAL5FBgC9VQYAvFUGADNkAICjpQYAN2QAgDtkAICmvQYAP2QAgENkAICltQYAqpkGAKuZBgBHZACAS2QAgK4BBgCvAQYArBEGAK0RBgConQIAqdECAKrRAgCrLQMArDUDAK09AwCuNQMAry0DAE9kAIBTZACAvmQCAFtkAIBfZACAY2QAgGdkAIBrZACAuOkDALnpAwC6iQMAu4UDALydAwC9gQMAvoEDAL+1AwCwVQMAsV0DALJVAwCz6QMAtPkDALX5AwC26QMAt+EDAIBtAwCBpQAAgq0AALNVAgBvZACAtbEDALaxAwBzZACAhOACAHdkAIC6nQMAu5UDALyNAwC9MQMAvjEDAL8xAwCjGQIAe2QAgIVwaQB/ZACAg2QAgKb9AwCl/QMAh2QAgKvZAwCq0QMAhkgMAIe8AwCvfQMArn0DAK19AwCswQMAi2QAgI9kAICTZACAl2QAgO+wBgDvxAMAm2QAgJ9kAIDjfAYA45QDAOG4BwDh3AEAo2QAgKdkAICrZACAr2QAgLNkAIC3ZACAhEQCAL5YDQCADQAAgTUAAII9AAC7ZACAv2QAgMNkAICGyAwAh1wNAMtkAIDPZACA02QAgNdkAIDbZACA32QAgONkAIDnZACA62QAgO9kAIDzZACA74AGAISsDQDh7AYA92QAgONcBgD7ZACA/2QAgANlAIAHZQCAs/UBAAtlAIAPZQCAE2UAgBdlAIC2RQEAteUBABtlAIC7LQEAuiEBAB9lAIAjZQCAv/UAAL71AAC9JQEAvC0BAKgtDgCpNQ4Aqj0OAKs1DgCsLQ4ArYUOAK6FDgCvuQ4Ax2QAgCdlAIArZQCAL2UAgIAZAACBGQAAggUAADNlAIC4WQ8AuVkPALp5DwC7eQ8AvGkPAL1pDwC+GQ8AvxkPALClDgCxqQ4AsrkOALOxDgC0cQ8AtXEPALZxDwC3cQ8Apb0OAL6IAwA7ZQCAph0OADdlAIA/ZQCAo60OAENlAICtfQ4ArHUOAK+tDwCurQ8AV2QAgEdlAICrdQ4AqnkOALO5DwBLZQCAhmgAAIcMAwBPZQCAtlEPALVZDwBTZQCAu3UPALp1DwBXZQCAW2UAgL9FDwC+RQ8AvVEPALxlDwCocQ4AqXEOAKpxDgCrcQ4ArJEOAK2RDgCukQ4Ar5EOAF9lAIBjZQCAZ2UAgGtlAIBvZQCAc2UAgHdlAIB7ZQCAuIUOALmNDgC6hQ4Au50OALyNDgC9vQ4AvrUOAL95AQCw8Q4AsfEOALLxDgCzxQ4AtMEOALXBDgC2wQ4At8EOAKP5DgB/ZQCAg2UAgIdlAICLZQCAphEOAKUZDgCPZQCAqzUOAKo1DgCTZQCAl2UAgK8FDgCuBQ4ArREOAKwlDgCADQAAgRUAAIIdAACbZQCAn2UAgKNlAICElAEAvpQBAIZABwCH5AAAq2UAgK9lAICzZQCAt2UAgLtlAIC/ZQCAqIkCAKmRAgCqlQIAq7kCAKzVAgCtxQIArsUCAK/1AgDDZQCAx2UAgMtlAIDPZQCAvnwDANNlAIDXZQCA22UAgLh9AwC5wQMAusEDALvBAwC8wQMAvckDAL7xAwC/8QMAsI0CALFFAwCyTQMAs0UDALRdAwC1RQMAtk0DALdFAwCzHQIA32UAgONlAIDnZQCA62UAgLZFAgC1XQIA72UAgLuBAwC6SQIA82UAgPdlAIC/gQMAvpkDAL2RAwC8mQMA+2UAgKNZAgD/ZQCAA2YAgKYBAgAHZgCAC2YAgKUZAgCqDQIAq8UDAA9mAIATZgCArt0DAK/FAwCs3QMArdUDAIDZAQCB7QEAguUBAO+4DgAbZgCA4cQBAISYAgDj1AAAH2YAgL7sBAAjZgCA7wgAACdmAIDhxA8AK2YAgONkDgCGAAUAh2gFAC9mAICzvQIAM2YAgLWtAgC2pQIAN2YAgDtmAIA/ZgCAukEBALtBAQC8RQEAvU0BAL5FAQC/+QEAQ2YAgEdmAIBLZgCAT2YAgFNmAIBXZgCAW2YAgO/gAQCEbAQA4dQOAF9mAIDjHA4AY2YAgGdmAIBrZgCAb2YAgKMxAgBzZgCAhCQHAHdmAIB7ZgCApikCAKUhAgB/ZgCAq80BAKrNAQCDZgCAi2YAgK91AQCuyQEArcEBAKzJAQCo6QUAqekFAKr5BQCr+QUArOkFAK3pBQCuOQYArzkGABdmAICCzQcAgfUHAID9BwCHZgCAj2YAgIYYAwCHkAMAuNEGALnZBgC64QYAu+EGALyRBgC9nQYAvpUGAL+JBgCwSQYAsUkGALJdBgCzVQYAtE0GALXxBgC28QYAt/EGALDhBwCx4QcAsgkHALMJBwC0GQcAtRkHALYJBwC3CQcAuDkHALkNBwC6GQcAuxkHALwJBwC9CQcAvn0HAL9xBwCTZgCAp2UAgJdmAICbZgCAn2YAgKNmAICnZgCAq2YAgKjxBwCpxQcAqsEHAKvdBwCsyQcArb0HAK6pBwCvoQcAsykGAK9mAICzZgCAt2YAgLtmAIC2XQYAtSEGAL9mAIC7RQYAukUGAMNmAIDHZgCAv70GAL69BgC9vQYAvL0GAMtmAICjbQYAz2YAgNNmAICmGQYA12YAgNtmAIClZQYAqgEGAKsBBgDfZgCA42YAgK75BgCv+QYArPkGAK35BgCobQYAqbEBAKpJAQCrRQEArF0BAK1FAQCuTQEAr0UBAOdmAICCHQAAgR0AAIAdAADrZgCA72YAgPNmAIC+VAEAuIEAALmNAAC6hQAAu5kAALyJAAC9vQAAvrUAAL99AACwPQEAseEAALLhAACz4QAAtOEAALXpAAC20QAAt9EAALsFAwC62QIAhiwCAIcsAwC/DQMAvgUDAL0VAwC8FQMAs+ECAPtmAID/ZgCAhCwDAANnAIC25QIAtfUCAAdnAICqnQIAq0EDAAtnAIAPZwCArkEDAK9JAwCsUQMArVEDABNnAICjpQIAF2cAgBtnAICmoQIAH2cAgCNnAIClsQIAqakAAKihAACrtQAAqr0AAK3dAACs3QAAr/EAAK79AAC+LBwAJ2cAgCtnAIAvZwCAM2cAgDdnAIA7ZwCAP2cAgLl9AAC4fQAAu80BALrNAQC93QEAvN0BAL/NAQC+zQEAsZUAALCJAACzTQAAspUAALVdAAC0XQAAt00AALZNAABDZwCAR2cAgEtnAIBPZwCAU2cAgFdnAIBbZwCAX2cAgIA5AACBOQAAggUAAGNnAIBrZwCAb2cAgIf4AgCGfB0A4bgEAL7IHADjQAYAc2cAgHdnAIB7ZwCAf2cAgINnAICHZwCAi2cAgI9nAICTZwCAl2cAgJtnAIDvsAcAn2cAgKNnAICnZwCAq2cAgO/IAACvZwCAs2cAgLdnAIDvQAYAu2cAgOH8BgC/ZwCA4xwGAMNnAIDhlAEAx2cAgONkBgCAEQAAgRkAAIIpAACz/QEAy2cAgLWdAQC2lQEAz2cAgNNnAICEbB0AuoUBALuZAQC8iQEAvVEBAL5RAQC/UQEAozEeAGdnAIDXZwCA22cAgN9nAICmWR4ApVEeAONnAICrVR4AqkkeAIYIAwCHbAMAr50eAK6dHgCtnR4ArEUeAOdnAICzCR8A62cAgO9nAIC2CR8A82cAgPdnAIC1CR8AugUfALsNHwD7ZwCA/2cAgL4FHwC/CR8AvBUfAL0NHwCw5R8Ase0fALLlHwCz/R8AtOUfALXpHwC2GR8AtxkfALgpHwC5NR8Auj0fALs1HwC8ER8AvR0fAL4JHwC/BR8AA2gAgAdoAID3ZgCAC2gAgA9oAIATaACAF2gAgBtoAICo0R8AqdEfAKqlHwCrvR8ArKUfAK2tHwCupR8Ar50fAKNNHgAfaACAI2gAgCdoAIAraACApk0eAKVNHgAvaACAq0keAKpBHgAzaACAN2gAgK9NHgCuQR4ArUkeAKxRHgCADQAAgRUAAIIdAAA7aACAP2gAgENoAICEtAEAvrQBAL/oAQBLaACAhkgHAIc0AACEvAYAT2gAgFNoAIC+tAYAqI0BAKmVAQCqlQEAq80BAKzZAQCt2QEArs0BAK/FAQBXaACAW2gAgF9oAIBjaACAZ2gAgGtoAIBvaACAc2gAgLgdAQC5wQAAusEAALvBAAC8wQAAvckAAL7xAAC/8QAAsIkBALGJAQCyKQEAsykBALQ9AQC1JQEAti0BALclAQC7bQIAum0CAHdoAIB7aACAv8ECAL7ZAgC93QIAvN0CALM9AgB/aACAg2gAgIdoAICE/AYAtnkCALVxAgCLaACAqikCAKspAgCPaACAk2gAgK6dAgCvhQIArJkCAK2ZAgCXaACAo3kCAJtoAICfaACApj0CAKNoAICnaACApTUCAIJtJwCDjSoAhqgFAIdsAwCGmS4Ah80vAIQRLgCFmS4AiiESAIspEgCraACAr2gAgI6RFgCPHRYAjBESAI0RFgCScRoAk+UaALNoAIDvlHYAlvEeAJflHgCUSRoAlRkeAJopAgCb4QIAu2gAgL9oAIDDaACA4SASAJzxAgDjIBYAnyEfAJ7BHwCdmRsAnC0bAJuhGwCavRcAmTkXAJixFwCXiRMAlqkTAJWpEwCUdS4AkzkvAJIxLwCRsS8AkDUrAI+tJgDjeB8A0gAAAOFcHwCCmQEAx2gAgIDxAQCB8QEAvqgHAMtoAIDPaACA02gAgIS8BgDvLB8A12gAgNtoAIDhpB4A48wAAON8HgDhvAEA32gAgONoAIDnaACAhJwGAOtoAIC+bAYA72gAgPNoAID3aACA7xAAAO8EHgD7aACA/2gAgANpAIAHaQCAC2kAgA9pAIATaQCAF2kAgBtpAICAPQAAgQkAAILJBwAfaQCAo/kDAKLxAwChMQMAoM0fALBJcQCxAXwAsgl8ALMhfQC0AXgAtRV4AEdoAIC3aACAI2kAgL4oDgCGDAAAh4wDACdpAIAraQCAL2kAgDNpAIA3aQCAoV0AAKJVAACjfQAApAEMAKUVDACm9QwApwEIAKghCACpxQgAqgF0AKsJdACsAXQArR11AK55cACveXAAqOUFAKnxBQCq8QUAqy0FAKw1BQCtPQUArjUFAK8tBQA7aQCAP2kAgENpAIBHaQCAS2kAgE9pAIBTaQCAV2kAgLj9BgC5jQYAuoUGALutBgC8uQYAvbkGAL6tBgC/pQYAsFUFALFdBQCyVQUAs+UGALT9BgC10QYAttEGALfRBgCzeQQAW2kAgF9pAIBjaQCAZ2kAgLa9BAC1vQQAa2kAgLuZBAC6kQQAb2kAgHNpAIC/FQcAvjkHAL0xBwC8gQQAd2kAgKM9BAB7aQCAf2kAgKb5BACDaQCAh2kAgKX5BACq1QQAq90EAItpAICPaQCArn0HAK9RBwCsxQQArXUHAKhpBwCpaQcAqnkHAKvZBgCs9QYArf0GAK71BgCv5QYAgMkAAIHJAACCBQAAk2kAgIZwDwCHNAAAm2kAgJ9pAIC4fQYAuQUGALoNBgC7BQYAvB0GAL0FBgC+DQYAvwUGALCdBgCxdQYAsn0GALN1BgC0UQYAtV0GALZVBgC3TQYAs/EEAKNpAICnaQCAq2kAgK9pAIC2fQUAtX0FALNpAIC7sQUAulkFALdpAIC7aQCAv5kFAL6VBQC9oQUAvKkFAL9pAICjtQQAw2kAgMdpAICmOQUAy2kAgM9pAIClOQUAqh0FAKv1BQDTaQCA12kAgK7RBQCv3QUArO0FAK3lBQCpuQIAqLECAKvJAgCqsQIArTUCAKw1AgCvNQIArjUCANtpAIDfaQCA42kAgOdpAIDraQCA72kAgPNpAID3aQCAuekDALjZAwC7iQMAuuEDAL2dAwC8nQMAv4EDAL6JAwCxVQIAsFUCALNVAgCyVQIAtfkDALTxAwC36QMAtvEDALM9AwD7aQCA/2kAgANqAIALagCAtrEDALW5AwAPagCAu5UDALqVAwCGiAwAh6ANAL85AgC+MQIAvYUDALyFAwATagCAo3kDABdqAIAbagCApvUDAB9qAIAjagCApf0DAKrRAwCr0QMAJ2oAgCtqAICudQIAr30CAKzBAwCtwQMAgIUAAIGNAACChQAA79AGAOOwBwDj9AQA4QgHAOHsBADvOAYA7yAEAL6kDAAvagCAM2oAgOGEAQA3agCA49wGADtqAIA/agCAhMANALPJAQBDagCAtdkBALbJAQBHagCAS2oAgE9qAIC6xQEAu60BALy5AQC9uQEAvq0BAL+lAQCwLQ4AsUUOALJBDgCzQQ4AtEUOALVNDgC2cQ4At3EOALiBDgC5gQ4AuoEOALuBDgC8gQ4AvYEOAL6BDgC/gQ4AB2oAgFNqAIBXagCAW2oAgJdpAIBfagCAY2oAgGdqAICo2Q0AqdkNAKptDgCrZQ4ArH0OAK1lDgCuZQ4Ar1UOAKOFDgCCLQAAgRUAAIAdAABragCApoUOAKWVDgBvagCAq+EOAKqJDgBzagCAd2oAgK/pDgCu4Q4ArfUOAKz1DgB7agCAs4UPAIZoAACHHAMAtoUPAH9qAICDagCAtZEPALqNDwC7SQ8Ah2oAgItqAIC+MQ8AvzEPALxJDwC9RQ8AqBEOAKkZDgCqSQ4Aq0UOAKxdDgCtQQ4ArkEOAK91DgCPagCAk2oAgJdqAICbagCAn2oAgKNqAICnagCAq2oAgLihDgC5oQ4Aug0BALsFAQC8HQEAvQEBAL4BAQC/AQEAsA0OALHJDgCy2Q4As9UOALSxDgC1sQ4AtqkOALehDgCjwQ4Ar2oAgLNqAIC3agCAu2oAgKbBDgCl1Q4Av2oAgKsNDgCqyQ4Aw2oAgMdqAICvdQ4ArnUOAK0BDgCsDQ4Ay2oAgM9qAIDTagCA12oAgIANAACBNQAAgj0AANtqAIDfagCA42oAgISEAQC+hAEAhjAHAIf4AADragCA72oAgKjBAgCp0QIAqtECAKvlAgCs/QIArTUDAK49AwCvNQMA82oAgPdqAID7agCA/2oAgANrAIAHawCAC2sAgA9rAIC40QMAudkDALrhAwC74QMAvJEDAL2RAwC+kQMAv5EDALBNAwCxVQMAsl0DALNVAwC0TQMAtfEDALbxAwC38QMAu7EDALqpAwATawCAvoQDAL8VAwC+qQMAvaEDALypAwCzeQIAF2sAgBtrAIAfawCAI2sAgLaVAwC1VQIAJ2sAgKrtAwCr9QMAK2sAgC9rAICu7QMAr1EDAKztAwCt5QMAM2sAgKM9AgA3awCAO2sAgKbRAwA/awCAQ2sAgKURAgBHawCAgiEAAIEVAACAFQAA7wQAAISUAgBLawCAT2sAgOPYAABTawCA4fgBAFtrAIBfawCAY2sAgGdrAIBrawCAhmAFAIcIBQBvawCAs20BAHNrAIC1fQEAtnUBAHdrAIB7awCAf2sAgLpRAQC7UQEAvPkBAL3RAQC+0QEAv9EBAINrAICjpQEAh2sAgItrAICmvQEAj2sAgJNrAICltQEAqpkBAKuZAQCXawCAm2sAgK4ZAQCvGQEArDEBAK0ZAQCfawCA4fQOAKNrAIDjFA4A9AAAAOF8DACnawCA41AKAKtrAICvawCAviAEAO8wDQCzawCAt2sAgIQ0BADvrA4AsDkGALE5BgCygQYAs6kGALS5BgC1uQYAtqkGALehBgC46QYAuekGALrJBgC7xQYAvN0GAL3BBgC+wQYAvz0HAFdrAICCHQAAgR0AAIAdAAC7awCAv2sAgMNrAIDnagCAqJkFAKmZBQCqSQYAq0kGAKxZBgCtWQYArkkGAK9JBgCorQcAqbUHAKq9BwCrtQcArK0HAK3dBwCuyQcAr8EHAMdrAIDLawCAhogDAIcQAwDPawCA02sAgNdrAIDbawCAuG0HALkFBwC6AQcAuxUHALwxBwC9MQcAvikHAL8pBwCwgQcAsYEHALJpBwCzZQcAtH0HALVhBwC2YQcAt1UHALM1BgDfawCA42sAgOdrAIDrawCAtl0GALUlBgDvawCAu0UGALpFBgDzawCA92sAgL+lBgC+uQYAvbEGALy9BgD7awCAo3EGAP9rAIADbACAphkGAAdsAIALbACApWEGAKoBBgCrAQYAD2wAgBNsAICu/QYAr+EGAKz5BgCt9QYAqCUBAKk1AQCqPQEAqzUBAKwtAQCtkQAArpEAAK+RAAAXbACAG2wAgB9sAIAjbACAJ2wAgIK9AwCBvQMAgL0DALiZAAC5rQAAuqUAALttAAC8dQAAvX0AAL51AAC/bQAAsPEAALH5AACywQAAs8EAALSxAAC1vQAAtrUAALepAAArbACAL2wAgDNsAICEgAIAvhwCADtsAICG+HwAh8wCAISsAwA/bACAQ2wAgEdsAIBLbACAT2wAgFNsAIBXbACAs/UCAFtsAIBfbACAkgAAAGNsAIC2UQMAteUCAGdsAIC7fQMAunUDAGtsAIBvbACAvzkDAL41AwC9VQMAvFUDAKM1AgBzbACAd2wAgHtsAIB/bACAppEDAKUlAgCDbACAq70DAKq1AwCHbACAi2wAgK/5AwCu9QMArZUDAKyVAwC+wAMAj2wAgJNsAICXbACAgA0AAIE1AACCPQAAm2wAgJ9sAICjbACAhsh8AIcAAwCrbACAr2wAgLNsAIC3bACAu2wAgL9sAIDDbACAx2wAgMtsAIDPbACA02wAgO/0AwCE7HwA4ZQBANdsAIDjMAMA22wAgN9sAIDjbACA52wAgLNpAQDrbACA72wAgPNsAID3bACAtmEBALVpAQD7bACAuykBALohAQD/bACAA20AgL8dAQC+HQEAvSUBALwtAQAHbQCAC20AgA9tAICjpQEAE20AgKWlAQCmrQEAvlR8AIaAfACH7HwAqu0BAKvlAQCs4QEArekBAK7RAQCv0QEAG20AgOGcBgCEBH8A4yQGAOPUBgAfbQCA4TAEACNtAIDvlAcAgnUAAIFhAACAaQAAJ20AgCttAIAvbQCA7+wGALiNfgC5lX4AupV+ALulfgC8vX4AvdF+AL7RfgC/0X4AsGV+ALFtfgCyeX4As3F+ALRZfgC1WX4Atr1+ALe1fgCoVX4AqWF+AKphfgCrYX4ArGF+AK1hfgCuYX4Ar2F+ADNtAICnbACAN2wAgDdtAIAXbQCAO20AgD9tAIBDbQCAqHF+AKlxfgCqcX4Aq3F+AKyRfwCtkX8ArpF/AK+RfwBHbQCAS20AgE9tAIBTbQCAV20AgFttAIBfbQCAY20AgLiFfwC5jX8AuoV/ALudfwC8jX8Avb1/AL61fwC/XX8AsPF/ALHxfwCy8X8As8V/ALTBfwC1wX8AtsF/ALfBfwCz+X8AZ20AgGttAIBvbQCAc20AgLYRfgC1GX4Ad20AgLs1fgC6NX4Ae20AgH9tAIC/BX4AvgV+AL0RfgC8JX4AghUAAKO9fwCAYQAAgWEAAKZVfgCDbQCAvpABAKVdfgCqcX4Aq3F+AIdtAICLbQCArkF+AK9BfgCsYX4ArVV+AKhBfgCpUX4AqlV+AKt9fgCsZX4ArW1+AK75AQCv8QEAhgAAAIc0AQCPbQCAk20AgJdtAICbbQCAn20AgKNtAIC4dQEAuX0BALp1AQC7yQAAvNkAAL3ZAAC+yQAAv8EAALCVAQCxnQEAspUBALNNAQC0VQEAtV0BALZVAQC3TQEAs919AKdtAICrbQCAr20AgLNtAIC27X0Ate19ALdtAIC7WQIAulECALttAIC/bQCAv5kCAL6RAgC9mQIAvEECAMNtAICjmX0Ax20AgMttAICmqX0Az20AgNNtAIClqX0AqhUCAKsdAgDXbQCA220AgK7VAgCv3QIArAUCAK3dAgDfbQCA420AgOdtAIDrbQCAgB0AAIEJAACCOQAA720AgPNtAIC+AAQA+20AgP9tAIADbgCAB24AgAtuAIAPbgCAhIwDABNuAICHCAMAhuwEABduAIDviAIAG24AgB9uAICEbAQA4zQCACNuAIDhVAEAJ24AgCtuAIAvbgCAM24AgKhtAgCprQIAqqUCAKu9AgCspQIAra0CAK6lAgCvGQEAvqwEADduAIA7bgCAP24AgENuAIBHbgCAS24AgE9uAIC4DQEAuREBALoRAQC7JQEAvD0BAL3VAQC+3QEAv9UBALBpAQCxaQEAsnkBALNxAQC0WQEAtVkBALY5AQC3NQEAsy0CAFNuAIBXbgCAW24AgF9uAIC2LQIAtS0CAGNuAIC7rQEAuq0BAGtuAIBvbgCAv50BAL6dAQC9pQEAvK0BAIBNAACBVQAAglUAAO9sAABzbgCA7+x/AO+8fgB3bgCA4RB/AOPUfwDj2H4A4ex/AHtuAIDhTH4Af24AgOMkfgD3bQCAZ24AgKsFBgCqBQYArQ0GAKwFBgCvNQYArjUGAIYAAwCHKAMAo4UFAINuAIClhQUAh24AgItuAICmhQUAs/EGAI9uAICTbgCAl24AgJtuAIC26QYAteEGAJ9uAIC7vQYAur0GAKNuAICnbgCAv4kGAL6BBgC9iQYAvJUGAKgpBgCpKQYAqjkGAKs5BgCsKQYArSkGAK5dBgCvTQYAq24AgK9uAICzbgCAt24AgLtuAIC/bgCAw24AgMduAIC46QcAuekHALr5BwC7+QcAvOkHAL3pBwC+XQcAv1UHALA5BgCxOQYAsgEGALMdBgC0BQYAtQ0GALYFBgC32QcAo7EHAIItAACBFQAAgB0AAMtuAICmqQcApaEHAM9uAICr/QcAqv0HANNuAICEpAIAr8kHAK7BBwCtyQcArNUHAL7MAQCzlQYA124AgNtuAIC2qQYA324AgONuAIC1rQYAulkBALshAQCGyAAAhwwBAL4hAQC/KQEAvDEBAL0xAQCoKQYAqSkGAKpZBgCrUQYArGEGAK1tBgCutQEAr6kBAITgAQDnbgCA624AgO9uAIDzbgCA924AgPtuAID/bgCAuGEBALlhAQC6YQEAu2EBALxhAQC9YQEAvmEBAL9hAQCw2QEAsaEBALKhAQCzoQEAtKEBALWpAQC2kQEAt5EBAKPRBQADbwCAB28AgAtvAIAPbwCApu0FAKXpBQATbwCAq2UCAKodAgAXbwCAG28AgK9tAgCuZQIArXUCAKx1AgAfbwCAI28AgCdvAIArbwCAL28AgDNvAIA3bwCAO28AgIA9AACBCQAAghkAAD9vAIBDbwCAS28AgL48AwBPbwCAhgAMAIcUAwBTbwCAs9UDAFdvAIC1PQMAtjUDAFtvAIBfbwCAv4wKALoRAwC7EQMAvLUAAL29AAC+tQAAv60AAGNvAIDjdAEAZ28AgOG8AQBrbwCAb28AgHNvAIB3bwCAe28AgH9vAICDbwCAh28AgItvAIDvdAIAj28AgJNvAICoTQIAqVECAKpRAgCrqQIArLkCAK25AgCuqQIAr6kCAIRsDQCXbwCAm28AgJ9vAICjbwCAp28AgKtvAIC+dA0AuG0BALkFAQC6DQEAuwUBALwdAQC9BQEAvg0BAL8FAQCw2QIAsdkCALJtAQCzZQEAtH0BALVlAQC2ZQEAt1UBAOG4AQDhUAcA47QAAON8BwCAqQAAgQkAAII5AACvbwCAs28AgLtvAIC/bwCAw28AgO4AAADHbwCA7wAAAO9kBgCGYAwAh+QMAKORAgDLbwCApXkCAM9vAIDTbwCApnECANdvAIDbbwCAq1UCAKpVAgCt+QEArPEBAK/pAQCu8QEAt28AgEdvAIDfbwCA428AgOdvAIDrbwCA728AgPNvAICoVQ4AqVkOAKqhDgCrvQ4ArK0OAK2VDgCu+Q4Ar/UOALCRDgCxkQ4AspEOALORDgC0sQ4AtbEOALaxDgC3sQ4AuJEOALmdDgC6lQ4Au0kPALxZDwC9WQ8AvkkPAL9JDwCzCQ4A928AgPtvAID/bwCAA3AAgLY1DgC1BQ4AB3AAgLt1DgC6dQ4AC3AAgA9wAIC/VQ4AvlUOAL1lDgC8ZQ4AE3AAgKNNDgAXcACAG3AAgKZxDgAfcACAI3AAgKVBDgCqMQ4AqzEOAISkAwC+pAMArhEOAK8RDgCsIQ4ArSEOAKilDgCprQ4AqqUOAKu5DgCs3Q4ArcEOAK7BDgCv/Q4AgO0BAIHxAQCC8QEAJ3AAgIaQAQCHtAEAK3AAgC9wAIC4yQEAuckBALrZAQC70QEAvPkBAL35AQC+mQEAv5UBALCFDgCxbQEAsmUBALN9AQC0ZQEAtW0BALZlAQC3+QEAsy0OADNwAIA3cACAO3AAgD9wAIC2QQ4AtVUOAENwAIC7qQEAukEOAEdwAIBLcACAv6kBAL6hAQC9qQEAvLEBAE9wAICjaQ4AU3AAgFdwAICmBQ4AW3AAgF9wAIClEQ4AqgUOAKvtAQBjcACAZ3AAgK7lAQCv7QEArPUBAK3tAQCoOQMAqTkDAKqNAwCrhQMArJ0DAK2FAwCuhQMAr7UDAGtwAIBvcACAc3AAgHdwAIB7cACAf3AAgINwAICHcACAuGEAALlhAAC6YQAAu2EAALxhAAC9YQAAvmEAAL9hAACwzQMAsaUDALKhAwCzoQMAtKUDALWtAwC2kQMAt5EDAIANAACBEQAAghEAAItwAIDv9AIAj3AAgJNwAIC+HAMA4xQCAISIAgDhgAEAm3AAgJ9wAICjcACAh8gDAIY8BAC7AQMAumkDAKdwAICrcACAvwkDAL4BAwC9FQMAvBUDALNlAwCvcACAs3AAgLdwAIC7cACAtmUDALV1AwC/cACAw3AAgMdwAIDLcACAo4kCAM9wAIClmQIApokCANNwAICELAIA13AAgKqFAgCr7QIArPkCAK35AgCu7QIAr+UCANtwAIDfcACAvkQFAIRMBQDjcACA53AAgOtwAIDvcACA83AAgPdwAID7cACA/3AAgIAZAACBGQAAggUAAANxAIDhGA8A4VwOAOO4DgDjdAEAC3EAgA9xAIATcQCAF3EAgIYABACHZAUAG3EAgB9xAIAjcQCAJ3EAgO98DgDvqAEAs3UBACtxAIAvcQCAM3EAgDdxAIC2MQEAtRUBADtxAIC7HQEAuhUBAD9xAIBDcQCAv+EAAL79AAC9/QAAvP0AAAdxAIBHcQCAS3EAgE9xAICXcACAU3EAgFdxAIBbcQCAqI0GAKmVBgCqnQYAq+UGAKz9BgCt0QYArtEGAK/RBgCwsQYAsbkGALJJBwCzSQcAtFkHALVFBwC2RQcAt3kHALghBwC5IQcAujkHALs5BwC8KQcAvSkHAL4ZBwC/GQcAozUGAF9xAIBjcQCAZ3EAgGtxAICmcQYApVUGAG9xAICrXQYAqlUGAHNxAIC+oAMAr6EHAK69BwCtvQcArL0HAIBRAACBWQAAgmEAALNVBwCF9AAAtX0HALZ1BwB3cQCAhgAcAIfkAQC6LQcAuyUHALw9BwC9JQcAviUHAL8VBwCokQYAqZEGAKqRBgCrkQYArLkGAK25BgCuqQYAr6kGAHtxAIB/cQCAg3EAgIdxAICiIQEAozUBAKA5BQChEQQAuEkBALlJAQC6XQEAu1UBALxNAQC90QEAvtEBAL/RAQCwpQYAsa0GALKlBgCzvQYAtK0GALWdBgC2lQYAt3kBAKMZBgCPnXkAi3EAgI9xAICTcQCApjkGAKUxBgCXcQCAq2kGAKphBgCbcQCAn3EAgK9ZBgCuaQYArWkGAKxxBgCeiQgAn8EFAJzJCQCdyQkAmqENAJu9DACYsQ0AmbkNAJahcQCXRXEAlEV1AJWxcQCSoXUAk7V1AJDleQCRzXkAil1yAItFcgCjcQCAvoAcAI51DgCPZQ4AjLlyAI11DgCCOXoAgzl6AKdxAICrcQCAhnF2AIeZdgCECXoAhW12AJptBwCbVQIAr3EAgLNxAIC3cQCA4ZAAAJxZAgDjCBoAkgkPAJNlCgC7cQCA7zgWAJZ1BgCXdQYAlH0KAJU1CwCpjRYAqIUWAKsBEACqMRYArXESAKy1EgCvuS4ArgEsAKF9AgC/cQCAo6EeAKKpHgClsRoApPUfAKflGwCmsRoAhMwDAIRMHADDcQCAx3EAgMtxAIDPcQCA03EAgNdxAICxASgAsNkuALONKgCy6SoAtfUmALQBJACEcB0A23EAgID9AQCBFQAAgh0AAL6AHADfcQCA43EAgIe4AgCGPB0A63EAgO9xAIDzcQCA93EAgPtxAID/cQCAA3IAgAdyAIALcgCAD3IAgBNyAIAXcgCA44ADABtyAIDhoAEAH3IAgO+UAwAjcgCAJ3IAgCtyAIAvcgCAM3IAgDdyAIA7cgCAP3IAgOE8BgBDcgCA49AGAEdyAIDhMAcAS3IAgOOsBgCAOQAAgRUAAIIdAADvHAYAT3IAgFNyAIC+uB8A7+gBALPpAgBbcgCAh8QcAIbsHABfcgCAtlkCALVRAgBjcgCAu00CALpNAgBncgCAa3IAgL+5AQC+2QEAvdEBALz1AQCjKR0A53EAgFdyAIBvcgCAc3IAgKaZHQClkR0Ad3IAgKuNHQCqjR0Ae3IAgH9yAICveR4ArhkeAK0RHgCsNR4Ag3IAgLNtHwCHcgCAi3IAgLZlHwCPcgCAk3IAgLVtHwC6IR8AuyEfAJdyAICbcgCAviUfAL8pHwC8MR8AvTEfAKihHwCpoR8AqqEfAKuhHwCsoR8AraEfAK6hHwCvoR8An3IAgKNyAICncgCAq3IAgK9yAICzcgCAt3IAgLtyAIC4rR8AubUfALq9HwC7tR8AvK0fAL1VHwC+UR8Av00fALChHwCxoR8AsqEfALOhHwC0pR8AtakfALadHwC3lR8AoykeAIIZAACBGQAAgLEBAL9yAICmIR4ApSkeAMNyAICrZR4AqmUeAIaIAACH/AEAr20eAK5hHgCtdR4ArHUeAMdyAICzmR4Ay3IAgM9yAIC2XQEA03IAgNdyAIC1sR4AukkBALtJAQDbcgCA33IAgL49AQC/IQEAvDkBAL01AQCoRR4AqVUeAKpVHgCrZR4ArH0eAK2ZAQCuiQEAr4EBAISsAADjcgCA53IAgOtyAIDvcgCA83IAgPdyAID7cgCAuK0BALllAQC6bQEAu2UBALx9AQC9ZQEAvm0BAL9lAQCwyQEAsckBALKpAQCzpQEAtL0BALWhAQC2oQEAt5UBALhpHAC5oRwAusEcALvBHAC8wRwAvcEcAL7BHAC/wRwAsIkfALGJHwCyIRwAswUcALQdHAC1fRwAtnUcALdtHACoYR8AqWEfAKphHwCrYR8ArNkfAK3ZHwCuyR8Ar8EfAP9yAIADcwCAB3MAgAtzAIAPcwCAE3MAgBdzAIAbcwCAH3MAgCNzAIC+AAQAo1EdACdzAICleR0AppUCACtzAIAvcwCAM3MAgKqBAgCrgQIArPECAK39AgCu9QIAr+kCADtzAIDh9AEAP3MAgON8AQCATQAAgXUAAIJ9AABDcwCAhsAEAIekBABHcwCAS3MAgE9zAIBTcwCAV3MAgO+MAgCoSQIAqUkCAKpdAgCrVQIArHkCAK15AgCuvQIAr7UCAISgBQBbcwCAX3MAgGNzAIC+vAQAZ3MAgGtzAIBvcwCAuC0BALk1AQC6PQEAuzUBALwtAQC91QEAvt0BAL/NAQCwzQIAsdUCALLdAgCz1QIAtM0CALUVAQC2HQEAtxUBAOGEHgDjbB8A41wfAOFYHgBzcwCAd3MAgHtzAIB/cwCAg3MAgIdzAICLcwCAj3MAgOkAAADv9B4A70weAJNzAICzlQIAl3MAgJtzAICfcwCAo3MAgLa5AgC1sQIAq3MAgLtRAgC6SQIAhsgEAIesBAC/kQEAvkkCAL1BAgC8SQIAN3MAgKNRBQCvcwCAp3MAgKZ9BQCzcwCAt3MAgKV1BQCqjQUAq5UFALtzAIC/cwCAro0FAK9VBgCsjQUArYUFAICJBwCBiQcAgpkHALORBgDDcwCAtbkGALapBgDHcwCAy3MAgM9zAIC6TQcAu0UHALxdBwC9QQcAvkEHAL9BBwCoQQYAqU0GAKpVBgCrZQYArH0GAK1lBgCubQYAr2UGANNzAIDXcwCA23MAgN9zAIDjcwCA53MAgOtzAIDvcwCAuFkHALlZBwC6aQcAu2kHALx5BwC9eQcAvmUHAL8ZBwCwxQcAsc0HALLFBwCz2QcAtMkHALXJBwC2aQcAt2kHAKPdBwDzcwCA93MAgPtzAID/cwCApuUHAKX1BwADdACAqwkGAKoBBgAHdACAC3QAgK8NBgCuDQYArQ0GAKwRBgCAbQAAgQkAAIIZAAAPdACAE3QAgISYAQC+kAEAF3QAgIbAAACH5AEAG3QAgB90AIAjdACAJ3QAgCt0AIAvdACAqF0GAKmNAQCqnQEAq5UBAKy5AQCtuQEArskBAK/BAQCEoAAAM3QAgDd0AIA7dACAP3QAgEN0AIBHdACAS3QAgLh5AQC5eQEAus0AALvFAAC83QAAvcUAAL7FAAC/9QAAsIEBALGBAQCySQEAs0kBALRZAQC1WQEAtkkBALdJAQCzFQIAT3QAgFN0AIBXdACAW3QAgLY5AgC1MQIAX3QAgLtFAgC6RQIAY3QAgGd0AIC/nQIAvp0CAL2dAgC8nQIAhXw+AKNRAgBrdACAb3QAgKZ9AgBzdACAd3QAgKV1AgCqAQIAqwECAHt0AIB/dACArtkCAK/ZAgCs2QIArdkCAIDpAACB6QAAggUAAIN0AIC+AAwAi3QAgIeoAwCGvAwAj3QAgJN0AICXdACAm3QAgJ90AICjdACAp3QAgKt0AICvdACAs3QAgLd0AIC7dACA42ABAL90AIDhoAEAw3QAgO+IAgDHdACAy3QAgM90AIDTdACA13QAgNt0AIDfdACAqGkCAKlpAgCqeQIAq3kCAKxpAgCtaQIArr0CAK+1AgC+rAwA43QAgOd0AIDrdACAgB0AAIEJAACCqQAA73QAgLhRAQC5WQEAumEBALthAQC8GQEAvRkBAL4NAQC/BQEAsM0CALHVAgCy3QIAs9UCALTNAgC1cQEAtnEBALdxAQDjxAAA4XwHAOF4BgDjvAYA83QAgIQYDQCGuAwAhzwNAL4sDwD7dACA/3QAgAN1AIDvEAAAB3UAgAt1AIDvdAYAD3UAgBN1AIAXdQCAs70CABt1AIC1rQIAtqUCAB91AIAjdQCAJ3UAgLpFAgC7XQIAvEUCAL1NAgC+RQIAv/kBAId0AIClfQ0ApnUNAPd0AIArdQCAL3UAgDN1AICjbQ0ArJUNAK2dDQCulQ0ArykOADd1AIA7dQCAqpUNAKuNDQCz5Q4AP3UAgEN1AIBHdQCAS3UAgLblDgC19Q4AT3UAgLuhDgC62Q4AU3UAgFd1AIC/pQ4AvrkOAL2xDgC8uQ4AqBUOAKklDgCqLQ4AqyUOAKw9DgCtJQ4Ari0OAK8lDgCADQAAgRUAAIIdAABbdQCAX3UAgGN1AICEMAMAZ3UAgLgpDgC5KQ4AujkOALs5DgC8KQ4AvSkOAL79DwC/9Q8AsF0OALElDgCyLQ4AsyUOALQ9DgC1IQ4AtiUOALcZDgCjpQ8Aa3UAgIYoAQCHTAEAb3UAgKalDwCltQ8Ac3UAgKvhDwCqmQ8Ad3UAgHt1AICv5Q8ArvkPAK3xDwCs+Q8Af3UAgLPpDgCDdQCAh3UAgLaRDgCLdQCAj3UAgLXlDgC6sQ4Au7kOAJN1AICXdQCAvmEBAL9hAQC8mQ4AvZkOAKglDgCpLQ4AqiUOAKs5DgCsKQ4ArVUOAK5dDgCvVQ4Am3UAgJ91AICjdQCAp3UAgKt1AICvdQCAs3UAgLd1AIC49QEAuYEBALqBAQC7gQEAvIEBAL2JAQC+sQEAv7EBALAxDgCxOQ4AsgkOALMJDgC04QEAteEBALbhAQC3zQEAo60NALt1AIC/dQCAw3UAgMd1AICm1Q0ApaENAMt1AICr/Q0AqvUNAM91AIDTdQCAryUCAK4lAgCt3Q0ArN0NAIBdAACBbQAAgmUAALNRAwC+nAMAtXkDALYZAwDbdQCAhOACAN91AIC6PQMAuzUDALwZAwC9GQMAvtkDAL/ZAwCohQMAqZUDAKqVAwCrpQMArL0DAK3VAwCu0QMAr9EDAIYABACHNAMAv6AzAON1AIDndQCA63UAgO91AIDzdQCAuHEDALlxAwC6cQMAu3EDALzVAAC93QAAvtUAAL/NAACwtQMAsb0DALKBAwCzgQMAtFEDALVRAwC2UQMAt1EDAO+oAwD3dQCA+3UAgP91AICEHAIAA3YAgAd2AIALdgCAviwFAA92AIATdgCAF3YAgONAAwAbdgCA4SgAAB92AICjXQIAI3YAgCd2AIArdgCAL3YAgKYVAgCldQIAM3YAgKs5AgCqMQIAN3YAgDt2AICv1QIArtUCAK0VAgCsFQIA4ygBAOEADwDhCA4A4wgOAID9AACBCQAAgjkAAD92AIBDdgCAS3YAgE92AIBTdgCA7+gOAFd2AIBbdgCA72QOALNtAQBfdgCAhugEAIcMBQBjdgCAtm0BALVtAQBndgCAu+0AALrtAABrdgCAb3YAgL/VAAC+6QAAveEAALzpAACoXQYAqWEGAKqlBgCrvQYArKUGAK2tBgCupQYArxkHAEd2AIBzdgCAd3YAgHt2AIB/dgCAg3YAgId2AICLdgCAuHUHALl5BwC6DQcAuwUHALwdBwC9BQcAvgUHAL81BwCwaQcAsWkHALJ9BwCzdQcAtG0HALVRBwC2UQcAt1EHAKMtBgCPdgCAk3YAgJd2AICbdgCApi0GAKUtBgCfdgCAq60HAKqtBwCjdgCAp3YAgK+VBwCuqQcAraEHAKypBwCADQAAgRUAAIIdAACrdgCAr3YAgLN2AICEVAMAvlwAALd2AIC7dgCAhugAAIdMAwC/dgCAw3YAgMd2AIDLdgCAz3YAgOMEBADTdgCA4bQFANd2AIDbdgCA33YAgON2AIDndgCA63YAgO92AIDzdgCA93YAgO/sBAD7dgCA/3YAgLPtBgADdwCAB3cAgAt3AIAPdwCAtpEGALXhBgATdwCAu40GALqNBgAXdwCAG3cAgL9BAQC+WQEAvVEBALxZAQCoJQYAqS0GAKolBgCrOQYArCkGAK1RBgCuSQYAr0EGAIDNAACBCQAAghkAAB93AIAjdwCAhCwBAL40AAArdwCAuP0BALlBAQC6QQEAu0EBALxBAQC9SQEAvnEBAL9xAQCwCQYAsQkGALLNAQCzxQEAtN0BALXFAQC2zQEAt8UBAIagPACHRAMAL3cAgKOhBQAzdwCApa0FAKbdBQA3dwCAO3cAgL4oPACqwQUAq8EFAKwVAgCtHQIArhUCAK8NAgC2QQMAP3cAgEN3AIC1sQIAR3cAgLOhAgBLdwCAT3cAgL5FAwC/TQMAvHUDAL1NAwC6ZQMAu20DAFN3AIBXdwCAW3cAgF93AIDXdQCAY3cAgGd3AIBrdwCAb3cAgHN3AICoRQIAqVUCAKpdAgCrVQIArE0CAK21AwCusQMAr60DALDVAwCx3QMAstUDALPtAwC09QMAtf0DALb1AwC37QMAuNkDALnZAwC6rQMAu6UDALy9AwC9pQMAvqUDAL+VAwCj9QMAd3cAgHt3AIB/dwCAg3cAgKYVAgCl5QMAh3cAgKs5AgCqMQIAi3cAgI93AICvGQIArhECAK0ZAgCsIQIAgGkAAIFpAACCBQAAk3cAgJt3AICfdwCAo3cAgO8cAACEbAIA4ZQBAKd3AIDjyAAAq3cAgK93AICGWDwAh1A9ALN3AIC3dwCAu3cAgISEPQC/dwCAw3cAgMd3AIDvuAEAvmw8AOF0BgDLdwCA42QBAM93AIDTdwCA13cAgNt3AICz0QEA33cAgON3AIDndwCA63cAgLaRAQC1+QEA73cAgLu9AQC6vQEA83cAgPd3AIC/dQEAvnUBAL2FAQC8hQEAqL09AKkNPgCqGT4AqxE+AKwxPgCtUT4ArlE+AK9NPgCXdwCAgh0AAIEdAACAHQAA+3cAgP93AIADeACAB3gAgLjVPgC53T4AutU+ALtJPwC8WT8AvVk/AL5JPwC/QT8AsDk+ALE5PgCyET4AsxE+ALTxPgC18T4AtvU+ALftPgCjkT4AC3gAgIYoAACHwAMAD3gAgKbRPgCluT4AE3gAgKv9PgCq/T4AF3gAgBt4AICvNT4ArjU+AK3FPgCsxT4AH3gAgLOdPwAjeACAJ3gAgLalPwAreACAL3gAgLWtPwC6aT8Au3U/ADN4AIA3eACAvlk/AL9FPwC8bT8AvWU/ADt4AIA/eACAQ3gAgEd4AIDjYDwAS3gAgOEAPQBPeACA7/w9AFN4AIBXeACAW3gAgF94AIBjeACAZ3gAgGt4AICjGT4AghkAAIEZAACAcQAAb3gAgKYhPgClKT4Ac3gAgKvxPgCq7T4AhCQBAL4kAQCvwT4Art0+AK3hPgCs6T4AqNE+AKnRPgCq0T4Aq+U+AKzhPgCt4T4Arhk+AK8ZPgCGAAAAh4QAAHt4AIB/eACAg3gAgId4AICLeACAj3gAgLh9PgC5AT4AugE+ALsBPgC8AT4AvQk+AL4xPgC/MT4AsGk+ALF1PgCyfT4As3U+ALRZPgC1RT4Atk0+ALdFPgCohQIAqZUCAKqVAgCrpQIArL0CAK3VAgCu0QIAr9ECAJN4AICXeACAm3gAgL8k5gGfeACAo3gAgKd4AICreACAuFUDALlZAwC6bQMAu2UDALx9AwC9ZQMAvm0DAL9lAwCwtQIAsb0CALKBAgCzgQIAtHEDALVxAwC2cQMAt3EDALMdAgCveACAs3gAgLd4AICEiAMAtlUCALU1AgAndwCAu3kCALpxAgC7eACAv3gAgL+1AwC+tQMAvVUCALxVAgDDeACAo1kCAMd4AIDLeACAphECAM94AIDTeACApXECAKo1AgCrPQIA13gAgNt4AICu8QMAr/EDAKwRAgCtEQIAqKkCAKmpAgCquQIAq7kCAKypAgCtqQIArjkBAK85AQCAzQEAgQkAAIIZAADfeACA43gAgL64BQDreACA73gAgLjpAQC56QEAuokBALuFAQC8nQEAvYEBAL6BAQC/tQEAsEkBALFVAQCyXQEAs1UBALRNAQC18QEAtvEBALfxAQDvFAAA83gAgIaoBQCH3AUA93gAgIRYBAD7eACA78Q+AP94AIDhxD4AA3kAgOMwPgDjyAAAB3kAgOEoAQALeQCAtn0CAA95AIATeQCAtXUCABd5AICzZQIAG3kAgB95AIC+3QEAv2EBALzdAQC91QEAutkBALvFAQAjeQCAJ3kAgKOxBQDneACAK3kAgC95AIAzeQCApqkFAKWhBQA3eQCAqxEGAKoNBgA7eQCAP3kAgK+1BgCuCQYArQEGAKwJBgBDeQCAR3kAgEt5AIBPeQCAgBkAAIEZAACCBQAAU3kAgL5sAwBXeQCAhsgAAIccAwBbeQCAX3kAgGN5AIBneQCAqLkHAKm5BwCqDQcAqx0HAKwJBwCtNQcArjEHAK8pBwCEqAMAa3kAgG95AIBzeQCAd3kAgHt5AIB/eQCAg3kAgLjJAAC5yQAAutkAALvRAAC8+QAAvfkAAL6ZAAC/mQAAsF0HALEhBwCyIQcAsz0HALQpBwC1KQcAtgEHALcBBwCzhQYAh3kAgIt5AICPeQCAk3kAgLa1BgC1gQYAl3kAgLvlBgC6mQYAm3kAgJ95AIC/7QYAvu0GAL3pBgC89QYAo3kAgKd5AICreQCAr3kAgLN5AIC3eQCAu3kAgO+QBAC/eQCA4dwGAMN5AIDj7AUAgCkAAIEVAACCEQAAvnwBAKMFBgDLeQCAhigAAIdMAQDPeQCApjUGAKUBBgDTeQCAq2UGAKoZBgDXeQCA23kAgK9tBgCubQYArWkGAKx1BgDfeQCAs70BAON5AIDneQCAtnkBAOt5AIDveQCAtXkBALpVAQC7XQEA83kAgPd5AIC++QAAv/kAALxFAQC9+QAAqHECAKlxAgCqcQIAq3ECAKy1AgCtvQIArrUCAK+tAgCE7AwA+3kAgP95AIADegCAB3oAgAt6AIAPegCAE3oAgLhpAwC5aQMAugkDALsJAwC8GQMAvRkDAL4JAwC/CQMAsNUCALHdAgCy1QIAs2kDALR5AwC1eQMAtmkDALdhAwAXegCAG3oAgB96AICj9QIAI3oAgKUxAgCmMQIAJ3oAgCt6AIAvegCAqh0CAKsVAgCsDQIArbEDAK6xAwCvsQMAgGEAAIFhAACCBQAAM3oAgIbwDACHYAMAvhAMADt6AIB3eACAP3oAgEN6AIBHegCAS3oAgE96AIBTegCAV3oAgKiFAgCplQIAqpUCAKulAgCsvQIArdUCAK7RAgCv0QIAW3oAgF96AIBjegCAZ3oAgGt6AIBvegCAc3oAgHd6AIC4dQEAuX0BALp1AQC7zQEAvNUBAL3dAQC+yQEAv8EBALC1AgCxvQIAsoECALOBAgC0VQEAtV0BALZVAQC3TQEA4RAGAIRIDADjDAYAe3oAgISYDAB/egCAg3oAgId6AICLegCAj3oAgJN6AICXegCAgXUAAIB1AADvIAEAgnUAAJt6AICfegCAo3oAgL7ADACFtA4A4RACAO9cAADjABYA4ZABAKt6AIDjWAEA7zwHAK96AICzegCAhgAIAIe4DACznQ0AN3oAgLd6AIC7egCAv3oAgLbVDQC1tQ0Aw3oAgLv5DQC68Q0Ax3oAgMt6AIC/GQ4AvhEOAL3VDQC81Q0Az3oAgKPZDQDTegCA13oAgKaRDQDbegCA33oAgKXxDQCqtQ0Aq70NAON6AIDnegCArlUOAK9dDgCskQ0ArZENAKhdDgCpYQ4AqmEOAKthDgCsYQ4ArWEOAK5hDgCvYQ4A63oAgO96AIDzegCA93oAgPt6AID/egCAA3sAgAd7AIC4TQ8AuVEPALpRDwC7UQ8AvHEPAL1xDwC+cQ8Av3EPALDBDwCxwQ8AssEPALPBDwC0wQ8AtcEPALbBDwC3wQ8As+kPAAt7AIC+gAEAD3sAgKd6AIC24Q8AtekPABN7AIC7BQ4AugUOABt7AIAXewCAvwUOAL4FDgC9FQ4AvBUOAIFNAACAQQAA72gNAIJRAACG8AcAh9QBAB97AIAjewCAJ3sAgIRwAQArewCAL3sAgOHgDgAzewCA40gNADd7AICjaQ8AO3sAgD97AIBDewCAR3sAgKZhDwClaQ8AS3sAgKuFDgCqhQ4AT3sAgFN7AICvhQ4AroUOAK2VDgCslQ4AV3sAgLMxDgBbewCAX3sAgLbBAQBjewCAZ3sAgLXRAQC6zQEAu6UBAGt7AIBvewCAvqUBAL+tAQC8sQEAvbEBAI/dJgCj8Q0Ac3sAgHd7AICmAQIAe3sAgH97AIClEQIAqg0CAKtlAgCDewCAviAEAK5lAgCvbQIArHECAK1xAgCfoQwAnnkKAJ1pCgCc0QgAm7E2AJp1NgCZ0TQAmOEyAJdtMgCWZTIAlTU/AJRhPgCTcT4AkjU7AJFxOgCQeToAgJUAAIGdAACCoQAAi3sAgO9EAgDhdA8Aj3sAgOMcDwDj1AEAk3sAgOHgAQDvXAEAo7UCAKJBAACh3Q4AoLkOALWpAwCXewCAhMAEALahAwCG8AUAh+QEALOFAwCbewCAvXEDALxpAwC/QQMAvnEDAJ97AIDHeQCAu3EDALp5AwCC3ScAgwE7AL6EBwC+wAYAhhE/AIcZPwCEETsAhV06AIp9PgCLJTMAo3sAgKd7AICOuTUAjxU3AIw1MwCNgTMAkqE3AJPZCQC+xBkAq3sAgJaxDQCXUQ8AlHkLAJVhCwCaBQ8Am5EBAK97AICzewCAt3sAgN0AAACcfQMAu3sAgOFIDwC/ewCA4xwOAMN7AIDHewCAy3sAgM97AIDTewCAsUEXALChFwCzqesBsgHoAbUB7AG0EesB74wOANd7AICpxR8AqAEcAKsBEACqkR8ArdkTAKzREwCv2RcArgUTAKHxAgDbewCAo8kHAKLBAgClARgApGUHAKehGwCm+RsAqCkFAKldBQCqVQUAq20FAKx5BQCteQUArm0FAK9hBQCHewCA33sAgON7AIDnewCAgA0AAIGxAACCsQAA63sAgLiJBQC5iQUAup0FALuVBQC8uQUAvbkFAL5RBgC/UQYAsOUFALHtBQCy5QUAs/0FALTtBQC13QUAttUFALe9BQCj3QUA73sAgPN7AICEDAAA93sAgKb5BQCl8QUA+3sAgKspBQCqIQUAhpgAAIegAACvGQUArikFAK0pBQCsMQUA/3sAgLNhBgADfACAB3wAgLYhBgALfACAD3wAgLUBBgC6rQcAu40HABN8AIAXfACAvo0HAL9xBwC8lQcAvY0HAL65BQC/uQUAvLkFAL25BQC6uQUAu7kFALi5BQC5uQUAtkkFALdJBQC0fQUAtXUFALJ5BQCzeQUAsBUFALF9BQCuXQUAr20FAKxFBQCtXQUAqqUKAKtdBQCovQoAqa0KABt8AIAffACAI3wAgCd8AIArfACAL3wAgDN8AIA3fACAqA0HAKkdBwCqLQcAq0kHAKxNBwCtZQcArrEGAK+xBgA7fACAP3wAgEN8AIBHfACAS3wAgE98AIBTfACAV3wAgLhVBgC5XQYAulUGALtxBgC8NQYAvfEBAL7xAQC/8QEAsK0GALGNBgCyhQYAs50GALSNBgC1cQYAtnUGALdtBgCjpQQAgi0AAIEVAACAHQAAW3wAgKblBAClxQQAX3wAgKtJBQCqaQUAY3wAgGt8AICvtQUArkkFAK1JBQCsUQUAhmAcAIcIAwBvfACAs4UCAHN8AIC1gQIAtoECAHd8AIB7fACAf3wAgLoJAwC7CQMAvBkDAL0ZAwC+CQMAvwkDAKxVAgCtXQIArmECAK9hAgCoDQIAqVUCAKpRAgCrUQIAhKwDAIN8AICHfACAi3wAgIT8HQCPfACAk3wAgJd8AIC8cQMAvXEDAL5xAwC/cQMAuHEDALlxAwC6cQMAu3EDALSRAwC1kQMAtpEDALeRAwCwkQMAsZEDALKRAwCzkQMAm3wAgJ98AICjfACAp3wAgKt8AIDhpAEAr3wAgOOAAQC+aBwAs3wAgLd8AIDv2AYAu3wAgL98AIDDfACAx3wAgKOJAwCCLQAAgRUAAIAdAADLfACApo0DAKWNAwDPfACAqwUCAKoFAgDTfACA23wAgK8FAgCuBQIArRUCAKwVAgCGIBwAh8QdAN98AIDjfACA53wAgOt8AIDvfACA72wGAPN8AIDhbAcA93wAgON0BwD7fACA/3wAgAN9AIAHfQCAs5EBAAt9AIAPfQCAE30AgBd9AIC2sQEAtbkBABt9AIC7VQEAukkBAB99AIAjfQCAv/UAAL71AAC9RQEAvEUBAKNRHgDXfACAJ30AgCt9AIAvfQCApnEeAKV5HgAzfQCAq5UeAKqJHgA3fQCAO30AgK81HwCuNR8ArYUeAKyFHgCAbQAAgRUAAIIdAADv/BkAP30AgEN9AIBHfQCAS30AgIbAAACHrAMAT30AgFN9AIBXfQCA4SwcAFt9AIDjzBwAqK0eAKnNHgCq2R4Aq9EeAKzxHgCt8R4Arj0eAK81HgCE7AAAX30AgGN9AIBnfQCAa30AgG99AIBzfQCAd30AgLjRHwC53R8Auu0fALvlHwC84R8AveEfAL7hHwC/4R8AsE0eALFRHgCyUR4As1EeALTxHwC18R8AtvEfALfxHwCobR4AqY0eAKqFHgCrnR4ArIUeAK2NHgCuuR4Ar7UeAHt9AIB/fQCAg30AgId9AICLfQCAj30AgJN9AICXfQCAuJ0eALmtHgC6pR4Au0UBALxdAQC9RQEAvkUBAL91AQCw0R4AsdEeALLRHgCz0R4AtLUeALW9HgC2tR4At60eALMNHgCbfQCAn30AgKN9AICnfQCAtg0eALUNHgCrfQCAuxUeALoVHgCvfQCAs30AgL95HgC+cR4AvQUeALwFHgCCbQAAo0keAIBVAACBZQAApkkeAL6cAQC7fQCApUkeAKpRHgCrUR4Ah3wAAIZMAACuNR4Arz0eAKxBHgCtQR4AqF0CAKltAgCqZQIAq30CAKxpAgCtsQIArrECAK+xAgCE7AQAv30AgMN9AIDHfQCAy30AgM99AIDTfQCA130AgLhxAwC5cQMAunEDALtxAwC81QMAvd0DAL7VAwC/zQMAsNECALHRAgCy0QIAs9ECALRRAwC1UQMAtlEDALdRAwCz7QIA230AgN99AIC+gAQA430AgLYxAgC14QIA530AgLsVAgC6FQIA630AgO99AIC/lQMAvpUDAL0FAgC8BQIA830AgKOpAgD3fQCA+30AgKZ1AgD/fQCAA34AgKWlAgCqUQIAq1ECAAd+AIALfgCArtEDAK/RAwCsQQIArUECAKjZAgCpIQEAqiEBAKshAQCsIQEArSEBAK4hAQCvIQEAD34AgBN+AIAXfgCAviAEABt+AIAffgCAI34AgCt+AIC4jQEAuZEBALqRAQC7pQEAvL0BAL11AAC+fQAAv3UAALDlAQCx7QEAsvkBALPxAQC02QEAtdkBALa5AQC3tQEA4RgeAC9+AIDjKB8AM34AgIGlAACApQAAN34AgIKlAACGAAQAh/QFADt+AIA/fgCAQ34AgEd+AIDvYB4AS34AgE9+AIBTfgCAhfD0AVd+AIBbfgCA42QBAF9+AIDhpAEAY34AgO/IAABnfgCAa34AgGd8AICE/AUAb34AgHN+AICzKQYAJ34AgHd+AIB7fgCAf34AgLYhBgC1KQYAg34AgLupBgC6oQYAh34AgIt+AIC/nQYAvp0GAL2lBgC8rQYA4bQHAI9+AIDjeAQAk34AgIB9AACBEQAAghUAAJd+AICGwAAAh1gDAJt+AICffgCAo34AgKd+AIDvDAQAq34AgKOpBgCvfgCAs34AgLd+AIC7fgCApqEGAKWpBgC/fgCAqykGAKohBgDDfgCAx34AgK8dBgCuHQYArSUGAKwtBgDLfgCAs0kHAM9+AIDTfgCAtn0HANd+AIDbfgCAtXUHALpdBwC7JQcA334AgON+AIC+IQcAvy0HALw9BwC9MQcAqD0GAKmBBgCqhQYAq5UGAKy5BgCtuQYArqkGAK+pBgDnfgCA634AgO9+AIDzfgCA934AgIK5AACBsQAAgLkAALitBgC5vQYAurUGALtFAQC8XQEAvUUBAL5FAQC/dQEAsN0GALGlBgCyrQYAs6EGALShBgC1rQYAtpkGALeVBgCjDQYA+34AgP9+AIADfwCAhJgCAKY5BgClMQYAvpwBAKthBgCqGQYAhggAAId8AQCvaQYArmUGAK11BgCseQYAC38AgLO1AQAPfwCAE38AgLZVAQAXfwCAG38AgLWhAQC6cQEAu3kBAB9/AIAjfwCAvjEBAL89AQC8UQEAvVEBAKhpAgCpaQIAqnkCAKt5AgCsbQIArZECAK6RAgCvkQIAJ38AgCt/AIAvfwCAM38AgDd/AIA7fwCAP38AgEN/AIC4mQIAua0CALqlAgC7bQMAvHUDAL19AwC+dQMAv20DALDxAgCx+QIAssECALPBAgC0sQIAtb0CALa1AgC3qQIAR38AgEt/AIBPfwCAo/0CAFN/AICl6QIAph0CAFd/AIBbfwCAX38AgKo5AgCrMQIArBkCAK0ZAgCueQIAr3UCAGN/AIBnfwCAa38AgIQADACAGQAAgQkAAII5AABvfwCAc38AgHt/AIB/fwCAvuAMAIN/AICHfwCAhlgNAIcMAwCowQIAqc0CAKrFAgCr2QIArMkCAK39AgCu9QIArz0BAIt/AICPfwCAk38AgJd/AICbfwCAn38AgKN/AIC+MAwAuMUBALnNAQC62QEAu9EBALzxAQC98QEAvpkBAL+ZAQCwRQEAsU0BALJFAQCzXQEAtEUBALVNAQC2RQEAt/0BAOE4BgCnfwCA42wGAKt/AICvfwCAs38AgLd/AIC7fwCAhKgNAL9/AIDDfwCAx38AgL6wDwDLfwCA72wGAM9/AIDTfwCAt30AgNd/AIDbfwCA41AAAN9/AIDhoAEA438AgO+EAADrfwCAhyANAIZMDwCAPQAAgSEAAIIlAADvfwCAs80NAHd/AIDnfwCA838AgPd/AIC2/Q0AtcENAPt/AIC7CQ4AugEOAP9/AIADgACAvwkOAL4BDgC9CQ4AvBEOAAeAAIDjmAwAC4AAgOH8DwAPgACAE4AAgBeAAIAbgACAH4AAgCOAAIAngACAK4AAgC+AAIDvYAwAM4AAgDeAAICjTQ0AO4AAgD+AAIBDgACAR4AAgKZ9DQClQQ0AS4AAgKuJDgCqgQ4AT4AAgFOAAICviQ4AroEOAK2JDgCskQ4Agm0AALM1DgCAVQAAgWUAALb1DwCE3AMAV4AAgLX9DwC60Q8Au9EPAIYABACH3AAAvn0PAL9lDwC8wQ8AvXkPAKjlDwCp7Q8AqvkPAKv5DwCsMQ4ArTEOAK4xDgCvMQ4AW4AAgF+AAIBjgACAZ4AAgGuAAIBvgACAc4AAgHeAAIC43Q4AueEOALrhDgC74Q4AvOUOAL3pDgC+mQ4Av5UOALBRDgCxUQ4AslEOALPpDgC0/Q4AteUOALbtDgC35Q4Ao3EPAHuAAIB/gACAg4AAgIeAAICmsQ4ApbkOAIuAAICrlQ4AqpUOAI+AAICTgACAryEOAK45DgCtPQ4ArIUOAJeAAICzyQEAm4AAgJ+AAIC2+QEAo4AAgKeAAIC1wQEAuqkBALu1AQCrgACAr4AAgL6tAQC/lQEAvK0BAL2lAQCo5Q0AqfkNAKoFAgCrHQIArA0CAK09AgCuNQIAr10CALOAAIC3gACAu4AAgL+AAICAGQAAgRkAAIIFAADDgACAuC0CALk1AgC6MQIAuzECALzVAgC93QIAvtUCAL/NAgCwKQIAsTUCALI9AgCzNQIAtC0CALUVAgC2HQIAtxUCAMuAAICEnAIAz4AAgKOBAgDTgACApYkCAKaxAgDXgACAhiAEAIfUAwCq4QIAq/0CAKzlAgCt7QIAruUCAK/dAgC29QMAvkQDAIWM/QG1/QMA24AAgLP9AwDfgACA44AAgL59AwC/TQMAvGUDAL19AwC6dQMAu30DAOeAAIDrgACA74AAgPOAAICEBAIAoyUCAPeAAIClJQIApi0CAPuAAID/gACAA4EAgKqtAgCrpQIArL0CAK2lAgCupQIAr5UCAAeBAIALgQCAD4EAgBOBAIAXgQCA48ADABuBAIDhrAEAH4EAgO9YAwAjgQCAJ4EAgIANAACB5QAAgu0AACuBAIDhYA8A40ABAOM4DgDheA4AL4EAgDOBAIC+lAUAO4EAgIYABACHZAUAP4EAgEOBAIBHgQCA7/wOAO98DgBLgQCAs1EBAE+BAIAHfwCAU4EAgFeBAIC2DQEAtQkBAFuBAIC74QAAuhkBAF+BAIBjgQCAv9EAAL7pAAC96QAAvPkAAMeAAIA3gQCAZ4EAgGuBAIBvgQCAc4EAgHeBAIB7gQCAqKEGAKmtBgCquQYAq7EGAKzhBgCt7QYAruUGAK/FBgCwvQYAsUUHALJNBwCzXQcAtE0HALV1BwC2fQcAtx0HALglBwC5LQcAuiUHALs9BwC8KQcAvRUHAL4RBwC/EQcAoxEGAH+BAICDgQCAh4EAgIuBAICmTQYApUkGAI+BAICroQcAqlkGAJOBAICXgQCAr5EHAK6pBwCtqQcArLkHAIANAACBFQAAgh0AAJuBAICfgQCAo4EAgISUAwC+lAMAp4EAgKuBAICGyAAAh4wAAK+BAICzgQCAt4EAgLuBAIConQYAqa0GAKqlBgCrvQYArK0GAK3RBgCu1QYAr80GAL+BAIDDgQCAx4EAgMuBAIDPgQCA04EAgNeBAIDbgQCAuF0BALnBAQC6wQEAu8EBALzBAQC9yQEAvvEBAL/xAQCwvQYAsY0GALKFBgCzZQEAtH0BALVlAQC2bQEAt2UBALMtBgDfgQCA44EAgOeBAIDrgQCAtlEGALUlBgDvgQCAu0kGALp5BgDzgQCA94EAgL+hAQC+uQEAvbEBALxRBgD7gQCAo2kGAP+BAIADggCAphUGAAeCAIALggCApWEGAKo9BgCrDQYAD4IAgBOCAICu/QEAr+UBAKwVBgCt9QEAutUHALvdBwC4wQcAucEHAL4xBAC/MQQAvPEHAL3xBwCyrQcAs7UHALCtBwCxpQcAtp0HALf1BwC0pQcAtZUHAKppBwCraQcAqGkHAKlpBwCuaQcAr2kHAKxpBwCtaQcAgLkDAIGNAwCChQMAhKgDAIZQ/AGHCAMAvjQDABuCAICoZQIAqXUCAKp9AgCrdQIArG0CAK21AwCuvQMAr7UDAB+CAIAjggCAJ4IAgCuCAIAvggCAM4IAgDeCAIA7ggCAuFEDALlZAwC6YQMAu2EDALwRAwC9HQMAvhUDAL8JAwCwzQMAsdUDALLdAwCz1QMAtM0DALVxAwC2cQMAt3EDAD+CAIBDggCAs/0DAEeCAIC17QMAS4IAgE+CAIC2PQIAU4IAgFeCAIC7GQIAugECAL0JAgC8AQIAv70CAL4BAgBbggCAX4IAgITE/QG+wPwBY4IAgGeCAIBrggCA79wDAG+CAIDhlAEAc4IAgOMQAwB3ggCAgu0AAIHtAACA7QAA4TgGAOE8BwDjQAEA45QGAHuCAIB/ggCAg4IAgIuCAICGgPwBh+j9AY+CAICTggCAl4IAgJuCAIDvnAEA79wGAKM1AwCfggCAo4IAgKeCAICrggCApvUCAKUlAwCvggCAq9ECAKrJAgCzggCAt4IAgK91AgCuyQIArcECAKzJAgCHggCAu4IAgL+CAIDDggCA76T9AceCAIDLggCAz4IAgON4/QHTggCA4UD8AdeCAIDbggCA34IAgOOCAIDnggCAs+X+AYItAACBFQAAgB0AAOuCAIC25f4BtfX+Ae+CAIC7Yf8Butn+AfOCAICE5AMAv2n/Ab5h/wG9df8BvHn/Aaj9/gGpJf4Bqi3+Aasl/gGsPf4BrSX+Aa4t/gGvJf4BviwAAPeCAICGiAAAh+wAAPuCAID/ggCAA4MAgAeDAIC4gf8BuYH/AbqZ/wG7mf8BvIn/Ab21/wG+sf8Bv63/AbBd/gGx5f8Bsu3/AbPh/wG05f8Bte3/AbbZ/wG32f8Bo6X/AQuDAIAPgwCAE4MAgBeDAICmpf8BpbX/ARuDAICrIf4Bqpn/AR+DAIAjgwCAryn+Aa4h/gGtNf4BrDn+ASeDAICz6f4BK4MAgC+DAIC2lf4BM4MAgDeDAIC16f4BurH+Abu5/gE7gwCAP4MAgL51AQC/fQEAvJH+Ab2R/gGoHf4BqS3+Aaol/gGrPf4BrCX+Aa1R/gGuUf4Br1H+AUODAIBHgwCAS4MAgE+DAIBTgwCAV4MAgFuDAIBfgwCAuNkBALnZAQC67QEAu+EBALzhAQC94QEAvuEBAL/hAQCwMf4BsTn+AbIB/gGzAf4BtPUBALX9AQC29QEAt+kBAKOt/QFjgwCAvkwDAGuDAIBvgwCAptH9AaWt/QFzgwCAq/39Aar1/QF3gwCAe4MAgK85AgCuMQIArdX9AazV/QGA+QMAgfkDAIJNAACFdCAAf4MAgITYAwCE1AQAg4MAgIZABACHVAMAh4MAgIuDAICPgwCAk4MAgJeDAIC+8AUAqDECAKkxAgCqMQIAqzECAKyVAwCtnQMArpUDAK+NAwCbgwCAn4MAgKODAICngwCAhHwHAKuDAICvgwCAs4MAgLipAwC5qQMAumkDALtpAwC8eQMAvXkDAL5pAwC/aQMAsP0DALHNAwCyxQMAs60DALS5AwC1uQMAtq0DALelAwC3gwCAu4MAgL+DAIDDgwCAx4MAgMuDAIDv6AMAz4MAgOGQAQDTgwCA42wDANuDAICAJQAAgSkAAIIdAADfgwCAs/kDAOODAICGaAcAh1wFAOeDAIC2XQIAtV0CAOuDAIC7SQIAunkCAO+DAIDzgwCAvz0CAL49AgC9OQIAvFECAPeDAIDhPP4BvkAGAOPwAQD7gwCA/4MAgAOEAIAHhACAC4QAgA+EAIAThACAF4IAgBeEAIAbhACAH4QAgO/kAQAjhACAJ4QAgKNxAwArhACApdUCAC+EAIAzhACAptUCADeEAIA7hACAq8ECAKrxAgCtsQIArNkCAK+1AgCutQIA4dz8AdeDAIDjUAQA74gEAID1BwCBCQAAgj0AAD+EAICEJAEAQ4QAgEeEAIBLhACAT4QAgOFMBADv5BwA43QEALNdBgBThACAhgAMAIfgAwBXhACAtgUGALV1BgBbhACAuxEGALoJBgBfhACAY4QAgL/VBgC+1QYAvQEGALwJBgCojQYAqZUGAKqVBgCrpQYArL0GAK3FBgCuxQYAr/UGAGeEAIBrhACAb4QAgHOEAIB3hACAe4QAgH+EAICDhACAuHUGALl9BgC6dQYAu80HALzVBwC93QcAvtUHAL/NBwCwjQYAsZUGALKdBgCzlQYAtFEGALVRBgC2UQYAt1EGAKMdBwCPFewBh4QAgIuEAICPhACApkUHAKU1BwCThACAq1EHAKpJBwCXhACAm4QAgK+VBwCulQcArUEHAKxJBwCeRfkBn6X5AZyR/QGdTfkBmlX9AZtd/QGYBfEBmZX+AZal8gGXYfEBlG31AZU19QGS4ekBk4X2AZBV7AGRXekBsbEdALClHQCziRkAskEcALUBJAC09RkAn4QAgKOEAICnhACAgqkDAIGhAwCAaQAAohUFAKMFAgCgFQYAob0FAKHFAQCrhACAo80NAKLlAQClAQgApN0NAKfRCQCm2QkAqQEUAKilCACrxRQAqs0VAK3REQCsARAArwEcAK51EQCCEe8BgynvAa+EAICzhACAhuH1AYcR9gGEOeoBhY3qAYp59gGL4fEBvqQMALuEAICO+f0BjzH+AYw98gGNYfIBkkn+AZOd/gGHCAwAhmwMAJax+gGX+QUAlFn6AZVZ+gGaYQYAm8EGAL+EAIDDhACAx4QAgMuEAICcyQEAz4QAgKitBQCpuQUAqs0FAKvdBQCszQUArf0FAK71BQCvHQUA04QAgNeEAIDbhACA34QAgOOEAIDnhACA64QAgO+EAIC4dQUAuX0FALoJBQC7CQUAvB0FAL0BBQC+AQUAvz0FALBxBQCxcQUAsnEFALNxBQC0UQUAtVEFALZRBQC3TQUAs0UEAPOEAID3hACA+4QAgP+EAIC2fQQAtUUEAAOFAIC7tQQAurUEAAeFAIALhQCAv5UEAL6VBAC9pQQAvKUEAA+FAICjAQQAE4UAgBeFAICmOQQAG4UAgB+FAIClAQQAqvEEAKvxBAAjhQCAhOwNAK7RBACv0QQArOEEAK3hBADh0AYAhAwMAOMoBwC+AAwAK4UAgO9EAwCGuAwAhywNAC+FAIDjlAEAM4UAgOH8AQBngwCAN4UAgO/IBgA7hQCAP4UAgEOFAICzjQMAR4UAgLWNAwBLhQCAT4UAgLa1AwBThQCAV4UAgLtBAwC6SQMAvUEDALxZAwC/QQMAvkkDAKNFDAC3hACAJ4UAgFuFAIBfhQCApn0MAKVFDABjhQCAq4kMAKqBDABnhQCAa4UAgK+JDACugQwArYkMAKyRDACAFQ8AgR0PAIIhDwCzIQ4Ab4UAgLUhDgC2JQ4Ac4UAgHeFAIB7hQCAusEOALvBDgC8wQ4AvcEOAL7BDgC/wQ4AqK0OAKntDgCq5Q4Aq/0OAKzlDgCt6Q4ArjkOAK85DgB/hQCAg4UAgIeFAICLhQCAgB0AAIEJAACCvQEAj4UAgLjNDwC51Q8AutUPALvlDwC8/Q8AvZUPAL6RDwC/kQ8AsEkOALFJDgCyWQ4As1kOALRJDgC1SQ4Atv0PALf1DwCjbQ8Ak4UAgL6EAQCbhQCAn4UAgKZpDwClbQ8Ao4UAgKuNDwCqjQ8AhogAAIdsAQCvjQ8Aro0PAK2NDwCsjQ8Ap4UAgLPtDgCrhQCAr4UAgLaRDgCzhQCAt4UAgLXhDgC6tQ4Au70OALuFAIC/hQCAvn0BAL9lAQC8mQ4AvZkOAKgRDgCpJQ4AqiEOAKs5DgCsLQ4ArVUOAK5dDgCvUQ4AhKgAAMOFAIDHhQCAy4UAgM+FAIDThQCA14UAgNuFAIC47QEAuZUBALqVAQC7rQEAvLUBAL11AQC+fQEAv3UBALA1DgCxPQ4AsgkOALMJDgC0/QEAteUBALblAQC31QEAo6kNAN+FAIDjhQCA54UAgOuFAICm1Q0ApaUNAO+FAICr+Q0AqvENAPOFAID3hQCAryECAK45AgCt3Q0ArN0NAIANAACBFQAAgh0AAPuFAID/hQCAA4YAgIeQAwCGfAQAvuwEAAuGAIAPhgCAE4YAgBeGAIAbhgCAH4YAgCOGAICyLQ4AszUOALAtDgCxJQ4Ati0OALedDwC0LQ4AtSUOALq9DwC7jQ8AuKUPALm9DwC+LQ8AvxUPALyVDwC9JQ8AJ4YAgCuGAIAvhgCAM4YAgDeGAIA7hgCAP4YAgEOGAICqpQ4Aq7UOAKjFDgCp3Q4Arp0OAK9VDgCspQ4ArZUOAKgNAgCpFQIAqhUCAKtNAgCsWQIArVkCAK5NAgCvRQIAhKgFAEeGAIBLhgCAT4YAgIS4BABThgCAV4YAgFuGAIC4/QIAuUEBALpBAQC7QQEAvEEBAL1JAQC+cQEAv3EBALAJAgCxCQIAss0CALPFAgC03QIAtcUCALbNAgC3xQIA4dQPAOMQDgDj9A4A4QwOAF+GAIBjhgCAZ4YAgGuGAIBvhgCAc4YAgL4kBAB7hgCA7AAAAO9EAADvzA4Af4YAgIJlAACz2QIAgFUAAIFtAAC2nQIAg4YAgIeGAIC1lQIAuokCALuJAgCGqAQAh+AEAL5dAgC/RQIAvF0CAL1VAgCjHQUAB4YAgHeGAICLhgCAj4YAgKZZBQClUQUAk4YAgKtNBQCqTQUAl4YAgJuGAICvgQUArpkFAK2RBQCsmQUAn4YAgLMpBgCjhgCAp4YAgLYpBgCrhgCAr4YAgLUpBgC6pQYAu60GALOGAIC3hgCAvqUGAL+tBgC8tQYAva0GAKjlBgCp7QYAquUGAKv9BgCs5QYAre0GAK7lBgCvXQYAu4YAgL+GAIDDhgCAx4YAgMuGAIDPhgCA04YAgNeGAIC46QcAuekHALr9BwC79QcAvO0HAL1FBwC+TQcAv0UHALAlBgCxLQYAsiUGALM9BgC0JQYAtS0GALYlBgC32QcAo20HAIItAACBFQAAgB0AANuGAICmbQcApW0HAN+GAICr6QcAquEHAOOGAIC+oAEAr+kHAK7hBwCt6QcArPEHAOeGAICzkQYAhugAAIcsAQC2QQEA64YAgO+GAIC1UQEAuk0BALslAQDzhgCA94YAgL4lAQC/LQEAvDEBAL0xAQCwrQEAscUBALLBAQCzwQEAtMUBALXNAQC28QEAt/EBALgBAQC5AQEAugEBALsBAQC8AQEAvQEBAL4BAQC/AQEA+4YAgP+GAIADhwCAB4cAgJeFAIALhwCAD4cAgBOHAICoTQYAqVkGAKo9BgCrNQYArP0BAK3lAQCu5QEAr9UBAKPVBQAXhwCAG4cAgB+HAIAjhwCApgUCAKUVAgAnhwCAq2ECAKoJAgArhwCAL4cAgK9pAgCuYQIArXUCAKx1AgAzhwCAN4cAgDuHAIA/hwCAQ4cAgOFkBQBHhwCA4+wFAIARAACBEQAAghEAAO/0BgBLhwCAT4cAgFOHAIC+MAMAhMQCAFuHAICz4QMAhMAcALVRAwBfhwCAY4cAgLZZAwBnhwCAa4cAgLtxAwC6eQMAvbUAALxpAwC/tQAAvrUAAG+HAIDhlAEAc4cAgONcAgCGcBwAh0QDAHeHAIB7hwCAf4cAgIOHAICHhwCAi4cAgI+HAICThwCAl4cAgO94AgCoVQIAqV0CAKphAgCrYQIArNECAK3RAgCu0QIAr9ECAJuHAICfhwCAo4cAgKeHAICrhwCAr4cAgLOHAIC3hwCAuGkBALlpAQC6CQEAuwkBALwZAQC9GQEAvgkBAL8FAQCwtQIAsb0CALK1AgCzaQEAtHkBALV5AQC2aQEAt2EBAOHEBwDjpAYA47gGAOF8BgCADQAAgTUAAII9AAC7hwCAv4cAgMOHAIC+4B0Ay4cAgM+HAIDvYAAA7+gGANOHAICjqQIA14cAgNuHAIDfhwCA44cAgKYRAgClGQIA54cAgKs5AgCqMQIAhkgcAIfMHACv/QEArv0BAK39AQCsIQIAqIUeAKmRHgCqkR4Aq60eAKy1HgCt1R4ArtEeAK/FHgDHhwCA64cAgO+HAIDzhwCA94cAgPuHAID/hwCAA4gAgLhhHwC5YR8AumEfALthHwC8YR8AvWEfAL5hHwC/YR8AsL0eALGFHgCyjR4As4UeALSdHgC1hR4Ato0eALeFHgCzGR4AB4gAgAuIAIAPiACAE4gAgLZVHgC1PR4AF4gAgLtBHgC6eR4AG4gAgB+IAIC/QR4AvlkeAL1RHgC8WR4AI4gAgKNdHgAniACAK4gAgKYRHgAviACAM4gAgKV5HgCqPR4AqwUeAISkAwC+qAMArh0eAK8FHgCsHR4ArRUeAKitHgCptR4AqrUeAKvJHgCs2R4ArdkeAK7JHgCvwR4AgO0BAIHxAQCC8QEAN4gAgIaQAACHdAEAO4gAgD+IAIC4yQEAuckBALrZAQC70QEAvPkBAL35AQC+mQEAv5UBALBFAQCxTQEAskUBALNdAQC0RQEAtU0BALZFAQC3+QEAsz0eAEOIAIBHiACAS4gAgE+IAIC2WR4AtVEeAFOIAIC7iQEAuoEBAFeIAIBbiACAv4kBAL6BAQC9iQEAvJEBAF+IAIBjiACAo3UeAGeIAIClGR4Aa4gAgG+IAICmER4AV4cAgHOIAICrwQEAqskBAK3BAQCs2QEAr8EBAK7JAQB3iACAe4gAgH+IAICDiACAh4gAgIQYAgCLiACAj4gAgJOIAICXiACAm4gAgJ+IAICjiACAq4gAgK+IAIC+cAMAgGkAAIFpAACCeQAAhAAEAIbwBACHdAMAs4gAgO8MHwC3iACA4aweALuIAIDj8B4Av4gAgMOIAIDHiACAy4gAgM+IAIDTiACA14gAgNuIAIDvVAIA34gAgOOIAIDniACA46QCAOuIAIDhgAEA74gAgPOIAID3iACA+4gAgP+IAICzRQMAA4kAgAeJAIALiQCAD4kAgLZFAwC1VQMAE4kAgLshAwC6SQMAvqAEABuJAIC/KQMAviEDAL01AwC8OQMAqDkCAKk5AgCqjQIAq4UCAKydAgCthQIAroUCAK+1AgCA7QEAgfUBAIL1AQAfiQCAhpAEAIcEBQAjiQCAJ4kAgLhFAQC5TQEAukUBALtdAQC8SQEAvUkBAL55AQC/eQEAsM0CALGlAgCyrQIAs6ECALSlAgC1rQIAtp0CALd9AQAriQCAL4kAgDOJAIA3iQCAO4kAgD+JAIBDiQCA74gBAITsBADhVB4AR4kAgONUAQBLiQCAT4kAgFOJAIBXiQCAo0UCAFuJAIBfiQCAY4kAgGeJAICmRQIApVUCAGuJAICrIQIAqkkCAG+JAIBziQCArykCAK4hAgCtNQIArDkCAKg1BgCpPQYAqlEGAKttBgCseQYArWUGAK5tBgCvZQYAF4kAgHeJAIB7iQCAf4kAgIAZAACBGQAAggUAAIOJAIC45QYAuekGALr5BgC7+QYAvOkGAL3pBgC+nQYAv5UGALAdBgCx5QYAsu0GALPlBgC0/QYAteEGALbhBgC34QYAs9kGAL7QAwCHiQCAi4kAgI+JAIC25QYAtfEGAJOJAIC7IQYAutkGAIaYAACHeAMAvyUGAL45BgC9MQYAvDkGAJeJAICjnQYAm4kAgJ+JAICmoQYAo4kAgKeJAICltQYAqp0GAKtlBgCriQCAr4kAgK59BgCvYQYArH0GAK11BgCo7QcAqSkGAKoxBgCrMQYArJEGAK2RBgCukQYAr5EGALOJAIC3iQCAu4kAgL+JAIDDiQCAx4kAgMuJAIDPiQCAuIUGALmNBgC6hQYAu50GALyNBgC9vQYAvrUGAL95AQCw8QYAsfEGALLxBgCzxQYAtMEGALXBBgC2wQYAt8EGALO5BgDTiQCA14kAgNuJAIDfiQCAthEGALUZBgDjiQCAuzUGALo1BgDniQCA64kAgL8FBgC+BQYAvREGALwlBgClQQYA74kAgPOJAICmSQYAgRUAAIB5AACj4QYAghUAAK1JBgCsfQYAr10GAK5dBgCENAEAp4gAgKttBgCqbQYAvswDAPuJAICzlQIA/4kAgLXZAgADigCAB4oAgLbRAgCGgAwAhzgDALvFAgC6xQIAvRUDALwVAwC/FQMAvhUDAAuKAIAPigCA71gGAIRAAwATigCAF4oAgBuKAIAfigCAI4oAgCeKAIArigCAL4oAgOE4BgAzigCA4yQGAL5wDACsSQIArUkCAK5dAgCvVQIAqB0CAKkFAgCqBQIAq10CAISoDAA3igCAO4oAgD+KAIC+vA0AQ4oAgEeKAIBLigCAvE0DAL1VAwC+VQMAv2UDALjpAwC56QMAul0DALtVAwC0yQMAtckDALbZAwC32QMAsBkCALEZAgCy2QMAs9kDAE+KAIDj5AAAU4oAgOG8AQBXigCAgj0AAIE9AACAPQAAW4oAgF+KAIBjigCAa4oAgG+KAIDvzAMAc4oAgHeKAICj3QMAe4oAgIboDACHYA0Af4oAgKaZAwClkQMAg4oAgKuNAwCqjQMAh4oAgIuKAICvXQIArl0CAK1dAgCsXQIAj4oAgJOKAICXigCAm4oAgJ+KAICjigCAp4oAgO/gAQCEvAwA4YwGAKuKAIDjHAYAr4oAgLOKAIC3igCAu4oAgLPVAQC/igCAw4oAgMeKAIDLigCAtpEBALWZAQDPigCAu70BALq9AQDTigCA24oAgL+dAQC+nQEAvZ0BALydAQCoBQ4AqQkOAKodDgCrFQ4ArFEOAK1RDgCuSQ4Ar0kOAGeKAICCzQ8AgfUPAID9DwDXigCA34oAgIYcAACHsAMAuOkOALnpDgC6/Q4Au/UOALztDgC9VQ8AvlEPAL9NDwCwOQ4AsTkOALIJDgCzCQ4AtBkOALUZDgC2DQ4At9kOAKOVDgDjigCA54oAgOuKAIDvigCAptEOAKXZDgDzigCAq/0OAKr9DgD3igCA+4oAgK/dDgCu3Q4Ard0OAKzdDgD/igCAs/0PAAOLAIAHiwCAtoEPAAuLAIAPiwCAtZkPALqNDwC7ZQ8AE4sAgBeLAIC+fQ8Av2UPALx9DwC9dQ8AqC0OAKk1DgCqMQ4AqzEOAKxVDgCtRQ4ArkUOAK91DgAbiwCAH4sAgCOLAIAniwCAK4sAgC+LAIAziwCAN4sAgLjpDgC59Q4Auv0OALv1DgC87Q4AvZEOAL6RDgC/kQ4AsA0OALHlDgCy7Q4As+UOALT9DgC15Q4Atu0OALflDgCjuQ4Agi0AAIEVAACAHQAAO4sAgKbFDgCl3Q4AP4sAgKshDgCqyQ4AQ4sAgL4sAQCvIQ4ArjkOAK0xDgCsOQ4AS4sAgLZVAQC1RQEAR4sAgLNVAQBPiwCAhngAAIdcAAC/OQEAvjEBAL0lAQC8JQEAuzEBALpZAQD3iQCAU4sAgFeLAIBbiwCAhAQDAKOJAgBfiwCApZkCAKaJAgBjiwCAvyg5AGeLAICqhQIAq+0CAKz5AgCt+QIAru0CAK/lAgDjWAIA78AOAOGIAQBriwCAb4sAgHOLAIB3iwCAe4sAgH+LAICDiwCAh4sAgIuLAIDvKAIA4ygOAI+LAIDhRA4AqbUCAKhpDQCrAQIAqgkCAK0BAgCsGQIArzECAK4BAgC+AAQAk4sAgJeLAICbiwCAn4sAgKOLAICniwCAq4sAgLnlAwC45QMAu+UDALrlAwC95QMAvOUDAL/lAwC+5QMAsSECALBJAgCzJQIAsiUCALUpAgC0IQIAtxUCALYVAgCowQIAqdECAKr1AgCrDQEArBUBAK0FAQCuBQEArzkBAK+LAICziwCAu4sAgL+LAIDDiwCAx4sAgMuLAIDPiwCAuC0BALk9AQC67QEAu+UBALz9AQC95QEAvu0BAL/lAQCwLQEAsTUBALI9AQCzNQEAtC0BALUVAQC2HQEAtxUBAIA9AQCBpQAAgq0AAO/YAACGsAUAh9gFANOLAIDv1A8AhGwEAOH0DgDXiwCA4xwPANuLAIDhlAEA34sAgOMMDgCzPQIA44sAgOeLAIDriwCA74sAgLbFAQC13QEA84sAgLuxAQC6qQEA94sAgPuLAIC/kQEAvqkBAL2hAQC8qQEAt4sAgP+LAICqRQYAq10GAKxFBgCtTQYArkUGAK99BgADjACAB4wAgAuMAICj0QUAD4wAgKUxBgCmKQYAE4wAgBeMAICCHQAAgR0AAIAdAAAbjACAH4wAgCOMAIC+lAMAJ4wAgCuMAICGSAMAh8wDAC+MAIAzjACAN4wAgDuMAICoqQcAqakHAKq5BwCruQcArKkHAK2pBwCuAQcArzUHAD+MAIBDjACAR4wAgEuMAIBPjACAU4wAgFeMAIBbjACAuC0HALnBAAC66QAAu+kAALz5AAC95QAAvuUAAL+dAACwUQcAsV0HALItBwCzJQcAtD0HALUlBwC2JQcAtxUHALMxBgBfjACAY4wAgGeMAIBrjACAtikGALUhBgBvjACAu5kGALqVBgBzjACAd4wAgL/hBgC++QYAvfEGALz5BgB7jACAo3UGAH+MAICDjACApm0GAIeMAICLjACApWUGAKrRBgCr3QYAj4wAgJOMAICuvQYAr6UGAKy9BgCttQYAqOUBAKn1AQCq/QEAq/UBAKztAQCtNQEArj0BAK81AQCA+QAAgc0AAILFAACEYAEAvngBAJuMAICHrAAAhpABALjRAAC52QAAuuEAALvhAAC8kQAAvZ0AAL6VAAC/iQAAsE0BALFVAQCyXQEAs1UBALRNAQC18QAAtvEAALfxAACzdQIAn4wAgKOMAICnjACAq4wAgLa1AgC1ZQIAr4wAgLuRAgC6iQIAs4wAgLeMAIC/NQMAvokCAL2BAgC8iQIAu4wAgKMxAgC/jACAhMADAKbxAgDDjACAx4wAgKUhAgCqzQIAq9UCAMuMAIDPjACArs0CAK9xAwCszQIArcUCAKuNAACqjQAAqY0AAKg5AwCvvQAArr0AAK2FAACsjQAAqgAAAKsAAADTjACA14wAgNuMAIDfjACA44wAgOeMAIC7fQAAun0AALl9AAC4fQAAv90BAL7dAQC93QEAvN0BALO5AACysQAAsaEAALCtAAC3XQAAtl0AALWVAAC0lQAA64wAgO+MAIDzjACA94wAgIE1AACADQAA+4wAgII1AAC+rD0A/4wAgAONAICFaD0AC40AgA+NAICGODwAh8ACALNJAQATjQCA0AAAABeNAIAbjQCAtkkBALVJAQAfjQCAuykBALolAQAjjQCAJ40AgL8dAQC+HQEAvSEBALwpAQDjNDYA4QwGAOGwAgDjPAYAK40AgC+NAIAzjQCAN40AgIQsPwC+oD8AO40AgD+NAIDvfDcAQ40AgEeNAIDvGAEAS40AgE+NAICGaD4Ah8w/AFONAIBXjQCAW40AgO+UAABfjQCA4ZQBAGONAIDjUAAAZ40AgILpPwCB6T8AgPE/AKMJPgCPASQAB40AgGuNAIBvjQCApgk+AKUJPgBzjQCAq2k+AKplPgB3jQCAe40AgK9dPgCuXT4ArWE+AKxpPgCeYTgAn3U4AJzBNACdtTkAmqU1AJt1NACYeTAAmXExAJYhLQCXhTEAlG0sAJVlLACSeSgAk6UtAJBRJACReSgAsQ0UALAFFACzARgAslUUALV5GAC0tRgAf40AgIONAICHjQCAi40AgI+NAICTjQCAotE8AKMlAQCgdTkAob08AKHJAACXjQCAowEEAKLlAAClHQQApPUEAKf5CACmAQgAqQEMAKhtCACrzQwAqs0MAK3REACsARAAr9URAK7ZEACCBSUAgy0lAJuNAICfjQCAhsEsAIcRLQCEHSkAhRUpAIopLQCLZSwAo40AgKeNAICOHTAAj8E0AIzZMACNHTEAkmE1AJPNNQCrjQCAr40AgJZhOQCXmTgAlKE4AJV9OQCaYT0AmwU9ALONAIC3jQCAu40AgL+NAICc6QAAw40AgMeNAIDLjQCAz40AgNONAICXjACA140AgNuNAIDfjQCAqJE+AKmRPgCq7T4Aq+E+AKzhPgCt6T4ArtE+AK/RPgCwUT4AsVE+ALJRPgCzUT4AtHk+ALV5PgC2bT4At2U+ALghPgC5IT4Aujk+ALs5PgC8KT4AvRU+AL4RPgC/DT4AgJkDAIGZAwCCBQAA440AgL5UAwDhsD0A640AgONAPgCEOAIA740AgPONAIDv9D8A940AgPuNAICGmAQAhxwDALMFPQCECAQA/40AgAOOAIAHjgCAtgk9ALUJPQALjgCAu/U9ALr1PQAPjgCAE44AgL/dPQC+3T0AveU9ALzlPQAXjgCAG44AgKPNPQC+xAQApcE9AB+OAIAjjgCApsE9ACeOAIArjgCAqz09AKo9PQCtLT0ArC09AK8VPQCuFT0AtmkCAC+OAIAzjgCAtWkCADeOAICzSQIAO44AgD+OAIC+qQMAv6kDALzBAwC9wQMAuvkDALv5AwBDjgCAR44AgKgtAwCpnQMAqpUDAKutAwCstQMArb0DAK61AwCv2QMAgA0AAIEVAACCHQAAS44AgE+OAIBTjgCAh7QFAIacBAC4MQIAuTECALo1AgC7zQIAvNUCAL3dAgC+1QIAv8kCALBpAgCxaQIAskECALNBAgC0OQIAtTkCALYRAgC3EQIAW44AgOM0PgBfjgCA4aw+AGOOAIDvfAMAZ44AgGuOAIBvjgCA45QDAHOOAIDhfD4Ad44AgO/oPgB7jgCAf44AgIOOAICHjgCAo1UDAIuOAICldQMAj44AgJOOAICmdQMAl44AgJuOAICr5QIAquUCAK3dAgCs3QIAr7UCAK61AgCoGQYAqSEGAKohBgCrPQYArCUGAK1dBgCuVQYAr00GAFeOAICfjgCAo44AgKeOAICrjgCAr44AgLOOAIC3jgCAuOUGALmBBgC6gQYAu50GALyJBgC9iQYAvqEGAL+hBgCwPQYAsQ0GALIFBgCz7QYAtPUGALXhBgC24QYAt90GALOpBgCCLQAAgRUAAIAdAAC7jgCAtt0GALWtBgC/jgCAu8kGALr5BgDDjgCAhOADAL8lBgC+MQYAvTkGALzRBgC+iAMAo+0GAOeNAIDHjgCAppkGAMuOAIDPjgCApekGAKq9BgCrjQYAhkgAAIdsAACudQYAr2EGAKyVBgCtfQYAqIEGAKmNBgCqmQYAq5UGAKyNBgCttQYArrEGAK+tBgDTjgCA144AgNuOAIDfjgCA444AgOeOAIDrjgCA744AgLilBgC5YQEAumEBALthAQC8YQEAvWEBAL5hAQC/YQEAsNkGALHZBgCyqQYAs6kGALS9BgC1oQYAtqEGALedBgCzEQYA844AgPeOAID7jgCA/44AgLY1BgC1BQYAA48AgLsdBgC6HQYAB48AgAuPAIC/ZQYAvnkGAL19BgC8fQYAD48AgKNVBgATjwCAF48AgKZxBgAbjwCAH48AgKVBBgCqWQYAq1kGACOPAIAnjwCArj0GAK8hBgCsOQYArTkGAKjVAgCp3QIAqikDAKspAwCsOQMArTkDAK4pAwCvKQMAK48AgC+PAIAzjwCAO48AgD+PAIBDjwCAvrgDAEePAIC47QMAuYUDALqBAwC7gQMAvIUDAL2NAwC+sQMAv7EDALBZAwCxWQMAsu0DALPlAwC0/QMAteUDALblAwC31QMAgKEAAIGhAACCoQAAvoAMAEuPAICEmAIAT48AgFOPAICGAAwAh/QDAFePAIBbjwCAX48AgGOPAIBnjwCAhLADALPhAwBrjwCAb48AgHOPAIB3jwCAtvkDALXxAwB7jwCAu90DALrdAwB/jwCAg48AgL9hAwC+eQMAvXEDALx5AwCHjwCAi48AgI+PAICjLQIAk48AgKU9AgCmNQIAl48AgJuPAICfjwCAqhECAKsRAgCstQIArb0CAK61AgCvrQIA48QDAOMQBwDhuAEA4WwHAIBxAACBcQAAggUAAKOPAICGwAwAh1QNAKuPAICvjwCA77ADAO8ABwCzjwCAt48AgLuPAIC/jwCAw48AgMePAIDLjwCAz48AgNOPAIDvpAEAhKANAOGABgDXjwCA4xABANuPAIDfjwCA448AgOePAICz9QEA648AgO+PAIDzjwCA948AgLZNAQC1SQEA+48AgLtRAQC6SQEA/48AgAOQAIC/OQEAvjEBAL1BAQC8SQEAqC0OAKk1DgCqPQ4AqzEOAKyBDgCtjQ4AroUOAK+1DgCnjwCAB5AAgAuQAIAPkACAgBkAAIEZAACCBQAAE5AAgLidDgC5rQ4AuqUOALtNDwC8VQ8AvV0PAL5JDwC/QQ8AsM0OALHVDgCy3Q4As9UOALS1DgC1vQ4AtrUOALetDgCjtQ4AvogDABeQAIAbkACAH5AAgKYNDgClCQ4AI5AAgKsRDgCqCQ4AhggAAIdsAwCveQ4ArnEOAK0BDgCsCQ4AJ5AAgCuQAIAvkACAs7UPADOQAIC1VQ8Atl0PADePAIA3kACAO5AAgLp5DwC7eQ8AvGkPAL1dDwC+SQ8Av0kPAKhpDgCpaQ4AqnEOAKtxDgCskQ4ArZEOAK6RDgCvkQ4AP5AAgEOQAIBHkACAS5AAgE+QAIBTkACAV5AAgFuQAIC4hQ4AuY0OALqFDgC7nQ4AvI0OAL29DgC+tQ4Av3kBALDxDgCx8Q4AsvEOALPFDgC0wQ4AtcEOALbBDgC3wQ4Ao/kOAF+QAIBjkACAZ5AAgGuQAICmEQ4ApRkOAG+QAICrNQ4AqjUOAHOQAIB3kACArwUOAK4FDgCtEQ4ArCUOAIANAACBFQAAgh0AAHuQAIB/kACAg5AAgISUAQC+lAEAhkAHAIf0AACLkACAj5AAgJOQAICXkACAm5AAgJ+QAICojQIAqZUCAKqVAgCrzQIArNUCAK3dAgCuyQIAr/0CAKOQAICnkACAq5AAgK+QAIC/ABQAs5AAgLeQAIC7kACAuH0DALnBAwC6wQMAu8EDALzBAwC9yQMAvvEDAL/xAwCwhQIAsUUDALJNAwCzRQMAtF0DALVFAwC2TQMAt0UDALMdAgC/kACAw5AAgMeQAIDLkACAtl0CALVdAgDPkACAu4EDALpBAgDTkACA15AAgL+BAwC+mQMAvZEDALyZAwDbkACAo1kCAN+QAIDjkACAphkCAOeQAIDrkACApRkCAKoFAgCrxQMA75AAgPOQAICu3QMAr8UDAKzdAwCt1QMA+5AAgOPMAACEBAIA4bwBAIDJAQCB/QEAgvUBAL4QBQD/kACAvigEAAORAIAHkQCAC5EAgO8QAAAPkQCAE5EAgIbgBACH9AIAF5EAgBuRAIDj/A8AH5EAgOHgDwAjkQCA7xQPACeRAIArkQCAL5EAgDORAIA3kQCAO5EAgD+RAIBDkQCAR5EAgEuRAIBPkQCAU5EAgFeRAIBbkQCA7+ABAIUEEgDh3A4AX5EAgOMcDgCAKQAAgR0AAIIFAABjkQCAszECAGuRAICEzAUAb5EAgHORAIC2KQIAtSECAHeRAIC7zQEAus0BAHuRAIB/kQCAv3UBAL7JAQC9wQEAvMkBAKjpBQCp6QUAqvkFAKv5BQCs6QUArekFAK45BgCvOQYA95AAgGeRAICGiAAAhwADAIORAICHkQCAi5EAgI+RAIC40QYAudkGALrhBgC74QYAvJEGAL2dBgC+lQYAv4kGALBJBgCxSQYAsl0GALNVBgC0TQYAtfEGALbxBgC38QYAo3EFAJORAICXkQCAm5EAgJ+RAICmaQUApWEFAKORAICrjQYAqo0GAKeRAICrkQCArzUGAK6JBgCtgQYArIkGAK+RAICzkQCAs+EHALeRAIC14QcAu5EAgL+RAIC25QcAh5AAgMORAIC7vQcAuqEHAL2VBwC8qQcAv5UHAL6VBwCoAQYAqSUGAKohBgCrIQYArCEGAK0tBgCuJQYAr1UGAMeRAICCHQAAgR0AAIAdAADLkQCAz5EAgNORAIC+MAEAuDkGALk5BgC6yQYAu8kGALzZBgC92QYAvskGAL/JBgCwLQYAsTEGALI1BgCzCQYAtBkGALUZBgC2CQYAtwkGAKOpBgCEjAIAhigfAIdEAQDbkQCApq0GAKWpBgDfkQCAq/UGAKrpBgDjkQCA55EAgK/dBgCu3QYArd0GAKzhBgDrkQCAsxUGAO+RAIDzkQCAtj0GAPeRAID7kQCAtTUGALrZAQC72QEA/5EAgAOSAIC+fQEAv2UBALx9AQC9dQEAqMUFAKnJBQCq2QUAq9EFAKz5BQCt+QUArikCAK8pAgAHkgCAC5IAgA+SAIATkgCAjAAAABeSAIAbkgCAH5IAgLjtAgC5hQIAuo0CALuBAgC8hQIAvY0CAL69AgC/fQMAsFkCALFZAgCy7QIAs+UCALT9AgC15QIAtuUCALfVAgCjUQUAI5IAgCeSAIArkgCAL5IAgKZ5BQClcQUAM5IAgKudAgCqnQIAN5IAgDuSAICvIQIArjkCAK0xAgCsOQIAghEAAD+SAICAZQAAgQkAAEOSAIC+mAMAS5IAgE+SAICEJAMAU5IAgIdoAwCGjBwAV5IAgFuSAIBfkgCAY5IAgGeSAIBrkgCAs6ECAITAHAC10QIAb5IAgHOSAIC21QIAd5IAgHuSAIC7wQIAuvUCAL0RAQC82QIAvxEBAL4ZAQB/kgCAg5IAgIeSAICLkgCAj5IAgJOSAICXkgCA77gGAJuSAIDhnAQAn5IAgON0BgCjkgCAp5IAgKuSAICvkgCAgPkAAIH5AACCBQAAs5IAgL5YHACEWB8A71wAAO9ABgDhkAEA4fwGAOM8AADjdAYAu5IAgL+SAICGmBwAh/QcAKNpAgC+DB8Aw5IAgMeSAIDLkgCAph0CAKUZAgDPkgCAqwkCAKo9AgDTkgCA15IAgK/ZAQCu0QEArdkBAKwRAgCokR0AqZkdAKqhHQCroR0ArNEdAK3dHQCu1R0Ar8kdAEeSAIC3kgCA25IAgN+SAIDjkgCA55IAgOuSAIDvkgCAuHkeALl5HgC6zR4Au8UeALzdHgC9xR4AvsUeAL/1HgCwuR0AsY0dALKFHQCzTR4AtFUeALVdHgC2VR4At0keALjNHwC51R8Aut0fALvVHwC88R8Avf0fAL7pHwC/6R8AsKUfALGxHwCysR8As40fALSVHwC19R8Atv0fALf1HwCoGR4AqRkeAKotHgCrPR4ArCUeAK0tHgCuJR4Ar90fAPOSAID3kgCA+5IAgP+SAIADkwCA15EAgAeTAIALkwCAs+UfAA+TAIATkwCAF5MAgBuTAIC27R8Ate0fAB+TAIC7NR4AuiEeACOTAIAnkwCAv3EeAL4RHgC9GR4AvCUeAIJpAACjoR8AgFkAAIFRAACmqR8AK5MAgC+TAIClqR8AqmUeAKtxHgCGAAQAh+wBAK5VHgCvNR4ArGEeAK1dHgCoMR4AqTEeAKpBHgCrQR4ArEEeAK1JHgCucR4Ar3EeADOTAIA3kwCAO5MAgD+TAIBDkwCAR5MAgEuTAIBPkwCAuCkBALkpAQC6OQEAuzUBALwtAQC90QAAvtEAAL/RAACwyQEAsckBALLZAQCz2QEAtMkBALXJAQC2GQEAtxkBALPJHQBTkwCAV5MAgFuTAIBfkwCAtskdALXJHQBjkwCAuw0CALoNAgBnkwCAa5MAgL8NAgC+DQIAvQ0CALwNAgBvkwCAo40dAHOTAIB3kwCApo0dAHuTAIB/kwCApY0dAKpJAgCrSQIAg5MAgIeTAICuSQIAr0kCAKxJAgCtSQIAgA0AAIERAACCEQAAi5MAgO/MAgCPkwCAk5MAgISQAgDjLAIAvigDAOHYAQCbkwCAhhAEAIfUAwCfkwCAo5MAgLNhAwCnkwCAq5MAgK+TAICzkwCAtnkDALVxAwC3kwCAu10DALpdAwC7kwCAv5MAgL/hAAC++QAAvfEAALz5AACjoQIAw5MAgMeTAIDLkwCAz5MAgKa5AgClsQIA05MAgKudAgCqnQIA15MAgNuTAICvIQEArjkBAK0xAQCsOQEA35MAgOOTAIDvZB8A55MAgOuTAIDvkwCA85MAgPeTAICADQAAgREAAIIVAAD7kwCA4eAcAP+TAIDjiB8AA5QAgISAAgC+jAUAh0gFAIYsBAALlACAD5QAgO+kHgDv9B4A4QAeAOFQHwDjLB4A47AeABOUAIAXlACAG5QAgB+UAIAjlACAJ5QAgISEBACzcQEAK5QAgLUdAQC2FQEAL5QAgDOUAIA3lACAugEBALsBAQC89QAAvf0AAL71AAC/7QAAqK0GAKm9BgCqtQYAq8kGAKzZBgCt2QYArskGAK/BBgA7lACAP5QAgEOUAIBHlACAS5QAgE+UAIBTlACAV5QAgLhtBwC5BQcAug0HALsBBwC8AQcAvQEHAL4BBwC/AQcAsIkGALGJBgCybQcAs2UHALR9BwC1ZQcAtmUHALdVBwCXkwCAozkGAFuUAIAHlACApl0GAF+UAIBjlACApVUGAKpJBgCrSQYAZ5QAgGuUAICuvQcAr6UHAKy9BwCttQcAgG0AAIEJAACCGQAAb5QAgHOUAIC+nAMAd5QAgHuUAICGQAAAh2AAAH+UAICDlACAh5QAgIuUAICPlACAk5QAgKiRBgCpkQYAqrkGAKu5BgCsqQYArakGAK7ZBgCv2QYAl5QAgJuUAICflACAo5QAgKeUAICrlACAr5QAgLOUAIC4cQEAuXEBALpxAQC7cQEAvNkBAL3BAQC+wQEAv/UBALCxBgCxuQYAsokGALOJBgC0UQEAtVEBALZRAQC3UQEAszEGALeUAIC7lACAv5QAgMOUAIC2KQYAtSEGAMeUAIC7fQYAunUGAMuUAIDPlACAv5UBAL6VAQC9XQYAvF0GANOUAICjdQYA15QAgNuUAICmbQYA35QAgOOUAIClZQYAqjEGAKs5BgCErAEAvqABAK7RAQCv0QEArBkGAK0ZBgCo3QIAqe0CAKrlAgCr/QIArOUCAK3tAgCu5QIArz0DAOuUAIDvlACA85QAgL5kDAD3lACA+5QAgP+UAIADlQCAuMkDALnJAwC62QMAu9EDALz5AwC9+QMAvpkDAL+VAwCwRQMAsU0DALJFAwCzXQMAtEUDALVNAwC2RQMAt/kDAIFVAwCASQMAs2UCAIJVAwC1ZQIAB5UAgAuVAIC2ZQIAhgAMAIfkAwC7gQMAuokDAL2BAwC8mQMAv4EDAL6JAwCjLQIAD5UAgBOVAIAXlQCAG5UAgKYtAgClLQIAH5UAgKvJAwCqwQMAI5UAgCeVAICvyQMArsEDAK3JAwCs0QMA49gGAOGsBwDhnAYA45wGACuVAICEWA0AL5UAgDOVAIA3lQCAO5UAgD+VAIBDlQCA7xwBAEeVAIBLlQCA70AGAIB5AACBFQAAghEAAIQADABPlQCA46wAAFOVAIDhpAEAW5UAgO9wAACGyAwAh6QNAF+VAIBjlQCAZ5UAgGuVAIC6yQUAu8kFALilBQC5zQUAvvkFAL/5BQC8zQUAvcUFALKlBQCzrQUAsBEGALERBgC2rQUAt50FALS1BQC1rQUAqmEGAKthBgConQYAqZUGAK5hBgCvYQYArHEGAK1xBgBvlQCAc5UAgHeVAIB7lQCAf5UAgIOVAIC+sAwAh5UAgKghDgCpIQ4AqiEOAKs9DgCsJQ4ArS0OAK4lDgCviQ4AV5UAgIuVAICPlQCAk5UAgJeVAICblQCAn5UAgKOVAIC4UQ8AuV0PALpVDwC7bQ8AvHUPAL19DwC+dQ8Av2kPALD5DgCxoQ4AsqEOALOhDgC0oQ4AtakOALaRDgC3kQ4As6kOAKeVAIDnlACAq5UAgK+VAIC2rQ4Ata0OALOVAIC7ZQ4Auj0OALeVAIC7lQCAv20OAL5lDgC9dQ4AvHUOAIIZAACj7Q4AgGUAAIEZAACm6Q4Av5UAgMOVAICl6Q4AqnkOAKshDgDHlQCAy5UAgK4hDgCvKQ4ArDEOAK0xDgCoYQ4AqXUOAKp9DgCrdQ4ArG0OAK31DgCu/Q4Ar/UOAIaAAQCHpAEAz5UAgNOVAIDXlQCA25UAgN+VAIDjlQCAuHUBALl9AQC6dQEAu8kBALzdAQC9xQEAvsUBAL/1AQCwjQ4AsZUOALKdDgCzkQ4AtFUBALVdAQC2VQEAt00BALP1DgDnlQCA65UAgO+VAIDzlQCAtnUOALXlDgD3lQCAu1EOALpJDgD7lQCA/5UAgL+ZAQC+kQEAvUUOALxJDgADlgCAo7EOAAeWAIALlgCApjEOAA+WAIATlgCApaEOAKoNDgCrFQ4AF5YAgBuWAICu1QEAr90BAKwNDgCtAQ4AqO0CAKktAwCqJQMAqz0DAKwlAwCtLQMAriUDAK+ZAwAflgCAI5YAgCeWAIArlgCAL5YAgDOWAIC+dAIAO5YAgLiNAwC5kQMAupEDALulAwC8vQMAvXUAAL59AAC/dQAAsOkDALHpAwCy+QMAs/EDALTZAwC12QMAtrkDALe1AwCArQAAgbUAAIK9AACzoQMAP5YAgLWhAwC2oQMAQ5YAgITgAgBHlgCAuiEDALshAwC8IQMAvSkDAL4RAwC/EQMAo+0DAIXABACFtG8AS5YAgE+WAICm7QMApe0DAFOWAICrbQMAqm0DAIZIBQCHbAMAr10DAK5dAwCtZQMArG0DAFeWAIDjAA4A71hsAOG0DwBblgCAX5YAgGOWAIBnlgCAoakDAKD9DwCjwQMAog0DAOHgAwDv4A8A4+QDAGuWAIBvlgCAc5YAgIQEBAC+BAQAd5YAgO+UAwB7lgCAf5YAgIOWAIDj1AMAh5YAgOFUAACLlgCAj5YAgJOWAICXlgCAgA0AAIEVAACCHQAAm5YAgJ+WAICjlgCAj5EbAO+cDgCE4AcA4dQOAKuWAIDj8A4Ar5YAgLOWAICGGAcAh5AEAJnlFwCY5RcAm+kLAJo5CwCd/QoAnPELAJ9VDwCeXQ8AkSkfAJDNGwCTJR8Aks0fAJXREwCUKRMAlxkXAJZ1EwCM4RAAjSUQAI4tEACP+QwAN5YAgKeWAICKORQAi5UUAITpGACFBRgAhuUYAIfxFAC3lgCAu5YAgIIxHACDFRwAnKkEAL+WAIDDlgCAx5YAgMuWAIDPlgCAmtEEAJt9BACUTQ0AleUIAJblCACXtQgA05YAgNeWAICSWQwAk1kMAKGRAADblgCAowF8AKKZAACluXwApJF8AKeZeACm4X0AqYF5AKiheACriXQAqgF0AK0BcACsWXQAr4VwAK6dcACx4WwAsAFsALMBaACyHWwAtfVoALT1aADflgCA45YAgOeWAIDrlgCA75YAgPOWAID3lgCA+5YAgP+WAIADlwCAqD0HAKmVBwCqlQcAq6kHAKzdBwCtxQcArsUHAK8dBgAHlwCAgh0AAIEdAACAHQAAC5cAgA+XAIATlwCAvmABALgZBgC5GQYAuikGALslBgC8IQYAvSEGAL4hBgC/IQYAsHEGALFxBgCycQYAs3EGALRNBgC1NQYAtj0GALctBgCzHQcAG5cAgIYoAACHqAAAH5cAgLZFBwC1VQcAI5cAgLu1BgC6tQYAJ5cAgCuXAIC/8QYAvokGAL2lBgC8pQYAL5cAgKNZBwAzlwCAN5cAgKYBBwA7lwCAP5cAgKURBwCq8QYAq/EGAEOXAIBHlwCArs0GAK+1BgCs4QYAreEGAKipBQCptQUAqr0FAKs9AgCsJQIArVECAK5RAgCvUQIAS5cAgE+XAIBTlwCAV5cAgIQ8AwBblwCAX5cAgGOXAIC4pQIAua0CALqlAgC7vQIAvKUCAL2tAgC+pQIAv30DALAxAgCxMQIAshkCALMZAgC09QIAta0CALalAgC3nQIAZ5cAgGuXAIBvlwCAszkFAHOXAIC1oQIAtt0CAHeXAIB7lwCAf5cAgLr5AgC7+QIAvMECAL3BAgC+PQIAv2UCAIOXAICmgQIApf0CAIuXAICjZQUAvlh8AIbYfACHnHwArzkCAK5hAgCtnQIArJ0CAKulAgCqpQIAj5cAgJOXAICohQIAqZUCAKqVAgCrpQIArL0CAK3VAgCu0QIAr9ECAIGFAQCAhQEAl5cAgILtAQCblwCAn5cAgKOXAICnlwCAuHUBALl9AQC6dQEAu80BALzVAQC93QEAvskBAL/BAQCwtQIAsb0CALKBAgCzgQIAtFEBALVRAQC2UQEAt1EBAKuXAICvlwCAs5cAgLeXAIDhMAYA4WQHAOMoBgDjxAYAhCB9ALuXAIDvbAAA7xgGAL+XAIDDlwCAx5cAgMuXAICzXQIAvkh8AM+XAIDTlwCA15cAgLYVAgC1dQIA25cAgLs5AgC6MQIA35cAgOOXAIC/1QEAvtUBAL0VAgC8FQIAo519AIeXAIDnlwCA65cAgO+XAICm1X0ApbV9APOXAICr+X0AqvF9APeXAID7lwCArxV+AK4VfgCt1X0ArNV9AIBNAACBVQAAglUAALOxfgD/lwCAtWV/ALZtfwADmACAhkADAIcEAwC66X8Au+l/ALz5fwC9+X8Avt1/AL/NfwAHmACAC5gAgBeXAIAPmACAE5gAgBeYAIAbmACAH5gAgKhtfgCpXX4AqlV+AKuFfwCsgX8ArYF/AK6BfwCvgX8AsEF/ALFBfwCyQX8As0F/ALR1fwC1ZX8Atm1/ALdlfwC4XX8AuS1/ALolfwC7PX8AvC1/AL0dfwC+FX8Av/UAAKP9fwAjmACAJ5gAgCuYAIAvmACApiF+AKUpfgAzmACAq6V+AKqlfgA3mACAO5gAgK+BfgCukX4ArbV+AKy1fgA/mACAQ5gAgEeYAIBLmACAT5gAgFOYAIBXmACAW5gAgIA9AACBCQAAghkAAF+YAIBjmACAhLgBAL6wAQBnmACAqK0BAKnVAQCq1QEAqw0BAKwVAQCtGQEArgkBAK8JAQCGAAQAhwQBAGuYAIBvmACAc5gAgHeYAIB7mACAf5gAgLjtAAC5hQAAuo0AALuFAAC8nQAAvYUAAL6NAAC/hQAAsHkBALF5AQCy7QAAs+UAALT9AAC15QAAtuUAALfVAACzXQIAg5gAgIeYAICLmACAj5gAgLaZAgC1nQIAk5gAgLu9AgC6vQIAl5gAgJuYAIC/IQMAvjkDAL0xAwC8OQMAvigDAKMZAgCfmACAo5gAgKbdAgCnmACAq5gAgKXZAgCq+QIAq/kCAK+YAICzmACArn0DAK9lAwCsfQMArXUDAL7IBAC3mACAu5gAgL7EBQC/mACAw5gAgMeYAIDLmACAgD0AAIEJAACCGQAAz5gAgNOYAICEOAMA25gAgN+YAIDveAIA45gAgIZIBACHVAMA55gAgOuYAIDvmACA85gAgPeYAID7mACA/5gAgAOZAIDjVAIAB5kAgOFAAQALmQCAD5kAgOMkfwATmQCA4Zx8ABeZAIAbmQCAH5kAgCOZAICEbAUAJ5kAgCuZAIAvmQCAM5kAgO8YfwA3mQCAO5kAgLPxAgA/mQCAQ5kAgEuZAIBPmQCAtukCALXhAgBTmQCAu3EBALppAQCHoAUAhswEAL85AQC+WQEAvVEBALxhAQDhQH8AV5kAgOM4fgCEwAQAgtkAAO8UAACApQAAgdkAAFuZAIDjwAAAX5kAgOHUAQBjmQCAZ5kAgO+EfgBrmQCAqs0BAKvVAQBvmQCAc5kAgK79AQCvnQEArMUBAK31AQB3mQCAo1UCAHuZAIB/mQCApk0CAIOZAICHmQCApUUCANeYAIBHmQCAi5kAgI+ZAICTmQCAl5kAgJuZAICfmQCAqJkGAKmZBgCq7QYAq/0GAKzlBgCt7QYAruUGAK/dBgCwpQYAsa0GALKlBgCzuQYAtK0GALVVBwC2UQcAt00HALh1BwC5fQcAunUHALtJBwC8WQcAvVkHAL5JBwC/RQcAs0UGAKOZAICnmQCAq5kAgK+ZAIC2TQYAtU0GALOZAIC7SQYAukEGAIYIAACHjAAAv7EHAL5JBgC9TQYAvFEGAIJdAACjAQYAgEUAAIFdAACmCQYAu5kAgL+ZAIClCQYAqgUGAKsNBgDDmQCAx5kAgK4NBgCv9QcArBUGAK0JBgCoTQYAqVUGAKpVBgCriQYArLEGAK29BgCuqQYAr6kGALeZAIDLmQCAz5kAgNOZAIDXmQCA25kAgN+ZAIDjmQCAuEkBALlJAQC6WQEAu1kBALxJAQC9SQEAvt0BAL/VAQCw3QYAsa0GALKlBgCzjQYAtJkGALWZBgC2jQYAt4UGALPdBgDnmQCA65kAgO+ZAIDzmQCAtj0GALU5BgD3mQCAu2kGALoZBgD7mQCA/5kAgL9dBgC+XQYAvVkGALxxBgADmgCAo5kGAAeaAIALmgCApnkGAA+aAIATmgCApX0GAKpdBgCrLQYAF5oAgBuaAICuGQYArxkGAKw1BgCtHQYAqNUCAKndAgCq4QIAq+ECAKw1AwCtPQMArjUDAK8tAwCAzQMAgQkAAIIZAAAfmgCAI5oAgIQYAgC+dAMAK5oAgLjpAwC56QMAuokDALuFAwC8nQMAvYEDAL6BAwC/tQMAsFUDALFdAwCyVQMAs+kDALT5AwC1+QMAtukDALfhAwCGIAwAhxADAC+aAIAzmgCAN5oAgDuaAIA/mgCA71wCAEOaAIDhFAAAR5oAgOOIAgC++AwAS5oAgE+aAIBTmgCAu/kDALrxAwC+gA0AV5oAgL9dAwC+XQMAvV0DALzhAwCzCQIAW5oAgF+aAIBjmgCAZ5oAgLbdAwC13QMAa5oAgKipBgCpqQYAqrkGAKu5BgCsqQYArakGAK4dBQCvFQUAb5oAgHOaAIB3mgCAe5oAgH+aAICDmgCAh5oAgIuaAIC4GQUAuS0FALolBQC7yQUAvNkFAL3FBQC+zQUAv8UFALBtBQCxdQUAsnUFALNFBQC0XQUAtT0FALY1BQC3KQUA4fQGAOFUBwDjFAYA47wGAIEJAACAqQAAj5oAgII5AACE7A0Ak5oAgIeIDACGDAwAm5oAgJ+aAIDvzAcA78QHAKMpAwCjmgCAp5oAgKuaAICvmgCApv0CAKX9AgCzmgCAq9kCAKrRAgC3mgCAu5oAgK99AgCufQIArX0CAKzBAgCoPQ4AqY0OAKqFDgCrnQ4ArIUOAK2NDgCuuQ4Ar7UOAJeaAIC/mgCAw5oAgMeaAIDLmgCAz5oAgNOaAIDXmgCAuL0OALllDwC6bQ8Au2UPALx9DwC9ZQ8Avm0PAL9lDwCw1Q4Asd0OALLVDgCzoQ4AtJUOALWdDgC2lQ4At40OALMNDgDbmgCA35oAgOOaAIDnmgCAtg0OALUNDgDrmgCAuxkOALoRDgDvmgCAJ5oAgL9ZDgC+UQ4AvXUOALwBDgDzmgCAo0kOAPeaAID7mgCApkkOAP+aAIADmwCApUkOAKpVDgCrXQ4AhKQDAAebAICuFQ4Arx0OAKxFDgCtMQ4AqLEOAKmxDgCqzQ4Aq8UOAKzdDgCtxQ4ArsUOAK/1DgCA7QEAgfEBAILxAQALmwCAhpABAIe0AQAPmwCAE5sAgLjFAQC5zQEAusUBALvdAQC8zQEAvf0BAL6ZAQC/lQEAsI0OALFBAQCyQQEAs0EBALRBAQC1QQEAtkEBALdBAQCzRQ4AF5sAgBubAIAfmwCAI5sAgLZFDgC1VQ4AJ5sAgLuFAQC6SQ4AK5sAgC+bAIC/hQEAvoUBAL2VAQC8lQEAM5sAgKMBDgA3mwCAO5sAgKYBDgA/mwCAQ5sAgKURDgCqDQ4Aq8EBAEebAIBLmwCArsEBAK/BAQCs0QEArdEBAKgtAwCpPQMAqjUDAKuJAwCsmQMArZkDAK6JAwCvgQMAT5sAgFObAIBXmwCAW5sAgF+bAIBjmwCAZ5sAgGubAIC4rQMAuWUAALptAAC7ZQAAvH0AAL1lAAC+bQAAv2UAALDJAwCxyQMAsqkDALOlAwC0vQMAtaEDALahAwC3lQMAgL0AAIEJAACCGQAAb5sAgHObAIC+2AMAe5sAgH+bAICErAIAg5sAgIfoAwCGDAQAh5sAgIubAICPmwCAk5sAgLP9AwCXmwCAm5sAgJ+bAICjmwCAtlkDALVRAwCnmwCAu00DALpNAwCrmwCAr5sAgL8lAwC+OQMAvTEDALw9AwCzmwCAt5sAgLubAIC/mwCA71gPAMObAIDHmwCAy5sAgOOQDgDPmwCA4bAPANObAIDXmwCA25sAgN+bAIDjmwCAgHUAAIF9AACCdQAAhBgFAO88AwDrmwCAvhQFAO+bAIDj0AMA85sAgOFAAAD3mwCAhtAEAIdYBQD7mwCA/5sAgAOcAIAHnACAC5wAgA+cAIATnACAF5wAgBucAIDvrA8AhOwEAOEQDgAfnACA41QBACOcAIAnnACAK5wAgC+cAICj/QIAM5wAgDecAIA7nACAP5wAgKZZAgClUQIAQ5wAgKtNAgCqTQIAR5wAgEucAICvJQIArjkCAK0xAgCsPQIAqJkGAKmZBgCqrQYAq70GAKylBgCtrQYArqUGAK/ZBgDnmwCAghEAAIEZAACAwQcAT5wAgFOcAIC+cAMAV5wAgLhJBwC5SQcAul0HALtVBwC8TQcAvXEHAL51BwC/bQcAsKkGALGpBgCyuQYAs7EGALSZBgC1mQYAtnkHALd5BwC1NQYAW5wAgF+cAIC2NQYAhjAAAIdcAwCzPQYAY5wAgL19BgC8dQYAv0UGAL5FBgB3mwCAZ5wAgLt1BgC6dQYAo2UGAGucAIBvnACAc5wAgHecAICmbQYApW0GAHucAICrLQYAqi0GAH+cAICDnACArx0GAK4dBgCtJQYArC0GAKhVBgCpWQYAqm0GAKthBgCsaQYArWkGAK6ZBgCvmQYAh5wAgIucAICPnACAk5wAgJecAICbnACAn5wAgKOcAIC4+QYAufkGALqNBgC7hQYAvJ0GAL2FBgC+hQYAv7UGALDpBgCx6QYAsvkGALP5BgC06QYAtd0GALbJBgC3yQYAs+UGAKecAICrnACAr5wAgLOcAIC26QYAteEGALecAIC7LQYAui0GALucAIC/nACAvxkGAL4tBgC9LQYAvC0GAIIVAACjoQYAgGEAAIFhAACmrQYAw5wAgL6QAQClpQYAqmkGAKtpBgCEpAEAy5wAgK5pBgCvXQYArGkGAK1pBgCohQIAqY0CAKqVAgCruQIArNUCAK3dAgCu1QIAr80CAIaAHACHZAMAz5wAgL5gAwDTnACA15wAgNucAIDfnACAuHUDALl9AwC6dQMAu8kDALzZAwC92QMAvskDAL/BAwCwvQIAsY0CALKFAgCzTQMAtFUDALVdAwC2VQMAt00DALMdAgDjnACAhAgDAOecAIDrnACAtl0CALVdAgDvnACAu0kCALp5AgDznACA95wAgL+ZAwC+kQMAvZkDALxRAgCwAAAAo1kCAPucAID/nACAphkCAAOdAIAHnQCApRkCAKo9AgCrDQIAC50AgA+dAICu1QMAr90DAKwVAgCt3QMAE50AgBedAIAbnQCA76wGAB+dAIAjnQCAJ50AgCudAIC+6BwAL50AgDOdAIA7nQCAP50AgOGABwBDnQCA42AGAIBdAACBYQAAgmEAALN9AQBHnQCAtW0BALZlAQBLnQCAhiAdAIdYHQC6+QEAu/EBALzZAQC92QEAvrEBAL+xAQDvoAAAT50AgFOdAIBXnQCAW50AgF+dAIBjnQCA71wBAIRsHADhzAYAZ50AgOMcBgDjSAAAa50AgOEwAQBvnQCAo/EBAHOdAICFABQAd50AgHudAICm6QEApeEBAH+dAICrfQEAqnUBAIOdAICHnQCArz0BAK49AQCtVQEArFUBAKjtHQCpLR4AqjkeAKs5HgCsKR4ArSkeAK6dHgCvkR4AN50AgIudAICPnQCAk50AgJedAICC+QAAgfEAAID9AAC4qR4AuakeALpJHwC7SR8AvFkfAL1FHwC+TR8Av0UfALDxHgCx+R4AssEeALPBHgC0uR4AtbkeALatHgC3pR4AsBEfALERHwCyER8AsyUfALQlHwC1KR8Atl0fALdRHwC4cR8AuXkfALpBHwC7QR8AvJUAAL2dAAC+lQAAv40AAJudAIDHnACAn50AgKOdAICnnQCAq50AgIb4AwCH0AAAqM0fAKnVHwCq0R8Aq70fAKytHwCtcR8ArnEfAK9xHwCzOR4Ar50AgLOdAIC3nQCAu50AgLaRHgC1RR4Av50AgLu1HgC6tR4Aw50AgMedAIC/jR4AvoEeAL2RHgC8pR4Ay50AgKN9HgDPnQCA050AgKbVHgDXnQCA250AgKUBHgCq8R4Aq/EeAN+dAIDjnQCArsUeAK/JHgCs4R4ArdUeAKhVAQCpgQAAqoEAAKuBAACsgQAArYkAAK6xAACvsQAA550AgOudAIDvnQCA850AgPedAID7nQCA/50AgAOeAIC4ZQAAuW0AALplAAC7fQAAvGUAAL1tAAC+ZQAAv90DALChAACxrQAAsqUAALO5AAC0qQAAtZ0AALaVAAC3XQAAB54AgIIdAACBHQAAgB0AAAueAIAPngCAE54AgL4UAgAbngCAhKgCAB+eAIAjngCAJ54AgCueAIAvngCAjwAAALNJAwAzngCAhugEAIesAgA3ngCAtkkDALVJAwA7ngCAuykDALolAwA/ngCAQ54AgL8ZAwC+LQMAvS0DALwxAwBHngCAo40DAEueAIBPngCApo0DAFOeAIBXngCApY0DAKrhAwCr7QMAW54AgF+eAICu6QMAr90DAKz1AwCt6QMAvoQDAGOeAIBnngCAa54AgG+eAIBzngCAd54AgHueAICAPQAAgQkAAIIZAAB/ngCAg54AgIueAICENAMAj54AgLMtAQCTngCAh8wCAIZMBQCXngCAti0BALUtAQCbngCAu0kBALp5AQCfngCAo54AgL+9AQC+vQEAvbkBALxRAQDheB8Ap54AgOPQHwCrngCAr54AgOGUAQCzngCA42gDALeeAIC7ngCAv54AgO+IAwDDngCAx54AgO+sHwDLngCAz54AgNOeAIDXngCA254AgN+eAIDjngCA554AgO9EHgDrngCA4dweAO+eAIDjHB4A854AgPueAID/ngCAA58AgIFpAACAZQAAo+UBAIJ9AACl5QEAB58AgIQUBACm5QEAvigEAAufAICrgQEAqrEBAK1xAQCsmQEAr3UBAK51AQCoIQYAqS0GAKolBgCrPQYArCUGAK0tBgCuXQYAr00GAIeeAID3ngCAhggDAIeMAwAPnwCAE58AgBefAIAbnwCAuOkGALnpBgC6jQYAu4UGALydBgC9hQYAvo0GAL+FBgCwPQYAsQ0GALIFBgCz7QYAtPkGALX5BgC27QYAt+UGALDNBwCx1QcAstEHALPtBwC09QcAtf0HALbpBwC36QcAuN0HALklBwC6LQcAuyUHALw9BwC9JQcAvi0HAL8lBwAfnwCAI58AgBeeAIAnnwCAK58AgC+fAIAznwCAN58AgKgVBgCpGQYAqu0HAKv9BwCs7QcArd0HAK7VBwCvuQcAswUGADufAIA/nwCAQ58AgEefAIC2PQYAtQUGAEufAIC7cQYAumkGAE+fAIBTnwCAv1kGAL5RBgC9WQYAvGUGAFefAICjQQYAW58AgF+fAICmeQYAY58AgIS0AQClQQYAqi0GAKs1BgC+gAEAa58AgK4VBgCvHQYArCEGAK0dBgCoNQYAqT0GAKo1BgCrWQYArHUGAK2lAQCurQEAr6UBAIDpAACB6QAAgv0AAL8kAQCGMA8Ah+QAAG+fAIBznwCAuMUAALnNAAC6xQAAu90AALzNAAC9/QAAvvUAAL+dAACw3QEAsSUBALItAQCzIQEAtCEBALUhAQC2IQEAtyEBALvBAgC6OQIAd58AgHufAIC/xQIAvsUCAL3VAgC82QIAs50FAH+fAICDnwCAh58AgIwAAAC2BQIAtd0FAIufAICqfQIAq4UCAI+fAICTnwCAroECAK+BAgCsnQIArZECAJefAICj2QUAm58AgJ+fAICmQQIAo58AgKefAIClmQUAgpFqAIORagCrnwCAr58AgIa5FgCH6RcAhBEWAIWZFgCKoRIAi6ESALOfAIC3nwCAjpEeAI9ZHgCMmRMAjREeAJJxGgCT5RoAu58AgO/oJACW8QYAlwUGAJTlGgCVGQYAmikCAJvFAgC/nwCAw58AgMefAIDhKBsAnN0CAOMgDwCfIQcAnsEHAJ01GwCcLRsAm6EbAJr5HwCZOR8AmLEfAJcBEgCWIRMAlSkTAJRRFgCTGRcAkjEXAJGxFwCQKWsAj1FrAOOsBwCEBA0A4RwHAIANAACBNQAAgj0AAMufAIDPnwCA058AgL4gDQDbnwCA358AgO9MBwCGWAwAh2ANAOOfAIDnnwCA658AgO+fAICEXA8A858AgO8IAADvhAYA4ZABAOGwBgDj4AAA42QGAPefAID7nwCA/58AgAOgAIAHoACAC6AAgL4ADwCEQA4AD6AAgBOgAIAXoACAG6AAgB+gAIAjoACAJ6AAgCugAICj1QMAotUDAKExAwCgLQcAZ58AgNefAIAvoACAM6AAgDegAICCmQAAgZEAAICZAACoTQ0AqZ0NAKqVDQCrJQ4ArD0OAK0RDgCuEQ4ArxEOALB9DgCxDQ4AsgUOALMtDgC0OQ4AtTkOALYtDgC3JQ4AuOkOALnpDgC6wQ4Au8EOALy5DgC9nQ4AvpUOAL+NDgCzPQ0AO6AAgD+gAIBDoACAR6AAgLaxDgC1lQ4AS6AAgLvpDgC6mQ4AhogAAIfkAAC/3Q4Avt0OAL3ZDgC88Q4AT6AAgKN5DQC+hAEAhIAGAKb1DgBToACAV6AAgKXRDgCq3Q4Aq60OAFugAIBfoACArpkOAK+ZDgCstQ4ArZ0OALIFNQCzGTQAsG0wALENNQBjoACAZ6AAgLQBKAC1PSkAa6AAgG+gAIBzoACAd6AAgHugAIB/oACAg6AAgIegAICiRQEAo9UBAIugAIChTQEAps0FAKcBOACkAQQApX0FAKoBPACrRT0AqEk5AKnlOQCudTEAr30xAKxdPQCtATAAqO0OAKn1DgCqCQ4AqwkOAKwZDgCtGQ4Arg0OAK8tDgCPoACAk6AAgJegAICboACAn6AAgKOgAICnoACAq6AAgLgdDgC5JQ4Aui0OALslDgC8PQ4Avd0BAL7VAQC/zQEAsFUOALFdDgCyVQ4Asy0OALQ1DgC1JQ4Ati0OALclDgCzgQ0Ar6AAgLOgAIC7oACAv6AAgLaZDQC1kQ0AvlQEALuZDQC6kQ0AhogEAIe8AwC/4Q0AvvENAL35DQC8gQ0AgkkAAKPFDQCA9QMAgUkAAKbdDQDDoACAx6AAgKXVDQCq1Q0Aq90NAMugAIDPoACArrUNAK+lDQCsxQ0Arb0NAKgdAgCpRQIAql0CAKtVAgCseQIArXkCAK6JAwCviQMA06AAgNegAIDboACA36AAgIT8BQDjoACA56AAgOugAIC4iQMAuWUDALptAwC7ZQMAvH0DAL1lAwC+bQMAv2UDALDBAwCxwQMAssEDALPBAwC0wQMAtcEDALbBAwC3wQMA76AAgPOgAID3oACA+6AAgP+gAIDhpAEAA6EAgOPADgC+aAQAB6EAgAuhAIDvHAEAD6EAgBOhAIAXoQCAG6EAgLOVAwAfoQCAI6EAgCuhAIAvoQCAtrkDALWxAwAzoQCAu0UCALpFAgCGqAQAh6QFAL9FAgC+RQIAvVUCALxVAgDh4A4A4SwMAOMIDgDj1A4AgK0AAIHRAACC0QAAN6EAgDuhAIA/oQCAQ6EAgEehAIBLoQCAT6EAgO+IDgDvLA4AoxUDAFOhAICFxCsAV6EAgFuhAICmOQMApTEDAF+hAICrxQIAqsUCAGOhAIBnoQCAr8UCAK7FAgCt1QIArNUCAKgNBgCpFQYAql0GAKtVBgCseQYArXkGAK65BgCvuQYAJ6EAgGuhAIBvoQCAc6EAgHehAIB7oQCAf6EAgIOhAIC4TQcAuVUHALpRBwC7aQcAvHkHAL1lBwC+bQcAv2UHALDJBgCxyQYAst0GALPVBgC0zQYAtXUHALZ9BwC3dQcAs9UGAIehAICLoQCAj6EAgJOhAIC2+QYAtfEGAJehAIC7DQYAug0GAIYIAACHLAAAv7EHAL4JBgC9AQYAvAkGAIJRAACjkQYAgEEAAIFBAACmvQYAm6EAgJ+hAICltQYAqkkGAKtJBgCjoQCAp6EAgK5NBgCv9QcArE0GAK1FBgCwsQYAsbEGALLNBgCzwQYAtMEGALXJBgC28QYAt/EGALgFAQC5DQEAugUBALsdAQC8BQEAvQ0BAL4FAQC/uQEAq6EAgK+hAICzoQCAt6EAgLuhAIC/oQCAt6AAgMOhAICoLQYAqTUGAKo1BgCr8QYArNEGAK3RBgCu0QYAr9EGALPdBgDHoQCAy6EAgM+hAIDToQCAtjEGALU5BgDXoQCAuxUGALoVBgDboQCA36EAgL9tBgC+ZQYAvXUGALx5BgDjoQCAo5kGAOehAIDroQCApnUGAO+hAIDzoQCApX0GAKpRBgCrUQYA96EAgPuhAICuIQYArykGAKw9BgCtMQYAqNUCAKndAgCq4QIAq+ECAKxRAwCtUQMArlEDAK9RAwD/oQCAA6IAgL7sAwALogCAD6IAgBOiAIAXogCAG6IAgLjpAwC56QMAuokDALuFAwC8nQMAvYEDAL6BAwC/tQMAsDEDALExAwCyNQMAs+kDALT5AwC1+QMAtukDALfhAwCAbQMAgaUAAIKtAACzZQIAH6IAgLXVAwC23QMAI6IAgITgAgAnogCAuvkDALv5AwC87QMAvTEDAL4xAwC/MQMAh+wDAIZkPACyAAAAK6IAgC+iAIDjCAQAM6IAgOHsBgA3ogCA7wAGADuiAIA/ogCAQ6IAgEeiAIBLogCAT6IAgFOiAIBXogCAW6IAgF+iAIDjoAMAY6IAgOGoAQBnogCA7/ADAIIdAACBHQAAgB0AAGuiAIBvogCAc6IAgHuiAIC+TD0Af6IAgKOhAwC+QDwApRECAIOiAICHogCAphkCAIRsAgCLogCAqz0CAKo9AgCt9QIArCkCAK/1AgCu9QIAhkA8AIe0PQCPogCAk6IAgJeiAICbogCAn6IAgO9EBgCjogCA4dQGAKeiAIDjDAcAq6IAgK+iAICzogCAt6IAgLP1AQC7ogCAv6IAgMOiAIDHogCAtkUBALXlAQDLogCAuzEBALopAQDPogCA06IAgL8dAQC+HQEAvRkBALwlAQCoLT4AqTU+AKo9PgCrNT4ArC0+AK2FPgCuhT4Ar7k+AHeiAIDXogCA26IAgN+iAICAGQAAgRkAAIIFAADjogCAuLk+ALm5PgC6ST8Au0k/ALxZPwC9WT8Avk0/AL9BPwCwrT4AsbU+ALKxPgCzjT4AtJk+ALWZPgC2iT4At4k+AKO1PgCEjAIA56IAgOuiAIDvogCApgU+AKWlPgDzogCAq3E+AKppPgCGCAAAh2gDAK9dPgCuXT4ArVk+AKxlPgD3ogCAs5E/APuiAID/ogCAtlk/AAOjAIAHowCAtbk/ALp1PwC7fT8AC6MAgA+jAIC+QT8Av0E/ALxZPwC9VT8AsJU+ALGdPgCyqT4As6U+ALShPgC1oT4AtqE+ALehPgC45T4Aue0+ALrlPgC7/T4AvO0+AL3dPgC+1T4AvxkBABOjAIAXowCAG6MAgB+jAIAjowCAB6IAgCejAIArowCAqF0+AKkhPgCqPT4AqzU+AKwVPgCt/T4ArvU+AK/tPgCj1T4AL6MAgDOjAIA3owCAO6MAgKYdPgCl/T4AP6MAgKs5PgCqMT4AQ6MAgEejAICvBT4ArgU+AK0RPgCsHT4AgREAAIANAABLowCAghkAAE+jAIBTowCAhJQBAL4QAACGQAcAhwABAFujAIBfowCAY6MAgGejAIBrowCAb6MAgKiNAgCplQIAqpUCAKvNAgCs2QIArdkCAK7NAgCvxQIAc6MAgHejAIB7owCAf6MAgIwAAACDowCAh6MAgIujAIC4HQMAucEDALrBAwC7wQMAvMEDAL3JAwC+8QMAv/EDALCJAgCxiQIAsikDALMpAwC0OQMAtTkDALYpAwC3JQMAsx0CAI+jAICTowCAl6MAgJujAIC2WQIAtVECAJ+jAIC7TQIAuk0CAKOjAICnowCAv/0DAL79AwC9/QMAvP0DAKujAICvowCAs6MAgLejAIDhDD4Au6MAgOOoPwC/owCAgT0AAIAxAADvUD8Agh0AAMOjAIC++AQAhhgFAIdMAwCEDAIA48wAAMujAIDhvAEAz6MAgNOjAIDXowCA26MAgN+jAICELAUA46MAgOejAIDrowCA7xAAAO+jAIDzowCAo90DAPejAID7owCA/6MAgAOkAICmmQMApZEDAAekAICrjQMAqo0DAAukAIAPpACArz0CAK49AgCtPQIArD0CABOkAIAXpACAG6QAgB+kAIAjpACAJ6QAgCukAIDvKD4AL6QAgOE8PgAzpACA4zgBAIApAACBFQAAghEAADukAICzMQIAvsgEAITABAA/pACAQ6QAgLYpAgC1IQIAR6QAgLvNAQC6zQEAS6QAgE+kAIC/dQEAvskBAL3BAQC8yQEAqOkFAKnpBQCq+QUAq/kFAKzpBQCt6QUArjkGAK85BgDHowCAN6QAgIaIAACHQAMAU6QAgFekAIBbpACAX6QAgLjRBgC52QYAuuEGALvhBgC8kQYAvZEGAL6RBgC/kQYAsEkGALFJBgCyXQYAs1UGALRNBgC18QYAtvEGALfxBgCjcQUAY6QAgGekAIBrpACAb6QAgKZpBQClYQUAc6QAgKuNBgCqjQYAd6QAgHukAICvNQYArokGAK2BBgCsiQYAf6QAgLPRBwCDpACAh6QAgLbxBwCLpACAj6QAgLXBBwC60QcAu90HAJOkAICXpACAvrkHAL+5BwC8xQcAvbkHALhpBgC5aQYAuokGALuJBgC8mQYAvZkGAL6JBgC/iQYAsBEGALEdBgCyFQYAs2kGALR5BgC1eQYAtmkGALdhBgCoSQYAqVUGAKpdBgCrVQYArE0GAK11BgCucQYAr3EGAFejAICCHQAAgR0AAIAdAACbpACAn6QAgKOkAIC+cAEAo5UGAKukAICGKAAAh0gBAK+kAICmtQYApYUGALOkAICrmQYAqpUGALekAIC7pACAr/0GAK79BgCt/QYArIEGAL+kAICzFQYAw6QAgMekAIC2PQYAy6QAgM+kAIC1NQYAutkBALvZAQDTpACA16QAgL59AQC/ZQEAvH0BAL11AQCovQUAqckFAKrZBQCr0QUArPkFAK35BQCuKQIArykCANukAIDfpACA46QAgOekAICMAAAA66QAgO+kAIDzpACAuO0CALmFAgC6gQIAu4ECALyFAgC9jQIAvrECAL+xAgCwWQIAsVkCALLtAgCz5QIAtP0CALXlAgC25QIAt9UCAKNRBQD3pACA+6QAgP+kAIADpQCApnkFAKVxBQAHpQCAq50CAKqdAgALpQCAD6UAgK8hAgCuOQIArTECAKw5AgCBbQAAgG0AABOlAICCBQAAvlwMABulAIAfpQCA79AGAITsAwDhHAUAI6UAgOP8BwAnpQCAK6UAgIbYDACHvAwAqIUCAKmVAgCqlQIAq6UCAKy9AgCt1QIArtECAK/RAgAvpQCAM6UAgDelAIA7pQCAP6UAgEOlAIBHpQCAS6UAgLh1AQC5fQEAunUBALvJAQC82QEAvdkBAL7JAQC/wQEAsLUCALG9AgCygQIAs4ECALRRAQC1UQEAtlEBALdRAQBPpQCAhAQNAFOlAIBXpQCAvhwMAFulAIDvHAAA76AGAOGQAQDhRAcA43AGAOOYBgBfpQCAY6UAgGelAIBrpQCAs10CAG+lAIBzpQCAd6UAgHulAIC2FQIAtXUCAH+lAIC7OQIAujECAIOlAICLpQCAv9UBAL7VAQC9FQIAvBUCAKOdDQAXpQCAh6UAgI+lAICTpQCAptUNAKW1DQCXpQCAq/kNAKrxDQCGCAMAh2ADAK8VDgCuFQ4ArdUNAKzVDQCAkQ8AgZkPAIKhDwCzpQ4Am6UAgLWhDgC2eQ8An6UAgKOlAICnpQCAukUPALtdDwC8RQ8AvU0PAL5FDwC//Q8AqFUOAKldDgCqYQ4Aq30OAKxlDgCttQ8Arr0PAK+1DwCrpQCAr6UAgLOlAIC3pQCAu6UAgL+lAIDDpQCAx6UAgLhVDwC5dQ8Aun0PALt1DwC8bQ8AvREPAL4RDwC/EQ8AsM0PALHVDwCy3Q8As9UPALTNDwC1dQ8AtnEPALdxDwCj6Q8Ay6UAgM+lAIDTpQCA16UAgKY1DgCl7Q8A26UAgKsRDgCqCQ4A36UAgOOlAICvsQ4ArgkOAK0BDgCsCQ4A56UAgIIdAACBHQAAgB0AAOulAIDvpQCA86UAgL6UAQCErAEA96UAgIfgAQCGzAAA+6UAgP+lAIADpgCAp6QAgKhtDgCpiQEAqpkBAKuRAQCswQEArckBAK75AQCv+QEAhKAAAAemAIALpgCAD6YAgBOmAIAXpgCAG6YAgB+mAIC4xQAAuc0AALrFAAC73QAAvM0AAL39AAC+9QAAv50AALBBAQCxQQEAskEBALNBAQC0QQEAtUEBALZBAQC3QQEAsxECACOmAIAnpgCAK6YAgC+mAIC2SQIAtUkCADOmAIC7hQIAuoUCADemAIA7pgCAv4UCAL6FAgC9lQIAvJUCAIU8GgCjVQIAP6YAgEOmAICmDQIAR6YAgEumAIClDQIAqsECAKvBAgBPpgCAU6YAgK7BAgCvwQIArNECAK3RAgCCGQAAV6YAgIAZAACBGQAAW6YAgF+mAIBjpgCAa6YAgL4ABABvpgCAc6YAgHemAIB7pgCAf6YAgIOmAICHpgCA7+gOAIumAICG6AQAh1ADAI+mAICTpgCA74ACAJemAIDhlAEAm6YAgONYAQCfpgCA4wAOAKOmAIDhaA0Ap6YAgKhxAgCpcQIAqnECAKupAgCsuQIArbkCAK6pAgCvqQIAhKwFAKumAICvpgCAs6YAgLemAIC7pgCAv6YAgMOmAIC4bQEAuQ0BALoFAQC7GQEAvAkBAL09AQC+NQEAv9kBALDZAgCx2QIAsm0BALNlAQC0fQEAtWUBALZlAQC3VQEA4WAPAOP0AADjHA4A4bwBAMemAICCOQAAgTEAAIA9AADLpgCAvigEAM+mAIDTpgCAvjwHAO8QAADv0A4A26YAgIbgBACHyAQA36YAgLO1AgDjpgCAtX0CALZ1AgDnpgCA66YAgO+mAIC6UQIAu1ECALz1AQC9/QEAvvUBAL/tAQBnpgCA16YAgKqxBQCrsQUArBUGAK0dBgCuFQYArw0GAPOmAID3pgCA+6YAgKNVBQD/pgCApZ0FAKaVBQADpwCAs+kGAAenAIALpwCAD6cAgBOnAIC24QYAtekGABenAIC7sQYAuqEGABunAIAfpwCAv50GAL6RBgC9pQYAvKkGAKgdBgCpIQYAqiEGAKshBgCsIQYArSEGAK4hBgCvIQYAI6cAgCenAIArpwCAL6cAgDOnAIA3pwCAO6cAgD+nAIC45QcAue0HALrlBwC7/QcAvOUHAL3tBwC+5QcAv00HALAlBgCxNQYAsj0GALMxBgC0FQYAtRkGALYNBgC3AQYAo6kHAIIVAACBtQEAgLUBAEOnAICmoQcApakHAEenAICr8QcAquEHAISgAgBLpwCAr90HAK7RBwCt5QcArOkHAE+nAICzlQYAhugAAIcYAQC2tQYAU6cAgFenAIC1vQYAukkBALtVAQBbpwCAX6cAgL45AQC/OQEAvEUBAL05AQCoPQYAqU0GAKpZBgCrUQYArHEGAK1xBgCuuQEAr7kBAISsAQBjpwCAZ6cAgGunAIBvpwCAc6cAgHenAIB7pwCAuKkBALmpAQC6aQEAu2kBALx5AQC9eQEAvmkBAL9pAQCwyQEAsdUBALLVAQCzqQEAtLkBALW5AQC2qQEAt6EBAKPRBQB/pwCAg6cAgIenAICLpwCApvEFAKX5BQCPpwCAqxECAKoNAgCTpwCAl6cAgK99AgCufQIArX0CAKwBAgCbpwCAn6cAgKOnAICnpwCAgTEAAIANAACrpwCAgjkAAK+nAICzpwCAviQDALunAIC/pwCAw6cAgIbYHACHTAMAx6cAgMunAIDPpwCAhMAcAOMgAQDTpwCA4cgBANenAIDvMAIA26cAgN+nAIDjpwCA56cAgOunAIDvpwCA86cAgLOVAwD3pwCA+6cAgP+nAIADqACAtrkDALWxAwAHqACAu1EDALpJAwALqACAD6gAgL/1AAC+SQMAvUEDALxJAwCoLQIAqUUCAKpdAgCrVQIArHkCAK15AgCuvQIAr7UCAL5oHQATqACAF6gAgBuoAICAHQAAgQkAAIKpAAAfqACAuFEBALlZAQC6YQEAu2EBALwRAQC9EQEAvhEBAL8RAQCwzQIAsdUCALLdAgCz1QIAtM0CALVxAQC2cQEAt3EBAOFYBgDhVAcA47AAAOO8BgAjqACAK6gAgIYYHACHVB0AL6gAgDOoAIA3qACAO6gAgL74HAA/qACA7/AGAO/gBgCjlQIAQ6gAgEeoAIBLqACAT6gAgKa5AgClsQIAU6gAgKtRAgCqSQIAV6gAgFuoAICv9QEArkkCAK1BAgCsSQIAqG0eAKl1HgCqfR4Aq40eAKyVHgCtnR4Aro0eAK+BHgAnqACAX6gAgGOoAIBnqACAa6gAgG+oAIBzqACAd6gAgLiJHgC5iR4AupkeALuRHgC8uR4AvbkeAL59HwC/dR8AsMUeALHNHgCyxR4As90eALTFHgC1zR4AtsUeALe5HgCz9R4Ae6gAgH+oAICDqACAh6gAgLYdHgC1HR4Ai6gAgLsJHgC6AR4Aj6gAgJOoAIC/CR4AvgEeAL0JHgC8ER4Agm0AAKOxHgCAVQAAgWUAAKZZHgCEmAMAv9ABAKVZHgCqRR4Aq00eAIYABACHmAEArkUeAK9NHgCsVR4ArU0eAJuoAICfqACAhCQAAKOoAICnqACAq6gAgLenAICXqACAqLUeAKmFHgCqjR4Aq4UeAKydHgCtgR4Arv0eAK/1HgCwjR4AsZUeALKVHgCzpR4AtL0eALVxAQC2cQEAt3EBALhRAQC5UQEAulEBALtRAQC89QEAvf0BAL71AQC/7QEAsyUeAL4IBwCvqACAs6gAgLeoAIC2IR4AtTUeALuoAIC7cR4AumkeAL+oAIDDqACAv5UBAL5ZHgC9UR4AvGEeAMeoAICjYR4Ay6gAgM+oAICmZR4A06gAgNeoAIClcR4Aqi0eAKs1HgDbqACA36gAgK4dHgCv0QEArCUeAK0VHgDhVBoA46gAgONcCgDnqACA66gAgO+oAIDzqACA96gAgPuoAIC+qAUA/6gAgAOpAICPMSoAC6kAgO/E+wAPqQCAk2EuAJIdLwCR2SoAkEkqAJfZEgCWdRIAlQ0TAJTBLgCbHRsAmkEWAJlJFgCYDRcAn3EeAJ4RGwCdcRoAnHkaAKOhAgCinQMAoZUfAKCJHgDjiAEA4wgeAOFoAADh/B4A79wBAO98HwC1if4AtAH8ALMB+gCylfoAsQH4ALAR9gCv4fYArgH0AK0l8gCs7fIAqwHwAKrpDwCp1Q4AqN0OAKcBDACmyQoApe0KAKQBCACj4QYAovEGAKHlAwATqQCAggErAIMBKwAXqQCAG6kAgIYxLwCHiS8AhIkrAIVFLgCKdRIAiwUTAIYIBQCHbAUAjhEXAI8RFwCMsRMAjV0WAJI9GgCTQRsAhMgFAIQABwCWUR8Al1EfAJRRGwCVORoAmn0eAJt9AgAfqQCAI6kAgIFZAQCAVQEAnFkDAIJRAQC+yAcAJ6kAgCupAIAvqQCAM6kAgDepAIA7qQCA79QeAD+pAIDhJB4AQ6kAgONoAQBHqQCAS6kAgE+pAIBTqQCAu2kCALpZAgBXqQCAW6kAgL8dAgC+HQIAvRkCALxxAgCz7QIAX6kAgGOpAIBnqQCAa6kAgLZ9AgC17QIAb6kAgKMNBQAHqQCAc6kAgHupAIB3qQCApp0FAKUNBQB/qQCAq4kFAKq5BQCGCAMAh3wDAK/9BQCu/QUArfkFAKyRBQCAsQcAgbkHAIJBAACzsQYAg6kAgLVZBwC2MQcAh6kAgIupAICPqQCAuuEHALvhBwC84QcAveEHAL7hBwC/3QcAqLUGAKm5BgCqdQYAq4UHAKydBwCt/QcArvUHAK8ZBwCTqQCAl6kAgJupAICfqQCAo6kAgKepAICrqQCAr6kAgLh1BwC5fQcAunUHALsFBwC8HQcAvTEHAL4xBwC/MQcAsGkHALFpBwCyeQcAs3kHALRpBwC1VQcAtlEHALdNBwCj/QcAs6kAgLepAIC7qQCAv6kAgKZ9BgClFQYAw6kAgKutBgCqrQYAx6kAgMupAICvkQYArq0GAK2tBgCsrQYAz6kAgNOpAIDXqQCA26kAgIAdAACBCQAAgjkAAN+pAIDjqQCA66kAgIbIAACHpAEA76kAgPOpAID3qQCA+6kAgKiNAQCpmQEAqtkBAKvRAQCs8QEArfEBAK45AQCvOQEAhKAAAP+pAIADqgCAB6oAgAuqAIAPqgCAE6oAgBeqAIC4zQAAudUAALrVAAC75QAAvP0AAL2VAAC+nQAAv5UAALBJAQCxSQEAslkBALNZAQC0SQEAtUkBALb9AAC39QAAugUEALsJBAC44QcAueEHAL4JBAC/CQQAvAkEAL0JBACyjQcAs+UHALC1BwCxhQcAtuUHALftBwC08QcAtfEHAKpNBwCrVQcAqEkHAKlJBwCu3QcAr8UHAKxNBwCt1QcAG6oAgB+qAIAjqgCAJ6oAgCuqAIAvqgCAM6oAgDeqAICz0QIAO6oAgD+qAIC+AAwAQ6oAgLbxAgC1+QIAR6oAgLsNAgC6DQIAS6oAgE+qAIC/DQIAvg0CAL0NAgC8DQIAghUAAKOVAgCAYQAAgWEAAKa1AgBTqgCAW6oAgKW9AgCqSQIAq0kCAIbIDACHrAwArkkCAK9JAgCsSQIArUkCAKhlAgCpdQIAqn0CAKt1AgCsbQIArbECAK6xAgCvsQIAhKANAF+qAIBjqgCAZ6oAgGuqAIBvqgCAc6oAgHeqAIC4MQEAuTEBALoxAQC7MQEAvNUBAL3dAQC+yQEAv8EBALDRAgCx0QIAstECALPRAgC0EQEAtREBALYRAQC3EQEA4bAGAHuqAIDj0AYAhEAPAH+qAIDhpAEAg6oAgOPABgCHqgCAi6oAgI+qAIDv1AYA7AAAAJOqAIDvZAcAl6oAgJuqAICfqgCAo6oAgLO5AgCnqgCAtakCALZ9AgCrqgCAr6oAgLOqAIC6WQIAu1kCALxJAgC9SQIAvpkBAL+ZAQCjdQ0AV6oAgLeqAIC7qgCAv6oAgKaxDQClZQ0Aw6oAgKuVDQCqlQ0AvqQDAMeqAICvVQ4ArlUOAK2FDQCshQ0AgE0AAIFVAACCVQAAs2UPAMuqAIC1ZQ8Atm0PAM+qAICGQAMAhxQDALrtDwC7/Q8AvOkPAL3VDwC+3Q8Av9UPAKhZDgCpoQ8AqqEPAKuhDwCsoQ8AraEPAK6hDwCvoQ8A06oAgNeqAIDbqgCA36oAgOOqAIDnqgCA66oAgO+qAIC4AQ8AuQEPALoBDwC7HQ8AvA0PAL01DwC+PQ8Av9UAALBlDwCxdQ8AsnEPALNNDwC0VQ8AtV0PALZNDwC3QQ8AoykOAPOqAID3qgCA+6oAgP+qAICmIQ4ApSkOAAOrAICrsQ4AqqEOAAerAIALqwCAr5kOAK6RDgCtmQ4ArKUOAA+rAIATqwCAF6sAgBurAIDvJA0AH6sAgCOrAIAnqwCA49AOACurAIDhGA4AL6sAgIAVAACBGQAAggUAADOrAICo0QEAqdkBAKopAQCrKQEArDkBAK05AQCuKQEArykBAL5oAQA7qwCAhsgBAIesAAA/qwCAQ6sAgEerAIBLqwCAuO0AALmFAAC6jQAAu4UAALydAAC9gQAAvoEAAL+BAACwWQEAsVkBALLtAACz5QAAtP0AALXlAAC25QAAt9UAALOhAgBPqwCAU6sAgFerAIBbqwCAtrkCALWxAgBfqwCAu50CALqdAgBjqwCAZ6sAgL8hAwC+OQMAvTEDALw5AwCF+PUAo+UCAGurAIBvqwCApv0CAHOrAIB3qwCApfUCAKrZAgCr2QIAe6sAgH+rAICufQMAr2UDAKx9AwCtdQMAuOkAALnpAAC6aQAAu2kAALx5AAC9ZQAAvm0AAL9lAACwsQAAsbkAALKBAACzgQAAtPkAALX5AAC27QAAt+UAAKhlAwCpdQMAqn0DAKt1AwCsbQMArdEAAK7RAACv0QAAg6sAgIerAICLqwCA56kAgI+rAICTqwCAl6sAgJurAICA/QEAgQkAAIIZAACfqwCAo6sAgL5EAgCrqwCAr6sAgISsAgCzqwCAh/gCAIasBQC3qwCAu6sAgL+rAIDDqwCAs/UCAMerAIDLqwCAz6sAgNOrAIC2UQEAteUCANerAIC7fQEAunUBANurAIDfqwCAvz0BAL49AQC9VQEAvFUBAOFwDwDjqwCA47gOAITABQDvyAAA56sAgOurAIDvqwCA4zwOAPOrAIDh0AEA96sAgIR0BwD7qwCA72gBAP+rAIADrACApXkCAKbNAQAHrACAgCEAAIEhAACC3QcAo2kCAKzJAQCtyQEArqEBAK+hAQALrACAD6wAgKrpAQCr4QEAp6sAgBOsAIC+QAIAF6wAgIYwAwCHMAMAG6wAgB+sAICoOQcAqTkHAKoNBwCrHQcArAUHAK0NBwCuBQcAr3kHALAJBwCxCQcAshkHALMRBwC0OQcAtTkHALbdBwC3yQcAuPkHALn5BwC6zQcAu8EHALzFBwC9yQcAvrkHAL+xBwCzpQcAI6wAgCesAIArrACAL6wAgLatBwC1rQcAM6wAgLvtBwC67QcAN6wAgDusAIC/3QcAvt0HAL3lBwC87QcAP6wAgKPhBwBDrACAR6wAgKbpBwBLrACAT6wAgKXpBwCqqQcAq6kHAFOsAIBXrACArpkHAK+ZBwCsqQcAraEHAFusAIBfrACAY6wAgGesAIBrrACAb6wAgHOsAIB3rACAgREAAIANAAB7rACAghkAAH+sAICDrACAvuQBAIesAICG4AAAhxgBAIusAICPrACAk6wAgJesAICbrACA77AEAJ+sAIDh1AYAo6wAgONcBACnrACAq6wAgK+sAICzrACAqJkBAKmZAQCqDQEAqwUBAKwdAQCtBQEArgUBAK81AQCEiAEAt6wAgLusAIC/rACAw6wAgMesAIDLrACAz6wAgLjBAAC5wQAAusEAALvBAAC8wQAAvcEAAL7BAAC/wQAAsE0BALElAQCyIQEAsyEBALQlAQC1LQEAthEBALcRAQDTrACA16wAgLONAgDbrACAtZ0CAN+sAIDjrACAto0CAOesAIDrrACAu+kCALqBAgC9/QIAvP0CAL/hAgC+6QIA76wAgKbVAgClxQIAvggDAKPVAgCCLQAAgRkAAIB5AACvuQIArrECAK2lAgCspQIAq7ECAKrZAgDzrACA+6wAgO80AgD/rACAhxgDAIYs/AADrQCAB60AgAutAIAPrQCAE60AgBetAIAbrQCAH60AgOMAAQAjrQCA4eABACetAIC6tQMAu70DACutAIAvrQCAvnkDAL95AwC8pQMAvXkDADerAICztQMAM60AgDetAIC2kQMAO60AgD+tAIC1pQMAqEkCAKlJAgCqWQIAq1kCAKxJAgCtdQIArnECAK9tAgC+aP0AvqT/AEOtAIBHrQCAS60AgE+tAIBTrQCAV60AgLj5AgC5+QIAukkBALtJAQC8XQEAvUEBAL5BAQC/fQEAsBUCALEdAgCyFQIAs8kCALTZAgC12QIAtskCALfJAgDjIAYA4bAGAOGAAQDjEAYAgA0AAIE1AACCPQAAW60AgF+tAIBjrQCAa60AgG+tAIDvcAAAc60AgHetAIDvTAEAhIz9AHutAICjmQIAf60AgKWJAgCDrQCAh60AgKa9AgCGwPwAh+T8AKuRAgCqmQIArVUCAKyJAgCvVQIArlUCAKh9/gCpgf4Aqpn+AKuZ/gCsif4ArYn+AK65/gCvuf4AZ60AgIutAICPrQCAk60AgJetAICbrQCAn60AgKOtAIC4tf4Aub3+ALph/wC7Yf8AvGH/AL1h/wC+Yf8Av2H/ALDJ/gCxyf4Ast3+ALPR/gC0uf4Atbn+ALaR/gC3kf4AsxH+AKetAICrrQCAr60AgLOtAIC2Cf4AtQH+ALetAIC7Df4Aug3+ALutAIC/rQCAv33+AL59/gC9Bf4AvAn+AMOtAICjVf4Ax60AgMutAICmTf4Az60AgNOtAIClRf4Aqkn+AKtJ/gCEKAMA160AgK45/gCvOf4ArE3+AK1B/gCAzQEAgdEBAILRAQCzuf4A260AgLXR/gC21f4A360AgIZgAQCHYAEAug0BALsFAQC8HQEAvQUBAL4NAQC/BQEA460AgOetAIDrrQCA760AgPOtAIDhwP0A960AgOOM/AD7rQCA/60AgAOuAIDvtPwAB64AgAuuAIAPrgCAE64AgKgp/gCpKf4Aqj3+AKs1/gCsVf4ArVn+AK5N/gCvRf4AF64AgBuuAIAfrgCAI64AgCeuAIArrgCAL64AgDOuAIC4SQEAuUkBALpZAQC7UQEAvHkBAL15AQC+GQEAvxUBALDFAQCxzQEAssUBALPdAQC0xQEAtc0BALbFAQC3eQEAN64AgDuuAIA/rgCAo7n9AEOuAICl0f0AptX9AITQAwBSrgCAvuACAKoNAgCrBQIArB0CAK0FAgCuDQIArwUCAIFJAACAQQAAowkDAIJdAAClGQMAVq4AgFquAICmEQMAhsAEAIfkAwCrDQMAqg0DAK0BAwCsHQMArwEDAK4JAwCw4QMAseEDALLhAwCz/QMAtOUDALXtAwC25QMAtz0DALgFAwC5DQMAugUDALsdAwC8BQMAvQ0DAL4FAwC/vQAAXq4AgGKuAIBmrgCAaq4AgPesAIBurgCAcq4AgHauAICo8QMAqfkDAKqpAwCrqQMArLkDAK25AwCuqQMAr6UDALNBAgB6rgCAfq4AgIKuAICGrgCAtlkCALVRAgCKrgCAu0UCALpFAgCOrgCAkq4AgL9JAgC+QQIAvUkCALxVAgCWrgCAmq4AgJ6uAICirgCA74wDAKauAICqrgCArq4AgONsAwCyrgCA4VAAALauAIC6rgCAvngFAMKuAICEcAIAgOUAAIHpAACC+QAAxq4AgIawBACHVAUAyq4AgO9A/gDOrgCA4Vz+ANKuAIDjVAEA1q4AgNquAIDergCA4q4AgLOZAQDmrgCA6q4AgO6uAIDyrgCAth0BALUdAQD2rgCAuz0BALo9AQD6rgCA/q4AgL/hAAC++QAAvfEAALz5AACoIQYAqVEGAKpRBgCrzQYArNUGAK3dBgCu1QYAr8kGAL6uAIACrwCABq8AgAqvAIAOrwCAEq8AgBavAIAarwCAuG0HALkFBwC6DQcAuwUHALwdBwC9AQcAvgEHAL8BBwCwuQYAsbkGALJtBwCzZQcAtH0HALVlBwC2ZQcAt1UHAKPZBgAerwCAIq8AgCavAIAqrwCApl0GAKVdBgCEnAIAq30GAKp9BgC+JAMALq8AgK+hBwCuuQcArbEHAKy5BwCASQAAgUkAAIJZAACzVQcAMq8AgLV9BwC2aQcANq8AgIZAAACHVAMAulUHALspBwC8OQcAvTkHAL4pBwC/IQcAo5kGADqvAIA+rwCAQq8AgEavAICmpQYApbEGAEqvAICr5QYAqpkGAE6vAIBSrwCAr+0GAK7lBgCt9QYArPUGAOE4BQBWrwCA4yQEAFqvAIBerwCAYq8AgGavAIBqrwCAbq8AgHKvAIB2rwCAeq8AgH6vAICCrwCA7/QEAIavAICo+QYAqQkGAKoRBgCrLQYArDkGAK0lBgCuLQYAryUGAIqvAICOrwCAkq8AgJavAICAGQAAgRkAAIIFAACarwCAuOUBALntAQC65QEAu/0BALzlAQC97QEAvuUBAL9ZAQCwXQYAsSEGALIhBgCzIQYAtCEGALUpBgC2EQYAtxEGAKjRAgCp2QIAqg0DAKsFAwCsHQMArQUDAK4FAwCvNQMAvmQCAKKvAICmrwCAqq8AgK6vAICyrwCAtq8AgLqvAIC4JQMAuS0DALolAwC7PQMAvCUDAL0pAwC++QMAv/kDALBNAwCxIQMAsiUDALM9AwC0JQMAtS0DALYlAwC3HQMAs4UDAITIAgC+rwCAhAgDAMKvAIC2hQMAtZUDAMavAIC75QMAuokDAIYIDACHnAMAv+kDAL7hAwC96QMAvPEDAIXsCgBHrgCAo80DAMqvAICl3QMAzq8AgNKvAICmzQMA1q8AgNqvAICrrQMAqsEDAK2hAwCsuQMAr6EDAK6pAwDerwCA4q8AgOavAIDqrwCA78gDAO6vAIDyrwCA9q8AgOO0AwD6rwCA4dABAP6vAICADQAAgXUAAIJ9AAACsACABrAAgAqwAICzZQEAvgQCALVlAQASsACAFrAAgLZlAQCGQA0Ah1gNALv1AQC6/QEAvaUBALy5AQC/mQEAvqUBABqwAIAesACAIrAAgIQADAAmsACAKrAAgC6wAIDvzAEAMrAAgOEsBgA2sACA4yABAOwAAAA6sACAPrAAgEKwAIBGsACAo+kBAEqwAIBOsACApukBAFKwAIBWsACApekBAKpxAQCreQEAWrAAgF6wAICuKQEArxUBAKw1AQCtKQEAqCUOAKktDgCqJQ4Aqz0OAKwlDgCtLQ4AriUOAK+VDgAOsACAYrAAgGawAIBqsACAbrAAgIKdAACBnQAAgJ0AALhFDwC5TQ8AukUPALtZDwC8SQ8AvUkPAL59DwC/cQ8AsPEOALH5DgCypQ4As7kOALSpDgC1lQ4Atp0OALd9DwCo1Q8Aqd0PAKoJDwCrCQ8ArBkPAK0FDwCuDQ8ArwUPAHKwAIB2sACAerAAgL6gAwB+sACAgrAAgId4AwCGEAAAuBUPALkdDwC6IQ8AuyEPALz1AAC9/QAAvvUAAL/tAACwQQ8AsU0PALJdDwCzVQ8AtE0PALU1DwC2MQ8AtzEPAIawAIDvsAwAirAAgI6wAICSsACAlrAAgJqwAICesACAorAAgKawAICqsACArrAAgLKwAIDjqA0AtrAAgOGMDQCzwQ4AurAAgL6wAIDCsACAxrAAgLbFDgC10Q4AyrAAgLvJDgC6xQ4AzrAAgNKwAIC/sQ4AvskOAL3BDgC8yQ4AowEOANawAIDasACA3rAAgOKwAICmBQ4ApREOAOawAICrCQ4AqgUOAOqwAICErAIAr3EOAK4JDgCtAQ4ArAkOAIBRAACBWQAAgmEAALPFAAC+zAEAtcUAALbNAADysACAhkAHAIcUAQC6yQAAu8kAALzZAAC92QAAvskAAL/FAACrDQMAqg0DAKkJAwCouQIArw0DAK4NAwCtDQMArA0DAL5gAwD2sACA+rAAgP6wAIACsQCABrEAgAqxAIC+MAUAuykDALoZAwC5GQMAuAEDAL/dAwC+3QMAvd0DALwxAwCzTQMAsk0DALFNAwCwTQMAtzkDALYxAwC1QQMAtE0DAA6xAICmkQMApZkDABKxAICjmQMAFrEAgBqxAIAesQCAr5kDAK6VAwCthQMArIUDAKuVAwCqlQMAnq8AgCKxAIAmsQCAKrEAgC6xAIAysQCANrEAgDqxAIA+sQCAQrEAgEaxAIBKsQCATrEAgFKxAICAHQAAgQkAAIL9AQBWsQCAvwgHAFqxAIBisQCA7yQAAGaxAICElAIAarEAgG6xAICH4AIAhgQFAL4AGABysQCAdrEAgOGQAQB6sQCA44AAAH6xAICCsQCAhrEAgLNlAQCKsQCAtWUBALZtAQCOsQCAkrEAgJaxAIC65QEAu/kBALzpAQC96QEAvsUBAL+9AQCasQCAnrEAgKKxAIC+xBkAprEAgKqxAICusQCA78gBALKxAIDh3A4AtrEAgOMwDgC6sQCAvrEAgMKxAICEMAQAgHkAAIEVAACCFQAAo+UBAMaxAICl5QEApu0BAMqxAICGQAYAh5AHAKplAQCreQEArGkBAK1pAQCuRQEArz0BAKjdBQCpIQYAqiEGAKshBgCsIQYArSEGAK4hBgCvnQYAXrEAgM6xAIDSsQCAhDABANaxAIDasQCA3rEAgOKxAIC4jQYAuZUGALqdBgC7lQYAvI0GAL21BgC+vQYAv7UGALDtBgCx8QYAsvEGALPxBgC0zQYAtbUGALa9BgC3tQYAqIkHAKmVBwCqkQcAq5EHAKy9BwCtpQcArqEHAK/dBwDmsQCA6rEAgO6xAIDysQCA9rEAgPqxAID+sQCAArIAgLhJBwC5VQcAul0HALtVBwC8cQcAvX0HAL5pBwC/aQcAsKUHALGtBwCyuQcAs7EHALSRBwC1kQcAtnkHALd5BwAGsgCACrIAgA6yAIASsgCA78gFAOHACQAWsgCA48AZAOMkBAAasgCA4dAGAO/cKACinQMAoxUBAKAZBQChjQUAs1kGAB6yAIAisgCAJrIAgCqyAIC2ZQYAtXUGAC6yAIC7KQYAuiEGADKyAIA2sgCAvxUGAL4VBgC9JQYAvC0GAKOZBgCPmfwAOrIAgEKyAIBGsgCApqUGAKW1BgBKsgCAq+kGAKrhBgCGKB8Ah5wAAK/VBgCu1QYAreUGAKztBgCebQkAn30HAJwNCwCd7QkAmvENAJs5DQCY5fAAmQ0PAJbh8QCX6fEAlMX1AJUN8wCSHfcAk/H1AJD9+QCR7fkAgh3/AIMB+gBOsgCAUrIAgIYV9gCHOfYAhAn6AIXx9ACKwfAAiyXyAFayAIBasgCAjuEMAI8VDgCMNfIAjQHzAJKtDgCTgQgAXrIAgGKyAICW6QQAl3UGAJR5CgCV8QoAmtEGAJvJAABmsgCAarIAgIEdAwCAHQMAnFkCAIL1AwCrARAAqpUWAKmNFgCojRYAr5UuAK4BLACt/RIArJkSAKOlHgCipR4AoY0CAO6wAICnGRoAppUaAKUBGACknR8AbrIAgHKyAIB2sgCAerIAgH6yAICCsgCAhrIAgIqyAICz5SoAsuUqALGtLwCw5S4AjrIAgJKyAIC1ASQAtBEqAKgpAwCpNQMAqj0DAKs1AwCsLQMArbUDAK69AwCvtQMAlrIAgJqyAICesgCAorIAgIAdAACBCQAAgrkAAKayAIC4TQIAuV0CALptAgC7CQIAvBkCAL0ZAgC+CQIAvwECALDNAwCx1QMAst0DALPVAwC0zQMAtXUCALZ9AgC3dQIAqrIAgITIHQCysgCAvgwfALayAIC6sgCA70gGAO9YBwDhWAYA4ZgGAOOUAQDjAAYAhhAcAId8HQC+9B4AvrIAgMKyAIC2ZQMAtfUDAMayAICz5QMAyrIAgM6yAIDSsgCAv+ECAL5ZAwC9UQMAvFkDALtBAwC6WQMA1rIAgNqyAIA+sgCArrIAgN6yAIDisgCA5rIAgOqyAIDusgCA8rIAgKitHQCptR0AqrUdAKslHgCsPR4ArR0eAK4VHgCvdR4AsA0eALEtHgCyJR4As40eALSVHgC1nR4AtpUeALeNHgC4tR4Aub0eALq1HgC7nR4AvIUeAL1VHwC+XR8Av1UfALMdHQD2sgCA+rIAgP6yAIACswCAtr0eALWVHgAGswCAu8keALrpHgAKswCADrMAgL95HgC+cR4AvXkeALzRHgCCKQAAo1kdAIAdAACBFQAApvkeABKzAIAWswCApdEeAKqtHgCrjR4AGrMAgITgAwCuNR4Arz0eAKyVHgCtPR4AqIkeAKmVHgCqnR4Aq7EeAKzRHgCt2R4Ars0eAK/FHgAeswCAIrMAgIaIAACHbAEAJrMAgCqzAIAuswCAMrMAgLhdAQC5wQEAusEBALvBAQC8wQEAvckBAL7xAQC/8QEAsL0eALGdHgCylR4As2UBALR9AQC1ZQEAtm0BALdlAQCqLR0AqzUdADazAIA6swCAri0dAK+VHACsLR0ArSUdAISMAQCjkR0APrMAgEKzAICmER0ARrMAgEqzAIClgR0As1UeAE6zAIBSswCAVrMAgFqzAIC2GR4AtRkeAF6zAIC7GR4AujkeAGKzAIBmswCAv+EBAL75AQC98QEAvAEeAGqzAIBuswCAcrMAgKOZHQB2swCApdUdAKbVHQB6swCAfrMAgIKzAICq9R0Aq9UdAKzNHQCtPQIArjUCAK8tAgCAZQAAgRUAAIIdAACEAAQAhrMAgIqzAICHcAMAhvwEAJKzAICWswCAmrMAgJ6zAICiswCAprMAgKqzAICuswCAvsgEALKzAIC2swCAurMAgL6zAIDCswCAxrMAgO/cHwDKswCA4ZQBAM6zAIDjHAEA0rMAgNazAIDaswCA3rMAgLt1AwC6aQMAvkgGAOKzAIC/HQMAvh0DAL0dAwC8ZQMAs9UDAOazAIDqswCA7rMAgPKzAIC2fQMAtcUDAIRwBQCoJQIAqTUCAKo9AgCrNQIArC0CAK2dAgCulQIAr7UCAIIVAAD2swCAgNkBAIEJAADEAAAA+rMAgAK0AIAGtACAuKkCALmpAgC6SQEAu0kBALxZAQC9RQEAvkUBAL99AQCwzQIAsdECALLRAgCzqQIAtLkCALW5AgC2qQIAt6ECAOEoHgDhNBwA43QBAOMYHgAKtACADrQAgIa4BACHVAUAhDgHABK0AIAWtACAGrQAgL6sBwAetACA78weAO/IGgCj9QIAIrQAgCa0AIAqtACALrQAgKZdAgCl5QIAMrQAgKtVAgCqSQIANrQAgDq0AICvPQIArj0CAK09AgCsRQIAqGEGAKlhBgCqYQYAq2EGAKxhBgCtYQYArmEGAK9hBgD+swCAPrQAgEK0AIBGtACASrQAgE60AIBStACAVrQAgLjxBgC58QYAuvEGALvxBgC8nQYAvbEGAL6xBgC/sQYAsOUGALHtBgCy5QYAs/0GALTlBgC17QYAttkGALfVBgCz6QYAWrQAgF60AIBitACAZrQAgLbhBgC16QYAarQAgLspBgC6IQYAbrQAgHK0AIC/KQYAviEGAL0pBgC8MQYAgl0AAKOtBgCARQAAgV0AAKalBgB2tACAerQAgKWtBgCqZQYAq20GAIYADACHQAMArmUGAK9tBgCsdQYArW0GAH60AIDvfAUAgrQAgIa0AICKtACAjrQAgJK0AICWtACAmrQAgJ60AICitACAprQAgKq0AIDjaAUArrQAgOF4BQCz0QYAsrQAgLa0AIC6tACAvrQAgLb9BgC1/QYAwrQAgLupBgC6oQYAxrQAgMq0AIC/mQYAvqkGAL2pBgC8sQYAqLkGAKm5BgCqGQYAqxkGAKw1BgCtPQYArjUGAK8pBgDOtACAgh0AAIEdAACAHQAA0rQAgNa0AIDatACA4rQAgLjpAQC56QEAuvkBALv5AQC86QEAvekBAL5dAQC/VQEAsCUGALEtBgCyJQYAsz0GALQtBgC1HQYAthUGALfZAQCGgAwAh+QCAOa0AICjnQUA6rQAgKWxBQCmsQUA7rQAgPK0AID2tACAqu0FAKvlBQCs/QUAreUFAK7lBQCv1QUAtk0DAPq0AICExAMAtUUDAP60AICzjQIAArUAgAa1AIC+SQMAv0kDALxJAwC9SQMAumkDALtpAwAKtQCADrUAgBK1AICmiQMApYEDABa1AICjSQIAGrUAgB61AIAitQCAr40DAK6NAwCtjQMArI0DAKutAwCqrQMAjrMAgCa1AIAqtQCALrUAgIW0PQAytQCANrUAgDq1AIA+tQCAQrUAgIA9AACBCQAAgh0AAEa1AIC+sAMASrUAgIc4AwCG3AwAUrUAgFa1AIBatQCAXrUAgGK1AIDvXAYAZrUAgGq1AIC+6AwA45QGAG61AIDh3AEAcrUAgHa1AIB6tQCAfrUAgLNRAQCCtQCAhrUAgIq1AICOtQCAtnEBALV5AQCStQCAuz0BALo9AQCWtQCAmrUAgL/9AQC+9QEAvQUBALwFAQCetQCAorUAgKa1AICEQAwAqrUAgK61AICytQCA76wHALa1AIDhJAYAurUAgONABwCGkAwAh/wMAMK1AIDGtQCAgFkAAIFlAACCYQAAo90BAMq1AICl9QEApv0BAM61AIDStQCA1rUAgKqxAQCrsQEArIkBAK2JAQCueQEAr3EBAN60AIBOtQCA2rUAgN61AIC+tQCA4rUAgOa1AIDqtQCAqJ0NAKktDgCqOQ4AqzEOAKwRDgCtEQ4Arn0OAK9tDgCwGQ4AsRkOALIxDgCzMQ4AtNEOALXZDgC2zQ4At8UOALj9DgC52Q4AuqkOALupDgC8vQ4AvaUOAL6tDgC/pQ4AqIEPAKmBDwCqgQ8Aq4EPAKyBDwCtjQ8AroUPAK+1DwDutQCA8rUAgPa1AID6tQCA/rUAgAK2AIAGtgCACrYAgLidDwC5rQ8AuqUPALtNDwC8VQ8AvV0PAL5JDwC/SQ8AsNEPALHRDwCy0Q8As9EPALS1DwC1vQ8AtrUPALetDwCzCQ4ADrYAgBK2AIAWtgCAGrYAgLYNDgC1CQ4AHrYAgLsVDgC6FQ4AIrYAgCa2AIC/eQ4AvnEOAL0FDgC8BQ4AghUAAKNNDgCAYQAAgWEAAKZJDgAqtgCAvhABAKVNDgCqUQ4Aq1EOAIQkAQAytgCArjUOAK89DgCsQQ4ArUEOAKg5DgCpOQ4AqlkOAKtRDgCscQ4ArXEOAK6RAQCvkQEAhgAAAIeEAAA2tgCAOrYAgD62AIBCtgCARrYAgEq2AIC4dQEAuX0BALp1AQC7yQAAvNkAAL3ZAAC+yQAAv8EAALD1AQCx/QEAsvUBALNNAQC0VQEAtV0BALZVAQC3TQEAuk0PALtVDwC4TQ8AuUUPAL59DwC/tQ8AvEUPAL11DwCyAQ8AswEPALAxDwCxMQ8AtgEPALcNDwC0EQ8AtREPAKqZDgCrRQ8AqOUOAKmZDgCuQQ8Ar0EPAKxRDwCtUQ8ATrYAgFK2AIBWtgCAWrYAgF62AIBitgCAZrYAgGq2AICzUQ0AbrYAgHK2AIB2tgCAerYAgLZxDQC1eQ0AfrYAgLu5AgC6sQIAgrYAgIa2AIC/GQIAvhECAL0ZAgC8oQIAirYAgKMVDQCOtgCAkrYAgKY1DQCWtgCAmrYAgKU9DQCq9QIAq/0CAIToAwCitgCArlUCAK9dAgCs5QIArV0CAKhtAgCprQIAqqUCAKu9AgCspQIAra0CAK6lAgCvfQEAgO0BAIHxAQCC8QEAvqAFAKa2AICqtgCAh2gFAIYcBQC4yQEAuckBALrZAQC70QEAvPkBAL35AQC+mQEAv5UBALAFAQCxDQEAsgUBALMdAQC0BQEAtQ0BALYFAQC3+QEA4WQPAOGcDwDjFA4A49QPAK62AIDhPA4AsrYAgOPkAAC+rAQAtrYAgLq2AIDvDAAAvrYAgMK2AIDvYA4A77QPAMa2AIDKtgCAhEQEALNhAgDOtgCAtWECALZhAgDStgCA1rYAgNq2AIC6jQEAu4UBALydAQC9hQEAvo0BAL+FAQCjrQUAnrYAgN62AIDitgCA5rYAgKatBQClrQUA6rYAgKtJBgCqQQYA7rYAgPK2AICvSQYArkEGAK1JBgCsUQYA9rYAgPq2AID+tgCAArcAgIAdAACBCQAAgjkAAAa3AIAKtwCADrcAgIbIAACHIAMAErcAgBa3AIAatwCAHrcAgKhtBgCptQcAqr0HAKsdBwCsCQcArTEHAK4xBwCvLQcAhKgDACK3AIAmtwCAKrcAgC63AIAytwCANrcAgDq3AIC4zQAAudUAALrVAAC75QAAvP0AAL2VAAC+nQAAv5UAALBVBwCxJQcAsi0HALM9BwC0LQcAtRUHALYdBwC39QAAPrcAgOG8BgBCtwCA4/QFAEa3AIBKtwCATrcAgFK3AIBWtwCAWrcAgF63AIBitwCAZrcAgGq3AIButwCA7+gEALN1BgCCLQAAgRUAAIAdAABytwCAtvEGALXBBgB2twCAu6EGALrRBgB6twCAvmwBAL+RBgC+qQYAvakGALy5BgCjtQYAgrcAgIYoAACHTAEAhrcAgKYxBgClAQYAircAgKthBgCqEQYAjrcAgJK3AICvUQYArmkGAK1pBgCseQYAlrcAgLO9AQCatwCAnrcAgLZ5AQCitwCAprcAgLV5AQC6VQEAu10BAKq3AICutwCAvvkAAL/lAAC8RQEAvf0AAKhxAgCpcQIAqnECAKtxAgCstQIArb0CAK61AgCvrQIAhOw8ALK3AIC2twCAurcAgL63AIDCtwCAxrcAgMq3AIC4XQMAuWUDALptAwC7ZQMAvH0DAL1lAwC+bQMAv2UDALDVAgCx3QIAstUCALNtAwC0eQMAtWUDALZtAwC3ZQMALrYAgM63AIDStwCAo/UCANa3AIClMQIApjECANq3AIDetwCA4rcAgKodAgCrFQIArA0CAK21AwCusQMAr60DAIBlAACBCQAAghkAAOa3AIDqtwCA8rcAgL4QPAD2twCAhsA8AIcgAwD6twCA/rcAgAK4AIAGuACACrgAgA64AICohQIAqZUCAKqVAgCrpQIArL0CAK3VAgCu0QIAr9ECABK4AIAWuACAGrgAgB64AIAiuACAJrgAgCq4AIAuuACAuHUBALl9AQC6dQEAu8kBALzZAQC9xQEAvsUBAL/9AQCwtQIAsb0CALKBAgCzgQIAtFUBALVdAQC2VQEAt00BAOGkBgAyuACA41AGAL6APACEHDwAvoA/ADa4AIA6uACAPrgAgEK4AIBGuACASrgAgE64AIBSuACA7+AGAFa4AICBfQAAgHEAAFq4AICCBQAAYrgAgGa4AIDvTAAAargAgOGQAQBuuACA41gBAHK4AIB2uACAergAgIZYPwCH/DwAs509AO63AIBeuACAfrgAgIK4AIC21T0AtbU9AIa4AIC7+T0AuvE9AIq4AICOuACAvxk+AL4RPgC91T0AvNU9AJK4AICj2T0AlrgAgJq4AICmkT0AnrgAgKK4AICl8T0AqrU9AKu9PQCmuACAqrgAgK5VPgCvXT4ArJE9AK2RPQCoVT4AqVk+AKphPgCrYT4ArGE+AK1hPgCuYT4Ar2E+AISoAwCuuACAsrgAgLa4AIC6uACAvrgAgMK4AIDGuACAuEU/ALldPwC6VT8Au20/ALx1PwC9fT8AvnU/AL9tPwCwwT8AscE/ALLBPwCzwT8AtME/ALXBPwC2wT8At8E/AIC5AQCBuQEAggUAAMq4AIDhgD4A0rgAgOMoPQDWuACAhoAAAIcEAQDvCD0A2rgAgN64AIDiuACA5rgAgOq4AICzqT8AzrgAgO64AIDyuACA9rgAgLahPwC1qT8A+rgAgLtFPgC6RT4A/rgAgAK5AIC/RT4AvkU+AL1VPgC8VT4Ao2k/AAa5AIAKuQCADrkAgBK5AICmYT8ApWk/ABa5AICrhT4AqoU+ABq5AIAeuQCAr4U+AK6FPgCtlT4ArJU+ACK5AICzGT4AJrkAgCq5AIC2IT4ALrkAgDK5AIC1MT4AuvEBALv5AQA2uQCAOrkAgL6xAQC/vQEAvNEBAL3RAQCo0T0AqdE9AKrVPQCr6T0ArP09AK3lPQCu7T0ArxECAID5AwCBzQMAgsUDAIQkAwC+AAQAQrkAgIesAwCGvAQAuBkCALktAgC6JQIAu+kCALz5AgC9+QIAvukCAL/pAgCwcQIAsXkCALJBAgCzQQIAtDECALU9AgC2NQIAtykCAKVtPQBGuQCASrkAgKZ9PQBOuQCAfrcAgKNFPQBSuQCArY0CAKyNAgCv4QIAru0CAKwAAABWuQCAq6UCAKqtAgDh+AEAWrkAgOP0AgCEwAQAXrkAgGK5AIBmuQCAarkAgG65AIByuQCAdrkAgHq5AIB+uQCAgrkAgO8wAgCGuQCAqBUCAKkZAgCqJQIAqz0CAKwlAgCtLQIAriUCAK9VAgCKuQCAjrkAgJK5AICWuQCAmrkAgJ65AICEsAQAorkAgLjRAgC52QIAuuECALvhAgC8kQIAvZ0CAL6VAgC/iQIAsC0CALE1AgCyNQIAswUCALQdAgC18QIAtvECALfxAgDheD8A4zQBAOMIPgDhbD4AgQkAAICpAACmuQCAgj0AAKq5AICyuQCAtrkAgL4gBAC6uQCA79g+AO/MPgC+uQCAwrkAgLPpAgCG6AQAh8AEALbpAgDGuQCAyrkAgLXpAgC6rQIAu7UCAM65AIDSuQCAvp0CAL9xAgC8pQIAvZUCAD65AICuuQCA1rkAgNq5AIDeuQCA4rkAgOa5AIDquQCAqBUGAKmhBgCqoQYAq70GAKytBgCtgQYArv0GAK/tBgCwlQYAsZ0GALKVBgCzrQYAtLUGALW9BgC2tQYAt60GALiVBgC5mQYAukkHALtJBwC8WQcAvVkHAL5JBwC/SQcArN0FAK3tBQCu5QUArwkFAO65AIDyuQCAqtUFAKvNBQD2uQCApZEFAKaRBQD6uQCA/rkAgAK6AIAGugCAo5EFALNJBgAKugCADroAgBK6AIAWugCAtmEGALVFBgAaugCAuzkGALoxBgC+ZAAAHroAgL8ZBgC+EQYAvRkGALwhBgCjiQcAgtkBAIHZAQCAwQEAIroAgKahBwClhQcAJroAgKv5BwCq8QcAhggBAId8AQCv2QcArtEHAK3ZBwCs4QcAKroAgLP1BgAuugCAMroAgLaFBgA2ugCAOroAgLWdBgC6jQYAu20BAD66AIBCugCAvmUBAL9tAQC8dQEAvW0BAKglBgCpLQYAqjkGAKsxBgCsUQYArUEGAK5BBgCvdQYARroAgEq6AIBOugCAUroAgFa6AIBaugCAXroAgGK6AIC4VQEAuWUBALplAQC7fQEAvGUBAL1tAQC+HQEAvxUBALANBgCx7QEAsuUBALP9AQC05QEAte0BALblAQC3bQEAo7EFAGa6AIBqugCAvkgDAL5YDACmwQUApdkFAG66AICrKQIAqskFAHK6AIB2ugCArykCAK4hAgCtKQIArDECAHq6AIB+ugCAgroAgIa6AICAGQAAgRkAAIIFAACKugCAhKwDAJK6AICHGAMAhswMAJa6AICaugCAnroAgKK6AICokQMAqZkDAKrJAwCrxQMArN0DAK3BAwCuwQMAr/UDAKa6AICqugCArroAgLK6AIC2ugCAuroAgL66AIDCugCAuH0DALnBAAC6wQAAu9EAALz5AAC9+QAAvpkAAL+ZAACwjQMAsUUDALJNAwCzRQMAtF0DALVFAwC2TQMAt0UDALNBAgDGugCAyroAgL8EDwDOugCAtkECALVVAgDSugCAu4ECALpJAgDWugCA2roAgL+BAgC+mQIAvZECALyZAgDeugCA4roAgOa6AIDqugCA76QDAO66AIDyugCA9roAgOMQAwD6ugCA4VgAAIQgDQCAKQAAgSkAAIIdAAACuwCA4VAGAOGgBwDjoAYA41AHAIWUDAAGuwCA70gbAAq7AIDhJAIADrsAgONwGgASuwCAFrsAgBq7AIDvqAEA7+gGAIagDwCHDA0Ao4kCAB67AIClnQIAIrsAgCa7AICmiQIAKrsAgC67AICrSQIAqoECAK1ZAgCsUQIAr0kCAK5RAgCoZQ4AqXUOAKp9DgCrdQ4ArG0OAK21DgCuvQ4Ar7UOAP66AIAyuwCANrsAgDq7AIA+uwCASbsAgE27AIBRuwCAuF0PALltDwC6ZQ8Auw0PALwVDwC9HQ8AvhUPAL8JDwCwzQ4AsdUOALLdDgCz1Q4AtM0OALVxDwC2cQ8At20PALP1DgBVuwCAWbsAgF27AIBhuwCAtjUOALXlDgBluwCAuxEOALoJDgBpuwCAbbsAgL+1DwC+CQ4AvQEOALwJDgCCFQAAo7EOAIBhAACBYQAApnEOAHG7AIC+EAEApaEOAKpNDgCrVQ4AebsAgIQgAQCuTQ4Ar/EPAKxNDgCtRQ4An0UIAJ4NCQCdDQkAnJkLAJt1NQCaETUAmZk3AJgNMQCXJTEAliUxAJWBPQCUDT0Ak4k/AJIVOACRPTkAkD05AI9lJQDvrA0AhgAEAIegAQB9uwCAgbsAgIW7AIDv6AEAibsAgOE0AgCNuwCA4zQBAJG7AIDjCAwAlbsAgOEIDQChoQEAmbsAgKMJBQCibQMApc0EAKQRBQCnHRkAph0ZAKmhHQCoORkAq+kcAKqpHQCtkREArAEQAK8BFACuUREAsfkVALDlFQCz6WkAsgFoALUBbAC0eWkAnbsAgKG7AICluwCAqbsAgK27AICxuwCAowkDAKIZDQCh/Q0AoP0NAIIlJgCDBToAtbsAgLm7AICGqTwAhzU+AIQdOgCFPTsAiok+AIslMgC9uwCAwbsAgI6xNACPMTYAjD0yAI0tMgCSJTYAk9EIAIREAwC+wAQAlhULAJdVDgCUXQoAlVUKAJplDgCbiQ4AxbsAgMm7AIDNuwCA0bsAgJyBAADVuwCAuLUCALm9AgC6tQIAuwkCALwZAgC9GQIAvgkCAL8BAgCwdQ0AsX0NALJJDQCzSQ0AtJUCALWdAgC2lQIAt40CAKi9DQCpUQ0AqlUNAKtpDQCsfQ0ArWUNAK5tDQCvEQ0AdbsAgILtAQCBHQAAgB0AANm7AIDduwCAjroAgL5wBQCznQwAhIwFAOG7AIDpuwCA7bsAgLalDAC1tQwA8bsAgLv5DAC68QwAhigFAIcgBQC/GQMAvhEDAL3dDAC83QwA9bsAgKPZDAD5uwCA/bsAgKbhDAABvACABbwAgKXxDACqtQwAq70MAAm8AIANvACArlUDAK9dAwCsmQwArZkMABG8AIAVvACAGbwAgB28AIAhvACAJbwAgCm8AIDvvAEALbwAgOF8DgAxvACA41ABADW8AIA5vACAPbwAgEG8AICzlQIARbwAgEm8AIBNvACAUbwAgLa9AgC1uQIAWbwAgLs5AgC6YQIAhsgEAIesBAC/GQIAvhECAL0ZAgC8IQIAo1UFAILVBwCBxQcAgMUHAF28AICmfQUApXkFAGG8AICr+QUAqqEFAGW8AIBpvACAr9kFAK7RBQCt2QUArOEFAG28AICzWQcAcbwAgHW8AIC2HQcAebwAgH28AIC1FQcAugkHALsJBwCBvACAhbwAgL75BwC/+QcAvPkHAL35BwDluwCAVbwAgIm8AICNvACAkbwAgJW8AICZvACAnbwAgKitBwCptQcAqrUHAKvtBwCs+QcArfkHAK7tBwCv5QcAsKkHALGpBwCySQcAs0kHALRZBwC1WQcAtkkHALdJBwC4eQcAuUUHALpBBwC7XQcAvEUHAL1NBwC+RQcAvzkHAKMdBgChvACApbwAgKm8AICtvACAplkGAKVRBgCxvACAq00GAKpNBgC1vACAubwAgK+9BgCuvQYArb0GAKy9BgCAbQAAgQkAAIIZAAC9vACAwbwAgISYAQC+kAEAxbwAgIYAHACHxAEAybwAgM28AIDRvACA1bwAgNm8AIDdvACAqF0GAKmVAQCqlQEAq6UBAKy9AQCt1QEArtEBAK/RAQDhvACA5bwAgOm8AIDtvACA8bwAgPW8AID5vACA/bwAgLhZAQC5WQEAus0AALvFAAC83QAAvcUAAL7FAAC/9QAAsLUBALG9AQCygQEAs4EBALR5AQC1eQEAtmkBALdpAQCzHQIAAb0AgAW9AIC+gBwACb0AgLZVAgC1NQIADb0AgLt5AgC6cQIAEb0AgBW9AIC/vQIAvr0CAL1VAgC8VQIAGb0AgKNZAgAdvQCAIb0AgKYRAgAlvQCAKb0AgKVxAgCqNQIAqz0CAC29AIAxvQCArvkCAK/5AgCsEQIArRECADm9AIA9vQCAvgQdAL4AHgBBvQCARb0AgEm9AIBNvQCAgPkAAIHNAACCxQAAhCADAIawHACHlAMAUb0AgFW9AIBZvQCAXb0AgGG9AIBlvQCA42wCAGm9AIDhoAEAbb0AgO8UAgBxvQCAdb0AgHm9AIB9vQCAgb0AgIW9AICJvQCA4fAGAOE0BgDjTAAA4xgGAI29AICRvQCAlb0AgJm9AICAPQAAgQkAAIIZAACdvQCAob0AgIS8HQDvmAAA7zgHALMxAgDRAAAAh9gdAIZsHACpvQCAtikCALUhAgCtvQCAu80CALrNAgCxvQCAtb0AgL/NAgC+zQIAvc0CALzNAgCyXQYAs2UGALANBgCxVQYAtn0GALedBQC0fQYAtXUGALqNBQC7zQUAuKUFALmFBQC+xQUAv8kFALzVBQC9zQUAub0AgL29AIDBvQCAxb0AgMm9AIDNvQCA0b0AgNW9AICqtQYAq70GAKgBBwCpvQYAroEGAK+NBgCsmQYArZUGAKNxHQDZvQCA3b0AgOG9AIDlvQCApmkdAKVhHQDpvQCAq40dAKqNHQDtvQCA8b0AgK+NHQCujR0ArY0dAKyNHQD1vQCAs9UeAPm9AID9vQCAts0eAAG+AIAFvgCAtcUeALqhHgC7oR4ACb4AgA2+AIC+pR4Av6keALyxHgC9sR4ANb0AgKW9AIARvgCAhAQDAID5AACB+QAAghEAABW+AICoIR4AqSEeAKo5HgCrOR4ArCkeAK0pHgCuAR4ArwEeALABHgCxAR4AsgEeALMBHgC0BR4AtQkeALY9HgC3NR4AuA0eALkVHgC6HR4AuxUeALwNHgC95R8Avu0fAL/lHwCjkR8AGb4AgIYoAQCHSAEAHb4AgKaJHwClgR8AIb4AgKvlHwCq5R8AJb4AgCm+AICv7R8AruEfAK31HwCs9R8ALb4AgLMtHgAxvgCANb4AgLaVHgA5vgCAPb4AgLWdHgC6sR4Au7EeAEG+AIBFvgCAvnUBAL99AQC8oR4AvaEeAKjRHgCp2R4AquEeAKvhHgCsUR4ArVEeAK5RHgCvUR4ASb4AgE2+AIBRvgCAVb4AgFm+AIBdvgCAYb4AgGW+AIC43QEAue0BALrlAQC7jQEAvJkBAL2ZAQC+jQEAv4UBALAxHgCxMR4AsjEeALMxHgC09QEAtf0BALb1AQC37QEAo2kdAGm+AIBtvgCAcb4AgHW+AICm0R0ApdkdAHm+AICr9R0AqvUdAH2+AICBvgCArzkCAK4xAgCt5R0ArOUdAIFpAACAWQAAvgAEAIJhAACJvgCAjb4AgJG+AICVvgCAhOwDAJm+AICHiAMAhuwEAJ2+AIChvgCApb4AgKm+AICohQMAqZUDAKqVAwCrpQMArL0DAK3VAwCu0QMAr9EDAK2+AICxvgCAtb4AgLm+AIC9vgCAwb4AgMW+AIDJvgCAuHEDALlxAwC6cQMAu3EDALzVAAC93QAAvtUAAL/NAACwtQMAsb0DALKBAwCzgQMAtFEDALVRAwC2UQMAt1EDAOFUHgDhrB8A45QBAOMoHgDjYAMAzb4AgOEIAADRvgCA75ADANW+AIDZvgCA3b4AgOG+AIDlvgCA70wfAO9MHwCzXQIA6b4AgO2+AIDxvgCA+b4AgLYVAgC1dQIA/b4AgLs5AgC6MQIAhCQFAL7gBAC/1QIAvtUCAL0VAgC8FQIAuJEdALmZHQC6oR0Au6EdALzRHQC93R0AvtUdAL/JHQCwCR4AsQkeALIZHgCzGR4AtAkeALUJHgC2vR0At7UdAKipHgCpqR4AqrkeAKu5HgCsqR4ArakeAK55HgCveR4AgKUAAIGtAACCpQAAAb8AgIbQBACH+AQABb8AgAm/AICFvgCA9b4AgA2/AIARvwCAFb8AgBm/AIAdvwCAIb8AgKhxBgCpcQYAqnEGAKtxBgCsVQYArUUGAK5NBgCvRQYAsD0GALHlBgCy7QYAs+UGALT9BgC15QYAtu0GALflBgC43QYAuXEHALp1BwC7SQcAvFkHAL1ZBwC+SQcAv0kHALPZBgAlvwCAKb8AgC2/AIAxvwCAtuUGALX9BgA1vwCAuwEGALrZBgA5vwCAPb8AgL8BBgC+GQYAvREGALwZBgBBvwCAo9kFAEW/AIBJvwCAppEFAE2/AIBRvwCApfEFAKq1BQCrvQUAVb8AgFm/AICuUQUAr1EFAKyRBQCtkQUAo1kHAIIZAACBGQAAgOEBAF2/AICmZQcApX0HAGG/AICrgQcAqlkHAISgAgC+rAEAr4EHAK6ZBwCtkQcArJkHAGW/AICzqQYAhugAAIcsAQC2WQEAab8AgG2/AIC1oQYAunUBALt9AQBxvwCAdb8AgL75AQC/+QEAvGUBAL35AQCo0QYAqdkGAKplBgCrdQYArG0GAK2dAQCulQEAr40BAITsAQB5vwCAfb8AgIG/AICFvwCAib8AgI2/AICRvwCAuGkBALlpAQC6CQEAuwUBALwdAQC9AQEAvgEBAL81AQCw9QEAsf0BALL1AQCzaQEAtHkBALV5AQC2aQEAt2EBAJW/AICZvwCAnb8AgKPhBQChvwCApekFAKYRAgClvwCAqb8AgK2/AICqPQIAqzUCAKwtAgCtsQIArrECAK+xAgCxvwCAtb8AgL4EAwCEAAwAub8AgL2/AIDBvwCAxb8AgIANAACBFQAAgh0AAMm/AIDNvwCA0b8AgIdEAwCG3AwAs+kDANm/AIDdvwCA4b8AgOW/AIC2PQMAtT0DAOm/AIC7GQMAuhEDAO2/AIDxvwCAv7kAAL6xAAC9uQAAvAEDAPW/AIDhlAEA+b8AgON8AQD9vwCAAcAAgAXAAIAJwACADcAAgBHAAIAVwACAGcAAgB3AAIAhwACAJcAAgO9MAgCoVQIAqV0CAKphAgCrYQIArLUCAK29AgCutQIAr60CAL5oDQApwACALcAAgDHAAIA1wACAgq0AAIGtAACArQAAuGEBALlhAQC6CQEAuwkBALwBAQC9AQEAvgEBAL8BAQCw1QIAsd0CALLVAgCzbQEAtHUBALV9AQC2aQEAt2EBAOFoBgDh8AcA47AAAOP0BgA5wACAPcAAgEHAAIBJwACATcAAgFHAAIBVwACAWcAAgL78DABdwACA72wAAO8oBgCjqQIAYcAAgIZoDACHBA0AZcAAgKZ9AgClfQIAacAAgKtZAgCqUQIAbcAAgHHAAICv+QEArvEBAK35AQCsQQIAqIUOAKmNDgCqhQ4Aq50OAKyNDgCtvQ4ArrUOAK/dDgBFwACAdcAAgHnAAIB9wACAgcAAgIXAAICJwACAjcAAgLitDgC5tQ4Aur0OALu1DgC8dQ8AvX0PAL51DwC/bQ8AsKkOALG1DgCyvQ4As7UOALStDgC1lQ4Atp0OALeVDgCzDQ4AkcAAgJXAAICZwACAncAAgLY9DgC1BQ4AocAAgLtxDgC6bQ4ApcAAgKnAAIC/UQ4AvmkOAL1hDgC8aQ4AghkAAKNJDgCAZQAAgRkAAKZ5DgCtwACAscAAgKVBDgCqKQ4AqzUOAIS8AwC1wACAri0OAK8VDgCsLQ4ArSUOAKidDgCppQ4Aqq0OAKulDgCsvQ4AraEOAK7dDgCvzQ4AhiABAIdkAQC5wACAvcAAgMHAAIDFwACAycAAgM3AAIC4eQEAuXkBALrNAQC7xQEAvN0BAL3FAQC+xQEAv/UBALC9DgCxjQ4AsoUOALNJAQC0WQEAtVkBALZJAQC3SQEAtS0OANHAAIDVwACAtjkOANnAAIDdwACAsz0OAOHAAIC9hQEAvEkOAL+FAQC+hQEA5cAAgNW/AIC7UQ4AumEOAKNlDgDpwACA7cAAgPHAAID1wACApmEOAKV1DgD5wACAqwkOAKo5DgD9wACAAcEAgK/dAQCu3QEArd0BAKwRDgAFwQCACcEAgO/QDwANwQCAEcEAgBXBAIAZwQCAHcEAgCHBAIC+aAMAKcEAgC3BAIDhVA4AMcEAgONkDgA1wQCAgFkAAIFZAACCaQAAhIwDAIbwBACHFAMAOcEAgD3BAIBBwQCARcEAgEnBAIBNwQCAUcEAgFXBAIBZwQCAXcEAgGHBAIBlwQCAacEAgG3BAIBxwQCAdcEAgHnBAIB9wQCAqIkDAKmJAwCqmQMAq5kDAKyJAwCtiQMArj0DAK81AwCwUQMAsVEDALJVAwCzfQMAtBUDALUdAwC2FQMAtw0DALg9AwC5DQMAugUDALvtAAC89QAAvfkAAL7pAAC/6QAAgcEAgIXBAICJwQCAsz0CAI3BAIC1LQIAtiUCAJHBAIC+aAUAmcEAgLq5AgC7uQIAvK0CAL2FAgC+/QIAv/UCAIBJAACBVQAAglUAAIQABQDvjAMAvhgEAId0BQCG/AQA4zwDAJ3BAIDhUAAAocEAgKXBAICpwQCArcEAgLHBAIC1wQCAucEAgL3BAIDBwQCAxcEAgMnBAIDNwQCA79QOAL4oBgDhdA4A0cEAgONUAQDVwQCA2cEAgN3BAIDhwQCAo/ECAOXBAIDpwQCA7cEAgPHBAICm6QIApeECAPXBAICrdQIAqnUCAPnBAID9wQCArzkCAK4xAgCtSQIArGECAKgpBgCpKQYAqj0GAKsxBgCsSQYArUkGAK55BgCveQYAlcEAgIIVAACBxQcAgMUHAAHCAICEaAMABcIAgAnCAIC4yQYAuckGALrZBgC72QYAvMkGAL3JBgC+WQcAv1kHALAJBgCxCQYAshkGALMZBgC0CQYAtQkGALb5BgC3+QYAs7UGAA3CAICGrAAAh0ADABHCAIC2yQYAtcEGABXCAIC7zQYAus0GABnCAIAdwgCAv80GAL7NBgC9zQYAvM0GACHCAICj8QYAJcIAgCnCAICmjQYALcIAgDHCAIClhQYAqokGAKuJBgA1wgCAOcIAgK6JBgCviQYArIkGAK2JBgCoJQYAqWEGAKplBgCrfQYArGUGAK1tBgCuZQYAr50GAD3CAIBBwgCARcIAgEnCAIBNwgCAUcIAgFXCAIBZwgCAuPUGALn9BgC69QYAu4kGALyZBgC9mQYAvokGAL+BBgCw5QYAse0GALLlBgCz/QYAtOUGALXtBgC20QYAt80GAF3CAIC2/QYAtf0GAGHCAICz/QYAZcIAgGnCAIBtwgCAvzkGAL4xBgC9OQYAvCEGALs5BgC6MQYAJcEAgHHCAICjrQYAgnkAAIFVAACAVQAAhFwBAKatBgClrQYAecIAgKtpBgCqYQYAhkh/AIfkAACvaQYArmEGAK1pBgCscQYAfcIAgO/cBwCBwgCAhcIAgInCAICNwgCAkcIAgJXCAICZwgCAhKADAJ3CAIC/JHkAocIAgONoBwClwgCA4XQGALPRAgCpwgCAvgQDAISAfQCtwgCAtvkCALXxAgCxwgCAu7UCALqpAgC1wgCAucIAgL9RAwC+mQIAvZECALylAgCpBQIAqLkCAKsVAgCqHQIArT0CAKw9AgCvUQIArl0CAL5ofQC9wgCAwcIAgMXCAIDJwgCAzcIAgNHCAIDVwgCAufEDALjpAwC78QMAuvkDAL1RAwC86QMAv00DAL5RAwCxNQIAsCkCALMBAgCyNQIAtdEDALQZAgC30QMAttkDAIIpAACjlQMAgB0AAIEVAACmvQMA2cIAgN3CAICltQMAqu0DAKvxAwDhwgCA6cIAgK7dAwCvFQIArOEDAK3VAwCGYH0Ah3h9ALNBAQCEAH8AtUEBAO3CAIDxwgCAtkkBAPXCAID5wgCAu0EBALpNAQC9SQEAvEUBAL8pAQC+OQEA/cIAgO/cBgABwwCABcMAgAnDAIANwwCAEcMAgO8wBgCELH4A4eAGABXDAIDjiAEAGcMAgON0AAAdwwCA4SwBAKPJAQAhwwCAJcMAgIVweQApwwCApsEBAKXJAQAtwwCAq8kBAKrFAQAxwwCANcMAgK+hAQCusQEArcEBAKzNAQCo3X0AqQV+AKoBfgCrAX4ArAF+AK0BfgCuAX4ArwF+AOXCAIA5wwCAPcMAgEHDAIBFwwCAgp0AAIGdAACAnQAAuC1+ALnhfgC64X4Au+F+ALzhfgC94X4AvuF+AL/hfgCwQX4AsU1+ALJZfgCzVX4AtDV+ALUlfgC2JX4AtxV+AKitfwCp0X8AqtF/AKvtfwCs9X8ArRV/AK4RfwCvEX8AScMAgE3DAIBRwwCAVcMAgIbwAwCHuAAAWcMAgF3DAIC4EX8AuRl/ALohfwC7IX8AvPUAAL39AAC+9QAAv+0AALBxfwCxcX8AsnF/ALNFfwC0QX8AtU1/ALY9fwC3NX8As1l+AGHDAIBlwwCAacMAgG3DAIC2lX4AtX1+AHHDAIC7tX4AurV+AHXDAIB5wwCAv4l+AL6FfgC9kX4AvKV+AH3DAICjHX4AgcMAgIXDAICm0X4AicMAgI3DAIClOX4AqvF+AKvxfgCRwwCAlcMAgK7BfgCvzX4ArOF+AK3VfgCwrQAAscUAALLBAACzwQAAtMUAALXNAAC28QAAt/EAALhhAAC5YQAAumEAALt9AAC8ZQAAvW0AAL5lAAC/vQMAmcMAgJ3DAIChwwCAdcIAgKXDAICpwwCArcMAgLHDAICoWQEAqVkBAKrtAACr5QAArP0AAK3lAACu5QAAr9UAALXDAICCHQAAgR0AAIAdAAC5wwCAvcMAgMHDAIC+VAIAhoAEAIfsAgDJwwCAzcMAgNHDAIDVwwCA2cMAgL54AwDjdH4A3cMAgOG4fQDhwwCA5cMAgOnDAIDtwwCA8cMAgPXDAID5wwCA/cMAgAHEAIDvwH4ABcQAgAnEAIANxACAs4UDABHEAIAVxACAGcQAgB3EAIC2hQMAtZUDACHEAIC74QMAuokDAL4kBgAlxACAv+kDAL7hAwC99QMAvPUDAIIpAACjwQMAgB0AAIEVAACmwQMAKcQAgC3EAICl0QMAqs0DAKulAwAxxACAheAFAK6lAwCvrQMArLEDAK2xAwDh+AMAOcQAgONcHwA9xACA7/QDAEHEAICGPAcAh6wCAON8fgBFxACA4YABAEnEAIBNxACAUcQAgO/kEwBVxACAs3EBAFnEAIBdxACAYcQAgGXEAIC2EQEAtWEBAGnEAIC7OQEAujEBAG3EAIBxxACAvxkBAL4RAQC9GQEAvCEBAHXEAIB5xACAfcQAgIHEAICFxACAicQAgI3EAIDvxH8AkcQAgOH8fgCVxACA4/B/AIANAACBdQAAgn0AAJnEAICdxACAocQAgKP5AQC+AAgApekBAKnEAICtxACAppkBAISoBQCxxACAq7EBAKq5AQCtkQEArKkBAK+RAQCumQEAqCkGAKkpBgCqOQYAqzkGAKwpBgCtUQYArlUGAK9NBgA1xACAhCABALXEAIClxACAo+EBAKKZBAChGQQAoPEFALg5BgC5OQYAus0GALvFBgC83QYAvcUGAL7FBgC/8QYAsDUGALE9BgCyNQYAsw0GALQVBgC1HQYAthUGALcJBgCPoWwAs5EHAIYoAQCHfAMAtqEHALnEAIC9xACAtbEHALrlBwC77QcAwcQAgMXEAIC+7QcAv90HALz1BwC97QcAn/l4AJ7leACdcXkAnCF8AJvxfACaYX0AmZlxAJjZcACX4XAAlnl0AJVtdACUbXQAk61pAJJxaACReWgAkB1uAIIhbQCD5W8AycQAgM3EAICGTWgAh5V1AISZaQCFmWkAiqV1AIu5dQDRxACA1cQAgI5xcACPgXwAjDlxAI05cQCSYX0Ak6l9ANnEAIDdxACAlml5AJeZBACU4XgAlX15AJpBBQCbyQUA4cQAgOXEAIDpxACA7cQAgJypAADxxACAo4ENAKKpAQChqQEA9cQAgKexCQCmAQgApU0NAKSZDQCrkRUAqoUVAKkBFACocQkArx0QAK7pEQCtvREArAEQALMBGACy8RwAscEdALDJHQDFwwCA+cQAgLXhGAC0/RkA/cQAgAHFAIAFxQCACcUAgIAdAACBCQAAgv0DAA3FAICjFQUAEcUAgIaIDACHPAMAGcUAgKYlBQClNQUAHcUAgKtpBQCqYQUAIcUAgCXFAICvWQUArmkFAK1pBQCscQUAKcUAgC3FAICEBAwAMcUAgDXFAIDhbAYAOcUAgOPsewA9xQCAQcUAgEXFAIDvqAYAScUAgE3FAIBRxQCAVcUAgKmNBQCogQUAq60FAKqZBQCtoQUArLkFAK+lBQCuqQUAhGgNAFnFAIBdxQCAYcUAgGXFAIBpxQCAbcUAgL70DAC5SQUAuEEFALtZBQC6QQUAvUkFALxBBQC/cQUAvn0FALGpBQCwoQUAs7kFALKhBQC1mQUAtKkFALd5BQC2kQUAqNUEAKndBACq7QQAqyUDAKyFAwCtjQMArrEDAK+xAwBxxQCAdcUAgHnFAIB9xQCAgBkAAIEZAACCBQAAgcUAgLgxAgC5MQIAujUCALvBAgC8hQIAvbUCAL69AgC/tQIAsGkCALFpAgCyQQIAs0ECALQ5AgC1OQIAthECALcRAgCGoAwAh0wNAInFAICNxQCA76QGAJHFAICVxQCA78wHAOOUAQDhpAYA4TgBAONcBgCZxQCAncUAgKHFAIClxQCAqcUAgK3FAICzLQQAscUAgLVFAwC1xQCAucUAgLZFAwC9xQCAwcUAgLvlAgC65QIAvd0CALzdAgC/tQIAvrUCABXFAICFxQCAxcUAgMnFAIDNxQCA0cUAgNXFAIDZxQCAqDEOAKk5DgCqAQ4AqwEOAKxxDgCtcQ4ArnUOAK9tDgCwGQ4AsSUOALItDgCzJQ4AtCEOALUhDgC2IQ4AtyEOALjFDgC5zQ4AusUOALvdDgC8xQ4Avc0OAL5ZDwC/WQ8As6kOAN3FAIDhxQCA5cUAgOnFAIC20Q4AtdkOAO3FAIC7wQ4Auv0OAPHFAIC+LAAAv8UOAL7FDgC90Q4AvNkOAIJpAACj7Q4AgFkAAIFRAACmlQ4A9cUAgPnFAIClnQ4AqrkOAKuFDgCGyAAAh6wAAK6BDgCvgQ4ArJ0OAK2VDgD9xQCAs5EOAAHGAIAFxgCAtqUOAAnGAIANxgCAta0OALrhDgC74Q4AEcYAgBXGAIC+6Q4Av9UOALz1DgC96Q4Ao6UKABnGAIAdxgCAIcYAgCXGAICmzQ0Apc0NACnGAICrbQwAqm0MAC3GAIAxxgCArz0MAK49DACtVQwArFUMAKgJDgCpCQ4Aqh0OAKsVDgCsIQ4ArSEOAK4hDgCvIQ4ANcYAgDnGAIA9xgCAQcYAgEXGAIBJxgCATcYAgFHGAIC4zQEAudUBALrdAQC71QEAvM0BAL1RAQC+UQEAv1EBALAhDgCxIQ4AsiUOALM5DgC0KQ4AtRUOALYdDgC39QEAVcYAgFnGAIBdxgCAo5kNAGHGAIClpQ0Apq0NAL7cAgCE7AMAacYAgKrpDQCr6Q0ArP0NAK3hDQCu4Q0Ar90NAIBFAACBTQAAglkAAKNFAwBtxgCApUEDAKZBAwBxxgCAhsAEAIcAAwCqLQMAqyUDAKw9AwCtJQMAriUDAK8VAwCoWQIAqYUDAKqBAwCrgQMArIUDAK2NAwCusQMAr7EDAHXGAIB5xgCAfcYAgIHGAICFxgCAicYAgI3GAICRxgCAuGUDALltAwC6ZQMAu30DALxlAwC9bQMAvmUDAL/dAACwpQMAsa0DALKlAwCzvQMAtK0DALWdAwC2lQMAt10DALMJAgCVxgCAmcYAgJ3GAIChxgCAtg0CALUNAgClxgCAu2kCALphAgCpxgCArcYAgL9ZAgC+aQIAvWkCALxxAgCxxgCAtcYAgLnGAIC9xgCA4aABAMHGAIDjaAMAxcYAgIEVAACAFQAA74wDAIIVAADJxgCAzcYAgNHGAIC+cAUA4RgOAOGUDwDjOA8A49QPAISUAgDZxgCA3cYAgOHGAIDlxgCA6cYAgO3GAIDxxgCA9cYAgPnGAIDv7AEA7/gPAIZgBACHBAUAs5UBAITMBQC1dQEA/cYAgAHHAIC2dQEABccAgAnHAIC7UQEAulkBAL31AAC8SQEAv/UAAL71AACoJQYAqVUGAKpVBgCrrQYArLUGAK29BgCutQYAr60GANXGAIANxwCAEccAgBXHAIAZxwCAHccAgCHHAIAlxwCAuGkHALlpBwC6CQcAuwkHALwZBwC9GQcAvg0HAL8BBwCw1QYAsd0GALLVBgCzaQcAtHkHALV5BwC2aQcAt2EHAKPdBgApxwCALccAgDHHAIA1xwCApj0GAKU9BgA5xwCAqxkGAKoRBgA9xwCAQccAgK+9BwCuvQcArb0HAKwBBgCAXQAAgW0AAIJlAACzUQcAvtgDALVxBwC2cQcARccAgIbgAACHFAMAul0HALs5BwC8KQcAvRUHAL4dBwC/2QAAqJUGAKmdBgCqlQYAq60GAKy1BgCtvQYArrUGAK+tBgBJxwCATccAgFHHAIBVxwCAWccAgF3HAIBhxwCAZccAgLhxAQC5cQEAunEBALtxAQC81QEAvd0BAL7VAQC/zQEAsNUGALGxBgCysQYAs40GALSVBgC1UQEAtlEBALdRAQBpxwCAoxkGAG3HAIBxxwCApjkGAGXGAIB1xwCApTkGAKoVBgCrcQYAeccAgH3HAICuVQYAr5EBAKxhBgCtXQYAgccAgIXHAICJxwCAjccAgJHHAICVxwCAmccAgJ3HAIChxwCApccAgKnHAICtxwCAgBkAAIEZAACCBQAAsccAgISAAgC+gAMAhwwDAIasHADhaAYAuccAgOOYBwC9xwCAwccAgMXHAIDvrAcAyccAgM3HAIDRxwCA1ccAgNnHAIDdxwCA4ccAgOXHAICzZQMA6ccAgLVlAwC2bQMA7ccAgPHHAID1xwCAuukDALvlAwC8/QMAve0DAL7RAwC/0QMA+ccAgP3HAIAByACABcgAgAnIAIANyACAEcgAgBXIAICogQMAqYEDAKqBAwCrgQMArIEDAK2BAwCugQMAr4EDALBBAwCxTQMAskUDALNVAwC0eQMAtXkDALYZAwC3GQMAuCkDALkpAwC6OQMAuzkDALwpAwC9KQMAvhkDAL8ZAwCBGQAAgBEAAKMhAgCCLQAApSECABnIAIAdyACApikCACHIAIApyACAq6ECAKqtAgCtqQIArLkCAK+VAgCulQIAhEwCAL5IHQCHZB0AhuwcAONAAwAtyACA4aABADHIAIDvnAMANcgAgDnIAIA9yACAQcgAgEXIAIBJyACATcgAgFHIAIBVyACAWcgAgF3IAIBhyACAZcgAgGnIAIDvtAEAhKgdAOF8BgBtyACA43AGAHHIAIB1yACAecgAgH3IAICz4QEAgcgAgIXIAICJyACAjcgAgLblAQC19QEAkcgAgLuhAQC62QEAvuQcAJnIAIC/rQEAvqUBAL2xAQC8uQEAqBUeAKkZHgCqKR4AqykeAKw9HgCtJR4Ari0eAK8lHgAlyACAgvkfAIH5HwCA4R8AlcgAgJ3IAICGHAAAh7ADALjBHgC5wR4AusEeALvBHgC8wR4AvcEeAL7BHgC/wR4AsF0eALElHgCyLR4AsyUeALQhHgC1KR4AthkeALcZHgCjoR4AocgAgKXIAICpyACArcgAgKalHgCltR4AscgAgKvhHgCqmR4AtcgAgLnIAICv7R4AruUeAK3xHgCs+R4AvcgAgLOZHwDByACAxcgAgLa9HwDJyACAzcgAgLW1HwC6mR8Au5kfANHIAIDVyACAvnkfAL95HwC8eR8AvXkfAKglHgCpUR4AqlUeAKtpHgCseR4ArXkeAK5pHgCvaR4A2cgAgN3IAIDhyACA5cgAgOnIAIDtyACA8cgAgPXIAIC42R4Aue0eALr5HgC7+R4AvOkeAL3pHgC+nR4Av5UeALAZHgCxGR4AsukeALPpHgC0+R4AtfkeALbpHgC36R4Ao90eAIIpAACBFQAAgB0AAPnIAICm+R4ApfEeAP3IAICr3R4Aqt0eALXHAIAByQCArz0eAK49HgCtPR4ArD0eAITIAgCzQQEAvgwBAAnJAIC2QQEADckAgBHJAIC1UQEAuk0BALslAQCGSAAAh1ABAL4lAQC/LQEAvDEBAL0xAQAVyQCAGckAgIQEAwC+gAQAHckAgO+oHwAhyQCAJckAgL8oMQDjdB8AKckAgOE4HgAtyQCAMckAgDXJAIA5yQCAPckAgEHJAICjzQIARckAgKXdAgBJyQCATckAgKbNAgBRyQCAVckAgKupAgCqwQIArb0CAKy9AgCvoQIArqkCAKm1AgCoaR0AqwECAKoJAgCtAQIArBkCAK8xAgCuAQIAhGwFAFnJAIBdyQCAYckAgGXJAICCnQEAgZ0BAICdAQC55QMAuOUDALvlAwC65QMAveUDALzlAwC/5QMAvuUDALEhAgCwSQIAsyUCALIlAgC1KQIAtCECALcVAgC2FQIAqM0CAKnRAgCq0QIAqw0BAKwVAQCtBQEArgEBAK8BAQBpyQCAbckAgHHJAIB5yQCAvvgEAH3JAICByQCAhckAgLgVAQC5HQEAuikBALspAQC89QEAvf0BAL71AQC/7QEAsEkBALFVAQCyXQEAs1UBALRNAQC1NQEAtj0BALcxAQCGoAUAh8gFAInJAIDvvAAAjckAgJHJAICVyQCA74weAIQsBwDh8B4AmckAgOMcHgCdyQCA4ZQBAKHJAIDjbAAAsxkCAKXJAICpyQCArckAgIQACAC2xQEAtd0BALHJAIC70QEAus0BALXJAIC5yQCAv7EBAL7JAQC9wQEAvMkBAKPZBQB1yQCAvckAgMHJAIDFyQCApgUGAKUdBgDJyQCAqxEGAKoNBgDNyQCA0ckAgK9xBgCuCQYArQEGAKwJBgDVyQCAgh0AAIEdAACAHQAA2ckAgN3JAIDhyQCA5ckAgIZAAwCHxAMA6ckAgO3JAIDxyQCA9ckAgPnJAID9yQCAqK0HAKmxBwCqsQcAq7EHAKwZBwCtBQcArg0HAK8FBwABygCABcoAgAnKAIANygCAEcoAgBXKAIAZygCAHcoAgLgtBwC5zQAAusUAALvdAAC8zQAAvf0AAL71AAC/nQAAsEkHALFVBwCyUQcAsykHALQ5BwC1OQcAtiUHALcVBwCzOQYAIcoAgCXKAIApygCALcoAgLaFBgC1kQYAMcoAgLuRBgC6jQYANcoAgDnKAIC//QYAvv0GAL39BgC8hQYAPcoAgKN9BgBBygCARcoAgKbBBgBJygCATcoAgKXVBgCqyQYAq9UGAFHKAIC+bAEArrkGAK+5BgCswQYArbkGAKjpAQCp6QEAqvkBAKv5AQCs6QEArekBAK45AQCvOQEAgPUAAIH9AACCwQAAVcoAgIYQAACHdAEAWcoAgAXJAIC4zQAAudUAALrVAAC75QAAvP0AAL2VAAC+kQAAv5EAALBJAQCxSQEAslkBALNZAQC0SQEAtUkBALb9AAC39QAA7/QGAF3KAIBhygCAZcoAgO8wAgBpygCAbcoAgHHKAIDj4AcAdcoAgOGAAQB5ygCA4ygGAH3KAIDhyAUAgcoAgLMxAgCFygCAicoAgJYAAACNygCAtikCALUhAgCRygCAu80CALrNAgCVygCAmcoAgL/NAgC+zQIAvc0CALzNAgCdygCAocoAgKXKAICj/QIAqcoAgKXtAgCm5QIArcoAgLHKAIC1ygCAqgECAKsBAgCsAQIArQECAK4BAgCvAQIAgA0AAIEVAACCHQAAucoAgL3KAIDBygCAvlQMAMnKAICGwAwAhyQDAM3KAIDRygCA1coAgNnKAIDdygCA4coAgKi5AgCpAQEAqgEBAKsBAQCsBQEArQ0BAK4FAQCvOQEAhKgNAOXKAIDpygCA7coAgPHKAID1ygCA+coAgP3KAIC4LQEAucUBALrNAQC7xQEAvMEBAL3JAQC++QEAv/kBALBNAQCxUQEAslUBALMpAQC0OQEAtSUBALYlAQC3FQEA4RgGAAHLAIDjOAcABcsAgAnLAIC+WAwADcsAgBHLAICEbA8AFcsAgL5gDwAZywCAHcsAgCHLAIDvcAYAJcsAgIAVAACBGQAAgi0AAITMDwDjYAYAKcsAgOGgAQAtywCA73QAADHLAICGyAwAh/wMADnLAIA9ywCAQcsAgEXLAICjCQ4AxcoAgDXLAIBJywCATcsAgKYNDgClDQ4AUcsAgKsVDgCqCQ4AVcsAgFnLAICvYQ4Arn0OAK19DgCsAQ4AXcsAgLOpDgBhywCAZcsAgLapDgBpywCAbcsAgLWpDgC6SQ8Au0kPAHHLAIB1ywCAvkkPAL9JDwC8SQ8AvUkPAKhdDgCpbQ4AqmUOAKt9DgCsZQ4ArW0OAK5lDgCvuQ8AecsAgH3LAICBywCAhcsAgInLAICNywCAkcsAgJXLAIC4UQ8AuV0PALpVDwC7aQ8AvH0PAL1lDwC+bQ8Av2EPALDJDwCxyQ8AstkPALPZDwC0yQ8AtckPALZ9DwC3cQ8AmcsAgLURDwC2EQ8AncsAgIARAACBGQAAgikAALMVDwC8HQ8AvWEPAL5hDwC/fQ8AocsAgKXLAIC6FQ8AuwkPAKOtDwCpywCAhugAAIfIAQCtywCApq0PAKWtDwCxywCAq00OAKpNDgC1ywCAucsAgK9NDgCuTQ4ArU0OAKxNDgCocQ4AqXEOAKpxDgCrcQ4ArJ0BAK2FAQCuhQEAr7UBAL7sAAC9ywCAwcsAgMXLAIDJywCAzcsAgNHLAIDVywCAuGEBALlhAQC6YQEAu2EBALxhAQC9YQEAvmEBAL9hAQCwzQEAsaUBALKhAQCzoQEAtKUBALWtAQC2kQEAt5EBALP5DQDZywCA3csAgOHLAIDlywCAtgUCALUVAgDpywCAu2ECALoJAgDtywCA8csAgL9pAgC+YQIAvXUCALx1AgD1ywCAo70NAPnLAID9ywCApkECAAHMAIAFzACApVECAKpNAgCrJQIACcwAgA3MAICuJQIAry0CAKwxAgCtMQIAge0AAIDtAADv0AEAgh0AABHMAIAZzACAhjgEAIdQAwAdzACAIcwAgCXMAIApzACA4eABAC3MAIDjZA8AMcwAgDXMAIA5zACAPcwAgLORAwBBzACAtbkDALZ9AwBFzACAScwAgE3MAIC6WQMAu1kDALxJAwC9SQMAvv0AAL/1AACoRQIAqVUCAKpVAgCrZQIArH0CAK2xAgCusQIAr7ECAL5oBQBRzACAVcwAgFnMAIBdzACAYcwAgGXMAIBpzACAuF0BALltAQC6ZQEAuw0BALwZAQC9GQEAvg0BAL8FAQCw0QIAsdECALLRAgCz0QIAtHUBALV9AQC2dQEAt20BAOF4DwDjNA4A47gOAOF8DgBtzACAccwAgHXMAIB5zACAfcwAgIHMAICJzACAjcwAgJHMAIDv5A4A79QOAJXMAICjnQIAgmEAAIFpAACAUQAAhJwFAKZxAgCltQIAmcwAgKtVAgCqVQIAhkgEAIfMBACv+QEArvEBAK1FAgCsRQIAqJUGAKmlBgCqrQYAq6UGAKy9BgCtoQYArqUGAK/dBgCFzACAncwAgKHMAIClzACAqcwAgK3MAICxzACAtcwAgLhtBwC5dQcAun0HALt1BwC8bQcAvcUHAL7NBwC/xQcAsKUGALGtBgCyuQYAs7EGALSRBgC1kQYAtl0HALdVBwCzJQYAucwAgL3MAIDBzACAxcwAgLYhBgC1NQYAycwAgLtpBgC6YQYAzcwAgNHMAIC/VQYAvlUGAL1lBgC8bQYA1cwAgKNhBgDZzACA3cwAgKZlBgDhzACA5cwAgKVxBgCqJQYAqy0GAOnMAIDtzACArhEGAK8RBgCsKQYArSEGAKipBgCpqQYAqrkGAKuxBgCszQYArTEBAK4xAQCvMQEAgMkBAIHJAQCCBQAA8cwAgL54AgCEeAIA9cwAgPnMAIC43QEAue0BALrlAQC7jQEAvJkBAL2ZAQC+jQEAv4UBALBRAQCxUQEAslEBALNRAQC09QEAtf0BALb1AQC37QEAszEGAP3MAICGKAAAh9wBAAHNAIC2sQEAtUUGAAXNAIC7lQEAupUBAAnNAIANzQCAvzkBAL4xAQC9hQEAvIUBABXMAICjdQYAEc0AgBXNAICm9QEAGc0AgB3NAIClAQYAqtEBAKvRAQAhzQCAJc0AgK51AQCvfQEArMEBAK3BAQApzQCALc0AgDHNAIA1zQCAOc0AgD3NAIBBzQCARc0AgEnNAIBNzQCAUc0AgFXNAIBZzQCAXc0AgGHNAIC+cAMAhQA8AOHEBgCERAIA44wHAIBhAACBYQAAgmEAAO9oAwCFRDwA4RACAGnNAIDj2CsAhlA9AIf0AwBtzQCA76QHAHHNAIDvQAIAdc0AgHnNAIB9zQCAgc0AgIXNAICJzQCAhDw8AI3NAICRzQCAlc0AgJnNAIDj7AIAnc0AgOEsAQCzUQMAoc0AgKXNAICpzQCArc0AgLZ5AwC1cQMAsc0AgLs5AwC6MQMAtc0AgLnNAIC/9QAAvvUAAL0VAwC8FQMAqD0CAKmBAgCqmQIAq5ECAKy5AgCtuQIArtECAK/RAgCEqD8Avqg/AL3NAIDBzQCAxc0AgMnNAIDNzQCA0c0AgLhRAQC5UQEAulEBALtRAQC8cQEAvXEBAL5xAQC/cQEAsLUCALG9AgCygQIAs4ECALRxAQC1cQEAtnEBALdxAQCAtQAAgb0AAIK1AADZzQCAhrA/AIfgPADdzQCA71QAAL4sPgDhVAYA4c0AgOOIAADlzQCA6c0AgO3NAIDxzQCAo1ECAPXNAIC/2CYA+c0AgP3NAICmeQIApXECAAHOAICrOQIAqjECAAXOAIAJzgCAr/UBAK71AQCtFQIArBUCAJAtJACRBSgAkg0oAJPZKACUhS0AlTUsAJbFLACXtTEAmAEwAJkVMACalTUAmyk0AJxtNACdmTUAnj04AJ81OABlzQCAttU+ALXFPgDVzQCAs9E+AA3OAIARzgCAFc4AgL/ZPgC+1T4AvcU+ALzFPgC71T4Auuk+ABnOAICPXSQAqeUJAKgVCACrBQwAqg0MAK0BEACsAQwAr0EQAK69EACh4QAAHc4AgKMBBACi4QAApZ0EAKSVBACnuQgApgEIAKD1OQChBT0Aouk8AKP1PQAhzgCAJc4AgCnOAIAtzgCAscEUALABFACzARgAsn0UALXVGAC01RgAMc4AgDXOAICCISUAgyklADnOAIA9zgCAhsUpAIeBLACEGSkAhRkpAIoBLQCL+S0AQc4AgEnOAICOATEAj4k0AIyRMACNHTEAkkU1AJMZNQCG6AcAh+wBAJZZOQCXYTgAlPU0AJVZOQCaoTwAm0U9AE3OAIBRzgCAgX0AAIB9AACcQTwAglUAAKjpPwCp/T8Aqgk/AKsFPwCsHT8ArQU/AK4NPwCvBT8AVc4AgFnOAIBdzgCAYc4AgGXOAIBpzgCAbc4AgHHOAIC4DT8AuRU/ALoVPwC7JT8AvD0/AL39PgC+9T4Av+0+ALB9PwCxQT8AskE/ALNBPwC0QT8AtU0/ALY9PwC3NT8Ao4E8AHXOAIB5zgCAfc4AgIHOAICmhTwApZU8AIXOAICrhTwAqrk8AInOAICNzgCAr4k8AK6FPACtlTwArJU8AITIAwCz7T0Akc4AgJXOAIC26T0Amc4AgJ3OAIC16T0Auq09ALu1PQChzgCApc4AgL6dPQC/IQIAvKU9AL2VPQCoDT0AqR09AKohPQCrPT0ArCU9AK0tPQCuJT0Ar1k9AIANAACBFQAAgh0AAKnOAICtzgCAsc4AgLnOAIC+uAMAuLkCALlhAgC6GQIAuxkCALwJAgC9CQIAviECAL8hAgCwLT0AsTU9ALI1PQCzBT0AtB09ALWhAgC2oQIAt6ECAKOpPAC9zgCAhigFAIfsAgDBzgCApq08AKWtPADFzgCAq/E8AKrpPADJzgCAzc4AgK9lAwCu2TwArdE8AKzhPADRzgCAsykCANXOAIDZzgCAtvkCAN3OAIDhzgCAtfkCALrVAgC73QIA5c4AgOnOAIC+eQEAv3kBALzFAgC9eQEA7c4AgPHOAICj5QIA9c4AgKU1AgD5zgCA/c4AgKY1AgABzwCABc8AgKsRAgCqGQIArbUBAKwJAgCvtQEArrUBAOPwPgDhrD8A4UA+AON8PwAJzwCADc8AgBHPAIAVzwCAgA0AAIERAACCEQAAGc8AgO+oPgAdzwCAIc8AgO8gPgCoLQUAqW0FAKplBQCrrQUArLUFAK29BQCutQUAr60FALXOAICE6AMAvuADACXPAICGEAMAh5gDACnPAIAtzwCAuGkGALlpBgC6AQYAuwEGALwFBgC9DQYAvjEGAL8xBgCw1QUAsd0FALLVBQCzaQYAtHkGALV5BgC2aQYAt2EGAKg5BgCpgQcAqpkHAKuRBwCsuQcArbkHAK7ZBwCv1QcAMc8AgDXPAIBFzgCAOc8AgD3PAIBBzwCARc8AgEnPAIC4VQcAuV0HALppBwC7aQcAvAEHAL0BBwC+AQcAvwEHALCtBwCxsQcAsrEHALOFBwC0nQcAtXUHALZ9BwC3cQcAsxEGAE3PAIBRzwCAVc8AgFnPAIC2OQYAtTEGAF3PAIC7dQYAumkGAGHPAIBlzwCAv7EGAL5ZBgC9UQYAvGUGAGnPAICjVQYAbc8AgHHPAICmfQYAdc8AgHnPAICldQYAqi0GAKsxBgB9zwCAgc8AgK4dBgCv9QYArCEGAK0VBgCouQEAqbkBAKopAQCrKQEArD0BAK0lAQCuLQEAryUBAIXPAICCHQAAgR0AAIAdAACJzwCAjc8AgJHPAIC+cAEAuIEAALmNAAC6hQAAu5kAALyJAAC9vQAAvrUAAL99AACwXQEAseEAALLhAACz4QAAtOEAALXpAAC20QAAt9EAAITIAgCzpQIAhzgDAIYoAgC2oQIAmc8AgJ3PAIC1sQIAup0CALshAwC+bAMAoc8AgL4hAwC/KQMAvDEDAL0xAwCj4QIApc8AgKnPAICtzwCAsc8AgKblAgCl9QIAtc8AgKtlAwCq2QIAuc8AgL3PAICvbQMArmUDAK11AwCsdQMAqZkAAKiRAACrzQAAqqEAAK3dAACs3QAAr8UAAK7NAAC+LA0Awc8AgMXPAIDJzwCAzc8AgNHPAIDVzwCA2c8AgLnBAQC4eQAAu8EBALrJAQC9wQEAvNkBAL/FAQC+xQEAsY0AALCNAACzQQAAskkAALVBAAC0WQAAt0EAALZJAADdzwCA4c8AgOXPAIDpzwCA7c8AgO9QBwDxzwCA9c8AgL74DwDjdAcA+c8AgOF8BACAGQAAgQkAAIJ5AAD9zwCAAdAAgLNpAQAJ0ACAhMQCALYdAQAN0ACAEdAAgLUVAQC6CQEAuwkBAIboDQCH6A0Avt0BAL/FAQC83QEAvdUBABXQAIAZ0ACAHdAAgCHQAIDv1AAAJdAAgCnQAIDvTAEA47ADAOG0BgDhgAEA45gBAC3QAIAx0ACANdAAgDnQAIA90ACAQdAAgKPlAQCEwA0ApZkBAEXQAIBJ0ACAppEBAE3QAIBR0ACAq4UBAKqFAQCtWQEArFEBAK9JAQCuUQEABdAAgFXQAIBZ0ACAXdAAgGHQAIBl0ACAadAAgG3QAICoaQ8AqXEPAKpxDwCrrQ8ArLUPAK29DwCutQ8Ar6kPALDZDwCx9Q8Asv0PALP1DwC07Q8AtZUPALadDwC3iQ8AuLkPALmFDwC6jQ8Au2kAALx5AAC9eQAAvmkAAL9pAACBnQAAgJ0AAHHQAICCBQAAddAAgHnQAIB90ACAgdAAgIaAAwCH9AMAhdAAgInQAICN0ACAkdAAgJXQAICVzwCAs5kPAJnQAICd0ACAodAAgKXQAIC2XQ8AtV0PAKnQAIC7UQ8Aun0PAK3QAICx0ACAvzEPAL5JDwC9QQ8AvEkPAKNZDgC10ACAudAAgL3QAIDB0ACApp0OAKWdDgDF0ACAq5EOAKq9DgDJ0ACAzdAAgK/xDgCuiQ4ArYEOAKyJDgDR0ACA1dAAgNnQAIDd0ACAgBkAAIEZAACCBQAA4dAAgISgAQDl0ACAh+gBAIYABADp0ACA7dAAgPHQAID10ACAqBUBAKkdAQCqFQEAqyUBAKw9AQCtJQEAri0BAK8lAQD50ACA/dAAgAHRAIAF0QCACdEAgA3RAIAR0QCAFdEAgLjJAAC5yQAAutkAALvRAAC8+QAAvfkAAL6ZAAC/mQAAsCUBALEtAQCyJQEAsz0BALQtAQC1HQEAthUBALf5AAAZ0QCAHdEAgCHRAICzkQIAJdEAgLW5AgC2qQIAKdEAgC3RAIAx0QCAuu0CALvlAgC8/QIAveUCAL7lAgC/1QIApvECADXRAIA50QCApeECAD3RAICjyQIAQdEAgEXRAICuvQIAr40CAKylAgCtvQIAqrUCAKu9AgBJ0QCATdEAgID5AACB+QAAggUAAFHRAIC+yAMAhBgDAFnRAIBd0QCAYdEAgGXRAIBp0QCAbdEAgHHRAIB10QCAhhgEAIecAwB50QCAfdEAgIHRAICF0QCAidEAgI3RAIDvsAIAkdEAgOGUAQCV0QCA42wCAJnRAICd0QCAodEAgKXRAICp0QCA79APAK3RAICx0QCAtdEAgLnRAIDhrAEAvdEAgONsAACAMQAAgT0AAIIdAADv9A4A42wOAMHRAIDhLA8AvnAFALM5AgCEDAUAhugEAIdgBQDcAAAAtvECALX5AgDJ0QCAu9UCALrVAgDN0QCA0dEAgL91AQC+dQEAvcUCALzFAgDV0QCA4fQOANnRAIDjUA4A3dEAgOHRAIDl0QCA6dEAgO3RAIDx0QCA9dEAgPnRAID90QCAAdIAgAXSAIDv5A8ApmUCAAnSAIAN0gCApW0CABHSAICjrQIAFdIAgBnSAICu4QEAr+EBAKxRAgCtUQIAqkECAKtBAgAd0gCAIdIAgKiZBgCpmQYAqqkGAKupBgCsuQYArbkGAK6pBgCvqQYAJdIAgIIdAACBHQAAgB0AACnSAIAt0gCAMdIAgL50AwC4rQYAubUGALq9BgC7tQYAvK0GAL1RBwC+UQcAv1EHALChBgCxoQYAsqEGALOhBgC0oQYAtaEGALalBgC3mQYAVdEAgLMlBgCExAMAxdEAgLY9BgA10gCAOdIAgLU1BgC6YQYAu2EGAIYIAACHiAAAvmEGAL9hBgC8cQYAvXEGAKNhBgA90gCAQdIAgEXSAIBJ0gCApnkGAKVxBgBN0gCAqyUGAKolBgBR0gCAVdIAgK8lBgCuJQYArTUGAKw1BgCoXQYAqW0GAKplBgCrjQYArJkGAK2FBgCujQYAr4UGAFnSAIBd0gCAYdIAgGXSAIBp0gCAbdIAgHHSAIB10gCAuIUGALmNBgC6mQYAu5UGALyNBgC9rQYAvqUGAL99AQCw/QYAscUGALLNBgCzxQYAtN0GALXFBgC2zQYAt8UGALPtBgB50gCAfdIAgIHSAICF0gCAtgUGALURBgCJ0gCAuwEGALo5BgCN0gCAkdIAgL8BBgC+GQYAvREGALwZBgCV0gCAo6kGAJnSAICd0gCApkEGAKHSAICElAEApVUGAKp9BgCrRQYAvqABAKnSAICuXQYAr0UGAKxdBgCtVQYAqJkCAKnBAgCqwQIAq8ECAKzBAgCtyQIArvECAK/xAgCB7QMAgO0DAK3SAICC+QMAhpAcAId0AwCx0gCAtdIAgLjFAwC5zQMAusUDALvdAwC8zQMAvf0DAL71AwC/nQMAsEEDALFBAwCyQQMAs0EDALRBAwC1QQMAtkEDALdBAwCzSQIAudIAgL3SAIDB0gCAxdIAgLZJAgC1SQIAydIAgLuFAwC6hQMAzdIAgNHSAIC/hQMAvoUDAL2VAwC8lQMA1dIAgKMNAgDZ0gCA3dIAgKYNAgDh0gCA5dIAgKUNAgCqwQMAq8EDAOnSAIDt0gCArsEDAK/BAwCs0QMArdEDAOOYAQDhpAcA4VgGAONYBgDhoAEA8dIAgOPQAAD10gCA+dIAgP3SAIDvOAAAAdMAgO/0AQAF0wCACdMAgO/4BgCAeQAAgRUAAIIdAACEAB0ADdMAgBHTAIC+EB0AGdMAgIbAHACHrB0AHdMAgCHTAIAl0wCAKdMAgC3TAIAx0wCAu8UFALqhBQC5qQUAuJEFAL/NBQC+zQUAvckFALzVBQCzHQYAsh0GALEdBgCwHQYAt6EFALa9BQC1vQUAtL0FAKu9BgCqvQYAqb0GAKi9BgCvfQYArn0GAK19BgCsfQYANdMAgDnTAIA90wCAQdMAgEXTAIBJ0wCATdMAgFHTAICo7R0AqS0eAKoxHgCrMR4ArJUeAK2dHgCulR4Ar40eABXTAIBV0wCAWdMAgF3TAIBh0wCAZdMAgGnTAIBt0wCAuKkeALmpHgC6XR8Au1EfALxxHwC9cR8AvnUfAL9pHwCw/R4Asc0eALLFHgCzrR4AtLkeALW5HgC2rR4At6UeALO5HgBx0wCAddMAgHnTAICl0gCAth0eALUdHgB90wCAuwkeALo5HgCB0wCAhOADAL99HgC+fR4AvXkeALwRHgCCaQAAo/0eAIBFAACBUQAAplkeAL6cAwCF0wCApVkeAKp9HgCrTR4AhkgAAIdsAACuOR4ArzkeAKxVHgCtPR4AqF0eAKltHgCqZR4Aq30eAKxlHgCtbR4ArmUeAK/9HgCJ0wCAjdMAgJHTAICV0wCAmdMAgJ3TAICh0wCApdMAgLhpAQC5aQEAunkBALt5AQC8aQEAvWkBAL7dAQC/1QEAsIUeALGNHgCyhR4As50eALSFHgC1jR4AtoUeALdZAQCz7R4AqdMAgK3TAICx0wCAtdMAgLbtHgC17R4AudMAgLtJHgC6QR4AvdMAgMHTAIC/SR4AvkEeAL1JHgC8UR4AxdMAgKOpHgDJ0wCAzdMAgKapHgDR0wCA1dMAgKWpHgCqBR4Aqw0eANnTAIDd0wCArgUeAK8NHgCsFR4ArQ0eAKghAwCpIQMAqiEDAKshAwCsIQMArSEDAK4hAwCvIQMA4dMAgOXTAIDp0wCAvmACAO3TAIDx0wCA+dMAgP3TAIC4iQMAuYkDALqdAwC7lQMAvLkDAL25AwC+eQAAv3kAALDlAwCx7QMAsuUDALP9AwC07QMAtd0DALbVAwC3vQMAgKkAAIG1AACCvQAAs6UDAAHUAIC1pQMAtq0DAAXUAICE4AIACdQAgLotAwC7JQMAvD0DAL0lAwC+JQMAvxUDAKPpAwAN1ACAhmgEAIeAAwAR1ACApuEDAKXpAwAV1ACAq2kDAKphAwAZ1ACAHdQAgK9ZAwCuaQMArWkDAKxxAwAh1ACAJdQAgCnUAIAt1ACAMdQAgOE8HwA11ACA40AeADnUAIA91ACAQdQAgO+MHgBF1ACASdQAgE3UAIBR1ACAVdQAgIIlAACBEQAAgB0AAFnUAIDj5AMAXdQAgOGsAQBh1ACA77ADAIRkAgC+YAUAhtAEAIdEBQBp1ACAbdQAgHHUAIB11ACAedQAgH3UAICB1ACAhdQAgInUAIDvsAEAhKQFAOHcHgCN1ACA4xABAJHUAICV1ACAmdQAgJ3UAICzUQEAodQAgKXUAICp1ACArdQAgLYRAQC1fQEAsdQAgLsNAQC6DQEAtdQAgLnUAIC//QAAvv0AAL39AAC8/QAAqDkGAKk5BgCqmQYAq5EGAKy1BgCt0QYArskGAK/BBgBl1ACAvdQAgMHUAIDF1ACAgA0AAIGxAACCsQAAydQAgLhhBwC5YQcAumEHALt9BwC8ZQcAvW0HAL5lBwC/HQcAsIkGALGJBgCyaQcAs2kHALR5BwC1eQcAtmkHALdlBwCjEQYAzdQAgNHUAIC+gAMA1dQAgKZRBgClPQYA2dQAgKtNBgCqTQYAhggAAId8AwCvvQcArr0HAK29BwCsvQcA3dQAgOHUAICzSQcA5dQAgLVZBwDp1ACA7dQAgLZRBwDx1ACA9dMAgLtBBwC6dQcAvUUHALxFBwC/RQcAvkUHAKh5BgCpeQYAqokGAKuJBgCsmQYArZkGAK6JBgCviQYA9dQAgPnUAID91ACAAdUAgAXVAIAJ1QCADdUAgBHVAIC4jQYAuZUGALqVBgC7pQYAvL0GAL1xAQC+cQEAv3EBALD5BgCxzQYAstkGALPZBgC0yQYAtckGALa9BgC3tQYAowEGABXVAIAZ1QCAHdUAgCHVAICmGQYApREGACXVAICrCQYAqj0GACnVAIAt1QCArw0GAK4NBgCtDQYArA0GADHVAIA11QCAOdUAgD3VAICAGQAAgRkAAIIFAABB1QCAhKwBAL6sAQCH6AAAhkwPAEnVAIBN1QCAUdUAgFXVAIConQIAqcUCAKrNAgCrwQIArMUCAK3NAgCu+QIArz0DAFnVAIBd1QCAYdUAgGXVAIC+PAwAadUAgG3VAIBx1QCAuMkDALnJAwC62QMAu9EDALz5AwC9+QMAvpkDAL+ZAwCwRQMAsU0DALJFAwCzXQMAtEUDALVNAwC2RQMAt/kDALNFAgB11QCAedUAgH3VAICB1QCAtk0CALVNAgCF1QCAu4kDALqBAwCJ1QCAjdUAgL+JAwC+gQMAvYkDALyRAwCR1QCAowECAJXVAICZ1QCApgkCAJ3VAICh1QCApQkCAKrFAwCrzQMApdUAgKnVAICuxQMAr80DAKzVAwCtzQMAgO0BAIEVAACCEQAAhAACAK3VAIDhpAEAsdUAgOPsAAC51QCAvdUAgMHVAIDvMAAAxdUAgMnVAIDN1QCA0dUAgIbgDACH9AIA1dUAgNnVAIDd1QCA4dUAgO/MBgDl1QCA4bAHAOnVAIDjEAYA7dUAgPHVAID11QCA+dUAgP3VAIAB1gCABdYAgAnWAIAN1gCAEdYAgBXWAIAZ1gCA7+gBAIUYDwDhzAYAHdYAgOMcBgCAKQAAgR0AAIIFAAAh1gCAszkCAITMDQCGaA8Ah/wMAOHQ0gO28QEAtfkBACnWAIC72QEAutEBAL7kDAAt1gCAv30BAL59AQC9fQEAvMEBAKjxDQCp8Q0AqvENAKvxDQCsMQ4ArTEOAK4xDgCvMQ4AtdUAgCXWAIAx1gCANdYAgDnWAIA91gCAQdYAgEXWAIC46Q4AuekOALqJDgC7hQ4AvJ0OAL2BDgC+gQ4Av7UOALBVDgCxXQ4AslUOALPpDgC0+Q4AtfkOALbpDgC34Q4Ao3kNAEnWAIBN1gCAUdYAgFXWAICmsQ4ApbkOAFnWAICrmQ4AqpEOAF3WAIBh1gCArz0OAK49DgCtPQ4ArIEOAGXWAICz7Q8AadYAgG3WAIC26Q8AcdYAgHXWAIC16Q8Auq0PALu1DwBF1QCAedYAgL6VDwC/mQ8AvK0PAL2hDwCoIQ4AqSEOAKohDgCrPQ4ArCUOAK0tDgCuJQ4Ar1UOAH3WAICB1gCAhdYAgInWAICAHQAAgQkAAIK9AACN1gCAuDkOALk5DgC6yQ4Au8kOALzZDgC92Q4AvskOAL/JDgCwLQ4AsTUOALI9DgCzMQ4AtBUOALUZDgC2CQ4AtwkOAKOpDgCR1gCAhIACAL6AAQCFAAQApq0OAKWtDgCZ1gCAq/EOAKrpDgCGKAcAhxgAAK/dDgCu0Q4AreUOAKzpDgCd1gCAs+0BAKHWAICl1gCAtuUBAKnWAICt1gCAte0BALplAQC7bQEAsdYAgLXWAIC+bQEAv10BALx1AQC9bQEAqN0NAKnpDQCqIQIAqyECAKwhAgCtIQIAriECAK8hAgC51gCAvdYAgMHWAIDF1gCAohECAKMRAgCgqQ4AodUCALiJAgC5iQIAup0CALuVAgC8vQIAvXUDAL59AwC/dQMAsOUCALHtAgCy5QIAs/0CALTtAgC13QIAttUCALe9AgCjqQIAj8UaAMnWAIDN1gCA0dYAgKahAgClqQIA1dYAgKspAgCqIQIA2dYAgN3WAICvGQIArikCAK0pAgCsMQIAniUOAJ/lDgCc6QoAnRUKAJpFFgCbRQoAmFkWAJlRFgCWcRIAl4ETAJRVEgCV7RIAktEeAJPZHgCQtRoAkVUeAISpHwCFJR8AhiUfAIexEwDh1gCA5dYAgIJZGwCDURsAjEUSAI2lFwCOpRcAj7kXAIA5+wHp1gCAijkTAIutEwCUmQsAlaEPAJZpDwCX3Q8A7dYAgO+cDwCSyQsAk30LAJxFAwDjeA4A8dYAgOGYDAD11gCAhHgCAJqRAwCbXQMA4QQAAL6IBQDj3OoD+dYAgP3WAIAB1wCA7+wAAO+MDgDhcA4A4fwOAOMwAADjeA4AgSEAAIA5AADvtO0DgikAALMJAgAJ1wCAhmgEAIcsBQAN1wCAtg0CALUNAgAR1wCAu8UBALrFAQAV1wCAGdcAgL99AQC+fQEAvdUBALzVAQCV1gCABdcAgB3XAIAh1wCAJdcAgCnXAIAt1wCAMdcAgKi9BQCp5QUAquEFAKvhBQCs5QUAre0FAK7RBQCv0QUAsGEGALFhBgCyYQYAs2EGALTZBgC12QYAtskGALfBBgC4yQYAuckGALp5BwC7eQcAvEUHAL0lBwC+EQcAvw0HAKNJBQA11wCAOdcAgD3XAIBB1wCApk0FAKVNBQBF1wCAq4UGAKqFBgBJ1wCATdcAgK89BgCuPQYArZUGAKyVBgBR1wCAVdcAgFnXAIBd1wCAYdcAgGXXAIBp1wCAbdcAgIA5AACBOQAAggUAAHHXAIC+uAMAhLgDAHnXAIB91wCAqMUGAKnVBgCq1QYAq+UGAKz9BgCtHQEArhUBAK8NAQB11wCAgdcAgIaIAQCHHAEAhdcAgInXAICN1wCAkdcAgLjpAQC56QEAuokBALuJAQC8mQEAvZkBAL6JAQC/iQEAsHUBALF9AQCydQEAs+kBALT5AQC1+QEAtukBALfhAQCzXQYAldcAgJnXAICd1wCAhLwBALadAQC1dQYAodcAgLu5AQC6sQEApdcAgKnXAIC/PQEAvj0BAL09AQC8oQEArdcAgKMZBgCx1wCAtdcAgKbZAQC51wCAvdcAgKUxBgCq9QEAq/0BAMHXAIDF1wCArnkBAK95AQCs5QEArXkBAKj5AgCp+QIAqi0DAKs9AwCsJQMArS0DAK4lAwCvmQMAydcAgM3XAIDR1wCA1dcAgIANAACBsQAAgrEAANnXAIC4lQMAuZ0DALqhAwC7oQMAvHEAAL1xAAC+cQAAv3EAALDpAwCx6QMAsvUDALPFAwC03QMAtbUDALaxAwC3sQMAvswDAN3XAIDh1wCA6dcAgO3XAIDx1wCA9dcAgO/kAgD51wCA4ZQBAP3XAIDjLAEAAdgAgAXYAICHGAMAhhz8A7tNAwC6TQMACdgAgA3YAIC/EQMAvnkDAL1xAwC8QQMAs8UDAITo/AMR2ACAFdgAgBnYAIC2zQMAtc0DAB3YAICkAfwDpSX/A6bZ/wOnAfgDIdgAgKEVAwCiHQMAoz0CAKwR9wOtAfADri3zA68B8wOoEfsDqZn7A6oB9AOrHfcDtAHoA7Vl6wO+xPwDhMT8A7AB7AOxVe8Dsk3vA7Nx7gMl2ACAKdgAgC3YAIAx2ACANdgAgDnYAIA92ACAQdgAgOFQBgDhNAQA42wBAOPoBgBF2ACASdgAgE3YAIBR2ACAgDUAAIE9AACCNQAAWdgAgF3YAIBh2ACA77ABAO/ABgCj5QIAZdgAgIbo/AOHfP0DadgAgKbtAgCl7QIAbdgAgKttAgCqbQIAcdgAgHXYAICvMQIArlkCAK1RAgCsYQIAqI3+A6mV/gOqnf4Dq5X+A6yx/gOtvf4Drqn+A6+p/gNV2ACAedgAgH3YAICB2ACAhdgAgInYAICN2ACAkdgAgLgl/wO5Lf8DuiX/A7s9/wO8Jf8DvS3/A74l/wO/zf8DsKn+A7Gp/gOygf4Ds4H+A7SB/gO1if4Dtmn/A7cd/wOV2ACA4SD8A5nYAIDjePwDndgAgKHYAICl2ACAqdgAgK3YAICx2ACAtdgAgLnYAICAHQAAgXEAAIJxAADvDP0Ds1X+A73YAIDB2ACAvkAAAMXYAIC2ff4DtXn+A8nYAIC7Lf4Dui3+A4boAACHrAAAvw3+A74F/gO9Ff4DvBX+A6OV/wPN2ACA0dgAgNXYAIDZ2ACApr3/A6W5/wPd2ACAq+3/A6rt/wPh2ACA5dgAgK/N/wOuxf8DrdX/A6zV/wPp2ACAs/H+A+3YAIDx2ACAto3+A/XYAID52ACAtY3+A7pFAQC7TQEA/dgAgAHZAIC+RQEAv00BALxVAQC9TQEAqC3+A6k1/gOqPf4Dq0n+A6xB/gOtSf4DrnH+A69x/gMF2QCACdkAgA3ZAIAR2QCAFdkAgBnZAIAd2QCAIdkAgLhJAQC5VQEAul0BALtVAQC8TQEAvXUBAL59AQC/dQEAsMUBALHNAQCyxQEAs90BALTFAQC1zQEAtsUBALd9AQCjtf0DJdkAgCnZAICExAMALdkAgKbJ/QOlyf0DMdkAgKsJAgCqAQIAOdkAgL7sAgCvCQIArgECAK0JAgCsEQIAgEkAAIFVAACCVQAAo0UDAD3ZAIClRQMApkUDAEHZAICGwAQAhxQDAKopAwCrJQMArD0DAK0hAwCuIQMArxUDAEXZAIBJ2QCATdkAgFHZAIBV2QCAWdkAgF3ZAIBh2QCAqH0CAKmhAwCqoQMAq6EDAKyhAwCtqQMArpEDAK+RAwCwgQMAsY0DALKFAwCzmQMAtIkDALW9AwC2tQMAt30DALhFAwC5TQMAukUDALtdAwC8RQMAvU0DAL5FAwC/+QAA5dcAgLMNAgBl2QCAadkAgLYNAgBt2QCAcdkAgLUNAgC6YQIAu20CAHXZAIB52QCAvmkCAL9dAgC8dQIAvWkCAH3ZAICB2QCAhdkAgInZAICN2QCA4aQBAJHZAIDjQAMAldkAgJnZAICd2QCA77gDAIAVAACBHQAAggUAAKHZAICEgAIAvsgFAIcYBQCGLAQAqdkAgK3ZAICx2QCA76gBALXZAIDhdP4DudkAgOPw/gO92QCAwdkAgMXZAIDJ2QCAzdkAgNHZAIDV2QCAs5EBANnZAIC1UQEAtlEBAN3ZAIDh2QCA5dkAgLp9AQC7dQEAvG0BAL39AAC+9QAAv+kAAKgpBgCpVQYAqlUGAKuNBgCslQYArZ0GAK6VBgCvjQYApdkAgOnZAIDt2QCA8dkAgPXZAID52QCA/dkAgAHaAIC4bQcAuQUHALoNBwC7BQcAvB0HAL0FBwC+AQcAvz0HALD1BgCx/QYAsvUGALNlBwC0fQcAtWEHALZhBwC3VQcA4xAFAAXaAIDh8AQACdoAgIAdAACBCQAAgjkAAA3aAIAR2gCAhOgDAL7gAwAV2gCA78wFABnaAICHOAAAhhgAAKOdBgAd2gCAIdoAgCXaAIAp2gCApl0GAKVdBgAt2gCAq3kGAKpxBgAx2gCANdoAgK/lBwCu+QcArfEHAKxhBgCokQYAqZEGAKqRBgCrrQYArLkGAK2lBgCurQYAr6UGADnaAIA92gCAQdoAgEXaAIBJ2gCATdoAgFHaAIBV2gCAuGUBALltAQC6ZQEAu30BALxlAQC9bQEAvmUBAL/ZAQCw3QYAsaUGALKtBgCzpQYAtKEGALWpBgC2mQYAt5kGALMZBgBZ2gCAXdoAgGHaAIBl2gCAtiUGALUxBgBp2gCAu2EGALoZBgBt2gCAcdoAgL9tBgC+ZQYAvXEGALx5BgB12gCAo10GAHnaAIB92gCApmEGAIHaAICEmAEApXUGAKpdBgCrJQYAvqQBAInaAICuIQYArykGAKw9BgCtNQYAqcUCAKixAgCrxQIAqsUCAK3NAgCsxQIAr/UCAK71AgCN2gCAkdoAgJXaAICZ2gCAndoAgKHaAICl2gCAqdoAgLnJAwC4wQMAu9kDALrBAwC9+QMAvMkDAL+ZAwC+8QMAsUUDALBFAwCzRQMAskUDALVFAwC0RQMAt0UDALZFAwCASQMAgUkDAIJdAwCzRQIAvtwMALVFAgC2RQIArdoAgIYADACH5AMAuokDALuJAwC8mQMAvZkDAL6JAwC/iQMAowkCALHaAIC12gCAudoAgL3aAICmCQIApQkCAMHaAICrxQMAqsUDAMXaAIDJ2gCAr8UDAK7FAwCt1QMArNUDAM3aAIDR2gCA1doAgDXZAIDvAAAA2doAgN3aAIDh2gCA4+gAAOXaAIDhjAEA6doAgO3aAIDx2gCA+doAgP3aAICAbQAAgXUAAIJ9AACEQAIAhvAMAId4DQAB2wCABdsAgAnbAIAN2wCAEdsAgBXbAIAZ2wCAHdsAgCHbAIAl2wCAKdsAgC3bAIAx2wCANdsAgDnbAIA92wCAQdsAgO/MAQCE7AwA4TAGAEXbAIDjGAEASdsAgE3bAIBR2wCAVdsAgLPlAQBZ2wCAhIQPAF3bAIBh2wCAtuUBALX1AQBp2wCAu30BALrZAQC+oAwAbdsAgL8hAQC+OQEAvTEBALw5AQCo7Q0AqSUOAKotDgCrJQ4ArD0OAK0lDgCuLQ4AryUOAPXaAICC9Q8AgeUPAIDpDwBl2wCAcdsAgIaYAACHDAMAuK0OALlFDwC6TQ8Au0UPALxFDwC9TQ8AvkUPAL95DwCwXQ4AsfkOALKtDgCzpQ4AtL0OALWlDgC2pQ4At5UOAHXbAIDv7AwAedsAgH3bAICB2wCAhdsAgInbAICN2wCAvugAAJHbAICV2wCAmdsAgJ3bAIDj6A0AodsAgOEEDACj5Q4ApdsAgKnbAICt2wCAsdsAgKblDgCl9Q4AtdsAgKt9DgCq2Q4AudsAgL3bAICvIQ4ArjkOAK0xDgCsOQ4AqDkOAKk5DgCqUQ4Aq1EOAKxxDgCtcQ4ArnEOAK9xDgDB2wCAxdsAgMnbAIDN2wCAgBkAAIEZAACCBQAA0dsAgLjRDgC50Q4AutEOALvlDgC84Q4AveEOAL7hDgC/4Q4AsBEOALERDgCyEQ4AsxEOALTxDgC18Q4AtvEOALfxDgCz2Q4A2dsAgIYoAACHuAAA3dsAgLbxDgC1+Q4A4dsAgLvVDgC61Q4A5dsAgOnbAIC/NQ4AvjUOAL3FDgC8xQ4A7dsAgKOdDgDx2wCA9dsAgKa1DgD52wCA/dsAgKW9DgCqkQ4Aq5EOAAHcAIAF3ACArnEOAK9xDgCsgQ4ArYEOAKjdDQCp6Q0Aqj0CAKuNAgCsmQIArZkCAK6JAgCviQIAvqwEAAncAIAN3ACAhCADABHcAIAV3ACAGdwAgB3cAIC4iQIAuYkCALqZAgC7kQIAvLkCAL25AgC+eQMAv3kDALD5AgCx+QIAss0CALPFAgC03QIAtcUCALbBAgC3uQIAs7UCACHcAIAl3ACAKdwAgC3cAIC2GQIAtRECADHcAIC7PQIAuj0CADXcAIA53ACAvwECAL4ZAgC9EQIAvBkCAD3cAICj8QIAQdwAgEncAICmXQIATdwAgFHcAIClVQIAqnkCAKt5AgCGSAUAh6wEAK5dAgCvRQIArF0CAK1VAgCohQIAqZUCAKqVAgCrpQIArL0CAK3VAgCu0QIAr9ECAFXcAIBZ3ACAXdwAgGHcAICB8QEAgJkBAIXaAICC9QEAuHkBALl5AQC6zQEAu8UBALzdAQC9xQEAvsUBAL/1AQCwtQIAsb0CALKBAgCzgQIAtFUBALVdAQC2SQEAt0kBAGXcAIBp3ACAbdwAgO/UAQCEEAUAcdwAgHXcAIDvjA4AvuwFAOHsDgB53ACA4xwOAH3cAIDhlAEAgdwAgONkDgCzXQIAhdwAgIncAICN3ACAkdwAgLYVAgC1dQIAldwAgLs5AgC6MQIAmdwAgJ3cAIC/2QEAvtEBAL0VAgC8FQIAo50FAEXcAICh3ACApdwAgKncAICm1QUApbUFAK3cAICr+QUAqvEFALHcAIC13ACArxkGAK4RBgCt1QUArNUFAIBRAACBWQAAgmEAALOVBgC53ACAtXEHALZxBwC93ACAhkADAIdUAwC67QcAu+UHALzlBwC97QcAvtEHAL/NBwDB3ACAxdwAgMncAIDN3ACA0dwAgNXcAIDvQAQA2dwAgOEwBwDd3ACA45QEAOHcAIDl3ACA6dwAgO3cAIDx3ACAoxkGAPXcAID53ACA/dwAgAHdAICm/QcApf0HAAXdAICraQcAqmEHAAndAIAN3QCAr0EHAK5dBwCtYQcArGkHAKjNBwCp0QcAqtEHAKstBgCsNQYArT0GAK41BgCvnQYAEd0AgBXdAIAZ3QCAHd0AgIAZAACBGQAAggUAACHdAIC4iQYAuYkGALqZBgC7kQYAvLkGAL25BgC+UQEAv1EBALDlBgCx7QYAsv0GALP1BgC02QYAtcUGALbBBgC3uQYAqNEBAKnZAQCqCQEAqwkBAKwZAQCtGQEArgkBAK8JAQCEYAEAvnwBAIeoAACGjAEAKd0AgC3dAIAx3QCANd0AgLgJAQC5CQEAuhkBALsRAQC8OQEAvTkBAL75AAC/+QAAsH0BALFBAQCyRQEAs10BALRFAQC1TQEAtkUBALc5AQA53QCAPd0AgEHdAICzjQIARd0AgLWdAgC2lQIASd0AgE3dAIBR3QCAurUCALuJAgC8nQIAvYUCAL6NAgC/hQIAps0CAFXdAIBZ3QCApcUCAF3dAICj1QIAYd0AgGXdAICu1QIAr90CAKzFAgCt3QIAqu0CAKvRAgCE9AMAad0AgKgxAwCpMQMAqjEDAKsxAwCskQAArZEAAK6RAACvjQAAbd0AgHHdAIB13QCAed0AgH3dAICB3QCAhd0AgIndAIC4vQAAuWUAALptAAC7ZQAAvH0AAL1lAAC+bQAAv2UAALD9AACxxQAAss0AALOpAAC0uQAAtaUAALahAAC3oQAAgL0BAIEJAACCGQAAjd0AgJHdAIC+WAIAhxQdAIacHQCEbB0A1dsAgJndAICd3QCAvrwcAKHdAICl3QCAqd0AgLP5AgCt3QCAsd0AgLXdAIC53QCAtlEBALVZAQC+3B8Au0EBALp5AQC93QCAwd0AgL8hAQC+PQEAvT0BALxZAQDhcAcAxd0AgOMIBgDJ3QCA78wAAM3dAIDR3QCA1d0AgOMQAADZ3QCA4dABAN3dAICGkBwAh/QcAO/gBgDh3QCAo3kCAOXdAIDp3QCA7d0AgPHdAICm0QEApdkBAPXdAICrwQEAqvkBAPndAID93QCAr6EBAK69AQCtvQEArNkBAJXdAICCFQAAgeUfAIDlHwAB3gCABd4AgAneAIAN3gCAqAkfAKkJHwCqHR8AqxUfAKwNHwCtcR8ArnEfAK9xHwCwER8AsS0fALIlHwCzyR8AtN0fALXBHwC2wR8At8EfALjFHwC5yR8AutUfALupHwC8uR8AvbkfAL6pHwC/oR8As7UfABHeAIAV3gCAGd4AgB3eAIC20R8AtaUfACHeAIC7yR8AuvUfACXeAIAp3gCAvyUfAL45HwC9PR8AvNEfAC3eAIAx3gCANd4AgDneAIA93gCA4WAfAEHeAIDjtBwARd4AgEneAIBN3gCA7wAdAFHeAIBV3gCAWd4AgF3eAICjNR4AYd4AgGXeAIBp3gCAbd4AgKZRHgClJR4Acd4AgKtJHgCqdR4AhKgCAHXeAICvpR4ArrkeAK29HgCsUR4AgE0AAIFVAACCVQAAs8kBAHneAIC12QEAtskBAH3eAICGoAAAhwQBALrFAQC7rQEAvLUBAL29AQC+tQEAv60BAKiZAQCpmQEAqg0BAKsFAQCsHQEArQUBAK4FAQCvNQEAgd4AgIXeAICJ3gCAjd4AgJHeAICV3gCAmd4AgJ3eAIC4JQEAuS0BALo5AQC7OQEAvCkBAL0pAQC+3QAAv9UAALBNAQCxJQEAsi0BALMlAQC0PQEAtSUBALYhAQC3HQEAod4AgKXeAICp3gCAo4kCAK3eAIClmQIApokCALHeAIC13gCAud4AgKqFAgCr7QIArPUCAK39AgCu9QIAr+0CAL3eAIDB3gCAxd4AgIRAAgDJ3gCAzd4AgNHeAIDV3gCAgA0AAIEVAACCHQAA2d4AgN3eAIDh3gCAh7QDAIbcBAC+zAMA6d4AgO3eAIDx3gCA7+gCAPXeAID53gCA/d4AgOP8AgAB3wCA4dABAAXfAIAJ3wCADd8AgBHfAIAV3wCAs2EDABnfAIAd3wCAId8AgCXfAIC2eQMAtXEDACnfAIC7XQMAul0DAC3fAIAx3wCAv+EAAL79AAC9/QAAvP0AALC5AgCxuQIAsgkBALMJAQC0GQEAtQUBALYFAQC3PQEAuAUBALllAQC6bQEAu2UBALxhAQC9YQEAvmEBAL9hAQCFXAcANd8AgDnfAIA93wCAJd0AgEHfAIBF3wCASd8AgKgxAgCpOQIAqskCAKvJAgCs2QIArdkCAK7JAgCvyQIAhMwFAOGAHgBN3wCA47weAOE4HgBR3wCA46AAAL4QBABZ3wCAXd8AgO8MHgBh3wCAZd8AgGnfAIBt3wCA73QeAKNhAgCCUQAAgUEAAICRAABx3wCApnkCAKVxAgB13wCAq10CAKpdAgCGyAQAhzwFAK/hAQCu/QEArf0BAKz9AQCohQYAqY0GAKqFBgCrmQYArIkGAK2JBgCuvQYAr7EGAFXfAIB53wCAfd8AgIHfAICF3wCAid8AgI3fAICR3wCAuJ0GALmtBgC6pQYAuwkHALwZBwC9GQcAvg0HAL8FBwCw0QYAsdEGALLRBgCz0QYAtLUGALW9BgC2tQYAt60GALMNBgCV3wCAmd8AgJ3fAICh3wCAtgkGALUBBgCl3wCAuxUGALoVBgCp3wCArd8AgL95BgC+cQYAvQUGALwFBgCx3wCA4aAEALXfAIDjXAUAgA0AAIE1AACCPQAAud8AgL3fAIDB3wCAhGADAL5sAAC/8AEAhZAAAMXfAIDvmAUAo40HAIQIAACGAAwAh4wAAMnfAICmiQcApYEHAM3fAICrlQcAqpUHANHfAIDV3wCAr/kHAK7xBwCthQcArIUHANnfAICz6QYA3d8AgOHfAIC26QYA5d8AgOnfAIC16QYAukUBALtNAQDt3wCA8d8AgL5FAQC/TQEAvFUBAL1NAQCoIQYAqSEGAKolBgCrPQYArCUGAK0tBgCuSQYAr0EGAPXfAID53wCA/d8AgAHgAIAF4ACACeAAgA3gAIAR4ACAuEkBALlJAQC6WQEAu1EBALx5AQC9eQEAvhkBAL8VAQCwxQEAsc0BALLFAQCz3QEAtMUBALXNAQC2xQEAt3kBABXgAIAZ4ACAHeAAgKOhBQAh4ACApaEFAKahBQAl4ACAjyHqAyngAICqDQIAqwUCAKwdAgCtBQIArg0CAK8FAgCX7RIAlmUSAJVFEQCUnRYAk3EWAJJVFQCReesDkFnqA59hBgCeNQUAnUUaAJxpGgCbVRkAmkUeAJlZHgCYRR0A4WAAAC3gAIDjTD4AMeAAgKOxAgCi1QEAobUHAKCJBgCxATgAsAk+ALOVOgCyjToAtbUmALQBJADvaDoAvjAMAKnJNgCowTYAqwEwAKrhNwCtzTMArPUyAK/5PgCuATwAoRkCADngAICjbQ4Aom0OAKX1CgCkAQgAp4ULAKaZCgCGAA0Ah0QNAIIJ6wODCesDhDHqA4UVFACGORcAh80XAISgDQA94ACAiiUQAIsNEwCMnRMAjQ0cAI4ZHwCPDR8A5d4AgO8AAwCSbRgAk0kbAJR9GwCVBQQAllkHAJdJBwBB4ACAReAAgJpFBgCbLQAAnFEDAONgAABJ4ACA4WwAAIClAQCBAQEAggUBAL4ADABN4ACAUeAAgFXgAIDviAEAWeAAgOFUBgBd4ACA41QBAGHgAIBl4ACAaeAAgG3gAICz6QIAceAAgHXgAIB54ACAfeAAgLadAgC1mQIAgeAAgLuJAgC6vQIAheAAgIngAIC/WQIAvlECAL1ZAgC8kQIAoykNAI3gAICR4ACAleAAgJngAICmXQ0ApVkNAJ3gAICrSQ0Aqn0NAKHgAICp4ACAr5kNAK6RDQCtmQ0ArFENAIBRAACBWQAAgmEAALMtDwCt4ACAtS0PALbJDwCx4ACAhkADAIcIAwC6yQ8Au8UPALzBDwC9wQ8AvsEPAL/BDwA14ACApeAAgLXgAIC54ACAveAAgMHgAIDF4ACAyeAAgKhFDgCpgQ8AqskPAKvJDwCsyQ8ArSUPAK4tDwCvJQ8AsGEPALFtDwCyeQ8As3kPALRpDwC1aQ8Ath0PALcVDwC4LQ8AuTUPALo1DwC7BQ8AvB0PAL3xAAC+8QAAv/EAAKNhDgDN4ACAhMQBANHgAIDV4ACApoUOAKVhDgDZ4ACAq4kOAKqFDgDd4ACA4eAAgK+NDgCujQ4ArY0OAKyNDgDl4ACA6eAAgO3gAIDx4ACA9eAAgPngAID94ACAAeEAgAXhAICCHQAAgR0AAIAdAAAJ4QCADeEAgBHhAIC+tAEAqK0BAKnVAQCq1QEAqwUBAKwdAQCtBQEArg0BAK8FAQCGgAEAhxgBABnhAIAd4QCAIeEAgCXhAIAp4QCALeEAgLiFAAC5jQAAuoUAALudAAC8hQAAvY0AAL6FAAC/vQAAsH0BALHhAACy5QAAs/0AALTtAAC13QAAttUAALe9AACzXQIAMeEAgDXhAIA54QCAPeEAgLaFAgC1lQIAQeEAgLslAwC6uQIAReEAgEnhAIC/GQMAvikDAL0pAwC8MQMAvswEAKMZAgBN4QCAUeEAgKbBAgBV4QCAWeEAgKXRAgCq/QIAq2EDAF3hAIBh4QCArm0DAK9dAwCsdQMArW0DAKgpAwCpKQMAqjkDAKs5AwCsKQMArSkDAK6dAACvlQAAZeEAgGnhAIBt4QCAceEAgHXhAICCqQEAga0BAICtAQC4mQAAua0AALqlAAC7bQAAvHUAAL19AAC+dQAAv20AALDtAACx9QAAsvUAALPFAAC03QAAtb0AALa1AAC3qQAA4XgBAOEcDgDjEAAA4zwOAHnhAIB94QCAvhQEAIHhAICErAIAieEAgId4BQCGDAUAjeEAgJHhAIDvvAAA70gOALPxAgCV4QCAmeEAgJ3hAICh4QCAtukCALXhAgCl4QCAu3EBALppAQCp4QCAhKAEAL85AQC+WQEAvVEBALxhAQCt4QCAhIwEALHhAICEADgAteEAgLnhAIC94QCAweEAgKqJDgCriQ4AqLkOAKmxDgCu/Q4Ar+EOAKz5DgCt9Q4Asq0OALNlDgCwkQ4AsaUOALZ9DgC3ZQ4AtH0OALV1DgC6XQ4Au+UNALhdDgC5VQ4AvuENAL/pDQC8/Q0AvfUNAKOxBQCF4QCAxeEAgMnhAIDN4QCApqkFAKWhBQDR4QCAqzEGAKopBgDV4QCA2eEAgK95BgCuGQYArREGAKwhBgDd4QCA4eEAgOXhAIDp4QCAgB0AAIEJAACCOQAA7eEAgPHhAID14QCAhsgAAIcMAwD54QCA/eEAgAHiAIAF4gCAqKUHAKm1BwCqvQcAq8kHAKzZBwCt2QcArskHAK/BBwC+oAAACeIAgA3iAIAR4gCAFeIAgBniAIAd4gCAIeIAgLjNAAC51QAAutUAALvlAAC8/QAAvZUAAL6dAAC/lQAAsIkHALFlBwCyYQcAs30HALRlBwC1bQcAtmUHALf1AACzNQYAJeIAgCniAIAt4gCAMeIAgLZZBgC1UQYANeIAgLuhBgC6TQYAOeIAgD3iAIC/qQYAvqEGAL2pBgC8tQYAQeIAgEXiAIDv8AUASeIAgE3iAIBR4gCAVeIAgFniAICAPQAAgQkAAIIdAABd4gCA4cgGAGHiAIDjSAQAZeIAgKO1BgBp4gCAhigAAIdAAQBt4gCAptkGAKXRBgBx4gCAqyEGAKrNBgB14gCAeeIAgK8pBgCuIQYArSkGAKw1BgB94gCAs70BAIHiAICF4gCAtnkBAIniAICN4gCAtXkBALpVAQC7XQEAkeIAgJXiAIC++QAAv/kAALxFAQC9+QAAqHECAKlxAgCqcQIAq3ECAKy1AgCtvQIArrUCAK+tAgC+rDwAmeIAgJ3iAICh4gCApeIAgKniAICt4gCAseIAgLhpAwC5aQMAugkDALsJAwC8HQMAvQUDAL4NAwC/BQMAsNUCALHdAgCy1QIAs2kDALR5AwC1eQMAtmkDALdhAwC14gCAueIAgL3iAICj9QIAweIAgKUxAgCmMQIAxeIAgMniAIDN4gCAqh0CAKsVAgCsDQIArbEDAK6xAwCvsQMA7xgCAIIVAACBbQAAgG0AANHiAIDZ4gCAhvg8AIcYAwDd4gCA4eIAgOXiAIDp4gCA42wHABXhAIDhaAEA7eIAgKiFAgCplQIAqpUCAKulAgCsvQIArdUCAK7RAgCv0QIA8eIAgPXiAID54gCA/eIAgAHjAIAF4wCACeMAgA3jAIC4dQEAuX0BALp1AQC7zQEAvNUBAL3dAQC+yQEAv8EBALC1AgCxvQIAsoECALOBAgC0VQEAtV0BALZVAQC3TQEA4bQGABHjAIDj9AYAFeMAgIQYPQAZ4wCAHeMAgCHjAIAl4wCAKeMAgC3jAIAx4wCANeMAgDnjAIDvWAYAPeMAgIF9AACAcQAAQeMAgIIFAABJ4wCATeMAgO+AAQC+VDwA4ZABAFHjAIDjfAYAVeMAgFnjAIBd4wCAhtg8AIf0PACjnT0A1eIAgEXjAIBh4wCAZeMAgKbVPQCltT0AaeMAgKv5PQCq8T0AbeMAgHHjAICvGT4ArhE+AK3VPQCs1T0AdeMAgLOhPgB54wCAfeMAgLatPgCB4wCAheMAgLWxPgC6ST8Au0k/AInjAICN4wCAvkk/AL9JPwC8ST8AvUk/AKhVPgCpZT4Aqm0+AKtlPgCsfT4ArWk+AK65PwCvuT8AkeMAgJXjAICZ4wCAneMAgKHjAICl4wCAqeMAgK3jAIC4VT8AuV0/ALpVPwC7bT8AvHU/AL19PwC+dT8Av20/ALDJPwCxyT8Astk/ALPZPwC0yT8Atck/ALZ9PwC3cT8AghUAAKPhPwCAsQEAgbEBAKbtPwCx4wCAvtABAKXxPwCqCT4Aqwk+AITkAQC14wCArgk+AK8JPgCsCT4ArQk+ALPdPAC54wCAhugAAIfMAQC94wCAtpU8ALX1PADB4wCAu7k8ALqxPADF4wCAyeMAgL9ZPwC+UT8AvZU8ALyVPACoUT4AqVE+AKptPgCrYT4ArGE+AK1hPgCulQEAr40BAISgAQDN4wCA0eMAgNXjAIDZ4wCA3eMAgOHjAIDl4wCAuKkBALmpAQC6aQEAu2kBALx5AQC9eQEAvmkBAL9pAQCw/QEAsc0BALLFAQCzrQEAtLkBALW5AQC2rQEAt6UBALPlPQDp4wCA7eMAgPHjAID14wCAtuE9ALXpPQD54wCAuwkCALo5AgD94wCAAeQAgL99AgC+fQIAvXkCALwRAgAF5ACAo6E9AAnkAIAN5ACApqU9ABHkAIAV5ACApa09AKp9AgCrTQIAGeQAgB3kAICuOQIArzkCAKxVAgCtPQIAgOkAAIHpAACCHQAAvsADAO/kAgAh5ACAh1QDAIY8BADjEAEAKeQAgOH4AQAt5ACAMeQAgDXkAIA55ACAPeQAgEHkAIBF5ACASeQAgLORAwBN5ACAtbkDALZ9AwBR5ACAVeQAgFnkAIC6WQMAu1kDALxJAwC9SQMAvv0AAL/1AACoRQIAqVUCAKpVAgCrZQIArH0CAK2xAgCusQIAr7ECAIRsBQBd5ACAYeQAgGXkAIBp5ACAbeQAgL5wBQBx5ACAuF0BALltAQC6ZQEAuw0BALwZAQC9GQEAvg0BAL8FAQCw0QIAsdECALLRAgCz0QIAtHUBALV9AQC2dQEAt20BAOFAPwDjvAAA4wg+AOFsPgB15ACAeeQAgH3kAICB5ACAheQAgInkAICN5ACAkeQAgL5sBwDvVAAA75w+AJnkAICjnQIAgmkAAIFhAACAaQAAneQAgKZxAgCltQIAoeQAgKtVAgCqVQIAhsgEAIfsBACv+QEArvEBAK1FAgCsRQIAqKUGAKmpBgCquQYAq7kGAKypBgCtqQYArtkGAK/ZBgCV5ACApeQAgKnkAICt5ACAseQAgLXkAIC55ACAveQAgLhxBwC5cQcAunUHALvdBwC8xQcAvc0HAL7FBwC//QcAsKkGALG1BgCytQYAs40GALSVBgC1UQcAtlEHALdRBwCzMQYAweQAgMXkAIDJ5ACAzeQAgLYpBgC1IQYA0eQAgLtxBgC6bQYA1eQAgNnkAIC/lQcAvlEGAL1ZBgC8YQYA3eQAgKN1BgDh5ACA5eQAgKZtBgDp5ACA7eQAgKVlBgCqKQYAqzUGAPHkAID15ACArhUGAK/RBwCsJQYArR0GAIANAACBFQAAgh0AAPnkAID95ACAAeUAgITcAQAF5QCAhoAAAIcgAQAJ5QCADeUAgBHlAIAV5QCAGeUAgB3lAIAh5QCA43QEACXlAIDhyAUAKeUAgC3lAIAx5QCANeUAgDnlAIA95QCAQeUAgEXlAIBJ5QCA77QEAE3lAIBR5QCAqD0GAKlVBgCqVQYAq6kBAKy5AQCtuQEArqkBAK+pAQCErAEAVeUAgFnlAIBd5QCAYeUAgGXlAIBp5QCAbeUAgLhtAQC5BQEAugEBALsBAQC8BQEAvQ0BAL4xAQC/MQEAsNkBALHZAQCybQEAs2UBALR9AQC1ZQEAtmUBALdVAQCBvQMAgL0DALPVBQCCGQAAtTkCAHHlAIC+VAMAtjECAHnlAIB95QCAuxUCALoVAgC9uQIAvLECAL+pAgC+sQIAgeUAgKZpAgClYQIAhAAMAKONBQCF5QCAhvgMAId8AwCv8QIArukCAK3hAgCs6QIAq00CAKpNAgCJ5QCAjeUAgJHlAICV5QCAmeUAgJ3lAIDjIAEAoeUAgOGgAQCl5QCA70ACAKnlAICt5QCAseUAgLXlAIC55QCAveUAgMHlAICz8QMAxeUAgCXkAIDJ5QCAzeUAgLbpAwC14QMA0eUAgLu1AwC6tQMA1eUAgNnlAIC/lQMAvpUDAL2lAwC8pQMAqCkCAKkpAgCqOQIAqzkCAKwpAgCtKQIArlkCAK9VAgCAzQEAgQkAAIIZAADd5QCA4eUAgL58DQCHtA0AhhwMALgxAgC5PQIAujUCALvpAgC8+QIAvfkCAL7pAgC/6QIAsDECALExAgCyMQIAszECALQRAgC1EQIAthECALcRAgDp5QCA7eUAgPHlAID15QCA+eUAgP3lAIAB5gCA79QGAAXmAIDhVAYACeYAgOOkAACsDBUADeYAgBHmAIAV5gCAo/ECABnmAIAd5gCAIeYAgCXmAICm6QIApeECACnmAICrtQIAqrUCAC3mAIAx5gCAr5UCAK6VAgCtpQIArKUCAKghDgCpIQ4AqkkOAKtZDgCsaQ4ArWkOAK6ZDgCvmQ4A5eUAgDXmAIA55gCAPeYAgEHmAIBF5gCASeYAgE3mAIC49Q4Auf0OALr1DgC7iQ4AvJ0OAL2FDgC+hQ4Av7UOALDpDgCx6Q4Asv0OALPxDgC01Q4Atd0OALbVDgC3zQ4As8EOAIIVAACBtQAAgLUAAFHmAIC26Q4AteEOAL4QAAC7LQ4Aui0OAIRkAwBV5gCAvxkOAL4RDgC9JQ4AvCkOAFnmAICjhQ4AhogAAIdsAwCmrQ4AXeYAgGHmAIClpQ4AqmkOAKtpDgBl5gCAaeYAgK5VDgCvXQ4ArG0OAK1hDgCziQ4AbeYAgHHmAIB15gCAeeYAgLaBDgC1iQ4AfeYAgLuVDgC6jQ4AgeYAgIXmAIC/+Q4AvvEOAL2FDgC8hQ4AieYAgI3mAICR5gCAleYAgOMMDQCZ5gCA4RgNAJ3mAIDvrAwAoeYAgKXmAICp5gCAreYAgLHmAIC15gCAueYAgKgBDgCpAQ4AqgEOAKsBDgCsAQ4ArQEOAK4BDgCvPQ4AgN0AAIEJAACCGQAAveYAgMHmAICEPAEAvnQAAMnmAIC4HQ4AuS0OALolDgC76QEAvPkBAL35AQC+6QEAv+kBALBJDgCxUQ4AslEOALNRDgC0NQ4AtT0OALY1DgC3LQ4Ao4kNAM3mAICGrAQAhzwDANHmAICmgQ0ApYkNANXmAICrlQ0Aqo0NANnmAIDd5gCAr/kNAK7xDQCthQ0ArIUNAOHmAICznQIAhEgDAL5ABAC2VQMA5eYAgOnmAIC1sQIAunEDALt5AwDt5gCA8eYAgL4xAwC/MQMAvFEDAL1RAwCwkQMAsZkDALKhAwCzoQMAtNEDALXRAwC20QMAt9EDALj1AwC5+QMAus0DALvFAwC83QMAvcUDAL7NAwC/xQMA9eYAgPnmAID95gCAAecAgIV8GQAF5wCACecAgHXlAICoIQIAqTECAKoxAgCrBQIArB0CAK3xAwCu8QMAr/EDAA3nAIAR5wCAFecAgBnnAIDvUAAAHecAgCHnAIAl5wCA44QAACnnAIDh+AEALecAgIAVAACBGQAAggUAADHnAICjmQMAOecAgIZoBACHYAUAPecAgKZRAgCltQMAQecAgKt9AgCqdQIARecAgEnnAICvNQIArjUCAK1VAgCsVQIATecAgFHnAIBV5wCAWecAgF3nAIBh5wCAZecAgO/4AQC+bAQA4YAOAGnnAIDjFAEAbecAgHHnAIB15wCAeecAgH3nAICB5wCAhecAgLPdAQCJ5wCAtf0BALb1AQCN5wCAkecAgJXnAIC6sQEAu4UBALydAQC9NQEAvj0BAL81AQCpBQYAqLkFAKsVBgCqHQYArT0GAKw9BgCvTQYArl0GADXnAICCHQAAgR0AAIAdAACZ5wCAnecAgKHnAICl5wCAuUEHALidBgC7QQcAukkHAL1FBwC8WQcAv0UHAL5FBwCxCQYAsD0GALOpBgCyAQYAtbkGALSxBgC3rQYAtrEGAKORBgCEjAIAhigAAIfAAwCp5wCAprkGAKWxBgCt5wCAq8kGAKr9BgCx5wCAtecAgK95BgCucQYArXkGAKzRBgC55wCAs5kHAL3nAIDB5wCAtlEHAMXnAIDJ5wCAtbEHALptBwC7dQcAzecAgNHnAIC+WQcAv0UHALxtBwC9ZQcA1ecAgNnnAIDd5wCA4ecAgOXnAIDp5wCA7ecAgO+oBQDx5wCA4TQFAPXnAIDjdAUA+ecAgP3nAIAB6ACABegAgKMdBgCCLQAAgRUAAIAdAAAJ6ACAptUGAKU1BgAN6ACAq/EGAKrpBgAR6ACAhCgBAK/BBgCu3QYAreEGAKzpBgCoxQYAqdUGAKrVBgCr5QYArP0GAK0VBgCuHQYArxUGAL7sAQAZ6ACAhggAAIcgAAAd6ACAIegAgCXoAIAp6ACAuH0GALkFBgC6DQYAuwUGALwBBgC9CQYAvjkGAL85BgCwbQYAsXUGALJ9BgCzdQYAtFkGALVFBgC2TQYAt0UGAKiRAgCpmQIAqqECAKuhAgCs0QIArd0CAK7VAgCvyQIALegAgDHoAIA16ACAvyweADnoAIA96ACAQegAgEXoAIC4VQMAuV0DALppAwC7ZQMAvGEDAL1hAwC+YQMAv2EDALC5AgCxjQIAsoUCALNtAwC0dQMAtX0DALZ1AwC3bQMASegAgE3oAICzIQIAUegAgLVRAgCEiAMAVegAgLZVAgDF5gCAvigcALtBAgC6dQIAvbEDALxZAgC/sQMAvrkDAKNpAgBZ6ACAXegAgGHoAIBl6ACAph0CAKUZAgBp6ACAqwkCAKo9AgBt6ACAcegAgK/5AwCu8QMArfkDAKwRAgCopQIAqbUCAKq9AgCrtQIArK0CAK01AQCuPQEArzUBAL4sHAB16ACAeegAgH3oAICB6ACAiegAgIdoHQCGHB0AuIUBALmNAQC6hQEAu50BALyNAQC9vQEAvrUBAL95AACwUQEAsVEBALJRAQCzUQEAtPEBALXxAQC29QEAt+UBAO/YAACCtQAAgaUAAIClAACN6ACAkegAgJXoAIDvxAYAmegAgOH0BgCd6ACA4zgBAOPMAACh6ACA4SgBAKXoAICp6ACAtuUBALV1AgCEQBwAs2UCAK3oAICx6ACAtegAgL9lAQC+ZQEAvdUBALzVAQC7xQEAusUBALnoAIC96ACAo7UdAIXoAIDB6ACAxegAgMnoAICmNR4ApaUdAM3oAICrFR4AqhUeANHoAIDV6ACAr7UeAK61HgCtBR4ArAUeANnoAIDd6ACA4egAgOXoAICADQAAgTUAAII9AADp6ACA7egAgPHoAIC1BQAAdxoAgOG0AgCs2AIAtQUAAHsaAICotR8AqRUfAKodHwCrFR8ArDEfAK09HwCuLR8AryEfAOG0AgCs2AIAtQUAAH8aAIDhtAIArNgCALUFAACDGgCAuNEAALnZAAC64QAAu+EAALyRAAC9kQAAvpEAAL+RAACwIR8AsTEfALIxHwCzMR8AtAkfALUJHwC28QAAt/EAAOG0AgCs3AIA71QdALUdAACHGgCA4bwCAKzQAgC1KQAAoyUBAKKRAwChFR0AoA0dAOGAHgCLGgCA47wdAOHEAgCz1R4AtQkAAKzYAgCPGgCA4bwCALb9HgC1+R4ArOACALu1HgC6pR4AtQUAAJMaAIC/jR4Avo0eAL2lHgC8pR4AoxUeAOG8AgCs0AIAtREAAI9pJQCmPR4ApTkeAJcaAICrdR4AqmUeAOG0AgCseAEAr00eAK5NHgCtZR4ArGUeAJvdFACa5RUAmQEXAJjhEACfcR8AnnkZAJ35GQCcARsAk+UtAJIRLwCRbSkAkG0pAJf5EQCW8REAlYUsAJSZLQC1JQAA4ZQCAILxJgCDjSoAhJUqAIXhLACGHS4Ah3kuAKy0AgCbGgCAilUvAIspEgCMORIAjRkTAI7xFACPHRYAtQUAAJ8aAICSVRcAk5EYAJRxGgCV+RoAlvkcAJd9HgCC4AMAlgsAgJpVHgCb2QAAnHUCAIYMAIC2DACAuIkKAKwBBACthQYAroEGAMwQAgDMfAMAuQwAgKMaAIDFDACAyAwAgMsMAIADCwCAgaUyAr8MAIAV6ACAmpUGAJtVIwK8kQYAvbEAAL6RBgC/rQYAuOkGALmVBgC6kQYApxoAgLTBBgC1zQYAts0GALfdBgCw/QYAseUGALKdAACz5QYAhVTHA6saAICH/AAAuAEKALMaAIDsDACAtxoAgIzlDACNpAEAzPACAMQNAIDHDQCAiRQAALgZCgCLDAAAIA4AgFkOAIC8DACAwgwAgBwKAICRwAEAzgwAgLhtCgDRDACA1wwAgN0MAIDgDACA4wwAgLsaAIAuDQCA6QwAgL8aAIDhpB4AMQ0AgONUHgCvJXMAzCgCAPIMAIDvDACA9QwAgPgMAID7DACAzIACAJS4AwD+DACAkhQCAO9gHgCQAAIAAQ0AgA0NAIC48QoAEA0AgKILAIATDQCAiSkLABYNAICvGgCAvDABAL/EAQC+7AEAGQ0AgMzsAgC4xQoAukQBALAJAIAcDQCAygYAgN8GAIDyBgCAIg0AgPoGAIAlDQCACgcAgC0HAIAYBwCA+QcAgC8HAICvDQCAOgcAgLUNAIBKBwCAtXkAAGoHAIC3cSoCdQcAgLFhAAB3BwCAsw0pApAHAIC96QAAowcAgP0HAICwBwCAuRkrAsYHAIC7WRQCIggAgF0JAIA/CACANQ4AgF4IAIA5AACAhAgAgHEAAIDKCACAKwAAgCMJAIA9AACAXwkAgEMAAIBhCQCASAgAgG0IAIBJAACAAwgAgFMAAIB8CQCAWQAAgCgNAIBfAACAuw0iAtYNAIDMFDYCHwAAgL9lAAC+EQAAvW0AAOgHAICAaQEAgXUBAIJxAQCD3SEChGkHAIWBBwCGgQcAh3EBAIihAQCJrQEAirUHAIuNBwCMlQcAjaUBAE8AAICPpQEAkOEBAJHtBwCSsSECk/0HAJSNBwCVUQYAlvEBAJfZAQCY0QEAmXUGAJp9BgCb1QEAnGkGAJ2ZFAKeUQYAn1EGAKB1FAKhuQYAokkBAKOFLQKkIQEApS0BAKZ1FAKntQYAqKERAqlRFAKqlQYAtyEAgMy8NQLNPDUCbQAAgKoDAICsAwCArwMAgMMhAIDKIQCA4SEAgOghAIDJAACADwAAgLihBgC6BgCAtwYAgMwAAIDUIQCAtQMAgN0FAIAYBgCAugUCALvVAgC46QUAuf0FAL7JAgC/5RcCvA0CAL0BAgCy4QUAs+EFALCNBQCxnQUAtuUFALfpBQC09QUAte0FAKo9BQCrwQUAqD0FAKk1BQCuzQUAr/UFAKzNBQCtxQUAoj0FAKMFBQCg1QIAoTkFAKYdBQCnBQUApB0FAKUVBQC/BgCAm8EFAD4GAIBVBgCAnt0FAJ8xBACcUQIAndUFAHIGAICJBgCApAMAgDYiAIDbAACAoAMAgJIHAIDxBwCA9QcAgJMJAIAFCACACQgAgJkLAICXCQCAsgoAgHIHAICOBwCAmgcAgKUHAICtBwCArQkAgAEPAIAYDwCAJQ8AgMwEMwLNsDACzCAzAs3gMALMEDACzGgwAsxYMALNjDACzGgxAs0UMQLM1DECzRQ2AsxwIALN0CcCzDA2AswkMQLMDDwCzWg/AswYPwLNND8CzBg9As3AMgLMRDwCzBg5Asw4MgLNqDICzIgyAs34MwLMfDMCzUAzAswoMwLNCDMCzMghAs0kJgLMrCYCzEA4AsyYJQLNyDoCzBwkAs0QJALMhDsCzag7AsysJQLNvDoCzKw4Asz4JwLM4DgCzXQ4Ai0PAID2BgCAZw0AgI4NAIDNICoCzBwrAqoGAIAyIgCAzKQgAs2gJwLMOCYCygQAgMw4OgLNPDsCzBA5As1gPgLMoAMAvj0NAL3tLALWBACAu1UjAgcJAIC5PSICzwYAgNwHAIClBACApg0AgLIEAIBvBQCA9AYAgL4EAIB1BQCAr70MAK6ZLgKtpQwAwgUAgKvFIgIDBgCAxAQAgCMGAIDQBACAyAUAgCkGAIBdBgCAowEYAqAEAIAaBwCAHQcAgJ9dDACeUQwAnUUMACcHAICbWSECsgcAgLQHAIC3BwCAuwcAgCoHAIDRBwCA0wcAgJMtJgLWBwCAbwgAgHIIAICPBQwAjnEMAI1lDAB8CACAi0UgAmMJAICJNS8CZgkAgGoJAIB/CACAcwkAgHYJAIC9AwCABiIAgIFdDACAYQwAgAABAIEYAACCAAQACiIAgIQQBwCFFAYAhuQIAIc8AgCILAUAiaQFAIoAeAAOIgCAjJAGABIiAIAaIgCAFiIAgLgRAACRxHsAkkh6AJNMeQAiIgCAzOgCAJbwCQC4OQAAkMAJACoiAICS8AkAzPgCAJS0CQC4DQAALiIAgMwcAgC4BQAAOiIAgMzkAgC4HQAAPiIAgEIiAIBJIgCAYCIAgKiMCACp5HsAZyIAgKt8BADM5AIAuA0AAHEiAIDMlAIAdSIAgLGUewC4CQAAuBUAAMz8AgC15AgAeSIAgMzYAgB9IgCAuAUAALroBQC7hAQAvAB8AL30fwC++H0Av/xyAIAJOgKBDToCggE6AoMFOgKEGToChR06AoYROgKHFToCiCk6AoktOgKKIToCiyU6Aow5OgLMhAIAjjE6Ao81OgKJIgCAkekPAIUiAIDMzAIAuBEAAJUiAIDM3AIAl+UPALgpAAC4MQAAzPgCAJkiAIC4HQAAzDwCAJ0iAIDMLAIAzIQCAKEiAIClIgCAzIQCAKQtDwClVQ8Apl0PAKkiAICoqToCqa06ArjRAACtIgCAuDUAALEiAIDMUAMAr7U6AswsAwDMFAMA1SIAgLMFDwC0HQ8AzAADALYJDwC3CQ8Avmh9ALhtAAC4dQAAzEgDALwpDwDZIgCAviUPAMxQAwCH5Q4AzOg6ArilAQC4uQEAzPA1As2kMwLMgCICzXwlAs2UNgLMBCkCzew7AsxkOgK4+QEAuMEBAInVDgCI1Q4Al7EOALgNAAC1IgCAuB0AALkiAIC4DQAAvSIAgMEiAICfaTsC3SIAgOEiAIC4MQAAzMQCAMz4AgDFIgCAySIAgLjlAADNIgCA0SIAgLjlAAC46QAAuOkAAMzMMwLlIgCAuCUAAMzoMwK4IQAA6SIAgO0iAIC4KQAAzNgCALgRAAC3TQ0Atk0NALU1DgC0NQ4AuGEAAPEiAICxGQ8AsCkOAL/1AwC+UQ0AvVkNALw1DAC7XQ0Aul0NALldDQC4XQ0AgL0KAIHFCgCCFQQAg8kKAMx8BQCF3QoAhtUKAIfNCgDMUAUAifEKAIq5CACLDQgAjBEIAI0VCACOtScCj+UKAJBpCACRbQgAknEIAJNtJALMJAUAlR0IAJaFCgDMHAUAzCgFAJk9CACaiQoAmw0IAJwRCACdFQgAzCwFAMwgBQCgZQoAoW0KAKJlCgDMvAUApLEEAMzoAgCmsQQAuEkHAKiBBAAbIwCAqpkIAKtdCgCsuQgArakEAB8jAICvNQgAsNEIALHxBADQAwCAs40IALQpKAK1IQoAtiEKALchCgC4IQsAuSUIAOkDAIC7KQsAvA0dAr3dDwC+MQsAvzELAIDdCgCpoQEAqrEBAAIEAIAbBACAhRkJAIaZCQCHlQkAiOEJAIklJQIuBACAQQQAgFQEAIBnBACAegQAgI0EAICQrQoAkUkFAJJtBQCTYQUAlGEFAJVtBQCWZQUAlxEFAJg1BQCZPQUAmjUFAJsNBQCcFQUAnR0FAJ4VBQCfCQUAoKkJAKH9BQCi9QUAowEFAKQFBQClDQUApgUFAKc9BQCoBQUAqQ0FAKoFBQCrGQUArIkJAK2pBQCutQkAr/0JALABCQCxfQUAsnUFALMBBQC0aQkAtQEFALYFBQC3PQUAuAUFALnhJQK6AQUAuwEFALzRJQK9PQkAvnkJAL9dCQCDMAUAoXgHAPMEAIDdAACApHgHAKVMBwATAQCAHAEAgIt8BAAgAQCAJAEAgIhIBAAoAQCALAEAgDABAIA0AQCA4QAAgOYAAICyDAcAswAHAOsAAIDwAACAtugHALfgBwD1AACA+gAAgLrgBwC7nAcAvIQHAL2QBwD/AACAn9F+AKPMBAAEAQCACQEAgIMABAAOAQCAhXQEAKUgBAAXAQCAiEwEAM0DAIDwBACAjwUAgK8tBwCNsAcArSEHAKwpBwDiBQCAHQYAgEMGAICwZQcAWgYAgHcGAICOBgCA0wMAgOwDAIAFBACAHgQAgDEEAIBEBACAVwQAgGoEAIC8fAQAgt0rAoPlKwKA/QoAgfkrAoaZCQCHmQkAhOEKAIXhCgCKiQkAi4kJAIiJCQCJiQkAjoUJAH0EAICM4QgAjY0JAJK5KwKTQScCkJkrApHFCwCWyQsAl3UnApTFDQCV0SQCmskLAJvZKgKYyQsAmXkHAJAEAID2BACAnP0LAKABAICkAQCAqAEAgKwBAICwAQCAtAEAgLgBAIC8AQCAwAEAgJw9fwCzFX8AqBEJAMQBAIDIAQCAzAEAgNABAICC2H4A1AEAgNgBAIDcAQCA4AEAgOQBAIDoAQCA7AEAgPABAID0AQCA+AEAgPwBAIAAAgCABAIAgDwJAIBTIgCAogYAgKD1VAKh2VQCoulUAqP1dQCk7XUApZ12AKaVdgCnvXYAqIV2AKntfACqwXwAqyF9AKwhfQCtHX0ArhV9AK8NfQCwdX0AsX19ALJ1fQCzRX4AtF1+ALVNfgC2RX4At3l+ALhJfgC5VX4Aul1+ALtVfgC8TX4AvTV+AL49fgC/LX4AlQcAgKwGAIDaBwCArwYAgLRtfwC1EQAAthUAAAkjAIC8oVgCvTF4AA8jAIDTMQCAPzoAgJ8qAIDDKgCAzyoAgN8qAIDnKgCA8yoAgPsqAIADKwCADysAgGorAICCKwCAkisAgKIrAICyKwCAwisAgOIrAIDmKwCA6isAgB4sAICAVX8AgWV/AIJtfwCDdX8AhJV/AIWdfwCGiX8Ah4F/AIiFfwCJjX8AioV/AIvtfwCM9X8Ajf1/AI7pfwCP6X8AkJl/AJGZfwCSqX8Ak6l/AJS5fwCVuX8Alql/AJepfwCYmX8AmVF+AJoZfgCbGX4AnA1+AJ31fgCe/X4An/V+AKANfgChFX4Aoh1+AKMVfgCkDX4ApTl+AKYpfgCnKX4AqBl+AKllfgCqbX4Aq2V+AKx9fgCtZX4Arm1+AK9lfgCwHX4AsSV+ALItfgCzJX4AtD1+ALUlfgC2WXcAt9V1ALj9eQC56XUAuvl1ALvZeQC86XUAvdV1AL7RdQC/2XUAgDF2AIE9dgCCSXYAg0V2AIRBdgCFTXYAhvl0AId9dgCIoQIAiU12AIpZdgCLuXoAjEl2AI2degCOsQIAjx16AJCRVgKRKXYAkoF2AJPNdgCU2XYAlel2AJbJdgCX0VkCmKF2AJllWgKa8XYAm01aApzRdgCdYXoAnoFWAp/VdgCgdX0AoY1aAqI1VwKjCXYApCF2AKUtdgCmiVoCp5laAqi5WgKpdXYAql13AEYsAIBWLACAXiwAgGIsAIBuLACAiiwAgI4sAICmLACAqiwAgLIsAIDCLACAXi0AgHItAICyLQCAxi0AgM4tAIDSLQCA4i0AgAUuAIAxLgCAPS4AgJlpCgBdLgCAaS4AgG0uAIBxLgCAiS4AgI0uAIC5LgCAgvx6AIPoegDFLgCAzS4AgIZwewCHcHsA1S4AgOUuAID0LgCA/C4AgCgvAIAsLwCANC8AgDgvAIBALwCASC8AgJJAfABYLwCAdC8AgJHkewDsLwCAADAAgAQwAICEMACAiDAAgKvUfACo6HwAqeB8AJwwAICgMACAqDAAgLAwAICisHwAuDAAgMQwAID6MACAzEBJAs0ASQLM/EoCzWhLAgoxAIAeMQCAmzEAgKcxAIC3MQCAwzEAgM8xAIDXMQCAsqh8ALOofADbMQCA3zEAgOMxAIDnMQCAtER8ALVYfACAfQcAgYUHAIKZBwCDmQcAhIkHAIWJBwCGvQcAh7UHAIiNBwCJ5QcAiu0HAIvlBwCM/QcAjeUHAI7tBwCP5QcAkJ0HAJGtBwCSpQcAk70HAJSlBwCVVQEAll0BAJdVAQCYbQEAmXUBAJp9AQCbdQEAnG0BAJ1VAQCeXQEAn1UBAKCtAQChtQEAor0BAKO1AQCkrQEApdUBAKbdAQCn1QEAqO0BAKn1AQCq/QEAq/UBAKztAQCt1QEArt0BAK/VAQCwrQEAsbUBALK9AQCztQEAtK0BALVVAQC2XQEAt1UBALhtAQC5dQEAun0BALt1AQC8bQEAvVUBAL5dAQC/VQEAnzIAgOcyAIDzMgCA9zIAgPsyAID/MgCABzMAgAszAIAfMwCAOzMAgEMzAICDMwCAhzMAgI8zAICTMwCAmzMAgJ8zAIDDMwCAxzMAgOMzAIDnMwCA6zMAgO8zAIADNACAJzQAgCs0AIAvNACAUzQAgJM0AICXNACAtzQAgMc0AIDPNACA7zQAgBM1AIBXNQCAXzUAgHM1AIB/NQCAhzUAgI81AICTNQCAlzUAgK81AICzNQCAzzUAgNc1AIDfNQCA4zUAgO81AID3NQCA+zUAgP81AIAHNgCACzYAgKs2AIC/NgCA8zYAgPc2AID/NgCAnpkMACs3AIAzNwCAOzcAgICtAwCBtQMAgr0DAIO1AwCErQMAhdUDAIbdAwCH1QMAiO0DAIn1AwCK/QMAi/UDAIztAwCN1QMAjt0DAI/VAwCQrQMAkbEDAJKxAwCTsQMAlAEMAJVVDgCWXQ4Al1UOAJhtDgCZdQ4Amn0OAJt1DgCcbQ4AnVUOAJ5dDgCfVQ4AoK0OAKG1DgCivQ4Ao7UOAKStDgCl1Q4Apt0OAKfVDgCo7Q4AqfUOAKr9DgCr9Q4ArO0OAK3VDgCu3Q4Ar9UOALCtDgCxtQ4Asr0OALO1DgC0rQ4AtVUOALZdDgC3VQ4AuG0OALl1DgC6fQ4Au3UOALxtDgC9VQ4Avl0OAL9VDgC8aQQAvXUEAL59BAC/dQQAuEUEALlNBAC6RQQAu3kEALRlBAC1bQQAtmUEALd9BACwHQQAsQ0EALIFBACzfQQArFUEAK1dBACuVQQAr2UEAKhVBACpXQQAqlUEAKtNBACkkQUApZEFAKaRBQCnkQUAoJEFAKGRBQCikQUAo5EFAJxRBQCdUQUAnlEFAJ9RBQCYUQUAmVEFAJpRBQCbUQUAlBEFAJURBQCWEQUAlxEFAJBFBwCRTQcAkkUHAJMRBQCMJQcAjS0HAI4lBwCPPQcAiAUHAIkNBwCKBQcAiz0HAIQlBwCFLQcAhiUHAIc9BwCARQcAgU0HAIJFBwCDPQcAfzcAgIM3AICLNwCAjzcAgJM3AIC/NwCAwzcAgMs3AIDfNwCA4zcAgP83AIAHOACACzgAgC84AIBPOACAYzgAgGc4AIBvOACAmzgAgJ84AICvOACA0zgAgN84AIDvOACABzkAgA85AIATOQCAFzkAgBs5AIAnOQCAKzkAgDM5AIBPOQCAUzkAgFc5AIBvOQCAczkAgHs5AICPOQCAkzkAgJc5AICfOQCAozkAgKc5AICrOQCArzkAgL85AIDXOQCA2zkAgOc5AIDrOQCA7zkAgPM5AID7OQCA/zkAgAM6AIAPOgCAFzoAgB86AIAjOgCAKzoAgC86AIAzOgCAOzoAgICtAQCBtQEAgr0BAIO1AQCErQEAhdUBAIbdAQCH1QEAiO0BAIn1AQCK/QEAi/UBAIztAQCN1QEAjt0BAI/VAQCQrQEAkbUBAJK9AQCTtQEAlK0BAJUNAABDOgCAQyMAgHIsAIB2LACAKyQAgIJ4AgCZBQAAuSMAgILQAgC+mAIAgIAAAIGYAACC5AYAg4gEAITUGwCFlBoAhhgfAJ7pAACIxB4AiQAQAIqoEwCLrBEAjAAoAI20KwCOuCoAj7wpAON0dADjoAIA47gCAJkdAAC9IwCAvnQCAJ4JAADv2AIAgmwCAMEjAICZDQAA4wQCAO/ocQDvcAIA79QCAL6MAADjBAMAxSMAgOMcAwDJIwCA4yADAM0jAIDjTAMA0SMAgO/kAwDVIwCA74QDANkjAIDv2AMA3SMAgO/sAwDhIwCA48gDAOPcAwDvoAQA4+gDAEM3AIDjqAMA5SMAgO+kBADXAAAA70wDAOkjAIDjAAQA7yADAO88AwDjeAQA70QDAO0jAIDxIwCA9SMAgPkjAIDvWAQA/SMAgO9cBADjgAQA44AEAAEkAIDjmAQA73QEAAUkAIAJJACADSQAgBEkAIAcJACA72AEAOOIBAAgJACAvQAAgMIAAIA3JACAJCQAgHMpAIAOJQCAcSUAgLQlAIDgJQCA46QEAO9EBAAKJgCAgAlLAoZIAQCe7QIAgnACAL58AgCeEQEAmR0BAIJcAgCPQAEAmSkBAI1sAQC+YAIAi3gBAJ45AQCCGAIAvggCAJfUAgCZUQEAldQCAJ5ZAQCT1AIAvlgCAJHUAgCCVAIAn7gCAJYAAACdqAIAmXEBAJu4AgCeaQEAmbQCAJlZAQCCmAIAppQCAJ6tAQCkeAIAgmgCAJ65AQCheAIAmbEBAK+EAgAvJgCAvlgCAIJwAgC+XAIARCYAgJmNAQCokAIAvnwCAJ7xAQC1sAIAmfEBAL5QAgCykAIAtoUCAJmFAQC4XQkAuYUCALqNAgBMJgCAu/QEAIJcAgCexQEAuPQEAL58AgCeXQYAgmACAJllBgC+cAIAgmwCAJ5xBgCZnQYAvnwCAJ6lBgCCYAIAmakGAL58AgCesQYAghwCAL4UAgCZyQYAvkwCAIJMAgCawQYAntkGAJ/ZBgCCvAIAvrACAJn1BgCC3AUAvogCAJrJBgCe5QYAn9EGAIK4AgCCqAIAmTEGAOP0AgDj/AIAmjkGAJ8lBgCeJQYAniEGAJ8hBgC+GAIAmR0GAJoVBgC+DAIAmXEGAO8kAgDv4AIAmnEGAJ4BBgCfCQYA4xQCAJkVBgDjLAIAgnwCAJ4BBgBoJgCA7/ACAJkFBgC+fAIAng0GAHAmAICCdAIA76wCAIKUAwCb0QcA4/ADAL5QAgCZ6QcAn80HAOM0AgCdXAIAnMkHAJ7FBwDv7AIAgmACAJnFBwC+cAIA78QCAJ7RBwCCEAIA43QCAL5sAgCZpQcAvlQCAJ69BwCZpQcAgmgCAOMMAgCekQcAmZkHAIJoAgDv2AIA78wCAL6IAgCCkAIAvpgCALsEAACeeQcAuQAAAJkpBgC/XAAA4/QCAL0EAACeOQYAs7gDAO8oAgCxeAMAgnwCALc8AACZAQYAtfwDAJ4JBgCrSAMAvgACAOO4AgCZIQYArxwDAIJ8AgCtQAMAvkQCAJ4NBgCCbAIAvpwBAJkxAQCCgAEApqgCAO98AgCCvAEA46QBAOPYAQDj5AEAntECAOPoAQCZ5QIAvnwCAJ7tAgDvwAIAmQkAAL5sAgB4JgCA7wwBAO8QAQDvKAEAnhEAAOO0AQDjHAIAmS0AAL5sAgCCfAIAglACAJ49AADjAAIAmeUAAIomAIC+aAIA7/wCAO+gAgDvwAIAnv0AAJnxAADjAAIAliYAgIKQAgCeJgCAvnwCAJ4ZAADjDAIAglgCAJkNAACCBAIA7+QCALQmAIDv2AIAvnQCAJ4VAACZZQEA42wCAJnJAQCanQEAvmQCAJ7dAQCfgQEA4wwCAIJYAgCZlQEAvrQDAO/0AgCalQEA78QCAIKsAwCwRSoAgvQDAIHIAgDj+AMAvjACAPImAIDj5AMAhMQCAL44AgCGEAIA78ACAIgwAgCeXQAAn1UAAJqxAADvxAIAj3QCAJlBAACePQAAn8UAAIKMAgCSGAIAldgDAAMnAICeyQAAn8kAAJr1AACY+AMAm2ADAJn9AACCqAMAJScAgDAnAICAJwCAlCcAgOPEAgC+IAIAvlACALYnAIDEJwCA4+ACAAwoAICZPQAAnn0AAO/cAgCaHQAAGigAgO/EAgCvrAMAghACALEAHACwlAMAmSkAALJMHACeTQAAn2UAAHcpAIDjaAIAeykAgL50AgCeCQAA7/ACAL08HACCbAIAv80fAJn9HwB/KQCAvnQCAJ4JAADjFAIAgmwCAIMpAICZDQAAvkQCAJ41AACCaAIAmQUAAI8pAIC+cAIA7yUAgJ4VAADv2AIA42wCAP8YAIADGQCA47AdACcaAIAHGQCAKxoAgC8aAIALGQCAMxoAgDcaAIA7GgCA77gCAD8aAIBDGgCA74ACALHFAAAPGQCAs9kAALLFAAC1yQAAtMEAALf5AAC2wQAAuWEAALghAAC7wQAAuskAAL3FAAC82QAAv30AAL7FAAATGQCARxoAgFMZAIAXGQCAGxkAgB8ZAIBnGQCA7xB4A7MZAIDh3E0DtxkAgONYeQO7GQCAvxkAgMMZAIDHGQCAgMkBAIHJAQCC2QEAg9kBAITJAQCFdQIAhgEEAIcdBQCIJQUAiTUFAIo9BQCLbQUAjHUFAI1lBQCObQUAj90BAJCpAQCRtQEAkr0BAJO1AQCUrQEAlVUDAJZdAwCXVQMAmG0DAJl1AwCafQMAm3EDAJxRAwCdUQMAnlEDAJ9NAwCgtQMAob0DAKK1AwCjxQMApMEDAKXBAwCmxQMAp/0DAKjFAwCpzQMAqsUDAKvZAwCsyQMArTUDAK49AwCvNQMAsE0DALFVAwCyXQMAs1UDALRNAwC1dQMAtn0DALd1AwC4TQMAuVUDALpdAwC7VQMAvE0DAL01AwC+PQMAvzEDAMsZAIDPGQCA0xkAgNcZAIDbGQCA3xkAgPAQAgDjGQCA5xkAgOsZAIDvGQCAgpwCAPMZAID3GQCA+xkAgP8ZAICc9TYAnf02AAMaAICRaAIArxkAgEsZAIC6ZdUATxkAgEsaAIBPGgCAUxoAgFcaAIDwSAIAWxoAgF8aAICRkAIAYxoAgFcZAIBnGgCAaxoAgFsZAIBfGQCAYxkAgGsZAIBvGQCAcxkAgHcZAIB7GQCAfxkAgIMZAICHGQCAixkAgI8ZAICTGQCAuo3VAJcZAICbGQCAnxkAgG8aAIBzGgCAoxkAgILEAgCnGQCAqxkAgAcaAIALGgCADxoAgPA4AwDh3NICExoAgONUxgIXGgCAGxoAgB8aAIAjGgCAqyoAgGItAICvKgCAxyoAgLMqAICjMwCAuyoAgO+k4AKjKgCA4xSHAuGQ8wLhjJ0C45D3AuGMqwLhkLYC4wBTA+OAogIjGQCA7S0AgO+0SwPvvLMC7ySLAnYtAIAIAgCA7ySXApEIBwAOAgCAFAIAgBoCAIAgAgCAJgIAgCwCAIAyAgCAOAIAgD4CAIBEAgCASgIAgFACAIBWAgCANAMAgDoDAIDhgHgC4wAeAuMUagLhEBMC4aAPAkADAIDjhA4C7yw/AkYDAIDhUDsC7zQ7AuM8IgJMAwCA7ygfAu8MEgJSAwCAKxkAgC8ZAIBYAwCAXgMAgDMZAIA3GQCAdgMAgIIDAICIAwCAjgMAgJQDAICaAwCAfAMAgGQDAIBtAwCAXAIAgDsZAIA/GQCAdAIAgGgCAIBDGQCARxkAgLwCAIB6AgCAmAIAgGICAICSAgCAbgIAgKQCAIDUAgCA8gIAgOwCAICAUQYAgVEGAIJRBgCDUQYAhHEGAIV9BgCGdQYAh20GAIhVBgCJXQYAiq0HAIuhBwCMoQcAjaEHAI6hBwDgAgCALgMAgMICAICSfRQAkwEUAJTNBwCV9QcAlv0HAJf1BwCYzQcAmdUHAJotFACb2QcAnM0HAJ2RBwCejQcAnykUAJklAQCYJQEAmyUBAJolAQCdJQEAnCUBACcZAICeJQEAkcEGAJDxBgCT1QYAkt0GAJU9AQCUPQEAlyUBAJYlAQCJ5QYAiOUGAIvlBgCK5QYAjeUGAIzlBgCP5QYAjuUGAIHlBgCAHQYAg+UGAILlBgCF5QYAhOUGAIflBgCG5QYAuaUDALilAwC7pQMAuqUDAL2lAwC8pQMAv6UDAL6lAwCxpQMAsKUDALOlAwCypQMAtaUDALSlAwC3pQMAtqUDAKmxAQCoqQEAq7EBAKq5AQCtbQEArKkBAK8dAQCuHQEAoakBAKDZAQCjsQEAoqEBAKWRAQCkqQEAp5EBAKaZAQDOAgCA5gIAgNoCAIAEAwCAsAIAgPgCAIAiAwCACgMAgJ4CAICAAgCAtgIAgMgCAID+AgCAhgIAgCgDAICqAgCAEAMAgIwCAIAWAwCAHAMAgBYtAID4LgCA1zQAgIcHAIAGBQCAFQUAgCQFAIAzBQCAQgUAgEsFAIBUBQCAXQUAgIIMAwBmBQCAkgUAgJsFAICkBQCA40huA6cFAIDhTG4DqgUAgO/0AQOtBQCAVzoAgLdMAIDnVQCAR2gAgHdxAICnegCAB40AgGefAICXqACA/roAgDXEAIBlzQCAldYAgMXfAIBCuwCAS64AgBelAID/KgCAlisAgKcqAIDrKgCATjEAgA4xAIBbNACA4iwAgBMzAICXNwCAbzQAgCosAICfNACAqzMAgB84AIBmKwCAkiwAgAcyAIA3OQCAKisAgLorAICrMQCAyS4AgNYsAIBmLACARS4AgDkuAID7MwCAJisAgLop0ACrNwCAgiwAgNotAICwBQCAswUAgLYFAIDh1D8D4VgaA+PcLwPjUA4D4RTyA+FA0wPjQOoD40DDA7kFAIDlBQCA73jrA+9c8gPoBQCA6wUAgO9E3gPvmCUD4bSLA+E8lwPjfKID45iLA+EwQQDhUKwD4xx/AOOIRgDuBQCA8QUAgO84ewDv4EEA9AUAgPcFAIDvzIoD7yCHA4C1GACByRgAghULAIMtCwCE4Q4AheEOAIbhDgCH/RgAiN0OAImZGgCK6RsAiy0dAIzFGwCN8RsAjn0QAI/hGgCQ6RsAkUUPAJJNDwCTRQ8AlF0PAJVFDwCWTQ8Al0UPAJh9DwCZRQ8Amk0PAJuZGgCcWQ8AnVkPAJ5JDwCfSQ8AoLkPAKG5DwCiyQ8Ao8kPAKS1CwClvQsAprULAKfVDwCo7Q8AqfUPAKr9DwCr9Q8ArO0PAK3VDwCu0Q8Ar9EPALCxDwCxsQ8AsrEPALOxDwC0cQ8AtXEPALZpDwC3aQ8AuAEPALkBDwC6GQ8AuxkPALzxAQC98QEAvvEBAL/xAQD6BQCA/QUAgAAGAIAgBgCA4QQAgIAFAIDTBQCADgYAgDQGAIBLBgCAaAYAgH8GAICWBgCA3QMAgPYDAIAPBACAEgcAgEQIAIBBCACAPwcAgD8kAIB4JACAqSQAgM4kAIC/JgCAyiYAgM4mAIDSJgCA1iYAgDUoAIB0KACAnCgAgKAoAIDFKACAzSgAgOkoAID7KACA/ygAgAMpAIAbKQCANikAgPAUNgBRKQCAHysAgEMkAIBQJACAXSQAgGokAIB8JACAiSQAgJskAICtJACAvSQAgNIkAIDcJACA6iQAgPQkAIABJQCAEiUAgBwlAIB1JQCAfCUAgColAICGJQCAgBEDAIERAwCCEQMAgxEDAIQxAwCFMQMAhjEDAIcxAwCIEQMAiREDAIoRAwCLEQMAjHEDAI1xAwCOcQMAj3EDAJARAwCREQMAkgEEAJMVAwCUDQMAlVUGAJZdBgCXVQYAmG0GAJl1BgCafQYAm3UGAJxtBgCdNQYAnj0GAJ81BgCgzQYAodUGAKLdBgCj1QYApPEDAKXxAwCm8QMAp/EDAKjRAwCp+QYAqikGAKspBgCsOQYArTkGAK7NAwCvxQMAsL0DALFFAwCyTQMAs0UDALRdAwC1RQMAtk0DALdFAwC4fQMAuUUDALpBAwC7fQYAvGUGAL1tBgC+ZQYAv1EDAKkVDwCoAQ8Aq00PAKpNDwCtRQ8ArEUPAK+pDQCusQ0AoXULAKBhCwCj6QsAok0LAKXhCwCk+QsApzkPAKZdCAC5mQ0AuJENALupDQC6kQ0AvbkNALyxDQA3JQCAvrENALHZDQCw0Q0As6kNALLRDQC1uQ0AtLENALepDQC2sQ0APiUAgE4lAIBhJQCAuCUAgMIlAICXJQCApyUAgNYlAICB5Q0AgOUNAIPlDQCC5Q0AheUNAITlDQCH4Q0Ahi0YAJlFDQCYuQ0Am0UNAJpFDQCdSQ0AnEUNAJ9xDQCefQ0AkYENAJC5DQCTgQ0AkokNAJWBDQCUmQ0Al4ENAJaJDQDmJACAJiUAgJMlAIDSJQCA5CUAgA4mAIAzJgCASCYAgPYlAIAAJgCAEiYAgB8mAIA3JgCAVCYAgF4mAIB8JgCAUCYAgGwmAIB0JgCAhiYAgJImAICaJgCAqSYAgOQmAICiJgCAuCYAgK0mAIDDJgCA2iYAgOgmAIAHJwCAFycAgCEnAIBVJwCAmCcAgO0nAIBVKQCAYykAgGcpAIBrKQCA9iYAgDQnAIBEJwCATicAgCknAIBZJwCAaScAgIQnAIB2JwCAnCcAgMgnAIDPJwCArCcAgNknAIDjJwCAuicAgB4oAIAQKACA8ScAgCsoAID4JwCAAigAgDkoAIBGKACAUCgAgFooAIBkKACAeCgAgIUoAICMKACApCgAgKsoAIC4KACA0SgAgNsoAIDtKACABykAgBQpAIAfKQCAKSkAgDopAIBBKQCAWSkAgMMDAIDmBACAhQUAgNgFAIATBgCAOQYAgFAGAIBtBgCAhAYAgJsGAIDjAwCA/AMAgBUEAIAoBACAOwQAgE4EAIBhBACAdAQAgIcEAICaBACAAAUAgA8FAIAeBQCALQUAgDwFAIBmCACAJwgAgMEGAID/BwCAIAkAgOM4EwA2CQCALQgAgDAIAIA0CACAJAcAgOkuAIDXMACA5i0AgMgwAIBSMQCAKgkAgO/kEwAJCQCA4g0AgNIIAICGCACAMQcAgEwHAID8BgCADQgAgJcIAIAtCQCADAkAgOYNAIDyDQCA3ggAgJwIAIAVBwCAiQgAgFUHAID/BgCAqQcAgJckAID2DQCA5QgAgCoIAICfCACAWwgAgBgJAID6DQCA6AgAgBcIAICiCACA6wgAgBoIAIDMCACApQgAgO8IAIAeCACAzwgAgKkIAID6CACAAAkAgIsHAICNCACAWQcAgAMHAIBACQCARAkAgEwJAIA5CQCAGwkAgP4NAID3CACAMAkAgA8JAIDqDQCA1QgAgJEIAIBgBwCAMwkAgBIJAIDuDQCA2AgAgJQIAIBjBwCAsAgAgGYHAIDjQBIA4/ATAOP0EwDj2BMA4wgNAOPoEgDjrBIA42QSAO/EDQDv2A0A7zQSAO9YEgDvQBIA79QSAO/IEgDvIBMA7AcAgMwGAIARCACAFAgAgNgGAIDUBgCAJAgAgAcHAIBqCACADAcAgHkIAIA0BwCANwcAgK0IAIC5CACAvAgAgOPQEADjsBAA46gQAON4EQDjTBAA4zAQAOPoEADj/BAA74QQAO+YEADvLBAA7yAQAO8EEADvCBAA73AQAO9MEADjmBMA4zAQAOMwEADjIBAA43wTAONAEwDjWBMA44ATAO/UEwDvqBMA74ATAO98EwDvRBMA7yQTAO8YEwDv7BAAgO08AIH1PACC/TwAg/U8AITtPACFFT0Ahh09AIcVPQCILT0AiTU9AIo9PQCLNT0AjC09AI0VPQCOHT0AjxU9AJBtPQCRdT0Akn09AJN1PQCUbT0AlRU9AJYdPQCXFT0AmC09AJk1PQCaPT0AmzU9AJwtPQCdFT0Anh09AJ8VPQCg7T0AofU9AKL9PQCj9T0ApO09AKUVPQCmHT0ApxU9AKgtPQCpNT0Aqj09AKs1PQCsLT0ArRU9AK4dPQCvFT0AsG09ALF1PQCyfT0As3U9ALRtPQC1FT0AthE9ALcRPQC4MT0AuTE9ALoxPQC7MT0AvBE9AL0RPQC+ET0AvxE9AIDxPACB/TwAgvU8AIMNPwCEFT8AhR0/AIYVPwCHDT8AiDU/AIk9PwCKNT8Aiw0/AIwVPwCNHT8AjhU/AI8NPwCQdT8AkX0/AJJ1PwCTDT8AlBU/AJUZPwCWCT8Alwk/AJg5PwCZOT8Amgk/AJsJPwCcGT8AnRk/AJ4JPwCfCT8AoPk/AKH5PwCiCT8Aowk/AKQZPwClGT8Apgk/AKcJPwCoOT8AqTk/AKoJPwCrCT8ArBk/AK0ZPwCuCT8Arwk/ALB5PwCxeT8Asgk/ALMJPwC0GT8AtRk/ALYJPwC3CT8AuDk/ALk5PwC6CT8Auwk/ALwZPwC9GT8Avgk/AL8JPwCA+TwAgfk8AIJJPQCDST0AhFk9AIVZPQCGST0Ah0k9AIh5PQCJeT0Aikk9AItJPQCMWT0AjVk9AI5JPQCPST0AkDk9AJE5PQCSAQQAk00GAJRVBgCVXQYAllUGAJdNBgCYdQYAmX0GAJp1BgCbTQYAnFUGAJ1dBgCeVQYAn00GAKC1BgChvQYAorUGAKPNBgCk1QYApd0GAKbVBgCnzQYAqPUGAKn9BgCq9QYAq80GAKzVBgCt3QYArtUGAK/NBgCwtQYAsb0GALK1BgCzTQYAtFUGALVdBgC2VQYAt00GALh1BgC5fQYAunUGALtNBgC8VQYAvV0GAL5VBgC/TQYArH0/AK2lPwCurT8Ar6U/AKh9PwCpZT8Aqm0/AKtlPwCkHT8ApUU/AKZNPwCnRT8AoB0/AKEFPwCiDT8AowU/ALydPwC9pT8Avq0/AL+lPwC4nT8AuYU/ALqNPwC7hT8AtN0/ALWlPwC2rT8At6U/ALDdPwCxxT8Ass0/ALPFPwCMYToAjWE6AI5hOgCPYToAiEE6AIlBOgCKQToAi0E6AIRhOgCFYToAhmE6AIdhOgCAAToAgQE6AIIBOgCDAToAnF04AJ3lPwCe7T8An+U/AJhdOACZRTgAmk04AJtFOACUuTgAlWU4AJZtOACXZTgAkAE6AJEBOgCSAToAkwE6AMMIAIDbCACA4QgAgPMIAIB5BwCAJQkAgHwHAICEBwCAVwkAgKAHAIDOBwCAwAcAgMQGAIDcBACAewUAgM4FAIAJBgCALwYAgEYGAIBjBgCAegYAgJEGAIDXAwCA8AMAgAkEAIAiBACANQQAgEgEAIBbBACAbgQAgIEEAICUBACA+gQAgAkFAIAYBQCAJwUAgDYFAIBFBQCATgUAgFcFAIBgBQCAaQUAgJUFAICeBQCAYAgAgFwOAIBfDgCASzoAgLgJAAC5CQAArwoAgBgLAIBHOgCATzoAgOYMAIBTOgCAHw0AgIc3AID+MACArzcAgGcyAIDLKgCAxiwAgPktAICaMDUAKi0AgPUtAIDkLwCA3zMAgJ80AwBvNQCAnSQpAJzxAQCd8QEAnvEBAJ/xAQCnNgCA4zYAgBc3AIArOACAgzEAgA8yAIC7MgCAUzMAgG82AIBXOACAgzkAgO8qAICaLACAlzEAgN8yAICjNgCA0zkAgKEuAICHMgCAkzYAgCc3AIAYMACAyzUAgO82AIAULwCAEjEAgCcyAIArMwCANzgAgDYrAIDOKwCAOiwAgIAwAICPMQCA2zIAgP8zAICbNgCAszYAgNc3AID/OACAszkAgM85AIA7NACArYwCAHs0AIAzNQCAUzYAgIs4AIBbNwCAqRUBAK4tAIAwLwCAsAAdALEEHgCyABgAswwZALToGgC1AOQAthjlALcc5wC4AOAAuSThALog4gC7AOwAvDTtAL007gC+OO8AvwDoAOs0AICrNQCAvwgAgA8zAICkUAMApQAMAKZYDgCnAAgAqGAKAKmQCwCqaBUAq2wXAKwAEACtdBEArnwSAK8AHABDNACApzcAgPc4AICqLQCAfS4AgIcxAIA7MgCAbzIAgCM1AIBLNQCAtzgAgDYsAIC3NQCA2isAgNYrAICnNACANzUAgGs2AIC/OACAdzcAgBwwAIBnNwCA1yoAgFEuAICILwCAPzMAgL8zAIBaLACASzQAgEYrAIBsLwCAtyoAgIDlAwCB7QMAgi0vAIPhAwCE4QMAheEDAIbhAwCH4S8AiN0vAInZAwCKbS8AiykCAIw5AgCNOQIAjikCAI8lAgCQcQIAkXECAJJxAgCTcQIAlBECAJURAgCWEQIAlxECAJgxAgCZMQIAmjECAJsxAgCcEQIAnRECAJ4RAgCfEQIAoPECAKHxAgCi8QIAo/ECAKQRAgClEQIAphECAKcRAgCoMQIAqTECAKoxAgCrMQIArBECAK0RAgCuEQIArxECALBxAgCxcQIAsnECALNxAgC0TSIAtRUCALYdAgC3FQIAuC0CALk1AgC6PQIAuzUCALwRAgC9EQIAvhECAL8RAgCVuQ4AlLEOAJfJDgCWsQ4AkbkOAJCxDgCTqQ4AkrEOAJ31DgCcZQ0An/UOAJ71DgCZ+Q4AmPEOAJvpDgCa8Q4AhQUOAIQFDgCHyQ4AhgEOAIGhDQCAlSAAg6UNAIKlDQCN+Q4AjPEOAI/JDgCO8Q4AifkOAIjxDgCL6Q4AivEOALWxAQC0qQEAt7EBALa5AQCxvSAAsLUBALOxAQCyuQEAvfEBALzpAQC/8QEAvvkBALnxAQC4iQEAu/EBALr5AQClNQ4ApDUOAKc1DgCmNQ4AoTUOAKA1DgCjNQ4AojUOAK31AQCs9QEAr/UBAK71AQCp9QEAqPUBAKv1AQCq9QEA+zEAgJgwAIAfNQCAriwAgJotAIALNACAczYAgEs3AIDHMQCA8zEAgCwwAIArNgCATDAAgLszAIALKwCAjisAgNIrAIBjMQCACzUAgAM2AIBXNwCAazgAgEIsAID2LACAJC8AgLQwAICLMgCATzQAgKc4AICLOQCA3zkAgPc5AID2MACAszEAgPs3AIDwLgCAzC8AgOgvAIB4MACAezIAgMcyAIB3MwCAmzQAgD81AICjNQCA6zcAgHs2AIATOACAjzgAgPYrAIAiLACACi0AgLcyAIADNwCAEC8AgIAvAIBEMACAvzEAgOc0AIAzMwCAGysAgGYtAIC1LgCAjC8AgIBdAwCB8SgAgmkDAINpAwCEeQMAhXkDAIZpAwCHaQMAiFkDAIlZAwCK1SoAi60DAIy1AwCNvQMAjrUDAI+tAwCQ1QMAkd0DAJLVAwCT7QMAlPUDAJX9AwCW9QMAl+0DAJjVAwCZ3QMAmtUDAJutAwCctQMAnb0DAJ61AwCfrQMAoFUDAKGZAwCiUQMAo1EDAKRxAwClcQMApnEDAKdxAwCoUQMAqVEDAKp1DACrVQMArE0DAK21AQCuvQEAr7UBALDNAQCx1QEAst0BALPVAQC0zQEAtfUBALb9AQC39QEAuM0BALnVAQC63QEAu9UBALzNAQC9tQEAvr0BAL+9DwBPMwCAazMAgHs1AICbNQCAczgAgPM4AIADOQCAPzkAgDorAICPNACAXzgAgNs4AICkLwCA9yoAgF4rAIBVLgCAdS4AgKQwAIDTMgCA2zMAgIc2AIAnOACA5jAAgLM4AIAaLACAMjEAgD4xAIAfMgCAVzIAgFszAIC3MwCANzQAgBs1AIBLOQCA+C8AgMM4AIBOKwCAmS4AgD8yAIDvNwCAXC8AgKwvAIBGMQCAyzgAgP4rAIDmLACAhS4AgM8wAIAiMQCAbzEAgAMyAIBXMwCAyzMAgGc1AIAHNwCAEzcAgOc4AIBqLACAWzIAgOosAIDXMgCAezMAgJc2AIDPOACAl50HAJadBwCVnQcAlJ0HAJOdBwCSnQcAke06AJCZBwCfmQcAnpEHAJ2ZBwCcgQcAm9UKAJqdBwCZnQcAmJ0HAIchBwCGGQcAhREHAIS1JACDBQcAggUHAIEVBwCAFQcAj+EHAI4ZBwCNEQcAjBkHAIsBBwCKGQcAiREHAIgZBwC3/SkAtpUBALWFAQC0hQEAs5UBALKVAQCxZQcAsGUHAL+RAQC+iQEAvYEBALyJAQC7kQEAuqkBALmhAQC4qQEApxkHAKYRBwClGQcApAEHAKMZBwCiEQcAoRkHAKBhBwCvFQcArhUHAK0FBwCsBQcAqxUHAKoVBwCpuSQAqCEHALs5AIDjOQCAOjEAgDcyAIDTNQCA0zQAgPc0AIAnMwCArzIAgHM3AIATKwCAOzYAgAIsAIDyKwCAAC8AgCAwAIADNQCAQS4AgBMyAIDyMACA9zcAgLs4AIAcLwCAbisAgEItAICWLQCA4jAAgN4rAIAvMwCA8zMAgFc0AIBzNACAdzQAgIs0AIALOQCA+zQAgJ82AIBjNwCAFzgAgEM4AIBfOQCAYzkAgGc5AIDLOQCAOzgAgNc4AIA+KwCAYisAgHYrAIAyLACAPiwAgH4sAIAyLQCATi0AgFYtAICSLQCAni0AgIEuAICYLwCAwC8AgMgvAICR5BgA4C8AgIwwAICANQMAgT0DAII1AwCDTQMAhFUDAIVdAwCGVQMAh00DAIjJKwCJcQMAitErAIt1AwCMbQMAjVUDALwwAIDqMACAkCUDAGcxAICSIQMAKzIAgEcyAICVOQMAlikDAJcpAwCYGQMAmRkDAJrpAwCb6QMAnPkDAJ35AwCe4SsAdzIAgKARAwDLMgCAoh0DAOsyAIBfMwCApQ0DAKYFAwA/NACAYzQAgF80AICqCQMAqwkDAKwZAwCtGQMArgkDAK8JAwCweQMAsXkDALIJAwCzCQMAtBkDALUZAwC2CQMAtwkDALg5AwC5OQMAugkDALsJAwC85S0AvR0DAL4VAwC/DQMAvY0fALxhAgC/nR8Avp0fALmRHwC4xQIAu5EfALqZHwC1VR8AtFUfALdVHwC2VR8AsVUfALBVHwCzVR8AslUfAK0dHwCsHR8AZzQAgGs0AICpHR8AqB0fAKsNHwCqDR8ApSEfAKRZHwCn8QIApikfAKFBHwCgWR8Ao0EfAKJJHwCdpR8AnKUfAJ+hHwCeqR8AmYUfAJiFHwCbhR8AmoUfAJWpHwCUoR8AlwUGAJahHwCRqTgAkPkAAJO5HwCSARwAjWEBAIzNOACPgQAAjmkBAIldAQCIUQEAi0UBAIpNAQCFrQEAhKEBAIeVAQCGvQEAgQ0CAIANAgCDxQEAgsUBAIc0AICrNACAvzQAgNs0AIBHNQCATzUAgGM1AICLNQCA2zUAgA82AIB3NgCAHzcAgDc3AIBrNwCAbzcAgLM3AIC3NwCADzgAgOs4AIAvOQCARzkAgJAvAICm6gCA8zUAgL8qAIDKKwCAiisAgDIrAIByKwCAnisAgC4sAIBKLACAHi0AgC4tAIBKLQCApi0AgPEtAID9LQCAGS4AgCkuAIAYLwCAIC8AgFAvAIBwLwCAoC8AgLgvAICoLwCAvC8AgPwvAIBUMACAYDAAgGgwAICQMACAFjEAgCoxAIBrMgCAYzIAgJMyAIAjNACA7zIAgCMzAIBvMwCAizMAgK8zAICAmQEAgZkBAIKpAQCDqQEAhLkBAIW5AQCGqQEAh6kBAIiZAQCJ1RwAipUBAIvVHACM8QEAjfEBAI7xAQCP8QEAkJEBAJEtHACS3RQAk5kBAJSJAQCVhTIAlnkYAJcxGgCYvQEAmYUBAJoVHwCbiQEAnPUfAJ2dAQCelQEAn40BAKDxHAChcQEAonEBAKNxAQCkkQMApZEDAKbtHACnlQMAqK0DAKm1AwCqvQMAq7UDAKytAwCtuQEArpkDAK+ZAwCwbRgAse0DALLVAQCz4QMAtOEDALXhAwC24QMAt+EDALjRAQC5pQMAun0cALupAwC8xQEAvSEXAL6xAwC/xQEA0zMAgNczAID3MwCABzQAgBs0AIAXNACARzQAgMM0AIDzNACAKzUAgFs1AIA/NgCAZzYAgNs2AIAjNwCALzcAgE83AIBTNwCAXzcAgHs3AIDzNwCAIzgAgFs4AIB7OACAxzgAgB85AIA7OQCAmzkAgD3qAIA46gCAauoAgOcpAIAPKgCAEyoAgOzqAIAZ6wCAkesAgCc6AIA3OgCASggAgFUIAIBYCACATQgAgFEIAIBaCQCA9w4AgOgOAIDtDgCA/A4AgPIOAIBRDwCA0A8AgIcPAIA1DwCAYA8AgG0PAIB1DwCAow8AgMgPAIC+DwCAww8AgLAPAIC3DwCABA8AgIBNAQCBRQMAglkBAINZAQCESQEAhUkBAIZ5AQCHVQMAiKkeAIlBAQCKZQMAi0UBAIxhAwCNWQEAjsU7AI9NAQCQNQEAkT0BAJI1AQCTzQEAlNUBAJXdAQCW1QEAl80BAJj1AQCZ/QEACQ8AgA4PAIAbDwCAKA8AgDAPAIA4DwCAQg8AgEcPAIBMDwCAVg8AgFsPAIBjDwCAcA8AgHgPAIB9DwCAgg8AgIoPAICPDwCAmQ8AgJ4PAICmDwCAgzQAgKsPAIDLDwCAPQ8AgCAPAIBoDwCAlA8AgBMPAIDjFgCA7BYAgO8WAID1FgCA6RYAgPIWAIDmFgCAGRcAgBwXAID7FgCAnc0GAPgWAICfwQYA/hYAgAEXAIAKFwCABxcAgJSZBgCVmQYAlukGAJfpBgANFwCABBcAgBMXAICTiQYAEBcAgB8XAIAlFwCAKxcAgCgXAIAuFwCAMRcAgDoXAICEzQYAhdUGAIbZBgA0FwCAgO0GAIHVBgCC3QYAg9UGALwZBwBdFwCAvhUHACIXAIC4GQcAuRkHALoJBwC7CQcAtN0HALUlBwC2LQcARhcAgLDdBwCxxQcAss0HALPFBwCsNQYArT0GAK41BgCvpQcAqDkGAKl9QgCqNQYAqy0GAKQ5BgClOQYApgkGAKcJBgCgIQYAoQFCAKI5QwCjKQYAgKEGAIGhBgBDFwCAg6UGAIS9BgBOFwCAhqkGAIepBgCImQYAieUGAIrtBgCL5QYAjP0GAI3lBgCO7QYAj+UGAJCdBgCRmQYAkqkGAJOtBgCUsQYAlbUGAJa9BgCXuQYAmIUGAJmBBgCagQYAm4UGAJyZBgCdnQYAnpUGAJ+RBgCgbQYAoWkGAKJ5BgCjfQYApGEGAKVlBgCmbQYAp2kGAKhVBgCpUQYAqlEGAKtVBgCsSQYArU0GAK5FBgCvQQYAsD0GALE5BgCyyQEAs80BALTRAQC11QEAttEBALfVAQC46QEAue0BALr5AQC7/QEAvOEBAL3lAQC+7QEAv+kBAIEVAgCAEQIAgxECAIIVAgCFDQIAhAkCAIcpAgCGLQIAiRUCAIgRAgCLEQIAihUCAI1xAgCMdQIAj30CAI55AgCRBQIAkAECAJMBAgCSBQIAlRkCAJQdAgCXFQIAlhECAJktAgCYKQIAmzkCAJo9AgCdIQIAnCUCAJ8tAgCeKQIAodkCAKDdAgCj0QIAotUCAKVpUwKkbVMCp8UCAKbBAgCp/QIAqPkCAKvFAgCqwQIArd0CAKzZAgCvPQIArjkCALEpUwKwLVMCVBcAgEAXAIBRFwCAWhcAgH8WAIDnDwCANxAAgBQQAIAoEACAIxAAgC0QAIAyEACAGRAAgFcXAICEnA4A7A8AgPEPAIAFEACAHhAAgF4QAIBjEACAbxAAgIUQAICUEACAmRAAgKQQAIC+EACA0RAAgNmIUAL1EACAJxEAgCwRAIA0EQCAQxEAgFIRAIBXEQCAXxEAgIIRAICpEQCAtREAgNURAIDaEQCA3xEAgBkSAIAsEgCAOBIAgFASAIDKEgCAIBMAgDkTAIA+EwCAURMAgGITAIB0EwCAeRMAgKATAICoEwCAvRMAgOQTAIDpEwCAQxQAgEgUAIBNFACAWRQAgGUUAIBqFACAchQAgH4UAICYFACAnRQAgNlcUAKlFACAqhQAgK8UAIC0FACAuRQAgL4UAIDRFACAn8kOAJ7NDgCdwV0CnBkNAJsFDQCaHQ0AmRENAJixDACXjQwAlqkMAJWlDACUoQwAk70MANYUAIDyFACADBUAgCYVAIAyFQCAShUAgE8VAIBcFQCAfRUAgKAVAIC6FQCAxhUAgMsVAIDTFQCA9BUAgA4WAIAdFgCAOhYAgD8WAIC/fQ4AvnkOAL11DgC8cQ4Au2kOALptDgC5YQ4AuGkOALdVDgC2UQ4AtVkOALRdDgCzXQ4AslkOALFRDgCwVQ4AryUOAK4hDgCtKQ4ArC0OAKsNDgCqCQ4AqQEOAKgFDgCnNQ4ApjEOAKU9DgCkOQ4AoyEOAKIlDgChNQ4AoDEOAIAFDgCBDQ4AggUOAIP1DwCEAQ4AhQEOAIYBDgCHAQ4AiAEOAIkBDgCKAQ4AiwEOAIwBDgCNAQ4AjgUOAI99DgCQBQ4AkQ0OAJIFDgCTHQ4AlAUOAJUNDgCWBQ4Alz0OAJgFDgCZDQ4AmgUOAJsdDgCcBQ4AnQ0OAJ4FDgCf/Q4AoAUOAKENDgCiBQ4Aox0OAKQFDgClDQ4ApgUOAKc9DgCoBQ4AqQ0OAKoFDgCrHQ4ArAUOAK0NDgCuBQ4Ar30OALAFDgCxDQ4AsgUOALMdDgC0BQ4AtQ0OALYFDgC3OQ4AuAkOALkJDgC6GQ4AuxkOALwJDgC9CQ4Avs0BAL/FAQCAPQIAgUUCAIJNAgCDRQIAhF0CAIVFAgCGTQIAh0UCAIh9AgCJRQIAik0CAItFAgCMXQIAjUUCAI5NAgCPRQIAkD0CAJFBAQCSQQEAk0EBAJRBAQCVQQEAlkEBAJdBAQCYQQEAmUEBAJpBAQCbQQEAnEEBAJ1BAQCeQQEAn0EBAKDBAQChwQEAosEBAKPBAQCkwQEApcEBAKaVDQCnxQEAqFkMAKm1DQCq9QEAq80BAKyRDQCt0QEArp0NAK+VDQCwqQEAsakBALL1DQCzvQEAtJENALWRDQC2rQEAt6UBALitDQC5mQEAurkNALu5DQC8OQ0AvTkNAL4hDQC/IQ0Ap1X8AEcWAIBMFgCAXxYAgGQWAICKFgCAlhYAgKIWAICxFgCAzhYAgNMWAID0EQCABRIAgIIWAICBAACAiwAAgJUAAICfAACAqQAAgLMAAID7DwCAABAAgAoQAIB7EACAgBAAgIoQAIDrEACA8BAAgB0RAIA5EQCAPhEAgEgRAIBXFQCAExYAgBgWAIAwFgCApxYAgKwWAIDEFgCA9g8AgA8QAICPEACAIhEAgN0SAIBFFQCANRYAgGkWAIDJFgCATREAgGoSAIClEgCAuBIAgBcUAIAjFACALxQAgJMTAICYEwCA1xMAgNwTAIADFACACBQAgG8SAIB0EgCAvRIAgIL5CwCD+QsAgO0LAIH5CwCGWQQAh1kEAIQtBACFWQQAiqUHAIutBwCIqQcAiXEEAI5JBACPSQQAjE0EAI2xBwCS1QcAk2UHAJB9BwCR3QcAlnkHAJdRCwCUwQcAlXkHAJptCwCbxQcAmGELAJnxBwCebQsAn1ULAJxtCwCdZQsAorELAKOxCwCgLQcAoaELAKbdCwCnzQsApKULAKU1BwCqxQsAq80LAKj1CwCpzQsArskLAK/JCwCsyQsArckLALJtBwCzTQsAsEkLALFJCwC2RQsAt00LALRVCwC1TQsAukkLALtJCwC4dQsAuUkLAL5JCwC/SQsAvEkLAL1JCwCAwQoAgcEKAILZCgCD2QoAhPkKAIX5CgCG6QoAh+kKAIjZCgCJHQUAihUFAIttBQCMdQUAjYUGAI5pBQCPaQUAkBkFAJEZBQCSIQUAkyEFAJQhBQCVIQUAlu0GAJdZBgCYaQYAmd0GAJp9BgCbdQYAnG0GAJ1VBgCexQYAn3EKAKAhBgChpQoAoi0GAKOxCgCkOQYApdkKAKZZBgCnHQoAqGUGAKltBgCqZQYAq1kKAKxJCgCt8QUArs0FAK8JBgCw4QYAsXkGALIZBgCzGQYAtAkGALUJBgC2OQYAtzkGALgJBgC5CQYAuhkGALsZBgC8CQYAvQkGAL75AwC/+QMAwhIAgMgRAIC8TQEAvUkBALqxCQC7sQkAuDUBALktAQC2XQkAtw0BALRdCQC1VQkAsv0FALOVCQCw8QUAsfkFAK5tAQCvdQEArAkBAK1lAQCqaQEAq2kBAKiRBQCpaQEApk0BAKdVAQCkTQEApUUBAKJtAQCjVQEAoG0BAKFlAQCejQEAn5UBAJyNAQCdhQEAmpEAAJuRAACYYQUAmWEFAJZRBQCXUQUAlEEFAJVBBQCSUQUAk1EFAJD5AQCRYQUAjvkBAI/5AQCMAQUAjfkBAIr9AQCL5QEAiP0BAIn1AQCG/QEAh8UBAIT9AQCF9QEAgv0BAIPlAQCA/QEAgfUBAPBQ/wDNEQCAnBEAgKERAIDkEQCA6REAgCwTAIAxEwCAZxMAgGwTAIB8EgCAiBIAgJsSAICgEgCASxIAgOISAIBdEwCAURAAgKkQAIDDEACAyhAAgNYQAID6EACAAREAgAgRAICHEQCAwREAgLoRAIAxEgCAHhIAgCUSAIBcEgCAVRIAgGMSAIDPEgCAJRMAgI0SAICBEgCAqhIAgLESAIBDEwCAVhMAgH4TAICFEwCAjBMAgK0TAIDCEwCAyRMAgO4TAID8EwCA9RMAgFIUAICDFACAihQAgBEVAIAfFQCAGBUAgPcUAIArFQCANxUAgIIVAICJFQCAmRUAgGEVAIBvFQCApRUAgKwVAIBoFQCAURYAgFgWAID5FQCAABYAgN8VAIDmFQCAKRYAgCIWAIC2FgCAdBAAgLcQAICwEACAlRn/AJQR/wCXKf8AlhH/AJEd/wCQHf8Akwn/AJIR/wCdFf8AnBX/AJ8V/wCeFf8AmRX/AJgR/wCbFf8AmhX/AKUJ/wCkDf8Apxn/AKYB/wChEf8AoOn/AKMd/wCiGf8ArT3/AKw5/wCvDf8Arg3/AKkl/wCoJf8AqyH/AKol/wC1df8AtHX/ALdx/wC2df8AsXn/ALBx/wCzdf8AsnX/AL0t/wC8Kf8Avz3/AL49/wC5Mf8AuEn/ALsx/wC6Of8AgNn+AIHZ/gCC6f4Ag+n+AIT1/gCF/f4AhvH+AIfx/gCIzf4AidX+AIrd/gCL1f4AjM3+AI01AQCOPQEAjzUBAOQQAIDdEACAkkUBAJNdAQCURQEAlU0BAJZFAQCXfQEAmEEBAJlBAQCaQQEAm0EBAJxBAQCdRQEAnk0BAJ9FAQCgvQEAocUAAKLNAACjjQMApJUDAKWdAwCmlQMAp40DAKi5AwCpuQMAqokDAKuJAwCsmQMArZkDAK6JAwCviQMAsPkDALH5AwCyiQMAs4kDALQB/gC1Df4AtpEDALeRAwC4sQMAubEDALqxAwC7sQMAvKkDAL2lAwC+mQMAv5kDABYRAIAPEQCAlREAgGQRAICOEQCAkQQHAD0SAIDWEgCAlBIAgEoTAIAFFQCAPhUAgJsWAICPFgCAvRYAgL8VAICRFACABxYAgNATAIDKFACA2BUAgLMVAID+FACAwxQAgGsRAICuEQCAdhUAgF4UAIBoEACARBIAgO0VAIAZEwCAdxQAgEgQAIA/EACAkBUAgOcSAID8EQCAtBMAgHEWAIDwEgCA9xIAgHIRAIAKEgCApgMAgBMjAIAXIwCAoAYAgMcAAIC1BgCAsSMAgLUjAIC/IQCAuyEAgOYHAIB+CQCAggkAgEcjAICtIwCAOyMAgD8jAIAjIwCAJyMAgCsjAICAaQEAgWkBAIJ5AQCDUQEAhKX8AIWZ/ACGbQEAh2UBAC8jAIAzIwCANyMAgN4HAIDiBwCA0QAAgNcAAICiAwCAqAMAgOAHAIDTAACA1QAAgL0GAIB5AACADRQAgH0AAICHAACAkQAAgBIUAICbAACAHhQAgKUAAIAqFACArwAAgDYUAIC5AACAOxQAgNUPAIBbEACAnhAAgKEQAIAxEQCAXBEAgKYRAIDSEQCA7hEAgPERAID5EQCAExIAgBYSAICwvQEAsUUBALJNAQCzRQEAtF0BALVFAQC2TQEAt0UBALh9AQC5RQEAuk0BALtFAQC8XQEAeRIAgMcSAIA2EwCAqAYAgLMGAIDMlLQAPBAAgHETAICdEwCApRMAgOETAIBAFACAbxQAgKIUAIDbFACAVBUAgNAVAIBEFgCAbhYAgJiNBgCZgbEAhxYAgN4UAIDjFACA6BQAgO0UAIDPAACAkNEGAJHRBgCS0QYAk9EGAJSpkgKVtQYAlr0GAJe1BgDZAACAswMAgOQHAICACQCAASMAgAUjAICHKQCAOyQAgHQkAICTJACApSQAgMokAIDJKACA5SgAgPcoAICOJgCAuCEGALkhBgC6IQYAuyEGALwhBgC9IQYAviEGAL95uACwIbEAsTUGALI9BgCzNQYAtCkGALVFsgC2TbIAtyEGAIC5uQCB+QcAgikGAIMpBgCEOQYAiykAgG8pAICHMQYAiBEGAIn5sACK9bAAi/GwAIztsACN7QcAjuEHAI/lBwCQ8QcAkfEHAJL1sAAvJACAlImTApXpBwCWnQcAl50HAJixBwCZ1bMAmt2zAJuxBwCckQcAnZEHAJ6RBwCfSQYAoLkGAKG5BgCiIbMAo80GAKSRAQClkQEApkGyADMkAICo5QYAqe0GAKrlBgCr/QYAhAkAgIcJAICQCQCAjQkAgLCVBgCxnQYAspUGAIoJAIC0sQYA8iEAgLa9BgC3tQYAuI0GALmVBgC6nQYAu5UGALyNBgC9dQYAvn0GAL91BgCCkaMCg5GjAoCFBQCBnaMChrmjAoeNowKEjaMChbWjAoqVowKLkaMCiLGjAomZowKOPQIAj6UFAIyNowKNMQIAktEFAJPRBQCQ2QUAkd0FAJbJBQCXzQUAlM0FAJXFBQCa/QUAm/kFAJjxBQCZ8QUAntEFAJ/VBQCc5QUAnd0FAKIlBQCjIQUAoCkFAKEpBQCmOQUApz0FAKQ9BQClNQUAqg0FAKsVBQCoAQUAqQEFAK4FBQCvDQUArAkFAK0JBQCyfQUAs3kFALBxBQCxcQUAtiUFALchBQC0ZQUAtSkFALoZBQC7HQUAuB0FALkVBQC+DQUAvwkFALwBBQC9AQUAhPwCAPUiAID6IQCAAiIAgPkiAID9IgCAkZQCAKQQDQCffBwAnmQIAJ08CADZzKACzMCFAsz8hQL2IQCA9egAgJmY1gD66ACAm/jWAP/oAICxNN0ABOkAgNmUoAKynPsApwUAAAnpAIDZhKAChywDAJE4AgAO6QCAjAwCAI20AgDwtAIAgqACABPpAIC++QAAkpACAJcdAAC5+AMAhAgCAPBAAgCRfAIAHekAgBjpAICljNYAIukAgI3IAgDwJAIAsGkAAJK4AgC4tAMAudgMAJa0AgCW/AIArz0AAJFUAgCvCCEAJ+kAgLpFAACRTAIALOkAgL1BAACWuAIArw0AAIsZDACKHQwAiREMAIgZDACPDQwAjgkMAI0FDACMBQwAgzEMAII1DACBOQwAgNELAIclDACGIQwAhS0MAIQtDACbaQwAmm0MAJlhDACYaQwAn30MAJ55DACddQwAnHUMAJNBDACSRQwAkUkMAJBxDACXVQwAllEMAJVdDACUXQwAq7kMAKq9DACpsQwAqLkMAK9J/gCuqQwAraUMAKylDACjkQwAopUMAKGZDACggQwAp4UMAKaBDACljQwApI0MALuZDAC6kQwAuZ0MALidDAC/2YgC8HwNALUcAgC8hQwAs6kMALKtDACxoQwAsDX+ALehDAC2uQwAtbUMALS1DACBwQsA8DgBAIM9CgCCUQ0AhV0KAIRdCgCHmQ0AhiUKAImVvwCIlb8Ai4G/AIoRCgCNcQoAjIULAI+9DQCOgbwAkbWdApCxvACTqZ0CkpGdApWtvACUrbwAl9W/AJbVvwCZxb8AmMW/AJuxnwKa0QsAnam+AJx1DQCfvQsAnnkNAKEBvwCg1QoAo5WxAKKFvQClgb0ApJm9AKeBvQCmqbEAqYm9AKiFvQCrmb0Aqom9AK3xvQCsjb0Ar+m9AK71vQCxIb8AsJ29ALNJyQCyrb0AtaG9ALS9vQC3ob0AtkmxALlpyQC4TbEAu8UKALrNsQC9wQoAvLEKAL8hCwC+dQ0AgP2+AIHVngKCZb4Ag8W+AISRvgCFnb4AhqW+AIeNvgCIrZECieW+AIopkgKLtb4AjBGSAo2VvgCOLbIAj8WeApDpvgCRsbUAkkGSApPxnwKU1b4AleW+AJbhvgCXTZICmGWSApl9kgKaub4Am7EIAJz9DgCdlQgAkRQDAJ/tDgCgFQ4AoT0IAKJ1CACjrQkApCUIAKUNDgCmBdYApwEOAKgBDgCpAQ4AqgEOAKsBDgCsAQ4ArQEOAK4BDgCvAQ4AsNUPALGV0wCyndMAs4XUALSJ0wC1idMAttnWALfZ1gC46dYAuenWALr51gC7+dYAvOnWAL3p1gC+udYAv7nWAIBJ1wCBSdcAglnXAINZ1wCESdcAhXXSAIZ90gCHddIAiE3SAIlV0gCKddcAi63XAIy11wCNvdcAjrXXAI9J0gCQOdIAkTnSAJLJ0gCTydIAlNnSAJXZ0gCWydIAl8nSAJj50gCZ+dIAmsnSAJvF0gCc4dcAnW0OAJ41DgCf4Q4AoNHZAKHB2wCiwdsAo93bAKTF2wClzdsApsXbAKf92wCoJdsAqWXbAKrN2wCrxdsArN3bAK3B2wCuwdsAr3nbALAJ2wCxCdsAshnbALMZ2wC0CdsAtbXYALbB3gC33d8AuOXfALn13wC6/d8Au63fALy13wC9pd8Avq3fAKQgBQCPFd4AjhXeAI0F3gCMBd4AixXeAIoV3gD+IQCAsCUAAIfd3ACG3dwAhd3cAITd3ACD2dwAgtXcADHpAIA26QCAP+kAgEjpAIBV6QCAnAXeAJtt3gCabd4AYukAgG/pAICXVd4AllXeAJVB3gCUSd4Ak1HeAJJp3gB86QCAiekAgKzpAICukQsArZkLAKyBCwCriQAAqp0LAKmhCwComQsAkukAgKZxCQClZQkApBEJAKNlmAKiDQkAoQ0JAKANCQC7vQAAtekAgL3BjAKf6QCAvx0AAL4RAAC9eQAAvF0AAIAFAADC6QCA0AoAgLMMAIBkDQCAag0AgHANAIB8DQCAhWgCAH8NAICH6AMAhiADAJ4tAQCfVQEAgg0AgIUNAICIDQCAlw0AgJ0NAICgDQCAow0AgCYiAIDNDQCA3A0AgJVAHQCURB4Al0gbAJYAGACRLAAAkFQBAJNYHwCSABwAnWASAJxkEwCfiBcAnmwRAJlwGQCYdBoAmwAQAJoAFAACDgCABQ4AgBQOAIAXDgCAIw4AgB4iAIA4DgCAOw4AgLgALAC5rC8AuqguAN0WAIAWFwCA4BYAgLoDAIC3AwCAygMAgO0EAICMBQCA3wUAgBoGAIBABgCAVwYAgHQGAICiuQEAo7kBAKCtAQChpQEAiwYAgDgBAICkgQEAPAEAgICBuACBDboAghW2AIMBugCEAboAhSG2AIYBugCHPboAiAm6AIkBugCKGboAixW6AIxxugCNfboAjmm6AI9lugCQobgAkSW6AJLJzgCTJboAlCG6AJXBtgCWIboAl/W2AJjpzgCZRbYAmrmaApsBugCcuboAnfW6AJ7xugCfwboAoBG6AKEJlQKiSboAo42WAqQJugCltZYCpjm6AKeJtgCoWZoCqQ26AKpdsQCrpZYCrA2bAq0xugCuCboArwW6ALDRlgKxwZYCstGWArMdugC0UbgAtd26ALbFtgC30boAuPG6ALnRtgC68boAu826ALzZugC90boAvsm6AL/FugCfCZcCnvGwAJ2huwCc9ZsCmwW3AJq1uwCZOZcCmIW7AJchlwKW5bsAzMS+AJS9uwCTjbsAkr27AJG5uwCQ9bsAjy27AI6VmwKNabcAjMXPAIv5twCKLbsAic23AIgtuwCHCbsAhuXPAIUJuwCEjbkAgym7AIIluwCBMbsAgD27AL8ptwC+/bsAvR23ALz9uwC7+bsAuhXPALn5uwC4fbkAt/m7ALb1uwC14bsAtO27ALOJuwCyhbsAsZ27ALCVuwCv4bsArt27AK39twCs3bsAq927AKrJtwCp0bsAqF25AKcxuwCm/ZcCpe2XAqT9lwKjSbsAokW7AKF9uwCgQZoCgJmkAIEliAKCqaQAgxmoAEABAICFvaQAhu2vAIcViAKInYUCiaGkAIqZpACLlaQAjCGIAo0xiAKOIYgCj+2kAJDBpgCRTaQAklWoAJNBpACUQaQAlWGoAJZBpACXfaQAmEmkAJlBpACaWaQAm1WkAJwxpACdPaQAnimkAJ8lpACgYaYAoeWkAKIJ0ACj5aQApOGkAKUBqACm4aQApzWoAKgp0ACphagAqnmEAqvBpACseaQArTWkAK4xpACvAaQAsFGkALFJiwKyCaQAs82IArRJpAC19YgCtnmkALfJqAC4GYQCuU2kALodrwC75YgCvE2FAr1xpAC+SaQAv0WkAIARiQKBAYkCghGJAoPdpQCEkacAhR2lAEQBAICHEaUAiDGlAIkRqQCKMaUAkQ0AgEgBAICNEaUAjgmlAI8FpQCQAaUAkQ2lAJIZpQCTFaUAlLGnAEwBAICW2dEAlzWlAJgRpQCZ8akAmhGlAJvFqQCc+dEAUAEAgJ6phQKfEaUAoEmlAKEFpQCiAaUAozGlAKQBpQClGYoCplmlAKediQKoOaUAqYWJAqoJpQCruakArEmFAq0dpQCuTa4Ar7WJArB9hAKxQaUAsnmlALN1pQC0wYkCtdGJArbBiQK3DaUAuGGnALntpQBUAQCAu+GlALzhpQC9wakAvuGlAFgBAICAKaYAgSGmAII5pgCDNaYAhFGmAFwBAICGSaYAzEyHAmABAIBkAQCAiqnSAItFpgCMQaYAjaGqAI5BpgCPlaoAkMnSAGgBAICSmYYCkyGmAJSZpgCV1aYAltGmAJfhpgCY8aYAmemJApqppgCbbYoCnOmmAJ1VigKe2aYAn2mqAKB5hgKhLaYAon2tAKOFigKkLYcCpRGmAKYppgCnJaYAqLGKAqmhigKqsYoCq32mAKwxpACtvaYArqWqAK+xpgCw0aYAsfGqALLRpgCz7aYAtPmmALXxpgC26aYAt+WmALihpgC5raYAurmmALu1pgC8EaQAvZWmAL550gC/laYAl7mnAJa1pwCVjacAlLGGApMZiwKS4awAkbGnAJDlhwKfLacAnjmrAGwBAICcraUAm+GnAJotiwKZPYsCmC2LAof9pwCGzacAhcmnAISFpwCDPacAgoWHAoF5qwCA1dMAj3WrAI7FpwCNSYsCjPWnAItxiwKKtacAifWIAojtpwC37acAtlWHArWpqwC0BdMAszmrALLtpwCxDasAsO2nAL+hiwK+ZacAvSWIAnABAIC7DacAdAEAgLk5pwC4dacAeAEAgKb1pwCl7acAfAEAgIABAICizacAhAEAgIgBAICviacArmXTAIwBAICsDaUAq6mnAKqlpwCpsacAkAEAgICRoACBiY8CgsmgAIMNjAKEiaAAhTWMAoa5oACHCawAiNmAAomNoACK3asAiyWMAoyNgQKNsaAAjomgAI+FoACQUYwCkUGMApJRjAKTnaAAlNGiAJVdoACWRawAl1GgAJhxoACZUawAmnGgAJtNoACcWaAAnVGgAJ5JoACfRaAAoMGgAKHNoACi2aAAo9WgAKRxogCl9aAAphnUAKf1oACo0aAAqTGsAKrRoACrBawArDnUAK2VrACuaYACr9GgALAJoACxRaAAskGgALNxoAC0QaAAtVmPArYZoAC33YwCuHmgALnFjAK6SaAAu/msALwJgAK9XaAAvg2rAL/1jAKAvYACgYGhAIK5oQCDtaEAhAGNAoURjQKGAY0Ch82hAIihowCJLaEAijWtAIshoQCMIaEAjQGtAI4hoQCPHaEAkGmhAJFhoQCSeaEAk3WhAJQRoQCVHaEAlgmhAJcFoQCYgaMAmQWhAJrp1QCbBaEAnAGhAJ3hrQCeAaEAn9WtAKAJ1QChpa0AolmBAqPhoQCkWaEApRWhAKYRoQCnIaEAqDGhAKkpjgKqaaEAq62NAqwpoQCtlY0CrhmhAK+prQCwOYECsW2hALI9qgCzxY0CtG2AArVRoQC2aaEAt2WhALjxjQK54Y0CuvGNArs9oQC8caMAvf2hAL7lrQC/8aEAgBGiAIExrgCCEaIAgy2iAIQ5ogCFMaIAhimiAIclogCIYaIAiW2iAIp5ogCLdaIAjNGgAI1VogCOudYAj1WiAJAxogCR0a4AkjGiAJPlrgCU2dYAlXWuAJaJggKXMaIAmKmiAJnlogCa4aIAm9GiAJzhogCd+Y0CnrmiAJ99jgKgGaIAoaWOAqIpogCjma4ApGmCAqU9ogCmbakAp5WOAqgdgwKpIaIAqhmiAKsVogCsoY4CrbGOAq6hjgKvbaIAsEGgALHNogCy1a4As8GiALTBogC14a4AtsGiALf9ogC4yaIAucGiALrZogC71aIAvLGiAL29ogC+qaIAv6WiAJMVrwCSpaMAkSmPApCVowCXGY8CluGoAJWxowCU5YMCm5mjAJqVowCZraMAmJGCAp/howCeLY8CnT2PApwtjwKD6a8Agj2jAIHdrwCAPaMAhz2jAIaFgwKFea8AhNXXAIvdowCK7aMAiemjAIilowCPcY8CjrWjAI31jAKM7aMAs+mjALIF1wCx6aMAsG2hALc5rwC27aMAtQ2vALTtowC7zaMAunWDArmJrwC4JdcAvw2jAL49owC9OaMAvHWjAKPNowCi2a8AocGjAKBNoQCn8aMAps2jAKXtrwCkzaMAq9mjAKrVowCpzaMAqMWjAK+powCupaMArbGjAKy9owC43YwCueWMArrxrQC78a0AvJmuAL2ZrgC+ua4Av7muALDZrQCx2a0AsqGuALOhrgC0ka4AtZGuALbNrQC3yYwCqMmuAKnJrgCq6a4Aq+muAKylrQCtoYwCroWMAq+9jAKgwa4AocGuAKKdrQCjmYwCpK2MAqWVjAKmga0Ap4GtAJh1rQCZcYwCmlWMApttjAKcaa0AnWmtAJ4RrgCfEa4AkH2MApFFjAKSUa0Ak1GtAJQ5rgCVOa4AlhmuAJcZrgCIwbwCiTm5AopRFQCLURUAYQ0AgJQBAICOLa0AjymMAoDdqwCBdRUAgqkBAIN5FQCE+bwChQW5Aob5ugKHCbkCbQ0AgJgBAIBzDQCAnAEAgHkNAIBLIwCAiw0AgNEGAIC+DQCAywcAgNANAIAPBwCACA4AgJcHAIAaDgCAnQcAgCYOAICGYAIAhYABAIRMOQCABwCAyAcAgE8HAIBSBwCAXQcAgJANAADhBgCAFSQAgOglAIA1LgCAiXg/AIgAPAC6LACA1i0AgD83AIAHKwCA0zAAgL8yAIAOLACAYC8AgKYrAICsMACA+isAgCc1AICbNwCAui0AgPIsAIBzMgCAEDAAgDwwAIAbOACAMDAAgAgwAIB/NACAuzQAgN4sAICvoAEAUzIAgKczAIASLACAPi0AgFM4AICPIwCAUyMAgMwouALNRLgCzIi4As38uALMkIQCzTSFAsywhQLNVIcCzBCCAs1QggLMoIICzYyCAswwgQLNJIECzBiBAs2EgQJdIwCAcSMAgJkjAIB7IwCAoyMAgGcjAICFIwCAZC8AgKo5AwCrOQMAziwAgNsqAIDMWLkCzaS6AqwZAwDTKgCAsHWvALFxjgKyVY4Cs22OArRprwC1aa8AthGsALcRrAC4AawAuQGsAOMqAIDP6QCALisAgEIrAIBKKwCAUisAgIJB2gCDra4AgCmsAIGtrgCGqa4Ah32iAISprgCFSaIAis0DAIs5xACIYdoAic2iAI6hAwCPoQMAjM0DAI3BAwCfua8AnrWvAJ2NrwCcsY4CmxmDAprhpACZsa8AmOWPApc1owCWha8AlQmDApS1rwCTMYMCkvWvAJG1gAKQra8Aj/2vAI7NrwCNya8AjIWvAIs9rwCKhY8CiXmjAIjV2wCHyaMAhh2vAIX9owCEHa8AgxmvAIL12wCBGa8AgJ2tAL+xFgC+qRYAvaEWALzBvgK7tRYAuom5ArmBuQK4sQoAt22sALa9AgC1iRYAtLEWALOpFgCysRYAsUUXALCtAgCv2bkCrs0CAK0xFwCszQIAqyUXAKodrACpKRcAqA0DAMyIgQLNSIECpQUXAKQFFwCjIa8Aou2DAqH9gwKg7YMCgAm4AoE9EQCCIQUAg/29AoT1qACFFb4ChiURAIfxiAKIoREAiaERAIoZBQCL0b0CjDG4Ao09vgKOsREAj7ERAJB5BQCRsb0CknWvAJPdEQCUEQUAlcERAJZRuAKXrb0CmGG+ApmRvQKaabgCm5G9ApyhBACdhRAAnrGrAJ+JEACggQUAoX0QAKKBBQCjlb4CpIEFAKVpEACmnREAp4URAKi9EQCphREAqrEFAKthqwCsnQ0Ara2+Aq6lvgKvmREAsI25ArHtEQCy5REAs/0RALT5pAC14REAtvkFALcxvQK4ya8AucmvALrhuAK71REAvNkFAL0FvQK+HagAv/2+AoA9EACB6YkCgokQAIOJEACEIQQAhem8AoYZuQKHFb8CiKkQAImpEACKEQQAi9m8AowNrgCNpRAAjnkEAI+pEACQSbkCkbW8ApJJvwKTubwClFG5ApWpvAKWiQUAl60RAJipqgCZkREAmmkEAJuVEQCceQQAnW2/Ap5pBACfgREAoIUQAKGdEACilRAAo60QAKSJBAClWaoAprUMAKeFvwKovb8CqYEQAKrluAKrhRAArJ0QAK2FEACukaUAr4kQALDhBACxKbwCsuGuALPhrgC02bkCte0QALbxBAC3LbwCuAWpALnlvwK61RAAuwGJArxxEAC9cRAAvskEAL8BvAKAAboCgQ28AoKBEwCDgRMAhCkHAIXhvwKGJa0Ah40TAIhhBwCJsRMAiiG6AovdvwKMMbwCjcG/Ao45ugKPwb8CkJEGAJG1EgCSgakAk7kSAJRRBwCVrRIAllEHAJdFvAKYcQcAmZkSAJptEwCbdRMAnG0TAJ1VEwCeYQcAn7GpAKCtDwChnbwCopW8AqOpEwCk3bsCpb0TAKa1EwCnrRMAqImmAKmREwCqiQcAq0G/AqyZrQCtma0ArrG6Aq+FEwCw6QcAsTW/ArItqgCzzbwCtO0TALU5igK2WRMAt1kTALjRBwC5Gb8Cuum6ArvlvAK8eRMAvXkTAL7BBwC/Cb8CngG9Ap/xvgKcAbsCnf2+AppRBgCbgRIAmCWsAJmNEgCWGQYAl9G+ApShEgCVoRIAkjG7ApM9vQKQCQYAkcG+Ao7BEgCPwRIAjHUSAI2hiwKKtasAi1W9AohxBgCJrb4Chmm7AoddEgCEQawAhUGsAIJRBgCDmb4CgFGnAIFJEgC+qawAv6msALypBgC9Yb4CurmnALuhEgC4tRIAua0SALbtugK3jRIAtLW9ArWJEgCynQ4As629ArChBgCxcagArt0SAK/lEgCszRIArdUSAKrBBgCrKRMAqNEGAKnFvQKm4QYApx0TAKQhqAClGRMAoiEHAKMFEwCg+bsCoQG+AoKhIwBWKwCAWisAgJMpAIDj6QCAh7EjAHorAIB+KwCAmisAgIsJJADU6QCAiWUkAI6NIwCPLSQAlykAgI0JJACSZSQAhisAgN7pAICRtSMAtisAgJf9IwCUrSMAvisAgBcrAICbeSQAxisAgJmRIwC56wCAn8EtAO4rAICdKdQAoiEjAJ8pAIAGLACAoR0jAAosAICnMSMApKEkABYsAICqiSQAoykAgKi5JACp5SQArg0jAK+tJACsiSQArYkkALLlJABOLACAsOkkALE1IwC2TSMAt30jALQtIwC1RSMAuvUkALv5JABSLACAuREjAL5BLQB6LACAvFUtAIYsAICADSUAgZUiAIKpIgCDoSIAhCklAIUpJQCGoSIAh7kiAIgxJQCJbSUAliwAgIsBJQCMASUAjQElAI6FIgCPJSUAkGElAJG9IgCSbSUAk/kiAJSlIgCVzSIAlsUiAJf1IgCY0SIAmZkiAJp9JQCbcSUAniwAgKIsAIDy6QCAozIAgLYsAIChFSIAoikiAKMhIgC+LACApaklAKYhIgCnOSIAqLElAKntJQD36QCAq4ElAKyBJQCtgSUArgUiAK+lJQCw4SUAsT0iALLtJQCzeSIAtCUiALVNIgC2RSIAt3UiALhRIgC5GSIAuv0lALvxJQDKLACA0iwAgNosAIACLQCAgLkvAIG5LwCCyS8Ag8kvAITZLwCF2S8AhskvAIfJLwCI+S8AifkvAIrZLwDuLACA+iwAgP4sAIAGLQCADi0AgJC1LwCRuS8AklkyAJNVMwCUYTMAlQEzAJYtMwCXtTMAmJ0zAJlxMACaSTAAm0kwAJw1MACdXTEAntUxAJ/JMQCgQTEAoUkxAKJZMQCjVTEApE0xAKV9MQCmZTEAp0k6AKilOwCpqTsAqr07AKuxmgCs0ZYArak7AK6dOwASLQCAsEGUALHNlgCy1ZoAs8GWALTBlgC14ZoAtsGWALf9lgC4yZYAucGWALrZlgC71ZYAvLGWAL29lgC+qZYAv6WWAMUAAACpTScAqiEnAKshJwCsIScArSEnAK6lIACvBScAGi0AgKG1IACiiSAAIi0AgKQJJwAmLQCANi0AgKeZIAA6LQCAubkgAEYtAIC7UScAai0AgG4tAIBSLQCAWi0AgLBBJwCxnSAAsk0nALYtAIC0hSAAte0gALblIAC31SAAiJEnAInNJwCKoScAi6EnAIyhJwCNoScAjiUgAI+FJwCArScAgTUgAIIJIACDASAAfi0AgIWJJwCGASAAhxkgAJhxIACZOSAAmt0nAJvRJwCcfS4AnYHXAJ5pLgCfaS4AkMEnAJEdIACSzScAk1kgAJQFIACVbSAAlmUgAJdVIACA+T0Agfk9AIJJPgCDST4AhFk+AIVZPgCGST4Ah0k+AIh5PgCCLQCAhi0AgHotAICOLQCAii0AgKItAID86QCAkB0+AJEtPgCoSBUAvi0AgMItAIDKLQCA3i0AgAEuAICiAAwAo2QPAKBoAwChbAAApgAUAKd0FwCkAAgApXwLAAHqAIAG6gCADS4AgBEuAIAVLgCACS4AgB0uAICnKQCAqykAgCUuAIAtLgCAC+oAgEkuAIBNLgCAWS4AgBDqAIBhLgCAZS4AgEQvAICvKQCAeS4AgJUuAICRLgCAGuoAgJ0uAIAf6gCAqS4AgKUuAICtLgCAvS4AgMEuAICzKQCAgG0/AIF5PwCCoT8Ag6E/AIShPwCFrcgAhq0/AIelPwCInT8AiSEFAIqVyACLJQUAjD0FAI0lBQCO+cgAjyUFAJBdBQCRZQUAkmEFAJN1BQCUFQUAlfU8AJYRBQCXDQUAmD0FAJkFBQCaAQUAmwnVAJwBBQCdgeUAngEFAJ/5BQCgCQUAoRUFAKJlPACjEQUApDUFAKXt1QCmXcgAp1XIAKjd1QCpYQUAqkEFAKtBwwCsRQUArXnIAK5FBQCvlQQAsLEEALGxBACyvQQAs7kEALSpBAC1qQQAtlkEALdNBAC4SQQAuRUEALodBAC7FQQAvA0EAL3ZBwC+yQcAv8kHAIA5BACBOQQAgrUEAIMtBQCEPQUAhSEFAIYtBQCHmdYAiBkFAIllBQCKYQUAi30FAIzlywCNYQUAjmEFAI9hBQCQdcsAkSkFAJL5BQCTaQIAlHkCAJV5AgCWaQIAl2kCAJhZAgCZWQIAmiUCAJs9AgCcJQIAnfE/AJ4hAgCfIQIAoAECAKEBAgCiAQIAowECAKQBAgClAQIApgECAKcBAgCoAQIAqQECAKoBAgCrBQIArB0CAK0FAgCuDQIAr2HAALB5AgCxeQIAsgkCALMJAgC0GQIAtdnlALYVAgC3DQIAuPXlALkxAgC6MQIAuzECALwRAgC9EQIAvhECAL8RAgCfpT4AnqU+AJ2VPgCclT4Am1XMAJqBPgCZiT4AmJE+AJdF9ACWrT4Ala0+AJTh5wCTvekAkrU+AJFNPgCQReQA0S4AgNkuAIDdLgCA4S4AgLcpAIAk6gCAuykAgAQvAIAILwCADC8AgOvrAIAu6gCA5zUAgIJh+wCBCT4AgBE+ADwvAIC/KQCAUeoAgCPrAIC7ZT4AumU+ALl1PgC4dT4At2HkALah8AC1TT4AtHE+ALNpPgCyYT4AsWk+ALCl7wCvDT4AronwAK3h9ACshfAAqxk+AKrJ9ACpbecAqBk+AKchPgCmWT4ApVE+AKRZPgCjQT4Aos3iAKFVPgCgVT4ATC8AgFQvAIDDKQCAaC8AgHgvAIB8LwCAhC8AgJQvAIDLKQCAxykAgDPqAICcLwCAsC8AgLQvAIDELwCA2C8AgNAvAIDULwCA3C8AgPAvAID0LwCADDAAgBQwAIAkMACAWAAAADgwAIBC6gCANDAAgCgwAIBAMACASDAAgFwwAIBH6gCAZDAAgFgwAIBQMACAzykAgGwwAIB0MACAfDAAgHAwAIDTKQCAlDAAgEzqAIDAMACAAjEAgN4wAIDfKQCA2ykAgNcpAICqKwCArisAgAYxAIDuMACAuzUAgEMqAIAaMQCALjEAgCYxAIBl6gCA4ykAgEIxAIA2MQCAXzEAgICJAQCBiQEAgpkBAIOZAQCEiQEAhYkBAIa5AQCHuQEAiN3EAImNAQCKhQEAi50BAIyFAQCNjQEAjoUBAI8BxwCQgQEAkYEBAJKBAQCTgQEAlIEBAJWBAQCWgQEAl4EBAJj1zACZsdkAmokBAJuJAQCcmQEAnZkBAJ6JAQCfiQEAoHkBAKF5AQCizccAo4kCAKSZAgClmQIApokCAKfh2wCotQIAqb0CAKq1AgCrjQIArJUCAK2dAgCulQIAr40CALD1AgCx/QIAsvUCALONAgC0lQIAtfXAALaRAgC3kQIAuLECALmxAgC6sQIAu7ECALyRAgC9kQIAvpECAL+RAgCYjQ0AmQUCAJoNAgCbBQIAnB0CAJ0FAgCeDQIAnwUCAJBt6QCRSQ0AkkUNAJNdDQCUsQ0AlbUNAJa9DQCXtQ0AiBENAIkRDQCKEQ0AixENAIwxDQCNMQ0AjsnPAI81DQCAkQ4AgZEOAIKRDgCDkQ4AhDENAIUxDQCGMQ0AhzENALjpAgC56QIAuvkCALv5AgC86QIAvekCAL4ZAgC/GQIAsFnHALGpzwCy5QIAs/0CALTlAgC17QIAtuUCALfZAgCoddoAqfECAKrxAgCrPccArO0CAK2VAgCunQIAr5UCAKD9AgChxQIAos0CAKMFxwCk2QIApdkCAKbJAgCnyQIAgAAAAG/qAIBrMQCASjEAgHMxAIB3MQCAezEAgH8xAICLMQCAdOoAgJMxAIDrKQCAnzEAgHnqAICjMQCA7ykAgK8xAIC7MQCAyzEAgH7qAIAV6gCAg+oAgOsxAICI6gCA9zEAgP8xAIDvMQCACzIAgBsyAIAjMgCALzIAgDMyAICN6gCAFzIAgEsyAIBPMgCA8ykAgF8yAICS6gCAQzIAgH8yAICX6gCAnOoAgIMyAICXMgCAjzIAgPcpAICbMgCAqzIAgKcyAICzMgCA2ekAgMMyAICh6gCAzzIAgKvqAIDjMgCAAzMAgLDqAIAXMwCAGzMAgLXqAIC66gCANzMAgEczAID7KQCASzMAgP8pAIBjMwCAZzMAgHMzAIB/MwCAAyoAgJczAIC/6gCAszMAgMTqAIAp6gCAzzMAgMnqAIDO6gCA0+oAgAcqAIALKgCA3eoAgNjqAIDi6gCA5+oAgA80AIATNACAHzQAgBcqAIAbKgCA8eoAgF4AAAAzNACAHyoAgPbqAID76gCAAOsAgKM0AIAjKgCArzQAgLM0AIAF6wCACusAgMs0AIAnKgCAD+sAgN80AIDjNACAKyoAgBTrAID/NACALyoAgA81AIAHNQCAFzUAgB7rAIAvNQCAMyoAgDs1AIBDNQCAUzUAgDcqAIAo6wCALesAgDsqAICADd8AgVUBAIJdAQCDVQEAhE0BAIV1AQCGfQEAh3kBAIgx3wCJod8AikEBAItBAQCMQQEAjUEBAI5FAQCP5cgAkFnfAJHFAQCSzQEAkwHCAJTZAQCV2QEAlt3AAJfNAQCY9QEAmWHdAJrxAQCb8QEAnNEBAJ3RAQCe3QEAn9UBAKAtAQChNQEAoj0BAKM1AQCkLQEApVUBAKZdAQCnVQEAqG0BAKl1AQCqfQEAq3UBAKxtAQCtVQEArl0BAK9VAQCwLQEAsTUBALI9AQCzNQEAtC0BALXRAgC20QIAt9ECALjxAgC58QIAuvXdALv1AgC87QIAvdUCAL7dAgC/1QIAntkFAJ/ZBQCc2QUAndkFAJrZBQCb2QUAmNkFAJnZBQCWmQUAl5kFAJTN3gCVmQUAkp0FAJOFBQCQnQUAkZUFAI7dBQCP5QUAjN0FAI3VBQCKwQUAi7XeAIjRBQCJ0QUAhuEFAIfhBQCEEQUAhREFAIIdxQCDAQUAgAkFAIExwAC+yQIAv8kCALzJAgC9yQIAuqkCALupAgC4ccgAuakCALaNAgC3lQIAtI0CALWFAgCyrQIAs5UCALBN3gCxpQIArtECAK/RAgCsxQIArcECAKrVAgCr3QIAqCUFAKndAgCmFQUApx0FAKQFBQClHQUAohUFAKMdBQCgGQUAoRHeAIAAAAAy6wCAazUAgDfrAIB3NQCAgzUAgDzrAIBB6wCAnzUAgEbrAICnNQCAVuoAgD8qAIC/NQCAwzUAgEcqAIDHNQCAIS4AgEvrAIBQ6wCAW+oAgGDqAIDrNQCAAzgAgEsqAIAXNgCAEzYAgBs2AIAmLACAHzYAgCM2AIAnNgCALzYAgFXrAIAzNgCARzYAgEs2AIA3NgCATzYAgGM2AIBDNgCAVzYAgFs2AIBfNgCAWusAgGTrAIBf6wCATyoAgH82AICDNgCAizYAgHjrAICPNgCAaesAgFMqAIBXKgCAbusAgHPrAIBbKgCArzYAgLc2AIC7NgCAxzYAgMM2AIDPNgCAyzYAgNM2AIDXNgCA3zYAgF8qAIDnNgCA6zYAgGMqAID7NgCAfesAgAs3AIAPNwCAZyoAgBs3AICbKQCAgusAgIfrAIBrKgCAbyoAgEc3AICM6wCAnzcAgKM3AIC7NwCAxzcAgJbrAIDo6QCAXQAAANM3AIDPNwCA2zcAgO3pAIDnNwCAm+sAgKDrAIAzOACAPzgAgEc4AICl6wCASzgAgHc4AICDOACAhzgAgH84AICTOACAlzgAgKrrAICjOACAcyoAgKs4AICv6wCAdyoAgOM4AICxLgCA+zgAgLTrAIC+6wCAeyoAgH8qAIAjOQCAw+sAgIMqAIDI6wCAgBkBAIEZAQCCKQEAgykBAIT99ACFPQEAhjUBAIctAQCIFQEAiR0BAIoVAQCLbQEAjHUBAI19AQCOdQEAj20BAJAJwwCRCcMAkkH5AJMZAQCUCQEAlQkBAJY5AQCXOQEAmAkBAJkJAQCaHQEAmxUBAJwNAQCd9QEAnv0BAJ8twwCgCQEAoQkBAKIZAQCjGQEApAkBAKUJAQCmOQEApzkBAKgJAQCpCQEAqhkBAKsZAQCsCQEArQkBAK55AQCveQEAsAkBALEJAQCyGQEAsxkBALQJAQC1CQEAtjkBALc5AQC4CQEAuQkBALoZAQC7GQEAvAkBAL0JAQC+kcMAv5XDAJyVHQCdnR0AnpUdAJ8lwgCYPdsAmZ0dAJqVHQCbjR0AlFkdAJVZHQCWqR0Al6kdAJBZHQCRWR0AkkkdAJNJHQCMGR0AjRkdAI4pHQCPKR0AiB0dAIkFHQCKDR0Aiy0VAIRdHQCFJR0Ahi0dAIclHQCAXR0AgUUdAIJNHQCDRR0AvIkCAL2JAgC+mQIAv5kCALhtHQC5lQIAup0CALupwAC0ZdcAtVUdALZdHQC3VR0AsFEdALFRHQCyUR0As1EdAKwRHQCtER0ArhEdAK8RHQCoER0AqREdAKoRHQCrER0ApFEdAKVRHQCmUR0Ap1EdAKBRHQChUR0AolEdAKNRHQCAeQAAgXkAAIKJAACDiQAAhJkAAIWZAACGiQAAh4kAAIi5AACJuQAAikXBAIuNAACMlQAAjZ0AAI6VAACPjQAAkPUAAJH9AACS9QAAk40AAJSVAACVnfsAlpEAAJeF+wCYrQAAmbUAAJq9AACbtQAAnJ37AJ2pAABDOQCAzesAgFs5AICHKgCAazkAgHc5AIB/OQCAhzkAgIsqAIDS6wCAtzkAgMM5AICPKgCAkyoAgMc5AIDX6wCAlyoAgNzrAIDh6wCA5usAgJsqAIAHOgCACzoAgBM6AIAbOgCA8OsAgLkAAAC4AAAAuwAAALoAAAC9AAAAvAAAAL8AAAC+AAAAACAAIMyBACDMgwAgzIQAIMyFACDMhgAgzIcAIMyIACDMiMyAACDMiMyBACDMiM2CACDMigAgzIsAIMyTACDMk8yAACDMk8yBACDMk82CACDMlAAgzJTMgAAgzJTMgQAgzJTNggAgzKcAIMyoACDMswAgzYIAIM2FACDZiwAg2YwAINmM2ZEAINmNACDZjdmRACDZjgAg2Y7ZkQAg2Y8AINmP2ZEAINmQACDZkNmRACDZkQAg2ZHZsAAg2ZIAIOOCmQAg44KaACEAISEAIT8AIgAjACQAJQAmACcAKAAoMSkAKDEwKQAoMTEpACgxMikAKDEzKQAoMTQpACgxNSkAKDE2KQAoMTcpACgxOCkAKDE5KQAoMikAKDIwKQAoMykAKDQpACg1KQAoNikAKDcpACg4KQAoOSkAKEEpAChCKQAoQykAKEQpAChFKQAoRikAKEcpAChIKQAoSSkAKEopAChLKQAoTCkAKE0pAChOKQAoTykAKFApAChRKQAoUikAKFMpAChUKQAoVSkAKFYpAChXKQAoWCkAKFkpAChaKQAoYSkAKGIpAChjKQAoZCkAKGUpAChmKQAoZykAKGgpAChpKQAoaikAKGspAChsKQAobSkAKG4pAChvKQAocCkAKHEpAChyKQAocykAKHQpACh1KQAodikAKHcpACh4KQAoeSkAKHopACjhhIApACjhhIIpACjhhIMpACjhhIUpACjhhIYpACjhhIcpACjhhIkpACjhhIspACjhhIwpACjhhI4pACjhhI8pACjhhJApACjhhJEpACjhhJIpACjkuIApACjkuIMpACjkuIkpACjkuZ0pACjkuowpACjkupQpACjku6MpACjkvIEpACjkvJEpACjlhaspACjlha0pACjlirQpACjljYEpACjljZQpACjlkI0pACjlkbwpACjlm5spACjlnJ8pACjlraYpACjml6UpACjmnIgpACjmnIkpACjmnKgpACjmoKopACjmsLQpACjngaspACjnibkpACjnm6MpACjnpL4pACjnpZ0pACjnpa0pACjoh6opACjoh7MpACjosqEpACjos4cpACjph5EpACjqsIApACjrgpgpACjri6QpACjrnbwpACjrp4gpACjrsJQpACjsgqwpACjslYQpACjsmKTsoIQpACjsmKTtm4QpACjsnpApACjso7wpACjssKgpACjsubQpACjtg4ApACjtjIwpACjtlZgpACkAKgArACwALQAuAC4uAC4uLgAvADAAMCwAMC4AMOKBhDMAMOeCuQAxADEsADEuADEwADEwLgAxMOaXpQAxMOaciAAxMOeCuQAxMQAxMS4AMTHml6UAMTHmnIgAMTHngrkAMTIAMTIuADEy5pelADEy5pyIADEy54K5ADEzADEzLgAxM+aXpQAxM+eCuQAxNAAxNC4AMTTml6UAMTTngrkAMTUAMTUuADE15pelADE154K5ADE2ADE2LgAxNuaXpQAxNueCuQAxNwAxNy4AMTfml6UAMTfngrkAMTgAMTguADE45pelADE454K5ADE5ADE5LgAxOeaXpQAxOeeCuQAx4oGEADHigYQxMAAx4oGEMgAx4oGEMwAx4oGENAAx4oGENQAx4oGENgAx4oGENwAx4oGEOAAx4oGEOQAx5pelADHmnIgAMeeCuQAyADIsADIuADIwADIwLgAyMOaXpQAyMOeCuQAyMQAyMeaXpQAyMeeCuQAyMgAyMuaXpQAyMueCuQAyMwAyM+aXpQAyM+eCuQAyNAAyNOaXpQAyNOeCuQAyNQAyNeaXpQAyNgAyNuaXpQAyNwAyN+aXpQAyOAAyOOaXpQAyOQAyOeaXpQAy4oGEMwAy4oGENQAy5pelADLmnIgAMueCuQAzADMsADMuADMwADMw5pelADMxADMx5pelADMyADMzADM0ADM1ADM2ADM3ADM4ADM5ADPigYQ0ADPigYQ1ADPigYQ4ADPml6UAM+aciAAz54K5ADQANCwANC4ANDAANDEANDIANDMANDQANDUANDYANDcANDgANDkANOKBhDUANOaXpQA05pyIADTngrkANQA1LAA1LgA1MAA14oGENgA14oGEOAA15pelADXmnIgANeeCuQA2ADYsADYuADbml6UANuaciAA254K5ADcANywANy4AN+KBhDgAN+aXpQA35pyIADfngrkAOAA4LAA4LgA45pelADjmnIgAOOeCuQA5ADksADkuADnml6UAOeaciAA554K5ADoAOjo9ADsAPAA9AD09AD09PQA+AD8APyEAPz8AQABBAEFVAEHiiJVtAEIAQnEAQwBDRABDby4AQ+KIlWtnAEQAREoARFoARHoARMW9AETFvgBFAEYARkFYAEcAR0IAR0h6AEdQYQBHeQBIAEhQAEhWAEhnAEh6AEkASUkASUlJAElKAElVAElWAElYAEoASwBLQgBLSwBLTQBMAExKAExURABMagBMwrcATQBNQgBNQwBNRABNSHoATVBhAE1SAE1WAE1XAE3OqQBOAE5KAE5qAE5vAE8AUABQSABQUE0AUFBWAFBSAFBURQBQYQBRAFIAUnMAUwBTRABTTQBTUwBTdgBUAFRFTABUSHoAVE0AVQBWAFZJAFZJSQBWSUlJAFbiiJVtAFcAV0MAV1oAV2IAWABYSQBYSUkAWQBaAFsAXABdAF4AXwBgAGEAYS5tLgBhL2MAYS9zAGHKvgBiAGJhcgBjAGMvbwBjL3UAY2FsAGNjAGNkAGNtAGNtMgBjbTMAZABkQgBkYQBkbABkbQBkbTIAZG0zAGR6AGTFvgBlAGVWAGVyZwBmAGZmAGZmaQBmZmwAZmkAZmwAZm0AZwBnYWwAaABoUGEAaGEAaQBpaQBpaWkAaWoAaW4AaXYAaXgAagBrAGtBAGtIegBrUGEAa1YAa1cAa2NhbABrZwBrbABrbQBrbTIAa20zAGt0AGvOqQBsAGxqAGxtAGxuAGxvZwBseABswrcAbQBtMgBtMwBtQQBtVgBtVwBtYgBtZwBtaWwAbWwAbW0AbW0yAG1tMwBtb2wAbXMAbeKIlXMAbeKIlXMyAG4AbkEAbkYAblYAblcAbmoAbm0AbnMAbwBvVgBwAHAubS4AcEEAcEYAcFYAcFcAcGMAcHMAcQByAHJhZAByYWTiiJVzAHJhZOKIlXMyAHMAc3IAc3QAdAB1AHYAdmkAdmlpAHZpaWkAdwB4AHhpAHhpaQB5AHoAewB8AH0AwqIAwqMAwqUAwqYAwqwAwrBDAMKwRgDCtwDDgADDgQDDggDDgwDDhADDhQDDhgDDhwDDiADDiQDDigDDiwDDjADDjQDDjgDDjwDDkQDDkgDDkwDDlADDlQDDlgDDmQDDmgDDmwDDnADDnQDDoADDoQDDogDDowDDpADDpQDDpwDDqADDqQDDqgDDqwDDrADDrQDDrgDDrwDDsADDsQDDsgDDswDDtADDtQDDtgDDuQDDugDDuwDDvADDvQDDvwDEgADEgQDEggDEgwDEhADEhQDEhgDEhwDEiADEiQDEigDEiwDEjADEjQDEjgDEjwDEkgDEkwDElADElQDElgDElwDEmADEmQDEmgDEmwDEnADEnQDEngDEnwDEoADEoQDEogDEowDEpADEpQDEpgDEpwDEqADEqQDEqgDEqwDErADErQDErgDErwDEsADEsQDEtADEtQDEtgDEtwDEuQDEugDEuwDEvADEvQDEvgDFgwDFhADFhQDFhgDFhwDFiADFiwDFjADFjQDFjgDFjwDFkADFkQDFkwDFlADFlQDFlgDFlwDFmADFmQDFmgDFmwDFnADFnQDFngDFnwDFoADFoQDFogDFowDFpADFpQDFqADFqQDFqgDFqwDFrADFrQDFrgDFrwDFsADFsQDFsgDFswDFtADFtQDFtgDFtwDFuADFuQDFugDFuwDFvADFvQDFvgDGjgDGkADGoADGoQDGqwDGrwDGsADHjQDHjgDHjwDHkADHkQDHkgDHkwDHlADHlQDHlgDHlwDHmADHmQDHmgDHmwDHnADHngDHnwDHoADHoQDHogDHowDHpgDHpwDHqADHqQDHqgDHqwDHrADHrQDHrgDHrwDHsADHtADHtQDHuADHuQDHugDHuwDHvADHvQDHvgDHvwDIgADIgQDIggDIgwDIhADIhQDIhgDIhwDIiADIiQDIigDIiwDIjADIjQDIjgDIjwDIkADIkQDIkgDIkwDIlADIlQDIlgDIlwDImADImQDImgDImwDIngDInwDIogDIpgDIpwDIqADIqQDIqgDIqwDIrADIrQDIrgDIrwDIsADIsQDIsgDIswDItwDJkADJkQDJkgDJlADJlQDJmQDJmwDJnADJnwDJoQDJowDJpQDJpgDJqADJqQDJqgDJqwDJrQDJrwDJsADJsQDJsgDJswDJtADJtQDJuADJuQDJuwDKgQDKggDKgwDKiQDKigDKiwDKjADKjQDKkADKkQDKkgDKlQDKnQDKnwDKuQDKvG4AzIAAzIEAzIjMgQDMkwDOhgDOiADOiQDOigDOjADOjgDOjwDOkADOkQDOkgDOkwDOlADOlQDOlgDOlwDOmADOmQDOmgDOmwDOnADOnQDOngDOnwDOoADOoQDOowDOpADOpQDOpgDOpwDOqADOqQDOqgDOqwDOrADOrQDOrgDOrwDOsADOsQDOsgDOswDOtADOtQDOtgDOtwDOuADOuQDOugDOuwDOvADOvEEAzrxGAM68VgDOvFcAzrxnAM68bADOvG0AzrxzAM69AM6+AM6/AM+AAM+BAM+CAM+DAM+EAM+FAM+GAM+HAM+IAM+JAM+KAM+LAM+MAM+NAM+OAM+cAM+dANCAANCBANCDANCHANCMANCNANCOANCZANC5ANC9ANGKANGMANGQANGRANGTANGXANGcANGdANGeANG2ANG3ANOBANOCANOQANORANOSANOTANOWANOXANOaANObANOcANOdANOeANOfANOiANOjANOkANOlANOmANOnANOqANOrANOsANOtANOuANOvANOwANOxANOyANOzANO0ANO1ANO4ANO5ANWl1oIA1bTVpQDVtNWrANW01a0A1bTVtgDVvtW2ANeQANeQ1rcA15DWuADXkNa8ANeQ15wA15EA15HWvADXkda/ANeSANeS1rwA15MA15PWvADXlADXlNa8ANeV1rkA15XWvADXlta8ANeY1rwA15nWtADXmda8ANea1rwA15sA15vWvADXm9a/ANecANec1rwA150A157WvADXoNa8ANeh1rwA16IA16PWvADXpNa8ANek1r8A16bWvADXp9a8ANeoANeo1rwA16nWvADXqda814EA16nWvNeCANep14EA16nXggDXqgDXqta8ANey1rcA2KEA2KIA2KMA2KQA2KUA2KYA2KbYpwDYptisANim2K0A2KbYrgDYptixANim2LIA2KbZhQDYptmGANim2YcA2KbZiADYptmJANim2YoA2KbbhgDYptuHANim24gA2KbbkADYptuVANinANin2YPYqNixANin2YTZhNmHANin2YsA2KfZtADYqADYqNisANio2K0A2KjYrdmKANio2K4A2KjYrtmKANio2LEA2KjYsgDYqNmFANio2YYA2KjZhwDYqNmJANio2YoA2KkA2KoA2KrYrADYqtis2YUA2KrYrNmJANiq2KzZigDYqtitANiq2K3YrADYqtit2YUA2KrYrgDYqtiu2YUA2KrYrtmJANiq2K7ZigDYqtixANiq2LIA2KrZhQDYqtmF2KwA2KrZhditANiq2YXYrgDYqtmF2YkA2KrZhdmKANiq2YYA2KrZhwDYqtmJANiq2YoA2KsA2KvYrADYq9ixANir2LIA2KvZhQDYq9mGANir2YcA2KvZiQDYq9mKANisANis2K0A2KzYrdmJANis2K3ZigDYrNmEINis2YTYp9mE2YcA2KzZhQDYrNmF2K0A2KzZhdmJANis2YXZigDYrNmJANis2YoA2K0A2K3YrADYrdis2YoA2K3ZhQDYrdmF2YkA2K3ZhdmKANit2YkA2K3ZigDYrgDYrtisANiu2K0A2K7ZhQDYrtmJANiu2YoA2K8A2LAA2LDZsADYsQDYsdiz2YjZhADYsdmwANix24zYp9mEANiyANizANiz2KwA2LPYrNitANiz2KzZiQDYs9itANiz2K3YrADYs9iuANiz2K7ZiQDYs9iu2YoA2LPYsQDYs9mFANiz2YXYrADYs9mF2K0A2LPZhdmFANiz2YcA2LPZiQDYs9mKANi0ANi02KwA2LTYrNmKANi02K0A2LTYrdmFANi02K3ZigDYtNiuANi02LEA2LTZhQDYtNmF2K4A2LTZhdmFANi02YcA2LTZiQDYtNmKANi1ANi12K0A2LXYrditANi12K3ZigDYtdiuANi12LEA2LXZhNi52YUA2LXZhNmJANi12YTZiSDYp9mE2YTZhyDYudmE2YrZhyDZiNiz2YTZhQDYtdmE25IA2LXZhQDYtdmF2YUA2LXZiQDYtdmKANi2ANi22KwA2LbYrQDYttit2YkA2LbYrdmKANi22K4A2LbYrtmFANi22LEA2LbZhQDYttmJANi22YoA2LcA2LfYrQDYt9mFANi32YXYrQDYt9mF2YUA2LfZhdmKANi32YkA2LfZigDYuADYuNmFANi5ANi52KwA2LnYrNmFANi52YTZitmHANi52YUA2LnZhdmFANi52YXZiQDYudmF2YoA2LnZiQDYudmKANi6ANi62KwA2LrZhQDYutmF2YUA2LrZhdmJANi62YXZigDYutmJANi62YoA2YDZiwDZgNmOANmA2Y7ZkQDZgNmPANmA2Y/ZkQDZgNmQANmA2ZDZkQDZgNmRANmA2ZIA2YEA2YHYrADZgditANmB2K4A2YHYrtmFANmB2YUA2YHZhdmKANmB2YkA2YHZigDZggDZgtitANmC2YTbkgDZgtmFANmC2YXYrQDZgtmF2YUA2YLZhdmKANmC2YkA2YLZigDZgwDZg9inANmD2KwA2YPYrQDZg9iuANmD2YQA2YPZhQDZg9mF2YUA2YPZhdmKANmD2YkA2YPZigDZhADZhNiiANmE2KMA2YTYpQDZhNinANmE2KwA2YTYrNisANmE2KzZhQDZhNis2YoA2YTYrQDZhNit2YUA2YTYrdmJANmE2K3ZigDZhNiuANmE2K7ZhQDZhNmFANmE2YXYrQDZhNmF2YoA2YTZhwDZhNmJANmE2YoA2YUA2YXYpwDZhdisANmF2KzYrQDZhdis2K4A2YXYrNmFANmF2KzZigDZhditANmF2K3YrADZhdit2YUA2YXYrdmF2K8A2YXYrdmKANmF2K4A2YXYrtisANmF2K7ZhQDZhdiu2YoA2YXZhQDZhdmF2YoA2YXZiQDZhdmKANmGANmG2KwA2YbYrNitANmG2KzZhQDZhtis2YkA2YbYrNmKANmG2K0A2YbYrdmFANmG2K3ZiQDZhtit2YoA2YbYrgDZhtixANmG2LIA2YbZhQDZhtmF2YkA2YbZhdmKANmG2YYA2YbZhwDZhtmJANmG2YoA2YcA2YfYrADZh9mFANmH2YXYrADZh9mF2YUA2YfZiQDZh9mKANmH2bAA2YgA2YjYs9mE2YUA2YjZtADZiQDZidmwANmKANmK2KwA2YrYrNmKANmK2K0A2YrYrdmKANmK2K4A2YrYsQDZitiyANmK2YUA2YrZhdmFANmK2YXZigDZitmGANmK2YcA2YrZiQDZitmKANmK2bQA2a4A2a8A2bEA2bkA2boA2bsA2b4A2b8A2oAA2oMA2oQA2oYA2ocA2ogA2owA2o0A2o4A2pEA2pgA2qEA2qQA2qYA2qkA2q0A2q8A2rEA2rMA2roA2rsA2r4A24AA24EA24IA24UA24YA24cA24fZtADbiADbiQDbiwDbjADbkADbkgDbkwDgpJXgpLwA4KSW4KS8AOCkl+CkvADgpJzgpLwA4KSh4KS8AOCkouCkvADgpKkA4KSr4KS8AOCkr+CkvADgpLEA4KS0AOCmoeCmvADgpqLgprwA4Kav4Ka8AOCniwDgp4wA4KiW4Ki8AOCol+CovADgqJzgqLwA4Kir4Ki8AOCosuCovADgqLjgqLwA4Kyh4Ky8AOCsouCsvADgrYgA4K2LAOCtjADgrpQA4K+KAOCviwDgr4wA4LGIAOCzgADgs4cA4LOIAOCzigDgs4sA4LWKAOC1iwDgtYwA4LeaAOC3nADgt50A4LeeAOC5jeC4sgDguqvgupkA4Lqr4LqhAOC7jeC6sgDgvIsA4L2A4L61AOC9guC+twDgvYzgvrcA4L2R4L63AOC9luC+twDgvZvgvrcA4L2x4L2yAOC9seC9tADgvbHgvoAA4L6Q4L61AOC+kuC+twDgvpzgvrcA4L6h4L63AOC+puC+twDgvqvgvrcA4L6y4L2x4L6AAOC+suC+gADgvrPgvbHgvoAA4L6z4L6AAOGApgDhg5wA4YSAAOGEgQDhhIIA4YSDAOGEhADhhIUA4YSGAOGEhwDhhIgA4YSJAOGEigDhhIsA4YSMAOGEjQDhhI4A4YSPAOGEkADhhJEA4YSSAOGElADhhJUA4YSaAOGEnADhhJ0A4YSeAOGEoADhhKEA4YSiAOGEowDhhKcA4YSpAOGEqwDhhKwA4YStAOGErgDhhK8A4YSyAOGEtgDhhYAA4YWHAOGFjADhhZcA4YWYAOGFmQDhhaAA4YWhAOGFogDhhaMA4YWkAOGFpQDhhaYA4YWnAOGFqADhhakA4YWqAOGFqwDhhawA4YWtAOGFrgDhha8A4YWwAOGFsQDhhbIA4YWzAOGFtADhhbUA4YaEAOGGhQDhhogA4YaRAOGGkgDhhpQA4YaeAOGGoQDhhqoA4YasAOGGrQDhhrAA4YaxAOGGsgDhhrMA4Ya0AOGGtQDhh4cA4YeIAOGHjADhh44A4YeTAOGHlwDhh5kA4YedAOGHnwDhh7EA4YeyAOGshgDhrIgA4ayKAOGsjADhrI4A4aySAOGsuwDhrL0A4a2AAOGtgQDhrYMA4bSCAOG0lgDhtJcA4bScAOG0nQDhtKUA4bW7AOG2hQDhuIAA4biBAOG4ggDhuIMA4biEAOG4hQDhuIYA4biHAOG4iADhuIkA4biKAOG4iwDhuIwA4biNAOG4jgDhuI8A4biQAOG4kQDhuJIA4biTAOG4lADhuJUA4biWAOG4lwDhuJgA4biZAOG4mgDhuJsA4bicAOG4nQDhuJ4A4bifAOG4oADhuKEA4biiAOG4owDhuKQA4bilAOG4pgDhuKcA4bioAOG4qQDhuKoA4birAOG4rADhuK0A4biuAOG4rwDhuLAA4bixAOG4sgDhuLMA4bi0AOG4tQDhuLYA4bi3AOG4uADhuLkA4bi6AOG4uwDhuLwA4bi9AOG4vgDhuL8A4bmAAOG5gQDhuYIA4bmDAOG5hADhuYUA4bmGAOG5hwDhuYgA4bmJAOG5igDhuYsA4bmMAOG5jQDhuY4A4bmPAOG5kADhuZEA4bmSAOG5kwDhuZQA4bmVAOG5lgDhuZcA4bmYAOG5mQDhuZoA4bmbAOG5nADhuZ0A4bmeAOG5nwDhuaAA4bmhAOG5ogDhuaMA4bmkAOG5pQDhuaYA4bmnAOG5qADhuakA4bmqAOG5qwDhuawA4bmtAOG5rgDhua8A4bmwAOG5sQDhubIA4bmzAOG5tADhubUA4bm2AOG5twDhubgA4bm5AOG5ugDhubsA4bm8AOG5vQDhub4A4bm/AOG6gADhuoEA4bqCAOG6gwDhuoQA4bqFAOG6hgDhuocA4bqIAOG6iQDhuooA4bqLAOG6jADhuo0A4bqOAOG6jwDhupAA4bqRAOG6kgDhupMA4bqUAOG6lQDhupYA4bqXAOG6mADhupkA4bqgAOG6oQDhuqIA4bqjAOG6pADhuqUA4bqmAOG6pwDhuqgA4bqpAOG6qgDhuqsA4bqsAOG6rQDhuq4A4bqvAOG6sADhurEA4bqyAOG6swDhurQA4bq1AOG6tgDhurcA4bq4AOG6uQDhuroA4bq7AOG6vADhur0A4bq+AOG6vwDhu4AA4buBAOG7ggDhu4MA4buEAOG7hQDhu4YA4buHAOG7iADhu4kA4buKAOG7iwDhu4wA4buNAOG7jgDhu48A4buQAOG7kQDhu5IA4buTAOG7lADhu5UA4buWAOG7lwDhu5gA4buZAOG7mgDhu5sA4bucAOG7nQDhu54A4bufAOG7oADhu6EA4buiAOG7owDhu6QA4bulAOG7pgDhu6cA4buoAOG7qQDhu6oA4burAOG7rADhu60A4buuAOG7rwDhu7AA4buxAOG7sgDhu7MA4bu0AOG7tQDhu7YA4bu3AOG7uADhu7kA4byAAOG8gQDhvIIA4byDAOG8hADhvIUA4byGAOG8hwDhvIgA4byJAOG8igDhvIsA4byMAOG8jQDhvI4A4byPAOG8kADhvJEA4bySAOG8kwDhvJQA4byVAOG8mADhvJkA4byaAOG8mwDhvJwA4bydAOG8oADhvKEA4byiAOG8owDhvKQA4bylAOG8pgDhvKcA4byoAOG8qQDhvKoA4byrAOG8rADhvK0A4byuAOG8rwDhvLAA4byxAOG8sgDhvLMA4by0AOG8tQDhvLYA4by3AOG8uADhvLkA4by6AOG8uwDhvLwA4by9AOG8vgDhvL8A4b2AAOG9gQDhvYIA4b2DAOG9hADhvYUA4b2IAOG9iQDhvYoA4b2LAOG9jADhvY0A4b2QAOG9kQDhvZIA4b2TAOG9lADhvZUA4b2WAOG9lwDhvZkA4b2bAOG9nQDhvZ8A4b2gAOG9oQDhvaIA4b2jAOG9pADhvaUA4b2mAOG9pwDhvagA4b2pAOG9qgDhvasA4b2sAOG9rQDhva4A4b2vAOG9sADhvbIA4b20AOG9tgDhvbgA4b26AOG9vADhvoAA4b6BAOG+ggDhvoMA4b6EAOG+hQDhvoYA4b6HAOG+iADhvokA4b6KAOG+iwDhvowA4b6NAOG+jgDhvo8A4b6QAOG+kQDhvpIA4b6TAOG+lADhvpUA4b6WAOG+lwDhvpgA4b6ZAOG+mgDhvpsA4b6cAOG+nQDhvp4A4b6fAOG+oADhvqEA4b6iAOG+owDhvqQA4b6lAOG+pgDhvqcA4b6oAOG+qQDhvqoA4b6rAOG+rADhvq0A4b6uAOG+rwDhvrAA4b6xAOG+sgDhvrMA4b60AOG+tgDhvrcA4b64AOG+uQDhvroA4b68AOG/ggDhv4MA4b+EAOG/hgDhv4cA4b+IAOG/igDhv4wA4b+QAOG/kQDhv5IA4b+WAOG/lwDhv5gA4b+ZAOG/mgDhv6AA4b+hAOG/ogDhv6QA4b+lAOG/pgDhv6cA4b+oAOG/qQDhv6oA4b+sAOG/sgDhv7MA4b+0AOG/tgDhv7cA4b+4AOG/ugDhv7wA4oCQAOKAkwDigJQA4oCy4oCyAOKAsuKAsuKAsgDigLLigLLigLLigLIA4oC14oC1AOKAteKAteKAtQDigqkA4oaQAOKGkQDihpIA4oaTAOKGmgDihpsA4oauAOKHjQDih44A4oePAOKIggDiiIQA4oiHAOKIiQDiiIwA4oiRAOKIkgDiiKQA4oimAOKIq+KIqwDiiKviiKviiKsA4oir4oir4oir4oirAOKIruKIrgDiiK7iiK7iiK4A4omBAOKJhADiiYcA4omJAOKJoADiiaIA4omtAOKJrgDiia8A4omwAOKJsQDiibQA4om1AOKJuADiibkA4oqAAOKKgQDiioQA4oqFAOKKiADiiokA4oqsAOKKrQDiiq4A4oqvAOKLoADii6EA4ouiAOKLowDii6oA4ourAOKLrADii60A4pSCAOKWoADil4sA4qaFAOKmhgDiq53MuADitaEA44CBAOOAggDjgIgA44CJAOOAigDjgIsA44CMAOOAjQDjgI4A44CPAOOAkADjgJEA44CSAOOAlADjgJRT44CVAOOAlOS4ieOAlQDjgJTkuozjgJUA44CU5Yud44CVAOOAlOWuieOAlQDjgJTmiZPjgJUA44CU5pWX44CVAOOAlOacrOOAlQDjgJTngrnjgJUA44CU55uX44CVAOOAlQDjgJYA44CXAOOBjADjgY4A44GQAOOBkgDjgZQA44GWAOOBmADjgZoA44GcAOOBngDjgaAA44GiAOOBpQDjgacA44GpAOOBsADjgbEA44GzAOOBtADjgbYA44G3AOOBuQDjgboA44G744GLAOOBvADjgb0A44KI44KKAOOClADjgpkA44KaAOOCngDjgqEA44KiAOOCouODkeODvOODiADjgqLjg6vjg5XjgqEA44Ki44Oz44Oa44KiAOOCouODvOODqwDjgqMA44KkAOOCpOODi+ODs+OCsADjgqTjg7Pjg4EA44KlAOOCpgDjgqbjgqnjg7MA44KnAOOCqADjgqjjgrnjgq/jg7zjg4kA44Ko44O844Kr44O8AOOCqQDjgqoA44Kq44Oz44K5AOOCquODvOODoADjgqsA44Kr44Kk44OqAOOCq+ODqeODg+ODiADjgqvjg63jg6rjg7wA44KsAOOCrOODreODswDjgqzjg7Pjg54A44KtAOOCreODpeODquODvADjgq3jg60A44Kt44Ot44Kw44Op44OgAOOCreODreODoeODvOODiOODqwDjgq3jg63jg6/jg4Pjg4gA44KuAOOCruOCrADjgq7jg4vjg7wA44Ku44Or44OA44O8AOOCrwDjgq/jg6vjgrzjgqTjg60A44Kv44Ot44O844ONAOOCsADjgrDjg6njg6AA44Kw44Op44Og44OI44OzAOOCsQDjgrHjg7zjgrkA44KyAOOCswDjgrPjgrMA44Kz44OIAOOCs+ODq+ODigDjgrPjg7zjg50A44K0AOOCtQDjgrXjgqTjgq/jg6sA44K144Oz44OB44O844OgAOOCtgDjgrcA44K344Oq44Oz44KwAOOCuADjgrkA44K6AOOCuwDjgrvjg7Pjg4EA44K744Oz44OIAOOCvADjgr0A44K+AOOCvwDjg4AA44OA44O844K5AOODgQDjg4IA44ODAOODhADjg4UA44OGAOODhwDjg4fjgrcA44OIAOODiOODswDjg4kA44OJ44OrAOODigDjg4rjg44A44OLAOODjADjg40A44OOAOODjuODg+ODiADjg48A44OP44Kk44OEAOODkADjg5Djg7zjg6zjg6sA44ORAOODkeODvOOCu+ODs+ODiADjg5Hjg7zjg4QA44OSAOODkwDjg5Pjg6sA44OUAOODlOOCouOCueODiOODqwDjg5Tjgq/jg6sA44OU44KzAOODlQDjg5XjgqHjg6njg4Pjg4kA44OV44Kj44O844OIAOODleODqeODswDjg5YA44OW44OD44K344Kn44OrAOODlwDjg5gA44OY44Kv44K/44O844OrAOODmOODq+ODhADjg5kA44OZ44O844K/AOODmgDjg5rjgr0A44Oa44OL44OSAOODmuODs+OCuQDjg5rjg7zjgrgA44ObAOODm+ODswDjg5vjg7zjg6sA44Ob44O844OzAOODnADjg5zjg6vjg4gA44OdAOODneOCpOODs+ODiADjg53jg7Pjg4kA44OeAOODnuOCpOOCr+ODrQDjg57jgqTjg6sA44Oe44OD44OPAOODnuODq+OCrwDjg57jg7Pjgrfjg6fjg7MA44OfAOODn+OCr+ODreODswDjg5/jg6oA44Of44Oq44OQ44O844OrAOODoADjg6EA44Oh44KsAOODoeOCrOODiOODswDjg6Hjg7zjg4jjg6sA44OiAOODowDjg6QA44Ok44O844OJAOODpOODvOODqwDjg6UA44OmAOODpuOCouODswDjg6cA44OoAOODqQDjg6oA44Oq44OD44OI44OrAOODquODqQDjg6sA44Or44OU44O8AOODq+ODvOODluODqwDjg6wA44Os44OgAOODrOODs+ODiOOCsuODswDjg60A44OvAOODr+ODg+ODiADjg7AA44OxAOODsgDjg7MA44O0AOODtwDjg7gA44O5AOODugDjg7sA44O8AOODvgDjkp4A45K5AOOSuwDjk58A45SVAOObrgDjm7wA456BAOOgrwDjoaIA46G8AOOjhwDjo6MA46ScAOOkugDjqK4A46msAOOrpADjrIgA46yZAOOtiQDjrp0A47CYAOOxjgDjtLMA47aWAOO6rADjurgA47ybAOO/vADkgIgA5ICYAOSAuQDkgYYA5IKWAOSDowDkhK8A5IiCAOSIpwDkiqAA5IyBAOSMtADkjZkA5I+VAOSPmQDkkIsA5JGrAOSUqwDklZ0A5JWhAOSVqwDkl5cA5Je5AOSYtQDkmr4A5JuHAOSmlQDkp6YA5KmuAOSptgDkqrIA5KyzAOSvjgDks44A5LOtAOSzuADktZYA5LiAAOS4gQDkuIMA5LiJAOS4igDkuIsA5LiNAOS4mQDkuKYA5LioAOS4rQDkuLIA5Li2AOS4uADkuLkA5Li9AOS4vwDkuYEA5LmZAOS5nQDkuoIA5LqFAOS6hgDkuowA5LqUAOS6oADkuqQA5LquAOS6ugDku4AA5LuMAOS7pADku6TlkowA5LyBAOS8kQDkvaAA5L6AAOS+hgDkvosA5L6uAOS+uwDkvr8A5YCCAOWAqwDlgboA5YKZAOWDjwDlg5oA5YOnAOWEqgDlhL8A5YWAAOWFhQDlhY0A5YWUAOWFpADlhaUA5YWnAOWFqADlhakA5YWrAOWFrQDlhbcA5YaAAOWGggDlho0A5YaSAOWGlQDlhpYA5YaXAOWGmQDlhqQA5YarAOWGrADlhrUA5Ya3AOWHiQDlh4wA5YecAOWHngDlh6AA5Ye1AOWIgADliIMA5YiHAOWIlwDliJ0A5YipAOWIugDliLsA5YmGAOWJjQDlibIA5Ym3AOWKiQDlipsA5YqjAOWKswDlirQA5YuHAOWLiQDli5IA5YueAOWLpADli7UA5Yu5AOWLugDljIUA5YyGAOWMlQDljJcA5YyaAOWMuADljLsA5Yy/AOWNgQDljYQA5Y2FAOWNiQDljZEA5Y2UAOWNmgDljZwA5Y2pAOWNsADljbMA5Y21AOWNvQDljb8A5Y6CAOWOtgDlj4MA5Y+IAOWPigDlj4wA5Y+fAOWPowDlj6UA5Y+rAOWPrwDlj7EA5Y+zAOWQhgDlkIgA5ZCNAOWQjwDlkJ0A5ZC4AOWQuQDlkYIA5ZGIAOWRqADlkp4A5ZKiAOWSvQDlk7YA5ZSQAOWVjwDllZMA5ZWVAOWVowDlloQA5ZaHAOWWmQDllp0A5ZarAOWWswDllrYA5ZeAAOWXggDll6IA5ZiGAOWZkQDlmagA5Zm0AOWblwDlm5sA5Zu5AOWclgDlnJcA5ZyfAOWcsADlnosA5Z+OAOWftADloI0A5aCxAOWgsgDloYAA5aGaAOWhngDloqgA5aKsAOWiswDlo5gA5aOfAOWjqwDlo64A5aOwAOWjsgDlo7cA5aSCAOWkhgDlpIoA5aSVAOWkmgDlpJwA5aSiAOWkpwDlpKfmraMA5aSpAOWlhADlpYgA5aWRAOWllADlpaIA5aWzAOWnmADlp6wA5aibAOWopwDlqaIA5ammAOWqtQDlrIgA5ayoAOWsvgDlrZAA5a2XAOWtpgDlroAA5a6FAOWulwDlr4MA5a+YAOWvpwDlr64A5a+zAOWvuADlr78A5bCGAOWwjwDlsKIA5bC4AOWwvwDlsaAA5bGiAOWxpADlsaUA5bGuAOWxsQDlso0A5bOAAOW0mQDltYMA5bWQAOW1qwDlta4A5bW8AOW2sgDltroA5bebAOW3oQDlt6IA5belAOW3pgDlt7EA5be9AOW3vgDluKgA5bi9AOW5qQDlubIA5bmz5oiQAOW5tADluboA5bm8AOW5vwDluqYA5bqwAOW6swDlurYA5buJAOW7igDlu5IA5buTAOW7mQDlu6wA5bu0AOW7vgDlvIQA5byLAOW8kwDlvKIA5b2QAOW9kwDlvaEA5b2iAOW9qQDlvasA5b2zAOW+iwDlvowA5b6XAOW+mgDlvqkA5b6tAOW/gwDlv40A5b+XAOW/tQDlv7kA5oCSAOaAnADmgbUA5oKBAOaClADmg4cA5oOYAOaDoQDmhIgA5oWEAOaFiADmhYwA5oWOAOaFoADmhagA5oW6AOaGjgDmhpAA5oakAOaGrwDmhrIA5oeeAOaHsgDmh7YA5oiAAOaIiADmiJAA5oibAOaIrgDmiLQA5oi2AOaJiwDmiZMA5omdAOaKlQDmirEA5ouJAOaLjwDmi5MA5ouUAOaLvADmi74A5oyHAOaMvQDmjZAA5o2VAOaNqADmjbsA5o6DAOaOoADmjqkA5o+EAOaPhQDmj6QA5pCcAOaQogDmkZIA5pGpAOaRtwDmkb4A5pKaAOaSnQDmk4QA5pSvAOaUtADmlY8A5pWWAOaVrADmlbgA5paHAOaWlwDmlpkA5pakAOaWsADmlrkA5peFAOaXoADml6IA5pejAOaXpQDmmI7msrsA5piTAOaYoADmmK3lkowA5pmJAOaZtADmmogA5pqRAOaanADmmrQA5puGAOabsADmm7QA5pu4AOacgADmnIgA5pyJAOaclwDmnJsA5pyhAOacqADmnY4A5p2TAOadlgDmnZ4A5p27AOaehQDmnpcA5p+zAOafugDmoJcA5qCfAOagqgDmoKrlvI/kvJrnpL4A5qGSAOaigQDmooUA5qKOAOaiqADmpJQA5qWCAOamowDmp6oA5qiCAOaokwDmqqgA5quTAOarmwDmrIQA5qygAOasoQDmrZQA5q2iAOatowDmrbIA5q23AOatuQDmrp8A5q6uAOauswDmrroA5q67AOaviwDmr40A5q+UAOavmwDmsI8A5rCUAOawtADmsY4A5rGnAOayiADmsr8A5rOMAOazjQDms6UA5rOoAOa0lgDmtJsA5rSeAOa0tADmtL4A5rWBAOa1qQDmtaoA5rW3AOa1uADmtoUA5reLAOa3mgDmt6oA5re5AOa4mgDmuK8A5rmuAOa6gADmupwA5rq6AOa7hwDmu4sA5ruRAOa7mwDmvI8A5ryUAOa8ogDmvKMA5r2uAOa/hgDmv6sA5r++AOeAmwDngJ4A54C5AOeBigDngasA54GwAOeBtwDngb0A54KZAOeCrQDng4gA54OZAOeEoQDnhYUA54WJAOeFrgDnhpwA54eOAOeHkADniJAA54ibAOeIqADniKoA54irAOeItQDniLYA54i7AOeIvwDniYcA54mQAOeJmQDniZsA54miAOeJuQDnioAA54qVAOeKrADniq8A54uAAOeLvADnjKoA5421AOeNugDnjoQA546HAOeOiQDnjosA546lAOeOsgDnj54A55CGAOeQiQDnkKIA55GHAOeRnADnkakA55GxAOeShQDnkokA55KYAOeTigDnk5wA55OmAOeUhgDnlJgA55SfAOeUpADnlKgA55SwAOeUsgDnlLMA55S3AOeUuwDnlL4A55WZAOeVpQDnlbAA55aLAOeWkgDnl6IA55iQAOeYnQDnmJ8A55mCAOeZqQDnmbYA55m9AOeargDnmr8A55uKAOebmwDnm6MA55unAOebrgDnm7QA55yBAOecngDnnJ8A552AAOedigDnnosA556nAOefmwDnn6IA55+zAOehjgDnoasA56KMAOeikQDno4oA56OMAOejuwDnpKoA56S6AOekvADnpL4A56WIAOeliQDnpZAA56WWAOelnQDnpZ4A56WlAOelvwDnpoEA56aNAOemjgDnpo8A56auAOemuADnpr4A56eKAOenmADnp6sA56icAOepgADnqYoA56mPAOeptADnqboA56qBAOeqsQDnq4sA56uuAOeruQDnrKAA566PAOevgADnr4YA56+JAOewvgDnsaAA57GzAOexuwDnspIA57K+AOezkgDns5YA57OjAOezpwDns6gA57O4AOe0gADntJAA57SiAOe0rwDntYIA57WbAOe1owDntqAA57a+AOe3hwDnt7QA57iCAOe4iQDnuLcA57mBAOe5hQDnvLYA57y+AOe9kQDnvbIA5725AOe9ugDnvoUA576KAOe+lQDnvpoA5769AOe/ugDogIEA6ICFAOiAjADogJIA6ICzAOiBhgDogaAA6IGvAOiBsADogb4A6IG/AOiCiQDogosA6IKtAOiCsgDohIMA6IS+AOiHmADoh6MA6IeoAOiHqgDoh60A6IezAOiHvADoiIEA6IiEAOiIjADoiJgA6IibAOiInwDoia4A6ImvAOiJsgDoibgA6Im5AOiKiwDoipEA6IqdAOiKsQDoirMA6Iq9AOiLpQDoi6YA6IydAOiMowDojLYA6I2SAOiNkwDojaMA6I6tAOiOvQDoj4kA6I+KAOiPjADoj5wA6I+nAOiPrwDoj7EA6JC9AOiRiQDokZcA6JOuAOiTsQDok7MA6JO8AOiUlgDolaQA6JeNAOiXugDomIYA6JiSAOiYrQDomL8A6JmNAOiZkADomZwA6JmnAOiZqQDomasA6JqIAOiaqQDom6IA6JyOAOicqADonasA6J25AOiehgDonroA6J+hAOiggQDooJ8A6KGAAOihjADooaAA6KGjAOijggDoo48A6KOXAOijngDoo6EA6KO4AOijugDopJAA6KWBAOilpADopb4A6KaGAOimiwDoppYA6KeSAOinowDoqIAA6KqgAOiqqgDoqr8A6KuLAOirkgDoq5YA6KutAOiruADoq74A6KyBAOisuQDorZgA6K6AAOiuigDosLcA6LGGAOixiADosZUA6LG4AOiynQDosqEA6LKpAOiyqwDos4EA6LOCAOizhwDos4gA6LOTAOi0iADotJsA6LWkAOi1sADotbcA6LazAOi2vADot4sA6LevAOi3sADouqsA6LuKAOi7lADovKYA6LyqAOi8uADovLsA6L2iAOi+mwDovp4A6L6wAOi+tQDovrYA6YCjAOmAuADpgYoA6YGpAOmBsgDpgbwA6YKPAOmCkQDpgpQA6YOOAOmDngDpg7EA6YO9AOmEkQDphJsA6YWJAOmFjQDphaoA6YaZAOmGtADph4YA6YeMAOmHjwDph5EA6Yi0AOmIuADpibYA6Ym8AOmLlwDpi5gA6YyEAOmNigDpj7kA6ZCVAOmVtwDploAA6ZaLAOmWrQDplrcA6ZicAOmYrgDpmYsA6ZmNAOmZtQDpmbgA6Zm8AOmahgDpmqMA6Zq2AOmatwDpmrgA6Zq5AOmbgwDpm6IA6ZujAOmbqADpm7YA6Zu3AOmcowDpnLIA6Z2IAOmdkQDpnZYA6Z2eAOmdogDpnakA6Z+LAOmfmwDpn6AA6Z+tAOmfswDpn78A6aCBAOmghQDpoIsA6aCYAOmgqQDpoLsA6aGeAOmiqADpo5sA6aOfAOmjogDpo68A6aO8AOmkqADppKkA6aaWAOmmmQDppqcA6aasAOmnggDpp7EA6ae+AOmpqgDpqqgA6auYAOmrnwDprJIA6aylAOmsrwDprLIA6ay8AOmtmgDpra8A6bGAAOmxlwDps6UA6bO9AOm1pwDptrQA6be6AOm4ngDpubUA6bm/AOm6lwDpup8A6bqlAOm6uwDpu4MA6buNAOm7jgDpu5EA6bu5AOm7vQDpu74A6byFAOm8jgDpvI8A6byTAOm8lgDpvKAA6by7AOm9gwDpvYoA6b2SAOm+jQDpvo4A6b6cAOm+nwDpvqAA6pynAOqdrwDqrLcA6q2SAOqwgADqsIEA6rCCAOqwgwDqsIQA6rCFAOqwhgDqsIcA6rCIAOqwiQDqsIoA6rCLAOqwjADqsI0A6rCOAOqwjwDqsJAA6rCRAOqwkgDqsJMA6rCUAOqwlQDqsJYA6rCXAOqwmADqsJkA6rCaAOqwmwDqsJwA6rCdAOqwngDqsJ8A6rCgAOqwoQDqsKIA6rCjAOqwpADqsKUA6rCmAOqwpwDqsKgA6rCpAOqwqgDqsKsA6rCsAOqwrQDqsK4A6rCvAOqwsADqsLEA6rCyAOqwswDqsLQA6rC1AOqwtgDqsLcA6rC4AOqwuQDqsLoA6rC7AOqwvADqsL0A6rC+AOqwvwDqsYAA6rGBAOqxggDqsYMA6rGEAOqxhQDqsYYA6rGHAOqxiADqsYkA6rGKAOqxiwDqsYwA6rGNAOqxjgDqsY8A6rGQAOqxkQDqsZIA6rGTAOqxlADqsZUA6rGWAOqxlwDqsZgA6rGZAOqxmgDqsZsA6rGcAOqxnQDqsZ4A6rGfAOqxoADqsaEA6rGiAOqxowDqsaQA6rGlAOqxpgDqsacA6rGoAOqxqQDqsaoA6rGrAOqxrADqsa0A6rGuAOqxrwDqsbAA6rGxAOqxsgDqsbMA6rG0AOqxtQDqsbYA6rG3AOqxuADqsbkA6rG6AOqxuwDqsbwA6rG9AOqxvgDqsb8A6rKAAOqygQDqsoIA6rKDAOqyhADqsoUA6rKGAOqyhwDqsogA6rKJAOqyigDqsosA6rKMAOqyjQDqso4A6rKPAOqykADqspEA6rKSAOqykwDqspQA6rKVAOqylgDqspcA6rKYAOqymQDqspoA6rKbAOqynADqsp0A6rKeAOqynwDqsqAA6rKhAOqyogDqsqMA6rKkAOqypQDqsqYA6rKnAOqyqADqsqkA6rKqAOqyqwDqsqwA6rKtAOqyrgDqsq8A6rKwAOqysQDqsrIA6rKzAOqytADqsrUA6rK2AOqytwDqsrgA6rK5AOqyugDqsrsA6rK8AOqyvQDqsr4A6rK/AOqzgADqs4EA6rOCAOqzgwDqs4QA6rOFAOqzhgDqs4cA6rOIAOqziQDqs4oA6rOLAOqzjADqs40A6rOOAOqzjwDqs5AA6rORAOqzkgDqs5MA6rOUAOqzlQDqs5YA6rOXAOqzmADqs5kA6rOaAOqzmwDqs5wA6rOdAOqzngDqs58A6rOgAOqzoQDqs6IA6rOjAOqzpADqs6UA6rOmAOqzpwDqs6gA6rOpAOqzqgDqs6sA6rOsAOqzrQDqs64A6rOvAOqzsADqs7EA6rOyAOqzswDqs7QA6rO1AOqztgDqs7cA6rO4AOqzuQDqs7oA6rO7AOqzvADqs70A6rO+AOqzvwDqtIAA6rSBAOq0ggDqtIMA6rSEAOq0hQDqtIYA6rSHAOq0iADqtIkA6rSKAOq0iwDqtIwA6rSNAOq0jgDqtI8A6rSQAOq0kQDqtJIA6rSTAOq0lADqtJUA6rSWAOq0lwDqtJgA6rSZAOq0mgDqtJsA6rScAOq0nQDqtJ4A6rSfAOq0oADqtKEA6rSiAOq0owDqtKQA6rSlAOq0pgDqtKcA6rSoAOq0qQDqtKoA6rSrAOq0rADqtK0A6rSuAOq0rwDqtLAA6rSxAOq0sgDqtLMA6rS0AOq0tQDqtLYA6rS3AOq0uADqtLkA6rS6AOq0uwDqtLwA6rS9AOq0vgDqtL8A6rWAAOq1gQDqtYIA6rWDAOq1hADqtYUA6rWGAOq1hwDqtYgA6rWJAOq1igDqtYsA6rWMAOq1jQDqtY4A6rWPAOq1kADqtZEA6rWSAOq1kwDqtZQA6rWVAOq1lgDqtZcA6rWYAOq1mQDqtZoA6rWbAOq1nADqtZ0A6rWeAOq1nwDqtaAA6rWhAOq1ogDqtaMA6rWkAOq1pQDqtaYA6rWnAOq1qADqtakA6rWqAOq1qwDqtawA6rWtAOq1rgDqta8A6rWwAOq1sQDqtbIA6rWzAOq1tADqtbUA6rW2AOq1twDqtbgA6rW5AOq1ugDqtbsA6rW8AOq1vQDqtb4A6rW/AOq2gADqtoEA6raCAOq2gwDqtoQA6raFAOq2hgDqtocA6raIAOq2iQDqtooA6raLAOq2jADqto0A6raOAOq2jwDqtpAA6raRAOq2kgDqtpMA6raUAOq2lQDqtpYA6raXAOq2mADqtpkA6raaAOq2mwDqtpwA6radAOq2ngDqtp8A6ragAOq2oQDqtqIA6rajAOq2pADqtqUA6ramAOq2pwDqtqgA6rapAOq2qgDqtqsA6rasAOq2rQDqtq4A6ravAOq2sADqtrEA6rayAOq2swDqtrQA6ra1AOq2tgDqtrcA6ra4AOq2uQDqtroA6ra7AOq2vADqtr0A6ra+AOq2vwDqt4AA6reBAOq3ggDqt4MA6reEAOq3hQDqt4YA6reHAOq3iADqt4kA6reKAOq3iwDqt4wA6reNAOq3jgDqt48A6reQAOq3kQDqt5IA6reTAOq3lADqt5UA6reWAOq3lwDqt5gA6reZAOq3mgDqt5sA6recAOq3nQDqt54A6refAOq3oADqt6EA6reiAOq3owDqt6QA6relAOq3pgDqt6cA6reoAOq3qQDqt6oA6rerAOq3rADqt60A6reuAOq3rwDqt7AA6rexAOq3sgDqt7MA6re0AOq3tQDqt7YA6re3AOq3uADqt7kA6re6AOq3uwDqt7wA6re9AOq3vgDqt78A6riAAOq4gQDquIIA6riDAOq4hADquIUA6riGAOq4hwDquIgA6riJAOq4igDquIsA6riMAOq4jQDquI4A6riPAOq4kADquJEA6riSAOq4kwDquJQA6riVAOq4lgDquJcA6riYAOq4mQDquJoA6ribAOq4nADquJ0A6rieAOq4nwDquKAA6rihAOq4ogDquKMA6rikAOq4pQDquKYA6rinAOq4qADquKkA6riqAOq4qwDquKwA6ritAOq4rgDquK8A6riwAOq4sQDquLIA6rizAOq4tADquLUA6ri2AOq4twDquLgA6ri5AOq4ugDquLsA6ri8AOq4vQDquL4A6ri/AOq5gADquYEA6rmCAOq5gwDquYQA6rmFAOq5hgDquYcA6rmIAOq5iQDquYoA6rmLAOq5jADquY0A6rmOAOq5jwDquZAA6rmRAOq5kgDquZMA6rmUAOq5lQDquZYA6rmXAOq5mADquZkA6rmaAOq5mwDquZwA6rmdAOq5ngDquZ8A6rmgAOq5oQDquaIA6rmjAOq5pADquaUA6rmmAOq5pwDquagA6rmpAOq5qgDquasA6rmsAOq5rQDqua4A6rmvAOq5sADqubEA6rmyAOq5swDqubQA6rm1AOq5tgDqubcA6rm4AOq5uQDquboA6rm7AOq5vADqub0A6rm+AOq5vwDquoAA6rqBAOq6ggDquoMA6rqEAOq6hQDquoYA6rqHAOq6iADquokA6rqKAOq6iwDquowA6rqNAOq6jgDquo8A6rqQAOq6kQDqupIA6rqTAOq6lADqupUA6rqWAOq6lwDqupgA6rqZAOq6mgDqupsA6rqcAOq6nQDqup4A6rqfAOq6oADquqEA6rqiAOq6owDquqQA6rqlAOq6pgDquqcA6rqoAOq6qQDquqoA6rqrAOq6rADquq0A6rquAOq6rwDqurAA6rqxAOq6sgDqurMA6rq0AOq6tQDqurYA6rq3AOq6uADqurkA6rq6AOq6uwDqurwA6rq9AOq6vgDqur8A6ruAAOq7gQDqu4IA6ruDAOq7hADqu4UA6ruGAOq7hwDqu4gA6ruJAOq7igDqu4sA6ruMAOq7jQDqu44A6ruPAOq7kADqu5EA6ruSAOq7kwDqu5QA6ruVAOq7lgDqu5cA6ruYAOq7mQDqu5oA6rubAOq7nADqu50A6rueAOq7nwDqu6AA6ruhAOq7ogDqu6MA6rukAOq7pQDqu6YA6runAOq7qADqu6kA6ruqAOq7qwDqu6wA6rutAOq7rgDqu68A6ruwAOq7sQDqu7IA6ruzAOq7tADqu7UA6ru2AOq7twDqu7gA6ru5AOq7ugDqu7sA6ru8AOq7vQDqu74A6ru/AOq8gADqvIEA6ryCAOq8gwDqvIQA6ryFAOq8hgDqvIcA6ryIAOq8iQDqvIoA6ryLAOq8jADqvI0A6ryOAOq8jwDqvJAA6ryRAOq8kgDqvJMA6ryUAOq8lQDqvJYA6ryXAOq8mADqvJkA6ryaAOq8mwDqvJwA6rydAOq8ngDqvJ8A6rygAOq8oQDqvKIA6ryjAOq8pADqvKUA6rymAOq8pwDqvKgA6rypAOq8qgDqvKsA6rysAOq8rQDqvK4A6ryvAOq8sADqvLEA6ryyAOq8swDqvLQA6ry1AOq8tgDqvLcA6ry4AOq8uQDqvLoA6ry7AOq8vADqvL0A6ry+AOq8vwDqvYAA6r2BAOq9ggDqvYMA6r2EAOq9hQDqvYYA6r2HAOq9iADqvYkA6r2KAOq9iwDqvYwA6r2NAOq9jgDqvY8A6r2QAOq9kQDqvZIA6r2TAOq9lADqvZUA6r2WAOq9lwDqvZgA6r2ZAOq9mgDqvZsA6r2cAOq9nQDqvZ4A6r2fAOq9oADqvaEA6r2iAOq9owDqvaQA6r2lAOq9pgDqvacA6r2oAOq9qQDqvaoA6r2rAOq9rADqva0A6r2uAOq9rwDqvbAA6r2xAOq9sgDqvbMA6r20AOq9tQDqvbYA6r23AOq9uADqvbkA6r26AOq9uwDqvbwA6r29AOq9vgDqvb8A6r6AAOq+gQDqvoIA6r6DAOq+hADqvoUA6r6GAOq+hwDqvogA6r6JAOq+igDqvosA6r6MAOq+jQDqvo4A6r6PAOq+kADqvpEA6r6SAOq+kwDqvpQA6r6VAOq+lgDqvpcA6r6YAOq+mQDqvpoA6r6bAOq+nADqvp0A6r6eAOq+nwDqvqAA6r6hAOq+ogDqvqMA6r6kAOq+pQDqvqYA6r6nAOq+qADqvqkA6r6qAOq+qwDqvqwA6r6tAOq+rgDqvq8A6r6wAOq+sQDqvrIA6r6zAOq+tADqvrUA6r62AOq+twDqvrgA6r65AOq+ugDqvrsA6r68AOq+vQDqvr4A6r6/AOq/gADqv4EA6r+CAOq/gwDqv4QA6r+FAOq/hgDqv4cA6r+IAOq/iQDqv4oA6r+LAOq/jADqv40A6r+OAOq/jwDqv5AA6r+RAOq/kgDqv5MA6r+UAOq/lQDqv5YA6r+XAOq/mADqv5kA6r+aAOq/mwDqv5wA6r+dAOq/ngDqv58A6r+gAOq/oQDqv6IA6r+jAOq/pADqv6UA6r+mAOq/pwDqv6gA6r+pAOq/qgDqv6sA6r+sAOq/rQDqv64A6r+vAOq/sADqv7EA6r+yAOq/swDqv7QA6r+1AOq/tgDqv7cA6r+4AOq/uQDqv7oA6r+7AOq/vADqv70A6r++AOq/vwDrgIAA64CBAOuAggDrgIMA64CEAOuAhQDrgIYA64CHAOuAiADrgIkA64CKAOuAiwDrgIwA64CNAOuAjgDrgI8A64CQAOuAkQDrgJIA64CTAOuAlADrgJUA64CWAOuAlwDrgJgA64CZAOuAmgDrgJsA64CcAOuAnQDrgJ4A64CfAOuAoADrgKEA64CiAOuAowDrgKQA64ClAOuApgDrgKcA64CoAOuAqQDrgKoA64CrAOuArADrgK0A64CuAOuArwDrgLAA64CxAOuAsgDrgLMA64C0AOuAtQDrgLYA64C3AOuAuADrgLkA64C6AOuAuwDrgLwA64C9AOuAvgDrgL8A64GAAOuBgQDrgYIA64GDAOuBhADrgYUA64GGAOuBhwDrgYgA64GJAOuBigDrgYsA64GMAOuBjQDrgY4A64GPAOuBkADrgZEA64GSAOuBkwDrgZQA64GVAOuBlgDrgZcA64GYAOuBmQDrgZoA64GbAOuBnADrgZ0A64GeAOuBnwDrgaAA64GhAOuBogDrgaMA64GkAOuBpQDrgaYA64GnAOuBqADrgakA64GqAOuBqwDrgawA64GtAOuBrgDrga8A64GwAOuBsQDrgbIA64GzAOuBtADrgbUA64G2AOuBtwDrgbgA64G5AOuBugDrgbsA64G8AOuBvQDrgb4A64G/AOuCgADrgoEA64KCAOuCgwDrgoQA64KFAOuChgDrgocA64KIAOuCiQDrgooA64KLAOuCjADrgo0A64KOAOuCjwDrgpAA64KRAOuCkgDrgpMA64KUAOuClQDrgpYA64KXAOuCmADrgpkA64KaAOuCmwDrgpwA64KdAOuCngDrgp8A64KgAOuCoQDrgqIA64KjAOuCpADrgqUA64KmAOuCpwDrgqgA64KpAOuCqgDrgqsA64KsAOuCrQDrgq4A64KvAOuCsADrgrEA64KyAOuCswDrgrQA64K1AOuCtgDrgrcA64K4AOuCuQDrgroA64K7AOuCvADrgr0A64K+AOuCvwDrg4AA64OBAOuDggDrg4MA64OEAOuDhQDrg4YA64OHAOuDiADrg4kA64OKAOuDiwDrg4wA64ONAOuDjgDrg48A64OQAOuDkQDrg5IA64OTAOuDlADrg5UA64OWAOuDlwDrg5gA64OZAOuDmgDrg5sA64OcAOuDnQDrg54A64OfAOuDoADrg6EA64OiAOuDowDrg6QA64OlAOuDpgDrg6cA64OoAOuDqQDrg6oA64OrAOuDrADrg60A64OuAOuDrwDrg7AA64OxAOuDsgDrg7MA64O0AOuDtQDrg7YA64O3AOuDuADrg7kA64O6AOuDuwDrg7wA64O9AOuDvgDrg78A64SAAOuEgQDrhIIA64SDAOuEhADrhIUA64SGAOuEhwDrhIgA64SJAOuEigDrhIsA64SMAOuEjQDrhI4A64SPAOuEkADrhJEA64SSAOuEkwDrhJQA64SVAOuElgDrhJcA64SYAOuEmQDrhJoA64SbAOuEnADrhJ0A64SeAOuEnwDrhKAA64ShAOuEogDrhKMA64SkAOuEpQDrhKYA64SnAOuEqADrhKkA64SqAOuEqwDrhKwA64StAOuErgDrhK8A64SwAOuEsQDrhLIA64SzAOuEtADrhLUA64S2AOuEtwDrhLgA64S5AOuEugDrhLsA64S8AOuEvQDrhL4A64S/AOuFgADrhYEA64WCAOuFgwDrhYQA64WFAOuFhgDrhYcA64WIAOuFiQDrhYoA64WLAOuFjADrhY0A64WOAOuFjwDrhZAA64WRAOuFkgDrhZMA64WUAOuFlQDrhZYA64WXAOuFmADrhZkA64WaAOuFmwDrhZwA64WdAOuFngDrhZ8A64WgAOuFoQDrhaIA64WjAOuFpADrhaUA64WmAOuFpwDrhagA64WpAOuFqgDrhasA64WsAOuFrQDrha4A64WvAOuFsADrhbEA64WyAOuFswDrhbQA64W1AOuFtgDrhbcA64W4AOuFuQDrhboA64W7AOuFvADrhb0A64W+AOuFvwDrhoAA64aBAOuGggDrhoMA64aEAOuGhQDrhoYA64aHAOuGiADrhokA64aKAOuGiwDrhowA64aNAOuGjgDrho8A64aQAOuGkQDrhpIA64aTAOuGlADrhpUA64aWAOuGlwDrhpgA64aZAOuGmgDrhpsA64acAOuGnQDrhp4A64afAOuGoADrhqEA64aiAOuGowDrhqQA64alAOuGpgDrhqcA64aoAOuGqQDrhqoA64arAOuGrADrhq0A64auAOuGrwDrhrAA64axAOuGsgDrhrMA64a0AOuGtQDrhrYA64a3AOuGuADrhrkA64a6AOuGuwDrhrwA64a9AOuGvgDrhr8A64eAAOuHgQDrh4IA64eDAOuHhADrh4UA64eGAOuHhwDrh4gA64eJAOuHigDrh4sA64eMAOuHjQDrh44A64ePAOuHkADrh5EA64eSAOuHkwDrh5QA64eVAOuHlgDrh5cA64eYAOuHmQDrh5oA64ebAOuHnADrh50A64eeAOuHnwDrh6AA64ehAOuHogDrh6MA64ekAOuHpQDrh6YA64enAOuHqADrh6kA64eqAOuHqwDrh6wA64etAOuHrgDrh68A64ewAOuHsQDrh7IA64ezAOuHtADrh7UA64e2AOuHtwDrh7gA64e5AOuHugDrh7sA64e8AOuHvQDrh74A64e/AOuIgADriIEA64iCAOuIgwDriIQA64iFAOuIhgDriIcA64iIAOuIiQDriIoA64iLAOuIjADriI0A64iOAOuIjwDriJAA64iRAOuIkgDriJMA64iUAOuIlQDriJYA64iXAOuImADriJkA64iaAOuImwDriJwA64idAOuIngDriJ8A64igAOuIoQDriKIA64ijAOuIpADriKUA64imAOuIpwDriKgA64ipAOuIqgDriKsA64isAOuIrQDriK4A64ivAOuIsADriLEA64iyAOuIswDriLQA64i1AOuItgDriLcA64i4AOuIuQDriLoA64i7AOuIvADriL0A64i+AOuIvwDriYAA64mBAOuJggDriYMA64mEAOuJhQDriYYA64mHAOuJiADriYkA64mKAOuJiwDriYwA64mNAOuJjgDriY8A64mQAOuJkQDriZIA64mTAOuJlADriZUA64mWAOuJlwDriZgA64mZAOuJmgDriZsA64mcAOuJnQDriZ4A64mfAOuJoADriaEA64miAOuJowDriaQA64mlAOuJpgDriacA64moAOuJqQDriaoA64mrAOuJrADria0A64muAOuJrwDribAA64mxAOuJsgDribMA64m0AOuJtQDribYA64m3AOuJuADribkA64m6AOuJuwDribwA64m9AOuJvgDrib8A64qAAOuKgQDrioIA64qDAOuKhADrioUA64qGAOuKhwDriogA64qJAOuKigDriosA64qMAOuKjQDrio4A64qPAOuKkADripEA64qSAOuKkwDripQA64qVAOuKlgDripcA64qYAOuKmQDripoA64qbAOuKnADrip0A64qeAOuKnwDriqAA64qhAOuKogDriqMA64qkAOuKpQDriqYA64qnAOuKqADriqkA64qqAOuKqwDriqwA64qtAOuKrgDriq8A64qwAOuKsQDrirIA64qzAOuKtADrirUA64q2AOuKtwDrirgA64q5AOuKugDrirsA64q8AOuKvQDrir4A64q/AOuLgADri4EA64uCAOuLgwDri4QA64uFAOuLhgDri4cA64uIAOuLiQDri4oA64uLAOuLjADri40A64uOAOuLjwDri5AA64uRAOuLkgDri5MA64uUAOuLlQDri5YA64uXAOuLmADri5kA64uaAOuLmwDri5wA64udAOuLngDri58A64ugAOuLoQDri6IA64ujAOuLpADri6UA64umAOuLpwDri6gA64upAOuLqgDri6sA64usAOuLrQDri64A64uvAOuLsADri7EA64uyAOuLswDri7QA64u1AOuLtgDri7cA64u4AOuLuQDri7oA64u7AOuLvADri70A64u+AOuLvwDrjIAA64yBAOuMggDrjIMA64yEAOuMhQDrjIYA64yHAOuMiADrjIkA64yKAOuMiwDrjIwA64yNAOuMjgDrjI8A64yQAOuMkQDrjJIA64yTAOuMlADrjJUA64yWAOuMlwDrjJgA64yZAOuMmgDrjJsA64ycAOuMnQDrjJ4A64yfAOuMoADrjKEA64yiAOuMowDrjKQA64ylAOuMpgDrjKcA64yoAOuMqQDrjKoA64yrAOuMrADrjK0A64yuAOuMrwDrjLAA64yxAOuMsgDrjLMA64y0AOuMtQDrjLYA64y3AOuMuADrjLkA64y6AOuMuwDrjLwA64y9AOuMvgDrjL8A642AAOuNgQDrjYIA642DAOuNhADrjYUA642GAOuNhwDrjYgA642JAOuNigDrjYsA642MAOuNjQDrjY4A642PAOuNkADrjZEA642SAOuNkwDrjZQA642VAOuNlgDrjZcA642YAOuNmQDrjZoA642bAOuNnADrjZ0A642eAOuNnwDrjaAA642hAOuNogDrjaMA642kAOuNpQDrjaYA642nAOuNqADrjakA642qAOuNqwDrjawA642tAOuNrgDrja8A642wAOuNsQDrjbIA642zAOuNtADrjbUA6422AOuNtwDrjbgA6425AOuNugDrjbsA6428AOuNvQDrjb4A642/AOuOgADrjoEA646CAOuOgwDrjoQA646FAOuOhgDrjocA646IAOuOiQDrjooA646LAOuOjADrjo0A646OAOuOjwDrjpAA646RAOuOkgDrjpMA646UAOuOlQDrjpYA646XAOuOmADrjpkA646aAOuOmwDrjpwA646dAOuOngDrjp8A646gAOuOoQDrjqIA646jAOuOpADrjqUA646mAOuOpwDrjqgA646pAOuOqgDrjqsA646sAOuOrQDrjq4A646vAOuOsADrjrEA646yAOuOswDrjrQA6461AOuOtgDrjrcA6464AOuOuQDrjroA6467AOuOvADrjr0A646+AOuOvwDrj4AA64+BAOuPggDrj4MA64+EAOuPhQDrj4YA64+HAOuPiADrj4kA64+KAOuPiwDrj4wA64+NAOuPjgDrj48A64+QAOuPkQDrj5IA64+TAOuPlADrj5UA64+WAOuPlwDrj5gA64+ZAOuPmgDrj5sA64+cAOuPnQDrj54A64+fAOuPoADrj6EA64+iAOuPowDrj6QA64+lAOuPpgDrj6cA64+oAOuPqQDrj6oA64+rAOuPrADrj60A64+uAOuPrwDrj7AA64+xAOuPsgDrj7MA64+0AOuPtQDrj7YA64+3AOuPuADrj7kA64+6AOuPuwDrj7wA64+9AOuPvgDrj78A65CAAOuQgQDrkIIA65CDAOuQhADrkIUA65CGAOuQhwDrkIgA65CJAOuQigDrkIsA65CMAOuQjQDrkI4A65CPAOuQkADrkJEA65CSAOuQkwDrkJQA65CVAOuQlgDrkJcA65CYAOuQmQDrkJoA65CbAOuQnADrkJ0A65CeAOuQnwDrkKAA65ChAOuQogDrkKMA65CkAOuQpQDrkKYA65CnAOuQqADrkKkA65CqAOuQqwDrkKwA65CtAOuQrgDrkK8A65CwAOuQsQDrkLIA65CzAOuQtADrkLUA65C2AOuQtwDrkLgA65C5AOuQugDrkLsA65C8AOuQvQDrkL4A65C/AOuRgADrkYEA65GCAOuRgwDrkYQA65GFAOuRhgDrkYcA65GIAOuRiQDrkYoA65GLAOuRjADrkY0A65GOAOuRjwDrkZAA65GRAOuRkgDrkZMA65GUAOuRlQDrkZYA65GXAOuRmADrkZkA65GaAOuRmwDrkZwA65GdAOuRngDrkZ8A65GgAOuRoQDrkaIA65GjAOuRpADrkaUA65GmAOuRpwDrkagA65GpAOuRqgDrkasA65GsAOuRrQDrka4A65GvAOuRsADrkbEA65GyAOuRswDrkbQA65G1AOuRtgDrkbcA65G4AOuRuQDrkboA65G7AOuRvADrkb0A65G+AOuRvwDrkoAA65KBAOuSggDrkoMA65KEAOuShQDrkoYA65KHAOuSiADrkokA65KKAOuSiwDrkowA65KNAOuSjgDrko8A65KQAOuSkQDrkpIA65KTAOuSlADrkpUA65KWAOuSlwDrkpgA65KZAOuSmgDrkpsA65KcAOuSnQDrkp4A65KfAOuSoADrkqEA65KiAOuSowDrkqQA65KlAOuSpgDrkqcA65KoAOuSqQDrkqoA65KrAOuSrADrkq0A65KuAOuSrwDrkrAA65KxAOuSsgDrkrMA65K0AOuStQDrkrYA65K3AOuSuADrkrkA65K6AOuSuwDrkrwA65K9AOuSvgDrkr8A65OAAOuTgQDrk4IA65ODAOuThADrk4UA65OGAOuThwDrk4gA65OJAOuTigDrk4sA65OMAOuTjQDrk44A65OPAOuTkADrk5EA65OSAOuTkwDrk5QA65OVAOuTlgDrk5cA65OYAOuTmQDrk5oA65ObAOuTnADrk50A65OeAOuTnwDrk6AA65OhAOuTogDrk6MA65OkAOuTpQDrk6YA65OnAOuTqADrk6kA65OqAOuTqwDrk6wA65OtAOuTrgDrk68A65OwAOuTsQDrk7IA65OzAOuTtADrk7UA65O2AOuTtwDrk7gA65O5AOuTugDrk7sA65O8AOuTvQDrk74A65O/AOuUgADrlIEA65SCAOuUgwDrlIQA65SFAOuUhgDrlIcA65SIAOuUiQDrlIoA65SLAOuUjADrlI0A65SOAOuUjwDrlJAA65SRAOuUkgDrlJMA65SUAOuUlQDrlJYA65SXAOuUmADrlJkA65SaAOuUmwDrlJwA65SdAOuUngDrlJ8A65SgAOuUoQDrlKIA65SjAOuUpADrlKUA65SmAOuUpwDrlKgA65SpAOuUqgDrlKsA65SsAOuUrQDrlK4A65SvAOuUsADrlLEA65SyAOuUswDrlLQA65S1AOuUtgDrlLcA65S4AOuUuQDrlLoA65S7AOuUvADrlL0A65S+AOuUvwDrlYAA65WBAOuVggDrlYMA65WEAOuVhQDrlYYA65WHAOuViADrlYkA65WKAOuViwDrlYwA65WNAOuVjgDrlY8A65WQAOuVkQDrlZIA65WTAOuVlADrlZUA65WWAOuVlwDrlZgA65WZAOuVmgDrlZsA65WcAOuVnQDrlZ4A65WfAOuVoADrlaEA65WiAOuVowDrlaQA65WlAOuVpgDrlacA65WoAOuVqQDrlaoA65WrAOuVrADrla0A65WuAOuVrwDrlbAA65WxAOuVsgDrlbMA65W0AOuVtQDrlbYA65W3AOuVuADrlbkA65W6AOuVuwDrlbwA65W9AOuVvgDrlb8A65aAAOuWgQDrloIA65aDAOuWhADrloUA65aGAOuWhwDrlogA65aJAOuWigDrlosA65aMAOuWjQDrlo4A65aPAOuWkADrlpEA65aSAOuWkwDrlpQA65aVAOuWlgDrlpcA65aYAOuWmQDrlpoA65abAOuWnADrlp0A65aeAOuWnwDrlqAA65ahAOuWogDrlqMA65akAOuWpQDrlqYA65anAOuWqADrlqkA65aqAOuWqwDrlqwA65atAOuWrgDrlq8A65awAOuWsQDrlrIA65azAOuWtADrlrUA65a2AOuWtwDrlrgA65a5AOuWugDrlrsA65a8AOuWvQDrlr4A65a/AOuXgADrl4EA65eCAOuXgwDrl4QA65eFAOuXhgDrl4cA65eIAOuXiQDrl4oA65eLAOuXjADrl40A65eOAOuXjwDrl5AA65eRAOuXkgDrl5MA65eUAOuXlQDrl5YA65eXAOuXmADrl5kA65eaAOuXmwDrl5wA65edAOuXngDrl58A65egAOuXoQDrl6IA65ejAOuXpADrl6UA65emAOuXpwDrl6gA65epAOuXqgDrl6sA65esAOuXrQDrl64A65evAOuXsADrl7EA65eyAOuXswDrl7QA65e1AOuXtgDrl7cA65e4AOuXuQDrl7oA65e7AOuXvADrl70A65e+AOuXvwDrmIAA65iBAOuYggDrmIMA65iEAOuYhQDrmIYA65iHAOuYiADrmIkA65iKAOuYiwDrmIwA65iNAOuYjgDrmI8A65iQAOuYkQDrmJIA65iTAOuYlADrmJUA65iWAOuYlwDrmJgA65iZAOuYmgDrmJsA65icAOuYnQDrmJ4A65ifAOuYoADrmKEA65iiAOuYowDrmKQA65ilAOuYpgDrmKcA65ioAOuYqQDrmKoA65irAOuYrADrmK0A65iuAOuYrwDrmLAA65ixAOuYsgDrmLMA65i0AOuYtQDrmLYA65i3AOuYuADrmLkA65i6AOuYuwDrmLwA65i9AOuYvgDrmL8A65mAAOuZgQDrmYIA65mDAOuZhADrmYUA65mGAOuZhwDrmYgA65mJAOuZigDrmYsA65mMAOuZjQDrmY4A65mPAOuZkADrmZEA65mSAOuZkwDrmZQA65mVAOuZlgDrmZcA65mYAOuZmQDrmZoA65mbAOuZnADrmZ0A65meAOuZnwDrmaAA65mhAOuZogDrmaMA65mkAOuZpQDrmaYA65mnAOuZqADrmakA65mqAOuZqwDrmawA65mtAOuZrgDrma8A65mwAOuZsQDrmbIA65mzAOuZtADrmbUA65m2AOuZtwDrmbgA65m5AOuZugDrmbsA65m8AOuZvQDrmb4A65m/AOuagADrmoEA65qCAOuagwDrmoQA65qFAOuahgDrmocA65qIAOuaiQDrmooA65qLAOuajADrmo0A65qOAOuajwDrmpAA65qRAOuakgDrmpMA65qUAOualQDrmpYA65qXAOuamADrmpkA65qaAOuamwDrmpwA65qdAOuangDrmp8A65qgAOuaoQDrmqIA65qjAOuapADrmqUA65qmAOuapwDrmqgA65qpAOuaqgDrmqsA65qsAOuarQDrmq4A65qvAOuasADrmrEA65qyAOuaswDrmrQA65q1AOuatgDrmrcA65q4AOuauQDrmroA65q7AOuavADrmr0A65q+AOuavwDrm4AA65uBAOubggDrm4MA65uEAOubhQDrm4YA65uHAOubiADrm4kA65uKAOubiwDrm4wA65uNAOubjgDrm48A65uQAOubkQDrm5IA65uTAOublADrm5UA65uWAOublwDrm5gA65uZAOubmgDrm5sA65ucAOubnQDrm54A65ufAOuboADrm6EA65uiAOubowDrm6QA65ulAOubpgDrm6cA65uoAOubqQDrm6oA65urAOubrADrm60A65uuAOubrwDrm7AA65uxAOubsgDrm7MA65u0AOubtQDrm7YA65u3AOubuADrm7kA65u6AOubuwDrm7wA65u9AOubvgDrm78A65yAAOucgQDrnIIA65yDAOuchADrnIUA65yGAOuchwDrnIgA65yJAOucigDrnIsA65yMAOucjQDrnI4A65yPAOuckADrnJEA65ySAOuckwDrnJQA65yVAOuclgDrnJcA65yYAOucmQDrnJoA65ybAOucnADrnJ0A65yeAOucnwDrnKAA65yhAOucogDrnKMA65ykAOucpQDrnKYA65ynAOucqADrnKkA65yqAOucqwDrnKwA65ytAOucrgDrnK8A65ywAOucsQDrnLIA65yzAOuctADrnLUA65y2AOuctwDrnLgA65y5AOucugDrnLsA65y8AOucvQDrnL4A65y/AOudgADrnYEA652CAOudgwDrnYQA652FAOudhgDrnYcA652IAOudiQDrnYoA652LAOudjADrnY0A652OAOudjwDrnZAA652RAOudkgDrnZMA652UAOudlQDrnZYA652XAOudmADrnZkA652aAOudmwDrnZwA652dAOudngDrnZ8A652gAOudoQDrnaIA652jAOudpADrnaUA652mAOudpwDrnagA652pAOudqgDrnasA652sAOudrQDrna4A652vAOudsADrnbEA652yAOudswDrnbQA6521AOudtgDrnbcA6524AOuduQDrnboA6527AOudvADrnb0A652+AOudvwDrnoAA656BAOueggDrnoMA656EAOuehQDrnoYA656HAOueiADrnokA656KAOueiwDrnowA656NAOuejgDrno8A656QAOuekQDrnpIA656TAOuelADrnpUA656WAOuelwDrnpgA656ZAOuemgDrnpsA656cAOuenQDrnp4A656fAOueoADrnqEA656iAOueowDrnqQA656lAOuepgDrnqcA656oAOueqQDrnqoA656rAOuerADrnq0A656uAOuerwDrnrAA656xAOuesgDrnrMA6560AOuetQDrnrYA6563AOueuADrnrkA6566AOueuwDrnrwA6569AOuevgDrnr8A65+AAOufgQDrn4IA65+DAOufhADrn4UA65+GAOufhwDrn4gA65+JAOufigDrn4sA65+MAOufjQDrn44A65+PAOufkADrn5EA65+SAOufkwDrn5QA65+VAOuflgDrn5cA65+YAOufmQDrn5oA65+bAOufnADrn50A65+eAOufnwDrn6AA65+hAOufogDrn6MA65+kAOufpQDrn6YA65+nAOufqADrn6kA65+qAOufqwDrn6wA65+tAOufrgDrn68A65+wAOufsQDrn7IA65+zAOuftADrn7UA65+2AOuftwDrn7gA65+5AOufugDrn7sA65+8AOufvQDrn74A65+/AOuggADroIEA66CCAOuggwDroIQA66CFAOughgDroIcA66CIAOugiQDroIoA66CLAOugjADroI0A66COAOugjwDroJAA66CRAOugkgDroJMA66CUAOuglQDroJYA66CXAOugmADroJkA66CaAOugmwDroJwA66CdAOugngDroJ8A66CgAOugoQDroKIA66CjAOugpADroKUA66CmAOugpwDroKgA66CpAOugqgDroKsA66CsAOugrQDroK4A66CvAOugsADroLEA66CyAOugswDroLQA66C1AOugtgDroLcA66C4AOuguQDroLoA66C7AOugvADroL0A66C+AOugvwDroYAA66GBAOuhggDroYMA66GEAOuhhQDroYYA66GHAOuhiADroYkA66GKAOuhiwDroYwA66GNAOuhjgDroY8A66GQAOuhkQDroZIA66GTAOuhlADroZUA66GWAOuhlwDroZgA66GZAOuhmgDroZsA66GcAOuhnQDroZ4A66GfAOuhoADroaEA66GiAOuhowDroaQA66GlAOuhpgDroacA66GoAOuhqQDroaoA66GrAOuhrADroa0A66GuAOuhrwDrobAA66GxAOuhsgDrobMA66G0AOuhtQDrobYA66G3AOuhuADrobkA66G6AOuhuwDrobwA66G9AOuhvgDrob8A66KAAOuigQDrooIA66KDAOuihADrooUA66KGAOuihwDroogA66KJAOuiigDroosA66KMAOuijQDroo4A66KPAOuikADropEA66KSAOuikwDropQA66KVAOuilgDropcA66KYAOuimQDropoA66KbAOuinADrop0A66KeAOuinwDroqAA66KhAOuiogDroqMA66KkAOuipQDroqYA66KnAOuiqADroqkA66KqAOuiqwDroqwA66KtAOuirgDroq8A66KwAOuisQDrorIA66KzAOuitADrorUA66K2AOuitwDrorgA66K5AOuiugDrorsA66K8AOuivQDror4A66K/AOujgADro4EA66OCAOujgwDro4QA66OFAOujhgDro4cA66OIAOujiQDro4oA66OLAOujjADro40A66OOAOujjwDro5AA66ORAOujkgDro5MA66OUAOujlQDro5YA66OXAOujmADro5kA66OaAOujmwDro5wA66OdAOujngDro58A66OgAOujoQDro6IA66OjAOujpADro6UA66OmAOujpwDro6gA66OpAOujqgDro6sA66OsAOujrQDro64A66OvAOujsADro7EA66OyAOujswDro7QA66O1AOujtgDro7cA66O4AOujuQDro7oA66O7AOujvADro70A66O+AOujvwDrpIAA66SBAOukggDrpIMA66SEAOukhQDrpIYA66SHAOukiADrpIkA66SKAOukiwDrpIwA66SNAOukjgDrpI8A66SQAOukkQDrpJIA66STAOuklADrpJUA66SWAOuklwDrpJgA66SZAOukmgDrpJsA66ScAOuknQDrpJ4A66SfAOukoADrpKEA66SiAOukowDrpKQA66SlAOukpgDrpKcA66SoAOukqQDrpKoA66SrAOukrADrpK0A66SuAOukrwDrpLAA66SxAOuksgDrpLMA66S0AOuktQDrpLYA66S3AOukuADrpLkA66S6AOukuwDrpLwA66S9AOukvgDrpL8A66WAAOulgQDrpYIA66WDAOulhADrpYUA66WGAOulhwDrpYgA66WJAOuligDrpYsA66WMAOuljQDrpY4A66WPAOulkADrpZEA66WSAOulkwDrpZQA66WVAOullgDrpZcA66WYAOulmQDrpZoA66WbAOulnADrpZ0A66WeAOulnwDrpaAA66WhAOulogDrpaMA66WkAOulpQDrpaYA66WnAOulqADrpakA66WqAOulqwDrpawA66WtAOulrgDrpa8A66WwAOulsQDrpbIA66WzAOultADrpbUA66W2AOultwDrpbgA66W5AOulugDrpbsA66W8AOulvQDrpb4A66W/AOumgADrpoEA66aCAOumgwDrpoQA66aFAOumhgDrpocA66aIAOumiQDrpooA66aLAOumjADrpo0A66aOAOumjwDrppAA66aRAOumkgDrppMA66aUAOumlQDrppYA66aXAOummADrppkA66aaAOummwDrppwA66adAOumngDrpp8A66agAOumoQDrpqIA66ajAOumpADrpqUA66amAOumpwDrpqgA66apAOumqgDrpqsA66asAOumrQDrpq4A66avAOumsADrprEA66ayAOumswDrprQA66a1AOumtgDrprcA66a4AOumuQDrproA66a7AOumvADrpr0A66a+AOumvwDrp4AA66eBAOunggDrp4MA66eEAOunhQDrp4YA66eHAOuniADrp4kA66eKAOuniwDrp4wA66eNAOunjgDrp48A66eQAOunkQDrp5IA66eTAOunlADrp5UA66eWAOunlwDrp5gA66eZAOunmgDrp5sA66ecAOunnQDrp54A66efAOunoADrp6EA66eiAOunowDrp6QA66elAOunpgDrp6cA66eoAOunqQDrp6oA66erAOunrADrp60A66euAOunrwDrp7AA66exAOunsgDrp7MA66e0AOuntQDrp7YA66e3AOunuADrp7kA66e6AOunuwDrp7wA66e9AOunvgDrp78A66iAAOuogQDrqIIA66iDAOuohADrqIUA66iGAOuohwDrqIgA66iJAOuoigDrqIsA66iMAOuojQDrqI4A66iPAOuokADrqJEA66iSAOuokwDrqJQA66iVAOuolgDrqJcA66iYAOuomQDrqJoA66ibAOuonADrqJ0A66ieAOuonwDrqKAA66ihAOuoogDrqKMA66ikAOuopQDrqKYA66inAOuoqADrqKkA66iqAOuoqwDrqKwA66itAOuorgDrqK8A66iwAOuosQDrqLIA66izAOuotADrqLUA66i2AOuotwDrqLgA66i5AOuougDrqLsA66i8AOuovQDrqL4A66i/AOupgADrqYEA66mCAOupgwDrqYQA66mFAOuphgDrqYcA66mIAOupiQDrqYoA66mLAOupjADrqY0A66mOAOupjwDrqZAA66mRAOupkgDrqZMA66mUAOuplQDrqZYA66mXAOupmADrqZkA66maAOupmwDrqZwA66mdAOupngDrqZ8A66mgAOupoQDrqaIA66mjAOuppADrqaUA66mmAOuppwDrqagA66mpAOupqgDrqasA66msAOuprQDrqa4A66mvAOupsADrqbEA66myAOupswDrqbQA66m1AOuptgDrqbcA66m4AOupuQDrqboA66m7AOupvADrqb0A66m+AOupvwDrqoAA66qBAOuqggDrqoMA66qEAOuqhQDrqoYA66qHAOuqiADrqokA66qKAOuqiwDrqowA66qNAOuqjgDrqo8A66qQAOuqkQDrqpIA66qTAOuqlADrqpUA66qWAOuqlwDrqpgA66qZAOuqmgDrqpsA66qcAOuqnQDrqp4A66qfAOuqoADrqqEA66qiAOuqowDrqqQA66qlAOuqpgDrqqcA66qoAOuqqQDrqqoA66qrAOuqrADrqq0A66quAOuqrwDrqrAA66qxAOuqsgDrqrMA66q0AOuqtQDrqrYA66q3AOuquADrqrkA66q6AOuquwDrqrwA66q9AOuqvgDrqr8A66uAAOurgQDrq4IA66uDAOurhADrq4UA66uGAOurhwDrq4gA66uJAOurigDrq4sA66uMAOurjQDrq44A66uPAOurkADrq5EA66uSAOurkwDrq5QA66uVAOurlgDrq5cA66uYAOurmQDrq5oA66ubAOurnADrq50A66ueAOurnwDrq6AA66uhAOurogDrq6MA66ukAOurpQDrq6YA66unAOurqADrq6kA66uqAOurqwDrq6wA66utAOurrgDrq68A66uwAOursQDrq7IA66uzAOurtADrq7UA66u2AOurtwDrq7gA66u5AOurugDrq7sA66u8AOurvQDrq74A66u/AOusgADrrIEA66yCAOusgwDrrIQA66yFAOushgDrrIcA66yIAOusiQDrrIoA66yLAOusjADrrI0A66yOAOusjwDrrJAA66yRAOuskgDrrJMA66yUAOuslQDrrJYA66yXAOusmADrrJkA66yaAOusmwDrrJwA66ydAOusngDrrJ8A66ygAOusoQDrrKIA66yjAOuspADrrKUA66ymAOuspwDrrKgA66ypAOusqgDrrKsA66ysAOusrQDrrK4A66yvAOussADrrLEA66yyAOusswDrrLQA66y1AOustgDrrLcA66y4AOusuQDrrLoA66y7AOusvADrrL0A66y+AOusvwDrrYAA662BAOutggDrrYMA662EAOuthQDrrYYA662HAOutiADrrYkA662KAOutiwDrrYwA662NAOutjgDrrY8A662QAOutkQDrrZIA662TAOutlADrrZUA662WAOutlwDrrZgA662ZAOutmgDrrZsA662cAOutnQDrrZ4A662fAOutoADrraEA662iAOutowDrraQA662lAOutpgDrracA662oAOutqQDrraoA662rAOutrADrra0A662uAOutrwDrrbAA662xAOutsgDrrbMA6620AOuttQDrrbYA6623AOutuADrrbkA6626AOutuwDrrbwA6629AOutvgDrrb8A666AAOuugQDrroIA666DAOuuhADrroUA666GAOuuhwDrrogA666JAOuuigDrrosA666MAOuujQDrro4A666PAOuukADrrpEA666SAOuukwDrrpQA666VAOuulgDrrpcA666YAOuumQDrrpoA666bAOuunADrrp0A666eAOuunwDrrqAA666hAOuuogDrrqMA666kAOuupQDrrqYA666nAOuuqADrrqkA666qAOuuqwDrrqwA666tAOuurgDrrq8A666wAOuusQDrrrIA666zAOuutADrrrUA6662AOuutwDrrrgA6665AOuuugDrrrsA6668AOuuvQDrrr4A666/AOuvgADrr4EA66+CAOuvgwDrr4QA66+FAOuvhgDrr4cA66+IAOuviQDrr4oA66+LAOuvjADrr40A66+OAOuvjwDrr5AA66+RAOuvkgDrr5MA66+UAOuvlQDrr5YA66+XAOuvmADrr5kA66+aAOuvmwDrr5wA66+dAOuvngDrr58A66+gAOuvoQDrr6IA66+jAOuvpADrr6UA66+mAOuvpwDrr6gA66+pAOuvqgDrr6sA66+sAOuvrQDrr64A66+vAOuvsADrr7EA66+yAOuvswDrr7QA66+1AOuvtgDrr7cA66+4AOuvuQDrr7oA66+7AOuvvADrr70A66++AOuvvwDrsIAA67CBAOuwggDrsIMA67CEAOuwhQDrsIYA67CHAOuwiADrsIkA67CKAOuwiwDrsIwA67CNAOuwjgDrsI8A67CQAOuwkQDrsJIA67CTAOuwlADrsJUA67CWAOuwlwDrsJgA67CZAOuwmgDrsJsA67CcAOuwnQDrsJ4A67CfAOuwoADrsKEA67CiAOuwowDrsKQA67ClAOuwpgDrsKcA67CoAOuwqQDrsKoA67CrAOuwrADrsK0A67CuAOuwrwDrsLAA67CxAOuwsgDrsLMA67C0AOuwtQDrsLYA67C3AOuwuADrsLkA67C6AOuwuwDrsLwA67C9AOuwvgDrsL8A67GAAOuxgQDrsYIA67GDAOuxhADrsYUA67GGAOuxhwDrsYgA67GJAOuxigDrsYsA67GMAOuxjQDrsY4A67GPAOuxkADrsZEA67GSAOuxkwDrsZQA67GVAOuxlgDrsZcA67GYAOuxmQDrsZoA67GbAOuxnADrsZ0A67GeAOuxnwDrsaAA67GhAOuxogDrsaMA67GkAOuxpQDrsaYA67GnAOuxqADrsakA67GqAOuxqwDrsawA67GtAOuxrgDrsa8A67GwAOuxsQDrsbIA67GzAOuxtADrsbUA67G2AOuxtwDrsbgA67G5AOuxugDrsbsA67G8AOuxvQDrsb4A67G/AOuygADrsoEA67KCAOuygwDrsoQA67KFAOuyhgDrsocA67KIAOuyiQDrsooA67KLAOuyjADrso0A67KOAOuyjwDrspAA67KRAOuykgDrspMA67KUAOuylQDrspYA67KXAOuymADrspkA67KaAOuymwDrspwA67KdAOuyngDrsp8A67KgAOuyoQDrsqIA67KjAOuypADrsqUA67KmAOuypwDrsqgA67KpAOuyqgDrsqsA67KsAOuyrQDrsq4A67KvAOuysADrsrEA67KyAOuyswDrsrQA67K1AOuytgDrsrcA67K4AOuyuQDrsroA67K7AOuyvADrsr0A67K+AOuyvwDrs4AA67OBAOuzggDrs4MA67OEAOuzhQDrs4YA67OHAOuziADrs4kA67OKAOuziwDrs4wA67ONAOuzjgDrs48A67OQAOuzkQDrs5IA67OTAOuzlADrs5UA67OWAOuzlwDrs5gA67OZAOuzmgDrs5sA67OcAOuznQDrs54A67OfAOuzoADrs6EA67OiAOuzowDrs6QA67OlAOuzpgDrs6cA67OoAOuzqQDrs6oA67OrAOuzrADrs60A67OuAOuzrwDrs7AA67OxAOuzsgDrs7MA67O0AOuztQDrs7YA67O3AOuzuADrs7kA67O6AOuzuwDrs7wA67O9AOuzvgDrs78A67SAAOu0gQDrtIIA67SDAOu0hADrtIUA67SGAOu0hwDrtIgA67SJAOu0igDrtIsA67SMAOu0jQDrtI4A67SPAOu0kADrtJEA67SSAOu0kwDrtJQA67SVAOu0lgDrtJcA67SYAOu0mQDrtJoA67SbAOu0nADrtJ0A67SeAOu0nwDrtKAA67ShAOu0ogDrtKMA67SkAOu0pQDrtKYA67SnAOu0qADrtKkA67SqAOu0qwDrtKwA67StAOu0rgDrtK8A67SwAOu0sQDrtLIA67SzAOu0tADrtLUA67S2AOu0twDrtLgA67S5AOu0ugDrtLsA67S8AOu0vQDrtL4A67S/AOu1gADrtYEA67WCAOu1gwDrtYQA67WFAOu1hgDrtYcA67WIAOu1iQDrtYoA67WLAOu1jADrtY0A67WOAOu1jwDrtZAA67WRAOu1kgDrtZMA67WUAOu1lQDrtZYA67WXAOu1mADrtZkA67WaAOu1mwDrtZwA67WdAOu1ngDrtZ8A67WgAOu1oQDrtaIA67WjAOu1pADrtaUA67WmAOu1pwDrtagA67WpAOu1qgDrtasA67WsAOu1rQDrta4A67WvAOu1sADrtbEA67WyAOu1swDrtbQA67W1AOu1tgDrtbcA67W4AOu1uQDrtboA67W7AOu1vADrtb0A67W+AOu1vwDrtoAA67aBAOu2ggDrtoMA67aEAOu2hQDrtoYA67aHAOu2iADrtokA67aKAOu2iwDrtowA67aNAOu2jgDrto8A67aQAOu2kQDrtpIA67aTAOu2lADrtpUA67aWAOu2lwDrtpgA67aZAOu2mgDrtpsA67acAOu2nQDrtp4A67afAOu2oADrtqEA67aiAOu2owDrtqQA67alAOu2pgDrtqcA67aoAOu2qQDrtqoA67arAOu2rADrtq0A67auAOu2rwDrtrAA67axAOu2sgDrtrMA67a0AOu2tQDrtrYA67a3AOu2uADrtrkA67a6AOu2uwDrtrwA67a9AOu2vgDrtr8A67eAAOu3gQDrt4IA67eDAOu3hADrt4UA67eGAOu3hwDrt4gA67eJAOu3igDrt4sA67eMAOu3jQDrt44A67ePAOu3kADrt5EA67eSAOu3kwDrt5QA67eVAOu3lgDrt5cA67eYAOu3mQDrt5oA67ebAOu3nADrt50A67eeAOu3nwDrt6AA67ehAOu3ogDrt6MA67ekAOu3pQDrt6YA67enAOu3qADrt6kA67eqAOu3qwDrt6wA67etAOu3rgDrt68A67ewAOu3sQDrt7IA67ezAOu3tADrt7UA67e2AOu3twDrt7gA67e5AOu3ugDrt7sA67e8AOu3vQDrt74A67e/AOu4gADruIEA67iCAOu4gwDruIQA67iFAOu4hgDruIcA67iIAOu4iQDruIoA67iLAOu4jADruI0A67iOAOu4jwDruJAA67iRAOu4kgDruJMA67iUAOu4lQDruJYA67iXAOu4mADruJkA67iaAOu4mwDruJwA67idAOu4ngDruJ8A67igAOu4oQDruKIA67ijAOu4pADruKUA67imAOu4pwDruKgA67ipAOu4qgDruKsA67isAOu4rQDruK4A67ivAOu4sADruLEA67iyAOu4swDruLQA67i1AOu4tgDruLcA67i4AOu4uQDruLoA67i7AOu4vADruL0A67i+AOu4vwDruYAA67mBAOu5ggDruYMA67mEAOu5hQDruYYA67mHAOu5iADruYkA67mKAOu5iwDruYwA67mNAOu5jgDruY8A67mQAOu5kQDruZIA67mTAOu5lADruZUA67mWAOu5lwDruZgA67mZAOu5mgDruZsA67mcAOu5nQDruZ4A67mfAOu5oADruaEA67miAOu5owDruaQA67mlAOu5pgDruacA67moAOu5qQDruaoA67mrAOu5rADrua0A67muAOu5rwDrubAA67mxAOu5sgDrubMA67m0AOu5tQDrubYA67m3AOu5uADrubkA67m6AOu5uwDrubwA67m9AOu5vgDrub8A67qAAOu6gQDruoIA67qDAOu6hADruoUA67qGAOu6hwDruogA67qJAOu6igDruosA67qMAOu6jQDruo4A67qPAOu6kADrupEA67qSAOu6kwDrupQA67qVAOu6lgDrupcA67qYAOu6mQDrupoA67qbAOu6nADrup0A67qeAOu6nwDruqAA67qhAOu6ogDruqMA67qkAOu6pQDruqYA67qnAOu6qADruqkA67qqAOu6qwDruqwA67qtAOu6rgDruq8A67qwAOu6sQDrurIA67qzAOu6tADrurUA67q2AOu6twDrurgA67q5AOu6ugDrursA67q8AOu6vQDrur4A67q/AOu7gADru4EA67uCAOu7gwDru4QA67uFAOu7hgDru4cA67uIAOu7iQDru4oA67uLAOu7jADru40A67uOAOu7jwDru5AA67uRAOu7kgDru5MA67uUAOu7lQDru5YA67uXAOu7mADru5kA67uaAOu7mwDru5wA67udAOu7ngDru58A67ugAOu7oQDru6IA67ujAOu7pADru6UA67umAOu7pwDru6gA67upAOu7qgDru6sA67usAOu7rQDru64A67uvAOu7sADru7EA67uyAOu7swDru7QA67u1AOu7tgDru7cA67u4AOu7uQDru7oA67u7AOu7vADru70A67u+AOu7vwDrvIAA67yBAOu8ggDrvIMA67yEAOu8hQDrvIYA67yHAOu8iADrvIkA67yKAOu8iwDrvIwA67yNAOu8jgDrvI8A67yQAOu8kQDrvJIA67yTAOu8lADrvJUA67yWAOu8lwDrvJgA67yZAOu8mgDrvJsA67ycAOu8nQDrvJ4A67yfAOu8oADrvKEA67yiAOu8owDrvKQA67ylAOu8pgDrvKcA67yoAOu8qQDrvKoA67yrAOu8rADrvK0A67yuAOu8rwDrvLAA67yxAOu8sgDrvLMA67y0AOu8tQDrvLYA67y3AOu8uADrvLkA67y6AOu8uwDrvLwA67y9AOu8vgDrvL8A672AAOu9gQDrvYIA672DAOu9hADrvYUA672GAOu9hwDrvYgA672JAOu9igDrvYsA672MAOu9jQDrvY4A672PAOu9kADrvZEA672SAOu9kwDrvZQA672VAOu9lgDrvZcA672YAOu9mQDrvZoA672bAOu9nADrvZ0A672eAOu9nwDrvaAA672hAOu9ogDrvaMA672kAOu9pQDrvaYA672nAOu9qADrvakA672qAOu9qwDrvawA672tAOu9rgDrva8A672wAOu9sQDrvbIA672zAOu9tADrvbUA6722AOu9twDrvbgA6725AOu9ugDrvbsA6728AOu9vQDrvb4A672/AOu+gADrvoEA676CAOu+gwDrvoQA676FAOu+hgDrvocA676IAOu+iQDrvooA676LAOu+jADrvo0A676OAOu+jwDrvpAA676RAOu+kgDrvpMA676UAOu+lQDrvpYA676XAOu+mADrvpkA676aAOu+mwDrvpwA676dAOu+ngDrvp8A676gAOu+oQDrvqIA676jAOu+pADrvqUA676mAOu+pwDrvqgA676pAOu+qgDrvqsA676sAOu+rQDrvq4A676vAOu+sADrvrEA676yAOu+swDrvrQA6761AOu+tgDrvrcA6764AOu+uQDrvroA6767AOu+vADrvr0A676+AOu+vwDrv4AA67+BAOu/ggDrv4MA67+EAOu/hQDrv4YA67+HAOu/iADrv4kA67+KAOu/iwDrv4wA67+NAOu/jgDrv48A67+QAOu/kQDrv5IA67+TAOu/lADrv5UA67+WAOu/lwDrv5gA67+ZAOu/mgDrv5sA67+cAOu/nQDrv54A67+fAOu/oADrv6EA67+iAOu/owDrv6QA67+lAOu/pgDrv6cA67+oAOu/qQDrv6oA67+rAOu/rADrv60A67+uAOu/rwDrv7AA67+xAOu/sgDrv7MA67+0AOu/tQDrv7YA67+3AOu/uADrv7kA67+6AOu/uwDrv7wA67+9AOu/vgDrv78A7ICAAOyAgQDsgIIA7ICDAOyAhADsgIUA7ICGAOyAhwDsgIgA7ICJAOyAigDsgIsA7ICMAOyAjQDsgI4A7ICPAOyAkADsgJEA7ICSAOyAkwDsgJQA7ICVAOyAlgDsgJcA7ICYAOyAmQDsgJoA7ICbAOyAnADsgJ0A7ICeAOyAnwDsgKAA7IChAOyAogDsgKMA7ICkAOyApQDsgKYA7ICnAOyAqADsgKkA7ICqAOyAqwDsgKwA7ICtAOyArgDsgK8A7ICwAOyAsQDsgLIA7ICzAOyAtADsgLUA7IC2AOyAtwDsgLgA7IC5AOyAugDsgLsA7IC8AOyAvQDsgL4A7IC/AOyBgADsgYEA7IGCAOyBgwDsgYQA7IGFAOyBhgDsgYcA7IGIAOyBiQDsgYoA7IGLAOyBjADsgY0A7IGOAOyBjwDsgZAA7IGRAOyBkgDsgZMA7IGUAOyBlQDsgZYA7IGXAOyBmADsgZkA7IGaAOyBmwDsgZwA7IGdAOyBngDsgZ8A7IGgAOyBoQDsgaIA7IGjAOyBpADsgaUA7IGmAOyBpwDsgagA7IGpAOyBqgDsgasA7IGsAOyBrQDsga4A7IGvAOyBsADsgbEA7IGyAOyBswDsgbQA7IG1AOyBtgDsgbcA7IG4AOyBuQDsgboA7IG7AOyBvADsgb0A7IG+AOyBvwDsgoAA7IKBAOyCggDsgoMA7IKEAOyChQDsgoYA7IKHAOyCiADsgokA7IKKAOyCiwDsgowA7IKNAOyCjgDsgo8A7IKQAOyCkQDsgpIA7IKTAOyClADsgpUA7IKWAOyClwDsgpgA7IKZAOyCmgDsgpsA7IKcAOyCnQDsgp4A7IKfAOyCoADsgqEA7IKiAOyCowDsgqQA7IKlAOyCpgDsgqcA7IKoAOyCqQDsgqoA7IKrAOyCrADsgq0A7IKuAOyCrwDsgrAA7IKxAOyCsgDsgrMA7IK0AOyCtQDsgrYA7IK3AOyCuADsgrkA7IK6AOyCuwDsgrwA7IK9AOyCvgDsgr8A7IOAAOyDgQDsg4IA7IODAOyDhADsg4UA7IOGAOyDhwDsg4gA7IOJAOyDigDsg4sA7IOMAOyDjQDsg44A7IOPAOyDkADsg5EA7IOSAOyDkwDsg5QA7IOVAOyDlgDsg5cA7IOYAOyDmQDsg5oA7IObAOyDnADsg50A7IOeAOyDnwDsg6AA7IOhAOyDogDsg6MA7IOkAOyDpQDsg6YA7IOnAOyDqADsg6kA7IOqAOyDqwDsg6wA7IOtAOyDrgDsg68A7IOwAOyDsQDsg7IA7IOzAOyDtADsg7UA7IO2AOyDtwDsg7gA7IO5AOyDugDsg7sA7IO8AOyDvQDsg74A7IO/AOyEgADshIEA7ISCAOyEgwDshIQA7ISFAOyEhgDshIcA7ISIAOyEiQDshIoA7ISLAOyEjADshI0A7ISOAOyEjwDshJAA7ISRAOyEkgDshJMA7ISUAOyElQDshJYA7ISXAOyEmADshJkA7ISaAOyEmwDshJwA7ISdAOyEngDshJ8A7ISgAOyEoQDshKIA7ISjAOyEpADshKUA7ISmAOyEpwDshKgA7ISpAOyEqgDshKsA7ISsAOyErQDshK4A7ISvAOyEsADshLEA7ISyAOyEswDshLQA7IS1AOyEtgDshLcA7IS4AOyEuQDshLoA7IS7AOyEvADshL0A7IS+AOyEvwDshYAA7IWBAOyFggDshYMA7IWEAOyFhQDshYYA7IWHAOyFiADshYkA7IWKAOyFiwDshYwA7IWNAOyFjgDshY8A7IWQAOyFkQDshZIA7IWTAOyFlADshZUA7IWWAOyFlwDshZgA7IWZAOyFmgDshZsA7IWcAOyFnQDshZ4A7IWfAOyFoADshaEA7IWiAOyFowDshaQA7IWlAOyFpgDshacA7IWoAOyFqQDshaoA7IWrAOyFrADsha0A7IWuAOyFrwDshbAA7IWxAOyFsgDshbMA7IW0AOyFtQDshbYA7IW3AOyFuADshbkA7IW6AOyFuwDshbwA7IW9AOyFvgDshb8A7IaAAOyGgQDshoIA7IaDAOyGhADshoUA7IaGAOyGhwDshogA7IaJAOyGigDshosA7IaMAOyGjQDsho4A7IaPAOyGkADshpEA7IaSAOyGkwDshpQA7IaVAOyGlgDshpcA7IaYAOyGmQDshpoA7IabAOyGnADshp0A7IaeAOyGnwDshqAA7IahAOyGogDshqMA7IakAOyGpQDshqYA7IanAOyGqADshqkA7IaqAOyGqwDshqwA7IatAOyGrgDshq8A7IawAOyGsQDshrIA7IazAOyGtADshrUA7Ia2AOyGtwDshrgA7Ia5AOyGugDshrsA7Ia8AOyGvQDshr4A7Ia/AOyHgADsh4EA7IeCAOyHgwDsh4QA7IeFAOyHhgDsh4cA7IeIAOyHiQDsh4oA7IeLAOyHjADsh40A7IeOAOyHjwDsh5AA7IeRAOyHkgDsh5MA7IeUAOyHlQDsh5YA7IeXAOyHmADsh5kA7IeaAOyHmwDsh5wA7IedAOyHngDsh58A7IegAOyHoQDsh6IA7IejAOyHpADsh6UA7IemAOyHpwDsh6gA7IepAOyHqgDsh6sA7IesAOyHrQDsh64A7IevAOyHsADsh7EA7IeyAOyHswDsh7QA7Ie1AOyHtgDsh7cA7Ie4AOyHuQDsh7oA7Ie7AOyHvADsh70A7Ie+AOyHvwDsiIAA7IiBAOyIggDsiIMA7IiEAOyIhQDsiIYA7IiHAOyIiADsiIkA7IiKAOyIiwDsiIwA7IiNAOyIjgDsiI8A7IiQAOyIkQDsiJIA7IiTAOyIlADsiJUA7IiWAOyIlwDsiJgA7IiZAOyImgDsiJsA7IicAOyInQDsiJ4A7IifAOyIoADsiKEA7IiiAOyIowDsiKQA7IilAOyIpgDsiKcA7IioAOyIqQDsiKoA7IirAOyIrADsiK0A7IiuAOyIrwDsiLAA7IixAOyIsgDsiLMA7Ii0AOyItQDsiLYA7Ii3AOyIuADsiLkA7Ii6AOyIuwDsiLwA7Ii9AOyIvgDsiL8A7ImAAOyJgQDsiYIA7ImDAOyJhADsiYUA7ImGAOyJhwDsiYgA7ImJAOyJigDsiYsA7ImMAOyJjQDsiY4A7ImPAOyJkADsiZEA7ImSAOyJkwDsiZQA7ImVAOyJlgDsiZcA7ImYAOyJmQDsiZoA7ImbAOyJnADsiZ0A7ImeAOyJnwDsiaAA7ImhAOyJogDsiaMA7ImkAOyJpQDsiaYA7ImnAOyJqADsiakA7ImqAOyJqwDsiawA7ImtAOyJrgDsia8A7ImwAOyJsQDsibIA7ImzAOyJtADsibUA7Im2AOyJtwDsibgA7Im5AOyJugDsibsA7Im8AOyJvQDsib4A7Im/AOyKgADsioEA7IqCAOyKgwDsioQA7IqFAOyKhgDsiocA7IqIAOyKiQDsiooA7IqLAOyKjADsio0A7IqOAOyKjwDsipAA7IqRAOyKkgDsipMA7IqUAOyKlQDsipYA7IqXAOyKmADsipkA7IqaAOyKmwDsipwA7IqdAOyKngDsip8A7IqgAOyKoQDsiqIA7IqjAOyKpADsiqUA7IqmAOyKpwDsiqgA7IqpAOyKqgDsiqsA7IqsAOyKrQDsiq4A7IqvAOyKsADsirEA7IqyAOyKswDsirQA7Iq1AOyKtgDsircA7Iq4AOyKuQDsiroA7Iq7AOyKvADsir0A7Iq+AOyKvwDsi4AA7IuBAOyLggDsi4MA7IuEAOyLhQDsi4YA7IuHAOyLiADsi4kA7IuKAOyLiwDsi4wA7IuNAOyLjgDsi48A7IuQAOyLkQDsi5IA7IuTAOyLlADsi5UA7IuWAOyLlwDsi5gA7IuZAOyLmgDsi5sA7IucAOyLnQDsi54A7IufAOyLoADsi6EA7IuiAOyLowDsi6QA7IulAOyLpgDsi6cA7IuoAOyLqQDsi6oA7IurAOyLrADsi60A7IuuAOyLrwDsi7AA7IuxAOyLsgDsi7MA7Iu0AOyLtQDsi7YA7Iu3AOyLuADsi7kA7Iu6AOyLuwDsi7wA7Iu9AOyLvgDsi78A7IyAAOyMgQDsjIIA7IyDAOyMhADsjIUA7IyGAOyMhwDsjIgA7IyJAOyMigDsjIsA7IyMAOyMjQDsjI4A7IyPAOyMkADsjJEA7IySAOyMkwDsjJQA7IyVAOyMlgDsjJcA7IyYAOyMmQDsjJoA7IybAOyMnADsjJ0A7IyeAOyMnwDsjKAA7IyhAOyMogDsjKMA7IykAOyMpQDsjKYA7IynAOyMqADsjKkA7IyqAOyMqwDsjKwA7IytAOyMrgDsjK8A7IywAOyMsQDsjLIA7IyzAOyMtADsjLUA7Iy2AOyMtwDsjLgA7Iy5AOyMugDsjLsA7Iy8AOyMvQDsjL4A7Iy/AOyNgADsjYEA7I2CAOyNgwDsjYQA7I2FAOyNhgDsjYcA7I2IAOyNiQDsjYoA7I2LAOyNjADsjY0A7I2OAOyNjwDsjZAA7I2RAOyNkgDsjZMA7I2UAOyNlQDsjZYA7I2XAOyNmADsjZkA7I2aAOyNmwDsjZwA7I2dAOyNngDsjZ8A7I2gAOyNoQDsjaIA7I2jAOyNpADsjaUA7I2mAOyNpwDsjagA7I2pAOyNqgDsjasA7I2sAOyNrQDsja4A7I2vAOyNsADsjbEA7I2yAOyNswDsjbQA7I21AOyNtgDsjbcA7I24AOyNuQDsjboA7I27AOyNvADsjb0A7I2+AOyNvwDsjoAA7I6BAOyOggDsjoMA7I6EAOyOhQDsjoYA7I6HAOyOiADsjokA7I6KAOyOiwDsjowA7I6NAOyOjgDsjo8A7I6QAOyOkQDsjpIA7I6TAOyOlADsjpUA7I6WAOyOlwDsjpgA7I6ZAOyOmgDsjpsA7I6cAOyOnQDsjp4A7I6fAOyOoADsjqEA7I6iAOyOowDsjqQA7I6lAOyOpgDsjqcA7I6oAOyOqQDsjqoA7I6rAOyOrADsjq0A7I6uAOyOrwDsjrAA7I6xAOyOsgDsjrMA7I60AOyOtQDsjrYA7I63AOyOuADsjrkA7I66AOyOuwDsjrwA7I69AOyOvgDsjr8A7I+AAOyPgQDsj4IA7I+DAOyPhADsj4UA7I+GAOyPhwDsj4gA7I+JAOyPigDsj4sA7I+MAOyPjQDsj44A7I+PAOyPkADsj5EA7I+SAOyPkwDsj5QA7I+VAOyPlgDsj5cA7I+YAOyPmQDsj5oA7I+bAOyPnADsj50A7I+eAOyPnwDsj6AA7I+hAOyPogDsj6MA7I+kAOyPpQDsj6YA7I+nAOyPqADsj6kA7I+qAOyPqwDsj6wA7I+tAOyPrgDsj68A7I+wAOyPsQDsj7IA7I+zAOyPtADsj7UA7I+2AOyPtwDsj7gA7I+5AOyPugDsj7sA7I+8AOyPvQDsj74A7I+/AOyQgADskIEA7JCCAOyQgwDskIQA7JCFAOyQhgDskIcA7JCIAOyQiQDskIoA7JCLAOyQjADskI0A7JCOAOyQjwDskJAA7JCRAOyQkgDskJMA7JCUAOyQlQDskJYA7JCXAOyQmADskJkA7JCaAOyQmwDskJwA7JCdAOyQngDskJ8A7JCgAOyQoQDskKIA7JCjAOyQpADskKUA7JCmAOyQpwDskKgA7JCpAOyQqgDskKsA7JCsAOyQrQDskK4A7JCvAOyQsADskLEA7JCyAOyQswDskLQA7JC1AOyQtgDskLcA7JC4AOyQuQDskLoA7JC7AOyQvADskL0A7JC+AOyQvwDskYAA7JGBAOyRggDskYMA7JGEAOyRhQDskYYA7JGHAOyRiADskYkA7JGKAOyRiwDskYwA7JGNAOyRjgDskY8A7JGQAOyRkQDskZIA7JGTAOyRlADskZUA7JGWAOyRlwDskZgA7JGZAOyRmgDskZsA7JGcAOyRnQDskZ4A7JGfAOyRoADskaEA7JGiAOyRowDskaQA7JGlAOyRpgDskacA7JGoAOyRqQDskaoA7JGrAOyRrADska0A7JGuAOyRrwDskbAA7JGxAOyRsgDskbMA7JG0AOyRtQDskbYA7JG3AOyRuADskbkA7JG6AOyRuwDskbwA7JG9AOyRvgDskb8A7JKAAOySgQDskoIA7JKDAOyShADskoUA7JKGAOyShwDskogA7JKJAOySigDskosA7JKMAOySjQDsko4A7JKPAOySkADskpEA7JKSAOySkwDskpQA7JKVAOySlgDskpcA7JKYAOySmQDskpoA7JKbAOySnADskp0A7JKeAOySnwDskqAA7JKhAOySogDskqMA7JKkAOySpQDskqYA7JKnAOySqADskqkA7JKqAOySqwDskqwA7JKtAOySrgDskq8A7JKwAOySsQDskrIA7JKzAOyStADskrUA7JK2AOyStwDskrgA7JK5AOySugDskrsA7JK8AOySvQDskr4A7JK/AOyTgADsk4EA7JOCAOyTgwDsk4QA7JOFAOyThgDsk4cA7JOIAOyTiQDsk4oA7JOLAOyTjADsk40A7JOOAOyTjwDsk5AA7JORAOyTkgDsk5MA7JOUAOyTlQDsk5YA7JOXAOyTmADsk5kA7JOaAOyTmwDsk5wA7JOdAOyTngDsk58A7JOgAOyToQDsk6IA7JOjAOyTpADsk6UA7JOmAOyTpwDsk6gA7JOpAOyTqgDsk6sA7JOsAOyTrQDsk64A7JOvAOyTsADsk7EA7JOyAOyTswDsk7QA7JO1AOyTtgDsk7cA7JO4AOyTuQDsk7oA7JO7AOyTvADsk70A7JO+AOyTvwDslIAA7JSBAOyUggDslIMA7JSEAOyUhQDslIYA7JSHAOyUiADslIkA7JSKAOyUiwDslIwA7JSNAOyUjgDslI8A7JSQAOyUkQDslJIA7JSTAOyUlADslJUA7JSWAOyUlwDslJgA7JSZAOyUmgDslJsA7JScAOyUnQDslJ4A7JSfAOyUoADslKEA7JSiAOyUowDslKQA7JSlAOyUpgDslKcA7JSoAOyUqQDslKoA7JSrAOyUrADslK0A7JSuAOyUrwDslLAA7JSxAOyUsgDslLMA7JS0AOyUtQDslLYA7JS3AOyUuADslLkA7JS6AOyUuwDslLwA7JS9AOyUvgDslL8A7JWAAOyVgQDslYIA7JWDAOyVhADslYUA7JWGAOyVhwDslYgA7JWJAOyVigDslYsA7JWMAOyVjQDslY4A7JWPAOyVkADslZEA7JWSAOyVkwDslZQA7JWVAOyVlgDslZcA7JWYAOyVmQDslZoA7JWbAOyVnADslZ0A7JWeAOyVnwDslaAA7JWhAOyVogDslaMA7JWkAOyVpQDslaYA7JWnAOyVqADslakA7JWqAOyVqwDslawA7JWtAOyVrgDsla8A7JWwAOyVsQDslbIA7JWzAOyVtADslbUA7JW2AOyVtwDslbgA7JW5AOyVugDslbsA7JW8AOyVvQDslb4A7JW/AOyWgADsloEA7JaCAOyWgwDsloQA7JaFAOyWhgDslocA7JaIAOyWiQDslooA7JaLAOyWjADslo0A7JaOAOyWjwDslpAA7JaRAOyWkgDslpMA7JaUAOyWlQDslpYA7JaXAOyWmADslpkA7JaaAOyWmwDslpwA7JadAOyWngDslp8A7JagAOyWoQDslqIA7JajAOyWpADslqUA7JamAOyWpwDslqgA7JapAOyWqgDslqsA7JasAOyWrQDslq4A7JavAOyWsADslrEA7JayAOyWswDslrQA7Ja1AOyWtgDslrcA7Ja4AOyWuQDslroA7Ja7AOyWvADslr0A7Ja+AOyWvwDsl4AA7JeBAOyXggDsl4MA7JeEAOyXhQDsl4YA7JeHAOyXiADsl4kA7JeKAOyXiwDsl4wA7JeNAOyXjgDsl48A7JeQAOyXkQDsl5IA7JeTAOyXlADsl5UA7JeWAOyXlwDsl5gA7JeZAOyXmgDsl5sA7JecAOyXnQDsl54A7JefAOyXoADsl6EA7JeiAOyXowDsl6QA7JelAOyXpgDsl6cA7JeoAOyXqQDsl6oA7JerAOyXrADsl60A7JeuAOyXrwDsl7AA7JexAOyXsgDsl7MA7Je0AOyXtQDsl7YA7Je3AOyXuADsl7kA7Je6AOyXuwDsl7wA7Je9AOyXvgDsl78A7JiAAOyYgQDsmIIA7JiDAOyYhADsmIUA7JiGAOyYhwDsmIgA7JiJAOyYigDsmIsA7JiMAOyYjQDsmI4A7JiPAOyYkADsmJEA7JiSAOyYkwDsmJQA7JiVAOyYlgDsmJcA7JiYAOyYmQDsmJoA7JibAOyYnADsmJ0A7JieAOyYnwDsmKAA7JihAOyYogDsmKMA7JikAOyYpQDsmKYA7JinAOyYqADsmKkA7JiqAOyYqwDsmKwA7JitAOyYrgDsmK8A7JiwAOyYsQDsmLIA7JizAOyYtADsmLUA7Ji2AOyYtwDsmLgA7Ji5AOyYugDsmLsA7Ji8AOyYvQDsmL4A7Ji/AOyZgADsmYEA7JmCAOyZgwDsmYQA7JmFAOyZhgDsmYcA7JmIAOyZiQDsmYoA7JmLAOyZjADsmY0A7JmOAOyZjwDsmZAA7JmRAOyZkgDsmZMA7JmUAOyZlQDsmZYA7JmXAOyZmADsmZkA7JmaAOyZmwDsmZwA7JmdAOyZngDsmZ8A7JmgAOyZoQDsmaIA7JmjAOyZpADsmaUA7JmmAOyZpwDsmagA7JmpAOyZqgDsmasA7JmsAOyZrQDsma4A7JmvAOyZsADsmbEA7JmyAOyZswDsmbQA7Jm1AOyZtgDsmbcA7Jm4AOyZuQDsmboA7Jm7AOyZvADsmb0A7Jm+AOyZvwDsmoAA7JqBAOyaggDsmoMA7JqEAOyahQDsmoYA7JqHAOyaiADsmokA7JqKAOyaiwDsmowA7JqNAOyajgDsmo8A7JqQAOyakQDsmpIA7JqTAOyalADsmpUA7JqWAOyalwDsmpgA7JqZAOyamgDsmpsA7JqcAOyanQDsmp4A7JqfAOyaoADsmqEA7JqiAOyaowDsmqQA7JqlAOyapgDsmqcA7JqoAOyaqQDsmqoA7JqrAOyarADsmq0A7JquAOyarwDsmrAA7JqxAOyasgDsmrMA7Jq0AOyatQDsmrYA7Jq3AOyauADsmrkA7Jq6AOyauwDsmrwA7Jq9AOyavgDsmr8A7JuAAOybgQDsm4IA7JuDAOybhADsm4UA7JuGAOybhwDsm4gA7JuJAOybigDsm4sA7JuMAOybjQDsm44A7JuPAOybkADsm5EA7JuSAOybkwDsm5QA7JuVAOyblgDsm5cA7JuYAOybmQDsm5oA7JubAOybnADsm50A7JueAOybnwDsm6AA7JuhAOybogDsm6MA7JukAOybpQDsm6YA7JunAOybqADsm6kA7JuqAOybqwDsm6wA7JutAOybrgDsm68A7JuwAOybsQDsm7IA7JuzAOybtADsm7UA7Ju2AOybtwDsm7gA7Ju5AOybugDsm7sA7Ju8AOybvQDsm74A7Ju/AOycgADsnIEA7JyCAOycgwDsnIQA7JyFAOychgDsnIcA7JyIAOyciQDsnIoA7JyLAOycjADsnI0A7JyOAOycjwDsnJAA7JyRAOyckgDsnJMA7JyUAOyclQDsnJYA7JyXAOycmADsnJkA7JyaAOycmwDsnJwA7JydAOycngDsnJ8A7JygAOycoQDsnKIA7JyjAOycpADsnKUA7JymAOycpwDsnKgA7JypAOycqgDsnKsA7JysAOycrQDsnK4A7JyvAOycsADsnLEA7JyyAOycswDsnLQA7Jy1AOyctgDsnLcA7Jy4AOycuQDsnLoA7Jy7AOycvADsnL0A7Jy+AOycvwDsnYAA7J2BAOydggDsnYMA7J2EAOydhQDsnYYA7J2HAOydiADsnYkA7J2KAOydiwDsnYwA7J2NAOydjgDsnY8A7J2QAOydkQDsnZIA7J2TAOydlADsnZUA7J2WAOydlwDsnZgA7J2ZAOydmgDsnZsA7J2cAOydnQDsnZ4A7J2fAOydoADsnaEA7J2iAOydowDsnaQA7J2lAOydpgDsnacA7J2oAOydqQDsnaoA7J2rAOydrADsna0A7J2uAOydrwDsnbAA7J2xAOydsgDsnbMA7J20AOydtQDsnbYA7J23AOyduADsnbkA7J26AOyduwDsnbwA7J29AOydvgDsnb8A7J6AAOyegQDsnoIA7J6DAOyehADsnoUA7J6GAOyehwDsnogA7J6JAOyeigDsnosA7J6MAOyejQDsno4A7J6PAOyekADsnpEA7J6SAOyekwDsnpQA7J6VAOyelgDsnpcA7J6YAOyemQDsnpoA7J6bAOyenADsnp0A7J6eAOyenwDsnqAA7J6hAOyeogDsnqMA7J6kAOyepQDsnqYA7J6nAOyeqADsnqkA7J6qAOyeqwDsnqwA7J6tAOyergDsnq8A7J6wAOyesQDsnrIA7J6zAOyetADsnrUA7J62AOyetwDsnrgA7J65AOyeugDsnrsA7J68AOyevQDsnr4A7J6/AOyfgADsn4EA7J+CAOyfgwDsn4QA7J+FAOyfhgDsn4cA7J+IAOyfiQDsn4oA7J+LAOyfjADsn40A7J+OAOyfjwDsn5AA7J+RAOyfkgDsn5MA7J+UAOyflQDsn5YA7J+XAOyfmADsn5kA7J+aAOyfmwDsn5wA7J+dAOyfngDsn58A7J+gAOyfoQDsn6IA7J+jAOyfpADsn6UA7J+mAOyfpwDsn6gA7J+pAOyfqgDsn6sA7J+sAOyfrQDsn64A7J+vAOyfsADsn7EA7J+yAOyfswDsn7QA7J+1AOyftgDsn7cA7J+4AOyfuQDsn7oA7J+7AOyfvADsn70A7J++AOyfvwDsoIAA7KCBAOygggDsoIMA7KCEAOyghQDsoIYA7KCHAOygiADsoIkA7KCKAOygiwDsoIwA7KCNAOygjgDsoI8A7KCQAOygkQDsoJIA7KCTAOyglADsoJUA7KCWAOyglwDsoJgA7KCZAOygmgDsoJsA7KCcAOygnQDsoJ4A7KCfAOygoADsoKEA7KCiAOygowDsoKQA7KClAOygpgDsoKcA7KCoAOygqQDsoKoA7KCrAOygrADsoK0A7KCuAOygrwDsoLAA7KCxAOygsgDsoLMA7KC0AOygtQDsoLYA7KC3AOyguADsoLkA7KC6AOyguwDsoLwA7KC9AOygvgDsoL8A7KGAAOyhgQDsoYIA7KGDAOyhhADsoYUA7KGGAOyhhwDsoYgA7KGJAOyhigDsoYsA7KGMAOyhjQDsoY4A7KGPAOyhkADsoZEA7KGSAOyhkwDsoZQA7KGVAOyhlgDsoZcA7KGYAOyhmQDsoZoA7KGbAOyhnADsoZ0A7KGeAOyhnwDsoaAA7KGhAOyhogDsoaMA7KGkAOyhpQDsoaYA7KGnAOyhqADsoakA7KGqAOyhqwDsoawA7KGtAOyhrgDsoa8A7KGwAOyhsQDsobIA7KGzAOyhtADsobUA7KG2AOyhtwDsobgA7KG5AOyhugDsobsA7KG8AOyhvQDsob4A7KG/AOyigADsooEA7KKCAOyigwDsooQA7KKFAOyihgDsoocA7KKIAOyiiQDsoooA7KKLAOyijADsoo0A7KKOAOyijwDsopAA7KKRAOyikgDsopMA7KKUAOyilQDsopYA7KKXAOyimADsopkA7KKaAOyimwDsopwA7KKdAOyingDsop8A7KKgAOyioQDsoqIA7KKjAOyipADsoqUA7KKmAOyipwDsoqgA7KKpAOyiqgDsoqsA7KKsAOyirQDsoq4A7KKvAOyisADsorEA7KKyAOyiswDsorQA7KK1AOyitgDsorcA7KK4AOyiuQDsoroA7KK7AOyivADsor0A7KK+AOyivwDso4AA7KOBAOyjggDso4MA7KOEAOyjhQDso4YA7KOHAOyjiADso4kA7KOKAOyjiwDso4wA7KONAOyjjgDso48A7KOQAOyjkQDso5IA7KOTAOyjlADso5UA7KOWAOyjlwDso5gA7KOZAOyjmgDso5sA7KOcAOyjnQDso54A7KOfAOyjoADso6EA7KOiAOyjowDso6QA7KOlAOyjpgDso6cA7KOoAOyjqQDso6oA7KOrAOyjrADso60A7KOuAOyjrwDso7AA7KOxAOyjsgDso7MA7KO0AOyjtQDso7YA7KO3AOyjuADso7kA7KO6AOyjuwDso7wA7KO87J2YAOyjvQDso74A7KO/AOykgADspIEA7KSCAOykgwDspIQA7KSFAOykhgDspIcA7KSIAOykiQDspIoA7KSLAOykjADspI0A7KSOAOykjwDspJAA7KSRAOykkgDspJMA7KSUAOyklQDspJYA7KSXAOykmADspJkA7KSaAOykmwDspJwA7KSdAOykngDspJ8A7KSgAOykoQDspKIA7KSjAOykpADspKUA7KSmAOykpwDspKgA7KSpAOykqgDspKsA7KSsAOykrQDspK4A7KSvAOyksADspLEA7KSyAOykswDspLQA7KS1AOyktgDspLcA7KS4AOykuQDspLoA7KS7AOykvADspL0A7KS+AOykvwDspYAA7KWBAOylggDspYMA7KWEAOylhQDspYYA7KWHAOyliADspYkA7KWKAOyliwDspYwA7KWNAOyljgDspY8A7KWQAOylkQDspZIA7KWTAOyllADspZUA7KWWAOyllwDspZgA7KWZAOylmgDspZsA7KWcAOylnQDspZ4A7KWfAOyloADspaEA7KWiAOylowDspaQA7KWlAOylpgDspacA7KWoAOylqQDspaoA7KWrAOylrADspa0A7KWuAOylrwDspbAA7KWxAOylsgDspbMA7KW0AOyltQDspbYA7KW3AOyluADspbkA7KW6AOyluwDspbwA7KW9AOylvgDspb8A7KaAAOymgQDspoIA7KaDAOymhADspoUA7KaGAOymhwDspogA7KaJAOymigDsposA7KaMAOymjQDspo4A7KaPAOymkADsppEA7KaSAOymkwDsppQA7KaVAOymlgDsppcA7KaYAOymmQDsppoA7KabAOymnADspp0A7KaeAOymnwDspqAA7KahAOymogDspqMA7KakAOympQDspqYA7KanAOymqADspqkA7KaqAOymqwDspqwA7KatAOymrgDspq8A7KawAOymsQDsprIA7KazAOymtADsprUA7Ka2AOymtwDsprgA7Ka5AOymugDsprsA7Ka8AOymvQDspr4A7Ka/AOyngADsp4EA7KeCAOyngwDsp4QA7KeFAOynhgDsp4cA7KeIAOyniQDsp4oA7KeLAOynjADsp40A7KeOAOynjwDsp5AA7KeRAOynkgDsp5MA7KeUAOynlQDsp5YA7KeXAOynmADsp5kA7KeaAOynmwDsp5wA7KedAOynngDsp58A7KegAOynoQDsp6IA7KejAOynpADsp6UA7KemAOynpwDsp6gA7KepAOynqgDsp6sA7KesAOynrQDsp64A7KevAOynsADsp7EA7KeyAOynswDsp7QA7Ke1AOyntgDsp7cA7Ke4AOynuQDsp7oA7Ke7AOynvADsp70A7Ke+AOynvwDsqIAA7KiBAOyoggDsqIMA7KiEAOyohQDsqIYA7KiHAOyoiADsqIkA7KiKAOyoiwDsqIwA7KiNAOyojgDsqI8A7KiQAOyokQDsqJIA7KiTAOyolADsqJUA7KiWAOyolwDsqJgA7KiZAOyomgDsqJsA7KicAOyonQDsqJ4A7KifAOyooADsqKEA7KiiAOyoowDsqKQA7KilAOyopgDsqKcA7KioAOyoqQDsqKoA7KirAOyorADsqK0A7KiuAOyorwDsqLAA7KixAOyosgDsqLMA7Ki0AOyotQDsqLYA7Ki3AOyouADsqLkA7Ki6AOyouwDsqLwA7Ki9AOyovgDsqL8A7KmAAOypgQDsqYIA7KmDAOyphADsqYUA7KmGAOyphwDsqYgA7KmJAOypigDsqYsA7KmMAOypjQDsqY4A7KmPAOypkADsqZEA7KmSAOypkwDsqZQA7KmVAOyplgDsqZcA7KmYAOypmQDsqZoA7KmbAOypnADsqZ0A7KmeAOypnwDsqaAA7KmhAOypogDsqaMA7KmkAOyppQDsqaYA7KmnAOypqADsqakA7KmqAOypqwDsqawA7KmtAOyprgDsqa8A7KmwAOypsQDsqbIA7KmzAOyptADsqbUA7Km2AOyptwDsqbgA7Km5AOypugDsqbsA7Km8AOypvQDsqb4A7Km/AOyqgADsqoEA7KqCAOyqgwDsqoQA7KqFAOyqhgDsqocA7KqIAOyqiQDsqooA7KqLAOyqjADsqo0A7KqOAOyqjwDsqpAA7KqRAOyqkgDsqpMA7KqUAOyqlQDsqpYA7KqXAOyqmADsqpkA7KqaAOyqmwDsqpwA7KqdAOyqngDsqp8A7KqgAOyqoQDsqqIA7KqjAOyqpADsqqUA7KqmAOyqpwDsqqgA7KqpAOyqqgDsqqsA7KqsAOyqrQDsqq4A7KqvAOyqsADsqrEA7KqyAOyqswDsqrQA7Kq1AOyqtgDsqrcA7Kq4AOyquQDsqroA7Kq7AOyqvADsqr0A7Kq+AOyqvwDsq4AA7KuBAOyrggDsq4MA7KuEAOyrhQDsq4YA7KuHAOyriADsq4kA7KuKAOyriwDsq4wA7KuNAOyrjgDsq48A7KuQAOyrkQDsq5IA7KuTAOyrlADsq5UA7KuWAOyrlwDsq5gA7KuZAOyrmgDsq5sA7KucAOyrnQDsq54A7KufAOyroADsq6EA7KuiAOyrowDsq6QA7KulAOyrpgDsq6cA7KuoAOyrqQDsq6oA7KurAOyrrADsq60A7KuuAOyrrwDsq7AA7KuxAOyrsgDsq7MA7Ku0AOyrtQDsq7YA7Ku3AOyruADsq7kA7Ku6AOyruwDsq7wA7Ku9AOyrvgDsq78A7KyAAOysgQDsrIIA7KyDAOyshADsrIUA7KyGAOyshwDsrIgA7KyJAOysigDsrIsA7KyMAOysjQDsrI4A7KyPAOyskADsrJEA7KySAOyskwDsrJQA7KyVAOyslgDsrJcA7KyYAOysmQDsrJoA7KybAOysnADsrJ0A7KyeAOysnwDsrKAA7KyhAOysogDsrKMA7KykAOyspQDsrKYA7KynAOysqADsrKkA7KyqAOysqwDsrKwA7KytAOysrgDsrK8A7KywAOyssQDsrLIA7KyzAOystADsrLUA7Ky2AOystwDsrLgA7Ky5AOysugDsrLsA7Ky8AOysvQDsrL4A7Ky/AOytgADsrYEA7K2CAOytgwDsrYQA7K2FAOythgDsrYcA7K2IAOytiQDsrYoA7K2LAOytjADsrY0A7K2OAOytjwDsrZAA7K2RAOytkgDsrZMA7K2UAOytlQDsrZYA7K2XAOytmADsrZkA7K2aAOytmwDsrZwA7K2dAOytngDsrZ8A7K2gAOytoQDsraIA7K2jAOytpADsraUA7K2mAOytpwDsragA7K2pAOytqgDsrasA7K2sAOytrQDsra4A7K2vAOytsADsrbEA7K2yAOytswDsrbQA7K21AOyttgDsrbcA7K24AOytuQDsrboA7K27AOytvADsrb0A7K2+AOytvwDsroAA7K6BAOyuggDsroMA7K6EAOyuhQDsroYA7K6HAOyuiADsrokA7K6KAOyuiwDsrowA7K6NAOyujgDsro8A7K6QAOyukQDsrpIA7K6TAOyulADsrpUA7K6WAOyulwDsrpgA7K6ZAOyumgDsrpsA7K6cAOyunQDsrp4A7K6fAOyuoADsrqEA7K6iAOyuowDsrqQA7K6lAOyupgDsrqcA7K6oAOyuqQDsrqoA7K6rAOyurADsrq0A7K6uAOyurwDsrrAA7K6xAOyusgDsrrMA7K60AOyutQDsrrYA7K63AOyuuADsrrkA7K66AOyuuwDsrrwA7K69AOyuvgDsrr8A7K+AAOyvgQDsr4IA7K+DAOyvhADsr4UA7K+GAOyvhwDsr4gA7K+JAOyvigDsr4sA7K+MAOyvjQDsr44A7K+PAOyvkADsr5EA7K+SAOyvkwDsr5QA7K+VAOyvlgDsr5cA7K+YAOyvmQDsr5oA7K+bAOyvnADsr50A7K+eAOyvnwDsr6AA7K+hAOyvogDsr6MA7K+kAOyvpQDsr6YA7K+nAOyvqADsr6kA7K+qAOyvqwDsr6wA7K+tAOyvrgDsr68A7K+wAOyvsQDsr7IA7K+zAOyvtADsr7UA7K+2AOyvtwDsr7gA7K+5AOyvugDsr7sA7K+8AOyvvQDsr74A7K+/AOywgADssIEA7LCCAOywgwDssIQA7LCFAOywhgDssIcA7LCIAOywiQDssIoA7LCLAOywjADssI0A7LCOAOywjwDssJAA7LCRAOywkgDssJMA7LCUAOywlQDssJYA7LCXAOywmADssJkA7LCaAOywmwDssJwA7LCdAOywngDssJ8A7LCgAOywoQDssKIA7LCjAOywpADssKUA7LCmAOywpwDssKgA7LCpAOywqgDssKsA7LCsAOywrQDssK4A7LCvAOywsADssLEA7LCyAOywswDssLQA7LC1AOywtgDssLcA7LC4AOywuOqzoADssLkA7LC6AOywuwDssLwA7LC9AOywvgDssL8A7LGAAOyxgQDssYIA7LGDAOyxhADssYUA7LGGAOyxhwDssYgA7LGJAOyxigDssYsA7LGMAOyxjQDssY4A7LGPAOyxkADssZEA7LGSAOyxkwDssZQA7LGVAOyxlgDssZcA7LGYAOyxmQDssZoA7LGbAOyxnADssZ0A7LGeAOyxnwDssaAA7LGhAOyxogDssaMA7LGkAOyxpQDssaYA7LGnAOyxqADssakA7LGqAOyxqwDssawA7LGtAOyxrgDssa8A7LGwAOyxsQDssbIA7LGzAOyxtADssbUA7LG2AOyxtwDssbgA7LG5AOyxugDssbsA7LG8AOyxvQDssb4A7LG/AOyygADssoEA7LKCAOyygwDssoQA7LKFAOyyhgDssocA7LKIAOyyiQDssooA7LKLAOyyjADsso0A7LKOAOyyjwDsspAA7LKRAOyykgDsspMA7LKUAOyylQDsspYA7LKXAOyymADsspkA7LKaAOyymwDsspwA7LKdAOyyngDssp8A7LKgAOyyoQDssqIA7LKjAOyypADssqUA7LKmAOyypwDssqgA7LKpAOyyqgDssqsA7LKsAOyyrQDssq4A7LKvAOyysADssrEA7LKyAOyyswDssrQA7LK1AOyytgDssrcA7LK4AOyyuQDssroA7LK7AOyyvADssr0A7LK+AOyyvwDss4AA7LOBAOyzggDss4MA7LOEAOyzhQDss4YA7LOHAOyziADss4kA7LOKAOyziwDss4wA7LONAOyzjgDss48A7LOQAOyzkQDss5IA7LOTAOyzlADss5UA7LOWAOyzlwDss5gA7LOZAOyzmgDss5sA7LOcAOyznQDss54A7LOfAOyzoADss6EA7LOiAOyzowDss6QA7LOlAOyzpgDss6cA7LOoAOyzqQDss6oA7LOrAOyzrADss60A7LOuAOyzrwDss7AA7LOxAOyzsgDss7MA7LO0AOyztQDss7YA7LO3AOyzuADss7kA7LO6AOyzuwDss7wA7LO9AOyzvgDss78A7LSAAOy0gQDstIIA7LSDAOy0hADstIUA7LSGAOy0hwDstIgA7LSJAOy0igDstIsA7LSMAOy0jQDstI4A7LSPAOy0kADstJEA7LSSAOy0kwDstJQA7LSVAOy0lgDstJcA7LSYAOy0mQDstJoA7LSbAOy0nADstJ0A7LSeAOy0nwDstKAA7LShAOy0ogDstKMA7LSkAOy0pQDstKYA7LSnAOy0qADstKkA7LSqAOy0qwDstKwA7LStAOy0rgDstK8A7LSwAOy0sQDstLIA7LSzAOy0tADstLUA7LS2AOy0twDstLgA7LS5AOy0ugDstLsA7LS8AOy0vQDstL4A7LS/AOy1gADstYEA7LWCAOy1gwDstYQA7LWFAOy1hgDstYcA7LWIAOy1iQDstYoA7LWLAOy1jADstY0A7LWOAOy1jwDstZAA7LWRAOy1kgDstZMA7LWUAOy1lQDstZYA7LWXAOy1mADstZkA7LWaAOy1mwDstZwA7LWdAOy1ngDstZ8A7LWgAOy1oQDstaIA7LWjAOy1pADstaUA7LWmAOy1pwDstagA7LWpAOy1qgDstasA7LWsAOy1rQDsta4A7LWvAOy1sADstbEA7LWyAOy1swDstbQA7LW1AOy1tgDstbcA7LW4AOy1uQDstboA7LW7AOy1vADstb0A7LW+AOy1vwDstoAA7LaBAOy2ggDstoMA7LaEAOy2hQDstoYA7LaHAOy2iADstokA7LaKAOy2iwDstowA7LaNAOy2jgDsto8A7LaQAOy2kQDstpIA7LaTAOy2lADstpUA7LaWAOy2lwDstpgA7LaZAOy2mgDstpsA7LacAOy2nQDstp4A7LafAOy2oADstqEA7LaiAOy2owDstqQA7LalAOy2pgDstqcA7LaoAOy2qQDstqoA7LarAOy2rADstq0A7LauAOy2rwDstrAA7LaxAOy2sgDstrMA7La0AOy2tQDstrYA7La3AOy2uADstrkA7La6AOy2uwDstrwA7La9AOy2vgDstr8A7LeAAOy3gQDst4IA7LeDAOy3hADst4UA7LeGAOy3hwDst4gA7LeJAOy3igDst4sA7LeMAOy3jQDst44A7LePAOy3kADst5EA7LeSAOy3kwDst5QA7LeVAOy3lgDst5cA7LeYAOy3mQDst5oA7LebAOy3nADst50A7LeeAOy3nwDst6AA7LehAOy3ogDst6MA7LekAOy3pQDst6YA7LenAOy3qADst6kA7LeqAOy3qwDst6wA7LetAOy3rgDst68A7LewAOy3sQDst7IA7LezAOy3tADst7UA7Le2AOy3twDst7gA7Le5AOy3ugDst7sA7Le8AOy3vQDst74A7Le/AOy4gADsuIEA7LiCAOy4gwDsuIQA7LiFAOy4hgDsuIcA7LiIAOy4iQDsuIoA7LiLAOy4jADsuI0A7LiOAOy4jwDsuJAA7LiRAOy4kgDsuJMA7LiUAOy4lQDsuJYA7LiXAOy4mADsuJkA7LiaAOy4mwDsuJwA7LidAOy4ngDsuJ8A7LigAOy4oQDsuKIA7LijAOy4pADsuKUA7LimAOy4pwDsuKgA7LipAOy4qgDsuKsA7LisAOy4rQDsuK4A7LivAOy4sADsuLEA7LiyAOy4swDsuLQA7Li1AOy4tgDsuLcA7Li4AOy4uQDsuLoA7Li7AOy4vADsuL0A7Li+AOy4vwDsuYAA7LmBAOy5ggDsuYMA7LmEAOy5hQDsuYYA7LmHAOy5iADsuYkA7LmKAOy5iwDsuYwA7LmNAOy5jgDsuY8A7LmQAOy5kQDsuZIA7LmTAOy5lADsuZUA7LmWAOy5lwDsuZgA7LmZAOy5mgDsuZsA7LmcAOy5nQDsuZ4A7LmfAOy5oADsuaEA7LmiAOy5owDsuaQA7LmlAOy5pgDsuacA7LmoAOy5qQDsuaoA7LmrAOy5rADsua0A7LmuAOy5rwDsubAA7LmxAOy5sgDsubMA7Lm0AOy5tQDsubYA7Lm3AOy5uADsubkA7Lm6AOy5uwDsubwA7Lm9AOy5vgDsub8A7LqAAOy6gQDsuoIA7LqDAOy6hADsuoUA7LqGAOy6hwDsuogA7LqJAOy6igDsuosA7LqMAOy6jQDsuo4A7LqPAOy6kADsupEA7LqSAOy6kwDsupQA7LqVAOy6lgDsupcA7LqYAOy6mQDsupoA7LqbAOy6nADsup0A7LqeAOy6nwDsuqAA7LqhAOy6ogDsuqMA7LqkAOy6pQDsuqYA7LqnAOy6qADsuqkA7LqqAOy6qwDsuqwA7LqtAOy6rgDsuq8A7LqwAOy6sQDsurIA7LqzAOy6tADsurUA7Lq2AOy6twDsurgA7Lq5AOy6ugDsursA7Lq8AOy6vQDsur4A7Lq/AOy7gADsu4EA7LuCAOy7gwDsu4QA7LuFAOy7hgDsu4cA7LuIAOy7iQDsu4oA7LuLAOy7jADsu40A7LuOAOy7jwDsu5AA7LuRAOy7kgDsu5MA7LuUAOy7lQDsu5YA7LuXAOy7mADsu5kA7LuaAOy7mwDsu5wA7LudAOy7ngDsu58A7LugAOy7oQDsu6IA7LujAOy7pADsu6UA7LumAOy7pwDsu6gA7LupAOy7qgDsu6sA7LusAOy7rQDsu64A7LuvAOy7sADsu7EA7LuyAOy7swDsu7QA7Lu1AOy7tgDsu7cA7Lu4AOy7uQDsu7oA7Lu7AOy7vADsu70A7Lu+AOy7vwDsvIAA7LyBAOy8ggDsvIMA7LyEAOy8hQDsvIYA7LyHAOy8iADsvIkA7LyKAOy8iwDsvIwA7LyNAOy8jgDsvI8A7LyQAOy8kQDsvJIA7LyTAOy8lADsvJUA7LyWAOy8lwDsvJgA7LyZAOy8mgDsvJsA7LycAOy8nQDsvJ4A7LyfAOy8oADsvKEA7LyiAOy8owDsvKQA7LylAOy8pgDsvKcA7LyoAOy8qQDsvKoA7LyrAOy8rADsvK0A7LyuAOy8rwDsvLAA7LyxAOy8sgDsvLMA7Ly0AOy8tQDsvLYA7Ly3AOy8uADsvLkA7Ly6AOy8uwDsvLwA7Ly9AOy8vgDsvL8A7L2AAOy9gQDsvYIA7L2DAOy9hADsvYUA7L2GAOy9hwDsvYgA7L2JAOy9igDsvYsA7L2MAOy9jQDsvY4A7L2PAOy9kADsvZEA7L2SAOy9kwDsvZQA7L2VAOy9lgDsvZcA7L2YAOy9mQDsvZoA7L2bAOy9nADsvZ0A7L2eAOy9nwDsvaAA7L2hAOy9ogDsvaMA7L2kAOy9pQDsvaYA7L2nAOy9qADsvakA7L2qAOy9qwDsvawA7L2tAOy9rgDsva8A7L2wAOy9sQDsvbIA7L2zAOy9tADsvbUA7L22AOy9twDsvbgA7L25AOy9ugDsvbsA7L28AOy9vQDsvb4A7L2/AOy+gADsvoEA7L6CAOy+gwDsvoQA7L6FAOy+hgDsvocA7L6IAOy+iQDsvooA7L6LAOy+jADsvo0A7L6OAOy+jwDsvpAA7L6RAOy+kgDsvpMA7L6UAOy+lQDsvpYA7L6XAOy+mADsvpkA7L6aAOy+mwDsvpwA7L6dAOy+ngDsvp8A7L6gAOy+oQDsvqIA7L6jAOy+pADsvqUA7L6mAOy+pwDsvqgA7L6pAOy+qgDsvqsA7L6sAOy+rQDsvq4A7L6vAOy+sADsvrEA7L6yAOy+swDsvrQA7L61AOy+tgDsvrcA7L64AOy+uQDsvroA7L67AOy+vADsvr0A7L6+AOy+vwDsv4AA7L+BAOy/ggDsv4MA7L+EAOy/hQDsv4YA7L+HAOy/iADsv4kA7L+KAOy/iwDsv4wA7L+NAOy/jgDsv48A7L+QAOy/kQDsv5IA7L+TAOy/lADsv5UA7L+WAOy/lwDsv5gA7L+ZAOy/mgDsv5sA7L+cAOy/nQDsv54A7L+fAOy/oADsv6EA7L+iAOy/owDsv6QA7L+lAOy/pgDsv6cA7L+oAOy/qQDsv6oA7L+rAOy/rADsv60A7L+uAOy/rwDsv7AA7L+xAOy/sgDsv7MA7L+0AOy/tQDsv7YA7L+3AOy/uADsv7kA7L+6AOy/uwDsv7wA7L+9AOy/vgDsv78A7YCAAO2AgQDtgIIA7YCDAO2AhADtgIUA7YCGAO2AhwDtgIgA7YCJAO2AigDtgIsA7YCMAO2AjQDtgI4A7YCPAO2AkADtgJEA7YCSAO2AkwDtgJQA7YCVAO2AlgDtgJcA7YCYAO2AmQDtgJoA7YCbAO2AnADtgJ0A7YCeAO2AnwDtgKAA7YChAO2AogDtgKMA7YCkAO2ApQDtgKYA7YCnAO2AqADtgKkA7YCqAO2AqwDtgKwA7YCtAO2ArgDtgK8A7YCwAO2AsQDtgLIA7YCzAO2AtADtgLUA7YC2AO2AtwDtgLgA7YC5AO2AugDtgLsA7YC8AO2AvQDtgL4A7YC/AO2BgADtgYEA7YGCAO2BgwDtgYQA7YGFAO2BhgDtgYcA7YGIAO2BiQDtgYoA7YGLAO2BjADtgY0A7YGOAO2BjwDtgZAA7YGRAO2BkgDtgZMA7YGUAO2BlQDtgZYA7YGXAO2BmADtgZkA7YGaAO2BmwDtgZwA7YGdAO2BngDtgZ8A7YGgAO2BoQDtgaIA7YGjAO2BpADtgaUA7YGmAO2BpwDtgagA7YGpAO2BqgDtgasA7YGsAO2BrQDtga4A7YGvAO2BsADtgbEA7YGyAO2BswDtgbQA7YG1AO2BtgDtgbcA7YG4AO2BuQDtgboA7YG7AO2BvADtgb0A7YG+AO2BvwDtgoAA7YKBAO2CggDtgoMA7YKEAO2ChQDtgoYA7YKHAO2CiADtgokA7YKKAO2CiwDtgowA7YKNAO2CjgDtgo8A7YKQAO2CkQDtgpIA7YKTAO2ClADtgpUA7YKWAO2ClwDtgpgA7YKZAO2CmgDtgpsA7YKcAO2CnQDtgp4A7YKfAO2CoADtgqEA7YKiAO2CowDtgqQA7YKlAO2CpgDtgqcA7YKoAO2CqQDtgqoA7YKrAO2CrADtgq0A7YKuAO2CrwDtgrAA7YKxAO2CsgDtgrMA7YK0AO2CtQDtgrYA7YK3AO2CuADtgrkA7YK6AO2CuwDtgrwA7YK9AO2CvgDtgr8A7YOAAO2DgQDtg4IA7YODAO2DhADtg4UA7YOGAO2DhwDtg4gA7YOJAO2DigDtg4sA7YOMAO2DjQDtg44A7YOPAO2DkADtg5EA7YOSAO2DkwDtg5QA7YOVAO2DlgDtg5cA7YOYAO2DmQDtg5oA7YObAO2DnADtg50A7YOeAO2DnwDtg6AA7YOhAO2DogDtg6MA7YOkAO2DpQDtg6YA7YOnAO2DqADtg6kA7YOqAO2DqwDtg6wA7YOtAO2DrgDtg68A7YOwAO2DsQDtg7IA7YOzAO2DtADtg7UA7YO2AO2DtwDtg7gA7YO5AO2DugDtg7sA7YO8AO2DvQDtg74A7YO/AO2EgADthIEA7YSCAO2EgwDthIQA7YSFAO2EhgDthIcA7YSIAO2EiQDthIoA7YSLAO2EjADthI0A7YSOAO2EjwDthJAA7YSRAO2EkgDthJMA7YSUAO2ElQDthJYA7YSXAO2EmADthJkA7YSaAO2EmwDthJwA7YSdAO2EngDthJ8A7YSgAO2EoQDthKIA7YSjAO2EpADthKUA7YSmAO2EpwDthKgA7YSpAO2EqgDthKsA7YSsAO2ErQDthK4A7YSvAO2EsADthLEA7YSyAO2EswDthLQA7YS1AO2EtgDthLcA7YS4AO2EuQDthLoA7YS7AO2EvADthL0A7YS+AO2EvwDthYAA7YWBAO2FggDthYMA7YWEAO2FhQDthYYA7YWHAO2FiADthYkA7YWKAO2FiwDthYwA7YWNAO2FjgDthY8A7YWQAO2FkQDthZIA7YWTAO2FlADthZUA7YWWAO2FlwDthZgA7YWZAO2FmgDthZsA7YWcAO2FnQDthZ4A7YWfAO2FoADthaEA7YWiAO2FowDthaQA7YWlAO2FpgDthacA7YWoAO2FqQDthaoA7YWrAO2FrADtha0A7YWuAO2FrwDthbAA7YWxAO2FsgDthbMA7YW0AO2FtQDthbYA7YW3AO2FuADthbkA7YW6AO2FuwDthbwA7YW9AO2FvgDthb8A7YaAAO2GgQDthoIA7YaDAO2GhADthoUA7YaGAO2GhwDthogA7YaJAO2GigDthosA7YaMAO2GjQDtho4A7YaPAO2GkADthpEA7YaSAO2GkwDthpQA7YaVAO2GlgDthpcA7YaYAO2GmQDthpoA7YabAO2GnADthp0A7YaeAO2GnwDthqAA7YahAO2GogDthqMA7YakAO2GpQDthqYA7YanAO2GqADthqkA7YaqAO2GqwDthqwA7YatAO2GrgDthq8A7YawAO2GsQDthrIA7YazAO2GtADthrUA7Ya2AO2GtwDthrgA7Ya5AO2GugDthrsA7Ya8AO2GvQDthr4A7Ya/AO2HgADth4EA7YeCAO2HgwDth4QA7YeFAO2HhgDth4cA7YeIAO2HiQDth4oA7YeLAO2HjADth40A7YeOAO2HjwDth5AA7YeRAO2HkgDth5MA7YeUAO2HlQDth5YA7YeXAO2HmADth5kA7YeaAO2HmwDth5wA7YedAO2HngDth58A7YegAO2HoQDth6IA7YejAO2HpADth6UA7YemAO2HpwDth6gA7YepAO2HqgDth6sA7YesAO2HrQDth64A7YevAO2HsADth7EA7YeyAO2HswDth7QA7Ye1AO2HtgDth7cA7Ye4AO2HuQDth7oA7Ye7AO2HvADth70A7Ye+AO2HvwDtiIAA7YiBAO2IggDtiIMA7YiEAO2IhQDtiIYA7YiHAO2IiADtiIkA7YiKAO2IiwDtiIwA7YiNAO2IjgDtiI8A7YiQAO2IkQDtiJIA7YiTAO2IlADtiJUA7YiWAO2IlwDtiJgA7YiZAO2ImgDtiJsA7YicAO2InQDtiJ4A7YifAO2IoADtiKEA7YiiAO2IowDtiKQA7YilAO2IpgDtiKcA7YioAO2IqQDtiKoA7YirAO2IrADtiK0A7YiuAO2IrwDtiLAA7YixAO2IsgDtiLMA7Yi0AO2ItQDtiLYA7Yi3AO2IuADtiLkA7Yi6AO2IuwDtiLwA7Yi9AO2IvgDtiL8A7YmAAO2JgQDtiYIA7YmDAO2JhADtiYUA7YmGAO2JhwDtiYgA7YmJAO2JigDtiYsA7YmMAO2JjQDtiY4A7YmPAO2JkADtiZEA7YmSAO2JkwDtiZQA7YmVAO2JlgDtiZcA7YmYAO2JmQDtiZoA7YmbAO2JnADtiZ0A7YmeAO2JnwDtiaAA7YmhAO2JogDtiaMA7YmkAO2JpQDtiaYA7YmnAO2JqADtiakA7YmqAO2JqwDtiawA7YmtAO2JrgDtia8A7YmwAO2JsQDtibIA7YmzAO2JtADtibUA7Ym2AO2JtwDtibgA7Ym5AO2JugDtibsA7Ym8AO2JvQDtib4A7Ym/AO2KgADtioEA7YqCAO2KgwDtioQA7YqFAO2KhgDtiocA7YqIAO2KiQDtiooA7YqLAO2KjADtio0A7YqOAO2KjwDtipAA7YqRAO2KkgDtipMA7YqUAO2KlQDtipYA7YqXAO2KmADtipkA7YqaAO2KmwDtipwA7YqdAO2KngDtip8A7YqgAO2KoQDtiqIA7YqjAO2KpADtiqUA7YqmAO2KpwDtiqgA7YqpAO2KqgDtiqsA7YqsAO2KrQDtiq4A7YqvAO2KsADtirEA7YqyAO2KswDtirQA7Yq1AO2KtgDtircA7Yq4AO2KuQDtiroA7Yq7AO2KvADtir0A7Yq+AO2KvwDti4AA7YuBAO2LggDti4MA7YuEAO2LhQDti4YA7YuHAO2LiADti4kA7YuKAO2LiwDti4wA7YuNAO2LjgDti48A7YuQAO2LkQDti5IA7YuTAO2LlADti5UA7YuWAO2LlwDti5gA7YuZAO2LmgDti5sA7YucAO2LnQDti54A7YufAO2LoADti6EA7YuiAO2LowDti6QA7YulAO2LpgDti6cA7YuoAO2LqQDti6oA7YurAO2LrADti60A7YuuAO2LrwDti7AA7YuxAO2LsgDti7MA7Yu0AO2LtQDti7YA7Yu3AO2LuADti7kA7Yu6AO2LuwDti7wA7Yu9AO2LvgDti78A7YyAAO2MgQDtjIIA7YyDAO2MhADtjIUA7YyGAO2MhwDtjIgA7YyJAO2MigDtjIsA7YyMAO2MjQDtjI4A7YyPAO2MkADtjJEA7YySAO2MkwDtjJQA7YyVAO2MlgDtjJcA7YyYAO2MmQDtjJoA7YybAO2MnADtjJ0A7YyeAO2MnwDtjKAA7YyhAO2MogDtjKMA7YykAO2MpQDtjKYA7YynAO2MqADtjKkA7YyqAO2MqwDtjKwA7YytAO2MrgDtjK8A7YywAO2MsQDtjLIA7YyzAO2MtADtjLUA7Yy2AO2MtwDtjLgA7Yy5AO2MugDtjLsA7Yy8AO2MvQDtjL4A7Yy/AO2NgADtjYEA7Y2CAO2NgwDtjYQA7Y2FAO2NhgDtjYcA7Y2IAO2NiQDtjYoA7Y2LAO2NjADtjY0A7Y2OAO2NjwDtjZAA7Y2RAO2NkgDtjZMA7Y2UAO2NlQDtjZYA7Y2XAO2NmADtjZkA7Y2aAO2NmwDtjZwA7Y2dAO2NngDtjZ8A7Y2gAO2NoQDtjaIA7Y2jAO2NpADtjaUA7Y2mAO2NpwDtjagA7Y2pAO2NqgDtjasA7Y2sAO2NrQDtja4A7Y2vAO2NsADtjbEA7Y2yAO2NswDtjbQA7Y21AO2NtgDtjbcA7Y24AO2NuQDtjboA7Y27AO2NvADtjb0A7Y2+AO2NvwDtjoAA7Y6BAO2OggDtjoMA7Y6EAO2OhQDtjoYA7Y6HAO2OiADtjokA7Y6KAO2OiwDtjowA7Y6NAO2OjgDtjo8A7Y6QAO2OkQDtjpIA7Y6TAO2OlADtjpUA7Y6WAO2OlwDtjpgA7Y6ZAO2OmgDtjpsA7Y6cAO2OnQDtjp4A7Y6fAO2OoADtjqEA7Y6iAO2OowDtjqQA7Y6lAO2OpgDtjqcA7Y6oAO2OqQDtjqoA7Y6rAO2OrADtjq0A7Y6uAO2OrwDtjrAA7Y6xAO2OsgDtjrMA7Y60AO2OtQDtjrYA7Y63AO2OuADtjrkA7Y66AO2OuwDtjrwA7Y69AO2OvgDtjr8A7Y+AAO2PgQDtj4IA7Y+DAO2PhADtj4UA7Y+GAO2PhwDtj4gA7Y+JAO2PigDtj4sA7Y+MAO2PjQDtj44A7Y+PAO2PkADtj5EA7Y+SAO2PkwDtj5QA7Y+VAO2PlgDtj5cA7Y+YAO2PmQDtj5oA7Y+bAO2PnADtj50A7Y+eAO2PnwDtj6AA7Y+hAO2PogDtj6MA7Y+kAO2PpQDtj6YA7Y+nAO2PqADtj6kA7Y+qAO2PqwDtj6wA7Y+tAO2PrgDtj68A7Y+wAO2PsQDtj7IA7Y+zAO2PtADtj7UA7Y+2AO2PtwDtj7gA7Y+5AO2PugDtj7sA7Y+8AO2PvQDtj74A7Y+/AO2QgADtkIEA7ZCCAO2QgwDtkIQA7ZCFAO2QhgDtkIcA7ZCIAO2QiQDtkIoA7ZCLAO2QjADtkI0A7ZCOAO2QjwDtkJAA7ZCRAO2QkgDtkJMA7ZCUAO2QlQDtkJYA7ZCXAO2QmADtkJkA7ZCaAO2QmwDtkJwA7ZCdAO2QngDtkJ8A7ZCgAO2QoQDtkKIA7ZCjAO2QpADtkKUA7ZCmAO2QpwDtkKgA7ZCpAO2QqgDtkKsA7ZCsAO2QrQDtkK4A7ZCvAO2QsADtkLEA7ZCyAO2QswDtkLQA7ZC1AO2QtgDtkLcA7ZC4AO2QuQDtkLoA7ZC7AO2QvADtkL0A7ZC+AO2QvwDtkYAA7ZGBAO2RggDtkYMA7ZGEAO2RhQDtkYYA7ZGHAO2RiADtkYkA7ZGKAO2RiwDtkYwA7ZGNAO2RjgDtkY8A7ZGQAO2RkQDtkZIA7ZGTAO2RlADtkZUA7ZGWAO2RlwDtkZgA7ZGZAO2RmgDtkZsA7ZGcAO2RnQDtkZ4A7ZGfAO2RoADtkaEA7ZGiAO2RowDtkaQA7ZGlAO2RpgDtkacA7ZGoAO2RqQDtkaoA7ZGrAO2RrADtka0A7ZGuAO2RrwDtkbAA7ZGxAO2RsgDtkbMA7ZG0AO2RtQDtkbYA7ZG3AO2RuADtkbkA7ZG6AO2RuwDtkbwA7ZG9AO2RvgDtkb8A7ZKAAO2SgQDtkoIA7ZKDAO2ShADtkoUA7ZKGAO2ShwDtkogA7ZKJAO2SigDtkosA7ZKMAO2SjQDtko4A7ZKPAO2SkADtkpEA7ZKSAO2SkwDtkpQA7ZKVAO2SlgDtkpcA7ZKYAO2SmQDtkpoA7ZKbAO2SnADtkp0A7ZKeAO2SnwDtkqAA7ZKhAO2SogDtkqMA7ZKkAO2SpQDtkqYA7ZKnAO2SqADtkqkA7ZKqAO2SqwDtkqwA7ZKtAO2SrgDtkq8A7ZKwAO2SsQDtkrIA7ZKzAO2StADtkrUA7ZK2AO2StwDtkrgA7ZK5AO2SugDtkrsA7ZK8AO2SvQDtkr4A7ZK/AO2TgADtk4EA7ZOCAO2TgwDtk4QA7ZOFAO2ThgDtk4cA7ZOIAO2TiQDtk4oA7ZOLAO2TjADtk40A7ZOOAO2TjwDtk5AA7ZORAO2TkgDtk5MA7ZOUAO2TlQDtk5YA7ZOXAO2TmADtk5kA7ZOaAO2TmwDtk5wA7ZOdAO2TngDtk58A7ZOgAO2ToQDtk6IA7ZOjAO2TpADtk6UA7ZOmAO2TpwDtk6gA7ZOpAO2TqgDtk6sA7ZOsAO2TrQDtk64A7ZOvAO2TsADtk7EA7ZOyAO2TswDtk7QA7ZO1AO2TtgDtk7cA7ZO4AO2TuQDtk7oA7ZO7AO2TvADtk70A7ZO+AO2TvwDtlIAA7ZSBAO2UggDtlIMA7ZSEAO2UhQDtlIYA7ZSHAO2UiADtlIkA7ZSKAO2UiwDtlIwA7ZSNAO2UjgDtlI8A7ZSQAO2UkQDtlJIA7ZSTAO2UlADtlJUA7ZSWAO2UlwDtlJgA7ZSZAO2UmgDtlJsA7ZScAO2UnQDtlJ4A7ZSfAO2UoADtlKEA7ZSiAO2UowDtlKQA7ZSlAO2UpgDtlKcA7ZSoAO2UqQDtlKoA7ZSrAO2UrADtlK0A7ZSuAO2UrwDtlLAA7ZSxAO2UsgDtlLMA7ZS0AO2UtQDtlLYA7ZS3AO2UuADtlLkA7ZS6AO2UuwDtlLwA7ZS9AO2UvgDtlL8A7ZWAAO2VgQDtlYIA7ZWDAO2VhADtlYUA7ZWGAO2VhwDtlYgA7ZWJAO2VigDtlYsA7ZWMAO2VjQDtlY4A7ZWPAO2VkADtlZEA7ZWSAO2VkwDtlZQA7ZWVAO2VlgDtlZcA7ZWYAO2VmQDtlZoA7ZWbAO2VnADtlZ0A7ZWeAO2VnwDtlaAA7ZWhAO2VogDtlaMA7ZWkAO2VpQDtlaYA7ZWnAO2VqADtlakA7ZWqAO2VqwDtlawA7ZWtAO2VrgDtla8A7ZWwAO2VsQDtlbIA7ZWzAO2VtADtlbUA7ZW2AO2VtwDtlbgA7ZW5AO2VugDtlbsA7ZW8AO2VvQDtlb4A7ZW/AO2WgADtloEA7ZaCAO2WgwDtloQA7ZaFAO2WhgDtlocA7ZaIAO2WiQDtlooA7ZaLAO2WjADtlo0A7ZaOAO2WjwDtlpAA7ZaRAO2WkgDtlpMA7ZaUAO2WlQDtlpYA7ZaXAO2WmADtlpkA7ZaaAO2WmwDtlpwA7ZadAO2WngDtlp8A7ZagAO2WoQDtlqIA7ZajAO2WpADtlqUA7ZamAO2WpwDtlqgA7ZapAO2WqgDtlqsA7ZasAO2WrQDtlq4A7ZavAO2WsADtlrEA7ZayAO2WswDtlrQA7Za1AO2WtgDtlrcA7Za4AO2WuQDtlroA7Za7AO2WvADtlr0A7Za+AO2WvwDtl4AA7ZeBAO2XggDtl4MA7ZeEAO2XhQDtl4YA7ZeHAO2XiADtl4kA7ZeKAO2XiwDtl4wA7ZeNAO2XjgDtl48A7ZeQAO2XkQDtl5IA7ZeTAO2XlADtl5UA7ZeWAO2XlwDtl5gA7ZeZAO2XmgDtl5sA7ZecAO2XnQDtl54A7ZefAO2XoADtl6EA7ZeiAO2XowDtl6QA7ZelAO2XpgDtl6cA7ZeoAO2XqQDtl6oA7ZerAO2XrADtl60A7ZeuAO2XrwDtl7AA7ZexAO2XsgDtl7MA7Ze0AO2XtQDtl7YA7Ze3AO2XuADtl7kA7Ze6AO2XuwDtl7wA7Ze9AO2XvgDtl78A7ZiAAO2YgQDtmIIA7ZiDAO2YhADtmIUA7ZiGAO2YhwDtmIgA7ZiJAO2YigDtmIsA7ZiMAO2YjQDtmI4A7ZiPAO2YkADtmJEA7ZiSAO2YkwDtmJQA7ZiVAO2YlgDtmJcA7ZiYAO2YmQDtmJoA7ZibAO2YnADtmJ0A7ZieAO2YnwDtmKAA7ZihAO2YogDtmKMA7ZikAO2YpQDtmKYA7ZinAO2YqADtmKkA7ZiqAO2YqwDtmKwA7ZitAO2YrgDtmK8A7ZiwAO2YsQDtmLIA7ZizAO2YtADtmLUA7Zi2AO2YtwDtmLgA7Zi5AO2YugDtmLsA7Zi8AO2YvQDtmL4A7Zi/AO2ZgADtmYEA7ZmCAO2ZgwDtmYQA7ZmFAO2ZhgDtmYcA7ZmIAO2ZiQDtmYoA7ZmLAO2ZjADtmY0A7ZmOAO2ZjwDtmZAA7ZmRAO2ZkgDtmZMA7ZmUAO2ZlQDtmZYA7ZmXAO2ZmADtmZkA7ZmaAO2ZmwDtmZwA7ZmdAO2ZngDtmZ8A7ZmgAO2ZoQDtmaIA7ZmjAO2ZpADtmaUA7ZmmAO2ZpwDtmagA7ZmpAO2ZqgDtmasA7ZmsAO2ZrQDtma4A7ZmvAO2ZsADtmbEA7ZmyAO2ZswDtmbQA7Zm1AO2ZtgDtmbcA7Zm4AO2ZuQDtmboA7Zm7AO2ZvADtmb0A7Zm+AO2ZvwDtmoAA7ZqBAO2aggDtmoMA7ZqEAO2ahQDtmoYA7ZqHAO2aiADtmokA7ZqKAO2aiwDtmowA7ZqNAO2ajgDtmo8A7ZqQAO2akQDtmpIA7ZqTAO2alADtmpUA7ZqWAO2alwDtmpgA7ZqZAO2amgDtmpsA7ZqcAO2anQDtmp4A7ZqfAO2aoADtmqEA7ZqiAO2aowDtmqQA7ZqlAO2apgDtmqcA7ZqoAO2aqQDtmqoA7ZqrAO2arADtmq0A7ZquAO2arwDtmrAA7ZqxAO2asgDtmrMA7Zq0AO2atQDtmrYA7Zq3AO2auADtmrkA7Zq6AO2auwDtmrwA7Zq9AO2avgDtmr8A7ZuAAO2bgQDtm4IA7ZuDAO2bhADtm4UA7ZuGAO2bhwDtm4gA7ZuJAO2bigDtm4sA7ZuMAO2bjQDtm44A7ZuPAO2bkADtm5EA7ZuSAO2bkwDtm5QA7ZuVAO2blgDtm5cA7ZuYAO2bmQDtm5oA7ZubAO2bnADtm50A7ZueAO2bnwDtm6AA7ZuhAO2bogDtm6MA7ZukAO2bpQDtm6YA7ZunAO2bqADtm6kA7ZuqAO2bqwDtm6wA7ZutAO2brgDtm68A7ZuwAO2bsQDtm7IA7ZuzAO2btADtm7UA7Zu2AO2btwDtm7gA7Zu5AO2bugDtm7sA7Zu8AO2bvQDtm74A7Zu/AO2cgADtnIEA7ZyCAO2cgwDtnIQA7ZyFAO2chgDtnIcA7ZyIAO2ciQDtnIoA7ZyLAO2cjADtnI0A7ZyOAO2cjwDtnJAA7ZyRAO2ckgDtnJMA7ZyUAO2clQDtnJYA7ZyXAO2cmADtnJkA7ZyaAO2cmwDtnJwA7ZydAO2cngDtnJ8A7ZygAO2coQDtnKIA7ZyjAO2cpADtnKUA7ZymAO2cpwDtnKgA7ZypAO2cqgDtnKsA7ZysAO2crQDtnK4A7ZyvAO2csADtnLEA7ZyyAO2cswDtnLQA7Zy1AO2ctgDtnLcA7Zy4AO2cuQDtnLoA7Zy7AO2cvADtnL0A7Zy+AO2cvwDtnYAA7Z2BAO2dggDtnYMA7Z2EAO2dhQDtnYYA7Z2HAO2diADtnYkA7Z2KAO2diwDtnYwA7Z2NAO2djgDtnY8A7Z2QAO2dkQDtnZIA7Z2TAO2dlADtnZUA7Z2WAO2dlwDtnZgA7Z2ZAO2dmgDtnZsA7Z2cAO2dnQDtnZ4A7Z2fAO2doADtnaEA7Z2iAO2dowDtnaQA7Z2lAO2dpgDtnacA7Z2oAO2dqQDtnaoA7Z2rAO2drADtna0A7Z2uAO2drwDtnbAA7Z2xAO2dsgDtnbMA7Z20AO2dtQDtnbYA7Z23AO2duADtnbkA7Z26AO2duwDtnbwA7Z29AO2dvgDtnb8A7Z6AAO2egQDtnoIA7Z6DAO2ehADtnoUA7Z6GAO2ehwDtnogA7Z6JAO2eigDtnosA7Z6MAO2ejQDtno4A7Z6PAO2ekADtnpEA7Z6SAO2ekwDtnpQA7Z6VAO2elgDtnpcA7Z6YAO2emQDtnpoA7Z6bAO2enADtnp0A7Z6eAO2enwDtnqAA7Z6hAO2eogDtnqMA8JGCmgDwkYKcAPCRgqsA8JGErgDwkYSvAPCRjYsA8JGNjADwkZK7APCRkrwA8JGSvgDwkZa6APCRlrsA8JGkuADwnYWX8J2FpQDwnYWY8J2FpQDwnYWY8J2FpfCdha4A8J2FmPCdhaXwnYWvAPCdhZjwnYWl8J2FsADwnYWY8J2FpfCdhbEA8J2FmPCdhaXwnYWyAPCdhrnwnYWlAPCdhrnwnYWl8J2FrgDwnYa58J2FpfCdha8A8J2GuvCdhaUA8J2GuvCdhaXwnYWuAPCdhrrwnYWl8J2FrwDwoISiAPCglJwA8KCUpQDwoJWLAPCgmLoA8KCghADwoKOeAPCgqKwA8KCtowDwoZOkAPChmqgA8KGbqgDwoaeIAPChrJgA8KG0iwDwobekAPCht6YA8KKGgwDwooafAPCijLEA8KKblADwoqGEAPCioYoA8KKsjADwoq+xAPCjgIoA8KOKuADwo42fAPCjjpMA8KOOnADwo4+DAPCjj5UA8KORrQDwo5qjAPCjoqcA8KOqjQDwo6u6APCjsrwA8KO0ngDwo7uRAPCjvZ4A8KO+jgDwpImjAPCki64A8KSOqwDwpJiIAPCknLUA8KSglADwpLC2APCkspIA8KS+oQDwpL64APClgYQA8KWDsgDwpYOzAPClhJkA8KWEswDwpYmJAPClkJ0A8KWYpgDwpZqaAPClm4UA8KWlvADwpaqnAPClrqsA8KWygADwpbOQAPClvoYA8KaHmgDwpoioAPCmiYcA8KaLmQDwpoy+APCmk5oA8KaUowDwppaoAPCmnqcA8KaetQDwpqy8APCmsLYA8KazlQDwprWrAPCmvKwA8Ka+sQDwp4OSAPCnj4oA8KeZpwDwp6KuAPCnpaYA8KeyqADwp7uTAPCnvK8A8KiXkgDwqJetAPConK4A8KivugDwqLW3APCphYUA8KmHnwDwqYiaAPCpkIoA8KmSlgDwqZa2APCprLAA8KqDjgDwqoSFAPCqiI4A8KqKkQDwqo6SAPCqmIAA" - }, - { - "type": "Replace", - "pattern": { - "Regex": " {2,}" - }, - "content": "▁" - } - ] - }, - "pre_tokenizer": { - "type": "Metaspace", - "replacement": "▁", - "prepend_scheme": "always", - "split": true - }, - "post_processor": null, - "decoder": { - "type": "Metaspace", - "replacement": "▁", - "prepend_scheme": "always", - "split": true - }, - "model": { - "type": "BPE", - "dropout": null, - "unk_token": "", - "continuing_subword_prefix": null, - "end_of_word_suffix": null, - "fuse_unk": true, - "byte_fallback": true, - "ignore_merges": false, - "vocab": { - "": 0, - "▁t": 1, - "▁th": 2, - "▁a": 3, - "in": 4, - "re": 5, - "▁the": 6, - "▁w": 7, - "▁s": 8, - "▁o": 9, - "er": 10, - "ou": 11, - "at": 12, - "nd": 13, - "it": 14, - "▁h": 15, - "▁c": 16, - "▁b": 17, - "is": 18, - "en": 19, - "on": 20, - "ing": 21, - "▁f": 22, - "▁to": 23, - "▁m": 24, - "es": 25, - "▁p": 26, - "or": 27, - "an": 28, - "▁d": 29, - "ll": 30, - "▁I": 31, - "ed": 32, - "▁and": 33, - "▁l": 34, - "▁of": 35, - "▁in": 36, - "▁y": 37, - "ar": 38, - "▁g": 39, - "▁you": 40, - "as": 41, - "om": 42, - "▁n": 43, - "ve": 44, - "▁that": 45, - "le": 46, - "ic": 47, - "us": 48, - "ow": 49, - "et": 50, - "al": 51, - "▁e": 52, - "ut": 53, - "▁it": 54, - "ot": 55, - "▁be": 56, - "▁T": 57, - "ion": 58, - "▁is": 59, - "▁wh": 60, - "▁re": 61, - "▁on": 62, - "▁we": 63, - "ent": 64, - "▁A": 65, - "ay": 66, - "▁ha": 67, - "▁Th": 68, - "id": 69, - "▁S": 70, - "ac": 71, - "gh": 72, - "ver": 73, - "ke": 74, - "▁for": 75, - "im": 76, - "ly": 77, - "ur": 78, - "ld": 79, - "▁he": 80, - "▁st": 81, - "all": 82, - "ro": 83, - "st": 84, - "se": 85, - "ct": 86, - "ith": 87, - "ir": 88, - "am": 89, - "▁this": 90, - "if": 91, - "▁W": 92, - "oo": 93, - "ri": 94, - "▁was": 95, - "ght": 96, - "▁u": 97, - "▁with": 98, - "ad": 99, - "ch": 100, - "▁se": 101, - "▁k": 102, - "▁an": 103, - "▁The": 104, - "▁li": 105, - "▁do": 106, - "▁B": 107, - "▁have": 108, - "▁as": 109, - "th": 110, - "▁are": 111, - "▁sh": 112, - "ust": 113, - "ce": 114, - "ally": 115, - "ill": 116, - "▁H": 117, - "▁j": 118, - "ter": 119, - "▁go": 120, - "▁And": 121, - "ation": 122, - "▁C": 123, - "▁so": 124, - "ome": 125, - "▁not": 126, - "op": 127, - "il": 128, - "ore": 129, - "▁ne": 130, - "▁can": 131, - "▁me": 132, - "▁at": 133, - "ould": 134, - "ant": 135, - "▁M": 136, - "▁like": 137, - "ere": 138, - "▁they": 139, - "ra": 140, - "ers": 141, - "▁ab": 142, - "▁de": 143, - "▁kn": 144, - "ge": 145, - "▁Y": 146, - "▁ch": 147, - "ul": 148, - "pp": 149, - "▁or": 150, - "▁al": 151, - "▁con": 152, - "▁com": 153, - "ess": 154, - "▁su": 155, - "out": 156, - "▁your": 157, - "▁So": 158, - "ate": 159, - "▁one": 160, - "▁all": 161, - "▁ex": 162, - "est": 163, - "▁fr": 164, - "▁just": 165, - "▁pro": 166, - "▁know": 167, - "▁O": 168, - "ain": 169, - "▁but": 170, - "ol": 171, - "ive": 172, - "▁v": 173, - "use": 174, - "very": 175, - "art": 176, - "qu": 177, - "▁my": 178, - "el": 179, - "▁N": 180, - "nt": 181, - "▁It": 182, - "▁what": 183, - "ab": 184, - "▁P": 185, - "▁wor": 186, - "▁out": 187, - "▁there": 188, - "▁up": 189, - "um": 190, - "▁from": 191, - "pe": 192, - "▁tw": 193, - "▁r": 194, - "and": 195, - "ight": 196, - "ort": 197, - "un": 198, - "▁L": 199, - "ist": 200, - "▁about": 201, - "ide": 202, - "ig": 203, - "ake": 204, - "▁D": 205, - "em": 206, - "os": 207, - "king": 208, - "rou": 209, - "ind": 210, - "our": 211, - "res": 212, - "▁We": 213, - "▁get": 214, - "▁E": 215, - "▁G": 216, - "ack": 217, - "▁le": 218, - "ity": 219, - "od": 220, - "▁F": 221, - "ard": 222, - "▁pl": 223, - "▁our": 224, - "▁int": 225, - "ment": 226, - "▁will": 227, - "ies": 228, - "▁by": 229, - "ink": 230, - "ca": 231, - "▁if": 232, - "red": 233, - "her": 234, - "ie": 235, - "▁us": 236, - "▁some": 237, - "▁don": 238, - "ven": 239, - "ood": 240, - "ast": 241, - "▁R": 242, - "▁his": 243, - "▁tim": 244, - "▁tr": 245, - "▁more": 246, - "ich": 247, - "ous": 248, - "ame": 249, - "▁going": 250, - "▁had": 251, - "▁them": 252, - "ook": 253, - "▁pe": 254, - "▁Wh": 255, - "▁You": 256, - "▁But": 257, - "ine": 258, - "▁here": 259, - "▁would": 260, - "cause": 261, - "right": 262, - "so": 263, - "ost": 264, - "ure": 265, - "▁has": 266, - "ect": 267, - "▁think": 268, - "▁fe": 269, - "ong": 270, - "▁see": 271, - "▁when": 272, - "▁who": 273, - "▁were": 274, - "▁really": 275, - "▁their": 276, - "▁want": 277, - "one": 278, - "ople": 279, - "▁then": 280, - "▁time": 281, - "▁sa": 282, - "ap": 283, - "▁te": 284, - "▁He": 285, - "▁ye": 286, - "ck": 287, - "▁her": 288, - "▁thing": 289, - "▁right": 290, - "▁which": 291, - "itt": 292, - "ice": 293, - "act": 294, - "▁people": 295, - "ty": 296, - "▁two": 297, - "▁J": 298, - "▁im": 299, - "ther": 300, - "ci": 301, - "ose": 302, - "▁cl": 303, - "▁qu": 304, - "▁man": 305, - "▁also": 306, - "ree": 307, - "▁en": 308, - "ud": 309, - "▁how": 310, - "reat": 311, - "ak": 312, - "hing": 313, - "ag": 314, - "▁any": 315, - "ff": 316, - "ace": 317, - "per": 318, - "▁because": 319, - "▁very": 320, - "own": 321, - "▁ad": 322, - "▁act": 323, - "▁been": 324, - "▁now": 325, - "▁ag": 326, - "▁into": 327, - "▁comp": 328, - "ars": 329, - "ions": 330, - "are": 331, - "ite": 332, - "iv": 333, - "▁these": 334, - "ays": 335, - "ep": 336, - "▁This": 337, - "▁she": 338, - "ans": 339, - "ah": 340, - "een": 341, - "▁over": 342, - "ry": 343, - "▁lo": 344, - "age": 345, - "▁pr": 346, - "▁sp": 347, - "ue": 348, - "▁co": 349, - "ick": 350, - "ber": 351, - "▁did": 352, - "ip": 353, - "ach": 354, - "▁back": 355, - "▁no": 356, - "▁cont": 357, - "▁other": 358, - "▁every": 359, - "pt": 360, - "▁need": 361, - "▁him": 362, - "▁U": 363, - "▁In": 364, - "▁work": 365, - "irst": 366, - "▁part": 367, - "▁look": 368, - "ittle": 369, - "ble": 370, - "iz": 371, - "▁un": 372, - "▁make": 373, - "omet": 374, - "nder": 375, - "ish": 376, - "na": 377, - "▁little": 378, - "▁off": 379, - "▁than": 380, - "▁got": 381, - "ually": 382, - "▁per": 383, - "▁good": 384, - "▁way": 385, - "▁could": 386, - "▁ac": 387, - "▁imp": 388, - "able": 389, - "▁where": 390, - "iff": 391, - "▁That": 392, - "▁res": 393, - "ount": 394, - "pl": 395, - "ance": 396, - "▁first": 397, - "▁ro": 398, - "▁pre": 399, - "ass": 400, - "▁say": 401, - "int": 402, - "ated": 403, - "ire": 404, - "uch": 405, - "ase": 406, - "▁somet": 407, - "ound": 408, - "▁down": 409, - "▁diff": 410, - "sel": 411, - "▁gu": 412, - "▁am": 413, - "ress": 414, - "▁lot": 415, - "ence": 416, - "▁dis": 417, - "orm": 418, - "ix": 419, - "▁po": 420, - "ving": 421, - "enty": 422, - "▁K": 423, - "▁spe": 424, - "und": 425, - "he": 426, - "▁much": 427, - "▁ar": 428, - "round": 429, - "▁app": 430, - "co": 431, - "ark": 432, - "▁new": 433, - "ater": 434, - "ult": 435, - "end": 436, - "▁even": 437, - "▁start": 438, - "ations": 439, - "rough": 440, - "ile": 441, - "fter": 442, - "▁well": 443, - "be": 444, - "▁They": 445, - "▁three": 446, - "ign": 447, - "ild": 448, - "▁said": 449, - "ough": 450, - "ang": 451, - "▁too": 452, - "ade": 453, - "▁bl": 454, - "ens": 455, - "▁inc": 456, - "ia": 457, - "▁those": 458, - "▁mo": 459, - "▁take": 460, - "▁through": 461, - "▁fl": 462, - "▁kind": 463, - "▁things": 464, - "▁bet": 465, - "▁only": 466, - "▁St": 467, - "▁let": 468, - "cess": 469, - "▁Ch": 470, - "ary": 471, - "vel": 472, - "▁If": 473, - "xt": 474, - "other": 475, - "av": 476, - "ical": 477, - "ord": 478, - "▁again": 479, - "▁something": 480, - "onna": 481, - "fore": 482, - "▁may": 483, - "ting": 484, - "▁bu": 485, - "▁differe": 486, - "urn": 487, - "▁gonna": 488, - "▁does": 489, - "uct": 490, - "og": 491, - "▁twenty": 492, - "▁gr": 493, - "▁Ye": 494, - "wn": 495, - "▁should": 496, - "▁comm": 497, - "ition": 498, - "▁under": 499, - "▁hel": 500, - "ory": 501, - "▁fo": 502, - "▁use": 503, - "igh": 504, - "ife": 505, - "▁actually": 506, - "▁tal": 507, - "▁call": 508, - "ents": 509, - "ious": 510, - "ull": 511, - "▁There": 512, - "▁Yeah": 513, - "▁most": 514, - "▁ke": 515, - "ors": 516, - "ved": 517, - "ys": 518, - "▁sc": 519, - "▁happ": 520, - "ope": 521, - "▁help": 522, - "atch": 523, - "▁What": 524, - "▁rem": 525, - "ple": 526, - "▁Now": 527, - "▁br": 528, - "ool": 529, - "oth": 530, - "▁four": 531, - "self": 532, - "▁str": 533, - "ne": 534, - "thing": 535, - "▁put": 536, - "ial": 537, - "▁great": 538, - "ail": 539, - "ub": 540, - "ning": 541, - "▁sm": 542, - "▁feel": 543, - "▁five": 544, - "ody": 545, - "undred": 546, - "iss": 547, - "ank": 548, - "get": 549, - "aking": 550, - "▁many": 551, - "▁hundred": 552, - "▁years": 553, - "▁being": 554, - "▁come": 555, - "▁mean": 556, - "ily": 557, - "▁different": 558, - "▁after": 559, - "▁ser": 560, - "▁show": 561, - "form": 562, - "ful": 563, - "oy": 564, - "▁six": 565, - "▁vide": 566, - "▁V": 567, - "▁its": 568, - "▁point": 569, - "▁day": 570, - "▁des": 571, - "ons": 572, - "▁bit": 573, - "▁bel": 574, - "▁before": 575, - "▁aw": 576, - "▁end": 577, - "▁Oh": 578, - "▁still": 579, - "ath": 580, - "▁long": 581, - "▁'": 582, - "ise": 583, - "ob": 584, - "day": 585, - "▁add": 586, - "ft": 587, - "ves": 588, - "ces": 589, - "ady": 590, - "▁cr": 591, - "▁around": 592, - "▁try": 593, - "les": 594, - "vers": 595, - "kay": 596, - "ian": 597, - "ates": 598, - "▁find": 599, - "ward": 600, - "▁As": 601, - "▁eight": 602, - "lic": 603, - "▁same": 604, - "▁pos": 605, - "▁em": 606, - "▁made": 607, - "▁supp": 608, - "▁life": 609, - "▁Be": 610, - "pect": 611, - "▁dec": 612, - "▁play": 613, - "ange": 614, - "▁att": 615, - "▁pers": 616, - "ways": 617, - "▁high": 618, - "▁hand": 619, - "▁next": 620, - "▁cons": 621, - "▁own": 622, - "▁inv": 623, - "ower": 624, - "▁ind": 625, - "ert": 626, - "ng": 627, - "ave": 628, - "▁year": 629, - "▁big": 630, - "ating": 631, - "▁world": 632, - "▁rel": 633, - "▁sure": 634, - "▁tra": 635, - "ew": 636, - "ered": 637, - "▁fin": 638, - "▁Well": 639, - "▁sl": 640, - "▁doing": 641, - "bs": 642, - "▁set": 643, - "▁rec": 644, - "ual": 645, - "cial": 646, - "▁ph": 647, - "erm": 648, - "▁love": 649, - "ph": 650, - "▁real": 651, - "▁last": 652, - "ict": 653, - "▁bo": 654, - "▁ra": 655, - "ible": 656, - "▁wr": 657, - "mer": 658, - "▁count": 659, - "ities": 660, - "▁always": 661, - "inet": 662, - "ments": 663, - "uc": 664, - "▁might": 665, - "▁inter": 666, - "▁video": 667, - "gin": 668, - "▁tell": 669, - "▁never": 670, - "vent": 671, - "▁import": 672, - "ied": 673, - "▁sy": 674, - "▁How": 675, - "ically": 676, - "ought": 677, - "▁thir": 678, - "▁rep": 679, - "ks": 680, - "ib": 681, - "▁fam": 682, - "ject": 683, - "▁bas": 684, - "▁She": 685, - "▁give": 686, - "akes": 687, - "▁ninet": 688, - "▁reg": 689, - "▁min": 690, - "▁op": 691, - "▁def": 692, - "▁didn": 693, - "te": 694, - "▁cour": 695, - "▁why": 696, - "▁ent": 697, - "▁place": 698, - "▁ins": 699, - "▁car": 700, - "ather": 701, - "▁person": 702, - "ular": 703, - "▁inst": 704, - "▁prod": 705, - "lect": 706, - "▁Al": 707, - "▁today": 708, - "▁bec": 709, - "▁sur": 710, - "▁All": 711, - "▁another": 712, - "▁bus": 713, - "▁keep": 714, - "ell": 715, - "ese": 716, - "riend": 717, - "▁quest": 718, - "▁talk": 719, - "als": 720, - "ings": 721, - "▁mon": 722, - "cond": 723, - "old": 724, - "▁acc": 725, - "▁la": 726, - "▁num": 727, - "ident": 728, - "▁che": 729, - "iness": 730, - "▁turn": 731, - "▁ear": 732, - "▁No": 733, - "ousand": 734, - "▁better": 735, - "ific": 736, - "▁loo": 737, - "▁gl": 738, - "oc": 739, - "▁important": 740, - "ited": 741, - "▁An": 742, - "▁thousand": 743, - "ility": 744, - "llow": 745, - "▁used": 746, - "▁gen": 747, - "▁sim": 748, - "li": 749, - "▁happen": 750, - "▁Un": 751, - "▁Let": 752, - "air": 753, - "ock": 754, - "ably": 755, - "gg": 756, - "▁watch": 757, - "▁For": 758, - "▁sw": 759, - "ren": 760, - "ute": 761, - "ever": 762, - "▁pol": 763, - "▁sch": 764, - "▁When": 765, - "▁such": 766, - "▁fif": 767, - "▁home": 768, - "▁cle": 769, - "▁contin": 770, - "ouse": 771, - "▁friend": 772, - "uring": 773, - "▁Okay": 774, - "gr": 775, - "▁able": 776, - "▁stud": 777, - "▁eff": 778, - "hip": 779, - "body": 780, - "▁top": 781, - "ness": 782, - "▁exper": 783, - "▁pret": 784, - "▁both": 785, - "▁done": 786, - "cri": 787, - "▁mark": 788, - "▁while": 789, - "▁old": 790, - "ros": 791, - "ont": 792, - "▁second": 793, - "ative": 794, - "▁thought": 795, - "▁best": 796, - "▁found": 797, - "iew": 798, - "▁belie": 799, - "▁each": 800, - "erest": 801, - "▁tri": 802, - "▁eas": 803, - "▁ca": 804, - "▁fact": 805, - "▁care": 806, - "▁fun": 807, - "atter": 808, - "ures": 809, - "▁head": 810, - "▁lear": 811, - "▁water": 812, - "▁hard": 813, - "▁few": 814, - "▁side": 815, - "ween": 816, - "▁exp": 817, - "▁away": 818, - "its": 819, - "▁ext": 820, - "lud": 821, - "▁run": 822, - "▁trans": 823, - "ince": 824, - "▁sk": 825, - "▁open": 826, - "cus": 827, - "▁between": 828, - "▁called": 829, - "▁wee": 830, - "▁pretty": 831, - "ason": 832, - "▁far": 833, - "ember": 834, - "omm": 835, - "▁interest": 836, - "any": 837, - "ner": 838, - "uff": 839, - "▁pres": 840, - "▁cur": 841, - "▁child": 842, - "ee": 843, - "▁toget": 844, - "▁together": 845, - "olog": 846, - "▁God": 847, - "ond": 848, - "▁char": 849, - "▁looking": 850, - "stem": 851, - "az": 852, - "cent": 853, - "▁ob": 854, - "▁ass": 855, - "land": 856, - "▁doesn": 857, - "▁business": 858, - "▁course": 859, - "▁ten": 860, - "ps": 861, - "arch": 862, - "ced": 863, - "ms": 864, - "ize": 865, - "nce": 866, - "▁ref": 867, - "▁name": 868, - "ross": 869, - "▁grow": 870, - "oney": 871, - "▁went": 872, - "ics": 873, - "teen": 874, - "▁cou": 875, - "▁prob": 876, - "▁ret": 877, - "▁guys": 878, - "▁came": 879, - "ash": 880, - "led": 881, - "▁Eur": 882, - "ues": 883, - "▁ide": 884, - "gan": 885, - "▁everything": 886, - "▁getting": 887, - "▁ask": 888, - "▁cor": 889, - "▁build": 890, - "▁sign": 891, - "▁small": 892, - "uck": 893, - "▁el": 894, - "▁col": 895, - "▁Is": 896, - "ational": 897, - "stand": 898, - "cy": 899, - "▁conf": 900, - "der": 901, - "▁bre": 902, - "▁cap": 903, - "▁mod": 904, - "ets": 905, - "ike": 906, - "▁number": 907, - "▁comple": 908, - "ertain": 909, - "▁ever": 910, - "▁coll": 911, - "▁hum": 912, - "▁Europe": 913, - "▁cre": 914, - "▁met": 915, - "▁exam": 916, - "▁move": 917, - "▁pass": 918, - "▁left": 919, - "▁system": 920, - "▁includ": 921, - "▁Thank": 922, - "cept": 923, - "▁wom": 924, - "▁product": 925, - "ten": 926, - "▁rest": 927, - "▁probably": 928, - "▁dri": 929, - "▁Do": 930, - "▁gener": 931, - "▁anything": 932, - "▁lar": 933, - "▁My": 934, - "▁school": 935, - "▁lead": 936, - "▁sub": 937, - "▁ty": 938, - "▁plan": 939, - "▁seem": 940, - "▁whole": 941, - "irect": 942, - "▁light": 943, - "▁must": 944, - "▁mom": 945, - "▁opp": 946, - "▁support": 947, - "▁family": 948, - "ices": 949, - "amp": 950, - "▁proble": 951, - "▁dr": 952, - "ready": 953, - "▁using": 954, - "ense": 955, - "▁prov": 956, - "ush": 957, - "ax": 958, - "▁power": 959, - "▁Re": 960, - "alth": 961, - "▁ev": 962, - "▁stand": 963, - "▁war": 964, - "ts": 965, - "▁": 966, - "e": 967, - "t": 968, - "o": 969, - "a": 970, - "n": 971, - "i": 972, - "s": 973, - "r": 974, - "h": 975, - "l": 976, - "d": 977, - "u": 978, - "c": 979, - "m": 980, - "y": 981, - "g": 982, - "w": 983, - "f": 984, - "p": 985, - ".": 986, - "b": 987, - ",": 988, - "v": 989, - "k": 990, - "'": 991, - "I": 992, - "T": 993, - "A": 994, - "S": 995, - "x": 996, - "W": 997, - "j": 998, - "B": 999, - "C": 1000, - "H": 1001, - "?": 1002, - "M": 1003, - "O": 1004, - "Y": 1005, - "N": 1006, - "P": 1007, - "E": 1008, - "q": 1009, - "L": 1010, - "D": 1011, - "z": 1012, - "G": 1013, - "F": 1014, - "R": 1015, - "!": 1016, - "J": 1017, - "U": 1018, - "K": 1019, - "V": 1020, - "Q": 1021, - "Z": 1022, - "X": 1023 - }, - "merges": [ - "▁ t", - "▁t h", - "▁ th", - "▁ a", - "i n", - "r e", - "▁th e", - "▁t he", - "▁ w", - "▁ s", - "▁ o", - "e r", - "o u", - "a t", - "n d", - "i t", - "▁ h", - "▁ c", - "▁ b", - "i s", - "e n", - "o n", - "in g", - "i ng", - "▁ f", - "▁t o", - "▁ m", - "e s", - "▁ p", - "o r", - "a n", - "▁ d", - "l l", - "▁ I", - "e d", - "▁an d", - "▁a nd", - "▁ and", - "▁ l", - "▁o f", - "▁ in", - "▁ y", - "a r", - "▁ g", - "▁y ou", - "a s", - "o m", - "▁ n", - "v e", - "▁th at", - "l e", - "i c", - "u s", - "o w", - "e t", - "a l", - "▁ e", - "u t", - "▁ it", - "o t", - "▁b e", - "▁ be", - "▁ T", - "i on", - "▁ is", - "▁w h", - "▁r e", - "▁ re", - "▁o n", - "▁ on", - "▁w e", - "en t", - "e nt", - "▁ A", - "a y", - "▁h a", - "▁T h", - "i d", - "▁ S", - "a c", - "g h", - "ve r", - "v er", - "k e", - "▁fo r", - "▁f or", - "i m", - "l y", - "u r", - "l d", - "▁h e", - "▁ he", - "▁s t", - "▁ st", - "al l", - "a ll", - "r o", - "s t", - "s e", - "c t", - "it h", - "i th", - "i r", - "a m", - "▁th is", - "i f", - "▁ W", - "o o", - "r i", - "▁w as", - "gh t", - "▁ u", - "▁w ith", - "a d", - "c h", - "▁s e", - "▁ se", - "▁ k", - "▁a n", - "▁ an", - "▁Th e", - "▁T he", - "▁l i", - "▁ li", - "▁d o", - "▁ B", - "▁ha ve", - "▁h ave", - "▁a s", - "▁ as", - "t h", - "▁ar e", - "▁a re", - "▁ are", - "▁s h", - "us t", - "u st", - "c e", - "all y", - "al ly", - "il l", - "i ll", - "▁ H", - "▁ j", - "te r", - "t er", - "▁g o", - "▁An d", - "▁A nd", - "at ion", - "▁ C", - "▁s o", - "▁ so", - "om e", - "▁no t", - "▁n ot", - "o p", - "i l", - "or e", - "o re", - "▁n e", - "▁ ne", - "▁ca n", - "▁c an", - "▁m e", - "▁a t", - "▁ at", - "ou ld", - "an t", - "a nt", - "▁ M", - "▁li ke", - "▁l ike", - "er e", - "e re", - "▁the y", - "r a", - "er s", - "▁a b", - "▁ ab", - "▁d e", - "▁k n", - "g e", - "▁ Y", - "▁c h", - "▁ ch", - "u l", - "p p", - "▁o r", - "▁ or", - "▁a l", - "▁ al", - "▁co n", - "▁c on", - "▁co m", - "▁c om", - "es s", - "▁s u", - "ou t", - "o ut", - "▁you r", - "▁y our", - "▁S o", - "at e", - "a te", - "▁on e", - "▁o ne", - "▁ one", - "▁al l", - "▁a ll", - "▁ all", - "▁e x", - "es t", - "e st", - "▁f r", - "▁j ust", - "▁pr o", - "▁p ro", - "▁kn ow", - "▁ O", - "a in", - "▁bu t", - "▁b ut", - "o l", - "iv e", - "i ve", - "▁ v", - "us e", - "u se", - "ver y", - "ve ry", - "ar t", - "q u", - "▁m y", - "e l", - "▁ N", - "n t", - "▁I t", - "▁wh at", - "a b", - "▁ P", - "▁w or", - "▁o ut", - "▁ out", - "▁the re", - "▁th ere", - "▁u p", - "u m", - "▁fr om", - "p e", - "▁t w", - "▁ r", - "an d", - "a nd", - "igh t", - "i ght", - "or t", - "u n", - "▁ L", - "is t", - "i st", - "▁ab out", - "id e", - "i g", - "ak e", - "a ke", - "▁ D", - "e m", - "o s", - "k ing", - "ro u", - "r ou", - "in d", - "i nd", - "ou r", - "o ur", - "re s", - "r es", - "▁W e", - "▁g et", - "▁ get", - "▁ E", - "▁ G", - "ac k", - "a ck", - "▁l e", - "▁ le", - "it y", - "i ty", - "o d", - "▁ F", - "ar d", - "▁p l", - "▁ pl", - "▁o ur", - "▁ our", - "▁in t", - "▁ int", - "m ent", - "▁w ill", - "ie s", - "i es", - "▁b y", - "in k", - "c a", - "▁ if", - "re d", - "r ed", - "he r", - "h er", - "i e", - "▁u s", - "▁ us", - "▁s ome", - "▁do n", - "▁d on", - "ve n", - "v en", - "oo d", - "o od", - "as t", - "a st", - "▁ R", - "▁h is", - "▁t im", - "▁t r", - "▁mo re", - "▁m ore", - "ic h", - "i ch", - "ou s", - "o us", - "am e", - "▁go ing", - "▁ha d", - "▁h ad", - "▁the m", - "▁th em", - "oo k", - "▁p e", - "▁ pe", - "▁W h", - "▁Y ou", - "▁B ut", - "in e", - "i ne", - "▁her e", - "▁he re", - "▁h ere", - "▁w ould", - "ca use", - "ri ght", - "r ight", - "s o", - "os t", - "o st", - "ur e", - "u re", - "▁ha s", - "▁h as", - "e ct", - "▁th ink", - "▁f e", - "on g", - "o ng", - "▁se e", - "▁s ee", - "▁wh en", - "▁wh o", - "▁we re", - "▁w ere", - "▁real ly", - "▁re ally", - "▁the ir", - "▁w ant", - "on e", - "o ne", - "op le", - "o ple", - "▁the n", - "▁th en", - "▁tim e", - "▁s a", - "a p", - "▁t e", - "▁ te", - "▁H e", - "▁y e", - "c k", - "▁he r", - "▁h er", - "▁ her", - "▁th ing", - "▁t hing", - "▁ thing", - "▁r ight", - "▁ right", - "▁wh ich", - "it t", - "ic e", - "i ce", - "ac t", - "a ct", - "▁pe ople", - "t y", - "▁tw o", - "▁ J", - "▁ im", - "th er", - "t her", - "c i", - "os e", - "o se", - "▁c l", - "▁ qu", - "▁m an", - "▁al so", - "re e", - "r ee", - "▁e n", - "▁ en", - "u d", - "▁h ow", - "re at", - "a k", - "h ing", - "a g", - "▁an y", - "▁ any", - "f f", - "ac e", - "a ce", - "pe r", - "p er", - "▁be cause", - "▁ very", - "ow n", - "o wn", - "▁a d", - "▁ ad", - "▁ac t", - "▁a ct", - "▁ act", - "▁be en", - "▁b een", - "▁no w", - "▁n ow", - "▁a g", - "▁ ag", - "▁int o", - "▁com p", - "ar s", - "ion s", - "i ons", - "ar e", - "a re", - "it e", - "i te", - "i v", - "▁the se", - "▁th ese", - "ay s", - "a ys", - "e p", - "▁Th is", - "▁sh e", - "▁s he", - "an s", - "a h", - "ee n", - "e en", - "▁o ver", - "r y", - "▁l o", - "ag e", - "a ge", - "▁p r", - "▁s p", - "u e", - "▁c o", - "▁ co", - "ic k", - "i ck", - "be r", - "b er", - "▁d id", - "i p", - "ac h", - "a ch", - "▁b ack", - "▁n o", - "▁con t", - "▁co nt", - "▁c ont", - "▁o ther", - "▁ other", - "▁ever y", - "▁e very", - "p t", - "▁ne ed", - "▁h im", - "▁ U", - "▁I n", - "▁wor k", - "ir st", - "▁p art", - "▁loo k", - "▁l ook", - "itt le", - "b le", - "i z", - "▁u n", - "▁ un", - "▁m ake", - "ome t", - "om et", - "nd er", - "n der", - "is h", - "n a", - "▁l ittle", - "▁of f", - "▁o ff", - "▁th an", - "▁go t", - "▁g ot", - "ual ly", - "u ally", - "▁pe r", - "▁p er", - "▁ per", - "▁go od", - "▁g ood", - "▁w ay", - "▁cou ld", - "▁c ould", - "▁a c", - "▁ ac", - "▁im p", - "ab le", - "a ble", - "▁wh ere", - "if f", - "i ff", - "▁Th at", - "▁re s", - "▁r es", - "▁ res", - "ou nt", - "p l", - "an ce", - "a nce", - "▁f irst", - "▁r o", - "▁ ro", - "▁pr e", - "▁p re", - "as s", - "▁sa y", - "▁s ay", - "in t", - "i nt", - "ate d", - "at ed", - "ir e", - "i re", - "uc h", - "u ch", - "as e", - "a se", - "▁some t", - "▁s omet", - "ou nd", - "o und", - "▁do wn", - "▁d own", - "▁d iff", - "se l", - "s el", - "▁g u", - "▁a m", - "▁ am", - "res s", - "r ess", - "▁lo t", - "▁l ot", - "en ce", - "e nce", - "▁d is", - "or m", - "i x", - "▁p o", - "v ing", - "ent y", - "en ty", - "▁ K", - "▁sp e", - "▁s pe", - "un d", - "u nd", - "h e", - "▁m uch", - "▁a r", - "▁ ar", - "rou nd", - "ro und", - "r ound", - "▁a pp", - "c o", - "ar k", - "▁ne w", - "▁n ew", - "ate r", - "at er", - "a ter", - "ul t", - "en d", - "e nd", - "▁ev en", - "▁e ven", - "▁st art", - "ation s", - "at ions", - "rou gh", - "r ough", - "il e", - "i le", - "ft er", - "f ter", - "▁we ll", - "▁w ell", - "b e", - "▁The y", - "▁th ree", - "ig n", - "il d", - "i ld", - "▁sa id", - "ou gh", - "an g", - "a ng", - "▁to o", - "▁t oo", - "ad e", - "▁b l", - "en s", - "▁in c", - "i a", - "▁th ose", - "▁m o", - "▁t ake", - "▁th rough", - "▁f l", - "▁k ind", - "▁thing s", - "▁th ings", - "▁be t", - "▁b et", - "▁on ly", - "▁S t", - "▁le t", - "▁l et", - "ces s", - "c ess", - "▁C h", - "ar y", - "a ry", - "ve l", - "v el", - "▁I f", - "x t", - "oth er", - "ot her", - "o ther", - "a v", - "ic al", - "or d", - "▁ag ain", - "▁somet hing", - "▁some thing", - "on na", - "f ore", - "▁m ay", - "t ing", - "▁b u", - "▁diff ere", - "ur n", - "▁g onna", - "▁do es", - "uc t", - "u ct", - "o g", - "▁tw enty", - "▁g r", - "▁ gr", - "▁Y e", - "w n", - "▁sh ould", - "▁com m", - "▁c omm", - "it ion", - "▁un der", - "▁u nder", - "▁he l", - "▁h el", - "or y", - "o ry", - "▁f o", - "▁us e", - "▁u se", - "▁ use", - "ig h", - "i gh", - "if e", - "▁act ually", - "▁t al", - "▁ca ll", - "▁c all", - "ent s", - "en ts", - "i ous", - "ul l", - "u ll", - "▁The re", - "▁Th ere", - "▁Ye ah", - "▁mo st", - "▁m ost", - "▁k e", - "▁ ke", - "or s", - "ve d", - "v ed", - "y s", - "▁s c", - "▁ha pp", - "op e", - "o pe", - "▁hel p", - "at ch", - "▁Wh at", - "▁re m", - "▁r em", - "pl e", - "p le", - "▁No w", - "▁N ow", - "▁b r", - "oo l", - "o ol", - "ot h", - "o th", - "▁fo ur", - "▁f our", - "sel f", - "▁st r", - "n e", - "th ing", - "t hing", - "▁p ut", - "ia l", - "i al", - "▁g reat", - "a il", - "u b", - "n ing", - "▁s m", - "▁fe el", - "▁f ive", - "od y", - "und red", - "is s", - "an k", - "ge t", - "g et", - "ak ing", - "a king", - "▁man y", - "▁m any", - "▁h undred", - "▁year s", - "▁ye ars", - "▁be ing", - "▁com e", - "▁c ome", - "▁me an", - "il y", - "i ly", - "▁differe nt", - "▁a fter", - "▁se r", - "▁s er", - "▁sh ow", - "f orm", - "f ul", - "o y", - "▁s ix", - "▁v ide", - "▁ V", - "▁it s", - "▁ its", - "▁po int", - "▁d ay", - "▁ day", - "▁de s", - "▁d es", - "on s", - "▁b it", - "▁be l", - "▁b el", - "▁be fore", - "▁a w", - "▁en d", - "▁e nd", - "▁ end", - "▁O h", - "▁st ill", - "at h", - "a th", - "▁lo ng", - "▁l ong", - "▁ '", - "is e", - "i se", - "o b", - "d ay", - "▁ad d", - "f t", - "ve s", - "v es", - "ce s", - "c es", - "ad y", - "▁c r", - "▁ar ound", - "▁a round", - "▁tr y", - "▁t ry", - "le s", - "l es", - "ver s", - "v ers", - "k ay", - "ia n", - "i an", - "ate s", - "at es", - "▁fin d", - "▁f ind", - "w ard", - "▁A s", - "▁e ight", - "li c", - "l ic", - "▁s ame", - "▁po s", - "▁p os", - "▁e m", - "▁ em", - "▁m ade", - "▁su pp", - "▁l ife", - "▁B e", - "pe ct", - "p ect", - "▁de c", - "▁pl ay", - "ang e", - "an ge", - "▁at t", - "▁per s", - "▁p ers", - "w ays", - "▁h igh", - "▁ha nd", - "▁h and", - "▁ne xt", - "▁con s", - "▁c ons", - "▁o wn", - "▁ own", - "▁in v", - "ow er", - "▁in d", - "▁ ind", - "er t", - "n g", - "av e", - "a ve", - "▁ye ar", - "▁b ig", - "at ing", - "a ting", - "▁wor ld", - "▁re l", - "▁r el", - "▁sur e", - "▁su re", - "▁s ure", - "▁tr a", - "▁t ra", - "e w", - "ere d", - "er ed", - "e red", - "▁f in", - "▁We ll", - "▁W ell", - "▁s l", - "▁do ing", - "b s", - "▁se t", - "▁s et", - "▁re c", - "u al", - "ci al", - "c ial", - "▁p h", - "▁ ph", - "er m", - "▁lo ve", - "p h", - "▁re al", - "▁la st", - "▁l ast", - "ic t", - "i ct", - "▁b o", - "▁r a", - "▁ ra", - "ib le", - "i ble", - "▁w r", - "m er", - "▁cou nt", - "▁c ount", - "it ies", - "▁al ways", - "ine t", - "in et", - "ment s", - "m ents", - "u c", - "▁m ight", - "▁int er", - "▁in ter", - "▁vide o", - "g in", - "▁te ll", - "▁t ell", - "▁ne ver", - "▁n ever", - "ven t", - "ve nt", - "v ent", - "▁imp ort", - "ie d", - "i ed", - "▁s y", - "▁H ow", - "ical ly", - "ic ally", - "ough t", - "ou ght", - "▁th ir", - "▁re p", - "▁r ep", - "k s", - "i b", - "▁f am", - "j ect", - "▁b as", - "▁S he", - "▁g ive", - "ake s", - "ak es", - "▁n inet", - "▁re g", - "▁m in", - "▁o p", - "▁ op", - "▁de f", - "▁did n", - "t e", - "▁cou r", - "▁co ur", - "▁c our", - "▁wh y", - "▁en t", - "▁e nt", - "▁ ent", - "▁pl ace", - "▁in s", - "▁ca r", - "▁c ar", - "ath er", - "at her", - "a ther", - "▁pers on", - "ul ar", - "▁ins t", - "▁in st", - "▁pro d", - "▁pr od", - "le ct", - "l ect", - "▁A l", - "▁to day", - "▁be c", - "▁su r", - "▁s ur", - "▁Al l", - "▁A ll", - "▁an other", - "▁bu s", - "▁b us", - "▁ke ep", - "el l", - "e ll", - "es e", - "e se", - "ri end", - "▁qu est", - "▁tal k", - "al s", - "ing s", - "▁mo n", - "▁m on", - "co nd", - "c ond", - "ol d", - "o ld", - "▁ac c", - "▁l a", - "▁n um", - "ide nt", - "id ent", - "▁ch e", - "▁c he", - "in ess", - "i ness", - "▁t urn", - "▁e ar", - "▁N o", - "ous and", - "▁bet ter", - "if ic", - "▁lo o", - "▁l oo", - "▁g l", - "o c", - "▁import ant", - "ite d", - "it ed", - "▁A n", - "▁th ousand", - "il ity", - "ll ow", - "▁use d", - "▁us ed", - "▁g en", - "▁s im", - "l i", - "▁happ en", - "▁U n", - "▁L et", - "a ir", - "oc k", - "o ck", - "ab ly", - "g g", - "▁w atch", - "▁F or", - "▁s w", - "re n", - "r en", - "ut e", - "u te", - "e ver", - "▁po l", - "▁p ol", - "▁sc h", - "▁s ch", - "▁Wh en", - "▁su ch", - "▁s uch", - "▁f if", - "▁h ome", - "▁cl e", - "▁c le", - "▁cont in", - "ous e", - "ou se", - "o use", - "▁f riend", - "ur ing", - "▁O kay", - "g r", - "▁ab le", - "▁a ble", - "▁ able", - "▁st ud", - "▁e ff", - "h ip", - "b ody", - "▁to p", - "▁t op", - "n ess", - "▁exp er", - "▁ex per", - "▁pre t", - "▁pr et", - "▁bo th", - "▁b oth", - "▁don e", - "▁do ne", - "▁d one", - "c ri", - "▁m ark", - "▁wh ile", - "▁o ld", - "▁ old", - "ro s", - "r os", - "on t", - "o nt", - "▁se cond", - "at ive", - "▁th ought", - "▁be st", - "▁b est", - "▁fo und", - "▁f ound", - "ie w", - "i ew", - "▁bel ie", - "▁e ach", - "ere st", - "er est", - "▁tr i", - "▁t ri", - "▁e as", - "▁c a", - "▁ ca", - "▁f act", - "▁car e", - "▁ca re", - "▁c are", - "▁f un", - "at ter", - "ure s", - "ur es", - "u res", - "▁he ad", - "▁le ar", - "▁w ater", - "▁h ard", - "▁fe w", - "▁f ew", - "▁s ide", - "w een", - "▁ex p", - "▁aw ay", - "it s", - "i ts", - "▁ex t", - "▁e xt", - "l ud", - "▁r un", - "▁tr ans", - "in ce", - "i nce", - "▁s k", - "▁op en", - "c us", - "▁bet ween", - "▁call ed", - "▁we e", - "▁w ee", - "▁pret ty", - "as on", - "▁f ar", - "em ber", - "om m", - "▁inter est", - "▁int erest", - "an y", - "ne r", - "n er", - "u ff", - "▁pre s", - "▁pr es", - "▁p res", - "▁c ur", - "▁ch ild", - "e e", - "▁to get", - "▁toget her", - "ol og", - "▁G od", - "on d", - "o nd", - "▁ch ar", - "▁look ing", - "▁loo king", - "st em", - "a z", - "ce nt", - "c ent", - "▁o b", - "▁ ob", - "▁as s", - "▁ ass", - "l and", - "▁does n", - "▁bus iness", - "▁cour se", - "▁te n", - "▁t en", - "▁ ten", - "p s", - "ar ch", - "ce d", - "c ed", - "m s", - "iz e", - "n ce", - "▁re f", - "▁n ame", - "ros s", - "▁gr ow", - "one y", - "▁we nt", - "▁w ent", - "ic s", - "te en", - "t een", - "▁co u", - "▁c ou", - "▁pro b", - "▁pr ob", - "▁re t", - "▁r et", - "▁gu ys", - "▁c ame", - "as h", - "le d", - "l ed", - "▁E ur", - "ue s", - "u es", - "▁ ide", - "g an", - "▁every thing", - "▁get ting", - "▁as k", - "▁co r", - "▁c or", - "▁bu ild", - "▁s ign", - "▁sm all", - "uc k", - "u ck", - "▁e l", - "▁ el", - "▁co l", - "▁c ol", - "▁I s", - "ation al", - "st and", - "c y", - "▁con f", - "d er", - "▁br e", - "▁b re", - "▁ca p", - "▁c ap", - "▁mo d", - "▁m od", - "et s", - "e ts", - "i ke", - "▁num ber", - "▁comp le", - "▁com ple", - "ert ain", - "▁ev er", - "▁e ver", - "▁ ever", - "▁col l", - "▁co ll", - "▁h um", - "▁Eur ope", - "▁cr e", - "▁c re", - "▁me t", - "▁m et", - "▁ex am", - "▁mo ve", - "▁p ass", - "▁le ft", - "▁sy stem", - "▁inc lud", - "▁Th ank", - "ce pt", - "▁w om", - "▁prod uct", - "te n", - "t en", - "▁res t", - "▁re st", - "▁r est", - "▁prob ably", - "▁dr i", - "▁d ri", - "▁D o", - "▁gen er", - "▁any thing", - "▁la r", - "▁l ar", - "▁M y", - "▁sch ool", - "▁le ad", - "▁su b", - "▁s ub", - "▁t y", - "▁ ty", - "▁pl an", - "▁see m", - "▁se em", - "▁who le", - "ire ct", - "ir ect", - "▁li ght", - "▁l ight", - "▁m ust", - "▁mo m", - "▁m om", - "▁op p", - "▁o pp", - "▁supp ort", - "▁fam ily", - "ice s", - "ic es", - "i ces", - "am p", - "▁prob le", - "▁pro ble", - "▁d r", - "re ady", - "▁us ing", - "ens e", - "en se", - "▁pro v", - "us h", - "a x", - "▁p ower", - "▁R e", - "al th", - "▁e v", - "▁st and", - "▁ stand", - "▁w ar", - "t s" - ] - } -} \ No newline at end of file diff --git a/models/parakeet-ctc-110m-coreml/tokenizer_config.json b/models/parakeet-ctc-110m-coreml/tokenizer_config.json deleted file mode 100644 index d9eaa0fb393ab6e3ac6890fe99efb6191cded97c..0000000000000000000000000000000000000000 --- a/models/parakeet-ctc-110m-coreml/tokenizer_config.json +++ /dev/null @@ -1,27 +0,0 @@ -{ - "added_tokens_decoder": { - "0": { - "content": "", - "lstrip": false, - "normalized": false, - "rstrip": false, - "single_word": false, - "special": true - }, - "1024": { - "content": "", - "lstrip": false, - "normalized": false, - "rstrip": false, - "single_word": false, - "special": true - } - }, - "clean_up_tokenization_spaces": false, - "extra_special_tokens": {}, - "model_max_length": 1000000000000000019884624838656, - "pad_token": "", - "processor_class": "ParakeetProcessor", - "tokenizer_class": "PreTrainedTokenizer", - "unk_token": "" -} diff --git a/models/parakeet-ctc-110m-coreml/vocab.json b/models/parakeet-ctc-110m-coreml/vocab.json deleted file mode 100644 index 33b95831722563db5362641673b37885e25fb47a..0000000000000000000000000000000000000000 --- a/models/parakeet-ctc-110m-coreml/vocab.json +++ /dev/null @@ -1 +0,0 @@ -{"0": "", "1": "▁t", "2": "▁th", "3": "▁a", "4": "in", "5": "re", "6": "▁the", "7": "▁w", "8": "▁s", "9": "▁o", "10": "er", "11": "ou", "12": "at", "13": "nd", "14": "it", "15": "▁h", "16": "▁c", "17": "▁b", "18": "is", "19": "en", "20": "on", "21": "ing", "22": "▁f", "23": "▁to", "24": "▁m", "25": "es", "26": "▁p", "27": "or", "28": "an", "29": "▁d", "30": "ll", "31": "▁I", "32": "ed", "33": "▁and", "34": "▁l", "35": "▁of", "36": "▁in", "37": "▁y", "38": "ar", "39": "▁g", "40": "▁you", "41": "as", "42": "om", "43": "▁n", "44": "ve", "45": "▁that", "46": "le", "47": "ic", "48": "us", "49": "ow", "50": "et", "51": "al", "52": "▁e", "53": "ut", "54": "▁it", "55": "ot", "56": "▁be", "57": "▁T", "58": "ion", "59": "▁is", "60": "▁wh", "61": "▁re", "62": "▁on", "63": "▁we", "64": "ent", "65": "▁A", "66": "ay", "67": "▁ha", "68": "▁Th", "69": "id", "70": "▁S", "71": "ac", "72": "gh", "73": "ver", "74": "ke", "75": "▁for", "76": "im", "77": "ly", "78": "ur", "79": "ld", "80": "▁he", "81": "▁st", "82": "all", "83": "ro", "84": "st", "85": "se", "86": "ct", "87": "ith", "88": "ir", "89": "am", "90": "▁this", "91": "if", "92": "▁W", "93": "oo", "94": "ri", "95": "▁was", "96": "ght", "97": "▁u", "98": "▁with", "99": "ad", "100": "ch", "101": "▁se", "102": "▁k", "103": "▁an", "104": "▁The", "105": "▁li", "106": "▁do", "107": "▁B", "108": "▁have", "109": "▁as", "110": "th", "111": "▁are", "112": "▁sh", "113": "ust", "114": "ce", "115": "ally", "116": "ill", "117": "▁H", "118": "▁j", "119": "ter", "120": "▁go", "121": "▁And", "122": "ation", "123": "▁C", "124": "▁so", "125": "ome", "126": "▁not", "127": "op", "128": "il", "129": "ore", "130": "▁ne", "131": "▁can", "132": "▁me", "133": "▁at", "134": "ould", "135": "ant", "136": "▁M", "137": "▁like", "138": "ere", "139": "▁they", "140": "ra", "141": "ers", "142": "▁ab", "143": "▁de", "144": "▁kn", "145": "ge", "146": "▁Y", "147": "▁ch", "148": "ul", "149": "pp", "150": "▁or", "151": "▁al", "152": "▁con", "153": "▁com", "154": "ess", "155": "▁su", "156": "out", "157": "▁your", "158": "▁So", "159": "ate", "160": "▁one", "161": "▁all", "162": "▁ex", "163": "est", "164": "▁fr", "165": "▁just", "166": "▁pro", "167": "▁know", "168": "▁O", "169": "ain", "170": "▁but", "171": "ol", "172": "ive", "173": "▁v", "174": "use", "175": "very", "176": "art", "177": "qu", "178": "▁my", "179": "el", "180": "▁N", "181": "nt", "182": "▁It", "183": "▁what", "184": "ab", "185": "▁P", "186": "▁wor", "187": "▁out", "188": "▁there", "189": "▁up", "190": "um", "191": "▁from", "192": "pe", "193": "▁tw", "194": "▁r", "195": "and", "196": "ight", "197": "ort", "198": "un", "199": "▁L", "200": "ist", "201": "▁about", "202": "ide", "203": "ig", "204": "ake", "205": "▁D", "206": "em", "207": "os", "208": "king", "209": "rou", "210": "ind", "211": "our", "212": "res", "213": "▁We", "214": "▁get", "215": "▁E", "216": "▁G", "217": "ack", "218": "▁le", "219": "ity", "220": "od", "221": "▁F", "222": "ard", "223": "▁pl", "224": "▁our", "225": "▁int", "226": "ment", "227": "▁will", "228": "ies", "229": "▁by", "230": "ink", "231": "ca", "232": "▁if", "233": "red", "234": "her", "235": "ie", "236": "▁us", "237": "▁some", "238": "▁don", "239": "ven", "240": "ood", "241": "ast", "242": "▁R", "243": "▁his", "244": "▁tim", "245": "▁tr", "246": "▁more", "247": "ich", "248": "ous", "249": "ame", "250": "▁going", "251": "▁had", "252": "▁them", "253": "ook", "254": "▁pe", "255": "▁Wh", "256": "▁You", "257": "▁But", "258": "ine", "259": "▁here", "260": "▁would", "261": "cause", "262": "right", "263": "so", "264": "ost", "265": "ure", "266": "▁has", "267": "ect", "268": "▁think", "269": "▁fe", "270": "ong", "271": "▁see", "272": "▁when", "273": "▁who", "274": "▁were", "275": "▁really", "276": "▁their", "277": "▁want", "278": "one", "279": "ople", "280": "▁then", "281": "▁time", "282": "▁sa", "283": "ap", "284": "▁te", "285": "▁He", "286": "▁ye", "287": "ck", "288": "▁her", "289": "▁thing", "290": "▁right", "291": "▁which", "292": "itt", "293": "ice", "294": "act", "295": "▁people", "296": "ty", "297": "▁two", "298": "▁J", "299": "▁im", "300": "ther", "301": "ci", "302": "ose", "303": "▁cl", "304": "▁qu", "305": "▁man", "306": "▁also", "307": "ree", "308": "▁en", "309": "ud", "310": "▁how", "311": "reat", "312": "ak", "313": "hing", "314": "ag", "315": "▁any", "316": "ff", "317": "ace", "318": "per", "319": "▁because", "320": "▁very", "321": "own", "322": "▁ad", "323": "▁act", "324": "▁been", "325": "▁now", "326": "▁ag", "327": "▁into", "328": "▁comp", "329": "ars", "330": "ions", "331": "are", "332": "ite", "333": "iv", "334": "▁these", "335": "ays", "336": "ep", "337": "▁This", "338": "▁she", "339": "ans", "340": "ah", "341": "een", "342": "▁over", "343": "ry", "344": "▁lo", "345": "age", "346": "▁pr", "347": "▁sp", "348": "ue", "349": "▁co", "350": "ick", "351": "ber", "352": "▁did", "353": "ip", "354": "ach", "355": "▁back", "356": "▁no", "357": "▁cont", "358": "▁other", "359": "▁every", "360": "pt", "361": "▁need", "362": "▁him", "363": "▁U", "364": "▁In", "365": "▁work", "366": "irst", "367": "▁part", "368": "▁look", "369": "ittle", "370": "ble", "371": "iz", "372": "▁un", "373": "▁make", "374": "omet", "375": "nder", "376": "ish", "377": "na", "378": "▁little", "379": "▁off", "380": "▁than", "381": "▁got", "382": "ually", "383": "▁per", "384": "▁good", "385": "▁way", "386": "▁could", "387": "▁ac", "388": "▁imp", "389": "able", "390": "▁where", "391": "iff", "392": "▁That", "393": "▁res", "394": "ount", "395": "pl", "396": "ance", "397": "▁first", "398": "▁ro", "399": "▁pre", "400": "ass", "401": "▁say", "402": "int", "403": "ated", "404": "ire", "405": "uch", "406": "ase", "407": "▁somet", "408": "ound", "409": "▁down", "410": "▁diff", "411": "sel", "412": "▁gu", "413": "▁am", "414": "ress", "415": "▁lot", "416": "ence", "417": "▁dis", "418": "orm", "419": "ix", "420": "▁po", "421": "ving", "422": "enty", "423": "▁K", "424": "▁spe", "425": "und", "426": "he", "427": "▁much", "428": "▁ar", "429": "round", "430": "▁app", "431": "co", "432": "ark", "433": "▁new", "434": "ater", "435": "ult", "436": "end", "437": "▁even", "438": "▁start", "439": "ations", "440": "rough", "441": "ile", "442": "fter", "443": "▁well", "444": "be", "445": "▁They", "446": "▁three", "447": "ign", "448": "ild", "449": "▁said", "450": "ough", "451": "ang", "452": "▁too", "453": "ade", "454": "▁bl", "455": "ens", "456": "▁inc", "457": "ia", "458": "▁those", "459": "▁mo", "460": "▁take", "461": "▁through", "462": "▁fl", "463": "▁kind", "464": "▁things", "465": "▁bet", "466": "▁only", "467": "▁St", "468": "▁let", "469": "cess", "470": "▁Ch", "471": "ary", "472": "vel", "473": "▁If", "474": "xt", "475": "other", "476": "av", "477": "ical", "478": "ord", "479": "▁again", "480": "▁something", "481": "onna", "482": "fore", "483": "▁may", "484": "ting", "485": "▁bu", "486": "▁differe", "487": "urn", "488": "▁gonna", "489": "▁does", "490": "uct", "491": "og", "492": "▁twenty", "493": "▁gr", "494": "▁Ye", "495": "wn", "496": "▁should", "497": "▁comm", "498": "ition", "499": "▁under", "500": "▁hel", "501": "ory", "502": "▁fo", "503": "▁use", "504": "igh", "505": "ife", "506": "▁actually", "507": "▁tal", "508": "▁call", "509": "ents", "510": "ious", "511": "ull", "512": "▁There", "513": "▁Yeah", "514": "▁most", "515": "▁ke", "516": "ors", "517": "ved", "518": "ys", "519": "▁sc", "520": "▁happ", "521": "ope", "522": "▁help", "523": "atch", "524": "▁What", "525": "▁rem", "526": "ple", "527": "▁Now", "528": "▁br", "529": "ool", "530": "oth", "531": "▁four", "532": "self", "533": "▁str", "534": "ne", "535": "thing", "536": "▁put", "537": "ial", "538": "▁great", "539": "ail", "540": "ub", "541": "ning", "542": "▁sm", "543": "▁feel", "544": "▁five", "545": "ody", "546": "undred", "547": "iss", "548": "ank", "549": "get", "550": "aking", "551": "▁many", "552": "▁hundred", "553": "▁years", "554": "▁being", "555": "▁come", "556": "▁mean", "557": "ily", "558": "▁different", "559": "▁after", "560": "▁ser", "561": "▁show", "562": "form", "563": "ful", "564": "oy", "565": "▁six", "566": "▁vide", "567": "▁V", "568": "▁its", "569": "▁point", "570": "▁day", "571": "▁des", "572": "ons", "573": "▁bit", "574": "▁bel", "575": "▁before", "576": "▁aw", "577": "▁end", "578": "▁Oh", "579": "▁still", "580": "ath", "581": "▁long", "582": "▁'", "583": "ise", "584": "ob", "585": "day", "586": "▁add", "587": "ft", "588": "ves", "589": "ces", "590": "ady", "591": "▁cr", "592": "▁around", "593": "▁try", "594": "les", "595": "vers", "596": "kay", "597": "ian", "598": "ates", "599": "▁find", "600": "ward", "601": "▁As", "602": "▁eight", "603": "lic", "604": "▁same", "605": "▁pos", "606": "▁em", "607": "▁made", "608": "▁supp", "609": "▁life", "610": "▁Be", "611": "pect", "612": "▁dec", "613": "▁play", "614": "ange", "615": "▁att", "616": "▁pers", "617": "ways", "618": "▁high", "619": "▁hand", "620": "▁next", "621": "▁cons", "622": "▁own", "623": "▁inv", "624": "ower", "625": "▁ind", "626": "ert", "627": "ng", "628": "ave", "629": "▁year", "630": "▁big", "631": "ating", "632": "▁world", "633": "▁rel", "634": "▁sure", "635": "▁tra", "636": "ew", "637": "ered", "638": "▁fin", "639": "▁Well", "640": "▁sl", "641": "▁doing", "642": "bs", "643": "▁set", "644": "▁rec", "645": "ual", "646": "cial", "647": "▁ph", "648": "erm", "649": "▁love", "650": "ph", "651": "▁real", "652": "▁last", "653": "ict", "654": "▁bo", "655": "▁ra", "656": "ible", "657": "▁wr", "658": "mer", "659": "▁count", "660": "ities", "661": "▁always", "662": "inet", "663": "ments", "664": "uc", "665": "▁might", "666": "▁inter", "667": "▁video", "668": "gin", "669": "▁tell", "670": "▁never", "671": "vent", "672": "▁import", "673": "ied", "674": "▁sy", "675": "▁How", "676": "ically", "677": "ought", "678": "▁thir", "679": "▁rep", "680": "ks", "681": "ib", "682": "▁fam", "683": "ject", "684": "▁bas", "685": "▁She", "686": "▁give", "687": "akes", "688": "▁ninet", "689": "▁reg", "690": "▁min", "691": "▁op", "692": "▁def", "693": "▁didn", "694": "te", "695": "▁cour", "696": "▁why", "697": "▁ent", "698": "▁place", "699": "▁ins", "700": "▁car", "701": "ather", "702": "▁person", "703": "ular", "704": "▁inst", "705": "▁prod", "706": "lect", "707": "▁Al", "708": "▁today", "709": "▁bec", "710": "▁sur", "711": "▁All", "712": "▁another", "713": "▁bus", "714": "▁keep", "715": "ell", "716": "ese", "717": "riend", "718": "▁quest", "719": "▁talk", "720": "als", "721": "ings", "722": "▁mon", "723": "cond", "724": "old", "725": "▁acc", "726": "▁la", "727": "▁num", "728": "ident", "729": "▁che", "730": "iness", "731": "▁turn", "732": "▁ear", "733": "▁No", "734": "ousand", "735": "▁better", "736": "ific", "737": "▁loo", "738": "▁gl", "739": "oc", "740": "▁important", "741": "ited", "742": "▁An", "743": "▁thousand", "744": "ility", "745": "llow", "746": "▁used", "747": "▁gen", "748": "▁sim", "749": "li", "750": "▁happen", "751": "▁Un", "752": "▁Let", "753": "air", "754": "ock", "755": "ably", "756": "gg", "757": "▁watch", "758": "▁For", "759": "▁sw", "760": "ren", "761": "ute", "762": "ever", "763": "▁pol", "764": "▁sch", "765": "▁When", "766": "▁such", "767": "▁fif", "768": "▁home", "769": "▁cle", "770": "▁contin", "771": "ouse", "772": "▁friend", "773": "uring", "774": "▁Okay", "775": "gr", "776": "▁able", "777": "▁stud", "778": "▁eff", "779": "hip", "780": "body", "781": "▁top", "782": "ness", "783": "▁exper", "784": "▁pret", "785": "▁both", "786": "▁done", "787": "cri", "788": "▁mark", "789": "▁while", "790": "▁old", "791": "ros", "792": "ont", "793": "▁second", "794": "ative", "795": "▁thought", "796": "▁best", "797": "▁found", "798": "iew", "799": "▁belie", "800": "▁each", "801": "erest", "802": "▁tri", "803": "▁eas", "804": "▁ca", "805": "▁fact", "806": "▁care", "807": "▁fun", "808": "atter", "809": "ures", "810": "▁head", "811": "▁lear", "812": "▁water", "813": "▁hard", "814": "▁few", "815": "▁side", "816": "ween", "817": "▁exp", "818": "▁away", "819": "its", "820": "▁ext", "821": "lud", "822": "▁run", "823": "▁trans", "824": "ince", "825": "▁sk", "826": "▁open", "827": "cus", "828": "▁between", "829": "▁called", "830": "▁wee", "831": "▁pretty", "832": "ason", "833": "▁far", "834": "ember", "835": "omm", "836": "▁interest", "837": "any", "838": "ner", "839": "uff", "840": "▁pres", "841": "▁cur", "842": "▁child", "843": "ee", "844": "▁toget", "845": "▁together", "846": "olog", "847": "▁God", "848": "ond", "849": "▁char", "850": "▁looking", "851": "stem", "852": "az", "853": "cent", "854": "▁ob", "855": "▁ass", "856": "land", "857": "▁doesn", "858": "▁business", "859": "▁course", "860": "▁ten", "861": "ps", "862": "arch", "863": "ced", "864": "ms", "865": "ize", "866": "nce", "867": "▁ref", "868": "▁name", "869": "ross", "870": "▁grow", "871": "oney", "872": "▁went", "873": "ics", "874": "teen", "875": "▁cou", "876": "▁prob", "877": "▁ret", "878": "▁guys", "879": "▁came", "880": "ash", "881": "led", "882": "▁Eur", "883": "ues", "884": "▁ide", "885": "gan", "886": "▁everything", "887": "▁getting", "888": "▁ask", "889": "▁cor", "890": "▁build", "891": "▁sign", "892": "▁small", "893": "uck", "894": "▁el", "895": "▁col", "896": "▁Is", "897": "ational", "898": "stand", "899": "cy", "900": "▁conf", "901": "der", "902": "▁bre", "903": "▁cap", "904": "▁mod", "905": "ets", "906": "ike", "907": "▁number", "908": "▁comple", "909": "ertain", "910": "▁ever", "911": "▁coll", "912": "▁hum", "913": "▁Europe", "914": "▁cre", "915": "▁met", "916": "▁exam", "917": "▁move", "918": "▁pass", "919": "▁left", "920": "▁system", "921": "▁includ", "922": "▁Thank", "923": "cept", "924": "▁wom", "925": "▁product", "926": "ten", "927": "▁rest", "928": "▁probably", "929": "▁dri", "930": "▁Do", "931": "▁gener", "932": "▁anything", "933": "▁lar", "934": "▁My", "935": "▁school", "936": "▁lead", "937": "▁sub", "938": "▁ty", "939": "▁plan", "940": "▁seem", "941": "▁whole", "942": "irect", "943": "▁light", "944": "▁must", "945": "▁mom", "946": "▁opp", "947": "▁support", "948": "▁family", "949": "ices", "950": "amp", "951": "▁proble", "952": "▁dr", "953": "ready", "954": "▁using", "955": "ense", "956": "▁prov", "957": "ush", "958": "ax", "959": "▁power", "960": "▁Re", "961": "alth", "962": "▁ev", "963": "▁stand", "964": "▁war", "965": "ts", "966": "▁", "967": "e", "968": "t", "969": "o", "970": "a", "971": "n", "972": "i", "973": "s", "974": "r", "975": "h", "976": "l", "977": "d", "978": "u", "979": "c", "980": "m", "981": "y", "982": "g", "983": "w", "984": "f", "985": "p", "986": ".", "987": "b", "988": ",", "989": "v", "990": "k", "991": "'", "992": "I", "993": "T", "994": "A", "995": "S", "996": "x", "997": "W", "998": "j", "999": "B", "1000": "C", "1001": "H", "1002": "?", "1003": "M", "1004": "O", "1005": "Y", "1006": "N", "1007": "P", "1008": "E", "1009": "q", "1010": "L", "1011": "D", "1012": "z", "1013": "G", "1014": "F", "1015": "R", "1016": "!", "1017": "J", "1018": "U", "1019": "K", "1020": "V", "1021": "Q", "1022": "Z", "1023": "X"} \ No newline at end of file diff --git a/models/parakeet-tdt-0.6b-v2-coreml/.DS_Store b/models/parakeet-tdt-0.6b-v2-coreml/.DS_Store deleted file mode 100644 index ad6e41b077109ad8b0188aa879c7d5c2f9afdae2..0000000000000000000000000000000000000000 Binary files a/models/parakeet-tdt-0.6b-v2-coreml/.DS_Store and /dev/null differ diff --git a/models/parakeet-tdt-0.6b-v2-coreml/Decoder.mlmodelc/analytics/coremldata.bin b/models/parakeet-tdt-0.6b-v2-coreml/Decoder.mlmodelc/analytics/coremldata.bin deleted file mode 100644 index 7e6fa4eb79131e62089f252a08b860e723f5114f..0000000000000000000000000000000000000000 --- a/models/parakeet-tdt-0.6b-v2-coreml/Decoder.mlmodelc/analytics/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:46de1a6fe2e49d19a2125bc91acf020df7f2aea84ba821532aade8427a440b05 -size 243 diff --git a/models/parakeet-tdt-0.6b-v2-coreml/Decoder.mlmodelc/coremldata.bin b/models/parakeet-tdt-0.6b-v2-coreml/Decoder.mlmodelc/coremldata.bin deleted file mode 100644 index 79b160cb7e3eb3213fb22cde5fbf6ac7f324c91b..0000000000000000000000000000000000000000 --- a/models/parakeet-tdt-0.6b-v2-coreml/Decoder.mlmodelc/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:d200ca07694a347f6d02a3886a062ae839831e094e443222f2e48a14945966a8 -size 554 diff --git a/models/parakeet-tdt-0.6b-v2-coreml/Decoder.mlmodelc/metadata.json b/models/parakeet-tdt-0.6b-v2-coreml/Decoder.mlmodelc/metadata.json deleted file mode 100644 index e3a08d21a5d1577db2bedd1749ea0ddae5d21b73..0000000000000000000000000000000000000000 --- a/models/parakeet-tdt-0.6b-v2-coreml/Decoder.mlmodelc/metadata.json +++ /dev/null @@ -1,123 +0,0 @@ -[ - { - "metadataOutputVersion" : "3.0", - "shortDescription" : "Parakeet decoder (RNNT prediction network)", - "outputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 640 × 1)", - "shortDescription" : "", - "shape" : "[1, 640, 1]", - "name" : "decoder", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 2 × 1 × 640)", - "shortDescription" : "", - "shape" : "[2, 1, 640]", - "name" : "h_out", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 2 × 1 × 640)", - "shortDescription" : "", - "shape" : "[2, 1, 640]", - "name" : "c_out", - "type" : "MultiArray" - } - ], - "storagePrecision" : "Float16", - "modelParameters" : [ - - ], - "author" : "Fluid Inference", - "specificationVersion" : 8, - "mlProgramOperationTypeHistogram" : { - "Select" : 1, - "Ios17.squeeze" : 4, - "Ios17.gather" : 1, - "Ios17.cast" : 8, - "Ios17.lstm" : 2, - "Split" : 2, - "Ios17.add" : 1, - "Ios17.transpose" : 2, - "Ios17.greaterEqual" : 1, - "Identity" : 1, - "Stack" : 2 - }, - "computePrecision" : "Mixed (Float16, Float32, Int16, Int32)", - "isUpdatable" : "0", - "stateSchema" : [ - - ], - "availability" : { - "macOS" : "14.0", - "tvOS" : "17.0", - "visionOS" : "1.0", - "watchOS" : "10.0", - "iOS" : "17.0", - "macCatalyst" : "17.0" - }, - "modelType" : { - "name" : "MLModelType_mlProgram" - }, - "inputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1 × 1)", - "shortDescription" : "", - "shape" : "[1, 1]", - "name" : "targets", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1)", - "shortDescription" : "", - "shape" : "[1]", - "name" : "target_length", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 2 × 1 × 640)", - "shortDescription" : "", - "shape" : "[2, 1, 640]", - "name" : "h_in", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 2 × 1 × 640)", - "shortDescription" : "", - "shape" : "[2, 1, 640]", - "name" : "c_in", - "type" : "MultiArray" - } - ], - "userDefinedMetadata" : { - "com.github.apple.coremltools.conversion_date" : "2025-09-25", - "com.github.apple.coremltools.source" : "torch==2.7.0", - "com.github.apple.coremltools.version" : "9.0b1", - "com.github.apple.coremltools.source_dialect" : "TorchScript" - }, - "generatedClassName" : "parakeet_decoder", - "method" : "predict" - } -] \ No newline at end of file diff --git a/models/parakeet-tdt-0.6b-v2-coreml/Decoder.mlmodelc/model.mil b/models/parakeet-tdt-0.6b-v2-coreml/Decoder.mlmodelc/model.mil deleted file mode 100644 index 167cc9db34dd7828e0e61eab40d6a09f8a234d30..0000000000000000000000000000000000000000 --- a/models/parakeet-tdt-0.6b-v2-coreml/Decoder.mlmodelc/model.mil +++ /dev/null @@ -1,73 +0,0 @@ -program(1.0) -[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3500.14.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.7.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "9.0b1"}})] -{ - func main(tensor c_in, tensor h_in, tensor target_length, tensor targets) { - tensor y_batch_dims_0 = const()[name = tensor("y_batch_dims_0"), val = tensor(0)]; - tensor y_validate_indices_0 = const()[name = tensor("y_validate_indices_0"), val = tensor(false)]; - tensor module_prediction_embed_weight_to_fp16 = const()[name = tensor("module_prediction_embed_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; - tensor targets_to_int16_dtype_0 = const()[name = tensor("targets_to_int16_dtype_0"), val = tensor("int16")]; - tensor cast_1_dtype_0 = const()[name = tensor("cast_1_dtype_0"), val = tensor("int32")]; - tensor greater_equal_0_y_0 = const()[name = tensor("greater_equal_0_y_0"), val = tensor(0)]; - tensor targets_to_int16 = cast(dtype = targets_to_int16_dtype_0, x = targets)[name = tensor("cast_9")]; - tensor cast_1 = cast(dtype = cast_1_dtype_0, x = targets_to_int16)[name = tensor("cast_8")]; - tensor greater_equal_0 = greater_equal(x = cast_1, y = greater_equal_0_y_0)[name = tensor("greater_equal_0")]; - tensor slice_by_index_0 = const()[name = tensor("slice_by_index_0"), val = tensor(1025)]; - tensor add_2 = add(x = cast_1, y = slice_by_index_0)[name = tensor("add_2")]; - tensor select_0 = select(a = cast_1, b = add_2, cond = greater_equal_0)[name = tensor("select_0")]; - tensor y_cast_fp16_cast_uint16_axis_0 = const()[name = tensor("y_cast_fp16_cast_uint16_axis_0"), val = tensor(0)]; - tensor select_0_to_int16_dtype_0 = const()[name = tensor("select_0_to_int16_dtype_0"), val = tensor("int16")]; - tensor select_0_to_int16 = cast(dtype = select_0_to_int16_dtype_0, x = select_0)[name = tensor("cast_7")]; - tensor y_cast_fp16_cast_uint16_cast_uint16 = gather(axis = y_cast_fp16_cast_uint16_axis_0, batch_dims = y_batch_dims_0, indices = select_0_to_int16, validate_indices = y_validate_indices_0, x = module_prediction_embed_weight_to_fp16)[name = tensor("y_cast_fp16_cast_uint16_cast_uint16")]; - tensor input_3_perm_0 = const()[name = tensor("input_3_perm_0"), val = tensor([1, 0, 2])]; - tensor split_0_num_splits_0 = const()[name = tensor("split_0_num_splits_0"), val = tensor(2)]; - tensor split_0_axis_0 = const()[name = tensor("split_0_axis_0"), val = tensor(0)]; - tensor h_in_to_fp16_dtype_0 = const()[name = tensor("h_in_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor h_in_to_fp16 = cast(dtype = h_in_to_fp16_dtype_0, x = h_in)[name = tensor("cast_6")]; - tensor split_0_cast_fp16_0, tensor split_0_cast_fp16_1 = split(axis = split_0_axis_0, num_splits = split_0_num_splits_0, x = h_in_to_fp16)[name = tensor("split_0_cast_fp16")]; - tensor split_1_num_splits_0 = const()[name = tensor("split_1_num_splits_0"), val = tensor(2)]; - tensor split_1_axis_0 = const()[name = tensor("split_1_axis_0"), val = tensor(0)]; - tensor c_in_to_fp16_dtype_0 = const()[name = tensor("c_in_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor c_in_to_fp16 = cast(dtype = c_in_to_fp16_dtype_0, x = c_in)[name = tensor("cast_5")]; - tensor split_1_cast_fp16_0, tensor split_1_cast_fp16_1 = split(axis = split_1_axis_0, num_splits = split_1_num_splits_0, x = c_in_to_fp16)[name = tensor("split_1_cast_fp16")]; - tensor input_lstm_layer_0_lstm_h0_squeeze_axes_0 = const()[name = tensor("input_lstm_layer_0_lstm_h0_squeeze_axes_0"), val = tensor([0])]; - tensor input_lstm_layer_0_lstm_h0_squeeze_cast_fp16 = squeeze(axes = input_lstm_layer_0_lstm_h0_squeeze_axes_0, x = split_0_cast_fp16_0)[name = tensor("input_lstm_layer_0_lstm_h0_squeeze_cast_fp16")]; - tensor input_lstm_layer_0_lstm_c0_squeeze_axes_0 = const()[name = tensor("input_lstm_layer_0_lstm_c0_squeeze_axes_0"), val = tensor([0])]; - tensor input_lstm_layer_0_lstm_c0_squeeze_cast_fp16 = squeeze(axes = input_lstm_layer_0_lstm_c0_squeeze_axes_0, x = split_1_cast_fp16_0)[name = tensor("input_lstm_layer_0_lstm_c0_squeeze_cast_fp16")]; - tensor input_lstm_layer_0_direction_0 = const()[name = tensor("input_lstm_layer_0_direction_0"), val = tensor("forward")]; - tensor input_lstm_layer_0_output_sequence_0 = const()[name = tensor("input_lstm_layer_0_output_sequence_0"), val = tensor(true)]; - tensor input_lstm_layer_0_recurrent_activation_0 = const()[name = tensor("input_lstm_layer_0_recurrent_activation_0"), val = tensor("sigmoid")]; - tensor input_lstm_layer_0_cell_activation_0 = const()[name = tensor("input_lstm_layer_0_cell_activation_0"), val = tensor("tanh")]; - tensor input_lstm_layer_0_activation_0 = const()[name = tensor("input_lstm_layer_0_activation_0"), val = tensor("tanh")]; - tensor concat_1_to_fp16 = const()[name = tensor("concat_1_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1312128)))]; - tensor concat_2_to_fp16 = const()[name = tensor("concat_2_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4588992)))]; - tensor concat_0_to_fp16 = const()[name = tensor("concat_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7865856)))]; - tensor input_3_cast_fp16 = transpose(perm = input_3_perm_0, x = y_cast_fp16_cast_uint16_cast_uint16)[name = tensor("transpose_2")]; - tensor input_lstm_layer_0_cast_fp16_0, tensor input_lstm_layer_0_cast_fp16_1, tensor input_lstm_layer_0_cast_fp16_2 = lstm(activation = input_lstm_layer_0_activation_0, bias = concat_0_to_fp16, cell_activation = input_lstm_layer_0_cell_activation_0, direction = input_lstm_layer_0_direction_0, initial_c = input_lstm_layer_0_lstm_c0_squeeze_cast_fp16, initial_h = input_lstm_layer_0_lstm_h0_squeeze_cast_fp16, output_sequence = input_lstm_layer_0_output_sequence_0, recurrent_activation = input_lstm_layer_0_recurrent_activation_0, weight_hh = concat_2_to_fp16, weight_ih = concat_1_to_fp16, x = input_3_cast_fp16)[name = tensor("input_lstm_layer_0_cast_fp16")]; - tensor input_lstm_h0_squeeze_axes_0 = const()[name = tensor("input_lstm_h0_squeeze_axes_0"), val = tensor([0])]; - tensor input_lstm_h0_squeeze_cast_fp16 = squeeze(axes = input_lstm_h0_squeeze_axes_0, x = split_0_cast_fp16_1)[name = tensor("input_lstm_h0_squeeze_cast_fp16")]; - tensor input_lstm_c0_squeeze_axes_0 = const()[name = tensor("input_lstm_c0_squeeze_axes_0"), val = tensor([0])]; - tensor input_lstm_c0_squeeze_cast_fp16 = squeeze(axes = input_lstm_c0_squeeze_axes_0, x = split_1_cast_fp16_1)[name = tensor("input_lstm_c0_squeeze_cast_fp16")]; - tensor input_direction_0 = const()[name = tensor("input_direction_0"), val = tensor("forward")]; - tensor input_output_sequence_0 = const()[name = tensor("input_output_sequence_0"), val = tensor(true)]; - tensor input_recurrent_activation_0 = const()[name = tensor("input_recurrent_activation_0"), val = tensor("sigmoid")]; - tensor input_cell_activation_0 = const()[name = tensor("input_cell_activation_0"), val = tensor("tanh")]; - tensor input_activation_0 = const()[name = tensor("input_activation_0"), val = tensor("tanh")]; - tensor concat_4_to_fp16 = const()[name = tensor("concat_4_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7871040)))]; - tensor concat_5_to_fp16 = const()[name = tensor("concat_5_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11147904)))]; - tensor concat_3_to_fp16 = const()[name = tensor("concat_3_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(14424768)))]; - tensor input_cast_fp16_0, tensor input_cast_fp16_1, tensor input_cast_fp16_2 = lstm(activation = input_activation_0, bias = concat_3_to_fp16, cell_activation = input_cell_activation_0, direction = input_direction_0, initial_c = input_lstm_c0_squeeze_cast_fp16, initial_h = input_lstm_h0_squeeze_cast_fp16, output_sequence = input_output_sequence_0, recurrent_activation = input_recurrent_activation_0, weight_hh = concat_5_to_fp16, weight_ih = concat_4_to_fp16, x = input_lstm_layer_0_cast_fp16_0)[name = tensor("input_cast_fp16")]; - tensor obj_3_axis_0 = const()[name = tensor("obj_3_axis_0"), val = tensor(0)]; - tensor obj_3_cast_fp16 = stack(axis = obj_3_axis_0, values = (input_lstm_layer_0_cast_fp16_1, input_cast_fp16_1))[name = tensor("obj_3_cast_fp16")]; - tensor obj_3_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("obj_3_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; - tensor obj_axis_0 = const()[name = tensor("obj_axis_0"), val = tensor(0)]; - tensor obj_cast_fp16 = stack(axis = obj_axis_0, values = (input_lstm_layer_0_cast_fp16_2, input_cast_fp16_2))[name = tensor("obj_cast_fp16")]; - tensor obj_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("obj_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; - tensor transpose_0_perm_0 = const()[name = tensor("transpose_0_perm_0"), val = tensor([1, 2, 0])]; - tensor transpose_0_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("transpose_0_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; - tensor transpose_0_cast_fp16 = transpose(perm = transpose_0_perm_0, x = input_cast_fp16_0)[name = tensor("transpose_1")]; - tensor decoder = cast(dtype = transpose_0_cast_fp16_to_fp32_dtype_0, x = transpose_0_cast_fp16)[name = tensor("cast_2")]; - tensor c_out = cast(dtype = obj_cast_fp16_to_fp32_dtype_0, x = obj_cast_fp16)[name = tensor("cast_3")]; - tensor h_out = cast(dtype = obj_3_cast_fp16_to_fp32_dtype_0, x = obj_3_cast_fp16)[name = tensor("cast_4")]; - tensor target_length_tmp = identity(x = target_length)[name = tensor("target_length_tmp")]; - } -> (decoder, h_out, c_out); -} \ No newline at end of file diff --git a/models/parakeet-tdt-0.6b-v2-coreml/Decoder.mlmodelc/weights/weight.bin b/models/parakeet-tdt-0.6b-v2-coreml/Decoder.mlmodelc/weights/weight.bin deleted file mode 100644 index a895491b78de520abc21d90321ffd90c95483662..0000000000000000000000000000000000000000 --- a/models/parakeet-tdt-0.6b-v2-coreml/Decoder.mlmodelc/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:27d26890221d82322c1092fd99d7b40578e435d5cf4b83c887c42603caf97aba -size 14429952 diff --git a/models/parakeet-tdt-0.6b-v2-coreml/Encoder.mlmodelc/analytics/coremldata.bin b/models/parakeet-tdt-0.6b-v2-coreml/Encoder.mlmodelc/analytics/coremldata.bin deleted file mode 100644 index 4738782a8e36389621ff4b3b90dd2a661b35146c..0000000000000000000000000000000000000000 --- a/models/parakeet-tdt-0.6b-v2-coreml/Encoder.mlmodelc/analytics/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:42e638870d73f26b332918a3496ce36793fbb413a81cbd3d16ba01328637a105 -size 243 diff --git a/models/parakeet-tdt-0.6b-v2-coreml/Encoder.mlmodelc/coremldata.bin b/models/parakeet-tdt-0.6b-v2-coreml/Encoder.mlmodelc/coremldata.bin deleted file mode 100644 index e016ed32c20e1563399f15adad451a442644b3d7..0000000000000000000000000000000000000000 --- a/models/parakeet-tdt-0.6b-v2-coreml/Encoder.mlmodelc/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:4def7aa848599ad0e17a8b9a982edcdbf33cf92e1f4b798de32e2ca0bc74b030 -size 485 diff --git a/models/parakeet-tdt-0.6b-v2-coreml/Encoder.mlmodelc/metadata.json b/models/parakeet-tdt-0.6b-v2-coreml/Encoder.mlmodelc/metadata.json deleted file mode 100644 index c87f9eb24e76b8e518dc1fbead46c8c0029d3925..0000000000000000000000000000000000000000 --- a/models/parakeet-tdt-0.6b-v2-coreml/Encoder.mlmodelc/metadata.json +++ /dev/null @@ -1,106 +0,0 @@ -[ - { - "metadataOutputVersion" : "3.0", - "shortDescription" : "enc6bit-palettize quantized - encoder", - "outputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 1024 × 188)", - "shortDescription" : "", - "shape" : "[1, 1024, 188]", - "name" : "encoder", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1)", - "shortDescription" : "", - "shape" : "[1]", - "name" : "encoder_length", - "type" : "MultiArray" - } - ], - "storagePrecision" : "Mixed (Float16, Palettized (6 bits))", - "modelParameters" : [ - - ], - "author" : "Fluid Inference", - "specificationVersion" : 8, - "mlProgramOperationTypeHistogram" : { - "Ios17.floor" : 3, - "Ios17.logicalAnd" : 2, - "Ios17.reshape" : 145, - "Ios16.softmax" : 24, - "Ios17.matmul" : 72, - "Ios17.transpose" : 172, - "Split" : 24, - "Ios17.expandDims" : 5, - "Select" : 72, - "Ios17.add" : 174, - "Tile" : 1, - "Ios17.sliceByIndex" : 48, - "Ios16.sigmoid" : 24, - "Pad" : 48, - "Ios17.logicalNot" : 2, - "Ios17.layerNorm" : 120, - "Ios16.silu" : 72, - "Ios17.less" : 1, - "Ios16.constexprLutToDense" : 294, - "Ios17.conv" : 77, - "Ios16.relu" : 3, - "Ios17.linear" : 193, - "Ios17.cast" : 4, - "Ios17.mul" : 99 - }, - "computePrecision" : "Mixed (Float16, Float32, Int32)", - "isUpdatable" : "0", - "stateSchema" : [ - - ], - "availability" : { - "macOS" : "14.0", - "tvOS" : "17.0", - "visionOS" : "1.0", - "watchOS" : "10.0", - "iOS" : "17.0", - "macCatalyst" : "17.0" - }, - "modelType" : { - "name" : "MLModelType_mlProgram" - }, - "inputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 128 × 1501)", - "shortDescription" : "", - "shape" : "[1, 128, 1501]", - "name" : "mel", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1)", - "shortDescription" : "", - "shape" : "[1]", - "name" : "mel_length", - "type" : "MultiArray" - } - ], - "userDefinedMetadata" : { - "com.github.apple.coremltools.conversion_date" : "2025-09-25", - "com.github.apple.coremltools.source" : "torch==2.7.0", - "com.github.apple.coremltools.version" : "9.0b1", - "com.github.apple.coremltools.source_dialect" : "TorchScript" - }, - "generatedClassName" : "parakeet_encoder", - "method" : "predict" - } -] \ No newline at end of file diff --git a/models/parakeet-tdt-0.6b-v2-coreml/Encoder.mlmodelc/model.mil b/models/parakeet-tdt-0.6b-v2-coreml/Encoder.mlmodelc/model.mil deleted file mode 100644 index 93d1301ccde7c2ca447af6c01ab05bfda5a5dd2c..0000000000000000000000000000000000000000 --- a/models/parakeet-tdt-0.6b-v2-coreml/Encoder.mlmodelc/model.mil +++ /dev/null @@ -1,3404 +0,0 @@ -program(1.0) -[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3500.14.1"}, {"coremlc-version", "3500.32.1"}})] -{ - func main(tensor mel, tensor mel_length) { - tensor var_30 = const()[name = tensor("op_30"), val = tensor(-1)]; - tensor x_1_perm_0 = const()[name = tensor("x_1_perm_0"), val = tensor([0, 2, 1])]; - tensor mel_to_fp16_dtype_0 = const()[name = tensor("mel_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor var_86_to_fp16_dtype_0 = const()[name = tensor("op_86_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor var_87_promoted_to_fp16 = const()[name = tensor("op_87_promoted_to_fp16"), val = tensor(-0x1p+0)]; - tensor mel_length_to_fp16 = cast(dtype = var_86_to_fp16_dtype_0, x = mel_length)[name = tensor("cast_3")]; - tensor var_88_cast_fp16 = add(x = mel_length_to_fp16, y = var_87_promoted_to_fp16)[name = tensor("op_88_cast_fp16")]; - tensor _inversed_90_y_0_to_fp16 = const()[name = tensor("_inversed_90_y_0_to_fp16"), val = tensor(0x1p-1)]; - tensor _inversed_90_cast_fp16 = mul(x = var_88_cast_fp16, y = _inversed_90_y_0_to_fp16)[name = tensor("_inversed_90_cast_fp16")]; - tensor var_91_to_fp16 = const()[name = tensor("op_91_to_fp16"), val = tensor(0x1p+0)]; - tensor lengths_1_cast_fp16 = add(x = _inversed_90_cast_fp16, y = var_91_to_fp16)[name = tensor("lengths_1_cast_fp16")]; - tensor lengths_3_cast_fp16 = floor(x = lengths_1_cast_fp16)[name = tensor("lengths_3_cast_fp16")]; - tensor var_95_promoted_to_fp16 = const()[name = tensor("op_95_promoted_to_fp16"), val = tensor(-0x1p+0)]; - tensor var_96_cast_fp16 = add(x = lengths_3_cast_fp16, y = var_95_promoted_to_fp16)[name = tensor("op_96_cast_fp16")]; - tensor _inversed_98_y_0_to_fp16 = const()[name = tensor("_inversed_98_y_0_to_fp16"), val = tensor(0x1p-1)]; - tensor _inversed_98_cast_fp16 = mul(x = var_96_cast_fp16, y = _inversed_98_y_0_to_fp16)[name = tensor("_inversed_98_cast_fp16")]; - tensor var_99_to_fp16 = const()[name = tensor("op_99_to_fp16"), val = tensor(0x1p+0)]; - tensor lengths_7_cast_fp16 = add(x = _inversed_98_cast_fp16, y = var_99_to_fp16)[name = tensor("lengths_7_cast_fp16")]; - tensor lengths_9_cast_fp16 = floor(x = lengths_7_cast_fp16)[name = tensor("lengths_9_cast_fp16")]; - tensor var_103_promoted_to_fp16 = const()[name = tensor("op_103_promoted_to_fp16"), val = tensor(-0x1p+0)]; - tensor var_104_cast_fp16 = add(x = lengths_9_cast_fp16, y = var_103_promoted_to_fp16)[name = tensor("op_104_cast_fp16")]; - tensor _inversed_106_y_0_to_fp16 = const()[name = tensor("_inversed_106_y_0_to_fp16"), val = tensor(0x1p-1)]; - tensor _inversed_106_cast_fp16 = mul(x = var_104_cast_fp16, y = _inversed_106_y_0_to_fp16)[name = tensor("_inversed_106_cast_fp16")]; - tensor var_107_to_fp16 = const()[name = tensor("op_107_to_fp16"), val = tensor(0x1p+0)]; - tensor lengths_13_cast_fp16 = add(x = _inversed_106_cast_fp16, y = var_107_to_fp16)[name = tensor("lengths_13_cast_fp16")]; - tensor lengths_cast_fp16 = floor(x = lengths_13_cast_fp16)[name = tensor("lengths_cast_fp16")]; - tensor input_1_axes_0 = const()[name = tensor("input_1_axes_0"), val = tensor([1])]; - tensor mel_to_fp16 = cast(dtype = mel_to_fp16_dtype_0, x = mel)[name = tensor("cast_2")]; - tensor x_1_cast_fp16 = transpose(perm = x_1_perm_0, x = mel_to_fp16)[name = tensor("transpose_315")]; - tensor input_1_cast_fp16 = expand_dims(axes = input_1_axes_0, x = x_1_cast_fp16)[name = tensor("input_1_cast_fp16")]; - tensor input_3_pad_type_0 = const()[name = tensor("input_3_pad_type_0"), val = tensor("custom")]; - tensor input_3_pad_0 = const()[name = tensor("input_3_pad_0"), val = tensor([1, 1, 1, 1])]; - tensor input_3_strides_0 = const()[name = tensor("input_3_strides_0"), val = tensor([2, 2])]; - tensor input_3_dilations_0 = const()[name = tensor("input_3_dilations_0"), val = tensor([1, 1])]; - tensor input_3_groups_0 = const()[name = tensor("input_3_groups_0"), val = tensor(1)]; - tensor module_pre_encode_conv_0_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1856))), name = tensor("module_pre_encode_conv_0_weight_to_fp16_palettized"), shape = tensor([256, 1, 3, 3])]; - tensor module_pre_encode_conv_0_bias_to_fp16 = const()[name = tensor("module_pre_encode_conv_0_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2048)))]; - tensor input_3_cast_fp16 = conv(bias = module_pre_encode_conv_0_bias_to_fp16, dilations = input_3_dilations_0, groups = input_3_groups_0, pad = input_3_pad_0, pad_type = input_3_pad_type_0, strides = input_3_strides_0, weight = module_pre_encode_conv_0_weight_to_fp16_palettized, x = input_1_cast_fp16)[name = tensor("input_3_cast_fp16")]; - tensor input_5_cast_fp16 = relu(x = input_3_cast_fp16)[name = tensor("input_5_cast_fp16")]; - tensor input_7_pad_type_0 = const()[name = tensor("input_7_pad_type_0"), val = tensor("custom")]; - tensor input_7_pad_0 = const()[name = tensor("input_7_pad_0"), val = tensor([1, 1, 1, 1])]; - tensor input_7_strides_0 = const()[name = tensor("input_7_strides_0"), val = tensor([2, 2])]; - tensor input_7_groups_0 = const()[name = tensor("input_7_groups_0"), val = tensor(256)]; - tensor input_7_dilations_0 = const()[name = tensor("input_7_dilations_0"), val = tensor([1, 1])]; - tensor module_pre_encode_conv_2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2624))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4416))), name = tensor("module_pre_encode_conv_2_weight_to_fp16_palettized"), shape = tensor([256, 1, 3, 3])]; - tensor module_pre_encode_conv_2_bias_to_fp16 = const()[name = tensor("module_pre_encode_conv_2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4608)))]; - tensor input_7_cast_fp16 = conv(bias = module_pre_encode_conv_2_bias_to_fp16, dilations = input_7_dilations_0, groups = input_7_groups_0, pad = input_7_pad_0, pad_type = input_7_pad_type_0, strides = input_7_strides_0, weight = module_pre_encode_conv_2_weight_to_fp16_palettized, x = input_5_cast_fp16)[name = tensor("input_7_cast_fp16")]; - tensor input_9_pad_type_0 = const()[name = tensor("input_9_pad_type_0"), val = tensor("valid")]; - tensor input_9_strides_0 = const()[name = tensor("input_9_strides_0"), val = tensor([1, 1])]; - tensor input_9_pad_0 = const()[name = tensor("input_9_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor input_9_dilations_0 = const()[name = tensor("input_9_dilations_0"), val = tensor([1, 1])]; - tensor input_9_groups_0 = const()[name = tensor("input_9_groups_0"), val = tensor(1)]; - tensor module_pre_encode_conv_3_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(5184))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(54400))), name = tensor("module_pre_encode_conv_3_weight_to_fp16_palettized"), shape = tensor([256, 256, 1, 1])]; - tensor module_pre_encode_conv_3_bias_to_fp16 = const()[name = tensor("module_pre_encode_conv_3_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(54592)))]; - tensor input_9_cast_fp16 = conv(bias = module_pre_encode_conv_3_bias_to_fp16, dilations = input_9_dilations_0, groups = input_9_groups_0, pad = input_9_pad_0, pad_type = input_9_pad_type_0, strides = input_9_strides_0, weight = module_pre_encode_conv_3_weight_to_fp16_palettized, x = input_7_cast_fp16)[name = tensor("input_9_cast_fp16")]; - tensor input_11_cast_fp16 = relu(x = input_9_cast_fp16)[name = tensor("input_11_cast_fp16")]; - tensor input_13_pad_type_0 = const()[name = tensor("input_13_pad_type_0"), val = tensor("custom")]; - tensor input_13_pad_0 = const()[name = tensor("input_13_pad_0"), val = tensor([1, 1, 1, 1])]; - tensor input_13_strides_0 = const()[name = tensor("input_13_strides_0"), val = tensor([2, 2])]; - tensor input_13_groups_0 = const()[name = tensor("input_13_groups_0"), val = tensor(256)]; - tensor input_13_dilations_0 = const()[name = tensor("input_13_dilations_0"), val = tensor([1, 1])]; - tensor module_pre_encode_conv_5_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(55168))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(56960))), name = tensor("module_pre_encode_conv_5_weight_to_fp16_palettized"), shape = tensor([256, 1, 3, 3])]; - tensor module_pre_encode_conv_5_bias_to_fp16 = const()[name = tensor("module_pre_encode_conv_5_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57152)))]; - tensor input_13_cast_fp16 = conv(bias = module_pre_encode_conv_5_bias_to_fp16, dilations = input_13_dilations_0, groups = input_13_groups_0, pad = input_13_pad_0, pad_type = input_13_pad_type_0, strides = input_13_strides_0, weight = module_pre_encode_conv_5_weight_to_fp16_palettized, x = input_11_cast_fp16)[name = tensor("input_13_cast_fp16")]; - tensor input_15_pad_type_0 = const()[name = tensor("input_15_pad_type_0"), val = tensor("valid")]; - tensor input_15_strides_0 = const()[name = tensor("input_15_strides_0"), val = tensor([1, 1])]; - tensor input_15_pad_0 = const()[name = tensor("input_15_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor input_15_dilations_0 = const()[name = tensor("input_15_dilations_0"), val = tensor([1, 1])]; - tensor input_15_groups_0 = const()[name = tensor("input_15_groups_0"), val = tensor(1)]; - tensor module_pre_encode_conv_6_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57728))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(106944))), name = tensor("module_pre_encode_conv_6_weight_to_fp16_palettized"), shape = tensor([256, 256, 1, 1])]; - tensor module_pre_encode_conv_6_bias_to_fp16 = const()[name = tensor("module_pre_encode_conv_6_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(107136)))]; - tensor input_15_cast_fp16 = conv(bias = module_pre_encode_conv_6_bias_to_fp16, dilations = input_15_dilations_0, groups = input_15_groups_0, pad = input_15_pad_0, pad_type = input_15_pad_type_0, strides = input_15_strides_0, weight = module_pre_encode_conv_6_weight_to_fp16_palettized, x = input_13_cast_fp16)[name = tensor("input_15_cast_fp16")]; - tensor x_3_cast_fp16 = relu(x = input_15_cast_fp16)[name = tensor("x_3_cast_fp16")]; - tensor var_157_perm_0 = const()[name = tensor("op_157_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_158 = const()[name = tensor("op_158"), val = tensor([1, 188, -1])]; - tensor var_157_cast_fp16 = transpose(perm = var_157_perm_0, x = x_3_cast_fp16)[name = tensor("transpose_314")]; - tensor input_17_cast_fp16 = reshape(shape = var_158, x = var_157_cast_fp16)[name = tensor("input_17_cast_fp16")]; - tensor module_pre_encode_out_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(107712))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3253504))), name = tensor("module_pre_encode_out_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor module_pre_encode_out_bias_to_fp16 = const()[name = tensor("module_pre_encode_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3253696)))]; - tensor linear_0_cast_fp16 = linear(bias = module_pre_encode_out_bias_to_fp16, weight = module_pre_encode_out_weight_to_fp16_palettized, x = input_17_cast_fp16)[name = tensor("linear_0_cast_fp16")]; - tensor padding_length_dtype_0 = const()[name = tensor("padding_length_dtype_0"), val = tensor("int32")]; - tensor expand_dims_0 = const()[name = tensor("expand_dims_0"), val = tensor([[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187]])]; - tensor var_196_axes_0 = const()[name = tensor("op_196_axes_0"), val = tensor([-1])]; - tensor encoder_length = cast(dtype = padding_length_dtype_0, x = lengths_cast_fp16)[name = tensor("cast_1")]; - tensor var_196 = expand_dims(axes = var_196_axes_0, x = encoder_length)[name = tensor("op_196")]; - tensor pad_mask_1 = less(x = expand_dims_0, y = var_196)[name = tensor("pad_mask_1")]; - tensor var_198_axes_0 = const()[name = tensor("op_198_axes_0"), val = tensor([1])]; - tensor var_198 = expand_dims(axes = var_198_axes_0, x = pad_mask_1)[name = tensor("op_198")]; - tensor var_199 = const()[name = tensor("op_199"), val = tensor([1, 188, 1])]; - tensor pad_mask_for_att_mask_1 = tile(reps = var_199, x = var_198)[name = tensor("pad_mask_for_att_mask_1")]; - tensor var_201_perm_0 = const()[name = tensor("op_201_perm_0"), val = tensor([0, 2, 1])]; - tensor var_201 = transpose(perm = var_201_perm_0, x = pad_mask_for_att_mask_1)[name = tensor("transpose_313")]; - tensor pad_mask_for_att_mask = logical_and(x = pad_mask_for_att_mask_1, y = var_201)[name = tensor("pad_mask_for_att_mask")]; - tensor const_7 = const()[name = tensor("const_7"), val = tensor([[[true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true]]])]; - tensor att_mask = logical_and(x = pad_mask_for_att_mask, y = const_7)[name = tensor("att_mask")]; - tensor mask_1 = logical_not(x = att_mask)[name = tensor("mask_1")]; - tensor pad_mask = logical_not(x = pad_mask_1)[name = tensor("pad_mask")]; - tensor input_21_axes_0 = const()[name = tensor("input_21_axes_0"), val = tensor([-1])]; - tensor module_layers_0_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_0_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3255808)))]; - tensor module_layers_0_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_0_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3257920)))]; - tensor var_9_to_fp16 = const()[name = tensor("op_9_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_21_cast_fp16 = layer_norm(axes = input_21_axes_0, beta = module_layers_0_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_0_norm_feed_forward1_weight_to_fp16, x = linear_0_cast_fp16)[name = tensor("input_21_cast_fp16")]; - tensor module_layers_0_feed_forward1_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3260032))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(6405824))), name = tensor("module_layers_0_feed_forward1_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_1_bias_0_to_fp16 = const()[name = tensor("linear_1_bias_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(6406016)))]; - tensor linear_1_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_0_feed_forward1_linear1_weight_to_fp16_palettized, x = input_21_cast_fp16)[name = tensor("linear_1_cast_fp16")]; - tensor input_25_cast_fp16 = silu(x = linear_1_cast_fp16)[name = tensor("input_25_cast_fp16")]; - tensor module_layers_0_feed_forward1_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(6414272))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9560064))), name = tensor("module_layers_0_feed_forward1_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_2_bias_0_to_fp16 = const()[name = tensor("linear_2_bias_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9560256)))]; - tensor linear_2_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_0_feed_forward1_linear2_weight_to_fp16_palettized, x = input_25_cast_fp16)[name = tensor("linear_2_cast_fp16")]; - tensor var_232_to_fp16 = const()[name = tensor("op_232_to_fp16"), val = tensor(0x1p-1)]; - tensor var_233_cast_fp16 = mul(x = linear_2_cast_fp16, y = var_232_to_fp16)[name = tensor("op_233_cast_fp16")]; - tensor input_31_cast_fp16 = add(x = linear_0_cast_fp16, y = var_233_cast_fp16)[name = tensor("input_31_cast_fp16")]; - tensor query_1_axes_0 = const()[name = tensor("query_1_axes_0"), val = tensor([-1])]; - tensor module_layers_0_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_0_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9562368)))]; - tensor module_layers_0_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_0_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9564480)))]; - tensor query_1_cast_fp16 = layer_norm(axes = query_1_axes_0, beta = module_layers_0_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_0_norm_self_att_weight_to_fp16, x = input_31_cast_fp16)[name = tensor("query_1_cast_fp16")]; - tensor module_layers_0_self_attn_linear_q_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9566592))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10353088))), name = tensor("module_layers_0_self_attn_linear_q_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_3_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_0_self_attn_linear_q_weight_to_fp16_palettized, x = query_1_cast_fp16)[name = tensor("linear_3_cast_fp16")]; - tensor var_249 = const()[name = tensor("op_249"), val = tensor([1, -1, 8, 128])]; - tensor q_1_cast_fp16 = reshape(shape = var_249, x = linear_3_cast_fp16)[name = tensor("q_1_cast_fp16")]; - tensor module_layers_0_self_attn_linear_k_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10353280))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11139776))), name = tensor("module_layers_0_self_attn_linear_k_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_4_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_0_self_attn_linear_k_weight_to_fp16_palettized, x = query_1_cast_fp16)[name = tensor("linear_4_cast_fp16")]; - tensor var_253 = const()[name = tensor("op_253"), val = tensor([1, -1, 8, 128])]; - tensor k_1_cast_fp16 = reshape(shape = var_253, x = linear_4_cast_fp16)[name = tensor("k_1_cast_fp16")]; - tensor module_layers_0_self_attn_linear_v_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11139968))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11926464))), name = tensor("module_layers_0_self_attn_linear_v_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_5_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_0_self_attn_linear_v_weight_to_fp16_palettized, x = query_1_cast_fp16)[name = tensor("linear_5_cast_fp16")]; - tensor var_257 = const()[name = tensor("op_257"), val = tensor([1, -1, 8, 128])]; - tensor v_1_cast_fp16 = reshape(shape = var_257, x = linear_5_cast_fp16)[name = tensor("v_1_cast_fp16")]; - tensor value_3_perm_0 = const()[name = tensor("value_3_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor module_layers_0_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_0_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11926656)))]; - tensor var_269_cast_fp16 = add(x = q_1_cast_fp16, y = module_layers_0_self_attn_pos_bias_u_to_fp16)[name = tensor("op_269_cast_fp16")]; - tensor module_layers_0_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_0_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11928768)))]; - tensor var_271_cast_fp16 = add(x = q_1_cast_fp16, y = module_layers_0_self_attn_pos_bias_v_to_fp16)[name = tensor("op_271_cast_fp16")]; - tensor q_with_bias_v_1_perm_0 = const()[name = tensor("q_with_bias_v_1_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor x_7_transpose_x_0 = const()[name = tensor("x_7_transpose_x_0"), val = tensor(false)]; - tensor x_7_transpose_y_0 = const()[name = tensor("x_7_transpose_y_0"), val = tensor(false)]; - tensor op_273_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11930880))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12218944))), name = tensor("op_273_to_fp16_palettized"), shape = tensor([1, 8, 128, 375])]; - tensor q_with_bias_v_1_cast_fp16 = transpose(perm = q_with_bias_v_1_perm_0, x = var_271_cast_fp16)[name = tensor("transpose_312")]; - tensor x_7_cast_fp16 = matmul(transpose_x = x_7_transpose_x_0, transpose_y = x_7_transpose_y_0, x = q_with_bias_v_1_cast_fp16, y = op_273_to_fp16_palettized)[name = tensor("x_7_cast_fp16")]; - tensor x_9_pad_0 = const()[name = tensor("x_9_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_9_mode_0 = const()[name = tensor("x_9_mode_0"), val = tensor("constant")]; - tensor const_14_to_fp16 = const()[name = tensor("const_14_to_fp16"), val = tensor(0x0p+0)]; - tensor x_9_cast_fp16 = pad(constant_val = const_14_to_fp16, mode = x_9_mode_0, pad = x_9_pad_0, x = x_7_cast_fp16)[name = tensor("x_9_cast_fp16")]; - tensor var_281 = const()[name = tensor("op_281"), val = tensor([1, 8, -1, 188])]; - tensor x_11_cast_fp16 = reshape(shape = var_281, x = x_9_cast_fp16)[name = tensor("x_11_cast_fp16")]; - tensor var_285_begin_0 = const()[name = tensor("op_285_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_285_end_0 = const()[name = tensor("op_285_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_285_end_mask_0 = const()[name = tensor("op_285_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_285_cast_fp16 = slice_by_index(begin = var_285_begin_0, end = var_285_end_0, end_mask = var_285_end_mask_0, x = x_11_cast_fp16)[name = tensor("op_285_cast_fp16")]; - tensor var_286 = const()[name = tensor("op_286"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_1_cast_fp16 = reshape(shape = var_286, x = var_285_cast_fp16)[name = tensor("matrix_bd_1_cast_fp16")]; - tensor matrix_ac_1_transpose_x_0 = const()[name = tensor("matrix_ac_1_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_1_transpose_y_0 = const()[name = tensor("matrix_ac_1_transpose_y_0"), val = tensor(false)]; - tensor transpose_96_perm_0 = const()[name = tensor("transpose_96_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_97_perm_0 = const()[name = tensor("transpose_97_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_97 = transpose(perm = transpose_97_perm_0, x = k_1_cast_fp16)[name = tensor("transpose_310")]; - tensor transpose_96 = transpose(perm = transpose_96_perm_0, x = var_269_cast_fp16)[name = tensor("transpose_311")]; - tensor matrix_ac_1_cast_fp16 = matmul(transpose_x = matrix_ac_1_transpose_x_0, transpose_y = matrix_ac_1_transpose_y_0, x = transpose_96, y = transpose_97)[name = tensor("matrix_ac_1_cast_fp16")]; - tensor matrix_bd_3_begin_0 = const()[name = tensor("matrix_bd_3_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_3_end_0 = const()[name = tensor("matrix_bd_3_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_3_end_mask_0 = const()[name = tensor("matrix_bd_3_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_3_cast_fp16 = slice_by_index(begin = matrix_bd_3_begin_0, end = matrix_bd_3_end_0, end_mask = matrix_bd_3_end_mask_0, x = matrix_bd_1_cast_fp16)[name = tensor("matrix_bd_3_cast_fp16")]; - tensor var_295_cast_fp16 = add(x = matrix_ac_1_cast_fp16, y = matrix_bd_3_cast_fp16)[name = tensor("op_295_cast_fp16")]; - tensor _inversed_scores_1_y_0_to_fp16 = const()[name = tensor("_inversed_scores_1_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; - tensor _inversed_scores_1_cast_fp16 = mul(x = var_295_cast_fp16, y = _inversed_scores_1_y_0_to_fp16)[name = tensor("_inversed_scores_1_cast_fp16")]; - tensor mask_3_axes_0 = const()[name = tensor("mask_3_axes_0"), val = tensor([1])]; - tensor mask_3 = expand_dims(axes = mask_3_axes_0, x = mask_1)[name = tensor("mask_3")]; - tensor var_12_to_fp16 = const()[name = tensor("op_12_to_fp16"), val = tensor(-0x1.388p+13)]; - tensor scores_3_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_1_cast_fp16, cond = mask_3)[name = tensor("scores_3_cast_fp16")]; - tensor var_301_cast_fp16 = softmax(axis = var_30, x = scores_3_cast_fp16)[name = tensor("op_301_cast_fp16")]; - tensor var_11_to_fp16 = const()[name = tensor("op_11_to_fp16"), val = tensor(0x0p+0)]; - tensor input_33_cast_fp16 = select(a = var_11_to_fp16, b = var_301_cast_fp16, cond = mask_3)[name = tensor("input_33_cast_fp16")]; - tensor x_13_transpose_x_0 = const()[name = tensor("x_13_transpose_x_0"), val = tensor(false)]; - tensor x_13_transpose_y_0 = const()[name = tensor("x_13_transpose_y_0"), val = tensor(false)]; - tensor value_3_cast_fp16 = transpose(perm = value_3_perm_0, x = v_1_cast_fp16)[name = tensor("transpose_309")]; - tensor x_13_cast_fp16 = matmul(transpose_x = x_13_transpose_x_0, transpose_y = x_13_transpose_y_0, x = input_33_cast_fp16, y = value_3_cast_fp16)[name = tensor("x_13_cast_fp16")]; - tensor var_305_perm_0 = const()[name = tensor("op_305_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_306 = const()[name = tensor("op_306"), val = tensor([1, -1, 1024])]; - tensor var_305_cast_fp16 = transpose(perm = var_305_perm_0, x = x_13_cast_fp16)[name = tensor("transpose_308")]; - tensor input_35_cast_fp16 = reshape(shape = var_306, x = var_305_cast_fp16)[name = tensor("input_35_cast_fp16")]; - tensor module_layers_0_self_attn_linear_out_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12219136))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(13005632))), name = tensor("module_layers_0_self_attn_linear_out_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_7_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_0_self_attn_linear_out_weight_to_fp16_palettized, x = input_35_cast_fp16)[name = tensor("linear_7_cast_fp16")]; - tensor input_39_cast_fp16 = add(x = input_31_cast_fp16, y = linear_7_cast_fp16)[name = tensor("input_39_cast_fp16")]; - tensor x_17_axes_0 = const()[name = tensor("x_17_axes_0"), val = tensor([-1])]; - tensor module_layers_0_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_0_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(13005824)))]; - tensor module_layers_0_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_0_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(13007936)))]; - tensor x_17_cast_fp16 = layer_norm(axes = x_17_axes_0, beta = module_layers_0_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_0_norm_conv_weight_to_fp16, x = input_39_cast_fp16)[name = tensor("x_17_cast_fp16")]; - tensor input_41_perm_0 = const()[name = tensor("input_41_perm_0"), val = tensor([0, 2, 1])]; - tensor input_43_pad_type_0 = const()[name = tensor("input_43_pad_type_0"), val = tensor("valid")]; - tensor input_43_strides_0 = const()[name = tensor("input_43_strides_0"), val = tensor([1])]; - tensor input_43_pad_0 = const()[name = tensor("input_43_pad_0"), val = tensor([0, 0])]; - tensor input_43_dilations_0 = const()[name = tensor("input_43_dilations_0"), val = tensor([1])]; - tensor input_43_groups_0 = const()[name = tensor("input_43_groups_0"), val = tensor(1)]; - tensor module_layers_0_conv_pointwise_conv1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(13010048))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(14582976))), name = tensor("module_layers_0_conv_pointwise_conv1_weight_to_fp16_palettized"), shape = tensor([2048, 1024, 1])]; - tensor input_41_cast_fp16 = transpose(perm = input_41_perm_0, x = x_17_cast_fp16)[name = tensor("transpose_307")]; - tensor input_43_cast_fp16 = conv(dilations = input_43_dilations_0, groups = input_43_groups_0, pad = input_43_pad_0, pad_type = input_43_pad_type_0, strides = input_43_strides_0, weight = module_layers_0_conv_pointwise_conv1_weight_to_fp16_palettized, x = input_41_cast_fp16)[name = tensor("input_43_cast_fp16")]; - tensor x_19_split_num_splits_0 = const()[name = tensor("x_19_split_num_splits_0"), val = tensor(2)]; - tensor x_19_split_axis_0 = const()[name = tensor("x_19_split_axis_0"), val = tensor(1)]; - tensor x_19_split_cast_fp16_0, tensor x_19_split_cast_fp16_1 = split(axis = x_19_split_axis_0, num_splits = x_19_split_num_splits_0, x = input_43_cast_fp16)[name = tensor("x_19_split_cast_fp16")]; - tensor x_19_split_1_sigmoid_cast_fp16 = sigmoid(x = x_19_split_cast_fp16_1)[name = tensor("x_19_split_1_sigmoid_cast_fp16")]; - tensor x_19_cast_fp16 = mul(x = x_19_split_cast_fp16_0, y = x_19_split_1_sigmoid_cast_fp16)[name = tensor("x_19_cast_fp16")]; - tensor var_328_axes_0 = const()[name = tensor("op_328_axes_0"), val = tensor([1])]; - tensor var_328 = expand_dims(axes = var_328_axes_0, x = pad_mask)[name = tensor("op_328")]; - tensor input_45_cast_fp16 = select(a = var_11_to_fp16, b = x_19_cast_fp16, cond = var_328)[name = tensor("input_45_cast_fp16")]; - tensor input_47_pad_0 = const()[name = tensor("input_47_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_47_mode_0 = const()[name = tensor("input_47_mode_0"), val = tensor("constant")]; - tensor const_17_to_fp16 = const()[name = tensor("const_17_to_fp16"), val = tensor(0x0p+0)]; - tensor input_47_cast_fp16 = pad(constant_val = const_17_to_fp16, mode = input_47_mode_0, pad = input_47_pad_0, x = input_45_cast_fp16)[name = tensor("input_47_cast_fp16")]; - tensor input_49_pad_type_0 = const()[name = tensor("input_49_pad_type_0"), val = tensor("valid")]; - tensor input_49_groups_0 = const()[name = tensor("input_49_groups_0"), val = tensor(1024)]; - tensor input_49_strides_0 = const()[name = tensor("input_49_strides_0"), val = tensor([1])]; - tensor input_49_pad_0 = const()[name = tensor("input_49_pad_0"), val = tensor([0, 0])]; - tensor input_49_dilations_0 = const()[name = tensor("input_49_dilations_0"), val = tensor([1])]; - tensor const_248_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(14583168))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(14590144))), name = tensor("const_248_to_fp16_palettized"), shape = tensor([1024, 1, 9])]; - tensor const_249_to_fp16 = const()[name = tensor("const_249_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(14590336)))]; - tensor input_51_cast_fp16 = conv(bias = const_249_to_fp16, dilations = input_49_dilations_0, groups = input_49_groups_0, pad = input_49_pad_0, pad_type = input_49_pad_type_0, strides = input_49_strides_0, weight = const_248_to_fp16_palettized, x = input_47_cast_fp16)[name = tensor("input_51_cast_fp16")]; - tensor input_53_cast_fp16 = silu(x = input_51_cast_fp16)[name = tensor("input_53_cast_fp16")]; - tensor x_21_pad_type_0 = const()[name = tensor("x_21_pad_type_0"), val = tensor("valid")]; - tensor x_21_strides_0 = const()[name = tensor("x_21_strides_0"), val = tensor([1])]; - tensor x_21_pad_0 = const()[name = tensor("x_21_pad_0"), val = tensor([0, 0])]; - tensor x_21_dilations_0 = const()[name = tensor("x_21_dilations_0"), val = tensor([1])]; - tensor x_21_groups_0 = const()[name = tensor("x_21_groups_0"), val = tensor(1)]; - tensor module_layers_0_conv_pointwise_conv2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(14592448))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15378944))), name = tensor("module_layers_0_conv_pointwise_conv2_weight_to_fp16_palettized"), shape = tensor([1024, 1024, 1])]; - tensor x_21_cast_fp16 = conv(dilations = x_21_dilations_0, groups = x_21_groups_0, pad = x_21_pad_0, pad_type = x_21_pad_type_0, strides = x_21_strides_0, weight = module_layers_0_conv_pointwise_conv2_weight_to_fp16_palettized, x = input_53_cast_fp16)[name = tensor("x_21_cast_fp16")]; - tensor input_55_perm_0 = const()[name = tensor("input_55_perm_0"), val = tensor([0, 2, 1])]; - tensor input_55_cast_fp16 = transpose(perm = input_55_perm_0, x = x_21_cast_fp16)[name = tensor("transpose_306")]; - tensor input_57_cast_fp16 = add(x = input_39_cast_fp16, y = input_55_cast_fp16)[name = tensor("input_57_cast_fp16")]; - tensor input_59_axes_0 = const()[name = tensor("input_59_axes_0"), val = tensor([-1])]; - tensor module_layers_0_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_0_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15379136)))]; - tensor module_layers_0_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_0_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15381248)))]; - tensor input_59_cast_fp16 = layer_norm(axes = input_59_axes_0, beta = module_layers_0_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_0_norm_feed_forward2_weight_to_fp16, x = input_57_cast_fp16)[name = tensor("input_59_cast_fp16")]; - tensor module_layers_0_feed_forward2_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15383360))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(18529152))), name = tensor("module_layers_0_feed_forward2_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_8_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_0_feed_forward2_linear1_weight_to_fp16_palettized, x = input_59_cast_fp16)[name = tensor("linear_8_cast_fp16")]; - tensor input_63_cast_fp16 = silu(x = linear_8_cast_fp16)[name = tensor("input_63_cast_fp16")]; - tensor module_layers_0_feed_forward2_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(18529344))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21675136))), name = tensor("module_layers_0_feed_forward2_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_9_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_0_feed_forward2_linear2_weight_to_fp16_palettized, x = input_63_cast_fp16)[name = tensor("linear_9_cast_fp16")]; - tensor var_366_to_fp16 = const()[name = tensor("op_366_to_fp16"), val = tensor(0x1p-1)]; - tensor var_367_cast_fp16 = mul(x = linear_9_cast_fp16, y = var_366_to_fp16)[name = tensor("op_367_cast_fp16")]; - tensor input_69_cast_fp16 = add(x = input_57_cast_fp16, y = var_367_cast_fp16)[name = tensor("input_69_cast_fp16")]; - tensor input_71_axes_0 = const()[name = tensor("input_71_axes_0"), val = tensor([-1])]; - tensor module_layers_0_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_0_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21675328)))]; - tensor module_layers_0_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_0_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21677440)))]; - tensor input_71_cast_fp16 = layer_norm(axes = input_71_axes_0, beta = module_layers_0_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_0_norm_out_weight_to_fp16, x = input_69_cast_fp16)[name = tensor("input_71_cast_fp16")]; - tensor input_73_axes_0 = const()[name = tensor("input_73_axes_0"), val = tensor([-1])]; - tensor module_layers_1_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_1_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21679552)))]; - tensor module_layers_1_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_1_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21681664)))]; - tensor input_73_cast_fp16 = layer_norm(axes = input_73_axes_0, beta = module_layers_1_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_1_norm_feed_forward1_weight_to_fp16, x = input_71_cast_fp16)[name = tensor("input_73_cast_fp16")]; - tensor module_layers_1_feed_forward1_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21683776))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(24829568))), name = tensor("module_layers_1_feed_forward1_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_10_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_1_feed_forward1_linear1_weight_to_fp16_palettized, x = input_73_cast_fp16)[name = tensor("linear_10_cast_fp16")]; - tensor input_77_cast_fp16 = silu(x = linear_10_cast_fp16)[name = tensor("input_77_cast_fp16")]; - tensor module_layers_1_feed_forward1_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(24829760))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27975552))), name = tensor("module_layers_1_feed_forward1_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_11_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_1_feed_forward1_linear2_weight_to_fp16_palettized, x = input_77_cast_fp16)[name = tensor("linear_11_cast_fp16")]; - tensor var_395_to_fp16 = const()[name = tensor("op_395_to_fp16"), val = tensor(0x1p-1)]; - tensor var_396_cast_fp16 = mul(x = linear_11_cast_fp16, y = var_395_to_fp16)[name = tensor("op_396_cast_fp16")]; - tensor input_83_cast_fp16 = add(x = input_71_cast_fp16, y = var_396_cast_fp16)[name = tensor("input_83_cast_fp16")]; - tensor query_3_axes_0 = const()[name = tensor("query_3_axes_0"), val = tensor([-1])]; - tensor module_layers_1_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_1_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27975744)))]; - tensor module_layers_1_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_1_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27977856)))]; - tensor query_3_cast_fp16 = layer_norm(axes = query_3_axes_0, beta = module_layers_1_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_1_norm_self_att_weight_to_fp16, x = input_83_cast_fp16)[name = tensor("query_3_cast_fp16")]; - tensor module_layers_1_self_attn_linear_q_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27979968))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(28766464))), name = tensor("module_layers_1_self_attn_linear_q_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_12_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_1_self_attn_linear_q_weight_to_fp16_palettized, x = query_3_cast_fp16)[name = tensor("linear_12_cast_fp16")]; - tensor var_412 = const()[name = tensor("op_412"), val = tensor([1, -1, 8, 128])]; - tensor q_7_cast_fp16 = reshape(shape = var_412, x = linear_12_cast_fp16)[name = tensor("q_7_cast_fp16")]; - tensor module_layers_1_self_attn_linear_k_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(28766656))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(29553152))), name = tensor("module_layers_1_self_attn_linear_k_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_13_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_1_self_attn_linear_k_weight_to_fp16_palettized, x = query_3_cast_fp16)[name = tensor("linear_13_cast_fp16")]; - tensor var_416 = const()[name = tensor("op_416"), val = tensor([1, -1, 8, 128])]; - tensor k_5_cast_fp16 = reshape(shape = var_416, x = linear_13_cast_fp16)[name = tensor("k_5_cast_fp16")]; - tensor module_layers_1_self_attn_linear_v_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(29553344))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(30339840))), name = tensor("module_layers_1_self_attn_linear_v_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_14_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_1_self_attn_linear_v_weight_to_fp16_palettized, x = query_3_cast_fp16)[name = tensor("linear_14_cast_fp16")]; - tensor var_420 = const()[name = tensor("op_420"), val = tensor([1, -1, 8, 128])]; - tensor v_3_cast_fp16 = reshape(shape = var_420, x = linear_14_cast_fp16)[name = tensor("v_3_cast_fp16")]; - tensor value_5_perm_0 = const()[name = tensor("value_5_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor module_layers_1_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_1_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(30340032)))]; - tensor var_432_cast_fp16 = add(x = q_7_cast_fp16, y = module_layers_1_self_attn_pos_bias_u_to_fp16)[name = tensor("op_432_cast_fp16")]; - tensor module_layers_1_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_1_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(30342144)))]; - tensor var_434_cast_fp16 = add(x = q_7_cast_fp16, y = module_layers_1_self_attn_pos_bias_v_to_fp16)[name = tensor("op_434_cast_fp16")]; - tensor q_with_bias_v_3_perm_0 = const()[name = tensor("q_with_bias_v_3_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor x_29_transpose_x_0 = const()[name = tensor("x_29_transpose_x_0"), val = tensor(false)]; - tensor x_29_transpose_y_0 = const()[name = tensor("x_29_transpose_y_0"), val = tensor(false)]; - tensor op_436_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(30344256))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(30632320))), name = tensor("op_436_to_fp16_palettized"), shape = tensor([1, 8, 128, 375])]; - tensor q_with_bias_v_3_cast_fp16 = transpose(perm = q_with_bias_v_3_perm_0, x = var_434_cast_fp16)[name = tensor("transpose_305")]; - tensor x_29_cast_fp16 = matmul(transpose_x = x_29_transpose_x_0, transpose_y = x_29_transpose_y_0, x = q_with_bias_v_3_cast_fp16, y = op_436_to_fp16_palettized)[name = tensor("x_29_cast_fp16")]; - tensor x_31_pad_0 = const()[name = tensor("x_31_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_31_mode_0 = const()[name = tensor("x_31_mode_0"), val = tensor("constant")]; - tensor const_24_to_fp16 = const()[name = tensor("const_24_to_fp16"), val = tensor(0x0p+0)]; - tensor x_31_cast_fp16 = pad(constant_val = const_24_to_fp16, mode = x_31_mode_0, pad = x_31_pad_0, x = x_29_cast_fp16)[name = tensor("x_31_cast_fp16")]; - tensor var_444 = const()[name = tensor("op_444"), val = tensor([1, 8, -1, 188])]; - tensor x_33_cast_fp16 = reshape(shape = var_444, x = x_31_cast_fp16)[name = tensor("x_33_cast_fp16")]; - tensor var_448_begin_0 = const()[name = tensor("op_448_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_448_end_0 = const()[name = tensor("op_448_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_448_end_mask_0 = const()[name = tensor("op_448_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_448_cast_fp16 = slice_by_index(begin = var_448_begin_0, end = var_448_end_0, end_mask = var_448_end_mask_0, x = x_33_cast_fp16)[name = tensor("op_448_cast_fp16")]; - tensor var_449 = const()[name = tensor("op_449"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_5_cast_fp16 = reshape(shape = var_449, x = var_448_cast_fp16)[name = tensor("matrix_bd_5_cast_fp16")]; - tensor matrix_ac_3_transpose_x_0 = const()[name = tensor("matrix_ac_3_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_3_transpose_y_0 = const()[name = tensor("matrix_ac_3_transpose_y_0"), val = tensor(false)]; - tensor transpose_98_perm_0 = const()[name = tensor("transpose_98_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_99_perm_0 = const()[name = tensor("transpose_99_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_99 = transpose(perm = transpose_99_perm_0, x = k_5_cast_fp16)[name = tensor("transpose_303")]; - tensor transpose_98 = transpose(perm = transpose_98_perm_0, x = var_432_cast_fp16)[name = tensor("transpose_304")]; - tensor matrix_ac_3_cast_fp16 = matmul(transpose_x = matrix_ac_3_transpose_x_0, transpose_y = matrix_ac_3_transpose_y_0, x = transpose_98, y = transpose_99)[name = tensor("matrix_ac_3_cast_fp16")]; - tensor matrix_bd_7_begin_0 = const()[name = tensor("matrix_bd_7_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_7_end_0 = const()[name = tensor("matrix_bd_7_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_7_end_mask_0 = const()[name = tensor("matrix_bd_7_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_7_cast_fp16 = slice_by_index(begin = matrix_bd_7_begin_0, end = matrix_bd_7_end_0, end_mask = matrix_bd_7_end_mask_0, x = matrix_bd_5_cast_fp16)[name = tensor("matrix_bd_7_cast_fp16")]; - tensor var_458_cast_fp16 = add(x = matrix_ac_3_cast_fp16, y = matrix_bd_7_cast_fp16)[name = tensor("op_458_cast_fp16")]; - tensor _inversed_scores_5_y_0_to_fp16 = const()[name = tensor("_inversed_scores_5_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; - tensor _inversed_scores_5_cast_fp16 = mul(x = var_458_cast_fp16, y = _inversed_scores_5_y_0_to_fp16)[name = tensor("_inversed_scores_5_cast_fp16")]; - tensor scores_7_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_5_cast_fp16, cond = mask_3)[name = tensor("scores_7_cast_fp16")]; - tensor var_464_cast_fp16 = softmax(axis = var_30, x = scores_7_cast_fp16)[name = tensor("op_464_cast_fp16")]; - tensor input_85_cast_fp16 = select(a = var_11_to_fp16, b = var_464_cast_fp16, cond = mask_3)[name = tensor("input_85_cast_fp16")]; - tensor x_35_transpose_x_0 = const()[name = tensor("x_35_transpose_x_0"), val = tensor(false)]; - tensor x_35_transpose_y_0 = const()[name = tensor("x_35_transpose_y_0"), val = tensor(false)]; - tensor value_5_cast_fp16 = transpose(perm = value_5_perm_0, x = v_3_cast_fp16)[name = tensor("transpose_302")]; - tensor x_35_cast_fp16 = matmul(transpose_x = x_35_transpose_x_0, transpose_y = x_35_transpose_y_0, x = input_85_cast_fp16, y = value_5_cast_fp16)[name = tensor("x_35_cast_fp16")]; - tensor var_468_perm_0 = const()[name = tensor("op_468_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_469 = const()[name = tensor("op_469"), val = tensor([1, -1, 1024])]; - tensor var_468_cast_fp16 = transpose(perm = var_468_perm_0, x = x_35_cast_fp16)[name = tensor("transpose_301")]; - tensor input_87_cast_fp16 = reshape(shape = var_469, x = var_468_cast_fp16)[name = tensor("input_87_cast_fp16")]; - tensor module_layers_1_self_attn_linear_out_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(30632512))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(31419008))), name = tensor("module_layers_1_self_attn_linear_out_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_16_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_1_self_attn_linear_out_weight_to_fp16_palettized, x = input_87_cast_fp16)[name = tensor("linear_16_cast_fp16")]; - tensor input_91_cast_fp16 = add(x = input_83_cast_fp16, y = linear_16_cast_fp16)[name = tensor("input_91_cast_fp16")]; - tensor x_39_axes_0 = const()[name = tensor("x_39_axes_0"), val = tensor([-1])]; - tensor module_layers_1_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_1_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(31419200)))]; - tensor module_layers_1_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_1_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(31421312)))]; - tensor x_39_cast_fp16 = layer_norm(axes = x_39_axes_0, beta = module_layers_1_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_1_norm_conv_weight_to_fp16, x = input_91_cast_fp16)[name = tensor("x_39_cast_fp16")]; - tensor input_93_perm_0 = const()[name = tensor("input_93_perm_0"), val = tensor([0, 2, 1])]; - tensor input_95_pad_type_0 = const()[name = tensor("input_95_pad_type_0"), val = tensor("valid")]; - tensor input_95_strides_0 = const()[name = tensor("input_95_strides_0"), val = tensor([1])]; - tensor input_95_pad_0 = const()[name = tensor("input_95_pad_0"), val = tensor([0, 0])]; - tensor input_95_dilations_0 = const()[name = tensor("input_95_dilations_0"), val = tensor([1])]; - tensor input_95_groups_0 = const()[name = tensor("input_95_groups_0"), val = tensor(1)]; - tensor module_layers_1_conv_pointwise_conv1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(31423424))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32996352))), name = tensor("module_layers_1_conv_pointwise_conv1_weight_to_fp16_palettized"), shape = tensor([2048, 1024, 1])]; - tensor input_93_cast_fp16 = transpose(perm = input_93_perm_0, x = x_39_cast_fp16)[name = tensor("transpose_300")]; - tensor input_95_cast_fp16 = conv(dilations = input_95_dilations_0, groups = input_95_groups_0, pad = input_95_pad_0, pad_type = input_95_pad_type_0, strides = input_95_strides_0, weight = module_layers_1_conv_pointwise_conv1_weight_to_fp16_palettized, x = input_93_cast_fp16)[name = tensor("input_95_cast_fp16")]; - tensor x_41_split_num_splits_0 = const()[name = tensor("x_41_split_num_splits_0"), val = tensor(2)]; - tensor x_41_split_axis_0 = const()[name = tensor("x_41_split_axis_0"), val = tensor(1)]; - tensor x_41_split_cast_fp16_0, tensor x_41_split_cast_fp16_1 = split(axis = x_41_split_axis_0, num_splits = x_41_split_num_splits_0, x = input_95_cast_fp16)[name = tensor("x_41_split_cast_fp16")]; - tensor x_41_split_1_sigmoid_cast_fp16 = sigmoid(x = x_41_split_cast_fp16_1)[name = tensor("x_41_split_1_sigmoid_cast_fp16")]; - tensor x_41_cast_fp16 = mul(x = x_41_split_cast_fp16_0, y = x_41_split_1_sigmoid_cast_fp16)[name = tensor("x_41_cast_fp16")]; - tensor input_97_cast_fp16 = select(a = var_11_to_fp16, b = x_41_cast_fp16, cond = var_328)[name = tensor("input_97_cast_fp16")]; - tensor input_99_pad_0 = const()[name = tensor("input_99_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_99_mode_0 = const()[name = tensor("input_99_mode_0"), val = tensor("constant")]; - tensor const_27_to_fp16 = const()[name = tensor("const_27_to_fp16"), val = tensor(0x0p+0)]; - tensor input_99_cast_fp16 = pad(constant_val = const_27_to_fp16, mode = input_99_mode_0, pad = input_99_pad_0, x = input_97_cast_fp16)[name = tensor("input_99_cast_fp16")]; - tensor input_101_pad_type_0 = const()[name = tensor("input_101_pad_type_0"), val = tensor("valid")]; - tensor input_101_groups_0 = const()[name = tensor("input_101_groups_0"), val = tensor(1024)]; - tensor input_101_strides_0 = const()[name = tensor("input_101_strides_0"), val = tensor([1])]; - tensor input_101_pad_0 = const()[name = tensor("input_101_pad_0"), val = tensor([0, 0])]; - tensor input_101_dilations_0 = const()[name = tensor("input_101_dilations_0"), val = tensor([1])]; - tensor const_250_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32996544))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33003520))), name = tensor("const_250_to_fp16_palettized"), shape = tensor([1024, 1, 9])]; - tensor const_251_to_fp16 = const()[name = tensor("const_251_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33003712)))]; - tensor input_103_cast_fp16 = conv(bias = const_251_to_fp16, dilations = input_101_dilations_0, groups = input_101_groups_0, pad = input_101_pad_0, pad_type = input_101_pad_type_0, strides = input_101_strides_0, weight = const_250_to_fp16_palettized, x = input_99_cast_fp16)[name = tensor("input_103_cast_fp16")]; - tensor input_105_cast_fp16 = silu(x = input_103_cast_fp16)[name = tensor("input_105_cast_fp16")]; - tensor x_43_pad_type_0 = const()[name = tensor("x_43_pad_type_0"), val = tensor("valid")]; - tensor x_43_strides_0 = const()[name = tensor("x_43_strides_0"), val = tensor([1])]; - tensor x_43_pad_0 = const()[name = tensor("x_43_pad_0"), val = tensor([0, 0])]; - tensor x_43_dilations_0 = const()[name = tensor("x_43_dilations_0"), val = tensor([1])]; - tensor x_43_groups_0 = const()[name = tensor("x_43_groups_0"), val = tensor(1)]; - tensor module_layers_1_conv_pointwise_conv2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33005824))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33792320))), name = tensor("module_layers_1_conv_pointwise_conv2_weight_to_fp16_palettized"), shape = tensor([1024, 1024, 1])]; - tensor x_43_cast_fp16 = conv(dilations = x_43_dilations_0, groups = x_43_groups_0, pad = x_43_pad_0, pad_type = x_43_pad_type_0, strides = x_43_strides_0, weight = module_layers_1_conv_pointwise_conv2_weight_to_fp16_palettized, x = input_105_cast_fp16)[name = tensor("x_43_cast_fp16")]; - tensor input_107_perm_0 = const()[name = tensor("input_107_perm_0"), val = tensor([0, 2, 1])]; - tensor input_107_cast_fp16 = transpose(perm = input_107_perm_0, x = x_43_cast_fp16)[name = tensor("transpose_299")]; - tensor input_109_cast_fp16 = add(x = input_91_cast_fp16, y = input_107_cast_fp16)[name = tensor("input_109_cast_fp16")]; - tensor input_111_axes_0 = const()[name = tensor("input_111_axes_0"), val = tensor([-1])]; - tensor module_layers_1_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_1_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33792512)))]; - tensor module_layers_1_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_1_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33794624)))]; - tensor input_111_cast_fp16 = layer_norm(axes = input_111_axes_0, beta = module_layers_1_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_1_norm_feed_forward2_weight_to_fp16, x = input_109_cast_fp16)[name = tensor("input_111_cast_fp16")]; - tensor module_layers_1_feed_forward2_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33796736))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36942528))), name = tensor("module_layers_1_feed_forward2_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_17_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_1_feed_forward2_linear1_weight_to_fp16_palettized, x = input_111_cast_fp16)[name = tensor("linear_17_cast_fp16")]; - tensor input_115_cast_fp16 = silu(x = linear_17_cast_fp16)[name = tensor("input_115_cast_fp16")]; - tensor module_layers_1_feed_forward2_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36942720))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40088512))), name = tensor("module_layers_1_feed_forward2_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_18_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_1_feed_forward2_linear2_weight_to_fp16_palettized, x = input_115_cast_fp16)[name = tensor("linear_18_cast_fp16")]; - tensor var_529_to_fp16 = const()[name = tensor("op_529_to_fp16"), val = tensor(0x1p-1)]; - tensor var_530_cast_fp16 = mul(x = linear_18_cast_fp16, y = var_529_to_fp16)[name = tensor("op_530_cast_fp16")]; - tensor input_121_cast_fp16 = add(x = input_109_cast_fp16, y = var_530_cast_fp16)[name = tensor("input_121_cast_fp16")]; - tensor input_123_axes_0 = const()[name = tensor("input_123_axes_0"), val = tensor([-1])]; - tensor module_layers_1_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_1_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40088704)))]; - tensor module_layers_1_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_1_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40090816)))]; - tensor input_123_cast_fp16 = layer_norm(axes = input_123_axes_0, beta = module_layers_1_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_1_norm_out_weight_to_fp16, x = input_121_cast_fp16)[name = tensor("input_123_cast_fp16")]; - tensor input_125_axes_0 = const()[name = tensor("input_125_axes_0"), val = tensor([-1])]; - tensor module_layers_2_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_2_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40092928)))]; - tensor module_layers_2_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_2_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40095040)))]; - tensor input_125_cast_fp16 = layer_norm(axes = input_125_axes_0, beta = module_layers_2_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_2_norm_feed_forward1_weight_to_fp16, x = input_123_cast_fp16)[name = tensor("input_125_cast_fp16")]; - tensor module_layers_2_feed_forward1_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40097152))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(43242944))), name = tensor("module_layers_2_feed_forward1_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_19_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_2_feed_forward1_linear1_weight_to_fp16_palettized, x = input_125_cast_fp16)[name = tensor("linear_19_cast_fp16")]; - tensor input_129_cast_fp16 = silu(x = linear_19_cast_fp16)[name = tensor("input_129_cast_fp16")]; - tensor module_layers_2_feed_forward1_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(43243136))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46388928))), name = tensor("module_layers_2_feed_forward1_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_20_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_2_feed_forward1_linear2_weight_to_fp16_palettized, x = input_129_cast_fp16)[name = tensor("linear_20_cast_fp16")]; - tensor var_558_to_fp16 = const()[name = tensor("op_558_to_fp16"), val = tensor(0x1p-1)]; - tensor var_559_cast_fp16 = mul(x = linear_20_cast_fp16, y = var_558_to_fp16)[name = tensor("op_559_cast_fp16")]; - tensor input_135_cast_fp16 = add(x = input_123_cast_fp16, y = var_559_cast_fp16)[name = tensor("input_135_cast_fp16")]; - tensor query_5_axes_0 = const()[name = tensor("query_5_axes_0"), val = tensor([-1])]; - tensor module_layers_2_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_2_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46389120)))]; - tensor module_layers_2_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_2_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46391232)))]; - tensor query_5_cast_fp16 = layer_norm(axes = query_5_axes_0, beta = module_layers_2_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_2_norm_self_att_weight_to_fp16, x = input_135_cast_fp16)[name = tensor("query_5_cast_fp16")]; - tensor module_layers_2_self_attn_linear_q_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46393344))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47179840))), name = tensor("module_layers_2_self_attn_linear_q_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_21_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_2_self_attn_linear_q_weight_to_fp16_palettized, x = query_5_cast_fp16)[name = tensor("linear_21_cast_fp16")]; - tensor var_575 = const()[name = tensor("op_575"), val = tensor([1, -1, 8, 128])]; - tensor q_13_cast_fp16 = reshape(shape = var_575, x = linear_21_cast_fp16)[name = tensor("q_13_cast_fp16")]; - tensor module_layers_2_self_attn_linear_k_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47180032))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47966528))), name = tensor("module_layers_2_self_attn_linear_k_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_22_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_2_self_attn_linear_k_weight_to_fp16_palettized, x = query_5_cast_fp16)[name = tensor("linear_22_cast_fp16")]; - tensor var_579 = const()[name = tensor("op_579"), val = tensor([1, -1, 8, 128])]; - tensor k_9_cast_fp16 = reshape(shape = var_579, x = linear_22_cast_fp16)[name = tensor("k_9_cast_fp16")]; - tensor module_layers_2_self_attn_linear_v_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47966720))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48753216))), name = tensor("module_layers_2_self_attn_linear_v_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_23_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_2_self_attn_linear_v_weight_to_fp16_palettized, x = query_5_cast_fp16)[name = tensor("linear_23_cast_fp16")]; - tensor var_583 = const()[name = tensor("op_583"), val = tensor([1, -1, 8, 128])]; - tensor v_5_cast_fp16 = reshape(shape = var_583, x = linear_23_cast_fp16)[name = tensor("v_5_cast_fp16")]; - tensor value_7_perm_0 = const()[name = tensor("value_7_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor module_layers_2_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_2_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48753408)))]; - tensor var_595_cast_fp16 = add(x = q_13_cast_fp16, y = module_layers_2_self_attn_pos_bias_u_to_fp16)[name = tensor("op_595_cast_fp16")]; - tensor module_layers_2_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_2_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48755520)))]; - tensor var_597_cast_fp16 = add(x = q_13_cast_fp16, y = module_layers_2_self_attn_pos_bias_v_to_fp16)[name = tensor("op_597_cast_fp16")]; - tensor q_with_bias_v_5_perm_0 = const()[name = tensor("q_with_bias_v_5_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor x_51_transpose_x_0 = const()[name = tensor("x_51_transpose_x_0"), val = tensor(false)]; - tensor x_51_transpose_y_0 = const()[name = tensor("x_51_transpose_y_0"), val = tensor(false)]; - tensor op_599_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48757632))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(49045696))), name = tensor("op_599_to_fp16_palettized"), shape = tensor([1, 8, 128, 375])]; - tensor q_with_bias_v_5_cast_fp16 = transpose(perm = q_with_bias_v_5_perm_0, x = var_597_cast_fp16)[name = tensor("transpose_298")]; - tensor x_51_cast_fp16 = matmul(transpose_x = x_51_transpose_x_0, transpose_y = x_51_transpose_y_0, x = q_with_bias_v_5_cast_fp16, y = op_599_to_fp16_palettized)[name = tensor("x_51_cast_fp16")]; - tensor x_53_pad_0 = const()[name = tensor("x_53_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_53_mode_0 = const()[name = tensor("x_53_mode_0"), val = tensor("constant")]; - tensor const_34_to_fp16 = const()[name = tensor("const_34_to_fp16"), val = tensor(0x0p+0)]; - tensor x_53_cast_fp16 = pad(constant_val = const_34_to_fp16, mode = x_53_mode_0, pad = x_53_pad_0, x = x_51_cast_fp16)[name = tensor("x_53_cast_fp16")]; - tensor var_607 = const()[name = tensor("op_607"), val = tensor([1, 8, -1, 188])]; - tensor x_55_cast_fp16 = reshape(shape = var_607, x = x_53_cast_fp16)[name = tensor("x_55_cast_fp16")]; - tensor var_611_begin_0 = const()[name = tensor("op_611_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_611_end_0 = const()[name = tensor("op_611_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_611_end_mask_0 = const()[name = tensor("op_611_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_611_cast_fp16 = slice_by_index(begin = var_611_begin_0, end = var_611_end_0, end_mask = var_611_end_mask_0, x = x_55_cast_fp16)[name = tensor("op_611_cast_fp16")]; - tensor var_612 = const()[name = tensor("op_612"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_9_cast_fp16 = reshape(shape = var_612, x = var_611_cast_fp16)[name = tensor("matrix_bd_9_cast_fp16")]; - tensor matrix_ac_5_transpose_x_0 = const()[name = tensor("matrix_ac_5_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_5_transpose_y_0 = const()[name = tensor("matrix_ac_5_transpose_y_0"), val = tensor(false)]; - tensor transpose_100_perm_0 = const()[name = tensor("transpose_100_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_101_perm_0 = const()[name = tensor("transpose_101_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_101 = transpose(perm = transpose_101_perm_0, x = k_9_cast_fp16)[name = tensor("transpose_296")]; - tensor transpose_100 = transpose(perm = transpose_100_perm_0, x = var_595_cast_fp16)[name = tensor("transpose_297")]; - tensor matrix_ac_5_cast_fp16 = matmul(transpose_x = matrix_ac_5_transpose_x_0, transpose_y = matrix_ac_5_transpose_y_0, x = transpose_100, y = transpose_101)[name = tensor("matrix_ac_5_cast_fp16")]; - tensor matrix_bd_11_begin_0 = const()[name = tensor("matrix_bd_11_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_11_end_0 = const()[name = tensor("matrix_bd_11_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_11_end_mask_0 = const()[name = tensor("matrix_bd_11_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_11_cast_fp16 = slice_by_index(begin = matrix_bd_11_begin_0, end = matrix_bd_11_end_0, end_mask = matrix_bd_11_end_mask_0, x = matrix_bd_9_cast_fp16)[name = tensor("matrix_bd_11_cast_fp16")]; - tensor var_621_cast_fp16 = add(x = matrix_ac_5_cast_fp16, y = matrix_bd_11_cast_fp16)[name = tensor("op_621_cast_fp16")]; - tensor _inversed_scores_9_y_0_to_fp16 = const()[name = tensor("_inversed_scores_9_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; - tensor _inversed_scores_9_cast_fp16 = mul(x = var_621_cast_fp16, y = _inversed_scores_9_y_0_to_fp16)[name = tensor("_inversed_scores_9_cast_fp16")]; - tensor scores_11_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_9_cast_fp16, cond = mask_3)[name = tensor("scores_11_cast_fp16")]; - tensor var_627_cast_fp16 = softmax(axis = var_30, x = scores_11_cast_fp16)[name = tensor("op_627_cast_fp16")]; - tensor input_137_cast_fp16 = select(a = var_11_to_fp16, b = var_627_cast_fp16, cond = mask_3)[name = tensor("input_137_cast_fp16")]; - tensor x_57_transpose_x_0 = const()[name = tensor("x_57_transpose_x_0"), val = tensor(false)]; - tensor x_57_transpose_y_0 = const()[name = tensor("x_57_transpose_y_0"), val = tensor(false)]; - tensor value_7_cast_fp16 = transpose(perm = value_7_perm_0, x = v_5_cast_fp16)[name = tensor("transpose_295")]; - tensor x_57_cast_fp16 = matmul(transpose_x = x_57_transpose_x_0, transpose_y = x_57_transpose_y_0, x = input_137_cast_fp16, y = value_7_cast_fp16)[name = tensor("x_57_cast_fp16")]; - tensor var_631_perm_0 = const()[name = tensor("op_631_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_632 = const()[name = tensor("op_632"), val = tensor([1, -1, 1024])]; - tensor var_631_cast_fp16 = transpose(perm = var_631_perm_0, x = x_57_cast_fp16)[name = tensor("transpose_294")]; - tensor input_139_cast_fp16 = reshape(shape = var_632, x = var_631_cast_fp16)[name = tensor("input_139_cast_fp16")]; - tensor module_layers_2_self_attn_linear_out_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(49045888))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(49832384))), name = tensor("module_layers_2_self_attn_linear_out_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_25_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_2_self_attn_linear_out_weight_to_fp16_palettized, x = input_139_cast_fp16)[name = tensor("linear_25_cast_fp16")]; - tensor input_143_cast_fp16 = add(x = input_135_cast_fp16, y = linear_25_cast_fp16)[name = tensor("input_143_cast_fp16")]; - tensor x_61_axes_0 = const()[name = tensor("x_61_axes_0"), val = tensor([-1])]; - tensor module_layers_2_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_2_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(49832576)))]; - tensor module_layers_2_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_2_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(49834688)))]; - tensor x_61_cast_fp16 = layer_norm(axes = x_61_axes_0, beta = module_layers_2_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_2_norm_conv_weight_to_fp16, x = input_143_cast_fp16)[name = tensor("x_61_cast_fp16")]; - tensor input_145_perm_0 = const()[name = tensor("input_145_perm_0"), val = tensor([0, 2, 1])]; - tensor input_147_pad_type_0 = const()[name = tensor("input_147_pad_type_0"), val = tensor("valid")]; - tensor input_147_strides_0 = const()[name = tensor("input_147_strides_0"), val = tensor([1])]; - tensor input_147_pad_0 = const()[name = tensor("input_147_pad_0"), val = tensor([0, 0])]; - tensor input_147_dilations_0 = const()[name = tensor("input_147_dilations_0"), val = tensor([1])]; - tensor input_147_groups_0 = const()[name = tensor("input_147_groups_0"), val = tensor(1)]; - tensor module_layers_2_conv_pointwise_conv1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(49836800))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(51409728))), name = tensor("module_layers_2_conv_pointwise_conv1_weight_to_fp16_palettized"), shape = tensor([2048, 1024, 1])]; - tensor input_145_cast_fp16 = transpose(perm = input_145_perm_0, x = x_61_cast_fp16)[name = tensor("transpose_293")]; - tensor input_147_cast_fp16 = conv(dilations = input_147_dilations_0, groups = input_147_groups_0, pad = input_147_pad_0, pad_type = input_147_pad_type_0, strides = input_147_strides_0, weight = module_layers_2_conv_pointwise_conv1_weight_to_fp16_palettized, x = input_145_cast_fp16)[name = tensor("input_147_cast_fp16")]; - tensor x_63_split_num_splits_0 = const()[name = tensor("x_63_split_num_splits_0"), val = tensor(2)]; - tensor x_63_split_axis_0 = const()[name = tensor("x_63_split_axis_0"), val = tensor(1)]; - tensor x_63_split_cast_fp16_0, tensor x_63_split_cast_fp16_1 = split(axis = x_63_split_axis_0, num_splits = x_63_split_num_splits_0, x = input_147_cast_fp16)[name = tensor("x_63_split_cast_fp16")]; - tensor x_63_split_1_sigmoid_cast_fp16 = sigmoid(x = x_63_split_cast_fp16_1)[name = tensor("x_63_split_1_sigmoid_cast_fp16")]; - tensor x_63_cast_fp16 = mul(x = x_63_split_cast_fp16_0, y = x_63_split_1_sigmoid_cast_fp16)[name = tensor("x_63_cast_fp16")]; - tensor input_149_cast_fp16 = select(a = var_11_to_fp16, b = x_63_cast_fp16, cond = var_328)[name = tensor("input_149_cast_fp16")]; - tensor input_151_pad_0 = const()[name = tensor("input_151_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_151_mode_0 = const()[name = tensor("input_151_mode_0"), val = tensor("constant")]; - tensor const_37_to_fp16 = const()[name = tensor("const_37_to_fp16"), val = tensor(0x0p+0)]; - tensor input_151_cast_fp16 = pad(constant_val = const_37_to_fp16, mode = input_151_mode_0, pad = input_151_pad_0, x = input_149_cast_fp16)[name = tensor("input_151_cast_fp16")]; - tensor input_153_pad_type_0 = const()[name = tensor("input_153_pad_type_0"), val = tensor("valid")]; - tensor input_153_groups_0 = const()[name = tensor("input_153_groups_0"), val = tensor(1024)]; - tensor input_153_strides_0 = const()[name = tensor("input_153_strides_0"), val = tensor([1])]; - tensor input_153_pad_0 = const()[name = tensor("input_153_pad_0"), val = tensor([0, 0])]; - tensor input_153_dilations_0 = const()[name = tensor("input_153_dilations_0"), val = tensor([1])]; - tensor const_252_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(51409920))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(51416896))), name = tensor("const_252_to_fp16_palettized"), shape = tensor([1024, 1, 9])]; - tensor const_253_to_fp16 = const()[name = tensor("const_253_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(51417088)))]; - tensor input_155_cast_fp16 = conv(bias = const_253_to_fp16, dilations = input_153_dilations_0, groups = input_153_groups_0, pad = input_153_pad_0, pad_type = input_153_pad_type_0, strides = input_153_strides_0, weight = const_252_to_fp16_palettized, x = input_151_cast_fp16)[name = tensor("input_155_cast_fp16")]; - tensor input_157_cast_fp16 = silu(x = input_155_cast_fp16)[name = tensor("input_157_cast_fp16")]; - tensor x_65_pad_type_0 = const()[name = tensor("x_65_pad_type_0"), val = tensor("valid")]; - tensor x_65_strides_0 = const()[name = tensor("x_65_strides_0"), val = tensor([1])]; - tensor x_65_pad_0 = const()[name = tensor("x_65_pad_0"), val = tensor([0, 0])]; - tensor x_65_dilations_0 = const()[name = tensor("x_65_dilations_0"), val = tensor([1])]; - tensor x_65_groups_0 = const()[name = tensor("x_65_groups_0"), val = tensor(1)]; - tensor module_layers_2_conv_pointwise_conv2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(51419200))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52205696))), name = tensor("module_layers_2_conv_pointwise_conv2_weight_to_fp16_palettized"), shape = tensor([1024, 1024, 1])]; - tensor x_65_cast_fp16 = conv(dilations = x_65_dilations_0, groups = x_65_groups_0, pad = x_65_pad_0, pad_type = x_65_pad_type_0, strides = x_65_strides_0, weight = module_layers_2_conv_pointwise_conv2_weight_to_fp16_palettized, x = input_157_cast_fp16)[name = tensor("x_65_cast_fp16")]; - tensor input_159_perm_0 = const()[name = tensor("input_159_perm_0"), val = tensor([0, 2, 1])]; - tensor input_159_cast_fp16 = transpose(perm = input_159_perm_0, x = x_65_cast_fp16)[name = tensor("transpose_292")]; - tensor input_161_cast_fp16 = add(x = input_143_cast_fp16, y = input_159_cast_fp16)[name = tensor("input_161_cast_fp16")]; - tensor input_163_axes_0 = const()[name = tensor("input_163_axes_0"), val = tensor([-1])]; - tensor module_layers_2_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_2_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52205888)))]; - tensor module_layers_2_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_2_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52208000)))]; - tensor input_163_cast_fp16 = layer_norm(axes = input_163_axes_0, beta = module_layers_2_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_2_norm_feed_forward2_weight_to_fp16, x = input_161_cast_fp16)[name = tensor("input_163_cast_fp16")]; - tensor module_layers_2_feed_forward2_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52210112))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(55355904))), name = tensor("module_layers_2_feed_forward2_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_26_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_2_feed_forward2_linear1_weight_to_fp16_palettized, x = input_163_cast_fp16)[name = tensor("linear_26_cast_fp16")]; - tensor input_167_cast_fp16 = silu(x = linear_26_cast_fp16)[name = tensor("input_167_cast_fp16")]; - tensor module_layers_2_feed_forward2_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(55356096))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58501888))), name = tensor("module_layers_2_feed_forward2_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_27_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_2_feed_forward2_linear2_weight_to_fp16_palettized, x = input_167_cast_fp16)[name = tensor("linear_27_cast_fp16")]; - tensor var_692_to_fp16 = const()[name = tensor("op_692_to_fp16"), val = tensor(0x1p-1)]; - tensor var_693_cast_fp16 = mul(x = linear_27_cast_fp16, y = var_692_to_fp16)[name = tensor("op_693_cast_fp16")]; - tensor input_173_cast_fp16 = add(x = input_161_cast_fp16, y = var_693_cast_fp16)[name = tensor("input_173_cast_fp16")]; - tensor input_175_axes_0 = const()[name = tensor("input_175_axes_0"), val = tensor([-1])]; - tensor module_layers_2_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_2_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58502080)))]; - tensor module_layers_2_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_2_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58504192)))]; - tensor input_175_cast_fp16 = layer_norm(axes = input_175_axes_0, beta = module_layers_2_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_2_norm_out_weight_to_fp16, x = input_173_cast_fp16)[name = tensor("input_175_cast_fp16")]; - tensor input_177_axes_0 = const()[name = tensor("input_177_axes_0"), val = tensor([-1])]; - tensor module_layers_3_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_3_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58506304)))]; - tensor module_layers_3_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_3_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58508416)))]; - tensor input_177_cast_fp16 = layer_norm(axes = input_177_axes_0, beta = module_layers_3_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_3_norm_feed_forward1_weight_to_fp16, x = input_175_cast_fp16)[name = tensor("input_177_cast_fp16")]; - tensor module_layers_3_feed_forward1_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58510528))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61656320))), name = tensor("module_layers_3_feed_forward1_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_28_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_3_feed_forward1_linear1_weight_to_fp16_palettized, x = input_177_cast_fp16)[name = tensor("linear_28_cast_fp16")]; - tensor input_181_cast_fp16 = silu(x = linear_28_cast_fp16)[name = tensor("input_181_cast_fp16")]; - tensor module_layers_3_feed_forward1_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61656512))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64802304))), name = tensor("module_layers_3_feed_forward1_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_29_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_3_feed_forward1_linear2_weight_to_fp16_palettized, x = input_181_cast_fp16)[name = tensor("linear_29_cast_fp16")]; - tensor var_721_to_fp16 = const()[name = tensor("op_721_to_fp16"), val = tensor(0x1p-1)]; - tensor var_722_cast_fp16 = mul(x = linear_29_cast_fp16, y = var_721_to_fp16)[name = tensor("op_722_cast_fp16")]; - tensor input_187_cast_fp16 = add(x = input_175_cast_fp16, y = var_722_cast_fp16)[name = tensor("input_187_cast_fp16")]; - tensor query_7_axes_0 = const()[name = tensor("query_7_axes_0"), val = tensor([-1])]; - tensor module_layers_3_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_3_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64802496)))]; - tensor module_layers_3_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_3_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64804608)))]; - tensor query_7_cast_fp16 = layer_norm(axes = query_7_axes_0, beta = module_layers_3_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_3_norm_self_att_weight_to_fp16, x = input_187_cast_fp16)[name = tensor("query_7_cast_fp16")]; - tensor module_layers_3_self_attn_linear_q_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64806720))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65593216))), name = tensor("module_layers_3_self_attn_linear_q_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_30_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_3_self_attn_linear_q_weight_to_fp16_palettized, x = query_7_cast_fp16)[name = tensor("linear_30_cast_fp16")]; - tensor var_738 = const()[name = tensor("op_738"), val = tensor([1, -1, 8, 128])]; - tensor q_19_cast_fp16 = reshape(shape = var_738, x = linear_30_cast_fp16)[name = tensor("q_19_cast_fp16")]; - tensor module_layers_3_self_attn_linear_k_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65593408))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(66379904))), name = tensor("module_layers_3_self_attn_linear_k_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_31_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_3_self_attn_linear_k_weight_to_fp16_palettized, x = query_7_cast_fp16)[name = tensor("linear_31_cast_fp16")]; - tensor var_742 = const()[name = tensor("op_742"), val = tensor([1, -1, 8, 128])]; - tensor k_13_cast_fp16 = reshape(shape = var_742, x = linear_31_cast_fp16)[name = tensor("k_13_cast_fp16")]; - tensor module_layers_3_self_attn_linear_v_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(66380096))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67166592))), name = tensor("module_layers_3_self_attn_linear_v_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_32_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_3_self_attn_linear_v_weight_to_fp16_palettized, x = query_7_cast_fp16)[name = tensor("linear_32_cast_fp16")]; - tensor var_746 = const()[name = tensor("op_746"), val = tensor([1, -1, 8, 128])]; - tensor v_7_cast_fp16 = reshape(shape = var_746, x = linear_32_cast_fp16)[name = tensor("v_7_cast_fp16")]; - tensor value_9_perm_0 = const()[name = tensor("value_9_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor module_layers_3_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_3_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67166784)))]; - tensor var_758_cast_fp16 = add(x = q_19_cast_fp16, y = module_layers_3_self_attn_pos_bias_u_to_fp16)[name = tensor("op_758_cast_fp16")]; - tensor module_layers_3_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_3_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67168896)))]; - tensor var_760_cast_fp16 = add(x = q_19_cast_fp16, y = module_layers_3_self_attn_pos_bias_v_to_fp16)[name = tensor("op_760_cast_fp16")]; - tensor q_with_bias_v_7_perm_0 = const()[name = tensor("q_with_bias_v_7_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor x_73_transpose_x_0 = const()[name = tensor("x_73_transpose_x_0"), val = tensor(false)]; - tensor x_73_transpose_y_0 = const()[name = tensor("x_73_transpose_y_0"), val = tensor(false)]; - tensor op_762_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67171008))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67459072))), name = tensor("op_762_to_fp16_palettized"), shape = tensor([1, 8, 128, 375])]; - tensor q_with_bias_v_7_cast_fp16 = transpose(perm = q_with_bias_v_7_perm_0, x = var_760_cast_fp16)[name = tensor("transpose_291")]; - tensor x_73_cast_fp16 = matmul(transpose_x = x_73_transpose_x_0, transpose_y = x_73_transpose_y_0, x = q_with_bias_v_7_cast_fp16, y = op_762_to_fp16_palettized)[name = tensor("x_73_cast_fp16")]; - tensor x_75_pad_0 = const()[name = tensor("x_75_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_75_mode_0 = const()[name = tensor("x_75_mode_0"), val = tensor("constant")]; - tensor const_44_to_fp16 = const()[name = tensor("const_44_to_fp16"), val = tensor(0x0p+0)]; - tensor x_75_cast_fp16 = pad(constant_val = const_44_to_fp16, mode = x_75_mode_0, pad = x_75_pad_0, x = x_73_cast_fp16)[name = tensor("x_75_cast_fp16")]; - tensor var_770 = const()[name = tensor("op_770"), val = tensor([1, 8, -1, 188])]; - tensor x_77_cast_fp16 = reshape(shape = var_770, x = x_75_cast_fp16)[name = tensor("x_77_cast_fp16")]; - tensor var_774_begin_0 = const()[name = tensor("op_774_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_774_end_0 = const()[name = tensor("op_774_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_774_end_mask_0 = const()[name = tensor("op_774_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_774_cast_fp16 = slice_by_index(begin = var_774_begin_0, end = var_774_end_0, end_mask = var_774_end_mask_0, x = x_77_cast_fp16)[name = tensor("op_774_cast_fp16")]; - tensor var_775 = const()[name = tensor("op_775"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_13_cast_fp16 = reshape(shape = var_775, x = var_774_cast_fp16)[name = tensor("matrix_bd_13_cast_fp16")]; - tensor matrix_ac_7_transpose_x_0 = const()[name = tensor("matrix_ac_7_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_7_transpose_y_0 = const()[name = tensor("matrix_ac_7_transpose_y_0"), val = tensor(false)]; - tensor transpose_102_perm_0 = const()[name = tensor("transpose_102_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_103_perm_0 = const()[name = tensor("transpose_103_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_103 = transpose(perm = transpose_103_perm_0, x = k_13_cast_fp16)[name = tensor("transpose_289")]; - tensor transpose_102 = transpose(perm = transpose_102_perm_0, x = var_758_cast_fp16)[name = tensor("transpose_290")]; - tensor matrix_ac_7_cast_fp16 = matmul(transpose_x = matrix_ac_7_transpose_x_0, transpose_y = matrix_ac_7_transpose_y_0, x = transpose_102, y = transpose_103)[name = tensor("matrix_ac_7_cast_fp16")]; - tensor matrix_bd_15_begin_0 = const()[name = tensor("matrix_bd_15_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_15_end_0 = const()[name = tensor("matrix_bd_15_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_15_end_mask_0 = const()[name = tensor("matrix_bd_15_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_15_cast_fp16 = slice_by_index(begin = matrix_bd_15_begin_0, end = matrix_bd_15_end_0, end_mask = matrix_bd_15_end_mask_0, x = matrix_bd_13_cast_fp16)[name = tensor("matrix_bd_15_cast_fp16")]; - tensor var_784_cast_fp16 = add(x = matrix_ac_7_cast_fp16, y = matrix_bd_15_cast_fp16)[name = tensor("op_784_cast_fp16")]; - tensor _inversed_scores_13_y_0_to_fp16 = const()[name = tensor("_inversed_scores_13_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; - tensor _inversed_scores_13_cast_fp16 = mul(x = var_784_cast_fp16, y = _inversed_scores_13_y_0_to_fp16)[name = tensor("_inversed_scores_13_cast_fp16")]; - tensor scores_15_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_13_cast_fp16, cond = mask_3)[name = tensor("scores_15_cast_fp16")]; - tensor var_790_cast_fp16 = softmax(axis = var_30, x = scores_15_cast_fp16)[name = tensor("op_790_cast_fp16")]; - tensor input_189_cast_fp16 = select(a = var_11_to_fp16, b = var_790_cast_fp16, cond = mask_3)[name = tensor("input_189_cast_fp16")]; - tensor x_79_transpose_x_0 = const()[name = tensor("x_79_transpose_x_0"), val = tensor(false)]; - tensor x_79_transpose_y_0 = const()[name = tensor("x_79_transpose_y_0"), val = tensor(false)]; - tensor value_9_cast_fp16 = transpose(perm = value_9_perm_0, x = v_7_cast_fp16)[name = tensor("transpose_288")]; - tensor x_79_cast_fp16 = matmul(transpose_x = x_79_transpose_x_0, transpose_y = x_79_transpose_y_0, x = input_189_cast_fp16, y = value_9_cast_fp16)[name = tensor("x_79_cast_fp16")]; - tensor var_794_perm_0 = const()[name = tensor("op_794_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_795 = const()[name = tensor("op_795"), val = tensor([1, -1, 1024])]; - tensor var_794_cast_fp16 = transpose(perm = var_794_perm_0, x = x_79_cast_fp16)[name = tensor("transpose_287")]; - tensor input_191_cast_fp16 = reshape(shape = var_795, x = var_794_cast_fp16)[name = tensor("input_191_cast_fp16")]; - tensor module_layers_3_self_attn_linear_out_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67459264))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(68245760))), name = tensor("module_layers_3_self_attn_linear_out_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_34_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_3_self_attn_linear_out_weight_to_fp16_palettized, x = input_191_cast_fp16)[name = tensor("linear_34_cast_fp16")]; - tensor input_195_cast_fp16 = add(x = input_187_cast_fp16, y = linear_34_cast_fp16)[name = tensor("input_195_cast_fp16")]; - tensor x_83_axes_0 = const()[name = tensor("x_83_axes_0"), val = tensor([-1])]; - tensor module_layers_3_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_3_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(68245952)))]; - tensor module_layers_3_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_3_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(68248064)))]; - tensor x_83_cast_fp16 = layer_norm(axes = x_83_axes_0, beta = module_layers_3_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_3_norm_conv_weight_to_fp16, x = input_195_cast_fp16)[name = tensor("x_83_cast_fp16")]; - tensor input_197_perm_0 = const()[name = tensor("input_197_perm_0"), val = tensor([0, 2, 1])]; - tensor input_199_pad_type_0 = const()[name = tensor("input_199_pad_type_0"), val = tensor("valid")]; - tensor input_199_strides_0 = const()[name = tensor("input_199_strides_0"), val = tensor([1])]; - tensor input_199_pad_0 = const()[name = tensor("input_199_pad_0"), val = tensor([0, 0])]; - tensor input_199_dilations_0 = const()[name = tensor("input_199_dilations_0"), val = tensor([1])]; - tensor input_199_groups_0 = const()[name = tensor("input_199_groups_0"), val = tensor(1)]; - tensor module_layers_3_conv_pointwise_conv1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(68250176))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(69823104))), name = tensor("module_layers_3_conv_pointwise_conv1_weight_to_fp16_palettized"), shape = tensor([2048, 1024, 1])]; - tensor input_197_cast_fp16 = transpose(perm = input_197_perm_0, x = x_83_cast_fp16)[name = tensor("transpose_286")]; - tensor input_199_cast_fp16 = conv(dilations = input_199_dilations_0, groups = input_199_groups_0, pad = input_199_pad_0, pad_type = input_199_pad_type_0, strides = input_199_strides_0, weight = module_layers_3_conv_pointwise_conv1_weight_to_fp16_palettized, x = input_197_cast_fp16)[name = tensor("input_199_cast_fp16")]; - tensor x_85_split_num_splits_0 = const()[name = tensor("x_85_split_num_splits_0"), val = tensor(2)]; - tensor x_85_split_axis_0 = const()[name = tensor("x_85_split_axis_0"), val = tensor(1)]; - tensor x_85_split_cast_fp16_0, tensor x_85_split_cast_fp16_1 = split(axis = x_85_split_axis_0, num_splits = x_85_split_num_splits_0, x = input_199_cast_fp16)[name = tensor("x_85_split_cast_fp16")]; - tensor x_85_split_1_sigmoid_cast_fp16 = sigmoid(x = x_85_split_cast_fp16_1)[name = tensor("x_85_split_1_sigmoid_cast_fp16")]; - tensor x_85_cast_fp16 = mul(x = x_85_split_cast_fp16_0, y = x_85_split_1_sigmoid_cast_fp16)[name = tensor("x_85_cast_fp16")]; - tensor input_201_cast_fp16 = select(a = var_11_to_fp16, b = x_85_cast_fp16, cond = var_328)[name = tensor("input_201_cast_fp16")]; - tensor input_203_pad_0 = const()[name = tensor("input_203_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_203_mode_0 = const()[name = tensor("input_203_mode_0"), val = tensor("constant")]; - tensor const_47_to_fp16 = const()[name = tensor("const_47_to_fp16"), val = tensor(0x0p+0)]; - tensor input_203_cast_fp16 = pad(constant_val = const_47_to_fp16, mode = input_203_mode_0, pad = input_203_pad_0, x = input_201_cast_fp16)[name = tensor("input_203_cast_fp16")]; - tensor input_205_pad_type_0 = const()[name = tensor("input_205_pad_type_0"), val = tensor("valid")]; - tensor input_205_groups_0 = const()[name = tensor("input_205_groups_0"), val = tensor(1024)]; - tensor input_205_strides_0 = const()[name = tensor("input_205_strides_0"), val = tensor([1])]; - tensor input_205_pad_0 = const()[name = tensor("input_205_pad_0"), val = tensor([0, 0])]; - tensor input_205_dilations_0 = const()[name = tensor("input_205_dilations_0"), val = tensor([1])]; - tensor const_254_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(69823296))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(69830272))), name = tensor("const_254_to_fp16_palettized"), shape = tensor([1024, 1, 9])]; - tensor const_255_to_fp16 = const()[name = tensor("const_255_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(69830464)))]; - tensor input_207_cast_fp16 = conv(bias = const_255_to_fp16, dilations = input_205_dilations_0, groups = input_205_groups_0, pad = input_205_pad_0, pad_type = input_205_pad_type_0, strides = input_205_strides_0, weight = const_254_to_fp16_palettized, x = input_203_cast_fp16)[name = tensor("input_207_cast_fp16")]; - tensor input_209_cast_fp16 = silu(x = input_207_cast_fp16)[name = tensor("input_209_cast_fp16")]; - tensor x_87_pad_type_0 = const()[name = tensor("x_87_pad_type_0"), val = tensor("valid")]; - tensor x_87_strides_0 = const()[name = tensor("x_87_strides_0"), val = tensor([1])]; - tensor x_87_pad_0 = const()[name = tensor("x_87_pad_0"), val = tensor([0, 0])]; - tensor x_87_dilations_0 = const()[name = tensor("x_87_dilations_0"), val = tensor([1])]; - tensor x_87_groups_0 = const()[name = tensor("x_87_groups_0"), val = tensor(1)]; - tensor module_layers_3_conv_pointwise_conv2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(69832576))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70619072))), name = tensor("module_layers_3_conv_pointwise_conv2_weight_to_fp16_palettized"), shape = tensor([1024, 1024, 1])]; - tensor x_87_cast_fp16 = conv(dilations = x_87_dilations_0, groups = x_87_groups_0, pad = x_87_pad_0, pad_type = x_87_pad_type_0, strides = x_87_strides_0, weight = module_layers_3_conv_pointwise_conv2_weight_to_fp16_palettized, x = input_209_cast_fp16)[name = tensor("x_87_cast_fp16")]; - tensor input_211_perm_0 = const()[name = tensor("input_211_perm_0"), val = tensor([0, 2, 1])]; - tensor input_211_cast_fp16 = transpose(perm = input_211_perm_0, x = x_87_cast_fp16)[name = tensor("transpose_285")]; - tensor input_213_cast_fp16 = add(x = input_195_cast_fp16, y = input_211_cast_fp16)[name = tensor("input_213_cast_fp16")]; - tensor input_215_axes_0 = const()[name = tensor("input_215_axes_0"), val = tensor([-1])]; - tensor module_layers_3_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_3_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70619264)))]; - tensor module_layers_3_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_3_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70621376)))]; - tensor input_215_cast_fp16 = layer_norm(axes = input_215_axes_0, beta = module_layers_3_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_3_norm_feed_forward2_weight_to_fp16, x = input_213_cast_fp16)[name = tensor("input_215_cast_fp16")]; - tensor module_layers_3_feed_forward2_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70623488))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73769280))), name = tensor("module_layers_3_feed_forward2_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_35_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_3_feed_forward2_linear1_weight_to_fp16_palettized, x = input_215_cast_fp16)[name = tensor("linear_35_cast_fp16")]; - tensor input_219_cast_fp16 = silu(x = linear_35_cast_fp16)[name = tensor("input_219_cast_fp16")]; - tensor module_layers_3_feed_forward2_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73769472))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(76915264))), name = tensor("module_layers_3_feed_forward2_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_36_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_3_feed_forward2_linear2_weight_to_fp16_palettized, x = input_219_cast_fp16)[name = tensor("linear_36_cast_fp16")]; - tensor var_855_to_fp16 = const()[name = tensor("op_855_to_fp16"), val = tensor(0x1p-1)]; - tensor var_856_cast_fp16 = mul(x = linear_36_cast_fp16, y = var_855_to_fp16)[name = tensor("op_856_cast_fp16")]; - tensor input_225_cast_fp16 = add(x = input_213_cast_fp16, y = var_856_cast_fp16)[name = tensor("input_225_cast_fp16")]; - tensor input_227_axes_0 = const()[name = tensor("input_227_axes_0"), val = tensor([-1])]; - tensor module_layers_3_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_3_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(76915456)))]; - tensor module_layers_3_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_3_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(76917568)))]; - tensor input_227_cast_fp16 = layer_norm(axes = input_227_axes_0, beta = module_layers_3_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_3_norm_out_weight_to_fp16, x = input_225_cast_fp16)[name = tensor("input_227_cast_fp16")]; - tensor input_229_axes_0 = const()[name = tensor("input_229_axes_0"), val = tensor([-1])]; - tensor module_layers_4_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_4_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(76919680)))]; - tensor module_layers_4_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_4_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(76921792)))]; - tensor input_229_cast_fp16 = layer_norm(axes = input_229_axes_0, beta = module_layers_4_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_4_norm_feed_forward1_weight_to_fp16, x = input_227_cast_fp16)[name = tensor("input_229_cast_fp16")]; - tensor module_layers_4_feed_forward1_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(76923904))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(80069696))), name = tensor("module_layers_4_feed_forward1_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_37_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_4_feed_forward1_linear1_weight_to_fp16_palettized, x = input_229_cast_fp16)[name = tensor("linear_37_cast_fp16")]; - tensor input_233_cast_fp16 = silu(x = linear_37_cast_fp16)[name = tensor("input_233_cast_fp16")]; - tensor module_layers_4_feed_forward1_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(80069888))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83215680))), name = tensor("module_layers_4_feed_forward1_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_38_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_4_feed_forward1_linear2_weight_to_fp16_palettized, x = input_233_cast_fp16)[name = tensor("linear_38_cast_fp16")]; - tensor var_884_to_fp16 = const()[name = tensor("op_884_to_fp16"), val = tensor(0x1p-1)]; - tensor var_885_cast_fp16 = mul(x = linear_38_cast_fp16, y = var_884_to_fp16)[name = tensor("op_885_cast_fp16")]; - tensor input_239_cast_fp16 = add(x = input_227_cast_fp16, y = var_885_cast_fp16)[name = tensor("input_239_cast_fp16")]; - tensor query_9_axes_0 = const()[name = tensor("query_9_axes_0"), val = tensor([-1])]; - tensor module_layers_4_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_4_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83215872)))]; - tensor module_layers_4_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_4_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83217984)))]; - tensor query_9_cast_fp16 = layer_norm(axes = query_9_axes_0, beta = module_layers_4_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_4_norm_self_att_weight_to_fp16, x = input_239_cast_fp16)[name = tensor("query_9_cast_fp16")]; - tensor module_layers_4_self_attn_linear_q_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83220096))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84006592))), name = tensor("module_layers_4_self_attn_linear_q_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_39_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_4_self_attn_linear_q_weight_to_fp16_palettized, x = query_9_cast_fp16)[name = tensor("linear_39_cast_fp16")]; - tensor var_901 = const()[name = tensor("op_901"), val = tensor([1, -1, 8, 128])]; - tensor q_25_cast_fp16 = reshape(shape = var_901, x = linear_39_cast_fp16)[name = tensor("q_25_cast_fp16")]; - tensor module_layers_4_self_attn_linear_k_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84006784))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84793280))), name = tensor("module_layers_4_self_attn_linear_k_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_40_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_4_self_attn_linear_k_weight_to_fp16_palettized, x = query_9_cast_fp16)[name = tensor("linear_40_cast_fp16")]; - tensor var_905 = const()[name = tensor("op_905"), val = tensor([1, -1, 8, 128])]; - tensor k_17_cast_fp16 = reshape(shape = var_905, x = linear_40_cast_fp16)[name = tensor("k_17_cast_fp16")]; - tensor module_layers_4_self_attn_linear_v_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84793472))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85579968))), name = tensor("module_layers_4_self_attn_linear_v_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_41_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_4_self_attn_linear_v_weight_to_fp16_palettized, x = query_9_cast_fp16)[name = tensor("linear_41_cast_fp16")]; - tensor var_909 = const()[name = tensor("op_909"), val = tensor([1, -1, 8, 128])]; - tensor v_9_cast_fp16 = reshape(shape = var_909, x = linear_41_cast_fp16)[name = tensor("v_9_cast_fp16")]; - tensor value_11_perm_0 = const()[name = tensor("value_11_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor module_layers_4_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_4_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85580160)))]; - tensor var_921_cast_fp16 = add(x = q_25_cast_fp16, y = module_layers_4_self_attn_pos_bias_u_to_fp16)[name = tensor("op_921_cast_fp16")]; - tensor module_layers_4_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_4_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85582272)))]; - tensor var_923_cast_fp16 = add(x = q_25_cast_fp16, y = module_layers_4_self_attn_pos_bias_v_to_fp16)[name = tensor("op_923_cast_fp16")]; - tensor q_with_bias_v_9_perm_0 = const()[name = tensor("q_with_bias_v_9_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor x_95_transpose_x_0 = const()[name = tensor("x_95_transpose_x_0"), val = tensor(false)]; - tensor x_95_transpose_y_0 = const()[name = tensor("x_95_transpose_y_0"), val = tensor(false)]; - tensor op_925_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85584384))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85872448))), name = tensor("op_925_to_fp16_palettized"), shape = tensor([1, 8, 128, 375])]; - tensor q_with_bias_v_9_cast_fp16 = transpose(perm = q_with_bias_v_9_perm_0, x = var_923_cast_fp16)[name = tensor("transpose_284")]; - tensor x_95_cast_fp16 = matmul(transpose_x = x_95_transpose_x_0, transpose_y = x_95_transpose_y_0, x = q_with_bias_v_9_cast_fp16, y = op_925_to_fp16_palettized)[name = tensor("x_95_cast_fp16")]; - tensor x_97_pad_0 = const()[name = tensor("x_97_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_97_mode_0 = const()[name = tensor("x_97_mode_0"), val = tensor("constant")]; - tensor const_54_to_fp16 = const()[name = tensor("const_54_to_fp16"), val = tensor(0x0p+0)]; - tensor x_97_cast_fp16 = pad(constant_val = const_54_to_fp16, mode = x_97_mode_0, pad = x_97_pad_0, x = x_95_cast_fp16)[name = tensor("x_97_cast_fp16")]; - tensor var_933 = const()[name = tensor("op_933"), val = tensor([1, 8, -1, 188])]; - tensor x_99_cast_fp16 = reshape(shape = var_933, x = x_97_cast_fp16)[name = tensor("x_99_cast_fp16")]; - tensor var_937_begin_0 = const()[name = tensor("op_937_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_937_end_0 = const()[name = tensor("op_937_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_937_end_mask_0 = const()[name = tensor("op_937_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_937_cast_fp16 = slice_by_index(begin = var_937_begin_0, end = var_937_end_0, end_mask = var_937_end_mask_0, x = x_99_cast_fp16)[name = tensor("op_937_cast_fp16")]; - tensor var_938 = const()[name = tensor("op_938"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_17_cast_fp16 = reshape(shape = var_938, x = var_937_cast_fp16)[name = tensor("matrix_bd_17_cast_fp16")]; - tensor matrix_ac_9_transpose_x_0 = const()[name = tensor("matrix_ac_9_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_9_transpose_y_0 = const()[name = tensor("matrix_ac_9_transpose_y_0"), val = tensor(false)]; - tensor transpose_104_perm_0 = const()[name = tensor("transpose_104_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_105_perm_0 = const()[name = tensor("transpose_105_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_105 = transpose(perm = transpose_105_perm_0, x = k_17_cast_fp16)[name = tensor("transpose_282")]; - tensor transpose_104 = transpose(perm = transpose_104_perm_0, x = var_921_cast_fp16)[name = tensor("transpose_283")]; - tensor matrix_ac_9_cast_fp16 = matmul(transpose_x = matrix_ac_9_transpose_x_0, transpose_y = matrix_ac_9_transpose_y_0, x = transpose_104, y = transpose_105)[name = tensor("matrix_ac_9_cast_fp16")]; - tensor matrix_bd_19_begin_0 = const()[name = tensor("matrix_bd_19_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_19_end_0 = const()[name = tensor("matrix_bd_19_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_19_end_mask_0 = const()[name = tensor("matrix_bd_19_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_19_cast_fp16 = slice_by_index(begin = matrix_bd_19_begin_0, end = matrix_bd_19_end_0, end_mask = matrix_bd_19_end_mask_0, x = matrix_bd_17_cast_fp16)[name = tensor("matrix_bd_19_cast_fp16")]; - tensor var_947_cast_fp16 = add(x = matrix_ac_9_cast_fp16, y = matrix_bd_19_cast_fp16)[name = tensor("op_947_cast_fp16")]; - tensor _inversed_scores_17_y_0_to_fp16 = const()[name = tensor("_inversed_scores_17_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; - tensor _inversed_scores_17_cast_fp16 = mul(x = var_947_cast_fp16, y = _inversed_scores_17_y_0_to_fp16)[name = tensor("_inversed_scores_17_cast_fp16")]; - tensor scores_19_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_17_cast_fp16, cond = mask_3)[name = tensor("scores_19_cast_fp16")]; - tensor var_953_cast_fp16 = softmax(axis = var_30, x = scores_19_cast_fp16)[name = tensor("op_953_cast_fp16")]; - tensor input_241_cast_fp16 = select(a = var_11_to_fp16, b = var_953_cast_fp16, cond = mask_3)[name = tensor("input_241_cast_fp16")]; - tensor x_101_transpose_x_0 = const()[name = tensor("x_101_transpose_x_0"), val = tensor(false)]; - tensor x_101_transpose_y_0 = const()[name = tensor("x_101_transpose_y_0"), val = tensor(false)]; - tensor value_11_cast_fp16 = transpose(perm = value_11_perm_0, x = v_9_cast_fp16)[name = tensor("transpose_281")]; - tensor x_101_cast_fp16 = matmul(transpose_x = x_101_transpose_x_0, transpose_y = x_101_transpose_y_0, x = input_241_cast_fp16, y = value_11_cast_fp16)[name = tensor("x_101_cast_fp16")]; - tensor var_957_perm_0 = const()[name = tensor("op_957_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_958 = const()[name = tensor("op_958"), val = tensor([1, -1, 1024])]; - tensor var_957_cast_fp16 = transpose(perm = var_957_perm_0, x = x_101_cast_fp16)[name = tensor("transpose_280")]; - tensor input_243_cast_fp16 = reshape(shape = var_958, x = var_957_cast_fp16)[name = tensor("input_243_cast_fp16")]; - tensor module_layers_4_self_attn_linear_out_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85872640))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86659136))), name = tensor("module_layers_4_self_attn_linear_out_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_43_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_4_self_attn_linear_out_weight_to_fp16_palettized, x = input_243_cast_fp16)[name = tensor("linear_43_cast_fp16")]; - tensor input_247_cast_fp16 = add(x = input_239_cast_fp16, y = linear_43_cast_fp16)[name = tensor("input_247_cast_fp16")]; - tensor x_105_axes_0 = const()[name = tensor("x_105_axes_0"), val = tensor([-1])]; - tensor module_layers_4_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_4_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86659328)))]; - tensor module_layers_4_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_4_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86661440)))]; - tensor x_105_cast_fp16 = layer_norm(axes = x_105_axes_0, beta = module_layers_4_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_4_norm_conv_weight_to_fp16, x = input_247_cast_fp16)[name = tensor("x_105_cast_fp16")]; - tensor input_249_perm_0 = const()[name = tensor("input_249_perm_0"), val = tensor([0, 2, 1])]; - tensor input_251_pad_type_0 = const()[name = tensor("input_251_pad_type_0"), val = tensor("valid")]; - tensor input_251_strides_0 = const()[name = tensor("input_251_strides_0"), val = tensor([1])]; - tensor input_251_pad_0 = const()[name = tensor("input_251_pad_0"), val = tensor([0, 0])]; - tensor input_251_dilations_0 = const()[name = tensor("input_251_dilations_0"), val = tensor([1])]; - tensor input_251_groups_0 = const()[name = tensor("input_251_groups_0"), val = tensor(1)]; - tensor module_layers_4_conv_pointwise_conv1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86663552))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(88236480))), name = tensor("module_layers_4_conv_pointwise_conv1_weight_to_fp16_palettized"), shape = tensor([2048, 1024, 1])]; - tensor input_249_cast_fp16 = transpose(perm = input_249_perm_0, x = x_105_cast_fp16)[name = tensor("transpose_279")]; - tensor input_251_cast_fp16 = conv(dilations = input_251_dilations_0, groups = input_251_groups_0, pad = input_251_pad_0, pad_type = input_251_pad_type_0, strides = input_251_strides_0, weight = module_layers_4_conv_pointwise_conv1_weight_to_fp16_palettized, x = input_249_cast_fp16)[name = tensor("input_251_cast_fp16")]; - tensor x_107_split_num_splits_0 = const()[name = tensor("x_107_split_num_splits_0"), val = tensor(2)]; - tensor x_107_split_axis_0 = const()[name = tensor("x_107_split_axis_0"), val = tensor(1)]; - tensor x_107_split_cast_fp16_0, tensor x_107_split_cast_fp16_1 = split(axis = x_107_split_axis_0, num_splits = x_107_split_num_splits_0, x = input_251_cast_fp16)[name = tensor("x_107_split_cast_fp16")]; - tensor x_107_split_1_sigmoid_cast_fp16 = sigmoid(x = x_107_split_cast_fp16_1)[name = tensor("x_107_split_1_sigmoid_cast_fp16")]; - tensor x_107_cast_fp16 = mul(x = x_107_split_cast_fp16_0, y = x_107_split_1_sigmoid_cast_fp16)[name = tensor("x_107_cast_fp16")]; - tensor input_253_cast_fp16 = select(a = var_11_to_fp16, b = x_107_cast_fp16, cond = var_328)[name = tensor("input_253_cast_fp16")]; - tensor input_255_pad_0 = const()[name = tensor("input_255_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_255_mode_0 = const()[name = tensor("input_255_mode_0"), val = tensor("constant")]; - tensor const_57_to_fp16 = const()[name = tensor("const_57_to_fp16"), val = tensor(0x0p+0)]; - tensor input_255_cast_fp16 = pad(constant_val = const_57_to_fp16, mode = input_255_mode_0, pad = input_255_pad_0, x = input_253_cast_fp16)[name = tensor("input_255_cast_fp16")]; - tensor input_257_pad_type_0 = const()[name = tensor("input_257_pad_type_0"), val = tensor("valid")]; - tensor input_257_groups_0 = const()[name = tensor("input_257_groups_0"), val = tensor(1024)]; - tensor input_257_strides_0 = const()[name = tensor("input_257_strides_0"), val = tensor([1])]; - tensor input_257_pad_0 = const()[name = tensor("input_257_pad_0"), val = tensor([0, 0])]; - tensor input_257_dilations_0 = const()[name = tensor("input_257_dilations_0"), val = tensor([1])]; - tensor const_256_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(88236672))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(88243648))), name = tensor("const_256_to_fp16_palettized"), shape = tensor([1024, 1, 9])]; - tensor const_257_to_fp16 = const()[name = tensor("const_257_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(88243840)))]; - tensor input_259_cast_fp16 = conv(bias = const_257_to_fp16, dilations = input_257_dilations_0, groups = input_257_groups_0, pad = input_257_pad_0, pad_type = input_257_pad_type_0, strides = input_257_strides_0, weight = const_256_to_fp16_palettized, x = input_255_cast_fp16)[name = tensor("input_259_cast_fp16")]; - tensor input_261_cast_fp16 = silu(x = input_259_cast_fp16)[name = tensor("input_261_cast_fp16")]; - tensor x_109_pad_type_0 = const()[name = tensor("x_109_pad_type_0"), val = tensor("valid")]; - tensor x_109_strides_0 = const()[name = tensor("x_109_strides_0"), val = tensor([1])]; - tensor x_109_pad_0 = const()[name = tensor("x_109_pad_0"), val = tensor([0, 0])]; - tensor x_109_dilations_0 = const()[name = tensor("x_109_dilations_0"), val = tensor([1])]; - tensor x_109_groups_0 = const()[name = tensor("x_109_groups_0"), val = tensor(1)]; - tensor module_layers_4_conv_pointwise_conv2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(88245952))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(89032448))), name = tensor("module_layers_4_conv_pointwise_conv2_weight_to_fp16_palettized"), shape = tensor([1024, 1024, 1])]; - tensor x_109_cast_fp16 = conv(dilations = x_109_dilations_0, groups = x_109_groups_0, pad = x_109_pad_0, pad_type = x_109_pad_type_0, strides = x_109_strides_0, weight = module_layers_4_conv_pointwise_conv2_weight_to_fp16_palettized, x = input_261_cast_fp16)[name = tensor("x_109_cast_fp16")]; - tensor input_263_perm_0 = const()[name = tensor("input_263_perm_0"), val = tensor([0, 2, 1])]; - tensor input_263_cast_fp16 = transpose(perm = input_263_perm_0, x = x_109_cast_fp16)[name = tensor("transpose_278")]; - tensor input_265_cast_fp16 = add(x = input_247_cast_fp16, y = input_263_cast_fp16)[name = tensor("input_265_cast_fp16")]; - tensor input_267_axes_0 = const()[name = tensor("input_267_axes_0"), val = tensor([-1])]; - tensor module_layers_4_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_4_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(89032640)))]; - tensor module_layers_4_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_4_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(89034752)))]; - tensor input_267_cast_fp16 = layer_norm(axes = input_267_axes_0, beta = module_layers_4_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_4_norm_feed_forward2_weight_to_fp16, x = input_265_cast_fp16)[name = tensor("input_267_cast_fp16")]; - tensor module_layers_4_feed_forward2_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(89036864))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92182656))), name = tensor("module_layers_4_feed_forward2_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_44_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_4_feed_forward2_linear1_weight_to_fp16_palettized, x = input_267_cast_fp16)[name = tensor("linear_44_cast_fp16")]; - tensor input_271_cast_fp16 = silu(x = linear_44_cast_fp16)[name = tensor("input_271_cast_fp16")]; - tensor module_layers_4_feed_forward2_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92182848))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95328640))), name = tensor("module_layers_4_feed_forward2_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_45_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_4_feed_forward2_linear2_weight_to_fp16_palettized, x = input_271_cast_fp16)[name = tensor("linear_45_cast_fp16")]; - tensor var_1018_to_fp16 = const()[name = tensor("op_1018_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1019_cast_fp16 = mul(x = linear_45_cast_fp16, y = var_1018_to_fp16)[name = tensor("op_1019_cast_fp16")]; - tensor input_277_cast_fp16 = add(x = input_265_cast_fp16, y = var_1019_cast_fp16)[name = tensor("input_277_cast_fp16")]; - tensor input_279_axes_0 = const()[name = tensor("input_279_axes_0"), val = tensor([-1])]; - tensor module_layers_4_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_4_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95328832)))]; - tensor module_layers_4_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_4_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95330944)))]; - tensor input_279_cast_fp16 = layer_norm(axes = input_279_axes_0, beta = module_layers_4_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_4_norm_out_weight_to_fp16, x = input_277_cast_fp16)[name = tensor("input_279_cast_fp16")]; - tensor input_281_axes_0 = const()[name = tensor("input_281_axes_0"), val = tensor([-1])]; - tensor module_layers_5_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_5_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95333056)))]; - tensor module_layers_5_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_5_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95335168)))]; - tensor input_281_cast_fp16 = layer_norm(axes = input_281_axes_0, beta = module_layers_5_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_5_norm_feed_forward1_weight_to_fp16, x = input_279_cast_fp16)[name = tensor("input_281_cast_fp16")]; - tensor module_layers_5_feed_forward1_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95337280))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98483072))), name = tensor("module_layers_5_feed_forward1_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_46_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_5_feed_forward1_linear1_weight_to_fp16_palettized, x = input_281_cast_fp16)[name = tensor("linear_46_cast_fp16")]; - tensor input_285_cast_fp16 = silu(x = linear_46_cast_fp16)[name = tensor("input_285_cast_fp16")]; - tensor module_layers_5_feed_forward1_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98483264))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(101629056))), name = tensor("module_layers_5_feed_forward1_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_47_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_5_feed_forward1_linear2_weight_to_fp16_palettized, x = input_285_cast_fp16)[name = tensor("linear_47_cast_fp16")]; - tensor var_1047_to_fp16 = const()[name = tensor("op_1047_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1048_cast_fp16 = mul(x = linear_47_cast_fp16, y = var_1047_to_fp16)[name = tensor("op_1048_cast_fp16")]; - tensor input_291_cast_fp16 = add(x = input_279_cast_fp16, y = var_1048_cast_fp16)[name = tensor("input_291_cast_fp16")]; - tensor query_11_axes_0 = const()[name = tensor("query_11_axes_0"), val = tensor([-1])]; - tensor module_layers_5_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_5_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(101629248)))]; - tensor module_layers_5_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_5_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(101631360)))]; - tensor query_11_cast_fp16 = layer_norm(axes = query_11_axes_0, beta = module_layers_5_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_5_norm_self_att_weight_to_fp16, x = input_291_cast_fp16)[name = tensor("query_11_cast_fp16")]; - tensor module_layers_5_self_attn_linear_q_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(101633472))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102419968))), name = tensor("module_layers_5_self_attn_linear_q_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_48_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_5_self_attn_linear_q_weight_to_fp16_palettized, x = query_11_cast_fp16)[name = tensor("linear_48_cast_fp16")]; - tensor var_1064 = const()[name = tensor("op_1064"), val = tensor([1, -1, 8, 128])]; - tensor q_31_cast_fp16 = reshape(shape = var_1064, x = linear_48_cast_fp16)[name = tensor("q_31_cast_fp16")]; - tensor module_layers_5_self_attn_linear_k_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102420160))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(103206656))), name = tensor("module_layers_5_self_attn_linear_k_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_49_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_5_self_attn_linear_k_weight_to_fp16_palettized, x = query_11_cast_fp16)[name = tensor("linear_49_cast_fp16")]; - tensor var_1068 = const()[name = tensor("op_1068"), val = tensor([1, -1, 8, 128])]; - tensor k_21_cast_fp16 = reshape(shape = var_1068, x = linear_49_cast_fp16)[name = tensor("k_21_cast_fp16")]; - tensor module_layers_5_self_attn_linear_v_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(103206848))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(103993344))), name = tensor("module_layers_5_self_attn_linear_v_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_50_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_5_self_attn_linear_v_weight_to_fp16_palettized, x = query_11_cast_fp16)[name = tensor("linear_50_cast_fp16")]; - tensor var_1072 = const()[name = tensor("op_1072"), val = tensor([1, -1, 8, 128])]; - tensor v_11_cast_fp16 = reshape(shape = var_1072, x = linear_50_cast_fp16)[name = tensor("v_11_cast_fp16")]; - tensor value_13_perm_0 = const()[name = tensor("value_13_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor module_layers_5_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_5_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(103993536)))]; - tensor var_1084_cast_fp16 = add(x = q_31_cast_fp16, y = module_layers_5_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1084_cast_fp16")]; - tensor module_layers_5_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_5_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(103995648)))]; - tensor var_1086_cast_fp16 = add(x = q_31_cast_fp16, y = module_layers_5_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1086_cast_fp16")]; - tensor q_with_bias_v_11_perm_0 = const()[name = tensor("q_with_bias_v_11_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor x_117_transpose_x_0 = const()[name = tensor("x_117_transpose_x_0"), val = tensor(false)]; - tensor x_117_transpose_y_0 = const()[name = tensor("x_117_transpose_y_0"), val = tensor(false)]; - tensor op_1088_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(103997760))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(104285824))), name = tensor("op_1088_to_fp16_palettized"), shape = tensor([1, 8, 128, 375])]; - tensor q_with_bias_v_11_cast_fp16 = transpose(perm = q_with_bias_v_11_perm_0, x = var_1086_cast_fp16)[name = tensor("transpose_277")]; - tensor x_117_cast_fp16 = matmul(transpose_x = x_117_transpose_x_0, transpose_y = x_117_transpose_y_0, x = q_with_bias_v_11_cast_fp16, y = op_1088_to_fp16_palettized)[name = tensor("x_117_cast_fp16")]; - tensor x_119_pad_0 = const()[name = tensor("x_119_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_119_mode_0 = const()[name = tensor("x_119_mode_0"), val = tensor("constant")]; - tensor const_64_to_fp16 = const()[name = tensor("const_64_to_fp16"), val = tensor(0x0p+0)]; - tensor x_119_cast_fp16 = pad(constant_val = const_64_to_fp16, mode = x_119_mode_0, pad = x_119_pad_0, x = x_117_cast_fp16)[name = tensor("x_119_cast_fp16")]; - tensor var_1096 = const()[name = tensor("op_1096"), val = tensor([1, 8, -1, 188])]; - tensor x_121_cast_fp16 = reshape(shape = var_1096, x = x_119_cast_fp16)[name = tensor("x_121_cast_fp16")]; - tensor var_1100_begin_0 = const()[name = tensor("op_1100_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_1100_end_0 = const()[name = tensor("op_1100_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_1100_end_mask_0 = const()[name = tensor("op_1100_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_1100_cast_fp16 = slice_by_index(begin = var_1100_begin_0, end = var_1100_end_0, end_mask = var_1100_end_mask_0, x = x_121_cast_fp16)[name = tensor("op_1100_cast_fp16")]; - tensor var_1101 = const()[name = tensor("op_1101"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_21_cast_fp16 = reshape(shape = var_1101, x = var_1100_cast_fp16)[name = tensor("matrix_bd_21_cast_fp16")]; - tensor matrix_ac_11_transpose_x_0 = const()[name = tensor("matrix_ac_11_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_11_transpose_y_0 = const()[name = tensor("matrix_ac_11_transpose_y_0"), val = tensor(false)]; - tensor transpose_106_perm_0 = const()[name = tensor("transpose_106_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_107_perm_0 = const()[name = tensor("transpose_107_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_107 = transpose(perm = transpose_107_perm_0, x = k_21_cast_fp16)[name = tensor("transpose_275")]; - tensor transpose_106 = transpose(perm = transpose_106_perm_0, x = var_1084_cast_fp16)[name = tensor("transpose_276")]; - tensor matrix_ac_11_cast_fp16 = matmul(transpose_x = matrix_ac_11_transpose_x_0, transpose_y = matrix_ac_11_transpose_y_0, x = transpose_106, y = transpose_107)[name = tensor("matrix_ac_11_cast_fp16")]; - tensor matrix_bd_23_begin_0 = const()[name = tensor("matrix_bd_23_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_23_end_0 = const()[name = tensor("matrix_bd_23_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_23_end_mask_0 = const()[name = tensor("matrix_bd_23_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_23_cast_fp16 = slice_by_index(begin = matrix_bd_23_begin_0, end = matrix_bd_23_end_0, end_mask = matrix_bd_23_end_mask_0, x = matrix_bd_21_cast_fp16)[name = tensor("matrix_bd_23_cast_fp16")]; - tensor var_1110_cast_fp16 = add(x = matrix_ac_11_cast_fp16, y = matrix_bd_23_cast_fp16)[name = tensor("op_1110_cast_fp16")]; - tensor _inversed_scores_21_y_0_to_fp16 = const()[name = tensor("_inversed_scores_21_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; - tensor _inversed_scores_21_cast_fp16 = mul(x = var_1110_cast_fp16, y = _inversed_scores_21_y_0_to_fp16)[name = tensor("_inversed_scores_21_cast_fp16")]; - tensor scores_23_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_21_cast_fp16, cond = mask_3)[name = tensor("scores_23_cast_fp16")]; - tensor var_1116_cast_fp16 = softmax(axis = var_30, x = scores_23_cast_fp16)[name = tensor("op_1116_cast_fp16")]; - tensor input_293_cast_fp16 = select(a = var_11_to_fp16, b = var_1116_cast_fp16, cond = mask_3)[name = tensor("input_293_cast_fp16")]; - tensor x_123_transpose_x_0 = const()[name = tensor("x_123_transpose_x_0"), val = tensor(false)]; - tensor x_123_transpose_y_0 = const()[name = tensor("x_123_transpose_y_0"), val = tensor(false)]; - tensor value_13_cast_fp16 = transpose(perm = value_13_perm_0, x = v_11_cast_fp16)[name = tensor("transpose_274")]; - tensor x_123_cast_fp16 = matmul(transpose_x = x_123_transpose_x_0, transpose_y = x_123_transpose_y_0, x = input_293_cast_fp16, y = value_13_cast_fp16)[name = tensor("x_123_cast_fp16")]; - tensor var_1120_perm_0 = const()[name = tensor("op_1120_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_1121 = const()[name = tensor("op_1121"), val = tensor([1, -1, 1024])]; - tensor var_1120_cast_fp16 = transpose(perm = var_1120_perm_0, x = x_123_cast_fp16)[name = tensor("transpose_273")]; - tensor input_295_cast_fp16 = reshape(shape = var_1121, x = var_1120_cast_fp16)[name = tensor("input_295_cast_fp16")]; - tensor module_layers_5_self_attn_linear_out_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(104286016))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(105072512))), name = tensor("module_layers_5_self_attn_linear_out_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_52_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_5_self_attn_linear_out_weight_to_fp16_palettized, x = input_295_cast_fp16)[name = tensor("linear_52_cast_fp16")]; - tensor input_299_cast_fp16 = add(x = input_291_cast_fp16, y = linear_52_cast_fp16)[name = tensor("input_299_cast_fp16")]; - tensor x_127_axes_0 = const()[name = tensor("x_127_axes_0"), val = tensor([-1])]; - tensor module_layers_5_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_5_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(105072704)))]; - tensor module_layers_5_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_5_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(105074816)))]; - tensor x_127_cast_fp16 = layer_norm(axes = x_127_axes_0, beta = module_layers_5_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_5_norm_conv_weight_to_fp16, x = input_299_cast_fp16)[name = tensor("x_127_cast_fp16")]; - tensor input_301_perm_0 = const()[name = tensor("input_301_perm_0"), val = tensor([0, 2, 1])]; - tensor input_303_pad_type_0 = const()[name = tensor("input_303_pad_type_0"), val = tensor("valid")]; - tensor input_303_strides_0 = const()[name = tensor("input_303_strides_0"), val = tensor([1])]; - tensor input_303_pad_0 = const()[name = tensor("input_303_pad_0"), val = tensor([0, 0])]; - tensor input_303_dilations_0 = const()[name = tensor("input_303_dilations_0"), val = tensor([1])]; - tensor input_303_groups_0 = const()[name = tensor("input_303_groups_0"), val = tensor(1)]; - tensor module_layers_5_conv_pointwise_conv1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(105076928))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(106649856))), name = tensor("module_layers_5_conv_pointwise_conv1_weight_to_fp16_palettized"), shape = tensor([2048, 1024, 1])]; - tensor input_301_cast_fp16 = transpose(perm = input_301_perm_0, x = x_127_cast_fp16)[name = tensor("transpose_272")]; - tensor input_303_cast_fp16 = conv(dilations = input_303_dilations_0, groups = input_303_groups_0, pad = input_303_pad_0, pad_type = input_303_pad_type_0, strides = input_303_strides_0, weight = module_layers_5_conv_pointwise_conv1_weight_to_fp16_palettized, x = input_301_cast_fp16)[name = tensor("input_303_cast_fp16")]; - tensor x_129_split_num_splits_0 = const()[name = tensor("x_129_split_num_splits_0"), val = tensor(2)]; - tensor x_129_split_axis_0 = const()[name = tensor("x_129_split_axis_0"), val = tensor(1)]; - tensor x_129_split_cast_fp16_0, tensor x_129_split_cast_fp16_1 = split(axis = x_129_split_axis_0, num_splits = x_129_split_num_splits_0, x = input_303_cast_fp16)[name = tensor("x_129_split_cast_fp16")]; - tensor x_129_split_1_sigmoid_cast_fp16 = sigmoid(x = x_129_split_cast_fp16_1)[name = tensor("x_129_split_1_sigmoid_cast_fp16")]; - tensor x_129_cast_fp16 = mul(x = x_129_split_cast_fp16_0, y = x_129_split_1_sigmoid_cast_fp16)[name = tensor("x_129_cast_fp16")]; - tensor input_305_cast_fp16 = select(a = var_11_to_fp16, b = x_129_cast_fp16, cond = var_328)[name = tensor("input_305_cast_fp16")]; - tensor input_307_pad_0 = const()[name = tensor("input_307_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_307_mode_0 = const()[name = tensor("input_307_mode_0"), val = tensor("constant")]; - tensor const_67_to_fp16 = const()[name = tensor("const_67_to_fp16"), val = tensor(0x0p+0)]; - tensor input_307_cast_fp16 = pad(constant_val = const_67_to_fp16, mode = input_307_mode_0, pad = input_307_pad_0, x = input_305_cast_fp16)[name = tensor("input_307_cast_fp16")]; - tensor input_309_pad_type_0 = const()[name = tensor("input_309_pad_type_0"), val = tensor("valid")]; - tensor input_309_groups_0 = const()[name = tensor("input_309_groups_0"), val = tensor(1024)]; - tensor input_309_strides_0 = const()[name = tensor("input_309_strides_0"), val = tensor([1])]; - tensor input_309_pad_0 = const()[name = tensor("input_309_pad_0"), val = tensor([0, 0])]; - tensor input_309_dilations_0 = const()[name = tensor("input_309_dilations_0"), val = tensor([1])]; - tensor const_258_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(106650048))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(106657024))), name = tensor("const_258_to_fp16_palettized"), shape = tensor([1024, 1, 9])]; - tensor const_259_to_fp16 = const()[name = tensor("const_259_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(106657216)))]; - tensor input_311_cast_fp16 = conv(bias = const_259_to_fp16, dilations = input_309_dilations_0, groups = input_309_groups_0, pad = input_309_pad_0, pad_type = input_309_pad_type_0, strides = input_309_strides_0, weight = const_258_to_fp16_palettized, x = input_307_cast_fp16)[name = tensor("input_311_cast_fp16")]; - tensor input_313_cast_fp16 = silu(x = input_311_cast_fp16)[name = tensor("input_313_cast_fp16")]; - tensor x_131_pad_type_0 = const()[name = tensor("x_131_pad_type_0"), val = tensor("valid")]; - tensor x_131_strides_0 = const()[name = tensor("x_131_strides_0"), val = tensor([1])]; - tensor x_131_pad_0 = const()[name = tensor("x_131_pad_0"), val = tensor([0, 0])]; - tensor x_131_dilations_0 = const()[name = tensor("x_131_dilations_0"), val = tensor([1])]; - tensor x_131_groups_0 = const()[name = tensor("x_131_groups_0"), val = tensor(1)]; - tensor module_layers_5_conv_pointwise_conv2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(106659328))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(107445824))), name = tensor("module_layers_5_conv_pointwise_conv2_weight_to_fp16_palettized"), shape = tensor([1024, 1024, 1])]; - tensor x_131_cast_fp16 = conv(dilations = x_131_dilations_0, groups = x_131_groups_0, pad = x_131_pad_0, pad_type = x_131_pad_type_0, strides = x_131_strides_0, weight = module_layers_5_conv_pointwise_conv2_weight_to_fp16_palettized, x = input_313_cast_fp16)[name = tensor("x_131_cast_fp16")]; - tensor input_315_perm_0 = const()[name = tensor("input_315_perm_0"), val = tensor([0, 2, 1])]; - tensor input_315_cast_fp16 = transpose(perm = input_315_perm_0, x = x_131_cast_fp16)[name = tensor("transpose_271")]; - tensor input_317_cast_fp16 = add(x = input_299_cast_fp16, y = input_315_cast_fp16)[name = tensor("input_317_cast_fp16")]; - tensor input_319_axes_0 = const()[name = tensor("input_319_axes_0"), val = tensor([-1])]; - tensor module_layers_5_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_5_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(107446016)))]; - tensor module_layers_5_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_5_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(107448128)))]; - tensor input_319_cast_fp16 = layer_norm(axes = input_319_axes_0, beta = module_layers_5_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_5_norm_feed_forward2_weight_to_fp16, x = input_317_cast_fp16)[name = tensor("input_319_cast_fp16")]; - tensor module_layers_5_feed_forward2_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(107450240))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110596032))), name = tensor("module_layers_5_feed_forward2_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_53_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_5_feed_forward2_linear1_weight_to_fp16_palettized, x = input_319_cast_fp16)[name = tensor("linear_53_cast_fp16")]; - tensor input_323_cast_fp16 = silu(x = linear_53_cast_fp16)[name = tensor("input_323_cast_fp16")]; - tensor module_layers_5_feed_forward2_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110596224))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(113742016))), name = tensor("module_layers_5_feed_forward2_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_54_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_5_feed_forward2_linear2_weight_to_fp16_palettized, x = input_323_cast_fp16)[name = tensor("linear_54_cast_fp16")]; - tensor var_1181_to_fp16 = const()[name = tensor("op_1181_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1182_cast_fp16 = mul(x = linear_54_cast_fp16, y = var_1181_to_fp16)[name = tensor("op_1182_cast_fp16")]; - tensor input_329_cast_fp16 = add(x = input_317_cast_fp16, y = var_1182_cast_fp16)[name = tensor("input_329_cast_fp16")]; - tensor input_331_axes_0 = const()[name = tensor("input_331_axes_0"), val = tensor([-1])]; - tensor module_layers_5_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_5_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(113742208)))]; - tensor module_layers_5_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_5_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(113744320)))]; - tensor input_331_cast_fp16 = layer_norm(axes = input_331_axes_0, beta = module_layers_5_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_5_norm_out_weight_to_fp16, x = input_329_cast_fp16)[name = tensor("input_331_cast_fp16")]; - tensor input_333_axes_0 = const()[name = tensor("input_333_axes_0"), val = tensor([-1])]; - tensor module_layers_6_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_6_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(113746432)))]; - tensor module_layers_6_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_6_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(113748544)))]; - tensor input_333_cast_fp16 = layer_norm(axes = input_333_axes_0, beta = module_layers_6_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_6_norm_feed_forward1_weight_to_fp16, x = input_331_cast_fp16)[name = tensor("input_333_cast_fp16")]; - tensor module_layers_6_feed_forward1_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(113750656))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(116896448))), name = tensor("module_layers_6_feed_forward1_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_55_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_6_feed_forward1_linear1_weight_to_fp16_palettized, x = input_333_cast_fp16)[name = tensor("linear_55_cast_fp16")]; - tensor input_337_cast_fp16 = silu(x = linear_55_cast_fp16)[name = tensor("input_337_cast_fp16")]; - tensor module_layers_6_feed_forward1_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(116896640))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(120042432))), name = tensor("module_layers_6_feed_forward1_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_56_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_6_feed_forward1_linear2_weight_to_fp16_palettized, x = input_337_cast_fp16)[name = tensor("linear_56_cast_fp16")]; - tensor var_1210_to_fp16 = const()[name = tensor("op_1210_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1211_cast_fp16 = mul(x = linear_56_cast_fp16, y = var_1210_to_fp16)[name = tensor("op_1211_cast_fp16")]; - tensor input_343_cast_fp16 = add(x = input_331_cast_fp16, y = var_1211_cast_fp16)[name = tensor("input_343_cast_fp16")]; - tensor query_13_axes_0 = const()[name = tensor("query_13_axes_0"), val = tensor([-1])]; - tensor module_layers_6_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_6_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(120042624)))]; - tensor module_layers_6_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_6_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(120044736)))]; - tensor query_13_cast_fp16 = layer_norm(axes = query_13_axes_0, beta = module_layers_6_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_6_norm_self_att_weight_to_fp16, x = input_343_cast_fp16)[name = tensor("query_13_cast_fp16")]; - tensor module_layers_6_self_attn_linear_q_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(120046848))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(120833344))), name = tensor("module_layers_6_self_attn_linear_q_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_57_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_6_self_attn_linear_q_weight_to_fp16_palettized, x = query_13_cast_fp16)[name = tensor("linear_57_cast_fp16")]; - tensor var_1227 = const()[name = tensor("op_1227"), val = tensor([1, -1, 8, 128])]; - tensor q_37_cast_fp16 = reshape(shape = var_1227, x = linear_57_cast_fp16)[name = tensor("q_37_cast_fp16")]; - tensor module_layers_6_self_attn_linear_k_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(120833536))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121620032))), name = tensor("module_layers_6_self_attn_linear_k_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_58_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_6_self_attn_linear_k_weight_to_fp16_palettized, x = query_13_cast_fp16)[name = tensor("linear_58_cast_fp16")]; - tensor var_1231 = const()[name = tensor("op_1231"), val = tensor([1, -1, 8, 128])]; - tensor k_25_cast_fp16 = reshape(shape = var_1231, x = linear_58_cast_fp16)[name = tensor("k_25_cast_fp16")]; - tensor module_layers_6_self_attn_linear_v_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121620224))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(122406720))), name = tensor("module_layers_6_self_attn_linear_v_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_59_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_6_self_attn_linear_v_weight_to_fp16_palettized, x = query_13_cast_fp16)[name = tensor("linear_59_cast_fp16")]; - tensor var_1235 = const()[name = tensor("op_1235"), val = tensor([1, -1, 8, 128])]; - tensor v_13_cast_fp16 = reshape(shape = var_1235, x = linear_59_cast_fp16)[name = tensor("v_13_cast_fp16")]; - tensor value_15_perm_0 = const()[name = tensor("value_15_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor module_layers_6_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_6_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(122406912)))]; - tensor var_1247_cast_fp16 = add(x = q_37_cast_fp16, y = module_layers_6_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1247_cast_fp16")]; - tensor module_layers_6_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_6_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(122409024)))]; - tensor var_1249_cast_fp16 = add(x = q_37_cast_fp16, y = module_layers_6_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1249_cast_fp16")]; - tensor q_with_bias_v_13_perm_0 = const()[name = tensor("q_with_bias_v_13_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor x_139_transpose_x_0 = const()[name = tensor("x_139_transpose_x_0"), val = tensor(false)]; - tensor x_139_transpose_y_0 = const()[name = tensor("x_139_transpose_y_0"), val = tensor(false)]; - tensor op_1251_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(122411136))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(122699200))), name = tensor("op_1251_to_fp16_palettized"), shape = tensor([1, 8, 128, 375])]; - tensor q_with_bias_v_13_cast_fp16 = transpose(perm = q_with_bias_v_13_perm_0, x = var_1249_cast_fp16)[name = tensor("transpose_270")]; - tensor x_139_cast_fp16 = matmul(transpose_x = x_139_transpose_x_0, transpose_y = x_139_transpose_y_0, x = q_with_bias_v_13_cast_fp16, y = op_1251_to_fp16_palettized)[name = tensor("x_139_cast_fp16")]; - tensor x_141_pad_0 = const()[name = tensor("x_141_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_141_mode_0 = const()[name = tensor("x_141_mode_0"), val = tensor("constant")]; - tensor const_74_to_fp16 = const()[name = tensor("const_74_to_fp16"), val = tensor(0x0p+0)]; - tensor x_141_cast_fp16 = pad(constant_val = const_74_to_fp16, mode = x_141_mode_0, pad = x_141_pad_0, x = x_139_cast_fp16)[name = tensor("x_141_cast_fp16")]; - tensor var_1259 = const()[name = tensor("op_1259"), val = tensor([1, 8, -1, 188])]; - tensor x_143_cast_fp16 = reshape(shape = var_1259, x = x_141_cast_fp16)[name = tensor("x_143_cast_fp16")]; - tensor var_1263_begin_0 = const()[name = tensor("op_1263_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_1263_end_0 = const()[name = tensor("op_1263_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_1263_end_mask_0 = const()[name = tensor("op_1263_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_1263_cast_fp16 = slice_by_index(begin = var_1263_begin_0, end = var_1263_end_0, end_mask = var_1263_end_mask_0, x = x_143_cast_fp16)[name = tensor("op_1263_cast_fp16")]; - tensor var_1264 = const()[name = tensor("op_1264"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_25_cast_fp16 = reshape(shape = var_1264, x = var_1263_cast_fp16)[name = tensor("matrix_bd_25_cast_fp16")]; - tensor matrix_ac_13_transpose_x_0 = const()[name = tensor("matrix_ac_13_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_13_transpose_y_0 = const()[name = tensor("matrix_ac_13_transpose_y_0"), val = tensor(false)]; - tensor transpose_108_perm_0 = const()[name = tensor("transpose_108_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_109_perm_0 = const()[name = tensor("transpose_109_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_109 = transpose(perm = transpose_109_perm_0, x = k_25_cast_fp16)[name = tensor("transpose_268")]; - tensor transpose_108 = transpose(perm = transpose_108_perm_0, x = var_1247_cast_fp16)[name = tensor("transpose_269")]; - tensor matrix_ac_13_cast_fp16 = matmul(transpose_x = matrix_ac_13_transpose_x_0, transpose_y = matrix_ac_13_transpose_y_0, x = transpose_108, y = transpose_109)[name = tensor("matrix_ac_13_cast_fp16")]; - tensor matrix_bd_27_begin_0 = const()[name = tensor("matrix_bd_27_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_27_end_0 = const()[name = tensor("matrix_bd_27_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_27_end_mask_0 = const()[name = tensor("matrix_bd_27_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_27_cast_fp16 = slice_by_index(begin = matrix_bd_27_begin_0, end = matrix_bd_27_end_0, end_mask = matrix_bd_27_end_mask_0, x = matrix_bd_25_cast_fp16)[name = tensor("matrix_bd_27_cast_fp16")]; - tensor var_1273_cast_fp16 = add(x = matrix_ac_13_cast_fp16, y = matrix_bd_27_cast_fp16)[name = tensor("op_1273_cast_fp16")]; - tensor _inversed_scores_25_y_0_to_fp16 = const()[name = tensor("_inversed_scores_25_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; - tensor _inversed_scores_25_cast_fp16 = mul(x = var_1273_cast_fp16, y = _inversed_scores_25_y_0_to_fp16)[name = tensor("_inversed_scores_25_cast_fp16")]; - tensor scores_27_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_25_cast_fp16, cond = mask_3)[name = tensor("scores_27_cast_fp16")]; - tensor var_1279_cast_fp16 = softmax(axis = var_30, x = scores_27_cast_fp16)[name = tensor("op_1279_cast_fp16")]; - tensor input_345_cast_fp16 = select(a = var_11_to_fp16, b = var_1279_cast_fp16, cond = mask_3)[name = tensor("input_345_cast_fp16")]; - tensor x_145_transpose_x_0 = const()[name = tensor("x_145_transpose_x_0"), val = tensor(false)]; - tensor x_145_transpose_y_0 = const()[name = tensor("x_145_transpose_y_0"), val = tensor(false)]; - tensor value_15_cast_fp16 = transpose(perm = value_15_perm_0, x = v_13_cast_fp16)[name = tensor("transpose_267")]; - tensor x_145_cast_fp16 = matmul(transpose_x = x_145_transpose_x_0, transpose_y = x_145_transpose_y_0, x = input_345_cast_fp16, y = value_15_cast_fp16)[name = tensor("x_145_cast_fp16")]; - tensor var_1283_perm_0 = const()[name = tensor("op_1283_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_1284 = const()[name = tensor("op_1284"), val = tensor([1, -1, 1024])]; - tensor var_1283_cast_fp16 = transpose(perm = var_1283_perm_0, x = x_145_cast_fp16)[name = tensor("transpose_266")]; - tensor input_347_cast_fp16 = reshape(shape = var_1284, x = var_1283_cast_fp16)[name = tensor("input_347_cast_fp16")]; - tensor module_layers_6_self_attn_linear_out_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(122699392))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123485888))), name = tensor("module_layers_6_self_attn_linear_out_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_61_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_6_self_attn_linear_out_weight_to_fp16_palettized, x = input_347_cast_fp16)[name = tensor("linear_61_cast_fp16")]; - tensor input_351_cast_fp16 = add(x = input_343_cast_fp16, y = linear_61_cast_fp16)[name = tensor("input_351_cast_fp16")]; - tensor x_149_axes_0 = const()[name = tensor("x_149_axes_0"), val = tensor([-1])]; - tensor module_layers_6_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_6_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123486080)))]; - tensor module_layers_6_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_6_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123488192)))]; - tensor x_149_cast_fp16 = layer_norm(axes = x_149_axes_0, beta = module_layers_6_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_6_norm_conv_weight_to_fp16, x = input_351_cast_fp16)[name = tensor("x_149_cast_fp16")]; - tensor input_353_perm_0 = const()[name = tensor("input_353_perm_0"), val = tensor([0, 2, 1])]; - tensor input_355_pad_type_0 = const()[name = tensor("input_355_pad_type_0"), val = tensor("valid")]; - tensor input_355_strides_0 = const()[name = tensor("input_355_strides_0"), val = tensor([1])]; - tensor input_355_pad_0 = const()[name = tensor("input_355_pad_0"), val = tensor([0, 0])]; - tensor input_355_dilations_0 = const()[name = tensor("input_355_dilations_0"), val = tensor([1])]; - tensor input_355_groups_0 = const()[name = tensor("input_355_groups_0"), val = tensor(1)]; - tensor module_layers_6_conv_pointwise_conv1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123490304))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(125063232))), name = tensor("module_layers_6_conv_pointwise_conv1_weight_to_fp16_palettized"), shape = tensor([2048, 1024, 1])]; - tensor input_353_cast_fp16 = transpose(perm = input_353_perm_0, x = x_149_cast_fp16)[name = tensor("transpose_265")]; - tensor input_355_cast_fp16 = conv(dilations = input_355_dilations_0, groups = input_355_groups_0, pad = input_355_pad_0, pad_type = input_355_pad_type_0, strides = input_355_strides_0, weight = module_layers_6_conv_pointwise_conv1_weight_to_fp16_palettized, x = input_353_cast_fp16)[name = tensor("input_355_cast_fp16")]; - tensor x_151_split_num_splits_0 = const()[name = tensor("x_151_split_num_splits_0"), val = tensor(2)]; - tensor x_151_split_axis_0 = const()[name = tensor("x_151_split_axis_0"), val = tensor(1)]; - tensor x_151_split_cast_fp16_0, tensor x_151_split_cast_fp16_1 = split(axis = x_151_split_axis_0, num_splits = x_151_split_num_splits_0, x = input_355_cast_fp16)[name = tensor("x_151_split_cast_fp16")]; - tensor x_151_split_1_sigmoid_cast_fp16 = sigmoid(x = x_151_split_cast_fp16_1)[name = tensor("x_151_split_1_sigmoid_cast_fp16")]; - tensor x_151_cast_fp16 = mul(x = x_151_split_cast_fp16_0, y = x_151_split_1_sigmoid_cast_fp16)[name = tensor("x_151_cast_fp16")]; - tensor input_357_cast_fp16 = select(a = var_11_to_fp16, b = x_151_cast_fp16, cond = var_328)[name = tensor("input_357_cast_fp16")]; - tensor input_359_pad_0 = const()[name = tensor("input_359_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_359_mode_0 = const()[name = tensor("input_359_mode_0"), val = tensor("constant")]; - tensor const_77_to_fp16 = const()[name = tensor("const_77_to_fp16"), val = tensor(0x0p+0)]; - tensor input_359_cast_fp16 = pad(constant_val = const_77_to_fp16, mode = input_359_mode_0, pad = input_359_pad_0, x = input_357_cast_fp16)[name = tensor("input_359_cast_fp16")]; - tensor input_361_pad_type_0 = const()[name = tensor("input_361_pad_type_0"), val = tensor("valid")]; - tensor input_361_groups_0 = const()[name = tensor("input_361_groups_0"), val = tensor(1024)]; - tensor input_361_strides_0 = const()[name = tensor("input_361_strides_0"), val = tensor([1])]; - tensor input_361_pad_0 = const()[name = tensor("input_361_pad_0"), val = tensor([0, 0])]; - tensor input_361_dilations_0 = const()[name = tensor("input_361_dilations_0"), val = tensor([1])]; - tensor const_260_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(125063424))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(125070400))), name = tensor("const_260_to_fp16_palettized"), shape = tensor([1024, 1, 9])]; - tensor const_261_to_fp16 = const()[name = tensor("const_261_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(125070592)))]; - tensor input_363_cast_fp16 = conv(bias = const_261_to_fp16, dilations = input_361_dilations_0, groups = input_361_groups_0, pad = input_361_pad_0, pad_type = input_361_pad_type_0, strides = input_361_strides_0, weight = const_260_to_fp16_palettized, x = input_359_cast_fp16)[name = tensor("input_363_cast_fp16")]; - tensor input_365_cast_fp16 = silu(x = input_363_cast_fp16)[name = tensor("input_365_cast_fp16")]; - tensor x_153_pad_type_0 = const()[name = tensor("x_153_pad_type_0"), val = tensor("valid")]; - tensor x_153_strides_0 = const()[name = tensor("x_153_strides_0"), val = tensor([1])]; - tensor x_153_pad_0 = const()[name = tensor("x_153_pad_0"), val = tensor([0, 0])]; - tensor x_153_dilations_0 = const()[name = tensor("x_153_dilations_0"), val = tensor([1])]; - tensor x_153_groups_0 = const()[name = tensor("x_153_groups_0"), val = tensor(1)]; - tensor module_layers_6_conv_pointwise_conv2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(125072704))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(125859200))), name = tensor("module_layers_6_conv_pointwise_conv2_weight_to_fp16_palettized"), shape = tensor([1024, 1024, 1])]; - tensor x_153_cast_fp16 = conv(dilations = x_153_dilations_0, groups = x_153_groups_0, pad = x_153_pad_0, pad_type = x_153_pad_type_0, strides = x_153_strides_0, weight = module_layers_6_conv_pointwise_conv2_weight_to_fp16_palettized, x = input_365_cast_fp16)[name = tensor("x_153_cast_fp16")]; - tensor input_367_perm_0 = const()[name = tensor("input_367_perm_0"), val = tensor([0, 2, 1])]; - tensor input_367_cast_fp16 = transpose(perm = input_367_perm_0, x = x_153_cast_fp16)[name = tensor("transpose_264")]; - tensor input_369_cast_fp16 = add(x = input_351_cast_fp16, y = input_367_cast_fp16)[name = tensor("input_369_cast_fp16")]; - tensor input_371_axes_0 = const()[name = tensor("input_371_axes_0"), val = tensor([-1])]; - tensor module_layers_6_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_6_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(125859392)))]; - tensor module_layers_6_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_6_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(125861504)))]; - tensor input_371_cast_fp16 = layer_norm(axes = input_371_axes_0, beta = module_layers_6_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_6_norm_feed_forward2_weight_to_fp16, x = input_369_cast_fp16)[name = tensor("input_371_cast_fp16")]; - tensor module_layers_6_feed_forward2_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(125863616))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(129009408))), name = tensor("module_layers_6_feed_forward2_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_62_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_6_feed_forward2_linear1_weight_to_fp16_palettized, x = input_371_cast_fp16)[name = tensor("linear_62_cast_fp16")]; - tensor input_375_cast_fp16 = silu(x = linear_62_cast_fp16)[name = tensor("input_375_cast_fp16")]; - tensor module_layers_6_feed_forward2_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(129009600))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132155392))), name = tensor("module_layers_6_feed_forward2_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_63_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_6_feed_forward2_linear2_weight_to_fp16_palettized, x = input_375_cast_fp16)[name = tensor("linear_63_cast_fp16")]; - tensor var_1344_to_fp16 = const()[name = tensor("op_1344_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1345_cast_fp16 = mul(x = linear_63_cast_fp16, y = var_1344_to_fp16)[name = tensor("op_1345_cast_fp16")]; - tensor input_381_cast_fp16 = add(x = input_369_cast_fp16, y = var_1345_cast_fp16)[name = tensor("input_381_cast_fp16")]; - tensor input_383_axes_0 = const()[name = tensor("input_383_axes_0"), val = tensor([-1])]; - tensor module_layers_6_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_6_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132155584)))]; - tensor module_layers_6_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_6_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132157696)))]; - tensor input_383_cast_fp16 = layer_norm(axes = input_383_axes_0, beta = module_layers_6_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_6_norm_out_weight_to_fp16, x = input_381_cast_fp16)[name = tensor("input_383_cast_fp16")]; - tensor input_385_axes_0 = const()[name = tensor("input_385_axes_0"), val = tensor([-1])]; - tensor module_layers_7_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_7_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132159808)))]; - tensor module_layers_7_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_7_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132161920)))]; - tensor input_385_cast_fp16 = layer_norm(axes = input_385_axes_0, beta = module_layers_7_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_7_norm_feed_forward1_weight_to_fp16, x = input_383_cast_fp16)[name = tensor("input_385_cast_fp16")]; - tensor module_layers_7_feed_forward1_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132164032))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(135309824))), name = tensor("module_layers_7_feed_forward1_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_64_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_7_feed_forward1_linear1_weight_to_fp16_palettized, x = input_385_cast_fp16)[name = tensor("linear_64_cast_fp16")]; - tensor input_389_cast_fp16 = silu(x = linear_64_cast_fp16)[name = tensor("input_389_cast_fp16")]; - tensor module_layers_7_feed_forward1_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(135310016))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(138455808))), name = tensor("module_layers_7_feed_forward1_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_65_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_7_feed_forward1_linear2_weight_to_fp16_palettized, x = input_389_cast_fp16)[name = tensor("linear_65_cast_fp16")]; - tensor var_1373_to_fp16 = const()[name = tensor("op_1373_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1374_cast_fp16 = mul(x = linear_65_cast_fp16, y = var_1373_to_fp16)[name = tensor("op_1374_cast_fp16")]; - tensor input_395_cast_fp16 = add(x = input_383_cast_fp16, y = var_1374_cast_fp16)[name = tensor("input_395_cast_fp16")]; - tensor query_15_axes_0 = const()[name = tensor("query_15_axes_0"), val = tensor([-1])]; - tensor module_layers_7_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_7_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(138456000)))]; - tensor module_layers_7_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_7_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(138458112)))]; - tensor query_15_cast_fp16 = layer_norm(axes = query_15_axes_0, beta = module_layers_7_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_7_norm_self_att_weight_to_fp16, x = input_395_cast_fp16)[name = tensor("query_15_cast_fp16")]; - tensor module_layers_7_self_attn_linear_q_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(138460224))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(139246720))), name = tensor("module_layers_7_self_attn_linear_q_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_66_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_7_self_attn_linear_q_weight_to_fp16_palettized, x = query_15_cast_fp16)[name = tensor("linear_66_cast_fp16")]; - tensor var_1390 = const()[name = tensor("op_1390"), val = tensor([1, -1, 8, 128])]; - tensor q_43_cast_fp16 = reshape(shape = var_1390, x = linear_66_cast_fp16)[name = tensor("q_43_cast_fp16")]; - tensor module_layers_7_self_attn_linear_k_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(139246912))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140033408))), name = tensor("module_layers_7_self_attn_linear_k_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_67_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_7_self_attn_linear_k_weight_to_fp16_palettized, x = query_15_cast_fp16)[name = tensor("linear_67_cast_fp16")]; - tensor var_1394 = const()[name = tensor("op_1394"), val = tensor([1, -1, 8, 128])]; - tensor k_29_cast_fp16 = reshape(shape = var_1394, x = linear_67_cast_fp16)[name = tensor("k_29_cast_fp16")]; - tensor module_layers_7_self_attn_linear_v_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140033600))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140820096))), name = tensor("module_layers_7_self_attn_linear_v_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_68_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_7_self_attn_linear_v_weight_to_fp16_palettized, x = query_15_cast_fp16)[name = tensor("linear_68_cast_fp16")]; - tensor var_1398 = const()[name = tensor("op_1398"), val = tensor([1, -1, 8, 128])]; - tensor v_15_cast_fp16 = reshape(shape = var_1398, x = linear_68_cast_fp16)[name = tensor("v_15_cast_fp16")]; - tensor value_17_perm_0 = const()[name = tensor("value_17_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor module_layers_7_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_7_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140820288)))]; - tensor var_1410_cast_fp16 = add(x = q_43_cast_fp16, y = module_layers_7_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1410_cast_fp16")]; - tensor module_layers_7_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_7_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140822400)))]; - tensor var_1412_cast_fp16 = add(x = q_43_cast_fp16, y = module_layers_7_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1412_cast_fp16")]; - tensor q_with_bias_v_15_perm_0 = const()[name = tensor("q_with_bias_v_15_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor x_161_transpose_x_0 = const()[name = tensor("x_161_transpose_x_0"), val = tensor(false)]; - tensor x_161_transpose_y_0 = const()[name = tensor("x_161_transpose_y_0"), val = tensor(false)]; - tensor op_1414_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140824512))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141112576))), name = tensor("op_1414_to_fp16_palettized"), shape = tensor([1, 8, 128, 375])]; - tensor q_with_bias_v_15_cast_fp16 = transpose(perm = q_with_bias_v_15_perm_0, x = var_1412_cast_fp16)[name = tensor("transpose_263")]; - tensor x_161_cast_fp16 = matmul(transpose_x = x_161_transpose_x_0, transpose_y = x_161_transpose_y_0, x = q_with_bias_v_15_cast_fp16, y = op_1414_to_fp16_palettized)[name = tensor("x_161_cast_fp16")]; - tensor x_163_pad_0 = const()[name = tensor("x_163_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_163_mode_0 = const()[name = tensor("x_163_mode_0"), val = tensor("constant")]; - tensor const_84_to_fp16 = const()[name = tensor("const_84_to_fp16"), val = tensor(0x0p+0)]; - tensor x_163_cast_fp16 = pad(constant_val = const_84_to_fp16, mode = x_163_mode_0, pad = x_163_pad_0, x = x_161_cast_fp16)[name = tensor("x_163_cast_fp16")]; - tensor var_1422 = const()[name = tensor("op_1422"), val = tensor([1, 8, -1, 188])]; - tensor x_165_cast_fp16 = reshape(shape = var_1422, x = x_163_cast_fp16)[name = tensor("x_165_cast_fp16")]; - tensor var_1426_begin_0 = const()[name = tensor("op_1426_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_1426_end_0 = const()[name = tensor("op_1426_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_1426_end_mask_0 = const()[name = tensor("op_1426_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_1426_cast_fp16 = slice_by_index(begin = var_1426_begin_0, end = var_1426_end_0, end_mask = var_1426_end_mask_0, x = x_165_cast_fp16)[name = tensor("op_1426_cast_fp16")]; - tensor var_1427 = const()[name = tensor("op_1427"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_29_cast_fp16 = reshape(shape = var_1427, x = var_1426_cast_fp16)[name = tensor("matrix_bd_29_cast_fp16")]; - tensor matrix_ac_15_transpose_x_0 = const()[name = tensor("matrix_ac_15_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_15_transpose_y_0 = const()[name = tensor("matrix_ac_15_transpose_y_0"), val = tensor(false)]; - tensor transpose_110_perm_0 = const()[name = tensor("transpose_110_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_111_perm_0 = const()[name = tensor("transpose_111_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_111 = transpose(perm = transpose_111_perm_0, x = k_29_cast_fp16)[name = tensor("transpose_261")]; - tensor transpose_110 = transpose(perm = transpose_110_perm_0, x = var_1410_cast_fp16)[name = tensor("transpose_262")]; - tensor matrix_ac_15_cast_fp16 = matmul(transpose_x = matrix_ac_15_transpose_x_0, transpose_y = matrix_ac_15_transpose_y_0, x = transpose_110, y = transpose_111)[name = tensor("matrix_ac_15_cast_fp16")]; - tensor matrix_bd_31_begin_0 = const()[name = tensor("matrix_bd_31_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_31_end_0 = const()[name = tensor("matrix_bd_31_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_31_end_mask_0 = const()[name = tensor("matrix_bd_31_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_31_cast_fp16 = slice_by_index(begin = matrix_bd_31_begin_0, end = matrix_bd_31_end_0, end_mask = matrix_bd_31_end_mask_0, x = matrix_bd_29_cast_fp16)[name = tensor("matrix_bd_31_cast_fp16")]; - tensor var_1436_cast_fp16 = add(x = matrix_ac_15_cast_fp16, y = matrix_bd_31_cast_fp16)[name = tensor("op_1436_cast_fp16")]; - tensor _inversed_scores_29_y_0_to_fp16 = const()[name = tensor("_inversed_scores_29_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; - tensor _inversed_scores_29_cast_fp16 = mul(x = var_1436_cast_fp16, y = _inversed_scores_29_y_0_to_fp16)[name = tensor("_inversed_scores_29_cast_fp16")]; - tensor scores_31_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_29_cast_fp16, cond = mask_3)[name = tensor("scores_31_cast_fp16")]; - tensor var_1442_cast_fp16 = softmax(axis = var_30, x = scores_31_cast_fp16)[name = tensor("op_1442_cast_fp16")]; - tensor input_397_cast_fp16 = select(a = var_11_to_fp16, b = var_1442_cast_fp16, cond = mask_3)[name = tensor("input_397_cast_fp16")]; - tensor x_167_transpose_x_0 = const()[name = tensor("x_167_transpose_x_0"), val = tensor(false)]; - tensor x_167_transpose_y_0 = const()[name = tensor("x_167_transpose_y_0"), val = tensor(false)]; - tensor value_17_cast_fp16 = transpose(perm = value_17_perm_0, x = v_15_cast_fp16)[name = tensor("transpose_260")]; - tensor x_167_cast_fp16 = matmul(transpose_x = x_167_transpose_x_0, transpose_y = x_167_transpose_y_0, x = input_397_cast_fp16, y = value_17_cast_fp16)[name = tensor("x_167_cast_fp16")]; - tensor var_1446_perm_0 = const()[name = tensor("op_1446_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_1447 = const()[name = tensor("op_1447"), val = tensor([1, -1, 1024])]; - tensor var_1446_cast_fp16 = transpose(perm = var_1446_perm_0, x = x_167_cast_fp16)[name = tensor("transpose_259")]; - tensor input_399_cast_fp16 = reshape(shape = var_1447, x = var_1446_cast_fp16)[name = tensor("input_399_cast_fp16")]; - tensor module_layers_7_self_attn_linear_out_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141112768))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141899264))), name = tensor("module_layers_7_self_attn_linear_out_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_70_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_7_self_attn_linear_out_weight_to_fp16_palettized, x = input_399_cast_fp16)[name = tensor("linear_70_cast_fp16")]; - tensor input_403_cast_fp16 = add(x = input_395_cast_fp16, y = linear_70_cast_fp16)[name = tensor("input_403_cast_fp16")]; - tensor x_171_axes_0 = const()[name = tensor("x_171_axes_0"), val = tensor([-1])]; - tensor module_layers_7_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_7_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141899456)))]; - tensor module_layers_7_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_7_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141901568)))]; - tensor x_171_cast_fp16 = layer_norm(axes = x_171_axes_0, beta = module_layers_7_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_7_norm_conv_weight_to_fp16, x = input_403_cast_fp16)[name = tensor("x_171_cast_fp16")]; - tensor input_405_perm_0 = const()[name = tensor("input_405_perm_0"), val = tensor([0, 2, 1])]; - tensor input_407_pad_type_0 = const()[name = tensor("input_407_pad_type_0"), val = tensor("valid")]; - tensor input_407_strides_0 = const()[name = tensor("input_407_strides_0"), val = tensor([1])]; - tensor input_407_pad_0 = const()[name = tensor("input_407_pad_0"), val = tensor([0, 0])]; - tensor input_407_dilations_0 = const()[name = tensor("input_407_dilations_0"), val = tensor([1])]; - tensor input_407_groups_0 = const()[name = tensor("input_407_groups_0"), val = tensor(1)]; - tensor module_layers_7_conv_pointwise_conv1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141903680))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(143476608))), name = tensor("module_layers_7_conv_pointwise_conv1_weight_to_fp16_palettized"), shape = tensor([2048, 1024, 1])]; - tensor input_405_cast_fp16 = transpose(perm = input_405_perm_0, x = x_171_cast_fp16)[name = tensor("transpose_258")]; - tensor input_407_cast_fp16 = conv(dilations = input_407_dilations_0, groups = input_407_groups_0, pad = input_407_pad_0, pad_type = input_407_pad_type_0, strides = input_407_strides_0, weight = module_layers_7_conv_pointwise_conv1_weight_to_fp16_palettized, x = input_405_cast_fp16)[name = tensor("input_407_cast_fp16")]; - tensor x_173_split_num_splits_0 = const()[name = tensor("x_173_split_num_splits_0"), val = tensor(2)]; - tensor x_173_split_axis_0 = const()[name = tensor("x_173_split_axis_0"), val = tensor(1)]; - tensor x_173_split_cast_fp16_0, tensor x_173_split_cast_fp16_1 = split(axis = x_173_split_axis_0, num_splits = x_173_split_num_splits_0, x = input_407_cast_fp16)[name = tensor("x_173_split_cast_fp16")]; - tensor x_173_split_1_sigmoid_cast_fp16 = sigmoid(x = x_173_split_cast_fp16_1)[name = tensor("x_173_split_1_sigmoid_cast_fp16")]; - tensor x_173_cast_fp16 = mul(x = x_173_split_cast_fp16_0, y = x_173_split_1_sigmoid_cast_fp16)[name = tensor("x_173_cast_fp16")]; - tensor input_409_cast_fp16 = select(a = var_11_to_fp16, b = x_173_cast_fp16, cond = var_328)[name = tensor("input_409_cast_fp16")]; - tensor input_411_pad_0 = const()[name = tensor("input_411_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_411_mode_0 = const()[name = tensor("input_411_mode_0"), val = tensor("constant")]; - tensor const_87_to_fp16 = const()[name = tensor("const_87_to_fp16"), val = tensor(0x0p+0)]; - tensor input_411_cast_fp16 = pad(constant_val = const_87_to_fp16, mode = input_411_mode_0, pad = input_411_pad_0, x = input_409_cast_fp16)[name = tensor("input_411_cast_fp16")]; - tensor input_413_pad_type_0 = const()[name = tensor("input_413_pad_type_0"), val = tensor("valid")]; - tensor input_413_groups_0 = const()[name = tensor("input_413_groups_0"), val = tensor(1024)]; - tensor input_413_strides_0 = const()[name = tensor("input_413_strides_0"), val = tensor([1])]; - tensor input_413_pad_0 = const()[name = tensor("input_413_pad_0"), val = tensor([0, 0])]; - tensor input_413_dilations_0 = const()[name = tensor("input_413_dilations_0"), val = tensor([1])]; - tensor const_262_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(143476800))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(143483776))), name = tensor("const_262_to_fp16_palettized"), shape = tensor([1024, 1, 9])]; - tensor const_263_to_fp16 = const()[name = tensor("const_263_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(143483968)))]; - tensor input_415_cast_fp16 = conv(bias = const_263_to_fp16, dilations = input_413_dilations_0, groups = input_413_groups_0, pad = input_413_pad_0, pad_type = input_413_pad_type_0, strides = input_413_strides_0, weight = const_262_to_fp16_palettized, x = input_411_cast_fp16)[name = tensor("input_415_cast_fp16")]; - tensor input_417_cast_fp16 = silu(x = input_415_cast_fp16)[name = tensor("input_417_cast_fp16")]; - tensor x_175_pad_type_0 = const()[name = tensor("x_175_pad_type_0"), val = tensor("valid")]; - tensor x_175_strides_0 = const()[name = tensor("x_175_strides_0"), val = tensor([1])]; - tensor x_175_pad_0 = const()[name = tensor("x_175_pad_0"), val = tensor([0, 0])]; - tensor x_175_dilations_0 = const()[name = tensor("x_175_dilations_0"), val = tensor([1])]; - tensor x_175_groups_0 = const()[name = tensor("x_175_groups_0"), val = tensor(1)]; - tensor module_layers_7_conv_pointwise_conv2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(143486080))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(144272576))), name = tensor("module_layers_7_conv_pointwise_conv2_weight_to_fp16_palettized"), shape = tensor([1024, 1024, 1])]; - tensor x_175_cast_fp16 = conv(dilations = x_175_dilations_0, groups = x_175_groups_0, pad = x_175_pad_0, pad_type = x_175_pad_type_0, strides = x_175_strides_0, weight = module_layers_7_conv_pointwise_conv2_weight_to_fp16_palettized, x = input_417_cast_fp16)[name = tensor("x_175_cast_fp16")]; - tensor input_419_perm_0 = const()[name = tensor("input_419_perm_0"), val = tensor([0, 2, 1])]; - tensor input_419_cast_fp16 = transpose(perm = input_419_perm_0, x = x_175_cast_fp16)[name = tensor("transpose_257")]; - tensor input_421_cast_fp16 = add(x = input_403_cast_fp16, y = input_419_cast_fp16)[name = tensor("input_421_cast_fp16")]; - tensor input_423_axes_0 = const()[name = tensor("input_423_axes_0"), val = tensor([-1])]; - tensor module_layers_7_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_7_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(144272768)))]; - tensor module_layers_7_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_7_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(144274880)))]; - tensor input_423_cast_fp16 = layer_norm(axes = input_423_axes_0, beta = module_layers_7_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_7_norm_feed_forward2_weight_to_fp16, x = input_421_cast_fp16)[name = tensor("input_423_cast_fp16")]; - tensor module_layers_7_feed_forward2_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(144276992))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147422784))), name = tensor("module_layers_7_feed_forward2_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_71_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_7_feed_forward2_linear1_weight_to_fp16_palettized, x = input_423_cast_fp16)[name = tensor("linear_71_cast_fp16")]; - tensor input_427_cast_fp16 = silu(x = linear_71_cast_fp16)[name = tensor("input_427_cast_fp16")]; - tensor module_layers_7_feed_forward2_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147422976))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(150568768))), name = tensor("module_layers_7_feed_forward2_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_72_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_7_feed_forward2_linear2_weight_to_fp16_palettized, x = input_427_cast_fp16)[name = tensor("linear_72_cast_fp16")]; - tensor var_1507_to_fp16 = const()[name = tensor("op_1507_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1508_cast_fp16 = mul(x = linear_72_cast_fp16, y = var_1507_to_fp16)[name = tensor("op_1508_cast_fp16")]; - tensor input_433_cast_fp16 = add(x = input_421_cast_fp16, y = var_1508_cast_fp16)[name = tensor("input_433_cast_fp16")]; - tensor input_435_axes_0 = const()[name = tensor("input_435_axes_0"), val = tensor([-1])]; - tensor module_layers_7_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_7_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(150568960)))]; - tensor module_layers_7_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_7_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(150571072)))]; - tensor input_435_cast_fp16 = layer_norm(axes = input_435_axes_0, beta = module_layers_7_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_7_norm_out_weight_to_fp16, x = input_433_cast_fp16)[name = tensor("input_435_cast_fp16")]; - tensor input_437_axes_0 = const()[name = tensor("input_437_axes_0"), val = tensor([-1])]; - tensor module_layers_8_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_8_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(150573184)))]; - tensor module_layers_8_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_8_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(150575296)))]; - tensor input_437_cast_fp16 = layer_norm(axes = input_437_axes_0, beta = module_layers_8_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_8_norm_feed_forward1_weight_to_fp16, x = input_435_cast_fp16)[name = tensor("input_437_cast_fp16")]; - tensor module_layers_8_feed_forward1_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(150577408))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(153723200))), name = tensor("module_layers_8_feed_forward1_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_73_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_8_feed_forward1_linear1_weight_to_fp16_palettized, x = input_437_cast_fp16)[name = tensor("linear_73_cast_fp16")]; - tensor input_441_cast_fp16 = silu(x = linear_73_cast_fp16)[name = tensor("input_441_cast_fp16")]; - tensor module_layers_8_feed_forward1_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(153723392))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(156869184))), name = tensor("module_layers_8_feed_forward1_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_74_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_8_feed_forward1_linear2_weight_to_fp16_palettized, x = input_441_cast_fp16)[name = tensor("linear_74_cast_fp16")]; - tensor var_1536_to_fp16 = const()[name = tensor("op_1536_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1537_cast_fp16 = mul(x = linear_74_cast_fp16, y = var_1536_to_fp16)[name = tensor("op_1537_cast_fp16")]; - tensor input_447_cast_fp16 = add(x = input_435_cast_fp16, y = var_1537_cast_fp16)[name = tensor("input_447_cast_fp16")]; - tensor query_17_axes_0 = const()[name = tensor("query_17_axes_0"), val = tensor([-1])]; - tensor module_layers_8_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_8_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(156869376)))]; - tensor module_layers_8_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_8_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(156871488)))]; - tensor query_17_cast_fp16 = layer_norm(axes = query_17_axes_0, beta = module_layers_8_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_8_norm_self_att_weight_to_fp16, x = input_447_cast_fp16)[name = tensor("query_17_cast_fp16")]; - tensor module_layers_8_self_attn_linear_q_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(156873600))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157660096))), name = tensor("module_layers_8_self_attn_linear_q_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_75_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_8_self_attn_linear_q_weight_to_fp16_palettized, x = query_17_cast_fp16)[name = tensor("linear_75_cast_fp16")]; - tensor var_1553 = const()[name = tensor("op_1553"), val = tensor([1, -1, 8, 128])]; - tensor q_49_cast_fp16 = reshape(shape = var_1553, x = linear_75_cast_fp16)[name = tensor("q_49_cast_fp16")]; - tensor module_layers_8_self_attn_linear_k_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157660288))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158446784))), name = tensor("module_layers_8_self_attn_linear_k_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_76_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_8_self_attn_linear_k_weight_to_fp16_palettized, x = query_17_cast_fp16)[name = tensor("linear_76_cast_fp16")]; - tensor var_1557 = const()[name = tensor("op_1557"), val = tensor([1, -1, 8, 128])]; - tensor k_33_cast_fp16 = reshape(shape = var_1557, x = linear_76_cast_fp16)[name = tensor("k_33_cast_fp16")]; - tensor module_layers_8_self_attn_linear_v_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158446976))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159233472))), name = tensor("module_layers_8_self_attn_linear_v_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_77_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_8_self_attn_linear_v_weight_to_fp16_palettized, x = query_17_cast_fp16)[name = tensor("linear_77_cast_fp16")]; - tensor var_1561 = const()[name = tensor("op_1561"), val = tensor([1, -1, 8, 128])]; - tensor v_17_cast_fp16 = reshape(shape = var_1561, x = linear_77_cast_fp16)[name = tensor("v_17_cast_fp16")]; - tensor value_19_perm_0 = const()[name = tensor("value_19_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor module_layers_8_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_8_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159233664)))]; - tensor var_1573_cast_fp16 = add(x = q_49_cast_fp16, y = module_layers_8_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1573_cast_fp16")]; - tensor module_layers_8_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_8_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159235776)))]; - tensor var_1575_cast_fp16 = add(x = q_49_cast_fp16, y = module_layers_8_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1575_cast_fp16")]; - tensor q_with_bias_v_17_perm_0 = const()[name = tensor("q_with_bias_v_17_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor x_183_transpose_x_0 = const()[name = tensor("x_183_transpose_x_0"), val = tensor(false)]; - tensor x_183_transpose_y_0 = const()[name = tensor("x_183_transpose_y_0"), val = tensor(false)]; - tensor op_1577_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159237888))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159525952))), name = tensor("op_1577_to_fp16_palettized"), shape = tensor([1, 8, 128, 375])]; - tensor q_with_bias_v_17_cast_fp16 = transpose(perm = q_with_bias_v_17_perm_0, x = var_1575_cast_fp16)[name = tensor("transpose_256")]; - tensor x_183_cast_fp16 = matmul(transpose_x = x_183_transpose_x_0, transpose_y = x_183_transpose_y_0, x = q_with_bias_v_17_cast_fp16, y = op_1577_to_fp16_palettized)[name = tensor("x_183_cast_fp16")]; - tensor x_185_pad_0 = const()[name = tensor("x_185_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_185_mode_0 = const()[name = tensor("x_185_mode_0"), val = tensor("constant")]; - tensor const_94_to_fp16 = const()[name = tensor("const_94_to_fp16"), val = tensor(0x0p+0)]; - tensor x_185_cast_fp16 = pad(constant_val = const_94_to_fp16, mode = x_185_mode_0, pad = x_185_pad_0, x = x_183_cast_fp16)[name = tensor("x_185_cast_fp16")]; - tensor var_1585 = const()[name = tensor("op_1585"), val = tensor([1, 8, -1, 188])]; - tensor x_187_cast_fp16 = reshape(shape = var_1585, x = x_185_cast_fp16)[name = tensor("x_187_cast_fp16")]; - tensor var_1589_begin_0 = const()[name = tensor("op_1589_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_1589_end_0 = const()[name = tensor("op_1589_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_1589_end_mask_0 = const()[name = tensor("op_1589_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_1589_cast_fp16 = slice_by_index(begin = var_1589_begin_0, end = var_1589_end_0, end_mask = var_1589_end_mask_0, x = x_187_cast_fp16)[name = tensor("op_1589_cast_fp16")]; - tensor var_1590 = const()[name = tensor("op_1590"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_33_cast_fp16 = reshape(shape = var_1590, x = var_1589_cast_fp16)[name = tensor("matrix_bd_33_cast_fp16")]; - tensor matrix_ac_17_transpose_x_0 = const()[name = tensor("matrix_ac_17_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_17_transpose_y_0 = const()[name = tensor("matrix_ac_17_transpose_y_0"), val = tensor(false)]; - tensor transpose_112_perm_0 = const()[name = tensor("transpose_112_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_113_perm_0 = const()[name = tensor("transpose_113_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_113 = transpose(perm = transpose_113_perm_0, x = k_33_cast_fp16)[name = tensor("transpose_254")]; - tensor transpose_112 = transpose(perm = transpose_112_perm_0, x = var_1573_cast_fp16)[name = tensor("transpose_255")]; - tensor matrix_ac_17_cast_fp16 = matmul(transpose_x = matrix_ac_17_transpose_x_0, transpose_y = matrix_ac_17_transpose_y_0, x = transpose_112, y = transpose_113)[name = tensor("matrix_ac_17_cast_fp16")]; - tensor matrix_bd_35_begin_0 = const()[name = tensor("matrix_bd_35_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_35_end_0 = const()[name = tensor("matrix_bd_35_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_35_end_mask_0 = const()[name = tensor("matrix_bd_35_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_35_cast_fp16 = slice_by_index(begin = matrix_bd_35_begin_0, end = matrix_bd_35_end_0, end_mask = matrix_bd_35_end_mask_0, x = matrix_bd_33_cast_fp16)[name = tensor("matrix_bd_35_cast_fp16")]; - tensor var_1599_cast_fp16 = add(x = matrix_ac_17_cast_fp16, y = matrix_bd_35_cast_fp16)[name = tensor("op_1599_cast_fp16")]; - tensor _inversed_scores_33_y_0_to_fp16 = const()[name = tensor("_inversed_scores_33_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; - tensor _inversed_scores_33_cast_fp16 = mul(x = var_1599_cast_fp16, y = _inversed_scores_33_y_0_to_fp16)[name = tensor("_inversed_scores_33_cast_fp16")]; - tensor scores_35_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_33_cast_fp16, cond = mask_3)[name = tensor("scores_35_cast_fp16")]; - tensor var_1605_cast_fp16 = softmax(axis = var_30, x = scores_35_cast_fp16)[name = tensor("op_1605_cast_fp16")]; - tensor input_449_cast_fp16 = select(a = var_11_to_fp16, b = var_1605_cast_fp16, cond = mask_3)[name = tensor("input_449_cast_fp16")]; - tensor x_189_transpose_x_0 = const()[name = tensor("x_189_transpose_x_0"), val = tensor(false)]; - tensor x_189_transpose_y_0 = const()[name = tensor("x_189_transpose_y_0"), val = tensor(false)]; - tensor value_19_cast_fp16 = transpose(perm = value_19_perm_0, x = v_17_cast_fp16)[name = tensor("transpose_253")]; - tensor x_189_cast_fp16 = matmul(transpose_x = x_189_transpose_x_0, transpose_y = x_189_transpose_y_0, x = input_449_cast_fp16, y = value_19_cast_fp16)[name = tensor("x_189_cast_fp16")]; - tensor var_1609_perm_0 = const()[name = tensor("op_1609_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_1610 = const()[name = tensor("op_1610"), val = tensor([1, -1, 1024])]; - tensor var_1609_cast_fp16 = transpose(perm = var_1609_perm_0, x = x_189_cast_fp16)[name = tensor("transpose_252")]; - tensor input_451_cast_fp16 = reshape(shape = var_1610, x = var_1609_cast_fp16)[name = tensor("input_451_cast_fp16")]; - tensor module_layers_8_self_attn_linear_out_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159526144))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160312640))), name = tensor("module_layers_8_self_attn_linear_out_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_79_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_8_self_attn_linear_out_weight_to_fp16_palettized, x = input_451_cast_fp16)[name = tensor("linear_79_cast_fp16")]; - tensor input_455_cast_fp16 = add(x = input_447_cast_fp16, y = linear_79_cast_fp16)[name = tensor("input_455_cast_fp16")]; - tensor x_193_axes_0 = const()[name = tensor("x_193_axes_0"), val = tensor([-1])]; - tensor module_layers_8_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_8_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160312832)))]; - tensor module_layers_8_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_8_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160314944)))]; - tensor x_193_cast_fp16 = layer_norm(axes = x_193_axes_0, beta = module_layers_8_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_8_norm_conv_weight_to_fp16, x = input_455_cast_fp16)[name = tensor("x_193_cast_fp16")]; - tensor input_457_perm_0 = const()[name = tensor("input_457_perm_0"), val = tensor([0, 2, 1])]; - tensor input_459_pad_type_0 = const()[name = tensor("input_459_pad_type_0"), val = tensor("valid")]; - tensor input_459_strides_0 = const()[name = tensor("input_459_strides_0"), val = tensor([1])]; - tensor input_459_pad_0 = const()[name = tensor("input_459_pad_0"), val = tensor([0, 0])]; - tensor input_459_dilations_0 = const()[name = tensor("input_459_dilations_0"), val = tensor([1])]; - tensor input_459_groups_0 = const()[name = tensor("input_459_groups_0"), val = tensor(1)]; - tensor module_layers_8_conv_pointwise_conv1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160317056))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161889984))), name = tensor("module_layers_8_conv_pointwise_conv1_weight_to_fp16_palettized"), shape = tensor([2048, 1024, 1])]; - tensor input_457_cast_fp16 = transpose(perm = input_457_perm_0, x = x_193_cast_fp16)[name = tensor("transpose_251")]; - tensor input_459_cast_fp16 = conv(dilations = input_459_dilations_0, groups = input_459_groups_0, pad = input_459_pad_0, pad_type = input_459_pad_type_0, strides = input_459_strides_0, weight = module_layers_8_conv_pointwise_conv1_weight_to_fp16_palettized, x = input_457_cast_fp16)[name = tensor("input_459_cast_fp16")]; - tensor x_195_split_num_splits_0 = const()[name = tensor("x_195_split_num_splits_0"), val = tensor(2)]; - tensor x_195_split_axis_0 = const()[name = tensor("x_195_split_axis_0"), val = tensor(1)]; - tensor x_195_split_cast_fp16_0, tensor x_195_split_cast_fp16_1 = split(axis = x_195_split_axis_0, num_splits = x_195_split_num_splits_0, x = input_459_cast_fp16)[name = tensor("x_195_split_cast_fp16")]; - tensor x_195_split_1_sigmoid_cast_fp16 = sigmoid(x = x_195_split_cast_fp16_1)[name = tensor("x_195_split_1_sigmoid_cast_fp16")]; - tensor x_195_cast_fp16 = mul(x = x_195_split_cast_fp16_0, y = x_195_split_1_sigmoid_cast_fp16)[name = tensor("x_195_cast_fp16")]; - tensor input_461_cast_fp16 = select(a = var_11_to_fp16, b = x_195_cast_fp16, cond = var_328)[name = tensor("input_461_cast_fp16")]; - tensor input_463_pad_0 = const()[name = tensor("input_463_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_463_mode_0 = const()[name = tensor("input_463_mode_0"), val = tensor("constant")]; - tensor const_97_to_fp16 = const()[name = tensor("const_97_to_fp16"), val = tensor(0x0p+0)]; - tensor input_463_cast_fp16 = pad(constant_val = const_97_to_fp16, mode = input_463_mode_0, pad = input_463_pad_0, x = input_461_cast_fp16)[name = tensor("input_463_cast_fp16")]; - tensor input_465_pad_type_0 = const()[name = tensor("input_465_pad_type_0"), val = tensor("valid")]; - tensor input_465_groups_0 = const()[name = tensor("input_465_groups_0"), val = tensor(1024)]; - tensor input_465_strides_0 = const()[name = tensor("input_465_strides_0"), val = tensor([1])]; - tensor input_465_pad_0 = const()[name = tensor("input_465_pad_0"), val = tensor([0, 0])]; - tensor input_465_dilations_0 = const()[name = tensor("input_465_dilations_0"), val = tensor([1])]; - tensor const_264_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161890176))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161897152))), name = tensor("const_264_to_fp16_palettized"), shape = tensor([1024, 1, 9])]; - tensor const_265_to_fp16 = const()[name = tensor("const_265_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161897344)))]; - tensor input_467_cast_fp16 = conv(bias = const_265_to_fp16, dilations = input_465_dilations_0, groups = input_465_groups_0, pad = input_465_pad_0, pad_type = input_465_pad_type_0, strides = input_465_strides_0, weight = const_264_to_fp16_palettized, x = input_463_cast_fp16)[name = tensor("input_467_cast_fp16")]; - tensor input_469_cast_fp16 = silu(x = input_467_cast_fp16)[name = tensor("input_469_cast_fp16")]; - tensor x_197_pad_type_0 = const()[name = tensor("x_197_pad_type_0"), val = tensor("valid")]; - tensor x_197_strides_0 = const()[name = tensor("x_197_strides_0"), val = tensor([1])]; - tensor x_197_pad_0 = const()[name = tensor("x_197_pad_0"), val = tensor([0, 0])]; - tensor x_197_dilations_0 = const()[name = tensor("x_197_dilations_0"), val = tensor([1])]; - tensor x_197_groups_0 = const()[name = tensor("x_197_groups_0"), val = tensor(1)]; - tensor module_layers_8_conv_pointwise_conv2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161899456))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(162685952))), name = tensor("module_layers_8_conv_pointwise_conv2_weight_to_fp16_palettized"), shape = tensor([1024, 1024, 1])]; - tensor x_197_cast_fp16 = conv(dilations = x_197_dilations_0, groups = x_197_groups_0, pad = x_197_pad_0, pad_type = x_197_pad_type_0, strides = x_197_strides_0, weight = module_layers_8_conv_pointwise_conv2_weight_to_fp16_palettized, x = input_469_cast_fp16)[name = tensor("x_197_cast_fp16")]; - tensor input_471_perm_0 = const()[name = tensor("input_471_perm_0"), val = tensor([0, 2, 1])]; - tensor input_471_cast_fp16 = transpose(perm = input_471_perm_0, x = x_197_cast_fp16)[name = tensor("transpose_250")]; - tensor input_473_cast_fp16 = add(x = input_455_cast_fp16, y = input_471_cast_fp16)[name = tensor("input_473_cast_fp16")]; - tensor input_475_axes_0 = const()[name = tensor("input_475_axes_0"), val = tensor([-1])]; - tensor module_layers_8_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_8_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(162686144)))]; - tensor module_layers_8_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_8_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(162688256)))]; - tensor input_475_cast_fp16 = layer_norm(axes = input_475_axes_0, beta = module_layers_8_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_8_norm_feed_forward2_weight_to_fp16, x = input_473_cast_fp16)[name = tensor("input_475_cast_fp16")]; - tensor module_layers_8_feed_forward2_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(162690368))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165836160))), name = tensor("module_layers_8_feed_forward2_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_80_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_8_feed_forward2_linear1_weight_to_fp16_palettized, x = input_475_cast_fp16)[name = tensor("linear_80_cast_fp16")]; - tensor input_479_cast_fp16 = silu(x = linear_80_cast_fp16)[name = tensor("input_479_cast_fp16")]; - tensor module_layers_8_feed_forward2_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165836352))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(168982144))), name = tensor("module_layers_8_feed_forward2_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_81_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_8_feed_forward2_linear2_weight_to_fp16_palettized, x = input_479_cast_fp16)[name = tensor("linear_81_cast_fp16")]; - tensor var_1670_to_fp16 = const()[name = tensor("op_1670_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1671_cast_fp16 = mul(x = linear_81_cast_fp16, y = var_1670_to_fp16)[name = tensor("op_1671_cast_fp16")]; - tensor input_485_cast_fp16 = add(x = input_473_cast_fp16, y = var_1671_cast_fp16)[name = tensor("input_485_cast_fp16")]; - tensor input_487_axes_0 = const()[name = tensor("input_487_axes_0"), val = tensor([-1])]; - tensor module_layers_8_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_8_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(168982336)))]; - tensor module_layers_8_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_8_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(168984448)))]; - tensor input_487_cast_fp16 = layer_norm(axes = input_487_axes_0, beta = module_layers_8_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_8_norm_out_weight_to_fp16, x = input_485_cast_fp16)[name = tensor("input_487_cast_fp16")]; - tensor input_489_axes_0 = const()[name = tensor("input_489_axes_0"), val = tensor([-1])]; - tensor module_layers_9_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_9_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(168986560)))]; - tensor module_layers_9_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_9_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(168988672)))]; - tensor input_489_cast_fp16 = layer_norm(axes = input_489_axes_0, beta = module_layers_9_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_9_norm_feed_forward1_weight_to_fp16, x = input_487_cast_fp16)[name = tensor("input_489_cast_fp16")]; - tensor module_layers_9_feed_forward1_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(168990784))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(172136576))), name = tensor("module_layers_9_feed_forward1_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_82_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_9_feed_forward1_linear1_weight_to_fp16_palettized, x = input_489_cast_fp16)[name = tensor("linear_82_cast_fp16")]; - tensor input_493_cast_fp16 = silu(x = linear_82_cast_fp16)[name = tensor("input_493_cast_fp16")]; - tensor module_layers_9_feed_forward1_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(172136768))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(175282560))), name = tensor("module_layers_9_feed_forward1_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_83_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_9_feed_forward1_linear2_weight_to_fp16_palettized, x = input_493_cast_fp16)[name = tensor("linear_83_cast_fp16")]; - tensor var_1699_to_fp16 = const()[name = tensor("op_1699_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1700_cast_fp16 = mul(x = linear_83_cast_fp16, y = var_1699_to_fp16)[name = tensor("op_1700_cast_fp16")]; - tensor input_499_cast_fp16 = add(x = input_487_cast_fp16, y = var_1700_cast_fp16)[name = tensor("input_499_cast_fp16")]; - tensor query_19_axes_0 = const()[name = tensor("query_19_axes_0"), val = tensor([-1])]; - tensor module_layers_9_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_9_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(175282752)))]; - tensor module_layers_9_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_9_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(175284864)))]; - tensor query_19_cast_fp16 = layer_norm(axes = query_19_axes_0, beta = module_layers_9_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_9_norm_self_att_weight_to_fp16, x = input_499_cast_fp16)[name = tensor("query_19_cast_fp16")]; - tensor module_layers_9_self_attn_linear_q_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(175286976))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(176073472))), name = tensor("module_layers_9_self_attn_linear_q_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_84_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_9_self_attn_linear_q_weight_to_fp16_palettized, x = query_19_cast_fp16)[name = tensor("linear_84_cast_fp16")]; - tensor var_1716 = const()[name = tensor("op_1716"), val = tensor([1, -1, 8, 128])]; - tensor q_55_cast_fp16 = reshape(shape = var_1716, x = linear_84_cast_fp16)[name = tensor("q_55_cast_fp16")]; - tensor module_layers_9_self_attn_linear_k_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(176073664))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(176860160))), name = tensor("module_layers_9_self_attn_linear_k_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_85_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_9_self_attn_linear_k_weight_to_fp16_palettized, x = query_19_cast_fp16)[name = tensor("linear_85_cast_fp16")]; - tensor var_1720 = const()[name = tensor("op_1720"), val = tensor([1, -1, 8, 128])]; - tensor k_37_cast_fp16 = reshape(shape = var_1720, x = linear_85_cast_fp16)[name = tensor("k_37_cast_fp16")]; - tensor module_layers_9_self_attn_linear_v_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(176860352))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(177646848))), name = tensor("module_layers_9_self_attn_linear_v_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_86_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_9_self_attn_linear_v_weight_to_fp16_palettized, x = query_19_cast_fp16)[name = tensor("linear_86_cast_fp16")]; - tensor var_1724 = const()[name = tensor("op_1724"), val = tensor([1, -1, 8, 128])]; - tensor v_19_cast_fp16 = reshape(shape = var_1724, x = linear_86_cast_fp16)[name = tensor("v_19_cast_fp16")]; - tensor value_21_perm_0 = const()[name = tensor("value_21_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor module_layers_9_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_9_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(177647040)))]; - tensor var_1736_cast_fp16 = add(x = q_55_cast_fp16, y = module_layers_9_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1736_cast_fp16")]; - tensor module_layers_9_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_9_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(177649152)))]; - tensor var_1738_cast_fp16 = add(x = q_55_cast_fp16, y = module_layers_9_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1738_cast_fp16")]; - tensor q_with_bias_v_19_perm_0 = const()[name = tensor("q_with_bias_v_19_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor x_205_transpose_x_0 = const()[name = tensor("x_205_transpose_x_0"), val = tensor(false)]; - tensor x_205_transpose_y_0 = const()[name = tensor("x_205_transpose_y_0"), val = tensor(false)]; - tensor op_1740_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(177651264))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(177939328))), name = tensor("op_1740_to_fp16_palettized"), shape = tensor([1, 8, 128, 375])]; - tensor q_with_bias_v_19_cast_fp16 = transpose(perm = q_with_bias_v_19_perm_0, x = var_1738_cast_fp16)[name = tensor("transpose_249")]; - tensor x_205_cast_fp16 = matmul(transpose_x = x_205_transpose_x_0, transpose_y = x_205_transpose_y_0, x = q_with_bias_v_19_cast_fp16, y = op_1740_to_fp16_palettized)[name = tensor("x_205_cast_fp16")]; - tensor x_207_pad_0 = const()[name = tensor("x_207_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_207_mode_0 = const()[name = tensor("x_207_mode_0"), val = tensor("constant")]; - tensor const_104_to_fp16 = const()[name = tensor("const_104_to_fp16"), val = tensor(0x0p+0)]; - tensor x_207_cast_fp16 = pad(constant_val = const_104_to_fp16, mode = x_207_mode_0, pad = x_207_pad_0, x = x_205_cast_fp16)[name = tensor("x_207_cast_fp16")]; - tensor var_1748 = const()[name = tensor("op_1748"), val = tensor([1, 8, -1, 188])]; - tensor x_209_cast_fp16 = reshape(shape = var_1748, x = x_207_cast_fp16)[name = tensor("x_209_cast_fp16")]; - tensor var_1752_begin_0 = const()[name = tensor("op_1752_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_1752_end_0 = const()[name = tensor("op_1752_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_1752_end_mask_0 = const()[name = tensor("op_1752_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_1752_cast_fp16 = slice_by_index(begin = var_1752_begin_0, end = var_1752_end_0, end_mask = var_1752_end_mask_0, x = x_209_cast_fp16)[name = tensor("op_1752_cast_fp16")]; - tensor var_1753 = const()[name = tensor("op_1753"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_37_cast_fp16 = reshape(shape = var_1753, x = var_1752_cast_fp16)[name = tensor("matrix_bd_37_cast_fp16")]; - tensor matrix_ac_19_transpose_x_0 = const()[name = tensor("matrix_ac_19_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_19_transpose_y_0 = const()[name = tensor("matrix_ac_19_transpose_y_0"), val = tensor(false)]; - tensor transpose_114_perm_0 = const()[name = tensor("transpose_114_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_115_perm_0 = const()[name = tensor("transpose_115_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_115 = transpose(perm = transpose_115_perm_0, x = k_37_cast_fp16)[name = tensor("transpose_247")]; - tensor transpose_114 = transpose(perm = transpose_114_perm_0, x = var_1736_cast_fp16)[name = tensor("transpose_248")]; - tensor matrix_ac_19_cast_fp16 = matmul(transpose_x = matrix_ac_19_transpose_x_0, transpose_y = matrix_ac_19_transpose_y_0, x = transpose_114, y = transpose_115)[name = tensor("matrix_ac_19_cast_fp16")]; - tensor matrix_bd_39_begin_0 = const()[name = tensor("matrix_bd_39_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_39_end_0 = const()[name = tensor("matrix_bd_39_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_39_end_mask_0 = const()[name = tensor("matrix_bd_39_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_39_cast_fp16 = slice_by_index(begin = matrix_bd_39_begin_0, end = matrix_bd_39_end_0, end_mask = matrix_bd_39_end_mask_0, x = matrix_bd_37_cast_fp16)[name = tensor("matrix_bd_39_cast_fp16")]; - tensor var_1762_cast_fp16 = add(x = matrix_ac_19_cast_fp16, y = matrix_bd_39_cast_fp16)[name = tensor("op_1762_cast_fp16")]; - tensor _inversed_scores_37_y_0_to_fp16 = const()[name = tensor("_inversed_scores_37_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; - tensor _inversed_scores_37_cast_fp16 = mul(x = var_1762_cast_fp16, y = _inversed_scores_37_y_0_to_fp16)[name = tensor("_inversed_scores_37_cast_fp16")]; - tensor scores_39_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_37_cast_fp16, cond = mask_3)[name = tensor("scores_39_cast_fp16")]; - tensor var_1768_cast_fp16 = softmax(axis = var_30, x = scores_39_cast_fp16)[name = tensor("op_1768_cast_fp16")]; - tensor input_501_cast_fp16 = select(a = var_11_to_fp16, b = var_1768_cast_fp16, cond = mask_3)[name = tensor("input_501_cast_fp16")]; - tensor x_211_transpose_x_0 = const()[name = tensor("x_211_transpose_x_0"), val = tensor(false)]; - tensor x_211_transpose_y_0 = const()[name = tensor("x_211_transpose_y_0"), val = tensor(false)]; - tensor value_21_cast_fp16 = transpose(perm = value_21_perm_0, x = v_19_cast_fp16)[name = tensor("transpose_246")]; - tensor x_211_cast_fp16 = matmul(transpose_x = x_211_transpose_x_0, transpose_y = x_211_transpose_y_0, x = input_501_cast_fp16, y = value_21_cast_fp16)[name = tensor("x_211_cast_fp16")]; - tensor var_1772_perm_0 = const()[name = tensor("op_1772_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_1773 = const()[name = tensor("op_1773"), val = tensor([1, -1, 1024])]; - tensor var_1772_cast_fp16 = transpose(perm = var_1772_perm_0, x = x_211_cast_fp16)[name = tensor("transpose_245")]; - tensor input_503_cast_fp16 = reshape(shape = var_1773, x = var_1772_cast_fp16)[name = tensor("input_503_cast_fp16")]; - tensor module_layers_9_self_attn_linear_out_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(177939520))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(178726016))), name = tensor("module_layers_9_self_attn_linear_out_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_88_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_9_self_attn_linear_out_weight_to_fp16_palettized, x = input_503_cast_fp16)[name = tensor("linear_88_cast_fp16")]; - tensor input_507_cast_fp16 = add(x = input_499_cast_fp16, y = linear_88_cast_fp16)[name = tensor("input_507_cast_fp16")]; - tensor x_215_axes_0 = const()[name = tensor("x_215_axes_0"), val = tensor([-1])]; - tensor module_layers_9_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_9_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(178726208)))]; - tensor module_layers_9_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_9_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(178728320)))]; - tensor x_215_cast_fp16 = layer_norm(axes = x_215_axes_0, beta = module_layers_9_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_9_norm_conv_weight_to_fp16, x = input_507_cast_fp16)[name = tensor("x_215_cast_fp16")]; - tensor input_509_perm_0 = const()[name = tensor("input_509_perm_0"), val = tensor([0, 2, 1])]; - tensor input_511_pad_type_0 = const()[name = tensor("input_511_pad_type_0"), val = tensor("valid")]; - tensor input_511_strides_0 = const()[name = tensor("input_511_strides_0"), val = tensor([1])]; - tensor input_511_pad_0 = const()[name = tensor("input_511_pad_0"), val = tensor([0, 0])]; - tensor input_511_dilations_0 = const()[name = tensor("input_511_dilations_0"), val = tensor([1])]; - tensor input_511_groups_0 = const()[name = tensor("input_511_groups_0"), val = tensor(1)]; - tensor module_layers_9_conv_pointwise_conv1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(178730432))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(180303360))), name = tensor("module_layers_9_conv_pointwise_conv1_weight_to_fp16_palettized"), shape = tensor([2048, 1024, 1])]; - tensor input_509_cast_fp16 = transpose(perm = input_509_perm_0, x = x_215_cast_fp16)[name = tensor("transpose_244")]; - tensor input_511_cast_fp16 = conv(dilations = input_511_dilations_0, groups = input_511_groups_0, pad = input_511_pad_0, pad_type = input_511_pad_type_0, strides = input_511_strides_0, weight = module_layers_9_conv_pointwise_conv1_weight_to_fp16_palettized, x = input_509_cast_fp16)[name = tensor("input_511_cast_fp16")]; - tensor x_217_split_num_splits_0 = const()[name = tensor("x_217_split_num_splits_0"), val = tensor(2)]; - tensor x_217_split_axis_0 = const()[name = tensor("x_217_split_axis_0"), val = tensor(1)]; - tensor x_217_split_cast_fp16_0, tensor x_217_split_cast_fp16_1 = split(axis = x_217_split_axis_0, num_splits = x_217_split_num_splits_0, x = input_511_cast_fp16)[name = tensor("x_217_split_cast_fp16")]; - tensor x_217_split_1_sigmoid_cast_fp16 = sigmoid(x = x_217_split_cast_fp16_1)[name = tensor("x_217_split_1_sigmoid_cast_fp16")]; - tensor x_217_cast_fp16 = mul(x = x_217_split_cast_fp16_0, y = x_217_split_1_sigmoid_cast_fp16)[name = tensor("x_217_cast_fp16")]; - tensor input_513_cast_fp16 = select(a = var_11_to_fp16, b = x_217_cast_fp16, cond = var_328)[name = tensor("input_513_cast_fp16")]; - tensor input_515_pad_0 = const()[name = tensor("input_515_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_515_mode_0 = const()[name = tensor("input_515_mode_0"), val = tensor("constant")]; - tensor const_107_to_fp16 = const()[name = tensor("const_107_to_fp16"), val = tensor(0x0p+0)]; - tensor input_515_cast_fp16 = pad(constant_val = const_107_to_fp16, mode = input_515_mode_0, pad = input_515_pad_0, x = input_513_cast_fp16)[name = tensor("input_515_cast_fp16")]; - tensor input_517_pad_type_0 = const()[name = tensor("input_517_pad_type_0"), val = tensor("valid")]; - tensor input_517_groups_0 = const()[name = tensor("input_517_groups_0"), val = tensor(1024)]; - tensor input_517_strides_0 = const()[name = tensor("input_517_strides_0"), val = tensor([1])]; - tensor input_517_pad_0 = const()[name = tensor("input_517_pad_0"), val = tensor([0, 0])]; - tensor input_517_dilations_0 = const()[name = tensor("input_517_dilations_0"), val = tensor([1])]; - tensor const_266_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(180303552))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(180310528))), name = tensor("const_266_to_fp16_palettized"), shape = tensor([1024, 1, 9])]; - tensor const_267_to_fp16 = const()[name = tensor("const_267_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(180310720)))]; - tensor input_519_cast_fp16 = conv(bias = const_267_to_fp16, dilations = input_517_dilations_0, groups = input_517_groups_0, pad = input_517_pad_0, pad_type = input_517_pad_type_0, strides = input_517_strides_0, weight = const_266_to_fp16_palettized, x = input_515_cast_fp16)[name = tensor("input_519_cast_fp16")]; - tensor input_521_cast_fp16 = silu(x = input_519_cast_fp16)[name = tensor("input_521_cast_fp16")]; - tensor x_219_pad_type_0 = const()[name = tensor("x_219_pad_type_0"), val = tensor("valid")]; - tensor x_219_strides_0 = const()[name = tensor("x_219_strides_0"), val = tensor([1])]; - tensor x_219_pad_0 = const()[name = tensor("x_219_pad_0"), val = tensor([0, 0])]; - tensor x_219_dilations_0 = const()[name = tensor("x_219_dilations_0"), val = tensor([1])]; - tensor x_219_groups_0 = const()[name = tensor("x_219_groups_0"), val = tensor(1)]; - tensor module_layers_9_conv_pointwise_conv2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(180312832))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(181099328))), name = tensor("module_layers_9_conv_pointwise_conv2_weight_to_fp16_palettized"), shape = tensor([1024, 1024, 1])]; - tensor x_219_cast_fp16 = conv(dilations = x_219_dilations_0, groups = x_219_groups_0, pad = x_219_pad_0, pad_type = x_219_pad_type_0, strides = x_219_strides_0, weight = module_layers_9_conv_pointwise_conv2_weight_to_fp16_palettized, x = input_521_cast_fp16)[name = tensor("x_219_cast_fp16")]; - tensor input_523_perm_0 = const()[name = tensor("input_523_perm_0"), val = tensor([0, 2, 1])]; - tensor input_523_cast_fp16 = transpose(perm = input_523_perm_0, x = x_219_cast_fp16)[name = tensor("transpose_243")]; - tensor input_525_cast_fp16 = add(x = input_507_cast_fp16, y = input_523_cast_fp16)[name = tensor("input_525_cast_fp16")]; - tensor input_527_axes_0 = const()[name = tensor("input_527_axes_0"), val = tensor([-1])]; - tensor module_layers_9_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_9_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(181099520)))]; - tensor module_layers_9_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_9_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(181101632)))]; - tensor input_527_cast_fp16 = layer_norm(axes = input_527_axes_0, beta = module_layers_9_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_9_norm_feed_forward2_weight_to_fp16, x = input_525_cast_fp16)[name = tensor("input_527_cast_fp16")]; - tensor module_layers_9_feed_forward2_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(181103744))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184249536))), name = tensor("module_layers_9_feed_forward2_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_89_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_9_feed_forward2_linear1_weight_to_fp16_palettized, x = input_527_cast_fp16)[name = tensor("linear_89_cast_fp16")]; - tensor input_531_cast_fp16 = silu(x = linear_89_cast_fp16)[name = tensor("input_531_cast_fp16")]; - tensor module_layers_9_feed_forward2_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184249728))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(187395520))), name = tensor("module_layers_9_feed_forward2_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_90_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_9_feed_forward2_linear2_weight_to_fp16_palettized, x = input_531_cast_fp16)[name = tensor("linear_90_cast_fp16")]; - tensor var_1833_to_fp16 = const()[name = tensor("op_1833_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1834_cast_fp16 = mul(x = linear_90_cast_fp16, y = var_1833_to_fp16)[name = tensor("op_1834_cast_fp16")]; - tensor input_537_cast_fp16 = add(x = input_525_cast_fp16, y = var_1834_cast_fp16)[name = tensor("input_537_cast_fp16")]; - tensor input_539_axes_0 = const()[name = tensor("input_539_axes_0"), val = tensor([-1])]; - tensor module_layers_9_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_9_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(187395712)))]; - tensor module_layers_9_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_9_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(187397824)))]; - tensor input_539_cast_fp16 = layer_norm(axes = input_539_axes_0, beta = module_layers_9_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_9_norm_out_weight_to_fp16, x = input_537_cast_fp16)[name = tensor("input_539_cast_fp16")]; - tensor input_541_axes_0 = const()[name = tensor("input_541_axes_0"), val = tensor([-1])]; - tensor module_layers_10_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_10_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(187399936)))]; - tensor module_layers_10_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_10_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(187402048)))]; - tensor input_541_cast_fp16 = layer_norm(axes = input_541_axes_0, beta = module_layers_10_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_10_norm_feed_forward1_weight_to_fp16, x = input_539_cast_fp16)[name = tensor("input_541_cast_fp16")]; - tensor module_layers_10_feed_forward1_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(187404160))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190549952))), name = tensor("module_layers_10_feed_forward1_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_91_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_10_feed_forward1_linear1_weight_to_fp16_palettized, x = input_541_cast_fp16)[name = tensor("linear_91_cast_fp16")]; - tensor input_545_cast_fp16 = silu(x = linear_91_cast_fp16)[name = tensor("input_545_cast_fp16")]; - tensor module_layers_10_feed_forward1_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190550144))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(193695936))), name = tensor("module_layers_10_feed_forward1_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_92_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_10_feed_forward1_linear2_weight_to_fp16_palettized, x = input_545_cast_fp16)[name = tensor("linear_92_cast_fp16")]; - tensor var_1862_to_fp16 = const()[name = tensor("op_1862_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1863_cast_fp16 = mul(x = linear_92_cast_fp16, y = var_1862_to_fp16)[name = tensor("op_1863_cast_fp16")]; - tensor input_551_cast_fp16 = add(x = input_539_cast_fp16, y = var_1863_cast_fp16)[name = tensor("input_551_cast_fp16")]; - tensor query_21_axes_0 = const()[name = tensor("query_21_axes_0"), val = tensor([-1])]; - tensor module_layers_10_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_10_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(193696128)))]; - tensor module_layers_10_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_10_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(193698240)))]; - tensor query_21_cast_fp16 = layer_norm(axes = query_21_axes_0, beta = module_layers_10_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_10_norm_self_att_weight_to_fp16, x = input_551_cast_fp16)[name = tensor("query_21_cast_fp16")]; - tensor module_layers_10_self_attn_linear_q_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(193700352))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194486848))), name = tensor("module_layers_10_self_attn_linear_q_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_93_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_10_self_attn_linear_q_weight_to_fp16_palettized, x = query_21_cast_fp16)[name = tensor("linear_93_cast_fp16")]; - tensor var_1879 = const()[name = tensor("op_1879"), val = tensor([1, -1, 8, 128])]; - tensor q_61_cast_fp16 = reshape(shape = var_1879, x = linear_93_cast_fp16)[name = tensor("q_61_cast_fp16")]; - tensor module_layers_10_self_attn_linear_k_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194487040))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195273536))), name = tensor("module_layers_10_self_attn_linear_k_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_94_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_10_self_attn_linear_k_weight_to_fp16_palettized, x = query_21_cast_fp16)[name = tensor("linear_94_cast_fp16")]; - tensor var_1883 = const()[name = tensor("op_1883"), val = tensor([1, -1, 8, 128])]; - tensor k_41_cast_fp16 = reshape(shape = var_1883, x = linear_94_cast_fp16)[name = tensor("k_41_cast_fp16")]; - tensor module_layers_10_self_attn_linear_v_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195273728))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196060224))), name = tensor("module_layers_10_self_attn_linear_v_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_95_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_10_self_attn_linear_v_weight_to_fp16_palettized, x = query_21_cast_fp16)[name = tensor("linear_95_cast_fp16")]; - tensor var_1887 = const()[name = tensor("op_1887"), val = tensor([1, -1, 8, 128])]; - tensor v_21_cast_fp16 = reshape(shape = var_1887, x = linear_95_cast_fp16)[name = tensor("v_21_cast_fp16")]; - tensor value_23_perm_0 = const()[name = tensor("value_23_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor module_layers_10_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_10_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196060416)))]; - tensor var_1899_cast_fp16 = add(x = q_61_cast_fp16, y = module_layers_10_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1899_cast_fp16")]; - tensor module_layers_10_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_10_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196062528)))]; - tensor var_1901_cast_fp16 = add(x = q_61_cast_fp16, y = module_layers_10_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1901_cast_fp16")]; - tensor q_with_bias_v_21_perm_0 = const()[name = tensor("q_with_bias_v_21_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor x_227_transpose_x_0 = const()[name = tensor("x_227_transpose_x_0"), val = tensor(false)]; - tensor x_227_transpose_y_0 = const()[name = tensor("x_227_transpose_y_0"), val = tensor(false)]; - tensor op_1903_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196064640))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196352704))), name = tensor("op_1903_to_fp16_palettized"), shape = tensor([1, 8, 128, 375])]; - tensor q_with_bias_v_21_cast_fp16 = transpose(perm = q_with_bias_v_21_perm_0, x = var_1901_cast_fp16)[name = tensor("transpose_242")]; - tensor x_227_cast_fp16 = matmul(transpose_x = x_227_transpose_x_0, transpose_y = x_227_transpose_y_0, x = q_with_bias_v_21_cast_fp16, y = op_1903_to_fp16_palettized)[name = tensor("x_227_cast_fp16")]; - tensor x_229_pad_0 = const()[name = tensor("x_229_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_229_mode_0 = const()[name = tensor("x_229_mode_0"), val = tensor("constant")]; - tensor const_114_to_fp16 = const()[name = tensor("const_114_to_fp16"), val = tensor(0x0p+0)]; - tensor x_229_cast_fp16 = pad(constant_val = const_114_to_fp16, mode = x_229_mode_0, pad = x_229_pad_0, x = x_227_cast_fp16)[name = tensor("x_229_cast_fp16")]; - tensor var_1911 = const()[name = tensor("op_1911"), val = tensor([1, 8, -1, 188])]; - tensor x_231_cast_fp16 = reshape(shape = var_1911, x = x_229_cast_fp16)[name = tensor("x_231_cast_fp16")]; - tensor var_1915_begin_0 = const()[name = tensor("op_1915_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_1915_end_0 = const()[name = tensor("op_1915_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_1915_end_mask_0 = const()[name = tensor("op_1915_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_1915_cast_fp16 = slice_by_index(begin = var_1915_begin_0, end = var_1915_end_0, end_mask = var_1915_end_mask_0, x = x_231_cast_fp16)[name = tensor("op_1915_cast_fp16")]; - tensor var_1916 = const()[name = tensor("op_1916"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_41_cast_fp16 = reshape(shape = var_1916, x = var_1915_cast_fp16)[name = tensor("matrix_bd_41_cast_fp16")]; - tensor matrix_ac_21_transpose_x_0 = const()[name = tensor("matrix_ac_21_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_21_transpose_y_0 = const()[name = tensor("matrix_ac_21_transpose_y_0"), val = tensor(false)]; - tensor transpose_116_perm_0 = const()[name = tensor("transpose_116_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_117_perm_0 = const()[name = tensor("transpose_117_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_117 = transpose(perm = transpose_117_perm_0, x = k_41_cast_fp16)[name = tensor("transpose_240")]; - tensor transpose_116 = transpose(perm = transpose_116_perm_0, x = var_1899_cast_fp16)[name = tensor("transpose_241")]; - tensor matrix_ac_21_cast_fp16 = matmul(transpose_x = matrix_ac_21_transpose_x_0, transpose_y = matrix_ac_21_transpose_y_0, x = transpose_116, y = transpose_117)[name = tensor("matrix_ac_21_cast_fp16")]; - tensor matrix_bd_43_begin_0 = const()[name = tensor("matrix_bd_43_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_43_end_0 = const()[name = tensor("matrix_bd_43_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_43_end_mask_0 = const()[name = tensor("matrix_bd_43_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_43_cast_fp16 = slice_by_index(begin = matrix_bd_43_begin_0, end = matrix_bd_43_end_0, end_mask = matrix_bd_43_end_mask_0, x = matrix_bd_41_cast_fp16)[name = tensor("matrix_bd_43_cast_fp16")]; - tensor var_1925_cast_fp16 = add(x = matrix_ac_21_cast_fp16, y = matrix_bd_43_cast_fp16)[name = tensor("op_1925_cast_fp16")]; - tensor _inversed_scores_41_y_0_to_fp16 = const()[name = tensor("_inversed_scores_41_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; - tensor _inversed_scores_41_cast_fp16 = mul(x = var_1925_cast_fp16, y = _inversed_scores_41_y_0_to_fp16)[name = tensor("_inversed_scores_41_cast_fp16")]; - tensor scores_43_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_41_cast_fp16, cond = mask_3)[name = tensor("scores_43_cast_fp16")]; - tensor var_1931_cast_fp16 = softmax(axis = var_30, x = scores_43_cast_fp16)[name = tensor("op_1931_cast_fp16")]; - tensor input_553_cast_fp16 = select(a = var_11_to_fp16, b = var_1931_cast_fp16, cond = mask_3)[name = tensor("input_553_cast_fp16")]; - tensor x_233_transpose_x_0 = const()[name = tensor("x_233_transpose_x_0"), val = tensor(false)]; - tensor x_233_transpose_y_0 = const()[name = tensor("x_233_transpose_y_0"), val = tensor(false)]; - tensor value_23_cast_fp16 = transpose(perm = value_23_perm_0, x = v_21_cast_fp16)[name = tensor("transpose_239")]; - tensor x_233_cast_fp16 = matmul(transpose_x = x_233_transpose_x_0, transpose_y = x_233_transpose_y_0, x = input_553_cast_fp16, y = value_23_cast_fp16)[name = tensor("x_233_cast_fp16")]; - tensor var_1935_perm_0 = const()[name = tensor("op_1935_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_1936 = const()[name = tensor("op_1936"), val = tensor([1, -1, 1024])]; - tensor var_1935_cast_fp16 = transpose(perm = var_1935_perm_0, x = x_233_cast_fp16)[name = tensor("transpose_238")]; - tensor input_555_cast_fp16 = reshape(shape = var_1936, x = var_1935_cast_fp16)[name = tensor("input_555_cast_fp16")]; - tensor module_layers_10_self_attn_linear_out_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196352896))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(197139392))), name = tensor("module_layers_10_self_attn_linear_out_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_97_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_10_self_attn_linear_out_weight_to_fp16_palettized, x = input_555_cast_fp16)[name = tensor("linear_97_cast_fp16")]; - tensor input_559_cast_fp16 = add(x = input_551_cast_fp16, y = linear_97_cast_fp16)[name = tensor("input_559_cast_fp16")]; - tensor x_237_axes_0 = const()[name = tensor("x_237_axes_0"), val = tensor([-1])]; - tensor module_layers_10_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_10_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(197139584)))]; - tensor module_layers_10_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_10_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(197141696)))]; - tensor x_237_cast_fp16 = layer_norm(axes = x_237_axes_0, beta = module_layers_10_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_10_norm_conv_weight_to_fp16, x = input_559_cast_fp16)[name = tensor("x_237_cast_fp16")]; - tensor input_561_perm_0 = const()[name = tensor("input_561_perm_0"), val = tensor([0, 2, 1])]; - tensor input_563_pad_type_0 = const()[name = tensor("input_563_pad_type_0"), val = tensor("valid")]; - tensor input_563_strides_0 = const()[name = tensor("input_563_strides_0"), val = tensor([1])]; - tensor input_563_pad_0 = const()[name = tensor("input_563_pad_0"), val = tensor([0, 0])]; - tensor input_563_dilations_0 = const()[name = tensor("input_563_dilations_0"), val = tensor([1])]; - tensor input_563_groups_0 = const()[name = tensor("input_563_groups_0"), val = tensor(1)]; - tensor module_layers_10_conv_pointwise_conv1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(197143808))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198716736))), name = tensor("module_layers_10_conv_pointwise_conv1_weight_to_fp16_palettized"), shape = tensor([2048, 1024, 1])]; - tensor input_561_cast_fp16 = transpose(perm = input_561_perm_0, x = x_237_cast_fp16)[name = tensor("transpose_237")]; - tensor input_563_cast_fp16 = conv(dilations = input_563_dilations_0, groups = input_563_groups_0, pad = input_563_pad_0, pad_type = input_563_pad_type_0, strides = input_563_strides_0, weight = module_layers_10_conv_pointwise_conv1_weight_to_fp16_palettized, x = input_561_cast_fp16)[name = tensor("input_563_cast_fp16")]; - tensor x_239_split_num_splits_0 = const()[name = tensor("x_239_split_num_splits_0"), val = tensor(2)]; - tensor x_239_split_axis_0 = const()[name = tensor("x_239_split_axis_0"), val = tensor(1)]; - tensor x_239_split_cast_fp16_0, tensor x_239_split_cast_fp16_1 = split(axis = x_239_split_axis_0, num_splits = x_239_split_num_splits_0, x = input_563_cast_fp16)[name = tensor("x_239_split_cast_fp16")]; - tensor x_239_split_1_sigmoid_cast_fp16 = sigmoid(x = x_239_split_cast_fp16_1)[name = tensor("x_239_split_1_sigmoid_cast_fp16")]; - tensor x_239_cast_fp16 = mul(x = x_239_split_cast_fp16_0, y = x_239_split_1_sigmoid_cast_fp16)[name = tensor("x_239_cast_fp16")]; - tensor input_565_cast_fp16 = select(a = var_11_to_fp16, b = x_239_cast_fp16, cond = var_328)[name = tensor("input_565_cast_fp16")]; - tensor input_567_pad_0 = const()[name = tensor("input_567_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_567_mode_0 = const()[name = tensor("input_567_mode_0"), val = tensor("constant")]; - tensor const_117_to_fp16 = const()[name = tensor("const_117_to_fp16"), val = tensor(0x0p+0)]; - tensor input_567_cast_fp16 = pad(constant_val = const_117_to_fp16, mode = input_567_mode_0, pad = input_567_pad_0, x = input_565_cast_fp16)[name = tensor("input_567_cast_fp16")]; - tensor input_569_pad_type_0 = const()[name = tensor("input_569_pad_type_0"), val = tensor("valid")]; - tensor input_569_groups_0 = const()[name = tensor("input_569_groups_0"), val = tensor(1024)]; - tensor input_569_strides_0 = const()[name = tensor("input_569_strides_0"), val = tensor([1])]; - tensor input_569_pad_0 = const()[name = tensor("input_569_pad_0"), val = tensor([0, 0])]; - tensor input_569_dilations_0 = const()[name = tensor("input_569_dilations_0"), val = tensor([1])]; - tensor const_268_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198716928))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198723904))), name = tensor("const_268_to_fp16_palettized"), shape = tensor([1024, 1, 9])]; - tensor const_269_to_fp16 = const()[name = tensor("const_269_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198724096)))]; - tensor input_571_cast_fp16 = conv(bias = const_269_to_fp16, dilations = input_569_dilations_0, groups = input_569_groups_0, pad = input_569_pad_0, pad_type = input_569_pad_type_0, strides = input_569_strides_0, weight = const_268_to_fp16_palettized, x = input_567_cast_fp16)[name = tensor("input_571_cast_fp16")]; - tensor input_573_cast_fp16 = silu(x = input_571_cast_fp16)[name = tensor("input_573_cast_fp16")]; - tensor x_241_pad_type_0 = const()[name = tensor("x_241_pad_type_0"), val = tensor("valid")]; - tensor x_241_strides_0 = const()[name = tensor("x_241_strides_0"), val = tensor([1])]; - tensor x_241_pad_0 = const()[name = tensor("x_241_pad_0"), val = tensor([0, 0])]; - tensor x_241_dilations_0 = const()[name = tensor("x_241_dilations_0"), val = tensor([1])]; - tensor x_241_groups_0 = const()[name = tensor("x_241_groups_0"), val = tensor(1)]; - tensor module_layers_10_conv_pointwise_conv2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198726208))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(199512704))), name = tensor("module_layers_10_conv_pointwise_conv2_weight_to_fp16_palettized"), shape = tensor([1024, 1024, 1])]; - tensor x_241_cast_fp16 = conv(dilations = x_241_dilations_0, groups = x_241_groups_0, pad = x_241_pad_0, pad_type = x_241_pad_type_0, strides = x_241_strides_0, weight = module_layers_10_conv_pointwise_conv2_weight_to_fp16_palettized, x = input_573_cast_fp16)[name = tensor("x_241_cast_fp16")]; - tensor input_575_perm_0 = const()[name = tensor("input_575_perm_0"), val = tensor([0, 2, 1])]; - tensor input_575_cast_fp16 = transpose(perm = input_575_perm_0, x = x_241_cast_fp16)[name = tensor("transpose_236")]; - tensor input_577_cast_fp16 = add(x = input_559_cast_fp16, y = input_575_cast_fp16)[name = tensor("input_577_cast_fp16")]; - tensor input_579_axes_0 = const()[name = tensor("input_579_axes_0"), val = tensor([-1])]; - tensor module_layers_10_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_10_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(199512896)))]; - tensor module_layers_10_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_10_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(199515008)))]; - tensor input_579_cast_fp16 = layer_norm(axes = input_579_axes_0, beta = module_layers_10_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_10_norm_feed_forward2_weight_to_fp16, x = input_577_cast_fp16)[name = tensor("input_579_cast_fp16")]; - tensor module_layers_10_feed_forward2_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(199517120))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202662912))), name = tensor("module_layers_10_feed_forward2_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_98_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_10_feed_forward2_linear1_weight_to_fp16_palettized, x = input_579_cast_fp16)[name = tensor("linear_98_cast_fp16")]; - tensor input_583_cast_fp16 = silu(x = linear_98_cast_fp16)[name = tensor("input_583_cast_fp16")]; - tensor module_layers_10_feed_forward2_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202663104))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(205808896))), name = tensor("module_layers_10_feed_forward2_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_99_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_10_feed_forward2_linear2_weight_to_fp16_palettized, x = input_583_cast_fp16)[name = tensor("linear_99_cast_fp16")]; - tensor var_1996_to_fp16 = const()[name = tensor("op_1996_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1997_cast_fp16 = mul(x = linear_99_cast_fp16, y = var_1996_to_fp16)[name = tensor("op_1997_cast_fp16")]; - tensor input_589_cast_fp16 = add(x = input_577_cast_fp16, y = var_1997_cast_fp16)[name = tensor("input_589_cast_fp16")]; - tensor input_591_axes_0 = const()[name = tensor("input_591_axes_0"), val = tensor([-1])]; - tensor module_layers_10_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_10_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(205809088)))]; - tensor module_layers_10_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_10_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(205811200)))]; - tensor input_591_cast_fp16 = layer_norm(axes = input_591_axes_0, beta = module_layers_10_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_10_norm_out_weight_to_fp16, x = input_589_cast_fp16)[name = tensor("input_591_cast_fp16")]; - tensor input_593_axes_0 = const()[name = tensor("input_593_axes_0"), val = tensor([-1])]; - tensor module_layers_11_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_11_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(205813312)))]; - tensor module_layers_11_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_11_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(205815424)))]; - tensor input_593_cast_fp16 = layer_norm(axes = input_593_axes_0, beta = module_layers_11_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_11_norm_feed_forward1_weight_to_fp16, x = input_591_cast_fp16)[name = tensor("input_593_cast_fp16")]; - tensor module_layers_11_feed_forward1_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(205817536))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208963328))), name = tensor("module_layers_11_feed_forward1_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_100_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_11_feed_forward1_linear1_weight_to_fp16_palettized, x = input_593_cast_fp16)[name = tensor("linear_100_cast_fp16")]; - tensor input_597_cast_fp16 = silu(x = linear_100_cast_fp16)[name = tensor("input_597_cast_fp16")]; - tensor module_layers_11_feed_forward1_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208963520))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(212109312))), name = tensor("module_layers_11_feed_forward1_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_101_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_11_feed_forward1_linear2_weight_to_fp16_palettized, x = input_597_cast_fp16)[name = tensor("linear_101_cast_fp16")]; - tensor var_2025_to_fp16 = const()[name = tensor("op_2025_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2026_cast_fp16 = mul(x = linear_101_cast_fp16, y = var_2025_to_fp16)[name = tensor("op_2026_cast_fp16")]; - tensor input_603_cast_fp16 = add(x = input_591_cast_fp16, y = var_2026_cast_fp16)[name = tensor("input_603_cast_fp16")]; - tensor query_23_axes_0 = const()[name = tensor("query_23_axes_0"), val = tensor([-1])]; - tensor module_layers_11_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_11_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(212109504)))]; - tensor module_layers_11_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_11_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(212111616)))]; - tensor query_23_cast_fp16 = layer_norm(axes = query_23_axes_0, beta = module_layers_11_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_11_norm_self_att_weight_to_fp16, x = input_603_cast_fp16)[name = tensor("query_23_cast_fp16")]; - tensor module_layers_11_self_attn_linear_q_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(212113728))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(212900224))), name = tensor("module_layers_11_self_attn_linear_q_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_102_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_11_self_attn_linear_q_weight_to_fp16_palettized, x = query_23_cast_fp16)[name = tensor("linear_102_cast_fp16")]; - tensor var_2042 = const()[name = tensor("op_2042"), val = tensor([1, -1, 8, 128])]; - tensor q_67_cast_fp16 = reshape(shape = var_2042, x = linear_102_cast_fp16)[name = tensor("q_67_cast_fp16")]; - tensor module_layers_11_self_attn_linear_k_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(212900416))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(213686912))), name = tensor("module_layers_11_self_attn_linear_k_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_103_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_11_self_attn_linear_k_weight_to_fp16_palettized, x = query_23_cast_fp16)[name = tensor("linear_103_cast_fp16")]; - tensor var_2046 = const()[name = tensor("op_2046"), val = tensor([1, -1, 8, 128])]; - tensor k_45_cast_fp16 = reshape(shape = var_2046, x = linear_103_cast_fp16)[name = tensor("k_45_cast_fp16")]; - tensor module_layers_11_self_attn_linear_v_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(213687104))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(214473600))), name = tensor("module_layers_11_self_attn_linear_v_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_104_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_11_self_attn_linear_v_weight_to_fp16_palettized, x = query_23_cast_fp16)[name = tensor("linear_104_cast_fp16")]; - tensor var_2050 = const()[name = tensor("op_2050"), val = tensor([1, -1, 8, 128])]; - tensor v_23_cast_fp16 = reshape(shape = var_2050, x = linear_104_cast_fp16)[name = tensor("v_23_cast_fp16")]; - tensor value_25_perm_0 = const()[name = tensor("value_25_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor module_layers_11_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_11_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(214473792)))]; - tensor var_2062_cast_fp16 = add(x = q_67_cast_fp16, y = module_layers_11_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2062_cast_fp16")]; - tensor module_layers_11_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_11_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(214475904)))]; - tensor var_2064_cast_fp16 = add(x = q_67_cast_fp16, y = module_layers_11_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2064_cast_fp16")]; - tensor q_with_bias_v_23_perm_0 = const()[name = tensor("q_with_bias_v_23_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor x_249_transpose_x_0 = const()[name = tensor("x_249_transpose_x_0"), val = tensor(false)]; - tensor x_249_transpose_y_0 = const()[name = tensor("x_249_transpose_y_0"), val = tensor(false)]; - tensor op_2066_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(214478016))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(214766080))), name = tensor("op_2066_to_fp16_palettized"), shape = tensor([1, 8, 128, 375])]; - tensor q_with_bias_v_23_cast_fp16 = transpose(perm = q_with_bias_v_23_perm_0, x = var_2064_cast_fp16)[name = tensor("transpose_235")]; - tensor x_249_cast_fp16 = matmul(transpose_x = x_249_transpose_x_0, transpose_y = x_249_transpose_y_0, x = q_with_bias_v_23_cast_fp16, y = op_2066_to_fp16_palettized)[name = tensor("x_249_cast_fp16")]; - tensor x_251_pad_0 = const()[name = tensor("x_251_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_251_mode_0 = const()[name = tensor("x_251_mode_0"), val = tensor("constant")]; - tensor const_124_to_fp16 = const()[name = tensor("const_124_to_fp16"), val = tensor(0x0p+0)]; - tensor x_251_cast_fp16 = pad(constant_val = const_124_to_fp16, mode = x_251_mode_0, pad = x_251_pad_0, x = x_249_cast_fp16)[name = tensor("x_251_cast_fp16")]; - tensor var_2074 = const()[name = tensor("op_2074"), val = tensor([1, 8, -1, 188])]; - tensor x_253_cast_fp16 = reshape(shape = var_2074, x = x_251_cast_fp16)[name = tensor("x_253_cast_fp16")]; - tensor var_2078_begin_0 = const()[name = tensor("op_2078_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_2078_end_0 = const()[name = tensor("op_2078_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_2078_end_mask_0 = const()[name = tensor("op_2078_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_2078_cast_fp16 = slice_by_index(begin = var_2078_begin_0, end = var_2078_end_0, end_mask = var_2078_end_mask_0, x = x_253_cast_fp16)[name = tensor("op_2078_cast_fp16")]; - tensor var_2079 = const()[name = tensor("op_2079"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_45_cast_fp16 = reshape(shape = var_2079, x = var_2078_cast_fp16)[name = tensor("matrix_bd_45_cast_fp16")]; - tensor matrix_ac_23_transpose_x_0 = const()[name = tensor("matrix_ac_23_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_23_transpose_y_0 = const()[name = tensor("matrix_ac_23_transpose_y_0"), val = tensor(false)]; - tensor transpose_118_perm_0 = const()[name = tensor("transpose_118_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_119_perm_0 = const()[name = tensor("transpose_119_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_119 = transpose(perm = transpose_119_perm_0, x = k_45_cast_fp16)[name = tensor("transpose_233")]; - tensor transpose_118 = transpose(perm = transpose_118_perm_0, x = var_2062_cast_fp16)[name = tensor("transpose_234")]; - tensor matrix_ac_23_cast_fp16 = matmul(transpose_x = matrix_ac_23_transpose_x_0, transpose_y = matrix_ac_23_transpose_y_0, x = transpose_118, y = transpose_119)[name = tensor("matrix_ac_23_cast_fp16")]; - tensor matrix_bd_47_begin_0 = const()[name = tensor("matrix_bd_47_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_47_end_0 = const()[name = tensor("matrix_bd_47_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_47_end_mask_0 = const()[name = tensor("matrix_bd_47_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_47_cast_fp16 = slice_by_index(begin = matrix_bd_47_begin_0, end = matrix_bd_47_end_0, end_mask = matrix_bd_47_end_mask_0, x = matrix_bd_45_cast_fp16)[name = tensor("matrix_bd_47_cast_fp16")]; - tensor var_2088_cast_fp16 = add(x = matrix_ac_23_cast_fp16, y = matrix_bd_47_cast_fp16)[name = tensor("op_2088_cast_fp16")]; - tensor _inversed_scores_45_y_0_to_fp16 = const()[name = tensor("_inversed_scores_45_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; - tensor _inversed_scores_45_cast_fp16 = mul(x = var_2088_cast_fp16, y = _inversed_scores_45_y_0_to_fp16)[name = tensor("_inversed_scores_45_cast_fp16")]; - tensor scores_47_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_45_cast_fp16, cond = mask_3)[name = tensor("scores_47_cast_fp16")]; - tensor var_2094_cast_fp16 = softmax(axis = var_30, x = scores_47_cast_fp16)[name = tensor("op_2094_cast_fp16")]; - tensor input_605_cast_fp16 = select(a = var_11_to_fp16, b = var_2094_cast_fp16, cond = mask_3)[name = tensor("input_605_cast_fp16")]; - tensor x_255_transpose_x_0 = const()[name = tensor("x_255_transpose_x_0"), val = tensor(false)]; - tensor x_255_transpose_y_0 = const()[name = tensor("x_255_transpose_y_0"), val = tensor(false)]; - tensor value_25_cast_fp16 = transpose(perm = value_25_perm_0, x = v_23_cast_fp16)[name = tensor("transpose_232")]; - tensor x_255_cast_fp16 = matmul(transpose_x = x_255_transpose_x_0, transpose_y = x_255_transpose_y_0, x = input_605_cast_fp16, y = value_25_cast_fp16)[name = tensor("x_255_cast_fp16")]; - tensor var_2098_perm_0 = const()[name = tensor("op_2098_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_2099 = const()[name = tensor("op_2099"), val = tensor([1, -1, 1024])]; - tensor var_2098_cast_fp16 = transpose(perm = var_2098_perm_0, x = x_255_cast_fp16)[name = tensor("transpose_231")]; - tensor input_607_cast_fp16 = reshape(shape = var_2099, x = var_2098_cast_fp16)[name = tensor("input_607_cast_fp16")]; - tensor module_layers_11_self_attn_linear_out_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(214766272))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(215552768))), name = tensor("module_layers_11_self_attn_linear_out_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_106_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_11_self_attn_linear_out_weight_to_fp16_palettized, x = input_607_cast_fp16)[name = tensor("linear_106_cast_fp16")]; - tensor input_611_cast_fp16 = add(x = input_603_cast_fp16, y = linear_106_cast_fp16)[name = tensor("input_611_cast_fp16")]; - tensor x_259_axes_0 = const()[name = tensor("x_259_axes_0"), val = tensor([-1])]; - tensor module_layers_11_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_11_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(215552960)))]; - tensor module_layers_11_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_11_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(215555072)))]; - tensor x_259_cast_fp16 = layer_norm(axes = x_259_axes_0, beta = module_layers_11_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_11_norm_conv_weight_to_fp16, x = input_611_cast_fp16)[name = tensor("x_259_cast_fp16")]; - tensor input_613_perm_0 = const()[name = tensor("input_613_perm_0"), val = tensor([0, 2, 1])]; - tensor input_615_pad_type_0 = const()[name = tensor("input_615_pad_type_0"), val = tensor("valid")]; - tensor input_615_strides_0 = const()[name = tensor("input_615_strides_0"), val = tensor([1])]; - tensor input_615_pad_0 = const()[name = tensor("input_615_pad_0"), val = tensor([0, 0])]; - tensor input_615_dilations_0 = const()[name = tensor("input_615_dilations_0"), val = tensor([1])]; - tensor input_615_groups_0 = const()[name = tensor("input_615_groups_0"), val = tensor(1)]; - tensor module_layers_11_conv_pointwise_conv1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(215557184))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(217130112))), name = tensor("module_layers_11_conv_pointwise_conv1_weight_to_fp16_palettized"), shape = tensor([2048, 1024, 1])]; - tensor input_613_cast_fp16 = transpose(perm = input_613_perm_0, x = x_259_cast_fp16)[name = tensor("transpose_230")]; - tensor input_615_cast_fp16 = conv(dilations = input_615_dilations_0, groups = input_615_groups_0, pad = input_615_pad_0, pad_type = input_615_pad_type_0, strides = input_615_strides_0, weight = module_layers_11_conv_pointwise_conv1_weight_to_fp16_palettized, x = input_613_cast_fp16)[name = tensor("input_615_cast_fp16")]; - tensor x_261_split_num_splits_0 = const()[name = tensor("x_261_split_num_splits_0"), val = tensor(2)]; - tensor x_261_split_axis_0 = const()[name = tensor("x_261_split_axis_0"), val = tensor(1)]; - tensor x_261_split_cast_fp16_0, tensor x_261_split_cast_fp16_1 = split(axis = x_261_split_axis_0, num_splits = x_261_split_num_splits_0, x = input_615_cast_fp16)[name = tensor("x_261_split_cast_fp16")]; - tensor x_261_split_1_sigmoid_cast_fp16 = sigmoid(x = x_261_split_cast_fp16_1)[name = tensor("x_261_split_1_sigmoid_cast_fp16")]; - tensor x_261_cast_fp16 = mul(x = x_261_split_cast_fp16_0, y = x_261_split_1_sigmoid_cast_fp16)[name = tensor("x_261_cast_fp16")]; - tensor input_617_cast_fp16 = select(a = var_11_to_fp16, b = x_261_cast_fp16, cond = var_328)[name = tensor("input_617_cast_fp16")]; - tensor input_619_pad_0 = const()[name = tensor("input_619_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_619_mode_0 = const()[name = tensor("input_619_mode_0"), val = tensor("constant")]; - tensor const_127_to_fp16 = const()[name = tensor("const_127_to_fp16"), val = tensor(0x0p+0)]; - tensor input_619_cast_fp16 = pad(constant_val = const_127_to_fp16, mode = input_619_mode_0, pad = input_619_pad_0, x = input_617_cast_fp16)[name = tensor("input_619_cast_fp16")]; - tensor input_621_pad_type_0 = const()[name = tensor("input_621_pad_type_0"), val = tensor("valid")]; - tensor input_621_groups_0 = const()[name = tensor("input_621_groups_0"), val = tensor(1024)]; - tensor input_621_strides_0 = const()[name = tensor("input_621_strides_0"), val = tensor([1])]; - tensor input_621_pad_0 = const()[name = tensor("input_621_pad_0"), val = tensor([0, 0])]; - tensor input_621_dilations_0 = const()[name = tensor("input_621_dilations_0"), val = tensor([1])]; - tensor const_270_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(217130304))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(217137280))), name = tensor("const_270_to_fp16_palettized"), shape = tensor([1024, 1, 9])]; - tensor const_271_to_fp16 = const()[name = tensor("const_271_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(217137472)))]; - tensor input_623_cast_fp16 = conv(bias = const_271_to_fp16, dilations = input_621_dilations_0, groups = input_621_groups_0, pad = input_621_pad_0, pad_type = input_621_pad_type_0, strides = input_621_strides_0, weight = const_270_to_fp16_palettized, x = input_619_cast_fp16)[name = tensor("input_623_cast_fp16")]; - tensor input_625_cast_fp16 = silu(x = input_623_cast_fp16)[name = tensor("input_625_cast_fp16")]; - tensor x_263_pad_type_0 = const()[name = tensor("x_263_pad_type_0"), val = tensor("valid")]; - tensor x_263_strides_0 = const()[name = tensor("x_263_strides_0"), val = tensor([1])]; - tensor x_263_pad_0 = const()[name = tensor("x_263_pad_0"), val = tensor([0, 0])]; - tensor x_263_dilations_0 = const()[name = tensor("x_263_dilations_0"), val = tensor([1])]; - tensor x_263_groups_0 = const()[name = tensor("x_263_groups_0"), val = tensor(1)]; - tensor module_layers_11_conv_pointwise_conv2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(217139584))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(217926080))), name = tensor("module_layers_11_conv_pointwise_conv2_weight_to_fp16_palettized"), shape = tensor([1024, 1024, 1])]; - tensor x_263_cast_fp16 = conv(dilations = x_263_dilations_0, groups = x_263_groups_0, pad = x_263_pad_0, pad_type = x_263_pad_type_0, strides = x_263_strides_0, weight = module_layers_11_conv_pointwise_conv2_weight_to_fp16_palettized, x = input_625_cast_fp16)[name = tensor("x_263_cast_fp16")]; - tensor input_627_perm_0 = const()[name = tensor("input_627_perm_0"), val = tensor([0, 2, 1])]; - tensor input_627_cast_fp16 = transpose(perm = input_627_perm_0, x = x_263_cast_fp16)[name = tensor("transpose_229")]; - tensor input_629_cast_fp16 = add(x = input_611_cast_fp16, y = input_627_cast_fp16)[name = tensor("input_629_cast_fp16")]; - tensor input_631_axes_0 = const()[name = tensor("input_631_axes_0"), val = tensor([-1])]; - tensor module_layers_11_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_11_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(217926272)))]; - tensor module_layers_11_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_11_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(217928384)))]; - tensor input_631_cast_fp16 = layer_norm(axes = input_631_axes_0, beta = module_layers_11_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_11_norm_feed_forward2_weight_to_fp16, x = input_629_cast_fp16)[name = tensor("input_631_cast_fp16")]; - tensor module_layers_11_feed_forward2_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(217930496))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(221076288))), name = tensor("module_layers_11_feed_forward2_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_107_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_11_feed_forward2_linear1_weight_to_fp16_palettized, x = input_631_cast_fp16)[name = tensor("linear_107_cast_fp16")]; - tensor input_635_cast_fp16 = silu(x = linear_107_cast_fp16)[name = tensor("input_635_cast_fp16")]; - tensor module_layers_11_feed_forward2_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(221076480))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(224222272))), name = tensor("module_layers_11_feed_forward2_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_108_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_11_feed_forward2_linear2_weight_to_fp16_palettized, x = input_635_cast_fp16)[name = tensor("linear_108_cast_fp16")]; - tensor var_2159_to_fp16 = const()[name = tensor("op_2159_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2160_cast_fp16 = mul(x = linear_108_cast_fp16, y = var_2159_to_fp16)[name = tensor("op_2160_cast_fp16")]; - tensor input_641_cast_fp16 = add(x = input_629_cast_fp16, y = var_2160_cast_fp16)[name = tensor("input_641_cast_fp16")]; - tensor input_643_axes_0 = const()[name = tensor("input_643_axes_0"), val = tensor([-1])]; - tensor module_layers_11_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_11_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(224222464)))]; - tensor module_layers_11_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_11_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(224224576)))]; - tensor input_643_cast_fp16 = layer_norm(axes = input_643_axes_0, beta = module_layers_11_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_11_norm_out_weight_to_fp16, x = input_641_cast_fp16)[name = tensor("input_643_cast_fp16")]; - tensor input_645_axes_0 = const()[name = tensor("input_645_axes_0"), val = tensor([-1])]; - tensor module_layers_12_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_12_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(224226688)))]; - tensor module_layers_12_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_12_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(224228800)))]; - tensor input_645_cast_fp16 = layer_norm(axes = input_645_axes_0, beta = module_layers_12_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_12_norm_feed_forward1_weight_to_fp16, x = input_643_cast_fp16)[name = tensor("input_645_cast_fp16")]; - tensor module_layers_12_feed_forward1_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(224230912))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(227376704))), name = tensor("module_layers_12_feed_forward1_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_109_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_12_feed_forward1_linear1_weight_to_fp16_palettized, x = input_645_cast_fp16)[name = tensor("linear_109_cast_fp16")]; - tensor input_649_cast_fp16 = silu(x = linear_109_cast_fp16)[name = tensor("input_649_cast_fp16")]; - tensor module_layers_12_feed_forward1_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(227376896))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(230522688))), name = tensor("module_layers_12_feed_forward1_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_110_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_12_feed_forward1_linear2_weight_to_fp16_palettized, x = input_649_cast_fp16)[name = tensor("linear_110_cast_fp16")]; - tensor var_2188_to_fp16 = const()[name = tensor("op_2188_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2189_cast_fp16 = mul(x = linear_110_cast_fp16, y = var_2188_to_fp16)[name = tensor("op_2189_cast_fp16")]; - tensor input_655_cast_fp16 = add(x = input_643_cast_fp16, y = var_2189_cast_fp16)[name = tensor("input_655_cast_fp16")]; - tensor query_25_axes_0 = const()[name = tensor("query_25_axes_0"), val = tensor([-1])]; - tensor module_layers_12_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_12_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(230522880)))]; - tensor module_layers_12_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_12_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(230524992)))]; - tensor query_25_cast_fp16 = layer_norm(axes = query_25_axes_0, beta = module_layers_12_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_12_norm_self_att_weight_to_fp16, x = input_655_cast_fp16)[name = tensor("query_25_cast_fp16")]; - tensor module_layers_12_self_attn_linear_q_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(230527104))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(231313600))), name = tensor("module_layers_12_self_attn_linear_q_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_111_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_12_self_attn_linear_q_weight_to_fp16_palettized, x = query_25_cast_fp16)[name = tensor("linear_111_cast_fp16")]; - tensor var_2205 = const()[name = tensor("op_2205"), val = tensor([1, -1, 8, 128])]; - tensor q_73_cast_fp16 = reshape(shape = var_2205, x = linear_111_cast_fp16)[name = tensor("q_73_cast_fp16")]; - tensor module_layers_12_self_attn_linear_k_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(231313792))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(232100288))), name = tensor("module_layers_12_self_attn_linear_k_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_112_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_12_self_attn_linear_k_weight_to_fp16_palettized, x = query_25_cast_fp16)[name = tensor("linear_112_cast_fp16")]; - tensor var_2209 = const()[name = tensor("op_2209"), val = tensor([1, -1, 8, 128])]; - tensor k_49_cast_fp16 = reshape(shape = var_2209, x = linear_112_cast_fp16)[name = tensor("k_49_cast_fp16")]; - tensor module_layers_12_self_attn_linear_v_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(232100480))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(232886976))), name = tensor("module_layers_12_self_attn_linear_v_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_113_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_12_self_attn_linear_v_weight_to_fp16_palettized, x = query_25_cast_fp16)[name = tensor("linear_113_cast_fp16")]; - tensor var_2213 = const()[name = tensor("op_2213"), val = tensor([1, -1, 8, 128])]; - tensor v_25_cast_fp16 = reshape(shape = var_2213, x = linear_113_cast_fp16)[name = tensor("v_25_cast_fp16")]; - tensor value_27_perm_0 = const()[name = tensor("value_27_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor module_layers_12_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_12_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(232887168)))]; - tensor var_2225_cast_fp16 = add(x = q_73_cast_fp16, y = module_layers_12_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2225_cast_fp16")]; - tensor module_layers_12_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_12_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(232889280)))]; - tensor var_2227_cast_fp16 = add(x = q_73_cast_fp16, y = module_layers_12_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2227_cast_fp16")]; - tensor q_with_bias_v_25_perm_0 = const()[name = tensor("q_with_bias_v_25_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor x_271_transpose_x_0 = const()[name = tensor("x_271_transpose_x_0"), val = tensor(false)]; - tensor x_271_transpose_y_0 = const()[name = tensor("x_271_transpose_y_0"), val = tensor(false)]; - tensor op_2229_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(232891392))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(233179456))), name = tensor("op_2229_to_fp16_palettized"), shape = tensor([1, 8, 128, 375])]; - tensor q_with_bias_v_25_cast_fp16 = transpose(perm = q_with_bias_v_25_perm_0, x = var_2227_cast_fp16)[name = tensor("transpose_228")]; - tensor x_271_cast_fp16 = matmul(transpose_x = x_271_transpose_x_0, transpose_y = x_271_transpose_y_0, x = q_with_bias_v_25_cast_fp16, y = op_2229_to_fp16_palettized)[name = tensor("x_271_cast_fp16")]; - tensor x_273_pad_0 = const()[name = tensor("x_273_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_273_mode_0 = const()[name = tensor("x_273_mode_0"), val = tensor("constant")]; - tensor const_134_to_fp16 = const()[name = tensor("const_134_to_fp16"), val = tensor(0x0p+0)]; - tensor x_273_cast_fp16 = pad(constant_val = const_134_to_fp16, mode = x_273_mode_0, pad = x_273_pad_0, x = x_271_cast_fp16)[name = tensor("x_273_cast_fp16")]; - tensor var_2237 = const()[name = tensor("op_2237"), val = tensor([1, 8, -1, 188])]; - tensor x_275_cast_fp16 = reshape(shape = var_2237, x = x_273_cast_fp16)[name = tensor("x_275_cast_fp16")]; - tensor var_2241_begin_0 = const()[name = tensor("op_2241_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_2241_end_0 = const()[name = tensor("op_2241_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_2241_end_mask_0 = const()[name = tensor("op_2241_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_2241_cast_fp16 = slice_by_index(begin = var_2241_begin_0, end = var_2241_end_0, end_mask = var_2241_end_mask_0, x = x_275_cast_fp16)[name = tensor("op_2241_cast_fp16")]; - tensor var_2242 = const()[name = tensor("op_2242"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_49_cast_fp16 = reshape(shape = var_2242, x = var_2241_cast_fp16)[name = tensor("matrix_bd_49_cast_fp16")]; - tensor matrix_ac_25_transpose_x_0 = const()[name = tensor("matrix_ac_25_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_25_transpose_y_0 = const()[name = tensor("matrix_ac_25_transpose_y_0"), val = tensor(false)]; - tensor transpose_120_perm_0 = const()[name = tensor("transpose_120_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_121_perm_0 = const()[name = tensor("transpose_121_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_121 = transpose(perm = transpose_121_perm_0, x = k_49_cast_fp16)[name = tensor("transpose_226")]; - tensor transpose_120 = transpose(perm = transpose_120_perm_0, x = var_2225_cast_fp16)[name = tensor("transpose_227")]; - tensor matrix_ac_25_cast_fp16 = matmul(transpose_x = matrix_ac_25_transpose_x_0, transpose_y = matrix_ac_25_transpose_y_0, x = transpose_120, y = transpose_121)[name = tensor("matrix_ac_25_cast_fp16")]; - tensor matrix_bd_51_begin_0 = const()[name = tensor("matrix_bd_51_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_51_end_0 = const()[name = tensor("matrix_bd_51_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_51_end_mask_0 = const()[name = tensor("matrix_bd_51_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_51_cast_fp16 = slice_by_index(begin = matrix_bd_51_begin_0, end = matrix_bd_51_end_0, end_mask = matrix_bd_51_end_mask_0, x = matrix_bd_49_cast_fp16)[name = tensor("matrix_bd_51_cast_fp16")]; - tensor var_2251_cast_fp16 = add(x = matrix_ac_25_cast_fp16, y = matrix_bd_51_cast_fp16)[name = tensor("op_2251_cast_fp16")]; - tensor _inversed_scores_49_y_0_to_fp16 = const()[name = tensor("_inversed_scores_49_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; - tensor _inversed_scores_49_cast_fp16 = mul(x = var_2251_cast_fp16, y = _inversed_scores_49_y_0_to_fp16)[name = tensor("_inversed_scores_49_cast_fp16")]; - tensor scores_51_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_49_cast_fp16, cond = mask_3)[name = tensor("scores_51_cast_fp16")]; - tensor var_2257_cast_fp16 = softmax(axis = var_30, x = scores_51_cast_fp16)[name = tensor("op_2257_cast_fp16")]; - tensor input_657_cast_fp16 = select(a = var_11_to_fp16, b = var_2257_cast_fp16, cond = mask_3)[name = tensor("input_657_cast_fp16")]; - tensor x_277_transpose_x_0 = const()[name = tensor("x_277_transpose_x_0"), val = tensor(false)]; - tensor x_277_transpose_y_0 = const()[name = tensor("x_277_transpose_y_0"), val = tensor(false)]; - tensor value_27_cast_fp16 = transpose(perm = value_27_perm_0, x = v_25_cast_fp16)[name = tensor("transpose_225")]; - tensor x_277_cast_fp16 = matmul(transpose_x = x_277_transpose_x_0, transpose_y = x_277_transpose_y_0, x = input_657_cast_fp16, y = value_27_cast_fp16)[name = tensor("x_277_cast_fp16")]; - tensor var_2261_perm_0 = const()[name = tensor("op_2261_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_2262 = const()[name = tensor("op_2262"), val = tensor([1, -1, 1024])]; - tensor var_2261_cast_fp16 = transpose(perm = var_2261_perm_0, x = x_277_cast_fp16)[name = tensor("transpose_224")]; - tensor input_659_cast_fp16 = reshape(shape = var_2262, x = var_2261_cast_fp16)[name = tensor("input_659_cast_fp16")]; - tensor module_layers_12_self_attn_linear_out_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(233179648))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(233966144))), name = tensor("module_layers_12_self_attn_linear_out_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_115_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_12_self_attn_linear_out_weight_to_fp16_palettized, x = input_659_cast_fp16)[name = tensor("linear_115_cast_fp16")]; - tensor input_663_cast_fp16 = add(x = input_655_cast_fp16, y = linear_115_cast_fp16)[name = tensor("input_663_cast_fp16")]; - tensor x_281_axes_0 = const()[name = tensor("x_281_axes_0"), val = tensor([-1])]; - tensor module_layers_12_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_12_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(233966336)))]; - tensor module_layers_12_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_12_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(233968448)))]; - tensor x_281_cast_fp16 = layer_norm(axes = x_281_axes_0, beta = module_layers_12_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_12_norm_conv_weight_to_fp16, x = input_663_cast_fp16)[name = tensor("x_281_cast_fp16")]; - tensor input_665_perm_0 = const()[name = tensor("input_665_perm_0"), val = tensor([0, 2, 1])]; - tensor input_667_pad_type_0 = const()[name = tensor("input_667_pad_type_0"), val = tensor("valid")]; - tensor input_667_strides_0 = const()[name = tensor("input_667_strides_0"), val = tensor([1])]; - tensor input_667_pad_0 = const()[name = tensor("input_667_pad_0"), val = tensor([0, 0])]; - tensor input_667_dilations_0 = const()[name = tensor("input_667_dilations_0"), val = tensor([1])]; - tensor input_667_groups_0 = const()[name = tensor("input_667_groups_0"), val = tensor(1)]; - tensor module_layers_12_conv_pointwise_conv1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(233970560))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(235543488))), name = tensor("module_layers_12_conv_pointwise_conv1_weight_to_fp16_palettized"), shape = tensor([2048, 1024, 1])]; - tensor input_665_cast_fp16 = transpose(perm = input_665_perm_0, x = x_281_cast_fp16)[name = tensor("transpose_223")]; - tensor input_667_cast_fp16 = conv(dilations = input_667_dilations_0, groups = input_667_groups_0, pad = input_667_pad_0, pad_type = input_667_pad_type_0, strides = input_667_strides_0, weight = module_layers_12_conv_pointwise_conv1_weight_to_fp16_palettized, x = input_665_cast_fp16)[name = tensor("input_667_cast_fp16")]; - tensor x_283_split_num_splits_0 = const()[name = tensor("x_283_split_num_splits_0"), val = tensor(2)]; - tensor x_283_split_axis_0 = const()[name = tensor("x_283_split_axis_0"), val = tensor(1)]; - tensor x_283_split_cast_fp16_0, tensor x_283_split_cast_fp16_1 = split(axis = x_283_split_axis_0, num_splits = x_283_split_num_splits_0, x = input_667_cast_fp16)[name = tensor("x_283_split_cast_fp16")]; - tensor x_283_split_1_sigmoid_cast_fp16 = sigmoid(x = x_283_split_cast_fp16_1)[name = tensor("x_283_split_1_sigmoid_cast_fp16")]; - tensor x_283_cast_fp16 = mul(x = x_283_split_cast_fp16_0, y = x_283_split_1_sigmoid_cast_fp16)[name = tensor("x_283_cast_fp16")]; - tensor input_669_cast_fp16 = select(a = var_11_to_fp16, b = x_283_cast_fp16, cond = var_328)[name = tensor("input_669_cast_fp16")]; - tensor input_671_pad_0 = const()[name = tensor("input_671_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_671_mode_0 = const()[name = tensor("input_671_mode_0"), val = tensor("constant")]; - tensor const_137_to_fp16 = const()[name = tensor("const_137_to_fp16"), val = tensor(0x0p+0)]; - tensor input_671_cast_fp16 = pad(constant_val = const_137_to_fp16, mode = input_671_mode_0, pad = input_671_pad_0, x = input_669_cast_fp16)[name = tensor("input_671_cast_fp16")]; - tensor input_673_pad_type_0 = const()[name = tensor("input_673_pad_type_0"), val = tensor("valid")]; - tensor input_673_groups_0 = const()[name = tensor("input_673_groups_0"), val = tensor(1024)]; - tensor input_673_strides_0 = const()[name = tensor("input_673_strides_0"), val = tensor([1])]; - tensor input_673_pad_0 = const()[name = tensor("input_673_pad_0"), val = tensor([0, 0])]; - tensor input_673_dilations_0 = const()[name = tensor("input_673_dilations_0"), val = tensor([1])]; - tensor const_272_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(235543680))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(235550656))), name = tensor("const_272_to_fp16_palettized"), shape = tensor([1024, 1, 9])]; - tensor const_273_to_fp16 = const()[name = tensor("const_273_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(235550848)))]; - tensor input_675_cast_fp16 = conv(bias = const_273_to_fp16, dilations = input_673_dilations_0, groups = input_673_groups_0, pad = input_673_pad_0, pad_type = input_673_pad_type_0, strides = input_673_strides_0, weight = const_272_to_fp16_palettized, x = input_671_cast_fp16)[name = tensor("input_675_cast_fp16")]; - tensor input_677_cast_fp16 = silu(x = input_675_cast_fp16)[name = tensor("input_677_cast_fp16")]; - tensor x_285_pad_type_0 = const()[name = tensor("x_285_pad_type_0"), val = tensor("valid")]; - tensor x_285_strides_0 = const()[name = tensor("x_285_strides_0"), val = tensor([1])]; - tensor x_285_pad_0 = const()[name = tensor("x_285_pad_0"), val = tensor([0, 0])]; - tensor x_285_dilations_0 = const()[name = tensor("x_285_dilations_0"), val = tensor([1])]; - tensor x_285_groups_0 = const()[name = tensor("x_285_groups_0"), val = tensor(1)]; - tensor module_layers_12_conv_pointwise_conv2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(235552960))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(236339456))), name = tensor("module_layers_12_conv_pointwise_conv2_weight_to_fp16_palettized"), shape = tensor([1024, 1024, 1])]; - tensor x_285_cast_fp16 = conv(dilations = x_285_dilations_0, groups = x_285_groups_0, pad = x_285_pad_0, pad_type = x_285_pad_type_0, strides = x_285_strides_0, weight = module_layers_12_conv_pointwise_conv2_weight_to_fp16_palettized, x = input_677_cast_fp16)[name = tensor("x_285_cast_fp16")]; - tensor input_679_perm_0 = const()[name = tensor("input_679_perm_0"), val = tensor([0, 2, 1])]; - tensor input_679_cast_fp16 = transpose(perm = input_679_perm_0, x = x_285_cast_fp16)[name = tensor("transpose_222")]; - tensor input_681_cast_fp16 = add(x = input_663_cast_fp16, y = input_679_cast_fp16)[name = tensor("input_681_cast_fp16")]; - tensor input_683_axes_0 = const()[name = tensor("input_683_axes_0"), val = tensor([-1])]; - tensor module_layers_12_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_12_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(236339648)))]; - tensor module_layers_12_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_12_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(236341760)))]; - tensor input_683_cast_fp16 = layer_norm(axes = input_683_axes_0, beta = module_layers_12_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_12_norm_feed_forward2_weight_to_fp16, x = input_681_cast_fp16)[name = tensor("input_683_cast_fp16")]; - tensor module_layers_12_feed_forward2_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(236343872))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(239489664))), name = tensor("module_layers_12_feed_forward2_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_116_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_12_feed_forward2_linear1_weight_to_fp16_palettized, x = input_683_cast_fp16)[name = tensor("linear_116_cast_fp16")]; - tensor input_687_cast_fp16 = silu(x = linear_116_cast_fp16)[name = tensor("input_687_cast_fp16")]; - tensor module_layers_12_feed_forward2_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(239489856))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(242635648))), name = tensor("module_layers_12_feed_forward2_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_117_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_12_feed_forward2_linear2_weight_to_fp16_palettized, x = input_687_cast_fp16)[name = tensor("linear_117_cast_fp16")]; - tensor var_2322_to_fp16 = const()[name = tensor("op_2322_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2323_cast_fp16 = mul(x = linear_117_cast_fp16, y = var_2322_to_fp16)[name = tensor("op_2323_cast_fp16")]; - tensor input_693_cast_fp16 = add(x = input_681_cast_fp16, y = var_2323_cast_fp16)[name = tensor("input_693_cast_fp16")]; - tensor input_695_axes_0 = const()[name = tensor("input_695_axes_0"), val = tensor([-1])]; - tensor module_layers_12_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_12_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(242635840)))]; - tensor module_layers_12_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_12_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(242637952)))]; - tensor input_695_cast_fp16 = layer_norm(axes = input_695_axes_0, beta = module_layers_12_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_12_norm_out_weight_to_fp16, x = input_693_cast_fp16)[name = tensor("input_695_cast_fp16")]; - tensor input_697_axes_0 = const()[name = tensor("input_697_axes_0"), val = tensor([-1])]; - tensor module_layers_13_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_13_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(242640064)))]; - tensor module_layers_13_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_13_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(242642176)))]; - tensor input_697_cast_fp16 = layer_norm(axes = input_697_axes_0, beta = module_layers_13_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_13_norm_feed_forward1_weight_to_fp16, x = input_695_cast_fp16)[name = tensor("input_697_cast_fp16")]; - tensor module_layers_13_feed_forward1_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(242644288))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(245790080))), name = tensor("module_layers_13_feed_forward1_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_118_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_13_feed_forward1_linear1_weight_to_fp16_palettized, x = input_697_cast_fp16)[name = tensor("linear_118_cast_fp16")]; - tensor input_701_cast_fp16 = silu(x = linear_118_cast_fp16)[name = tensor("input_701_cast_fp16")]; - tensor module_layers_13_feed_forward1_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(245790272))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(248936064))), name = tensor("module_layers_13_feed_forward1_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_119_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_13_feed_forward1_linear2_weight_to_fp16_palettized, x = input_701_cast_fp16)[name = tensor("linear_119_cast_fp16")]; - tensor var_2351_to_fp16 = const()[name = tensor("op_2351_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2352_cast_fp16 = mul(x = linear_119_cast_fp16, y = var_2351_to_fp16)[name = tensor("op_2352_cast_fp16")]; - tensor input_707_cast_fp16 = add(x = input_695_cast_fp16, y = var_2352_cast_fp16)[name = tensor("input_707_cast_fp16")]; - tensor query_27_axes_0 = const()[name = tensor("query_27_axes_0"), val = tensor([-1])]; - tensor module_layers_13_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_13_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(248936256)))]; - tensor module_layers_13_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_13_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(248938368)))]; - tensor query_27_cast_fp16 = layer_norm(axes = query_27_axes_0, beta = module_layers_13_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_13_norm_self_att_weight_to_fp16, x = input_707_cast_fp16)[name = tensor("query_27_cast_fp16")]; - tensor module_layers_13_self_attn_linear_q_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(248940480))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(249726976))), name = tensor("module_layers_13_self_attn_linear_q_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_120_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_13_self_attn_linear_q_weight_to_fp16_palettized, x = query_27_cast_fp16)[name = tensor("linear_120_cast_fp16")]; - tensor var_2368 = const()[name = tensor("op_2368"), val = tensor([1, -1, 8, 128])]; - tensor q_79_cast_fp16 = reshape(shape = var_2368, x = linear_120_cast_fp16)[name = tensor("q_79_cast_fp16")]; - tensor module_layers_13_self_attn_linear_k_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(249727168))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(250513664))), name = tensor("module_layers_13_self_attn_linear_k_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_121_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_13_self_attn_linear_k_weight_to_fp16_palettized, x = query_27_cast_fp16)[name = tensor("linear_121_cast_fp16")]; - tensor var_2372 = const()[name = tensor("op_2372"), val = tensor([1, -1, 8, 128])]; - tensor k_53_cast_fp16 = reshape(shape = var_2372, x = linear_121_cast_fp16)[name = tensor("k_53_cast_fp16")]; - tensor module_layers_13_self_attn_linear_v_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(250513856))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(251300352))), name = tensor("module_layers_13_self_attn_linear_v_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_122_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_13_self_attn_linear_v_weight_to_fp16_palettized, x = query_27_cast_fp16)[name = tensor("linear_122_cast_fp16")]; - tensor var_2376 = const()[name = tensor("op_2376"), val = tensor([1, -1, 8, 128])]; - tensor v_27_cast_fp16 = reshape(shape = var_2376, x = linear_122_cast_fp16)[name = tensor("v_27_cast_fp16")]; - tensor value_29_perm_0 = const()[name = tensor("value_29_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor module_layers_13_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_13_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(251300544)))]; - tensor var_2388_cast_fp16 = add(x = q_79_cast_fp16, y = module_layers_13_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2388_cast_fp16")]; - tensor module_layers_13_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_13_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(251302656)))]; - tensor var_2390_cast_fp16 = add(x = q_79_cast_fp16, y = module_layers_13_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2390_cast_fp16")]; - tensor q_with_bias_v_27_perm_0 = const()[name = tensor("q_with_bias_v_27_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor x_293_transpose_x_0 = const()[name = tensor("x_293_transpose_x_0"), val = tensor(false)]; - tensor x_293_transpose_y_0 = const()[name = tensor("x_293_transpose_y_0"), val = tensor(false)]; - tensor op_2392_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(251304768))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(251592832))), name = tensor("op_2392_to_fp16_palettized"), shape = tensor([1, 8, 128, 375])]; - tensor q_with_bias_v_27_cast_fp16 = transpose(perm = q_with_bias_v_27_perm_0, x = var_2390_cast_fp16)[name = tensor("transpose_221")]; - tensor x_293_cast_fp16 = matmul(transpose_x = x_293_transpose_x_0, transpose_y = x_293_transpose_y_0, x = q_with_bias_v_27_cast_fp16, y = op_2392_to_fp16_palettized)[name = tensor("x_293_cast_fp16")]; - tensor x_295_pad_0 = const()[name = tensor("x_295_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_295_mode_0 = const()[name = tensor("x_295_mode_0"), val = tensor("constant")]; - tensor const_144_to_fp16 = const()[name = tensor("const_144_to_fp16"), val = tensor(0x0p+0)]; - tensor x_295_cast_fp16 = pad(constant_val = const_144_to_fp16, mode = x_295_mode_0, pad = x_295_pad_0, x = x_293_cast_fp16)[name = tensor("x_295_cast_fp16")]; - tensor var_2400 = const()[name = tensor("op_2400"), val = tensor([1, 8, -1, 188])]; - tensor x_297_cast_fp16 = reshape(shape = var_2400, x = x_295_cast_fp16)[name = tensor("x_297_cast_fp16")]; - tensor var_2404_begin_0 = const()[name = tensor("op_2404_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_2404_end_0 = const()[name = tensor("op_2404_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_2404_end_mask_0 = const()[name = tensor("op_2404_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_2404_cast_fp16 = slice_by_index(begin = var_2404_begin_0, end = var_2404_end_0, end_mask = var_2404_end_mask_0, x = x_297_cast_fp16)[name = tensor("op_2404_cast_fp16")]; - tensor var_2405 = const()[name = tensor("op_2405"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_53_cast_fp16 = reshape(shape = var_2405, x = var_2404_cast_fp16)[name = tensor("matrix_bd_53_cast_fp16")]; - tensor matrix_ac_27_transpose_x_0 = const()[name = tensor("matrix_ac_27_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_27_transpose_y_0 = const()[name = tensor("matrix_ac_27_transpose_y_0"), val = tensor(false)]; - tensor transpose_122_perm_0 = const()[name = tensor("transpose_122_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_123_perm_0 = const()[name = tensor("transpose_123_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_123 = transpose(perm = transpose_123_perm_0, x = k_53_cast_fp16)[name = tensor("transpose_219")]; - tensor transpose_122 = transpose(perm = transpose_122_perm_0, x = var_2388_cast_fp16)[name = tensor("transpose_220")]; - tensor matrix_ac_27_cast_fp16 = matmul(transpose_x = matrix_ac_27_transpose_x_0, transpose_y = matrix_ac_27_transpose_y_0, x = transpose_122, y = transpose_123)[name = tensor("matrix_ac_27_cast_fp16")]; - tensor matrix_bd_55_begin_0 = const()[name = tensor("matrix_bd_55_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_55_end_0 = const()[name = tensor("matrix_bd_55_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_55_end_mask_0 = const()[name = tensor("matrix_bd_55_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_55_cast_fp16 = slice_by_index(begin = matrix_bd_55_begin_0, end = matrix_bd_55_end_0, end_mask = matrix_bd_55_end_mask_0, x = matrix_bd_53_cast_fp16)[name = tensor("matrix_bd_55_cast_fp16")]; - tensor var_2414_cast_fp16 = add(x = matrix_ac_27_cast_fp16, y = matrix_bd_55_cast_fp16)[name = tensor("op_2414_cast_fp16")]; - tensor _inversed_scores_53_y_0_to_fp16 = const()[name = tensor("_inversed_scores_53_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; - tensor _inversed_scores_53_cast_fp16 = mul(x = var_2414_cast_fp16, y = _inversed_scores_53_y_0_to_fp16)[name = tensor("_inversed_scores_53_cast_fp16")]; - tensor scores_55_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_53_cast_fp16, cond = mask_3)[name = tensor("scores_55_cast_fp16")]; - tensor var_2420_cast_fp16 = softmax(axis = var_30, x = scores_55_cast_fp16)[name = tensor("op_2420_cast_fp16")]; - tensor input_709_cast_fp16 = select(a = var_11_to_fp16, b = var_2420_cast_fp16, cond = mask_3)[name = tensor("input_709_cast_fp16")]; - tensor x_299_transpose_x_0 = const()[name = tensor("x_299_transpose_x_0"), val = tensor(false)]; - tensor x_299_transpose_y_0 = const()[name = tensor("x_299_transpose_y_0"), val = tensor(false)]; - tensor value_29_cast_fp16 = transpose(perm = value_29_perm_0, x = v_27_cast_fp16)[name = tensor("transpose_218")]; - tensor x_299_cast_fp16 = matmul(transpose_x = x_299_transpose_x_0, transpose_y = x_299_transpose_y_0, x = input_709_cast_fp16, y = value_29_cast_fp16)[name = tensor("x_299_cast_fp16")]; - tensor var_2424_perm_0 = const()[name = tensor("op_2424_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_2425 = const()[name = tensor("op_2425"), val = tensor([1, -1, 1024])]; - tensor var_2424_cast_fp16 = transpose(perm = var_2424_perm_0, x = x_299_cast_fp16)[name = tensor("transpose_217")]; - tensor input_711_cast_fp16 = reshape(shape = var_2425, x = var_2424_cast_fp16)[name = tensor("input_711_cast_fp16")]; - tensor module_layers_13_self_attn_linear_out_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(251593024))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(252379520))), name = tensor("module_layers_13_self_attn_linear_out_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_124_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_13_self_attn_linear_out_weight_to_fp16_palettized, x = input_711_cast_fp16)[name = tensor("linear_124_cast_fp16")]; - tensor input_715_cast_fp16 = add(x = input_707_cast_fp16, y = linear_124_cast_fp16)[name = tensor("input_715_cast_fp16")]; - tensor x_303_axes_0 = const()[name = tensor("x_303_axes_0"), val = tensor([-1])]; - tensor module_layers_13_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_13_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(252379712)))]; - tensor module_layers_13_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_13_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(252381824)))]; - tensor x_303_cast_fp16 = layer_norm(axes = x_303_axes_0, beta = module_layers_13_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_13_norm_conv_weight_to_fp16, x = input_715_cast_fp16)[name = tensor("x_303_cast_fp16")]; - tensor input_717_perm_0 = const()[name = tensor("input_717_perm_0"), val = tensor([0, 2, 1])]; - tensor input_719_pad_type_0 = const()[name = tensor("input_719_pad_type_0"), val = tensor("valid")]; - tensor input_719_strides_0 = const()[name = tensor("input_719_strides_0"), val = tensor([1])]; - tensor input_719_pad_0 = const()[name = tensor("input_719_pad_0"), val = tensor([0, 0])]; - tensor input_719_dilations_0 = const()[name = tensor("input_719_dilations_0"), val = tensor([1])]; - tensor input_719_groups_0 = const()[name = tensor("input_719_groups_0"), val = tensor(1)]; - tensor module_layers_13_conv_pointwise_conv1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(252383936))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(253956864))), name = tensor("module_layers_13_conv_pointwise_conv1_weight_to_fp16_palettized"), shape = tensor([2048, 1024, 1])]; - tensor input_717_cast_fp16 = transpose(perm = input_717_perm_0, x = x_303_cast_fp16)[name = tensor("transpose_216")]; - tensor input_719_cast_fp16 = conv(dilations = input_719_dilations_0, groups = input_719_groups_0, pad = input_719_pad_0, pad_type = input_719_pad_type_0, strides = input_719_strides_0, weight = module_layers_13_conv_pointwise_conv1_weight_to_fp16_palettized, x = input_717_cast_fp16)[name = tensor("input_719_cast_fp16")]; - tensor x_305_split_num_splits_0 = const()[name = tensor("x_305_split_num_splits_0"), val = tensor(2)]; - tensor x_305_split_axis_0 = const()[name = tensor("x_305_split_axis_0"), val = tensor(1)]; - tensor x_305_split_cast_fp16_0, tensor x_305_split_cast_fp16_1 = split(axis = x_305_split_axis_0, num_splits = x_305_split_num_splits_0, x = input_719_cast_fp16)[name = tensor("x_305_split_cast_fp16")]; - tensor x_305_split_1_sigmoid_cast_fp16 = sigmoid(x = x_305_split_cast_fp16_1)[name = tensor("x_305_split_1_sigmoid_cast_fp16")]; - tensor x_305_cast_fp16 = mul(x = x_305_split_cast_fp16_0, y = x_305_split_1_sigmoid_cast_fp16)[name = tensor("x_305_cast_fp16")]; - tensor input_721_cast_fp16 = select(a = var_11_to_fp16, b = x_305_cast_fp16, cond = var_328)[name = tensor("input_721_cast_fp16")]; - tensor input_723_pad_0 = const()[name = tensor("input_723_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_723_mode_0 = const()[name = tensor("input_723_mode_0"), val = tensor("constant")]; - tensor const_147_to_fp16 = const()[name = tensor("const_147_to_fp16"), val = tensor(0x0p+0)]; - tensor input_723_cast_fp16 = pad(constant_val = const_147_to_fp16, mode = input_723_mode_0, pad = input_723_pad_0, x = input_721_cast_fp16)[name = tensor("input_723_cast_fp16")]; - tensor input_725_pad_type_0 = const()[name = tensor("input_725_pad_type_0"), val = tensor("valid")]; - tensor input_725_groups_0 = const()[name = tensor("input_725_groups_0"), val = tensor(1024)]; - tensor input_725_strides_0 = const()[name = tensor("input_725_strides_0"), val = tensor([1])]; - tensor input_725_pad_0 = const()[name = tensor("input_725_pad_0"), val = tensor([0, 0])]; - tensor input_725_dilations_0 = const()[name = tensor("input_725_dilations_0"), val = tensor([1])]; - tensor const_274_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(253957056))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(253964032))), name = tensor("const_274_to_fp16_palettized"), shape = tensor([1024, 1, 9])]; - tensor const_275_to_fp16 = const()[name = tensor("const_275_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(253964224)))]; - tensor input_727_cast_fp16 = conv(bias = const_275_to_fp16, dilations = input_725_dilations_0, groups = input_725_groups_0, pad = input_725_pad_0, pad_type = input_725_pad_type_0, strides = input_725_strides_0, weight = const_274_to_fp16_palettized, x = input_723_cast_fp16)[name = tensor("input_727_cast_fp16")]; - tensor input_729_cast_fp16 = silu(x = input_727_cast_fp16)[name = tensor("input_729_cast_fp16")]; - tensor x_307_pad_type_0 = const()[name = tensor("x_307_pad_type_0"), val = tensor("valid")]; - tensor x_307_strides_0 = const()[name = tensor("x_307_strides_0"), val = tensor([1])]; - tensor x_307_pad_0 = const()[name = tensor("x_307_pad_0"), val = tensor([0, 0])]; - tensor x_307_dilations_0 = const()[name = tensor("x_307_dilations_0"), val = tensor([1])]; - tensor x_307_groups_0 = const()[name = tensor("x_307_groups_0"), val = tensor(1)]; - tensor module_layers_13_conv_pointwise_conv2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(253966336))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(254752832))), name = tensor("module_layers_13_conv_pointwise_conv2_weight_to_fp16_palettized"), shape = tensor([1024, 1024, 1])]; - tensor x_307_cast_fp16 = conv(dilations = x_307_dilations_0, groups = x_307_groups_0, pad = x_307_pad_0, pad_type = x_307_pad_type_0, strides = x_307_strides_0, weight = module_layers_13_conv_pointwise_conv2_weight_to_fp16_palettized, x = input_729_cast_fp16)[name = tensor("x_307_cast_fp16")]; - tensor input_731_perm_0 = const()[name = tensor("input_731_perm_0"), val = tensor([0, 2, 1])]; - tensor input_731_cast_fp16 = transpose(perm = input_731_perm_0, x = x_307_cast_fp16)[name = tensor("transpose_215")]; - tensor input_733_cast_fp16 = add(x = input_715_cast_fp16, y = input_731_cast_fp16)[name = tensor("input_733_cast_fp16")]; - tensor input_735_axes_0 = const()[name = tensor("input_735_axes_0"), val = tensor([-1])]; - tensor module_layers_13_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_13_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(254753024)))]; - tensor module_layers_13_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_13_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(254755136)))]; - tensor input_735_cast_fp16 = layer_norm(axes = input_735_axes_0, beta = module_layers_13_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_13_norm_feed_forward2_weight_to_fp16, x = input_733_cast_fp16)[name = tensor("input_735_cast_fp16")]; - tensor module_layers_13_feed_forward2_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(254757248))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(257903040))), name = tensor("module_layers_13_feed_forward2_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_125_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_13_feed_forward2_linear1_weight_to_fp16_palettized, x = input_735_cast_fp16)[name = tensor("linear_125_cast_fp16")]; - tensor input_739_cast_fp16 = silu(x = linear_125_cast_fp16)[name = tensor("input_739_cast_fp16")]; - tensor module_layers_13_feed_forward2_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(257903232))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(261049024))), name = tensor("module_layers_13_feed_forward2_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_126_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_13_feed_forward2_linear2_weight_to_fp16_palettized, x = input_739_cast_fp16)[name = tensor("linear_126_cast_fp16")]; - tensor var_2485_to_fp16 = const()[name = tensor("op_2485_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2486_cast_fp16 = mul(x = linear_126_cast_fp16, y = var_2485_to_fp16)[name = tensor("op_2486_cast_fp16")]; - tensor input_745_cast_fp16 = add(x = input_733_cast_fp16, y = var_2486_cast_fp16)[name = tensor("input_745_cast_fp16")]; - tensor input_747_axes_0 = const()[name = tensor("input_747_axes_0"), val = tensor([-1])]; - tensor module_layers_13_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_13_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(261049216)))]; - tensor module_layers_13_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_13_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(261051328)))]; - tensor input_747_cast_fp16 = layer_norm(axes = input_747_axes_0, beta = module_layers_13_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_13_norm_out_weight_to_fp16, x = input_745_cast_fp16)[name = tensor("input_747_cast_fp16")]; - tensor input_749_axes_0 = const()[name = tensor("input_749_axes_0"), val = tensor([-1])]; - tensor module_layers_14_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_14_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(261053440)))]; - tensor module_layers_14_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_14_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(261055552)))]; - tensor input_749_cast_fp16 = layer_norm(axes = input_749_axes_0, beta = module_layers_14_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_14_norm_feed_forward1_weight_to_fp16, x = input_747_cast_fp16)[name = tensor("input_749_cast_fp16")]; - tensor module_layers_14_feed_forward1_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(261057664))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(264203456))), name = tensor("module_layers_14_feed_forward1_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_127_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_14_feed_forward1_linear1_weight_to_fp16_palettized, x = input_749_cast_fp16)[name = tensor("linear_127_cast_fp16")]; - tensor input_753_cast_fp16 = silu(x = linear_127_cast_fp16)[name = tensor("input_753_cast_fp16")]; - tensor module_layers_14_feed_forward1_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(264203648))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(267349440))), name = tensor("module_layers_14_feed_forward1_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_128_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_14_feed_forward1_linear2_weight_to_fp16_palettized, x = input_753_cast_fp16)[name = tensor("linear_128_cast_fp16")]; - tensor var_2514_to_fp16 = const()[name = tensor("op_2514_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2515_cast_fp16 = mul(x = linear_128_cast_fp16, y = var_2514_to_fp16)[name = tensor("op_2515_cast_fp16")]; - tensor input_759_cast_fp16 = add(x = input_747_cast_fp16, y = var_2515_cast_fp16)[name = tensor("input_759_cast_fp16")]; - tensor query_29_axes_0 = const()[name = tensor("query_29_axes_0"), val = tensor([-1])]; - tensor module_layers_14_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_14_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(267349632)))]; - tensor module_layers_14_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_14_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(267351744)))]; - tensor query_29_cast_fp16 = layer_norm(axes = query_29_axes_0, beta = module_layers_14_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_14_norm_self_att_weight_to_fp16, x = input_759_cast_fp16)[name = tensor("query_29_cast_fp16")]; - tensor module_layers_14_self_attn_linear_q_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(267353856))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(268140352))), name = tensor("module_layers_14_self_attn_linear_q_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_129_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_14_self_attn_linear_q_weight_to_fp16_palettized, x = query_29_cast_fp16)[name = tensor("linear_129_cast_fp16")]; - tensor var_2531 = const()[name = tensor("op_2531"), val = tensor([1, -1, 8, 128])]; - tensor q_85_cast_fp16 = reshape(shape = var_2531, x = linear_129_cast_fp16)[name = tensor("q_85_cast_fp16")]; - tensor module_layers_14_self_attn_linear_k_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(268140544))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(268927040))), name = tensor("module_layers_14_self_attn_linear_k_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_130_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_14_self_attn_linear_k_weight_to_fp16_palettized, x = query_29_cast_fp16)[name = tensor("linear_130_cast_fp16")]; - tensor var_2535 = const()[name = tensor("op_2535"), val = tensor([1, -1, 8, 128])]; - tensor k_57_cast_fp16 = reshape(shape = var_2535, x = linear_130_cast_fp16)[name = tensor("k_57_cast_fp16")]; - tensor module_layers_14_self_attn_linear_v_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(268927232))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(269713728))), name = tensor("module_layers_14_self_attn_linear_v_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_131_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_14_self_attn_linear_v_weight_to_fp16_palettized, x = query_29_cast_fp16)[name = tensor("linear_131_cast_fp16")]; - tensor var_2539 = const()[name = tensor("op_2539"), val = tensor([1, -1, 8, 128])]; - tensor v_29_cast_fp16 = reshape(shape = var_2539, x = linear_131_cast_fp16)[name = tensor("v_29_cast_fp16")]; - tensor value_31_perm_0 = const()[name = tensor("value_31_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor module_layers_14_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_14_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(269713920)))]; - tensor var_2551_cast_fp16 = add(x = q_85_cast_fp16, y = module_layers_14_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2551_cast_fp16")]; - tensor module_layers_14_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_14_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(269716032)))]; - tensor var_2553_cast_fp16 = add(x = q_85_cast_fp16, y = module_layers_14_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2553_cast_fp16")]; - tensor q_with_bias_v_29_perm_0 = const()[name = tensor("q_with_bias_v_29_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor x_315_transpose_x_0 = const()[name = tensor("x_315_transpose_x_0"), val = tensor(false)]; - tensor x_315_transpose_y_0 = const()[name = tensor("x_315_transpose_y_0"), val = tensor(false)]; - tensor op_2555_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(269718144))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(270006208))), name = tensor("op_2555_to_fp16_palettized"), shape = tensor([1, 8, 128, 375])]; - tensor q_with_bias_v_29_cast_fp16 = transpose(perm = q_with_bias_v_29_perm_0, x = var_2553_cast_fp16)[name = tensor("transpose_214")]; - tensor x_315_cast_fp16 = matmul(transpose_x = x_315_transpose_x_0, transpose_y = x_315_transpose_y_0, x = q_with_bias_v_29_cast_fp16, y = op_2555_to_fp16_palettized)[name = tensor("x_315_cast_fp16")]; - tensor x_317_pad_0 = const()[name = tensor("x_317_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_317_mode_0 = const()[name = tensor("x_317_mode_0"), val = tensor("constant")]; - tensor const_154_to_fp16 = const()[name = tensor("const_154_to_fp16"), val = tensor(0x0p+0)]; - tensor x_317_cast_fp16 = pad(constant_val = const_154_to_fp16, mode = x_317_mode_0, pad = x_317_pad_0, x = x_315_cast_fp16)[name = tensor("x_317_cast_fp16")]; - tensor var_2563 = const()[name = tensor("op_2563"), val = tensor([1, 8, -1, 188])]; - tensor x_319_cast_fp16 = reshape(shape = var_2563, x = x_317_cast_fp16)[name = tensor("x_319_cast_fp16")]; - tensor var_2567_begin_0 = const()[name = tensor("op_2567_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_2567_end_0 = const()[name = tensor("op_2567_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_2567_end_mask_0 = const()[name = tensor("op_2567_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_2567_cast_fp16 = slice_by_index(begin = var_2567_begin_0, end = var_2567_end_0, end_mask = var_2567_end_mask_0, x = x_319_cast_fp16)[name = tensor("op_2567_cast_fp16")]; - tensor var_2568 = const()[name = tensor("op_2568"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_57_cast_fp16 = reshape(shape = var_2568, x = var_2567_cast_fp16)[name = tensor("matrix_bd_57_cast_fp16")]; - tensor matrix_ac_29_transpose_x_0 = const()[name = tensor("matrix_ac_29_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_29_transpose_y_0 = const()[name = tensor("matrix_ac_29_transpose_y_0"), val = tensor(false)]; - tensor transpose_124_perm_0 = const()[name = tensor("transpose_124_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_125_perm_0 = const()[name = tensor("transpose_125_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_125 = transpose(perm = transpose_125_perm_0, x = k_57_cast_fp16)[name = tensor("transpose_212")]; - tensor transpose_124 = transpose(perm = transpose_124_perm_0, x = var_2551_cast_fp16)[name = tensor("transpose_213")]; - tensor matrix_ac_29_cast_fp16 = matmul(transpose_x = matrix_ac_29_transpose_x_0, transpose_y = matrix_ac_29_transpose_y_0, x = transpose_124, y = transpose_125)[name = tensor("matrix_ac_29_cast_fp16")]; - tensor matrix_bd_59_begin_0 = const()[name = tensor("matrix_bd_59_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_59_end_0 = const()[name = tensor("matrix_bd_59_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_59_end_mask_0 = const()[name = tensor("matrix_bd_59_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_59_cast_fp16 = slice_by_index(begin = matrix_bd_59_begin_0, end = matrix_bd_59_end_0, end_mask = matrix_bd_59_end_mask_0, x = matrix_bd_57_cast_fp16)[name = tensor("matrix_bd_59_cast_fp16")]; - tensor var_2577_cast_fp16 = add(x = matrix_ac_29_cast_fp16, y = matrix_bd_59_cast_fp16)[name = tensor("op_2577_cast_fp16")]; - tensor _inversed_scores_57_y_0_to_fp16 = const()[name = tensor("_inversed_scores_57_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; - tensor _inversed_scores_57_cast_fp16 = mul(x = var_2577_cast_fp16, y = _inversed_scores_57_y_0_to_fp16)[name = tensor("_inversed_scores_57_cast_fp16")]; - tensor scores_59_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_57_cast_fp16, cond = mask_3)[name = tensor("scores_59_cast_fp16")]; - tensor var_2583_cast_fp16 = softmax(axis = var_30, x = scores_59_cast_fp16)[name = tensor("op_2583_cast_fp16")]; - tensor input_761_cast_fp16 = select(a = var_11_to_fp16, b = var_2583_cast_fp16, cond = mask_3)[name = tensor("input_761_cast_fp16")]; - tensor x_321_transpose_x_0 = const()[name = tensor("x_321_transpose_x_0"), val = tensor(false)]; - tensor x_321_transpose_y_0 = const()[name = tensor("x_321_transpose_y_0"), val = tensor(false)]; - tensor value_31_cast_fp16 = transpose(perm = value_31_perm_0, x = v_29_cast_fp16)[name = tensor("transpose_211")]; - tensor x_321_cast_fp16 = matmul(transpose_x = x_321_transpose_x_0, transpose_y = x_321_transpose_y_0, x = input_761_cast_fp16, y = value_31_cast_fp16)[name = tensor("x_321_cast_fp16")]; - tensor var_2587_perm_0 = const()[name = tensor("op_2587_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_2588 = const()[name = tensor("op_2588"), val = tensor([1, -1, 1024])]; - tensor var_2587_cast_fp16 = transpose(perm = var_2587_perm_0, x = x_321_cast_fp16)[name = tensor("transpose_210")]; - tensor input_763_cast_fp16 = reshape(shape = var_2588, x = var_2587_cast_fp16)[name = tensor("input_763_cast_fp16")]; - tensor module_layers_14_self_attn_linear_out_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(270006400))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(270792896))), name = tensor("module_layers_14_self_attn_linear_out_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_133_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_14_self_attn_linear_out_weight_to_fp16_palettized, x = input_763_cast_fp16)[name = tensor("linear_133_cast_fp16")]; - tensor input_767_cast_fp16 = add(x = input_759_cast_fp16, y = linear_133_cast_fp16)[name = tensor("input_767_cast_fp16")]; - tensor x_325_axes_0 = const()[name = tensor("x_325_axes_0"), val = tensor([-1])]; - tensor module_layers_14_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_14_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(270793088)))]; - tensor module_layers_14_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_14_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(270795200)))]; - tensor x_325_cast_fp16 = layer_norm(axes = x_325_axes_0, beta = module_layers_14_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_14_norm_conv_weight_to_fp16, x = input_767_cast_fp16)[name = tensor("x_325_cast_fp16")]; - tensor input_769_perm_0 = const()[name = tensor("input_769_perm_0"), val = tensor([0, 2, 1])]; - tensor input_771_pad_type_0 = const()[name = tensor("input_771_pad_type_0"), val = tensor("valid")]; - tensor input_771_strides_0 = const()[name = tensor("input_771_strides_0"), val = tensor([1])]; - tensor input_771_pad_0 = const()[name = tensor("input_771_pad_0"), val = tensor([0, 0])]; - tensor input_771_dilations_0 = const()[name = tensor("input_771_dilations_0"), val = tensor([1])]; - tensor input_771_groups_0 = const()[name = tensor("input_771_groups_0"), val = tensor(1)]; - tensor module_layers_14_conv_pointwise_conv1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(270797312))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(272370240))), name = tensor("module_layers_14_conv_pointwise_conv1_weight_to_fp16_palettized"), shape = tensor([2048, 1024, 1])]; - tensor input_769_cast_fp16 = transpose(perm = input_769_perm_0, x = x_325_cast_fp16)[name = tensor("transpose_209")]; - tensor input_771_cast_fp16 = conv(dilations = input_771_dilations_0, groups = input_771_groups_0, pad = input_771_pad_0, pad_type = input_771_pad_type_0, strides = input_771_strides_0, weight = module_layers_14_conv_pointwise_conv1_weight_to_fp16_palettized, x = input_769_cast_fp16)[name = tensor("input_771_cast_fp16")]; - tensor x_327_split_num_splits_0 = const()[name = tensor("x_327_split_num_splits_0"), val = tensor(2)]; - tensor x_327_split_axis_0 = const()[name = tensor("x_327_split_axis_0"), val = tensor(1)]; - tensor x_327_split_cast_fp16_0, tensor x_327_split_cast_fp16_1 = split(axis = x_327_split_axis_0, num_splits = x_327_split_num_splits_0, x = input_771_cast_fp16)[name = tensor("x_327_split_cast_fp16")]; - tensor x_327_split_1_sigmoid_cast_fp16 = sigmoid(x = x_327_split_cast_fp16_1)[name = tensor("x_327_split_1_sigmoid_cast_fp16")]; - tensor x_327_cast_fp16 = mul(x = x_327_split_cast_fp16_0, y = x_327_split_1_sigmoid_cast_fp16)[name = tensor("x_327_cast_fp16")]; - tensor input_773_cast_fp16 = select(a = var_11_to_fp16, b = x_327_cast_fp16, cond = var_328)[name = tensor("input_773_cast_fp16")]; - tensor input_775_pad_0 = const()[name = tensor("input_775_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_775_mode_0 = const()[name = tensor("input_775_mode_0"), val = tensor("constant")]; - tensor const_157_to_fp16 = const()[name = tensor("const_157_to_fp16"), val = tensor(0x0p+0)]; - tensor input_775_cast_fp16 = pad(constant_val = const_157_to_fp16, mode = input_775_mode_0, pad = input_775_pad_0, x = input_773_cast_fp16)[name = tensor("input_775_cast_fp16")]; - tensor input_777_pad_type_0 = const()[name = tensor("input_777_pad_type_0"), val = tensor("valid")]; - tensor input_777_groups_0 = const()[name = tensor("input_777_groups_0"), val = tensor(1024)]; - tensor input_777_strides_0 = const()[name = tensor("input_777_strides_0"), val = tensor([1])]; - tensor input_777_pad_0 = const()[name = tensor("input_777_pad_0"), val = tensor([0, 0])]; - tensor input_777_dilations_0 = const()[name = tensor("input_777_dilations_0"), val = tensor([1])]; - tensor const_276_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(272370432))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(272377408))), name = tensor("const_276_to_fp16_palettized"), shape = tensor([1024, 1, 9])]; - tensor const_277_to_fp16 = const()[name = tensor("const_277_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(272377600)))]; - tensor input_779_cast_fp16 = conv(bias = const_277_to_fp16, dilations = input_777_dilations_0, groups = input_777_groups_0, pad = input_777_pad_0, pad_type = input_777_pad_type_0, strides = input_777_strides_0, weight = const_276_to_fp16_palettized, x = input_775_cast_fp16)[name = tensor("input_779_cast_fp16")]; - tensor input_781_cast_fp16 = silu(x = input_779_cast_fp16)[name = tensor("input_781_cast_fp16")]; - tensor x_329_pad_type_0 = const()[name = tensor("x_329_pad_type_0"), val = tensor("valid")]; - tensor x_329_strides_0 = const()[name = tensor("x_329_strides_0"), val = tensor([1])]; - tensor x_329_pad_0 = const()[name = tensor("x_329_pad_0"), val = tensor([0, 0])]; - tensor x_329_dilations_0 = const()[name = tensor("x_329_dilations_0"), val = tensor([1])]; - tensor x_329_groups_0 = const()[name = tensor("x_329_groups_0"), val = tensor(1)]; - tensor module_layers_14_conv_pointwise_conv2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(272379712))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(273166208))), name = tensor("module_layers_14_conv_pointwise_conv2_weight_to_fp16_palettized"), shape = tensor([1024, 1024, 1])]; - tensor x_329_cast_fp16 = conv(dilations = x_329_dilations_0, groups = x_329_groups_0, pad = x_329_pad_0, pad_type = x_329_pad_type_0, strides = x_329_strides_0, weight = module_layers_14_conv_pointwise_conv2_weight_to_fp16_palettized, x = input_781_cast_fp16)[name = tensor("x_329_cast_fp16")]; - tensor input_783_perm_0 = const()[name = tensor("input_783_perm_0"), val = tensor([0, 2, 1])]; - tensor input_783_cast_fp16 = transpose(perm = input_783_perm_0, x = x_329_cast_fp16)[name = tensor("transpose_208")]; - tensor input_785_cast_fp16 = add(x = input_767_cast_fp16, y = input_783_cast_fp16)[name = tensor("input_785_cast_fp16")]; - tensor input_787_axes_0 = const()[name = tensor("input_787_axes_0"), val = tensor([-1])]; - tensor module_layers_14_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_14_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(273166400)))]; - tensor module_layers_14_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_14_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(273168512)))]; - tensor input_787_cast_fp16 = layer_norm(axes = input_787_axes_0, beta = module_layers_14_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_14_norm_feed_forward2_weight_to_fp16, x = input_785_cast_fp16)[name = tensor("input_787_cast_fp16")]; - tensor module_layers_14_feed_forward2_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(273170624))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(276316416))), name = tensor("module_layers_14_feed_forward2_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_134_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_14_feed_forward2_linear1_weight_to_fp16_palettized, x = input_787_cast_fp16)[name = tensor("linear_134_cast_fp16")]; - tensor input_791_cast_fp16 = silu(x = linear_134_cast_fp16)[name = tensor("input_791_cast_fp16")]; - tensor module_layers_14_feed_forward2_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(276316608))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(279462400))), name = tensor("module_layers_14_feed_forward2_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_135_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_14_feed_forward2_linear2_weight_to_fp16_palettized, x = input_791_cast_fp16)[name = tensor("linear_135_cast_fp16")]; - tensor var_2648_to_fp16 = const()[name = tensor("op_2648_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2649_cast_fp16 = mul(x = linear_135_cast_fp16, y = var_2648_to_fp16)[name = tensor("op_2649_cast_fp16")]; - tensor input_797_cast_fp16 = add(x = input_785_cast_fp16, y = var_2649_cast_fp16)[name = tensor("input_797_cast_fp16")]; - tensor input_799_axes_0 = const()[name = tensor("input_799_axes_0"), val = tensor([-1])]; - tensor module_layers_14_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_14_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(279462592)))]; - tensor module_layers_14_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_14_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(279464704)))]; - tensor input_799_cast_fp16 = layer_norm(axes = input_799_axes_0, beta = module_layers_14_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_14_norm_out_weight_to_fp16, x = input_797_cast_fp16)[name = tensor("input_799_cast_fp16")]; - tensor input_801_axes_0 = const()[name = tensor("input_801_axes_0"), val = tensor([-1])]; - tensor module_layers_15_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_15_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(279466816)))]; - tensor module_layers_15_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_15_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(279468928)))]; - tensor input_801_cast_fp16 = layer_norm(axes = input_801_axes_0, beta = module_layers_15_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_15_norm_feed_forward1_weight_to_fp16, x = input_799_cast_fp16)[name = tensor("input_801_cast_fp16")]; - tensor module_layers_15_feed_forward1_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(279471040))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(282616832))), name = tensor("module_layers_15_feed_forward1_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_136_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_15_feed_forward1_linear1_weight_to_fp16_palettized, x = input_801_cast_fp16)[name = tensor("linear_136_cast_fp16")]; - tensor input_805_cast_fp16 = silu(x = linear_136_cast_fp16)[name = tensor("input_805_cast_fp16")]; - tensor module_layers_15_feed_forward1_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(282617024))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(285762816))), name = tensor("module_layers_15_feed_forward1_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_137_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_15_feed_forward1_linear2_weight_to_fp16_palettized, x = input_805_cast_fp16)[name = tensor("linear_137_cast_fp16")]; - tensor var_2677_to_fp16 = const()[name = tensor("op_2677_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2678_cast_fp16 = mul(x = linear_137_cast_fp16, y = var_2677_to_fp16)[name = tensor("op_2678_cast_fp16")]; - tensor input_811_cast_fp16 = add(x = input_799_cast_fp16, y = var_2678_cast_fp16)[name = tensor("input_811_cast_fp16")]; - tensor query_31_axes_0 = const()[name = tensor("query_31_axes_0"), val = tensor([-1])]; - tensor module_layers_15_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_15_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(285763008)))]; - tensor module_layers_15_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_15_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(285765120)))]; - tensor query_31_cast_fp16 = layer_norm(axes = query_31_axes_0, beta = module_layers_15_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_15_norm_self_att_weight_to_fp16, x = input_811_cast_fp16)[name = tensor("query_31_cast_fp16")]; - tensor module_layers_15_self_attn_linear_q_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(285767232))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(286553728))), name = tensor("module_layers_15_self_attn_linear_q_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_138_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_15_self_attn_linear_q_weight_to_fp16_palettized, x = query_31_cast_fp16)[name = tensor("linear_138_cast_fp16")]; - tensor var_2694 = const()[name = tensor("op_2694"), val = tensor([1, -1, 8, 128])]; - tensor q_91_cast_fp16 = reshape(shape = var_2694, x = linear_138_cast_fp16)[name = tensor("q_91_cast_fp16")]; - tensor module_layers_15_self_attn_linear_k_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(286553920))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(287340416))), name = tensor("module_layers_15_self_attn_linear_k_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_139_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_15_self_attn_linear_k_weight_to_fp16_palettized, x = query_31_cast_fp16)[name = tensor("linear_139_cast_fp16")]; - tensor var_2698 = const()[name = tensor("op_2698"), val = tensor([1, -1, 8, 128])]; - tensor k_61_cast_fp16 = reshape(shape = var_2698, x = linear_139_cast_fp16)[name = tensor("k_61_cast_fp16")]; - tensor module_layers_15_self_attn_linear_v_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(287340608))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(288127104))), name = tensor("module_layers_15_self_attn_linear_v_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_140_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_15_self_attn_linear_v_weight_to_fp16_palettized, x = query_31_cast_fp16)[name = tensor("linear_140_cast_fp16")]; - tensor var_2702 = const()[name = tensor("op_2702"), val = tensor([1, -1, 8, 128])]; - tensor v_31_cast_fp16 = reshape(shape = var_2702, x = linear_140_cast_fp16)[name = tensor("v_31_cast_fp16")]; - tensor value_33_perm_0 = const()[name = tensor("value_33_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor module_layers_15_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_15_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(288127296)))]; - tensor var_2714_cast_fp16 = add(x = q_91_cast_fp16, y = module_layers_15_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2714_cast_fp16")]; - tensor module_layers_15_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_15_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(288129408)))]; - tensor var_2716_cast_fp16 = add(x = q_91_cast_fp16, y = module_layers_15_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2716_cast_fp16")]; - tensor q_with_bias_v_31_perm_0 = const()[name = tensor("q_with_bias_v_31_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor x_337_transpose_x_0 = const()[name = tensor("x_337_transpose_x_0"), val = tensor(false)]; - tensor x_337_transpose_y_0 = const()[name = tensor("x_337_transpose_y_0"), val = tensor(false)]; - tensor op_2718_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(288131520))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(288419584))), name = tensor("op_2718_to_fp16_palettized"), shape = tensor([1, 8, 128, 375])]; - tensor q_with_bias_v_31_cast_fp16 = transpose(perm = q_with_bias_v_31_perm_0, x = var_2716_cast_fp16)[name = tensor("transpose_207")]; - tensor x_337_cast_fp16 = matmul(transpose_x = x_337_transpose_x_0, transpose_y = x_337_transpose_y_0, x = q_with_bias_v_31_cast_fp16, y = op_2718_to_fp16_palettized)[name = tensor("x_337_cast_fp16")]; - tensor x_339_pad_0 = const()[name = tensor("x_339_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_339_mode_0 = const()[name = tensor("x_339_mode_0"), val = tensor("constant")]; - tensor const_164_to_fp16 = const()[name = tensor("const_164_to_fp16"), val = tensor(0x0p+0)]; - tensor x_339_cast_fp16 = pad(constant_val = const_164_to_fp16, mode = x_339_mode_0, pad = x_339_pad_0, x = x_337_cast_fp16)[name = tensor("x_339_cast_fp16")]; - tensor var_2726 = const()[name = tensor("op_2726"), val = tensor([1, 8, -1, 188])]; - tensor x_341_cast_fp16 = reshape(shape = var_2726, x = x_339_cast_fp16)[name = tensor("x_341_cast_fp16")]; - tensor var_2730_begin_0 = const()[name = tensor("op_2730_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_2730_end_0 = const()[name = tensor("op_2730_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_2730_end_mask_0 = const()[name = tensor("op_2730_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_2730_cast_fp16 = slice_by_index(begin = var_2730_begin_0, end = var_2730_end_0, end_mask = var_2730_end_mask_0, x = x_341_cast_fp16)[name = tensor("op_2730_cast_fp16")]; - tensor var_2731 = const()[name = tensor("op_2731"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_61_cast_fp16 = reshape(shape = var_2731, x = var_2730_cast_fp16)[name = tensor("matrix_bd_61_cast_fp16")]; - tensor matrix_ac_31_transpose_x_0 = const()[name = tensor("matrix_ac_31_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_31_transpose_y_0 = const()[name = tensor("matrix_ac_31_transpose_y_0"), val = tensor(false)]; - tensor transpose_126_perm_0 = const()[name = tensor("transpose_126_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_127_perm_0 = const()[name = tensor("transpose_127_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_127 = transpose(perm = transpose_127_perm_0, x = k_61_cast_fp16)[name = tensor("transpose_205")]; - tensor transpose_126 = transpose(perm = transpose_126_perm_0, x = var_2714_cast_fp16)[name = tensor("transpose_206")]; - tensor matrix_ac_31_cast_fp16 = matmul(transpose_x = matrix_ac_31_transpose_x_0, transpose_y = matrix_ac_31_transpose_y_0, x = transpose_126, y = transpose_127)[name = tensor("matrix_ac_31_cast_fp16")]; - tensor matrix_bd_63_begin_0 = const()[name = tensor("matrix_bd_63_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_63_end_0 = const()[name = tensor("matrix_bd_63_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_63_end_mask_0 = const()[name = tensor("matrix_bd_63_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_63_cast_fp16 = slice_by_index(begin = matrix_bd_63_begin_0, end = matrix_bd_63_end_0, end_mask = matrix_bd_63_end_mask_0, x = matrix_bd_61_cast_fp16)[name = tensor("matrix_bd_63_cast_fp16")]; - tensor var_2740_cast_fp16 = add(x = matrix_ac_31_cast_fp16, y = matrix_bd_63_cast_fp16)[name = tensor("op_2740_cast_fp16")]; - tensor _inversed_scores_61_y_0_to_fp16 = const()[name = tensor("_inversed_scores_61_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; - tensor _inversed_scores_61_cast_fp16 = mul(x = var_2740_cast_fp16, y = _inversed_scores_61_y_0_to_fp16)[name = tensor("_inversed_scores_61_cast_fp16")]; - tensor scores_63_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_61_cast_fp16, cond = mask_3)[name = tensor("scores_63_cast_fp16")]; - tensor var_2746_cast_fp16 = softmax(axis = var_30, x = scores_63_cast_fp16)[name = tensor("op_2746_cast_fp16")]; - tensor input_813_cast_fp16 = select(a = var_11_to_fp16, b = var_2746_cast_fp16, cond = mask_3)[name = tensor("input_813_cast_fp16")]; - tensor x_343_transpose_x_0 = const()[name = tensor("x_343_transpose_x_0"), val = tensor(false)]; - tensor x_343_transpose_y_0 = const()[name = tensor("x_343_transpose_y_0"), val = tensor(false)]; - tensor value_33_cast_fp16 = transpose(perm = value_33_perm_0, x = v_31_cast_fp16)[name = tensor("transpose_204")]; - tensor x_343_cast_fp16 = matmul(transpose_x = x_343_transpose_x_0, transpose_y = x_343_transpose_y_0, x = input_813_cast_fp16, y = value_33_cast_fp16)[name = tensor("x_343_cast_fp16")]; - tensor var_2750_perm_0 = const()[name = tensor("op_2750_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_2751 = const()[name = tensor("op_2751"), val = tensor([1, -1, 1024])]; - tensor var_2750_cast_fp16 = transpose(perm = var_2750_perm_0, x = x_343_cast_fp16)[name = tensor("transpose_203")]; - tensor input_815_cast_fp16 = reshape(shape = var_2751, x = var_2750_cast_fp16)[name = tensor("input_815_cast_fp16")]; - tensor module_layers_15_self_attn_linear_out_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(288419776))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(289206272))), name = tensor("module_layers_15_self_attn_linear_out_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_142_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_15_self_attn_linear_out_weight_to_fp16_palettized, x = input_815_cast_fp16)[name = tensor("linear_142_cast_fp16")]; - tensor input_819_cast_fp16 = add(x = input_811_cast_fp16, y = linear_142_cast_fp16)[name = tensor("input_819_cast_fp16")]; - tensor x_347_axes_0 = const()[name = tensor("x_347_axes_0"), val = tensor([-1])]; - tensor module_layers_15_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_15_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(289206464)))]; - tensor module_layers_15_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_15_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(289208576)))]; - tensor x_347_cast_fp16 = layer_norm(axes = x_347_axes_0, beta = module_layers_15_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_15_norm_conv_weight_to_fp16, x = input_819_cast_fp16)[name = tensor("x_347_cast_fp16")]; - tensor input_821_perm_0 = const()[name = tensor("input_821_perm_0"), val = tensor([0, 2, 1])]; - tensor input_823_pad_type_0 = const()[name = tensor("input_823_pad_type_0"), val = tensor("valid")]; - tensor input_823_strides_0 = const()[name = tensor("input_823_strides_0"), val = tensor([1])]; - tensor input_823_pad_0 = const()[name = tensor("input_823_pad_0"), val = tensor([0, 0])]; - tensor input_823_dilations_0 = const()[name = tensor("input_823_dilations_0"), val = tensor([1])]; - tensor input_823_groups_0 = const()[name = tensor("input_823_groups_0"), val = tensor(1)]; - tensor module_layers_15_conv_pointwise_conv1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(289210688))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(290783616))), name = tensor("module_layers_15_conv_pointwise_conv1_weight_to_fp16_palettized"), shape = tensor([2048, 1024, 1])]; - tensor input_821_cast_fp16 = transpose(perm = input_821_perm_0, x = x_347_cast_fp16)[name = tensor("transpose_202")]; - tensor input_823_cast_fp16 = conv(dilations = input_823_dilations_0, groups = input_823_groups_0, pad = input_823_pad_0, pad_type = input_823_pad_type_0, strides = input_823_strides_0, weight = module_layers_15_conv_pointwise_conv1_weight_to_fp16_palettized, x = input_821_cast_fp16)[name = tensor("input_823_cast_fp16")]; - tensor x_349_split_num_splits_0 = const()[name = tensor("x_349_split_num_splits_0"), val = tensor(2)]; - tensor x_349_split_axis_0 = const()[name = tensor("x_349_split_axis_0"), val = tensor(1)]; - tensor x_349_split_cast_fp16_0, tensor x_349_split_cast_fp16_1 = split(axis = x_349_split_axis_0, num_splits = x_349_split_num_splits_0, x = input_823_cast_fp16)[name = tensor("x_349_split_cast_fp16")]; - tensor x_349_split_1_sigmoid_cast_fp16 = sigmoid(x = x_349_split_cast_fp16_1)[name = tensor("x_349_split_1_sigmoid_cast_fp16")]; - tensor x_349_cast_fp16 = mul(x = x_349_split_cast_fp16_0, y = x_349_split_1_sigmoid_cast_fp16)[name = tensor("x_349_cast_fp16")]; - tensor input_825_cast_fp16 = select(a = var_11_to_fp16, b = x_349_cast_fp16, cond = var_328)[name = tensor("input_825_cast_fp16")]; - tensor input_827_pad_0 = const()[name = tensor("input_827_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_827_mode_0 = const()[name = tensor("input_827_mode_0"), val = tensor("constant")]; - tensor const_167_to_fp16 = const()[name = tensor("const_167_to_fp16"), val = tensor(0x0p+0)]; - tensor input_827_cast_fp16 = pad(constant_val = const_167_to_fp16, mode = input_827_mode_0, pad = input_827_pad_0, x = input_825_cast_fp16)[name = tensor("input_827_cast_fp16")]; - tensor input_829_pad_type_0 = const()[name = tensor("input_829_pad_type_0"), val = tensor("valid")]; - tensor input_829_groups_0 = const()[name = tensor("input_829_groups_0"), val = tensor(1024)]; - tensor input_829_strides_0 = const()[name = tensor("input_829_strides_0"), val = tensor([1])]; - tensor input_829_pad_0 = const()[name = tensor("input_829_pad_0"), val = tensor([0, 0])]; - tensor input_829_dilations_0 = const()[name = tensor("input_829_dilations_0"), val = tensor([1])]; - tensor const_278_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(290783808))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(290790784))), name = tensor("const_278_to_fp16_palettized"), shape = tensor([1024, 1, 9])]; - tensor const_279_to_fp16 = const()[name = tensor("const_279_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(290790976)))]; - tensor input_831_cast_fp16 = conv(bias = const_279_to_fp16, dilations = input_829_dilations_0, groups = input_829_groups_0, pad = input_829_pad_0, pad_type = input_829_pad_type_0, strides = input_829_strides_0, weight = const_278_to_fp16_palettized, x = input_827_cast_fp16)[name = tensor("input_831_cast_fp16")]; - tensor input_833_cast_fp16 = silu(x = input_831_cast_fp16)[name = tensor("input_833_cast_fp16")]; - tensor x_351_pad_type_0 = const()[name = tensor("x_351_pad_type_0"), val = tensor("valid")]; - tensor x_351_strides_0 = const()[name = tensor("x_351_strides_0"), val = tensor([1])]; - tensor x_351_pad_0 = const()[name = tensor("x_351_pad_0"), val = tensor([0, 0])]; - tensor x_351_dilations_0 = const()[name = tensor("x_351_dilations_0"), val = tensor([1])]; - tensor x_351_groups_0 = const()[name = tensor("x_351_groups_0"), val = tensor(1)]; - tensor module_layers_15_conv_pointwise_conv2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(290793088))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(291579584))), name = tensor("module_layers_15_conv_pointwise_conv2_weight_to_fp16_palettized"), shape = tensor([1024, 1024, 1])]; - tensor x_351_cast_fp16 = conv(dilations = x_351_dilations_0, groups = x_351_groups_0, pad = x_351_pad_0, pad_type = x_351_pad_type_0, strides = x_351_strides_0, weight = module_layers_15_conv_pointwise_conv2_weight_to_fp16_palettized, x = input_833_cast_fp16)[name = tensor("x_351_cast_fp16")]; - tensor input_835_perm_0 = const()[name = tensor("input_835_perm_0"), val = tensor([0, 2, 1])]; - tensor input_835_cast_fp16 = transpose(perm = input_835_perm_0, x = x_351_cast_fp16)[name = tensor("transpose_201")]; - tensor input_837_cast_fp16 = add(x = input_819_cast_fp16, y = input_835_cast_fp16)[name = tensor("input_837_cast_fp16")]; - tensor input_839_axes_0 = const()[name = tensor("input_839_axes_0"), val = tensor([-1])]; - tensor module_layers_15_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_15_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(291579776)))]; - tensor module_layers_15_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_15_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(291581888)))]; - tensor input_839_cast_fp16 = layer_norm(axes = input_839_axes_0, beta = module_layers_15_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_15_norm_feed_forward2_weight_to_fp16, x = input_837_cast_fp16)[name = tensor("input_839_cast_fp16")]; - tensor module_layers_15_feed_forward2_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(291584000))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(294729792))), name = tensor("module_layers_15_feed_forward2_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_143_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_15_feed_forward2_linear1_weight_to_fp16_palettized, x = input_839_cast_fp16)[name = tensor("linear_143_cast_fp16")]; - tensor input_843_cast_fp16 = silu(x = linear_143_cast_fp16)[name = tensor("input_843_cast_fp16")]; - tensor module_layers_15_feed_forward2_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(294729984))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(297875776))), name = tensor("module_layers_15_feed_forward2_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_144_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_15_feed_forward2_linear2_weight_to_fp16_palettized, x = input_843_cast_fp16)[name = tensor("linear_144_cast_fp16")]; - tensor var_2811_to_fp16 = const()[name = tensor("op_2811_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2812_cast_fp16 = mul(x = linear_144_cast_fp16, y = var_2811_to_fp16)[name = tensor("op_2812_cast_fp16")]; - tensor input_849_cast_fp16 = add(x = input_837_cast_fp16, y = var_2812_cast_fp16)[name = tensor("input_849_cast_fp16")]; - tensor input_851_axes_0 = const()[name = tensor("input_851_axes_0"), val = tensor([-1])]; - tensor module_layers_15_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_15_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(297875968)))]; - tensor module_layers_15_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_15_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(297878080)))]; - tensor input_851_cast_fp16 = layer_norm(axes = input_851_axes_0, beta = module_layers_15_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_15_norm_out_weight_to_fp16, x = input_849_cast_fp16)[name = tensor("input_851_cast_fp16")]; - tensor input_853_axes_0 = const()[name = tensor("input_853_axes_0"), val = tensor([-1])]; - tensor module_layers_16_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_16_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(297880192)))]; - tensor module_layers_16_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_16_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(297882304)))]; - tensor input_853_cast_fp16 = layer_norm(axes = input_853_axes_0, beta = module_layers_16_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_16_norm_feed_forward1_weight_to_fp16, x = input_851_cast_fp16)[name = tensor("input_853_cast_fp16")]; - tensor module_layers_16_feed_forward1_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(297884416))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(301030208))), name = tensor("module_layers_16_feed_forward1_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_145_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_16_feed_forward1_linear1_weight_to_fp16_palettized, x = input_853_cast_fp16)[name = tensor("linear_145_cast_fp16")]; - tensor input_857_cast_fp16 = silu(x = linear_145_cast_fp16)[name = tensor("input_857_cast_fp16")]; - tensor module_layers_16_feed_forward1_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(301030400))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(304176192))), name = tensor("module_layers_16_feed_forward1_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_146_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_16_feed_forward1_linear2_weight_to_fp16_palettized, x = input_857_cast_fp16)[name = tensor("linear_146_cast_fp16")]; - tensor var_2840_to_fp16 = const()[name = tensor("op_2840_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2841_cast_fp16 = mul(x = linear_146_cast_fp16, y = var_2840_to_fp16)[name = tensor("op_2841_cast_fp16")]; - tensor input_863_cast_fp16 = add(x = input_851_cast_fp16, y = var_2841_cast_fp16)[name = tensor("input_863_cast_fp16")]; - tensor query_33_axes_0 = const()[name = tensor("query_33_axes_0"), val = tensor([-1])]; - tensor module_layers_16_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_16_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(304176384)))]; - tensor module_layers_16_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_16_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(304178496)))]; - tensor query_33_cast_fp16 = layer_norm(axes = query_33_axes_0, beta = module_layers_16_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_16_norm_self_att_weight_to_fp16, x = input_863_cast_fp16)[name = tensor("query_33_cast_fp16")]; - tensor module_layers_16_self_attn_linear_q_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(304180608))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(304967104))), name = tensor("module_layers_16_self_attn_linear_q_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_147_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_16_self_attn_linear_q_weight_to_fp16_palettized, x = query_33_cast_fp16)[name = tensor("linear_147_cast_fp16")]; - tensor var_2857 = const()[name = tensor("op_2857"), val = tensor([1, -1, 8, 128])]; - tensor q_97_cast_fp16 = reshape(shape = var_2857, x = linear_147_cast_fp16)[name = tensor("q_97_cast_fp16")]; - tensor module_layers_16_self_attn_linear_k_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(304967296))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(305753792))), name = tensor("module_layers_16_self_attn_linear_k_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_148_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_16_self_attn_linear_k_weight_to_fp16_palettized, x = query_33_cast_fp16)[name = tensor("linear_148_cast_fp16")]; - tensor var_2861 = const()[name = tensor("op_2861"), val = tensor([1, -1, 8, 128])]; - tensor k_65_cast_fp16 = reshape(shape = var_2861, x = linear_148_cast_fp16)[name = tensor("k_65_cast_fp16")]; - tensor module_layers_16_self_attn_linear_v_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(305753984))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(306540480))), name = tensor("module_layers_16_self_attn_linear_v_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_149_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_16_self_attn_linear_v_weight_to_fp16_palettized, x = query_33_cast_fp16)[name = tensor("linear_149_cast_fp16")]; - tensor var_2865 = const()[name = tensor("op_2865"), val = tensor([1, -1, 8, 128])]; - tensor v_33_cast_fp16 = reshape(shape = var_2865, x = linear_149_cast_fp16)[name = tensor("v_33_cast_fp16")]; - tensor value_35_perm_0 = const()[name = tensor("value_35_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor module_layers_16_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_16_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(306540672)))]; - tensor var_2877_cast_fp16 = add(x = q_97_cast_fp16, y = module_layers_16_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2877_cast_fp16")]; - tensor module_layers_16_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_16_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(306542784)))]; - tensor var_2879_cast_fp16 = add(x = q_97_cast_fp16, y = module_layers_16_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2879_cast_fp16")]; - tensor q_with_bias_v_33_perm_0 = const()[name = tensor("q_with_bias_v_33_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor x_359_transpose_x_0 = const()[name = tensor("x_359_transpose_x_0"), val = tensor(false)]; - tensor x_359_transpose_y_0 = const()[name = tensor("x_359_transpose_y_0"), val = tensor(false)]; - tensor op_2881_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(306544896))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(306832960))), name = tensor("op_2881_to_fp16_palettized"), shape = tensor([1, 8, 128, 375])]; - tensor q_with_bias_v_33_cast_fp16 = transpose(perm = q_with_bias_v_33_perm_0, x = var_2879_cast_fp16)[name = tensor("transpose_200")]; - tensor x_359_cast_fp16 = matmul(transpose_x = x_359_transpose_x_0, transpose_y = x_359_transpose_y_0, x = q_with_bias_v_33_cast_fp16, y = op_2881_to_fp16_palettized)[name = tensor("x_359_cast_fp16")]; - tensor x_361_pad_0 = const()[name = tensor("x_361_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_361_mode_0 = const()[name = tensor("x_361_mode_0"), val = tensor("constant")]; - tensor const_174_to_fp16 = const()[name = tensor("const_174_to_fp16"), val = tensor(0x0p+0)]; - tensor x_361_cast_fp16 = pad(constant_val = const_174_to_fp16, mode = x_361_mode_0, pad = x_361_pad_0, x = x_359_cast_fp16)[name = tensor("x_361_cast_fp16")]; - tensor var_2889 = const()[name = tensor("op_2889"), val = tensor([1, 8, -1, 188])]; - tensor x_363_cast_fp16 = reshape(shape = var_2889, x = x_361_cast_fp16)[name = tensor("x_363_cast_fp16")]; - tensor var_2893_begin_0 = const()[name = tensor("op_2893_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_2893_end_0 = const()[name = tensor("op_2893_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_2893_end_mask_0 = const()[name = tensor("op_2893_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_2893_cast_fp16 = slice_by_index(begin = var_2893_begin_0, end = var_2893_end_0, end_mask = var_2893_end_mask_0, x = x_363_cast_fp16)[name = tensor("op_2893_cast_fp16")]; - tensor var_2894 = const()[name = tensor("op_2894"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_65_cast_fp16 = reshape(shape = var_2894, x = var_2893_cast_fp16)[name = tensor("matrix_bd_65_cast_fp16")]; - tensor matrix_ac_33_transpose_x_0 = const()[name = tensor("matrix_ac_33_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_33_transpose_y_0 = const()[name = tensor("matrix_ac_33_transpose_y_0"), val = tensor(false)]; - tensor transpose_128_perm_0 = const()[name = tensor("transpose_128_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_129_perm_0 = const()[name = tensor("transpose_129_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_129 = transpose(perm = transpose_129_perm_0, x = k_65_cast_fp16)[name = tensor("transpose_198")]; - tensor transpose_128 = transpose(perm = transpose_128_perm_0, x = var_2877_cast_fp16)[name = tensor("transpose_199")]; - tensor matrix_ac_33_cast_fp16 = matmul(transpose_x = matrix_ac_33_transpose_x_0, transpose_y = matrix_ac_33_transpose_y_0, x = transpose_128, y = transpose_129)[name = tensor("matrix_ac_33_cast_fp16")]; - tensor matrix_bd_67_begin_0 = const()[name = tensor("matrix_bd_67_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_67_end_0 = const()[name = tensor("matrix_bd_67_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_67_end_mask_0 = const()[name = tensor("matrix_bd_67_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_67_cast_fp16 = slice_by_index(begin = matrix_bd_67_begin_0, end = matrix_bd_67_end_0, end_mask = matrix_bd_67_end_mask_0, x = matrix_bd_65_cast_fp16)[name = tensor("matrix_bd_67_cast_fp16")]; - tensor var_2903_cast_fp16 = add(x = matrix_ac_33_cast_fp16, y = matrix_bd_67_cast_fp16)[name = tensor("op_2903_cast_fp16")]; - tensor _inversed_scores_65_y_0_to_fp16 = const()[name = tensor("_inversed_scores_65_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; - tensor _inversed_scores_65_cast_fp16 = mul(x = var_2903_cast_fp16, y = _inversed_scores_65_y_0_to_fp16)[name = tensor("_inversed_scores_65_cast_fp16")]; - tensor scores_67_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_65_cast_fp16, cond = mask_3)[name = tensor("scores_67_cast_fp16")]; - tensor var_2909_cast_fp16 = softmax(axis = var_30, x = scores_67_cast_fp16)[name = tensor("op_2909_cast_fp16")]; - tensor input_865_cast_fp16 = select(a = var_11_to_fp16, b = var_2909_cast_fp16, cond = mask_3)[name = tensor("input_865_cast_fp16")]; - tensor x_365_transpose_x_0 = const()[name = tensor("x_365_transpose_x_0"), val = tensor(false)]; - tensor x_365_transpose_y_0 = const()[name = tensor("x_365_transpose_y_0"), val = tensor(false)]; - tensor value_35_cast_fp16 = transpose(perm = value_35_perm_0, x = v_33_cast_fp16)[name = tensor("transpose_197")]; - tensor x_365_cast_fp16 = matmul(transpose_x = x_365_transpose_x_0, transpose_y = x_365_transpose_y_0, x = input_865_cast_fp16, y = value_35_cast_fp16)[name = tensor("x_365_cast_fp16")]; - tensor var_2913_perm_0 = const()[name = tensor("op_2913_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_2914 = const()[name = tensor("op_2914"), val = tensor([1, -1, 1024])]; - tensor var_2913_cast_fp16 = transpose(perm = var_2913_perm_0, x = x_365_cast_fp16)[name = tensor("transpose_196")]; - tensor input_867_cast_fp16 = reshape(shape = var_2914, x = var_2913_cast_fp16)[name = tensor("input_867_cast_fp16")]; - tensor module_layers_16_self_attn_linear_out_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(306833152))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(307619648))), name = tensor("module_layers_16_self_attn_linear_out_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_151_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_16_self_attn_linear_out_weight_to_fp16_palettized, x = input_867_cast_fp16)[name = tensor("linear_151_cast_fp16")]; - tensor input_871_cast_fp16 = add(x = input_863_cast_fp16, y = linear_151_cast_fp16)[name = tensor("input_871_cast_fp16")]; - tensor x_369_axes_0 = const()[name = tensor("x_369_axes_0"), val = tensor([-1])]; - tensor module_layers_16_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_16_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(307619840)))]; - tensor module_layers_16_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_16_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(307621952)))]; - tensor x_369_cast_fp16 = layer_norm(axes = x_369_axes_0, beta = module_layers_16_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_16_norm_conv_weight_to_fp16, x = input_871_cast_fp16)[name = tensor("x_369_cast_fp16")]; - tensor input_873_perm_0 = const()[name = tensor("input_873_perm_0"), val = tensor([0, 2, 1])]; - tensor input_875_pad_type_0 = const()[name = tensor("input_875_pad_type_0"), val = tensor("valid")]; - tensor input_875_strides_0 = const()[name = tensor("input_875_strides_0"), val = tensor([1])]; - tensor input_875_pad_0 = const()[name = tensor("input_875_pad_0"), val = tensor([0, 0])]; - tensor input_875_dilations_0 = const()[name = tensor("input_875_dilations_0"), val = tensor([1])]; - tensor input_875_groups_0 = const()[name = tensor("input_875_groups_0"), val = tensor(1)]; - tensor module_layers_16_conv_pointwise_conv1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(307624064))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(309196992))), name = tensor("module_layers_16_conv_pointwise_conv1_weight_to_fp16_palettized"), shape = tensor([2048, 1024, 1])]; - tensor input_873_cast_fp16 = transpose(perm = input_873_perm_0, x = x_369_cast_fp16)[name = tensor("transpose_195")]; - tensor input_875_cast_fp16 = conv(dilations = input_875_dilations_0, groups = input_875_groups_0, pad = input_875_pad_0, pad_type = input_875_pad_type_0, strides = input_875_strides_0, weight = module_layers_16_conv_pointwise_conv1_weight_to_fp16_palettized, x = input_873_cast_fp16)[name = tensor("input_875_cast_fp16")]; - tensor x_371_split_num_splits_0 = const()[name = tensor("x_371_split_num_splits_0"), val = tensor(2)]; - tensor x_371_split_axis_0 = const()[name = tensor("x_371_split_axis_0"), val = tensor(1)]; - tensor x_371_split_cast_fp16_0, tensor x_371_split_cast_fp16_1 = split(axis = x_371_split_axis_0, num_splits = x_371_split_num_splits_0, x = input_875_cast_fp16)[name = tensor("x_371_split_cast_fp16")]; - tensor x_371_split_1_sigmoid_cast_fp16 = sigmoid(x = x_371_split_cast_fp16_1)[name = tensor("x_371_split_1_sigmoid_cast_fp16")]; - tensor x_371_cast_fp16 = mul(x = x_371_split_cast_fp16_0, y = x_371_split_1_sigmoid_cast_fp16)[name = tensor("x_371_cast_fp16")]; - tensor input_877_cast_fp16 = select(a = var_11_to_fp16, b = x_371_cast_fp16, cond = var_328)[name = tensor("input_877_cast_fp16")]; - tensor input_879_pad_0 = const()[name = tensor("input_879_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_879_mode_0 = const()[name = tensor("input_879_mode_0"), val = tensor("constant")]; - tensor const_177_to_fp16 = const()[name = tensor("const_177_to_fp16"), val = tensor(0x0p+0)]; - tensor input_879_cast_fp16 = pad(constant_val = const_177_to_fp16, mode = input_879_mode_0, pad = input_879_pad_0, x = input_877_cast_fp16)[name = tensor("input_879_cast_fp16")]; - tensor input_881_pad_type_0 = const()[name = tensor("input_881_pad_type_0"), val = tensor("valid")]; - tensor input_881_groups_0 = const()[name = tensor("input_881_groups_0"), val = tensor(1024)]; - tensor input_881_strides_0 = const()[name = tensor("input_881_strides_0"), val = tensor([1])]; - tensor input_881_pad_0 = const()[name = tensor("input_881_pad_0"), val = tensor([0, 0])]; - tensor input_881_dilations_0 = const()[name = tensor("input_881_dilations_0"), val = tensor([1])]; - tensor const_280_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(309197184))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(309204160))), name = tensor("const_280_to_fp16_palettized"), shape = tensor([1024, 1, 9])]; - tensor const_281_to_fp16 = const()[name = tensor("const_281_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(309204352)))]; - tensor input_883_cast_fp16 = conv(bias = const_281_to_fp16, dilations = input_881_dilations_0, groups = input_881_groups_0, pad = input_881_pad_0, pad_type = input_881_pad_type_0, strides = input_881_strides_0, weight = const_280_to_fp16_palettized, x = input_879_cast_fp16)[name = tensor("input_883_cast_fp16")]; - tensor input_885_cast_fp16 = silu(x = input_883_cast_fp16)[name = tensor("input_885_cast_fp16")]; - tensor x_373_pad_type_0 = const()[name = tensor("x_373_pad_type_0"), val = tensor("valid")]; - tensor x_373_strides_0 = const()[name = tensor("x_373_strides_0"), val = tensor([1])]; - tensor x_373_pad_0 = const()[name = tensor("x_373_pad_0"), val = tensor([0, 0])]; - tensor x_373_dilations_0 = const()[name = tensor("x_373_dilations_0"), val = tensor([1])]; - tensor x_373_groups_0 = const()[name = tensor("x_373_groups_0"), val = tensor(1)]; - tensor module_layers_16_conv_pointwise_conv2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(309206464))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(309992960))), name = tensor("module_layers_16_conv_pointwise_conv2_weight_to_fp16_palettized"), shape = tensor([1024, 1024, 1])]; - tensor x_373_cast_fp16 = conv(dilations = x_373_dilations_0, groups = x_373_groups_0, pad = x_373_pad_0, pad_type = x_373_pad_type_0, strides = x_373_strides_0, weight = module_layers_16_conv_pointwise_conv2_weight_to_fp16_palettized, x = input_885_cast_fp16)[name = tensor("x_373_cast_fp16")]; - tensor input_887_perm_0 = const()[name = tensor("input_887_perm_0"), val = tensor([0, 2, 1])]; - tensor input_887_cast_fp16 = transpose(perm = input_887_perm_0, x = x_373_cast_fp16)[name = tensor("transpose_194")]; - tensor input_889_cast_fp16 = add(x = input_871_cast_fp16, y = input_887_cast_fp16)[name = tensor("input_889_cast_fp16")]; - tensor input_891_axes_0 = const()[name = tensor("input_891_axes_0"), val = tensor([-1])]; - tensor module_layers_16_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_16_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(309993152)))]; - tensor module_layers_16_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_16_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(309995264)))]; - tensor input_891_cast_fp16 = layer_norm(axes = input_891_axes_0, beta = module_layers_16_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_16_norm_feed_forward2_weight_to_fp16, x = input_889_cast_fp16)[name = tensor("input_891_cast_fp16")]; - tensor module_layers_16_feed_forward2_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(309997376))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(313143168))), name = tensor("module_layers_16_feed_forward2_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_152_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_16_feed_forward2_linear1_weight_to_fp16_palettized, x = input_891_cast_fp16)[name = tensor("linear_152_cast_fp16")]; - tensor input_895_cast_fp16 = silu(x = linear_152_cast_fp16)[name = tensor("input_895_cast_fp16")]; - tensor module_layers_16_feed_forward2_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(313143360))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(316289152))), name = tensor("module_layers_16_feed_forward2_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_153_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_16_feed_forward2_linear2_weight_to_fp16_palettized, x = input_895_cast_fp16)[name = tensor("linear_153_cast_fp16")]; - tensor var_2974_to_fp16 = const()[name = tensor("op_2974_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2975_cast_fp16 = mul(x = linear_153_cast_fp16, y = var_2974_to_fp16)[name = tensor("op_2975_cast_fp16")]; - tensor input_901_cast_fp16 = add(x = input_889_cast_fp16, y = var_2975_cast_fp16)[name = tensor("input_901_cast_fp16")]; - tensor input_903_axes_0 = const()[name = tensor("input_903_axes_0"), val = tensor([-1])]; - tensor module_layers_16_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_16_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(316289344)))]; - tensor module_layers_16_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_16_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(316291456)))]; - tensor input_903_cast_fp16 = layer_norm(axes = input_903_axes_0, beta = module_layers_16_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_16_norm_out_weight_to_fp16, x = input_901_cast_fp16)[name = tensor("input_903_cast_fp16")]; - tensor input_905_axes_0 = const()[name = tensor("input_905_axes_0"), val = tensor([-1])]; - tensor module_layers_17_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_17_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(316293568)))]; - tensor module_layers_17_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_17_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(316295680)))]; - tensor input_905_cast_fp16 = layer_norm(axes = input_905_axes_0, beta = module_layers_17_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_17_norm_feed_forward1_weight_to_fp16, x = input_903_cast_fp16)[name = tensor("input_905_cast_fp16")]; - tensor module_layers_17_feed_forward1_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(316297792))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(319443584))), name = tensor("module_layers_17_feed_forward1_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_154_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_17_feed_forward1_linear1_weight_to_fp16_palettized, x = input_905_cast_fp16)[name = tensor("linear_154_cast_fp16")]; - tensor input_909_cast_fp16 = silu(x = linear_154_cast_fp16)[name = tensor("input_909_cast_fp16")]; - tensor module_layers_17_feed_forward1_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(319443776))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(322589568))), name = tensor("module_layers_17_feed_forward1_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_155_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_17_feed_forward1_linear2_weight_to_fp16_palettized, x = input_909_cast_fp16)[name = tensor("linear_155_cast_fp16")]; - tensor var_3003_to_fp16 = const()[name = tensor("op_3003_to_fp16"), val = tensor(0x1p-1)]; - tensor var_3004_cast_fp16 = mul(x = linear_155_cast_fp16, y = var_3003_to_fp16)[name = tensor("op_3004_cast_fp16")]; - tensor input_915_cast_fp16 = add(x = input_903_cast_fp16, y = var_3004_cast_fp16)[name = tensor("input_915_cast_fp16")]; - tensor query_35_axes_0 = const()[name = tensor("query_35_axes_0"), val = tensor([-1])]; - tensor module_layers_17_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_17_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(322589760)))]; - tensor module_layers_17_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_17_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(322591872)))]; - tensor query_35_cast_fp16 = layer_norm(axes = query_35_axes_0, beta = module_layers_17_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_17_norm_self_att_weight_to_fp16, x = input_915_cast_fp16)[name = tensor("query_35_cast_fp16")]; - tensor module_layers_17_self_attn_linear_q_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(322593984))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(323380480))), name = tensor("module_layers_17_self_attn_linear_q_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_156_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_17_self_attn_linear_q_weight_to_fp16_palettized, x = query_35_cast_fp16)[name = tensor("linear_156_cast_fp16")]; - tensor var_3020 = const()[name = tensor("op_3020"), val = tensor([1, -1, 8, 128])]; - tensor q_103_cast_fp16 = reshape(shape = var_3020, x = linear_156_cast_fp16)[name = tensor("q_103_cast_fp16")]; - tensor module_layers_17_self_attn_linear_k_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(323380672))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(324167168))), name = tensor("module_layers_17_self_attn_linear_k_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_157_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_17_self_attn_linear_k_weight_to_fp16_palettized, x = query_35_cast_fp16)[name = tensor("linear_157_cast_fp16")]; - tensor var_3024 = const()[name = tensor("op_3024"), val = tensor([1, -1, 8, 128])]; - tensor k_69_cast_fp16 = reshape(shape = var_3024, x = linear_157_cast_fp16)[name = tensor("k_69_cast_fp16")]; - tensor module_layers_17_self_attn_linear_v_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(324167360))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(324953856))), name = tensor("module_layers_17_self_attn_linear_v_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_158_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_17_self_attn_linear_v_weight_to_fp16_palettized, x = query_35_cast_fp16)[name = tensor("linear_158_cast_fp16")]; - tensor var_3028 = const()[name = tensor("op_3028"), val = tensor([1, -1, 8, 128])]; - tensor v_35_cast_fp16 = reshape(shape = var_3028, x = linear_158_cast_fp16)[name = tensor("v_35_cast_fp16")]; - tensor value_37_perm_0 = const()[name = tensor("value_37_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor module_layers_17_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_17_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(324954048)))]; - tensor var_3040_cast_fp16 = add(x = q_103_cast_fp16, y = module_layers_17_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3040_cast_fp16")]; - tensor module_layers_17_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_17_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(324956160)))]; - tensor var_3042_cast_fp16 = add(x = q_103_cast_fp16, y = module_layers_17_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3042_cast_fp16")]; - tensor q_with_bias_v_35_perm_0 = const()[name = tensor("q_with_bias_v_35_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor x_381_transpose_x_0 = const()[name = tensor("x_381_transpose_x_0"), val = tensor(false)]; - tensor x_381_transpose_y_0 = const()[name = tensor("x_381_transpose_y_0"), val = tensor(false)]; - tensor op_3044_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(324958272))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(325246336))), name = tensor("op_3044_to_fp16_palettized"), shape = tensor([1, 8, 128, 375])]; - tensor q_with_bias_v_35_cast_fp16 = transpose(perm = q_with_bias_v_35_perm_0, x = var_3042_cast_fp16)[name = tensor("transpose_193")]; - tensor x_381_cast_fp16 = matmul(transpose_x = x_381_transpose_x_0, transpose_y = x_381_transpose_y_0, x = q_with_bias_v_35_cast_fp16, y = op_3044_to_fp16_palettized)[name = tensor("x_381_cast_fp16")]; - tensor x_383_pad_0 = const()[name = tensor("x_383_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_383_mode_0 = const()[name = tensor("x_383_mode_0"), val = tensor("constant")]; - tensor const_184_to_fp16 = const()[name = tensor("const_184_to_fp16"), val = tensor(0x0p+0)]; - tensor x_383_cast_fp16 = pad(constant_val = const_184_to_fp16, mode = x_383_mode_0, pad = x_383_pad_0, x = x_381_cast_fp16)[name = tensor("x_383_cast_fp16")]; - tensor var_3052 = const()[name = tensor("op_3052"), val = tensor([1, 8, -1, 188])]; - tensor x_385_cast_fp16 = reshape(shape = var_3052, x = x_383_cast_fp16)[name = tensor("x_385_cast_fp16")]; - tensor var_3056_begin_0 = const()[name = tensor("op_3056_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_3056_end_0 = const()[name = tensor("op_3056_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_3056_end_mask_0 = const()[name = tensor("op_3056_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_3056_cast_fp16 = slice_by_index(begin = var_3056_begin_0, end = var_3056_end_0, end_mask = var_3056_end_mask_0, x = x_385_cast_fp16)[name = tensor("op_3056_cast_fp16")]; - tensor var_3057 = const()[name = tensor("op_3057"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_69_cast_fp16 = reshape(shape = var_3057, x = var_3056_cast_fp16)[name = tensor("matrix_bd_69_cast_fp16")]; - tensor matrix_ac_35_transpose_x_0 = const()[name = tensor("matrix_ac_35_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_35_transpose_y_0 = const()[name = tensor("matrix_ac_35_transpose_y_0"), val = tensor(false)]; - tensor transpose_130_perm_0 = const()[name = tensor("transpose_130_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_131_perm_0 = const()[name = tensor("transpose_131_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_131 = transpose(perm = transpose_131_perm_0, x = k_69_cast_fp16)[name = tensor("transpose_191")]; - tensor transpose_130 = transpose(perm = transpose_130_perm_0, x = var_3040_cast_fp16)[name = tensor("transpose_192")]; - tensor matrix_ac_35_cast_fp16 = matmul(transpose_x = matrix_ac_35_transpose_x_0, transpose_y = matrix_ac_35_transpose_y_0, x = transpose_130, y = transpose_131)[name = tensor("matrix_ac_35_cast_fp16")]; - tensor matrix_bd_71_begin_0 = const()[name = tensor("matrix_bd_71_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_71_end_0 = const()[name = tensor("matrix_bd_71_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_71_end_mask_0 = const()[name = tensor("matrix_bd_71_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_71_cast_fp16 = slice_by_index(begin = matrix_bd_71_begin_0, end = matrix_bd_71_end_0, end_mask = matrix_bd_71_end_mask_0, x = matrix_bd_69_cast_fp16)[name = tensor("matrix_bd_71_cast_fp16")]; - tensor var_3066_cast_fp16 = add(x = matrix_ac_35_cast_fp16, y = matrix_bd_71_cast_fp16)[name = tensor("op_3066_cast_fp16")]; - tensor _inversed_scores_69_y_0_to_fp16 = const()[name = tensor("_inversed_scores_69_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; - tensor _inversed_scores_69_cast_fp16 = mul(x = var_3066_cast_fp16, y = _inversed_scores_69_y_0_to_fp16)[name = tensor("_inversed_scores_69_cast_fp16")]; - tensor scores_71_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_69_cast_fp16, cond = mask_3)[name = tensor("scores_71_cast_fp16")]; - tensor var_3072_cast_fp16 = softmax(axis = var_30, x = scores_71_cast_fp16)[name = tensor("op_3072_cast_fp16")]; - tensor input_917_cast_fp16 = select(a = var_11_to_fp16, b = var_3072_cast_fp16, cond = mask_3)[name = tensor("input_917_cast_fp16")]; - tensor x_387_transpose_x_0 = const()[name = tensor("x_387_transpose_x_0"), val = tensor(false)]; - tensor x_387_transpose_y_0 = const()[name = tensor("x_387_transpose_y_0"), val = tensor(false)]; - tensor value_37_cast_fp16 = transpose(perm = value_37_perm_0, x = v_35_cast_fp16)[name = tensor("transpose_190")]; - tensor x_387_cast_fp16 = matmul(transpose_x = x_387_transpose_x_0, transpose_y = x_387_transpose_y_0, x = input_917_cast_fp16, y = value_37_cast_fp16)[name = tensor("x_387_cast_fp16")]; - tensor var_3076_perm_0 = const()[name = tensor("op_3076_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_3077 = const()[name = tensor("op_3077"), val = tensor([1, -1, 1024])]; - tensor var_3076_cast_fp16 = transpose(perm = var_3076_perm_0, x = x_387_cast_fp16)[name = tensor("transpose_189")]; - tensor input_919_cast_fp16 = reshape(shape = var_3077, x = var_3076_cast_fp16)[name = tensor("input_919_cast_fp16")]; - tensor module_layers_17_self_attn_linear_out_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(325246528))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(326033024))), name = tensor("module_layers_17_self_attn_linear_out_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_160_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_17_self_attn_linear_out_weight_to_fp16_palettized, x = input_919_cast_fp16)[name = tensor("linear_160_cast_fp16")]; - tensor input_923_cast_fp16 = add(x = input_915_cast_fp16, y = linear_160_cast_fp16)[name = tensor("input_923_cast_fp16")]; - tensor x_391_axes_0 = const()[name = tensor("x_391_axes_0"), val = tensor([-1])]; - tensor module_layers_17_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_17_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(326033216)))]; - tensor module_layers_17_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_17_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(326035328)))]; - tensor x_391_cast_fp16 = layer_norm(axes = x_391_axes_0, beta = module_layers_17_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_17_norm_conv_weight_to_fp16, x = input_923_cast_fp16)[name = tensor("x_391_cast_fp16")]; - tensor input_925_perm_0 = const()[name = tensor("input_925_perm_0"), val = tensor([0, 2, 1])]; - tensor input_927_pad_type_0 = const()[name = tensor("input_927_pad_type_0"), val = tensor("valid")]; - tensor input_927_strides_0 = const()[name = tensor("input_927_strides_0"), val = tensor([1])]; - tensor input_927_pad_0 = const()[name = tensor("input_927_pad_0"), val = tensor([0, 0])]; - tensor input_927_dilations_0 = const()[name = tensor("input_927_dilations_0"), val = tensor([1])]; - tensor input_927_groups_0 = const()[name = tensor("input_927_groups_0"), val = tensor(1)]; - tensor module_layers_17_conv_pointwise_conv1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(326037440))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(327610368))), name = tensor("module_layers_17_conv_pointwise_conv1_weight_to_fp16_palettized"), shape = tensor([2048, 1024, 1])]; - tensor input_925_cast_fp16 = transpose(perm = input_925_perm_0, x = x_391_cast_fp16)[name = tensor("transpose_188")]; - tensor input_927_cast_fp16 = conv(dilations = input_927_dilations_0, groups = input_927_groups_0, pad = input_927_pad_0, pad_type = input_927_pad_type_0, strides = input_927_strides_0, weight = module_layers_17_conv_pointwise_conv1_weight_to_fp16_palettized, x = input_925_cast_fp16)[name = tensor("input_927_cast_fp16")]; - tensor x_393_split_num_splits_0 = const()[name = tensor("x_393_split_num_splits_0"), val = tensor(2)]; - tensor x_393_split_axis_0 = const()[name = tensor("x_393_split_axis_0"), val = tensor(1)]; - tensor x_393_split_cast_fp16_0, tensor x_393_split_cast_fp16_1 = split(axis = x_393_split_axis_0, num_splits = x_393_split_num_splits_0, x = input_927_cast_fp16)[name = tensor("x_393_split_cast_fp16")]; - tensor x_393_split_1_sigmoid_cast_fp16 = sigmoid(x = x_393_split_cast_fp16_1)[name = tensor("x_393_split_1_sigmoid_cast_fp16")]; - tensor x_393_cast_fp16 = mul(x = x_393_split_cast_fp16_0, y = x_393_split_1_sigmoid_cast_fp16)[name = tensor("x_393_cast_fp16")]; - tensor input_929_cast_fp16 = select(a = var_11_to_fp16, b = x_393_cast_fp16, cond = var_328)[name = tensor("input_929_cast_fp16")]; - tensor input_931_pad_0 = const()[name = tensor("input_931_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_931_mode_0 = const()[name = tensor("input_931_mode_0"), val = tensor("constant")]; - tensor const_187_to_fp16 = const()[name = tensor("const_187_to_fp16"), val = tensor(0x0p+0)]; - tensor input_931_cast_fp16 = pad(constant_val = const_187_to_fp16, mode = input_931_mode_0, pad = input_931_pad_0, x = input_929_cast_fp16)[name = tensor("input_931_cast_fp16")]; - tensor input_933_pad_type_0 = const()[name = tensor("input_933_pad_type_0"), val = tensor("valid")]; - tensor input_933_groups_0 = const()[name = tensor("input_933_groups_0"), val = tensor(1024)]; - tensor input_933_strides_0 = const()[name = tensor("input_933_strides_0"), val = tensor([1])]; - tensor input_933_pad_0 = const()[name = tensor("input_933_pad_0"), val = tensor([0, 0])]; - tensor input_933_dilations_0 = const()[name = tensor("input_933_dilations_0"), val = tensor([1])]; - tensor const_282_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(327610560))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(327617536))), name = tensor("const_282_to_fp16_palettized"), shape = tensor([1024, 1, 9])]; - tensor const_283_to_fp16 = const()[name = tensor("const_283_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(327617728)))]; - tensor input_935_cast_fp16 = conv(bias = const_283_to_fp16, dilations = input_933_dilations_0, groups = input_933_groups_0, pad = input_933_pad_0, pad_type = input_933_pad_type_0, strides = input_933_strides_0, weight = const_282_to_fp16_palettized, x = input_931_cast_fp16)[name = tensor("input_935_cast_fp16")]; - tensor input_937_cast_fp16 = silu(x = input_935_cast_fp16)[name = tensor("input_937_cast_fp16")]; - tensor x_395_pad_type_0 = const()[name = tensor("x_395_pad_type_0"), val = tensor("valid")]; - tensor x_395_strides_0 = const()[name = tensor("x_395_strides_0"), val = tensor([1])]; - tensor x_395_pad_0 = const()[name = tensor("x_395_pad_0"), val = tensor([0, 0])]; - tensor x_395_dilations_0 = const()[name = tensor("x_395_dilations_0"), val = tensor([1])]; - tensor x_395_groups_0 = const()[name = tensor("x_395_groups_0"), val = tensor(1)]; - tensor module_layers_17_conv_pointwise_conv2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(327619840))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(328406336))), name = tensor("module_layers_17_conv_pointwise_conv2_weight_to_fp16_palettized"), shape = tensor([1024, 1024, 1])]; - tensor x_395_cast_fp16 = conv(dilations = x_395_dilations_0, groups = x_395_groups_0, pad = x_395_pad_0, pad_type = x_395_pad_type_0, strides = x_395_strides_0, weight = module_layers_17_conv_pointwise_conv2_weight_to_fp16_palettized, x = input_937_cast_fp16)[name = tensor("x_395_cast_fp16")]; - tensor input_939_perm_0 = const()[name = tensor("input_939_perm_0"), val = tensor([0, 2, 1])]; - tensor input_939_cast_fp16 = transpose(perm = input_939_perm_0, x = x_395_cast_fp16)[name = tensor("transpose_187")]; - tensor input_941_cast_fp16 = add(x = input_923_cast_fp16, y = input_939_cast_fp16)[name = tensor("input_941_cast_fp16")]; - tensor input_943_axes_0 = const()[name = tensor("input_943_axes_0"), val = tensor([-1])]; - tensor module_layers_17_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_17_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(328406528)))]; - tensor module_layers_17_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_17_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(328408640)))]; - tensor input_943_cast_fp16 = layer_norm(axes = input_943_axes_0, beta = module_layers_17_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_17_norm_feed_forward2_weight_to_fp16, x = input_941_cast_fp16)[name = tensor("input_943_cast_fp16")]; - tensor module_layers_17_feed_forward2_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(328410752))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(331556544))), name = tensor("module_layers_17_feed_forward2_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_161_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_17_feed_forward2_linear1_weight_to_fp16_palettized, x = input_943_cast_fp16)[name = tensor("linear_161_cast_fp16")]; - tensor input_947_cast_fp16 = silu(x = linear_161_cast_fp16)[name = tensor("input_947_cast_fp16")]; - tensor module_layers_17_feed_forward2_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(331556736))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(334702528))), name = tensor("module_layers_17_feed_forward2_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_162_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_17_feed_forward2_linear2_weight_to_fp16_palettized, x = input_947_cast_fp16)[name = tensor("linear_162_cast_fp16")]; - tensor var_3137_to_fp16 = const()[name = tensor("op_3137_to_fp16"), val = tensor(0x1p-1)]; - tensor var_3138_cast_fp16 = mul(x = linear_162_cast_fp16, y = var_3137_to_fp16)[name = tensor("op_3138_cast_fp16")]; - tensor input_953_cast_fp16 = add(x = input_941_cast_fp16, y = var_3138_cast_fp16)[name = tensor("input_953_cast_fp16")]; - tensor input_955_axes_0 = const()[name = tensor("input_955_axes_0"), val = tensor([-1])]; - tensor module_layers_17_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_17_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(334702720)))]; - tensor module_layers_17_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_17_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(334704832)))]; - tensor input_955_cast_fp16 = layer_norm(axes = input_955_axes_0, beta = module_layers_17_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_17_norm_out_weight_to_fp16, x = input_953_cast_fp16)[name = tensor("input_955_cast_fp16")]; - tensor input_957_axes_0 = const()[name = tensor("input_957_axes_0"), val = tensor([-1])]; - tensor module_layers_18_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_18_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(334706944)))]; - tensor module_layers_18_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_18_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(334709056)))]; - tensor input_957_cast_fp16 = layer_norm(axes = input_957_axes_0, beta = module_layers_18_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_18_norm_feed_forward1_weight_to_fp16, x = input_955_cast_fp16)[name = tensor("input_957_cast_fp16")]; - tensor module_layers_18_feed_forward1_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(334711168))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(337856960))), name = tensor("module_layers_18_feed_forward1_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_163_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_18_feed_forward1_linear1_weight_to_fp16_palettized, x = input_957_cast_fp16)[name = tensor("linear_163_cast_fp16")]; - tensor input_961_cast_fp16 = silu(x = linear_163_cast_fp16)[name = tensor("input_961_cast_fp16")]; - tensor module_layers_18_feed_forward1_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(337857152))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(341002944))), name = tensor("module_layers_18_feed_forward1_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_164_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_18_feed_forward1_linear2_weight_to_fp16_palettized, x = input_961_cast_fp16)[name = tensor("linear_164_cast_fp16")]; - tensor var_3166_to_fp16 = const()[name = tensor("op_3166_to_fp16"), val = tensor(0x1p-1)]; - tensor var_3167_cast_fp16 = mul(x = linear_164_cast_fp16, y = var_3166_to_fp16)[name = tensor("op_3167_cast_fp16")]; - tensor input_967_cast_fp16 = add(x = input_955_cast_fp16, y = var_3167_cast_fp16)[name = tensor("input_967_cast_fp16")]; - tensor query_37_axes_0 = const()[name = tensor("query_37_axes_0"), val = tensor([-1])]; - tensor module_layers_18_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_18_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(341003136)))]; - tensor module_layers_18_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_18_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(341005248)))]; - tensor query_37_cast_fp16 = layer_norm(axes = query_37_axes_0, beta = module_layers_18_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_18_norm_self_att_weight_to_fp16, x = input_967_cast_fp16)[name = tensor("query_37_cast_fp16")]; - tensor module_layers_18_self_attn_linear_q_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(341007360))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(341793856))), name = tensor("module_layers_18_self_attn_linear_q_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_165_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_18_self_attn_linear_q_weight_to_fp16_palettized, x = query_37_cast_fp16)[name = tensor("linear_165_cast_fp16")]; - tensor var_3183 = const()[name = tensor("op_3183"), val = tensor([1, -1, 8, 128])]; - tensor q_109_cast_fp16 = reshape(shape = var_3183, x = linear_165_cast_fp16)[name = tensor("q_109_cast_fp16")]; - tensor module_layers_18_self_attn_linear_k_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(341794048))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(342580544))), name = tensor("module_layers_18_self_attn_linear_k_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_166_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_18_self_attn_linear_k_weight_to_fp16_palettized, x = query_37_cast_fp16)[name = tensor("linear_166_cast_fp16")]; - tensor var_3187 = const()[name = tensor("op_3187"), val = tensor([1, -1, 8, 128])]; - tensor k_73_cast_fp16 = reshape(shape = var_3187, x = linear_166_cast_fp16)[name = tensor("k_73_cast_fp16")]; - tensor module_layers_18_self_attn_linear_v_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(342580736))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(343367232))), name = tensor("module_layers_18_self_attn_linear_v_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_167_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_18_self_attn_linear_v_weight_to_fp16_palettized, x = query_37_cast_fp16)[name = tensor("linear_167_cast_fp16")]; - tensor var_3191 = const()[name = tensor("op_3191"), val = tensor([1, -1, 8, 128])]; - tensor v_37_cast_fp16 = reshape(shape = var_3191, x = linear_167_cast_fp16)[name = tensor("v_37_cast_fp16")]; - tensor value_39_perm_0 = const()[name = tensor("value_39_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor module_layers_18_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_18_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(343367424)))]; - tensor var_3203_cast_fp16 = add(x = q_109_cast_fp16, y = module_layers_18_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3203_cast_fp16")]; - tensor module_layers_18_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_18_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(343369536)))]; - tensor var_3205_cast_fp16 = add(x = q_109_cast_fp16, y = module_layers_18_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3205_cast_fp16")]; - tensor q_with_bias_v_37_perm_0 = const()[name = tensor("q_with_bias_v_37_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor x_403_transpose_x_0 = const()[name = tensor("x_403_transpose_x_0"), val = tensor(false)]; - tensor x_403_transpose_y_0 = const()[name = tensor("x_403_transpose_y_0"), val = tensor(false)]; - tensor op_3207_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(343371648))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(343659712))), name = tensor("op_3207_to_fp16_palettized"), shape = tensor([1, 8, 128, 375])]; - tensor q_with_bias_v_37_cast_fp16 = transpose(perm = q_with_bias_v_37_perm_0, x = var_3205_cast_fp16)[name = tensor("transpose_186")]; - tensor x_403_cast_fp16 = matmul(transpose_x = x_403_transpose_x_0, transpose_y = x_403_transpose_y_0, x = q_with_bias_v_37_cast_fp16, y = op_3207_to_fp16_palettized)[name = tensor("x_403_cast_fp16")]; - tensor x_405_pad_0 = const()[name = tensor("x_405_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_405_mode_0 = const()[name = tensor("x_405_mode_0"), val = tensor("constant")]; - tensor const_194_to_fp16 = const()[name = tensor("const_194_to_fp16"), val = tensor(0x0p+0)]; - tensor x_405_cast_fp16 = pad(constant_val = const_194_to_fp16, mode = x_405_mode_0, pad = x_405_pad_0, x = x_403_cast_fp16)[name = tensor("x_405_cast_fp16")]; - tensor var_3215 = const()[name = tensor("op_3215"), val = tensor([1, 8, -1, 188])]; - tensor x_407_cast_fp16 = reshape(shape = var_3215, x = x_405_cast_fp16)[name = tensor("x_407_cast_fp16")]; - tensor var_3219_begin_0 = const()[name = tensor("op_3219_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_3219_end_0 = const()[name = tensor("op_3219_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_3219_end_mask_0 = const()[name = tensor("op_3219_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_3219_cast_fp16 = slice_by_index(begin = var_3219_begin_0, end = var_3219_end_0, end_mask = var_3219_end_mask_0, x = x_407_cast_fp16)[name = tensor("op_3219_cast_fp16")]; - tensor var_3220 = const()[name = tensor("op_3220"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_73_cast_fp16 = reshape(shape = var_3220, x = var_3219_cast_fp16)[name = tensor("matrix_bd_73_cast_fp16")]; - tensor matrix_ac_37_transpose_x_0 = const()[name = tensor("matrix_ac_37_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_37_transpose_y_0 = const()[name = tensor("matrix_ac_37_transpose_y_0"), val = tensor(false)]; - tensor transpose_132_perm_0 = const()[name = tensor("transpose_132_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_133_perm_0 = const()[name = tensor("transpose_133_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_133 = transpose(perm = transpose_133_perm_0, x = k_73_cast_fp16)[name = tensor("transpose_184")]; - tensor transpose_132 = transpose(perm = transpose_132_perm_0, x = var_3203_cast_fp16)[name = tensor("transpose_185")]; - tensor matrix_ac_37_cast_fp16 = matmul(transpose_x = matrix_ac_37_transpose_x_0, transpose_y = matrix_ac_37_transpose_y_0, x = transpose_132, y = transpose_133)[name = tensor("matrix_ac_37_cast_fp16")]; - tensor matrix_bd_75_begin_0 = const()[name = tensor("matrix_bd_75_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_75_end_0 = const()[name = tensor("matrix_bd_75_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_75_end_mask_0 = const()[name = tensor("matrix_bd_75_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_75_cast_fp16 = slice_by_index(begin = matrix_bd_75_begin_0, end = matrix_bd_75_end_0, end_mask = matrix_bd_75_end_mask_0, x = matrix_bd_73_cast_fp16)[name = tensor("matrix_bd_75_cast_fp16")]; - tensor var_3229_cast_fp16 = add(x = matrix_ac_37_cast_fp16, y = matrix_bd_75_cast_fp16)[name = tensor("op_3229_cast_fp16")]; - tensor _inversed_scores_73_y_0_to_fp16 = const()[name = tensor("_inversed_scores_73_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; - tensor _inversed_scores_73_cast_fp16 = mul(x = var_3229_cast_fp16, y = _inversed_scores_73_y_0_to_fp16)[name = tensor("_inversed_scores_73_cast_fp16")]; - tensor scores_75_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_73_cast_fp16, cond = mask_3)[name = tensor("scores_75_cast_fp16")]; - tensor var_3235_cast_fp16 = softmax(axis = var_30, x = scores_75_cast_fp16)[name = tensor("op_3235_cast_fp16")]; - tensor input_969_cast_fp16 = select(a = var_11_to_fp16, b = var_3235_cast_fp16, cond = mask_3)[name = tensor("input_969_cast_fp16")]; - tensor x_409_transpose_x_0 = const()[name = tensor("x_409_transpose_x_0"), val = tensor(false)]; - tensor x_409_transpose_y_0 = const()[name = tensor("x_409_transpose_y_0"), val = tensor(false)]; - tensor value_39_cast_fp16 = transpose(perm = value_39_perm_0, x = v_37_cast_fp16)[name = tensor("transpose_183")]; - tensor x_409_cast_fp16 = matmul(transpose_x = x_409_transpose_x_0, transpose_y = x_409_transpose_y_0, x = input_969_cast_fp16, y = value_39_cast_fp16)[name = tensor("x_409_cast_fp16")]; - tensor var_3239_perm_0 = const()[name = tensor("op_3239_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_3240 = const()[name = tensor("op_3240"), val = tensor([1, -1, 1024])]; - tensor var_3239_cast_fp16 = transpose(perm = var_3239_perm_0, x = x_409_cast_fp16)[name = tensor("transpose_182")]; - tensor input_971_cast_fp16 = reshape(shape = var_3240, x = var_3239_cast_fp16)[name = tensor("input_971_cast_fp16")]; - tensor module_layers_18_self_attn_linear_out_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(343659904))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(344446400))), name = tensor("module_layers_18_self_attn_linear_out_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_169_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_18_self_attn_linear_out_weight_to_fp16_palettized, x = input_971_cast_fp16)[name = tensor("linear_169_cast_fp16")]; - tensor input_975_cast_fp16 = add(x = input_967_cast_fp16, y = linear_169_cast_fp16)[name = tensor("input_975_cast_fp16")]; - tensor x_413_axes_0 = const()[name = tensor("x_413_axes_0"), val = tensor([-1])]; - tensor module_layers_18_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_18_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(344446592)))]; - tensor module_layers_18_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_18_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(344448704)))]; - tensor x_413_cast_fp16 = layer_norm(axes = x_413_axes_0, beta = module_layers_18_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_18_norm_conv_weight_to_fp16, x = input_975_cast_fp16)[name = tensor("x_413_cast_fp16")]; - tensor input_977_perm_0 = const()[name = tensor("input_977_perm_0"), val = tensor([0, 2, 1])]; - tensor input_979_pad_type_0 = const()[name = tensor("input_979_pad_type_0"), val = tensor("valid")]; - tensor input_979_strides_0 = const()[name = tensor("input_979_strides_0"), val = tensor([1])]; - tensor input_979_pad_0 = const()[name = tensor("input_979_pad_0"), val = tensor([0, 0])]; - tensor input_979_dilations_0 = const()[name = tensor("input_979_dilations_0"), val = tensor([1])]; - tensor input_979_groups_0 = const()[name = tensor("input_979_groups_0"), val = tensor(1)]; - tensor module_layers_18_conv_pointwise_conv1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(344450816))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(346023744))), name = tensor("module_layers_18_conv_pointwise_conv1_weight_to_fp16_palettized"), shape = tensor([2048, 1024, 1])]; - tensor input_977_cast_fp16 = transpose(perm = input_977_perm_0, x = x_413_cast_fp16)[name = tensor("transpose_181")]; - tensor input_979_cast_fp16 = conv(dilations = input_979_dilations_0, groups = input_979_groups_0, pad = input_979_pad_0, pad_type = input_979_pad_type_0, strides = input_979_strides_0, weight = module_layers_18_conv_pointwise_conv1_weight_to_fp16_palettized, x = input_977_cast_fp16)[name = tensor("input_979_cast_fp16")]; - tensor x_415_split_num_splits_0 = const()[name = tensor("x_415_split_num_splits_0"), val = tensor(2)]; - tensor x_415_split_axis_0 = const()[name = tensor("x_415_split_axis_0"), val = tensor(1)]; - tensor x_415_split_cast_fp16_0, tensor x_415_split_cast_fp16_1 = split(axis = x_415_split_axis_0, num_splits = x_415_split_num_splits_0, x = input_979_cast_fp16)[name = tensor("x_415_split_cast_fp16")]; - tensor x_415_split_1_sigmoid_cast_fp16 = sigmoid(x = x_415_split_cast_fp16_1)[name = tensor("x_415_split_1_sigmoid_cast_fp16")]; - tensor x_415_cast_fp16 = mul(x = x_415_split_cast_fp16_0, y = x_415_split_1_sigmoid_cast_fp16)[name = tensor("x_415_cast_fp16")]; - tensor input_981_cast_fp16 = select(a = var_11_to_fp16, b = x_415_cast_fp16, cond = var_328)[name = tensor("input_981_cast_fp16")]; - tensor input_983_pad_0 = const()[name = tensor("input_983_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_983_mode_0 = const()[name = tensor("input_983_mode_0"), val = tensor("constant")]; - tensor const_197_to_fp16 = const()[name = tensor("const_197_to_fp16"), val = tensor(0x0p+0)]; - tensor input_983_cast_fp16 = pad(constant_val = const_197_to_fp16, mode = input_983_mode_0, pad = input_983_pad_0, x = input_981_cast_fp16)[name = tensor("input_983_cast_fp16")]; - tensor input_985_pad_type_0 = const()[name = tensor("input_985_pad_type_0"), val = tensor("valid")]; - tensor input_985_groups_0 = const()[name = tensor("input_985_groups_0"), val = tensor(1024)]; - tensor input_985_strides_0 = const()[name = tensor("input_985_strides_0"), val = tensor([1])]; - tensor input_985_pad_0 = const()[name = tensor("input_985_pad_0"), val = tensor([0, 0])]; - tensor input_985_dilations_0 = const()[name = tensor("input_985_dilations_0"), val = tensor([1])]; - tensor const_284_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(346023936))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(346030912))), name = tensor("const_284_to_fp16_palettized"), shape = tensor([1024, 1, 9])]; - tensor const_285_to_fp16 = const()[name = tensor("const_285_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(346031104)))]; - tensor input_987_cast_fp16 = conv(bias = const_285_to_fp16, dilations = input_985_dilations_0, groups = input_985_groups_0, pad = input_985_pad_0, pad_type = input_985_pad_type_0, strides = input_985_strides_0, weight = const_284_to_fp16_palettized, x = input_983_cast_fp16)[name = tensor("input_987_cast_fp16")]; - tensor input_989_cast_fp16 = silu(x = input_987_cast_fp16)[name = tensor("input_989_cast_fp16")]; - tensor x_417_pad_type_0 = const()[name = tensor("x_417_pad_type_0"), val = tensor("valid")]; - tensor x_417_strides_0 = const()[name = tensor("x_417_strides_0"), val = tensor([1])]; - tensor x_417_pad_0 = const()[name = tensor("x_417_pad_0"), val = tensor([0, 0])]; - tensor x_417_dilations_0 = const()[name = tensor("x_417_dilations_0"), val = tensor([1])]; - tensor x_417_groups_0 = const()[name = tensor("x_417_groups_0"), val = tensor(1)]; - tensor module_layers_18_conv_pointwise_conv2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(346033216))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(346819712))), name = tensor("module_layers_18_conv_pointwise_conv2_weight_to_fp16_palettized"), shape = tensor([1024, 1024, 1])]; - tensor x_417_cast_fp16 = conv(dilations = x_417_dilations_0, groups = x_417_groups_0, pad = x_417_pad_0, pad_type = x_417_pad_type_0, strides = x_417_strides_0, weight = module_layers_18_conv_pointwise_conv2_weight_to_fp16_palettized, x = input_989_cast_fp16)[name = tensor("x_417_cast_fp16")]; - tensor input_991_perm_0 = const()[name = tensor("input_991_perm_0"), val = tensor([0, 2, 1])]; - tensor input_991_cast_fp16 = transpose(perm = input_991_perm_0, x = x_417_cast_fp16)[name = tensor("transpose_180")]; - tensor input_993_cast_fp16 = add(x = input_975_cast_fp16, y = input_991_cast_fp16)[name = tensor("input_993_cast_fp16")]; - tensor input_995_axes_0 = const()[name = tensor("input_995_axes_0"), val = tensor([-1])]; - tensor module_layers_18_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_18_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(346819904)))]; - tensor module_layers_18_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_18_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(346822016)))]; - tensor input_995_cast_fp16 = layer_norm(axes = input_995_axes_0, beta = module_layers_18_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_18_norm_feed_forward2_weight_to_fp16, x = input_993_cast_fp16)[name = tensor("input_995_cast_fp16")]; - tensor module_layers_18_feed_forward2_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(346824128))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(349969920))), name = tensor("module_layers_18_feed_forward2_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_170_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_18_feed_forward2_linear1_weight_to_fp16_palettized, x = input_995_cast_fp16)[name = tensor("linear_170_cast_fp16")]; - tensor input_999_cast_fp16 = silu(x = linear_170_cast_fp16)[name = tensor("input_999_cast_fp16")]; - tensor module_layers_18_feed_forward2_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(349970112))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(353115904))), name = tensor("module_layers_18_feed_forward2_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_171_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_18_feed_forward2_linear2_weight_to_fp16_palettized, x = input_999_cast_fp16)[name = tensor("linear_171_cast_fp16")]; - tensor var_3300_to_fp16 = const()[name = tensor("op_3300_to_fp16"), val = tensor(0x1p-1)]; - tensor var_3301_cast_fp16 = mul(x = linear_171_cast_fp16, y = var_3300_to_fp16)[name = tensor("op_3301_cast_fp16")]; - tensor input_1005_cast_fp16 = add(x = input_993_cast_fp16, y = var_3301_cast_fp16)[name = tensor("input_1005_cast_fp16")]; - tensor input_1007_axes_0 = const()[name = tensor("input_1007_axes_0"), val = tensor([-1])]; - tensor module_layers_18_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_18_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(353116096)))]; - tensor module_layers_18_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_18_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(353118208)))]; - tensor input_1007_cast_fp16 = layer_norm(axes = input_1007_axes_0, beta = module_layers_18_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_18_norm_out_weight_to_fp16, x = input_1005_cast_fp16)[name = tensor("input_1007_cast_fp16")]; - tensor input_1009_axes_0 = const()[name = tensor("input_1009_axes_0"), val = tensor([-1])]; - tensor module_layers_19_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_19_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(353120320)))]; - tensor module_layers_19_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_19_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(353122432)))]; - tensor input_1009_cast_fp16 = layer_norm(axes = input_1009_axes_0, beta = module_layers_19_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_19_norm_feed_forward1_weight_to_fp16, x = input_1007_cast_fp16)[name = tensor("input_1009_cast_fp16")]; - tensor module_layers_19_feed_forward1_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(353124544))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(356270336))), name = tensor("module_layers_19_feed_forward1_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_172_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_19_feed_forward1_linear1_weight_to_fp16_palettized, x = input_1009_cast_fp16)[name = tensor("linear_172_cast_fp16")]; - tensor input_1013_cast_fp16 = silu(x = linear_172_cast_fp16)[name = tensor("input_1013_cast_fp16")]; - tensor module_layers_19_feed_forward1_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(356270528))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(359416320))), name = tensor("module_layers_19_feed_forward1_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_173_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_19_feed_forward1_linear2_weight_to_fp16_palettized, x = input_1013_cast_fp16)[name = tensor("linear_173_cast_fp16")]; - tensor var_3329_to_fp16 = const()[name = tensor("op_3329_to_fp16"), val = tensor(0x1p-1)]; - tensor var_3330_cast_fp16 = mul(x = linear_173_cast_fp16, y = var_3329_to_fp16)[name = tensor("op_3330_cast_fp16")]; - tensor input_1019_cast_fp16 = add(x = input_1007_cast_fp16, y = var_3330_cast_fp16)[name = tensor("input_1019_cast_fp16")]; - tensor query_39_axes_0 = const()[name = tensor("query_39_axes_0"), val = tensor([-1])]; - tensor module_layers_19_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_19_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(359416512)))]; - tensor module_layers_19_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_19_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(359418624)))]; - tensor query_39_cast_fp16 = layer_norm(axes = query_39_axes_0, beta = module_layers_19_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_19_norm_self_att_weight_to_fp16, x = input_1019_cast_fp16)[name = tensor("query_39_cast_fp16")]; - tensor module_layers_19_self_attn_linear_q_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(359420736))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(360207232))), name = tensor("module_layers_19_self_attn_linear_q_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_174_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_19_self_attn_linear_q_weight_to_fp16_palettized, x = query_39_cast_fp16)[name = tensor("linear_174_cast_fp16")]; - tensor var_3346 = const()[name = tensor("op_3346"), val = tensor([1, -1, 8, 128])]; - tensor q_115_cast_fp16 = reshape(shape = var_3346, x = linear_174_cast_fp16)[name = tensor("q_115_cast_fp16")]; - tensor module_layers_19_self_attn_linear_k_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(360207424))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(360993920))), name = tensor("module_layers_19_self_attn_linear_k_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_175_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_19_self_attn_linear_k_weight_to_fp16_palettized, x = query_39_cast_fp16)[name = tensor("linear_175_cast_fp16")]; - tensor var_3350 = const()[name = tensor("op_3350"), val = tensor([1, -1, 8, 128])]; - tensor k_77_cast_fp16 = reshape(shape = var_3350, x = linear_175_cast_fp16)[name = tensor("k_77_cast_fp16")]; - tensor module_layers_19_self_attn_linear_v_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(360994112))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(361780608))), name = tensor("module_layers_19_self_attn_linear_v_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_176_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_19_self_attn_linear_v_weight_to_fp16_palettized, x = query_39_cast_fp16)[name = tensor("linear_176_cast_fp16")]; - tensor var_3354 = const()[name = tensor("op_3354"), val = tensor([1, -1, 8, 128])]; - tensor v_39_cast_fp16 = reshape(shape = var_3354, x = linear_176_cast_fp16)[name = tensor("v_39_cast_fp16")]; - tensor value_41_perm_0 = const()[name = tensor("value_41_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor module_layers_19_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_19_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(361780800)))]; - tensor var_3366_cast_fp16 = add(x = q_115_cast_fp16, y = module_layers_19_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3366_cast_fp16")]; - tensor module_layers_19_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_19_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(361782912)))]; - tensor var_3368_cast_fp16 = add(x = q_115_cast_fp16, y = module_layers_19_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3368_cast_fp16")]; - tensor q_with_bias_v_39_perm_0 = const()[name = tensor("q_with_bias_v_39_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor x_425_transpose_x_0 = const()[name = tensor("x_425_transpose_x_0"), val = tensor(false)]; - tensor x_425_transpose_y_0 = const()[name = tensor("x_425_transpose_y_0"), val = tensor(false)]; - tensor op_3370_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(361785024))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(362073088))), name = tensor("op_3370_to_fp16_palettized"), shape = tensor([1, 8, 128, 375])]; - tensor q_with_bias_v_39_cast_fp16 = transpose(perm = q_with_bias_v_39_perm_0, x = var_3368_cast_fp16)[name = tensor("transpose_179")]; - tensor x_425_cast_fp16 = matmul(transpose_x = x_425_transpose_x_0, transpose_y = x_425_transpose_y_0, x = q_with_bias_v_39_cast_fp16, y = op_3370_to_fp16_palettized)[name = tensor("x_425_cast_fp16")]; - tensor x_427_pad_0 = const()[name = tensor("x_427_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_427_mode_0 = const()[name = tensor("x_427_mode_0"), val = tensor("constant")]; - tensor const_204_to_fp16 = const()[name = tensor("const_204_to_fp16"), val = tensor(0x0p+0)]; - tensor x_427_cast_fp16 = pad(constant_val = const_204_to_fp16, mode = x_427_mode_0, pad = x_427_pad_0, x = x_425_cast_fp16)[name = tensor("x_427_cast_fp16")]; - tensor var_3378 = const()[name = tensor("op_3378"), val = tensor([1, 8, -1, 188])]; - tensor x_429_cast_fp16 = reshape(shape = var_3378, x = x_427_cast_fp16)[name = tensor("x_429_cast_fp16")]; - tensor var_3382_begin_0 = const()[name = tensor("op_3382_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_3382_end_0 = const()[name = tensor("op_3382_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_3382_end_mask_0 = const()[name = tensor("op_3382_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_3382_cast_fp16 = slice_by_index(begin = var_3382_begin_0, end = var_3382_end_0, end_mask = var_3382_end_mask_0, x = x_429_cast_fp16)[name = tensor("op_3382_cast_fp16")]; - tensor var_3383 = const()[name = tensor("op_3383"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_77_cast_fp16 = reshape(shape = var_3383, x = var_3382_cast_fp16)[name = tensor("matrix_bd_77_cast_fp16")]; - tensor matrix_ac_39_transpose_x_0 = const()[name = tensor("matrix_ac_39_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_39_transpose_y_0 = const()[name = tensor("matrix_ac_39_transpose_y_0"), val = tensor(false)]; - tensor transpose_134_perm_0 = const()[name = tensor("transpose_134_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_135_perm_0 = const()[name = tensor("transpose_135_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_135 = transpose(perm = transpose_135_perm_0, x = k_77_cast_fp16)[name = tensor("transpose_177")]; - tensor transpose_134 = transpose(perm = transpose_134_perm_0, x = var_3366_cast_fp16)[name = tensor("transpose_178")]; - tensor matrix_ac_39_cast_fp16 = matmul(transpose_x = matrix_ac_39_transpose_x_0, transpose_y = matrix_ac_39_transpose_y_0, x = transpose_134, y = transpose_135)[name = tensor("matrix_ac_39_cast_fp16")]; - tensor matrix_bd_79_begin_0 = const()[name = tensor("matrix_bd_79_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_79_end_0 = const()[name = tensor("matrix_bd_79_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_79_end_mask_0 = const()[name = tensor("matrix_bd_79_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_79_cast_fp16 = slice_by_index(begin = matrix_bd_79_begin_0, end = matrix_bd_79_end_0, end_mask = matrix_bd_79_end_mask_0, x = matrix_bd_77_cast_fp16)[name = tensor("matrix_bd_79_cast_fp16")]; - tensor var_3392_cast_fp16 = add(x = matrix_ac_39_cast_fp16, y = matrix_bd_79_cast_fp16)[name = tensor("op_3392_cast_fp16")]; - tensor _inversed_scores_77_y_0_to_fp16 = const()[name = tensor("_inversed_scores_77_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; - tensor _inversed_scores_77_cast_fp16 = mul(x = var_3392_cast_fp16, y = _inversed_scores_77_y_0_to_fp16)[name = tensor("_inversed_scores_77_cast_fp16")]; - tensor scores_79_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_77_cast_fp16, cond = mask_3)[name = tensor("scores_79_cast_fp16")]; - tensor var_3398_cast_fp16 = softmax(axis = var_30, x = scores_79_cast_fp16)[name = tensor("op_3398_cast_fp16")]; - tensor input_1021_cast_fp16 = select(a = var_11_to_fp16, b = var_3398_cast_fp16, cond = mask_3)[name = tensor("input_1021_cast_fp16")]; - tensor x_431_transpose_x_0 = const()[name = tensor("x_431_transpose_x_0"), val = tensor(false)]; - tensor x_431_transpose_y_0 = const()[name = tensor("x_431_transpose_y_0"), val = tensor(false)]; - tensor value_41_cast_fp16 = transpose(perm = value_41_perm_0, x = v_39_cast_fp16)[name = tensor("transpose_176")]; - tensor x_431_cast_fp16 = matmul(transpose_x = x_431_transpose_x_0, transpose_y = x_431_transpose_y_0, x = input_1021_cast_fp16, y = value_41_cast_fp16)[name = tensor("x_431_cast_fp16")]; - tensor var_3402_perm_0 = const()[name = tensor("op_3402_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_3403 = const()[name = tensor("op_3403"), val = tensor([1, -1, 1024])]; - tensor var_3402_cast_fp16 = transpose(perm = var_3402_perm_0, x = x_431_cast_fp16)[name = tensor("transpose_175")]; - tensor input_1023_cast_fp16 = reshape(shape = var_3403, x = var_3402_cast_fp16)[name = tensor("input_1023_cast_fp16")]; - tensor module_layers_19_self_attn_linear_out_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(362073280))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(362859776))), name = tensor("module_layers_19_self_attn_linear_out_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_178_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_19_self_attn_linear_out_weight_to_fp16_palettized, x = input_1023_cast_fp16)[name = tensor("linear_178_cast_fp16")]; - tensor input_1027_cast_fp16 = add(x = input_1019_cast_fp16, y = linear_178_cast_fp16)[name = tensor("input_1027_cast_fp16")]; - tensor x_435_axes_0 = const()[name = tensor("x_435_axes_0"), val = tensor([-1])]; - tensor module_layers_19_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_19_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(362859968)))]; - tensor module_layers_19_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_19_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(362862080)))]; - tensor x_435_cast_fp16 = layer_norm(axes = x_435_axes_0, beta = module_layers_19_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_19_norm_conv_weight_to_fp16, x = input_1027_cast_fp16)[name = tensor("x_435_cast_fp16")]; - tensor input_1029_perm_0 = const()[name = tensor("input_1029_perm_0"), val = tensor([0, 2, 1])]; - tensor input_1031_pad_type_0 = const()[name = tensor("input_1031_pad_type_0"), val = tensor("valid")]; - tensor input_1031_strides_0 = const()[name = tensor("input_1031_strides_0"), val = tensor([1])]; - tensor input_1031_pad_0 = const()[name = tensor("input_1031_pad_0"), val = tensor([0, 0])]; - tensor input_1031_dilations_0 = const()[name = tensor("input_1031_dilations_0"), val = tensor([1])]; - tensor input_1031_groups_0 = const()[name = tensor("input_1031_groups_0"), val = tensor(1)]; - tensor module_layers_19_conv_pointwise_conv1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(362864192))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(364437120))), name = tensor("module_layers_19_conv_pointwise_conv1_weight_to_fp16_palettized"), shape = tensor([2048, 1024, 1])]; - tensor input_1029_cast_fp16 = transpose(perm = input_1029_perm_0, x = x_435_cast_fp16)[name = tensor("transpose_174")]; - tensor input_1031_cast_fp16 = conv(dilations = input_1031_dilations_0, groups = input_1031_groups_0, pad = input_1031_pad_0, pad_type = input_1031_pad_type_0, strides = input_1031_strides_0, weight = module_layers_19_conv_pointwise_conv1_weight_to_fp16_palettized, x = input_1029_cast_fp16)[name = tensor("input_1031_cast_fp16")]; - tensor x_437_split_num_splits_0 = const()[name = tensor("x_437_split_num_splits_0"), val = tensor(2)]; - tensor x_437_split_axis_0 = const()[name = tensor("x_437_split_axis_0"), val = tensor(1)]; - tensor x_437_split_cast_fp16_0, tensor x_437_split_cast_fp16_1 = split(axis = x_437_split_axis_0, num_splits = x_437_split_num_splits_0, x = input_1031_cast_fp16)[name = tensor("x_437_split_cast_fp16")]; - tensor x_437_split_1_sigmoid_cast_fp16 = sigmoid(x = x_437_split_cast_fp16_1)[name = tensor("x_437_split_1_sigmoid_cast_fp16")]; - tensor x_437_cast_fp16 = mul(x = x_437_split_cast_fp16_0, y = x_437_split_1_sigmoid_cast_fp16)[name = tensor("x_437_cast_fp16")]; - tensor input_1033_cast_fp16 = select(a = var_11_to_fp16, b = x_437_cast_fp16, cond = var_328)[name = tensor("input_1033_cast_fp16")]; - tensor input_1035_pad_0 = const()[name = tensor("input_1035_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_1035_mode_0 = const()[name = tensor("input_1035_mode_0"), val = tensor("constant")]; - tensor const_207_to_fp16 = const()[name = tensor("const_207_to_fp16"), val = tensor(0x0p+0)]; - tensor input_1035_cast_fp16 = pad(constant_val = const_207_to_fp16, mode = input_1035_mode_0, pad = input_1035_pad_0, x = input_1033_cast_fp16)[name = tensor("input_1035_cast_fp16")]; - tensor input_1037_pad_type_0 = const()[name = tensor("input_1037_pad_type_0"), val = tensor("valid")]; - tensor input_1037_groups_0 = const()[name = tensor("input_1037_groups_0"), val = tensor(1024)]; - tensor input_1037_strides_0 = const()[name = tensor("input_1037_strides_0"), val = tensor([1])]; - tensor input_1037_pad_0 = const()[name = tensor("input_1037_pad_0"), val = tensor([0, 0])]; - tensor input_1037_dilations_0 = const()[name = tensor("input_1037_dilations_0"), val = tensor([1])]; - tensor const_286_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(364437312))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(364444288))), name = tensor("const_286_to_fp16_palettized"), shape = tensor([1024, 1, 9])]; - tensor const_287_to_fp16 = const()[name = tensor("const_287_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(364444480)))]; - tensor input_1039_cast_fp16 = conv(bias = const_287_to_fp16, dilations = input_1037_dilations_0, groups = input_1037_groups_0, pad = input_1037_pad_0, pad_type = input_1037_pad_type_0, strides = input_1037_strides_0, weight = const_286_to_fp16_palettized, x = input_1035_cast_fp16)[name = tensor("input_1039_cast_fp16")]; - tensor input_1041_cast_fp16 = silu(x = input_1039_cast_fp16)[name = tensor("input_1041_cast_fp16")]; - tensor x_439_pad_type_0 = const()[name = tensor("x_439_pad_type_0"), val = tensor("valid")]; - tensor x_439_strides_0 = const()[name = tensor("x_439_strides_0"), val = tensor([1])]; - tensor x_439_pad_0 = const()[name = tensor("x_439_pad_0"), val = tensor([0, 0])]; - tensor x_439_dilations_0 = const()[name = tensor("x_439_dilations_0"), val = tensor([1])]; - tensor x_439_groups_0 = const()[name = tensor("x_439_groups_0"), val = tensor(1)]; - tensor module_layers_19_conv_pointwise_conv2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(364446592))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(365233088))), name = tensor("module_layers_19_conv_pointwise_conv2_weight_to_fp16_palettized"), shape = tensor([1024, 1024, 1])]; - tensor x_439_cast_fp16 = conv(dilations = x_439_dilations_0, groups = x_439_groups_0, pad = x_439_pad_0, pad_type = x_439_pad_type_0, strides = x_439_strides_0, weight = module_layers_19_conv_pointwise_conv2_weight_to_fp16_palettized, x = input_1041_cast_fp16)[name = tensor("x_439_cast_fp16")]; - tensor input_1043_perm_0 = const()[name = tensor("input_1043_perm_0"), val = tensor([0, 2, 1])]; - tensor input_1043_cast_fp16 = transpose(perm = input_1043_perm_0, x = x_439_cast_fp16)[name = tensor("transpose_173")]; - tensor input_1045_cast_fp16 = add(x = input_1027_cast_fp16, y = input_1043_cast_fp16)[name = tensor("input_1045_cast_fp16")]; - tensor input_1047_axes_0 = const()[name = tensor("input_1047_axes_0"), val = tensor([-1])]; - tensor module_layers_19_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_19_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(365233280)))]; - tensor module_layers_19_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_19_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(365235392)))]; - tensor input_1047_cast_fp16 = layer_norm(axes = input_1047_axes_0, beta = module_layers_19_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_19_norm_feed_forward2_weight_to_fp16, x = input_1045_cast_fp16)[name = tensor("input_1047_cast_fp16")]; - tensor module_layers_19_feed_forward2_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(365237504))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(368383296))), name = tensor("module_layers_19_feed_forward2_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_179_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_19_feed_forward2_linear1_weight_to_fp16_palettized, x = input_1047_cast_fp16)[name = tensor("linear_179_cast_fp16")]; - tensor input_1051_cast_fp16 = silu(x = linear_179_cast_fp16)[name = tensor("input_1051_cast_fp16")]; - tensor module_layers_19_feed_forward2_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(368383488))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(371529280))), name = tensor("module_layers_19_feed_forward2_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_180_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_19_feed_forward2_linear2_weight_to_fp16_palettized, x = input_1051_cast_fp16)[name = tensor("linear_180_cast_fp16")]; - tensor var_3463_to_fp16 = const()[name = tensor("op_3463_to_fp16"), val = tensor(0x1p-1)]; - tensor var_3464_cast_fp16 = mul(x = linear_180_cast_fp16, y = var_3463_to_fp16)[name = tensor("op_3464_cast_fp16")]; - tensor input_1057_cast_fp16 = add(x = input_1045_cast_fp16, y = var_3464_cast_fp16)[name = tensor("input_1057_cast_fp16")]; - tensor input_1059_axes_0 = const()[name = tensor("input_1059_axes_0"), val = tensor([-1])]; - tensor module_layers_19_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_19_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(371529472)))]; - tensor module_layers_19_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_19_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(371531584)))]; - tensor input_1059_cast_fp16 = layer_norm(axes = input_1059_axes_0, beta = module_layers_19_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_19_norm_out_weight_to_fp16, x = input_1057_cast_fp16)[name = tensor("input_1059_cast_fp16")]; - tensor input_1061_axes_0 = const()[name = tensor("input_1061_axes_0"), val = tensor([-1])]; - tensor module_layers_20_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_20_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(371533696)))]; - tensor module_layers_20_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_20_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(371535808)))]; - tensor input_1061_cast_fp16 = layer_norm(axes = input_1061_axes_0, beta = module_layers_20_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_20_norm_feed_forward1_weight_to_fp16, x = input_1059_cast_fp16)[name = tensor("input_1061_cast_fp16")]; - tensor module_layers_20_feed_forward1_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(371537920))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(374683712))), name = tensor("module_layers_20_feed_forward1_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_181_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_20_feed_forward1_linear1_weight_to_fp16_palettized, x = input_1061_cast_fp16)[name = tensor("linear_181_cast_fp16")]; - tensor input_1065_cast_fp16 = silu(x = linear_181_cast_fp16)[name = tensor("input_1065_cast_fp16")]; - tensor module_layers_20_feed_forward1_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(374683904))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(377829696))), name = tensor("module_layers_20_feed_forward1_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_182_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_20_feed_forward1_linear2_weight_to_fp16_palettized, x = input_1065_cast_fp16)[name = tensor("linear_182_cast_fp16")]; - tensor var_3492_to_fp16 = const()[name = tensor("op_3492_to_fp16"), val = tensor(0x1p-1)]; - tensor var_3493_cast_fp16 = mul(x = linear_182_cast_fp16, y = var_3492_to_fp16)[name = tensor("op_3493_cast_fp16")]; - tensor input_1071_cast_fp16 = add(x = input_1059_cast_fp16, y = var_3493_cast_fp16)[name = tensor("input_1071_cast_fp16")]; - tensor query_41_axes_0 = const()[name = tensor("query_41_axes_0"), val = tensor([-1])]; - tensor module_layers_20_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_20_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(377829888)))]; - tensor module_layers_20_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_20_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(377832000)))]; - tensor query_41_cast_fp16 = layer_norm(axes = query_41_axes_0, beta = module_layers_20_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_20_norm_self_att_weight_to_fp16, x = input_1071_cast_fp16)[name = tensor("query_41_cast_fp16")]; - tensor module_layers_20_self_attn_linear_q_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(377834112))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(378620608))), name = tensor("module_layers_20_self_attn_linear_q_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_183_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_20_self_attn_linear_q_weight_to_fp16_palettized, x = query_41_cast_fp16)[name = tensor("linear_183_cast_fp16")]; - tensor var_3509 = const()[name = tensor("op_3509"), val = tensor([1, -1, 8, 128])]; - tensor q_121_cast_fp16 = reshape(shape = var_3509, x = linear_183_cast_fp16)[name = tensor("q_121_cast_fp16")]; - tensor module_layers_20_self_attn_linear_k_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(378620800))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(379407296))), name = tensor("module_layers_20_self_attn_linear_k_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_184_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_20_self_attn_linear_k_weight_to_fp16_palettized, x = query_41_cast_fp16)[name = tensor("linear_184_cast_fp16")]; - tensor var_3513 = const()[name = tensor("op_3513"), val = tensor([1, -1, 8, 128])]; - tensor k_81_cast_fp16 = reshape(shape = var_3513, x = linear_184_cast_fp16)[name = tensor("k_81_cast_fp16")]; - tensor module_layers_20_self_attn_linear_v_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(379407488))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(380193984))), name = tensor("module_layers_20_self_attn_linear_v_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_185_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_20_self_attn_linear_v_weight_to_fp16_palettized, x = query_41_cast_fp16)[name = tensor("linear_185_cast_fp16")]; - tensor var_3517 = const()[name = tensor("op_3517"), val = tensor([1, -1, 8, 128])]; - tensor v_41_cast_fp16 = reshape(shape = var_3517, x = linear_185_cast_fp16)[name = tensor("v_41_cast_fp16")]; - tensor value_43_perm_0 = const()[name = tensor("value_43_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor module_layers_20_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_20_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(380194176)))]; - tensor var_3529_cast_fp16 = add(x = q_121_cast_fp16, y = module_layers_20_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3529_cast_fp16")]; - tensor module_layers_20_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_20_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(380196288)))]; - tensor var_3531_cast_fp16 = add(x = q_121_cast_fp16, y = module_layers_20_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3531_cast_fp16")]; - tensor q_with_bias_v_41_perm_0 = const()[name = tensor("q_with_bias_v_41_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor x_447_transpose_x_0 = const()[name = tensor("x_447_transpose_x_0"), val = tensor(false)]; - tensor x_447_transpose_y_0 = const()[name = tensor("x_447_transpose_y_0"), val = tensor(false)]; - tensor op_3533_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(380198400))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(380486464))), name = tensor("op_3533_to_fp16_palettized"), shape = tensor([1, 8, 128, 375])]; - tensor q_with_bias_v_41_cast_fp16 = transpose(perm = q_with_bias_v_41_perm_0, x = var_3531_cast_fp16)[name = tensor("transpose_172")]; - tensor x_447_cast_fp16 = matmul(transpose_x = x_447_transpose_x_0, transpose_y = x_447_transpose_y_0, x = q_with_bias_v_41_cast_fp16, y = op_3533_to_fp16_palettized)[name = tensor("x_447_cast_fp16")]; - tensor x_449_pad_0 = const()[name = tensor("x_449_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_449_mode_0 = const()[name = tensor("x_449_mode_0"), val = tensor("constant")]; - tensor const_214_to_fp16 = const()[name = tensor("const_214_to_fp16"), val = tensor(0x0p+0)]; - tensor x_449_cast_fp16 = pad(constant_val = const_214_to_fp16, mode = x_449_mode_0, pad = x_449_pad_0, x = x_447_cast_fp16)[name = tensor("x_449_cast_fp16")]; - tensor var_3541 = const()[name = tensor("op_3541"), val = tensor([1, 8, -1, 188])]; - tensor x_451_cast_fp16 = reshape(shape = var_3541, x = x_449_cast_fp16)[name = tensor("x_451_cast_fp16")]; - tensor var_3545_begin_0 = const()[name = tensor("op_3545_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_3545_end_0 = const()[name = tensor("op_3545_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_3545_end_mask_0 = const()[name = tensor("op_3545_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_3545_cast_fp16 = slice_by_index(begin = var_3545_begin_0, end = var_3545_end_0, end_mask = var_3545_end_mask_0, x = x_451_cast_fp16)[name = tensor("op_3545_cast_fp16")]; - tensor var_3546 = const()[name = tensor("op_3546"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_81_cast_fp16 = reshape(shape = var_3546, x = var_3545_cast_fp16)[name = tensor("matrix_bd_81_cast_fp16")]; - tensor matrix_ac_41_transpose_x_0 = const()[name = tensor("matrix_ac_41_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_41_transpose_y_0 = const()[name = tensor("matrix_ac_41_transpose_y_0"), val = tensor(false)]; - tensor transpose_136_perm_0 = const()[name = tensor("transpose_136_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_137_perm_0 = const()[name = tensor("transpose_137_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_137 = transpose(perm = transpose_137_perm_0, x = k_81_cast_fp16)[name = tensor("transpose_170")]; - tensor transpose_136 = transpose(perm = transpose_136_perm_0, x = var_3529_cast_fp16)[name = tensor("transpose_171")]; - tensor matrix_ac_41_cast_fp16 = matmul(transpose_x = matrix_ac_41_transpose_x_0, transpose_y = matrix_ac_41_transpose_y_0, x = transpose_136, y = transpose_137)[name = tensor("matrix_ac_41_cast_fp16")]; - tensor matrix_bd_83_begin_0 = const()[name = tensor("matrix_bd_83_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_83_end_0 = const()[name = tensor("matrix_bd_83_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_83_end_mask_0 = const()[name = tensor("matrix_bd_83_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_83_cast_fp16 = slice_by_index(begin = matrix_bd_83_begin_0, end = matrix_bd_83_end_0, end_mask = matrix_bd_83_end_mask_0, x = matrix_bd_81_cast_fp16)[name = tensor("matrix_bd_83_cast_fp16")]; - tensor var_3555_cast_fp16 = add(x = matrix_ac_41_cast_fp16, y = matrix_bd_83_cast_fp16)[name = tensor("op_3555_cast_fp16")]; - tensor _inversed_scores_81_y_0_to_fp16 = const()[name = tensor("_inversed_scores_81_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; - tensor _inversed_scores_81_cast_fp16 = mul(x = var_3555_cast_fp16, y = _inversed_scores_81_y_0_to_fp16)[name = tensor("_inversed_scores_81_cast_fp16")]; - tensor scores_83_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_81_cast_fp16, cond = mask_3)[name = tensor("scores_83_cast_fp16")]; - tensor var_3561_cast_fp16 = softmax(axis = var_30, x = scores_83_cast_fp16)[name = tensor("op_3561_cast_fp16")]; - tensor input_1073_cast_fp16 = select(a = var_11_to_fp16, b = var_3561_cast_fp16, cond = mask_3)[name = tensor("input_1073_cast_fp16")]; - tensor x_453_transpose_x_0 = const()[name = tensor("x_453_transpose_x_0"), val = tensor(false)]; - tensor x_453_transpose_y_0 = const()[name = tensor("x_453_transpose_y_0"), val = tensor(false)]; - tensor value_43_cast_fp16 = transpose(perm = value_43_perm_0, x = v_41_cast_fp16)[name = tensor("transpose_169")]; - tensor x_453_cast_fp16 = matmul(transpose_x = x_453_transpose_x_0, transpose_y = x_453_transpose_y_0, x = input_1073_cast_fp16, y = value_43_cast_fp16)[name = tensor("x_453_cast_fp16")]; - tensor var_3565_perm_0 = const()[name = tensor("op_3565_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_3566 = const()[name = tensor("op_3566"), val = tensor([1, -1, 1024])]; - tensor var_3565_cast_fp16 = transpose(perm = var_3565_perm_0, x = x_453_cast_fp16)[name = tensor("transpose_168")]; - tensor input_1075_cast_fp16 = reshape(shape = var_3566, x = var_3565_cast_fp16)[name = tensor("input_1075_cast_fp16")]; - tensor module_layers_20_self_attn_linear_out_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(380486656))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(381273152))), name = tensor("module_layers_20_self_attn_linear_out_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_187_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_20_self_attn_linear_out_weight_to_fp16_palettized, x = input_1075_cast_fp16)[name = tensor("linear_187_cast_fp16")]; - tensor input_1079_cast_fp16 = add(x = input_1071_cast_fp16, y = linear_187_cast_fp16)[name = tensor("input_1079_cast_fp16")]; - tensor x_457_axes_0 = const()[name = tensor("x_457_axes_0"), val = tensor([-1])]; - tensor module_layers_20_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_20_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(381273344)))]; - tensor module_layers_20_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_20_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(381275456)))]; - tensor x_457_cast_fp16 = layer_norm(axes = x_457_axes_0, beta = module_layers_20_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_20_norm_conv_weight_to_fp16, x = input_1079_cast_fp16)[name = tensor("x_457_cast_fp16")]; - tensor input_1081_perm_0 = const()[name = tensor("input_1081_perm_0"), val = tensor([0, 2, 1])]; - tensor input_1083_pad_type_0 = const()[name = tensor("input_1083_pad_type_0"), val = tensor("valid")]; - tensor input_1083_strides_0 = const()[name = tensor("input_1083_strides_0"), val = tensor([1])]; - tensor input_1083_pad_0 = const()[name = tensor("input_1083_pad_0"), val = tensor([0, 0])]; - tensor input_1083_dilations_0 = const()[name = tensor("input_1083_dilations_0"), val = tensor([1])]; - tensor input_1083_groups_0 = const()[name = tensor("input_1083_groups_0"), val = tensor(1)]; - tensor module_layers_20_conv_pointwise_conv1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(381277568))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(382850496))), name = tensor("module_layers_20_conv_pointwise_conv1_weight_to_fp16_palettized"), shape = tensor([2048, 1024, 1])]; - tensor input_1081_cast_fp16 = transpose(perm = input_1081_perm_0, x = x_457_cast_fp16)[name = tensor("transpose_167")]; - tensor input_1083_cast_fp16 = conv(dilations = input_1083_dilations_0, groups = input_1083_groups_0, pad = input_1083_pad_0, pad_type = input_1083_pad_type_0, strides = input_1083_strides_0, weight = module_layers_20_conv_pointwise_conv1_weight_to_fp16_palettized, x = input_1081_cast_fp16)[name = tensor("input_1083_cast_fp16")]; - tensor x_459_split_num_splits_0 = const()[name = tensor("x_459_split_num_splits_0"), val = tensor(2)]; - tensor x_459_split_axis_0 = const()[name = tensor("x_459_split_axis_0"), val = tensor(1)]; - tensor x_459_split_cast_fp16_0, tensor x_459_split_cast_fp16_1 = split(axis = x_459_split_axis_0, num_splits = x_459_split_num_splits_0, x = input_1083_cast_fp16)[name = tensor("x_459_split_cast_fp16")]; - tensor x_459_split_1_sigmoid_cast_fp16 = sigmoid(x = x_459_split_cast_fp16_1)[name = tensor("x_459_split_1_sigmoid_cast_fp16")]; - tensor x_459_cast_fp16 = mul(x = x_459_split_cast_fp16_0, y = x_459_split_1_sigmoid_cast_fp16)[name = tensor("x_459_cast_fp16")]; - tensor input_1085_cast_fp16 = select(a = var_11_to_fp16, b = x_459_cast_fp16, cond = var_328)[name = tensor("input_1085_cast_fp16")]; - tensor input_1087_pad_0 = const()[name = tensor("input_1087_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_1087_mode_0 = const()[name = tensor("input_1087_mode_0"), val = tensor("constant")]; - tensor const_217_to_fp16 = const()[name = tensor("const_217_to_fp16"), val = tensor(0x0p+0)]; - tensor input_1087_cast_fp16 = pad(constant_val = const_217_to_fp16, mode = input_1087_mode_0, pad = input_1087_pad_0, x = input_1085_cast_fp16)[name = tensor("input_1087_cast_fp16")]; - tensor input_1089_pad_type_0 = const()[name = tensor("input_1089_pad_type_0"), val = tensor("valid")]; - tensor input_1089_groups_0 = const()[name = tensor("input_1089_groups_0"), val = tensor(1024)]; - tensor input_1089_strides_0 = const()[name = tensor("input_1089_strides_0"), val = tensor([1])]; - tensor input_1089_pad_0 = const()[name = tensor("input_1089_pad_0"), val = tensor([0, 0])]; - tensor input_1089_dilations_0 = const()[name = tensor("input_1089_dilations_0"), val = tensor([1])]; - tensor const_288_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(382850688))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(382857664))), name = tensor("const_288_to_fp16_palettized"), shape = tensor([1024, 1, 9])]; - tensor const_289_to_fp16 = const()[name = tensor("const_289_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(382857856)))]; - tensor input_1091_cast_fp16 = conv(bias = const_289_to_fp16, dilations = input_1089_dilations_0, groups = input_1089_groups_0, pad = input_1089_pad_0, pad_type = input_1089_pad_type_0, strides = input_1089_strides_0, weight = const_288_to_fp16_palettized, x = input_1087_cast_fp16)[name = tensor("input_1091_cast_fp16")]; - tensor input_1093_cast_fp16 = silu(x = input_1091_cast_fp16)[name = tensor("input_1093_cast_fp16")]; - tensor x_461_pad_type_0 = const()[name = tensor("x_461_pad_type_0"), val = tensor("valid")]; - tensor x_461_strides_0 = const()[name = tensor("x_461_strides_0"), val = tensor([1])]; - tensor x_461_pad_0 = const()[name = tensor("x_461_pad_0"), val = tensor([0, 0])]; - tensor x_461_dilations_0 = const()[name = tensor("x_461_dilations_0"), val = tensor([1])]; - tensor x_461_groups_0 = const()[name = tensor("x_461_groups_0"), val = tensor(1)]; - tensor module_layers_20_conv_pointwise_conv2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(382859968))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(383646464))), name = tensor("module_layers_20_conv_pointwise_conv2_weight_to_fp16_palettized"), shape = tensor([1024, 1024, 1])]; - tensor x_461_cast_fp16 = conv(dilations = x_461_dilations_0, groups = x_461_groups_0, pad = x_461_pad_0, pad_type = x_461_pad_type_0, strides = x_461_strides_0, weight = module_layers_20_conv_pointwise_conv2_weight_to_fp16_palettized, x = input_1093_cast_fp16)[name = tensor("x_461_cast_fp16")]; - tensor input_1095_perm_0 = const()[name = tensor("input_1095_perm_0"), val = tensor([0, 2, 1])]; - tensor input_1095_cast_fp16 = transpose(perm = input_1095_perm_0, x = x_461_cast_fp16)[name = tensor("transpose_166")]; - tensor input_1097_cast_fp16 = add(x = input_1079_cast_fp16, y = input_1095_cast_fp16)[name = tensor("input_1097_cast_fp16")]; - tensor input_1099_axes_0 = const()[name = tensor("input_1099_axes_0"), val = tensor([-1])]; - tensor module_layers_20_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_20_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(383646656)))]; - tensor module_layers_20_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_20_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(383648768)))]; - tensor input_1099_cast_fp16 = layer_norm(axes = input_1099_axes_0, beta = module_layers_20_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_20_norm_feed_forward2_weight_to_fp16, x = input_1097_cast_fp16)[name = tensor("input_1099_cast_fp16")]; - tensor module_layers_20_feed_forward2_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(383650880))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(386796672))), name = tensor("module_layers_20_feed_forward2_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_188_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_20_feed_forward2_linear1_weight_to_fp16_palettized, x = input_1099_cast_fp16)[name = tensor("linear_188_cast_fp16")]; - tensor input_1103_cast_fp16 = silu(x = linear_188_cast_fp16)[name = tensor("input_1103_cast_fp16")]; - tensor module_layers_20_feed_forward2_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(386796864))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(389942656))), name = tensor("module_layers_20_feed_forward2_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_189_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_20_feed_forward2_linear2_weight_to_fp16_palettized, x = input_1103_cast_fp16)[name = tensor("linear_189_cast_fp16")]; - tensor var_3626_to_fp16 = const()[name = tensor("op_3626_to_fp16"), val = tensor(0x1p-1)]; - tensor var_3627_cast_fp16 = mul(x = linear_189_cast_fp16, y = var_3626_to_fp16)[name = tensor("op_3627_cast_fp16")]; - tensor input_1109_cast_fp16 = add(x = input_1097_cast_fp16, y = var_3627_cast_fp16)[name = tensor("input_1109_cast_fp16")]; - tensor input_1111_axes_0 = const()[name = tensor("input_1111_axes_0"), val = tensor([-1])]; - tensor module_layers_20_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_20_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(389942848)))]; - tensor module_layers_20_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_20_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(389944960)))]; - tensor input_1111_cast_fp16 = layer_norm(axes = input_1111_axes_0, beta = module_layers_20_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_20_norm_out_weight_to_fp16, x = input_1109_cast_fp16)[name = tensor("input_1111_cast_fp16")]; - tensor input_1113_axes_0 = const()[name = tensor("input_1113_axes_0"), val = tensor([-1])]; - tensor module_layers_21_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_21_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(389947072)))]; - tensor module_layers_21_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_21_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(389949184)))]; - tensor input_1113_cast_fp16 = layer_norm(axes = input_1113_axes_0, beta = module_layers_21_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_21_norm_feed_forward1_weight_to_fp16, x = input_1111_cast_fp16)[name = tensor("input_1113_cast_fp16")]; - tensor module_layers_21_feed_forward1_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(389951296))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(393097088))), name = tensor("module_layers_21_feed_forward1_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_190_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_21_feed_forward1_linear1_weight_to_fp16_palettized, x = input_1113_cast_fp16)[name = tensor("linear_190_cast_fp16")]; - tensor input_1117_cast_fp16 = silu(x = linear_190_cast_fp16)[name = tensor("input_1117_cast_fp16")]; - tensor module_layers_21_feed_forward1_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(393097280))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(396243072))), name = tensor("module_layers_21_feed_forward1_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_191_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_21_feed_forward1_linear2_weight_to_fp16_palettized, x = input_1117_cast_fp16)[name = tensor("linear_191_cast_fp16")]; - tensor var_3655_to_fp16 = const()[name = tensor("op_3655_to_fp16"), val = tensor(0x1p-1)]; - tensor var_3656_cast_fp16 = mul(x = linear_191_cast_fp16, y = var_3655_to_fp16)[name = tensor("op_3656_cast_fp16")]; - tensor input_1123_cast_fp16 = add(x = input_1111_cast_fp16, y = var_3656_cast_fp16)[name = tensor("input_1123_cast_fp16")]; - tensor query_43_axes_0 = const()[name = tensor("query_43_axes_0"), val = tensor([-1])]; - tensor module_layers_21_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_21_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(396243264)))]; - tensor module_layers_21_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_21_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(396245376)))]; - tensor query_43_cast_fp16 = layer_norm(axes = query_43_axes_0, beta = module_layers_21_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_21_norm_self_att_weight_to_fp16, x = input_1123_cast_fp16)[name = tensor("query_43_cast_fp16")]; - tensor module_layers_21_self_attn_linear_q_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(396247488))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(397033984))), name = tensor("module_layers_21_self_attn_linear_q_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_192_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_21_self_attn_linear_q_weight_to_fp16_palettized, x = query_43_cast_fp16)[name = tensor("linear_192_cast_fp16")]; - tensor var_3672 = const()[name = tensor("op_3672"), val = tensor([1, -1, 8, 128])]; - tensor q_127_cast_fp16 = reshape(shape = var_3672, x = linear_192_cast_fp16)[name = tensor("q_127_cast_fp16")]; - tensor module_layers_21_self_attn_linear_k_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(397034176))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(397820672))), name = tensor("module_layers_21_self_attn_linear_k_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_193_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_21_self_attn_linear_k_weight_to_fp16_palettized, x = query_43_cast_fp16)[name = tensor("linear_193_cast_fp16")]; - tensor var_3676 = const()[name = tensor("op_3676"), val = tensor([1, -1, 8, 128])]; - tensor k_85_cast_fp16 = reshape(shape = var_3676, x = linear_193_cast_fp16)[name = tensor("k_85_cast_fp16")]; - tensor module_layers_21_self_attn_linear_v_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(397820864))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(398607360))), name = tensor("module_layers_21_self_attn_linear_v_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_194_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_21_self_attn_linear_v_weight_to_fp16_palettized, x = query_43_cast_fp16)[name = tensor("linear_194_cast_fp16")]; - tensor var_3680 = const()[name = tensor("op_3680"), val = tensor([1, -1, 8, 128])]; - tensor v_43_cast_fp16 = reshape(shape = var_3680, x = linear_194_cast_fp16)[name = tensor("v_43_cast_fp16")]; - tensor value_45_perm_0 = const()[name = tensor("value_45_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor module_layers_21_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_21_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(398607552)))]; - tensor var_3692_cast_fp16 = add(x = q_127_cast_fp16, y = module_layers_21_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3692_cast_fp16")]; - tensor module_layers_21_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_21_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(398609664)))]; - tensor var_3694_cast_fp16 = add(x = q_127_cast_fp16, y = module_layers_21_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3694_cast_fp16")]; - tensor q_with_bias_v_43_perm_0 = const()[name = tensor("q_with_bias_v_43_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor x_469_transpose_x_0 = const()[name = tensor("x_469_transpose_x_0"), val = tensor(false)]; - tensor x_469_transpose_y_0 = const()[name = tensor("x_469_transpose_y_0"), val = tensor(false)]; - tensor op_3696_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(398611776))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(398899840))), name = tensor("op_3696_to_fp16_palettized"), shape = tensor([1, 8, 128, 375])]; - tensor q_with_bias_v_43_cast_fp16 = transpose(perm = q_with_bias_v_43_perm_0, x = var_3694_cast_fp16)[name = tensor("transpose_165")]; - tensor x_469_cast_fp16 = matmul(transpose_x = x_469_transpose_x_0, transpose_y = x_469_transpose_y_0, x = q_with_bias_v_43_cast_fp16, y = op_3696_to_fp16_palettized)[name = tensor("x_469_cast_fp16")]; - tensor x_471_pad_0 = const()[name = tensor("x_471_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_471_mode_0 = const()[name = tensor("x_471_mode_0"), val = tensor("constant")]; - tensor const_224_to_fp16 = const()[name = tensor("const_224_to_fp16"), val = tensor(0x0p+0)]; - tensor x_471_cast_fp16 = pad(constant_val = const_224_to_fp16, mode = x_471_mode_0, pad = x_471_pad_0, x = x_469_cast_fp16)[name = tensor("x_471_cast_fp16")]; - tensor var_3704 = const()[name = tensor("op_3704"), val = tensor([1, 8, -1, 188])]; - tensor x_473_cast_fp16 = reshape(shape = var_3704, x = x_471_cast_fp16)[name = tensor("x_473_cast_fp16")]; - tensor var_3708_begin_0 = const()[name = tensor("op_3708_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_3708_end_0 = const()[name = tensor("op_3708_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_3708_end_mask_0 = const()[name = tensor("op_3708_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_3708_cast_fp16 = slice_by_index(begin = var_3708_begin_0, end = var_3708_end_0, end_mask = var_3708_end_mask_0, x = x_473_cast_fp16)[name = tensor("op_3708_cast_fp16")]; - tensor var_3709 = const()[name = tensor("op_3709"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_85_cast_fp16 = reshape(shape = var_3709, x = var_3708_cast_fp16)[name = tensor("matrix_bd_85_cast_fp16")]; - tensor matrix_ac_43_transpose_x_0 = const()[name = tensor("matrix_ac_43_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_43_transpose_y_0 = const()[name = tensor("matrix_ac_43_transpose_y_0"), val = tensor(false)]; - tensor transpose_138_perm_0 = const()[name = tensor("transpose_138_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_139_perm_0 = const()[name = tensor("transpose_139_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_139 = transpose(perm = transpose_139_perm_0, x = k_85_cast_fp16)[name = tensor("transpose_163")]; - tensor transpose_138 = transpose(perm = transpose_138_perm_0, x = var_3692_cast_fp16)[name = tensor("transpose_164")]; - tensor matrix_ac_43_cast_fp16 = matmul(transpose_x = matrix_ac_43_transpose_x_0, transpose_y = matrix_ac_43_transpose_y_0, x = transpose_138, y = transpose_139)[name = tensor("matrix_ac_43_cast_fp16")]; - tensor matrix_bd_87_begin_0 = const()[name = tensor("matrix_bd_87_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_87_end_0 = const()[name = tensor("matrix_bd_87_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_87_end_mask_0 = const()[name = tensor("matrix_bd_87_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_87_cast_fp16 = slice_by_index(begin = matrix_bd_87_begin_0, end = matrix_bd_87_end_0, end_mask = matrix_bd_87_end_mask_0, x = matrix_bd_85_cast_fp16)[name = tensor("matrix_bd_87_cast_fp16")]; - tensor var_3718_cast_fp16 = add(x = matrix_ac_43_cast_fp16, y = matrix_bd_87_cast_fp16)[name = tensor("op_3718_cast_fp16")]; - tensor _inversed_scores_85_y_0_to_fp16 = const()[name = tensor("_inversed_scores_85_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; - tensor _inversed_scores_85_cast_fp16 = mul(x = var_3718_cast_fp16, y = _inversed_scores_85_y_0_to_fp16)[name = tensor("_inversed_scores_85_cast_fp16")]; - tensor scores_87_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_85_cast_fp16, cond = mask_3)[name = tensor("scores_87_cast_fp16")]; - tensor var_3724_cast_fp16 = softmax(axis = var_30, x = scores_87_cast_fp16)[name = tensor("op_3724_cast_fp16")]; - tensor input_1125_cast_fp16 = select(a = var_11_to_fp16, b = var_3724_cast_fp16, cond = mask_3)[name = tensor("input_1125_cast_fp16")]; - tensor x_475_transpose_x_0 = const()[name = tensor("x_475_transpose_x_0"), val = tensor(false)]; - tensor x_475_transpose_y_0 = const()[name = tensor("x_475_transpose_y_0"), val = tensor(false)]; - tensor value_45_cast_fp16 = transpose(perm = value_45_perm_0, x = v_43_cast_fp16)[name = tensor("transpose_162")]; - tensor x_475_cast_fp16 = matmul(transpose_x = x_475_transpose_x_0, transpose_y = x_475_transpose_y_0, x = input_1125_cast_fp16, y = value_45_cast_fp16)[name = tensor("x_475_cast_fp16")]; - tensor var_3728_perm_0 = const()[name = tensor("op_3728_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_3729 = const()[name = tensor("op_3729"), val = tensor([1, -1, 1024])]; - tensor var_3728_cast_fp16 = transpose(perm = var_3728_perm_0, x = x_475_cast_fp16)[name = tensor("transpose_161")]; - tensor input_1127_cast_fp16 = reshape(shape = var_3729, x = var_3728_cast_fp16)[name = tensor("input_1127_cast_fp16")]; - tensor module_layers_21_self_attn_linear_out_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(398900032))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(399686528))), name = tensor("module_layers_21_self_attn_linear_out_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_196_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_21_self_attn_linear_out_weight_to_fp16_palettized, x = input_1127_cast_fp16)[name = tensor("linear_196_cast_fp16")]; - tensor input_1131_cast_fp16 = add(x = input_1123_cast_fp16, y = linear_196_cast_fp16)[name = tensor("input_1131_cast_fp16")]; - tensor x_479_axes_0 = const()[name = tensor("x_479_axes_0"), val = tensor([-1])]; - tensor module_layers_21_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_21_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(399686720)))]; - tensor module_layers_21_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_21_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(399688832)))]; - tensor x_479_cast_fp16 = layer_norm(axes = x_479_axes_0, beta = module_layers_21_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_21_norm_conv_weight_to_fp16, x = input_1131_cast_fp16)[name = tensor("x_479_cast_fp16")]; - tensor input_1133_perm_0 = const()[name = tensor("input_1133_perm_0"), val = tensor([0, 2, 1])]; - tensor input_1135_pad_type_0 = const()[name = tensor("input_1135_pad_type_0"), val = tensor("valid")]; - tensor input_1135_strides_0 = const()[name = tensor("input_1135_strides_0"), val = tensor([1])]; - tensor input_1135_pad_0 = const()[name = tensor("input_1135_pad_0"), val = tensor([0, 0])]; - tensor input_1135_dilations_0 = const()[name = tensor("input_1135_dilations_0"), val = tensor([1])]; - tensor input_1135_groups_0 = const()[name = tensor("input_1135_groups_0"), val = tensor(1)]; - tensor module_layers_21_conv_pointwise_conv1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(399690944))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(401263872))), name = tensor("module_layers_21_conv_pointwise_conv1_weight_to_fp16_palettized"), shape = tensor([2048, 1024, 1])]; - tensor input_1133_cast_fp16 = transpose(perm = input_1133_perm_0, x = x_479_cast_fp16)[name = tensor("transpose_160")]; - tensor input_1135_cast_fp16 = conv(dilations = input_1135_dilations_0, groups = input_1135_groups_0, pad = input_1135_pad_0, pad_type = input_1135_pad_type_0, strides = input_1135_strides_0, weight = module_layers_21_conv_pointwise_conv1_weight_to_fp16_palettized, x = input_1133_cast_fp16)[name = tensor("input_1135_cast_fp16")]; - tensor x_481_split_num_splits_0 = const()[name = tensor("x_481_split_num_splits_0"), val = tensor(2)]; - tensor x_481_split_axis_0 = const()[name = tensor("x_481_split_axis_0"), val = tensor(1)]; - tensor x_481_split_cast_fp16_0, tensor x_481_split_cast_fp16_1 = split(axis = x_481_split_axis_0, num_splits = x_481_split_num_splits_0, x = input_1135_cast_fp16)[name = tensor("x_481_split_cast_fp16")]; - tensor x_481_split_1_sigmoid_cast_fp16 = sigmoid(x = x_481_split_cast_fp16_1)[name = tensor("x_481_split_1_sigmoid_cast_fp16")]; - tensor x_481_cast_fp16 = mul(x = x_481_split_cast_fp16_0, y = x_481_split_1_sigmoid_cast_fp16)[name = tensor("x_481_cast_fp16")]; - tensor input_1137_cast_fp16 = select(a = var_11_to_fp16, b = x_481_cast_fp16, cond = var_328)[name = tensor("input_1137_cast_fp16")]; - tensor input_1139_pad_0 = const()[name = tensor("input_1139_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_1139_mode_0 = const()[name = tensor("input_1139_mode_0"), val = tensor("constant")]; - tensor const_227_to_fp16 = const()[name = tensor("const_227_to_fp16"), val = tensor(0x0p+0)]; - tensor input_1139_cast_fp16 = pad(constant_val = const_227_to_fp16, mode = input_1139_mode_0, pad = input_1139_pad_0, x = input_1137_cast_fp16)[name = tensor("input_1139_cast_fp16")]; - tensor input_1141_pad_type_0 = const()[name = tensor("input_1141_pad_type_0"), val = tensor("valid")]; - tensor input_1141_groups_0 = const()[name = tensor("input_1141_groups_0"), val = tensor(1024)]; - tensor input_1141_strides_0 = const()[name = tensor("input_1141_strides_0"), val = tensor([1])]; - tensor input_1141_pad_0 = const()[name = tensor("input_1141_pad_0"), val = tensor([0, 0])]; - tensor input_1141_dilations_0 = const()[name = tensor("input_1141_dilations_0"), val = tensor([1])]; - tensor const_290_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(401264064))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(401271040))), name = tensor("const_290_to_fp16_palettized"), shape = tensor([1024, 1, 9])]; - tensor const_291_to_fp16 = const()[name = tensor("const_291_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(401271232)))]; - tensor input_1143_cast_fp16 = conv(bias = const_291_to_fp16, dilations = input_1141_dilations_0, groups = input_1141_groups_0, pad = input_1141_pad_0, pad_type = input_1141_pad_type_0, strides = input_1141_strides_0, weight = const_290_to_fp16_palettized, x = input_1139_cast_fp16)[name = tensor("input_1143_cast_fp16")]; - tensor input_1145_cast_fp16 = silu(x = input_1143_cast_fp16)[name = tensor("input_1145_cast_fp16")]; - tensor x_483_pad_type_0 = const()[name = tensor("x_483_pad_type_0"), val = tensor("valid")]; - tensor x_483_strides_0 = const()[name = tensor("x_483_strides_0"), val = tensor([1])]; - tensor x_483_pad_0 = const()[name = tensor("x_483_pad_0"), val = tensor([0, 0])]; - tensor x_483_dilations_0 = const()[name = tensor("x_483_dilations_0"), val = tensor([1])]; - tensor x_483_groups_0 = const()[name = tensor("x_483_groups_0"), val = tensor(1)]; - tensor module_layers_21_conv_pointwise_conv2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(401273344))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(402059840))), name = tensor("module_layers_21_conv_pointwise_conv2_weight_to_fp16_palettized"), shape = tensor([1024, 1024, 1])]; - tensor x_483_cast_fp16 = conv(dilations = x_483_dilations_0, groups = x_483_groups_0, pad = x_483_pad_0, pad_type = x_483_pad_type_0, strides = x_483_strides_0, weight = module_layers_21_conv_pointwise_conv2_weight_to_fp16_palettized, x = input_1145_cast_fp16)[name = tensor("x_483_cast_fp16")]; - tensor input_1147_perm_0 = const()[name = tensor("input_1147_perm_0"), val = tensor([0, 2, 1])]; - tensor input_1147_cast_fp16 = transpose(perm = input_1147_perm_0, x = x_483_cast_fp16)[name = tensor("transpose_159")]; - tensor input_1149_cast_fp16 = add(x = input_1131_cast_fp16, y = input_1147_cast_fp16)[name = tensor("input_1149_cast_fp16")]; - tensor input_1151_axes_0 = const()[name = tensor("input_1151_axes_0"), val = tensor([-1])]; - tensor module_layers_21_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_21_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(402060032)))]; - tensor module_layers_21_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_21_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(402062144)))]; - tensor input_1151_cast_fp16 = layer_norm(axes = input_1151_axes_0, beta = module_layers_21_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_21_norm_feed_forward2_weight_to_fp16, x = input_1149_cast_fp16)[name = tensor("input_1151_cast_fp16")]; - tensor module_layers_21_feed_forward2_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(402064256))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(405210048))), name = tensor("module_layers_21_feed_forward2_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_197_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_21_feed_forward2_linear1_weight_to_fp16_palettized, x = input_1151_cast_fp16)[name = tensor("linear_197_cast_fp16")]; - tensor input_1155_cast_fp16 = silu(x = linear_197_cast_fp16)[name = tensor("input_1155_cast_fp16")]; - tensor module_layers_21_feed_forward2_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(405210240))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(408356032))), name = tensor("module_layers_21_feed_forward2_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_198_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_21_feed_forward2_linear2_weight_to_fp16_palettized, x = input_1155_cast_fp16)[name = tensor("linear_198_cast_fp16")]; - tensor var_3789_to_fp16 = const()[name = tensor("op_3789_to_fp16"), val = tensor(0x1p-1)]; - tensor var_3790_cast_fp16 = mul(x = linear_198_cast_fp16, y = var_3789_to_fp16)[name = tensor("op_3790_cast_fp16")]; - tensor input_1161_cast_fp16 = add(x = input_1149_cast_fp16, y = var_3790_cast_fp16)[name = tensor("input_1161_cast_fp16")]; - tensor input_1163_axes_0 = const()[name = tensor("input_1163_axes_0"), val = tensor([-1])]; - tensor module_layers_21_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_21_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(408356224)))]; - tensor module_layers_21_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_21_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(408358336)))]; - tensor input_1163_cast_fp16 = layer_norm(axes = input_1163_axes_0, beta = module_layers_21_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_21_norm_out_weight_to_fp16, x = input_1161_cast_fp16)[name = tensor("input_1163_cast_fp16")]; - tensor input_1165_axes_0 = const()[name = tensor("input_1165_axes_0"), val = tensor([-1])]; - tensor module_layers_22_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_22_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(408360448)))]; - tensor module_layers_22_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_22_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(408362560)))]; - tensor input_1165_cast_fp16 = layer_norm(axes = input_1165_axes_0, beta = module_layers_22_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_22_norm_feed_forward1_weight_to_fp16, x = input_1163_cast_fp16)[name = tensor("input_1165_cast_fp16")]; - tensor module_layers_22_feed_forward1_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(408364672))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(411510464))), name = tensor("module_layers_22_feed_forward1_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_199_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_22_feed_forward1_linear1_weight_to_fp16_palettized, x = input_1165_cast_fp16)[name = tensor("linear_199_cast_fp16")]; - tensor input_1169_cast_fp16 = silu(x = linear_199_cast_fp16)[name = tensor("input_1169_cast_fp16")]; - tensor module_layers_22_feed_forward1_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(411510656))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(414656448))), name = tensor("module_layers_22_feed_forward1_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_200_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_22_feed_forward1_linear2_weight_to_fp16_palettized, x = input_1169_cast_fp16)[name = tensor("linear_200_cast_fp16")]; - tensor var_3818_to_fp16 = const()[name = tensor("op_3818_to_fp16"), val = tensor(0x1p-1)]; - tensor var_3819_cast_fp16 = mul(x = linear_200_cast_fp16, y = var_3818_to_fp16)[name = tensor("op_3819_cast_fp16")]; - tensor input_1175_cast_fp16 = add(x = input_1163_cast_fp16, y = var_3819_cast_fp16)[name = tensor("input_1175_cast_fp16")]; - tensor query_45_axes_0 = const()[name = tensor("query_45_axes_0"), val = tensor([-1])]; - tensor module_layers_22_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_22_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(414656640)))]; - tensor module_layers_22_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_22_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(414658752)))]; - tensor query_45_cast_fp16 = layer_norm(axes = query_45_axes_0, beta = module_layers_22_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_22_norm_self_att_weight_to_fp16, x = input_1175_cast_fp16)[name = tensor("query_45_cast_fp16")]; - tensor module_layers_22_self_attn_linear_q_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(414660864))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(415447360))), name = tensor("module_layers_22_self_attn_linear_q_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_201_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_22_self_attn_linear_q_weight_to_fp16_palettized, x = query_45_cast_fp16)[name = tensor("linear_201_cast_fp16")]; - tensor var_3835 = const()[name = tensor("op_3835"), val = tensor([1, -1, 8, 128])]; - tensor q_133_cast_fp16 = reshape(shape = var_3835, x = linear_201_cast_fp16)[name = tensor("q_133_cast_fp16")]; - tensor module_layers_22_self_attn_linear_k_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(415447552))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(416234048))), name = tensor("module_layers_22_self_attn_linear_k_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_202_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_22_self_attn_linear_k_weight_to_fp16_palettized, x = query_45_cast_fp16)[name = tensor("linear_202_cast_fp16")]; - tensor var_3839 = const()[name = tensor("op_3839"), val = tensor([1, -1, 8, 128])]; - tensor k_89_cast_fp16 = reshape(shape = var_3839, x = linear_202_cast_fp16)[name = tensor("k_89_cast_fp16")]; - tensor module_layers_22_self_attn_linear_v_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(416234240))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(417020736))), name = tensor("module_layers_22_self_attn_linear_v_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_203_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_22_self_attn_linear_v_weight_to_fp16_palettized, x = query_45_cast_fp16)[name = tensor("linear_203_cast_fp16")]; - tensor var_3843 = const()[name = tensor("op_3843"), val = tensor([1, -1, 8, 128])]; - tensor v_45_cast_fp16 = reshape(shape = var_3843, x = linear_203_cast_fp16)[name = tensor("v_45_cast_fp16")]; - tensor value_47_perm_0 = const()[name = tensor("value_47_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor module_layers_22_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_22_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(417020928)))]; - tensor var_3855_cast_fp16 = add(x = q_133_cast_fp16, y = module_layers_22_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3855_cast_fp16")]; - tensor module_layers_22_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_22_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(417023040)))]; - tensor var_3857_cast_fp16 = add(x = q_133_cast_fp16, y = module_layers_22_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3857_cast_fp16")]; - tensor q_with_bias_v_45_perm_0 = const()[name = tensor("q_with_bias_v_45_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor x_491_transpose_x_0 = const()[name = tensor("x_491_transpose_x_0"), val = tensor(false)]; - tensor x_491_transpose_y_0 = const()[name = tensor("x_491_transpose_y_0"), val = tensor(false)]; - tensor op_3859_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(417025152))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(417313216))), name = tensor("op_3859_to_fp16_palettized"), shape = tensor([1, 8, 128, 375])]; - tensor q_with_bias_v_45_cast_fp16 = transpose(perm = q_with_bias_v_45_perm_0, x = var_3857_cast_fp16)[name = tensor("transpose_158")]; - tensor x_491_cast_fp16 = matmul(transpose_x = x_491_transpose_x_0, transpose_y = x_491_transpose_y_0, x = q_with_bias_v_45_cast_fp16, y = op_3859_to_fp16_palettized)[name = tensor("x_491_cast_fp16")]; - tensor x_493_pad_0 = const()[name = tensor("x_493_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_493_mode_0 = const()[name = tensor("x_493_mode_0"), val = tensor("constant")]; - tensor const_234_to_fp16 = const()[name = tensor("const_234_to_fp16"), val = tensor(0x0p+0)]; - tensor x_493_cast_fp16 = pad(constant_val = const_234_to_fp16, mode = x_493_mode_0, pad = x_493_pad_0, x = x_491_cast_fp16)[name = tensor("x_493_cast_fp16")]; - tensor var_3867 = const()[name = tensor("op_3867"), val = tensor([1, 8, -1, 188])]; - tensor x_495_cast_fp16 = reshape(shape = var_3867, x = x_493_cast_fp16)[name = tensor("x_495_cast_fp16")]; - tensor var_3871_begin_0 = const()[name = tensor("op_3871_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_3871_end_0 = const()[name = tensor("op_3871_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_3871_end_mask_0 = const()[name = tensor("op_3871_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_3871_cast_fp16 = slice_by_index(begin = var_3871_begin_0, end = var_3871_end_0, end_mask = var_3871_end_mask_0, x = x_495_cast_fp16)[name = tensor("op_3871_cast_fp16")]; - tensor var_3872 = const()[name = tensor("op_3872"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_89_cast_fp16 = reshape(shape = var_3872, x = var_3871_cast_fp16)[name = tensor("matrix_bd_89_cast_fp16")]; - tensor matrix_ac_45_transpose_x_0 = const()[name = tensor("matrix_ac_45_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_45_transpose_y_0 = const()[name = tensor("matrix_ac_45_transpose_y_0"), val = tensor(false)]; - tensor transpose_140_perm_0 = const()[name = tensor("transpose_140_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_141_perm_0 = const()[name = tensor("transpose_141_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_141 = transpose(perm = transpose_141_perm_0, x = k_89_cast_fp16)[name = tensor("transpose_156")]; - tensor transpose_140 = transpose(perm = transpose_140_perm_0, x = var_3855_cast_fp16)[name = tensor("transpose_157")]; - tensor matrix_ac_45_cast_fp16 = matmul(transpose_x = matrix_ac_45_transpose_x_0, transpose_y = matrix_ac_45_transpose_y_0, x = transpose_140, y = transpose_141)[name = tensor("matrix_ac_45_cast_fp16")]; - tensor matrix_bd_91_begin_0 = const()[name = tensor("matrix_bd_91_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_91_end_0 = const()[name = tensor("matrix_bd_91_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_91_end_mask_0 = const()[name = tensor("matrix_bd_91_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_91_cast_fp16 = slice_by_index(begin = matrix_bd_91_begin_0, end = matrix_bd_91_end_0, end_mask = matrix_bd_91_end_mask_0, x = matrix_bd_89_cast_fp16)[name = tensor("matrix_bd_91_cast_fp16")]; - tensor var_3881_cast_fp16 = add(x = matrix_ac_45_cast_fp16, y = matrix_bd_91_cast_fp16)[name = tensor("op_3881_cast_fp16")]; - tensor _inversed_scores_89_y_0_to_fp16 = const()[name = tensor("_inversed_scores_89_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; - tensor _inversed_scores_89_cast_fp16 = mul(x = var_3881_cast_fp16, y = _inversed_scores_89_y_0_to_fp16)[name = tensor("_inversed_scores_89_cast_fp16")]; - tensor scores_91_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_89_cast_fp16, cond = mask_3)[name = tensor("scores_91_cast_fp16")]; - tensor var_3887_cast_fp16 = softmax(axis = var_30, x = scores_91_cast_fp16)[name = tensor("op_3887_cast_fp16")]; - tensor input_1177_cast_fp16 = select(a = var_11_to_fp16, b = var_3887_cast_fp16, cond = mask_3)[name = tensor("input_1177_cast_fp16")]; - tensor x_497_transpose_x_0 = const()[name = tensor("x_497_transpose_x_0"), val = tensor(false)]; - tensor x_497_transpose_y_0 = const()[name = tensor("x_497_transpose_y_0"), val = tensor(false)]; - tensor value_47_cast_fp16 = transpose(perm = value_47_perm_0, x = v_45_cast_fp16)[name = tensor("transpose_155")]; - tensor x_497_cast_fp16 = matmul(transpose_x = x_497_transpose_x_0, transpose_y = x_497_transpose_y_0, x = input_1177_cast_fp16, y = value_47_cast_fp16)[name = tensor("x_497_cast_fp16")]; - tensor var_3891_perm_0 = const()[name = tensor("op_3891_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_3892 = const()[name = tensor("op_3892"), val = tensor([1, -1, 1024])]; - tensor var_3891_cast_fp16 = transpose(perm = var_3891_perm_0, x = x_497_cast_fp16)[name = tensor("transpose_154")]; - tensor input_1179_cast_fp16 = reshape(shape = var_3892, x = var_3891_cast_fp16)[name = tensor("input_1179_cast_fp16")]; - tensor module_layers_22_self_attn_linear_out_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(417313408))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(418099904))), name = tensor("module_layers_22_self_attn_linear_out_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_205_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_22_self_attn_linear_out_weight_to_fp16_palettized, x = input_1179_cast_fp16)[name = tensor("linear_205_cast_fp16")]; - tensor input_1183_cast_fp16 = add(x = input_1175_cast_fp16, y = linear_205_cast_fp16)[name = tensor("input_1183_cast_fp16")]; - tensor x_501_axes_0 = const()[name = tensor("x_501_axes_0"), val = tensor([-1])]; - tensor module_layers_22_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_22_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(418100096)))]; - tensor module_layers_22_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_22_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(418102208)))]; - tensor x_501_cast_fp16 = layer_norm(axes = x_501_axes_0, beta = module_layers_22_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_22_norm_conv_weight_to_fp16, x = input_1183_cast_fp16)[name = tensor("x_501_cast_fp16")]; - tensor input_1185_perm_0 = const()[name = tensor("input_1185_perm_0"), val = tensor([0, 2, 1])]; - tensor input_1187_pad_type_0 = const()[name = tensor("input_1187_pad_type_0"), val = tensor("valid")]; - tensor input_1187_strides_0 = const()[name = tensor("input_1187_strides_0"), val = tensor([1])]; - tensor input_1187_pad_0 = const()[name = tensor("input_1187_pad_0"), val = tensor([0, 0])]; - tensor input_1187_dilations_0 = const()[name = tensor("input_1187_dilations_0"), val = tensor([1])]; - tensor input_1187_groups_0 = const()[name = tensor("input_1187_groups_0"), val = tensor(1)]; - tensor module_layers_22_conv_pointwise_conv1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(418104320))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(419677248))), name = tensor("module_layers_22_conv_pointwise_conv1_weight_to_fp16_palettized"), shape = tensor([2048, 1024, 1])]; - tensor input_1185_cast_fp16 = transpose(perm = input_1185_perm_0, x = x_501_cast_fp16)[name = tensor("transpose_153")]; - tensor input_1187_cast_fp16 = conv(dilations = input_1187_dilations_0, groups = input_1187_groups_0, pad = input_1187_pad_0, pad_type = input_1187_pad_type_0, strides = input_1187_strides_0, weight = module_layers_22_conv_pointwise_conv1_weight_to_fp16_palettized, x = input_1185_cast_fp16)[name = tensor("input_1187_cast_fp16")]; - tensor x_503_split_num_splits_0 = const()[name = tensor("x_503_split_num_splits_0"), val = tensor(2)]; - tensor x_503_split_axis_0 = const()[name = tensor("x_503_split_axis_0"), val = tensor(1)]; - tensor x_503_split_cast_fp16_0, tensor x_503_split_cast_fp16_1 = split(axis = x_503_split_axis_0, num_splits = x_503_split_num_splits_0, x = input_1187_cast_fp16)[name = tensor("x_503_split_cast_fp16")]; - tensor x_503_split_1_sigmoid_cast_fp16 = sigmoid(x = x_503_split_cast_fp16_1)[name = tensor("x_503_split_1_sigmoid_cast_fp16")]; - tensor x_503_cast_fp16 = mul(x = x_503_split_cast_fp16_0, y = x_503_split_1_sigmoid_cast_fp16)[name = tensor("x_503_cast_fp16")]; - tensor input_1189_cast_fp16 = select(a = var_11_to_fp16, b = x_503_cast_fp16, cond = var_328)[name = tensor("input_1189_cast_fp16")]; - tensor input_1191_pad_0 = const()[name = tensor("input_1191_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_1191_mode_0 = const()[name = tensor("input_1191_mode_0"), val = tensor("constant")]; - tensor const_237_to_fp16 = const()[name = tensor("const_237_to_fp16"), val = tensor(0x0p+0)]; - tensor input_1191_cast_fp16 = pad(constant_val = const_237_to_fp16, mode = input_1191_mode_0, pad = input_1191_pad_0, x = input_1189_cast_fp16)[name = tensor("input_1191_cast_fp16")]; - tensor input_1193_pad_type_0 = const()[name = tensor("input_1193_pad_type_0"), val = tensor("valid")]; - tensor input_1193_groups_0 = const()[name = tensor("input_1193_groups_0"), val = tensor(1024)]; - tensor input_1193_strides_0 = const()[name = tensor("input_1193_strides_0"), val = tensor([1])]; - tensor input_1193_pad_0 = const()[name = tensor("input_1193_pad_0"), val = tensor([0, 0])]; - tensor input_1193_dilations_0 = const()[name = tensor("input_1193_dilations_0"), val = tensor([1])]; - tensor const_292_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(419677440))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(419684416))), name = tensor("const_292_to_fp16_palettized"), shape = tensor([1024, 1, 9])]; - tensor const_293_to_fp16 = const()[name = tensor("const_293_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(419684608)))]; - tensor input_1195_cast_fp16 = conv(bias = const_293_to_fp16, dilations = input_1193_dilations_0, groups = input_1193_groups_0, pad = input_1193_pad_0, pad_type = input_1193_pad_type_0, strides = input_1193_strides_0, weight = const_292_to_fp16_palettized, x = input_1191_cast_fp16)[name = tensor("input_1195_cast_fp16")]; - tensor input_1197_cast_fp16 = silu(x = input_1195_cast_fp16)[name = tensor("input_1197_cast_fp16")]; - tensor x_505_pad_type_0 = const()[name = tensor("x_505_pad_type_0"), val = tensor("valid")]; - tensor x_505_strides_0 = const()[name = tensor("x_505_strides_0"), val = tensor([1])]; - tensor x_505_pad_0 = const()[name = tensor("x_505_pad_0"), val = tensor([0, 0])]; - tensor x_505_dilations_0 = const()[name = tensor("x_505_dilations_0"), val = tensor([1])]; - tensor x_505_groups_0 = const()[name = tensor("x_505_groups_0"), val = tensor(1)]; - tensor module_layers_22_conv_pointwise_conv2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(419686720))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(420473216))), name = tensor("module_layers_22_conv_pointwise_conv2_weight_to_fp16_palettized"), shape = tensor([1024, 1024, 1])]; - tensor x_505_cast_fp16 = conv(dilations = x_505_dilations_0, groups = x_505_groups_0, pad = x_505_pad_0, pad_type = x_505_pad_type_0, strides = x_505_strides_0, weight = module_layers_22_conv_pointwise_conv2_weight_to_fp16_palettized, x = input_1197_cast_fp16)[name = tensor("x_505_cast_fp16")]; - tensor input_1199_perm_0 = const()[name = tensor("input_1199_perm_0"), val = tensor([0, 2, 1])]; - tensor input_1199_cast_fp16 = transpose(perm = input_1199_perm_0, x = x_505_cast_fp16)[name = tensor("transpose_152")]; - tensor input_1201_cast_fp16 = add(x = input_1183_cast_fp16, y = input_1199_cast_fp16)[name = tensor("input_1201_cast_fp16")]; - tensor input_1203_axes_0 = const()[name = tensor("input_1203_axes_0"), val = tensor([-1])]; - tensor module_layers_22_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_22_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(420473408)))]; - tensor module_layers_22_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_22_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(420475520)))]; - tensor input_1203_cast_fp16 = layer_norm(axes = input_1203_axes_0, beta = module_layers_22_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_22_norm_feed_forward2_weight_to_fp16, x = input_1201_cast_fp16)[name = tensor("input_1203_cast_fp16")]; - tensor module_layers_22_feed_forward2_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(420477632))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(423623424))), name = tensor("module_layers_22_feed_forward2_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_206_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_22_feed_forward2_linear1_weight_to_fp16_palettized, x = input_1203_cast_fp16)[name = tensor("linear_206_cast_fp16")]; - tensor input_1207_cast_fp16 = silu(x = linear_206_cast_fp16)[name = tensor("input_1207_cast_fp16")]; - tensor module_layers_22_feed_forward2_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(423623616))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(426769408))), name = tensor("module_layers_22_feed_forward2_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_207_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_22_feed_forward2_linear2_weight_to_fp16_palettized, x = input_1207_cast_fp16)[name = tensor("linear_207_cast_fp16")]; - tensor var_3952_to_fp16 = const()[name = tensor("op_3952_to_fp16"), val = tensor(0x1p-1)]; - tensor var_3953_cast_fp16 = mul(x = linear_207_cast_fp16, y = var_3952_to_fp16)[name = tensor("op_3953_cast_fp16")]; - tensor input_1213_cast_fp16 = add(x = input_1201_cast_fp16, y = var_3953_cast_fp16)[name = tensor("input_1213_cast_fp16")]; - tensor input_1215_axes_0 = const()[name = tensor("input_1215_axes_0"), val = tensor([-1])]; - tensor module_layers_22_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_22_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(426769600)))]; - tensor module_layers_22_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_22_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(426771712)))]; - tensor input_1215_cast_fp16 = layer_norm(axes = input_1215_axes_0, beta = module_layers_22_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_22_norm_out_weight_to_fp16, x = input_1213_cast_fp16)[name = tensor("input_1215_cast_fp16")]; - tensor input_1217_axes_0 = const()[name = tensor("input_1217_axes_0"), val = tensor([-1])]; - tensor module_layers_23_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_23_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(426773824)))]; - tensor module_layers_23_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_23_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(426775936)))]; - tensor input_1217_cast_fp16 = layer_norm(axes = input_1217_axes_0, beta = module_layers_23_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_23_norm_feed_forward1_weight_to_fp16, x = input_1215_cast_fp16)[name = tensor("input_1217_cast_fp16")]; - tensor module_layers_23_feed_forward1_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(426778048))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(429923840))), name = tensor("module_layers_23_feed_forward1_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_208_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_23_feed_forward1_linear1_weight_to_fp16_palettized, x = input_1217_cast_fp16)[name = tensor("linear_208_cast_fp16")]; - tensor input_1221_cast_fp16 = silu(x = linear_208_cast_fp16)[name = tensor("input_1221_cast_fp16")]; - tensor module_layers_23_feed_forward1_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(429924032))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(433069824))), name = tensor("module_layers_23_feed_forward1_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_209_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_23_feed_forward1_linear2_weight_to_fp16_palettized, x = input_1221_cast_fp16)[name = tensor("linear_209_cast_fp16")]; - tensor var_3981_to_fp16 = const()[name = tensor("op_3981_to_fp16"), val = tensor(0x1p-1)]; - tensor var_3982_cast_fp16 = mul(x = linear_209_cast_fp16, y = var_3981_to_fp16)[name = tensor("op_3982_cast_fp16")]; - tensor input_1227_cast_fp16 = add(x = input_1215_cast_fp16, y = var_3982_cast_fp16)[name = tensor("input_1227_cast_fp16")]; - tensor query_axes_0 = const()[name = tensor("query_axes_0"), val = tensor([-1])]; - tensor module_layers_23_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_23_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(433070016)))]; - tensor module_layers_23_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_23_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(433072128)))]; - tensor query_cast_fp16 = layer_norm(axes = query_axes_0, beta = module_layers_23_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_23_norm_self_att_weight_to_fp16, x = input_1227_cast_fp16)[name = tensor("query_cast_fp16")]; - tensor module_layers_23_self_attn_linear_q_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(433074240))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(433860736))), name = tensor("module_layers_23_self_attn_linear_q_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_210_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_23_self_attn_linear_q_weight_to_fp16_palettized, x = query_cast_fp16)[name = tensor("linear_210_cast_fp16")]; - tensor var_3998 = const()[name = tensor("op_3998"), val = tensor([1, -1, 8, 128])]; - tensor q_139_cast_fp16 = reshape(shape = var_3998, x = linear_210_cast_fp16)[name = tensor("q_139_cast_fp16")]; - tensor module_layers_23_self_attn_linear_k_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(433860928))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(434647424))), name = tensor("module_layers_23_self_attn_linear_k_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_211_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_23_self_attn_linear_k_weight_to_fp16_palettized, x = query_cast_fp16)[name = tensor("linear_211_cast_fp16")]; - tensor var_4002 = const()[name = tensor("op_4002"), val = tensor([1, -1, 8, 128])]; - tensor k_93_cast_fp16 = reshape(shape = var_4002, x = linear_211_cast_fp16)[name = tensor("k_93_cast_fp16")]; - tensor module_layers_23_self_attn_linear_v_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(434647616))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(435434112))), name = tensor("module_layers_23_self_attn_linear_v_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_212_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_23_self_attn_linear_v_weight_to_fp16_palettized, x = query_cast_fp16)[name = tensor("linear_212_cast_fp16")]; - tensor var_4006 = const()[name = tensor("op_4006"), val = tensor([1, -1, 8, 128])]; - tensor v_cast_fp16 = reshape(shape = var_4006, x = linear_212_cast_fp16)[name = tensor("v_cast_fp16")]; - tensor value_perm_0 = const()[name = tensor("value_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor module_layers_23_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_23_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(435434304)))]; - tensor var_4018_cast_fp16 = add(x = q_139_cast_fp16, y = module_layers_23_self_attn_pos_bias_u_to_fp16)[name = tensor("op_4018_cast_fp16")]; - tensor module_layers_23_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_23_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(435436416)))]; - tensor var_4020_cast_fp16 = add(x = q_139_cast_fp16, y = module_layers_23_self_attn_pos_bias_v_to_fp16)[name = tensor("op_4020_cast_fp16")]; - tensor q_with_bias_v_perm_0 = const()[name = tensor("q_with_bias_v_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor x_513_transpose_x_0 = const()[name = tensor("x_513_transpose_x_0"), val = tensor(false)]; - tensor x_513_transpose_y_0 = const()[name = tensor("x_513_transpose_y_0"), val = tensor(false)]; - tensor op_4022_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(435438528))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(435726592))), name = tensor("op_4022_to_fp16_palettized"), shape = tensor([1, 8, 128, 375])]; - tensor q_with_bias_v_cast_fp16 = transpose(perm = q_with_bias_v_perm_0, x = var_4020_cast_fp16)[name = tensor("transpose_151")]; - tensor x_513_cast_fp16 = matmul(transpose_x = x_513_transpose_x_0, transpose_y = x_513_transpose_y_0, x = q_with_bias_v_cast_fp16, y = op_4022_to_fp16_palettized)[name = tensor("x_513_cast_fp16")]; - tensor x_515_pad_0 = const()[name = tensor("x_515_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_515_mode_0 = const()[name = tensor("x_515_mode_0"), val = tensor("constant")]; - tensor const_244_to_fp16 = const()[name = tensor("const_244_to_fp16"), val = tensor(0x0p+0)]; - tensor x_515_cast_fp16 = pad(constant_val = const_244_to_fp16, mode = x_515_mode_0, pad = x_515_pad_0, x = x_513_cast_fp16)[name = tensor("x_515_cast_fp16")]; - tensor var_4030 = const()[name = tensor("op_4030"), val = tensor([1, 8, -1, 188])]; - tensor x_517_cast_fp16 = reshape(shape = var_4030, x = x_515_cast_fp16)[name = tensor("x_517_cast_fp16")]; - tensor var_4034_begin_0 = const()[name = tensor("op_4034_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_4034_end_0 = const()[name = tensor("op_4034_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_4034_end_mask_0 = const()[name = tensor("op_4034_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_4034_cast_fp16 = slice_by_index(begin = var_4034_begin_0, end = var_4034_end_0, end_mask = var_4034_end_mask_0, x = x_517_cast_fp16)[name = tensor("op_4034_cast_fp16")]; - tensor var_4035 = const()[name = tensor("op_4035"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_93_cast_fp16 = reshape(shape = var_4035, x = var_4034_cast_fp16)[name = tensor("matrix_bd_93_cast_fp16")]; - tensor matrix_ac_transpose_x_0 = const()[name = tensor("matrix_ac_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_transpose_y_0 = const()[name = tensor("matrix_ac_transpose_y_0"), val = tensor(false)]; - tensor transpose_142_perm_0 = const()[name = tensor("transpose_142_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_143_perm_0 = const()[name = tensor("transpose_143_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_143 = transpose(perm = transpose_143_perm_0, x = k_93_cast_fp16)[name = tensor("transpose_149")]; - tensor transpose_142 = transpose(perm = transpose_142_perm_0, x = var_4018_cast_fp16)[name = tensor("transpose_150")]; - tensor matrix_ac_cast_fp16 = matmul(transpose_x = matrix_ac_transpose_x_0, transpose_y = matrix_ac_transpose_y_0, x = transpose_142, y = transpose_143)[name = tensor("matrix_ac_cast_fp16")]; - tensor matrix_bd_begin_0 = const()[name = tensor("matrix_bd_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_end_0 = const()[name = tensor("matrix_bd_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_end_mask_0 = const()[name = tensor("matrix_bd_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_cast_fp16 = slice_by_index(begin = matrix_bd_begin_0, end = matrix_bd_end_0, end_mask = matrix_bd_end_mask_0, x = matrix_bd_93_cast_fp16)[name = tensor("matrix_bd_cast_fp16")]; - tensor var_4044_cast_fp16 = add(x = matrix_ac_cast_fp16, y = matrix_bd_cast_fp16)[name = tensor("op_4044_cast_fp16")]; - tensor _inversed_scores_93_y_0_to_fp16 = const()[name = tensor("_inversed_scores_93_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; - tensor _inversed_scores_93_cast_fp16 = mul(x = var_4044_cast_fp16, y = _inversed_scores_93_y_0_to_fp16)[name = tensor("_inversed_scores_93_cast_fp16")]; - tensor scores_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_93_cast_fp16, cond = mask_3)[name = tensor("scores_cast_fp16")]; - tensor var_4050_cast_fp16 = softmax(axis = var_30, x = scores_cast_fp16)[name = tensor("op_4050_cast_fp16")]; - tensor input_1229_cast_fp16 = select(a = var_11_to_fp16, b = var_4050_cast_fp16, cond = mask_3)[name = tensor("input_1229_cast_fp16")]; - tensor x_519_transpose_x_0 = const()[name = tensor("x_519_transpose_x_0"), val = tensor(false)]; - tensor x_519_transpose_y_0 = const()[name = tensor("x_519_transpose_y_0"), val = tensor(false)]; - tensor value_cast_fp16 = transpose(perm = value_perm_0, x = v_cast_fp16)[name = tensor("transpose_148")]; - tensor x_519_cast_fp16 = matmul(transpose_x = x_519_transpose_x_0, transpose_y = x_519_transpose_y_0, x = input_1229_cast_fp16, y = value_cast_fp16)[name = tensor("x_519_cast_fp16")]; - tensor var_4054_perm_0 = const()[name = tensor("op_4054_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_4055 = const()[name = tensor("op_4055"), val = tensor([1, -1, 1024])]; - tensor var_4054_cast_fp16 = transpose(perm = var_4054_perm_0, x = x_519_cast_fp16)[name = tensor("transpose_147")]; - tensor input_1231_cast_fp16 = reshape(shape = var_4055, x = var_4054_cast_fp16)[name = tensor("input_1231_cast_fp16")]; - tensor module_layers_23_self_attn_linear_out_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(435726784))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(436513280))), name = tensor("module_layers_23_self_attn_linear_out_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_214_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_23_self_attn_linear_out_weight_to_fp16_palettized, x = input_1231_cast_fp16)[name = tensor("linear_214_cast_fp16")]; - tensor input_1235_cast_fp16 = add(x = input_1227_cast_fp16, y = linear_214_cast_fp16)[name = tensor("input_1235_cast_fp16")]; - tensor x_523_axes_0 = const()[name = tensor("x_523_axes_0"), val = tensor([-1])]; - tensor module_layers_23_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_23_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(436513472)))]; - tensor module_layers_23_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_23_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(436515584)))]; - tensor x_523_cast_fp16 = layer_norm(axes = x_523_axes_0, beta = module_layers_23_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_23_norm_conv_weight_to_fp16, x = input_1235_cast_fp16)[name = tensor("x_523_cast_fp16")]; - tensor input_1237_perm_0 = const()[name = tensor("input_1237_perm_0"), val = tensor([0, 2, 1])]; - tensor input_1239_pad_type_0 = const()[name = tensor("input_1239_pad_type_0"), val = tensor("valid")]; - tensor input_1239_strides_0 = const()[name = tensor("input_1239_strides_0"), val = tensor([1])]; - tensor input_1239_pad_0 = const()[name = tensor("input_1239_pad_0"), val = tensor([0, 0])]; - tensor input_1239_dilations_0 = const()[name = tensor("input_1239_dilations_0"), val = tensor([1])]; - tensor input_1239_groups_0 = const()[name = tensor("input_1239_groups_0"), val = tensor(1)]; - tensor module_layers_23_conv_pointwise_conv1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(436517696))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(438090624))), name = tensor("module_layers_23_conv_pointwise_conv1_weight_to_fp16_palettized"), shape = tensor([2048, 1024, 1])]; - tensor input_1237_cast_fp16 = transpose(perm = input_1237_perm_0, x = x_523_cast_fp16)[name = tensor("transpose_146")]; - tensor input_1239_cast_fp16 = conv(dilations = input_1239_dilations_0, groups = input_1239_groups_0, pad = input_1239_pad_0, pad_type = input_1239_pad_type_0, strides = input_1239_strides_0, weight = module_layers_23_conv_pointwise_conv1_weight_to_fp16_palettized, x = input_1237_cast_fp16)[name = tensor("input_1239_cast_fp16")]; - tensor x_525_split_num_splits_0 = const()[name = tensor("x_525_split_num_splits_0"), val = tensor(2)]; - tensor x_525_split_axis_0 = const()[name = tensor("x_525_split_axis_0"), val = tensor(1)]; - tensor x_525_split_cast_fp16_0, tensor x_525_split_cast_fp16_1 = split(axis = x_525_split_axis_0, num_splits = x_525_split_num_splits_0, x = input_1239_cast_fp16)[name = tensor("x_525_split_cast_fp16")]; - tensor x_525_split_1_sigmoid_cast_fp16 = sigmoid(x = x_525_split_cast_fp16_1)[name = tensor("x_525_split_1_sigmoid_cast_fp16")]; - tensor x_525_cast_fp16 = mul(x = x_525_split_cast_fp16_0, y = x_525_split_1_sigmoid_cast_fp16)[name = tensor("x_525_cast_fp16")]; - tensor input_1241_cast_fp16 = select(a = var_11_to_fp16, b = x_525_cast_fp16, cond = var_328)[name = tensor("input_1241_cast_fp16")]; - tensor input_1243_pad_0 = const()[name = tensor("input_1243_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_1243_mode_0 = const()[name = tensor("input_1243_mode_0"), val = tensor("constant")]; - tensor const_247_to_fp16 = const()[name = tensor("const_247_to_fp16"), val = tensor(0x0p+0)]; - tensor input_1243_cast_fp16 = pad(constant_val = const_247_to_fp16, mode = input_1243_mode_0, pad = input_1243_pad_0, x = input_1241_cast_fp16)[name = tensor("input_1243_cast_fp16")]; - tensor input_1245_pad_type_0 = const()[name = tensor("input_1245_pad_type_0"), val = tensor("valid")]; - tensor input_1245_groups_0 = const()[name = tensor("input_1245_groups_0"), val = tensor(1024)]; - tensor input_1245_strides_0 = const()[name = tensor("input_1245_strides_0"), val = tensor([1])]; - tensor input_1245_pad_0 = const()[name = tensor("input_1245_pad_0"), val = tensor([0, 0])]; - tensor input_1245_dilations_0 = const()[name = tensor("input_1245_dilations_0"), val = tensor([1])]; - tensor const_294_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(438090816))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(438097792))), name = tensor("const_294_to_fp16_palettized"), shape = tensor([1024, 1, 9])]; - tensor const_295_to_fp16 = const()[name = tensor("const_295_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(438097984)))]; - tensor input_1247_cast_fp16 = conv(bias = const_295_to_fp16, dilations = input_1245_dilations_0, groups = input_1245_groups_0, pad = input_1245_pad_0, pad_type = input_1245_pad_type_0, strides = input_1245_strides_0, weight = const_294_to_fp16_palettized, x = input_1243_cast_fp16)[name = tensor("input_1247_cast_fp16")]; - tensor input_1249_cast_fp16 = silu(x = input_1247_cast_fp16)[name = tensor("input_1249_cast_fp16")]; - tensor x_527_pad_type_0 = const()[name = tensor("x_527_pad_type_0"), val = tensor("valid")]; - tensor x_527_strides_0 = const()[name = tensor("x_527_strides_0"), val = tensor([1])]; - tensor x_527_pad_0 = const()[name = tensor("x_527_pad_0"), val = tensor([0, 0])]; - tensor x_527_dilations_0 = const()[name = tensor("x_527_dilations_0"), val = tensor([1])]; - tensor x_527_groups_0 = const()[name = tensor("x_527_groups_0"), val = tensor(1)]; - tensor module_layers_23_conv_pointwise_conv2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(438100096))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(438886592))), name = tensor("module_layers_23_conv_pointwise_conv2_weight_to_fp16_palettized"), shape = tensor([1024, 1024, 1])]; - tensor x_527_cast_fp16 = conv(dilations = x_527_dilations_0, groups = x_527_groups_0, pad = x_527_pad_0, pad_type = x_527_pad_type_0, strides = x_527_strides_0, weight = module_layers_23_conv_pointwise_conv2_weight_to_fp16_palettized, x = input_1249_cast_fp16)[name = tensor("x_527_cast_fp16")]; - tensor input_1251_perm_0 = const()[name = tensor("input_1251_perm_0"), val = tensor([0, 2, 1])]; - tensor input_1251_cast_fp16 = transpose(perm = input_1251_perm_0, x = x_527_cast_fp16)[name = tensor("transpose_145")]; - tensor input_1253_cast_fp16 = add(x = input_1235_cast_fp16, y = input_1251_cast_fp16)[name = tensor("input_1253_cast_fp16")]; - tensor input_1255_axes_0 = const()[name = tensor("input_1255_axes_0"), val = tensor([-1])]; - tensor module_layers_23_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_23_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(438886784)))]; - tensor module_layers_23_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_23_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(438888896)))]; - tensor input_1255_cast_fp16 = layer_norm(axes = input_1255_axes_0, beta = module_layers_23_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_23_norm_feed_forward2_weight_to_fp16, x = input_1253_cast_fp16)[name = tensor("input_1255_cast_fp16")]; - tensor module_layers_23_feed_forward2_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(438891008))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(442036800))), name = tensor("module_layers_23_feed_forward2_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_215_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_23_feed_forward2_linear1_weight_to_fp16_palettized, x = input_1255_cast_fp16)[name = tensor("linear_215_cast_fp16")]; - tensor input_1259_cast_fp16 = silu(x = linear_215_cast_fp16)[name = tensor("input_1259_cast_fp16")]; - tensor module_layers_23_feed_forward2_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(442036992))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(445182784))), name = tensor("module_layers_23_feed_forward2_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_216_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_23_feed_forward2_linear2_weight_to_fp16_palettized, x = input_1259_cast_fp16)[name = tensor("linear_216_cast_fp16")]; - tensor var_4115_to_fp16 = const()[name = tensor("op_4115_to_fp16"), val = tensor(0x1p-1)]; - tensor var_4116_cast_fp16 = mul(x = linear_216_cast_fp16, y = var_4115_to_fp16)[name = tensor("op_4116_cast_fp16")]; - tensor input_cast_fp16 = add(x = input_1253_cast_fp16, y = var_4116_cast_fp16)[name = tensor("input_cast_fp16")]; - tensor audio_signal_axes_0 = const()[name = tensor("audio_signal_axes_0"), val = tensor([-1])]; - tensor module_layers_23_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_23_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(445182976)))]; - tensor module_layers_23_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_23_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(445185088)))]; - tensor audio_signal_cast_fp16 = layer_norm(axes = audio_signal_axes_0, beta = module_layers_23_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_23_norm_out_weight_to_fp16, x = input_cast_fp16)[name = tensor("audio_signal_cast_fp16")]; - tensor obj_1_perm_0 = const()[name = tensor("obj_1_perm_0"), val = tensor([0, 2, 1])]; - tensor obj_1_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("obj_1_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; - tensor obj_1_cast_fp16 = transpose(perm = obj_1_perm_0, x = audio_signal_cast_fp16)[name = tensor("transpose_144")]; - tensor encoder = cast(dtype = obj_1_cast_fp16_to_fp32_dtype_0, x = obj_1_cast_fp16)[name = tensor("cast_0")]; - } -> (encoder, encoder_length); -} \ No newline at end of file diff --git a/models/parakeet-tdt-0.6b-v2-coreml/Encoder.mlmodelc/weights/weight.bin b/models/parakeet-tdt-0.6b-v2-coreml/Encoder.mlmodelc/weights/weight.bin deleted file mode 100644 index e6313f81d9d94af46347f28087f51a60eb28164c..0000000000000000000000000000000000000000 --- a/models/parakeet-tdt-0.6b-v2-coreml/Encoder.mlmodelc/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:4adc7ad44f9d05e1bffeb2b06d3bb02861a5c7602dff63a6b494aed3bf8a6c3e -size 445187200 diff --git a/models/parakeet-tdt-0.6b-v2-coreml/JointDecision.mlmodelc/analytics/coremldata.bin b/models/parakeet-tdt-0.6b-v2-coreml/JointDecision.mlmodelc/analytics/coremldata.bin deleted file mode 100644 index 5007841fb604e70990e535a2a002ee18f93e83e4..0000000000000000000000000000000000000000 --- a/models/parakeet-tdt-0.6b-v2-coreml/JointDecision.mlmodelc/analytics/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:f1183ba213bb94a918c8d2cad19ab045320618f97f6ca662245b3936d7b090f7 -size 243 diff --git a/models/parakeet-tdt-0.6b-v2-coreml/JointDecision.mlmodelc/coremldata.bin b/models/parakeet-tdt-0.6b-v2-coreml/JointDecision.mlmodelc/coremldata.bin deleted file mode 100644 index 72290f46bfb535192ed1ed5ca1fec941874efbca..0000000000000000000000000000000000000000 --- a/models/parakeet-tdt-0.6b-v2-coreml/JointDecision.mlmodelc/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:e2c6752f1c8cf2d3f6f26ec93195c9bfa759ad59edf9f806696a138154f96f11 -size 534 diff --git a/models/parakeet-tdt-0.6b-v2-coreml/JointDecision.mlmodelc/metadata.json b/models/parakeet-tdt-0.6b-v2-coreml/JointDecision.mlmodelc/metadata.json deleted file mode 100644 index db3b2bc33cfda6f7cd1073d21182446180e53f94..0000000000000000000000000000000000000000 --- a/models/parakeet-tdt-0.6b-v2-coreml/JointDecision.mlmodelc/metadata.json +++ /dev/null @@ -1,103 +0,0 @@ -[ - { - "metadataOutputVersion" : "3.0", - "shortDescription" : "Parakeet single-step joint decision (current frame)", - "outputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1 × 1 × 1)", - "shortDescription" : "", - "shape" : "[1, 1, 1]", - "name" : "token_id", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 1 × 1)", - "shortDescription" : "", - "shape" : "[1, 1, 1]", - "name" : "token_prob", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1 × 1 × 1)", - "shortDescription" : "", - "shape" : "[1, 1, 1]", - "name" : "duration", - "type" : "MultiArray" - } - ], - "storagePrecision" : "Float16", - "modelParameters" : [ - - ], - "author" : "Fluid Inference", - "specificationVersion" : 8, - "mlProgramOperationTypeHistogram" : { - "Ios17.reduceArgmax" : 2, - "Ios17.squeeze" : 1, - "Ios17.cast" : 4, - "Ios17.linear" : 3, - "Ios17.transpose" : 2, - "Ios17.sliceByIndex" : 2, - "Ios17.add" : 1, - "Ios16.relu" : 1, - "Ios16.softmax" : 1, - "Ios17.gatherAlongAxis" : 1, - "Ios17.expandDims" : 3 - }, - "computePrecision" : "Mixed (Float16, Float32, Int16, Int32)", - "isUpdatable" : "0", - "stateSchema" : [ - - ], - "availability" : { - "macOS" : "14.0", - "tvOS" : "17.0", - "visionOS" : "1.0", - "watchOS" : "10.0", - "iOS" : "17.0", - "macCatalyst" : "17.0" - }, - "modelType" : { - "name" : "MLModelType_mlProgram" - }, - "inputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 1024 × 1)", - "shortDescription" : "", - "shape" : "[1, 1024, 1]", - "name" : "encoder_step", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 640 × 1)", - "shortDescription" : "", - "shape" : "[1, 640, 1]", - "name" : "decoder_step", - "type" : "MultiArray" - } - ], - "userDefinedMetadata" : { - "com.github.apple.coremltools.conversion_date" : "2025-09-25", - "com.github.apple.coremltools.source" : "torch==2.7.0", - "com.github.apple.coremltools.version" : "9.0b1", - "com.github.apple.coremltools.source_dialect" : "TorchScript" - }, - "generatedClassName" : "parakeet_joint_decision_single_step", - "method" : "predict" - } -] \ No newline at end of file diff --git a/models/parakeet-tdt-0.6b-v2-coreml/JointDecision.mlmodelc/model.mil b/models/parakeet-tdt-0.6b-v2-coreml/JointDecision.mlmodelc/model.mil deleted file mode 100644 index 339c35b303b074de7fe047bb099311a4cc08cf53..0000000000000000000000000000000000000000 --- a/models/parakeet-tdt-0.6b-v2-coreml/JointDecision.mlmodelc/model.mil +++ /dev/null @@ -1,58 +0,0 @@ -program(1.0) -[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3500.14.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.7.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "9.0b1"}})] -{ - func main(tensor decoder_step, tensor encoder_step) { - tensor input_1_perm_0 = const()[name = tensor("input_1_perm_0"), val = tensor([0, 2, 1])]; - tensor encoder_step_to_fp16_dtype_0 = const()[name = tensor("encoder_step_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor input_3_perm_0 = const()[name = tensor("input_3_perm_0"), val = tensor([0, 2, 1])]; - tensor decoder_step_to_fp16_dtype_0 = const()[name = tensor("decoder_step_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor joint_module_enc_weight_to_fp16 = const()[name = tensor("joint_module_enc_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; - tensor joint_module_enc_bias_to_fp16 = const()[name = tensor("joint_module_enc_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1310848)))]; - tensor encoder_step_to_fp16 = cast(dtype = encoder_step_to_fp16_dtype_0, x = encoder_step)[name = tensor("cast_3")]; - tensor input_1_cast_fp16 = transpose(perm = input_1_perm_0, x = encoder_step_to_fp16)[name = tensor("transpose_1")]; - tensor linear_0_cast_fp16 = linear(bias = joint_module_enc_bias_to_fp16, weight = joint_module_enc_weight_to_fp16, x = input_1_cast_fp16)[name = tensor("linear_0_cast_fp16")]; - tensor joint_module_pred_weight_to_fp16 = const()[name = tensor("joint_module_pred_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1312192)))]; - tensor joint_module_pred_bias_to_fp16 = const()[name = tensor("joint_module_pred_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2131456)))]; - tensor decoder_step_to_fp16 = cast(dtype = decoder_step_to_fp16_dtype_0, x = decoder_step)[name = tensor("cast_2")]; - tensor input_3_cast_fp16 = transpose(perm = input_3_perm_0, x = decoder_step_to_fp16)[name = tensor("transpose_0")]; - tensor linear_1_cast_fp16 = linear(bias = joint_module_pred_bias_to_fp16, weight = joint_module_pred_weight_to_fp16, x = input_3_cast_fp16)[name = tensor("linear_1_cast_fp16")]; - tensor var_23_axes_0 = const()[name = tensor("op_23_axes_0"), val = tensor([2])]; - tensor var_23_cast_fp16 = expand_dims(axes = var_23_axes_0, x = linear_0_cast_fp16)[name = tensor("op_23_cast_fp16")]; - tensor var_24_axes_0 = const()[name = tensor("op_24_axes_0"), val = tensor([1])]; - tensor var_24_cast_fp16 = expand_dims(axes = var_24_axes_0, x = linear_1_cast_fp16)[name = tensor("op_24_cast_fp16")]; - tensor input_5_cast_fp16 = add(x = var_23_cast_fp16, y = var_24_cast_fp16)[name = tensor("input_5_cast_fp16")]; - tensor input_7_cast_fp16 = relu(x = input_5_cast_fp16)[name = tensor("input_7_cast_fp16")]; - tensor joint_module_joint_net_2_weight_to_fp16 = const()[name = tensor("joint_module_joint_net_2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2132800)))]; - tensor joint_module_joint_net_2_bias_to_fp16 = const()[name = tensor("joint_module_joint_net_2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3451264)))]; - tensor linear_2_cast_fp16 = linear(bias = joint_module_joint_net_2_bias_to_fp16, weight = joint_module_joint_net_2_weight_to_fp16, x = input_7_cast_fp16)[name = tensor("linear_2_cast_fp16")]; - tensor token_logits_begin_0 = const()[name = tensor("token_logits_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor token_logits_end_0 = const()[name = tensor("token_logits_end_0"), val = tensor([1, 1, 1, 1025])]; - tensor token_logits_end_mask_0 = const()[name = tensor("token_logits_end_mask_0"), val = tensor([true, true, true, false])]; - tensor token_logits_cast_fp16 = slice_by_index(begin = token_logits_begin_0, end = token_logits_end_0, end_mask = token_logits_end_mask_0, x = linear_2_cast_fp16)[name = tensor("token_logits_cast_fp16")]; - tensor duration_logits_begin_0 = const()[name = tensor("duration_logits_begin_0"), val = tensor([0, 0, 0, 1025])]; - tensor duration_logits_end_0 = const()[name = tensor("duration_logits_end_0"), val = tensor([1, 1, 1, 1030])]; - tensor duration_logits_end_mask_0 = const()[name = tensor("duration_logits_end_mask_0"), val = tensor([true, true, true, true])]; - tensor duration_logits_cast_fp16 = slice_by_index(begin = duration_logits_begin_0, end = duration_logits_end_0, end_mask = duration_logits_end_mask_0, x = linear_2_cast_fp16)[name = tensor("duration_logits_cast_fp16")]; - tensor var_43_axis_0 = const()[name = tensor("op_43_axis_0"), val = tensor(-1)]; - tensor var_43_keep_dims_0 = const()[name = tensor("op_43_keep_dims_0"), val = tensor(false)]; - tensor var_43_output_dtype_0 = const()[name = tensor("op_43_output_dtype_0"), val = tensor("int32")]; - tensor token_id = reduce_argmax(axis = var_43_axis_0, keep_dims = var_43_keep_dims_0, output_dtype = var_43_output_dtype_0, x = token_logits_cast_fp16)[name = tensor("op_43_cast_fp16")]; - tensor var_49 = const()[name = tensor("op_49"), val = tensor(-1)]; - tensor token_probs_all_cast_fp16 = softmax(axis = var_49, x = token_logits_cast_fp16)[name = tensor("token_probs_all_cast_fp16")]; - tensor var_58_axes_0 = const()[name = tensor("op_58_axes_0"), val = tensor([-1])]; - tensor var_58 = expand_dims(axes = var_58_axes_0, x = token_id)[name = tensor("op_58")]; - tensor var_59 = const()[name = tensor("op_59"), val = tensor(-1)]; - tensor var_61_validate_indices_0 = const()[name = tensor("op_61_validate_indices_0"), val = tensor(false)]; - tensor var_58_to_int16_dtype_0 = const()[name = tensor("op_58_to_int16_dtype_0"), val = tensor("int16")]; - tensor var_58_to_int16 = cast(dtype = var_58_to_int16_dtype_0, x = var_58)[name = tensor("cast_1")]; - tensor var_61_cast_fp16_cast_int16 = gather_along_axis(axis = var_59, indices = var_58_to_int16, validate_indices = var_61_validate_indices_0, x = token_probs_all_cast_fp16)[name = tensor("op_61_cast_fp16_cast_int16")]; - tensor var_63_axes_0 = const()[name = tensor("op_63_axes_0"), val = tensor([-1])]; - tensor var_63_cast_fp16 = squeeze(axes = var_63_axes_0, x = var_61_cast_fp16_cast_int16)[name = tensor("op_63_cast_fp16")]; - tensor var_63_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("op_63_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; - tensor var_66_axis_0 = const()[name = tensor("op_66_axis_0"), val = tensor(-1)]; - tensor var_66_keep_dims_0 = const()[name = tensor("op_66_keep_dims_0"), val = tensor(false)]; - tensor var_66_output_dtype_0 = const()[name = tensor("op_66_output_dtype_0"), val = tensor("int32")]; - tensor duration = reduce_argmax(axis = var_66_axis_0, keep_dims = var_66_keep_dims_0, output_dtype = var_66_output_dtype_0, x = duration_logits_cast_fp16)[name = tensor("op_66_cast_fp16")]; - tensor token_prob = cast(dtype = var_63_cast_fp16_to_fp32_dtype_0, x = var_63_cast_fp16)[name = tensor("cast_0")]; - } -> (token_id, token_prob, duration); -} \ No newline at end of file diff --git a/models/parakeet-tdt-0.6b-v2-coreml/JointDecision.mlmodelc/weights/weight.bin b/models/parakeet-tdt-0.6b-v2-coreml/JointDecision.mlmodelc/weights/weight.bin deleted file mode 100644 index 8764a151a786b38079c36a409fe5144ddf8b95a9..0000000000000000000000000000000000000000 --- a/models/parakeet-tdt-0.6b-v2-coreml/JointDecision.mlmodelc/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:ca22a65903a05e64137677da608077578a8606090a598abf4875fa6199aaa19d -size 3453388 diff --git a/models/parakeet-tdt-0.6b-v2-coreml/Preprocessor.mlmodelc/analytics/coremldata.bin b/models/parakeet-tdt-0.6b-v2-coreml/Preprocessor.mlmodelc/analytics/coremldata.bin deleted file mode 100644 index 84e7c818c15d2650b6c55b20e618fba3e6289763..0000000000000000000000000000000000000000 --- a/models/parakeet-tdt-0.6b-v2-coreml/Preprocessor.mlmodelc/analytics/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:03ab3c1327a054c54c07a40325db967ec574f2c91dcc8192bfa44aa561bcf2d8 -size 243 diff --git a/models/parakeet-tdt-0.6b-v2-coreml/Preprocessor.mlmodelc/coremldata.bin b/models/parakeet-tdt-0.6b-v2-coreml/Preprocessor.mlmodelc/coremldata.bin deleted file mode 100644 index b53fff2bc93fbe4b468811db15aa17767f3cb33c..0000000000000000000000000000000000000000 --- a/models/parakeet-tdt-0.6b-v2-coreml/Preprocessor.mlmodelc/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:d88ea1fc349459c9e100d6a96688c5b29a1f0d865f544be103001724b986b6d6 -size 494 diff --git a/models/parakeet-tdt-0.6b-v2-coreml/Preprocessor.mlmodelc/metadata.json b/models/parakeet-tdt-0.6b-v2-coreml/Preprocessor.mlmodelc/metadata.json deleted file mode 100644 index 887f7d4adc6d7f933686f7b984e23a47b45f57e3..0000000000000000000000000000000000000000 --- a/models/parakeet-tdt-0.6b-v2-coreml/Preprocessor.mlmodelc/metadata.json +++ /dev/null @@ -1,110 +0,0 @@ -[ - { - "metadataOutputVersion" : "3.0", - "shortDescription" : "int8-linear quantized - preprocessor", - "outputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32)", - "shortDescription" : "", - "shape" : "[]", - "name" : "mel", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1)", - "shortDescription" : "", - "shape" : "[1]", - "name" : "mel_length", - "type" : "MultiArray" - } - ], - "storagePrecision" : "Int8", - "modelParameters" : [ - - ], - "author" : "Fluid Inference", - "specificationVersion" : 8, - "mlProgramOperationTypeHistogram" : { - "Range1d" : 2, - "Ios17.reshape" : 2, - "Ios17.matmul" : 1, - "Ios17.expandDims" : 10, - "Select" : 3, - "Ios17.add" : 4, - "Tile" : 2, - "Ios17.sliceByIndex" : 3, - "Ios16.reduceSum" : 4, - "Shape" : 3, - "Ios17.gather" : 3, - "Pad" : 1, - "Ios17.log" : 1, - "Ios16.constexprAffineDequantize" : 3, - "Ios17.conv" : 2, - "Ios17.sub" : 4, - "Ios17.pow" : 2, - "Ios17.cast" : 10, - "Ios17.realDiv" : 4, - "Stack" : 1, - "Ios17.concat" : 3, - "Ios17.floorDiv" : 1, - "Ios17.less" : 1, - "Ios17.sqrt" : 1, - "Ios17.greaterEqual" : 1, - "Ios17.mul" : 1 - }, - "computePrecision" : "Mixed (Float16, Float32, Int32, UInt16)", - "isUpdatable" : "0", - "stateSchema" : [ - - ], - "availability" : { - "macOS" : "14.0", - "tvOS" : "17.0", - "visionOS" : "1.0", - "watchOS" : "10.0", - "iOS" : "17.0", - "macCatalyst" : "17.0" - }, - "modelType" : { - "name" : "MLModelType_mlProgram" - }, - "inputSchema" : [ - { - "dataType" : "Float32", - "hasShapeFlexibility" : "1", - "isOptional" : "0", - "shapeFlexibility" : "1 × 1...240000", - "shapeRange" : "[[1, 1], [1, 240000]]", - "formattedType" : "MultiArray (Float32 1 × 1)", - "type" : "MultiArray", - "shape" : "[1, 1]", - "name" : "audio_signal", - "shortDescription" : "" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1)", - "shortDescription" : "", - "shape" : "[1]", - "name" : "audio_length", - "type" : "MultiArray" - } - ], - "userDefinedMetadata" : { - "com.github.apple.coremltools.conversion_date" : "2025-09-25", - "com.github.apple.coremltools.source" : "torch==2.7.0", - "com.github.apple.coremltools.version" : "9.0b1", - "com.github.apple.coremltools.source_dialect" : "TorchScript" - }, - "generatedClassName" : "parakeet_preprocessor", - "method" : "predict" - } -] \ No newline at end of file diff --git a/models/parakeet-tdt-0.6b-v2-coreml/Preprocessor.mlmodelc/model.mil b/models/parakeet-tdt-0.6b-v2-coreml/Preprocessor.mlmodelc/model.mil deleted file mode 100644 index e94cce4bc7fbf55e55788449bf067ffe09a5ca91..0000000000000000000000000000000000000000 --- a/models/parakeet-tdt-0.6b-v2-coreml/Preprocessor.mlmodelc/model.mil +++ /dev/null @@ -1,169 +0,0 @@ -program(1.0) -[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3500.14.1"}, {"coremlc-version", "3500.32.1"}})] -{ - func main(tensor audio_length, tensor audio_signal) [FlexibleShapeInformation = tuple, dict, tensor>>, tuple, dict, list, ?>>>>((("DefaultShapes", {{"audio_signal", [1, 1]}}), ("RangeDims", {{"audio_signal", [[1, 1], [1, 240000]]}})))] { - tensor var_9 = const()[name = tensor("op_9"), val = tensor(1)]; - tensor var_10 = const()[name = tensor("op_10"), val = tensor(160)]; - tensor var_34 = const()[name = tensor("op_34"), val = tensor(512)]; - tensor var_35 = add(x = audio_length, y = var_34)[name = tensor("op_35")]; - tensor var_36 = const()[name = tensor("op_36"), val = tensor(512)]; - tensor var_37 = sub(x = var_35, y = var_36)[name = tensor("op_37")]; - tensor floor_div_0 = floor_div(x = var_37, y = var_10)[name = tensor("floor_div_0")]; - tensor var_38_to_fp16_dtype_0 = const()[name = tensor("op_38_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor var_39_promoted_to_fp16 = const()[name = tensor("op_39_promoted_to_fp16"), val = tensor(0x1p+0)]; - tensor floor_div_0_to_fp16 = cast(dtype = var_38_to_fp16_dtype_0, x = floor_div_0)[name = tensor("cast_9")]; - tensor seq_len_1_cast_fp16 = add(x = floor_div_0_to_fp16, y = var_39_promoted_to_fp16)[name = tensor("seq_len_1_cast_fp16")]; - tensor seq_len_dtype_0 = const()[name = tensor("seq_len_dtype_0"), val = tensor("int32")]; - tensor var_43_begin_0 = const()[name = tensor("op_43_begin_0"), val = tensor([0, 0])]; - tensor var_43_end_0 = const()[name = tensor("op_43_end_0"), val = tensor([1, 1])]; - tensor var_43_end_mask_0 = const()[name = tensor("op_43_end_mask_0"), val = tensor([true, false])]; - tensor var_43_squeeze_mask_0 = const()[name = tensor("op_43_squeeze_mask_0"), val = tensor([false, true])]; - tensor audio_signal_to_fp16_dtype_0 = const()[name = tensor("audio_signal_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor audio_signal_to_fp16 = cast(dtype = audio_signal_to_fp16_dtype_0, x = audio_signal)[name = tensor("cast_8")]; - tensor var_43_cast_fp16 = slice_by_index(begin = var_43_begin_0, end = var_43_end_0, end_mask = var_43_end_mask_0, squeeze_mask = var_43_squeeze_mask_0, x = audio_signal_to_fp16)[name = tensor("op_43_cast_fp16")]; - tensor var_44_axes_0 = const()[name = tensor("op_44_axes_0"), val = tensor([1])]; - tensor var_44_cast_fp16 = expand_dims(axes = var_44_axes_0, x = var_43_cast_fp16)[name = tensor("op_44_cast_fp16")]; - tensor var_46_begin_0 = const()[name = tensor("op_46_begin_0"), val = tensor([0, 1])]; - tensor var_46_end_0 = const()[name = tensor("op_46_end_0"), val = tensor([1, 0])]; - tensor var_46_end_mask_0 = const()[name = tensor("op_46_end_mask_0"), val = tensor([true, true])]; - tensor var_46_cast_fp16 = slice_by_index(begin = var_46_begin_0, end = var_46_end_0, end_mask = var_46_end_mask_0, x = audio_signal_to_fp16)[name = tensor("op_46_cast_fp16")]; - tensor var_48_begin_0 = const()[name = tensor("op_48_begin_0"), val = tensor([0, 0])]; - tensor var_48_end_0 = const()[name = tensor("op_48_end_0"), val = tensor([1, -1])]; - tensor var_48_end_mask_0 = const()[name = tensor("op_48_end_mask_0"), val = tensor([true, false])]; - tensor var_48_cast_fp16 = slice_by_index(begin = var_48_begin_0, end = var_48_end_0, end_mask = var_48_end_mask_0, x = audio_signal_to_fp16)[name = tensor("op_48_cast_fp16")]; - tensor var_49_to_fp16 = const()[name = tensor("op_49_to_fp16"), val = tensor(0x1.f0cp-1)]; - tensor var_50_cast_fp16 = mul(x = var_48_cast_fp16, y = var_49_to_fp16)[name = tensor("op_50_cast_fp16")]; - tensor var_51_cast_fp16 = sub(x = var_46_cast_fp16, y = var_50_cast_fp16)[name = tensor("op_51_cast_fp16")]; - tensor input_1_interleave_0 = const()[name = tensor("input_1_interleave_0"), val = tensor(false)]; - tensor input_1_cast_fp16 = concat(axis = var_9, interleave = input_1_interleave_0, values = (var_44_cast_fp16, var_51_cast_fp16))[name = tensor("input_1_cast_fp16")]; - tensor concat_0x = const()[name = tensor("concat_0x"), val = tensor([1, 1, -1])]; - tensor input_3_cast_fp16 = reshape(shape = concat_0x, x = input_1_cast_fp16)[name = tensor("input_3_cast_fp16")]; - tensor input_5_pad_0 = const()[name = tensor("input_5_pad_0"), val = tensor([0, 0, 0, 0, 256, 256])]; - tensor input_5_mode_0 = const()[name = tensor("input_5_mode_0"), val = tensor("reflect")]; - tensor const_1_to_fp16 = const()[name = tensor("const_1_to_fp16"), val = tensor(0x0p+0)]; - tensor input_5_cast_fp16 = pad(constant_val = const_1_to_fp16, mode = input_5_mode_0, pad = input_5_pad_0, x = input_3_cast_fp16)[name = tensor("input_5_cast_fp16")]; - tensor concat_1x = const()[name = tensor("concat_1x"), val = tensor([1, -1])]; - tensor input_cast_fp16 = reshape(shape = concat_1x, x = input_5_cast_fp16)[name = tensor("input_cast_fp16")]; - tensor expand_dims_3 = const()[name = tensor("expand_dims_3"), val = tensor([160])]; - tensor expand_dims_4_axes_0 = const()[name = tensor("expand_dims_4_axes_0"), val = tensor([1])]; - tensor expand_dims_4_cast_fp16 = expand_dims(axes = expand_dims_4_axes_0, x = input_cast_fp16)[name = tensor("expand_dims_4_cast_fp16")]; - tensor conv_0_pad_type_0 = const()[name = tensor("conv_0_pad_type_0"), val = tensor("valid")]; - tensor conv_0_pad_0 = const()[name = tensor("conv_0_pad_0"), val = tensor([0, 0])]; - tensor conv_0_dilations_0 = const()[name = tensor("conv_0_dilations_0"), val = tensor([1])]; - tensor conv_0_groups_0 = const()[name = tensor("conv_0_groups_0"), val = tensor(1)]; - tensor expand_dims_1_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("expand_dims_1_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132096))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(131712)))]; - tensor conv_0_cast_fp16 = conv(dilations = conv_0_dilations_0, groups = conv_0_groups_0, pad = conv_0_pad_0, pad_type = conv_0_pad_type_0, strides = expand_dims_3, weight = expand_dims_1_to_fp16_quantized, x = expand_dims_4_cast_fp16)[name = tensor("conv_0_cast_fp16")]; - tensor conv_1_pad_type_0 = const()[name = tensor("conv_1_pad_type_0"), val = tensor("valid")]; - tensor conv_1_pad_0 = const()[name = tensor("conv_1_pad_0"), val = tensor([0, 0])]; - tensor conv_1_dilations_0 = const()[name = tensor("conv_1_dilations_0"), val = tensor([1])]; - tensor conv_1_groups_0 = const()[name = tensor("conv_1_groups_0"), val = tensor(1)]; - tensor expand_dims_2_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("expand_dims_2_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132736))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(264768))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(264384)))]; - tensor conv_1_cast_fp16 = conv(dilations = conv_1_dilations_0, groups = conv_1_groups_0, pad = conv_1_pad_0, pad_type = conv_1_pad_type_0, strides = expand_dims_3, weight = expand_dims_2_to_fp16_quantized, x = expand_dims_4_cast_fp16)[name = tensor("conv_1_cast_fp16")]; - tensor stack_0_axis_0 = const()[name = tensor("stack_0_axis_0"), val = tensor(-1)]; - tensor stack_0_cast_fp16 = stack(axis = stack_0_axis_0, values = (conv_0_cast_fp16, conv_1_cast_fp16))[name = tensor("stack_0_cast_fp16")]; - tensor var_17_promoted_to_fp16 = const()[name = tensor("op_17_promoted_to_fp16"), val = tensor(0x1p+1)]; - tensor var_67_cast_fp16 = pow(x = stack_0_cast_fp16, y = var_17_promoted_to_fp16)[name = tensor("op_67_cast_fp16")]; - tensor var_69_axes_0 = const()[name = tensor("op_69_axes_0"), val = tensor([-1])]; - tensor var_69_keep_dims_0 = const()[name = tensor("op_69_keep_dims_0"), val = tensor(false)]; - tensor var_69_cast_fp16 = reduce_sum(axes = var_69_axes_0, keep_dims = var_69_keep_dims_0, x = var_67_cast_fp16)[name = tensor("op_69_cast_fp16")]; - tensor x_11_transpose_x_0 = const()[name = tensor("x_11_transpose_x_0"), val = tensor(false)]; - tensor x_11_transpose_y_0 = const()[name = tensor("x_11_transpose_y_0"), val = tensor(false)]; - tensor const_2_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(1), name = tensor("const_2_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(265408))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(298560))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(298368)))]; - tensor x_11_cast_fp16 = matmul(transpose_x = x_11_transpose_x_0, transpose_y = x_11_transpose_y_0, x = const_2_to_fp16_quantized, y = var_69_cast_fp16)[name = tensor("x_11_cast_fp16")]; - tensor var_76_to_fp16 = const()[name = tensor("op_76_to_fp16"), val = tensor(0x1p-24)]; - tensor var_77_cast_fp16 = add(x = x_11_cast_fp16, y = var_76_to_fp16)[name = tensor("op_77_cast_fp16")]; - tensor x_13_epsilon_0 = const()[name = tensor("x_13_epsilon_0"), val = tensor(0x1p-149)]; - tensor x_13_cast_fp16 = log(epsilon = x_13_epsilon_0, x = var_77_cast_fp16)[name = tensor("x_13_cast_fp16")]; - tensor var_79_shape_cast_fp16 = shape(x = x_13_cast_fp16)[name = tensor("op_79_shape_cast_fp16")]; - tensor gather_4 = const()[name = tensor("gather_4"), val = tensor(1)]; - tensor gather_5_axis_0 = const()[name = tensor("gather_5_axis_0"), val = tensor(0)]; - tensor gather_5_batch_dims_0 = const()[name = tensor("gather_5_batch_dims_0"), val = tensor(0)]; - tensor gather_5_validate_indices_0 = const()[name = tensor("gather_5_validate_indices_0"), val = tensor(false)]; - tensor var_79_shape_cast_fp16_to_uint16_dtype_0 = const()[name = tensor("op_79_shape_cast_fp16_to_uint16_dtype_0"), val = tensor("uint16")]; - tensor gather_5_indices_0_to_uint16 = const()[name = tensor("gather_5_indices_0_to_uint16"), val = tensor(2)]; - tensor var_79_shape_cast_fp16_to_uint16 = cast(dtype = var_79_shape_cast_fp16_to_uint16_dtype_0, x = var_79_shape_cast_fp16)[name = tensor("cast_7")]; - tensor gather_5_cast_uint16 = gather(axis = gather_5_axis_0, batch_dims = gather_5_batch_dims_0, indices = gather_5_indices_0_to_uint16, validate_indices = gather_5_validate_indices_0, x = var_79_shape_cast_fp16_to_uint16)[name = tensor("gather_5_cast_uint16")]; - tensor gather_5_cast_uint16_to_int32_dtype_0 = const()[name = tensor("gather_5_cast_uint16_to_int32_dtype_0"), val = tensor("int32")]; - tensor const_3 = const()[name = tensor("const_3"), val = tensor(0)]; - tensor const_4 = const()[name = tensor("const_4"), val = tensor(1)]; - tensor gather_5_cast_uint16_to_int32 = cast(dtype = gather_5_cast_uint16_to_int32_dtype_0, x = gather_5_cast_uint16)[name = tensor("cast_6")]; - tensor var_81 = range_1d(end = gather_5_cast_uint16_to_int32, start = const_3, step = const_4)[name = tensor("op_81")]; - tensor var_82_axes_0 = const()[name = tensor("op_82_axes_0"), val = tensor([0])]; - tensor var_82 = expand_dims(axes = var_82_axes_0, x = var_81)[name = tensor("op_82")]; - tensor concat_2_axis_0 = const()[name = tensor("concat_2_axis_0"), val = tensor(0)]; - tensor concat_2_interleave_0 = const()[name = tensor("concat_2_interleave_0"), val = tensor(false)]; - tensor concat_2 = concat(axis = concat_2_axis_0, interleave = concat_2_interleave_0, values = (gather_4, gather_5_cast_uint16_to_int32))[name = tensor("concat_2")]; - tensor shape_0 = shape(x = var_82)[name = tensor("shape_0")]; - tensor real_div_0 = real_div(x = concat_2, y = shape_0)[name = tensor("real_div_0")]; - tensor time_steps = tile(reps = real_div_0, x = var_82)[name = tensor("time_steps")]; - tensor var_85_axes_0 = const()[name = tensor("op_85_axes_0"), val = tensor([1])]; - tensor mel_length = cast(dtype = seq_len_dtype_0, x = seq_len_1_cast_fp16)[name = tensor("cast_5")]; - tensor var_85 = expand_dims(axes = var_85_axes_0, x = mel_length)[name = tensor("op_85")]; - tensor valid_mask = less(x = time_steps, y = var_85)[name = tensor("valid_mask")]; - tensor var_87_axes_0 = const()[name = tensor("op_87_axes_0"), val = tensor([1])]; - tensor var_87 = expand_dims(axes = var_87_axes_0, x = valid_mask)[name = tensor("op_87")]; - tensor var_24_to_fp16 = const()[name = tensor("op_24_to_fp16"), val = tensor(0x0p+0)]; - tensor var_88_cast_fp16 = select(a = x_13_cast_fp16, b = var_24_to_fp16, cond = var_87)[name = tensor("op_88_cast_fp16")]; - tensor x_mean_numerator_axes_0 = const()[name = tensor("x_mean_numerator_axes_0"), val = tensor([2])]; - tensor x_mean_numerator_keep_dims_0 = const()[name = tensor("x_mean_numerator_keep_dims_0"), val = tensor(false)]; - tensor x_mean_numerator_cast_fp16 = reduce_sum(axes = x_mean_numerator_axes_0, keep_dims = x_mean_numerator_keep_dims_0, x = var_88_cast_fp16)[name = tensor("x_mean_numerator_cast_fp16")]; - tensor x_mean_denominator_axes_0 = const()[name = tensor("x_mean_denominator_axes_0"), val = tensor([1])]; - tensor x_mean_denominator_keep_dims_0 = const()[name = tensor("x_mean_denominator_keep_dims_0"), val = tensor(false)]; - tensor cast_3_to_fp16_dtype_0 = const()[name = tensor("cast_3_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor valid_mask_to_fp16 = cast(dtype = cast_3_to_fp16_dtype_0, x = valid_mask)[name = tensor("cast_4")]; - tensor x_mean_denominator_cast_fp16 = reduce_sum(axes = x_mean_denominator_axes_0, keep_dims = x_mean_denominator_keep_dims_0, x = valid_mask_to_fp16)[name = tensor("x_mean_denominator_cast_fp16")]; - tensor var_93_axes_0 = const()[name = tensor("op_93_axes_0"), val = tensor([1])]; - tensor var_93_cast_fp16 = expand_dims(axes = var_93_axes_0, x = x_mean_denominator_cast_fp16)[name = tensor("op_93_cast_fp16")]; - tensor x_mean_cast_fp16 = real_div(x = x_mean_numerator_cast_fp16, y = var_93_cast_fp16)[name = tensor("x_mean_cast_fp16")]; - tensor var_96_axes_0 = const()[name = tensor("op_96_axes_0"), val = tensor([2])]; - tensor var_96_cast_fp16 = expand_dims(axes = var_96_axes_0, x = x_mean_cast_fp16)[name = tensor("op_96_cast_fp16")]; - tensor var_97_cast_fp16 = sub(x = x_13_cast_fp16, y = var_96_cast_fp16)[name = tensor("op_97_cast_fp16")]; - tensor var_98_cast_fp16 = select(a = var_97_cast_fp16, b = var_24_to_fp16, cond = var_87)[name = tensor("op_98_cast_fp16")]; - tensor var_17_promoted_1_to_fp16 = const()[name = tensor("op_17_promoted_1_to_fp16"), val = tensor(0x1p+1)]; - tensor var_99_cast_fp16 = pow(x = var_98_cast_fp16, y = var_17_promoted_1_to_fp16)[name = tensor("op_99_cast_fp16")]; - tensor var_101_axes_0 = const()[name = tensor("op_101_axes_0"), val = tensor([2])]; - tensor var_101_keep_dims_0 = const()[name = tensor("op_101_keep_dims_0"), val = tensor(false)]; - tensor var_101_cast_fp16 = reduce_sum(axes = var_101_axes_0, keep_dims = var_101_keep_dims_0, x = var_99_cast_fp16)[name = tensor("op_101_cast_fp16")]; - tensor var_103_to_fp16 = const()[name = tensor("op_103_to_fp16"), val = tensor(0x1p+0)]; - tensor var_104_cast_fp16 = sub(x = var_93_cast_fp16, y = var_103_to_fp16)[name = tensor("op_104_cast_fp16")]; - tensor var_105_cast_fp16 = real_div(x = var_101_cast_fp16, y = var_104_cast_fp16)[name = tensor("op_105_cast_fp16")]; - tensor x_std_1_cast_fp16 = sqrt(x = var_105_cast_fp16)[name = tensor("x_std_1_cast_fp16")]; - tensor var_25_to_fp16 = const()[name = tensor("op_25_to_fp16"), val = tensor(0x1.5p-17)]; - tensor x_std_cast_fp16 = add(x = x_std_1_cast_fp16, y = var_25_to_fp16)[name = tensor("x_std_cast_fp16")]; - tensor var_110_axes_0 = const()[name = tensor("op_110_axes_0"), val = tensor([2])]; - tensor var_110_cast_fp16 = expand_dims(axes = var_110_axes_0, x = x_std_cast_fp16)[name = tensor("op_110_cast_fp16")]; - tensor x_cast_fp16 = real_div(x = var_97_cast_fp16, y = var_110_cast_fp16)[name = tensor("x_cast_fp16")]; - tensor var_112_shape_cast_fp16 = shape(x = x_cast_fp16)[name = tensor("op_112_shape_cast_fp16")]; - tensor gather_6_batch_dims_0 = const()[name = tensor("gather_6_batch_dims_0"), val = tensor(0)]; - tensor gather_6_validate_indices_0 = const()[name = tensor("gather_6_validate_indices_0"), val = tensor(false)]; - tensor var_112_shape_cast_fp16_to_uint16_dtype_0 = const()[name = tensor("op_112_shape_cast_fp16_to_uint16_dtype_0"), val = tensor("uint16")]; - tensor gather_6_cast_uint16_axis_0 = const()[name = tensor("gather_6_cast_uint16_axis_0"), val = tensor(0)]; - tensor select_0_to_uint16 = const()[name = tensor("select_0_to_uint16"), val = tensor(2)]; - tensor var_112_shape_cast_fp16_to_uint16 = cast(dtype = var_112_shape_cast_fp16_to_uint16_dtype_0, x = var_112_shape_cast_fp16)[name = tensor("cast_3")]; - tensor gather_6_cast_uint16_cast_uint16 = gather(axis = gather_6_cast_uint16_axis_0, batch_dims = gather_6_batch_dims_0, indices = select_0_to_uint16, validate_indices = gather_6_validate_indices_0, x = var_112_shape_cast_fp16_to_uint16)[name = tensor("gather_6_cast_uint16_cast_uint16")]; - tensor gather_6_cast_uint16_to_int32_dtype_0 = const()[name = tensor("gather_6_cast_uint16_to_int32_dtype_0"), val = tensor("int32")]; - tensor const_5 = const()[name = tensor("const_5"), val = tensor(0)]; - tensor const_6 = const()[name = tensor("const_6"), val = tensor(1)]; - tensor gather_6_cast_uint16_to_int32 = cast(dtype = gather_6_cast_uint16_to_int32_dtype_0, x = gather_6_cast_uint16_cast_uint16)[name = tensor("cast_2")]; - tensor mask_1 = range_1d(end = gather_6_cast_uint16_to_int32, start = const_5, step = const_6)[name = tensor("mask_1")]; - tensor gather_7_axis_0 = const()[name = tensor("gather_7_axis_0"), val = tensor(0)]; - tensor gather_7_batch_dims_0 = const()[name = tensor("gather_7_batch_dims_0"), val = tensor(0)]; - tensor gather_7_validate_indices_0 = const()[name = tensor("gather_7_validate_indices_0"), val = tensor(false)]; - tensor gather_7_indices_0_to_uint16 = const()[name = tensor("gather_7_indices_0_to_uint16"), val = tensor(0)]; - tensor gather_7_cast_uint16 = gather(axis = gather_7_axis_0, batch_dims = gather_7_batch_dims_0, indices = gather_7_indices_0_to_uint16, validate_indices = gather_7_validate_indices_0, x = var_112_shape_cast_fp16_to_uint16)[name = tensor("gather_7_cast_uint16")]; - tensor gather_7_cast_uint16_to_int32_dtype_0 = const()[name = tensor("gather_7_cast_uint16_to_int32_dtype_0"), val = tensor("int32")]; - tensor concat_3_axis_0 = const()[name = tensor("concat_3_axis_0"), val = tensor(0)]; - tensor concat_3_interleave_0 = const()[name = tensor("concat_3_interleave_0"), val = tensor(false)]; - tensor gather_7_cast_uint16_to_int32 = cast(dtype = gather_7_cast_uint16_to_int32_dtype_0, x = gather_7_cast_uint16)[name = tensor("cast_1")]; - tensor concat_3 = concat(axis = concat_3_axis_0, interleave = concat_3_interleave_0, values = (gather_7_cast_uint16_to_int32, var_9))[name = tensor("concat_3")]; - tensor expand_dims_0_axes_0 = const()[name = tensor("expand_dims_0_axes_0"), val = tensor([0])]; - tensor expand_dims_0 = expand_dims(axes = expand_dims_0_axes_0, x = mask_1)[name = tensor("expand_dims_0")]; - tensor var_116 = tile(reps = concat_3, x = expand_dims_0)[name = tensor("op_116")]; - tensor mask = greater_equal(x = var_116, y = var_85)[name = tensor("mask")]; - tensor var_119_axes_0 = const()[name = tensor("op_119_axes_0"), val = tensor([1])]; - tensor var_119 = expand_dims(axes = var_119_axes_0, x = mask)[name = tensor("op_119")]; - tensor processed_signal_cast_fp16 = select(a = var_24_to_fp16, b = x_cast_fp16, cond = var_119)[name = tensor("processed_signal_cast_fp16")]; - tensor processed_signal_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("processed_signal_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; - tensor mel = cast(dtype = processed_signal_cast_fp16_to_fp32_dtype_0, x = processed_signal_cast_fp16)[name = tensor("cast_0")]; - } -> (mel, mel_length); -} \ No newline at end of file diff --git a/models/parakeet-tdt-0.6b-v2-coreml/Preprocessor.mlmodelc/weights/weight.bin b/models/parakeet-tdt-0.6b-v2-coreml/Preprocessor.mlmodelc/weights/weight.bin deleted file mode 100644 index a9af1ac1e407ef336eebc2fd4da9de57e6d172ad..0000000000000000000000000000000000000000 --- a/models/parakeet-tdt-0.6b-v2-coreml/Preprocessor.mlmodelc/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:a5f7df6c7f47147ae9486fe18cc7792f9a44d093ec3c6a11e91ef2dc363c48dc -size 298880 diff --git a/models/parakeet-tdt-0.6b-v2-coreml/config.json b/models/parakeet-tdt-0.6b-v2-coreml/config.json deleted file mode 100644 index 0967ef424bce6791893e9a57bb952f80fd536e93..0000000000000000000000000000000000000000 --- a/models/parakeet-tdt-0.6b-v2-coreml/config.json +++ /dev/null @@ -1 +0,0 @@ -{} diff --git a/models/parakeet-tdt-0.6b-v2-coreml/parakeet_vocab.json b/models/parakeet-tdt-0.6b-v2-coreml/parakeet_vocab.json deleted file mode 100644 index a111d3bd29d44781c4588447c3eac1ea549cb1f1..0000000000000000000000000000000000000000 --- a/models/parakeet-tdt-0.6b-v2-coreml/parakeet_vocab.json +++ /dev/null @@ -1,1033 +0,0 @@ -{ - "479" : "▁happ", - "493" : "ial", - "386" : "▁diff", - "398" : "▁little", - "168" : "▁ch", - "872" : "2", - "227" : "ie", - "251" : "▁R", - "488" : "▁day", - "38" : "ed", - "610" : "▁might", - "127" : "▁can", - "70" : "ct", - "903" : "Ă ", - "329" : "▁been", - "694" : "ren", - "1005" : "Ćœ", - "661" : "▁reg", - "597" : "▁around", - "442" : "ys", - "764" : "▁number", - "705" : "▁trans", - "275" : "▁had", - "271" : "ak", - "1024" : "▁warming", - "564" : "▁play", - "32" : "▁y", - "463" : "▁under", - "684" : "ute", - "529" : "▁quest", - "719" : "ock", - "235" : "her", - "861" : "q", - "987" : "Č", - "593" : "urn", - "555" : "ath", - "216" : "▁think", - "211" : "el", - "500" : "▁bit", - "353" : "ry", - "490" : "▁Yeah", - "750" : "▁sim", - "283" : "▁these", - "320" : "▁sp", - "945" : "ÎČ", - "179" : "ca", - "934" : "т", - "258" : "ear", - "870" : "R", - "974" : "Ί", - "157" : "▁H", - "280" : "ine", - "381" : "▁well", - "21" : "▁m", - "668" : "▁la", - "429" : "▁bo", - "812" : "▁ty", - "537" : "ual", - "743" : "ss", - "119" : "▁kn", - "863" : "P", - "855" : "0", - "333" : "ple", - "850" : "x", - "959" : "я", - "467" : "▁bu", - "262" : "▁would", - "835" : "g", - "254" : "▁who", - "233" : "▁here", - "993" : "ĆŒ", - "889" : "Z", - "231" : "ack", - "788" : "ution", - "466" : "ning", - "332" : "▁did", - "388" : "cc", - "260" : "▁E", - "839" : ",", - "582" : "▁three", - "1000" : "Ă”", - "73" : "ver", - "486" : "ign", - "541" : "▁ser", - "800" : "▁exper", - "192" : "um", - "199" : "▁from", - "223" : "▁because", - "68" : "ent", - "608" : "▁tra", - "383" : "▁pre", - "774" : "▁place", - "194" : "▁your", - "148" : "ra", - "955" : "Ü", - "158" : "▁fr", - "670" : "▁sub", - "906" : "π", - "829" : "l", - "202" : "pe", - "948" : "ÎŽ", - "480" : "ater", - "342" : "ree", - "958" : "ƛ", - "967" : "Á", - "171" : "ate", - "91" : "am", - "114" : "ation", - "390" : "int", - "1013" : "À", - "794" : "▁read", - "1011" : "ф", - "355" : "ble", - "921" : "Ć«", - "234" : "▁pe", - "573" : "▁id", - "785" : "gan", - "335" : "▁other", - "963" : "Ö", - "570" : "get", - "2" : "▁th", - "273" : "ide", - "574" : "▁Oh", - "807" : "▁open", - "155" : "▁The", - "506" : "vers", - "729" : "▁sort", - "318" : "▁got", - "550" : "ank", - "896" : "Ăł", - "563" : "ict", - "1010" : "Đ·", - "520" : "▁gonna", - "1020" : "Κ", - "548" : "ved", - "848" : "S", - "578" : "▁rel", - "828" : "r", - "361" : "▁cont", - "379" : "ish", - "259" : "▁tim", - "441" : "▁imp", - "44" : "as", - "620" : "ways", - "34" : "▁I", - "87" : "st", - "13" : "nd", - "350" : "▁every", - "866" : "D", - "545" : "▁V", - "182" : "if", - "615" : "cial", - "69" : "ke", - "276" : "so", - "732" : "ics", - "749" : "als", - "605" : "aking", - "857" : "-", - "446" : "▁down", - "41" : "ar", - "203" : "un", - "816" : "▁met", - "543" : "▁ind", - "926" : "ĂŠ", - "64" : "▁on", - "246" : "▁them", - "172" : "qu", - "229" : "▁tr", - "775" : "▁gener", - "592" : "ower", - "617" : "▁give", - "399" : "ical", - "319" : "ag", - "669" : "▁last", - "400" : "▁gr", - "389" : "ittle", - "533" : "▁fu", - "460" : "uct", - "825" : "i", - "992" : "Ç", - "432" : "▁much", - "638" : "ction", - "700" : "▁love", - "988" : "έ", - "658" : "▁inv", - "549" : "▁still", - "322" : "act", - "143" : "il", - "966" : "ĐČ", - "245" : "▁It", - "445" : "▁yeah", - "524" : "self", - "742" : "ata", - "380" : "kay", - "176" : "▁su", - "624" : "▁gen", - "766" : "▁saying", - "42" : "▁that", - "136" : "▁ab", - "224" : "▁by", - "811" : "▁getting", - "76" : "id", - "430" : "be", - "124" : "▁know", - "88" : "ch", - "420" : "▁she", - "822" : "o", - "991" : "Å", - "623" : "ily", - "230" : "▁then", - "65" : "▁T", - "86" : "▁st", - "873" : "J", - "464" : "▁br", - "165" : "pp", - "325" : "iz", - "237" : "▁F", - "939" : "č", - "250" : "▁say", - "423" : "ord", - "516" : "▁fl", - "1030" : "▁urge", - "977" : "Îč", - "358" : "▁kind", - "471" : "co", - "7" : "▁w", - "252" : "▁people", - "12" : "er", - "932" : "ψ", - "632" : "▁cour", - "440" : "▁comm", - "664" : "▁cr", - "637" : "▁num", - "588" : "▁To", - "286" : "▁sa", - "292" : "pt", - "772" : "▁better", - "396" : "▁him", - "1007" : "Γ", - "26" : "▁d", - "450" : "ens", - "780" : "cept", - "936" : "ω", - "123" : "▁And", - "349" : "▁need", - "536" : "we", - "882" : "8", - "546" : "he", - "854" : "?", - "875" : ":", - "712" : "▁exam", - "437" : "ary", - "394" : "ip", - "344" : "own", - "709" : "▁fin", - "457" : "▁K", - "996" : "χ", - "990" : "ы", - "89" : "▁li", - "878" : "3", - "503" : "fe", - "61" : "et", - "782" : "▁understand", - "126" : "op", - "135" : "▁at", - "831" : "u", - "456" : "ody", - "594" : "▁okay", - "695" : "erest", - "324" : "▁also", - "51" : "▁it", - "542" : "▁rem", - "174" : "▁ex", - "1006" : "ț", - "128" : "▁or", - "357" : "ue", - "641" : "▁um", - "802" : "▁ele", - "213" : "▁some", - "606" : "▁pos", - "929" : "Μ", - "885" : "!", - "784" : "▁thought", - "716" : "▁stud", - "232" : "▁pl", - "522" : "ces", - "184" : "▁if", - "37" : "is", - "867" : "N", - "653" : "uc", - "413" : "ult", - "167" : "ess", - "22" : "en", - "439" : "ving", - "104" : "▁r", - "755" : "oc", - "63" : "▁re", - "622" : "ward", - "1003" : "Ɓ", - "468" : "▁use", - "879" : "K", - "838" : "p", - "11" : "ou", - "54" : "le", - "118" : "▁not", - "558" : "ft", - "858" : "M", - "579" : "▁before", - "652" : "les", - "98" : "▁do", - "1026" : "▁issue", - "360" : "▁back", - "790" : "ason", - "731" : "▁today", - "706" : "▁count", - "650" : "▁didn", - "348" : "▁where", - "371" : "ep", - "898" : "ĂŒ", - "384" : "▁two", - "121" : "▁B", - "720" : "▁used", - "120" : "ight", - "1022" : "Ο", - "345" : "▁tw", - "307" : "▁work", - "655" : "ating", - "745" : "ween", - "596" : "▁bel", - "852" : "B", - "986" : "ĂŹ", - "197" : "▁get", - "79" : "▁he", - "504" : "▁doing", - "644" : "▁own", - "791" : "▁problem", - "931" : "Îł", - "33" : "▁l", - "187" : "ab", - "771" : "▁between", - "783" : "▁fun", - "823" : "a", - "723" : "▁No", - "303" : "▁mo", - "482" : "ition", - "164" : "▁M", - "346" : "▁part", - "103" : "ad", - "902" : "ç", - "781" : "ull", - "411" : "▁actually", - "598" : "ful", - "1002" : "ħ", - "351" : "pl", - "343" : "▁into", - "908" : "Ăș", - "337" : "ite", - "894" : "ĂĄ", - "912" : "Ä«", - "864" : "z", - "925" : "τ", - "153" : "▁con", - "144" : "▁but", - "433" : "▁per", - "711" : "▁pol", - "631" : "▁sm", - "431" : "ount", - "999" : "Í", - "946" : "Ăž", - "141" : "use", - "77" : "▁for", - "733" : "▁vide", - "730" : "▁For", - "418" : "ations", - "693" : "▁always", - "642" : "ood", - "1015" : "Ā", - "154" : "▁all", - "659" : "ably", - "1027" : "▁stay", - "777" : "▁ins", - "740" : "▁keep", - "96" : "▁so", - "806" : "▁partic", - "313" : "▁im", - "507" : "av", - "880" : "4", - "792" : "▁doesn", - "426" : "▁am", - "1001" : "ě", - "526" : "▁If", - "465" : "▁take", - "436" : "vel", - "116" : "ere", - "576" : "ever", - "568" : "oth", - "166" : "▁com", - "151" : "ul", - "798" : "ah", - "779" : "cond", - "556" : "▁end", - "161" : "ea", - "39" : "▁g", - "765" : "ention", - "100" : "th", - "461" : "▁only", - "590" : "▁hel", - "505" : "▁St", - "957" : "Ƅ", - "876" : "5", - "580" : "▁feel", - "321" : "ans", - "972" : "ь", - "718" : "▁car", - "851" : "W", - "960" : "đ", - "469" : "▁Ch", - "291" : "one", - "83" : "ly", - "295" : "▁has", - "84" : "▁go", - "981" : "Ƒ", - "923" : "λ", - "52" : "▁be", - "821" : "t", - "323" : "▁te", - "50" : "al", - "760" : "ense", - "149" : "ore", - "306" : "▁le", - "403" : "▁thr", - "628" : "ob", - "299" : "▁look", - "406" : "▁This", - "0" : "", - "75" : "all", - "475" : "▁call", - "341" : "reat", - "519" : "ents", - "66" : "▁A", - "186" : "nt", - "928" : "ĐŸ", - "244" : "ople", - "393" : "ence", - "834" : "m", - "833" : "y", - "414" : "able", - "297" : "▁very", - "364" : "▁pr", - "865" : "L", - "35" : "it", - "339" : "omet", - "27" : "es", - "150" : "▁there", - "715" : "ell", - "677" : "▁import", - "681" : "▁ear", - "820" : "e", - "139" : "▁So", - "449" : "na", - "302" : "▁time", - "532" : "▁What", - "133" : "ck", - "970" : "Îż", - "773" : "cus", - "228" : "▁us", - "552" : "▁wr", - "680" : "▁made", - "871" : "E", - "874" : "U", - "895" : "ÂŁ", - "846" : "T", - "55" : "ion", - "492" : "ile", - "787" : "cy", - "105" : "ir", - "662" : "lic", - "629" : "▁tell", - "367" : "▁good", - "514" : "form", - "656" : "olog", - "334" : "ually", - "294" : "ong", - "485" : "ade", - "682" : "▁ac", - "106" : "▁was", - "1028" : "▁together", - "910" : "ĂŁ", - "142" : "ter", - "922" : "Δ", - "175" : "very", - "314" : "▁ag", - "327" : "▁That", - "826" : "s", - "183" : "ive", - "979" : "Đł", - "751" : "vern", - "997" : "э", - "278" : "eah", - "517" : "fter", - "311" : "per", - "535" : "▁show", - "918" : "^", - "954" : "ĆĄ", - "722" : "stand", - "6" : "re", - "93" : "ce", - "435" : "▁differe", - "746" : "▁stuff", - "915" : "ρ", - "602" : "▁supp", - "209" : "▁L", - "767" : "▁commun", - "769" : "akes", - "375" : "▁lot", - "859" : "H", - "397" : "▁make", - "340" : "ber", - "886" : "%", - "736" : "▁Al", - "600" : "ise", - "938" : "ć", - "478" : "ting", - "962" : "ĐŒ", - "138" : "ol", - "125" : "ome", - "309" : "are", - "673" : "▁inst", - "97" : "▁have", - "562" : "ject", - "678" : "ific", - "257" : "ect", - "17" : "on", - "953" : "с", - "933" : "ē", - "844" : "'", - "949" : "η", - "190" : "▁v", - "980" : "Đș", - "452" : "▁fo", - "247" : "ame", - "612" : "▁help", - "501" : "▁spe", - "604" : "ange", - "654" : "ib", - "842" : "k", - "815" : "ave", - "687" : "▁form", - "222" : "res", - "421" : "sel", - "477" : "other", - "308" : "▁their", - "212" : "▁N", - "737" : "▁important", - "90" : "▁u", - "630" : "▁Now", - "425" : "ia", - "14" : "▁i", - "315" : "▁J", - "331" : "▁fe", - "304" : "▁ar", - "29" : "ll", - "499" : "▁sc", - "919" : "€", - "301" : "itt", - "201" : "ri", - "137" : "ould", - "289" : "▁man", - "81" : "▁this", - "1008" : "П", - "312" : "ions", - "725" : "ks", - "837" : "f", - "458" : "▁through", - "714" : "▁maybe", - "487" : "thing", - "424" : "▁may", - "31" : "▁and", - "78" : "ro", - "961" : "Đ»", - "956" : "Ă„", - "198" : "cause", - "95" : "im", - "899" : "ñ", - "111" : "ally", - "523" : "▁There", - "538" : "ons", - "797" : "▁el", - "726" : "▁interest", - "53" : "▁wh", - "296" : "▁any", - "489" : "fore", - "911" : "φ", - "248" : "▁We", - "639" : "▁add", - "108" : "▁W", - "521" : "▁point", - "416" : "▁dis", - "739" : "▁run", - "747" : "ract", - "3" : "▁a", - "530" : "▁most", - "734" : "▁bec", - "338" : "age", - "544" : "▁pers", - "113" : "▁se", - "691" : "▁able", - "847" : "A", - "651" : "stem", - "115" : "od", - "527" : "▁same", - "575" : "ves", - "696" : "▁As", - "277" : "▁qu", - "728" : "ited", - "640" : "▁set", - "502" : "ub", - "497" : "▁try", - "881" : "V", - "219" : "▁G", - "561" : "ph", - "759" : "▁All", - "177" : "ain", - "515" : "ors", - "71" : "▁S", - "509" : "ian", - "803" : "▁cou", - "25" : "an", - "583" : "iss", - "417" : "▁first", - "840" : "b", - "768" : "▁An", - "419" : "▁something", - "569" : "▁acc", - "607" : "atch", - "534" : "ug", - "195" : "▁my", - "832" : "c", - "634" : "cess", - "809" : "▁everything", - "1" : "▁t", - "557" : "▁bas", - "481" : "▁inc", - "57" : "ot", - "518" : "ail", - "924" : "α", - "173" : "▁lo", - "905" : "ÎŒ", - "762" : "▁probably", - "626" : "▁dec", - "647" : "▁its", - "415" : "orm", - "917" : "ĂŽ", - "560" : "body", - "474" : "▁put", - "572" : "▁em", - "689" : "▁system", - "909" : "Ξ", - "408" : "▁res", - "862" : "1", - "830" : "d", - "748" : "▁question", - "285" : "▁now", - "717" : "▁prod", - "60" : "▁e", - "818" : "oney", - "814" : "▁Because", - "893" : "Ă­", - "587" : "▁uh", - "377" : "▁things", - "454" : "▁ro", - "205" : "▁up", - "849" : "j", - "152" : "out", - "994" : "ÎŻ", - "789" : "ope", - "1018" : "Ćș", - "973" : "ĆŸ", - "950" : "Đż", - "901" : "Ăš", - "288" : "▁Wh", - "710" : "▁prob", - "581" : "igh", - "1009" : "ĐŽ", - "47" : "us", - "756" : "ness", - "648" : "▁God", - "43" : "om", - "952" : "Đœ", - "744" : "▁never", - "688" : "▁guys", - "272" : "▁co", - "36" : "▁in", - "405" : "ated", - "741" : "▁fact", - "56" : "ut", - "290" : "ous", - "770" : "▁belie", - "943" : "ĂČ", - "284" : "▁how", - "697" : "▁mod", - "671" : "▁att", - "453" : "▁comp", - "690" : "ew", - "101" : "▁an", - "264" : "ven", - "200" : "▁don", - "279" : "▁were", - "99" : "ht", - "305" : "hing", - "1014" : "Î", - "525" : "▁many", - "738" : "▁such", - "940" : "Δ", - "995" : "ζ", - "491" : "ark", - "18" : "▁h", - "947" : "Ƃ", - "702" : "▁ask", - "49" : "ow", - "808" : "▁gl", - "484" : "▁should", - "907" : "Ă€", - "916" : "Ăą", - "447" : "ang", - "107" : "▁as", - "843" : "v", - "1025" : "▁global", - "267" : "▁really", - "892" : "\/", - "703" : "old", - "679" : "ix", - "601" : "▁ob", - "498" : "ious", - "298" : "▁But", - "300" : "iv", - "565" : "▁Is", - "282" : "ther", - "249" : "our", - "539" : "▁Be", - "356" : "ap", - "528" : "▁sy", - "470" : "xt", - "373" : "ick", - "853" : "C", - "215" : "and", - "869" : "F", - "10" : "at", - "1004" : "Ɠ", - "686" : "ative", - "553" : "ought", - "976" : "ę", - "473" : "ild", - "4" : "in", - "352" : "▁ad", - "951" : "Ă«", - "265" : "▁our", - "427" : "▁her", - "676" : "▁rep", - "982" : "Ú", - "243" : "ies", - "48" : "ic", - "978" : "б", - "28" : "or", - "566" : "ates", - "965" : "ș", - "9" : "▁s", - "92" : "ur", - "16" : "▁c", - "30" : "▁of", - "15" : "▁b", - "547" : "▁str", - "645" : "▁life", - "24" : "▁p", - "310" : "▁his", - "472" : "ory", - "189" : "▁going", - "699" : "ings", - "1029" : "▁bipartisan", - "178" : "▁one", - "67" : "▁ha", - "824" : "n", - "422" : "▁let", - "888" : "$", - "649" : "pect", - "635" : "nds", - "540" : "ically", - "511" : "red", - "827" : "h", - "19" : "ing", - "969" : "Ăč", - "985" : "υ", - "225" : "ake", - "239" : "ard", - "618" : "ike", - "240" : "▁right", - "585" : "ne", - "378" : "▁In", - "616" : "▁world", - "130" : "▁me", - "434" : "▁even", - "599" : "te", - "897" : "ā", - "206" : "▁P", - "336" : "▁U", - "82" : "ld", - "877" : "9", - "368" : "▁than", - "1016" : "ė", - "181" : "ist", - "438" : "▁app", - "611" : "ert", - "567" : "▁ph", - "374" : "way", - "395" : "ase", - "621" : "▁min", - "188" : "▁about", - "663" : "▁stu", - "571" : "▁years", - "94" : "ith", - "758" : "ize", - "180" : "art", - "40" : "▁you", - "904" : "Âż", - "845" : "I", - "268" : "▁more", - "253" : "▁see", - "683" : "▁def", - "856" : "O", - "370" : "▁gu", - "551" : "▁rec", - "8" : "▁o", - "428" : "▁said", - "675" : "▁happen", - "102" : "▁with", - "776" : "▁ca", - "363" : "▁somet", - "757" : "arch", - "914" : "ĂȘ", - "208" : "ort", - "660" : "▁sure", - "207" : "▁out", - "613" : "ost", - "942" : "Đž", - "761" : "blem", - "117" : "▁like", - "129" : "▁sh", - "293" : "ff", - "407" : "▁off", - "595" : "▁long", - "366" : "ire", - "614" : "▁too", - "256" : "ure", - "328" : "▁cl", - "793" : "ational", - "724" : "▁mon", - "989" : "х", - "134" : "▁what", - "45" : "▁n", - "707" : "ility", - "62" : "ay", - "801" : "▁four", - "392" : "▁those", - "160" : "ge", - "369" : "ace", - "236" : "▁will", - "80" : "se", - "708" : "▁high", - "217" : "em", - "944" : "р", - "672" : "▁op", - "754" : "▁Of", - "58" : "▁we", - "131" : "ill", - "810" : "▁eff", - "146" : "▁ne", - "998" : "Æ", - "496" : "▁being", - "975" : "у", - "625" : "▁find", - "132" : "ant", - "884" : "7", - "382" : "▁could", - "459" : "▁start", - "147" : "▁de", - "448" : "▁mean", - "1019" : "Κ", - "241" : "▁thing", - "819" : "▁", - "698" : "▁done", - "444" : "ress", - "46" : "ve", - "971" : "ч", - "646" : "ities", - "713" : "▁pres", - "483" : "▁different", - "577" : "▁inter", - "372" : "og", - "786" : "iew", - "692" : "ied", - "362" : "iff", - "1021" : "ÎŹ", - "586" : "▁why", - "636" : "▁big", - "1023" : "ό", - "162" : "▁Y", - "685" : "▁next", - "193" : "ok", - "221" : "▁D", - "665" : "▁ev", - "159" : "▁pro", - "59" : "▁is", - "1017" : "Ć ", - "74" : "▁Th", - "813" : "▁Am", - "5" : "▁the", - "218" : "oug", - "376" : "▁un", - "270" : "ose", - "140" : "▁C", - "122" : "▁they", - "412" : "▁talk", - "913" : "σ", - "326" : "ice", - "451" : "▁does", - "559" : "erm", - "23" : "▁f", - "210" : "ment", - "778" : "▁ass", - "984" : "Îș", - "316" : "▁no", - "666" : "ments", - "494" : "▁come", - "887" : "Q", - "891" : "Ă©", - "404" : "uch", - "920" : "É", - "455" : "▁bl", - "170" : "est", - "72" : "ig", - "443" : "▁again", - "727" : "▁ent", - "983" : "Ω", - "410" : "ance", - "281" : "▁act", - "763" : "hip", - "220" : "os", - "261" : "▁You", - "214" : "ich", - "930" : "Ăź", - "805" : "▁called", - "156" : "ers", - "753" : "▁course", - "868" : "G", - "238" : "▁which", - "836" : "w", - "704" : "ered", - "513" : "▁ke", - "964" : "Ă»", - "589" : "▁cons", - "409" : "ac", - "609" : "gr", - "476" : "▁new", - "619" : "▁Okay", - "287" : "ud", - "603" : "ady", - "633" : "▁real", - "508" : "ty", - "1012" : "ÂĄ", - "401" : "▁year", - "110" : "ust", - "196" : "ind", - "242" : "▁want", - "817" : "▁Like", - "890" : "X", - "191" : "▁wor", - "402" : "ass", - "266" : "ci", - "347" : "alk", - "387" : "ach", - "804" : "ont", - "701" : "ism", - "841" : ".", - "317" : "▁en", - "385" : "irst", - "255" : "ast", - "510" : "onna", - "20" : "▁to", - "667" : "▁another", - "591" : "▁after", - "354" : "▁over", - "112" : "▁j", - "674" : "▁sl", - "935" : "ß", - "204" : "ity", - "735" : "▁Well", - "900" : "ö", - "163" : "▁O", - "169" : "▁al", - "531" : "▁great", - "185" : "ink", - "752" : "ather", - "795" : "▁trying", - "796" : "▁sch", - "721" : "oy", - "657" : "▁person", - "799" : "atter", - "554" : "day", - "274" : "ough", - "109" : "▁are", - "643" : "ible", - "968" : "Ø", - "269" : "ound", - "145" : "▁just", - "263" : "▁when", - "85" : "▁k", - "365" : "nder", - "359" : "▁po", - "937" : "ĂŻ", - "226" : "▁int", - "584" : "▁des", - "512" : "wn", - "495" : "▁They", - "391" : "▁He", - "883" : "6", - "627" : "ular", - "927" : "а", - "941" : "Đ”", - "330" : "▁way", - "860" : "Y", - "462" : "▁bet" -} \ No newline at end of file diff --git a/parakeet-ctc-110m-coreml/.DS_Store b/parakeet-ctc-110m-coreml/.DS_Store deleted file mode 100644 index 287c48c58388c597f097f3f208a17a86407f7b8c..0000000000000000000000000000000000000000 Binary files a/parakeet-ctc-110m-coreml/.DS_Store and /dev/null differ diff --git a/parakeet-ctc-110m-coreml/AudioEncoder.mlmodelc/analytics/coremldata.bin b/parakeet-ctc-110m-coreml/AudioEncoder.mlmodelc/analytics/coremldata.bin deleted file mode 100644 index 00fb4ad280710b3595d52ae9835e2227af4ed23b..0000000000000000000000000000000000000000 --- a/parakeet-ctc-110m-coreml/AudioEncoder.mlmodelc/analytics/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:8906c823e9bb3bf6b16d9f0308f98cd70573526333ad85dd767dc3f9ae6b25fa -size 243 diff --git a/parakeet-ctc-110m-coreml/AudioEncoder.mlmodelc/coremldata.bin b/parakeet-ctc-110m-coreml/AudioEncoder.mlmodelc/coremldata.bin deleted file mode 100644 index 8cde3a55d03d57b8be4a9eedab5488ee3d5d482d..0000000000000000000000000000000000000000 --- a/parakeet-ctc-110m-coreml/AudioEncoder.mlmodelc/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:a88b002b58193b4c31211754cdfdf220a85f9651dc61caf336ab84400cbc191a -size 505 diff --git a/parakeet-ctc-110m-coreml/AudioEncoder.mlmodelc/metadata.json b/parakeet-ctc-110m-coreml/AudioEncoder.mlmodelc/metadata.json deleted file mode 100644 index d8cc3094ea923273b3870c1418c982eb8c518704..0000000000000000000000000000000000000000 --- a/parakeet-ctc-110m-coreml/AudioEncoder.mlmodelc/metadata.json +++ /dev/null @@ -1,118 +0,0 @@ -[ - { - "metadataOutputVersion" : "3.0", - "storagePrecision" : "Mixed (Float16, Palettized (6 bits), Sparse)", - "outputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float16", - "formattedType" : "MultiArray (Float16 1 × 512 × 1 × 188)", - "shortDescription" : "", - "shape" : "[1, 512, 1, 188]", - "name" : "encoder_output_embeds", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float16", - "formattedType" : "MultiArray (Float16 1 × 640 × 1 × 188)", - "shortDescription" : "", - "shape" : "[1, 640, 1, 188]", - "name" : "joint_projected_encoder_output_embeds", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float16", - "formattedType" : "MultiArray (Float16 1 × 1025 × 1 × 188)", - "shortDescription" : "", - "shape" : "[1, 1025, 1, 188]", - "name" : "ctc_head_raw_output", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float16", - "formattedType" : "MultiArray (Float16 1 × 1025 × 1 × 188)", - "shortDescription" : "", - "shape" : "[1, 1025, 1, 188]", - "name" : "ctc_head_output", - "type" : "MultiArray" - } - ], - "modelParameters" : [ - - ], - "specificationVersion" : 8, - "mlProgramOperationTypeHistogram" : { - "Pad" : 17, - "Ios17.mul" : 86, - "Split" : 17, - "Ios17.transpose" : 1, - "Ios17.sub" : 1, - "Ios16.constexprLutToDense" : 190, - "Ios17.conv" : 402, - "Ios17.matmul" : 51, - "Ios17.log" : 1, - "Ios16.sigmoid" : 17, - "Ios17.add" : 309, - "Ios17.sliceByIndex" : 34, - "Ios16.constexprSparseToDense" : 190, - "Ios16.relu" : 3, - "Ios17.batchNorm" : 85, - "Ios16.softmax" : 18, - "Ios17.reshape" : 137, - "Ios17.layerNorm" : 85, - "Ios16.silu" : 51 - }, - "computePrecision" : "Mixed (Float16, Float32, Int32)", - "isUpdatable" : "0", - "stateSchema" : [ - - ], - "availability" : { - "macOS" : "14.0", - "tvOS" : "17.0", - "visionOS" : "1.0", - "watchOS" : "10.0", - "iOS" : "17.0", - "macCatalyst" : "17.0" - }, - "modelType" : { - "name" : "MLModelType_mlProgram" - }, - "userDefinedMetadata" : { - "com.github.apple.coremltools.source_dialect" : "TorchScript", - "com.github.apple.coremltools.source" : "torch==2.5.0", - "com.github.apple.coremltools.version" : "8.3.0" - }, - "inputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float16", - "formattedType" : "MultiArray (Float16 1 × 1 × 1501 × 80)", - "shortDescription" : "", - "shape" : "[1, 1, 1501, 80]", - "name" : "melspectrogram_features", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float16", - "formattedType" : "MultiArray (Float16 1 × 1 × 1 × 1)", - "shortDescription" : "", - "shape" : "[1, 1, 1, 1]", - "name" : "input_1", - "type" : "MultiArray" - } - ], - "generatedClassName" : "AudioEncoder_mixedBitPalettized_6_bit", - "method" : "predict" - } -] \ No newline at end of file diff --git a/parakeet-ctc-110m-coreml/AudioEncoder.mlmodelc/model.mil b/parakeet-ctc-110m-coreml/AudioEncoder.mlmodelc/model.mil deleted file mode 100644 index f394aeed7e0f3e8b25ae16e713f2a80558435158..0000000000000000000000000000000000000000 --- a/parakeet-ctc-110m-coreml/AudioEncoder.mlmodelc/model.mil +++ /dev/null @@ -1,4887 +0,0 @@ -program(1.0) -[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3404.16.1"}, {"coremlc-version", "3404.23.1"}})] -{ - func main(tensor input_1, tensor melspectrogram_features) { - tensor pos_emb_to_fp16 = const()[name = tensor("pos_emb_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; - tensor var_42_cast_fp16 = sub(x = pos_emb_to_fp16, y = input_1)[name = tensor("op_42_cast_fp16")]; - tensor obj_3_cast_fp16 = mul(x = var_42_cast_fp16, y = input_1)[name = tensor("obj_3_cast_fp16")]; - tensor input_1_pad_type_0 = const()[name = tensor("input_1_pad_type_0"), val = tensor("custom")]; - tensor input_1_pad_0 = const()[name = tensor("input_1_pad_0"), val = tensor([1, 1, 1, 1])]; - tensor input_1_strides_0 = const()[name = tensor("input_1_strides_0"), val = tensor([2, 2])]; - tensor input_1_dilations_0 = const()[name = tensor("input_1_dilations_0"), val = tensor([1, 1])]; - tensor input_1_groups_0 = const()[name = tensor("input_1_groups_0"), val = tensor(1)]; - tensor pre_encode_conv_0_weight_to_fp16 = const()[name = tensor("pre_encode_conv_0_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(384128)))]; - tensor pre_encode_conv_0_bias_to_fp16 = const()[name = tensor("pre_encode_conv_0_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(388800)))]; - tensor input_1_cast_fp16 = conv(bias = pre_encode_conv_0_bias_to_fp16, dilations = input_1_dilations_0, groups = input_1_groups_0, pad = input_1_pad_0, pad_type = input_1_pad_type_0, strides = input_1_strides_0, weight = pre_encode_conv_0_weight_to_fp16, x = melspectrogram_features)[name = tensor("input_1_cast_fp16")]; - tensor input_3_cast_fp16 = relu(x = input_1_cast_fp16)[name = tensor("input_3_cast_fp16")]; - tensor input_5_pad_type_0 = const()[name = tensor("input_5_pad_type_0"), val = tensor("custom")]; - tensor input_5_pad_0 = const()[name = tensor("input_5_pad_0"), val = tensor([1, 1, 1, 1])]; - tensor input_5_strides_0 = const()[name = tensor("input_5_strides_0"), val = tensor([2, 2])]; - tensor input_5_groups_0 = const()[name = tensor("input_5_groups_0"), val = tensor(256)]; - tensor input_5_dilations_0 = const()[name = tensor("input_5_dilations_0"), val = tensor([1, 1])]; - tensor pre_encode_conv_2_weight_to_fp16 = const()[name = tensor("pre_encode_conv_2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(389376)))]; - tensor pre_encode_conv_2_bias_to_fp16 = const()[name = tensor("pre_encode_conv_2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(394048)))]; - tensor input_5_cast_fp16 = conv(bias = pre_encode_conv_2_bias_to_fp16, dilations = input_5_dilations_0, groups = input_5_groups_0, pad = input_5_pad_0, pad_type = input_5_pad_type_0, strides = input_5_strides_0, weight = pre_encode_conv_2_weight_to_fp16, x = input_3_cast_fp16)[name = tensor("input_5_cast_fp16")]; - tensor input_7_pad_type_0 = const()[name = tensor("input_7_pad_type_0"), val = tensor("valid")]; - tensor input_7_strides_0 = const()[name = tensor("input_7_strides_0"), val = tensor([1, 1])]; - tensor input_7_pad_0 = const()[name = tensor("input_7_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor input_7_dilations_0 = const()[name = tensor("input_7_dilations_0"), val = tensor([1, 1])]; - tensor input_7_groups_0 = const()[name = tensor("input_7_groups_0"), val = tensor(1)]; - tensor pre_encode_conv_3_weight_to_fp16 = const()[name = tensor("pre_encode_conv_3_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(394624)))]; - tensor pre_encode_conv_3_bias_to_fp16 = const()[name = tensor("pre_encode_conv_3_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(525760)))]; - tensor input_7_cast_fp16 = conv(bias = pre_encode_conv_3_bias_to_fp16, dilations = input_7_dilations_0, groups = input_7_groups_0, pad = input_7_pad_0, pad_type = input_7_pad_type_0, strides = input_7_strides_0, weight = pre_encode_conv_3_weight_to_fp16, x = input_5_cast_fp16)[name = tensor("input_7_cast_fp16")]; - tensor input_9_cast_fp16 = relu(x = input_7_cast_fp16)[name = tensor("input_9_cast_fp16")]; - tensor input_11_pad_type_0 = const()[name = tensor("input_11_pad_type_0"), val = tensor("custom")]; - tensor input_11_pad_0 = const()[name = tensor("input_11_pad_0"), val = tensor([1, 1, 1, 1])]; - tensor input_11_strides_0 = const()[name = tensor("input_11_strides_0"), val = tensor([2, 2])]; - tensor input_11_groups_0 = const()[name = tensor("input_11_groups_0"), val = tensor(256)]; - tensor input_11_dilations_0 = const()[name = tensor("input_11_dilations_0"), val = tensor([1, 1])]; - tensor pre_encode_conv_5_weight_to_fp16 = const()[name = tensor("pre_encode_conv_5_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(526336)))]; - tensor pre_encode_conv_5_bias_to_fp16 = const()[name = tensor("pre_encode_conv_5_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(531008)))]; - tensor input_11_cast_fp16 = conv(bias = pre_encode_conv_5_bias_to_fp16, dilations = input_11_dilations_0, groups = input_11_groups_0, pad = input_11_pad_0, pad_type = input_11_pad_type_0, strides = input_11_strides_0, weight = pre_encode_conv_5_weight_to_fp16, x = input_9_cast_fp16)[name = tensor("input_11_cast_fp16")]; - tensor input_13_pad_type_0 = const()[name = tensor("input_13_pad_type_0"), val = tensor("valid")]; - tensor input_13_strides_0 = const()[name = tensor("input_13_strides_0"), val = tensor([1, 1])]; - tensor input_13_pad_0 = const()[name = tensor("input_13_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor input_13_dilations_0 = const()[name = tensor("input_13_dilations_0"), val = tensor([1, 1])]; - tensor input_13_groups_0 = const()[name = tensor("input_13_groups_0"), val = tensor(1)]; - tensor pre_encode_conv_6_weight_to_fp16 = const()[name = tensor("pre_encode_conv_6_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(531584)))]; - tensor pre_encode_conv_6_bias_to_fp16 = const()[name = tensor("pre_encode_conv_6_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(662720)))]; - tensor input_13_cast_fp16 = conv(bias = pre_encode_conv_6_bias_to_fp16, dilations = input_13_dilations_0, groups = input_13_groups_0, pad = input_13_pad_0, pad_type = input_13_pad_type_0, strides = input_13_strides_0, weight = pre_encode_conv_6_weight_to_fp16, x = input_11_cast_fp16)[name = tensor("input_13_cast_fp16")]; - tensor x_1_cast_fp16 = relu(x = input_13_cast_fp16)[name = tensor("x_1_cast_fp16")]; - tensor var_103_perm_0 = const()[name = tensor("op_103_perm_0"), val = tensor([0, 1, 3, 2])]; - tensor var_106 = const()[name = tensor("op_106"), val = tensor([1, 2560, 1, 188])]; - tensor var_103_cast_fp16 = transpose(perm = var_103_perm_0, x = x_1_cast_fp16)[name = tensor("transpose_0")]; - tensor input_15_cast_fp16 = reshape(shape = var_106, x = var_103_cast_fp16)[name = tensor("input_15_cast_fp16")]; - tensor var_116_pad_type_0 = const()[name = tensor("op_116_pad_type_0"), val = tensor("valid")]; - tensor var_116_strides_0 = const()[name = tensor("op_116_strides_0"), val = tensor([1, 1])]; - tensor var_116_pad_0 = const()[name = tensor("op_116_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_116_dilations_0 = const()[name = tensor("op_116_dilations_0"), val = tensor([1, 1])]; - tensor var_116_groups_0 = const()[name = tensor("op_116_groups_0"), val = tensor(1)]; - tensor pre_encode_out_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(663296))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1646400))), name = tensor("pre_encode_out_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 2560, 1, 1])]; - tensor pre_encode_out_inlier_module_bias_to_fp16 = const()[name = tensor("pre_encode_out_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1646592)))]; - tensor var_116_cast_fp16 = conv(bias = pre_encode_out_inlier_module_bias_to_fp16, dilations = var_116_dilations_0, groups = var_116_groups_0, pad = var_116_pad_0, pad_type = var_116_pad_type_0, strides = var_116_strides_0, weight = pre_encode_out_inlier_module_weight_to_fp16_palettized, x = input_15_cast_fp16)[name = tensor("op_116_cast_fp16")]; - tensor var_122_pad_type_0 = const()[name = tensor("op_122_pad_type_0"), val = tensor("valid")]; - tensor var_122_strides_0 = const()[name = tensor("op_122_strides_0"), val = tensor([1, 1])]; - tensor var_122_pad_0 = const()[name = tensor("op_122_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_122_dilations_0 = const()[name = tensor("op_122_dilations_0"), val = tensor([1, 1])]; - tensor var_122_groups_0 = const()[name = tensor("op_122_groups_0"), val = tensor(1)]; - tensor pre_encode_out_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1696256))), name = tensor("pre_encode_out_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1647680))), shape = tensor([512, 2560, 1, 1])]; - tensor var_122_cast_fp16 = conv(dilations = var_122_dilations_0, groups = var_122_groups_0, pad = var_122_pad_0, pad_type = var_122_pad_type_0, strides = var_122_strides_0, weight = pre_encode_out_outlier_module_weight_to_fp16_sparsified, x = input_15_cast_fp16)[name = tensor("op_122_cast_fp16")]; - tensor inputs_1_cast_fp16 = add(x = var_116_cast_fp16, y = var_122_cast_fp16)[name = tensor("inputs_1_cast_fp16")]; - tensor var_128 = const()[name = tensor("op_128"), val = tensor(3)]; - tensor out_1_axes_0 = const()[name = tensor("out_1_axes_0"), val = tensor([1])]; - tensor var_159_to_fp16 = const()[name = tensor("op_159_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_1_cast_fp16 = layer_norm(axes = out_1_axes_0, epsilon = var_159_to_fp16, x = inputs_1_cast_fp16)[name = tensor("out_1_cast_fp16")]; - tensor input_17_mean_0_to_fp16 = const()[name = tensor("input_17_mean_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1860160)))]; - tensor input_17_variance_0_to_fp16 = const()[name = tensor("input_17_variance_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1861248)))]; - tensor input_17_gamma_0_to_fp16 = const()[name = tensor("input_17_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1862336)))]; - tensor input_17_beta_0_to_fp16 = const()[name = tensor("input_17_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1863424)))]; - tensor input_17_epsilon_0_to_fp16 = const()[name = tensor("input_17_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_17_cast_fp16 = batch_norm(beta = input_17_beta_0_to_fp16, epsilon = input_17_epsilon_0_to_fp16, gamma = input_17_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_1_cast_fp16)[name = tensor("input_17_cast_fp16")]; - tensor var_179_pad_type_0 = const()[name = tensor("op_179_pad_type_0"), val = tensor("valid")]; - tensor var_179_strides_0 = const()[name = tensor("op_179_strides_0"), val = tensor([1, 1])]; - tensor var_179_pad_0 = const()[name = tensor("op_179_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_179_dilations_0 = const()[name = tensor("op_179_dilations_0"), val = tensor([1, 1])]; - tensor var_179_groups_0 = const()[name = tensor("op_179_groups_0"), val = tensor(1)]; - tensor layers_0_feed_forward1_fc1_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1864512))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2651008))), name = tensor("layers_0_feed_forward1_fc1_inlier_module_weight_to_fp16_palettized"), shape = tensor([2048, 512, 1, 1])]; - tensor layers_0_feed_forward1_fc1_inlier_module_bias_to_fp16 = const()[name = tensor("layers_0_feed_forward1_fc1_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2651200)))]; - tensor var_179_cast_fp16 = conv(bias = layers_0_feed_forward1_fc1_inlier_module_bias_to_fp16, dilations = var_179_dilations_0, groups = var_179_groups_0, pad = var_179_pad_0, pad_type = var_179_pad_type_0, strides = var_179_strides_0, weight = layers_0_feed_forward1_fc1_inlier_module_weight_to_fp16_palettized, x = input_17_cast_fp16)[name = tensor("op_179_cast_fp16")]; - tensor var_185_pad_type_0 = const()[name = tensor("op_185_pad_type_0"), val = tensor("valid")]; - tensor var_185_strides_0 = const()[name = tensor("op_185_strides_0"), val = tensor([1, 1])]; - tensor var_185_pad_0 = const()[name = tensor("op_185_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_185_dilations_0 = const()[name = tensor("op_185_dilations_0"), val = tensor([1, 1])]; - tensor var_185_groups_0 = const()[name = tensor("op_185_groups_0"), val = tensor(1)]; - tensor layers_0_feed_forward1_fc1_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2688192))), name = tensor("layers_0_feed_forward1_fc1_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2655360))), shape = tensor([2048, 512, 1, 1])]; - tensor var_185_cast_fp16 = conv(dilations = var_185_dilations_0, groups = var_185_groups_0, pad = var_185_pad_0, pad_type = var_185_pad_type_0, strides = var_185_strides_0, weight = layers_0_feed_forward1_fc1_outlier_module_weight_to_fp16_sparsified, x = input_17_cast_fp16)[name = tensor("op_185_cast_fp16")]; - tensor input_19_cast_fp16 = add(x = var_179_cast_fp16, y = var_185_cast_fp16)[name = tensor("input_19_cast_fp16")]; - tensor input_21_cast_fp16 = silu(x = input_19_cast_fp16)[name = tensor("input_21_cast_fp16")]; - tensor var_196_pad_type_0 = const()[name = tensor("op_196_pad_type_0"), val = tensor("valid")]; - tensor var_196_strides_0 = const()[name = tensor("op_196_strides_0"), val = tensor([1, 1])]; - tensor var_196_pad_0 = const()[name = tensor("op_196_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_196_dilations_0 = const()[name = tensor("op_196_dilations_0"), val = tensor([1, 1])]; - tensor var_196_groups_0 = const()[name = tensor("op_196_groups_0"), val = tensor(1)]; - tensor layers_0_feed_forward1_fc2_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2819328))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3605824))), name = tensor("layers_0_feed_forward1_fc2_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 2048, 1, 1])]; - tensor layers_0_feed_forward1_fc2_inlier_module_bias_to_fp16 = const()[name = tensor("layers_0_feed_forward1_fc2_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3606016)))]; - tensor var_196_cast_fp16 = conv(bias = layers_0_feed_forward1_fc2_inlier_module_bias_to_fp16, dilations = var_196_dilations_0, groups = var_196_groups_0, pad = var_196_pad_0, pad_type = var_196_pad_type_0, strides = var_196_strides_0, weight = layers_0_feed_forward1_fc2_inlier_module_weight_to_fp16_palettized, x = input_21_cast_fp16)[name = tensor("op_196_cast_fp16")]; - tensor var_202_pad_type_0 = const()[name = tensor("op_202_pad_type_0"), val = tensor("valid")]; - tensor var_202_strides_0 = const()[name = tensor("op_202_strides_0"), val = tensor([1, 1])]; - tensor var_202_pad_0 = const()[name = tensor("op_202_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_202_dilations_0 = const()[name = tensor("op_202_dilations_0"), val = tensor([1, 1])]; - tensor var_202_groups_0 = const()[name = tensor("op_202_groups_0"), val = tensor(1)]; - tensor layers_0_feed_forward1_fc2_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3635200))), name = tensor("layers_0_feed_forward1_fc2_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3607104))), shape = tensor([512, 2048, 1, 1])]; - tensor var_202_cast_fp16 = conv(dilations = var_202_dilations_0, groups = var_202_groups_0, pad = var_202_pad_0, pad_type = var_202_pad_type_0, strides = var_202_strides_0, weight = layers_0_feed_forward1_fc2_outlier_module_weight_to_fp16_sparsified, x = input_21_cast_fp16)[name = tensor("op_202_cast_fp16")]; - tensor x_3_cast_fp16 = add(x = var_196_cast_fp16, y = var_202_cast_fp16)[name = tensor("x_3_cast_fp16")]; - tensor var_204_to_fp16 = const()[name = tensor("op_204_to_fp16"), val = tensor(0x1p-1)]; - tensor var_205_cast_fp16 = mul(x = x_3_cast_fp16, y = var_204_to_fp16)[name = tensor("op_205_cast_fp16")]; - tensor inputs_3_cast_fp16 = add(x = inputs_1_cast_fp16, y = var_205_cast_fp16)[name = tensor("inputs_3_cast_fp16")]; - tensor out_3_axes_0 = const()[name = tensor("out_3_axes_0"), val = tensor([1])]; - tensor var_215_to_fp16 = const()[name = tensor("op_215_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_3_cast_fp16 = layer_norm(axes = out_3_axes_0, epsilon = var_215_to_fp16, x = inputs_3_cast_fp16)[name = tensor("out_3_cast_fp16")]; - tensor obj_1_gamma_0_to_fp16 = const()[name = tensor("obj_1_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3766336)))]; - tensor obj_1_beta_0_to_fp16 = const()[name = tensor("obj_1_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3767424)))]; - tensor obj_1_epsilon_0_to_fp16 = const()[name = tensor("obj_1_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor obj_1_cast_fp16 = batch_norm(beta = obj_1_beta_0_to_fp16, epsilon = obj_1_epsilon_0_to_fp16, gamma = obj_1_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_3_cast_fp16)[name = tensor("obj_1_cast_fp16")]; - tensor var_240_pad_type_0 = const()[name = tensor("op_240_pad_type_0"), val = tensor("valid")]; - tensor var_240_strides_0 = const()[name = tensor("op_240_strides_0"), val = tensor([1, 1])]; - tensor var_240_pad_0 = const()[name = tensor("op_240_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_240_dilations_0 = const()[name = tensor("op_240_dilations_0"), val = tensor([1, 1])]; - tensor var_240_groups_0 = const()[name = tensor("op_240_groups_0"), val = tensor(1)]; - tensor layers_0_self_attn_q_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3768512))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3965184))), name = tensor("layers_0_self_attn_q_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_0_self_attn_q_proj_inlier_module_bias_to_fp16 = const()[name = tensor("layers_0_self_attn_q_proj_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3965376)))]; - tensor var_240_cast_fp16 = conv(bias = layers_0_self_attn_q_proj_inlier_module_bias_to_fp16, dilations = var_240_dilations_0, groups = var_240_groups_0, pad = var_240_pad_0, pad_type = var_240_pad_type_0, strides = var_240_strides_0, weight = layers_0_self_attn_q_proj_inlier_module_weight_to_fp16_palettized, x = obj_1_cast_fp16)[name = tensor("op_240_cast_fp16")]; - tensor var_246_pad_type_0 = const()[name = tensor("op_246_pad_type_0"), val = tensor("valid")]; - tensor var_246_strides_0 = const()[name = tensor("op_246_strides_0"), val = tensor([1, 1])]; - tensor var_246_pad_0 = const()[name = tensor("op_246_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_246_dilations_0 = const()[name = tensor("op_246_dilations_0"), val = tensor([1, 1])]; - tensor var_246_groups_0 = const()[name = tensor("op_246_groups_0"), val = tensor(1)]; - tensor layers_0_self_attn_q_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3977984))), name = tensor("layers_0_self_attn_q_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3966464))), shape = tensor([512, 512, 1, 1])]; - tensor var_246_cast_fp16 = conv(dilations = var_246_dilations_0, groups = var_246_groups_0, pad = var_246_pad_0, pad_type = var_246_pad_type_0, strides = var_246_strides_0, weight = layers_0_self_attn_q_proj_outlier_module_weight_to_fp16_sparsified, x = obj_1_cast_fp16)[name = tensor("op_246_cast_fp16")]; - tensor query_1_cast_fp16 = add(x = var_240_cast_fp16, y = var_246_cast_fp16)[name = tensor("query_1_cast_fp16")]; - tensor var_255_pad_type_0 = const()[name = tensor("op_255_pad_type_0"), val = tensor("valid")]; - tensor var_255_strides_0 = const()[name = tensor("op_255_strides_0"), val = tensor([1, 1])]; - tensor var_255_pad_0 = const()[name = tensor("op_255_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_255_dilations_0 = const()[name = tensor("op_255_dilations_0"), val = tensor([1, 1])]; - tensor var_255_groups_0 = const()[name = tensor("op_255_groups_0"), val = tensor(1)]; - tensor layers_0_self_attn_k_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4010816))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4207488))), name = tensor("layers_0_self_attn_k_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor var_255_cast_fp16 = conv(dilations = var_255_dilations_0, groups = var_255_groups_0, pad = var_255_pad_0, pad_type = var_255_pad_type_0, strides = var_255_strides_0, weight = layers_0_self_attn_k_proj_inlier_module_weight_to_fp16_palettized, x = obj_1_cast_fp16)[name = tensor("op_255_cast_fp16")]; - tensor var_261_pad_type_0 = const()[name = tensor("op_261_pad_type_0"), val = tensor("valid")]; - tensor var_261_strides_0 = const()[name = tensor("op_261_strides_0"), val = tensor([1, 1])]; - tensor var_261_pad_0 = const()[name = tensor("op_261_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_261_dilations_0 = const()[name = tensor("op_261_dilations_0"), val = tensor([1, 1])]; - tensor var_261_groups_0 = const()[name = tensor("op_261_groups_0"), val = tensor(1)]; - tensor layers_0_self_attn_k_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4220160))), name = tensor("layers_0_self_attn_k_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4207680))), shape = tensor([512, 512, 1, 1])]; - tensor var_261_cast_fp16 = conv(dilations = var_261_dilations_0, groups = var_261_groups_0, pad = var_261_pad_0, pad_type = var_261_pad_type_0, strides = var_261_strides_0, weight = layers_0_self_attn_k_proj_outlier_module_weight_to_fp16_sparsified, x = obj_1_cast_fp16)[name = tensor("op_261_cast_fp16")]; - tensor key_1_cast_fp16 = add(x = var_255_cast_fp16, y = var_261_cast_fp16)[name = tensor("key_1_cast_fp16")]; - tensor var_271_pad_type_0 = const()[name = tensor("op_271_pad_type_0"), val = tensor("valid")]; - tensor var_271_strides_0 = const()[name = tensor("op_271_strides_0"), val = tensor([1, 1])]; - tensor var_271_pad_0 = const()[name = tensor("op_271_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_271_dilations_0 = const()[name = tensor("op_271_dilations_0"), val = tensor([1, 1])]; - tensor var_271_groups_0 = const()[name = tensor("op_271_groups_0"), val = tensor(1)]; - tensor layers_0_self_attn_v_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4252992))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4449664))), name = tensor("layers_0_self_attn_v_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_0_self_attn_v_proj_inlier_module_bias_to_fp16 = const()[name = tensor("layers_0_self_attn_v_proj_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4449856)))]; - tensor var_271_cast_fp16 = conv(bias = layers_0_self_attn_v_proj_inlier_module_bias_to_fp16, dilations = var_271_dilations_0, groups = var_271_groups_0, pad = var_271_pad_0, pad_type = var_271_pad_type_0, strides = var_271_strides_0, weight = layers_0_self_attn_v_proj_inlier_module_weight_to_fp16_palettized, x = obj_1_cast_fp16)[name = tensor("op_271_cast_fp16")]; - tensor var_277_pad_type_0 = const()[name = tensor("op_277_pad_type_0"), val = tensor("valid")]; - tensor var_277_strides_0 = const()[name = tensor("op_277_strides_0"), val = tensor([1, 1])]; - tensor var_277_pad_0 = const()[name = tensor("op_277_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_277_dilations_0 = const()[name = tensor("op_277_dilations_0"), val = tensor([1, 1])]; - tensor var_277_groups_0 = const()[name = tensor("op_277_groups_0"), val = tensor(1)]; - tensor layers_0_self_attn_v_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4461312))), name = tensor("layers_0_self_attn_v_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4450944))), shape = tensor([512, 512, 1, 1])]; - tensor var_277_cast_fp16 = conv(dilations = var_277_dilations_0, groups = var_277_groups_0, pad = var_277_pad_0, pad_type = var_277_pad_type_0, strides = var_277_strides_0, weight = layers_0_self_attn_v_proj_outlier_module_weight_to_fp16_sparsified, x = obj_1_cast_fp16)[name = tensor("op_277_cast_fp16")]; - tensor value_1_cast_fp16 = add(x = var_271_cast_fp16, y = var_277_cast_fp16)[name = tensor("value_1_cast_fp16")]; - tensor var_280_to_fp16 = const()[name = tensor("op_280_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4494144)))]; - tensor query_3_cast_fp16 = add(x = query_1_cast_fp16, y = var_280_to_fp16)[name = tensor("query_3_cast_fp16")]; - tensor var_283_to_fp16 = const()[name = tensor("op_283_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4495232)))]; - tensor q_with_bias_v_1_cast_fp16 = add(x = query_1_cast_fp16, y = var_283_to_fp16)[name = tensor("q_with_bias_v_1_cast_fp16")]; - tensor var_293_pad_type_0 = const()[name = tensor("op_293_pad_type_0"), val = tensor("valid")]; - tensor var_293_strides_0 = const()[name = tensor("op_293_strides_0"), val = tensor([1, 1])]; - tensor var_293_pad_0 = const()[name = tensor("op_293_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_293_dilations_0 = const()[name = tensor("op_293_dilations_0"), val = tensor([1, 1])]; - tensor var_293_groups_0 = const()[name = tensor("op_293_groups_0"), val = tensor(1)]; - tensor layers_0_self_attn_linear_pos_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4496320))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4692992))), name = tensor("layers_0_self_attn_linear_pos_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor var_293_cast_fp16 = conv(bias = input_17_mean_0_to_fp16, dilations = var_293_dilations_0, groups = var_293_groups_0, pad = var_293_pad_0, pad_type = var_293_pad_type_0, strides = var_293_strides_0, weight = layers_0_self_attn_linear_pos_inlier_module_weight_to_fp16_palettized, x = obj_3_cast_fp16)[name = tensor("op_293_cast_fp16")]; - tensor var_299_pad_type_0 = const()[name = tensor("op_299_pad_type_0"), val = tensor("valid")]; - tensor var_299_strides_0 = const()[name = tensor("op_299_strides_0"), val = tensor([1, 1])]; - tensor var_299_pad_0 = const()[name = tensor("op_299_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_299_dilations_0 = const()[name = tensor("op_299_dilations_0"), val = tensor([1, 1])]; - tensor var_299_groups_0 = const()[name = tensor("op_299_groups_0"), val = tensor(1)]; - tensor layers_0_self_attn_linear_pos_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4712704))), name = tensor("layers_0_self_attn_linear_pos_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4693184))), shape = tensor([512, 512, 1, 1])]; - tensor var_299_cast_fp16 = conv(dilations = var_299_dilations_0, groups = var_299_groups_0, pad = var_299_pad_0, pad_type = var_299_pad_type_0, strides = var_299_strides_0, weight = layers_0_self_attn_linear_pos_outlier_module_weight_to_fp16_sparsified, x = obj_3_cast_fp16)[name = tensor("op_299_cast_fp16")]; - tensor p_1_cast_fp16 = add(x = var_293_cast_fp16, y = var_299_cast_fp16)[name = tensor("p_1_cast_fp16")]; - tensor var_303 = const()[name = tensor("op_303"), val = tensor([1, 8, 64, 188])]; - tensor var_304_cast_fp16 = reshape(shape = var_303, x = q_with_bias_v_1_cast_fp16)[name = tensor("op_304_cast_fp16")]; - tensor var_305 = const()[name = tensor("op_305"), val = tensor([1, 8, 64, -1])]; - tensor var_306_cast_fp16 = reshape(shape = var_305, x = p_1_cast_fp16)[name = tensor("op_306_cast_fp16")]; - tensor matrix_bd_1_transpose_x_0 = const()[name = tensor("matrix_bd_1_transpose_x_0"), val = tensor(true)]; - tensor matrix_bd_1_transpose_y_0 = const()[name = tensor("matrix_bd_1_transpose_y_0"), val = tensor(false)]; - tensor matrix_bd_1_cast_fp16 = matmul(transpose_x = matrix_bd_1_transpose_x_0, transpose_y = matrix_bd_1_transpose_y_0, x = var_304_cast_fp16, y = var_306_cast_fp16)[name = tensor("matrix_bd_1_cast_fp16")]; - tensor matrix_bd_3_pad_0 = const()[name = tensor("matrix_bd_3_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor matrix_bd_3_mode_0 = const()[name = tensor("matrix_bd_3_mode_0"), val = tensor("constant")]; - tensor const_10_to_fp16 = const()[name = tensor("const_10_to_fp16"), val = tensor(0x0p+0)]; - tensor matrix_bd_3_cast_fp16 = pad(constant_val = const_10_to_fp16, mode = matrix_bd_3_mode_0, pad = matrix_bd_3_pad_0, x = matrix_bd_1_cast_fp16)[name = tensor("matrix_bd_3_cast_fp16")]; - tensor var_315 = const()[name = tensor("op_315"), val = tensor([1, 8, -1, 188])]; - tensor matrix_bd_5_cast_fp16 = reshape(shape = var_315, x = matrix_bd_3_cast_fp16)[name = tensor("matrix_bd_5_cast_fp16")]; - tensor var_319_begin_0 = const()[name = tensor("op_319_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_319_end_0 = const()[name = tensor("op_319_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_319_end_mask_0 = const()[name = tensor("op_319_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_319_cast_fp16 = slice_by_index(begin = var_319_begin_0, end = var_319_end_0, end_mask = var_319_end_mask_0, x = matrix_bd_5_cast_fp16)[name = tensor("op_319_cast_fp16")]; - tensor var_320 = const()[name = tensor("op_320"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_7_cast_fp16 = reshape(shape = var_320, x = var_319_cast_fp16)[name = tensor("matrix_bd_7_cast_fp16")]; - tensor var_325_begin_0 = const()[name = tensor("op_325_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor var_325_end_0 = const()[name = tensor("op_325_end_0"), val = tensor([1, 8, 188, 188])]; - tensor var_325_end_mask_0 = const()[name = tensor("op_325_end_mask_0"), val = tensor([true, true, true, false])]; - tensor var_325_cast_fp16 = slice_by_index(begin = var_325_begin_0, end = var_325_end_0, end_mask = var_325_end_mask_0, x = matrix_bd_7_cast_fp16)[name = tensor("op_325_cast_fp16")]; - tensor var_326_to_fp16 = const()[name = tensor("op_326_to_fp16"), val = tensor(0x1p-3)]; - tensor qk_mask_1_cast_fp16 = mul(x = var_325_cast_fp16, y = var_326_to_fp16)[name = tensor("qk_mask_1_cast_fp16")]; - tensor var_330 = const()[name = tensor("op_330"), val = tensor([1, 8, 64, 188])]; - tensor mh_q_1_cast_fp16 = reshape(shape = var_330, x = query_3_cast_fp16)[name = tensor("mh_q_1_cast_fp16")]; - tensor var_332_to_fp16 = const()[name = tensor("op_332_to_fp16"), val = tensor(0x1p-3)]; - tensor var_333_cast_fp16 = mul(x = mh_q_1_cast_fp16, y = var_332_to_fp16)[name = tensor("op_333_cast_fp16")]; - tensor var_336 = const()[name = tensor("op_336"), val = tensor([1, 8, 64, 188])]; - tensor var_337_cast_fp16 = reshape(shape = var_336, x = key_1_cast_fp16)[name = tensor("op_337_cast_fp16")]; - tensor mh_w_1_transpose_x_0 = const()[name = tensor("mh_w_1_transpose_x_0"), val = tensor(true)]; - tensor mh_w_1_transpose_y_0 = const()[name = tensor("mh_w_1_transpose_y_0"), val = tensor(false)]; - tensor mh_w_1_cast_fp16 = matmul(transpose_x = mh_w_1_transpose_x_0, transpose_y = mh_w_1_transpose_y_0, x = var_333_cast_fp16, y = var_337_cast_fp16)[name = tensor("mh_w_1_cast_fp16")]; - tensor mh_w_3_cast_fp16 = add(x = mh_w_1_cast_fp16, y = qk_mask_1_cast_fp16)[name = tensor("mh_w_3_cast_fp16")]; - tensor var_341_cast_fp16 = softmax(axis = var_128, x = mh_w_3_cast_fp16)[name = tensor("op_341_cast_fp16")]; - tensor var_342 = const()[name = tensor("op_342"), val = tensor([1, 8, 64, 188])]; - tensor var_343_cast_fp16 = reshape(shape = var_342, x = value_1_cast_fp16)[name = tensor("op_343_cast_fp16")]; - tensor attn_1_transpose_x_0 = const()[name = tensor("attn_1_transpose_x_0"), val = tensor(false)]; - tensor attn_1_transpose_y_0 = const()[name = tensor("attn_1_transpose_y_0"), val = tensor(true)]; - tensor attn_1_cast_fp16 = matmul(transpose_x = attn_1_transpose_x_0, transpose_y = attn_1_transpose_y_0, x = var_343_cast_fp16, y = var_341_cast_fp16)[name = tensor("attn_1_cast_fp16")]; - tensor var_346 = const()[name = tensor("op_346"), val = tensor([1, 512, 1, 188])]; - tensor input_23_cast_fp16 = reshape(shape = var_346, x = attn_1_cast_fp16)[name = tensor("input_23_cast_fp16")]; - tensor var_356_pad_type_0 = const()[name = tensor("op_356_pad_type_0"), val = tensor("valid")]; - tensor var_356_strides_0 = const()[name = tensor("op_356_strides_0"), val = tensor([1, 1])]; - tensor var_356_pad_0 = const()[name = tensor("op_356_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_356_dilations_0 = const()[name = tensor("op_356_dilations_0"), val = tensor([1, 1])]; - tensor var_356_groups_0 = const()[name = tensor("op_356_groups_0"), val = tensor(1)]; - tensor layers_0_self_attn_o_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4745536))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4942208))), name = tensor("layers_0_self_attn_o_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_0_self_attn_o_proj_inlier_module_bias_to_fp16 = const()[name = tensor("layers_0_self_attn_o_proj_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4942400)))]; - tensor var_356_cast_fp16 = conv(bias = layers_0_self_attn_o_proj_inlier_module_bias_to_fp16, dilations = var_356_dilations_0, groups = var_356_groups_0, pad = var_356_pad_0, pad_type = var_356_pad_type_0, strides = var_356_strides_0, weight = layers_0_self_attn_o_proj_inlier_module_weight_to_fp16_palettized, x = input_23_cast_fp16)[name = tensor("op_356_cast_fp16")]; - tensor var_362_pad_type_0 = const()[name = tensor("op_362_pad_type_0"), val = tensor("valid")]; - tensor var_362_strides_0 = const()[name = tensor("op_362_strides_0"), val = tensor([1, 1])]; - tensor var_362_pad_0 = const()[name = tensor("op_362_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_362_dilations_0 = const()[name = tensor("op_362_dilations_0"), val = tensor([1, 1])]; - tensor var_362_groups_0 = const()[name = tensor("op_362_groups_0"), val = tensor(1)]; - tensor layers_0_self_attn_o_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4952448))), name = tensor("layers_0_self_attn_o_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4943488))), shape = tensor([512, 512, 1, 1])]; - tensor var_362_cast_fp16 = conv(dilations = var_362_dilations_0, groups = var_362_groups_0, pad = var_362_pad_0, pad_type = var_362_pad_type_0, strides = var_362_strides_0, weight = layers_0_self_attn_o_proj_outlier_module_weight_to_fp16_sparsified, x = input_23_cast_fp16)[name = tensor("op_362_cast_fp16")]; - tensor obj_5_cast_fp16 = add(x = var_356_cast_fp16, y = var_362_cast_fp16)[name = tensor("obj_5_cast_fp16")]; - tensor inputs_5_cast_fp16 = add(x = inputs_3_cast_fp16, y = obj_5_cast_fp16)[name = tensor("inputs_5_cast_fp16")]; - tensor out_5_axes_0 = const()[name = tensor("out_5_axes_0"), val = tensor([1])]; - tensor var_373_to_fp16 = const()[name = tensor("op_373_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_5_cast_fp16 = layer_norm(axes = out_5_axes_0, epsilon = var_373_to_fp16, x = inputs_5_cast_fp16)[name = tensor("out_5_cast_fp16")]; - tensor input_25_gamma_0_to_fp16 = const()[name = tensor("input_25_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4985280)))]; - tensor input_25_beta_0_to_fp16 = const()[name = tensor("input_25_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4986368)))]; - tensor input_25_epsilon_0_to_fp16 = const()[name = tensor("input_25_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_25_cast_fp16 = batch_norm(beta = input_25_beta_0_to_fp16, epsilon = input_25_epsilon_0_to_fp16, gamma = input_25_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_5_cast_fp16)[name = tensor("input_25_cast_fp16")]; - tensor var_395_pad_type_0 = const()[name = tensor("op_395_pad_type_0"), val = tensor("valid")]; - tensor var_395_strides_0 = const()[name = tensor("op_395_strides_0"), val = tensor([1, 1])]; - tensor var_395_pad_0 = const()[name = tensor("op_395_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_395_dilations_0 = const()[name = tensor("op_395_dilations_0"), val = tensor([1, 1])]; - tensor var_395_groups_0 = const()[name = tensor("op_395_groups_0"), val = tensor(1)]; - tensor layers_0_conv_pointwise_conv1_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4987456))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(5380736))), name = tensor("layers_0_conv_pointwise_conv1_inlier_module_weight_to_fp16_palettized"), shape = tensor([1024, 512, 1, 1])]; - tensor layers_0_conv_pointwise_conv1_inlier_module_bias_to_fp16 = const()[name = tensor("layers_0_conv_pointwise_conv1_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(5380928)))]; - tensor var_395_cast_fp16 = conv(bias = layers_0_conv_pointwise_conv1_inlier_module_bias_to_fp16, dilations = var_395_dilations_0, groups = var_395_groups_0, pad = var_395_pad_0, pad_type = var_395_pad_type_0, strides = var_395_strides_0, weight = layers_0_conv_pointwise_conv1_inlier_module_weight_to_fp16_palettized, x = input_25_cast_fp16)[name = tensor("op_395_cast_fp16")]; - tensor var_401_pad_type_0 = const()[name = tensor("op_401_pad_type_0"), val = tensor("valid")]; - tensor var_401_strides_0 = const()[name = tensor("op_401_strides_0"), val = tensor([1, 1])]; - tensor var_401_pad_0 = const()[name = tensor("op_401_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_401_dilations_0 = const()[name = tensor("op_401_dilations_0"), val = tensor([1, 1])]; - tensor var_401_groups_0 = const()[name = tensor("op_401_groups_0"), val = tensor(1)]; - tensor layers_0_conv_pointwise_conv1_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(5401344))), name = tensor("layers_0_conv_pointwise_conv1_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(5383040))), shape = tensor([1024, 512, 1, 1])]; - tensor var_401_cast_fp16 = conv(dilations = var_401_dilations_0, groups = var_401_groups_0, pad = var_401_pad_0, pad_type = var_401_pad_type_0, strides = var_401_strides_0, weight = layers_0_conv_pointwise_conv1_outlier_module_weight_to_fp16_sparsified, x = input_25_cast_fp16)[name = tensor("op_401_cast_fp16")]; - tensor input_27_cast_fp16 = add(x = var_395_cast_fp16, y = var_401_cast_fp16)[name = tensor("input_27_cast_fp16")]; - tensor input_29_split_num_splits_0 = const()[name = tensor("input_29_split_num_splits_0"), val = tensor(2)]; - tensor input_29_split_axis_0 = const()[name = tensor("input_29_split_axis_0"), val = tensor(1)]; - tensor input_29_split_cast_fp16_0, tensor input_29_split_cast_fp16_1 = split(axis = input_29_split_axis_0, num_splits = input_29_split_num_splits_0, x = input_27_cast_fp16)[name = tensor("input_29_split_cast_fp16")]; - tensor input_29_split_1_sigmoid_cast_fp16 = sigmoid(x = input_29_split_cast_fp16_1)[name = tensor("input_29_split_1_sigmoid_cast_fp16")]; - tensor input_29_cast_fp16 = mul(x = input_29_split_cast_fp16_0, y = input_29_split_1_sigmoid_cast_fp16)[name = tensor("input_29_cast_fp16")]; - tensor input_31_pad_type_0 = const()[name = tensor("input_31_pad_type_0"), val = tensor("custom")]; - tensor input_31_pad_0 = const()[name = tensor("input_31_pad_0"), val = tensor([0, 0, 4, 4])]; - tensor input_31_groups_0 = const()[name = tensor("input_31_groups_0"), val = tensor(512)]; - tensor input_31_strides_0 = const()[name = tensor("input_31_strides_0"), val = tensor([1, 1])]; - tensor input_31_dilations_0 = const()[name = tensor("input_31_dilations_0"), val = tensor([1, 1])]; - tensor const_191_to_fp16 = const()[name = tensor("const_191_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(5466944)))]; - tensor const_192_to_fp16 = const()[name = tensor("const_192_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(5476224)))]; - tensor input_33_cast_fp16 = conv(bias = const_192_to_fp16, dilations = input_31_dilations_0, groups = input_31_groups_0, pad = input_31_pad_0, pad_type = input_31_pad_type_0, strides = input_31_strides_0, weight = const_191_to_fp16, x = input_29_cast_fp16)[name = tensor("input_33_cast_fp16")]; - tensor input_35_cast_fp16 = silu(x = input_33_cast_fp16)[name = tensor("input_35_cast_fp16")]; - tensor var_425_pad_type_0 = const()[name = tensor("op_425_pad_type_0"), val = tensor("valid")]; - tensor var_425_strides_0 = const()[name = tensor("op_425_strides_0"), val = tensor([1, 1])]; - tensor var_425_pad_0 = const()[name = tensor("op_425_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_425_dilations_0 = const()[name = tensor("op_425_dilations_0"), val = tensor([1, 1])]; - tensor var_425_groups_0 = const()[name = tensor("op_425_groups_0"), val = tensor(1)]; - tensor layers_0_conv_pointwise_conv2_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(5477312))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(5673984))), name = tensor("layers_0_conv_pointwise_conv2_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_0_conv_pointwise_conv2_inlier_module_bias_to_fp16 = const()[name = tensor("layers_0_conv_pointwise_conv2_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(5674176)))]; - tensor var_425_cast_fp16 = conv(bias = layers_0_conv_pointwise_conv2_inlier_module_bias_to_fp16, dilations = var_425_dilations_0, groups = var_425_groups_0, pad = var_425_pad_0, pad_type = var_425_pad_type_0, strides = var_425_strides_0, weight = layers_0_conv_pointwise_conv2_inlier_module_weight_to_fp16_palettized, x = input_35_cast_fp16)[name = tensor("op_425_cast_fp16")]; - tensor var_431_pad_type_0 = const()[name = tensor("op_431_pad_type_0"), val = tensor("valid")]; - tensor var_431_strides_0 = const()[name = tensor("op_431_strides_0"), val = tensor([1, 1])]; - tensor var_431_pad_0 = const()[name = tensor("op_431_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_431_dilations_0 = const()[name = tensor("op_431_dilations_0"), val = tensor([1, 1])]; - tensor var_431_groups_0 = const()[name = tensor("op_431_groups_0"), val = tensor(1)]; - tensor layers_0_conv_pointwise_conv2_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(5683840))), name = tensor("layers_0_conv_pointwise_conv2_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(5675264))), shape = tensor([512, 512, 1, 1])]; - tensor var_431_cast_fp16 = conv(dilations = var_431_dilations_0, groups = var_431_groups_0, pad = var_431_pad_0, pad_type = var_431_pad_type_0, strides = var_431_strides_0, weight = layers_0_conv_pointwise_conv2_outlier_module_weight_to_fp16_sparsified, x = input_35_cast_fp16)[name = tensor("op_431_cast_fp16")]; - tensor x_5_cast_fp16 = add(x = var_425_cast_fp16, y = var_431_cast_fp16)[name = tensor("x_5_cast_fp16")]; - tensor inputs_7_cast_fp16 = add(x = inputs_5_cast_fp16, y = x_5_cast_fp16)[name = tensor("inputs_7_cast_fp16")]; - tensor out_7_axes_0 = const()[name = tensor("out_7_axes_0"), val = tensor([1])]; - tensor var_442_to_fp16 = const()[name = tensor("op_442_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_7_cast_fp16 = layer_norm(axes = out_7_axes_0, epsilon = var_442_to_fp16, x = inputs_7_cast_fp16)[name = tensor("out_7_cast_fp16")]; - tensor input_37_gamma_0_to_fp16 = const()[name = tensor("input_37_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(5716672)))]; - tensor input_37_beta_0_to_fp16 = const()[name = tensor("input_37_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(5717760)))]; - tensor input_37_epsilon_0_to_fp16 = const()[name = tensor("input_37_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_37_cast_fp16 = batch_norm(beta = input_37_beta_0_to_fp16, epsilon = input_37_epsilon_0_to_fp16, gamma = input_37_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_7_cast_fp16)[name = tensor("input_37_cast_fp16")]; - tensor var_462_pad_type_0 = const()[name = tensor("op_462_pad_type_0"), val = tensor("valid")]; - tensor var_462_strides_0 = const()[name = tensor("op_462_strides_0"), val = tensor([1, 1])]; - tensor var_462_pad_0 = const()[name = tensor("op_462_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_462_dilations_0 = const()[name = tensor("op_462_dilations_0"), val = tensor([1, 1])]; - tensor var_462_groups_0 = const()[name = tensor("op_462_groups_0"), val = tensor(1)]; - tensor layers_0_feed_forward2_fc1_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(5718848))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(6505344))), name = tensor("layers_0_feed_forward2_fc1_inlier_module_weight_to_fp16_palettized"), shape = tensor([2048, 512, 1, 1])]; - tensor layers_0_feed_forward2_fc1_inlier_module_bias_to_fp16 = const()[name = tensor("layers_0_feed_forward2_fc1_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(6505536)))]; - tensor var_462_cast_fp16 = conv(bias = layers_0_feed_forward2_fc1_inlier_module_bias_to_fp16, dilations = var_462_dilations_0, groups = var_462_groups_0, pad = var_462_pad_0, pad_type = var_462_pad_type_0, strides = var_462_strides_0, weight = layers_0_feed_forward2_fc1_inlier_module_weight_to_fp16_palettized, x = input_37_cast_fp16)[name = tensor("op_462_cast_fp16")]; - tensor var_468_pad_type_0 = const()[name = tensor("op_468_pad_type_0"), val = tensor("valid")]; - tensor var_468_strides_0 = const()[name = tensor("op_468_strides_0"), val = tensor([1, 1])]; - tensor var_468_pad_0 = const()[name = tensor("op_468_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_468_dilations_0 = const()[name = tensor("op_468_dilations_0"), val = tensor([1, 1])]; - tensor var_468_groups_0 = const()[name = tensor("op_468_groups_0"), val = tensor(1)]; - tensor layers_0_feed_forward2_fc1_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(6541376))), name = tensor("layers_0_feed_forward2_fc1_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(6509696))), shape = tensor([2048, 512, 1, 1])]; - tensor var_468_cast_fp16 = conv(dilations = var_468_dilations_0, groups = var_468_groups_0, pad = var_468_pad_0, pad_type = var_468_pad_type_0, strides = var_468_strides_0, weight = layers_0_feed_forward2_fc1_outlier_module_weight_to_fp16_sparsified, x = input_37_cast_fp16)[name = tensor("op_468_cast_fp16")]; - tensor input_39_cast_fp16 = add(x = var_462_cast_fp16, y = var_468_cast_fp16)[name = tensor("input_39_cast_fp16")]; - tensor input_41_cast_fp16 = silu(x = input_39_cast_fp16)[name = tensor("input_41_cast_fp16")]; - tensor var_479_pad_type_0 = const()[name = tensor("op_479_pad_type_0"), val = tensor("valid")]; - tensor var_479_strides_0 = const()[name = tensor("op_479_strides_0"), val = tensor([1, 1])]; - tensor var_479_pad_0 = const()[name = tensor("op_479_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_479_dilations_0 = const()[name = tensor("op_479_dilations_0"), val = tensor([1, 1])]; - tensor var_479_groups_0 = const()[name = tensor("op_479_groups_0"), val = tensor(1)]; - tensor layers_0_feed_forward2_fc2_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(6672512))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7459008))), name = tensor("layers_0_feed_forward2_fc2_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 2048, 1, 1])]; - tensor layers_0_feed_forward2_fc2_inlier_module_bias_to_fp16 = const()[name = tensor("layers_0_feed_forward2_fc2_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7459200)))]; - tensor var_479_cast_fp16 = conv(bias = layers_0_feed_forward2_fc2_inlier_module_bias_to_fp16, dilations = var_479_dilations_0, groups = var_479_groups_0, pad = var_479_pad_0, pad_type = var_479_pad_type_0, strides = var_479_strides_0, weight = layers_0_feed_forward2_fc2_inlier_module_weight_to_fp16_palettized, x = input_41_cast_fp16)[name = tensor("op_479_cast_fp16")]; - tensor var_485_pad_type_0 = const()[name = tensor("op_485_pad_type_0"), val = tensor("valid")]; - tensor var_485_strides_0 = const()[name = tensor("op_485_strides_0"), val = tensor([1, 1])]; - tensor var_485_pad_0 = const()[name = tensor("op_485_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_485_dilations_0 = const()[name = tensor("op_485_dilations_0"), val = tensor([1, 1])]; - tensor var_485_groups_0 = const()[name = tensor("op_485_groups_0"), val = tensor(1)]; - tensor layers_0_feed_forward2_fc2_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7490688))), name = tensor("layers_0_feed_forward2_fc2_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7460288))), shape = tensor([512, 2048, 1, 1])]; - tensor var_485_cast_fp16 = conv(dilations = var_485_dilations_0, groups = var_485_groups_0, pad = var_485_pad_0, pad_type = var_485_pad_type_0, strides = var_485_strides_0, weight = layers_0_feed_forward2_fc2_outlier_module_weight_to_fp16_sparsified, x = input_41_cast_fp16)[name = tensor("op_485_cast_fp16")]; - tensor x_7_cast_fp16 = add(x = var_479_cast_fp16, y = var_485_cast_fp16)[name = tensor("x_7_cast_fp16")]; - tensor var_487_to_fp16 = const()[name = tensor("op_487_to_fp16"), val = tensor(0x1p-1)]; - tensor var_488_cast_fp16 = mul(x = x_7_cast_fp16, y = var_487_to_fp16)[name = tensor("op_488_cast_fp16")]; - tensor inputs_9_cast_fp16 = add(x = inputs_7_cast_fp16, y = var_488_cast_fp16)[name = tensor("inputs_9_cast_fp16")]; - tensor out_9_axes_0 = const()[name = tensor("out_9_axes_0"), val = tensor([1])]; - tensor var_498_to_fp16 = const()[name = tensor("op_498_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_9_cast_fp16 = layer_norm(axes = out_9_axes_0, epsilon = var_498_to_fp16, x = inputs_9_cast_fp16)[name = tensor("out_9_cast_fp16")]; - tensor inputs_11_gamma_0_to_fp16 = const()[name = tensor("inputs_11_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7621824)))]; - tensor inputs_11_beta_0_to_fp16 = const()[name = tensor("inputs_11_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7622912)))]; - tensor inputs_11_epsilon_0_to_fp16 = const()[name = tensor("inputs_11_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor inputs_11_cast_fp16 = batch_norm(beta = inputs_11_beta_0_to_fp16, epsilon = inputs_11_epsilon_0_to_fp16, gamma = inputs_11_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_9_cast_fp16)[name = tensor("inputs_11_cast_fp16")]; - tensor var_512 = const()[name = tensor("op_512"), val = tensor(3)]; - tensor out_11_axes_0 = const()[name = tensor("out_11_axes_0"), val = tensor([1])]; - tensor var_543_to_fp16 = const()[name = tensor("op_543_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_11_cast_fp16 = layer_norm(axes = out_11_axes_0, epsilon = var_543_to_fp16, x = inputs_11_cast_fp16)[name = tensor("out_11_cast_fp16")]; - tensor input_43_gamma_0_to_fp16 = const()[name = tensor("input_43_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7624000)))]; - tensor input_43_beta_0_to_fp16 = const()[name = tensor("input_43_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7625088)))]; - tensor input_43_epsilon_0_to_fp16 = const()[name = tensor("input_43_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_43_cast_fp16 = batch_norm(beta = input_43_beta_0_to_fp16, epsilon = input_43_epsilon_0_to_fp16, gamma = input_43_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_11_cast_fp16)[name = tensor("input_43_cast_fp16")]; - tensor var_563_pad_type_0 = const()[name = tensor("op_563_pad_type_0"), val = tensor("valid")]; - tensor var_563_strides_0 = const()[name = tensor("op_563_strides_0"), val = tensor([1, 1])]; - tensor var_563_pad_0 = const()[name = tensor("op_563_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_563_dilations_0 = const()[name = tensor("op_563_dilations_0"), val = tensor([1, 1])]; - tensor var_563_groups_0 = const()[name = tensor("op_563_groups_0"), val = tensor(1)]; - tensor layers_1_feed_forward1_fc1_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7626176))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8412672))), name = tensor("layers_1_feed_forward1_fc1_inlier_module_weight_to_fp16_palettized"), shape = tensor([2048, 512, 1, 1])]; - tensor layers_1_feed_forward1_fc1_inlier_module_bias_to_fp16 = const()[name = tensor("layers_1_feed_forward1_fc1_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8412864)))]; - tensor var_563_cast_fp16 = conv(bias = layers_1_feed_forward1_fc1_inlier_module_bias_to_fp16, dilations = var_563_dilations_0, groups = var_563_groups_0, pad = var_563_pad_0, pad_type = var_563_pad_type_0, strides = var_563_strides_0, weight = layers_1_feed_forward1_fc1_inlier_module_weight_to_fp16_palettized, x = input_43_cast_fp16)[name = tensor("op_563_cast_fp16")]; - tensor var_569_pad_type_0 = const()[name = tensor("op_569_pad_type_0"), val = tensor("valid")]; - tensor var_569_strides_0 = const()[name = tensor("op_569_strides_0"), val = tensor([1, 1])]; - tensor var_569_pad_0 = const()[name = tensor("op_569_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_569_dilations_0 = const()[name = tensor("op_569_dilations_0"), val = tensor([1, 1])]; - tensor var_569_groups_0 = const()[name = tensor("op_569_groups_0"), val = tensor(1)]; - tensor layers_1_feed_forward1_fc1_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8451136))), name = tensor("layers_1_feed_forward1_fc1_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8417024))), shape = tensor([2048, 512, 1, 1])]; - tensor var_569_cast_fp16 = conv(dilations = var_569_dilations_0, groups = var_569_groups_0, pad = var_569_pad_0, pad_type = var_569_pad_type_0, strides = var_569_strides_0, weight = layers_1_feed_forward1_fc1_outlier_module_weight_to_fp16_sparsified, x = input_43_cast_fp16)[name = tensor("op_569_cast_fp16")]; - tensor input_45_cast_fp16 = add(x = var_563_cast_fp16, y = var_569_cast_fp16)[name = tensor("input_45_cast_fp16")]; - tensor input_47_cast_fp16 = silu(x = input_45_cast_fp16)[name = tensor("input_47_cast_fp16")]; - tensor var_580_pad_type_0 = const()[name = tensor("op_580_pad_type_0"), val = tensor("valid")]; - tensor var_580_strides_0 = const()[name = tensor("op_580_strides_0"), val = tensor([1, 1])]; - tensor var_580_pad_0 = const()[name = tensor("op_580_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_580_dilations_0 = const()[name = tensor("op_580_dilations_0"), val = tensor([1, 1])]; - tensor var_580_groups_0 = const()[name = tensor("op_580_groups_0"), val = tensor(1)]; - tensor layers_1_feed_forward1_fc2_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8582272))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9368768))), name = tensor("layers_1_feed_forward1_fc2_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 2048, 1, 1])]; - tensor layers_1_feed_forward1_fc2_inlier_module_bias_to_fp16 = const()[name = tensor("layers_1_feed_forward1_fc2_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9368960)))]; - tensor var_580_cast_fp16 = conv(bias = layers_1_feed_forward1_fc2_inlier_module_bias_to_fp16, dilations = var_580_dilations_0, groups = var_580_groups_0, pad = var_580_pad_0, pad_type = var_580_pad_type_0, strides = var_580_strides_0, weight = layers_1_feed_forward1_fc2_inlier_module_weight_to_fp16_palettized, x = input_47_cast_fp16)[name = tensor("op_580_cast_fp16")]; - tensor var_586_pad_type_0 = const()[name = tensor("op_586_pad_type_0"), val = tensor("valid")]; - tensor var_586_strides_0 = const()[name = tensor("op_586_strides_0"), val = tensor([1, 1])]; - tensor var_586_pad_0 = const()[name = tensor("op_586_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_586_dilations_0 = const()[name = tensor("op_586_dilations_0"), val = tensor([1, 1])]; - tensor var_586_groups_0 = const()[name = tensor("op_586_groups_0"), val = tensor(1)]; - tensor layers_1_feed_forward1_fc2_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9406336))), name = tensor("layers_1_feed_forward1_fc2_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9370048))), shape = tensor([512, 2048, 1, 1])]; - tensor var_586_cast_fp16 = conv(dilations = var_586_dilations_0, groups = var_586_groups_0, pad = var_586_pad_0, pad_type = var_586_pad_type_0, strides = var_586_strides_0, weight = layers_1_feed_forward1_fc2_outlier_module_weight_to_fp16_sparsified, x = input_47_cast_fp16)[name = tensor("op_586_cast_fp16")]; - tensor x_9_cast_fp16 = add(x = var_580_cast_fp16, y = var_586_cast_fp16)[name = tensor("x_9_cast_fp16")]; - tensor var_588_to_fp16 = const()[name = tensor("op_588_to_fp16"), val = tensor(0x1p-1)]; - tensor var_589_cast_fp16 = mul(x = x_9_cast_fp16, y = var_588_to_fp16)[name = tensor("op_589_cast_fp16")]; - tensor inputs_13_cast_fp16 = add(x = inputs_11_cast_fp16, y = var_589_cast_fp16)[name = tensor("inputs_13_cast_fp16")]; - tensor out_13_axes_0 = const()[name = tensor("out_13_axes_0"), val = tensor([1])]; - tensor var_599_to_fp16 = const()[name = tensor("op_599_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_13_cast_fp16 = layer_norm(axes = out_13_axes_0, epsilon = var_599_to_fp16, x = inputs_13_cast_fp16)[name = tensor("out_13_cast_fp16")]; - tensor obj_7_gamma_0_to_fp16 = const()[name = tensor("obj_7_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9537472)))]; - tensor obj_7_beta_0_to_fp16 = const()[name = tensor("obj_7_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9538560)))]; - tensor obj_7_epsilon_0_to_fp16 = const()[name = tensor("obj_7_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor obj_7_cast_fp16 = batch_norm(beta = obj_7_beta_0_to_fp16, epsilon = obj_7_epsilon_0_to_fp16, gamma = obj_7_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_13_cast_fp16)[name = tensor("obj_7_cast_fp16")]; - tensor var_624_pad_type_0 = const()[name = tensor("op_624_pad_type_0"), val = tensor("valid")]; - tensor var_624_strides_0 = const()[name = tensor("op_624_strides_0"), val = tensor([1, 1])]; - tensor var_624_pad_0 = const()[name = tensor("op_624_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_624_dilations_0 = const()[name = tensor("op_624_dilations_0"), val = tensor([1, 1])]; - tensor var_624_groups_0 = const()[name = tensor("op_624_groups_0"), val = tensor(1)]; - tensor layers_1_self_attn_q_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9539648))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9736320))), name = tensor("layers_1_self_attn_q_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_1_self_attn_q_proj_inlier_module_bias_to_fp16 = const()[name = tensor("layers_1_self_attn_q_proj_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9736512)))]; - tensor var_624_cast_fp16 = conv(bias = layers_1_self_attn_q_proj_inlier_module_bias_to_fp16, dilations = var_624_dilations_0, groups = var_624_groups_0, pad = var_624_pad_0, pad_type = var_624_pad_type_0, strides = var_624_strides_0, weight = layers_1_self_attn_q_proj_inlier_module_weight_to_fp16_palettized, x = obj_7_cast_fp16)[name = tensor("op_624_cast_fp16")]; - tensor var_630_pad_type_0 = const()[name = tensor("op_630_pad_type_0"), val = tensor("valid")]; - tensor var_630_strides_0 = const()[name = tensor("op_630_strides_0"), val = tensor([1, 1])]; - tensor var_630_pad_0 = const()[name = tensor("op_630_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_630_dilations_0 = const()[name = tensor("op_630_dilations_0"), val = tensor([1, 1])]; - tensor var_630_groups_0 = const()[name = tensor("op_630_groups_0"), val = tensor(1)]; - tensor layers_1_self_attn_q_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9746432))), name = tensor("layers_1_self_attn_q_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9737600))), shape = tensor([512, 512, 1, 1])]; - tensor var_630_cast_fp16 = conv(dilations = var_630_dilations_0, groups = var_630_groups_0, pad = var_630_pad_0, pad_type = var_630_pad_type_0, strides = var_630_strides_0, weight = layers_1_self_attn_q_proj_outlier_module_weight_to_fp16_sparsified, x = obj_7_cast_fp16)[name = tensor("op_630_cast_fp16")]; - tensor query_5_cast_fp16 = add(x = var_624_cast_fp16, y = var_630_cast_fp16)[name = tensor("query_5_cast_fp16")]; - tensor var_639_pad_type_0 = const()[name = tensor("op_639_pad_type_0"), val = tensor("valid")]; - tensor var_639_strides_0 = const()[name = tensor("op_639_strides_0"), val = tensor([1, 1])]; - tensor var_639_pad_0 = const()[name = tensor("op_639_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_639_dilations_0 = const()[name = tensor("op_639_dilations_0"), val = tensor([1, 1])]; - tensor var_639_groups_0 = const()[name = tensor("op_639_groups_0"), val = tensor(1)]; - tensor layers_1_self_attn_k_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9779264))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9975936))), name = tensor("layers_1_self_attn_k_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor var_639_cast_fp16 = conv(dilations = var_639_dilations_0, groups = var_639_groups_0, pad = var_639_pad_0, pad_type = var_639_pad_type_0, strides = var_639_strides_0, weight = layers_1_self_attn_k_proj_inlier_module_weight_to_fp16_palettized, x = obj_7_cast_fp16)[name = tensor("op_639_cast_fp16")]; - tensor var_645_pad_type_0 = const()[name = tensor("op_645_pad_type_0"), val = tensor("valid")]; - tensor var_645_strides_0 = const()[name = tensor("op_645_strides_0"), val = tensor([1, 1])]; - tensor var_645_pad_0 = const()[name = tensor("op_645_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_645_dilations_0 = const()[name = tensor("op_645_dilations_0"), val = tensor([1, 1])]; - tensor var_645_groups_0 = const()[name = tensor("op_645_groups_0"), val = tensor(1)]; - tensor layers_1_self_attn_k_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9985344))), name = tensor("layers_1_self_attn_k_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9976128))), shape = tensor([512, 512, 1, 1])]; - tensor var_645_cast_fp16 = conv(dilations = var_645_dilations_0, groups = var_645_groups_0, pad = var_645_pad_0, pad_type = var_645_pad_type_0, strides = var_645_strides_0, weight = layers_1_self_attn_k_proj_outlier_module_weight_to_fp16_sparsified, x = obj_7_cast_fp16)[name = tensor("op_645_cast_fp16")]; - tensor key_3_cast_fp16 = add(x = var_639_cast_fp16, y = var_645_cast_fp16)[name = tensor("key_3_cast_fp16")]; - tensor var_655_pad_type_0 = const()[name = tensor("op_655_pad_type_0"), val = tensor("valid")]; - tensor var_655_strides_0 = const()[name = tensor("op_655_strides_0"), val = tensor([1, 1])]; - tensor var_655_pad_0 = const()[name = tensor("op_655_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_655_dilations_0 = const()[name = tensor("op_655_dilations_0"), val = tensor([1, 1])]; - tensor var_655_groups_0 = const()[name = tensor("op_655_groups_0"), val = tensor(1)]; - tensor layers_1_self_attn_v_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10018176))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10214848))), name = tensor("layers_1_self_attn_v_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_1_self_attn_v_proj_inlier_module_bias_to_fp16 = const()[name = tensor("layers_1_self_attn_v_proj_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10215040)))]; - tensor var_655_cast_fp16 = conv(bias = layers_1_self_attn_v_proj_inlier_module_bias_to_fp16, dilations = var_655_dilations_0, groups = var_655_groups_0, pad = var_655_pad_0, pad_type = var_655_pad_type_0, strides = var_655_strides_0, weight = layers_1_self_attn_v_proj_inlier_module_weight_to_fp16_palettized, x = obj_7_cast_fp16)[name = tensor("op_655_cast_fp16")]; - tensor var_661_pad_type_0 = const()[name = tensor("op_661_pad_type_0"), val = tensor("valid")]; - tensor var_661_strides_0 = const()[name = tensor("op_661_strides_0"), val = tensor([1, 1])]; - tensor var_661_pad_0 = const()[name = tensor("op_661_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_661_dilations_0 = const()[name = tensor("op_661_dilations_0"), val = tensor([1, 1])]; - tensor var_661_groups_0 = const()[name = tensor("op_661_groups_0"), val = tensor(1)]; - tensor layers_1_self_attn_v_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10223872))), name = tensor("layers_1_self_attn_v_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10216128))), shape = tensor([512, 512, 1, 1])]; - tensor var_661_cast_fp16 = conv(dilations = var_661_dilations_0, groups = var_661_groups_0, pad = var_661_pad_0, pad_type = var_661_pad_type_0, strides = var_661_strides_0, weight = layers_1_self_attn_v_proj_outlier_module_weight_to_fp16_sparsified, x = obj_7_cast_fp16)[name = tensor("op_661_cast_fp16")]; - tensor value_3_cast_fp16 = add(x = var_655_cast_fp16, y = var_661_cast_fp16)[name = tensor("value_3_cast_fp16")]; - tensor var_664_to_fp16 = const()[name = tensor("op_664_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10256704)))]; - tensor query_7_cast_fp16 = add(x = query_5_cast_fp16, y = var_664_to_fp16)[name = tensor("query_7_cast_fp16")]; - tensor var_667_to_fp16 = const()[name = tensor("op_667_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10257792)))]; - tensor q_with_bias_v_3_cast_fp16 = add(x = query_5_cast_fp16, y = var_667_to_fp16)[name = tensor("q_with_bias_v_3_cast_fp16")]; - tensor var_677_pad_type_0 = const()[name = tensor("op_677_pad_type_0"), val = tensor("valid")]; - tensor var_677_strides_0 = const()[name = tensor("op_677_strides_0"), val = tensor([1, 1])]; - tensor var_677_pad_0 = const()[name = tensor("op_677_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_677_dilations_0 = const()[name = tensor("op_677_dilations_0"), val = tensor([1, 1])]; - tensor var_677_groups_0 = const()[name = tensor("op_677_groups_0"), val = tensor(1)]; - tensor layers_1_self_attn_linear_pos_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10258880))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10455552))), name = tensor("layers_1_self_attn_linear_pos_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor var_677_cast_fp16 = conv(bias = input_17_mean_0_to_fp16, dilations = var_677_dilations_0, groups = var_677_groups_0, pad = var_677_pad_0, pad_type = var_677_pad_type_0, strides = var_677_strides_0, weight = layers_1_self_attn_linear_pos_inlier_module_weight_to_fp16_palettized, x = obj_3_cast_fp16)[name = tensor("op_677_cast_fp16")]; - tensor var_683_pad_type_0 = const()[name = tensor("op_683_pad_type_0"), val = tensor("valid")]; - tensor var_683_strides_0 = const()[name = tensor("op_683_strides_0"), val = tensor([1, 1])]; - tensor var_683_pad_0 = const()[name = tensor("op_683_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_683_dilations_0 = const()[name = tensor("op_683_dilations_0"), val = tensor([1, 1])]; - tensor var_683_groups_0 = const()[name = tensor("op_683_groups_0"), val = tensor(1)]; - tensor layers_1_self_attn_linear_pos_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10477632))), name = tensor("layers_1_self_attn_linear_pos_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10455744))), shape = tensor([512, 512, 1, 1])]; - tensor var_683_cast_fp16 = conv(dilations = var_683_dilations_0, groups = var_683_groups_0, pad = var_683_pad_0, pad_type = var_683_pad_type_0, strides = var_683_strides_0, weight = layers_1_self_attn_linear_pos_outlier_module_weight_to_fp16_sparsified, x = obj_3_cast_fp16)[name = tensor("op_683_cast_fp16")]; - tensor p_3_cast_fp16 = add(x = var_677_cast_fp16, y = var_683_cast_fp16)[name = tensor("p_3_cast_fp16")]; - tensor var_687 = const()[name = tensor("op_687"), val = tensor([1, 8, 64, 188])]; - tensor var_688_cast_fp16 = reshape(shape = var_687, x = q_with_bias_v_3_cast_fp16)[name = tensor("op_688_cast_fp16")]; - tensor var_689 = const()[name = tensor("op_689"), val = tensor([1, 8, 64, -1])]; - tensor var_690_cast_fp16 = reshape(shape = var_689, x = p_3_cast_fp16)[name = tensor("op_690_cast_fp16")]; - tensor matrix_bd_9_transpose_x_0 = const()[name = tensor("matrix_bd_9_transpose_x_0"), val = tensor(true)]; - tensor matrix_bd_9_transpose_y_0 = const()[name = tensor("matrix_bd_9_transpose_y_0"), val = tensor(false)]; - tensor matrix_bd_9_cast_fp16 = matmul(transpose_x = matrix_bd_9_transpose_x_0, transpose_y = matrix_bd_9_transpose_y_0, x = var_688_cast_fp16, y = var_690_cast_fp16)[name = tensor("matrix_bd_9_cast_fp16")]; - tensor matrix_bd_11_pad_0 = const()[name = tensor("matrix_bd_11_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor matrix_bd_11_mode_0 = const()[name = tensor("matrix_bd_11_mode_0"), val = tensor("constant")]; - tensor const_21_to_fp16 = const()[name = tensor("const_21_to_fp16"), val = tensor(0x0p+0)]; - tensor matrix_bd_11_cast_fp16 = pad(constant_val = const_21_to_fp16, mode = matrix_bd_11_mode_0, pad = matrix_bd_11_pad_0, x = matrix_bd_9_cast_fp16)[name = tensor("matrix_bd_11_cast_fp16")]; - tensor var_699 = const()[name = tensor("op_699"), val = tensor([1, 8, -1, 188])]; - tensor matrix_bd_13_cast_fp16 = reshape(shape = var_699, x = matrix_bd_11_cast_fp16)[name = tensor("matrix_bd_13_cast_fp16")]; - tensor var_703_begin_0 = const()[name = tensor("op_703_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_703_end_0 = const()[name = tensor("op_703_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_703_end_mask_0 = const()[name = tensor("op_703_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_703_cast_fp16 = slice_by_index(begin = var_703_begin_0, end = var_703_end_0, end_mask = var_703_end_mask_0, x = matrix_bd_13_cast_fp16)[name = tensor("op_703_cast_fp16")]; - tensor var_704 = const()[name = tensor("op_704"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_15_cast_fp16 = reshape(shape = var_704, x = var_703_cast_fp16)[name = tensor("matrix_bd_15_cast_fp16")]; - tensor var_709_begin_0 = const()[name = tensor("op_709_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor var_709_end_0 = const()[name = tensor("op_709_end_0"), val = tensor([1, 8, 188, 188])]; - tensor var_709_end_mask_0 = const()[name = tensor("op_709_end_mask_0"), val = tensor([true, true, true, false])]; - tensor var_709_cast_fp16 = slice_by_index(begin = var_709_begin_0, end = var_709_end_0, end_mask = var_709_end_mask_0, x = matrix_bd_15_cast_fp16)[name = tensor("op_709_cast_fp16")]; - tensor var_710_to_fp16 = const()[name = tensor("op_710_to_fp16"), val = tensor(0x1p-3)]; - tensor qk_mask_3_cast_fp16 = mul(x = var_709_cast_fp16, y = var_710_to_fp16)[name = tensor("qk_mask_3_cast_fp16")]; - tensor var_714 = const()[name = tensor("op_714"), val = tensor([1, 8, 64, 188])]; - tensor mh_q_3_cast_fp16 = reshape(shape = var_714, x = query_7_cast_fp16)[name = tensor("mh_q_3_cast_fp16")]; - tensor var_716_to_fp16 = const()[name = tensor("op_716_to_fp16"), val = tensor(0x1p-3)]; - tensor var_717_cast_fp16 = mul(x = mh_q_3_cast_fp16, y = var_716_to_fp16)[name = tensor("op_717_cast_fp16")]; - tensor var_720 = const()[name = tensor("op_720"), val = tensor([1, 8, 64, 188])]; - tensor var_721_cast_fp16 = reshape(shape = var_720, x = key_3_cast_fp16)[name = tensor("op_721_cast_fp16")]; - tensor mh_w_5_transpose_x_0 = const()[name = tensor("mh_w_5_transpose_x_0"), val = tensor(true)]; - tensor mh_w_5_transpose_y_0 = const()[name = tensor("mh_w_5_transpose_y_0"), val = tensor(false)]; - tensor mh_w_5_cast_fp16 = matmul(transpose_x = mh_w_5_transpose_x_0, transpose_y = mh_w_5_transpose_y_0, x = var_717_cast_fp16, y = var_721_cast_fp16)[name = tensor("mh_w_5_cast_fp16")]; - tensor mh_w_7_cast_fp16 = add(x = mh_w_5_cast_fp16, y = qk_mask_3_cast_fp16)[name = tensor("mh_w_7_cast_fp16")]; - tensor var_725_cast_fp16 = softmax(axis = var_512, x = mh_w_7_cast_fp16)[name = tensor("op_725_cast_fp16")]; - tensor var_726 = const()[name = tensor("op_726"), val = tensor([1, 8, 64, 188])]; - tensor var_727_cast_fp16 = reshape(shape = var_726, x = value_3_cast_fp16)[name = tensor("op_727_cast_fp16")]; - tensor attn_3_transpose_x_0 = const()[name = tensor("attn_3_transpose_x_0"), val = tensor(false)]; - tensor attn_3_transpose_y_0 = const()[name = tensor("attn_3_transpose_y_0"), val = tensor(true)]; - tensor attn_3_cast_fp16 = matmul(transpose_x = attn_3_transpose_x_0, transpose_y = attn_3_transpose_y_0, x = var_727_cast_fp16, y = var_725_cast_fp16)[name = tensor("attn_3_cast_fp16")]; - tensor var_730 = const()[name = tensor("op_730"), val = tensor([1, 512, 1, 188])]; - tensor input_49_cast_fp16 = reshape(shape = var_730, x = attn_3_cast_fp16)[name = tensor("input_49_cast_fp16")]; - tensor var_740_pad_type_0 = const()[name = tensor("op_740_pad_type_0"), val = tensor("valid")]; - tensor var_740_strides_0 = const()[name = tensor("op_740_strides_0"), val = tensor([1, 1])]; - tensor var_740_pad_0 = const()[name = tensor("op_740_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_740_dilations_0 = const()[name = tensor("op_740_dilations_0"), val = tensor([1, 1])]; - tensor var_740_groups_0 = const()[name = tensor("op_740_groups_0"), val = tensor(1)]; - tensor layers_1_self_attn_o_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10510464))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10707136))), name = tensor("layers_1_self_attn_o_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_1_self_attn_o_proj_inlier_module_bias_to_fp16 = const()[name = tensor("layers_1_self_attn_o_proj_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10707328)))]; - tensor var_740_cast_fp16 = conv(bias = layers_1_self_attn_o_proj_inlier_module_bias_to_fp16, dilations = var_740_dilations_0, groups = var_740_groups_0, pad = var_740_pad_0, pad_type = var_740_pad_type_0, strides = var_740_strides_0, weight = layers_1_self_attn_o_proj_inlier_module_weight_to_fp16_palettized, x = input_49_cast_fp16)[name = tensor("op_740_cast_fp16")]; - tensor var_746_pad_type_0 = const()[name = tensor("op_746_pad_type_0"), val = tensor("valid")]; - tensor var_746_strides_0 = const()[name = tensor("op_746_strides_0"), val = tensor([1, 1])]; - tensor var_746_pad_0 = const()[name = tensor("op_746_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_746_dilations_0 = const()[name = tensor("op_746_dilations_0"), val = tensor([1, 1])]; - tensor var_746_groups_0 = const()[name = tensor("op_746_groups_0"), val = tensor(1)]; - tensor layers_1_self_attn_o_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10716288))), name = tensor("layers_1_self_attn_o_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10708416))), shape = tensor([512, 512, 1, 1])]; - tensor var_746_cast_fp16 = conv(dilations = var_746_dilations_0, groups = var_746_groups_0, pad = var_746_pad_0, pad_type = var_746_pad_type_0, strides = var_746_strides_0, weight = layers_1_self_attn_o_proj_outlier_module_weight_to_fp16_sparsified, x = input_49_cast_fp16)[name = tensor("op_746_cast_fp16")]; - tensor obj_9_cast_fp16 = add(x = var_740_cast_fp16, y = var_746_cast_fp16)[name = tensor("obj_9_cast_fp16")]; - tensor inputs_15_cast_fp16 = add(x = inputs_13_cast_fp16, y = obj_9_cast_fp16)[name = tensor("inputs_15_cast_fp16")]; - tensor out_15_axes_0 = const()[name = tensor("out_15_axes_0"), val = tensor([1])]; - tensor var_757_to_fp16 = const()[name = tensor("op_757_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_15_cast_fp16 = layer_norm(axes = out_15_axes_0, epsilon = var_757_to_fp16, x = inputs_15_cast_fp16)[name = tensor("out_15_cast_fp16")]; - tensor input_51_gamma_0_to_fp16 = const()[name = tensor("input_51_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10749120)))]; - tensor input_51_beta_0_to_fp16 = const()[name = tensor("input_51_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10750208)))]; - tensor input_51_epsilon_0_to_fp16 = const()[name = tensor("input_51_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_51_cast_fp16 = batch_norm(beta = input_51_beta_0_to_fp16, epsilon = input_51_epsilon_0_to_fp16, gamma = input_51_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_15_cast_fp16)[name = tensor("input_51_cast_fp16")]; - tensor var_779_pad_type_0 = const()[name = tensor("op_779_pad_type_0"), val = tensor("valid")]; - tensor var_779_strides_0 = const()[name = tensor("op_779_strides_0"), val = tensor([1, 1])]; - tensor var_779_pad_0 = const()[name = tensor("op_779_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_779_dilations_0 = const()[name = tensor("op_779_dilations_0"), val = tensor([1, 1])]; - tensor var_779_groups_0 = const()[name = tensor("op_779_groups_0"), val = tensor(1)]; - tensor layers_1_conv_pointwise_conv1_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10751296))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11144576))), name = tensor("layers_1_conv_pointwise_conv1_inlier_module_weight_to_fp16_palettized"), shape = tensor([1024, 512, 1, 1])]; - tensor layers_1_conv_pointwise_conv1_inlier_module_bias_to_fp16 = const()[name = tensor("layers_1_conv_pointwise_conv1_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11144768)))]; - tensor var_779_cast_fp16 = conv(bias = layers_1_conv_pointwise_conv1_inlier_module_bias_to_fp16, dilations = var_779_dilations_0, groups = var_779_groups_0, pad = var_779_pad_0, pad_type = var_779_pad_type_0, strides = var_779_strides_0, weight = layers_1_conv_pointwise_conv1_inlier_module_weight_to_fp16_palettized, x = input_51_cast_fp16)[name = tensor("op_779_cast_fp16")]; - tensor var_785_pad_type_0 = const()[name = tensor("op_785_pad_type_0"), val = tensor("valid")]; - tensor var_785_strides_0 = const()[name = tensor("op_785_strides_0"), val = tensor([1, 1])]; - tensor var_785_pad_0 = const()[name = tensor("op_785_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_785_dilations_0 = const()[name = tensor("op_785_dilations_0"), val = tensor([1, 1])]; - tensor var_785_groups_0 = const()[name = tensor("op_785_groups_0"), val = tensor(1)]; - tensor layers_1_conv_pointwise_conv1_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11165248))), name = tensor("layers_1_conv_pointwise_conv1_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11146880))), shape = tensor([1024, 512, 1, 1])]; - tensor var_785_cast_fp16 = conv(dilations = var_785_dilations_0, groups = var_785_groups_0, pad = var_785_pad_0, pad_type = var_785_pad_type_0, strides = var_785_strides_0, weight = layers_1_conv_pointwise_conv1_outlier_module_weight_to_fp16_sparsified, x = input_51_cast_fp16)[name = tensor("op_785_cast_fp16")]; - tensor input_53_cast_fp16 = add(x = var_779_cast_fp16, y = var_785_cast_fp16)[name = tensor("input_53_cast_fp16")]; - tensor input_55_split_num_splits_0 = const()[name = tensor("input_55_split_num_splits_0"), val = tensor(2)]; - tensor input_55_split_axis_0 = const()[name = tensor("input_55_split_axis_0"), val = tensor(1)]; - tensor input_55_split_cast_fp16_0, tensor input_55_split_cast_fp16_1 = split(axis = input_55_split_axis_0, num_splits = input_55_split_num_splits_0, x = input_53_cast_fp16)[name = tensor("input_55_split_cast_fp16")]; - tensor input_55_split_1_sigmoid_cast_fp16 = sigmoid(x = input_55_split_cast_fp16_1)[name = tensor("input_55_split_1_sigmoid_cast_fp16")]; - tensor input_55_cast_fp16 = mul(x = input_55_split_cast_fp16_0, y = input_55_split_1_sigmoid_cast_fp16)[name = tensor("input_55_cast_fp16")]; - tensor input_57_pad_type_0 = const()[name = tensor("input_57_pad_type_0"), val = tensor("custom")]; - tensor input_57_pad_0 = const()[name = tensor("input_57_pad_0"), val = tensor([0, 0, 4, 4])]; - tensor input_57_groups_0 = const()[name = tensor("input_57_groups_0"), val = tensor(512)]; - tensor input_57_strides_0 = const()[name = tensor("input_57_strides_0"), val = tensor([1, 1])]; - tensor input_57_dilations_0 = const()[name = tensor("input_57_dilations_0"), val = tensor([1, 1])]; - tensor const_193_to_fp16 = const()[name = tensor("const_193_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11230848)))]; - tensor const_194_to_fp16 = const()[name = tensor("const_194_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11240128)))]; - tensor input_59_cast_fp16 = conv(bias = const_194_to_fp16, dilations = input_57_dilations_0, groups = input_57_groups_0, pad = input_57_pad_0, pad_type = input_57_pad_type_0, strides = input_57_strides_0, weight = const_193_to_fp16, x = input_55_cast_fp16)[name = tensor("input_59_cast_fp16")]; - tensor input_61_cast_fp16 = silu(x = input_59_cast_fp16)[name = tensor("input_61_cast_fp16")]; - tensor var_809_pad_type_0 = const()[name = tensor("op_809_pad_type_0"), val = tensor("valid")]; - tensor var_809_strides_0 = const()[name = tensor("op_809_strides_0"), val = tensor([1, 1])]; - tensor var_809_pad_0 = const()[name = tensor("op_809_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_809_dilations_0 = const()[name = tensor("op_809_dilations_0"), val = tensor([1, 1])]; - tensor var_809_groups_0 = const()[name = tensor("op_809_groups_0"), val = tensor(1)]; - tensor layers_1_conv_pointwise_conv2_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11241216))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11437888))), name = tensor("layers_1_conv_pointwise_conv2_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_1_conv_pointwise_conv2_inlier_module_bias_to_fp16 = const()[name = tensor("layers_1_conv_pointwise_conv2_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11438080)))]; - tensor var_809_cast_fp16 = conv(bias = layers_1_conv_pointwise_conv2_inlier_module_bias_to_fp16, dilations = var_809_dilations_0, groups = var_809_groups_0, pad = var_809_pad_0, pad_type = var_809_pad_type_0, strides = var_809_strides_0, weight = layers_1_conv_pointwise_conv2_inlier_module_weight_to_fp16_palettized, x = input_61_cast_fp16)[name = tensor("op_809_cast_fp16")]; - tensor var_815_pad_type_0 = const()[name = tensor("op_815_pad_type_0"), val = tensor("valid")]; - tensor var_815_strides_0 = const()[name = tensor("op_815_strides_0"), val = tensor([1, 1])]; - tensor var_815_pad_0 = const()[name = tensor("op_815_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_815_dilations_0 = const()[name = tensor("op_815_dilations_0"), val = tensor([1, 1])]; - tensor var_815_groups_0 = const()[name = tensor("op_815_groups_0"), val = tensor(1)]; - tensor layers_1_conv_pointwise_conv2_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11447872))), name = tensor("layers_1_conv_pointwise_conv2_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11439168))), shape = tensor([512, 512, 1, 1])]; - tensor var_815_cast_fp16 = conv(dilations = var_815_dilations_0, groups = var_815_groups_0, pad = var_815_pad_0, pad_type = var_815_pad_type_0, strides = var_815_strides_0, weight = layers_1_conv_pointwise_conv2_outlier_module_weight_to_fp16_sparsified, x = input_61_cast_fp16)[name = tensor("op_815_cast_fp16")]; - tensor x_11_cast_fp16 = add(x = var_809_cast_fp16, y = var_815_cast_fp16)[name = tensor("x_11_cast_fp16")]; - tensor inputs_17_cast_fp16 = add(x = inputs_15_cast_fp16, y = x_11_cast_fp16)[name = tensor("inputs_17_cast_fp16")]; - tensor out_17_axes_0 = const()[name = tensor("out_17_axes_0"), val = tensor([1])]; - tensor var_826_to_fp16 = const()[name = tensor("op_826_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_17_cast_fp16 = layer_norm(axes = out_17_axes_0, epsilon = var_826_to_fp16, x = inputs_17_cast_fp16)[name = tensor("out_17_cast_fp16")]; - tensor input_63_gamma_0_to_fp16 = const()[name = tensor("input_63_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11480704)))]; - tensor input_63_beta_0_to_fp16 = const()[name = tensor("input_63_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11481792)))]; - tensor input_63_epsilon_0_to_fp16 = const()[name = tensor("input_63_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_63_cast_fp16 = batch_norm(beta = input_63_beta_0_to_fp16, epsilon = input_63_epsilon_0_to_fp16, gamma = input_63_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_17_cast_fp16)[name = tensor("input_63_cast_fp16")]; - tensor var_846_pad_type_0 = const()[name = tensor("op_846_pad_type_0"), val = tensor("valid")]; - tensor var_846_strides_0 = const()[name = tensor("op_846_strides_0"), val = tensor([1, 1])]; - tensor var_846_pad_0 = const()[name = tensor("op_846_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_846_dilations_0 = const()[name = tensor("op_846_dilations_0"), val = tensor([1, 1])]; - tensor var_846_groups_0 = const()[name = tensor("op_846_groups_0"), val = tensor(1)]; - tensor layers_1_feed_forward2_fc1_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11482880))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12269376))), name = tensor("layers_1_feed_forward2_fc1_inlier_module_weight_to_fp16_palettized"), shape = tensor([2048, 512, 1, 1])]; - tensor layers_1_feed_forward2_fc1_inlier_module_bias_to_fp16 = const()[name = tensor("layers_1_feed_forward2_fc1_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12269568)))]; - tensor var_846_cast_fp16 = conv(bias = layers_1_feed_forward2_fc1_inlier_module_bias_to_fp16, dilations = var_846_dilations_0, groups = var_846_groups_0, pad = var_846_pad_0, pad_type = var_846_pad_type_0, strides = var_846_strides_0, weight = layers_1_feed_forward2_fc1_inlier_module_weight_to_fp16_palettized, x = input_63_cast_fp16)[name = tensor("op_846_cast_fp16")]; - tensor var_852_pad_type_0 = const()[name = tensor("op_852_pad_type_0"), val = tensor("valid")]; - tensor var_852_strides_0 = const()[name = tensor("op_852_strides_0"), val = tensor([1, 1])]; - tensor var_852_pad_0 = const()[name = tensor("op_852_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_852_dilations_0 = const()[name = tensor("op_852_dilations_0"), val = tensor([1, 1])]; - tensor var_852_groups_0 = const()[name = tensor("op_852_groups_0"), val = tensor(1)]; - tensor layers_1_feed_forward2_fc1_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12306112))), name = tensor("layers_1_feed_forward2_fc1_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12273728))), shape = tensor([2048, 512, 1, 1])]; - tensor var_852_cast_fp16 = conv(dilations = var_852_dilations_0, groups = var_852_groups_0, pad = var_852_pad_0, pad_type = var_852_pad_type_0, strides = var_852_strides_0, weight = layers_1_feed_forward2_fc1_outlier_module_weight_to_fp16_sparsified, x = input_63_cast_fp16)[name = tensor("op_852_cast_fp16")]; - tensor input_65_cast_fp16 = add(x = var_846_cast_fp16, y = var_852_cast_fp16)[name = tensor("input_65_cast_fp16")]; - tensor input_67_cast_fp16 = silu(x = input_65_cast_fp16)[name = tensor("input_67_cast_fp16")]; - tensor var_863_pad_type_0 = const()[name = tensor("op_863_pad_type_0"), val = tensor("valid")]; - tensor var_863_strides_0 = const()[name = tensor("op_863_strides_0"), val = tensor([1, 1])]; - tensor var_863_pad_0 = const()[name = tensor("op_863_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_863_dilations_0 = const()[name = tensor("op_863_dilations_0"), val = tensor([1, 1])]; - tensor var_863_groups_0 = const()[name = tensor("op_863_groups_0"), val = tensor(1)]; - tensor layers_1_feed_forward2_fc2_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12437248))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(13223744))), name = tensor("layers_1_feed_forward2_fc2_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 2048, 1, 1])]; - tensor layers_1_feed_forward2_fc2_inlier_module_bias_to_fp16 = const()[name = tensor("layers_1_feed_forward2_fc2_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(13223936)))]; - tensor var_863_cast_fp16 = conv(bias = layers_1_feed_forward2_fc2_inlier_module_bias_to_fp16, dilations = var_863_dilations_0, groups = var_863_groups_0, pad = var_863_pad_0, pad_type = var_863_pad_type_0, strides = var_863_strides_0, weight = layers_1_feed_forward2_fc2_inlier_module_weight_to_fp16_palettized, x = input_67_cast_fp16)[name = tensor("op_863_cast_fp16")]; - tensor var_869_pad_type_0 = const()[name = tensor("op_869_pad_type_0"), val = tensor("valid")]; - tensor var_869_strides_0 = const()[name = tensor("op_869_strides_0"), val = tensor([1, 1])]; - tensor var_869_pad_0 = const()[name = tensor("op_869_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_869_dilations_0 = const()[name = tensor("op_869_dilations_0"), val = tensor([1, 1])]; - tensor var_869_groups_0 = const()[name = tensor("op_869_groups_0"), val = tensor(1)]; - tensor layers_1_feed_forward2_fc2_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(13260736))), name = tensor("layers_1_feed_forward2_fc2_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(13225024))), shape = tensor([512, 2048, 1, 1])]; - tensor var_869_cast_fp16 = conv(dilations = var_869_dilations_0, groups = var_869_groups_0, pad = var_869_pad_0, pad_type = var_869_pad_type_0, strides = var_869_strides_0, weight = layers_1_feed_forward2_fc2_outlier_module_weight_to_fp16_sparsified, x = input_67_cast_fp16)[name = tensor("op_869_cast_fp16")]; - tensor x_13_cast_fp16 = add(x = var_863_cast_fp16, y = var_869_cast_fp16)[name = tensor("x_13_cast_fp16")]; - tensor var_871_to_fp16 = const()[name = tensor("op_871_to_fp16"), val = tensor(0x1p-1)]; - tensor var_872_cast_fp16 = mul(x = x_13_cast_fp16, y = var_871_to_fp16)[name = tensor("op_872_cast_fp16")]; - tensor inputs_19_cast_fp16 = add(x = inputs_17_cast_fp16, y = var_872_cast_fp16)[name = tensor("inputs_19_cast_fp16")]; - tensor out_19_axes_0 = const()[name = tensor("out_19_axes_0"), val = tensor([1])]; - tensor var_882_to_fp16 = const()[name = tensor("op_882_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_19_cast_fp16 = layer_norm(axes = out_19_axes_0, epsilon = var_882_to_fp16, x = inputs_19_cast_fp16)[name = tensor("out_19_cast_fp16")]; - tensor inputs_21_gamma_0_to_fp16 = const()[name = tensor("inputs_21_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(13391872)))]; - tensor inputs_21_beta_0_to_fp16 = const()[name = tensor("inputs_21_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(13392960)))]; - tensor inputs_21_epsilon_0_to_fp16 = const()[name = tensor("inputs_21_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor inputs_21_cast_fp16 = batch_norm(beta = inputs_21_beta_0_to_fp16, epsilon = inputs_21_epsilon_0_to_fp16, gamma = inputs_21_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_19_cast_fp16)[name = tensor("inputs_21_cast_fp16")]; - tensor var_896 = const()[name = tensor("op_896"), val = tensor(3)]; - tensor out_21_axes_0 = const()[name = tensor("out_21_axes_0"), val = tensor([1])]; - tensor var_927_to_fp16 = const()[name = tensor("op_927_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_21_cast_fp16 = layer_norm(axes = out_21_axes_0, epsilon = var_927_to_fp16, x = inputs_21_cast_fp16)[name = tensor("out_21_cast_fp16")]; - tensor input_69_gamma_0_to_fp16 = const()[name = tensor("input_69_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(13394048)))]; - tensor input_69_beta_0_to_fp16 = const()[name = tensor("input_69_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(13395136)))]; - tensor input_69_epsilon_0_to_fp16 = const()[name = tensor("input_69_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_69_cast_fp16 = batch_norm(beta = input_69_beta_0_to_fp16, epsilon = input_69_epsilon_0_to_fp16, gamma = input_69_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_21_cast_fp16)[name = tensor("input_69_cast_fp16")]; - tensor var_947_pad_type_0 = const()[name = tensor("op_947_pad_type_0"), val = tensor("valid")]; - tensor var_947_strides_0 = const()[name = tensor("op_947_strides_0"), val = tensor([1, 1])]; - tensor var_947_pad_0 = const()[name = tensor("op_947_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_947_dilations_0 = const()[name = tensor("op_947_dilations_0"), val = tensor([1, 1])]; - tensor var_947_groups_0 = const()[name = tensor("op_947_groups_0"), val = tensor(1)]; - tensor layers_2_feed_forward1_fc1_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(13396224))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(14182720))), name = tensor("layers_2_feed_forward1_fc1_inlier_module_weight_to_fp16_palettized"), shape = tensor([2048, 512, 1, 1])]; - tensor layers_2_feed_forward1_fc1_inlier_module_bias_to_fp16 = const()[name = tensor("layers_2_feed_forward1_fc1_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(14182912)))]; - tensor var_947_cast_fp16 = conv(bias = layers_2_feed_forward1_fc1_inlier_module_bias_to_fp16, dilations = var_947_dilations_0, groups = var_947_groups_0, pad = var_947_pad_0, pad_type = var_947_pad_type_0, strides = var_947_strides_0, weight = layers_2_feed_forward1_fc1_inlier_module_weight_to_fp16_palettized, x = input_69_cast_fp16)[name = tensor("op_947_cast_fp16")]; - tensor var_953_pad_type_0 = const()[name = tensor("op_953_pad_type_0"), val = tensor("valid")]; - tensor var_953_strides_0 = const()[name = tensor("op_953_strides_0"), val = tensor([1, 1])]; - tensor var_953_pad_0 = const()[name = tensor("op_953_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_953_dilations_0 = const()[name = tensor("op_953_dilations_0"), val = tensor([1, 1])]; - tensor var_953_groups_0 = const()[name = tensor("op_953_groups_0"), val = tensor(1)]; - tensor layers_2_feed_forward1_fc1_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(14219904))), name = tensor("layers_2_feed_forward1_fc1_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(14187072))), shape = tensor([2048, 512, 1, 1])]; - tensor var_953_cast_fp16 = conv(dilations = var_953_dilations_0, groups = var_953_groups_0, pad = var_953_pad_0, pad_type = var_953_pad_type_0, strides = var_953_strides_0, weight = layers_2_feed_forward1_fc1_outlier_module_weight_to_fp16_sparsified, x = input_69_cast_fp16)[name = tensor("op_953_cast_fp16")]; - tensor input_71_cast_fp16 = add(x = var_947_cast_fp16, y = var_953_cast_fp16)[name = tensor("input_71_cast_fp16")]; - tensor input_73_cast_fp16 = silu(x = input_71_cast_fp16)[name = tensor("input_73_cast_fp16")]; - tensor var_964_pad_type_0 = const()[name = tensor("op_964_pad_type_0"), val = tensor("valid")]; - tensor var_964_strides_0 = const()[name = tensor("op_964_strides_0"), val = tensor([1, 1])]; - tensor var_964_pad_0 = const()[name = tensor("op_964_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_964_dilations_0 = const()[name = tensor("op_964_dilations_0"), val = tensor([1, 1])]; - tensor var_964_groups_0 = const()[name = tensor("op_964_groups_0"), val = tensor(1)]; - tensor layers_2_feed_forward1_fc2_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(14351040))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15137536))), name = tensor("layers_2_feed_forward1_fc2_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 2048, 1, 1])]; - tensor layers_2_feed_forward1_fc2_inlier_module_bias_to_fp16 = const()[name = tensor("layers_2_feed_forward1_fc2_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15137728)))]; - tensor var_964_cast_fp16 = conv(bias = layers_2_feed_forward1_fc2_inlier_module_bias_to_fp16, dilations = var_964_dilations_0, groups = var_964_groups_0, pad = var_964_pad_0, pad_type = var_964_pad_type_0, strides = var_964_strides_0, weight = layers_2_feed_forward1_fc2_inlier_module_weight_to_fp16_palettized, x = input_73_cast_fp16)[name = tensor("op_964_cast_fp16")]; - tensor var_970_pad_type_0 = const()[name = tensor("op_970_pad_type_0"), val = tensor("valid")]; - tensor var_970_strides_0 = const()[name = tensor("op_970_strides_0"), val = tensor([1, 1])]; - tensor var_970_pad_0 = const()[name = tensor("op_970_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_970_dilations_0 = const()[name = tensor("op_970_dilations_0"), val = tensor([1, 1])]; - tensor var_970_groups_0 = const()[name = tensor("op_970_groups_0"), val = tensor(1)]; - tensor layers_2_feed_forward1_fc2_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15173824))), name = tensor("layers_2_feed_forward1_fc2_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15138816))), shape = tensor([512, 2048, 1, 1])]; - tensor var_970_cast_fp16 = conv(dilations = var_970_dilations_0, groups = var_970_groups_0, pad = var_970_pad_0, pad_type = var_970_pad_type_0, strides = var_970_strides_0, weight = layers_2_feed_forward1_fc2_outlier_module_weight_to_fp16_sparsified, x = input_73_cast_fp16)[name = tensor("op_970_cast_fp16")]; - tensor x_15_cast_fp16 = add(x = var_964_cast_fp16, y = var_970_cast_fp16)[name = tensor("x_15_cast_fp16")]; - tensor var_972_to_fp16 = const()[name = tensor("op_972_to_fp16"), val = tensor(0x1p-1)]; - tensor var_973_cast_fp16 = mul(x = x_15_cast_fp16, y = var_972_to_fp16)[name = tensor("op_973_cast_fp16")]; - tensor inputs_23_cast_fp16 = add(x = inputs_21_cast_fp16, y = var_973_cast_fp16)[name = tensor("inputs_23_cast_fp16")]; - tensor out_23_axes_0 = const()[name = tensor("out_23_axes_0"), val = tensor([1])]; - tensor var_983_to_fp16 = const()[name = tensor("op_983_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_23_cast_fp16 = layer_norm(axes = out_23_axes_0, epsilon = var_983_to_fp16, x = inputs_23_cast_fp16)[name = tensor("out_23_cast_fp16")]; - tensor obj_11_gamma_0_to_fp16 = const()[name = tensor("obj_11_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15304960)))]; - tensor obj_11_beta_0_to_fp16 = const()[name = tensor("obj_11_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15306048)))]; - tensor obj_11_epsilon_0_to_fp16 = const()[name = tensor("obj_11_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor obj_11_cast_fp16 = batch_norm(beta = obj_11_beta_0_to_fp16, epsilon = obj_11_epsilon_0_to_fp16, gamma = obj_11_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_23_cast_fp16)[name = tensor("obj_11_cast_fp16")]; - tensor var_1008_pad_type_0 = const()[name = tensor("op_1008_pad_type_0"), val = tensor("valid")]; - tensor var_1008_strides_0 = const()[name = tensor("op_1008_strides_0"), val = tensor([1, 1])]; - tensor var_1008_pad_0 = const()[name = tensor("op_1008_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1008_dilations_0 = const()[name = tensor("op_1008_dilations_0"), val = tensor([1, 1])]; - tensor var_1008_groups_0 = const()[name = tensor("op_1008_groups_0"), val = tensor(1)]; - tensor layers_2_self_attn_q_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15307136))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15503808))), name = tensor("layers_2_self_attn_q_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_2_self_attn_q_proj_inlier_module_bias_to_fp16 = const()[name = tensor("layers_2_self_attn_q_proj_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15504000)))]; - tensor var_1008_cast_fp16 = conv(bias = layers_2_self_attn_q_proj_inlier_module_bias_to_fp16, dilations = var_1008_dilations_0, groups = var_1008_groups_0, pad = var_1008_pad_0, pad_type = var_1008_pad_type_0, strides = var_1008_strides_0, weight = layers_2_self_attn_q_proj_inlier_module_weight_to_fp16_palettized, x = obj_11_cast_fp16)[name = tensor("op_1008_cast_fp16")]; - tensor var_1014_pad_type_0 = const()[name = tensor("op_1014_pad_type_0"), val = tensor("valid")]; - tensor var_1014_strides_0 = const()[name = tensor("op_1014_strides_0"), val = tensor([1, 1])]; - tensor var_1014_pad_0 = const()[name = tensor("op_1014_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1014_dilations_0 = const()[name = tensor("op_1014_dilations_0"), val = tensor([1, 1])]; - tensor var_1014_groups_0 = const()[name = tensor("op_1014_groups_0"), val = tensor(1)]; - tensor layers_2_self_attn_q_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15513472))), name = tensor("layers_2_self_attn_q_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15505088))), shape = tensor([512, 512, 1, 1])]; - tensor var_1014_cast_fp16 = conv(dilations = var_1014_dilations_0, groups = var_1014_groups_0, pad = var_1014_pad_0, pad_type = var_1014_pad_type_0, strides = var_1014_strides_0, weight = layers_2_self_attn_q_proj_outlier_module_weight_to_fp16_sparsified, x = obj_11_cast_fp16)[name = tensor("op_1014_cast_fp16")]; - tensor query_9_cast_fp16 = add(x = var_1008_cast_fp16, y = var_1014_cast_fp16)[name = tensor("query_9_cast_fp16")]; - tensor var_1023_pad_type_0 = const()[name = tensor("op_1023_pad_type_0"), val = tensor("valid")]; - tensor var_1023_strides_0 = const()[name = tensor("op_1023_strides_0"), val = tensor([1, 1])]; - tensor var_1023_pad_0 = const()[name = tensor("op_1023_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1023_dilations_0 = const()[name = tensor("op_1023_dilations_0"), val = tensor([1, 1])]; - tensor var_1023_groups_0 = const()[name = tensor("op_1023_groups_0"), val = tensor(1)]; - tensor layers_2_self_attn_k_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15546304))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15742976))), name = tensor("layers_2_self_attn_k_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor var_1023_cast_fp16 = conv(dilations = var_1023_dilations_0, groups = var_1023_groups_0, pad = var_1023_pad_0, pad_type = var_1023_pad_type_0, strides = var_1023_strides_0, weight = layers_2_self_attn_k_proj_inlier_module_weight_to_fp16_palettized, x = obj_11_cast_fp16)[name = tensor("op_1023_cast_fp16")]; - tensor var_1029_pad_type_0 = const()[name = tensor("op_1029_pad_type_0"), val = tensor("valid")]; - tensor var_1029_strides_0 = const()[name = tensor("op_1029_strides_0"), val = tensor([1, 1])]; - tensor var_1029_pad_0 = const()[name = tensor("op_1029_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1029_dilations_0 = const()[name = tensor("op_1029_dilations_0"), val = tensor([1, 1])]; - tensor var_1029_groups_0 = const()[name = tensor("op_1029_groups_0"), val = tensor(1)]; - tensor layers_2_self_attn_k_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15751936))), name = tensor("layers_2_self_attn_k_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15743168))), shape = tensor([512, 512, 1, 1])]; - tensor var_1029_cast_fp16 = conv(dilations = var_1029_dilations_0, groups = var_1029_groups_0, pad = var_1029_pad_0, pad_type = var_1029_pad_type_0, strides = var_1029_strides_0, weight = layers_2_self_attn_k_proj_outlier_module_weight_to_fp16_sparsified, x = obj_11_cast_fp16)[name = tensor("op_1029_cast_fp16")]; - tensor key_5_cast_fp16 = add(x = var_1023_cast_fp16, y = var_1029_cast_fp16)[name = tensor("key_5_cast_fp16")]; - tensor var_1039_pad_type_0 = const()[name = tensor("op_1039_pad_type_0"), val = tensor("valid")]; - tensor var_1039_strides_0 = const()[name = tensor("op_1039_strides_0"), val = tensor([1, 1])]; - tensor var_1039_pad_0 = const()[name = tensor("op_1039_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1039_dilations_0 = const()[name = tensor("op_1039_dilations_0"), val = tensor([1, 1])]; - tensor var_1039_groups_0 = const()[name = tensor("op_1039_groups_0"), val = tensor(1)]; - tensor layers_2_self_attn_v_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15784768))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15981440))), name = tensor("layers_2_self_attn_v_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_2_self_attn_v_proj_inlier_module_bias_to_fp16 = const()[name = tensor("layers_2_self_attn_v_proj_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15981632)))]; - tensor var_1039_cast_fp16 = conv(bias = layers_2_self_attn_v_proj_inlier_module_bias_to_fp16, dilations = var_1039_dilations_0, groups = var_1039_groups_0, pad = var_1039_pad_0, pad_type = var_1039_pad_type_0, strides = var_1039_strides_0, weight = layers_2_self_attn_v_proj_inlier_module_weight_to_fp16_palettized, x = obj_11_cast_fp16)[name = tensor("op_1039_cast_fp16")]; - tensor var_1045_pad_type_0 = const()[name = tensor("op_1045_pad_type_0"), val = tensor("valid")]; - tensor var_1045_strides_0 = const()[name = tensor("op_1045_strides_0"), val = tensor([1, 1])]; - tensor var_1045_pad_0 = const()[name = tensor("op_1045_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1045_dilations_0 = const()[name = tensor("op_1045_dilations_0"), val = tensor([1, 1])]; - tensor var_1045_groups_0 = const()[name = tensor("op_1045_groups_0"), val = tensor(1)]; - tensor layers_2_self_attn_v_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15990208))), name = tensor("layers_2_self_attn_v_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15982720))), shape = tensor([512, 512, 1, 1])]; - tensor var_1045_cast_fp16 = conv(dilations = var_1045_dilations_0, groups = var_1045_groups_0, pad = var_1045_pad_0, pad_type = var_1045_pad_type_0, strides = var_1045_strides_0, weight = layers_2_self_attn_v_proj_outlier_module_weight_to_fp16_sparsified, x = obj_11_cast_fp16)[name = tensor("op_1045_cast_fp16")]; - tensor value_5_cast_fp16 = add(x = var_1039_cast_fp16, y = var_1045_cast_fp16)[name = tensor("value_5_cast_fp16")]; - tensor var_1048_to_fp16 = const()[name = tensor("op_1048_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16023040)))]; - tensor query_11_cast_fp16 = add(x = query_9_cast_fp16, y = var_1048_to_fp16)[name = tensor("query_11_cast_fp16")]; - tensor var_1051_to_fp16 = const()[name = tensor("op_1051_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16024128)))]; - tensor q_with_bias_v_5_cast_fp16 = add(x = query_9_cast_fp16, y = var_1051_to_fp16)[name = tensor("q_with_bias_v_5_cast_fp16")]; - tensor var_1061_pad_type_0 = const()[name = tensor("op_1061_pad_type_0"), val = tensor("valid")]; - tensor var_1061_strides_0 = const()[name = tensor("op_1061_strides_0"), val = tensor([1, 1])]; - tensor var_1061_pad_0 = const()[name = tensor("op_1061_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1061_dilations_0 = const()[name = tensor("op_1061_dilations_0"), val = tensor([1, 1])]; - tensor var_1061_groups_0 = const()[name = tensor("op_1061_groups_0"), val = tensor(1)]; - tensor layers_2_self_attn_linear_pos_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16025216))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16221888))), name = tensor("layers_2_self_attn_linear_pos_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor var_1061_cast_fp16 = conv(bias = input_17_mean_0_to_fp16, dilations = var_1061_dilations_0, groups = var_1061_groups_0, pad = var_1061_pad_0, pad_type = var_1061_pad_type_0, strides = var_1061_strides_0, weight = layers_2_self_attn_linear_pos_inlier_module_weight_to_fp16_palettized, x = obj_3_cast_fp16)[name = tensor("op_1061_cast_fp16")]; - tensor var_1067_pad_type_0 = const()[name = tensor("op_1067_pad_type_0"), val = tensor("valid")]; - tensor var_1067_strides_0 = const()[name = tensor("op_1067_strides_0"), val = tensor([1, 1])]; - tensor var_1067_pad_0 = const()[name = tensor("op_1067_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1067_dilations_0 = const()[name = tensor("op_1067_dilations_0"), val = tensor([1, 1])]; - tensor var_1067_groups_0 = const()[name = tensor("op_1067_groups_0"), val = tensor(1)]; - tensor layers_2_self_attn_linear_pos_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16243968))), name = tensor("layers_2_self_attn_linear_pos_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16222080))), shape = tensor([512, 512, 1, 1])]; - tensor var_1067_cast_fp16 = conv(dilations = var_1067_dilations_0, groups = var_1067_groups_0, pad = var_1067_pad_0, pad_type = var_1067_pad_type_0, strides = var_1067_strides_0, weight = layers_2_self_attn_linear_pos_outlier_module_weight_to_fp16_sparsified, x = obj_3_cast_fp16)[name = tensor("op_1067_cast_fp16")]; - tensor p_5_cast_fp16 = add(x = var_1061_cast_fp16, y = var_1067_cast_fp16)[name = tensor("p_5_cast_fp16")]; - tensor var_1071 = const()[name = tensor("op_1071"), val = tensor([1, 8, 64, 188])]; - tensor var_1072_cast_fp16 = reshape(shape = var_1071, x = q_with_bias_v_5_cast_fp16)[name = tensor("op_1072_cast_fp16")]; - tensor var_1073 = const()[name = tensor("op_1073"), val = tensor([1, 8, 64, -1])]; - tensor var_1074_cast_fp16 = reshape(shape = var_1073, x = p_5_cast_fp16)[name = tensor("op_1074_cast_fp16")]; - tensor matrix_bd_17_transpose_x_0 = const()[name = tensor("matrix_bd_17_transpose_x_0"), val = tensor(true)]; - tensor matrix_bd_17_transpose_y_0 = const()[name = tensor("matrix_bd_17_transpose_y_0"), val = tensor(false)]; - tensor matrix_bd_17_cast_fp16 = matmul(transpose_x = matrix_bd_17_transpose_x_0, transpose_y = matrix_bd_17_transpose_y_0, x = var_1072_cast_fp16, y = var_1074_cast_fp16)[name = tensor("matrix_bd_17_cast_fp16")]; - tensor matrix_bd_19_pad_0 = const()[name = tensor("matrix_bd_19_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor matrix_bd_19_mode_0 = const()[name = tensor("matrix_bd_19_mode_0"), val = tensor("constant")]; - tensor const_32_to_fp16 = const()[name = tensor("const_32_to_fp16"), val = tensor(0x0p+0)]; - tensor matrix_bd_19_cast_fp16 = pad(constant_val = const_32_to_fp16, mode = matrix_bd_19_mode_0, pad = matrix_bd_19_pad_0, x = matrix_bd_17_cast_fp16)[name = tensor("matrix_bd_19_cast_fp16")]; - tensor var_1083 = const()[name = tensor("op_1083"), val = tensor([1, 8, -1, 188])]; - tensor matrix_bd_21_cast_fp16 = reshape(shape = var_1083, x = matrix_bd_19_cast_fp16)[name = tensor("matrix_bd_21_cast_fp16")]; - tensor var_1087_begin_0 = const()[name = tensor("op_1087_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_1087_end_0 = const()[name = tensor("op_1087_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_1087_end_mask_0 = const()[name = tensor("op_1087_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_1087_cast_fp16 = slice_by_index(begin = var_1087_begin_0, end = var_1087_end_0, end_mask = var_1087_end_mask_0, x = matrix_bd_21_cast_fp16)[name = tensor("op_1087_cast_fp16")]; - tensor var_1088 = const()[name = tensor("op_1088"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_23_cast_fp16 = reshape(shape = var_1088, x = var_1087_cast_fp16)[name = tensor("matrix_bd_23_cast_fp16")]; - tensor var_1093_begin_0 = const()[name = tensor("op_1093_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1093_end_0 = const()[name = tensor("op_1093_end_0"), val = tensor([1, 8, 188, 188])]; - tensor var_1093_end_mask_0 = const()[name = tensor("op_1093_end_mask_0"), val = tensor([true, true, true, false])]; - tensor var_1093_cast_fp16 = slice_by_index(begin = var_1093_begin_0, end = var_1093_end_0, end_mask = var_1093_end_mask_0, x = matrix_bd_23_cast_fp16)[name = tensor("op_1093_cast_fp16")]; - tensor var_1094_to_fp16 = const()[name = tensor("op_1094_to_fp16"), val = tensor(0x1p-3)]; - tensor qk_mask_5_cast_fp16 = mul(x = var_1093_cast_fp16, y = var_1094_to_fp16)[name = tensor("qk_mask_5_cast_fp16")]; - tensor var_1098 = const()[name = tensor("op_1098"), val = tensor([1, 8, 64, 188])]; - tensor mh_q_5_cast_fp16 = reshape(shape = var_1098, x = query_11_cast_fp16)[name = tensor("mh_q_5_cast_fp16")]; - tensor var_1100_to_fp16 = const()[name = tensor("op_1100_to_fp16"), val = tensor(0x1p-3)]; - tensor var_1101_cast_fp16 = mul(x = mh_q_5_cast_fp16, y = var_1100_to_fp16)[name = tensor("op_1101_cast_fp16")]; - tensor var_1104 = const()[name = tensor("op_1104"), val = tensor([1, 8, 64, 188])]; - tensor var_1105_cast_fp16 = reshape(shape = var_1104, x = key_5_cast_fp16)[name = tensor("op_1105_cast_fp16")]; - tensor mh_w_9_transpose_x_0 = const()[name = tensor("mh_w_9_transpose_x_0"), val = tensor(true)]; - tensor mh_w_9_transpose_y_0 = const()[name = tensor("mh_w_9_transpose_y_0"), val = tensor(false)]; - tensor mh_w_9_cast_fp16 = matmul(transpose_x = mh_w_9_transpose_x_0, transpose_y = mh_w_9_transpose_y_0, x = var_1101_cast_fp16, y = var_1105_cast_fp16)[name = tensor("mh_w_9_cast_fp16")]; - tensor mh_w_11_cast_fp16 = add(x = mh_w_9_cast_fp16, y = qk_mask_5_cast_fp16)[name = tensor("mh_w_11_cast_fp16")]; - tensor var_1109_cast_fp16 = softmax(axis = var_896, x = mh_w_11_cast_fp16)[name = tensor("op_1109_cast_fp16")]; - tensor var_1110 = const()[name = tensor("op_1110"), val = tensor([1, 8, 64, 188])]; - tensor var_1111_cast_fp16 = reshape(shape = var_1110, x = value_5_cast_fp16)[name = tensor("op_1111_cast_fp16")]; - tensor attn_5_transpose_x_0 = const()[name = tensor("attn_5_transpose_x_0"), val = tensor(false)]; - tensor attn_5_transpose_y_0 = const()[name = tensor("attn_5_transpose_y_0"), val = tensor(true)]; - tensor attn_5_cast_fp16 = matmul(transpose_x = attn_5_transpose_x_0, transpose_y = attn_5_transpose_y_0, x = var_1111_cast_fp16, y = var_1109_cast_fp16)[name = tensor("attn_5_cast_fp16")]; - tensor var_1114 = const()[name = tensor("op_1114"), val = tensor([1, 512, 1, 188])]; - tensor input_75_cast_fp16 = reshape(shape = var_1114, x = attn_5_cast_fp16)[name = tensor("input_75_cast_fp16")]; - tensor var_1124_pad_type_0 = const()[name = tensor("op_1124_pad_type_0"), val = tensor("valid")]; - tensor var_1124_strides_0 = const()[name = tensor("op_1124_strides_0"), val = tensor([1, 1])]; - tensor var_1124_pad_0 = const()[name = tensor("op_1124_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1124_dilations_0 = const()[name = tensor("op_1124_dilations_0"), val = tensor([1, 1])]; - tensor var_1124_groups_0 = const()[name = tensor("op_1124_groups_0"), val = tensor(1)]; - tensor layers_2_self_attn_o_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16276800))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16473472))), name = tensor("layers_2_self_attn_o_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_2_self_attn_o_proj_inlier_module_bias_to_fp16 = const()[name = tensor("layers_2_self_attn_o_proj_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16473664)))]; - tensor var_1124_cast_fp16 = conv(bias = layers_2_self_attn_o_proj_inlier_module_bias_to_fp16, dilations = var_1124_dilations_0, groups = var_1124_groups_0, pad = var_1124_pad_0, pad_type = var_1124_pad_type_0, strides = var_1124_strides_0, weight = layers_2_self_attn_o_proj_inlier_module_weight_to_fp16_palettized, x = input_75_cast_fp16)[name = tensor("op_1124_cast_fp16")]; - tensor var_1130_pad_type_0 = const()[name = tensor("op_1130_pad_type_0"), val = tensor("valid")]; - tensor var_1130_strides_0 = const()[name = tensor("op_1130_strides_0"), val = tensor([1, 1])]; - tensor var_1130_pad_0 = const()[name = tensor("op_1130_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1130_dilations_0 = const()[name = tensor("op_1130_dilations_0"), val = tensor([1, 1])]; - tensor var_1130_groups_0 = const()[name = tensor("op_1130_groups_0"), val = tensor(1)]; - tensor layers_2_self_attn_o_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16482560))), name = tensor("layers_2_self_attn_o_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16474752))), shape = tensor([512, 512, 1, 1])]; - tensor var_1130_cast_fp16 = conv(dilations = var_1130_dilations_0, groups = var_1130_groups_0, pad = var_1130_pad_0, pad_type = var_1130_pad_type_0, strides = var_1130_strides_0, weight = layers_2_self_attn_o_proj_outlier_module_weight_to_fp16_sparsified, x = input_75_cast_fp16)[name = tensor("op_1130_cast_fp16")]; - tensor obj_13_cast_fp16 = add(x = var_1124_cast_fp16, y = var_1130_cast_fp16)[name = tensor("obj_13_cast_fp16")]; - tensor inputs_25_cast_fp16 = add(x = inputs_23_cast_fp16, y = obj_13_cast_fp16)[name = tensor("inputs_25_cast_fp16")]; - tensor out_25_axes_0 = const()[name = tensor("out_25_axes_0"), val = tensor([1])]; - tensor var_1141_to_fp16 = const()[name = tensor("op_1141_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_25_cast_fp16 = layer_norm(axes = out_25_axes_0, epsilon = var_1141_to_fp16, x = inputs_25_cast_fp16)[name = tensor("out_25_cast_fp16")]; - tensor input_77_gamma_0_to_fp16 = const()[name = tensor("input_77_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16515392)))]; - tensor input_77_beta_0_to_fp16 = const()[name = tensor("input_77_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16516480)))]; - tensor input_77_epsilon_0_to_fp16 = const()[name = tensor("input_77_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_77_cast_fp16 = batch_norm(beta = input_77_beta_0_to_fp16, epsilon = input_77_epsilon_0_to_fp16, gamma = input_77_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_25_cast_fp16)[name = tensor("input_77_cast_fp16")]; - tensor var_1163_pad_type_0 = const()[name = tensor("op_1163_pad_type_0"), val = tensor("valid")]; - tensor var_1163_strides_0 = const()[name = tensor("op_1163_strides_0"), val = tensor([1, 1])]; - tensor var_1163_pad_0 = const()[name = tensor("op_1163_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1163_dilations_0 = const()[name = tensor("op_1163_dilations_0"), val = tensor([1, 1])]; - tensor var_1163_groups_0 = const()[name = tensor("op_1163_groups_0"), val = tensor(1)]; - tensor layers_2_conv_pointwise_conv1_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16517568))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16910848))), name = tensor("layers_2_conv_pointwise_conv1_inlier_module_weight_to_fp16_palettized"), shape = tensor([1024, 512, 1, 1])]; - tensor layers_2_conv_pointwise_conv1_inlier_module_bias_to_fp16 = const()[name = tensor("layers_2_conv_pointwise_conv1_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16911040)))]; - tensor var_1163_cast_fp16 = conv(bias = layers_2_conv_pointwise_conv1_inlier_module_bias_to_fp16, dilations = var_1163_dilations_0, groups = var_1163_groups_0, pad = var_1163_pad_0, pad_type = var_1163_pad_type_0, strides = var_1163_strides_0, weight = layers_2_conv_pointwise_conv1_inlier_module_weight_to_fp16_palettized, x = input_77_cast_fp16)[name = tensor("op_1163_cast_fp16")]; - tensor var_1169_pad_type_0 = const()[name = tensor("op_1169_pad_type_0"), val = tensor("valid")]; - tensor var_1169_strides_0 = const()[name = tensor("op_1169_strides_0"), val = tensor([1, 1])]; - tensor var_1169_pad_0 = const()[name = tensor("op_1169_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1169_dilations_0 = const()[name = tensor("op_1169_dilations_0"), val = tensor([1, 1])]; - tensor var_1169_groups_0 = const()[name = tensor("op_1169_groups_0"), val = tensor(1)]; - tensor layers_2_conv_pointwise_conv1_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16930752))), name = tensor("layers_2_conv_pointwise_conv1_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16913152))), shape = tensor([1024, 512, 1, 1])]; - tensor var_1169_cast_fp16 = conv(dilations = var_1169_dilations_0, groups = var_1169_groups_0, pad = var_1169_pad_0, pad_type = var_1169_pad_type_0, strides = var_1169_strides_0, weight = layers_2_conv_pointwise_conv1_outlier_module_weight_to_fp16_sparsified, x = input_77_cast_fp16)[name = tensor("op_1169_cast_fp16")]; - tensor input_79_cast_fp16 = add(x = var_1163_cast_fp16, y = var_1169_cast_fp16)[name = tensor("input_79_cast_fp16")]; - tensor input_81_split_num_splits_0 = const()[name = tensor("input_81_split_num_splits_0"), val = tensor(2)]; - tensor input_81_split_axis_0 = const()[name = tensor("input_81_split_axis_0"), val = tensor(1)]; - tensor input_81_split_cast_fp16_0, tensor input_81_split_cast_fp16_1 = split(axis = input_81_split_axis_0, num_splits = input_81_split_num_splits_0, x = input_79_cast_fp16)[name = tensor("input_81_split_cast_fp16")]; - tensor input_81_split_1_sigmoid_cast_fp16 = sigmoid(x = input_81_split_cast_fp16_1)[name = tensor("input_81_split_1_sigmoid_cast_fp16")]; - tensor input_81_cast_fp16 = mul(x = input_81_split_cast_fp16_0, y = input_81_split_1_sigmoid_cast_fp16)[name = tensor("input_81_cast_fp16")]; - tensor input_83_pad_type_0 = const()[name = tensor("input_83_pad_type_0"), val = tensor("custom")]; - tensor input_83_pad_0 = const()[name = tensor("input_83_pad_0"), val = tensor([0, 0, 4, 4])]; - tensor input_83_groups_0 = const()[name = tensor("input_83_groups_0"), val = tensor(512)]; - tensor input_83_strides_0 = const()[name = tensor("input_83_strides_0"), val = tensor([1, 1])]; - tensor input_83_dilations_0 = const()[name = tensor("input_83_dilations_0"), val = tensor([1, 1])]; - tensor const_195_to_fp16 = const()[name = tensor("const_195_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16996352)))]; - tensor const_196_to_fp16 = const()[name = tensor("const_196_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(17005632)))]; - tensor input_85_cast_fp16 = conv(bias = const_196_to_fp16, dilations = input_83_dilations_0, groups = input_83_groups_0, pad = input_83_pad_0, pad_type = input_83_pad_type_0, strides = input_83_strides_0, weight = const_195_to_fp16, x = input_81_cast_fp16)[name = tensor("input_85_cast_fp16")]; - tensor input_87_cast_fp16 = silu(x = input_85_cast_fp16)[name = tensor("input_87_cast_fp16")]; - tensor var_1193_pad_type_0 = const()[name = tensor("op_1193_pad_type_0"), val = tensor("valid")]; - tensor var_1193_strides_0 = const()[name = tensor("op_1193_strides_0"), val = tensor([1, 1])]; - tensor var_1193_pad_0 = const()[name = tensor("op_1193_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1193_dilations_0 = const()[name = tensor("op_1193_dilations_0"), val = tensor([1, 1])]; - tensor var_1193_groups_0 = const()[name = tensor("op_1193_groups_0"), val = tensor(1)]; - tensor layers_2_conv_pointwise_conv2_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(17006720))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(17203392))), name = tensor("layers_2_conv_pointwise_conv2_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_2_conv_pointwise_conv2_inlier_module_bias_to_fp16 = const()[name = tensor("layers_2_conv_pointwise_conv2_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(17203584)))]; - tensor var_1193_cast_fp16 = conv(bias = layers_2_conv_pointwise_conv2_inlier_module_bias_to_fp16, dilations = var_1193_dilations_0, groups = var_1193_groups_0, pad = var_1193_pad_0, pad_type = var_1193_pad_type_0, strides = var_1193_strides_0, weight = layers_2_conv_pointwise_conv2_inlier_module_weight_to_fp16_palettized, x = input_87_cast_fp16)[name = tensor("op_1193_cast_fp16")]; - tensor var_1199_pad_type_0 = const()[name = tensor("op_1199_pad_type_0"), val = tensor("valid")]; - tensor var_1199_strides_0 = const()[name = tensor("op_1199_strides_0"), val = tensor([1, 1])]; - tensor var_1199_pad_0 = const()[name = tensor("op_1199_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1199_dilations_0 = const()[name = tensor("op_1199_dilations_0"), val = tensor([1, 1])]; - tensor var_1199_groups_0 = const()[name = tensor("op_1199_groups_0"), val = tensor(1)]; - tensor layers_2_conv_pointwise_conv2_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(17213248))), name = tensor("layers_2_conv_pointwise_conv2_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(17204672))), shape = tensor([512, 512, 1, 1])]; - tensor var_1199_cast_fp16 = conv(dilations = var_1199_dilations_0, groups = var_1199_groups_0, pad = var_1199_pad_0, pad_type = var_1199_pad_type_0, strides = var_1199_strides_0, weight = layers_2_conv_pointwise_conv2_outlier_module_weight_to_fp16_sparsified, x = input_87_cast_fp16)[name = tensor("op_1199_cast_fp16")]; - tensor x_17_cast_fp16 = add(x = var_1193_cast_fp16, y = var_1199_cast_fp16)[name = tensor("x_17_cast_fp16")]; - tensor inputs_27_cast_fp16 = add(x = inputs_25_cast_fp16, y = x_17_cast_fp16)[name = tensor("inputs_27_cast_fp16")]; - tensor out_27_axes_0 = const()[name = tensor("out_27_axes_0"), val = tensor([1])]; - tensor var_1210_to_fp16 = const()[name = tensor("op_1210_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_27_cast_fp16 = layer_norm(axes = out_27_axes_0, epsilon = var_1210_to_fp16, x = inputs_27_cast_fp16)[name = tensor("out_27_cast_fp16")]; - tensor input_89_gamma_0_to_fp16 = const()[name = tensor("input_89_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(17246080)))]; - tensor input_89_beta_0_to_fp16 = const()[name = tensor("input_89_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(17247168)))]; - tensor input_89_epsilon_0_to_fp16 = const()[name = tensor("input_89_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_89_cast_fp16 = batch_norm(beta = input_89_beta_0_to_fp16, epsilon = input_89_epsilon_0_to_fp16, gamma = input_89_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_27_cast_fp16)[name = tensor("input_89_cast_fp16")]; - tensor var_1230_pad_type_0 = const()[name = tensor("op_1230_pad_type_0"), val = tensor("valid")]; - tensor var_1230_strides_0 = const()[name = tensor("op_1230_strides_0"), val = tensor([1, 1])]; - tensor var_1230_pad_0 = const()[name = tensor("op_1230_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1230_dilations_0 = const()[name = tensor("op_1230_dilations_0"), val = tensor([1, 1])]; - tensor var_1230_groups_0 = const()[name = tensor("op_1230_groups_0"), val = tensor(1)]; - tensor layers_2_feed_forward2_fc1_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(17248256))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(18034752))), name = tensor("layers_2_feed_forward2_fc1_inlier_module_weight_to_fp16_palettized"), shape = tensor([2048, 512, 1, 1])]; - tensor layers_2_feed_forward2_fc1_inlier_module_bias_to_fp16 = const()[name = tensor("layers_2_feed_forward2_fc1_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(18034944)))]; - tensor var_1230_cast_fp16 = conv(bias = layers_2_feed_forward2_fc1_inlier_module_bias_to_fp16, dilations = var_1230_dilations_0, groups = var_1230_groups_0, pad = var_1230_pad_0, pad_type = var_1230_pad_type_0, strides = var_1230_strides_0, weight = layers_2_feed_forward2_fc1_inlier_module_weight_to_fp16_palettized, x = input_89_cast_fp16)[name = tensor("op_1230_cast_fp16")]; - tensor var_1236_pad_type_0 = const()[name = tensor("op_1236_pad_type_0"), val = tensor("valid")]; - tensor var_1236_strides_0 = const()[name = tensor("op_1236_strides_0"), val = tensor([1, 1])]; - tensor var_1236_pad_0 = const()[name = tensor("op_1236_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1236_dilations_0 = const()[name = tensor("op_1236_dilations_0"), val = tensor([1, 1])]; - tensor var_1236_groups_0 = const()[name = tensor("op_1236_groups_0"), val = tensor(1)]; - tensor layers_2_feed_forward2_fc1_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(18070784))), name = tensor("layers_2_feed_forward2_fc1_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(18039104))), shape = tensor([2048, 512, 1, 1])]; - tensor var_1236_cast_fp16 = conv(dilations = var_1236_dilations_0, groups = var_1236_groups_0, pad = var_1236_pad_0, pad_type = var_1236_pad_type_0, strides = var_1236_strides_0, weight = layers_2_feed_forward2_fc1_outlier_module_weight_to_fp16_sparsified, x = input_89_cast_fp16)[name = tensor("op_1236_cast_fp16")]; - tensor input_91_cast_fp16 = add(x = var_1230_cast_fp16, y = var_1236_cast_fp16)[name = tensor("input_91_cast_fp16")]; - tensor input_93_cast_fp16 = silu(x = input_91_cast_fp16)[name = tensor("input_93_cast_fp16")]; - tensor var_1247_pad_type_0 = const()[name = tensor("op_1247_pad_type_0"), val = tensor("valid")]; - tensor var_1247_strides_0 = const()[name = tensor("op_1247_strides_0"), val = tensor([1, 1])]; - tensor var_1247_pad_0 = const()[name = tensor("op_1247_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1247_dilations_0 = const()[name = tensor("op_1247_dilations_0"), val = tensor([1, 1])]; - tensor var_1247_groups_0 = const()[name = tensor("op_1247_groups_0"), val = tensor(1)]; - tensor layers_2_feed_forward2_fc2_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(18201920))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(18988416))), name = tensor("layers_2_feed_forward2_fc2_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 2048, 1, 1])]; - tensor layers_2_feed_forward2_fc2_inlier_module_bias_to_fp16 = const()[name = tensor("layers_2_feed_forward2_fc2_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(18988608)))]; - tensor var_1247_cast_fp16 = conv(bias = layers_2_feed_forward2_fc2_inlier_module_bias_to_fp16, dilations = var_1247_dilations_0, groups = var_1247_groups_0, pad = var_1247_pad_0, pad_type = var_1247_pad_type_0, strides = var_1247_strides_0, weight = layers_2_feed_forward2_fc2_inlier_module_weight_to_fp16_palettized, x = input_93_cast_fp16)[name = tensor("op_1247_cast_fp16")]; - tensor var_1253_pad_type_0 = const()[name = tensor("op_1253_pad_type_0"), val = tensor("valid")]; - tensor var_1253_strides_0 = const()[name = tensor("op_1253_strides_0"), val = tensor([1, 1])]; - tensor var_1253_pad_0 = const()[name = tensor("op_1253_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1253_dilations_0 = const()[name = tensor("op_1253_dilations_0"), val = tensor([1, 1])]; - tensor var_1253_groups_0 = const()[name = tensor("op_1253_groups_0"), val = tensor(1)]; - tensor layers_2_feed_forward2_fc2_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19025600))), name = tensor("layers_2_feed_forward2_fc2_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(18989696))), shape = tensor([512, 2048, 1, 1])]; - tensor var_1253_cast_fp16 = conv(dilations = var_1253_dilations_0, groups = var_1253_groups_0, pad = var_1253_pad_0, pad_type = var_1253_pad_type_0, strides = var_1253_strides_0, weight = layers_2_feed_forward2_fc2_outlier_module_weight_to_fp16_sparsified, x = input_93_cast_fp16)[name = tensor("op_1253_cast_fp16")]; - tensor x_19_cast_fp16 = add(x = var_1247_cast_fp16, y = var_1253_cast_fp16)[name = tensor("x_19_cast_fp16")]; - tensor var_1255_to_fp16 = const()[name = tensor("op_1255_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1256_cast_fp16 = mul(x = x_19_cast_fp16, y = var_1255_to_fp16)[name = tensor("op_1256_cast_fp16")]; - tensor inputs_29_cast_fp16 = add(x = inputs_27_cast_fp16, y = var_1256_cast_fp16)[name = tensor("inputs_29_cast_fp16")]; - tensor out_29_axes_0 = const()[name = tensor("out_29_axes_0"), val = tensor([1])]; - tensor var_1266_to_fp16 = const()[name = tensor("op_1266_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_29_cast_fp16 = layer_norm(axes = out_29_axes_0, epsilon = var_1266_to_fp16, x = inputs_29_cast_fp16)[name = tensor("out_29_cast_fp16")]; - tensor inputs_31_gamma_0_to_fp16 = const()[name = tensor("inputs_31_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19156736)))]; - tensor inputs_31_beta_0_to_fp16 = const()[name = tensor("inputs_31_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19157824)))]; - tensor inputs_31_epsilon_0_to_fp16 = const()[name = tensor("inputs_31_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor inputs_31_cast_fp16 = batch_norm(beta = inputs_31_beta_0_to_fp16, epsilon = inputs_31_epsilon_0_to_fp16, gamma = inputs_31_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_29_cast_fp16)[name = tensor("inputs_31_cast_fp16")]; - tensor var_1280 = const()[name = tensor("op_1280"), val = tensor(3)]; - tensor out_31_axes_0 = const()[name = tensor("out_31_axes_0"), val = tensor([1])]; - tensor var_1311_to_fp16 = const()[name = tensor("op_1311_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_31_cast_fp16 = layer_norm(axes = out_31_axes_0, epsilon = var_1311_to_fp16, x = inputs_31_cast_fp16)[name = tensor("out_31_cast_fp16")]; - tensor input_95_gamma_0_to_fp16 = const()[name = tensor("input_95_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19158912)))]; - tensor input_95_beta_0_to_fp16 = const()[name = tensor("input_95_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19160000)))]; - tensor input_95_epsilon_0_to_fp16 = const()[name = tensor("input_95_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_95_cast_fp16 = batch_norm(beta = input_95_beta_0_to_fp16, epsilon = input_95_epsilon_0_to_fp16, gamma = input_95_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_31_cast_fp16)[name = tensor("input_95_cast_fp16")]; - tensor var_1331_pad_type_0 = const()[name = tensor("op_1331_pad_type_0"), val = tensor("valid")]; - tensor var_1331_strides_0 = const()[name = tensor("op_1331_strides_0"), val = tensor([1, 1])]; - tensor var_1331_pad_0 = const()[name = tensor("op_1331_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1331_dilations_0 = const()[name = tensor("op_1331_dilations_0"), val = tensor([1, 1])]; - tensor var_1331_groups_0 = const()[name = tensor("op_1331_groups_0"), val = tensor(1)]; - tensor layers_3_feed_forward1_fc1_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19161088))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19947584))), name = tensor("layers_3_feed_forward1_fc1_inlier_module_weight_to_fp16_palettized"), shape = tensor([2048, 512, 1, 1])]; - tensor layers_3_feed_forward1_fc1_inlier_module_bias_to_fp16 = const()[name = tensor("layers_3_feed_forward1_fc1_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19947776)))]; - tensor var_1331_cast_fp16 = conv(bias = layers_3_feed_forward1_fc1_inlier_module_bias_to_fp16, dilations = var_1331_dilations_0, groups = var_1331_groups_0, pad = var_1331_pad_0, pad_type = var_1331_pad_type_0, strides = var_1331_strides_0, weight = layers_3_feed_forward1_fc1_inlier_module_weight_to_fp16_palettized, x = input_95_cast_fp16)[name = tensor("op_1331_cast_fp16")]; - tensor var_1337_pad_type_0 = const()[name = tensor("op_1337_pad_type_0"), val = tensor("valid")]; - tensor var_1337_strides_0 = const()[name = tensor("op_1337_strides_0"), val = tensor([1, 1])]; - tensor var_1337_pad_0 = const()[name = tensor("op_1337_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1337_dilations_0 = const()[name = tensor("op_1337_dilations_0"), val = tensor([1, 1])]; - tensor var_1337_groups_0 = const()[name = tensor("op_1337_groups_0"), val = tensor(1)]; - tensor layers_3_feed_forward1_fc1_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19984128))), name = tensor("layers_3_feed_forward1_fc1_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19951936))), shape = tensor([2048, 512, 1, 1])]; - tensor var_1337_cast_fp16 = conv(dilations = var_1337_dilations_0, groups = var_1337_groups_0, pad = var_1337_pad_0, pad_type = var_1337_pad_type_0, strides = var_1337_strides_0, weight = layers_3_feed_forward1_fc1_outlier_module_weight_to_fp16_sparsified, x = input_95_cast_fp16)[name = tensor("op_1337_cast_fp16")]; - tensor input_97_cast_fp16 = add(x = var_1331_cast_fp16, y = var_1337_cast_fp16)[name = tensor("input_97_cast_fp16")]; - tensor input_99_cast_fp16 = silu(x = input_97_cast_fp16)[name = tensor("input_99_cast_fp16")]; - tensor var_1348_pad_type_0 = const()[name = tensor("op_1348_pad_type_0"), val = tensor("valid")]; - tensor var_1348_strides_0 = const()[name = tensor("op_1348_strides_0"), val = tensor([1, 1])]; - tensor var_1348_pad_0 = const()[name = tensor("op_1348_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1348_dilations_0 = const()[name = tensor("op_1348_dilations_0"), val = tensor([1, 1])]; - tensor var_1348_groups_0 = const()[name = tensor("op_1348_groups_0"), val = tensor(1)]; - tensor layers_3_feed_forward1_fc2_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(20115264))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(20901760))), name = tensor("layers_3_feed_forward1_fc2_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 2048, 1, 1])]; - tensor layers_3_feed_forward1_fc2_inlier_module_bias_to_fp16 = const()[name = tensor("layers_3_feed_forward1_fc2_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(20901952)))]; - tensor var_1348_cast_fp16 = conv(bias = layers_3_feed_forward1_fc2_inlier_module_bias_to_fp16, dilations = var_1348_dilations_0, groups = var_1348_groups_0, pad = var_1348_pad_0, pad_type = var_1348_pad_type_0, strides = var_1348_strides_0, weight = layers_3_feed_forward1_fc2_inlier_module_weight_to_fp16_palettized, x = input_99_cast_fp16)[name = tensor("op_1348_cast_fp16")]; - tensor var_1354_pad_type_0 = const()[name = tensor("op_1354_pad_type_0"), val = tensor("valid")]; - tensor var_1354_strides_0 = const()[name = tensor("op_1354_strides_0"), val = tensor([1, 1])]; - tensor var_1354_pad_0 = const()[name = tensor("op_1354_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1354_dilations_0 = const()[name = tensor("op_1354_dilations_0"), val = tensor([1, 1])]; - tensor var_1354_groups_0 = const()[name = tensor("op_1354_groups_0"), val = tensor(1)]; - tensor layers_3_feed_forward1_fc2_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(20938176))), name = tensor("layers_3_feed_forward1_fc2_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(20903040))), shape = tensor([512, 2048, 1, 1])]; - tensor var_1354_cast_fp16 = conv(dilations = var_1354_dilations_0, groups = var_1354_groups_0, pad = var_1354_pad_0, pad_type = var_1354_pad_type_0, strides = var_1354_strides_0, weight = layers_3_feed_forward1_fc2_outlier_module_weight_to_fp16_sparsified, x = input_99_cast_fp16)[name = tensor("op_1354_cast_fp16")]; - tensor x_21_cast_fp16 = add(x = var_1348_cast_fp16, y = var_1354_cast_fp16)[name = tensor("x_21_cast_fp16")]; - tensor var_1356_to_fp16 = const()[name = tensor("op_1356_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1357_cast_fp16 = mul(x = x_21_cast_fp16, y = var_1356_to_fp16)[name = tensor("op_1357_cast_fp16")]; - tensor inputs_33_cast_fp16 = add(x = inputs_31_cast_fp16, y = var_1357_cast_fp16)[name = tensor("inputs_33_cast_fp16")]; - tensor out_33_axes_0 = const()[name = tensor("out_33_axes_0"), val = tensor([1])]; - tensor var_1367_to_fp16 = const()[name = tensor("op_1367_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_33_cast_fp16 = layer_norm(axes = out_33_axes_0, epsilon = var_1367_to_fp16, x = inputs_33_cast_fp16)[name = tensor("out_33_cast_fp16")]; - tensor obj_15_gamma_0_to_fp16 = const()[name = tensor("obj_15_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21069312)))]; - tensor obj_15_beta_0_to_fp16 = const()[name = tensor("obj_15_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21070400)))]; - tensor obj_15_epsilon_0_to_fp16 = const()[name = tensor("obj_15_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor obj_15_cast_fp16 = batch_norm(beta = obj_15_beta_0_to_fp16, epsilon = obj_15_epsilon_0_to_fp16, gamma = obj_15_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_33_cast_fp16)[name = tensor("obj_15_cast_fp16")]; - tensor var_1392_pad_type_0 = const()[name = tensor("op_1392_pad_type_0"), val = tensor("valid")]; - tensor var_1392_strides_0 = const()[name = tensor("op_1392_strides_0"), val = tensor([1, 1])]; - tensor var_1392_pad_0 = const()[name = tensor("op_1392_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1392_dilations_0 = const()[name = tensor("op_1392_dilations_0"), val = tensor([1, 1])]; - tensor var_1392_groups_0 = const()[name = tensor("op_1392_groups_0"), val = tensor(1)]; - tensor layers_3_self_attn_q_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21071488))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21268160))), name = tensor("layers_3_self_attn_q_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_3_self_attn_q_proj_inlier_module_bias_to_fp16 = const()[name = tensor("layers_3_self_attn_q_proj_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21268352)))]; - tensor var_1392_cast_fp16 = conv(bias = layers_3_self_attn_q_proj_inlier_module_bias_to_fp16, dilations = var_1392_dilations_0, groups = var_1392_groups_0, pad = var_1392_pad_0, pad_type = var_1392_pad_type_0, strides = var_1392_strides_0, weight = layers_3_self_attn_q_proj_inlier_module_weight_to_fp16_palettized, x = obj_15_cast_fp16)[name = tensor("op_1392_cast_fp16")]; - tensor var_1398_pad_type_0 = const()[name = tensor("op_1398_pad_type_0"), val = tensor("valid")]; - tensor var_1398_strides_0 = const()[name = tensor("op_1398_strides_0"), val = tensor([1, 1])]; - tensor var_1398_pad_0 = const()[name = tensor("op_1398_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1398_dilations_0 = const()[name = tensor("op_1398_dilations_0"), val = tensor([1, 1])]; - tensor var_1398_groups_0 = const()[name = tensor("op_1398_groups_0"), val = tensor(1)]; - tensor layers_3_self_attn_q_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21278208))), name = tensor("layers_3_self_attn_q_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21269440))), shape = tensor([512, 512, 1, 1])]; - tensor var_1398_cast_fp16 = conv(dilations = var_1398_dilations_0, groups = var_1398_groups_0, pad = var_1398_pad_0, pad_type = var_1398_pad_type_0, strides = var_1398_strides_0, weight = layers_3_self_attn_q_proj_outlier_module_weight_to_fp16_sparsified, x = obj_15_cast_fp16)[name = tensor("op_1398_cast_fp16")]; - tensor query_13_cast_fp16 = add(x = var_1392_cast_fp16, y = var_1398_cast_fp16)[name = tensor("query_13_cast_fp16")]; - tensor var_1407_pad_type_0 = const()[name = tensor("op_1407_pad_type_0"), val = tensor("valid")]; - tensor var_1407_strides_0 = const()[name = tensor("op_1407_strides_0"), val = tensor([1, 1])]; - tensor var_1407_pad_0 = const()[name = tensor("op_1407_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1407_dilations_0 = const()[name = tensor("op_1407_dilations_0"), val = tensor([1, 1])]; - tensor var_1407_groups_0 = const()[name = tensor("op_1407_groups_0"), val = tensor(1)]; - tensor layers_3_self_attn_k_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21311040))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21507712))), name = tensor("layers_3_self_attn_k_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor var_1407_cast_fp16 = conv(dilations = var_1407_dilations_0, groups = var_1407_groups_0, pad = var_1407_pad_0, pad_type = var_1407_pad_type_0, strides = var_1407_strides_0, weight = layers_3_self_attn_k_proj_inlier_module_weight_to_fp16_palettized, x = obj_15_cast_fp16)[name = tensor("op_1407_cast_fp16")]; - tensor var_1413_pad_type_0 = const()[name = tensor("op_1413_pad_type_0"), val = tensor("valid")]; - tensor var_1413_strides_0 = const()[name = tensor("op_1413_strides_0"), val = tensor([1, 1])]; - tensor var_1413_pad_0 = const()[name = tensor("op_1413_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1413_dilations_0 = const()[name = tensor("op_1413_dilations_0"), val = tensor([1, 1])]; - tensor var_1413_groups_0 = const()[name = tensor("op_1413_groups_0"), val = tensor(1)]; - tensor layers_3_self_attn_k_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21517824))), name = tensor("layers_3_self_attn_k_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21507904))), shape = tensor([512, 512, 1, 1])]; - tensor var_1413_cast_fp16 = conv(dilations = var_1413_dilations_0, groups = var_1413_groups_0, pad = var_1413_pad_0, pad_type = var_1413_pad_type_0, strides = var_1413_strides_0, weight = layers_3_self_attn_k_proj_outlier_module_weight_to_fp16_sparsified, x = obj_15_cast_fp16)[name = tensor("op_1413_cast_fp16")]; - tensor key_7_cast_fp16 = add(x = var_1407_cast_fp16, y = var_1413_cast_fp16)[name = tensor("key_7_cast_fp16")]; - tensor var_1423_pad_type_0 = const()[name = tensor("op_1423_pad_type_0"), val = tensor("valid")]; - tensor var_1423_strides_0 = const()[name = tensor("op_1423_strides_0"), val = tensor([1, 1])]; - tensor var_1423_pad_0 = const()[name = tensor("op_1423_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1423_dilations_0 = const()[name = tensor("op_1423_dilations_0"), val = tensor([1, 1])]; - tensor var_1423_groups_0 = const()[name = tensor("op_1423_groups_0"), val = tensor(1)]; - tensor layers_3_self_attn_v_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21550656))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21747328))), name = tensor("layers_3_self_attn_v_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_3_self_attn_v_proj_inlier_module_bias_to_fp16 = const()[name = tensor("layers_3_self_attn_v_proj_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21747520)))]; - tensor var_1423_cast_fp16 = conv(bias = layers_3_self_attn_v_proj_inlier_module_bias_to_fp16, dilations = var_1423_dilations_0, groups = var_1423_groups_0, pad = var_1423_pad_0, pad_type = var_1423_pad_type_0, strides = var_1423_strides_0, weight = layers_3_self_attn_v_proj_inlier_module_weight_to_fp16_palettized, x = obj_15_cast_fp16)[name = tensor("op_1423_cast_fp16")]; - tensor var_1429_pad_type_0 = const()[name = tensor("op_1429_pad_type_0"), val = tensor("valid")]; - tensor var_1429_strides_0 = const()[name = tensor("op_1429_strides_0"), val = tensor([1, 1])]; - tensor var_1429_pad_0 = const()[name = tensor("op_1429_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1429_dilations_0 = const()[name = tensor("op_1429_dilations_0"), val = tensor([1, 1])]; - tensor var_1429_groups_0 = const()[name = tensor("op_1429_groups_0"), val = tensor(1)]; - tensor layers_3_self_attn_v_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21756352))), name = tensor("layers_3_self_attn_v_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21748608))), shape = tensor([512, 512, 1, 1])]; - tensor var_1429_cast_fp16 = conv(dilations = var_1429_dilations_0, groups = var_1429_groups_0, pad = var_1429_pad_0, pad_type = var_1429_pad_type_0, strides = var_1429_strides_0, weight = layers_3_self_attn_v_proj_outlier_module_weight_to_fp16_sparsified, x = obj_15_cast_fp16)[name = tensor("op_1429_cast_fp16")]; - tensor value_7_cast_fp16 = add(x = var_1423_cast_fp16, y = var_1429_cast_fp16)[name = tensor("value_7_cast_fp16")]; - tensor var_1432_to_fp16 = const()[name = tensor("op_1432_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21789184)))]; - tensor query_15_cast_fp16 = add(x = query_13_cast_fp16, y = var_1432_to_fp16)[name = tensor("query_15_cast_fp16")]; - tensor var_1435_to_fp16 = const()[name = tensor("op_1435_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21790272)))]; - tensor q_with_bias_v_7_cast_fp16 = add(x = query_13_cast_fp16, y = var_1435_to_fp16)[name = tensor("q_with_bias_v_7_cast_fp16")]; - tensor var_1445_pad_type_0 = const()[name = tensor("op_1445_pad_type_0"), val = tensor("valid")]; - tensor var_1445_strides_0 = const()[name = tensor("op_1445_strides_0"), val = tensor([1, 1])]; - tensor var_1445_pad_0 = const()[name = tensor("op_1445_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1445_dilations_0 = const()[name = tensor("op_1445_dilations_0"), val = tensor([1, 1])]; - tensor var_1445_groups_0 = const()[name = tensor("op_1445_groups_0"), val = tensor(1)]; - tensor layers_3_self_attn_linear_pos_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21791360))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21988032))), name = tensor("layers_3_self_attn_linear_pos_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor var_1445_cast_fp16 = conv(bias = input_17_mean_0_to_fp16, dilations = var_1445_dilations_0, groups = var_1445_groups_0, pad = var_1445_pad_0, pad_type = var_1445_pad_type_0, strides = var_1445_strides_0, weight = layers_3_self_attn_linear_pos_inlier_module_weight_to_fp16_palettized, x = obj_3_cast_fp16)[name = tensor("op_1445_cast_fp16")]; - tensor var_1451_pad_type_0 = const()[name = tensor("op_1451_pad_type_0"), val = tensor("valid")]; - tensor var_1451_strides_0 = const()[name = tensor("op_1451_strides_0"), val = tensor([1, 1])]; - tensor var_1451_pad_0 = const()[name = tensor("op_1451_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1451_dilations_0 = const()[name = tensor("op_1451_dilations_0"), val = tensor([1, 1])]; - tensor var_1451_groups_0 = const()[name = tensor("op_1451_groups_0"), val = tensor(1)]; - tensor layers_3_self_attn_linear_pos_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22007936))), name = tensor("layers_3_self_attn_linear_pos_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21988224))), shape = tensor([512, 512, 1, 1])]; - tensor var_1451_cast_fp16 = conv(dilations = var_1451_dilations_0, groups = var_1451_groups_0, pad = var_1451_pad_0, pad_type = var_1451_pad_type_0, strides = var_1451_strides_0, weight = layers_3_self_attn_linear_pos_outlier_module_weight_to_fp16_sparsified, x = obj_3_cast_fp16)[name = tensor("op_1451_cast_fp16")]; - tensor p_7_cast_fp16 = add(x = var_1445_cast_fp16, y = var_1451_cast_fp16)[name = tensor("p_7_cast_fp16")]; - tensor var_1455 = const()[name = tensor("op_1455"), val = tensor([1, 8, 64, 188])]; - tensor var_1456_cast_fp16 = reshape(shape = var_1455, x = q_with_bias_v_7_cast_fp16)[name = tensor("op_1456_cast_fp16")]; - tensor var_1457 = const()[name = tensor("op_1457"), val = tensor([1, 8, 64, -1])]; - tensor var_1458_cast_fp16 = reshape(shape = var_1457, x = p_7_cast_fp16)[name = tensor("op_1458_cast_fp16")]; - tensor matrix_bd_25_transpose_x_0 = const()[name = tensor("matrix_bd_25_transpose_x_0"), val = tensor(true)]; - tensor matrix_bd_25_transpose_y_0 = const()[name = tensor("matrix_bd_25_transpose_y_0"), val = tensor(false)]; - tensor matrix_bd_25_cast_fp16 = matmul(transpose_x = matrix_bd_25_transpose_x_0, transpose_y = matrix_bd_25_transpose_y_0, x = var_1456_cast_fp16, y = var_1458_cast_fp16)[name = tensor("matrix_bd_25_cast_fp16")]; - tensor matrix_bd_27_pad_0 = const()[name = tensor("matrix_bd_27_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor matrix_bd_27_mode_0 = const()[name = tensor("matrix_bd_27_mode_0"), val = tensor("constant")]; - tensor const_43_to_fp16 = const()[name = tensor("const_43_to_fp16"), val = tensor(0x0p+0)]; - tensor matrix_bd_27_cast_fp16 = pad(constant_val = const_43_to_fp16, mode = matrix_bd_27_mode_0, pad = matrix_bd_27_pad_0, x = matrix_bd_25_cast_fp16)[name = tensor("matrix_bd_27_cast_fp16")]; - tensor var_1467 = const()[name = tensor("op_1467"), val = tensor([1, 8, -1, 188])]; - tensor matrix_bd_29_cast_fp16 = reshape(shape = var_1467, x = matrix_bd_27_cast_fp16)[name = tensor("matrix_bd_29_cast_fp16")]; - tensor var_1471_begin_0 = const()[name = tensor("op_1471_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_1471_end_0 = const()[name = tensor("op_1471_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_1471_end_mask_0 = const()[name = tensor("op_1471_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_1471_cast_fp16 = slice_by_index(begin = var_1471_begin_0, end = var_1471_end_0, end_mask = var_1471_end_mask_0, x = matrix_bd_29_cast_fp16)[name = tensor("op_1471_cast_fp16")]; - tensor var_1472 = const()[name = tensor("op_1472"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_31_cast_fp16 = reshape(shape = var_1472, x = var_1471_cast_fp16)[name = tensor("matrix_bd_31_cast_fp16")]; - tensor var_1477_begin_0 = const()[name = tensor("op_1477_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1477_end_0 = const()[name = tensor("op_1477_end_0"), val = tensor([1, 8, 188, 188])]; - tensor var_1477_end_mask_0 = const()[name = tensor("op_1477_end_mask_0"), val = tensor([true, true, true, false])]; - tensor var_1477_cast_fp16 = slice_by_index(begin = var_1477_begin_0, end = var_1477_end_0, end_mask = var_1477_end_mask_0, x = matrix_bd_31_cast_fp16)[name = tensor("op_1477_cast_fp16")]; - tensor var_1478_to_fp16 = const()[name = tensor("op_1478_to_fp16"), val = tensor(0x1p-3)]; - tensor qk_mask_7_cast_fp16 = mul(x = var_1477_cast_fp16, y = var_1478_to_fp16)[name = tensor("qk_mask_7_cast_fp16")]; - tensor var_1482 = const()[name = tensor("op_1482"), val = tensor([1, 8, 64, 188])]; - tensor mh_q_7_cast_fp16 = reshape(shape = var_1482, x = query_15_cast_fp16)[name = tensor("mh_q_7_cast_fp16")]; - tensor var_1484_to_fp16 = const()[name = tensor("op_1484_to_fp16"), val = tensor(0x1p-3)]; - tensor var_1485_cast_fp16 = mul(x = mh_q_7_cast_fp16, y = var_1484_to_fp16)[name = tensor("op_1485_cast_fp16")]; - tensor var_1488 = const()[name = tensor("op_1488"), val = tensor([1, 8, 64, 188])]; - tensor var_1489_cast_fp16 = reshape(shape = var_1488, x = key_7_cast_fp16)[name = tensor("op_1489_cast_fp16")]; - tensor mh_w_13_transpose_x_0 = const()[name = tensor("mh_w_13_transpose_x_0"), val = tensor(true)]; - tensor mh_w_13_transpose_y_0 = const()[name = tensor("mh_w_13_transpose_y_0"), val = tensor(false)]; - tensor mh_w_13_cast_fp16 = matmul(transpose_x = mh_w_13_transpose_x_0, transpose_y = mh_w_13_transpose_y_0, x = var_1485_cast_fp16, y = var_1489_cast_fp16)[name = tensor("mh_w_13_cast_fp16")]; - tensor mh_w_15_cast_fp16 = add(x = mh_w_13_cast_fp16, y = qk_mask_7_cast_fp16)[name = tensor("mh_w_15_cast_fp16")]; - tensor var_1493_cast_fp16 = softmax(axis = var_1280, x = mh_w_15_cast_fp16)[name = tensor("op_1493_cast_fp16")]; - tensor var_1494 = const()[name = tensor("op_1494"), val = tensor([1, 8, 64, 188])]; - tensor var_1495_cast_fp16 = reshape(shape = var_1494, x = value_7_cast_fp16)[name = tensor("op_1495_cast_fp16")]; - tensor attn_7_transpose_x_0 = const()[name = tensor("attn_7_transpose_x_0"), val = tensor(false)]; - tensor attn_7_transpose_y_0 = const()[name = tensor("attn_7_transpose_y_0"), val = tensor(true)]; - tensor attn_7_cast_fp16 = matmul(transpose_x = attn_7_transpose_x_0, transpose_y = attn_7_transpose_y_0, x = var_1495_cast_fp16, y = var_1493_cast_fp16)[name = tensor("attn_7_cast_fp16")]; - tensor var_1498 = const()[name = tensor("op_1498"), val = tensor([1, 512, 1, 188])]; - tensor input_101_cast_fp16 = reshape(shape = var_1498, x = attn_7_cast_fp16)[name = tensor("input_101_cast_fp16")]; - tensor var_1508_pad_type_0 = const()[name = tensor("op_1508_pad_type_0"), val = tensor("valid")]; - tensor var_1508_strides_0 = const()[name = tensor("op_1508_strides_0"), val = tensor([1, 1])]; - tensor var_1508_pad_0 = const()[name = tensor("op_1508_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1508_dilations_0 = const()[name = tensor("op_1508_dilations_0"), val = tensor([1, 1])]; - tensor var_1508_groups_0 = const()[name = tensor("op_1508_groups_0"), val = tensor(1)]; - tensor layers_3_self_attn_o_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22040768))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22237440))), name = tensor("layers_3_self_attn_o_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_3_self_attn_o_proj_inlier_module_bias_to_fp16 = const()[name = tensor("layers_3_self_attn_o_proj_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22237632)))]; - tensor var_1508_cast_fp16 = conv(bias = layers_3_self_attn_o_proj_inlier_module_bias_to_fp16, dilations = var_1508_dilations_0, groups = var_1508_groups_0, pad = var_1508_pad_0, pad_type = var_1508_pad_type_0, strides = var_1508_strides_0, weight = layers_3_self_attn_o_proj_inlier_module_weight_to_fp16_palettized, x = input_101_cast_fp16)[name = tensor("op_1508_cast_fp16")]; - tensor var_1514_pad_type_0 = const()[name = tensor("op_1514_pad_type_0"), val = tensor("valid")]; - tensor var_1514_strides_0 = const()[name = tensor("op_1514_strides_0"), val = tensor([1, 1])]; - tensor var_1514_pad_0 = const()[name = tensor("op_1514_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1514_dilations_0 = const()[name = tensor("op_1514_dilations_0"), val = tensor([1, 1])]; - tensor var_1514_groups_0 = const()[name = tensor("op_1514_groups_0"), val = tensor(1)]; - tensor layers_3_self_attn_o_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22247104))), name = tensor("layers_3_self_attn_o_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22238720))), shape = tensor([512, 512, 1, 1])]; - tensor var_1514_cast_fp16 = conv(dilations = var_1514_dilations_0, groups = var_1514_groups_0, pad = var_1514_pad_0, pad_type = var_1514_pad_type_0, strides = var_1514_strides_0, weight = layers_3_self_attn_o_proj_outlier_module_weight_to_fp16_sparsified, x = input_101_cast_fp16)[name = tensor("op_1514_cast_fp16")]; - tensor obj_17_cast_fp16 = add(x = var_1508_cast_fp16, y = var_1514_cast_fp16)[name = tensor("obj_17_cast_fp16")]; - tensor inputs_35_cast_fp16 = add(x = inputs_33_cast_fp16, y = obj_17_cast_fp16)[name = tensor("inputs_35_cast_fp16")]; - tensor out_35_axes_0 = const()[name = tensor("out_35_axes_0"), val = tensor([1])]; - tensor var_1525_to_fp16 = const()[name = tensor("op_1525_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_35_cast_fp16 = layer_norm(axes = out_35_axes_0, epsilon = var_1525_to_fp16, x = inputs_35_cast_fp16)[name = tensor("out_35_cast_fp16")]; - tensor input_103_gamma_0_to_fp16 = const()[name = tensor("input_103_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22279936)))]; - tensor input_103_beta_0_to_fp16 = const()[name = tensor("input_103_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22281024)))]; - tensor input_103_epsilon_0_to_fp16 = const()[name = tensor("input_103_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_103_cast_fp16 = batch_norm(beta = input_103_beta_0_to_fp16, epsilon = input_103_epsilon_0_to_fp16, gamma = input_103_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_35_cast_fp16)[name = tensor("input_103_cast_fp16")]; - tensor var_1547_pad_type_0 = const()[name = tensor("op_1547_pad_type_0"), val = tensor("valid")]; - tensor var_1547_strides_0 = const()[name = tensor("op_1547_strides_0"), val = tensor([1, 1])]; - tensor var_1547_pad_0 = const()[name = tensor("op_1547_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1547_dilations_0 = const()[name = tensor("op_1547_dilations_0"), val = tensor([1, 1])]; - tensor var_1547_groups_0 = const()[name = tensor("op_1547_groups_0"), val = tensor(1)]; - tensor layers_3_conv_pointwise_conv1_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22282112))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22675392))), name = tensor("layers_3_conv_pointwise_conv1_inlier_module_weight_to_fp16_palettized"), shape = tensor([1024, 512, 1, 1])]; - tensor layers_3_conv_pointwise_conv1_inlier_module_bias_to_fp16 = const()[name = tensor("layers_3_conv_pointwise_conv1_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22675584)))]; - tensor var_1547_cast_fp16 = conv(bias = layers_3_conv_pointwise_conv1_inlier_module_bias_to_fp16, dilations = var_1547_dilations_0, groups = var_1547_groups_0, pad = var_1547_pad_0, pad_type = var_1547_pad_type_0, strides = var_1547_strides_0, weight = layers_3_conv_pointwise_conv1_inlier_module_weight_to_fp16_palettized, x = input_103_cast_fp16)[name = tensor("op_1547_cast_fp16")]; - tensor var_1553_pad_type_0 = const()[name = tensor("op_1553_pad_type_0"), val = tensor("valid")]; - tensor var_1553_strides_0 = const()[name = tensor("op_1553_strides_0"), val = tensor([1, 1])]; - tensor var_1553_pad_0 = const()[name = tensor("op_1553_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1553_dilations_0 = const()[name = tensor("op_1553_dilations_0"), val = tensor([1, 1])]; - tensor var_1553_groups_0 = const()[name = tensor("op_1553_groups_0"), val = tensor(1)]; - tensor layers_3_conv_pointwise_conv1_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22695360))), name = tensor("layers_3_conv_pointwise_conv1_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22677696))), shape = tensor([1024, 512, 1, 1])]; - tensor var_1553_cast_fp16 = conv(dilations = var_1553_dilations_0, groups = var_1553_groups_0, pad = var_1553_pad_0, pad_type = var_1553_pad_type_0, strides = var_1553_strides_0, weight = layers_3_conv_pointwise_conv1_outlier_module_weight_to_fp16_sparsified, x = input_103_cast_fp16)[name = tensor("op_1553_cast_fp16")]; - tensor input_105_cast_fp16 = add(x = var_1547_cast_fp16, y = var_1553_cast_fp16)[name = tensor("input_105_cast_fp16")]; - tensor input_107_split_num_splits_0 = const()[name = tensor("input_107_split_num_splits_0"), val = tensor(2)]; - tensor input_107_split_axis_0 = const()[name = tensor("input_107_split_axis_0"), val = tensor(1)]; - tensor input_107_split_cast_fp16_0, tensor input_107_split_cast_fp16_1 = split(axis = input_107_split_axis_0, num_splits = input_107_split_num_splits_0, x = input_105_cast_fp16)[name = tensor("input_107_split_cast_fp16")]; - tensor input_107_split_1_sigmoid_cast_fp16 = sigmoid(x = input_107_split_cast_fp16_1)[name = tensor("input_107_split_1_sigmoid_cast_fp16")]; - tensor input_107_cast_fp16 = mul(x = input_107_split_cast_fp16_0, y = input_107_split_1_sigmoid_cast_fp16)[name = tensor("input_107_cast_fp16")]; - tensor input_109_pad_type_0 = const()[name = tensor("input_109_pad_type_0"), val = tensor("custom")]; - tensor input_109_pad_0 = const()[name = tensor("input_109_pad_0"), val = tensor([0, 0, 4, 4])]; - tensor input_109_groups_0 = const()[name = tensor("input_109_groups_0"), val = tensor(512)]; - tensor input_109_strides_0 = const()[name = tensor("input_109_strides_0"), val = tensor([1, 1])]; - tensor input_109_dilations_0 = const()[name = tensor("input_109_dilations_0"), val = tensor([1, 1])]; - tensor const_197_to_fp16 = const()[name = tensor("const_197_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22760960)))]; - tensor const_198_to_fp16 = const()[name = tensor("const_198_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22770240)))]; - tensor input_111_cast_fp16 = conv(bias = const_198_to_fp16, dilations = input_109_dilations_0, groups = input_109_groups_0, pad = input_109_pad_0, pad_type = input_109_pad_type_0, strides = input_109_strides_0, weight = const_197_to_fp16, x = input_107_cast_fp16)[name = tensor("input_111_cast_fp16")]; - tensor input_113_cast_fp16 = silu(x = input_111_cast_fp16)[name = tensor("input_113_cast_fp16")]; - tensor var_1577_pad_type_0 = const()[name = tensor("op_1577_pad_type_0"), val = tensor("valid")]; - tensor var_1577_strides_0 = const()[name = tensor("op_1577_strides_0"), val = tensor([1, 1])]; - tensor var_1577_pad_0 = const()[name = tensor("op_1577_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1577_dilations_0 = const()[name = tensor("op_1577_dilations_0"), val = tensor([1, 1])]; - tensor var_1577_groups_0 = const()[name = tensor("op_1577_groups_0"), val = tensor(1)]; - tensor layers_3_conv_pointwise_conv2_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22771328))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22968000))), name = tensor("layers_3_conv_pointwise_conv2_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_3_conv_pointwise_conv2_inlier_module_bias_to_fp16 = const()[name = tensor("layers_3_conv_pointwise_conv2_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22968192)))]; - tensor var_1577_cast_fp16 = conv(bias = layers_3_conv_pointwise_conv2_inlier_module_bias_to_fp16, dilations = var_1577_dilations_0, groups = var_1577_groups_0, pad = var_1577_pad_0, pad_type = var_1577_pad_type_0, strides = var_1577_strides_0, weight = layers_3_conv_pointwise_conv2_inlier_module_weight_to_fp16_palettized, x = input_113_cast_fp16)[name = tensor("op_1577_cast_fp16")]; - tensor var_1583_pad_type_0 = const()[name = tensor("op_1583_pad_type_0"), val = tensor("valid")]; - tensor var_1583_strides_0 = const()[name = tensor("op_1583_strides_0"), val = tensor([1, 1])]; - tensor var_1583_pad_0 = const()[name = tensor("op_1583_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1583_dilations_0 = const()[name = tensor("op_1583_dilations_0"), val = tensor([1, 1])]; - tensor var_1583_groups_0 = const()[name = tensor("op_1583_groups_0"), val = tensor(1)]; - tensor layers_3_conv_pointwise_conv2_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22978048))), name = tensor("layers_3_conv_pointwise_conv2_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22969280))), shape = tensor([512, 512, 1, 1])]; - tensor var_1583_cast_fp16 = conv(dilations = var_1583_dilations_0, groups = var_1583_groups_0, pad = var_1583_pad_0, pad_type = var_1583_pad_type_0, strides = var_1583_strides_0, weight = layers_3_conv_pointwise_conv2_outlier_module_weight_to_fp16_sparsified, x = input_113_cast_fp16)[name = tensor("op_1583_cast_fp16")]; - tensor x_23_cast_fp16 = add(x = var_1577_cast_fp16, y = var_1583_cast_fp16)[name = tensor("x_23_cast_fp16")]; - tensor inputs_37_cast_fp16 = add(x = inputs_35_cast_fp16, y = x_23_cast_fp16)[name = tensor("inputs_37_cast_fp16")]; - tensor out_37_axes_0 = const()[name = tensor("out_37_axes_0"), val = tensor([1])]; - tensor var_1594_to_fp16 = const()[name = tensor("op_1594_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_37_cast_fp16 = layer_norm(axes = out_37_axes_0, epsilon = var_1594_to_fp16, x = inputs_37_cast_fp16)[name = tensor("out_37_cast_fp16")]; - tensor input_115_gamma_0_to_fp16 = const()[name = tensor("input_115_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23010880)))]; - tensor input_115_beta_0_to_fp16 = const()[name = tensor("input_115_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23011968)))]; - tensor input_115_epsilon_0_to_fp16 = const()[name = tensor("input_115_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_115_cast_fp16 = batch_norm(beta = input_115_beta_0_to_fp16, epsilon = input_115_epsilon_0_to_fp16, gamma = input_115_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_37_cast_fp16)[name = tensor("input_115_cast_fp16")]; - tensor var_1614_pad_type_0 = const()[name = tensor("op_1614_pad_type_0"), val = tensor("valid")]; - tensor var_1614_strides_0 = const()[name = tensor("op_1614_strides_0"), val = tensor([1, 1])]; - tensor var_1614_pad_0 = const()[name = tensor("op_1614_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1614_dilations_0 = const()[name = tensor("op_1614_dilations_0"), val = tensor([1, 1])]; - tensor var_1614_groups_0 = const()[name = tensor("op_1614_groups_0"), val = tensor(1)]; - tensor layers_3_feed_forward2_fc1_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23013056))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23799552))), name = tensor("layers_3_feed_forward2_fc1_inlier_module_weight_to_fp16_palettized"), shape = tensor([2048, 512, 1, 1])]; - tensor layers_3_feed_forward2_fc1_inlier_module_bias_to_fp16 = const()[name = tensor("layers_3_feed_forward2_fc1_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23799744)))]; - tensor var_1614_cast_fp16 = conv(bias = layers_3_feed_forward2_fc1_inlier_module_bias_to_fp16, dilations = var_1614_dilations_0, groups = var_1614_groups_0, pad = var_1614_pad_0, pad_type = var_1614_pad_type_0, strides = var_1614_strides_0, weight = layers_3_feed_forward2_fc1_inlier_module_weight_to_fp16_palettized, x = input_115_cast_fp16)[name = tensor("op_1614_cast_fp16")]; - tensor var_1620_pad_type_0 = const()[name = tensor("op_1620_pad_type_0"), val = tensor("valid")]; - tensor var_1620_strides_0 = const()[name = tensor("op_1620_strides_0"), val = tensor([1, 1])]; - tensor var_1620_pad_0 = const()[name = tensor("op_1620_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1620_dilations_0 = const()[name = tensor("op_1620_dilations_0"), val = tensor([1, 1])]; - tensor var_1620_groups_0 = const()[name = tensor("op_1620_groups_0"), val = tensor(1)]; - tensor layers_3_feed_forward2_fc1_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23835840))), name = tensor("layers_3_feed_forward2_fc1_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23803904))), shape = tensor([2048, 512, 1, 1])]; - tensor var_1620_cast_fp16 = conv(dilations = var_1620_dilations_0, groups = var_1620_groups_0, pad = var_1620_pad_0, pad_type = var_1620_pad_type_0, strides = var_1620_strides_0, weight = layers_3_feed_forward2_fc1_outlier_module_weight_to_fp16_sparsified, x = input_115_cast_fp16)[name = tensor("op_1620_cast_fp16")]; - tensor input_117_cast_fp16 = add(x = var_1614_cast_fp16, y = var_1620_cast_fp16)[name = tensor("input_117_cast_fp16")]; - tensor input_119_cast_fp16 = silu(x = input_117_cast_fp16)[name = tensor("input_119_cast_fp16")]; - tensor var_1631_pad_type_0 = const()[name = tensor("op_1631_pad_type_0"), val = tensor("valid")]; - tensor var_1631_strides_0 = const()[name = tensor("op_1631_strides_0"), val = tensor([1, 1])]; - tensor var_1631_pad_0 = const()[name = tensor("op_1631_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1631_dilations_0 = const()[name = tensor("op_1631_dilations_0"), val = tensor([1, 1])]; - tensor var_1631_groups_0 = const()[name = tensor("op_1631_groups_0"), val = tensor(1)]; - tensor layers_3_feed_forward2_fc2_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23966976))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(24753472))), name = tensor("layers_3_feed_forward2_fc2_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 2048, 1, 1])]; - tensor layers_3_feed_forward2_fc2_inlier_module_bias_to_fp16 = const()[name = tensor("layers_3_feed_forward2_fc2_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(24753664)))]; - tensor var_1631_cast_fp16 = conv(bias = layers_3_feed_forward2_fc2_inlier_module_bias_to_fp16, dilations = var_1631_dilations_0, groups = var_1631_groups_0, pad = var_1631_pad_0, pad_type = var_1631_pad_type_0, strides = var_1631_strides_0, weight = layers_3_feed_forward2_fc2_inlier_module_weight_to_fp16_palettized, x = input_119_cast_fp16)[name = tensor("op_1631_cast_fp16")]; - tensor var_1637_pad_type_0 = const()[name = tensor("op_1637_pad_type_0"), val = tensor("valid")]; - tensor var_1637_strides_0 = const()[name = tensor("op_1637_strides_0"), val = tensor([1, 1])]; - tensor var_1637_pad_0 = const()[name = tensor("op_1637_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1637_dilations_0 = const()[name = tensor("op_1637_dilations_0"), val = tensor([1, 1])]; - tensor var_1637_groups_0 = const()[name = tensor("op_1637_groups_0"), val = tensor(1)]; - tensor layers_3_feed_forward2_fc2_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(24790848))), name = tensor("layers_3_feed_forward2_fc2_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(24754752))), shape = tensor([512, 2048, 1, 1])]; - tensor var_1637_cast_fp16 = conv(dilations = var_1637_dilations_0, groups = var_1637_groups_0, pad = var_1637_pad_0, pad_type = var_1637_pad_type_0, strides = var_1637_strides_0, weight = layers_3_feed_forward2_fc2_outlier_module_weight_to_fp16_sparsified, x = input_119_cast_fp16)[name = tensor("op_1637_cast_fp16")]; - tensor x_25_cast_fp16 = add(x = var_1631_cast_fp16, y = var_1637_cast_fp16)[name = tensor("x_25_cast_fp16")]; - tensor var_1639_to_fp16 = const()[name = tensor("op_1639_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1640_cast_fp16 = mul(x = x_25_cast_fp16, y = var_1639_to_fp16)[name = tensor("op_1640_cast_fp16")]; - tensor inputs_39_cast_fp16 = add(x = inputs_37_cast_fp16, y = var_1640_cast_fp16)[name = tensor("inputs_39_cast_fp16")]; - tensor out_39_axes_0 = const()[name = tensor("out_39_axes_0"), val = tensor([1])]; - tensor var_1650_to_fp16 = const()[name = tensor("op_1650_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_39_cast_fp16 = layer_norm(axes = out_39_axes_0, epsilon = var_1650_to_fp16, x = inputs_39_cast_fp16)[name = tensor("out_39_cast_fp16")]; - tensor inputs_41_gamma_0_to_fp16 = const()[name = tensor("inputs_41_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(24921984)))]; - tensor inputs_41_beta_0_to_fp16 = const()[name = tensor("inputs_41_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(24923072)))]; - tensor inputs_41_epsilon_0_to_fp16 = const()[name = tensor("inputs_41_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor inputs_41_cast_fp16 = batch_norm(beta = inputs_41_beta_0_to_fp16, epsilon = inputs_41_epsilon_0_to_fp16, gamma = inputs_41_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_39_cast_fp16)[name = tensor("inputs_41_cast_fp16")]; - tensor var_1664 = const()[name = tensor("op_1664"), val = tensor(3)]; - tensor out_41_axes_0 = const()[name = tensor("out_41_axes_0"), val = tensor([1])]; - tensor var_1695_to_fp16 = const()[name = tensor("op_1695_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_41_cast_fp16 = layer_norm(axes = out_41_axes_0, epsilon = var_1695_to_fp16, x = inputs_41_cast_fp16)[name = tensor("out_41_cast_fp16")]; - tensor input_121_gamma_0_to_fp16 = const()[name = tensor("input_121_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(24924160)))]; - tensor input_121_beta_0_to_fp16 = const()[name = tensor("input_121_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(24925248)))]; - tensor input_121_epsilon_0_to_fp16 = const()[name = tensor("input_121_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_121_cast_fp16 = batch_norm(beta = input_121_beta_0_to_fp16, epsilon = input_121_epsilon_0_to_fp16, gamma = input_121_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_41_cast_fp16)[name = tensor("input_121_cast_fp16")]; - tensor var_1715_pad_type_0 = const()[name = tensor("op_1715_pad_type_0"), val = tensor("valid")]; - tensor var_1715_strides_0 = const()[name = tensor("op_1715_strides_0"), val = tensor([1, 1])]; - tensor var_1715_pad_0 = const()[name = tensor("op_1715_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1715_dilations_0 = const()[name = tensor("op_1715_dilations_0"), val = tensor([1, 1])]; - tensor var_1715_groups_0 = const()[name = tensor("op_1715_groups_0"), val = tensor(1)]; - tensor layers_4_feed_forward1_fc1_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(24926336))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(25712832))), name = tensor("layers_4_feed_forward1_fc1_inlier_module_weight_to_fp16_palettized"), shape = tensor([2048, 512, 1, 1])]; - tensor layers_4_feed_forward1_fc1_inlier_module_bias_to_fp16 = const()[name = tensor("layers_4_feed_forward1_fc1_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(25713024)))]; - tensor var_1715_cast_fp16 = conv(bias = layers_4_feed_forward1_fc1_inlier_module_bias_to_fp16, dilations = var_1715_dilations_0, groups = var_1715_groups_0, pad = var_1715_pad_0, pad_type = var_1715_pad_type_0, strides = var_1715_strides_0, weight = layers_4_feed_forward1_fc1_inlier_module_weight_to_fp16_palettized, x = input_121_cast_fp16)[name = tensor("op_1715_cast_fp16")]; - tensor var_1721_pad_type_0 = const()[name = tensor("op_1721_pad_type_0"), val = tensor("valid")]; - tensor var_1721_strides_0 = const()[name = tensor("op_1721_strides_0"), val = tensor([1, 1])]; - tensor var_1721_pad_0 = const()[name = tensor("op_1721_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1721_dilations_0 = const()[name = tensor("op_1721_dilations_0"), val = tensor([1, 1])]; - tensor var_1721_groups_0 = const()[name = tensor("op_1721_groups_0"), val = tensor(1)]; - tensor layers_4_feed_forward1_fc1_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(25749248))), name = tensor("layers_4_feed_forward1_fc1_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(25717184))), shape = tensor([2048, 512, 1, 1])]; - tensor var_1721_cast_fp16 = conv(dilations = var_1721_dilations_0, groups = var_1721_groups_0, pad = var_1721_pad_0, pad_type = var_1721_pad_type_0, strides = var_1721_strides_0, weight = layers_4_feed_forward1_fc1_outlier_module_weight_to_fp16_sparsified, x = input_121_cast_fp16)[name = tensor("op_1721_cast_fp16")]; - tensor input_123_cast_fp16 = add(x = var_1715_cast_fp16, y = var_1721_cast_fp16)[name = tensor("input_123_cast_fp16")]; - tensor input_125_cast_fp16 = silu(x = input_123_cast_fp16)[name = tensor("input_125_cast_fp16")]; - tensor var_1732_pad_type_0 = const()[name = tensor("op_1732_pad_type_0"), val = tensor("valid")]; - tensor var_1732_strides_0 = const()[name = tensor("op_1732_strides_0"), val = tensor([1, 1])]; - tensor var_1732_pad_0 = const()[name = tensor("op_1732_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1732_dilations_0 = const()[name = tensor("op_1732_dilations_0"), val = tensor([1, 1])]; - tensor var_1732_groups_0 = const()[name = tensor("op_1732_groups_0"), val = tensor(1)]; - tensor layers_4_feed_forward1_fc2_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(25880384))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(26666880))), name = tensor("layers_4_feed_forward1_fc2_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 2048, 1, 1])]; - tensor layers_4_feed_forward1_fc2_inlier_module_bias_to_fp16 = const()[name = tensor("layers_4_feed_forward1_fc2_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(26667072)))]; - tensor var_1732_cast_fp16 = conv(bias = layers_4_feed_forward1_fc2_inlier_module_bias_to_fp16, dilations = var_1732_dilations_0, groups = var_1732_groups_0, pad = var_1732_pad_0, pad_type = var_1732_pad_type_0, strides = var_1732_strides_0, weight = layers_4_feed_forward1_fc2_inlier_module_weight_to_fp16_palettized, x = input_125_cast_fp16)[name = tensor("op_1732_cast_fp16")]; - tensor var_1738_pad_type_0 = const()[name = tensor("op_1738_pad_type_0"), val = tensor("valid")]; - tensor var_1738_strides_0 = const()[name = tensor("op_1738_strides_0"), val = tensor([1, 1])]; - tensor var_1738_pad_0 = const()[name = tensor("op_1738_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1738_dilations_0 = const()[name = tensor("op_1738_dilations_0"), val = tensor([1, 1])]; - tensor var_1738_groups_0 = const()[name = tensor("op_1738_groups_0"), val = tensor(1)]; - tensor layers_4_feed_forward1_fc2_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(26704896))), name = tensor("layers_4_feed_forward1_fc2_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(26668160))), shape = tensor([512, 2048, 1, 1])]; - tensor var_1738_cast_fp16 = conv(dilations = var_1738_dilations_0, groups = var_1738_groups_0, pad = var_1738_pad_0, pad_type = var_1738_pad_type_0, strides = var_1738_strides_0, weight = layers_4_feed_forward1_fc2_outlier_module_weight_to_fp16_sparsified, x = input_125_cast_fp16)[name = tensor("op_1738_cast_fp16")]; - tensor x_27_cast_fp16 = add(x = var_1732_cast_fp16, y = var_1738_cast_fp16)[name = tensor("x_27_cast_fp16")]; - tensor var_1740_to_fp16 = const()[name = tensor("op_1740_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1741_cast_fp16 = mul(x = x_27_cast_fp16, y = var_1740_to_fp16)[name = tensor("op_1741_cast_fp16")]; - tensor inputs_43_cast_fp16 = add(x = inputs_41_cast_fp16, y = var_1741_cast_fp16)[name = tensor("inputs_43_cast_fp16")]; - tensor out_43_axes_0 = const()[name = tensor("out_43_axes_0"), val = tensor([1])]; - tensor var_1751_to_fp16 = const()[name = tensor("op_1751_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_43_cast_fp16 = layer_norm(axes = out_43_axes_0, epsilon = var_1751_to_fp16, x = inputs_43_cast_fp16)[name = tensor("out_43_cast_fp16")]; - tensor obj_19_gamma_0_to_fp16 = const()[name = tensor("obj_19_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(26836032)))]; - tensor obj_19_beta_0_to_fp16 = const()[name = tensor("obj_19_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(26837120)))]; - tensor obj_19_epsilon_0_to_fp16 = const()[name = tensor("obj_19_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor obj_19_cast_fp16 = batch_norm(beta = obj_19_beta_0_to_fp16, epsilon = obj_19_epsilon_0_to_fp16, gamma = obj_19_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_43_cast_fp16)[name = tensor("obj_19_cast_fp16")]; - tensor var_1776_pad_type_0 = const()[name = tensor("op_1776_pad_type_0"), val = tensor("valid")]; - tensor var_1776_strides_0 = const()[name = tensor("op_1776_strides_0"), val = tensor([1, 1])]; - tensor var_1776_pad_0 = const()[name = tensor("op_1776_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1776_dilations_0 = const()[name = tensor("op_1776_dilations_0"), val = tensor([1, 1])]; - tensor var_1776_groups_0 = const()[name = tensor("op_1776_groups_0"), val = tensor(1)]; - tensor layers_4_self_attn_q_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(26838208))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27034880))), name = tensor("layers_4_self_attn_q_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_4_self_attn_q_proj_inlier_module_bias_to_fp16 = const()[name = tensor("layers_4_self_attn_q_proj_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27035072)))]; - tensor var_1776_cast_fp16 = conv(bias = layers_4_self_attn_q_proj_inlier_module_bias_to_fp16, dilations = var_1776_dilations_0, groups = var_1776_groups_0, pad = var_1776_pad_0, pad_type = var_1776_pad_type_0, strides = var_1776_strides_0, weight = layers_4_self_attn_q_proj_inlier_module_weight_to_fp16_palettized, x = obj_19_cast_fp16)[name = tensor("op_1776_cast_fp16")]; - tensor var_1782_pad_type_0 = const()[name = tensor("op_1782_pad_type_0"), val = tensor("valid")]; - tensor var_1782_strides_0 = const()[name = tensor("op_1782_strides_0"), val = tensor([1, 1])]; - tensor var_1782_pad_0 = const()[name = tensor("op_1782_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1782_dilations_0 = const()[name = tensor("op_1782_dilations_0"), val = tensor([1, 1])]; - tensor var_1782_groups_0 = const()[name = tensor("op_1782_groups_0"), val = tensor(1)]; - tensor layers_4_self_attn_q_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27044608))), name = tensor("layers_4_self_attn_q_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27036160))), shape = tensor([512, 512, 1, 1])]; - tensor var_1782_cast_fp16 = conv(dilations = var_1782_dilations_0, groups = var_1782_groups_0, pad = var_1782_pad_0, pad_type = var_1782_pad_type_0, strides = var_1782_strides_0, weight = layers_4_self_attn_q_proj_outlier_module_weight_to_fp16_sparsified, x = obj_19_cast_fp16)[name = tensor("op_1782_cast_fp16")]; - tensor query_17_cast_fp16 = add(x = var_1776_cast_fp16, y = var_1782_cast_fp16)[name = tensor("query_17_cast_fp16")]; - tensor var_1791_pad_type_0 = const()[name = tensor("op_1791_pad_type_0"), val = tensor("valid")]; - tensor var_1791_strides_0 = const()[name = tensor("op_1791_strides_0"), val = tensor([1, 1])]; - tensor var_1791_pad_0 = const()[name = tensor("op_1791_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1791_dilations_0 = const()[name = tensor("op_1791_dilations_0"), val = tensor([1, 1])]; - tensor var_1791_groups_0 = const()[name = tensor("op_1791_groups_0"), val = tensor(1)]; - tensor layers_4_self_attn_k_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27077440))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27274112))), name = tensor("layers_4_self_attn_k_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor var_1791_cast_fp16 = conv(dilations = var_1791_dilations_0, groups = var_1791_groups_0, pad = var_1791_pad_0, pad_type = var_1791_pad_type_0, strides = var_1791_strides_0, weight = layers_4_self_attn_k_proj_inlier_module_weight_to_fp16_palettized, x = obj_19_cast_fp16)[name = tensor("op_1791_cast_fp16")]; - tensor var_1797_pad_type_0 = const()[name = tensor("op_1797_pad_type_0"), val = tensor("valid")]; - tensor var_1797_strides_0 = const()[name = tensor("op_1797_strides_0"), val = tensor([1, 1])]; - tensor var_1797_pad_0 = const()[name = tensor("op_1797_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1797_dilations_0 = const()[name = tensor("op_1797_dilations_0"), val = tensor([1, 1])]; - tensor var_1797_groups_0 = const()[name = tensor("op_1797_groups_0"), val = tensor(1)]; - tensor layers_4_self_attn_k_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27283968))), name = tensor("layers_4_self_attn_k_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27274304))), shape = tensor([512, 512, 1, 1])]; - tensor var_1797_cast_fp16 = conv(dilations = var_1797_dilations_0, groups = var_1797_groups_0, pad = var_1797_pad_0, pad_type = var_1797_pad_type_0, strides = var_1797_strides_0, weight = layers_4_self_attn_k_proj_outlier_module_weight_to_fp16_sparsified, x = obj_19_cast_fp16)[name = tensor("op_1797_cast_fp16")]; - tensor key_9_cast_fp16 = add(x = var_1791_cast_fp16, y = var_1797_cast_fp16)[name = tensor("key_9_cast_fp16")]; - tensor var_1807_pad_type_0 = const()[name = tensor("op_1807_pad_type_0"), val = tensor("valid")]; - tensor var_1807_strides_0 = const()[name = tensor("op_1807_strides_0"), val = tensor([1, 1])]; - tensor var_1807_pad_0 = const()[name = tensor("op_1807_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1807_dilations_0 = const()[name = tensor("op_1807_dilations_0"), val = tensor([1, 1])]; - tensor var_1807_groups_0 = const()[name = tensor("op_1807_groups_0"), val = tensor(1)]; - tensor layers_4_self_attn_v_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27316800))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27513472))), name = tensor("layers_4_self_attn_v_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_4_self_attn_v_proj_inlier_module_bias_to_fp16 = const()[name = tensor("layers_4_self_attn_v_proj_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27513664)))]; - tensor var_1807_cast_fp16 = conv(bias = layers_4_self_attn_v_proj_inlier_module_bias_to_fp16, dilations = var_1807_dilations_0, groups = var_1807_groups_0, pad = var_1807_pad_0, pad_type = var_1807_pad_type_0, strides = var_1807_strides_0, weight = layers_4_self_attn_v_proj_inlier_module_weight_to_fp16_palettized, x = obj_19_cast_fp16)[name = tensor("op_1807_cast_fp16")]; - tensor var_1813_pad_type_0 = const()[name = tensor("op_1813_pad_type_0"), val = tensor("valid")]; - tensor var_1813_strides_0 = const()[name = tensor("op_1813_strides_0"), val = tensor([1, 1])]; - tensor var_1813_pad_0 = const()[name = tensor("op_1813_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1813_dilations_0 = const()[name = tensor("op_1813_dilations_0"), val = tensor([1, 1])]; - tensor var_1813_groups_0 = const()[name = tensor("op_1813_groups_0"), val = tensor(1)]; - tensor layers_4_self_attn_v_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27522816))), name = tensor("layers_4_self_attn_v_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27514752))), shape = tensor([512, 512, 1, 1])]; - tensor var_1813_cast_fp16 = conv(dilations = var_1813_dilations_0, groups = var_1813_groups_0, pad = var_1813_pad_0, pad_type = var_1813_pad_type_0, strides = var_1813_strides_0, weight = layers_4_self_attn_v_proj_outlier_module_weight_to_fp16_sparsified, x = obj_19_cast_fp16)[name = tensor("op_1813_cast_fp16")]; - tensor value_9_cast_fp16 = add(x = var_1807_cast_fp16, y = var_1813_cast_fp16)[name = tensor("value_9_cast_fp16")]; - tensor var_1816_to_fp16 = const()[name = tensor("op_1816_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27555648)))]; - tensor query_19_cast_fp16 = add(x = query_17_cast_fp16, y = var_1816_to_fp16)[name = tensor("query_19_cast_fp16")]; - tensor var_1819_to_fp16 = const()[name = tensor("op_1819_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27556736)))]; - tensor q_with_bias_v_9_cast_fp16 = add(x = query_17_cast_fp16, y = var_1819_to_fp16)[name = tensor("q_with_bias_v_9_cast_fp16")]; - tensor var_1829_pad_type_0 = const()[name = tensor("op_1829_pad_type_0"), val = tensor("valid")]; - tensor var_1829_strides_0 = const()[name = tensor("op_1829_strides_0"), val = tensor([1, 1])]; - tensor var_1829_pad_0 = const()[name = tensor("op_1829_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1829_dilations_0 = const()[name = tensor("op_1829_dilations_0"), val = tensor([1, 1])]; - tensor var_1829_groups_0 = const()[name = tensor("op_1829_groups_0"), val = tensor(1)]; - tensor layers_4_self_attn_linear_pos_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27557824))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27754496))), name = tensor("layers_4_self_attn_linear_pos_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor var_1829_cast_fp16 = conv(bias = input_17_mean_0_to_fp16, dilations = var_1829_dilations_0, groups = var_1829_groups_0, pad = var_1829_pad_0, pad_type = var_1829_pad_type_0, strides = var_1829_strides_0, weight = layers_4_self_attn_linear_pos_inlier_module_weight_to_fp16_palettized, x = obj_3_cast_fp16)[name = tensor("op_1829_cast_fp16")]; - tensor var_1835_pad_type_0 = const()[name = tensor("op_1835_pad_type_0"), val = tensor("valid")]; - tensor var_1835_strides_0 = const()[name = tensor("op_1835_strides_0"), val = tensor([1, 1])]; - tensor var_1835_pad_0 = const()[name = tensor("op_1835_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1835_dilations_0 = const()[name = tensor("op_1835_dilations_0"), val = tensor([1, 1])]; - tensor var_1835_groups_0 = const()[name = tensor("op_1835_groups_0"), val = tensor(1)]; - tensor layers_4_self_attn_linear_pos_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27776128))), name = tensor("layers_4_self_attn_linear_pos_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27754688))), shape = tensor([512, 512, 1, 1])]; - tensor var_1835_cast_fp16 = conv(dilations = var_1835_dilations_0, groups = var_1835_groups_0, pad = var_1835_pad_0, pad_type = var_1835_pad_type_0, strides = var_1835_strides_0, weight = layers_4_self_attn_linear_pos_outlier_module_weight_to_fp16_sparsified, x = obj_3_cast_fp16)[name = tensor("op_1835_cast_fp16")]; - tensor p_9_cast_fp16 = add(x = var_1829_cast_fp16, y = var_1835_cast_fp16)[name = tensor("p_9_cast_fp16")]; - tensor var_1839 = const()[name = tensor("op_1839"), val = tensor([1, 8, 64, 188])]; - tensor var_1840_cast_fp16 = reshape(shape = var_1839, x = q_with_bias_v_9_cast_fp16)[name = tensor("op_1840_cast_fp16")]; - tensor var_1841 = const()[name = tensor("op_1841"), val = tensor([1, 8, 64, -1])]; - tensor var_1842_cast_fp16 = reshape(shape = var_1841, x = p_9_cast_fp16)[name = tensor("op_1842_cast_fp16")]; - tensor matrix_bd_33_transpose_x_0 = const()[name = tensor("matrix_bd_33_transpose_x_0"), val = tensor(true)]; - tensor matrix_bd_33_transpose_y_0 = const()[name = tensor("matrix_bd_33_transpose_y_0"), val = tensor(false)]; - tensor matrix_bd_33_cast_fp16 = matmul(transpose_x = matrix_bd_33_transpose_x_0, transpose_y = matrix_bd_33_transpose_y_0, x = var_1840_cast_fp16, y = var_1842_cast_fp16)[name = tensor("matrix_bd_33_cast_fp16")]; - tensor matrix_bd_35_pad_0 = const()[name = tensor("matrix_bd_35_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor matrix_bd_35_mode_0 = const()[name = tensor("matrix_bd_35_mode_0"), val = tensor("constant")]; - tensor const_54_to_fp16 = const()[name = tensor("const_54_to_fp16"), val = tensor(0x0p+0)]; - tensor matrix_bd_35_cast_fp16 = pad(constant_val = const_54_to_fp16, mode = matrix_bd_35_mode_0, pad = matrix_bd_35_pad_0, x = matrix_bd_33_cast_fp16)[name = tensor("matrix_bd_35_cast_fp16")]; - tensor var_1851 = const()[name = tensor("op_1851"), val = tensor([1, 8, -1, 188])]; - tensor matrix_bd_37_cast_fp16 = reshape(shape = var_1851, x = matrix_bd_35_cast_fp16)[name = tensor("matrix_bd_37_cast_fp16")]; - tensor var_1855_begin_0 = const()[name = tensor("op_1855_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_1855_end_0 = const()[name = tensor("op_1855_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_1855_end_mask_0 = const()[name = tensor("op_1855_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_1855_cast_fp16 = slice_by_index(begin = var_1855_begin_0, end = var_1855_end_0, end_mask = var_1855_end_mask_0, x = matrix_bd_37_cast_fp16)[name = tensor("op_1855_cast_fp16")]; - tensor var_1856 = const()[name = tensor("op_1856"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_39_cast_fp16 = reshape(shape = var_1856, x = var_1855_cast_fp16)[name = tensor("matrix_bd_39_cast_fp16")]; - tensor var_1861_begin_0 = const()[name = tensor("op_1861_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1861_end_0 = const()[name = tensor("op_1861_end_0"), val = tensor([1, 8, 188, 188])]; - tensor var_1861_end_mask_0 = const()[name = tensor("op_1861_end_mask_0"), val = tensor([true, true, true, false])]; - tensor var_1861_cast_fp16 = slice_by_index(begin = var_1861_begin_0, end = var_1861_end_0, end_mask = var_1861_end_mask_0, x = matrix_bd_39_cast_fp16)[name = tensor("op_1861_cast_fp16")]; - tensor var_1862_to_fp16 = const()[name = tensor("op_1862_to_fp16"), val = tensor(0x1p-3)]; - tensor qk_mask_9_cast_fp16 = mul(x = var_1861_cast_fp16, y = var_1862_to_fp16)[name = tensor("qk_mask_9_cast_fp16")]; - tensor var_1866 = const()[name = tensor("op_1866"), val = tensor([1, 8, 64, 188])]; - tensor mh_q_9_cast_fp16 = reshape(shape = var_1866, x = query_19_cast_fp16)[name = tensor("mh_q_9_cast_fp16")]; - tensor var_1868_to_fp16 = const()[name = tensor("op_1868_to_fp16"), val = tensor(0x1p-3)]; - tensor var_1869_cast_fp16 = mul(x = mh_q_9_cast_fp16, y = var_1868_to_fp16)[name = tensor("op_1869_cast_fp16")]; - tensor var_1872 = const()[name = tensor("op_1872"), val = tensor([1, 8, 64, 188])]; - tensor var_1873_cast_fp16 = reshape(shape = var_1872, x = key_9_cast_fp16)[name = tensor("op_1873_cast_fp16")]; - tensor mh_w_17_transpose_x_0 = const()[name = tensor("mh_w_17_transpose_x_0"), val = tensor(true)]; - tensor mh_w_17_transpose_y_0 = const()[name = tensor("mh_w_17_transpose_y_0"), val = tensor(false)]; - tensor mh_w_17_cast_fp16 = matmul(transpose_x = mh_w_17_transpose_x_0, transpose_y = mh_w_17_transpose_y_0, x = var_1869_cast_fp16, y = var_1873_cast_fp16)[name = tensor("mh_w_17_cast_fp16")]; - tensor mh_w_19_cast_fp16 = add(x = mh_w_17_cast_fp16, y = qk_mask_9_cast_fp16)[name = tensor("mh_w_19_cast_fp16")]; - tensor var_1877_cast_fp16 = softmax(axis = var_1664, x = mh_w_19_cast_fp16)[name = tensor("op_1877_cast_fp16")]; - tensor var_1878 = const()[name = tensor("op_1878"), val = tensor([1, 8, 64, 188])]; - tensor var_1879_cast_fp16 = reshape(shape = var_1878, x = value_9_cast_fp16)[name = tensor("op_1879_cast_fp16")]; - tensor attn_9_transpose_x_0 = const()[name = tensor("attn_9_transpose_x_0"), val = tensor(false)]; - tensor attn_9_transpose_y_0 = const()[name = tensor("attn_9_transpose_y_0"), val = tensor(true)]; - tensor attn_9_cast_fp16 = matmul(transpose_x = attn_9_transpose_x_0, transpose_y = attn_9_transpose_y_0, x = var_1879_cast_fp16, y = var_1877_cast_fp16)[name = tensor("attn_9_cast_fp16")]; - tensor var_1882 = const()[name = tensor("op_1882"), val = tensor([1, 512, 1, 188])]; - tensor input_127_cast_fp16 = reshape(shape = var_1882, x = attn_9_cast_fp16)[name = tensor("input_127_cast_fp16")]; - tensor var_1892_pad_type_0 = const()[name = tensor("op_1892_pad_type_0"), val = tensor("valid")]; - tensor var_1892_strides_0 = const()[name = tensor("op_1892_strides_0"), val = tensor([1, 1])]; - tensor var_1892_pad_0 = const()[name = tensor("op_1892_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1892_dilations_0 = const()[name = tensor("op_1892_dilations_0"), val = tensor([1, 1])]; - tensor var_1892_groups_0 = const()[name = tensor("op_1892_groups_0"), val = tensor(1)]; - tensor layers_4_self_attn_o_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27808960))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(28005632))), name = tensor("layers_4_self_attn_o_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_4_self_attn_o_proj_inlier_module_bias_to_fp16 = const()[name = tensor("layers_4_self_attn_o_proj_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(28005824)))]; - tensor var_1892_cast_fp16 = conv(bias = layers_4_self_attn_o_proj_inlier_module_bias_to_fp16, dilations = var_1892_dilations_0, groups = var_1892_groups_0, pad = var_1892_pad_0, pad_type = var_1892_pad_type_0, strides = var_1892_strides_0, weight = layers_4_self_attn_o_proj_inlier_module_weight_to_fp16_palettized, x = input_127_cast_fp16)[name = tensor("op_1892_cast_fp16")]; - tensor var_1898_pad_type_0 = const()[name = tensor("op_1898_pad_type_0"), val = tensor("valid")]; - tensor var_1898_strides_0 = const()[name = tensor("op_1898_strides_0"), val = tensor([1, 1])]; - tensor var_1898_pad_0 = const()[name = tensor("op_1898_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1898_dilations_0 = const()[name = tensor("op_1898_dilations_0"), val = tensor([1, 1])]; - tensor var_1898_groups_0 = const()[name = tensor("op_1898_groups_0"), val = tensor(1)]; - tensor layers_4_self_attn_o_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(28014848))), name = tensor("layers_4_self_attn_o_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(28006912))), shape = tensor([512, 512, 1, 1])]; - tensor var_1898_cast_fp16 = conv(dilations = var_1898_dilations_0, groups = var_1898_groups_0, pad = var_1898_pad_0, pad_type = var_1898_pad_type_0, strides = var_1898_strides_0, weight = layers_4_self_attn_o_proj_outlier_module_weight_to_fp16_sparsified, x = input_127_cast_fp16)[name = tensor("op_1898_cast_fp16")]; - tensor obj_21_cast_fp16 = add(x = var_1892_cast_fp16, y = var_1898_cast_fp16)[name = tensor("obj_21_cast_fp16")]; - tensor inputs_45_cast_fp16 = add(x = inputs_43_cast_fp16, y = obj_21_cast_fp16)[name = tensor("inputs_45_cast_fp16")]; - tensor out_45_axes_0 = const()[name = tensor("out_45_axes_0"), val = tensor([1])]; - tensor var_1909_to_fp16 = const()[name = tensor("op_1909_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_45_cast_fp16 = layer_norm(axes = out_45_axes_0, epsilon = var_1909_to_fp16, x = inputs_45_cast_fp16)[name = tensor("out_45_cast_fp16")]; - tensor input_129_gamma_0_to_fp16 = const()[name = tensor("input_129_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(28047680)))]; - tensor input_129_beta_0_to_fp16 = const()[name = tensor("input_129_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(28048768)))]; - tensor input_129_epsilon_0_to_fp16 = const()[name = tensor("input_129_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_129_cast_fp16 = batch_norm(beta = input_129_beta_0_to_fp16, epsilon = input_129_epsilon_0_to_fp16, gamma = input_129_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_45_cast_fp16)[name = tensor("input_129_cast_fp16")]; - tensor var_1931_pad_type_0 = const()[name = tensor("op_1931_pad_type_0"), val = tensor("valid")]; - tensor var_1931_strides_0 = const()[name = tensor("op_1931_strides_0"), val = tensor([1, 1])]; - tensor var_1931_pad_0 = const()[name = tensor("op_1931_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1931_dilations_0 = const()[name = tensor("op_1931_dilations_0"), val = tensor([1, 1])]; - tensor var_1931_groups_0 = const()[name = tensor("op_1931_groups_0"), val = tensor(1)]; - tensor layers_4_conv_pointwise_conv1_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(28049856))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(28443136))), name = tensor("layers_4_conv_pointwise_conv1_inlier_module_weight_to_fp16_palettized"), shape = tensor([1024, 512, 1, 1])]; - tensor layers_4_conv_pointwise_conv1_inlier_module_bias_to_fp16 = const()[name = tensor("layers_4_conv_pointwise_conv1_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(28443328)))]; - tensor var_1931_cast_fp16 = conv(bias = layers_4_conv_pointwise_conv1_inlier_module_bias_to_fp16, dilations = var_1931_dilations_0, groups = var_1931_groups_0, pad = var_1931_pad_0, pad_type = var_1931_pad_type_0, strides = var_1931_strides_0, weight = layers_4_conv_pointwise_conv1_inlier_module_weight_to_fp16_palettized, x = input_129_cast_fp16)[name = tensor("op_1931_cast_fp16")]; - tensor var_1937_pad_type_0 = const()[name = tensor("op_1937_pad_type_0"), val = tensor("valid")]; - tensor var_1937_strides_0 = const()[name = tensor("op_1937_strides_0"), val = tensor([1, 1])]; - tensor var_1937_pad_0 = const()[name = tensor("op_1937_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1937_dilations_0 = const()[name = tensor("op_1937_dilations_0"), val = tensor([1, 1])]; - tensor var_1937_groups_0 = const()[name = tensor("op_1937_groups_0"), val = tensor(1)]; - tensor layers_4_conv_pointwise_conv1_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(28463360))), name = tensor("layers_4_conv_pointwise_conv1_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(28445440))), shape = tensor([1024, 512, 1, 1])]; - tensor var_1937_cast_fp16 = conv(dilations = var_1937_dilations_0, groups = var_1937_groups_0, pad = var_1937_pad_0, pad_type = var_1937_pad_type_0, strides = var_1937_strides_0, weight = layers_4_conv_pointwise_conv1_outlier_module_weight_to_fp16_sparsified, x = input_129_cast_fp16)[name = tensor("op_1937_cast_fp16")]; - tensor input_131_cast_fp16 = add(x = var_1931_cast_fp16, y = var_1937_cast_fp16)[name = tensor("input_131_cast_fp16")]; - tensor input_133_split_num_splits_0 = const()[name = tensor("input_133_split_num_splits_0"), val = tensor(2)]; - tensor input_133_split_axis_0 = const()[name = tensor("input_133_split_axis_0"), val = tensor(1)]; - tensor input_133_split_cast_fp16_0, tensor input_133_split_cast_fp16_1 = split(axis = input_133_split_axis_0, num_splits = input_133_split_num_splits_0, x = input_131_cast_fp16)[name = tensor("input_133_split_cast_fp16")]; - tensor input_133_split_1_sigmoid_cast_fp16 = sigmoid(x = input_133_split_cast_fp16_1)[name = tensor("input_133_split_1_sigmoid_cast_fp16")]; - tensor input_133_cast_fp16 = mul(x = input_133_split_cast_fp16_0, y = input_133_split_1_sigmoid_cast_fp16)[name = tensor("input_133_cast_fp16")]; - tensor input_135_pad_type_0 = const()[name = tensor("input_135_pad_type_0"), val = tensor("custom")]; - tensor input_135_pad_0 = const()[name = tensor("input_135_pad_0"), val = tensor([0, 0, 4, 4])]; - tensor input_135_groups_0 = const()[name = tensor("input_135_groups_0"), val = tensor(512)]; - tensor input_135_strides_0 = const()[name = tensor("input_135_strides_0"), val = tensor([1, 1])]; - tensor input_135_dilations_0 = const()[name = tensor("input_135_dilations_0"), val = tensor([1, 1])]; - tensor const_199_to_fp16 = const()[name = tensor("const_199_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(28528960)))]; - tensor const_200_to_fp16 = const()[name = tensor("const_200_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(28538240)))]; - tensor input_137_cast_fp16 = conv(bias = const_200_to_fp16, dilations = input_135_dilations_0, groups = input_135_groups_0, pad = input_135_pad_0, pad_type = input_135_pad_type_0, strides = input_135_strides_0, weight = const_199_to_fp16, x = input_133_cast_fp16)[name = tensor("input_137_cast_fp16")]; - tensor input_139_cast_fp16 = silu(x = input_137_cast_fp16)[name = tensor("input_139_cast_fp16")]; - tensor var_1961_pad_type_0 = const()[name = tensor("op_1961_pad_type_0"), val = tensor("valid")]; - tensor var_1961_strides_0 = const()[name = tensor("op_1961_strides_0"), val = tensor([1, 1])]; - tensor var_1961_pad_0 = const()[name = tensor("op_1961_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1961_dilations_0 = const()[name = tensor("op_1961_dilations_0"), val = tensor([1, 1])]; - tensor var_1961_groups_0 = const()[name = tensor("op_1961_groups_0"), val = tensor(1)]; - tensor layers_4_conv_pointwise_conv2_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(28539328))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(28736000))), name = tensor("layers_4_conv_pointwise_conv2_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_4_conv_pointwise_conv2_inlier_module_bias_to_fp16 = const()[name = tensor("layers_4_conv_pointwise_conv2_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(28736192)))]; - tensor var_1961_cast_fp16 = conv(bias = layers_4_conv_pointwise_conv2_inlier_module_bias_to_fp16, dilations = var_1961_dilations_0, groups = var_1961_groups_0, pad = var_1961_pad_0, pad_type = var_1961_pad_type_0, strides = var_1961_strides_0, weight = layers_4_conv_pointwise_conv2_inlier_module_weight_to_fp16_palettized, x = input_139_cast_fp16)[name = tensor("op_1961_cast_fp16")]; - tensor var_1967_pad_type_0 = const()[name = tensor("op_1967_pad_type_0"), val = tensor("valid")]; - tensor var_1967_strides_0 = const()[name = tensor("op_1967_strides_0"), val = tensor([1, 1])]; - tensor var_1967_pad_0 = const()[name = tensor("op_1967_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1967_dilations_0 = const()[name = tensor("op_1967_dilations_0"), val = tensor([1, 1])]; - tensor var_1967_groups_0 = const()[name = tensor("op_1967_groups_0"), val = tensor(1)]; - tensor layers_4_conv_pointwise_conv2_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(28745856))), name = tensor("layers_4_conv_pointwise_conv2_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(28737280))), shape = tensor([512, 512, 1, 1])]; - tensor var_1967_cast_fp16 = conv(dilations = var_1967_dilations_0, groups = var_1967_groups_0, pad = var_1967_pad_0, pad_type = var_1967_pad_type_0, strides = var_1967_strides_0, weight = layers_4_conv_pointwise_conv2_outlier_module_weight_to_fp16_sparsified, x = input_139_cast_fp16)[name = tensor("op_1967_cast_fp16")]; - tensor x_29_cast_fp16 = add(x = var_1961_cast_fp16, y = var_1967_cast_fp16)[name = tensor("x_29_cast_fp16")]; - tensor inputs_47_cast_fp16 = add(x = inputs_45_cast_fp16, y = x_29_cast_fp16)[name = tensor("inputs_47_cast_fp16")]; - tensor out_47_axes_0 = const()[name = tensor("out_47_axes_0"), val = tensor([1])]; - tensor var_1978_to_fp16 = const()[name = tensor("op_1978_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_47_cast_fp16 = layer_norm(axes = out_47_axes_0, epsilon = var_1978_to_fp16, x = inputs_47_cast_fp16)[name = tensor("out_47_cast_fp16")]; - tensor input_141_gamma_0_to_fp16 = const()[name = tensor("input_141_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(28778688)))]; - tensor input_141_beta_0_to_fp16 = const()[name = tensor("input_141_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(28779776)))]; - tensor input_141_epsilon_0_to_fp16 = const()[name = tensor("input_141_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_141_cast_fp16 = batch_norm(beta = input_141_beta_0_to_fp16, epsilon = input_141_epsilon_0_to_fp16, gamma = input_141_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_47_cast_fp16)[name = tensor("input_141_cast_fp16")]; - tensor var_1998_pad_type_0 = const()[name = tensor("op_1998_pad_type_0"), val = tensor("valid")]; - tensor var_1998_strides_0 = const()[name = tensor("op_1998_strides_0"), val = tensor([1, 1])]; - tensor var_1998_pad_0 = const()[name = tensor("op_1998_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_1998_dilations_0 = const()[name = tensor("op_1998_dilations_0"), val = tensor([1, 1])]; - tensor var_1998_groups_0 = const()[name = tensor("op_1998_groups_0"), val = tensor(1)]; - tensor layers_4_feed_forward2_fc1_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(28780864))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(29567360))), name = tensor("layers_4_feed_forward2_fc1_inlier_module_weight_to_fp16_palettized"), shape = tensor([2048, 512, 1, 1])]; - tensor layers_4_feed_forward2_fc1_inlier_module_bias_to_fp16 = const()[name = tensor("layers_4_feed_forward2_fc1_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(29567552)))]; - tensor var_1998_cast_fp16 = conv(bias = layers_4_feed_forward2_fc1_inlier_module_bias_to_fp16, dilations = var_1998_dilations_0, groups = var_1998_groups_0, pad = var_1998_pad_0, pad_type = var_1998_pad_type_0, strides = var_1998_strides_0, weight = layers_4_feed_forward2_fc1_inlier_module_weight_to_fp16_palettized, x = input_141_cast_fp16)[name = tensor("op_1998_cast_fp16")]; - tensor var_2004_pad_type_0 = const()[name = tensor("op_2004_pad_type_0"), val = tensor("valid")]; - tensor var_2004_strides_0 = const()[name = tensor("op_2004_strides_0"), val = tensor([1, 1])]; - tensor var_2004_pad_0 = const()[name = tensor("op_2004_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2004_dilations_0 = const()[name = tensor("op_2004_dilations_0"), val = tensor([1, 1])]; - tensor var_2004_groups_0 = const()[name = tensor("op_2004_groups_0"), val = tensor(1)]; - tensor layers_4_feed_forward2_fc1_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(29603136))), name = tensor("layers_4_feed_forward2_fc1_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(29571712))), shape = tensor([2048, 512, 1, 1])]; - tensor var_2004_cast_fp16 = conv(dilations = var_2004_dilations_0, groups = var_2004_groups_0, pad = var_2004_pad_0, pad_type = var_2004_pad_type_0, strides = var_2004_strides_0, weight = layers_4_feed_forward2_fc1_outlier_module_weight_to_fp16_sparsified, x = input_141_cast_fp16)[name = tensor("op_2004_cast_fp16")]; - tensor input_143_cast_fp16 = add(x = var_1998_cast_fp16, y = var_2004_cast_fp16)[name = tensor("input_143_cast_fp16")]; - tensor input_145_cast_fp16 = silu(x = input_143_cast_fp16)[name = tensor("input_145_cast_fp16")]; - tensor var_2015_pad_type_0 = const()[name = tensor("op_2015_pad_type_0"), val = tensor("valid")]; - tensor var_2015_strides_0 = const()[name = tensor("op_2015_strides_0"), val = tensor([1, 1])]; - tensor var_2015_pad_0 = const()[name = tensor("op_2015_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2015_dilations_0 = const()[name = tensor("op_2015_dilations_0"), val = tensor([1, 1])]; - tensor var_2015_groups_0 = const()[name = tensor("op_2015_groups_0"), val = tensor(1)]; - tensor layers_4_feed_forward2_fc2_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(29734272))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(30520768))), name = tensor("layers_4_feed_forward2_fc2_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 2048, 1, 1])]; - tensor layers_4_feed_forward2_fc2_inlier_module_bias_to_fp16 = const()[name = tensor("layers_4_feed_forward2_fc2_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(30520960)))]; - tensor var_2015_cast_fp16 = conv(bias = layers_4_feed_forward2_fc2_inlier_module_bias_to_fp16, dilations = var_2015_dilations_0, groups = var_2015_groups_0, pad = var_2015_pad_0, pad_type = var_2015_pad_type_0, strides = var_2015_strides_0, weight = layers_4_feed_forward2_fc2_inlier_module_weight_to_fp16_palettized, x = input_145_cast_fp16)[name = tensor("op_2015_cast_fp16")]; - tensor var_2021_pad_type_0 = const()[name = tensor("op_2021_pad_type_0"), val = tensor("valid")]; - tensor var_2021_strides_0 = const()[name = tensor("op_2021_strides_0"), val = tensor([1, 1])]; - tensor var_2021_pad_0 = const()[name = tensor("op_2021_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2021_dilations_0 = const()[name = tensor("op_2021_dilations_0"), val = tensor([1, 1])]; - tensor var_2021_groups_0 = const()[name = tensor("op_2021_groups_0"), val = tensor(1)]; - tensor layers_4_feed_forward2_fc2_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(30556864))), name = tensor("layers_4_feed_forward2_fc2_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(30522048))), shape = tensor([512, 2048, 1, 1])]; - tensor var_2021_cast_fp16 = conv(dilations = var_2021_dilations_0, groups = var_2021_groups_0, pad = var_2021_pad_0, pad_type = var_2021_pad_type_0, strides = var_2021_strides_0, weight = layers_4_feed_forward2_fc2_outlier_module_weight_to_fp16_sparsified, x = input_145_cast_fp16)[name = tensor("op_2021_cast_fp16")]; - tensor x_31_cast_fp16 = add(x = var_2015_cast_fp16, y = var_2021_cast_fp16)[name = tensor("x_31_cast_fp16")]; - tensor var_2023_to_fp16 = const()[name = tensor("op_2023_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2024_cast_fp16 = mul(x = x_31_cast_fp16, y = var_2023_to_fp16)[name = tensor("op_2024_cast_fp16")]; - tensor inputs_49_cast_fp16 = add(x = inputs_47_cast_fp16, y = var_2024_cast_fp16)[name = tensor("inputs_49_cast_fp16")]; - tensor out_49_axes_0 = const()[name = tensor("out_49_axes_0"), val = tensor([1])]; - tensor var_2034_to_fp16 = const()[name = tensor("op_2034_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_49_cast_fp16 = layer_norm(axes = out_49_axes_0, epsilon = var_2034_to_fp16, x = inputs_49_cast_fp16)[name = tensor("out_49_cast_fp16")]; - tensor inputs_51_gamma_0_to_fp16 = const()[name = tensor("inputs_51_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(30688000)))]; - tensor inputs_51_beta_0_to_fp16 = const()[name = tensor("inputs_51_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(30689088)))]; - tensor inputs_51_epsilon_0_to_fp16 = const()[name = tensor("inputs_51_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor inputs_51_cast_fp16 = batch_norm(beta = inputs_51_beta_0_to_fp16, epsilon = inputs_51_epsilon_0_to_fp16, gamma = inputs_51_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_49_cast_fp16)[name = tensor("inputs_51_cast_fp16")]; - tensor var_2048 = const()[name = tensor("op_2048"), val = tensor(3)]; - tensor out_51_axes_0 = const()[name = tensor("out_51_axes_0"), val = tensor([1])]; - tensor var_2079_to_fp16 = const()[name = tensor("op_2079_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_51_cast_fp16 = layer_norm(axes = out_51_axes_0, epsilon = var_2079_to_fp16, x = inputs_51_cast_fp16)[name = tensor("out_51_cast_fp16")]; - tensor input_147_gamma_0_to_fp16 = const()[name = tensor("input_147_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(30690176)))]; - tensor input_147_beta_0_to_fp16 = const()[name = tensor("input_147_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(30691264)))]; - tensor input_147_epsilon_0_to_fp16 = const()[name = tensor("input_147_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_147_cast_fp16 = batch_norm(beta = input_147_beta_0_to_fp16, epsilon = input_147_epsilon_0_to_fp16, gamma = input_147_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_51_cast_fp16)[name = tensor("input_147_cast_fp16")]; - tensor var_2099_pad_type_0 = const()[name = tensor("op_2099_pad_type_0"), val = tensor("valid")]; - tensor var_2099_strides_0 = const()[name = tensor("op_2099_strides_0"), val = tensor([1, 1])]; - tensor var_2099_pad_0 = const()[name = tensor("op_2099_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2099_dilations_0 = const()[name = tensor("op_2099_dilations_0"), val = tensor([1, 1])]; - tensor var_2099_groups_0 = const()[name = tensor("op_2099_groups_0"), val = tensor(1)]; - tensor layers_5_feed_forward1_fc1_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(30692352))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(31478848))), name = tensor("layers_5_feed_forward1_fc1_inlier_module_weight_to_fp16_palettized"), shape = tensor([2048, 512, 1, 1])]; - tensor layers_5_feed_forward1_fc1_inlier_module_bias_to_fp16 = const()[name = tensor("layers_5_feed_forward1_fc1_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(31479040)))]; - tensor var_2099_cast_fp16 = conv(bias = layers_5_feed_forward1_fc1_inlier_module_bias_to_fp16, dilations = var_2099_dilations_0, groups = var_2099_groups_0, pad = var_2099_pad_0, pad_type = var_2099_pad_type_0, strides = var_2099_strides_0, weight = layers_5_feed_forward1_fc1_inlier_module_weight_to_fp16_palettized, x = input_147_cast_fp16)[name = tensor("op_2099_cast_fp16")]; - tensor var_2105_pad_type_0 = const()[name = tensor("op_2105_pad_type_0"), val = tensor("valid")]; - tensor var_2105_strides_0 = const()[name = tensor("op_2105_strides_0"), val = tensor([1, 1])]; - tensor var_2105_pad_0 = const()[name = tensor("op_2105_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2105_dilations_0 = const()[name = tensor("op_2105_dilations_0"), val = tensor([1, 1])]; - tensor var_2105_groups_0 = const()[name = tensor("op_2105_groups_0"), val = tensor(1)]; - tensor layers_5_feed_forward1_fc1_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(31515392))), name = tensor("layers_5_feed_forward1_fc1_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(31483200))), shape = tensor([2048, 512, 1, 1])]; - tensor var_2105_cast_fp16 = conv(dilations = var_2105_dilations_0, groups = var_2105_groups_0, pad = var_2105_pad_0, pad_type = var_2105_pad_type_0, strides = var_2105_strides_0, weight = layers_5_feed_forward1_fc1_outlier_module_weight_to_fp16_sparsified, x = input_147_cast_fp16)[name = tensor("op_2105_cast_fp16")]; - tensor input_149_cast_fp16 = add(x = var_2099_cast_fp16, y = var_2105_cast_fp16)[name = tensor("input_149_cast_fp16")]; - tensor input_151_cast_fp16 = silu(x = input_149_cast_fp16)[name = tensor("input_151_cast_fp16")]; - tensor var_2116_pad_type_0 = const()[name = tensor("op_2116_pad_type_0"), val = tensor("valid")]; - tensor var_2116_strides_0 = const()[name = tensor("op_2116_strides_0"), val = tensor([1, 1])]; - tensor var_2116_pad_0 = const()[name = tensor("op_2116_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2116_dilations_0 = const()[name = tensor("op_2116_dilations_0"), val = tensor([1, 1])]; - tensor var_2116_groups_0 = const()[name = tensor("op_2116_groups_0"), val = tensor(1)]; - tensor layers_5_feed_forward1_fc2_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(31646528))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32433024))), name = tensor("layers_5_feed_forward1_fc2_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 2048, 1, 1])]; - tensor layers_5_feed_forward1_fc2_inlier_module_bias_to_fp16 = const()[name = tensor("layers_5_feed_forward1_fc2_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32433216)))]; - tensor var_2116_cast_fp16 = conv(bias = layers_5_feed_forward1_fc2_inlier_module_bias_to_fp16, dilations = var_2116_dilations_0, groups = var_2116_groups_0, pad = var_2116_pad_0, pad_type = var_2116_pad_type_0, strides = var_2116_strides_0, weight = layers_5_feed_forward1_fc2_inlier_module_weight_to_fp16_palettized, x = input_151_cast_fp16)[name = tensor("op_2116_cast_fp16")]; - tensor var_2122_pad_type_0 = const()[name = tensor("op_2122_pad_type_0"), val = tensor("valid")]; - tensor var_2122_strides_0 = const()[name = tensor("op_2122_strides_0"), val = tensor([1, 1])]; - tensor var_2122_pad_0 = const()[name = tensor("op_2122_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2122_dilations_0 = const()[name = tensor("op_2122_dilations_0"), val = tensor([1, 1])]; - tensor var_2122_groups_0 = const()[name = tensor("op_2122_groups_0"), val = tensor(1)]; - tensor layers_5_feed_forward1_fc2_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32470784))), name = tensor("layers_5_feed_forward1_fc2_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32434304))), shape = tensor([512, 2048, 1, 1])]; - tensor var_2122_cast_fp16 = conv(dilations = var_2122_dilations_0, groups = var_2122_groups_0, pad = var_2122_pad_0, pad_type = var_2122_pad_type_0, strides = var_2122_strides_0, weight = layers_5_feed_forward1_fc2_outlier_module_weight_to_fp16_sparsified, x = input_151_cast_fp16)[name = tensor("op_2122_cast_fp16")]; - tensor x_33_cast_fp16 = add(x = var_2116_cast_fp16, y = var_2122_cast_fp16)[name = tensor("x_33_cast_fp16")]; - tensor var_2124_to_fp16 = const()[name = tensor("op_2124_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2125_cast_fp16 = mul(x = x_33_cast_fp16, y = var_2124_to_fp16)[name = tensor("op_2125_cast_fp16")]; - tensor inputs_53_cast_fp16 = add(x = inputs_51_cast_fp16, y = var_2125_cast_fp16)[name = tensor("inputs_53_cast_fp16")]; - tensor out_53_axes_0 = const()[name = tensor("out_53_axes_0"), val = tensor([1])]; - tensor var_2135_to_fp16 = const()[name = tensor("op_2135_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_53_cast_fp16 = layer_norm(axes = out_53_axes_0, epsilon = var_2135_to_fp16, x = inputs_53_cast_fp16)[name = tensor("out_53_cast_fp16")]; - tensor obj_23_gamma_0_to_fp16 = const()[name = tensor("obj_23_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32601920)))]; - tensor obj_23_beta_0_to_fp16 = const()[name = tensor("obj_23_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32603008)))]; - tensor obj_23_epsilon_0_to_fp16 = const()[name = tensor("obj_23_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor obj_23_cast_fp16 = batch_norm(beta = obj_23_beta_0_to_fp16, epsilon = obj_23_epsilon_0_to_fp16, gamma = obj_23_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_53_cast_fp16)[name = tensor("obj_23_cast_fp16")]; - tensor var_2160_pad_type_0 = const()[name = tensor("op_2160_pad_type_0"), val = tensor("valid")]; - tensor var_2160_strides_0 = const()[name = tensor("op_2160_strides_0"), val = tensor([1, 1])]; - tensor var_2160_pad_0 = const()[name = tensor("op_2160_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2160_dilations_0 = const()[name = tensor("op_2160_dilations_0"), val = tensor([1, 1])]; - tensor var_2160_groups_0 = const()[name = tensor("op_2160_groups_0"), val = tensor(1)]; - tensor layers_5_self_attn_q_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32604096))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32800768))), name = tensor("layers_5_self_attn_q_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_5_self_attn_q_proj_inlier_module_bias_to_fp16 = const()[name = tensor("layers_5_self_attn_q_proj_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32800960)))]; - tensor var_2160_cast_fp16 = conv(bias = layers_5_self_attn_q_proj_inlier_module_bias_to_fp16, dilations = var_2160_dilations_0, groups = var_2160_groups_0, pad = var_2160_pad_0, pad_type = var_2160_pad_type_0, strides = var_2160_strides_0, weight = layers_5_self_attn_q_proj_inlier_module_weight_to_fp16_palettized, x = obj_23_cast_fp16)[name = tensor("op_2160_cast_fp16")]; - tensor var_2166_pad_type_0 = const()[name = tensor("op_2166_pad_type_0"), val = tensor("valid")]; - tensor var_2166_strides_0 = const()[name = tensor("op_2166_strides_0"), val = tensor([1, 1])]; - tensor var_2166_pad_0 = const()[name = tensor("op_2166_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2166_dilations_0 = const()[name = tensor("op_2166_dilations_0"), val = tensor([1, 1])]; - tensor var_2166_groups_0 = const()[name = tensor("op_2166_groups_0"), val = tensor(1)]; - tensor layers_5_self_attn_q_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32810752))), name = tensor("layers_5_self_attn_q_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32802048))), shape = tensor([512, 512, 1, 1])]; - tensor var_2166_cast_fp16 = conv(dilations = var_2166_dilations_0, groups = var_2166_groups_0, pad = var_2166_pad_0, pad_type = var_2166_pad_type_0, strides = var_2166_strides_0, weight = layers_5_self_attn_q_proj_outlier_module_weight_to_fp16_sparsified, x = obj_23_cast_fp16)[name = tensor("op_2166_cast_fp16")]; - tensor query_21_cast_fp16 = add(x = var_2160_cast_fp16, y = var_2166_cast_fp16)[name = tensor("query_21_cast_fp16")]; - tensor var_2175_pad_type_0 = const()[name = tensor("op_2175_pad_type_0"), val = tensor("valid")]; - tensor var_2175_strides_0 = const()[name = tensor("op_2175_strides_0"), val = tensor([1, 1])]; - tensor var_2175_pad_0 = const()[name = tensor("op_2175_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2175_dilations_0 = const()[name = tensor("op_2175_dilations_0"), val = tensor([1, 1])]; - tensor var_2175_groups_0 = const()[name = tensor("op_2175_groups_0"), val = tensor(1)]; - tensor layers_5_self_attn_k_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32843584))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33040256))), name = tensor("layers_5_self_attn_k_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor var_2175_cast_fp16 = conv(dilations = var_2175_dilations_0, groups = var_2175_groups_0, pad = var_2175_pad_0, pad_type = var_2175_pad_type_0, strides = var_2175_strides_0, weight = layers_5_self_attn_k_proj_inlier_module_weight_to_fp16_palettized, x = obj_23_cast_fp16)[name = tensor("op_2175_cast_fp16")]; - tensor var_2181_pad_type_0 = const()[name = tensor("op_2181_pad_type_0"), val = tensor("valid")]; - tensor var_2181_strides_0 = const()[name = tensor("op_2181_strides_0"), val = tensor([1, 1])]; - tensor var_2181_pad_0 = const()[name = tensor("op_2181_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2181_dilations_0 = const()[name = tensor("op_2181_dilations_0"), val = tensor([1, 1])]; - tensor var_2181_groups_0 = const()[name = tensor("op_2181_groups_0"), val = tensor(1)]; - tensor layers_5_self_attn_k_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33051712))), name = tensor("layers_5_self_attn_k_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33040448))), shape = tensor([512, 512, 1, 1])]; - tensor var_2181_cast_fp16 = conv(dilations = var_2181_dilations_0, groups = var_2181_groups_0, pad = var_2181_pad_0, pad_type = var_2181_pad_type_0, strides = var_2181_strides_0, weight = layers_5_self_attn_k_proj_outlier_module_weight_to_fp16_sparsified, x = obj_23_cast_fp16)[name = tensor("op_2181_cast_fp16")]; - tensor key_11_cast_fp16 = add(x = var_2175_cast_fp16, y = var_2181_cast_fp16)[name = tensor("key_11_cast_fp16")]; - tensor var_2191_pad_type_0 = const()[name = tensor("op_2191_pad_type_0"), val = tensor("valid")]; - tensor var_2191_strides_0 = const()[name = tensor("op_2191_strides_0"), val = tensor([1, 1])]; - tensor var_2191_pad_0 = const()[name = tensor("op_2191_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2191_dilations_0 = const()[name = tensor("op_2191_dilations_0"), val = tensor([1, 1])]; - tensor var_2191_groups_0 = const()[name = tensor("op_2191_groups_0"), val = tensor(1)]; - tensor layers_5_self_attn_v_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33084544))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33281216))), name = tensor("layers_5_self_attn_v_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_5_self_attn_v_proj_inlier_module_bias_to_fp16 = const()[name = tensor("layers_5_self_attn_v_proj_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33281408)))]; - tensor var_2191_cast_fp16 = conv(bias = layers_5_self_attn_v_proj_inlier_module_bias_to_fp16, dilations = var_2191_dilations_0, groups = var_2191_groups_0, pad = var_2191_pad_0, pad_type = var_2191_pad_type_0, strides = var_2191_strides_0, weight = layers_5_self_attn_v_proj_inlier_module_weight_to_fp16_palettized, x = obj_23_cast_fp16)[name = tensor("op_2191_cast_fp16")]; - tensor var_2197_pad_type_0 = const()[name = tensor("op_2197_pad_type_0"), val = tensor("valid")]; - tensor var_2197_strides_0 = const()[name = tensor("op_2197_strides_0"), val = tensor([1, 1])]; - tensor var_2197_pad_0 = const()[name = tensor("op_2197_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2197_dilations_0 = const()[name = tensor("op_2197_dilations_0"), val = tensor([1, 1])]; - tensor var_2197_groups_0 = const()[name = tensor("op_2197_groups_0"), val = tensor(1)]; - tensor layers_5_self_attn_v_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33290816))), name = tensor("layers_5_self_attn_v_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33282496))), shape = tensor([512, 512, 1, 1])]; - tensor var_2197_cast_fp16 = conv(dilations = var_2197_dilations_0, groups = var_2197_groups_0, pad = var_2197_pad_0, pad_type = var_2197_pad_type_0, strides = var_2197_strides_0, weight = layers_5_self_attn_v_proj_outlier_module_weight_to_fp16_sparsified, x = obj_23_cast_fp16)[name = tensor("op_2197_cast_fp16")]; - tensor value_11_cast_fp16 = add(x = var_2191_cast_fp16, y = var_2197_cast_fp16)[name = tensor("value_11_cast_fp16")]; - tensor var_2200_to_fp16 = const()[name = tensor("op_2200_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33323648)))]; - tensor query_23_cast_fp16 = add(x = query_21_cast_fp16, y = var_2200_to_fp16)[name = tensor("query_23_cast_fp16")]; - tensor var_2203_to_fp16 = const()[name = tensor("op_2203_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33324736)))]; - tensor q_with_bias_v_11_cast_fp16 = add(x = query_21_cast_fp16, y = var_2203_to_fp16)[name = tensor("q_with_bias_v_11_cast_fp16")]; - tensor var_2213_pad_type_0 = const()[name = tensor("op_2213_pad_type_0"), val = tensor("valid")]; - tensor var_2213_strides_0 = const()[name = tensor("op_2213_strides_0"), val = tensor([1, 1])]; - tensor var_2213_pad_0 = const()[name = tensor("op_2213_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2213_dilations_0 = const()[name = tensor("op_2213_dilations_0"), val = tensor([1, 1])]; - tensor var_2213_groups_0 = const()[name = tensor("op_2213_groups_0"), val = tensor(1)]; - tensor layers_5_self_attn_linear_pos_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33325824))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33522496))), name = tensor("layers_5_self_attn_linear_pos_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor var_2213_cast_fp16 = conv(bias = input_17_mean_0_to_fp16, dilations = var_2213_dilations_0, groups = var_2213_groups_0, pad = var_2213_pad_0, pad_type = var_2213_pad_type_0, strides = var_2213_strides_0, weight = layers_5_self_attn_linear_pos_inlier_module_weight_to_fp16_palettized, x = obj_3_cast_fp16)[name = tensor("op_2213_cast_fp16")]; - tensor var_2219_pad_type_0 = const()[name = tensor("op_2219_pad_type_0"), val = tensor("valid")]; - tensor var_2219_strides_0 = const()[name = tensor("op_2219_strides_0"), val = tensor([1, 1])]; - tensor var_2219_pad_0 = const()[name = tensor("op_2219_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2219_dilations_0 = const()[name = tensor("op_2219_dilations_0"), val = tensor([1, 1])]; - tensor var_2219_groups_0 = const()[name = tensor("op_2219_groups_0"), val = tensor(1)]; - tensor layers_5_self_attn_linear_pos_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33541568))), name = tensor("layers_5_self_attn_linear_pos_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33522688))), shape = tensor([512, 512, 1, 1])]; - tensor var_2219_cast_fp16 = conv(dilations = var_2219_dilations_0, groups = var_2219_groups_0, pad = var_2219_pad_0, pad_type = var_2219_pad_type_0, strides = var_2219_strides_0, weight = layers_5_self_attn_linear_pos_outlier_module_weight_to_fp16_sparsified, x = obj_3_cast_fp16)[name = tensor("op_2219_cast_fp16")]; - tensor p_11_cast_fp16 = add(x = var_2213_cast_fp16, y = var_2219_cast_fp16)[name = tensor("p_11_cast_fp16")]; - tensor var_2223 = const()[name = tensor("op_2223"), val = tensor([1, 8, 64, 188])]; - tensor var_2224_cast_fp16 = reshape(shape = var_2223, x = q_with_bias_v_11_cast_fp16)[name = tensor("op_2224_cast_fp16")]; - tensor var_2225 = const()[name = tensor("op_2225"), val = tensor([1, 8, 64, -1])]; - tensor var_2226_cast_fp16 = reshape(shape = var_2225, x = p_11_cast_fp16)[name = tensor("op_2226_cast_fp16")]; - tensor matrix_bd_41_transpose_x_0 = const()[name = tensor("matrix_bd_41_transpose_x_0"), val = tensor(true)]; - tensor matrix_bd_41_transpose_y_0 = const()[name = tensor("matrix_bd_41_transpose_y_0"), val = tensor(false)]; - tensor matrix_bd_41_cast_fp16 = matmul(transpose_x = matrix_bd_41_transpose_x_0, transpose_y = matrix_bd_41_transpose_y_0, x = var_2224_cast_fp16, y = var_2226_cast_fp16)[name = tensor("matrix_bd_41_cast_fp16")]; - tensor matrix_bd_43_pad_0 = const()[name = tensor("matrix_bd_43_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor matrix_bd_43_mode_0 = const()[name = tensor("matrix_bd_43_mode_0"), val = tensor("constant")]; - tensor const_65_to_fp16 = const()[name = tensor("const_65_to_fp16"), val = tensor(0x0p+0)]; - tensor matrix_bd_43_cast_fp16 = pad(constant_val = const_65_to_fp16, mode = matrix_bd_43_mode_0, pad = matrix_bd_43_pad_0, x = matrix_bd_41_cast_fp16)[name = tensor("matrix_bd_43_cast_fp16")]; - tensor var_2235 = const()[name = tensor("op_2235"), val = tensor([1, 8, -1, 188])]; - tensor matrix_bd_45_cast_fp16 = reshape(shape = var_2235, x = matrix_bd_43_cast_fp16)[name = tensor("matrix_bd_45_cast_fp16")]; - tensor var_2239_begin_0 = const()[name = tensor("op_2239_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_2239_end_0 = const()[name = tensor("op_2239_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_2239_end_mask_0 = const()[name = tensor("op_2239_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_2239_cast_fp16 = slice_by_index(begin = var_2239_begin_0, end = var_2239_end_0, end_mask = var_2239_end_mask_0, x = matrix_bd_45_cast_fp16)[name = tensor("op_2239_cast_fp16")]; - tensor var_2240 = const()[name = tensor("op_2240"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_47_cast_fp16 = reshape(shape = var_2240, x = var_2239_cast_fp16)[name = tensor("matrix_bd_47_cast_fp16")]; - tensor var_2245_begin_0 = const()[name = tensor("op_2245_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2245_end_0 = const()[name = tensor("op_2245_end_0"), val = tensor([1, 8, 188, 188])]; - tensor var_2245_end_mask_0 = const()[name = tensor("op_2245_end_mask_0"), val = tensor([true, true, true, false])]; - tensor var_2245_cast_fp16 = slice_by_index(begin = var_2245_begin_0, end = var_2245_end_0, end_mask = var_2245_end_mask_0, x = matrix_bd_47_cast_fp16)[name = tensor("op_2245_cast_fp16")]; - tensor var_2246_to_fp16 = const()[name = tensor("op_2246_to_fp16"), val = tensor(0x1p-3)]; - tensor qk_mask_11_cast_fp16 = mul(x = var_2245_cast_fp16, y = var_2246_to_fp16)[name = tensor("qk_mask_11_cast_fp16")]; - tensor var_2250 = const()[name = tensor("op_2250"), val = tensor([1, 8, 64, 188])]; - tensor mh_q_11_cast_fp16 = reshape(shape = var_2250, x = query_23_cast_fp16)[name = tensor("mh_q_11_cast_fp16")]; - tensor var_2252_to_fp16 = const()[name = tensor("op_2252_to_fp16"), val = tensor(0x1p-3)]; - tensor var_2253_cast_fp16 = mul(x = mh_q_11_cast_fp16, y = var_2252_to_fp16)[name = tensor("op_2253_cast_fp16")]; - tensor var_2256 = const()[name = tensor("op_2256"), val = tensor([1, 8, 64, 188])]; - tensor var_2257_cast_fp16 = reshape(shape = var_2256, x = key_11_cast_fp16)[name = tensor("op_2257_cast_fp16")]; - tensor mh_w_21_transpose_x_0 = const()[name = tensor("mh_w_21_transpose_x_0"), val = tensor(true)]; - tensor mh_w_21_transpose_y_0 = const()[name = tensor("mh_w_21_transpose_y_0"), val = tensor(false)]; - tensor mh_w_21_cast_fp16 = matmul(transpose_x = mh_w_21_transpose_x_0, transpose_y = mh_w_21_transpose_y_0, x = var_2253_cast_fp16, y = var_2257_cast_fp16)[name = tensor("mh_w_21_cast_fp16")]; - tensor mh_w_23_cast_fp16 = add(x = mh_w_21_cast_fp16, y = qk_mask_11_cast_fp16)[name = tensor("mh_w_23_cast_fp16")]; - tensor var_2261_cast_fp16 = softmax(axis = var_2048, x = mh_w_23_cast_fp16)[name = tensor("op_2261_cast_fp16")]; - tensor var_2262 = const()[name = tensor("op_2262"), val = tensor([1, 8, 64, 188])]; - tensor var_2263_cast_fp16 = reshape(shape = var_2262, x = value_11_cast_fp16)[name = tensor("op_2263_cast_fp16")]; - tensor attn_11_transpose_x_0 = const()[name = tensor("attn_11_transpose_x_0"), val = tensor(false)]; - tensor attn_11_transpose_y_0 = const()[name = tensor("attn_11_transpose_y_0"), val = tensor(true)]; - tensor attn_11_cast_fp16 = matmul(transpose_x = attn_11_transpose_x_0, transpose_y = attn_11_transpose_y_0, x = var_2263_cast_fp16, y = var_2261_cast_fp16)[name = tensor("attn_11_cast_fp16")]; - tensor var_2266 = const()[name = tensor("op_2266"), val = tensor([1, 512, 1, 188])]; - tensor input_153_cast_fp16 = reshape(shape = var_2266, x = attn_11_cast_fp16)[name = tensor("input_153_cast_fp16")]; - tensor var_2276_pad_type_0 = const()[name = tensor("op_2276_pad_type_0"), val = tensor("valid")]; - tensor var_2276_strides_0 = const()[name = tensor("op_2276_strides_0"), val = tensor([1, 1])]; - tensor var_2276_pad_0 = const()[name = tensor("op_2276_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2276_dilations_0 = const()[name = tensor("op_2276_dilations_0"), val = tensor([1, 1])]; - tensor var_2276_groups_0 = const()[name = tensor("op_2276_groups_0"), val = tensor(1)]; - tensor layers_5_self_attn_o_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33574400))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33771072))), name = tensor("layers_5_self_attn_o_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_5_self_attn_o_proj_inlier_module_bias_to_fp16 = const()[name = tensor("layers_5_self_attn_o_proj_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33771264)))]; - tensor var_2276_cast_fp16 = conv(bias = layers_5_self_attn_o_proj_inlier_module_bias_to_fp16, dilations = var_2276_dilations_0, groups = var_2276_groups_0, pad = var_2276_pad_0, pad_type = var_2276_pad_type_0, strides = var_2276_strides_0, weight = layers_5_self_attn_o_proj_inlier_module_weight_to_fp16_palettized, x = input_153_cast_fp16)[name = tensor("op_2276_cast_fp16")]; - tensor var_2282_pad_type_0 = const()[name = tensor("op_2282_pad_type_0"), val = tensor("valid")]; - tensor var_2282_strides_0 = const()[name = tensor("op_2282_strides_0"), val = tensor([1, 1])]; - tensor var_2282_pad_0 = const()[name = tensor("op_2282_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2282_dilations_0 = const()[name = tensor("op_2282_dilations_0"), val = tensor([1, 1])]; - tensor var_2282_groups_0 = const()[name = tensor("op_2282_groups_0"), val = tensor(1)]; - tensor layers_5_self_attn_o_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33780736))), name = tensor("layers_5_self_attn_o_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33772352))), shape = tensor([512, 512, 1, 1])]; - tensor var_2282_cast_fp16 = conv(dilations = var_2282_dilations_0, groups = var_2282_groups_0, pad = var_2282_pad_0, pad_type = var_2282_pad_type_0, strides = var_2282_strides_0, weight = layers_5_self_attn_o_proj_outlier_module_weight_to_fp16_sparsified, x = input_153_cast_fp16)[name = tensor("op_2282_cast_fp16")]; - tensor obj_25_cast_fp16 = add(x = var_2276_cast_fp16, y = var_2282_cast_fp16)[name = tensor("obj_25_cast_fp16")]; - tensor inputs_55_cast_fp16 = add(x = inputs_53_cast_fp16, y = obj_25_cast_fp16)[name = tensor("inputs_55_cast_fp16")]; - tensor out_55_axes_0 = const()[name = tensor("out_55_axes_0"), val = tensor([1])]; - tensor var_2293_to_fp16 = const()[name = tensor("op_2293_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_55_cast_fp16 = layer_norm(axes = out_55_axes_0, epsilon = var_2293_to_fp16, x = inputs_55_cast_fp16)[name = tensor("out_55_cast_fp16")]; - tensor input_155_gamma_0_to_fp16 = const()[name = tensor("input_155_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33813568)))]; - tensor input_155_beta_0_to_fp16 = const()[name = tensor("input_155_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33814656)))]; - tensor input_155_epsilon_0_to_fp16 = const()[name = tensor("input_155_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_155_cast_fp16 = batch_norm(beta = input_155_beta_0_to_fp16, epsilon = input_155_epsilon_0_to_fp16, gamma = input_155_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_55_cast_fp16)[name = tensor("input_155_cast_fp16")]; - tensor var_2315_pad_type_0 = const()[name = tensor("op_2315_pad_type_0"), val = tensor("valid")]; - tensor var_2315_strides_0 = const()[name = tensor("op_2315_strides_0"), val = tensor([1, 1])]; - tensor var_2315_pad_0 = const()[name = tensor("op_2315_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2315_dilations_0 = const()[name = tensor("op_2315_dilations_0"), val = tensor([1, 1])]; - tensor var_2315_groups_0 = const()[name = tensor("op_2315_groups_0"), val = tensor(1)]; - tensor layers_5_conv_pointwise_conv1_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33815744))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34209024))), name = tensor("layers_5_conv_pointwise_conv1_inlier_module_weight_to_fp16_palettized"), shape = tensor([1024, 512, 1, 1])]; - tensor layers_5_conv_pointwise_conv1_inlier_module_bias_to_fp16 = const()[name = tensor("layers_5_conv_pointwise_conv1_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34209216)))]; - tensor var_2315_cast_fp16 = conv(bias = layers_5_conv_pointwise_conv1_inlier_module_bias_to_fp16, dilations = var_2315_dilations_0, groups = var_2315_groups_0, pad = var_2315_pad_0, pad_type = var_2315_pad_type_0, strides = var_2315_strides_0, weight = layers_5_conv_pointwise_conv1_inlier_module_weight_to_fp16_palettized, x = input_155_cast_fp16)[name = tensor("op_2315_cast_fp16")]; - tensor var_2321_pad_type_0 = const()[name = tensor("op_2321_pad_type_0"), val = tensor("valid")]; - tensor var_2321_strides_0 = const()[name = tensor("op_2321_strides_0"), val = tensor([1, 1])]; - tensor var_2321_pad_0 = const()[name = tensor("op_2321_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2321_dilations_0 = const()[name = tensor("op_2321_dilations_0"), val = tensor([1, 1])]; - tensor var_2321_groups_0 = const()[name = tensor("op_2321_groups_0"), val = tensor(1)]; - tensor layers_5_conv_pointwise_conv1_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34228480))), name = tensor("layers_5_conv_pointwise_conv1_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34211328))), shape = tensor([1024, 512, 1, 1])]; - tensor var_2321_cast_fp16 = conv(dilations = var_2321_dilations_0, groups = var_2321_groups_0, pad = var_2321_pad_0, pad_type = var_2321_pad_type_0, strides = var_2321_strides_0, weight = layers_5_conv_pointwise_conv1_outlier_module_weight_to_fp16_sparsified, x = input_155_cast_fp16)[name = tensor("op_2321_cast_fp16")]; - tensor input_157_cast_fp16 = add(x = var_2315_cast_fp16, y = var_2321_cast_fp16)[name = tensor("input_157_cast_fp16")]; - tensor input_159_split_num_splits_0 = const()[name = tensor("input_159_split_num_splits_0"), val = tensor(2)]; - tensor input_159_split_axis_0 = const()[name = tensor("input_159_split_axis_0"), val = tensor(1)]; - tensor input_159_split_cast_fp16_0, tensor input_159_split_cast_fp16_1 = split(axis = input_159_split_axis_0, num_splits = input_159_split_num_splits_0, x = input_157_cast_fp16)[name = tensor("input_159_split_cast_fp16")]; - tensor input_159_split_1_sigmoid_cast_fp16 = sigmoid(x = input_159_split_cast_fp16_1)[name = tensor("input_159_split_1_sigmoid_cast_fp16")]; - tensor input_159_cast_fp16 = mul(x = input_159_split_cast_fp16_0, y = input_159_split_1_sigmoid_cast_fp16)[name = tensor("input_159_cast_fp16")]; - tensor input_161_pad_type_0 = const()[name = tensor("input_161_pad_type_0"), val = tensor("custom")]; - tensor input_161_pad_0 = const()[name = tensor("input_161_pad_0"), val = tensor([0, 0, 4, 4])]; - tensor input_161_groups_0 = const()[name = tensor("input_161_groups_0"), val = tensor(512)]; - tensor input_161_strides_0 = const()[name = tensor("input_161_strides_0"), val = tensor([1, 1])]; - tensor input_161_dilations_0 = const()[name = tensor("input_161_dilations_0"), val = tensor([1, 1])]; - tensor const_201_to_fp16 = const()[name = tensor("const_201_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34294080)))]; - tensor const_202_to_fp16 = const()[name = tensor("const_202_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34303360)))]; - tensor input_163_cast_fp16 = conv(bias = const_202_to_fp16, dilations = input_161_dilations_0, groups = input_161_groups_0, pad = input_161_pad_0, pad_type = input_161_pad_type_0, strides = input_161_strides_0, weight = const_201_to_fp16, x = input_159_cast_fp16)[name = tensor("input_163_cast_fp16")]; - tensor input_165_cast_fp16 = silu(x = input_163_cast_fp16)[name = tensor("input_165_cast_fp16")]; - tensor var_2345_pad_type_0 = const()[name = tensor("op_2345_pad_type_0"), val = tensor("valid")]; - tensor var_2345_strides_0 = const()[name = tensor("op_2345_strides_0"), val = tensor([1, 1])]; - tensor var_2345_pad_0 = const()[name = tensor("op_2345_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2345_dilations_0 = const()[name = tensor("op_2345_dilations_0"), val = tensor([1, 1])]; - tensor var_2345_groups_0 = const()[name = tensor("op_2345_groups_0"), val = tensor(1)]; - tensor layers_5_conv_pointwise_conv2_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34304448))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34501120))), name = tensor("layers_5_conv_pointwise_conv2_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_5_conv_pointwise_conv2_inlier_module_bias_to_fp16 = const()[name = tensor("layers_5_conv_pointwise_conv2_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34501312)))]; - tensor var_2345_cast_fp16 = conv(bias = layers_5_conv_pointwise_conv2_inlier_module_bias_to_fp16, dilations = var_2345_dilations_0, groups = var_2345_groups_0, pad = var_2345_pad_0, pad_type = var_2345_pad_type_0, strides = var_2345_strides_0, weight = layers_5_conv_pointwise_conv2_inlier_module_weight_to_fp16_palettized, x = input_165_cast_fp16)[name = tensor("op_2345_cast_fp16")]; - tensor var_2351_pad_type_0 = const()[name = tensor("op_2351_pad_type_0"), val = tensor("valid")]; - tensor var_2351_strides_0 = const()[name = tensor("op_2351_strides_0"), val = tensor([1, 1])]; - tensor var_2351_pad_0 = const()[name = tensor("op_2351_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2351_dilations_0 = const()[name = tensor("op_2351_dilations_0"), val = tensor([1, 1])]; - tensor var_2351_groups_0 = const()[name = tensor("op_2351_groups_0"), val = tensor(1)]; - tensor layers_5_conv_pointwise_conv2_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34510656))), name = tensor("layers_5_conv_pointwise_conv2_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34502400))), shape = tensor([512, 512, 1, 1])]; - tensor var_2351_cast_fp16 = conv(dilations = var_2351_dilations_0, groups = var_2351_groups_0, pad = var_2351_pad_0, pad_type = var_2351_pad_type_0, strides = var_2351_strides_0, weight = layers_5_conv_pointwise_conv2_outlier_module_weight_to_fp16_sparsified, x = input_165_cast_fp16)[name = tensor("op_2351_cast_fp16")]; - tensor x_35_cast_fp16 = add(x = var_2345_cast_fp16, y = var_2351_cast_fp16)[name = tensor("x_35_cast_fp16")]; - tensor inputs_57_cast_fp16 = add(x = inputs_55_cast_fp16, y = x_35_cast_fp16)[name = tensor("inputs_57_cast_fp16")]; - tensor out_57_axes_0 = const()[name = tensor("out_57_axes_0"), val = tensor([1])]; - tensor var_2362_to_fp16 = const()[name = tensor("op_2362_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_57_cast_fp16 = layer_norm(axes = out_57_axes_0, epsilon = var_2362_to_fp16, x = inputs_57_cast_fp16)[name = tensor("out_57_cast_fp16")]; - tensor input_167_gamma_0_to_fp16 = const()[name = tensor("input_167_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34543488)))]; - tensor input_167_beta_0_to_fp16 = const()[name = tensor("input_167_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34544576)))]; - tensor input_167_epsilon_0_to_fp16 = const()[name = tensor("input_167_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_167_cast_fp16 = batch_norm(beta = input_167_beta_0_to_fp16, epsilon = input_167_epsilon_0_to_fp16, gamma = input_167_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_57_cast_fp16)[name = tensor("input_167_cast_fp16")]; - tensor var_2382_pad_type_0 = const()[name = tensor("op_2382_pad_type_0"), val = tensor("valid")]; - tensor var_2382_strides_0 = const()[name = tensor("op_2382_strides_0"), val = tensor([1, 1])]; - tensor var_2382_pad_0 = const()[name = tensor("op_2382_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2382_dilations_0 = const()[name = tensor("op_2382_dilations_0"), val = tensor([1, 1])]; - tensor var_2382_groups_0 = const()[name = tensor("op_2382_groups_0"), val = tensor(1)]; - tensor layers_5_feed_forward2_fc1_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34545664))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35332160))), name = tensor("layers_5_feed_forward2_fc1_inlier_module_weight_to_fp16_palettized"), shape = tensor([2048, 512, 1, 1])]; - tensor layers_5_feed_forward2_fc1_inlier_module_bias_to_fp16 = const()[name = tensor("layers_5_feed_forward2_fc1_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35332352)))]; - tensor var_2382_cast_fp16 = conv(bias = layers_5_feed_forward2_fc1_inlier_module_bias_to_fp16, dilations = var_2382_dilations_0, groups = var_2382_groups_0, pad = var_2382_pad_0, pad_type = var_2382_pad_type_0, strides = var_2382_strides_0, weight = layers_5_feed_forward2_fc1_inlier_module_weight_to_fp16_palettized, x = input_167_cast_fp16)[name = tensor("op_2382_cast_fp16")]; - tensor var_2388_pad_type_0 = const()[name = tensor("op_2388_pad_type_0"), val = tensor("valid")]; - tensor var_2388_strides_0 = const()[name = tensor("op_2388_strides_0"), val = tensor([1, 1])]; - tensor var_2388_pad_0 = const()[name = tensor("op_2388_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2388_dilations_0 = const()[name = tensor("op_2388_dilations_0"), val = tensor([1, 1])]; - tensor var_2388_groups_0 = const()[name = tensor("op_2388_groups_0"), val = tensor(1)]; - tensor layers_5_feed_forward2_fc1_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35369792))), name = tensor("layers_5_feed_forward2_fc1_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35336512))), shape = tensor([2048, 512, 1, 1])]; - tensor var_2388_cast_fp16 = conv(dilations = var_2388_dilations_0, groups = var_2388_groups_0, pad = var_2388_pad_0, pad_type = var_2388_pad_type_0, strides = var_2388_strides_0, weight = layers_5_feed_forward2_fc1_outlier_module_weight_to_fp16_sparsified, x = input_167_cast_fp16)[name = tensor("op_2388_cast_fp16")]; - tensor input_169_cast_fp16 = add(x = var_2382_cast_fp16, y = var_2388_cast_fp16)[name = tensor("input_169_cast_fp16")]; - tensor input_171_cast_fp16 = silu(x = input_169_cast_fp16)[name = tensor("input_171_cast_fp16")]; - tensor var_2399_pad_type_0 = const()[name = tensor("op_2399_pad_type_0"), val = tensor("valid")]; - tensor var_2399_strides_0 = const()[name = tensor("op_2399_strides_0"), val = tensor([1, 1])]; - tensor var_2399_pad_0 = const()[name = tensor("op_2399_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2399_dilations_0 = const()[name = tensor("op_2399_dilations_0"), val = tensor([1, 1])]; - tensor var_2399_groups_0 = const()[name = tensor("op_2399_groups_0"), val = tensor(1)]; - tensor layers_5_feed_forward2_fc2_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35500928))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36287424))), name = tensor("layers_5_feed_forward2_fc2_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 2048, 1, 1])]; - tensor layers_5_feed_forward2_fc2_inlier_module_bias_to_fp16 = const()[name = tensor("layers_5_feed_forward2_fc2_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36287616)))]; - tensor var_2399_cast_fp16 = conv(bias = layers_5_feed_forward2_fc2_inlier_module_bias_to_fp16, dilations = var_2399_dilations_0, groups = var_2399_groups_0, pad = var_2399_pad_0, pad_type = var_2399_pad_type_0, strides = var_2399_strides_0, weight = layers_5_feed_forward2_fc2_inlier_module_weight_to_fp16_palettized, x = input_171_cast_fp16)[name = tensor("op_2399_cast_fp16")]; - tensor var_2405_pad_type_0 = const()[name = tensor("op_2405_pad_type_0"), val = tensor("valid")]; - tensor var_2405_strides_0 = const()[name = tensor("op_2405_strides_0"), val = tensor([1, 1])]; - tensor var_2405_pad_0 = const()[name = tensor("op_2405_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2405_dilations_0 = const()[name = tensor("op_2405_dilations_0"), val = tensor([1, 1])]; - tensor var_2405_groups_0 = const()[name = tensor("op_2405_groups_0"), val = tensor(1)]; - tensor layers_5_feed_forward2_fc2_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36328256))), name = tensor("layers_5_feed_forward2_fc2_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36288704))), shape = tensor([512, 2048, 1, 1])]; - tensor var_2405_cast_fp16 = conv(dilations = var_2405_dilations_0, groups = var_2405_groups_0, pad = var_2405_pad_0, pad_type = var_2405_pad_type_0, strides = var_2405_strides_0, weight = layers_5_feed_forward2_fc2_outlier_module_weight_to_fp16_sparsified, x = input_171_cast_fp16)[name = tensor("op_2405_cast_fp16")]; - tensor x_37_cast_fp16 = add(x = var_2399_cast_fp16, y = var_2405_cast_fp16)[name = tensor("x_37_cast_fp16")]; - tensor var_2407_to_fp16 = const()[name = tensor("op_2407_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2408_cast_fp16 = mul(x = x_37_cast_fp16, y = var_2407_to_fp16)[name = tensor("op_2408_cast_fp16")]; - tensor inputs_59_cast_fp16 = add(x = inputs_57_cast_fp16, y = var_2408_cast_fp16)[name = tensor("inputs_59_cast_fp16")]; - tensor out_59_axes_0 = const()[name = tensor("out_59_axes_0"), val = tensor([1])]; - tensor var_2418_to_fp16 = const()[name = tensor("op_2418_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_59_cast_fp16 = layer_norm(axes = out_59_axes_0, epsilon = var_2418_to_fp16, x = inputs_59_cast_fp16)[name = tensor("out_59_cast_fp16")]; - tensor inputs_61_gamma_0_to_fp16 = const()[name = tensor("inputs_61_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36459392)))]; - tensor inputs_61_beta_0_to_fp16 = const()[name = tensor("inputs_61_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36460480)))]; - tensor inputs_61_epsilon_0_to_fp16 = const()[name = tensor("inputs_61_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor inputs_61_cast_fp16 = batch_norm(beta = inputs_61_beta_0_to_fp16, epsilon = inputs_61_epsilon_0_to_fp16, gamma = inputs_61_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_59_cast_fp16)[name = tensor("inputs_61_cast_fp16")]; - tensor var_2432 = const()[name = tensor("op_2432"), val = tensor(3)]; - tensor out_61_axes_0 = const()[name = tensor("out_61_axes_0"), val = tensor([1])]; - tensor var_2463_to_fp16 = const()[name = tensor("op_2463_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_61_cast_fp16 = layer_norm(axes = out_61_axes_0, epsilon = var_2463_to_fp16, x = inputs_61_cast_fp16)[name = tensor("out_61_cast_fp16")]; - tensor input_173_gamma_0_to_fp16 = const()[name = tensor("input_173_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36461568)))]; - tensor input_173_beta_0_to_fp16 = const()[name = tensor("input_173_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36462656)))]; - tensor input_173_epsilon_0_to_fp16 = const()[name = tensor("input_173_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_173_cast_fp16 = batch_norm(beta = input_173_beta_0_to_fp16, epsilon = input_173_epsilon_0_to_fp16, gamma = input_173_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_61_cast_fp16)[name = tensor("input_173_cast_fp16")]; - tensor var_2483_pad_type_0 = const()[name = tensor("op_2483_pad_type_0"), val = tensor("valid")]; - tensor var_2483_strides_0 = const()[name = tensor("op_2483_strides_0"), val = tensor([1, 1])]; - tensor var_2483_pad_0 = const()[name = tensor("op_2483_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2483_dilations_0 = const()[name = tensor("op_2483_dilations_0"), val = tensor([1, 1])]; - tensor var_2483_groups_0 = const()[name = tensor("op_2483_groups_0"), val = tensor(1)]; - tensor layers_6_feed_forward1_fc1_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36463744))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(37250240))), name = tensor("layers_6_feed_forward1_fc1_inlier_module_weight_to_fp16_palettized"), shape = tensor([2048, 512, 1, 1])]; - tensor layers_6_feed_forward1_fc1_inlier_module_bias_to_fp16 = const()[name = tensor("layers_6_feed_forward1_fc1_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(37250432)))]; - tensor var_2483_cast_fp16 = conv(bias = layers_6_feed_forward1_fc1_inlier_module_bias_to_fp16, dilations = var_2483_dilations_0, groups = var_2483_groups_0, pad = var_2483_pad_0, pad_type = var_2483_pad_type_0, strides = var_2483_strides_0, weight = layers_6_feed_forward1_fc1_inlier_module_weight_to_fp16_palettized, x = input_173_cast_fp16)[name = tensor("op_2483_cast_fp16")]; - tensor var_2489_pad_type_0 = const()[name = tensor("op_2489_pad_type_0"), val = tensor("valid")]; - tensor var_2489_strides_0 = const()[name = tensor("op_2489_strides_0"), val = tensor([1, 1])]; - tensor var_2489_pad_0 = const()[name = tensor("op_2489_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2489_dilations_0 = const()[name = tensor("op_2489_dilations_0"), val = tensor([1, 1])]; - tensor var_2489_groups_0 = const()[name = tensor("op_2489_groups_0"), val = tensor(1)]; - tensor layers_6_feed_forward1_fc1_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(37288576))), name = tensor("layers_6_feed_forward1_fc1_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(37254592))), shape = tensor([2048, 512, 1, 1])]; - tensor var_2489_cast_fp16 = conv(dilations = var_2489_dilations_0, groups = var_2489_groups_0, pad = var_2489_pad_0, pad_type = var_2489_pad_type_0, strides = var_2489_strides_0, weight = layers_6_feed_forward1_fc1_outlier_module_weight_to_fp16_sparsified, x = input_173_cast_fp16)[name = tensor("op_2489_cast_fp16")]; - tensor input_175_cast_fp16 = add(x = var_2483_cast_fp16, y = var_2489_cast_fp16)[name = tensor("input_175_cast_fp16")]; - tensor input_177_cast_fp16 = silu(x = input_175_cast_fp16)[name = tensor("input_177_cast_fp16")]; - tensor var_2500_pad_type_0 = const()[name = tensor("op_2500_pad_type_0"), val = tensor("valid")]; - tensor var_2500_strides_0 = const()[name = tensor("op_2500_strides_0"), val = tensor([1, 1])]; - tensor var_2500_pad_0 = const()[name = tensor("op_2500_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2500_dilations_0 = const()[name = tensor("op_2500_dilations_0"), val = tensor([1, 1])]; - tensor var_2500_groups_0 = const()[name = tensor("op_2500_groups_0"), val = tensor(1)]; - tensor layers_6_feed_forward1_fc2_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(37419712))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(38206208))), name = tensor("layers_6_feed_forward1_fc2_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 2048, 1, 1])]; - tensor layers_6_feed_forward1_fc2_inlier_module_bias_to_fp16 = const()[name = tensor("layers_6_feed_forward1_fc2_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(38206400)))]; - tensor var_2500_cast_fp16 = conv(bias = layers_6_feed_forward1_fc2_inlier_module_bias_to_fp16, dilations = var_2500_dilations_0, groups = var_2500_groups_0, pad = var_2500_pad_0, pad_type = var_2500_pad_type_0, strides = var_2500_strides_0, weight = layers_6_feed_forward1_fc2_inlier_module_weight_to_fp16_palettized, x = input_177_cast_fp16)[name = tensor("op_2500_cast_fp16")]; - tensor var_2506_pad_type_0 = const()[name = tensor("op_2506_pad_type_0"), val = tensor("valid")]; - tensor var_2506_strides_0 = const()[name = tensor("op_2506_strides_0"), val = tensor([1, 1])]; - tensor var_2506_pad_0 = const()[name = tensor("op_2506_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2506_dilations_0 = const()[name = tensor("op_2506_dilations_0"), val = tensor([1, 1])]; - tensor var_2506_groups_0 = const()[name = tensor("op_2506_groups_0"), val = tensor(1)]; - tensor layers_6_feed_forward1_fc2_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(38248448))), name = tensor("layers_6_feed_forward1_fc2_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(38207488))), shape = tensor([512, 2048, 1, 1])]; - tensor var_2506_cast_fp16 = conv(dilations = var_2506_dilations_0, groups = var_2506_groups_0, pad = var_2506_pad_0, pad_type = var_2506_pad_type_0, strides = var_2506_strides_0, weight = layers_6_feed_forward1_fc2_outlier_module_weight_to_fp16_sparsified, x = input_177_cast_fp16)[name = tensor("op_2506_cast_fp16")]; - tensor x_39_cast_fp16 = add(x = var_2500_cast_fp16, y = var_2506_cast_fp16)[name = tensor("x_39_cast_fp16")]; - tensor var_2508_to_fp16 = const()[name = tensor("op_2508_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2509_cast_fp16 = mul(x = x_39_cast_fp16, y = var_2508_to_fp16)[name = tensor("op_2509_cast_fp16")]; - tensor inputs_63_cast_fp16 = add(x = inputs_61_cast_fp16, y = var_2509_cast_fp16)[name = tensor("inputs_63_cast_fp16")]; - tensor out_63_axes_0 = const()[name = tensor("out_63_axes_0"), val = tensor([1])]; - tensor var_2519_to_fp16 = const()[name = tensor("op_2519_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_63_cast_fp16 = layer_norm(axes = out_63_axes_0, epsilon = var_2519_to_fp16, x = inputs_63_cast_fp16)[name = tensor("out_63_cast_fp16")]; - tensor obj_27_gamma_0_to_fp16 = const()[name = tensor("obj_27_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(38379584)))]; - tensor obj_27_beta_0_to_fp16 = const()[name = tensor("obj_27_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(38380672)))]; - tensor obj_27_epsilon_0_to_fp16 = const()[name = tensor("obj_27_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor obj_27_cast_fp16 = batch_norm(beta = obj_27_beta_0_to_fp16, epsilon = obj_27_epsilon_0_to_fp16, gamma = obj_27_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_63_cast_fp16)[name = tensor("obj_27_cast_fp16")]; - tensor var_2544_pad_type_0 = const()[name = tensor("op_2544_pad_type_0"), val = tensor("valid")]; - tensor var_2544_strides_0 = const()[name = tensor("op_2544_strides_0"), val = tensor([1, 1])]; - tensor var_2544_pad_0 = const()[name = tensor("op_2544_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2544_dilations_0 = const()[name = tensor("op_2544_dilations_0"), val = tensor([1, 1])]; - tensor var_2544_groups_0 = const()[name = tensor("op_2544_groups_0"), val = tensor(1)]; - tensor layers_6_self_attn_q_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(38381760))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(38578432))), name = tensor("layers_6_self_attn_q_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_6_self_attn_q_proj_inlier_module_bias_to_fp16 = const()[name = tensor("layers_6_self_attn_q_proj_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(38578624)))]; - tensor var_2544_cast_fp16 = conv(bias = layers_6_self_attn_q_proj_inlier_module_bias_to_fp16, dilations = var_2544_dilations_0, groups = var_2544_groups_0, pad = var_2544_pad_0, pad_type = var_2544_pad_type_0, strides = var_2544_strides_0, weight = layers_6_self_attn_q_proj_inlier_module_weight_to_fp16_palettized, x = obj_27_cast_fp16)[name = tensor("op_2544_cast_fp16")]; - tensor var_2550_pad_type_0 = const()[name = tensor("op_2550_pad_type_0"), val = tensor("valid")]; - tensor var_2550_strides_0 = const()[name = tensor("op_2550_strides_0"), val = tensor([1, 1])]; - tensor var_2550_pad_0 = const()[name = tensor("op_2550_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2550_dilations_0 = const()[name = tensor("op_2550_dilations_0"), val = tensor([1, 1])]; - tensor var_2550_groups_0 = const()[name = tensor("op_2550_groups_0"), val = tensor(1)]; - tensor layers_6_self_attn_q_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(38588992))), name = tensor("layers_6_self_attn_q_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(38579712))), shape = tensor([512, 512, 1, 1])]; - tensor var_2550_cast_fp16 = conv(dilations = var_2550_dilations_0, groups = var_2550_groups_0, pad = var_2550_pad_0, pad_type = var_2550_pad_type_0, strides = var_2550_strides_0, weight = layers_6_self_attn_q_proj_outlier_module_weight_to_fp16_sparsified, x = obj_27_cast_fp16)[name = tensor("op_2550_cast_fp16")]; - tensor query_25_cast_fp16 = add(x = var_2544_cast_fp16, y = var_2550_cast_fp16)[name = tensor("query_25_cast_fp16")]; - tensor var_2559_pad_type_0 = const()[name = tensor("op_2559_pad_type_0"), val = tensor("valid")]; - tensor var_2559_strides_0 = const()[name = tensor("op_2559_strides_0"), val = tensor([1, 1])]; - tensor var_2559_pad_0 = const()[name = tensor("op_2559_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2559_dilations_0 = const()[name = tensor("op_2559_dilations_0"), val = tensor([1, 1])]; - tensor var_2559_groups_0 = const()[name = tensor("op_2559_groups_0"), val = tensor(1)]; - tensor layers_6_self_attn_k_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(38621824))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(38818496))), name = tensor("layers_6_self_attn_k_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor var_2559_cast_fp16 = conv(dilations = var_2559_dilations_0, groups = var_2559_groups_0, pad = var_2559_pad_0, pad_type = var_2559_pad_type_0, strides = var_2559_strides_0, weight = layers_6_self_attn_k_proj_inlier_module_weight_to_fp16_palettized, x = obj_27_cast_fp16)[name = tensor("op_2559_cast_fp16")]; - tensor var_2565_pad_type_0 = const()[name = tensor("op_2565_pad_type_0"), val = tensor("valid")]; - tensor var_2565_strides_0 = const()[name = tensor("op_2565_strides_0"), val = tensor([1, 1])]; - tensor var_2565_pad_0 = const()[name = tensor("op_2565_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2565_dilations_0 = const()[name = tensor("op_2565_dilations_0"), val = tensor([1, 1])]; - tensor var_2565_groups_0 = const()[name = tensor("op_2565_groups_0"), val = tensor(1)]; - tensor layers_6_self_attn_k_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(38829632))), name = tensor("layers_6_self_attn_k_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(38818688))), shape = tensor([512, 512, 1, 1])]; - tensor var_2565_cast_fp16 = conv(dilations = var_2565_dilations_0, groups = var_2565_groups_0, pad = var_2565_pad_0, pad_type = var_2565_pad_type_0, strides = var_2565_strides_0, weight = layers_6_self_attn_k_proj_outlier_module_weight_to_fp16_sparsified, x = obj_27_cast_fp16)[name = tensor("op_2565_cast_fp16")]; - tensor key_13_cast_fp16 = add(x = var_2559_cast_fp16, y = var_2565_cast_fp16)[name = tensor("key_13_cast_fp16")]; - tensor var_2575_pad_type_0 = const()[name = tensor("op_2575_pad_type_0"), val = tensor("valid")]; - tensor var_2575_strides_0 = const()[name = tensor("op_2575_strides_0"), val = tensor([1, 1])]; - tensor var_2575_pad_0 = const()[name = tensor("op_2575_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2575_dilations_0 = const()[name = tensor("op_2575_dilations_0"), val = tensor([1, 1])]; - tensor var_2575_groups_0 = const()[name = tensor("op_2575_groups_0"), val = tensor(1)]; - tensor layers_6_self_attn_v_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(38862464))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(39059136))), name = tensor("layers_6_self_attn_v_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_6_self_attn_v_proj_inlier_module_bias_to_fp16 = const()[name = tensor("layers_6_self_attn_v_proj_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(39059328)))]; - tensor var_2575_cast_fp16 = conv(bias = layers_6_self_attn_v_proj_inlier_module_bias_to_fp16, dilations = var_2575_dilations_0, groups = var_2575_groups_0, pad = var_2575_pad_0, pad_type = var_2575_pad_type_0, strides = var_2575_strides_0, weight = layers_6_self_attn_v_proj_inlier_module_weight_to_fp16_palettized, x = obj_27_cast_fp16)[name = tensor("op_2575_cast_fp16")]; - tensor var_2581_pad_type_0 = const()[name = tensor("op_2581_pad_type_0"), val = tensor("valid")]; - tensor var_2581_strides_0 = const()[name = tensor("op_2581_strides_0"), val = tensor([1, 1])]; - tensor var_2581_pad_0 = const()[name = tensor("op_2581_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2581_dilations_0 = const()[name = tensor("op_2581_dilations_0"), val = tensor([1, 1])]; - tensor var_2581_groups_0 = const()[name = tensor("op_2581_groups_0"), val = tensor(1)]; - tensor layers_6_self_attn_v_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(39068864))), name = tensor("layers_6_self_attn_v_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(39060416))), shape = tensor([512, 512, 1, 1])]; - tensor var_2581_cast_fp16 = conv(dilations = var_2581_dilations_0, groups = var_2581_groups_0, pad = var_2581_pad_0, pad_type = var_2581_pad_type_0, strides = var_2581_strides_0, weight = layers_6_self_attn_v_proj_outlier_module_weight_to_fp16_sparsified, x = obj_27_cast_fp16)[name = tensor("op_2581_cast_fp16")]; - tensor value_13_cast_fp16 = add(x = var_2575_cast_fp16, y = var_2581_cast_fp16)[name = tensor("value_13_cast_fp16")]; - tensor var_2584_to_fp16 = const()[name = tensor("op_2584_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(39101696)))]; - tensor query_27_cast_fp16 = add(x = query_25_cast_fp16, y = var_2584_to_fp16)[name = tensor("query_27_cast_fp16")]; - tensor var_2587_to_fp16 = const()[name = tensor("op_2587_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(39102784)))]; - tensor q_with_bias_v_13_cast_fp16 = add(x = query_25_cast_fp16, y = var_2587_to_fp16)[name = tensor("q_with_bias_v_13_cast_fp16")]; - tensor var_2597_pad_type_0 = const()[name = tensor("op_2597_pad_type_0"), val = tensor("valid")]; - tensor var_2597_strides_0 = const()[name = tensor("op_2597_strides_0"), val = tensor([1, 1])]; - tensor var_2597_pad_0 = const()[name = tensor("op_2597_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2597_dilations_0 = const()[name = tensor("op_2597_dilations_0"), val = tensor([1, 1])]; - tensor var_2597_groups_0 = const()[name = tensor("op_2597_groups_0"), val = tensor(1)]; - tensor layers_6_self_attn_linear_pos_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(39103872))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(39300544))), name = tensor("layers_6_self_attn_linear_pos_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor var_2597_cast_fp16 = conv(bias = input_17_mean_0_to_fp16, dilations = var_2597_dilations_0, groups = var_2597_groups_0, pad = var_2597_pad_0, pad_type = var_2597_pad_type_0, strides = var_2597_strides_0, weight = layers_6_self_attn_linear_pos_inlier_module_weight_to_fp16_palettized, x = obj_3_cast_fp16)[name = tensor("op_2597_cast_fp16")]; - tensor var_2603_pad_type_0 = const()[name = tensor("op_2603_pad_type_0"), val = tensor("valid")]; - tensor var_2603_strides_0 = const()[name = tensor("op_2603_strides_0"), val = tensor([1, 1])]; - tensor var_2603_pad_0 = const()[name = tensor("op_2603_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2603_dilations_0 = const()[name = tensor("op_2603_dilations_0"), val = tensor([1, 1])]; - tensor var_2603_groups_0 = const()[name = tensor("op_2603_groups_0"), val = tensor(1)]; - tensor layers_6_self_attn_linear_pos_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(39321024))), name = tensor("layers_6_self_attn_linear_pos_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(39300736))), shape = tensor([512, 512, 1, 1])]; - tensor var_2603_cast_fp16 = conv(dilations = var_2603_dilations_0, groups = var_2603_groups_0, pad = var_2603_pad_0, pad_type = var_2603_pad_type_0, strides = var_2603_strides_0, weight = layers_6_self_attn_linear_pos_outlier_module_weight_to_fp16_sparsified, x = obj_3_cast_fp16)[name = tensor("op_2603_cast_fp16")]; - tensor p_13_cast_fp16 = add(x = var_2597_cast_fp16, y = var_2603_cast_fp16)[name = tensor("p_13_cast_fp16")]; - tensor var_2607 = const()[name = tensor("op_2607"), val = tensor([1, 8, 64, 188])]; - tensor var_2608_cast_fp16 = reshape(shape = var_2607, x = q_with_bias_v_13_cast_fp16)[name = tensor("op_2608_cast_fp16")]; - tensor var_2609 = const()[name = tensor("op_2609"), val = tensor([1, 8, 64, -1])]; - tensor var_2610_cast_fp16 = reshape(shape = var_2609, x = p_13_cast_fp16)[name = tensor("op_2610_cast_fp16")]; - tensor matrix_bd_49_transpose_x_0 = const()[name = tensor("matrix_bd_49_transpose_x_0"), val = tensor(true)]; - tensor matrix_bd_49_transpose_y_0 = const()[name = tensor("matrix_bd_49_transpose_y_0"), val = tensor(false)]; - tensor matrix_bd_49_cast_fp16 = matmul(transpose_x = matrix_bd_49_transpose_x_0, transpose_y = matrix_bd_49_transpose_y_0, x = var_2608_cast_fp16, y = var_2610_cast_fp16)[name = tensor("matrix_bd_49_cast_fp16")]; - tensor matrix_bd_51_pad_0 = const()[name = tensor("matrix_bd_51_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor matrix_bd_51_mode_0 = const()[name = tensor("matrix_bd_51_mode_0"), val = tensor("constant")]; - tensor const_76_to_fp16 = const()[name = tensor("const_76_to_fp16"), val = tensor(0x0p+0)]; - tensor matrix_bd_51_cast_fp16 = pad(constant_val = const_76_to_fp16, mode = matrix_bd_51_mode_0, pad = matrix_bd_51_pad_0, x = matrix_bd_49_cast_fp16)[name = tensor("matrix_bd_51_cast_fp16")]; - tensor var_2619 = const()[name = tensor("op_2619"), val = tensor([1, 8, -1, 188])]; - tensor matrix_bd_53_cast_fp16 = reshape(shape = var_2619, x = matrix_bd_51_cast_fp16)[name = tensor("matrix_bd_53_cast_fp16")]; - tensor var_2623_begin_0 = const()[name = tensor("op_2623_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_2623_end_0 = const()[name = tensor("op_2623_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_2623_end_mask_0 = const()[name = tensor("op_2623_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_2623_cast_fp16 = slice_by_index(begin = var_2623_begin_0, end = var_2623_end_0, end_mask = var_2623_end_mask_0, x = matrix_bd_53_cast_fp16)[name = tensor("op_2623_cast_fp16")]; - tensor var_2624 = const()[name = tensor("op_2624"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_55_cast_fp16 = reshape(shape = var_2624, x = var_2623_cast_fp16)[name = tensor("matrix_bd_55_cast_fp16")]; - tensor var_2629_begin_0 = const()[name = tensor("op_2629_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2629_end_0 = const()[name = tensor("op_2629_end_0"), val = tensor([1, 8, 188, 188])]; - tensor var_2629_end_mask_0 = const()[name = tensor("op_2629_end_mask_0"), val = tensor([true, true, true, false])]; - tensor var_2629_cast_fp16 = slice_by_index(begin = var_2629_begin_0, end = var_2629_end_0, end_mask = var_2629_end_mask_0, x = matrix_bd_55_cast_fp16)[name = tensor("op_2629_cast_fp16")]; - tensor var_2630_to_fp16 = const()[name = tensor("op_2630_to_fp16"), val = tensor(0x1p-3)]; - tensor qk_mask_13_cast_fp16 = mul(x = var_2629_cast_fp16, y = var_2630_to_fp16)[name = tensor("qk_mask_13_cast_fp16")]; - tensor var_2634 = const()[name = tensor("op_2634"), val = tensor([1, 8, 64, 188])]; - tensor mh_q_13_cast_fp16 = reshape(shape = var_2634, x = query_27_cast_fp16)[name = tensor("mh_q_13_cast_fp16")]; - tensor var_2636_to_fp16 = const()[name = tensor("op_2636_to_fp16"), val = tensor(0x1p-3)]; - tensor var_2637_cast_fp16 = mul(x = mh_q_13_cast_fp16, y = var_2636_to_fp16)[name = tensor("op_2637_cast_fp16")]; - tensor var_2640 = const()[name = tensor("op_2640"), val = tensor([1, 8, 64, 188])]; - tensor var_2641_cast_fp16 = reshape(shape = var_2640, x = key_13_cast_fp16)[name = tensor("op_2641_cast_fp16")]; - tensor mh_w_25_transpose_x_0 = const()[name = tensor("mh_w_25_transpose_x_0"), val = tensor(true)]; - tensor mh_w_25_transpose_y_0 = const()[name = tensor("mh_w_25_transpose_y_0"), val = tensor(false)]; - tensor mh_w_25_cast_fp16 = matmul(transpose_x = mh_w_25_transpose_x_0, transpose_y = mh_w_25_transpose_y_0, x = var_2637_cast_fp16, y = var_2641_cast_fp16)[name = tensor("mh_w_25_cast_fp16")]; - tensor mh_w_27_cast_fp16 = add(x = mh_w_25_cast_fp16, y = qk_mask_13_cast_fp16)[name = tensor("mh_w_27_cast_fp16")]; - tensor var_2645_cast_fp16 = softmax(axis = var_2432, x = mh_w_27_cast_fp16)[name = tensor("op_2645_cast_fp16")]; - tensor var_2646 = const()[name = tensor("op_2646"), val = tensor([1, 8, 64, 188])]; - tensor var_2647_cast_fp16 = reshape(shape = var_2646, x = value_13_cast_fp16)[name = tensor("op_2647_cast_fp16")]; - tensor attn_13_transpose_x_0 = const()[name = tensor("attn_13_transpose_x_0"), val = tensor(false)]; - tensor attn_13_transpose_y_0 = const()[name = tensor("attn_13_transpose_y_0"), val = tensor(true)]; - tensor attn_13_cast_fp16 = matmul(transpose_x = attn_13_transpose_x_0, transpose_y = attn_13_transpose_y_0, x = var_2647_cast_fp16, y = var_2645_cast_fp16)[name = tensor("attn_13_cast_fp16")]; - tensor var_2650 = const()[name = tensor("op_2650"), val = tensor([1, 512, 1, 188])]; - tensor input_179_cast_fp16 = reshape(shape = var_2650, x = attn_13_cast_fp16)[name = tensor("input_179_cast_fp16")]; - tensor var_2660_pad_type_0 = const()[name = tensor("op_2660_pad_type_0"), val = tensor("valid")]; - tensor var_2660_strides_0 = const()[name = tensor("op_2660_strides_0"), val = tensor([1, 1])]; - tensor var_2660_pad_0 = const()[name = tensor("op_2660_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2660_dilations_0 = const()[name = tensor("op_2660_dilations_0"), val = tensor([1, 1])]; - tensor var_2660_groups_0 = const()[name = tensor("op_2660_groups_0"), val = tensor(1)]; - tensor layers_6_self_attn_o_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(39353856))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(39550528))), name = tensor("layers_6_self_attn_o_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_6_self_attn_o_proj_inlier_module_bias_to_fp16 = const()[name = tensor("layers_6_self_attn_o_proj_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(39550720)))]; - tensor var_2660_cast_fp16 = conv(bias = layers_6_self_attn_o_proj_inlier_module_bias_to_fp16, dilations = var_2660_dilations_0, groups = var_2660_groups_0, pad = var_2660_pad_0, pad_type = var_2660_pad_type_0, strides = var_2660_strides_0, weight = layers_6_self_attn_o_proj_inlier_module_weight_to_fp16_palettized, x = input_179_cast_fp16)[name = tensor("op_2660_cast_fp16")]; - tensor var_2666_pad_type_0 = const()[name = tensor("op_2666_pad_type_0"), val = tensor("valid")]; - tensor var_2666_strides_0 = const()[name = tensor("op_2666_strides_0"), val = tensor([1, 1])]; - tensor var_2666_pad_0 = const()[name = tensor("op_2666_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2666_dilations_0 = const()[name = tensor("op_2666_dilations_0"), val = tensor([1, 1])]; - tensor var_2666_groups_0 = const()[name = tensor("op_2666_groups_0"), val = tensor(1)]; - tensor layers_6_self_attn_o_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(39559936))), name = tensor("layers_6_self_attn_o_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(39551808))), shape = tensor([512, 512, 1, 1])]; - tensor var_2666_cast_fp16 = conv(dilations = var_2666_dilations_0, groups = var_2666_groups_0, pad = var_2666_pad_0, pad_type = var_2666_pad_type_0, strides = var_2666_strides_0, weight = layers_6_self_attn_o_proj_outlier_module_weight_to_fp16_sparsified, x = input_179_cast_fp16)[name = tensor("op_2666_cast_fp16")]; - tensor obj_29_cast_fp16 = add(x = var_2660_cast_fp16, y = var_2666_cast_fp16)[name = tensor("obj_29_cast_fp16")]; - tensor inputs_65_cast_fp16 = add(x = inputs_63_cast_fp16, y = obj_29_cast_fp16)[name = tensor("inputs_65_cast_fp16")]; - tensor out_65_axes_0 = const()[name = tensor("out_65_axes_0"), val = tensor([1])]; - tensor var_2677_to_fp16 = const()[name = tensor("op_2677_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_65_cast_fp16 = layer_norm(axes = out_65_axes_0, epsilon = var_2677_to_fp16, x = inputs_65_cast_fp16)[name = tensor("out_65_cast_fp16")]; - tensor input_181_gamma_0_to_fp16 = const()[name = tensor("input_181_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(39592768)))]; - tensor input_181_beta_0_to_fp16 = const()[name = tensor("input_181_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(39593856)))]; - tensor input_181_epsilon_0_to_fp16 = const()[name = tensor("input_181_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_181_cast_fp16 = batch_norm(beta = input_181_beta_0_to_fp16, epsilon = input_181_epsilon_0_to_fp16, gamma = input_181_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_65_cast_fp16)[name = tensor("input_181_cast_fp16")]; - tensor var_2699_pad_type_0 = const()[name = tensor("op_2699_pad_type_0"), val = tensor("valid")]; - tensor var_2699_strides_0 = const()[name = tensor("op_2699_strides_0"), val = tensor([1, 1])]; - tensor var_2699_pad_0 = const()[name = tensor("op_2699_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2699_dilations_0 = const()[name = tensor("op_2699_dilations_0"), val = tensor([1, 1])]; - tensor var_2699_groups_0 = const()[name = tensor("op_2699_groups_0"), val = tensor(1)]; - tensor layers_6_conv_pointwise_conv1_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(39594944))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(39988224))), name = tensor("layers_6_conv_pointwise_conv1_inlier_module_weight_to_fp16_palettized"), shape = tensor([1024, 512, 1, 1])]; - tensor layers_6_conv_pointwise_conv1_inlier_module_bias_to_fp16 = const()[name = tensor("layers_6_conv_pointwise_conv1_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(39988416)))]; - tensor var_2699_cast_fp16 = conv(bias = layers_6_conv_pointwise_conv1_inlier_module_bias_to_fp16, dilations = var_2699_dilations_0, groups = var_2699_groups_0, pad = var_2699_pad_0, pad_type = var_2699_pad_type_0, strides = var_2699_strides_0, weight = layers_6_conv_pointwise_conv1_inlier_module_weight_to_fp16_palettized, x = input_181_cast_fp16)[name = tensor("op_2699_cast_fp16")]; - tensor var_2705_pad_type_0 = const()[name = tensor("op_2705_pad_type_0"), val = tensor("valid")]; - tensor var_2705_strides_0 = const()[name = tensor("op_2705_strides_0"), val = tensor([1, 1])]; - tensor var_2705_pad_0 = const()[name = tensor("op_2705_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2705_dilations_0 = const()[name = tensor("op_2705_dilations_0"), val = tensor([1, 1])]; - tensor var_2705_groups_0 = const()[name = tensor("op_2705_groups_0"), val = tensor(1)]; - tensor layers_6_conv_pointwise_conv1_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40007936))), name = tensor("layers_6_conv_pointwise_conv1_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(39990528))), shape = tensor([1024, 512, 1, 1])]; - tensor var_2705_cast_fp16 = conv(dilations = var_2705_dilations_0, groups = var_2705_groups_0, pad = var_2705_pad_0, pad_type = var_2705_pad_type_0, strides = var_2705_strides_0, weight = layers_6_conv_pointwise_conv1_outlier_module_weight_to_fp16_sparsified, x = input_181_cast_fp16)[name = tensor("op_2705_cast_fp16")]; - tensor input_183_cast_fp16 = add(x = var_2699_cast_fp16, y = var_2705_cast_fp16)[name = tensor("input_183_cast_fp16")]; - tensor input_185_split_num_splits_0 = const()[name = tensor("input_185_split_num_splits_0"), val = tensor(2)]; - tensor input_185_split_axis_0 = const()[name = tensor("input_185_split_axis_0"), val = tensor(1)]; - tensor input_185_split_cast_fp16_0, tensor input_185_split_cast_fp16_1 = split(axis = input_185_split_axis_0, num_splits = input_185_split_num_splits_0, x = input_183_cast_fp16)[name = tensor("input_185_split_cast_fp16")]; - tensor input_185_split_1_sigmoid_cast_fp16 = sigmoid(x = input_185_split_cast_fp16_1)[name = tensor("input_185_split_1_sigmoid_cast_fp16")]; - tensor input_185_cast_fp16 = mul(x = input_185_split_cast_fp16_0, y = input_185_split_1_sigmoid_cast_fp16)[name = tensor("input_185_cast_fp16")]; - tensor input_187_pad_type_0 = const()[name = tensor("input_187_pad_type_0"), val = tensor("custom")]; - tensor input_187_pad_0 = const()[name = tensor("input_187_pad_0"), val = tensor([0, 0, 4, 4])]; - tensor input_187_groups_0 = const()[name = tensor("input_187_groups_0"), val = tensor(512)]; - tensor input_187_strides_0 = const()[name = tensor("input_187_strides_0"), val = tensor([1, 1])]; - tensor input_187_dilations_0 = const()[name = tensor("input_187_dilations_0"), val = tensor([1, 1])]; - tensor const_203_to_fp16 = const()[name = tensor("const_203_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40073536)))]; - tensor const_204_to_fp16 = const()[name = tensor("const_204_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40082816)))]; - tensor input_189_cast_fp16 = conv(bias = const_204_to_fp16, dilations = input_187_dilations_0, groups = input_187_groups_0, pad = input_187_pad_0, pad_type = input_187_pad_type_0, strides = input_187_strides_0, weight = const_203_to_fp16, x = input_185_cast_fp16)[name = tensor("input_189_cast_fp16")]; - tensor input_191_cast_fp16 = silu(x = input_189_cast_fp16)[name = tensor("input_191_cast_fp16")]; - tensor var_2729_pad_type_0 = const()[name = tensor("op_2729_pad_type_0"), val = tensor("valid")]; - tensor var_2729_strides_0 = const()[name = tensor("op_2729_strides_0"), val = tensor([1, 1])]; - tensor var_2729_pad_0 = const()[name = tensor("op_2729_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2729_dilations_0 = const()[name = tensor("op_2729_dilations_0"), val = tensor([1, 1])]; - tensor var_2729_groups_0 = const()[name = tensor("op_2729_groups_0"), val = tensor(1)]; - tensor layers_6_conv_pointwise_conv2_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40083904))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40280576))), name = tensor("layers_6_conv_pointwise_conv2_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_6_conv_pointwise_conv2_inlier_module_bias_to_fp16 = const()[name = tensor("layers_6_conv_pointwise_conv2_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40280768)))]; - tensor var_2729_cast_fp16 = conv(bias = layers_6_conv_pointwise_conv2_inlier_module_bias_to_fp16, dilations = var_2729_dilations_0, groups = var_2729_groups_0, pad = var_2729_pad_0, pad_type = var_2729_pad_type_0, strides = var_2729_strides_0, weight = layers_6_conv_pointwise_conv2_inlier_module_weight_to_fp16_palettized, x = input_191_cast_fp16)[name = tensor("op_2729_cast_fp16")]; - tensor var_2735_pad_type_0 = const()[name = tensor("op_2735_pad_type_0"), val = tensor("valid")]; - tensor var_2735_strides_0 = const()[name = tensor("op_2735_strides_0"), val = tensor([1, 1])]; - tensor var_2735_pad_0 = const()[name = tensor("op_2735_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2735_dilations_0 = const()[name = tensor("op_2735_dilations_0"), val = tensor([1, 1])]; - tensor var_2735_groups_0 = const()[name = tensor("op_2735_groups_0"), val = tensor(1)]; - tensor layers_6_conv_pointwise_conv2_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40290496))), name = tensor("layers_6_conv_pointwise_conv2_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40281856))), shape = tensor([512, 512, 1, 1])]; - tensor var_2735_cast_fp16 = conv(dilations = var_2735_dilations_0, groups = var_2735_groups_0, pad = var_2735_pad_0, pad_type = var_2735_pad_type_0, strides = var_2735_strides_0, weight = layers_6_conv_pointwise_conv2_outlier_module_weight_to_fp16_sparsified, x = input_191_cast_fp16)[name = tensor("op_2735_cast_fp16")]; - tensor x_41_cast_fp16 = add(x = var_2729_cast_fp16, y = var_2735_cast_fp16)[name = tensor("x_41_cast_fp16")]; - tensor inputs_67_cast_fp16 = add(x = inputs_65_cast_fp16, y = x_41_cast_fp16)[name = tensor("inputs_67_cast_fp16")]; - tensor out_67_axes_0 = const()[name = tensor("out_67_axes_0"), val = tensor([1])]; - tensor var_2746_to_fp16 = const()[name = tensor("op_2746_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_67_cast_fp16 = layer_norm(axes = out_67_axes_0, epsilon = var_2746_to_fp16, x = inputs_67_cast_fp16)[name = tensor("out_67_cast_fp16")]; - tensor input_193_gamma_0_to_fp16 = const()[name = tensor("input_193_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40323328)))]; - tensor input_193_beta_0_to_fp16 = const()[name = tensor("input_193_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40324416)))]; - tensor input_193_epsilon_0_to_fp16 = const()[name = tensor("input_193_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_193_cast_fp16 = batch_norm(beta = input_193_beta_0_to_fp16, epsilon = input_193_epsilon_0_to_fp16, gamma = input_193_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_67_cast_fp16)[name = tensor("input_193_cast_fp16")]; - tensor var_2766_pad_type_0 = const()[name = tensor("op_2766_pad_type_0"), val = tensor("valid")]; - tensor var_2766_strides_0 = const()[name = tensor("op_2766_strides_0"), val = tensor([1, 1])]; - tensor var_2766_pad_0 = const()[name = tensor("op_2766_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2766_dilations_0 = const()[name = tensor("op_2766_dilations_0"), val = tensor([1, 1])]; - tensor var_2766_groups_0 = const()[name = tensor("op_2766_groups_0"), val = tensor(1)]; - tensor layers_6_feed_forward2_fc1_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40325504))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(41112000))), name = tensor("layers_6_feed_forward2_fc1_inlier_module_weight_to_fp16_palettized"), shape = tensor([2048, 512, 1, 1])]; - tensor layers_6_feed_forward2_fc1_inlier_module_bias_to_fp16 = const()[name = tensor("layers_6_feed_forward2_fc1_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(41112192)))]; - tensor var_2766_cast_fp16 = conv(bias = layers_6_feed_forward2_fc1_inlier_module_bias_to_fp16, dilations = var_2766_dilations_0, groups = var_2766_groups_0, pad = var_2766_pad_0, pad_type = var_2766_pad_type_0, strides = var_2766_strides_0, weight = layers_6_feed_forward2_fc1_inlier_module_weight_to_fp16_palettized, x = input_193_cast_fp16)[name = tensor("op_2766_cast_fp16")]; - tensor var_2772_pad_type_0 = const()[name = tensor("op_2772_pad_type_0"), val = tensor("valid")]; - tensor var_2772_strides_0 = const()[name = tensor("op_2772_strides_0"), val = tensor([1, 1])]; - tensor var_2772_pad_0 = const()[name = tensor("op_2772_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2772_dilations_0 = const()[name = tensor("op_2772_dilations_0"), val = tensor([1, 1])]; - tensor var_2772_groups_0 = const()[name = tensor("op_2772_groups_0"), val = tensor(1)]; - tensor layers_6_feed_forward2_fc1_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(41149888))), name = tensor("layers_6_feed_forward2_fc1_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(41116352))), shape = tensor([2048, 512, 1, 1])]; - tensor var_2772_cast_fp16 = conv(dilations = var_2772_dilations_0, groups = var_2772_groups_0, pad = var_2772_pad_0, pad_type = var_2772_pad_type_0, strides = var_2772_strides_0, weight = layers_6_feed_forward2_fc1_outlier_module_weight_to_fp16_sparsified, x = input_193_cast_fp16)[name = tensor("op_2772_cast_fp16")]; - tensor input_195_cast_fp16 = add(x = var_2766_cast_fp16, y = var_2772_cast_fp16)[name = tensor("input_195_cast_fp16")]; - tensor input_197_cast_fp16 = silu(x = input_195_cast_fp16)[name = tensor("input_197_cast_fp16")]; - tensor var_2783_pad_type_0 = const()[name = tensor("op_2783_pad_type_0"), val = tensor("valid")]; - tensor var_2783_strides_0 = const()[name = tensor("op_2783_strides_0"), val = tensor([1, 1])]; - tensor var_2783_pad_0 = const()[name = tensor("op_2783_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2783_dilations_0 = const()[name = tensor("op_2783_dilations_0"), val = tensor([1, 1])]; - tensor var_2783_groups_0 = const()[name = tensor("op_2783_groups_0"), val = tensor(1)]; - tensor layers_6_feed_forward2_fc2_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(41281024))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(42067520))), name = tensor("layers_6_feed_forward2_fc2_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 2048, 1, 1])]; - tensor layers_6_feed_forward2_fc2_inlier_module_bias_to_fp16 = const()[name = tensor("layers_6_feed_forward2_fc2_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(42067712)))]; - tensor var_2783_cast_fp16 = conv(bias = layers_6_feed_forward2_fc2_inlier_module_bias_to_fp16, dilations = var_2783_dilations_0, groups = var_2783_groups_0, pad = var_2783_pad_0, pad_type = var_2783_pad_type_0, strides = var_2783_strides_0, weight = layers_6_feed_forward2_fc2_inlier_module_weight_to_fp16_palettized, x = input_197_cast_fp16)[name = tensor("op_2783_cast_fp16")]; - tensor var_2789_pad_type_0 = const()[name = tensor("op_2789_pad_type_0"), val = tensor("valid")]; - tensor var_2789_strides_0 = const()[name = tensor("op_2789_strides_0"), val = tensor([1, 1])]; - tensor var_2789_pad_0 = const()[name = tensor("op_2789_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2789_dilations_0 = const()[name = tensor("op_2789_dilations_0"), val = tensor([1, 1])]; - tensor var_2789_groups_0 = const()[name = tensor("op_2789_groups_0"), val = tensor(1)]; - tensor layers_6_feed_forward2_fc2_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(42107840))), name = tensor("layers_6_feed_forward2_fc2_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(42068800))), shape = tensor([512, 2048, 1, 1])]; - tensor var_2789_cast_fp16 = conv(dilations = var_2789_dilations_0, groups = var_2789_groups_0, pad = var_2789_pad_0, pad_type = var_2789_pad_type_0, strides = var_2789_strides_0, weight = layers_6_feed_forward2_fc2_outlier_module_weight_to_fp16_sparsified, x = input_197_cast_fp16)[name = tensor("op_2789_cast_fp16")]; - tensor x_43_cast_fp16 = add(x = var_2783_cast_fp16, y = var_2789_cast_fp16)[name = tensor("x_43_cast_fp16")]; - tensor var_2791_to_fp16 = const()[name = tensor("op_2791_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2792_cast_fp16 = mul(x = x_43_cast_fp16, y = var_2791_to_fp16)[name = tensor("op_2792_cast_fp16")]; - tensor inputs_69_cast_fp16 = add(x = inputs_67_cast_fp16, y = var_2792_cast_fp16)[name = tensor("inputs_69_cast_fp16")]; - tensor out_69_axes_0 = const()[name = tensor("out_69_axes_0"), val = tensor([1])]; - tensor var_2802_to_fp16 = const()[name = tensor("op_2802_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_69_cast_fp16 = layer_norm(axes = out_69_axes_0, epsilon = var_2802_to_fp16, x = inputs_69_cast_fp16)[name = tensor("out_69_cast_fp16")]; - tensor inputs_71_gamma_0_to_fp16 = const()[name = tensor("inputs_71_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(42238976)))]; - tensor inputs_71_beta_0_to_fp16 = const()[name = tensor("inputs_71_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(42240064)))]; - tensor inputs_71_epsilon_0_to_fp16 = const()[name = tensor("inputs_71_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor inputs_71_cast_fp16 = batch_norm(beta = inputs_71_beta_0_to_fp16, epsilon = inputs_71_epsilon_0_to_fp16, gamma = inputs_71_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_69_cast_fp16)[name = tensor("inputs_71_cast_fp16")]; - tensor var_2816 = const()[name = tensor("op_2816"), val = tensor(3)]; - tensor out_71_axes_0 = const()[name = tensor("out_71_axes_0"), val = tensor([1])]; - tensor var_2847_to_fp16 = const()[name = tensor("op_2847_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_71_cast_fp16 = layer_norm(axes = out_71_axes_0, epsilon = var_2847_to_fp16, x = inputs_71_cast_fp16)[name = tensor("out_71_cast_fp16")]; - tensor input_199_gamma_0_to_fp16 = const()[name = tensor("input_199_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(42241152)))]; - tensor input_199_beta_0_to_fp16 = const()[name = tensor("input_199_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(42242240)))]; - tensor input_199_epsilon_0_to_fp16 = const()[name = tensor("input_199_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_199_cast_fp16 = batch_norm(beta = input_199_beta_0_to_fp16, epsilon = input_199_epsilon_0_to_fp16, gamma = input_199_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_71_cast_fp16)[name = tensor("input_199_cast_fp16")]; - tensor var_2867_pad_type_0 = const()[name = tensor("op_2867_pad_type_0"), val = tensor("valid")]; - tensor var_2867_strides_0 = const()[name = tensor("op_2867_strides_0"), val = tensor([1, 1])]; - tensor var_2867_pad_0 = const()[name = tensor("op_2867_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2867_dilations_0 = const()[name = tensor("op_2867_dilations_0"), val = tensor([1, 1])]; - tensor var_2867_groups_0 = const()[name = tensor("op_2867_groups_0"), val = tensor(1)]; - tensor layers_7_feed_forward1_fc1_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(42243328))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(43029824))), name = tensor("layers_7_feed_forward1_fc1_inlier_module_weight_to_fp16_palettized"), shape = tensor([2048, 512, 1, 1])]; - tensor layers_7_feed_forward1_fc1_inlier_module_bias_to_fp16 = const()[name = tensor("layers_7_feed_forward1_fc1_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(43030016)))]; - tensor var_2867_cast_fp16 = conv(bias = layers_7_feed_forward1_fc1_inlier_module_bias_to_fp16, dilations = var_2867_dilations_0, groups = var_2867_groups_0, pad = var_2867_pad_0, pad_type = var_2867_pad_type_0, strides = var_2867_strides_0, weight = layers_7_feed_forward1_fc1_inlier_module_weight_to_fp16_palettized, x = input_199_cast_fp16)[name = tensor("op_2867_cast_fp16")]; - tensor var_2873_pad_type_0 = const()[name = tensor("op_2873_pad_type_0"), val = tensor("valid")]; - tensor var_2873_strides_0 = const()[name = tensor("op_2873_strides_0"), val = tensor([1, 1])]; - tensor var_2873_pad_0 = const()[name = tensor("op_2873_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2873_dilations_0 = const()[name = tensor("op_2873_dilations_0"), val = tensor([1, 1])]; - tensor var_2873_groups_0 = const()[name = tensor("op_2873_groups_0"), val = tensor(1)]; - tensor layers_7_feed_forward1_fc1_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(43068288))), name = tensor("layers_7_feed_forward1_fc1_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(43034176))), shape = tensor([2048, 512, 1, 1])]; - tensor var_2873_cast_fp16 = conv(dilations = var_2873_dilations_0, groups = var_2873_groups_0, pad = var_2873_pad_0, pad_type = var_2873_pad_type_0, strides = var_2873_strides_0, weight = layers_7_feed_forward1_fc1_outlier_module_weight_to_fp16_sparsified, x = input_199_cast_fp16)[name = tensor("op_2873_cast_fp16")]; - tensor input_201_cast_fp16 = add(x = var_2867_cast_fp16, y = var_2873_cast_fp16)[name = tensor("input_201_cast_fp16")]; - tensor input_203_cast_fp16 = silu(x = input_201_cast_fp16)[name = tensor("input_203_cast_fp16")]; - tensor var_2884_pad_type_0 = const()[name = tensor("op_2884_pad_type_0"), val = tensor("valid")]; - tensor var_2884_strides_0 = const()[name = tensor("op_2884_strides_0"), val = tensor([1, 1])]; - tensor var_2884_pad_0 = const()[name = tensor("op_2884_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2884_dilations_0 = const()[name = tensor("op_2884_dilations_0"), val = tensor([1, 1])]; - tensor var_2884_groups_0 = const()[name = tensor("op_2884_groups_0"), val = tensor(1)]; - tensor layers_7_feed_forward1_fc2_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(43199424))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(43985920))), name = tensor("layers_7_feed_forward1_fc2_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 2048, 1, 1])]; - tensor layers_7_feed_forward1_fc2_inlier_module_bias_to_fp16 = const()[name = tensor("layers_7_feed_forward1_fc2_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(43986112)))]; - tensor var_2884_cast_fp16 = conv(bias = layers_7_feed_forward1_fc2_inlier_module_bias_to_fp16, dilations = var_2884_dilations_0, groups = var_2884_groups_0, pad = var_2884_pad_0, pad_type = var_2884_pad_type_0, strides = var_2884_strides_0, weight = layers_7_feed_forward1_fc2_inlier_module_weight_to_fp16_palettized, x = input_203_cast_fp16)[name = tensor("op_2884_cast_fp16")]; - tensor var_2890_pad_type_0 = const()[name = tensor("op_2890_pad_type_0"), val = tensor("valid")]; - tensor var_2890_strides_0 = const()[name = tensor("op_2890_strides_0"), val = tensor([1, 1])]; - tensor var_2890_pad_0 = const()[name = tensor("op_2890_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2890_dilations_0 = const()[name = tensor("op_2890_dilations_0"), val = tensor([1, 1])]; - tensor var_2890_groups_0 = const()[name = tensor("op_2890_groups_0"), val = tensor(1)]; - tensor layers_7_feed_forward1_fc2_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44027328))), name = tensor("layers_7_feed_forward1_fc2_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(43987200))), shape = tensor([512, 2048, 1, 1])]; - tensor var_2890_cast_fp16 = conv(dilations = var_2890_dilations_0, groups = var_2890_groups_0, pad = var_2890_pad_0, pad_type = var_2890_pad_type_0, strides = var_2890_strides_0, weight = layers_7_feed_forward1_fc2_outlier_module_weight_to_fp16_sparsified, x = input_203_cast_fp16)[name = tensor("op_2890_cast_fp16")]; - tensor x_45_cast_fp16 = add(x = var_2884_cast_fp16, y = var_2890_cast_fp16)[name = tensor("x_45_cast_fp16")]; - tensor var_2892_to_fp16 = const()[name = tensor("op_2892_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2893_cast_fp16 = mul(x = x_45_cast_fp16, y = var_2892_to_fp16)[name = tensor("op_2893_cast_fp16")]; - tensor inputs_73_cast_fp16 = add(x = inputs_71_cast_fp16, y = var_2893_cast_fp16)[name = tensor("inputs_73_cast_fp16")]; - tensor out_73_axes_0 = const()[name = tensor("out_73_axes_0"), val = tensor([1])]; - tensor var_2903_to_fp16 = const()[name = tensor("op_2903_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_73_cast_fp16 = layer_norm(axes = out_73_axes_0, epsilon = var_2903_to_fp16, x = inputs_73_cast_fp16)[name = tensor("out_73_cast_fp16")]; - tensor obj_31_gamma_0_to_fp16 = const()[name = tensor("obj_31_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44158464)))]; - tensor obj_31_beta_0_to_fp16 = const()[name = tensor("obj_31_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44159552)))]; - tensor obj_31_epsilon_0_to_fp16 = const()[name = tensor("obj_31_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor obj_31_cast_fp16 = batch_norm(beta = obj_31_beta_0_to_fp16, epsilon = obj_31_epsilon_0_to_fp16, gamma = obj_31_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_73_cast_fp16)[name = tensor("obj_31_cast_fp16")]; - tensor var_2928_pad_type_0 = const()[name = tensor("op_2928_pad_type_0"), val = tensor("valid")]; - tensor var_2928_strides_0 = const()[name = tensor("op_2928_strides_0"), val = tensor([1, 1])]; - tensor var_2928_pad_0 = const()[name = tensor("op_2928_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2928_dilations_0 = const()[name = tensor("op_2928_dilations_0"), val = tensor([1, 1])]; - tensor var_2928_groups_0 = const()[name = tensor("op_2928_groups_0"), val = tensor(1)]; - tensor layers_7_self_attn_q_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44160640))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44357312))), name = tensor("layers_7_self_attn_q_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_7_self_attn_q_proj_inlier_module_bias_to_fp16 = const()[name = tensor("layers_7_self_attn_q_proj_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44357504)))]; - tensor var_2928_cast_fp16 = conv(bias = layers_7_self_attn_q_proj_inlier_module_bias_to_fp16, dilations = var_2928_dilations_0, groups = var_2928_groups_0, pad = var_2928_pad_0, pad_type = var_2928_pad_type_0, strides = var_2928_strides_0, weight = layers_7_self_attn_q_proj_inlier_module_weight_to_fp16_palettized, x = obj_31_cast_fp16)[name = tensor("op_2928_cast_fp16")]; - tensor var_2934_pad_type_0 = const()[name = tensor("op_2934_pad_type_0"), val = tensor("valid")]; - tensor var_2934_strides_0 = const()[name = tensor("op_2934_strides_0"), val = tensor([1, 1])]; - tensor var_2934_pad_0 = const()[name = tensor("op_2934_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2934_dilations_0 = const()[name = tensor("op_2934_dilations_0"), val = tensor([1, 1])]; - tensor var_2934_groups_0 = const()[name = tensor("op_2934_groups_0"), val = tensor(1)]; - tensor layers_7_self_attn_q_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44368832))), name = tensor("layers_7_self_attn_q_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44358592))), shape = tensor([512, 512, 1, 1])]; - tensor var_2934_cast_fp16 = conv(dilations = var_2934_dilations_0, groups = var_2934_groups_0, pad = var_2934_pad_0, pad_type = var_2934_pad_type_0, strides = var_2934_strides_0, weight = layers_7_self_attn_q_proj_outlier_module_weight_to_fp16_sparsified, x = obj_31_cast_fp16)[name = tensor("op_2934_cast_fp16")]; - tensor query_29_cast_fp16 = add(x = var_2928_cast_fp16, y = var_2934_cast_fp16)[name = tensor("query_29_cast_fp16")]; - tensor var_2943_pad_type_0 = const()[name = tensor("op_2943_pad_type_0"), val = tensor("valid")]; - tensor var_2943_strides_0 = const()[name = tensor("op_2943_strides_0"), val = tensor([1, 1])]; - tensor var_2943_pad_0 = const()[name = tensor("op_2943_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2943_dilations_0 = const()[name = tensor("op_2943_dilations_0"), val = tensor([1, 1])]; - tensor var_2943_groups_0 = const()[name = tensor("op_2943_groups_0"), val = tensor(1)]; - tensor layers_7_self_attn_k_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44401664))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44598336))), name = tensor("layers_7_self_attn_k_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor var_2943_cast_fp16 = conv(dilations = var_2943_dilations_0, groups = var_2943_groups_0, pad = var_2943_pad_0, pad_type = var_2943_pad_type_0, strides = var_2943_strides_0, weight = layers_7_self_attn_k_proj_inlier_module_weight_to_fp16_palettized, x = obj_31_cast_fp16)[name = tensor("op_2943_cast_fp16")]; - tensor var_2949_pad_type_0 = const()[name = tensor("op_2949_pad_type_0"), val = tensor("valid")]; - tensor var_2949_strides_0 = const()[name = tensor("op_2949_strides_0"), val = tensor([1, 1])]; - tensor var_2949_pad_0 = const()[name = tensor("op_2949_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2949_dilations_0 = const()[name = tensor("op_2949_dilations_0"), val = tensor([1, 1])]; - tensor var_2949_groups_0 = const()[name = tensor("op_2949_groups_0"), val = tensor(1)]; - tensor layers_7_self_attn_k_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44611072))), name = tensor("layers_7_self_attn_k_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44598528))), shape = tensor([512, 512, 1, 1])]; - tensor var_2949_cast_fp16 = conv(dilations = var_2949_dilations_0, groups = var_2949_groups_0, pad = var_2949_pad_0, pad_type = var_2949_pad_type_0, strides = var_2949_strides_0, weight = layers_7_self_attn_k_proj_outlier_module_weight_to_fp16_sparsified, x = obj_31_cast_fp16)[name = tensor("op_2949_cast_fp16")]; - tensor key_15_cast_fp16 = add(x = var_2943_cast_fp16, y = var_2949_cast_fp16)[name = tensor("key_15_cast_fp16")]; - tensor var_2959_pad_type_0 = const()[name = tensor("op_2959_pad_type_0"), val = tensor("valid")]; - tensor var_2959_strides_0 = const()[name = tensor("op_2959_strides_0"), val = tensor([1, 1])]; - tensor var_2959_pad_0 = const()[name = tensor("op_2959_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2959_dilations_0 = const()[name = tensor("op_2959_dilations_0"), val = tensor([1, 1])]; - tensor var_2959_groups_0 = const()[name = tensor("op_2959_groups_0"), val = tensor(1)]; - tensor layers_7_self_attn_v_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44643904))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44840576))), name = tensor("layers_7_self_attn_v_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_7_self_attn_v_proj_inlier_module_bias_to_fp16 = const()[name = tensor("layers_7_self_attn_v_proj_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44840768)))]; - tensor var_2959_cast_fp16 = conv(bias = layers_7_self_attn_v_proj_inlier_module_bias_to_fp16, dilations = var_2959_dilations_0, groups = var_2959_groups_0, pad = var_2959_pad_0, pad_type = var_2959_pad_type_0, strides = var_2959_strides_0, weight = layers_7_self_attn_v_proj_inlier_module_weight_to_fp16_palettized, x = obj_31_cast_fp16)[name = tensor("op_2959_cast_fp16")]; - tensor var_2965_pad_type_0 = const()[name = tensor("op_2965_pad_type_0"), val = tensor("valid")]; - tensor var_2965_strides_0 = const()[name = tensor("op_2965_strides_0"), val = tensor([1, 1])]; - tensor var_2965_pad_0 = const()[name = tensor("op_2965_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2965_dilations_0 = const()[name = tensor("op_2965_dilations_0"), val = tensor([1, 1])]; - tensor var_2965_groups_0 = const()[name = tensor("op_2965_groups_0"), val = tensor(1)]; - tensor layers_7_self_attn_v_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44851456))), name = tensor("layers_7_self_attn_v_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44841856))), shape = tensor([512, 512, 1, 1])]; - tensor var_2965_cast_fp16 = conv(dilations = var_2965_dilations_0, groups = var_2965_groups_0, pad = var_2965_pad_0, pad_type = var_2965_pad_type_0, strides = var_2965_strides_0, weight = layers_7_self_attn_v_proj_outlier_module_weight_to_fp16_sparsified, x = obj_31_cast_fp16)[name = tensor("op_2965_cast_fp16")]; - tensor value_15_cast_fp16 = add(x = var_2959_cast_fp16, y = var_2965_cast_fp16)[name = tensor("value_15_cast_fp16")]; - tensor var_2968_to_fp16 = const()[name = tensor("op_2968_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44884288)))]; - tensor query_31_cast_fp16 = add(x = query_29_cast_fp16, y = var_2968_to_fp16)[name = tensor("query_31_cast_fp16")]; - tensor var_2971_to_fp16 = const()[name = tensor("op_2971_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44885376)))]; - tensor q_with_bias_v_15_cast_fp16 = add(x = query_29_cast_fp16, y = var_2971_to_fp16)[name = tensor("q_with_bias_v_15_cast_fp16")]; - tensor var_2981_pad_type_0 = const()[name = tensor("op_2981_pad_type_0"), val = tensor("valid")]; - tensor var_2981_strides_0 = const()[name = tensor("op_2981_strides_0"), val = tensor([1, 1])]; - tensor var_2981_pad_0 = const()[name = tensor("op_2981_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2981_dilations_0 = const()[name = tensor("op_2981_dilations_0"), val = tensor([1, 1])]; - tensor var_2981_groups_0 = const()[name = tensor("op_2981_groups_0"), val = tensor(1)]; - tensor layers_7_self_attn_linear_pos_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44886464))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45083136))), name = tensor("layers_7_self_attn_linear_pos_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor var_2981_cast_fp16 = conv(bias = input_17_mean_0_to_fp16, dilations = var_2981_dilations_0, groups = var_2981_groups_0, pad = var_2981_pad_0, pad_type = var_2981_pad_type_0, strides = var_2981_strides_0, weight = layers_7_self_attn_linear_pos_inlier_module_weight_to_fp16_palettized, x = obj_3_cast_fp16)[name = tensor("op_2981_cast_fp16")]; - tensor var_2987_pad_type_0 = const()[name = tensor("op_2987_pad_type_0"), val = tensor("valid")]; - tensor var_2987_strides_0 = const()[name = tensor("op_2987_strides_0"), val = tensor([1, 1])]; - tensor var_2987_pad_0 = const()[name = tensor("op_2987_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_2987_dilations_0 = const()[name = tensor("op_2987_dilations_0"), val = tensor([1, 1])]; - tensor var_2987_groups_0 = const()[name = tensor("op_2987_groups_0"), val = tensor(1)]; - tensor layers_7_self_attn_linear_pos_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45102592))), name = tensor("layers_7_self_attn_linear_pos_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45083328))), shape = tensor([512, 512, 1, 1])]; - tensor var_2987_cast_fp16 = conv(dilations = var_2987_dilations_0, groups = var_2987_groups_0, pad = var_2987_pad_0, pad_type = var_2987_pad_type_0, strides = var_2987_strides_0, weight = layers_7_self_attn_linear_pos_outlier_module_weight_to_fp16_sparsified, x = obj_3_cast_fp16)[name = tensor("op_2987_cast_fp16")]; - tensor p_15_cast_fp16 = add(x = var_2981_cast_fp16, y = var_2987_cast_fp16)[name = tensor("p_15_cast_fp16")]; - tensor var_2991 = const()[name = tensor("op_2991"), val = tensor([1, 8, 64, 188])]; - tensor var_2992_cast_fp16 = reshape(shape = var_2991, x = q_with_bias_v_15_cast_fp16)[name = tensor("op_2992_cast_fp16")]; - tensor var_2993 = const()[name = tensor("op_2993"), val = tensor([1, 8, 64, -1])]; - tensor var_2994_cast_fp16 = reshape(shape = var_2993, x = p_15_cast_fp16)[name = tensor("op_2994_cast_fp16")]; - tensor matrix_bd_57_transpose_x_0 = const()[name = tensor("matrix_bd_57_transpose_x_0"), val = tensor(true)]; - tensor matrix_bd_57_transpose_y_0 = const()[name = tensor("matrix_bd_57_transpose_y_0"), val = tensor(false)]; - tensor matrix_bd_57_cast_fp16 = matmul(transpose_x = matrix_bd_57_transpose_x_0, transpose_y = matrix_bd_57_transpose_y_0, x = var_2992_cast_fp16, y = var_2994_cast_fp16)[name = tensor("matrix_bd_57_cast_fp16")]; - tensor matrix_bd_59_pad_0 = const()[name = tensor("matrix_bd_59_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor matrix_bd_59_mode_0 = const()[name = tensor("matrix_bd_59_mode_0"), val = tensor("constant")]; - tensor const_87_to_fp16 = const()[name = tensor("const_87_to_fp16"), val = tensor(0x0p+0)]; - tensor matrix_bd_59_cast_fp16 = pad(constant_val = const_87_to_fp16, mode = matrix_bd_59_mode_0, pad = matrix_bd_59_pad_0, x = matrix_bd_57_cast_fp16)[name = tensor("matrix_bd_59_cast_fp16")]; - tensor var_3003 = const()[name = tensor("op_3003"), val = tensor([1, 8, -1, 188])]; - tensor matrix_bd_61_cast_fp16 = reshape(shape = var_3003, x = matrix_bd_59_cast_fp16)[name = tensor("matrix_bd_61_cast_fp16")]; - tensor var_3007_begin_0 = const()[name = tensor("op_3007_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_3007_end_0 = const()[name = tensor("op_3007_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_3007_end_mask_0 = const()[name = tensor("op_3007_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_3007_cast_fp16 = slice_by_index(begin = var_3007_begin_0, end = var_3007_end_0, end_mask = var_3007_end_mask_0, x = matrix_bd_61_cast_fp16)[name = tensor("op_3007_cast_fp16")]; - tensor var_3008 = const()[name = tensor("op_3008"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_63_cast_fp16 = reshape(shape = var_3008, x = var_3007_cast_fp16)[name = tensor("matrix_bd_63_cast_fp16")]; - tensor var_3013_begin_0 = const()[name = tensor("op_3013_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3013_end_0 = const()[name = tensor("op_3013_end_0"), val = tensor([1, 8, 188, 188])]; - tensor var_3013_end_mask_0 = const()[name = tensor("op_3013_end_mask_0"), val = tensor([true, true, true, false])]; - tensor var_3013_cast_fp16 = slice_by_index(begin = var_3013_begin_0, end = var_3013_end_0, end_mask = var_3013_end_mask_0, x = matrix_bd_63_cast_fp16)[name = tensor("op_3013_cast_fp16")]; - tensor var_3014_to_fp16 = const()[name = tensor("op_3014_to_fp16"), val = tensor(0x1p-3)]; - tensor qk_mask_15_cast_fp16 = mul(x = var_3013_cast_fp16, y = var_3014_to_fp16)[name = tensor("qk_mask_15_cast_fp16")]; - tensor var_3018 = const()[name = tensor("op_3018"), val = tensor([1, 8, 64, 188])]; - tensor mh_q_15_cast_fp16 = reshape(shape = var_3018, x = query_31_cast_fp16)[name = tensor("mh_q_15_cast_fp16")]; - tensor var_3020_to_fp16 = const()[name = tensor("op_3020_to_fp16"), val = tensor(0x1p-3)]; - tensor var_3021_cast_fp16 = mul(x = mh_q_15_cast_fp16, y = var_3020_to_fp16)[name = tensor("op_3021_cast_fp16")]; - tensor var_3024 = const()[name = tensor("op_3024"), val = tensor([1, 8, 64, 188])]; - tensor var_3025_cast_fp16 = reshape(shape = var_3024, x = key_15_cast_fp16)[name = tensor("op_3025_cast_fp16")]; - tensor mh_w_29_transpose_x_0 = const()[name = tensor("mh_w_29_transpose_x_0"), val = tensor(true)]; - tensor mh_w_29_transpose_y_0 = const()[name = tensor("mh_w_29_transpose_y_0"), val = tensor(false)]; - tensor mh_w_29_cast_fp16 = matmul(transpose_x = mh_w_29_transpose_x_0, transpose_y = mh_w_29_transpose_y_0, x = var_3021_cast_fp16, y = var_3025_cast_fp16)[name = tensor("mh_w_29_cast_fp16")]; - tensor mh_w_31_cast_fp16 = add(x = mh_w_29_cast_fp16, y = qk_mask_15_cast_fp16)[name = tensor("mh_w_31_cast_fp16")]; - tensor var_3029_cast_fp16 = softmax(axis = var_2816, x = mh_w_31_cast_fp16)[name = tensor("op_3029_cast_fp16")]; - tensor var_3030 = const()[name = tensor("op_3030"), val = tensor([1, 8, 64, 188])]; - tensor var_3031_cast_fp16 = reshape(shape = var_3030, x = value_15_cast_fp16)[name = tensor("op_3031_cast_fp16")]; - tensor attn_15_transpose_x_0 = const()[name = tensor("attn_15_transpose_x_0"), val = tensor(false)]; - tensor attn_15_transpose_y_0 = const()[name = tensor("attn_15_transpose_y_0"), val = tensor(true)]; - tensor attn_15_cast_fp16 = matmul(transpose_x = attn_15_transpose_x_0, transpose_y = attn_15_transpose_y_0, x = var_3031_cast_fp16, y = var_3029_cast_fp16)[name = tensor("attn_15_cast_fp16")]; - tensor var_3034 = const()[name = tensor("op_3034"), val = tensor([1, 512, 1, 188])]; - tensor input_205_cast_fp16 = reshape(shape = var_3034, x = attn_15_cast_fp16)[name = tensor("input_205_cast_fp16")]; - tensor var_3044_pad_type_0 = const()[name = tensor("op_3044_pad_type_0"), val = tensor("valid")]; - tensor var_3044_strides_0 = const()[name = tensor("op_3044_strides_0"), val = tensor([1, 1])]; - tensor var_3044_pad_0 = const()[name = tensor("op_3044_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3044_dilations_0 = const()[name = tensor("op_3044_dilations_0"), val = tensor([1, 1])]; - tensor var_3044_groups_0 = const()[name = tensor("op_3044_groups_0"), val = tensor(1)]; - tensor layers_7_self_attn_o_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45135424))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45332096))), name = tensor("layers_7_self_attn_o_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_7_self_attn_o_proj_inlier_module_bias_to_fp16 = const()[name = tensor("layers_7_self_attn_o_proj_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45332288)))]; - tensor var_3044_cast_fp16 = conv(bias = layers_7_self_attn_o_proj_inlier_module_bias_to_fp16, dilations = var_3044_dilations_0, groups = var_3044_groups_0, pad = var_3044_pad_0, pad_type = var_3044_pad_type_0, strides = var_3044_strides_0, weight = layers_7_self_attn_o_proj_inlier_module_weight_to_fp16_palettized, x = input_205_cast_fp16)[name = tensor("op_3044_cast_fp16")]; - tensor var_3050_pad_type_0 = const()[name = tensor("op_3050_pad_type_0"), val = tensor("valid")]; - tensor var_3050_strides_0 = const()[name = tensor("op_3050_strides_0"), val = tensor([1, 1])]; - tensor var_3050_pad_0 = const()[name = tensor("op_3050_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3050_dilations_0 = const()[name = tensor("op_3050_dilations_0"), val = tensor([1, 1])]; - tensor var_3050_groups_0 = const()[name = tensor("op_3050_groups_0"), val = tensor(1)]; - tensor layers_7_self_attn_o_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45342656))), name = tensor("layers_7_self_attn_o_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45333376))), shape = tensor([512, 512, 1, 1])]; - tensor var_3050_cast_fp16 = conv(dilations = var_3050_dilations_0, groups = var_3050_groups_0, pad = var_3050_pad_0, pad_type = var_3050_pad_type_0, strides = var_3050_strides_0, weight = layers_7_self_attn_o_proj_outlier_module_weight_to_fp16_sparsified, x = input_205_cast_fp16)[name = tensor("op_3050_cast_fp16")]; - tensor obj_33_cast_fp16 = add(x = var_3044_cast_fp16, y = var_3050_cast_fp16)[name = tensor("obj_33_cast_fp16")]; - tensor inputs_75_cast_fp16 = add(x = inputs_73_cast_fp16, y = obj_33_cast_fp16)[name = tensor("inputs_75_cast_fp16")]; - tensor out_75_axes_0 = const()[name = tensor("out_75_axes_0"), val = tensor([1])]; - tensor var_3061_to_fp16 = const()[name = tensor("op_3061_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_75_cast_fp16 = layer_norm(axes = out_75_axes_0, epsilon = var_3061_to_fp16, x = inputs_75_cast_fp16)[name = tensor("out_75_cast_fp16")]; - tensor input_207_gamma_0_to_fp16 = const()[name = tensor("input_207_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45375488)))]; - tensor input_207_beta_0_to_fp16 = const()[name = tensor("input_207_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45376576)))]; - tensor input_207_epsilon_0_to_fp16 = const()[name = tensor("input_207_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_207_cast_fp16 = batch_norm(beta = input_207_beta_0_to_fp16, epsilon = input_207_epsilon_0_to_fp16, gamma = input_207_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_75_cast_fp16)[name = tensor("input_207_cast_fp16")]; - tensor var_3083_pad_type_0 = const()[name = tensor("op_3083_pad_type_0"), val = tensor("valid")]; - tensor var_3083_strides_0 = const()[name = tensor("op_3083_strides_0"), val = tensor([1, 1])]; - tensor var_3083_pad_0 = const()[name = tensor("op_3083_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3083_dilations_0 = const()[name = tensor("op_3083_dilations_0"), val = tensor([1, 1])]; - tensor var_3083_groups_0 = const()[name = tensor("op_3083_groups_0"), val = tensor(1)]; - tensor layers_7_conv_pointwise_conv1_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45377664))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45770944))), name = tensor("layers_7_conv_pointwise_conv1_inlier_module_weight_to_fp16_palettized"), shape = tensor([1024, 512, 1, 1])]; - tensor layers_7_conv_pointwise_conv1_inlier_module_bias_to_fp16 = const()[name = tensor("layers_7_conv_pointwise_conv1_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45771136)))]; - tensor var_3083_cast_fp16 = conv(bias = layers_7_conv_pointwise_conv1_inlier_module_bias_to_fp16, dilations = var_3083_dilations_0, groups = var_3083_groups_0, pad = var_3083_pad_0, pad_type = var_3083_pad_type_0, strides = var_3083_strides_0, weight = layers_7_conv_pointwise_conv1_inlier_module_weight_to_fp16_palettized, x = input_207_cast_fp16)[name = tensor("op_3083_cast_fp16")]; - tensor var_3089_pad_type_0 = const()[name = tensor("op_3089_pad_type_0"), val = tensor("valid")]; - tensor var_3089_strides_0 = const()[name = tensor("op_3089_strides_0"), val = tensor([1, 1])]; - tensor var_3089_pad_0 = const()[name = tensor("op_3089_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3089_dilations_0 = const()[name = tensor("op_3089_dilations_0"), val = tensor([1, 1])]; - tensor var_3089_groups_0 = const()[name = tensor("op_3089_groups_0"), val = tensor(1)]; - tensor layers_7_conv_pointwise_conv1_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45791168))), name = tensor("layers_7_conv_pointwise_conv1_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45773248))), shape = tensor([1024, 512, 1, 1])]; - tensor var_3089_cast_fp16 = conv(dilations = var_3089_dilations_0, groups = var_3089_groups_0, pad = var_3089_pad_0, pad_type = var_3089_pad_type_0, strides = var_3089_strides_0, weight = layers_7_conv_pointwise_conv1_outlier_module_weight_to_fp16_sparsified, x = input_207_cast_fp16)[name = tensor("op_3089_cast_fp16")]; - tensor input_209_cast_fp16 = add(x = var_3083_cast_fp16, y = var_3089_cast_fp16)[name = tensor("input_209_cast_fp16")]; - tensor input_211_split_num_splits_0 = const()[name = tensor("input_211_split_num_splits_0"), val = tensor(2)]; - tensor input_211_split_axis_0 = const()[name = tensor("input_211_split_axis_0"), val = tensor(1)]; - tensor input_211_split_cast_fp16_0, tensor input_211_split_cast_fp16_1 = split(axis = input_211_split_axis_0, num_splits = input_211_split_num_splits_0, x = input_209_cast_fp16)[name = tensor("input_211_split_cast_fp16")]; - tensor input_211_split_1_sigmoid_cast_fp16 = sigmoid(x = input_211_split_cast_fp16_1)[name = tensor("input_211_split_1_sigmoid_cast_fp16")]; - tensor input_211_cast_fp16 = mul(x = input_211_split_cast_fp16_0, y = input_211_split_1_sigmoid_cast_fp16)[name = tensor("input_211_cast_fp16")]; - tensor input_213_pad_type_0 = const()[name = tensor("input_213_pad_type_0"), val = tensor("custom")]; - tensor input_213_pad_0 = const()[name = tensor("input_213_pad_0"), val = tensor([0, 0, 4, 4])]; - tensor input_213_groups_0 = const()[name = tensor("input_213_groups_0"), val = tensor(512)]; - tensor input_213_strides_0 = const()[name = tensor("input_213_strides_0"), val = tensor([1, 1])]; - tensor input_213_dilations_0 = const()[name = tensor("input_213_dilations_0"), val = tensor([1, 1])]; - tensor const_205_to_fp16 = const()[name = tensor("const_205_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45856768)))]; - tensor const_206_to_fp16 = const()[name = tensor("const_206_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45866048)))]; - tensor input_215_cast_fp16 = conv(bias = const_206_to_fp16, dilations = input_213_dilations_0, groups = input_213_groups_0, pad = input_213_pad_0, pad_type = input_213_pad_type_0, strides = input_213_strides_0, weight = const_205_to_fp16, x = input_211_cast_fp16)[name = tensor("input_215_cast_fp16")]; - tensor input_217_cast_fp16 = silu(x = input_215_cast_fp16)[name = tensor("input_217_cast_fp16")]; - tensor var_3113_pad_type_0 = const()[name = tensor("op_3113_pad_type_0"), val = tensor("valid")]; - tensor var_3113_strides_0 = const()[name = tensor("op_3113_strides_0"), val = tensor([1, 1])]; - tensor var_3113_pad_0 = const()[name = tensor("op_3113_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3113_dilations_0 = const()[name = tensor("op_3113_dilations_0"), val = tensor([1, 1])]; - tensor var_3113_groups_0 = const()[name = tensor("op_3113_groups_0"), val = tensor(1)]; - tensor layers_7_conv_pointwise_conv2_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45867136))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46063808))), name = tensor("layers_7_conv_pointwise_conv2_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_7_conv_pointwise_conv2_inlier_module_bias_to_fp16 = const()[name = tensor("layers_7_conv_pointwise_conv2_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46064000)))]; - tensor var_3113_cast_fp16 = conv(bias = layers_7_conv_pointwise_conv2_inlier_module_bias_to_fp16, dilations = var_3113_dilations_0, groups = var_3113_groups_0, pad = var_3113_pad_0, pad_type = var_3113_pad_type_0, strides = var_3113_strides_0, weight = layers_7_conv_pointwise_conv2_inlier_module_weight_to_fp16_palettized, x = input_217_cast_fp16)[name = tensor("op_3113_cast_fp16")]; - tensor var_3119_pad_type_0 = const()[name = tensor("op_3119_pad_type_0"), val = tensor("valid")]; - tensor var_3119_strides_0 = const()[name = tensor("op_3119_strides_0"), val = tensor([1, 1])]; - tensor var_3119_pad_0 = const()[name = tensor("op_3119_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3119_dilations_0 = const()[name = tensor("op_3119_dilations_0"), val = tensor([1, 1])]; - tensor var_3119_groups_0 = const()[name = tensor("op_3119_groups_0"), val = tensor(1)]; - tensor layers_7_conv_pointwise_conv2_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46073536))), name = tensor("layers_7_conv_pointwise_conv2_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46065088))), shape = tensor([512, 512, 1, 1])]; - tensor var_3119_cast_fp16 = conv(dilations = var_3119_dilations_0, groups = var_3119_groups_0, pad = var_3119_pad_0, pad_type = var_3119_pad_type_0, strides = var_3119_strides_0, weight = layers_7_conv_pointwise_conv2_outlier_module_weight_to_fp16_sparsified, x = input_217_cast_fp16)[name = tensor("op_3119_cast_fp16")]; - tensor x_47_cast_fp16 = add(x = var_3113_cast_fp16, y = var_3119_cast_fp16)[name = tensor("x_47_cast_fp16")]; - tensor inputs_77_cast_fp16 = add(x = inputs_75_cast_fp16, y = x_47_cast_fp16)[name = tensor("inputs_77_cast_fp16")]; - tensor out_77_axes_0 = const()[name = tensor("out_77_axes_0"), val = tensor([1])]; - tensor var_3130_to_fp16 = const()[name = tensor("op_3130_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_77_cast_fp16 = layer_norm(axes = out_77_axes_0, epsilon = var_3130_to_fp16, x = inputs_77_cast_fp16)[name = tensor("out_77_cast_fp16")]; - tensor input_219_gamma_0_to_fp16 = const()[name = tensor("input_219_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46106368)))]; - tensor input_219_beta_0_to_fp16 = const()[name = tensor("input_219_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46107456)))]; - tensor input_219_epsilon_0_to_fp16 = const()[name = tensor("input_219_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_219_cast_fp16 = batch_norm(beta = input_219_beta_0_to_fp16, epsilon = input_219_epsilon_0_to_fp16, gamma = input_219_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_77_cast_fp16)[name = tensor("input_219_cast_fp16")]; - tensor var_3150_pad_type_0 = const()[name = tensor("op_3150_pad_type_0"), val = tensor("valid")]; - tensor var_3150_strides_0 = const()[name = tensor("op_3150_strides_0"), val = tensor([1, 1])]; - tensor var_3150_pad_0 = const()[name = tensor("op_3150_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3150_dilations_0 = const()[name = tensor("op_3150_dilations_0"), val = tensor([1, 1])]; - tensor var_3150_groups_0 = const()[name = tensor("op_3150_groups_0"), val = tensor(1)]; - tensor layers_7_feed_forward2_fc1_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46108544))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46895040))), name = tensor("layers_7_feed_forward2_fc1_inlier_module_weight_to_fp16_palettized"), shape = tensor([2048, 512, 1, 1])]; - tensor layers_7_feed_forward2_fc1_inlier_module_bias_to_fp16 = const()[name = tensor("layers_7_feed_forward2_fc1_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46895232)))]; - tensor var_3150_cast_fp16 = conv(bias = layers_7_feed_forward2_fc1_inlier_module_bias_to_fp16, dilations = var_3150_dilations_0, groups = var_3150_groups_0, pad = var_3150_pad_0, pad_type = var_3150_pad_type_0, strides = var_3150_strides_0, weight = layers_7_feed_forward2_fc1_inlier_module_weight_to_fp16_palettized, x = input_219_cast_fp16)[name = tensor("op_3150_cast_fp16")]; - tensor var_3156_pad_type_0 = const()[name = tensor("op_3156_pad_type_0"), val = tensor("valid")]; - tensor var_3156_strides_0 = const()[name = tensor("op_3156_strides_0"), val = tensor([1, 1])]; - tensor var_3156_pad_0 = const()[name = tensor("op_3156_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3156_dilations_0 = const()[name = tensor("op_3156_dilations_0"), val = tensor([1, 1])]; - tensor var_3156_groups_0 = const()[name = tensor("op_3156_groups_0"), val = tensor(1)]; - tensor layers_7_feed_forward2_fc1_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46930816))), name = tensor("layers_7_feed_forward2_fc1_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46899392))), shape = tensor([2048, 512, 1, 1])]; - tensor var_3156_cast_fp16 = conv(dilations = var_3156_dilations_0, groups = var_3156_groups_0, pad = var_3156_pad_0, pad_type = var_3156_pad_type_0, strides = var_3156_strides_0, weight = layers_7_feed_forward2_fc1_outlier_module_weight_to_fp16_sparsified, x = input_219_cast_fp16)[name = tensor("op_3156_cast_fp16")]; - tensor input_221_cast_fp16 = add(x = var_3150_cast_fp16, y = var_3156_cast_fp16)[name = tensor("input_221_cast_fp16")]; - tensor input_223_cast_fp16 = silu(x = input_221_cast_fp16)[name = tensor("input_223_cast_fp16")]; - tensor var_3167_pad_type_0 = const()[name = tensor("op_3167_pad_type_0"), val = tensor("valid")]; - tensor var_3167_strides_0 = const()[name = tensor("op_3167_strides_0"), val = tensor([1, 1])]; - tensor var_3167_pad_0 = const()[name = tensor("op_3167_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3167_dilations_0 = const()[name = tensor("op_3167_dilations_0"), val = tensor([1, 1])]; - tensor var_3167_groups_0 = const()[name = tensor("op_3167_groups_0"), val = tensor(1)]; - tensor layers_7_feed_forward2_fc2_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47061952))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47848448))), name = tensor("layers_7_feed_forward2_fc2_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 2048, 1, 1])]; - tensor layers_7_feed_forward2_fc2_inlier_module_bias_to_fp16 = const()[name = tensor("layers_7_feed_forward2_fc2_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47848640)))]; - tensor var_3167_cast_fp16 = conv(bias = layers_7_feed_forward2_fc2_inlier_module_bias_to_fp16, dilations = var_3167_dilations_0, groups = var_3167_groups_0, pad = var_3167_pad_0, pad_type = var_3167_pad_type_0, strides = var_3167_strides_0, weight = layers_7_feed_forward2_fc2_inlier_module_weight_to_fp16_palettized, x = input_223_cast_fp16)[name = tensor("op_3167_cast_fp16")]; - tensor var_3173_pad_type_0 = const()[name = tensor("op_3173_pad_type_0"), val = tensor("valid")]; - tensor var_3173_strides_0 = const()[name = tensor("op_3173_strides_0"), val = tensor([1, 1])]; - tensor var_3173_pad_0 = const()[name = tensor("op_3173_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3173_dilations_0 = const()[name = tensor("op_3173_dilations_0"), val = tensor([1, 1])]; - tensor var_3173_groups_0 = const()[name = tensor("op_3173_groups_0"), val = tensor(1)]; - tensor layers_7_feed_forward2_fc2_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47885632))), name = tensor("layers_7_feed_forward2_fc2_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47849728))), shape = tensor([512, 2048, 1, 1])]; - tensor var_3173_cast_fp16 = conv(dilations = var_3173_dilations_0, groups = var_3173_groups_0, pad = var_3173_pad_0, pad_type = var_3173_pad_type_0, strides = var_3173_strides_0, weight = layers_7_feed_forward2_fc2_outlier_module_weight_to_fp16_sparsified, x = input_223_cast_fp16)[name = tensor("op_3173_cast_fp16")]; - tensor x_49_cast_fp16 = add(x = var_3167_cast_fp16, y = var_3173_cast_fp16)[name = tensor("x_49_cast_fp16")]; - tensor var_3175_to_fp16 = const()[name = tensor("op_3175_to_fp16"), val = tensor(0x1p-1)]; - tensor var_3176_cast_fp16 = mul(x = x_49_cast_fp16, y = var_3175_to_fp16)[name = tensor("op_3176_cast_fp16")]; - tensor inputs_79_cast_fp16 = add(x = inputs_77_cast_fp16, y = var_3176_cast_fp16)[name = tensor("inputs_79_cast_fp16")]; - tensor out_79_axes_0 = const()[name = tensor("out_79_axes_0"), val = tensor([1])]; - tensor var_3186_to_fp16 = const()[name = tensor("op_3186_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_79_cast_fp16 = layer_norm(axes = out_79_axes_0, epsilon = var_3186_to_fp16, x = inputs_79_cast_fp16)[name = tensor("out_79_cast_fp16")]; - tensor inputs_81_gamma_0_to_fp16 = const()[name = tensor("inputs_81_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48016768)))]; - tensor inputs_81_beta_0_to_fp16 = const()[name = tensor("inputs_81_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48017856)))]; - tensor inputs_81_epsilon_0_to_fp16 = const()[name = tensor("inputs_81_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor inputs_81_cast_fp16 = batch_norm(beta = inputs_81_beta_0_to_fp16, epsilon = inputs_81_epsilon_0_to_fp16, gamma = inputs_81_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_79_cast_fp16)[name = tensor("inputs_81_cast_fp16")]; - tensor var_3200 = const()[name = tensor("op_3200"), val = tensor(3)]; - tensor out_81_axes_0 = const()[name = tensor("out_81_axes_0"), val = tensor([1])]; - tensor var_3231_to_fp16 = const()[name = tensor("op_3231_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_81_cast_fp16 = layer_norm(axes = out_81_axes_0, epsilon = var_3231_to_fp16, x = inputs_81_cast_fp16)[name = tensor("out_81_cast_fp16")]; - tensor input_225_gamma_0_to_fp16 = const()[name = tensor("input_225_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48018944)))]; - tensor input_225_beta_0_to_fp16 = const()[name = tensor("input_225_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48020032)))]; - tensor input_225_epsilon_0_to_fp16 = const()[name = tensor("input_225_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_225_cast_fp16 = batch_norm(beta = input_225_beta_0_to_fp16, epsilon = input_225_epsilon_0_to_fp16, gamma = input_225_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_81_cast_fp16)[name = tensor("input_225_cast_fp16")]; - tensor var_3251_pad_type_0 = const()[name = tensor("op_3251_pad_type_0"), val = tensor("valid")]; - tensor var_3251_strides_0 = const()[name = tensor("op_3251_strides_0"), val = tensor([1, 1])]; - tensor var_3251_pad_0 = const()[name = tensor("op_3251_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3251_dilations_0 = const()[name = tensor("op_3251_dilations_0"), val = tensor([1, 1])]; - tensor var_3251_groups_0 = const()[name = tensor("op_3251_groups_0"), val = tensor(1)]; - tensor layers_8_feed_forward1_fc1_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48021120))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48807616))), name = tensor("layers_8_feed_forward1_fc1_inlier_module_weight_to_fp16_palettized"), shape = tensor([2048, 512, 1, 1])]; - tensor layers_8_feed_forward1_fc1_inlier_module_bias_to_fp16 = const()[name = tensor("layers_8_feed_forward1_fc1_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48807808)))]; - tensor var_3251_cast_fp16 = conv(bias = layers_8_feed_forward1_fc1_inlier_module_bias_to_fp16, dilations = var_3251_dilations_0, groups = var_3251_groups_0, pad = var_3251_pad_0, pad_type = var_3251_pad_type_0, strides = var_3251_strides_0, weight = layers_8_feed_forward1_fc1_inlier_module_weight_to_fp16_palettized, x = input_225_cast_fp16)[name = tensor("op_3251_cast_fp16")]; - tensor var_3257_pad_type_0 = const()[name = tensor("op_3257_pad_type_0"), val = tensor("valid")]; - tensor var_3257_strides_0 = const()[name = tensor("op_3257_strides_0"), val = tensor([1, 1])]; - tensor var_3257_pad_0 = const()[name = tensor("op_3257_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3257_dilations_0 = const()[name = tensor("op_3257_dilations_0"), val = tensor([1, 1])]; - tensor var_3257_groups_0 = const()[name = tensor("op_3257_groups_0"), val = tensor(1)]; - tensor layers_8_feed_forward1_fc1_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48844544))), name = tensor("layers_8_feed_forward1_fc1_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48811968))), shape = tensor([2048, 512, 1, 1])]; - tensor var_3257_cast_fp16 = conv(dilations = var_3257_dilations_0, groups = var_3257_groups_0, pad = var_3257_pad_0, pad_type = var_3257_pad_type_0, strides = var_3257_strides_0, weight = layers_8_feed_forward1_fc1_outlier_module_weight_to_fp16_sparsified, x = input_225_cast_fp16)[name = tensor("op_3257_cast_fp16")]; - tensor input_227_cast_fp16 = add(x = var_3251_cast_fp16, y = var_3257_cast_fp16)[name = tensor("input_227_cast_fp16")]; - tensor input_229_cast_fp16 = silu(x = input_227_cast_fp16)[name = tensor("input_229_cast_fp16")]; - tensor var_3268_pad_type_0 = const()[name = tensor("op_3268_pad_type_0"), val = tensor("valid")]; - tensor var_3268_strides_0 = const()[name = tensor("op_3268_strides_0"), val = tensor([1, 1])]; - tensor var_3268_pad_0 = const()[name = tensor("op_3268_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3268_dilations_0 = const()[name = tensor("op_3268_dilations_0"), val = tensor([1, 1])]; - tensor var_3268_groups_0 = const()[name = tensor("op_3268_groups_0"), val = tensor(1)]; - tensor layers_8_feed_forward1_fc2_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48975680))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(49762176))), name = tensor("layers_8_feed_forward1_fc2_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 2048, 1, 1])]; - tensor layers_8_feed_forward1_fc2_inlier_module_bias_to_fp16 = const()[name = tensor("layers_8_feed_forward1_fc2_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(49762368)))]; - tensor var_3268_cast_fp16 = conv(bias = layers_8_feed_forward1_fc2_inlier_module_bias_to_fp16, dilations = var_3268_dilations_0, groups = var_3268_groups_0, pad = var_3268_pad_0, pad_type = var_3268_pad_type_0, strides = var_3268_strides_0, weight = layers_8_feed_forward1_fc2_inlier_module_weight_to_fp16_palettized, x = input_229_cast_fp16)[name = tensor("op_3268_cast_fp16")]; - tensor var_3274_pad_type_0 = const()[name = tensor("op_3274_pad_type_0"), val = tensor("valid")]; - tensor var_3274_strides_0 = const()[name = tensor("op_3274_strides_0"), val = tensor([1, 1])]; - tensor var_3274_pad_0 = const()[name = tensor("op_3274_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3274_dilations_0 = const()[name = tensor("op_3274_dilations_0"), val = tensor([1, 1])]; - tensor var_3274_groups_0 = const()[name = tensor("op_3274_groups_0"), val = tensor(1)]; - tensor layers_8_feed_forward1_fc2_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(49800128))), name = tensor("layers_8_feed_forward1_fc2_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(49763456))), shape = tensor([512, 2048, 1, 1])]; - tensor var_3274_cast_fp16 = conv(dilations = var_3274_dilations_0, groups = var_3274_groups_0, pad = var_3274_pad_0, pad_type = var_3274_pad_type_0, strides = var_3274_strides_0, weight = layers_8_feed_forward1_fc2_outlier_module_weight_to_fp16_sparsified, x = input_229_cast_fp16)[name = tensor("op_3274_cast_fp16")]; - tensor x_51_cast_fp16 = add(x = var_3268_cast_fp16, y = var_3274_cast_fp16)[name = tensor("x_51_cast_fp16")]; - tensor var_3276_to_fp16 = const()[name = tensor("op_3276_to_fp16"), val = tensor(0x1p-1)]; - tensor var_3277_cast_fp16 = mul(x = x_51_cast_fp16, y = var_3276_to_fp16)[name = tensor("op_3277_cast_fp16")]; - tensor inputs_83_cast_fp16 = add(x = inputs_81_cast_fp16, y = var_3277_cast_fp16)[name = tensor("inputs_83_cast_fp16")]; - tensor out_83_axes_0 = const()[name = tensor("out_83_axes_0"), val = tensor([1])]; - tensor var_3287_to_fp16 = const()[name = tensor("op_3287_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_83_cast_fp16 = layer_norm(axes = out_83_axes_0, epsilon = var_3287_to_fp16, x = inputs_83_cast_fp16)[name = tensor("out_83_cast_fp16")]; - tensor obj_35_gamma_0_to_fp16 = const()[name = tensor("obj_35_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(49931264)))]; - tensor obj_35_beta_0_to_fp16 = const()[name = tensor("obj_35_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(49932352)))]; - tensor obj_35_epsilon_0_to_fp16 = const()[name = tensor("obj_35_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor obj_35_cast_fp16 = batch_norm(beta = obj_35_beta_0_to_fp16, epsilon = obj_35_epsilon_0_to_fp16, gamma = obj_35_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_83_cast_fp16)[name = tensor("obj_35_cast_fp16")]; - tensor var_3312_pad_type_0 = const()[name = tensor("op_3312_pad_type_0"), val = tensor("valid")]; - tensor var_3312_strides_0 = const()[name = tensor("op_3312_strides_0"), val = tensor([1, 1])]; - tensor var_3312_pad_0 = const()[name = tensor("op_3312_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3312_dilations_0 = const()[name = tensor("op_3312_dilations_0"), val = tensor([1, 1])]; - tensor var_3312_groups_0 = const()[name = tensor("op_3312_groups_0"), val = tensor(1)]; - tensor layers_8_self_attn_q_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(49933440))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(50130112))), name = tensor("layers_8_self_attn_q_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_8_self_attn_q_proj_inlier_module_bias_to_fp16 = const()[name = tensor("layers_8_self_attn_q_proj_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(50130304)))]; - tensor var_3312_cast_fp16 = conv(bias = layers_8_self_attn_q_proj_inlier_module_bias_to_fp16, dilations = var_3312_dilations_0, groups = var_3312_groups_0, pad = var_3312_pad_0, pad_type = var_3312_pad_type_0, strides = var_3312_strides_0, weight = layers_8_self_attn_q_proj_inlier_module_weight_to_fp16_palettized, x = obj_35_cast_fp16)[name = tensor("op_3312_cast_fp16")]; - tensor var_3318_pad_type_0 = const()[name = tensor("op_3318_pad_type_0"), val = tensor("valid")]; - tensor var_3318_strides_0 = const()[name = tensor("op_3318_strides_0"), val = tensor([1, 1])]; - tensor var_3318_pad_0 = const()[name = tensor("op_3318_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3318_dilations_0 = const()[name = tensor("op_3318_dilations_0"), val = tensor([1, 1])]; - tensor var_3318_groups_0 = const()[name = tensor("op_3318_groups_0"), val = tensor(1)]; - tensor layers_8_self_attn_q_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(50142912))), name = tensor("layers_8_self_attn_q_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(50131392))), shape = tensor([512, 512, 1, 1])]; - tensor var_3318_cast_fp16 = conv(dilations = var_3318_dilations_0, groups = var_3318_groups_0, pad = var_3318_pad_0, pad_type = var_3318_pad_type_0, strides = var_3318_strides_0, weight = layers_8_self_attn_q_proj_outlier_module_weight_to_fp16_sparsified, x = obj_35_cast_fp16)[name = tensor("op_3318_cast_fp16")]; - tensor query_33_cast_fp16 = add(x = var_3312_cast_fp16, y = var_3318_cast_fp16)[name = tensor("query_33_cast_fp16")]; - tensor var_3327_pad_type_0 = const()[name = tensor("op_3327_pad_type_0"), val = tensor("valid")]; - tensor var_3327_strides_0 = const()[name = tensor("op_3327_strides_0"), val = tensor([1, 1])]; - tensor var_3327_pad_0 = const()[name = tensor("op_3327_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3327_dilations_0 = const()[name = tensor("op_3327_dilations_0"), val = tensor([1, 1])]; - tensor var_3327_groups_0 = const()[name = tensor("op_3327_groups_0"), val = tensor(1)]; - tensor layers_8_self_attn_k_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(50175744))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(50372416))), name = tensor("layers_8_self_attn_k_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor var_3327_cast_fp16 = conv(dilations = var_3327_dilations_0, groups = var_3327_groups_0, pad = var_3327_pad_0, pad_type = var_3327_pad_type_0, strides = var_3327_strides_0, weight = layers_8_self_attn_k_proj_inlier_module_weight_to_fp16_palettized, x = obj_35_cast_fp16)[name = tensor("op_3327_cast_fp16")]; - tensor var_3333_pad_type_0 = const()[name = tensor("op_3333_pad_type_0"), val = tensor("valid")]; - tensor var_3333_strides_0 = const()[name = tensor("op_3333_strides_0"), val = tensor([1, 1])]; - tensor var_3333_pad_0 = const()[name = tensor("op_3333_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3333_dilations_0 = const()[name = tensor("op_3333_dilations_0"), val = tensor([1, 1])]; - tensor var_3333_groups_0 = const()[name = tensor("op_3333_groups_0"), val = tensor(1)]; - tensor layers_8_self_attn_k_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(50381632))), name = tensor("layers_8_self_attn_k_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(50372608))), shape = tensor([512, 512, 1, 1])]; - tensor var_3333_cast_fp16 = conv(dilations = var_3333_dilations_0, groups = var_3333_groups_0, pad = var_3333_pad_0, pad_type = var_3333_pad_type_0, strides = var_3333_strides_0, weight = layers_8_self_attn_k_proj_outlier_module_weight_to_fp16_sparsified, x = obj_35_cast_fp16)[name = tensor("op_3333_cast_fp16")]; - tensor key_17_cast_fp16 = add(x = var_3327_cast_fp16, y = var_3333_cast_fp16)[name = tensor("key_17_cast_fp16")]; - tensor var_3343_pad_type_0 = const()[name = tensor("op_3343_pad_type_0"), val = tensor("valid")]; - tensor var_3343_strides_0 = const()[name = tensor("op_3343_strides_0"), val = tensor([1, 1])]; - tensor var_3343_pad_0 = const()[name = tensor("op_3343_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3343_dilations_0 = const()[name = tensor("op_3343_dilations_0"), val = tensor([1, 1])]; - tensor var_3343_groups_0 = const()[name = tensor("op_3343_groups_0"), val = tensor(1)]; - tensor layers_8_self_attn_v_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(50414464))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(50611136))), name = tensor("layers_8_self_attn_v_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_8_self_attn_v_proj_inlier_module_bias_to_fp16 = const()[name = tensor("layers_8_self_attn_v_proj_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(50611328)))]; - tensor var_3343_cast_fp16 = conv(bias = layers_8_self_attn_v_proj_inlier_module_bias_to_fp16, dilations = var_3343_dilations_0, groups = var_3343_groups_0, pad = var_3343_pad_0, pad_type = var_3343_pad_type_0, strides = var_3343_strides_0, weight = layers_8_self_attn_v_proj_inlier_module_weight_to_fp16_palettized, x = obj_35_cast_fp16)[name = tensor("op_3343_cast_fp16")]; - tensor var_3349_pad_type_0 = const()[name = tensor("op_3349_pad_type_0"), val = tensor("valid")]; - tensor var_3349_strides_0 = const()[name = tensor("op_3349_strides_0"), val = tensor([1, 1])]; - tensor var_3349_pad_0 = const()[name = tensor("op_3349_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3349_dilations_0 = const()[name = tensor("op_3349_dilations_0"), val = tensor([1, 1])]; - tensor var_3349_groups_0 = const()[name = tensor("op_3349_groups_0"), val = tensor(1)]; - tensor layers_8_self_attn_v_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(50621184))), name = tensor("layers_8_self_attn_v_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(50612416))), shape = tensor([512, 512, 1, 1])]; - tensor var_3349_cast_fp16 = conv(dilations = var_3349_dilations_0, groups = var_3349_groups_0, pad = var_3349_pad_0, pad_type = var_3349_pad_type_0, strides = var_3349_strides_0, weight = layers_8_self_attn_v_proj_outlier_module_weight_to_fp16_sparsified, x = obj_35_cast_fp16)[name = tensor("op_3349_cast_fp16")]; - tensor value_17_cast_fp16 = add(x = var_3343_cast_fp16, y = var_3349_cast_fp16)[name = tensor("value_17_cast_fp16")]; - tensor var_3352_to_fp16 = const()[name = tensor("op_3352_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(50654016)))]; - tensor query_35_cast_fp16 = add(x = query_33_cast_fp16, y = var_3352_to_fp16)[name = tensor("query_35_cast_fp16")]; - tensor var_3355_to_fp16 = const()[name = tensor("op_3355_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(50655104)))]; - tensor q_with_bias_v_17_cast_fp16 = add(x = query_33_cast_fp16, y = var_3355_to_fp16)[name = tensor("q_with_bias_v_17_cast_fp16")]; - tensor var_3365_pad_type_0 = const()[name = tensor("op_3365_pad_type_0"), val = tensor("valid")]; - tensor var_3365_strides_0 = const()[name = tensor("op_3365_strides_0"), val = tensor([1, 1])]; - tensor var_3365_pad_0 = const()[name = tensor("op_3365_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3365_dilations_0 = const()[name = tensor("op_3365_dilations_0"), val = tensor([1, 1])]; - tensor var_3365_groups_0 = const()[name = tensor("op_3365_groups_0"), val = tensor(1)]; - tensor layers_8_self_attn_linear_pos_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(50656192))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(50852864))), name = tensor("layers_8_self_attn_linear_pos_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor var_3365_cast_fp16 = conv(bias = input_17_mean_0_to_fp16, dilations = var_3365_dilations_0, groups = var_3365_groups_0, pad = var_3365_pad_0, pad_type = var_3365_pad_type_0, strides = var_3365_strides_0, weight = layers_8_self_attn_linear_pos_inlier_module_weight_to_fp16_palettized, x = obj_3_cast_fp16)[name = tensor("op_3365_cast_fp16")]; - tensor var_3371_pad_type_0 = const()[name = tensor("op_3371_pad_type_0"), val = tensor("valid")]; - tensor var_3371_strides_0 = const()[name = tensor("op_3371_strides_0"), val = tensor([1, 1])]; - tensor var_3371_pad_0 = const()[name = tensor("op_3371_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3371_dilations_0 = const()[name = tensor("op_3371_dilations_0"), val = tensor([1, 1])]; - tensor var_3371_groups_0 = const()[name = tensor("op_3371_groups_0"), val = tensor(1)]; - tensor layers_8_self_attn_linear_pos_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(50867840))), name = tensor("layers_8_self_attn_linear_pos_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(50853056))), shape = tensor([512, 512, 1, 1])]; - tensor var_3371_cast_fp16 = conv(dilations = var_3371_dilations_0, groups = var_3371_groups_0, pad = var_3371_pad_0, pad_type = var_3371_pad_type_0, strides = var_3371_strides_0, weight = layers_8_self_attn_linear_pos_outlier_module_weight_to_fp16_sparsified, x = obj_3_cast_fp16)[name = tensor("op_3371_cast_fp16")]; - tensor p_17_cast_fp16 = add(x = var_3365_cast_fp16, y = var_3371_cast_fp16)[name = tensor("p_17_cast_fp16")]; - tensor var_3375 = const()[name = tensor("op_3375"), val = tensor([1, 8, 64, 188])]; - tensor var_3376_cast_fp16 = reshape(shape = var_3375, x = q_with_bias_v_17_cast_fp16)[name = tensor("op_3376_cast_fp16")]; - tensor var_3377 = const()[name = tensor("op_3377"), val = tensor([1, 8, 64, -1])]; - tensor var_3378_cast_fp16 = reshape(shape = var_3377, x = p_17_cast_fp16)[name = tensor("op_3378_cast_fp16")]; - tensor matrix_bd_65_transpose_x_0 = const()[name = tensor("matrix_bd_65_transpose_x_0"), val = tensor(true)]; - tensor matrix_bd_65_transpose_y_0 = const()[name = tensor("matrix_bd_65_transpose_y_0"), val = tensor(false)]; - tensor matrix_bd_65_cast_fp16 = matmul(transpose_x = matrix_bd_65_transpose_x_0, transpose_y = matrix_bd_65_transpose_y_0, x = var_3376_cast_fp16, y = var_3378_cast_fp16)[name = tensor("matrix_bd_65_cast_fp16")]; - tensor matrix_bd_67_pad_0 = const()[name = tensor("matrix_bd_67_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor matrix_bd_67_mode_0 = const()[name = tensor("matrix_bd_67_mode_0"), val = tensor("constant")]; - tensor const_98_to_fp16 = const()[name = tensor("const_98_to_fp16"), val = tensor(0x0p+0)]; - tensor matrix_bd_67_cast_fp16 = pad(constant_val = const_98_to_fp16, mode = matrix_bd_67_mode_0, pad = matrix_bd_67_pad_0, x = matrix_bd_65_cast_fp16)[name = tensor("matrix_bd_67_cast_fp16")]; - tensor var_3387 = const()[name = tensor("op_3387"), val = tensor([1, 8, -1, 188])]; - tensor matrix_bd_69_cast_fp16 = reshape(shape = var_3387, x = matrix_bd_67_cast_fp16)[name = tensor("matrix_bd_69_cast_fp16")]; - tensor var_3391_begin_0 = const()[name = tensor("op_3391_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_3391_end_0 = const()[name = tensor("op_3391_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_3391_end_mask_0 = const()[name = tensor("op_3391_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_3391_cast_fp16 = slice_by_index(begin = var_3391_begin_0, end = var_3391_end_0, end_mask = var_3391_end_mask_0, x = matrix_bd_69_cast_fp16)[name = tensor("op_3391_cast_fp16")]; - tensor var_3392 = const()[name = tensor("op_3392"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_71_cast_fp16 = reshape(shape = var_3392, x = var_3391_cast_fp16)[name = tensor("matrix_bd_71_cast_fp16")]; - tensor var_3397_begin_0 = const()[name = tensor("op_3397_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3397_end_0 = const()[name = tensor("op_3397_end_0"), val = tensor([1, 8, 188, 188])]; - tensor var_3397_end_mask_0 = const()[name = tensor("op_3397_end_mask_0"), val = tensor([true, true, true, false])]; - tensor var_3397_cast_fp16 = slice_by_index(begin = var_3397_begin_0, end = var_3397_end_0, end_mask = var_3397_end_mask_0, x = matrix_bd_71_cast_fp16)[name = tensor("op_3397_cast_fp16")]; - tensor var_3398_to_fp16 = const()[name = tensor("op_3398_to_fp16"), val = tensor(0x1p-3)]; - tensor qk_mask_17_cast_fp16 = mul(x = var_3397_cast_fp16, y = var_3398_to_fp16)[name = tensor("qk_mask_17_cast_fp16")]; - tensor var_3402 = const()[name = tensor("op_3402"), val = tensor([1, 8, 64, 188])]; - tensor mh_q_17_cast_fp16 = reshape(shape = var_3402, x = query_35_cast_fp16)[name = tensor("mh_q_17_cast_fp16")]; - tensor var_3404_to_fp16 = const()[name = tensor("op_3404_to_fp16"), val = tensor(0x1p-3)]; - tensor var_3405_cast_fp16 = mul(x = mh_q_17_cast_fp16, y = var_3404_to_fp16)[name = tensor("op_3405_cast_fp16")]; - tensor var_3408 = const()[name = tensor("op_3408"), val = tensor([1, 8, 64, 188])]; - tensor var_3409_cast_fp16 = reshape(shape = var_3408, x = key_17_cast_fp16)[name = tensor("op_3409_cast_fp16")]; - tensor mh_w_33_transpose_x_0 = const()[name = tensor("mh_w_33_transpose_x_0"), val = tensor(true)]; - tensor mh_w_33_transpose_y_0 = const()[name = tensor("mh_w_33_transpose_y_0"), val = tensor(false)]; - tensor mh_w_33_cast_fp16 = matmul(transpose_x = mh_w_33_transpose_x_0, transpose_y = mh_w_33_transpose_y_0, x = var_3405_cast_fp16, y = var_3409_cast_fp16)[name = tensor("mh_w_33_cast_fp16")]; - tensor mh_w_35_cast_fp16 = add(x = mh_w_33_cast_fp16, y = qk_mask_17_cast_fp16)[name = tensor("mh_w_35_cast_fp16")]; - tensor var_3413_cast_fp16 = softmax(axis = var_3200, x = mh_w_35_cast_fp16)[name = tensor("op_3413_cast_fp16")]; - tensor var_3414 = const()[name = tensor("op_3414"), val = tensor([1, 8, 64, 188])]; - tensor var_3415_cast_fp16 = reshape(shape = var_3414, x = value_17_cast_fp16)[name = tensor("op_3415_cast_fp16")]; - tensor attn_17_transpose_x_0 = const()[name = tensor("attn_17_transpose_x_0"), val = tensor(false)]; - tensor attn_17_transpose_y_0 = const()[name = tensor("attn_17_transpose_y_0"), val = tensor(true)]; - tensor attn_17_cast_fp16 = matmul(transpose_x = attn_17_transpose_x_0, transpose_y = attn_17_transpose_y_0, x = var_3415_cast_fp16, y = var_3413_cast_fp16)[name = tensor("attn_17_cast_fp16")]; - tensor var_3418 = const()[name = tensor("op_3418"), val = tensor([1, 512, 1, 188])]; - tensor input_231_cast_fp16 = reshape(shape = var_3418, x = attn_17_cast_fp16)[name = tensor("input_231_cast_fp16")]; - tensor var_3428_pad_type_0 = const()[name = tensor("op_3428_pad_type_0"), val = tensor("valid")]; - tensor var_3428_strides_0 = const()[name = tensor("op_3428_strides_0"), val = tensor([1, 1])]; - tensor var_3428_pad_0 = const()[name = tensor("op_3428_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3428_dilations_0 = const()[name = tensor("op_3428_dilations_0"), val = tensor([1, 1])]; - tensor var_3428_groups_0 = const()[name = tensor("op_3428_groups_0"), val = tensor(1)]; - tensor layers_8_self_attn_o_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(50900672))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(51097344))), name = tensor("layers_8_self_attn_o_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_8_self_attn_o_proj_inlier_module_bias_to_fp16 = const()[name = tensor("layers_8_self_attn_o_proj_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(51097536)))]; - tensor var_3428_cast_fp16 = conv(bias = layers_8_self_attn_o_proj_inlier_module_bias_to_fp16, dilations = var_3428_dilations_0, groups = var_3428_groups_0, pad = var_3428_pad_0, pad_type = var_3428_pad_type_0, strides = var_3428_strides_0, weight = layers_8_self_attn_o_proj_inlier_module_weight_to_fp16_palettized, x = input_231_cast_fp16)[name = tensor("op_3428_cast_fp16")]; - tensor var_3434_pad_type_0 = const()[name = tensor("op_3434_pad_type_0"), val = tensor("valid")]; - tensor var_3434_strides_0 = const()[name = tensor("op_3434_strides_0"), val = tensor([1, 1])]; - tensor var_3434_pad_0 = const()[name = tensor("op_3434_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3434_dilations_0 = const()[name = tensor("op_3434_dilations_0"), val = tensor([1, 1])]; - tensor var_3434_groups_0 = const()[name = tensor("op_3434_groups_0"), val = tensor(1)]; - tensor layers_8_self_attn_o_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(51107392))), name = tensor("layers_8_self_attn_o_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(51098624))), shape = tensor([512, 512, 1, 1])]; - tensor var_3434_cast_fp16 = conv(dilations = var_3434_dilations_0, groups = var_3434_groups_0, pad = var_3434_pad_0, pad_type = var_3434_pad_type_0, strides = var_3434_strides_0, weight = layers_8_self_attn_o_proj_outlier_module_weight_to_fp16_sparsified, x = input_231_cast_fp16)[name = tensor("op_3434_cast_fp16")]; - tensor obj_37_cast_fp16 = add(x = var_3428_cast_fp16, y = var_3434_cast_fp16)[name = tensor("obj_37_cast_fp16")]; - tensor inputs_85_cast_fp16 = add(x = inputs_83_cast_fp16, y = obj_37_cast_fp16)[name = tensor("inputs_85_cast_fp16")]; - tensor out_85_axes_0 = const()[name = tensor("out_85_axes_0"), val = tensor([1])]; - tensor var_3445_to_fp16 = const()[name = tensor("op_3445_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_85_cast_fp16 = layer_norm(axes = out_85_axes_0, epsilon = var_3445_to_fp16, x = inputs_85_cast_fp16)[name = tensor("out_85_cast_fp16")]; - tensor input_233_gamma_0_to_fp16 = const()[name = tensor("input_233_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(51140224)))]; - tensor input_233_beta_0_to_fp16 = const()[name = tensor("input_233_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(51141312)))]; - tensor input_233_epsilon_0_to_fp16 = const()[name = tensor("input_233_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_233_cast_fp16 = batch_norm(beta = input_233_beta_0_to_fp16, epsilon = input_233_epsilon_0_to_fp16, gamma = input_233_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_85_cast_fp16)[name = tensor("input_233_cast_fp16")]; - tensor var_3467_pad_type_0 = const()[name = tensor("op_3467_pad_type_0"), val = tensor("valid")]; - tensor var_3467_strides_0 = const()[name = tensor("op_3467_strides_0"), val = tensor([1, 1])]; - tensor var_3467_pad_0 = const()[name = tensor("op_3467_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3467_dilations_0 = const()[name = tensor("op_3467_dilations_0"), val = tensor([1, 1])]; - tensor var_3467_groups_0 = const()[name = tensor("op_3467_groups_0"), val = tensor(1)]; - tensor layers_8_conv_pointwise_conv1_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(51142400))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(51535680))), name = tensor("layers_8_conv_pointwise_conv1_inlier_module_weight_to_fp16_palettized"), shape = tensor([1024, 512, 1, 1])]; - tensor layers_8_conv_pointwise_conv1_inlier_module_bias_to_fp16 = const()[name = tensor("layers_8_conv_pointwise_conv1_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(51535872)))]; - tensor var_3467_cast_fp16 = conv(bias = layers_8_conv_pointwise_conv1_inlier_module_bias_to_fp16, dilations = var_3467_dilations_0, groups = var_3467_groups_0, pad = var_3467_pad_0, pad_type = var_3467_pad_type_0, strides = var_3467_strides_0, weight = layers_8_conv_pointwise_conv1_inlier_module_weight_to_fp16_palettized, x = input_233_cast_fp16)[name = tensor("op_3467_cast_fp16")]; - tensor var_3473_pad_type_0 = const()[name = tensor("op_3473_pad_type_0"), val = tensor("valid")]; - tensor var_3473_strides_0 = const()[name = tensor("op_3473_strides_0"), val = tensor([1, 1])]; - tensor var_3473_pad_0 = const()[name = tensor("op_3473_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3473_dilations_0 = const()[name = tensor("op_3473_dilations_0"), val = tensor([1, 1])]; - tensor var_3473_groups_0 = const()[name = tensor("op_3473_groups_0"), val = tensor(1)]; - tensor layers_8_conv_pointwise_conv1_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(51555520))), name = tensor("layers_8_conv_pointwise_conv1_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(51537984))), shape = tensor([1024, 512, 1, 1])]; - tensor var_3473_cast_fp16 = conv(dilations = var_3473_dilations_0, groups = var_3473_groups_0, pad = var_3473_pad_0, pad_type = var_3473_pad_type_0, strides = var_3473_strides_0, weight = layers_8_conv_pointwise_conv1_outlier_module_weight_to_fp16_sparsified, x = input_233_cast_fp16)[name = tensor("op_3473_cast_fp16")]; - tensor input_235_cast_fp16 = add(x = var_3467_cast_fp16, y = var_3473_cast_fp16)[name = tensor("input_235_cast_fp16")]; - tensor input_237_split_num_splits_0 = const()[name = tensor("input_237_split_num_splits_0"), val = tensor(2)]; - tensor input_237_split_axis_0 = const()[name = tensor("input_237_split_axis_0"), val = tensor(1)]; - tensor input_237_split_cast_fp16_0, tensor input_237_split_cast_fp16_1 = split(axis = input_237_split_axis_0, num_splits = input_237_split_num_splits_0, x = input_235_cast_fp16)[name = tensor("input_237_split_cast_fp16")]; - tensor input_237_split_1_sigmoid_cast_fp16 = sigmoid(x = input_237_split_cast_fp16_1)[name = tensor("input_237_split_1_sigmoid_cast_fp16")]; - tensor input_237_cast_fp16 = mul(x = input_237_split_cast_fp16_0, y = input_237_split_1_sigmoid_cast_fp16)[name = tensor("input_237_cast_fp16")]; - tensor input_239_pad_type_0 = const()[name = tensor("input_239_pad_type_0"), val = tensor("custom")]; - tensor input_239_pad_0 = const()[name = tensor("input_239_pad_0"), val = tensor([0, 0, 4, 4])]; - tensor input_239_groups_0 = const()[name = tensor("input_239_groups_0"), val = tensor(512)]; - tensor input_239_strides_0 = const()[name = tensor("input_239_strides_0"), val = tensor([1, 1])]; - tensor input_239_dilations_0 = const()[name = tensor("input_239_dilations_0"), val = tensor([1, 1])]; - tensor const_207_to_fp16 = const()[name = tensor("const_207_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(51621120)))]; - tensor const_208_to_fp16 = const()[name = tensor("const_208_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(51630400)))]; - tensor input_241_cast_fp16 = conv(bias = const_208_to_fp16, dilations = input_239_dilations_0, groups = input_239_groups_0, pad = input_239_pad_0, pad_type = input_239_pad_type_0, strides = input_239_strides_0, weight = const_207_to_fp16, x = input_237_cast_fp16)[name = tensor("input_241_cast_fp16")]; - tensor input_243_cast_fp16 = silu(x = input_241_cast_fp16)[name = tensor("input_243_cast_fp16")]; - tensor var_3497_pad_type_0 = const()[name = tensor("op_3497_pad_type_0"), val = tensor("valid")]; - tensor var_3497_strides_0 = const()[name = tensor("op_3497_strides_0"), val = tensor([1, 1])]; - tensor var_3497_pad_0 = const()[name = tensor("op_3497_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3497_dilations_0 = const()[name = tensor("op_3497_dilations_0"), val = tensor([1, 1])]; - tensor var_3497_groups_0 = const()[name = tensor("op_3497_groups_0"), val = tensor(1)]; - tensor layers_8_conv_pointwise_conv2_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(51631488))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(51828160))), name = tensor("layers_8_conv_pointwise_conv2_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_8_conv_pointwise_conv2_inlier_module_bias_to_fp16 = const()[name = tensor("layers_8_conv_pointwise_conv2_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(51828352)))]; - tensor var_3497_cast_fp16 = conv(bias = layers_8_conv_pointwise_conv2_inlier_module_bias_to_fp16, dilations = var_3497_dilations_0, groups = var_3497_groups_0, pad = var_3497_pad_0, pad_type = var_3497_pad_type_0, strides = var_3497_strides_0, weight = layers_8_conv_pointwise_conv2_inlier_module_weight_to_fp16_palettized, x = input_243_cast_fp16)[name = tensor("op_3497_cast_fp16")]; - tensor var_3503_pad_type_0 = const()[name = tensor("op_3503_pad_type_0"), val = tensor("valid")]; - tensor var_3503_strides_0 = const()[name = tensor("op_3503_strides_0"), val = tensor([1, 1])]; - tensor var_3503_pad_0 = const()[name = tensor("op_3503_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3503_dilations_0 = const()[name = tensor("op_3503_dilations_0"), val = tensor([1, 1])]; - tensor var_3503_groups_0 = const()[name = tensor("op_3503_groups_0"), val = tensor(1)]; - tensor layers_8_conv_pointwise_conv2_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(51837312))), name = tensor("layers_8_conv_pointwise_conv2_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(51829440))), shape = tensor([512, 512, 1, 1])]; - tensor var_3503_cast_fp16 = conv(dilations = var_3503_dilations_0, groups = var_3503_groups_0, pad = var_3503_pad_0, pad_type = var_3503_pad_type_0, strides = var_3503_strides_0, weight = layers_8_conv_pointwise_conv2_outlier_module_weight_to_fp16_sparsified, x = input_243_cast_fp16)[name = tensor("op_3503_cast_fp16")]; - tensor x_53_cast_fp16 = add(x = var_3497_cast_fp16, y = var_3503_cast_fp16)[name = tensor("x_53_cast_fp16")]; - tensor inputs_87_cast_fp16 = add(x = inputs_85_cast_fp16, y = x_53_cast_fp16)[name = tensor("inputs_87_cast_fp16")]; - tensor out_87_axes_0 = const()[name = tensor("out_87_axes_0"), val = tensor([1])]; - tensor var_3514_to_fp16 = const()[name = tensor("op_3514_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_87_cast_fp16 = layer_norm(axes = out_87_axes_0, epsilon = var_3514_to_fp16, x = inputs_87_cast_fp16)[name = tensor("out_87_cast_fp16")]; - tensor input_245_gamma_0_to_fp16 = const()[name = tensor("input_245_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(51870144)))]; - tensor input_245_beta_0_to_fp16 = const()[name = tensor("input_245_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(51871232)))]; - tensor input_245_epsilon_0_to_fp16 = const()[name = tensor("input_245_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_245_cast_fp16 = batch_norm(beta = input_245_beta_0_to_fp16, epsilon = input_245_epsilon_0_to_fp16, gamma = input_245_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_87_cast_fp16)[name = tensor("input_245_cast_fp16")]; - tensor var_3534_pad_type_0 = const()[name = tensor("op_3534_pad_type_0"), val = tensor("valid")]; - tensor var_3534_strides_0 = const()[name = tensor("op_3534_strides_0"), val = tensor([1, 1])]; - tensor var_3534_pad_0 = const()[name = tensor("op_3534_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3534_dilations_0 = const()[name = tensor("op_3534_dilations_0"), val = tensor([1, 1])]; - tensor var_3534_groups_0 = const()[name = tensor("op_3534_groups_0"), val = tensor(1)]; - tensor layers_8_feed_forward2_fc1_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(51872320))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52658816))), name = tensor("layers_8_feed_forward2_fc1_inlier_module_weight_to_fp16_palettized"), shape = tensor([2048, 512, 1, 1])]; - tensor layers_8_feed_forward2_fc1_inlier_module_bias_to_fp16 = const()[name = tensor("layers_8_feed_forward2_fc1_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52659008)))]; - tensor var_3534_cast_fp16 = conv(bias = layers_8_feed_forward2_fc1_inlier_module_bias_to_fp16, dilations = var_3534_dilations_0, groups = var_3534_groups_0, pad = var_3534_pad_0, pad_type = var_3534_pad_type_0, strides = var_3534_strides_0, weight = layers_8_feed_forward2_fc1_inlier_module_weight_to_fp16_palettized, x = input_245_cast_fp16)[name = tensor("op_3534_cast_fp16")]; - tensor var_3540_pad_type_0 = const()[name = tensor("op_3540_pad_type_0"), val = tensor("valid")]; - tensor var_3540_strides_0 = const()[name = tensor("op_3540_strides_0"), val = tensor([1, 1])]; - tensor var_3540_pad_0 = const()[name = tensor("op_3540_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3540_dilations_0 = const()[name = tensor("op_3540_dilations_0"), val = tensor([1, 1])]; - tensor var_3540_groups_0 = const()[name = tensor("op_3540_groups_0"), val = tensor(1)]; - tensor layers_8_feed_forward2_fc1_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52694080))), name = tensor("layers_8_feed_forward2_fc1_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52663168))), shape = tensor([2048, 512, 1, 1])]; - tensor var_3540_cast_fp16 = conv(dilations = var_3540_dilations_0, groups = var_3540_groups_0, pad = var_3540_pad_0, pad_type = var_3540_pad_type_0, strides = var_3540_strides_0, weight = layers_8_feed_forward2_fc1_outlier_module_weight_to_fp16_sparsified, x = input_245_cast_fp16)[name = tensor("op_3540_cast_fp16")]; - tensor input_247_cast_fp16 = add(x = var_3534_cast_fp16, y = var_3540_cast_fp16)[name = tensor("input_247_cast_fp16")]; - tensor input_249_cast_fp16 = silu(x = input_247_cast_fp16)[name = tensor("input_249_cast_fp16")]; - tensor var_3551_pad_type_0 = const()[name = tensor("op_3551_pad_type_0"), val = tensor("valid")]; - tensor var_3551_strides_0 = const()[name = tensor("op_3551_strides_0"), val = tensor([1, 1])]; - tensor var_3551_pad_0 = const()[name = tensor("op_3551_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3551_dilations_0 = const()[name = tensor("op_3551_dilations_0"), val = tensor([1, 1])]; - tensor var_3551_groups_0 = const()[name = tensor("op_3551_groups_0"), val = tensor(1)]; - tensor layers_8_feed_forward2_fc2_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52825216))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(53611712))), name = tensor("layers_8_feed_forward2_fc2_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 2048, 1, 1])]; - tensor layers_8_feed_forward2_fc2_inlier_module_bias_to_fp16 = const()[name = tensor("layers_8_feed_forward2_fc2_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(53611904)))]; - tensor var_3551_cast_fp16 = conv(bias = layers_8_feed_forward2_fc2_inlier_module_bias_to_fp16, dilations = var_3551_dilations_0, groups = var_3551_groups_0, pad = var_3551_pad_0, pad_type = var_3551_pad_type_0, strides = var_3551_strides_0, weight = layers_8_feed_forward2_fc2_inlier_module_weight_to_fp16_palettized, x = input_249_cast_fp16)[name = tensor("op_3551_cast_fp16")]; - tensor var_3557_pad_type_0 = const()[name = tensor("op_3557_pad_type_0"), val = tensor("valid")]; - tensor var_3557_strides_0 = const()[name = tensor("op_3557_strides_0"), val = tensor([1, 1])]; - tensor var_3557_pad_0 = const()[name = tensor("op_3557_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3557_dilations_0 = const()[name = tensor("op_3557_dilations_0"), val = tensor([1, 1])]; - tensor var_3557_groups_0 = const()[name = tensor("op_3557_groups_0"), val = tensor(1)]; - tensor layers_8_feed_forward2_fc2_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(53646336))), name = tensor("layers_8_feed_forward2_fc2_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(53612992))), shape = tensor([512, 2048, 1, 1])]; - tensor var_3557_cast_fp16 = conv(dilations = var_3557_dilations_0, groups = var_3557_groups_0, pad = var_3557_pad_0, pad_type = var_3557_pad_type_0, strides = var_3557_strides_0, weight = layers_8_feed_forward2_fc2_outlier_module_weight_to_fp16_sparsified, x = input_249_cast_fp16)[name = tensor("op_3557_cast_fp16")]; - tensor x_55_cast_fp16 = add(x = var_3551_cast_fp16, y = var_3557_cast_fp16)[name = tensor("x_55_cast_fp16")]; - tensor var_3559_to_fp16 = const()[name = tensor("op_3559_to_fp16"), val = tensor(0x1p-1)]; - tensor var_3560_cast_fp16 = mul(x = x_55_cast_fp16, y = var_3559_to_fp16)[name = tensor("op_3560_cast_fp16")]; - tensor inputs_89_cast_fp16 = add(x = inputs_87_cast_fp16, y = var_3560_cast_fp16)[name = tensor("inputs_89_cast_fp16")]; - tensor out_89_axes_0 = const()[name = tensor("out_89_axes_0"), val = tensor([1])]; - tensor var_3570_to_fp16 = const()[name = tensor("op_3570_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_89_cast_fp16 = layer_norm(axes = out_89_axes_0, epsilon = var_3570_to_fp16, x = inputs_89_cast_fp16)[name = tensor("out_89_cast_fp16")]; - tensor inputs_91_gamma_0_to_fp16 = const()[name = tensor("inputs_91_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(53777472)))]; - tensor inputs_91_beta_0_to_fp16 = const()[name = tensor("inputs_91_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(53778560)))]; - tensor inputs_91_epsilon_0_to_fp16 = const()[name = tensor("inputs_91_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor inputs_91_cast_fp16 = batch_norm(beta = inputs_91_beta_0_to_fp16, epsilon = inputs_91_epsilon_0_to_fp16, gamma = inputs_91_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_89_cast_fp16)[name = tensor("inputs_91_cast_fp16")]; - tensor var_3584 = const()[name = tensor("op_3584"), val = tensor(3)]; - tensor out_91_axes_0 = const()[name = tensor("out_91_axes_0"), val = tensor([1])]; - tensor var_3615_to_fp16 = const()[name = tensor("op_3615_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_91_cast_fp16 = layer_norm(axes = out_91_axes_0, epsilon = var_3615_to_fp16, x = inputs_91_cast_fp16)[name = tensor("out_91_cast_fp16")]; - tensor input_251_gamma_0_to_fp16 = const()[name = tensor("input_251_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(53779648)))]; - tensor input_251_beta_0_to_fp16 = const()[name = tensor("input_251_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(53780736)))]; - tensor input_251_epsilon_0_to_fp16 = const()[name = tensor("input_251_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_251_cast_fp16 = batch_norm(beta = input_251_beta_0_to_fp16, epsilon = input_251_epsilon_0_to_fp16, gamma = input_251_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_91_cast_fp16)[name = tensor("input_251_cast_fp16")]; - tensor var_3635_pad_type_0 = const()[name = tensor("op_3635_pad_type_0"), val = tensor("valid")]; - tensor var_3635_strides_0 = const()[name = tensor("op_3635_strides_0"), val = tensor([1, 1])]; - tensor var_3635_pad_0 = const()[name = tensor("op_3635_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3635_dilations_0 = const()[name = tensor("op_3635_dilations_0"), val = tensor([1, 1])]; - tensor var_3635_groups_0 = const()[name = tensor("op_3635_groups_0"), val = tensor(1)]; - tensor layers_9_feed_forward1_fc1_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(53781824))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(54568320))), name = tensor("layers_9_feed_forward1_fc1_inlier_module_weight_to_fp16_palettized"), shape = tensor([2048, 512, 1, 1])]; - tensor layers_9_feed_forward1_fc1_inlier_module_bias_to_fp16 = const()[name = tensor("layers_9_feed_forward1_fc1_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(54568512)))]; - tensor var_3635_cast_fp16 = conv(bias = layers_9_feed_forward1_fc1_inlier_module_bias_to_fp16, dilations = var_3635_dilations_0, groups = var_3635_groups_0, pad = var_3635_pad_0, pad_type = var_3635_pad_type_0, strides = var_3635_strides_0, weight = layers_9_feed_forward1_fc1_inlier_module_weight_to_fp16_palettized, x = input_251_cast_fp16)[name = tensor("op_3635_cast_fp16")]; - tensor var_3641_pad_type_0 = const()[name = tensor("op_3641_pad_type_0"), val = tensor("valid")]; - tensor var_3641_strides_0 = const()[name = tensor("op_3641_strides_0"), val = tensor([1, 1])]; - tensor var_3641_pad_0 = const()[name = tensor("op_3641_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3641_dilations_0 = const()[name = tensor("op_3641_dilations_0"), val = tensor([1, 1])]; - tensor var_3641_groups_0 = const()[name = tensor("op_3641_groups_0"), val = tensor(1)]; - tensor layers_9_feed_forward1_fc1_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(54604800))), name = tensor("layers_9_feed_forward1_fc1_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(54572672))), shape = tensor([2048, 512, 1, 1])]; - tensor var_3641_cast_fp16 = conv(dilations = var_3641_dilations_0, groups = var_3641_groups_0, pad = var_3641_pad_0, pad_type = var_3641_pad_type_0, strides = var_3641_strides_0, weight = layers_9_feed_forward1_fc1_outlier_module_weight_to_fp16_sparsified, x = input_251_cast_fp16)[name = tensor("op_3641_cast_fp16")]; - tensor input_253_cast_fp16 = add(x = var_3635_cast_fp16, y = var_3641_cast_fp16)[name = tensor("input_253_cast_fp16")]; - tensor input_255_cast_fp16 = silu(x = input_253_cast_fp16)[name = tensor("input_255_cast_fp16")]; - tensor var_3652_pad_type_0 = const()[name = tensor("op_3652_pad_type_0"), val = tensor("valid")]; - tensor var_3652_strides_0 = const()[name = tensor("op_3652_strides_0"), val = tensor([1, 1])]; - tensor var_3652_pad_0 = const()[name = tensor("op_3652_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3652_dilations_0 = const()[name = tensor("op_3652_dilations_0"), val = tensor([1, 1])]; - tensor var_3652_groups_0 = const()[name = tensor("op_3652_groups_0"), val = tensor(1)]; - tensor layers_9_feed_forward1_fc2_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(54735936))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(55522432))), name = tensor("layers_9_feed_forward1_fc2_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 2048, 1, 1])]; - tensor layers_9_feed_forward1_fc2_inlier_module_bias_to_fp16 = const()[name = tensor("layers_9_feed_forward1_fc2_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(55522624)))]; - tensor var_3652_cast_fp16 = conv(bias = layers_9_feed_forward1_fc2_inlier_module_bias_to_fp16, dilations = var_3652_dilations_0, groups = var_3652_groups_0, pad = var_3652_pad_0, pad_type = var_3652_pad_type_0, strides = var_3652_strides_0, weight = layers_9_feed_forward1_fc2_inlier_module_weight_to_fp16_palettized, x = input_255_cast_fp16)[name = tensor("op_3652_cast_fp16")]; - tensor var_3658_pad_type_0 = const()[name = tensor("op_3658_pad_type_0"), val = tensor("valid")]; - tensor var_3658_strides_0 = const()[name = tensor("op_3658_strides_0"), val = tensor([1, 1])]; - tensor var_3658_pad_0 = const()[name = tensor("op_3658_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3658_dilations_0 = const()[name = tensor("op_3658_dilations_0"), val = tensor([1, 1])]; - tensor var_3658_groups_0 = const()[name = tensor("op_3658_groups_0"), val = tensor(1)]; - tensor layers_9_feed_forward1_fc2_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(55559424))), name = tensor("layers_9_feed_forward1_fc2_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(55523712))), shape = tensor([512, 2048, 1, 1])]; - tensor var_3658_cast_fp16 = conv(dilations = var_3658_dilations_0, groups = var_3658_groups_0, pad = var_3658_pad_0, pad_type = var_3658_pad_type_0, strides = var_3658_strides_0, weight = layers_9_feed_forward1_fc2_outlier_module_weight_to_fp16_sparsified, x = input_255_cast_fp16)[name = tensor("op_3658_cast_fp16")]; - tensor x_57_cast_fp16 = add(x = var_3652_cast_fp16, y = var_3658_cast_fp16)[name = tensor("x_57_cast_fp16")]; - tensor var_3660_to_fp16 = const()[name = tensor("op_3660_to_fp16"), val = tensor(0x1p-1)]; - tensor var_3661_cast_fp16 = mul(x = x_57_cast_fp16, y = var_3660_to_fp16)[name = tensor("op_3661_cast_fp16")]; - tensor inputs_93_cast_fp16 = add(x = inputs_91_cast_fp16, y = var_3661_cast_fp16)[name = tensor("inputs_93_cast_fp16")]; - tensor out_93_axes_0 = const()[name = tensor("out_93_axes_0"), val = tensor([1])]; - tensor var_3671_to_fp16 = const()[name = tensor("op_3671_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_93_cast_fp16 = layer_norm(axes = out_93_axes_0, epsilon = var_3671_to_fp16, x = inputs_93_cast_fp16)[name = tensor("out_93_cast_fp16")]; - tensor obj_39_gamma_0_to_fp16 = const()[name = tensor("obj_39_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(55690560)))]; - tensor obj_39_beta_0_to_fp16 = const()[name = tensor("obj_39_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(55691648)))]; - tensor obj_39_epsilon_0_to_fp16 = const()[name = tensor("obj_39_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor obj_39_cast_fp16 = batch_norm(beta = obj_39_beta_0_to_fp16, epsilon = obj_39_epsilon_0_to_fp16, gamma = obj_39_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_93_cast_fp16)[name = tensor("obj_39_cast_fp16")]; - tensor var_3696_pad_type_0 = const()[name = tensor("op_3696_pad_type_0"), val = tensor("valid")]; - tensor var_3696_strides_0 = const()[name = tensor("op_3696_strides_0"), val = tensor([1, 1])]; - tensor var_3696_pad_0 = const()[name = tensor("op_3696_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3696_dilations_0 = const()[name = tensor("op_3696_dilations_0"), val = tensor([1, 1])]; - tensor var_3696_groups_0 = const()[name = tensor("op_3696_groups_0"), val = tensor(1)]; - tensor layers_9_self_attn_q_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(55692736))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(55889408))), name = tensor("layers_9_self_attn_q_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_9_self_attn_q_proj_inlier_module_bias_to_fp16 = const()[name = tensor("layers_9_self_attn_q_proj_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(55889600)))]; - tensor var_3696_cast_fp16 = conv(bias = layers_9_self_attn_q_proj_inlier_module_bias_to_fp16, dilations = var_3696_dilations_0, groups = var_3696_groups_0, pad = var_3696_pad_0, pad_type = var_3696_pad_type_0, strides = var_3696_strides_0, weight = layers_9_self_attn_q_proj_inlier_module_weight_to_fp16_palettized, x = obj_39_cast_fp16)[name = tensor("op_3696_cast_fp16")]; - tensor var_3702_pad_type_0 = const()[name = tensor("op_3702_pad_type_0"), val = tensor("valid")]; - tensor var_3702_strides_0 = const()[name = tensor("op_3702_strides_0"), val = tensor([1, 1])]; - tensor var_3702_pad_0 = const()[name = tensor("op_3702_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3702_dilations_0 = const()[name = tensor("op_3702_dilations_0"), val = tensor([1, 1])]; - tensor var_3702_groups_0 = const()[name = tensor("op_3702_groups_0"), val = tensor(1)]; - tensor layers_9_self_attn_q_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(55900672))), name = tensor("layers_9_self_attn_q_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(55890688))), shape = tensor([512, 512, 1, 1])]; - tensor var_3702_cast_fp16 = conv(dilations = var_3702_dilations_0, groups = var_3702_groups_0, pad = var_3702_pad_0, pad_type = var_3702_pad_type_0, strides = var_3702_strides_0, weight = layers_9_self_attn_q_proj_outlier_module_weight_to_fp16_sparsified, x = obj_39_cast_fp16)[name = tensor("op_3702_cast_fp16")]; - tensor query_37_cast_fp16 = add(x = var_3696_cast_fp16, y = var_3702_cast_fp16)[name = tensor("query_37_cast_fp16")]; - tensor var_3711_pad_type_0 = const()[name = tensor("op_3711_pad_type_0"), val = tensor("valid")]; - tensor var_3711_strides_0 = const()[name = tensor("op_3711_strides_0"), val = tensor([1, 1])]; - tensor var_3711_pad_0 = const()[name = tensor("op_3711_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3711_dilations_0 = const()[name = tensor("op_3711_dilations_0"), val = tensor([1, 1])]; - tensor var_3711_groups_0 = const()[name = tensor("op_3711_groups_0"), val = tensor(1)]; - tensor layers_9_self_attn_k_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(55933504))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(56130176))), name = tensor("layers_9_self_attn_k_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor var_3711_cast_fp16 = conv(dilations = var_3711_dilations_0, groups = var_3711_groups_0, pad = var_3711_pad_0, pad_type = var_3711_pad_type_0, strides = var_3711_strides_0, weight = layers_9_self_attn_k_proj_inlier_module_weight_to_fp16_palettized, x = obj_39_cast_fp16)[name = tensor("op_3711_cast_fp16")]; - tensor var_3717_pad_type_0 = const()[name = tensor("op_3717_pad_type_0"), val = tensor("valid")]; - tensor var_3717_strides_0 = const()[name = tensor("op_3717_strides_0"), val = tensor([1, 1])]; - tensor var_3717_pad_0 = const()[name = tensor("op_3717_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3717_dilations_0 = const()[name = tensor("op_3717_dilations_0"), val = tensor([1, 1])]; - tensor var_3717_groups_0 = const()[name = tensor("op_3717_groups_0"), val = tensor(1)]; - tensor layers_9_self_attn_k_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(56138432))), name = tensor("layers_9_self_attn_k_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(56130368))), shape = tensor([512, 512, 1, 1])]; - tensor var_3717_cast_fp16 = conv(dilations = var_3717_dilations_0, groups = var_3717_groups_0, pad = var_3717_pad_0, pad_type = var_3717_pad_type_0, strides = var_3717_strides_0, weight = layers_9_self_attn_k_proj_outlier_module_weight_to_fp16_sparsified, x = obj_39_cast_fp16)[name = tensor("op_3717_cast_fp16")]; - tensor key_19_cast_fp16 = add(x = var_3711_cast_fp16, y = var_3717_cast_fp16)[name = tensor("key_19_cast_fp16")]; - tensor var_3727_pad_type_0 = const()[name = tensor("op_3727_pad_type_0"), val = tensor("valid")]; - tensor var_3727_strides_0 = const()[name = tensor("op_3727_strides_0"), val = tensor([1, 1])]; - tensor var_3727_pad_0 = const()[name = tensor("op_3727_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3727_dilations_0 = const()[name = tensor("op_3727_dilations_0"), val = tensor([1, 1])]; - tensor var_3727_groups_0 = const()[name = tensor("op_3727_groups_0"), val = tensor(1)]; - tensor layers_9_self_attn_v_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(56171264))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(56367936))), name = tensor("layers_9_self_attn_v_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_9_self_attn_v_proj_inlier_module_bias_to_fp16 = const()[name = tensor("layers_9_self_attn_v_proj_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(56368128)))]; - tensor var_3727_cast_fp16 = conv(bias = layers_9_self_attn_v_proj_inlier_module_bias_to_fp16, dilations = var_3727_dilations_0, groups = var_3727_groups_0, pad = var_3727_pad_0, pad_type = var_3727_pad_type_0, strides = var_3727_strides_0, weight = layers_9_self_attn_v_proj_inlier_module_weight_to_fp16_palettized, x = obj_39_cast_fp16)[name = tensor("op_3727_cast_fp16")]; - tensor var_3733_pad_type_0 = const()[name = tensor("op_3733_pad_type_0"), val = tensor("valid")]; - tensor var_3733_strides_0 = const()[name = tensor("op_3733_strides_0"), val = tensor([1, 1])]; - tensor var_3733_pad_0 = const()[name = tensor("op_3733_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3733_dilations_0 = const()[name = tensor("op_3733_dilations_0"), val = tensor([1, 1])]; - tensor var_3733_groups_0 = const()[name = tensor("op_3733_groups_0"), val = tensor(1)]; - tensor layers_9_self_attn_v_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(56376768))), name = tensor("layers_9_self_attn_v_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(56369216))), shape = tensor([512, 512, 1, 1])]; - tensor var_3733_cast_fp16 = conv(dilations = var_3733_dilations_0, groups = var_3733_groups_0, pad = var_3733_pad_0, pad_type = var_3733_pad_type_0, strides = var_3733_strides_0, weight = layers_9_self_attn_v_proj_outlier_module_weight_to_fp16_sparsified, x = obj_39_cast_fp16)[name = tensor("op_3733_cast_fp16")]; - tensor value_19_cast_fp16 = add(x = var_3727_cast_fp16, y = var_3733_cast_fp16)[name = tensor("value_19_cast_fp16")]; - tensor var_3736_to_fp16 = const()[name = tensor("op_3736_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(56409600)))]; - tensor query_39_cast_fp16 = add(x = query_37_cast_fp16, y = var_3736_to_fp16)[name = tensor("query_39_cast_fp16")]; - tensor var_3739_to_fp16 = const()[name = tensor("op_3739_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(56410688)))]; - tensor q_with_bias_v_19_cast_fp16 = add(x = query_37_cast_fp16, y = var_3739_to_fp16)[name = tensor("q_with_bias_v_19_cast_fp16")]; - tensor var_3749_pad_type_0 = const()[name = tensor("op_3749_pad_type_0"), val = tensor("valid")]; - tensor var_3749_strides_0 = const()[name = tensor("op_3749_strides_0"), val = tensor([1, 1])]; - tensor var_3749_pad_0 = const()[name = tensor("op_3749_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3749_dilations_0 = const()[name = tensor("op_3749_dilations_0"), val = tensor([1, 1])]; - tensor var_3749_groups_0 = const()[name = tensor("op_3749_groups_0"), val = tensor(1)]; - tensor layers_9_self_attn_linear_pos_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(56411776))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(56608448))), name = tensor("layers_9_self_attn_linear_pos_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor var_3749_cast_fp16 = conv(bias = input_17_mean_0_to_fp16, dilations = var_3749_dilations_0, groups = var_3749_groups_0, pad = var_3749_pad_0, pad_type = var_3749_pad_type_0, strides = var_3749_strides_0, weight = layers_9_self_attn_linear_pos_inlier_module_weight_to_fp16_palettized, x = obj_3_cast_fp16)[name = tensor("op_3749_cast_fp16")]; - tensor var_3755_pad_type_0 = const()[name = tensor("op_3755_pad_type_0"), val = tensor("valid")]; - tensor var_3755_strides_0 = const()[name = tensor("op_3755_strides_0"), val = tensor([1, 1])]; - tensor var_3755_pad_0 = const()[name = tensor("op_3755_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3755_dilations_0 = const()[name = tensor("op_3755_dilations_0"), val = tensor([1, 1])]; - tensor var_3755_groups_0 = const()[name = tensor("op_3755_groups_0"), val = tensor(1)]; - tensor layers_9_self_attn_linear_pos_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(56624256))), name = tensor("layers_9_self_attn_linear_pos_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(56608640))), shape = tensor([512, 512, 1, 1])]; - tensor var_3755_cast_fp16 = conv(dilations = var_3755_dilations_0, groups = var_3755_groups_0, pad = var_3755_pad_0, pad_type = var_3755_pad_type_0, strides = var_3755_strides_0, weight = layers_9_self_attn_linear_pos_outlier_module_weight_to_fp16_sparsified, x = obj_3_cast_fp16)[name = tensor("op_3755_cast_fp16")]; - tensor p_19_cast_fp16 = add(x = var_3749_cast_fp16, y = var_3755_cast_fp16)[name = tensor("p_19_cast_fp16")]; - tensor var_3759 = const()[name = tensor("op_3759"), val = tensor([1, 8, 64, 188])]; - tensor var_3760_cast_fp16 = reshape(shape = var_3759, x = q_with_bias_v_19_cast_fp16)[name = tensor("op_3760_cast_fp16")]; - tensor var_3761 = const()[name = tensor("op_3761"), val = tensor([1, 8, 64, -1])]; - tensor var_3762_cast_fp16 = reshape(shape = var_3761, x = p_19_cast_fp16)[name = tensor("op_3762_cast_fp16")]; - tensor matrix_bd_73_transpose_x_0 = const()[name = tensor("matrix_bd_73_transpose_x_0"), val = tensor(true)]; - tensor matrix_bd_73_transpose_y_0 = const()[name = tensor("matrix_bd_73_transpose_y_0"), val = tensor(false)]; - tensor matrix_bd_73_cast_fp16 = matmul(transpose_x = matrix_bd_73_transpose_x_0, transpose_y = matrix_bd_73_transpose_y_0, x = var_3760_cast_fp16, y = var_3762_cast_fp16)[name = tensor("matrix_bd_73_cast_fp16")]; - tensor matrix_bd_75_pad_0 = const()[name = tensor("matrix_bd_75_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor matrix_bd_75_mode_0 = const()[name = tensor("matrix_bd_75_mode_0"), val = tensor("constant")]; - tensor const_109_to_fp16 = const()[name = tensor("const_109_to_fp16"), val = tensor(0x0p+0)]; - tensor matrix_bd_75_cast_fp16 = pad(constant_val = const_109_to_fp16, mode = matrix_bd_75_mode_0, pad = matrix_bd_75_pad_0, x = matrix_bd_73_cast_fp16)[name = tensor("matrix_bd_75_cast_fp16")]; - tensor var_3771 = const()[name = tensor("op_3771"), val = tensor([1, 8, -1, 188])]; - tensor matrix_bd_77_cast_fp16 = reshape(shape = var_3771, x = matrix_bd_75_cast_fp16)[name = tensor("matrix_bd_77_cast_fp16")]; - tensor var_3775_begin_0 = const()[name = tensor("op_3775_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_3775_end_0 = const()[name = tensor("op_3775_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_3775_end_mask_0 = const()[name = tensor("op_3775_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_3775_cast_fp16 = slice_by_index(begin = var_3775_begin_0, end = var_3775_end_0, end_mask = var_3775_end_mask_0, x = matrix_bd_77_cast_fp16)[name = tensor("op_3775_cast_fp16")]; - tensor var_3776 = const()[name = tensor("op_3776"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_79_cast_fp16 = reshape(shape = var_3776, x = var_3775_cast_fp16)[name = tensor("matrix_bd_79_cast_fp16")]; - tensor var_3781_begin_0 = const()[name = tensor("op_3781_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3781_end_0 = const()[name = tensor("op_3781_end_0"), val = tensor([1, 8, 188, 188])]; - tensor var_3781_end_mask_0 = const()[name = tensor("op_3781_end_mask_0"), val = tensor([true, true, true, false])]; - tensor var_3781_cast_fp16 = slice_by_index(begin = var_3781_begin_0, end = var_3781_end_0, end_mask = var_3781_end_mask_0, x = matrix_bd_79_cast_fp16)[name = tensor("op_3781_cast_fp16")]; - tensor var_3782_to_fp16 = const()[name = tensor("op_3782_to_fp16"), val = tensor(0x1p-3)]; - tensor qk_mask_19_cast_fp16 = mul(x = var_3781_cast_fp16, y = var_3782_to_fp16)[name = tensor("qk_mask_19_cast_fp16")]; - tensor var_3786 = const()[name = tensor("op_3786"), val = tensor([1, 8, 64, 188])]; - tensor mh_q_19_cast_fp16 = reshape(shape = var_3786, x = query_39_cast_fp16)[name = tensor("mh_q_19_cast_fp16")]; - tensor var_3788_to_fp16 = const()[name = tensor("op_3788_to_fp16"), val = tensor(0x1p-3)]; - tensor var_3789_cast_fp16 = mul(x = mh_q_19_cast_fp16, y = var_3788_to_fp16)[name = tensor("op_3789_cast_fp16")]; - tensor var_3792 = const()[name = tensor("op_3792"), val = tensor([1, 8, 64, 188])]; - tensor var_3793_cast_fp16 = reshape(shape = var_3792, x = key_19_cast_fp16)[name = tensor("op_3793_cast_fp16")]; - tensor mh_w_37_transpose_x_0 = const()[name = tensor("mh_w_37_transpose_x_0"), val = tensor(true)]; - tensor mh_w_37_transpose_y_0 = const()[name = tensor("mh_w_37_transpose_y_0"), val = tensor(false)]; - tensor mh_w_37_cast_fp16 = matmul(transpose_x = mh_w_37_transpose_x_0, transpose_y = mh_w_37_transpose_y_0, x = var_3789_cast_fp16, y = var_3793_cast_fp16)[name = tensor("mh_w_37_cast_fp16")]; - tensor mh_w_39_cast_fp16 = add(x = mh_w_37_cast_fp16, y = qk_mask_19_cast_fp16)[name = tensor("mh_w_39_cast_fp16")]; - tensor var_3797_cast_fp16 = softmax(axis = var_3584, x = mh_w_39_cast_fp16)[name = tensor("op_3797_cast_fp16")]; - tensor var_3798 = const()[name = tensor("op_3798"), val = tensor([1, 8, 64, 188])]; - tensor var_3799_cast_fp16 = reshape(shape = var_3798, x = value_19_cast_fp16)[name = tensor("op_3799_cast_fp16")]; - tensor attn_19_transpose_x_0 = const()[name = tensor("attn_19_transpose_x_0"), val = tensor(false)]; - tensor attn_19_transpose_y_0 = const()[name = tensor("attn_19_transpose_y_0"), val = tensor(true)]; - tensor attn_19_cast_fp16 = matmul(transpose_x = attn_19_transpose_x_0, transpose_y = attn_19_transpose_y_0, x = var_3799_cast_fp16, y = var_3797_cast_fp16)[name = tensor("attn_19_cast_fp16")]; - tensor var_3802 = const()[name = tensor("op_3802"), val = tensor([1, 512, 1, 188])]; - tensor input_257_cast_fp16 = reshape(shape = var_3802, x = attn_19_cast_fp16)[name = tensor("input_257_cast_fp16")]; - tensor var_3812_pad_type_0 = const()[name = tensor("op_3812_pad_type_0"), val = tensor("valid")]; - tensor var_3812_strides_0 = const()[name = tensor("op_3812_strides_0"), val = tensor([1, 1])]; - tensor var_3812_pad_0 = const()[name = tensor("op_3812_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3812_dilations_0 = const()[name = tensor("op_3812_dilations_0"), val = tensor([1, 1])]; - tensor var_3812_groups_0 = const()[name = tensor("op_3812_groups_0"), val = tensor(1)]; - tensor layers_9_self_attn_o_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(56657088))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(56853760))), name = tensor("layers_9_self_attn_o_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_9_self_attn_o_proj_inlier_module_bias_to_fp16 = const()[name = tensor("layers_9_self_attn_o_proj_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(56853952)))]; - tensor var_3812_cast_fp16 = conv(bias = layers_9_self_attn_o_proj_inlier_module_bias_to_fp16, dilations = var_3812_dilations_0, groups = var_3812_groups_0, pad = var_3812_pad_0, pad_type = var_3812_pad_type_0, strides = var_3812_strides_0, weight = layers_9_self_attn_o_proj_inlier_module_weight_to_fp16_palettized, x = input_257_cast_fp16)[name = tensor("op_3812_cast_fp16")]; - tensor var_3818_pad_type_0 = const()[name = tensor("op_3818_pad_type_0"), val = tensor("valid")]; - tensor var_3818_strides_0 = const()[name = tensor("op_3818_strides_0"), val = tensor([1, 1])]; - tensor var_3818_pad_0 = const()[name = tensor("op_3818_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3818_dilations_0 = const()[name = tensor("op_3818_dilations_0"), val = tensor([1, 1])]; - tensor var_3818_groups_0 = const()[name = tensor("op_3818_groups_0"), val = tensor(1)]; - tensor layers_9_self_attn_o_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(56863616))), name = tensor("layers_9_self_attn_o_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(56855040))), shape = tensor([512, 512, 1, 1])]; - tensor var_3818_cast_fp16 = conv(dilations = var_3818_dilations_0, groups = var_3818_groups_0, pad = var_3818_pad_0, pad_type = var_3818_pad_type_0, strides = var_3818_strides_0, weight = layers_9_self_attn_o_proj_outlier_module_weight_to_fp16_sparsified, x = input_257_cast_fp16)[name = tensor("op_3818_cast_fp16")]; - tensor obj_41_cast_fp16 = add(x = var_3812_cast_fp16, y = var_3818_cast_fp16)[name = tensor("obj_41_cast_fp16")]; - tensor inputs_95_cast_fp16 = add(x = inputs_93_cast_fp16, y = obj_41_cast_fp16)[name = tensor("inputs_95_cast_fp16")]; - tensor out_95_axes_0 = const()[name = tensor("out_95_axes_0"), val = tensor([1])]; - tensor var_3829_to_fp16 = const()[name = tensor("op_3829_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_95_cast_fp16 = layer_norm(axes = out_95_axes_0, epsilon = var_3829_to_fp16, x = inputs_95_cast_fp16)[name = tensor("out_95_cast_fp16")]; - tensor input_259_gamma_0_to_fp16 = const()[name = tensor("input_259_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(56896448)))]; - tensor input_259_beta_0_to_fp16 = const()[name = tensor("input_259_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(56897536)))]; - tensor input_259_epsilon_0_to_fp16 = const()[name = tensor("input_259_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_259_cast_fp16 = batch_norm(beta = input_259_beta_0_to_fp16, epsilon = input_259_epsilon_0_to_fp16, gamma = input_259_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_95_cast_fp16)[name = tensor("input_259_cast_fp16")]; - tensor var_3851_pad_type_0 = const()[name = tensor("op_3851_pad_type_0"), val = tensor("valid")]; - tensor var_3851_strides_0 = const()[name = tensor("op_3851_strides_0"), val = tensor([1, 1])]; - tensor var_3851_pad_0 = const()[name = tensor("op_3851_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3851_dilations_0 = const()[name = tensor("op_3851_dilations_0"), val = tensor([1, 1])]; - tensor var_3851_groups_0 = const()[name = tensor("op_3851_groups_0"), val = tensor(1)]; - tensor layers_9_conv_pointwise_conv1_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(56898624))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57291904))), name = tensor("layers_9_conv_pointwise_conv1_inlier_module_weight_to_fp16_palettized"), shape = tensor([1024, 512, 1, 1])]; - tensor layers_9_conv_pointwise_conv1_inlier_module_bias_to_fp16 = const()[name = tensor("layers_9_conv_pointwise_conv1_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57292096)))]; - tensor var_3851_cast_fp16 = conv(bias = layers_9_conv_pointwise_conv1_inlier_module_bias_to_fp16, dilations = var_3851_dilations_0, groups = var_3851_groups_0, pad = var_3851_pad_0, pad_type = var_3851_pad_type_0, strides = var_3851_strides_0, weight = layers_9_conv_pointwise_conv1_inlier_module_weight_to_fp16_palettized, x = input_259_cast_fp16)[name = tensor("op_3851_cast_fp16")]; - tensor var_3857_pad_type_0 = const()[name = tensor("op_3857_pad_type_0"), val = tensor("valid")]; - tensor var_3857_strides_0 = const()[name = tensor("op_3857_strides_0"), val = tensor([1, 1])]; - tensor var_3857_pad_0 = const()[name = tensor("op_3857_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3857_dilations_0 = const()[name = tensor("op_3857_dilations_0"), val = tensor([1, 1])]; - tensor var_3857_groups_0 = const()[name = tensor("op_3857_groups_0"), val = tensor(1)]; - tensor layers_9_conv_pointwise_conv1_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57311296))), name = tensor("layers_9_conv_pointwise_conv1_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57294208))), shape = tensor([1024, 512, 1, 1])]; - tensor var_3857_cast_fp16 = conv(dilations = var_3857_dilations_0, groups = var_3857_groups_0, pad = var_3857_pad_0, pad_type = var_3857_pad_type_0, strides = var_3857_strides_0, weight = layers_9_conv_pointwise_conv1_outlier_module_weight_to_fp16_sparsified, x = input_259_cast_fp16)[name = tensor("op_3857_cast_fp16")]; - tensor input_261_cast_fp16 = add(x = var_3851_cast_fp16, y = var_3857_cast_fp16)[name = tensor("input_261_cast_fp16")]; - tensor input_263_split_num_splits_0 = const()[name = tensor("input_263_split_num_splits_0"), val = tensor(2)]; - tensor input_263_split_axis_0 = const()[name = tensor("input_263_split_axis_0"), val = tensor(1)]; - tensor input_263_split_cast_fp16_0, tensor input_263_split_cast_fp16_1 = split(axis = input_263_split_axis_0, num_splits = input_263_split_num_splits_0, x = input_261_cast_fp16)[name = tensor("input_263_split_cast_fp16")]; - tensor input_263_split_1_sigmoid_cast_fp16 = sigmoid(x = input_263_split_cast_fp16_1)[name = tensor("input_263_split_1_sigmoid_cast_fp16")]; - tensor input_263_cast_fp16 = mul(x = input_263_split_cast_fp16_0, y = input_263_split_1_sigmoid_cast_fp16)[name = tensor("input_263_cast_fp16")]; - tensor input_265_pad_type_0 = const()[name = tensor("input_265_pad_type_0"), val = tensor("custom")]; - tensor input_265_pad_0 = const()[name = tensor("input_265_pad_0"), val = tensor([0, 0, 4, 4])]; - tensor input_265_groups_0 = const()[name = tensor("input_265_groups_0"), val = tensor(512)]; - tensor input_265_strides_0 = const()[name = tensor("input_265_strides_0"), val = tensor([1, 1])]; - tensor input_265_dilations_0 = const()[name = tensor("input_265_dilations_0"), val = tensor([1, 1])]; - tensor const_209_to_fp16 = const()[name = tensor("const_209_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57376896)))]; - tensor const_210_to_fp16 = const()[name = tensor("const_210_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57386176)))]; - tensor input_267_cast_fp16 = conv(bias = const_210_to_fp16, dilations = input_265_dilations_0, groups = input_265_groups_0, pad = input_265_pad_0, pad_type = input_265_pad_type_0, strides = input_265_strides_0, weight = const_209_to_fp16, x = input_263_cast_fp16)[name = tensor("input_267_cast_fp16")]; - tensor input_269_cast_fp16 = silu(x = input_267_cast_fp16)[name = tensor("input_269_cast_fp16")]; - tensor var_3881_pad_type_0 = const()[name = tensor("op_3881_pad_type_0"), val = tensor("valid")]; - tensor var_3881_strides_0 = const()[name = tensor("op_3881_strides_0"), val = tensor([1, 1])]; - tensor var_3881_pad_0 = const()[name = tensor("op_3881_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3881_dilations_0 = const()[name = tensor("op_3881_dilations_0"), val = tensor([1, 1])]; - tensor var_3881_groups_0 = const()[name = tensor("op_3881_groups_0"), val = tensor(1)]; - tensor layers_9_conv_pointwise_conv2_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57387264))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57583936))), name = tensor("layers_9_conv_pointwise_conv2_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_9_conv_pointwise_conv2_inlier_module_bias_to_fp16 = const()[name = tensor("layers_9_conv_pointwise_conv2_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57584128)))]; - tensor var_3881_cast_fp16 = conv(bias = layers_9_conv_pointwise_conv2_inlier_module_bias_to_fp16, dilations = var_3881_dilations_0, groups = var_3881_groups_0, pad = var_3881_pad_0, pad_type = var_3881_pad_type_0, strides = var_3881_strides_0, weight = layers_9_conv_pointwise_conv2_inlier_module_weight_to_fp16_palettized, x = input_269_cast_fp16)[name = tensor("op_3881_cast_fp16")]; - tensor var_3887_pad_type_0 = const()[name = tensor("op_3887_pad_type_0"), val = tensor("valid")]; - tensor var_3887_strides_0 = const()[name = tensor("op_3887_strides_0"), val = tensor([1, 1])]; - tensor var_3887_pad_0 = const()[name = tensor("op_3887_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3887_dilations_0 = const()[name = tensor("op_3887_dilations_0"), val = tensor([1, 1])]; - tensor var_3887_groups_0 = const()[name = tensor("op_3887_groups_0"), val = tensor(1)]; - tensor layers_9_conv_pointwise_conv2_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57593408))), name = tensor("layers_9_conv_pointwise_conv2_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57585216))), shape = tensor([512, 512, 1, 1])]; - tensor var_3887_cast_fp16 = conv(dilations = var_3887_dilations_0, groups = var_3887_groups_0, pad = var_3887_pad_0, pad_type = var_3887_pad_type_0, strides = var_3887_strides_0, weight = layers_9_conv_pointwise_conv2_outlier_module_weight_to_fp16_sparsified, x = input_269_cast_fp16)[name = tensor("op_3887_cast_fp16")]; - tensor x_59_cast_fp16 = add(x = var_3881_cast_fp16, y = var_3887_cast_fp16)[name = tensor("x_59_cast_fp16")]; - tensor inputs_97_cast_fp16 = add(x = inputs_95_cast_fp16, y = x_59_cast_fp16)[name = tensor("inputs_97_cast_fp16")]; - tensor out_97_axes_0 = const()[name = tensor("out_97_axes_0"), val = tensor([1])]; - tensor var_3898_to_fp16 = const()[name = tensor("op_3898_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_97_cast_fp16 = layer_norm(axes = out_97_axes_0, epsilon = var_3898_to_fp16, x = inputs_97_cast_fp16)[name = tensor("out_97_cast_fp16")]; - tensor input_271_gamma_0_to_fp16 = const()[name = tensor("input_271_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57626240)))]; - tensor input_271_beta_0_to_fp16 = const()[name = tensor("input_271_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57627328)))]; - tensor input_271_epsilon_0_to_fp16 = const()[name = tensor("input_271_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_271_cast_fp16 = batch_norm(beta = input_271_beta_0_to_fp16, epsilon = input_271_epsilon_0_to_fp16, gamma = input_271_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_97_cast_fp16)[name = tensor("input_271_cast_fp16")]; - tensor var_3918_pad_type_0 = const()[name = tensor("op_3918_pad_type_0"), val = tensor("valid")]; - tensor var_3918_strides_0 = const()[name = tensor("op_3918_strides_0"), val = tensor([1, 1])]; - tensor var_3918_pad_0 = const()[name = tensor("op_3918_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3918_dilations_0 = const()[name = tensor("op_3918_dilations_0"), val = tensor([1, 1])]; - tensor var_3918_groups_0 = const()[name = tensor("op_3918_groups_0"), val = tensor(1)]; - tensor layers_9_feed_forward2_fc1_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57628416))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58414912))), name = tensor("layers_9_feed_forward2_fc1_inlier_module_weight_to_fp16_palettized"), shape = tensor([2048, 512, 1, 1])]; - tensor layers_9_feed_forward2_fc1_inlier_module_bias_to_fp16 = const()[name = tensor("layers_9_feed_forward2_fc1_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58415104)))]; - tensor var_3918_cast_fp16 = conv(bias = layers_9_feed_forward2_fc1_inlier_module_bias_to_fp16, dilations = var_3918_dilations_0, groups = var_3918_groups_0, pad = var_3918_pad_0, pad_type = var_3918_pad_type_0, strides = var_3918_strides_0, weight = layers_9_feed_forward2_fc1_inlier_module_weight_to_fp16_palettized, x = input_271_cast_fp16)[name = tensor("op_3918_cast_fp16")]; - tensor var_3924_pad_type_0 = const()[name = tensor("op_3924_pad_type_0"), val = tensor("valid")]; - tensor var_3924_strides_0 = const()[name = tensor("op_3924_strides_0"), val = tensor([1, 1])]; - tensor var_3924_pad_0 = const()[name = tensor("op_3924_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3924_dilations_0 = const()[name = tensor("op_3924_dilations_0"), val = tensor([1, 1])]; - tensor var_3924_groups_0 = const()[name = tensor("op_3924_groups_0"), val = tensor(1)]; - tensor layers_9_feed_forward2_fc1_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58451456))), name = tensor("layers_9_feed_forward2_fc1_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58419264))), shape = tensor([2048, 512, 1, 1])]; - tensor var_3924_cast_fp16 = conv(dilations = var_3924_dilations_0, groups = var_3924_groups_0, pad = var_3924_pad_0, pad_type = var_3924_pad_type_0, strides = var_3924_strides_0, weight = layers_9_feed_forward2_fc1_outlier_module_weight_to_fp16_sparsified, x = input_271_cast_fp16)[name = tensor("op_3924_cast_fp16")]; - tensor input_273_cast_fp16 = add(x = var_3918_cast_fp16, y = var_3924_cast_fp16)[name = tensor("input_273_cast_fp16")]; - tensor input_275_cast_fp16 = silu(x = input_273_cast_fp16)[name = tensor("input_275_cast_fp16")]; - tensor var_3935_pad_type_0 = const()[name = tensor("op_3935_pad_type_0"), val = tensor("valid")]; - tensor var_3935_strides_0 = const()[name = tensor("op_3935_strides_0"), val = tensor([1, 1])]; - tensor var_3935_pad_0 = const()[name = tensor("op_3935_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3935_dilations_0 = const()[name = tensor("op_3935_dilations_0"), val = tensor([1, 1])]; - tensor var_3935_groups_0 = const()[name = tensor("op_3935_groups_0"), val = tensor(1)]; - tensor layers_9_feed_forward2_fc2_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58582592))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59369088))), name = tensor("layers_9_feed_forward2_fc2_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 2048, 1, 1])]; - tensor layers_9_feed_forward2_fc2_inlier_module_bias_to_fp16 = const()[name = tensor("layers_9_feed_forward2_fc2_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59369280)))]; - tensor var_3935_cast_fp16 = conv(bias = layers_9_feed_forward2_fc2_inlier_module_bias_to_fp16, dilations = var_3935_dilations_0, groups = var_3935_groups_0, pad = var_3935_pad_0, pad_type = var_3935_pad_type_0, strides = var_3935_strides_0, weight = layers_9_feed_forward2_fc2_inlier_module_weight_to_fp16_palettized, x = input_275_cast_fp16)[name = tensor("op_3935_cast_fp16")]; - tensor var_3941_pad_type_0 = const()[name = tensor("op_3941_pad_type_0"), val = tensor("valid")]; - tensor var_3941_strides_0 = const()[name = tensor("op_3941_strides_0"), val = tensor([1, 1])]; - tensor var_3941_pad_0 = const()[name = tensor("op_3941_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_3941_dilations_0 = const()[name = tensor("op_3941_dilations_0"), val = tensor([1, 1])]; - tensor var_3941_groups_0 = const()[name = tensor("op_3941_groups_0"), val = tensor(1)]; - tensor layers_9_feed_forward2_fc2_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59408640))), name = tensor("layers_9_feed_forward2_fc2_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59370368))), shape = tensor([512, 2048, 1, 1])]; - tensor var_3941_cast_fp16 = conv(dilations = var_3941_dilations_0, groups = var_3941_groups_0, pad = var_3941_pad_0, pad_type = var_3941_pad_type_0, strides = var_3941_strides_0, weight = layers_9_feed_forward2_fc2_outlier_module_weight_to_fp16_sparsified, x = input_275_cast_fp16)[name = tensor("op_3941_cast_fp16")]; - tensor x_61_cast_fp16 = add(x = var_3935_cast_fp16, y = var_3941_cast_fp16)[name = tensor("x_61_cast_fp16")]; - tensor var_3943_to_fp16 = const()[name = tensor("op_3943_to_fp16"), val = tensor(0x1p-1)]; - tensor var_3944_cast_fp16 = mul(x = x_61_cast_fp16, y = var_3943_to_fp16)[name = tensor("op_3944_cast_fp16")]; - tensor inputs_99_cast_fp16 = add(x = inputs_97_cast_fp16, y = var_3944_cast_fp16)[name = tensor("inputs_99_cast_fp16")]; - tensor out_99_axes_0 = const()[name = tensor("out_99_axes_0"), val = tensor([1])]; - tensor var_3954_to_fp16 = const()[name = tensor("op_3954_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_99_cast_fp16 = layer_norm(axes = out_99_axes_0, epsilon = var_3954_to_fp16, x = inputs_99_cast_fp16)[name = tensor("out_99_cast_fp16")]; - tensor inputs_101_gamma_0_to_fp16 = const()[name = tensor("inputs_101_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59539776)))]; - tensor inputs_101_beta_0_to_fp16 = const()[name = tensor("inputs_101_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59540864)))]; - tensor inputs_101_epsilon_0_to_fp16 = const()[name = tensor("inputs_101_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor inputs_101_cast_fp16 = batch_norm(beta = inputs_101_beta_0_to_fp16, epsilon = inputs_101_epsilon_0_to_fp16, gamma = inputs_101_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_99_cast_fp16)[name = tensor("inputs_101_cast_fp16")]; - tensor var_3968 = const()[name = tensor("op_3968"), val = tensor(3)]; - tensor out_101_axes_0 = const()[name = tensor("out_101_axes_0"), val = tensor([1])]; - tensor var_3999_to_fp16 = const()[name = tensor("op_3999_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_101_cast_fp16 = layer_norm(axes = out_101_axes_0, epsilon = var_3999_to_fp16, x = inputs_101_cast_fp16)[name = tensor("out_101_cast_fp16")]; - tensor input_277_gamma_0_to_fp16 = const()[name = tensor("input_277_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59541952)))]; - tensor input_277_beta_0_to_fp16 = const()[name = tensor("input_277_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59543040)))]; - tensor input_277_epsilon_0_to_fp16 = const()[name = tensor("input_277_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_277_cast_fp16 = batch_norm(beta = input_277_beta_0_to_fp16, epsilon = input_277_epsilon_0_to_fp16, gamma = input_277_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_101_cast_fp16)[name = tensor("input_277_cast_fp16")]; - tensor var_4019_pad_type_0 = const()[name = tensor("op_4019_pad_type_0"), val = tensor("valid")]; - tensor var_4019_strides_0 = const()[name = tensor("op_4019_strides_0"), val = tensor([1, 1])]; - tensor var_4019_pad_0 = const()[name = tensor("op_4019_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4019_dilations_0 = const()[name = tensor("op_4019_dilations_0"), val = tensor([1, 1])]; - tensor var_4019_groups_0 = const()[name = tensor("op_4019_groups_0"), val = tensor(1)]; - tensor layers_10_feed_forward1_fc1_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59544128))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60330624))), name = tensor("layers_10_feed_forward1_fc1_inlier_module_weight_to_fp16_palettized"), shape = tensor([2048, 512, 1, 1])]; - tensor layers_10_feed_forward1_fc1_inlier_module_bias_to_fp16 = const()[name = tensor("layers_10_feed_forward1_fc1_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60330816)))]; - tensor var_4019_cast_fp16 = conv(bias = layers_10_feed_forward1_fc1_inlier_module_bias_to_fp16, dilations = var_4019_dilations_0, groups = var_4019_groups_0, pad = var_4019_pad_0, pad_type = var_4019_pad_type_0, strides = var_4019_strides_0, weight = layers_10_feed_forward1_fc1_inlier_module_weight_to_fp16_palettized, x = input_277_cast_fp16)[name = tensor("op_4019_cast_fp16")]; - tensor var_4025_pad_type_0 = const()[name = tensor("op_4025_pad_type_0"), val = tensor("valid")]; - tensor var_4025_strides_0 = const()[name = tensor("op_4025_strides_0"), val = tensor([1, 1])]; - tensor var_4025_pad_0 = const()[name = tensor("op_4025_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4025_dilations_0 = const()[name = tensor("op_4025_dilations_0"), val = tensor([1, 1])]; - tensor var_4025_groups_0 = const()[name = tensor("op_4025_groups_0"), val = tensor(1)]; - tensor layers_10_feed_forward1_fc1_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60369216))), name = tensor("layers_10_feed_forward1_fc1_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60334976))), shape = tensor([2048, 512, 1, 1])]; - tensor var_4025_cast_fp16 = conv(dilations = var_4025_dilations_0, groups = var_4025_groups_0, pad = var_4025_pad_0, pad_type = var_4025_pad_type_0, strides = var_4025_strides_0, weight = layers_10_feed_forward1_fc1_outlier_module_weight_to_fp16_sparsified, x = input_277_cast_fp16)[name = tensor("op_4025_cast_fp16")]; - tensor input_279_cast_fp16 = add(x = var_4019_cast_fp16, y = var_4025_cast_fp16)[name = tensor("input_279_cast_fp16")]; - tensor input_281_cast_fp16 = silu(x = input_279_cast_fp16)[name = tensor("input_281_cast_fp16")]; - tensor var_4036_pad_type_0 = const()[name = tensor("op_4036_pad_type_0"), val = tensor("valid")]; - tensor var_4036_strides_0 = const()[name = tensor("op_4036_strides_0"), val = tensor([1, 1])]; - tensor var_4036_pad_0 = const()[name = tensor("op_4036_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4036_dilations_0 = const()[name = tensor("op_4036_dilations_0"), val = tensor([1, 1])]; - tensor var_4036_groups_0 = const()[name = tensor("op_4036_groups_0"), val = tensor(1)]; - tensor layers_10_feed_forward1_fc2_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60500352))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61286848))), name = tensor("layers_10_feed_forward1_fc2_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 2048, 1, 1])]; - tensor layers_10_feed_forward1_fc2_inlier_module_bias_to_fp16 = const()[name = tensor("layers_10_feed_forward1_fc2_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61287040)))]; - tensor var_4036_cast_fp16 = conv(bias = layers_10_feed_forward1_fc2_inlier_module_bias_to_fp16, dilations = var_4036_dilations_0, groups = var_4036_groups_0, pad = var_4036_pad_0, pad_type = var_4036_pad_type_0, strides = var_4036_strides_0, weight = layers_10_feed_forward1_fc2_inlier_module_weight_to_fp16_palettized, x = input_281_cast_fp16)[name = tensor("op_4036_cast_fp16")]; - tensor var_4042_pad_type_0 = const()[name = tensor("op_4042_pad_type_0"), val = tensor("valid")]; - tensor var_4042_strides_0 = const()[name = tensor("op_4042_strides_0"), val = tensor([1, 1])]; - tensor var_4042_pad_0 = const()[name = tensor("op_4042_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4042_dilations_0 = const()[name = tensor("op_4042_dilations_0"), val = tensor([1, 1])]; - tensor var_4042_groups_0 = const()[name = tensor("op_4042_groups_0"), val = tensor(1)]; - tensor layers_10_feed_forward1_fc2_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61328448))), name = tensor("layers_10_feed_forward1_fc2_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61288128))), shape = tensor([512, 2048, 1, 1])]; - tensor var_4042_cast_fp16 = conv(dilations = var_4042_dilations_0, groups = var_4042_groups_0, pad = var_4042_pad_0, pad_type = var_4042_pad_type_0, strides = var_4042_strides_0, weight = layers_10_feed_forward1_fc2_outlier_module_weight_to_fp16_sparsified, x = input_281_cast_fp16)[name = tensor("op_4042_cast_fp16")]; - tensor x_63_cast_fp16 = add(x = var_4036_cast_fp16, y = var_4042_cast_fp16)[name = tensor("x_63_cast_fp16")]; - tensor var_4044_to_fp16 = const()[name = tensor("op_4044_to_fp16"), val = tensor(0x1p-1)]; - tensor var_4045_cast_fp16 = mul(x = x_63_cast_fp16, y = var_4044_to_fp16)[name = tensor("op_4045_cast_fp16")]; - tensor inputs_103_cast_fp16 = add(x = inputs_101_cast_fp16, y = var_4045_cast_fp16)[name = tensor("inputs_103_cast_fp16")]; - tensor out_103_axes_0 = const()[name = tensor("out_103_axes_0"), val = tensor([1])]; - tensor var_4055_to_fp16 = const()[name = tensor("op_4055_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_103_cast_fp16 = layer_norm(axes = out_103_axes_0, epsilon = var_4055_to_fp16, x = inputs_103_cast_fp16)[name = tensor("out_103_cast_fp16")]; - tensor obj_43_gamma_0_to_fp16 = const()[name = tensor("obj_43_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61459584)))]; - tensor obj_43_beta_0_to_fp16 = const()[name = tensor("obj_43_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61460672)))]; - tensor obj_43_epsilon_0_to_fp16 = const()[name = tensor("obj_43_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor obj_43_cast_fp16 = batch_norm(beta = obj_43_beta_0_to_fp16, epsilon = obj_43_epsilon_0_to_fp16, gamma = obj_43_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_103_cast_fp16)[name = tensor("obj_43_cast_fp16")]; - tensor var_4080_pad_type_0 = const()[name = tensor("op_4080_pad_type_0"), val = tensor("valid")]; - tensor var_4080_strides_0 = const()[name = tensor("op_4080_strides_0"), val = tensor([1, 1])]; - tensor var_4080_pad_0 = const()[name = tensor("op_4080_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4080_dilations_0 = const()[name = tensor("op_4080_dilations_0"), val = tensor([1, 1])]; - tensor var_4080_groups_0 = const()[name = tensor("op_4080_groups_0"), val = tensor(1)]; - tensor layers_10_self_attn_q_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61461760))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61658432))), name = tensor("layers_10_self_attn_q_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_10_self_attn_q_proj_inlier_module_bias_to_fp16 = const()[name = tensor("layers_10_self_attn_q_proj_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61658624)))]; - tensor var_4080_cast_fp16 = conv(bias = layers_10_self_attn_q_proj_inlier_module_bias_to_fp16, dilations = var_4080_dilations_0, groups = var_4080_groups_0, pad = var_4080_pad_0, pad_type = var_4080_pad_type_0, strides = var_4080_strides_0, weight = layers_10_self_attn_q_proj_inlier_module_weight_to_fp16_palettized, x = obj_43_cast_fp16)[name = tensor("op_4080_cast_fp16")]; - tensor var_4086_pad_type_0 = const()[name = tensor("op_4086_pad_type_0"), val = tensor("valid")]; - tensor var_4086_strides_0 = const()[name = tensor("op_4086_strides_0"), val = tensor([1, 1])]; - tensor var_4086_pad_0 = const()[name = tensor("op_4086_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4086_dilations_0 = const()[name = tensor("op_4086_dilations_0"), val = tensor([1, 1])]; - tensor var_4086_groups_0 = const()[name = tensor("op_4086_groups_0"), val = tensor(1)]; - tensor layers_10_self_attn_q_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61669056))), name = tensor("layers_10_self_attn_q_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61659712))), shape = tensor([512, 512, 1, 1])]; - tensor var_4086_cast_fp16 = conv(dilations = var_4086_dilations_0, groups = var_4086_groups_0, pad = var_4086_pad_0, pad_type = var_4086_pad_type_0, strides = var_4086_strides_0, weight = layers_10_self_attn_q_proj_outlier_module_weight_to_fp16_sparsified, x = obj_43_cast_fp16)[name = tensor("op_4086_cast_fp16")]; - tensor query_41_cast_fp16 = add(x = var_4080_cast_fp16, y = var_4086_cast_fp16)[name = tensor("query_41_cast_fp16")]; - tensor var_4095_pad_type_0 = const()[name = tensor("op_4095_pad_type_0"), val = tensor("valid")]; - tensor var_4095_strides_0 = const()[name = tensor("op_4095_strides_0"), val = tensor([1, 1])]; - tensor var_4095_pad_0 = const()[name = tensor("op_4095_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4095_dilations_0 = const()[name = tensor("op_4095_dilations_0"), val = tensor([1, 1])]; - tensor var_4095_groups_0 = const()[name = tensor("op_4095_groups_0"), val = tensor(1)]; - tensor layers_10_self_attn_k_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61701888))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61898560))), name = tensor("layers_10_self_attn_k_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor var_4095_cast_fp16 = conv(dilations = var_4095_dilations_0, groups = var_4095_groups_0, pad = var_4095_pad_0, pad_type = var_4095_pad_type_0, strides = var_4095_strides_0, weight = layers_10_self_attn_k_proj_inlier_module_weight_to_fp16_palettized, x = obj_43_cast_fp16)[name = tensor("op_4095_cast_fp16")]; - tensor var_4101_pad_type_0 = const()[name = tensor("op_4101_pad_type_0"), val = tensor("valid")]; - tensor var_4101_strides_0 = const()[name = tensor("op_4101_strides_0"), val = tensor([1, 1])]; - tensor var_4101_pad_0 = const()[name = tensor("op_4101_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4101_dilations_0 = const()[name = tensor("op_4101_dilations_0"), val = tensor([1, 1])]; - tensor var_4101_groups_0 = const()[name = tensor("op_4101_groups_0"), val = tensor(1)]; - tensor layers_10_self_attn_k_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61906368))), name = tensor("layers_10_self_attn_k_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61898752))), shape = tensor([512, 512, 1, 1])]; - tensor var_4101_cast_fp16 = conv(dilations = var_4101_dilations_0, groups = var_4101_groups_0, pad = var_4101_pad_0, pad_type = var_4101_pad_type_0, strides = var_4101_strides_0, weight = layers_10_self_attn_k_proj_outlier_module_weight_to_fp16_sparsified, x = obj_43_cast_fp16)[name = tensor("op_4101_cast_fp16")]; - tensor key_21_cast_fp16 = add(x = var_4095_cast_fp16, y = var_4101_cast_fp16)[name = tensor("key_21_cast_fp16")]; - tensor var_4111_pad_type_0 = const()[name = tensor("op_4111_pad_type_0"), val = tensor("valid")]; - tensor var_4111_strides_0 = const()[name = tensor("op_4111_strides_0"), val = tensor([1, 1])]; - tensor var_4111_pad_0 = const()[name = tensor("op_4111_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4111_dilations_0 = const()[name = tensor("op_4111_dilations_0"), val = tensor([1, 1])]; - tensor var_4111_groups_0 = const()[name = tensor("op_4111_groups_0"), val = tensor(1)]; - tensor layers_10_self_attn_v_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61939200))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(62135872))), name = tensor("layers_10_self_attn_v_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_10_self_attn_v_proj_inlier_module_bias_to_fp16 = const()[name = tensor("layers_10_self_attn_v_proj_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(62136064)))]; - tensor var_4111_cast_fp16 = conv(bias = layers_10_self_attn_v_proj_inlier_module_bias_to_fp16, dilations = var_4111_dilations_0, groups = var_4111_groups_0, pad = var_4111_pad_0, pad_type = var_4111_pad_type_0, strides = var_4111_strides_0, weight = layers_10_self_attn_v_proj_inlier_module_weight_to_fp16_palettized, x = obj_43_cast_fp16)[name = tensor("op_4111_cast_fp16")]; - tensor var_4117_pad_type_0 = const()[name = tensor("op_4117_pad_type_0"), val = tensor("valid")]; - tensor var_4117_strides_0 = const()[name = tensor("op_4117_strides_0"), val = tensor([1, 1])]; - tensor var_4117_pad_0 = const()[name = tensor("op_4117_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4117_dilations_0 = const()[name = tensor("op_4117_dilations_0"), val = tensor([1, 1])]; - tensor var_4117_groups_0 = const()[name = tensor("op_4117_groups_0"), val = tensor(1)]; - tensor layers_10_self_attn_v_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(62145024))), name = tensor("layers_10_self_attn_v_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(62137152))), shape = tensor([512, 512, 1, 1])]; - tensor var_4117_cast_fp16 = conv(dilations = var_4117_dilations_0, groups = var_4117_groups_0, pad = var_4117_pad_0, pad_type = var_4117_pad_type_0, strides = var_4117_strides_0, weight = layers_10_self_attn_v_proj_outlier_module_weight_to_fp16_sparsified, x = obj_43_cast_fp16)[name = tensor("op_4117_cast_fp16")]; - tensor value_21_cast_fp16 = add(x = var_4111_cast_fp16, y = var_4117_cast_fp16)[name = tensor("value_21_cast_fp16")]; - tensor var_4120_to_fp16 = const()[name = tensor("op_4120_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(62177856)))]; - tensor query_43_cast_fp16 = add(x = query_41_cast_fp16, y = var_4120_to_fp16)[name = tensor("query_43_cast_fp16")]; - tensor var_4123_to_fp16 = const()[name = tensor("op_4123_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(62178944)))]; - tensor q_with_bias_v_21_cast_fp16 = add(x = query_41_cast_fp16, y = var_4123_to_fp16)[name = tensor("q_with_bias_v_21_cast_fp16")]; - tensor var_4133_pad_type_0 = const()[name = tensor("op_4133_pad_type_0"), val = tensor("valid")]; - tensor var_4133_strides_0 = const()[name = tensor("op_4133_strides_0"), val = tensor([1, 1])]; - tensor var_4133_pad_0 = const()[name = tensor("op_4133_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4133_dilations_0 = const()[name = tensor("op_4133_dilations_0"), val = tensor([1, 1])]; - tensor var_4133_groups_0 = const()[name = tensor("op_4133_groups_0"), val = tensor(1)]; - tensor layers_10_self_attn_linear_pos_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(62180032))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(62376704))), name = tensor("layers_10_self_attn_linear_pos_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor var_4133_cast_fp16 = conv(bias = input_17_mean_0_to_fp16, dilations = var_4133_dilations_0, groups = var_4133_groups_0, pad = var_4133_pad_0, pad_type = var_4133_pad_type_0, strides = var_4133_strides_0, weight = layers_10_self_attn_linear_pos_inlier_module_weight_to_fp16_palettized, x = obj_3_cast_fp16)[name = tensor("op_4133_cast_fp16")]; - tensor var_4139_pad_type_0 = const()[name = tensor("op_4139_pad_type_0"), val = tensor("valid")]; - tensor var_4139_strides_0 = const()[name = tensor("op_4139_strides_0"), val = tensor([1, 1])]; - tensor var_4139_pad_0 = const()[name = tensor("op_4139_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4139_dilations_0 = const()[name = tensor("op_4139_dilations_0"), val = tensor([1, 1])]; - tensor var_4139_groups_0 = const()[name = tensor("op_4139_groups_0"), val = tensor(1)]; - tensor layers_10_self_attn_linear_pos_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(62393344))), name = tensor("layers_10_self_attn_linear_pos_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(62376896))), shape = tensor([512, 512, 1, 1])]; - tensor var_4139_cast_fp16 = conv(dilations = var_4139_dilations_0, groups = var_4139_groups_0, pad = var_4139_pad_0, pad_type = var_4139_pad_type_0, strides = var_4139_strides_0, weight = layers_10_self_attn_linear_pos_outlier_module_weight_to_fp16_sparsified, x = obj_3_cast_fp16)[name = tensor("op_4139_cast_fp16")]; - tensor p_21_cast_fp16 = add(x = var_4133_cast_fp16, y = var_4139_cast_fp16)[name = tensor("p_21_cast_fp16")]; - tensor var_4143 = const()[name = tensor("op_4143"), val = tensor([1, 8, 64, 188])]; - tensor var_4144_cast_fp16 = reshape(shape = var_4143, x = q_with_bias_v_21_cast_fp16)[name = tensor("op_4144_cast_fp16")]; - tensor var_4145 = const()[name = tensor("op_4145"), val = tensor([1, 8, 64, -1])]; - tensor var_4146_cast_fp16 = reshape(shape = var_4145, x = p_21_cast_fp16)[name = tensor("op_4146_cast_fp16")]; - tensor matrix_bd_81_transpose_x_0 = const()[name = tensor("matrix_bd_81_transpose_x_0"), val = tensor(true)]; - tensor matrix_bd_81_transpose_y_0 = const()[name = tensor("matrix_bd_81_transpose_y_0"), val = tensor(false)]; - tensor matrix_bd_81_cast_fp16 = matmul(transpose_x = matrix_bd_81_transpose_x_0, transpose_y = matrix_bd_81_transpose_y_0, x = var_4144_cast_fp16, y = var_4146_cast_fp16)[name = tensor("matrix_bd_81_cast_fp16")]; - tensor matrix_bd_83_pad_0 = const()[name = tensor("matrix_bd_83_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor matrix_bd_83_mode_0 = const()[name = tensor("matrix_bd_83_mode_0"), val = tensor("constant")]; - tensor const_120_to_fp16 = const()[name = tensor("const_120_to_fp16"), val = tensor(0x0p+0)]; - tensor matrix_bd_83_cast_fp16 = pad(constant_val = const_120_to_fp16, mode = matrix_bd_83_mode_0, pad = matrix_bd_83_pad_0, x = matrix_bd_81_cast_fp16)[name = tensor("matrix_bd_83_cast_fp16")]; - tensor var_4155 = const()[name = tensor("op_4155"), val = tensor([1, 8, -1, 188])]; - tensor matrix_bd_85_cast_fp16 = reshape(shape = var_4155, x = matrix_bd_83_cast_fp16)[name = tensor("matrix_bd_85_cast_fp16")]; - tensor var_4159_begin_0 = const()[name = tensor("op_4159_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_4159_end_0 = const()[name = tensor("op_4159_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_4159_end_mask_0 = const()[name = tensor("op_4159_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_4159_cast_fp16 = slice_by_index(begin = var_4159_begin_0, end = var_4159_end_0, end_mask = var_4159_end_mask_0, x = matrix_bd_85_cast_fp16)[name = tensor("op_4159_cast_fp16")]; - tensor var_4160 = const()[name = tensor("op_4160"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_87_cast_fp16 = reshape(shape = var_4160, x = var_4159_cast_fp16)[name = tensor("matrix_bd_87_cast_fp16")]; - tensor var_4165_begin_0 = const()[name = tensor("op_4165_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4165_end_0 = const()[name = tensor("op_4165_end_0"), val = tensor([1, 8, 188, 188])]; - tensor var_4165_end_mask_0 = const()[name = tensor("op_4165_end_mask_0"), val = tensor([true, true, true, false])]; - tensor var_4165_cast_fp16 = slice_by_index(begin = var_4165_begin_0, end = var_4165_end_0, end_mask = var_4165_end_mask_0, x = matrix_bd_87_cast_fp16)[name = tensor("op_4165_cast_fp16")]; - tensor var_4166_to_fp16 = const()[name = tensor("op_4166_to_fp16"), val = tensor(0x1p-3)]; - tensor qk_mask_21_cast_fp16 = mul(x = var_4165_cast_fp16, y = var_4166_to_fp16)[name = tensor("qk_mask_21_cast_fp16")]; - tensor var_4170 = const()[name = tensor("op_4170"), val = tensor([1, 8, 64, 188])]; - tensor mh_q_21_cast_fp16 = reshape(shape = var_4170, x = query_43_cast_fp16)[name = tensor("mh_q_21_cast_fp16")]; - tensor var_4172_to_fp16 = const()[name = tensor("op_4172_to_fp16"), val = tensor(0x1p-3)]; - tensor var_4173_cast_fp16 = mul(x = mh_q_21_cast_fp16, y = var_4172_to_fp16)[name = tensor("op_4173_cast_fp16")]; - tensor var_4176 = const()[name = tensor("op_4176"), val = tensor([1, 8, 64, 188])]; - tensor var_4177_cast_fp16 = reshape(shape = var_4176, x = key_21_cast_fp16)[name = tensor("op_4177_cast_fp16")]; - tensor mh_w_41_transpose_x_0 = const()[name = tensor("mh_w_41_transpose_x_0"), val = tensor(true)]; - tensor mh_w_41_transpose_y_0 = const()[name = tensor("mh_w_41_transpose_y_0"), val = tensor(false)]; - tensor mh_w_41_cast_fp16 = matmul(transpose_x = mh_w_41_transpose_x_0, transpose_y = mh_w_41_transpose_y_0, x = var_4173_cast_fp16, y = var_4177_cast_fp16)[name = tensor("mh_w_41_cast_fp16")]; - tensor mh_w_43_cast_fp16 = add(x = mh_w_41_cast_fp16, y = qk_mask_21_cast_fp16)[name = tensor("mh_w_43_cast_fp16")]; - tensor var_4181_cast_fp16 = softmax(axis = var_3968, x = mh_w_43_cast_fp16)[name = tensor("op_4181_cast_fp16")]; - tensor var_4182 = const()[name = tensor("op_4182"), val = tensor([1, 8, 64, 188])]; - tensor var_4183_cast_fp16 = reshape(shape = var_4182, x = value_21_cast_fp16)[name = tensor("op_4183_cast_fp16")]; - tensor attn_21_transpose_x_0 = const()[name = tensor("attn_21_transpose_x_0"), val = tensor(false)]; - tensor attn_21_transpose_y_0 = const()[name = tensor("attn_21_transpose_y_0"), val = tensor(true)]; - tensor attn_21_cast_fp16 = matmul(transpose_x = attn_21_transpose_x_0, transpose_y = attn_21_transpose_y_0, x = var_4183_cast_fp16, y = var_4181_cast_fp16)[name = tensor("attn_21_cast_fp16")]; - tensor var_4186 = const()[name = tensor("op_4186"), val = tensor([1, 512, 1, 188])]; - tensor input_283_cast_fp16 = reshape(shape = var_4186, x = attn_21_cast_fp16)[name = tensor("input_283_cast_fp16")]; - tensor var_4196_pad_type_0 = const()[name = tensor("op_4196_pad_type_0"), val = tensor("valid")]; - tensor var_4196_strides_0 = const()[name = tensor("op_4196_strides_0"), val = tensor([1, 1])]; - tensor var_4196_pad_0 = const()[name = tensor("op_4196_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4196_dilations_0 = const()[name = tensor("op_4196_dilations_0"), val = tensor([1, 1])]; - tensor var_4196_groups_0 = const()[name = tensor("op_4196_groups_0"), val = tensor(1)]; - tensor layers_10_self_attn_o_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(62426176))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(62622848))), name = tensor("layers_10_self_attn_o_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_10_self_attn_o_proj_inlier_module_bias_to_fp16 = const()[name = tensor("layers_10_self_attn_o_proj_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(62623040)))]; - tensor var_4196_cast_fp16 = conv(bias = layers_10_self_attn_o_proj_inlier_module_bias_to_fp16, dilations = var_4196_dilations_0, groups = var_4196_groups_0, pad = var_4196_pad_0, pad_type = var_4196_pad_type_0, strides = var_4196_strides_0, weight = layers_10_self_attn_o_proj_inlier_module_weight_to_fp16_palettized, x = input_283_cast_fp16)[name = tensor("op_4196_cast_fp16")]; - tensor var_4202_pad_type_0 = const()[name = tensor("op_4202_pad_type_0"), val = tensor("valid")]; - tensor var_4202_strides_0 = const()[name = tensor("op_4202_strides_0"), val = tensor([1, 1])]; - tensor var_4202_pad_0 = const()[name = tensor("op_4202_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4202_dilations_0 = const()[name = tensor("op_4202_dilations_0"), val = tensor([1, 1])]; - tensor var_4202_groups_0 = const()[name = tensor("op_4202_groups_0"), val = tensor(1)]; - tensor layers_10_self_attn_o_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(62632832))), name = tensor("layers_10_self_attn_o_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(62624128))), shape = tensor([512, 512, 1, 1])]; - tensor var_4202_cast_fp16 = conv(dilations = var_4202_dilations_0, groups = var_4202_groups_0, pad = var_4202_pad_0, pad_type = var_4202_pad_type_0, strides = var_4202_strides_0, weight = layers_10_self_attn_o_proj_outlier_module_weight_to_fp16_sparsified, x = input_283_cast_fp16)[name = tensor("op_4202_cast_fp16")]; - tensor obj_45_cast_fp16 = add(x = var_4196_cast_fp16, y = var_4202_cast_fp16)[name = tensor("obj_45_cast_fp16")]; - tensor inputs_105_cast_fp16 = add(x = inputs_103_cast_fp16, y = obj_45_cast_fp16)[name = tensor("inputs_105_cast_fp16")]; - tensor out_105_axes_0 = const()[name = tensor("out_105_axes_0"), val = tensor([1])]; - tensor var_4213_to_fp16 = const()[name = tensor("op_4213_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_105_cast_fp16 = layer_norm(axes = out_105_axes_0, epsilon = var_4213_to_fp16, x = inputs_105_cast_fp16)[name = tensor("out_105_cast_fp16")]; - tensor input_285_gamma_0_to_fp16 = const()[name = tensor("input_285_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(62665664)))]; - tensor input_285_beta_0_to_fp16 = const()[name = tensor("input_285_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(62666752)))]; - tensor input_285_epsilon_0_to_fp16 = const()[name = tensor("input_285_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_285_cast_fp16 = batch_norm(beta = input_285_beta_0_to_fp16, epsilon = input_285_epsilon_0_to_fp16, gamma = input_285_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_105_cast_fp16)[name = tensor("input_285_cast_fp16")]; - tensor var_4235_pad_type_0 = const()[name = tensor("op_4235_pad_type_0"), val = tensor("valid")]; - tensor var_4235_strides_0 = const()[name = tensor("op_4235_strides_0"), val = tensor([1, 1])]; - tensor var_4235_pad_0 = const()[name = tensor("op_4235_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4235_dilations_0 = const()[name = tensor("op_4235_dilations_0"), val = tensor([1, 1])]; - tensor var_4235_groups_0 = const()[name = tensor("op_4235_groups_0"), val = tensor(1)]; - tensor layers_10_conv_pointwise_conv1_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(62667840))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(63061120))), name = tensor("layers_10_conv_pointwise_conv1_inlier_module_weight_to_fp16_palettized"), shape = tensor([1024, 512, 1, 1])]; - tensor layers_10_conv_pointwise_conv1_inlier_module_bias_to_fp16 = const()[name = tensor("layers_10_conv_pointwise_conv1_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(63061312)))]; - tensor var_4235_cast_fp16 = conv(bias = layers_10_conv_pointwise_conv1_inlier_module_bias_to_fp16, dilations = var_4235_dilations_0, groups = var_4235_groups_0, pad = var_4235_pad_0, pad_type = var_4235_pad_type_0, strides = var_4235_strides_0, weight = layers_10_conv_pointwise_conv1_inlier_module_weight_to_fp16_palettized, x = input_285_cast_fp16)[name = tensor("op_4235_cast_fp16")]; - tensor var_4241_pad_type_0 = const()[name = tensor("op_4241_pad_type_0"), val = tensor("valid")]; - tensor var_4241_strides_0 = const()[name = tensor("op_4241_strides_0"), val = tensor([1, 1])]; - tensor var_4241_pad_0 = const()[name = tensor("op_4241_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4241_dilations_0 = const()[name = tensor("op_4241_dilations_0"), val = tensor([1, 1])]; - tensor var_4241_groups_0 = const()[name = tensor("op_4241_groups_0"), val = tensor(1)]; - tensor layers_10_conv_pointwise_conv1_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(63080960))), name = tensor("layers_10_conv_pointwise_conv1_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(63063424))), shape = tensor([1024, 512, 1, 1])]; - tensor var_4241_cast_fp16 = conv(dilations = var_4241_dilations_0, groups = var_4241_groups_0, pad = var_4241_pad_0, pad_type = var_4241_pad_type_0, strides = var_4241_strides_0, weight = layers_10_conv_pointwise_conv1_outlier_module_weight_to_fp16_sparsified, x = input_285_cast_fp16)[name = tensor("op_4241_cast_fp16")]; - tensor input_287_cast_fp16 = add(x = var_4235_cast_fp16, y = var_4241_cast_fp16)[name = tensor("input_287_cast_fp16")]; - tensor input_289_split_num_splits_0 = const()[name = tensor("input_289_split_num_splits_0"), val = tensor(2)]; - tensor input_289_split_axis_0 = const()[name = tensor("input_289_split_axis_0"), val = tensor(1)]; - tensor input_289_split_cast_fp16_0, tensor input_289_split_cast_fp16_1 = split(axis = input_289_split_axis_0, num_splits = input_289_split_num_splits_0, x = input_287_cast_fp16)[name = tensor("input_289_split_cast_fp16")]; - tensor input_289_split_1_sigmoid_cast_fp16 = sigmoid(x = input_289_split_cast_fp16_1)[name = tensor("input_289_split_1_sigmoid_cast_fp16")]; - tensor input_289_cast_fp16 = mul(x = input_289_split_cast_fp16_0, y = input_289_split_1_sigmoid_cast_fp16)[name = tensor("input_289_cast_fp16")]; - tensor input_291_pad_type_0 = const()[name = tensor("input_291_pad_type_0"), val = tensor("custom")]; - tensor input_291_pad_0 = const()[name = tensor("input_291_pad_0"), val = tensor([0, 0, 4, 4])]; - tensor input_291_groups_0 = const()[name = tensor("input_291_groups_0"), val = tensor(512)]; - tensor input_291_strides_0 = const()[name = tensor("input_291_strides_0"), val = tensor([1, 1])]; - tensor input_291_dilations_0 = const()[name = tensor("input_291_dilations_0"), val = tensor([1, 1])]; - tensor const_211_to_fp16 = const()[name = tensor("const_211_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(63146560)))]; - tensor const_212_to_fp16 = const()[name = tensor("const_212_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(63155840)))]; - tensor input_293_cast_fp16 = conv(bias = const_212_to_fp16, dilations = input_291_dilations_0, groups = input_291_groups_0, pad = input_291_pad_0, pad_type = input_291_pad_type_0, strides = input_291_strides_0, weight = const_211_to_fp16, x = input_289_cast_fp16)[name = tensor("input_293_cast_fp16")]; - tensor input_295_cast_fp16 = silu(x = input_293_cast_fp16)[name = tensor("input_295_cast_fp16")]; - tensor var_4265_pad_type_0 = const()[name = tensor("op_4265_pad_type_0"), val = tensor("valid")]; - tensor var_4265_strides_0 = const()[name = tensor("op_4265_strides_0"), val = tensor([1, 1])]; - tensor var_4265_pad_0 = const()[name = tensor("op_4265_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4265_dilations_0 = const()[name = tensor("op_4265_dilations_0"), val = tensor([1, 1])]; - tensor var_4265_groups_0 = const()[name = tensor("op_4265_groups_0"), val = tensor(1)]; - tensor layers_10_conv_pointwise_conv2_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(63156928))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(63353600))), name = tensor("layers_10_conv_pointwise_conv2_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_10_conv_pointwise_conv2_inlier_module_bias_to_fp16 = const()[name = tensor("layers_10_conv_pointwise_conv2_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(63353792)))]; - tensor var_4265_cast_fp16 = conv(bias = layers_10_conv_pointwise_conv2_inlier_module_bias_to_fp16, dilations = var_4265_dilations_0, groups = var_4265_groups_0, pad = var_4265_pad_0, pad_type = var_4265_pad_type_0, strides = var_4265_strides_0, weight = layers_10_conv_pointwise_conv2_inlier_module_weight_to_fp16_palettized, x = input_295_cast_fp16)[name = tensor("op_4265_cast_fp16")]; - tensor var_4271_pad_type_0 = const()[name = tensor("op_4271_pad_type_0"), val = tensor("valid")]; - tensor var_4271_strides_0 = const()[name = tensor("op_4271_strides_0"), val = tensor([1, 1])]; - tensor var_4271_pad_0 = const()[name = tensor("op_4271_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4271_dilations_0 = const()[name = tensor("op_4271_dilations_0"), val = tensor([1, 1])]; - tensor var_4271_groups_0 = const()[name = tensor("op_4271_groups_0"), val = tensor(1)]; - tensor layers_10_conv_pointwise_conv2_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(63363712))), name = tensor("layers_10_conv_pointwise_conv2_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(63354880))), shape = tensor([512, 512, 1, 1])]; - tensor var_4271_cast_fp16 = conv(dilations = var_4271_dilations_0, groups = var_4271_groups_0, pad = var_4271_pad_0, pad_type = var_4271_pad_type_0, strides = var_4271_strides_0, weight = layers_10_conv_pointwise_conv2_outlier_module_weight_to_fp16_sparsified, x = input_295_cast_fp16)[name = tensor("op_4271_cast_fp16")]; - tensor x_65_cast_fp16 = add(x = var_4265_cast_fp16, y = var_4271_cast_fp16)[name = tensor("x_65_cast_fp16")]; - tensor inputs_107_cast_fp16 = add(x = inputs_105_cast_fp16, y = x_65_cast_fp16)[name = tensor("inputs_107_cast_fp16")]; - tensor out_107_axes_0 = const()[name = tensor("out_107_axes_0"), val = tensor([1])]; - tensor var_4282_to_fp16 = const()[name = tensor("op_4282_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_107_cast_fp16 = layer_norm(axes = out_107_axes_0, epsilon = var_4282_to_fp16, x = inputs_107_cast_fp16)[name = tensor("out_107_cast_fp16")]; - tensor input_297_gamma_0_to_fp16 = const()[name = tensor("input_297_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(63396544)))]; - tensor input_297_beta_0_to_fp16 = const()[name = tensor("input_297_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(63397632)))]; - tensor input_297_epsilon_0_to_fp16 = const()[name = tensor("input_297_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_297_cast_fp16 = batch_norm(beta = input_297_beta_0_to_fp16, epsilon = input_297_epsilon_0_to_fp16, gamma = input_297_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_107_cast_fp16)[name = tensor("input_297_cast_fp16")]; - tensor var_4302_pad_type_0 = const()[name = tensor("op_4302_pad_type_0"), val = tensor("valid")]; - tensor var_4302_strides_0 = const()[name = tensor("op_4302_strides_0"), val = tensor([1, 1])]; - tensor var_4302_pad_0 = const()[name = tensor("op_4302_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4302_dilations_0 = const()[name = tensor("op_4302_dilations_0"), val = tensor([1, 1])]; - tensor var_4302_groups_0 = const()[name = tensor("op_4302_groups_0"), val = tensor(1)]; - tensor layers_10_feed_forward2_fc1_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(63398720))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64185216))), name = tensor("layers_10_feed_forward2_fc1_inlier_module_weight_to_fp16_palettized"), shape = tensor([2048, 512, 1, 1])]; - tensor layers_10_feed_forward2_fc1_inlier_module_bias_to_fp16 = const()[name = tensor("layers_10_feed_forward2_fc1_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64185408)))]; - tensor var_4302_cast_fp16 = conv(bias = layers_10_feed_forward2_fc1_inlier_module_bias_to_fp16, dilations = var_4302_dilations_0, groups = var_4302_groups_0, pad = var_4302_pad_0, pad_type = var_4302_pad_type_0, strides = var_4302_strides_0, weight = layers_10_feed_forward2_fc1_inlier_module_weight_to_fp16_palettized, x = input_297_cast_fp16)[name = tensor("op_4302_cast_fp16")]; - tensor var_4308_pad_type_0 = const()[name = tensor("op_4308_pad_type_0"), val = tensor("valid")]; - tensor var_4308_strides_0 = const()[name = tensor("op_4308_strides_0"), val = tensor([1, 1])]; - tensor var_4308_pad_0 = const()[name = tensor("op_4308_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4308_dilations_0 = const()[name = tensor("op_4308_dilations_0"), val = tensor([1, 1])]; - tensor var_4308_groups_0 = const()[name = tensor("op_4308_groups_0"), val = tensor(1)]; - tensor layers_10_feed_forward2_fc1_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64224256))), name = tensor("layers_10_feed_forward2_fc1_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64189568))), shape = tensor([2048, 512, 1, 1])]; - tensor var_4308_cast_fp16 = conv(dilations = var_4308_dilations_0, groups = var_4308_groups_0, pad = var_4308_pad_0, pad_type = var_4308_pad_type_0, strides = var_4308_strides_0, weight = layers_10_feed_forward2_fc1_outlier_module_weight_to_fp16_sparsified, x = input_297_cast_fp16)[name = tensor("op_4308_cast_fp16")]; - tensor input_299_cast_fp16 = add(x = var_4302_cast_fp16, y = var_4308_cast_fp16)[name = tensor("input_299_cast_fp16")]; - tensor input_301_cast_fp16 = silu(x = input_299_cast_fp16)[name = tensor("input_301_cast_fp16")]; - tensor var_4319_pad_type_0 = const()[name = tensor("op_4319_pad_type_0"), val = tensor("valid")]; - tensor var_4319_strides_0 = const()[name = tensor("op_4319_strides_0"), val = tensor([1, 1])]; - tensor var_4319_pad_0 = const()[name = tensor("op_4319_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4319_dilations_0 = const()[name = tensor("op_4319_dilations_0"), val = tensor([1, 1])]; - tensor var_4319_groups_0 = const()[name = tensor("op_4319_groups_0"), val = tensor(1)]; - tensor layers_10_feed_forward2_fc2_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64355392))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65141888))), name = tensor("layers_10_feed_forward2_fc2_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 2048, 1, 1])]; - tensor layers_10_feed_forward2_fc2_inlier_module_bias_to_fp16 = const()[name = tensor("layers_10_feed_forward2_fc2_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65142080)))]; - tensor var_4319_cast_fp16 = conv(bias = layers_10_feed_forward2_fc2_inlier_module_bias_to_fp16, dilations = var_4319_dilations_0, groups = var_4319_groups_0, pad = var_4319_pad_0, pad_type = var_4319_pad_type_0, strides = var_4319_strides_0, weight = layers_10_feed_forward2_fc2_inlier_module_weight_to_fp16_palettized, x = input_301_cast_fp16)[name = tensor("op_4319_cast_fp16")]; - tensor var_4325_pad_type_0 = const()[name = tensor("op_4325_pad_type_0"), val = tensor("valid")]; - tensor var_4325_strides_0 = const()[name = tensor("op_4325_strides_0"), val = tensor([1, 1])]; - tensor var_4325_pad_0 = const()[name = tensor("op_4325_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4325_dilations_0 = const()[name = tensor("op_4325_dilations_0"), val = tensor([1, 1])]; - tensor var_4325_groups_0 = const()[name = tensor("op_4325_groups_0"), val = tensor(1)]; - tensor layers_10_feed_forward2_fc2_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65185344))), name = tensor("layers_10_feed_forward2_fc2_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65143168))), shape = tensor([512, 2048, 1, 1])]; - tensor var_4325_cast_fp16 = conv(dilations = var_4325_dilations_0, groups = var_4325_groups_0, pad = var_4325_pad_0, pad_type = var_4325_pad_type_0, strides = var_4325_strides_0, weight = layers_10_feed_forward2_fc2_outlier_module_weight_to_fp16_sparsified, x = input_301_cast_fp16)[name = tensor("op_4325_cast_fp16")]; - tensor x_67_cast_fp16 = add(x = var_4319_cast_fp16, y = var_4325_cast_fp16)[name = tensor("x_67_cast_fp16")]; - tensor var_4327_to_fp16 = const()[name = tensor("op_4327_to_fp16"), val = tensor(0x1p-1)]; - tensor var_4328_cast_fp16 = mul(x = x_67_cast_fp16, y = var_4327_to_fp16)[name = tensor("op_4328_cast_fp16")]; - tensor inputs_109_cast_fp16 = add(x = inputs_107_cast_fp16, y = var_4328_cast_fp16)[name = tensor("inputs_109_cast_fp16")]; - tensor out_109_axes_0 = const()[name = tensor("out_109_axes_0"), val = tensor([1])]; - tensor var_4338_to_fp16 = const()[name = tensor("op_4338_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_109_cast_fp16 = layer_norm(axes = out_109_axes_0, epsilon = var_4338_to_fp16, x = inputs_109_cast_fp16)[name = tensor("out_109_cast_fp16")]; - tensor inputs_111_gamma_0_to_fp16 = const()[name = tensor("inputs_111_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65316480)))]; - tensor inputs_111_beta_0_to_fp16 = const()[name = tensor("inputs_111_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65317568)))]; - tensor inputs_111_epsilon_0_to_fp16 = const()[name = tensor("inputs_111_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor inputs_111_cast_fp16 = batch_norm(beta = inputs_111_beta_0_to_fp16, epsilon = inputs_111_epsilon_0_to_fp16, gamma = inputs_111_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_109_cast_fp16)[name = tensor("inputs_111_cast_fp16")]; - tensor var_4352 = const()[name = tensor("op_4352"), val = tensor(3)]; - tensor out_111_axes_0 = const()[name = tensor("out_111_axes_0"), val = tensor([1])]; - tensor var_4383_to_fp16 = const()[name = tensor("op_4383_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_111_cast_fp16 = layer_norm(axes = out_111_axes_0, epsilon = var_4383_to_fp16, x = inputs_111_cast_fp16)[name = tensor("out_111_cast_fp16")]; - tensor input_303_gamma_0_to_fp16 = const()[name = tensor("input_303_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65318656)))]; - tensor input_303_beta_0_to_fp16 = const()[name = tensor("input_303_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65319744)))]; - tensor input_303_epsilon_0_to_fp16 = const()[name = tensor("input_303_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_303_cast_fp16 = batch_norm(beta = input_303_beta_0_to_fp16, epsilon = input_303_epsilon_0_to_fp16, gamma = input_303_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_111_cast_fp16)[name = tensor("input_303_cast_fp16")]; - tensor var_4403_pad_type_0 = const()[name = tensor("op_4403_pad_type_0"), val = tensor("valid")]; - tensor var_4403_strides_0 = const()[name = tensor("op_4403_strides_0"), val = tensor([1, 1])]; - tensor var_4403_pad_0 = const()[name = tensor("op_4403_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4403_dilations_0 = const()[name = tensor("op_4403_dilations_0"), val = tensor([1, 1])]; - tensor var_4403_groups_0 = const()[name = tensor("op_4403_groups_0"), val = tensor(1)]; - tensor layers_11_feed_forward1_fc1_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65320832))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(66107328))), name = tensor("layers_11_feed_forward1_fc1_inlier_module_weight_to_fp16_palettized"), shape = tensor([2048, 512, 1, 1])]; - tensor layers_11_feed_forward1_fc1_inlier_module_bias_to_fp16 = const()[name = tensor("layers_11_feed_forward1_fc1_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(66107520)))]; - tensor var_4403_cast_fp16 = conv(bias = layers_11_feed_forward1_fc1_inlier_module_bias_to_fp16, dilations = var_4403_dilations_0, groups = var_4403_groups_0, pad = var_4403_pad_0, pad_type = var_4403_pad_type_0, strides = var_4403_strides_0, weight = layers_11_feed_forward1_fc1_inlier_module_weight_to_fp16_palettized, x = input_303_cast_fp16)[name = tensor("op_4403_cast_fp16")]; - tensor var_4409_pad_type_0 = const()[name = tensor("op_4409_pad_type_0"), val = tensor("valid")]; - tensor var_4409_strides_0 = const()[name = tensor("op_4409_strides_0"), val = tensor([1, 1])]; - tensor var_4409_pad_0 = const()[name = tensor("op_4409_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4409_dilations_0 = const()[name = tensor("op_4409_dilations_0"), val = tensor([1, 1])]; - tensor var_4409_groups_0 = const()[name = tensor("op_4409_groups_0"), val = tensor(1)]; - tensor layers_11_feed_forward1_fc1_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(66147264))), name = tensor("layers_11_feed_forward1_fc1_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(66111680))), shape = tensor([2048, 512, 1, 1])]; - tensor var_4409_cast_fp16 = conv(dilations = var_4409_dilations_0, groups = var_4409_groups_0, pad = var_4409_pad_0, pad_type = var_4409_pad_type_0, strides = var_4409_strides_0, weight = layers_11_feed_forward1_fc1_outlier_module_weight_to_fp16_sparsified, x = input_303_cast_fp16)[name = tensor("op_4409_cast_fp16")]; - tensor input_305_cast_fp16 = add(x = var_4403_cast_fp16, y = var_4409_cast_fp16)[name = tensor("input_305_cast_fp16")]; - tensor input_307_cast_fp16 = silu(x = input_305_cast_fp16)[name = tensor("input_307_cast_fp16")]; - tensor var_4420_pad_type_0 = const()[name = tensor("op_4420_pad_type_0"), val = tensor("valid")]; - tensor var_4420_strides_0 = const()[name = tensor("op_4420_strides_0"), val = tensor([1, 1])]; - tensor var_4420_pad_0 = const()[name = tensor("op_4420_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4420_dilations_0 = const()[name = tensor("op_4420_dilations_0"), val = tensor([1, 1])]; - tensor var_4420_groups_0 = const()[name = tensor("op_4420_groups_0"), val = tensor(1)]; - tensor layers_11_feed_forward1_fc2_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(66278400))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67064896))), name = tensor("layers_11_feed_forward1_fc2_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 2048, 1, 1])]; - tensor layers_11_feed_forward1_fc2_inlier_module_bias_to_fp16 = const()[name = tensor("layers_11_feed_forward1_fc2_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67065088)))]; - tensor var_4420_cast_fp16 = conv(bias = layers_11_feed_forward1_fc2_inlier_module_bias_to_fp16, dilations = var_4420_dilations_0, groups = var_4420_groups_0, pad = var_4420_pad_0, pad_type = var_4420_pad_type_0, strides = var_4420_strides_0, weight = layers_11_feed_forward1_fc2_inlier_module_weight_to_fp16_palettized, x = input_307_cast_fp16)[name = tensor("op_4420_cast_fp16")]; - tensor var_4426_pad_type_0 = const()[name = tensor("op_4426_pad_type_0"), val = tensor("valid")]; - tensor var_4426_strides_0 = const()[name = tensor("op_4426_strides_0"), val = tensor([1, 1])]; - tensor var_4426_pad_0 = const()[name = tensor("op_4426_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4426_dilations_0 = const()[name = tensor("op_4426_dilations_0"), val = tensor([1, 1])]; - tensor var_4426_groups_0 = const()[name = tensor("op_4426_groups_0"), val = tensor(1)]; - tensor layers_11_feed_forward1_fc2_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67105920))), name = tensor("layers_11_feed_forward1_fc2_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67066176))), shape = tensor([512, 2048, 1, 1])]; - tensor var_4426_cast_fp16 = conv(dilations = var_4426_dilations_0, groups = var_4426_groups_0, pad = var_4426_pad_0, pad_type = var_4426_pad_type_0, strides = var_4426_strides_0, weight = layers_11_feed_forward1_fc2_outlier_module_weight_to_fp16_sparsified, x = input_307_cast_fp16)[name = tensor("op_4426_cast_fp16")]; - tensor x_69_cast_fp16 = add(x = var_4420_cast_fp16, y = var_4426_cast_fp16)[name = tensor("x_69_cast_fp16")]; - tensor var_4428_to_fp16 = const()[name = tensor("op_4428_to_fp16"), val = tensor(0x1p-1)]; - tensor var_4429_cast_fp16 = mul(x = x_69_cast_fp16, y = var_4428_to_fp16)[name = tensor("op_4429_cast_fp16")]; - tensor inputs_113_cast_fp16 = add(x = inputs_111_cast_fp16, y = var_4429_cast_fp16)[name = tensor("inputs_113_cast_fp16")]; - tensor out_113_axes_0 = const()[name = tensor("out_113_axes_0"), val = tensor([1])]; - tensor var_4439_to_fp16 = const()[name = tensor("op_4439_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_113_cast_fp16 = layer_norm(axes = out_113_axes_0, epsilon = var_4439_to_fp16, x = inputs_113_cast_fp16)[name = tensor("out_113_cast_fp16")]; - tensor obj_47_gamma_0_to_fp16 = const()[name = tensor("obj_47_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67237056)))]; - tensor obj_47_beta_0_to_fp16 = const()[name = tensor("obj_47_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67238144)))]; - tensor obj_47_epsilon_0_to_fp16 = const()[name = tensor("obj_47_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor obj_47_cast_fp16 = batch_norm(beta = obj_47_beta_0_to_fp16, epsilon = obj_47_epsilon_0_to_fp16, gamma = obj_47_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_113_cast_fp16)[name = tensor("obj_47_cast_fp16")]; - tensor var_4464_pad_type_0 = const()[name = tensor("op_4464_pad_type_0"), val = tensor("valid")]; - tensor var_4464_strides_0 = const()[name = tensor("op_4464_strides_0"), val = tensor([1, 1])]; - tensor var_4464_pad_0 = const()[name = tensor("op_4464_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4464_dilations_0 = const()[name = tensor("op_4464_dilations_0"), val = tensor([1, 1])]; - tensor var_4464_groups_0 = const()[name = tensor("op_4464_groups_0"), val = tensor(1)]; - tensor layers_11_self_attn_q_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67239232))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67435904))), name = tensor("layers_11_self_attn_q_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_11_self_attn_q_proj_inlier_module_bias_to_fp16 = const()[name = tensor("layers_11_self_attn_q_proj_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67436096)))]; - tensor var_4464_cast_fp16 = conv(bias = layers_11_self_attn_q_proj_inlier_module_bias_to_fp16, dilations = var_4464_dilations_0, groups = var_4464_groups_0, pad = var_4464_pad_0, pad_type = var_4464_pad_type_0, strides = var_4464_strides_0, weight = layers_11_self_attn_q_proj_inlier_module_weight_to_fp16_palettized, x = obj_47_cast_fp16)[name = tensor("op_4464_cast_fp16")]; - tensor var_4470_pad_type_0 = const()[name = tensor("op_4470_pad_type_0"), val = tensor("valid")]; - tensor var_4470_strides_0 = const()[name = tensor("op_4470_strides_0"), val = tensor([1, 1])]; - tensor var_4470_pad_0 = const()[name = tensor("op_4470_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4470_dilations_0 = const()[name = tensor("op_4470_dilations_0"), val = tensor([1, 1])]; - tensor var_4470_groups_0 = const()[name = tensor("op_4470_groups_0"), val = tensor(1)]; - tensor layers_11_self_attn_q_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67446528))), name = tensor("layers_11_self_attn_q_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67437184))), shape = tensor([512, 512, 1, 1])]; - tensor var_4470_cast_fp16 = conv(dilations = var_4470_dilations_0, groups = var_4470_groups_0, pad = var_4470_pad_0, pad_type = var_4470_pad_type_0, strides = var_4470_strides_0, weight = layers_11_self_attn_q_proj_outlier_module_weight_to_fp16_sparsified, x = obj_47_cast_fp16)[name = tensor("op_4470_cast_fp16")]; - tensor query_45_cast_fp16 = add(x = var_4464_cast_fp16, y = var_4470_cast_fp16)[name = tensor("query_45_cast_fp16")]; - tensor var_4479_pad_type_0 = const()[name = tensor("op_4479_pad_type_0"), val = tensor("valid")]; - tensor var_4479_strides_0 = const()[name = tensor("op_4479_strides_0"), val = tensor([1, 1])]; - tensor var_4479_pad_0 = const()[name = tensor("op_4479_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4479_dilations_0 = const()[name = tensor("op_4479_dilations_0"), val = tensor([1, 1])]; - tensor var_4479_groups_0 = const()[name = tensor("op_4479_groups_0"), val = tensor(1)]; - tensor layers_11_self_attn_k_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67479360))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67676032))), name = tensor("layers_11_self_attn_k_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor var_4479_cast_fp16 = conv(dilations = var_4479_dilations_0, groups = var_4479_groups_0, pad = var_4479_pad_0, pad_type = var_4479_pad_type_0, strides = var_4479_strides_0, weight = layers_11_self_attn_k_proj_inlier_module_weight_to_fp16_palettized, x = obj_47_cast_fp16)[name = tensor("op_4479_cast_fp16")]; - tensor var_4485_pad_type_0 = const()[name = tensor("op_4485_pad_type_0"), val = tensor("valid")]; - tensor var_4485_strides_0 = const()[name = tensor("op_4485_strides_0"), val = tensor([1, 1])]; - tensor var_4485_pad_0 = const()[name = tensor("op_4485_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4485_dilations_0 = const()[name = tensor("op_4485_dilations_0"), val = tensor([1, 1])]; - tensor var_4485_groups_0 = const()[name = tensor("op_4485_groups_0"), val = tensor(1)]; - tensor layers_11_self_attn_k_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67685952))), name = tensor("layers_11_self_attn_k_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67676224))), shape = tensor([512, 512, 1, 1])]; - tensor var_4485_cast_fp16 = conv(dilations = var_4485_dilations_0, groups = var_4485_groups_0, pad = var_4485_pad_0, pad_type = var_4485_pad_type_0, strides = var_4485_strides_0, weight = layers_11_self_attn_k_proj_outlier_module_weight_to_fp16_sparsified, x = obj_47_cast_fp16)[name = tensor("op_4485_cast_fp16")]; - tensor key_23_cast_fp16 = add(x = var_4479_cast_fp16, y = var_4485_cast_fp16)[name = tensor("key_23_cast_fp16")]; - tensor var_4495_pad_type_0 = const()[name = tensor("op_4495_pad_type_0"), val = tensor("valid")]; - tensor var_4495_strides_0 = const()[name = tensor("op_4495_strides_0"), val = tensor([1, 1])]; - tensor var_4495_pad_0 = const()[name = tensor("op_4495_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4495_dilations_0 = const()[name = tensor("op_4495_dilations_0"), val = tensor([1, 1])]; - tensor var_4495_groups_0 = const()[name = tensor("op_4495_groups_0"), val = tensor(1)]; - tensor layers_11_self_attn_v_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67718784))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67915456))), name = tensor("layers_11_self_attn_v_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_11_self_attn_v_proj_inlier_module_bias_to_fp16 = const()[name = tensor("layers_11_self_attn_v_proj_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67915648)))]; - tensor var_4495_cast_fp16 = conv(bias = layers_11_self_attn_v_proj_inlier_module_bias_to_fp16, dilations = var_4495_dilations_0, groups = var_4495_groups_0, pad = var_4495_pad_0, pad_type = var_4495_pad_type_0, strides = var_4495_strides_0, weight = layers_11_self_attn_v_proj_inlier_module_weight_to_fp16_palettized, x = obj_47_cast_fp16)[name = tensor("op_4495_cast_fp16")]; - tensor var_4501_pad_type_0 = const()[name = tensor("op_4501_pad_type_0"), val = tensor("valid")]; - tensor var_4501_strides_0 = const()[name = tensor("op_4501_strides_0"), val = tensor([1, 1])]; - tensor var_4501_pad_0 = const()[name = tensor("op_4501_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4501_dilations_0 = const()[name = tensor("op_4501_dilations_0"), val = tensor([1, 1])]; - tensor var_4501_groups_0 = const()[name = tensor("op_4501_groups_0"), val = tensor(1)]; - tensor layers_11_self_attn_v_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67925056))), name = tensor("layers_11_self_attn_v_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67916736))), shape = tensor([512, 512, 1, 1])]; - tensor var_4501_cast_fp16 = conv(dilations = var_4501_dilations_0, groups = var_4501_groups_0, pad = var_4501_pad_0, pad_type = var_4501_pad_type_0, strides = var_4501_strides_0, weight = layers_11_self_attn_v_proj_outlier_module_weight_to_fp16_sparsified, x = obj_47_cast_fp16)[name = tensor("op_4501_cast_fp16")]; - tensor value_23_cast_fp16 = add(x = var_4495_cast_fp16, y = var_4501_cast_fp16)[name = tensor("value_23_cast_fp16")]; - tensor var_4504_to_fp16 = const()[name = tensor("op_4504_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67957888)))]; - tensor query_47_cast_fp16 = add(x = query_45_cast_fp16, y = var_4504_to_fp16)[name = tensor("query_47_cast_fp16")]; - tensor var_4507_to_fp16 = const()[name = tensor("op_4507_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67958976)))]; - tensor q_with_bias_v_23_cast_fp16 = add(x = query_45_cast_fp16, y = var_4507_to_fp16)[name = tensor("q_with_bias_v_23_cast_fp16")]; - tensor var_4517_pad_type_0 = const()[name = tensor("op_4517_pad_type_0"), val = tensor("valid")]; - tensor var_4517_strides_0 = const()[name = tensor("op_4517_strides_0"), val = tensor([1, 1])]; - tensor var_4517_pad_0 = const()[name = tensor("op_4517_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4517_dilations_0 = const()[name = tensor("op_4517_dilations_0"), val = tensor([1, 1])]; - tensor var_4517_groups_0 = const()[name = tensor("op_4517_groups_0"), val = tensor(1)]; - tensor layers_11_self_attn_linear_pos_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67960064))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(68156736))), name = tensor("layers_11_self_attn_linear_pos_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor var_4517_cast_fp16 = conv(bias = input_17_mean_0_to_fp16, dilations = var_4517_dilations_0, groups = var_4517_groups_0, pad = var_4517_pad_0, pad_type = var_4517_pad_type_0, strides = var_4517_strides_0, weight = layers_11_self_attn_linear_pos_inlier_module_weight_to_fp16_palettized, x = obj_3_cast_fp16)[name = tensor("op_4517_cast_fp16")]; - tensor var_4523_pad_type_0 = const()[name = tensor("op_4523_pad_type_0"), val = tensor("valid")]; - tensor var_4523_strides_0 = const()[name = tensor("op_4523_strides_0"), val = tensor([1, 1])]; - tensor var_4523_pad_0 = const()[name = tensor("op_4523_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4523_dilations_0 = const()[name = tensor("op_4523_dilations_0"), val = tensor([1, 1])]; - tensor var_4523_groups_0 = const()[name = tensor("op_4523_groups_0"), val = tensor(1)]; - tensor layers_11_self_attn_linear_pos_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(68178304))), name = tensor("layers_11_self_attn_linear_pos_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(68156928))), shape = tensor([512, 512, 1, 1])]; - tensor var_4523_cast_fp16 = conv(dilations = var_4523_dilations_0, groups = var_4523_groups_0, pad = var_4523_pad_0, pad_type = var_4523_pad_type_0, strides = var_4523_strides_0, weight = layers_11_self_attn_linear_pos_outlier_module_weight_to_fp16_sparsified, x = obj_3_cast_fp16)[name = tensor("op_4523_cast_fp16")]; - tensor p_23_cast_fp16 = add(x = var_4517_cast_fp16, y = var_4523_cast_fp16)[name = tensor("p_23_cast_fp16")]; - tensor var_4527 = const()[name = tensor("op_4527"), val = tensor([1, 8, 64, 188])]; - tensor var_4528_cast_fp16 = reshape(shape = var_4527, x = q_with_bias_v_23_cast_fp16)[name = tensor("op_4528_cast_fp16")]; - tensor var_4529 = const()[name = tensor("op_4529"), val = tensor([1, 8, 64, -1])]; - tensor var_4530_cast_fp16 = reshape(shape = var_4529, x = p_23_cast_fp16)[name = tensor("op_4530_cast_fp16")]; - tensor matrix_bd_89_transpose_x_0 = const()[name = tensor("matrix_bd_89_transpose_x_0"), val = tensor(true)]; - tensor matrix_bd_89_transpose_y_0 = const()[name = tensor("matrix_bd_89_transpose_y_0"), val = tensor(false)]; - tensor matrix_bd_89_cast_fp16 = matmul(transpose_x = matrix_bd_89_transpose_x_0, transpose_y = matrix_bd_89_transpose_y_0, x = var_4528_cast_fp16, y = var_4530_cast_fp16)[name = tensor("matrix_bd_89_cast_fp16")]; - tensor matrix_bd_91_pad_0 = const()[name = tensor("matrix_bd_91_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor matrix_bd_91_mode_0 = const()[name = tensor("matrix_bd_91_mode_0"), val = tensor("constant")]; - tensor const_131_to_fp16 = const()[name = tensor("const_131_to_fp16"), val = tensor(0x0p+0)]; - tensor matrix_bd_91_cast_fp16 = pad(constant_val = const_131_to_fp16, mode = matrix_bd_91_mode_0, pad = matrix_bd_91_pad_0, x = matrix_bd_89_cast_fp16)[name = tensor("matrix_bd_91_cast_fp16")]; - tensor var_4539 = const()[name = tensor("op_4539"), val = tensor([1, 8, -1, 188])]; - tensor matrix_bd_93_cast_fp16 = reshape(shape = var_4539, x = matrix_bd_91_cast_fp16)[name = tensor("matrix_bd_93_cast_fp16")]; - tensor var_4543_begin_0 = const()[name = tensor("op_4543_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_4543_end_0 = const()[name = tensor("op_4543_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_4543_end_mask_0 = const()[name = tensor("op_4543_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_4543_cast_fp16 = slice_by_index(begin = var_4543_begin_0, end = var_4543_end_0, end_mask = var_4543_end_mask_0, x = matrix_bd_93_cast_fp16)[name = tensor("op_4543_cast_fp16")]; - tensor var_4544 = const()[name = tensor("op_4544"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_95_cast_fp16 = reshape(shape = var_4544, x = var_4543_cast_fp16)[name = tensor("matrix_bd_95_cast_fp16")]; - tensor var_4549_begin_0 = const()[name = tensor("op_4549_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4549_end_0 = const()[name = tensor("op_4549_end_0"), val = tensor([1, 8, 188, 188])]; - tensor var_4549_end_mask_0 = const()[name = tensor("op_4549_end_mask_0"), val = tensor([true, true, true, false])]; - tensor var_4549_cast_fp16 = slice_by_index(begin = var_4549_begin_0, end = var_4549_end_0, end_mask = var_4549_end_mask_0, x = matrix_bd_95_cast_fp16)[name = tensor("op_4549_cast_fp16")]; - tensor var_4550_to_fp16 = const()[name = tensor("op_4550_to_fp16"), val = tensor(0x1p-3)]; - tensor qk_mask_23_cast_fp16 = mul(x = var_4549_cast_fp16, y = var_4550_to_fp16)[name = tensor("qk_mask_23_cast_fp16")]; - tensor var_4554 = const()[name = tensor("op_4554"), val = tensor([1, 8, 64, 188])]; - tensor mh_q_23_cast_fp16 = reshape(shape = var_4554, x = query_47_cast_fp16)[name = tensor("mh_q_23_cast_fp16")]; - tensor var_4556_to_fp16 = const()[name = tensor("op_4556_to_fp16"), val = tensor(0x1p-3)]; - tensor var_4557_cast_fp16 = mul(x = mh_q_23_cast_fp16, y = var_4556_to_fp16)[name = tensor("op_4557_cast_fp16")]; - tensor var_4560 = const()[name = tensor("op_4560"), val = tensor([1, 8, 64, 188])]; - tensor var_4561_cast_fp16 = reshape(shape = var_4560, x = key_23_cast_fp16)[name = tensor("op_4561_cast_fp16")]; - tensor mh_w_45_transpose_x_0 = const()[name = tensor("mh_w_45_transpose_x_0"), val = tensor(true)]; - tensor mh_w_45_transpose_y_0 = const()[name = tensor("mh_w_45_transpose_y_0"), val = tensor(false)]; - tensor mh_w_45_cast_fp16 = matmul(transpose_x = mh_w_45_transpose_x_0, transpose_y = mh_w_45_transpose_y_0, x = var_4557_cast_fp16, y = var_4561_cast_fp16)[name = tensor("mh_w_45_cast_fp16")]; - tensor mh_w_47_cast_fp16 = add(x = mh_w_45_cast_fp16, y = qk_mask_23_cast_fp16)[name = tensor("mh_w_47_cast_fp16")]; - tensor var_4565_cast_fp16 = softmax(axis = var_4352, x = mh_w_47_cast_fp16)[name = tensor("op_4565_cast_fp16")]; - tensor var_4566 = const()[name = tensor("op_4566"), val = tensor([1, 8, 64, 188])]; - tensor var_4567_cast_fp16 = reshape(shape = var_4566, x = value_23_cast_fp16)[name = tensor("op_4567_cast_fp16")]; - tensor attn_23_transpose_x_0 = const()[name = tensor("attn_23_transpose_x_0"), val = tensor(false)]; - tensor attn_23_transpose_y_0 = const()[name = tensor("attn_23_transpose_y_0"), val = tensor(true)]; - tensor attn_23_cast_fp16 = matmul(transpose_x = attn_23_transpose_x_0, transpose_y = attn_23_transpose_y_0, x = var_4567_cast_fp16, y = var_4565_cast_fp16)[name = tensor("attn_23_cast_fp16")]; - tensor var_4570 = const()[name = tensor("op_4570"), val = tensor([1, 512, 1, 188])]; - tensor input_309_cast_fp16 = reshape(shape = var_4570, x = attn_23_cast_fp16)[name = tensor("input_309_cast_fp16")]; - tensor var_4580_pad_type_0 = const()[name = tensor("op_4580_pad_type_0"), val = tensor("valid")]; - tensor var_4580_strides_0 = const()[name = tensor("op_4580_strides_0"), val = tensor([1, 1])]; - tensor var_4580_pad_0 = const()[name = tensor("op_4580_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4580_dilations_0 = const()[name = tensor("op_4580_dilations_0"), val = tensor([1, 1])]; - tensor var_4580_groups_0 = const()[name = tensor("op_4580_groups_0"), val = tensor(1)]; - tensor layers_11_self_attn_o_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(68211136))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(68407808))), name = tensor("layers_11_self_attn_o_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_11_self_attn_o_proj_inlier_module_bias_to_fp16 = const()[name = tensor("layers_11_self_attn_o_proj_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(68408000)))]; - tensor var_4580_cast_fp16 = conv(bias = layers_11_self_attn_o_proj_inlier_module_bias_to_fp16, dilations = var_4580_dilations_0, groups = var_4580_groups_0, pad = var_4580_pad_0, pad_type = var_4580_pad_type_0, strides = var_4580_strides_0, weight = layers_11_self_attn_o_proj_inlier_module_weight_to_fp16_palettized, x = input_309_cast_fp16)[name = tensor("op_4580_cast_fp16")]; - tensor var_4586_pad_type_0 = const()[name = tensor("op_4586_pad_type_0"), val = tensor("valid")]; - tensor var_4586_strides_0 = const()[name = tensor("op_4586_strides_0"), val = tensor([1, 1])]; - tensor var_4586_pad_0 = const()[name = tensor("op_4586_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4586_dilations_0 = const()[name = tensor("op_4586_dilations_0"), val = tensor([1, 1])]; - tensor var_4586_groups_0 = const()[name = tensor("op_4586_groups_0"), val = tensor(1)]; - tensor layers_11_self_attn_o_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(68418048))), name = tensor("layers_11_self_attn_o_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(68409088))), shape = tensor([512, 512, 1, 1])]; - tensor var_4586_cast_fp16 = conv(dilations = var_4586_dilations_0, groups = var_4586_groups_0, pad = var_4586_pad_0, pad_type = var_4586_pad_type_0, strides = var_4586_strides_0, weight = layers_11_self_attn_o_proj_outlier_module_weight_to_fp16_sparsified, x = input_309_cast_fp16)[name = tensor("op_4586_cast_fp16")]; - tensor obj_49_cast_fp16 = add(x = var_4580_cast_fp16, y = var_4586_cast_fp16)[name = tensor("obj_49_cast_fp16")]; - tensor inputs_115_cast_fp16 = add(x = inputs_113_cast_fp16, y = obj_49_cast_fp16)[name = tensor("inputs_115_cast_fp16")]; - tensor out_115_axes_0 = const()[name = tensor("out_115_axes_0"), val = tensor([1])]; - tensor var_4597_to_fp16 = const()[name = tensor("op_4597_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_115_cast_fp16 = layer_norm(axes = out_115_axes_0, epsilon = var_4597_to_fp16, x = inputs_115_cast_fp16)[name = tensor("out_115_cast_fp16")]; - tensor input_311_gamma_0_to_fp16 = const()[name = tensor("input_311_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(68450880)))]; - tensor input_311_beta_0_to_fp16 = const()[name = tensor("input_311_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(68451968)))]; - tensor input_311_epsilon_0_to_fp16 = const()[name = tensor("input_311_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_311_cast_fp16 = batch_norm(beta = input_311_beta_0_to_fp16, epsilon = input_311_epsilon_0_to_fp16, gamma = input_311_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_115_cast_fp16)[name = tensor("input_311_cast_fp16")]; - tensor var_4619_pad_type_0 = const()[name = tensor("op_4619_pad_type_0"), val = tensor("valid")]; - tensor var_4619_strides_0 = const()[name = tensor("op_4619_strides_0"), val = tensor([1, 1])]; - tensor var_4619_pad_0 = const()[name = tensor("op_4619_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4619_dilations_0 = const()[name = tensor("op_4619_dilations_0"), val = tensor([1, 1])]; - tensor var_4619_groups_0 = const()[name = tensor("op_4619_groups_0"), val = tensor(1)]; - tensor layers_11_conv_pointwise_conv1_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(68453056))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(68846336))), name = tensor("layers_11_conv_pointwise_conv1_inlier_module_weight_to_fp16_palettized"), shape = tensor([1024, 512, 1, 1])]; - tensor layers_11_conv_pointwise_conv1_inlier_module_bias_to_fp16 = const()[name = tensor("layers_11_conv_pointwise_conv1_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(68846528)))]; - tensor var_4619_cast_fp16 = conv(bias = layers_11_conv_pointwise_conv1_inlier_module_bias_to_fp16, dilations = var_4619_dilations_0, groups = var_4619_groups_0, pad = var_4619_pad_0, pad_type = var_4619_pad_type_0, strides = var_4619_strides_0, weight = layers_11_conv_pointwise_conv1_inlier_module_weight_to_fp16_palettized, x = input_311_cast_fp16)[name = tensor("op_4619_cast_fp16")]; - tensor var_4625_pad_type_0 = const()[name = tensor("op_4625_pad_type_0"), val = tensor("valid")]; - tensor var_4625_strides_0 = const()[name = tensor("op_4625_strides_0"), val = tensor([1, 1])]; - tensor var_4625_pad_0 = const()[name = tensor("op_4625_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4625_dilations_0 = const()[name = tensor("op_4625_dilations_0"), val = tensor([1, 1])]; - tensor var_4625_groups_0 = const()[name = tensor("op_4625_groups_0"), val = tensor(1)]; - tensor layers_11_conv_pointwise_conv1_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(68866112))), name = tensor("layers_11_conv_pointwise_conv1_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(68848640))), shape = tensor([1024, 512, 1, 1])]; - tensor var_4625_cast_fp16 = conv(dilations = var_4625_dilations_0, groups = var_4625_groups_0, pad = var_4625_pad_0, pad_type = var_4625_pad_type_0, strides = var_4625_strides_0, weight = layers_11_conv_pointwise_conv1_outlier_module_weight_to_fp16_sparsified, x = input_311_cast_fp16)[name = tensor("op_4625_cast_fp16")]; - tensor input_313_cast_fp16 = add(x = var_4619_cast_fp16, y = var_4625_cast_fp16)[name = tensor("input_313_cast_fp16")]; - tensor input_315_split_num_splits_0 = const()[name = tensor("input_315_split_num_splits_0"), val = tensor(2)]; - tensor input_315_split_axis_0 = const()[name = tensor("input_315_split_axis_0"), val = tensor(1)]; - tensor input_315_split_cast_fp16_0, tensor input_315_split_cast_fp16_1 = split(axis = input_315_split_axis_0, num_splits = input_315_split_num_splits_0, x = input_313_cast_fp16)[name = tensor("input_315_split_cast_fp16")]; - tensor input_315_split_1_sigmoid_cast_fp16 = sigmoid(x = input_315_split_cast_fp16_1)[name = tensor("input_315_split_1_sigmoid_cast_fp16")]; - tensor input_315_cast_fp16 = mul(x = input_315_split_cast_fp16_0, y = input_315_split_1_sigmoid_cast_fp16)[name = tensor("input_315_cast_fp16")]; - tensor input_317_pad_type_0 = const()[name = tensor("input_317_pad_type_0"), val = tensor("custom")]; - tensor input_317_pad_0 = const()[name = tensor("input_317_pad_0"), val = tensor([0, 0, 4, 4])]; - tensor input_317_groups_0 = const()[name = tensor("input_317_groups_0"), val = tensor(512)]; - tensor input_317_strides_0 = const()[name = tensor("input_317_strides_0"), val = tensor([1, 1])]; - tensor input_317_dilations_0 = const()[name = tensor("input_317_dilations_0"), val = tensor([1, 1])]; - tensor const_213_to_fp16 = const()[name = tensor("const_213_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(68931712)))]; - tensor const_214_to_fp16 = const()[name = tensor("const_214_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(68940992)))]; - tensor input_319_cast_fp16 = conv(bias = const_214_to_fp16, dilations = input_317_dilations_0, groups = input_317_groups_0, pad = input_317_pad_0, pad_type = input_317_pad_type_0, strides = input_317_strides_0, weight = const_213_to_fp16, x = input_315_cast_fp16)[name = tensor("input_319_cast_fp16")]; - tensor input_321_cast_fp16 = silu(x = input_319_cast_fp16)[name = tensor("input_321_cast_fp16")]; - tensor var_4649_pad_type_0 = const()[name = tensor("op_4649_pad_type_0"), val = tensor("valid")]; - tensor var_4649_strides_0 = const()[name = tensor("op_4649_strides_0"), val = tensor([1, 1])]; - tensor var_4649_pad_0 = const()[name = tensor("op_4649_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4649_dilations_0 = const()[name = tensor("op_4649_dilations_0"), val = tensor([1, 1])]; - tensor var_4649_groups_0 = const()[name = tensor("op_4649_groups_0"), val = tensor(1)]; - tensor layers_11_conv_pointwise_conv2_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(68942080))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(69138752))), name = tensor("layers_11_conv_pointwise_conv2_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_11_conv_pointwise_conv2_inlier_module_bias_to_fp16 = const()[name = tensor("layers_11_conv_pointwise_conv2_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(69138944)))]; - tensor var_4649_cast_fp16 = conv(bias = layers_11_conv_pointwise_conv2_inlier_module_bias_to_fp16, dilations = var_4649_dilations_0, groups = var_4649_groups_0, pad = var_4649_pad_0, pad_type = var_4649_pad_type_0, strides = var_4649_strides_0, weight = layers_11_conv_pointwise_conv2_inlier_module_weight_to_fp16_palettized, x = input_321_cast_fp16)[name = tensor("op_4649_cast_fp16")]; - tensor var_4655_pad_type_0 = const()[name = tensor("op_4655_pad_type_0"), val = tensor("valid")]; - tensor var_4655_strides_0 = const()[name = tensor("op_4655_strides_0"), val = tensor([1, 1])]; - tensor var_4655_pad_0 = const()[name = tensor("op_4655_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4655_dilations_0 = const()[name = tensor("op_4655_dilations_0"), val = tensor([1, 1])]; - tensor var_4655_groups_0 = const()[name = tensor("op_4655_groups_0"), val = tensor(1)]; - tensor layers_11_conv_pointwise_conv2_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(69149120))), name = tensor("layers_11_conv_pointwise_conv2_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(69140032))), shape = tensor([512, 512, 1, 1])]; - tensor var_4655_cast_fp16 = conv(dilations = var_4655_dilations_0, groups = var_4655_groups_0, pad = var_4655_pad_0, pad_type = var_4655_pad_type_0, strides = var_4655_strides_0, weight = layers_11_conv_pointwise_conv2_outlier_module_weight_to_fp16_sparsified, x = input_321_cast_fp16)[name = tensor("op_4655_cast_fp16")]; - tensor x_71_cast_fp16 = add(x = var_4649_cast_fp16, y = var_4655_cast_fp16)[name = tensor("x_71_cast_fp16")]; - tensor inputs_117_cast_fp16 = add(x = inputs_115_cast_fp16, y = x_71_cast_fp16)[name = tensor("inputs_117_cast_fp16")]; - tensor out_117_axes_0 = const()[name = tensor("out_117_axes_0"), val = tensor([1])]; - tensor var_4666_to_fp16 = const()[name = tensor("op_4666_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_117_cast_fp16 = layer_norm(axes = out_117_axes_0, epsilon = var_4666_to_fp16, x = inputs_117_cast_fp16)[name = tensor("out_117_cast_fp16")]; - tensor input_323_gamma_0_to_fp16 = const()[name = tensor("input_323_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(69181952)))]; - tensor input_323_beta_0_to_fp16 = const()[name = tensor("input_323_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(69183040)))]; - tensor input_323_epsilon_0_to_fp16 = const()[name = tensor("input_323_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_323_cast_fp16 = batch_norm(beta = input_323_beta_0_to_fp16, epsilon = input_323_epsilon_0_to_fp16, gamma = input_323_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_117_cast_fp16)[name = tensor("input_323_cast_fp16")]; - tensor var_4686_pad_type_0 = const()[name = tensor("op_4686_pad_type_0"), val = tensor("valid")]; - tensor var_4686_strides_0 = const()[name = tensor("op_4686_strides_0"), val = tensor([1, 1])]; - tensor var_4686_pad_0 = const()[name = tensor("op_4686_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4686_dilations_0 = const()[name = tensor("op_4686_dilations_0"), val = tensor([1, 1])]; - tensor var_4686_groups_0 = const()[name = tensor("op_4686_groups_0"), val = tensor(1)]; - tensor layers_11_feed_forward2_fc1_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(69184128))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(69970624))), name = tensor("layers_11_feed_forward2_fc1_inlier_module_weight_to_fp16_palettized"), shape = tensor([2048, 512, 1, 1])]; - tensor layers_11_feed_forward2_fc1_inlier_module_bias_to_fp16 = const()[name = tensor("layers_11_feed_forward2_fc1_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(69970816)))]; - tensor var_4686_cast_fp16 = conv(bias = layers_11_feed_forward2_fc1_inlier_module_bias_to_fp16, dilations = var_4686_dilations_0, groups = var_4686_groups_0, pad = var_4686_pad_0, pad_type = var_4686_pad_type_0, strides = var_4686_strides_0, weight = layers_11_feed_forward2_fc1_inlier_module_weight_to_fp16_palettized, x = input_323_cast_fp16)[name = tensor("op_4686_cast_fp16")]; - tensor var_4692_pad_type_0 = const()[name = tensor("op_4692_pad_type_0"), val = tensor("valid")]; - tensor var_4692_strides_0 = const()[name = tensor("op_4692_strides_0"), val = tensor([1, 1])]; - tensor var_4692_pad_0 = const()[name = tensor("op_4692_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4692_dilations_0 = const()[name = tensor("op_4692_dilations_0"), val = tensor([1, 1])]; - tensor var_4692_groups_0 = const()[name = tensor("op_4692_groups_0"), val = tensor(1)]; - tensor layers_11_feed_forward2_fc1_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70009856))), name = tensor("layers_11_feed_forward2_fc1_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(69974976))), shape = tensor([2048, 512, 1, 1])]; - tensor var_4692_cast_fp16 = conv(dilations = var_4692_dilations_0, groups = var_4692_groups_0, pad = var_4692_pad_0, pad_type = var_4692_pad_type_0, strides = var_4692_strides_0, weight = layers_11_feed_forward2_fc1_outlier_module_weight_to_fp16_sparsified, x = input_323_cast_fp16)[name = tensor("op_4692_cast_fp16")]; - tensor input_325_cast_fp16 = add(x = var_4686_cast_fp16, y = var_4692_cast_fp16)[name = tensor("input_325_cast_fp16")]; - tensor input_327_cast_fp16 = silu(x = input_325_cast_fp16)[name = tensor("input_327_cast_fp16")]; - tensor var_4703_pad_type_0 = const()[name = tensor("op_4703_pad_type_0"), val = tensor("valid")]; - tensor var_4703_strides_0 = const()[name = tensor("op_4703_strides_0"), val = tensor([1, 1])]; - tensor var_4703_pad_0 = const()[name = tensor("op_4703_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4703_dilations_0 = const()[name = tensor("op_4703_dilations_0"), val = tensor([1, 1])]; - tensor var_4703_groups_0 = const()[name = tensor("op_4703_groups_0"), val = tensor(1)]; - tensor layers_11_feed_forward2_fc2_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70140992))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70927488))), name = tensor("layers_11_feed_forward2_fc2_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 2048, 1, 1])]; - tensor layers_11_feed_forward2_fc2_inlier_module_bias_to_fp16 = const()[name = tensor("layers_11_feed_forward2_fc2_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70927680)))]; - tensor var_4703_cast_fp16 = conv(bias = layers_11_feed_forward2_fc2_inlier_module_bias_to_fp16, dilations = var_4703_dilations_0, groups = var_4703_groups_0, pad = var_4703_pad_0, pad_type = var_4703_pad_type_0, strides = var_4703_strides_0, weight = layers_11_feed_forward2_fc2_inlier_module_weight_to_fp16_palettized, x = input_327_cast_fp16)[name = tensor("op_4703_cast_fp16")]; - tensor var_4709_pad_type_0 = const()[name = tensor("op_4709_pad_type_0"), val = tensor("valid")]; - tensor var_4709_strides_0 = const()[name = tensor("op_4709_strides_0"), val = tensor([1, 1])]; - tensor var_4709_pad_0 = const()[name = tensor("op_4709_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4709_dilations_0 = const()[name = tensor("op_4709_dilations_0"), val = tensor([1, 1])]; - tensor var_4709_groups_0 = const()[name = tensor("op_4709_groups_0"), val = tensor(1)]; - tensor layers_11_feed_forward2_fc2_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70970688))), name = tensor("layers_11_feed_forward2_fc2_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70928768))), shape = tensor([512, 2048, 1, 1])]; - tensor var_4709_cast_fp16 = conv(dilations = var_4709_dilations_0, groups = var_4709_groups_0, pad = var_4709_pad_0, pad_type = var_4709_pad_type_0, strides = var_4709_strides_0, weight = layers_11_feed_forward2_fc2_outlier_module_weight_to_fp16_sparsified, x = input_327_cast_fp16)[name = tensor("op_4709_cast_fp16")]; - tensor x_73_cast_fp16 = add(x = var_4703_cast_fp16, y = var_4709_cast_fp16)[name = tensor("x_73_cast_fp16")]; - tensor var_4711_to_fp16 = const()[name = tensor("op_4711_to_fp16"), val = tensor(0x1p-1)]; - tensor var_4712_cast_fp16 = mul(x = x_73_cast_fp16, y = var_4711_to_fp16)[name = tensor("op_4712_cast_fp16")]; - tensor inputs_119_cast_fp16 = add(x = inputs_117_cast_fp16, y = var_4712_cast_fp16)[name = tensor("inputs_119_cast_fp16")]; - tensor out_119_axes_0 = const()[name = tensor("out_119_axes_0"), val = tensor([1])]; - tensor var_4722_to_fp16 = const()[name = tensor("op_4722_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_119_cast_fp16 = layer_norm(axes = out_119_axes_0, epsilon = var_4722_to_fp16, x = inputs_119_cast_fp16)[name = tensor("out_119_cast_fp16")]; - tensor inputs_121_gamma_0_to_fp16 = const()[name = tensor("inputs_121_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71101824)))]; - tensor inputs_121_beta_0_to_fp16 = const()[name = tensor("inputs_121_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71102912)))]; - tensor inputs_121_epsilon_0_to_fp16 = const()[name = tensor("inputs_121_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor inputs_121_cast_fp16 = batch_norm(beta = inputs_121_beta_0_to_fp16, epsilon = inputs_121_epsilon_0_to_fp16, gamma = inputs_121_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_119_cast_fp16)[name = tensor("inputs_121_cast_fp16")]; - tensor var_4736 = const()[name = tensor("op_4736"), val = tensor(3)]; - tensor out_121_axes_0 = const()[name = tensor("out_121_axes_0"), val = tensor([1])]; - tensor var_4767_to_fp16 = const()[name = tensor("op_4767_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_121_cast_fp16 = layer_norm(axes = out_121_axes_0, epsilon = var_4767_to_fp16, x = inputs_121_cast_fp16)[name = tensor("out_121_cast_fp16")]; - tensor input_329_gamma_0_to_fp16 = const()[name = tensor("input_329_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71104000)))]; - tensor input_329_beta_0_to_fp16 = const()[name = tensor("input_329_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71105088)))]; - tensor input_329_epsilon_0_to_fp16 = const()[name = tensor("input_329_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_329_cast_fp16 = batch_norm(beta = input_329_beta_0_to_fp16, epsilon = input_329_epsilon_0_to_fp16, gamma = input_329_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_121_cast_fp16)[name = tensor("input_329_cast_fp16")]; - tensor var_4787_pad_type_0 = const()[name = tensor("op_4787_pad_type_0"), val = tensor("valid")]; - tensor var_4787_strides_0 = const()[name = tensor("op_4787_strides_0"), val = tensor([1, 1])]; - tensor var_4787_pad_0 = const()[name = tensor("op_4787_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4787_dilations_0 = const()[name = tensor("op_4787_dilations_0"), val = tensor([1, 1])]; - tensor var_4787_groups_0 = const()[name = tensor("op_4787_groups_0"), val = tensor(1)]; - tensor layers_12_feed_forward1_fc1_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71106176))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71892672))), name = tensor("layers_12_feed_forward1_fc1_inlier_module_weight_to_fp16_palettized"), shape = tensor([2048, 512, 1, 1])]; - tensor layers_12_feed_forward1_fc1_inlier_module_bias_to_fp16 = const()[name = tensor("layers_12_feed_forward1_fc1_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71892864)))]; - tensor var_4787_cast_fp16 = conv(bias = layers_12_feed_forward1_fc1_inlier_module_bias_to_fp16, dilations = var_4787_dilations_0, groups = var_4787_groups_0, pad = var_4787_pad_0, pad_type = var_4787_pad_type_0, strides = var_4787_strides_0, weight = layers_12_feed_forward1_fc1_inlier_module_weight_to_fp16_palettized, x = input_329_cast_fp16)[name = tensor("op_4787_cast_fp16")]; - tensor var_4793_pad_type_0 = const()[name = tensor("op_4793_pad_type_0"), val = tensor("valid")]; - tensor var_4793_strides_0 = const()[name = tensor("op_4793_strides_0"), val = tensor([1, 1])]; - tensor var_4793_pad_0 = const()[name = tensor("op_4793_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4793_dilations_0 = const()[name = tensor("op_4793_dilations_0"), val = tensor([1, 1])]; - tensor var_4793_groups_0 = const()[name = tensor("op_4793_groups_0"), val = tensor(1)]; - tensor layers_12_feed_forward1_fc1_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71931776))), name = tensor("layers_12_feed_forward1_fc1_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71897024))), shape = tensor([2048, 512, 1, 1])]; - tensor var_4793_cast_fp16 = conv(dilations = var_4793_dilations_0, groups = var_4793_groups_0, pad = var_4793_pad_0, pad_type = var_4793_pad_type_0, strides = var_4793_strides_0, weight = layers_12_feed_forward1_fc1_outlier_module_weight_to_fp16_sparsified, x = input_329_cast_fp16)[name = tensor("op_4793_cast_fp16")]; - tensor input_331_cast_fp16 = add(x = var_4787_cast_fp16, y = var_4793_cast_fp16)[name = tensor("input_331_cast_fp16")]; - tensor input_333_cast_fp16 = silu(x = input_331_cast_fp16)[name = tensor("input_333_cast_fp16")]; - tensor var_4804_pad_type_0 = const()[name = tensor("op_4804_pad_type_0"), val = tensor("valid")]; - tensor var_4804_strides_0 = const()[name = tensor("op_4804_strides_0"), val = tensor([1, 1])]; - tensor var_4804_pad_0 = const()[name = tensor("op_4804_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4804_dilations_0 = const()[name = tensor("op_4804_dilations_0"), val = tensor([1, 1])]; - tensor var_4804_groups_0 = const()[name = tensor("op_4804_groups_0"), val = tensor(1)]; - tensor layers_12_feed_forward1_fc2_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(72062912))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(72849408))), name = tensor("layers_12_feed_forward1_fc2_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 2048, 1, 1])]; - tensor layers_12_feed_forward1_fc2_inlier_module_bias_to_fp16 = const()[name = tensor("layers_12_feed_forward1_fc2_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(72849600)))]; - tensor var_4804_cast_fp16 = conv(bias = layers_12_feed_forward1_fc2_inlier_module_bias_to_fp16, dilations = var_4804_dilations_0, groups = var_4804_groups_0, pad = var_4804_pad_0, pad_type = var_4804_pad_type_0, strides = var_4804_strides_0, weight = layers_12_feed_forward1_fc2_inlier_module_weight_to_fp16_palettized, x = input_333_cast_fp16)[name = tensor("op_4804_cast_fp16")]; - tensor var_4810_pad_type_0 = const()[name = tensor("op_4810_pad_type_0"), val = tensor("valid")]; - tensor var_4810_strides_0 = const()[name = tensor("op_4810_strides_0"), val = tensor([1, 1])]; - tensor var_4810_pad_0 = const()[name = tensor("op_4810_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4810_dilations_0 = const()[name = tensor("op_4810_dilations_0"), val = tensor([1, 1])]; - tensor var_4810_groups_0 = const()[name = tensor("op_4810_groups_0"), val = tensor(1)]; - tensor layers_12_feed_forward1_fc2_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(72890432))), name = tensor("layers_12_feed_forward1_fc2_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(72850688))), shape = tensor([512, 2048, 1, 1])]; - tensor var_4810_cast_fp16 = conv(dilations = var_4810_dilations_0, groups = var_4810_groups_0, pad = var_4810_pad_0, pad_type = var_4810_pad_type_0, strides = var_4810_strides_0, weight = layers_12_feed_forward1_fc2_outlier_module_weight_to_fp16_sparsified, x = input_333_cast_fp16)[name = tensor("op_4810_cast_fp16")]; - tensor x_75_cast_fp16 = add(x = var_4804_cast_fp16, y = var_4810_cast_fp16)[name = tensor("x_75_cast_fp16")]; - tensor var_4812_to_fp16 = const()[name = tensor("op_4812_to_fp16"), val = tensor(0x1p-1)]; - tensor var_4813_cast_fp16 = mul(x = x_75_cast_fp16, y = var_4812_to_fp16)[name = tensor("op_4813_cast_fp16")]; - tensor inputs_123_cast_fp16 = add(x = inputs_121_cast_fp16, y = var_4813_cast_fp16)[name = tensor("inputs_123_cast_fp16")]; - tensor out_123_axes_0 = const()[name = tensor("out_123_axes_0"), val = tensor([1])]; - tensor var_4823_to_fp16 = const()[name = tensor("op_4823_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_123_cast_fp16 = layer_norm(axes = out_123_axes_0, epsilon = var_4823_to_fp16, x = inputs_123_cast_fp16)[name = tensor("out_123_cast_fp16")]; - tensor obj_51_gamma_0_to_fp16 = const()[name = tensor("obj_51_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73021568)))]; - tensor obj_51_beta_0_to_fp16 = const()[name = tensor("obj_51_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73022656)))]; - tensor obj_51_epsilon_0_to_fp16 = const()[name = tensor("obj_51_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor obj_51_cast_fp16 = batch_norm(beta = obj_51_beta_0_to_fp16, epsilon = obj_51_epsilon_0_to_fp16, gamma = obj_51_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_123_cast_fp16)[name = tensor("obj_51_cast_fp16")]; - tensor var_4848_pad_type_0 = const()[name = tensor("op_4848_pad_type_0"), val = tensor("valid")]; - tensor var_4848_strides_0 = const()[name = tensor("op_4848_strides_0"), val = tensor([1, 1])]; - tensor var_4848_pad_0 = const()[name = tensor("op_4848_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4848_dilations_0 = const()[name = tensor("op_4848_dilations_0"), val = tensor([1, 1])]; - tensor var_4848_groups_0 = const()[name = tensor("op_4848_groups_0"), val = tensor(1)]; - tensor layers_12_self_attn_q_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73023744))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73220416))), name = tensor("layers_12_self_attn_q_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_12_self_attn_q_proj_inlier_module_bias_to_fp16 = const()[name = tensor("layers_12_self_attn_q_proj_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73220608)))]; - tensor var_4848_cast_fp16 = conv(bias = layers_12_self_attn_q_proj_inlier_module_bias_to_fp16, dilations = var_4848_dilations_0, groups = var_4848_groups_0, pad = var_4848_pad_0, pad_type = var_4848_pad_type_0, strides = var_4848_strides_0, weight = layers_12_self_attn_q_proj_inlier_module_weight_to_fp16_palettized, x = obj_51_cast_fp16)[name = tensor("op_4848_cast_fp16")]; - tensor var_4854_pad_type_0 = const()[name = tensor("op_4854_pad_type_0"), val = tensor("valid")]; - tensor var_4854_strides_0 = const()[name = tensor("op_4854_strides_0"), val = tensor([1, 1])]; - tensor var_4854_pad_0 = const()[name = tensor("op_4854_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4854_dilations_0 = const()[name = tensor("op_4854_dilations_0"), val = tensor([1, 1])]; - tensor var_4854_groups_0 = const()[name = tensor("op_4854_groups_0"), val = tensor(1)]; - tensor layers_12_self_attn_q_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73230784))), name = tensor("layers_12_self_attn_q_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73221696))), shape = tensor([512, 512, 1, 1])]; - tensor var_4854_cast_fp16 = conv(dilations = var_4854_dilations_0, groups = var_4854_groups_0, pad = var_4854_pad_0, pad_type = var_4854_pad_type_0, strides = var_4854_strides_0, weight = layers_12_self_attn_q_proj_outlier_module_weight_to_fp16_sparsified, x = obj_51_cast_fp16)[name = tensor("op_4854_cast_fp16")]; - tensor query_49_cast_fp16 = add(x = var_4848_cast_fp16, y = var_4854_cast_fp16)[name = tensor("query_49_cast_fp16")]; - tensor var_4863_pad_type_0 = const()[name = tensor("op_4863_pad_type_0"), val = tensor("valid")]; - tensor var_4863_strides_0 = const()[name = tensor("op_4863_strides_0"), val = tensor([1, 1])]; - tensor var_4863_pad_0 = const()[name = tensor("op_4863_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4863_dilations_0 = const()[name = tensor("op_4863_dilations_0"), val = tensor([1, 1])]; - tensor var_4863_groups_0 = const()[name = tensor("op_4863_groups_0"), val = tensor(1)]; - tensor layers_12_self_attn_k_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73263616))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73460288))), name = tensor("layers_12_self_attn_k_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor var_4863_cast_fp16 = conv(dilations = var_4863_dilations_0, groups = var_4863_groups_0, pad = var_4863_pad_0, pad_type = var_4863_pad_type_0, strides = var_4863_strides_0, weight = layers_12_self_attn_k_proj_inlier_module_weight_to_fp16_palettized, x = obj_51_cast_fp16)[name = tensor("op_4863_cast_fp16")]; - tensor var_4869_pad_type_0 = const()[name = tensor("op_4869_pad_type_0"), val = tensor("valid")]; - tensor var_4869_strides_0 = const()[name = tensor("op_4869_strides_0"), val = tensor([1, 1])]; - tensor var_4869_pad_0 = const()[name = tensor("op_4869_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4869_dilations_0 = const()[name = tensor("op_4869_dilations_0"), val = tensor([1, 1])]; - tensor var_4869_groups_0 = const()[name = tensor("op_4869_groups_0"), val = tensor(1)]; - tensor layers_12_self_attn_k_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73468160))), name = tensor("layers_12_self_attn_k_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73460480))), shape = tensor([512, 512, 1, 1])]; - tensor var_4869_cast_fp16 = conv(dilations = var_4869_dilations_0, groups = var_4869_groups_0, pad = var_4869_pad_0, pad_type = var_4869_pad_type_0, strides = var_4869_strides_0, weight = layers_12_self_attn_k_proj_outlier_module_weight_to_fp16_sparsified, x = obj_51_cast_fp16)[name = tensor("op_4869_cast_fp16")]; - tensor key_25_cast_fp16 = add(x = var_4863_cast_fp16, y = var_4869_cast_fp16)[name = tensor("key_25_cast_fp16")]; - tensor var_4879_pad_type_0 = const()[name = tensor("op_4879_pad_type_0"), val = tensor("valid")]; - tensor var_4879_strides_0 = const()[name = tensor("op_4879_strides_0"), val = tensor([1, 1])]; - tensor var_4879_pad_0 = const()[name = tensor("op_4879_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4879_dilations_0 = const()[name = tensor("op_4879_dilations_0"), val = tensor([1, 1])]; - tensor var_4879_groups_0 = const()[name = tensor("op_4879_groups_0"), val = tensor(1)]; - tensor layers_12_self_attn_v_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73500992))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73697664))), name = tensor("layers_12_self_attn_v_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_12_self_attn_v_proj_inlier_module_bias_to_fp16 = const()[name = tensor("layers_12_self_attn_v_proj_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73697856)))]; - tensor var_4879_cast_fp16 = conv(bias = layers_12_self_attn_v_proj_inlier_module_bias_to_fp16, dilations = var_4879_dilations_0, groups = var_4879_groups_0, pad = var_4879_pad_0, pad_type = var_4879_pad_type_0, strides = var_4879_strides_0, weight = layers_12_self_attn_v_proj_inlier_module_weight_to_fp16_palettized, x = obj_51_cast_fp16)[name = tensor("op_4879_cast_fp16")]; - tensor var_4885_pad_type_0 = const()[name = tensor("op_4885_pad_type_0"), val = tensor("valid")]; - tensor var_4885_strides_0 = const()[name = tensor("op_4885_strides_0"), val = tensor([1, 1])]; - tensor var_4885_pad_0 = const()[name = tensor("op_4885_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4885_dilations_0 = const()[name = tensor("op_4885_dilations_0"), val = tensor([1, 1])]; - tensor var_4885_groups_0 = const()[name = tensor("op_4885_groups_0"), val = tensor(1)]; - tensor layers_12_self_attn_v_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73707712))), name = tensor("layers_12_self_attn_v_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73698944))), shape = tensor([512, 512, 1, 1])]; - tensor var_4885_cast_fp16 = conv(dilations = var_4885_dilations_0, groups = var_4885_groups_0, pad = var_4885_pad_0, pad_type = var_4885_pad_type_0, strides = var_4885_strides_0, weight = layers_12_self_attn_v_proj_outlier_module_weight_to_fp16_sparsified, x = obj_51_cast_fp16)[name = tensor("op_4885_cast_fp16")]; - tensor value_25_cast_fp16 = add(x = var_4879_cast_fp16, y = var_4885_cast_fp16)[name = tensor("value_25_cast_fp16")]; - tensor var_4888_to_fp16 = const()[name = tensor("op_4888_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73740544)))]; - tensor query_51_cast_fp16 = add(x = query_49_cast_fp16, y = var_4888_to_fp16)[name = tensor("query_51_cast_fp16")]; - tensor var_4891_to_fp16 = const()[name = tensor("op_4891_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73741632)))]; - tensor q_with_bias_v_25_cast_fp16 = add(x = query_49_cast_fp16, y = var_4891_to_fp16)[name = tensor("q_with_bias_v_25_cast_fp16")]; - tensor var_4901_pad_type_0 = const()[name = tensor("op_4901_pad_type_0"), val = tensor("valid")]; - tensor var_4901_strides_0 = const()[name = tensor("op_4901_strides_0"), val = tensor([1, 1])]; - tensor var_4901_pad_0 = const()[name = tensor("op_4901_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4901_dilations_0 = const()[name = tensor("op_4901_dilations_0"), val = tensor([1, 1])]; - tensor var_4901_groups_0 = const()[name = tensor("op_4901_groups_0"), val = tensor(1)]; - tensor layers_12_self_attn_linear_pos_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73742720))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73939392))), name = tensor("layers_12_self_attn_linear_pos_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor var_4901_cast_fp16 = conv(bias = input_17_mean_0_to_fp16, dilations = var_4901_dilations_0, groups = var_4901_groups_0, pad = var_4901_pad_0, pad_type = var_4901_pad_type_0, strides = var_4901_strides_0, weight = layers_12_self_attn_linear_pos_inlier_module_weight_to_fp16_palettized, x = obj_3_cast_fp16)[name = tensor("op_4901_cast_fp16")]; - tensor var_4907_pad_type_0 = const()[name = tensor("op_4907_pad_type_0"), val = tensor("valid")]; - tensor var_4907_strides_0 = const()[name = tensor("op_4907_strides_0"), val = tensor([1, 1])]; - tensor var_4907_pad_0 = const()[name = tensor("op_4907_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4907_dilations_0 = const()[name = tensor("op_4907_dilations_0"), val = tensor([1, 1])]; - tensor var_4907_groups_0 = const()[name = tensor("op_4907_groups_0"), val = tensor(1)]; - tensor layers_12_self_attn_linear_pos_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73961280))), name = tensor("layers_12_self_attn_linear_pos_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73939584))), shape = tensor([512, 512, 1, 1])]; - tensor var_4907_cast_fp16 = conv(dilations = var_4907_dilations_0, groups = var_4907_groups_0, pad = var_4907_pad_0, pad_type = var_4907_pad_type_0, strides = var_4907_strides_0, weight = layers_12_self_attn_linear_pos_outlier_module_weight_to_fp16_sparsified, x = obj_3_cast_fp16)[name = tensor("op_4907_cast_fp16")]; - tensor p_25_cast_fp16 = add(x = var_4901_cast_fp16, y = var_4907_cast_fp16)[name = tensor("p_25_cast_fp16")]; - tensor var_4911 = const()[name = tensor("op_4911"), val = tensor([1, 8, 64, 188])]; - tensor var_4912_cast_fp16 = reshape(shape = var_4911, x = q_with_bias_v_25_cast_fp16)[name = tensor("op_4912_cast_fp16")]; - tensor var_4913 = const()[name = tensor("op_4913"), val = tensor([1, 8, 64, -1])]; - tensor var_4914_cast_fp16 = reshape(shape = var_4913, x = p_25_cast_fp16)[name = tensor("op_4914_cast_fp16")]; - tensor matrix_bd_97_transpose_x_0 = const()[name = tensor("matrix_bd_97_transpose_x_0"), val = tensor(true)]; - tensor matrix_bd_97_transpose_y_0 = const()[name = tensor("matrix_bd_97_transpose_y_0"), val = tensor(false)]; - tensor matrix_bd_97_cast_fp16 = matmul(transpose_x = matrix_bd_97_transpose_x_0, transpose_y = matrix_bd_97_transpose_y_0, x = var_4912_cast_fp16, y = var_4914_cast_fp16)[name = tensor("matrix_bd_97_cast_fp16")]; - tensor matrix_bd_99_pad_0 = const()[name = tensor("matrix_bd_99_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor matrix_bd_99_mode_0 = const()[name = tensor("matrix_bd_99_mode_0"), val = tensor("constant")]; - tensor const_142_to_fp16 = const()[name = tensor("const_142_to_fp16"), val = tensor(0x0p+0)]; - tensor matrix_bd_99_cast_fp16 = pad(constant_val = const_142_to_fp16, mode = matrix_bd_99_mode_0, pad = matrix_bd_99_pad_0, x = matrix_bd_97_cast_fp16)[name = tensor("matrix_bd_99_cast_fp16")]; - tensor var_4923 = const()[name = tensor("op_4923"), val = tensor([1, 8, -1, 188])]; - tensor matrix_bd_101_cast_fp16 = reshape(shape = var_4923, x = matrix_bd_99_cast_fp16)[name = tensor("matrix_bd_101_cast_fp16")]; - tensor var_4927_begin_0 = const()[name = tensor("op_4927_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_4927_end_0 = const()[name = tensor("op_4927_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_4927_end_mask_0 = const()[name = tensor("op_4927_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_4927_cast_fp16 = slice_by_index(begin = var_4927_begin_0, end = var_4927_end_0, end_mask = var_4927_end_mask_0, x = matrix_bd_101_cast_fp16)[name = tensor("op_4927_cast_fp16")]; - tensor var_4928 = const()[name = tensor("op_4928"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_103_cast_fp16 = reshape(shape = var_4928, x = var_4927_cast_fp16)[name = tensor("matrix_bd_103_cast_fp16")]; - tensor var_4933_begin_0 = const()[name = tensor("op_4933_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4933_end_0 = const()[name = tensor("op_4933_end_0"), val = tensor([1, 8, 188, 188])]; - tensor var_4933_end_mask_0 = const()[name = tensor("op_4933_end_mask_0"), val = tensor([true, true, true, false])]; - tensor var_4933_cast_fp16 = slice_by_index(begin = var_4933_begin_0, end = var_4933_end_0, end_mask = var_4933_end_mask_0, x = matrix_bd_103_cast_fp16)[name = tensor("op_4933_cast_fp16")]; - tensor var_4934_to_fp16 = const()[name = tensor("op_4934_to_fp16"), val = tensor(0x1p-3)]; - tensor qk_mask_25_cast_fp16 = mul(x = var_4933_cast_fp16, y = var_4934_to_fp16)[name = tensor("qk_mask_25_cast_fp16")]; - tensor var_4938 = const()[name = tensor("op_4938"), val = tensor([1, 8, 64, 188])]; - tensor mh_q_25_cast_fp16 = reshape(shape = var_4938, x = query_51_cast_fp16)[name = tensor("mh_q_25_cast_fp16")]; - tensor var_4940_to_fp16 = const()[name = tensor("op_4940_to_fp16"), val = tensor(0x1p-3)]; - tensor var_4941_cast_fp16 = mul(x = mh_q_25_cast_fp16, y = var_4940_to_fp16)[name = tensor("op_4941_cast_fp16")]; - tensor var_4944 = const()[name = tensor("op_4944"), val = tensor([1, 8, 64, 188])]; - tensor var_4945_cast_fp16 = reshape(shape = var_4944, x = key_25_cast_fp16)[name = tensor("op_4945_cast_fp16")]; - tensor mh_w_49_transpose_x_0 = const()[name = tensor("mh_w_49_transpose_x_0"), val = tensor(true)]; - tensor mh_w_49_transpose_y_0 = const()[name = tensor("mh_w_49_transpose_y_0"), val = tensor(false)]; - tensor mh_w_49_cast_fp16 = matmul(transpose_x = mh_w_49_transpose_x_0, transpose_y = mh_w_49_transpose_y_0, x = var_4941_cast_fp16, y = var_4945_cast_fp16)[name = tensor("mh_w_49_cast_fp16")]; - tensor mh_w_51_cast_fp16 = add(x = mh_w_49_cast_fp16, y = qk_mask_25_cast_fp16)[name = tensor("mh_w_51_cast_fp16")]; - tensor var_4949_cast_fp16 = softmax(axis = var_4736, x = mh_w_51_cast_fp16)[name = tensor("op_4949_cast_fp16")]; - tensor var_4950 = const()[name = tensor("op_4950"), val = tensor([1, 8, 64, 188])]; - tensor var_4951_cast_fp16 = reshape(shape = var_4950, x = value_25_cast_fp16)[name = tensor("op_4951_cast_fp16")]; - tensor attn_25_transpose_x_0 = const()[name = tensor("attn_25_transpose_x_0"), val = tensor(false)]; - tensor attn_25_transpose_y_0 = const()[name = tensor("attn_25_transpose_y_0"), val = tensor(true)]; - tensor attn_25_cast_fp16 = matmul(transpose_x = attn_25_transpose_x_0, transpose_y = attn_25_transpose_y_0, x = var_4951_cast_fp16, y = var_4949_cast_fp16)[name = tensor("attn_25_cast_fp16")]; - tensor var_4954 = const()[name = tensor("op_4954"), val = tensor([1, 512, 1, 188])]; - tensor input_335_cast_fp16 = reshape(shape = var_4954, x = attn_25_cast_fp16)[name = tensor("input_335_cast_fp16")]; - tensor var_4964_pad_type_0 = const()[name = tensor("op_4964_pad_type_0"), val = tensor("valid")]; - tensor var_4964_strides_0 = const()[name = tensor("op_4964_strides_0"), val = tensor([1, 1])]; - tensor var_4964_pad_0 = const()[name = tensor("op_4964_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4964_dilations_0 = const()[name = tensor("op_4964_dilations_0"), val = tensor([1, 1])]; - tensor var_4964_groups_0 = const()[name = tensor("op_4964_groups_0"), val = tensor(1)]; - tensor layers_12_self_attn_o_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73994112))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(74190784))), name = tensor("layers_12_self_attn_o_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_12_self_attn_o_proj_inlier_module_bias_to_fp16 = const()[name = tensor("layers_12_self_attn_o_proj_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(74190976)))]; - tensor var_4964_cast_fp16 = conv(bias = layers_12_self_attn_o_proj_inlier_module_bias_to_fp16, dilations = var_4964_dilations_0, groups = var_4964_groups_0, pad = var_4964_pad_0, pad_type = var_4964_pad_type_0, strides = var_4964_strides_0, weight = layers_12_self_attn_o_proj_inlier_module_weight_to_fp16_palettized, x = input_335_cast_fp16)[name = tensor("op_4964_cast_fp16")]; - tensor var_4970_pad_type_0 = const()[name = tensor("op_4970_pad_type_0"), val = tensor("valid")]; - tensor var_4970_strides_0 = const()[name = tensor("op_4970_strides_0"), val = tensor([1, 1])]; - tensor var_4970_pad_0 = const()[name = tensor("op_4970_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_4970_dilations_0 = const()[name = tensor("op_4970_dilations_0"), val = tensor([1, 1])]; - tensor var_4970_groups_0 = const()[name = tensor("op_4970_groups_0"), val = tensor(1)]; - tensor layers_12_self_attn_o_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(74201984))), name = tensor("layers_12_self_attn_o_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(74192064))), shape = tensor([512, 512, 1, 1])]; - tensor var_4970_cast_fp16 = conv(dilations = var_4970_dilations_0, groups = var_4970_groups_0, pad = var_4970_pad_0, pad_type = var_4970_pad_type_0, strides = var_4970_strides_0, weight = layers_12_self_attn_o_proj_outlier_module_weight_to_fp16_sparsified, x = input_335_cast_fp16)[name = tensor("op_4970_cast_fp16")]; - tensor obj_53_cast_fp16 = add(x = var_4964_cast_fp16, y = var_4970_cast_fp16)[name = tensor("obj_53_cast_fp16")]; - tensor inputs_125_cast_fp16 = add(x = inputs_123_cast_fp16, y = obj_53_cast_fp16)[name = tensor("inputs_125_cast_fp16")]; - tensor out_125_axes_0 = const()[name = tensor("out_125_axes_0"), val = tensor([1])]; - tensor var_4981_to_fp16 = const()[name = tensor("op_4981_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_125_cast_fp16 = layer_norm(axes = out_125_axes_0, epsilon = var_4981_to_fp16, x = inputs_125_cast_fp16)[name = tensor("out_125_cast_fp16")]; - tensor input_337_gamma_0_to_fp16 = const()[name = tensor("input_337_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(74234816)))]; - tensor input_337_beta_0_to_fp16 = const()[name = tensor("input_337_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(74235904)))]; - tensor input_337_epsilon_0_to_fp16 = const()[name = tensor("input_337_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_337_cast_fp16 = batch_norm(beta = input_337_beta_0_to_fp16, epsilon = input_337_epsilon_0_to_fp16, gamma = input_337_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_125_cast_fp16)[name = tensor("input_337_cast_fp16")]; - tensor var_5003_pad_type_0 = const()[name = tensor("op_5003_pad_type_0"), val = tensor("valid")]; - tensor var_5003_strides_0 = const()[name = tensor("op_5003_strides_0"), val = tensor([1, 1])]; - tensor var_5003_pad_0 = const()[name = tensor("op_5003_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5003_dilations_0 = const()[name = tensor("op_5003_dilations_0"), val = tensor([1, 1])]; - tensor var_5003_groups_0 = const()[name = tensor("op_5003_groups_0"), val = tensor(1)]; - tensor layers_12_conv_pointwise_conv1_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(74236992))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(74630272))), name = tensor("layers_12_conv_pointwise_conv1_inlier_module_weight_to_fp16_palettized"), shape = tensor([1024, 512, 1, 1])]; - tensor layers_12_conv_pointwise_conv1_inlier_module_bias_to_fp16 = const()[name = tensor("layers_12_conv_pointwise_conv1_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(74630464)))]; - tensor var_5003_cast_fp16 = conv(bias = layers_12_conv_pointwise_conv1_inlier_module_bias_to_fp16, dilations = var_5003_dilations_0, groups = var_5003_groups_0, pad = var_5003_pad_0, pad_type = var_5003_pad_type_0, strides = var_5003_strides_0, weight = layers_12_conv_pointwise_conv1_inlier_module_weight_to_fp16_palettized, x = input_337_cast_fp16)[name = tensor("op_5003_cast_fp16")]; - tensor var_5009_pad_type_0 = const()[name = tensor("op_5009_pad_type_0"), val = tensor("valid")]; - tensor var_5009_strides_0 = const()[name = tensor("op_5009_strides_0"), val = tensor([1, 1])]; - tensor var_5009_pad_0 = const()[name = tensor("op_5009_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5009_dilations_0 = const()[name = tensor("op_5009_dilations_0"), val = tensor([1, 1])]; - tensor var_5009_groups_0 = const()[name = tensor("op_5009_groups_0"), val = tensor(1)]; - tensor layers_12_conv_pointwise_conv1_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(74651328))), name = tensor("layers_12_conv_pointwise_conv1_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(74632576))), shape = tensor([1024, 512, 1, 1])]; - tensor var_5009_cast_fp16 = conv(dilations = var_5009_dilations_0, groups = var_5009_groups_0, pad = var_5009_pad_0, pad_type = var_5009_pad_type_0, strides = var_5009_strides_0, weight = layers_12_conv_pointwise_conv1_outlier_module_weight_to_fp16_sparsified, x = input_337_cast_fp16)[name = tensor("op_5009_cast_fp16")]; - tensor input_339_cast_fp16 = add(x = var_5003_cast_fp16, y = var_5009_cast_fp16)[name = tensor("input_339_cast_fp16")]; - tensor input_341_split_num_splits_0 = const()[name = tensor("input_341_split_num_splits_0"), val = tensor(2)]; - tensor input_341_split_axis_0 = const()[name = tensor("input_341_split_axis_0"), val = tensor(1)]; - tensor input_341_split_cast_fp16_0, tensor input_341_split_cast_fp16_1 = split(axis = input_341_split_axis_0, num_splits = input_341_split_num_splits_0, x = input_339_cast_fp16)[name = tensor("input_341_split_cast_fp16")]; - tensor input_341_split_1_sigmoid_cast_fp16 = sigmoid(x = input_341_split_cast_fp16_1)[name = tensor("input_341_split_1_sigmoid_cast_fp16")]; - tensor input_341_cast_fp16 = mul(x = input_341_split_cast_fp16_0, y = input_341_split_1_sigmoid_cast_fp16)[name = tensor("input_341_cast_fp16")]; - tensor input_343_pad_type_0 = const()[name = tensor("input_343_pad_type_0"), val = tensor("custom")]; - tensor input_343_pad_0 = const()[name = tensor("input_343_pad_0"), val = tensor([0, 0, 4, 4])]; - tensor input_343_groups_0 = const()[name = tensor("input_343_groups_0"), val = tensor(512)]; - tensor input_343_strides_0 = const()[name = tensor("input_343_strides_0"), val = tensor([1, 1])]; - tensor input_343_dilations_0 = const()[name = tensor("input_343_dilations_0"), val = tensor([1, 1])]; - tensor const_215_to_fp16 = const()[name = tensor("const_215_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(74716928)))]; - tensor const_216_to_fp16 = const()[name = tensor("const_216_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(74726208)))]; - tensor input_345_cast_fp16 = conv(bias = const_216_to_fp16, dilations = input_343_dilations_0, groups = input_343_groups_0, pad = input_343_pad_0, pad_type = input_343_pad_type_0, strides = input_343_strides_0, weight = const_215_to_fp16, x = input_341_cast_fp16)[name = tensor("input_345_cast_fp16")]; - tensor input_347_cast_fp16 = silu(x = input_345_cast_fp16)[name = tensor("input_347_cast_fp16")]; - tensor var_5033_pad_type_0 = const()[name = tensor("op_5033_pad_type_0"), val = tensor("valid")]; - tensor var_5033_strides_0 = const()[name = tensor("op_5033_strides_0"), val = tensor([1, 1])]; - tensor var_5033_pad_0 = const()[name = tensor("op_5033_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5033_dilations_0 = const()[name = tensor("op_5033_dilations_0"), val = tensor([1, 1])]; - tensor var_5033_groups_0 = const()[name = tensor("op_5033_groups_0"), val = tensor(1)]; - tensor layers_12_conv_pointwise_conv2_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(74727296))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(74923968))), name = tensor("layers_12_conv_pointwise_conv2_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_12_conv_pointwise_conv2_inlier_module_bias_to_fp16 = const()[name = tensor("layers_12_conv_pointwise_conv2_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(74924160)))]; - tensor var_5033_cast_fp16 = conv(bias = layers_12_conv_pointwise_conv2_inlier_module_bias_to_fp16, dilations = var_5033_dilations_0, groups = var_5033_groups_0, pad = var_5033_pad_0, pad_type = var_5033_pad_type_0, strides = var_5033_strides_0, weight = layers_12_conv_pointwise_conv2_inlier_module_weight_to_fp16_palettized, x = input_347_cast_fp16)[name = tensor("op_5033_cast_fp16")]; - tensor var_5039_pad_type_0 = const()[name = tensor("op_5039_pad_type_0"), val = tensor("valid")]; - tensor var_5039_strides_0 = const()[name = tensor("op_5039_strides_0"), val = tensor([1, 1])]; - tensor var_5039_pad_0 = const()[name = tensor("op_5039_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5039_dilations_0 = const()[name = tensor("op_5039_dilations_0"), val = tensor([1, 1])]; - tensor var_5039_groups_0 = const()[name = tensor("op_5039_groups_0"), val = tensor(1)]; - tensor layers_12_conv_pointwise_conv2_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(74934784))), name = tensor("layers_12_conv_pointwise_conv2_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(74925248))), shape = tensor([512, 512, 1, 1])]; - tensor var_5039_cast_fp16 = conv(dilations = var_5039_dilations_0, groups = var_5039_groups_0, pad = var_5039_pad_0, pad_type = var_5039_pad_type_0, strides = var_5039_strides_0, weight = layers_12_conv_pointwise_conv2_outlier_module_weight_to_fp16_sparsified, x = input_347_cast_fp16)[name = tensor("op_5039_cast_fp16")]; - tensor x_77_cast_fp16 = add(x = var_5033_cast_fp16, y = var_5039_cast_fp16)[name = tensor("x_77_cast_fp16")]; - tensor inputs_127_cast_fp16 = add(x = inputs_125_cast_fp16, y = x_77_cast_fp16)[name = tensor("inputs_127_cast_fp16")]; - tensor out_127_axes_0 = const()[name = tensor("out_127_axes_0"), val = tensor([1])]; - tensor var_5050_to_fp16 = const()[name = tensor("op_5050_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_127_cast_fp16 = layer_norm(axes = out_127_axes_0, epsilon = var_5050_to_fp16, x = inputs_127_cast_fp16)[name = tensor("out_127_cast_fp16")]; - tensor input_349_gamma_0_to_fp16 = const()[name = tensor("input_349_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(74967616)))]; - tensor input_349_beta_0_to_fp16 = const()[name = tensor("input_349_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(74968704)))]; - tensor input_349_epsilon_0_to_fp16 = const()[name = tensor("input_349_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_349_cast_fp16 = batch_norm(beta = input_349_beta_0_to_fp16, epsilon = input_349_epsilon_0_to_fp16, gamma = input_349_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_127_cast_fp16)[name = tensor("input_349_cast_fp16")]; - tensor var_5070_pad_type_0 = const()[name = tensor("op_5070_pad_type_0"), val = tensor("valid")]; - tensor var_5070_strides_0 = const()[name = tensor("op_5070_strides_0"), val = tensor([1, 1])]; - tensor var_5070_pad_0 = const()[name = tensor("op_5070_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5070_dilations_0 = const()[name = tensor("op_5070_dilations_0"), val = tensor([1, 1])]; - tensor var_5070_groups_0 = const()[name = tensor("op_5070_groups_0"), val = tensor(1)]; - tensor layers_12_feed_forward2_fc1_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(74969792))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(75756288))), name = tensor("layers_12_feed_forward2_fc1_inlier_module_weight_to_fp16_palettized"), shape = tensor([2048, 512, 1, 1])]; - tensor layers_12_feed_forward2_fc1_inlier_module_bias_to_fp16 = const()[name = tensor("layers_12_feed_forward2_fc1_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(75756480)))]; - tensor var_5070_cast_fp16 = conv(bias = layers_12_feed_forward2_fc1_inlier_module_bias_to_fp16, dilations = var_5070_dilations_0, groups = var_5070_groups_0, pad = var_5070_pad_0, pad_type = var_5070_pad_type_0, strides = var_5070_strides_0, weight = layers_12_feed_forward2_fc1_inlier_module_weight_to_fp16_palettized, x = input_349_cast_fp16)[name = tensor("op_5070_cast_fp16")]; - tensor var_5076_pad_type_0 = const()[name = tensor("op_5076_pad_type_0"), val = tensor("valid")]; - tensor var_5076_strides_0 = const()[name = tensor("op_5076_strides_0"), val = tensor([1, 1])]; - tensor var_5076_pad_0 = const()[name = tensor("op_5076_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5076_dilations_0 = const()[name = tensor("op_5076_dilations_0"), val = tensor([1, 1])]; - tensor var_5076_groups_0 = const()[name = tensor("op_5076_groups_0"), val = tensor(1)]; - tensor layers_12_feed_forward2_fc1_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(75797184))), name = tensor("layers_12_feed_forward2_fc1_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(75760640))), shape = tensor([2048, 512, 1, 1])]; - tensor var_5076_cast_fp16 = conv(dilations = var_5076_dilations_0, groups = var_5076_groups_0, pad = var_5076_pad_0, pad_type = var_5076_pad_type_0, strides = var_5076_strides_0, weight = layers_12_feed_forward2_fc1_outlier_module_weight_to_fp16_sparsified, x = input_349_cast_fp16)[name = tensor("op_5076_cast_fp16")]; - tensor input_351_cast_fp16 = add(x = var_5070_cast_fp16, y = var_5076_cast_fp16)[name = tensor("input_351_cast_fp16")]; - tensor input_353_cast_fp16 = silu(x = input_351_cast_fp16)[name = tensor("input_353_cast_fp16")]; - tensor var_5087_pad_type_0 = const()[name = tensor("op_5087_pad_type_0"), val = tensor("valid")]; - tensor var_5087_strides_0 = const()[name = tensor("op_5087_strides_0"), val = tensor([1, 1])]; - tensor var_5087_pad_0 = const()[name = tensor("op_5087_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5087_dilations_0 = const()[name = tensor("op_5087_dilations_0"), val = tensor([1, 1])]; - tensor var_5087_groups_0 = const()[name = tensor("op_5087_groups_0"), val = tensor(1)]; - tensor layers_12_feed_forward2_fc2_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(75928320))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(76714816))), name = tensor("layers_12_feed_forward2_fc2_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 2048, 1, 1])]; - tensor layers_12_feed_forward2_fc2_inlier_module_bias_to_fp16 = const()[name = tensor("layers_12_feed_forward2_fc2_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(76715008)))]; - tensor var_5087_cast_fp16 = conv(bias = layers_12_feed_forward2_fc2_inlier_module_bias_to_fp16, dilations = var_5087_dilations_0, groups = var_5087_groups_0, pad = var_5087_pad_0, pad_type = var_5087_pad_type_0, strides = var_5087_strides_0, weight = layers_12_feed_forward2_fc2_inlier_module_weight_to_fp16_palettized, x = input_353_cast_fp16)[name = tensor("op_5087_cast_fp16")]; - tensor var_5093_pad_type_0 = const()[name = tensor("op_5093_pad_type_0"), val = tensor("valid")]; - tensor var_5093_strides_0 = const()[name = tensor("op_5093_strides_0"), val = tensor([1, 1])]; - tensor var_5093_pad_0 = const()[name = tensor("op_5093_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5093_dilations_0 = const()[name = tensor("op_5093_dilations_0"), val = tensor([1, 1])]; - tensor var_5093_groups_0 = const()[name = tensor("op_5093_groups_0"), val = tensor(1)]; - tensor layers_12_feed_forward2_fc2_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(76758592))), name = tensor("layers_12_feed_forward2_fc2_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(76716096))), shape = tensor([512, 2048, 1, 1])]; - tensor var_5093_cast_fp16 = conv(dilations = var_5093_dilations_0, groups = var_5093_groups_0, pad = var_5093_pad_0, pad_type = var_5093_pad_type_0, strides = var_5093_strides_0, weight = layers_12_feed_forward2_fc2_outlier_module_weight_to_fp16_sparsified, x = input_353_cast_fp16)[name = tensor("op_5093_cast_fp16")]; - tensor x_79_cast_fp16 = add(x = var_5087_cast_fp16, y = var_5093_cast_fp16)[name = tensor("x_79_cast_fp16")]; - tensor var_5095_to_fp16 = const()[name = tensor("op_5095_to_fp16"), val = tensor(0x1p-1)]; - tensor var_5096_cast_fp16 = mul(x = x_79_cast_fp16, y = var_5095_to_fp16)[name = tensor("op_5096_cast_fp16")]; - tensor inputs_129_cast_fp16 = add(x = inputs_127_cast_fp16, y = var_5096_cast_fp16)[name = tensor("inputs_129_cast_fp16")]; - tensor out_129_axes_0 = const()[name = tensor("out_129_axes_0"), val = tensor([1])]; - tensor var_5106_to_fp16 = const()[name = tensor("op_5106_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_129_cast_fp16 = layer_norm(axes = out_129_axes_0, epsilon = var_5106_to_fp16, x = inputs_129_cast_fp16)[name = tensor("out_129_cast_fp16")]; - tensor inputs_131_gamma_0_to_fp16 = const()[name = tensor("inputs_131_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(76889728)))]; - tensor inputs_131_beta_0_to_fp16 = const()[name = tensor("inputs_131_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(76890816)))]; - tensor inputs_131_epsilon_0_to_fp16 = const()[name = tensor("inputs_131_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor inputs_131_cast_fp16 = batch_norm(beta = inputs_131_beta_0_to_fp16, epsilon = inputs_131_epsilon_0_to_fp16, gamma = inputs_131_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_129_cast_fp16)[name = tensor("inputs_131_cast_fp16")]; - tensor var_5120 = const()[name = tensor("op_5120"), val = tensor(3)]; - tensor out_131_axes_0 = const()[name = tensor("out_131_axes_0"), val = tensor([1])]; - tensor var_5151_to_fp16 = const()[name = tensor("op_5151_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_131_cast_fp16 = layer_norm(axes = out_131_axes_0, epsilon = var_5151_to_fp16, x = inputs_131_cast_fp16)[name = tensor("out_131_cast_fp16")]; - tensor input_355_gamma_0_to_fp16 = const()[name = tensor("input_355_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(76891904)))]; - tensor input_355_beta_0_to_fp16 = const()[name = tensor("input_355_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(76892992)))]; - tensor input_355_epsilon_0_to_fp16 = const()[name = tensor("input_355_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_355_cast_fp16 = batch_norm(beta = input_355_beta_0_to_fp16, epsilon = input_355_epsilon_0_to_fp16, gamma = input_355_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_131_cast_fp16)[name = tensor("input_355_cast_fp16")]; - tensor var_5171_pad_type_0 = const()[name = tensor("op_5171_pad_type_0"), val = tensor("valid")]; - tensor var_5171_strides_0 = const()[name = tensor("op_5171_strides_0"), val = tensor([1, 1])]; - tensor var_5171_pad_0 = const()[name = tensor("op_5171_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5171_dilations_0 = const()[name = tensor("op_5171_dilations_0"), val = tensor([1, 1])]; - tensor var_5171_groups_0 = const()[name = tensor("op_5171_groups_0"), val = tensor(1)]; - tensor layers_13_feed_forward1_fc1_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(76894080))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(77680576))), name = tensor("layers_13_feed_forward1_fc1_inlier_module_weight_to_fp16_palettized"), shape = tensor([2048, 512, 1, 1])]; - tensor layers_13_feed_forward1_fc1_inlier_module_bias_to_fp16 = const()[name = tensor("layers_13_feed_forward1_fc1_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(77680768)))]; - tensor var_5171_cast_fp16 = conv(bias = layers_13_feed_forward1_fc1_inlier_module_bias_to_fp16, dilations = var_5171_dilations_0, groups = var_5171_groups_0, pad = var_5171_pad_0, pad_type = var_5171_pad_type_0, strides = var_5171_strides_0, weight = layers_13_feed_forward1_fc1_inlier_module_weight_to_fp16_palettized, x = input_355_cast_fp16)[name = tensor("op_5171_cast_fp16")]; - tensor var_5177_pad_type_0 = const()[name = tensor("op_5177_pad_type_0"), val = tensor("valid")]; - tensor var_5177_strides_0 = const()[name = tensor("op_5177_strides_0"), val = tensor([1, 1])]; - tensor var_5177_pad_0 = const()[name = tensor("op_5177_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5177_dilations_0 = const()[name = tensor("op_5177_dilations_0"), val = tensor([1, 1])]; - tensor var_5177_groups_0 = const()[name = tensor("op_5177_groups_0"), val = tensor(1)]; - tensor layers_13_feed_forward1_fc1_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(77722240))), name = tensor("layers_13_feed_forward1_fc1_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(77684928))), shape = tensor([2048, 512, 1, 1])]; - tensor var_5177_cast_fp16 = conv(dilations = var_5177_dilations_0, groups = var_5177_groups_0, pad = var_5177_pad_0, pad_type = var_5177_pad_type_0, strides = var_5177_strides_0, weight = layers_13_feed_forward1_fc1_outlier_module_weight_to_fp16_sparsified, x = input_355_cast_fp16)[name = tensor("op_5177_cast_fp16")]; - tensor input_357_cast_fp16 = add(x = var_5171_cast_fp16, y = var_5177_cast_fp16)[name = tensor("input_357_cast_fp16")]; - tensor input_359_cast_fp16 = silu(x = input_357_cast_fp16)[name = tensor("input_359_cast_fp16")]; - tensor var_5188_pad_type_0 = const()[name = tensor("op_5188_pad_type_0"), val = tensor("valid")]; - tensor var_5188_strides_0 = const()[name = tensor("op_5188_strides_0"), val = tensor([1, 1])]; - tensor var_5188_pad_0 = const()[name = tensor("op_5188_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5188_dilations_0 = const()[name = tensor("op_5188_dilations_0"), val = tensor([1, 1])]; - tensor var_5188_groups_0 = const()[name = tensor("op_5188_groups_0"), val = tensor(1)]; - tensor layers_13_feed_forward1_fc2_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(77853376))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(78639872))), name = tensor("layers_13_feed_forward1_fc2_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 2048, 1, 1])]; - tensor layers_13_feed_forward1_fc2_inlier_module_bias_to_fp16 = const()[name = tensor("layers_13_feed_forward1_fc2_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(78640064)))]; - tensor var_5188_cast_fp16 = conv(bias = layers_13_feed_forward1_fc2_inlier_module_bias_to_fp16, dilations = var_5188_dilations_0, groups = var_5188_groups_0, pad = var_5188_pad_0, pad_type = var_5188_pad_type_0, strides = var_5188_strides_0, weight = layers_13_feed_forward1_fc2_inlier_module_weight_to_fp16_palettized, x = input_359_cast_fp16)[name = tensor("op_5188_cast_fp16")]; - tensor var_5194_pad_type_0 = const()[name = tensor("op_5194_pad_type_0"), val = tensor("valid")]; - tensor var_5194_strides_0 = const()[name = tensor("op_5194_strides_0"), val = tensor([1, 1])]; - tensor var_5194_pad_0 = const()[name = tensor("op_5194_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5194_dilations_0 = const()[name = tensor("op_5194_dilations_0"), val = tensor([1, 1])]; - tensor var_5194_groups_0 = const()[name = tensor("op_5194_groups_0"), val = tensor(1)]; - tensor layers_13_feed_forward1_fc2_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(78683136))), name = tensor("layers_13_feed_forward1_fc2_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(78641152))), shape = tensor([512, 2048, 1, 1])]; - tensor var_5194_cast_fp16 = conv(dilations = var_5194_dilations_0, groups = var_5194_groups_0, pad = var_5194_pad_0, pad_type = var_5194_pad_type_0, strides = var_5194_strides_0, weight = layers_13_feed_forward1_fc2_outlier_module_weight_to_fp16_sparsified, x = input_359_cast_fp16)[name = tensor("op_5194_cast_fp16")]; - tensor x_81_cast_fp16 = add(x = var_5188_cast_fp16, y = var_5194_cast_fp16)[name = tensor("x_81_cast_fp16")]; - tensor var_5196_to_fp16 = const()[name = tensor("op_5196_to_fp16"), val = tensor(0x1p-1)]; - tensor var_5197_cast_fp16 = mul(x = x_81_cast_fp16, y = var_5196_to_fp16)[name = tensor("op_5197_cast_fp16")]; - tensor inputs_133_cast_fp16 = add(x = inputs_131_cast_fp16, y = var_5197_cast_fp16)[name = tensor("inputs_133_cast_fp16")]; - tensor out_133_axes_0 = const()[name = tensor("out_133_axes_0"), val = tensor([1])]; - tensor var_5207_to_fp16 = const()[name = tensor("op_5207_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_133_cast_fp16 = layer_norm(axes = out_133_axes_0, epsilon = var_5207_to_fp16, x = inputs_133_cast_fp16)[name = tensor("out_133_cast_fp16")]; - tensor obj_55_gamma_0_to_fp16 = const()[name = tensor("obj_55_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(78814272)))]; - tensor obj_55_beta_0_to_fp16 = const()[name = tensor("obj_55_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(78815360)))]; - tensor obj_55_epsilon_0_to_fp16 = const()[name = tensor("obj_55_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor obj_55_cast_fp16 = batch_norm(beta = obj_55_beta_0_to_fp16, epsilon = obj_55_epsilon_0_to_fp16, gamma = obj_55_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_133_cast_fp16)[name = tensor("obj_55_cast_fp16")]; - tensor var_5232_pad_type_0 = const()[name = tensor("op_5232_pad_type_0"), val = tensor("valid")]; - tensor var_5232_strides_0 = const()[name = tensor("op_5232_strides_0"), val = tensor([1, 1])]; - tensor var_5232_pad_0 = const()[name = tensor("op_5232_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5232_dilations_0 = const()[name = tensor("op_5232_dilations_0"), val = tensor([1, 1])]; - tensor var_5232_groups_0 = const()[name = tensor("op_5232_groups_0"), val = tensor(1)]; - tensor layers_13_self_attn_q_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(78816448))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(79013120))), name = tensor("layers_13_self_attn_q_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_13_self_attn_q_proj_inlier_module_bias_to_fp16 = const()[name = tensor("layers_13_self_attn_q_proj_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(79013312)))]; - tensor var_5232_cast_fp16 = conv(bias = layers_13_self_attn_q_proj_inlier_module_bias_to_fp16, dilations = var_5232_dilations_0, groups = var_5232_groups_0, pad = var_5232_pad_0, pad_type = var_5232_pad_type_0, strides = var_5232_strides_0, weight = layers_13_self_attn_q_proj_inlier_module_weight_to_fp16_palettized, x = obj_55_cast_fp16)[name = tensor("op_5232_cast_fp16")]; - tensor var_5238_pad_type_0 = const()[name = tensor("op_5238_pad_type_0"), val = tensor("valid")]; - tensor var_5238_strides_0 = const()[name = tensor("op_5238_strides_0"), val = tensor([1, 1])]; - tensor var_5238_pad_0 = const()[name = tensor("op_5238_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5238_dilations_0 = const()[name = tensor("op_5238_dilations_0"), val = tensor([1, 1])]; - tensor var_5238_groups_0 = const()[name = tensor("op_5238_groups_0"), val = tensor(1)]; - tensor layers_13_self_attn_q_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(79024320))), name = tensor("layers_13_self_attn_q_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(79014400))), shape = tensor([512, 512, 1, 1])]; - tensor var_5238_cast_fp16 = conv(dilations = var_5238_dilations_0, groups = var_5238_groups_0, pad = var_5238_pad_0, pad_type = var_5238_pad_type_0, strides = var_5238_strides_0, weight = layers_13_self_attn_q_proj_outlier_module_weight_to_fp16_sparsified, x = obj_55_cast_fp16)[name = tensor("op_5238_cast_fp16")]; - tensor query_53_cast_fp16 = add(x = var_5232_cast_fp16, y = var_5238_cast_fp16)[name = tensor("query_53_cast_fp16")]; - tensor var_5247_pad_type_0 = const()[name = tensor("op_5247_pad_type_0"), val = tensor("valid")]; - tensor var_5247_strides_0 = const()[name = tensor("op_5247_strides_0"), val = tensor([1, 1])]; - tensor var_5247_pad_0 = const()[name = tensor("op_5247_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5247_dilations_0 = const()[name = tensor("op_5247_dilations_0"), val = tensor([1, 1])]; - tensor var_5247_groups_0 = const()[name = tensor("op_5247_groups_0"), val = tensor(1)]; - tensor layers_13_self_attn_k_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(79057152))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(79253824))), name = tensor("layers_13_self_attn_k_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor var_5247_cast_fp16 = conv(dilations = var_5247_dilations_0, groups = var_5247_groups_0, pad = var_5247_pad_0, pad_type = var_5247_pad_type_0, strides = var_5247_strides_0, weight = layers_13_self_attn_k_proj_inlier_module_weight_to_fp16_palettized, x = obj_55_cast_fp16)[name = tensor("op_5247_cast_fp16")]; - tensor var_5253_pad_type_0 = const()[name = tensor("op_5253_pad_type_0"), val = tensor("valid")]; - tensor var_5253_strides_0 = const()[name = tensor("op_5253_strides_0"), val = tensor([1, 1])]; - tensor var_5253_pad_0 = const()[name = tensor("op_5253_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5253_dilations_0 = const()[name = tensor("op_5253_dilations_0"), val = tensor([1, 1])]; - tensor var_5253_groups_0 = const()[name = tensor("op_5253_groups_0"), val = tensor(1)]; - tensor layers_13_self_attn_k_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(79263680))), name = tensor("layers_13_self_attn_k_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(79254016))), shape = tensor([512, 512, 1, 1])]; - tensor var_5253_cast_fp16 = conv(dilations = var_5253_dilations_0, groups = var_5253_groups_0, pad = var_5253_pad_0, pad_type = var_5253_pad_type_0, strides = var_5253_strides_0, weight = layers_13_self_attn_k_proj_outlier_module_weight_to_fp16_sparsified, x = obj_55_cast_fp16)[name = tensor("op_5253_cast_fp16")]; - tensor key_27_cast_fp16 = add(x = var_5247_cast_fp16, y = var_5253_cast_fp16)[name = tensor("key_27_cast_fp16")]; - tensor var_5263_pad_type_0 = const()[name = tensor("op_5263_pad_type_0"), val = tensor("valid")]; - tensor var_5263_strides_0 = const()[name = tensor("op_5263_strides_0"), val = tensor([1, 1])]; - tensor var_5263_pad_0 = const()[name = tensor("op_5263_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5263_dilations_0 = const()[name = tensor("op_5263_dilations_0"), val = tensor([1, 1])]; - tensor var_5263_groups_0 = const()[name = tensor("op_5263_groups_0"), val = tensor(1)]; - tensor layers_13_self_attn_v_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(79296512))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(79493184))), name = tensor("layers_13_self_attn_v_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_13_self_attn_v_proj_inlier_module_bias_to_fp16 = const()[name = tensor("layers_13_self_attn_v_proj_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(79493376)))]; - tensor var_5263_cast_fp16 = conv(bias = layers_13_self_attn_v_proj_inlier_module_bias_to_fp16, dilations = var_5263_dilations_0, groups = var_5263_groups_0, pad = var_5263_pad_0, pad_type = var_5263_pad_type_0, strides = var_5263_strides_0, weight = layers_13_self_attn_v_proj_inlier_module_weight_to_fp16_palettized, x = obj_55_cast_fp16)[name = tensor("op_5263_cast_fp16")]; - tensor var_5269_pad_type_0 = const()[name = tensor("op_5269_pad_type_0"), val = tensor("valid")]; - tensor var_5269_strides_0 = const()[name = tensor("op_5269_strides_0"), val = tensor([1, 1])]; - tensor var_5269_pad_0 = const()[name = tensor("op_5269_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5269_dilations_0 = const()[name = tensor("op_5269_dilations_0"), val = tensor([1, 1])]; - tensor var_5269_groups_0 = const()[name = tensor("op_5269_groups_0"), val = tensor(1)]; - tensor layers_13_self_attn_v_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(79502784))), name = tensor("layers_13_self_attn_v_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(79494464))), shape = tensor([512, 512, 1, 1])]; - tensor var_5269_cast_fp16 = conv(dilations = var_5269_dilations_0, groups = var_5269_groups_0, pad = var_5269_pad_0, pad_type = var_5269_pad_type_0, strides = var_5269_strides_0, weight = layers_13_self_attn_v_proj_outlier_module_weight_to_fp16_sparsified, x = obj_55_cast_fp16)[name = tensor("op_5269_cast_fp16")]; - tensor value_27_cast_fp16 = add(x = var_5263_cast_fp16, y = var_5269_cast_fp16)[name = tensor("value_27_cast_fp16")]; - tensor var_5272_to_fp16 = const()[name = tensor("op_5272_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(79535616)))]; - tensor query_55_cast_fp16 = add(x = query_53_cast_fp16, y = var_5272_to_fp16)[name = tensor("query_55_cast_fp16")]; - tensor var_5275_to_fp16 = const()[name = tensor("op_5275_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(79536704)))]; - tensor q_with_bias_v_27_cast_fp16 = add(x = query_53_cast_fp16, y = var_5275_to_fp16)[name = tensor("q_with_bias_v_27_cast_fp16")]; - tensor var_5285_pad_type_0 = const()[name = tensor("op_5285_pad_type_0"), val = tensor("valid")]; - tensor var_5285_strides_0 = const()[name = tensor("op_5285_strides_0"), val = tensor([1, 1])]; - tensor var_5285_pad_0 = const()[name = tensor("op_5285_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5285_dilations_0 = const()[name = tensor("op_5285_dilations_0"), val = tensor([1, 1])]; - tensor var_5285_groups_0 = const()[name = tensor("op_5285_groups_0"), val = tensor(1)]; - tensor layers_13_self_attn_linear_pos_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(79537792))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(79734464))), name = tensor("layers_13_self_attn_linear_pos_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor var_5285_cast_fp16 = conv(bias = input_17_mean_0_to_fp16, dilations = var_5285_dilations_0, groups = var_5285_groups_0, pad = var_5285_pad_0, pad_type = var_5285_pad_type_0, strides = var_5285_strides_0, weight = layers_13_self_attn_linear_pos_inlier_module_weight_to_fp16_palettized, x = obj_3_cast_fp16)[name = tensor("op_5285_cast_fp16")]; - tensor var_5291_pad_type_0 = const()[name = tensor("op_5291_pad_type_0"), val = tensor("valid")]; - tensor var_5291_strides_0 = const()[name = tensor("op_5291_strides_0"), val = tensor([1, 1])]; - tensor var_5291_pad_0 = const()[name = tensor("op_5291_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5291_dilations_0 = const()[name = tensor("op_5291_dilations_0"), val = tensor([1, 1])]; - tensor var_5291_groups_0 = const()[name = tensor("op_5291_groups_0"), val = tensor(1)]; - tensor layers_13_self_attn_linear_pos_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(79756544))), name = tensor("layers_13_self_attn_linear_pos_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(79734656))), shape = tensor([512, 512, 1, 1])]; - tensor var_5291_cast_fp16 = conv(dilations = var_5291_dilations_0, groups = var_5291_groups_0, pad = var_5291_pad_0, pad_type = var_5291_pad_type_0, strides = var_5291_strides_0, weight = layers_13_self_attn_linear_pos_outlier_module_weight_to_fp16_sparsified, x = obj_3_cast_fp16)[name = tensor("op_5291_cast_fp16")]; - tensor p_27_cast_fp16 = add(x = var_5285_cast_fp16, y = var_5291_cast_fp16)[name = tensor("p_27_cast_fp16")]; - tensor var_5295 = const()[name = tensor("op_5295"), val = tensor([1, 8, 64, 188])]; - tensor var_5296_cast_fp16 = reshape(shape = var_5295, x = q_with_bias_v_27_cast_fp16)[name = tensor("op_5296_cast_fp16")]; - tensor var_5297 = const()[name = tensor("op_5297"), val = tensor([1, 8, 64, -1])]; - tensor var_5298_cast_fp16 = reshape(shape = var_5297, x = p_27_cast_fp16)[name = tensor("op_5298_cast_fp16")]; - tensor matrix_bd_105_transpose_x_0 = const()[name = tensor("matrix_bd_105_transpose_x_0"), val = tensor(true)]; - tensor matrix_bd_105_transpose_y_0 = const()[name = tensor("matrix_bd_105_transpose_y_0"), val = tensor(false)]; - tensor matrix_bd_105_cast_fp16 = matmul(transpose_x = matrix_bd_105_transpose_x_0, transpose_y = matrix_bd_105_transpose_y_0, x = var_5296_cast_fp16, y = var_5298_cast_fp16)[name = tensor("matrix_bd_105_cast_fp16")]; - tensor matrix_bd_107_pad_0 = const()[name = tensor("matrix_bd_107_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor matrix_bd_107_mode_0 = const()[name = tensor("matrix_bd_107_mode_0"), val = tensor("constant")]; - tensor const_153_to_fp16 = const()[name = tensor("const_153_to_fp16"), val = tensor(0x0p+0)]; - tensor matrix_bd_107_cast_fp16 = pad(constant_val = const_153_to_fp16, mode = matrix_bd_107_mode_0, pad = matrix_bd_107_pad_0, x = matrix_bd_105_cast_fp16)[name = tensor("matrix_bd_107_cast_fp16")]; - tensor var_5307 = const()[name = tensor("op_5307"), val = tensor([1, 8, -1, 188])]; - tensor matrix_bd_109_cast_fp16 = reshape(shape = var_5307, x = matrix_bd_107_cast_fp16)[name = tensor("matrix_bd_109_cast_fp16")]; - tensor var_5311_begin_0 = const()[name = tensor("op_5311_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_5311_end_0 = const()[name = tensor("op_5311_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_5311_end_mask_0 = const()[name = tensor("op_5311_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_5311_cast_fp16 = slice_by_index(begin = var_5311_begin_0, end = var_5311_end_0, end_mask = var_5311_end_mask_0, x = matrix_bd_109_cast_fp16)[name = tensor("op_5311_cast_fp16")]; - tensor var_5312 = const()[name = tensor("op_5312"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_111_cast_fp16 = reshape(shape = var_5312, x = var_5311_cast_fp16)[name = tensor("matrix_bd_111_cast_fp16")]; - tensor var_5317_begin_0 = const()[name = tensor("op_5317_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5317_end_0 = const()[name = tensor("op_5317_end_0"), val = tensor([1, 8, 188, 188])]; - tensor var_5317_end_mask_0 = const()[name = tensor("op_5317_end_mask_0"), val = tensor([true, true, true, false])]; - tensor var_5317_cast_fp16 = slice_by_index(begin = var_5317_begin_0, end = var_5317_end_0, end_mask = var_5317_end_mask_0, x = matrix_bd_111_cast_fp16)[name = tensor("op_5317_cast_fp16")]; - tensor var_5318_to_fp16 = const()[name = tensor("op_5318_to_fp16"), val = tensor(0x1p-3)]; - tensor qk_mask_27_cast_fp16 = mul(x = var_5317_cast_fp16, y = var_5318_to_fp16)[name = tensor("qk_mask_27_cast_fp16")]; - tensor var_5322 = const()[name = tensor("op_5322"), val = tensor([1, 8, 64, 188])]; - tensor mh_q_27_cast_fp16 = reshape(shape = var_5322, x = query_55_cast_fp16)[name = tensor("mh_q_27_cast_fp16")]; - tensor var_5324_to_fp16 = const()[name = tensor("op_5324_to_fp16"), val = tensor(0x1p-3)]; - tensor var_5325_cast_fp16 = mul(x = mh_q_27_cast_fp16, y = var_5324_to_fp16)[name = tensor("op_5325_cast_fp16")]; - tensor var_5328 = const()[name = tensor("op_5328"), val = tensor([1, 8, 64, 188])]; - tensor var_5329_cast_fp16 = reshape(shape = var_5328, x = key_27_cast_fp16)[name = tensor("op_5329_cast_fp16")]; - tensor mh_w_53_transpose_x_0 = const()[name = tensor("mh_w_53_transpose_x_0"), val = tensor(true)]; - tensor mh_w_53_transpose_y_0 = const()[name = tensor("mh_w_53_transpose_y_0"), val = tensor(false)]; - tensor mh_w_53_cast_fp16 = matmul(transpose_x = mh_w_53_transpose_x_0, transpose_y = mh_w_53_transpose_y_0, x = var_5325_cast_fp16, y = var_5329_cast_fp16)[name = tensor("mh_w_53_cast_fp16")]; - tensor mh_w_55_cast_fp16 = add(x = mh_w_53_cast_fp16, y = qk_mask_27_cast_fp16)[name = tensor("mh_w_55_cast_fp16")]; - tensor var_5333_cast_fp16 = softmax(axis = var_5120, x = mh_w_55_cast_fp16)[name = tensor("op_5333_cast_fp16")]; - tensor var_5334 = const()[name = tensor("op_5334"), val = tensor([1, 8, 64, 188])]; - tensor var_5335_cast_fp16 = reshape(shape = var_5334, x = value_27_cast_fp16)[name = tensor("op_5335_cast_fp16")]; - tensor attn_27_transpose_x_0 = const()[name = tensor("attn_27_transpose_x_0"), val = tensor(false)]; - tensor attn_27_transpose_y_0 = const()[name = tensor("attn_27_transpose_y_0"), val = tensor(true)]; - tensor attn_27_cast_fp16 = matmul(transpose_x = attn_27_transpose_x_0, transpose_y = attn_27_transpose_y_0, x = var_5335_cast_fp16, y = var_5333_cast_fp16)[name = tensor("attn_27_cast_fp16")]; - tensor var_5338 = const()[name = tensor("op_5338"), val = tensor([1, 512, 1, 188])]; - tensor input_361_cast_fp16 = reshape(shape = var_5338, x = attn_27_cast_fp16)[name = tensor("input_361_cast_fp16")]; - tensor var_5348_pad_type_0 = const()[name = tensor("op_5348_pad_type_0"), val = tensor("valid")]; - tensor var_5348_strides_0 = const()[name = tensor("op_5348_strides_0"), val = tensor([1, 1])]; - tensor var_5348_pad_0 = const()[name = tensor("op_5348_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5348_dilations_0 = const()[name = tensor("op_5348_dilations_0"), val = tensor([1, 1])]; - tensor var_5348_groups_0 = const()[name = tensor("op_5348_groups_0"), val = tensor(1)]; - tensor layers_13_self_attn_o_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(79789376))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(79986048))), name = tensor("layers_13_self_attn_o_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_13_self_attn_o_proj_inlier_module_bias_to_fp16 = const()[name = tensor("layers_13_self_attn_o_proj_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(79986240)))]; - tensor var_5348_cast_fp16 = conv(bias = layers_13_self_attn_o_proj_inlier_module_bias_to_fp16, dilations = var_5348_dilations_0, groups = var_5348_groups_0, pad = var_5348_pad_0, pad_type = var_5348_pad_type_0, strides = var_5348_strides_0, weight = layers_13_self_attn_o_proj_inlier_module_weight_to_fp16_palettized, x = input_361_cast_fp16)[name = tensor("op_5348_cast_fp16")]; - tensor var_5354_pad_type_0 = const()[name = tensor("op_5354_pad_type_0"), val = tensor("valid")]; - tensor var_5354_strides_0 = const()[name = tensor("op_5354_strides_0"), val = tensor([1, 1])]; - tensor var_5354_pad_0 = const()[name = tensor("op_5354_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5354_dilations_0 = const()[name = tensor("op_5354_dilations_0"), val = tensor([1, 1])]; - tensor var_5354_groups_0 = const()[name = tensor("op_5354_groups_0"), val = tensor(1)]; - tensor layers_13_self_attn_o_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(79996608))), name = tensor("layers_13_self_attn_o_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(79987328))), shape = tensor([512, 512, 1, 1])]; - tensor var_5354_cast_fp16 = conv(dilations = var_5354_dilations_0, groups = var_5354_groups_0, pad = var_5354_pad_0, pad_type = var_5354_pad_type_0, strides = var_5354_strides_0, weight = layers_13_self_attn_o_proj_outlier_module_weight_to_fp16_sparsified, x = input_361_cast_fp16)[name = tensor("op_5354_cast_fp16")]; - tensor obj_57_cast_fp16 = add(x = var_5348_cast_fp16, y = var_5354_cast_fp16)[name = tensor("obj_57_cast_fp16")]; - tensor inputs_135_cast_fp16 = add(x = inputs_133_cast_fp16, y = obj_57_cast_fp16)[name = tensor("inputs_135_cast_fp16")]; - tensor out_135_axes_0 = const()[name = tensor("out_135_axes_0"), val = tensor([1])]; - tensor var_5365_to_fp16 = const()[name = tensor("op_5365_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_135_cast_fp16 = layer_norm(axes = out_135_axes_0, epsilon = var_5365_to_fp16, x = inputs_135_cast_fp16)[name = tensor("out_135_cast_fp16")]; - tensor input_363_gamma_0_to_fp16 = const()[name = tensor("input_363_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(80029440)))]; - tensor input_363_beta_0_to_fp16 = const()[name = tensor("input_363_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(80030528)))]; - tensor input_363_epsilon_0_to_fp16 = const()[name = tensor("input_363_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_363_cast_fp16 = batch_norm(beta = input_363_beta_0_to_fp16, epsilon = input_363_epsilon_0_to_fp16, gamma = input_363_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_135_cast_fp16)[name = tensor("input_363_cast_fp16")]; - tensor var_5387_pad_type_0 = const()[name = tensor("op_5387_pad_type_0"), val = tensor("valid")]; - tensor var_5387_strides_0 = const()[name = tensor("op_5387_strides_0"), val = tensor([1, 1])]; - tensor var_5387_pad_0 = const()[name = tensor("op_5387_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5387_dilations_0 = const()[name = tensor("op_5387_dilations_0"), val = tensor([1, 1])]; - tensor var_5387_groups_0 = const()[name = tensor("op_5387_groups_0"), val = tensor(1)]; - tensor layers_13_conv_pointwise_conv1_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(80031616))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(80424896))), name = tensor("layers_13_conv_pointwise_conv1_inlier_module_weight_to_fp16_palettized"), shape = tensor([1024, 512, 1, 1])]; - tensor layers_13_conv_pointwise_conv1_inlier_module_bias_to_fp16 = const()[name = tensor("layers_13_conv_pointwise_conv1_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(80425088)))]; - tensor var_5387_cast_fp16 = conv(bias = layers_13_conv_pointwise_conv1_inlier_module_bias_to_fp16, dilations = var_5387_dilations_0, groups = var_5387_groups_0, pad = var_5387_pad_0, pad_type = var_5387_pad_type_0, strides = var_5387_strides_0, weight = layers_13_conv_pointwise_conv1_inlier_module_weight_to_fp16_palettized, x = input_363_cast_fp16)[name = tensor("op_5387_cast_fp16")]; - tensor var_5393_pad_type_0 = const()[name = tensor("op_5393_pad_type_0"), val = tensor("valid")]; - tensor var_5393_strides_0 = const()[name = tensor("op_5393_strides_0"), val = tensor([1, 1])]; - tensor var_5393_pad_0 = const()[name = tensor("op_5393_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5393_dilations_0 = const()[name = tensor("op_5393_dilations_0"), val = tensor([1, 1])]; - tensor var_5393_groups_0 = const()[name = tensor("op_5393_groups_0"), val = tensor(1)]; - tensor layers_13_conv_pointwise_conv1_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(80446784))), name = tensor("layers_13_conv_pointwise_conv1_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(80427200))), shape = tensor([1024, 512, 1, 1])]; - tensor var_5393_cast_fp16 = conv(dilations = var_5393_dilations_0, groups = var_5393_groups_0, pad = var_5393_pad_0, pad_type = var_5393_pad_type_0, strides = var_5393_strides_0, weight = layers_13_conv_pointwise_conv1_outlier_module_weight_to_fp16_sparsified, x = input_363_cast_fp16)[name = tensor("op_5393_cast_fp16")]; - tensor input_365_cast_fp16 = add(x = var_5387_cast_fp16, y = var_5393_cast_fp16)[name = tensor("input_365_cast_fp16")]; - tensor input_367_split_num_splits_0 = const()[name = tensor("input_367_split_num_splits_0"), val = tensor(2)]; - tensor input_367_split_axis_0 = const()[name = tensor("input_367_split_axis_0"), val = tensor(1)]; - tensor input_367_split_cast_fp16_0, tensor input_367_split_cast_fp16_1 = split(axis = input_367_split_axis_0, num_splits = input_367_split_num_splits_0, x = input_365_cast_fp16)[name = tensor("input_367_split_cast_fp16")]; - tensor input_367_split_1_sigmoid_cast_fp16 = sigmoid(x = input_367_split_cast_fp16_1)[name = tensor("input_367_split_1_sigmoid_cast_fp16")]; - tensor input_367_cast_fp16 = mul(x = input_367_split_cast_fp16_0, y = input_367_split_1_sigmoid_cast_fp16)[name = tensor("input_367_cast_fp16")]; - tensor input_369_pad_type_0 = const()[name = tensor("input_369_pad_type_0"), val = tensor("custom")]; - tensor input_369_pad_0 = const()[name = tensor("input_369_pad_0"), val = tensor([0, 0, 4, 4])]; - tensor input_369_groups_0 = const()[name = tensor("input_369_groups_0"), val = tensor(512)]; - tensor input_369_strides_0 = const()[name = tensor("input_369_strides_0"), val = tensor([1, 1])]; - tensor input_369_dilations_0 = const()[name = tensor("input_369_dilations_0"), val = tensor([1, 1])]; - tensor const_217_to_fp16 = const()[name = tensor("const_217_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(80512384)))]; - tensor const_218_to_fp16 = const()[name = tensor("const_218_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(80521664)))]; - tensor input_371_cast_fp16 = conv(bias = const_218_to_fp16, dilations = input_369_dilations_0, groups = input_369_groups_0, pad = input_369_pad_0, pad_type = input_369_pad_type_0, strides = input_369_strides_0, weight = const_217_to_fp16, x = input_367_cast_fp16)[name = tensor("input_371_cast_fp16")]; - tensor input_373_cast_fp16 = silu(x = input_371_cast_fp16)[name = tensor("input_373_cast_fp16")]; - tensor var_5417_pad_type_0 = const()[name = tensor("op_5417_pad_type_0"), val = tensor("valid")]; - tensor var_5417_strides_0 = const()[name = tensor("op_5417_strides_0"), val = tensor([1, 1])]; - tensor var_5417_pad_0 = const()[name = tensor("op_5417_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5417_dilations_0 = const()[name = tensor("op_5417_dilations_0"), val = tensor([1, 1])]; - tensor var_5417_groups_0 = const()[name = tensor("op_5417_groups_0"), val = tensor(1)]; - tensor layers_13_conv_pointwise_conv2_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(80522752))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(80719424))), name = tensor("layers_13_conv_pointwise_conv2_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_13_conv_pointwise_conv2_inlier_module_bias_to_fp16 = const()[name = tensor("layers_13_conv_pointwise_conv2_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(80719616)))]; - tensor var_5417_cast_fp16 = conv(bias = layers_13_conv_pointwise_conv2_inlier_module_bias_to_fp16, dilations = var_5417_dilations_0, groups = var_5417_groups_0, pad = var_5417_pad_0, pad_type = var_5417_pad_type_0, strides = var_5417_strides_0, weight = layers_13_conv_pointwise_conv2_inlier_module_weight_to_fp16_palettized, x = input_373_cast_fp16)[name = tensor("op_5417_cast_fp16")]; - tensor var_5423_pad_type_0 = const()[name = tensor("op_5423_pad_type_0"), val = tensor("valid")]; - tensor var_5423_strides_0 = const()[name = tensor("op_5423_strides_0"), val = tensor([1, 1])]; - tensor var_5423_pad_0 = const()[name = tensor("op_5423_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5423_dilations_0 = const()[name = tensor("op_5423_dilations_0"), val = tensor([1, 1])]; - tensor var_5423_groups_0 = const()[name = tensor("op_5423_groups_0"), val = tensor(1)]; - tensor layers_13_conv_pointwise_conv2_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(80730112))), name = tensor("layers_13_conv_pointwise_conv2_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(80720704))), shape = tensor([512, 512, 1, 1])]; - tensor var_5423_cast_fp16 = conv(dilations = var_5423_dilations_0, groups = var_5423_groups_0, pad = var_5423_pad_0, pad_type = var_5423_pad_type_0, strides = var_5423_strides_0, weight = layers_13_conv_pointwise_conv2_outlier_module_weight_to_fp16_sparsified, x = input_373_cast_fp16)[name = tensor("op_5423_cast_fp16")]; - tensor x_83_cast_fp16 = add(x = var_5417_cast_fp16, y = var_5423_cast_fp16)[name = tensor("x_83_cast_fp16")]; - tensor inputs_137_cast_fp16 = add(x = inputs_135_cast_fp16, y = x_83_cast_fp16)[name = tensor("inputs_137_cast_fp16")]; - tensor out_137_axes_0 = const()[name = tensor("out_137_axes_0"), val = tensor([1])]; - tensor var_5434_to_fp16 = const()[name = tensor("op_5434_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_137_cast_fp16 = layer_norm(axes = out_137_axes_0, epsilon = var_5434_to_fp16, x = inputs_137_cast_fp16)[name = tensor("out_137_cast_fp16")]; - tensor input_375_gamma_0_to_fp16 = const()[name = tensor("input_375_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(80762944)))]; - tensor input_375_beta_0_to_fp16 = const()[name = tensor("input_375_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(80764032)))]; - tensor input_375_epsilon_0_to_fp16 = const()[name = tensor("input_375_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_375_cast_fp16 = batch_norm(beta = input_375_beta_0_to_fp16, epsilon = input_375_epsilon_0_to_fp16, gamma = input_375_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_137_cast_fp16)[name = tensor("input_375_cast_fp16")]; - tensor var_5454_pad_type_0 = const()[name = tensor("op_5454_pad_type_0"), val = tensor("valid")]; - tensor var_5454_strides_0 = const()[name = tensor("op_5454_strides_0"), val = tensor([1, 1])]; - tensor var_5454_pad_0 = const()[name = tensor("op_5454_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5454_dilations_0 = const()[name = tensor("op_5454_dilations_0"), val = tensor([1, 1])]; - tensor var_5454_groups_0 = const()[name = tensor("op_5454_groups_0"), val = tensor(1)]; - tensor layers_13_feed_forward2_fc1_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(80765120))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(81551616))), name = tensor("layers_13_feed_forward2_fc1_inlier_module_weight_to_fp16_palettized"), shape = tensor([2048, 512, 1, 1])]; - tensor layers_13_feed_forward2_fc1_inlier_module_bias_to_fp16 = const()[name = tensor("layers_13_feed_forward2_fc1_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(81551808)))]; - tensor var_5454_cast_fp16 = conv(bias = layers_13_feed_forward2_fc1_inlier_module_bias_to_fp16, dilations = var_5454_dilations_0, groups = var_5454_groups_0, pad = var_5454_pad_0, pad_type = var_5454_pad_type_0, strides = var_5454_strides_0, weight = layers_13_feed_forward2_fc1_inlier_module_weight_to_fp16_palettized, x = input_375_cast_fp16)[name = tensor("op_5454_cast_fp16")]; - tensor var_5460_pad_type_0 = const()[name = tensor("op_5460_pad_type_0"), val = tensor("valid")]; - tensor var_5460_strides_0 = const()[name = tensor("op_5460_strides_0"), val = tensor([1, 1])]; - tensor var_5460_pad_0 = const()[name = tensor("op_5460_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5460_dilations_0 = const()[name = tensor("op_5460_dilations_0"), val = tensor([1, 1])]; - tensor var_5460_groups_0 = const()[name = tensor("op_5460_groups_0"), val = tensor(1)]; - tensor layers_13_feed_forward2_fc1_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(81591104))), name = tensor("layers_13_feed_forward2_fc1_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(81555968))), shape = tensor([2048, 512, 1, 1])]; - tensor var_5460_cast_fp16 = conv(dilations = var_5460_dilations_0, groups = var_5460_groups_0, pad = var_5460_pad_0, pad_type = var_5460_pad_type_0, strides = var_5460_strides_0, weight = layers_13_feed_forward2_fc1_outlier_module_weight_to_fp16_sparsified, x = input_375_cast_fp16)[name = tensor("op_5460_cast_fp16")]; - tensor input_377_cast_fp16 = add(x = var_5454_cast_fp16, y = var_5460_cast_fp16)[name = tensor("input_377_cast_fp16")]; - tensor input_379_cast_fp16 = silu(x = input_377_cast_fp16)[name = tensor("input_379_cast_fp16")]; - tensor var_5471_pad_type_0 = const()[name = tensor("op_5471_pad_type_0"), val = tensor("valid")]; - tensor var_5471_strides_0 = const()[name = tensor("op_5471_strides_0"), val = tensor([1, 1])]; - tensor var_5471_pad_0 = const()[name = tensor("op_5471_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5471_dilations_0 = const()[name = tensor("op_5471_dilations_0"), val = tensor([1, 1])]; - tensor var_5471_groups_0 = const()[name = tensor("op_5471_groups_0"), val = tensor(1)]; - tensor layers_13_feed_forward2_fc2_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(81722240))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82508736))), name = tensor("layers_13_feed_forward2_fc2_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 2048, 1, 1])]; - tensor layers_13_feed_forward2_fc2_inlier_module_bias_to_fp16 = const()[name = tensor("layers_13_feed_forward2_fc2_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82508928)))]; - tensor var_5471_cast_fp16 = conv(bias = layers_13_feed_forward2_fc2_inlier_module_bias_to_fp16, dilations = var_5471_dilations_0, groups = var_5471_groups_0, pad = var_5471_pad_0, pad_type = var_5471_pad_type_0, strides = var_5471_strides_0, weight = layers_13_feed_forward2_fc2_inlier_module_weight_to_fp16_palettized, x = input_379_cast_fp16)[name = tensor("op_5471_cast_fp16")]; - tensor var_5477_pad_type_0 = const()[name = tensor("op_5477_pad_type_0"), val = tensor("valid")]; - tensor var_5477_strides_0 = const()[name = tensor("op_5477_strides_0"), val = tensor([1, 1])]; - tensor var_5477_pad_0 = const()[name = tensor("op_5477_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5477_dilations_0 = const()[name = tensor("op_5477_dilations_0"), val = tensor([1, 1])]; - tensor var_5477_groups_0 = const()[name = tensor("op_5477_groups_0"), val = tensor(1)]; - tensor layers_13_feed_forward2_fc2_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82552320))), name = tensor("layers_13_feed_forward2_fc2_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82510016))), shape = tensor([512, 2048, 1, 1])]; - tensor var_5477_cast_fp16 = conv(dilations = var_5477_dilations_0, groups = var_5477_groups_0, pad = var_5477_pad_0, pad_type = var_5477_pad_type_0, strides = var_5477_strides_0, weight = layers_13_feed_forward2_fc2_outlier_module_weight_to_fp16_sparsified, x = input_379_cast_fp16)[name = tensor("op_5477_cast_fp16")]; - tensor x_85_cast_fp16 = add(x = var_5471_cast_fp16, y = var_5477_cast_fp16)[name = tensor("x_85_cast_fp16")]; - tensor var_5479_to_fp16 = const()[name = tensor("op_5479_to_fp16"), val = tensor(0x1p-1)]; - tensor var_5480_cast_fp16 = mul(x = x_85_cast_fp16, y = var_5479_to_fp16)[name = tensor("op_5480_cast_fp16")]; - tensor inputs_139_cast_fp16 = add(x = inputs_137_cast_fp16, y = var_5480_cast_fp16)[name = tensor("inputs_139_cast_fp16")]; - tensor out_139_axes_0 = const()[name = tensor("out_139_axes_0"), val = tensor([1])]; - tensor var_5490_to_fp16 = const()[name = tensor("op_5490_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_139_cast_fp16 = layer_norm(axes = out_139_axes_0, epsilon = var_5490_to_fp16, x = inputs_139_cast_fp16)[name = tensor("out_139_cast_fp16")]; - tensor inputs_141_gamma_0_to_fp16 = const()[name = tensor("inputs_141_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82683456)))]; - tensor inputs_141_beta_0_to_fp16 = const()[name = tensor("inputs_141_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82684544)))]; - tensor inputs_141_epsilon_0_to_fp16 = const()[name = tensor("inputs_141_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor inputs_141_cast_fp16 = batch_norm(beta = inputs_141_beta_0_to_fp16, epsilon = inputs_141_epsilon_0_to_fp16, gamma = inputs_141_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_139_cast_fp16)[name = tensor("inputs_141_cast_fp16")]; - tensor var_5504 = const()[name = tensor("op_5504"), val = tensor(3)]; - tensor out_141_axes_0 = const()[name = tensor("out_141_axes_0"), val = tensor([1])]; - tensor var_5535_to_fp16 = const()[name = tensor("op_5535_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_141_cast_fp16 = layer_norm(axes = out_141_axes_0, epsilon = var_5535_to_fp16, x = inputs_141_cast_fp16)[name = tensor("out_141_cast_fp16")]; - tensor input_381_gamma_0_to_fp16 = const()[name = tensor("input_381_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82685632)))]; - tensor input_381_beta_0_to_fp16 = const()[name = tensor("input_381_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82686720)))]; - tensor input_381_epsilon_0_to_fp16 = const()[name = tensor("input_381_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_381_cast_fp16 = batch_norm(beta = input_381_beta_0_to_fp16, epsilon = input_381_epsilon_0_to_fp16, gamma = input_381_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_141_cast_fp16)[name = tensor("input_381_cast_fp16")]; - tensor var_5555_pad_type_0 = const()[name = tensor("op_5555_pad_type_0"), val = tensor("valid")]; - tensor var_5555_strides_0 = const()[name = tensor("op_5555_strides_0"), val = tensor([1, 1])]; - tensor var_5555_pad_0 = const()[name = tensor("op_5555_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5555_dilations_0 = const()[name = tensor("op_5555_dilations_0"), val = tensor([1, 1])]; - tensor var_5555_groups_0 = const()[name = tensor("op_5555_groups_0"), val = tensor(1)]; - tensor layers_14_feed_forward1_fc1_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82687808))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83474304))), name = tensor("layers_14_feed_forward1_fc1_inlier_module_weight_to_fp16_palettized"), shape = tensor([2048, 512, 1, 1])]; - tensor layers_14_feed_forward1_fc1_inlier_module_bias_to_fp16 = const()[name = tensor("layers_14_feed_forward1_fc1_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83474496)))]; - tensor var_5555_cast_fp16 = conv(bias = layers_14_feed_forward1_fc1_inlier_module_bias_to_fp16, dilations = var_5555_dilations_0, groups = var_5555_groups_0, pad = var_5555_pad_0, pad_type = var_5555_pad_type_0, strides = var_5555_strides_0, weight = layers_14_feed_forward1_fc1_inlier_module_weight_to_fp16_palettized, x = input_381_cast_fp16)[name = tensor("op_5555_cast_fp16")]; - tensor var_5561_pad_type_0 = const()[name = tensor("op_5561_pad_type_0"), val = tensor("valid")]; - tensor var_5561_strides_0 = const()[name = tensor("op_5561_strides_0"), val = tensor([1, 1])]; - tensor var_5561_pad_0 = const()[name = tensor("op_5561_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5561_dilations_0 = const()[name = tensor("op_5561_dilations_0"), val = tensor([1, 1])]; - tensor var_5561_groups_0 = const()[name = tensor("op_5561_groups_0"), val = tensor(1)]; - tensor layers_14_feed_forward1_fc1_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83513984))), name = tensor("layers_14_feed_forward1_fc1_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83478656))), shape = tensor([2048, 512, 1, 1])]; - tensor var_5561_cast_fp16 = conv(dilations = var_5561_dilations_0, groups = var_5561_groups_0, pad = var_5561_pad_0, pad_type = var_5561_pad_type_0, strides = var_5561_strides_0, weight = layers_14_feed_forward1_fc1_outlier_module_weight_to_fp16_sparsified, x = input_381_cast_fp16)[name = tensor("op_5561_cast_fp16")]; - tensor input_383_cast_fp16 = add(x = var_5555_cast_fp16, y = var_5561_cast_fp16)[name = tensor("input_383_cast_fp16")]; - tensor input_385_cast_fp16 = silu(x = input_383_cast_fp16)[name = tensor("input_385_cast_fp16")]; - tensor var_5572_pad_type_0 = const()[name = tensor("op_5572_pad_type_0"), val = tensor("valid")]; - tensor var_5572_strides_0 = const()[name = tensor("op_5572_strides_0"), val = tensor([1, 1])]; - tensor var_5572_pad_0 = const()[name = tensor("op_5572_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5572_dilations_0 = const()[name = tensor("op_5572_dilations_0"), val = tensor([1, 1])]; - tensor var_5572_groups_0 = const()[name = tensor("op_5572_groups_0"), val = tensor(1)]; - tensor layers_14_feed_forward1_fc2_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83645120))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84431616))), name = tensor("layers_14_feed_forward1_fc2_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 2048, 1, 1])]; - tensor layers_14_feed_forward1_fc2_inlier_module_bias_to_fp16 = const()[name = tensor("layers_14_feed_forward1_fc2_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84431808)))]; - tensor var_5572_cast_fp16 = conv(bias = layers_14_feed_forward1_fc2_inlier_module_bias_to_fp16, dilations = var_5572_dilations_0, groups = var_5572_groups_0, pad = var_5572_pad_0, pad_type = var_5572_pad_type_0, strides = var_5572_strides_0, weight = layers_14_feed_forward1_fc2_inlier_module_weight_to_fp16_palettized, x = input_385_cast_fp16)[name = tensor("op_5572_cast_fp16")]; - tensor var_5578_pad_type_0 = const()[name = tensor("op_5578_pad_type_0"), val = tensor("valid")]; - tensor var_5578_strides_0 = const()[name = tensor("op_5578_strides_0"), val = tensor([1, 1])]; - tensor var_5578_pad_0 = const()[name = tensor("op_5578_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5578_dilations_0 = const()[name = tensor("op_5578_dilations_0"), val = tensor([1, 1])]; - tensor var_5578_groups_0 = const()[name = tensor("op_5578_groups_0"), val = tensor(1)]; - tensor layers_14_feed_forward1_fc2_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84471232))), name = tensor("layers_14_feed_forward1_fc2_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84432896))), shape = tensor([512, 2048, 1, 1])]; - tensor var_5578_cast_fp16 = conv(dilations = var_5578_dilations_0, groups = var_5578_groups_0, pad = var_5578_pad_0, pad_type = var_5578_pad_type_0, strides = var_5578_strides_0, weight = layers_14_feed_forward1_fc2_outlier_module_weight_to_fp16_sparsified, x = input_385_cast_fp16)[name = tensor("op_5578_cast_fp16")]; - tensor x_87_cast_fp16 = add(x = var_5572_cast_fp16, y = var_5578_cast_fp16)[name = tensor("x_87_cast_fp16")]; - tensor var_5580_to_fp16 = const()[name = tensor("op_5580_to_fp16"), val = tensor(0x1p-1)]; - tensor var_5581_cast_fp16 = mul(x = x_87_cast_fp16, y = var_5580_to_fp16)[name = tensor("op_5581_cast_fp16")]; - tensor inputs_143_cast_fp16 = add(x = inputs_141_cast_fp16, y = var_5581_cast_fp16)[name = tensor("inputs_143_cast_fp16")]; - tensor out_143_axes_0 = const()[name = tensor("out_143_axes_0"), val = tensor([1])]; - tensor var_5591_to_fp16 = const()[name = tensor("op_5591_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_143_cast_fp16 = layer_norm(axes = out_143_axes_0, epsilon = var_5591_to_fp16, x = inputs_143_cast_fp16)[name = tensor("out_143_cast_fp16")]; - tensor obj_59_gamma_0_to_fp16 = const()[name = tensor("obj_59_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84602368)))]; - tensor obj_59_beta_0_to_fp16 = const()[name = tensor("obj_59_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84603456)))]; - tensor obj_59_epsilon_0_to_fp16 = const()[name = tensor("obj_59_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor obj_59_cast_fp16 = batch_norm(beta = obj_59_beta_0_to_fp16, epsilon = obj_59_epsilon_0_to_fp16, gamma = obj_59_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_143_cast_fp16)[name = tensor("obj_59_cast_fp16")]; - tensor var_5616_pad_type_0 = const()[name = tensor("op_5616_pad_type_0"), val = tensor("valid")]; - tensor var_5616_strides_0 = const()[name = tensor("op_5616_strides_0"), val = tensor([1, 1])]; - tensor var_5616_pad_0 = const()[name = tensor("op_5616_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5616_dilations_0 = const()[name = tensor("op_5616_dilations_0"), val = tensor([1, 1])]; - tensor var_5616_groups_0 = const()[name = tensor("op_5616_groups_0"), val = tensor(1)]; - tensor layers_14_self_attn_q_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84604544))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84801216))), name = tensor("layers_14_self_attn_q_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_14_self_attn_q_proj_inlier_module_bias_to_fp16 = const()[name = tensor("layers_14_self_attn_q_proj_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84801408)))]; - tensor var_5616_cast_fp16 = conv(bias = layers_14_self_attn_q_proj_inlier_module_bias_to_fp16, dilations = var_5616_dilations_0, groups = var_5616_groups_0, pad = var_5616_pad_0, pad_type = var_5616_pad_type_0, strides = var_5616_strides_0, weight = layers_14_self_attn_q_proj_inlier_module_weight_to_fp16_palettized, x = obj_59_cast_fp16)[name = tensor("op_5616_cast_fp16")]; - tensor var_5622_pad_type_0 = const()[name = tensor("op_5622_pad_type_0"), val = tensor("valid")]; - tensor var_5622_strides_0 = const()[name = tensor("op_5622_strides_0"), val = tensor([1, 1])]; - tensor var_5622_pad_0 = const()[name = tensor("op_5622_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5622_dilations_0 = const()[name = tensor("op_5622_dilations_0"), val = tensor([1, 1])]; - tensor var_5622_groups_0 = const()[name = tensor("op_5622_groups_0"), val = tensor(1)]; - tensor layers_14_self_attn_q_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84809984))), name = tensor("layers_14_self_attn_q_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84802496))), shape = tensor([512, 512, 1, 1])]; - tensor var_5622_cast_fp16 = conv(dilations = var_5622_dilations_0, groups = var_5622_groups_0, pad = var_5622_pad_0, pad_type = var_5622_pad_type_0, strides = var_5622_strides_0, weight = layers_14_self_attn_q_proj_outlier_module_weight_to_fp16_sparsified, x = obj_59_cast_fp16)[name = tensor("op_5622_cast_fp16")]; - tensor query_57_cast_fp16 = add(x = var_5616_cast_fp16, y = var_5622_cast_fp16)[name = tensor("query_57_cast_fp16")]; - tensor var_5631_pad_type_0 = const()[name = tensor("op_5631_pad_type_0"), val = tensor("valid")]; - tensor var_5631_strides_0 = const()[name = tensor("op_5631_strides_0"), val = tensor([1, 1])]; - tensor var_5631_pad_0 = const()[name = tensor("op_5631_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5631_dilations_0 = const()[name = tensor("op_5631_dilations_0"), val = tensor([1, 1])]; - tensor var_5631_groups_0 = const()[name = tensor("op_5631_groups_0"), val = tensor(1)]; - tensor layers_14_self_attn_k_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84842816))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85039488))), name = tensor("layers_14_self_attn_k_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor var_5631_cast_fp16 = conv(dilations = var_5631_dilations_0, groups = var_5631_groups_0, pad = var_5631_pad_0, pad_type = var_5631_pad_type_0, strides = var_5631_strides_0, weight = layers_14_self_attn_k_proj_inlier_module_weight_to_fp16_palettized, x = obj_59_cast_fp16)[name = tensor("op_5631_cast_fp16")]; - tensor var_5637_pad_type_0 = const()[name = tensor("op_5637_pad_type_0"), val = tensor("valid")]; - tensor var_5637_strides_0 = const()[name = tensor("op_5637_strides_0"), val = tensor([1, 1])]; - tensor var_5637_pad_0 = const()[name = tensor("op_5637_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5637_dilations_0 = const()[name = tensor("op_5637_dilations_0"), val = tensor([1, 1])]; - tensor var_5637_groups_0 = const()[name = tensor("op_5637_groups_0"), val = tensor(1)]; - tensor layers_14_self_attn_k_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85047232))), name = tensor("layers_14_self_attn_k_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85039680))), shape = tensor([512, 512, 1, 1])]; - tensor var_5637_cast_fp16 = conv(dilations = var_5637_dilations_0, groups = var_5637_groups_0, pad = var_5637_pad_0, pad_type = var_5637_pad_type_0, strides = var_5637_strides_0, weight = layers_14_self_attn_k_proj_outlier_module_weight_to_fp16_sparsified, x = obj_59_cast_fp16)[name = tensor("op_5637_cast_fp16")]; - tensor key_29_cast_fp16 = add(x = var_5631_cast_fp16, y = var_5637_cast_fp16)[name = tensor("key_29_cast_fp16")]; - tensor var_5647_pad_type_0 = const()[name = tensor("op_5647_pad_type_0"), val = tensor("valid")]; - tensor var_5647_strides_0 = const()[name = tensor("op_5647_strides_0"), val = tensor([1, 1])]; - tensor var_5647_pad_0 = const()[name = tensor("op_5647_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5647_dilations_0 = const()[name = tensor("op_5647_dilations_0"), val = tensor([1, 1])]; - tensor var_5647_groups_0 = const()[name = tensor("op_5647_groups_0"), val = tensor(1)]; - tensor layers_14_self_attn_v_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85080064))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85276736))), name = tensor("layers_14_self_attn_v_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_14_self_attn_v_proj_inlier_module_bias_to_fp16 = const()[name = tensor("layers_14_self_attn_v_proj_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85276928)))]; - tensor var_5647_cast_fp16 = conv(bias = layers_14_self_attn_v_proj_inlier_module_bias_to_fp16, dilations = var_5647_dilations_0, groups = var_5647_groups_0, pad = var_5647_pad_0, pad_type = var_5647_pad_type_0, strides = var_5647_strides_0, weight = layers_14_self_attn_v_proj_inlier_module_weight_to_fp16_palettized, x = obj_59_cast_fp16)[name = tensor("op_5647_cast_fp16")]; - tensor var_5653_pad_type_0 = const()[name = tensor("op_5653_pad_type_0"), val = tensor("valid")]; - tensor var_5653_strides_0 = const()[name = tensor("op_5653_strides_0"), val = tensor([1, 1])]; - tensor var_5653_pad_0 = const()[name = tensor("op_5653_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5653_dilations_0 = const()[name = tensor("op_5653_dilations_0"), val = tensor([1, 1])]; - tensor var_5653_groups_0 = const()[name = tensor("op_5653_groups_0"), val = tensor(1)]; - tensor layers_14_self_attn_v_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85285760))), name = tensor("layers_14_self_attn_v_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85278016))), shape = tensor([512, 512, 1, 1])]; - tensor var_5653_cast_fp16 = conv(dilations = var_5653_dilations_0, groups = var_5653_groups_0, pad = var_5653_pad_0, pad_type = var_5653_pad_type_0, strides = var_5653_strides_0, weight = layers_14_self_attn_v_proj_outlier_module_weight_to_fp16_sparsified, x = obj_59_cast_fp16)[name = tensor("op_5653_cast_fp16")]; - tensor value_29_cast_fp16 = add(x = var_5647_cast_fp16, y = var_5653_cast_fp16)[name = tensor("value_29_cast_fp16")]; - tensor var_5656_to_fp16 = const()[name = tensor("op_5656_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85318592)))]; - tensor query_59_cast_fp16 = add(x = query_57_cast_fp16, y = var_5656_to_fp16)[name = tensor("query_59_cast_fp16")]; - tensor var_5659_to_fp16 = const()[name = tensor("op_5659_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85319680)))]; - tensor q_with_bias_v_29_cast_fp16 = add(x = query_57_cast_fp16, y = var_5659_to_fp16)[name = tensor("q_with_bias_v_29_cast_fp16")]; - tensor var_5669_pad_type_0 = const()[name = tensor("op_5669_pad_type_0"), val = tensor("valid")]; - tensor var_5669_strides_0 = const()[name = tensor("op_5669_strides_0"), val = tensor([1, 1])]; - tensor var_5669_pad_0 = const()[name = tensor("op_5669_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5669_dilations_0 = const()[name = tensor("op_5669_dilations_0"), val = tensor([1, 1])]; - tensor var_5669_groups_0 = const()[name = tensor("op_5669_groups_0"), val = tensor(1)]; - tensor layers_14_self_attn_linear_pos_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85320768))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85517440))), name = tensor("layers_14_self_attn_linear_pos_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor var_5669_cast_fp16 = conv(bias = input_17_mean_0_to_fp16, dilations = var_5669_dilations_0, groups = var_5669_groups_0, pad = var_5669_pad_0, pad_type = var_5669_pad_type_0, strides = var_5669_strides_0, weight = layers_14_self_attn_linear_pos_inlier_module_weight_to_fp16_palettized, x = obj_3_cast_fp16)[name = tensor("op_5669_cast_fp16")]; - tensor var_5675_pad_type_0 = const()[name = tensor("op_5675_pad_type_0"), val = tensor("valid")]; - tensor var_5675_strides_0 = const()[name = tensor("op_5675_strides_0"), val = tensor([1, 1])]; - tensor var_5675_pad_0 = const()[name = tensor("op_5675_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5675_dilations_0 = const()[name = tensor("op_5675_dilations_0"), val = tensor([1, 1])]; - tensor var_5675_groups_0 = const()[name = tensor("op_5675_groups_0"), val = tensor(1)]; - tensor layers_14_self_attn_linear_pos_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85539840))), name = tensor("layers_14_self_attn_linear_pos_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85517632))), shape = tensor([512, 512, 1, 1])]; - tensor var_5675_cast_fp16 = conv(dilations = var_5675_dilations_0, groups = var_5675_groups_0, pad = var_5675_pad_0, pad_type = var_5675_pad_type_0, strides = var_5675_strides_0, weight = layers_14_self_attn_linear_pos_outlier_module_weight_to_fp16_sparsified, x = obj_3_cast_fp16)[name = tensor("op_5675_cast_fp16")]; - tensor p_29_cast_fp16 = add(x = var_5669_cast_fp16, y = var_5675_cast_fp16)[name = tensor("p_29_cast_fp16")]; - tensor var_5679 = const()[name = tensor("op_5679"), val = tensor([1, 8, 64, 188])]; - tensor var_5680_cast_fp16 = reshape(shape = var_5679, x = q_with_bias_v_29_cast_fp16)[name = tensor("op_5680_cast_fp16")]; - tensor var_5681 = const()[name = tensor("op_5681"), val = tensor([1, 8, 64, -1])]; - tensor var_5682_cast_fp16 = reshape(shape = var_5681, x = p_29_cast_fp16)[name = tensor("op_5682_cast_fp16")]; - tensor matrix_bd_113_transpose_x_0 = const()[name = tensor("matrix_bd_113_transpose_x_0"), val = tensor(true)]; - tensor matrix_bd_113_transpose_y_0 = const()[name = tensor("matrix_bd_113_transpose_y_0"), val = tensor(false)]; - tensor matrix_bd_113_cast_fp16 = matmul(transpose_x = matrix_bd_113_transpose_x_0, transpose_y = matrix_bd_113_transpose_y_0, x = var_5680_cast_fp16, y = var_5682_cast_fp16)[name = tensor("matrix_bd_113_cast_fp16")]; - tensor matrix_bd_115_pad_0 = const()[name = tensor("matrix_bd_115_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor matrix_bd_115_mode_0 = const()[name = tensor("matrix_bd_115_mode_0"), val = tensor("constant")]; - tensor const_164_to_fp16 = const()[name = tensor("const_164_to_fp16"), val = tensor(0x0p+0)]; - tensor matrix_bd_115_cast_fp16 = pad(constant_val = const_164_to_fp16, mode = matrix_bd_115_mode_0, pad = matrix_bd_115_pad_0, x = matrix_bd_113_cast_fp16)[name = tensor("matrix_bd_115_cast_fp16")]; - tensor var_5691 = const()[name = tensor("op_5691"), val = tensor([1, 8, -1, 188])]; - tensor matrix_bd_117_cast_fp16 = reshape(shape = var_5691, x = matrix_bd_115_cast_fp16)[name = tensor("matrix_bd_117_cast_fp16")]; - tensor var_5695_begin_0 = const()[name = tensor("op_5695_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_5695_end_0 = const()[name = tensor("op_5695_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_5695_end_mask_0 = const()[name = tensor("op_5695_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_5695_cast_fp16 = slice_by_index(begin = var_5695_begin_0, end = var_5695_end_0, end_mask = var_5695_end_mask_0, x = matrix_bd_117_cast_fp16)[name = tensor("op_5695_cast_fp16")]; - tensor var_5696 = const()[name = tensor("op_5696"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_119_cast_fp16 = reshape(shape = var_5696, x = var_5695_cast_fp16)[name = tensor("matrix_bd_119_cast_fp16")]; - tensor var_5701_begin_0 = const()[name = tensor("op_5701_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5701_end_0 = const()[name = tensor("op_5701_end_0"), val = tensor([1, 8, 188, 188])]; - tensor var_5701_end_mask_0 = const()[name = tensor("op_5701_end_mask_0"), val = tensor([true, true, true, false])]; - tensor var_5701_cast_fp16 = slice_by_index(begin = var_5701_begin_0, end = var_5701_end_0, end_mask = var_5701_end_mask_0, x = matrix_bd_119_cast_fp16)[name = tensor("op_5701_cast_fp16")]; - tensor var_5702_to_fp16 = const()[name = tensor("op_5702_to_fp16"), val = tensor(0x1p-3)]; - tensor qk_mask_29_cast_fp16 = mul(x = var_5701_cast_fp16, y = var_5702_to_fp16)[name = tensor("qk_mask_29_cast_fp16")]; - tensor var_5706 = const()[name = tensor("op_5706"), val = tensor([1, 8, 64, 188])]; - tensor mh_q_29_cast_fp16 = reshape(shape = var_5706, x = query_59_cast_fp16)[name = tensor("mh_q_29_cast_fp16")]; - tensor var_5708_to_fp16 = const()[name = tensor("op_5708_to_fp16"), val = tensor(0x1p-3)]; - tensor var_5709_cast_fp16 = mul(x = mh_q_29_cast_fp16, y = var_5708_to_fp16)[name = tensor("op_5709_cast_fp16")]; - tensor var_5712 = const()[name = tensor("op_5712"), val = tensor([1, 8, 64, 188])]; - tensor var_5713_cast_fp16 = reshape(shape = var_5712, x = key_29_cast_fp16)[name = tensor("op_5713_cast_fp16")]; - tensor mh_w_57_transpose_x_0 = const()[name = tensor("mh_w_57_transpose_x_0"), val = tensor(true)]; - tensor mh_w_57_transpose_y_0 = const()[name = tensor("mh_w_57_transpose_y_0"), val = tensor(false)]; - tensor mh_w_57_cast_fp16 = matmul(transpose_x = mh_w_57_transpose_x_0, transpose_y = mh_w_57_transpose_y_0, x = var_5709_cast_fp16, y = var_5713_cast_fp16)[name = tensor("mh_w_57_cast_fp16")]; - tensor mh_w_59_cast_fp16 = add(x = mh_w_57_cast_fp16, y = qk_mask_29_cast_fp16)[name = tensor("mh_w_59_cast_fp16")]; - tensor var_5717_cast_fp16 = softmax(axis = var_5504, x = mh_w_59_cast_fp16)[name = tensor("op_5717_cast_fp16")]; - tensor var_5718 = const()[name = tensor("op_5718"), val = tensor([1, 8, 64, 188])]; - tensor var_5719_cast_fp16 = reshape(shape = var_5718, x = value_29_cast_fp16)[name = tensor("op_5719_cast_fp16")]; - tensor attn_29_transpose_x_0 = const()[name = tensor("attn_29_transpose_x_0"), val = tensor(false)]; - tensor attn_29_transpose_y_0 = const()[name = tensor("attn_29_transpose_y_0"), val = tensor(true)]; - tensor attn_29_cast_fp16 = matmul(transpose_x = attn_29_transpose_x_0, transpose_y = attn_29_transpose_y_0, x = var_5719_cast_fp16, y = var_5717_cast_fp16)[name = tensor("attn_29_cast_fp16")]; - tensor var_5722 = const()[name = tensor("op_5722"), val = tensor([1, 512, 1, 188])]; - tensor input_387_cast_fp16 = reshape(shape = var_5722, x = attn_29_cast_fp16)[name = tensor("input_387_cast_fp16")]; - tensor var_5732_pad_type_0 = const()[name = tensor("op_5732_pad_type_0"), val = tensor("valid")]; - tensor var_5732_strides_0 = const()[name = tensor("op_5732_strides_0"), val = tensor([1, 1])]; - tensor var_5732_pad_0 = const()[name = tensor("op_5732_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5732_dilations_0 = const()[name = tensor("op_5732_dilations_0"), val = tensor([1, 1])]; - tensor var_5732_groups_0 = const()[name = tensor("op_5732_groups_0"), val = tensor(1)]; - tensor layers_14_self_attn_o_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85572672))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85769344))), name = tensor("layers_14_self_attn_o_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_14_self_attn_o_proj_inlier_module_bias_to_fp16 = const()[name = tensor("layers_14_self_attn_o_proj_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85769536)))]; - tensor var_5732_cast_fp16 = conv(bias = layers_14_self_attn_o_proj_inlier_module_bias_to_fp16, dilations = var_5732_dilations_0, groups = var_5732_groups_0, pad = var_5732_pad_0, pad_type = var_5732_pad_type_0, strides = var_5732_strides_0, weight = layers_14_self_attn_o_proj_inlier_module_weight_to_fp16_palettized, x = input_387_cast_fp16)[name = tensor("op_5732_cast_fp16")]; - tensor var_5738_pad_type_0 = const()[name = tensor("op_5738_pad_type_0"), val = tensor("valid")]; - tensor var_5738_strides_0 = const()[name = tensor("op_5738_strides_0"), val = tensor([1, 1])]; - tensor var_5738_pad_0 = const()[name = tensor("op_5738_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5738_dilations_0 = const()[name = tensor("op_5738_dilations_0"), val = tensor([1, 1])]; - tensor var_5738_groups_0 = const()[name = tensor("op_5738_groups_0"), val = tensor(1)]; - tensor layers_14_self_attn_o_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85779392))), name = tensor("layers_14_self_attn_o_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85770624))), shape = tensor([512, 512, 1, 1])]; - tensor var_5738_cast_fp16 = conv(dilations = var_5738_dilations_0, groups = var_5738_groups_0, pad = var_5738_pad_0, pad_type = var_5738_pad_type_0, strides = var_5738_strides_0, weight = layers_14_self_attn_o_proj_outlier_module_weight_to_fp16_sparsified, x = input_387_cast_fp16)[name = tensor("op_5738_cast_fp16")]; - tensor obj_61_cast_fp16 = add(x = var_5732_cast_fp16, y = var_5738_cast_fp16)[name = tensor("obj_61_cast_fp16")]; - tensor inputs_145_cast_fp16 = add(x = inputs_143_cast_fp16, y = obj_61_cast_fp16)[name = tensor("inputs_145_cast_fp16")]; - tensor out_145_axes_0 = const()[name = tensor("out_145_axes_0"), val = tensor([1])]; - tensor var_5749_to_fp16 = const()[name = tensor("op_5749_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_145_cast_fp16 = layer_norm(axes = out_145_axes_0, epsilon = var_5749_to_fp16, x = inputs_145_cast_fp16)[name = tensor("out_145_cast_fp16")]; - tensor input_389_gamma_0_to_fp16 = const()[name = tensor("input_389_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85812224)))]; - tensor input_389_beta_0_to_fp16 = const()[name = tensor("input_389_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85813312)))]; - tensor input_389_epsilon_0_to_fp16 = const()[name = tensor("input_389_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_389_cast_fp16 = batch_norm(beta = input_389_beta_0_to_fp16, epsilon = input_389_epsilon_0_to_fp16, gamma = input_389_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_145_cast_fp16)[name = tensor("input_389_cast_fp16")]; - tensor var_5771_pad_type_0 = const()[name = tensor("op_5771_pad_type_0"), val = tensor("valid")]; - tensor var_5771_strides_0 = const()[name = tensor("op_5771_strides_0"), val = tensor([1, 1])]; - tensor var_5771_pad_0 = const()[name = tensor("op_5771_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5771_dilations_0 = const()[name = tensor("op_5771_dilations_0"), val = tensor([1, 1])]; - tensor var_5771_groups_0 = const()[name = tensor("op_5771_groups_0"), val = tensor(1)]; - tensor layers_14_conv_pointwise_conv1_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85814400))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86207680))), name = tensor("layers_14_conv_pointwise_conv1_inlier_module_weight_to_fp16_palettized"), shape = tensor([1024, 512, 1, 1])]; - tensor layers_14_conv_pointwise_conv1_inlier_module_bias_to_fp16 = const()[name = tensor("layers_14_conv_pointwise_conv1_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86207872)))]; - tensor var_5771_cast_fp16 = conv(bias = layers_14_conv_pointwise_conv1_inlier_module_bias_to_fp16, dilations = var_5771_dilations_0, groups = var_5771_groups_0, pad = var_5771_pad_0, pad_type = var_5771_pad_type_0, strides = var_5771_strides_0, weight = layers_14_conv_pointwise_conv1_inlier_module_weight_to_fp16_palettized, x = input_389_cast_fp16)[name = tensor("op_5771_cast_fp16")]; - tensor var_5777_pad_type_0 = const()[name = tensor("op_5777_pad_type_0"), val = tensor("valid")]; - tensor var_5777_strides_0 = const()[name = tensor("op_5777_strides_0"), val = tensor([1, 1])]; - tensor var_5777_pad_0 = const()[name = tensor("op_5777_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5777_dilations_0 = const()[name = tensor("op_5777_dilations_0"), val = tensor([1, 1])]; - tensor var_5777_groups_0 = const()[name = tensor("op_5777_groups_0"), val = tensor(1)]; - tensor layers_14_conv_pointwise_conv1_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86229248))), name = tensor("layers_14_conv_pointwise_conv1_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86209984))), shape = tensor([1024, 512, 1, 1])]; - tensor var_5777_cast_fp16 = conv(dilations = var_5777_dilations_0, groups = var_5777_groups_0, pad = var_5777_pad_0, pad_type = var_5777_pad_type_0, strides = var_5777_strides_0, weight = layers_14_conv_pointwise_conv1_outlier_module_weight_to_fp16_sparsified, x = input_389_cast_fp16)[name = tensor("op_5777_cast_fp16")]; - tensor input_391_cast_fp16 = add(x = var_5771_cast_fp16, y = var_5777_cast_fp16)[name = tensor("input_391_cast_fp16")]; - tensor input_393_split_num_splits_0 = const()[name = tensor("input_393_split_num_splits_0"), val = tensor(2)]; - tensor input_393_split_axis_0 = const()[name = tensor("input_393_split_axis_0"), val = tensor(1)]; - tensor input_393_split_cast_fp16_0, tensor input_393_split_cast_fp16_1 = split(axis = input_393_split_axis_0, num_splits = input_393_split_num_splits_0, x = input_391_cast_fp16)[name = tensor("input_393_split_cast_fp16")]; - tensor input_393_split_1_sigmoid_cast_fp16 = sigmoid(x = input_393_split_cast_fp16_1)[name = tensor("input_393_split_1_sigmoid_cast_fp16")]; - tensor input_393_cast_fp16 = mul(x = input_393_split_cast_fp16_0, y = input_393_split_1_sigmoid_cast_fp16)[name = tensor("input_393_cast_fp16")]; - tensor input_395_pad_type_0 = const()[name = tensor("input_395_pad_type_0"), val = tensor("custom")]; - tensor input_395_pad_0 = const()[name = tensor("input_395_pad_0"), val = tensor([0, 0, 4, 4])]; - tensor input_395_groups_0 = const()[name = tensor("input_395_groups_0"), val = tensor(512)]; - tensor input_395_strides_0 = const()[name = tensor("input_395_strides_0"), val = tensor([1, 1])]; - tensor input_395_dilations_0 = const()[name = tensor("input_395_dilations_0"), val = tensor([1, 1])]; - tensor const_219_to_fp16 = const()[name = tensor("const_219_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86294848)))]; - tensor const_220_to_fp16 = const()[name = tensor("const_220_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86304128)))]; - tensor input_397_cast_fp16 = conv(bias = const_220_to_fp16, dilations = input_395_dilations_0, groups = input_395_groups_0, pad = input_395_pad_0, pad_type = input_395_pad_type_0, strides = input_395_strides_0, weight = const_219_to_fp16, x = input_393_cast_fp16)[name = tensor("input_397_cast_fp16")]; - tensor input_399_cast_fp16 = silu(x = input_397_cast_fp16)[name = tensor("input_399_cast_fp16")]; - tensor var_5801_pad_type_0 = const()[name = tensor("op_5801_pad_type_0"), val = tensor("valid")]; - tensor var_5801_strides_0 = const()[name = tensor("op_5801_strides_0"), val = tensor([1, 1])]; - tensor var_5801_pad_0 = const()[name = tensor("op_5801_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5801_dilations_0 = const()[name = tensor("op_5801_dilations_0"), val = tensor([1, 1])]; - tensor var_5801_groups_0 = const()[name = tensor("op_5801_groups_0"), val = tensor(1)]; - tensor layers_14_conv_pointwise_conv2_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86305216))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86501888))), name = tensor("layers_14_conv_pointwise_conv2_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_14_conv_pointwise_conv2_inlier_module_bias_to_fp16 = const()[name = tensor("layers_14_conv_pointwise_conv2_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86502080)))]; - tensor var_5801_cast_fp16 = conv(bias = layers_14_conv_pointwise_conv2_inlier_module_bias_to_fp16, dilations = var_5801_dilations_0, groups = var_5801_groups_0, pad = var_5801_pad_0, pad_type = var_5801_pad_type_0, strides = var_5801_strides_0, weight = layers_14_conv_pointwise_conv2_inlier_module_weight_to_fp16_palettized, x = input_399_cast_fp16)[name = tensor("op_5801_cast_fp16")]; - tensor var_5807_pad_type_0 = const()[name = tensor("op_5807_pad_type_0"), val = tensor("valid")]; - tensor var_5807_strides_0 = const()[name = tensor("op_5807_strides_0"), val = tensor([1, 1])]; - tensor var_5807_pad_0 = const()[name = tensor("op_5807_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5807_dilations_0 = const()[name = tensor("op_5807_dilations_0"), val = tensor([1, 1])]; - tensor var_5807_groups_0 = const()[name = tensor("op_5807_groups_0"), val = tensor(1)]; - tensor layers_14_conv_pointwise_conv2_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86511872))), name = tensor("layers_14_conv_pointwise_conv2_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86503168))), shape = tensor([512, 512, 1, 1])]; - tensor var_5807_cast_fp16 = conv(dilations = var_5807_dilations_0, groups = var_5807_groups_0, pad = var_5807_pad_0, pad_type = var_5807_pad_type_0, strides = var_5807_strides_0, weight = layers_14_conv_pointwise_conv2_outlier_module_weight_to_fp16_sparsified, x = input_399_cast_fp16)[name = tensor("op_5807_cast_fp16")]; - tensor x_89_cast_fp16 = add(x = var_5801_cast_fp16, y = var_5807_cast_fp16)[name = tensor("x_89_cast_fp16")]; - tensor inputs_147_cast_fp16 = add(x = inputs_145_cast_fp16, y = x_89_cast_fp16)[name = tensor("inputs_147_cast_fp16")]; - tensor out_147_axes_0 = const()[name = tensor("out_147_axes_0"), val = tensor([1])]; - tensor var_5818_to_fp16 = const()[name = tensor("op_5818_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_147_cast_fp16 = layer_norm(axes = out_147_axes_0, epsilon = var_5818_to_fp16, x = inputs_147_cast_fp16)[name = tensor("out_147_cast_fp16")]; - tensor input_401_gamma_0_to_fp16 = const()[name = tensor("input_401_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86544704)))]; - tensor input_401_beta_0_to_fp16 = const()[name = tensor("input_401_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86545792)))]; - tensor input_401_epsilon_0_to_fp16 = const()[name = tensor("input_401_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_401_cast_fp16 = batch_norm(beta = input_401_beta_0_to_fp16, epsilon = input_401_epsilon_0_to_fp16, gamma = input_401_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_147_cast_fp16)[name = tensor("input_401_cast_fp16")]; - tensor var_5838_pad_type_0 = const()[name = tensor("op_5838_pad_type_0"), val = tensor("valid")]; - tensor var_5838_strides_0 = const()[name = tensor("op_5838_strides_0"), val = tensor([1, 1])]; - tensor var_5838_pad_0 = const()[name = tensor("op_5838_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5838_dilations_0 = const()[name = tensor("op_5838_dilations_0"), val = tensor([1, 1])]; - tensor var_5838_groups_0 = const()[name = tensor("op_5838_groups_0"), val = tensor(1)]; - tensor layers_14_feed_forward2_fc1_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86546880))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(87333376))), name = tensor("layers_14_feed_forward2_fc1_inlier_module_weight_to_fp16_palettized"), shape = tensor([2048, 512, 1, 1])]; - tensor layers_14_feed_forward2_fc1_inlier_module_bias_to_fp16 = const()[name = tensor("layers_14_feed_forward2_fc1_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(87333568)))]; - tensor var_5838_cast_fp16 = conv(bias = layers_14_feed_forward2_fc1_inlier_module_bias_to_fp16, dilations = var_5838_dilations_0, groups = var_5838_groups_0, pad = var_5838_pad_0, pad_type = var_5838_pad_type_0, strides = var_5838_strides_0, weight = layers_14_feed_forward2_fc1_inlier_module_weight_to_fp16_palettized, x = input_401_cast_fp16)[name = tensor("op_5838_cast_fp16")]; - tensor var_5844_pad_type_0 = const()[name = tensor("op_5844_pad_type_0"), val = tensor("valid")]; - tensor var_5844_strides_0 = const()[name = tensor("op_5844_strides_0"), val = tensor([1, 1])]; - tensor var_5844_pad_0 = const()[name = tensor("op_5844_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5844_dilations_0 = const()[name = tensor("op_5844_dilations_0"), val = tensor([1, 1])]; - tensor var_5844_groups_0 = const()[name = tensor("op_5844_groups_0"), val = tensor(1)]; - tensor layers_14_feed_forward2_fc1_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(87370176))), name = tensor("layers_14_feed_forward2_fc1_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(87337728))), shape = tensor([2048, 512, 1, 1])]; - tensor var_5844_cast_fp16 = conv(dilations = var_5844_dilations_0, groups = var_5844_groups_0, pad = var_5844_pad_0, pad_type = var_5844_pad_type_0, strides = var_5844_strides_0, weight = layers_14_feed_forward2_fc1_outlier_module_weight_to_fp16_sparsified, x = input_401_cast_fp16)[name = tensor("op_5844_cast_fp16")]; - tensor input_403_cast_fp16 = add(x = var_5838_cast_fp16, y = var_5844_cast_fp16)[name = tensor("input_403_cast_fp16")]; - tensor input_405_cast_fp16 = silu(x = input_403_cast_fp16)[name = tensor("input_405_cast_fp16")]; - tensor var_5855_pad_type_0 = const()[name = tensor("op_5855_pad_type_0"), val = tensor("valid")]; - tensor var_5855_strides_0 = const()[name = tensor("op_5855_strides_0"), val = tensor([1, 1])]; - tensor var_5855_pad_0 = const()[name = tensor("op_5855_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5855_dilations_0 = const()[name = tensor("op_5855_dilations_0"), val = tensor([1, 1])]; - tensor var_5855_groups_0 = const()[name = tensor("op_5855_groups_0"), val = tensor(1)]; - tensor layers_14_feed_forward2_fc2_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(87501312))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(88287808))), name = tensor("layers_14_feed_forward2_fc2_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 2048, 1, 1])]; - tensor layers_14_feed_forward2_fc2_inlier_module_bias_to_fp16 = const()[name = tensor("layers_14_feed_forward2_fc2_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(88288000)))]; - tensor var_5855_cast_fp16 = conv(bias = layers_14_feed_forward2_fc2_inlier_module_bias_to_fp16, dilations = var_5855_dilations_0, groups = var_5855_groups_0, pad = var_5855_pad_0, pad_type = var_5855_pad_type_0, strides = var_5855_strides_0, weight = layers_14_feed_forward2_fc2_inlier_module_weight_to_fp16_palettized, x = input_405_cast_fp16)[name = tensor("op_5855_cast_fp16")]; - tensor var_5861_pad_type_0 = const()[name = tensor("op_5861_pad_type_0"), val = tensor("valid")]; - tensor var_5861_strides_0 = const()[name = tensor("op_5861_strides_0"), val = tensor([1, 1])]; - tensor var_5861_pad_0 = const()[name = tensor("op_5861_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5861_dilations_0 = const()[name = tensor("op_5861_dilations_0"), val = tensor([1, 1])]; - tensor var_5861_groups_0 = const()[name = tensor("op_5861_groups_0"), val = tensor(1)]; - tensor layers_14_feed_forward2_fc2_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(88327936))), name = tensor("layers_14_feed_forward2_fc2_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(88289088))), shape = tensor([512, 2048, 1, 1])]; - tensor var_5861_cast_fp16 = conv(dilations = var_5861_dilations_0, groups = var_5861_groups_0, pad = var_5861_pad_0, pad_type = var_5861_pad_type_0, strides = var_5861_strides_0, weight = layers_14_feed_forward2_fc2_outlier_module_weight_to_fp16_sparsified, x = input_405_cast_fp16)[name = tensor("op_5861_cast_fp16")]; - tensor x_91_cast_fp16 = add(x = var_5855_cast_fp16, y = var_5861_cast_fp16)[name = tensor("x_91_cast_fp16")]; - tensor var_5863_to_fp16 = const()[name = tensor("op_5863_to_fp16"), val = tensor(0x1p-1)]; - tensor var_5864_cast_fp16 = mul(x = x_91_cast_fp16, y = var_5863_to_fp16)[name = tensor("op_5864_cast_fp16")]; - tensor inputs_149_cast_fp16 = add(x = inputs_147_cast_fp16, y = var_5864_cast_fp16)[name = tensor("inputs_149_cast_fp16")]; - tensor out_149_axes_0 = const()[name = tensor("out_149_axes_0"), val = tensor([1])]; - tensor var_5874_to_fp16 = const()[name = tensor("op_5874_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_149_cast_fp16 = layer_norm(axes = out_149_axes_0, epsilon = var_5874_to_fp16, x = inputs_149_cast_fp16)[name = tensor("out_149_cast_fp16")]; - tensor inputs_151_gamma_0_to_fp16 = const()[name = tensor("inputs_151_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(88459072)))]; - tensor inputs_151_beta_0_to_fp16 = const()[name = tensor("inputs_151_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(88460160)))]; - tensor inputs_151_epsilon_0_to_fp16 = const()[name = tensor("inputs_151_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor inputs_151_cast_fp16 = batch_norm(beta = inputs_151_beta_0_to_fp16, epsilon = inputs_151_epsilon_0_to_fp16, gamma = inputs_151_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_149_cast_fp16)[name = tensor("inputs_151_cast_fp16")]; - tensor var_5888 = const()[name = tensor("op_5888"), val = tensor(3)]; - tensor out_151_axes_0 = const()[name = tensor("out_151_axes_0"), val = tensor([1])]; - tensor var_5919_to_fp16 = const()[name = tensor("op_5919_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_151_cast_fp16 = layer_norm(axes = out_151_axes_0, epsilon = var_5919_to_fp16, x = inputs_151_cast_fp16)[name = tensor("out_151_cast_fp16")]; - tensor input_407_gamma_0_to_fp16 = const()[name = tensor("input_407_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(88461248)))]; - tensor input_407_beta_0_to_fp16 = const()[name = tensor("input_407_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(88462336)))]; - tensor input_407_epsilon_0_to_fp16 = const()[name = tensor("input_407_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_407_cast_fp16 = batch_norm(beta = input_407_beta_0_to_fp16, epsilon = input_407_epsilon_0_to_fp16, gamma = input_407_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_151_cast_fp16)[name = tensor("input_407_cast_fp16")]; - tensor var_5939_pad_type_0 = const()[name = tensor("op_5939_pad_type_0"), val = tensor("valid")]; - tensor var_5939_strides_0 = const()[name = tensor("op_5939_strides_0"), val = tensor([1, 1])]; - tensor var_5939_pad_0 = const()[name = tensor("op_5939_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5939_dilations_0 = const()[name = tensor("op_5939_dilations_0"), val = tensor([1, 1])]; - tensor var_5939_groups_0 = const()[name = tensor("op_5939_groups_0"), val = tensor(1)]; - tensor layers_15_feed_forward1_fc1_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(88463424))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(89249920))), name = tensor("layers_15_feed_forward1_fc1_inlier_module_weight_to_fp16_palettized"), shape = tensor([2048, 512, 1, 1])]; - tensor layers_15_feed_forward1_fc1_inlier_module_bias_to_fp16 = const()[name = tensor("layers_15_feed_forward1_fc1_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(89250112)))]; - tensor var_5939_cast_fp16 = conv(bias = layers_15_feed_forward1_fc1_inlier_module_bias_to_fp16, dilations = var_5939_dilations_0, groups = var_5939_groups_0, pad = var_5939_pad_0, pad_type = var_5939_pad_type_0, strides = var_5939_strides_0, weight = layers_15_feed_forward1_fc1_inlier_module_weight_to_fp16_palettized, x = input_407_cast_fp16)[name = tensor("op_5939_cast_fp16")]; - tensor var_5945_pad_type_0 = const()[name = tensor("op_5945_pad_type_0"), val = tensor("valid")]; - tensor var_5945_strides_0 = const()[name = tensor("op_5945_strides_0"), val = tensor([1, 1])]; - tensor var_5945_pad_0 = const()[name = tensor("op_5945_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5945_dilations_0 = const()[name = tensor("op_5945_dilations_0"), val = tensor([1, 1])]; - tensor var_5945_groups_0 = const()[name = tensor("op_5945_groups_0"), val = tensor(1)]; - tensor layers_15_feed_forward1_fc1_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(89285760))), name = tensor("layers_15_feed_forward1_fc1_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(89254272))), shape = tensor([2048, 512, 1, 1])]; - tensor var_5945_cast_fp16 = conv(dilations = var_5945_dilations_0, groups = var_5945_groups_0, pad = var_5945_pad_0, pad_type = var_5945_pad_type_0, strides = var_5945_strides_0, weight = layers_15_feed_forward1_fc1_outlier_module_weight_to_fp16_sparsified, x = input_407_cast_fp16)[name = tensor("op_5945_cast_fp16")]; - tensor input_409_cast_fp16 = add(x = var_5939_cast_fp16, y = var_5945_cast_fp16)[name = tensor("input_409_cast_fp16")]; - tensor input_411_cast_fp16 = silu(x = input_409_cast_fp16)[name = tensor("input_411_cast_fp16")]; - tensor var_5956_pad_type_0 = const()[name = tensor("op_5956_pad_type_0"), val = tensor("valid")]; - tensor var_5956_strides_0 = const()[name = tensor("op_5956_strides_0"), val = tensor([1, 1])]; - tensor var_5956_pad_0 = const()[name = tensor("op_5956_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5956_dilations_0 = const()[name = tensor("op_5956_dilations_0"), val = tensor([1, 1])]; - tensor var_5956_groups_0 = const()[name = tensor("op_5956_groups_0"), val = tensor(1)]; - tensor layers_15_feed_forward1_fc2_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(89416896))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90203392))), name = tensor("layers_15_feed_forward1_fc2_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 2048, 1, 1])]; - tensor layers_15_feed_forward1_fc2_inlier_module_bias_to_fp16 = const()[name = tensor("layers_15_feed_forward1_fc2_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90203584)))]; - tensor var_5956_cast_fp16 = conv(bias = layers_15_feed_forward1_fc2_inlier_module_bias_to_fp16, dilations = var_5956_dilations_0, groups = var_5956_groups_0, pad = var_5956_pad_0, pad_type = var_5956_pad_type_0, strides = var_5956_strides_0, weight = layers_15_feed_forward1_fc2_inlier_module_weight_to_fp16_palettized, x = input_411_cast_fp16)[name = tensor("op_5956_cast_fp16")]; - tensor var_5962_pad_type_0 = const()[name = tensor("op_5962_pad_type_0"), val = tensor("valid")]; - tensor var_5962_strides_0 = const()[name = tensor("op_5962_strides_0"), val = tensor([1, 1])]; - tensor var_5962_pad_0 = const()[name = tensor("op_5962_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_5962_dilations_0 = const()[name = tensor("op_5962_dilations_0"), val = tensor([1, 1])]; - tensor var_5962_groups_0 = const()[name = tensor("op_5962_groups_0"), val = tensor(1)]; - tensor layers_15_feed_forward1_fc2_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90244800))), name = tensor("layers_15_feed_forward1_fc2_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90204672))), shape = tensor([512, 2048, 1, 1])]; - tensor var_5962_cast_fp16 = conv(dilations = var_5962_dilations_0, groups = var_5962_groups_0, pad = var_5962_pad_0, pad_type = var_5962_pad_type_0, strides = var_5962_strides_0, weight = layers_15_feed_forward1_fc2_outlier_module_weight_to_fp16_sparsified, x = input_411_cast_fp16)[name = tensor("op_5962_cast_fp16")]; - tensor x_93_cast_fp16 = add(x = var_5956_cast_fp16, y = var_5962_cast_fp16)[name = tensor("x_93_cast_fp16")]; - tensor var_5964_to_fp16 = const()[name = tensor("op_5964_to_fp16"), val = tensor(0x1p-1)]; - tensor var_5965_cast_fp16 = mul(x = x_93_cast_fp16, y = var_5964_to_fp16)[name = tensor("op_5965_cast_fp16")]; - tensor inputs_153_cast_fp16 = add(x = inputs_151_cast_fp16, y = var_5965_cast_fp16)[name = tensor("inputs_153_cast_fp16")]; - tensor out_153_axes_0 = const()[name = tensor("out_153_axes_0"), val = tensor([1])]; - tensor var_5975_to_fp16 = const()[name = tensor("op_5975_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_153_cast_fp16 = layer_norm(axes = out_153_axes_0, epsilon = var_5975_to_fp16, x = inputs_153_cast_fp16)[name = tensor("out_153_cast_fp16")]; - tensor obj_63_gamma_0_to_fp16 = const()[name = tensor("obj_63_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90375936)))]; - tensor obj_63_beta_0_to_fp16 = const()[name = tensor("obj_63_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90377024)))]; - tensor obj_63_epsilon_0_to_fp16 = const()[name = tensor("obj_63_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor obj_63_cast_fp16 = batch_norm(beta = obj_63_beta_0_to_fp16, epsilon = obj_63_epsilon_0_to_fp16, gamma = obj_63_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_153_cast_fp16)[name = tensor("obj_63_cast_fp16")]; - tensor var_6000_pad_type_0 = const()[name = tensor("op_6000_pad_type_0"), val = tensor("valid")]; - tensor var_6000_strides_0 = const()[name = tensor("op_6000_strides_0"), val = tensor([1, 1])]; - tensor var_6000_pad_0 = const()[name = tensor("op_6000_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_6000_dilations_0 = const()[name = tensor("op_6000_dilations_0"), val = tensor([1, 1])]; - tensor var_6000_groups_0 = const()[name = tensor("op_6000_groups_0"), val = tensor(1)]; - tensor layers_15_self_attn_q_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90378112))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90574784))), name = tensor("layers_15_self_attn_q_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_15_self_attn_q_proj_inlier_module_bias_to_fp16 = const()[name = tensor("layers_15_self_attn_q_proj_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90574976)))]; - tensor var_6000_cast_fp16 = conv(bias = layers_15_self_attn_q_proj_inlier_module_bias_to_fp16, dilations = var_6000_dilations_0, groups = var_6000_groups_0, pad = var_6000_pad_0, pad_type = var_6000_pad_type_0, strides = var_6000_strides_0, weight = layers_15_self_attn_q_proj_inlier_module_weight_to_fp16_palettized, x = obj_63_cast_fp16)[name = tensor("op_6000_cast_fp16")]; - tensor var_6006_pad_type_0 = const()[name = tensor("op_6006_pad_type_0"), val = tensor("valid")]; - tensor var_6006_strides_0 = const()[name = tensor("op_6006_strides_0"), val = tensor([1, 1])]; - tensor var_6006_pad_0 = const()[name = tensor("op_6006_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_6006_dilations_0 = const()[name = tensor("op_6006_dilations_0"), val = tensor([1, 1])]; - tensor var_6006_groups_0 = const()[name = tensor("op_6006_groups_0"), val = tensor(1)]; - tensor layers_15_self_attn_q_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90583680))), name = tensor("layers_15_self_attn_q_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90576064))), shape = tensor([512, 512, 1, 1])]; - tensor var_6006_cast_fp16 = conv(dilations = var_6006_dilations_0, groups = var_6006_groups_0, pad = var_6006_pad_0, pad_type = var_6006_pad_type_0, strides = var_6006_strides_0, weight = layers_15_self_attn_q_proj_outlier_module_weight_to_fp16_sparsified, x = obj_63_cast_fp16)[name = tensor("op_6006_cast_fp16")]; - tensor query_61_cast_fp16 = add(x = var_6000_cast_fp16, y = var_6006_cast_fp16)[name = tensor("query_61_cast_fp16")]; - tensor var_6015_pad_type_0 = const()[name = tensor("op_6015_pad_type_0"), val = tensor("valid")]; - tensor var_6015_strides_0 = const()[name = tensor("op_6015_strides_0"), val = tensor([1, 1])]; - tensor var_6015_pad_0 = const()[name = tensor("op_6015_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_6015_dilations_0 = const()[name = tensor("op_6015_dilations_0"), val = tensor([1, 1])]; - tensor var_6015_groups_0 = const()[name = tensor("op_6015_groups_0"), val = tensor(1)]; - tensor layers_15_self_attn_k_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90616512))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90813184))), name = tensor("layers_15_self_attn_k_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor var_6015_cast_fp16 = conv(dilations = var_6015_dilations_0, groups = var_6015_groups_0, pad = var_6015_pad_0, pad_type = var_6015_pad_type_0, strides = var_6015_strides_0, weight = layers_15_self_attn_k_proj_inlier_module_weight_to_fp16_palettized, x = obj_63_cast_fp16)[name = tensor("op_6015_cast_fp16")]; - tensor var_6021_pad_type_0 = const()[name = tensor("op_6021_pad_type_0"), val = tensor("valid")]; - tensor var_6021_strides_0 = const()[name = tensor("op_6021_strides_0"), val = tensor([1, 1])]; - tensor var_6021_pad_0 = const()[name = tensor("op_6021_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_6021_dilations_0 = const()[name = tensor("op_6021_dilations_0"), val = tensor([1, 1])]; - tensor var_6021_groups_0 = const()[name = tensor("op_6021_groups_0"), val = tensor(1)]; - tensor layers_15_self_attn_k_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90821184))), name = tensor("layers_15_self_attn_k_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90813376))), shape = tensor([512, 512, 1, 1])]; - tensor var_6021_cast_fp16 = conv(dilations = var_6021_dilations_0, groups = var_6021_groups_0, pad = var_6021_pad_0, pad_type = var_6021_pad_type_0, strides = var_6021_strides_0, weight = layers_15_self_attn_k_proj_outlier_module_weight_to_fp16_sparsified, x = obj_63_cast_fp16)[name = tensor("op_6021_cast_fp16")]; - tensor key_31_cast_fp16 = add(x = var_6015_cast_fp16, y = var_6021_cast_fp16)[name = tensor("key_31_cast_fp16")]; - tensor var_6031_pad_type_0 = const()[name = tensor("op_6031_pad_type_0"), val = tensor("valid")]; - tensor var_6031_strides_0 = const()[name = tensor("op_6031_strides_0"), val = tensor([1, 1])]; - tensor var_6031_pad_0 = const()[name = tensor("op_6031_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_6031_dilations_0 = const()[name = tensor("op_6031_dilations_0"), val = tensor([1, 1])]; - tensor var_6031_groups_0 = const()[name = tensor("op_6031_groups_0"), val = tensor(1)]; - tensor layers_15_self_attn_v_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90854016))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(91050688))), name = tensor("layers_15_self_attn_v_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_15_self_attn_v_proj_inlier_module_bias_to_fp16 = const()[name = tensor("layers_15_self_attn_v_proj_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(91050880)))]; - tensor var_6031_cast_fp16 = conv(bias = layers_15_self_attn_v_proj_inlier_module_bias_to_fp16, dilations = var_6031_dilations_0, groups = var_6031_groups_0, pad = var_6031_pad_0, pad_type = var_6031_pad_type_0, strides = var_6031_strides_0, weight = layers_15_self_attn_v_proj_inlier_module_weight_to_fp16_palettized, x = obj_63_cast_fp16)[name = tensor("op_6031_cast_fp16")]; - tensor var_6037_pad_type_0 = const()[name = tensor("op_6037_pad_type_0"), val = tensor("valid")]; - tensor var_6037_strides_0 = const()[name = tensor("op_6037_strides_0"), val = tensor([1, 1])]; - tensor var_6037_pad_0 = const()[name = tensor("op_6037_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_6037_dilations_0 = const()[name = tensor("op_6037_dilations_0"), val = tensor([1, 1])]; - tensor var_6037_groups_0 = const()[name = tensor("op_6037_groups_0"), val = tensor(1)]; - tensor layers_15_self_attn_v_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(91060672))), name = tensor("layers_15_self_attn_v_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(91051968))), shape = tensor([512, 512, 1, 1])]; - tensor var_6037_cast_fp16 = conv(dilations = var_6037_dilations_0, groups = var_6037_groups_0, pad = var_6037_pad_0, pad_type = var_6037_pad_type_0, strides = var_6037_strides_0, weight = layers_15_self_attn_v_proj_outlier_module_weight_to_fp16_sparsified, x = obj_63_cast_fp16)[name = tensor("op_6037_cast_fp16")]; - tensor value_31_cast_fp16 = add(x = var_6031_cast_fp16, y = var_6037_cast_fp16)[name = tensor("value_31_cast_fp16")]; - tensor var_6040_to_fp16 = const()[name = tensor("op_6040_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(91093504)))]; - tensor query_63_cast_fp16 = add(x = query_61_cast_fp16, y = var_6040_to_fp16)[name = tensor("query_63_cast_fp16")]; - tensor var_6043_to_fp16 = const()[name = tensor("op_6043_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(91094592)))]; - tensor q_with_bias_v_31_cast_fp16 = add(x = query_61_cast_fp16, y = var_6043_to_fp16)[name = tensor("q_with_bias_v_31_cast_fp16")]; - tensor var_6053_pad_type_0 = const()[name = tensor("op_6053_pad_type_0"), val = tensor("valid")]; - tensor var_6053_strides_0 = const()[name = tensor("op_6053_strides_0"), val = tensor([1, 1])]; - tensor var_6053_pad_0 = const()[name = tensor("op_6053_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_6053_dilations_0 = const()[name = tensor("op_6053_dilations_0"), val = tensor([1, 1])]; - tensor var_6053_groups_0 = const()[name = tensor("op_6053_groups_0"), val = tensor(1)]; - tensor layers_15_self_attn_linear_pos_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(91095680))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(91292352))), name = tensor("layers_15_self_attn_linear_pos_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor var_6053_cast_fp16 = conv(bias = input_17_mean_0_to_fp16, dilations = var_6053_dilations_0, groups = var_6053_groups_0, pad = var_6053_pad_0, pad_type = var_6053_pad_type_0, strides = var_6053_strides_0, weight = layers_15_self_attn_linear_pos_inlier_module_weight_to_fp16_palettized, x = obj_3_cast_fp16)[name = tensor("op_6053_cast_fp16")]; - tensor var_6059_pad_type_0 = const()[name = tensor("op_6059_pad_type_0"), val = tensor("valid")]; - tensor var_6059_strides_0 = const()[name = tensor("op_6059_strides_0"), val = tensor([1, 1])]; - tensor var_6059_pad_0 = const()[name = tensor("op_6059_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_6059_dilations_0 = const()[name = tensor("op_6059_dilations_0"), val = tensor([1, 1])]; - tensor var_6059_groups_0 = const()[name = tensor("op_6059_groups_0"), val = tensor(1)]; - tensor layers_15_self_attn_linear_pos_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(91315264))), name = tensor("layers_15_self_attn_linear_pos_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(91292544))), shape = tensor([512, 512, 1, 1])]; - tensor var_6059_cast_fp16 = conv(dilations = var_6059_dilations_0, groups = var_6059_groups_0, pad = var_6059_pad_0, pad_type = var_6059_pad_type_0, strides = var_6059_strides_0, weight = layers_15_self_attn_linear_pos_outlier_module_weight_to_fp16_sparsified, x = obj_3_cast_fp16)[name = tensor("op_6059_cast_fp16")]; - tensor p_31_cast_fp16 = add(x = var_6053_cast_fp16, y = var_6059_cast_fp16)[name = tensor("p_31_cast_fp16")]; - tensor var_6063 = const()[name = tensor("op_6063"), val = tensor([1, 8, 64, 188])]; - tensor var_6064_cast_fp16 = reshape(shape = var_6063, x = q_with_bias_v_31_cast_fp16)[name = tensor("op_6064_cast_fp16")]; - tensor var_6065 = const()[name = tensor("op_6065"), val = tensor([1, 8, 64, -1])]; - tensor var_6066_cast_fp16 = reshape(shape = var_6065, x = p_31_cast_fp16)[name = tensor("op_6066_cast_fp16")]; - tensor matrix_bd_121_transpose_x_0 = const()[name = tensor("matrix_bd_121_transpose_x_0"), val = tensor(true)]; - tensor matrix_bd_121_transpose_y_0 = const()[name = tensor("matrix_bd_121_transpose_y_0"), val = tensor(false)]; - tensor matrix_bd_121_cast_fp16 = matmul(transpose_x = matrix_bd_121_transpose_x_0, transpose_y = matrix_bd_121_transpose_y_0, x = var_6064_cast_fp16, y = var_6066_cast_fp16)[name = tensor("matrix_bd_121_cast_fp16")]; - tensor matrix_bd_123_pad_0 = const()[name = tensor("matrix_bd_123_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor matrix_bd_123_mode_0 = const()[name = tensor("matrix_bd_123_mode_0"), val = tensor("constant")]; - tensor const_175_to_fp16 = const()[name = tensor("const_175_to_fp16"), val = tensor(0x0p+0)]; - tensor matrix_bd_123_cast_fp16 = pad(constant_val = const_175_to_fp16, mode = matrix_bd_123_mode_0, pad = matrix_bd_123_pad_0, x = matrix_bd_121_cast_fp16)[name = tensor("matrix_bd_123_cast_fp16")]; - tensor var_6075 = const()[name = tensor("op_6075"), val = tensor([1, 8, -1, 188])]; - tensor matrix_bd_125_cast_fp16 = reshape(shape = var_6075, x = matrix_bd_123_cast_fp16)[name = tensor("matrix_bd_125_cast_fp16")]; - tensor var_6079_begin_0 = const()[name = tensor("op_6079_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_6079_end_0 = const()[name = tensor("op_6079_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_6079_end_mask_0 = const()[name = tensor("op_6079_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_6079_cast_fp16 = slice_by_index(begin = var_6079_begin_0, end = var_6079_end_0, end_mask = var_6079_end_mask_0, x = matrix_bd_125_cast_fp16)[name = tensor("op_6079_cast_fp16")]; - tensor var_6080 = const()[name = tensor("op_6080"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_127_cast_fp16 = reshape(shape = var_6080, x = var_6079_cast_fp16)[name = tensor("matrix_bd_127_cast_fp16")]; - tensor var_6085_begin_0 = const()[name = tensor("op_6085_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor var_6085_end_0 = const()[name = tensor("op_6085_end_0"), val = tensor([1, 8, 188, 188])]; - tensor var_6085_end_mask_0 = const()[name = tensor("op_6085_end_mask_0"), val = tensor([true, true, true, false])]; - tensor var_6085_cast_fp16 = slice_by_index(begin = var_6085_begin_0, end = var_6085_end_0, end_mask = var_6085_end_mask_0, x = matrix_bd_127_cast_fp16)[name = tensor("op_6085_cast_fp16")]; - tensor var_6086_to_fp16 = const()[name = tensor("op_6086_to_fp16"), val = tensor(0x1p-3)]; - tensor qk_mask_31_cast_fp16 = mul(x = var_6085_cast_fp16, y = var_6086_to_fp16)[name = tensor("qk_mask_31_cast_fp16")]; - tensor var_6090 = const()[name = tensor("op_6090"), val = tensor([1, 8, 64, 188])]; - tensor mh_q_31_cast_fp16 = reshape(shape = var_6090, x = query_63_cast_fp16)[name = tensor("mh_q_31_cast_fp16")]; - tensor var_6092_to_fp16 = const()[name = tensor("op_6092_to_fp16"), val = tensor(0x1p-3)]; - tensor var_6093_cast_fp16 = mul(x = mh_q_31_cast_fp16, y = var_6092_to_fp16)[name = tensor("op_6093_cast_fp16")]; - tensor var_6096 = const()[name = tensor("op_6096"), val = tensor([1, 8, 64, 188])]; - tensor var_6097_cast_fp16 = reshape(shape = var_6096, x = key_31_cast_fp16)[name = tensor("op_6097_cast_fp16")]; - tensor mh_w_61_transpose_x_0 = const()[name = tensor("mh_w_61_transpose_x_0"), val = tensor(true)]; - tensor mh_w_61_transpose_y_0 = const()[name = tensor("mh_w_61_transpose_y_0"), val = tensor(false)]; - tensor mh_w_61_cast_fp16 = matmul(transpose_x = mh_w_61_transpose_x_0, transpose_y = mh_w_61_transpose_y_0, x = var_6093_cast_fp16, y = var_6097_cast_fp16)[name = tensor("mh_w_61_cast_fp16")]; - tensor mh_w_63_cast_fp16 = add(x = mh_w_61_cast_fp16, y = qk_mask_31_cast_fp16)[name = tensor("mh_w_63_cast_fp16")]; - tensor var_6101_cast_fp16 = softmax(axis = var_5888, x = mh_w_63_cast_fp16)[name = tensor("op_6101_cast_fp16")]; - tensor var_6102 = const()[name = tensor("op_6102"), val = tensor([1, 8, 64, 188])]; - tensor var_6103_cast_fp16 = reshape(shape = var_6102, x = value_31_cast_fp16)[name = tensor("op_6103_cast_fp16")]; - tensor attn_31_transpose_x_0 = const()[name = tensor("attn_31_transpose_x_0"), val = tensor(false)]; - tensor attn_31_transpose_y_0 = const()[name = tensor("attn_31_transpose_y_0"), val = tensor(true)]; - tensor attn_31_cast_fp16 = matmul(transpose_x = attn_31_transpose_x_0, transpose_y = attn_31_transpose_y_0, x = var_6103_cast_fp16, y = var_6101_cast_fp16)[name = tensor("attn_31_cast_fp16")]; - tensor var_6106 = const()[name = tensor("op_6106"), val = tensor([1, 512, 1, 188])]; - tensor input_413_cast_fp16 = reshape(shape = var_6106, x = attn_31_cast_fp16)[name = tensor("input_413_cast_fp16")]; - tensor var_6116_pad_type_0 = const()[name = tensor("op_6116_pad_type_0"), val = tensor("valid")]; - tensor var_6116_strides_0 = const()[name = tensor("op_6116_strides_0"), val = tensor([1, 1])]; - tensor var_6116_pad_0 = const()[name = tensor("op_6116_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_6116_dilations_0 = const()[name = tensor("op_6116_dilations_0"), val = tensor([1, 1])]; - tensor var_6116_groups_0 = const()[name = tensor("op_6116_groups_0"), val = tensor(1)]; - tensor layers_15_self_attn_o_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(91348096))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(91544768))), name = tensor("layers_15_self_attn_o_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_15_self_attn_o_proj_inlier_module_bias_to_fp16 = const()[name = tensor("layers_15_self_attn_o_proj_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(91544960)))]; - tensor var_6116_cast_fp16 = conv(bias = layers_15_self_attn_o_proj_inlier_module_bias_to_fp16, dilations = var_6116_dilations_0, groups = var_6116_groups_0, pad = var_6116_pad_0, pad_type = var_6116_pad_type_0, strides = var_6116_strides_0, weight = layers_15_self_attn_o_proj_inlier_module_weight_to_fp16_palettized, x = input_413_cast_fp16)[name = tensor("op_6116_cast_fp16")]; - tensor var_6122_pad_type_0 = const()[name = tensor("op_6122_pad_type_0"), val = tensor("valid")]; - tensor var_6122_strides_0 = const()[name = tensor("op_6122_strides_0"), val = tensor([1, 1])]; - tensor var_6122_pad_0 = const()[name = tensor("op_6122_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_6122_dilations_0 = const()[name = tensor("op_6122_dilations_0"), val = tensor([1, 1])]; - tensor var_6122_groups_0 = const()[name = tensor("op_6122_groups_0"), val = tensor(1)]; - tensor layers_15_self_attn_o_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(91554496))), name = tensor("layers_15_self_attn_o_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(91546048))), shape = tensor([512, 512, 1, 1])]; - tensor var_6122_cast_fp16 = conv(dilations = var_6122_dilations_0, groups = var_6122_groups_0, pad = var_6122_pad_0, pad_type = var_6122_pad_type_0, strides = var_6122_strides_0, weight = layers_15_self_attn_o_proj_outlier_module_weight_to_fp16_sparsified, x = input_413_cast_fp16)[name = tensor("op_6122_cast_fp16")]; - tensor obj_65_cast_fp16 = add(x = var_6116_cast_fp16, y = var_6122_cast_fp16)[name = tensor("obj_65_cast_fp16")]; - tensor inputs_155_cast_fp16 = add(x = inputs_153_cast_fp16, y = obj_65_cast_fp16)[name = tensor("inputs_155_cast_fp16")]; - tensor out_155_axes_0 = const()[name = tensor("out_155_axes_0"), val = tensor([1])]; - tensor var_6133_to_fp16 = const()[name = tensor("op_6133_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_155_cast_fp16 = layer_norm(axes = out_155_axes_0, epsilon = var_6133_to_fp16, x = inputs_155_cast_fp16)[name = tensor("out_155_cast_fp16")]; - tensor input_415_gamma_0_to_fp16 = const()[name = tensor("input_415_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(91587328)))]; - tensor input_415_beta_0_to_fp16 = const()[name = tensor("input_415_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(91588416)))]; - tensor input_415_epsilon_0_to_fp16 = const()[name = tensor("input_415_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_415_cast_fp16 = batch_norm(beta = input_415_beta_0_to_fp16, epsilon = input_415_epsilon_0_to_fp16, gamma = input_415_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_155_cast_fp16)[name = tensor("input_415_cast_fp16")]; - tensor var_6155_pad_type_0 = const()[name = tensor("op_6155_pad_type_0"), val = tensor("valid")]; - tensor var_6155_strides_0 = const()[name = tensor("op_6155_strides_0"), val = tensor([1, 1])]; - tensor var_6155_pad_0 = const()[name = tensor("op_6155_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_6155_dilations_0 = const()[name = tensor("op_6155_dilations_0"), val = tensor([1, 1])]; - tensor var_6155_groups_0 = const()[name = tensor("op_6155_groups_0"), val = tensor(1)]; - tensor layers_15_conv_pointwise_conv1_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(91589504))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(91982784))), name = tensor("layers_15_conv_pointwise_conv1_inlier_module_weight_to_fp16_palettized"), shape = tensor([1024, 512, 1, 1])]; - tensor layers_15_conv_pointwise_conv1_inlier_module_bias_to_fp16 = const()[name = tensor("layers_15_conv_pointwise_conv1_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(91982976)))]; - tensor var_6155_cast_fp16 = conv(bias = layers_15_conv_pointwise_conv1_inlier_module_bias_to_fp16, dilations = var_6155_dilations_0, groups = var_6155_groups_0, pad = var_6155_pad_0, pad_type = var_6155_pad_type_0, strides = var_6155_strides_0, weight = layers_15_conv_pointwise_conv1_inlier_module_weight_to_fp16_palettized, x = input_415_cast_fp16)[name = tensor("op_6155_cast_fp16")]; - tensor var_6161_pad_type_0 = const()[name = tensor("op_6161_pad_type_0"), val = tensor("valid")]; - tensor var_6161_strides_0 = const()[name = tensor("op_6161_strides_0"), val = tensor([1, 1])]; - tensor var_6161_pad_0 = const()[name = tensor("op_6161_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_6161_dilations_0 = const()[name = tensor("op_6161_dilations_0"), val = tensor([1, 1])]; - tensor var_6161_groups_0 = const()[name = tensor("op_6161_groups_0"), val = tensor(1)]; - tensor layers_15_conv_pointwise_conv1_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92003456))), name = tensor("layers_15_conv_pointwise_conv1_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(91985088))), shape = tensor([1024, 512, 1, 1])]; - tensor var_6161_cast_fp16 = conv(dilations = var_6161_dilations_0, groups = var_6161_groups_0, pad = var_6161_pad_0, pad_type = var_6161_pad_type_0, strides = var_6161_strides_0, weight = layers_15_conv_pointwise_conv1_outlier_module_weight_to_fp16_sparsified, x = input_415_cast_fp16)[name = tensor("op_6161_cast_fp16")]; - tensor input_417_cast_fp16 = add(x = var_6155_cast_fp16, y = var_6161_cast_fp16)[name = tensor("input_417_cast_fp16")]; - tensor input_419_split_num_splits_0 = const()[name = tensor("input_419_split_num_splits_0"), val = tensor(2)]; - tensor input_419_split_axis_0 = const()[name = tensor("input_419_split_axis_0"), val = tensor(1)]; - tensor input_419_split_cast_fp16_0, tensor input_419_split_cast_fp16_1 = split(axis = input_419_split_axis_0, num_splits = input_419_split_num_splits_0, x = input_417_cast_fp16)[name = tensor("input_419_split_cast_fp16")]; - tensor input_419_split_1_sigmoid_cast_fp16 = sigmoid(x = input_419_split_cast_fp16_1)[name = tensor("input_419_split_1_sigmoid_cast_fp16")]; - tensor input_419_cast_fp16 = mul(x = input_419_split_cast_fp16_0, y = input_419_split_1_sigmoid_cast_fp16)[name = tensor("input_419_cast_fp16")]; - tensor input_421_pad_type_0 = const()[name = tensor("input_421_pad_type_0"), val = tensor("custom")]; - tensor input_421_pad_0 = const()[name = tensor("input_421_pad_0"), val = tensor([0, 0, 4, 4])]; - tensor input_421_groups_0 = const()[name = tensor("input_421_groups_0"), val = tensor(512)]; - tensor input_421_strides_0 = const()[name = tensor("input_421_strides_0"), val = tensor([1, 1])]; - tensor input_421_dilations_0 = const()[name = tensor("input_421_dilations_0"), val = tensor([1, 1])]; - tensor const_221_to_fp16 = const()[name = tensor("const_221_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92069056)))]; - tensor const_222_to_fp16 = const()[name = tensor("const_222_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92078336)))]; - tensor input_423_cast_fp16 = conv(bias = const_222_to_fp16, dilations = input_421_dilations_0, groups = input_421_groups_0, pad = input_421_pad_0, pad_type = input_421_pad_type_0, strides = input_421_strides_0, weight = const_221_to_fp16, x = input_419_cast_fp16)[name = tensor("input_423_cast_fp16")]; - tensor input_425_cast_fp16 = silu(x = input_423_cast_fp16)[name = tensor("input_425_cast_fp16")]; - tensor var_6185_pad_type_0 = const()[name = tensor("op_6185_pad_type_0"), val = tensor("valid")]; - tensor var_6185_strides_0 = const()[name = tensor("op_6185_strides_0"), val = tensor([1, 1])]; - tensor var_6185_pad_0 = const()[name = tensor("op_6185_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_6185_dilations_0 = const()[name = tensor("op_6185_dilations_0"), val = tensor([1, 1])]; - tensor var_6185_groups_0 = const()[name = tensor("op_6185_groups_0"), val = tensor(1)]; - tensor layers_15_conv_pointwise_conv2_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92079424))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92276096))), name = tensor("layers_15_conv_pointwise_conv2_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_15_conv_pointwise_conv2_inlier_module_bias_to_fp16 = const()[name = tensor("layers_15_conv_pointwise_conv2_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92276288)))]; - tensor var_6185_cast_fp16 = conv(bias = layers_15_conv_pointwise_conv2_inlier_module_bias_to_fp16, dilations = var_6185_dilations_0, groups = var_6185_groups_0, pad = var_6185_pad_0, pad_type = var_6185_pad_type_0, strides = var_6185_strides_0, weight = layers_15_conv_pointwise_conv2_inlier_module_weight_to_fp16_palettized, x = input_425_cast_fp16)[name = tensor("op_6185_cast_fp16")]; - tensor var_6191_pad_type_0 = const()[name = tensor("op_6191_pad_type_0"), val = tensor("valid")]; - tensor var_6191_strides_0 = const()[name = tensor("op_6191_strides_0"), val = tensor([1, 1])]; - tensor var_6191_pad_0 = const()[name = tensor("op_6191_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_6191_dilations_0 = const()[name = tensor("op_6191_dilations_0"), val = tensor([1, 1])]; - tensor var_6191_groups_0 = const()[name = tensor("op_6191_groups_0"), val = tensor(1)]; - tensor layers_15_conv_pointwise_conv2_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92285568))), name = tensor("layers_15_conv_pointwise_conv2_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92277376))), shape = tensor([512, 512, 1, 1])]; - tensor var_6191_cast_fp16 = conv(dilations = var_6191_dilations_0, groups = var_6191_groups_0, pad = var_6191_pad_0, pad_type = var_6191_pad_type_0, strides = var_6191_strides_0, weight = layers_15_conv_pointwise_conv2_outlier_module_weight_to_fp16_sparsified, x = input_425_cast_fp16)[name = tensor("op_6191_cast_fp16")]; - tensor x_95_cast_fp16 = add(x = var_6185_cast_fp16, y = var_6191_cast_fp16)[name = tensor("x_95_cast_fp16")]; - tensor inputs_157_cast_fp16 = add(x = inputs_155_cast_fp16, y = x_95_cast_fp16)[name = tensor("inputs_157_cast_fp16")]; - tensor out_157_axes_0 = const()[name = tensor("out_157_axes_0"), val = tensor([1])]; - tensor var_6202_to_fp16 = const()[name = tensor("op_6202_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_157_cast_fp16 = layer_norm(axes = out_157_axes_0, epsilon = var_6202_to_fp16, x = inputs_157_cast_fp16)[name = tensor("out_157_cast_fp16")]; - tensor input_427_gamma_0_to_fp16 = const()[name = tensor("input_427_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92318400)))]; - tensor input_427_beta_0_to_fp16 = const()[name = tensor("input_427_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92319488)))]; - tensor input_427_epsilon_0_to_fp16 = const()[name = tensor("input_427_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_427_cast_fp16 = batch_norm(beta = input_427_beta_0_to_fp16, epsilon = input_427_epsilon_0_to_fp16, gamma = input_427_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_157_cast_fp16)[name = tensor("input_427_cast_fp16")]; - tensor var_6222_pad_type_0 = const()[name = tensor("op_6222_pad_type_0"), val = tensor("valid")]; - tensor var_6222_strides_0 = const()[name = tensor("op_6222_strides_0"), val = tensor([1, 1])]; - tensor var_6222_pad_0 = const()[name = tensor("op_6222_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_6222_dilations_0 = const()[name = tensor("op_6222_dilations_0"), val = tensor([1, 1])]; - tensor var_6222_groups_0 = const()[name = tensor("op_6222_groups_0"), val = tensor(1)]; - tensor layers_15_feed_forward2_fc1_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92320576))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(93107072))), name = tensor("layers_15_feed_forward2_fc1_inlier_module_weight_to_fp16_palettized"), shape = tensor([2048, 512, 1, 1])]; - tensor layers_15_feed_forward2_fc1_inlier_module_bias_to_fp16 = const()[name = tensor("layers_15_feed_forward2_fc1_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(93107264)))]; - tensor var_6222_cast_fp16 = conv(bias = layers_15_feed_forward2_fc1_inlier_module_bias_to_fp16, dilations = var_6222_dilations_0, groups = var_6222_groups_0, pad = var_6222_pad_0, pad_type = var_6222_pad_type_0, strides = var_6222_strides_0, weight = layers_15_feed_forward2_fc1_inlier_module_weight_to_fp16_palettized, x = input_427_cast_fp16)[name = tensor("op_6222_cast_fp16")]; - tensor var_6228_pad_type_0 = const()[name = tensor("op_6228_pad_type_0"), val = tensor("valid")]; - tensor var_6228_strides_0 = const()[name = tensor("op_6228_strides_0"), val = tensor([1, 1])]; - tensor var_6228_pad_0 = const()[name = tensor("op_6228_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_6228_dilations_0 = const()[name = tensor("op_6228_dilations_0"), val = tensor([1, 1])]; - tensor var_6228_groups_0 = const()[name = tensor("op_6228_groups_0"), val = tensor(1)]; - tensor layers_15_feed_forward2_fc1_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(93144704))), name = tensor("layers_15_feed_forward2_fc1_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(93111424))), shape = tensor([2048, 512, 1, 1])]; - tensor var_6228_cast_fp16 = conv(dilations = var_6228_dilations_0, groups = var_6228_groups_0, pad = var_6228_pad_0, pad_type = var_6228_pad_type_0, strides = var_6228_strides_0, weight = layers_15_feed_forward2_fc1_outlier_module_weight_to_fp16_sparsified, x = input_427_cast_fp16)[name = tensor("op_6228_cast_fp16")]; - tensor input_429_cast_fp16 = add(x = var_6222_cast_fp16, y = var_6228_cast_fp16)[name = tensor("input_429_cast_fp16")]; - tensor input_431_cast_fp16 = silu(x = input_429_cast_fp16)[name = tensor("input_431_cast_fp16")]; - tensor var_6239_pad_type_0 = const()[name = tensor("op_6239_pad_type_0"), val = tensor("valid")]; - tensor var_6239_strides_0 = const()[name = tensor("op_6239_strides_0"), val = tensor([1, 1])]; - tensor var_6239_pad_0 = const()[name = tensor("op_6239_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_6239_dilations_0 = const()[name = tensor("op_6239_dilations_0"), val = tensor([1, 1])]; - tensor var_6239_groups_0 = const()[name = tensor("op_6239_groups_0"), val = tensor(1)]; - tensor layers_15_feed_forward2_fc2_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(93275840))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(94062336))), name = tensor("layers_15_feed_forward2_fc2_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 2048, 1, 1])]; - tensor layers_15_feed_forward2_fc2_inlier_module_bias_to_fp16 = const()[name = tensor("layers_15_feed_forward2_fc2_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(94062528)))]; - tensor var_6239_cast_fp16 = conv(bias = layers_15_feed_forward2_fc2_inlier_module_bias_to_fp16, dilations = var_6239_dilations_0, groups = var_6239_groups_0, pad = var_6239_pad_0, pad_type = var_6239_pad_type_0, strides = var_6239_strides_0, weight = layers_15_feed_forward2_fc2_inlier_module_weight_to_fp16_palettized, x = input_431_cast_fp16)[name = tensor("op_6239_cast_fp16")]; - tensor var_6245_pad_type_0 = const()[name = tensor("op_6245_pad_type_0"), val = tensor("valid")]; - tensor var_6245_strides_0 = const()[name = tensor("op_6245_strides_0"), val = tensor([1, 1])]; - tensor var_6245_pad_0 = const()[name = tensor("op_6245_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_6245_dilations_0 = const()[name = tensor("op_6245_dilations_0"), val = tensor([1, 1])]; - tensor var_6245_groups_0 = const()[name = tensor("op_6245_groups_0"), val = tensor(1)]; - tensor layers_15_feed_forward2_fc2_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(94104576))), name = tensor("layers_15_feed_forward2_fc2_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(94063616))), shape = tensor([512, 2048, 1, 1])]; - tensor var_6245_cast_fp16 = conv(dilations = var_6245_dilations_0, groups = var_6245_groups_0, pad = var_6245_pad_0, pad_type = var_6245_pad_type_0, strides = var_6245_strides_0, weight = layers_15_feed_forward2_fc2_outlier_module_weight_to_fp16_sparsified, x = input_431_cast_fp16)[name = tensor("op_6245_cast_fp16")]; - tensor x_97_cast_fp16 = add(x = var_6239_cast_fp16, y = var_6245_cast_fp16)[name = tensor("x_97_cast_fp16")]; - tensor var_6247_to_fp16 = const()[name = tensor("op_6247_to_fp16"), val = tensor(0x1p-1)]; - tensor var_6248_cast_fp16 = mul(x = x_97_cast_fp16, y = var_6247_to_fp16)[name = tensor("op_6248_cast_fp16")]; - tensor inputs_159_cast_fp16 = add(x = inputs_157_cast_fp16, y = var_6248_cast_fp16)[name = tensor("inputs_159_cast_fp16")]; - tensor out_159_axes_0 = const()[name = tensor("out_159_axes_0"), val = tensor([1])]; - tensor var_6258_to_fp16 = const()[name = tensor("op_6258_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_159_cast_fp16 = layer_norm(axes = out_159_axes_0, epsilon = var_6258_to_fp16, x = inputs_159_cast_fp16)[name = tensor("out_159_cast_fp16")]; - tensor inputs_161_gamma_0_to_fp16 = const()[name = tensor("inputs_161_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(94235712)))]; - tensor inputs_161_beta_0_to_fp16 = const()[name = tensor("inputs_161_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(94236800)))]; - tensor inputs_161_epsilon_0_to_fp16 = const()[name = tensor("inputs_161_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor inputs_161_cast_fp16 = batch_norm(beta = inputs_161_beta_0_to_fp16, epsilon = inputs_161_epsilon_0_to_fp16, gamma = inputs_161_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_159_cast_fp16)[name = tensor("inputs_161_cast_fp16")]; - tensor var_6272 = const()[name = tensor("op_6272"), val = tensor(3)]; - tensor out_161_axes_0 = const()[name = tensor("out_161_axes_0"), val = tensor([1])]; - tensor var_6303_to_fp16 = const()[name = tensor("op_6303_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_161_cast_fp16 = layer_norm(axes = out_161_axes_0, epsilon = var_6303_to_fp16, x = inputs_161_cast_fp16)[name = tensor("out_161_cast_fp16")]; - tensor input_433_gamma_0_to_fp16 = const()[name = tensor("input_433_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(94237888)))]; - tensor input_433_beta_0_to_fp16 = const()[name = tensor("input_433_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(94238976)))]; - tensor input_433_epsilon_0_to_fp16 = const()[name = tensor("input_433_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_433_cast_fp16 = batch_norm(beta = input_433_beta_0_to_fp16, epsilon = input_433_epsilon_0_to_fp16, gamma = input_433_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_161_cast_fp16)[name = tensor("input_433_cast_fp16")]; - tensor var_6323_pad_type_0 = const()[name = tensor("op_6323_pad_type_0"), val = tensor("valid")]; - tensor var_6323_strides_0 = const()[name = tensor("op_6323_strides_0"), val = tensor([1, 1])]; - tensor var_6323_pad_0 = const()[name = tensor("op_6323_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_6323_dilations_0 = const()[name = tensor("op_6323_dilations_0"), val = tensor([1, 1])]; - tensor var_6323_groups_0 = const()[name = tensor("op_6323_groups_0"), val = tensor(1)]; - tensor layers_16_feed_forward1_fc1_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(94240064))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95026560))), name = tensor("layers_16_feed_forward1_fc1_inlier_module_weight_to_fp16_palettized"), shape = tensor([2048, 512, 1, 1])]; - tensor layers_16_feed_forward1_fc1_inlier_module_bias_to_fp16 = const()[name = tensor("layers_16_feed_forward1_fc1_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95026752)))]; - tensor var_6323_cast_fp16 = conv(bias = layers_16_feed_forward1_fc1_inlier_module_bias_to_fp16, dilations = var_6323_dilations_0, groups = var_6323_groups_0, pad = var_6323_pad_0, pad_type = var_6323_pad_type_0, strides = var_6323_strides_0, weight = layers_16_feed_forward1_fc1_inlier_module_weight_to_fp16_palettized, x = input_433_cast_fp16)[name = tensor("op_6323_cast_fp16")]; - tensor var_6329_pad_type_0 = const()[name = tensor("op_6329_pad_type_0"), val = tensor("valid")]; - tensor var_6329_strides_0 = const()[name = tensor("op_6329_strides_0"), val = tensor([1, 1])]; - tensor var_6329_pad_0 = const()[name = tensor("op_6329_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_6329_dilations_0 = const()[name = tensor("op_6329_dilations_0"), val = tensor([1, 1])]; - tensor var_6329_groups_0 = const()[name = tensor("op_6329_groups_0"), val = tensor(1)]; - tensor layers_16_feed_forward1_fc1_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95064576))), name = tensor("layers_16_feed_forward1_fc1_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95030912))), shape = tensor([2048, 512, 1, 1])]; - tensor var_6329_cast_fp16 = conv(dilations = var_6329_dilations_0, groups = var_6329_groups_0, pad = var_6329_pad_0, pad_type = var_6329_pad_type_0, strides = var_6329_strides_0, weight = layers_16_feed_forward1_fc1_outlier_module_weight_to_fp16_sparsified, x = input_433_cast_fp16)[name = tensor("op_6329_cast_fp16")]; - tensor input_435_cast_fp16 = add(x = var_6323_cast_fp16, y = var_6329_cast_fp16)[name = tensor("input_435_cast_fp16")]; - tensor input_437_cast_fp16 = silu(x = input_435_cast_fp16)[name = tensor("input_437_cast_fp16")]; - tensor var_6340_pad_type_0 = const()[name = tensor("op_6340_pad_type_0"), val = tensor("valid")]; - tensor var_6340_strides_0 = const()[name = tensor("op_6340_strides_0"), val = tensor([1, 1])]; - tensor var_6340_pad_0 = const()[name = tensor("op_6340_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_6340_dilations_0 = const()[name = tensor("op_6340_dilations_0"), val = tensor([1, 1])]; - tensor var_6340_groups_0 = const()[name = tensor("op_6340_groups_0"), val = tensor(1)]; - tensor layers_16_feed_forward1_fc2_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95195712))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95982208))), name = tensor("layers_16_feed_forward1_fc2_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 2048, 1, 1])]; - tensor layers_16_feed_forward1_fc2_inlier_module_bias_to_fp16 = const()[name = tensor("layers_16_feed_forward1_fc2_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95982400)))]; - tensor var_6340_cast_fp16 = conv(bias = layers_16_feed_forward1_fc2_inlier_module_bias_to_fp16, dilations = var_6340_dilations_0, groups = var_6340_groups_0, pad = var_6340_pad_0, pad_type = var_6340_pad_type_0, strides = var_6340_strides_0, weight = layers_16_feed_forward1_fc2_inlier_module_weight_to_fp16_palettized, x = input_437_cast_fp16)[name = tensor("op_6340_cast_fp16")]; - tensor var_6346_pad_type_0 = const()[name = tensor("op_6346_pad_type_0"), val = tensor("valid")]; - tensor var_6346_strides_0 = const()[name = tensor("op_6346_strides_0"), val = tensor([1, 1])]; - tensor var_6346_pad_0 = const()[name = tensor("op_6346_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_6346_dilations_0 = const()[name = tensor("op_6346_dilations_0"), val = tensor([1, 1])]; - tensor var_6346_groups_0 = const()[name = tensor("op_6346_groups_0"), val = tensor(1)]; - tensor layers_16_feed_forward1_fc2_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96026752))), name = tensor("layers_16_feed_forward1_fc2_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95983488))), shape = tensor([512, 2048, 1, 1])]; - tensor var_6346_cast_fp16 = conv(dilations = var_6346_dilations_0, groups = var_6346_groups_0, pad = var_6346_pad_0, pad_type = var_6346_pad_type_0, strides = var_6346_strides_0, weight = layers_16_feed_forward1_fc2_outlier_module_weight_to_fp16_sparsified, x = input_437_cast_fp16)[name = tensor("op_6346_cast_fp16")]; - tensor x_99_cast_fp16 = add(x = var_6340_cast_fp16, y = var_6346_cast_fp16)[name = tensor("x_99_cast_fp16")]; - tensor var_6348_to_fp16 = const()[name = tensor("op_6348_to_fp16"), val = tensor(0x1p-1)]; - tensor var_6349_cast_fp16 = mul(x = x_99_cast_fp16, y = var_6348_to_fp16)[name = tensor("op_6349_cast_fp16")]; - tensor inputs_163_cast_fp16 = add(x = inputs_161_cast_fp16, y = var_6349_cast_fp16)[name = tensor("inputs_163_cast_fp16")]; - tensor out_163_axes_0 = const()[name = tensor("out_163_axes_0"), val = tensor([1])]; - tensor var_6359_to_fp16 = const()[name = tensor("op_6359_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_163_cast_fp16 = layer_norm(axes = out_163_axes_0, epsilon = var_6359_to_fp16, x = inputs_163_cast_fp16)[name = tensor("out_163_cast_fp16")]; - tensor obj_67_gamma_0_to_fp16 = const()[name = tensor("obj_67_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96157888)))]; - tensor obj_67_beta_0_to_fp16 = const()[name = tensor("obj_67_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96158976)))]; - tensor obj_67_epsilon_0_to_fp16 = const()[name = tensor("obj_67_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor obj_67_cast_fp16 = batch_norm(beta = obj_67_beta_0_to_fp16, epsilon = obj_67_epsilon_0_to_fp16, gamma = obj_67_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_163_cast_fp16)[name = tensor("obj_67_cast_fp16")]; - tensor var_6384_pad_type_0 = const()[name = tensor("op_6384_pad_type_0"), val = tensor("valid")]; - tensor var_6384_strides_0 = const()[name = tensor("op_6384_strides_0"), val = tensor([1, 1])]; - tensor var_6384_pad_0 = const()[name = tensor("op_6384_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_6384_dilations_0 = const()[name = tensor("op_6384_dilations_0"), val = tensor([1, 1])]; - tensor var_6384_groups_0 = const()[name = tensor("op_6384_groups_0"), val = tensor(1)]; - tensor layers_16_self_attn_q_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96160064))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96356736))), name = tensor("layers_16_self_attn_q_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_16_self_attn_q_proj_inlier_module_bias_to_fp16 = const()[name = tensor("layers_16_self_attn_q_proj_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96356928)))]; - tensor var_6384_cast_fp16 = conv(bias = layers_16_self_attn_q_proj_inlier_module_bias_to_fp16, dilations = var_6384_dilations_0, groups = var_6384_groups_0, pad = var_6384_pad_0, pad_type = var_6384_pad_type_0, strides = var_6384_strides_0, weight = layers_16_self_attn_q_proj_inlier_module_weight_to_fp16_palettized, x = obj_67_cast_fp16)[name = tensor("op_6384_cast_fp16")]; - tensor var_6390_pad_type_0 = const()[name = tensor("op_6390_pad_type_0"), val = tensor("valid")]; - tensor var_6390_strides_0 = const()[name = tensor("op_6390_strides_0"), val = tensor([1, 1])]; - tensor var_6390_pad_0 = const()[name = tensor("op_6390_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_6390_dilations_0 = const()[name = tensor("op_6390_dilations_0"), val = tensor([1, 1])]; - tensor var_6390_groups_0 = const()[name = tensor("op_6390_groups_0"), val = tensor(1)]; - tensor layers_16_self_attn_q_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96365312))), name = tensor("layers_16_self_attn_q_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96358016))), shape = tensor([512, 512, 1, 1])]; - tensor var_6390_cast_fp16 = conv(dilations = var_6390_dilations_0, groups = var_6390_groups_0, pad = var_6390_pad_0, pad_type = var_6390_pad_type_0, strides = var_6390_strides_0, weight = layers_16_self_attn_q_proj_outlier_module_weight_to_fp16_sparsified, x = obj_67_cast_fp16)[name = tensor("op_6390_cast_fp16")]; - tensor query_65_cast_fp16 = add(x = var_6384_cast_fp16, y = var_6390_cast_fp16)[name = tensor("query_65_cast_fp16")]; - tensor var_6399_pad_type_0 = const()[name = tensor("op_6399_pad_type_0"), val = tensor("valid")]; - tensor var_6399_strides_0 = const()[name = tensor("op_6399_strides_0"), val = tensor([1, 1])]; - tensor var_6399_pad_0 = const()[name = tensor("op_6399_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_6399_dilations_0 = const()[name = tensor("op_6399_dilations_0"), val = tensor([1, 1])]; - tensor var_6399_groups_0 = const()[name = tensor("op_6399_groups_0"), val = tensor(1)]; - tensor layers_16_self_attn_k_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96398144))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96594816))), name = tensor("layers_16_self_attn_k_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor var_6399_cast_fp16 = conv(dilations = var_6399_dilations_0, groups = var_6399_groups_0, pad = var_6399_pad_0, pad_type = var_6399_pad_type_0, strides = var_6399_strides_0, weight = layers_16_self_attn_k_proj_inlier_module_weight_to_fp16_palettized, x = obj_67_cast_fp16)[name = tensor("op_6399_cast_fp16")]; - tensor var_6405_pad_type_0 = const()[name = tensor("op_6405_pad_type_0"), val = tensor("valid")]; - tensor var_6405_strides_0 = const()[name = tensor("op_6405_strides_0"), val = tensor([1, 1])]; - tensor var_6405_pad_0 = const()[name = tensor("op_6405_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_6405_dilations_0 = const()[name = tensor("op_6405_dilations_0"), val = tensor([1, 1])]; - tensor var_6405_groups_0 = const()[name = tensor("op_6405_groups_0"), val = tensor(1)]; - tensor layers_16_self_attn_k_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96602560))), name = tensor("layers_16_self_attn_k_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96595008))), shape = tensor([512, 512, 1, 1])]; - tensor var_6405_cast_fp16 = conv(dilations = var_6405_dilations_0, groups = var_6405_groups_0, pad = var_6405_pad_0, pad_type = var_6405_pad_type_0, strides = var_6405_strides_0, weight = layers_16_self_attn_k_proj_outlier_module_weight_to_fp16_sparsified, x = obj_67_cast_fp16)[name = tensor("op_6405_cast_fp16")]; - tensor key_cast_fp16 = add(x = var_6399_cast_fp16, y = var_6405_cast_fp16)[name = tensor("key_cast_fp16")]; - tensor var_6415_pad_type_0 = const()[name = tensor("op_6415_pad_type_0"), val = tensor("valid")]; - tensor var_6415_strides_0 = const()[name = tensor("op_6415_strides_0"), val = tensor([1, 1])]; - tensor var_6415_pad_0 = const()[name = tensor("op_6415_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_6415_dilations_0 = const()[name = tensor("op_6415_dilations_0"), val = tensor([1, 1])]; - tensor var_6415_groups_0 = const()[name = tensor("op_6415_groups_0"), val = tensor(1)]; - tensor layers_16_self_attn_v_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96635392))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96832064))), name = tensor("layers_16_self_attn_v_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_16_self_attn_v_proj_inlier_module_bias_to_fp16 = const()[name = tensor("layers_16_self_attn_v_proj_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96832256)))]; - tensor var_6415_cast_fp16 = conv(bias = layers_16_self_attn_v_proj_inlier_module_bias_to_fp16, dilations = var_6415_dilations_0, groups = var_6415_groups_0, pad = var_6415_pad_0, pad_type = var_6415_pad_type_0, strides = var_6415_strides_0, weight = layers_16_self_attn_v_proj_inlier_module_weight_to_fp16_palettized, x = obj_67_cast_fp16)[name = tensor("op_6415_cast_fp16")]; - tensor var_6421_pad_type_0 = const()[name = tensor("op_6421_pad_type_0"), val = tensor("valid")]; - tensor var_6421_strides_0 = const()[name = tensor("op_6421_strides_0"), val = tensor([1, 1])]; - tensor var_6421_pad_0 = const()[name = tensor("op_6421_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_6421_dilations_0 = const()[name = tensor("op_6421_dilations_0"), val = tensor([1, 1])]; - tensor var_6421_groups_0 = const()[name = tensor("op_6421_groups_0"), val = tensor(1)]; - tensor layers_16_self_attn_v_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96841664))), name = tensor("layers_16_self_attn_v_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96833344))), shape = tensor([512, 512, 1, 1])]; - tensor var_6421_cast_fp16 = conv(dilations = var_6421_dilations_0, groups = var_6421_groups_0, pad = var_6421_pad_0, pad_type = var_6421_pad_type_0, strides = var_6421_strides_0, weight = layers_16_self_attn_v_proj_outlier_module_weight_to_fp16_sparsified, x = obj_67_cast_fp16)[name = tensor("op_6421_cast_fp16")]; - tensor value_cast_fp16 = add(x = var_6415_cast_fp16, y = var_6421_cast_fp16)[name = tensor("value_cast_fp16")]; - tensor var_6424_to_fp16 = const()[name = tensor("op_6424_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96874496)))]; - tensor query_cast_fp16 = add(x = query_65_cast_fp16, y = var_6424_to_fp16)[name = tensor("query_cast_fp16")]; - tensor var_6427_to_fp16 = const()[name = tensor("op_6427_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96875584)))]; - tensor q_with_bias_v_cast_fp16 = add(x = query_65_cast_fp16, y = var_6427_to_fp16)[name = tensor("q_with_bias_v_cast_fp16")]; - tensor var_6437_pad_type_0 = const()[name = tensor("op_6437_pad_type_0"), val = tensor("valid")]; - tensor var_6437_strides_0 = const()[name = tensor("op_6437_strides_0"), val = tensor([1, 1])]; - tensor var_6437_pad_0 = const()[name = tensor("op_6437_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_6437_dilations_0 = const()[name = tensor("op_6437_dilations_0"), val = tensor([1, 1])]; - tensor var_6437_groups_0 = const()[name = tensor("op_6437_groups_0"), val = tensor(1)]; - tensor layers_16_self_attn_linear_pos_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96876672))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(97073344))), name = tensor("layers_16_self_attn_linear_pos_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor var_6437_cast_fp16 = conv(bias = input_17_mean_0_to_fp16, dilations = var_6437_dilations_0, groups = var_6437_groups_0, pad = var_6437_pad_0, pad_type = var_6437_pad_type_0, strides = var_6437_strides_0, weight = layers_16_self_attn_linear_pos_inlier_module_weight_to_fp16_palettized, x = obj_3_cast_fp16)[name = tensor("op_6437_cast_fp16")]; - tensor var_6443_pad_type_0 = const()[name = tensor("op_6443_pad_type_0"), val = tensor("valid")]; - tensor var_6443_strides_0 = const()[name = tensor("op_6443_strides_0"), val = tensor([1, 1])]; - tensor var_6443_pad_0 = const()[name = tensor("op_6443_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_6443_dilations_0 = const()[name = tensor("op_6443_dilations_0"), val = tensor([1, 1])]; - tensor var_6443_groups_0 = const()[name = tensor("op_6443_groups_0"), val = tensor(1)]; - tensor layers_16_self_attn_linear_pos_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(97094016))), name = tensor("layers_16_self_attn_linear_pos_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(97073536))), shape = tensor([512, 512, 1, 1])]; - tensor var_6443_cast_fp16 = conv(dilations = var_6443_dilations_0, groups = var_6443_groups_0, pad = var_6443_pad_0, pad_type = var_6443_pad_type_0, strides = var_6443_strides_0, weight = layers_16_self_attn_linear_pos_outlier_module_weight_to_fp16_sparsified, x = obj_3_cast_fp16)[name = tensor("op_6443_cast_fp16")]; - tensor p_cast_fp16 = add(x = var_6437_cast_fp16, y = var_6443_cast_fp16)[name = tensor("p_cast_fp16")]; - tensor var_6447 = const()[name = tensor("op_6447"), val = tensor([1, 8, 64, 188])]; - tensor var_6448_cast_fp16 = reshape(shape = var_6447, x = q_with_bias_v_cast_fp16)[name = tensor("op_6448_cast_fp16")]; - tensor var_6449 = const()[name = tensor("op_6449"), val = tensor([1, 8, 64, -1])]; - tensor var_6450_cast_fp16 = reshape(shape = var_6449, x = p_cast_fp16)[name = tensor("op_6450_cast_fp16")]; - tensor matrix_bd_129_transpose_x_0 = const()[name = tensor("matrix_bd_129_transpose_x_0"), val = tensor(true)]; - tensor matrix_bd_129_transpose_y_0 = const()[name = tensor("matrix_bd_129_transpose_y_0"), val = tensor(false)]; - tensor matrix_bd_129_cast_fp16 = matmul(transpose_x = matrix_bd_129_transpose_x_0, transpose_y = matrix_bd_129_transpose_y_0, x = var_6448_cast_fp16, y = var_6450_cast_fp16)[name = tensor("matrix_bd_129_cast_fp16")]; - tensor matrix_bd_131_pad_0 = const()[name = tensor("matrix_bd_131_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor matrix_bd_131_mode_0 = const()[name = tensor("matrix_bd_131_mode_0"), val = tensor("constant")]; - tensor const_186_to_fp16 = const()[name = tensor("const_186_to_fp16"), val = tensor(0x0p+0)]; - tensor matrix_bd_131_cast_fp16 = pad(constant_val = const_186_to_fp16, mode = matrix_bd_131_mode_0, pad = matrix_bd_131_pad_0, x = matrix_bd_129_cast_fp16)[name = tensor("matrix_bd_131_cast_fp16")]; - tensor var_6459 = const()[name = tensor("op_6459"), val = tensor([1, 8, -1, 188])]; - tensor matrix_bd_133_cast_fp16 = reshape(shape = var_6459, x = matrix_bd_131_cast_fp16)[name = tensor("matrix_bd_133_cast_fp16")]; - tensor var_6463_begin_0 = const()[name = tensor("op_6463_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_6463_end_0 = const()[name = tensor("op_6463_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_6463_end_mask_0 = const()[name = tensor("op_6463_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_6463_cast_fp16 = slice_by_index(begin = var_6463_begin_0, end = var_6463_end_0, end_mask = var_6463_end_mask_0, x = matrix_bd_133_cast_fp16)[name = tensor("op_6463_cast_fp16")]; - tensor var_6464 = const()[name = tensor("op_6464"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_cast_fp16 = reshape(shape = var_6464, x = var_6463_cast_fp16)[name = tensor("matrix_bd_cast_fp16")]; - tensor var_6469_begin_0 = const()[name = tensor("op_6469_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor var_6469_end_0 = const()[name = tensor("op_6469_end_0"), val = tensor([1, 8, 188, 188])]; - tensor var_6469_end_mask_0 = const()[name = tensor("op_6469_end_mask_0"), val = tensor([true, true, true, false])]; - tensor var_6469_cast_fp16 = slice_by_index(begin = var_6469_begin_0, end = var_6469_end_0, end_mask = var_6469_end_mask_0, x = matrix_bd_cast_fp16)[name = tensor("op_6469_cast_fp16")]; - tensor var_6470_to_fp16 = const()[name = tensor("op_6470_to_fp16"), val = tensor(0x1p-3)]; - tensor qk_mask_cast_fp16 = mul(x = var_6469_cast_fp16, y = var_6470_to_fp16)[name = tensor("qk_mask_cast_fp16")]; - tensor var_6474 = const()[name = tensor("op_6474"), val = tensor([1, 8, 64, 188])]; - tensor mh_q_cast_fp16 = reshape(shape = var_6474, x = query_cast_fp16)[name = tensor("mh_q_cast_fp16")]; - tensor var_6476_to_fp16 = const()[name = tensor("op_6476_to_fp16"), val = tensor(0x1p-3)]; - tensor var_6477_cast_fp16 = mul(x = mh_q_cast_fp16, y = var_6476_to_fp16)[name = tensor("op_6477_cast_fp16")]; - tensor var_6480 = const()[name = tensor("op_6480"), val = tensor([1, 8, 64, 188])]; - tensor var_6481_cast_fp16 = reshape(shape = var_6480, x = key_cast_fp16)[name = tensor("op_6481_cast_fp16")]; - tensor mh_w_65_transpose_x_0 = const()[name = tensor("mh_w_65_transpose_x_0"), val = tensor(true)]; - tensor mh_w_65_transpose_y_0 = const()[name = tensor("mh_w_65_transpose_y_0"), val = tensor(false)]; - tensor mh_w_65_cast_fp16 = matmul(transpose_x = mh_w_65_transpose_x_0, transpose_y = mh_w_65_transpose_y_0, x = var_6477_cast_fp16, y = var_6481_cast_fp16)[name = tensor("mh_w_65_cast_fp16")]; - tensor mh_w_cast_fp16 = add(x = mh_w_65_cast_fp16, y = qk_mask_cast_fp16)[name = tensor("mh_w_cast_fp16")]; - tensor var_6485_cast_fp16 = softmax(axis = var_6272, x = mh_w_cast_fp16)[name = tensor("op_6485_cast_fp16")]; - tensor var_6486 = const()[name = tensor("op_6486"), val = tensor([1, 8, 64, 188])]; - tensor var_6487_cast_fp16 = reshape(shape = var_6486, x = value_cast_fp16)[name = tensor("op_6487_cast_fp16")]; - tensor attn_transpose_x_0 = const()[name = tensor("attn_transpose_x_0"), val = tensor(false)]; - tensor attn_transpose_y_0 = const()[name = tensor("attn_transpose_y_0"), val = tensor(true)]; - tensor attn_cast_fp16 = matmul(transpose_x = attn_transpose_x_0, transpose_y = attn_transpose_y_0, x = var_6487_cast_fp16, y = var_6485_cast_fp16)[name = tensor("attn_cast_fp16")]; - tensor var_6490 = const()[name = tensor("op_6490"), val = tensor([1, 512, 1, 188])]; - tensor input_439_cast_fp16 = reshape(shape = var_6490, x = attn_cast_fp16)[name = tensor("input_439_cast_fp16")]; - tensor var_6500_pad_type_0 = const()[name = tensor("op_6500_pad_type_0"), val = tensor("valid")]; - tensor var_6500_strides_0 = const()[name = tensor("op_6500_strides_0"), val = tensor([1, 1])]; - tensor var_6500_pad_0 = const()[name = tensor("op_6500_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_6500_dilations_0 = const()[name = tensor("op_6500_dilations_0"), val = tensor([1, 1])]; - tensor var_6500_groups_0 = const()[name = tensor("op_6500_groups_0"), val = tensor(1)]; - tensor layers_16_self_attn_o_proj_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(97126848))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(97323520))), name = tensor("layers_16_self_attn_o_proj_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_16_self_attn_o_proj_inlier_module_bias_to_fp16 = const()[name = tensor("layers_16_self_attn_o_proj_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(97323712)))]; - tensor var_6500_cast_fp16 = conv(bias = layers_16_self_attn_o_proj_inlier_module_bias_to_fp16, dilations = var_6500_dilations_0, groups = var_6500_groups_0, pad = var_6500_pad_0, pad_type = var_6500_pad_type_0, strides = var_6500_strides_0, weight = layers_16_self_attn_o_proj_inlier_module_weight_to_fp16_palettized, x = input_439_cast_fp16)[name = tensor("op_6500_cast_fp16")]; - tensor var_6506_pad_type_0 = const()[name = tensor("op_6506_pad_type_0"), val = tensor("valid")]; - tensor var_6506_strides_0 = const()[name = tensor("op_6506_strides_0"), val = tensor([1, 1])]; - tensor var_6506_pad_0 = const()[name = tensor("op_6506_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_6506_dilations_0 = const()[name = tensor("op_6506_dilations_0"), val = tensor([1, 1])]; - tensor var_6506_groups_0 = const()[name = tensor("op_6506_groups_0"), val = tensor(1)]; - tensor layers_16_self_attn_o_proj_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(97333248))), name = tensor("layers_16_self_attn_o_proj_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(97324800))), shape = tensor([512, 512, 1, 1])]; - tensor var_6506_cast_fp16 = conv(dilations = var_6506_dilations_0, groups = var_6506_groups_0, pad = var_6506_pad_0, pad_type = var_6506_pad_type_0, strides = var_6506_strides_0, weight = layers_16_self_attn_o_proj_outlier_module_weight_to_fp16_sparsified, x = input_439_cast_fp16)[name = tensor("op_6506_cast_fp16")]; - tensor obj_cast_fp16 = add(x = var_6500_cast_fp16, y = var_6506_cast_fp16)[name = tensor("obj_cast_fp16")]; - tensor inputs_165_cast_fp16 = add(x = inputs_163_cast_fp16, y = obj_cast_fp16)[name = tensor("inputs_165_cast_fp16")]; - tensor out_165_axes_0 = const()[name = tensor("out_165_axes_0"), val = tensor([1])]; - tensor var_6517_to_fp16 = const()[name = tensor("op_6517_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_165_cast_fp16 = layer_norm(axes = out_165_axes_0, epsilon = var_6517_to_fp16, x = inputs_165_cast_fp16)[name = tensor("out_165_cast_fp16")]; - tensor input_441_gamma_0_to_fp16 = const()[name = tensor("input_441_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(97366080)))]; - tensor input_441_beta_0_to_fp16 = const()[name = tensor("input_441_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(97367168)))]; - tensor input_441_epsilon_0_to_fp16 = const()[name = tensor("input_441_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_441_cast_fp16 = batch_norm(beta = input_441_beta_0_to_fp16, epsilon = input_441_epsilon_0_to_fp16, gamma = input_441_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_165_cast_fp16)[name = tensor("input_441_cast_fp16")]; - tensor var_6539_pad_type_0 = const()[name = tensor("op_6539_pad_type_0"), val = tensor("valid")]; - tensor var_6539_strides_0 = const()[name = tensor("op_6539_strides_0"), val = tensor([1, 1])]; - tensor var_6539_pad_0 = const()[name = tensor("op_6539_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_6539_dilations_0 = const()[name = tensor("op_6539_dilations_0"), val = tensor([1, 1])]; - tensor var_6539_groups_0 = const()[name = tensor("op_6539_groups_0"), val = tensor(1)]; - tensor layers_16_conv_pointwise_conv1_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(97368256))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(97761536))), name = tensor("layers_16_conv_pointwise_conv1_inlier_module_weight_to_fp16_palettized"), shape = tensor([1024, 512, 1, 1])]; - tensor layers_16_conv_pointwise_conv1_inlier_module_bias_to_fp16 = const()[name = tensor("layers_16_conv_pointwise_conv1_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(97761728)))]; - tensor var_6539_cast_fp16 = conv(bias = layers_16_conv_pointwise_conv1_inlier_module_bias_to_fp16, dilations = var_6539_dilations_0, groups = var_6539_groups_0, pad = var_6539_pad_0, pad_type = var_6539_pad_type_0, strides = var_6539_strides_0, weight = layers_16_conv_pointwise_conv1_inlier_module_weight_to_fp16_palettized, x = input_441_cast_fp16)[name = tensor("op_6539_cast_fp16")]; - tensor var_6545_pad_type_0 = const()[name = tensor("op_6545_pad_type_0"), val = tensor("valid")]; - tensor var_6545_strides_0 = const()[name = tensor("op_6545_strides_0"), val = tensor([1, 1])]; - tensor var_6545_pad_0 = const()[name = tensor("op_6545_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_6545_dilations_0 = const()[name = tensor("op_6545_dilations_0"), val = tensor([1, 1])]; - tensor var_6545_groups_0 = const()[name = tensor("op_6545_groups_0"), val = tensor(1)]; - tensor layers_16_conv_pointwise_conv1_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(97782720))), name = tensor("layers_16_conv_pointwise_conv1_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(97763840))), shape = tensor([1024, 512, 1, 1])]; - tensor var_6545_cast_fp16 = conv(dilations = var_6545_dilations_0, groups = var_6545_groups_0, pad = var_6545_pad_0, pad_type = var_6545_pad_type_0, strides = var_6545_strides_0, weight = layers_16_conv_pointwise_conv1_outlier_module_weight_to_fp16_sparsified, x = input_441_cast_fp16)[name = tensor("op_6545_cast_fp16")]; - tensor input_443_cast_fp16 = add(x = var_6539_cast_fp16, y = var_6545_cast_fp16)[name = tensor("input_443_cast_fp16")]; - tensor input_445_split_num_splits_0 = const()[name = tensor("input_445_split_num_splits_0"), val = tensor(2)]; - tensor input_445_split_axis_0 = const()[name = tensor("input_445_split_axis_0"), val = tensor(1)]; - tensor input_445_split_cast_fp16_0, tensor input_445_split_cast_fp16_1 = split(axis = input_445_split_axis_0, num_splits = input_445_split_num_splits_0, x = input_443_cast_fp16)[name = tensor("input_445_split_cast_fp16")]; - tensor input_445_split_1_sigmoid_cast_fp16 = sigmoid(x = input_445_split_cast_fp16_1)[name = tensor("input_445_split_1_sigmoid_cast_fp16")]; - tensor input_445_cast_fp16 = mul(x = input_445_split_cast_fp16_0, y = input_445_split_1_sigmoid_cast_fp16)[name = tensor("input_445_cast_fp16")]; - tensor input_447_pad_type_0 = const()[name = tensor("input_447_pad_type_0"), val = tensor("custom")]; - tensor input_447_pad_0 = const()[name = tensor("input_447_pad_0"), val = tensor([0, 0, 4, 4])]; - tensor input_447_groups_0 = const()[name = tensor("input_447_groups_0"), val = tensor(512)]; - tensor input_447_strides_0 = const()[name = tensor("input_447_strides_0"), val = tensor([1, 1])]; - tensor input_447_dilations_0 = const()[name = tensor("input_447_dilations_0"), val = tensor([1, 1])]; - tensor const_223_to_fp16 = const()[name = tensor("const_223_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(97848320)))]; - tensor const_224_to_fp16 = const()[name = tensor("const_224_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(97857600)))]; - tensor input_449_cast_fp16 = conv(bias = const_224_to_fp16, dilations = input_447_dilations_0, groups = input_447_groups_0, pad = input_447_pad_0, pad_type = input_447_pad_type_0, strides = input_447_strides_0, weight = const_223_to_fp16, x = input_445_cast_fp16)[name = tensor("input_449_cast_fp16")]; - tensor input_451_cast_fp16 = silu(x = input_449_cast_fp16)[name = tensor("input_451_cast_fp16")]; - tensor var_6569_pad_type_0 = const()[name = tensor("op_6569_pad_type_0"), val = tensor("valid")]; - tensor var_6569_strides_0 = const()[name = tensor("op_6569_strides_0"), val = tensor([1, 1])]; - tensor var_6569_pad_0 = const()[name = tensor("op_6569_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_6569_dilations_0 = const()[name = tensor("op_6569_dilations_0"), val = tensor([1, 1])]; - tensor var_6569_groups_0 = const()[name = tensor("op_6569_groups_0"), val = tensor(1)]; - tensor layers_16_conv_pointwise_conv2_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(97858688))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98055360))), name = tensor("layers_16_conv_pointwise_conv2_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 512, 1, 1])]; - tensor layers_16_conv_pointwise_conv2_inlier_module_bias_to_fp16 = const()[name = tensor("layers_16_conv_pointwise_conv2_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98055552)))]; - tensor var_6569_cast_fp16 = conv(bias = layers_16_conv_pointwise_conv2_inlier_module_bias_to_fp16, dilations = var_6569_dilations_0, groups = var_6569_groups_0, pad = var_6569_pad_0, pad_type = var_6569_pad_type_0, strides = var_6569_strides_0, weight = layers_16_conv_pointwise_conv2_inlier_module_weight_to_fp16_palettized, x = input_451_cast_fp16)[name = tensor("op_6569_cast_fp16")]; - tensor var_6575_pad_type_0 = const()[name = tensor("op_6575_pad_type_0"), val = tensor("valid")]; - tensor var_6575_strides_0 = const()[name = tensor("op_6575_strides_0"), val = tensor([1, 1])]; - tensor var_6575_pad_0 = const()[name = tensor("op_6575_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_6575_dilations_0 = const()[name = tensor("op_6575_dilations_0"), val = tensor([1, 1])]; - tensor var_6575_groups_0 = const()[name = tensor("op_6575_groups_0"), val = tensor(1)]; - tensor layers_16_conv_pointwise_conv2_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98064256))), name = tensor("layers_16_conv_pointwise_conv2_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98056640))), shape = tensor([512, 512, 1, 1])]; - tensor var_6575_cast_fp16 = conv(dilations = var_6575_dilations_0, groups = var_6575_groups_0, pad = var_6575_pad_0, pad_type = var_6575_pad_type_0, strides = var_6575_strides_0, weight = layers_16_conv_pointwise_conv2_outlier_module_weight_to_fp16_sparsified, x = input_451_cast_fp16)[name = tensor("op_6575_cast_fp16")]; - tensor x_101_cast_fp16 = add(x = var_6569_cast_fp16, y = var_6575_cast_fp16)[name = tensor("x_101_cast_fp16")]; - tensor inputs_167_cast_fp16 = add(x = inputs_165_cast_fp16, y = x_101_cast_fp16)[name = tensor("inputs_167_cast_fp16")]; - tensor out_167_axes_0 = const()[name = tensor("out_167_axes_0"), val = tensor([1])]; - tensor var_6586_to_fp16 = const()[name = tensor("op_6586_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_167_cast_fp16 = layer_norm(axes = out_167_axes_0, epsilon = var_6586_to_fp16, x = inputs_167_cast_fp16)[name = tensor("out_167_cast_fp16")]; - tensor input_453_gamma_0_to_fp16 = const()[name = tensor("input_453_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98097088)))]; - tensor input_453_beta_0_to_fp16 = const()[name = tensor("input_453_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98098176)))]; - tensor input_453_epsilon_0_to_fp16 = const()[name = tensor("input_453_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_453_cast_fp16 = batch_norm(beta = input_453_beta_0_to_fp16, epsilon = input_453_epsilon_0_to_fp16, gamma = input_453_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_167_cast_fp16)[name = tensor("input_453_cast_fp16")]; - tensor var_6606_pad_type_0 = const()[name = tensor("op_6606_pad_type_0"), val = tensor("valid")]; - tensor var_6606_strides_0 = const()[name = tensor("op_6606_strides_0"), val = tensor([1, 1])]; - tensor var_6606_pad_0 = const()[name = tensor("op_6606_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_6606_dilations_0 = const()[name = tensor("op_6606_dilations_0"), val = tensor([1, 1])]; - tensor var_6606_groups_0 = const()[name = tensor("op_6606_groups_0"), val = tensor(1)]; - tensor layers_16_feed_forward2_fc1_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98099264))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98885760))), name = tensor("layers_16_feed_forward2_fc1_inlier_module_weight_to_fp16_palettized"), shape = tensor([2048, 512, 1, 1])]; - tensor layers_16_feed_forward2_fc1_inlier_module_bias_to_fp16 = const()[name = tensor("layers_16_feed_forward2_fc1_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98885952)))]; - tensor var_6606_cast_fp16 = conv(bias = layers_16_feed_forward2_fc1_inlier_module_bias_to_fp16, dilations = var_6606_dilations_0, groups = var_6606_groups_0, pad = var_6606_pad_0, pad_type = var_6606_pad_type_0, strides = var_6606_strides_0, weight = layers_16_feed_forward2_fc1_inlier_module_weight_to_fp16_palettized, x = input_453_cast_fp16)[name = tensor("op_6606_cast_fp16")]; - tensor var_6612_pad_type_0 = const()[name = tensor("op_6612_pad_type_0"), val = tensor("valid")]; - tensor var_6612_strides_0 = const()[name = tensor("op_6612_strides_0"), val = tensor([1, 1])]; - tensor var_6612_pad_0 = const()[name = tensor("op_6612_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_6612_dilations_0 = const()[name = tensor("op_6612_dilations_0"), val = tensor([1, 1])]; - tensor var_6612_groups_0 = const()[name = tensor("op_6612_groups_0"), val = tensor(1)]; - tensor layers_16_feed_forward2_fc1_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98919552))), name = tensor("layers_16_feed_forward2_fc1_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98890112))), shape = tensor([2048, 512, 1, 1])]; - tensor var_6612_cast_fp16 = conv(dilations = var_6612_dilations_0, groups = var_6612_groups_0, pad = var_6612_pad_0, pad_type = var_6612_pad_type_0, strides = var_6612_strides_0, weight = layers_16_feed_forward2_fc1_outlier_module_weight_to_fp16_sparsified, x = input_453_cast_fp16)[name = tensor("op_6612_cast_fp16")]; - tensor input_455_cast_fp16 = add(x = var_6606_cast_fp16, y = var_6612_cast_fp16)[name = tensor("input_455_cast_fp16")]; - tensor input_457_cast_fp16 = silu(x = input_455_cast_fp16)[name = tensor("input_457_cast_fp16")]; - tensor var_6623_pad_type_0 = const()[name = tensor("op_6623_pad_type_0"), val = tensor("valid")]; - tensor var_6623_strides_0 = const()[name = tensor("op_6623_strides_0"), val = tensor([1, 1])]; - tensor var_6623_pad_0 = const()[name = tensor("op_6623_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_6623_dilations_0 = const()[name = tensor("op_6623_dilations_0"), val = tensor([1, 1])]; - tensor var_6623_groups_0 = const()[name = tensor("op_6623_groups_0"), val = tensor(1)]; - tensor layers_16_feed_forward2_fc2_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(99050688))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(99837184))), name = tensor("layers_16_feed_forward2_fc2_inlier_module_weight_to_fp16_palettized"), shape = tensor([512, 2048, 1, 1])]; - tensor layers_16_feed_forward2_fc2_inlier_module_bias_to_fp16 = const()[name = tensor("layers_16_feed_forward2_fc2_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(99837376)))]; - tensor var_6623_cast_fp16 = conv(bias = layers_16_feed_forward2_fc2_inlier_module_bias_to_fp16, dilations = var_6623_dilations_0, groups = var_6623_groups_0, pad = var_6623_pad_0, pad_type = var_6623_pad_type_0, strides = var_6623_strides_0, weight = layers_16_feed_forward2_fc2_inlier_module_weight_to_fp16_palettized, x = input_457_cast_fp16)[name = tensor("op_6623_cast_fp16")]; - tensor var_6629_pad_type_0 = const()[name = tensor("op_6629_pad_type_0"), val = tensor("valid")]; - tensor var_6629_strides_0 = const()[name = tensor("op_6629_strides_0"), val = tensor([1, 1])]; - tensor var_6629_pad_0 = const()[name = tensor("op_6629_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_6629_dilations_0 = const()[name = tensor("op_6629_dilations_0"), val = tensor([1, 1])]; - tensor var_6629_groups_0 = const()[name = tensor("op_6629_groups_0"), val = tensor(1)]; - tensor layers_16_feed_forward2_fc2_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(99870784))), name = tensor("layers_16_feed_forward2_fc2_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(99838464))), shape = tensor([512, 2048, 1, 1])]; - tensor var_6629_cast_fp16 = conv(dilations = var_6629_dilations_0, groups = var_6629_groups_0, pad = var_6629_pad_0, pad_type = var_6629_pad_type_0, strides = var_6629_strides_0, weight = layers_16_feed_forward2_fc2_outlier_module_weight_to_fp16_sparsified, x = input_457_cast_fp16)[name = tensor("op_6629_cast_fp16")]; - tensor x_cast_fp16 = add(x = var_6623_cast_fp16, y = var_6629_cast_fp16)[name = tensor("x_cast_fp16")]; - tensor var_6631_to_fp16 = const()[name = tensor("op_6631_to_fp16"), val = tensor(0x1p-1)]; - tensor var_6632_cast_fp16 = mul(x = x_cast_fp16, y = var_6631_to_fp16)[name = tensor("op_6632_cast_fp16")]; - tensor inputs_cast_fp16 = add(x = inputs_167_cast_fp16, y = var_6632_cast_fp16)[name = tensor("inputs_cast_fp16")]; - tensor out_169_axes_0 = const()[name = tensor("out_169_axes_0"), val = tensor([1])]; - tensor var_6642_to_fp16 = const()[name = tensor("op_6642_to_fp16"), val = tensor(0x1.5p-17)]; - tensor out_169_cast_fp16 = layer_norm(axes = out_169_axes_0, epsilon = var_6642_to_fp16, x = inputs_cast_fp16)[name = tensor("out_169_cast_fp16")]; - tensor encoder_output_embeds_type_fp32_gamma_0_to_fp16 = const()[name = tensor("encoder_output_embeds_type_fp32_gamma_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(100001920)))]; - tensor encoder_output_embeds_type_fp32_beta_0_to_fp16 = const()[name = tensor("encoder_output_embeds_type_fp32_beta_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(100003008)))]; - tensor encoder_output_embeds_type_fp32_epsilon_0_to_fp16 = const()[name = tensor("encoder_output_embeds_type_fp32_epsilon_0_to_fp16"), val = tensor(0x1.5p-17)]; - tensor encoder_output_embeds = batch_norm(beta = encoder_output_embeds_type_fp32_beta_0_to_fp16, epsilon = encoder_output_embeds_type_fp32_epsilon_0_to_fp16, gamma = encoder_output_embeds_type_fp32_gamma_0_to_fp16, mean = input_17_mean_0_to_fp16, variance = input_17_variance_0_to_fp16, x = out_169_cast_fp16)[name = tensor("encoder_output_embeds_type_fp32_cast_fp16")]; - tensor var_6665_pad_type_0 = const()[name = tensor("op_6665_pad_type_0"), val = tensor("valid")]; - tensor var_6665_strides_0 = const()[name = tensor("op_6665_strides_0"), val = tensor([1, 1])]; - tensor var_6665_pad_0 = const()[name = tensor("op_6665_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_6665_dilations_0 = const()[name = tensor("op_6665_dilations_0"), val = tensor([1, 1])]; - tensor var_6665_groups_0 = const()[name = tensor("op_6665_groups_0"), val = tensor(1)]; - tensor out_projection_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(100004096))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(100249920))), name = tensor("out_projection_inlier_module_weight_to_fp16_palettized"), shape = tensor([640, 512, 1, 1])]; - tensor out_projection_inlier_module_bias_to_fp16 = const()[name = tensor("out_projection_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(100250112)))]; - tensor var_6665_cast_fp16 = conv(bias = out_projection_inlier_module_bias_to_fp16, dilations = var_6665_dilations_0, groups = var_6665_groups_0, pad = var_6665_pad_0, pad_type = var_6665_pad_type_0, strides = var_6665_strides_0, weight = out_projection_inlier_module_weight_to_fp16_palettized, x = encoder_output_embeds)[name = tensor("op_6665_cast_fp16")]; - tensor var_6671_pad_type_0 = const()[name = tensor("op_6671_pad_type_0"), val = tensor("valid")]; - tensor var_6671_strides_0 = const()[name = tensor("op_6671_strides_0"), val = tensor([1, 1])]; - tensor var_6671_pad_0 = const()[name = tensor("op_6671_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_6671_dilations_0 = const()[name = tensor("op_6671_dilations_0"), val = tensor([1, 1])]; - tensor var_6671_groups_0 = const()[name = tensor("op_6671_groups_0"), val = tensor(1)]; - tensor out_projection_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(100261568))), name = tensor("out_projection_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(100251456))), shape = tensor([640, 512, 1, 1])]; - tensor var_6671_cast_fp16 = conv(dilations = var_6671_dilations_0, groups = var_6671_groups_0, pad = var_6671_pad_0, pad_type = var_6671_pad_type_0, strides = var_6671_strides_0, weight = out_projection_outlier_module_weight_to_fp16_sparsified, x = encoder_output_embeds)[name = tensor("op_6671_cast_fp16")]; - tensor joint_projected_encoder_output_embeds = add(x = var_6665_cast_fp16, y = var_6671_cast_fp16)[name = tensor("op_6672_cast_fp16")]; - tensor var_6686_pad_type_0 = const()[name = tensor("op_6686_pad_type_0"), val = tensor("valid")]; - tensor var_6686_strides_0 = const()[name = tensor("op_6686_strides_0"), val = tensor([1, 1])]; - tensor var_6686_pad_0 = const()[name = tensor("op_6686_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_6686_dilations_0 = const()[name = tensor("op_6686_dilations_0"), val = tensor([1, 1])]; - tensor var_6686_groups_0 = const()[name = tensor("op_6686_groups_0"), val = tensor(1)]; - tensor ctc_head_inlier_module_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(100302592))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(100696256))), name = tensor("ctc_head_inlier_module_weight_to_fp16_palettized"), shape = tensor([1025, 512, 1, 1])]; - tensor ctc_head_inlier_module_bias_to_fp16 = const()[name = tensor("ctc_head_inlier_module_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(100696448)))]; - tensor var_6686_cast_fp16 = conv(bias = ctc_head_inlier_module_bias_to_fp16, dilations = var_6686_dilations_0, groups = var_6686_groups_0, pad = var_6686_pad_0, pad_type = var_6686_pad_type_0, strides = var_6686_strides_0, weight = ctc_head_inlier_module_weight_to_fp16_palettized, x = encoder_output_embeds)[name = tensor("op_6686_cast_fp16")]; - tensor var_6692_pad_type_0 = const()[name = tensor("op_6692_pad_type_0"), val = tensor("valid")]; - tensor var_6692_strides_0 = const()[name = tensor("op_6692_strides_0"), val = tensor([1, 1])]; - tensor var_6692_pad_0 = const()[name = tensor("op_6692_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor var_6692_dilations_0 = const()[name = tensor("op_6692_dilations_0"), val = tensor([1, 1])]; - tensor var_6692_groups_0 = const()[name = tensor("op_6692_groups_0"), val = tensor(1)]; - tensor ctc_head_outlier_module_weight_to_fp16_sparsified = constexpr_sparse_to_dense()[mask = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(100712640))), name = tensor("ctc_head_outlier_module_weight_to_fp16_sparsified"), nonzero_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(100698624))), shape = tensor([1025, 512, 1, 1])]; - tensor var_6692_cast_fp16 = conv(dilations = var_6692_dilations_0, groups = var_6692_groups_0, pad = var_6692_pad_0, pad_type = var_6692_pad_type_0, strides = var_6692_strides_0, weight = ctc_head_outlier_module_weight_to_fp16_sparsified, x = encoder_output_embeds)[name = tensor("op_6692_cast_fp16")]; - tensor ctc_head_raw_output = add(x = var_6686_cast_fp16, y = var_6692_cast_fp16)[name = tensor("op_6693_cast_fp16")]; - tensor var_6715 = const()[name = tensor("op_6715"), val = tensor(1)]; - tensor var_6717_softmax_cast_fp16 = softmax(axis = var_6715, x = ctc_head_raw_output)[name = tensor("op_6717_softmax_cast_fp16")]; - tensor var_6717_epsilon_0 = const()[name = tensor("op_6717_epsilon_0"), val = tensor(0x1p-149)]; - tensor ctc_head_output = log(epsilon = var_6717_epsilon_0, x = var_6717_softmax_cast_fp16)[name = tensor("op_6717_cast_fp16")]; - } -> (encoder_output_embeds, joint_projected_encoder_output_embeds, ctc_head_raw_output, ctc_head_output); -} \ No newline at end of file diff --git a/parakeet-ctc-110m-coreml/AudioEncoder.mlmodelc/weights/weight.bin b/parakeet-ctc-110m-coreml/AudioEncoder.mlmodelc/weights/weight.bin deleted file mode 100644 index 5a2bb38f8adc0042481f17648abc50b67e583d27..0000000000000000000000000000000000000000 --- a/parakeet-ctc-110m-coreml/AudioEncoder.mlmodelc/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:af0734b4a5d7465ad9e8bb170f0c53c5e6b91ebb75a9bdf88d3f59ae4ad6aebd -size 100778304 diff --git a/parakeet-ctc-110m-coreml/CTCHead.mlmodelc/analytics/coremldata.bin b/parakeet-ctc-110m-coreml/CTCHead.mlmodelc/analytics/coremldata.bin deleted file mode 100644 index 4093e2df453ec508a533d293a574198b2372845b..0000000000000000000000000000000000000000 --- a/parakeet-ctc-110m-coreml/CTCHead.mlmodelc/analytics/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:fc681823d92eca3dbece3a30c975afa7251eedae0e718b07ffbf1a8b4313b87e -size 243 diff --git a/parakeet-ctc-110m-coreml/CTCHead.mlmodelc/coremldata.bin b/parakeet-ctc-110m-coreml/CTCHead.mlmodelc/coremldata.bin deleted file mode 100644 index db4ae8de920867977480465fffece41bc606c0a2..0000000000000000000000000000000000000000 --- a/parakeet-ctc-110m-coreml/CTCHead.mlmodelc/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:2ebec8fc38c063de4b2159e21b1f981309fa5947c24d7e4883aca20f7c15fbb9 -size 377 diff --git a/parakeet-ctc-110m-coreml/CTCHead.mlmodelc/metadata.json b/parakeet-ctc-110m-coreml/CTCHead.mlmodelc/metadata.json deleted file mode 100644 index 4abe30bf0b35b72d2c9c013b61cc78b4f26a85e1..0000000000000000000000000000000000000000 --- a/parakeet-ctc-110m-coreml/CTCHead.mlmodelc/metadata.json +++ /dev/null @@ -1,66 +0,0 @@ -[ - { - "metadataOutputVersion" : "3.0", - "shortDescription" : "Parakeet 110M CTC decoder head", - "outputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 188 × 1025)", - "shortDescription" : "", - "shape" : "[1, 188, 1025]", - "name" : "ctc_logits", - "type" : "MultiArray" - } - ], - "storagePrecision" : "Float16", - "modelParameters" : [ - - ], - "author" : "Fluid Inference", - "specificationVersion" : 8, - "mlProgramOperationTypeHistogram" : { - "Ios17.cast" : 2, - "Ios17.conv" : 1, - "Ios17.transpose" : 1, - "Ios16.softmax" : 1, - "Ios17.log" : 1 - }, - "computePrecision" : "Mixed (Float16, Float32, Int32)", - "isUpdatable" : "0", - "stateSchema" : [ - - ], - "availability" : { - "macOS" : "14.0", - "tvOS" : "17.0", - "visionOS" : "1.0", - "watchOS" : "10.0", - "iOS" : "17.0", - "macCatalyst" : "17.0" - }, - "modelType" : { - "name" : "MLModelType_mlProgram" - }, - "inputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 512 × 188)", - "shortDescription" : "", - "shape" : "[1, 512, 188]", - "name" : "encoder_output", - "type" : "MultiArray" - } - ], - "userDefinedMetadata" : { - "com.github.apple.coremltools.source_dialect" : "TorchScript", - "com.github.apple.coremltools.source" : "torch==2.9.0", - "com.github.apple.coremltools.version" : "8.3.0" - }, - "generatedClassName" : "parakeet_ctc_head", - "method" : "predict" - } -] \ No newline at end of file diff --git a/parakeet-ctc-110m-coreml/CTCHead.mlmodelc/model.mil b/parakeet-ctc-110m-coreml/CTCHead.mlmodelc/model.mil deleted file mode 100644 index 67b3b5f87ce33caed71f9233828f90775bb8ac9d..0000000000000000000000000000000000000000 --- a/parakeet-ctc-110m-coreml/CTCHead.mlmodelc/model.mil +++ /dev/null @@ -1,24 +0,0 @@ -program(1.0) -[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3510.2.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.9.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "8.3.0"}})] -{ - func main(tensor encoder_output) { - tensor var_4 = const()[name = tensor("op_4"), val = tensor(-1)]; - tensor var_18_pad_type_0 = const()[name = tensor("op_18_pad_type_0"), val = tensor("valid")]; - tensor var_18_strides_0 = const()[name = tensor("op_18_strides_0"), val = tensor([1])]; - tensor var_18_pad_0 = const()[name = tensor("op_18_pad_0"), val = tensor([0, 0])]; - tensor var_18_dilations_0 = const()[name = tensor("op_18_dilations_0"), val = tensor([1])]; - tensor var_18_groups_0 = const()[name = tensor("op_18_groups_0"), val = tensor(1)]; - tensor encoder_output_to_fp16_dtype_0 = const()[name = tensor("encoder_output_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor module_decoder_layers_0_weight_to_fp16 = const()[name = tensor("module_decoder_layers_0_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; - tensor module_decoder_layers_0_bias_to_fp16 = const()[name = tensor("module_decoder_layers_0_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1049728)))]; - tensor encoder_output_to_fp16 = cast(dtype = encoder_output_to_fp16_dtype_0, x = encoder_output)[name = tensor("cast_1")]; - tensor var_18_cast_fp16 = conv(bias = module_decoder_layers_0_bias_to_fp16, dilations = var_18_dilations_0, groups = var_18_groups_0, pad = var_18_pad_0, pad_type = var_18_pad_type_0, strides = var_18_strides_0, weight = module_decoder_layers_0_weight_to_fp16, x = encoder_output_to_fp16)[name = tensor("op_18_cast_fp16")]; - tensor input_perm_0 = const()[name = tensor("input_perm_0"), val = tensor([0, 2, 1])]; - tensor input_cast_fp16 = transpose(perm = input_perm_0, x = var_18_cast_fp16)[name = tensor("transpose_0")]; - tensor out_objects_softmax_cast_fp16 = softmax(axis = var_4, x = input_cast_fp16)[name = tensor("out_objects_softmax_cast_fp16")]; - tensor out_objects_epsilon_0 = const()[name = tensor("out_objects_epsilon_0"), val = tensor(0x1p-149)]; - tensor out_objects_cast_fp16 = log(epsilon = out_objects_epsilon_0, x = out_objects_softmax_cast_fp16)[name = tensor("out_objects_cast_fp16")]; - tensor out_objects_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("out_objects_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; - tensor ctc_logits = cast(dtype = out_objects_cast_fp16_to_fp32_dtype_0, x = out_objects_cast_fp16)[name = tensor("cast_0")]; - } -> (ctc_logits); -} \ No newline at end of file diff --git a/parakeet-ctc-110m-coreml/CTCHead.mlmodelc/weights/weight.bin b/parakeet-ctc-110m-coreml/CTCHead.mlmodelc/weights/weight.bin deleted file mode 100644 index 23cedd570125191064c6111997f08c48898245f7..0000000000000000000000000000000000000000 --- a/parakeet-ctc-110m-coreml/CTCHead.mlmodelc/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:fb9bead064427ffcb7529c0e3f378e421b4dde8e6d81447b6d1ca3352ca850e1 -size 1051842 diff --git a/parakeet-ctc-110m-coreml/Decoder.mlmodelc/analytics/coremldata.bin b/parakeet-ctc-110m-coreml/Decoder.mlmodelc/analytics/coremldata.bin deleted file mode 100644 index dd015548a03cc453e6477a24374e8fee955e75ef..0000000000000000000000000000000000000000 --- a/parakeet-ctc-110m-coreml/Decoder.mlmodelc/analytics/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:990455f6431342750254f66edf27bfb41be62a7ba17a18e1dd6afd4f5f56e9eb -size 243 diff --git a/parakeet-ctc-110m-coreml/Decoder.mlmodelc/coremldata.bin b/parakeet-ctc-110m-coreml/Decoder.mlmodelc/coremldata.bin deleted file mode 100644 index cbea0affef0c3bb63be750e134dc0c9ed70aadb1..0000000000000000000000000000000000000000 --- a/parakeet-ctc-110m-coreml/Decoder.mlmodelc/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:29009727821ad8551ab5fe9271e93c597d92a9714f64b94aa533a9ceb6e22b93 -size 498 diff --git a/parakeet-ctc-110m-coreml/Decoder.mlmodelc/metadata.json b/parakeet-ctc-110m-coreml/Decoder.mlmodelc/metadata.json deleted file mode 100644 index 786f4bb2823b1aa7ebb55fe335c9385ee0065b7c..0000000000000000000000000000000000000000 --- a/parakeet-ctc-110m-coreml/Decoder.mlmodelc/metadata.json +++ /dev/null @@ -1,118 +0,0 @@ -[ - { - "metadataOutputVersion" : "3.0", - "shortDescription" : "Parakeet 110M decoder (RNNT prediction network)", - "outputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 640 × 1)", - "shortDescription" : "", - "shape" : "[1, 640, 1]", - "name" : "decoder", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 1 × 640)", - "shortDescription" : "", - "shape" : "[1, 1, 640]", - "name" : "h_out", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 1 × 640)", - "shortDescription" : "", - "shape" : "[1, 1, 640]", - "name" : "c_out", - "type" : "MultiArray" - } - ], - "storagePrecision" : "Float16", - "modelParameters" : [ - - ], - "author" : "Fluid Inference", - "specificationVersion" : 8, - "mlProgramOperationTypeHistogram" : { - "Ios17.squeeze" : 2, - "Ios17.gather" : 1, - "Ios17.cast" : 6, - "Ios17.lstm" : 1, - "Ios17.transpose" : 2, - "Identity" : 1, - "Ios17.expandDims" : 2 - }, - "computePrecision" : "Mixed (Float16, Float32, Int16, Int32)", - "isUpdatable" : "0", - "stateSchema" : [ - - ], - "availability" : { - "macOS" : "14.0", - "tvOS" : "17.0", - "visionOS" : "1.0", - "watchOS" : "10.0", - "iOS" : "17.0", - "macCatalyst" : "17.0" - }, - "modelType" : { - "name" : "MLModelType_mlProgram" - }, - "inputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1 × 1)", - "shortDescription" : "", - "shape" : "[1, 1]", - "name" : "targets", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1)", - "shortDescription" : "", - "shape" : "[1]", - "name" : "target_length", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 1 × 640)", - "shortDescription" : "", - "shape" : "[1, 1, 640]", - "name" : "h_in", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 1 × 640)", - "shortDescription" : "", - "shape" : "[1, 1, 640]", - "name" : "c_in", - "type" : "MultiArray" - } - ], - "userDefinedMetadata" : { - "com.github.apple.coremltools.version" : "8.3.0", - "com.github.apple.coremltools.source_dialect" : "TorchScript", - "com.github.apple.coremltools.source" : "torch==2.9.0" - }, - "generatedClassName" : "parakeet_decoder", - "method" : "predict" - } -] \ No newline at end of file diff --git a/parakeet-ctc-110m-coreml/Decoder.mlmodelc/model.mil b/parakeet-ctc-110m-coreml/Decoder.mlmodelc/model.mil deleted file mode 100644 index f69c7247525bc9d95f9341acc4f11eca709c7d00..0000000000000000000000000000000000000000 --- a/parakeet-ctc-110m-coreml/Decoder.mlmodelc/model.mil +++ /dev/null @@ -1,45 +0,0 @@ -program(1.0) -[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3510.2.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.9.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "8.3.0"}})] -{ - func main(tensor c_in, tensor h_in, tensor target_length, tensor targets) { - tensor y_axis_0 = const()[name = tensor("y_axis_0"), val = tensor(0)]; - tensor y_batch_dims_0 = const()[name = tensor("y_batch_dims_0"), val = tensor(0)]; - tensor y_validate_indices_0 = const()[name = tensor("y_validate_indices_0"), val = tensor(false)]; - tensor module_prediction_embed_weight_to_fp16 = const()[name = tensor("module_prediction_embed_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; - tensor targets_to_int16_dtype_0 = const()[name = tensor("targets_to_int16_dtype_0"), val = tensor("int16")]; - tensor targets_to_int16 = cast(dtype = targets_to_int16_dtype_0, x = targets)[name = tensor("cast_8")]; - tensor y_cast_fp16_cast_uint16 = gather(axis = y_axis_0, batch_dims = y_batch_dims_0, indices = targets_to_int16, validate_indices = y_validate_indices_0, x = module_prediction_embed_weight_to_fp16)[name = tensor("y_cast_fp16_cast_uint16")]; - tensor input_3_perm_0 = const()[name = tensor("input_3_perm_0"), val = tensor([1, 0, 2])]; - tensor input_lstm_h0_squeeze_axes_0 = const()[name = tensor("input_lstm_h0_squeeze_axes_0"), val = tensor([0])]; - tensor h_in_to_fp16_dtype_0 = const()[name = tensor("h_in_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor h_in_to_fp16 = cast(dtype = h_in_to_fp16_dtype_0, x = h_in)[name = tensor("cast_7")]; - tensor input_lstm_h0_squeeze_cast_fp16 = squeeze(axes = input_lstm_h0_squeeze_axes_0, x = h_in_to_fp16)[name = tensor("input_lstm_h0_squeeze_cast_fp16")]; - tensor input_lstm_c0_squeeze_axes_0 = const()[name = tensor("input_lstm_c0_squeeze_axes_0"), val = tensor([0])]; - tensor c_in_to_fp16_dtype_0 = const()[name = tensor("c_in_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor c_in_to_fp16 = cast(dtype = c_in_to_fp16_dtype_0, x = c_in)[name = tensor("cast_6")]; - tensor input_lstm_c0_squeeze_cast_fp16 = squeeze(axes = input_lstm_c0_squeeze_axes_0, x = c_in_to_fp16)[name = tensor("input_lstm_c0_squeeze_cast_fp16")]; - tensor input_direction_0 = const()[name = tensor("input_direction_0"), val = tensor("forward")]; - tensor input_output_sequence_0 = const()[name = tensor("input_output_sequence_0"), val = tensor(true)]; - tensor input_recurrent_activation_0 = const()[name = tensor("input_recurrent_activation_0"), val = tensor("sigmoid")]; - tensor input_cell_activation_0 = const()[name = tensor("input_cell_activation_0"), val = tensor("tanh")]; - tensor input_activation_0 = const()[name = tensor("input_activation_0"), val = tensor("tanh")]; - tensor concat_1_to_fp16 = const()[name = tensor("concat_1_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1312128)))]; - tensor concat_2_to_fp16 = const()[name = tensor("concat_2_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4588992)))]; - tensor concat_0_to_fp16 = const()[name = tensor("concat_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7865856)))]; - tensor input_3_cast_fp16 = transpose(perm = input_3_perm_0, x = y_cast_fp16_cast_uint16)[name = tensor("transpose_2")]; - tensor input_cast_fp16_0, tensor input_cast_fp16_1, tensor input_cast_fp16_2 = lstm(activation = input_activation_0, bias = concat_0_to_fp16, cell_activation = input_cell_activation_0, direction = input_direction_0, initial_c = input_lstm_c0_squeeze_cast_fp16, initial_h = input_lstm_h0_squeeze_cast_fp16, output_sequence = input_output_sequence_0, recurrent_activation = input_recurrent_activation_0, weight_hh = concat_2_to_fp16, weight_ih = concat_1_to_fp16, x = input_3_cast_fp16)[name = tensor("input_cast_fp16")]; - tensor obj_3_axes_0 = const()[name = tensor("obj_3_axes_0"), val = tensor([0])]; - tensor obj_3_cast_fp16 = expand_dims(axes = obj_3_axes_0, x = input_cast_fp16_1)[name = tensor("obj_3_cast_fp16")]; - tensor obj_3_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("obj_3_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; - tensor obj_axes_0 = const()[name = tensor("obj_axes_0"), val = tensor([0])]; - tensor obj_cast_fp16 = expand_dims(axes = obj_axes_0, x = input_cast_fp16_2)[name = tensor("obj_cast_fp16")]; - tensor obj_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("obj_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; - tensor transpose_0_perm_0 = const()[name = tensor("transpose_0_perm_0"), val = tensor([1, 2, 0])]; - tensor transpose_0_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("transpose_0_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; - tensor transpose_0_cast_fp16 = transpose(perm = transpose_0_perm_0, x = input_cast_fp16_0)[name = tensor("transpose_1")]; - tensor decoder = cast(dtype = transpose_0_cast_fp16_to_fp32_dtype_0, x = transpose_0_cast_fp16)[name = tensor("cast_3")]; - tensor c_out = cast(dtype = obj_cast_fp16_to_fp32_dtype_0, x = obj_cast_fp16)[name = tensor("cast_4")]; - tensor h_out = cast(dtype = obj_3_cast_fp16_to_fp32_dtype_0, x = obj_3_cast_fp16)[name = tensor("cast_5")]; - tensor target_length_tmp = identity(x = target_length)[name = tensor("target_length_tmp")]; - } -> (decoder, h_out, c_out); -} \ No newline at end of file diff --git a/parakeet-ctc-110m-coreml/Decoder.mlmodelc/weights/weight.bin b/parakeet-ctc-110m-coreml/Decoder.mlmodelc/weights/weight.bin deleted file mode 100644 index 4543ad4ea89b21ebed0de0d5ec3975d07eace7fb..0000000000000000000000000000000000000000 --- a/parakeet-ctc-110m-coreml/Decoder.mlmodelc/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:dd90b58597ee2c172c672dffe13b1110898ba07394c1a15efc96cc8c6b18411b -size 7871040 diff --git a/parakeet-ctc-110m-coreml/Encoder.mlmodelc/analytics/coremldata.bin b/parakeet-ctc-110m-coreml/Encoder.mlmodelc/analytics/coremldata.bin deleted file mode 100644 index d2787d4e5376d7f894d2b0a5a09f497a9b486c73..0000000000000000000000000000000000000000 --- a/parakeet-ctc-110m-coreml/Encoder.mlmodelc/analytics/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:b7ae65e2af616df46066b7efca2d7c19941666ac0685f4ed005666890a052b0d -size 243 diff --git a/parakeet-ctc-110m-coreml/Encoder.mlmodelc/coremldata.bin b/parakeet-ctc-110m-coreml/Encoder.mlmodelc/coremldata.bin deleted file mode 100644 index 9d9cdbfd392e8f7391b47c1c96f0fcbaab440389..0000000000000000000000000000000000000000 --- a/parakeet-ctc-110m-coreml/Encoder.mlmodelc/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:0713c2d6ac5f8f6fb9582be250351ebd8efc925f71f4261191165f1406f2ee5d -size 437 diff --git a/parakeet-ctc-110m-coreml/Encoder.mlmodelc/metadata.json b/parakeet-ctc-110m-coreml/Encoder.mlmodelc/metadata.json deleted file mode 100644 index 8e3f6cc9ddca206205987b2c828d2e595e19bfc2..0000000000000000000000000000000000000000 --- a/parakeet-ctc-110m-coreml/Encoder.mlmodelc/metadata.json +++ /dev/null @@ -1,105 +0,0 @@ -[ - { - "metadataOutputVersion" : "3.0", - "shortDescription" : "Parakeet 110M encoder (15 s window)", - "outputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 512 × 188)", - "shortDescription" : "", - "shape" : "[1, 512, 188]", - "name" : "encoder_output", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1)", - "shortDescription" : "", - "shape" : "[1]", - "name" : "encoder_length", - "type" : "MultiArray" - } - ], - "storagePrecision" : "Float16", - "modelParameters" : [ - - ], - "author" : "Fluid Inference", - "specificationVersion" : 8, - "mlProgramOperationTypeHistogram" : { - "Ios17.logicalAnd" : 2, - "Ios17.reshape" : 103, - "Ios16.softmax" : 17, - "Ios17.matmul" : 51, - "Ios17.transpose" : 123, - "Split" : 17, - "Ios17.expandDims" : 17, - "Select" : 51, - "Ios17.add" : 128, - "Tile" : 8, - "Ios17.sliceByIndex" : 34, - "Ios16.sigmoid" : 17, - "Pad" : 34, - "Ios17.logicalNot" : 2, - "Ios17.layerNorm" : 85, - "Ios16.silu" : 51, - "Ios17.less" : 5, - "Ios17.sub" : 3, - "Ios17.conv" : 56, - "Ios16.relu" : 3, - "Ios17.linear" : 137, - "Ios17.cast" : 11, - "Ios17.floorDiv" : 3, - "Ios17.mul" : 77 - }, - "computePrecision" : "Mixed (Float16, Float32, Int32)", - "isUpdatable" : "0", - "stateSchema" : [ - - ], - "availability" : { - "macOS" : "14.0", - "tvOS" : "17.0", - "visionOS" : "1.0", - "watchOS" : "10.0", - "iOS" : "17.0", - "macCatalyst" : "17.0" - }, - "modelType" : { - "name" : "MLModelType_mlProgram" - }, - "inputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 80 × 1501)", - "shortDescription" : "", - "shape" : "[1, 80, 1501]", - "name" : "mel_features", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1)", - "shortDescription" : "", - "shape" : "[1]", - "name" : "mel_length", - "type" : "MultiArray" - } - ], - "userDefinedMetadata" : { - "com.github.apple.coremltools.source_dialect" : "TorchScript", - "com.github.apple.coremltools.source" : "torch==2.9.0", - "com.github.apple.coremltools.version" : "8.3.0" - }, - "generatedClassName" : "parakeet_encoder", - "method" : "predict" - } -] \ No newline at end of file diff --git a/parakeet-ctc-110m-coreml/Encoder.mlmodelc/model.mil b/parakeet-ctc-110m-coreml/Encoder.mlmodelc/model.mil deleted file mode 100644 index 344bd1abf313c03c9ed9287c00893739f5d361bd..0000000000000000000000000000000000000000 --- a/parakeet-ctc-110m-coreml/Encoder.mlmodelc/model.mil +++ /dev/null @@ -1,2690 +0,0 @@ -program(1.0) -[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3510.2.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.9.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "8.3.0"}})] -{ - func main(tensor mel_features, tensor mel_length) { - tensor var_23 = const()[name = tensor("op_23"), val = tensor(-1)]; - tensor x_1_perm_0 = const()[name = tensor("x_1_perm_0"), val = tensor([0, 2, 1])]; - tensor mel_features_to_fp16_dtype_0 = const()[name = tensor("mel_features_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor tensor_1_axes_0 = const()[name = tensor("tensor_1_axes_0"), val = tensor([1])]; - tensor mel_features_to_fp16 = cast(dtype = mel_features_to_fp16_dtype_0, x = mel_features)[name = tensor("cast_182")]; - tensor x_1_cast_fp16 = transpose(perm = x_1_perm_0, x = mel_features_to_fp16)[name = tensor("transpose_207")]; - tensor tensor_1_cast_fp16 = expand_dims(axes = tensor_1_axes_0, x = x_1_cast_fp16)[name = tensor("tensor_1_cast_fp16")]; - tensor expand_dims_0 = const()[name = tensor("expand_dims_0"), val = tensor([[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390, 391, 392, 393, 394, 395, 396, 397, 398, 399, 400, 401, 402, 403, 404, 405, 406, 407, 408, 409, 410, 411, 412, 413, 414, 415, 416, 417, 418, 419, 420, 421, 422, 423, 424, 425, 426, 427, 428, 429, 430, 431, 432, 433, 434, 435, 436, 437, 438, 439, 440, 441, 442, 443, 444, 445, 446, 447, 448, 449, 450, 451, 452, 453, 454, 455, 456, 457, 458, 459, 460, 461, 462, 463, 464, 465, 466, 467, 468, 469, 470, 471, 472, 473, 474, 475, 476, 477, 478, 479, 480, 481, 482, 483, 484, 485, 486, 487, 488, 489, 490, 491, 492, 493, 494, 495, 496, 497, 498, 499, 500, 501, 502, 503, 504, 505, 506, 507, 508, 509, 510, 511, 512, 513, 514, 515, 516, 517, 518, 519, 520, 521, 522, 523, 524, 525, 526, 527, 528, 529, 530, 531, 532, 533, 534, 535, 536, 537, 538, 539, 540, 541, 542, 543, 544, 545, 546, 547, 548, 549, 550, 551, 552, 553, 554, 555, 556, 557, 558, 559, 560, 561, 562, 563, 564, 565, 566, 567, 568, 569, 570, 571, 572, 573, 574, 575, 576, 577, 578, 579, 580, 581, 582, 583, 584, 585, 586, 587, 588, 589, 590, 591, 592, 593, 594, 595, 596, 597, 598, 599, 600, 601, 602, 603, 604, 605, 606, 607, 608, 609, 610, 611, 612, 613, 614, 615, 616, 617, 618, 619, 620, 621, 622, 623, 624, 625, 626, 627, 628, 629, 630, 631, 632, 633, 634, 635, 636, 637, 638, 639, 640, 641, 642, 643, 644, 645, 646, 647, 648, 649, 650, 651, 652, 653, 654, 655, 656, 657, 658, 659, 660, 661, 662, 663, 664, 665, 666, 667, 668, 669, 670, 671, 672, 673, 674, 675, 676, 677, 678, 679, 680, 681, 682, 683, 684, 685, 686, 687, 688, 689, 690, 691, 692, 693, 694, 695, 696, 697, 698, 699, 700, 701, 702, 703, 704, 705, 706, 707, 708, 709, 710, 711, 712, 713, 714, 715, 716, 717, 718, 719, 720, 721, 722, 723, 724, 725, 726, 727, 728, 729, 730, 731, 732, 733, 734, 735, 736, 737, 738, 739, 740, 741, 742, 743, 744, 745, 746, 747, 748, 749, 750, 751, 752, 753, 754, 755, 756, 757, 758, 759, 760, 761, 762, 763, 764, 765, 766, 767, 768, 769, 770, 771, 772, 773, 774, 775, 776, 777, 778, 779, 780, 781, 782, 783, 784, 785, 786, 787, 788, 789, 790, 791, 792, 793, 794, 795, 796, 797, 798, 799, 800, 801, 802, 803, 804, 805, 806, 807, 808, 809, 810, 811, 812, 813, 814, 815, 816, 817, 818, 819, 820, 821, 822, 823, 824, 825, 826, 827, 828, 829, 830, 831, 832, 833, 834, 835, 836, 837, 838, 839, 840, 841, 842, 843, 844, 845, 846, 847, 848, 849, 850, 851, 852, 853, 854, 855, 856, 857, 858, 859, 860, 861, 862, 863, 864, 865, 866, 867, 868, 869, 870, 871, 872, 873, 874, 875, 876, 877, 878, 879, 880, 881, 882, 883, 884, 885, 886, 887, 888, 889, 890, 891, 892, 893, 894, 895, 896, 897, 898, 899, 900, 901, 902, 903, 904, 905, 906, 907, 908, 909, 910, 911, 912, 913, 914, 915, 916, 917, 918, 919, 920, 921, 922, 923, 924, 925, 926, 927, 928, 929, 930, 931, 932, 933, 934, 935, 936, 937, 938, 939, 940, 941, 942, 943, 944, 945, 946, 947, 948, 949, 950, 951, 952, 953, 954, 955, 956, 957, 958, 959, 960, 961, 962, 963, 964, 965, 966, 967, 968, 969, 970, 971, 972, 973, 974, 975, 976, 977, 978, 979, 980, 981, 982, 983, 984, 985, 986, 987, 988, 989, 990, 991, 992, 993, 994, 995, 996, 997, 998, 999, 1000, 1001, 1002, 1003, 1004, 1005, 1006, 1007, 1008, 1009, 1010, 1011, 1012, 1013, 1014, 1015, 1016, 1017, 1018, 1019, 1020, 1021, 1022, 1023, 1024, 1025, 1026, 1027, 1028, 1029, 1030, 1031, 1032, 1033, 1034, 1035, 1036, 1037, 1038, 1039, 1040, 1041, 1042, 1043, 1044, 1045, 1046, 1047, 1048, 1049, 1050, 1051, 1052, 1053, 1054, 1055, 1056, 1057, 1058, 1059, 1060, 1061, 1062, 1063, 1064, 1065, 1066, 1067, 1068, 1069, 1070, 1071, 1072, 1073, 1074, 1075, 1076, 1077, 1078, 1079, 1080, 1081, 1082, 1083, 1084, 1085, 1086, 1087, 1088, 1089, 1090, 1091, 1092, 1093, 1094, 1095, 1096, 1097, 1098, 1099, 1100, 1101, 1102, 1103, 1104, 1105, 1106, 1107, 1108, 1109, 1110, 1111, 1112, 1113, 1114, 1115, 1116, 1117, 1118, 1119, 1120, 1121, 1122, 1123, 1124, 1125, 1126, 1127, 1128, 1129, 1130, 1131, 1132, 1133, 1134, 1135, 1136, 1137, 1138, 1139, 1140, 1141, 1142, 1143, 1144, 1145, 1146, 1147, 1148, 1149, 1150, 1151, 1152, 1153, 1154, 1155, 1156, 1157, 1158, 1159, 1160, 1161, 1162, 1163, 1164, 1165, 1166, 1167, 1168, 1169, 1170, 1171, 1172, 1173, 1174, 1175, 1176, 1177, 1178, 1179, 1180, 1181, 1182, 1183, 1184, 1185, 1186, 1187, 1188, 1189, 1190, 1191, 1192, 1193, 1194, 1195, 1196, 1197, 1198, 1199, 1200, 1201, 1202, 1203, 1204, 1205, 1206, 1207, 1208, 1209, 1210, 1211, 1212, 1213, 1214, 1215, 1216, 1217, 1218, 1219, 1220, 1221, 1222, 1223, 1224, 1225, 1226, 1227, 1228, 1229, 1230, 1231, 1232, 1233, 1234, 1235, 1236, 1237, 1238, 1239, 1240, 1241, 1242, 1243, 1244, 1245, 1246, 1247, 1248, 1249, 1250, 1251, 1252, 1253, 1254, 1255, 1256, 1257, 1258, 1259, 1260, 1261, 1262, 1263, 1264, 1265, 1266, 1267, 1268, 1269, 1270, 1271, 1272, 1273, 1274, 1275, 1276, 1277, 1278, 1279, 1280, 1281, 1282, 1283, 1284, 1285, 1286, 1287, 1288, 1289, 1290, 1291, 1292, 1293, 1294, 1295, 1296, 1297, 1298, 1299, 1300, 1301, 1302, 1303, 1304, 1305, 1306, 1307, 1308, 1309, 1310, 1311, 1312, 1313, 1314, 1315, 1316, 1317, 1318, 1319, 1320, 1321, 1322, 1323, 1324, 1325, 1326, 1327, 1328, 1329, 1330, 1331, 1332, 1333, 1334, 1335, 1336, 1337, 1338, 1339, 1340, 1341, 1342, 1343, 1344, 1345, 1346, 1347, 1348, 1349, 1350, 1351, 1352, 1353, 1354, 1355, 1356, 1357, 1358, 1359, 1360, 1361, 1362, 1363, 1364, 1365, 1366, 1367, 1368, 1369, 1370, 1371, 1372, 1373, 1374, 1375, 1376, 1377, 1378, 1379, 1380, 1381, 1382, 1383, 1384, 1385, 1386, 1387, 1388, 1389, 1390, 1391, 1392, 1393, 1394, 1395, 1396, 1397, 1398, 1399, 1400, 1401, 1402, 1403, 1404, 1405, 1406, 1407, 1408, 1409, 1410, 1411, 1412, 1413, 1414, 1415, 1416, 1417, 1418, 1419, 1420, 1421, 1422, 1423, 1424, 1425, 1426, 1427, 1428, 1429, 1430, 1431, 1432, 1433, 1434, 1435, 1436, 1437, 1438, 1439, 1440, 1441, 1442, 1443, 1444, 1445, 1446, 1447, 1448, 1449, 1450, 1451, 1452, 1453, 1454, 1455, 1456, 1457, 1458, 1459, 1460, 1461, 1462, 1463, 1464, 1465, 1466, 1467, 1468, 1469, 1470, 1471, 1472, 1473, 1474, 1475, 1476, 1477, 1478, 1479, 1480, 1481, 1482, 1483, 1484, 1485, 1486, 1487, 1488, 1489, 1490, 1491, 1492, 1493, 1494, 1495, 1496, 1497, 1498, 1499, 1500]])]; - tensor var_87_axes_0 = const()[name = tensor("op_87_axes_0"), val = tensor([1])]; - tensor var_87 = expand_dims(axes = var_87_axes_0, x = mel_length)[name = tensor("op_87")]; - tensor time_mask_1 = less(x = expand_dims_0, y = var_87)[name = tensor("time_mask_1")]; - tensor var_89_axes_0 = const()[name = tensor("op_89_axes_0"), val = tensor([-1])]; - tensor var_89 = expand_dims(axes = var_89_axes_0, x = time_mask_1)[name = tensor("op_89")]; - tensor var_91_reps_0 = const()[name = tensor("op_91_reps_0"), val = tensor([1, 1, 80])]; - tensor var_91 = tile(reps = var_91_reps_0, x = var_89)[name = tensor("op_91")]; - tensor var_97_axes_0 = const()[name = tensor("op_97_axes_0"), val = tensor([1])]; - tensor cast_3_to_fp16_dtype_0 = const()[name = tensor("cast_3_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor var_91_to_fp16 = cast(dtype = cast_3_to_fp16_dtype_0, x = var_91)[name = tensor("cast_181")]; - tensor var_97_cast_fp16 = expand_dims(axes = var_97_axes_0, x = var_91_to_fp16)[name = tensor("op_97_cast_fp16")]; - tensor input_1_cast_fp16 = mul(x = tensor_1_cast_fp16, y = var_97_cast_fp16)[name = tensor("input_1_cast_fp16")]; - tensor tensor_3_pad_type_0 = const()[name = tensor("tensor_3_pad_type_0"), val = tensor("custom")]; - tensor tensor_3_pad_0 = const()[name = tensor("tensor_3_pad_0"), val = tensor([1, 1, 1, 1])]; - tensor tensor_3_strides_0 = const()[name = tensor("tensor_3_strides_0"), val = tensor([2, 2])]; - tensor tensor_3_dilations_0 = const()[name = tensor("tensor_3_dilations_0"), val = tensor([1, 1])]; - tensor tensor_3_groups_0 = const()[name = tensor("tensor_3_groups_0"), val = tensor(1)]; - tensor module_pre_encode_conv_0_weight_to_fp16 = const()[name = tensor("module_pre_encode_conv_0_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; - tensor module_pre_encode_conv_0_bias_to_fp16 = const()[name = tensor("module_pre_encode_conv_0_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4736)))]; - tensor tensor_3_cast_fp16 = conv(bias = module_pre_encode_conv_0_bias_to_fp16, dilations = tensor_3_dilations_0, groups = tensor_3_groups_0, pad = tensor_3_pad_0, pad_type = tensor_3_pad_type_0, strides = tensor_3_strides_0, weight = module_pre_encode_conv_0_weight_to_fp16, x = input_1_cast_fp16)[name = tensor("tensor_3_cast_fp16")]; - tensor cast_1_to_fp16_dtype_0 = const()[name = tensor("cast_1_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor var_108_promoted_to_fp16 = const()[name = tensor("op_108_promoted_to_fp16"), val = tensor(0x1p+0)]; - tensor mel_length_to_fp16 = cast(dtype = cast_1_to_fp16_dtype_0, x = mel_length)[name = tensor("cast_180")]; - tensor var_109_cast_fp16 = add(x = mel_length_to_fp16, y = var_108_promoted_to_fp16)[name = tensor("op_109_cast_fp16")]; - tensor var_110_promoted_to_fp16 = const()[name = tensor("op_110_promoted_to_fp16"), val = tensor(0x1p+0)]; - tensor var_111_cast_fp16 = add(x = var_109_cast_fp16, y = var_110_promoted_to_fp16)[name = tensor("op_111_cast_fp16")]; - tensor var_112_promoted_to_fp16 = const()[name = tensor("op_112_promoted_to_fp16"), val = tensor(0x1.8p+1)]; - tensor var_113_cast_fp16 = sub(x = var_111_cast_fp16, y = var_112_promoted_to_fp16)[name = tensor("op_113_cast_fp16")]; - tensor var_21_promoted_to_fp16 = const()[name = tensor("op_21_promoted_to_fp16"), val = tensor(0x1p+1)]; - tensor floor_div_0_cast_fp16 = floor_div(x = var_113_cast_fp16, y = var_21_promoted_to_fp16)[name = tensor("floor_div_0_cast_fp16")]; - tensor var_115_promoted_to_fp16 = const()[name = tensor("op_115_promoted_to_fp16"), val = tensor(0x1p+0)]; - tensor current_lengths_3_cast_fp16 = add(x = floor_div_0_cast_fp16, y = var_115_promoted_to_fp16)[name = tensor("current_lengths_3_cast_fp16")]; - tensor cast_4_dtype_0 = const()[name = tensor("cast_4_dtype_0"), val = tensor("int32")]; - tensor expand_dims_1 = const()[name = tensor("expand_dims_1"), val = tensor([[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390, 391, 392, 393, 394, 395, 396, 397, 398, 399, 400, 401, 402, 403, 404, 405, 406, 407, 408, 409, 410, 411, 412, 413, 414, 415, 416, 417, 418, 419, 420, 421, 422, 423, 424, 425, 426, 427, 428, 429, 430, 431, 432, 433, 434, 435, 436, 437, 438, 439, 440, 441, 442, 443, 444, 445, 446, 447, 448, 449, 450, 451, 452, 453, 454, 455, 456, 457, 458, 459, 460, 461, 462, 463, 464, 465, 466, 467, 468, 469, 470, 471, 472, 473, 474, 475, 476, 477, 478, 479, 480, 481, 482, 483, 484, 485, 486, 487, 488, 489, 490, 491, 492, 493, 494, 495, 496, 497, 498, 499, 500, 501, 502, 503, 504, 505, 506, 507, 508, 509, 510, 511, 512, 513, 514, 515, 516, 517, 518, 519, 520, 521, 522, 523, 524, 525, 526, 527, 528, 529, 530, 531, 532, 533, 534, 535, 536, 537, 538, 539, 540, 541, 542, 543, 544, 545, 546, 547, 548, 549, 550, 551, 552, 553, 554, 555, 556, 557, 558, 559, 560, 561, 562, 563, 564, 565, 566, 567, 568, 569, 570, 571, 572, 573, 574, 575, 576, 577, 578, 579, 580, 581, 582, 583, 584, 585, 586, 587, 588, 589, 590, 591, 592, 593, 594, 595, 596, 597, 598, 599, 600, 601, 602, 603, 604, 605, 606, 607, 608, 609, 610, 611, 612, 613, 614, 615, 616, 617, 618, 619, 620, 621, 622, 623, 624, 625, 626, 627, 628, 629, 630, 631, 632, 633, 634, 635, 636, 637, 638, 639, 640, 641, 642, 643, 644, 645, 646, 647, 648, 649, 650, 651, 652, 653, 654, 655, 656, 657, 658, 659, 660, 661, 662, 663, 664, 665, 666, 667, 668, 669, 670, 671, 672, 673, 674, 675, 676, 677, 678, 679, 680, 681, 682, 683, 684, 685, 686, 687, 688, 689, 690, 691, 692, 693, 694, 695, 696, 697, 698, 699, 700, 701, 702, 703, 704, 705, 706, 707, 708, 709, 710, 711, 712, 713, 714, 715, 716, 717, 718, 719, 720, 721, 722, 723, 724, 725, 726, 727, 728, 729, 730, 731, 732, 733, 734, 735, 736, 737, 738, 739, 740, 741, 742, 743, 744, 745, 746, 747, 748, 749, 750]])]; - tensor var_124_axes_0 = const()[name = tensor("op_124_axes_0"), val = tensor([1])]; - tensor current_lengths_3_cast_fp16_to_int32 = cast(dtype = cast_4_dtype_0, x = current_lengths_3_cast_fp16)[name = tensor("cast_179")]; - tensor var_124 = expand_dims(axes = var_124_axes_0, x = current_lengths_3_cast_fp16_to_int32)[name = tensor("op_124")]; - tensor time_mask_3 = less(x = expand_dims_1, y = var_124)[name = tensor("time_mask_3")]; - tensor var_126_axes_0 = const()[name = tensor("op_126_axes_0"), val = tensor([-1])]; - tensor var_126 = expand_dims(axes = var_126_axes_0, x = time_mask_3)[name = tensor("op_126")]; - tensor var_128_reps_0 = const()[name = tensor("op_128_reps_0"), val = tensor([1, 1, 40])]; - tensor var_128 = tile(reps = var_128_reps_0, x = var_126)[name = tensor("op_128")]; - tensor var_134_axes_0 = const()[name = tensor("op_134_axes_0"), val = tensor([1])]; - tensor cast_5_to_fp16_dtype_0 = const()[name = tensor("cast_5_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor var_128_to_fp16 = cast(dtype = cast_5_to_fp16_dtype_0, x = var_128)[name = tensor("cast_178")]; - tensor var_134_cast_fp16 = expand_dims(axes = var_134_axes_0, x = var_128_to_fp16)[name = tensor("op_134_cast_fp16")]; - tensor expanded_mask_3_reps_0 = const()[name = tensor("expanded_mask_3_reps_0"), val = tensor([1, 256, 1, 1])]; - tensor expanded_mask_3_cast_fp16 = tile(reps = expanded_mask_3_reps_0, x = var_134_cast_fp16)[name = tensor("expanded_mask_3_cast_fp16")]; - tensor input_3_cast_fp16 = mul(x = tensor_3_cast_fp16, y = expanded_mask_3_cast_fp16)[name = tensor("input_3_cast_fp16")]; - tensor tensor_5_cast_fp16 = relu(x = input_3_cast_fp16)[name = tensor("tensor_5_cast_fp16")]; - tensor input_5_cast_fp16 = mul(x = tensor_5_cast_fp16, y = expanded_mask_3_cast_fp16)[name = tensor("input_5_cast_fp16")]; - tensor tensor_7_pad_type_0 = const()[name = tensor("tensor_7_pad_type_0"), val = tensor("custom")]; - tensor tensor_7_pad_0 = const()[name = tensor("tensor_7_pad_0"), val = tensor([1, 1, 1, 1])]; - tensor tensor_7_strides_0 = const()[name = tensor("tensor_7_strides_0"), val = tensor([2, 2])]; - tensor tensor_7_groups_0 = const()[name = tensor("tensor_7_groups_0"), val = tensor(256)]; - tensor tensor_7_dilations_0 = const()[name = tensor("tensor_7_dilations_0"), val = tensor([1, 1])]; - tensor module_pre_encode_conv_2_weight_to_fp16 = const()[name = tensor("module_pre_encode_conv_2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(5312)))]; - tensor module_pre_encode_conv_2_bias_to_fp16 = const()[name = tensor("module_pre_encode_conv_2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9984)))]; - tensor tensor_7_cast_fp16 = conv(bias = module_pre_encode_conv_2_bias_to_fp16, dilations = tensor_7_dilations_0, groups = tensor_7_groups_0, pad = tensor_7_pad_0, pad_type = tensor_7_pad_type_0, strides = tensor_7_strides_0, weight = module_pre_encode_conv_2_weight_to_fp16, x = input_5_cast_fp16)[name = tensor("tensor_7_cast_fp16")]; - tensor var_154_promoted_to_fp16 = const()[name = tensor("op_154_promoted_to_fp16"), val = tensor(0x1p+0)]; - tensor var_155_cast_fp16 = add(x = current_lengths_3_cast_fp16, y = var_154_promoted_to_fp16)[name = tensor("op_155_cast_fp16")]; - tensor var_156_promoted_to_fp16 = const()[name = tensor("op_156_promoted_to_fp16"), val = tensor(0x1p+0)]; - tensor var_157_cast_fp16 = add(x = var_155_cast_fp16, y = var_156_promoted_to_fp16)[name = tensor("op_157_cast_fp16")]; - tensor var_158_promoted_to_fp16 = const()[name = tensor("op_158_promoted_to_fp16"), val = tensor(0x1.8p+1)]; - tensor var_159_cast_fp16 = sub(x = var_157_cast_fp16, y = var_158_promoted_to_fp16)[name = tensor("op_159_cast_fp16")]; - tensor var_21_promoted_1_to_fp16 = const()[name = tensor("op_21_promoted_1_to_fp16"), val = tensor(0x1p+1)]; - tensor floor_div_1_cast_fp16 = floor_div(x = var_159_cast_fp16, y = var_21_promoted_1_to_fp16)[name = tensor("floor_div_1_cast_fp16")]; - tensor var_161_promoted_to_fp16 = const()[name = tensor("op_161_promoted_to_fp16"), val = tensor(0x1p+0)]; - tensor current_lengths_5_cast_fp16 = add(x = floor_div_1_cast_fp16, y = var_161_promoted_to_fp16)[name = tensor("current_lengths_5_cast_fp16")]; - tensor cast_6_dtype_0 = const()[name = tensor("cast_6_dtype_0"), val = tensor("int32")]; - tensor expand_dims_2 = const()[name = tensor("expand_dims_2"), val = tensor([[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375]])]; - tensor var_170_axes_0 = const()[name = tensor("op_170_axes_0"), val = tensor([1])]; - tensor current_lengths_5_cast_fp16_to_int32 = cast(dtype = cast_6_dtype_0, x = current_lengths_5_cast_fp16)[name = tensor("cast_177")]; - tensor var_170 = expand_dims(axes = var_170_axes_0, x = current_lengths_5_cast_fp16_to_int32)[name = tensor("op_170")]; - tensor time_mask_5 = less(x = expand_dims_2, y = var_170)[name = tensor("time_mask_5")]; - tensor var_172_axes_0 = const()[name = tensor("op_172_axes_0"), val = tensor([-1])]; - tensor var_172 = expand_dims(axes = var_172_axes_0, x = time_mask_5)[name = tensor("op_172")]; - tensor var_174_reps_0 = const()[name = tensor("op_174_reps_0"), val = tensor([1, 1, 20])]; - tensor var_174 = tile(reps = var_174_reps_0, x = var_172)[name = tensor("op_174")]; - tensor var_180_axes_0 = const()[name = tensor("op_180_axes_0"), val = tensor([1])]; - tensor cast_7_to_fp16_dtype_0 = const()[name = tensor("cast_7_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor var_174_to_fp16 = cast(dtype = cast_7_to_fp16_dtype_0, x = var_174)[name = tensor("cast_176")]; - tensor var_180_cast_fp16 = expand_dims(axes = var_180_axes_0, x = var_174_to_fp16)[name = tensor("op_180_cast_fp16")]; - tensor expanded_mask_7_reps_0 = const()[name = tensor("expanded_mask_7_reps_0"), val = tensor([1, 256, 1, 1])]; - tensor expanded_mask_7_cast_fp16 = tile(reps = expanded_mask_7_reps_0, x = var_180_cast_fp16)[name = tensor("expanded_mask_7_cast_fp16")]; - tensor input_7_cast_fp16 = mul(x = tensor_7_cast_fp16, y = expanded_mask_7_cast_fp16)[name = tensor("input_7_cast_fp16")]; - tensor tensor_9_pad_type_0 = const()[name = tensor("tensor_9_pad_type_0"), val = tensor("valid")]; - tensor tensor_9_strides_0 = const()[name = tensor("tensor_9_strides_0"), val = tensor([1, 1])]; - tensor tensor_9_pad_0 = const()[name = tensor("tensor_9_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor tensor_9_dilations_0 = const()[name = tensor("tensor_9_dilations_0"), val = tensor([1, 1])]; - tensor tensor_9_groups_0 = const()[name = tensor("tensor_9_groups_0"), val = tensor(1)]; - tensor module_pre_encode_conv_3_weight_to_fp16 = const()[name = tensor("module_pre_encode_conv_3_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10560)))]; - tensor module_pre_encode_conv_3_bias_to_fp16 = const()[name = tensor("module_pre_encode_conv_3_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141696)))]; - tensor tensor_9_cast_fp16 = conv(bias = module_pre_encode_conv_3_bias_to_fp16, dilations = tensor_9_dilations_0, groups = tensor_9_groups_0, pad = tensor_9_pad_0, pad_type = tensor_9_pad_type_0, strides = tensor_9_strides_0, weight = module_pre_encode_conv_3_weight_to_fp16, x = input_7_cast_fp16)[name = tensor("tensor_9_cast_fp16")]; - tensor input_9_cast_fp16 = mul(x = tensor_9_cast_fp16, y = expanded_mask_7_cast_fp16)[name = tensor("input_9_cast_fp16")]; - tensor tensor_11_cast_fp16 = relu(x = input_9_cast_fp16)[name = tensor("tensor_11_cast_fp16")]; - tensor input_11_cast_fp16 = mul(x = tensor_11_cast_fp16, y = expanded_mask_7_cast_fp16)[name = tensor("input_11_cast_fp16")]; - tensor tensor_13_pad_type_0 = const()[name = tensor("tensor_13_pad_type_0"), val = tensor("custom")]; - tensor tensor_13_pad_0 = const()[name = tensor("tensor_13_pad_0"), val = tensor([1, 1, 1, 1])]; - tensor tensor_13_strides_0 = const()[name = tensor("tensor_13_strides_0"), val = tensor([2, 2])]; - tensor tensor_13_groups_0 = const()[name = tensor("tensor_13_groups_0"), val = tensor(256)]; - tensor tensor_13_dilations_0 = const()[name = tensor("tensor_13_dilations_0"), val = tensor([1, 1])]; - tensor module_pre_encode_conv_5_weight_to_fp16 = const()[name = tensor("module_pre_encode_conv_5_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(142272)))]; - tensor module_pre_encode_conv_5_bias_to_fp16 = const()[name = tensor("module_pre_encode_conv_5_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146944)))]; - tensor tensor_13_cast_fp16 = conv(bias = module_pre_encode_conv_5_bias_to_fp16, dilations = tensor_13_dilations_0, groups = tensor_13_groups_0, pad = tensor_13_pad_0, pad_type = tensor_13_pad_type_0, strides = tensor_13_strides_0, weight = module_pre_encode_conv_5_weight_to_fp16, x = input_11_cast_fp16)[name = tensor("tensor_13_cast_fp16")]; - tensor var_215_promoted_to_fp16 = const()[name = tensor("op_215_promoted_to_fp16"), val = tensor(0x1p+0)]; - tensor var_216_cast_fp16 = add(x = current_lengths_5_cast_fp16, y = var_215_promoted_to_fp16)[name = tensor("op_216_cast_fp16")]; - tensor var_217_promoted_to_fp16 = const()[name = tensor("op_217_promoted_to_fp16"), val = tensor(0x1p+0)]; - tensor var_218_cast_fp16 = add(x = var_216_cast_fp16, y = var_217_promoted_to_fp16)[name = tensor("op_218_cast_fp16")]; - tensor var_219_promoted_to_fp16 = const()[name = tensor("op_219_promoted_to_fp16"), val = tensor(0x1.8p+1)]; - tensor var_220_cast_fp16 = sub(x = var_218_cast_fp16, y = var_219_promoted_to_fp16)[name = tensor("op_220_cast_fp16")]; - tensor var_21_promoted_2_to_fp16 = const()[name = tensor("op_21_promoted_2_to_fp16"), val = tensor(0x1p+1)]; - tensor floor_div_2_cast_fp16 = floor_div(x = var_220_cast_fp16, y = var_21_promoted_2_to_fp16)[name = tensor("floor_div_2_cast_fp16")]; - tensor var_222_promoted_to_fp16 = const()[name = tensor("op_222_promoted_to_fp16"), val = tensor(0x1p+0)]; - tensor current_lengths_cast_fp16 = add(x = floor_div_2_cast_fp16, y = var_222_promoted_to_fp16)[name = tensor("current_lengths_cast_fp16")]; - tensor cast_8_dtype_0 = const()[name = tensor("cast_8_dtype_0"), val = tensor("int32")]; - tensor expand_dims_3 = const()[name = tensor("expand_dims_3"), val = tensor([[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187]])]; - tensor var_231_axes_0 = const()[name = tensor("op_231_axes_0"), val = tensor([1])]; - tensor current_lengths_cast_fp16_to_int32 = cast(dtype = cast_8_dtype_0, x = current_lengths_cast_fp16)[name = tensor("cast_175")]; - tensor var_231 = expand_dims(axes = var_231_axes_0, x = current_lengths_cast_fp16_to_int32)[name = tensor("op_231")]; - tensor time_mask = less(x = expand_dims_3, y = var_231)[name = tensor("time_mask")]; - tensor var_233_axes_0 = const()[name = tensor("op_233_axes_0"), val = tensor([-1])]; - tensor var_233 = expand_dims(axes = var_233_axes_0, x = time_mask)[name = tensor("op_233")]; - tensor var_235_reps_0 = const()[name = tensor("op_235_reps_0"), val = tensor([1, 1, 10])]; - tensor var_235 = tile(reps = var_235_reps_0, x = var_233)[name = tensor("op_235")]; - tensor var_241_axes_0 = const()[name = tensor("op_241_axes_0"), val = tensor([1])]; - tensor cast_9_to_fp16_dtype_0 = const()[name = tensor("cast_9_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor var_235_to_fp16 = cast(dtype = cast_9_to_fp16_dtype_0, x = var_235)[name = tensor("cast_174")]; - tensor var_241_cast_fp16 = expand_dims(axes = var_241_axes_0, x = var_235_to_fp16)[name = tensor("op_241_cast_fp16")]; - tensor expanded_mask_13_reps_0 = const()[name = tensor("expanded_mask_13_reps_0"), val = tensor([1, 256, 1, 1])]; - tensor expanded_mask_13_cast_fp16 = tile(reps = expanded_mask_13_reps_0, x = var_241_cast_fp16)[name = tensor("expanded_mask_13_cast_fp16")]; - tensor input_13_cast_fp16 = mul(x = tensor_13_cast_fp16, y = expanded_mask_13_cast_fp16)[name = tensor("input_13_cast_fp16")]; - tensor tensor_15_pad_type_0 = const()[name = tensor("tensor_15_pad_type_0"), val = tensor("valid")]; - tensor tensor_15_strides_0 = const()[name = tensor("tensor_15_strides_0"), val = tensor([1, 1])]; - tensor tensor_15_pad_0 = const()[name = tensor("tensor_15_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor tensor_15_dilations_0 = const()[name = tensor("tensor_15_dilations_0"), val = tensor([1, 1])]; - tensor tensor_15_groups_0 = const()[name = tensor("tensor_15_groups_0"), val = tensor(1)]; - tensor module_pre_encode_conv_6_weight_to_fp16 = const()[name = tensor("module_pre_encode_conv_6_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147520)))]; - tensor module_pre_encode_conv_6_bias_to_fp16 = const()[name = tensor("module_pre_encode_conv_6_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(278656)))]; - tensor tensor_15_cast_fp16 = conv(bias = module_pre_encode_conv_6_bias_to_fp16, dilations = tensor_15_dilations_0, groups = tensor_15_groups_0, pad = tensor_15_pad_0, pad_type = tensor_15_pad_type_0, strides = tensor_15_strides_0, weight = module_pre_encode_conv_6_weight_to_fp16, x = input_13_cast_fp16)[name = tensor("tensor_15_cast_fp16")]; - tensor input_15_cast_fp16 = mul(x = tensor_15_cast_fp16, y = expanded_mask_13_cast_fp16)[name = tensor("input_15_cast_fp16")]; - tensor tensor_cast_fp16 = relu(x = input_15_cast_fp16)[name = tensor("tensor_cast_fp16")]; - tensor x_3_cast_fp16 = mul(x = tensor_cast_fp16, y = expanded_mask_13_cast_fp16)[name = tensor("x_3_cast_fp16")]; - tensor var_275_perm_0 = const()[name = tensor("op_275_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_276 = const()[name = tensor("op_276"), val = tensor([1, 188, -1])]; - tensor var_275_cast_fp16 = transpose(perm = var_275_perm_0, x = x_3_cast_fp16)[name = tensor("transpose_206")]; - tensor input_17_cast_fp16 = reshape(shape = var_276, x = var_275_cast_fp16)[name = tensor("input_17_cast_fp16")]; - tensor module_pre_encode_out_weight_to_fp16 = const()[name = tensor("module_pre_encode_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(279232)))]; - tensor module_pre_encode_out_bias_to_fp16 = const()[name = tensor("module_pre_encode_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2900736)))]; - tensor linear_0_cast_fp16 = linear(bias = module_pre_encode_out_bias_to_fp16, weight = module_pre_encode_out_weight_to_fp16, x = input_17_cast_fp16)[name = tensor("linear_0_cast_fp16")]; - tensor cast_12_dtype_0 = const()[name = tensor("cast_12_dtype_0"), val = tensor("int32")]; - tensor var_314_axes_0 = const()[name = tensor("op_314_axes_0"), val = tensor([-1])]; - tensor encoder_length = cast(dtype = cast_12_dtype_0, x = current_lengths_cast_fp16)[name = tensor("cast_173")]; - tensor var_314 = expand_dims(axes = var_314_axes_0, x = encoder_length)[name = tensor("op_314")]; - tensor pad_mask_1 = less(x = expand_dims_3, y = var_314)[name = tensor("pad_mask_1")]; - tensor var_316_axes_0 = const()[name = tensor("op_316_axes_0"), val = tensor([1])]; - tensor var_316 = expand_dims(axes = var_316_axes_0, x = pad_mask_1)[name = tensor("op_316")]; - tensor var_317 = const()[name = tensor("op_317"), val = tensor([1, 188, 1])]; - tensor pad_mask_for_att_mask_1 = tile(reps = var_317, x = var_316)[name = tensor("pad_mask_for_att_mask_1")]; - tensor var_319_perm_0 = const()[name = tensor("op_319_perm_0"), val = tensor([0, 2, 1])]; - tensor var_319 = transpose(perm = var_319_perm_0, x = pad_mask_for_att_mask_1)[name = tensor("transpose_205")]; - tensor pad_mask_for_att_mask = logical_and(x = pad_mask_for_att_mask_1, y = var_319)[name = tensor("pad_mask_for_att_mask")]; - tensor const_63 = const()[name = tensor("const_63"), val = tensor([[[true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true]]])]; - tensor att_mask = logical_and(x = pad_mask_for_att_mask, y = const_63)[name = tensor("att_mask")]; - tensor mask_9 = logical_not(x = att_mask)[name = tensor("mask_9")]; - tensor pad_mask = logical_not(x = pad_mask_1)[name = tensor("pad_mask")]; - tensor input_21_axes_0 = const()[name = tensor("input_21_axes_0"), val = tensor([-1])]; - tensor module_layers_0_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_0_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2901824)))]; - tensor module_layers_0_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_0_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2902912)))]; - tensor var_9_to_fp16 = const()[name = tensor("op_9_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_21_cast_fp16 = layer_norm(axes = input_21_axes_0, beta = module_layers_0_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_0_norm_feed_forward1_weight_to_fp16, x = linear_0_cast_fp16)[name = tensor("input_21_cast_fp16")]; - tensor module_layers_0_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_0_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2904000)))]; - tensor module_layers_0_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_0_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(5001216)))]; - tensor linear_1_cast_fp16 = linear(bias = module_layers_0_feed_forward1_linear1_bias_to_fp16, weight = module_layers_0_feed_forward1_linear1_weight_to_fp16, x = input_21_cast_fp16)[name = tensor("linear_1_cast_fp16")]; - tensor input_25_cast_fp16 = silu(x = linear_1_cast_fp16)[name = tensor("input_25_cast_fp16")]; - tensor module_layers_0_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_0_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(5005376)))]; - tensor module_layers_0_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_0_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7102592)))]; - tensor linear_2_cast_fp16 = linear(bias = module_layers_0_feed_forward1_linear2_bias_to_fp16, weight = module_layers_0_feed_forward1_linear2_weight_to_fp16, x = input_25_cast_fp16)[name = tensor("linear_2_cast_fp16")]; - tensor var_352_to_fp16 = const()[name = tensor("op_352_to_fp16"), val = tensor(0x1p-1)]; - tensor var_353_cast_fp16 = mul(x = linear_2_cast_fp16, y = var_352_to_fp16)[name = tensor("op_353_cast_fp16")]; - tensor input_31_cast_fp16 = add(x = linear_0_cast_fp16, y = var_353_cast_fp16)[name = tensor("input_31_cast_fp16")]; - tensor query_1_axes_0 = const()[name = tensor("query_1_axes_0"), val = tensor([-1])]; - tensor module_layers_0_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_0_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7103680)))]; - tensor module_layers_0_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_0_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7104768)))]; - tensor query_1_cast_fp16 = layer_norm(axes = query_1_axes_0, beta = module_layers_0_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_0_norm_self_att_weight_to_fp16, x = input_31_cast_fp16)[name = tensor("query_1_cast_fp16")]; - tensor module_layers_0_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_0_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7105856)))]; - tensor module_layers_0_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_0_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7630208)))]; - tensor linear_3_cast_fp16 = linear(bias = module_layers_0_self_attn_linear_q_bias_to_fp16, weight = module_layers_0_self_attn_linear_q_weight_to_fp16, x = query_1_cast_fp16)[name = tensor("linear_3_cast_fp16")]; - tensor var_370 = const()[name = tensor("op_370"), val = tensor([1, -1, 8, 64])]; - tensor q_1_cast_fp16 = reshape(shape = var_370, x = linear_3_cast_fp16)[name = tensor("q_1_cast_fp16")]; - tensor module_layers_0_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_0_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7631296)))]; - tensor module_layers_0_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_0_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8155648)))]; - tensor linear_4_cast_fp16 = linear(bias = module_layers_0_self_attn_linear_k_bias_to_fp16, weight = module_layers_0_self_attn_linear_k_weight_to_fp16, x = query_1_cast_fp16)[name = tensor("linear_4_cast_fp16")]; - tensor var_375 = const()[name = tensor("op_375"), val = tensor([1, -1, 8, 64])]; - tensor k_1_cast_fp16 = reshape(shape = var_375, x = linear_4_cast_fp16)[name = tensor("k_1_cast_fp16")]; - tensor module_layers_0_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_0_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8156736)))]; - tensor module_layers_0_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_0_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8681088)))]; - tensor linear_5_cast_fp16 = linear(bias = module_layers_0_self_attn_linear_v_bias_to_fp16, weight = module_layers_0_self_attn_linear_v_weight_to_fp16, x = query_1_cast_fp16)[name = tensor("linear_5_cast_fp16")]; - tensor var_380 = const()[name = tensor("op_380"), val = tensor([1, -1, 8, 64])]; - tensor v_1_cast_fp16 = reshape(shape = var_380, x = linear_5_cast_fp16)[name = tensor("v_1_cast_fp16")]; - tensor value_3_perm_0 = const()[name = tensor("value_3_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_0_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_0_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8682176)))]; - tensor var_392_cast_fp16 = add(x = q_1_cast_fp16, y = module_layers_0_self_attn_pos_bias_u_to_fp16)[name = tensor("op_392_cast_fp16")]; - tensor module_layers_0_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_0_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8683264)))]; - tensor var_394_cast_fp16 = add(x = q_1_cast_fp16, y = module_layers_0_self_attn_pos_bias_v_to_fp16)[name = tensor("op_394_cast_fp16")]; - tensor q_with_bias_v_1_perm_0 = const()[name = tensor("q_with_bias_v_1_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_7_transpose_x_0 = const()[name = tensor("x_7_transpose_x_0"), val = tensor(false)]; - tensor x_7_transpose_y_0 = const()[name = tensor("x_7_transpose_y_0"), val = tensor(false)]; - tensor var_396_to_fp16 = const()[name = tensor("op_396_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8684352)))]; - tensor q_with_bias_v_1_cast_fp16 = transpose(perm = q_with_bias_v_1_perm_0, x = var_394_cast_fp16)[name = tensor("transpose_203")]; - tensor x_7_cast_fp16 = matmul(transpose_x = x_7_transpose_x_0, transpose_y = x_7_transpose_y_0, x = q_with_bias_v_1_cast_fp16, y = var_396_to_fp16)[name = tensor("x_7_cast_fp16")]; - tensor x_9_pad_0 = const()[name = tensor("x_9_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_9_mode_0 = const()[name = tensor("x_9_mode_0"), val = tensor("constant")]; - tensor const_70_to_fp16 = const()[name = tensor("const_70_to_fp16"), val = tensor(0x0p+0)]; - tensor x_9_cast_fp16 = pad(constant_val = const_70_to_fp16, mode = x_9_mode_0, pad = x_9_pad_0, x = x_7_cast_fp16)[name = tensor("x_9_cast_fp16")]; - tensor var_404 = const()[name = tensor("op_404"), val = tensor([1, 8, -1, 188])]; - tensor x_11_cast_fp16 = reshape(shape = var_404, x = x_9_cast_fp16)[name = tensor("x_11_cast_fp16")]; - tensor var_408_begin_0 = const()[name = tensor("op_408_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_408_end_0 = const()[name = tensor("op_408_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_408_end_mask_0 = const()[name = tensor("op_408_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_408_cast_fp16 = slice_by_index(begin = var_408_begin_0, end = var_408_end_0, end_mask = var_408_end_mask_0, x = x_11_cast_fp16)[name = tensor("op_408_cast_fp16")]; - tensor var_409 = const()[name = tensor("op_409"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_1_cast_fp16 = reshape(shape = var_409, x = var_408_cast_fp16)[name = tensor("matrix_bd_1_cast_fp16")]; - tensor matrix_ac_1_transpose_x_0 = const()[name = tensor("matrix_ac_1_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_1_transpose_y_0 = const()[name = tensor("matrix_ac_1_transpose_y_0"), val = tensor(false)]; - tensor transpose_51_perm_0 = const()[name = tensor("transpose_51_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_52_perm_0 = const()[name = tensor("transpose_52_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_52 = transpose(perm = transpose_52_perm_0, x = k_1_cast_fp16)[name = tensor("transpose_201")]; - tensor transpose_51 = transpose(perm = transpose_51_perm_0, x = var_392_cast_fp16)[name = tensor("transpose_202")]; - tensor matrix_ac_1_cast_fp16 = matmul(transpose_x = matrix_ac_1_transpose_x_0, transpose_y = matrix_ac_1_transpose_y_0, x = transpose_51, y = transpose_52)[name = tensor("matrix_ac_1_cast_fp16")]; - tensor matrix_bd_3_begin_0 = const()[name = tensor("matrix_bd_3_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_3_end_0 = const()[name = tensor("matrix_bd_3_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_3_end_mask_0 = const()[name = tensor("matrix_bd_3_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_3_cast_fp16 = slice_by_index(begin = matrix_bd_3_begin_0, end = matrix_bd_3_end_0, end_mask = matrix_bd_3_end_mask_0, x = matrix_bd_1_cast_fp16)[name = tensor("matrix_bd_3_cast_fp16")]; - tensor var_418_cast_fp16 = add(x = matrix_ac_1_cast_fp16, y = matrix_bd_3_cast_fp16)[name = tensor("op_418_cast_fp16")]; - tensor _inversed_scores_1_y_0_to_fp16 = const()[name = tensor("_inversed_scores_1_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_1_cast_fp16 = mul(x = var_418_cast_fp16, y = _inversed_scores_1_y_0_to_fp16)[name = tensor("_inversed_scores_1_cast_fp16")]; - tensor mask_11_axes_0 = const()[name = tensor("mask_11_axes_0"), val = tensor([1])]; - tensor mask_11 = expand_dims(axes = mask_11_axes_0, x = mask_9)[name = tensor("mask_11")]; - tensor var_12_to_fp16 = const()[name = tensor("op_12_to_fp16"), val = tensor(-0x1.388p+13)]; - tensor scores_3_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_1_cast_fp16, cond = mask_11)[name = tensor("scores_3_cast_fp16")]; - tensor var_424_cast_fp16 = softmax(axis = var_23, x = scores_3_cast_fp16)[name = tensor("op_424_cast_fp16")]; - tensor var_11_to_fp16 = const()[name = tensor("op_11_to_fp16"), val = tensor(0x0p+0)]; - tensor input_33_cast_fp16 = select(a = var_11_to_fp16, b = var_424_cast_fp16, cond = mask_11)[name = tensor("input_33_cast_fp16")]; - tensor x_13_transpose_x_0 = const()[name = tensor("x_13_transpose_x_0"), val = tensor(false)]; - tensor x_13_transpose_y_0 = const()[name = tensor("x_13_transpose_y_0"), val = tensor(false)]; - tensor value_3_cast_fp16 = transpose(perm = value_3_perm_0, x = v_1_cast_fp16)[name = tensor("transpose_204")]; - tensor x_13_cast_fp16 = matmul(transpose_x = x_13_transpose_x_0, transpose_y = x_13_transpose_y_0, x = input_33_cast_fp16, y = value_3_cast_fp16)[name = tensor("x_13_cast_fp16")]; - tensor var_428_perm_0 = const()[name = tensor("op_428_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_429 = const()[name = tensor("op_429"), val = tensor([1, -1, 512])]; - tensor var_428_cast_fp16 = transpose(perm = var_428_perm_0, x = x_13_cast_fp16)[name = tensor("transpose_200")]; - tensor input_35_cast_fp16 = reshape(shape = var_429, x = var_428_cast_fp16)[name = tensor("input_35_cast_fp16")]; - tensor module_layers_0_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_0_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9068416)))]; - tensor module_layers_0_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_0_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9592768)))]; - tensor linear_7_cast_fp16 = linear(bias = module_layers_0_self_attn_linear_out_bias_to_fp16, weight = module_layers_0_self_attn_linear_out_weight_to_fp16, x = input_35_cast_fp16)[name = tensor("linear_7_cast_fp16")]; - tensor input_39_cast_fp16 = add(x = input_31_cast_fp16, y = linear_7_cast_fp16)[name = tensor("input_39_cast_fp16")]; - tensor x_17_axes_0 = const()[name = tensor("x_17_axes_0"), val = tensor([-1])]; - tensor module_layers_0_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_0_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9593856)))]; - tensor module_layers_0_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_0_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9594944)))]; - tensor x_17_cast_fp16 = layer_norm(axes = x_17_axes_0, beta = module_layers_0_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_0_norm_conv_weight_to_fp16, x = input_39_cast_fp16)[name = tensor("x_17_cast_fp16")]; - tensor input_41_perm_0 = const()[name = tensor("input_41_perm_0"), val = tensor([0, 2, 1])]; - tensor input_43_pad_type_0 = const()[name = tensor("input_43_pad_type_0"), val = tensor("valid")]; - tensor input_43_strides_0 = const()[name = tensor("input_43_strides_0"), val = tensor([1])]; - tensor input_43_pad_0 = const()[name = tensor("input_43_pad_0"), val = tensor([0, 0])]; - tensor input_43_dilations_0 = const()[name = tensor("input_43_dilations_0"), val = tensor([1])]; - tensor input_43_groups_0 = const()[name = tensor("input_43_groups_0"), val = tensor(1)]; - tensor module_layers_0_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_0_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9596032)))]; - tensor module_layers_0_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_0_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10644672)))]; - tensor input_41_cast_fp16 = transpose(perm = input_41_perm_0, x = x_17_cast_fp16)[name = tensor("transpose_199")]; - tensor input_43_cast_fp16 = conv(bias = module_layers_0_conv_pointwise_conv1_bias_to_fp16, dilations = input_43_dilations_0, groups = input_43_groups_0, pad = input_43_pad_0, pad_type = input_43_pad_type_0, strides = input_43_strides_0, weight = module_layers_0_conv_pointwise_conv1_weight_to_fp16, x = input_41_cast_fp16)[name = tensor("input_43_cast_fp16")]; - tensor x_19_split_num_splits_0 = const()[name = tensor("x_19_split_num_splits_0"), val = tensor(2)]; - tensor x_19_split_axis_0 = const()[name = tensor("x_19_split_axis_0"), val = tensor(1)]; - tensor x_19_split_cast_fp16_0, tensor x_19_split_cast_fp16_1 = split(axis = x_19_split_axis_0, num_splits = x_19_split_num_splits_0, x = input_43_cast_fp16)[name = tensor("x_19_split_cast_fp16")]; - tensor x_19_split_1_sigmoid_cast_fp16 = sigmoid(x = x_19_split_cast_fp16_1)[name = tensor("x_19_split_1_sigmoid_cast_fp16")]; - tensor x_19_cast_fp16 = mul(x = x_19_split_cast_fp16_0, y = x_19_split_1_sigmoid_cast_fp16)[name = tensor("x_19_cast_fp16")]; - tensor var_453_axes_0 = const()[name = tensor("op_453_axes_0"), val = tensor([1])]; - tensor var_453 = expand_dims(axes = var_453_axes_0, x = pad_mask)[name = tensor("op_453")]; - tensor input_45_cast_fp16 = select(a = var_11_to_fp16, b = x_19_cast_fp16, cond = var_453)[name = tensor("input_45_cast_fp16")]; - tensor input_47_pad_0 = const()[name = tensor("input_47_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_47_mode_0 = const()[name = tensor("input_47_mode_0"), val = tensor("constant")]; - tensor const_73_to_fp16 = const()[name = tensor("const_73_to_fp16"), val = tensor(0x0p+0)]; - tensor input_47_cast_fp16 = pad(constant_val = const_73_to_fp16, mode = input_47_mode_0, pad = input_47_pad_0, x = input_45_cast_fp16)[name = tensor("input_47_cast_fp16")]; - tensor input_49_pad_type_0 = const()[name = tensor("input_49_pad_type_0"), val = tensor("valid")]; - tensor input_49_groups_0 = const()[name = tensor("input_49_groups_0"), val = tensor(512)]; - tensor input_49_strides_0 = const()[name = tensor("input_49_strides_0"), val = tensor([1])]; - tensor input_49_pad_0 = const()[name = tensor("input_49_pad_0"), val = tensor([0, 0])]; - tensor input_49_dilations_0 = const()[name = tensor("input_49_dilations_0"), val = tensor([1])]; - tensor const_234_to_fp16 = const()[name = tensor("const_234_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10646784)))]; - tensor const_235_to_fp16 = const()[name = tensor("const_235_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10656064)))]; - tensor input_51_cast_fp16 = conv(bias = const_235_to_fp16, dilations = input_49_dilations_0, groups = input_49_groups_0, pad = input_49_pad_0, pad_type = input_49_pad_type_0, strides = input_49_strides_0, weight = const_234_to_fp16, x = input_47_cast_fp16)[name = tensor("input_51_cast_fp16")]; - tensor input_53_cast_fp16 = silu(x = input_51_cast_fp16)[name = tensor("input_53_cast_fp16")]; - tensor x_21_pad_type_0 = const()[name = tensor("x_21_pad_type_0"), val = tensor("valid")]; - tensor x_21_strides_0 = const()[name = tensor("x_21_strides_0"), val = tensor([1])]; - tensor x_21_pad_0 = const()[name = tensor("x_21_pad_0"), val = tensor([0, 0])]; - tensor x_21_dilations_0 = const()[name = tensor("x_21_dilations_0"), val = tensor([1])]; - tensor x_21_groups_0 = const()[name = tensor("x_21_groups_0"), val = tensor(1)]; - tensor module_layers_0_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_0_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10657152)))]; - tensor module_layers_0_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_0_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11181504)))]; - tensor x_21_cast_fp16 = conv(bias = module_layers_0_conv_pointwise_conv2_bias_to_fp16, dilations = x_21_dilations_0, groups = x_21_groups_0, pad = x_21_pad_0, pad_type = x_21_pad_type_0, strides = x_21_strides_0, weight = module_layers_0_conv_pointwise_conv2_weight_to_fp16, x = input_53_cast_fp16)[name = tensor("x_21_cast_fp16")]; - tensor input_55_perm_0 = const()[name = tensor("input_55_perm_0"), val = tensor([0, 2, 1])]; - tensor input_55_cast_fp16 = transpose(perm = input_55_perm_0, x = x_21_cast_fp16)[name = tensor("transpose_198")]; - tensor input_57_cast_fp16 = add(x = input_39_cast_fp16, y = input_55_cast_fp16)[name = tensor("input_57_cast_fp16")]; - tensor input_59_axes_0 = const()[name = tensor("input_59_axes_0"), val = tensor([-1])]; - tensor module_layers_0_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_0_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11182592)))]; - tensor module_layers_0_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_0_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11183680)))]; - tensor input_59_cast_fp16 = layer_norm(axes = input_59_axes_0, beta = module_layers_0_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_0_norm_feed_forward2_weight_to_fp16, x = input_57_cast_fp16)[name = tensor("input_59_cast_fp16")]; - tensor module_layers_0_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_0_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11184768)))]; - tensor module_layers_0_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_0_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(13281984)))]; - tensor linear_8_cast_fp16 = linear(bias = module_layers_0_feed_forward2_linear1_bias_to_fp16, weight = module_layers_0_feed_forward2_linear1_weight_to_fp16, x = input_59_cast_fp16)[name = tensor("linear_8_cast_fp16")]; - tensor input_63_cast_fp16 = silu(x = linear_8_cast_fp16)[name = tensor("input_63_cast_fp16")]; - tensor module_layers_0_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_0_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(13286144)))]; - tensor module_layers_0_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_0_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15383360)))]; - tensor linear_9_cast_fp16 = linear(bias = module_layers_0_feed_forward2_linear2_bias_to_fp16, weight = module_layers_0_feed_forward2_linear2_weight_to_fp16, x = input_63_cast_fp16)[name = tensor("linear_9_cast_fp16")]; - tensor var_495_to_fp16 = const()[name = tensor("op_495_to_fp16"), val = tensor(0x1p-1)]; - tensor var_496_cast_fp16 = mul(x = linear_9_cast_fp16, y = var_495_to_fp16)[name = tensor("op_496_cast_fp16")]; - tensor input_69_cast_fp16 = add(x = input_57_cast_fp16, y = var_496_cast_fp16)[name = tensor("input_69_cast_fp16")]; - tensor input_71_axes_0 = const()[name = tensor("input_71_axes_0"), val = tensor([-1])]; - tensor module_layers_0_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_0_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15384448)))]; - tensor module_layers_0_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_0_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15385536)))]; - tensor input_71_cast_fp16 = layer_norm(axes = input_71_axes_0, beta = module_layers_0_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_0_norm_out_weight_to_fp16, x = input_69_cast_fp16)[name = tensor("input_71_cast_fp16")]; - tensor input_73_axes_0 = const()[name = tensor("input_73_axes_0"), val = tensor([-1])]; - tensor module_layers_1_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_1_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15386624)))]; - tensor module_layers_1_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_1_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15387712)))]; - tensor input_73_cast_fp16 = layer_norm(axes = input_73_axes_0, beta = module_layers_1_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_1_norm_feed_forward1_weight_to_fp16, x = input_71_cast_fp16)[name = tensor("input_73_cast_fp16")]; - tensor module_layers_1_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_1_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15388800)))]; - tensor module_layers_1_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_1_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(17486016)))]; - tensor linear_10_cast_fp16 = linear(bias = module_layers_1_feed_forward1_linear1_bias_to_fp16, weight = module_layers_1_feed_forward1_linear1_weight_to_fp16, x = input_73_cast_fp16)[name = tensor("linear_10_cast_fp16")]; - tensor input_77_cast_fp16 = silu(x = linear_10_cast_fp16)[name = tensor("input_77_cast_fp16")]; - tensor module_layers_1_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_1_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(17490176)))]; - tensor module_layers_1_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_1_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19587392)))]; - tensor linear_11_cast_fp16 = linear(bias = module_layers_1_feed_forward1_linear2_bias_to_fp16, weight = module_layers_1_feed_forward1_linear2_weight_to_fp16, x = input_77_cast_fp16)[name = tensor("linear_11_cast_fp16")]; - tensor var_526_to_fp16 = const()[name = tensor("op_526_to_fp16"), val = tensor(0x1p-1)]; - tensor var_527_cast_fp16 = mul(x = linear_11_cast_fp16, y = var_526_to_fp16)[name = tensor("op_527_cast_fp16")]; - tensor input_83_cast_fp16 = add(x = input_71_cast_fp16, y = var_527_cast_fp16)[name = tensor("input_83_cast_fp16")]; - tensor query_3_axes_0 = const()[name = tensor("query_3_axes_0"), val = tensor([-1])]; - tensor module_layers_1_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_1_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19588480)))]; - tensor module_layers_1_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_1_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19589568)))]; - tensor query_3_cast_fp16 = layer_norm(axes = query_3_axes_0, beta = module_layers_1_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_1_norm_self_att_weight_to_fp16, x = input_83_cast_fp16)[name = tensor("query_3_cast_fp16")]; - tensor module_layers_1_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_1_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19590656)))]; - tensor module_layers_1_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_1_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(20115008)))]; - tensor linear_12_cast_fp16 = linear(bias = module_layers_1_self_attn_linear_q_bias_to_fp16, weight = module_layers_1_self_attn_linear_q_weight_to_fp16, x = query_3_cast_fp16)[name = tensor("linear_12_cast_fp16")]; - tensor var_544 = const()[name = tensor("op_544"), val = tensor([1, -1, 8, 64])]; - tensor q_7_cast_fp16 = reshape(shape = var_544, x = linear_12_cast_fp16)[name = tensor("q_7_cast_fp16")]; - tensor module_layers_1_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_1_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(20116096)))]; - tensor module_layers_1_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_1_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(20640448)))]; - tensor linear_13_cast_fp16 = linear(bias = module_layers_1_self_attn_linear_k_bias_to_fp16, weight = module_layers_1_self_attn_linear_k_weight_to_fp16, x = query_3_cast_fp16)[name = tensor("linear_13_cast_fp16")]; - tensor var_549 = const()[name = tensor("op_549"), val = tensor([1, -1, 8, 64])]; - tensor k_5_cast_fp16 = reshape(shape = var_549, x = linear_13_cast_fp16)[name = tensor("k_5_cast_fp16")]; - tensor module_layers_1_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_1_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(20641536)))]; - tensor module_layers_1_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_1_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21165888)))]; - tensor linear_14_cast_fp16 = linear(bias = module_layers_1_self_attn_linear_v_bias_to_fp16, weight = module_layers_1_self_attn_linear_v_weight_to_fp16, x = query_3_cast_fp16)[name = tensor("linear_14_cast_fp16")]; - tensor var_554 = const()[name = tensor("op_554"), val = tensor([1, -1, 8, 64])]; - tensor v_3_cast_fp16 = reshape(shape = var_554, x = linear_14_cast_fp16)[name = tensor("v_3_cast_fp16")]; - tensor value_5_perm_0 = const()[name = tensor("value_5_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_1_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_1_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21166976)))]; - tensor var_566_cast_fp16 = add(x = q_7_cast_fp16, y = module_layers_1_self_attn_pos_bias_u_to_fp16)[name = tensor("op_566_cast_fp16")]; - tensor module_layers_1_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_1_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21168064)))]; - tensor var_568_cast_fp16 = add(x = q_7_cast_fp16, y = module_layers_1_self_attn_pos_bias_v_to_fp16)[name = tensor("op_568_cast_fp16")]; - tensor q_with_bias_v_3_perm_0 = const()[name = tensor("q_with_bias_v_3_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_29_transpose_x_0 = const()[name = tensor("x_29_transpose_x_0"), val = tensor(false)]; - tensor x_29_transpose_y_0 = const()[name = tensor("x_29_transpose_y_0"), val = tensor(false)]; - tensor var_570_to_fp16 = const()[name = tensor("op_570_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21169152)))]; - tensor q_with_bias_v_3_cast_fp16 = transpose(perm = q_with_bias_v_3_perm_0, x = var_568_cast_fp16)[name = tensor("transpose_196")]; - tensor x_29_cast_fp16 = matmul(transpose_x = x_29_transpose_x_0, transpose_y = x_29_transpose_y_0, x = q_with_bias_v_3_cast_fp16, y = var_570_to_fp16)[name = tensor("x_29_cast_fp16")]; - tensor x_31_pad_0 = const()[name = tensor("x_31_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_31_mode_0 = const()[name = tensor("x_31_mode_0"), val = tensor("constant")]; - tensor const_80_to_fp16 = const()[name = tensor("const_80_to_fp16"), val = tensor(0x0p+0)]; - tensor x_31_cast_fp16 = pad(constant_val = const_80_to_fp16, mode = x_31_mode_0, pad = x_31_pad_0, x = x_29_cast_fp16)[name = tensor("x_31_cast_fp16")]; - tensor var_578 = const()[name = tensor("op_578"), val = tensor([1, 8, -1, 188])]; - tensor x_33_cast_fp16 = reshape(shape = var_578, x = x_31_cast_fp16)[name = tensor("x_33_cast_fp16")]; - tensor var_582_begin_0 = const()[name = tensor("op_582_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_582_end_0 = const()[name = tensor("op_582_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_582_end_mask_0 = const()[name = tensor("op_582_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_582_cast_fp16 = slice_by_index(begin = var_582_begin_0, end = var_582_end_0, end_mask = var_582_end_mask_0, x = x_33_cast_fp16)[name = tensor("op_582_cast_fp16")]; - tensor var_583 = const()[name = tensor("op_583"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_5_cast_fp16 = reshape(shape = var_583, x = var_582_cast_fp16)[name = tensor("matrix_bd_5_cast_fp16")]; - tensor matrix_ac_3_transpose_x_0 = const()[name = tensor("matrix_ac_3_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_3_transpose_y_0 = const()[name = tensor("matrix_ac_3_transpose_y_0"), val = tensor(false)]; - tensor transpose_53_perm_0 = const()[name = tensor("transpose_53_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_54_perm_0 = const()[name = tensor("transpose_54_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_54 = transpose(perm = transpose_54_perm_0, x = k_5_cast_fp16)[name = tensor("transpose_194")]; - tensor transpose_53 = transpose(perm = transpose_53_perm_0, x = var_566_cast_fp16)[name = tensor("transpose_195")]; - tensor matrix_ac_3_cast_fp16 = matmul(transpose_x = matrix_ac_3_transpose_x_0, transpose_y = matrix_ac_3_transpose_y_0, x = transpose_53, y = transpose_54)[name = tensor("matrix_ac_3_cast_fp16")]; - tensor matrix_bd_7_begin_0 = const()[name = tensor("matrix_bd_7_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_7_end_0 = const()[name = tensor("matrix_bd_7_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_7_end_mask_0 = const()[name = tensor("matrix_bd_7_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_7_cast_fp16 = slice_by_index(begin = matrix_bd_7_begin_0, end = matrix_bd_7_end_0, end_mask = matrix_bd_7_end_mask_0, x = matrix_bd_5_cast_fp16)[name = tensor("matrix_bd_7_cast_fp16")]; - tensor var_592_cast_fp16 = add(x = matrix_ac_3_cast_fp16, y = matrix_bd_7_cast_fp16)[name = tensor("op_592_cast_fp16")]; - tensor _inversed_scores_5_y_0_to_fp16 = const()[name = tensor("_inversed_scores_5_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_5_cast_fp16 = mul(x = var_592_cast_fp16, y = _inversed_scores_5_y_0_to_fp16)[name = tensor("_inversed_scores_5_cast_fp16")]; - tensor scores_7_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_5_cast_fp16, cond = mask_11)[name = tensor("scores_7_cast_fp16")]; - tensor var_598_cast_fp16 = softmax(axis = var_23, x = scores_7_cast_fp16)[name = tensor("op_598_cast_fp16")]; - tensor input_85_cast_fp16 = select(a = var_11_to_fp16, b = var_598_cast_fp16, cond = mask_11)[name = tensor("input_85_cast_fp16")]; - tensor x_35_transpose_x_0 = const()[name = tensor("x_35_transpose_x_0"), val = tensor(false)]; - tensor x_35_transpose_y_0 = const()[name = tensor("x_35_transpose_y_0"), val = tensor(false)]; - tensor value_5_cast_fp16 = transpose(perm = value_5_perm_0, x = v_3_cast_fp16)[name = tensor("transpose_197")]; - tensor x_35_cast_fp16 = matmul(transpose_x = x_35_transpose_x_0, transpose_y = x_35_transpose_y_0, x = input_85_cast_fp16, y = value_5_cast_fp16)[name = tensor("x_35_cast_fp16")]; - tensor var_602_perm_0 = const()[name = tensor("op_602_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_603 = const()[name = tensor("op_603"), val = tensor([1, -1, 512])]; - tensor var_602_cast_fp16 = transpose(perm = var_602_perm_0, x = x_35_cast_fp16)[name = tensor("transpose_193")]; - tensor input_87_cast_fp16 = reshape(shape = var_603, x = var_602_cast_fp16)[name = tensor("input_87_cast_fp16")]; - tensor module_layers_1_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_1_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21553216)))]; - tensor module_layers_1_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_1_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22077568)))]; - tensor linear_16_cast_fp16 = linear(bias = module_layers_1_self_attn_linear_out_bias_to_fp16, weight = module_layers_1_self_attn_linear_out_weight_to_fp16, x = input_87_cast_fp16)[name = tensor("linear_16_cast_fp16")]; - tensor input_91_cast_fp16 = add(x = input_83_cast_fp16, y = linear_16_cast_fp16)[name = tensor("input_91_cast_fp16")]; - tensor x_39_axes_0 = const()[name = tensor("x_39_axes_0"), val = tensor([-1])]; - tensor module_layers_1_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_1_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22078656)))]; - tensor module_layers_1_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_1_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22079744)))]; - tensor x_39_cast_fp16 = layer_norm(axes = x_39_axes_0, beta = module_layers_1_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_1_norm_conv_weight_to_fp16, x = input_91_cast_fp16)[name = tensor("x_39_cast_fp16")]; - tensor input_93_perm_0 = const()[name = tensor("input_93_perm_0"), val = tensor([0, 2, 1])]; - tensor input_95_pad_type_0 = const()[name = tensor("input_95_pad_type_0"), val = tensor("valid")]; - tensor input_95_strides_0 = const()[name = tensor("input_95_strides_0"), val = tensor([1])]; - tensor input_95_pad_0 = const()[name = tensor("input_95_pad_0"), val = tensor([0, 0])]; - tensor input_95_dilations_0 = const()[name = tensor("input_95_dilations_0"), val = tensor([1])]; - tensor input_95_groups_0 = const()[name = tensor("input_95_groups_0"), val = tensor(1)]; - tensor module_layers_1_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_1_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22080832)))]; - tensor module_layers_1_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_1_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23129472)))]; - tensor input_93_cast_fp16 = transpose(perm = input_93_perm_0, x = x_39_cast_fp16)[name = tensor("transpose_192")]; - tensor input_95_cast_fp16 = conv(bias = module_layers_1_conv_pointwise_conv1_bias_to_fp16, dilations = input_95_dilations_0, groups = input_95_groups_0, pad = input_95_pad_0, pad_type = input_95_pad_type_0, strides = input_95_strides_0, weight = module_layers_1_conv_pointwise_conv1_weight_to_fp16, x = input_93_cast_fp16)[name = tensor("input_95_cast_fp16")]; - tensor x_41_split_num_splits_0 = const()[name = tensor("x_41_split_num_splits_0"), val = tensor(2)]; - tensor x_41_split_axis_0 = const()[name = tensor("x_41_split_axis_0"), val = tensor(1)]; - tensor x_41_split_cast_fp16_0, tensor x_41_split_cast_fp16_1 = split(axis = x_41_split_axis_0, num_splits = x_41_split_num_splits_0, x = input_95_cast_fp16)[name = tensor("x_41_split_cast_fp16")]; - tensor x_41_split_1_sigmoid_cast_fp16 = sigmoid(x = x_41_split_cast_fp16_1)[name = tensor("x_41_split_1_sigmoid_cast_fp16")]; - tensor x_41_cast_fp16 = mul(x = x_41_split_cast_fp16_0, y = x_41_split_1_sigmoid_cast_fp16)[name = tensor("x_41_cast_fp16")]; - tensor input_97_cast_fp16 = select(a = var_11_to_fp16, b = x_41_cast_fp16, cond = var_453)[name = tensor("input_97_cast_fp16")]; - tensor input_99_pad_0 = const()[name = tensor("input_99_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_99_mode_0 = const()[name = tensor("input_99_mode_0"), val = tensor("constant")]; - tensor const_83_to_fp16 = const()[name = tensor("const_83_to_fp16"), val = tensor(0x0p+0)]; - tensor input_99_cast_fp16 = pad(constant_val = const_83_to_fp16, mode = input_99_mode_0, pad = input_99_pad_0, x = input_97_cast_fp16)[name = tensor("input_99_cast_fp16")]; - tensor input_101_pad_type_0 = const()[name = tensor("input_101_pad_type_0"), val = tensor("valid")]; - tensor input_101_groups_0 = const()[name = tensor("input_101_groups_0"), val = tensor(512)]; - tensor input_101_strides_0 = const()[name = tensor("input_101_strides_0"), val = tensor([1])]; - tensor input_101_pad_0 = const()[name = tensor("input_101_pad_0"), val = tensor([0, 0])]; - tensor input_101_dilations_0 = const()[name = tensor("input_101_dilations_0"), val = tensor([1])]; - tensor const_236_to_fp16 = const()[name = tensor("const_236_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23131584)))]; - tensor const_237_to_fp16 = const()[name = tensor("const_237_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23140864)))]; - tensor input_103_cast_fp16 = conv(bias = const_237_to_fp16, dilations = input_101_dilations_0, groups = input_101_groups_0, pad = input_101_pad_0, pad_type = input_101_pad_type_0, strides = input_101_strides_0, weight = const_236_to_fp16, x = input_99_cast_fp16)[name = tensor("input_103_cast_fp16")]; - tensor input_105_cast_fp16 = silu(x = input_103_cast_fp16)[name = tensor("input_105_cast_fp16")]; - tensor x_43_pad_type_0 = const()[name = tensor("x_43_pad_type_0"), val = tensor("valid")]; - tensor x_43_strides_0 = const()[name = tensor("x_43_strides_0"), val = tensor([1])]; - tensor x_43_pad_0 = const()[name = tensor("x_43_pad_0"), val = tensor([0, 0])]; - tensor x_43_dilations_0 = const()[name = tensor("x_43_dilations_0"), val = tensor([1])]; - tensor x_43_groups_0 = const()[name = tensor("x_43_groups_0"), val = tensor(1)]; - tensor module_layers_1_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_1_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23141952)))]; - tensor module_layers_1_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_1_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23666304)))]; - tensor x_43_cast_fp16 = conv(bias = module_layers_1_conv_pointwise_conv2_bias_to_fp16, dilations = x_43_dilations_0, groups = x_43_groups_0, pad = x_43_pad_0, pad_type = x_43_pad_type_0, strides = x_43_strides_0, weight = module_layers_1_conv_pointwise_conv2_weight_to_fp16, x = input_105_cast_fp16)[name = tensor("x_43_cast_fp16")]; - tensor input_107_perm_0 = const()[name = tensor("input_107_perm_0"), val = tensor([0, 2, 1])]; - tensor input_107_cast_fp16 = transpose(perm = input_107_perm_0, x = x_43_cast_fp16)[name = tensor("transpose_191")]; - tensor input_109_cast_fp16 = add(x = input_91_cast_fp16, y = input_107_cast_fp16)[name = tensor("input_109_cast_fp16")]; - tensor input_111_axes_0 = const()[name = tensor("input_111_axes_0"), val = tensor([-1])]; - tensor module_layers_1_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_1_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23667392)))]; - tensor module_layers_1_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_1_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23668480)))]; - tensor input_111_cast_fp16 = layer_norm(axes = input_111_axes_0, beta = module_layers_1_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_1_norm_feed_forward2_weight_to_fp16, x = input_109_cast_fp16)[name = tensor("input_111_cast_fp16")]; - tensor module_layers_1_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_1_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23669568)))]; - tensor module_layers_1_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_1_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(25766784)))]; - tensor linear_17_cast_fp16 = linear(bias = module_layers_1_feed_forward2_linear1_bias_to_fp16, weight = module_layers_1_feed_forward2_linear1_weight_to_fp16, x = input_111_cast_fp16)[name = tensor("linear_17_cast_fp16")]; - tensor input_115_cast_fp16 = silu(x = linear_17_cast_fp16)[name = tensor("input_115_cast_fp16")]; - tensor module_layers_1_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_1_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(25770944)))]; - tensor module_layers_1_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_1_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27868160)))]; - tensor linear_18_cast_fp16 = linear(bias = module_layers_1_feed_forward2_linear2_bias_to_fp16, weight = module_layers_1_feed_forward2_linear2_weight_to_fp16, x = input_115_cast_fp16)[name = tensor("linear_18_cast_fp16")]; - tensor var_669_to_fp16 = const()[name = tensor("op_669_to_fp16"), val = tensor(0x1p-1)]; - tensor var_670_cast_fp16 = mul(x = linear_18_cast_fp16, y = var_669_to_fp16)[name = tensor("op_670_cast_fp16")]; - tensor input_121_cast_fp16 = add(x = input_109_cast_fp16, y = var_670_cast_fp16)[name = tensor("input_121_cast_fp16")]; - tensor input_123_axes_0 = const()[name = tensor("input_123_axes_0"), val = tensor([-1])]; - tensor module_layers_1_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_1_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27869248)))]; - tensor module_layers_1_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_1_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27870336)))]; - tensor input_123_cast_fp16 = layer_norm(axes = input_123_axes_0, beta = module_layers_1_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_1_norm_out_weight_to_fp16, x = input_121_cast_fp16)[name = tensor("input_123_cast_fp16")]; - tensor input_125_axes_0 = const()[name = tensor("input_125_axes_0"), val = tensor([-1])]; - tensor module_layers_2_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_2_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27871424)))]; - tensor module_layers_2_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_2_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27872512)))]; - tensor input_125_cast_fp16 = layer_norm(axes = input_125_axes_0, beta = module_layers_2_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_2_norm_feed_forward1_weight_to_fp16, x = input_123_cast_fp16)[name = tensor("input_125_cast_fp16")]; - tensor module_layers_2_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_2_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27873600)))]; - tensor module_layers_2_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_2_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(29970816)))]; - tensor linear_19_cast_fp16 = linear(bias = module_layers_2_feed_forward1_linear1_bias_to_fp16, weight = module_layers_2_feed_forward1_linear1_weight_to_fp16, x = input_125_cast_fp16)[name = tensor("linear_19_cast_fp16")]; - tensor input_129_cast_fp16 = silu(x = linear_19_cast_fp16)[name = tensor("input_129_cast_fp16")]; - tensor module_layers_2_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_2_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(29974976)))]; - tensor module_layers_2_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_2_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32072192)))]; - tensor linear_20_cast_fp16 = linear(bias = module_layers_2_feed_forward1_linear2_bias_to_fp16, weight = module_layers_2_feed_forward1_linear2_weight_to_fp16, x = input_129_cast_fp16)[name = tensor("linear_20_cast_fp16")]; - tensor var_700_to_fp16 = const()[name = tensor("op_700_to_fp16"), val = tensor(0x1p-1)]; - tensor var_701_cast_fp16 = mul(x = linear_20_cast_fp16, y = var_700_to_fp16)[name = tensor("op_701_cast_fp16")]; - tensor input_135_cast_fp16 = add(x = input_123_cast_fp16, y = var_701_cast_fp16)[name = tensor("input_135_cast_fp16")]; - tensor query_5_axes_0 = const()[name = tensor("query_5_axes_0"), val = tensor([-1])]; - tensor module_layers_2_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_2_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32073280)))]; - tensor module_layers_2_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_2_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32074368)))]; - tensor query_5_cast_fp16 = layer_norm(axes = query_5_axes_0, beta = module_layers_2_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_2_norm_self_att_weight_to_fp16, x = input_135_cast_fp16)[name = tensor("query_5_cast_fp16")]; - tensor module_layers_2_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_2_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32075456)))]; - tensor module_layers_2_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_2_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32599808)))]; - tensor linear_21_cast_fp16 = linear(bias = module_layers_2_self_attn_linear_q_bias_to_fp16, weight = module_layers_2_self_attn_linear_q_weight_to_fp16, x = query_5_cast_fp16)[name = tensor("linear_21_cast_fp16")]; - tensor var_718 = const()[name = tensor("op_718"), val = tensor([1, -1, 8, 64])]; - tensor q_13_cast_fp16 = reshape(shape = var_718, x = linear_21_cast_fp16)[name = tensor("q_13_cast_fp16")]; - tensor module_layers_2_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_2_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32600896)))]; - tensor module_layers_2_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_2_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33125248)))]; - tensor linear_22_cast_fp16 = linear(bias = module_layers_2_self_attn_linear_k_bias_to_fp16, weight = module_layers_2_self_attn_linear_k_weight_to_fp16, x = query_5_cast_fp16)[name = tensor("linear_22_cast_fp16")]; - tensor var_723 = const()[name = tensor("op_723"), val = tensor([1, -1, 8, 64])]; - tensor k_9_cast_fp16 = reshape(shape = var_723, x = linear_22_cast_fp16)[name = tensor("k_9_cast_fp16")]; - tensor module_layers_2_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_2_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33126336)))]; - tensor module_layers_2_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_2_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33650688)))]; - tensor linear_23_cast_fp16 = linear(bias = module_layers_2_self_attn_linear_v_bias_to_fp16, weight = module_layers_2_self_attn_linear_v_weight_to_fp16, x = query_5_cast_fp16)[name = tensor("linear_23_cast_fp16")]; - tensor var_728 = const()[name = tensor("op_728"), val = tensor([1, -1, 8, 64])]; - tensor v_5_cast_fp16 = reshape(shape = var_728, x = linear_23_cast_fp16)[name = tensor("v_5_cast_fp16")]; - tensor value_7_perm_0 = const()[name = tensor("value_7_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_2_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_2_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33651776)))]; - tensor var_740_cast_fp16 = add(x = q_13_cast_fp16, y = module_layers_2_self_attn_pos_bias_u_to_fp16)[name = tensor("op_740_cast_fp16")]; - tensor module_layers_2_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_2_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33652864)))]; - tensor var_742_cast_fp16 = add(x = q_13_cast_fp16, y = module_layers_2_self_attn_pos_bias_v_to_fp16)[name = tensor("op_742_cast_fp16")]; - tensor q_with_bias_v_5_perm_0 = const()[name = tensor("q_with_bias_v_5_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_51_transpose_x_0 = const()[name = tensor("x_51_transpose_x_0"), val = tensor(false)]; - tensor x_51_transpose_y_0 = const()[name = tensor("x_51_transpose_y_0"), val = tensor(false)]; - tensor var_744_to_fp16 = const()[name = tensor("op_744_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33653952)))]; - tensor q_with_bias_v_5_cast_fp16 = transpose(perm = q_with_bias_v_5_perm_0, x = var_742_cast_fp16)[name = tensor("transpose_189")]; - tensor x_51_cast_fp16 = matmul(transpose_x = x_51_transpose_x_0, transpose_y = x_51_transpose_y_0, x = q_with_bias_v_5_cast_fp16, y = var_744_to_fp16)[name = tensor("x_51_cast_fp16")]; - tensor x_53_pad_0 = const()[name = tensor("x_53_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_53_mode_0 = const()[name = tensor("x_53_mode_0"), val = tensor("constant")]; - tensor const_90_to_fp16 = const()[name = tensor("const_90_to_fp16"), val = tensor(0x0p+0)]; - tensor x_53_cast_fp16 = pad(constant_val = const_90_to_fp16, mode = x_53_mode_0, pad = x_53_pad_0, x = x_51_cast_fp16)[name = tensor("x_53_cast_fp16")]; - tensor var_752 = const()[name = tensor("op_752"), val = tensor([1, 8, -1, 188])]; - tensor x_55_cast_fp16 = reshape(shape = var_752, x = x_53_cast_fp16)[name = tensor("x_55_cast_fp16")]; - tensor var_756_begin_0 = const()[name = tensor("op_756_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_756_end_0 = const()[name = tensor("op_756_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_756_end_mask_0 = const()[name = tensor("op_756_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_756_cast_fp16 = slice_by_index(begin = var_756_begin_0, end = var_756_end_0, end_mask = var_756_end_mask_0, x = x_55_cast_fp16)[name = tensor("op_756_cast_fp16")]; - tensor var_757 = const()[name = tensor("op_757"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_9_cast_fp16 = reshape(shape = var_757, x = var_756_cast_fp16)[name = tensor("matrix_bd_9_cast_fp16")]; - tensor matrix_ac_5_transpose_x_0 = const()[name = tensor("matrix_ac_5_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_5_transpose_y_0 = const()[name = tensor("matrix_ac_5_transpose_y_0"), val = tensor(false)]; - tensor transpose_55_perm_0 = const()[name = tensor("transpose_55_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_56_perm_0 = const()[name = tensor("transpose_56_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_56 = transpose(perm = transpose_56_perm_0, x = k_9_cast_fp16)[name = tensor("transpose_187")]; - tensor transpose_55 = transpose(perm = transpose_55_perm_0, x = var_740_cast_fp16)[name = tensor("transpose_188")]; - tensor matrix_ac_5_cast_fp16 = matmul(transpose_x = matrix_ac_5_transpose_x_0, transpose_y = matrix_ac_5_transpose_y_0, x = transpose_55, y = transpose_56)[name = tensor("matrix_ac_5_cast_fp16")]; - tensor matrix_bd_11_begin_0 = const()[name = tensor("matrix_bd_11_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_11_end_0 = const()[name = tensor("matrix_bd_11_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_11_end_mask_0 = const()[name = tensor("matrix_bd_11_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_11_cast_fp16 = slice_by_index(begin = matrix_bd_11_begin_0, end = matrix_bd_11_end_0, end_mask = matrix_bd_11_end_mask_0, x = matrix_bd_9_cast_fp16)[name = tensor("matrix_bd_11_cast_fp16")]; - tensor var_766_cast_fp16 = add(x = matrix_ac_5_cast_fp16, y = matrix_bd_11_cast_fp16)[name = tensor("op_766_cast_fp16")]; - tensor _inversed_scores_9_y_0_to_fp16 = const()[name = tensor("_inversed_scores_9_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_9_cast_fp16 = mul(x = var_766_cast_fp16, y = _inversed_scores_9_y_0_to_fp16)[name = tensor("_inversed_scores_9_cast_fp16")]; - tensor scores_11_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_9_cast_fp16, cond = mask_11)[name = tensor("scores_11_cast_fp16")]; - tensor var_772_cast_fp16 = softmax(axis = var_23, x = scores_11_cast_fp16)[name = tensor("op_772_cast_fp16")]; - tensor input_137_cast_fp16 = select(a = var_11_to_fp16, b = var_772_cast_fp16, cond = mask_11)[name = tensor("input_137_cast_fp16")]; - tensor x_57_transpose_x_0 = const()[name = tensor("x_57_transpose_x_0"), val = tensor(false)]; - tensor x_57_transpose_y_0 = const()[name = tensor("x_57_transpose_y_0"), val = tensor(false)]; - tensor value_7_cast_fp16 = transpose(perm = value_7_perm_0, x = v_5_cast_fp16)[name = tensor("transpose_190")]; - tensor x_57_cast_fp16 = matmul(transpose_x = x_57_transpose_x_0, transpose_y = x_57_transpose_y_0, x = input_137_cast_fp16, y = value_7_cast_fp16)[name = tensor("x_57_cast_fp16")]; - tensor var_776_perm_0 = const()[name = tensor("op_776_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_777 = const()[name = tensor("op_777"), val = tensor([1, -1, 512])]; - tensor var_776_cast_fp16 = transpose(perm = var_776_perm_0, x = x_57_cast_fp16)[name = tensor("transpose_186")]; - tensor input_139_cast_fp16 = reshape(shape = var_777, x = var_776_cast_fp16)[name = tensor("input_139_cast_fp16")]; - tensor module_layers_2_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_2_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34038016)))]; - tensor module_layers_2_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_2_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34562368)))]; - tensor linear_25_cast_fp16 = linear(bias = module_layers_2_self_attn_linear_out_bias_to_fp16, weight = module_layers_2_self_attn_linear_out_weight_to_fp16, x = input_139_cast_fp16)[name = tensor("linear_25_cast_fp16")]; - tensor input_143_cast_fp16 = add(x = input_135_cast_fp16, y = linear_25_cast_fp16)[name = tensor("input_143_cast_fp16")]; - tensor x_61_axes_0 = const()[name = tensor("x_61_axes_0"), val = tensor([-1])]; - tensor module_layers_2_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_2_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34563456)))]; - tensor module_layers_2_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_2_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34564544)))]; - tensor x_61_cast_fp16 = layer_norm(axes = x_61_axes_0, beta = module_layers_2_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_2_norm_conv_weight_to_fp16, x = input_143_cast_fp16)[name = tensor("x_61_cast_fp16")]; - tensor input_145_perm_0 = const()[name = tensor("input_145_perm_0"), val = tensor([0, 2, 1])]; - tensor input_147_pad_type_0 = const()[name = tensor("input_147_pad_type_0"), val = tensor("valid")]; - tensor input_147_strides_0 = const()[name = tensor("input_147_strides_0"), val = tensor([1])]; - tensor input_147_pad_0 = const()[name = tensor("input_147_pad_0"), val = tensor([0, 0])]; - tensor input_147_dilations_0 = const()[name = tensor("input_147_dilations_0"), val = tensor([1])]; - tensor input_147_groups_0 = const()[name = tensor("input_147_groups_0"), val = tensor(1)]; - tensor module_layers_2_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_2_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34565632)))]; - tensor module_layers_2_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_2_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35614272)))]; - tensor input_145_cast_fp16 = transpose(perm = input_145_perm_0, x = x_61_cast_fp16)[name = tensor("transpose_185")]; - tensor input_147_cast_fp16 = conv(bias = module_layers_2_conv_pointwise_conv1_bias_to_fp16, dilations = input_147_dilations_0, groups = input_147_groups_0, pad = input_147_pad_0, pad_type = input_147_pad_type_0, strides = input_147_strides_0, weight = module_layers_2_conv_pointwise_conv1_weight_to_fp16, x = input_145_cast_fp16)[name = tensor("input_147_cast_fp16")]; - tensor x_63_split_num_splits_0 = const()[name = tensor("x_63_split_num_splits_0"), val = tensor(2)]; - tensor x_63_split_axis_0 = const()[name = tensor("x_63_split_axis_0"), val = tensor(1)]; - tensor x_63_split_cast_fp16_0, tensor x_63_split_cast_fp16_1 = split(axis = x_63_split_axis_0, num_splits = x_63_split_num_splits_0, x = input_147_cast_fp16)[name = tensor("x_63_split_cast_fp16")]; - tensor x_63_split_1_sigmoid_cast_fp16 = sigmoid(x = x_63_split_cast_fp16_1)[name = tensor("x_63_split_1_sigmoid_cast_fp16")]; - tensor x_63_cast_fp16 = mul(x = x_63_split_cast_fp16_0, y = x_63_split_1_sigmoid_cast_fp16)[name = tensor("x_63_cast_fp16")]; - tensor input_149_cast_fp16 = select(a = var_11_to_fp16, b = x_63_cast_fp16, cond = var_453)[name = tensor("input_149_cast_fp16")]; - tensor input_151_pad_0 = const()[name = tensor("input_151_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_151_mode_0 = const()[name = tensor("input_151_mode_0"), val = tensor("constant")]; - tensor const_93_to_fp16 = const()[name = tensor("const_93_to_fp16"), val = tensor(0x0p+0)]; - tensor input_151_cast_fp16 = pad(constant_val = const_93_to_fp16, mode = input_151_mode_0, pad = input_151_pad_0, x = input_149_cast_fp16)[name = tensor("input_151_cast_fp16")]; - tensor input_153_pad_type_0 = const()[name = tensor("input_153_pad_type_0"), val = tensor("valid")]; - tensor input_153_groups_0 = const()[name = tensor("input_153_groups_0"), val = tensor(512)]; - tensor input_153_strides_0 = const()[name = tensor("input_153_strides_0"), val = tensor([1])]; - tensor input_153_pad_0 = const()[name = tensor("input_153_pad_0"), val = tensor([0, 0])]; - tensor input_153_dilations_0 = const()[name = tensor("input_153_dilations_0"), val = tensor([1])]; - tensor const_238_to_fp16 = const()[name = tensor("const_238_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35616384)))]; - tensor const_239_to_fp16 = const()[name = tensor("const_239_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35625664)))]; - tensor input_155_cast_fp16 = conv(bias = const_239_to_fp16, dilations = input_153_dilations_0, groups = input_153_groups_0, pad = input_153_pad_0, pad_type = input_153_pad_type_0, strides = input_153_strides_0, weight = const_238_to_fp16, x = input_151_cast_fp16)[name = tensor("input_155_cast_fp16")]; - tensor input_157_cast_fp16 = silu(x = input_155_cast_fp16)[name = tensor("input_157_cast_fp16")]; - tensor x_65_pad_type_0 = const()[name = tensor("x_65_pad_type_0"), val = tensor("valid")]; - tensor x_65_strides_0 = const()[name = tensor("x_65_strides_0"), val = tensor([1])]; - tensor x_65_pad_0 = const()[name = tensor("x_65_pad_0"), val = tensor([0, 0])]; - tensor x_65_dilations_0 = const()[name = tensor("x_65_dilations_0"), val = tensor([1])]; - tensor x_65_groups_0 = const()[name = tensor("x_65_groups_0"), val = tensor(1)]; - tensor module_layers_2_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_2_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35626752)))]; - tensor module_layers_2_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_2_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36151104)))]; - tensor x_65_cast_fp16 = conv(bias = module_layers_2_conv_pointwise_conv2_bias_to_fp16, dilations = x_65_dilations_0, groups = x_65_groups_0, pad = x_65_pad_0, pad_type = x_65_pad_type_0, strides = x_65_strides_0, weight = module_layers_2_conv_pointwise_conv2_weight_to_fp16, x = input_157_cast_fp16)[name = tensor("x_65_cast_fp16")]; - tensor input_159_perm_0 = const()[name = tensor("input_159_perm_0"), val = tensor([0, 2, 1])]; - tensor input_159_cast_fp16 = transpose(perm = input_159_perm_0, x = x_65_cast_fp16)[name = tensor("transpose_184")]; - tensor input_161_cast_fp16 = add(x = input_143_cast_fp16, y = input_159_cast_fp16)[name = tensor("input_161_cast_fp16")]; - tensor input_163_axes_0 = const()[name = tensor("input_163_axes_0"), val = tensor([-1])]; - tensor module_layers_2_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_2_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36152192)))]; - tensor module_layers_2_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_2_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36153280)))]; - tensor input_163_cast_fp16 = layer_norm(axes = input_163_axes_0, beta = module_layers_2_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_2_norm_feed_forward2_weight_to_fp16, x = input_161_cast_fp16)[name = tensor("input_163_cast_fp16")]; - tensor module_layers_2_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_2_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36154368)))]; - tensor module_layers_2_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_2_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(38251584)))]; - tensor linear_26_cast_fp16 = linear(bias = module_layers_2_feed_forward2_linear1_bias_to_fp16, weight = module_layers_2_feed_forward2_linear1_weight_to_fp16, x = input_163_cast_fp16)[name = tensor("linear_26_cast_fp16")]; - tensor input_167_cast_fp16 = silu(x = linear_26_cast_fp16)[name = tensor("input_167_cast_fp16")]; - tensor module_layers_2_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_2_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(38255744)))]; - tensor module_layers_2_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_2_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40352960)))]; - tensor linear_27_cast_fp16 = linear(bias = module_layers_2_feed_forward2_linear2_bias_to_fp16, weight = module_layers_2_feed_forward2_linear2_weight_to_fp16, x = input_167_cast_fp16)[name = tensor("linear_27_cast_fp16")]; - tensor var_843_to_fp16 = const()[name = tensor("op_843_to_fp16"), val = tensor(0x1p-1)]; - tensor var_844_cast_fp16 = mul(x = linear_27_cast_fp16, y = var_843_to_fp16)[name = tensor("op_844_cast_fp16")]; - tensor input_173_cast_fp16 = add(x = input_161_cast_fp16, y = var_844_cast_fp16)[name = tensor("input_173_cast_fp16")]; - tensor input_175_axes_0 = const()[name = tensor("input_175_axes_0"), val = tensor([-1])]; - tensor module_layers_2_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_2_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40354048)))]; - tensor module_layers_2_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_2_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40355136)))]; - tensor input_175_cast_fp16 = layer_norm(axes = input_175_axes_0, beta = module_layers_2_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_2_norm_out_weight_to_fp16, x = input_173_cast_fp16)[name = tensor("input_175_cast_fp16")]; - tensor input_177_axes_0 = const()[name = tensor("input_177_axes_0"), val = tensor([-1])]; - tensor module_layers_3_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_3_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40356224)))]; - tensor module_layers_3_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_3_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40357312)))]; - tensor input_177_cast_fp16 = layer_norm(axes = input_177_axes_0, beta = module_layers_3_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_3_norm_feed_forward1_weight_to_fp16, x = input_175_cast_fp16)[name = tensor("input_177_cast_fp16")]; - tensor module_layers_3_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_3_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40358400)))]; - tensor module_layers_3_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_3_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(42455616)))]; - tensor linear_28_cast_fp16 = linear(bias = module_layers_3_feed_forward1_linear1_bias_to_fp16, weight = module_layers_3_feed_forward1_linear1_weight_to_fp16, x = input_177_cast_fp16)[name = tensor("linear_28_cast_fp16")]; - tensor input_181_cast_fp16 = silu(x = linear_28_cast_fp16)[name = tensor("input_181_cast_fp16")]; - tensor module_layers_3_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_3_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(42459776)))]; - tensor module_layers_3_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_3_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44556992)))]; - tensor linear_29_cast_fp16 = linear(bias = module_layers_3_feed_forward1_linear2_bias_to_fp16, weight = module_layers_3_feed_forward1_linear2_weight_to_fp16, x = input_181_cast_fp16)[name = tensor("linear_29_cast_fp16")]; - tensor var_874_to_fp16 = const()[name = tensor("op_874_to_fp16"), val = tensor(0x1p-1)]; - tensor var_875_cast_fp16 = mul(x = linear_29_cast_fp16, y = var_874_to_fp16)[name = tensor("op_875_cast_fp16")]; - tensor input_187_cast_fp16 = add(x = input_175_cast_fp16, y = var_875_cast_fp16)[name = tensor("input_187_cast_fp16")]; - tensor query_7_axes_0 = const()[name = tensor("query_7_axes_0"), val = tensor([-1])]; - tensor module_layers_3_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_3_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44558080)))]; - tensor module_layers_3_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_3_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44559168)))]; - tensor query_7_cast_fp16 = layer_norm(axes = query_7_axes_0, beta = module_layers_3_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_3_norm_self_att_weight_to_fp16, x = input_187_cast_fp16)[name = tensor("query_7_cast_fp16")]; - tensor module_layers_3_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_3_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44560256)))]; - tensor module_layers_3_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_3_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45084608)))]; - tensor linear_30_cast_fp16 = linear(bias = module_layers_3_self_attn_linear_q_bias_to_fp16, weight = module_layers_3_self_attn_linear_q_weight_to_fp16, x = query_7_cast_fp16)[name = tensor("linear_30_cast_fp16")]; - tensor var_892 = const()[name = tensor("op_892"), val = tensor([1, -1, 8, 64])]; - tensor q_19_cast_fp16 = reshape(shape = var_892, x = linear_30_cast_fp16)[name = tensor("q_19_cast_fp16")]; - tensor module_layers_3_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_3_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45085696)))]; - tensor module_layers_3_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_3_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45610048)))]; - tensor linear_31_cast_fp16 = linear(bias = module_layers_3_self_attn_linear_k_bias_to_fp16, weight = module_layers_3_self_attn_linear_k_weight_to_fp16, x = query_7_cast_fp16)[name = tensor("linear_31_cast_fp16")]; - tensor var_897 = const()[name = tensor("op_897"), val = tensor([1, -1, 8, 64])]; - tensor k_13_cast_fp16 = reshape(shape = var_897, x = linear_31_cast_fp16)[name = tensor("k_13_cast_fp16")]; - tensor module_layers_3_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_3_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45611136)))]; - tensor module_layers_3_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_3_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46135488)))]; - tensor linear_32_cast_fp16 = linear(bias = module_layers_3_self_attn_linear_v_bias_to_fp16, weight = module_layers_3_self_attn_linear_v_weight_to_fp16, x = query_7_cast_fp16)[name = tensor("linear_32_cast_fp16")]; - tensor var_902 = const()[name = tensor("op_902"), val = tensor([1, -1, 8, 64])]; - tensor v_7_cast_fp16 = reshape(shape = var_902, x = linear_32_cast_fp16)[name = tensor("v_7_cast_fp16")]; - tensor value_9_perm_0 = const()[name = tensor("value_9_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_3_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_3_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46136576)))]; - tensor var_914_cast_fp16 = add(x = q_19_cast_fp16, y = module_layers_3_self_attn_pos_bias_u_to_fp16)[name = tensor("op_914_cast_fp16")]; - tensor module_layers_3_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_3_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46137664)))]; - tensor var_916_cast_fp16 = add(x = q_19_cast_fp16, y = module_layers_3_self_attn_pos_bias_v_to_fp16)[name = tensor("op_916_cast_fp16")]; - tensor q_with_bias_v_7_perm_0 = const()[name = tensor("q_with_bias_v_7_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_73_transpose_x_0 = const()[name = tensor("x_73_transpose_x_0"), val = tensor(false)]; - tensor x_73_transpose_y_0 = const()[name = tensor("x_73_transpose_y_0"), val = tensor(false)]; - tensor var_918_to_fp16 = const()[name = tensor("op_918_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46138752)))]; - tensor q_with_bias_v_7_cast_fp16 = transpose(perm = q_with_bias_v_7_perm_0, x = var_916_cast_fp16)[name = tensor("transpose_182")]; - tensor x_73_cast_fp16 = matmul(transpose_x = x_73_transpose_x_0, transpose_y = x_73_transpose_y_0, x = q_with_bias_v_7_cast_fp16, y = var_918_to_fp16)[name = tensor("x_73_cast_fp16")]; - tensor x_75_pad_0 = const()[name = tensor("x_75_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_75_mode_0 = const()[name = tensor("x_75_mode_0"), val = tensor("constant")]; - tensor const_100_to_fp16 = const()[name = tensor("const_100_to_fp16"), val = tensor(0x0p+0)]; - tensor x_75_cast_fp16 = pad(constant_val = const_100_to_fp16, mode = x_75_mode_0, pad = x_75_pad_0, x = x_73_cast_fp16)[name = tensor("x_75_cast_fp16")]; - tensor var_926 = const()[name = tensor("op_926"), val = tensor([1, 8, -1, 188])]; - tensor x_77_cast_fp16 = reshape(shape = var_926, x = x_75_cast_fp16)[name = tensor("x_77_cast_fp16")]; - tensor var_930_begin_0 = const()[name = tensor("op_930_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_930_end_0 = const()[name = tensor("op_930_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_930_end_mask_0 = const()[name = tensor("op_930_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_930_cast_fp16 = slice_by_index(begin = var_930_begin_0, end = var_930_end_0, end_mask = var_930_end_mask_0, x = x_77_cast_fp16)[name = tensor("op_930_cast_fp16")]; - tensor var_931 = const()[name = tensor("op_931"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_13_cast_fp16 = reshape(shape = var_931, x = var_930_cast_fp16)[name = tensor("matrix_bd_13_cast_fp16")]; - tensor matrix_ac_7_transpose_x_0 = const()[name = tensor("matrix_ac_7_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_7_transpose_y_0 = const()[name = tensor("matrix_ac_7_transpose_y_0"), val = tensor(false)]; - tensor transpose_57_perm_0 = const()[name = tensor("transpose_57_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_58_perm_0 = const()[name = tensor("transpose_58_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_58 = transpose(perm = transpose_58_perm_0, x = k_13_cast_fp16)[name = tensor("transpose_180")]; - tensor transpose_57 = transpose(perm = transpose_57_perm_0, x = var_914_cast_fp16)[name = tensor("transpose_181")]; - tensor matrix_ac_7_cast_fp16 = matmul(transpose_x = matrix_ac_7_transpose_x_0, transpose_y = matrix_ac_7_transpose_y_0, x = transpose_57, y = transpose_58)[name = tensor("matrix_ac_7_cast_fp16")]; - tensor matrix_bd_15_begin_0 = const()[name = tensor("matrix_bd_15_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_15_end_0 = const()[name = tensor("matrix_bd_15_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_15_end_mask_0 = const()[name = tensor("matrix_bd_15_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_15_cast_fp16 = slice_by_index(begin = matrix_bd_15_begin_0, end = matrix_bd_15_end_0, end_mask = matrix_bd_15_end_mask_0, x = matrix_bd_13_cast_fp16)[name = tensor("matrix_bd_15_cast_fp16")]; - tensor var_940_cast_fp16 = add(x = matrix_ac_7_cast_fp16, y = matrix_bd_15_cast_fp16)[name = tensor("op_940_cast_fp16")]; - tensor _inversed_scores_13_y_0_to_fp16 = const()[name = tensor("_inversed_scores_13_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_13_cast_fp16 = mul(x = var_940_cast_fp16, y = _inversed_scores_13_y_0_to_fp16)[name = tensor("_inversed_scores_13_cast_fp16")]; - tensor scores_15_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_13_cast_fp16, cond = mask_11)[name = tensor("scores_15_cast_fp16")]; - tensor var_946_cast_fp16 = softmax(axis = var_23, x = scores_15_cast_fp16)[name = tensor("op_946_cast_fp16")]; - tensor input_189_cast_fp16 = select(a = var_11_to_fp16, b = var_946_cast_fp16, cond = mask_11)[name = tensor("input_189_cast_fp16")]; - tensor x_79_transpose_x_0 = const()[name = tensor("x_79_transpose_x_0"), val = tensor(false)]; - tensor x_79_transpose_y_0 = const()[name = tensor("x_79_transpose_y_0"), val = tensor(false)]; - tensor value_9_cast_fp16 = transpose(perm = value_9_perm_0, x = v_7_cast_fp16)[name = tensor("transpose_183")]; - tensor x_79_cast_fp16 = matmul(transpose_x = x_79_transpose_x_0, transpose_y = x_79_transpose_y_0, x = input_189_cast_fp16, y = value_9_cast_fp16)[name = tensor("x_79_cast_fp16")]; - tensor var_950_perm_0 = const()[name = tensor("op_950_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_951 = const()[name = tensor("op_951"), val = tensor([1, -1, 512])]; - tensor var_950_cast_fp16 = transpose(perm = var_950_perm_0, x = x_79_cast_fp16)[name = tensor("transpose_179")]; - tensor input_191_cast_fp16 = reshape(shape = var_951, x = var_950_cast_fp16)[name = tensor("input_191_cast_fp16")]; - tensor module_layers_3_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_3_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46522816)))]; - tensor module_layers_3_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_3_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47047168)))]; - tensor linear_34_cast_fp16 = linear(bias = module_layers_3_self_attn_linear_out_bias_to_fp16, weight = module_layers_3_self_attn_linear_out_weight_to_fp16, x = input_191_cast_fp16)[name = tensor("linear_34_cast_fp16")]; - tensor input_195_cast_fp16 = add(x = input_187_cast_fp16, y = linear_34_cast_fp16)[name = tensor("input_195_cast_fp16")]; - tensor x_83_axes_0 = const()[name = tensor("x_83_axes_0"), val = tensor([-1])]; - tensor module_layers_3_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_3_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47048256)))]; - tensor module_layers_3_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_3_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47049344)))]; - tensor x_83_cast_fp16 = layer_norm(axes = x_83_axes_0, beta = module_layers_3_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_3_norm_conv_weight_to_fp16, x = input_195_cast_fp16)[name = tensor("x_83_cast_fp16")]; - tensor input_197_perm_0 = const()[name = tensor("input_197_perm_0"), val = tensor([0, 2, 1])]; - tensor input_199_pad_type_0 = const()[name = tensor("input_199_pad_type_0"), val = tensor("valid")]; - tensor input_199_strides_0 = const()[name = tensor("input_199_strides_0"), val = tensor([1])]; - tensor input_199_pad_0 = const()[name = tensor("input_199_pad_0"), val = tensor([0, 0])]; - tensor input_199_dilations_0 = const()[name = tensor("input_199_dilations_0"), val = tensor([1])]; - tensor input_199_groups_0 = const()[name = tensor("input_199_groups_0"), val = tensor(1)]; - tensor module_layers_3_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_3_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47050432)))]; - tensor module_layers_3_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_3_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48099072)))]; - tensor input_197_cast_fp16 = transpose(perm = input_197_perm_0, x = x_83_cast_fp16)[name = tensor("transpose_178")]; - tensor input_199_cast_fp16 = conv(bias = module_layers_3_conv_pointwise_conv1_bias_to_fp16, dilations = input_199_dilations_0, groups = input_199_groups_0, pad = input_199_pad_0, pad_type = input_199_pad_type_0, strides = input_199_strides_0, weight = module_layers_3_conv_pointwise_conv1_weight_to_fp16, x = input_197_cast_fp16)[name = tensor("input_199_cast_fp16")]; - tensor x_85_split_num_splits_0 = const()[name = tensor("x_85_split_num_splits_0"), val = tensor(2)]; - tensor x_85_split_axis_0 = const()[name = tensor("x_85_split_axis_0"), val = tensor(1)]; - tensor x_85_split_cast_fp16_0, tensor x_85_split_cast_fp16_1 = split(axis = x_85_split_axis_0, num_splits = x_85_split_num_splits_0, x = input_199_cast_fp16)[name = tensor("x_85_split_cast_fp16")]; - tensor x_85_split_1_sigmoid_cast_fp16 = sigmoid(x = x_85_split_cast_fp16_1)[name = tensor("x_85_split_1_sigmoid_cast_fp16")]; - tensor x_85_cast_fp16 = mul(x = x_85_split_cast_fp16_0, y = x_85_split_1_sigmoid_cast_fp16)[name = tensor("x_85_cast_fp16")]; - tensor input_201_cast_fp16 = select(a = var_11_to_fp16, b = x_85_cast_fp16, cond = var_453)[name = tensor("input_201_cast_fp16")]; - tensor input_203_pad_0 = const()[name = tensor("input_203_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_203_mode_0 = const()[name = tensor("input_203_mode_0"), val = tensor("constant")]; - tensor const_103_to_fp16 = const()[name = tensor("const_103_to_fp16"), val = tensor(0x0p+0)]; - tensor input_203_cast_fp16 = pad(constant_val = const_103_to_fp16, mode = input_203_mode_0, pad = input_203_pad_0, x = input_201_cast_fp16)[name = tensor("input_203_cast_fp16")]; - tensor input_205_pad_type_0 = const()[name = tensor("input_205_pad_type_0"), val = tensor("valid")]; - tensor input_205_groups_0 = const()[name = tensor("input_205_groups_0"), val = tensor(512)]; - tensor input_205_strides_0 = const()[name = tensor("input_205_strides_0"), val = tensor([1])]; - tensor input_205_pad_0 = const()[name = tensor("input_205_pad_0"), val = tensor([0, 0])]; - tensor input_205_dilations_0 = const()[name = tensor("input_205_dilations_0"), val = tensor([1])]; - tensor const_240_to_fp16 = const()[name = tensor("const_240_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48101184)))]; - tensor const_241_to_fp16 = const()[name = tensor("const_241_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48110464)))]; - tensor input_207_cast_fp16 = conv(bias = const_241_to_fp16, dilations = input_205_dilations_0, groups = input_205_groups_0, pad = input_205_pad_0, pad_type = input_205_pad_type_0, strides = input_205_strides_0, weight = const_240_to_fp16, x = input_203_cast_fp16)[name = tensor("input_207_cast_fp16")]; - tensor input_209_cast_fp16 = silu(x = input_207_cast_fp16)[name = tensor("input_209_cast_fp16")]; - tensor x_87_pad_type_0 = const()[name = tensor("x_87_pad_type_0"), val = tensor("valid")]; - tensor x_87_strides_0 = const()[name = tensor("x_87_strides_0"), val = tensor([1])]; - tensor x_87_pad_0 = const()[name = tensor("x_87_pad_0"), val = tensor([0, 0])]; - tensor x_87_dilations_0 = const()[name = tensor("x_87_dilations_0"), val = tensor([1])]; - tensor x_87_groups_0 = const()[name = tensor("x_87_groups_0"), val = tensor(1)]; - tensor module_layers_3_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_3_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48111552)))]; - tensor module_layers_3_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_3_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48635904)))]; - tensor x_87_cast_fp16 = conv(bias = module_layers_3_conv_pointwise_conv2_bias_to_fp16, dilations = x_87_dilations_0, groups = x_87_groups_0, pad = x_87_pad_0, pad_type = x_87_pad_type_0, strides = x_87_strides_0, weight = module_layers_3_conv_pointwise_conv2_weight_to_fp16, x = input_209_cast_fp16)[name = tensor("x_87_cast_fp16")]; - tensor input_211_perm_0 = const()[name = tensor("input_211_perm_0"), val = tensor([0, 2, 1])]; - tensor input_211_cast_fp16 = transpose(perm = input_211_perm_0, x = x_87_cast_fp16)[name = tensor("transpose_177")]; - tensor input_213_cast_fp16 = add(x = input_195_cast_fp16, y = input_211_cast_fp16)[name = tensor("input_213_cast_fp16")]; - tensor input_215_axes_0 = const()[name = tensor("input_215_axes_0"), val = tensor([-1])]; - tensor module_layers_3_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_3_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48636992)))]; - tensor module_layers_3_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_3_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48638080)))]; - tensor input_215_cast_fp16 = layer_norm(axes = input_215_axes_0, beta = module_layers_3_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_3_norm_feed_forward2_weight_to_fp16, x = input_213_cast_fp16)[name = tensor("input_215_cast_fp16")]; - tensor module_layers_3_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_3_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48639168)))]; - tensor module_layers_3_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_3_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(50736384)))]; - tensor linear_35_cast_fp16 = linear(bias = module_layers_3_feed_forward2_linear1_bias_to_fp16, weight = module_layers_3_feed_forward2_linear1_weight_to_fp16, x = input_215_cast_fp16)[name = tensor("linear_35_cast_fp16")]; - tensor input_219_cast_fp16 = silu(x = linear_35_cast_fp16)[name = tensor("input_219_cast_fp16")]; - tensor module_layers_3_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_3_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(50740544)))]; - tensor module_layers_3_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_3_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52837760)))]; - tensor linear_36_cast_fp16 = linear(bias = module_layers_3_feed_forward2_linear2_bias_to_fp16, weight = module_layers_3_feed_forward2_linear2_weight_to_fp16, x = input_219_cast_fp16)[name = tensor("linear_36_cast_fp16")]; - tensor var_1017_to_fp16 = const()[name = tensor("op_1017_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1018_cast_fp16 = mul(x = linear_36_cast_fp16, y = var_1017_to_fp16)[name = tensor("op_1018_cast_fp16")]; - tensor input_225_cast_fp16 = add(x = input_213_cast_fp16, y = var_1018_cast_fp16)[name = tensor("input_225_cast_fp16")]; - tensor input_227_axes_0 = const()[name = tensor("input_227_axes_0"), val = tensor([-1])]; - tensor module_layers_3_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_3_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52838848)))]; - tensor module_layers_3_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_3_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52839936)))]; - tensor input_227_cast_fp16 = layer_norm(axes = input_227_axes_0, beta = module_layers_3_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_3_norm_out_weight_to_fp16, x = input_225_cast_fp16)[name = tensor("input_227_cast_fp16")]; - tensor input_229_axes_0 = const()[name = tensor("input_229_axes_0"), val = tensor([-1])]; - tensor module_layers_4_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_4_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52841024)))]; - tensor module_layers_4_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_4_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52842112)))]; - tensor input_229_cast_fp16 = layer_norm(axes = input_229_axes_0, beta = module_layers_4_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_4_norm_feed_forward1_weight_to_fp16, x = input_227_cast_fp16)[name = tensor("input_229_cast_fp16")]; - tensor module_layers_4_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_4_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52843200)))]; - tensor module_layers_4_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_4_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(54940416)))]; - tensor linear_37_cast_fp16 = linear(bias = module_layers_4_feed_forward1_linear1_bias_to_fp16, weight = module_layers_4_feed_forward1_linear1_weight_to_fp16, x = input_229_cast_fp16)[name = tensor("linear_37_cast_fp16")]; - tensor input_233_cast_fp16 = silu(x = linear_37_cast_fp16)[name = tensor("input_233_cast_fp16")]; - tensor module_layers_4_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_4_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(54944576)))]; - tensor module_layers_4_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_4_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57041792)))]; - tensor linear_38_cast_fp16 = linear(bias = module_layers_4_feed_forward1_linear2_bias_to_fp16, weight = module_layers_4_feed_forward1_linear2_weight_to_fp16, x = input_233_cast_fp16)[name = tensor("linear_38_cast_fp16")]; - tensor var_1048_to_fp16 = const()[name = tensor("op_1048_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1049_cast_fp16 = mul(x = linear_38_cast_fp16, y = var_1048_to_fp16)[name = tensor("op_1049_cast_fp16")]; - tensor input_239_cast_fp16 = add(x = input_227_cast_fp16, y = var_1049_cast_fp16)[name = tensor("input_239_cast_fp16")]; - tensor query_9_axes_0 = const()[name = tensor("query_9_axes_0"), val = tensor([-1])]; - tensor module_layers_4_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_4_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57042880)))]; - tensor module_layers_4_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_4_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57043968)))]; - tensor query_9_cast_fp16 = layer_norm(axes = query_9_axes_0, beta = module_layers_4_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_4_norm_self_att_weight_to_fp16, x = input_239_cast_fp16)[name = tensor("query_9_cast_fp16")]; - tensor module_layers_4_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_4_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57045056)))]; - tensor module_layers_4_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_4_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57569408)))]; - tensor linear_39_cast_fp16 = linear(bias = module_layers_4_self_attn_linear_q_bias_to_fp16, weight = module_layers_4_self_attn_linear_q_weight_to_fp16, x = query_9_cast_fp16)[name = tensor("linear_39_cast_fp16")]; - tensor var_1066 = const()[name = tensor("op_1066"), val = tensor([1, -1, 8, 64])]; - tensor q_25_cast_fp16 = reshape(shape = var_1066, x = linear_39_cast_fp16)[name = tensor("q_25_cast_fp16")]; - tensor module_layers_4_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_4_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57570496)))]; - tensor module_layers_4_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_4_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58094848)))]; - tensor linear_40_cast_fp16 = linear(bias = module_layers_4_self_attn_linear_k_bias_to_fp16, weight = module_layers_4_self_attn_linear_k_weight_to_fp16, x = query_9_cast_fp16)[name = tensor("linear_40_cast_fp16")]; - tensor var_1071 = const()[name = tensor("op_1071"), val = tensor([1, -1, 8, 64])]; - tensor k_17_cast_fp16 = reshape(shape = var_1071, x = linear_40_cast_fp16)[name = tensor("k_17_cast_fp16")]; - tensor module_layers_4_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_4_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58095936)))]; - tensor module_layers_4_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_4_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58620288)))]; - tensor linear_41_cast_fp16 = linear(bias = module_layers_4_self_attn_linear_v_bias_to_fp16, weight = module_layers_4_self_attn_linear_v_weight_to_fp16, x = query_9_cast_fp16)[name = tensor("linear_41_cast_fp16")]; - tensor var_1076 = const()[name = tensor("op_1076"), val = tensor([1, -1, 8, 64])]; - tensor v_9_cast_fp16 = reshape(shape = var_1076, x = linear_41_cast_fp16)[name = tensor("v_9_cast_fp16")]; - tensor value_11_perm_0 = const()[name = tensor("value_11_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_4_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_4_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58621376)))]; - tensor var_1088_cast_fp16 = add(x = q_25_cast_fp16, y = module_layers_4_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1088_cast_fp16")]; - tensor module_layers_4_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_4_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58622464)))]; - tensor var_1090_cast_fp16 = add(x = q_25_cast_fp16, y = module_layers_4_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1090_cast_fp16")]; - tensor q_with_bias_v_9_perm_0 = const()[name = tensor("q_with_bias_v_9_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_95_transpose_x_0 = const()[name = tensor("x_95_transpose_x_0"), val = tensor(false)]; - tensor x_95_transpose_y_0 = const()[name = tensor("x_95_transpose_y_0"), val = tensor(false)]; - tensor var_1092_to_fp16 = const()[name = tensor("op_1092_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58623552)))]; - tensor q_with_bias_v_9_cast_fp16 = transpose(perm = q_with_bias_v_9_perm_0, x = var_1090_cast_fp16)[name = tensor("transpose_175")]; - tensor x_95_cast_fp16 = matmul(transpose_x = x_95_transpose_x_0, transpose_y = x_95_transpose_y_0, x = q_with_bias_v_9_cast_fp16, y = var_1092_to_fp16)[name = tensor("x_95_cast_fp16")]; - tensor x_97_pad_0 = const()[name = tensor("x_97_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_97_mode_0 = const()[name = tensor("x_97_mode_0"), val = tensor("constant")]; - tensor const_110_to_fp16 = const()[name = tensor("const_110_to_fp16"), val = tensor(0x0p+0)]; - tensor x_97_cast_fp16 = pad(constant_val = const_110_to_fp16, mode = x_97_mode_0, pad = x_97_pad_0, x = x_95_cast_fp16)[name = tensor("x_97_cast_fp16")]; - tensor var_1100 = const()[name = tensor("op_1100"), val = tensor([1, 8, -1, 188])]; - tensor x_99_cast_fp16 = reshape(shape = var_1100, x = x_97_cast_fp16)[name = tensor("x_99_cast_fp16")]; - tensor var_1104_begin_0 = const()[name = tensor("op_1104_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_1104_end_0 = const()[name = tensor("op_1104_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_1104_end_mask_0 = const()[name = tensor("op_1104_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_1104_cast_fp16 = slice_by_index(begin = var_1104_begin_0, end = var_1104_end_0, end_mask = var_1104_end_mask_0, x = x_99_cast_fp16)[name = tensor("op_1104_cast_fp16")]; - tensor var_1105 = const()[name = tensor("op_1105"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_17_cast_fp16 = reshape(shape = var_1105, x = var_1104_cast_fp16)[name = tensor("matrix_bd_17_cast_fp16")]; - tensor matrix_ac_9_transpose_x_0 = const()[name = tensor("matrix_ac_9_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_9_transpose_y_0 = const()[name = tensor("matrix_ac_9_transpose_y_0"), val = tensor(false)]; - tensor transpose_59_perm_0 = const()[name = tensor("transpose_59_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_60_perm_0 = const()[name = tensor("transpose_60_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_60 = transpose(perm = transpose_60_perm_0, x = k_17_cast_fp16)[name = tensor("transpose_173")]; - tensor transpose_59 = transpose(perm = transpose_59_perm_0, x = var_1088_cast_fp16)[name = tensor("transpose_174")]; - tensor matrix_ac_9_cast_fp16 = matmul(transpose_x = matrix_ac_9_transpose_x_0, transpose_y = matrix_ac_9_transpose_y_0, x = transpose_59, y = transpose_60)[name = tensor("matrix_ac_9_cast_fp16")]; - tensor matrix_bd_19_begin_0 = const()[name = tensor("matrix_bd_19_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_19_end_0 = const()[name = tensor("matrix_bd_19_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_19_end_mask_0 = const()[name = tensor("matrix_bd_19_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_19_cast_fp16 = slice_by_index(begin = matrix_bd_19_begin_0, end = matrix_bd_19_end_0, end_mask = matrix_bd_19_end_mask_0, x = matrix_bd_17_cast_fp16)[name = tensor("matrix_bd_19_cast_fp16")]; - tensor var_1114_cast_fp16 = add(x = matrix_ac_9_cast_fp16, y = matrix_bd_19_cast_fp16)[name = tensor("op_1114_cast_fp16")]; - tensor _inversed_scores_17_y_0_to_fp16 = const()[name = tensor("_inversed_scores_17_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_17_cast_fp16 = mul(x = var_1114_cast_fp16, y = _inversed_scores_17_y_0_to_fp16)[name = tensor("_inversed_scores_17_cast_fp16")]; - tensor scores_19_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_17_cast_fp16, cond = mask_11)[name = tensor("scores_19_cast_fp16")]; - tensor var_1120_cast_fp16 = softmax(axis = var_23, x = scores_19_cast_fp16)[name = tensor("op_1120_cast_fp16")]; - tensor input_241_cast_fp16 = select(a = var_11_to_fp16, b = var_1120_cast_fp16, cond = mask_11)[name = tensor("input_241_cast_fp16")]; - tensor x_101_transpose_x_0 = const()[name = tensor("x_101_transpose_x_0"), val = tensor(false)]; - tensor x_101_transpose_y_0 = const()[name = tensor("x_101_transpose_y_0"), val = tensor(false)]; - tensor value_11_cast_fp16 = transpose(perm = value_11_perm_0, x = v_9_cast_fp16)[name = tensor("transpose_176")]; - tensor x_101_cast_fp16 = matmul(transpose_x = x_101_transpose_x_0, transpose_y = x_101_transpose_y_0, x = input_241_cast_fp16, y = value_11_cast_fp16)[name = tensor("x_101_cast_fp16")]; - tensor var_1124_perm_0 = const()[name = tensor("op_1124_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_1125 = const()[name = tensor("op_1125"), val = tensor([1, -1, 512])]; - tensor var_1124_cast_fp16 = transpose(perm = var_1124_perm_0, x = x_101_cast_fp16)[name = tensor("transpose_172")]; - tensor input_243_cast_fp16 = reshape(shape = var_1125, x = var_1124_cast_fp16)[name = tensor("input_243_cast_fp16")]; - tensor module_layers_4_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_4_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59007616)))]; - tensor module_layers_4_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_4_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59531968)))]; - tensor linear_43_cast_fp16 = linear(bias = module_layers_4_self_attn_linear_out_bias_to_fp16, weight = module_layers_4_self_attn_linear_out_weight_to_fp16, x = input_243_cast_fp16)[name = tensor("linear_43_cast_fp16")]; - tensor input_247_cast_fp16 = add(x = input_239_cast_fp16, y = linear_43_cast_fp16)[name = tensor("input_247_cast_fp16")]; - tensor x_105_axes_0 = const()[name = tensor("x_105_axes_0"), val = tensor([-1])]; - tensor module_layers_4_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_4_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59533056)))]; - tensor module_layers_4_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_4_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59534144)))]; - tensor x_105_cast_fp16 = layer_norm(axes = x_105_axes_0, beta = module_layers_4_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_4_norm_conv_weight_to_fp16, x = input_247_cast_fp16)[name = tensor("x_105_cast_fp16")]; - tensor input_249_perm_0 = const()[name = tensor("input_249_perm_0"), val = tensor([0, 2, 1])]; - tensor input_251_pad_type_0 = const()[name = tensor("input_251_pad_type_0"), val = tensor("valid")]; - tensor input_251_strides_0 = const()[name = tensor("input_251_strides_0"), val = tensor([1])]; - tensor input_251_pad_0 = const()[name = tensor("input_251_pad_0"), val = tensor([0, 0])]; - tensor input_251_dilations_0 = const()[name = tensor("input_251_dilations_0"), val = tensor([1])]; - tensor input_251_groups_0 = const()[name = tensor("input_251_groups_0"), val = tensor(1)]; - tensor module_layers_4_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_4_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59535232)))]; - tensor module_layers_4_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_4_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60583872)))]; - tensor input_249_cast_fp16 = transpose(perm = input_249_perm_0, x = x_105_cast_fp16)[name = tensor("transpose_171")]; - tensor input_251_cast_fp16 = conv(bias = module_layers_4_conv_pointwise_conv1_bias_to_fp16, dilations = input_251_dilations_0, groups = input_251_groups_0, pad = input_251_pad_0, pad_type = input_251_pad_type_0, strides = input_251_strides_0, weight = module_layers_4_conv_pointwise_conv1_weight_to_fp16, x = input_249_cast_fp16)[name = tensor("input_251_cast_fp16")]; - tensor x_107_split_num_splits_0 = const()[name = tensor("x_107_split_num_splits_0"), val = tensor(2)]; - tensor x_107_split_axis_0 = const()[name = tensor("x_107_split_axis_0"), val = tensor(1)]; - tensor x_107_split_cast_fp16_0, tensor x_107_split_cast_fp16_1 = split(axis = x_107_split_axis_0, num_splits = x_107_split_num_splits_0, x = input_251_cast_fp16)[name = tensor("x_107_split_cast_fp16")]; - tensor x_107_split_1_sigmoid_cast_fp16 = sigmoid(x = x_107_split_cast_fp16_1)[name = tensor("x_107_split_1_sigmoid_cast_fp16")]; - tensor x_107_cast_fp16 = mul(x = x_107_split_cast_fp16_0, y = x_107_split_1_sigmoid_cast_fp16)[name = tensor("x_107_cast_fp16")]; - tensor input_253_cast_fp16 = select(a = var_11_to_fp16, b = x_107_cast_fp16, cond = var_453)[name = tensor("input_253_cast_fp16")]; - tensor input_255_pad_0 = const()[name = tensor("input_255_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_255_mode_0 = const()[name = tensor("input_255_mode_0"), val = tensor("constant")]; - tensor const_113_to_fp16 = const()[name = tensor("const_113_to_fp16"), val = tensor(0x0p+0)]; - tensor input_255_cast_fp16 = pad(constant_val = const_113_to_fp16, mode = input_255_mode_0, pad = input_255_pad_0, x = input_253_cast_fp16)[name = tensor("input_255_cast_fp16")]; - tensor input_257_pad_type_0 = const()[name = tensor("input_257_pad_type_0"), val = tensor("valid")]; - tensor input_257_groups_0 = const()[name = tensor("input_257_groups_0"), val = tensor(512)]; - tensor input_257_strides_0 = const()[name = tensor("input_257_strides_0"), val = tensor([1])]; - tensor input_257_pad_0 = const()[name = tensor("input_257_pad_0"), val = tensor([0, 0])]; - tensor input_257_dilations_0 = const()[name = tensor("input_257_dilations_0"), val = tensor([1])]; - tensor const_242_to_fp16 = const()[name = tensor("const_242_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60585984)))]; - tensor const_243_to_fp16 = const()[name = tensor("const_243_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60595264)))]; - tensor input_259_cast_fp16 = conv(bias = const_243_to_fp16, dilations = input_257_dilations_0, groups = input_257_groups_0, pad = input_257_pad_0, pad_type = input_257_pad_type_0, strides = input_257_strides_0, weight = const_242_to_fp16, x = input_255_cast_fp16)[name = tensor("input_259_cast_fp16")]; - tensor input_261_cast_fp16 = silu(x = input_259_cast_fp16)[name = tensor("input_261_cast_fp16")]; - tensor x_109_pad_type_0 = const()[name = tensor("x_109_pad_type_0"), val = tensor("valid")]; - tensor x_109_strides_0 = const()[name = tensor("x_109_strides_0"), val = tensor([1])]; - tensor x_109_pad_0 = const()[name = tensor("x_109_pad_0"), val = tensor([0, 0])]; - tensor x_109_dilations_0 = const()[name = tensor("x_109_dilations_0"), val = tensor([1])]; - tensor x_109_groups_0 = const()[name = tensor("x_109_groups_0"), val = tensor(1)]; - tensor module_layers_4_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_4_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60596352)))]; - tensor module_layers_4_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_4_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61120704)))]; - tensor x_109_cast_fp16 = conv(bias = module_layers_4_conv_pointwise_conv2_bias_to_fp16, dilations = x_109_dilations_0, groups = x_109_groups_0, pad = x_109_pad_0, pad_type = x_109_pad_type_0, strides = x_109_strides_0, weight = module_layers_4_conv_pointwise_conv2_weight_to_fp16, x = input_261_cast_fp16)[name = tensor("x_109_cast_fp16")]; - tensor input_263_perm_0 = const()[name = tensor("input_263_perm_0"), val = tensor([0, 2, 1])]; - tensor input_263_cast_fp16 = transpose(perm = input_263_perm_0, x = x_109_cast_fp16)[name = tensor("transpose_170")]; - tensor input_265_cast_fp16 = add(x = input_247_cast_fp16, y = input_263_cast_fp16)[name = tensor("input_265_cast_fp16")]; - tensor input_267_axes_0 = const()[name = tensor("input_267_axes_0"), val = tensor([-1])]; - tensor module_layers_4_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_4_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61121792)))]; - tensor module_layers_4_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_4_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61122880)))]; - tensor input_267_cast_fp16 = layer_norm(axes = input_267_axes_0, beta = module_layers_4_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_4_norm_feed_forward2_weight_to_fp16, x = input_265_cast_fp16)[name = tensor("input_267_cast_fp16")]; - tensor module_layers_4_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_4_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61123968)))]; - tensor module_layers_4_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_4_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(63221184)))]; - tensor linear_44_cast_fp16 = linear(bias = module_layers_4_feed_forward2_linear1_bias_to_fp16, weight = module_layers_4_feed_forward2_linear1_weight_to_fp16, x = input_267_cast_fp16)[name = tensor("linear_44_cast_fp16")]; - tensor input_271_cast_fp16 = silu(x = linear_44_cast_fp16)[name = tensor("input_271_cast_fp16")]; - tensor module_layers_4_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_4_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(63225344)))]; - tensor module_layers_4_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_4_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65322560)))]; - tensor linear_45_cast_fp16 = linear(bias = module_layers_4_feed_forward2_linear2_bias_to_fp16, weight = module_layers_4_feed_forward2_linear2_weight_to_fp16, x = input_271_cast_fp16)[name = tensor("linear_45_cast_fp16")]; - tensor var_1191_to_fp16 = const()[name = tensor("op_1191_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1192_cast_fp16 = mul(x = linear_45_cast_fp16, y = var_1191_to_fp16)[name = tensor("op_1192_cast_fp16")]; - tensor input_277_cast_fp16 = add(x = input_265_cast_fp16, y = var_1192_cast_fp16)[name = tensor("input_277_cast_fp16")]; - tensor input_279_axes_0 = const()[name = tensor("input_279_axes_0"), val = tensor([-1])]; - tensor module_layers_4_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_4_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65323648)))]; - tensor module_layers_4_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_4_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65324736)))]; - tensor input_279_cast_fp16 = layer_norm(axes = input_279_axes_0, beta = module_layers_4_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_4_norm_out_weight_to_fp16, x = input_277_cast_fp16)[name = tensor("input_279_cast_fp16")]; - tensor input_281_axes_0 = const()[name = tensor("input_281_axes_0"), val = tensor([-1])]; - tensor module_layers_5_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_5_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65325824)))]; - tensor module_layers_5_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_5_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65326912)))]; - tensor input_281_cast_fp16 = layer_norm(axes = input_281_axes_0, beta = module_layers_5_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_5_norm_feed_forward1_weight_to_fp16, x = input_279_cast_fp16)[name = tensor("input_281_cast_fp16")]; - tensor module_layers_5_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_5_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65328000)))]; - tensor module_layers_5_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_5_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67425216)))]; - tensor linear_46_cast_fp16 = linear(bias = module_layers_5_feed_forward1_linear1_bias_to_fp16, weight = module_layers_5_feed_forward1_linear1_weight_to_fp16, x = input_281_cast_fp16)[name = tensor("linear_46_cast_fp16")]; - tensor input_285_cast_fp16 = silu(x = linear_46_cast_fp16)[name = tensor("input_285_cast_fp16")]; - tensor module_layers_5_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_5_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67429376)))]; - tensor module_layers_5_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_5_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(69526592)))]; - tensor linear_47_cast_fp16 = linear(bias = module_layers_5_feed_forward1_linear2_bias_to_fp16, weight = module_layers_5_feed_forward1_linear2_weight_to_fp16, x = input_285_cast_fp16)[name = tensor("linear_47_cast_fp16")]; - tensor var_1222_to_fp16 = const()[name = tensor("op_1222_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1223_cast_fp16 = mul(x = linear_47_cast_fp16, y = var_1222_to_fp16)[name = tensor("op_1223_cast_fp16")]; - tensor input_291_cast_fp16 = add(x = input_279_cast_fp16, y = var_1223_cast_fp16)[name = tensor("input_291_cast_fp16")]; - tensor query_11_axes_0 = const()[name = tensor("query_11_axes_0"), val = tensor([-1])]; - tensor module_layers_5_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_5_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(69527680)))]; - tensor module_layers_5_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_5_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(69528768)))]; - tensor query_11_cast_fp16 = layer_norm(axes = query_11_axes_0, beta = module_layers_5_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_5_norm_self_att_weight_to_fp16, x = input_291_cast_fp16)[name = tensor("query_11_cast_fp16")]; - tensor module_layers_5_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_5_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(69529856)))]; - tensor module_layers_5_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_5_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70054208)))]; - tensor linear_48_cast_fp16 = linear(bias = module_layers_5_self_attn_linear_q_bias_to_fp16, weight = module_layers_5_self_attn_linear_q_weight_to_fp16, x = query_11_cast_fp16)[name = tensor("linear_48_cast_fp16")]; - tensor var_1240 = const()[name = tensor("op_1240"), val = tensor([1, -1, 8, 64])]; - tensor q_31_cast_fp16 = reshape(shape = var_1240, x = linear_48_cast_fp16)[name = tensor("q_31_cast_fp16")]; - tensor module_layers_5_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_5_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70055296)))]; - tensor module_layers_5_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_5_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70579648)))]; - tensor linear_49_cast_fp16 = linear(bias = module_layers_5_self_attn_linear_k_bias_to_fp16, weight = module_layers_5_self_attn_linear_k_weight_to_fp16, x = query_11_cast_fp16)[name = tensor("linear_49_cast_fp16")]; - tensor var_1245 = const()[name = tensor("op_1245"), val = tensor([1, -1, 8, 64])]; - tensor k_21_cast_fp16 = reshape(shape = var_1245, x = linear_49_cast_fp16)[name = tensor("k_21_cast_fp16")]; - tensor module_layers_5_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_5_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70580736)))]; - tensor module_layers_5_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_5_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71105088)))]; - tensor linear_50_cast_fp16 = linear(bias = module_layers_5_self_attn_linear_v_bias_to_fp16, weight = module_layers_5_self_attn_linear_v_weight_to_fp16, x = query_11_cast_fp16)[name = tensor("linear_50_cast_fp16")]; - tensor var_1250 = const()[name = tensor("op_1250"), val = tensor([1, -1, 8, 64])]; - tensor v_11_cast_fp16 = reshape(shape = var_1250, x = linear_50_cast_fp16)[name = tensor("v_11_cast_fp16")]; - tensor value_13_perm_0 = const()[name = tensor("value_13_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_5_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_5_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71106176)))]; - tensor var_1262_cast_fp16 = add(x = q_31_cast_fp16, y = module_layers_5_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1262_cast_fp16")]; - tensor module_layers_5_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_5_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71107264)))]; - tensor var_1264_cast_fp16 = add(x = q_31_cast_fp16, y = module_layers_5_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1264_cast_fp16")]; - tensor q_with_bias_v_11_perm_0 = const()[name = tensor("q_with_bias_v_11_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_117_transpose_x_0 = const()[name = tensor("x_117_transpose_x_0"), val = tensor(false)]; - tensor x_117_transpose_y_0 = const()[name = tensor("x_117_transpose_y_0"), val = tensor(false)]; - tensor var_1266_to_fp16 = const()[name = tensor("op_1266_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71108352)))]; - tensor q_with_bias_v_11_cast_fp16 = transpose(perm = q_with_bias_v_11_perm_0, x = var_1264_cast_fp16)[name = tensor("transpose_168")]; - tensor x_117_cast_fp16 = matmul(transpose_x = x_117_transpose_x_0, transpose_y = x_117_transpose_y_0, x = q_with_bias_v_11_cast_fp16, y = var_1266_to_fp16)[name = tensor("x_117_cast_fp16")]; - tensor x_119_pad_0 = const()[name = tensor("x_119_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_119_mode_0 = const()[name = tensor("x_119_mode_0"), val = tensor("constant")]; - tensor const_120_to_fp16 = const()[name = tensor("const_120_to_fp16"), val = tensor(0x0p+0)]; - tensor x_119_cast_fp16 = pad(constant_val = const_120_to_fp16, mode = x_119_mode_0, pad = x_119_pad_0, x = x_117_cast_fp16)[name = tensor("x_119_cast_fp16")]; - tensor var_1274 = const()[name = tensor("op_1274"), val = tensor([1, 8, -1, 188])]; - tensor x_121_cast_fp16 = reshape(shape = var_1274, x = x_119_cast_fp16)[name = tensor("x_121_cast_fp16")]; - tensor var_1278_begin_0 = const()[name = tensor("op_1278_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_1278_end_0 = const()[name = tensor("op_1278_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_1278_end_mask_0 = const()[name = tensor("op_1278_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_1278_cast_fp16 = slice_by_index(begin = var_1278_begin_0, end = var_1278_end_0, end_mask = var_1278_end_mask_0, x = x_121_cast_fp16)[name = tensor("op_1278_cast_fp16")]; - tensor var_1279 = const()[name = tensor("op_1279"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_21_cast_fp16 = reshape(shape = var_1279, x = var_1278_cast_fp16)[name = tensor("matrix_bd_21_cast_fp16")]; - tensor matrix_ac_11_transpose_x_0 = const()[name = tensor("matrix_ac_11_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_11_transpose_y_0 = const()[name = tensor("matrix_ac_11_transpose_y_0"), val = tensor(false)]; - tensor transpose_61_perm_0 = const()[name = tensor("transpose_61_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_62_perm_0 = const()[name = tensor("transpose_62_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_62 = transpose(perm = transpose_62_perm_0, x = k_21_cast_fp16)[name = tensor("transpose_166")]; - tensor transpose_61 = transpose(perm = transpose_61_perm_0, x = var_1262_cast_fp16)[name = tensor("transpose_167")]; - tensor matrix_ac_11_cast_fp16 = matmul(transpose_x = matrix_ac_11_transpose_x_0, transpose_y = matrix_ac_11_transpose_y_0, x = transpose_61, y = transpose_62)[name = tensor("matrix_ac_11_cast_fp16")]; - tensor matrix_bd_23_begin_0 = const()[name = tensor("matrix_bd_23_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_23_end_0 = const()[name = tensor("matrix_bd_23_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_23_end_mask_0 = const()[name = tensor("matrix_bd_23_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_23_cast_fp16 = slice_by_index(begin = matrix_bd_23_begin_0, end = matrix_bd_23_end_0, end_mask = matrix_bd_23_end_mask_0, x = matrix_bd_21_cast_fp16)[name = tensor("matrix_bd_23_cast_fp16")]; - tensor var_1288_cast_fp16 = add(x = matrix_ac_11_cast_fp16, y = matrix_bd_23_cast_fp16)[name = tensor("op_1288_cast_fp16")]; - tensor _inversed_scores_21_y_0_to_fp16 = const()[name = tensor("_inversed_scores_21_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_21_cast_fp16 = mul(x = var_1288_cast_fp16, y = _inversed_scores_21_y_0_to_fp16)[name = tensor("_inversed_scores_21_cast_fp16")]; - tensor scores_23_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_21_cast_fp16, cond = mask_11)[name = tensor("scores_23_cast_fp16")]; - tensor var_1294_cast_fp16 = softmax(axis = var_23, x = scores_23_cast_fp16)[name = tensor("op_1294_cast_fp16")]; - tensor input_293_cast_fp16 = select(a = var_11_to_fp16, b = var_1294_cast_fp16, cond = mask_11)[name = tensor("input_293_cast_fp16")]; - tensor x_123_transpose_x_0 = const()[name = tensor("x_123_transpose_x_0"), val = tensor(false)]; - tensor x_123_transpose_y_0 = const()[name = tensor("x_123_transpose_y_0"), val = tensor(false)]; - tensor value_13_cast_fp16 = transpose(perm = value_13_perm_0, x = v_11_cast_fp16)[name = tensor("transpose_169")]; - tensor x_123_cast_fp16 = matmul(transpose_x = x_123_transpose_x_0, transpose_y = x_123_transpose_y_0, x = input_293_cast_fp16, y = value_13_cast_fp16)[name = tensor("x_123_cast_fp16")]; - tensor var_1298_perm_0 = const()[name = tensor("op_1298_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_1299 = const()[name = tensor("op_1299"), val = tensor([1, -1, 512])]; - tensor var_1298_cast_fp16 = transpose(perm = var_1298_perm_0, x = x_123_cast_fp16)[name = tensor("transpose_165")]; - tensor input_295_cast_fp16 = reshape(shape = var_1299, x = var_1298_cast_fp16)[name = tensor("input_295_cast_fp16")]; - tensor module_layers_5_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_5_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71492416)))]; - tensor module_layers_5_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_5_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(72016768)))]; - tensor linear_52_cast_fp16 = linear(bias = module_layers_5_self_attn_linear_out_bias_to_fp16, weight = module_layers_5_self_attn_linear_out_weight_to_fp16, x = input_295_cast_fp16)[name = tensor("linear_52_cast_fp16")]; - tensor input_299_cast_fp16 = add(x = input_291_cast_fp16, y = linear_52_cast_fp16)[name = tensor("input_299_cast_fp16")]; - tensor x_127_axes_0 = const()[name = tensor("x_127_axes_0"), val = tensor([-1])]; - tensor module_layers_5_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_5_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(72017856)))]; - tensor module_layers_5_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_5_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(72018944)))]; - tensor x_127_cast_fp16 = layer_norm(axes = x_127_axes_0, beta = module_layers_5_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_5_norm_conv_weight_to_fp16, x = input_299_cast_fp16)[name = tensor("x_127_cast_fp16")]; - tensor input_301_perm_0 = const()[name = tensor("input_301_perm_0"), val = tensor([0, 2, 1])]; - tensor input_303_pad_type_0 = const()[name = tensor("input_303_pad_type_0"), val = tensor("valid")]; - tensor input_303_strides_0 = const()[name = tensor("input_303_strides_0"), val = tensor([1])]; - tensor input_303_pad_0 = const()[name = tensor("input_303_pad_0"), val = tensor([0, 0])]; - tensor input_303_dilations_0 = const()[name = tensor("input_303_dilations_0"), val = tensor([1])]; - tensor input_303_groups_0 = const()[name = tensor("input_303_groups_0"), val = tensor(1)]; - tensor module_layers_5_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_5_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(72020032)))]; - tensor module_layers_5_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_5_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73068672)))]; - tensor input_301_cast_fp16 = transpose(perm = input_301_perm_0, x = x_127_cast_fp16)[name = tensor("transpose_164")]; - tensor input_303_cast_fp16 = conv(bias = module_layers_5_conv_pointwise_conv1_bias_to_fp16, dilations = input_303_dilations_0, groups = input_303_groups_0, pad = input_303_pad_0, pad_type = input_303_pad_type_0, strides = input_303_strides_0, weight = module_layers_5_conv_pointwise_conv1_weight_to_fp16, x = input_301_cast_fp16)[name = tensor("input_303_cast_fp16")]; - tensor x_129_split_num_splits_0 = const()[name = tensor("x_129_split_num_splits_0"), val = tensor(2)]; - tensor x_129_split_axis_0 = const()[name = tensor("x_129_split_axis_0"), val = tensor(1)]; - tensor x_129_split_cast_fp16_0, tensor x_129_split_cast_fp16_1 = split(axis = x_129_split_axis_0, num_splits = x_129_split_num_splits_0, x = input_303_cast_fp16)[name = tensor("x_129_split_cast_fp16")]; - tensor x_129_split_1_sigmoid_cast_fp16 = sigmoid(x = x_129_split_cast_fp16_1)[name = tensor("x_129_split_1_sigmoid_cast_fp16")]; - tensor x_129_cast_fp16 = mul(x = x_129_split_cast_fp16_0, y = x_129_split_1_sigmoid_cast_fp16)[name = tensor("x_129_cast_fp16")]; - tensor input_305_cast_fp16 = select(a = var_11_to_fp16, b = x_129_cast_fp16, cond = var_453)[name = tensor("input_305_cast_fp16")]; - tensor input_307_pad_0 = const()[name = tensor("input_307_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_307_mode_0 = const()[name = tensor("input_307_mode_0"), val = tensor("constant")]; - tensor const_123_to_fp16 = const()[name = tensor("const_123_to_fp16"), val = tensor(0x0p+0)]; - tensor input_307_cast_fp16 = pad(constant_val = const_123_to_fp16, mode = input_307_mode_0, pad = input_307_pad_0, x = input_305_cast_fp16)[name = tensor("input_307_cast_fp16")]; - tensor input_309_pad_type_0 = const()[name = tensor("input_309_pad_type_0"), val = tensor("valid")]; - tensor input_309_groups_0 = const()[name = tensor("input_309_groups_0"), val = tensor(512)]; - tensor input_309_strides_0 = const()[name = tensor("input_309_strides_0"), val = tensor([1])]; - tensor input_309_pad_0 = const()[name = tensor("input_309_pad_0"), val = tensor([0, 0])]; - tensor input_309_dilations_0 = const()[name = tensor("input_309_dilations_0"), val = tensor([1])]; - tensor const_244_to_fp16 = const()[name = tensor("const_244_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73070784)))]; - tensor const_245_to_fp16 = const()[name = tensor("const_245_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73080064)))]; - tensor input_311_cast_fp16 = conv(bias = const_245_to_fp16, dilations = input_309_dilations_0, groups = input_309_groups_0, pad = input_309_pad_0, pad_type = input_309_pad_type_0, strides = input_309_strides_0, weight = const_244_to_fp16, x = input_307_cast_fp16)[name = tensor("input_311_cast_fp16")]; - tensor input_313_cast_fp16 = silu(x = input_311_cast_fp16)[name = tensor("input_313_cast_fp16")]; - tensor x_131_pad_type_0 = const()[name = tensor("x_131_pad_type_0"), val = tensor("valid")]; - tensor x_131_strides_0 = const()[name = tensor("x_131_strides_0"), val = tensor([1])]; - tensor x_131_pad_0 = const()[name = tensor("x_131_pad_0"), val = tensor([0, 0])]; - tensor x_131_dilations_0 = const()[name = tensor("x_131_dilations_0"), val = tensor([1])]; - tensor x_131_groups_0 = const()[name = tensor("x_131_groups_0"), val = tensor(1)]; - tensor module_layers_5_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_5_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73081152)))]; - tensor module_layers_5_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_5_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73605504)))]; - tensor x_131_cast_fp16 = conv(bias = module_layers_5_conv_pointwise_conv2_bias_to_fp16, dilations = x_131_dilations_0, groups = x_131_groups_0, pad = x_131_pad_0, pad_type = x_131_pad_type_0, strides = x_131_strides_0, weight = module_layers_5_conv_pointwise_conv2_weight_to_fp16, x = input_313_cast_fp16)[name = tensor("x_131_cast_fp16")]; - tensor input_315_perm_0 = const()[name = tensor("input_315_perm_0"), val = tensor([0, 2, 1])]; - tensor input_315_cast_fp16 = transpose(perm = input_315_perm_0, x = x_131_cast_fp16)[name = tensor("transpose_163")]; - tensor input_317_cast_fp16 = add(x = input_299_cast_fp16, y = input_315_cast_fp16)[name = tensor("input_317_cast_fp16")]; - tensor input_319_axes_0 = const()[name = tensor("input_319_axes_0"), val = tensor([-1])]; - tensor module_layers_5_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_5_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73606592)))]; - tensor module_layers_5_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_5_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73607680)))]; - tensor input_319_cast_fp16 = layer_norm(axes = input_319_axes_0, beta = module_layers_5_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_5_norm_feed_forward2_weight_to_fp16, x = input_317_cast_fp16)[name = tensor("input_319_cast_fp16")]; - tensor module_layers_5_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_5_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73608768)))]; - tensor module_layers_5_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_5_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(75705984)))]; - tensor linear_53_cast_fp16 = linear(bias = module_layers_5_feed_forward2_linear1_bias_to_fp16, weight = module_layers_5_feed_forward2_linear1_weight_to_fp16, x = input_319_cast_fp16)[name = tensor("linear_53_cast_fp16")]; - tensor input_323_cast_fp16 = silu(x = linear_53_cast_fp16)[name = tensor("input_323_cast_fp16")]; - tensor module_layers_5_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_5_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(75710144)))]; - tensor module_layers_5_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_5_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(77807360)))]; - tensor linear_54_cast_fp16 = linear(bias = module_layers_5_feed_forward2_linear2_bias_to_fp16, weight = module_layers_5_feed_forward2_linear2_weight_to_fp16, x = input_323_cast_fp16)[name = tensor("linear_54_cast_fp16")]; - tensor var_1365_to_fp16 = const()[name = tensor("op_1365_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1366_cast_fp16 = mul(x = linear_54_cast_fp16, y = var_1365_to_fp16)[name = tensor("op_1366_cast_fp16")]; - tensor input_329_cast_fp16 = add(x = input_317_cast_fp16, y = var_1366_cast_fp16)[name = tensor("input_329_cast_fp16")]; - tensor input_331_axes_0 = const()[name = tensor("input_331_axes_0"), val = tensor([-1])]; - tensor module_layers_5_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_5_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(77808448)))]; - tensor module_layers_5_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_5_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(77809536)))]; - tensor input_331_cast_fp16 = layer_norm(axes = input_331_axes_0, beta = module_layers_5_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_5_norm_out_weight_to_fp16, x = input_329_cast_fp16)[name = tensor("input_331_cast_fp16")]; - tensor input_333_axes_0 = const()[name = tensor("input_333_axes_0"), val = tensor([-1])]; - tensor module_layers_6_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_6_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(77810624)))]; - tensor module_layers_6_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_6_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(77811712)))]; - tensor input_333_cast_fp16 = layer_norm(axes = input_333_axes_0, beta = module_layers_6_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_6_norm_feed_forward1_weight_to_fp16, x = input_331_cast_fp16)[name = tensor("input_333_cast_fp16")]; - tensor module_layers_6_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_6_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(77812800)))]; - tensor module_layers_6_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_6_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(79910016)))]; - tensor linear_55_cast_fp16 = linear(bias = module_layers_6_feed_forward1_linear1_bias_to_fp16, weight = module_layers_6_feed_forward1_linear1_weight_to_fp16, x = input_333_cast_fp16)[name = tensor("linear_55_cast_fp16")]; - tensor input_337_cast_fp16 = silu(x = linear_55_cast_fp16)[name = tensor("input_337_cast_fp16")]; - tensor module_layers_6_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_6_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(79914176)))]; - tensor module_layers_6_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_6_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82011392)))]; - tensor linear_56_cast_fp16 = linear(bias = module_layers_6_feed_forward1_linear2_bias_to_fp16, weight = module_layers_6_feed_forward1_linear2_weight_to_fp16, x = input_337_cast_fp16)[name = tensor("linear_56_cast_fp16")]; - tensor var_1396_to_fp16 = const()[name = tensor("op_1396_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1397_cast_fp16 = mul(x = linear_56_cast_fp16, y = var_1396_to_fp16)[name = tensor("op_1397_cast_fp16")]; - tensor input_343_cast_fp16 = add(x = input_331_cast_fp16, y = var_1397_cast_fp16)[name = tensor("input_343_cast_fp16")]; - tensor query_13_axes_0 = const()[name = tensor("query_13_axes_0"), val = tensor([-1])]; - tensor module_layers_6_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_6_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82012480)))]; - tensor module_layers_6_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_6_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82013568)))]; - tensor query_13_cast_fp16 = layer_norm(axes = query_13_axes_0, beta = module_layers_6_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_6_norm_self_att_weight_to_fp16, x = input_343_cast_fp16)[name = tensor("query_13_cast_fp16")]; - tensor module_layers_6_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_6_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82014656)))]; - tensor module_layers_6_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_6_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82539008)))]; - tensor linear_57_cast_fp16 = linear(bias = module_layers_6_self_attn_linear_q_bias_to_fp16, weight = module_layers_6_self_attn_linear_q_weight_to_fp16, x = query_13_cast_fp16)[name = tensor("linear_57_cast_fp16")]; - tensor var_1414 = const()[name = tensor("op_1414"), val = tensor([1, -1, 8, 64])]; - tensor q_37_cast_fp16 = reshape(shape = var_1414, x = linear_57_cast_fp16)[name = tensor("q_37_cast_fp16")]; - tensor module_layers_6_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_6_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82540096)))]; - tensor module_layers_6_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_6_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83064448)))]; - tensor linear_58_cast_fp16 = linear(bias = module_layers_6_self_attn_linear_k_bias_to_fp16, weight = module_layers_6_self_attn_linear_k_weight_to_fp16, x = query_13_cast_fp16)[name = tensor("linear_58_cast_fp16")]; - tensor var_1419 = const()[name = tensor("op_1419"), val = tensor([1, -1, 8, 64])]; - tensor k_25_cast_fp16 = reshape(shape = var_1419, x = linear_58_cast_fp16)[name = tensor("k_25_cast_fp16")]; - tensor module_layers_6_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_6_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83065536)))]; - tensor module_layers_6_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_6_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83589888)))]; - tensor linear_59_cast_fp16 = linear(bias = module_layers_6_self_attn_linear_v_bias_to_fp16, weight = module_layers_6_self_attn_linear_v_weight_to_fp16, x = query_13_cast_fp16)[name = tensor("linear_59_cast_fp16")]; - tensor var_1424 = const()[name = tensor("op_1424"), val = tensor([1, -1, 8, 64])]; - tensor v_13_cast_fp16 = reshape(shape = var_1424, x = linear_59_cast_fp16)[name = tensor("v_13_cast_fp16")]; - tensor value_15_perm_0 = const()[name = tensor("value_15_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_6_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_6_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83590976)))]; - tensor var_1436_cast_fp16 = add(x = q_37_cast_fp16, y = module_layers_6_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1436_cast_fp16")]; - tensor module_layers_6_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_6_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83592064)))]; - tensor var_1438_cast_fp16 = add(x = q_37_cast_fp16, y = module_layers_6_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1438_cast_fp16")]; - tensor q_with_bias_v_13_perm_0 = const()[name = tensor("q_with_bias_v_13_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_139_transpose_x_0 = const()[name = tensor("x_139_transpose_x_0"), val = tensor(false)]; - tensor x_139_transpose_y_0 = const()[name = tensor("x_139_transpose_y_0"), val = tensor(false)]; - tensor var_1440_to_fp16 = const()[name = tensor("op_1440_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83593152)))]; - tensor q_with_bias_v_13_cast_fp16 = transpose(perm = q_with_bias_v_13_perm_0, x = var_1438_cast_fp16)[name = tensor("transpose_161")]; - tensor x_139_cast_fp16 = matmul(transpose_x = x_139_transpose_x_0, transpose_y = x_139_transpose_y_0, x = q_with_bias_v_13_cast_fp16, y = var_1440_to_fp16)[name = tensor("x_139_cast_fp16")]; - tensor x_141_pad_0 = const()[name = tensor("x_141_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_141_mode_0 = const()[name = tensor("x_141_mode_0"), val = tensor("constant")]; - tensor const_130_to_fp16 = const()[name = tensor("const_130_to_fp16"), val = tensor(0x0p+0)]; - tensor x_141_cast_fp16 = pad(constant_val = const_130_to_fp16, mode = x_141_mode_0, pad = x_141_pad_0, x = x_139_cast_fp16)[name = tensor("x_141_cast_fp16")]; - tensor var_1448 = const()[name = tensor("op_1448"), val = tensor([1, 8, -1, 188])]; - tensor x_143_cast_fp16 = reshape(shape = var_1448, x = x_141_cast_fp16)[name = tensor("x_143_cast_fp16")]; - tensor var_1452_begin_0 = const()[name = tensor("op_1452_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_1452_end_0 = const()[name = tensor("op_1452_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_1452_end_mask_0 = const()[name = tensor("op_1452_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_1452_cast_fp16 = slice_by_index(begin = var_1452_begin_0, end = var_1452_end_0, end_mask = var_1452_end_mask_0, x = x_143_cast_fp16)[name = tensor("op_1452_cast_fp16")]; - tensor var_1453 = const()[name = tensor("op_1453"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_25_cast_fp16 = reshape(shape = var_1453, x = var_1452_cast_fp16)[name = tensor("matrix_bd_25_cast_fp16")]; - tensor matrix_ac_13_transpose_x_0 = const()[name = tensor("matrix_ac_13_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_13_transpose_y_0 = const()[name = tensor("matrix_ac_13_transpose_y_0"), val = tensor(false)]; - tensor transpose_63_perm_0 = const()[name = tensor("transpose_63_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_64_perm_0 = const()[name = tensor("transpose_64_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_64 = transpose(perm = transpose_64_perm_0, x = k_25_cast_fp16)[name = tensor("transpose_159")]; - tensor transpose_63 = transpose(perm = transpose_63_perm_0, x = var_1436_cast_fp16)[name = tensor("transpose_160")]; - tensor matrix_ac_13_cast_fp16 = matmul(transpose_x = matrix_ac_13_transpose_x_0, transpose_y = matrix_ac_13_transpose_y_0, x = transpose_63, y = transpose_64)[name = tensor("matrix_ac_13_cast_fp16")]; - tensor matrix_bd_27_begin_0 = const()[name = tensor("matrix_bd_27_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_27_end_0 = const()[name = tensor("matrix_bd_27_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_27_end_mask_0 = const()[name = tensor("matrix_bd_27_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_27_cast_fp16 = slice_by_index(begin = matrix_bd_27_begin_0, end = matrix_bd_27_end_0, end_mask = matrix_bd_27_end_mask_0, x = matrix_bd_25_cast_fp16)[name = tensor("matrix_bd_27_cast_fp16")]; - tensor var_1462_cast_fp16 = add(x = matrix_ac_13_cast_fp16, y = matrix_bd_27_cast_fp16)[name = tensor("op_1462_cast_fp16")]; - tensor _inversed_scores_25_y_0_to_fp16 = const()[name = tensor("_inversed_scores_25_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_25_cast_fp16 = mul(x = var_1462_cast_fp16, y = _inversed_scores_25_y_0_to_fp16)[name = tensor("_inversed_scores_25_cast_fp16")]; - tensor scores_27_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_25_cast_fp16, cond = mask_11)[name = tensor("scores_27_cast_fp16")]; - tensor var_1468_cast_fp16 = softmax(axis = var_23, x = scores_27_cast_fp16)[name = tensor("op_1468_cast_fp16")]; - tensor input_345_cast_fp16 = select(a = var_11_to_fp16, b = var_1468_cast_fp16, cond = mask_11)[name = tensor("input_345_cast_fp16")]; - tensor x_145_transpose_x_0 = const()[name = tensor("x_145_transpose_x_0"), val = tensor(false)]; - tensor x_145_transpose_y_0 = const()[name = tensor("x_145_transpose_y_0"), val = tensor(false)]; - tensor value_15_cast_fp16 = transpose(perm = value_15_perm_0, x = v_13_cast_fp16)[name = tensor("transpose_162")]; - tensor x_145_cast_fp16 = matmul(transpose_x = x_145_transpose_x_0, transpose_y = x_145_transpose_y_0, x = input_345_cast_fp16, y = value_15_cast_fp16)[name = tensor("x_145_cast_fp16")]; - tensor var_1472_perm_0 = const()[name = tensor("op_1472_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_1473 = const()[name = tensor("op_1473"), val = tensor([1, -1, 512])]; - tensor var_1472_cast_fp16 = transpose(perm = var_1472_perm_0, x = x_145_cast_fp16)[name = tensor("transpose_158")]; - tensor input_347_cast_fp16 = reshape(shape = var_1473, x = var_1472_cast_fp16)[name = tensor("input_347_cast_fp16")]; - tensor module_layers_6_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_6_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83977216)))]; - tensor module_layers_6_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_6_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84501568)))]; - tensor linear_61_cast_fp16 = linear(bias = module_layers_6_self_attn_linear_out_bias_to_fp16, weight = module_layers_6_self_attn_linear_out_weight_to_fp16, x = input_347_cast_fp16)[name = tensor("linear_61_cast_fp16")]; - tensor input_351_cast_fp16 = add(x = input_343_cast_fp16, y = linear_61_cast_fp16)[name = tensor("input_351_cast_fp16")]; - tensor x_149_axes_0 = const()[name = tensor("x_149_axes_0"), val = tensor([-1])]; - tensor module_layers_6_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_6_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84502656)))]; - tensor module_layers_6_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_6_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84503744)))]; - tensor x_149_cast_fp16 = layer_norm(axes = x_149_axes_0, beta = module_layers_6_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_6_norm_conv_weight_to_fp16, x = input_351_cast_fp16)[name = tensor("x_149_cast_fp16")]; - tensor input_353_perm_0 = const()[name = tensor("input_353_perm_0"), val = tensor([0, 2, 1])]; - tensor input_355_pad_type_0 = const()[name = tensor("input_355_pad_type_0"), val = tensor("valid")]; - tensor input_355_strides_0 = const()[name = tensor("input_355_strides_0"), val = tensor([1])]; - tensor input_355_pad_0 = const()[name = tensor("input_355_pad_0"), val = tensor([0, 0])]; - tensor input_355_dilations_0 = const()[name = tensor("input_355_dilations_0"), val = tensor([1])]; - tensor input_355_groups_0 = const()[name = tensor("input_355_groups_0"), val = tensor(1)]; - tensor module_layers_6_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_6_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84504832)))]; - tensor module_layers_6_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_6_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85553472)))]; - tensor input_353_cast_fp16 = transpose(perm = input_353_perm_0, x = x_149_cast_fp16)[name = tensor("transpose_157")]; - tensor input_355_cast_fp16 = conv(bias = module_layers_6_conv_pointwise_conv1_bias_to_fp16, dilations = input_355_dilations_0, groups = input_355_groups_0, pad = input_355_pad_0, pad_type = input_355_pad_type_0, strides = input_355_strides_0, weight = module_layers_6_conv_pointwise_conv1_weight_to_fp16, x = input_353_cast_fp16)[name = tensor("input_355_cast_fp16")]; - tensor x_151_split_num_splits_0 = const()[name = tensor("x_151_split_num_splits_0"), val = tensor(2)]; - tensor x_151_split_axis_0 = const()[name = tensor("x_151_split_axis_0"), val = tensor(1)]; - tensor x_151_split_cast_fp16_0, tensor x_151_split_cast_fp16_1 = split(axis = x_151_split_axis_0, num_splits = x_151_split_num_splits_0, x = input_355_cast_fp16)[name = tensor("x_151_split_cast_fp16")]; - tensor x_151_split_1_sigmoid_cast_fp16 = sigmoid(x = x_151_split_cast_fp16_1)[name = tensor("x_151_split_1_sigmoid_cast_fp16")]; - tensor x_151_cast_fp16 = mul(x = x_151_split_cast_fp16_0, y = x_151_split_1_sigmoid_cast_fp16)[name = tensor("x_151_cast_fp16")]; - tensor input_357_cast_fp16 = select(a = var_11_to_fp16, b = x_151_cast_fp16, cond = var_453)[name = tensor("input_357_cast_fp16")]; - tensor input_359_pad_0 = const()[name = tensor("input_359_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_359_mode_0 = const()[name = tensor("input_359_mode_0"), val = tensor("constant")]; - tensor const_133_to_fp16 = const()[name = tensor("const_133_to_fp16"), val = tensor(0x0p+0)]; - tensor input_359_cast_fp16 = pad(constant_val = const_133_to_fp16, mode = input_359_mode_0, pad = input_359_pad_0, x = input_357_cast_fp16)[name = tensor("input_359_cast_fp16")]; - tensor input_361_pad_type_0 = const()[name = tensor("input_361_pad_type_0"), val = tensor("valid")]; - tensor input_361_groups_0 = const()[name = tensor("input_361_groups_0"), val = tensor(512)]; - tensor input_361_strides_0 = const()[name = tensor("input_361_strides_0"), val = tensor([1])]; - tensor input_361_pad_0 = const()[name = tensor("input_361_pad_0"), val = tensor([0, 0])]; - tensor input_361_dilations_0 = const()[name = tensor("input_361_dilations_0"), val = tensor([1])]; - tensor const_246_to_fp16 = const()[name = tensor("const_246_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85555584)))]; - tensor const_247_to_fp16 = const()[name = tensor("const_247_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85564864)))]; - tensor input_363_cast_fp16 = conv(bias = const_247_to_fp16, dilations = input_361_dilations_0, groups = input_361_groups_0, pad = input_361_pad_0, pad_type = input_361_pad_type_0, strides = input_361_strides_0, weight = const_246_to_fp16, x = input_359_cast_fp16)[name = tensor("input_363_cast_fp16")]; - tensor input_365_cast_fp16 = silu(x = input_363_cast_fp16)[name = tensor("input_365_cast_fp16")]; - tensor x_153_pad_type_0 = const()[name = tensor("x_153_pad_type_0"), val = tensor("valid")]; - tensor x_153_strides_0 = const()[name = tensor("x_153_strides_0"), val = tensor([1])]; - tensor x_153_pad_0 = const()[name = tensor("x_153_pad_0"), val = tensor([0, 0])]; - tensor x_153_dilations_0 = const()[name = tensor("x_153_dilations_0"), val = tensor([1])]; - tensor x_153_groups_0 = const()[name = tensor("x_153_groups_0"), val = tensor(1)]; - tensor module_layers_6_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_6_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85565952)))]; - tensor module_layers_6_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_6_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86090304)))]; - tensor x_153_cast_fp16 = conv(bias = module_layers_6_conv_pointwise_conv2_bias_to_fp16, dilations = x_153_dilations_0, groups = x_153_groups_0, pad = x_153_pad_0, pad_type = x_153_pad_type_0, strides = x_153_strides_0, weight = module_layers_6_conv_pointwise_conv2_weight_to_fp16, x = input_365_cast_fp16)[name = tensor("x_153_cast_fp16")]; - tensor input_367_perm_0 = const()[name = tensor("input_367_perm_0"), val = tensor([0, 2, 1])]; - tensor input_367_cast_fp16 = transpose(perm = input_367_perm_0, x = x_153_cast_fp16)[name = tensor("transpose_156")]; - tensor input_369_cast_fp16 = add(x = input_351_cast_fp16, y = input_367_cast_fp16)[name = tensor("input_369_cast_fp16")]; - tensor input_371_axes_0 = const()[name = tensor("input_371_axes_0"), val = tensor([-1])]; - tensor module_layers_6_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_6_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86091392)))]; - tensor module_layers_6_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_6_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86092480)))]; - tensor input_371_cast_fp16 = layer_norm(axes = input_371_axes_0, beta = module_layers_6_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_6_norm_feed_forward2_weight_to_fp16, x = input_369_cast_fp16)[name = tensor("input_371_cast_fp16")]; - tensor module_layers_6_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_6_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86093568)))]; - tensor module_layers_6_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_6_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(88190784)))]; - tensor linear_62_cast_fp16 = linear(bias = module_layers_6_feed_forward2_linear1_bias_to_fp16, weight = module_layers_6_feed_forward2_linear1_weight_to_fp16, x = input_371_cast_fp16)[name = tensor("linear_62_cast_fp16")]; - tensor input_375_cast_fp16 = silu(x = linear_62_cast_fp16)[name = tensor("input_375_cast_fp16")]; - tensor module_layers_6_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_6_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(88194944)))]; - tensor module_layers_6_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_6_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90292160)))]; - tensor linear_63_cast_fp16 = linear(bias = module_layers_6_feed_forward2_linear2_bias_to_fp16, weight = module_layers_6_feed_forward2_linear2_weight_to_fp16, x = input_375_cast_fp16)[name = tensor("linear_63_cast_fp16")]; - tensor var_1539_to_fp16 = const()[name = tensor("op_1539_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1540_cast_fp16 = mul(x = linear_63_cast_fp16, y = var_1539_to_fp16)[name = tensor("op_1540_cast_fp16")]; - tensor input_381_cast_fp16 = add(x = input_369_cast_fp16, y = var_1540_cast_fp16)[name = tensor("input_381_cast_fp16")]; - tensor input_383_axes_0 = const()[name = tensor("input_383_axes_0"), val = tensor([-1])]; - tensor module_layers_6_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_6_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90293248)))]; - tensor module_layers_6_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_6_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90294336)))]; - tensor input_383_cast_fp16 = layer_norm(axes = input_383_axes_0, beta = module_layers_6_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_6_norm_out_weight_to_fp16, x = input_381_cast_fp16)[name = tensor("input_383_cast_fp16")]; - tensor input_385_axes_0 = const()[name = tensor("input_385_axes_0"), val = tensor([-1])]; - tensor module_layers_7_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_7_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90295424)))]; - tensor module_layers_7_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_7_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90296512)))]; - tensor input_385_cast_fp16 = layer_norm(axes = input_385_axes_0, beta = module_layers_7_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_7_norm_feed_forward1_weight_to_fp16, x = input_383_cast_fp16)[name = tensor("input_385_cast_fp16")]; - tensor module_layers_7_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_7_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90297600)))]; - tensor module_layers_7_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_7_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92394816)))]; - tensor linear_64_cast_fp16 = linear(bias = module_layers_7_feed_forward1_linear1_bias_to_fp16, weight = module_layers_7_feed_forward1_linear1_weight_to_fp16, x = input_385_cast_fp16)[name = tensor("linear_64_cast_fp16")]; - tensor input_389_cast_fp16 = silu(x = linear_64_cast_fp16)[name = tensor("input_389_cast_fp16")]; - tensor module_layers_7_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_7_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92398976)))]; - tensor module_layers_7_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_7_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(94496192)))]; - tensor linear_65_cast_fp16 = linear(bias = module_layers_7_feed_forward1_linear2_bias_to_fp16, weight = module_layers_7_feed_forward1_linear2_weight_to_fp16, x = input_389_cast_fp16)[name = tensor("linear_65_cast_fp16")]; - tensor var_1570_to_fp16 = const()[name = tensor("op_1570_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1571_cast_fp16 = mul(x = linear_65_cast_fp16, y = var_1570_to_fp16)[name = tensor("op_1571_cast_fp16")]; - tensor input_395_cast_fp16 = add(x = input_383_cast_fp16, y = var_1571_cast_fp16)[name = tensor("input_395_cast_fp16")]; - tensor query_15_axes_0 = const()[name = tensor("query_15_axes_0"), val = tensor([-1])]; - tensor module_layers_7_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_7_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(94497280)))]; - tensor module_layers_7_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_7_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(94498368)))]; - tensor query_15_cast_fp16 = layer_norm(axes = query_15_axes_0, beta = module_layers_7_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_7_norm_self_att_weight_to_fp16, x = input_395_cast_fp16)[name = tensor("query_15_cast_fp16")]; - tensor module_layers_7_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_7_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(94499456)))]; - tensor module_layers_7_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_7_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95023808)))]; - tensor linear_66_cast_fp16 = linear(bias = module_layers_7_self_attn_linear_q_bias_to_fp16, weight = module_layers_7_self_attn_linear_q_weight_to_fp16, x = query_15_cast_fp16)[name = tensor("linear_66_cast_fp16")]; - tensor var_1588 = const()[name = tensor("op_1588"), val = tensor([1, -1, 8, 64])]; - tensor q_43_cast_fp16 = reshape(shape = var_1588, x = linear_66_cast_fp16)[name = tensor("q_43_cast_fp16")]; - tensor module_layers_7_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_7_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95024896)))]; - tensor module_layers_7_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_7_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95549248)))]; - tensor linear_67_cast_fp16 = linear(bias = module_layers_7_self_attn_linear_k_bias_to_fp16, weight = module_layers_7_self_attn_linear_k_weight_to_fp16, x = query_15_cast_fp16)[name = tensor("linear_67_cast_fp16")]; - tensor var_1593 = const()[name = tensor("op_1593"), val = tensor([1, -1, 8, 64])]; - tensor k_29_cast_fp16 = reshape(shape = var_1593, x = linear_67_cast_fp16)[name = tensor("k_29_cast_fp16")]; - tensor module_layers_7_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_7_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95550336)))]; - tensor module_layers_7_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_7_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96074688)))]; - tensor linear_68_cast_fp16 = linear(bias = module_layers_7_self_attn_linear_v_bias_to_fp16, weight = module_layers_7_self_attn_linear_v_weight_to_fp16, x = query_15_cast_fp16)[name = tensor("linear_68_cast_fp16")]; - tensor var_1598 = const()[name = tensor("op_1598"), val = tensor([1, -1, 8, 64])]; - tensor v_15_cast_fp16 = reshape(shape = var_1598, x = linear_68_cast_fp16)[name = tensor("v_15_cast_fp16")]; - tensor value_17_perm_0 = const()[name = tensor("value_17_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_7_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_7_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96075776)))]; - tensor var_1610_cast_fp16 = add(x = q_43_cast_fp16, y = module_layers_7_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1610_cast_fp16")]; - tensor module_layers_7_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_7_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96076864)))]; - tensor var_1612_cast_fp16 = add(x = q_43_cast_fp16, y = module_layers_7_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1612_cast_fp16")]; - tensor q_with_bias_v_15_perm_0 = const()[name = tensor("q_with_bias_v_15_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_161_transpose_x_0 = const()[name = tensor("x_161_transpose_x_0"), val = tensor(false)]; - tensor x_161_transpose_y_0 = const()[name = tensor("x_161_transpose_y_0"), val = tensor(false)]; - tensor var_1614_to_fp16 = const()[name = tensor("op_1614_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96077952)))]; - tensor q_with_bias_v_15_cast_fp16 = transpose(perm = q_with_bias_v_15_perm_0, x = var_1612_cast_fp16)[name = tensor("transpose_154")]; - tensor x_161_cast_fp16 = matmul(transpose_x = x_161_transpose_x_0, transpose_y = x_161_transpose_y_0, x = q_with_bias_v_15_cast_fp16, y = var_1614_to_fp16)[name = tensor("x_161_cast_fp16")]; - tensor x_163_pad_0 = const()[name = tensor("x_163_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_163_mode_0 = const()[name = tensor("x_163_mode_0"), val = tensor("constant")]; - tensor const_140_to_fp16 = const()[name = tensor("const_140_to_fp16"), val = tensor(0x0p+0)]; - tensor x_163_cast_fp16 = pad(constant_val = const_140_to_fp16, mode = x_163_mode_0, pad = x_163_pad_0, x = x_161_cast_fp16)[name = tensor("x_163_cast_fp16")]; - tensor var_1622 = const()[name = tensor("op_1622"), val = tensor([1, 8, -1, 188])]; - tensor x_165_cast_fp16 = reshape(shape = var_1622, x = x_163_cast_fp16)[name = tensor("x_165_cast_fp16")]; - tensor var_1626_begin_0 = const()[name = tensor("op_1626_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_1626_end_0 = const()[name = tensor("op_1626_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_1626_end_mask_0 = const()[name = tensor("op_1626_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_1626_cast_fp16 = slice_by_index(begin = var_1626_begin_0, end = var_1626_end_0, end_mask = var_1626_end_mask_0, x = x_165_cast_fp16)[name = tensor("op_1626_cast_fp16")]; - tensor var_1627 = const()[name = tensor("op_1627"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_29_cast_fp16 = reshape(shape = var_1627, x = var_1626_cast_fp16)[name = tensor("matrix_bd_29_cast_fp16")]; - tensor matrix_ac_15_transpose_x_0 = const()[name = tensor("matrix_ac_15_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_15_transpose_y_0 = const()[name = tensor("matrix_ac_15_transpose_y_0"), val = tensor(false)]; - tensor transpose_65_perm_0 = const()[name = tensor("transpose_65_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_66_perm_0 = const()[name = tensor("transpose_66_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_66 = transpose(perm = transpose_66_perm_0, x = k_29_cast_fp16)[name = tensor("transpose_152")]; - tensor transpose_65 = transpose(perm = transpose_65_perm_0, x = var_1610_cast_fp16)[name = tensor("transpose_153")]; - tensor matrix_ac_15_cast_fp16 = matmul(transpose_x = matrix_ac_15_transpose_x_0, transpose_y = matrix_ac_15_transpose_y_0, x = transpose_65, y = transpose_66)[name = tensor("matrix_ac_15_cast_fp16")]; - tensor matrix_bd_31_begin_0 = const()[name = tensor("matrix_bd_31_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_31_end_0 = const()[name = tensor("matrix_bd_31_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_31_end_mask_0 = const()[name = tensor("matrix_bd_31_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_31_cast_fp16 = slice_by_index(begin = matrix_bd_31_begin_0, end = matrix_bd_31_end_0, end_mask = matrix_bd_31_end_mask_0, x = matrix_bd_29_cast_fp16)[name = tensor("matrix_bd_31_cast_fp16")]; - tensor var_1636_cast_fp16 = add(x = matrix_ac_15_cast_fp16, y = matrix_bd_31_cast_fp16)[name = tensor("op_1636_cast_fp16")]; - tensor _inversed_scores_29_y_0_to_fp16 = const()[name = tensor("_inversed_scores_29_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_29_cast_fp16 = mul(x = var_1636_cast_fp16, y = _inversed_scores_29_y_0_to_fp16)[name = tensor("_inversed_scores_29_cast_fp16")]; - tensor scores_31_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_29_cast_fp16, cond = mask_11)[name = tensor("scores_31_cast_fp16")]; - tensor var_1642_cast_fp16 = softmax(axis = var_23, x = scores_31_cast_fp16)[name = tensor("op_1642_cast_fp16")]; - tensor input_397_cast_fp16 = select(a = var_11_to_fp16, b = var_1642_cast_fp16, cond = mask_11)[name = tensor("input_397_cast_fp16")]; - tensor x_167_transpose_x_0 = const()[name = tensor("x_167_transpose_x_0"), val = tensor(false)]; - tensor x_167_transpose_y_0 = const()[name = tensor("x_167_transpose_y_0"), val = tensor(false)]; - tensor value_17_cast_fp16 = transpose(perm = value_17_perm_0, x = v_15_cast_fp16)[name = tensor("transpose_155")]; - tensor x_167_cast_fp16 = matmul(transpose_x = x_167_transpose_x_0, transpose_y = x_167_transpose_y_0, x = input_397_cast_fp16, y = value_17_cast_fp16)[name = tensor("x_167_cast_fp16")]; - tensor var_1646_perm_0 = const()[name = tensor("op_1646_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_1647 = const()[name = tensor("op_1647"), val = tensor([1, -1, 512])]; - tensor var_1646_cast_fp16 = transpose(perm = var_1646_perm_0, x = x_167_cast_fp16)[name = tensor("transpose_151")]; - tensor input_399_cast_fp16 = reshape(shape = var_1647, x = var_1646_cast_fp16)[name = tensor("input_399_cast_fp16")]; - tensor module_layers_7_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_7_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96462016)))]; - tensor module_layers_7_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_7_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96986368)))]; - tensor linear_70_cast_fp16 = linear(bias = module_layers_7_self_attn_linear_out_bias_to_fp16, weight = module_layers_7_self_attn_linear_out_weight_to_fp16, x = input_399_cast_fp16)[name = tensor("linear_70_cast_fp16")]; - tensor input_403_cast_fp16 = add(x = input_395_cast_fp16, y = linear_70_cast_fp16)[name = tensor("input_403_cast_fp16")]; - tensor x_171_axes_0 = const()[name = tensor("x_171_axes_0"), val = tensor([-1])]; - tensor module_layers_7_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_7_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96987456)))]; - tensor module_layers_7_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_7_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96988544)))]; - tensor x_171_cast_fp16 = layer_norm(axes = x_171_axes_0, beta = module_layers_7_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_7_norm_conv_weight_to_fp16, x = input_403_cast_fp16)[name = tensor("x_171_cast_fp16")]; - tensor input_405_perm_0 = const()[name = tensor("input_405_perm_0"), val = tensor([0, 2, 1])]; - tensor input_407_pad_type_0 = const()[name = tensor("input_407_pad_type_0"), val = tensor("valid")]; - tensor input_407_strides_0 = const()[name = tensor("input_407_strides_0"), val = tensor([1])]; - tensor input_407_pad_0 = const()[name = tensor("input_407_pad_0"), val = tensor([0, 0])]; - tensor input_407_dilations_0 = const()[name = tensor("input_407_dilations_0"), val = tensor([1])]; - tensor input_407_groups_0 = const()[name = tensor("input_407_groups_0"), val = tensor(1)]; - tensor module_layers_7_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_7_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96989632)))]; - tensor module_layers_7_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_7_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98038272)))]; - tensor input_405_cast_fp16 = transpose(perm = input_405_perm_0, x = x_171_cast_fp16)[name = tensor("transpose_150")]; - tensor input_407_cast_fp16 = conv(bias = module_layers_7_conv_pointwise_conv1_bias_to_fp16, dilations = input_407_dilations_0, groups = input_407_groups_0, pad = input_407_pad_0, pad_type = input_407_pad_type_0, strides = input_407_strides_0, weight = module_layers_7_conv_pointwise_conv1_weight_to_fp16, x = input_405_cast_fp16)[name = tensor("input_407_cast_fp16")]; - tensor x_173_split_num_splits_0 = const()[name = tensor("x_173_split_num_splits_0"), val = tensor(2)]; - tensor x_173_split_axis_0 = const()[name = tensor("x_173_split_axis_0"), val = tensor(1)]; - tensor x_173_split_cast_fp16_0, tensor x_173_split_cast_fp16_1 = split(axis = x_173_split_axis_0, num_splits = x_173_split_num_splits_0, x = input_407_cast_fp16)[name = tensor("x_173_split_cast_fp16")]; - tensor x_173_split_1_sigmoid_cast_fp16 = sigmoid(x = x_173_split_cast_fp16_1)[name = tensor("x_173_split_1_sigmoid_cast_fp16")]; - tensor x_173_cast_fp16 = mul(x = x_173_split_cast_fp16_0, y = x_173_split_1_sigmoid_cast_fp16)[name = tensor("x_173_cast_fp16")]; - tensor input_409_cast_fp16 = select(a = var_11_to_fp16, b = x_173_cast_fp16, cond = var_453)[name = tensor("input_409_cast_fp16")]; - tensor input_411_pad_0 = const()[name = tensor("input_411_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_411_mode_0 = const()[name = tensor("input_411_mode_0"), val = tensor("constant")]; - tensor const_143_to_fp16 = const()[name = tensor("const_143_to_fp16"), val = tensor(0x0p+0)]; - tensor input_411_cast_fp16 = pad(constant_val = const_143_to_fp16, mode = input_411_mode_0, pad = input_411_pad_0, x = input_409_cast_fp16)[name = tensor("input_411_cast_fp16")]; - tensor input_413_pad_type_0 = const()[name = tensor("input_413_pad_type_0"), val = tensor("valid")]; - tensor input_413_groups_0 = const()[name = tensor("input_413_groups_0"), val = tensor(512)]; - tensor input_413_strides_0 = const()[name = tensor("input_413_strides_0"), val = tensor([1])]; - tensor input_413_pad_0 = const()[name = tensor("input_413_pad_0"), val = tensor([0, 0])]; - tensor input_413_dilations_0 = const()[name = tensor("input_413_dilations_0"), val = tensor([1])]; - tensor const_248_to_fp16 = const()[name = tensor("const_248_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98040384)))]; - tensor const_249_to_fp16 = const()[name = tensor("const_249_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98049664)))]; - tensor input_415_cast_fp16 = conv(bias = const_249_to_fp16, dilations = input_413_dilations_0, groups = input_413_groups_0, pad = input_413_pad_0, pad_type = input_413_pad_type_0, strides = input_413_strides_0, weight = const_248_to_fp16, x = input_411_cast_fp16)[name = tensor("input_415_cast_fp16")]; - tensor input_417_cast_fp16 = silu(x = input_415_cast_fp16)[name = tensor("input_417_cast_fp16")]; - tensor x_175_pad_type_0 = const()[name = tensor("x_175_pad_type_0"), val = tensor("valid")]; - tensor x_175_strides_0 = const()[name = tensor("x_175_strides_0"), val = tensor([1])]; - tensor x_175_pad_0 = const()[name = tensor("x_175_pad_0"), val = tensor([0, 0])]; - tensor x_175_dilations_0 = const()[name = tensor("x_175_dilations_0"), val = tensor([1])]; - tensor x_175_groups_0 = const()[name = tensor("x_175_groups_0"), val = tensor(1)]; - tensor module_layers_7_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_7_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98050752)))]; - tensor module_layers_7_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_7_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98575104)))]; - tensor x_175_cast_fp16 = conv(bias = module_layers_7_conv_pointwise_conv2_bias_to_fp16, dilations = x_175_dilations_0, groups = x_175_groups_0, pad = x_175_pad_0, pad_type = x_175_pad_type_0, strides = x_175_strides_0, weight = module_layers_7_conv_pointwise_conv2_weight_to_fp16, x = input_417_cast_fp16)[name = tensor("x_175_cast_fp16")]; - tensor input_419_perm_0 = const()[name = tensor("input_419_perm_0"), val = tensor([0, 2, 1])]; - tensor input_419_cast_fp16 = transpose(perm = input_419_perm_0, x = x_175_cast_fp16)[name = tensor("transpose_149")]; - tensor input_421_cast_fp16 = add(x = input_403_cast_fp16, y = input_419_cast_fp16)[name = tensor("input_421_cast_fp16")]; - tensor input_423_axes_0 = const()[name = tensor("input_423_axes_0"), val = tensor([-1])]; - tensor module_layers_7_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_7_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98576192)))]; - tensor module_layers_7_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_7_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98577280)))]; - tensor input_423_cast_fp16 = layer_norm(axes = input_423_axes_0, beta = module_layers_7_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_7_norm_feed_forward2_weight_to_fp16, x = input_421_cast_fp16)[name = tensor("input_423_cast_fp16")]; - tensor module_layers_7_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_7_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98578368)))]; - tensor module_layers_7_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_7_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(100675584)))]; - tensor linear_71_cast_fp16 = linear(bias = module_layers_7_feed_forward2_linear1_bias_to_fp16, weight = module_layers_7_feed_forward2_linear1_weight_to_fp16, x = input_423_cast_fp16)[name = tensor("linear_71_cast_fp16")]; - tensor input_427_cast_fp16 = silu(x = linear_71_cast_fp16)[name = tensor("input_427_cast_fp16")]; - tensor module_layers_7_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_7_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(100679744)))]; - tensor module_layers_7_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_7_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102776960)))]; - tensor linear_72_cast_fp16 = linear(bias = module_layers_7_feed_forward2_linear2_bias_to_fp16, weight = module_layers_7_feed_forward2_linear2_weight_to_fp16, x = input_427_cast_fp16)[name = tensor("linear_72_cast_fp16")]; - tensor var_1713_to_fp16 = const()[name = tensor("op_1713_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1714_cast_fp16 = mul(x = linear_72_cast_fp16, y = var_1713_to_fp16)[name = tensor("op_1714_cast_fp16")]; - tensor input_433_cast_fp16 = add(x = input_421_cast_fp16, y = var_1714_cast_fp16)[name = tensor("input_433_cast_fp16")]; - tensor input_435_axes_0 = const()[name = tensor("input_435_axes_0"), val = tensor([-1])]; - tensor module_layers_7_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_7_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102778048)))]; - tensor module_layers_7_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_7_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102779136)))]; - tensor input_435_cast_fp16 = layer_norm(axes = input_435_axes_0, beta = module_layers_7_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_7_norm_out_weight_to_fp16, x = input_433_cast_fp16)[name = tensor("input_435_cast_fp16")]; - tensor input_437_axes_0 = const()[name = tensor("input_437_axes_0"), val = tensor([-1])]; - tensor module_layers_8_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_8_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102780224)))]; - tensor module_layers_8_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_8_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102781312)))]; - tensor input_437_cast_fp16 = layer_norm(axes = input_437_axes_0, beta = module_layers_8_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_8_norm_feed_forward1_weight_to_fp16, x = input_435_cast_fp16)[name = tensor("input_437_cast_fp16")]; - tensor module_layers_8_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_8_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102782400)))]; - tensor module_layers_8_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_8_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(104879616)))]; - tensor linear_73_cast_fp16 = linear(bias = module_layers_8_feed_forward1_linear1_bias_to_fp16, weight = module_layers_8_feed_forward1_linear1_weight_to_fp16, x = input_437_cast_fp16)[name = tensor("linear_73_cast_fp16")]; - tensor input_441_cast_fp16 = silu(x = linear_73_cast_fp16)[name = tensor("input_441_cast_fp16")]; - tensor module_layers_8_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_8_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(104883776)))]; - tensor module_layers_8_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_8_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(106980992)))]; - tensor linear_74_cast_fp16 = linear(bias = module_layers_8_feed_forward1_linear2_bias_to_fp16, weight = module_layers_8_feed_forward1_linear2_weight_to_fp16, x = input_441_cast_fp16)[name = tensor("linear_74_cast_fp16")]; - tensor var_1744_to_fp16 = const()[name = tensor("op_1744_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1745_cast_fp16 = mul(x = linear_74_cast_fp16, y = var_1744_to_fp16)[name = tensor("op_1745_cast_fp16")]; - tensor input_447_cast_fp16 = add(x = input_435_cast_fp16, y = var_1745_cast_fp16)[name = tensor("input_447_cast_fp16")]; - tensor query_17_axes_0 = const()[name = tensor("query_17_axes_0"), val = tensor([-1])]; - tensor module_layers_8_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_8_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(106982080)))]; - tensor module_layers_8_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_8_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(106983168)))]; - tensor query_17_cast_fp16 = layer_norm(axes = query_17_axes_0, beta = module_layers_8_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_8_norm_self_att_weight_to_fp16, x = input_447_cast_fp16)[name = tensor("query_17_cast_fp16")]; - tensor module_layers_8_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_8_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(106984256)))]; - tensor module_layers_8_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_8_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(107508608)))]; - tensor linear_75_cast_fp16 = linear(bias = module_layers_8_self_attn_linear_q_bias_to_fp16, weight = module_layers_8_self_attn_linear_q_weight_to_fp16, x = query_17_cast_fp16)[name = tensor("linear_75_cast_fp16")]; - tensor var_1762 = const()[name = tensor("op_1762"), val = tensor([1, -1, 8, 64])]; - tensor q_49_cast_fp16 = reshape(shape = var_1762, x = linear_75_cast_fp16)[name = tensor("q_49_cast_fp16")]; - tensor module_layers_8_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_8_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(107509696)))]; - tensor module_layers_8_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_8_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108034048)))]; - tensor linear_76_cast_fp16 = linear(bias = module_layers_8_self_attn_linear_k_bias_to_fp16, weight = module_layers_8_self_attn_linear_k_weight_to_fp16, x = query_17_cast_fp16)[name = tensor("linear_76_cast_fp16")]; - tensor var_1767 = const()[name = tensor("op_1767"), val = tensor([1, -1, 8, 64])]; - tensor k_33_cast_fp16 = reshape(shape = var_1767, x = linear_76_cast_fp16)[name = tensor("k_33_cast_fp16")]; - tensor module_layers_8_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_8_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108035136)))]; - tensor module_layers_8_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_8_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108559488)))]; - tensor linear_77_cast_fp16 = linear(bias = module_layers_8_self_attn_linear_v_bias_to_fp16, weight = module_layers_8_self_attn_linear_v_weight_to_fp16, x = query_17_cast_fp16)[name = tensor("linear_77_cast_fp16")]; - tensor var_1772 = const()[name = tensor("op_1772"), val = tensor([1, -1, 8, 64])]; - tensor v_17_cast_fp16 = reshape(shape = var_1772, x = linear_77_cast_fp16)[name = tensor("v_17_cast_fp16")]; - tensor value_19_perm_0 = const()[name = tensor("value_19_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_8_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_8_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108560576)))]; - tensor var_1784_cast_fp16 = add(x = q_49_cast_fp16, y = module_layers_8_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1784_cast_fp16")]; - tensor module_layers_8_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_8_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108561664)))]; - tensor var_1786_cast_fp16 = add(x = q_49_cast_fp16, y = module_layers_8_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1786_cast_fp16")]; - tensor q_with_bias_v_17_perm_0 = const()[name = tensor("q_with_bias_v_17_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_183_transpose_x_0 = const()[name = tensor("x_183_transpose_x_0"), val = tensor(false)]; - tensor x_183_transpose_y_0 = const()[name = tensor("x_183_transpose_y_0"), val = tensor(false)]; - tensor var_1788_to_fp16 = const()[name = tensor("op_1788_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108562752)))]; - tensor q_with_bias_v_17_cast_fp16 = transpose(perm = q_with_bias_v_17_perm_0, x = var_1786_cast_fp16)[name = tensor("transpose_147")]; - tensor x_183_cast_fp16 = matmul(transpose_x = x_183_transpose_x_0, transpose_y = x_183_transpose_y_0, x = q_with_bias_v_17_cast_fp16, y = var_1788_to_fp16)[name = tensor("x_183_cast_fp16")]; - tensor x_185_pad_0 = const()[name = tensor("x_185_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_185_mode_0 = const()[name = tensor("x_185_mode_0"), val = tensor("constant")]; - tensor const_150_to_fp16 = const()[name = tensor("const_150_to_fp16"), val = tensor(0x0p+0)]; - tensor x_185_cast_fp16 = pad(constant_val = const_150_to_fp16, mode = x_185_mode_0, pad = x_185_pad_0, x = x_183_cast_fp16)[name = tensor("x_185_cast_fp16")]; - tensor var_1796 = const()[name = tensor("op_1796"), val = tensor([1, 8, -1, 188])]; - tensor x_187_cast_fp16 = reshape(shape = var_1796, x = x_185_cast_fp16)[name = tensor("x_187_cast_fp16")]; - tensor var_1800_begin_0 = const()[name = tensor("op_1800_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_1800_end_0 = const()[name = tensor("op_1800_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_1800_end_mask_0 = const()[name = tensor("op_1800_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_1800_cast_fp16 = slice_by_index(begin = var_1800_begin_0, end = var_1800_end_0, end_mask = var_1800_end_mask_0, x = x_187_cast_fp16)[name = tensor("op_1800_cast_fp16")]; - tensor var_1801 = const()[name = tensor("op_1801"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_33_cast_fp16 = reshape(shape = var_1801, x = var_1800_cast_fp16)[name = tensor("matrix_bd_33_cast_fp16")]; - tensor matrix_ac_17_transpose_x_0 = const()[name = tensor("matrix_ac_17_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_17_transpose_y_0 = const()[name = tensor("matrix_ac_17_transpose_y_0"), val = tensor(false)]; - tensor transpose_67_perm_0 = const()[name = tensor("transpose_67_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_68_perm_0 = const()[name = tensor("transpose_68_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_68 = transpose(perm = transpose_68_perm_0, x = k_33_cast_fp16)[name = tensor("transpose_145")]; - tensor transpose_67 = transpose(perm = transpose_67_perm_0, x = var_1784_cast_fp16)[name = tensor("transpose_146")]; - tensor matrix_ac_17_cast_fp16 = matmul(transpose_x = matrix_ac_17_transpose_x_0, transpose_y = matrix_ac_17_transpose_y_0, x = transpose_67, y = transpose_68)[name = tensor("matrix_ac_17_cast_fp16")]; - tensor matrix_bd_35_begin_0 = const()[name = tensor("matrix_bd_35_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_35_end_0 = const()[name = tensor("matrix_bd_35_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_35_end_mask_0 = const()[name = tensor("matrix_bd_35_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_35_cast_fp16 = slice_by_index(begin = matrix_bd_35_begin_0, end = matrix_bd_35_end_0, end_mask = matrix_bd_35_end_mask_0, x = matrix_bd_33_cast_fp16)[name = tensor("matrix_bd_35_cast_fp16")]; - tensor var_1810_cast_fp16 = add(x = matrix_ac_17_cast_fp16, y = matrix_bd_35_cast_fp16)[name = tensor("op_1810_cast_fp16")]; - tensor _inversed_scores_33_y_0_to_fp16 = const()[name = tensor("_inversed_scores_33_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_33_cast_fp16 = mul(x = var_1810_cast_fp16, y = _inversed_scores_33_y_0_to_fp16)[name = tensor("_inversed_scores_33_cast_fp16")]; - tensor scores_35_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_33_cast_fp16, cond = mask_11)[name = tensor("scores_35_cast_fp16")]; - tensor var_1816_cast_fp16 = softmax(axis = var_23, x = scores_35_cast_fp16)[name = tensor("op_1816_cast_fp16")]; - tensor input_449_cast_fp16 = select(a = var_11_to_fp16, b = var_1816_cast_fp16, cond = mask_11)[name = tensor("input_449_cast_fp16")]; - tensor x_189_transpose_x_0 = const()[name = tensor("x_189_transpose_x_0"), val = tensor(false)]; - tensor x_189_transpose_y_0 = const()[name = tensor("x_189_transpose_y_0"), val = tensor(false)]; - tensor value_19_cast_fp16 = transpose(perm = value_19_perm_0, x = v_17_cast_fp16)[name = tensor("transpose_148")]; - tensor x_189_cast_fp16 = matmul(transpose_x = x_189_transpose_x_0, transpose_y = x_189_transpose_y_0, x = input_449_cast_fp16, y = value_19_cast_fp16)[name = tensor("x_189_cast_fp16")]; - tensor var_1820_perm_0 = const()[name = tensor("op_1820_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_1821 = const()[name = tensor("op_1821"), val = tensor([1, -1, 512])]; - tensor var_1820_cast_fp16 = transpose(perm = var_1820_perm_0, x = x_189_cast_fp16)[name = tensor("transpose_144")]; - tensor input_451_cast_fp16 = reshape(shape = var_1821, x = var_1820_cast_fp16)[name = tensor("input_451_cast_fp16")]; - tensor module_layers_8_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_8_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108946816)))]; - tensor module_layers_8_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_8_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109471168)))]; - tensor linear_79_cast_fp16 = linear(bias = module_layers_8_self_attn_linear_out_bias_to_fp16, weight = module_layers_8_self_attn_linear_out_weight_to_fp16, x = input_451_cast_fp16)[name = tensor("linear_79_cast_fp16")]; - tensor input_455_cast_fp16 = add(x = input_447_cast_fp16, y = linear_79_cast_fp16)[name = tensor("input_455_cast_fp16")]; - tensor x_193_axes_0 = const()[name = tensor("x_193_axes_0"), val = tensor([-1])]; - tensor module_layers_8_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_8_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109472256)))]; - tensor module_layers_8_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_8_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109473344)))]; - tensor x_193_cast_fp16 = layer_norm(axes = x_193_axes_0, beta = module_layers_8_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_8_norm_conv_weight_to_fp16, x = input_455_cast_fp16)[name = tensor("x_193_cast_fp16")]; - tensor input_457_perm_0 = const()[name = tensor("input_457_perm_0"), val = tensor([0, 2, 1])]; - tensor input_459_pad_type_0 = const()[name = tensor("input_459_pad_type_0"), val = tensor("valid")]; - tensor input_459_strides_0 = const()[name = tensor("input_459_strides_0"), val = tensor([1])]; - tensor input_459_pad_0 = const()[name = tensor("input_459_pad_0"), val = tensor([0, 0])]; - tensor input_459_dilations_0 = const()[name = tensor("input_459_dilations_0"), val = tensor([1])]; - tensor input_459_groups_0 = const()[name = tensor("input_459_groups_0"), val = tensor(1)]; - tensor module_layers_8_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_8_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109474432)))]; - tensor module_layers_8_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_8_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110523072)))]; - tensor input_457_cast_fp16 = transpose(perm = input_457_perm_0, x = x_193_cast_fp16)[name = tensor("transpose_143")]; - tensor input_459_cast_fp16 = conv(bias = module_layers_8_conv_pointwise_conv1_bias_to_fp16, dilations = input_459_dilations_0, groups = input_459_groups_0, pad = input_459_pad_0, pad_type = input_459_pad_type_0, strides = input_459_strides_0, weight = module_layers_8_conv_pointwise_conv1_weight_to_fp16, x = input_457_cast_fp16)[name = tensor("input_459_cast_fp16")]; - tensor x_195_split_num_splits_0 = const()[name = tensor("x_195_split_num_splits_0"), val = tensor(2)]; - tensor x_195_split_axis_0 = const()[name = tensor("x_195_split_axis_0"), val = tensor(1)]; - tensor x_195_split_cast_fp16_0, tensor x_195_split_cast_fp16_1 = split(axis = x_195_split_axis_0, num_splits = x_195_split_num_splits_0, x = input_459_cast_fp16)[name = tensor("x_195_split_cast_fp16")]; - tensor x_195_split_1_sigmoid_cast_fp16 = sigmoid(x = x_195_split_cast_fp16_1)[name = tensor("x_195_split_1_sigmoid_cast_fp16")]; - tensor x_195_cast_fp16 = mul(x = x_195_split_cast_fp16_0, y = x_195_split_1_sigmoid_cast_fp16)[name = tensor("x_195_cast_fp16")]; - tensor input_461_cast_fp16 = select(a = var_11_to_fp16, b = x_195_cast_fp16, cond = var_453)[name = tensor("input_461_cast_fp16")]; - tensor input_463_pad_0 = const()[name = tensor("input_463_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_463_mode_0 = const()[name = tensor("input_463_mode_0"), val = tensor("constant")]; - tensor const_153_to_fp16 = const()[name = tensor("const_153_to_fp16"), val = tensor(0x0p+0)]; - tensor input_463_cast_fp16 = pad(constant_val = const_153_to_fp16, mode = input_463_mode_0, pad = input_463_pad_0, x = input_461_cast_fp16)[name = tensor("input_463_cast_fp16")]; - tensor input_465_pad_type_0 = const()[name = tensor("input_465_pad_type_0"), val = tensor("valid")]; - tensor input_465_groups_0 = const()[name = tensor("input_465_groups_0"), val = tensor(512)]; - tensor input_465_strides_0 = const()[name = tensor("input_465_strides_0"), val = tensor([1])]; - tensor input_465_pad_0 = const()[name = tensor("input_465_pad_0"), val = tensor([0, 0])]; - tensor input_465_dilations_0 = const()[name = tensor("input_465_dilations_0"), val = tensor([1])]; - tensor const_250_to_fp16 = const()[name = tensor("const_250_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110525184)))]; - tensor const_251_to_fp16 = const()[name = tensor("const_251_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110534464)))]; - tensor input_467_cast_fp16 = conv(bias = const_251_to_fp16, dilations = input_465_dilations_0, groups = input_465_groups_0, pad = input_465_pad_0, pad_type = input_465_pad_type_0, strides = input_465_strides_0, weight = const_250_to_fp16, x = input_463_cast_fp16)[name = tensor("input_467_cast_fp16")]; - tensor input_469_cast_fp16 = silu(x = input_467_cast_fp16)[name = tensor("input_469_cast_fp16")]; - tensor x_197_pad_type_0 = const()[name = tensor("x_197_pad_type_0"), val = tensor("valid")]; - tensor x_197_strides_0 = const()[name = tensor("x_197_strides_0"), val = tensor([1])]; - tensor x_197_pad_0 = const()[name = tensor("x_197_pad_0"), val = tensor([0, 0])]; - tensor x_197_dilations_0 = const()[name = tensor("x_197_dilations_0"), val = tensor([1])]; - tensor x_197_groups_0 = const()[name = tensor("x_197_groups_0"), val = tensor(1)]; - tensor module_layers_8_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_8_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110535552)))]; - tensor module_layers_8_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_8_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(111059904)))]; - tensor x_197_cast_fp16 = conv(bias = module_layers_8_conv_pointwise_conv2_bias_to_fp16, dilations = x_197_dilations_0, groups = x_197_groups_0, pad = x_197_pad_0, pad_type = x_197_pad_type_0, strides = x_197_strides_0, weight = module_layers_8_conv_pointwise_conv2_weight_to_fp16, x = input_469_cast_fp16)[name = tensor("x_197_cast_fp16")]; - tensor input_471_perm_0 = const()[name = tensor("input_471_perm_0"), val = tensor([0, 2, 1])]; - tensor input_471_cast_fp16 = transpose(perm = input_471_perm_0, x = x_197_cast_fp16)[name = tensor("transpose_142")]; - tensor input_473_cast_fp16 = add(x = input_455_cast_fp16, y = input_471_cast_fp16)[name = tensor("input_473_cast_fp16")]; - tensor input_475_axes_0 = const()[name = tensor("input_475_axes_0"), val = tensor([-1])]; - tensor module_layers_8_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_8_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(111060992)))]; - tensor module_layers_8_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_8_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(111062080)))]; - tensor input_475_cast_fp16 = layer_norm(axes = input_475_axes_0, beta = module_layers_8_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_8_norm_feed_forward2_weight_to_fp16, x = input_473_cast_fp16)[name = tensor("input_475_cast_fp16")]; - tensor module_layers_8_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_8_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(111063168)))]; - tensor module_layers_8_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_8_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(113160384)))]; - tensor linear_80_cast_fp16 = linear(bias = module_layers_8_feed_forward2_linear1_bias_to_fp16, weight = module_layers_8_feed_forward2_linear1_weight_to_fp16, x = input_475_cast_fp16)[name = tensor("linear_80_cast_fp16")]; - tensor input_479_cast_fp16 = silu(x = linear_80_cast_fp16)[name = tensor("input_479_cast_fp16")]; - tensor module_layers_8_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_8_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(113164544)))]; - tensor module_layers_8_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_8_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(115261760)))]; - tensor linear_81_cast_fp16 = linear(bias = module_layers_8_feed_forward2_linear2_bias_to_fp16, weight = module_layers_8_feed_forward2_linear2_weight_to_fp16, x = input_479_cast_fp16)[name = tensor("linear_81_cast_fp16")]; - tensor var_1887_to_fp16 = const()[name = tensor("op_1887_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1888_cast_fp16 = mul(x = linear_81_cast_fp16, y = var_1887_to_fp16)[name = tensor("op_1888_cast_fp16")]; - tensor input_485_cast_fp16 = add(x = input_473_cast_fp16, y = var_1888_cast_fp16)[name = tensor("input_485_cast_fp16")]; - tensor input_487_axes_0 = const()[name = tensor("input_487_axes_0"), val = tensor([-1])]; - tensor module_layers_8_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_8_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(115262848)))]; - tensor module_layers_8_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_8_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(115263936)))]; - tensor input_487_cast_fp16 = layer_norm(axes = input_487_axes_0, beta = module_layers_8_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_8_norm_out_weight_to_fp16, x = input_485_cast_fp16)[name = tensor("input_487_cast_fp16")]; - tensor input_489_axes_0 = const()[name = tensor("input_489_axes_0"), val = tensor([-1])]; - tensor module_layers_9_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_9_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(115265024)))]; - tensor module_layers_9_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_9_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(115266112)))]; - tensor input_489_cast_fp16 = layer_norm(axes = input_489_axes_0, beta = module_layers_9_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_9_norm_feed_forward1_weight_to_fp16, x = input_487_cast_fp16)[name = tensor("input_489_cast_fp16")]; - tensor module_layers_9_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_9_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(115267200)))]; - tensor module_layers_9_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_9_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(117364416)))]; - tensor linear_82_cast_fp16 = linear(bias = module_layers_9_feed_forward1_linear1_bias_to_fp16, weight = module_layers_9_feed_forward1_linear1_weight_to_fp16, x = input_489_cast_fp16)[name = tensor("linear_82_cast_fp16")]; - tensor input_493_cast_fp16 = silu(x = linear_82_cast_fp16)[name = tensor("input_493_cast_fp16")]; - tensor module_layers_9_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_9_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(117368576)))]; - tensor module_layers_9_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_9_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(119465792)))]; - tensor linear_83_cast_fp16 = linear(bias = module_layers_9_feed_forward1_linear2_bias_to_fp16, weight = module_layers_9_feed_forward1_linear2_weight_to_fp16, x = input_493_cast_fp16)[name = tensor("linear_83_cast_fp16")]; - tensor var_1918_to_fp16 = const()[name = tensor("op_1918_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1919_cast_fp16 = mul(x = linear_83_cast_fp16, y = var_1918_to_fp16)[name = tensor("op_1919_cast_fp16")]; - tensor input_499_cast_fp16 = add(x = input_487_cast_fp16, y = var_1919_cast_fp16)[name = tensor("input_499_cast_fp16")]; - tensor query_19_axes_0 = const()[name = tensor("query_19_axes_0"), val = tensor([-1])]; - tensor module_layers_9_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_9_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(119466880)))]; - tensor module_layers_9_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_9_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(119467968)))]; - tensor query_19_cast_fp16 = layer_norm(axes = query_19_axes_0, beta = module_layers_9_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_9_norm_self_att_weight_to_fp16, x = input_499_cast_fp16)[name = tensor("query_19_cast_fp16")]; - tensor module_layers_9_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_9_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(119469056)))]; - tensor module_layers_9_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_9_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(119993408)))]; - tensor linear_84_cast_fp16 = linear(bias = module_layers_9_self_attn_linear_q_bias_to_fp16, weight = module_layers_9_self_attn_linear_q_weight_to_fp16, x = query_19_cast_fp16)[name = tensor("linear_84_cast_fp16")]; - tensor var_1936 = const()[name = tensor("op_1936"), val = tensor([1, -1, 8, 64])]; - tensor q_55_cast_fp16 = reshape(shape = var_1936, x = linear_84_cast_fp16)[name = tensor("q_55_cast_fp16")]; - tensor module_layers_9_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_9_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(119994496)))]; - tensor module_layers_9_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_9_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(120518848)))]; - tensor linear_85_cast_fp16 = linear(bias = module_layers_9_self_attn_linear_k_bias_to_fp16, weight = module_layers_9_self_attn_linear_k_weight_to_fp16, x = query_19_cast_fp16)[name = tensor("linear_85_cast_fp16")]; - tensor var_1941 = const()[name = tensor("op_1941"), val = tensor([1, -1, 8, 64])]; - tensor k_37_cast_fp16 = reshape(shape = var_1941, x = linear_85_cast_fp16)[name = tensor("k_37_cast_fp16")]; - tensor module_layers_9_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_9_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(120519936)))]; - tensor module_layers_9_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_9_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121044288)))]; - tensor linear_86_cast_fp16 = linear(bias = module_layers_9_self_attn_linear_v_bias_to_fp16, weight = module_layers_9_self_attn_linear_v_weight_to_fp16, x = query_19_cast_fp16)[name = tensor("linear_86_cast_fp16")]; - tensor var_1946 = const()[name = tensor("op_1946"), val = tensor([1, -1, 8, 64])]; - tensor v_19_cast_fp16 = reshape(shape = var_1946, x = linear_86_cast_fp16)[name = tensor("v_19_cast_fp16")]; - tensor value_21_perm_0 = const()[name = tensor("value_21_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_9_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_9_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121045376)))]; - tensor var_1958_cast_fp16 = add(x = q_55_cast_fp16, y = module_layers_9_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1958_cast_fp16")]; - tensor module_layers_9_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_9_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121046464)))]; - tensor var_1960_cast_fp16 = add(x = q_55_cast_fp16, y = module_layers_9_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1960_cast_fp16")]; - tensor q_with_bias_v_19_perm_0 = const()[name = tensor("q_with_bias_v_19_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_205_transpose_x_0 = const()[name = tensor("x_205_transpose_x_0"), val = tensor(false)]; - tensor x_205_transpose_y_0 = const()[name = tensor("x_205_transpose_y_0"), val = tensor(false)]; - tensor var_1962_to_fp16 = const()[name = tensor("op_1962_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121047552)))]; - tensor q_with_bias_v_19_cast_fp16 = transpose(perm = q_with_bias_v_19_perm_0, x = var_1960_cast_fp16)[name = tensor("transpose_140")]; - tensor x_205_cast_fp16 = matmul(transpose_x = x_205_transpose_x_0, transpose_y = x_205_transpose_y_0, x = q_with_bias_v_19_cast_fp16, y = var_1962_to_fp16)[name = tensor("x_205_cast_fp16")]; - tensor x_207_pad_0 = const()[name = tensor("x_207_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_207_mode_0 = const()[name = tensor("x_207_mode_0"), val = tensor("constant")]; - tensor const_160_to_fp16 = const()[name = tensor("const_160_to_fp16"), val = tensor(0x0p+0)]; - tensor x_207_cast_fp16 = pad(constant_val = const_160_to_fp16, mode = x_207_mode_0, pad = x_207_pad_0, x = x_205_cast_fp16)[name = tensor("x_207_cast_fp16")]; - tensor var_1970 = const()[name = tensor("op_1970"), val = tensor([1, 8, -1, 188])]; - tensor x_209_cast_fp16 = reshape(shape = var_1970, x = x_207_cast_fp16)[name = tensor("x_209_cast_fp16")]; - tensor var_1974_begin_0 = const()[name = tensor("op_1974_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_1974_end_0 = const()[name = tensor("op_1974_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_1974_end_mask_0 = const()[name = tensor("op_1974_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_1974_cast_fp16 = slice_by_index(begin = var_1974_begin_0, end = var_1974_end_0, end_mask = var_1974_end_mask_0, x = x_209_cast_fp16)[name = tensor("op_1974_cast_fp16")]; - tensor var_1975 = const()[name = tensor("op_1975"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_37_cast_fp16 = reshape(shape = var_1975, x = var_1974_cast_fp16)[name = tensor("matrix_bd_37_cast_fp16")]; - tensor matrix_ac_19_transpose_x_0 = const()[name = tensor("matrix_ac_19_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_19_transpose_y_0 = const()[name = tensor("matrix_ac_19_transpose_y_0"), val = tensor(false)]; - tensor transpose_69_perm_0 = const()[name = tensor("transpose_69_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_70_perm_0 = const()[name = tensor("transpose_70_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_70 = transpose(perm = transpose_70_perm_0, x = k_37_cast_fp16)[name = tensor("transpose_138")]; - tensor transpose_69 = transpose(perm = transpose_69_perm_0, x = var_1958_cast_fp16)[name = tensor("transpose_139")]; - tensor matrix_ac_19_cast_fp16 = matmul(transpose_x = matrix_ac_19_transpose_x_0, transpose_y = matrix_ac_19_transpose_y_0, x = transpose_69, y = transpose_70)[name = tensor("matrix_ac_19_cast_fp16")]; - tensor matrix_bd_39_begin_0 = const()[name = tensor("matrix_bd_39_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_39_end_0 = const()[name = tensor("matrix_bd_39_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_39_end_mask_0 = const()[name = tensor("matrix_bd_39_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_39_cast_fp16 = slice_by_index(begin = matrix_bd_39_begin_0, end = matrix_bd_39_end_0, end_mask = matrix_bd_39_end_mask_0, x = matrix_bd_37_cast_fp16)[name = tensor("matrix_bd_39_cast_fp16")]; - tensor var_1984_cast_fp16 = add(x = matrix_ac_19_cast_fp16, y = matrix_bd_39_cast_fp16)[name = tensor("op_1984_cast_fp16")]; - tensor _inversed_scores_37_y_0_to_fp16 = const()[name = tensor("_inversed_scores_37_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_37_cast_fp16 = mul(x = var_1984_cast_fp16, y = _inversed_scores_37_y_0_to_fp16)[name = tensor("_inversed_scores_37_cast_fp16")]; - tensor scores_39_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_37_cast_fp16, cond = mask_11)[name = tensor("scores_39_cast_fp16")]; - tensor var_1990_cast_fp16 = softmax(axis = var_23, x = scores_39_cast_fp16)[name = tensor("op_1990_cast_fp16")]; - tensor input_501_cast_fp16 = select(a = var_11_to_fp16, b = var_1990_cast_fp16, cond = mask_11)[name = tensor("input_501_cast_fp16")]; - tensor x_211_transpose_x_0 = const()[name = tensor("x_211_transpose_x_0"), val = tensor(false)]; - tensor x_211_transpose_y_0 = const()[name = tensor("x_211_transpose_y_0"), val = tensor(false)]; - tensor value_21_cast_fp16 = transpose(perm = value_21_perm_0, x = v_19_cast_fp16)[name = tensor("transpose_141")]; - tensor x_211_cast_fp16 = matmul(transpose_x = x_211_transpose_x_0, transpose_y = x_211_transpose_y_0, x = input_501_cast_fp16, y = value_21_cast_fp16)[name = tensor("x_211_cast_fp16")]; - tensor var_1994_perm_0 = const()[name = tensor("op_1994_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_1995 = const()[name = tensor("op_1995"), val = tensor([1, -1, 512])]; - tensor var_1994_cast_fp16 = transpose(perm = var_1994_perm_0, x = x_211_cast_fp16)[name = tensor("transpose_137")]; - tensor input_503_cast_fp16 = reshape(shape = var_1995, x = var_1994_cast_fp16)[name = tensor("input_503_cast_fp16")]; - tensor module_layers_9_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_9_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121431616)))]; - tensor module_layers_9_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_9_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121955968)))]; - tensor linear_88_cast_fp16 = linear(bias = module_layers_9_self_attn_linear_out_bias_to_fp16, weight = module_layers_9_self_attn_linear_out_weight_to_fp16, x = input_503_cast_fp16)[name = tensor("linear_88_cast_fp16")]; - tensor input_507_cast_fp16 = add(x = input_499_cast_fp16, y = linear_88_cast_fp16)[name = tensor("input_507_cast_fp16")]; - tensor x_215_axes_0 = const()[name = tensor("x_215_axes_0"), val = tensor([-1])]; - tensor module_layers_9_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_9_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121957056)))]; - tensor module_layers_9_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_9_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121958144)))]; - tensor x_215_cast_fp16 = layer_norm(axes = x_215_axes_0, beta = module_layers_9_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_9_norm_conv_weight_to_fp16, x = input_507_cast_fp16)[name = tensor("x_215_cast_fp16")]; - tensor input_509_perm_0 = const()[name = tensor("input_509_perm_0"), val = tensor([0, 2, 1])]; - tensor input_511_pad_type_0 = const()[name = tensor("input_511_pad_type_0"), val = tensor("valid")]; - tensor input_511_strides_0 = const()[name = tensor("input_511_strides_0"), val = tensor([1])]; - tensor input_511_pad_0 = const()[name = tensor("input_511_pad_0"), val = tensor([0, 0])]; - tensor input_511_dilations_0 = const()[name = tensor("input_511_dilations_0"), val = tensor([1])]; - tensor input_511_groups_0 = const()[name = tensor("input_511_groups_0"), val = tensor(1)]; - tensor module_layers_9_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_9_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121959232)))]; - tensor module_layers_9_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_9_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123007872)))]; - tensor input_509_cast_fp16 = transpose(perm = input_509_perm_0, x = x_215_cast_fp16)[name = tensor("transpose_136")]; - tensor input_511_cast_fp16 = conv(bias = module_layers_9_conv_pointwise_conv1_bias_to_fp16, dilations = input_511_dilations_0, groups = input_511_groups_0, pad = input_511_pad_0, pad_type = input_511_pad_type_0, strides = input_511_strides_0, weight = module_layers_9_conv_pointwise_conv1_weight_to_fp16, x = input_509_cast_fp16)[name = tensor("input_511_cast_fp16")]; - tensor x_217_split_num_splits_0 = const()[name = tensor("x_217_split_num_splits_0"), val = tensor(2)]; - tensor x_217_split_axis_0 = const()[name = tensor("x_217_split_axis_0"), val = tensor(1)]; - tensor x_217_split_cast_fp16_0, tensor x_217_split_cast_fp16_1 = split(axis = x_217_split_axis_0, num_splits = x_217_split_num_splits_0, x = input_511_cast_fp16)[name = tensor("x_217_split_cast_fp16")]; - tensor x_217_split_1_sigmoid_cast_fp16 = sigmoid(x = x_217_split_cast_fp16_1)[name = tensor("x_217_split_1_sigmoid_cast_fp16")]; - tensor x_217_cast_fp16 = mul(x = x_217_split_cast_fp16_0, y = x_217_split_1_sigmoid_cast_fp16)[name = tensor("x_217_cast_fp16")]; - tensor input_513_cast_fp16 = select(a = var_11_to_fp16, b = x_217_cast_fp16, cond = var_453)[name = tensor("input_513_cast_fp16")]; - tensor input_515_pad_0 = const()[name = tensor("input_515_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_515_mode_0 = const()[name = tensor("input_515_mode_0"), val = tensor("constant")]; - tensor const_163_to_fp16 = const()[name = tensor("const_163_to_fp16"), val = tensor(0x0p+0)]; - tensor input_515_cast_fp16 = pad(constant_val = const_163_to_fp16, mode = input_515_mode_0, pad = input_515_pad_0, x = input_513_cast_fp16)[name = tensor("input_515_cast_fp16")]; - tensor input_517_pad_type_0 = const()[name = tensor("input_517_pad_type_0"), val = tensor("valid")]; - tensor input_517_groups_0 = const()[name = tensor("input_517_groups_0"), val = tensor(512)]; - tensor input_517_strides_0 = const()[name = tensor("input_517_strides_0"), val = tensor([1])]; - tensor input_517_pad_0 = const()[name = tensor("input_517_pad_0"), val = tensor([0, 0])]; - tensor input_517_dilations_0 = const()[name = tensor("input_517_dilations_0"), val = tensor([1])]; - tensor const_252_to_fp16 = const()[name = tensor("const_252_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123009984)))]; - tensor const_253_to_fp16 = const()[name = tensor("const_253_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123019264)))]; - tensor input_519_cast_fp16 = conv(bias = const_253_to_fp16, dilations = input_517_dilations_0, groups = input_517_groups_0, pad = input_517_pad_0, pad_type = input_517_pad_type_0, strides = input_517_strides_0, weight = const_252_to_fp16, x = input_515_cast_fp16)[name = tensor("input_519_cast_fp16")]; - tensor input_521_cast_fp16 = silu(x = input_519_cast_fp16)[name = tensor("input_521_cast_fp16")]; - tensor x_219_pad_type_0 = const()[name = tensor("x_219_pad_type_0"), val = tensor("valid")]; - tensor x_219_strides_0 = const()[name = tensor("x_219_strides_0"), val = tensor([1])]; - tensor x_219_pad_0 = const()[name = tensor("x_219_pad_0"), val = tensor([0, 0])]; - tensor x_219_dilations_0 = const()[name = tensor("x_219_dilations_0"), val = tensor([1])]; - tensor x_219_groups_0 = const()[name = tensor("x_219_groups_0"), val = tensor(1)]; - tensor module_layers_9_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_9_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123020352)))]; - tensor module_layers_9_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_9_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123544704)))]; - tensor x_219_cast_fp16 = conv(bias = module_layers_9_conv_pointwise_conv2_bias_to_fp16, dilations = x_219_dilations_0, groups = x_219_groups_0, pad = x_219_pad_0, pad_type = x_219_pad_type_0, strides = x_219_strides_0, weight = module_layers_9_conv_pointwise_conv2_weight_to_fp16, x = input_521_cast_fp16)[name = tensor("x_219_cast_fp16")]; - tensor input_523_perm_0 = const()[name = tensor("input_523_perm_0"), val = tensor([0, 2, 1])]; - tensor input_523_cast_fp16 = transpose(perm = input_523_perm_0, x = x_219_cast_fp16)[name = tensor("transpose_135")]; - tensor input_525_cast_fp16 = add(x = input_507_cast_fp16, y = input_523_cast_fp16)[name = tensor("input_525_cast_fp16")]; - tensor input_527_axes_0 = const()[name = tensor("input_527_axes_0"), val = tensor([-1])]; - tensor module_layers_9_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_9_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123545792)))]; - tensor module_layers_9_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_9_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123546880)))]; - tensor input_527_cast_fp16 = layer_norm(axes = input_527_axes_0, beta = module_layers_9_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_9_norm_feed_forward2_weight_to_fp16, x = input_525_cast_fp16)[name = tensor("input_527_cast_fp16")]; - tensor module_layers_9_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_9_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123547968)))]; - tensor module_layers_9_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_9_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(125645184)))]; - tensor linear_89_cast_fp16 = linear(bias = module_layers_9_feed_forward2_linear1_bias_to_fp16, weight = module_layers_9_feed_forward2_linear1_weight_to_fp16, x = input_527_cast_fp16)[name = tensor("linear_89_cast_fp16")]; - tensor input_531_cast_fp16 = silu(x = linear_89_cast_fp16)[name = tensor("input_531_cast_fp16")]; - tensor module_layers_9_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_9_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(125649344)))]; - tensor module_layers_9_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_9_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127746560)))]; - tensor linear_90_cast_fp16 = linear(bias = module_layers_9_feed_forward2_linear2_bias_to_fp16, weight = module_layers_9_feed_forward2_linear2_weight_to_fp16, x = input_531_cast_fp16)[name = tensor("linear_90_cast_fp16")]; - tensor var_2061_to_fp16 = const()[name = tensor("op_2061_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2062_cast_fp16 = mul(x = linear_90_cast_fp16, y = var_2061_to_fp16)[name = tensor("op_2062_cast_fp16")]; - tensor input_537_cast_fp16 = add(x = input_525_cast_fp16, y = var_2062_cast_fp16)[name = tensor("input_537_cast_fp16")]; - tensor input_539_axes_0 = const()[name = tensor("input_539_axes_0"), val = tensor([-1])]; - tensor module_layers_9_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_9_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127747648)))]; - tensor module_layers_9_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_9_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127748736)))]; - tensor input_539_cast_fp16 = layer_norm(axes = input_539_axes_0, beta = module_layers_9_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_9_norm_out_weight_to_fp16, x = input_537_cast_fp16)[name = tensor("input_539_cast_fp16")]; - tensor input_541_axes_0 = const()[name = tensor("input_541_axes_0"), val = tensor([-1])]; - tensor module_layers_10_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_10_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127749824)))]; - tensor module_layers_10_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_10_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127750912)))]; - tensor input_541_cast_fp16 = layer_norm(axes = input_541_axes_0, beta = module_layers_10_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_10_norm_feed_forward1_weight_to_fp16, x = input_539_cast_fp16)[name = tensor("input_541_cast_fp16")]; - tensor module_layers_10_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_10_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127752000)))]; - tensor module_layers_10_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_10_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(129849216)))]; - tensor linear_91_cast_fp16 = linear(bias = module_layers_10_feed_forward1_linear1_bias_to_fp16, weight = module_layers_10_feed_forward1_linear1_weight_to_fp16, x = input_541_cast_fp16)[name = tensor("linear_91_cast_fp16")]; - tensor input_545_cast_fp16 = silu(x = linear_91_cast_fp16)[name = tensor("input_545_cast_fp16")]; - tensor module_layers_10_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_10_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(129853376)))]; - tensor module_layers_10_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_10_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(131950592)))]; - tensor linear_92_cast_fp16 = linear(bias = module_layers_10_feed_forward1_linear2_bias_to_fp16, weight = module_layers_10_feed_forward1_linear2_weight_to_fp16, x = input_545_cast_fp16)[name = tensor("linear_92_cast_fp16")]; - tensor var_2092_to_fp16 = const()[name = tensor("op_2092_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2093_cast_fp16 = mul(x = linear_92_cast_fp16, y = var_2092_to_fp16)[name = tensor("op_2093_cast_fp16")]; - tensor input_551_cast_fp16 = add(x = input_539_cast_fp16, y = var_2093_cast_fp16)[name = tensor("input_551_cast_fp16")]; - tensor query_21_axes_0 = const()[name = tensor("query_21_axes_0"), val = tensor([-1])]; - tensor module_layers_10_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_10_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(131951680)))]; - tensor module_layers_10_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_10_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(131952768)))]; - tensor query_21_cast_fp16 = layer_norm(axes = query_21_axes_0, beta = module_layers_10_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_10_norm_self_att_weight_to_fp16, x = input_551_cast_fp16)[name = tensor("query_21_cast_fp16")]; - tensor module_layers_10_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_10_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(131953856)))]; - tensor module_layers_10_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_10_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132478208)))]; - tensor linear_93_cast_fp16 = linear(bias = module_layers_10_self_attn_linear_q_bias_to_fp16, weight = module_layers_10_self_attn_linear_q_weight_to_fp16, x = query_21_cast_fp16)[name = tensor("linear_93_cast_fp16")]; - tensor var_2110 = const()[name = tensor("op_2110"), val = tensor([1, -1, 8, 64])]; - tensor q_61_cast_fp16 = reshape(shape = var_2110, x = linear_93_cast_fp16)[name = tensor("q_61_cast_fp16")]; - tensor module_layers_10_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_10_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132479296)))]; - tensor module_layers_10_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_10_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133003648)))]; - tensor linear_94_cast_fp16 = linear(bias = module_layers_10_self_attn_linear_k_bias_to_fp16, weight = module_layers_10_self_attn_linear_k_weight_to_fp16, x = query_21_cast_fp16)[name = tensor("linear_94_cast_fp16")]; - tensor var_2115 = const()[name = tensor("op_2115"), val = tensor([1, -1, 8, 64])]; - tensor k_41_cast_fp16 = reshape(shape = var_2115, x = linear_94_cast_fp16)[name = tensor("k_41_cast_fp16")]; - tensor module_layers_10_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_10_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133004736)))]; - tensor module_layers_10_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_10_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133529088)))]; - tensor linear_95_cast_fp16 = linear(bias = module_layers_10_self_attn_linear_v_bias_to_fp16, weight = module_layers_10_self_attn_linear_v_weight_to_fp16, x = query_21_cast_fp16)[name = tensor("linear_95_cast_fp16")]; - tensor var_2120 = const()[name = tensor("op_2120"), val = tensor([1, -1, 8, 64])]; - tensor v_21_cast_fp16 = reshape(shape = var_2120, x = linear_95_cast_fp16)[name = tensor("v_21_cast_fp16")]; - tensor value_23_perm_0 = const()[name = tensor("value_23_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_10_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_10_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133530176)))]; - tensor var_2132_cast_fp16 = add(x = q_61_cast_fp16, y = module_layers_10_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2132_cast_fp16")]; - tensor module_layers_10_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_10_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133531264)))]; - tensor var_2134_cast_fp16 = add(x = q_61_cast_fp16, y = module_layers_10_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2134_cast_fp16")]; - tensor q_with_bias_v_21_perm_0 = const()[name = tensor("q_with_bias_v_21_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_227_transpose_x_0 = const()[name = tensor("x_227_transpose_x_0"), val = tensor(false)]; - tensor x_227_transpose_y_0 = const()[name = tensor("x_227_transpose_y_0"), val = tensor(false)]; - tensor var_2136_to_fp16 = const()[name = tensor("op_2136_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133532352)))]; - tensor q_with_bias_v_21_cast_fp16 = transpose(perm = q_with_bias_v_21_perm_0, x = var_2134_cast_fp16)[name = tensor("transpose_133")]; - tensor x_227_cast_fp16 = matmul(transpose_x = x_227_transpose_x_0, transpose_y = x_227_transpose_y_0, x = q_with_bias_v_21_cast_fp16, y = var_2136_to_fp16)[name = tensor("x_227_cast_fp16")]; - tensor x_229_pad_0 = const()[name = tensor("x_229_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_229_mode_0 = const()[name = tensor("x_229_mode_0"), val = tensor("constant")]; - tensor const_170_to_fp16 = const()[name = tensor("const_170_to_fp16"), val = tensor(0x0p+0)]; - tensor x_229_cast_fp16 = pad(constant_val = const_170_to_fp16, mode = x_229_mode_0, pad = x_229_pad_0, x = x_227_cast_fp16)[name = tensor("x_229_cast_fp16")]; - tensor var_2144 = const()[name = tensor("op_2144"), val = tensor([1, 8, -1, 188])]; - tensor x_231_cast_fp16 = reshape(shape = var_2144, x = x_229_cast_fp16)[name = tensor("x_231_cast_fp16")]; - tensor var_2148_begin_0 = const()[name = tensor("op_2148_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_2148_end_0 = const()[name = tensor("op_2148_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_2148_end_mask_0 = const()[name = tensor("op_2148_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_2148_cast_fp16 = slice_by_index(begin = var_2148_begin_0, end = var_2148_end_0, end_mask = var_2148_end_mask_0, x = x_231_cast_fp16)[name = tensor("op_2148_cast_fp16")]; - tensor var_2149 = const()[name = tensor("op_2149"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_41_cast_fp16 = reshape(shape = var_2149, x = var_2148_cast_fp16)[name = tensor("matrix_bd_41_cast_fp16")]; - tensor matrix_ac_21_transpose_x_0 = const()[name = tensor("matrix_ac_21_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_21_transpose_y_0 = const()[name = tensor("matrix_ac_21_transpose_y_0"), val = tensor(false)]; - tensor transpose_71_perm_0 = const()[name = tensor("transpose_71_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_72_perm_0 = const()[name = tensor("transpose_72_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_72 = transpose(perm = transpose_72_perm_0, x = k_41_cast_fp16)[name = tensor("transpose_131")]; - tensor transpose_71 = transpose(perm = transpose_71_perm_0, x = var_2132_cast_fp16)[name = tensor("transpose_132")]; - tensor matrix_ac_21_cast_fp16 = matmul(transpose_x = matrix_ac_21_transpose_x_0, transpose_y = matrix_ac_21_transpose_y_0, x = transpose_71, y = transpose_72)[name = tensor("matrix_ac_21_cast_fp16")]; - tensor matrix_bd_43_begin_0 = const()[name = tensor("matrix_bd_43_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_43_end_0 = const()[name = tensor("matrix_bd_43_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_43_end_mask_0 = const()[name = tensor("matrix_bd_43_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_43_cast_fp16 = slice_by_index(begin = matrix_bd_43_begin_0, end = matrix_bd_43_end_0, end_mask = matrix_bd_43_end_mask_0, x = matrix_bd_41_cast_fp16)[name = tensor("matrix_bd_43_cast_fp16")]; - tensor var_2158_cast_fp16 = add(x = matrix_ac_21_cast_fp16, y = matrix_bd_43_cast_fp16)[name = tensor("op_2158_cast_fp16")]; - tensor _inversed_scores_41_y_0_to_fp16 = const()[name = tensor("_inversed_scores_41_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_41_cast_fp16 = mul(x = var_2158_cast_fp16, y = _inversed_scores_41_y_0_to_fp16)[name = tensor("_inversed_scores_41_cast_fp16")]; - tensor scores_43_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_41_cast_fp16, cond = mask_11)[name = tensor("scores_43_cast_fp16")]; - tensor var_2164_cast_fp16 = softmax(axis = var_23, x = scores_43_cast_fp16)[name = tensor("op_2164_cast_fp16")]; - tensor input_553_cast_fp16 = select(a = var_11_to_fp16, b = var_2164_cast_fp16, cond = mask_11)[name = tensor("input_553_cast_fp16")]; - tensor x_233_transpose_x_0 = const()[name = tensor("x_233_transpose_x_0"), val = tensor(false)]; - tensor x_233_transpose_y_0 = const()[name = tensor("x_233_transpose_y_0"), val = tensor(false)]; - tensor value_23_cast_fp16 = transpose(perm = value_23_perm_0, x = v_21_cast_fp16)[name = tensor("transpose_134")]; - tensor x_233_cast_fp16 = matmul(transpose_x = x_233_transpose_x_0, transpose_y = x_233_transpose_y_0, x = input_553_cast_fp16, y = value_23_cast_fp16)[name = tensor("x_233_cast_fp16")]; - tensor var_2168_perm_0 = const()[name = tensor("op_2168_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_2169 = const()[name = tensor("op_2169"), val = tensor([1, -1, 512])]; - tensor var_2168_cast_fp16 = transpose(perm = var_2168_perm_0, x = x_233_cast_fp16)[name = tensor("transpose_130")]; - tensor input_555_cast_fp16 = reshape(shape = var_2169, x = var_2168_cast_fp16)[name = tensor("input_555_cast_fp16")]; - tensor module_layers_10_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_10_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133916416)))]; - tensor module_layers_10_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_10_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134440768)))]; - tensor linear_97_cast_fp16 = linear(bias = module_layers_10_self_attn_linear_out_bias_to_fp16, weight = module_layers_10_self_attn_linear_out_weight_to_fp16, x = input_555_cast_fp16)[name = tensor("linear_97_cast_fp16")]; - tensor input_559_cast_fp16 = add(x = input_551_cast_fp16, y = linear_97_cast_fp16)[name = tensor("input_559_cast_fp16")]; - tensor x_237_axes_0 = const()[name = tensor("x_237_axes_0"), val = tensor([-1])]; - tensor module_layers_10_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_10_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134441856)))]; - tensor module_layers_10_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_10_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134442944)))]; - tensor x_237_cast_fp16 = layer_norm(axes = x_237_axes_0, beta = module_layers_10_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_10_norm_conv_weight_to_fp16, x = input_559_cast_fp16)[name = tensor("x_237_cast_fp16")]; - tensor input_561_perm_0 = const()[name = tensor("input_561_perm_0"), val = tensor([0, 2, 1])]; - tensor input_563_pad_type_0 = const()[name = tensor("input_563_pad_type_0"), val = tensor("valid")]; - tensor input_563_strides_0 = const()[name = tensor("input_563_strides_0"), val = tensor([1])]; - tensor input_563_pad_0 = const()[name = tensor("input_563_pad_0"), val = tensor([0, 0])]; - tensor input_563_dilations_0 = const()[name = tensor("input_563_dilations_0"), val = tensor([1])]; - tensor input_563_groups_0 = const()[name = tensor("input_563_groups_0"), val = tensor(1)]; - tensor module_layers_10_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_10_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134444032)))]; - tensor module_layers_10_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_10_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(135492672)))]; - tensor input_561_cast_fp16 = transpose(perm = input_561_perm_0, x = x_237_cast_fp16)[name = tensor("transpose_129")]; - tensor input_563_cast_fp16 = conv(bias = module_layers_10_conv_pointwise_conv1_bias_to_fp16, dilations = input_563_dilations_0, groups = input_563_groups_0, pad = input_563_pad_0, pad_type = input_563_pad_type_0, strides = input_563_strides_0, weight = module_layers_10_conv_pointwise_conv1_weight_to_fp16, x = input_561_cast_fp16)[name = tensor("input_563_cast_fp16")]; - tensor x_239_split_num_splits_0 = const()[name = tensor("x_239_split_num_splits_0"), val = tensor(2)]; - tensor x_239_split_axis_0 = const()[name = tensor("x_239_split_axis_0"), val = tensor(1)]; - tensor x_239_split_cast_fp16_0, tensor x_239_split_cast_fp16_1 = split(axis = x_239_split_axis_0, num_splits = x_239_split_num_splits_0, x = input_563_cast_fp16)[name = tensor("x_239_split_cast_fp16")]; - tensor x_239_split_1_sigmoid_cast_fp16 = sigmoid(x = x_239_split_cast_fp16_1)[name = tensor("x_239_split_1_sigmoid_cast_fp16")]; - tensor x_239_cast_fp16 = mul(x = x_239_split_cast_fp16_0, y = x_239_split_1_sigmoid_cast_fp16)[name = tensor("x_239_cast_fp16")]; - tensor input_565_cast_fp16 = select(a = var_11_to_fp16, b = x_239_cast_fp16, cond = var_453)[name = tensor("input_565_cast_fp16")]; - tensor input_567_pad_0 = const()[name = tensor("input_567_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_567_mode_0 = const()[name = tensor("input_567_mode_0"), val = tensor("constant")]; - tensor const_173_to_fp16 = const()[name = tensor("const_173_to_fp16"), val = tensor(0x0p+0)]; - tensor input_567_cast_fp16 = pad(constant_val = const_173_to_fp16, mode = input_567_mode_0, pad = input_567_pad_0, x = input_565_cast_fp16)[name = tensor("input_567_cast_fp16")]; - tensor input_569_pad_type_0 = const()[name = tensor("input_569_pad_type_0"), val = tensor("valid")]; - tensor input_569_groups_0 = const()[name = tensor("input_569_groups_0"), val = tensor(512)]; - tensor input_569_strides_0 = const()[name = tensor("input_569_strides_0"), val = tensor([1])]; - tensor input_569_pad_0 = const()[name = tensor("input_569_pad_0"), val = tensor([0, 0])]; - tensor input_569_dilations_0 = const()[name = tensor("input_569_dilations_0"), val = tensor([1])]; - tensor const_254_to_fp16 = const()[name = tensor("const_254_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(135494784)))]; - tensor const_255_to_fp16 = const()[name = tensor("const_255_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(135504064)))]; - tensor input_571_cast_fp16 = conv(bias = const_255_to_fp16, dilations = input_569_dilations_0, groups = input_569_groups_0, pad = input_569_pad_0, pad_type = input_569_pad_type_0, strides = input_569_strides_0, weight = const_254_to_fp16, x = input_567_cast_fp16)[name = tensor("input_571_cast_fp16")]; - tensor input_573_cast_fp16 = silu(x = input_571_cast_fp16)[name = tensor("input_573_cast_fp16")]; - tensor x_241_pad_type_0 = const()[name = tensor("x_241_pad_type_0"), val = tensor("valid")]; - tensor x_241_strides_0 = const()[name = tensor("x_241_strides_0"), val = tensor([1])]; - tensor x_241_pad_0 = const()[name = tensor("x_241_pad_0"), val = tensor([0, 0])]; - tensor x_241_dilations_0 = const()[name = tensor("x_241_dilations_0"), val = tensor([1])]; - tensor x_241_groups_0 = const()[name = tensor("x_241_groups_0"), val = tensor(1)]; - tensor module_layers_10_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_10_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(135505152)))]; - tensor module_layers_10_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_10_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136029504)))]; - tensor x_241_cast_fp16 = conv(bias = module_layers_10_conv_pointwise_conv2_bias_to_fp16, dilations = x_241_dilations_0, groups = x_241_groups_0, pad = x_241_pad_0, pad_type = x_241_pad_type_0, strides = x_241_strides_0, weight = module_layers_10_conv_pointwise_conv2_weight_to_fp16, x = input_573_cast_fp16)[name = tensor("x_241_cast_fp16")]; - tensor input_575_perm_0 = const()[name = tensor("input_575_perm_0"), val = tensor([0, 2, 1])]; - tensor input_575_cast_fp16 = transpose(perm = input_575_perm_0, x = x_241_cast_fp16)[name = tensor("transpose_128")]; - tensor input_577_cast_fp16 = add(x = input_559_cast_fp16, y = input_575_cast_fp16)[name = tensor("input_577_cast_fp16")]; - tensor input_579_axes_0 = const()[name = tensor("input_579_axes_0"), val = tensor([-1])]; - tensor module_layers_10_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_10_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136030592)))]; - tensor module_layers_10_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_10_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136031680)))]; - tensor input_579_cast_fp16 = layer_norm(axes = input_579_axes_0, beta = module_layers_10_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_10_norm_feed_forward2_weight_to_fp16, x = input_577_cast_fp16)[name = tensor("input_579_cast_fp16")]; - tensor module_layers_10_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_10_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136032768)))]; - tensor module_layers_10_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_10_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(138129984)))]; - tensor linear_98_cast_fp16 = linear(bias = module_layers_10_feed_forward2_linear1_bias_to_fp16, weight = module_layers_10_feed_forward2_linear1_weight_to_fp16, x = input_579_cast_fp16)[name = tensor("linear_98_cast_fp16")]; - tensor input_583_cast_fp16 = silu(x = linear_98_cast_fp16)[name = tensor("input_583_cast_fp16")]; - tensor module_layers_10_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_10_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(138134144)))]; - tensor module_layers_10_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_10_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140231360)))]; - tensor linear_99_cast_fp16 = linear(bias = module_layers_10_feed_forward2_linear2_bias_to_fp16, weight = module_layers_10_feed_forward2_linear2_weight_to_fp16, x = input_583_cast_fp16)[name = tensor("linear_99_cast_fp16")]; - tensor var_2235_to_fp16 = const()[name = tensor("op_2235_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2236_cast_fp16 = mul(x = linear_99_cast_fp16, y = var_2235_to_fp16)[name = tensor("op_2236_cast_fp16")]; - tensor input_589_cast_fp16 = add(x = input_577_cast_fp16, y = var_2236_cast_fp16)[name = tensor("input_589_cast_fp16")]; - tensor input_591_axes_0 = const()[name = tensor("input_591_axes_0"), val = tensor([-1])]; - tensor module_layers_10_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_10_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140232448)))]; - tensor module_layers_10_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_10_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140233536)))]; - tensor input_591_cast_fp16 = layer_norm(axes = input_591_axes_0, beta = module_layers_10_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_10_norm_out_weight_to_fp16, x = input_589_cast_fp16)[name = tensor("input_591_cast_fp16")]; - tensor input_593_axes_0 = const()[name = tensor("input_593_axes_0"), val = tensor([-1])]; - tensor module_layers_11_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_11_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140234624)))]; - tensor module_layers_11_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_11_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140235712)))]; - tensor input_593_cast_fp16 = layer_norm(axes = input_593_axes_0, beta = module_layers_11_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_11_norm_feed_forward1_weight_to_fp16, x = input_591_cast_fp16)[name = tensor("input_593_cast_fp16")]; - tensor module_layers_11_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_11_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140236800)))]; - tensor module_layers_11_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_11_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(142334016)))]; - tensor linear_100_cast_fp16 = linear(bias = module_layers_11_feed_forward1_linear1_bias_to_fp16, weight = module_layers_11_feed_forward1_linear1_weight_to_fp16, x = input_593_cast_fp16)[name = tensor("linear_100_cast_fp16")]; - tensor input_597_cast_fp16 = silu(x = linear_100_cast_fp16)[name = tensor("input_597_cast_fp16")]; - tensor module_layers_11_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_11_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(142338176)))]; - tensor module_layers_11_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_11_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(144435392)))]; - tensor linear_101_cast_fp16 = linear(bias = module_layers_11_feed_forward1_linear2_bias_to_fp16, weight = module_layers_11_feed_forward1_linear2_weight_to_fp16, x = input_597_cast_fp16)[name = tensor("linear_101_cast_fp16")]; - tensor var_2266_to_fp16 = const()[name = tensor("op_2266_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2267_cast_fp16 = mul(x = linear_101_cast_fp16, y = var_2266_to_fp16)[name = tensor("op_2267_cast_fp16")]; - tensor input_603_cast_fp16 = add(x = input_591_cast_fp16, y = var_2267_cast_fp16)[name = tensor("input_603_cast_fp16")]; - tensor query_23_axes_0 = const()[name = tensor("query_23_axes_0"), val = tensor([-1])]; - tensor module_layers_11_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_11_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(144436480)))]; - tensor module_layers_11_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_11_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(144437568)))]; - tensor query_23_cast_fp16 = layer_norm(axes = query_23_axes_0, beta = module_layers_11_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_11_norm_self_att_weight_to_fp16, x = input_603_cast_fp16)[name = tensor("query_23_cast_fp16")]; - tensor module_layers_11_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_11_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(144438656)))]; - tensor module_layers_11_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_11_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(144963008)))]; - tensor linear_102_cast_fp16 = linear(bias = module_layers_11_self_attn_linear_q_bias_to_fp16, weight = module_layers_11_self_attn_linear_q_weight_to_fp16, x = query_23_cast_fp16)[name = tensor("linear_102_cast_fp16")]; - tensor var_2284 = const()[name = tensor("op_2284"), val = tensor([1, -1, 8, 64])]; - tensor q_67_cast_fp16 = reshape(shape = var_2284, x = linear_102_cast_fp16)[name = tensor("q_67_cast_fp16")]; - tensor module_layers_11_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_11_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(144964096)))]; - tensor module_layers_11_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_11_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145488448)))]; - tensor linear_103_cast_fp16 = linear(bias = module_layers_11_self_attn_linear_k_bias_to_fp16, weight = module_layers_11_self_attn_linear_k_weight_to_fp16, x = query_23_cast_fp16)[name = tensor("linear_103_cast_fp16")]; - tensor var_2289 = const()[name = tensor("op_2289"), val = tensor([1, -1, 8, 64])]; - tensor k_45_cast_fp16 = reshape(shape = var_2289, x = linear_103_cast_fp16)[name = tensor("k_45_cast_fp16")]; - tensor module_layers_11_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_11_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145489536)))]; - tensor module_layers_11_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_11_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146013888)))]; - tensor linear_104_cast_fp16 = linear(bias = module_layers_11_self_attn_linear_v_bias_to_fp16, weight = module_layers_11_self_attn_linear_v_weight_to_fp16, x = query_23_cast_fp16)[name = tensor("linear_104_cast_fp16")]; - tensor var_2294 = const()[name = tensor("op_2294"), val = tensor([1, -1, 8, 64])]; - tensor v_23_cast_fp16 = reshape(shape = var_2294, x = linear_104_cast_fp16)[name = tensor("v_23_cast_fp16")]; - tensor value_25_perm_0 = const()[name = tensor("value_25_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_11_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_11_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146014976)))]; - tensor var_2306_cast_fp16 = add(x = q_67_cast_fp16, y = module_layers_11_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2306_cast_fp16")]; - tensor module_layers_11_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_11_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146016064)))]; - tensor var_2308_cast_fp16 = add(x = q_67_cast_fp16, y = module_layers_11_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2308_cast_fp16")]; - tensor q_with_bias_v_23_perm_0 = const()[name = tensor("q_with_bias_v_23_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_249_transpose_x_0 = const()[name = tensor("x_249_transpose_x_0"), val = tensor(false)]; - tensor x_249_transpose_y_0 = const()[name = tensor("x_249_transpose_y_0"), val = tensor(false)]; - tensor var_2310_to_fp16 = const()[name = tensor("op_2310_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146017152)))]; - tensor q_with_bias_v_23_cast_fp16 = transpose(perm = q_with_bias_v_23_perm_0, x = var_2308_cast_fp16)[name = tensor("transpose_126")]; - tensor x_249_cast_fp16 = matmul(transpose_x = x_249_transpose_x_0, transpose_y = x_249_transpose_y_0, x = q_with_bias_v_23_cast_fp16, y = var_2310_to_fp16)[name = tensor("x_249_cast_fp16")]; - tensor x_251_pad_0 = const()[name = tensor("x_251_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_251_mode_0 = const()[name = tensor("x_251_mode_0"), val = tensor("constant")]; - tensor const_180_to_fp16 = const()[name = tensor("const_180_to_fp16"), val = tensor(0x0p+0)]; - tensor x_251_cast_fp16 = pad(constant_val = const_180_to_fp16, mode = x_251_mode_0, pad = x_251_pad_0, x = x_249_cast_fp16)[name = tensor("x_251_cast_fp16")]; - tensor var_2318 = const()[name = tensor("op_2318"), val = tensor([1, 8, -1, 188])]; - tensor x_253_cast_fp16 = reshape(shape = var_2318, x = x_251_cast_fp16)[name = tensor("x_253_cast_fp16")]; - tensor var_2322_begin_0 = const()[name = tensor("op_2322_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_2322_end_0 = const()[name = tensor("op_2322_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_2322_end_mask_0 = const()[name = tensor("op_2322_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_2322_cast_fp16 = slice_by_index(begin = var_2322_begin_0, end = var_2322_end_0, end_mask = var_2322_end_mask_0, x = x_253_cast_fp16)[name = tensor("op_2322_cast_fp16")]; - tensor var_2323 = const()[name = tensor("op_2323"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_45_cast_fp16 = reshape(shape = var_2323, x = var_2322_cast_fp16)[name = tensor("matrix_bd_45_cast_fp16")]; - tensor matrix_ac_23_transpose_x_0 = const()[name = tensor("matrix_ac_23_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_23_transpose_y_0 = const()[name = tensor("matrix_ac_23_transpose_y_0"), val = tensor(false)]; - tensor transpose_73_perm_0 = const()[name = tensor("transpose_73_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_74_perm_0 = const()[name = tensor("transpose_74_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_74 = transpose(perm = transpose_74_perm_0, x = k_45_cast_fp16)[name = tensor("transpose_124")]; - tensor transpose_73 = transpose(perm = transpose_73_perm_0, x = var_2306_cast_fp16)[name = tensor("transpose_125")]; - tensor matrix_ac_23_cast_fp16 = matmul(transpose_x = matrix_ac_23_transpose_x_0, transpose_y = matrix_ac_23_transpose_y_0, x = transpose_73, y = transpose_74)[name = tensor("matrix_ac_23_cast_fp16")]; - tensor matrix_bd_47_begin_0 = const()[name = tensor("matrix_bd_47_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_47_end_0 = const()[name = tensor("matrix_bd_47_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_47_end_mask_0 = const()[name = tensor("matrix_bd_47_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_47_cast_fp16 = slice_by_index(begin = matrix_bd_47_begin_0, end = matrix_bd_47_end_0, end_mask = matrix_bd_47_end_mask_0, x = matrix_bd_45_cast_fp16)[name = tensor("matrix_bd_47_cast_fp16")]; - tensor var_2332_cast_fp16 = add(x = matrix_ac_23_cast_fp16, y = matrix_bd_47_cast_fp16)[name = tensor("op_2332_cast_fp16")]; - tensor _inversed_scores_45_y_0_to_fp16 = const()[name = tensor("_inversed_scores_45_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_45_cast_fp16 = mul(x = var_2332_cast_fp16, y = _inversed_scores_45_y_0_to_fp16)[name = tensor("_inversed_scores_45_cast_fp16")]; - tensor scores_47_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_45_cast_fp16, cond = mask_11)[name = tensor("scores_47_cast_fp16")]; - tensor var_2338_cast_fp16 = softmax(axis = var_23, x = scores_47_cast_fp16)[name = tensor("op_2338_cast_fp16")]; - tensor input_605_cast_fp16 = select(a = var_11_to_fp16, b = var_2338_cast_fp16, cond = mask_11)[name = tensor("input_605_cast_fp16")]; - tensor x_255_transpose_x_0 = const()[name = tensor("x_255_transpose_x_0"), val = tensor(false)]; - tensor x_255_transpose_y_0 = const()[name = tensor("x_255_transpose_y_0"), val = tensor(false)]; - tensor value_25_cast_fp16 = transpose(perm = value_25_perm_0, x = v_23_cast_fp16)[name = tensor("transpose_127")]; - tensor x_255_cast_fp16 = matmul(transpose_x = x_255_transpose_x_0, transpose_y = x_255_transpose_y_0, x = input_605_cast_fp16, y = value_25_cast_fp16)[name = tensor("x_255_cast_fp16")]; - tensor var_2342_perm_0 = const()[name = tensor("op_2342_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_2343 = const()[name = tensor("op_2343"), val = tensor([1, -1, 512])]; - tensor var_2342_cast_fp16 = transpose(perm = var_2342_perm_0, x = x_255_cast_fp16)[name = tensor("transpose_123")]; - tensor input_607_cast_fp16 = reshape(shape = var_2343, x = var_2342_cast_fp16)[name = tensor("input_607_cast_fp16")]; - tensor module_layers_11_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_11_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146401216)))]; - tensor module_layers_11_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_11_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146925568)))]; - tensor linear_106_cast_fp16 = linear(bias = module_layers_11_self_attn_linear_out_bias_to_fp16, weight = module_layers_11_self_attn_linear_out_weight_to_fp16, x = input_607_cast_fp16)[name = tensor("linear_106_cast_fp16")]; - tensor input_611_cast_fp16 = add(x = input_603_cast_fp16, y = linear_106_cast_fp16)[name = tensor("input_611_cast_fp16")]; - tensor x_259_axes_0 = const()[name = tensor("x_259_axes_0"), val = tensor([-1])]; - tensor module_layers_11_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_11_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146926656)))]; - tensor module_layers_11_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_11_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146927744)))]; - tensor x_259_cast_fp16 = layer_norm(axes = x_259_axes_0, beta = module_layers_11_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_11_norm_conv_weight_to_fp16, x = input_611_cast_fp16)[name = tensor("x_259_cast_fp16")]; - tensor input_613_perm_0 = const()[name = tensor("input_613_perm_0"), val = tensor([0, 2, 1])]; - tensor input_615_pad_type_0 = const()[name = tensor("input_615_pad_type_0"), val = tensor("valid")]; - tensor input_615_strides_0 = const()[name = tensor("input_615_strides_0"), val = tensor([1])]; - tensor input_615_pad_0 = const()[name = tensor("input_615_pad_0"), val = tensor([0, 0])]; - tensor input_615_dilations_0 = const()[name = tensor("input_615_dilations_0"), val = tensor([1])]; - tensor input_615_groups_0 = const()[name = tensor("input_615_groups_0"), val = tensor(1)]; - tensor module_layers_11_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_11_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146928832)))]; - tensor module_layers_11_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_11_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147977472)))]; - tensor input_613_cast_fp16 = transpose(perm = input_613_perm_0, x = x_259_cast_fp16)[name = tensor("transpose_122")]; - tensor input_615_cast_fp16 = conv(bias = module_layers_11_conv_pointwise_conv1_bias_to_fp16, dilations = input_615_dilations_0, groups = input_615_groups_0, pad = input_615_pad_0, pad_type = input_615_pad_type_0, strides = input_615_strides_0, weight = module_layers_11_conv_pointwise_conv1_weight_to_fp16, x = input_613_cast_fp16)[name = tensor("input_615_cast_fp16")]; - tensor x_261_split_num_splits_0 = const()[name = tensor("x_261_split_num_splits_0"), val = tensor(2)]; - tensor x_261_split_axis_0 = const()[name = tensor("x_261_split_axis_0"), val = tensor(1)]; - tensor x_261_split_cast_fp16_0, tensor x_261_split_cast_fp16_1 = split(axis = x_261_split_axis_0, num_splits = x_261_split_num_splits_0, x = input_615_cast_fp16)[name = tensor("x_261_split_cast_fp16")]; - tensor x_261_split_1_sigmoid_cast_fp16 = sigmoid(x = x_261_split_cast_fp16_1)[name = tensor("x_261_split_1_sigmoid_cast_fp16")]; - tensor x_261_cast_fp16 = mul(x = x_261_split_cast_fp16_0, y = x_261_split_1_sigmoid_cast_fp16)[name = tensor("x_261_cast_fp16")]; - tensor input_617_cast_fp16 = select(a = var_11_to_fp16, b = x_261_cast_fp16, cond = var_453)[name = tensor("input_617_cast_fp16")]; - tensor input_619_pad_0 = const()[name = tensor("input_619_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_619_mode_0 = const()[name = tensor("input_619_mode_0"), val = tensor("constant")]; - tensor const_183_to_fp16 = const()[name = tensor("const_183_to_fp16"), val = tensor(0x0p+0)]; - tensor input_619_cast_fp16 = pad(constant_val = const_183_to_fp16, mode = input_619_mode_0, pad = input_619_pad_0, x = input_617_cast_fp16)[name = tensor("input_619_cast_fp16")]; - tensor input_621_pad_type_0 = const()[name = tensor("input_621_pad_type_0"), val = tensor("valid")]; - tensor input_621_groups_0 = const()[name = tensor("input_621_groups_0"), val = tensor(512)]; - tensor input_621_strides_0 = const()[name = tensor("input_621_strides_0"), val = tensor([1])]; - tensor input_621_pad_0 = const()[name = tensor("input_621_pad_0"), val = tensor([0, 0])]; - tensor input_621_dilations_0 = const()[name = tensor("input_621_dilations_0"), val = tensor([1])]; - tensor const_256_to_fp16 = const()[name = tensor("const_256_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147979584)))]; - tensor const_257_to_fp16 = const()[name = tensor("const_257_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147988864)))]; - tensor input_623_cast_fp16 = conv(bias = const_257_to_fp16, dilations = input_621_dilations_0, groups = input_621_groups_0, pad = input_621_pad_0, pad_type = input_621_pad_type_0, strides = input_621_strides_0, weight = const_256_to_fp16, x = input_619_cast_fp16)[name = tensor("input_623_cast_fp16")]; - tensor input_625_cast_fp16 = silu(x = input_623_cast_fp16)[name = tensor("input_625_cast_fp16")]; - tensor x_263_pad_type_0 = const()[name = tensor("x_263_pad_type_0"), val = tensor("valid")]; - tensor x_263_strides_0 = const()[name = tensor("x_263_strides_0"), val = tensor([1])]; - tensor x_263_pad_0 = const()[name = tensor("x_263_pad_0"), val = tensor([0, 0])]; - tensor x_263_dilations_0 = const()[name = tensor("x_263_dilations_0"), val = tensor([1])]; - tensor x_263_groups_0 = const()[name = tensor("x_263_groups_0"), val = tensor(1)]; - tensor module_layers_11_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_11_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147989952)))]; - tensor module_layers_11_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_11_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(148514304)))]; - tensor x_263_cast_fp16 = conv(bias = module_layers_11_conv_pointwise_conv2_bias_to_fp16, dilations = x_263_dilations_0, groups = x_263_groups_0, pad = x_263_pad_0, pad_type = x_263_pad_type_0, strides = x_263_strides_0, weight = module_layers_11_conv_pointwise_conv2_weight_to_fp16, x = input_625_cast_fp16)[name = tensor("x_263_cast_fp16")]; - tensor input_627_perm_0 = const()[name = tensor("input_627_perm_0"), val = tensor([0, 2, 1])]; - tensor input_627_cast_fp16 = transpose(perm = input_627_perm_0, x = x_263_cast_fp16)[name = tensor("transpose_121")]; - tensor input_629_cast_fp16 = add(x = input_611_cast_fp16, y = input_627_cast_fp16)[name = tensor("input_629_cast_fp16")]; - tensor input_631_axes_0 = const()[name = tensor("input_631_axes_0"), val = tensor([-1])]; - tensor module_layers_11_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_11_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(148515392)))]; - tensor module_layers_11_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_11_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(148516480)))]; - tensor input_631_cast_fp16 = layer_norm(axes = input_631_axes_0, beta = module_layers_11_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_11_norm_feed_forward2_weight_to_fp16, x = input_629_cast_fp16)[name = tensor("input_631_cast_fp16")]; - tensor module_layers_11_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_11_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(148517568)))]; - tensor module_layers_11_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_11_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(150614784)))]; - tensor linear_107_cast_fp16 = linear(bias = module_layers_11_feed_forward2_linear1_bias_to_fp16, weight = module_layers_11_feed_forward2_linear1_weight_to_fp16, x = input_631_cast_fp16)[name = tensor("linear_107_cast_fp16")]; - tensor input_635_cast_fp16 = silu(x = linear_107_cast_fp16)[name = tensor("input_635_cast_fp16")]; - tensor module_layers_11_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_11_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(150618944)))]; - tensor module_layers_11_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_11_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(152716160)))]; - tensor linear_108_cast_fp16 = linear(bias = module_layers_11_feed_forward2_linear2_bias_to_fp16, weight = module_layers_11_feed_forward2_linear2_weight_to_fp16, x = input_635_cast_fp16)[name = tensor("linear_108_cast_fp16")]; - tensor var_2409_to_fp16 = const()[name = tensor("op_2409_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2410_cast_fp16 = mul(x = linear_108_cast_fp16, y = var_2409_to_fp16)[name = tensor("op_2410_cast_fp16")]; - tensor input_641_cast_fp16 = add(x = input_629_cast_fp16, y = var_2410_cast_fp16)[name = tensor("input_641_cast_fp16")]; - tensor input_643_axes_0 = const()[name = tensor("input_643_axes_0"), val = tensor([-1])]; - tensor module_layers_11_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_11_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(152717248)))]; - tensor module_layers_11_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_11_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(152718336)))]; - tensor input_643_cast_fp16 = layer_norm(axes = input_643_axes_0, beta = module_layers_11_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_11_norm_out_weight_to_fp16, x = input_641_cast_fp16)[name = tensor("input_643_cast_fp16")]; - tensor input_645_axes_0 = const()[name = tensor("input_645_axes_0"), val = tensor([-1])]; - tensor module_layers_12_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_12_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(152719424)))]; - tensor module_layers_12_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_12_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(152720512)))]; - tensor input_645_cast_fp16 = layer_norm(axes = input_645_axes_0, beta = module_layers_12_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_12_norm_feed_forward1_weight_to_fp16, x = input_643_cast_fp16)[name = tensor("input_645_cast_fp16")]; - tensor module_layers_12_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_12_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(152721600)))]; - tensor module_layers_12_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_12_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(154818816)))]; - tensor linear_109_cast_fp16 = linear(bias = module_layers_12_feed_forward1_linear1_bias_to_fp16, weight = module_layers_12_feed_forward1_linear1_weight_to_fp16, x = input_645_cast_fp16)[name = tensor("linear_109_cast_fp16")]; - tensor input_649_cast_fp16 = silu(x = linear_109_cast_fp16)[name = tensor("input_649_cast_fp16")]; - tensor module_layers_12_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_12_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(154822976)))]; - tensor module_layers_12_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_12_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(156920192)))]; - tensor linear_110_cast_fp16 = linear(bias = module_layers_12_feed_forward1_linear2_bias_to_fp16, weight = module_layers_12_feed_forward1_linear2_weight_to_fp16, x = input_649_cast_fp16)[name = tensor("linear_110_cast_fp16")]; - tensor var_2440_to_fp16 = const()[name = tensor("op_2440_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2441_cast_fp16 = mul(x = linear_110_cast_fp16, y = var_2440_to_fp16)[name = tensor("op_2441_cast_fp16")]; - tensor input_655_cast_fp16 = add(x = input_643_cast_fp16, y = var_2441_cast_fp16)[name = tensor("input_655_cast_fp16")]; - tensor query_25_axes_0 = const()[name = tensor("query_25_axes_0"), val = tensor([-1])]; - tensor module_layers_12_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_12_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(156921280)))]; - tensor module_layers_12_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_12_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(156922368)))]; - tensor query_25_cast_fp16 = layer_norm(axes = query_25_axes_0, beta = module_layers_12_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_12_norm_self_att_weight_to_fp16, x = input_655_cast_fp16)[name = tensor("query_25_cast_fp16")]; - tensor module_layers_12_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_12_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(156923456)))]; - tensor module_layers_12_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_12_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157447808)))]; - tensor linear_111_cast_fp16 = linear(bias = module_layers_12_self_attn_linear_q_bias_to_fp16, weight = module_layers_12_self_attn_linear_q_weight_to_fp16, x = query_25_cast_fp16)[name = tensor("linear_111_cast_fp16")]; - tensor var_2458 = const()[name = tensor("op_2458"), val = tensor([1, -1, 8, 64])]; - tensor q_73_cast_fp16 = reshape(shape = var_2458, x = linear_111_cast_fp16)[name = tensor("q_73_cast_fp16")]; - tensor module_layers_12_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_12_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157448896)))]; - tensor module_layers_12_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_12_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157973248)))]; - tensor linear_112_cast_fp16 = linear(bias = module_layers_12_self_attn_linear_k_bias_to_fp16, weight = module_layers_12_self_attn_linear_k_weight_to_fp16, x = query_25_cast_fp16)[name = tensor("linear_112_cast_fp16")]; - tensor var_2463 = const()[name = tensor("op_2463"), val = tensor([1, -1, 8, 64])]; - tensor k_49_cast_fp16 = reshape(shape = var_2463, x = linear_112_cast_fp16)[name = tensor("k_49_cast_fp16")]; - tensor module_layers_12_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_12_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157974336)))]; - tensor module_layers_12_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_12_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158498688)))]; - tensor linear_113_cast_fp16 = linear(bias = module_layers_12_self_attn_linear_v_bias_to_fp16, weight = module_layers_12_self_attn_linear_v_weight_to_fp16, x = query_25_cast_fp16)[name = tensor("linear_113_cast_fp16")]; - tensor var_2468 = const()[name = tensor("op_2468"), val = tensor([1, -1, 8, 64])]; - tensor v_25_cast_fp16 = reshape(shape = var_2468, x = linear_113_cast_fp16)[name = tensor("v_25_cast_fp16")]; - tensor value_27_perm_0 = const()[name = tensor("value_27_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_12_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_12_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158499776)))]; - tensor var_2480_cast_fp16 = add(x = q_73_cast_fp16, y = module_layers_12_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2480_cast_fp16")]; - tensor module_layers_12_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_12_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158500864)))]; - tensor var_2482_cast_fp16 = add(x = q_73_cast_fp16, y = module_layers_12_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2482_cast_fp16")]; - tensor q_with_bias_v_25_perm_0 = const()[name = tensor("q_with_bias_v_25_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_271_transpose_x_0 = const()[name = tensor("x_271_transpose_x_0"), val = tensor(false)]; - tensor x_271_transpose_y_0 = const()[name = tensor("x_271_transpose_y_0"), val = tensor(false)]; - tensor var_2484_to_fp16 = const()[name = tensor("op_2484_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158501952)))]; - tensor q_with_bias_v_25_cast_fp16 = transpose(perm = q_with_bias_v_25_perm_0, x = var_2482_cast_fp16)[name = tensor("transpose_119")]; - tensor x_271_cast_fp16 = matmul(transpose_x = x_271_transpose_x_0, transpose_y = x_271_transpose_y_0, x = q_with_bias_v_25_cast_fp16, y = var_2484_to_fp16)[name = tensor("x_271_cast_fp16")]; - tensor x_273_pad_0 = const()[name = tensor("x_273_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_273_mode_0 = const()[name = tensor("x_273_mode_0"), val = tensor("constant")]; - tensor const_190_to_fp16 = const()[name = tensor("const_190_to_fp16"), val = tensor(0x0p+0)]; - tensor x_273_cast_fp16 = pad(constant_val = const_190_to_fp16, mode = x_273_mode_0, pad = x_273_pad_0, x = x_271_cast_fp16)[name = tensor("x_273_cast_fp16")]; - tensor var_2492 = const()[name = tensor("op_2492"), val = tensor([1, 8, -1, 188])]; - tensor x_275_cast_fp16 = reshape(shape = var_2492, x = x_273_cast_fp16)[name = tensor("x_275_cast_fp16")]; - tensor var_2496_begin_0 = const()[name = tensor("op_2496_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_2496_end_0 = const()[name = tensor("op_2496_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_2496_end_mask_0 = const()[name = tensor("op_2496_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_2496_cast_fp16 = slice_by_index(begin = var_2496_begin_0, end = var_2496_end_0, end_mask = var_2496_end_mask_0, x = x_275_cast_fp16)[name = tensor("op_2496_cast_fp16")]; - tensor var_2497 = const()[name = tensor("op_2497"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_49_cast_fp16 = reshape(shape = var_2497, x = var_2496_cast_fp16)[name = tensor("matrix_bd_49_cast_fp16")]; - tensor matrix_ac_25_transpose_x_0 = const()[name = tensor("matrix_ac_25_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_25_transpose_y_0 = const()[name = tensor("matrix_ac_25_transpose_y_0"), val = tensor(false)]; - tensor transpose_75_perm_0 = const()[name = tensor("transpose_75_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_76_perm_0 = const()[name = tensor("transpose_76_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_76 = transpose(perm = transpose_76_perm_0, x = k_49_cast_fp16)[name = tensor("transpose_117")]; - tensor transpose_75 = transpose(perm = transpose_75_perm_0, x = var_2480_cast_fp16)[name = tensor("transpose_118")]; - tensor matrix_ac_25_cast_fp16 = matmul(transpose_x = matrix_ac_25_transpose_x_0, transpose_y = matrix_ac_25_transpose_y_0, x = transpose_75, y = transpose_76)[name = tensor("matrix_ac_25_cast_fp16")]; - tensor matrix_bd_51_begin_0 = const()[name = tensor("matrix_bd_51_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_51_end_0 = const()[name = tensor("matrix_bd_51_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_51_end_mask_0 = const()[name = tensor("matrix_bd_51_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_51_cast_fp16 = slice_by_index(begin = matrix_bd_51_begin_0, end = matrix_bd_51_end_0, end_mask = matrix_bd_51_end_mask_0, x = matrix_bd_49_cast_fp16)[name = tensor("matrix_bd_51_cast_fp16")]; - tensor var_2506_cast_fp16 = add(x = matrix_ac_25_cast_fp16, y = matrix_bd_51_cast_fp16)[name = tensor("op_2506_cast_fp16")]; - tensor _inversed_scores_49_y_0_to_fp16 = const()[name = tensor("_inversed_scores_49_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_49_cast_fp16 = mul(x = var_2506_cast_fp16, y = _inversed_scores_49_y_0_to_fp16)[name = tensor("_inversed_scores_49_cast_fp16")]; - tensor scores_51_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_49_cast_fp16, cond = mask_11)[name = tensor("scores_51_cast_fp16")]; - tensor var_2512_cast_fp16 = softmax(axis = var_23, x = scores_51_cast_fp16)[name = tensor("op_2512_cast_fp16")]; - tensor input_657_cast_fp16 = select(a = var_11_to_fp16, b = var_2512_cast_fp16, cond = mask_11)[name = tensor("input_657_cast_fp16")]; - tensor x_277_transpose_x_0 = const()[name = tensor("x_277_transpose_x_0"), val = tensor(false)]; - tensor x_277_transpose_y_0 = const()[name = tensor("x_277_transpose_y_0"), val = tensor(false)]; - tensor value_27_cast_fp16 = transpose(perm = value_27_perm_0, x = v_25_cast_fp16)[name = tensor("transpose_120")]; - tensor x_277_cast_fp16 = matmul(transpose_x = x_277_transpose_x_0, transpose_y = x_277_transpose_y_0, x = input_657_cast_fp16, y = value_27_cast_fp16)[name = tensor("x_277_cast_fp16")]; - tensor var_2516_perm_0 = const()[name = tensor("op_2516_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_2517 = const()[name = tensor("op_2517"), val = tensor([1, -1, 512])]; - tensor var_2516_cast_fp16 = transpose(perm = var_2516_perm_0, x = x_277_cast_fp16)[name = tensor("transpose_116")]; - tensor input_659_cast_fp16 = reshape(shape = var_2517, x = var_2516_cast_fp16)[name = tensor("input_659_cast_fp16")]; - tensor module_layers_12_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_12_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158886016)))]; - tensor module_layers_12_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_12_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159410368)))]; - tensor linear_115_cast_fp16 = linear(bias = module_layers_12_self_attn_linear_out_bias_to_fp16, weight = module_layers_12_self_attn_linear_out_weight_to_fp16, x = input_659_cast_fp16)[name = tensor("linear_115_cast_fp16")]; - tensor input_663_cast_fp16 = add(x = input_655_cast_fp16, y = linear_115_cast_fp16)[name = tensor("input_663_cast_fp16")]; - tensor x_281_axes_0 = const()[name = tensor("x_281_axes_0"), val = tensor([-1])]; - tensor module_layers_12_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_12_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159411456)))]; - tensor module_layers_12_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_12_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159412544)))]; - tensor x_281_cast_fp16 = layer_norm(axes = x_281_axes_0, beta = module_layers_12_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_12_norm_conv_weight_to_fp16, x = input_663_cast_fp16)[name = tensor("x_281_cast_fp16")]; - tensor input_665_perm_0 = const()[name = tensor("input_665_perm_0"), val = tensor([0, 2, 1])]; - tensor input_667_pad_type_0 = const()[name = tensor("input_667_pad_type_0"), val = tensor("valid")]; - tensor input_667_strides_0 = const()[name = tensor("input_667_strides_0"), val = tensor([1])]; - tensor input_667_pad_0 = const()[name = tensor("input_667_pad_0"), val = tensor([0, 0])]; - tensor input_667_dilations_0 = const()[name = tensor("input_667_dilations_0"), val = tensor([1])]; - tensor input_667_groups_0 = const()[name = tensor("input_667_groups_0"), val = tensor(1)]; - tensor module_layers_12_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_12_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159413632)))]; - tensor module_layers_12_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_12_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160462272)))]; - tensor input_665_cast_fp16 = transpose(perm = input_665_perm_0, x = x_281_cast_fp16)[name = tensor("transpose_115")]; - tensor input_667_cast_fp16 = conv(bias = module_layers_12_conv_pointwise_conv1_bias_to_fp16, dilations = input_667_dilations_0, groups = input_667_groups_0, pad = input_667_pad_0, pad_type = input_667_pad_type_0, strides = input_667_strides_0, weight = module_layers_12_conv_pointwise_conv1_weight_to_fp16, x = input_665_cast_fp16)[name = tensor("input_667_cast_fp16")]; - tensor x_283_split_num_splits_0 = const()[name = tensor("x_283_split_num_splits_0"), val = tensor(2)]; - tensor x_283_split_axis_0 = const()[name = tensor("x_283_split_axis_0"), val = tensor(1)]; - tensor x_283_split_cast_fp16_0, tensor x_283_split_cast_fp16_1 = split(axis = x_283_split_axis_0, num_splits = x_283_split_num_splits_0, x = input_667_cast_fp16)[name = tensor("x_283_split_cast_fp16")]; - tensor x_283_split_1_sigmoid_cast_fp16 = sigmoid(x = x_283_split_cast_fp16_1)[name = tensor("x_283_split_1_sigmoid_cast_fp16")]; - tensor x_283_cast_fp16 = mul(x = x_283_split_cast_fp16_0, y = x_283_split_1_sigmoid_cast_fp16)[name = tensor("x_283_cast_fp16")]; - tensor input_669_cast_fp16 = select(a = var_11_to_fp16, b = x_283_cast_fp16, cond = var_453)[name = tensor("input_669_cast_fp16")]; - tensor input_671_pad_0 = const()[name = tensor("input_671_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_671_mode_0 = const()[name = tensor("input_671_mode_0"), val = tensor("constant")]; - tensor const_193_to_fp16 = const()[name = tensor("const_193_to_fp16"), val = tensor(0x0p+0)]; - tensor input_671_cast_fp16 = pad(constant_val = const_193_to_fp16, mode = input_671_mode_0, pad = input_671_pad_0, x = input_669_cast_fp16)[name = tensor("input_671_cast_fp16")]; - tensor input_673_pad_type_0 = const()[name = tensor("input_673_pad_type_0"), val = tensor("valid")]; - tensor input_673_groups_0 = const()[name = tensor("input_673_groups_0"), val = tensor(512)]; - tensor input_673_strides_0 = const()[name = tensor("input_673_strides_0"), val = tensor([1])]; - tensor input_673_pad_0 = const()[name = tensor("input_673_pad_0"), val = tensor([0, 0])]; - tensor input_673_dilations_0 = const()[name = tensor("input_673_dilations_0"), val = tensor([1])]; - tensor const_258_to_fp16 = const()[name = tensor("const_258_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160464384)))]; - tensor const_259_to_fp16 = const()[name = tensor("const_259_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160473664)))]; - tensor input_675_cast_fp16 = conv(bias = const_259_to_fp16, dilations = input_673_dilations_0, groups = input_673_groups_0, pad = input_673_pad_0, pad_type = input_673_pad_type_0, strides = input_673_strides_0, weight = const_258_to_fp16, x = input_671_cast_fp16)[name = tensor("input_675_cast_fp16")]; - tensor input_677_cast_fp16 = silu(x = input_675_cast_fp16)[name = tensor("input_677_cast_fp16")]; - tensor x_285_pad_type_0 = const()[name = tensor("x_285_pad_type_0"), val = tensor("valid")]; - tensor x_285_strides_0 = const()[name = tensor("x_285_strides_0"), val = tensor([1])]; - tensor x_285_pad_0 = const()[name = tensor("x_285_pad_0"), val = tensor([0, 0])]; - tensor x_285_dilations_0 = const()[name = tensor("x_285_dilations_0"), val = tensor([1])]; - tensor x_285_groups_0 = const()[name = tensor("x_285_groups_0"), val = tensor(1)]; - tensor module_layers_12_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_12_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160474752)))]; - tensor module_layers_12_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_12_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160999104)))]; - tensor x_285_cast_fp16 = conv(bias = module_layers_12_conv_pointwise_conv2_bias_to_fp16, dilations = x_285_dilations_0, groups = x_285_groups_0, pad = x_285_pad_0, pad_type = x_285_pad_type_0, strides = x_285_strides_0, weight = module_layers_12_conv_pointwise_conv2_weight_to_fp16, x = input_677_cast_fp16)[name = tensor("x_285_cast_fp16")]; - tensor input_679_perm_0 = const()[name = tensor("input_679_perm_0"), val = tensor([0, 2, 1])]; - tensor input_679_cast_fp16 = transpose(perm = input_679_perm_0, x = x_285_cast_fp16)[name = tensor("transpose_114")]; - tensor input_681_cast_fp16 = add(x = input_663_cast_fp16, y = input_679_cast_fp16)[name = tensor("input_681_cast_fp16")]; - tensor input_683_axes_0 = const()[name = tensor("input_683_axes_0"), val = tensor([-1])]; - tensor module_layers_12_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_12_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161000192)))]; - tensor module_layers_12_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_12_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161001280)))]; - tensor input_683_cast_fp16 = layer_norm(axes = input_683_axes_0, beta = module_layers_12_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_12_norm_feed_forward2_weight_to_fp16, x = input_681_cast_fp16)[name = tensor("input_683_cast_fp16")]; - tensor module_layers_12_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_12_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161002368)))]; - tensor module_layers_12_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_12_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(163099584)))]; - tensor linear_116_cast_fp16 = linear(bias = module_layers_12_feed_forward2_linear1_bias_to_fp16, weight = module_layers_12_feed_forward2_linear1_weight_to_fp16, x = input_683_cast_fp16)[name = tensor("linear_116_cast_fp16")]; - tensor input_687_cast_fp16 = silu(x = linear_116_cast_fp16)[name = tensor("input_687_cast_fp16")]; - tensor module_layers_12_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_12_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(163103744)))]; - tensor module_layers_12_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_12_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165200960)))]; - tensor linear_117_cast_fp16 = linear(bias = module_layers_12_feed_forward2_linear2_bias_to_fp16, weight = module_layers_12_feed_forward2_linear2_weight_to_fp16, x = input_687_cast_fp16)[name = tensor("linear_117_cast_fp16")]; - tensor var_2583_to_fp16 = const()[name = tensor("op_2583_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2584_cast_fp16 = mul(x = linear_117_cast_fp16, y = var_2583_to_fp16)[name = tensor("op_2584_cast_fp16")]; - tensor input_693_cast_fp16 = add(x = input_681_cast_fp16, y = var_2584_cast_fp16)[name = tensor("input_693_cast_fp16")]; - tensor input_695_axes_0 = const()[name = tensor("input_695_axes_0"), val = tensor([-1])]; - tensor module_layers_12_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_12_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165202048)))]; - tensor module_layers_12_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_12_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165203136)))]; - tensor input_695_cast_fp16 = layer_norm(axes = input_695_axes_0, beta = module_layers_12_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_12_norm_out_weight_to_fp16, x = input_693_cast_fp16)[name = tensor("input_695_cast_fp16")]; - tensor input_697_axes_0 = const()[name = tensor("input_697_axes_0"), val = tensor([-1])]; - tensor module_layers_13_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_13_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165204224)))]; - tensor module_layers_13_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_13_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165205312)))]; - tensor input_697_cast_fp16 = layer_norm(axes = input_697_axes_0, beta = module_layers_13_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_13_norm_feed_forward1_weight_to_fp16, x = input_695_cast_fp16)[name = tensor("input_697_cast_fp16")]; - tensor module_layers_13_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_13_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165206400)))]; - tensor module_layers_13_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_13_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(167303616)))]; - tensor linear_118_cast_fp16 = linear(bias = module_layers_13_feed_forward1_linear1_bias_to_fp16, weight = module_layers_13_feed_forward1_linear1_weight_to_fp16, x = input_697_cast_fp16)[name = tensor("linear_118_cast_fp16")]; - tensor input_701_cast_fp16 = silu(x = linear_118_cast_fp16)[name = tensor("input_701_cast_fp16")]; - tensor module_layers_13_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_13_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(167307776)))]; - tensor module_layers_13_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_13_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169404992)))]; - tensor linear_119_cast_fp16 = linear(bias = module_layers_13_feed_forward1_linear2_bias_to_fp16, weight = module_layers_13_feed_forward1_linear2_weight_to_fp16, x = input_701_cast_fp16)[name = tensor("linear_119_cast_fp16")]; - tensor var_2614_to_fp16 = const()[name = tensor("op_2614_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2615_cast_fp16 = mul(x = linear_119_cast_fp16, y = var_2614_to_fp16)[name = tensor("op_2615_cast_fp16")]; - tensor input_707_cast_fp16 = add(x = input_695_cast_fp16, y = var_2615_cast_fp16)[name = tensor("input_707_cast_fp16")]; - tensor query_27_axes_0 = const()[name = tensor("query_27_axes_0"), val = tensor([-1])]; - tensor module_layers_13_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_13_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169406080)))]; - tensor module_layers_13_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_13_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169407168)))]; - tensor query_27_cast_fp16 = layer_norm(axes = query_27_axes_0, beta = module_layers_13_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_13_norm_self_att_weight_to_fp16, x = input_707_cast_fp16)[name = tensor("query_27_cast_fp16")]; - tensor module_layers_13_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_13_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169408256)))]; - tensor module_layers_13_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_13_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169932608)))]; - tensor linear_120_cast_fp16 = linear(bias = module_layers_13_self_attn_linear_q_bias_to_fp16, weight = module_layers_13_self_attn_linear_q_weight_to_fp16, x = query_27_cast_fp16)[name = tensor("linear_120_cast_fp16")]; - tensor var_2632 = const()[name = tensor("op_2632"), val = tensor([1, -1, 8, 64])]; - tensor q_79_cast_fp16 = reshape(shape = var_2632, x = linear_120_cast_fp16)[name = tensor("q_79_cast_fp16")]; - tensor module_layers_13_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_13_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169933696)))]; - tensor module_layers_13_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_13_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170458048)))]; - tensor linear_121_cast_fp16 = linear(bias = module_layers_13_self_attn_linear_k_bias_to_fp16, weight = module_layers_13_self_attn_linear_k_weight_to_fp16, x = query_27_cast_fp16)[name = tensor("linear_121_cast_fp16")]; - tensor var_2637 = const()[name = tensor("op_2637"), val = tensor([1, -1, 8, 64])]; - tensor k_53_cast_fp16 = reshape(shape = var_2637, x = linear_121_cast_fp16)[name = tensor("k_53_cast_fp16")]; - tensor module_layers_13_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_13_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170459136)))]; - tensor module_layers_13_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_13_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170983488)))]; - tensor linear_122_cast_fp16 = linear(bias = module_layers_13_self_attn_linear_v_bias_to_fp16, weight = module_layers_13_self_attn_linear_v_weight_to_fp16, x = query_27_cast_fp16)[name = tensor("linear_122_cast_fp16")]; - tensor var_2642 = const()[name = tensor("op_2642"), val = tensor([1, -1, 8, 64])]; - tensor v_27_cast_fp16 = reshape(shape = var_2642, x = linear_122_cast_fp16)[name = tensor("v_27_cast_fp16")]; - tensor value_29_perm_0 = const()[name = tensor("value_29_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_13_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_13_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170984576)))]; - tensor var_2654_cast_fp16 = add(x = q_79_cast_fp16, y = module_layers_13_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2654_cast_fp16")]; - tensor module_layers_13_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_13_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170985664)))]; - tensor var_2656_cast_fp16 = add(x = q_79_cast_fp16, y = module_layers_13_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2656_cast_fp16")]; - tensor q_with_bias_v_27_perm_0 = const()[name = tensor("q_with_bias_v_27_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_293_transpose_x_0 = const()[name = tensor("x_293_transpose_x_0"), val = tensor(false)]; - tensor x_293_transpose_y_0 = const()[name = tensor("x_293_transpose_y_0"), val = tensor(false)]; - tensor var_2658_to_fp16 = const()[name = tensor("op_2658_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170986752)))]; - tensor q_with_bias_v_27_cast_fp16 = transpose(perm = q_with_bias_v_27_perm_0, x = var_2656_cast_fp16)[name = tensor("transpose_112")]; - tensor x_293_cast_fp16 = matmul(transpose_x = x_293_transpose_x_0, transpose_y = x_293_transpose_y_0, x = q_with_bias_v_27_cast_fp16, y = var_2658_to_fp16)[name = tensor("x_293_cast_fp16")]; - tensor x_295_pad_0 = const()[name = tensor("x_295_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_295_mode_0 = const()[name = tensor("x_295_mode_0"), val = tensor("constant")]; - tensor const_200_to_fp16 = const()[name = tensor("const_200_to_fp16"), val = tensor(0x0p+0)]; - tensor x_295_cast_fp16 = pad(constant_val = const_200_to_fp16, mode = x_295_mode_0, pad = x_295_pad_0, x = x_293_cast_fp16)[name = tensor("x_295_cast_fp16")]; - tensor var_2666 = const()[name = tensor("op_2666"), val = tensor([1, 8, -1, 188])]; - tensor x_297_cast_fp16 = reshape(shape = var_2666, x = x_295_cast_fp16)[name = tensor("x_297_cast_fp16")]; - tensor var_2670_begin_0 = const()[name = tensor("op_2670_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_2670_end_0 = const()[name = tensor("op_2670_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_2670_end_mask_0 = const()[name = tensor("op_2670_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_2670_cast_fp16 = slice_by_index(begin = var_2670_begin_0, end = var_2670_end_0, end_mask = var_2670_end_mask_0, x = x_297_cast_fp16)[name = tensor("op_2670_cast_fp16")]; - tensor var_2671 = const()[name = tensor("op_2671"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_53_cast_fp16 = reshape(shape = var_2671, x = var_2670_cast_fp16)[name = tensor("matrix_bd_53_cast_fp16")]; - tensor matrix_ac_27_transpose_x_0 = const()[name = tensor("matrix_ac_27_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_27_transpose_y_0 = const()[name = tensor("matrix_ac_27_transpose_y_0"), val = tensor(false)]; - tensor transpose_77_perm_0 = const()[name = tensor("transpose_77_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_78_perm_0 = const()[name = tensor("transpose_78_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_78 = transpose(perm = transpose_78_perm_0, x = k_53_cast_fp16)[name = tensor("transpose_110")]; - tensor transpose_77 = transpose(perm = transpose_77_perm_0, x = var_2654_cast_fp16)[name = tensor("transpose_111")]; - tensor matrix_ac_27_cast_fp16 = matmul(transpose_x = matrix_ac_27_transpose_x_0, transpose_y = matrix_ac_27_transpose_y_0, x = transpose_77, y = transpose_78)[name = tensor("matrix_ac_27_cast_fp16")]; - tensor matrix_bd_55_begin_0 = const()[name = tensor("matrix_bd_55_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_55_end_0 = const()[name = tensor("matrix_bd_55_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_55_end_mask_0 = const()[name = tensor("matrix_bd_55_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_55_cast_fp16 = slice_by_index(begin = matrix_bd_55_begin_0, end = matrix_bd_55_end_0, end_mask = matrix_bd_55_end_mask_0, x = matrix_bd_53_cast_fp16)[name = tensor("matrix_bd_55_cast_fp16")]; - tensor var_2680_cast_fp16 = add(x = matrix_ac_27_cast_fp16, y = matrix_bd_55_cast_fp16)[name = tensor("op_2680_cast_fp16")]; - tensor _inversed_scores_53_y_0_to_fp16 = const()[name = tensor("_inversed_scores_53_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_53_cast_fp16 = mul(x = var_2680_cast_fp16, y = _inversed_scores_53_y_0_to_fp16)[name = tensor("_inversed_scores_53_cast_fp16")]; - tensor scores_55_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_53_cast_fp16, cond = mask_11)[name = tensor("scores_55_cast_fp16")]; - tensor var_2686_cast_fp16 = softmax(axis = var_23, x = scores_55_cast_fp16)[name = tensor("op_2686_cast_fp16")]; - tensor input_709_cast_fp16 = select(a = var_11_to_fp16, b = var_2686_cast_fp16, cond = mask_11)[name = tensor("input_709_cast_fp16")]; - tensor x_299_transpose_x_0 = const()[name = tensor("x_299_transpose_x_0"), val = tensor(false)]; - tensor x_299_transpose_y_0 = const()[name = tensor("x_299_transpose_y_0"), val = tensor(false)]; - tensor value_29_cast_fp16 = transpose(perm = value_29_perm_0, x = v_27_cast_fp16)[name = tensor("transpose_113")]; - tensor x_299_cast_fp16 = matmul(transpose_x = x_299_transpose_x_0, transpose_y = x_299_transpose_y_0, x = input_709_cast_fp16, y = value_29_cast_fp16)[name = tensor("x_299_cast_fp16")]; - tensor var_2690_perm_0 = const()[name = tensor("op_2690_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_2691 = const()[name = tensor("op_2691"), val = tensor([1, -1, 512])]; - tensor var_2690_cast_fp16 = transpose(perm = var_2690_perm_0, x = x_299_cast_fp16)[name = tensor("transpose_109")]; - tensor input_711_cast_fp16 = reshape(shape = var_2691, x = var_2690_cast_fp16)[name = tensor("input_711_cast_fp16")]; - tensor module_layers_13_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_13_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171370816)))]; - tensor module_layers_13_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_13_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171895168)))]; - tensor linear_124_cast_fp16 = linear(bias = module_layers_13_self_attn_linear_out_bias_to_fp16, weight = module_layers_13_self_attn_linear_out_weight_to_fp16, x = input_711_cast_fp16)[name = tensor("linear_124_cast_fp16")]; - tensor input_715_cast_fp16 = add(x = input_707_cast_fp16, y = linear_124_cast_fp16)[name = tensor("input_715_cast_fp16")]; - tensor x_303_axes_0 = const()[name = tensor("x_303_axes_0"), val = tensor([-1])]; - tensor module_layers_13_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_13_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171896256)))]; - tensor module_layers_13_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_13_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171897344)))]; - tensor x_303_cast_fp16 = layer_norm(axes = x_303_axes_0, beta = module_layers_13_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_13_norm_conv_weight_to_fp16, x = input_715_cast_fp16)[name = tensor("x_303_cast_fp16")]; - tensor input_717_perm_0 = const()[name = tensor("input_717_perm_0"), val = tensor([0, 2, 1])]; - tensor input_719_pad_type_0 = const()[name = tensor("input_719_pad_type_0"), val = tensor("valid")]; - tensor input_719_strides_0 = const()[name = tensor("input_719_strides_0"), val = tensor([1])]; - tensor input_719_pad_0 = const()[name = tensor("input_719_pad_0"), val = tensor([0, 0])]; - tensor input_719_dilations_0 = const()[name = tensor("input_719_dilations_0"), val = tensor([1])]; - tensor input_719_groups_0 = const()[name = tensor("input_719_groups_0"), val = tensor(1)]; - tensor module_layers_13_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_13_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171898432)))]; - tensor module_layers_13_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_13_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(172947072)))]; - tensor input_717_cast_fp16 = transpose(perm = input_717_perm_0, x = x_303_cast_fp16)[name = tensor("transpose_108")]; - tensor input_719_cast_fp16 = conv(bias = module_layers_13_conv_pointwise_conv1_bias_to_fp16, dilations = input_719_dilations_0, groups = input_719_groups_0, pad = input_719_pad_0, pad_type = input_719_pad_type_0, strides = input_719_strides_0, weight = module_layers_13_conv_pointwise_conv1_weight_to_fp16, x = input_717_cast_fp16)[name = tensor("input_719_cast_fp16")]; - tensor x_305_split_num_splits_0 = const()[name = tensor("x_305_split_num_splits_0"), val = tensor(2)]; - tensor x_305_split_axis_0 = const()[name = tensor("x_305_split_axis_0"), val = tensor(1)]; - tensor x_305_split_cast_fp16_0, tensor x_305_split_cast_fp16_1 = split(axis = x_305_split_axis_0, num_splits = x_305_split_num_splits_0, x = input_719_cast_fp16)[name = tensor("x_305_split_cast_fp16")]; - tensor x_305_split_1_sigmoid_cast_fp16 = sigmoid(x = x_305_split_cast_fp16_1)[name = tensor("x_305_split_1_sigmoid_cast_fp16")]; - tensor x_305_cast_fp16 = mul(x = x_305_split_cast_fp16_0, y = x_305_split_1_sigmoid_cast_fp16)[name = tensor("x_305_cast_fp16")]; - tensor input_721_cast_fp16 = select(a = var_11_to_fp16, b = x_305_cast_fp16, cond = var_453)[name = tensor("input_721_cast_fp16")]; - tensor input_723_pad_0 = const()[name = tensor("input_723_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_723_mode_0 = const()[name = tensor("input_723_mode_0"), val = tensor("constant")]; - tensor const_203_to_fp16 = const()[name = tensor("const_203_to_fp16"), val = tensor(0x0p+0)]; - tensor input_723_cast_fp16 = pad(constant_val = const_203_to_fp16, mode = input_723_mode_0, pad = input_723_pad_0, x = input_721_cast_fp16)[name = tensor("input_723_cast_fp16")]; - tensor input_725_pad_type_0 = const()[name = tensor("input_725_pad_type_0"), val = tensor("valid")]; - tensor input_725_groups_0 = const()[name = tensor("input_725_groups_0"), val = tensor(512)]; - tensor input_725_strides_0 = const()[name = tensor("input_725_strides_0"), val = tensor([1])]; - tensor input_725_pad_0 = const()[name = tensor("input_725_pad_0"), val = tensor([0, 0])]; - tensor input_725_dilations_0 = const()[name = tensor("input_725_dilations_0"), val = tensor([1])]; - tensor const_260_to_fp16 = const()[name = tensor("const_260_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(172949184)))]; - tensor const_261_to_fp16 = const()[name = tensor("const_261_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(172958464)))]; - tensor input_727_cast_fp16 = conv(bias = const_261_to_fp16, dilations = input_725_dilations_0, groups = input_725_groups_0, pad = input_725_pad_0, pad_type = input_725_pad_type_0, strides = input_725_strides_0, weight = const_260_to_fp16, x = input_723_cast_fp16)[name = tensor("input_727_cast_fp16")]; - tensor input_729_cast_fp16 = silu(x = input_727_cast_fp16)[name = tensor("input_729_cast_fp16")]; - tensor x_307_pad_type_0 = const()[name = tensor("x_307_pad_type_0"), val = tensor("valid")]; - tensor x_307_strides_0 = const()[name = tensor("x_307_strides_0"), val = tensor([1])]; - tensor x_307_pad_0 = const()[name = tensor("x_307_pad_0"), val = tensor([0, 0])]; - tensor x_307_dilations_0 = const()[name = tensor("x_307_dilations_0"), val = tensor([1])]; - tensor x_307_groups_0 = const()[name = tensor("x_307_groups_0"), val = tensor(1)]; - tensor module_layers_13_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_13_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(172959552)))]; - tensor module_layers_13_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_13_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(173483904)))]; - tensor x_307_cast_fp16 = conv(bias = module_layers_13_conv_pointwise_conv2_bias_to_fp16, dilations = x_307_dilations_0, groups = x_307_groups_0, pad = x_307_pad_0, pad_type = x_307_pad_type_0, strides = x_307_strides_0, weight = module_layers_13_conv_pointwise_conv2_weight_to_fp16, x = input_729_cast_fp16)[name = tensor("x_307_cast_fp16")]; - tensor input_731_perm_0 = const()[name = tensor("input_731_perm_0"), val = tensor([0, 2, 1])]; - tensor input_731_cast_fp16 = transpose(perm = input_731_perm_0, x = x_307_cast_fp16)[name = tensor("transpose_107")]; - tensor input_733_cast_fp16 = add(x = input_715_cast_fp16, y = input_731_cast_fp16)[name = tensor("input_733_cast_fp16")]; - tensor input_735_axes_0 = const()[name = tensor("input_735_axes_0"), val = tensor([-1])]; - tensor module_layers_13_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_13_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(173484992)))]; - tensor module_layers_13_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_13_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(173486080)))]; - tensor input_735_cast_fp16 = layer_norm(axes = input_735_axes_0, beta = module_layers_13_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_13_norm_feed_forward2_weight_to_fp16, x = input_733_cast_fp16)[name = tensor("input_735_cast_fp16")]; - tensor module_layers_13_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_13_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(173487168)))]; - tensor module_layers_13_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_13_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(175584384)))]; - tensor linear_125_cast_fp16 = linear(bias = module_layers_13_feed_forward2_linear1_bias_to_fp16, weight = module_layers_13_feed_forward2_linear1_weight_to_fp16, x = input_735_cast_fp16)[name = tensor("linear_125_cast_fp16")]; - tensor input_739_cast_fp16 = silu(x = linear_125_cast_fp16)[name = tensor("input_739_cast_fp16")]; - tensor module_layers_13_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_13_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(175588544)))]; - tensor module_layers_13_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_13_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(177685760)))]; - tensor linear_126_cast_fp16 = linear(bias = module_layers_13_feed_forward2_linear2_bias_to_fp16, weight = module_layers_13_feed_forward2_linear2_weight_to_fp16, x = input_739_cast_fp16)[name = tensor("linear_126_cast_fp16")]; - tensor var_2757_to_fp16 = const()[name = tensor("op_2757_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2758_cast_fp16 = mul(x = linear_126_cast_fp16, y = var_2757_to_fp16)[name = tensor("op_2758_cast_fp16")]; - tensor input_745_cast_fp16 = add(x = input_733_cast_fp16, y = var_2758_cast_fp16)[name = tensor("input_745_cast_fp16")]; - tensor input_747_axes_0 = const()[name = tensor("input_747_axes_0"), val = tensor([-1])]; - tensor module_layers_13_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_13_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(177686848)))]; - tensor module_layers_13_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_13_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(177687936)))]; - tensor input_747_cast_fp16 = layer_norm(axes = input_747_axes_0, beta = module_layers_13_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_13_norm_out_weight_to_fp16, x = input_745_cast_fp16)[name = tensor("input_747_cast_fp16")]; - tensor input_749_axes_0 = const()[name = tensor("input_749_axes_0"), val = tensor([-1])]; - tensor module_layers_14_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_14_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(177689024)))]; - tensor module_layers_14_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_14_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(177690112)))]; - tensor input_749_cast_fp16 = layer_norm(axes = input_749_axes_0, beta = module_layers_14_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_14_norm_feed_forward1_weight_to_fp16, x = input_747_cast_fp16)[name = tensor("input_749_cast_fp16")]; - tensor module_layers_14_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_14_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(177691200)))]; - tensor module_layers_14_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_14_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(179788416)))]; - tensor linear_127_cast_fp16 = linear(bias = module_layers_14_feed_forward1_linear1_bias_to_fp16, weight = module_layers_14_feed_forward1_linear1_weight_to_fp16, x = input_749_cast_fp16)[name = tensor("linear_127_cast_fp16")]; - tensor input_753_cast_fp16 = silu(x = linear_127_cast_fp16)[name = tensor("input_753_cast_fp16")]; - tensor module_layers_14_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_14_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(179792576)))]; - tensor module_layers_14_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_14_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(181889792)))]; - tensor linear_128_cast_fp16 = linear(bias = module_layers_14_feed_forward1_linear2_bias_to_fp16, weight = module_layers_14_feed_forward1_linear2_weight_to_fp16, x = input_753_cast_fp16)[name = tensor("linear_128_cast_fp16")]; - tensor var_2788_to_fp16 = const()[name = tensor("op_2788_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2789_cast_fp16 = mul(x = linear_128_cast_fp16, y = var_2788_to_fp16)[name = tensor("op_2789_cast_fp16")]; - tensor input_759_cast_fp16 = add(x = input_747_cast_fp16, y = var_2789_cast_fp16)[name = tensor("input_759_cast_fp16")]; - tensor query_29_axes_0 = const()[name = tensor("query_29_axes_0"), val = tensor([-1])]; - tensor module_layers_14_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_14_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(181890880)))]; - tensor module_layers_14_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_14_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(181891968)))]; - tensor query_29_cast_fp16 = layer_norm(axes = query_29_axes_0, beta = module_layers_14_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_14_norm_self_att_weight_to_fp16, x = input_759_cast_fp16)[name = tensor("query_29_cast_fp16")]; - tensor module_layers_14_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_14_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(181893056)))]; - tensor module_layers_14_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_14_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182417408)))]; - tensor linear_129_cast_fp16 = linear(bias = module_layers_14_self_attn_linear_q_bias_to_fp16, weight = module_layers_14_self_attn_linear_q_weight_to_fp16, x = query_29_cast_fp16)[name = tensor("linear_129_cast_fp16")]; - tensor var_2806 = const()[name = tensor("op_2806"), val = tensor([1, -1, 8, 64])]; - tensor q_85_cast_fp16 = reshape(shape = var_2806, x = linear_129_cast_fp16)[name = tensor("q_85_cast_fp16")]; - tensor module_layers_14_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_14_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182418496)))]; - tensor module_layers_14_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_14_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182942848)))]; - tensor linear_130_cast_fp16 = linear(bias = module_layers_14_self_attn_linear_k_bias_to_fp16, weight = module_layers_14_self_attn_linear_k_weight_to_fp16, x = query_29_cast_fp16)[name = tensor("linear_130_cast_fp16")]; - tensor var_2811 = const()[name = tensor("op_2811"), val = tensor([1, -1, 8, 64])]; - tensor k_57_cast_fp16 = reshape(shape = var_2811, x = linear_130_cast_fp16)[name = tensor("k_57_cast_fp16")]; - tensor module_layers_14_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_14_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182943936)))]; - tensor module_layers_14_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_14_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183468288)))]; - tensor linear_131_cast_fp16 = linear(bias = module_layers_14_self_attn_linear_v_bias_to_fp16, weight = module_layers_14_self_attn_linear_v_weight_to_fp16, x = query_29_cast_fp16)[name = tensor("linear_131_cast_fp16")]; - tensor var_2816 = const()[name = tensor("op_2816"), val = tensor([1, -1, 8, 64])]; - tensor v_29_cast_fp16 = reshape(shape = var_2816, x = linear_131_cast_fp16)[name = tensor("v_29_cast_fp16")]; - tensor value_31_perm_0 = const()[name = tensor("value_31_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_14_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_14_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183469376)))]; - tensor var_2828_cast_fp16 = add(x = q_85_cast_fp16, y = module_layers_14_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2828_cast_fp16")]; - tensor module_layers_14_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_14_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183470464)))]; - tensor var_2830_cast_fp16 = add(x = q_85_cast_fp16, y = module_layers_14_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2830_cast_fp16")]; - tensor q_with_bias_v_29_perm_0 = const()[name = tensor("q_with_bias_v_29_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_315_transpose_x_0 = const()[name = tensor("x_315_transpose_x_0"), val = tensor(false)]; - tensor x_315_transpose_y_0 = const()[name = tensor("x_315_transpose_y_0"), val = tensor(false)]; - tensor var_2832_to_fp16 = const()[name = tensor("op_2832_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183471552)))]; - tensor q_with_bias_v_29_cast_fp16 = transpose(perm = q_with_bias_v_29_perm_0, x = var_2830_cast_fp16)[name = tensor("transpose_105")]; - tensor x_315_cast_fp16 = matmul(transpose_x = x_315_transpose_x_0, transpose_y = x_315_transpose_y_0, x = q_with_bias_v_29_cast_fp16, y = var_2832_to_fp16)[name = tensor("x_315_cast_fp16")]; - tensor x_317_pad_0 = const()[name = tensor("x_317_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_317_mode_0 = const()[name = tensor("x_317_mode_0"), val = tensor("constant")]; - tensor const_210_to_fp16 = const()[name = tensor("const_210_to_fp16"), val = tensor(0x0p+0)]; - tensor x_317_cast_fp16 = pad(constant_val = const_210_to_fp16, mode = x_317_mode_0, pad = x_317_pad_0, x = x_315_cast_fp16)[name = tensor("x_317_cast_fp16")]; - tensor var_2840 = const()[name = tensor("op_2840"), val = tensor([1, 8, -1, 188])]; - tensor x_319_cast_fp16 = reshape(shape = var_2840, x = x_317_cast_fp16)[name = tensor("x_319_cast_fp16")]; - tensor var_2844_begin_0 = const()[name = tensor("op_2844_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_2844_end_0 = const()[name = tensor("op_2844_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_2844_end_mask_0 = const()[name = tensor("op_2844_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_2844_cast_fp16 = slice_by_index(begin = var_2844_begin_0, end = var_2844_end_0, end_mask = var_2844_end_mask_0, x = x_319_cast_fp16)[name = tensor("op_2844_cast_fp16")]; - tensor var_2845 = const()[name = tensor("op_2845"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_57_cast_fp16 = reshape(shape = var_2845, x = var_2844_cast_fp16)[name = tensor("matrix_bd_57_cast_fp16")]; - tensor matrix_ac_29_transpose_x_0 = const()[name = tensor("matrix_ac_29_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_29_transpose_y_0 = const()[name = tensor("matrix_ac_29_transpose_y_0"), val = tensor(false)]; - tensor transpose_79_perm_0 = const()[name = tensor("transpose_79_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_80_perm_0 = const()[name = tensor("transpose_80_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_80 = transpose(perm = transpose_80_perm_0, x = k_57_cast_fp16)[name = tensor("transpose_103")]; - tensor transpose_79 = transpose(perm = transpose_79_perm_0, x = var_2828_cast_fp16)[name = tensor("transpose_104")]; - tensor matrix_ac_29_cast_fp16 = matmul(transpose_x = matrix_ac_29_transpose_x_0, transpose_y = matrix_ac_29_transpose_y_0, x = transpose_79, y = transpose_80)[name = tensor("matrix_ac_29_cast_fp16")]; - tensor matrix_bd_59_begin_0 = const()[name = tensor("matrix_bd_59_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_59_end_0 = const()[name = tensor("matrix_bd_59_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_59_end_mask_0 = const()[name = tensor("matrix_bd_59_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_59_cast_fp16 = slice_by_index(begin = matrix_bd_59_begin_0, end = matrix_bd_59_end_0, end_mask = matrix_bd_59_end_mask_0, x = matrix_bd_57_cast_fp16)[name = tensor("matrix_bd_59_cast_fp16")]; - tensor var_2854_cast_fp16 = add(x = matrix_ac_29_cast_fp16, y = matrix_bd_59_cast_fp16)[name = tensor("op_2854_cast_fp16")]; - tensor _inversed_scores_57_y_0_to_fp16 = const()[name = tensor("_inversed_scores_57_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_57_cast_fp16 = mul(x = var_2854_cast_fp16, y = _inversed_scores_57_y_0_to_fp16)[name = tensor("_inversed_scores_57_cast_fp16")]; - tensor scores_59_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_57_cast_fp16, cond = mask_11)[name = tensor("scores_59_cast_fp16")]; - tensor var_2860_cast_fp16 = softmax(axis = var_23, x = scores_59_cast_fp16)[name = tensor("op_2860_cast_fp16")]; - tensor input_761_cast_fp16 = select(a = var_11_to_fp16, b = var_2860_cast_fp16, cond = mask_11)[name = tensor("input_761_cast_fp16")]; - tensor x_321_transpose_x_0 = const()[name = tensor("x_321_transpose_x_0"), val = tensor(false)]; - tensor x_321_transpose_y_0 = const()[name = tensor("x_321_transpose_y_0"), val = tensor(false)]; - tensor value_31_cast_fp16 = transpose(perm = value_31_perm_0, x = v_29_cast_fp16)[name = tensor("transpose_106")]; - tensor x_321_cast_fp16 = matmul(transpose_x = x_321_transpose_x_0, transpose_y = x_321_transpose_y_0, x = input_761_cast_fp16, y = value_31_cast_fp16)[name = tensor("x_321_cast_fp16")]; - tensor var_2864_perm_0 = const()[name = tensor("op_2864_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_2865 = const()[name = tensor("op_2865"), val = tensor([1, -1, 512])]; - tensor var_2864_cast_fp16 = transpose(perm = var_2864_perm_0, x = x_321_cast_fp16)[name = tensor("transpose_102")]; - tensor input_763_cast_fp16 = reshape(shape = var_2865, x = var_2864_cast_fp16)[name = tensor("input_763_cast_fp16")]; - tensor module_layers_14_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_14_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183855616)))]; - tensor module_layers_14_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_14_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184379968)))]; - tensor linear_133_cast_fp16 = linear(bias = module_layers_14_self_attn_linear_out_bias_to_fp16, weight = module_layers_14_self_attn_linear_out_weight_to_fp16, x = input_763_cast_fp16)[name = tensor("linear_133_cast_fp16")]; - tensor input_767_cast_fp16 = add(x = input_759_cast_fp16, y = linear_133_cast_fp16)[name = tensor("input_767_cast_fp16")]; - tensor x_325_axes_0 = const()[name = tensor("x_325_axes_0"), val = tensor([-1])]; - tensor module_layers_14_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_14_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184381056)))]; - tensor module_layers_14_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_14_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184382144)))]; - tensor x_325_cast_fp16 = layer_norm(axes = x_325_axes_0, beta = module_layers_14_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_14_norm_conv_weight_to_fp16, x = input_767_cast_fp16)[name = tensor("x_325_cast_fp16")]; - tensor input_769_perm_0 = const()[name = tensor("input_769_perm_0"), val = tensor([0, 2, 1])]; - tensor input_771_pad_type_0 = const()[name = tensor("input_771_pad_type_0"), val = tensor("valid")]; - tensor input_771_strides_0 = const()[name = tensor("input_771_strides_0"), val = tensor([1])]; - tensor input_771_pad_0 = const()[name = tensor("input_771_pad_0"), val = tensor([0, 0])]; - tensor input_771_dilations_0 = const()[name = tensor("input_771_dilations_0"), val = tensor([1])]; - tensor input_771_groups_0 = const()[name = tensor("input_771_groups_0"), val = tensor(1)]; - tensor module_layers_14_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_14_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184383232)))]; - tensor module_layers_14_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_14_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185431872)))]; - tensor input_769_cast_fp16 = transpose(perm = input_769_perm_0, x = x_325_cast_fp16)[name = tensor("transpose_101")]; - tensor input_771_cast_fp16 = conv(bias = module_layers_14_conv_pointwise_conv1_bias_to_fp16, dilations = input_771_dilations_0, groups = input_771_groups_0, pad = input_771_pad_0, pad_type = input_771_pad_type_0, strides = input_771_strides_0, weight = module_layers_14_conv_pointwise_conv1_weight_to_fp16, x = input_769_cast_fp16)[name = tensor("input_771_cast_fp16")]; - tensor x_327_split_num_splits_0 = const()[name = tensor("x_327_split_num_splits_0"), val = tensor(2)]; - tensor x_327_split_axis_0 = const()[name = tensor("x_327_split_axis_0"), val = tensor(1)]; - tensor x_327_split_cast_fp16_0, tensor x_327_split_cast_fp16_1 = split(axis = x_327_split_axis_0, num_splits = x_327_split_num_splits_0, x = input_771_cast_fp16)[name = tensor("x_327_split_cast_fp16")]; - tensor x_327_split_1_sigmoid_cast_fp16 = sigmoid(x = x_327_split_cast_fp16_1)[name = tensor("x_327_split_1_sigmoid_cast_fp16")]; - tensor x_327_cast_fp16 = mul(x = x_327_split_cast_fp16_0, y = x_327_split_1_sigmoid_cast_fp16)[name = tensor("x_327_cast_fp16")]; - tensor input_773_cast_fp16 = select(a = var_11_to_fp16, b = x_327_cast_fp16, cond = var_453)[name = tensor("input_773_cast_fp16")]; - tensor input_775_pad_0 = const()[name = tensor("input_775_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_775_mode_0 = const()[name = tensor("input_775_mode_0"), val = tensor("constant")]; - tensor const_213_to_fp16 = const()[name = tensor("const_213_to_fp16"), val = tensor(0x0p+0)]; - tensor input_775_cast_fp16 = pad(constant_val = const_213_to_fp16, mode = input_775_mode_0, pad = input_775_pad_0, x = input_773_cast_fp16)[name = tensor("input_775_cast_fp16")]; - tensor input_777_pad_type_0 = const()[name = tensor("input_777_pad_type_0"), val = tensor("valid")]; - tensor input_777_groups_0 = const()[name = tensor("input_777_groups_0"), val = tensor(512)]; - tensor input_777_strides_0 = const()[name = tensor("input_777_strides_0"), val = tensor([1])]; - tensor input_777_pad_0 = const()[name = tensor("input_777_pad_0"), val = tensor([0, 0])]; - tensor input_777_dilations_0 = const()[name = tensor("input_777_dilations_0"), val = tensor([1])]; - tensor const_262_to_fp16 = const()[name = tensor("const_262_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185433984)))]; - tensor const_263_to_fp16 = const()[name = tensor("const_263_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185443264)))]; - tensor input_779_cast_fp16 = conv(bias = const_263_to_fp16, dilations = input_777_dilations_0, groups = input_777_groups_0, pad = input_777_pad_0, pad_type = input_777_pad_type_0, strides = input_777_strides_0, weight = const_262_to_fp16, x = input_775_cast_fp16)[name = tensor("input_779_cast_fp16")]; - tensor input_781_cast_fp16 = silu(x = input_779_cast_fp16)[name = tensor("input_781_cast_fp16")]; - tensor x_329_pad_type_0 = const()[name = tensor("x_329_pad_type_0"), val = tensor("valid")]; - tensor x_329_strides_0 = const()[name = tensor("x_329_strides_0"), val = tensor([1])]; - tensor x_329_pad_0 = const()[name = tensor("x_329_pad_0"), val = tensor([0, 0])]; - tensor x_329_dilations_0 = const()[name = tensor("x_329_dilations_0"), val = tensor([1])]; - tensor x_329_groups_0 = const()[name = tensor("x_329_groups_0"), val = tensor(1)]; - tensor module_layers_14_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_14_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185444352)))]; - tensor module_layers_14_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_14_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185968704)))]; - tensor x_329_cast_fp16 = conv(bias = module_layers_14_conv_pointwise_conv2_bias_to_fp16, dilations = x_329_dilations_0, groups = x_329_groups_0, pad = x_329_pad_0, pad_type = x_329_pad_type_0, strides = x_329_strides_0, weight = module_layers_14_conv_pointwise_conv2_weight_to_fp16, x = input_781_cast_fp16)[name = tensor("x_329_cast_fp16")]; - tensor input_783_perm_0 = const()[name = tensor("input_783_perm_0"), val = tensor([0, 2, 1])]; - tensor input_783_cast_fp16 = transpose(perm = input_783_perm_0, x = x_329_cast_fp16)[name = tensor("transpose_100")]; - tensor input_785_cast_fp16 = add(x = input_767_cast_fp16, y = input_783_cast_fp16)[name = tensor("input_785_cast_fp16")]; - tensor input_787_axes_0 = const()[name = tensor("input_787_axes_0"), val = tensor([-1])]; - tensor module_layers_14_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_14_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185969792)))]; - tensor module_layers_14_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_14_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185970880)))]; - tensor input_787_cast_fp16 = layer_norm(axes = input_787_axes_0, beta = module_layers_14_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_14_norm_feed_forward2_weight_to_fp16, x = input_785_cast_fp16)[name = tensor("input_787_cast_fp16")]; - tensor module_layers_14_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_14_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185971968)))]; - tensor module_layers_14_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_14_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188069184)))]; - tensor linear_134_cast_fp16 = linear(bias = module_layers_14_feed_forward2_linear1_bias_to_fp16, weight = module_layers_14_feed_forward2_linear1_weight_to_fp16, x = input_787_cast_fp16)[name = tensor("linear_134_cast_fp16")]; - tensor input_791_cast_fp16 = silu(x = linear_134_cast_fp16)[name = tensor("input_791_cast_fp16")]; - tensor module_layers_14_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_14_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188073344)))]; - tensor module_layers_14_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_14_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190170560)))]; - tensor linear_135_cast_fp16 = linear(bias = module_layers_14_feed_forward2_linear2_bias_to_fp16, weight = module_layers_14_feed_forward2_linear2_weight_to_fp16, x = input_791_cast_fp16)[name = tensor("linear_135_cast_fp16")]; - tensor var_2931_to_fp16 = const()[name = tensor("op_2931_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2932_cast_fp16 = mul(x = linear_135_cast_fp16, y = var_2931_to_fp16)[name = tensor("op_2932_cast_fp16")]; - tensor input_797_cast_fp16 = add(x = input_785_cast_fp16, y = var_2932_cast_fp16)[name = tensor("input_797_cast_fp16")]; - tensor input_799_axes_0 = const()[name = tensor("input_799_axes_0"), val = tensor([-1])]; - tensor module_layers_14_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_14_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190171648)))]; - tensor module_layers_14_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_14_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190172736)))]; - tensor input_799_cast_fp16 = layer_norm(axes = input_799_axes_0, beta = module_layers_14_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_14_norm_out_weight_to_fp16, x = input_797_cast_fp16)[name = tensor("input_799_cast_fp16")]; - tensor input_801_axes_0 = const()[name = tensor("input_801_axes_0"), val = tensor([-1])]; - tensor module_layers_15_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_15_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190173824)))]; - tensor module_layers_15_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_15_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190174912)))]; - tensor input_801_cast_fp16 = layer_norm(axes = input_801_axes_0, beta = module_layers_15_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_15_norm_feed_forward1_weight_to_fp16, x = input_799_cast_fp16)[name = tensor("input_801_cast_fp16")]; - tensor module_layers_15_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_15_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190176000)))]; - tensor module_layers_15_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_15_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(192273216)))]; - tensor linear_136_cast_fp16 = linear(bias = module_layers_15_feed_forward1_linear1_bias_to_fp16, weight = module_layers_15_feed_forward1_linear1_weight_to_fp16, x = input_801_cast_fp16)[name = tensor("linear_136_cast_fp16")]; - tensor input_805_cast_fp16 = silu(x = linear_136_cast_fp16)[name = tensor("input_805_cast_fp16")]; - tensor module_layers_15_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_15_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(192277376)))]; - tensor module_layers_15_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_15_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194374592)))]; - tensor linear_137_cast_fp16 = linear(bias = module_layers_15_feed_forward1_linear2_bias_to_fp16, weight = module_layers_15_feed_forward1_linear2_weight_to_fp16, x = input_805_cast_fp16)[name = tensor("linear_137_cast_fp16")]; - tensor var_2962_to_fp16 = const()[name = tensor("op_2962_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2963_cast_fp16 = mul(x = linear_137_cast_fp16, y = var_2962_to_fp16)[name = tensor("op_2963_cast_fp16")]; - tensor input_811_cast_fp16 = add(x = input_799_cast_fp16, y = var_2963_cast_fp16)[name = tensor("input_811_cast_fp16")]; - tensor query_31_axes_0 = const()[name = tensor("query_31_axes_0"), val = tensor([-1])]; - tensor module_layers_15_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_15_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194375680)))]; - tensor module_layers_15_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_15_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194376768)))]; - tensor query_31_cast_fp16 = layer_norm(axes = query_31_axes_0, beta = module_layers_15_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_15_norm_self_att_weight_to_fp16, x = input_811_cast_fp16)[name = tensor("query_31_cast_fp16")]; - tensor module_layers_15_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_15_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194377856)))]; - tensor module_layers_15_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_15_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194902208)))]; - tensor linear_138_cast_fp16 = linear(bias = module_layers_15_self_attn_linear_q_bias_to_fp16, weight = module_layers_15_self_attn_linear_q_weight_to_fp16, x = query_31_cast_fp16)[name = tensor("linear_138_cast_fp16")]; - tensor var_2980 = const()[name = tensor("op_2980"), val = tensor([1, -1, 8, 64])]; - tensor q_91_cast_fp16 = reshape(shape = var_2980, x = linear_138_cast_fp16)[name = tensor("q_91_cast_fp16")]; - tensor module_layers_15_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_15_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194903296)))]; - tensor module_layers_15_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_15_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195427648)))]; - tensor linear_139_cast_fp16 = linear(bias = module_layers_15_self_attn_linear_k_bias_to_fp16, weight = module_layers_15_self_attn_linear_k_weight_to_fp16, x = query_31_cast_fp16)[name = tensor("linear_139_cast_fp16")]; - tensor var_2985 = const()[name = tensor("op_2985"), val = tensor([1, -1, 8, 64])]; - tensor k_61_cast_fp16 = reshape(shape = var_2985, x = linear_139_cast_fp16)[name = tensor("k_61_cast_fp16")]; - tensor module_layers_15_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_15_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195428736)))]; - tensor module_layers_15_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_15_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195953088)))]; - tensor linear_140_cast_fp16 = linear(bias = module_layers_15_self_attn_linear_v_bias_to_fp16, weight = module_layers_15_self_attn_linear_v_weight_to_fp16, x = query_31_cast_fp16)[name = tensor("linear_140_cast_fp16")]; - tensor var_2990 = const()[name = tensor("op_2990"), val = tensor([1, -1, 8, 64])]; - tensor v_31_cast_fp16 = reshape(shape = var_2990, x = linear_140_cast_fp16)[name = tensor("v_31_cast_fp16")]; - tensor value_33_perm_0 = const()[name = tensor("value_33_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_15_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_15_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195954176)))]; - tensor var_3002_cast_fp16 = add(x = q_91_cast_fp16, y = module_layers_15_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3002_cast_fp16")]; - tensor module_layers_15_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_15_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195955264)))]; - tensor var_3004_cast_fp16 = add(x = q_91_cast_fp16, y = module_layers_15_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3004_cast_fp16")]; - tensor q_with_bias_v_31_perm_0 = const()[name = tensor("q_with_bias_v_31_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_337_transpose_x_0 = const()[name = tensor("x_337_transpose_x_0"), val = tensor(false)]; - tensor x_337_transpose_y_0 = const()[name = tensor("x_337_transpose_y_0"), val = tensor(false)]; - tensor var_3006_to_fp16 = const()[name = tensor("op_3006_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195956352)))]; - tensor q_with_bias_v_31_cast_fp16 = transpose(perm = q_with_bias_v_31_perm_0, x = var_3004_cast_fp16)[name = tensor("transpose_98")]; - tensor x_337_cast_fp16 = matmul(transpose_x = x_337_transpose_x_0, transpose_y = x_337_transpose_y_0, x = q_with_bias_v_31_cast_fp16, y = var_3006_to_fp16)[name = tensor("x_337_cast_fp16")]; - tensor x_339_pad_0 = const()[name = tensor("x_339_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_339_mode_0 = const()[name = tensor("x_339_mode_0"), val = tensor("constant")]; - tensor const_220_to_fp16 = const()[name = tensor("const_220_to_fp16"), val = tensor(0x0p+0)]; - tensor x_339_cast_fp16 = pad(constant_val = const_220_to_fp16, mode = x_339_mode_0, pad = x_339_pad_0, x = x_337_cast_fp16)[name = tensor("x_339_cast_fp16")]; - tensor var_3014 = const()[name = tensor("op_3014"), val = tensor([1, 8, -1, 188])]; - tensor x_341_cast_fp16 = reshape(shape = var_3014, x = x_339_cast_fp16)[name = tensor("x_341_cast_fp16")]; - tensor var_3018_begin_0 = const()[name = tensor("op_3018_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_3018_end_0 = const()[name = tensor("op_3018_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_3018_end_mask_0 = const()[name = tensor("op_3018_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_3018_cast_fp16 = slice_by_index(begin = var_3018_begin_0, end = var_3018_end_0, end_mask = var_3018_end_mask_0, x = x_341_cast_fp16)[name = tensor("op_3018_cast_fp16")]; - tensor var_3019 = const()[name = tensor("op_3019"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_61_cast_fp16 = reshape(shape = var_3019, x = var_3018_cast_fp16)[name = tensor("matrix_bd_61_cast_fp16")]; - tensor matrix_ac_31_transpose_x_0 = const()[name = tensor("matrix_ac_31_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_31_transpose_y_0 = const()[name = tensor("matrix_ac_31_transpose_y_0"), val = tensor(false)]; - tensor transpose_81_perm_0 = const()[name = tensor("transpose_81_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_82_perm_0 = const()[name = tensor("transpose_82_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_82 = transpose(perm = transpose_82_perm_0, x = k_61_cast_fp16)[name = tensor("transpose_96")]; - tensor transpose_81 = transpose(perm = transpose_81_perm_0, x = var_3002_cast_fp16)[name = tensor("transpose_97")]; - tensor matrix_ac_31_cast_fp16 = matmul(transpose_x = matrix_ac_31_transpose_x_0, transpose_y = matrix_ac_31_transpose_y_0, x = transpose_81, y = transpose_82)[name = tensor("matrix_ac_31_cast_fp16")]; - tensor matrix_bd_63_begin_0 = const()[name = tensor("matrix_bd_63_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_63_end_0 = const()[name = tensor("matrix_bd_63_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_63_end_mask_0 = const()[name = tensor("matrix_bd_63_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_63_cast_fp16 = slice_by_index(begin = matrix_bd_63_begin_0, end = matrix_bd_63_end_0, end_mask = matrix_bd_63_end_mask_0, x = matrix_bd_61_cast_fp16)[name = tensor("matrix_bd_63_cast_fp16")]; - tensor var_3028_cast_fp16 = add(x = matrix_ac_31_cast_fp16, y = matrix_bd_63_cast_fp16)[name = tensor("op_3028_cast_fp16")]; - tensor _inversed_scores_61_y_0_to_fp16 = const()[name = tensor("_inversed_scores_61_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_61_cast_fp16 = mul(x = var_3028_cast_fp16, y = _inversed_scores_61_y_0_to_fp16)[name = tensor("_inversed_scores_61_cast_fp16")]; - tensor scores_63_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_61_cast_fp16, cond = mask_11)[name = tensor("scores_63_cast_fp16")]; - tensor var_3034_cast_fp16 = softmax(axis = var_23, x = scores_63_cast_fp16)[name = tensor("op_3034_cast_fp16")]; - tensor input_813_cast_fp16 = select(a = var_11_to_fp16, b = var_3034_cast_fp16, cond = mask_11)[name = tensor("input_813_cast_fp16")]; - tensor x_343_transpose_x_0 = const()[name = tensor("x_343_transpose_x_0"), val = tensor(false)]; - tensor x_343_transpose_y_0 = const()[name = tensor("x_343_transpose_y_0"), val = tensor(false)]; - tensor value_33_cast_fp16 = transpose(perm = value_33_perm_0, x = v_31_cast_fp16)[name = tensor("transpose_99")]; - tensor x_343_cast_fp16 = matmul(transpose_x = x_343_transpose_x_0, transpose_y = x_343_transpose_y_0, x = input_813_cast_fp16, y = value_33_cast_fp16)[name = tensor("x_343_cast_fp16")]; - tensor var_3038_perm_0 = const()[name = tensor("op_3038_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_3039 = const()[name = tensor("op_3039"), val = tensor([1, -1, 512])]; - tensor var_3038_cast_fp16 = transpose(perm = var_3038_perm_0, x = x_343_cast_fp16)[name = tensor("transpose_95")]; - tensor input_815_cast_fp16 = reshape(shape = var_3039, x = var_3038_cast_fp16)[name = tensor("input_815_cast_fp16")]; - tensor module_layers_15_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_15_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196340416)))]; - tensor module_layers_15_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_15_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196864768)))]; - tensor linear_142_cast_fp16 = linear(bias = module_layers_15_self_attn_linear_out_bias_to_fp16, weight = module_layers_15_self_attn_linear_out_weight_to_fp16, x = input_815_cast_fp16)[name = tensor("linear_142_cast_fp16")]; - tensor input_819_cast_fp16 = add(x = input_811_cast_fp16, y = linear_142_cast_fp16)[name = tensor("input_819_cast_fp16")]; - tensor x_347_axes_0 = const()[name = tensor("x_347_axes_0"), val = tensor([-1])]; - tensor module_layers_15_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_15_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196865856)))]; - tensor module_layers_15_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_15_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196866944)))]; - tensor x_347_cast_fp16 = layer_norm(axes = x_347_axes_0, beta = module_layers_15_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_15_norm_conv_weight_to_fp16, x = input_819_cast_fp16)[name = tensor("x_347_cast_fp16")]; - tensor input_821_perm_0 = const()[name = tensor("input_821_perm_0"), val = tensor([0, 2, 1])]; - tensor input_823_pad_type_0 = const()[name = tensor("input_823_pad_type_0"), val = tensor("valid")]; - tensor input_823_strides_0 = const()[name = tensor("input_823_strides_0"), val = tensor([1])]; - tensor input_823_pad_0 = const()[name = tensor("input_823_pad_0"), val = tensor([0, 0])]; - tensor input_823_dilations_0 = const()[name = tensor("input_823_dilations_0"), val = tensor([1])]; - tensor input_823_groups_0 = const()[name = tensor("input_823_groups_0"), val = tensor(1)]; - tensor module_layers_15_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_15_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196868032)))]; - tensor module_layers_15_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_15_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(197916672)))]; - tensor input_821_cast_fp16 = transpose(perm = input_821_perm_0, x = x_347_cast_fp16)[name = tensor("transpose_94")]; - tensor input_823_cast_fp16 = conv(bias = module_layers_15_conv_pointwise_conv1_bias_to_fp16, dilations = input_823_dilations_0, groups = input_823_groups_0, pad = input_823_pad_0, pad_type = input_823_pad_type_0, strides = input_823_strides_0, weight = module_layers_15_conv_pointwise_conv1_weight_to_fp16, x = input_821_cast_fp16)[name = tensor("input_823_cast_fp16")]; - tensor x_349_split_num_splits_0 = const()[name = tensor("x_349_split_num_splits_0"), val = tensor(2)]; - tensor x_349_split_axis_0 = const()[name = tensor("x_349_split_axis_0"), val = tensor(1)]; - tensor x_349_split_cast_fp16_0, tensor x_349_split_cast_fp16_1 = split(axis = x_349_split_axis_0, num_splits = x_349_split_num_splits_0, x = input_823_cast_fp16)[name = tensor("x_349_split_cast_fp16")]; - tensor x_349_split_1_sigmoid_cast_fp16 = sigmoid(x = x_349_split_cast_fp16_1)[name = tensor("x_349_split_1_sigmoid_cast_fp16")]; - tensor x_349_cast_fp16 = mul(x = x_349_split_cast_fp16_0, y = x_349_split_1_sigmoid_cast_fp16)[name = tensor("x_349_cast_fp16")]; - tensor input_825_cast_fp16 = select(a = var_11_to_fp16, b = x_349_cast_fp16, cond = var_453)[name = tensor("input_825_cast_fp16")]; - tensor input_827_pad_0 = const()[name = tensor("input_827_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_827_mode_0 = const()[name = tensor("input_827_mode_0"), val = tensor("constant")]; - tensor const_223_to_fp16 = const()[name = tensor("const_223_to_fp16"), val = tensor(0x0p+0)]; - tensor input_827_cast_fp16 = pad(constant_val = const_223_to_fp16, mode = input_827_mode_0, pad = input_827_pad_0, x = input_825_cast_fp16)[name = tensor("input_827_cast_fp16")]; - tensor input_829_pad_type_0 = const()[name = tensor("input_829_pad_type_0"), val = tensor("valid")]; - tensor input_829_groups_0 = const()[name = tensor("input_829_groups_0"), val = tensor(512)]; - tensor input_829_strides_0 = const()[name = tensor("input_829_strides_0"), val = tensor([1])]; - tensor input_829_pad_0 = const()[name = tensor("input_829_pad_0"), val = tensor([0, 0])]; - tensor input_829_dilations_0 = const()[name = tensor("input_829_dilations_0"), val = tensor([1])]; - tensor const_264_to_fp16 = const()[name = tensor("const_264_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(197918784)))]; - tensor const_265_to_fp16 = const()[name = tensor("const_265_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(197928064)))]; - tensor input_831_cast_fp16 = conv(bias = const_265_to_fp16, dilations = input_829_dilations_0, groups = input_829_groups_0, pad = input_829_pad_0, pad_type = input_829_pad_type_0, strides = input_829_strides_0, weight = const_264_to_fp16, x = input_827_cast_fp16)[name = tensor("input_831_cast_fp16")]; - tensor input_833_cast_fp16 = silu(x = input_831_cast_fp16)[name = tensor("input_833_cast_fp16")]; - tensor x_351_pad_type_0 = const()[name = tensor("x_351_pad_type_0"), val = tensor("valid")]; - tensor x_351_strides_0 = const()[name = tensor("x_351_strides_0"), val = tensor([1])]; - tensor x_351_pad_0 = const()[name = tensor("x_351_pad_0"), val = tensor([0, 0])]; - tensor x_351_dilations_0 = const()[name = tensor("x_351_dilations_0"), val = tensor([1])]; - tensor x_351_groups_0 = const()[name = tensor("x_351_groups_0"), val = tensor(1)]; - tensor module_layers_15_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_15_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(197929152)))]; - tensor module_layers_15_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_15_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198453504)))]; - tensor x_351_cast_fp16 = conv(bias = module_layers_15_conv_pointwise_conv2_bias_to_fp16, dilations = x_351_dilations_0, groups = x_351_groups_0, pad = x_351_pad_0, pad_type = x_351_pad_type_0, strides = x_351_strides_0, weight = module_layers_15_conv_pointwise_conv2_weight_to_fp16, x = input_833_cast_fp16)[name = tensor("x_351_cast_fp16")]; - tensor input_835_perm_0 = const()[name = tensor("input_835_perm_0"), val = tensor([0, 2, 1])]; - tensor input_835_cast_fp16 = transpose(perm = input_835_perm_0, x = x_351_cast_fp16)[name = tensor("transpose_93")]; - tensor input_837_cast_fp16 = add(x = input_819_cast_fp16, y = input_835_cast_fp16)[name = tensor("input_837_cast_fp16")]; - tensor input_839_axes_0 = const()[name = tensor("input_839_axes_0"), val = tensor([-1])]; - tensor module_layers_15_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_15_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198454592)))]; - tensor module_layers_15_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_15_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198455680)))]; - tensor input_839_cast_fp16 = layer_norm(axes = input_839_axes_0, beta = module_layers_15_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_15_norm_feed_forward2_weight_to_fp16, x = input_837_cast_fp16)[name = tensor("input_839_cast_fp16")]; - tensor module_layers_15_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_15_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198456768)))]; - tensor module_layers_15_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_15_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(200553984)))]; - tensor linear_143_cast_fp16 = linear(bias = module_layers_15_feed_forward2_linear1_bias_to_fp16, weight = module_layers_15_feed_forward2_linear1_weight_to_fp16, x = input_839_cast_fp16)[name = tensor("linear_143_cast_fp16")]; - tensor input_843_cast_fp16 = silu(x = linear_143_cast_fp16)[name = tensor("input_843_cast_fp16")]; - tensor module_layers_15_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_15_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(200558144)))]; - tensor module_layers_15_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_15_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202655360)))]; - tensor linear_144_cast_fp16 = linear(bias = module_layers_15_feed_forward2_linear2_bias_to_fp16, weight = module_layers_15_feed_forward2_linear2_weight_to_fp16, x = input_843_cast_fp16)[name = tensor("linear_144_cast_fp16")]; - tensor var_3105_to_fp16 = const()[name = tensor("op_3105_to_fp16"), val = tensor(0x1p-1)]; - tensor var_3106_cast_fp16 = mul(x = linear_144_cast_fp16, y = var_3105_to_fp16)[name = tensor("op_3106_cast_fp16")]; - tensor input_849_cast_fp16 = add(x = input_837_cast_fp16, y = var_3106_cast_fp16)[name = tensor("input_849_cast_fp16")]; - tensor input_851_axes_0 = const()[name = tensor("input_851_axes_0"), val = tensor([-1])]; - tensor module_layers_15_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_15_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202656448)))]; - tensor module_layers_15_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_15_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202657536)))]; - tensor input_851_cast_fp16 = layer_norm(axes = input_851_axes_0, beta = module_layers_15_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_15_norm_out_weight_to_fp16, x = input_849_cast_fp16)[name = tensor("input_851_cast_fp16")]; - tensor input_853_axes_0 = const()[name = tensor("input_853_axes_0"), val = tensor([-1])]; - tensor module_layers_16_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_16_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202658624)))]; - tensor module_layers_16_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_16_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202659712)))]; - tensor input_853_cast_fp16 = layer_norm(axes = input_853_axes_0, beta = module_layers_16_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_16_norm_feed_forward1_weight_to_fp16, x = input_851_cast_fp16)[name = tensor("input_853_cast_fp16")]; - tensor module_layers_16_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_16_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202660800)))]; - tensor module_layers_16_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_16_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(204758016)))]; - tensor linear_145_cast_fp16 = linear(bias = module_layers_16_feed_forward1_linear1_bias_to_fp16, weight = module_layers_16_feed_forward1_linear1_weight_to_fp16, x = input_853_cast_fp16)[name = tensor("linear_145_cast_fp16")]; - tensor input_857_cast_fp16 = silu(x = linear_145_cast_fp16)[name = tensor("input_857_cast_fp16")]; - tensor module_layers_16_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_16_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(204762176)))]; - tensor module_layers_16_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_16_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(206859392)))]; - tensor linear_146_cast_fp16 = linear(bias = module_layers_16_feed_forward1_linear2_bias_to_fp16, weight = module_layers_16_feed_forward1_linear2_weight_to_fp16, x = input_857_cast_fp16)[name = tensor("linear_146_cast_fp16")]; - tensor var_3136_to_fp16 = const()[name = tensor("op_3136_to_fp16"), val = tensor(0x1p-1)]; - tensor var_3137_cast_fp16 = mul(x = linear_146_cast_fp16, y = var_3136_to_fp16)[name = tensor("op_3137_cast_fp16")]; - tensor input_863_cast_fp16 = add(x = input_851_cast_fp16, y = var_3137_cast_fp16)[name = tensor("input_863_cast_fp16")]; - tensor query_axes_0 = const()[name = tensor("query_axes_0"), val = tensor([-1])]; - tensor module_layers_16_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_16_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(206860480)))]; - tensor module_layers_16_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_16_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(206861568)))]; - tensor query_cast_fp16 = layer_norm(axes = query_axes_0, beta = module_layers_16_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_16_norm_self_att_weight_to_fp16, x = input_863_cast_fp16)[name = tensor("query_cast_fp16")]; - tensor module_layers_16_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_16_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(206862656)))]; - tensor module_layers_16_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_16_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(207387008)))]; - tensor linear_147_cast_fp16 = linear(bias = module_layers_16_self_attn_linear_q_bias_to_fp16, weight = module_layers_16_self_attn_linear_q_weight_to_fp16, x = query_cast_fp16)[name = tensor("linear_147_cast_fp16")]; - tensor var_3154 = const()[name = tensor("op_3154"), val = tensor([1, -1, 8, 64])]; - tensor q_97_cast_fp16 = reshape(shape = var_3154, x = linear_147_cast_fp16)[name = tensor("q_97_cast_fp16")]; - tensor module_layers_16_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_16_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(207388096)))]; - tensor module_layers_16_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_16_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(207912448)))]; - tensor linear_148_cast_fp16 = linear(bias = module_layers_16_self_attn_linear_k_bias_to_fp16, weight = module_layers_16_self_attn_linear_k_weight_to_fp16, x = query_cast_fp16)[name = tensor("linear_148_cast_fp16")]; - tensor var_3159 = const()[name = tensor("op_3159"), val = tensor([1, -1, 8, 64])]; - tensor k_65_cast_fp16 = reshape(shape = var_3159, x = linear_148_cast_fp16)[name = tensor("k_65_cast_fp16")]; - tensor module_layers_16_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_16_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(207913536)))]; - tensor module_layers_16_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_16_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208437888)))]; - tensor linear_149_cast_fp16 = linear(bias = module_layers_16_self_attn_linear_v_bias_to_fp16, weight = module_layers_16_self_attn_linear_v_weight_to_fp16, x = query_cast_fp16)[name = tensor("linear_149_cast_fp16")]; - tensor var_3164 = const()[name = tensor("op_3164"), val = tensor([1, -1, 8, 64])]; - tensor v_cast_fp16 = reshape(shape = var_3164, x = linear_149_cast_fp16)[name = tensor("v_cast_fp16")]; - tensor value_perm_0 = const()[name = tensor("value_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_16_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_16_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208438976)))]; - tensor var_3176_cast_fp16 = add(x = q_97_cast_fp16, y = module_layers_16_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3176_cast_fp16")]; - tensor module_layers_16_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_16_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208440064)))]; - tensor var_3178_cast_fp16 = add(x = q_97_cast_fp16, y = module_layers_16_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3178_cast_fp16")]; - tensor q_with_bias_v_perm_0 = const()[name = tensor("q_with_bias_v_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_359_transpose_x_0 = const()[name = tensor("x_359_transpose_x_0"), val = tensor(false)]; - tensor x_359_transpose_y_0 = const()[name = tensor("x_359_transpose_y_0"), val = tensor(false)]; - tensor var_3180_to_fp16 = const()[name = tensor("op_3180_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208441152)))]; - tensor q_with_bias_v_cast_fp16 = transpose(perm = q_with_bias_v_perm_0, x = var_3178_cast_fp16)[name = tensor("transpose_91")]; - tensor x_359_cast_fp16 = matmul(transpose_x = x_359_transpose_x_0, transpose_y = x_359_transpose_y_0, x = q_with_bias_v_cast_fp16, y = var_3180_to_fp16)[name = tensor("x_359_cast_fp16")]; - tensor x_361_pad_0 = const()[name = tensor("x_361_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_361_mode_0 = const()[name = tensor("x_361_mode_0"), val = tensor("constant")]; - tensor const_230_to_fp16 = const()[name = tensor("const_230_to_fp16"), val = tensor(0x0p+0)]; - tensor x_361_cast_fp16 = pad(constant_val = const_230_to_fp16, mode = x_361_mode_0, pad = x_361_pad_0, x = x_359_cast_fp16)[name = tensor("x_361_cast_fp16")]; - tensor var_3188 = const()[name = tensor("op_3188"), val = tensor([1, 8, -1, 188])]; - tensor x_363_cast_fp16 = reshape(shape = var_3188, x = x_361_cast_fp16)[name = tensor("x_363_cast_fp16")]; - tensor var_3192_begin_0 = const()[name = tensor("op_3192_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_3192_end_0 = const()[name = tensor("op_3192_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_3192_end_mask_0 = const()[name = tensor("op_3192_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_3192_cast_fp16 = slice_by_index(begin = var_3192_begin_0, end = var_3192_end_0, end_mask = var_3192_end_mask_0, x = x_363_cast_fp16)[name = tensor("op_3192_cast_fp16")]; - tensor var_3193 = const()[name = tensor("op_3193"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_65_cast_fp16 = reshape(shape = var_3193, x = var_3192_cast_fp16)[name = tensor("matrix_bd_65_cast_fp16")]; - tensor matrix_ac_transpose_x_0 = const()[name = tensor("matrix_ac_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_transpose_y_0 = const()[name = tensor("matrix_ac_transpose_y_0"), val = tensor(false)]; - tensor transpose_83_perm_0 = const()[name = tensor("transpose_83_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_84_perm_0 = const()[name = tensor("transpose_84_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_84 = transpose(perm = transpose_84_perm_0, x = k_65_cast_fp16)[name = tensor("transpose_89")]; - tensor transpose_83 = transpose(perm = transpose_83_perm_0, x = var_3176_cast_fp16)[name = tensor("transpose_90")]; - tensor matrix_ac_cast_fp16 = matmul(transpose_x = matrix_ac_transpose_x_0, transpose_y = matrix_ac_transpose_y_0, x = transpose_83, y = transpose_84)[name = tensor("matrix_ac_cast_fp16")]; - tensor matrix_bd_begin_0 = const()[name = tensor("matrix_bd_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_end_0 = const()[name = tensor("matrix_bd_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_end_mask_0 = const()[name = tensor("matrix_bd_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_cast_fp16 = slice_by_index(begin = matrix_bd_begin_0, end = matrix_bd_end_0, end_mask = matrix_bd_end_mask_0, x = matrix_bd_65_cast_fp16)[name = tensor("matrix_bd_cast_fp16")]; - tensor var_3202_cast_fp16 = add(x = matrix_ac_cast_fp16, y = matrix_bd_cast_fp16)[name = tensor("op_3202_cast_fp16")]; - tensor _inversed_scores_65_y_0_to_fp16 = const()[name = tensor("_inversed_scores_65_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_65_cast_fp16 = mul(x = var_3202_cast_fp16, y = _inversed_scores_65_y_0_to_fp16)[name = tensor("_inversed_scores_65_cast_fp16")]; - tensor scores_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_65_cast_fp16, cond = mask_11)[name = tensor("scores_cast_fp16")]; - tensor var_3208_cast_fp16 = softmax(axis = var_23, x = scores_cast_fp16)[name = tensor("op_3208_cast_fp16")]; - tensor input_865_cast_fp16 = select(a = var_11_to_fp16, b = var_3208_cast_fp16, cond = mask_11)[name = tensor("input_865_cast_fp16")]; - tensor x_365_transpose_x_0 = const()[name = tensor("x_365_transpose_x_0"), val = tensor(false)]; - tensor x_365_transpose_y_0 = const()[name = tensor("x_365_transpose_y_0"), val = tensor(false)]; - tensor value_cast_fp16 = transpose(perm = value_perm_0, x = v_cast_fp16)[name = tensor("transpose_92")]; - tensor x_365_cast_fp16 = matmul(transpose_x = x_365_transpose_x_0, transpose_y = x_365_transpose_y_0, x = input_865_cast_fp16, y = value_cast_fp16)[name = tensor("x_365_cast_fp16")]; - tensor var_3212_perm_0 = const()[name = tensor("op_3212_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_3213 = const()[name = tensor("op_3213"), val = tensor([1, -1, 512])]; - tensor var_3212_cast_fp16 = transpose(perm = var_3212_perm_0, x = x_365_cast_fp16)[name = tensor("transpose_88")]; - tensor input_867_cast_fp16 = reshape(shape = var_3213, x = var_3212_cast_fp16)[name = tensor("input_867_cast_fp16")]; - tensor module_layers_16_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_16_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208825216)))]; - tensor module_layers_16_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_16_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(209349568)))]; - tensor linear_151_cast_fp16 = linear(bias = module_layers_16_self_attn_linear_out_bias_to_fp16, weight = module_layers_16_self_attn_linear_out_weight_to_fp16, x = input_867_cast_fp16)[name = tensor("linear_151_cast_fp16")]; - tensor input_871_cast_fp16 = add(x = input_863_cast_fp16, y = linear_151_cast_fp16)[name = tensor("input_871_cast_fp16")]; - tensor x_369_axes_0 = const()[name = tensor("x_369_axes_0"), val = tensor([-1])]; - tensor module_layers_16_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_16_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(209350656)))]; - tensor module_layers_16_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_16_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(209351744)))]; - tensor x_369_cast_fp16 = layer_norm(axes = x_369_axes_0, beta = module_layers_16_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_16_norm_conv_weight_to_fp16, x = input_871_cast_fp16)[name = tensor("x_369_cast_fp16")]; - tensor input_873_perm_0 = const()[name = tensor("input_873_perm_0"), val = tensor([0, 2, 1])]; - tensor input_875_pad_type_0 = const()[name = tensor("input_875_pad_type_0"), val = tensor("valid")]; - tensor input_875_strides_0 = const()[name = tensor("input_875_strides_0"), val = tensor([1])]; - tensor input_875_pad_0 = const()[name = tensor("input_875_pad_0"), val = tensor([0, 0])]; - tensor input_875_dilations_0 = const()[name = tensor("input_875_dilations_0"), val = tensor([1])]; - tensor input_875_groups_0 = const()[name = tensor("input_875_groups_0"), val = tensor(1)]; - tensor module_layers_16_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_16_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(209352832)))]; - tensor module_layers_16_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_16_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210401472)))]; - tensor input_873_cast_fp16 = transpose(perm = input_873_perm_0, x = x_369_cast_fp16)[name = tensor("transpose_87")]; - tensor input_875_cast_fp16 = conv(bias = module_layers_16_conv_pointwise_conv1_bias_to_fp16, dilations = input_875_dilations_0, groups = input_875_groups_0, pad = input_875_pad_0, pad_type = input_875_pad_type_0, strides = input_875_strides_0, weight = module_layers_16_conv_pointwise_conv1_weight_to_fp16, x = input_873_cast_fp16)[name = tensor("input_875_cast_fp16")]; - tensor x_371_split_num_splits_0 = const()[name = tensor("x_371_split_num_splits_0"), val = tensor(2)]; - tensor x_371_split_axis_0 = const()[name = tensor("x_371_split_axis_0"), val = tensor(1)]; - tensor x_371_split_cast_fp16_0, tensor x_371_split_cast_fp16_1 = split(axis = x_371_split_axis_0, num_splits = x_371_split_num_splits_0, x = input_875_cast_fp16)[name = tensor("x_371_split_cast_fp16")]; - tensor x_371_split_1_sigmoid_cast_fp16 = sigmoid(x = x_371_split_cast_fp16_1)[name = tensor("x_371_split_1_sigmoid_cast_fp16")]; - tensor x_371_cast_fp16 = mul(x = x_371_split_cast_fp16_0, y = x_371_split_1_sigmoid_cast_fp16)[name = tensor("x_371_cast_fp16")]; - tensor input_877_cast_fp16 = select(a = var_11_to_fp16, b = x_371_cast_fp16, cond = var_453)[name = tensor("input_877_cast_fp16")]; - tensor input_879_pad_0 = const()[name = tensor("input_879_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_879_mode_0 = const()[name = tensor("input_879_mode_0"), val = tensor("constant")]; - tensor const_233_to_fp16 = const()[name = tensor("const_233_to_fp16"), val = tensor(0x0p+0)]; - tensor input_879_cast_fp16 = pad(constant_val = const_233_to_fp16, mode = input_879_mode_0, pad = input_879_pad_0, x = input_877_cast_fp16)[name = tensor("input_879_cast_fp16")]; - tensor input_881_pad_type_0 = const()[name = tensor("input_881_pad_type_0"), val = tensor("valid")]; - tensor input_881_groups_0 = const()[name = tensor("input_881_groups_0"), val = tensor(512)]; - tensor input_881_strides_0 = const()[name = tensor("input_881_strides_0"), val = tensor([1])]; - tensor input_881_pad_0 = const()[name = tensor("input_881_pad_0"), val = tensor([0, 0])]; - tensor input_881_dilations_0 = const()[name = tensor("input_881_dilations_0"), val = tensor([1])]; - tensor const_266_to_fp16 = const()[name = tensor("const_266_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210403584)))]; - tensor const_267_to_fp16 = const()[name = tensor("const_267_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210412864)))]; - tensor input_883_cast_fp16 = conv(bias = const_267_to_fp16, dilations = input_881_dilations_0, groups = input_881_groups_0, pad = input_881_pad_0, pad_type = input_881_pad_type_0, strides = input_881_strides_0, weight = const_266_to_fp16, x = input_879_cast_fp16)[name = tensor("input_883_cast_fp16")]; - tensor input_885_cast_fp16 = silu(x = input_883_cast_fp16)[name = tensor("input_885_cast_fp16")]; - tensor x_373_pad_type_0 = const()[name = tensor("x_373_pad_type_0"), val = tensor("valid")]; - tensor x_373_strides_0 = const()[name = tensor("x_373_strides_0"), val = tensor([1])]; - tensor x_373_pad_0 = const()[name = tensor("x_373_pad_0"), val = tensor([0, 0])]; - tensor x_373_dilations_0 = const()[name = tensor("x_373_dilations_0"), val = tensor([1])]; - tensor x_373_groups_0 = const()[name = tensor("x_373_groups_0"), val = tensor(1)]; - tensor module_layers_16_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_16_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210413952)))]; - tensor module_layers_16_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_16_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210938304)))]; - tensor x_373_cast_fp16 = conv(bias = module_layers_16_conv_pointwise_conv2_bias_to_fp16, dilations = x_373_dilations_0, groups = x_373_groups_0, pad = x_373_pad_0, pad_type = x_373_pad_type_0, strides = x_373_strides_0, weight = module_layers_16_conv_pointwise_conv2_weight_to_fp16, x = input_885_cast_fp16)[name = tensor("x_373_cast_fp16")]; - tensor input_887_perm_0 = const()[name = tensor("input_887_perm_0"), val = tensor([0, 2, 1])]; - tensor input_887_cast_fp16 = transpose(perm = input_887_perm_0, x = x_373_cast_fp16)[name = tensor("transpose_86")]; - tensor input_889_cast_fp16 = add(x = input_871_cast_fp16, y = input_887_cast_fp16)[name = tensor("input_889_cast_fp16")]; - tensor input_891_axes_0 = const()[name = tensor("input_891_axes_0"), val = tensor([-1])]; - tensor module_layers_16_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_16_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210939392)))]; - tensor module_layers_16_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_16_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210940480)))]; - tensor input_891_cast_fp16 = layer_norm(axes = input_891_axes_0, beta = module_layers_16_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_16_norm_feed_forward2_weight_to_fp16, x = input_889_cast_fp16)[name = tensor("input_891_cast_fp16")]; - tensor module_layers_16_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_16_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210941568)))]; - tensor module_layers_16_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_16_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(213038784)))]; - tensor linear_152_cast_fp16 = linear(bias = module_layers_16_feed_forward2_linear1_bias_to_fp16, weight = module_layers_16_feed_forward2_linear1_weight_to_fp16, x = input_891_cast_fp16)[name = tensor("linear_152_cast_fp16")]; - tensor input_895_cast_fp16 = silu(x = linear_152_cast_fp16)[name = tensor("input_895_cast_fp16")]; - tensor module_layers_16_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_16_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(213042944)))]; - tensor module_layers_16_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_16_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(215140160)))]; - tensor linear_153_cast_fp16 = linear(bias = module_layers_16_feed_forward2_linear2_bias_to_fp16, weight = module_layers_16_feed_forward2_linear2_weight_to_fp16, x = input_895_cast_fp16)[name = tensor("linear_153_cast_fp16")]; - tensor var_3279_to_fp16 = const()[name = tensor("op_3279_to_fp16"), val = tensor(0x1p-1)]; - tensor var_3280_cast_fp16 = mul(x = linear_153_cast_fp16, y = var_3279_to_fp16)[name = tensor("op_3280_cast_fp16")]; - tensor input_cast_fp16 = add(x = input_889_cast_fp16, y = var_3280_cast_fp16)[name = tensor("input_cast_fp16")]; - tensor audio_signal_axes_0 = const()[name = tensor("audio_signal_axes_0"), val = tensor([-1])]; - tensor module_layers_16_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_16_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(215141248)))]; - tensor module_layers_16_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_16_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(215142336)))]; - tensor audio_signal_cast_fp16 = layer_norm(axes = audio_signal_axes_0, beta = module_layers_16_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_16_norm_out_weight_to_fp16, x = input_cast_fp16)[name = tensor("audio_signal_cast_fp16")]; - tensor obj_1_perm_0 = const()[name = tensor("obj_1_perm_0"), val = tensor([0, 2, 1])]; - tensor obj_1_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("obj_1_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; - tensor obj_1_cast_fp16 = transpose(perm = obj_1_perm_0, x = audio_signal_cast_fp16)[name = tensor("transpose_85")]; - tensor encoder_output = cast(dtype = obj_1_cast_fp16_to_fp32_dtype_0, x = obj_1_cast_fp16)[name = tensor("cast_172")]; - } -> (encoder_output, encoder_length); -} \ No newline at end of file diff --git a/parakeet-ctc-110m-coreml/Encoder.mlmodelc/weights/weight.bin b/parakeet-ctc-110m-coreml/Encoder.mlmodelc/weights/weight.bin deleted file mode 100644 index c5bffa9667270bb21c093e55898f4fb2e299e426..0000000000000000000000000000000000000000 --- a/parakeet-ctc-110m-coreml/Encoder.mlmodelc/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:cecf7994b2758397d992802a4f6e5d656e3a1aeb7bbedc2aa430b1316d62474c -size 215143424 diff --git a/parakeet-ctc-110m-coreml/JointDecision.mlmodelc/analytics/coremldata.bin b/parakeet-ctc-110m-coreml/JointDecision.mlmodelc/analytics/coremldata.bin deleted file mode 100644 index 4b4b92e5e8e2b45c67e95237edd74d11f323a65a..0000000000000000000000000000000000000000 --- a/parakeet-ctc-110m-coreml/JointDecision.mlmodelc/analytics/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:983ba26dd9276b8d2d4f75f3475aefb1817c542df87dbd0fdac95bd63647494f -size 243 diff --git a/parakeet-ctc-110m-coreml/JointDecision.mlmodelc/coremldata.bin b/parakeet-ctc-110m-coreml/JointDecision.mlmodelc/coremldata.bin deleted file mode 100644 index 74e1d92e032200815d325fa2c7c84ee8562d6aad..0000000000000000000000000000000000000000 --- a/parakeet-ctc-110m-coreml/JointDecision.mlmodelc/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:0800e3bdf4ecb1bd46fd27e1826d33125cd574f9ae1e15dd9ff70ea42944ca2d -size 476 diff --git a/parakeet-ctc-110m-coreml/JointDecision.mlmodelc/metadata.json b/parakeet-ctc-110m-coreml/JointDecision.mlmodelc/metadata.json deleted file mode 100644 index fe49b70e7e98f7ec9524130b0027886d862f6f8a..0000000000000000000000000000000000000000 --- a/parakeet-ctc-110m-coreml/JointDecision.mlmodelc/metadata.json +++ /dev/null @@ -1,102 +0,0 @@ -[ - { - "metadataOutputVersion" : "3.0", - "shortDescription" : "Parakeet 110M joint + decision head (split, softmax, argmax)", - "outputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1 × 188 × 1)", - "shortDescription" : "", - "shape" : "[1, 188, 1]", - "name" : "token_id", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 188 × 1)", - "shortDescription" : "", - "shape" : "[1, 188, 1]", - "name" : "token_prob", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1 × 188 × 1)", - "shortDescription" : "", - "shape" : "[1, 188, 1]", - "name" : "duration", - "type" : "MultiArray" - } - ], - "storagePrecision" : "Float16", - "modelParameters" : [ - - ], - "author" : "Fluid Inference", - "specificationVersion" : 8, - "mlProgramOperationTypeHistogram" : { - "Ios17.reduceArgmax" : 2, - "Ios17.squeeze" : 1, - "Ios17.cast" : 4, - "Ios17.linear" : 3, - "Ios17.transpose" : 2, - "Ios17.sliceByIndex" : 2, - "Ios17.add" : 1, - "Ios16.relu" : 1, - "Ios16.softmax" : 1, - "Ios17.gatherAlongAxis" : 1, - "Ios17.expandDims" : 3 - }, - "computePrecision" : "Mixed (Float16, Float32, Int16, Int32)", - "isUpdatable" : "0", - "stateSchema" : [ - - ], - "availability" : { - "macOS" : "14.0", - "tvOS" : "17.0", - "visionOS" : "1.0", - "watchOS" : "10.0", - "iOS" : "17.0", - "macCatalyst" : "17.0" - }, - "modelType" : { - "name" : "MLModelType_mlProgram" - }, - "inputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 512 × 188)", - "shortDescription" : "", - "shape" : "[1, 512, 188]", - "name" : "encoder", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 640 × 1)", - "shortDescription" : "", - "shape" : "[1, 640, 1]", - "name" : "decoder", - "type" : "MultiArray" - } - ], - "userDefinedMetadata" : { - "com.github.apple.coremltools.version" : "8.3.0", - "com.github.apple.coremltools.source_dialect" : "TorchScript", - "com.github.apple.coremltools.source" : "torch==2.9.0" - }, - "generatedClassName" : "parakeet_joint_decision", - "method" : "predict" - } -] \ No newline at end of file diff --git a/parakeet-ctc-110m-coreml/JointDecision.mlmodelc/model.mil b/parakeet-ctc-110m-coreml/JointDecision.mlmodelc/model.mil deleted file mode 100644 index edfc5050f47ffbe1f5344799cef5dc0b21c392e0..0000000000000000000000000000000000000000 --- a/parakeet-ctc-110m-coreml/JointDecision.mlmodelc/model.mil +++ /dev/null @@ -1,58 +0,0 @@ -program(1.0) -[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3510.2.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.9.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "8.3.0"}})] -{ - func main(tensor decoder, tensor encoder) { - tensor input_1_perm_0 = const()[name = tensor("input_1_perm_0"), val = tensor([0, 2, 1])]; - tensor encoder_to_fp16_dtype_0 = const()[name = tensor("encoder_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor input_3_perm_0 = const()[name = tensor("input_3_perm_0"), val = tensor([0, 2, 1])]; - tensor decoder_to_fp16_dtype_0 = const()[name = tensor("decoder_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor joint_module_enc_weight_to_fp16 = const()[name = tensor("joint_module_enc_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; - tensor joint_module_enc_bias_to_fp16 = const()[name = tensor("joint_module_enc_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(655488)))]; - tensor encoder_to_fp16 = cast(dtype = encoder_to_fp16_dtype_0, x = encoder)[name = tensor("cast_6")]; - tensor input_1_cast_fp16 = transpose(perm = input_1_perm_0, x = encoder_to_fp16)[name = tensor("transpose_1")]; - tensor linear_0_cast_fp16 = linear(bias = joint_module_enc_bias_to_fp16, weight = joint_module_enc_weight_to_fp16, x = input_1_cast_fp16)[name = tensor("linear_0_cast_fp16")]; - tensor joint_module_pred_weight_to_fp16 = const()[name = tensor("joint_module_pred_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(656832)))]; - tensor joint_module_pred_bias_to_fp16 = const()[name = tensor("joint_module_pred_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1476096)))]; - tensor decoder_to_fp16 = cast(dtype = decoder_to_fp16_dtype_0, x = decoder)[name = tensor("cast_5")]; - tensor input_3_cast_fp16 = transpose(perm = input_3_perm_0, x = decoder_to_fp16)[name = tensor("transpose_0")]; - tensor linear_1_cast_fp16 = linear(bias = joint_module_pred_bias_to_fp16, weight = joint_module_pred_weight_to_fp16, x = input_3_cast_fp16)[name = tensor("linear_1_cast_fp16")]; - tensor var_23_axes_0 = const()[name = tensor("op_23_axes_0"), val = tensor([2])]; - tensor var_23_cast_fp16 = expand_dims(axes = var_23_axes_0, x = linear_0_cast_fp16)[name = tensor("op_23_cast_fp16")]; - tensor var_24_axes_0 = const()[name = tensor("op_24_axes_0"), val = tensor([1])]; - tensor var_24_cast_fp16 = expand_dims(axes = var_24_axes_0, x = linear_1_cast_fp16)[name = tensor("op_24_cast_fp16")]; - tensor input_5_cast_fp16 = add(x = var_23_cast_fp16, y = var_24_cast_fp16)[name = tensor("input_5_cast_fp16")]; - tensor input_7_cast_fp16 = relu(x = input_5_cast_fp16)[name = tensor("input_7_cast_fp16")]; - tensor joint_module_joint_net_2_weight_to_fp16 = const()[name = tensor("joint_module_joint_net_2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1477440)))]; - tensor joint_module_joint_net_2_bias_to_fp16 = const()[name = tensor("joint_module_joint_net_2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2795904)))]; - tensor linear_2_cast_fp16 = linear(bias = joint_module_joint_net_2_bias_to_fp16, weight = joint_module_joint_net_2_weight_to_fp16, x = input_7_cast_fp16)[name = tensor("linear_2_cast_fp16")]; - tensor token_logits_begin_0 = const()[name = tensor("token_logits_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor token_logits_end_0 = const()[name = tensor("token_logits_end_0"), val = tensor([1, 188, 1, 1025])]; - tensor token_logits_end_mask_0 = const()[name = tensor("token_logits_end_mask_0"), val = tensor([true, true, true, false])]; - tensor token_logits_cast_fp16 = slice_by_index(begin = token_logits_begin_0, end = token_logits_end_0, end_mask = token_logits_end_mask_0, x = linear_2_cast_fp16)[name = tensor("token_logits_cast_fp16")]; - tensor duration_logits_begin_0 = const()[name = tensor("duration_logits_begin_0"), val = tensor([0, 0, 0, 1025])]; - tensor duration_logits_end_0 = const()[name = tensor("duration_logits_end_0"), val = tensor([1, 188, 1, 1030])]; - tensor duration_logits_end_mask_0 = const()[name = tensor("duration_logits_end_mask_0"), val = tensor([true, true, true, true])]; - tensor duration_logits_cast_fp16 = slice_by_index(begin = duration_logits_begin_0, end = duration_logits_end_0, end_mask = duration_logits_end_mask_0, x = linear_2_cast_fp16)[name = tensor("duration_logits_cast_fp16")]; - tensor var_43_axis_0 = const()[name = tensor("op_43_axis_0"), val = tensor(-1)]; - tensor var_43_keep_dims_0 = const()[name = tensor("op_43_keep_dims_0"), val = tensor(false)]; - tensor var_43_output_dtype_0 = const()[name = tensor("op_43_output_dtype_0"), val = tensor("int32")]; - tensor token_id = reduce_argmax(axis = var_43_axis_0, keep_dims = var_43_keep_dims_0, output_dtype = var_43_output_dtype_0, x = token_logits_cast_fp16)[name = tensor("op_43_cast_fp16")]; - tensor var_49 = const()[name = tensor("op_49"), val = tensor(-1)]; - tensor token_probs_all_cast_fp16 = softmax(axis = var_49, x = token_logits_cast_fp16)[name = tensor("token_probs_all_cast_fp16")]; - tensor var_58_axes_0 = const()[name = tensor("op_58_axes_0"), val = tensor([-1])]; - tensor var_58 = expand_dims(axes = var_58_axes_0, x = token_id)[name = tensor("op_58")]; - tensor var_59 = const()[name = tensor("op_59"), val = tensor(-1)]; - tensor var_61_validate_indices_0 = const()[name = tensor("op_61_validate_indices_0"), val = tensor(false)]; - tensor var_58_to_int16_dtype_0 = const()[name = tensor("op_58_to_int16_dtype_0"), val = tensor("int16")]; - tensor var_58_to_int16 = cast(dtype = var_58_to_int16_dtype_0, x = var_58)[name = tensor("cast_4")]; - tensor var_61_cast_fp16_cast_int16 = gather_along_axis(axis = var_59, indices = var_58_to_int16, validate_indices = var_61_validate_indices_0, x = token_probs_all_cast_fp16)[name = tensor("op_61_cast_fp16_cast_int16")]; - tensor var_63_axes_0 = const()[name = tensor("op_63_axes_0"), val = tensor([-1])]; - tensor var_63_cast_fp16 = squeeze(axes = var_63_axes_0, x = var_61_cast_fp16_cast_int16)[name = tensor("op_63_cast_fp16")]; - tensor var_63_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("op_63_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; - tensor var_66_axis_0 = const()[name = tensor("op_66_axis_0"), val = tensor(-1)]; - tensor var_66_keep_dims_0 = const()[name = tensor("op_66_keep_dims_0"), val = tensor(false)]; - tensor var_66_output_dtype_0 = const()[name = tensor("op_66_output_dtype_0"), val = tensor("int32")]; - tensor duration = reduce_argmax(axis = var_66_axis_0, keep_dims = var_66_keep_dims_0, output_dtype = var_66_output_dtype_0, x = duration_logits_cast_fp16)[name = tensor("op_66_cast_fp16")]; - tensor token_prob = cast(dtype = var_63_cast_fp16_to_fp32_dtype_0, x = var_63_cast_fp16)[name = tensor("cast_3")]; - } -> (token_id, token_prob, duration); -} \ No newline at end of file diff --git a/parakeet-ctc-110m-coreml/JointDecision.mlmodelc/weights/weight.bin b/parakeet-ctc-110m-coreml/JointDecision.mlmodelc/weights/weight.bin deleted file mode 100644 index 747ec9d24e2d3e008bc869d2c404bb3d2653325c..0000000000000000000000000000000000000000 --- a/parakeet-ctc-110m-coreml/JointDecision.mlmodelc/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:b3f771cb65b190f1873e39629676ed79b65a8361522f451b37bdba8b1106e6ff -size 2798028 diff --git a/parakeet-ctc-110m-coreml/JointDecisionSingleStep.mlmodelc/analytics/coremldata.bin b/parakeet-ctc-110m-coreml/JointDecisionSingleStep.mlmodelc/analytics/coremldata.bin deleted file mode 100644 index ab3e67b973cf846c0af5e1f075f1b4b7ffb77a00..0000000000000000000000000000000000000000 --- a/parakeet-ctc-110m-coreml/JointDecisionSingleStep.mlmodelc/analytics/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:c7c11c6bb985fab7f835ba687a575f1eb04f4c93b0783155d634adbc49f0e797 -size 243 diff --git a/parakeet-ctc-110m-coreml/JointDecisionSingleStep.mlmodelc/coremldata.bin b/parakeet-ctc-110m-coreml/JointDecisionSingleStep.mlmodelc/coremldata.bin deleted file mode 100644 index bc71cc6aa73de6959f4c6718500197ebbf364633..0000000000000000000000000000000000000000 --- a/parakeet-ctc-110m-coreml/JointDecisionSingleStep.mlmodelc/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:1af2cb9bcc13eec83ce006e4f1c2cf158393745cd9187428333fbcb6917da244 -size 535 diff --git a/parakeet-ctc-110m-coreml/JointDecisionSingleStep.mlmodelc/metadata.json b/parakeet-ctc-110m-coreml/JointDecisionSingleStep.mlmodelc/metadata.json deleted file mode 100644 index b96db3615aa1d5e110c61c1f4694def827cb0a1f..0000000000000000000000000000000000000000 --- a/parakeet-ctc-110m-coreml/JointDecisionSingleStep.mlmodelc/metadata.json +++ /dev/null @@ -1,123 +0,0 @@ -[ - { - "metadataOutputVersion" : "3.0", - "shortDescription" : "Parakeet 110M single-step joint decision (current frame)", - "outputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1 × 1 × 1)", - "shortDescription" : "", - "shape" : "[1, 1, 1]", - "name" : "token_id", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 1 × 1)", - "shortDescription" : "", - "shape" : "[1, 1, 1]", - "name" : "token_prob", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1 × 1 × 1)", - "shortDescription" : "", - "shape" : "[1, 1, 1]", - "name" : "duration", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1 × 1 × 1 × 64)", - "shortDescription" : "", - "shape" : "[1, 1, 1, 64]", - "name" : "top_k_ids", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 1 × 1 × 64)", - "shortDescription" : "", - "shape" : "[1, 1, 1, 64]", - "name" : "top_k_logits", - "type" : "MultiArray" - } - ], - "storagePrecision" : "Float16", - "modelParameters" : [ - - ], - "author" : "Fluid Inference", - "specificationVersion" : 8, - "mlProgramOperationTypeHistogram" : { - "Ios17.reduceArgmax" : 2, - "Ios17.linear" : 3, - "Ios17.transpose" : 2, - "Ios17.sliceByIndex" : 2, - "Ios17.add" : 1, - "Ios17.topk" : 1, - "Ios16.relu" : 1, - "Ios16.softmax" : 1, - "Ios17.expandDims" : 3, - "Ios17.squeeze" : 1, - "Ios17.cast" : 6, - "Ios17.gatherAlongAxis" : 1 - }, - "computePrecision" : "Mixed (Float16, Float32, Int16, Int32, UInt16)", - "isUpdatable" : "0", - "stateSchema" : [ - - ], - "availability" : { - "macOS" : "14.0", - "tvOS" : "17.0", - "visionOS" : "1.0", - "watchOS" : "10.0", - "iOS" : "17.0", - "macCatalyst" : "17.0" - }, - "modelType" : { - "name" : "MLModelType_mlProgram" - }, - "inputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 512 × 1)", - "shortDescription" : "", - "shape" : "[1, 512, 1]", - "name" : "encoder_step", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 640 × 1)", - "shortDescription" : "", - "shape" : "[1, 640, 1]", - "name" : "decoder_step", - "type" : "MultiArray" - } - ], - "userDefinedMetadata" : { - "com.github.apple.coremltools.source_dialect" : "TorchScript", - "com.github.apple.coremltools.source" : "torch==2.9.0", - "com.github.apple.coremltools.version" : "8.3.0" - }, - "generatedClassName" : "parakeet_joint_decision_single_step", - "method" : "predict" - } -] \ No newline at end of file diff --git a/parakeet-ctc-110m-coreml/JointDecisionSingleStep.mlmodelc/model.mil b/parakeet-ctc-110m-coreml/JointDecisionSingleStep.mlmodelc/model.mil deleted file mode 100644 index 42bd4ac5def601f37ece871ff09bd6242de025aa..0000000000000000000000000000000000000000 --- a/parakeet-ctc-110m-coreml/JointDecisionSingleStep.mlmodelc/model.mil +++ /dev/null @@ -1,69 +0,0 @@ -program(1.0) -[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3510.2.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.9.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "8.3.0"}})] -{ - func main(tensor decoder_step, tensor encoder_step) { - tensor input_1_perm_0 = const()[name = tensor("input_1_perm_0"), val = tensor([0, 2, 1])]; - tensor encoder_step_to_fp16_dtype_0 = const()[name = tensor("encoder_step_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor input_3_perm_0 = const()[name = tensor("input_3_perm_0"), val = tensor([0, 2, 1])]; - tensor decoder_step_to_fp16_dtype_0 = const()[name = tensor("decoder_step_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor joint_module_enc_weight_to_fp16 = const()[name = tensor("joint_module_enc_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; - tensor joint_module_enc_bias_to_fp16 = const()[name = tensor("joint_module_enc_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(655488)))]; - tensor encoder_step_to_fp16 = cast(dtype = encoder_step_to_fp16_dtype_0, x = encoder_step)[name = tensor("cast_9")]; - tensor input_1_cast_fp16 = transpose(perm = input_1_perm_0, x = encoder_step_to_fp16)[name = tensor("transpose_1")]; - tensor linear_0_cast_fp16 = linear(bias = joint_module_enc_bias_to_fp16, weight = joint_module_enc_weight_to_fp16, x = input_1_cast_fp16)[name = tensor("linear_0_cast_fp16")]; - tensor joint_module_pred_weight_to_fp16 = const()[name = tensor("joint_module_pred_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(656832)))]; - tensor joint_module_pred_bias_to_fp16 = const()[name = tensor("joint_module_pred_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1476096)))]; - tensor decoder_step_to_fp16 = cast(dtype = decoder_step_to_fp16_dtype_0, x = decoder_step)[name = tensor("cast_8")]; - tensor input_3_cast_fp16 = transpose(perm = input_3_perm_0, x = decoder_step_to_fp16)[name = tensor("transpose_0")]; - tensor linear_1_cast_fp16 = linear(bias = joint_module_pred_bias_to_fp16, weight = joint_module_pred_weight_to_fp16, x = input_3_cast_fp16)[name = tensor("linear_1_cast_fp16")]; - tensor var_23_axes_0 = const()[name = tensor("op_23_axes_0"), val = tensor([2])]; - tensor var_23_cast_fp16 = expand_dims(axes = var_23_axes_0, x = linear_0_cast_fp16)[name = tensor("op_23_cast_fp16")]; - tensor var_24_axes_0 = const()[name = tensor("op_24_axes_0"), val = tensor([1])]; - tensor var_24_cast_fp16 = expand_dims(axes = var_24_axes_0, x = linear_1_cast_fp16)[name = tensor("op_24_cast_fp16")]; - tensor input_5_cast_fp16 = add(x = var_23_cast_fp16, y = var_24_cast_fp16)[name = tensor("input_5_cast_fp16")]; - tensor input_7_cast_fp16 = relu(x = input_5_cast_fp16)[name = tensor("input_7_cast_fp16")]; - tensor joint_module_joint_net_2_weight_to_fp16 = const()[name = tensor("joint_module_joint_net_2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1477440)))]; - tensor joint_module_joint_net_2_bias_to_fp16 = const()[name = tensor("joint_module_joint_net_2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2795904)))]; - tensor linear_2_cast_fp16 = linear(bias = joint_module_joint_net_2_bias_to_fp16, weight = joint_module_joint_net_2_weight_to_fp16, x = input_7_cast_fp16)[name = tensor("linear_2_cast_fp16")]; - tensor token_logits_begin_0 = const()[name = tensor("token_logits_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor token_logits_end_0 = const()[name = tensor("token_logits_end_0"), val = tensor([1, 1, 1, 1025])]; - tensor token_logits_end_mask_0 = const()[name = tensor("token_logits_end_mask_0"), val = tensor([true, true, true, false])]; - tensor token_logits_cast_fp16 = slice_by_index(begin = token_logits_begin_0, end = token_logits_end_0, end_mask = token_logits_end_mask_0, x = linear_2_cast_fp16)[name = tensor("token_logits_cast_fp16")]; - tensor duration_logits_begin_0 = const()[name = tensor("duration_logits_begin_0"), val = tensor([0, 0, 0, 1025])]; - tensor duration_logits_end_0 = const()[name = tensor("duration_logits_end_0"), val = tensor([1, 1, 1, 1030])]; - tensor duration_logits_end_mask_0 = const()[name = tensor("duration_logits_end_mask_0"), val = tensor([true, true, true, true])]; - tensor duration_logits_cast_fp16 = slice_by_index(begin = duration_logits_begin_0, end = duration_logits_end_0, end_mask = duration_logits_end_mask_0, x = linear_2_cast_fp16)[name = tensor("duration_logits_cast_fp16")]; - tensor var_43_axis_0 = const()[name = tensor("op_43_axis_0"), val = tensor(-1)]; - tensor var_43_keep_dims_0 = const()[name = tensor("op_43_keep_dims_0"), val = tensor(false)]; - tensor var_43_output_dtype_0 = const()[name = tensor("op_43_output_dtype_0"), val = tensor("int32")]; - tensor token_id = reduce_argmax(axis = var_43_axis_0, keep_dims = var_43_keep_dims_0, output_dtype = var_43_output_dtype_0, x = token_logits_cast_fp16)[name = tensor("op_43_cast_fp16")]; - tensor var_49 = const()[name = tensor("op_49"), val = tensor(-1)]; - tensor token_probs_all_cast_fp16 = softmax(axis = var_49, x = token_logits_cast_fp16)[name = tensor("token_probs_all_cast_fp16")]; - tensor var_58_axes_0 = const()[name = tensor("op_58_axes_0"), val = tensor([-1])]; - tensor var_58 = expand_dims(axes = var_58_axes_0, x = token_id)[name = tensor("op_58")]; - tensor var_59 = const()[name = tensor("op_59"), val = tensor(-1)]; - tensor var_61_validate_indices_0 = const()[name = tensor("op_61_validate_indices_0"), val = tensor(false)]; - tensor var_58_to_int16_dtype_0 = const()[name = tensor("op_58_to_int16_dtype_0"), val = tensor("int16")]; - tensor var_58_to_int16 = cast(dtype = var_58_to_int16_dtype_0, x = var_58)[name = tensor("cast_7")]; - tensor var_61_cast_fp16_cast_int16 = gather_along_axis(axis = var_59, indices = var_58_to_int16, validate_indices = var_61_validate_indices_0, x = token_probs_all_cast_fp16)[name = tensor("op_61_cast_fp16_cast_int16")]; - tensor var_63_axes_0 = const()[name = tensor("op_63_axes_0"), val = tensor([-1])]; - tensor var_63_cast_fp16 = squeeze(axes = var_63_axes_0, x = var_61_cast_fp16_cast_int16)[name = tensor("op_63_cast_fp16")]; - tensor var_63_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("op_63_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; - tensor var_66_axis_0 = const()[name = tensor("op_66_axis_0"), val = tensor(-1)]; - tensor var_66_keep_dims_0 = const()[name = tensor("op_66_keep_dims_0"), val = tensor(false)]; - tensor var_66_output_dtype_0 = const()[name = tensor("op_66_output_dtype_0"), val = tensor("int32")]; - tensor duration = reduce_argmax(axis = var_66_axis_0, keep_dims = var_66_keep_dims_0, output_dtype = var_66_output_dtype_0, x = duration_logits_cast_fp16)[name = tensor("op_66_cast_fp16")]; - tensor var_72 = const()[name = tensor("op_72"), val = tensor(64)]; - tensor var_76_axis_0 = const()[name = tensor("op_76_axis_0"), val = tensor(-1)]; - tensor var_76_ascending_0 = const()[name = tensor("op_76_ascending_0"), val = tensor(false)]; - tensor var_76_sort_0 = const()[name = tensor("op_76_sort_0"), val = tensor(true)]; - tensor var_76_return_indices_0 = const()[name = tensor("op_76_return_indices_0"), val = tensor(true)]; - tensor var_76_cast_fp16_cast_int16_output_indices_dtype_0 = const()[name = tensor("op_76_cast_fp16_cast_int16_output_indices_dtype_0"), val = tensor("uint16")]; - tensor var_76_cast_fp16_cast_int16_0, tensor var_76_cast_fp16_cast_int16_1 = topk(ascending = var_76_ascending_0, axis = var_76_axis_0, k = var_72, output_indices_dtype = var_76_cast_fp16_cast_int16_output_indices_dtype_0, return_indices = var_76_return_indices_0, sort = var_76_sort_0, x = token_logits_cast_fp16)[name = tensor("op_76_cast_fp16_cast_int16")]; - tensor var_76_cast_fp16_cast_int16_1_to_int32_dtype_0 = const()[name = tensor("op_76_cast_fp16_cast_int16_1_to_int32_dtype_0"), val = tensor("int32")]; - tensor var_76_cast_fp16_0_to_fp32_dtype_0 = const()[name = tensor("op_76_cast_fp16_0_to_fp32_dtype_0"), val = tensor("fp32")]; - tensor top_k_logits = cast(dtype = var_76_cast_fp16_0_to_fp32_dtype_0, x = var_76_cast_fp16_cast_int16_0)[name = tensor("cast_4")]; - tensor top_k_ids = cast(dtype = var_76_cast_fp16_cast_int16_1_to_int32_dtype_0, x = var_76_cast_fp16_cast_int16_1)[name = tensor("cast_5")]; - tensor token_prob = cast(dtype = var_63_cast_fp16_to_fp32_dtype_0, x = var_63_cast_fp16)[name = tensor("cast_6")]; - } -> (token_id, token_prob, duration, top_k_ids, top_k_logits); -} \ No newline at end of file diff --git a/parakeet-ctc-110m-coreml/JointDecisionSingleStep.mlmodelc/weights/weight.bin b/parakeet-ctc-110m-coreml/JointDecisionSingleStep.mlmodelc/weights/weight.bin deleted file mode 100644 index 747ec9d24e2d3e008bc869d2c404bb3d2653325c..0000000000000000000000000000000000000000 --- a/parakeet-ctc-110m-coreml/JointDecisionSingleStep.mlmodelc/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:b3f771cb65b190f1873e39629676ed79b65a8361522f451b37bdba8b1106e6ff -size 2798028 diff --git a/parakeet-ctc-110m-coreml/MelSpectrogram.mlmodelc/analytics/coremldata.bin b/parakeet-ctc-110m-coreml/MelSpectrogram.mlmodelc/analytics/coremldata.bin deleted file mode 100644 index 9f2576cb0d17fa1f6d2f472a8331a3b4e3fa2a39..0000000000000000000000000000000000000000 --- a/parakeet-ctc-110m-coreml/MelSpectrogram.mlmodelc/analytics/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:22f2a8cba1de25c984050566b534a1d8caf22a82f9fe6c1c6f3149a0dd7e8ae3 -size 243 diff --git a/parakeet-ctc-110m-coreml/MelSpectrogram.mlmodelc/coremldata.bin b/parakeet-ctc-110m-coreml/MelSpectrogram.mlmodelc/coremldata.bin deleted file mode 100644 index 55e0cbd61f46a2d6e3d4ba0691275ec7eb44730d..0000000000000000000000000000000000000000 --- a/parakeet-ctc-110m-coreml/MelSpectrogram.mlmodelc/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:3a32ec67c76aa0aa2faef518413c311493e89aeb7fa11289fa4b8653ab8a160c -size 330 diff --git a/parakeet-ctc-110m-coreml/MelSpectrogram.mlmodelc/metadata.json b/parakeet-ctc-110m-coreml/MelSpectrogram.mlmodelc/metadata.json deleted file mode 100644 index 944c0c11d0e59dff561e97c44e5e47127a9e953b..0000000000000000000000000000000000000000 --- a/parakeet-ctc-110m-coreml/MelSpectrogram.mlmodelc/metadata.json +++ /dev/null @@ -1,76 +0,0 @@ -[ - { - "metadataOutputVersion" : "3.0", - "storagePrecision" : "Float16", - "outputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float16", - "formattedType" : "MultiArray (Float16 1 × 1 × 1501 × 80)", - "shortDescription" : "", - "shape" : "[1, 1, 1501, 80]", - "name" : "melspectrogram_features", - "type" : "MultiArray" - } - ], - "modelParameters" : [ - - ], - "specificationVersion" : 8, - "mlProgramOperationTypeHistogram" : { - "Ios17.mul" : 2, - "Ios17.sqrt" : 1, - "Ios17.square" : 3, - "Ios17.transpose" : 1, - "Ios17.sub" : 2, - "Ios17.matmul" : 1, - "Ios17.conv" : 2, - "Ios17.log" : 1, - "Ios17.sliceByIndex" : 2, - "Ios17.add" : 3, - "Ios16.reduceMean" : 2, - "Ios17.realDiv" : 1, - "Ios17.expandDims" : 4, - "Ios17.squeeze" : 2, - "Ios17.reshape" : 2, - "Identity" : 1, - "Pad" : 2 - }, - "computePrecision" : "Mixed (Float16, Float32, Int32)", - "isUpdatable" : "0", - "stateSchema" : [ - - ], - "availability" : { - "macOS" : "14.0", - "tvOS" : "17.0", - "visionOS" : "1.0", - "watchOS" : "10.0", - "iOS" : "17.0", - "macCatalyst" : "17.0" - }, - "modelType" : { - "name" : "MLModelType_mlProgram" - }, - "userDefinedMetadata" : { - "com.github.apple.coremltools.source_dialect" : "TorchScript", - "com.github.apple.coremltools.source" : "torch==2.5.0", - "com.github.apple.coremltools.version" : "8.3.0" - }, - "inputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float16", - "formattedType" : "MultiArray (Float16 240000)", - "shortDescription" : "", - "shape" : "[240000]", - "name" : "audio", - "type" : "MultiArray" - } - ], - "generatedClassName" : "MelSpectrogram", - "method" : "predict" - } -] \ No newline at end of file diff --git a/parakeet-ctc-110m-coreml/MelSpectrogram.mlmodelc/model.mil b/parakeet-ctc-110m-coreml/MelSpectrogram.mlmodelc/model.mil deleted file mode 100644 index d1ef7d81e60edf7bc921ade3f5b602a471e8609b..0000000000000000000000000000000000000000 --- a/parakeet-ctc-110m-coreml/MelSpectrogram.mlmodelc/model.mil +++ /dev/null @@ -1,82 +0,0 @@ -program(1.0) -[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3404.16.1"}, {"coremlc-version", "3404.23.1"}, {"coremltools-component-torch", "2.5.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "8.3.0"}})] -{ - func main(tensor audio) { - tensor var_8_begin_0 = const()[name = tensor("op_8_begin_0"), val = tensor([1])]; - tensor var_8_end_0 = const()[name = tensor("op_8_end_0"), val = tensor([240000])]; - tensor var_8_end_mask_0 = const()[name = tensor("op_8_end_mask_0"), val = tensor([true])]; - tensor var_8_cast_fp16 = slice_by_index(begin = var_8_begin_0, end = var_8_end_0, end_mask = var_8_end_mask_0, x = audio)[name = tensor("op_8_cast_fp16")]; - tensor var_13_begin_0 = const()[name = tensor("op_13_begin_0"), val = tensor([0])]; - tensor var_13_end_0 = const()[name = tensor("op_13_end_0"), val = tensor([239999])]; - tensor var_13_end_mask_0 = const()[name = tensor("op_13_end_mask_0"), val = tensor([false])]; - tensor var_13_cast_fp16 = slice_by_index(begin = var_13_begin_0, end = var_13_end_0, end_mask = var_13_end_mask_0, x = audio)[name = tensor("op_13_cast_fp16")]; - tensor var_14_to_fp16 = const()[name = tensor("op_14_to_fp16"), val = tensor(0x1.f0cp-1)]; - tensor var_15_cast_fp16 = mul(x = var_13_cast_fp16, y = var_14_to_fp16)[name = tensor("op_15_cast_fp16")]; - tensor input_1_cast_fp16 = sub(x = var_8_cast_fp16, y = var_15_cast_fp16)[name = tensor("input_1_cast_fp16")]; - tensor input_3_pad_0 = const()[name = tensor("input_3_pad_0"), val = tensor([1, 0])]; - tensor input_3_mode_0 = const()[name = tensor("input_3_mode_0"), val = tensor("constant")]; - tensor const_0_to_fp16 = const()[name = tensor("const_0_to_fp16"), val = tensor(0x0p+0)]; - tensor input_3_cast_fp16 = pad(constant_val = const_0_to_fp16, mode = input_3_mode_0, pad = input_3_pad_0, x = input_1_cast_fp16)[name = tensor("input_3_cast_fp16")]; - tensor var_30 = const()[name = tensor("op_30"), val = tensor([1, 1, 240000])]; - tensor input_5_cast_fp16 = reshape(shape = var_30, x = input_3_cast_fp16)[name = tensor("input_5_cast_fp16")]; - tensor input_7_pad_0 = const()[name = tensor("input_7_pad_0"), val = tensor([0, 0, 0, 0, 256, 256])]; - tensor input_7_mode_0 = const()[name = tensor("input_7_mode_0"), val = tensor("reflect")]; - tensor const_2_to_fp16 = const()[name = tensor("const_2_to_fp16"), val = tensor(0x0p+0)]; - tensor input_7_cast_fp16 = pad(constant_val = const_2_to_fp16, mode = input_7_mode_0, pad = input_7_pad_0, x = input_5_cast_fp16)[name = tensor("input_7_cast_fp16")]; - tensor var_42 = const()[name = tensor("op_42"), val = tensor([240512])]; - tensor input_cast_fp16 = reshape(shape = var_42, x = input_7_cast_fp16)[name = tensor("input_cast_fp16")]; - tensor expand_dims_0_axes_0 = const()[name = tensor("expand_dims_0_axes_0"), val = tensor([0])]; - tensor expand_dims_0_cast_fp16 = expand_dims(axes = expand_dims_0_axes_0, x = input_cast_fp16)[name = tensor("expand_dims_0_cast_fp16")]; - tensor expand_dims_3 = const()[name = tensor("expand_dims_3"), val = tensor([160])]; - tensor expand_dims_4_axes_0 = const()[name = tensor("expand_dims_4_axes_0"), val = tensor([1])]; - tensor expand_dims_4_cast_fp16 = expand_dims(axes = expand_dims_4_axes_0, x = expand_dims_0_cast_fp16)[name = tensor("expand_dims_4_cast_fp16")]; - tensor conv_0_pad_type_0 = const()[name = tensor("conv_0_pad_type_0"), val = tensor("valid")]; - tensor conv_0_pad_0 = const()[name = tensor("conv_0_pad_0"), val = tensor([0, 0])]; - tensor conv_0_dilations_0 = const()[name = tensor("conv_0_dilations_0"), val = tensor([1])]; - tensor conv_0_groups_0 = const()[name = tensor("conv_0_groups_0"), val = tensor(1)]; - tensor expand_dims_1_to_fp16 = const()[name = tensor("expand_dims_1_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; - tensor conv_0_cast_fp16 = conv(dilations = conv_0_dilations_0, groups = conv_0_groups_0, pad = conv_0_pad_0, pad_type = conv_0_pad_type_0, strides = expand_dims_3, weight = expand_dims_1_to_fp16, x = expand_dims_4_cast_fp16)[name = tensor("conv_0_cast_fp16")]; - tensor conv_1_pad_type_0 = const()[name = tensor("conv_1_pad_type_0"), val = tensor("valid")]; - tensor conv_1_pad_0 = const()[name = tensor("conv_1_pad_0"), val = tensor([0, 0])]; - tensor conv_1_dilations_0 = const()[name = tensor("conv_1_dilations_0"), val = tensor([1])]; - tensor conv_1_groups_0 = const()[name = tensor("conv_1_groups_0"), val = tensor(1)]; - tensor expand_dims_2_to_fp16 = const()[name = tensor("expand_dims_2_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(263296)))]; - tensor conv_1_cast_fp16 = conv(dilations = conv_1_dilations_0, groups = conv_1_groups_0, pad = conv_1_pad_0, pad_type = conv_1_pad_type_0, strides = expand_dims_3, weight = expand_dims_2_to_fp16, x = expand_dims_4_cast_fp16)[name = tensor("conv_1_cast_fp16")]; - tensor squeeze_0_axes_0 = const()[name = tensor("squeeze_0_axes_0"), val = tensor([0])]; - tensor squeeze_0_cast_fp16 = squeeze(axes = squeeze_0_axes_0, x = conv_0_cast_fp16)[name = tensor("squeeze_0_cast_fp16")]; - tensor squeeze_1_axes_0 = const()[name = tensor("squeeze_1_axes_0"), val = tensor([0])]; - tensor squeeze_1_cast_fp16 = squeeze(axes = squeeze_1_axes_0, x = conv_1_cast_fp16)[name = tensor("squeeze_1_cast_fp16")]; - tensor square_1_cast_fp16 = square(x = squeeze_0_cast_fp16)[name = tensor("square_1_cast_fp16")]; - tensor square_2_cast_fp16 = square(x = squeeze_1_cast_fp16)[name = tensor("square_2_cast_fp16")]; - tensor add_1_cast_fp16 = add(x = square_1_cast_fp16, y = square_2_cast_fp16)[name = tensor("add_1_cast_fp16")]; - tensor magnitudes_cast_fp16 = identity(x = add_1_cast_fp16)[name = tensor("magnitudes_cast_fp16")]; - tensor mel_spec_1_transpose_x_0 = const()[name = tensor("mel_spec_1_transpose_x_0"), val = tensor(false)]; - tensor mel_spec_1_transpose_y_0 = const()[name = tensor("mel_spec_1_transpose_y_0"), val = tensor(false)]; - tensor mel_filters_to_fp16 = const()[name = tensor("mel_filters_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(526528)))]; - tensor mel_spec_1_cast_fp16 = matmul(transpose_x = mel_spec_1_transpose_x_0, transpose_y = mel_spec_1_transpose_y_0, x = mel_filters_to_fp16, y = magnitudes_cast_fp16)[name = tensor("mel_spec_1_cast_fp16")]; - tensor var_56_to_fp16 = const()[name = tensor("op_56_to_fp16"), val = tensor(0x1p-24)]; - tensor mel_spec_3_cast_fp16 = add(x = mel_spec_1_cast_fp16, y = var_56_to_fp16)[name = tensor("mel_spec_3_cast_fp16")]; - tensor mel_spec_5_epsilon_0 = const()[name = tensor("mel_spec_5_epsilon_0"), val = tensor(0x1p-149)]; - tensor mel_spec_5_cast_fp16 = log(epsilon = mel_spec_5_epsilon_0, x = mel_spec_3_cast_fp16)[name = tensor("mel_spec_5_cast_fp16")]; - tensor per_feature_mean_axes_0 = const()[name = tensor("per_feature_mean_axes_0"), val = tensor([-1])]; - tensor per_feature_mean_keep_dims_0 = const()[name = tensor("per_feature_mean_keep_dims_0"), val = tensor(true)]; - tensor per_feature_mean_cast_fp16 = reduce_mean(axes = per_feature_mean_axes_0, keep_dims = per_feature_mean_keep_dims_0, x = mel_spec_5_cast_fp16)[name = tensor("per_feature_mean_cast_fp16")]; - tensor sub_0_cast_fp16 = sub(x = mel_spec_5_cast_fp16, y = per_feature_mean_cast_fp16)[name = tensor("sub_0_cast_fp16")]; - tensor square_0_cast_fp16 = square(x = sub_0_cast_fp16)[name = tensor("square_0_cast_fp16")]; - tensor reduce_mean_1_axes_0 = const()[name = tensor("reduce_mean_1_axes_0"), val = tensor([-1])]; - tensor reduce_mean_1_keep_dims_0 = const()[name = tensor("reduce_mean_1_keep_dims_0"), val = tensor(true)]; - tensor reduce_mean_1_cast_fp16 = reduce_mean(axes = reduce_mean_1_axes_0, keep_dims = reduce_mean_1_keep_dims_0, x = square_0_cast_fp16)[name = tensor("reduce_mean_1_cast_fp16")]; - tensor real_div_0_to_fp16 = const()[name = tensor("real_div_0_to_fp16"), val = tensor(0x1.004p+0)]; - tensor mul_0_cast_fp16 = mul(x = reduce_mean_1_cast_fp16, y = real_div_0_to_fp16)[name = tensor("mul_0_cast_fp16")]; - tensor sqrt_0_cast_fp16 = sqrt(x = mul_0_cast_fp16)[name = tensor("sqrt_0_cast_fp16")]; - tensor var_70_to_fp16 = const()[name = tensor("op_70_to_fp16"), val = tensor(0x1.5p-17)]; - tensor per_feature_std_cast_fp16 = add(x = sqrt_0_cast_fp16, y = var_70_to_fp16)[name = tensor("per_feature_std_cast_fp16")]; - tensor mel_spec_cast_fp16 = real_div(x = sub_0_cast_fp16, y = per_feature_std_cast_fp16)[name = tensor("mel_spec_cast_fp16")]; - tensor var_75_perm_0 = const()[name = tensor("op_75_perm_0"), val = tensor([1, 0])]; - tensor var_77_axes_0 = const()[name = tensor("op_77_axes_0"), val = tensor([0])]; - tensor var_75_cast_fp16 = transpose(perm = var_75_perm_0, x = mel_spec_cast_fp16)[name = tensor("transpose_0")]; - tensor var_77_cast_fp16 = expand_dims(axes = var_77_axes_0, x = var_75_cast_fp16)[name = tensor("op_77_cast_fp16")]; - tensor var_79_axes_0 = const()[name = tensor("op_79_axes_0"), val = tensor([1])]; - tensor melspectrogram_features = expand_dims(axes = var_79_axes_0, x = var_77_cast_fp16)[name = tensor("op_79_cast_fp16")]; - } -> (melspectrogram_features); -} \ No newline at end of file diff --git a/parakeet-ctc-110m-coreml/MelSpectrogram.mlmodelc/weights/weight.bin b/parakeet-ctc-110m-coreml/MelSpectrogram.mlmodelc/weights/weight.bin deleted file mode 100644 index 09bc101712342c386fbfbcae32f80a96569376e3..0000000000000000000000000000000000000000 --- a/parakeet-ctc-110m-coreml/MelSpectrogram.mlmodelc/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:0a89c055bfde9022029d3cc59a23e949385e063974460d8eaec3a7614c3eaaa8 -size 567712 diff --git a/parakeet-ctc-110m-coreml/Preprocessor.mlmodelc/analytics/coremldata.bin b/parakeet-ctc-110m-coreml/Preprocessor.mlmodelc/analytics/coremldata.bin deleted file mode 100644 index d34457003aea6648633a14e7cd6134edc9c30f17..0000000000000000000000000000000000000000 --- a/parakeet-ctc-110m-coreml/Preprocessor.mlmodelc/analytics/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:a1ac15543fbb9301fba5f018b147e44d767479dec352aaa91dfe7bcf65949693 -size 243 diff --git a/parakeet-ctc-110m-coreml/Preprocessor.mlmodelc/coremldata.bin b/parakeet-ctc-110m-coreml/Preprocessor.mlmodelc/coremldata.bin deleted file mode 100644 index cd81f47cc6672a441cbe9556757a30ffc0ff0bc7..0000000000000000000000000000000000000000 --- a/parakeet-ctc-110m-coreml/Preprocessor.mlmodelc/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:4940877938cc1b6d8830bbdd68ac8a49377cc57d75b61308883da5235b6a1914 -size 439 diff --git a/parakeet-ctc-110m-coreml/Preprocessor.mlmodelc/metadata.json b/parakeet-ctc-110m-coreml/Preprocessor.mlmodelc/metadata.json deleted file mode 100644 index 6087d5e76f7dd060cf4f135f1f3ae0bbbeb2f30a..0000000000000000000000000000000000000000 --- a/parakeet-ctc-110m-coreml/Preprocessor.mlmodelc/metadata.json +++ /dev/null @@ -1,112 +0,0 @@ -[ - { - "metadataOutputVersion" : "3.0", - "shortDescription" : "Parakeet 110M preprocessor (15 s window)", - "outputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32)", - "shortDescription" : "", - "shape" : "[]", - "name" : "mel_features", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1)", - "shortDescription" : "", - "shape" : "[1]", - "name" : "mel_length", - "type" : "MultiArray" - } - ], - "storagePrecision" : "Float16", - "modelParameters" : [ - - ], - "author" : "Fluid Inference", - "specificationVersion" : 8, - "mlProgramOperationTypeHistogram" : { - "Range1d" : 3, - "Ios17.equal" : 1, - "Ios17.notEqual" : 1, - "Ios17.reshape" : 2, - "Identity" : 1, - "Ios17.matmul" : 1, - "Select" : 6, - "Ios17.expandDims" : 12, - "Ios17.add" : 3, - "Tile" : 2, - "Ios17.sliceByIndex" : 3, - "Ios16.reduceSum" : 4, - "Shape" : 4, - "Ios17.gather" : 4, - "Ios17.logicalNot" : 1, - "Pad" : 1, - "Ios17.log" : 1, - "Ios17.less" : 2, - "Ios17.sub" : 4, - "Ios17.conv" : 2, - "Ios17.pow" : 2, - "Ios17.cast" : 10, - "Ios17.concat" : 3, - "Stack" : 1, - "Ios17.floorDiv" : 1, - "Ios17.realDiv" : 4, - "Ios17.sqrt" : 1, - "Ios17.greaterEqual" : 1, - "Ios17.mul" : 1 - }, - "computePrecision" : "Mixed (Float16, Float32, Int16, Int32, UInt16)", - "isUpdatable" : "0", - "stateSchema" : [ - - ], - "availability" : { - "macOS" : "14.0", - "tvOS" : "17.0", - "visionOS" : "1.0", - "watchOS" : "10.0", - "iOS" : "17.0", - "macCatalyst" : "17.0" - }, - "modelType" : { - "name" : "MLModelType_mlProgram" - }, - "inputSchema" : [ - { - "dataType" : "Float32", - "hasShapeFlexibility" : "1", - "isOptional" : "0", - "shapeFlexibility" : "1 × 1...240000", - "shapeRange" : "[[1, 1], [1, 240000]]", - "formattedType" : "MultiArray (Float32 1 × 1)", - "type" : "MultiArray", - "shape" : "[1, 1]", - "name" : "audio", - "shortDescription" : "" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1)", - "shortDescription" : "", - "shape" : "[1]", - "name" : "audio_length", - "type" : "MultiArray" - } - ], - "userDefinedMetadata" : { - "com.github.apple.coremltools.source_dialect" : "TorchScript", - "com.github.apple.coremltools.source" : "torch==2.9.0", - "com.github.apple.coremltools.version" : "8.3.0" - }, - "generatedClassName" : "parakeet_preprocessor", - "method" : "predict" - } -] \ No newline at end of file diff --git a/parakeet-ctc-110m-coreml/Preprocessor.mlmodelc/model.mil b/parakeet-ctc-110m-coreml/Preprocessor.mlmodelc/model.mil deleted file mode 100644 index ae325b6071a8ee5dad74a2e516e60889d89771d8..0000000000000000000000000000000000000000 --- a/parakeet-ctc-110m-coreml/Preprocessor.mlmodelc/model.mil +++ /dev/null @@ -1,191 +0,0 @@ -program(1.0) -[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3510.2.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.9.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "8.3.0"}})] -{ - func main(tensor audio, tensor audio_length) [FlexibleShapeInformation = tuple, dict, tensor>>, tuple, dict, list, ?>>>>((("DefaultShapes", {{"audio", [1, 1]}}), ("RangeDims", {{"audio", [[1, 1], [1, 240000]]}})))] { - tensor var_9 = const()[name = tensor("op_9"), val = tensor(1)]; - tensor var_10 = const()[name = tensor("op_10"), val = tensor(160)]; - tensor var_12 = const()[name = tensor("op_12"), val = tensor(0)]; - tensor var_34 = const()[name = tensor("op_34"), val = tensor(512)]; - tensor var_35 = add(x = audio_length, y = var_34)[name = tensor("op_35")]; - tensor var_36 = const()[name = tensor("op_36"), val = tensor(512)]; - tensor var_37 = sub(x = var_35, y = var_36)[name = tensor("op_37")]; - tensor floor_div_0 = floor_div(x = var_37, y = var_10)[name = tensor("floor_div_0")]; - tensor var_40 = equal(x = audio_length, y = var_12)[name = tensor("op_40")]; - tensor var_41 = const()[name = tensor("op_41"), val = tensor([0])]; - tensor mel_length = select(a = var_41, b = floor_div_0, cond = var_40)[name = tensor("seq_len")]; - tensor audio_to_fp16_dtype_0 = const()[name = tensor("audio_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor audio_to_fp16 = cast(dtype = audio_to_fp16_dtype_0, x = audio)[name = tensor("cast_27")]; - tensor var_43_shape_cast_fp16 = shape(x = audio_to_fp16)[name = tensor("op_43_shape_cast_fp16")]; - tensor gather_0_axis_0 = const()[name = tensor("gather_0_axis_0"), val = tensor(0)]; - tensor gather_0_batch_dims_0 = const()[name = tensor("gather_0_batch_dims_0"), val = tensor(0)]; - tensor gather_0_validate_indices_0 = const()[name = tensor("gather_0_validate_indices_0"), val = tensor(false)]; - tensor var_43_shape_cast_fp16_to_int16_dtype_0 = const()[name = tensor("op_43_shape_cast_fp16_to_int16_dtype_0"), val = tensor("int16")]; - tensor select_0_to_uint16 = const()[name = tensor("select_0_to_uint16"), val = tensor(1)]; - tensor var_43_shape_cast_fp16_to_int16 = cast(dtype = var_43_shape_cast_fp16_to_int16_dtype_0, x = var_43_shape_cast_fp16)[name = tensor("cast_26")]; - tensor gather_0_cast_uint16 = gather(axis = gather_0_axis_0, batch_dims = gather_0_batch_dims_0, indices = select_0_to_uint16, validate_indices = gather_0_validate_indices_0, x = var_43_shape_cast_fp16_to_int16)[name = tensor("gather_0_cast_uint16")]; - tensor gather_0_cast_uint16_to_int32_dtype_0 = const()[name = tensor("gather_0_cast_uint16_to_int32_dtype_0"), val = tensor("int32")]; - tensor const_0 = const()[name = tensor("const_0"), val = tensor(0)]; - tensor const_1 = const()[name = tensor("const_1"), val = tensor(1)]; - tensor gather_0_cast_uint16_to_int32 = cast(dtype = gather_0_cast_uint16_to_int32_dtype_0, x = gather_0_cast_uint16)[name = tensor("cast_25")]; - tensor var_44 = range_1d(end = gather_0_cast_uint16_to_int32, start = const_0, step = const_1)[name = tensor("op_44")]; - tensor var_45_axes_0 = const()[name = tensor("op_45_axes_0"), val = tensor([0])]; - tensor var_45 = expand_dims(axes = var_45_axes_0, x = var_44)[name = tensor("op_45")]; - tensor var_46_axes_0 = const()[name = tensor("op_46_axes_0"), val = tensor([1])]; - tensor var_46 = expand_dims(axes = var_46_axes_0, x = audio_length)[name = tensor("op_46")]; - tensor timemask = less(x = var_45, y = var_46)[name = tensor("timemask")]; - tensor var_49_begin_0 = const()[name = tensor("op_49_begin_0"), val = tensor([0, 0])]; - tensor var_49_end_0 = const()[name = tensor("op_49_end_0"), val = tensor([1, 1])]; - tensor var_49_end_mask_0 = const()[name = tensor("op_49_end_mask_0"), val = tensor([true, false])]; - tensor var_49_squeeze_mask_0 = const()[name = tensor("op_49_squeeze_mask_0"), val = tensor([false, true])]; - tensor var_49_cast_fp16 = slice_by_index(begin = var_49_begin_0, end = var_49_end_0, end_mask = var_49_end_mask_0, squeeze_mask = var_49_squeeze_mask_0, x = audio_to_fp16)[name = tensor("op_49_cast_fp16")]; - tensor var_50_axes_0 = const()[name = tensor("op_50_axes_0"), val = tensor([1])]; - tensor var_50_cast_fp16 = expand_dims(axes = var_50_axes_0, x = var_49_cast_fp16)[name = tensor("op_50_cast_fp16")]; - tensor var_52_begin_0 = const()[name = tensor("op_52_begin_0"), val = tensor([0, 1])]; - tensor var_52_end_0 = const()[name = tensor("op_52_end_0"), val = tensor([1, 0])]; - tensor var_52_end_mask_0 = const()[name = tensor("op_52_end_mask_0"), val = tensor([true, true])]; - tensor var_52_cast_fp16 = slice_by_index(begin = var_52_begin_0, end = var_52_end_0, end_mask = var_52_end_mask_0, x = audio_to_fp16)[name = tensor("op_52_cast_fp16")]; - tensor var_54_begin_0 = const()[name = tensor("op_54_begin_0"), val = tensor([0, 0])]; - tensor var_54_end_0 = const()[name = tensor("op_54_end_0"), val = tensor([1, -1])]; - tensor var_54_end_mask_0 = const()[name = tensor("op_54_end_mask_0"), val = tensor([true, false])]; - tensor var_54_cast_fp16 = slice_by_index(begin = var_54_begin_0, end = var_54_end_0, end_mask = var_54_end_mask_0, x = audio_to_fp16)[name = tensor("op_54_cast_fp16")]; - tensor var_55_to_fp16 = const()[name = tensor("op_55_to_fp16"), val = tensor(0x1.f0cp-1)]; - tensor var_56_cast_fp16 = mul(x = var_54_cast_fp16, y = var_55_to_fp16)[name = tensor("op_56_cast_fp16")]; - tensor var_57_cast_fp16 = sub(x = var_52_cast_fp16, y = var_56_cast_fp16)[name = tensor("op_57_cast_fp16")]; - tensor x_3_interleave_0 = const()[name = tensor("x_3_interleave_0"), val = tensor(false)]; - tensor x_3_cast_fp16 = concat(axis = var_9, interleave = x_3_interleave_0, values = (var_50_cast_fp16, var_57_cast_fp16))[name = tensor("x_3_cast_fp16")]; - tensor var_60 = logical_not(x = timemask)[name = tensor("op_60")]; - tensor var_16_to_fp16 = const()[name = tensor("op_16_to_fp16"), val = tensor(0x0p+0)]; - tensor input_1_cast_fp16 = select(a = var_16_to_fp16, b = x_3_cast_fp16, cond = var_60)[name = tensor("input_1_cast_fp16")]; - tensor concat_1x = const()[name = tensor("concat_1x"), val = tensor([1, 1, -1])]; - tensor input_3_cast_fp16 = reshape(shape = concat_1x, x = input_1_cast_fp16)[name = tensor("input_3_cast_fp16")]; - tensor input_5_pad_0 = const()[name = tensor("input_5_pad_0"), val = tensor([0, 0, 0, 0, 256, 256])]; - tensor input_5_mode_0 = const()[name = tensor("input_5_mode_0"), val = tensor("constant")]; - tensor const_3_to_fp16 = const()[name = tensor("const_3_to_fp16"), val = tensor(0x0p+0)]; - tensor input_5_cast_fp16 = pad(constant_val = const_3_to_fp16, mode = input_5_mode_0, pad = input_5_pad_0, x = input_3_cast_fp16)[name = tensor("input_5_cast_fp16")]; - tensor concat_2x = const()[name = tensor("concat_2x"), val = tensor([1, -1])]; - tensor input_cast_fp16 = reshape(shape = concat_2x, x = input_5_cast_fp16)[name = tensor("input_cast_fp16")]; - tensor expand_dims_3 = const()[name = tensor("expand_dims_3"), val = tensor([160])]; - tensor expand_dims_4_axes_0 = const()[name = tensor("expand_dims_4_axes_0"), val = tensor([1])]; - tensor expand_dims_4_cast_fp16 = expand_dims(axes = expand_dims_4_axes_0, x = input_cast_fp16)[name = tensor("expand_dims_4_cast_fp16")]; - tensor conv_0_pad_type_0 = const()[name = tensor("conv_0_pad_type_0"), val = tensor("valid")]; - tensor conv_0_pad_0 = const()[name = tensor("conv_0_pad_0"), val = tensor([0, 0])]; - tensor conv_0_dilations_0 = const()[name = tensor("conv_0_dilations_0"), val = tensor([1])]; - tensor conv_0_groups_0 = const()[name = tensor("conv_0_groups_0"), val = tensor(1)]; - tensor expand_dims_1_to_fp16 = const()[name = tensor("expand_dims_1_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; - tensor conv_0_cast_fp16 = conv(dilations = conv_0_dilations_0, groups = conv_0_groups_0, pad = conv_0_pad_0, pad_type = conv_0_pad_type_0, strides = expand_dims_3, weight = expand_dims_1_to_fp16, x = expand_dims_4_cast_fp16)[name = tensor("conv_0_cast_fp16")]; - tensor conv_1_pad_type_0 = const()[name = tensor("conv_1_pad_type_0"), val = tensor("valid")]; - tensor conv_1_pad_0 = const()[name = tensor("conv_1_pad_0"), val = tensor([0, 0])]; - tensor conv_1_dilations_0 = const()[name = tensor("conv_1_dilations_0"), val = tensor([1])]; - tensor conv_1_groups_0 = const()[name = tensor("conv_1_groups_0"), val = tensor(1)]; - tensor expand_dims_2_to_fp16 = const()[name = tensor("expand_dims_2_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(263296)))]; - tensor conv_1_cast_fp16 = conv(dilations = conv_1_dilations_0, groups = conv_1_groups_0, pad = conv_1_pad_0, pad_type = conv_1_pad_type_0, strides = expand_dims_3, weight = expand_dims_2_to_fp16, x = expand_dims_4_cast_fp16)[name = tensor("conv_1_cast_fp16")]; - tensor stack_0_axis_0 = const()[name = tensor("stack_0_axis_0"), val = tensor(-1)]; - tensor stack_0_cast_fp16 = stack(axis = stack_0_axis_0, values = (conv_0_cast_fp16, conv_1_cast_fp16))[name = tensor("stack_0_cast_fp16")]; - tensor var_19_promoted_to_fp16 = const()[name = tensor("op_19_promoted_to_fp16"), val = tensor(0x1p+1)]; - tensor var_75_cast_fp16 = pow(x = stack_0_cast_fp16, y = var_19_promoted_to_fp16)[name = tensor("op_75_cast_fp16")]; - tensor var_77_axes_0 = const()[name = tensor("op_77_axes_0"), val = tensor([-1])]; - tensor var_77_keep_dims_0 = const()[name = tensor("op_77_keep_dims_0"), val = tensor(false)]; - tensor var_77_cast_fp16 = reduce_sum(axes = var_77_axes_0, keep_dims = var_77_keep_dims_0, x = var_75_cast_fp16)[name = tensor("op_77_cast_fp16")]; - tensor x_11_cast_fp16 = identity(x = var_77_cast_fp16)[name = tensor("x_11_cast_fp16")]; - tensor x_13_transpose_x_0 = const()[name = tensor("x_13_transpose_x_0"), val = tensor(false)]; - tensor x_13_transpose_y_0 = const()[name = tensor("x_13_transpose_y_0"), val = tensor(false)]; - tensor const_4_to_fp16 = const()[name = tensor("const_4_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(526528)))]; - tensor x_13_cast_fp16 = matmul(transpose_x = x_13_transpose_x_0, transpose_y = x_13_transpose_y_0, x = const_4_to_fp16, y = x_11_cast_fp16)[name = tensor("x_13_cast_fp16")]; - tensor var_84_to_fp16 = const()[name = tensor("op_84_to_fp16"), val = tensor(0x1p-24)]; - tensor var_85_cast_fp16 = add(x = x_13_cast_fp16, y = var_84_to_fp16)[name = tensor("op_85_cast_fp16")]; - tensor x_15_epsilon_0 = const()[name = tensor("x_15_epsilon_0"), val = tensor(0x1p-149)]; - tensor x_15_cast_fp16 = log(epsilon = x_15_epsilon_0, x = var_85_cast_fp16)[name = tensor("x_15_cast_fp16")]; - tensor var_87_shape_cast_fp16 = shape(x = x_15_cast_fp16)[name = tensor("op_87_shape_cast_fp16")]; - tensor gather_5 = const()[name = tensor("gather_5"), val = tensor(1)]; - tensor gather_6_axis_0 = const()[name = tensor("gather_6_axis_0"), val = tensor(0)]; - tensor gather_6_batch_dims_0 = const()[name = tensor("gather_6_batch_dims_0"), val = tensor(0)]; - tensor gather_6_validate_indices_0 = const()[name = tensor("gather_6_validate_indices_0"), val = tensor(false)]; - tensor var_87_shape_cast_fp16_to_uint16_dtype_0 = const()[name = tensor("op_87_shape_cast_fp16_to_uint16_dtype_0"), val = tensor("uint16")]; - tensor select_6_to_uint16 = const()[name = tensor("select_6_to_uint16"), val = tensor(2)]; - tensor var_87_shape_cast_fp16_to_uint16 = cast(dtype = var_87_shape_cast_fp16_to_uint16_dtype_0, x = var_87_shape_cast_fp16)[name = tensor("cast_24")]; - tensor gather_6_cast_uint16 = gather(axis = gather_6_axis_0, batch_dims = gather_6_batch_dims_0, indices = select_6_to_uint16, validate_indices = gather_6_validate_indices_0, x = var_87_shape_cast_fp16_to_uint16)[name = tensor("gather_6_cast_uint16")]; - tensor gather_6_cast_uint16_to_int32_dtype_0 = const()[name = tensor("gather_6_cast_uint16_to_int32_dtype_0"), val = tensor("int32")]; - tensor const_5 = const()[name = tensor("const_5"), val = tensor(0)]; - tensor const_6 = const()[name = tensor("const_6"), val = tensor(1)]; - tensor gather_6_cast_uint16_to_int32 = cast(dtype = gather_6_cast_uint16_to_int32_dtype_0, x = gather_6_cast_uint16)[name = tensor("cast_23")]; - tensor var_89 = range_1d(end = gather_6_cast_uint16_to_int32, start = const_5, step = const_6)[name = tensor("op_89")]; - tensor var_90_axes_0 = const()[name = tensor("op_90_axes_0"), val = tensor([0])]; - tensor var_90 = expand_dims(axes = var_90_axes_0, x = var_89)[name = tensor("op_90")]; - tensor concat_3_axis_0 = const()[name = tensor("concat_3_axis_0"), val = tensor(0)]; - tensor concat_3_interleave_0 = const()[name = tensor("concat_3_interleave_0"), val = tensor(false)]; - tensor concat_3 = concat(axis = concat_3_axis_0, interleave = concat_3_interleave_0, values = (gather_5, gather_6_cast_uint16_to_int32))[name = tensor("concat_3")]; - tensor shape_8 = shape(x = var_90)[name = tensor("shape_8")]; - tensor real_div_0 = real_div(x = concat_3, y = shape_8)[name = tensor("real_div_0")]; - tensor time_steps = tile(reps = real_div_0, x = var_90)[name = tensor("time_steps")]; - tensor var_93_axes_0 = const()[name = tensor("op_93_axes_0"), val = tensor([1])]; - tensor var_93 = expand_dims(axes = var_93_axes_0, x = mel_length)[name = tensor("op_93")]; - tensor valid_mask = less(x = time_steps, y = var_93)[name = tensor("valid_mask")]; - tensor var_95_axes_0 = const()[name = tensor("op_95_axes_0"), val = tensor([1])]; - tensor var_95 = expand_dims(axes = var_95_axes_0, x = valid_mask)[name = tensor("op_95")]; - tensor var_96_cast_fp16 = select(a = x_15_cast_fp16, b = var_16_to_fp16, cond = var_95)[name = tensor("op_96_cast_fp16")]; - tensor x_mean_numerator_axes_0 = const()[name = tensor("x_mean_numerator_axes_0"), val = tensor([2])]; - tensor x_mean_numerator_keep_dims_0 = const()[name = tensor("x_mean_numerator_keep_dims_0"), val = tensor(false)]; - tensor x_mean_numerator_cast_fp16 = reduce_sum(axes = x_mean_numerator_axes_0, keep_dims = x_mean_numerator_keep_dims_0, x = var_96_cast_fp16)[name = tensor("x_mean_numerator_cast_fp16")]; - tensor x_mean_denominator_axes_0 = const()[name = tensor("x_mean_denominator_axes_0"), val = tensor([1])]; - tensor x_mean_denominator_keep_dims_0 = const()[name = tensor("x_mean_denominator_keep_dims_0"), val = tensor(false)]; - tensor cast_6_to_fp16_dtype_0 = const()[name = tensor("cast_6_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor valid_mask_to_fp16 = cast(dtype = cast_6_to_fp16_dtype_0, x = valid_mask)[name = tensor("cast_22")]; - tensor x_mean_denominator_cast_fp16 = reduce_sum(axes = x_mean_denominator_axes_0, keep_dims = x_mean_denominator_keep_dims_0, x = valid_mask_to_fp16)[name = tensor("x_mean_denominator_cast_fp16")]; - tensor var_101_axes_0 = const()[name = tensor("op_101_axes_0"), val = tensor([1])]; - tensor var_101_cast_fp16 = expand_dims(axes = var_101_axes_0, x = x_mean_denominator_cast_fp16)[name = tensor("op_101_cast_fp16")]; - tensor x_mean_cast_fp16 = real_div(x = x_mean_numerator_cast_fp16, y = var_101_cast_fp16)[name = tensor("x_mean_cast_fp16")]; - tensor var_104_axes_0 = const()[name = tensor("op_104_axes_0"), val = tensor([2])]; - tensor var_104_cast_fp16 = expand_dims(axes = var_104_axes_0, x = x_mean_cast_fp16)[name = tensor("op_104_cast_fp16")]; - tensor var_105_cast_fp16 = sub(x = x_15_cast_fp16, y = var_104_cast_fp16)[name = tensor("op_105_cast_fp16")]; - tensor var_106_cast_fp16 = select(a = var_105_cast_fp16, b = var_16_to_fp16, cond = var_95)[name = tensor("op_106_cast_fp16")]; - tensor var_19_promoted_1_to_fp16 = const()[name = tensor("op_19_promoted_1_to_fp16"), val = tensor(0x1p+1)]; - tensor var_107_cast_fp16 = pow(x = var_106_cast_fp16, y = var_19_promoted_1_to_fp16)[name = tensor("op_107_cast_fp16")]; - tensor var_109_axes_0 = const()[name = tensor("op_109_axes_0"), val = tensor([2])]; - tensor var_109_keep_dims_0 = const()[name = tensor("op_109_keep_dims_0"), val = tensor(false)]; - tensor var_109_cast_fp16 = reduce_sum(axes = var_109_axes_0, keep_dims = var_109_keep_dims_0, x = var_107_cast_fp16)[name = tensor("op_109_cast_fp16")]; - tensor var_111_to_fp16 = const()[name = tensor("op_111_to_fp16"), val = tensor(0x1p+0)]; - tensor var_112_cast_fp16 = sub(x = var_101_cast_fp16, y = var_111_to_fp16)[name = tensor("op_112_cast_fp16")]; - tensor var_113_cast_fp16 = real_div(x = var_109_cast_fp16, y = var_112_cast_fp16)[name = tensor("op_113_cast_fp16")]; - tensor x_std_1_cast_fp16 = sqrt(x = var_113_cast_fp16)[name = tensor("x_std_1_cast_fp16")]; - tensor var_115_cast_fp16 = not_equal(x = x_std_1_cast_fp16, y = x_std_1_cast_fp16)[name = tensor("op_115_cast_fp16")]; - tensor x_std_3_cast_fp16 = select(a = var_16_to_fp16, b = x_std_1_cast_fp16, cond = var_115_cast_fp16)[name = tensor("x_std_3_cast_fp16")]; - tensor var_25_to_fp16 = const()[name = tensor("op_25_to_fp16"), val = tensor(0x1.5p-17)]; - tensor x_std_cast_fp16 = add(x = x_std_3_cast_fp16, y = var_25_to_fp16)[name = tensor("x_std_cast_fp16")]; - tensor var_120_axes_0 = const()[name = tensor("op_120_axes_0"), val = tensor([2])]; - tensor var_120_cast_fp16 = expand_dims(axes = var_120_axes_0, x = x_std_cast_fp16)[name = tensor("op_120_cast_fp16")]; - tensor x_cast_fp16 = real_div(x = var_105_cast_fp16, y = var_120_cast_fp16)[name = tensor("x_cast_fp16")]; - tensor var_122_shape_cast_fp16 = shape(x = x_cast_fp16)[name = tensor("op_122_shape_cast_fp16")]; - tensor gather_7_axis_0 = const()[name = tensor("gather_7_axis_0"), val = tensor(0)]; - tensor gather_7_batch_dims_0 = const()[name = tensor("gather_7_batch_dims_0"), val = tensor(0)]; - tensor gather_7_validate_indices_0 = const()[name = tensor("gather_7_validate_indices_0"), val = tensor(false)]; - tensor var_122_shape_cast_fp16_to_uint16_dtype_0 = const()[name = tensor("op_122_shape_cast_fp16_to_uint16_dtype_0"), val = tensor("uint16")]; - tensor select_7_to_uint16 = const()[name = tensor("select_7_to_uint16"), val = tensor(2)]; - tensor var_122_shape_cast_fp16_to_uint16 = cast(dtype = var_122_shape_cast_fp16_to_uint16_dtype_0, x = var_122_shape_cast_fp16)[name = tensor("cast_21")]; - tensor gather_7_cast_uint16 = gather(axis = gather_7_axis_0, batch_dims = gather_7_batch_dims_0, indices = select_7_to_uint16, validate_indices = gather_7_validate_indices_0, x = var_122_shape_cast_fp16_to_uint16)[name = tensor("gather_7_cast_uint16")]; - tensor gather_7_cast_uint16_to_int32_dtype_0 = const()[name = tensor("gather_7_cast_uint16_to_int32_dtype_0"), val = tensor("int32")]; - tensor const_7 = const()[name = tensor("const_7"), val = tensor(0)]; - tensor const_8 = const()[name = tensor("const_8"), val = tensor(1)]; - tensor gather_7_cast_uint16_to_int32 = cast(dtype = gather_7_cast_uint16_to_int32_dtype_0, x = gather_7_cast_uint16)[name = tensor("cast_20")]; - tensor mask_1 = range_1d(end = gather_7_cast_uint16_to_int32, start = const_7, step = const_8)[name = tensor("mask_1")]; - tensor gather_8_axis_0 = const()[name = tensor("gather_8_axis_0"), val = tensor(0)]; - tensor gather_8_batch_dims_0 = const()[name = tensor("gather_8_batch_dims_0"), val = tensor(0)]; - tensor gather_8_validate_indices_0 = const()[name = tensor("gather_8_validate_indices_0"), val = tensor(false)]; - tensor select_8_to_uint16 = const()[name = tensor("select_8_to_uint16"), val = tensor(0)]; - tensor gather_8_cast_uint16 = gather(axis = gather_8_axis_0, batch_dims = gather_8_batch_dims_0, indices = select_8_to_uint16, validate_indices = gather_8_validate_indices_0, x = var_122_shape_cast_fp16_to_uint16)[name = tensor("gather_8_cast_uint16")]; - tensor gather_8_cast_uint16_to_int32_dtype_0 = const()[name = tensor("gather_8_cast_uint16_to_int32_dtype_0"), val = tensor("int32")]; - tensor concat_4_axis_0 = const()[name = tensor("concat_4_axis_0"), val = tensor(0)]; - tensor concat_4_interleave_0 = const()[name = tensor("concat_4_interleave_0"), val = tensor(false)]; - tensor gather_8_cast_uint16_to_int32 = cast(dtype = gather_8_cast_uint16_to_int32_dtype_0, x = gather_8_cast_uint16)[name = tensor("cast_19")]; - tensor concat_4 = concat(axis = concat_4_axis_0, interleave = concat_4_interleave_0, values = (gather_8_cast_uint16_to_int32, var_9))[name = tensor("concat_4")]; - tensor expand_dims_0_axes_0 = const()[name = tensor("expand_dims_0_axes_0"), val = tensor([0])]; - tensor expand_dims_0 = expand_dims(axes = expand_dims_0_axes_0, x = mask_1)[name = tensor("expand_dims_0")]; - tensor var_126 = tile(reps = concat_4, x = expand_dims_0)[name = tensor("op_126")]; - tensor mask = greater_equal(x = var_126, y = var_93)[name = tensor("mask")]; - tensor var_129_axes_0 = const()[name = tensor("op_129_axes_0"), val = tensor([1])]; - tensor var_129 = expand_dims(axes = var_129_axes_0, x = mask)[name = tensor("op_129")]; - tensor cast_15_to_fp16 = const()[name = tensor("cast_15_to_fp16"), val = tensor(0x0p+0)]; - tensor processed_signal_cast_fp16 = select(a = cast_15_to_fp16, b = x_cast_fp16, cond = var_129)[name = tensor("processed_signal_cast_fp16")]; - tensor processed_signal_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("processed_signal_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; - tensor mel_features = cast(dtype = processed_signal_cast_fp16_to_fp32_dtype_0, x = processed_signal_cast_fp16)[name = tensor("cast_18")]; - } -> (mel_features, mel_length); -} \ No newline at end of file diff --git a/parakeet-ctc-110m-coreml/Preprocessor.mlmodelc/weights/weight.bin b/parakeet-ctc-110m-coreml/Preprocessor.mlmodelc/weights/weight.bin deleted file mode 100644 index d0c6ac87e5f78315662149af9ab4ea3658497dbe..0000000000000000000000000000000000000000 --- a/parakeet-ctc-110m-coreml/Preprocessor.mlmodelc/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:c062338de852a26607ce4101f74e6895de3a4134a57b07232bd72bfc6f1d7f1a -size 567712 diff --git a/parakeet-ctc-110m-coreml/config.json b/parakeet-ctc-110m-coreml/config.json deleted file mode 100644 index aaa7174ca014092519980c51451ac7ca3832a352..0000000000000000000000000000000000000000 --- a/parakeet-ctc-110m-coreml/config.json +++ /dev/null @@ -1,7 +0,0 @@ -{ - "bos_token_id": 1, - "eos_token_id": 2, - "nemo_model_type": "parakeet", - "pad_token_id": 0, - "vocab_size": 1024, -} diff --git a/parakeet-ctc-110m-coreml/metadata.json b/parakeet-ctc-110m-coreml/metadata.json deleted file mode 100644 index 6db872ae8a7d463c95dc180210ace0fce019f4a6..0000000000000000000000000000000000000000 --- a/parakeet-ctc-110m-coreml/metadata.json +++ /dev/null @@ -1,247 +0,0 @@ -{ - "model_id": "nvidia/parakeet-tdt_ctc-110m", - "model_type": "hybrid_rnnt_ctc", - "sample_rate": 16000, - "max_audio_seconds": 15.0, - "max_audio_samples": 240000, - "max_symbol_steps": 1, - "vocab_size": 1024, - "joint_extra_outputs": 5, - "encoder_dim": 512, - "decoder_dim": 640, - "decoder_hidden": 640, - "decoder_layers": 1, - "blank_id": 1024, - "checkpoint": { - "type": "pretrained", - "model_id": "nvidia/parakeet-tdt_ctc-110m" - }, - "coreml": { - "compute_units": "CPU_ONLY", - "compute_precision": "FLOAT32" - }, - "components": { - "preprocessor": { - "inputs": { - "audio_signal": [ - 1, - 240000 - ], - "audio_length": [ - 1 - ] - }, - "outputs": { - "mel": [ - 1, - 80, - 1501 - ], - "mel_length": [ - 1 - ] - }, - "path": "parakeet_preprocessor.mlpackage" - }, - "encoder": { - "inputs": { - "mel": [ - 1, - 80, - 1501 - ], - "mel_length": [ - 1 - ] - }, - "outputs": { - "encoder": [ - 1, - 512, - 188 - ], - "encoder_length": [ - 1 - ] - }, - "path": "parakeet_encoder.mlpackage" - }, - "ctc_head": { - "inputs": { - "encoder": [ - 1, - 512, - 188 - ] - }, - "outputs": { - "log_probs": [ - 1, - 188, - 1025 - ] - }, - "path": "parakeet_ctc_head.mlpackage" - }, - "mel_encoder": { - "inputs": { - "audio_signal": [ - 1, - 240000 - ], - "audio_length": [ - 1 - ] - }, - "outputs": { - "encoder": [ - 1, - 512, - 188 - ], - "encoder_length": [ - 1 - ] - }, - "path": "parakeet_mel_encoder.mlpackage" - }, - "decoder": { - "inputs": { - "targets": [ - 1, - 1 - ], - "target_length": [ - 1 - ], - "h_in": [ - 1, - 1, - 640 - ], - "c_in": [ - 1, - 1, - 640 - ] - }, - "outputs": { - "decoder": [ - 1, - 640, - 1 - ], - "h_out": [ - 1, - 1, - 640 - ], - "c_out": [ - 1, - 1, - 640 - ] - }, - "path": "parakeet_decoder.mlpackage" - }, - "joint": { - "inputs": { - "encoder": [ - 1, - 512, - 188 - ], - "decoder": [ - 1, - 640, - 1 - ] - }, - "outputs": { - "logits": [ - 1, - 188, - 1, - 1030 - ] - }, - "path": "parakeet_joint.mlpackage" - }, - "joint_decision": { - "inputs": { - "encoder": [ - 1, - 512, - 188 - ], - "decoder": [ - 1, - 640, - 1 - ] - }, - "outputs": { - "token_id": [ - 1, - 188, - 1 - ], - "token_prob": [ - 1, - 188, - 1 - ], - "duration": [ - 1, - 188, - 1 - ] - }, - "path": "parakeet_joint_decision.mlpackage" - }, - "joint_decision_single_step": { - "inputs": { - "encoder_step": [ - 1, - 512, - 1 - ], - "decoder_step": [ - 1, - 640, - 1 - ] - }, - "outputs": { - "token_id": [ - 1, - 1, - 1 - ], - "token_prob": [ - 1, - 1, - 1 - ], - "duration": [ - 1, - 1, - 1 - ], - "top_k_ids": [ - 1, - 1, - 1, - 64 - ], - "top_k_logits": [ - 1, - 1, - 1, - 64 - ] - }, - "path": "parakeet_joint_decision_single_step.mlpackage" - } - } -} \ No newline at end of file diff --git a/parakeet-ctc-110m-coreml/preprocessor_config.json b/parakeet-ctc-110m-coreml/preprocessor_config.json deleted file mode 100644 index 54683c2afb1e841dfe59d03ad92c060a6cf50c0d..0000000000000000000000000000000000000000 --- a/parakeet-ctc-110m-coreml/preprocessor_config.json +++ /dev/null @@ -1,13 +0,0 @@ -{ - "feature_extractor_type": "ParakeetFeatureExtractor", - "feature_size": 80, - "hop_length": 160, - "n_fft": 512, - "padding_side": "right", - "padding_value": 0.0, - "preemphasis": 0.97, - "processor_class": "ParakeetProcessor", - "return_attention_mask": true, - "sampling_rate": 16000, - "win_length": 400 -} diff --git a/parakeet-ctc-110m-coreml/special_tokens_map.json b/parakeet-ctc-110m-coreml/special_tokens_map.json deleted file mode 100644 index 815fbdc56abc60a0f29e51ff2373c998d7779805..0000000000000000000000000000000000000000 --- a/parakeet-ctc-110m-coreml/special_tokens_map.json +++ /dev/null @@ -1,16 +0,0 @@ -{ - "pad_token": { - "content": "", - "lstrip": false, - "normalized": false, - "rstrip": false, - "single_word": false - }, - "unk_token": { - "content": "", - "lstrip": false, - "normalized": false, - "rstrip": false, - "single_word": false - } -} diff --git a/parakeet-ctc-110m-coreml/tokenizer.json b/parakeet-ctc-110m-coreml/tokenizer.json deleted file mode 100644 index 21e87f185d71820f5089456aa9c79c97c0a0ce8b..0000000000000000000000000000000000000000 --- a/parakeet-ctc-110m-coreml/tokenizer.json +++ /dev/null @@ -1,2456 +0,0 @@ -{ - "version": "1.0", - "truncation": null, - "padding": null, - "added_tokens": [ - { - "id": 0, - "content": "", - "single_word": false, - "lstrip": false, - "rstrip": false, - "normalized": false, - "special": true - }, - { - "id": 1024, - "content": "", - "single_word": false, - "lstrip": false, - "rstrip": false, - "normalized": false, - "special": true - } - ], - "normalizer": { - "type": "Sequence", - "normalizers": [ - { - "type": "Precompiled", - "precompiled_charsmap": "ALQCAACEAAAAAACAAQAAgMz8AgC4BQAAjSIAgMzkAgC4PQAAgSIAgMzsAgC4BQAAkSIAgMw8AADNvAAAngkAgKEJAICkCQCAgx0AAIAZAACBGQAAQx0AgDsdAIBTHQCASx0AgIAxAACBMQAApwkAgIkxAAA9WAMAPEgDAEMKAIA+aAMAAYUAAIQBAQADjQAAAokAAAWVAAAEkQAAB50AAAaZAAAJqQAACKEAAAutAAAKpQAADbkAAAy9AAAPvQAADrkAABHFAAAQwQAAE80AABLJAAAV1QAAFNEAABfdAAAW2QAAGeUAABjhAAAb7QAAGukAAB31AAAc8QAAH/0AAB75AABhOAkAax0AgGNADgBi8AgAZSgPAGSADgBn2A8AZvAPAGlwDABoMAwAa/AMAGrYDABtSA0AbBwNAG8QEgBubA0ASQoAgHAMEwBzqBMAcuwTAHUoEAB0TBAAd9ARAHYUEAB50BYAePQQAGMdAIB69BYAex0AgHMdAIB/fQEAiQwAgEGAAgDhCwCAQxgAAELAAABFSAAARGAAAEeQBgBGhAEASSgGAEhsAQBLOAcASvAHAE1wBwBMRAcAT/AEAE7MBACqCQCAUCwFAFOgCgBSEAUAVQAKAFRQCgBX0AgAVhALAFlICABYuAgAhBEAAFo8CACA9QAAgZ0AANsLAIAzHQCAg2kCAIJFAgCBNQIAgDUCAIdtAwCGVQMAgTkAAIRlAgAaDACAigEEAInVAwCI7QMAjwkAAKsLAIAsDACAjAkAADIMAICJMQMAkQkAAMzYAABbHQCAgx0AgMMaAIBPCgCAgGUDAIENAwCGPQAAgx0DAMwQAgDNhAEAgikAAMx0AwCjgQYAyxoAgICxAgCBsQIA0xoAgIEpAAClwQAA2xoAgMzoAwDNYAIAVQoAgKjxAABbCgCAYQoAgGcKAIDjGgCAgWkAAMzcBACCEQEA6xoAgG0KAIDzGgCAAxsAgAsbAID7GgCAtgkAgMygBADN3AQAzAgBALkJAICrHQCAhhEBAOEAKwDgfCcA44hIAuIMOAKjHQCAh5EBALsdAICzHQCAgNkBAIE1AADMxAIA6kRkApsdAIATGwCA7/hnAoERBwCC8QEA8BiGAolVAACB5QEAGxsAgIfhAQCAbQAAgQ0AAIN5AAB5CgCAgXkAAICVAQDMOAEAzRQBAIzBAQB/CgCAvwkAgKMVAQDDlBcAwpwUAMWEFwDEUBcAx+wXAMaAEgCTHQCAiwoAgMvQFgDK4BYAzRQWADgMAIDPvCAAzpwZANHMJADQ2CUA0+gkALFRAQA+DACAp90HAMMdAIDWvCQA2cgnANjUIgDb+CcAMxsAgIftBwCFCgCAzPgEACMbAIArHQCAh8kGALMJAICR3QcAvAkAgCsbAIBzCgCAOxsAgIsdAICPDACAjPkGAA4MAICA1QYAgcEGAMzEAgDNBAUAglEAAIN1BwCArQYAgbkGAIY1BwCHKQcAhEEAAJEKAICn7QAAQxsAgIjpBwCJzQcAlwoAgI/BBwCM3QcAnQoAgO0LAICnXQYAsJ0AAKMKAICpCgCAo0EGAEsbAIBbGwCAgAwAgFMbAIBjGwCArXEGAGsbAIDCCQCAzPgDAM0sAwDFCQCAo+UAAMgJAICMTQAAtQoAgKfxAAC7CgCAsT0GAIedAACGlQAAqB0HAISJAADBCgCAgqkAAIHVAACtAQcAzQoAgJE9AACCmQEAywkAgM0MBQDMCAUAgT0AAIeFAQCIvQEAexsAgMsdAICxCwCAjJEBAEQMAIBKDACA0x0AgID1AQCBhQEAgoEBAIOdAQCEiQEAxwoAgIapAQCHXQAAiG0AAIlNAABzGwCAzBACAIxdAACCDQAA0woAgI9JAACw6QAAgxsAgPMLAICjKQEAgCUBAIFVAQCLGwCApzUBAMykAQDNEAIA2QoAgJMbAICBNQAA3woAgK4JAQDrCgCAzOgBAM0oAgCbGwCAo/EAAIQFAACjGwCA5QoAgLMbAICotQAAqxsAgIFdAAC7GwCAzPwBAM3AAQDDGwCAyxsAgIGFAwAUDACAgeUDAPEKAICH6QMAzgkAgIylAwDTGwCA/QoAgK0JAIDbGwCAgZkDAIHdAwCMvQMAzSQBAMwgAQDMEAIAzTACAIH5AACHUQAAgFUAAIFZAAD3CgCAg0kAAIxBAADrGwCA4xsAgNEJAICBfQAAgHEAAMwgAwDNsAMAo30DANQJAICjEQMA8x0AgIEtAQCx/QAApzEDAK1BAwDrHQCAo20DAAMeAID7HQCA8xsAgKdtAwCANQAAgR0AALFtAwCILQAAmwwAgKeVAACBcQAAgFkAAINxAACj9QAAgVEAAK2BAAD7GwCAsQkDAIldAACEPQAAzDgBAISdAQCBGQAAgAkAAIRlAAADHACAzNAHAMzwBwALHACAkYkAAMxMBgDNBAYAzHAGAM10BgDMQAcAmy0PAMyoBwDNrAcAhg0AAIdVDwCEQQ8ADAsAgIIBDACDVQ8AgDUBAIHZAQCnDACAj+kAAIztAACVDACA4x0AgIv1AACIbQ8AiQ0AABILAIC3CwCAgiUAAFAMAICBQQAAVgwAgBseAIATHgCAKx4AgCMeAIAzHgCACx4AgIApAACBKQAA/wsAgBMcAICEeQAAGxwAgIFNAQCAoQEAGwsAgKP9DwDMOAIAzUgDACMcAICBWQAAzXwCAMykDQAnCwCAXAwAgKjJDwCHOQAA2gkAgImhDwAGCwCAkREAAKEMAIDdCQCAnAsAgGIMAICAuQ8AgbkPANsdAICDjQ8A+QsAgCscAICEBQAAMxwAgCELAIA7HACALQsAgIGdDwCHIQAAh7UPAMyoAgDN6AIAzLQMAM3cDACmzQAAp8UAAFMcAICPgQ8AjIkPAKPlAAAzCwCAQxwAgD8LAICxyQAAhwUAAFscAIBLHACAhz0AAGMcAIB0DACAOQsAgKMFDwCB+QAAzKgDAGscAIBLCwCAjEkAAKPxAABzHACAegwAgEULAICnlQAAgxwAgHscAIDMrAMAzcgAAOAJAICHaQAA4wkAgIG9AACCeQAA5gkAgIe5AQBRCwCAkaUAAIEdAACjHACAVwsAgIgFAACrHACAm5EAAF0LAIDpCQCAjJEBANULAIDJCwCAwwsAgM8LAICDRQAAgrkBAIG5AQCApQEAQx4AgIZxAABjCwCAhEkAAIsVAACKPQAAiTkAAIhFAACP+QAAaQsAgL0LAICMBQAAp1EBAKZJAQBoDACAsHkAAKNZAQCMqQAAgKkAAIGpAACBlQAAgJUAAK1xAQBuDACApQsAgISNAABTHgCASx4AgKMhAABjHgCAWx4AgGseAICBbQAAgG0AALEFAQCkOQAAOx4AgIscAIBvCwCAqAUAAJscAICTHACArQkAAMywAQCBvQMAgL0DAIPNAwCzHACAuxwAgMMcAIDMvAEAzYQBAInpAwDMHAEAgdkCAIDFAgDNOAEAzDwBAMxoAgDNRAIAg00AAMscAICH2QAAhy0AAIBFAACBEQAAggUAAHULAIDbHACA0xwAgOMcAIDMOAIAiBUAAIjhAACAbQAAgTkAAMyEAgDNUAEAo0UDAIQ5AQDrHACA8xwAgMzcAwDNSAIAcx4AgOwJAIB7CwCAix4AgK0MAICBbQAA+xwAgIELAICj0QAAgx4AgHseAIDMiAQAgXUAAIB1AACECwCAo7UAAMwABADNVAIAAx0AgIoLAICETQEAkAsAgAsdAIATHQCAzNAOAMwsAQDMAAUAzVwFAO8JAIDyCQCAzJgOAIHBAADMzA8AzDwOAMwIAQDNnA4AzNQPAM14DwDMPA4AzTgOAIHlAQCA5QEAg+UBAILlAQDXCQCAhOUBAIfhAQBHHQCAiaUBAIjZAQCByQcAPx0AgFcdAIBPHQCAzDQBAPgJAICA3QAAgekAAEYKAICD/QAAgM0AAIH5AACBEQcAbx0AgGcdAICJ0QAAzCgBAH8dAIB3HQCA5AsAgMw0AQDeCwCAgF0AAIFlAACjAQEAg2EAAIFxAACASQAANx0AgB0MAICuCwCAiVUAAC8MAIA1DACAXx0AgIcdAIDHGgCAUgoAgIIdAACDeQcAgBkHAIEZBwCGIQAAhykAAISRBwD1CQCAimkAALHZBgCIaQAAifUHAEwKAICP3QcAjNkHAIwMAID7CQCALx0AgP4JAICRoQcAgEEHAIFBBwCHBQAAzxoAgIKRBwDXGgCA3xoAgKOVBgCGhQcAp+0AAMyQAgDN4AUAsekAAKPBAABYCgCAXgoAgGQKAIBqCgCAAQoAgKVlBwDnGgCAzLgDAKhVBwDvGgCAcAoAgPcaAIAHGwCADxsAgP8aAIAECgCAo60AAAcKAICMJQYACgoAgIxNAACvHQCAgm0AAIE9BgCCAQYAgWUAAKcdAICHZQAAvx0AgIcRBgCHrQEAtx0AgMxQAgDNxAIAgeEBAIDJAQCD4QEAkYkAAID9AQCB1QEAnx0AgIydAQCJNQAAdgoAgIB1AACBXQAAhi0AAIc1AACEfQAAFxsAgIKFAQCDfQAAgJ0BAIGRAQAfGwCAj+kAAIzhAAB8CgCAggoAgA0KAICIDQAAifkAAKc5AQCXHQCAjgoAgDsMAICjJQEAQQwAgLBZAACPHQCAggUAAMcdAICtFQEAkgwAgDcbAICGBQAAiAoAgCcbAIAvGwCAp2kAAIANAQCBAQEAhzEAAKNJAACxGQEAzBACAD8bAIARDACAlAoAgK1RAADM1AEAzfgBAKhBAABHGwCAzTgBAMw8AQCB7QMAmgoAgKAKAICMDQAA8AsAgKYKAICBxQMAzGgCAKwKAICCxQMATxsAgITJAwCHKQAAhjEAAF8bAICCbQAAgwwAgFcbAICHYQAAZxsAgG8bAIAbHQCAzKgDAM2sAgCB+QAAiC0AABAKAIATCgCAFgoAgIw1AAC4CgCAvgoAgLHVAADECgCAfxsAgM8dAIC0CwCAzDABAEcMAIBNDACA1x0AgMwEAQDKCgCAdxsAgKelAADWCgCAo40AAMwUAgCAuQAAgbkAAKeFAAALDACAgmUAAIcbAICMNQAA9gsAgMzsHADN/AMAjxsAgK6tAADcCgCAlxsAgMzABgDN0AYAsL0BAMyQBwDiCgCAgckBAMwYHQDNIAIAhBEAAO4KAIDNuAYAzKwGAKcbAIDoCgCAgSkAALcbAICvGwCAo+0BAMxAHQDNEAIAvxsAgMcbAICBCQAAzxsAgMxAHQDN0AIAqNkBABcMAIDMkAcAzBwBAMxgBgDNZAYA9AoAgB8KAIDXGwCAkSkBAAALAICBzR8A3xsAgPoKAIDvGwCA5xsAgMzEBgDNwAYAgTEAAIDZAAAiCgCAJQoAgIK5AQCDRQEAgLkBAIG5AQCGXQEA9x0AgIRdAQDvHQCAzcAAAMzwAACIARwAiXkBAAceAICPVQEAjGEBAP8dAICB3R4AgRUfAJ8bAICBXR8AjIEfAIdBHwDMGAMAzWgDAIBNHwCBpR8AKAoAgIOpHwCMFR8AjNEeACsKAICHtR8AgJUfAIGZHwCBEQAAg70fAICFHwCBiR8A9xsAgIQ9AACeDACAiZkfAP8bAICIBQAACQsAgAccAICADQAAgf0AAA8cAICj2R8Ao3keAKOFAAAPCwCArTUfAKdhHgCnqR8ApAwAgIQNAACqDACAozUfAC4KAICtiR8AhHEAAKchHwCxPR4AsYUfAJgMAIDnHQCAFQsAgLoLAIDMtBwAzbAcAFMMAICxQR8AWQwAgJ8LAIAfHgCAFx4AgC8eAIAnHgCAgLkeAIG5HgCCIQEAgzUBAIRhAQA3HgCAhokBAIe9AQCIkQEAiekBAN8dAICL/QEAjOUBAIINAAAPHgCAj90BAIO5AQCRrQEAgb0BAIC9AQCAoQEAgaEBAPwLAIACDACAhD0AABccAICJlQEAm4EBAIHNHgCAzR4AzPwCAM3wAgCB5QAAHxwAgIHtAACjpQAAzJABAM1cAgCHHQAAHgsAgKj5AAAnHACAKgsAgF8MAIBlDACALxwAgIQFAAA3HACAo9UAACQLAIA/HACAgVEAAMz0AQDN0AEAMAsAgIc9AABXHACANgsAgEccAIBCCwCAhwUAAF8cAIBPHACAh/EDAIHZAwCBmQMAgZEAAGccAIB3DACAjPkDAMwkAQCHuQMAgfkDADwLAIDMZAIAgskDAIyZAwBvHACAh9EDAI+RAwCB3QYAkfUDAMwABADN7AMAh2UAAB8dAIBOCwCAdxwAgH0MAIBICwCAzBgBAIg5AACHHACAfxwAgMxcAwCMJQAAMQoAgMwsAQCx/QAAozkDADQKAIA3CgCApxwAgKdZAwDMdAMAiAkAAKNRAwCvHACAYAsAgINtDQCnnQAApq0AAKOdAACxDQMAzCgBANgLAICntQAAprUAAMwLAIDMMAEAgdUHAMYLAIDMKAEA0gsAgEceAIBmCwCArYkAAGwLAICAzQEAgd0BAMxEAQDNnB4AhPUBAMALAIDMWAEAzUwBAIDtAQCB/QEAg7UAAGsMAICM3QEAcQwAgMwIHgCM8QYAzDgBAM08AQBXHgCAiREAAIEFBgBPHgCAZx4AgF8eAIBvHgCAgz0AAIAhAACBOQAAgDkAAIEhAAA/HgCAjxwAgMwoAQCB2QYAcgsAgIH9BgDMJAEAnxwAgJccAIC3HACAgCEBAIE1AQCjBQAAvxwAgMccAIDPHACAzIwFAM1AAgC3HAMAeAsAgIfNBwDfHACA1xwAgCMdAIDNiAAAzJAAAIzdBQCjhQAAGQoAgMzgAgDnHACAiNUHAIFNAACATQAAVAsAgO8cAIBaCwCAkTkHADoKAICIxQcAqAsAgIrJBwD3HACAmz0AAIflBwB3HgCAgYUHAICFBwA9CgCAgvkHAILVBgCDRQAAgMkGAIHdBgCG4QYAfgsAgIRRAACPHgCAipUGAIuZBgCIeQAAiZ0GALAMAICPWQcAjG0HAP8cAIDMgAMAzSQCALARBwBACgCAhx4AgCcdAIB/HgCAhwsAgICNAACBnQAAzOwDAM3oBAAHHQCAjQsAgKNJBwCTCwCADx0AgKO9BwAXHQCAGwAAgOoHAIALAACApKUHAOsEAICKBQCAAwAAgKhhBwDfDQCAZQAAgMgDAIAeCQCArWkHAIAtAQCBPQEAgl0BAINRAQCEYQEAuAQAgKwEAICHYQEAiK0BAIm1AQCKvQEAjykVALwFAIAgDACAzHgCAM3YBQCB3QEAgXEAAOcLAICC/QEAhBkAACYMAICH7QEAIwwAgMw0BADNMAQA6gsAgJ9pFQApDACAjMkBAM34BADM8AIAsUkBACEHAICB1QAAoxUBAKCZFQB2CACARgcAgIT1AADMKAQAzSwEAMYIAICveQEAqH0BADcNAICqaQEAVQkAgLQlAQC1KQEAowkBAAUMAIDqBgCA7gYAgLIFAQCzPQEAvPUAAL39AAC+2QAAOwgAgLgBAQC5AQEAugEBADwHAIBDBwCAhgwAALOdAwCyiQMAtggAgIC9AwBsBwCAbwcAgBUJAIDkBgCA5wYAgDgIAICJhQMAzOQHAL+hAwAIDACA2gwAgIxlAADN5AwAzCQMAIlBAACIVQAAi0UAAIpFAACFtQMAhLUDAIeVAwCGgQMABA0AgAcNAIAKDQCAmCwAABMAAICmyAAAzYwGAMyoBgCFaQAAFwAAgDEAAIBpAACAzPADAAcAAIA1AACA1AwAgLGVAAArDQCAs5UAALKVAAA7DQCAPg0AgEYNAIBBDQCANA0AgHUAAICmBgCAJQAAgJsJAIAjIQCAv1UDAEkNAIAfIQCAGyEAgGcgAIC4bAAAlGUNAJIAAgCcrQEAnaUBAJqJAQCbiQEAmJkBAJmJAQDMIAYAzQQGAMxABgDNXAYAzDwHAM04BwDMvAcAhXUAAIABDwCBDQ8AbyAAgLqZAQCFBQAAdyAAgF8gAIC+hQEAgSkPAIAlDwBrIACAgiEPAIUpAAC0pQEAhREAAHMgAICziQ8AsoUPALHJAQCwAQwAt4EPALbtAQC17QEAtO0BAIFlAQCAZQEAg2EBALi1DwDMPAsAhHkBAIDhDwCB3Q8AeyAAgGMgAIDMyAQAzbgEAIWtAACFFQAAJyEAgD8hAIDM6BkAzbQZAKRdAQBMDQCAok0CAKPxDwCgVQEAod0PAIIIAIBxCQCAPgkAgPMeAIBvCQCA+x4AgHoJAID3HgCAtAgAgJMNAACzHgCA/x4AgITVDACF6Q4AlGkAAIfdDgC7HgCAmbQCAMMeAIDLHgCAtx4AgEMhAIC/HgCAn3QBAMceAICRGA0AgI0OAIGBDgCGhQ4AlYwDAISJDgCXRAIAghEAAKm4AACA0QAAge0AAM8eAIBPDQCA6x4AgIVZDwCDiQAAoTQNAIFFDgCASQ4A7x4AgKU0AQCFYQ8AzPAUACMfAIC5xAUAzMgDAM3cAwCA3QAAgcEAACsfAIC/kAUAhREAALHsBwCA9QAAgcEAAKcgAIC1jAYAMx8AgLdABgCA3Q4AgekOAMwoAgDNtAIAgM0OAIH5DgCFKQAAg4UBAIB1AQCBsQEAgPEBAIHVAQCvIACAOx8AgIUFAAC3IACAgJkBAIG9AQCCfQAAk9UBAJThAQCFDQAAnyAAgCcfAICACQAAgRkAAC8fAICTrQEAlC0AAKsgAICFDQAANx8AgIUFAACzIACAPx8AgIUpAACCGQAAhTUAAIDxAACB4QAAuyAAgKMgAIBHIQCAhQUAAGchAICDdQEAgO0BAIEpAQDM8AEAzbABAFINAIBjIQCAXyEAgKkNAIBjHwCAax8AgIA9AACBDQAAcx8AgHsfAICALQAAgR0AAIIVAABnHwCAzSwBAG8fAIB3HwCAfx8AgIjFAwCrIQCAzJACAM28AgCE7QMAVQ0AgIb5AwCjHwCAgIEDAIH9AwCAPQAAgTUAAIFJAACAQQAAzdwBAIJBAACrHwCApx8AgK8fAIDNMAEAlJ0DAJMhAIDN8AEAzAwBAIG5AwCAxQMAg6EDAJOlAwCArQAAgdUAAICdAACBqQAAjyEAgFgNAICBwQAAgMkAAIC1AACBgQAAiyEAgINpBADMcAMAzbQDAIchAIDNPAEArA0AgJMBBADNjAIAzPQCAIANAACBNQAAlNkGANcfAIDbHwCA3x8AgMwIAQDNHAEAgREAAIApAACvIQCAghkAAICRAQCBkQEAzWgFAMyUAgDMEAkAzSgWAMxYDgDNeA4AzBQNAM3YCgDMKAwAzYwNAMzgFwDM4AoAzDgLAM30CACFEQAAWw0AgIBRBwCBUQcA5yAAgM2QDgCFBQAA7yAAgMzYDgDN7AEA9yAAgM0ADgCFGQAAzfAPAM08DgDNVA4AzGgBAM1sAQDfIACAZAgAgJSZBwDMwDsAgGEBAIHZAACFKQAAzWQOAMx4AQDNfAEAga0HAICtBwCFZQAAgp0HAIBRAQCBUQEAlOEHAM3AAACEeQEAk8UHAIZhAQDrIACAiCEBAIUNAADzIACAzRgBAMzYAADNtAAAgN0HAIHNBwCfHwCAhQkAANMfAID7IACAAyAAgOMgAIALIACAEyAAgBsgAIAPIACAByAAgLMhAIAXIACAHyAAgMy4AgDNHAMAgGUAAIF1AACCfQAAIyAAgIUJAACFQQAAByEAgK8NAICAmQYAgSEHAIUZAACDfQAADyEAgIVZAAADIQCA/yAAgIDNAACB2QAAkx4AgIURAACE6QAAmx4AgIblAABHIACAgDUAAIENAACjHgCAhR0AAE8gAICrHgCAhQUAAFcgAICAVQAAgW0AAIJ9AACTRQAAlA0AAIUNAAA/IACAlx4AgIAJAACBEQAAnx4AgIUdAABLIACApx4AgIUFAABTIACAgOkBAIHxAQCCBQAArx4AgIUJAACFCQAAWyAAgEMgAICAbQEAgXkBAIIZAACDpQEAEyEAgIV1AACFBQAAFyEAgAshAIAnIACAzMgCAM3cAgCyDQCA0x4AgIA5AACBOQAA2x4AgOMeAIDXHgCA3x4AgIAdAACBDQAA5x4AgCsgAICAxQAAgdUAAM3AAADMJAIAgNUAAIHFAACFOQAAg8kAACshAIC1DQCAgNUAAIEJAACFBQAAMyEAgAMfAICHIACAgAkAAIERAAALHwCAk5kAAJS5AAATHwCAhWUAAIU9AACPIACAk10AABsfAICFEQAAzXAFAMx0BQCUATwAlyAAgH8gAIDNKAEAiyAAgJMgAICFGQAAmyAAgIMgAIA7IQCALyEAgC8gAICFJQAAhTkAAMz4AgDNxAMAzTwBALgNAICBlQMAgI0DAM3EAQCCpQMAhVEAAIVJAADMKAEAzSwBAM04AQDMPAEAgGk+AIFpPgBPIQCASyEAgM04PADMVDwAgdE8AJOdPgDMSAEAzcgCAM00AQBTIQCAlLk+AF4NAICAoT4AgaE+AIKhPgCIjTwAWyEAgIWtAACALQAAgSEAAIXVPwCbHwCAgO0AAIHxAACGpQAASx8AgISpAADNJAEAzSgBAFMfAICI+T4AhfE/AFsfAIBPHwCAhcU/AM0wAQDNEAEAzfQGAIDdAQCB6QEAzbwGAM1wBgDM4AYAzVwBAMxoBgDNkAYAzWQGAM14BgDMrAcAzagHAMzoBwDNyAcAgk0/AIP9AgCANQIAgekCAFcfAIBfHwCAgAU9AIV9AQBXIQCAMyAAgM0UAQAvDgCAge0BAIDhAQDNPAEAgs0BAM0sAQCCdQEAgW0BAIBZAQCAZQEAgcUAAIsfAIDNJAEAzTgBAILxAACB+QAAgFkBAIApAACBcQAAzBgBAM18AQDNLAEAkx8AgIEdAACAHQAAjx8AgJcfAIB3IQCAzSQBAMzkPQDNXA8AzegAAMwMAQCA1QEAgckBAIKZAACD5T8ADx8AgBcfAIAfHwCANyEAgCkOAIB7IQCAQx8AgDcgAIBHHwCAMg4AgIBNPwCBQT8Agx8AgG8hAICHHwCAayEAgIAlPwCBKT8Ak5E/AIN9AAAsDgCAlEEAAMzYAgDNrAIAcyEAgJNVAACACQAAgR0AALsNAICDIQCAlEEAALMfAICAnQAAgaEAAIAdAACBEQAAhKUAALsfAICGpQAAwx8AgIjxAACC0QAAgdkAAIDNAACAJQAAgSkAAIIFAADLHwCAtx8AgL8fAIDHHwCAk7EAAJQRAADPHwCAgB0AAIEVAACAJQAAgS0AAII9AAB/IQCAgO0AAIHRAACCFQAAg4EAAIHQPQA7IACAzCACAM3cAQCFeAIAlyEAgDUOAICfIQCAiRgDAOMfAICALQAAgTUAAIAJAACBbQAA6x8AgMcgAICRsQAAkKkAAJPdOwCSAQQAlaUAAJSVOwDzHwCAlqEAAIUJAACTQQAAzyAAgPsfAICFBQAA1yAAgJT1AAC/IACAgLkAAIHdAACC5QAA5x8AgO8fAICF6QAAgAkAAIE1AACFBQAAyyAAgPcfAICFHQAA0yAAgP8fAICFBQAA2yAAgLHBBQCwxQMAwyAAgLLFAwC12QUAtM0DAKMhAICFOQAAuf0DAKchAICbIQCAwQ0AgNMNAIAdDgCABx8AgAsOAIDZDQCAzIgCABEOAIDN4D4AzZABAMwkAQB2DQCAlA0AgEcOAICDDgCAgLEAAM3UPgDN5D4AiQ4AgMy8PgDNuD4AgNEDAIHtAwCC/QMAhmkAAEQOAICFnQMAzTwBAD4OAIDM6AIAzTw/AIjlAADNGAEAjw4AgIhBAABBDgCAfQ4AgM0sAQCbDgCAgNUAAKEOAICG4QAAhukAAE0OAIDNJAEApw4AgM0QAQCI0QAAiCkAAMz4AgBTDgCAzfgCAMwkAQCtDgCAhS0DAMygPgDNbD4AgNUDAIHNAwCCAQMAg/kDAMxkAwDNzAIASg4AgM0kAQDMDAIAzQgCAIERAADMnAMAzLA+AM20PgDMxD4AzcA+AMyAPgDNuD4Asw4AgMyEAgDMmD8AzVA+AMwgPgDNoD4AzQw/AM0wPwDNeD8AzQQ/AIhZAADFDgCAzfgBAMzEAQBQDgCAyw4AgNEOAIDMFAIAzAgBAM3IAQCIBQAA1w4AgN0OAIDMKAIAvw4AgIgNAACG0QAAgB0BAITNAACI9QAAzDwCAIQ1AQDMRAIAhikBAIYOAICIZQEAjA4AgKdEBQBoDgCAi+0AAIjtAACBDQAAiCUAAIZlAADMcAIAzXQCAMwwAgDN2AUAYg4AgJIOAICAOQAAZQ4AgMzgBQCADgCAzCgBAM0UAQCGJQAAiFUAAA4OAICGhDAAyg0AgIDVBwCG/QcAng4AgMwkAgCIPQAApA4AgHEOAICIPQAAqg4AgMxIAgDNeAIAVg4AgLAOAICXwAUAlnAFAJUYBQCAaQAAk1gFAIE5AACIZQAAkPg8AIZZAACeqAUAhEUAAG4OAIDM1AIAmrQFAIBdAACYrAUAp+wEAIgRAADM2AIAzdwCAKO8BAC2DgCAzGACAMgOAIB0DgCAzg4AgK0IBADUDgCAq/QEAMwsAgCIBQAA2g4AgLfoAwC2HAQAtSgEAMwAAgCzKAQAi3kAAIh9AACwdAQAhkEAAL6kAwCEdQAAiB0AAOAOAIC6TAMAzNwDALj8AwCDqAIAiA0AAMIOAICIFQAAh5QCAMw4AgBrDgCAzAQCAIvcAgCPDQAAdw4AgI8ZAADMIAIAeg4AgI3wAgCIdQAAmCADAJksAwCVDgCAmg0AgMxMAgCWcAMAzCQCAIg9AACYDgCAzCwCAIgFAAC5DgCAzCQCAIgNAAC8DgCAh/UAAKjUAwCpxAMA4w4AgNlgAgDYDwCA2w8AgOEPAICUNQAAkzEAANloAgDeDwCA2UwCAJQFAADkDwCAlSEAAJQpAABWEACAehYAgEkXAIDYFgCA2WACAD0XAIC12AMAtPADAJQ1AADZWAIAYBcAgJQFAADZVAIAlA0AADcXAIDgdAEAisgAALwVAACIyAAA4IACAI0XAICBoAAApOwCAKTIAgCoXAAAvA0AAJ8XAIDghAIAvAUAAKMXAICk+AIA4PQCALDMAwCV0AAAYxcAgLPgAwCmyAIAp2ACAJLYAABqFwCAvsEAAHEXAICXwQAAeBcAgH8XAICGFwCAzXg/AMy8PwC+gA0AkRcAgLx4DAC9gA0AuvQMALtUDAC49AwAmBcAgLwXAIC3uAwAwBcAgLWMDACyoAMAs6AMAKcXAICxQAMArnACAK9kAwC4BQMArUgDAK4XAIC1FwCAqEQDAKnYAwDgFwCAp9gDAKRoAgCliAMAtjUDALc9AwCSyAIAtT0DAJldAQCYTQEAm2UBAJppAQCdZQEAnGUBAJ+FAQCemQEAh5wCAL6tAACWpQAAl70AAMw0BQDNjDcAzLg4AM2sOACflQEAth0AAJ2ZAQCc9QEAs7EBAK54AgDnFwCAxBcAgJk9AADLFwCAmxkAAJoJAADSFwCA2RcAgOBIAgCeCQAArFwCAK30AgAAGACA/BcAgAQYAIDuFwCAh2ADAPUXAICvVAIAvhEAAJcFAAAIGACA4KwCAAwYAICG+AMAh+wDAOC0AgAUGACAr0gCAK6QAgDgPAIAvg0AABAYAICXGQAA4NgCAIaEAwCWEQAAvwAMAJ1tAACcYQAAGBgAgLFMAgCzUAIAlQ0AABwYAICGnAMA4MgCALMEAgCCBQAAKBgAgLNQAgCVDQAALBgAgCAYAIAkGACA4LQCAIaMAwCH3AMAvg0AAJVpAACWeQAAMBgAgLToAgC1UAIAlwUAADgYAIDg1AIAtPQCAL4ZAADgoAIANBgAgODUAgCZjAMAt9QCAIoFAAA8GACAQBgAgIoVAAC3NAIAjx0AAEQYAIBIGACAswUAAEwYAICzBQAAYRgAgJwJAACdCQAAUxgAgFoYAICMBQAAaBgAgHMYAIB6GACAgRgAgJ9JAACIGACAjxgAgGwYAICWGACAnRgAgN8YAIDVGACA8BgAgOYYAICkGACAg8kBAIH5AQCyGACAuRgAgMAYAIDHGACAzhgAgKsYAICAtAIApYgDAOEIAgCuHQAA9xgAgLwJAACN9QEA+xgAgOEAAgCSlQEA45QQAJNFAACXiQEAhRQAAId4AQCGAAQAVzoAgFs6AIBfOgCAYzoAgGc6AICdeQAA74xoAJyhAQBrOgCAbzoAgKKZAABzOgCAdzoAgHs6AIB/OgCAp4kAAIM6AICHOgCAqUkBAIs6AICsqQAAjzoAgJM6AICXOgCAsyUBAJs6AICfOgCAozoAgLchAQC2OQEAtTEBAKc6AICrOgCAufkAALkRAQC4GQEArzoAgLM6AIC3OgCAuzoAgICwAQCEiAIAvzoAgIPIAQCEVAMAhFwEAMM6AICEXAUAgN0DAIEtAACCMQAAvjwCAMs6AIDPOgCAh4gDAIacBACzLQMA0zoAgNc6AIC+AAQAvhwFALbRAwC12QMA2zoAgLv5AwC68QMAmljTAYTgBwC/xQMAvtkDAL3dAwC83QMAvgAYAKUFAwCmDQMA3zoAgIQcGADjOgCA5zoAgKPxAwCsAQMArQEDAK4FAwCvGQMArKQbAq3cGgKqLQMAqyUDAL5MGQC+SBoA6zoAgL6AGwC04BoCtdQdArYwHgLvCAIA7zoAgOGgAQC6OBoC4/gCALoAAAC9ZBwCvvQcAr8AEAKRBNMBkOT2AeBEAQCSCD4C8zoAgPc6AID7OgCA/zoAgL6sHAADOwCABzsAgAs7AIAPOwCAEzsAgBc7AIAbOwCAgbBtAICAAQCDHFIAgth3AIUgmgCEkL4AhwjPAIaM5gCJbDcBiOAsAYsYfgGK2BMBjeClAYzwWgGP/OsBjliPAbDVFwCxAWgAso1rALOdawC0SWsAtZVvAB87AIDgcAEAIzsAgCc7AIArOwCALzsAgIAZAACBGQAAggUAADM7AIA7OwCAoaUCAKJJBwCjQQcApEEGAKXVGwCm3RsAp8EaAKgBHACp4R8AqkkfAKsBEACs9RMAra0TAK4BFACv+RcAqDEGAKkxBgCqTQYAq0UGAKxNBgCtmQYAro0GAK+FBgCGgAMAhxgDAD87AIBDOwCARzsAgEs7AIBPOwCAUzsAgLhtBwC5dQcAun0HALt1BwC8bQcAvc0HAL75BwC/+QcAsKkGALGFBgCyeQcAs3kHALRpBwC1aQcAtl0HALdVBwDHOgCAs8EGAFc7AIA3OwCAth0GAFs7AIBfOwCAtcEGALppBgC7RQYAYzsAgGc7AIC+qQcAv6kHALypBwC9qQcAo4UGAGs7AIBvOwCAczsAgHc7AICmWQYApYUGAHs7AICrAQYAqi0GAH87AICDOwCAr+0HAK7tBwCt7QcArO0HAKjBBgCpLQEAqiUBAKs9AQCsJQEArS0BAK4lAQCvlQEAhzsAgIs7AICPOwCAkzsAgJc7AICCvQAAgb0AAIC9AAC4nQEAua0BALqlAQC7bQAAvHUAAL19AAC+dQAAv20AALD1AQCx/QEAssEBALPBAQC0tQEAtb0BALa1AQC3rQEAmzsAgJ87AICjOwCAs6EBAKc7AIC1oQEAtqEBAKs7AICGgAEAh8QBALo9AQC7NQEAvBkBAL0ZAQC+fQEAv3UBAKPtAQCvOwCAszsAgLc7AIC7OwCApu0BAKXtAQC/OwCAq3kBAKpxAQDDOwCAxzsAgK85AQCuMQEArVUBAKxVAQDLOwCAzzsAgNM7AIDXOwCA2zsAgOGsAQDfOwCA42AGAOM7AIDnOwCA6zsAgO9UBgDvOwCA8zsAgL60GgD3OwCA+zsAgP87AICGaBwAh4wDAAM8AIAHPACACzwAgA88AICAOQAAgTkAAIIFAAATPACAGzwAgB88AIAjPACAJzwAgKgdAwCpQQMAqkEDAKtBAwCsQQMArUkDAK5xAwCvcQMAhCAdACs8AIAvPACAMzwAgDc8AIA7PACAPzwAgEM8AIC46QAAufUAALr9AAC78QAAvJEAAL2RAAC+iQAAv4kAALDhAACx4QAAsuEAALPhAAC04QAAte0AALbZAAC32QAA4wwHAOEgBwDhMAEA4wgHAEc8AIBLPACATzwAgFM8AIBXPACAWzwAgF88AIBjPACA75gHAGc8AIBrPACA74gHALOJAgBvPACAczwAgL6AGgB3PACAtokCALWJAgB7PACAu2UBALplAQB/PACAgzwAgL9pAQC+ZQEAvXUBALx1AQC3PQYAtj0GALU9BgC0IQYAszUGALI1BgCxAQYAsAkGAL9ZBgC+UQYAvVkGALxNBgC7bQYAunkGALlxBgC4eQYAgJ0AAIGtAACCpQAAizwAgI88AICTPACAlzwAgJs8AICvcQYArmkGAK1tBgCsbQYAq4EGAKqZBgCpkQYAqJkGABc8AICHPACAnzwAgKPFHQCjPACApcUdAKbFHQCnPACAhgADAIdkAwCqKR4AqykeAKw5HgCtOR4ArikeAK8lHgCzOR4AqzwAgK88AICzPACAtzwAgLb9HgC1/R4AuzwAgLvZHgC60R4AvzwAgMM8AIC/aR8AvmEfAL1pHwC8wR4AqPEeAKnxHgCq8R4Aq/EeAKw1HgCtPR4ArjUeAK8tHgDHPACAyzwAgM88AIDTPACA1zwAgNs8AIDfPACA4zwAgLjlHwC57R8AuuUfALv5HwC86R8AvZEfAL6RHwC/jR8AsFUeALFdHgCyVR4As/0fALTlHwC17R8AtuUfALfdHwCjeR8A5zwAgOs8AIDvPACA8zwAgKa9HwClvR8A9zwAgKuZHwCqkR8AhogAAIdMAQCvKR4AriEeAK0pHgCsgR8AgEkAAIFJAACCWQAAs5keAPs8AIC1iR4AtlEBAP88AIADPQCABz0AgLotAQC7JQEAvD0BAL0lAQC+JQEAvxUBAKhNHgCpVR4Aql0eAKtVHgCsTR4ArZ0BAK6JAQCvgQEAhKwBAAs9AIAPPQCAEz0AgBc9AIAbPQCAHz0AgCM9AIC4ZQEAuW0BALplAQC7fQEAvGUBAL1tAQC+ZQEAv9kAALClAQCxrQEAsqUBALO9AQC0rQEAtZ0BALaVAQC3XQEAo9UdACc9AIArPQCALz0AgDM9AICmHQIApcUdADc9AICraQIAqmECADs9AIA/PQCAr1kCAK5pAgCtaQIArHECAEM9AIBHPQCASz0AgE89AIBTPQCAVz0AgFs9AIBfPQCAgDkAAIE5AACCBQAAYz0AgGs9AIBvPQCAh0ADAIZcBACETAQAcz0AgHc9AICEBAUA4yABAHs9AIDhqAEAfz0AgO+UGgCDPQCAhz0AgIs9AICPPQCAkz0AgJc9AICbPQCAs6EDAJ89AICjPQCApz0AgKs9AIC2fQMAtX0DAK89AIC7WQMAulEDALM9AIC3PQCAv/0AAL79AAC9/QAAvEEDAKhRAgCpWQIAqmkCAKtpAgCstQIArb0CAK61AgCvrQIAhKgHALs9AIC/PQCAwz0AgIKpAADHPQCAgKkAAIGpAAC4aQEAuWkBALoJAQC7CQEAvBkBAL0ZAQC+CQEAvwkBALDVAgCx3QIAstUCALNpAQC0eQEAtXkBALZpAQC3YQEA4bgBAOHUHwDjOB8A4wwbAMs9AIDPPQCA0z0AgNs9AIDfPQCA4z0AgOc9AIDrPQCAvjwJAO89AIDvhBsA74QbAKOhAgDzPQCAhugEAIe8BQD3PQCApn0CAKV9AgD7PQCAq1kCAKpRAgD/PQCAAz4AgK/9AQCu/QEArf0BAKxBAgCzhQYA1z0AgAc+AIALPgCADz4AgLaJBgC1jQYAEz4AgLuRBgC6iQYAFz4AgBs+AIC/9QYAvokGAL2BBgC8iQYAHz4AgCM+AIAnPgCAKz4AgC8+AIAzPgCANz4AgO+EHQA7PgCA4QAEAD8+AIDj/AQAgBEAAIEdAACCBQAAQz4AgKjxBgCp8QYAqg0GAKsFBgCsBQYArQkGAK49BgCvNQYARz4AgEs+AICGiAAAhxADAE8+AIBTPgCAVz4AgFs+AIC4EQYAuRkGALohBgC7IQYAvPUHAL39BwC+9QcAv+kHALBNBgCxVQYAsl0GALNVBgC0TQYAtTEGALYxBgC3MQYAo4UHAF8+AIBjPgCAZz4AgGs+AICmiQcApY0HAG8+AICrkQcAqokHAHM+AIB3PgCAr/UHAK6JBwCtgQcArIkHAHs+AICz4QYAfz4AgIM+AIC25QYAhz4AgIs+AIC18QYAur0GALuNBgCPPgCAkz4AgL59AQC/ZQEAvJUGAL11AQCoHQYAqSUGAKotBgCrJQYArD0GAK0hBgCuXQYAr00GAJc+AICbPgCAnz4AgKM+AICnPgCAgrkDAIGxAwCAuQMAuO0BALmFAQC6jQEAu4UBALydAQC9hQEAvo0BAL+FAQCwPQYAsQ0GALIFBgCz5QEAtP0BALXlAQC25QEAt9UBAKOlBQCrPgCArz4AgLM+AIC7PgCApqEFAKW1BQC/PgCAq8kFAKr5BQCGCAwAhxwDAK8hAgCuOQIArTECAKzRBQDDPgCAs/ECAMc+AIDLPgCAtlUDAM8+AIDTPgCAteECALpxAwC7eQMA1z4AgNs+AIC+MQMAvz0DALxRAwC9UQMAqCUCAKk1AgCqPQIAqzUCAKwtAgCtkQMArpEDAK+RAwDfPgCA4z4AgOc+AIDrPgCArAAAAO8+AIDzPgCA9z4AgLiZAwC5rQMAuqUDALttAwC8dQMAvX0DAL51AwC/bQMAsPEDALH5AwCywQMAs8EDALSxAwC1vQMAtrUDALepAwD7PgCA/z4AgAM/AIAHPwCACz8AgA8/AIATPwCA76gaAL5oDADhlAEAFz8AgOMcBgCADQAAgXEAAIJxAAAbPwCAo/UDAB8/AIAjPwCAhEwCACs/AICmUQIApeUDAC8/AICrfQIAqnUCAIbIDACHLA0ArzkCAK41AgCtVQIArFUCAOFQBgAzPwCA4xQHAITADAA3PwCAOz8AgD8/AIBDPwCARz8AgEs/AIBPPwCAUz8AgFc/AIBbPwCA73gbAL74DwBfPwCAYz8AgGc/AICzjQEAaz8AgLWZAQC2jQEAbz8AgGc9AIBzPwCAuoUBALtNAQC8VQEAvV0BAL5VAQC/SQEAo0EOACc/AIB3PwCAez8AgH8/AICmQQ4ApVUOAIM/AICrgQ4AqkkOAIc/AICLPwCAr4UOAK6ZDgCtkQ4ArJkOAIBtAACBCQAAgh0AAI8/AIDvGAkAkz8AgJc/AICbPwCA4zwNAJ8/AIDhWAwAoz8AgIbQAACHvAMApz8AgKs/AICokQ4AqZkOAKrJDgCrxQ4ArN0OAK3BDgCuwQ4Ar/UOAIToAACvPwCAsz8AgLc/AIC7PwCAvz8AgMM/AIDHPwCAuMEPALnBDwC6wQ8Au8EPALzBDwC9wQ8AvsEPAL/1DwCwjQ4AsUUOALJNDgCzRQ4AtF0OALVBDgC2QQ4At0EOAKhRDgCpWQ4Aqo0OAKudDgCshQ4ArY0OAK6FDgCvvQ4Ayz8AgM8/AIDTPwCA1z8AgNs/AIDfPwCA4z8AgOc/AIC4kQ4AuZkOALqtDgC7RQEAvF0BAL1FAQC+RQEAv3UBALDFDgCxzQ4AssUOALPdDgC0xQ4AtbUOALa9DgC3tQ4AswUOAOs/AIDvPwCA8z8AgPc/AIC2DQ4AtQ0OAPs/AIC7CQ4AugEOAP8/AIADQACAv3EOAL4BDgC9CQ4AvBEOAIJtAACjQQ4AgFUAAIFlAACmSQ4AC0AAgA9AAIClSQ4AqkUOAKtNDgCGSAAAh3gAAK5FDgCvNQ4ArFUOAK1NDgCoXQIAqWECAKplAgCrdQIArG0CAK2xAgCusQIAr7ECAITsBAATQACAF0AAgBtAAIAfQACAI0AAgCdAAIArQACAuHEDALlxAwC6cQMAu3EDALzVAwC93QMAvtUDAL/NAwCw0QIAsdECALLRAgCz0QIAtFEDALVRAwC2UQMAt1EDAC9AAICz6QIAM0AAgL6ABAC2NQIAN0AAgDtAAIC14QIAuhECALsRAgA/QACAQ0AAgL6RAwC/kQMAvAECAL0BAgBHQACAS0AAgKOlAgBPQACApa0CAFNAAIBXQACApnkCAFtAAIBfQACAq10CAKpdAgCtTQIArE0CAK/dAwCu3QMAqNUCAKndAgCqLQEAqyUBAKw9AQCtJQEAri0BAK8lAQBjQACAZ0AAgGtAAIBvQACAc0AAgHtAAIB/QACAg0AAgLiFAQC5iQEAup0BALuVAQC8sQEAvbEBAL55AAC/eQAAsF0BALHlAQCy4QEAs/kBALTpAQC13QEAttUBALe9AQDh8A4Ah0AAgOMUDgCLQACAgb0AAIC9AACPQACAgq0AAIYABACH7AUAk0AAgJdAAICbQACAn0AAgO9gDgCjQACAp0AAgKtAAICFXH0Ar0AAgLNAAIDjZAEAt0AAgOG0AQC7QACA76AOAL9AAIC3PgCAhPgFAMNAAIDHQACAy0AAgLMlBgB3QACAz0AAgNNAAIDXQACAtiUGALU1BgDbQACAu6EGALoZBgDfQACA40AAgL+ZBgC+rQYAva0GALy1BgCCbQAA7zAEAIBVAACBZQAAvlwDAOdAAICG+AAAh2wDAOtAAIDvQACA80AAgPdAAID7QACA40QEAP9AAIDhjAcAo6UGAANBAIAHQQCAC0EAgA9BAICmpQYApbUGABNBAICrIQYAqpkGABdBAIAbQQCArxkGAK4tBgCtLQYArDUGAB9BAICz+QcAI0EAgCdBAIC2SQcAK0EAgC9BAIC1UQcAulEHALtRBwAzQQCAN0EAgL41BwC/OQcAvEUHAL09BwCoNQYAqT0GAKo1BgCriQYArJ0GAK2NBgCusQYAr7EGADtBAIA/QQCAQ0EAgEdBAICADQAAgbEAAIKxAABLQQCAuKEGALmtBgC6vQYAu7UGALytBgC9XQEAvlUBAL9NAQCw0QYAsdEGALLVBgCzrQYAtLUGALW5BgC2qQYAt6UGAKO9BgBPQQCAU0EAgISEAgC+kAEApg0GAKUVBgBbQQCAqxUGAKoVBgCGCAAAh3wBAK99BgCucQYArXkGAKwBBgBfQQCAs60BAGNBAIBnQQCAtqkBAGtBAIBvQQCAta0BALptAQC7dQEAc0EAgHdBAIC+XQEAvzUBALxlAQC9VQEAqGECAKlhAgCqYQIAq2ECAKxhAgCtbQIArp0CAK+VAgB7QQCAf0EAgINBAICHQQCAi0EAgI9BAICTQQCAl0EAgLiVAgC5nQIAuqECALuhAgC8cQMAvXEDAL5xAwC/cQMAsO0CALH1AgCy9QIAs8UCALTdAgC1tQIAtrECALexAgCbQQCAn0EAgKNBAICj5QIAp0EAgKXlAgCm4QIAq0EAgK9BAICzQQCAqiUCAKs9AgCsLQIArR0CAK4VAgCvfQIAt0EAgLtBAIC/QQCAhEB8AIAVAACBHQAAggUAAMNBAIC+7HwAy0EAgIZIfQCHCAMAz0EAgNNBAIDXQQCA20EAgKidAgCpxQIAqsECAKvBAgCsxQIArc0CAK7xAgCv8QIA30EAgONBAIDnQQCA60EAgMkAAADvQQCA80EAgPdBAIC4wQEAucEBALrBAQC73QEAvM0BAL31AQC+/QEAv50BALBBAQCxQQEAskEBALNBAQC0QQEAtUEBALZBAQC3QQEA4TgGAPtBAIDjaAYA/0EAgANCAIAHQgCAC0IAgISUfQC+rHwAD0IAgBNCAIAXQgCAvrh/ABtCAIDvEAEAH0IAgCNCAIAnQgCAK0IAgC9CAIDhkAEAM0IAgONEAAA7QgCAgS0AAIAtAADvgAAAgjkAAD9CAIBDQgCAB0AAgEdCAIDhsH8Ax0EAgOPUfABLQgCAN0IAgE9CAICGuAAAh9QCAFNCAIBXQgCAW0IAgF9CAIBjQgCAZ0IAgO8gfABrQgCAs4l9AG9CAIBzQgCAd0IAgHtCAIC2jX0AtY19AH9CAIC7RX4AukV+AINCAICHQgCAv0V+AL5FfgC9VX4AvFV+AKNJfQCLQgCAj0IAgJNCAICXQgCApk19AKVNfQCbQgCAq4V+AKqFfgCfQgCAo0IAgK+FfgCuhX4ArZV+AKyVfgCCbQAAszF+AIBVAACBZQAAtvF/AITcAwCnQgCAtSF+ALrNfwC70X8AhgAEAIfUAAC+dX8Av3l/ALzBfwC9wX8AqOV/AKn1fwCq/X8Aq/V/AKztfwCtNX4Arj1+AK81fgCrQgCAr0IAgLNCAIC3QgCAu0IAgL9CAIDDQgCAx0IAgLjZfgC54X4AuuF+ALvhfgC85X4Avel+AL6ZfgC/mX4AsE1+ALFRfgCyUX4As1F+ALT1fgC1+X4Atul+ALfpfgCjdX8Ay0IAgM9CAIDTQgCA10IAgKa1fgClZX8A20IAgKuVfgCqiX4A30IAgONCAICvPX4ArjF+AK2FfgCshX4A50IAgLMxfgDrQgCA70IAgLbFAQDzQgCA90IAgLXRAQC6yQEAu8kBAPtCAID/QgCAvs0BAL+xAQC8yQEAvckBAKjdfQCp9X0Aqv19AKvxfQCsHQIArQECAK45AgCvOQIAA0MAgAdDAIALQwCAD0MAgIIFAAATQwCAgBEAAIERAAC4EQIAuRkCALohAgC7IQIAvNUCAL3dAgC+1QIAv80CALBJAgCxSQIAslkCALNZAgC0TQIAtTECALYxAgC3MQIAvgADAKNxfQCEiAIAvoAEAKaFAgAbQwCAH0MAgKWRAgCqiQIAq4kCAIYoBACHDAMAro0CAK/xAgCsiQIArYkCACNDAICEyAMAhcwFALPlAwAnQwCAteUDALbtAwArQwCAL0MAgDNDAIC6bQMAu2UDALx9AwC9ZQMAvmUDAL9VAwA3QwCAO0MAgL8ABACjJQIAP0MAgKUlAgCmLQIAQ0MAgEdDAIBLQwCAqq0CAKulAgCsvQIAraUCAK6lAgCvlQIAT0MAgFNDAIBXQwCAW0MAgF9DAIDjzAMAY0MAgOGsAQBnQwCA7xwDAGtDAIBvQwCAc0MAgHdDAIB7QwCAf0MAgOFwfwBXQQCA4wR+AINDAICLQwCA4ZQBAI9DAIDjWAEAgNkAAIHZAACCJQAA7+R+AJNDAICXQwCA7+B+AJtDAICzAQEAn0MAgIboBwCHLAQAo0MAgLY1AQC1BQEAp0MAgLvxAAC64QAAq0MAgK9DAIC/sQAAvtEAAL3ZAAC84QAAF0MAgIdDAICzQwCAt0MAgKEBBACgEQQAoxkAAKLFBACotQYAqb0GAKrpBgCr/QYArO0GAK3VBgCu3QYArz0HALBFBwCxVQcAslUHALNtBwC0dQcAtRUHALYdBwC3FQcAuC0HALk1BwC6MQcAuw0HALwZBwC9GQcAvgkHAL8JBwCjQQYAu0MAgL9DAIDDQwCAx0MAgKZ1BgClRQYAy0MAgKuxBwCqoQcAj8ltAM9DAICv8QcArpEHAK2ZBwCsoQcAld11AJTBdACXzXAAli1zAJFdaACQVWgAk9l0AJJNaQCd5XgAnB17AJ9tBwCeuXgAmR1/AJhVcACboXwAmvl8AIJhbACDhWkA00MAgNdDAICGEXUAhxF1AISVaQCFjWgAij10AIvFcgDbQwCA30MAgI7dfgCPMX0AjD1xAI2dcQCSGX0Ak716AONDAIDvkAkAltUGAJdRBQCUXXkAlQl5AJpxBQCbvQUA50MAgOtDAIDvQwCA4agFAJx5AQDjuAgAoYUBAPNDAICjqQ0AogEMAKUBCACkOQ0Ap6kJAKa9CQCppRUAqAEUAKsBFACq/RUArbkRAKyxEQCvARwArqEQALH9HACw5R0As+kZALIBGAC1ASQAtH0ZAIQUAAC+FAAAgI0AAIGVAACCbQAA+0MAgIZQDwCHZAAA/0MAgANEAIC61QcAu90HALjBBwC5wQcAvjEEAL8xBAC88QcAvfEHALKtBwCztQcAsK0HALGlBwC2nQcAt/UHALSlBwC1lQcAqmkHAKtpBwCoaQcAqWkHAK5pBwCvaQcArGkHAK1pBwAHRACAC0QAgA9EAIATRACAF0QAgBtEAIAfRACAI0QAgKgRBQCpHQUAqjkFAKs5BQCsLQUArVEFAK5JBQCvQQUAJ0QAgCtEAIAvRACAM0QAgDdEAIA7RACAP0QAgENEAIC4XQIAuWkCALrBAwC7wQMAvPkDAL35AwC+kQMAv7UDALAJBQCxCQUAsuECALPhAgC0dQIAtX0CALZ1AgC3bQIAs7EEAIQAAgC+BA0AR0QAgEtEAIC20QQAtaUEAE9EAIC7zQQAus0EAFNEAIBXRACAv7kDAL6xAwC9NQMAvDUDAFtEAICj9QQAX0QAgGNEAICmlQQAa0QAgG9EAICl4QQAqokEAKuJBACHqA0AhswMAK71AwCv/QMArHEDAK1xAwDhUAYA4TQHAONAAADjWAcAgNEAAIHdAACC1QAAc0QAgHdEAIB7RACAf0QAgINEAICHRACAi0QAgO+cAADvyAcAj0QAgJNEAICzNQIAl0QAgLW1AQCbRACAn0QAgLa1AQC+7AwAo0QAgLuRAQC6mQEAvVEBALyJAQC/UQEAvlkBAKjtDQCp/Q0AqvUNAKttDgCsdQ4ArX0OAK51DgCvbQ4AZ0QAgKdEAICrRACAr0QAgLNEAIC3RACAu0QAgL9EAIC49Q4Auf0OALr1DgC7QQ8AvEEPAL1JDwC+cQ8Av3EPALAVDgCxHQ4AshUOALPNDgC01Q4Atd0OALbVDgC3zQ4Ao30NAMNEAIDHRACAy0QAgM9EAICm/Q4Apf0OANNEAICr2Q4AqtEOAISoAgDXRACArxkOAK4RDgCtGQ4ArMEOAIBNAACBVQAAglUAALNRDwDbRACAtXEPALZxDwDfRACAhuAAAIcEAwC6XQ8Auy0PALw1DwC9OQ8Avi0PAL8lDwCoVQ4AqV0OAKqVDgCrrQ4ArLUOAK29DgCutQ4Ar60OAONEAIDnRACA60QAgO9EAIDzRACA90QAgPtEAID/RACAuGkBALlpAQC6eQEAu3kBALxpAQC9aQEAvt0BAL/VAQCw1Q4AsaUOALKtDgCzoQ4AtKUOALWtDgC2nQ4At1kBAKMdDgADRQCAB0UAgPdDAIALRQCApj0OAKU9DgAPRQCAq2EOAKoRDgATRQCAF0UAgK9pDgCuYQ4ArXUOAKx5DgAbRQCAH0UAgCNFAIAnRQCAK0UAgC9FAIAzRQCAN0UAgIANAACBFQAAgh0AADtFAIA/RQCAQ0UAgIR4AQC+FAAA4xQPAEtFAIDh4A0AhAADAIawBACHFAMAT0UAgFNFAIBXRQCAW0UAgF9FAIBjRQCA78APAGdFAIBrRQCAb0UAgHNFAIB3RQCAe0UAgLNtAwB/RQCAtX0DALZ1AwCDRQCAh0UAgItFAIC6UQMAu1EDALz1AwC9/QMAvukDAL/hAwCPRQCAk0UAgJdFAICbRQCAn0UAgKNFAICnRQCAq0UAgKhxAgCpeQIAqokDAKuJAwCsmQMArZkDAK6JAwCviQMAsPkDALH5AwCyTQMAs0UDALRBAwC1SQMAtnEDALdxAwC4IQMAuSEDALohAwC7IQMAvCEDAL0hAwC+IQMAvyEDAICdAQCBEQAAghEAAIQEBQDvFAAAr0UAgLNFAIC+EAUA48gAALtFAIDh0AEAv0UAgMNFAIDHRQCAy0UAgM9FAICqeQIAq3kCAIboBACHYAUArsECAK/JAgCs3QIArdUCANNFAICjRQIA10UAgNtFAICmXQIA30UAgONFAIClVQIA50UAgOtFAIDvRQCA80UAgPdFAID7RQCA/0UAgO+EDgC+rAQA4dAOAANGAIDjFAEAB0YAgAtGAIAPRgCAE0YAgLPdAQAXRgCAG0YAgB9GAIAjRgCAtv0BALX9AQArRgCAu90BALrdAQCE4AQAL0YAgL+hAQC+vQEAvb0BALy9AQCoBQYAqR0GAKoVBgCrLQYArDUGAK09BgCuNQYArykGALdFAICC9QcAgeUHAIDlBwAnRgCAM0YAgIYcAACHsAMAuCUGALnFBgC6zQYAu8UGALzdBgC9xQYAvs0GAL/FBgCwWQYAsVkGALIpBgCzKQYAtDkGALUlBgC2JQYAtx0GAKOdBgA3RgCAO0YAgD9GAIBDRgCApr0GAKW9BgBHRgCAq50GAKqdBgBLRgCAT0YAgK/hBgCu/QYArf0GAKz9BgBTRgCAs/UHAFdGAIBbRgCAtu0HAF9GAIBjRgCAteUHALqNBwC7kQcAZ0YAgGtGAIC+dQcAv30HALyBBwC9fQcAqCUGAKkpBgCqOQYAqzkGAKwpBgCtKQYArnkGAK91BgBvRgCAc0YAgHdGAIB7RgCAf0YAgINGAICHRgCAi0YAgLjVBgC53QYAuuEGALv9BgC85QYAve0GAL7lBgC/mQYAsA0GALERBgCyEQYAs+0GALT1BgC1/QYAtvUGALftBgCjsQYAgi0AAIEVAACAsQAAR0UAgKapBgCloQYAj0YAgKvVBgCqyQYAk0YAgL5oAQCvOQYArjEGAK05BgCsxQYAm0YAgLPxAQCGaAAAh3wBALZdAQCfRgCAo0YAgLVVAQC6SQEAu0kBAKdGAICrRgCAvj0BAL8hAQC8OQEAvTUBAK9GAICzRgCAhAQDAL6AHAC3RgCA4RwGALtGAIDjAAYAvwguAL9GAIDDRgCA78gHAMdGAIDLRgCAz0YAgNNGAIDXRgCA20YAgKN9AgDfRgCApdkCAONGAIDnRgCAptECAOtGAIDvRgCAq8UCAKrFAgCtuQIArLUCAK+tAgCusQIAqW0FAKhZBQCrDQIAqrkCAK0dAgCsHQIArwUCAK4NAgC+aB0A80YAgPdGAID7RgCAgB0AAIEJAACCmQEA/0YAgLnhAwC4KQIAu+EDALrpAwC94QMAvPkDAL/hAwC+6QMAsU0CALBNAgCzIQIAsi0CALUlAgC0OQIAtxECALYlAgCowQIAqdECAKrRAgCr5QIArP0CAK0VAQCuHQEArw0BAANHAIALRwCAD0cAgBNHAIAXRwCAG0cAgB9HAIAjRwCAuAUBALkJAQC6HQEAuxUBALwxAQC9MQEAvv0BAL/1AQCweQEAsUEBALJBAQCzXQEAtEUBALVNAQC2RQEAtz0BAIagHQCHxB0AJ0cAgO/YAAArRwCAL0cAgDNHAIDvxAYAhGwcAOH0BgA3RwCA47AGADtHAIDhlAEAP0cAgONEBgCzGQIAQ0cAgEdHAIBLRwCAhewsALbVAQC1NQIAT0cAgLvFAQC6/QEAU0cAgFdHAIC/yQEAvsEBAL3JAQC81QEAo9kdAAdHAIBbRwCAX0cAgGNHAICmFR4ApfUdAGdHAICrBR4Aqj0eAGtHAIBvRwCArwkeAK4BHgCtCR4ArBUeAIBpAACBaQAAggUAAHNHAIB3RwCAe0cAgIcQAwCGfAMAf0cAgINHAICHRwCAi0cAgI9HAICTRwCAl0cAgJtHAICopR8Aqa0fAKqlHwCrvR8ArKUfAK2tHwCupR8ArxUfAJ9HAICjRwCAp0cAgKtHAICvRwCAs0cAgLdHAIC7RwCAuA0fALkZHwC6IR8AuyEfALzZAAC92QAAvskAAL/BAACwcR8AsXEfALJxHwCzRR8AtEEfALVNHwC2PR8AtzUfALMtHgC/RwCAw0cAgMdHAIDLRwCAti0eALUtHgDPRwCAu7UeALq1HgDTRwCA10cAgL+JHgC+hR4AvZEeALylHgCCKQAAo2keAIAdAACBFQAApmkeANtHAIDfRwCApWkeAKrxHgCr8R4A40cAgITgAQCuwR4Ar80eAKzhHgCt1R4AqNUBAKnlAQCq7QEAq+UBAKz9AQCt5QEAru0BAK/lAQC+oAEAl0YAgOdHAIDrRwCAhhAAAId0AQDvRwCA80cAgLh9AQC5wQAAusEAALvBAAC8wQAAvckAAL7xAAC/8QAAsJ0BALFFAQCyTQEAs0UBALRdAQC1RQEAtk0BALdFAQD3RwCA+0cAgP9HAIADSACAB0gAgO80AgDv7B4AC0gAgOHwHQDj4AIA4zAeAOGEAQAPSACAE0gAgBdIAIAbSACAsyUCAJQAAAAfSACAI0gAgCdIAIC2JQIAtTUCACtIAIC7wQIAuhkCAC9IAIAzSACAv8ECAL7ZAgC90QIAvNkCADdIAIA7SACAP0gAgKPpAgBDSACApfkCAKbpAgBHSACAS0gAgE9IAICq1QIAqw0CAKwVAgCtHQIArhUCAK8NAgCAYQAAgWEAAIIFAABTSACAW0gAgIQABAC+FAQAX0gAgIbABACHUAMAY0gAgGdIAIBrSACAb0gAgHNIAIB3SACAqK0CAKm9AgCqtQIAqw0BAKwVAQCtHQEArhUBAK8NAQCE7AQAe0gAgH9IAICDSACAh0gAgItIAICPSACAk0gAgLgdAQC5LQEAuiUBALvNAQC81QEAvd0BAL7JAQC/wQEAsH0BALFVAQCyXQEAs1UBALRNAQC1PQEAtjUBALctAQDhGB4Al0gAgOM4HgCbSACAn0gAgKNIAICnSACAq0gAgK9IAICzSACAvmAEALdIAICBdQAAgHUAAO/gHwCCbQAAu0gAgL9IAICG6AQAh3wFAMNIAIDhkAEAy0gAgOOgAADPSACA00gAgNdIAIDvtAAA20gAgN9IAIDjSACA50gAgLUFBgBXSACAx0gAgLYFBgDrSACA70gAgLOlBQDzSACAvRkGALwRBgC/YQYAvhEGAPdIAID7SACAuwkGALohBgCj/QUA/0gAgANJAIAHSQCAC0kAgKZdBgClXQYAD0kAgKtRBgCqeQYAE0kAgBdJAICvOQYArkkGAK1BBgCsSQYAqFEGAKlZBgCqYQYAq2EGAKxhBgCtYQYArmEGAK9hBgAbSQCAH0kAgCNJAIAnSQCAgA0AAIGxAQCCsQEAK0kAgLhNBwC5VQcAul0HALtVBwC8TQcAvXUHAL59BwC/cQcAsMUHALHNBwCyxQcAs90HALTFBwC1zQcAtsUHALd5BwCz6QcAL0kAgDNJAICEwAEAvtgBALbhBwC16QcAN0kAgLsJBgC6AQYAhogAAIesAQC/CQYAvgEGAL0JBgC8EQYAO0kAgKOtBwA/SQCAQ0kAgKalBwBHSQCAS0kAgKWtBwCqRQYAq00GAE9JAIBTSQCArkUGAK9NBgCsVQYArU0GAKhZBgCpZQYAqm0GAKtlBgCsYQYArWEGAK5hBgCvYQYAhKwBAFdJAIBbSQCAX0kAgGNJAIBnSQCAa0kAgG9JAIC4kQEAuZkBALqhAQC7oQEAvHEBAL1xAQC+cQEAv3EBALDxAQCx8QEAsvUBALPdAQC0xQEAtbEBALaxAQC3sQEAs+UFAHNJAIB3SQCAe0kAgH9JAIC24QUAtekFAINJAIC7NQIAujUCAIdJAICLSQCAv3UCAL4BAgC9CQIAvCECAI9JAICjoQUAk0kAgJdJAICmpQUAm0kAgJ9JAIClrQUAqnECAKtxAgCjSQCAvigDAK5FAgCvMQIArGUCAK1NAgCA1QAAgd0AAILhAACrSQCA4yABAK9JAIDhqAEAs0kAgO80AgC3SQCAhggMAIdoAwCsAAAAu0kAgL9JAIDDSQCAs40DAMdJAIDLSQCAhIAMAM9JAIC2vQMAtYEDANNJAIC7TQMAuk0DANdJAIDbSQCAv00DAL5NAwC9TQMAvE0DAKhBAgCpTQIAqkUCAKtZAgCsSQIArX0CAK51AgCvuQIAvmgNAN9JAIDjSQCA50kAgIRsDADrSQCA70kAgPNJAIC4TQEAuVUBALpVAQC7ZQEAvH0BAL0VAQC+EQEAvxEBALDJAgCxyQIAstkCALPZAgC0yQIAtckCALZ9AQC3dQEA4XgHAOOYAADjuAYA4VwGAPdJAID7SQCA/0kAgANKAIAHSgCAC0oAgA9KAIATSgCA7AAAAO9cAADv6AYAG0oAgIFpAACAYQAAo4UCAIJhAACliQIAH0oAgCNKAICmtQIAhkAMAIfEDACrRQIAqkUCAK1FAgCsRQIAr0UCAK5FAgCojQ4AqZEOAKqVDgCrqQ4ArKUOAK2tDgCupQ4Ar9kOABdKAIAnSgCAK0oAgC9KAIAzSgCAN0oAgDtKAIA/SgCAuHUPALl9DwC6dQ8Au90PALzFDwC9zQ8AvsUPAL/9DwCwqQ4AsbUOALK1DgCzhQ4AtJ0OALVRDwC2UQ8At1EPALMdDgBDSgCAR0oAgEtKAIBPSgCAti0OALUtDgBTSgCAu3EOALptDgBXSgCAW0oAgL+VDwC+WQ4AvVEOALxhDgBfSgCAo1kOAGNKAIBnSgCApmkOAGtKAIBvSgCApWkOAKopDgCrNQ4Ac0oAgHdKAICuHQ4Ar9EPAKwlDgCtFQ4AqL0OAKnRDgCq0Q4AqykBAKw5AQCtOQEArikBAK8pAQCADQAAgRUAAIIdAAB7SgCAf0oAgINKAIC+dAIAh0oAgLjtAQC5hQEAuoEBALuBAQC8hQEAvY0BAL6xAQC/sQEAsFkBALFZAQCy7QEAs+UBALT9AQC15QEAtuUBALfVAQCLSgCAtqkBALWhAQCPSgCAs0kOAJNKAICGOAAAh9wBAL8xAQC+KQEAvSEBALwpAQC7jQEAuo0BAKdJAICXSgCAoxkOAJtKAICfSgCAo0oAgKdKAICm+QEApfEBAKtKAICr3QEAqt0BAK9KAICzSgCAr2EBAK55AQCtcQEArHkBALdKAIDv3A8Au0oAgL9KAIDDSgCAx0oAgMtKAIDPSgCA00oAgNdKAIDbSgCA30oAgONKAIDj6A4A50oAgOGMDgCAEQAAgREAAIIRAACEQAIA60oAgO9KAIDzSgCAvhADAIbABACHRAMA+0oAgP9KAIADSwCAB0sAgAtLAIAPSwCA7yQCABNLAIAXSwCAG0sAgB9LAIAjSwCAJ0sAgCtLAICE7AQAL0sAgDNLAIA3SwCA4+wCADtLAIDhOAEAP0sAgLNVAwBDSwCAR0sAgEtLAIBPSwCAth0DALUdAwBTSwCAuwkDALo5AwBXSwCAW0sAgL/9AAC+/QAAvfkAALwRAwCogQIAqYkCAKqdAgCrsQIArNUCAK3dAgCu1QIAr80CAIDNAQCBCQAAghkAAF9LAIBjSwCAa0sAgL5wBQBvSwCAuFkBALlZAQC6aQEAu2kBALx5AQC9eQEAvmkBAL9lAQCwvQIAsY0CALKFAgCzbQEAtHkBALV5AQC2aQEAt2kBAIYgBACHCAUAc0sAgHdLAIB7SwCAf0sAgINLAIDvXAAAhOwEAOFcDgCHSwCA44wOAItLAICPSwCAk0sAgJdLAICjVQIAm0sAgJ9LAICjSwCAp0sAgKYdAgClHQIAq0sAgKsJAgCqOQIAr0sAgLNLAICv/QEArv0BAK35AQCsEQIAqGkGAKlpBgCqeQYAq3kGAKxpBgCtaQYArp0GAK+VBgBnSwCAt0sAgLtLAIC/SwCAw0sAgMdLAIDLSwCAz0sAgLj1BgC5+QYAuo0GALuFBgC8nQYAvYUGAL6FBgC/tQYAsO0GALH1BgCy/QYAs/UGALTtBgC10QYAttEGALfRBgCz8QYAghUAAIG1AACAtQAA00sAgLbpBgC14QYAvtQDALsxBgC6KQYA10sAgNtLAIC/FQYAvikGAL0hBgC8KQYA30sAgKO1BgCGyAAAh8gAAKatBgDjSwCA50sAgKWlBgCqbQYAq3UGAOtLAIDvSwCArm0GAK9RBgCsbQYArWUGAKg1BgCpOQYAqoEGAKuBBgCsgQYArYEGAK6BBgCvtQYA80sAgPdLAID7SwCA/0sAgANMAIAHTACAC0wAgA9MAIC4nQYAua0GALqlBgC7aQEAvHkBAL15AQC+aQEAv2kBALDRBgCx0QYAstEGALPRBgC0tQYAtb0GALa1BgC3rQYAswkGABNMAIAXTACAG0wAgB9MAIC2AQYAtQkGACNMAIC7FQYAuhUGACdMAIArTACAv3kGAL5xBgC9BQYAvAUGAC9MAICjTQYAM0wAgPdKAICmRQYAN0wAgDtMAIClTQYAqlEGAKtRBgA/TACAQ0wAgK41BgCvPQYArEEGAK1BBgCB6QMAgN0DAISIAwCC4QMAhrA8AIeIAgC+VAMAS0wAgE9MAIBTTACAV0wAgFtMAIBfTACAY0wAgGdMAIBrTACA4/AGAG9MAIDhMAYAhAA8AHNMAIB3TACAe0wAgH9MAICDTACAhTQ9AIdMAICLTACA77AHAI9MAICTTACAl0wAgJtMAICfTACAo0wAgL7EPACnTACAgp0BAIGdAQCAnQEAqA0CAKllAgCqfQIAq3UCAKxZAgCtWQIArpkDAK+ZAwCw6QMAsekDALL5AwCz+QMAtOkDALXpAwC2XQMAt1UDALhtAwC5dQMAunUDALtFAwC8XQMAvTUDAL4xAwC/KQMAq0wAgK9MAICzTACAu0wAgOFgAwDv9AMA40QCAL9MAIDDTACA4zwDAO/0NwDh/AEAx0wAgMtMAIDPTACA00wAgIZkPwCHaD0AhTQhALOZAwDXTACAtb0DALa1AwDbTACA30wAgONMAIC6QQIAu0ECALxBAgC9QQIAvkECAL9BAgDnTACA60wAgO9MAIDzTACA90wAgPtMAID/TACA7/gBAIRoPADhPAYAA00AgOMcBgAHTQCAC00AgA9NAIATTQCAoxUDABdNAIAbTQCAH00AgCNNAICmOQMApTEDACtNAICrzQIAqs0CAL5kPgAvTQCAr80CAK7NAgCtzQIArM0CAKgdPgCpJT4Aqi0+AKslPgCsPT4ArSU+AK4tPgCvJT4At0wAgIL1PwCB5T8AgOU/ACdNAIAzTQCAhgAEAIecAwC4LT4AuTE+ALoxPgC7MT4AvNE+AL3RPgC+0T4Av80+ALBdPgCxIT4Asjk+ALM5PgC0KT4AtSk+ALYZPgC3FT4As6U+ADdNAIA7TQCAP00AgENNAIC2pT4AtbU+AEdNAIC75T4Aupk+AEtNAIBPTQCAv+0+AL7tPgC97T4AvO0+AFNNAICj4T4AV00AgFtNAICm4T4AX00AgGNNAICl8T4Aqt0+AKuhPgBnTQCAa00AgK6pPgCvqT4ArKk+AK2pPgCPBSUAsyU+AG9NAIBzTQCAtik+AHdNAIB7TQCAtSk+ALp9PgC7RT4Af00AgINNAIC+tT4Av70+ALxdPgC9vT4An304AJ5lOQCd8TgAnFE0AJtZNQCaUTUAmfEwAJgNMQCXZTEAlsEwAJVZLQCUTS0Ak+EsAJLZKQCRWSkAkPEoALSlGQC13RgAh00AgIQIAACwkRUAsQEVALIBGACzvRkAgA0AAIGtAwCCpQMAi00AgKNhAACiHT0AoZk9AKBxPACkxQUApUEEAKYBCACn4QkAR0wAgKH1AQCi6QEAo90FAKwBEACtxREArtkRAK85EACoZQgAqQEMAKrZDQCrCQ0AijEuAIuhMwCPTQCAk00AgI65MwCPETYAjB0yAI1NMgCCJSYAg6krAL5kAwCEYAQAhqEvAIcVLgCEGSoAhZEqAJphPgCb7T4AhsgEAIfcAwCbTQCA4Vw+AJyJAwDjAD4Akmk2AJN5NwCfTQCA7xg+AJZNOwCXuT8AlME7AJVdOgCpnT0AqIk9AKu5PQCqrT0Arak9AKyhPQCvyT0ArqE9AL7oBACjTQCAp00AgKtNAICvTQCAs00AgLdNAIC7TQCAuVk9ALhRPQC7eT0AumU9AL1pPQC8YT0Avx09AL5hPQCxgT0AsLk9ALNpPQCyiT0AtXk9ALRxPQC3aT0AtnE9AKMhPAC/TQCAw00AgMdNAIDLTQCApi08AKUtPADPTQCAq0E8AKp5PADTTQCA100AgK+5PACusTwArbk8AKxZPADbTQCA300AgLN9AwDjTQCAtdkDAOdNAIDrTQCAttEDAO9NAIDzTQCAu8UDALrFAwC9uQMAvLUDAL+tAwC+sQMA900AgPtNAID/TQCA71wDAIAVAACBHQAAgjEAAO+MPgCE7AQA4fw+AANOAIDjHD4AC04AgOGUAQAPTgCA4yAAAKP1AwATTgCAh+gEAIZsBAAXTgCAplkDAKVRAwAbTgCAq00DAKpNAwAfTgCAI04AgK8lAwCuOQMArTEDAKw9AwCXTQCAB04AgCdOAIArTgCAL04AgDNOAIA3TgCAO04AgKhxBgCpTQYAqo0GAKuFBgCsnQYArYUGAK6NBgCvhQYAsP0GALFBBwCyQQcAs0EHALRBBwC1SQcAtnEHALdxBwC4IQcAuSEHALolBwC7OQcAvCkHAL0VBwC+HQcAv/0HALMlBgA/TgCAQ04AgEdOAIBLTgCAtiUGALU1BgBPTgCAu6UHALoZBgBTTgCAV04AgL+tBwC+pQcAvbUHALy1BwBbTgCAo2EGAF9OAIBjTgCApmEGAGdOAIBrTgCApXEGAKpdBgCr4QcAb04AgHNOAICu4QcAr+kHAKzxBwCt8QcAqLEGAKm9BgCqzQYAq90GAKzNBgCt/QYArvUGAK8VAQCA+QEAgc0BAILFAQC+ZAIAhpAAAIcAAQB7TgCAf04AgLjRAQC52QEAuuEBALvhAQC8kQEAvZ0BAL6VAQC/iQEAsG0BALF1AQCyfQEAs3UBALRtAQC18QEAtvEBALfxAQCzRQYAd04AgINOAICHTgCAi04AgLZ9BgC1RQYAj04AgLuxAQC6qQEAk04AgJdOAIC/NQEAvqkBAL2hAQC8qQEAm04AgKMBBgCfTgCAo04AgKY5BgCnTgCAq04AgKUBBgCq7QEAq/UBAK9OAICzTgCAru0BAK9xAQCs7QEAreUBAOEoAQC3TgCA41ACALtOAIC/TgCAw04AgMdOAIDLTgCAz04AgNNOAIDXTgCA204AgIFxAACAGQAA75wCAIJ5AADfTgCA404AgITIAgCzxQMA604AgLXFAwC2xQMAvhADAIbADACHRAwAuqkDALulAwC8vQMAvaEDAL6hAwC/lQMArhEGAK8ZBgCsAQYArQEGAKqlBgCrEQYAqEU5AKlxOQDvTgCA804AgPdOAID7TgCA/04AgANPAIAHTwCAC08AgL7tBwC/TQcAvNEHAL3lBwC63QcAu8EHALg1BgC51QcAtjkGALcNBgC0JQYAtTkGALIxBgCzPQYAsFEGALFRBgCoOQIAqTkCAKqBAgCrgQIArIECAK2JAgCusQIAr7ECAIRsDQAPTwCAvmANABNPAIAXTwCAG08AgB9PAIAjTwCAuE0BALlVAQC6XQEAu1UBALxNAQC9dQEAvn0BAL91AQCwoQIAsa0CALKlAgCzuQIAtKkCALWdAgC2lQIAt3kBAOFUBgDh1AcA4zgGAOOwBwAnTwCAK08AgC9PAIAzTwCAhOQMADdPAIA7TwCAP08AgENPAIBHTwCA72wAAO/kBwCjSQIAS08AgE9PAIBTTwCAW08AgKZJAgClSQIAX08AgKspAgCqJQIAhkgMAIfcDACvGQIAri0CAK0tAgCsMQIAqFEOAKmlDgCqrQ4Aq6UOAKy9DgCtpQ4Arq0OAK+lDgCA5Q8Age0PAILlDwBXTwCAY08AgGdPAIBrTwCAb08AgLjVDwC53Q8AutUPALvpDwC8+Q8AvfkPAL7pDwC/6Q8AsN0OALFBDwCyRQ8As10PALRFDwC1TQ8AtkUPALftDwCzJQ4Ac08AgHdPAIB7TwCAf08AgLYlDgC1NQ4Ag08AgLuFDwC6GQ4Ah08AgItPAIC/iQ8AvoEPAL2JDwC8kQ8Aj08AgKNhDgCTTwCAl08AgKZhDgCbTwCAn08AgKVxDgCqXQ4Aq8EPAKNPAICnTwCArsUPAK/NDwCs1Q8Arc0PAKjRDgCp2Q4AqjkBAKs5AQCsKQEArSkBAK6dAQCvlQEAq08AgK9PAICzTwCAt08AgIANAACBtQAAgr0AALtPAIC4lQEAuZ0BALqhAQC7oQEAvHEAAL1xAAC+cQAAv3EAALDtAQCx9QEAsvUBALPFAQC03QEAtbUBALaxAQC3sQEAv08AgMNPAICzuQEAvsACALWpAQDHTwCAy08AgLahAQCGgAEAh8QBALs5AQC6IQEAvRkBALwpAQC/eQEAvhEBAKPxAQDPTwCA504AgNNPAIDXTwCApukBAKXhAQDbTwCAq3EBAKppAQDfTwCA408AgK8xAQCuWQEArVEBAKxhAQDnTwCA608AgO9PAIDzTwCA4agBAPdPAIDjQAIA+08AgL8oFQD/TwCA73QCAANQAIAHUACAC1AAgA9QAIATUACAF1AAgON0DwCEiAMA4TQOABtQAIAfUACAI1AAgCdQAICADQAAgRUAAIIRAAArUACAL1AAgO+kDwAzUACAO1AAgKgZAwCpQQMAqkUDAKtdAwCsTQMArX0DAK51AwCvnQAAhaQVAL58AwCGCAQAhxwDAD9QAIBDUACAR1AAgEtQAIC49QAAuf0AALr1AAC7jQAAvIEAAL2BAAC+gQAAv4EAALDlAACx7QAAsuUAALP5AAC07QAAtdEAALbVAAC3zQAAT1AAgFNQAIBXUACAs8ECAFtQAIC1yQIAtvECAF9QAIBjUACAZ1AAgLotAQC7JQEAvD0BAL0hAQC+JQEAvxkBAKapAgCESAIAa1AAgKWRAgBvUACAo5kCAHNQAIB3UACArn0BAK9BAQCsZQEArXkBAKp1AQCrfQEAe1AAgH9QAICDUACAh1AAgItQAICPUACA7+QAAJNQAICXUACAm1AAgOMQDgCfUACA4VgOAKNQAICALQAAgREAAIIVAAC+sAUAs3UBAKtQAICHFAUAhmwEAK9QAIC21QAAtWUBALNQAIC7/QAAuvUAALdQAIC7UACAv6EAAL69AAC93QAAvN0AAKh9BgCptQYAqr0GAKu1BgCsrQYArRUHAK4dBwCvFQcAp1AAgL9QAIDDUACAx1AAgMtQAIDPUACA01AAgNdQAIC4OQcAuTkHALrJBwC7yQcAvNkHAL3ZBwC+zQcAv8UHALBxBwCxeQcAskkHALNJBwC0OQcAtSUHALYhBwC3IQcAozUGANtQAIDfUACA41AAgOdQAICmlQcApSUGAOtQAICrvQcAqrUHAO9QAIDzUACAr+EHAK79BwCtnQcArJ0HAPdQAID7UACA/1AAgANRAIAHUQCAgj0AAIE9AACAPQAAC1EAgA9RAIATUQCAhKADAL6kAwAXUQCAhvgAAIfgAACoxQYAqdUGAKrVBgCr5QYArP0GAK0xAQCuMQEArzEBABtRAIAfUQCAI1EAgCdRAIArUQCAL1EAgDNRAIA3UQCAuN0BALntAQC65QEAu40BALyVAQC9nQEAvpUBAL+NAQCwUQEAsVEBALJRAQCzUQEAtPUBALX9AQC29QEAt+0BALNdBgA7UQCAP1EAgENRAIBHUQCAtrEBALV1BgBLUQCAu5UBALqVAQBPUQCAU1EAgL85AQC+MQEAvYUBALyFAQClLQYAV1EAgFtRAICm6QEAX1EAgGNRAICjBQYAZ1EAgK3dAQCs3QEAr2EBAK5pAQBrUQCAN1AAgKvNAQCqzQEAb1EAgHNRAICExAMAvwD0AHdRAICCPQAAgT0AAIA9AAB7UQCAf1EAgINRAIC+YAMAi1EAgI9RAICTUQCAl1EAgIbgHACHAAMA7wwHAJtRAICfUQCAo1EAgKdRAICrUQCAr1EAgLNRAIC3UQCAu1EAgOHABgC/UQCA4ywHAMNRAIDHUQCAy1EAgM9RAIDTUQCA11EAgNtRAIDfUQCA41EAgKiBAwCpgQMAqoEDAKuBAwCsgQMArYEDAK6BAwCvgQMAsEUDALFNAwCyRQMAs10DALRNAwC1fQMAtnUDALcZAwC4KQMAuTUDALo9AwC7MQMAvAEDAL31AAC+/QAAv+0AALMpAgDnUQCA61EAgO9RAIDzUQCAtiECALUpAgCEUB0Au6kCALqhAgD7UQCA/1EAgL+ZAgC+qQIAvakCALyxAgCBTQAAgE0AAO+cAwCCXQAAhvAcAId4HQC+EB0AA1IAgAdSAIALUgCAD1IAgBNSAIDhkAEAF1IAgONgAwAbUgCAH1IAgCNSAIAnUgCAK1IAgC9SAIAzUgCAN1IAgO+UAQCE7BwA4XAGADtSAIDjUAEAP1IAgENSAIBHUgCAS1IAgKPpAgBPUgCAU1IAgFdSAIBbUgCApuECAKXpAgBfUgCAq2kCAKphAgBjUgCAvqgcAK9ZAgCuaQIArWkCAKxxAgCoMR4AqTEeAKoxHgCrMR4ArF0eAK1FHgCuTR4Ar0UeAPdRAICCzR8AgfUfAID9HwBnUgCAa1IAgIYcAACH+AMAuMUeALnNHgC6xR4Au90eALzFHgC9zR4AvsUeAL9ZHwCwPR4AsQUeALINHgCzBR4AtB0eALUBHgC2BR4At/0eALO5HgBvUgCAc1IAgHdSAIB7UgCAtsUeALXVHgB/UgCAu8EeALr5HgCDUgCAh1IAgL/FHgC+2R4AvdEeALzZHgCLUgCAo/0eAI9SAICTUgCApoEeAJdSAICbUgCApZEeAKq9HgCrhR4An1IAgKNSAICunR4Ar4EeAKydHgCtlR4AqCkeAKkpHgCqVR4Aq20eAKx1HgCtfR4ArnUeAK9pHgCnUgCAq1IAgK9SAICzUgCAt1IAgLtSAIC/UgCAw1IAgLjpHgC59R4Auv0eALv1HgC87R4AvZEeAL6RHgC/kR4AsB0eALHlHgCy7R4As+UeALT9HgC15R4Atu0eALflHgCz3R4Ax1IAgMtSAIDPUgCA01IAgLb9HgC1/R4AhFgBALshHgC62R4AvigAANtSAIC/IR4AvjkeAL0xHgC8OR4AgU0AAIBNAACjlR4Agl0AAKW1HgDXUgCA31IAgKa1HgCHUQCA41IAgKtpHgCqkR4ArXkeAKxxHgCvaR4ArnEeAIYABACHRAMAs4ECAOdSAIC1gQIA61IAgO9SAIC2gQIAiAAAAPNSAIC74QIAuu0CAL3lAgC8+QIAv9ECAL7lAgD3UgCA+1IAgIREAwC+jAMA4UgCAP9SAIDjAAIA7/wfAANTAIDhPB4A79wCAONgHwAHUwCAC1MAgA9TAIATUwCAqQUCAKixAgCrBQIAqgUCAK0NAgCsBQIArzUCAK41AgCEbAUAF1MAgBtTAIAfUwCAI1MAgCdTAIArUwCAL1MAgLnpAwC44QMAu/kDALrhAwC96QMAvOEDAL9dAwC+4QMAsSkCALAlAgCzPQIAsiECALUZAgC0LQIAt9kDALYRAgAzUwCAN1MAgDtTAICjhQMAP1MAgKWFAwCmhQMAQ1MAgEtTAIBPUwCAqukDAKvlAwCs/QMAreEDAK7hAwCv1QMAgEkAAIFVAACCVQAAo6kCAL6YBAClQQEApkEBAFNTAICG4AUAh+AFAKotAQCrOQEArBEBAK0FAQCuDQEArwUBAFdTAIBbUwCAX1MAgO/cAABjUwCAZ1MAgGtTAIDviB4AhCwHAOHsHgBvUwCA4xweAHNTAIDhlAEAd1MAgOMwAACzJQIAhWDmAHtTAIB/UwCAg1MAgLbNAQC1zQEAh1MAgLu1AQC6oQEAi1MAgI9TAIC/iQEAvoEBAL2JAQC8nQEAR1MAgJNTAICXUwCAm1MAgJ9TAICjUwCAp1MAgKtTAICoAQcAqQEHAKp1BwCrrQcArLUHAK29BwCuqQcAr6kHALDZBwCx7QcAsvkHALP1BwC0mQcAtZkHALaJBwC3gQcAuIkHALmJBwC6bQAAu2UAALx9AAC9ZQAAvm0AAL9lAACBCQAAgJkAAK9TAICCHQAAs1MAgLdTAIC7UwCAv1MAgKgNBQCpfQUAqk0FAKuhBgCspQYAra0GAK6dBgCv/QYAsIUGALGRBgCyqQYAs70GALSlBgC1rQYAtqUGALd5BgC4SQYAuUkGALpZBgC7WQYAvEkGAL1JBgC++QcAv/kHALNdBgDDUwCAhigCAIcsAQDHUwCAtp0GALWdBgDLUwCAu4kGALq9BgDPUwCA01MAgL/9BgC+/QYAvYEGALyNBgDXUwCAoxkGANtTAIDfUwCAptkGAONTAIDnUwCApdkGAKr5BgCrzQYA61MAgO9TAICuuQYAr7kGAKzJBgCtxQYAqBkBAKkZAQCqjQAAq50AAKyNAACtvQAArrUAAK/dAADzUwCA91MAgPtTAID/UwCAA1QAgAdUAIALVACAD1QAgLhpAAC5aQAAunkAALt5AAC8aQAAvWkAAL7dAwC/1QMAsKkAALGpAACyvQAAs7UAALSZAAC1mQAAtlkAALdZAAC+LAIAE1QAgBdUAIAbVACAH1QAgCNUAIArVACAL1QAgIAtAACBNQAAgj0AADNUAICGkAwAh+gCADdUAIA7VACAs0UDAD9UAIBDVACAR1QAgEtUAIC2fQMAtUUDAE9UAIC7LQMAui0DAFNUAIBXVACAvx0DAL4dAwC9IQMAvCkDAKvNAwCqzQMAW1QAgF9UAICv/QMArv0DAK3BAwCsyQMAo6UDAGNUAIBnVACAa1QAgG9UAICmnQMApaUDAHNUAIB3VACAe1QAgH9UAICDVACAh1QAgII9AACBPQAAgD0AAItUAICPVACAk1QAgIRgAwCG0AwAhzADAJtUAICfVACAvkQCAKNUAICnVACAq1QAgOEAAACvVACA46gGALNUAICE7AwAt1QAgO/QAwC7VACAv1QAgMNUAIDHVACAy1QAgLNtAQDPVACA01QAgNdUAIDbVACAthEBALVlAQDfVACAuz0BALo1AQDjVACA51QAgL/9AQC+/QEAvRUBALwVAQDrVACA4fwGAO9UAIDjPAcA81QAgPdUAID7VACA/1QAgANVAIC+bAwAC1UAgA9VAIATVQCAF1UAgBtVAIDvFAYAgV0AAIBdAACj5QEAgm0AAKXtAQAfVQCAI1UAgKaZAQCHqAwAhuQMAKu1AQCqvQEArZ0BAKydAQCvdQEArnUBAKgZDgCpGQ4AqiUOAKs1DgCsLQ4ArVEOAK5RDgCvUQ4Al1QAgAdVAIAnVQCAK1UAgC9VAIAzVQCAN1UAgDtVAIC47Q4AufUOALr1DgC7jQ4AvJUOAL2dDgC+lQ4Av40OALAxDgCxOQ4AsgEOALMBDgC0+Q4AtfkOALbdDgC31Q4AqHkOAKl5DgCqjQ8Aq4UPAKydDwCtgQ8AroUPAK+5DwA/VQCAQ1UAgEdVAIBLVQCAT1UAgFNVAIBXVQCAW1UAgLiRDwC5mQ8AuqEPALuhDwC8UQ8AvV0PAL5JDwC/SQ8AsM0PALHVDwCy3Q8As9UPALTNDwC1sQ8AtrEPALexDwCzBQ4AX1UAgGNVAIBnVQCAa1UAgLYBDgC1FQ4Ab1UAgLsRDgC6CQ4Ac1UAgISgAQC/dQ4AvgkOAL0BDgC8CQ4AgmkAAKNBDgCAWQAAgVEAAKZFDgC+WAEAd1UAgKVRDgCqTQ4Aq1UOAIbIAACHrAEArk0OAK8xDgCsTQ4ArUUOAHtVAIB/VQCAg1UAgIdVAICLVQCAj1UAgCdUAICTVQCAqAkOAKkJDgCqGQ4AqxkOAKwJDgCtYQ4ArmEOAK+VAQCw7QEAsfUBALL9AQCz9QEAtO0BALV1AQC2fQEAt3UBALhNAQC5VQEAul0BALtVAQC8TQEAvfEAAL7xAAC/8QAAl1UAgJtVAICfVQCAo1UAgKdVAIDj6A4Aq1UAgOE0DgC+AAQA79wPAK9VAICzVQCAt1UAgLtVAIC/VQCAw1UAgLPxDQDHVQCAy1UAgM9VAIDTVQCAtoENALXhDQDXVQCAu1ECALpJAgDbVQCA31UAgL/RAgC+SQIAvUECALxJAgCjMQ0A41UAgISIAwDrVQCA71UAgKZBDQClIQ0A81UAgKuRAgCqiQIA91UAgPtVAICvEQIArokCAK2BAgCsiQIAgKkAAIGpAACCTQAA/1UAgOFkEgDjTAIA4wgLAOGsAQADVgCA7zwCAO8YFgAHVgCAhlAGAIdIAwALVgCAD1YAgKiBAgCpgQIAqoECAKuBAgCsgQIArYECAK6FAgCvHQEAE1YAgBdWAIAbVgCAH1YAgCNWAIAnVgCAK1YAgIS4BQC4dQEAuX0BALp1AQC7CQEAvBkBAL0ZAQC+CQEAvwEBALBlAQCxbQEAsmUBALN9AQC0aQEAtV0BALZVAQC3TQEAL1YAgDNWAIA3VgCAO1YAgD9WAIBDVgCA7zQAAO/ADgDhXA4A4UwPAOOUAADjnA4AR1YAgIJlAACBfQAAgH0AAEtWAIBPVgCAvsQHALNFAgBTVgCAtUUCALZNAgBbVgCAhkAGAIeQBAC67QEAu+UBALz9AQC95QEAvuEBAL/VAQCflQgAngUIAJ3dDQCcPQwAmzEMAJr1DQCZ7RAAmD0QAJfVEQCWsRUAlQUUAJTlFQCTtRkAkjEYAJE5GACQDRwAj2EcAOdVAICz1QYAX1YAgLX9BgBXVgCAY1YAgLaRBgBnVgCAa1YAgLuVBgC6lQYAvVUHALxVBwC/VQcAvlUHAG9WAIBzVgCAqo0GAKuFBgCsnQYArYUGAK6BBgCvtQYAhKgAAHdWAIB7VgCAoyUFAH9WAIClJQUApi0FAINWAICHVgCAi1YAgI9WAICTVgCAl1YAgJtWAICfVgCAo1YAgKdWAICrVgCAr1YAgLNWAICjqQUAotEEAKHZBACgZQUAgiEdAIM1HQC3VgCAu1YAgIaVGACH3RQAhBkZAIUZGQCKDRUAi7EUAL9WAIDDVgCAjsURAI/VDACMzRAAjR0RAJJhDQCTdQ0AvkwAAMtWAICWxQkAl80EAJSNDACVXQkAmkEFAJtBBQCGyP8Ah0wAAIFZAACAeQAAnCEEAIJRAAChxQEAz1YAgKMB/ACi2QEApRX9AKS1/QCnufkApgH4AKkJ+AColfkAqwX1AKqt9QCtsfEArAHwAK8d8ACurfEAseHtALAB7ACzAegAsv3sALVd6QC09ekA01YAgNdWAIDbVgCA31YAgONWAIDnVgCA61YAgO9WAIDzVgCA91YAgKiNBACplQQAqpUEAKulBACsvQQArdkEAK75BACv8QQAhGz8APtWAID/VgCAA1cAgAdXAIALVwCAD1cAgBNXAIC4eQUAucUFALrNBQC7xQUAvN0FAL3FBQC+zQUAv+0FALCZBACxmQQAskkFALNJBQC0WQUAtVkFALZJBQC3SQUAox0EAL7M/AAXVwCAG1cAgB9XAICmWQQApTUEACNXAICrXQQAql0EACdXAIArVwCAr50FAK6dBQCtnQUArJ0FAC9XAICznQIAM1cAgDtXAIC2UQIAP1cAgENXAIC1uQIAukkCALtVAgCGSP0Ah8D8AL41AgC/PQIAvEUCAL09AgCo3QQAqUkDAKpRAwCrbQMArHUDAK2VAwCunQMAr7kDAICNAQCB5QEAguEBAEdXAIBLVwCAT1cAgFNXAIBXVwCAuJUDALmdAwC6lQMAu60DALy1AwC9vQMAvrUDAL9VAgCwyQMAsdUDALLVAwCzrQMAtLUDALW9AwC2tQMAt60DAFtXAIBfVwCAo9EDAGNXAICl9QMAZ1cAgGtXAICmHQMAb1cAgHNXAICrGQMAqgUDAK1xAwCsCQMAr3EDAK55AwDhKAcAd1cAgOPkBgB7VwCA4SgGAH9XAIDjaAEAg1cAgIdXAICLVwCA71gAAI9XAICTVwCAl1cAgO/IBgCbVwCAqE39AKmB/QCq0f0Aq9H9AKzx/QCt8f0ArvH9AK/x/QA3VwCAghEAAIEZAACA0f8An1cAgKNXAICEdAMAvnQDALh1/gC5ff4AunX+ALvF/gC83f4AvcX+AL7F/gC/9f4AsJH9ALGR/QCykf0As5H9ALRV/gC1Xf4AtlX+ALdN/gCzWf0Ap1cAgIasAACHRAMAq1cAgLZx/QC1ef0Ar1cAgLtV/QC6Vf0As1cAgLdXAIC/mf4AvpH+AL1F/QC8Rf0Au1cAgKMd/QC/VwCAw1cAgKY1/QDHVwCAy1cAgKU9/QCqEf0AqxH9AM9XAIDTVwCArtX+AK/d/gCsAf0ArQH9AKjN/wCp0f8AqtH/AKsh/gCsIf4ArSH+AK4h/gCvIf4A11cAgNtXAIDfVwCA41cAgOdXAIDrVwCA71cAgPNXAIC4jf4AuZH+ALqV/gC7rf4AvLX+AL25/gC+qf4Av6n+ALDh/gCx4f4AsuX+ALP5/gC06f4AtdX+ALbd/gC3uf4As1n/APdXAIDHVgCA+1cAgP9XAIC2of4Atan+AANYAIC7Jf4AuiX+AAdYAIALWACAvxH+AL4t/gC9Lf4AvDH+AIIZAACjHf8AgGUAAIEZAACm5f4AD1gAgBNYAICl7f4AqmH+AKth/gCEZAEAviAAAK5p/gCvVf4ArHX+AK1p/gAbWACA4zT+AB9YAIDhfP0AhrAEAIcIAwAjWACAJ1gAgCtYAIAvWACAhCQDAIQkBAAzWACA70j+ADdYAIA7WACAs+kCAD9YAIC+RAQAvkAFAENYAIC2nQIAtZkCAEdYAIC7iQIAur0CAEtYAIBPWACAv1kDAL5RAwC9WQMAvJECAKkdAgCoFQIAqyUCAKolAgCtWQIArFUCAK9NAgCuUQIAvmQGAFNYAIBXWACAW1gAgF9YAIBjWACAZ1gAgGtYAIC5+QMAuPEDALtNAwC68QMAvUEDALxZAwC/cQMAvkEDALEJAgCwPQIAs8kDALIBAgC12QMAtNEDALfJAwC20QMA4ZABAG9YAIDj8AAAc1gAgHdYAICCPQAAgT0AAIA9AAB7WACAf1gAgINYAICLWACAj1gAgJNYAIDvLAAAl1gAgKPpAwCbWACAhugEAIdgBQCfWACApp0DAKWZAwCjWACAq4kDAKq9AwCnWACAq1gAgK9ZAgCuUQIArVkCAKyRAwCvWACAs1gAgLdYAIC7WACAv1gAgMNYAIDHWACA71gBAISgBADhVP8Ay1gAgOOEAQDPWACA01gAgNdYAIDbWACAs9kBAN9YAICFzBkA41gAgOdYAIC28QEAtfkBAOtYAIC7pQEAutkBAO9YAIDzWACAv50BAL6dAQC9pQEAvK0BAKgBBgCpDQYAqhEGAKsRBgCsMQYArTEGAK4pBgCvJQYAh1gAgILJBwCBwQcAgPEHAPdYAID7WACAhhwAAIf8AwC47QYAufUGALr9BgC79QYAvO0GAL1RBwC+VQcAv00HALBdBgCxIQYAsjkGALMxBgC0GQYAtRkGALbdBgC31QYAo5kGAP9YAIADWQCAB1kAgAtZAICmsQYApbkGAA9ZAICr5QYAqpkGABNZAIAXWQCAr90GAK7dBgCt5QYArO0GABtZAICz8QcAH1kAgCNZAIC2gQcAJ1kAgCtZAIC1mQcAuo0HALtlBwAvWQCAM1kAgL59BwC/ZQcAvH0HAL11BwCoLQYAqTUGAKo9BgCrMQYArFUGAK1FBgCuRQYAr3UGADdZAIA7WQCAP1kAgENZAIBHWQCAS1kAgE9ZAIBTWQCAuOkGALn1BgC6/QYAu/UGALztBgC9kQYAvpUGAL+NBgCwDQYAseUGALLtBgCz5QYAtP0GALXlBgC27QYAt+UGAKO1BgBXWQCAW1kAgF9ZAIBjWQCApsUGAKXdBgAXWACAqyEGAKrJBgBnWQCAa1kAgK8hBgCuOQYArTEGAKw5BgCASQAAgUkAAIJZAACzRQEAb1kAgLVFAQC2RQEAc1kAgIZAAACHZAAAuikBALslAQC8PQEAvSEBAL4hAQC/FQEAd1kAgHtZAICEBAMAvgAMAOMoBgDv4AIA4RAGAH9ZAIDvkAYA4zwCAINZAIDh1AEAh1kAgItZAICPWQCAk1kAgJdZAICbWQCAo8ECAJ9ZAIClwQIAo1kAgKdZAICmwQIAq1kAgK9ZAICroQIAqq0CAK2lAgCsuQIAr5ECAK6lAgCpBQIAqLECAKsFAgCqBQIArQ0CAKwFAgCvNQIArjUCAISoDACzWQCAt1kAgLtZAIC/WQCAw1kAgMdZAIDLWQCAuekDALjhAwC7+QMAuuEDAL3pAwC84QMAv10DAL7hAwCxKQIAsCUCALM9AgCyIQIAtRkCALQtAgC32QMAthECAKitAgCp1QIAqtUCAKsNAQCsFQEArQkBAK4xAQCvLQEAz1kAgNNZAIDbWQCA31kAgONZAIDnWQCA61kAgO9ZAIC4IQEAuSEBALrtAQC75QEAvP0BAL3lAQC+7QEAv+UBALBVAQCxXQEAslUBALMtAQC0NQEAtTkBALYtAQC3JQEAgD0BAIGlAACCrQAA79QHAPNZAID3WQCA+1kAgO8oBwC+LAwA4fQGAP9ZAIDjkAcAA1oAgOGUAQAHWgCA4wwGALMdAgALWgCAh0QNAIZMDQAPWgCAtskBALXdAQATWgCAu9kBALrRAQAXWgCAG1oAgL+9AQC+sQEAvbkBALzBAQDXWQCAH1oAgCNaAIAnWgCAK1oAgC9aAIAzWgCAN1oAgKgJDwCpCQ8AqhkPAKsZDwCsCQ8ArQkPAK6pDwCvqQ8AsNkPALHtDwCy+Q8As/UPALSVDwC1hQ8AtoUPALe1DwC4jQ8AuWEAALphAAC7YQAAvGEAAL1hAAC+YQAAv2EAAKNdDQCCLQAAgRUAAIAdAAA7WgCApokOAKWdDgA/WgCAq5kOAKqRDgBDWgCAR1oAgK/9DgCu8Q4ArfkOAKyBDgBLWgCAs/UPAIboAwCHvAMAtu0PAE9aAIBTWgCAteUPALp5DwC7TQ8AV1oAgFtaAIC+NQ8AvyUPALxJDwC9RQ8AozEOAF9aAIBjWgCAZ1oAgGtaAICmKQ4ApSEOAG9aAICriQ4Aqr0OAHNaAIB3WgCAr+EOAK7xDgCtgQ4ArI0OAHtaAIB/WgCAg1oAgIdaAICLWgCAj1oAgJNaAICXWgCAm1oAgJ9aAICjWgCAp1oAgIANAACB1QAAgt0AAKtaAICoQQEAqVEBAKpRAQCrZQEArH0BAK2RAACukQAAr5EAAK9aAICzWgCAhGQBAL5kAQCGkAEAh4QAALtaAIC/WgCAuJEAALmRAAC6kQAAu5EAALyxAAC9sQAAvrEAAL+xAACw8QAAsfkAALLBAACzwQAAtLEAALWxAAC2sQAAt7EAALPZAgDDWgCAvnADAL5EBADHWgCAthEDALX1AgDLWgCAuz0DALo1AwDPWgCA01oAgL91AwC+dQMAvRUDALwVAwDXWgCAo50CANtaAIDfWgCAplUDAONaAIDnWgCApbECAKpxAwCreQMA61oAgO9aAICuMQMArzEDAKxRAwCtUQMAqDkDAKk5AwCqjQAAq50AAKyNAACtvQAArrUAAK/dAADzWgCA91oAgPtaAID/WgCAA1sAgAdbAIALWwCAD1sAgLhpAAC5aQAAunkAALt5AAC8aQAAvWkAAL7ZAQC/2QEAsKkAALGpAACyvQAAs7UAALSZAAC1mQAAtlkAALdZAAATWwCAF1sAgBtbAIAfWwCA70QAACNbAICGmAUAh+QCAOOYAACEqAIA4fgBACtbAICAOQAAgTkAAIItAAAvWwCAs0UBADNbAIA3WwCAO1sAgD9bAIC2fQEAtUUBAENbAIC7LQEAui0BAEdbAIBLWwCAvx0BAL4dAQC9IQEAvCkBAE9bAIDhUA4AU1sAgOM8DwBXWwCAW1sAgF9bAIBjWwCAZ1sAgGtbAIDjAAAAb1sAgHNbAIB3WwCAhPQFAO/kDgCuqQEAr6kBAKydAQCtlQEAqpkBAKuZAQB7WwCAf1sAgKbJAQCDWwCAh1sAgKXxAQCC/QcAo/EBAID9BwCB9QcAJ1sAgItbAICPWwCAk1sAgJdbAICbWwCAhrgDAIeQAwCoDQcAqRkHAKptBwCrZQcArH0HAK1lBwCuZQcAr1UHALAtBwCxxQcAssEHALPdBwC0xQcAtc0HALbFBwC3/QcAuMUHALnJBwC62QcAu9kHALypBwC9qQcAvp0HAL+VBwCzxQcAn1sAgKNbAICnWwCAq1sAgLbFBwC11QcAr1sAgLshBwC6yQcAs1sAgLdbAIC/KQcAviEHAL0pBwC8NQcAu1sAgKOBBwC/WwCAw1sAgKaBBwDHWwCAy1sAgKWRBwCqjQcAq2UHAM9bAIDTWwCArmUHAK9tBwCscQcArW0HAKgVAQCpgQEAqoEBAKuBAQCsgQEArYkBAK6xAQCvsQEA11sAgNtbAIDfWwCA41sAgOdbAIDrWwCA71sAgPNbAIC4ZQAAuW0AALplAAC7fQAAvGUAAL1tAAC+ZQAAv90AALChAQCxrQEAsqUBALO5AQC0qQEAtZ0BALaVAQC3XQAA91sAgIIdAACBHQAAgB0AAPtbAID/WwCAA1wAgL5YAQCErAIAB1wAgIcIAQCGjAEAC1wAgLdaAIAPXACAE1wAgLNJAQAXXACAG1wAgB9cAIAjXACAtkkBALVJAQAnXACAuykBALolAQArXACAL1wAgL8ZAQC+LQEAvS0BALwxAQC+2AMAM1wAgO/4BgA3XACAO1wAgD9cAIDv4AIAQ1wAgOGUAQBHXACA43QCAEtcAIDhmAUAT1wAgOMMBwBTXACAV1wAgFtcAICjwQIAhIwDAKXBAgBfXACAY1wAgKbBAgBnXACAa1wAgKuhAgCqrQIAraUCAKy5AgCvkQIArqUCAKgxAwCpPQMAqjUDAKtJAwCsWQMArVkDAK5JAwCvQQMAgMUAAIEJAACCGQAAb1wAgHNcAIB7XACAh2wDAIYcHAC47QAAufEAALr1AAC7jQAAvJUAAL2BAAC+gQAAv70AALAJAwCxCQMAsu0AALPhAAC04QAAteEAALblAAC32QAAf1wAgINcAICHXACAs7ECAItcAIC13QIAttUCAI9cAICTXACAl1wAgLrBAgC7wQIAvDUBAL05AQC+KQEAvykBAKaNAgCbXACAn1wAgKWFAgCjXACAo+kCAKdcAICrXACArnEBAK9xAQCsbQEArWEBAKqZAgCrmQIAr1wAgLNcAIC3XACA4YQGALtcAIDjJAYAv1wAgOGUAQDDXACA4ywAAL7oHQDHXACAy1wAgO/IAACE/B0AvvAcAM9cAIDvSAcA01wAgNdcAIDbXACA31wAgIEdAACAHQAA41wAgIIFAACGQBwAh8QcAOtcAIDvXACA81wAgPdcAID7XACA/1wAgKi1HgCpBR8Aqg0fAKsFHwCsAR8ArQkfAK45HwCvOR8A51wAgANdAIAHXQCAC10AgA9dAIATXQCAF10AgBtdAIC4yR8AudUfALrRHwC76R8AvPkfAL3tHwC+mR8Av5kfALAlHwCxLR8AsjkfALM1HwC0LR8AtQ0fALYFHwC3/R8As4UfAB9dAIAjXQCAJ10AgCtdAIC2iR8AtYkfAC9dAIC76R8AuuEfADNdAIA3XQCAv8kfAL7pHwC94R8AvO0fADtdAICjwR8AP10AgENdAICmzR8AR10AgEtdAIClzR8AqqUfAKutHwBPXQCAU10AgK6tHwCvjR8ArKkfAK2lHwCo6R4AqekeAKr5HgCr+R4ArOkeAK3pHgCuPQEArzUBAID5AQCBzQEAgsUBAIRgAgBXXQCAW10AgIdoAQCGnAAAuNEBALnZAQC64QEAu+EBALyRAQC9nQEAvpUBAL+JAQCwTQEAsVUBALJdAQCzVQEAtE0BALXxAQC28QEAt/EBALNxHgBfXQCAY10AgGddAIBrXQCAtmkeALVhHgBvXQCAu5EBALqJAQBzXQCAd10AgL81AQC+iQEAvYEBALyJAQB7XQCAd1wAgKM5HgB/XQCApSkeAINdAICHXQCApiEeAItdAICPXQCAq9kBAKrBAQCtyQEArMEBAK99AQCuwQEAk10AgJddAICbXQCAn10AgKNdAICnXQCAq10AgK9dAICzXQCAt10AgLtdAIC/XQCAw10AgMtdAIDPXQCAvnADAOHkHgCESAIA4+gfAIQABACAeQAAgXkAAIJpAADTXQCAhsAEAIdEAwDXXQCA210AgN9dAIDjXQCA7yAfAOddAIDrXQCA710AgPNdAIDvSAIA910AgPtdAID/XQCAA14AgL7oBAAHXgCAC14AgA9eAIATXgCA4ZABABdeAIDj6AIAs0kDABteAIAfXgCAI14AgCdeAIC2SQMAtUkDACteAIC7LQMAuiUDAC9eAIAzXgCAvxUDAL4VAwC9IQMAvCkDAKg1AgCpgQIAqoECAKuBAgCsgQIArYkCAK6xAgCvsQIAgP0BAIHNAQCCxQEAO14AgIaQBACHBAUAP14AgIRwBAC4SQEAuUkBALpZAQC7WQEAvEkBAL1JAQC+eQEAv3kBALChAgCxqQIAsr0CALO1AgC0kQIAtZECALZ5AQC3eQEAQ14AgEdeAIBLXgCAT14AgFNeAIBXXgCAW14AgO/QHgC+6AQA4VweAF9eAIDjkAAAY14AgGdeAIBrXgCAb14AgKNJAgBzXgCAd14AgHteAIB/XgCApkkCAKVJAgCDXgCAqy0CAKolAgCHXgCAi14AgK8VAgCuFQIArSECAKwpAgCoNQYAqT0GAKpVBgCrZQYArH0GAK1lBgCubQYAr2EGADdeAICPXgCAk14AgJdeAICADQAAgbEAAIKxAACbXgCAuOkGALnpBgC6+QYAu/UGALyVBgC9nQYAvpUGAL+NBgCw4QYAseEGALLhBgCz/QYAtOUGALXtBgC25QYAt9kGALPdBgCfXgCAo14AgKdeAICrXgCAtuUGALX1BgCvXgCAuyUGALolBgCGmAAAh6wAAL8pBgC+IQYAvSkGALw1BgCzXgCAo5kGALdeAIC7XgCApqEGAL9eAIDDXgCApbEGAKphBgCrYQYAx14AgMteAICuZQYAr20GAKxxBgCtbQYAqC0GAKk9BgCqiQYAq4kGAKyZBgCtmQYArokGAK+JBgDPXgCA014AgNdeAIDbXgCA314AgONeAIDnXgCA614AgLiNBgC5lQYAupUGALulBgC8vQYAvXEBAL5xAQC/cQEAsPkGALHNBgCy2QYAs9kGALTJBgC1yQYAtr0GALe1BgCzAQYA714AgPNeAID3XgCA+14AgLYZBgC1EQYA/14AgLsJBgC6PQYAA18AgAdfAIC/DQYAvg0GAL0NBgC8DQYAC18AgKNFBgDHXQCAD18AgKZdBgATXwCAhFgAAKVVBgCqeQYAq00GAL5oAQAXXwCArkkGAK9JBgCsSQYArUkGAIDBAwCByQMAgt0DAKPNAgAbXwCApdkCAKbNAgAfXwCAhoANAIeUAwCqxQIAqw0DAKwVAwCtHQMArhUDAK8NAwDhnBcA4xgGAOMUAwDhNAYA7xgCACNfAIAnXwCAK18AgOPQAgAvXwCA4VACADNfAIA3XwCA7ywGAO/kJQA7XwCArE0CAK1RAgCuUQIAr2UCAKgBAgCpCQIAqlkCAKtVAgCE7A0AP18AgENfAIBHXwCAvvgNAEtfAIBPXwCAU18AgLxRAwC9WQMAvmEDAL9hAwC47QMAuVEDALpRAwC7UQMAtM0DALXVAwC23QMAt9UDALAdAgCx1QMAst0DALPVAwDjyAAAV18AgOG4AQBbXwCAhFQPAF9fAIBjXwCAZ18AgKHpAgCgFQYAo6UDAKINAwDvIAAAa18AgG9fAIBzXwCAd18AgHtfAICFNCYAs40DAH9fAIC1mQMAto0DAINfAICGwA8Ah5QNALqFAwC7TQIAvFUCAL1dAgC+VQIAv00CAItfAICPXwCAk18AgJdfAICbXwCAn18AgI/d6wDvxAYAvuAPAOGMBgCjXwCA44AGAID1AACB5QAAguUAAKdfAICZbR8AmMUfAJvJGwCaeRoAnXUaAJzFGwCf+QcAnhkGAJFpFgCQsesAk20XAJLNFwCV0RMAlGkSAJdREgCWzRMAg1XkAIJB5ACHXwCAq18AgIeNHQCGkRgAhTkYAISVGQCLERwAigUcAK9fAICzXwCAj4UVAI6ZEACNORAAjJUdAJNRFACSRRQAt18AgLtfAICXYQkAlnUIAJWdCQCU+RUAm0EMAJqtDQC/XwCAw18AgMdfAIDLXwCAz18AgJzxDAChbQ0A018AgKMBBACihQAApZkEAKSRBACnGTgApsUFAKkJOACoKTgAq4k8AKoBPACtATAArB08AK8pMACunTAAseE0ALABNACzASgAsv00ALXZKAC00SgA118AgNtfAIDfXwCA418AgOdfAIDrXwCAgB0AAIEJAACC2QEA718AgKgRDwCpGQ8Aql0PAKtVDwCsTQ8ArXEPAK51DwCvbQ8A818AgPtfAICGiAAAhxABAP9fAIADYACAB2AAgAtgAIC4TQ4AuVEOALpRDgC7UQ4AvGUOAL1tDgC+ZQ4Avx0OALAdDwCxwQ8AssEPALPBDwC0xQ8Atc0PALbFDwC3eQ4As9UPAA9gAIATYACAF2AAgBtgAIC28Q8AtcUPAB9gAIC7BQ8AutkPACNgAIAnYACAvwkPAL4BDwC9FQ8AvBUPACtgAICjkQ8AL2AAgDNgAICmtQ8AN2AAgDtgAIClgQ8Aqp0PAKtBDwA/YACAQ2AAgK5FDwCvTQ8ArFEPAK1RDwCogQ0AqYENAKqBDQCrgQ0ArIENAK2BDQCusQ0Ar6ENAEdgAIBLYACAT2AAgFNgAIBXYACAgrkAAIG9AACAvQAAuDUCALk9AgC6zQIAu5UCALyNAgC9tQIAvr0CAL+1AgCwbQIAsU0CALJFAgCzJQIAtD0CALUdAgC2FQIAtw0CAFtgAIBfYACAswENAGNgAIC1AQ0Aa2AAgISUAwC2CQ0AviwEAG9gAIC7gQIAuqECAL35AgC8mQIAv9ECAL7xAgBzYACAd2AAgHtgAICjRQ0Af2AAgKVFDQCmTQ0Ag2AAgIbgBACHpAQAquUCAKvFAgCs3QIArb0CAK61AgCvlQIAqCUCAKk1AgCqPQIAqzUCAKwtAgCtkQIArpECAK+RAgCHYACAi2AAgI9gAICTYACAzAAAAJdgAICbYACAn2AAgLiZAgC5rQIAuqUCALttAQC8dQEAvX0BAL51AQC/bQEAsPECALH5AgCywQIAs8ECALSxAgC1vQIAtrUCALepAgCjYACA44QOAKdgAIDh9A4Aq2AAgK9gAICzYACAt2AAgIQgBQC7YACAv2AAgMNgAIDHYACA7+wOAMtgAIDPYACAs/UCANNgAICG6AQAh4wEAL5cBAC2UQIAteUCANtgAIC7fQIAunUCAN9gAIDjYACAvzkCAL41AgC9VQIAvFUCAKM1BQBnYACA12AAgOdgAIDrYACAppEFAKUlBQDvYACAq70FAKq1BQDzYACA92AAgK/5BQCu9QUArZUFAKyVBQCA+QcAgfkHAIKNBwCzjQYA+2AAgLWdBgC2iQYA/2AAgANhAIAHYQCAuk0HALtFBwC8XQcAvUEHAL5BBwC/QQcAC2EAgA9hAID3XwCAE2EAgBdhAIAbYQCAH2EAgCNhAICoNQYAqQEGAKppBgCraQYArHkGAK1lBgCuZQYAr50HALDlBwCx7QcAsuUHALP5BwC06QcAtekHALZZBwC3VQcAuHEHALlxBwC6cQcAu3EHALxVBwC9XQcAvlUHAL9NBwCjwQcAJ2EAgCthAIAvYQCAM2EAgKbFBwCl0QcAN2EAgKsJBgCqAQYAO2EAgD9hAICvDQYArg0GAK0NBgCsEQYAgGkAAIFpAACCBQAAQ2EAgL6YAQCEmAEAR2EAgEthAICGADwAh8QBAE9hAIBTYQCAV2EAgFthAIBfYQCAY2EAgKhdBgCpbQYAqmUGAKuBAQCsgQEArYkBAK6xAQCvsQEAZ2EAgGthAIBvYQCAc2EAgHdhAIB7YQCAf2EAgINhAIC4VQEAuV0BALpVAQC7yQAAvNkAAL3ZAAC+yQAAv8EAALCxAQCxuQEAsokBALOJAQC0cQEAtXEBALZ1AQC3bQEAs+0FAIdhAICLYQCAj2EAgJNhAIC2CQIAtQkCAJdhAIC7fQIAunUCAJthAICfYQCAv7UCAL61AgC9XQIAvF0CAL5gAgCjqQUAo2EAgKdhAICmTQIAq2EAgK9hAIClTQIAqjECAKs5AgCzYQCAhOADAK7xAgCv8QIArBkCAK0ZAgC+iDwAu2EAgKotAwCrJQMArD0DAK0lAwCuLQMAryUDAID1AACB/QAAgsEAAKPBAwC/YQCApcEDAKbBAwDDYQCAhmA8AIdUAwDHYQCAy2EAgM9hAIDjqAIA02EAgOGkAQDXYQCA71wCANthAIDfYQCA42EAgOdhAIDrYQCA72EAgPNhAIDjjAcA92EAgOE8BAD7YQCA/2EAgANiAIAHYgCAhCACAAtiAIAPYgCAE2IAgBdiAIDvbAcAG2IAgB9iAICzLQIAhEQ9ACNiAIArYgCAL2IAgLYtAgC1LQIAM2IAgLvJAgC6wQIAN2IAgDtiAIC/yQIAvsECAL3JAgC80QIA4XgHAOPAAADjOAYA4VwGAICpAACBqQAAgtEAAD9iAIBDYgCAR2IAgL6kPABLYgCAT2IAgO8cAADvkAYAU2IAgIZgPACHBD0AV2IAgLNxAQBbYgCAtRkBALYJAQBfYgCAY2IAgGdiAIC6AQEAuwEBALwBAQC9AQEAvgEBAL8BAQCohT4AqbU+AKq1PgCrxT4ArN0+AK3FPgCuwT4Ar/0+AGtiAIBvYgCAc2IAgHdiAIB7YgCAf2IAgINiAICHYgCAuFE/ALlRPwC6UT8Au1E/ALx1PwC9fT8AvnU/AL9tPwCwiT4AsYk+ALKZPgCzmT4AtIk+ALWJPgC2eT8At3U/ALdhAICjOT4Ai2IAgCdiAICmQT4Aj2IAgJNiAIClUT4Aqkk+AKtJPgCXYgCAm2IAgK5JPgCvST4ArEk+AK1JPgCASQAAgVEAAIJRAACzkT8An2IAgLW5PwC2RT8Ao2IAgIZAAACHBAMAukU/ALtdPwC8TT8AvT0/AL4pPwC/IT8AqE0+AKlVPgCqVT4Aq2U+AKx9PgCtiT4Arrk+AK+5PgCnYgCAq2IAgK9iAICzYgCAt2IAgLtiAIC/YgCAw2IAgLhhAQC5YQEAumEBALthAQC8YQEAvWEBAL5hAQC/YQEAsM0+ALHVPgCy1T4As6U+ALShPgC1qT4Atpk+ALeZPgCj3T4Ax2IAgMtiAIDPYgCA02IAgKYJPgCl9T4A12IAgKsRPgCqCT4A22IAgN9iAICvbT4ArmU+AK1xPgCsAT4A42IAgOdiAIDrYgCA72IAgPNiAID3YgCA+2IAgP9iAICAOQAAgTkAAIIFAAADYwCAvrgBAIS4AQALYwCAD2MAgKitAgCp1QIAqtUCAKstAwCsNQMArT0DAK41AwCvLQMAE2MAgBdjAIAbYwCAH2MAgCNjAIAnYwCAK2MAgC9jAIC46QMAuekDALqJAwC7iQMAvJkDAL2ZAwC+iQMAv4kDALBVAwCxXQMAslUDALPpAwC0+QMAtfkDALbpAwC34QMAs10CADNjAICGKAQAh8wDADdjAIC2vQMAtb0DADtjAIC7mQMAupEDAD9jAIBDYwCAvz0DAL49AwC9PQMAvIEDAIUAFACjGQIAR2MAgEtjAICm+QMAT2MAgFNjAICl+QMAqtUDAKvdAwBXYwCAW2MAgK55AwCveQMArMUDAK15AwDjVD4A4dw/AOHQPgDjPD4AX2MAgO8cAABjYwCAZ2MAgGtjAIDjwAAAb2MAgOHUAQDvYD4Ac2MAgHtjAIDvRD8AgGEAAIFtAACCfQAAhAAFAIbwBACHnAUAvhAFAH9jAICDYwCAh2MAgItjAICPYwCAk2MAgJdjAICbYwCAn2MAgLiJPQC5iT0Aupk9ALuRPQC8uT0Avbk9AL7RPQC/0T0AsAU+ALENPgCyBT4Asx0+ALQFPgC1DT4AtgU+ALe5PQConT4Aqa0+AKqlPgCrvT4ArKU+AK2tPgCupT4Ar30+AISsBAC+rAQAo2MAgKdjAICrYwCAr2MAgLNjAIC3YwCAqPkFAKn5BQCqKQYAqykGAKw5BgCtOQYArikGAK8pBgB3YwCAu2MAgL9jAIDDYwCAx2MAgMtjAIDPYwCA02MAgLiNBgC5kQYAupEGALulBgC8vQYAvUUHAL5BBwC/QQcAsFkGALFZBgCy7QYAs/0GALTtBgC13QYAttUGALe1BgCzoQYA12MAgNtjAIDfYwCA42MAgLa5BgC1sQYA62MAgLudBgC6nQYA52MAgAdjAIC/GQYAvikGAL0pBgC8OQYAglEAAKPlBgCAQQAAgUEAAKb9BgDvYwCA82MAgKX1BgCq2QYAq9kGAIZIAACHbAAArm0GAK9dBgCsfQYArW0GAKg5BgCpWQYAqmkGAKtpBgCseQYArXkGAK5pBgCvaQYA92MAgPtjAID/YwCAA2QAgAdkAIALZACAD2QAgBNkAIC4ZQEAuW0BALplAQC7fQEAvGUBAL1tAQC+ZQEAv9kBALAZBgCxGQYAsoEGALOBBgC0gQYAtYEGALaBBgC3gQYAs+EGABdkAIAbZACAH2QAgCNkAIC2+QYAtfEGACdkAIC73QYAut0GACtkAIAvZACAv0UGAL5FBgC9VQYAvFUGADNkAICjpQYAN2QAgDtkAICmvQYAP2QAgENkAICltQYAqpkGAKuZBgBHZACAS2QAgK4BBgCvAQYArBEGAK0RBgConQIAqdECAKrRAgCrLQMArDUDAK09AwCuNQMAry0DAE9kAIBTZACAvmQCAFtkAIBfZACAY2QAgGdkAIBrZACAuOkDALnpAwC6iQMAu4UDALydAwC9gQMAvoEDAL+1AwCwVQMAsV0DALJVAwCz6QMAtPkDALX5AwC26QMAt+EDAIBtAwCBpQAAgq0AALNVAgBvZACAtbEDALaxAwBzZACAhOACAHdkAIC6nQMAu5UDALyNAwC9MQMAvjEDAL8xAwCjGQIAe2QAgIVwaQB/ZACAg2QAgKb9AwCl/QMAh2QAgKvZAwCq0QMAhkgMAIe8AwCvfQMArn0DAK19AwCswQMAi2QAgI9kAICTZACAl2QAgO+wBgDvxAMAm2QAgJ9kAIDjfAYA45QDAOG4BwDh3AEAo2QAgKdkAICrZACAr2QAgLNkAIC3ZACAhEQCAL5YDQCADQAAgTUAAII9AAC7ZACAv2QAgMNkAICGyAwAh1wNAMtkAIDPZACA02QAgNdkAIDbZACA32QAgONkAIDnZACA62QAgO9kAIDzZACA74AGAISsDQDh7AYA92QAgONcBgD7ZACA/2QAgANlAIAHZQCAs/UBAAtlAIAPZQCAE2UAgBdlAIC2RQEAteUBABtlAIC7LQEAuiEBAB9lAIAjZQCAv/UAAL71AAC9JQEAvC0BAKgtDgCpNQ4Aqj0OAKs1DgCsLQ4ArYUOAK6FDgCvuQ4Ax2QAgCdlAIArZQCAL2UAgIAZAACBGQAAggUAADNlAIC4WQ8AuVkPALp5DwC7eQ8AvGkPAL1pDwC+GQ8AvxkPALClDgCxqQ4AsrkOALOxDgC0cQ8AtXEPALZxDwC3cQ8Apb0OAL6IAwA7ZQCAph0OADdlAIA/ZQCAo60OAENlAICtfQ4ArHUOAK+tDwCurQ8AV2QAgEdlAICrdQ4AqnkOALO5DwBLZQCAhmgAAIcMAwBPZQCAtlEPALVZDwBTZQCAu3UPALp1DwBXZQCAW2UAgL9FDwC+RQ8AvVEPALxlDwCocQ4AqXEOAKpxDgCrcQ4ArJEOAK2RDgCukQ4Ar5EOAF9lAIBjZQCAZ2UAgGtlAIBvZQCAc2UAgHdlAIB7ZQCAuIUOALmNDgC6hQ4Au50OALyNDgC9vQ4AvrUOAL95AQCw8Q4AsfEOALLxDgCzxQ4AtMEOALXBDgC2wQ4At8EOAKP5DgB/ZQCAg2UAgIdlAICLZQCAphEOAKUZDgCPZQCAqzUOAKo1DgCTZQCAl2UAgK8FDgCuBQ4ArREOAKwlDgCADQAAgRUAAIIdAACbZQCAn2UAgKNlAICElAEAvpQBAIZABwCH5AAAq2UAgK9lAICzZQCAt2UAgLtlAIC/ZQCAqIkCAKmRAgCqlQIAq7kCAKzVAgCtxQIArsUCAK/1AgDDZQCAx2UAgMtlAIDPZQCAvnwDANNlAIDXZQCA22UAgLh9AwC5wQMAusEDALvBAwC8wQMAvckDAL7xAwC/8QMAsI0CALFFAwCyTQMAs0UDALRdAwC1RQMAtk0DALdFAwCzHQIA32UAgONlAIDnZQCA62UAgLZFAgC1XQIA72UAgLuBAwC6SQIA82UAgPdlAIC/gQMAvpkDAL2RAwC8mQMA+2UAgKNZAgD/ZQCAA2YAgKYBAgAHZgCAC2YAgKUZAgCqDQIAq8UDAA9mAIATZgCArt0DAK/FAwCs3QMArdUDAIDZAQCB7QEAguUBAO+4DgAbZgCA4cQBAISYAgDj1AAAH2YAgL7sBAAjZgCA7wgAACdmAIDhxA8AK2YAgONkDgCGAAUAh2gFAC9mAICzvQIAM2YAgLWtAgC2pQIAN2YAgDtmAIA/ZgCAukEBALtBAQC8RQEAvU0BAL5FAQC/+QEAQ2YAgEdmAIBLZgCAT2YAgFNmAIBXZgCAW2YAgO/gAQCEbAQA4dQOAF9mAIDjHA4AY2YAgGdmAIBrZgCAb2YAgKMxAgBzZgCAhCQHAHdmAIB7ZgCApikCAKUhAgB/ZgCAq80BAKrNAQCDZgCAi2YAgK91AQCuyQEArcEBAKzJAQCo6QUAqekFAKr5BQCr+QUArOkFAK3pBQCuOQYArzkGABdmAICCzQcAgfUHAID9BwCHZgCAj2YAgIYYAwCHkAMAuNEGALnZBgC64QYAu+EGALyRBgC9nQYAvpUGAL+JBgCwSQYAsUkGALJdBgCzVQYAtE0GALXxBgC28QYAt/EGALDhBwCx4QcAsgkHALMJBwC0GQcAtRkHALYJBwC3CQcAuDkHALkNBwC6GQcAuxkHALwJBwC9CQcAvn0HAL9xBwCTZgCAp2UAgJdmAICbZgCAn2YAgKNmAICnZgCAq2YAgKjxBwCpxQcAqsEHAKvdBwCsyQcArb0HAK6pBwCvoQcAsykGAK9mAICzZgCAt2YAgLtmAIC2XQYAtSEGAL9mAIC7RQYAukUGAMNmAIDHZgCAv70GAL69BgC9vQYAvL0GAMtmAICjbQYAz2YAgNNmAICmGQYA12YAgNtmAIClZQYAqgEGAKsBBgDfZgCA42YAgK75BgCv+QYArPkGAK35BgCobQYAqbEBAKpJAQCrRQEArF0BAK1FAQCuTQEAr0UBAOdmAICCHQAAgR0AAIAdAADrZgCA72YAgPNmAIC+VAEAuIEAALmNAAC6hQAAu5kAALyJAAC9vQAAvrUAAL99AACwPQEAseEAALLhAACz4QAAtOEAALXpAAC20QAAt9EAALsFAwC62QIAhiwCAIcsAwC/DQMAvgUDAL0VAwC8FQMAs+ECAPtmAID/ZgCAhCwDAANnAIC25QIAtfUCAAdnAICqnQIAq0EDAAtnAIAPZwCArkEDAK9JAwCsUQMArVEDABNnAICjpQIAF2cAgBtnAICmoQIAH2cAgCNnAIClsQIAqakAAKihAACrtQAAqr0AAK3dAACs3QAAr/EAAK79AAC+LBwAJ2cAgCtnAIAvZwCAM2cAgDdnAIA7ZwCAP2cAgLl9AAC4fQAAu80BALrNAQC93QEAvN0BAL/NAQC+zQEAsZUAALCJAACzTQAAspUAALVdAAC0XQAAt00AALZNAABDZwCAR2cAgEtnAIBPZwCAU2cAgFdnAIBbZwCAX2cAgIA5AACBOQAAggUAAGNnAIBrZwCAb2cAgIf4AgCGfB0A4bgEAL7IHADjQAYAc2cAgHdnAIB7ZwCAf2cAgINnAICHZwCAi2cAgI9nAICTZwCAl2cAgJtnAIDvsAcAn2cAgKNnAICnZwCAq2cAgO/IAACvZwCAs2cAgLdnAIDvQAYAu2cAgOH8BgC/ZwCA4xwGAMNnAIDhlAEAx2cAgONkBgCAEQAAgRkAAIIpAACz/QEAy2cAgLWdAQC2lQEAz2cAgNNnAICEbB0AuoUBALuZAQC8iQEAvVEBAL5RAQC/UQEAozEeAGdnAIDXZwCA22cAgN9nAICmWR4ApVEeAONnAICrVR4AqkkeAIYIAwCHbAMAr50eAK6dHgCtnR4ArEUeAOdnAICzCR8A62cAgO9nAIC2CR8A82cAgPdnAIC1CR8AugUfALsNHwD7ZwCA/2cAgL4FHwC/CR8AvBUfAL0NHwCw5R8Ase0fALLlHwCz/R8AtOUfALXpHwC2GR8AtxkfALgpHwC5NR8Auj0fALs1HwC8ER8AvR0fAL4JHwC/BR8AA2gAgAdoAID3ZgCAC2gAgA9oAIATaACAF2gAgBtoAICo0R8AqdEfAKqlHwCrvR8ArKUfAK2tHwCupR8Ar50fAKNNHgAfaACAI2gAgCdoAIAraACApk0eAKVNHgAvaACAq0keAKpBHgAzaACAN2gAgK9NHgCuQR4ArUkeAKxRHgCADQAAgRUAAIIdAAA7aACAP2gAgENoAICEtAEAvrQBAL/oAQBLaACAhkgHAIc0AACEvAYAT2gAgFNoAIC+tAYAqI0BAKmVAQCqlQEAq80BAKzZAQCt2QEArs0BAK/FAQBXaACAW2gAgF9oAIBjaACAZ2gAgGtoAIBvaACAc2gAgLgdAQC5wQAAusEAALvBAAC8wQAAvckAAL7xAAC/8QAAsIkBALGJAQCyKQEAsykBALQ9AQC1JQEAti0BALclAQC7bQIAum0CAHdoAIB7aACAv8ECAL7ZAgC93QIAvN0CALM9AgB/aACAg2gAgIdoAICE/AYAtnkCALVxAgCLaACAqikCAKspAgCPaACAk2gAgK6dAgCvhQIArJkCAK2ZAgCXaACAo3kCAJtoAICfaACApj0CAKNoAICnaACApTUCAIJtJwCDjSoAhqgFAIdsAwCGmS4Ah80vAIQRLgCFmS4AiiESAIspEgCraACAr2gAgI6RFgCPHRYAjBESAI0RFgCScRoAk+UaALNoAIDvlHYAlvEeAJflHgCUSRoAlRkeAJopAgCb4QIAu2gAgL9oAIDDaACA4SASAJzxAgDjIBYAnyEfAJ7BHwCdmRsAnC0bAJuhGwCavRcAmTkXAJixFwCXiRMAlqkTAJWpEwCUdS4AkzkvAJIxLwCRsS8AkDUrAI+tJgDjeB8A0gAAAOFcHwCCmQEAx2gAgIDxAQCB8QEAvqgHAMtoAIDPaACA02gAgIS8BgDvLB8A12gAgNtoAIDhpB4A48wAAON8HgDhvAEA32gAgONoAIDnaACAhJwGAOtoAIC+bAYA72gAgPNoAID3aACA7xAAAO8EHgD7aACA/2gAgANpAIAHaQCAC2kAgA9pAIATaQCAF2kAgBtpAICAPQAAgQkAAILJBwAfaQCAo/kDAKLxAwChMQMAoM0fALBJcQCxAXwAsgl8ALMhfQC0AXgAtRV4AEdoAIC3aACAI2kAgL4oDgCGDAAAh4wDACdpAIAraQCAL2kAgDNpAIA3aQCAoV0AAKJVAACjfQAApAEMAKUVDACm9QwApwEIAKghCACpxQgAqgF0AKsJdACsAXQArR11AK55cACveXAAqOUFAKnxBQCq8QUAqy0FAKw1BQCtPQUArjUFAK8tBQA7aQCAP2kAgENpAIBHaQCAS2kAgE9pAIBTaQCAV2kAgLj9BgC5jQYAuoUGALutBgC8uQYAvbkGAL6tBgC/pQYAsFUFALFdBQCyVQUAs+UGALT9BgC10QYAttEGALfRBgCzeQQAW2kAgF9pAIBjaQCAZ2kAgLa9BAC1vQQAa2kAgLuZBAC6kQQAb2kAgHNpAIC/FQcAvjkHAL0xBwC8gQQAd2kAgKM9BAB7aQCAf2kAgKb5BACDaQCAh2kAgKX5BACq1QQAq90EAItpAICPaQCArn0HAK9RBwCsxQQArXUHAKhpBwCpaQcAqnkHAKvZBgCs9QYArf0GAK71BgCv5QYAgMkAAIHJAACCBQAAk2kAgIZwDwCHNAAAm2kAgJ9pAIC4fQYAuQUGALoNBgC7BQYAvB0GAL0FBgC+DQYAvwUGALCdBgCxdQYAsn0GALN1BgC0UQYAtV0GALZVBgC3TQYAs/EEAKNpAICnaQCAq2kAgK9pAIC2fQUAtX0FALNpAIC7sQUAulkFALdpAIC7aQCAv5kFAL6VBQC9oQUAvKkFAL9pAICjtQQAw2kAgMdpAICmOQUAy2kAgM9pAIClOQUAqh0FAKv1BQDTaQCA12kAgK7RBQCv3QUArO0FAK3lBQCpuQIAqLECAKvJAgCqsQIArTUCAKw1AgCvNQIArjUCANtpAIDfaQCA42kAgOdpAIDraQCA72kAgPNpAID3aQCAuekDALjZAwC7iQMAuuEDAL2dAwC8nQMAv4EDAL6JAwCxVQIAsFUCALNVAgCyVQIAtfkDALTxAwC36QMAtvEDALM9AwD7aQCA/2kAgANqAIALagCAtrEDALW5AwAPagCAu5UDALqVAwCGiAwAh6ANAL85AgC+MQIAvYUDALyFAwATagCAo3kDABdqAIAbagCApvUDAB9qAIAjagCApf0DAKrRAwCr0QMAJ2oAgCtqAICudQIAr30CAKzBAwCtwQMAgIUAAIGNAACChQAA79AGAOOwBwDj9AQA4QgHAOHsBADvOAYA7yAEAL6kDAAvagCAM2oAgOGEAQA3agCA49wGADtqAIA/agCAhMANALPJAQBDagCAtdkBALbJAQBHagCAS2oAgE9qAIC6xQEAu60BALy5AQC9uQEAvq0BAL+lAQCwLQ4AsUUOALJBDgCzQQ4AtEUOALVNDgC2cQ4At3EOALiBDgC5gQ4AuoEOALuBDgC8gQ4AvYEOAL6BDgC/gQ4AB2oAgFNqAIBXagCAW2oAgJdpAIBfagCAY2oAgGdqAICo2Q0AqdkNAKptDgCrZQ4ArH0OAK1lDgCuZQ4Ar1UOAKOFDgCCLQAAgRUAAIAdAABragCApoUOAKWVDgBvagCAq+EOAKqJDgBzagCAd2oAgK/pDgCu4Q4ArfUOAKz1DgB7agCAs4UPAIZoAACHHAMAtoUPAH9qAICDagCAtZEPALqNDwC7SQ8Ah2oAgItqAIC+MQ8AvzEPALxJDwC9RQ8AqBEOAKkZDgCqSQ4Aq0UOAKxdDgCtQQ4ArkEOAK91DgCPagCAk2oAgJdqAICbagCAn2oAgKNqAICnagCAq2oAgLihDgC5oQ4Aug0BALsFAQC8HQEAvQEBAL4BAQC/AQEAsA0OALHJDgCy2Q4As9UOALSxDgC1sQ4AtqkOALehDgCjwQ4Ar2oAgLNqAIC3agCAu2oAgKbBDgCl1Q4Av2oAgKsNDgCqyQ4Aw2oAgMdqAICvdQ4ArnUOAK0BDgCsDQ4Ay2oAgM9qAIDTagCA12oAgIANAACBNQAAgj0AANtqAIDfagCA42oAgISEAQC+hAEAhjAHAIf4AADragCA72oAgKjBAgCp0QIAqtECAKvlAgCs/QIArTUDAK49AwCvNQMA82oAgPdqAID7agCA/2oAgANrAIAHawCAC2sAgA9rAIC40QMAudkDALrhAwC74QMAvJEDAL2RAwC+kQMAv5EDALBNAwCxVQMAsl0DALNVAwC0TQMAtfEDALbxAwC38QMAu7EDALqpAwATawCAvoQDAL8VAwC+qQMAvaEDALypAwCzeQIAF2sAgBtrAIAfawCAI2sAgLaVAwC1VQIAJ2sAgKrtAwCr9QMAK2sAgC9rAICu7QMAr1EDAKztAwCt5QMAM2sAgKM9AgA3awCAO2sAgKbRAwA/awCAQ2sAgKURAgBHawCAgiEAAIEVAACAFQAA7wQAAISUAgBLawCAT2sAgOPYAABTawCA4fgBAFtrAIBfawCAY2sAgGdrAIBrawCAhmAFAIcIBQBvawCAs20BAHNrAIC1fQEAtnUBAHdrAIB7awCAf2sAgLpRAQC7UQEAvPkBAL3RAQC+0QEAv9EBAINrAICjpQEAh2sAgItrAICmvQEAj2sAgJNrAICltQEAqpkBAKuZAQCXawCAm2sAgK4ZAQCvGQEArDEBAK0ZAQCfawCA4fQOAKNrAIDjFA4A9AAAAOF8DACnawCA41AKAKtrAICvawCAviAEAO8wDQCzawCAt2sAgIQ0BADvrA4AsDkGALE5BgCygQYAs6kGALS5BgC1uQYAtqkGALehBgC46QYAuekGALrJBgC7xQYAvN0GAL3BBgC+wQYAvz0HAFdrAICCHQAAgR0AAIAdAAC7awCAv2sAgMNrAIDnagCAqJkFAKmZBQCqSQYAq0kGAKxZBgCtWQYArkkGAK9JBgCorQcAqbUHAKq9BwCrtQcArK0HAK3dBwCuyQcAr8EHAMdrAIDLawCAhogDAIcQAwDPawCA02sAgNdrAIDbawCAuG0HALkFBwC6AQcAuxUHALwxBwC9MQcAvikHAL8pBwCwgQcAsYEHALJpBwCzZQcAtH0HALVhBwC2YQcAt1UHALM1BgDfawCA42sAgOdrAIDrawCAtl0GALUlBgDvawCAu0UGALpFBgDzawCA92sAgL+lBgC+uQYAvbEGALy9BgD7awCAo3EGAP9rAIADbACAphkGAAdsAIALbACApWEGAKoBBgCrAQYAD2wAgBNsAICu/QYAr+EGAKz5BgCt9QYAqCUBAKk1AQCqPQEAqzUBAKwtAQCtkQAArpEAAK+RAAAXbACAG2wAgB9sAIAjbACAJ2wAgIK9AwCBvQMAgL0DALiZAAC5rQAAuqUAALttAAC8dQAAvX0AAL51AAC/bQAAsPEAALH5AACywQAAs8EAALSxAAC1vQAAtrUAALepAAArbACAL2wAgDNsAICEgAIAvhwCADtsAICG+HwAh8wCAISsAwA/bACAQ2wAgEdsAIBLbACAT2wAgFNsAIBXbACAs/UCAFtsAIBfbACAkgAAAGNsAIC2UQMAteUCAGdsAIC7fQMAunUDAGtsAIBvbACAvzkDAL41AwC9VQMAvFUDAKM1AgBzbACAd2wAgHtsAIB/bACAppEDAKUlAgCDbACAq70DAKq1AwCHbACAi2wAgK/5AwCu9QMArZUDAKyVAwC+wAMAj2wAgJNsAICXbACAgA0AAIE1AACCPQAAm2wAgJ9sAICjbACAhsh8AIcAAwCrbACAr2wAgLNsAIC3bACAu2wAgL9sAIDDbACAx2wAgMtsAIDPbACA02wAgO/0AwCE7HwA4ZQBANdsAIDjMAMA22wAgN9sAIDjbACA52wAgLNpAQDrbACA72wAgPNsAID3bACAtmEBALVpAQD7bACAuykBALohAQD/bACAA20AgL8dAQC+HQEAvSUBALwtAQAHbQCAC20AgA9tAICjpQEAE20AgKWlAQCmrQEAvlR8AIaAfACH7HwAqu0BAKvlAQCs4QEArekBAK7RAQCv0QEAG20AgOGcBgCEBH8A4yQGAOPUBgAfbQCA4TAEACNtAIDvlAcAgnUAAIFhAACAaQAAJ20AgCttAIAvbQCA7+wGALiNfgC5lX4AupV+ALulfgC8vX4AvdF+AL7RfgC/0X4AsGV+ALFtfgCyeX4As3F+ALRZfgC1WX4Atr1+ALe1fgCoVX4AqWF+AKphfgCrYX4ArGF+AK1hfgCuYX4Ar2F+ADNtAICnbACAN2wAgDdtAIAXbQCAO20AgD9tAIBDbQCAqHF+AKlxfgCqcX4Aq3F+AKyRfwCtkX8ArpF/AK+RfwBHbQCAS20AgE9tAIBTbQCAV20AgFttAIBfbQCAY20AgLiFfwC5jX8AuoV/ALudfwC8jX8Avb1/AL61fwC/XX8AsPF/ALHxfwCy8X8As8V/ALTBfwC1wX8AtsF/ALfBfwCz+X8AZ20AgGttAIBvbQCAc20AgLYRfgC1GX4Ad20AgLs1fgC6NX4Ae20AgH9tAIC/BX4AvgV+AL0RfgC8JX4AghUAAKO9fwCAYQAAgWEAAKZVfgCDbQCAvpABAKVdfgCqcX4Aq3F+AIdtAICLbQCArkF+AK9BfgCsYX4ArVV+AKhBfgCpUX4AqlV+AKt9fgCsZX4ArW1+AK75AQCv8QEAhgAAAIc0AQCPbQCAk20AgJdtAICbbQCAn20AgKNtAIC4dQEAuX0BALp1AQC7yQAAvNkAAL3ZAAC+yQAAv8EAALCVAQCxnQEAspUBALNNAQC0VQEAtV0BALZVAQC3TQEAs919AKdtAICrbQCAr20AgLNtAIC27X0Ate19ALdtAIC7WQIAulECALttAIC/bQCAv5kCAL6RAgC9mQIAvEECAMNtAICjmX0Ax20AgMttAICmqX0Az20AgNNtAIClqX0AqhUCAKsdAgDXbQCA220AgK7VAgCv3QIArAUCAK3dAgDfbQCA420AgOdtAIDrbQCAgB0AAIEJAACCOQAA720AgPNtAIC+AAQA+20AgP9tAIADbgCAB24AgAtuAIAPbgCAhIwDABNuAICHCAMAhuwEABduAIDviAIAG24AgB9uAICEbAQA4zQCACNuAIDhVAEAJ24AgCtuAIAvbgCAM24AgKhtAgCprQIAqqUCAKu9AgCspQIAra0CAK6lAgCvGQEAvqwEADduAIA7bgCAP24AgENuAIBHbgCAS24AgE9uAIC4DQEAuREBALoRAQC7JQEAvD0BAL3VAQC+3QEAv9UBALBpAQCxaQEAsnkBALNxAQC0WQEAtVkBALY5AQC3NQEAsy0CAFNuAIBXbgCAW24AgF9uAIC2LQIAtS0CAGNuAIC7rQEAuq0BAGtuAIBvbgCAv50BAL6dAQC9pQEAvK0BAIBNAACBVQAAglUAAO9sAABzbgCA7+x/AO+8fgB3bgCA4RB/AOPUfwDj2H4A4ex/AHtuAIDhTH4Af24AgOMkfgD3bQCAZ24AgKsFBgCqBQYArQ0GAKwFBgCvNQYArjUGAIYAAwCHKAMAo4UFAINuAIClhQUAh24AgItuAICmhQUAs/EGAI9uAICTbgCAl24AgJtuAIC26QYAteEGAJ9uAIC7vQYAur0GAKNuAICnbgCAv4kGAL6BBgC9iQYAvJUGAKgpBgCpKQYAqjkGAKs5BgCsKQYArSkGAK5dBgCvTQYAq24AgK9uAICzbgCAt24AgLtuAIC/bgCAw24AgMduAIC46QcAuekHALr5BwC7+QcAvOkHAL3pBwC+XQcAv1UHALA5BgCxOQYAsgEGALMdBgC0BQYAtQ0GALYFBgC32QcAo7EHAIItAACBFQAAgB0AAMtuAICmqQcApaEHAM9uAICr/QcAqv0HANNuAICEpAIAr8kHAK7BBwCtyQcArNUHAL7MAQCzlQYA124AgNtuAIC2qQYA324AgONuAIC1rQYAulkBALshAQCGyAAAhwwBAL4hAQC/KQEAvDEBAL0xAQCoKQYAqSkGAKpZBgCrUQYArGEGAK1tBgCutQEAr6kBAITgAQDnbgCA624AgO9uAIDzbgCA924AgPtuAID/bgCAuGEBALlhAQC6YQEAu2EBALxhAQC9YQEAvmEBAL9hAQCw2QEAsaEBALKhAQCzoQEAtKEBALWpAQC2kQEAt5EBAKPRBQADbwCAB28AgAtvAIAPbwCApu0FAKXpBQATbwCAq2UCAKodAgAXbwCAG28AgK9tAgCuZQIArXUCAKx1AgAfbwCAI28AgCdvAIArbwCAL28AgDNvAIA3bwCAO28AgIA9AACBCQAAghkAAD9vAIBDbwCAS28AgL48AwBPbwCAhgAMAIcUAwBTbwCAs9UDAFdvAIC1PQMAtjUDAFtvAIBfbwCAv4wKALoRAwC7EQMAvLUAAL29AAC+tQAAv60AAGNvAIDjdAEAZ28AgOG8AQBrbwCAb28AgHNvAIB3bwCAe28AgH9vAICDbwCAh28AgItvAIDvdAIAj28AgJNvAICoTQIAqVECAKpRAgCrqQIArLkCAK25AgCuqQIAr6kCAIRsDQCXbwCAm28AgJ9vAICjbwCAp28AgKtvAIC+dA0AuG0BALkFAQC6DQEAuwUBALwdAQC9BQEAvg0BAL8FAQCw2QIAsdkCALJtAQCzZQEAtH0BALVlAQC2ZQEAt1UBAOG4AQDhUAcA47QAAON8BwCAqQAAgQkAAII5AACvbwCAs28AgLtvAIC/bwCAw28AgO4AAADHbwCA7wAAAO9kBgCGYAwAh+QMAKORAgDLbwCApXkCAM9vAIDTbwCApnECANdvAIDbbwCAq1UCAKpVAgCt+QEArPEBAK/pAQCu8QEAt28AgEdvAIDfbwCA428AgOdvAIDrbwCA728AgPNvAICoVQ4AqVkOAKqhDgCrvQ4ArK0OAK2VDgCu+Q4Ar/UOALCRDgCxkQ4AspEOALORDgC0sQ4AtbEOALaxDgC3sQ4AuJEOALmdDgC6lQ4Au0kPALxZDwC9WQ8AvkkPAL9JDwCzCQ4A928AgPtvAID/bwCAA3AAgLY1DgC1BQ4AB3AAgLt1DgC6dQ4AC3AAgA9wAIC/VQ4AvlUOAL1lDgC8ZQ4AE3AAgKNNDgAXcACAG3AAgKZxDgAfcACAI3AAgKVBDgCqMQ4AqzEOAISkAwC+pAMArhEOAK8RDgCsIQ4ArSEOAKilDgCprQ4AqqUOAKu5DgCs3Q4ArcEOAK7BDgCv/Q4AgO0BAIHxAQCC8QEAJ3AAgIaQAQCHtAEAK3AAgC9wAIC4yQEAuckBALrZAQC70QEAvPkBAL35AQC+mQEAv5UBALCFDgCxbQEAsmUBALN9AQC0ZQEAtW0BALZlAQC3+QEAsy0OADNwAIA3cACAO3AAgD9wAIC2QQ4AtVUOAENwAIC7qQEAukEOAEdwAIBLcACAv6kBAL6hAQC9qQEAvLEBAE9wAICjaQ4AU3AAgFdwAICmBQ4AW3AAgF9wAIClEQ4AqgUOAKvtAQBjcACAZ3AAgK7lAQCv7QEArPUBAK3tAQCoOQMAqTkDAKqNAwCrhQMArJ0DAK2FAwCuhQMAr7UDAGtwAIBvcACAc3AAgHdwAIB7cACAf3AAgINwAICHcACAuGEAALlhAAC6YQAAu2EAALxhAAC9YQAAvmEAAL9hAACwzQMAsaUDALKhAwCzoQMAtKUDALWtAwC2kQMAt5EDAIANAACBEQAAghEAAItwAIDv9AIAj3AAgJNwAIC+HAMA4xQCAISIAgDhgAEAm3AAgJ9wAICjcACAh8gDAIY8BAC7AQMAumkDAKdwAICrcACAvwkDAL4BAwC9FQMAvBUDALNlAwCvcACAs3AAgLdwAIC7cACAtmUDALV1AwC/cACAw3AAgMdwAIDLcACAo4kCAM9wAIClmQIApokCANNwAICELAIA13AAgKqFAgCr7QIArPkCAK35AgCu7QIAr+UCANtwAIDfcACAvkQFAIRMBQDjcACA53AAgOtwAIDvcACA83AAgPdwAID7cACA/3AAgIAZAACBGQAAggUAAANxAIDhGA8A4VwOAOO4DgDjdAEAC3EAgA9xAIATcQCAF3EAgIYABACHZAUAG3EAgB9xAIAjcQCAJ3EAgO98DgDvqAEAs3UBACtxAIAvcQCAM3EAgDdxAIC2MQEAtRUBADtxAIC7HQEAuhUBAD9xAIBDcQCAv+EAAL79AAC9/QAAvP0AAAdxAIBHcQCAS3EAgE9xAICXcACAU3EAgFdxAIBbcQCAqI0GAKmVBgCqnQYAq+UGAKz9BgCt0QYArtEGAK/RBgCwsQYAsbkGALJJBwCzSQcAtFkHALVFBwC2RQcAt3kHALghBwC5IQcAujkHALs5BwC8KQcAvSkHAL4ZBwC/GQcAozUGAF9xAIBjcQCAZ3EAgGtxAICmcQYApVUGAG9xAICrXQYAqlUGAHNxAIC+oAMAr6EHAK69BwCtvQcArL0HAIBRAACBWQAAgmEAALNVBwCF9AAAtX0HALZ1BwB3cQCAhgAcAIfkAQC6LQcAuyUHALw9BwC9JQcAviUHAL8VBwCokQYAqZEGAKqRBgCrkQYArLkGAK25BgCuqQYAr6kGAHtxAIB/cQCAg3EAgIdxAICiIQEAozUBAKA5BQChEQQAuEkBALlJAQC6XQEAu1UBALxNAQC90QEAvtEBAL/RAQCwpQYAsa0GALKlBgCzvQYAtK0GALWdBgC2lQYAt3kBAKMZBgCPnXkAi3EAgI9xAICTcQCApjkGAKUxBgCXcQCAq2kGAKphBgCbcQCAn3EAgK9ZBgCuaQYArWkGAKxxBgCeiQgAn8EFAJzJCQCdyQkAmqENAJu9DACYsQ0AmbkNAJahcQCXRXEAlEV1AJWxcQCSoXUAk7V1AJDleQCRzXkAil1yAItFcgCjcQCAvoAcAI51DgCPZQ4AjLlyAI11DgCCOXoAgzl6AKdxAICrcQCAhnF2AIeZdgCECXoAhW12AJptBwCbVQIAr3EAgLNxAIC3cQCA4ZAAAJxZAgDjCBoAkgkPAJNlCgC7cQCA7zgWAJZ1BgCXdQYAlH0KAJU1CwCpjRYAqIUWAKsBEACqMRYArXESAKy1EgCvuS4ArgEsAKF9AgC/cQCAo6EeAKKpHgClsRoApPUfAKflGwCmsRoAhMwDAIRMHADDcQCAx3EAgMtxAIDPcQCA03EAgNdxAICxASgAsNkuALONKgCy6SoAtfUmALQBJACEcB0A23EAgID9AQCBFQAAgh0AAL6AHADfcQCA43EAgIe4AgCGPB0A63EAgO9xAIDzcQCA93EAgPtxAID/cQCAA3IAgAdyAIALcgCAD3IAgBNyAIAXcgCA44ADABtyAIDhoAEAH3IAgO+UAwAjcgCAJ3IAgCtyAIAvcgCAM3IAgDdyAIA7cgCAP3IAgOE8BgBDcgCA49AGAEdyAIDhMAcAS3IAgOOsBgCAOQAAgRUAAIIdAADvHAYAT3IAgFNyAIC+uB8A7+gBALPpAgBbcgCAh8QcAIbsHABfcgCAtlkCALVRAgBjcgCAu00CALpNAgBncgCAa3IAgL+5AQC+2QEAvdEBALz1AQCjKR0A53EAgFdyAIBvcgCAc3IAgKaZHQClkR0Ad3IAgKuNHQCqjR0Ae3IAgH9yAICveR4ArhkeAK0RHgCsNR4Ag3IAgLNtHwCHcgCAi3IAgLZlHwCPcgCAk3IAgLVtHwC6IR8AuyEfAJdyAICbcgCAviUfAL8pHwC8MR8AvTEfAKihHwCpoR8AqqEfAKuhHwCsoR8AraEfAK6hHwCvoR8An3IAgKNyAICncgCAq3IAgK9yAICzcgCAt3IAgLtyAIC4rR8AubUfALq9HwC7tR8AvK0fAL1VHwC+UR8Av00fALChHwCxoR8AsqEfALOhHwC0pR8AtakfALadHwC3lR8AoykeAIIZAACBGQAAgLEBAL9yAICmIR4ApSkeAMNyAICrZR4AqmUeAIaIAACH/AEAr20eAK5hHgCtdR4ArHUeAMdyAICzmR4Ay3IAgM9yAIC2XQEA03IAgNdyAIC1sR4AukkBALtJAQDbcgCA33IAgL49AQC/IQEAvDkBAL01AQCoRR4AqVUeAKpVHgCrZR4ArH0eAK2ZAQCuiQEAr4EBAISsAADjcgCA53IAgOtyAIDvcgCA83IAgPdyAID7cgCAuK0BALllAQC6bQEAu2UBALx9AQC9ZQEAvm0BAL9lAQCwyQEAsckBALKpAQCzpQEAtL0BALWhAQC2oQEAt5UBALhpHAC5oRwAusEcALvBHAC8wRwAvcEcAL7BHAC/wRwAsIkfALGJHwCyIRwAswUcALQdHAC1fRwAtnUcALdtHACoYR8AqWEfAKphHwCrYR8ArNkfAK3ZHwCuyR8Ar8EfAP9yAIADcwCAB3MAgAtzAIAPcwCAE3MAgBdzAIAbcwCAH3MAgCNzAIC+AAQAo1EdACdzAICleR0AppUCACtzAIAvcwCAM3MAgKqBAgCrgQIArPECAK39AgCu9QIAr+kCADtzAIDh9AEAP3MAgON8AQCATQAAgXUAAIJ9AABDcwCAhsAEAIekBABHcwCAS3MAgE9zAIBTcwCAV3MAgO+MAgCoSQIAqUkCAKpdAgCrVQIArHkCAK15AgCuvQIAr7UCAISgBQBbcwCAX3MAgGNzAIC+vAQAZ3MAgGtzAIBvcwCAuC0BALk1AQC6PQEAuzUBALwtAQC91QEAvt0BAL/NAQCwzQIAsdUCALLdAgCz1QIAtM0CALUVAQC2HQEAtxUBAOGEHgDjbB8A41wfAOFYHgBzcwCAd3MAgHtzAIB/cwCAg3MAgIdzAICLcwCAj3MAgOkAAADv9B4A70weAJNzAICzlQIAl3MAgJtzAICfcwCAo3MAgLa5AgC1sQIAq3MAgLtRAgC6SQIAhsgEAIesBAC/kQEAvkkCAL1BAgC8SQIAN3MAgKNRBQCvcwCAp3MAgKZ9BQCzcwCAt3MAgKV1BQCqjQUAq5UFALtzAIC/cwCAro0FAK9VBgCsjQUArYUFAICJBwCBiQcAgpkHALORBgDDcwCAtbkGALapBgDHcwCAy3MAgM9zAIC6TQcAu0UHALxdBwC9QQcAvkEHAL9BBwCoQQYAqU0GAKpVBgCrZQYArH0GAK1lBgCubQYAr2UGANNzAIDXcwCA23MAgN9zAIDjcwCA53MAgOtzAIDvcwCAuFkHALlZBwC6aQcAu2kHALx5BwC9eQcAvmUHAL8ZBwCwxQcAsc0HALLFBwCz2QcAtMkHALXJBwC2aQcAt2kHAKPdBwDzcwCA93MAgPtzAID/cwCApuUHAKX1BwADdACAqwkGAKoBBgAHdACAC3QAgK8NBgCuDQYArQ0GAKwRBgCAbQAAgQkAAIIZAAAPdACAE3QAgISYAQC+kAEAF3QAgIbAAACH5AEAG3QAgB90AIAjdACAJ3QAgCt0AIAvdACAqF0GAKmNAQCqnQEAq5UBAKy5AQCtuQEArskBAK/BAQCEoAAAM3QAgDd0AIA7dACAP3QAgEN0AIBHdACAS3QAgLh5AQC5eQEAus0AALvFAAC83QAAvcUAAL7FAAC/9QAAsIEBALGBAQCySQEAs0kBALRZAQC1WQEAtkkBALdJAQCzFQIAT3QAgFN0AIBXdACAW3QAgLY5AgC1MQIAX3QAgLtFAgC6RQIAY3QAgGd0AIC/nQIAvp0CAL2dAgC8nQIAhXw+AKNRAgBrdACAb3QAgKZ9AgBzdACAd3QAgKV1AgCqAQIAqwECAHt0AIB/dACArtkCAK/ZAgCs2QIArdkCAIDpAACB6QAAggUAAIN0AIC+AAwAi3QAgIeoAwCGvAwAj3QAgJN0AICXdACAm3QAgJ90AICjdACAp3QAgKt0AICvdACAs3QAgLd0AIC7dACA42ABAL90AIDhoAEAw3QAgO+IAgDHdACAy3QAgM90AIDTdACA13QAgNt0AIDfdACAqGkCAKlpAgCqeQIAq3kCAKxpAgCtaQIArr0CAK+1AgC+rAwA43QAgOd0AIDrdACAgB0AAIEJAACCqQAA73QAgLhRAQC5WQEAumEBALthAQC8GQEAvRkBAL4NAQC/BQEAsM0CALHVAgCy3QIAs9UCALTNAgC1cQEAtnEBALdxAQDjxAAA4XwHAOF4BgDjvAYA83QAgIQYDQCGuAwAhzwNAL4sDwD7dACA/3QAgAN1AIDvEAAAB3UAgAt1AIDvdAYAD3UAgBN1AIAXdQCAs70CABt1AIC1rQIAtqUCAB91AIAjdQCAJ3UAgLpFAgC7XQIAvEUCAL1NAgC+RQIAv/kBAId0AIClfQ0ApnUNAPd0AIArdQCAL3UAgDN1AICjbQ0ArJUNAK2dDQCulQ0ArykOADd1AIA7dQCAqpUNAKuNDQCz5Q4AP3UAgEN1AIBHdQCAS3UAgLblDgC19Q4AT3UAgLuhDgC62Q4AU3UAgFd1AIC/pQ4AvrkOAL2xDgC8uQ4AqBUOAKklDgCqLQ4AqyUOAKw9DgCtJQ4Ari0OAK8lDgCADQAAgRUAAIIdAABbdQCAX3UAgGN1AICEMAMAZ3UAgLgpDgC5KQ4AujkOALs5DgC8KQ4AvSkOAL79DwC/9Q8AsF0OALElDgCyLQ4AsyUOALQ9DgC1IQ4AtiUOALcZDgCjpQ8Aa3UAgIYoAQCHTAEAb3UAgKalDwCltQ8Ac3UAgKvhDwCqmQ8Ad3UAgHt1AICv5Q8ArvkPAK3xDwCs+Q8Af3UAgLPpDgCDdQCAh3UAgLaRDgCLdQCAj3UAgLXlDgC6sQ4Au7kOAJN1AICXdQCAvmEBAL9hAQC8mQ4AvZkOAKglDgCpLQ4AqiUOAKs5DgCsKQ4ArVUOAK5dDgCvVQ4Am3UAgJ91AICjdQCAp3UAgKt1AICvdQCAs3UAgLd1AIC49QEAuYEBALqBAQC7gQEAvIEBAL2JAQC+sQEAv7EBALAxDgCxOQ4AsgkOALMJDgC04QEAteEBALbhAQC3zQEAo60NALt1AIC/dQCAw3UAgMd1AICm1Q0ApaENAMt1AICr/Q0AqvUNAM91AIDTdQCAryUCAK4lAgCt3Q0ArN0NAIBdAACBbQAAgmUAALNRAwC+nAMAtXkDALYZAwDbdQCAhOACAN91AIC6PQMAuzUDALwZAwC9GQMAvtkDAL/ZAwCohQMAqZUDAKqVAwCrpQMArL0DAK3VAwCu0QMAr9EDAIYABACHNAMAv6AzAON1AIDndQCA63UAgO91AIDzdQCAuHEDALlxAwC6cQMAu3EDALzVAAC93QAAvtUAAL/NAACwtQMAsb0DALKBAwCzgQMAtFEDALVRAwC2UQMAt1EDAO+oAwD3dQCA+3UAgP91AICEHAIAA3YAgAd2AIALdgCAviwFAA92AIATdgCAF3YAgONAAwAbdgCA4SgAAB92AICjXQIAI3YAgCd2AIArdgCAL3YAgKYVAgCldQIAM3YAgKs5AgCqMQIAN3YAgDt2AICv1QIArtUCAK0VAgCsFQIA4ygBAOEADwDhCA4A4wgOAID9AACBCQAAgjkAAD92AIBDdgCAS3YAgE92AIBTdgCA7+gOAFd2AIBbdgCA72QOALNtAQBfdgCAhugEAIcMBQBjdgCAtm0BALVtAQBndgCAu+0AALrtAABrdgCAb3YAgL/VAAC+6QAAveEAALzpAACoXQYAqWEGAKqlBgCrvQYArKUGAK2tBgCupQYArxkHAEd2AIBzdgCAd3YAgHt2AIB/dgCAg3YAgId2AICLdgCAuHUHALl5BwC6DQcAuwUHALwdBwC9BQcAvgUHAL81BwCwaQcAsWkHALJ9BwCzdQcAtG0HALVRBwC2UQcAt1EHAKMtBgCPdgCAk3YAgJd2AICbdgCApi0GAKUtBgCfdgCAq60HAKqtBwCjdgCAp3YAgK+VBwCuqQcAraEHAKypBwCADQAAgRUAAIIdAACrdgCAr3YAgLN2AICEVAMAvlwAALd2AIC7dgCAhugAAIdMAwC/dgCAw3YAgMd2AIDLdgCAz3YAgOMEBADTdgCA4bQFANd2AIDbdgCA33YAgON2AIDndgCA63YAgO92AIDzdgCA93YAgO/sBAD7dgCA/3YAgLPtBgADdwCAB3cAgAt3AIAPdwCAtpEGALXhBgATdwCAu40GALqNBgAXdwCAG3cAgL9BAQC+WQEAvVEBALxZAQCoJQYAqS0GAKolBgCrOQYArCkGAK1RBgCuSQYAr0EGAIDNAACBCQAAghkAAB93AIAjdwCAhCwBAL40AAArdwCAuP0BALlBAQC6QQEAu0EBALxBAQC9SQEAvnEBAL9xAQCwCQYAsQkGALLNAQCzxQEAtN0BALXFAQC2zQEAt8UBAIagPACHRAMAL3cAgKOhBQAzdwCApa0FAKbdBQA3dwCAO3cAgL4oPACqwQUAq8EFAKwVAgCtHQIArhUCAK8NAgC2QQMAP3cAgEN3AIC1sQIAR3cAgLOhAgBLdwCAT3cAgL5FAwC/TQMAvHUDAL1NAwC6ZQMAu20DAFN3AIBXdwCAW3cAgF93AIDXdQCAY3cAgGd3AIBrdwCAb3cAgHN3AICoRQIAqVUCAKpdAgCrVQIArE0CAK21AwCusQMAr60DALDVAwCx3QMAstUDALPtAwC09QMAtf0DALb1AwC37QMAuNkDALnZAwC6rQMAu6UDALy9AwC9pQMAvqUDAL+VAwCj9QMAd3cAgHt3AIB/dwCAg3cAgKYVAgCl5QMAh3cAgKs5AgCqMQIAi3cAgI93AICvGQIArhECAK0ZAgCsIQIAgGkAAIFpAACCBQAAk3cAgJt3AICfdwCAo3cAgO8cAACEbAIA4ZQBAKd3AIDjyAAAq3cAgK93AICGWDwAh1A9ALN3AIC3dwCAu3cAgISEPQC/dwCAw3cAgMd3AIDvuAEAvmw8AOF0BgDLdwCA42QBAM93AIDTdwCA13cAgNt3AICz0QEA33cAgON3AIDndwCA63cAgLaRAQC1+QEA73cAgLu9AQC6vQEA83cAgPd3AIC/dQEAvnUBAL2FAQC8hQEAqL09AKkNPgCqGT4AqxE+AKwxPgCtUT4ArlE+AK9NPgCXdwCAgh0AAIEdAACAHQAA+3cAgP93AIADeACAB3gAgLjVPgC53T4AutU+ALtJPwC8WT8AvVk/AL5JPwC/QT8AsDk+ALE5PgCyET4AsxE+ALTxPgC18T4AtvU+ALftPgCjkT4AC3gAgIYoAACHwAMAD3gAgKbRPgCluT4AE3gAgKv9PgCq/T4AF3gAgBt4AICvNT4ArjU+AK3FPgCsxT4AH3gAgLOdPwAjeACAJ3gAgLalPwAreACAL3gAgLWtPwC6aT8Au3U/ADN4AIA3eACAvlk/AL9FPwC8bT8AvWU/ADt4AIA/eACAQ3gAgEd4AIDjYDwAS3gAgOEAPQBPeACA7/w9AFN4AIBXeACAW3gAgF94AIBjeACAZ3gAgGt4AICjGT4AghkAAIEZAACAcQAAb3gAgKYhPgClKT4Ac3gAgKvxPgCq7T4AhCQBAL4kAQCvwT4Art0+AK3hPgCs6T4AqNE+AKnRPgCq0T4Aq+U+AKzhPgCt4T4Arhk+AK8ZPgCGAAAAh4QAAHt4AIB/eACAg3gAgId4AICLeACAj3gAgLh9PgC5AT4AugE+ALsBPgC8AT4AvQk+AL4xPgC/MT4AsGk+ALF1PgCyfT4As3U+ALRZPgC1RT4Atk0+ALdFPgCohQIAqZUCAKqVAgCrpQIArL0CAK3VAgCu0QIAr9ECAJN4AICXeACAm3gAgL8k5gGfeACAo3gAgKd4AICreACAuFUDALlZAwC6bQMAu2UDALx9AwC9ZQMAvm0DAL9lAwCwtQIAsb0CALKBAgCzgQIAtHEDALVxAwC2cQMAt3EDALMdAgCveACAs3gAgLd4AICEiAMAtlUCALU1AgAndwCAu3kCALpxAgC7eACAv3gAgL+1AwC+tQMAvVUCALxVAgDDeACAo1kCAMd4AIDLeACAphECAM94AIDTeACApXECAKo1AgCrPQIA13gAgNt4AICu8QMAr/EDAKwRAgCtEQIAqKkCAKmpAgCquQIAq7kCAKypAgCtqQIArjkBAK85AQCAzQEAgQkAAIIZAADfeACA43gAgL64BQDreACA73gAgLjpAQC56QEAuokBALuFAQC8nQEAvYEBAL6BAQC/tQEAsEkBALFVAQCyXQEAs1UBALRNAQC18QEAtvEBALfxAQDvFAAA83gAgIaoBQCH3AUA93gAgIRYBAD7eACA78Q+AP94AIDhxD4AA3kAgOMwPgDjyAAAB3kAgOEoAQALeQCAtn0CAA95AIATeQCAtXUCABd5AICzZQIAG3kAgB95AIC+3QEAv2EBALzdAQC91QEAutkBALvFAQAjeQCAJ3kAgKOxBQDneACAK3kAgC95AIAzeQCApqkFAKWhBQA3eQCAqxEGAKoNBgA7eQCAP3kAgK+1BgCuCQYArQEGAKwJBgBDeQCAR3kAgEt5AIBPeQCAgBkAAIEZAACCBQAAU3kAgL5sAwBXeQCAhsgAAIccAwBbeQCAX3kAgGN5AIBneQCAqLkHAKm5BwCqDQcAqx0HAKwJBwCtNQcArjEHAK8pBwCEqAMAa3kAgG95AIBzeQCAd3kAgHt5AIB/eQCAg3kAgLjJAAC5yQAAutkAALvRAAC8+QAAvfkAAL6ZAAC/mQAAsF0HALEhBwCyIQcAsz0HALQpBwC1KQcAtgEHALcBBwCzhQYAh3kAgIt5AICPeQCAk3kAgLa1BgC1gQYAl3kAgLvlBgC6mQYAm3kAgJ95AIC/7QYAvu0GAL3pBgC89QYAo3kAgKd5AICreQCAr3kAgLN5AIC3eQCAu3kAgO+QBAC/eQCA4dwGAMN5AIDj7AUAgCkAAIEVAACCEQAAvnwBAKMFBgDLeQCAhigAAIdMAQDPeQCApjUGAKUBBgDTeQCAq2UGAKoZBgDXeQCA23kAgK9tBgCubQYArWkGAKx1BgDfeQCAs70BAON5AIDneQCAtnkBAOt5AIDveQCAtXkBALpVAQC7XQEA83kAgPd5AIC++QAAv/kAALxFAQC9+QAAqHECAKlxAgCqcQIAq3ECAKy1AgCtvQIArrUCAK+tAgCE7AwA+3kAgP95AIADegCAB3oAgAt6AIAPegCAE3oAgLhpAwC5aQMAugkDALsJAwC8GQMAvRkDAL4JAwC/CQMAsNUCALHdAgCy1QIAs2kDALR5AwC1eQMAtmkDALdhAwAXegCAG3oAgB96AICj9QIAI3oAgKUxAgCmMQIAJ3oAgCt6AIAvegCAqh0CAKsVAgCsDQIArbEDAK6xAwCvsQMAgGEAAIFhAACCBQAAM3oAgIbwDACHYAMAvhAMADt6AIB3eACAP3oAgEN6AIBHegCAS3oAgE96AIBTegCAV3oAgKiFAgCplQIAqpUCAKulAgCsvQIArdUCAK7RAgCv0QIAW3oAgF96AIBjegCAZ3oAgGt6AIBvegCAc3oAgHd6AIC4dQEAuX0BALp1AQC7zQEAvNUBAL3dAQC+yQEAv8EBALC1AgCxvQIAsoECALOBAgC0VQEAtV0BALZVAQC3TQEA4RAGAIRIDADjDAYAe3oAgISYDAB/egCAg3oAgId6AICLegCAj3oAgJN6AICXegCAgXUAAIB1AADvIAEAgnUAAJt6AICfegCAo3oAgL7ADACFtA4A4RACAO9cAADjABYA4ZABAKt6AIDjWAEA7zwHAK96AICzegCAhgAIAIe4DACznQ0AN3oAgLd6AIC7egCAv3oAgLbVDQC1tQ0Aw3oAgLv5DQC68Q0Ax3oAgMt6AIC/GQ4AvhEOAL3VDQC81Q0Az3oAgKPZDQDTegCA13oAgKaRDQDbegCA33oAgKXxDQCqtQ0Aq70NAON6AIDnegCArlUOAK9dDgCskQ0ArZENAKhdDgCpYQ4AqmEOAKthDgCsYQ4ArWEOAK5hDgCvYQ4A63oAgO96AIDzegCA93oAgPt6AID/egCAA3sAgAd7AIC4TQ8AuVEPALpRDwC7UQ8AvHEPAL1xDwC+cQ8Av3EPALDBDwCxwQ8AssEPALPBDwC0wQ8AtcEPALbBDwC3wQ8As+kPAAt7AIC+gAEAD3sAgKd6AIC24Q8AtekPABN7AIC7BQ4AugUOABt7AIAXewCAvwUOAL4FDgC9FQ4AvBUOAIFNAACAQQAA72gNAIJRAACG8AcAh9QBAB97AIAjewCAJ3sAgIRwAQArewCAL3sAgOHgDgAzewCA40gNADd7AICjaQ8AO3sAgD97AIBDewCAR3sAgKZhDwClaQ8AS3sAgKuFDgCqhQ4AT3sAgFN7AICvhQ4AroUOAK2VDgCslQ4AV3sAgLMxDgBbewCAX3sAgLbBAQBjewCAZ3sAgLXRAQC6zQEAu6UBAGt7AIBvewCAvqUBAL+tAQC8sQEAvbEBAI/dJgCj8Q0Ac3sAgHd7AICmAQIAe3sAgH97AIClEQIAqg0CAKtlAgCDewCAviAEAK5lAgCvbQIArHECAK1xAgCfoQwAnnkKAJ1pCgCc0QgAm7E2AJp1NgCZ0TQAmOEyAJdtMgCWZTIAlTU/AJRhPgCTcT4AkjU7AJFxOgCQeToAgJUAAIGdAACCoQAAi3sAgO9EAgDhdA8Aj3sAgOMcDwDj1AEAk3sAgOHgAQDvXAEAo7UCAKJBAACh3Q4AoLkOALWpAwCXewCAhMAEALahAwCG8AUAh+QEALOFAwCbewCAvXEDALxpAwC/QQMAvnEDAJ97AIDHeQCAu3EDALp5AwCC3ScAgwE7AL6EBwC+wAYAhhE/AIcZPwCEETsAhV06AIp9PgCLJTMAo3sAgKd7AICOuTUAjxU3AIw1MwCNgTMAkqE3AJPZCQC+xBkAq3sAgJaxDQCXUQ8AlHkLAJVhCwCaBQ8Am5EBAK97AICzewCAt3sAgN0AAACcfQMAu3sAgOFIDwC/ewCA4xwOAMN7AIDHewCAy3sAgM97AIDTewCAsUEXALChFwCzqesBsgHoAbUB7AG0EesB74wOANd7AICpxR8AqAEcAKsBEACqkR8ArdkTAKzREwCv2RcArgUTAKHxAgDbewCAo8kHAKLBAgClARgApGUHAKehGwCm+RsAqCkFAKldBQCqVQUAq20FAKx5BQCteQUArm0FAK9hBQCHewCA33sAgON7AIDnewCAgA0AAIGxAACCsQAA63sAgLiJBQC5iQUAup0FALuVBQC8uQUAvbkFAL5RBgC/UQYAsOUFALHtBQCy5QUAs/0FALTtBQC13QUAttUFALe9BQCj3QUA73sAgPN7AICEDAAA93sAgKb5BQCl8QUA+3sAgKspBQCqIQUAhpgAAIegAACvGQUArikFAK0pBQCsMQUA/3sAgLNhBgADfACAB3wAgLYhBgALfACAD3wAgLUBBgC6rQcAu40HABN8AIAXfACAvo0HAL9xBwC8lQcAvY0HAL65BQC/uQUAvLkFAL25BQC6uQUAu7kFALi5BQC5uQUAtkkFALdJBQC0fQUAtXUFALJ5BQCzeQUAsBUFALF9BQCuXQUAr20FAKxFBQCtXQUAqqUKAKtdBQCovQoAqa0KABt8AIAffACAI3wAgCd8AIArfACAL3wAgDN8AIA3fACAqA0HAKkdBwCqLQcAq0kHAKxNBwCtZQcArrEGAK+xBgA7fACAP3wAgEN8AIBHfACAS3wAgE98AIBTfACAV3wAgLhVBgC5XQYAulUGALtxBgC8NQYAvfEBAL7xAQC/8QEAsK0GALGNBgCyhQYAs50GALSNBgC1cQYAtnUGALdtBgCjpQQAgi0AAIEVAACAHQAAW3wAgKblBAClxQQAX3wAgKtJBQCqaQUAY3wAgGt8AICvtQUArkkFAK1JBQCsUQUAhmAcAIcIAwBvfACAs4UCAHN8AIC1gQIAtoECAHd8AIB7fACAf3wAgLoJAwC7CQMAvBkDAL0ZAwC+CQMAvwkDAKxVAgCtXQIArmECAK9hAgCoDQIAqVUCAKpRAgCrUQIAhKwDAIN8AICHfACAi3wAgIT8HQCPfACAk3wAgJd8AIC8cQMAvXEDAL5xAwC/cQMAuHEDALlxAwC6cQMAu3EDALSRAwC1kQMAtpEDALeRAwCwkQMAsZEDALKRAwCzkQMAm3wAgJ98AICjfACAp3wAgKt8AIDhpAEAr3wAgOOAAQC+aBwAs3wAgLd8AIDv2AYAu3wAgL98AIDDfACAx3wAgKOJAwCCLQAAgRUAAIAdAADLfACApo0DAKWNAwDPfACAqwUCAKoFAgDTfACA23wAgK8FAgCuBQIArRUCAKwVAgCGIBwAh8QdAN98AIDjfACA53wAgOt8AIDvfACA72wGAPN8AIDhbAcA93wAgON0BwD7fACA/3wAgAN9AIAHfQCAs5EBAAt9AIAPfQCAE30AgBd9AIC2sQEAtbkBABt9AIC7VQEAukkBAB99AIAjfQCAv/UAAL71AAC9RQEAvEUBAKNRHgDXfACAJ30AgCt9AIAvfQCApnEeAKV5HgAzfQCAq5UeAKqJHgA3fQCAO30AgK81HwCuNR8ArYUeAKyFHgCAbQAAgRUAAIIdAADv/BkAP30AgEN9AIBHfQCAS30AgIbAAACHrAMAT30AgFN9AIBXfQCA4SwcAFt9AIDjzBwAqK0eAKnNHgCq2R4Aq9EeAKzxHgCt8R4Arj0eAK81HgCE7AAAX30AgGN9AIBnfQCAa30AgG99AIBzfQCAd30AgLjRHwC53R8Auu0fALvlHwC84R8AveEfAL7hHwC/4R8AsE0eALFRHgCyUR4As1EeALTxHwC18R8AtvEfALfxHwCobR4AqY0eAKqFHgCrnR4ArIUeAK2NHgCuuR4Ar7UeAHt9AIB/fQCAg30AgId9AICLfQCAj30AgJN9AICXfQCAuJ0eALmtHgC6pR4Au0UBALxdAQC9RQEAvkUBAL91AQCw0R4AsdEeALLRHgCz0R4AtLUeALW9HgC2tR4At60eALMNHgCbfQCAn30AgKN9AICnfQCAtg0eALUNHgCrfQCAuxUeALoVHgCvfQCAs30AgL95HgC+cR4AvQUeALwFHgCCbQAAo0keAIBVAACBZQAApkkeAL6cAQC7fQCApUkeAKpRHgCrUR4Ah3wAAIZMAACuNR4Arz0eAKxBHgCtQR4AqF0CAKltAgCqZQIAq30CAKxpAgCtsQIArrECAK+xAgCE7AQAv30AgMN9AIDHfQCAy30AgM99AIDTfQCA130AgLhxAwC5cQMAunEDALtxAwC81QMAvd0DAL7VAwC/zQMAsNECALHRAgCy0QIAs9ECALRRAwC1UQMAtlEDALdRAwCz7QIA230AgN99AIC+gAQA430AgLYxAgC14QIA530AgLsVAgC6FQIA630AgO99AIC/lQMAvpUDAL0FAgC8BQIA830AgKOpAgD3fQCA+30AgKZ1AgD/fQCAA34AgKWlAgCqUQIAq1ECAAd+AIALfgCArtEDAK/RAwCsQQIArUECAKjZAgCpIQEAqiEBAKshAQCsIQEArSEBAK4hAQCvIQEAD34AgBN+AIAXfgCAviAEABt+AIAffgCAI34AgCt+AIC4jQEAuZEBALqRAQC7pQEAvL0BAL11AAC+fQAAv3UAALDlAQCx7QEAsvkBALPxAQC02QEAtdkBALa5AQC3tQEA4RgeAC9+AIDjKB8AM34AgIGlAACApQAAN34AgIKlAACGAAQAh/QFADt+AIA/fgCAQ34AgEd+AIDvYB4AS34AgE9+AIBTfgCAhfD0AVd+AIBbfgCA42QBAF9+AIDhpAEAY34AgO/IAABnfgCAa34AgGd8AICE/AUAb34AgHN+AICzKQYAJ34AgHd+AIB7fgCAf34AgLYhBgC1KQYAg34AgLupBgC6oQYAh34AgIt+AIC/nQYAvp0GAL2lBgC8rQYA4bQHAI9+AIDjeAQAk34AgIB9AACBEQAAghUAAJd+AICGwAAAh1gDAJt+AICffgCAo34AgKd+AIDvDAQAq34AgKOpBgCvfgCAs34AgLd+AIC7fgCApqEGAKWpBgC/fgCAqykGAKohBgDDfgCAx34AgK8dBgCuHQYArSUGAKwtBgDLfgCAs0kHAM9+AIDTfgCAtn0HANd+AIDbfgCAtXUHALpdBwC7JQcA334AgON+AIC+IQcAvy0HALw9BwC9MQcAqD0GAKmBBgCqhQYAq5UGAKy5BgCtuQYArqkGAK+pBgDnfgCA634AgO9+AIDzfgCA934AgIK5AACBsQAAgLkAALitBgC5vQYAurUGALtFAQC8XQEAvUUBAL5FAQC/dQEAsN0GALGlBgCyrQYAs6EGALShBgC1rQYAtpkGALeVBgCjDQYA+34AgP9+AIADfwCAhJgCAKY5BgClMQYAvpwBAKthBgCqGQYAhggAAId8AQCvaQYArmUGAK11BgCseQYAC38AgLO1AQAPfwCAE38AgLZVAQAXfwCAG38AgLWhAQC6cQEAu3kBAB9/AIAjfwCAvjEBAL89AQC8UQEAvVEBAKhpAgCpaQIAqnkCAKt5AgCsbQIArZECAK6RAgCvkQIAJ38AgCt/AIAvfwCAM38AgDd/AIA7fwCAP38AgEN/AIC4mQIAua0CALqlAgC7bQMAvHUDAL19AwC+dQMAv20DALDxAgCx+QIAssECALPBAgC0sQIAtb0CALa1AgC3qQIAR38AgEt/AIBPfwCAo/0CAFN/AICl6QIAph0CAFd/AIBbfwCAX38AgKo5AgCrMQIArBkCAK0ZAgCueQIAr3UCAGN/AIBnfwCAa38AgIQADACAGQAAgQkAAII5AABvfwCAc38AgHt/AIB/fwCAvuAMAIN/AICHfwCAhlgNAIcMAwCowQIAqc0CAKrFAgCr2QIArMkCAK39AgCu9QIArz0BAIt/AICPfwCAk38AgJd/AICbfwCAn38AgKN/AIC+MAwAuMUBALnNAQC62QEAu9EBALzxAQC98QEAvpkBAL+ZAQCwRQEAsU0BALJFAQCzXQEAtEUBALVNAQC2RQEAt/0BAOE4BgCnfwCA42wGAKt/AICvfwCAs38AgLd/AIC7fwCAhKgNAL9/AIDDfwCAx38AgL6wDwDLfwCA72wGAM9/AIDTfwCAt30AgNd/AIDbfwCA41AAAN9/AIDhoAEA438AgO+EAADrfwCAhyANAIZMDwCAPQAAgSEAAIIlAADvfwCAs80NAHd/AIDnfwCA838AgPd/AIC2/Q0AtcENAPt/AIC7CQ4AugEOAP9/AIADgACAvwkOAL4BDgC9CQ4AvBEOAAeAAIDjmAwAC4AAgOH8DwAPgACAE4AAgBeAAIAbgACAH4AAgCOAAIAngACAK4AAgC+AAIDvYAwAM4AAgDeAAICjTQ0AO4AAgD+AAIBDgACAR4AAgKZ9DQClQQ0AS4AAgKuJDgCqgQ4AT4AAgFOAAICviQ4AroEOAK2JDgCskQ4Agm0AALM1DgCAVQAAgWUAALb1DwCE3AMAV4AAgLX9DwC60Q8Au9EPAIYABACH3AAAvn0PAL9lDwC8wQ8AvXkPAKjlDwCp7Q8AqvkPAKv5DwCsMQ4ArTEOAK4xDgCvMQ4AW4AAgF+AAIBjgACAZ4AAgGuAAIBvgACAc4AAgHeAAIC43Q4AueEOALrhDgC74Q4AvOUOAL3pDgC+mQ4Av5UOALBRDgCxUQ4AslEOALPpDgC0/Q4AteUOALbtDgC35Q4Ao3EPAHuAAIB/gACAg4AAgIeAAICmsQ4ApbkOAIuAAICrlQ4AqpUOAI+AAICTgACAryEOAK45DgCtPQ4ArIUOAJeAAICzyQEAm4AAgJ+AAIC2+QEAo4AAgKeAAIC1wQEAuqkBALu1AQCrgACAr4AAgL6tAQC/lQEAvK0BAL2lAQCo5Q0AqfkNAKoFAgCrHQIArA0CAK09AgCuNQIAr10CALOAAIC3gACAu4AAgL+AAICAGQAAgRkAAIIFAADDgACAuC0CALk1AgC6MQIAuzECALzVAgC93QIAvtUCAL/NAgCwKQIAsTUCALI9AgCzNQIAtC0CALUVAgC2HQIAtxUCAMuAAICEnAIAz4AAgKOBAgDTgACApYkCAKaxAgDXgACAhiAEAIfUAwCq4QIAq/0CAKzlAgCt7QIAruUCAK/dAgC29QMAvkQDAIWM/QG1/QMA24AAgLP9AwDfgACA44AAgL59AwC/TQMAvGUDAL19AwC6dQMAu30DAOeAAIDrgACA74AAgPOAAICEBAIAoyUCAPeAAIClJQIApi0CAPuAAID/gACAA4EAgKqtAgCrpQIArL0CAK2lAgCupQIAr5UCAAeBAIALgQCAD4EAgBOBAIAXgQCA48ADABuBAIDhrAEAH4EAgO9YAwAjgQCAJ4EAgIANAACB5QAAgu0AACuBAIDhYA8A40ABAOM4DgDheA4AL4EAgDOBAIC+lAUAO4EAgIYABACHZAUAP4EAgEOBAIBHgQCA7/wOAO98DgBLgQCAs1EBAE+BAIAHfwCAU4EAgFeBAIC2DQEAtQkBAFuBAIC74QAAuhkBAF+BAIBjgQCAv9EAAL7pAAC96QAAvPkAAMeAAIA3gQCAZ4EAgGuBAIBvgQCAc4EAgHeBAIB7gQCAqKEGAKmtBgCquQYAq7EGAKzhBgCt7QYAruUGAK/FBgCwvQYAsUUHALJNBwCzXQcAtE0HALV1BwC2fQcAtx0HALglBwC5LQcAuiUHALs9BwC8KQcAvRUHAL4RBwC/EQcAoxEGAH+BAICDgQCAh4EAgIuBAICmTQYApUkGAI+BAICroQcAqlkGAJOBAICXgQCAr5EHAK6pBwCtqQcArLkHAIANAACBFQAAgh0AAJuBAICfgQCAo4EAgISUAwC+lAMAp4EAgKuBAICGyAAAh4wAAK+BAICzgQCAt4EAgLuBAIConQYAqa0GAKqlBgCrvQYArK0GAK3RBgCu1QYAr80GAL+BAIDDgQCAx4EAgMuBAIDPgQCA04EAgNeBAIDbgQCAuF0BALnBAQC6wQEAu8EBALzBAQC9yQEAvvEBAL/xAQCwvQYAsY0GALKFBgCzZQEAtH0BALVlAQC2bQEAt2UBALMtBgDfgQCA44EAgOeBAIDrgQCAtlEGALUlBgDvgQCAu0kGALp5BgDzgQCA94EAgL+hAQC+uQEAvbEBALxRBgD7gQCAo2kGAP+BAIADggCAphUGAAeCAIALggCApWEGAKo9BgCrDQYAD4IAgBOCAICu/QEAr+UBAKwVBgCt9QEAutUHALvdBwC4wQcAucEHAL4xBAC/MQQAvPEHAL3xBwCyrQcAs7UHALCtBwCxpQcAtp0HALf1BwC0pQcAtZUHAKppBwCraQcAqGkHAKlpBwCuaQcAr2kHAKxpBwCtaQcAgLkDAIGNAwCChQMAhKgDAIZQ/AGHCAMAvjQDABuCAICoZQIAqXUCAKp9AgCrdQIArG0CAK21AwCuvQMAr7UDAB+CAIAjggCAJ4IAgCuCAIAvggCAM4IAgDeCAIA7ggCAuFEDALlZAwC6YQMAu2EDALwRAwC9HQMAvhUDAL8JAwCwzQMAsdUDALLdAwCz1QMAtM0DALVxAwC2cQMAt3EDAD+CAIBDggCAs/0DAEeCAIC17QMAS4IAgE+CAIC2PQIAU4IAgFeCAIC7GQIAugECAL0JAgC8AQIAv70CAL4BAgBbggCAX4IAgITE/QG+wPwBY4IAgGeCAIBrggCA79wDAG+CAIDhlAEAc4IAgOMQAwB3ggCAgu0AAIHtAACA7QAA4TgGAOE8BwDjQAEA45QGAHuCAIB/ggCAg4IAgIuCAICGgPwBh+j9AY+CAICTggCAl4IAgJuCAIDvnAEA79wGAKM1AwCfggCAo4IAgKeCAICrggCApvUCAKUlAwCvggCAq9ECAKrJAgCzggCAt4IAgK91AgCuyQIArcECAKzJAgCHggCAu4IAgL+CAIDDggCA76T9AceCAIDLggCAz4IAgON4/QHTggCA4UD8AdeCAIDbggCA34IAgOOCAIDnggCAs+X+AYItAACBFQAAgB0AAOuCAIC25f4BtfX+Ae+CAIC7Yf8Butn+AfOCAICE5AMAv2n/Ab5h/wG9df8BvHn/Aaj9/gGpJf4Bqi3+Aasl/gGsPf4BrSX+Aa4t/gGvJf4BviwAAPeCAICGiAAAh+wAAPuCAID/ggCAA4MAgAeDAIC4gf8BuYH/AbqZ/wG7mf8BvIn/Ab21/wG+sf8Bv63/AbBd/gGx5f8Bsu3/AbPh/wG05f8Bte3/AbbZ/wG32f8Bo6X/AQuDAIAPgwCAE4MAgBeDAICmpf8BpbX/ARuDAICrIf4Bqpn/AR+DAIAjgwCAryn+Aa4h/gGtNf4BrDn+ASeDAICz6f4BK4MAgC+DAIC2lf4BM4MAgDeDAIC16f4BurH+Abu5/gE7gwCAP4MAgL51AQC/fQEAvJH+Ab2R/gGoHf4BqS3+Aaol/gGrPf4BrCX+Aa1R/gGuUf4Br1H+AUODAIBHgwCAS4MAgE+DAIBTgwCAV4MAgFuDAIBfgwCAuNkBALnZAQC67QEAu+EBALzhAQC94QEAvuEBAL/hAQCwMf4BsTn+AbIB/gGzAf4BtPUBALX9AQC29QEAt+kBAKOt/QFjgwCAvkwDAGuDAIBvgwCAptH9AaWt/QFzgwCAq/39Aar1/QF3gwCAe4MAgK85AgCuMQIArdX9AazV/QGA+QMAgfkDAIJNAACFdCAAf4MAgITYAwCE1AQAg4MAgIZABACHVAMAh4MAgIuDAICPgwCAk4MAgJeDAIC+8AUAqDECAKkxAgCqMQIAqzECAKyVAwCtnQMArpUDAK+NAwCbgwCAn4MAgKODAICngwCAhHwHAKuDAICvgwCAs4MAgLipAwC5qQMAumkDALtpAwC8eQMAvXkDAL5pAwC/aQMAsP0DALHNAwCyxQMAs60DALS5AwC1uQMAtq0DALelAwC3gwCAu4MAgL+DAIDDgwCAx4MAgMuDAIDv6AMAz4MAgOGQAQDTgwCA42wDANuDAICAJQAAgSkAAIIdAADfgwCAs/kDAOODAICGaAcAh1wFAOeDAIC2XQIAtV0CAOuDAIC7SQIAunkCAO+DAIDzgwCAvz0CAL49AgC9OQIAvFECAPeDAIDhPP4BvkAGAOPwAQD7gwCA/4MAgAOEAIAHhACAC4QAgA+EAIAThACAF4IAgBeEAIAbhACAH4QAgO/kAQAjhACAJ4QAgKNxAwArhACApdUCAC+EAIAzhACAptUCADeEAIA7hACAq8ECAKrxAgCtsQIArNkCAK+1AgCutQIA4dz8AdeDAIDjUAQA74gEAID1BwCBCQAAgj0AAD+EAICEJAEAQ4QAgEeEAIBLhACAT4QAgOFMBADv5BwA43QEALNdBgBThACAhgAMAIfgAwBXhACAtgUGALV1BgBbhACAuxEGALoJBgBfhACAY4QAgL/VBgC+1QYAvQEGALwJBgCojQYAqZUGAKqVBgCrpQYArL0GAK3FBgCuxQYAr/UGAGeEAIBrhACAb4QAgHOEAIB3hACAe4QAgH+EAICDhACAuHUGALl9BgC6dQYAu80HALzVBwC93QcAvtUHAL/NBwCwjQYAsZUGALKdBgCzlQYAtFEGALVRBgC2UQYAt1EGAKMdBwCPFewBh4QAgIuEAICPhACApkUHAKU1BwCThACAq1EHAKpJBwCXhACAm4QAgK+VBwCulQcArUEHAKxJBwCeRfkBn6X5AZyR/QGdTfkBmlX9AZtd/QGYBfEBmZX+AZal8gGXYfEBlG31AZU19QGS4ekBk4X2AZBV7AGRXekBsbEdALClHQCziRkAskEcALUBJAC09RkAn4QAgKOEAICnhACAgqkDAIGhAwCAaQAAohUFAKMFAgCgFQYAob0FAKHFAQCrhACAo80NAKLlAQClAQgApN0NAKfRCQCm2QkAqQEUAKilCACrxRQAqs0VAK3REQCsARAArwEcAK51EQCCEe8BgynvAa+EAICzhACAhuH1AYcR9gGEOeoBhY3qAYp59gGL4fEBvqQMALuEAICO+f0BjzH+AYw98gGNYfIBkkn+AZOd/gGHCAwAhmwMAJax+gGX+QUAlFn6AZVZ+gGaYQYAm8EGAL+EAIDDhACAx4QAgMuEAICcyQEAz4QAgKitBQCpuQUAqs0FAKvdBQCszQUArf0FAK71BQCvHQUA04QAgNeEAIDbhACA34QAgOOEAIDnhACA64QAgO+EAIC4dQUAuX0FALoJBQC7CQUAvB0FAL0BBQC+AQUAvz0FALBxBQCxcQUAsnEFALNxBQC0UQUAtVEFALZRBQC3TQUAs0UEAPOEAID3hACA+4QAgP+EAIC2fQQAtUUEAAOFAIC7tQQAurUEAAeFAIALhQCAv5UEAL6VBAC9pQQAvKUEAA+FAICjAQQAE4UAgBeFAICmOQQAG4UAgB+FAIClAQQAqvEEAKvxBAAjhQCAhOwNAK7RBACv0QQArOEEAK3hBADh0AYAhAwMAOMoBwC+AAwAK4UAgO9EAwCGuAwAhywNAC+FAIDjlAEAM4UAgOH8AQBngwCAN4UAgO/IBgA7hQCAP4UAgEOFAICzjQMAR4UAgLWNAwBLhQCAT4UAgLa1AwBThQCAV4UAgLtBAwC6SQMAvUEDALxZAwC/QQMAvkkDAKNFDAC3hACAJ4UAgFuFAIBfhQCApn0MAKVFDABjhQCAq4kMAKqBDABnhQCAa4UAgK+JDACugQwArYkMAKyRDACAFQ8AgR0PAIIhDwCzIQ4Ab4UAgLUhDgC2JQ4Ac4UAgHeFAIB7hQCAusEOALvBDgC8wQ4AvcEOAL7BDgC/wQ4AqK0OAKntDgCq5Q4Aq/0OAKzlDgCt6Q4ArjkOAK85DgB/hQCAg4UAgIeFAICLhQCAgB0AAIEJAACCvQEAj4UAgLjNDwC51Q8AutUPALvlDwC8/Q8AvZUPAL6RDwC/kQ8AsEkOALFJDgCyWQ4As1kOALRJDgC1SQ4Atv0PALf1DwCjbQ8Ak4UAgL6EAQCbhQCAn4UAgKZpDwClbQ8Ao4UAgKuNDwCqjQ8AhogAAIdsAQCvjQ8Aro0PAK2NDwCsjQ8Ap4UAgLPtDgCrhQCAr4UAgLaRDgCzhQCAt4UAgLXhDgC6tQ4Au70OALuFAIC/hQCAvn0BAL9lAQC8mQ4AvZkOAKgRDgCpJQ4AqiEOAKs5DgCsLQ4ArVUOAK5dDgCvUQ4AhKgAAMOFAIDHhQCAy4UAgM+FAIDThQCA14UAgNuFAIC47QEAuZUBALqVAQC7rQEAvLUBAL11AQC+fQEAv3UBALA1DgCxPQ4AsgkOALMJDgC0/QEAteUBALblAQC31QEAo6kNAN+FAIDjhQCA54UAgOuFAICm1Q0ApaUNAO+FAICr+Q0AqvENAPOFAID3hQCAryECAK45AgCt3Q0ArN0NAIANAACBFQAAgh0AAPuFAID/hQCAA4YAgIeQAwCGfAQAvuwEAAuGAIAPhgCAE4YAgBeGAIAbhgCAH4YAgCOGAICyLQ4AszUOALAtDgCxJQ4Ati0OALedDwC0LQ4AtSUOALq9DwC7jQ8AuKUPALm9DwC+LQ8AvxUPALyVDwC9JQ8AJ4YAgCuGAIAvhgCAM4YAgDeGAIA7hgCAP4YAgEOGAICqpQ4Aq7UOAKjFDgCp3Q4Arp0OAK9VDgCspQ4ArZUOAKgNAgCpFQIAqhUCAKtNAgCsWQIArVkCAK5NAgCvRQIAhKgFAEeGAIBLhgCAT4YAgIS4BABThgCAV4YAgFuGAIC4/QIAuUEBALpBAQC7QQEAvEEBAL1JAQC+cQEAv3EBALAJAgCxCQIAss0CALPFAgC03QIAtcUCALbNAgC3xQIA4dQPAOMQDgDj9A4A4QwOAF+GAIBjhgCAZ4YAgGuGAIBvhgCAc4YAgL4kBAB7hgCA7AAAAO9EAADvzA4Af4YAgIJlAACz2QIAgFUAAIFtAAC2nQIAg4YAgIeGAIC1lQIAuokCALuJAgCGqAQAh+AEAL5dAgC/RQIAvF0CAL1VAgCjHQUAB4YAgHeGAICLhgCAj4YAgKZZBQClUQUAk4YAgKtNBQCqTQUAl4YAgJuGAICvgQUArpkFAK2RBQCsmQUAn4YAgLMpBgCjhgCAp4YAgLYpBgCrhgCAr4YAgLUpBgC6pQYAu60GALOGAIC3hgCAvqUGAL+tBgC8tQYAva0GAKjlBgCp7QYAquUGAKv9BgCs5QYAre0GAK7lBgCvXQYAu4YAgL+GAIDDhgCAx4YAgMuGAIDPhgCA04YAgNeGAIC46QcAuekHALr9BwC79QcAvO0HAL1FBwC+TQcAv0UHALAlBgCxLQYAsiUGALM9BgC0JQYAtS0GALYlBgC32QcAo20HAIItAACBFQAAgB0AANuGAICmbQcApW0HAN+GAICr6QcAquEHAOOGAIC+oAEAr+kHAK7hBwCt6QcArPEHAOeGAICzkQYAhugAAIcsAQC2QQEA64YAgO+GAIC1UQEAuk0BALslAQDzhgCA94YAgL4lAQC/LQEAvDEBAL0xAQCwrQEAscUBALLBAQCzwQEAtMUBALXNAQC28QEAt/EBALgBAQC5AQEAugEBALsBAQC8AQEAvQEBAL4BAQC/AQEA+4YAgP+GAIADhwCAB4cAgJeFAIALhwCAD4cAgBOHAICoTQYAqVkGAKo9BgCrNQYArP0BAK3lAQCu5QEAr9UBAKPVBQAXhwCAG4cAgB+HAIAjhwCApgUCAKUVAgAnhwCAq2ECAKoJAgArhwCAL4cAgK9pAgCuYQIArXUCAKx1AgAzhwCAN4cAgDuHAIA/hwCAQ4cAgOFkBQBHhwCA4+wFAIARAACBEQAAghEAAO/0BgBLhwCAT4cAgFOHAIC+MAMAhMQCAFuHAICz4QMAhMAcALVRAwBfhwCAY4cAgLZZAwBnhwCAa4cAgLtxAwC6eQMAvbUAALxpAwC/tQAAvrUAAG+HAIDhlAEAc4cAgONcAgCGcBwAh0QDAHeHAIB7hwCAf4cAgIOHAICHhwCAi4cAgI+HAICThwCAl4cAgO94AgCoVQIAqV0CAKphAgCrYQIArNECAK3RAgCu0QIAr9ECAJuHAICfhwCAo4cAgKeHAICrhwCAr4cAgLOHAIC3hwCAuGkBALlpAQC6CQEAuwkBALwZAQC9GQEAvgkBAL8FAQCwtQIAsb0CALK1AgCzaQEAtHkBALV5AQC2aQEAt2EBAOHEBwDjpAYA47gGAOF8BgCADQAAgTUAAII9AAC7hwCAv4cAgMOHAIC+4B0Ay4cAgM+HAIDvYAAA7+gGANOHAICjqQIA14cAgNuHAIDfhwCA44cAgKYRAgClGQIA54cAgKs5AgCqMQIAhkgcAIfMHACv/QEArv0BAK39AQCsIQIAqIUeAKmRHgCqkR4Aq60eAKy1HgCt1R4ArtEeAK/FHgDHhwCA64cAgO+HAIDzhwCA94cAgPuHAID/hwCAA4gAgLhhHwC5YR8AumEfALthHwC8YR8AvWEfAL5hHwC/YR8AsL0eALGFHgCyjR4As4UeALSdHgC1hR4Ato0eALeFHgCzGR4AB4gAgAuIAIAPiACAE4gAgLZVHgC1PR4AF4gAgLtBHgC6eR4AG4gAgB+IAIC/QR4AvlkeAL1RHgC8WR4AI4gAgKNdHgAniACAK4gAgKYRHgAviACAM4gAgKV5HgCqPR4AqwUeAISkAwC+qAMArh0eAK8FHgCsHR4ArRUeAKitHgCptR4AqrUeAKvJHgCs2R4ArdkeAK7JHgCvwR4AgO0BAIHxAQCC8QEAN4gAgIaQAACHdAEAO4gAgD+IAIC4yQEAuckBALrZAQC70QEAvPkBAL35AQC+mQEAv5UBALBFAQCxTQEAskUBALNdAQC0RQEAtU0BALZFAQC3+QEAsz0eAEOIAIBHiACAS4gAgE+IAIC2WR4AtVEeAFOIAIC7iQEAuoEBAFeIAIBbiACAv4kBAL6BAQC9iQEAvJEBAF+IAIBjiACAo3UeAGeIAIClGR4Aa4gAgG+IAICmER4AV4cAgHOIAICrwQEAqskBAK3BAQCs2QEAr8EBAK7JAQB3iACAe4gAgH+IAICDiACAh4gAgIQYAgCLiACAj4gAgJOIAICXiACAm4gAgJ+IAICjiACAq4gAgK+IAIC+cAMAgGkAAIFpAACCeQAAhAAEAIbwBACHdAMAs4gAgO8MHwC3iACA4aweALuIAIDj8B4Av4gAgMOIAIDHiACAy4gAgM+IAIDTiACA14gAgNuIAIDvVAIA34gAgOOIAIDniACA46QCAOuIAIDhgAEA74gAgPOIAID3iACA+4gAgP+IAICzRQMAA4kAgAeJAIALiQCAD4kAgLZFAwC1VQMAE4kAgLshAwC6SQMAvqAEABuJAIC/KQMAviEDAL01AwC8OQMAqDkCAKk5AgCqjQIAq4UCAKydAgCthQIAroUCAK+1AgCA7QEAgfUBAIL1AQAfiQCAhpAEAIcEBQAjiQCAJ4kAgLhFAQC5TQEAukUBALtdAQC8SQEAvUkBAL55AQC/eQEAsM0CALGlAgCyrQIAs6ECALSlAgC1rQIAtp0CALd9AQAriQCAL4kAgDOJAIA3iQCAO4kAgD+JAIBDiQCA74gBAITsBADhVB4AR4kAgONUAQBLiQCAT4kAgFOJAIBXiQCAo0UCAFuJAIBfiQCAY4kAgGeJAICmRQIApVUCAGuJAICrIQIAqkkCAG+JAIBziQCArykCAK4hAgCtNQIArDkCAKg1BgCpPQYAqlEGAKttBgCseQYArWUGAK5tBgCvZQYAF4kAgHeJAIB7iQCAf4kAgIAZAACBGQAAggUAAIOJAIC45QYAuekGALr5BgC7+QYAvOkGAL3pBgC+nQYAv5UGALAdBgCx5QYAsu0GALPlBgC0/QYAteEGALbhBgC34QYAs9kGAL7QAwCHiQCAi4kAgI+JAIC25QYAtfEGAJOJAIC7IQYAutkGAIaYAACHeAMAvyUGAL45BgC9MQYAvDkGAJeJAICjnQYAm4kAgJ+JAICmoQYAo4kAgKeJAICltQYAqp0GAKtlBgCriQCAr4kAgK59BgCvYQYArH0GAK11BgCo7QcAqSkGAKoxBgCrMQYArJEGAK2RBgCukQYAr5EGALOJAIC3iQCAu4kAgL+JAIDDiQCAx4kAgMuJAIDPiQCAuIUGALmNBgC6hQYAu50GALyNBgC9vQYAvrUGAL95AQCw8QYAsfEGALLxBgCzxQYAtMEGALXBBgC2wQYAt8EGALO5BgDTiQCA14kAgNuJAIDfiQCAthEGALUZBgDjiQCAuzUGALo1BgDniQCA64kAgL8FBgC+BQYAvREGALwlBgClQQYA74kAgPOJAICmSQYAgRUAAIB5AACj4QYAghUAAK1JBgCsfQYAr10GAK5dBgCENAEAp4gAgKttBgCqbQYAvswDAPuJAICzlQIA/4kAgLXZAgADigCAB4oAgLbRAgCGgAwAhzgDALvFAgC6xQIAvRUDALwVAwC/FQMAvhUDAAuKAIAPigCA71gGAIRAAwATigCAF4oAgBuKAIAfigCAI4oAgCeKAIArigCAL4oAgOE4BgAzigCA4yQGAL5wDACsSQIArUkCAK5dAgCvVQIAqB0CAKkFAgCqBQIAq10CAISoDAA3igCAO4oAgD+KAIC+vA0AQ4oAgEeKAIBLigCAvE0DAL1VAwC+VQMAv2UDALjpAwC56QMAul0DALtVAwC0yQMAtckDALbZAwC32QMAsBkCALEZAgCy2QMAs9kDAE+KAIDj5AAAU4oAgOG8AQBXigCAgj0AAIE9AACAPQAAW4oAgF+KAIBjigCAa4oAgG+KAIDvzAMAc4oAgHeKAICj3QMAe4oAgIboDACHYA0Af4oAgKaZAwClkQMAg4oAgKuNAwCqjQMAh4oAgIuKAICvXQIArl0CAK1dAgCsXQIAj4oAgJOKAICXigCAm4oAgJ+KAICjigCAp4oAgO/gAQCEvAwA4YwGAKuKAIDjHAYAr4oAgLOKAIC3igCAu4oAgLPVAQC/igCAw4oAgMeKAIDLigCAtpEBALWZAQDPigCAu70BALq9AQDTigCA24oAgL+dAQC+nQEAvZ0BALydAQCoBQ4AqQkOAKodDgCrFQ4ArFEOAK1RDgCuSQ4Ar0kOAGeKAICCzQ8AgfUPAID9DwDXigCA34oAgIYcAACHsAMAuOkOALnpDgC6/Q4Au/UOALztDgC9VQ8AvlEPAL9NDwCwOQ4AsTkOALIJDgCzCQ4AtBkOALUZDgC2DQ4At9kOAKOVDgDjigCA54oAgOuKAIDvigCAptEOAKXZDgDzigCAq/0OAKr9DgD3igCA+4oAgK/dDgCu3Q4Ard0OAKzdDgD/igCAs/0PAAOLAIAHiwCAtoEPAAuLAIAPiwCAtZkPALqNDwC7ZQ8AE4sAgBeLAIC+fQ8Av2UPALx9DwC9dQ8AqC0OAKk1DgCqMQ4AqzEOAKxVDgCtRQ4ArkUOAK91DgAbiwCAH4sAgCOLAIAniwCAK4sAgC+LAIAziwCAN4sAgLjpDgC59Q4Auv0OALv1DgC87Q4AvZEOAL6RDgC/kQ4AsA0OALHlDgCy7Q4As+UOALT9DgC15Q4Atu0OALflDgCjuQ4Agi0AAIEVAACAHQAAO4sAgKbFDgCl3Q4AP4sAgKshDgCqyQ4AQ4sAgL4sAQCvIQ4ArjkOAK0xDgCsOQ4AS4sAgLZVAQC1RQEAR4sAgLNVAQBPiwCAhngAAIdcAAC/OQEAvjEBAL0lAQC8JQEAuzEBALpZAQD3iQCAU4sAgFeLAIBbiwCAhAQDAKOJAgBfiwCApZkCAKaJAgBjiwCAvyg5AGeLAICqhQIAq+0CAKz5AgCt+QIAru0CAK/lAgDjWAIA78AOAOGIAQBriwCAb4sAgHOLAIB3iwCAe4sAgH+LAICDiwCAh4sAgIuLAIDvKAIA4ygOAI+LAIDhRA4AqbUCAKhpDQCrAQIAqgkCAK0BAgCsGQIArzECAK4BAgC+AAQAk4sAgJeLAICbiwCAn4sAgKOLAICniwCAq4sAgLnlAwC45QMAu+UDALrlAwC95QMAvOUDAL/lAwC+5QMAsSECALBJAgCzJQIAsiUCALUpAgC0IQIAtxUCALYVAgCowQIAqdECAKr1AgCrDQEArBUBAK0FAQCuBQEArzkBAK+LAICziwCAu4sAgL+LAIDDiwCAx4sAgMuLAIDPiwCAuC0BALk9AQC67QEAu+UBALz9AQC95QEAvu0BAL/lAQCwLQEAsTUBALI9AQCzNQEAtC0BALUVAQC2HQEAtxUBAIA9AQCBpQAAgq0AAO/YAACGsAUAh9gFANOLAIDv1A8AhGwEAOH0DgDXiwCA4xwPANuLAIDhlAEA34sAgOMMDgCzPQIA44sAgOeLAIDriwCA74sAgLbFAQC13QEA84sAgLuxAQC6qQEA94sAgPuLAIC/kQEAvqkBAL2hAQC8qQEAt4sAgP+LAICqRQYAq10GAKxFBgCtTQYArkUGAK99BgADjACAB4wAgAuMAICj0QUAD4wAgKUxBgCmKQYAE4wAgBeMAICCHQAAgR0AAIAdAAAbjACAH4wAgCOMAIC+lAMAJ4wAgCuMAICGSAMAh8wDAC+MAIAzjACAN4wAgDuMAICoqQcAqakHAKq5BwCruQcArKkHAK2pBwCuAQcArzUHAD+MAIBDjACAR4wAgEuMAIBPjACAU4wAgFeMAIBbjACAuC0HALnBAAC66QAAu+kAALz5AAC95QAAvuUAAL+dAACwUQcAsV0HALItBwCzJQcAtD0HALUlBwC2JQcAtxUHALMxBgBfjACAY4wAgGeMAIBrjACAtikGALUhBgBvjACAu5kGALqVBgBzjACAd4wAgL/hBgC++QYAvfEGALz5BgB7jACAo3UGAH+MAICDjACApm0GAIeMAICLjACApWUGAKrRBgCr3QYAj4wAgJOMAICuvQYAr6UGAKy9BgCttQYAqOUBAKn1AQCq/QEAq/UBAKztAQCtNQEArj0BAK81AQCA+QAAgc0AAILFAACEYAEAvngBAJuMAICHrAAAhpABALjRAAC52QAAuuEAALvhAAC8kQAAvZ0AAL6VAAC/iQAAsE0BALFVAQCyXQEAs1UBALRNAQC18QAAtvEAALfxAACzdQIAn4wAgKOMAICnjACAq4wAgLa1AgC1ZQIAr4wAgLuRAgC6iQIAs4wAgLeMAIC/NQMAvokCAL2BAgC8iQIAu4wAgKMxAgC/jACAhMADAKbxAgDDjACAx4wAgKUhAgCqzQIAq9UCAMuMAIDPjACArs0CAK9xAwCszQIArcUCAKuNAACqjQAAqY0AAKg5AwCvvQAArr0AAK2FAACsjQAAqgAAAKsAAADTjACA14wAgNuMAIDfjACA44wAgOeMAIC7fQAAun0AALl9AAC4fQAAv90BAL7dAQC93QEAvN0BALO5AACysQAAsaEAALCtAAC3XQAAtl0AALWVAAC0lQAA64wAgO+MAIDzjACA94wAgIE1AACADQAA+4wAgII1AAC+rD0A/4wAgAONAICFaD0AC40AgA+NAICGODwAh8ACALNJAQATjQCA0AAAABeNAIAbjQCAtkkBALVJAQAfjQCAuykBALolAQAjjQCAJ40AgL8dAQC+HQEAvSEBALwpAQDjNDYA4QwGAOGwAgDjPAYAK40AgC+NAIAzjQCAN40AgIQsPwC+oD8AO40AgD+NAIDvfDcAQ40AgEeNAIDvGAEAS40AgE+NAICGaD4Ah8w/AFONAIBXjQCAW40AgO+UAABfjQCA4ZQBAGONAIDjUAAAZ40AgILpPwCB6T8AgPE/AKMJPgCPASQAB40AgGuNAIBvjQCApgk+AKUJPgBzjQCAq2k+AKplPgB3jQCAe40AgK9dPgCuXT4ArWE+AKxpPgCeYTgAn3U4AJzBNACdtTkAmqU1AJt1NACYeTAAmXExAJYhLQCXhTEAlG0sAJVlLACSeSgAk6UtAJBRJACReSgAsQ0UALAFFACzARgAslUUALV5GAC0tRgAf40AgIONAICHjQCAi40AgI+NAICTjQCAotE8AKMlAQCgdTkAob08AKHJAACXjQCAowEEAKLlAAClHQQApPUEAKf5CACmAQgAqQEMAKhtCACrzQwAqs0MAK3REACsARAAr9URAK7ZEACCBSUAgy0lAJuNAICfjQCAhsEsAIcRLQCEHSkAhRUpAIopLQCLZSwAo40AgKeNAICOHTAAj8E0AIzZMACNHTEAkmE1AJPNNQCrjQCAr40AgJZhOQCXmTgAlKE4AJV9OQCaYT0AmwU9ALONAIC3jQCAu40AgL+NAICc6QAAw40AgMeNAIDLjQCAz40AgNONAICXjACA140AgNuNAIDfjQCAqJE+AKmRPgCq7T4Aq+E+AKzhPgCt6T4ArtE+AK/RPgCwUT4AsVE+ALJRPgCzUT4AtHk+ALV5PgC2bT4At2U+ALghPgC5IT4Aujk+ALs5PgC8KT4AvRU+AL4RPgC/DT4AgJkDAIGZAwCCBQAA440AgL5UAwDhsD0A640AgONAPgCEOAIA740AgPONAIDv9D8A940AgPuNAICGmAQAhxwDALMFPQCECAQA/40AgAOOAIAHjgCAtgk9ALUJPQALjgCAu/U9ALr1PQAPjgCAE44AgL/dPQC+3T0AveU9ALzlPQAXjgCAG44AgKPNPQC+xAQApcE9AB+OAIAjjgCApsE9ACeOAIArjgCAqz09AKo9PQCtLT0ArC09AK8VPQCuFT0AtmkCAC+OAIAzjgCAtWkCADeOAICzSQIAO44AgD+OAIC+qQMAv6kDALzBAwC9wQMAuvkDALv5AwBDjgCAR44AgKgtAwCpnQMAqpUDAKutAwCstQMArb0DAK61AwCv2QMAgA0AAIEVAACCHQAAS44AgE+OAIBTjgCAh7QFAIacBAC4MQIAuTECALo1AgC7zQIAvNUCAL3dAgC+1QIAv8kCALBpAgCxaQIAskECALNBAgC0OQIAtTkCALYRAgC3EQIAW44AgOM0PgBfjgCA4aw+AGOOAIDvfAMAZ44AgGuOAIBvjgCA45QDAHOOAIDhfD4Ad44AgO/oPgB7jgCAf44AgIOOAICHjgCAo1UDAIuOAICldQMAj44AgJOOAICmdQMAl44AgJuOAICr5QIAquUCAK3dAgCs3QIAr7UCAK61AgCoGQYAqSEGAKohBgCrPQYArCUGAK1dBgCuVQYAr00GAFeOAICfjgCAo44AgKeOAICrjgCAr44AgLOOAIC3jgCAuOUGALmBBgC6gQYAu50GALyJBgC9iQYAvqEGAL+hBgCwPQYAsQ0GALIFBgCz7QYAtPUGALXhBgC24QYAt90GALOpBgCCLQAAgRUAAIAdAAC7jgCAtt0GALWtBgC/jgCAu8kGALr5BgDDjgCAhOADAL8lBgC+MQYAvTkGALzRBgC+iAMAo+0GAOeNAIDHjgCAppkGAMuOAIDPjgCApekGAKq9BgCrjQYAhkgAAIdsAACudQYAr2EGAKyVBgCtfQYAqIEGAKmNBgCqmQYAq5UGAKyNBgCttQYArrEGAK+tBgDTjgCA144AgNuOAIDfjgCA444AgOeOAIDrjgCA744AgLilBgC5YQEAumEBALthAQC8YQEAvWEBAL5hAQC/YQEAsNkGALHZBgCyqQYAs6kGALS9BgC1oQYAtqEGALedBgCzEQYA844AgPeOAID7jgCA/44AgLY1BgC1BQYAA48AgLsdBgC6HQYAB48AgAuPAIC/ZQYAvnkGAL19BgC8fQYAD48AgKNVBgATjwCAF48AgKZxBgAbjwCAH48AgKVBBgCqWQYAq1kGACOPAIAnjwCArj0GAK8hBgCsOQYArTkGAKjVAgCp3QIAqikDAKspAwCsOQMArTkDAK4pAwCvKQMAK48AgC+PAIAzjwCAO48AgD+PAIBDjwCAvrgDAEePAIC47QMAuYUDALqBAwC7gQMAvIUDAL2NAwC+sQMAv7EDALBZAwCxWQMAsu0DALPlAwC0/QMAteUDALblAwC31QMAgKEAAIGhAACCoQAAvoAMAEuPAICEmAIAT48AgFOPAICGAAwAh/QDAFePAIBbjwCAX48AgGOPAIBnjwCAhLADALPhAwBrjwCAb48AgHOPAIB3jwCAtvkDALXxAwB7jwCAu90DALrdAwB/jwCAg48AgL9hAwC+eQMAvXEDALx5AwCHjwCAi48AgI+PAICjLQIAk48AgKU9AgCmNQIAl48AgJuPAICfjwCAqhECAKsRAgCstQIArb0CAK61AgCvrQIA48QDAOMQBwDhuAEA4WwHAIBxAACBcQAAggUAAKOPAICGwAwAh1QNAKuPAICvjwCA77ADAO8ABwCzjwCAt48AgLuPAIC/jwCAw48AgMePAIDLjwCAz48AgNOPAIDvpAEAhKANAOGABgDXjwCA4xABANuPAIDfjwCA448AgOePAICz9QEA648AgO+PAIDzjwCA948AgLZNAQC1SQEA+48AgLtRAQC6SQEA/48AgAOQAIC/OQEAvjEBAL1BAQC8SQEAqC0OAKk1DgCqPQ4AqzEOAKyBDgCtjQ4AroUOAK+1DgCnjwCAB5AAgAuQAIAPkACAgBkAAIEZAACCBQAAE5AAgLidDgC5rQ4AuqUOALtNDwC8VQ8AvV0PAL5JDwC/QQ8AsM0OALHVDgCy3Q4As9UOALS1DgC1vQ4AtrUOALetDgCjtQ4AvogDABeQAIAbkACAH5AAgKYNDgClCQ4AI5AAgKsRDgCqCQ4AhggAAIdsAwCveQ4ArnEOAK0BDgCsCQ4AJ5AAgCuQAIAvkACAs7UPADOQAIC1VQ8Atl0PADePAIA3kACAO5AAgLp5DwC7eQ8AvGkPAL1dDwC+SQ8Av0kPAKhpDgCpaQ4AqnEOAKtxDgCskQ4ArZEOAK6RDgCvkQ4AP5AAgEOQAIBHkACAS5AAgE+QAIBTkACAV5AAgFuQAIC4hQ4AuY0OALqFDgC7nQ4AvI0OAL29DgC+tQ4Av3kBALDxDgCx8Q4AsvEOALPFDgC0wQ4AtcEOALbBDgC3wQ4Ao/kOAF+QAIBjkACAZ5AAgGuQAICmEQ4ApRkOAG+QAICrNQ4AqjUOAHOQAIB3kACArwUOAK4FDgCtEQ4ArCUOAIANAACBFQAAgh0AAHuQAIB/kACAg5AAgISUAQC+lAEAhkAHAIf0AACLkACAj5AAgJOQAICXkACAm5AAgJ+QAICojQIAqZUCAKqVAgCrzQIArNUCAK3dAgCuyQIAr/0CAKOQAICnkACAq5AAgK+QAIC/ABQAs5AAgLeQAIC7kACAuH0DALnBAwC6wQMAu8EDALzBAwC9yQMAvvEDAL/xAwCwhQIAsUUDALJNAwCzRQMAtF0DALVFAwC2TQMAt0UDALMdAgC/kACAw5AAgMeQAIDLkACAtl0CALVdAgDPkACAu4EDALpBAgDTkACA15AAgL+BAwC+mQMAvZEDALyZAwDbkACAo1kCAN+QAIDjkACAphkCAOeQAIDrkACApRkCAKoFAgCrxQMA75AAgPOQAICu3QMAr8UDAKzdAwCt1QMA+5AAgOPMAACEBAIA4bwBAIDJAQCB/QEAgvUBAL4QBQD/kACAvigEAAORAIAHkQCAC5EAgO8QAAAPkQCAE5EAgIbgBACH9AIAF5EAgBuRAIDj/A8AH5EAgOHgDwAjkQCA7xQPACeRAIArkQCAL5EAgDORAIA3kQCAO5EAgD+RAIBDkQCAR5EAgEuRAIBPkQCAU5EAgFeRAIBbkQCA7+ABAIUEEgDh3A4AX5EAgOMcDgCAKQAAgR0AAIIFAABjkQCAszECAGuRAICEzAUAb5EAgHORAIC2KQIAtSECAHeRAIC7zQEAus0BAHuRAIB/kQCAv3UBAL7JAQC9wQEAvMkBAKjpBQCp6QUAqvkFAKv5BQCs6QUArekFAK45BgCvOQYA95AAgGeRAICGiAAAhwADAIORAICHkQCAi5EAgI+RAIC40QYAudkGALrhBgC74QYAvJEGAL2dBgC+lQYAv4kGALBJBgCxSQYAsl0GALNVBgC0TQYAtfEGALbxBgC38QYAo3EFAJORAICXkQCAm5EAgJ+RAICmaQUApWEFAKORAICrjQYAqo0GAKeRAICrkQCArzUGAK6JBgCtgQYArIkGAK+RAICzkQCAs+EHALeRAIC14QcAu5EAgL+RAIC25QcAh5AAgMORAIC7vQcAuqEHAL2VBwC8qQcAv5UHAL6VBwCoAQYAqSUGAKohBgCrIQYArCEGAK0tBgCuJQYAr1UGAMeRAICCHQAAgR0AAIAdAADLkQCAz5EAgNORAIC+MAEAuDkGALk5BgC6yQYAu8kGALzZBgC92QYAvskGAL/JBgCwLQYAsTEGALI1BgCzCQYAtBkGALUZBgC2CQYAtwkGAKOpBgCEjAIAhigfAIdEAQDbkQCApq0GAKWpBgDfkQCAq/UGAKrpBgDjkQCA55EAgK/dBgCu3QYArd0GAKzhBgDrkQCAsxUGAO+RAIDzkQCAtj0GAPeRAID7kQCAtTUGALrZAQC72QEA/5EAgAOSAIC+fQEAv2UBALx9AQC9dQEAqMUFAKnJBQCq2QUAq9EFAKz5BQCt+QUArikCAK8pAgAHkgCAC5IAgA+SAIATkgCAjAAAABeSAIAbkgCAH5IAgLjtAgC5hQIAuo0CALuBAgC8hQIAvY0CAL69AgC/fQMAsFkCALFZAgCy7QIAs+UCALT9AgC15QIAtuUCALfVAgCjUQUAI5IAgCeSAIArkgCAL5IAgKZ5BQClcQUAM5IAgKudAgCqnQIAN5IAgDuSAICvIQIArjkCAK0xAgCsOQIAghEAAD+SAICAZQAAgQkAAEOSAIC+mAMAS5IAgE+SAICEJAMAU5IAgIdoAwCGjBwAV5IAgFuSAIBfkgCAY5IAgGeSAIBrkgCAs6ECAITAHAC10QIAb5IAgHOSAIC21QIAd5IAgHuSAIC7wQIAuvUCAL0RAQC82QIAvxEBAL4ZAQB/kgCAg5IAgIeSAICLkgCAj5IAgJOSAICXkgCA77gGAJuSAIDhnAQAn5IAgON0BgCjkgCAp5IAgKuSAICvkgCAgPkAAIH5AACCBQAAs5IAgL5YHACEWB8A71wAAO9ABgDhkAEA4fwGAOM8AADjdAYAu5IAgL+SAICGmBwAh/QcAKNpAgC+DB8Aw5IAgMeSAIDLkgCAph0CAKUZAgDPkgCAqwkCAKo9AgDTkgCA15IAgK/ZAQCu0QEArdkBAKwRAgCokR0AqZkdAKqhHQCroR0ArNEdAK3dHQCu1R0Ar8kdAEeSAIC3kgCA25IAgN+SAIDjkgCA55IAgOuSAIDvkgCAuHkeALl5HgC6zR4Au8UeALzdHgC9xR4AvsUeAL/1HgCwuR0AsY0dALKFHQCzTR4AtFUeALVdHgC2VR4At0keALjNHwC51R8Aut0fALvVHwC88R8Avf0fAL7pHwC/6R8AsKUfALGxHwCysR8As40fALSVHwC19R8Atv0fALf1HwCoGR4AqRkeAKotHgCrPR4ArCUeAK0tHgCuJR4Ar90fAPOSAID3kgCA+5IAgP+SAIADkwCA15EAgAeTAIALkwCAs+UfAA+TAIATkwCAF5MAgBuTAIC27R8Ate0fAB+TAIC7NR4AuiEeACOTAIAnkwCAv3EeAL4RHgC9GR4AvCUeAIJpAACjoR8AgFkAAIFRAACmqR8AK5MAgC+TAIClqR8AqmUeAKtxHgCGAAQAh+wBAK5VHgCvNR4ArGEeAK1dHgCoMR4AqTEeAKpBHgCrQR4ArEEeAK1JHgCucR4Ar3EeADOTAIA3kwCAO5MAgD+TAIBDkwCAR5MAgEuTAIBPkwCAuCkBALkpAQC6OQEAuzUBALwtAQC90QAAvtEAAL/RAACwyQEAsckBALLZAQCz2QEAtMkBALXJAQC2GQEAtxkBALPJHQBTkwCAV5MAgFuTAIBfkwCAtskdALXJHQBjkwCAuw0CALoNAgBnkwCAa5MAgL8NAgC+DQIAvQ0CALwNAgBvkwCAo40dAHOTAIB3kwCApo0dAHuTAIB/kwCApY0dAKpJAgCrSQIAg5MAgIeTAICuSQIAr0kCAKxJAgCtSQIAgA0AAIERAACCEQAAi5MAgO/MAgCPkwCAk5MAgISQAgDjLAIAvigDAOHYAQCbkwCAhhAEAIfUAwCfkwCAo5MAgLNhAwCnkwCAq5MAgK+TAICzkwCAtnkDALVxAwC3kwCAu10DALpdAwC7kwCAv5MAgL/hAAC++QAAvfEAALz5AACjoQIAw5MAgMeTAIDLkwCAz5MAgKa5AgClsQIA05MAgKudAgCqnQIA15MAgNuTAICvIQEArjkBAK0xAQCsOQEA35MAgOOTAIDvZB8A55MAgOuTAIDvkwCA85MAgPeTAICADQAAgREAAIIVAAD7kwCA4eAcAP+TAIDjiB8AA5QAgISAAgC+jAUAh0gFAIYsBAALlACAD5QAgO+kHgDv9B4A4QAeAOFQHwDjLB4A47AeABOUAIAXlACAG5QAgB+UAIAjlACAJ5QAgISEBACzcQEAK5QAgLUdAQC2FQEAL5QAgDOUAIA3lACAugEBALsBAQC89QAAvf0AAL71AAC/7QAAqK0GAKm9BgCqtQYAq8kGAKzZBgCt2QYArskGAK/BBgA7lACAP5QAgEOUAIBHlACAS5QAgE+UAIBTlACAV5QAgLhtBwC5BQcAug0HALsBBwC8AQcAvQEHAL4BBwC/AQcAsIkGALGJBgCybQcAs2UHALR9BwC1ZQcAtmUHALdVBwCXkwCAozkGAFuUAIAHlACApl0GAF+UAIBjlACApVUGAKpJBgCrSQYAZ5QAgGuUAICuvQcAr6UHAKy9BwCttQcAgG0AAIEJAACCGQAAb5QAgHOUAIC+nAMAd5QAgHuUAICGQAAAh2AAAH+UAICDlACAh5QAgIuUAICPlACAk5QAgKiRBgCpkQYAqrkGAKu5BgCsqQYArakGAK7ZBgCv2QYAl5QAgJuUAICflACAo5QAgKeUAICrlACAr5QAgLOUAIC4cQEAuXEBALpxAQC7cQEAvNkBAL3BAQC+wQEAv/UBALCxBgCxuQYAsokGALOJBgC0UQEAtVEBALZRAQC3UQEAszEGALeUAIC7lACAv5QAgMOUAIC2KQYAtSEGAMeUAIC7fQYAunUGAMuUAIDPlACAv5UBAL6VAQC9XQYAvF0GANOUAICjdQYA15QAgNuUAICmbQYA35QAgOOUAIClZQYAqjEGAKs5BgCErAEAvqABAK7RAQCv0QEArBkGAK0ZBgCo3QIAqe0CAKrlAgCr/QIArOUCAK3tAgCu5QIArz0DAOuUAIDvlACA85QAgL5kDAD3lACA+5QAgP+UAIADlQCAuMkDALnJAwC62QMAu9EDALz5AwC9+QMAvpkDAL+VAwCwRQMAsU0DALJFAwCzXQMAtEUDALVNAwC2RQMAt/kDAIFVAwCASQMAs2UCAIJVAwC1ZQIAB5UAgAuVAIC2ZQIAhgAMAIfkAwC7gQMAuokDAL2BAwC8mQMAv4EDAL6JAwCjLQIAD5UAgBOVAIAXlQCAG5UAgKYtAgClLQIAH5UAgKvJAwCqwQMAI5UAgCeVAICvyQMArsEDAK3JAwCs0QMA49gGAOGsBwDhnAYA45wGACuVAICEWA0AL5UAgDOVAIA3lQCAO5UAgD+VAIBDlQCA7xwBAEeVAIBLlQCA70AGAIB5AACBFQAAghEAAIQADABPlQCA46wAAFOVAIDhpAEAW5UAgO9wAACGyAwAh6QNAF+VAIBjlQCAZ5UAgGuVAIC6yQUAu8kFALilBQC5zQUAvvkFAL/5BQC8zQUAvcUFALKlBQCzrQUAsBEGALERBgC2rQUAt50FALS1BQC1rQUAqmEGAKthBgConQYAqZUGAK5hBgCvYQYArHEGAK1xBgBvlQCAc5UAgHeVAIB7lQCAf5UAgIOVAIC+sAwAh5UAgKghDgCpIQ4AqiEOAKs9DgCsJQ4ArS0OAK4lDgCviQ4AV5UAgIuVAICPlQCAk5UAgJeVAICblQCAn5UAgKOVAIC4UQ8AuV0PALpVDwC7bQ8AvHUPAL19DwC+dQ8Av2kPALD5DgCxoQ4AsqEOALOhDgC0oQ4AtakOALaRDgC3kQ4As6kOAKeVAIDnlACAq5UAgK+VAIC2rQ4Ata0OALOVAIC7ZQ4Auj0OALeVAIC7lQCAv20OAL5lDgC9dQ4AvHUOAIIZAACj7Q4AgGUAAIEZAACm6Q4Av5UAgMOVAICl6Q4AqnkOAKshDgDHlQCAy5UAgK4hDgCvKQ4ArDEOAK0xDgCoYQ4AqXUOAKp9DgCrdQ4ArG0OAK31DgCu/Q4Ar/UOAIaAAQCHpAEAz5UAgNOVAIDXlQCA25UAgN+VAIDjlQCAuHUBALl9AQC6dQEAu8kBALzdAQC9xQEAvsUBAL/1AQCwjQ4AsZUOALKdDgCzkQ4AtFUBALVdAQC2VQEAt00BALP1DgDnlQCA65UAgO+VAIDzlQCAtnUOALXlDgD3lQCAu1EOALpJDgD7lQCA/5UAgL+ZAQC+kQEAvUUOALxJDgADlgCAo7EOAAeWAIALlgCApjEOAA+WAIATlgCApaEOAKoNDgCrFQ4AF5YAgBuWAICu1QEAr90BAKwNDgCtAQ4AqO0CAKktAwCqJQMAqz0DAKwlAwCtLQMAriUDAK+ZAwAflgCAI5YAgCeWAIArlgCAL5YAgDOWAIC+dAIAO5YAgLiNAwC5kQMAupEDALulAwC8vQMAvXUAAL59AAC/dQAAsOkDALHpAwCy+QMAs/EDALTZAwC12QMAtrkDALe1AwCArQAAgbUAAIK9AACzoQMAP5YAgLWhAwC2oQMAQ5YAgITgAgBHlgCAuiEDALshAwC8IQMAvSkDAL4RAwC/EQMAo+0DAIXABACFtG8AS5YAgE+WAICm7QMApe0DAFOWAICrbQMAqm0DAIZIBQCHbAMAr10DAK5dAwCtZQMArG0DAFeWAIDjAA4A71hsAOG0DwBblgCAX5YAgGOWAIBnlgCAoakDAKD9DwCjwQMAog0DAOHgAwDv4A8A4+QDAGuWAIBvlgCAc5YAgIQEBAC+BAQAd5YAgO+UAwB7lgCAf5YAgIOWAIDj1AMAh5YAgOFUAACLlgCAj5YAgJOWAICXlgCAgA0AAIEVAACCHQAAm5YAgJ+WAICjlgCAj5EbAO+cDgCE4AcA4dQOAKuWAIDj8A4Ar5YAgLOWAICGGAcAh5AEAJnlFwCY5RcAm+kLAJo5CwCd/QoAnPELAJ9VDwCeXQ8AkSkfAJDNGwCTJR8Aks0fAJXREwCUKRMAlxkXAJZ1EwCM4RAAjSUQAI4tEACP+QwAN5YAgKeWAICKORQAi5UUAITpGACFBRgAhuUYAIfxFAC3lgCAu5YAgIIxHACDFRwAnKkEAL+WAIDDlgCAx5YAgMuWAIDPlgCAmtEEAJt9BACUTQ0AleUIAJblCACXtQgA05YAgNeWAICSWQwAk1kMAKGRAADblgCAowF8AKKZAACluXwApJF8AKeZeACm4X0AqYF5AKiheACriXQAqgF0AK0BcACsWXQAr4VwAK6dcACx4WwAsAFsALMBaACyHWwAtfVoALT1aADflgCA45YAgOeWAIDrlgCA75YAgPOWAID3lgCA+5YAgP+WAIADlwCAqD0HAKmVBwCqlQcAq6kHAKzdBwCtxQcArsUHAK8dBgAHlwCAgh0AAIEdAACAHQAAC5cAgA+XAIATlwCAvmABALgZBgC5GQYAuikGALslBgC8IQYAvSEGAL4hBgC/IQYAsHEGALFxBgCycQYAs3EGALRNBgC1NQYAtj0GALctBgCzHQcAG5cAgIYoAACHqAAAH5cAgLZFBwC1VQcAI5cAgLu1BgC6tQYAJ5cAgCuXAIC/8QYAvokGAL2lBgC8pQYAL5cAgKNZBwAzlwCAN5cAgKYBBwA7lwCAP5cAgKURBwCq8QYAq/EGAEOXAIBHlwCArs0GAK+1BgCs4QYAreEGAKipBQCptQUAqr0FAKs9AgCsJQIArVECAK5RAgCvUQIAS5cAgE+XAIBTlwCAV5cAgIQ8AwBblwCAX5cAgGOXAIC4pQIAua0CALqlAgC7vQIAvKUCAL2tAgC+pQIAv30DALAxAgCxMQIAshkCALMZAgC09QIAta0CALalAgC3nQIAZ5cAgGuXAIBvlwCAszkFAHOXAIC1oQIAtt0CAHeXAIB7lwCAf5cAgLr5AgC7+QIAvMECAL3BAgC+PQIAv2UCAIOXAICmgQIApf0CAIuXAICjZQUAvlh8AIbYfACHnHwArzkCAK5hAgCtnQIArJ0CAKulAgCqpQIAj5cAgJOXAICohQIAqZUCAKqVAgCrpQIArL0CAK3VAgCu0QIAr9ECAIGFAQCAhQEAl5cAgILtAQCblwCAn5cAgKOXAICnlwCAuHUBALl9AQC6dQEAu80BALzVAQC93QEAvskBAL/BAQCwtQIAsb0CALKBAgCzgQIAtFEBALVRAQC2UQEAt1EBAKuXAICvlwCAs5cAgLeXAIDhMAYA4WQHAOMoBgDjxAYAhCB9ALuXAIDvbAAA7xgGAL+XAIDDlwCAx5cAgMuXAICzXQIAvkh8AM+XAIDTlwCA15cAgLYVAgC1dQIA25cAgLs5AgC6MQIA35cAgOOXAIC/1QEAvtUBAL0VAgC8FQIAo519AIeXAIDnlwCA65cAgO+XAICm1X0ApbV9APOXAICr+X0AqvF9APeXAID7lwCArxV+AK4VfgCt1X0ArNV9AIBNAACBVQAAglUAALOxfgD/lwCAtWV/ALZtfwADmACAhkADAIcEAwC66X8Au+l/ALz5fwC9+X8Avt1/AL/NfwAHmACAC5gAgBeXAIAPmACAE5gAgBeYAIAbmACAH5gAgKhtfgCpXX4AqlV+AKuFfwCsgX8ArYF/AK6BfwCvgX8AsEF/ALFBfwCyQX8As0F/ALR1fwC1ZX8Atm1/ALdlfwC4XX8AuS1/ALolfwC7PX8AvC1/AL0dfwC+FX8Av/UAAKP9fwAjmACAJ5gAgCuYAIAvmACApiF+AKUpfgAzmACAq6V+AKqlfgA3mACAO5gAgK+BfgCukX4ArbV+AKy1fgA/mACAQ5gAgEeYAIBLmACAT5gAgFOYAIBXmACAW5gAgIA9AACBCQAAghkAAF+YAIBjmACAhLgBAL6wAQBnmACAqK0BAKnVAQCq1QEAqw0BAKwVAQCtGQEArgkBAK8JAQCGAAQAhwQBAGuYAIBvmACAc5gAgHeYAIB7mACAf5gAgLjtAAC5hQAAuo0AALuFAAC8nQAAvYUAAL6NAAC/hQAAsHkBALF5AQCy7QAAs+UAALT9AAC15QAAtuUAALfVAACzXQIAg5gAgIeYAICLmACAj5gAgLaZAgC1nQIAk5gAgLu9AgC6vQIAl5gAgJuYAIC/IQMAvjkDAL0xAwC8OQMAvigDAKMZAgCfmACAo5gAgKbdAgCnmACAq5gAgKXZAgCq+QIAq/kCAK+YAICzmACArn0DAK9lAwCsfQMArXUDAL7IBAC3mACAu5gAgL7EBQC/mACAw5gAgMeYAIDLmACAgD0AAIEJAACCGQAAz5gAgNOYAICEOAMA25gAgN+YAIDveAIA45gAgIZIBACHVAMA55gAgOuYAIDvmACA85gAgPeYAID7mACA/5gAgAOZAIDjVAIAB5kAgOFAAQALmQCAD5kAgOMkfwATmQCA4Zx8ABeZAIAbmQCAH5kAgCOZAICEbAUAJ5kAgCuZAIAvmQCAM5kAgO8YfwA3mQCAO5kAgLPxAgA/mQCAQ5kAgEuZAIBPmQCAtukCALXhAgBTmQCAu3EBALppAQCHoAUAhswEAL85AQC+WQEAvVEBALxhAQDhQH8AV5kAgOM4fgCEwAQAgtkAAO8UAACApQAAgdkAAFuZAIDjwAAAX5kAgOHUAQBjmQCAZ5kAgO+EfgBrmQCAqs0BAKvVAQBvmQCAc5kAgK79AQCvnQEArMUBAK31AQB3mQCAo1UCAHuZAIB/mQCApk0CAIOZAICHmQCApUUCANeYAIBHmQCAi5kAgI+ZAICTmQCAl5kAgJuZAICfmQCAqJkGAKmZBgCq7QYAq/0GAKzlBgCt7QYAruUGAK/dBgCwpQYAsa0GALKlBgCzuQYAtK0GALVVBwC2UQcAt00HALh1BwC5fQcAunUHALtJBwC8WQcAvVkHAL5JBwC/RQcAs0UGAKOZAICnmQCAq5kAgK+ZAIC2TQYAtU0GALOZAIC7SQYAukEGAIYIAACHjAAAv7EHAL5JBgC9TQYAvFEGAIJdAACjAQYAgEUAAIFdAACmCQYAu5kAgL+ZAIClCQYAqgUGAKsNBgDDmQCAx5kAgK4NBgCv9QcArBUGAK0JBgCoTQYAqVUGAKpVBgCriQYArLEGAK29BgCuqQYAr6kGALeZAIDLmQCAz5kAgNOZAIDXmQCA25kAgN+ZAIDjmQCAuEkBALlJAQC6WQEAu1kBALxJAQC9SQEAvt0BAL/VAQCw3QYAsa0GALKlBgCzjQYAtJkGALWZBgC2jQYAt4UGALPdBgDnmQCA65kAgO+ZAIDzmQCAtj0GALU5BgD3mQCAu2kGALoZBgD7mQCA/5kAgL9dBgC+XQYAvVkGALxxBgADmgCAo5kGAAeaAIALmgCApnkGAA+aAIATmgCApX0GAKpdBgCrLQYAF5oAgBuaAICuGQYArxkGAKw1BgCtHQYAqNUCAKndAgCq4QIAq+ECAKw1AwCtPQMArjUDAK8tAwCAzQMAgQkAAIIZAAAfmgCAI5oAgIQYAgC+dAMAK5oAgLjpAwC56QMAuokDALuFAwC8nQMAvYEDAL6BAwC/tQMAsFUDALFdAwCyVQMAs+kDALT5AwC1+QMAtukDALfhAwCGIAwAhxADAC+aAIAzmgCAN5oAgDuaAIA/mgCA71wCAEOaAIDhFAAAR5oAgOOIAgC++AwAS5oAgE+aAIBTmgCAu/kDALrxAwC+gA0AV5oAgL9dAwC+XQMAvV0DALzhAwCzCQIAW5oAgF+aAIBjmgCAZ5oAgLbdAwC13QMAa5oAgKipBgCpqQYAqrkGAKu5BgCsqQYArakGAK4dBQCvFQUAb5oAgHOaAIB3mgCAe5oAgH+aAICDmgCAh5oAgIuaAIC4GQUAuS0FALolBQC7yQUAvNkFAL3FBQC+zQUAv8UFALBtBQCxdQUAsnUFALNFBQC0XQUAtT0FALY1BQC3KQUA4fQGAOFUBwDjFAYA47wGAIEJAACAqQAAj5oAgII5AACE7A0Ak5oAgIeIDACGDAwAm5oAgJ+aAIDvzAcA78QHAKMpAwCjmgCAp5oAgKuaAICvmgCApv0CAKX9AgCzmgCAq9kCAKrRAgC3mgCAu5oAgK99AgCufQIArX0CAKzBAgCoPQ4AqY0OAKqFDgCrnQ4ArIUOAK2NDgCuuQ4Ar7UOAJeaAIC/mgCAw5oAgMeaAIDLmgCAz5oAgNOaAIDXmgCAuL0OALllDwC6bQ8Au2UPALx9DwC9ZQ8Avm0PAL9lDwCw1Q4Asd0OALLVDgCzoQ4AtJUOALWdDgC2lQ4At40OALMNDgDbmgCA35oAgOOaAIDnmgCAtg0OALUNDgDrmgCAuxkOALoRDgDvmgCAJ5oAgL9ZDgC+UQ4AvXUOALwBDgDzmgCAo0kOAPeaAID7mgCApkkOAP+aAIADmwCApUkOAKpVDgCrXQ4AhKQDAAebAICuFQ4Arx0OAKxFDgCtMQ4AqLEOAKmxDgCqzQ4Aq8UOAKzdDgCtxQ4ArsUOAK/1DgCA7QEAgfEBAILxAQALmwCAhpABAIe0AQAPmwCAE5sAgLjFAQC5zQEAusUBALvdAQC8zQEAvf0BAL6ZAQC/lQEAsI0OALFBAQCyQQEAs0EBALRBAQC1QQEAtkEBALdBAQCzRQ4AF5sAgBubAIAfmwCAI5sAgLZFDgC1VQ4AJ5sAgLuFAQC6SQ4AK5sAgC+bAIC/hQEAvoUBAL2VAQC8lQEAM5sAgKMBDgA3mwCAO5sAgKYBDgA/mwCAQ5sAgKURDgCqDQ4Aq8EBAEebAIBLmwCArsEBAK/BAQCs0QEArdEBAKgtAwCpPQMAqjUDAKuJAwCsmQMArZkDAK6JAwCvgQMAT5sAgFObAIBXmwCAW5sAgF+bAIBjmwCAZ5sAgGubAIC4rQMAuWUAALptAAC7ZQAAvH0AAL1lAAC+bQAAv2UAALDJAwCxyQMAsqkDALOlAwC0vQMAtaEDALahAwC3lQMAgL0AAIEJAACCGQAAb5sAgHObAIC+2AMAe5sAgH+bAICErAIAg5sAgIfoAwCGDAQAh5sAgIubAICPmwCAk5sAgLP9AwCXmwCAm5sAgJ+bAICjmwCAtlkDALVRAwCnmwCAu00DALpNAwCrmwCAr5sAgL8lAwC+OQMAvTEDALw9AwCzmwCAt5sAgLubAIC/mwCA71gPAMObAIDHmwCAy5sAgOOQDgDPmwCA4bAPANObAIDXmwCA25sAgN+bAIDjmwCAgHUAAIF9AACCdQAAhBgFAO88AwDrmwCAvhQFAO+bAIDj0AMA85sAgOFAAAD3mwCAhtAEAIdYBQD7mwCA/5sAgAOcAIAHnACAC5wAgA+cAIATnACAF5wAgBucAIDvrA8AhOwEAOEQDgAfnACA41QBACOcAIAnnACAK5wAgC+cAICj/QIAM5wAgDecAIA7nACAP5wAgKZZAgClUQIAQ5wAgKtNAgCqTQIAR5wAgEucAICvJQIArjkCAK0xAgCsPQIAqJkGAKmZBgCqrQYAq70GAKylBgCtrQYArqUGAK/ZBgDnmwCAghEAAIEZAACAwQcAT5wAgFOcAIC+cAMAV5wAgLhJBwC5SQcAul0HALtVBwC8TQcAvXEHAL51BwC/bQcAsKkGALGpBgCyuQYAs7EGALSZBgC1mQYAtnkHALd5BwC1NQYAW5wAgF+cAIC2NQYAhjAAAIdcAwCzPQYAY5wAgL19BgC8dQYAv0UGAL5FBgB3mwCAZ5wAgLt1BgC6dQYAo2UGAGucAIBvnACAc5wAgHecAICmbQYApW0GAHucAICrLQYAqi0GAH+cAICDnACArx0GAK4dBgCtJQYArC0GAKhVBgCpWQYAqm0GAKthBgCsaQYArWkGAK6ZBgCvmQYAh5wAgIucAICPnACAk5wAgJecAICbnACAn5wAgKOcAIC4+QYAufkGALqNBgC7hQYAvJ0GAL2FBgC+hQYAv7UGALDpBgCx6QYAsvkGALP5BgC06QYAtd0GALbJBgC3yQYAs+UGAKecAICrnACAr5wAgLOcAIC26QYAteEGALecAIC7LQYAui0GALucAIC/nACAvxkGAL4tBgC9LQYAvC0GAIIVAACjoQYAgGEAAIFhAACmrQYAw5wAgL6QAQClpQYAqmkGAKtpBgCEpAEAy5wAgK5pBgCvXQYArGkGAK1pBgCohQIAqY0CAKqVAgCruQIArNUCAK3dAgCu1QIAr80CAIaAHACHZAMAz5wAgL5gAwDTnACA15wAgNucAIDfnACAuHUDALl9AwC6dQMAu8kDALzZAwC92QMAvskDAL/BAwCwvQIAsY0CALKFAgCzTQMAtFUDALVdAwC2VQMAt00DALMdAgDjnACAhAgDAOecAIDrnACAtl0CALVdAgDvnACAu0kCALp5AgDznACA95wAgL+ZAwC+kQMAvZkDALxRAgCwAAAAo1kCAPucAID/nACAphkCAAOdAIAHnQCApRkCAKo9AgCrDQIAC50AgA+dAICu1QMAr90DAKwVAgCt3QMAE50AgBedAIAbnQCA76wGAB+dAIAjnQCAJ50AgCudAIC+6BwAL50AgDOdAIA7nQCAP50AgOGABwBDnQCA42AGAIBdAACBYQAAgmEAALN9AQBHnQCAtW0BALZlAQBLnQCAhiAdAIdYHQC6+QEAu/EBALzZAQC92QEAvrEBAL+xAQDvoAAAT50AgFOdAIBXnQCAW50AgF+dAIBjnQCA71wBAIRsHADhzAYAZ50AgOMcBgDjSAAAa50AgOEwAQBvnQCAo/EBAHOdAICFABQAd50AgHudAICm6QEApeEBAH+dAICrfQEAqnUBAIOdAICHnQCArz0BAK49AQCtVQEArFUBAKjtHQCpLR4AqjkeAKs5HgCsKR4ArSkeAK6dHgCvkR4AN50AgIudAICPnQCAk50AgJedAICC+QAAgfEAAID9AAC4qR4AuakeALpJHwC7SR8AvFkfAL1FHwC+TR8Av0UfALDxHgCx+R4AssEeALPBHgC0uR4AtbkeALatHgC3pR4AsBEfALERHwCyER8AsyUfALQlHwC1KR8Atl0fALdRHwC4cR8AuXkfALpBHwC7QR8AvJUAAL2dAAC+lQAAv40AAJudAIDHnACAn50AgKOdAICnnQCAq50AgIb4AwCH0AAAqM0fAKnVHwCq0R8Aq70fAKytHwCtcR8ArnEfAK9xHwCzOR4Ar50AgLOdAIC3nQCAu50AgLaRHgC1RR4Av50AgLu1HgC6tR4Aw50AgMedAIC/jR4AvoEeAL2RHgC8pR4Ay50AgKN9HgDPnQCA050AgKbVHgDXnQCA250AgKUBHgCq8R4Aq/EeAN+dAIDjnQCArsUeAK/JHgCs4R4ArdUeAKhVAQCpgQAAqoEAAKuBAACsgQAArYkAAK6xAACvsQAA550AgOudAIDvnQCA850AgPedAID7nQCA/50AgAOeAIC4ZQAAuW0AALplAAC7fQAAvGUAAL1tAAC+ZQAAv90DALChAACxrQAAsqUAALO5AAC0qQAAtZ0AALaVAAC3XQAAB54AgIIdAACBHQAAgB0AAAueAIAPngCAE54AgL4UAgAbngCAhKgCAB+eAIAjngCAJ54AgCueAIAvngCAjwAAALNJAwAzngCAhugEAIesAgA3ngCAtkkDALVJAwA7ngCAuykDALolAwA/ngCAQ54AgL8ZAwC+LQMAvS0DALwxAwBHngCAo40DAEueAIBPngCApo0DAFOeAIBXngCApY0DAKrhAwCr7QMAW54AgF+eAICu6QMAr90DAKz1AwCt6QMAvoQDAGOeAIBnngCAa54AgG+eAIBzngCAd54AgHueAICAPQAAgQkAAIIZAAB/ngCAg54AgIueAICENAMAj54AgLMtAQCTngCAh8wCAIZMBQCXngCAti0BALUtAQCbngCAu0kBALp5AQCfngCAo54AgL+9AQC+vQEAvbkBALxRAQDheB8Ap54AgOPQHwCrngCAr54AgOGUAQCzngCA42gDALeeAIC7ngCAv54AgO+IAwDDngCAx54AgO+sHwDLngCAz54AgNOeAIDXngCA254AgN+eAIDjngCA554AgO9EHgDrngCA4dweAO+eAIDjHB4A854AgPueAID/ngCAA58AgIFpAACAZQAAo+UBAIJ9AACl5QEAB58AgIQUBACm5QEAvigEAAufAICrgQEAqrEBAK1xAQCsmQEAr3UBAK51AQCoIQYAqS0GAKolBgCrPQYArCUGAK0tBgCuXQYAr00GAIeeAID3ngCAhggDAIeMAwAPnwCAE58AgBefAIAbnwCAuOkGALnpBgC6jQYAu4UGALydBgC9hQYAvo0GAL+FBgCwPQYAsQ0GALIFBgCz7QYAtPkGALX5BgC27QYAt+UGALDNBwCx1QcAstEHALPtBwC09QcAtf0HALbpBwC36QcAuN0HALklBwC6LQcAuyUHALw9BwC9JQcAvi0HAL8lBwAfnwCAI58AgBeeAIAnnwCAK58AgC+fAIAznwCAN58AgKgVBgCpGQYAqu0HAKv9BwCs7QcArd0HAK7VBwCvuQcAswUGADufAIA/nwCAQ58AgEefAIC2PQYAtQUGAEufAIC7cQYAumkGAE+fAIBTnwCAv1kGAL5RBgC9WQYAvGUGAFefAICjQQYAW58AgF+fAICmeQYAY58AgIS0AQClQQYAqi0GAKs1BgC+gAEAa58AgK4VBgCvHQYArCEGAK0dBgCoNQYAqT0GAKo1BgCrWQYArHUGAK2lAQCurQEAr6UBAIDpAACB6QAAgv0AAL8kAQCGMA8Ah+QAAG+fAIBznwCAuMUAALnNAAC6xQAAu90AALzNAAC9/QAAvvUAAL+dAACw3QEAsSUBALItAQCzIQEAtCEBALUhAQC2IQEAtyEBALvBAgC6OQIAd58AgHufAIC/xQIAvsUCAL3VAgC82QIAs50FAH+fAICDnwCAh58AgIwAAAC2BQIAtd0FAIufAICqfQIAq4UCAI+fAICTnwCAroECAK+BAgCsnQIArZECAJefAICj2QUAm58AgJ+fAICmQQIAo58AgKefAIClmQUAgpFqAIORagCrnwCAr58AgIa5FgCH6RcAhBEWAIWZFgCKoRIAi6ESALOfAIC3nwCAjpEeAI9ZHgCMmRMAjREeAJJxGgCT5RoAu58AgO/oJACW8QYAlwUGAJTlGgCVGQYAmikCAJvFAgC/nwCAw58AgMefAIDhKBsAnN0CAOMgDwCfIQcAnsEHAJ01GwCcLRsAm6EbAJr5HwCZOR8AmLEfAJcBEgCWIRMAlSkTAJRRFgCTGRcAkjEXAJGxFwCQKWsAj1FrAOOsBwCEBA0A4RwHAIANAACBNQAAgj0AAMufAIDPnwCA058AgL4gDQDbnwCA358AgO9MBwCGWAwAh2ANAOOfAIDnnwCA658AgO+fAICEXA8A858AgO8IAADvhAYA4ZABAOGwBgDj4AAA42QGAPefAID7nwCA/58AgAOgAIAHoACAC6AAgL4ADwCEQA4AD6AAgBOgAIAXoACAG6AAgB+gAIAjoACAJ6AAgCugAICj1QMAotUDAKExAwCgLQcAZ58AgNefAIAvoACAM6AAgDegAICCmQAAgZEAAICZAACoTQ0AqZ0NAKqVDQCrJQ4ArD0OAK0RDgCuEQ4ArxEOALB9DgCxDQ4AsgUOALMtDgC0OQ4AtTkOALYtDgC3JQ4AuOkOALnpDgC6wQ4Au8EOALy5DgC9nQ4AvpUOAL+NDgCzPQ0AO6AAgD+gAIBDoACAR6AAgLaxDgC1lQ4AS6AAgLvpDgC6mQ4AhogAAIfkAAC/3Q4Avt0OAL3ZDgC88Q4AT6AAgKN5DQC+hAEAhIAGAKb1DgBToACAV6AAgKXRDgCq3Q4Aq60OAFugAIBfoACArpkOAK+ZDgCstQ4ArZ0OALIFNQCzGTQAsG0wALENNQBjoACAZ6AAgLQBKAC1PSkAa6AAgG+gAIBzoACAd6AAgHugAIB/oACAg6AAgIegAICiRQEAo9UBAIugAIChTQEAps0FAKcBOACkAQQApX0FAKoBPACrRT0AqEk5AKnlOQCudTEAr30xAKxdPQCtATAAqO0OAKn1DgCqCQ4AqwkOAKwZDgCtGQ4Arg0OAK8tDgCPoACAk6AAgJegAICboACAn6AAgKOgAICnoACAq6AAgLgdDgC5JQ4Aui0OALslDgC8PQ4Avd0BAL7VAQC/zQEAsFUOALFdDgCyVQ4Asy0OALQ1DgC1JQ4Ati0OALclDgCzgQ0Ar6AAgLOgAIC7oACAv6AAgLaZDQC1kQ0AvlQEALuZDQC6kQ0AhogEAIe8AwC/4Q0AvvENAL35DQC8gQ0AgkkAAKPFDQCA9QMAgUkAAKbdDQDDoACAx6AAgKXVDQCq1Q0Aq90NAMugAIDPoACArrUNAK+lDQCsxQ0Arb0NAKgdAgCpRQIAql0CAKtVAgCseQIArXkCAK6JAwCviQMA06AAgNegAIDboACA36AAgIT8BQDjoACA56AAgOugAIC4iQMAuWUDALptAwC7ZQMAvH0DAL1lAwC+bQMAv2UDALDBAwCxwQMAssEDALPBAwC0wQMAtcEDALbBAwC3wQMA76AAgPOgAID3oACA+6AAgP+gAIDhpAEAA6EAgOPADgC+aAQAB6EAgAuhAIDvHAEAD6EAgBOhAIAXoQCAG6EAgLOVAwAfoQCAI6EAgCuhAIAvoQCAtrkDALWxAwAzoQCAu0UCALpFAgCGqAQAh6QFAL9FAgC+RQIAvVUCALxVAgDh4A4A4SwMAOMIDgDj1A4AgK0AAIHRAACC0QAAN6EAgDuhAIA/oQCAQ6EAgEehAIBLoQCAT6EAgO+IDgDvLA4AoxUDAFOhAICFxCsAV6EAgFuhAICmOQMApTEDAF+hAICrxQIAqsUCAGOhAIBnoQCAr8UCAK7FAgCt1QIArNUCAKgNBgCpFQYAql0GAKtVBgCseQYArXkGAK65BgCvuQYAJ6EAgGuhAIBvoQCAc6EAgHehAIB7oQCAf6EAgIOhAIC4TQcAuVUHALpRBwC7aQcAvHkHAL1lBwC+bQcAv2UHALDJBgCxyQYAst0GALPVBgC0zQYAtXUHALZ9BwC3dQcAs9UGAIehAICLoQCAj6EAgJOhAIC2+QYAtfEGAJehAIC7DQYAug0GAIYIAACHLAAAv7EHAL4JBgC9AQYAvAkGAIJRAACjkQYAgEEAAIFBAACmvQYAm6EAgJ+hAICltQYAqkkGAKtJBgCjoQCAp6EAgK5NBgCv9QcArE0GAK1FBgCwsQYAsbEGALLNBgCzwQYAtMEGALXJBgC28QYAt/EGALgFAQC5DQEAugUBALsdAQC8BQEAvQ0BAL4FAQC/uQEAq6EAgK+hAICzoQCAt6EAgLuhAIC/oQCAt6AAgMOhAICoLQYAqTUGAKo1BgCr8QYArNEGAK3RBgCu0QYAr9EGALPdBgDHoQCAy6EAgM+hAIDToQCAtjEGALU5BgDXoQCAuxUGALoVBgDboQCA36EAgL9tBgC+ZQYAvXUGALx5BgDjoQCAo5kGAOehAIDroQCApnUGAO+hAIDzoQCApX0GAKpRBgCrUQYA96EAgPuhAICuIQYArykGAKw9BgCtMQYAqNUCAKndAgCq4QIAq+ECAKxRAwCtUQMArlEDAK9RAwD/oQCAA6IAgL7sAwALogCAD6IAgBOiAIAXogCAG6IAgLjpAwC56QMAuokDALuFAwC8nQMAvYEDAL6BAwC/tQMAsDEDALExAwCyNQMAs+kDALT5AwC1+QMAtukDALfhAwCAbQMAgaUAAIKtAACzZQIAH6IAgLXVAwC23QMAI6IAgITgAgAnogCAuvkDALv5AwC87QMAvTEDAL4xAwC/MQMAh+wDAIZkPACyAAAAK6IAgC+iAIDjCAQAM6IAgOHsBgA3ogCA7wAGADuiAIA/ogCAQ6IAgEeiAIBLogCAT6IAgFOiAIBXogCAW6IAgF+iAIDjoAMAY6IAgOGoAQBnogCA7/ADAIIdAACBHQAAgB0AAGuiAIBvogCAc6IAgHuiAIC+TD0Af6IAgKOhAwC+QDwApRECAIOiAICHogCAphkCAIRsAgCLogCAqz0CAKo9AgCt9QIArCkCAK/1AgCu9QIAhkA8AIe0PQCPogCAk6IAgJeiAICbogCAn6IAgO9EBgCjogCA4dQGAKeiAIDjDAcAq6IAgK+iAICzogCAt6IAgLP1AQC7ogCAv6IAgMOiAIDHogCAtkUBALXlAQDLogCAuzEBALopAQDPogCA06IAgL8dAQC+HQEAvRkBALwlAQCoLT4AqTU+AKo9PgCrNT4ArC0+AK2FPgCuhT4Ar7k+AHeiAIDXogCA26IAgN+iAICAGQAAgRkAAIIFAADjogCAuLk+ALm5PgC6ST8Au0k/ALxZPwC9WT8Avk0/AL9BPwCwrT4AsbU+ALKxPgCzjT4AtJk+ALWZPgC2iT4At4k+AKO1PgCEjAIA56IAgOuiAIDvogCApgU+AKWlPgDzogCAq3E+AKppPgCGCAAAh2gDAK9dPgCuXT4ArVk+AKxlPgD3ogCAs5E/APuiAID/ogCAtlk/AAOjAIAHowCAtbk/ALp1PwC7fT8AC6MAgA+jAIC+QT8Av0E/ALxZPwC9VT8AsJU+ALGdPgCyqT4As6U+ALShPgC1oT4AtqE+ALehPgC45T4Aue0+ALrlPgC7/T4AvO0+AL3dPgC+1T4AvxkBABOjAIAXowCAG6MAgB+jAIAjowCAB6IAgCejAIArowCAqF0+AKkhPgCqPT4AqzU+AKwVPgCt/T4ArvU+AK/tPgCj1T4AL6MAgDOjAIA3owCAO6MAgKYdPgCl/T4AP6MAgKs5PgCqMT4AQ6MAgEejAICvBT4ArgU+AK0RPgCsHT4AgREAAIANAABLowCAghkAAE+jAIBTowCAhJQBAL4QAACGQAcAhwABAFujAIBfowCAY6MAgGejAIBrowCAb6MAgKiNAgCplQIAqpUCAKvNAgCs2QIArdkCAK7NAgCvxQIAc6MAgHejAIB7owCAf6MAgIwAAACDowCAh6MAgIujAIC4HQMAucEDALrBAwC7wQMAvMEDAL3JAwC+8QMAv/EDALCJAgCxiQIAsikDALMpAwC0OQMAtTkDALYpAwC3JQMAsx0CAI+jAICTowCAl6MAgJujAIC2WQIAtVECAJ+jAIC7TQIAuk0CAKOjAICnowCAv/0DAL79AwC9/QMAvP0DAKujAICvowCAs6MAgLejAIDhDD4Au6MAgOOoPwC/owCAgT0AAIAxAADvUD8Agh0AAMOjAIC++AQAhhgFAIdMAwCEDAIA48wAAMujAIDhvAEAz6MAgNOjAIDXowCA26MAgN+jAICELAUA46MAgOejAIDrowCA7xAAAO+jAIDzowCAo90DAPejAID7owCA/6MAgAOkAICmmQMApZEDAAekAICrjQMAqo0DAAukAIAPpACArz0CAK49AgCtPQIArD0CABOkAIAXpACAG6QAgB+kAIAjpACAJ6QAgCukAIDvKD4AL6QAgOE8PgAzpACA4zgBAIApAACBFQAAghEAADukAICzMQIAvsgEAITABAA/pACAQ6QAgLYpAgC1IQIAR6QAgLvNAQC6zQEAS6QAgE+kAIC/dQEAvskBAL3BAQC8yQEAqOkFAKnpBQCq+QUAq/kFAKzpBQCt6QUArjkGAK85BgDHowCAN6QAgIaIAACHQAMAU6QAgFekAIBbpACAX6QAgLjRBgC52QYAuuEGALvhBgC8kQYAvZEGAL6RBgC/kQYAsEkGALFJBgCyXQYAs1UGALRNBgC18QYAtvEGALfxBgCjcQUAY6QAgGekAIBrpACAb6QAgKZpBQClYQUAc6QAgKuNBgCqjQYAd6QAgHukAICvNQYArokGAK2BBgCsiQYAf6QAgLPRBwCDpACAh6QAgLbxBwCLpACAj6QAgLXBBwC60QcAu90HAJOkAICXpACAvrkHAL+5BwC8xQcAvbkHALhpBgC5aQYAuokGALuJBgC8mQYAvZkGAL6JBgC/iQYAsBEGALEdBgCyFQYAs2kGALR5BgC1eQYAtmkGALdhBgCoSQYAqVUGAKpdBgCrVQYArE0GAK11BgCucQYAr3EGAFejAICCHQAAgR0AAIAdAACbpACAn6QAgKOkAIC+cAEAo5UGAKukAICGKAAAh0gBAK+kAICmtQYApYUGALOkAICrmQYAqpUGALekAIC7pACAr/0GAK79BgCt/QYArIEGAL+kAICzFQYAw6QAgMekAIC2PQYAy6QAgM+kAIC1NQYAutkBALvZAQDTpACA16QAgL59AQC/ZQEAvH0BAL11AQCovQUAqckFAKrZBQCr0QUArPkFAK35BQCuKQIArykCANukAIDfpACA46QAgOekAICMAAAA66QAgO+kAIDzpACAuO0CALmFAgC6gQIAu4ECALyFAgC9jQIAvrECAL+xAgCwWQIAsVkCALLtAgCz5QIAtP0CALXlAgC25QIAt9UCAKNRBQD3pACA+6QAgP+kAIADpQCApnkFAKVxBQAHpQCAq50CAKqdAgALpQCAD6UAgK8hAgCuOQIArTECAKw5AgCBbQAAgG0AABOlAICCBQAAvlwMABulAIAfpQCA79AGAITsAwDhHAUAI6UAgOP8BwAnpQCAK6UAgIbYDACHvAwAqIUCAKmVAgCqlQIAq6UCAKy9AgCt1QIArtECAK/RAgAvpQCAM6UAgDelAIA7pQCAP6UAgEOlAIBHpQCAS6UAgLh1AQC5fQEAunUBALvJAQC82QEAvdkBAL7JAQC/wQEAsLUCALG9AgCygQIAs4ECALRRAQC1UQEAtlEBALdRAQBPpQCAhAQNAFOlAIBXpQCAvhwMAFulAIDvHAAA76AGAOGQAQDhRAcA43AGAOOYBgBfpQCAY6UAgGelAIBrpQCAs10CAG+lAIBzpQCAd6UAgHulAIC2FQIAtXUCAH+lAIC7OQIAujECAIOlAICLpQCAv9UBAL7VAQC9FQIAvBUCAKOdDQAXpQCAh6UAgI+lAICTpQCAptUNAKW1DQCXpQCAq/kNAKrxDQCGCAMAh2ADAK8VDgCuFQ4ArdUNAKzVDQCAkQ8AgZkPAIKhDwCzpQ4Am6UAgLWhDgC2eQ8An6UAgKOlAICnpQCAukUPALtdDwC8RQ8AvU0PAL5FDwC//Q8AqFUOAKldDgCqYQ4Aq30OAKxlDgCttQ8Arr0PAK+1DwCrpQCAr6UAgLOlAIC3pQCAu6UAgL+lAIDDpQCAx6UAgLhVDwC5dQ8Aun0PALt1DwC8bQ8AvREPAL4RDwC/EQ8AsM0PALHVDwCy3Q8As9UPALTNDwC1dQ8AtnEPALdxDwCj6Q8Ay6UAgM+lAIDTpQCA16UAgKY1DgCl7Q8A26UAgKsRDgCqCQ4A36UAgOOlAICvsQ4ArgkOAK0BDgCsCQ4A56UAgIIdAACBHQAAgB0AAOulAIDvpQCA86UAgL6UAQCErAEA96UAgIfgAQCGzAAA+6UAgP+lAIADpgCAp6QAgKhtDgCpiQEAqpkBAKuRAQCswQEArckBAK75AQCv+QEAhKAAAAemAIALpgCAD6YAgBOmAIAXpgCAG6YAgB+mAIC4xQAAuc0AALrFAAC73QAAvM0AAL39AAC+9QAAv50AALBBAQCxQQEAskEBALNBAQC0QQEAtUEBALZBAQC3QQEAsxECACOmAIAnpgCAK6YAgC+mAIC2SQIAtUkCADOmAIC7hQIAuoUCADemAIA7pgCAv4UCAL6FAgC9lQIAvJUCAIU8GgCjVQIAP6YAgEOmAICmDQIAR6YAgEumAIClDQIAqsECAKvBAgBPpgCAU6YAgK7BAgCvwQIArNECAK3RAgCCGQAAV6YAgIAZAACBGQAAW6YAgF+mAIBjpgCAa6YAgL4ABABvpgCAc6YAgHemAIB7pgCAf6YAgIOmAICHpgCA7+gOAIumAICG6AQAh1ADAI+mAICTpgCA74ACAJemAIDhlAEAm6YAgONYAQCfpgCA4wAOAKOmAIDhaA0Ap6YAgKhxAgCpcQIAqnECAKupAgCsuQIArbkCAK6pAgCvqQIAhKwFAKumAICvpgCAs6YAgLemAIC7pgCAv6YAgMOmAIC4bQEAuQ0BALoFAQC7GQEAvAkBAL09AQC+NQEAv9kBALDZAgCx2QIAsm0BALNlAQC0fQEAtWUBALZlAQC3VQEA4WAPAOP0AADjHA4A4bwBAMemAICCOQAAgTEAAIA9AADLpgCAvigEAM+mAIDTpgCAvjwHAO8QAADv0A4A26YAgIbgBACHyAQA36YAgLO1AgDjpgCAtX0CALZ1AgDnpgCA66YAgO+mAIC6UQIAu1ECALz1AQC9/QEAvvUBAL/tAQBnpgCA16YAgKqxBQCrsQUArBUGAK0dBgCuFQYArw0GAPOmAID3pgCA+6YAgKNVBQD/pgCApZ0FAKaVBQADpwCAs+kGAAenAIALpwCAD6cAgBOnAIC24QYAtekGABenAIC7sQYAuqEGABunAIAfpwCAv50GAL6RBgC9pQYAvKkGAKgdBgCpIQYAqiEGAKshBgCsIQYArSEGAK4hBgCvIQYAI6cAgCenAIArpwCAL6cAgDOnAIA3pwCAO6cAgD+nAIC45QcAue0HALrlBwC7/QcAvOUHAL3tBwC+5QcAv00HALAlBgCxNQYAsj0GALMxBgC0FQYAtRkGALYNBgC3AQYAo6kHAIIVAACBtQEAgLUBAEOnAICmoQcApakHAEenAICr8QcAquEHAISgAgBLpwCAr90HAK7RBwCt5QcArOkHAE+nAICzlQYAhugAAIcYAQC2tQYAU6cAgFenAIC1vQYAukkBALtVAQBbpwCAX6cAgL45AQC/OQEAvEUBAL05AQCoPQYAqU0GAKpZBgCrUQYArHEGAK1xBgCuuQEAr7kBAISsAQBjpwCAZ6cAgGunAIBvpwCAc6cAgHenAIB7pwCAuKkBALmpAQC6aQEAu2kBALx5AQC9eQEAvmkBAL9pAQCwyQEAsdUBALLVAQCzqQEAtLkBALW5AQC2qQEAt6EBAKPRBQB/pwCAg6cAgIenAICLpwCApvEFAKX5BQCPpwCAqxECAKoNAgCTpwCAl6cAgK99AgCufQIArX0CAKwBAgCbpwCAn6cAgKOnAICnpwCAgTEAAIANAACrpwCAgjkAAK+nAICzpwCAviQDALunAIC/pwCAw6cAgIbYHACHTAMAx6cAgMunAIDPpwCAhMAcAOMgAQDTpwCA4cgBANenAIDvMAIA26cAgN+nAIDjpwCA56cAgOunAIDvpwCA86cAgLOVAwD3pwCA+6cAgP+nAIADqACAtrkDALWxAwAHqACAu1EDALpJAwALqACAD6gAgL/1AAC+SQMAvUEDALxJAwCoLQIAqUUCAKpdAgCrVQIArHkCAK15AgCuvQIAr7UCAL5oHQATqACAF6gAgBuoAICAHQAAgQkAAIKpAAAfqACAuFEBALlZAQC6YQEAu2EBALwRAQC9EQEAvhEBAL8RAQCwzQIAsdUCALLdAgCz1QIAtM0CALVxAQC2cQEAt3EBAOFYBgDhVAcA47AAAOO8BgAjqACAK6gAgIYYHACHVB0AL6gAgDOoAIA3qACAO6gAgL74HAA/qACA7/AGAO/gBgCjlQIAQ6gAgEeoAIBLqACAT6gAgKa5AgClsQIAU6gAgKtRAgCqSQIAV6gAgFuoAICv9QEArkkCAK1BAgCsSQIAqG0eAKl1HgCqfR4Aq40eAKyVHgCtnR4Aro0eAK+BHgAnqACAX6gAgGOoAIBnqACAa6gAgG+oAIBzqACAd6gAgLiJHgC5iR4AupkeALuRHgC8uR4AvbkeAL59HwC/dR8AsMUeALHNHgCyxR4As90eALTFHgC1zR4AtsUeALe5HgCz9R4Ae6gAgH+oAICDqACAh6gAgLYdHgC1HR4Ai6gAgLsJHgC6AR4Aj6gAgJOoAIC/CR4AvgEeAL0JHgC8ER4Agm0AAKOxHgCAVQAAgWUAAKZZHgCEmAMAv9ABAKVZHgCqRR4Aq00eAIYABACHmAEArkUeAK9NHgCsVR4ArU0eAJuoAICfqACAhCQAAKOoAICnqACAq6gAgLenAICXqACAqLUeAKmFHgCqjR4Aq4UeAKydHgCtgR4Arv0eAK/1HgCwjR4AsZUeALKVHgCzpR4AtL0eALVxAQC2cQEAt3EBALhRAQC5UQEAulEBALtRAQC89QEAvf0BAL71AQC/7QEAsyUeAL4IBwCvqACAs6gAgLeoAIC2IR4AtTUeALuoAIC7cR4AumkeAL+oAIDDqACAv5UBAL5ZHgC9UR4AvGEeAMeoAICjYR4Ay6gAgM+oAICmZR4A06gAgNeoAIClcR4Aqi0eAKs1HgDbqACA36gAgK4dHgCv0QEArCUeAK0VHgDhVBoA46gAgONcCgDnqACA66gAgO+oAIDzqACA96gAgPuoAIC+qAUA/6gAgAOpAICPMSoAC6kAgO/E+wAPqQCAk2EuAJIdLwCR2SoAkEkqAJfZEgCWdRIAlQ0TAJTBLgCbHRsAmkEWAJlJFgCYDRcAn3EeAJ4RGwCdcRoAnHkaAKOhAgCinQMAoZUfAKCJHgDjiAEA4wgeAOFoAADh/B4A79wBAO98HwC1if4AtAH8ALMB+gCylfoAsQH4ALAR9gCv4fYArgH0AK0l8gCs7fIAqwHwAKrpDwCp1Q4AqN0OAKcBDACmyQoApe0KAKQBCACj4QYAovEGAKHlAwATqQCAggErAIMBKwAXqQCAG6kAgIYxLwCHiS8AhIkrAIVFLgCKdRIAiwUTAIYIBQCHbAUAjhEXAI8RFwCMsRMAjV0WAJI9GgCTQRsAhMgFAIQABwCWUR8Al1EfAJRRGwCVORoAmn0eAJt9AgAfqQCAI6kAgIFZAQCAVQEAnFkDAIJRAQC+yAcAJ6kAgCupAIAvqQCAM6kAgDepAIA7qQCA79QeAD+pAIDhJB4AQ6kAgONoAQBHqQCAS6kAgE+pAIBTqQCAu2kCALpZAgBXqQCAW6kAgL8dAgC+HQIAvRkCALxxAgCz7QIAX6kAgGOpAIBnqQCAa6kAgLZ9AgC17QIAb6kAgKMNBQAHqQCAc6kAgHupAIB3qQCApp0FAKUNBQB/qQCAq4kFAKq5BQCGCAMAh3wDAK/9BQCu/QUArfkFAKyRBQCAsQcAgbkHAIJBAACzsQYAg6kAgLVZBwC2MQcAh6kAgIupAICPqQCAuuEHALvhBwC84QcAveEHAL7hBwC/3QcAqLUGAKm5BgCqdQYAq4UHAKydBwCt/QcArvUHAK8ZBwCTqQCAl6kAgJupAICfqQCAo6kAgKepAICrqQCAr6kAgLh1BwC5fQcAunUHALsFBwC8HQcAvTEHAL4xBwC/MQcAsGkHALFpBwCyeQcAs3kHALRpBwC1VQcAtlEHALdNBwCj/QcAs6kAgLepAIC7qQCAv6kAgKZ9BgClFQYAw6kAgKutBgCqrQYAx6kAgMupAICvkQYArq0GAK2tBgCsrQYAz6kAgNOpAIDXqQCA26kAgIAdAACBCQAAgjkAAN+pAIDjqQCA66kAgIbIAACHpAEA76kAgPOpAID3qQCA+6kAgKiNAQCpmQEAqtkBAKvRAQCs8QEArfEBAK45AQCvOQEAhKAAAP+pAIADqgCAB6oAgAuqAIAPqgCAE6oAgBeqAIC4zQAAudUAALrVAAC75QAAvP0AAL2VAAC+nQAAv5UAALBJAQCxSQEAslkBALNZAQC0SQEAtUkBALb9AAC39QAAugUEALsJBAC44QcAueEHAL4JBAC/CQQAvAkEAL0JBACyjQcAs+UHALC1BwCxhQcAtuUHALftBwC08QcAtfEHAKpNBwCrVQcAqEkHAKlJBwCu3QcAr8UHAKxNBwCt1QcAG6oAgB+qAIAjqgCAJ6oAgCuqAIAvqgCAM6oAgDeqAICz0QIAO6oAgD+qAIC+AAwAQ6oAgLbxAgC1+QIAR6oAgLsNAgC6DQIAS6oAgE+qAIC/DQIAvg0CAL0NAgC8DQIAghUAAKOVAgCAYQAAgWEAAKa1AgBTqgCAW6oAgKW9AgCqSQIAq0kCAIbIDACHrAwArkkCAK9JAgCsSQIArUkCAKhlAgCpdQIAqn0CAKt1AgCsbQIArbECAK6xAgCvsQIAhKANAF+qAIBjqgCAZ6oAgGuqAIBvqgCAc6oAgHeqAIC4MQEAuTEBALoxAQC7MQEAvNUBAL3dAQC+yQEAv8EBALDRAgCx0QIAstECALPRAgC0EQEAtREBALYRAQC3EQEA4bAGAHuqAIDj0AYAhEAPAH+qAIDhpAEAg6oAgOPABgCHqgCAi6oAgI+qAIDv1AYA7AAAAJOqAIDvZAcAl6oAgJuqAICfqgCAo6oAgLO5AgCnqgCAtakCALZ9AgCrqgCAr6oAgLOqAIC6WQIAu1kCALxJAgC9SQIAvpkBAL+ZAQCjdQ0AV6oAgLeqAIC7qgCAv6oAgKaxDQClZQ0Aw6oAgKuVDQCqlQ0AvqQDAMeqAICvVQ4ArlUOAK2FDQCshQ0AgE0AAIFVAACCVQAAs2UPAMuqAIC1ZQ8Atm0PAM+qAICGQAMAhxQDALrtDwC7/Q8AvOkPAL3VDwC+3Q8Av9UPAKhZDgCpoQ8AqqEPAKuhDwCsoQ8AraEPAK6hDwCvoQ8A06oAgNeqAIDbqgCA36oAgOOqAIDnqgCA66oAgO+qAIC4AQ8AuQEPALoBDwC7HQ8AvA0PAL01DwC+PQ8Av9UAALBlDwCxdQ8AsnEPALNNDwC0VQ8AtV0PALZNDwC3QQ8AoykOAPOqAID3qgCA+6oAgP+qAICmIQ4ApSkOAAOrAICrsQ4AqqEOAAerAIALqwCAr5kOAK6RDgCtmQ4ArKUOAA+rAIATqwCAF6sAgBurAIDvJA0AH6sAgCOrAIAnqwCA49AOACurAIDhGA4AL6sAgIAVAACBGQAAggUAADOrAICo0QEAqdkBAKopAQCrKQEArDkBAK05AQCuKQEArykBAL5oAQA7qwCAhsgBAIesAAA/qwCAQ6sAgEerAIBLqwCAuO0AALmFAAC6jQAAu4UAALydAAC9gQAAvoEAAL+BAACwWQEAsVkBALLtAACz5QAAtP0AALXlAAC25QAAt9UAALOhAgBPqwCAU6sAgFerAIBbqwCAtrkCALWxAgBfqwCAu50CALqdAgBjqwCAZ6sAgL8hAwC+OQMAvTEDALw5AwCF+PUAo+UCAGurAIBvqwCApv0CAHOrAIB3qwCApfUCAKrZAgCr2QIAe6sAgH+rAICufQMAr2UDAKx9AwCtdQMAuOkAALnpAAC6aQAAu2kAALx5AAC9ZQAAvm0AAL9lAACwsQAAsbkAALKBAACzgQAAtPkAALX5AAC27QAAt+UAAKhlAwCpdQMAqn0DAKt1AwCsbQMArdEAAK7RAACv0QAAg6sAgIerAICLqwCA56kAgI+rAICTqwCAl6sAgJurAICA/QEAgQkAAIIZAACfqwCAo6sAgL5EAgCrqwCAr6sAgISsAgCzqwCAh/gCAIasBQC3qwCAu6sAgL+rAIDDqwCAs/UCAMerAIDLqwCAz6sAgNOrAIC2UQEAteUCANerAIC7fQEAunUBANurAIDfqwCAvz0BAL49AQC9VQEAvFUBAOFwDwDjqwCA47gOAITABQDvyAAA56sAgOurAIDvqwCA4zwOAPOrAIDh0AEA96sAgIR0BwD7qwCA72gBAP+rAIADrACApXkCAKbNAQAHrACAgCEAAIEhAACC3QcAo2kCAKzJAQCtyQEArqEBAK+hAQALrACAD6wAgKrpAQCr4QEAp6sAgBOsAIC+QAIAF6wAgIYwAwCHMAMAG6wAgB+sAICoOQcAqTkHAKoNBwCrHQcArAUHAK0NBwCuBQcAr3kHALAJBwCxCQcAshkHALMRBwC0OQcAtTkHALbdBwC3yQcAuPkHALn5BwC6zQcAu8EHALzFBwC9yQcAvrkHAL+xBwCzpQcAI6wAgCesAIArrACAL6wAgLatBwC1rQcAM6wAgLvtBwC67QcAN6wAgDusAIC/3QcAvt0HAL3lBwC87QcAP6wAgKPhBwBDrACAR6wAgKbpBwBLrACAT6wAgKXpBwCqqQcAq6kHAFOsAIBXrACArpkHAK+ZBwCsqQcAraEHAFusAIBfrACAY6wAgGesAIBrrACAb6wAgHOsAIB3rACAgREAAIANAAB7rACAghkAAH+sAICDrACAvuQBAIesAICG4AAAhxgBAIusAICPrACAk6wAgJesAICbrACA77AEAJ+sAIDh1AYAo6wAgONcBACnrACAq6wAgK+sAICzrACAqJkBAKmZAQCqDQEAqwUBAKwdAQCtBQEArgUBAK81AQCEiAEAt6wAgLusAIC/rACAw6wAgMesAIDLrACAz6wAgLjBAAC5wQAAusEAALvBAAC8wQAAvcEAAL7BAAC/wQAAsE0BALElAQCyIQEAsyEBALQlAQC1LQEAthEBALcRAQDTrACA16wAgLONAgDbrACAtZ0CAN+sAIDjrACAto0CAOesAIDrrACAu+kCALqBAgC9/QIAvP0CAL/hAgC+6QIA76wAgKbVAgClxQIAvggDAKPVAgCCLQAAgRkAAIB5AACvuQIArrECAK2lAgCspQIAq7ECAKrZAgDzrACA+6wAgO80AgD/rACAhxgDAIYs/AADrQCAB60AgAutAIAPrQCAE60AgBetAIAbrQCAH60AgOMAAQAjrQCA4eABACetAIC6tQMAu70DACutAIAvrQCAvnkDAL95AwC8pQMAvXkDADerAICztQMAM60AgDetAIC2kQMAO60AgD+tAIC1pQMAqEkCAKlJAgCqWQIAq1kCAKxJAgCtdQIArnECAK9tAgC+aP0AvqT/AEOtAIBHrQCAS60AgE+tAIBTrQCAV60AgLj5AgC5+QIAukkBALtJAQC8XQEAvUEBAL5BAQC/fQEAsBUCALEdAgCyFQIAs8kCALTZAgC12QIAtskCALfJAgDjIAYA4bAGAOGAAQDjEAYAgA0AAIE1AACCPQAAW60AgF+tAIBjrQCAa60AgG+tAIDvcAAAc60AgHetAIDvTAEAhIz9AHutAICjmQIAf60AgKWJAgCDrQCAh60AgKa9AgCGwPwAh+T8AKuRAgCqmQIArVUCAKyJAgCvVQIArlUCAKh9/gCpgf4Aqpn+AKuZ/gCsif4ArYn+AK65/gCvuf4AZ60AgIutAICPrQCAk60AgJetAICbrQCAn60AgKOtAIC4tf4Aub3+ALph/wC7Yf8AvGH/AL1h/wC+Yf8Av2H/ALDJ/gCxyf4Ast3+ALPR/gC0uf4Atbn+ALaR/gC3kf4AsxH+AKetAICrrQCAr60AgLOtAIC2Cf4AtQH+ALetAIC7Df4Aug3+ALutAIC/rQCAv33+AL59/gC9Bf4AvAn+AMOtAICjVf4Ax60AgMutAICmTf4Az60AgNOtAIClRf4Aqkn+AKtJ/gCEKAMA160AgK45/gCvOf4ArE3+AK1B/gCAzQEAgdEBAILRAQCzuf4A260AgLXR/gC21f4A360AgIZgAQCHYAEAug0BALsFAQC8HQEAvQUBAL4NAQC/BQEA460AgOetAIDrrQCA760AgPOtAIDhwP0A960AgOOM/AD7rQCA/60AgAOuAIDvtPwAB64AgAuuAIAPrgCAE64AgKgp/gCpKf4Aqj3+AKs1/gCsVf4ArVn+AK5N/gCvRf4AF64AgBuuAIAfrgCAI64AgCeuAIArrgCAL64AgDOuAIC4SQEAuUkBALpZAQC7UQEAvHkBAL15AQC+GQEAvxUBALDFAQCxzQEAssUBALPdAQC0xQEAtc0BALbFAQC3eQEAN64AgDuuAIA/rgCAo7n9AEOuAICl0f0AptX9AITQAwBSrgCAvuACAKoNAgCrBQIArB0CAK0FAgCuDQIArwUCAIFJAACAQQAAowkDAIJdAAClGQMAVq4AgFquAICmEQMAhsAEAIfkAwCrDQMAqg0DAK0BAwCsHQMArwEDAK4JAwCw4QMAseEDALLhAwCz/QMAtOUDALXtAwC25QMAtz0DALgFAwC5DQMAugUDALsdAwC8BQMAvQ0DAL4FAwC/vQAAXq4AgGKuAIBmrgCAaq4AgPesAIBurgCAcq4AgHauAICo8QMAqfkDAKqpAwCrqQMArLkDAK25AwCuqQMAr6UDALNBAgB6rgCAfq4AgIKuAICGrgCAtlkCALVRAgCKrgCAu0UCALpFAgCOrgCAkq4AgL9JAgC+QQIAvUkCALxVAgCWrgCAmq4AgJ6uAICirgCA74wDAKauAICqrgCArq4AgONsAwCyrgCA4VAAALauAIC6rgCAvngFAMKuAICEcAIAgOUAAIHpAACC+QAAxq4AgIawBACHVAUAyq4AgO9A/gDOrgCA4Vz+ANKuAIDjVAEA1q4AgNquAIDergCA4q4AgLOZAQDmrgCA6q4AgO6uAIDyrgCAth0BALUdAQD2rgCAuz0BALo9AQD6rgCA/q4AgL/hAAC++QAAvfEAALz5AACoIQYAqVEGAKpRBgCrzQYArNUGAK3dBgCu1QYAr8kGAL6uAIACrwCABq8AgAqvAIAOrwCAEq8AgBavAIAarwCAuG0HALkFBwC6DQcAuwUHALwdBwC9AQcAvgEHAL8BBwCwuQYAsbkGALJtBwCzZQcAtH0HALVlBwC2ZQcAt1UHAKPZBgAerwCAIq8AgCavAIAqrwCApl0GAKVdBgCEnAIAq30GAKp9BgC+JAMALq8AgK+hBwCuuQcArbEHAKy5BwCASQAAgUkAAIJZAACzVQcAMq8AgLV9BwC2aQcANq8AgIZAAACHVAMAulUHALspBwC8OQcAvTkHAL4pBwC/IQcAo5kGADqvAIA+rwCAQq8AgEavAICmpQYApbEGAEqvAICr5QYAqpkGAE6vAIBSrwCAr+0GAK7lBgCt9QYArPUGAOE4BQBWrwCA4yQEAFqvAIBerwCAYq8AgGavAIBqrwCAbq8AgHKvAIB2rwCAeq8AgH6vAICCrwCA7/QEAIavAICo+QYAqQkGAKoRBgCrLQYArDkGAK0lBgCuLQYAryUGAIqvAICOrwCAkq8AgJavAICAGQAAgRkAAIIFAACarwCAuOUBALntAQC65QEAu/0BALzlAQC97QEAvuUBAL9ZAQCwXQYAsSEGALIhBgCzIQYAtCEGALUpBgC2EQYAtxEGAKjRAgCp2QIAqg0DAKsFAwCsHQMArQUDAK4FAwCvNQMAvmQCAKKvAICmrwCAqq8AgK6vAICyrwCAtq8AgLqvAIC4JQMAuS0DALolAwC7PQMAvCUDAL0pAwC++QMAv/kDALBNAwCxIQMAsiUDALM9AwC0JQMAtS0DALYlAwC3HQMAs4UDAITIAgC+rwCAhAgDAMKvAIC2hQMAtZUDAMavAIC75QMAuokDAIYIDACHnAMAv+kDAL7hAwC96QMAvPEDAIXsCgBHrgCAo80DAMqvAICl3QMAzq8AgNKvAICmzQMA1q8AgNqvAICrrQMAqsEDAK2hAwCsuQMAr6EDAK6pAwDerwCA4q8AgOavAIDqrwCA78gDAO6vAIDyrwCA9q8AgOO0AwD6rwCA4dABAP6vAICADQAAgXUAAIJ9AAACsACABrAAgAqwAICzZQEAvgQCALVlAQASsACAFrAAgLZlAQCGQA0Ah1gNALv1AQC6/QEAvaUBALy5AQC/mQEAvqUBABqwAIAesACAIrAAgIQADAAmsACAKrAAgC6wAIDvzAEAMrAAgOEsBgA2sACA4yABAOwAAAA6sACAPrAAgEKwAIBGsACAo+kBAEqwAIBOsACApukBAFKwAIBWsACApekBAKpxAQCreQEAWrAAgF6wAICuKQEArxUBAKw1AQCtKQEAqCUOAKktDgCqJQ4Aqz0OAKwlDgCtLQ4AriUOAK+VDgAOsACAYrAAgGawAIBqsACAbrAAgIKdAACBnQAAgJ0AALhFDwC5TQ8AukUPALtZDwC8SQ8AvUkPAL59DwC/cQ8AsPEOALH5DgCypQ4As7kOALSpDgC1lQ4Atp0OALd9DwCo1Q8Aqd0PAKoJDwCrCQ8ArBkPAK0FDwCuDQ8ArwUPAHKwAIB2sACAerAAgL6gAwB+sACAgrAAgId4AwCGEAAAuBUPALkdDwC6IQ8AuyEPALz1AAC9/QAAvvUAAL/tAACwQQ8AsU0PALJdDwCzVQ8AtE0PALU1DwC2MQ8AtzEPAIawAIDvsAwAirAAgI6wAICSsACAlrAAgJqwAICesACAorAAgKawAICqsACArrAAgLKwAIDjqA0AtrAAgOGMDQCzwQ4AurAAgL6wAIDCsACAxrAAgLbFDgC10Q4AyrAAgLvJDgC6xQ4AzrAAgNKwAIC/sQ4AvskOAL3BDgC8yQ4AowEOANawAIDasACA3rAAgOKwAICmBQ4ApREOAOawAICrCQ4AqgUOAOqwAICErAIAr3EOAK4JDgCtAQ4ArAkOAIBRAACBWQAAgmEAALPFAAC+zAEAtcUAALbNAADysACAhkAHAIcUAQC6yQAAu8kAALzZAAC92QAAvskAAL/FAACrDQMAqg0DAKkJAwCouQIArw0DAK4NAwCtDQMArA0DAL5gAwD2sACA+rAAgP6wAIACsQCABrEAgAqxAIC+MAUAuykDALoZAwC5GQMAuAEDAL/dAwC+3QMAvd0DALwxAwCzTQMAsk0DALFNAwCwTQMAtzkDALYxAwC1QQMAtE0DAA6xAICmkQMApZkDABKxAICjmQMAFrEAgBqxAIAesQCAr5kDAK6VAwCthQMArIUDAKuVAwCqlQMAnq8AgCKxAIAmsQCAKrEAgC6xAIAysQCANrEAgDqxAIA+sQCAQrEAgEaxAIBKsQCATrEAgFKxAICAHQAAgQkAAIL9AQBWsQCAvwgHAFqxAIBisQCA7yQAAGaxAICElAIAarEAgG6xAICH4AIAhgQFAL4AGABysQCAdrEAgOGQAQB6sQCA44AAAH6xAICCsQCAhrEAgLNlAQCKsQCAtWUBALZtAQCOsQCAkrEAgJaxAIC65QEAu/kBALzpAQC96QEAvsUBAL+9AQCasQCAnrEAgKKxAIC+xBkAprEAgKqxAICusQCA78gBALKxAIDh3A4AtrEAgOMwDgC6sQCAvrEAgMKxAICEMAQAgHkAAIEVAACCFQAAo+UBAMaxAICl5QEApu0BAMqxAICGQAYAh5AHAKplAQCreQEArGkBAK1pAQCuRQEArz0BAKjdBQCpIQYAqiEGAKshBgCsIQYArSEGAK4hBgCvnQYAXrEAgM6xAIDSsQCAhDABANaxAIDasQCA3rEAgOKxAIC4jQYAuZUGALqdBgC7lQYAvI0GAL21BgC+vQYAv7UGALDtBgCx8QYAsvEGALPxBgC0zQYAtbUGALa9BgC3tQYAqIkHAKmVBwCqkQcAq5EHAKy9BwCtpQcArqEHAK/dBwDmsQCA6rEAgO6xAIDysQCA9rEAgPqxAID+sQCAArIAgLhJBwC5VQcAul0HALtVBwC8cQcAvX0HAL5pBwC/aQcAsKUHALGtBwCyuQcAs7EHALSRBwC1kQcAtnkHALd5BwAGsgCACrIAgA6yAIASsgCA78gFAOHACQAWsgCA48AZAOMkBAAasgCA4dAGAO/cKACinQMAoxUBAKAZBQChjQUAs1kGAB6yAIAisgCAJrIAgCqyAIC2ZQYAtXUGAC6yAIC7KQYAuiEGADKyAIA2sgCAvxUGAL4VBgC9JQYAvC0GAKOZBgCPmfwAOrIAgEKyAIBGsgCApqUGAKW1BgBKsgCAq+kGAKrhBgCGKB8Ah5wAAK/VBgCu1QYAreUGAKztBgCebQkAn30HAJwNCwCd7QkAmvENAJs5DQCY5fAAmQ0PAJbh8QCX6fEAlMX1AJUN8wCSHfcAk/H1AJD9+QCR7fkAgh3/AIMB+gBOsgCAUrIAgIYV9gCHOfYAhAn6AIXx9ACKwfAAiyXyAFayAIBasgCAjuEMAI8VDgCMNfIAjQHzAJKtDgCTgQgAXrIAgGKyAICW6QQAl3UGAJR5CgCV8QoAmtEGAJvJAABmsgCAarIAgIEdAwCAHQMAnFkCAIL1AwCrARAAqpUWAKmNFgCojRYAr5UuAK4BLACt/RIArJkSAKOlHgCipR4AoY0CAO6wAICnGRoAppUaAKUBGACknR8AbrIAgHKyAIB2sgCAerIAgH6yAICCsgCAhrIAgIqyAICz5SoAsuUqALGtLwCw5S4AjrIAgJKyAIC1ASQAtBEqAKgpAwCpNQMAqj0DAKs1AwCsLQMArbUDAK69AwCvtQMAlrIAgJqyAICesgCAorIAgIAdAACBCQAAgrkAAKayAIC4TQIAuV0CALptAgC7CQIAvBkCAL0ZAgC+CQIAvwECALDNAwCx1QMAst0DALPVAwC0zQMAtXUCALZ9AgC3dQIAqrIAgITIHQCysgCAvgwfALayAIC6sgCA70gGAO9YBwDhWAYA4ZgGAOOUAQDjAAYAhhAcAId8HQC+9B4AvrIAgMKyAIC2ZQMAtfUDAMayAICz5QMAyrIAgM6yAIDSsgCAv+ECAL5ZAwC9UQMAvFkDALtBAwC6WQMA1rIAgNqyAIA+sgCArrIAgN6yAIDisgCA5rIAgOqyAIDusgCA8rIAgKitHQCptR0AqrUdAKslHgCsPR4ArR0eAK4VHgCvdR4AsA0eALEtHgCyJR4As40eALSVHgC1nR4AtpUeALeNHgC4tR4Aub0eALq1HgC7nR4AvIUeAL1VHwC+XR8Av1UfALMdHQD2sgCA+rIAgP6yAIACswCAtr0eALWVHgAGswCAu8keALrpHgAKswCADrMAgL95HgC+cR4AvXkeALzRHgCCKQAAo1kdAIAdAACBFQAApvkeABKzAIAWswCApdEeAKqtHgCrjR4AGrMAgITgAwCuNR4Arz0eAKyVHgCtPR4AqIkeAKmVHgCqnR4Aq7EeAKzRHgCt2R4Ars0eAK/FHgAeswCAIrMAgIaIAACHbAEAJrMAgCqzAIAuswCAMrMAgLhdAQC5wQEAusEBALvBAQC8wQEAvckBAL7xAQC/8QEAsL0eALGdHgCylR4As2UBALR9AQC1ZQEAtm0BALdlAQCqLR0AqzUdADazAIA6swCAri0dAK+VHACsLR0ArSUdAISMAQCjkR0APrMAgEKzAICmER0ARrMAgEqzAIClgR0As1UeAE6zAIBSswCAVrMAgFqzAIC2GR4AtRkeAF6zAIC7GR4AujkeAGKzAIBmswCAv+EBAL75AQC98QEAvAEeAGqzAIBuswCAcrMAgKOZHQB2swCApdUdAKbVHQB6swCAfrMAgIKzAICq9R0Aq9UdAKzNHQCtPQIArjUCAK8tAgCAZQAAgRUAAIIdAACEAAQAhrMAgIqzAICHcAMAhvwEAJKzAICWswCAmrMAgJ6zAICiswCAprMAgKqzAICuswCAvsgEALKzAIC2swCAurMAgL6zAIDCswCAxrMAgO/cHwDKswCA4ZQBAM6zAIDjHAEA0rMAgNazAIDaswCA3rMAgLt1AwC6aQMAvkgGAOKzAIC/HQMAvh0DAL0dAwC8ZQMAs9UDAOazAIDqswCA7rMAgPKzAIC2fQMAtcUDAIRwBQCoJQIAqTUCAKo9AgCrNQIArC0CAK2dAgCulQIAr7UCAIIVAAD2swCAgNkBAIEJAADEAAAA+rMAgAK0AIAGtACAuKkCALmpAgC6SQEAu0kBALxZAQC9RQEAvkUBAL99AQCwzQIAsdECALLRAgCzqQIAtLkCALW5AgC2qQIAt6ECAOEoHgDhNBwA43QBAOMYHgAKtACADrQAgIa4BACHVAUAhDgHABK0AIAWtACAGrQAgL6sBwAetACA78weAO/IGgCj9QIAIrQAgCa0AIAqtACALrQAgKZdAgCl5QIAMrQAgKtVAgCqSQIANrQAgDq0AICvPQIArj0CAK09AgCsRQIAqGEGAKlhBgCqYQYAq2EGAKxhBgCtYQYArmEGAK9hBgD+swCAPrQAgEK0AIBGtACASrQAgE60AIBStACAVrQAgLjxBgC58QYAuvEGALvxBgC8nQYAvbEGAL6xBgC/sQYAsOUGALHtBgCy5QYAs/0GALTlBgC17QYAttkGALfVBgCz6QYAWrQAgF60AIBitACAZrQAgLbhBgC16QYAarQAgLspBgC6IQYAbrQAgHK0AIC/KQYAviEGAL0pBgC8MQYAgl0AAKOtBgCARQAAgV0AAKalBgB2tACAerQAgKWtBgCqZQYAq20GAIYADACHQAMArmUGAK9tBgCsdQYArW0GAH60AIDvfAUAgrQAgIa0AICKtACAjrQAgJK0AICWtACAmrQAgJ60AICitACAprQAgKq0AIDjaAUArrQAgOF4BQCz0QYAsrQAgLa0AIC6tACAvrQAgLb9BgC1/QYAwrQAgLupBgC6oQYAxrQAgMq0AIC/mQYAvqkGAL2pBgC8sQYAqLkGAKm5BgCqGQYAqxkGAKw1BgCtPQYArjUGAK8pBgDOtACAgh0AAIEdAACAHQAA0rQAgNa0AIDatACA4rQAgLjpAQC56QEAuvkBALv5AQC86QEAvekBAL5dAQC/VQEAsCUGALEtBgCyJQYAsz0GALQtBgC1HQYAthUGALfZAQCGgAwAh+QCAOa0AICjnQUA6rQAgKWxBQCmsQUA7rQAgPK0AID2tACAqu0FAKvlBQCs/QUAreUFAK7lBQCv1QUAtk0DAPq0AICExAMAtUUDAP60AICzjQIAArUAgAa1AIC+SQMAv0kDALxJAwC9SQMAumkDALtpAwAKtQCADrUAgBK1AICmiQMApYEDABa1AICjSQIAGrUAgB61AIAitQCAr40DAK6NAwCtjQMArI0DAKutAwCqrQMAjrMAgCa1AIAqtQCALrUAgIW0PQAytQCANrUAgDq1AIA+tQCAQrUAgIA9AACBCQAAgh0AAEa1AIC+sAMASrUAgIc4AwCG3AwAUrUAgFa1AIBatQCAXrUAgGK1AIDvXAYAZrUAgGq1AIC+6AwA45QGAG61AIDh3AEAcrUAgHa1AIB6tQCAfrUAgLNRAQCCtQCAhrUAgIq1AICOtQCAtnEBALV5AQCStQCAuz0BALo9AQCWtQCAmrUAgL/9AQC+9QEAvQUBALwFAQCetQCAorUAgKa1AICEQAwAqrUAgK61AICytQCA76wHALa1AIDhJAYAurUAgONABwCGkAwAh/wMAMK1AIDGtQCAgFkAAIFlAACCYQAAo90BAMq1AICl9QEApv0BAM61AIDStQCA1rUAgKqxAQCrsQEArIkBAK2JAQCueQEAr3EBAN60AIBOtQCA2rUAgN61AIC+tQCA4rUAgOa1AIDqtQCAqJ0NAKktDgCqOQ4AqzEOAKwRDgCtEQ4Arn0OAK9tDgCwGQ4AsRkOALIxDgCzMQ4AtNEOALXZDgC2zQ4At8UOALj9DgC52Q4AuqkOALupDgC8vQ4AvaUOAL6tDgC/pQ4AqIEPAKmBDwCqgQ8Aq4EPAKyBDwCtjQ8AroUPAK+1DwDutQCA8rUAgPa1AID6tQCA/rUAgAK2AIAGtgCACrYAgLidDwC5rQ8AuqUPALtNDwC8VQ8AvV0PAL5JDwC/SQ8AsNEPALHRDwCy0Q8As9EPALS1DwC1vQ8AtrUPALetDwCzCQ4ADrYAgBK2AIAWtgCAGrYAgLYNDgC1CQ4AHrYAgLsVDgC6FQ4AIrYAgCa2AIC/eQ4AvnEOAL0FDgC8BQ4AghUAAKNNDgCAYQAAgWEAAKZJDgAqtgCAvhABAKVNDgCqUQ4Aq1EOAIQkAQAytgCArjUOAK89DgCsQQ4ArUEOAKg5DgCpOQ4AqlkOAKtRDgCscQ4ArXEOAK6RAQCvkQEAhgAAAIeEAAA2tgCAOrYAgD62AIBCtgCARrYAgEq2AIC4dQEAuX0BALp1AQC7yQAAvNkAAL3ZAAC+yQAAv8EAALD1AQCx/QEAsvUBALNNAQC0VQEAtV0BALZVAQC3TQEAuk0PALtVDwC4TQ8AuUUPAL59DwC/tQ8AvEUPAL11DwCyAQ8AswEPALAxDwCxMQ8AtgEPALcNDwC0EQ8AtREPAKqZDgCrRQ8AqOUOAKmZDgCuQQ8Ar0EPAKxRDwCtUQ8ATrYAgFK2AIBWtgCAWrYAgF62AIBitgCAZrYAgGq2AICzUQ0AbrYAgHK2AIB2tgCAerYAgLZxDQC1eQ0AfrYAgLu5AgC6sQIAgrYAgIa2AIC/GQIAvhECAL0ZAgC8oQIAirYAgKMVDQCOtgCAkrYAgKY1DQCWtgCAmrYAgKU9DQCq9QIAq/0CAIToAwCitgCArlUCAK9dAgCs5QIArV0CAKhtAgCprQIAqqUCAKu9AgCspQIAra0CAK6lAgCvfQEAgO0BAIHxAQCC8QEAvqAFAKa2AICqtgCAh2gFAIYcBQC4yQEAuckBALrZAQC70QEAvPkBAL35AQC+mQEAv5UBALAFAQCxDQEAsgUBALMdAQC0BQEAtQ0BALYFAQC3+QEA4WQPAOGcDwDjFA4A49QPAK62AIDhPA4AsrYAgOPkAAC+rAQAtrYAgLq2AIDvDAAAvrYAgMK2AIDvYA4A77QPAMa2AIDKtgCAhEQEALNhAgDOtgCAtWECALZhAgDStgCA1rYAgNq2AIC6jQEAu4UBALydAQC9hQEAvo0BAL+FAQCjrQUAnrYAgN62AIDitgCA5rYAgKatBQClrQUA6rYAgKtJBgCqQQYA7rYAgPK2AICvSQYArkEGAK1JBgCsUQYA9rYAgPq2AID+tgCAArcAgIAdAACBCQAAgjkAAAa3AIAKtwCADrcAgIbIAACHIAMAErcAgBa3AIAatwCAHrcAgKhtBgCptQcAqr0HAKsdBwCsCQcArTEHAK4xBwCvLQcAhKgDACK3AIAmtwCAKrcAgC63AIAytwCANrcAgDq3AIC4zQAAudUAALrVAAC75QAAvP0AAL2VAAC+nQAAv5UAALBVBwCxJQcAsi0HALM9BwC0LQcAtRUHALYdBwC39QAAPrcAgOG8BgBCtwCA4/QFAEa3AIBKtwCATrcAgFK3AIBWtwCAWrcAgF63AIBitwCAZrcAgGq3AIButwCA7+gEALN1BgCCLQAAgRUAAIAdAABytwCAtvEGALXBBgB2twCAu6EGALrRBgB6twCAvmwBAL+RBgC+qQYAvakGALy5BgCjtQYAgrcAgIYoAACHTAEAhrcAgKYxBgClAQYAircAgKthBgCqEQYAjrcAgJK3AICvUQYArmkGAK1pBgCseQYAlrcAgLO9AQCatwCAnrcAgLZ5AQCitwCAprcAgLV5AQC6VQEAu10BAKq3AICutwCAvvkAAL/lAAC8RQEAvf0AAKhxAgCpcQIAqnECAKtxAgCstQIArb0CAK61AgCvrQIAhOw8ALK3AIC2twCAurcAgL63AIDCtwCAxrcAgMq3AIC4XQMAuWUDALptAwC7ZQMAvH0DAL1lAwC+bQMAv2UDALDVAgCx3QIAstUCALNtAwC0eQMAtWUDALZtAwC3ZQMALrYAgM63AIDStwCAo/UCANa3AIClMQIApjECANq3AIDetwCA4rcAgKodAgCrFQIArA0CAK21AwCusQMAr60DAIBlAACBCQAAghkAAOa3AIDqtwCA8rcAgL4QPAD2twCAhsA8AIcgAwD6twCA/rcAgAK4AIAGuACACrgAgA64AICohQIAqZUCAKqVAgCrpQIArL0CAK3VAgCu0QIAr9ECABK4AIAWuACAGrgAgB64AIAiuACAJrgAgCq4AIAuuACAuHUBALl9AQC6dQEAu8kBALzZAQC9xQEAvsUBAL/9AQCwtQIAsb0CALKBAgCzgQIAtFUBALVdAQC2VQEAt00BAOGkBgAyuACA41AGAL6APACEHDwAvoA/ADa4AIA6uACAPrgAgEK4AIBGuACASrgAgE64AIBSuACA7+AGAFa4AICBfQAAgHEAAFq4AICCBQAAYrgAgGa4AIDvTAAAargAgOGQAQBuuACA41gBAHK4AIB2uACAergAgIZYPwCH/DwAs509AO63AIBeuACAfrgAgIK4AIC21T0AtbU9AIa4AIC7+T0AuvE9AIq4AICOuACAvxk+AL4RPgC91T0AvNU9AJK4AICj2T0AlrgAgJq4AICmkT0AnrgAgKK4AICl8T0AqrU9AKu9PQCmuACAqrgAgK5VPgCvXT4ArJE9AK2RPQCoVT4AqVk+AKphPgCrYT4ArGE+AK1hPgCuYT4Ar2E+AISoAwCuuACAsrgAgLa4AIC6uACAvrgAgMK4AIDGuACAuEU/ALldPwC6VT8Au20/ALx1PwC9fT8AvnU/AL9tPwCwwT8AscE/ALLBPwCzwT8AtME/ALXBPwC2wT8At8E/AIC5AQCBuQEAggUAAMq4AIDhgD4A0rgAgOMoPQDWuACAhoAAAIcEAQDvCD0A2rgAgN64AIDiuACA5rgAgOq4AICzqT8AzrgAgO64AIDyuACA9rgAgLahPwC1qT8A+rgAgLtFPgC6RT4A/rgAgAK5AIC/RT4AvkU+AL1VPgC8VT4Ao2k/AAa5AIAKuQCADrkAgBK5AICmYT8ApWk/ABa5AICrhT4AqoU+ABq5AIAeuQCAr4U+AK6FPgCtlT4ArJU+ACK5AICzGT4AJrkAgCq5AIC2IT4ALrkAgDK5AIC1MT4AuvEBALv5AQA2uQCAOrkAgL6xAQC/vQEAvNEBAL3RAQCo0T0AqdE9AKrVPQCr6T0ArP09AK3lPQCu7T0ArxECAID5AwCBzQMAgsUDAIQkAwC+AAQAQrkAgIesAwCGvAQAuBkCALktAgC6JQIAu+kCALz5AgC9+QIAvukCAL/pAgCwcQIAsXkCALJBAgCzQQIAtDECALU9AgC2NQIAtykCAKVtPQBGuQCASrkAgKZ9PQBOuQCAfrcAgKNFPQBSuQCArY0CAKyNAgCv4QIAru0CAKwAAABWuQCAq6UCAKqtAgDh+AEAWrkAgOP0AgCEwAQAXrkAgGK5AIBmuQCAarkAgG65AIByuQCAdrkAgHq5AIB+uQCAgrkAgO8wAgCGuQCAqBUCAKkZAgCqJQIAqz0CAKwlAgCtLQIAriUCAK9VAgCKuQCAjrkAgJK5AICWuQCAmrkAgJ65AICEsAQAorkAgLjRAgC52QIAuuECALvhAgC8kQIAvZ0CAL6VAgC/iQIAsC0CALE1AgCyNQIAswUCALQdAgC18QIAtvECALfxAgDheD8A4zQBAOMIPgDhbD4AgQkAAICpAACmuQCAgj0AAKq5AICyuQCAtrkAgL4gBAC6uQCA79g+AO/MPgC+uQCAwrkAgLPpAgCG6AQAh8AEALbpAgDGuQCAyrkAgLXpAgC6rQIAu7UCAM65AIDSuQCAvp0CAL9xAgC8pQIAvZUCAD65AICuuQCA1rkAgNq5AIDeuQCA4rkAgOa5AIDquQCAqBUGAKmhBgCqoQYAq70GAKytBgCtgQYArv0GAK/tBgCwlQYAsZ0GALKVBgCzrQYAtLUGALW9BgC2tQYAt60GALiVBgC5mQYAukkHALtJBwC8WQcAvVkHAL5JBwC/SQcArN0FAK3tBQCu5QUArwkFAO65AIDyuQCAqtUFAKvNBQD2uQCApZEFAKaRBQD6uQCA/rkAgAK6AIAGugCAo5EFALNJBgAKugCADroAgBK6AIAWugCAtmEGALVFBgAaugCAuzkGALoxBgC+ZAAAHroAgL8ZBgC+EQYAvRkGALwhBgCjiQcAgtkBAIHZAQCAwQEAIroAgKahBwClhQcAJroAgKv5BwCq8QcAhggBAId8AQCv2QcArtEHAK3ZBwCs4QcAKroAgLP1BgAuugCAMroAgLaFBgA2ugCAOroAgLWdBgC6jQYAu20BAD66AIBCugCAvmUBAL9tAQC8dQEAvW0BAKglBgCpLQYAqjkGAKsxBgCsUQYArUEGAK5BBgCvdQYARroAgEq6AIBOugCAUroAgFa6AIBaugCAXroAgGK6AIC4VQEAuWUBALplAQC7fQEAvGUBAL1tAQC+HQEAvxUBALANBgCx7QEAsuUBALP9AQC05QEAte0BALblAQC3bQEAo7EFAGa6AIBqugCAvkgDAL5YDACmwQUApdkFAG66AICrKQIAqskFAHK6AIB2ugCArykCAK4hAgCtKQIArDECAHq6AIB+ugCAgroAgIa6AICAGQAAgRkAAIIFAACKugCAhKwDAJK6AICHGAMAhswMAJa6AICaugCAnroAgKK6AICokQMAqZkDAKrJAwCrxQMArN0DAK3BAwCuwQMAr/UDAKa6AICqugCArroAgLK6AIC2ugCAuroAgL66AIDCugCAuH0DALnBAAC6wQAAu9EAALz5AAC9+QAAvpkAAL+ZAACwjQMAsUUDALJNAwCzRQMAtF0DALVFAwC2TQMAt0UDALNBAgDGugCAyroAgL8EDwDOugCAtkECALVVAgDSugCAu4ECALpJAgDWugCA2roAgL+BAgC+mQIAvZECALyZAgDeugCA4roAgOa6AIDqugCA76QDAO66AIDyugCA9roAgOMQAwD6ugCA4VgAAIQgDQCAKQAAgSkAAIIdAAACuwCA4VAGAOGgBwDjoAYA41AHAIWUDAAGuwCA70gbAAq7AIDhJAIADrsAgONwGgASuwCAFrsAgBq7AIDvqAEA7+gGAIagDwCHDA0Ao4kCAB67AIClnQIAIrsAgCa7AICmiQIAKrsAgC67AICrSQIAqoECAK1ZAgCsUQIAr0kCAK5RAgCoZQ4AqXUOAKp9DgCrdQ4ArG0OAK21DgCuvQ4Ar7UOAP66AIAyuwCANrsAgDq7AIA+uwCASbsAgE27AIBRuwCAuF0PALltDwC6ZQ8Auw0PALwVDwC9HQ8AvhUPAL8JDwCwzQ4AsdUOALLdDgCz1Q4AtM0OALVxDwC2cQ8At20PALP1DgBVuwCAWbsAgF27AIBhuwCAtjUOALXlDgBluwCAuxEOALoJDgBpuwCAbbsAgL+1DwC+CQ4AvQEOALwJDgCCFQAAo7EOAIBhAACBYQAApnEOAHG7AIC+EAEApaEOAKpNDgCrVQ4AebsAgIQgAQCuTQ4Ar/EPAKxNDgCtRQ4An0UIAJ4NCQCdDQkAnJkLAJt1NQCaETUAmZk3AJgNMQCXJTEAliUxAJWBPQCUDT0Ak4k/AJIVOACRPTkAkD05AI9lJQDvrA0AhgAEAIegAQB9uwCAgbsAgIW7AIDv6AEAibsAgOE0AgCNuwCA4zQBAJG7AIDjCAwAlbsAgOEIDQChoQEAmbsAgKMJBQCibQMApc0EAKQRBQCnHRkAph0ZAKmhHQCoORkAq+kcAKqpHQCtkREArAEQAK8BFACuUREAsfkVALDlFQCz6WkAsgFoALUBbAC0eWkAnbsAgKG7AICluwCAqbsAgK27AICxuwCAowkDAKIZDQCh/Q0AoP0NAIIlJgCDBToAtbsAgLm7AICGqTwAhzU+AIQdOgCFPTsAiok+AIslMgC9uwCAwbsAgI6xNACPMTYAjD0yAI0tMgCSJTYAk9EIAIREAwC+wAQAlhULAJdVDgCUXQoAlVUKAJplDgCbiQ4AxbsAgMm7AIDNuwCA0bsAgJyBAADVuwCAuLUCALm9AgC6tQIAuwkCALwZAgC9GQIAvgkCAL8BAgCwdQ0AsX0NALJJDQCzSQ0AtJUCALWdAgC2lQIAt40CAKi9DQCpUQ0AqlUNAKtpDQCsfQ0ArWUNAK5tDQCvEQ0AdbsAgILtAQCBHQAAgB0AANm7AIDduwCAjroAgL5wBQCznQwAhIwFAOG7AIDpuwCA7bsAgLalDAC1tQwA8bsAgLv5DAC68QwAhigFAIcgBQC/GQMAvhEDAL3dDAC83QwA9bsAgKPZDAD5uwCA/bsAgKbhDAABvACABbwAgKXxDACqtQwAq70MAAm8AIANvACArlUDAK9dAwCsmQwArZkMABG8AIAVvACAGbwAgB28AIAhvACAJbwAgCm8AIDvvAEALbwAgOF8DgAxvACA41ABADW8AIA5vACAPbwAgEG8AICzlQIARbwAgEm8AIBNvACAUbwAgLa9AgC1uQIAWbwAgLs5AgC6YQIAhsgEAIesBAC/GQIAvhECAL0ZAgC8IQIAo1UFAILVBwCBxQcAgMUHAF28AICmfQUApXkFAGG8AICr+QUAqqEFAGW8AIBpvACAr9kFAK7RBQCt2QUArOEFAG28AICzWQcAcbwAgHW8AIC2HQcAebwAgH28AIC1FQcAugkHALsJBwCBvACAhbwAgL75BwC/+QcAvPkHAL35BwDluwCAVbwAgIm8AICNvACAkbwAgJW8AICZvACAnbwAgKitBwCptQcAqrUHAKvtBwCs+QcArfkHAK7tBwCv5QcAsKkHALGpBwCySQcAs0kHALRZBwC1WQcAtkkHALdJBwC4eQcAuUUHALpBBwC7XQcAvEUHAL1NBwC+RQcAvzkHAKMdBgChvACApbwAgKm8AICtvACAplkGAKVRBgCxvACAq00GAKpNBgC1vACAubwAgK+9BgCuvQYArb0GAKy9BgCAbQAAgQkAAIIZAAC9vACAwbwAgISYAQC+kAEAxbwAgIYAHACHxAEAybwAgM28AIDRvACA1bwAgNm8AIDdvACAqF0GAKmVAQCqlQEAq6UBAKy9AQCt1QEArtEBAK/RAQDhvACA5bwAgOm8AIDtvACA8bwAgPW8AID5vACA/bwAgLhZAQC5WQEAus0AALvFAAC83QAAvcUAAL7FAAC/9QAAsLUBALG9AQCygQEAs4EBALR5AQC1eQEAtmkBALdpAQCzHQIAAb0AgAW9AIC+gBwACb0AgLZVAgC1NQIADb0AgLt5AgC6cQIAEb0AgBW9AIC/vQIAvr0CAL1VAgC8VQIAGb0AgKNZAgAdvQCAIb0AgKYRAgAlvQCAKb0AgKVxAgCqNQIAqz0CAC29AIAxvQCArvkCAK/5AgCsEQIArRECADm9AIA9vQCAvgQdAL4AHgBBvQCARb0AgEm9AIBNvQCAgPkAAIHNAACCxQAAhCADAIawHACHlAMAUb0AgFW9AIBZvQCAXb0AgGG9AIBlvQCA42wCAGm9AIDhoAEAbb0AgO8UAgBxvQCAdb0AgHm9AIB9vQCAgb0AgIW9AICJvQCA4fAGAOE0BgDjTAAA4xgGAI29AICRvQCAlb0AgJm9AICAPQAAgQkAAIIZAACdvQCAob0AgIS8HQDvmAAA7zgHALMxAgDRAAAAh9gdAIZsHACpvQCAtikCALUhAgCtvQCAu80CALrNAgCxvQCAtb0AgL/NAgC+zQIAvc0CALzNAgCyXQYAs2UGALANBgCxVQYAtn0GALedBQC0fQYAtXUGALqNBQC7zQUAuKUFALmFBQC+xQUAv8kFALzVBQC9zQUAub0AgL29AIDBvQCAxb0AgMm9AIDNvQCA0b0AgNW9AICqtQYAq70GAKgBBwCpvQYAroEGAK+NBgCsmQYArZUGAKNxHQDZvQCA3b0AgOG9AIDlvQCApmkdAKVhHQDpvQCAq40dAKqNHQDtvQCA8b0AgK+NHQCujR0ArY0dAKyNHQD1vQCAs9UeAPm9AID9vQCAts0eAAG+AIAFvgCAtcUeALqhHgC7oR4ACb4AgA2+AIC+pR4Av6keALyxHgC9sR4ANb0AgKW9AIARvgCAhAQDAID5AACB+QAAghEAABW+AICoIR4AqSEeAKo5HgCrOR4ArCkeAK0pHgCuAR4ArwEeALABHgCxAR4AsgEeALMBHgC0BR4AtQkeALY9HgC3NR4AuA0eALkVHgC6HR4AuxUeALwNHgC95R8Avu0fAL/lHwCjkR8AGb4AgIYoAQCHSAEAHb4AgKaJHwClgR8AIb4AgKvlHwCq5R8AJb4AgCm+AICv7R8AruEfAK31HwCs9R8ALb4AgLMtHgAxvgCANb4AgLaVHgA5vgCAPb4AgLWdHgC6sR4Au7EeAEG+AIBFvgCAvnUBAL99AQC8oR4AvaEeAKjRHgCp2R4AquEeAKvhHgCsUR4ArVEeAK5RHgCvUR4ASb4AgE2+AIBRvgCAVb4AgFm+AIBdvgCAYb4AgGW+AIC43QEAue0BALrlAQC7jQEAvJkBAL2ZAQC+jQEAv4UBALAxHgCxMR4AsjEeALMxHgC09QEAtf0BALb1AQC37QEAo2kdAGm+AIBtvgCAcb4AgHW+AICm0R0ApdkdAHm+AICr9R0AqvUdAH2+AICBvgCArzkCAK4xAgCt5R0ArOUdAIFpAACAWQAAvgAEAIJhAACJvgCAjb4AgJG+AICVvgCAhOwDAJm+AICHiAMAhuwEAJ2+AIChvgCApb4AgKm+AICohQMAqZUDAKqVAwCrpQMArL0DAK3VAwCu0QMAr9EDAK2+AICxvgCAtb4AgLm+AIC9vgCAwb4AgMW+AIDJvgCAuHEDALlxAwC6cQMAu3EDALzVAAC93QAAvtUAAL/NAACwtQMAsb0DALKBAwCzgQMAtFEDALVRAwC2UQMAt1EDAOFUHgDhrB8A45QBAOMoHgDjYAMAzb4AgOEIAADRvgCA75ADANW+AIDZvgCA3b4AgOG+AIDlvgCA70wfAO9MHwCzXQIA6b4AgO2+AIDxvgCA+b4AgLYVAgC1dQIA/b4AgLs5AgC6MQIAhCQFAL7gBAC/1QIAvtUCAL0VAgC8FQIAuJEdALmZHQC6oR0Au6EdALzRHQC93R0AvtUdAL/JHQCwCR4AsQkeALIZHgCzGR4AtAkeALUJHgC2vR0At7UdAKipHgCpqR4AqrkeAKu5HgCsqR4ArakeAK55HgCveR4AgKUAAIGtAACCpQAAAb8AgIbQBACH+AQABb8AgAm/AICFvgCA9b4AgA2/AIARvwCAFb8AgBm/AIAdvwCAIb8AgKhxBgCpcQYAqnEGAKtxBgCsVQYArUUGAK5NBgCvRQYAsD0GALHlBgCy7QYAs+UGALT9BgC15QYAtu0GALflBgC43QYAuXEHALp1BwC7SQcAvFkHAL1ZBwC+SQcAv0kHALPZBgAlvwCAKb8AgC2/AIAxvwCAtuUGALX9BgA1vwCAuwEGALrZBgA5vwCAPb8AgL8BBgC+GQYAvREGALwZBgBBvwCAo9kFAEW/AIBJvwCAppEFAE2/AIBRvwCApfEFAKq1BQCrvQUAVb8AgFm/AICuUQUAr1EFAKyRBQCtkQUAo1kHAIIZAACBGQAAgOEBAF2/AICmZQcApX0HAGG/AICrgQcAqlkHAISgAgC+rAEAr4EHAK6ZBwCtkQcArJkHAGW/AICzqQYAhugAAIcsAQC2WQEAab8AgG2/AIC1oQYAunUBALt9AQBxvwCAdb8AgL75AQC/+QEAvGUBAL35AQCo0QYAqdkGAKplBgCrdQYArG0GAK2dAQCulQEAr40BAITsAQB5vwCAfb8AgIG/AICFvwCAib8AgI2/AICRvwCAuGkBALlpAQC6CQEAuwUBALwdAQC9AQEAvgEBAL81AQCw9QEAsf0BALL1AQCzaQEAtHkBALV5AQC2aQEAt2EBAJW/AICZvwCAnb8AgKPhBQChvwCApekFAKYRAgClvwCAqb8AgK2/AICqPQIAqzUCAKwtAgCtsQIArrECAK+xAgCxvwCAtb8AgL4EAwCEAAwAub8AgL2/AIDBvwCAxb8AgIANAACBFQAAgh0AAMm/AIDNvwCA0b8AgIdEAwCG3AwAs+kDANm/AIDdvwCA4b8AgOW/AIC2PQMAtT0DAOm/AIC7GQMAuhEDAO2/AIDxvwCAv7kAAL6xAAC9uQAAvAEDAPW/AIDhlAEA+b8AgON8AQD9vwCAAcAAgAXAAIAJwACADcAAgBHAAIAVwACAGcAAgB3AAIAhwACAJcAAgO9MAgCoVQIAqV0CAKphAgCrYQIArLUCAK29AgCutQIAr60CAL5oDQApwACALcAAgDHAAIA1wACAgq0AAIGtAACArQAAuGEBALlhAQC6CQEAuwkBALwBAQC9AQEAvgEBAL8BAQCw1QIAsd0CALLVAgCzbQEAtHUBALV9AQC2aQEAt2EBAOFoBgDh8AcA47AAAOP0BgA5wACAPcAAgEHAAIBJwACATcAAgFHAAIBVwACAWcAAgL78DABdwACA72wAAO8oBgCjqQIAYcAAgIZoDACHBA0AZcAAgKZ9AgClfQIAacAAgKtZAgCqUQIAbcAAgHHAAICv+QEArvEBAK35AQCsQQIAqIUOAKmNDgCqhQ4Aq50OAKyNDgCtvQ4ArrUOAK/dDgBFwACAdcAAgHnAAIB9wACAgcAAgIXAAICJwACAjcAAgLitDgC5tQ4Aur0OALu1DgC8dQ8AvX0PAL51DwC/bQ8AsKkOALG1DgCyvQ4As7UOALStDgC1lQ4Atp0OALeVDgCzDQ4AkcAAgJXAAICZwACAncAAgLY9DgC1BQ4AocAAgLtxDgC6bQ4ApcAAgKnAAIC/UQ4AvmkOAL1hDgC8aQ4AghkAAKNJDgCAZQAAgRkAAKZ5DgCtwACAscAAgKVBDgCqKQ4AqzUOAIS8AwC1wACAri0OAK8VDgCsLQ4ArSUOAKidDgCppQ4Aqq0OAKulDgCsvQ4AraEOAK7dDgCvzQ4AhiABAIdkAQC5wACAvcAAgMHAAIDFwACAycAAgM3AAIC4eQEAuXkBALrNAQC7xQEAvN0BAL3FAQC+xQEAv/UBALC9DgCxjQ4AsoUOALNJAQC0WQEAtVkBALZJAQC3SQEAtS0OANHAAIDVwACAtjkOANnAAIDdwACAsz0OAOHAAIC9hQEAvEkOAL+FAQC+hQEA5cAAgNW/AIC7UQ4AumEOAKNlDgDpwACA7cAAgPHAAID1wACApmEOAKV1DgD5wACAqwkOAKo5DgD9wACAAcEAgK/dAQCu3QEArd0BAKwRDgAFwQCACcEAgO/QDwANwQCAEcEAgBXBAIAZwQCAHcEAgCHBAIC+aAMAKcEAgC3BAIDhVA4AMcEAgONkDgA1wQCAgFkAAIFZAACCaQAAhIwDAIbwBACHFAMAOcEAgD3BAIBBwQCARcEAgEnBAIBNwQCAUcEAgFXBAIBZwQCAXcEAgGHBAIBlwQCAacEAgG3BAIBxwQCAdcEAgHnBAIB9wQCAqIkDAKmJAwCqmQMAq5kDAKyJAwCtiQMArj0DAK81AwCwUQMAsVEDALJVAwCzfQMAtBUDALUdAwC2FQMAtw0DALg9AwC5DQMAugUDALvtAAC89QAAvfkAAL7pAAC/6QAAgcEAgIXBAICJwQCAsz0CAI3BAIC1LQIAtiUCAJHBAIC+aAUAmcEAgLq5AgC7uQIAvK0CAL2FAgC+/QIAv/UCAIBJAACBVQAAglUAAIQABQDvjAMAvhgEAId0BQCG/AQA4zwDAJ3BAIDhUAAAocEAgKXBAICpwQCArcEAgLHBAIC1wQCAucEAgL3BAIDBwQCAxcEAgMnBAIDNwQCA79QOAL4oBgDhdA4A0cEAgONUAQDVwQCA2cEAgN3BAIDhwQCAo/ECAOXBAIDpwQCA7cEAgPHBAICm6QIApeECAPXBAICrdQIAqnUCAPnBAID9wQCArzkCAK4xAgCtSQIArGECAKgpBgCpKQYAqj0GAKsxBgCsSQYArUkGAK55BgCveQYAlcEAgIIVAACBxQcAgMUHAAHCAICEaAMABcIAgAnCAIC4yQYAuckGALrZBgC72QYAvMkGAL3JBgC+WQcAv1kHALAJBgCxCQYAshkGALMZBgC0CQYAtQkGALb5BgC3+QYAs7UGAA3CAICGrAAAh0ADABHCAIC2yQYAtcEGABXCAIC7zQYAus0GABnCAIAdwgCAv80GAL7NBgC9zQYAvM0GACHCAICj8QYAJcIAgCnCAICmjQYALcIAgDHCAIClhQYAqokGAKuJBgA1wgCAOcIAgK6JBgCviQYArIkGAK2JBgCoJQYAqWEGAKplBgCrfQYArGUGAK1tBgCuZQYAr50GAD3CAIBBwgCARcIAgEnCAIBNwgCAUcIAgFXCAIBZwgCAuPUGALn9BgC69QYAu4kGALyZBgC9mQYAvokGAL+BBgCw5QYAse0GALLlBgCz/QYAtOUGALXtBgC20QYAt80GAF3CAIC2/QYAtf0GAGHCAICz/QYAZcIAgGnCAIBtwgCAvzkGAL4xBgC9OQYAvCEGALs5BgC6MQYAJcEAgHHCAICjrQYAgnkAAIFVAACAVQAAhFwBAKatBgClrQYAecIAgKtpBgCqYQYAhkh/AIfkAACvaQYArmEGAK1pBgCscQYAfcIAgO/cBwCBwgCAhcIAgInCAICNwgCAkcIAgJXCAICZwgCAhKADAJ3CAIC/JHkAocIAgONoBwClwgCA4XQGALPRAgCpwgCAvgQDAISAfQCtwgCAtvkCALXxAgCxwgCAu7UCALqpAgC1wgCAucIAgL9RAwC+mQIAvZECALylAgCpBQIAqLkCAKsVAgCqHQIArT0CAKw9AgCvUQIArl0CAL5ofQC9wgCAwcIAgMXCAIDJwgCAzcIAgNHCAIDVwgCAufEDALjpAwC78QMAuvkDAL1RAwC86QMAv00DAL5RAwCxNQIAsCkCALMBAgCyNQIAtdEDALQZAgC30QMAttkDAIIpAACjlQMAgB0AAIEVAACmvQMA2cIAgN3CAICltQMAqu0DAKvxAwDhwgCA6cIAgK7dAwCvFQIArOEDAK3VAwCGYH0Ah3h9ALNBAQCEAH8AtUEBAO3CAIDxwgCAtkkBAPXCAID5wgCAu0EBALpNAQC9SQEAvEUBAL8pAQC+OQEA/cIAgO/cBgABwwCABcMAgAnDAIANwwCAEcMAgO8wBgCELH4A4eAGABXDAIDjiAEAGcMAgON0AAAdwwCA4SwBAKPJAQAhwwCAJcMAgIVweQApwwCApsEBAKXJAQAtwwCAq8kBAKrFAQAxwwCANcMAgK+hAQCusQEArcEBAKzNAQCo3X0AqQV+AKoBfgCrAX4ArAF+AK0BfgCuAX4ArwF+AOXCAIA5wwCAPcMAgEHDAIBFwwCAgp0AAIGdAACAnQAAuC1+ALnhfgC64X4Au+F+ALzhfgC94X4AvuF+AL/hfgCwQX4AsU1+ALJZfgCzVX4AtDV+ALUlfgC2JX4AtxV+AKitfwCp0X8AqtF/AKvtfwCs9X8ArRV/AK4RfwCvEX8AScMAgE3DAIBRwwCAVcMAgIbwAwCHuAAAWcMAgF3DAIC4EX8AuRl/ALohfwC7IX8AvPUAAL39AAC+9QAAv+0AALBxfwCxcX8AsnF/ALNFfwC0QX8AtU1/ALY9fwC3NX8As1l+AGHDAIBlwwCAacMAgG3DAIC2lX4AtX1+AHHDAIC7tX4AurV+AHXDAIB5wwCAv4l+AL6FfgC9kX4AvKV+AH3DAICjHX4AgcMAgIXDAICm0X4AicMAgI3DAIClOX4AqvF+AKvxfgCRwwCAlcMAgK7BfgCvzX4ArOF+AK3VfgCwrQAAscUAALLBAACzwQAAtMUAALXNAAC28QAAt/EAALhhAAC5YQAAumEAALt9AAC8ZQAAvW0AAL5lAAC/vQMAmcMAgJ3DAIChwwCAdcIAgKXDAICpwwCArcMAgLHDAICoWQEAqVkBAKrtAACr5QAArP0AAK3lAACu5QAAr9UAALXDAICCHQAAgR0AAIAdAAC5wwCAvcMAgMHDAIC+VAIAhoAEAIfsAgDJwwCAzcMAgNHDAIDVwwCA2cMAgL54AwDjdH4A3cMAgOG4fQDhwwCA5cMAgOnDAIDtwwCA8cMAgPXDAID5wwCA/cMAgAHEAIDvwH4ABcQAgAnEAIANxACAs4UDABHEAIAVxACAGcQAgB3EAIC2hQMAtZUDACHEAIC74QMAuokDAL4kBgAlxACAv+kDAL7hAwC99QMAvPUDAIIpAACjwQMAgB0AAIEVAACmwQMAKcQAgC3EAICl0QMAqs0DAKulAwAxxACAheAFAK6lAwCvrQMArLEDAK2xAwDh+AMAOcQAgONcHwA9xACA7/QDAEHEAICGPAcAh6wCAON8fgBFxACA4YABAEnEAIBNxACAUcQAgO/kEwBVxACAs3EBAFnEAIBdxACAYcQAgGXEAIC2EQEAtWEBAGnEAIC7OQEAujEBAG3EAIBxxACAvxkBAL4RAQC9GQEAvCEBAHXEAIB5xACAfcQAgIHEAICFxACAicQAgI3EAIDvxH8AkcQAgOH8fgCVxACA4/B/AIANAACBdQAAgn0AAJnEAICdxACAocQAgKP5AQC+AAgApekBAKnEAICtxACAppkBAISoBQCxxACAq7EBAKq5AQCtkQEArKkBAK+RAQCumQEAqCkGAKkpBgCqOQYAqzkGAKwpBgCtUQYArlUGAK9NBgA1xACAhCABALXEAIClxACAo+EBAKKZBAChGQQAoPEFALg5BgC5OQYAus0GALvFBgC83QYAvcUGAL7FBgC/8QYAsDUGALE9BgCyNQYAsw0GALQVBgC1HQYAthUGALcJBgCPoWwAs5EHAIYoAQCHfAMAtqEHALnEAIC9xACAtbEHALrlBwC77QcAwcQAgMXEAIC+7QcAv90HALz1BwC97QcAn/l4AJ7leACdcXkAnCF8AJvxfACaYX0AmZlxAJjZcACX4XAAlnl0AJVtdACUbXQAk61pAJJxaACReWgAkB1uAIIhbQCD5W8AycQAgM3EAICGTWgAh5V1AISZaQCFmWkAiqV1AIu5dQDRxACA1cQAgI5xcACPgXwAjDlxAI05cQCSYX0Ak6l9ANnEAIDdxACAlml5AJeZBACU4XgAlX15AJpBBQCbyQUA4cQAgOXEAIDpxACA7cQAgJypAADxxACAo4ENAKKpAQChqQEA9cQAgKexCQCmAQgApU0NAKSZDQCrkRUAqoUVAKkBFACocQkArx0QAK7pEQCtvREArAEQALMBGACy8RwAscEdALDJHQDFwwCA+cQAgLXhGAC0/RkA/cQAgAHFAIAFxQCACcUAgIAdAACBCQAAgv0DAA3FAICjFQUAEcUAgIaIDACHPAMAGcUAgKYlBQClNQUAHcUAgKtpBQCqYQUAIcUAgCXFAICvWQUArmkFAK1pBQCscQUAKcUAgC3FAICEBAwAMcUAgDXFAIDhbAYAOcUAgOPsewA9xQCAQcUAgEXFAIDvqAYAScUAgE3FAIBRxQCAVcUAgKmNBQCogQUAq60FAKqZBQCtoQUArLkFAK+lBQCuqQUAhGgNAFnFAIBdxQCAYcUAgGXFAIBpxQCAbcUAgL70DAC5SQUAuEEFALtZBQC6QQUAvUkFALxBBQC/cQUAvn0FALGpBQCwoQUAs7kFALKhBQC1mQUAtKkFALd5BQC2kQUAqNUEAKndBACq7QQAqyUDAKyFAwCtjQMArrEDAK+xAwBxxQCAdcUAgHnFAIB9xQCAgBkAAIEZAACCBQAAgcUAgLgxAgC5MQIAujUCALvBAgC8hQIAvbUCAL69AgC/tQIAsGkCALFpAgCyQQIAs0ECALQ5AgC1OQIAthECALcRAgCGoAwAh0wNAInFAICNxQCA76QGAJHFAICVxQCA78wHAOOUAQDhpAYA4TgBAONcBgCZxQCAncUAgKHFAIClxQCAqcUAgK3FAICzLQQAscUAgLVFAwC1xQCAucUAgLZFAwC9xQCAwcUAgLvlAgC65QIAvd0CALzdAgC/tQIAvrUCABXFAICFxQCAxcUAgMnFAIDNxQCA0cUAgNXFAIDZxQCAqDEOAKk5DgCqAQ4AqwEOAKxxDgCtcQ4ArnUOAK9tDgCwGQ4AsSUOALItDgCzJQ4AtCEOALUhDgC2IQ4AtyEOALjFDgC5zQ4AusUOALvdDgC8xQ4Avc0OAL5ZDwC/WQ8As6kOAN3FAIDhxQCA5cUAgOnFAIC20Q4AtdkOAO3FAIC7wQ4Auv0OAPHFAIC+LAAAv8UOAL7FDgC90Q4AvNkOAIJpAACj7Q4AgFkAAIFRAACmlQ4A9cUAgPnFAIClnQ4AqrkOAKuFDgCGyAAAh6wAAK6BDgCvgQ4ArJ0OAK2VDgD9xQCAs5EOAAHGAIAFxgCAtqUOAAnGAIANxgCAta0OALrhDgC74Q4AEcYAgBXGAIC+6Q4Av9UOALz1DgC96Q4Ao6UKABnGAIAdxgCAIcYAgCXGAICmzQ0Apc0NACnGAICrbQwAqm0MAC3GAIAxxgCArz0MAK49DACtVQwArFUMAKgJDgCpCQ4Aqh0OAKsVDgCsIQ4ArSEOAK4hDgCvIQ4ANcYAgDnGAIA9xgCAQcYAgEXGAIBJxgCATcYAgFHGAIC4zQEAudUBALrdAQC71QEAvM0BAL1RAQC+UQEAv1EBALAhDgCxIQ4AsiUOALM5DgC0KQ4AtRUOALYdDgC39QEAVcYAgFnGAIBdxgCAo5kNAGHGAIClpQ0Apq0NAL7cAgCE7AMAacYAgKrpDQCr6Q0ArP0NAK3hDQCu4Q0Ar90NAIBFAACBTQAAglkAAKNFAwBtxgCApUEDAKZBAwBxxgCAhsAEAIcAAwCqLQMAqyUDAKw9AwCtJQMAriUDAK8VAwCoWQIAqYUDAKqBAwCrgQMArIUDAK2NAwCusQMAr7EDAHXGAIB5xgCAfcYAgIHGAICFxgCAicYAgI3GAICRxgCAuGUDALltAwC6ZQMAu30DALxlAwC9bQMAvmUDAL/dAACwpQMAsa0DALKlAwCzvQMAtK0DALWdAwC2lQMAt10DALMJAgCVxgCAmcYAgJ3GAIChxgCAtg0CALUNAgClxgCAu2kCALphAgCpxgCArcYAgL9ZAgC+aQIAvWkCALxxAgCxxgCAtcYAgLnGAIC9xgCA4aABAMHGAIDjaAMAxcYAgIEVAACAFQAA74wDAIIVAADJxgCAzcYAgNHGAIC+cAUA4RgOAOGUDwDjOA8A49QPAISUAgDZxgCA3cYAgOHGAIDlxgCA6cYAgO3GAIDxxgCA9cYAgPnGAIDv7AEA7/gPAIZgBACHBAUAs5UBAITMBQC1dQEA/cYAgAHHAIC2dQEABccAgAnHAIC7UQEAulkBAL31AAC8SQEAv/UAAL71AACoJQYAqVUGAKpVBgCrrQYArLUGAK29BgCutQYAr60GANXGAIANxwCAEccAgBXHAIAZxwCAHccAgCHHAIAlxwCAuGkHALlpBwC6CQcAuwkHALwZBwC9GQcAvg0HAL8BBwCw1QYAsd0GALLVBgCzaQcAtHkHALV5BwC2aQcAt2EHAKPdBgApxwCALccAgDHHAIA1xwCApj0GAKU9BgA5xwCAqxkGAKoRBgA9xwCAQccAgK+9BwCuvQcArb0HAKwBBgCAXQAAgW0AAIJlAACzUQcAvtgDALVxBwC2cQcARccAgIbgAACHFAMAul0HALs5BwC8KQcAvRUHAL4dBwC/2QAAqJUGAKmdBgCqlQYAq60GAKy1BgCtvQYArrUGAK+tBgBJxwCATccAgFHHAIBVxwCAWccAgF3HAIBhxwCAZccAgLhxAQC5cQEAunEBALtxAQC81QEAvd0BAL7VAQC/zQEAsNUGALGxBgCysQYAs40GALSVBgC1UQEAtlEBALdRAQBpxwCAoxkGAG3HAIBxxwCApjkGAGXGAIB1xwCApTkGAKoVBgCrcQYAeccAgH3HAICuVQYAr5EBAKxhBgCtXQYAgccAgIXHAICJxwCAjccAgJHHAICVxwCAmccAgJ3HAIChxwCApccAgKnHAICtxwCAgBkAAIEZAACCBQAAsccAgISAAgC+gAMAhwwDAIasHADhaAYAuccAgOOYBwC9xwCAwccAgMXHAIDvrAcAyccAgM3HAIDRxwCA1ccAgNnHAIDdxwCA4ccAgOXHAICzZQMA6ccAgLVlAwC2bQMA7ccAgPHHAID1xwCAuukDALvlAwC8/QMAve0DAL7RAwC/0QMA+ccAgP3HAIAByACABcgAgAnIAIANyACAEcgAgBXIAICogQMAqYEDAKqBAwCrgQMArIEDAK2BAwCugQMAr4EDALBBAwCxTQMAskUDALNVAwC0eQMAtXkDALYZAwC3GQMAuCkDALkpAwC6OQMAuzkDALwpAwC9KQMAvhkDAL8ZAwCBGQAAgBEAAKMhAgCCLQAApSECABnIAIAdyACApikCACHIAIApyACAq6ECAKqtAgCtqQIArLkCAK+VAgCulQIAhEwCAL5IHQCHZB0AhuwcAONAAwAtyACA4aABADHIAIDvnAMANcgAgDnIAIA9yACAQcgAgEXIAIBJyACATcgAgFHIAIBVyACAWcgAgF3IAIBhyACAZcgAgGnIAIDvtAEAhKgdAOF8BgBtyACA43AGAHHIAIB1yACAecgAgH3IAICz4QEAgcgAgIXIAICJyACAjcgAgLblAQC19QEAkcgAgLuhAQC62QEAvuQcAJnIAIC/rQEAvqUBAL2xAQC8uQEAqBUeAKkZHgCqKR4AqykeAKw9HgCtJR4Ari0eAK8lHgAlyACAgvkfAIH5HwCA4R8AlcgAgJ3IAICGHAAAh7ADALjBHgC5wR4AusEeALvBHgC8wR4AvcEeAL7BHgC/wR4AsF0eALElHgCyLR4AsyUeALQhHgC1KR4AthkeALcZHgCjoR4AocgAgKXIAICpyACArcgAgKalHgCltR4AscgAgKvhHgCqmR4AtcgAgLnIAICv7R4AruUeAK3xHgCs+R4AvcgAgLOZHwDByACAxcgAgLa9HwDJyACAzcgAgLW1HwC6mR8Au5kfANHIAIDVyACAvnkfAL95HwC8eR8AvXkfAKglHgCpUR4AqlUeAKtpHgCseR4ArXkeAK5pHgCvaR4A2cgAgN3IAIDhyACA5cgAgOnIAIDtyACA8cgAgPXIAIC42R4Aue0eALr5HgC7+R4AvOkeAL3pHgC+nR4Av5UeALAZHgCxGR4AsukeALPpHgC0+R4AtfkeALbpHgC36R4Ao90eAIIpAACBFQAAgB0AAPnIAICm+R4ApfEeAP3IAICr3R4Aqt0eALXHAIAByQCArz0eAK49HgCtPR4ArD0eAITIAgCzQQEAvgwBAAnJAIC2QQEADckAgBHJAIC1UQEAuk0BALslAQCGSAAAh1ABAL4lAQC/LQEAvDEBAL0xAQAVyQCAGckAgIQEAwC+gAQAHckAgO+oHwAhyQCAJckAgL8oMQDjdB8AKckAgOE4HgAtyQCAMckAgDXJAIA5yQCAPckAgEHJAICjzQIARckAgKXdAgBJyQCATckAgKbNAgBRyQCAVckAgKupAgCqwQIArb0CAKy9AgCvoQIArqkCAKm1AgCoaR0AqwECAKoJAgCtAQIArBkCAK8xAgCuAQIAhGwFAFnJAIBdyQCAYckAgGXJAICCnQEAgZ0BAICdAQC55QMAuOUDALvlAwC65QMAveUDALzlAwC/5QMAvuUDALEhAgCwSQIAsyUCALIlAgC1KQIAtCECALcVAgC2FQIAqM0CAKnRAgCq0QIAqw0BAKwVAQCtBQEArgEBAK8BAQBpyQCAbckAgHHJAIB5yQCAvvgEAH3JAICByQCAhckAgLgVAQC5HQEAuikBALspAQC89QEAvf0BAL71AQC/7QEAsEkBALFVAQCyXQEAs1UBALRNAQC1NQEAtj0BALcxAQCGoAUAh8gFAInJAIDvvAAAjckAgJHJAICVyQCA74weAIQsBwDh8B4AmckAgOMcHgCdyQCA4ZQBAKHJAIDjbAAAsxkCAKXJAICpyQCArckAgIQACAC2xQEAtd0BALHJAIC70QEAus0BALXJAIC5yQCAv7EBAL7JAQC9wQEAvMkBAKPZBQB1yQCAvckAgMHJAIDFyQCApgUGAKUdBgDJyQCAqxEGAKoNBgDNyQCA0ckAgK9xBgCuCQYArQEGAKwJBgDVyQCAgh0AAIEdAACAHQAA2ckAgN3JAIDhyQCA5ckAgIZAAwCHxAMA6ckAgO3JAIDxyQCA9ckAgPnJAID9yQCAqK0HAKmxBwCqsQcAq7EHAKwZBwCtBQcArg0HAK8FBwABygCABcoAgAnKAIANygCAEcoAgBXKAIAZygCAHcoAgLgtBwC5zQAAusUAALvdAAC8zQAAvf0AAL71AAC/nQAAsEkHALFVBwCyUQcAsykHALQ5BwC1OQcAtiUHALcVBwCzOQYAIcoAgCXKAIApygCALcoAgLaFBgC1kQYAMcoAgLuRBgC6jQYANcoAgDnKAIC//QYAvv0GAL39BgC8hQYAPcoAgKN9BgBBygCARcoAgKbBBgBJygCATcoAgKXVBgCqyQYAq9UGAFHKAIC+bAEArrkGAK+5BgCswQYArbkGAKjpAQCp6QEAqvkBAKv5AQCs6QEArekBAK45AQCvOQEAgPUAAIH9AACCwQAAVcoAgIYQAACHdAEAWcoAgAXJAIC4zQAAudUAALrVAAC75QAAvP0AAL2VAAC+kQAAv5EAALBJAQCxSQEAslkBALNZAQC0SQEAtUkBALb9AAC39QAA7/QGAF3KAIBhygCAZcoAgO8wAgBpygCAbcoAgHHKAIDj4AcAdcoAgOGAAQB5ygCA4ygGAH3KAIDhyAUAgcoAgLMxAgCFygCAicoAgJYAAACNygCAtikCALUhAgCRygCAu80CALrNAgCVygCAmcoAgL/NAgC+zQIAvc0CALzNAgCdygCAocoAgKXKAICj/QIAqcoAgKXtAgCm5QIArcoAgLHKAIC1ygCAqgECAKsBAgCsAQIArQECAK4BAgCvAQIAgA0AAIEVAACCHQAAucoAgL3KAIDBygCAvlQMAMnKAICGwAwAhyQDAM3KAIDRygCA1coAgNnKAIDdygCA4coAgKi5AgCpAQEAqgEBAKsBAQCsBQEArQ0BAK4FAQCvOQEAhKgNAOXKAIDpygCA7coAgPHKAID1ygCA+coAgP3KAIC4LQEAucUBALrNAQC7xQEAvMEBAL3JAQC++QEAv/kBALBNAQCxUQEAslUBALMpAQC0OQEAtSUBALYlAQC3FQEA4RgGAAHLAIDjOAcABcsAgAnLAIC+WAwADcsAgBHLAICEbA8AFcsAgL5gDwAZywCAHcsAgCHLAIDvcAYAJcsAgIAVAACBGQAAgi0AAITMDwDjYAYAKcsAgOGgAQAtywCA73QAADHLAICGyAwAh/wMADnLAIA9ywCAQcsAgEXLAICjCQ4AxcoAgDXLAIBJywCATcsAgKYNDgClDQ4AUcsAgKsVDgCqCQ4AVcsAgFnLAICvYQ4Arn0OAK19DgCsAQ4AXcsAgLOpDgBhywCAZcsAgLapDgBpywCAbcsAgLWpDgC6SQ8Au0kPAHHLAIB1ywCAvkkPAL9JDwC8SQ8AvUkPAKhdDgCpbQ4AqmUOAKt9DgCsZQ4ArW0OAK5lDgCvuQ8AecsAgH3LAICBywCAhcsAgInLAICNywCAkcsAgJXLAIC4UQ8AuV0PALpVDwC7aQ8AvH0PAL1lDwC+bQ8Av2EPALDJDwCxyQ8AstkPALPZDwC0yQ8AtckPALZ9DwC3cQ8AmcsAgLURDwC2EQ8AncsAgIARAACBGQAAgikAALMVDwC8HQ8AvWEPAL5hDwC/fQ8AocsAgKXLAIC6FQ8AuwkPAKOtDwCpywCAhugAAIfIAQCtywCApq0PAKWtDwCxywCAq00OAKpNDgC1ywCAucsAgK9NDgCuTQ4ArU0OAKxNDgCocQ4AqXEOAKpxDgCrcQ4ArJ0BAK2FAQCuhQEAr7UBAL7sAAC9ywCAwcsAgMXLAIDJywCAzcsAgNHLAIDVywCAuGEBALlhAQC6YQEAu2EBALxhAQC9YQEAvmEBAL9hAQCwzQEAsaUBALKhAQCzoQEAtKUBALWtAQC2kQEAt5EBALP5DQDZywCA3csAgOHLAIDlywCAtgUCALUVAgDpywCAu2ECALoJAgDtywCA8csAgL9pAgC+YQIAvXUCALx1AgD1ywCAo70NAPnLAID9ywCApkECAAHMAIAFzACApVECAKpNAgCrJQIACcwAgA3MAICuJQIAry0CAKwxAgCtMQIAge0AAIDtAADv0AEAgh0AABHMAIAZzACAhjgEAIdQAwAdzACAIcwAgCXMAIApzACA4eABAC3MAIDjZA8AMcwAgDXMAIA5zACAPcwAgLORAwBBzACAtbkDALZ9AwBFzACAScwAgE3MAIC6WQMAu1kDALxJAwC9SQMAvv0AAL/1AACoRQIAqVUCAKpVAgCrZQIArH0CAK2xAgCusQIAr7ECAL5oBQBRzACAVcwAgFnMAIBdzACAYcwAgGXMAIBpzACAuF0BALltAQC6ZQEAuw0BALwZAQC9GQEAvg0BAL8FAQCw0QIAsdECALLRAgCz0QIAtHUBALV9AQC2dQEAt20BAOF4DwDjNA4A47gOAOF8DgBtzACAccwAgHXMAIB5zACAfcwAgIHMAICJzACAjcwAgJHMAIDv5A4A79QOAJXMAICjnQIAgmEAAIFpAACAUQAAhJwFAKZxAgCltQIAmcwAgKtVAgCqVQIAhkgEAIfMBACv+QEArvEBAK1FAgCsRQIAqJUGAKmlBgCqrQYAq6UGAKy9BgCtoQYArqUGAK/dBgCFzACAncwAgKHMAIClzACAqcwAgK3MAICxzACAtcwAgLhtBwC5dQcAun0HALt1BwC8bQcAvcUHAL7NBwC/xQcAsKUGALGtBgCyuQYAs7EGALSRBgC1kQYAtl0HALdVBwCzJQYAucwAgL3MAIDBzACAxcwAgLYhBgC1NQYAycwAgLtpBgC6YQYAzcwAgNHMAIC/VQYAvlUGAL1lBgC8bQYA1cwAgKNhBgDZzACA3cwAgKZlBgDhzACA5cwAgKVxBgCqJQYAqy0GAOnMAIDtzACArhEGAK8RBgCsKQYArSEGAKipBgCpqQYAqrkGAKuxBgCszQYArTEBAK4xAQCvMQEAgMkBAIHJAQCCBQAA8cwAgL54AgCEeAIA9cwAgPnMAIC43QEAue0BALrlAQC7jQEAvJkBAL2ZAQC+jQEAv4UBALBRAQCxUQEAslEBALNRAQC09QEAtf0BALb1AQC37QEAszEGAP3MAICGKAAAh9wBAAHNAIC2sQEAtUUGAAXNAIC7lQEAupUBAAnNAIANzQCAvzkBAL4xAQC9hQEAvIUBABXMAICjdQYAEc0AgBXNAICm9QEAGc0AgB3NAIClAQYAqtEBAKvRAQAhzQCAJc0AgK51AQCvfQEArMEBAK3BAQApzQCALc0AgDHNAIA1zQCAOc0AgD3NAIBBzQCARc0AgEnNAIBNzQCAUc0AgFXNAIBZzQCAXc0AgGHNAIC+cAMAhQA8AOHEBgCERAIA44wHAIBhAACBYQAAgmEAAO9oAwCFRDwA4RACAGnNAIDj2CsAhlA9AIf0AwBtzQCA76QHAHHNAIDvQAIAdc0AgHnNAIB9zQCAgc0AgIXNAICJzQCAhDw8AI3NAICRzQCAlc0AgJnNAIDj7AIAnc0AgOEsAQCzUQMAoc0AgKXNAICpzQCArc0AgLZ5AwC1cQMAsc0AgLs5AwC6MQMAtc0AgLnNAIC/9QAAvvUAAL0VAwC8FQMAqD0CAKmBAgCqmQIAq5ECAKy5AgCtuQIArtECAK/RAgCEqD8Avqg/AL3NAIDBzQCAxc0AgMnNAIDNzQCA0c0AgLhRAQC5UQEAulEBALtRAQC8cQEAvXEBAL5xAQC/cQEAsLUCALG9AgCygQIAs4ECALRxAQC1cQEAtnEBALdxAQCAtQAAgb0AAIK1AADZzQCAhrA/AIfgPADdzQCA71QAAL4sPgDhVAYA4c0AgOOIAADlzQCA6c0AgO3NAIDxzQCAo1ECAPXNAIC/2CYA+c0AgP3NAICmeQIApXECAAHOAICrOQIAqjECAAXOAIAJzgCAr/UBAK71AQCtFQIArBUCAJAtJACRBSgAkg0oAJPZKACUhS0AlTUsAJbFLACXtTEAmAEwAJkVMACalTUAmyk0AJxtNACdmTUAnj04AJ81OABlzQCAttU+ALXFPgDVzQCAs9E+AA3OAIARzgCAFc4AgL/ZPgC+1T4AvcU+ALzFPgC71T4Auuk+ABnOAICPXSQAqeUJAKgVCACrBQwAqg0MAK0BEACsAQwAr0EQAK69EACh4QAAHc4AgKMBBACi4QAApZ0EAKSVBACnuQgApgEIAKD1OQChBT0Aouk8AKP1PQAhzgCAJc4AgCnOAIAtzgCAscEUALABFACzARgAsn0UALXVGAC01RgAMc4AgDXOAICCISUAgyklADnOAIA9zgCAhsUpAIeBLACEGSkAhRkpAIoBLQCL+S0AQc4AgEnOAICOATEAj4k0AIyRMACNHTEAkkU1AJMZNQCG6AcAh+wBAJZZOQCXYTgAlPU0AJVZOQCaoTwAm0U9AE3OAIBRzgCAgX0AAIB9AACcQTwAglUAAKjpPwCp/T8Aqgk/AKsFPwCsHT8ArQU/AK4NPwCvBT8AVc4AgFnOAIBdzgCAYc4AgGXOAIBpzgCAbc4AgHHOAIC4DT8AuRU/ALoVPwC7JT8AvD0/AL39PgC+9T4Av+0+ALB9PwCxQT8AskE/ALNBPwC0QT8AtU0/ALY9PwC3NT8Ao4E8AHXOAIB5zgCAfc4AgIHOAICmhTwApZU8AIXOAICrhTwAqrk8AInOAICNzgCAr4k8AK6FPACtlTwArJU8AITIAwCz7T0Akc4AgJXOAIC26T0Amc4AgJ3OAIC16T0Auq09ALu1PQChzgCApc4AgL6dPQC/IQIAvKU9AL2VPQCoDT0AqR09AKohPQCrPT0ArCU9AK0tPQCuJT0Ar1k9AIANAACBFQAAgh0AAKnOAICtzgCAsc4AgLnOAIC+uAMAuLkCALlhAgC6GQIAuxkCALwJAgC9CQIAviECAL8hAgCwLT0AsTU9ALI1PQCzBT0AtB09ALWhAgC2oQIAt6ECAKOpPAC9zgCAhigFAIfsAgDBzgCApq08AKWtPADFzgCAq/E8AKrpPADJzgCAzc4AgK9lAwCu2TwArdE8AKzhPADRzgCAsykCANXOAIDZzgCAtvkCAN3OAIDhzgCAtfkCALrVAgC73QIA5c4AgOnOAIC+eQEAv3kBALzFAgC9eQEA7c4AgPHOAICj5QIA9c4AgKU1AgD5zgCA/c4AgKY1AgABzwCABc8AgKsRAgCqGQIArbUBAKwJAgCvtQEArrUBAOPwPgDhrD8A4UA+AON8PwAJzwCADc8AgBHPAIAVzwCAgA0AAIERAACCEQAAGc8AgO+oPgAdzwCAIc8AgO8gPgCoLQUAqW0FAKplBQCrrQUArLUFAK29BQCutQUAr60FALXOAICE6AMAvuADACXPAICGEAMAh5gDACnPAIAtzwCAuGkGALlpBgC6AQYAuwEGALwFBgC9DQYAvjEGAL8xBgCw1QUAsd0FALLVBQCzaQYAtHkGALV5BgC2aQYAt2EGAKg5BgCpgQcAqpkHAKuRBwCsuQcArbkHAK7ZBwCv1QcAMc8AgDXPAIBFzgCAOc8AgD3PAIBBzwCARc8AgEnPAIC4VQcAuV0HALppBwC7aQcAvAEHAL0BBwC+AQcAvwEHALCtBwCxsQcAsrEHALOFBwC0nQcAtXUHALZ9BwC3cQcAsxEGAE3PAIBRzwCAVc8AgFnPAIC2OQYAtTEGAF3PAIC7dQYAumkGAGHPAIBlzwCAv7EGAL5ZBgC9UQYAvGUGAGnPAICjVQYAbc8AgHHPAICmfQYAdc8AgHnPAICldQYAqi0GAKsxBgB9zwCAgc8AgK4dBgCv9QYArCEGAK0VBgCouQEAqbkBAKopAQCrKQEArD0BAK0lAQCuLQEAryUBAIXPAICCHQAAgR0AAIAdAACJzwCAjc8AgJHPAIC+cAEAuIEAALmNAAC6hQAAu5kAALyJAAC9vQAAvrUAAL99AACwXQEAseEAALLhAACz4QAAtOEAALXpAAC20QAAt9EAAITIAgCzpQIAhzgDAIYoAgC2oQIAmc8AgJ3PAIC1sQIAup0CALshAwC+bAMAoc8AgL4hAwC/KQMAvDEDAL0xAwCj4QIApc8AgKnPAICtzwCAsc8AgKblAgCl9QIAtc8AgKtlAwCq2QIAuc8AgL3PAICvbQMArmUDAK11AwCsdQMAqZkAAKiRAACrzQAAqqEAAK3dAACs3QAAr8UAAK7NAAC+LA0Awc8AgMXPAIDJzwCAzc8AgNHPAIDVzwCA2c8AgLnBAQC4eQAAu8EBALrJAQC9wQEAvNkBAL/FAQC+xQEAsY0AALCNAACzQQAAskkAALVBAAC0WQAAt0EAALZJAADdzwCA4c8AgOXPAIDpzwCA7c8AgO9QBwDxzwCA9c8AgL74DwDjdAcA+c8AgOF8BACAGQAAgQkAAIJ5AAD9zwCAAdAAgLNpAQAJ0ACAhMQCALYdAQAN0ACAEdAAgLUVAQC6CQEAuwkBAIboDQCH6A0Avt0BAL/FAQC83QEAvdUBABXQAIAZ0ACAHdAAgCHQAIDv1AAAJdAAgCnQAIDvTAEA47ADAOG0BgDhgAEA45gBAC3QAIAx0ACANdAAgDnQAIA90ACAQdAAgKPlAQCEwA0ApZkBAEXQAIBJ0ACAppEBAE3QAIBR0ACAq4UBAKqFAQCtWQEArFEBAK9JAQCuUQEABdAAgFXQAIBZ0ACAXdAAgGHQAIBl0ACAadAAgG3QAICoaQ8AqXEPAKpxDwCrrQ8ArLUPAK29DwCutQ8Ar6kPALDZDwCx9Q8Asv0PALP1DwC07Q8AtZUPALadDwC3iQ8AuLkPALmFDwC6jQ8Au2kAALx5AAC9eQAAvmkAAL9pAACBnQAAgJ0AAHHQAICCBQAAddAAgHnQAIB90ACAgdAAgIaAAwCH9AMAhdAAgInQAICN0ACAkdAAgJXQAICVzwCAs5kPAJnQAICd0ACAodAAgKXQAIC2XQ8AtV0PAKnQAIC7UQ8Aun0PAK3QAICx0ACAvzEPAL5JDwC9QQ8AvEkPAKNZDgC10ACAudAAgL3QAIDB0ACApp0OAKWdDgDF0ACAq5EOAKq9DgDJ0ACAzdAAgK/xDgCuiQ4ArYEOAKyJDgDR0ACA1dAAgNnQAIDd0ACAgBkAAIEZAACCBQAA4dAAgISgAQDl0ACAh+gBAIYABADp0ACA7dAAgPHQAID10ACAqBUBAKkdAQCqFQEAqyUBAKw9AQCtJQEAri0BAK8lAQD50ACA/dAAgAHRAIAF0QCACdEAgA3RAIAR0QCAFdEAgLjJAAC5yQAAutkAALvRAAC8+QAAvfkAAL6ZAAC/mQAAsCUBALEtAQCyJQEAsz0BALQtAQC1HQEAthUBALf5AAAZ0QCAHdEAgCHRAICzkQIAJdEAgLW5AgC2qQIAKdEAgC3RAIAx0QCAuu0CALvlAgC8/QIAveUCAL7lAgC/1QIApvECADXRAIA50QCApeECAD3RAICjyQIAQdEAgEXRAICuvQIAr40CAKylAgCtvQIAqrUCAKu9AgBJ0QCATdEAgID5AACB+QAAggUAAFHRAIC+yAMAhBgDAFnRAIBd0QCAYdEAgGXRAIBp0QCAbdEAgHHRAIB10QCAhhgEAIecAwB50QCAfdEAgIHRAICF0QCAidEAgI3RAIDvsAIAkdEAgOGUAQCV0QCA42wCAJnRAICd0QCAodEAgKXRAICp0QCA79APAK3RAICx0QCAtdEAgLnRAIDhrAEAvdEAgONsAACAMQAAgT0AAIIdAADv9A4A42wOAMHRAIDhLA8AvnAFALM5AgCEDAUAhugEAIdgBQDcAAAAtvECALX5AgDJ0QCAu9UCALrVAgDN0QCA0dEAgL91AQC+dQEAvcUCALzFAgDV0QCA4fQOANnRAIDjUA4A3dEAgOHRAIDl0QCA6dEAgO3RAIDx0QCA9dEAgPnRAID90QCAAdIAgAXSAIDv5A8ApmUCAAnSAIAN0gCApW0CABHSAICjrQIAFdIAgBnSAICu4QEAr+EBAKxRAgCtUQIAqkECAKtBAgAd0gCAIdIAgKiZBgCpmQYAqqkGAKupBgCsuQYArbkGAK6pBgCvqQYAJdIAgIIdAACBHQAAgB0AACnSAIAt0gCAMdIAgL50AwC4rQYAubUGALq9BgC7tQYAvK0GAL1RBwC+UQcAv1EHALChBgCxoQYAsqEGALOhBgC0oQYAtaEGALalBgC3mQYAVdEAgLMlBgCExAMAxdEAgLY9BgA10gCAOdIAgLU1BgC6YQYAu2EGAIYIAACHiAAAvmEGAL9hBgC8cQYAvXEGAKNhBgA90gCAQdIAgEXSAIBJ0gCApnkGAKVxBgBN0gCAqyUGAKolBgBR0gCAVdIAgK8lBgCuJQYArTUGAKw1BgCoXQYAqW0GAKplBgCrjQYArJkGAK2FBgCujQYAr4UGAFnSAIBd0gCAYdIAgGXSAIBp0gCAbdIAgHHSAIB10gCAuIUGALmNBgC6mQYAu5UGALyNBgC9rQYAvqUGAL99AQCw/QYAscUGALLNBgCzxQYAtN0GALXFBgC2zQYAt8UGALPtBgB50gCAfdIAgIHSAICF0gCAtgUGALURBgCJ0gCAuwEGALo5BgCN0gCAkdIAgL8BBgC+GQYAvREGALwZBgCV0gCAo6kGAJnSAICd0gCApkEGAKHSAICElAEApVUGAKp9BgCrRQYAvqABAKnSAICuXQYAr0UGAKxdBgCtVQYAqJkCAKnBAgCqwQIAq8ECAKzBAgCtyQIArvECAK/xAgCB7QMAgO0DAK3SAICC+QMAhpAcAId0AwCx0gCAtdIAgLjFAwC5zQMAusUDALvdAwC8zQMAvf0DAL71AwC/nQMAsEEDALFBAwCyQQMAs0EDALRBAwC1QQMAtkEDALdBAwCzSQIAudIAgL3SAIDB0gCAxdIAgLZJAgC1SQIAydIAgLuFAwC6hQMAzdIAgNHSAIC/hQMAvoUDAL2VAwC8lQMA1dIAgKMNAgDZ0gCA3dIAgKYNAgDh0gCA5dIAgKUNAgCqwQMAq8EDAOnSAIDt0gCArsEDAK/BAwCs0QMArdEDAOOYAQDhpAcA4VgGAONYBgDhoAEA8dIAgOPQAAD10gCA+dIAgP3SAIDvOAAAAdMAgO/0AQAF0wCACdMAgO/4BgCAeQAAgRUAAIIdAACEAB0ADdMAgBHTAIC+EB0AGdMAgIbAHACHrB0AHdMAgCHTAIAl0wCAKdMAgC3TAIAx0wCAu8UFALqhBQC5qQUAuJEFAL/NBQC+zQUAvckFALzVBQCzHQYAsh0GALEdBgCwHQYAt6EFALa9BQC1vQUAtL0FAKu9BgCqvQYAqb0GAKi9BgCvfQYArn0GAK19BgCsfQYANdMAgDnTAIA90wCAQdMAgEXTAIBJ0wCATdMAgFHTAICo7R0AqS0eAKoxHgCrMR4ArJUeAK2dHgCulR4Ar40eABXTAIBV0wCAWdMAgF3TAIBh0wCAZdMAgGnTAIBt0wCAuKkeALmpHgC6XR8Au1EfALxxHwC9cR8AvnUfAL9pHwCw/R4Asc0eALLFHgCzrR4AtLkeALW5HgC2rR4At6UeALO5HgBx0wCAddMAgHnTAICl0gCAth0eALUdHgB90wCAuwkeALo5HgCB0wCAhOADAL99HgC+fR4AvXkeALwRHgCCaQAAo/0eAIBFAACBUQAAplkeAL6cAwCF0wCApVkeAKp9HgCrTR4AhkgAAIdsAACuOR4ArzkeAKxVHgCtPR4AqF0eAKltHgCqZR4Aq30eAKxlHgCtbR4ArmUeAK/9HgCJ0wCAjdMAgJHTAICV0wCAmdMAgJ3TAICh0wCApdMAgLhpAQC5aQEAunkBALt5AQC8aQEAvWkBAL7dAQC/1QEAsIUeALGNHgCyhR4As50eALSFHgC1jR4AtoUeALdZAQCz7R4AqdMAgK3TAICx0wCAtdMAgLbtHgC17R4AudMAgLtJHgC6QR4AvdMAgMHTAIC/SR4AvkEeAL1JHgC8UR4AxdMAgKOpHgDJ0wCAzdMAgKapHgDR0wCA1dMAgKWpHgCqBR4Aqw0eANnTAIDd0wCArgUeAK8NHgCsFR4ArQ0eAKghAwCpIQMAqiEDAKshAwCsIQMArSEDAK4hAwCvIQMA4dMAgOXTAIDp0wCAvmACAO3TAIDx0wCA+dMAgP3TAIC4iQMAuYkDALqdAwC7lQMAvLkDAL25AwC+eQAAv3kAALDlAwCx7QMAsuUDALP9AwC07QMAtd0DALbVAwC3vQMAgKkAAIG1AACCvQAAs6UDAAHUAIC1pQMAtq0DAAXUAICE4AIACdQAgLotAwC7JQMAvD0DAL0lAwC+JQMAvxUDAKPpAwAN1ACAhmgEAIeAAwAR1ACApuEDAKXpAwAV1ACAq2kDAKphAwAZ1ACAHdQAgK9ZAwCuaQMArWkDAKxxAwAh1ACAJdQAgCnUAIAt1ACAMdQAgOE8HwA11ACA40AeADnUAIA91ACAQdQAgO+MHgBF1ACASdQAgE3UAIBR1ACAVdQAgIIlAACBEQAAgB0AAFnUAIDj5AMAXdQAgOGsAQBh1ACA77ADAIRkAgC+YAUAhtAEAIdEBQBp1ACAbdQAgHHUAIB11ACAedQAgH3UAICB1ACAhdQAgInUAIDvsAEAhKQFAOHcHgCN1ACA4xABAJHUAICV1ACAmdQAgJ3UAICzUQEAodQAgKXUAICp1ACArdQAgLYRAQC1fQEAsdQAgLsNAQC6DQEAtdQAgLnUAIC//QAAvv0AAL39AAC8/QAAqDkGAKk5BgCqmQYAq5EGAKy1BgCt0QYArskGAK/BBgBl1ACAvdQAgMHUAIDF1ACAgA0AAIGxAACCsQAAydQAgLhhBwC5YQcAumEHALt9BwC8ZQcAvW0HAL5lBwC/HQcAsIkGALGJBgCyaQcAs2kHALR5BwC1eQcAtmkHALdlBwCjEQYAzdQAgNHUAIC+gAMA1dQAgKZRBgClPQYA2dQAgKtNBgCqTQYAhggAAId8AwCvvQcArr0HAK29BwCsvQcA3dQAgOHUAICzSQcA5dQAgLVZBwDp1ACA7dQAgLZRBwDx1ACA9dMAgLtBBwC6dQcAvUUHALxFBwC/RQcAvkUHAKh5BgCpeQYAqokGAKuJBgCsmQYArZkGAK6JBgCviQYA9dQAgPnUAID91ACAAdUAgAXVAIAJ1QCADdUAgBHVAIC4jQYAuZUGALqVBgC7pQYAvL0GAL1xAQC+cQEAv3EBALD5BgCxzQYAstkGALPZBgC0yQYAtckGALa9BgC3tQYAowEGABXVAIAZ1QCAHdUAgCHVAICmGQYApREGACXVAICrCQYAqj0GACnVAIAt1QCArw0GAK4NBgCtDQYArA0GADHVAIA11QCAOdUAgD3VAICAGQAAgRkAAIIFAABB1QCAhKwBAL6sAQCH6AAAhkwPAEnVAIBN1QCAUdUAgFXVAIConQIAqcUCAKrNAgCrwQIArMUCAK3NAgCu+QIArz0DAFnVAIBd1QCAYdUAgGXVAIC+PAwAadUAgG3VAIBx1QCAuMkDALnJAwC62QMAu9EDALz5AwC9+QMAvpkDAL+ZAwCwRQMAsU0DALJFAwCzXQMAtEUDALVNAwC2RQMAt/kDALNFAgB11QCAedUAgH3VAICB1QCAtk0CALVNAgCF1QCAu4kDALqBAwCJ1QCAjdUAgL+JAwC+gQMAvYkDALyRAwCR1QCAowECAJXVAICZ1QCApgkCAJ3VAICh1QCApQkCAKrFAwCrzQMApdUAgKnVAICuxQMAr80DAKzVAwCtzQMAgO0BAIEVAACCEQAAhAACAK3VAIDhpAEAsdUAgOPsAAC51QCAvdUAgMHVAIDvMAAAxdUAgMnVAIDN1QCA0dUAgIbgDACH9AIA1dUAgNnVAIDd1QCA4dUAgO/MBgDl1QCA4bAHAOnVAIDjEAYA7dUAgPHVAID11QCA+dUAgP3VAIAB1gCABdYAgAnWAIAN1gCAEdYAgBXWAIAZ1gCA7+gBAIUYDwDhzAYAHdYAgOMcBgCAKQAAgR0AAIIFAAAh1gCAszkCAITMDQCGaA8Ah/wMAOHQ0gO28QEAtfkBACnWAIC72QEAutEBAL7kDAAt1gCAv30BAL59AQC9fQEAvMEBAKjxDQCp8Q0AqvENAKvxDQCsMQ4ArTEOAK4xDgCvMQ4AtdUAgCXWAIAx1gCANdYAgDnWAIA91gCAQdYAgEXWAIC46Q4AuekOALqJDgC7hQ4AvJ0OAL2BDgC+gQ4Av7UOALBVDgCxXQ4AslUOALPpDgC0+Q4AtfkOALbpDgC34Q4Ao3kNAEnWAIBN1gCAUdYAgFXWAICmsQ4ApbkOAFnWAICrmQ4AqpEOAF3WAIBh1gCArz0OAK49DgCtPQ4ArIEOAGXWAICz7Q8AadYAgG3WAIC26Q8AcdYAgHXWAIC16Q8Auq0PALu1DwBF1QCAedYAgL6VDwC/mQ8AvK0PAL2hDwCoIQ4AqSEOAKohDgCrPQ4ArCUOAK0tDgCuJQ4Ar1UOAH3WAICB1gCAhdYAgInWAICAHQAAgQkAAIK9AACN1gCAuDkOALk5DgC6yQ4Au8kOALzZDgC92Q4AvskOAL/JDgCwLQ4AsTUOALI9DgCzMQ4AtBUOALUZDgC2CQ4AtwkOAKOpDgCR1gCAhIACAL6AAQCFAAQApq0OAKWtDgCZ1gCAq/EOAKrpDgCGKAcAhxgAAK/dDgCu0Q4AreUOAKzpDgCd1gCAs+0BAKHWAICl1gCAtuUBAKnWAICt1gCAte0BALplAQC7bQEAsdYAgLXWAIC+bQEAv10BALx1AQC9bQEAqN0NAKnpDQCqIQIAqyECAKwhAgCtIQIAriECAK8hAgC51gCAvdYAgMHWAIDF1gCAohECAKMRAgCgqQ4AodUCALiJAgC5iQIAup0CALuVAgC8vQIAvXUDAL59AwC/dQMAsOUCALHtAgCy5QIAs/0CALTtAgC13QIAttUCALe9AgCjqQIAj8UaAMnWAIDN1gCA0dYAgKahAgClqQIA1dYAgKspAgCqIQIA2dYAgN3WAICvGQIArikCAK0pAgCsMQIAniUOAJ/lDgCc6QoAnRUKAJpFFgCbRQoAmFkWAJlRFgCWcRIAl4ETAJRVEgCV7RIAktEeAJPZHgCQtRoAkVUeAISpHwCFJR8AhiUfAIexEwDh1gCA5dYAgIJZGwCDURsAjEUSAI2lFwCOpRcAj7kXAIA5+wHp1gCAijkTAIutEwCUmQsAlaEPAJZpDwCX3Q8A7dYAgO+cDwCSyQsAk30LAJxFAwDjeA4A8dYAgOGYDAD11gCAhHgCAJqRAwCbXQMA4QQAAL6IBQDj3OoD+dYAgP3WAIAB1wCA7+wAAO+MDgDhcA4A4fwOAOMwAADjeA4AgSEAAIA5AADvtO0DgikAALMJAgAJ1wCAhmgEAIcsBQAN1wCAtg0CALUNAgAR1wCAu8UBALrFAQAV1wCAGdcAgL99AQC+fQEAvdUBALzVAQCV1gCABdcAgB3XAIAh1wCAJdcAgCnXAIAt1wCAMdcAgKi9BQCp5QUAquEFAKvhBQCs5QUAre0FAK7RBQCv0QUAsGEGALFhBgCyYQYAs2EGALTZBgC12QYAtskGALfBBgC4yQYAuckGALp5BwC7eQcAvEUHAL0lBwC+EQcAvw0HAKNJBQA11wCAOdcAgD3XAIBB1wCApk0FAKVNBQBF1wCAq4UGAKqFBgBJ1wCATdcAgK89BgCuPQYArZUGAKyVBgBR1wCAVdcAgFnXAIBd1wCAYdcAgGXXAIBp1wCAbdcAgIA5AACBOQAAggUAAHHXAIC+uAMAhLgDAHnXAIB91wCAqMUGAKnVBgCq1QYAq+UGAKz9BgCtHQEArhUBAK8NAQB11wCAgdcAgIaIAQCHHAEAhdcAgInXAICN1wCAkdcAgLjpAQC56QEAuokBALuJAQC8mQEAvZkBAL6JAQC/iQEAsHUBALF9AQCydQEAs+kBALT5AQC1+QEAtukBALfhAQCzXQYAldcAgJnXAICd1wCAhLwBALadAQC1dQYAodcAgLu5AQC6sQEApdcAgKnXAIC/PQEAvj0BAL09AQC8oQEArdcAgKMZBgCx1wCAtdcAgKbZAQC51wCAvdcAgKUxBgCq9QEAq/0BAMHXAIDF1wCArnkBAK95AQCs5QEArXkBAKj5AgCp+QIAqi0DAKs9AwCsJQMArS0DAK4lAwCvmQMAydcAgM3XAIDR1wCA1dcAgIANAACBsQAAgrEAANnXAIC4lQMAuZ0DALqhAwC7oQMAvHEAAL1xAAC+cQAAv3EAALDpAwCx6QMAsvUDALPFAwC03QMAtbUDALaxAwC3sQMAvswDAN3XAIDh1wCA6dcAgO3XAIDx1wCA9dcAgO/kAgD51wCA4ZQBAP3XAIDjLAEAAdgAgAXYAICHGAMAhhz8A7tNAwC6TQMACdgAgA3YAIC/EQMAvnkDAL1xAwC8QQMAs8UDAITo/AMR2ACAFdgAgBnYAIC2zQMAtc0DAB3YAICkAfwDpSX/A6bZ/wOnAfgDIdgAgKEVAwCiHQMAoz0CAKwR9wOtAfADri3zA68B8wOoEfsDqZn7A6oB9AOrHfcDtAHoA7Vl6wO+xPwDhMT8A7AB7AOxVe8Dsk3vA7Nx7gMl2ACAKdgAgC3YAIAx2ACANdgAgDnYAIA92ACAQdgAgOFQBgDhNAQA42wBAOPoBgBF2ACASdgAgE3YAIBR2ACAgDUAAIE9AACCNQAAWdgAgF3YAIBh2ACA77ABAO/ABgCj5QIAZdgAgIbo/AOHfP0DadgAgKbtAgCl7QIAbdgAgKttAgCqbQIAcdgAgHXYAICvMQIArlkCAK1RAgCsYQIAqI3+A6mV/gOqnf4Dq5X+A6yx/gOtvf4Drqn+A6+p/gNV2ACAedgAgH3YAICB2ACAhdgAgInYAICN2ACAkdgAgLgl/wO5Lf8DuiX/A7s9/wO8Jf8DvS3/A74l/wO/zf8DsKn+A7Gp/gOygf4Ds4H+A7SB/gO1if4Dtmn/A7cd/wOV2ACA4SD8A5nYAIDjePwDndgAgKHYAICl2ACAqdgAgK3YAICx2ACAtdgAgLnYAICAHQAAgXEAAIJxAADvDP0Ds1X+A73YAIDB2ACAvkAAAMXYAIC2ff4DtXn+A8nYAIC7Lf4Dui3+A4boAACHrAAAvw3+A74F/gO9Ff4DvBX+A6OV/wPN2ACA0dgAgNXYAIDZ2ACApr3/A6W5/wPd2ACAq+3/A6rt/wPh2ACA5dgAgK/N/wOuxf8DrdX/A6zV/wPp2ACAs/H+A+3YAIDx2ACAto3+A/XYAID52ACAtY3+A7pFAQC7TQEA/dgAgAHZAIC+RQEAv00BALxVAQC9TQEAqC3+A6k1/gOqPf4Dq0n+A6xB/gOtSf4DrnH+A69x/gMF2QCACdkAgA3ZAIAR2QCAFdkAgBnZAIAd2QCAIdkAgLhJAQC5VQEAul0BALtVAQC8TQEAvXUBAL59AQC/dQEAsMUBALHNAQCyxQEAs90BALTFAQC1zQEAtsUBALd9AQCjtf0DJdkAgCnZAICExAMALdkAgKbJ/QOlyf0DMdkAgKsJAgCqAQIAOdkAgL7sAgCvCQIArgECAK0JAgCsEQIAgEkAAIFVAACCVQAAo0UDAD3ZAIClRQMApkUDAEHZAICGwAQAhxQDAKopAwCrJQMArD0DAK0hAwCuIQMArxUDAEXZAIBJ2QCATdkAgFHZAIBV2QCAWdkAgF3ZAIBh2QCAqH0CAKmhAwCqoQMAq6EDAKyhAwCtqQMArpEDAK+RAwCwgQMAsY0DALKFAwCzmQMAtIkDALW9AwC2tQMAt30DALhFAwC5TQMAukUDALtdAwC8RQMAvU0DAL5FAwC/+QAA5dcAgLMNAgBl2QCAadkAgLYNAgBt2QCAcdkAgLUNAgC6YQIAu20CAHXZAIB52QCAvmkCAL9dAgC8dQIAvWkCAH3ZAICB2QCAhdkAgInZAICN2QCA4aQBAJHZAIDjQAMAldkAgJnZAICd2QCA77gDAIAVAACBHQAAggUAAKHZAICEgAIAvsgFAIcYBQCGLAQAqdkAgK3ZAICx2QCA76gBALXZAIDhdP4DudkAgOPw/gO92QCAwdkAgMXZAIDJ2QCAzdkAgNHZAIDV2QCAs5EBANnZAIC1UQEAtlEBAN3ZAIDh2QCA5dkAgLp9AQC7dQEAvG0BAL39AAC+9QAAv+kAAKgpBgCpVQYAqlUGAKuNBgCslQYArZ0GAK6VBgCvjQYApdkAgOnZAIDt2QCA8dkAgPXZAID52QCA/dkAgAHaAIC4bQcAuQUHALoNBwC7BQcAvB0HAL0FBwC+AQcAvz0HALD1BgCx/QYAsvUGALNlBwC0fQcAtWEHALZhBwC3VQcA4xAFAAXaAIDh8AQACdoAgIAdAACBCQAAgjkAAA3aAIAR2gCAhOgDAL7gAwAV2gCA78wFABnaAICHOAAAhhgAAKOdBgAd2gCAIdoAgCXaAIAp2gCApl0GAKVdBgAt2gCAq3kGAKpxBgAx2gCANdoAgK/lBwCu+QcArfEHAKxhBgCokQYAqZEGAKqRBgCrrQYArLkGAK2lBgCurQYAr6UGADnaAIA92gCAQdoAgEXaAIBJ2gCATdoAgFHaAIBV2gCAuGUBALltAQC6ZQEAu30BALxlAQC9bQEAvmUBAL/ZAQCw3QYAsaUGALKtBgCzpQYAtKEGALWpBgC2mQYAt5kGALMZBgBZ2gCAXdoAgGHaAIBl2gCAtiUGALUxBgBp2gCAu2EGALoZBgBt2gCAcdoAgL9tBgC+ZQYAvXEGALx5BgB12gCAo10GAHnaAIB92gCApmEGAIHaAICEmAEApXUGAKpdBgCrJQYAvqQBAInaAICuIQYArykGAKw9BgCtNQYAqcUCAKixAgCrxQIAqsUCAK3NAgCsxQIAr/UCAK71AgCN2gCAkdoAgJXaAICZ2gCAndoAgKHaAICl2gCAqdoAgLnJAwC4wQMAu9kDALrBAwC9+QMAvMkDAL+ZAwC+8QMAsUUDALBFAwCzRQMAskUDALVFAwC0RQMAt0UDALZFAwCASQMAgUkDAIJdAwCzRQIAvtwMALVFAgC2RQIArdoAgIYADACH5AMAuokDALuJAwC8mQMAvZkDAL6JAwC/iQMAowkCALHaAIC12gCAudoAgL3aAICmCQIApQkCAMHaAICrxQMAqsUDAMXaAIDJ2gCAr8UDAK7FAwCt1QMArNUDAM3aAIDR2gCA1doAgDXZAIDvAAAA2doAgN3aAIDh2gCA4+gAAOXaAIDhjAEA6doAgO3aAIDx2gCA+doAgP3aAICAbQAAgXUAAIJ9AACEQAIAhvAMAId4DQAB2wCABdsAgAnbAIAN2wCAEdsAgBXbAIAZ2wCAHdsAgCHbAIAl2wCAKdsAgC3bAIAx2wCANdsAgDnbAIA92wCAQdsAgO/MAQCE7AwA4TAGAEXbAIDjGAEASdsAgE3bAIBR2wCAVdsAgLPlAQBZ2wCAhIQPAF3bAIBh2wCAtuUBALX1AQBp2wCAu30BALrZAQC+oAwAbdsAgL8hAQC+OQEAvTEBALw5AQCo7Q0AqSUOAKotDgCrJQ4ArD0OAK0lDgCuLQ4AryUOAPXaAICC9Q8AgeUPAIDpDwBl2wCAcdsAgIaYAACHDAMAuK0OALlFDwC6TQ8Au0UPALxFDwC9TQ8AvkUPAL95DwCwXQ4AsfkOALKtDgCzpQ4AtL0OALWlDgC2pQ4At5UOAHXbAIDv7AwAedsAgH3bAICB2wCAhdsAgInbAICN2wCAvugAAJHbAICV2wCAmdsAgJ3bAIDj6A0AodsAgOEEDACj5Q4ApdsAgKnbAICt2wCAsdsAgKblDgCl9Q4AtdsAgKt9DgCq2Q4AudsAgL3bAICvIQ4ArjkOAK0xDgCsOQ4AqDkOAKk5DgCqUQ4Aq1EOAKxxDgCtcQ4ArnEOAK9xDgDB2wCAxdsAgMnbAIDN2wCAgBkAAIEZAACCBQAA0dsAgLjRDgC50Q4AutEOALvlDgC84Q4AveEOAL7hDgC/4Q4AsBEOALERDgCyEQ4AsxEOALTxDgC18Q4AtvEOALfxDgCz2Q4A2dsAgIYoAACHuAAA3dsAgLbxDgC1+Q4A4dsAgLvVDgC61Q4A5dsAgOnbAIC/NQ4AvjUOAL3FDgC8xQ4A7dsAgKOdDgDx2wCA9dsAgKa1DgD52wCA/dsAgKW9DgCqkQ4Aq5EOAAHcAIAF3ACArnEOAK9xDgCsgQ4ArYEOAKjdDQCp6Q0Aqj0CAKuNAgCsmQIArZkCAK6JAgCviQIAvqwEAAncAIAN3ACAhCADABHcAIAV3ACAGdwAgB3cAIC4iQIAuYkCALqZAgC7kQIAvLkCAL25AgC+eQMAv3kDALD5AgCx+QIAss0CALPFAgC03QIAtcUCALbBAgC3uQIAs7UCACHcAIAl3ACAKdwAgC3cAIC2GQIAtRECADHcAIC7PQIAuj0CADXcAIA53ACAvwECAL4ZAgC9EQIAvBkCAD3cAICj8QIAQdwAgEncAICmXQIATdwAgFHcAIClVQIAqnkCAKt5AgCGSAUAh6wEAK5dAgCvRQIArF0CAK1VAgCohQIAqZUCAKqVAgCrpQIArL0CAK3VAgCu0QIAr9ECAFXcAIBZ3ACAXdwAgGHcAICB8QEAgJkBAIXaAICC9QEAuHkBALl5AQC6zQEAu8UBALzdAQC9xQEAvsUBAL/1AQCwtQIAsb0CALKBAgCzgQIAtFUBALVdAQC2SQEAt0kBAGXcAIBp3ACAbdwAgO/UAQCEEAUAcdwAgHXcAIDvjA4AvuwFAOHsDgB53ACA4xwOAH3cAIDhlAEAgdwAgONkDgCzXQIAhdwAgIncAICN3ACAkdwAgLYVAgC1dQIAldwAgLs5AgC6MQIAmdwAgJ3cAIC/2QEAvtEBAL0VAgC8FQIAo50FAEXcAICh3ACApdwAgKncAICm1QUApbUFAK3cAICr+QUAqvEFALHcAIC13ACArxkGAK4RBgCt1QUArNUFAIBRAACBWQAAgmEAALOVBgC53ACAtXEHALZxBwC93ACAhkADAIdUAwC67QcAu+UHALzlBwC97QcAvtEHAL/NBwDB3ACAxdwAgMncAIDN3ACA0dwAgNXcAIDvQAQA2dwAgOEwBwDd3ACA45QEAOHcAIDl3ACA6dwAgO3cAIDx3ACAoxkGAPXcAID53ACA/dwAgAHdAICm/QcApf0HAAXdAICraQcAqmEHAAndAIAN3QCAr0EHAK5dBwCtYQcArGkHAKjNBwCp0QcAqtEHAKstBgCsNQYArT0GAK41BgCvnQYAEd0AgBXdAIAZ3QCAHd0AgIAZAACBGQAAggUAACHdAIC4iQYAuYkGALqZBgC7kQYAvLkGAL25BgC+UQEAv1EBALDlBgCx7QYAsv0GALP1BgC02QYAtcUGALbBBgC3uQYAqNEBAKnZAQCqCQEAqwkBAKwZAQCtGQEArgkBAK8JAQCEYAEAvnwBAIeoAACGjAEAKd0AgC3dAIAx3QCANd0AgLgJAQC5CQEAuhkBALsRAQC8OQEAvTkBAL75AAC/+QAAsH0BALFBAQCyRQEAs10BALRFAQC1TQEAtkUBALc5AQA53QCAPd0AgEHdAICzjQIARd0AgLWdAgC2lQIASd0AgE3dAIBR3QCAurUCALuJAgC8nQIAvYUCAL6NAgC/hQIAps0CAFXdAIBZ3QCApcUCAF3dAICj1QIAYd0AgGXdAICu1QIAr90CAKzFAgCt3QIAqu0CAKvRAgCE9AMAad0AgKgxAwCpMQMAqjEDAKsxAwCskQAArZEAAK6RAACvjQAAbd0AgHHdAIB13QCAed0AgH3dAICB3QCAhd0AgIndAIC4vQAAuWUAALptAAC7ZQAAvH0AAL1lAAC+bQAAv2UAALD9AACxxQAAss0AALOpAAC0uQAAtaUAALahAAC3oQAAgL0BAIEJAACCGQAAjd0AgJHdAIC+WAIAhxQdAIacHQCEbB0A1dsAgJndAICd3QCAvrwcAKHdAICl3QCAqd0AgLP5AgCt3QCAsd0AgLXdAIC53QCAtlEBALVZAQC+3B8Au0EBALp5AQC93QCAwd0AgL8hAQC+PQEAvT0BALxZAQDhcAcAxd0AgOMIBgDJ3QCA78wAAM3dAIDR3QCA1d0AgOMQAADZ3QCA4dABAN3dAICGkBwAh/QcAO/gBgDh3QCAo3kCAOXdAIDp3QCA7d0AgPHdAICm0QEApdkBAPXdAICrwQEAqvkBAPndAID93QCAr6EBAK69AQCtvQEArNkBAJXdAICCFQAAgeUfAIDlHwAB3gCABd4AgAneAIAN3gCAqAkfAKkJHwCqHR8AqxUfAKwNHwCtcR8ArnEfAK9xHwCwER8AsS0fALIlHwCzyR8AtN0fALXBHwC2wR8At8EfALjFHwC5yR8AutUfALupHwC8uR8AvbkfAL6pHwC/oR8As7UfABHeAIAV3gCAGd4AgB3eAIC20R8AtaUfACHeAIC7yR8AuvUfACXeAIAp3gCAvyUfAL45HwC9PR8AvNEfAC3eAIAx3gCANd4AgDneAIA93gCA4WAfAEHeAIDjtBwARd4AgEneAIBN3gCA7wAdAFHeAIBV3gCAWd4AgF3eAICjNR4AYd4AgGXeAIBp3gCAbd4AgKZRHgClJR4Acd4AgKtJHgCqdR4AhKgCAHXeAICvpR4ArrkeAK29HgCsUR4AgE0AAIFVAACCVQAAs8kBAHneAIC12QEAtskBAH3eAICGoAAAhwQBALrFAQC7rQEAvLUBAL29AQC+tQEAv60BAKiZAQCpmQEAqg0BAKsFAQCsHQEArQUBAK4FAQCvNQEAgd4AgIXeAICJ3gCAjd4AgJHeAICV3gCAmd4AgJ3eAIC4JQEAuS0BALo5AQC7OQEAvCkBAL0pAQC+3QAAv9UAALBNAQCxJQEAsi0BALMlAQC0PQEAtSUBALYhAQC3HQEAod4AgKXeAICp3gCAo4kCAK3eAIClmQIApokCALHeAIC13gCAud4AgKqFAgCr7QIArPUCAK39AgCu9QIAr+0CAL3eAIDB3gCAxd4AgIRAAgDJ3gCAzd4AgNHeAIDV3gCAgA0AAIEVAACCHQAA2d4AgN3eAIDh3gCAh7QDAIbcBAC+zAMA6d4AgO3eAIDx3gCA7+gCAPXeAID53gCA/d4AgOP8AgAB3wCA4dABAAXfAIAJ3wCADd8AgBHfAIAV3wCAs2EDABnfAIAd3wCAId8AgCXfAIC2eQMAtXEDACnfAIC7XQMAul0DAC3fAIAx3wCAv+EAAL79AAC9/QAAvP0AALC5AgCxuQIAsgkBALMJAQC0GQEAtQUBALYFAQC3PQEAuAUBALllAQC6bQEAu2UBALxhAQC9YQEAvmEBAL9hAQCFXAcANd8AgDnfAIA93wCAJd0AgEHfAIBF3wCASd8AgKgxAgCpOQIAqskCAKvJAgCs2QIArdkCAK7JAgCvyQIAhMwFAOGAHgBN3wCA47weAOE4HgBR3wCA46AAAL4QBABZ3wCAXd8AgO8MHgBh3wCAZd8AgGnfAIBt3wCA73QeAKNhAgCCUQAAgUEAAICRAABx3wCApnkCAKVxAgB13wCAq10CAKpdAgCGyAQAhzwFAK/hAQCu/QEArf0BAKz9AQCohQYAqY0GAKqFBgCrmQYArIkGAK2JBgCuvQYAr7EGAFXfAIB53wCAfd8AgIHfAICF3wCAid8AgI3fAICR3wCAuJ0GALmtBgC6pQYAuwkHALwZBwC9GQcAvg0HAL8FBwCw0QYAsdEGALLRBgCz0QYAtLUGALW9BgC2tQYAt60GALMNBgCV3wCAmd8AgJ3fAICh3wCAtgkGALUBBgCl3wCAuxUGALoVBgCp3wCArd8AgL95BgC+cQYAvQUGALwFBgCx3wCA4aAEALXfAIDjXAUAgA0AAIE1AACCPQAAud8AgL3fAIDB3wCAhGADAL5sAAC/8AEAhZAAAMXfAIDvmAUAo40HAIQIAACGAAwAh4wAAMnfAICmiQcApYEHAM3fAICrlQcAqpUHANHfAIDV3wCAr/kHAK7xBwCthQcArIUHANnfAICz6QYA3d8AgOHfAIC26QYA5d8AgOnfAIC16QYAukUBALtNAQDt3wCA8d8AgL5FAQC/TQEAvFUBAL1NAQCoIQYAqSEGAKolBgCrPQYArCUGAK0tBgCuSQYAr0EGAPXfAID53wCA/d8AgAHgAIAF4ACACeAAgA3gAIAR4ACAuEkBALlJAQC6WQEAu1EBALx5AQC9eQEAvhkBAL8VAQCwxQEAsc0BALLFAQCz3QEAtMUBALXNAQC2xQEAt3kBABXgAIAZ4ACAHeAAgKOhBQAh4ACApaEFAKahBQAl4ACAjyHqAyngAICqDQIAqwUCAKwdAgCtBQIArg0CAK8FAgCX7RIAlmUSAJVFEQCUnRYAk3EWAJJVFQCReesDkFnqA59hBgCeNQUAnUUaAJxpGgCbVRkAmkUeAJlZHgCYRR0A4WAAAC3gAIDjTD4AMeAAgKOxAgCi1QEAobUHAKCJBgCxATgAsAk+ALOVOgCyjToAtbUmALQBJADvaDoAvjAMAKnJNgCowTYAqwEwAKrhNwCtzTMArPUyAK/5PgCuATwAoRkCADngAICjbQ4Aom0OAKX1CgCkAQgAp4ULAKaZCgCGAA0Ah0QNAIIJ6wODCesDhDHqA4UVFACGORcAh80XAISgDQA94ACAiiUQAIsNEwCMnRMAjQ0cAI4ZHwCPDR8A5d4AgO8AAwCSbRgAk0kbAJR9GwCVBQQAllkHAJdJBwBB4ACAReAAgJpFBgCbLQAAnFEDAONgAABJ4ACA4WwAAIClAQCBAQEAggUBAL4ADABN4ACAUeAAgFXgAIDviAEAWeAAgOFUBgBd4ACA41QBAGHgAIBl4ACAaeAAgG3gAICz6QIAceAAgHXgAIB54ACAfeAAgLadAgC1mQIAgeAAgLuJAgC6vQIAheAAgIngAIC/WQIAvlECAL1ZAgC8kQIAoykNAI3gAICR4ACAleAAgJngAICmXQ0ApVkNAJ3gAICrSQ0Aqn0NAKHgAICp4ACAr5kNAK6RDQCtmQ0ArFENAIBRAACBWQAAgmEAALMtDwCt4ACAtS0PALbJDwCx4ACAhkADAIcIAwC6yQ8Au8UPALzBDwC9wQ8AvsEPAL/BDwA14ACApeAAgLXgAIC54ACAveAAgMHgAIDF4ACAyeAAgKhFDgCpgQ8AqskPAKvJDwCsyQ8ArSUPAK4tDwCvJQ8AsGEPALFtDwCyeQ8As3kPALRpDwC1aQ8Ath0PALcVDwC4LQ8AuTUPALo1DwC7BQ8AvB0PAL3xAAC+8QAAv/EAAKNhDgDN4ACAhMQBANHgAIDV4ACApoUOAKVhDgDZ4ACAq4kOAKqFDgDd4ACA4eAAgK+NDgCujQ4ArY0OAKyNDgDl4ACA6eAAgO3gAIDx4ACA9eAAgPngAID94ACAAeEAgAXhAICCHQAAgR0AAIAdAAAJ4QCADeEAgBHhAIC+tAEAqK0BAKnVAQCq1QEAqwUBAKwdAQCtBQEArg0BAK8FAQCGgAEAhxgBABnhAIAd4QCAIeEAgCXhAIAp4QCALeEAgLiFAAC5jQAAuoUAALudAAC8hQAAvY0AAL6FAAC/vQAAsH0BALHhAACy5QAAs/0AALTtAAC13QAAttUAALe9AACzXQIAMeEAgDXhAIA54QCAPeEAgLaFAgC1lQIAQeEAgLslAwC6uQIAReEAgEnhAIC/GQMAvikDAL0pAwC8MQMAvswEAKMZAgBN4QCAUeEAgKbBAgBV4QCAWeEAgKXRAgCq/QIAq2EDAF3hAIBh4QCArm0DAK9dAwCsdQMArW0DAKgpAwCpKQMAqjkDAKs5AwCsKQMArSkDAK6dAACvlQAAZeEAgGnhAIBt4QCAceEAgHXhAICCqQEAga0BAICtAQC4mQAAua0AALqlAAC7bQAAvHUAAL19AAC+dQAAv20AALDtAACx9QAAsvUAALPFAAC03QAAtb0AALa1AAC3qQAA4XgBAOEcDgDjEAAA4zwOAHnhAIB94QCAvhQEAIHhAICErAIAieEAgId4BQCGDAUAjeEAgJHhAIDvvAAA70gOALPxAgCV4QCAmeEAgJ3hAICh4QCAtukCALXhAgCl4QCAu3EBALppAQCp4QCAhKAEAL85AQC+WQEAvVEBALxhAQCt4QCAhIwEALHhAICEADgAteEAgLnhAIC94QCAweEAgKqJDgCriQ4AqLkOAKmxDgCu/Q4Ar+EOAKz5DgCt9Q4Asq0OALNlDgCwkQ4AsaUOALZ9DgC3ZQ4AtH0OALV1DgC6XQ4Au+UNALhdDgC5VQ4AvuENAL/pDQC8/Q0AvfUNAKOxBQCF4QCAxeEAgMnhAIDN4QCApqkFAKWhBQDR4QCAqzEGAKopBgDV4QCA2eEAgK95BgCuGQYArREGAKwhBgDd4QCA4eEAgOXhAIDp4QCAgB0AAIEJAACCOQAA7eEAgPHhAID14QCAhsgAAIcMAwD54QCA/eEAgAHiAIAF4gCAqKUHAKm1BwCqvQcAq8kHAKzZBwCt2QcArskHAK/BBwC+oAAACeIAgA3iAIAR4gCAFeIAgBniAIAd4gCAIeIAgLjNAAC51QAAutUAALvlAAC8/QAAvZUAAL6dAAC/lQAAsIkHALFlBwCyYQcAs30HALRlBwC1bQcAtmUHALf1AACzNQYAJeIAgCniAIAt4gCAMeIAgLZZBgC1UQYANeIAgLuhBgC6TQYAOeIAgD3iAIC/qQYAvqEGAL2pBgC8tQYAQeIAgEXiAIDv8AUASeIAgE3iAIBR4gCAVeIAgFniAICAPQAAgQkAAIIdAABd4gCA4cgGAGHiAIDjSAQAZeIAgKO1BgBp4gCAhigAAIdAAQBt4gCAptkGAKXRBgBx4gCAqyEGAKrNBgB14gCAeeIAgK8pBgCuIQYArSkGAKw1BgB94gCAs70BAIHiAICF4gCAtnkBAIniAICN4gCAtXkBALpVAQC7XQEAkeIAgJXiAIC++QAAv/kAALxFAQC9+QAAqHECAKlxAgCqcQIAq3ECAKy1AgCtvQIArrUCAK+tAgC+rDwAmeIAgJ3iAICh4gCApeIAgKniAICt4gCAseIAgLhpAwC5aQMAugkDALsJAwC8HQMAvQUDAL4NAwC/BQMAsNUCALHdAgCy1QIAs2kDALR5AwC1eQMAtmkDALdhAwC14gCAueIAgL3iAICj9QIAweIAgKUxAgCmMQIAxeIAgMniAIDN4gCAqh0CAKsVAgCsDQIArbEDAK6xAwCvsQMA7xgCAIIVAACBbQAAgG0AANHiAIDZ4gCAhvg8AIcYAwDd4gCA4eIAgOXiAIDp4gCA42wHABXhAIDhaAEA7eIAgKiFAgCplQIAqpUCAKulAgCsvQIArdUCAK7RAgCv0QIA8eIAgPXiAID54gCA/eIAgAHjAIAF4wCACeMAgA3jAIC4dQEAuX0BALp1AQC7zQEAvNUBAL3dAQC+yQEAv8EBALC1AgCxvQIAsoECALOBAgC0VQEAtV0BALZVAQC3TQEA4bQGABHjAIDj9AYAFeMAgIQYPQAZ4wCAHeMAgCHjAIAl4wCAKeMAgC3jAIAx4wCANeMAgDnjAIDvWAYAPeMAgIF9AACAcQAAQeMAgIIFAABJ4wCATeMAgO+AAQC+VDwA4ZABAFHjAIDjfAYAVeMAgFnjAIBd4wCAhtg8AIf0PACjnT0A1eIAgEXjAIBh4wCAZeMAgKbVPQCltT0AaeMAgKv5PQCq8T0AbeMAgHHjAICvGT4ArhE+AK3VPQCs1T0AdeMAgLOhPgB54wCAfeMAgLatPgCB4wCAheMAgLWxPgC6ST8Au0k/AInjAICN4wCAvkk/AL9JPwC8ST8AvUk/AKhVPgCpZT4Aqm0+AKtlPgCsfT4ArWk+AK65PwCvuT8AkeMAgJXjAICZ4wCAneMAgKHjAICl4wCAqeMAgK3jAIC4VT8AuV0/ALpVPwC7bT8AvHU/AL19PwC+dT8Av20/ALDJPwCxyT8Astk/ALPZPwC0yT8Atck/ALZ9PwC3cT8AghUAAKPhPwCAsQEAgbEBAKbtPwCx4wCAvtABAKXxPwCqCT4Aqwk+AITkAQC14wCArgk+AK8JPgCsCT4ArQk+ALPdPAC54wCAhugAAIfMAQC94wCAtpU8ALX1PADB4wCAu7k8ALqxPADF4wCAyeMAgL9ZPwC+UT8AvZU8ALyVPACoUT4AqVE+AKptPgCrYT4ArGE+AK1hPgCulQEAr40BAISgAQDN4wCA0eMAgNXjAIDZ4wCA3eMAgOHjAIDl4wCAuKkBALmpAQC6aQEAu2kBALx5AQC9eQEAvmkBAL9pAQCw/QEAsc0BALLFAQCzrQEAtLkBALW5AQC2rQEAt6UBALPlPQDp4wCA7eMAgPHjAID14wCAtuE9ALXpPQD54wCAuwkCALo5AgD94wCAAeQAgL99AgC+fQIAvXkCALwRAgAF5ACAo6E9AAnkAIAN5ACApqU9ABHkAIAV5ACApa09AKp9AgCrTQIAGeQAgB3kAICuOQIArzkCAKxVAgCtPQIAgOkAAIHpAACCHQAAvsADAO/kAgAh5ACAh1QDAIY8BADjEAEAKeQAgOH4AQAt5ACAMeQAgDXkAIA55ACAPeQAgEHkAIBF5ACASeQAgLORAwBN5ACAtbkDALZ9AwBR5ACAVeQAgFnkAIC6WQMAu1kDALxJAwC9SQMAvv0AAL/1AACoRQIAqVUCAKpVAgCrZQIArH0CAK2xAgCusQIAr7ECAIRsBQBd5ACAYeQAgGXkAIBp5ACAbeQAgL5wBQBx5ACAuF0BALltAQC6ZQEAuw0BALwZAQC9GQEAvg0BAL8FAQCw0QIAsdECALLRAgCz0QIAtHUBALV9AQC2dQEAt20BAOFAPwDjvAAA4wg+AOFsPgB15ACAeeQAgH3kAICB5ACAheQAgInkAICN5ACAkeQAgL5sBwDvVAAA75w+AJnkAICjnQIAgmkAAIFhAACAaQAAneQAgKZxAgCltQIAoeQAgKtVAgCqVQIAhsgEAIfsBACv+QEArvEBAK1FAgCsRQIAqKUGAKmpBgCquQYAq7kGAKypBgCtqQYArtkGAK/ZBgCV5ACApeQAgKnkAICt5ACAseQAgLXkAIC55ACAveQAgLhxBwC5cQcAunUHALvdBwC8xQcAvc0HAL7FBwC//QcAsKkGALG1BgCytQYAs40GALSVBgC1UQcAtlEHALdRBwCzMQYAweQAgMXkAIDJ5ACAzeQAgLYpBgC1IQYA0eQAgLtxBgC6bQYA1eQAgNnkAIC/lQcAvlEGAL1ZBgC8YQYA3eQAgKN1BgDh5ACA5eQAgKZtBgDp5ACA7eQAgKVlBgCqKQYAqzUGAPHkAID15ACArhUGAK/RBwCsJQYArR0GAIANAACBFQAAgh0AAPnkAID95ACAAeUAgITcAQAF5QCAhoAAAIcgAQAJ5QCADeUAgBHlAIAV5QCAGeUAgB3lAIAh5QCA43QEACXlAIDhyAUAKeUAgC3lAIAx5QCANeUAgDnlAIA95QCAQeUAgEXlAIBJ5QCA77QEAE3lAIBR5QCAqD0GAKlVBgCqVQYAq6kBAKy5AQCtuQEArqkBAK+pAQCErAEAVeUAgFnlAIBd5QCAYeUAgGXlAIBp5QCAbeUAgLhtAQC5BQEAugEBALsBAQC8BQEAvQ0BAL4xAQC/MQEAsNkBALHZAQCybQEAs2UBALR9AQC1ZQEAtmUBALdVAQCBvQMAgL0DALPVBQCCGQAAtTkCAHHlAIC+VAMAtjECAHnlAIB95QCAuxUCALoVAgC9uQIAvLECAL+pAgC+sQIAgeUAgKZpAgClYQIAhAAMAKONBQCF5QCAhvgMAId8AwCv8QIArukCAK3hAgCs6QIAq00CAKpNAgCJ5QCAjeUAgJHlAICV5QCAmeUAgJ3lAIDjIAEAoeUAgOGgAQCl5QCA70ACAKnlAICt5QCAseUAgLXlAIC55QCAveUAgMHlAICz8QMAxeUAgCXkAIDJ5QCAzeUAgLbpAwC14QMA0eUAgLu1AwC6tQMA1eUAgNnlAIC/lQMAvpUDAL2lAwC8pQMAqCkCAKkpAgCqOQIAqzkCAKwpAgCtKQIArlkCAK9VAgCAzQEAgQkAAIIZAADd5QCA4eUAgL58DQCHtA0AhhwMALgxAgC5PQIAujUCALvpAgC8+QIAvfkCAL7pAgC/6QIAsDECALExAgCyMQIAszECALQRAgC1EQIAthECALcRAgDp5QCA7eUAgPHlAID15QCA+eUAgP3lAIAB5gCA79QGAAXmAIDhVAYACeYAgOOkAACsDBUADeYAgBHmAIAV5gCAo/ECABnmAIAd5gCAIeYAgCXmAICm6QIApeECACnmAICrtQIAqrUCAC3mAIAx5gCAr5UCAK6VAgCtpQIArKUCAKghDgCpIQ4AqkkOAKtZDgCsaQ4ArWkOAK6ZDgCvmQ4A5eUAgDXmAIA55gCAPeYAgEHmAIBF5gCASeYAgE3mAIC49Q4Auf0OALr1DgC7iQ4AvJ0OAL2FDgC+hQ4Av7UOALDpDgCx6Q4Asv0OALPxDgC01Q4Atd0OALbVDgC3zQ4As8EOAIIVAACBtQAAgLUAAFHmAIC26Q4AteEOAL4QAAC7LQ4Aui0OAIRkAwBV5gCAvxkOAL4RDgC9JQ4AvCkOAFnmAICjhQ4AhogAAIdsAwCmrQ4AXeYAgGHmAIClpQ4AqmkOAKtpDgBl5gCAaeYAgK5VDgCvXQ4ArG0OAK1hDgCziQ4AbeYAgHHmAIB15gCAeeYAgLaBDgC1iQ4AfeYAgLuVDgC6jQ4AgeYAgIXmAIC/+Q4AvvEOAL2FDgC8hQ4AieYAgI3mAICR5gCAleYAgOMMDQCZ5gCA4RgNAJ3mAIDvrAwAoeYAgKXmAICp5gCAreYAgLHmAIC15gCAueYAgKgBDgCpAQ4AqgEOAKsBDgCsAQ4ArQEOAK4BDgCvPQ4AgN0AAIEJAACCGQAAveYAgMHmAICEPAEAvnQAAMnmAIC4HQ4AuS0OALolDgC76QEAvPkBAL35AQC+6QEAv+kBALBJDgCxUQ4AslEOALNRDgC0NQ4AtT0OALY1DgC3LQ4Ao4kNAM3mAICGrAQAhzwDANHmAICmgQ0ApYkNANXmAICrlQ0Aqo0NANnmAIDd5gCAr/kNAK7xDQCthQ0ArIUNAOHmAICznQIAhEgDAL5ABAC2VQMA5eYAgOnmAIC1sQIAunEDALt5AwDt5gCA8eYAgL4xAwC/MQMAvFEDAL1RAwCwkQMAsZkDALKhAwCzoQMAtNEDALXRAwC20QMAt9EDALj1AwC5+QMAus0DALvFAwC83QMAvcUDAL7NAwC/xQMA9eYAgPnmAID95gCAAecAgIV8GQAF5wCACecAgHXlAICoIQIAqTECAKoxAgCrBQIArB0CAK3xAwCu8QMAr/EDAA3nAIAR5wCAFecAgBnnAIDvUAAAHecAgCHnAIAl5wCA44QAACnnAIDh+AEALecAgIAVAACBGQAAggUAADHnAICjmQMAOecAgIZoBACHYAUAPecAgKZRAgCltQMAQecAgKt9AgCqdQIARecAgEnnAICvNQIArjUCAK1VAgCsVQIATecAgFHnAIBV5wCAWecAgF3nAIBh5wCAZecAgO/4AQC+bAQA4YAOAGnnAIDjFAEAbecAgHHnAIB15wCAeecAgH3nAICB5wCAhecAgLPdAQCJ5wCAtf0BALb1AQCN5wCAkecAgJXnAIC6sQEAu4UBALydAQC9NQEAvj0BAL81AQCpBQYAqLkFAKsVBgCqHQYArT0GAKw9BgCvTQYArl0GADXnAICCHQAAgR0AAIAdAACZ5wCAnecAgKHnAICl5wCAuUEHALidBgC7QQcAukkHAL1FBwC8WQcAv0UHAL5FBwCxCQYAsD0GALOpBgCyAQYAtbkGALSxBgC3rQYAtrEGAKORBgCEjAIAhigAAIfAAwCp5wCAprkGAKWxBgCt5wCAq8kGAKr9BgCx5wCAtecAgK95BgCucQYArXkGAKzRBgC55wCAs5kHAL3nAIDB5wCAtlEHAMXnAIDJ5wCAtbEHALptBwC7dQcAzecAgNHnAIC+WQcAv0UHALxtBwC9ZQcA1ecAgNnnAIDd5wCA4ecAgOXnAIDp5wCA7ecAgO+oBQDx5wCA4TQFAPXnAIDjdAUA+ecAgP3nAIAB6ACABegAgKMdBgCCLQAAgRUAAIAdAAAJ6ACAptUGAKU1BgAN6ACAq/EGAKrpBgAR6ACAhCgBAK/BBgCu3QYAreEGAKzpBgCoxQYAqdUGAKrVBgCr5QYArP0GAK0VBgCuHQYArxUGAL7sAQAZ6ACAhggAAIcgAAAd6ACAIegAgCXoAIAp6ACAuH0GALkFBgC6DQYAuwUGALwBBgC9CQYAvjkGAL85BgCwbQYAsXUGALJ9BgCzdQYAtFkGALVFBgC2TQYAt0UGAKiRAgCpmQIAqqECAKuhAgCs0QIArd0CAK7VAgCvyQIALegAgDHoAIA16ACAvyweADnoAIA96ACAQegAgEXoAIC4VQMAuV0DALppAwC7ZQMAvGEDAL1hAwC+YQMAv2EDALC5AgCxjQIAsoUCALNtAwC0dQMAtX0DALZ1AwC3bQMASegAgE3oAICzIQIAUegAgLVRAgCEiAMAVegAgLZVAgDF5gCAvigcALtBAgC6dQIAvbEDALxZAgC/sQMAvrkDAKNpAgBZ6ACAXegAgGHoAIBl6ACAph0CAKUZAgBp6ACAqwkCAKo9AgBt6ACAcegAgK/5AwCu8QMArfkDAKwRAgCopQIAqbUCAKq9AgCrtQIArK0CAK01AQCuPQEArzUBAL4sHAB16ACAeegAgH3oAICB6ACAiegAgIdoHQCGHB0AuIUBALmNAQC6hQEAu50BALyNAQC9vQEAvrUBAL95AACwUQEAsVEBALJRAQCzUQEAtPEBALXxAQC29QEAt+UBAO/YAACCtQAAgaUAAIClAACN6ACAkegAgJXoAIDvxAYAmegAgOH0BgCd6ACA4zgBAOPMAACh6ACA4SgBAKXoAICp6ACAtuUBALV1AgCEQBwAs2UCAK3oAICx6ACAtegAgL9lAQC+ZQEAvdUBALzVAQC7xQEAusUBALnoAIC96ACAo7UdAIXoAIDB6ACAxegAgMnoAICmNR4ApaUdAM3oAICrFR4AqhUeANHoAIDV6ACAr7UeAK61HgCtBR4ArAUeANnoAIDd6ACA4egAgOXoAICADQAAgTUAAII9AADp6ACA7egAgPHoAIC1BQAAdxoAgOG0AgCs2AIAtQUAAHsaAICotR8AqRUfAKodHwCrFR8ArDEfAK09HwCuLR8AryEfAOG0AgCs2AIAtQUAAH8aAIDhtAIArNgCALUFAACDGgCAuNEAALnZAAC64QAAu+EAALyRAAC9kQAAvpEAAL+RAACwIR8AsTEfALIxHwCzMR8AtAkfALUJHwC28QAAt/EAAOG0AgCs3AIA71QdALUdAACHGgCA4bwCAKzQAgC1KQAAoyUBAKKRAwChFR0AoA0dAOGAHgCLGgCA47wdAOHEAgCz1R4AtQkAAKzYAgCPGgCA4bwCALb9HgC1+R4ArOACALu1HgC6pR4AtQUAAJMaAIC/jR4Avo0eAL2lHgC8pR4AoxUeAOG8AgCs0AIAtREAAI9pJQCmPR4ApTkeAJcaAICrdR4AqmUeAOG0AgCseAEAr00eAK5NHgCtZR4ArGUeAJvdFACa5RUAmQEXAJjhEACfcR8AnnkZAJ35GQCcARsAk+UtAJIRLwCRbSkAkG0pAJf5EQCW8REAlYUsAJSZLQC1JQAA4ZQCAILxJgCDjSoAhJUqAIXhLACGHS4Ah3kuAKy0AgCbGgCAilUvAIspEgCMORIAjRkTAI7xFACPHRYAtQUAAJ8aAICSVRcAk5EYAJRxGgCV+RoAlvkcAJd9HgCC4AMAlgsAgJpVHgCb2QAAnHUCAIYMAIC2DACAuIkKAKwBBACthQYAroEGAMwQAgDMfAMAuQwAgKMaAIDFDACAyAwAgMsMAIADCwCAgaUyAr8MAIAV6ACAmpUGAJtVIwK8kQYAvbEAAL6RBgC/rQYAuOkGALmVBgC6kQYApxoAgLTBBgC1zQYAts0GALfdBgCw/QYAseUGALKdAACz5QYAhVTHA6saAICH/AAAuAEKALMaAIDsDACAtxoAgIzlDACNpAEAzPACAMQNAIDHDQCAiRQAALgZCgCLDAAAIA4AgFkOAIC8DACAwgwAgBwKAICRwAEAzgwAgLhtCgDRDACA1wwAgN0MAIDgDACA4wwAgLsaAIAuDQCA6QwAgL8aAIDhpB4AMQ0AgONUHgCvJXMAzCgCAPIMAIDvDACA9QwAgPgMAID7DACAzIACAJS4AwD+DACAkhQCAO9gHgCQAAIAAQ0AgA0NAIC48QoAEA0AgKILAIATDQCAiSkLABYNAICvGgCAvDABAL/EAQC+7AEAGQ0AgMzsAgC4xQoAukQBALAJAIAcDQCAygYAgN8GAIDyBgCAIg0AgPoGAIAlDQCACgcAgC0HAIAYBwCA+QcAgC8HAICvDQCAOgcAgLUNAIBKBwCAtXkAAGoHAIC3cSoCdQcAgLFhAAB3BwCAsw0pApAHAIC96QAAowcAgP0HAICwBwCAuRkrAsYHAIC7WRQCIggAgF0JAIA/CACANQ4AgF4IAIA5AACAhAgAgHEAAIDKCACAKwAAgCMJAIA9AACAXwkAgEMAAIBhCQCASAgAgG0IAIBJAACAAwgAgFMAAIB8CQCAWQAAgCgNAIBfAACAuw0iAtYNAIDMFDYCHwAAgL9lAAC+EQAAvW0AAOgHAICAaQEAgXUBAIJxAQCD3SEChGkHAIWBBwCGgQcAh3EBAIihAQCJrQEAirUHAIuNBwCMlQcAjaUBAE8AAICPpQEAkOEBAJHtBwCSsSECk/0HAJSNBwCVUQYAlvEBAJfZAQCY0QEAmXUGAJp9BgCb1QEAnGkGAJ2ZFAKeUQYAn1EGAKB1FAKhuQYAokkBAKOFLQKkIQEApS0BAKZ1FAKntQYAqKERAqlRFAKqlQYAtyEAgMy8NQLNPDUCbQAAgKoDAICsAwCArwMAgMMhAIDKIQCA4SEAgOghAIDJAACADwAAgLihBgC6BgCAtwYAgMwAAIDUIQCAtQMAgN0FAIAYBgCAugUCALvVAgC46QUAuf0FAL7JAgC/5RcCvA0CAL0BAgCy4QUAs+EFALCNBQCxnQUAtuUFALfpBQC09QUAte0FAKo9BQCrwQUAqD0FAKk1BQCuzQUAr/UFAKzNBQCtxQUAoj0FAKMFBQCg1QIAoTkFAKYdBQCnBQUApB0FAKUVBQC/BgCAm8EFAD4GAIBVBgCAnt0FAJ8xBACcUQIAndUFAHIGAICJBgCApAMAgDYiAIDbAACAoAMAgJIHAIDxBwCA9QcAgJMJAIAFCACACQgAgJkLAICXCQCAsgoAgHIHAICOBwCAmgcAgKUHAICtBwCArQkAgAEPAIAYDwCAJQ8AgMwEMwLNsDACzCAzAs3gMALMEDACzGgwAsxYMALNjDACzGgxAs0UMQLM1DECzRQ2AsxwIALN0CcCzDA2AswkMQLMDDwCzWg/AswYPwLNND8CzBg9As3AMgLMRDwCzBg5Asw4MgLNqDICzIgyAs34MwLMfDMCzUAzAswoMwLNCDMCzMghAs0kJgLMrCYCzEA4AsyYJQLNyDoCzBwkAs0QJALMhDsCzag7AsysJQLNvDoCzKw4Asz4JwLM4DgCzXQ4Ai0PAID2BgCAZw0AgI4NAIDNICoCzBwrAqoGAIAyIgCAzKQgAs2gJwLMOCYCygQAgMw4OgLNPDsCzBA5As1gPgLMoAMAvj0NAL3tLALWBACAu1UjAgcJAIC5PSICzwYAgNwHAIClBACApg0AgLIEAIBvBQCA9AYAgL4EAIB1BQCAr70MAK6ZLgKtpQwAwgUAgKvFIgIDBgCAxAQAgCMGAIDQBACAyAUAgCkGAIBdBgCAowEYAqAEAIAaBwCAHQcAgJ9dDACeUQwAnUUMACcHAICbWSECsgcAgLQHAIC3BwCAuwcAgCoHAIDRBwCA0wcAgJMtJgLWBwCAbwgAgHIIAICPBQwAjnEMAI1lDAB8CACAi0UgAmMJAICJNS8CZgkAgGoJAIB/CACAcwkAgHYJAIC9AwCABiIAgIFdDACAYQwAgAABAIEYAACCAAQACiIAgIQQBwCFFAYAhuQIAIc8AgCILAUAiaQFAIoAeAAOIgCAjJAGABIiAIAaIgCAFiIAgLgRAACRxHsAkkh6AJNMeQAiIgCAzOgCAJbwCQC4OQAAkMAJACoiAICS8AkAzPgCAJS0CQC4DQAALiIAgMwcAgC4BQAAOiIAgMzkAgC4HQAAPiIAgEIiAIBJIgCAYCIAgKiMCACp5HsAZyIAgKt8BADM5AIAuA0AAHEiAIDMlAIAdSIAgLGUewC4CQAAuBUAAMz8AgC15AgAeSIAgMzYAgB9IgCAuAUAALroBQC7hAQAvAB8AL30fwC++H0Av/xyAIAJOgKBDToCggE6AoMFOgKEGToChR06AoYROgKHFToCiCk6AoktOgKKIToCiyU6Aow5OgLMhAIAjjE6Ao81OgKJIgCAkekPAIUiAIDMzAIAuBEAAJUiAIDM3AIAl+UPALgpAAC4MQAAzPgCAJkiAIC4HQAAzDwCAJ0iAIDMLAIAzIQCAKEiAIClIgCAzIQCAKQtDwClVQ8Apl0PAKkiAICoqToCqa06ArjRAACtIgCAuDUAALEiAIDMUAMAr7U6AswsAwDMFAMA1SIAgLMFDwC0HQ8AzAADALYJDwC3CQ8Avmh9ALhtAAC4dQAAzEgDALwpDwDZIgCAviUPAMxQAwCH5Q4AzOg6ArilAQC4uQEAzPA1As2kMwLMgCICzXwlAs2UNgLMBCkCzew7AsxkOgK4+QEAuMEBAInVDgCI1Q4Al7EOALgNAAC1IgCAuB0AALkiAIC4DQAAvSIAgMEiAICfaTsC3SIAgOEiAIC4MQAAzMQCAMz4AgDFIgCAySIAgLjlAADNIgCA0SIAgLjlAAC46QAAuOkAAMzMMwLlIgCAuCUAAMzoMwK4IQAA6SIAgO0iAIC4KQAAzNgCALgRAAC3TQ0Atk0NALU1DgC0NQ4AuGEAAPEiAICxGQ8AsCkOAL/1AwC+UQ0AvVkNALw1DAC7XQ0Aul0NALldDQC4XQ0AgL0KAIHFCgCCFQQAg8kKAMx8BQCF3QoAhtUKAIfNCgDMUAUAifEKAIq5CACLDQgAjBEIAI0VCACOtScCj+UKAJBpCACRbQgAknEIAJNtJALMJAUAlR0IAJaFCgDMHAUAzCgFAJk9CACaiQoAmw0IAJwRCACdFQgAzCwFAMwgBQCgZQoAoW0KAKJlCgDMvAUApLEEAMzoAgCmsQQAuEkHAKiBBAAbIwCAqpkIAKtdCgCsuQgArakEAB8jAICvNQgAsNEIALHxBADQAwCAs40IALQpKAK1IQoAtiEKALchCgC4IQsAuSUIAOkDAIC7KQsAvA0dAr3dDwC+MQsAvzELAIDdCgCpoQEAqrEBAAIEAIAbBACAhRkJAIaZCQCHlQkAiOEJAIklJQIuBACAQQQAgFQEAIBnBACAegQAgI0EAICQrQoAkUkFAJJtBQCTYQUAlGEFAJVtBQCWZQUAlxEFAJg1BQCZPQUAmjUFAJsNBQCcFQUAnR0FAJ4VBQCfCQUAoKkJAKH9BQCi9QUAowEFAKQFBQClDQUApgUFAKc9BQCoBQUAqQ0FAKoFBQCrGQUArIkJAK2pBQCutQkAr/0JALABCQCxfQUAsnUFALMBBQC0aQkAtQEFALYFBQC3PQUAuAUFALnhJQK6AQUAuwEFALzRJQK9PQkAvnkJAL9dCQCDMAUAoXgHAPMEAIDdAACApHgHAKVMBwATAQCAHAEAgIt8BAAgAQCAJAEAgIhIBAAoAQCALAEAgDABAIA0AQCA4QAAgOYAAICyDAcAswAHAOsAAIDwAACAtugHALfgBwD1AACA+gAAgLrgBwC7nAcAvIQHAL2QBwD/AACAn9F+AKPMBAAEAQCACQEAgIMABAAOAQCAhXQEAKUgBAAXAQCAiEwEAM0DAIDwBACAjwUAgK8tBwCNsAcArSEHAKwpBwDiBQCAHQYAgEMGAICwZQcAWgYAgHcGAICOBgCA0wMAgOwDAIAFBACAHgQAgDEEAIBEBACAVwQAgGoEAIC8fAQAgt0rAoPlKwKA/QoAgfkrAoaZCQCHmQkAhOEKAIXhCgCKiQkAi4kJAIiJCQCJiQkAjoUJAH0EAICM4QgAjY0JAJK5KwKTQScCkJkrApHFCwCWyQsAl3UnApTFDQCV0SQCmskLAJvZKgKYyQsAmXkHAJAEAID2BACAnP0LAKABAICkAQCAqAEAgKwBAICwAQCAtAEAgLgBAIC8AQCAwAEAgJw9fwCzFX8AqBEJAMQBAIDIAQCAzAEAgNABAICC2H4A1AEAgNgBAIDcAQCA4AEAgOQBAIDoAQCA7AEAgPABAID0AQCA+AEAgPwBAIAAAgCABAIAgDwJAIBTIgCAogYAgKD1VAKh2VQCoulUAqP1dQCk7XUApZ12AKaVdgCnvXYAqIV2AKntfACqwXwAqyF9AKwhfQCtHX0ArhV9AK8NfQCwdX0AsX19ALJ1fQCzRX4AtF1+ALVNfgC2RX4At3l+ALhJfgC5VX4Aul1+ALtVfgC8TX4AvTV+AL49fgC/LX4AlQcAgKwGAIDaBwCArwYAgLRtfwC1EQAAthUAAAkjAIC8oVgCvTF4AA8jAIDTMQCAPzoAgJ8qAIDDKgCAzyoAgN8qAIDnKgCA8yoAgPsqAIADKwCADysAgGorAICCKwCAkisAgKIrAICyKwCAwisAgOIrAIDmKwCA6isAgB4sAICAVX8AgWV/AIJtfwCDdX8AhJV/AIWdfwCGiX8Ah4F/AIiFfwCJjX8AioV/AIvtfwCM9X8Ajf1/AI7pfwCP6X8AkJl/AJGZfwCSqX8Ak6l/AJS5fwCVuX8Alql/AJepfwCYmX8AmVF+AJoZfgCbGX4AnA1+AJ31fgCe/X4An/V+AKANfgChFX4Aoh1+AKMVfgCkDX4ApTl+AKYpfgCnKX4AqBl+AKllfgCqbX4Aq2V+AKx9fgCtZX4Arm1+AK9lfgCwHX4AsSV+ALItfgCzJX4AtD1+ALUlfgC2WXcAt9V1ALj9eQC56XUAuvl1ALvZeQC86XUAvdV1AL7RdQC/2XUAgDF2AIE9dgCCSXYAg0V2AIRBdgCFTXYAhvl0AId9dgCIoQIAiU12AIpZdgCLuXoAjEl2AI2degCOsQIAjx16AJCRVgKRKXYAkoF2AJPNdgCU2XYAlel2AJbJdgCX0VkCmKF2AJllWgKa8XYAm01aApzRdgCdYXoAnoFWAp/VdgCgdX0AoY1aAqI1VwKjCXYApCF2AKUtdgCmiVoCp5laAqi5WgKpdXYAql13AEYsAIBWLACAXiwAgGIsAIBuLACAiiwAgI4sAICmLACAqiwAgLIsAIDCLACAXi0AgHItAICyLQCAxi0AgM4tAIDSLQCA4i0AgAUuAIAxLgCAPS4AgJlpCgBdLgCAaS4AgG0uAIBxLgCAiS4AgI0uAIC5LgCAgvx6AIPoegDFLgCAzS4AgIZwewCHcHsA1S4AgOUuAID0LgCA/C4AgCgvAIAsLwCANC8AgDgvAIBALwCASC8AgJJAfABYLwCAdC8AgJHkewDsLwCAADAAgAQwAICEMACAiDAAgKvUfACo6HwAqeB8AJwwAICgMACAqDAAgLAwAICisHwAuDAAgMQwAID6MACAzEBJAs0ASQLM/EoCzWhLAgoxAIAeMQCAmzEAgKcxAIC3MQCAwzEAgM8xAIDXMQCAsqh8ALOofADbMQCA3zEAgOMxAIDnMQCAtER8ALVYfACAfQcAgYUHAIKZBwCDmQcAhIkHAIWJBwCGvQcAh7UHAIiNBwCJ5QcAiu0HAIvlBwCM/QcAjeUHAI7tBwCP5QcAkJ0HAJGtBwCSpQcAk70HAJSlBwCVVQEAll0BAJdVAQCYbQEAmXUBAJp9AQCbdQEAnG0BAJ1VAQCeXQEAn1UBAKCtAQChtQEAor0BAKO1AQCkrQEApdUBAKbdAQCn1QEAqO0BAKn1AQCq/QEAq/UBAKztAQCt1QEArt0BAK/VAQCwrQEAsbUBALK9AQCztQEAtK0BALVVAQC2XQEAt1UBALhtAQC5dQEAun0BALt1AQC8bQEAvVUBAL5dAQC/VQEAnzIAgOcyAIDzMgCA9zIAgPsyAID/MgCABzMAgAszAIAfMwCAOzMAgEMzAICDMwCAhzMAgI8zAICTMwCAmzMAgJ8zAIDDMwCAxzMAgOMzAIDnMwCA6zMAgO8zAIADNACAJzQAgCs0AIAvNACAUzQAgJM0AICXNACAtzQAgMc0AIDPNACA7zQAgBM1AIBXNQCAXzUAgHM1AIB/NQCAhzUAgI81AICTNQCAlzUAgK81AICzNQCAzzUAgNc1AIDfNQCA4zUAgO81AID3NQCA+zUAgP81AIAHNgCACzYAgKs2AIC/NgCA8zYAgPc2AID/NgCAnpkMACs3AIAzNwCAOzcAgICtAwCBtQMAgr0DAIO1AwCErQMAhdUDAIbdAwCH1QMAiO0DAIn1AwCK/QMAi/UDAIztAwCN1QMAjt0DAI/VAwCQrQMAkbEDAJKxAwCTsQMAlAEMAJVVDgCWXQ4Al1UOAJhtDgCZdQ4Amn0OAJt1DgCcbQ4AnVUOAJ5dDgCfVQ4AoK0OAKG1DgCivQ4Ao7UOAKStDgCl1Q4Apt0OAKfVDgCo7Q4AqfUOAKr9DgCr9Q4ArO0OAK3VDgCu3Q4Ar9UOALCtDgCxtQ4Asr0OALO1DgC0rQ4AtVUOALZdDgC3VQ4AuG0OALl1DgC6fQ4Au3UOALxtDgC9VQ4Avl0OAL9VDgC8aQQAvXUEAL59BAC/dQQAuEUEALlNBAC6RQQAu3kEALRlBAC1bQQAtmUEALd9BACwHQQAsQ0EALIFBACzfQQArFUEAK1dBACuVQQAr2UEAKhVBACpXQQAqlUEAKtNBACkkQUApZEFAKaRBQCnkQUAoJEFAKGRBQCikQUAo5EFAJxRBQCdUQUAnlEFAJ9RBQCYUQUAmVEFAJpRBQCbUQUAlBEFAJURBQCWEQUAlxEFAJBFBwCRTQcAkkUHAJMRBQCMJQcAjS0HAI4lBwCPPQcAiAUHAIkNBwCKBQcAiz0HAIQlBwCFLQcAhiUHAIc9BwCARQcAgU0HAIJFBwCDPQcAfzcAgIM3AICLNwCAjzcAgJM3AIC/NwCAwzcAgMs3AIDfNwCA4zcAgP83AIAHOACACzgAgC84AIBPOACAYzgAgGc4AIBvOACAmzgAgJ84AICvOACA0zgAgN84AIDvOACABzkAgA85AIATOQCAFzkAgBs5AIAnOQCAKzkAgDM5AIBPOQCAUzkAgFc5AIBvOQCAczkAgHs5AICPOQCAkzkAgJc5AICfOQCAozkAgKc5AICrOQCArzkAgL85AIDXOQCA2zkAgOc5AIDrOQCA7zkAgPM5AID7OQCA/zkAgAM6AIAPOgCAFzoAgB86AIAjOgCAKzoAgC86AIAzOgCAOzoAgICtAQCBtQEAgr0BAIO1AQCErQEAhdUBAIbdAQCH1QEAiO0BAIn1AQCK/QEAi/UBAIztAQCN1QEAjt0BAI/VAQCQrQEAkbUBAJK9AQCTtQEAlK0BAJUNAABDOgCAQyMAgHIsAIB2LACAKyQAgIJ4AgCZBQAAuSMAgILQAgC+mAIAgIAAAIGYAACC5AYAg4gEAITUGwCFlBoAhhgfAJ7pAACIxB4AiQAQAIqoEwCLrBEAjAAoAI20KwCOuCoAj7wpAON0dADjoAIA47gCAJkdAAC9IwCAvnQCAJ4JAADv2AIAgmwCAMEjAICZDQAA4wQCAO/ocQDvcAIA79QCAL6MAADjBAMAxSMAgOMcAwDJIwCA4yADAM0jAIDjTAMA0SMAgO/kAwDVIwCA74QDANkjAIDv2AMA3SMAgO/sAwDhIwCA48gDAOPcAwDvoAQA4+gDAEM3AIDjqAMA5SMAgO+kBADXAAAA70wDAOkjAIDjAAQA7yADAO88AwDjeAQA70QDAO0jAIDxIwCA9SMAgPkjAIDvWAQA/SMAgO9cBADjgAQA44AEAAEkAIDjmAQA73QEAAUkAIAJJACADSQAgBEkAIAcJACA72AEAOOIBAAgJACAvQAAgMIAAIA3JACAJCQAgHMpAIAOJQCAcSUAgLQlAIDgJQCA46QEAO9EBAAKJgCAgAlLAoZIAQCe7QIAgnACAL58AgCeEQEAmR0BAIJcAgCPQAEAmSkBAI1sAQC+YAIAi3gBAJ45AQCCGAIAvggCAJfUAgCZUQEAldQCAJ5ZAQCT1AIAvlgCAJHUAgCCVAIAn7gCAJYAAACdqAIAmXEBAJu4AgCeaQEAmbQCAJlZAQCCmAIAppQCAJ6tAQCkeAIAgmgCAJ65AQCheAIAmbEBAK+EAgAvJgCAvlgCAIJwAgC+XAIARCYAgJmNAQCokAIAvnwCAJ7xAQC1sAIAmfEBAL5QAgCykAIAtoUCAJmFAQC4XQkAuYUCALqNAgBMJgCAu/QEAIJcAgCexQEAuPQEAL58AgCeXQYAgmACAJllBgC+cAIAgmwCAJ5xBgCZnQYAvnwCAJ6lBgCCYAIAmakGAL58AgCesQYAghwCAL4UAgCZyQYAvkwCAIJMAgCawQYAntkGAJ/ZBgCCvAIAvrACAJn1BgCC3AUAvogCAJrJBgCe5QYAn9EGAIK4AgCCqAIAmTEGAOP0AgDj/AIAmjkGAJ8lBgCeJQYAniEGAJ8hBgC+GAIAmR0GAJoVBgC+DAIAmXEGAO8kAgDv4AIAmnEGAJ4BBgCfCQYA4xQCAJkVBgDjLAIAgnwCAJ4BBgBoJgCA7/ACAJkFBgC+fAIAng0GAHAmAICCdAIA76wCAIKUAwCb0QcA4/ADAL5QAgCZ6QcAn80HAOM0AgCdXAIAnMkHAJ7FBwDv7AIAgmACAJnFBwC+cAIA78QCAJ7RBwCCEAIA43QCAL5sAgCZpQcAvlQCAJ69BwCZpQcAgmgCAOMMAgCekQcAmZkHAIJoAgDv2AIA78wCAL6IAgCCkAIAvpgCALsEAACeeQcAuQAAAJkpBgC/XAAA4/QCAL0EAACeOQYAs7gDAO8oAgCxeAMAgnwCALc8AACZAQYAtfwDAJ4JBgCrSAMAvgACAOO4AgCZIQYArxwDAIJ8AgCtQAMAvkQCAJ4NBgCCbAIAvpwBAJkxAQCCgAEApqgCAO98AgCCvAEA46QBAOPYAQDj5AEAntECAOPoAQCZ5QIAvnwCAJ7tAgDvwAIAmQkAAL5sAgB4JgCA7wwBAO8QAQDvKAEAnhEAAOO0AQDjHAIAmS0AAL5sAgCCfAIAglACAJ49AADjAAIAmeUAAIomAIC+aAIA7/wCAO+gAgDvwAIAnv0AAJnxAADjAAIAliYAgIKQAgCeJgCAvnwCAJ4ZAADjDAIAglgCAJkNAACCBAIA7+QCALQmAIDv2AIAvnQCAJ4VAACZZQEA42wCAJnJAQCanQEAvmQCAJ7dAQCfgQEA4wwCAIJYAgCZlQEAvrQDAO/0AgCalQEA78QCAIKsAwCwRSoAgvQDAIHIAgDj+AMAvjACAPImAIDj5AMAhMQCAL44AgCGEAIA78ACAIgwAgCeXQAAn1UAAJqxAADvxAIAj3QCAJlBAACePQAAn8UAAIKMAgCSGAIAldgDAAMnAICeyQAAn8kAAJr1AACY+AMAm2ADAJn9AACCqAMAJScAgDAnAICAJwCAlCcAgOPEAgC+IAIAvlACALYnAIDEJwCA4+ACAAwoAICZPQAAnn0AAO/cAgCaHQAAGigAgO/EAgCvrAMAghACALEAHACwlAMAmSkAALJMHACeTQAAn2UAAHcpAIDjaAIAeykAgL50AgCeCQAA7/ACAL08HACCbAIAv80fAJn9HwB/KQCAvnQCAJ4JAADjFAIAgmwCAIMpAICZDQAAvkQCAJ41AACCaAIAmQUAAI8pAIC+cAIA7yUAgJ4VAADv2AIA42wCAP8YAIADGQCA47AdACcaAIAHGQCAKxoAgC8aAIALGQCAMxoAgDcaAIA7GgCA77gCAD8aAIBDGgCA74ACALHFAAAPGQCAs9kAALLFAAC1yQAAtMEAALf5AAC2wQAAuWEAALghAAC7wQAAuskAAL3FAAC82QAAv30AAL7FAAATGQCARxoAgFMZAIAXGQCAGxkAgB8ZAIBnGQCA7xB4A7MZAIDh3E0DtxkAgONYeQO7GQCAvxkAgMMZAIDHGQCAgMkBAIHJAQCC2QEAg9kBAITJAQCFdQIAhgEEAIcdBQCIJQUAiTUFAIo9BQCLbQUAjHUFAI1lBQCObQUAj90BAJCpAQCRtQEAkr0BAJO1AQCUrQEAlVUDAJZdAwCXVQMAmG0DAJl1AwCafQMAm3EDAJxRAwCdUQMAnlEDAJ9NAwCgtQMAob0DAKK1AwCjxQMApMEDAKXBAwCmxQMAp/0DAKjFAwCpzQMAqsUDAKvZAwCsyQMArTUDAK49AwCvNQMAsE0DALFVAwCyXQMAs1UDALRNAwC1dQMAtn0DALd1AwC4TQMAuVUDALpdAwC7VQMAvE0DAL01AwC+PQMAvzEDAMsZAIDPGQCA0xkAgNcZAIDbGQCA3xkAgPAQAgDjGQCA5xkAgOsZAIDvGQCAgpwCAPMZAID3GQCA+xkAgP8ZAICc9TYAnf02AAMaAICRaAIArxkAgEsZAIC6ZdUATxkAgEsaAIBPGgCAUxoAgFcaAIDwSAIAWxoAgF8aAICRkAIAYxoAgFcZAIBnGgCAaxoAgFsZAIBfGQCAYxkAgGsZAIBvGQCAcxkAgHcZAIB7GQCAfxkAgIMZAICHGQCAixkAgI8ZAICTGQCAuo3VAJcZAICbGQCAnxkAgG8aAIBzGgCAoxkAgILEAgCnGQCAqxkAgAcaAIALGgCADxoAgPA4AwDh3NICExoAgONUxgIXGgCAGxoAgB8aAIAjGgCAqyoAgGItAICvKgCAxyoAgLMqAICjMwCAuyoAgO+k4AKjKgCA4xSHAuGQ8wLhjJ0C45D3AuGMqwLhkLYC4wBTA+OAogIjGQCA7S0AgO+0SwPvvLMC7ySLAnYtAIAIAgCA7ySXApEIBwAOAgCAFAIAgBoCAIAgAgCAJgIAgCwCAIAyAgCAOAIAgD4CAIBEAgCASgIAgFACAIBWAgCANAMAgDoDAIDhgHgC4wAeAuMUagLhEBMC4aAPAkADAIDjhA4C7yw/AkYDAIDhUDsC7zQ7AuM8IgJMAwCA7ygfAu8MEgJSAwCAKxkAgC8ZAIBYAwCAXgMAgDMZAIA3GQCAdgMAgIIDAICIAwCAjgMAgJQDAICaAwCAfAMAgGQDAIBtAwCAXAIAgDsZAIA/GQCAdAIAgGgCAIBDGQCARxkAgLwCAIB6AgCAmAIAgGICAICSAgCAbgIAgKQCAIDUAgCA8gIAgOwCAICAUQYAgVEGAIJRBgCDUQYAhHEGAIV9BgCGdQYAh20GAIhVBgCJXQYAiq0HAIuhBwCMoQcAjaEHAI6hBwDgAgCALgMAgMICAICSfRQAkwEUAJTNBwCV9QcAlv0HAJf1BwCYzQcAmdUHAJotFACb2QcAnM0HAJ2RBwCejQcAnykUAJklAQCYJQEAmyUBAJolAQCdJQEAnCUBACcZAICeJQEAkcEGAJDxBgCT1QYAkt0GAJU9AQCUPQEAlyUBAJYlAQCJ5QYAiOUGAIvlBgCK5QYAjeUGAIzlBgCP5QYAjuUGAIHlBgCAHQYAg+UGAILlBgCF5QYAhOUGAIflBgCG5QYAuaUDALilAwC7pQMAuqUDAL2lAwC8pQMAv6UDAL6lAwCxpQMAsKUDALOlAwCypQMAtaUDALSlAwC3pQMAtqUDAKmxAQCoqQEAq7EBAKq5AQCtbQEArKkBAK8dAQCuHQEAoakBAKDZAQCjsQEAoqEBAKWRAQCkqQEAp5EBAKaZAQDOAgCA5gIAgNoCAIAEAwCAsAIAgPgCAIAiAwCACgMAgJ4CAICAAgCAtgIAgMgCAID+AgCAhgIAgCgDAICqAgCAEAMAgIwCAIAWAwCAHAMAgBYtAID4LgCA1zQAgIcHAIAGBQCAFQUAgCQFAIAzBQCAQgUAgEsFAIBUBQCAXQUAgIIMAwBmBQCAkgUAgJsFAICkBQCA40huA6cFAIDhTG4DqgUAgO/0AQOtBQCAVzoAgLdMAIDnVQCAR2gAgHdxAICnegCAB40AgGefAICXqACA/roAgDXEAIBlzQCAldYAgMXfAIBCuwCAS64AgBelAID/KgCAlisAgKcqAIDrKgCATjEAgA4xAIBbNACA4iwAgBMzAICXNwCAbzQAgCosAICfNACAqzMAgB84AIBmKwCAkiwAgAcyAIA3OQCAKisAgLorAICrMQCAyS4AgNYsAIBmLACARS4AgDkuAID7MwCAJisAgLop0ACrNwCAgiwAgNotAICwBQCAswUAgLYFAIDh1D8D4VgaA+PcLwPjUA4D4RTyA+FA0wPjQOoD40DDA7kFAIDlBQCA73jrA+9c8gPoBQCA6wUAgO9E3gPvmCUD4bSLA+E8lwPjfKID45iLA+EwQQDhUKwD4xx/AOOIRgDuBQCA8QUAgO84ewDv4EEA9AUAgPcFAIDvzIoD7yCHA4C1GACByRgAghULAIMtCwCE4Q4AheEOAIbhDgCH/RgAiN0OAImZGgCK6RsAiy0dAIzFGwCN8RsAjn0QAI/hGgCQ6RsAkUUPAJJNDwCTRQ8AlF0PAJVFDwCWTQ8Al0UPAJh9DwCZRQ8Amk0PAJuZGgCcWQ8AnVkPAJ5JDwCfSQ8AoLkPAKG5DwCiyQ8Ao8kPAKS1CwClvQsAprULAKfVDwCo7Q8AqfUPAKr9DwCr9Q8ArO0PAK3VDwCu0Q8Ar9EPALCxDwCxsQ8AsrEPALOxDwC0cQ8AtXEPALZpDwC3aQ8AuAEPALkBDwC6GQ8AuxkPALzxAQC98QEAvvEBAL/xAQD6BQCA/QUAgAAGAIAgBgCA4QQAgIAFAIDTBQCADgYAgDQGAIBLBgCAaAYAgH8GAICWBgCA3QMAgPYDAIAPBACAEgcAgEQIAIBBCACAPwcAgD8kAIB4JACAqSQAgM4kAIC/JgCAyiYAgM4mAIDSJgCA1iYAgDUoAIB0KACAnCgAgKAoAIDFKACAzSgAgOkoAID7KACA/ygAgAMpAIAbKQCANikAgPAUNgBRKQCAHysAgEMkAIBQJACAXSQAgGokAIB8JACAiSQAgJskAICtJACAvSQAgNIkAIDcJACA6iQAgPQkAIABJQCAEiUAgBwlAIB1JQCAfCUAgColAICGJQCAgBEDAIERAwCCEQMAgxEDAIQxAwCFMQMAhjEDAIcxAwCIEQMAiREDAIoRAwCLEQMAjHEDAI1xAwCOcQMAj3EDAJARAwCREQMAkgEEAJMVAwCUDQMAlVUGAJZdBgCXVQYAmG0GAJl1BgCafQYAm3UGAJxtBgCdNQYAnj0GAJ81BgCgzQYAodUGAKLdBgCj1QYApPEDAKXxAwCm8QMAp/EDAKjRAwCp+QYAqikGAKspBgCsOQYArTkGAK7NAwCvxQMAsL0DALFFAwCyTQMAs0UDALRdAwC1RQMAtk0DALdFAwC4fQMAuUUDALpBAwC7fQYAvGUGAL1tBgC+ZQYAv1EDAKkVDwCoAQ8Aq00PAKpNDwCtRQ8ArEUPAK+pDQCusQ0AoXULAKBhCwCj6QsAok0LAKXhCwCk+QsApzkPAKZdCAC5mQ0AuJENALupDQC6kQ0AvbkNALyxDQA3JQCAvrENALHZDQCw0Q0As6kNALLRDQC1uQ0AtLENALepDQC2sQ0APiUAgE4lAIBhJQCAuCUAgMIlAICXJQCApyUAgNYlAICB5Q0AgOUNAIPlDQCC5Q0AheUNAITlDQCH4Q0Ahi0YAJlFDQCYuQ0Am0UNAJpFDQCdSQ0AnEUNAJ9xDQCefQ0AkYENAJC5DQCTgQ0AkokNAJWBDQCUmQ0Al4ENAJaJDQDmJACAJiUAgJMlAIDSJQCA5CUAgA4mAIAzJgCASCYAgPYlAIAAJgCAEiYAgB8mAIA3JgCAVCYAgF4mAIB8JgCAUCYAgGwmAIB0JgCAhiYAgJImAICaJgCAqSYAgOQmAICiJgCAuCYAgK0mAIDDJgCA2iYAgOgmAIAHJwCAFycAgCEnAIBVJwCAmCcAgO0nAIBVKQCAYykAgGcpAIBrKQCA9iYAgDQnAIBEJwCATicAgCknAIBZJwCAaScAgIQnAIB2JwCAnCcAgMgnAIDPJwCArCcAgNknAIDjJwCAuicAgB4oAIAQKACA8ScAgCsoAID4JwCAAigAgDkoAIBGKACAUCgAgFooAIBkKACAeCgAgIUoAICMKACApCgAgKsoAIC4KACA0SgAgNsoAIDtKACABykAgBQpAIAfKQCAKSkAgDopAIBBKQCAWSkAgMMDAIDmBACAhQUAgNgFAIATBgCAOQYAgFAGAIBtBgCAhAYAgJsGAIDjAwCA/AMAgBUEAIAoBACAOwQAgE4EAIBhBACAdAQAgIcEAICaBACAAAUAgA8FAIAeBQCALQUAgDwFAIBmCACAJwgAgMEGAID/BwCAIAkAgOM4EwA2CQCALQgAgDAIAIA0CACAJAcAgOkuAIDXMACA5i0AgMgwAIBSMQCAKgkAgO/kEwAJCQCA4g0AgNIIAICGCACAMQcAgEwHAID8BgCADQgAgJcIAIAtCQCADAkAgOYNAIDyDQCA3ggAgJwIAIAVBwCAiQgAgFUHAID/BgCAqQcAgJckAID2DQCA5QgAgCoIAICfCACAWwgAgBgJAID6DQCA6AgAgBcIAICiCACA6wgAgBoIAIDMCACApQgAgO8IAIAeCACAzwgAgKkIAID6CACAAAkAgIsHAICNCACAWQcAgAMHAIBACQCARAkAgEwJAIA5CQCAGwkAgP4NAID3CACAMAkAgA8JAIDqDQCA1QgAgJEIAIBgBwCAMwkAgBIJAIDuDQCA2AgAgJQIAIBjBwCAsAgAgGYHAIDjQBIA4/ATAOP0EwDj2BMA4wgNAOPoEgDjrBIA42QSAO/EDQDv2A0A7zQSAO9YEgDvQBIA79QSAO/IEgDvIBMA7AcAgMwGAIARCACAFAgAgNgGAIDUBgCAJAgAgAcHAIBqCACADAcAgHkIAIA0BwCANwcAgK0IAIC5CACAvAgAgOPQEADjsBAA46gQAON4EQDjTBAA4zAQAOPoEADj/BAA74QQAO+YEADvLBAA7yAQAO8EEADvCBAA73AQAO9MEADjmBMA4zAQAOMwEADjIBAA43wTAONAEwDjWBMA44ATAO/UEwDvqBMA74ATAO98EwDvRBMA7yQTAO8YEwDv7BAAgO08AIH1PACC/TwAg/U8AITtPACFFT0Ahh09AIcVPQCILT0AiTU9AIo9PQCLNT0AjC09AI0VPQCOHT0AjxU9AJBtPQCRdT0Akn09AJN1PQCUbT0AlRU9AJYdPQCXFT0AmC09AJk1PQCaPT0AmzU9AJwtPQCdFT0Anh09AJ8VPQCg7T0AofU9AKL9PQCj9T0ApO09AKUVPQCmHT0ApxU9AKgtPQCpNT0Aqj09AKs1PQCsLT0ArRU9AK4dPQCvFT0AsG09ALF1PQCyfT0As3U9ALRtPQC1FT0AthE9ALcRPQC4MT0AuTE9ALoxPQC7MT0AvBE9AL0RPQC+ET0AvxE9AIDxPACB/TwAgvU8AIMNPwCEFT8AhR0/AIYVPwCHDT8AiDU/AIk9PwCKNT8Aiw0/AIwVPwCNHT8AjhU/AI8NPwCQdT8AkX0/AJJ1PwCTDT8AlBU/AJUZPwCWCT8Alwk/AJg5PwCZOT8Amgk/AJsJPwCcGT8AnRk/AJ4JPwCfCT8AoPk/AKH5PwCiCT8Aowk/AKQZPwClGT8Apgk/AKcJPwCoOT8AqTk/AKoJPwCrCT8ArBk/AK0ZPwCuCT8Arwk/ALB5PwCxeT8Asgk/ALMJPwC0GT8AtRk/ALYJPwC3CT8AuDk/ALk5PwC6CT8Auwk/ALwZPwC9GT8Avgk/AL8JPwCA+TwAgfk8AIJJPQCDST0AhFk9AIVZPQCGST0Ah0k9AIh5PQCJeT0Aikk9AItJPQCMWT0AjVk9AI5JPQCPST0AkDk9AJE5PQCSAQQAk00GAJRVBgCVXQYAllUGAJdNBgCYdQYAmX0GAJp1BgCbTQYAnFUGAJ1dBgCeVQYAn00GAKC1BgChvQYAorUGAKPNBgCk1QYApd0GAKbVBgCnzQYAqPUGAKn9BgCq9QYAq80GAKzVBgCt3QYArtUGAK/NBgCwtQYAsb0GALK1BgCzTQYAtFUGALVdBgC2VQYAt00GALh1BgC5fQYAunUGALtNBgC8VQYAvV0GAL5VBgC/TQYArH0/AK2lPwCurT8Ar6U/AKh9PwCpZT8Aqm0/AKtlPwCkHT8ApUU/AKZNPwCnRT8AoB0/AKEFPwCiDT8AowU/ALydPwC9pT8Avq0/AL+lPwC4nT8AuYU/ALqNPwC7hT8AtN0/ALWlPwC2rT8At6U/ALDdPwCxxT8Ass0/ALPFPwCMYToAjWE6AI5hOgCPYToAiEE6AIlBOgCKQToAi0E6AIRhOgCFYToAhmE6AIdhOgCAAToAgQE6AIIBOgCDAToAnF04AJ3lPwCe7T8An+U/AJhdOACZRTgAmk04AJtFOACUuTgAlWU4AJZtOACXZTgAkAE6AJEBOgCSAToAkwE6AMMIAIDbCACA4QgAgPMIAIB5BwCAJQkAgHwHAICEBwCAVwkAgKAHAIDOBwCAwAcAgMQGAIDcBACAewUAgM4FAIAJBgCALwYAgEYGAIBjBgCAegYAgJEGAIDXAwCA8AMAgAkEAIAiBACANQQAgEgEAIBbBACAbgQAgIEEAICUBACA+gQAgAkFAIAYBQCAJwUAgDYFAIBFBQCATgUAgFcFAIBgBQCAaQUAgJUFAICeBQCAYAgAgFwOAIBfDgCASzoAgLgJAAC5CQAArwoAgBgLAIBHOgCATzoAgOYMAIBTOgCAHw0AgIc3AID+MACArzcAgGcyAIDLKgCAxiwAgPktAICaMDUAKi0AgPUtAIDkLwCA3zMAgJ80AwBvNQCAnSQpAJzxAQCd8QEAnvEBAJ/xAQCnNgCA4zYAgBc3AIArOACAgzEAgA8yAIC7MgCAUzMAgG82AIBXOACAgzkAgO8qAICaLACAlzEAgN8yAICjNgCA0zkAgKEuAICHMgCAkzYAgCc3AIAYMACAyzUAgO82AIAULwCAEjEAgCcyAIArMwCANzgAgDYrAIDOKwCAOiwAgIAwAICPMQCA2zIAgP8zAICbNgCAszYAgNc3AID/OACAszkAgM85AIA7NACArYwCAHs0AIAzNQCAUzYAgIs4AIBbNwCAqRUBAK4tAIAwLwCAsAAdALEEHgCyABgAswwZALToGgC1AOQAthjlALcc5wC4AOAAuSThALog4gC7AOwAvDTtAL007gC+OO8AvwDoAOs0AICrNQCAvwgAgA8zAICkUAMApQAMAKZYDgCnAAgAqGAKAKmQCwCqaBUAq2wXAKwAEACtdBEArnwSAK8AHABDNACApzcAgPc4AICqLQCAfS4AgIcxAIA7MgCAbzIAgCM1AIBLNQCAtzgAgDYsAIC3NQCA2isAgNYrAICnNACANzUAgGs2AIC/OACAdzcAgBwwAIBnNwCA1yoAgFEuAICILwCAPzMAgL8zAIBaLACASzQAgEYrAIBsLwCAtyoAgIDlAwCB7QMAgi0vAIPhAwCE4QMAheEDAIbhAwCH4S8AiN0vAInZAwCKbS8AiykCAIw5AgCNOQIAjikCAI8lAgCQcQIAkXECAJJxAgCTcQIAlBECAJURAgCWEQIAlxECAJgxAgCZMQIAmjECAJsxAgCcEQIAnRECAJ4RAgCfEQIAoPECAKHxAgCi8QIAo/ECAKQRAgClEQIAphECAKcRAgCoMQIAqTECAKoxAgCrMQIArBECAK0RAgCuEQIArxECALBxAgCxcQIAsnECALNxAgC0TSIAtRUCALYdAgC3FQIAuC0CALk1AgC6PQIAuzUCALwRAgC9EQIAvhECAL8RAgCVuQ4AlLEOAJfJDgCWsQ4AkbkOAJCxDgCTqQ4AkrEOAJ31DgCcZQ0An/UOAJ71DgCZ+Q4AmPEOAJvpDgCa8Q4AhQUOAIQFDgCHyQ4AhgEOAIGhDQCAlSAAg6UNAIKlDQCN+Q4AjPEOAI/JDgCO8Q4AifkOAIjxDgCL6Q4AivEOALWxAQC0qQEAt7EBALa5AQCxvSAAsLUBALOxAQCyuQEAvfEBALzpAQC/8QEAvvkBALnxAQC4iQEAu/EBALr5AQClNQ4ApDUOAKc1DgCmNQ4AoTUOAKA1DgCjNQ4AojUOAK31AQCs9QEAr/UBAK71AQCp9QEAqPUBAKv1AQCq9QEA+zEAgJgwAIAfNQCAriwAgJotAIALNACAczYAgEs3AIDHMQCA8zEAgCwwAIArNgCATDAAgLszAIALKwCAjisAgNIrAIBjMQCACzUAgAM2AIBXNwCAazgAgEIsAID2LACAJC8AgLQwAICLMgCATzQAgKc4AICLOQCA3zkAgPc5AID2MACAszEAgPs3AIDwLgCAzC8AgOgvAIB4MACAezIAgMcyAIB3MwCAmzQAgD81AICjNQCA6zcAgHs2AIATOACAjzgAgPYrAIAiLACACi0AgLcyAIADNwCAEC8AgIAvAIBEMACAvzEAgOc0AIAzMwCAGysAgGYtAIC1LgCAjC8AgIBdAwCB8SgAgmkDAINpAwCEeQMAhXkDAIZpAwCHaQMAiFkDAIlZAwCK1SoAi60DAIy1AwCNvQMAjrUDAI+tAwCQ1QMAkd0DAJLVAwCT7QMAlPUDAJX9AwCW9QMAl+0DAJjVAwCZ3QMAmtUDAJutAwCctQMAnb0DAJ61AwCfrQMAoFUDAKGZAwCiUQMAo1EDAKRxAwClcQMApnEDAKdxAwCoUQMAqVEDAKp1DACrVQMArE0DAK21AQCuvQEAr7UBALDNAQCx1QEAst0BALPVAQC0zQEAtfUBALb9AQC39QEAuM0BALnVAQC63QEAu9UBALzNAQC9tQEAvr0BAL+9DwBPMwCAazMAgHs1AICbNQCAczgAgPM4AIADOQCAPzkAgDorAICPNACAXzgAgNs4AICkLwCA9yoAgF4rAIBVLgCAdS4AgKQwAIDTMgCA2zMAgIc2AIAnOACA5jAAgLM4AIAaLACAMjEAgD4xAIAfMgCAVzIAgFszAIC3MwCANzQAgBs1AIBLOQCA+C8AgMM4AIBOKwCAmS4AgD8yAIDvNwCAXC8AgKwvAIBGMQCAyzgAgP4rAIDmLACAhS4AgM8wAIAiMQCAbzEAgAMyAIBXMwCAyzMAgGc1AIAHNwCAEzcAgOc4AIBqLACAWzIAgOosAIDXMgCAezMAgJc2AIDPOACAl50HAJadBwCVnQcAlJ0HAJOdBwCSnQcAke06AJCZBwCfmQcAnpEHAJ2ZBwCcgQcAm9UKAJqdBwCZnQcAmJ0HAIchBwCGGQcAhREHAIS1JACDBQcAggUHAIEVBwCAFQcAj+EHAI4ZBwCNEQcAjBkHAIsBBwCKGQcAiREHAIgZBwC3/SkAtpUBALWFAQC0hQEAs5UBALKVAQCxZQcAsGUHAL+RAQC+iQEAvYEBALyJAQC7kQEAuqkBALmhAQC4qQEApxkHAKYRBwClGQcApAEHAKMZBwCiEQcAoRkHAKBhBwCvFQcArhUHAK0FBwCsBQcAqxUHAKoVBwCpuSQAqCEHALs5AIDjOQCAOjEAgDcyAIDTNQCA0zQAgPc0AIAnMwCArzIAgHM3AIATKwCAOzYAgAIsAIDyKwCAAC8AgCAwAIADNQCAQS4AgBMyAIDyMACA9zcAgLs4AIAcLwCAbisAgEItAICWLQCA4jAAgN4rAIAvMwCA8zMAgFc0AIBzNACAdzQAgIs0AIALOQCA+zQAgJ82AIBjNwCAFzgAgEM4AIBfOQCAYzkAgGc5AIDLOQCAOzgAgNc4AIA+KwCAYisAgHYrAIAyLACAPiwAgH4sAIAyLQCATi0AgFYtAICSLQCAni0AgIEuAICYLwCAwC8AgMgvAICR5BgA4C8AgIwwAICANQMAgT0DAII1AwCDTQMAhFUDAIVdAwCGVQMAh00DAIjJKwCJcQMAitErAIt1AwCMbQMAjVUDALwwAIDqMACAkCUDAGcxAICSIQMAKzIAgEcyAICVOQMAlikDAJcpAwCYGQMAmRkDAJrpAwCb6QMAnPkDAJ35AwCe4SsAdzIAgKARAwDLMgCAoh0DAOsyAIBfMwCApQ0DAKYFAwA/NACAYzQAgF80AICqCQMAqwkDAKwZAwCtGQMArgkDAK8JAwCweQMAsXkDALIJAwCzCQMAtBkDALUZAwC2CQMAtwkDALg5AwC5OQMAugkDALsJAwC85S0AvR0DAL4VAwC/DQMAvY0fALxhAgC/nR8Avp0fALmRHwC4xQIAu5EfALqZHwC1VR8AtFUfALdVHwC2VR8AsVUfALBVHwCzVR8AslUfAK0dHwCsHR8AZzQAgGs0AICpHR8AqB0fAKsNHwCqDR8ApSEfAKRZHwCn8QIApikfAKFBHwCgWR8Ao0EfAKJJHwCdpR8AnKUfAJ+hHwCeqR8AmYUfAJiFHwCbhR8AmoUfAJWpHwCUoR8AlwUGAJahHwCRqTgAkPkAAJO5HwCSARwAjWEBAIzNOACPgQAAjmkBAIldAQCIUQEAi0UBAIpNAQCFrQEAhKEBAIeVAQCGvQEAgQ0CAIANAgCDxQEAgsUBAIc0AICrNACAvzQAgNs0AIBHNQCATzUAgGM1AICLNQCA2zUAgA82AIB3NgCAHzcAgDc3AIBrNwCAbzcAgLM3AIC3NwCADzgAgOs4AIAvOQCARzkAgJAvAICm6gCA8zUAgL8qAIDKKwCAiisAgDIrAIByKwCAnisAgC4sAIBKLACAHi0AgC4tAIBKLQCApi0AgPEtAID9LQCAGS4AgCkuAIAYLwCAIC8AgFAvAIBwLwCAoC8AgLgvAICoLwCAvC8AgPwvAIBUMACAYDAAgGgwAICQMACAFjEAgCoxAIBrMgCAYzIAgJMyAIAjNACA7zIAgCMzAIBvMwCAizMAgK8zAICAmQEAgZkBAIKpAQCDqQEAhLkBAIW5AQCGqQEAh6kBAIiZAQCJ1RwAipUBAIvVHACM8QEAjfEBAI7xAQCP8QEAkJEBAJEtHACS3RQAk5kBAJSJAQCVhTIAlnkYAJcxGgCYvQEAmYUBAJoVHwCbiQEAnPUfAJ2dAQCelQEAn40BAKDxHAChcQEAonEBAKNxAQCkkQMApZEDAKbtHACnlQMAqK0DAKm1AwCqvQMAq7UDAKytAwCtuQEArpkDAK+ZAwCwbRgAse0DALLVAQCz4QMAtOEDALXhAwC24QMAt+EDALjRAQC5pQMAun0cALupAwC8xQEAvSEXAL6xAwC/xQEA0zMAgNczAID3MwCABzQAgBs0AIAXNACARzQAgMM0AIDzNACAKzUAgFs1AIA/NgCAZzYAgNs2AIAjNwCALzcAgE83AIBTNwCAXzcAgHs3AIDzNwCAIzgAgFs4AIB7OACAxzgAgB85AIA7OQCAmzkAgD3qAIA46gCAauoAgOcpAIAPKgCAEyoAgOzqAIAZ6wCAkesAgCc6AIA3OgCASggAgFUIAIBYCACATQgAgFEIAIBaCQCA9w4AgOgOAIDtDgCA/A4AgPIOAIBRDwCA0A8AgIcPAIA1DwCAYA8AgG0PAIB1DwCAow8AgMgPAIC+DwCAww8AgLAPAIC3DwCABA8AgIBNAQCBRQMAglkBAINZAQCESQEAhUkBAIZ5AQCHVQMAiKkeAIlBAQCKZQMAi0UBAIxhAwCNWQEAjsU7AI9NAQCQNQEAkT0BAJI1AQCTzQEAlNUBAJXdAQCW1QEAl80BAJj1AQCZ/QEACQ8AgA4PAIAbDwCAKA8AgDAPAIA4DwCAQg8AgEcPAIBMDwCAVg8AgFsPAIBjDwCAcA8AgHgPAIB9DwCAgg8AgIoPAICPDwCAmQ8AgJ4PAICmDwCAgzQAgKsPAIDLDwCAPQ8AgCAPAIBoDwCAlA8AgBMPAIDjFgCA7BYAgO8WAID1FgCA6RYAgPIWAIDmFgCAGRcAgBwXAID7FgCAnc0GAPgWAICfwQYA/hYAgAEXAIAKFwCABxcAgJSZBgCVmQYAlukGAJfpBgANFwCABBcAgBMXAICTiQYAEBcAgB8XAIAlFwCAKxcAgCgXAIAuFwCAMRcAgDoXAICEzQYAhdUGAIbZBgA0FwCAgO0GAIHVBgCC3QYAg9UGALwZBwBdFwCAvhUHACIXAIC4GQcAuRkHALoJBwC7CQcAtN0HALUlBwC2LQcARhcAgLDdBwCxxQcAss0HALPFBwCsNQYArT0GAK41BgCvpQcAqDkGAKl9QgCqNQYAqy0GAKQ5BgClOQYApgkGAKcJBgCgIQYAoQFCAKI5QwCjKQYAgKEGAIGhBgBDFwCAg6UGAIS9BgBOFwCAhqkGAIepBgCImQYAieUGAIrtBgCL5QYAjP0GAI3lBgCO7QYAj+UGAJCdBgCRmQYAkqkGAJOtBgCUsQYAlbUGAJa9BgCXuQYAmIUGAJmBBgCagQYAm4UGAJyZBgCdnQYAnpUGAJ+RBgCgbQYAoWkGAKJ5BgCjfQYApGEGAKVlBgCmbQYAp2kGAKhVBgCpUQYAqlEGAKtVBgCsSQYArU0GAK5FBgCvQQYAsD0GALE5BgCyyQEAs80BALTRAQC11QEAttEBALfVAQC46QEAue0BALr5AQC7/QEAvOEBAL3lAQC+7QEAv+kBAIEVAgCAEQIAgxECAIIVAgCFDQIAhAkCAIcpAgCGLQIAiRUCAIgRAgCLEQIAihUCAI1xAgCMdQIAj30CAI55AgCRBQIAkAECAJMBAgCSBQIAlRkCAJQdAgCXFQIAlhECAJktAgCYKQIAmzkCAJo9AgCdIQIAnCUCAJ8tAgCeKQIAodkCAKDdAgCj0QIAotUCAKVpUwKkbVMCp8UCAKbBAgCp/QIAqPkCAKvFAgCqwQIArd0CAKzZAgCvPQIArjkCALEpUwKwLVMCVBcAgEAXAIBRFwCAWhcAgH8WAIDnDwCANxAAgBQQAIAoEACAIxAAgC0QAIAyEACAGRAAgFcXAICEnA4A7A8AgPEPAIAFEACAHhAAgF4QAIBjEACAbxAAgIUQAICUEACAmRAAgKQQAIC+EACA0RAAgNmIUAL1EACAJxEAgCwRAIA0EQCAQxEAgFIRAIBXEQCAXxEAgIIRAICpEQCAtREAgNURAIDaEQCA3xEAgBkSAIAsEgCAOBIAgFASAIDKEgCAIBMAgDkTAIA+EwCAURMAgGITAIB0EwCAeRMAgKATAICoEwCAvRMAgOQTAIDpEwCAQxQAgEgUAIBNFACAWRQAgGUUAIBqFACAchQAgH4UAICYFACAnRQAgNlcUAKlFACAqhQAgK8UAIC0FACAuRQAgL4UAIDRFACAn8kOAJ7NDgCdwV0CnBkNAJsFDQCaHQ0AmRENAJixDACXjQwAlqkMAJWlDACUoQwAk70MANYUAIDyFACADBUAgCYVAIAyFQCAShUAgE8VAIBcFQCAfRUAgKAVAIC6FQCAxhUAgMsVAIDTFQCA9BUAgA4WAIAdFgCAOhYAgD8WAIC/fQ4AvnkOAL11DgC8cQ4Au2kOALptDgC5YQ4AuGkOALdVDgC2UQ4AtVkOALRdDgCzXQ4AslkOALFRDgCwVQ4AryUOAK4hDgCtKQ4ArC0OAKsNDgCqCQ4AqQEOAKgFDgCnNQ4ApjEOAKU9DgCkOQ4AoyEOAKIlDgChNQ4AoDEOAIAFDgCBDQ4AggUOAIP1DwCEAQ4AhQEOAIYBDgCHAQ4AiAEOAIkBDgCKAQ4AiwEOAIwBDgCNAQ4AjgUOAI99DgCQBQ4AkQ0OAJIFDgCTHQ4AlAUOAJUNDgCWBQ4Alz0OAJgFDgCZDQ4AmgUOAJsdDgCcBQ4AnQ0OAJ4FDgCf/Q4AoAUOAKENDgCiBQ4Aox0OAKQFDgClDQ4ApgUOAKc9DgCoBQ4AqQ0OAKoFDgCrHQ4ArAUOAK0NDgCuBQ4Ar30OALAFDgCxDQ4AsgUOALMdDgC0BQ4AtQ0OALYFDgC3OQ4AuAkOALkJDgC6GQ4AuxkOALwJDgC9CQ4Avs0BAL/FAQCAPQIAgUUCAIJNAgCDRQIAhF0CAIVFAgCGTQIAh0UCAIh9AgCJRQIAik0CAItFAgCMXQIAjUUCAI5NAgCPRQIAkD0CAJFBAQCSQQEAk0EBAJRBAQCVQQEAlkEBAJdBAQCYQQEAmUEBAJpBAQCbQQEAnEEBAJ1BAQCeQQEAn0EBAKDBAQChwQEAosEBAKPBAQCkwQEApcEBAKaVDQCnxQEAqFkMAKm1DQCq9QEAq80BAKyRDQCt0QEArp0NAK+VDQCwqQEAsakBALL1DQCzvQEAtJENALWRDQC2rQEAt6UBALitDQC5mQEAurkNALu5DQC8OQ0AvTkNAL4hDQC/IQ0Ap1X8AEcWAIBMFgCAXxYAgGQWAICKFgCAlhYAgKIWAICxFgCAzhYAgNMWAID0EQCABRIAgIIWAICBAACAiwAAgJUAAICfAACAqQAAgLMAAID7DwCAABAAgAoQAIB7EACAgBAAgIoQAIDrEACA8BAAgB0RAIA5EQCAPhEAgEgRAIBXFQCAExYAgBgWAIAwFgCApxYAgKwWAIDEFgCA9g8AgA8QAICPEACAIhEAgN0SAIBFFQCANRYAgGkWAIDJFgCATREAgGoSAIClEgCAuBIAgBcUAIAjFACALxQAgJMTAICYEwCA1xMAgNwTAIADFACACBQAgG8SAIB0EgCAvRIAgIL5CwCD+QsAgO0LAIH5CwCGWQQAh1kEAIQtBACFWQQAiqUHAIutBwCIqQcAiXEEAI5JBACPSQQAjE0EAI2xBwCS1QcAk2UHAJB9BwCR3QcAlnkHAJdRCwCUwQcAlXkHAJptCwCbxQcAmGELAJnxBwCebQsAn1ULAJxtCwCdZQsAorELAKOxCwCgLQcAoaELAKbdCwCnzQsApKULAKU1BwCqxQsAq80LAKj1CwCpzQsArskLAK/JCwCsyQsArckLALJtBwCzTQsAsEkLALFJCwC2RQsAt00LALRVCwC1TQsAukkLALtJCwC4dQsAuUkLAL5JCwC/SQsAvEkLAL1JCwCAwQoAgcEKAILZCgCD2QoAhPkKAIX5CgCG6QoAh+kKAIjZCgCJHQUAihUFAIttBQCMdQUAjYUGAI5pBQCPaQUAkBkFAJEZBQCSIQUAkyEFAJQhBQCVIQUAlu0GAJdZBgCYaQYAmd0GAJp9BgCbdQYAnG0GAJ1VBgCexQYAn3EKAKAhBgChpQoAoi0GAKOxCgCkOQYApdkKAKZZBgCnHQoAqGUGAKltBgCqZQYAq1kKAKxJCgCt8QUArs0FAK8JBgCw4QYAsXkGALIZBgCzGQYAtAkGALUJBgC2OQYAtzkGALgJBgC5CQYAuhkGALsZBgC8CQYAvQkGAL75AwC/+QMAwhIAgMgRAIC8TQEAvUkBALqxCQC7sQkAuDUBALktAQC2XQkAtw0BALRdCQC1VQkAsv0FALOVCQCw8QUAsfkFAK5tAQCvdQEArAkBAK1lAQCqaQEAq2kBAKiRBQCpaQEApk0BAKdVAQCkTQEApUUBAKJtAQCjVQEAoG0BAKFlAQCejQEAn5UBAJyNAQCdhQEAmpEAAJuRAACYYQUAmWEFAJZRBQCXUQUAlEEFAJVBBQCSUQUAk1EFAJD5AQCRYQUAjvkBAI/5AQCMAQUAjfkBAIr9AQCL5QEAiP0BAIn1AQCG/QEAh8UBAIT9AQCF9QEAgv0BAIPlAQCA/QEAgfUBAPBQ/wDNEQCAnBEAgKERAIDkEQCA6REAgCwTAIAxEwCAZxMAgGwTAIB8EgCAiBIAgJsSAICgEgCASxIAgOISAIBdEwCAURAAgKkQAIDDEACAyhAAgNYQAID6EACAAREAgAgRAICHEQCAwREAgLoRAIAxEgCAHhIAgCUSAIBcEgCAVRIAgGMSAIDPEgCAJRMAgI0SAICBEgCAqhIAgLESAIBDEwCAVhMAgH4TAICFEwCAjBMAgK0TAIDCEwCAyRMAgO4TAID8EwCA9RMAgFIUAICDFACAihQAgBEVAIAfFQCAGBUAgPcUAIArFQCANxUAgIIVAICJFQCAmRUAgGEVAIBvFQCApRUAgKwVAIBoFQCAURYAgFgWAID5FQCAABYAgN8VAIDmFQCAKRYAgCIWAIC2FgCAdBAAgLcQAICwEACAlRn/AJQR/wCXKf8AlhH/AJEd/wCQHf8Akwn/AJIR/wCdFf8AnBX/AJ8V/wCeFf8AmRX/AJgR/wCbFf8AmhX/AKUJ/wCkDf8Apxn/AKYB/wChEf8AoOn/AKMd/wCiGf8ArT3/AKw5/wCvDf8Arg3/AKkl/wCoJf8AqyH/AKol/wC1df8AtHX/ALdx/wC2df8AsXn/ALBx/wCzdf8AsnX/AL0t/wC8Kf8Avz3/AL49/wC5Mf8AuEn/ALsx/wC6Of8AgNn+AIHZ/gCC6f4Ag+n+AIT1/gCF/f4AhvH+AIfx/gCIzf4AidX+AIrd/gCL1f4AjM3+AI01AQCOPQEAjzUBAOQQAIDdEACAkkUBAJNdAQCURQEAlU0BAJZFAQCXfQEAmEEBAJlBAQCaQQEAm0EBAJxBAQCdRQEAnk0BAJ9FAQCgvQEAocUAAKLNAACjjQMApJUDAKWdAwCmlQMAp40DAKi5AwCpuQMAqokDAKuJAwCsmQMArZkDAK6JAwCviQMAsPkDALH5AwCyiQMAs4kDALQB/gC1Df4AtpEDALeRAwC4sQMAubEDALqxAwC7sQMAvKkDAL2lAwC+mQMAv5kDABYRAIAPEQCAlREAgGQRAICOEQCAkQQHAD0SAIDWEgCAlBIAgEoTAIAFFQCAPhUAgJsWAICPFgCAvRYAgL8VAICRFACABxYAgNATAIDKFACA2BUAgLMVAID+FACAwxQAgGsRAICuEQCAdhUAgF4UAIBoEACARBIAgO0VAIAZEwCAdxQAgEgQAIA/EACAkBUAgOcSAID8EQCAtBMAgHEWAIDwEgCA9xIAgHIRAIAKEgCApgMAgBMjAIAXIwCAoAYAgMcAAIC1BgCAsSMAgLUjAIC/IQCAuyEAgOYHAIB+CQCAggkAgEcjAICtIwCAOyMAgD8jAIAjIwCAJyMAgCsjAICAaQEAgWkBAIJ5AQCDUQEAhKX8AIWZ/ACGbQEAh2UBAC8jAIAzIwCANyMAgN4HAIDiBwCA0QAAgNcAAICiAwCAqAMAgOAHAIDTAACA1QAAgL0GAIB5AACADRQAgH0AAICHAACAkQAAgBIUAICbAACAHhQAgKUAAIAqFACArwAAgDYUAIC5AACAOxQAgNUPAIBbEACAnhAAgKEQAIAxEQCAXBEAgKYRAIDSEQCA7hEAgPERAID5EQCAExIAgBYSAICwvQEAsUUBALJNAQCzRQEAtF0BALVFAQC2TQEAt0UBALh9AQC5RQEAuk0BALtFAQC8XQEAeRIAgMcSAIA2EwCAqAYAgLMGAIDMlLQAPBAAgHETAICdEwCApRMAgOETAIBAFACAbxQAgKIUAIDbFACAVBUAgNAVAIBEFgCAbhYAgJiNBgCZgbEAhxYAgN4UAIDjFACA6BQAgO0UAIDPAACAkNEGAJHRBgCS0QYAk9EGAJSpkgKVtQYAlr0GAJe1BgDZAACAswMAgOQHAICACQCAASMAgAUjAICHKQCAOyQAgHQkAICTJACApSQAgMokAIDJKACA5SgAgPcoAICOJgCAuCEGALkhBgC6IQYAuyEGALwhBgC9IQYAviEGAL95uACwIbEAsTUGALI9BgCzNQYAtCkGALVFsgC2TbIAtyEGAIC5uQCB+QcAgikGAIMpBgCEOQYAiykAgG8pAICHMQYAiBEGAIn5sACK9bAAi/GwAIztsACN7QcAjuEHAI/lBwCQ8QcAkfEHAJL1sAAvJACAlImTApXpBwCWnQcAl50HAJixBwCZ1bMAmt2zAJuxBwCckQcAnZEHAJ6RBwCfSQYAoLkGAKG5BgCiIbMAo80GAKSRAQClkQEApkGyADMkAICo5QYAqe0GAKrlBgCr/QYAhAkAgIcJAICQCQCAjQkAgLCVBgCxnQYAspUGAIoJAIC0sQYA8iEAgLa9BgC3tQYAuI0GALmVBgC6nQYAu5UGALyNBgC9dQYAvn0GAL91BgCCkaMCg5GjAoCFBQCBnaMChrmjAoeNowKEjaMChbWjAoqVowKLkaMCiLGjAomZowKOPQIAj6UFAIyNowKNMQIAktEFAJPRBQCQ2QUAkd0FAJbJBQCXzQUAlM0FAJXFBQCa/QUAm/kFAJjxBQCZ8QUAntEFAJ/VBQCc5QUAnd0FAKIlBQCjIQUAoCkFAKEpBQCmOQUApz0FAKQ9BQClNQUAqg0FAKsVBQCoAQUAqQEFAK4FBQCvDQUArAkFAK0JBQCyfQUAs3kFALBxBQCxcQUAtiUFALchBQC0ZQUAtSkFALoZBQC7HQUAuB0FALkVBQC+DQUAvwkFALwBBQC9AQUAhPwCAPUiAID6IQCAAiIAgPkiAID9IgCAkZQCAKQQDQCffBwAnmQIAJ08CADZzKACzMCFAsz8hQL2IQCA9egAgJmY1gD66ACAm/jWAP/oAICxNN0ABOkAgNmUoAKynPsApwUAAAnpAIDZhKAChywDAJE4AgAO6QCAjAwCAI20AgDwtAIAgqACABPpAIC++QAAkpACAJcdAAC5+AMAhAgCAPBAAgCRfAIAHekAgBjpAICljNYAIukAgI3IAgDwJAIAsGkAAJK4AgC4tAMAudgMAJa0AgCW/AIArz0AAJFUAgCvCCEAJ+kAgLpFAACRTAIALOkAgL1BAACWuAIArw0AAIsZDACKHQwAiREMAIgZDACPDQwAjgkMAI0FDACMBQwAgzEMAII1DACBOQwAgNELAIclDACGIQwAhS0MAIQtDACbaQwAmm0MAJlhDACYaQwAn30MAJ55DACddQwAnHUMAJNBDACSRQwAkUkMAJBxDACXVQwAllEMAJVdDACUXQwAq7kMAKq9DACpsQwAqLkMAK9J/gCuqQwAraUMAKylDACjkQwAopUMAKGZDACggQwAp4UMAKaBDACljQwApI0MALuZDAC6kQwAuZ0MALidDAC/2YgC8HwNALUcAgC8hQwAs6kMALKtDACxoQwAsDX+ALehDAC2uQwAtbUMALS1DACBwQsA8DgBAIM9CgCCUQ0AhV0KAIRdCgCHmQ0AhiUKAImVvwCIlb8Ai4G/AIoRCgCNcQoAjIULAI+9DQCOgbwAkbWdApCxvACTqZ0CkpGdApWtvACUrbwAl9W/AJbVvwCZxb8AmMW/AJuxnwKa0QsAnam+AJx1DQCfvQsAnnkNAKEBvwCg1QoAo5WxAKKFvQClgb0ApJm9AKeBvQCmqbEAqYm9AKiFvQCrmb0Aqom9AK3xvQCsjb0Ar+m9AK71vQCxIb8AsJ29ALNJyQCyrb0AtaG9ALS9vQC3ob0AtkmxALlpyQC4TbEAu8UKALrNsQC9wQoAvLEKAL8hCwC+dQ0AgP2+AIHVngKCZb4Ag8W+AISRvgCFnb4AhqW+AIeNvgCIrZECieW+AIopkgKLtb4AjBGSAo2VvgCOLbIAj8WeApDpvgCRsbUAkkGSApPxnwKU1b4AleW+AJbhvgCXTZICmGWSApl9kgKaub4Am7EIAJz9DgCdlQgAkRQDAJ/tDgCgFQ4AoT0IAKJ1CACjrQkApCUIAKUNDgCmBdYApwEOAKgBDgCpAQ4AqgEOAKsBDgCsAQ4ArQEOAK4BDgCvAQ4AsNUPALGV0wCyndMAs4XUALSJ0wC1idMAttnWALfZ1gC46dYAuenWALr51gC7+dYAvOnWAL3p1gC+udYAv7nWAIBJ1wCBSdcAglnXAINZ1wCESdcAhXXSAIZ90gCHddIAiE3SAIlV0gCKddcAi63XAIy11wCNvdcAjrXXAI9J0gCQOdIAkTnSAJLJ0gCTydIAlNnSAJXZ0gCWydIAl8nSAJj50gCZ+dIAmsnSAJvF0gCc4dcAnW0OAJ41DgCf4Q4AoNHZAKHB2wCiwdsAo93bAKTF2wClzdsApsXbAKf92wCoJdsAqWXbAKrN2wCrxdsArN3bAK3B2wCuwdsAr3nbALAJ2wCxCdsAshnbALMZ2wC0CdsAtbXYALbB3gC33d8AuOXfALn13wC6/d8Au63fALy13wC9pd8Avq3fAKQgBQCPFd4AjhXeAI0F3gCMBd4AixXeAIoV3gD+IQCAsCUAAIfd3ACG3dwAhd3cAITd3ACD2dwAgtXcADHpAIA26QCAP+kAgEjpAIBV6QCAnAXeAJtt3gCabd4AYukAgG/pAICXVd4AllXeAJVB3gCUSd4Ak1HeAJJp3gB86QCAiekAgKzpAICukQsArZkLAKyBCwCriQAAqp0LAKmhCwComQsAkukAgKZxCQClZQkApBEJAKNlmAKiDQkAoQ0JAKANCQC7vQAAtekAgL3BjAKf6QCAvx0AAL4RAAC9eQAAvF0AAIAFAADC6QCA0AoAgLMMAIBkDQCAag0AgHANAIB8DQCAhWgCAH8NAICH6AMAhiADAJ4tAQCfVQEAgg0AgIUNAICIDQCAlw0AgJ0NAICgDQCAow0AgCYiAIDNDQCA3A0AgJVAHQCURB4Al0gbAJYAGACRLAAAkFQBAJNYHwCSABwAnWASAJxkEwCfiBcAnmwRAJlwGQCYdBoAmwAQAJoAFAACDgCABQ4AgBQOAIAXDgCAIw4AgB4iAIA4DgCAOw4AgLgALAC5rC8AuqguAN0WAIAWFwCA4BYAgLoDAIC3AwCAygMAgO0EAICMBQCA3wUAgBoGAIBABgCAVwYAgHQGAICiuQEAo7kBAKCtAQChpQEAiwYAgDgBAICkgQEAPAEAgICBuACBDboAghW2AIMBugCEAboAhSG2AIYBugCHPboAiAm6AIkBugCKGboAixW6AIxxugCNfboAjmm6AI9lugCQobgAkSW6AJLJzgCTJboAlCG6AJXBtgCWIboAl/W2AJjpzgCZRbYAmrmaApsBugCcuboAnfW6AJ7xugCfwboAoBG6AKEJlQKiSboAo42WAqQJugCltZYCpjm6AKeJtgCoWZoCqQ26AKpdsQCrpZYCrA2bAq0xugCuCboArwW6ALDRlgKxwZYCstGWArMdugC0UbgAtd26ALbFtgC30boAuPG6ALnRtgC68boAu826ALzZugC90boAvsm6AL/FugCfCZcCnvGwAJ2huwCc9ZsCmwW3AJq1uwCZOZcCmIW7AJchlwKW5bsAzMS+AJS9uwCTjbsAkr27AJG5uwCQ9bsAjy27AI6VmwKNabcAjMXPAIv5twCKLbsAic23AIgtuwCHCbsAhuXPAIUJuwCEjbkAgym7AIIluwCBMbsAgD27AL8ptwC+/bsAvR23ALz9uwC7+bsAuhXPALn5uwC4fbkAt/m7ALb1uwC14bsAtO27ALOJuwCyhbsAsZ27ALCVuwCv4bsArt27AK39twCs3bsAq927AKrJtwCp0bsAqF25AKcxuwCm/ZcCpe2XAqT9lwKjSbsAokW7AKF9uwCgQZoCgJmkAIEliAKCqaQAgxmoAEABAICFvaQAhu2vAIcViAKInYUCiaGkAIqZpACLlaQAjCGIAo0xiAKOIYgCj+2kAJDBpgCRTaQAklWoAJNBpACUQaQAlWGoAJZBpACXfaQAmEmkAJlBpACaWaQAm1WkAJwxpACdPaQAnimkAJ8lpACgYaYAoeWkAKIJ0ACj5aQApOGkAKUBqACm4aQApzWoAKgp0ACphagAqnmEAqvBpACseaQArTWkAK4xpACvAaQAsFGkALFJiwKyCaQAs82IArRJpAC19YgCtnmkALfJqAC4GYQCuU2kALodrwC75YgCvE2FAr1xpAC+SaQAv0WkAIARiQKBAYkCghGJAoPdpQCEkacAhR2lAEQBAICHEaUAiDGlAIkRqQCKMaUAkQ0AgEgBAICNEaUAjgmlAI8FpQCQAaUAkQ2lAJIZpQCTFaUAlLGnAEwBAICW2dEAlzWlAJgRpQCZ8akAmhGlAJvFqQCc+dEAUAEAgJ6phQKfEaUAoEmlAKEFpQCiAaUAozGlAKQBpQClGYoCplmlAKediQKoOaUAqYWJAqoJpQCruakArEmFAq0dpQCuTa4Ar7WJArB9hAKxQaUAsnmlALN1pQC0wYkCtdGJArbBiQK3DaUAuGGnALntpQBUAQCAu+GlALzhpQC9wakAvuGlAFgBAICAKaYAgSGmAII5pgCDNaYAhFGmAFwBAICGSaYAzEyHAmABAIBkAQCAiqnSAItFpgCMQaYAjaGqAI5BpgCPlaoAkMnSAGgBAICSmYYCkyGmAJSZpgCV1aYAltGmAJfhpgCY8aYAmemJApqppgCbbYoCnOmmAJ1VigKe2aYAn2mqAKB5hgKhLaYAon2tAKOFigKkLYcCpRGmAKYppgCnJaYAqLGKAqmhigKqsYoCq32mAKwxpACtvaYArqWqAK+xpgCw0aYAsfGqALLRpgCz7aYAtPmmALXxpgC26aYAt+WmALihpgC5raYAurmmALu1pgC8EaQAvZWmAL550gC/laYAl7mnAJa1pwCVjacAlLGGApMZiwKS4awAkbGnAJDlhwKfLacAnjmrAGwBAICcraUAm+GnAJotiwKZPYsCmC2LAof9pwCGzacAhcmnAISFpwCDPacAgoWHAoF5qwCA1dMAj3WrAI7FpwCNSYsCjPWnAItxiwKKtacAifWIAojtpwC37acAtlWHArWpqwC0BdMAszmrALLtpwCxDasAsO2nAL+hiwK+ZacAvSWIAnABAIC7DacAdAEAgLk5pwC4dacAeAEAgKb1pwCl7acAfAEAgIABAICizacAhAEAgIgBAICviacArmXTAIwBAICsDaUAq6mnAKqlpwCpsacAkAEAgICRoACBiY8CgsmgAIMNjAKEiaAAhTWMAoa5oACHCawAiNmAAomNoACK3asAiyWMAoyNgQKNsaAAjomgAI+FoACQUYwCkUGMApJRjAKTnaAAlNGiAJVdoACWRawAl1GgAJhxoACZUawAmnGgAJtNoACcWaAAnVGgAJ5JoACfRaAAoMGgAKHNoACi2aAAo9WgAKRxogCl9aAAphnUAKf1oACo0aAAqTGsAKrRoACrBawArDnUAK2VrACuaYACr9GgALAJoACxRaAAskGgALNxoAC0QaAAtVmPArYZoAC33YwCuHmgALnFjAK6SaAAu/msALwJgAK9XaAAvg2rAL/1jAKAvYACgYGhAIK5oQCDtaEAhAGNAoURjQKGAY0Ch82hAIihowCJLaEAijWtAIshoQCMIaEAjQGtAI4hoQCPHaEAkGmhAJFhoQCSeaEAk3WhAJQRoQCVHaEAlgmhAJcFoQCYgaMAmQWhAJrp1QCbBaEAnAGhAJ3hrQCeAaEAn9WtAKAJ1QChpa0AolmBAqPhoQCkWaEApRWhAKYRoQCnIaEAqDGhAKkpjgKqaaEAq62NAqwpoQCtlY0CrhmhAK+prQCwOYECsW2hALI9qgCzxY0CtG2AArVRoQC2aaEAt2WhALjxjQK54Y0CuvGNArs9oQC8caMAvf2hAL7lrQC/8aEAgBGiAIExrgCCEaIAgy2iAIQ5ogCFMaIAhimiAIclogCIYaIAiW2iAIp5ogCLdaIAjNGgAI1VogCOudYAj1WiAJAxogCR0a4AkjGiAJPlrgCU2dYAlXWuAJaJggKXMaIAmKmiAJnlogCa4aIAm9GiAJzhogCd+Y0CnrmiAJ99jgKgGaIAoaWOAqIpogCjma4ApGmCAqU9ogCmbakAp5WOAqgdgwKpIaIAqhmiAKsVogCsoY4CrbGOAq6hjgKvbaIAsEGgALHNogCy1a4As8GiALTBogC14a4AtsGiALf9ogC4yaIAucGiALrZogC71aIAvLGiAL29ogC+qaIAv6WiAJMVrwCSpaMAkSmPApCVowCXGY8CluGoAJWxowCU5YMCm5mjAJqVowCZraMAmJGCAp/howCeLY8CnT2PApwtjwKD6a8Agj2jAIHdrwCAPaMAhz2jAIaFgwKFea8AhNXXAIvdowCK7aMAiemjAIilowCPcY8CjrWjAI31jAKM7aMAs+mjALIF1wCx6aMAsG2hALc5rwC27aMAtQ2vALTtowC7zaMAunWDArmJrwC4JdcAvw2jAL49owC9OaMAvHWjAKPNowCi2a8AocGjAKBNoQCn8aMAps2jAKXtrwCkzaMAq9mjAKrVowCpzaMAqMWjAK+powCupaMArbGjAKy9owC43YwCueWMArrxrQC78a0AvJmuAL2ZrgC+ua4Av7muALDZrQCx2a0AsqGuALOhrgC0ka4AtZGuALbNrQC3yYwCqMmuAKnJrgCq6a4Aq+muAKylrQCtoYwCroWMAq+9jAKgwa4AocGuAKKdrQCjmYwCpK2MAqWVjAKmga0Ap4GtAJh1rQCZcYwCmlWMApttjAKcaa0AnWmtAJ4RrgCfEa4AkH2MApFFjAKSUa0Ak1GtAJQ5rgCVOa4AlhmuAJcZrgCIwbwCiTm5AopRFQCLURUAYQ0AgJQBAICOLa0AjymMAoDdqwCBdRUAgqkBAIN5FQCE+bwChQW5Aob5ugKHCbkCbQ0AgJgBAIBzDQCAnAEAgHkNAIBLIwCAiw0AgNEGAIC+DQCAywcAgNANAIAPBwCACA4AgJcHAIAaDgCAnQcAgCYOAICGYAIAhYABAIRMOQCABwCAyAcAgE8HAIBSBwCAXQcAgJANAADhBgCAFSQAgOglAIA1LgCAiXg/AIgAPAC6LACA1i0AgD83AIAHKwCA0zAAgL8yAIAOLACAYC8AgKYrAICsMACA+isAgCc1AICbNwCAui0AgPIsAIBzMgCAEDAAgDwwAIAbOACAMDAAgAgwAIB/NACAuzQAgN4sAICvoAEAUzIAgKczAIASLACAPi0AgFM4AICPIwCAUyMAgMwouALNRLgCzIi4As38uALMkIQCzTSFAsywhQLNVIcCzBCCAs1QggLMoIICzYyCAswwgQLNJIECzBiBAs2EgQJdIwCAcSMAgJkjAIB7IwCAoyMAgGcjAICFIwCAZC8AgKo5AwCrOQMAziwAgNsqAIDMWLkCzaS6AqwZAwDTKgCAsHWvALFxjgKyVY4Cs22OArRprwC1aa8AthGsALcRrAC4AawAuQGsAOMqAIDP6QCALisAgEIrAIBKKwCAUisAgIJB2gCDra4AgCmsAIGtrgCGqa4Ah32iAISprgCFSaIAis0DAIs5xACIYdoAic2iAI6hAwCPoQMAjM0DAI3BAwCfua8AnrWvAJ2NrwCcsY4CmxmDAprhpACZsa8AmOWPApc1owCWha8AlQmDApS1rwCTMYMCkvWvAJG1gAKQra8Aj/2vAI7NrwCNya8AjIWvAIs9rwCKhY8CiXmjAIjV2wCHyaMAhh2vAIX9owCEHa8AgxmvAIL12wCBGa8AgJ2tAL+xFgC+qRYAvaEWALzBvgK7tRYAuom5ArmBuQK4sQoAt22sALa9AgC1iRYAtLEWALOpFgCysRYAsUUXALCtAgCv2bkCrs0CAK0xFwCszQIAqyUXAKodrACpKRcAqA0DAMyIgQLNSIECpQUXAKQFFwCjIa8Aou2DAqH9gwKg7YMCgAm4AoE9EQCCIQUAg/29AoT1qACFFb4ChiURAIfxiAKIoREAiaERAIoZBQCL0b0CjDG4Ao09vgKOsREAj7ERAJB5BQCRsb0CknWvAJPdEQCUEQUAlcERAJZRuAKXrb0CmGG+ApmRvQKaabgCm5G9ApyhBACdhRAAnrGrAJ+JEACggQUAoX0QAKKBBQCjlb4CpIEFAKVpEACmnREAp4URAKi9EQCphREAqrEFAKthqwCsnQ0Ara2+Aq6lvgKvmREAsI25ArHtEQCy5REAs/0RALT5pAC14REAtvkFALcxvQK4ya8AucmvALrhuAK71REAvNkFAL0FvQK+HagAv/2+AoA9EACB6YkCgokQAIOJEACEIQQAhem8AoYZuQKHFb8CiKkQAImpEACKEQQAi9m8AowNrgCNpRAAjnkEAI+pEACQSbkCkbW8ApJJvwKTubwClFG5ApWpvAKWiQUAl60RAJipqgCZkREAmmkEAJuVEQCceQQAnW2/Ap5pBACfgREAoIUQAKGdEACilRAAo60QAKSJBAClWaoAprUMAKeFvwKovb8CqYEQAKrluAKrhRAArJ0QAK2FEACukaUAr4kQALDhBACxKbwCsuGuALPhrgC02bkCte0QALbxBAC3LbwCuAWpALnlvwK61RAAuwGJArxxEAC9cRAAvskEAL8BvAKAAboCgQ28AoKBEwCDgRMAhCkHAIXhvwKGJa0Ah40TAIhhBwCJsRMAiiG6AovdvwKMMbwCjcG/Ao45ugKPwb8CkJEGAJG1EgCSgakAk7kSAJRRBwCVrRIAllEHAJdFvAKYcQcAmZkSAJptEwCbdRMAnG0TAJ1VEwCeYQcAn7GpAKCtDwChnbwCopW8AqOpEwCk3bsCpb0TAKa1EwCnrRMAqImmAKmREwCqiQcAq0G/AqyZrQCtma0ArrG6Aq+FEwCw6QcAsTW/ArItqgCzzbwCtO0TALU5igK2WRMAt1kTALjRBwC5Gb8Cuum6ArvlvAK8eRMAvXkTAL7BBwC/Cb8CngG9Ap/xvgKcAbsCnf2+AppRBgCbgRIAmCWsAJmNEgCWGQYAl9G+ApShEgCVoRIAkjG7ApM9vQKQCQYAkcG+Ao7BEgCPwRIAjHUSAI2hiwKKtasAi1W9AohxBgCJrb4Chmm7AoddEgCEQawAhUGsAIJRBgCDmb4CgFGnAIFJEgC+qawAv6msALypBgC9Yb4CurmnALuhEgC4tRIAua0SALbtugK3jRIAtLW9ArWJEgCynQ4As629ArChBgCxcagArt0SAK/lEgCszRIArdUSAKrBBgCrKRMAqNEGAKnFvQKm4QYApx0TAKQhqAClGRMAoiEHAKMFEwCg+bsCoQG+AoKhIwBWKwCAWisAgJMpAIDj6QCAh7EjAHorAIB+KwCAmisAgIsJJADU6QCAiWUkAI6NIwCPLSQAlykAgI0JJACSZSQAhisAgN7pAICRtSMAtisAgJf9IwCUrSMAvisAgBcrAICbeSQAxisAgJmRIwC56wCAn8EtAO4rAICdKdQAoiEjAJ8pAIAGLACAoR0jAAosAICnMSMApKEkABYsAICqiSQAoykAgKi5JACp5SQArg0jAK+tJACsiSQArYkkALLlJABOLACAsOkkALE1IwC2TSMAt30jALQtIwC1RSMAuvUkALv5JABSLACAuREjAL5BLQB6LACAvFUtAIYsAICADSUAgZUiAIKpIgCDoSIAhCklAIUpJQCGoSIAh7kiAIgxJQCJbSUAliwAgIsBJQCMASUAjQElAI6FIgCPJSUAkGElAJG9IgCSbSUAk/kiAJSlIgCVzSIAlsUiAJf1IgCY0SIAmZkiAJp9JQCbcSUAniwAgKIsAIDy6QCAozIAgLYsAIChFSIAoikiAKMhIgC+LACApaklAKYhIgCnOSIAqLElAKntJQD36QCAq4ElAKyBJQCtgSUArgUiAK+lJQCw4SUAsT0iALLtJQCzeSIAtCUiALVNIgC2RSIAt3UiALhRIgC5GSIAuv0lALvxJQDKLACA0iwAgNosAIACLQCAgLkvAIG5LwCCyS8Ag8kvAITZLwCF2S8AhskvAIfJLwCI+S8AifkvAIrZLwDuLACA+iwAgP4sAIAGLQCADi0AgJC1LwCRuS8AklkyAJNVMwCUYTMAlQEzAJYtMwCXtTMAmJ0zAJlxMACaSTAAm0kwAJw1MACdXTEAntUxAJ/JMQCgQTEAoUkxAKJZMQCjVTEApE0xAKV9MQCmZTEAp0k6AKilOwCpqTsAqr07AKuxmgCs0ZYArak7AK6dOwASLQCAsEGUALHNlgCy1ZoAs8GWALTBlgC14ZoAtsGWALf9lgC4yZYAucGWALrZlgC71ZYAvLGWAL29lgC+qZYAv6WWAMUAAACpTScAqiEnAKshJwCsIScArSEnAK6lIACvBScAGi0AgKG1IACiiSAAIi0AgKQJJwAmLQCANi0AgKeZIAA6LQCAubkgAEYtAIC7UScAai0AgG4tAIBSLQCAWi0AgLBBJwCxnSAAsk0nALYtAIC0hSAAte0gALblIAC31SAAiJEnAInNJwCKoScAi6EnAIyhJwCNoScAjiUgAI+FJwCArScAgTUgAIIJIACDASAAfi0AgIWJJwCGASAAhxkgAJhxIACZOSAAmt0nAJvRJwCcfS4AnYHXAJ5pLgCfaS4AkMEnAJEdIACSzScAk1kgAJQFIACVbSAAlmUgAJdVIACA+T0Agfk9AIJJPgCDST4AhFk+AIVZPgCGST4Ah0k+AIh5PgCCLQCAhi0AgHotAICOLQCAii0AgKItAID86QCAkB0+AJEtPgCoSBUAvi0AgMItAIDKLQCA3i0AgAEuAICiAAwAo2QPAKBoAwChbAAApgAUAKd0FwCkAAgApXwLAAHqAIAG6gCADS4AgBEuAIAVLgCACS4AgB0uAICnKQCAqykAgCUuAIAtLgCAC+oAgEkuAIBNLgCAWS4AgBDqAIBhLgCAZS4AgEQvAICvKQCAeS4AgJUuAICRLgCAGuoAgJ0uAIAf6gCAqS4AgKUuAICtLgCAvS4AgMEuAICzKQCAgG0/AIF5PwCCoT8Ag6E/AIShPwCFrcgAhq0/AIelPwCInT8AiSEFAIqVyACLJQUAjD0FAI0lBQCO+cgAjyUFAJBdBQCRZQUAkmEFAJN1BQCUFQUAlfU8AJYRBQCXDQUAmD0FAJkFBQCaAQUAmwnVAJwBBQCdgeUAngEFAJ/5BQCgCQUAoRUFAKJlPACjEQUApDUFAKXt1QCmXcgAp1XIAKjd1QCpYQUAqkEFAKtBwwCsRQUArXnIAK5FBQCvlQQAsLEEALGxBACyvQQAs7kEALSpBAC1qQQAtlkEALdNBAC4SQQAuRUEALodBAC7FQQAvA0EAL3ZBwC+yQcAv8kHAIA5BACBOQQAgrUEAIMtBQCEPQUAhSEFAIYtBQCHmdYAiBkFAIllBQCKYQUAi30FAIzlywCNYQUAjmEFAI9hBQCQdcsAkSkFAJL5BQCTaQIAlHkCAJV5AgCWaQIAl2kCAJhZAgCZWQIAmiUCAJs9AgCcJQIAnfE/AJ4hAgCfIQIAoAECAKEBAgCiAQIAowECAKQBAgClAQIApgECAKcBAgCoAQIAqQECAKoBAgCrBQIArB0CAK0FAgCuDQIAr2HAALB5AgCxeQIAsgkCALMJAgC0GQIAtdnlALYVAgC3DQIAuPXlALkxAgC6MQIAuzECALwRAgC9EQIAvhECAL8RAgCfpT4AnqU+AJ2VPgCclT4Am1XMAJqBPgCZiT4AmJE+AJdF9ACWrT4Ala0+AJTh5wCTvekAkrU+AJFNPgCQReQA0S4AgNkuAIDdLgCA4S4AgLcpAIAk6gCAuykAgAQvAIAILwCADC8AgOvrAIAu6gCA5zUAgIJh+wCBCT4AgBE+ADwvAIC/KQCAUeoAgCPrAIC7ZT4AumU+ALl1PgC4dT4At2HkALah8AC1TT4AtHE+ALNpPgCyYT4AsWk+ALCl7wCvDT4AronwAK3h9ACshfAAqxk+AKrJ9ACpbecAqBk+AKchPgCmWT4ApVE+AKRZPgCjQT4Aos3iAKFVPgCgVT4ATC8AgFQvAIDDKQCAaC8AgHgvAIB8LwCAhC8AgJQvAIDLKQCAxykAgDPqAICcLwCAsC8AgLQvAIDELwCA2C8AgNAvAIDULwCA3C8AgPAvAID0LwCADDAAgBQwAIAkMACAWAAAADgwAIBC6gCANDAAgCgwAIBAMACASDAAgFwwAIBH6gCAZDAAgFgwAIBQMACAzykAgGwwAIB0MACAfDAAgHAwAIDTKQCAlDAAgEzqAIDAMACAAjEAgN4wAIDfKQCA2ykAgNcpAICqKwCArisAgAYxAIDuMACAuzUAgEMqAIAaMQCALjEAgCYxAIBl6gCA4ykAgEIxAIA2MQCAXzEAgICJAQCBiQEAgpkBAIOZAQCEiQEAhYkBAIa5AQCHuQEAiN3EAImNAQCKhQEAi50BAIyFAQCNjQEAjoUBAI8BxwCQgQEAkYEBAJKBAQCTgQEAlIEBAJWBAQCWgQEAl4EBAJj1zACZsdkAmokBAJuJAQCcmQEAnZkBAJ6JAQCfiQEAoHkBAKF5AQCizccAo4kCAKSZAgClmQIApokCAKfh2wCotQIAqb0CAKq1AgCrjQIArJUCAK2dAgCulQIAr40CALD1AgCx/QIAsvUCALONAgC0lQIAtfXAALaRAgC3kQIAuLECALmxAgC6sQIAu7ECALyRAgC9kQIAvpECAL+RAgCYjQ0AmQUCAJoNAgCbBQIAnB0CAJ0FAgCeDQIAnwUCAJBt6QCRSQ0AkkUNAJNdDQCUsQ0AlbUNAJa9DQCXtQ0AiBENAIkRDQCKEQ0AixENAIwxDQCNMQ0AjsnPAI81DQCAkQ4AgZEOAIKRDgCDkQ4AhDENAIUxDQCGMQ0AhzENALjpAgC56QIAuvkCALv5AgC86QIAvekCAL4ZAgC/GQIAsFnHALGpzwCy5QIAs/0CALTlAgC17QIAtuUCALfZAgCoddoAqfECAKrxAgCrPccArO0CAK2VAgCunQIAr5UCAKD9AgChxQIAos0CAKMFxwCk2QIApdkCAKbJAgCnyQIAgAAAAG/qAIBrMQCASjEAgHMxAIB3MQCAezEAgH8xAICLMQCAdOoAgJMxAIDrKQCAnzEAgHnqAICjMQCA7ykAgK8xAIC7MQCAyzEAgH7qAIAV6gCAg+oAgOsxAICI6gCA9zEAgP8xAIDvMQCACzIAgBsyAIAjMgCALzIAgDMyAICN6gCAFzIAgEsyAIBPMgCA8ykAgF8yAICS6gCAQzIAgH8yAICX6gCAnOoAgIMyAICXMgCAjzIAgPcpAICbMgCAqzIAgKcyAICzMgCA2ekAgMMyAICh6gCAzzIAgKvqAIDjMgCAAzMAgLDqAIAXMwCAGzMAgLXqAIC66gCANzMAgEczAID7KQCASzMAgP8pAIBjMwCAZzMAgHMzAIB/MwCAAyoAgJczAIC/6gCAszMAgMTqAIAp6gCAzzMAgMnqAIDO6gCA0+oAgAcqAIALKgCA3eoAgNjqAIDi6gCA5+oAgA80AIATNACAHzQAgBcqAIAbKgCA8eoAgF4AAAAzNACAHyoAgPbqAID76gCAAOsAgKM0AIAjKgCArzQAgLM0AIAF6wCACusAgMs0AIAnKgCAD+sAgN80AIDjNACAKyoAgBTrAID/NACALyoAgA81AIAHNQCAFzUAgB7rAIAvNQCAMyoAgDs1AIBDNQCAUzUAgDcqAIAo6wCALesAgDsqAICADd8AgVUBAIJdAQCDVQEAhE0BAIV1AQCGfQEAh3kBAIgx3wCJod8AikEBAItBAQCMQQEAjUEBAI5FAQCP5cgAkFnfAJHFAQCSzQEAkwHCAJTZAQCV2QEAlt3AAJfNAQCY9QEAmWHdAJrxAQCb8QEAnNEBAJ3RAQCe3QEAn9UBAKAtAQChNQEAoj0BAKM1AQCkLQEApVUBAKZdAQCnVQEAqG0BAKl1AQCqfQEAq3UBAKxtAQCtVQEArl0BAK9VAQCwLQEAsTUBALI9AQCzNQEAtC0BALXRAgC20QIAt9ECALjxAgC58QIAuvXdALv1AgC87QIAvdUCAL7dAgC/1QIAntkFAJ/ZBQCc2QUAndkFAJrZBQCb2QUAmNkFAJnZBQCWmQUAl5kFAJTN3gCVmQUAkp0FAJOFBQCQnQUAkZUFAI7dBQCP5QUAjN0FAI3VBQCKwQUAi7XeAIjRBQCJ0QUAhuEFAIfhBQCEEQUAhREFAIIdxQCDAQUAgAkFAIExwAC+yQIAv8kCALzJAgC9yQIAuqkCALupAgC4ccgAuakCALaNAgC3lQIAtI0CALWFAgCyrQIAs5UCALBN3gCxpQIArtECAK/RAgCsxQIArcECAKrVAgCr3QIAqCUFAKndAgCmFQUApx0FAKQFBQClHQUAohUFAKMdBQCgGQUAoRHeAIAAAAAy6wCAazUAgDfrAIB3NQCAgzUAgDzrAIBB6wCAnzUAgEbrAICnNQCAVuoAgD8qAIC/NQCAwzUAgEcqAIDHNQCAIS4AgEvrAIBQ6wCAW+oAgGDqAIDrNQCAAzgAgEsqAIAXNgCAEzYAgBs2AIAmLACAHzYAgCM2AIAnNgCALzYAgFXrAIAzNgCARzYAgEs2AIA3NgCATzYAgGM2AIBDNgCAVzYAgFs2AIBfNgCAWusAgGTrAIBf6wCATyoAgH82AICDNgCAizYAgHjrAICPNgCAaesAgFMqAIBXKgCAbusAgHPrAIBbKgCArzYAgLc2AIC7NgCAxzYAgMM2AIDPNgCAyzYAgNM2AIDXNgCA3zYAgF8qAIDnNgCA6zYAgGMqAID7NgCAfesAgAs3AIAPNwCAZyoAgBs3AICbKQCAgusAgIfrAIBrKgCAbyoAgEc3AICM6wCAnzcAgKM3AIC7NwCAxzcAgJbrAIDo6QCAXQAAANM3AIDPNwCA2zcAgO3pAIDnNwCAm+sAgKDrAIAzOACAPzgAgEc4AICl6wCASzgAgHc4AICDOACAhzgAgH84AICTOACAlzgAgKrrAICjOACAcyoAgKs4AICv6wCAdyoAgOM4AICxLgCA+zgAgLTrAIC+6wCAeyoAgH8qAIAjOQCAw+sAgIMqAIDI6wCAgBkBAIEZAQCCKQEAgykBAIT99ACFPQEAhjUBAIctAQCIFQEAiR0BAIoVAQCLbQEAjHUBAI19AQCOdQEAj20BAJAJwwCRCcMAkkH5AJMZAQCUCQEAlQkBAJY5AQCXOQEAmAkBAJkJAQCaHQEAmxUBAJwNAQCd9QEAnv0BAJ8twwCgCQEAoQkBAKIZAQCjGQEApAkBAKUJAQCmOQEApzkBAKgJAQCpCQEAqhkBAKsZAQCsCQEArQkBAK55AQCveQEAsAkBALEJAQCyGQEAsxkBALQJAQC1CQEAtjkBALc5AQC4CQEAuQkBALoZAQC7GQEAvAkBAL0JAQC+kcMAv5XDAJyVHQCdnR0AnpUdAJ8lwgCYPdsAmZ0dAJqVHQCbjR0AlFkdAJVZHQCWqR0Al6kdAJBZHQCRWR0AkkkdAJNJHQCMGR0AjRkdAI4pHQCPKR0AiB0dAIkFHQCKDR0Aiy0VAIRdHQCFJR0Ahi0dAIclHQCAXR0AgUUdAIJNHQCDRR0AvIkCAL2JAgC+mQIAv5kCALhtHQC5lQIAup0CALupwAC0ZdcAtVUdALZdHQC3VR0AsFEdALFRHQCyUR0As1EdAKwRHQCtER0ArhEdAK8RHQCoER0AqREdAKoRHQCrER0ApFEdAKVRHQCmUR0Ap1EdAKBRHQChUR0AolEdAKNRHQCAeQAAgXkAAIKJAACDiQAAhJkAAIWZAACGiQAAh4kAAIi5AACJuQAAikXBAIuNAACMlQAAjZ0AAI6VAACPjQAAkPUAAJH9AACS9QAAk40AAJSVAACVnfsAlpEAAJeF+wCYrQAAmbUAAJq9AACbtQAAnJ37AJ2pAABDOQCAzesAgFs5AICHKgCAazkAgHc5AIB/OQCAhzkAgIsqAIDS6wCAtzkAgMM5AICPKgCAkyoAgMc5AIDX6wCAlyoAgNzrAIDh6wCA5usAgJsqAIAHOgCACzoAgBM6AIAbOgCA8OsAgLkAAAC4AAAAuwAAALoAAAC9AAAAvAAAAL8AAAC+AAAAACAAIMyBACDMgwAgzIQAIMyFACDMhgAgzIcAIMyIACDMiMyAACDMiMyBACDMiM2CACDMigAgzIsAIMyTACDMk8yAACDMk8yBACDMk82CACDMlAAgzJTMgAAgzJTMgQAgzJTNggAgzKcAIMyoACDMswAgzYIAIM2FACDZiwAg2YwAINmM2ZEAINmNACDZjdmRACDZjgAg2Y7ZkQAg2Y8AINmP2ZEAINmQACDZkNmRACDZkQAg2ZHZsAAg2ZIAIOOCmQAg44KaACEAISEAIT8AIgAjACQAJQAmACcAKAAoMSkAKDEwKQAoMTEpACgxMikAKDEzKQAoMTQpACgxNSkAKDE2KQAoMTcpACgxOCkAKDE5KQAoMikAKDIwKQAoMykAKDQpACg1KQAoNikAKDcpACg4KQAoOSkAKEEpAChCKQAoQykAKEQpAChFKQAoRikAKEcpAChIKQAoSSkAKEopAChLKQAoTCkAKE0pAChOKQAoTykAKFApAChRKQAoUikAKFMpAChUKQAoVSkAKFYpAChXKQAoWCkAKFkpAChaKQAoYSkAKGIpAChjKQAoZCkAKGUpAChmKQAoZykAKGgpAChpKQAoaikAKGspAChsKQAobSkAKG4pAChvKQAocCkAKHEpAChyKQAocykAKHQpACh1KQAodikAKHcpACh4KQAoeSkAKHopACjhhIApACjhhIIpACjhhIMpACjhhIUpACjhhIYpACjhhIcpACjhhIkpACjhhIspACjhhIwpACjhhI4pACjhhI8pACjhhJApACjhhJEpACjhhJIpACjkuIApACjkuIMpACjkuIkpACjkuZ0pACjkuowpACjkupQpACjku6MpACjkvIEpACjkvJEpACjlhaspACjlha0pACjlirQpACjljYEpACjljZQpACjlkI0pACjlkbwpACjlm5spACjlnJ8pACjlraYpACjml6UpACjmnIgpACjmnIkpACjmnKgpACjmoKopACjmsLQpACjngaspACjnibkpACjnm6MpACjnpL4pACjnpZ0pACjnpa0pACjoh6opACjoh7MpACjosqEpACjos4cpACjph5EpACjqsIApACjrgpgpACjri6QpACjrnbwpACjrp4gpACjrsJQpACjsgqwpACjslYQpACjsmKTsoIQpACjsmKTtm4QpACjsnpApACjso7wpACjssKgpACjsubQpACjtg4ApACjtjIwpACjtlZgpACkAKgArACwALQAuAC4uAC4uLgAvADAAMCwAMC4AMOKBhDMAMOeCuQAxADEsADEuADEwADEwLgAxMOaXpQAxMOaciAAxMOeCuQAxMQAxMS4AMTHml6UAMTHmnIgAMTHngrkAMTIAMTIuADEy5pelADEy5pyIADEy54K5ADEzADEzLgAxM+aXpQAxM+eCuQAxNAAxNC4AMTTml6UAMTTngrkAMTUAMTUuADE15pelADE154K5ADE2ADE2LgAxNuaXpQAxNueCuQAxNwAxNy4AMTfml6UAMTfngrkAMTgAMTguADE45pelADE454K5ADE5ADE5LgAxOeaXpQAxOeeCuQAx4oGEADHigYQxMAAx4oGEMgAx4oGEMwAx4oGENAAx4oGENQAx4oGENgAx4oGENwAx4oGEOAAx4oGEOQAx5pelADHmnIgAMeeCuQAyADIsADIuADIwADIwLgAyMOaXpQAyMOeCuQAyMQAyMeaXpQAyMeeCuQAyMgAyMuaXpQAyMueCuQAyMwAyM+aXpQAyM+eCuQAyNAAyNOaXpQAyNOeCuQAyNQAyNeaXpQAyNgAyNuaXpQAyNwAyN+aXpQAyOAAyOOaXpQAyOQAyOeaXpQAy4oGEMwAy4oGENQAy5pelADLmnIgAMueCuQAzADMsADMuADMwADMw5pelADMxADMx5pelADMyADMzADM0ADM1ADM2ADM3ADM4ADM5ADPigYQ0ADPigYQ1ADPigYQ4ADPml6UAM+aciAAz54K5ADQANCwANC4ANDAANDEANDIANDMANDQANDUANDYANDcANDgANDkANOKBhDUANOaXpQA05pyIADTngrkANQA1LAA1LgA1MAA14oGENgA14oGEOAA15pelADXmnIgANeeCuQA2ADYsADYuADbml6UANuaciAA254K5ADcANywANy4AN+KBhDgAN+aXpQA35pyIADfngrkAOAA4LAA4LgA45pelADjmnIgAOOeCuQA5ADksADkuADnml6UAOeaciAA554K5ADoAOjo9ADsAPAA9AD09AD09PQA+AD8APyEAPz8AQABBAEFVAEHiiJVtAEIAQnEAQwBDRABDby4AQ+KIlWtnAEQAREoARFoARHoARMW9AETFvgBFAEYARkFYAEcAR0IAR0h6AEdQYQBHeQBIAEhQAEhWAEhnAEh6AEkASUkASUlJAElKAElVAElWAElYAEoASwBLQgBLSwBLTQBMAExKAExURABMagBMwrcATQBNQgBNQwBNRABNSHoATVBhAE1SAE1WAE1XAE3OqQBOAE5KAE5qAE5vAE8AUABQSABQUE0AUFBWAFBSAFBURQBQYQBRAFIAUnMAUwBTRABTTQBTUwBTdgBUAFRFTABUSHoAVE0AVQBWAFZJAFZJSQBWSUlJAFbiiJVtAFcAV0MAV1oAV2IAWABYSQBYSUkAWQBaAFsAXABdAF4AXwBgAGEAYS5tLgBhL2MAYS9zAGHKvgBiAGJhcgBjAGMvbwBjL3UAY2FsAGNjAGNkAGNtAGNtMgBjbTMAZABkQgBkYQBkbABkbQBkbTIAZG0zAGR6AGTFvgBlAGVWAGVyZwBmAGZmAGZmaQBmZmwAZmkAZmwAZm0AZwBnYWwAaABoUGEAaGEAaQBpaQBpaWkAaWoAaW4AaXYAaXgAagBrAGtBAGtIegBrUGEAa1YAa1cAa2NhbABrZwBrbABrbQBrbTIAa20zAGt0AGvOqQBsAGxqAGxtAGxuAGxvZwBseABswrcAbQBtMgBtMwBtQQBtVgBtVwBtYgBtZwBtaWwAbWwAbW0AbW0yAG1tMwBtb2wAbXMAbeKIlXMAbeKIlXMyAG4AbkEAbkYAblYAblcAbmoAbm0AbnMAbwBvVgBwAHAubS4AcEEAcEYAcFYAcFcAcGMAcHMAcQByAHJhZAByYWTiiJVzAHJhZOKIlXMyAHMAc3IAc3QAdAB1AHYAdmkAdmlpAHZpaWkAdwB4AHhpAHhpaQB5AHoAewB8AH0AwqIAwqMAwqUAwqYAwqwAwrBDAMKwRgDCtwDDgADDgQDDggDDgwDDhADDhQDDhgDDhwDDiADDiQDDigDDiwDDjADDjQDDjgDDjwDDkQDDkgDDkwDDlADDlQDDlgDDmQDDmgDDmwDDnADDnQDDoADDoQDDogDDowDDpADDpQDDpwDDqADDqQDDqgDDqwDDrADDrQDDrgDDrwDDsADDsQDDsgDDswDDtADDtQDDtgDDuQDDugDDuwDDvADDvQDDvwDEgADEgQDEggDEgwDEhADEhQDEhgDEhwDEiADEiQDEigDEiwDEjADEjQDEjgDEjwDEkgDEkwDElADElQDElgDElwDEmADEmQDEmgDEmwDEnADEnQDEngDEnwDEoADEoQDEogDEowDEpADEpQDEpgDEpwDEqADEqQDEqgDEqwDErADErQDErgDErwDEsADEsQDEtADEtQDEtgDEtwDEuQDEugDEuwDEvADEvQDEvgDFgwDFhADFhQDFhgDFhwDFiADFiwDFjADFjQDFjgDFjwDFkADFkQDFkwDFlADFlQDFlgDFlwDFmADFmQDFmgDFmwDFnADFnQDFngDFnwDFoADFoQDFogDFowDFpADFpQDFqADFqQDFqgDFqwDFrADFrQDFrgDFrwDFsADFsQDFsgDFswDFtADFtQDFtgDFtwDFuADFuQDFugDFuwDFvADFvQDFvgDGjgDGkADGoADGoQDGqwDGrwDGsADHjQDHjgDHjwDHkADHkQDHkgDHkwDHlADHlQDHlgDHlwDHmADHmQDHmgDHmwDHnADHngDHnwDHoADHoQDHogDHowDHpgDHpwDHqADHqQDHqgDHqwDHrADHrQDHrgDHrwDHsADHtADHtQDHuADHuQDHugDHuwDHvADHvQDHvgDHvwDIgADIgQDIggDIgwDIhADIhQDIhgDIhwDIiADIiQDIigDIiwDIjADIjQDIjgDIjwDIkADIkQDIkgDIkwDIlADIlQDIlgDIlwDImADImQDImgDImwDIngDInwDIogDIpgDIpwDIqADIqQDIqgDIqwDIrADIrQDIrgDIrwDIsADIsQDIsgDIswDItwDJkADJkQDJkgDJlADJlQDJmQDJmwDJnADJnwDJoQDJowDJpQDJpgDJqADJqQDJqgDJqwDJrQDJrwDJsADJsQDJsgDJswDJtADJtQDJuADJuQDJuwDKgQDKggDKgwDKiQDKigDKiwDKjADKjQDKkADKkQDKkgDKlQDKnQDKnwDKuQDKvG4AzIAAzIEAzIjMgQDMkwDOhgDOiADOiQDOigDOjADOjgDOjwDOkADOkQDOkgDOkwDOlADOlQDOlgDOlwDOmADOmQDOmgDOmwDOnADOnQDOngDOnwDOoADOoQDOowDOpADOpQDOpgDOpwDOqADOqQDOqgDOqwDOrADOrQDOrgDOrwDOsADOsQDOsgDOswDOtADOtQDOtgDOtwDOuADOuQDOugDOuwDOvADOvEEAzrxGAM68VgDOvFcAzrxnAM68bADOvG0AzrxzAM69AM6+AM6/AM+AAM+BAM+CAM+DAM+EAM+FAM+GAM+HAM+IAM+JAM+KAM+LAM+MAM+NAM+OAM+cAM+dANCAANCBANCDANCHANCMANCNANCOANCZANC5ANC9ANGKANGMANGQANGRANGTANGXANGcANGdANGeANG2ANG3ANOBANOCANOQANORANOSANOTANOWANOXANOaANObANOcANOdANOeANOfANOiANOjANOkANOlANOmANOnANOqANOrANOsANOtANOuANOvANOwANOxANOyANOzANO0ANO1ANO4ANO5ANWl1oIA1bTVpQDVtNWrANW01a0A1bTVtgDVvtW2ANeQANeQ1rcA15DWuADXkNa8ANeQ15wA15EA15HWvADXkda/ANeSANeS1rwA15MA15PWvADXlADXlNa8ANeV1rkA15XWvADXlta8ANeY1rwA15nWtADXmda8ANea1rwA15sA15vWvADXm9a/ANecANec1rwA150A157WvADXoNa8ANeh1rwA16IA16PWvADXpNa8ANek1r8A16bWvADXp9a8ANeoANeo1rwA16nWvADXqda814EA16nWvNeCANep14EA16nXggDXqgDXqta8ANey1rcA2KEA2KIA2KMA2KQA2KUA2KYA2KbYpwDYptisANim2K0A2KbYrgDYptixANim2LIA2KbZhQDYptmGANim2YcA2KbZiADYptmJANim2YoA2KbbhgDYptuHANim24gA2KbbkADYptuVANinANin2YPYqNixANin2YTZhNmHANin2YsA2KfZtADYqADYqNisANio2K0A2KjYrdmKANio2K4A2KjYrtmKANio2LEA2KjYsgDYqNmFANio2YYA2KjZhwDYqNmJANio2YoA2KkA2KoA2KrYrADYqtis2YUA2KrYrNmJANiq2KzZigDYqtitANiq2K3YrADYqtit2YUA2KrYrgDYqtiu2YUA2KrYrtmJANiq2K7ZigDYqtixANiq2LIA2KrZhQDYqtmF2KwA2KrZhditANiq2YXYrgDYqtmF2YkA2KrZhdmKANiq2YYA2KrZhwDYqtmJANiq2YoA2KsA2KvYrADYq9ixANir2LIA2KvZhQDYq9mGANir2YcA2KvZiQDYq9mKANisANis2K0A2KzYrdmJANis2K3ZigDYrNmEINis2YTYp9mE2YcA2KzZhQDYrNmF2K0A2KzZhdmJANis2YXZigDYrNmJANis2YoA2K0A2K3YrADYrdis2YoA2K3ZhQDYrdmF2YkA2K3ZhdmKANit2YkA2K3ZigDYrgDYrtisANiu2K0A2K7ZhQDYrtmJANiu2YoA2K8A2LAA2LDZsADYsQDYsdiz2YjZhADYsdmwANix24zYp9mEANiyANizANiz2KwA2LPYrNitANiz2KzZiQDYs9itANiz2K3YrADYs9iuANiz2K7ZiQDYs9iu2YoA2LPYsQDYs9mFANiz2YXYrADYs9mF2K0A2LPZhdmFANiz2YcA2LPZiQDYs9mKANi0ANi02KwA2LTYrNmKANi02K0A2LTYrdmFANi02K3ZigDYtNiuANi02LEA2LTZhQDYtNmF2K4A2LTZhdmFANi02YcA2LTZiQDYtNmKANi1ANi12K0A2LXYrditANi12K3ZigDYtdiuANi12LEA2LXZhNi52YUA2LXZhNmJANi12YTZiSDYp9mE2YTZhyDYudmE2YrZhyDZiNiz2YTZhQDYtdmE25IA2LXZhQDYtdmF2YUA2LXZiQDYtdmKANi2ANi22KwA2LbYrQDYttit2YkA2LbYrdmKANi22K4A2LbYrtmFANi22LEA2LbZhQDYttmJANi22YoA2LcA2LfYrQDYt9mFANi32YXYrQDYt9mF2YUA2LfZhdmKANi32YkA2LfZigDYuADYuNmFANi5ANi52KwA2LnYrNmFANi52YTZitmHANi52YUA2LnZhdmFANi52YXZiQDYudmF2YoA2LnZiQDYudmKANi6ANi62KwA2LrZhQDYutmF2YUA2LrZhdmJANi62YXZigDYutmJANi62YoA2YDZiwDZgNmOANmA2Y7ZkQDZgNmPANmA2Y/ZkQDZgNmQANmA2ZDZkQDZgNmRANmA2ZIA2YEA2YHYrADZgditANmB2K4A2YHYrtmFANmB2YUA2YHZhdmKANmB2YkA2YHZigDZggDZgtitANmC2YTbkgDZgtmFANmC2YXYrQDZgtmF2YUA2YLZhdmKANmC2YkA2YLZigDZgwDZg9inANmD2KwA2YPYrQDZg9iuANmD2YQA2YPZhQDZg9mF2YUA2YPZhdmKANmD2YkA2YPZigDZhADZhNiiANmE2KMA2YTYpQDZhNinANmE2KwA2YTYrNisANmE2KzZhQDZhNis2YoA2YTYrQDZhNit2YUA2YTYrdmJANmE2K3ZigDZhNiuANmE2K7ZhQDZhNmFANmE2YXYrQDZhNmF2YoA2YTZhwDZhNmJANmE2YoA2YUA2YXYpwDZhdisANmF2KzYrQDZhdis2K4A2YXYrNmFANmF2KzZigDZhditANmF2K3YrADZhdit2YUA2YXYrdmF2K8A2YXYrdmKANmF2K4A2YXYrtisANmF2K7ZhQDZhdiu2YoA2YXZhQDZhdmF2YoA2YXZiQDZhdmKANmGANmG2KwA2YbYrNitANmG2KzZhQDZhtis2YkA2YbYrNmKANmG2K0A2YbYrdmFANmG2K3ZiQDZhtit2YoA2YbYrgDZhtixANmG2LIA2YbZhQDZhtmF2YkA2YbZhdmKANmG2YYA2YbZhwDZhtmJANmG2YoA2YcA2YfYrADZh9mFANmH2YXYrADZh9mF2YUA2YfZiQDZh9mKANmH2bAA2YgA2YjYs9mE2YUA2YjZtADZiQDZidmwANmKANmK2KwA2YrYrNmKANmK2K0A2YrYrdmKANmK2K4A2YrYsQDZitiyANmK2YUA2YrZhdmFANmK2YXZigDZitmGANmK2YcA2YrZiQDZitmKANmK2bQA2a4A2a8A2bEA2bkA2boA2bsA2b4A2b8A2oAA2oMA2oQA2oYA2ocA2ogA2owA2o0A2o4A2pEA2pgA2qEA2qQA2qYA2qkA2q0A2q8A2rEA2rMA2roA2rsA2r4A24AA24EA24IA24UA24YA24cA24fZtADbiADbiQDbiwDbjADbkADbkgDbkwDgpJXgpLwA4KSW4KS8AOCkl+CkvADgpJzgpLwA4KSh4KS8AOCkouCkvADgpKkA4KSr4KS8AOCkr+CkvADgpLEA4KS0AOCmoeCmvADgpqLgprwA4Kav4Ka8AOCniwDgp4wA4KiW4Ki8AOCol+CovADgqJzgqLwA4Kir4Ki8AOCosuCovADgqLjgqLwA4Kyh4Ky8AOCsouCsvADgrYgA4K2LAOCtjADgrpQA4K+KAOCviwDgr4wA4LGIAOCzgADgs4cA4LOIAOCzigDgs4sA4LWKAOC1iwDgtYwA4LeaAOC3nADgt50A4LeeAOC5jeC4sgDguqvgupkA4Lqr4LqhAOC7jeC6sgDgvIsA4L2A4L61AOC9guC+twDgvYzgvrcA4L2R4L63AOC9luC+twDgvZvgvrcA4L2x4L2yAOC9seC9tADgvbHgvoAA4L6Q4L61AOC+kuC+twDgvpzgvrcA4L6h4L63AOC+puC+twDgvqvgvrcA4L6y4L2x4L6AAOC+suC+gADgvrPgvbHgvoAA4L6z4L6AAOGApgDhg5wA4YSAAOGEgQDhhIIA4YSDAOGEhADhhIUA4YSGAOGEhwDhhIgA4YSJAOGEigDhhIsA4YSMAOGEjQDhhI4A4YSPAOGEkADhhJEA4YSSAOGElADhhJUA4YSaAOGEnADhhJ0A4YSeAOGEoADhhKEA4YSiAOGEowDhhKcA4YSpAOGEqwDhhKwA4YStAOGErgDhhK8A4YSyAOGEtgDhhYAA4YWHAOGFjADhhZcA4YWYAOGFmQDhhaAA4YWhAOGFogDhhaMA4YWkAOGFpQDhhaYA4YWnAOGFqADhhakA4YWqAOGFqwDhhawA4YWtAOGFrgDhha8A4YWwAOGFsQDhhbIA4YWzAOGFtADhhbUA4YaEAOGGhQDhhogA4YaRAOGGkgDhhpQA4YaeAOGGoQDhhqoA4YasAOGGrQDhhrAA4YaxAOGGsgDhhrMA4Ya0AOGGtQDhh4cA4YeIAOGHjADhh44A4YeTAOGHlwDhh5kA4YedAOGHnwDhh7EA4YeyAOGshgDhrIgA4ayKAOGsjADhrI4A4aySAOGsuwDhrL0A4a2AAOGtgQDhrYMA4bSCAOG0lgDhtJcA4bScAOG0nQDhtKUA4bW7AOG2hQDhuIAA4biBAOG4ggDhuIMA4biEAOG4hQDhuIYA4biHAOG4iADhuIkA4biKAOG4iwDhuIwA4biNAOG4jgDhuI8A4biQAOG4kQDhuJIA4biTAOG4lADhuJUA4biWAOG4lwDhuJgA4biZAOG4mgDhuJsA4bicAOG4nQDhuJ4A4bifAOG4oADhuKEA4biiAOG4owDhuKQA4bilAOG4pgDhuKcA4bioAOG4qQDhuKoA4birAOG4rADhuK0A4biuAOG4rwDhuLAA4bixAOG4sgDhuLMA4bi0AOG4tQDhuLYA4bi3AOG4uADhuLkA4bi6AOG4uwDhuLwA4bi9AOG4vgDhuL8A4bmAAOG5gQDhuYIA4bmDAOG5hADhuYUA4bmGAOG5hwDhuYgA4bmJAOG5igDhuYsA4bmMAOG5jQDhuY4A4bmPAOG5kADhuZEA4bmSAOG5kwDhuZQA4bmVAOG5lgDhuZcA4bmYAOG5mQDhuZoA4bmbAOG5nADhuZ0A4bmeAOG5nwDhuaAA4bmhAOG5ogDhuaMA4bmkAOG5pQDhuaYA4bmnAOG5qADhuakA4bmqAOG5qwDhuawA4bmtAOG5rgDhua8A4bmwAOG5sQDhubIA4bmzAOG5tADhubUA4bm2AOG5twDhubgA4bm5AOG5ugDhubsA4bm8AOG5vQDhub4A4bm/AOG6gADhuoEA4bqCAOG6gwDhuoQA4bqFAOG6hgDhuocA4bqIAOG6iQDhuooA4bqLAOG6jADhuo0A4bqOAOG6jwDhupAA4bqRAOG6kgDhupMA4bqUAOG6lQDhupYA4bqXAOG6mADhupkA4bqgAOG6oQDhuqIA4bqjAOG6pADhuqUA4bqmAOG6pwDhuqgA4bqpAOG6qgDhuqsA4bqsAOG6rQDhuq4A4bqvAOG6sADhurEA4bqyAOG6swDhurQA4bq1AOG6tgDhurcA4bq4AOG6uQDhuroA4bq7AOG6vADhur0A4bq+AOG6vwDhu4AA4buBAOG7ggDhu4MA4buEAOG7hQDhu4YA4buHAOG7iADhu4kA4buKAOG7iwDhu4wA4buNAOG7jgDhu48A4buQAOG7kQDhu5IA4buTAOG7lADhu5UA4buWAOG7lwDhu5gA4buZAOG7mgDhu5sA4bucAOG7nQDhu54A4bufAOG7oADhu6EA4buiAOG7owDhu6QA4bulAOG7pgDhu6cA4buoAOG7qQDhu6oA4burAOG7rADhu60A4buuAOG7rwDhu7AA4buxAOG7sgDhu7MA4bu0AOG7tQDhu7YA4bu3AOG7uADhu7kA4byAAOG8gQDhvIIA4byDAOG8hADhvIUA4byGAOG8hwDhvIgA4byJAOG8igDhvIsA4byMAOG8jQDhvI4A4byPAOG8kADhvJEA4bySAOG8kwDhvJQA4byVAOG8mADhvJkA4byaAOG8mwDhvJwA4bydAOG8oADhvKEA4byiAOG8owDhvKQA4bylAOG8pgDhvKcA4byoAOG8qQDhvKoA4byrAOG8rADhvK0A4byuAOG8rwDhvLAA4byxAOG8sgDhvLMA4by0AOG8tQDhvLYA4by3AOG8uADhvLkA4by6AOG8uwDhvLwA4by9AOG8vgDhvL8A4b2AAOG9gQDhvYIA4b2DAOG9hADhvYUA4b2IAOG9iQDhvYoA4b2LAOG9jADhvY0A4b2QAOG9kQDhvZIA4b2TAOG9lADhvZUA4b2WAOG9lwDhvZkA4b2bAOG9nQDhvZ8A4b2gAOG9oQDhvaIA4b2jAOG9pADhvaUA4b2mAOG9pwDhvagA4b2pAOG9qgDhvasA4b2sAOG9rQDhva4A4b2vAOG9sADhvbIA4b20AOG9tgDhvbgA4b26AOG9vADhvoAA4b6BAOG+ggDhvoMA4b6EAOG+hQDhvoYA4b6HAOG+iADhvokA4b6KAOG+iwDhvowA4b6NAOG+jgDhvo8A4b6QAOG+kQDhvpIA4b6TAOG+lADhvpUA4b6WAOG+lwDhvpgA4b6ZAOG+mgDhvpsA4b6cAOG+nQDhvp4A4b6fAOG+oADhvqEA4b6iAOG+owDhvqQA4b6lAOG+pgDhvqcA4b6oAOG+qQDhvqoA4b6rAOG+rADhvq0A4b6uAOG+rwDhvrAA4b6xAOG+sgDhvrMA4b60AOG+tgDhvrcA4b64AOG+uQDhvroA4b68AOG/ggDhv4MA4b+EAOG/hgDhv4cA4b+IAOG/igDhv4wA4b+QAOG/kQDhv5IA4b+WAOG/lwDhv5gA4b+ZAOG/mgDhv6AA4b+hAOG/ogDhv6QA4b+lAOG/pgDhv6cA4b+oAOG/qQDhv6oA4b+sAOG/sgDhv7MA4b+0AOG/tgDhv7cA4b+4AOG/ugDhv7wA4oCQAOKAkwDigJQA4oCy4oCyAOKAsuKAsuKAsgDigLLigLLigLLigLIA4oC14oC1AOKAteKAteKAtQDigqkA4oaQAOKGkQDihpIA4oaTAOKGmgDihpsA4oauAOKHjQDih44A4oePAOKIggDiiIQA4oiHAOKIiQDiiIwA4oiRAOKIkgDiiKQA4oimAOKIq+KIqwDiiKviiKviiKsA4oir4oir4oir4oirAOKIruKIrgDiiK7iiK7iiK4A4omBAOKJhADiiYcA4omJAOKJoADiiaIA4omtAOKJrgDiia8A4omwAOKJsQDiibQA4om1AOKJuADiibkA4oqAAOKKgQDiioQA4oqFAOKKiADiiokA4oqsAOKKrQDiiq4A4oqvAOKLoADii6EA4ouiAOKLowDii6oA4ourAOKLrADii60A4pSCAOKWoADil4sA4qaFAOKmhgDiq53MuADitaEA44CBAOOAggDjgIgA44CJAOOAigDjgIsA44CMAOOAjQDjgI4A44CPAOOAkADjgJEA44CSAOOAlADjgJRT44CVAOOAlOS4ieOAlQDjgJTkuozjgJUA44CU5Yud44CVAOOAlOWuieOAlQDjgJTmiZPjgJUA44CU5pWX44CVAOOAlOacrOOAlQDjgJTngrnjgJUA44CU55uX44CVAOOAlQDjgJYA44CXAOOBjADjgY4A44GQAOOBkgDjgZQA44GWAOOBmADjgZoA44GcAOOBngDjgaAA44GiAOOBpQDjgacA44GpAOOBsADjgbEA44GzAOOBtADjgbYA44G3AOOBuQDjgboA44G744GLAOOBvADjgb0A44KI44KKAOOClADjgpkA44KaAOOCngDjgqEA44KiAOOCouODkeODvOODiADjgqLjg6vjg5XjgqEA44Ki44Oz44Oa44KiAOOCouODvOODqwDjgqMA44KkAOOCpOODi+ODs+OCsADjgqTjg7Pjg4EA44KlAOOCpgDjgqbjgqnjg7MA44KnAOOCqADjgqjjgrnjgq/jg7zjg4kA44Ko44O844Kr44O8AOOCqQDjgqoA44Kq44Oz44K5AOOCquODvOODoADjgqsA44Kr44Kk44OqAOOCq+ODqeODg+ODiADjgqvjg63jg6rjg7wA44KsAOOCrOODreODswDjgqzjg7Pjg54A44KtAOOCreODpeODquODvADjgq3jg60A44Kt44Ot44Kw44Op44OgAOOCreODreODoeODvOODiOODqwDjgq3jg63jg6/jg4Pjg4gA44KuAOOCruOCrADjgq7jg4vjg7wA44Ku44Or44OA44O8AOOCrwDjgq/jg6vjgrzjgqTjg60A44Kv44Ot44O844ONAOOCsADjgrDjg6njg6AA44Kw44Op44Og44OI44OzAOOCsQDjgrHjg7zjgrkA44KyAOOCswDjgrPjgrMA44Kz44OIAOOCs+ODq+ODigDjgrPjg7zjg50A44K0AOOCtQDjgrXjgqTjgq/jg6sA44K144Oz44OB44O844OgAOOCtgDjgrcA44K344Oq44Oz44KwAOOCuADjgrkA44K6AOOCuwDjgrvjg7Pjg4EA44K744Oz44OIAOOCvADjgr0A44K+AOOCvwDjg4AA44OA44O844K5AOODgQDjg4IA44ODAOODhADjg4UA44OGAOODhwDjg4fjgrcA44OIAOODiOODswDjg4kA44OJ44OrAOODigDjg4rjg44A44OLAOODjADjg40A44OOAOODjuODg+ODiADjg48A44OP44Kk44OEAOODkADjg5Djg7zjg6zjg6sA44ORAOODkeODvOOCu+ODs+ODiADjg5Hjg7zjg4QA44OSAOODkwDjg5Pjg6sA44OUAOODlOOCouOCueODiOODqwDjg5Tjgq/jg6sA44OU44KzAOODlQDjg5XjgqHjg6njg4Pjg4kA44OV44Kj44O844OIAOODleODqeODswDjg5YA44OW44OD44K344Kn44OrAOODlwDjg5gA44OY44Kv44K/44O844OrAOODmOODq+ODhADjg5kA44OZ44O844K/AOODmgDjg5rjgr0A44Oa44OL44OSAOODmuODs+OCuQDjg5rjg7zjgrgA44ObAOODm+ODswDjg5vjg7zjg6sA44Ob44O844OzAOODnADjg5zjg6vjg4gA44OdAOODneOCpOODs+ODiADjg53jg7Pjg4kA44OeAOODnuOCpOOCr+ODrQDjg57jgqTjg6sA44Oe44OD44OPAOODnuODq+OCrwDjg57jg7Pjgrfjg6fjg7MA44OfAOODn+OCr+ODreODswDjg5/jg6oA44Of44Oq44OQ44O844OrAOODoADjg6EA44Oh44KsAOODoeOCrOODiOODswDjg6Hjg7zjg4jjg6sA44OiAOODowDjg6QA44Ok44O844OJAOODpOODvOODqwDjg6UA44OmAOODpuOCouODswDjg6cA44OoAOODqQDjg6oA44Oq44OD44OI44OrAOODquODqQDjg6sA44Or44OU44O8AOODq+ODvOODluODqwDjg6wA44Os44OgAOODrOODs+ODiOOCsuODswDjg60A44OvAOODr+ODg+ODiADjg7AA44OxAOODsgDjg7MA44O0AOODtwDjg7gA44O5AOODugDjg7sA44O8AOODvgDjkp4A45K5AOOSuwDjk58A45SVAOObrgDjm7wA456BAOOgrwDjoaIA46G8AOOjhwDjo6MA46ScAOOkugDjqK4A46msAOOrpADjrIgA46yZAOOtiQDjrp0A47CYAOOxjgDjtLMA47aWAOO6rADjurgA47ybAOO/vADkgIgA5ICYAOSAuQDkgYYA5IKWAOSDowDkhK8A5IiCAOSIpwDkiqAA5IyBAOSMtADkjZkA5I+VAOSPmQDkkIsA5JGrAOSUqwDklZ0A5JWhAOSVqwDkl5cA5Je5AOSYtQDkmr4A5JuHAOSmlQDkp6YA5KmuAOSptgDkqrIA5KyzAOSvjgDks44A5LOtAOSzuADktZYA5LiAAOS4gQDkuIMA5LiJAOS4igDkuIsA5LiNAOS4mQDkuKYA5LioAOS4rQDkuLIA5Li2AOS4uADkuLkA5Li9AOS4vwDkuYEA5LmZAOS5nQDkuoIA5LqFAOS6hgDkuowA5LqUAOS6oADkuqQA5LquAOS6ugDku4AA5LuMAOS7pADku6TlkowA5LyBAOS8kQDkvaAA5L6AAOS+hgDkvosA5L6uAOS+uwDkvr8A5YCCAOWAqwDlgboA5YKZAOWDjwDlg5oA5YOnAOWEqgDlhL8A5YWAAOWFhQDlhY0A5YWUAOWFpADlhaUA5YWnAOWFqADlhakA5YWrAOWFrQDlhbcA5YaAAOWGggDlho0A5YaSAOWGlQDlhpYA5YaXAOWGmQDlhqQA5YarAOWGrADlhrUA5Ya3AOWHiQDlh4wA5YecAOWHngDlh6AA5Ye1AOWIgADliIMA5YiHAOWIlwDliJ0A5YipAOWIugDliLsA5YmGAOWJjQDlibIA5Ym3AOWKiQDlipsA5YqjAOWKswDlirQA5YuHAOWLiQDli5IA5YueAOWLpADli7UA5Yu5AOWLugDljIUA5YyGAOWMlQDljJcA5YyaAOWMuADljLsA5Yy/AOWNgQDljYQA5Y2FAOWNiQDljZEA5Y2UAOWNmgDljZwA5Y2pAOWNsADljbMA5Y21AOWNvQDljb8A5Y6CAOWOtgDlj4MA5Y+IAOWPigDlj4wA5Y+fAOWPowDlj6UA5Y+rAOWPrwDlj7EA5Y+zAOWQhgDlkIgA5ZCNAOWQjwDlkJ0A5ZC4AOWQuQDlkYIA5ZGIAOWRqADlkp4A5ZKiAOWSvQDlk7YA5ZSQAOWVjwDllZMA5ZWVAOWVowDlloQA5ZaHAOWWmQDllp0A5ZarAOWWswDllrYA5ZeAAOWXggDll6IA5ZiGAOWZkQDlmagA5Zm0AOWblwDlm5sA5Zu5AOWclgDlnJcA5ZyfAOWcsADlnosA5Z+OAOWftADloI0A5aCxAOWgsgDloYAA5aGaAOWhngDloqgA5aKsAOWiswDlo5gA5aOfAOWjqwDlo64A5aOwAOWjsgDlo7cA5aSCAOWkhgDlpIoA5aSVAOWkmgDlpJwA5aSiAOWkpwDlpKfmraMA5aSpAOWlhADlpYgA5aWRAOWllADlpaIA5aWzAOWnmADlp6wA5aibAOWopwDlqaIA5ammAOWqtQDlrIgA5ayoAOWsvgDlrZAA5a2XAOWtpgDlroAA5a6FAOWulwDlr4MA5a+YAOWvpwDlr64A5a+zAOWvuADlr78A5bCGAOWwjwDlsKIA5bC4AOWwvwDlsaAA5bGiAOWxpADlsaUA5bGuAOWxsQDlso0A5bOAAOW0mQDltYMA5bWQAOW1qwDlta4A5bW8AOW2sgDltroA5bebAOW3oQDlt6IA5belAOW3pgDlt7EA5be9AOW3vgDluKgA5bi9AOW5qQDlubIA5bmz5oiQAOW5tADluboA5bm8AOW5vwDluqYA5bqwAOW6swDlurYA5buJAOW7igDlu5IA5buTAOW7mQDlu6wA5bu0AOW7vgDlvIQA5byLAOW8kwDlvKIA5b2QAOW9kwDlvaEA5b2iAOW9qQDlvasA5b2zAOW+iwDlvowA5b6XAOW+mgDlvqkA5b6tAOW/gwDlv40A5b+XAOW/tQDlv7kA5oCSAOaAnADmgbUA5oKBAOaClADmg4cA5oOYAOaDoQDmhIgA5oWEAOaFiADmhYwA5oWOAOaFoADmhagA5oW6AOaGjgDmhpAA5oakAOaGrwDmhrIA5oeeAOaHsgDmh7YA5oiAAOaIiADmiJAA5oibAOaIrgDmiLQA5oi2AOaJiwDmiZMA5omdAOaKlQDmirEA5ouJAOaLjwDmi5MA5ouUAOaLvADmi74A5oyHAOaMvQDmjZAA5o2VAOaNqADmjbsA5o6DAOaOoADmjqkA5o+EAOaPhQDmj6QA5pCcAOaQogDmkZIA5pGpAOaRtwDmkb4A5pKaAOaSnQDmk4QA5pSvAOaUtADmlY8A5pWWAOaVrADmlbgA5paHAOaWlwDmlpkA5pakAOaWsADmlrkA5peFAOaXoADml6IA5pejAOaXpQDmmI7msrsA5piTAOaYoADmmK3lkowA5pmJAOaZtADmmogA5pqRAOaanADmmrQA5puGAOabsADmm7QA5pu4AOacgADmnIgA5pyJAOaclwDmnJsA5pyhAOacqADmnY4A5p2TAOadlgDmnZ4A5p27AOaehQDmnpcA5p+zAOafugDmoJcA5qCfAOagqgDmoKrlvI/kvJrnpL4A5qGSAOaigQDmooUA5qKOAOaiqADmpJQA5qWCAOamowDmp6oA5qiCAOaokwDmqqgA5quTAOarmwDmrIQA5qygAOasoQDmrZQA5q2iAOatowDmrbIA5q23AOatuQDmrp8A5q6uAOauswDmrroA5q67AOaviwDmr40A5q+UAOavmwDmsI8A5rCUAOawtADmsY4A5rGnAOayiADmsr8A5rOMAOazjQDms6UA5rOoAOa0lgDmtJsA5rSeAOa0tADmtL4A5rWBAOa1qQDmtaoA5rW3AOa1uADmtoUA5reLAOa3mgDmt6oA5re5AOa4mgDmuK8A5rmuAOa6gADmupwA5rq6AOa7hwDmu4sA5ruRAOa7mwDmvI8A5ryUAOa8ogDmvKMA5r2uAOa/hgDmv6sA5r++AOeAmwDngJ4A54C5AOeBigDngasA54GwAOeBtwDngb0A54KZAOeCrQDng4gA54OZAOeEoQDnhYUA54WJAOeFrgDnhpwA54eOAOeHkADniJAA54ibAOeIqADniKoA54irAOeItQDniLYA54i7AOeIvwDniYcA54mQAOeJmQDniZsA54miAOeJuQDnioAA54qVAOeKrADniq8A54uAAOeLvADnjKoA5421AOeNugDnjoQA546HAOeOiQDnjosA546lAOeOsgDnj54A55CGAOeQiQDnkKIA55GHAOeRnADnkakA55GxAOeShQDnkokA55KYAOeTigDnk5wA55OmAOeUhgDnlJgA55SfAOeUpADnlKgA55SwAOeUsgDnlLMA55S3AOeUuwDnlL4A55WZAOeVpQDnlbAA55aLAOeWkgDnl6IA55iQAOeYnQDnmJ8A55mCAOeZqQDnmbYA55m9AOeargDnmr8A55uKAOebmwDnm6MA55unAOebrgDnm7QA55yBAOecngDnnJ8A552AAOedigDnnosA556nAOefmwDnn6IA55+zAOehjgDnoasA56KMAOeikQDno4oA56OMAOejuwDnpKoA56S6AOekvADnpL4A56WIAOeliQDnpZAA56WWAOelnQDnpZ4A56WlAOelvwDnpoEA56aNAOemjgDnpo8A56auAOemuADnpr4A56eKAOenmADnp6sA56icAOepgADnqYoA56mPAOeptADnqboA56qBAOeqsQDnq4sA56uuAOeruQDnrKAA566PAOevgADnr4YA56+JAOewvgDnsaAA57GzAOexuwDnspIA57K+AOezkgDns5YA57OjAOezpwDns6gA57O4AOe0gADntJAA57SiAOe0rwDntYIA57WbAOe1owDntqAA57a+AOe3hwDnt7QA57iCAOe4iQDnuLcA57mBAOe5hQDnvLYA57y+AOe9kQDnvbIA5725AOe9ugDnvoUA576KAOe+lQDnvpoA5769AOe/ugDogIEA6ICFAOiAjADogJIA6ICzAOiBhgDogaAA6IGvAOiBsADogb4A6IG/AOiCiQDogosA6IKtAOiCsgDohIMA6IS+AOiHmADoh6MA6IeoAOiHqgDoh60A6IezAOiHvADoiIEA6IiEAOiIjADoiJgA6IibAOiInwDoia4A6ImvAOiJsgDoibgA6Im5AOiKiwDoipEA6IqdAOiKsQDoirMA6Iq9AOiLpQDoi6YA6IydAOiMowDojLYA6I2SAOiNkwDojaMA6I6tAOiOvQDoj4kA6I+KAOiPjADoj5wA6I+nAOiPrwDoj7EA6JC9AOiRiQDokZcA6JOuAOiTsQDok7MA6JO8AOiUlgDolaQA6JeNAOiXugDomIYA6JiSAOiYrQDomL8A6JmNAOiZkADomZwA6JmnAOiZqQDomasA6JqIAOiaqQDom6IA6JyOAOicqADonasA6J25AOiehgDonroA6J+hAOiggQDooJ8A6KGAAOihjADooaAA6KGjAOijggDoo48A6KOXAOijngDoo6EA6KO4AOijugDopJAA6KWBAOilpADopb4A6KaGAOimiwDoppYA6KeSAOinowDoqIAA6KqgAOiqqgDoqr8A6KuLAOirkgDoq5YA6KutAOiruADoq74A6KyBAOisuQDorZgA6K6AAOiuigDosLcA6LGGAOixiADosZUA6LG4AOiynQDosqEA6LKpAOiyqwDos4EA6LOCAOizhwDos4gA6LOTAOi0iADotJsA6LWkAOi1sADotbcA6LazAOi2vADot4sA6LevAOi3sADouqsA6LuKAOi7lADovKYA6LyqAOi8uADovLsA6L2iAOi+mwDovp4A6L6wAOi+tQDovrYA6YCjAOmAuADpgYoA6YGpAOmBsgDpgbwA6YKPAOmCkQDpgpQA6YOOAOmDngDpg7EA6YO9AOmEkQDphJsA6YWJAOmFjQDphaoA6YaZAOmGtADph4YA6YeMAOmHjwDph5EA6Yi0AOmIuADpibYA6Ym8AOmLlwDpi5gA6YyEAOmNigDpj7kA6ZCVAOmVtwDploAA6ZaLAOmWrQDplrcA6ZicAOmYrgDpmYsA6ZmNAOmZtQDpmbgA6Zm8AOmahgDpmqMA6Zq2AOmatwDpmrgA6Zq5AOmbgwDpm6IA6ZujAOmbqADpm7YA6Zu3AOmcowDpnLIA6Z2IAOmdkQDpnZYA6Z2eAOmdogDpnakA6Z+LAOmfmwDpn6AA6Z+tAOmfswDpn78A6aCBAOmghQDpoIsA6aCYAOmgqQDpoLsA6aGeAOmiqADpo5sA6aOfAOmjogDpo68A6aO8AOmkqADppKkA6aaWAOmmmQDppqcA6aasAOmnggDpp7EA6ae+AOmpqgDpqqgA6auYAOmrnwDprJIA6aylAOmsrwDprLIA6ay8AOmtmgDpra8A6bGAAOmxlwDps6UA6bO9AOm1pwDptrQA6be6AOm4ngDpubUA6bm/AOm6lwDpup8A6bqlAOm6uwDpu4MA6buNAOm7jgDpu5EA6bu5AOm7vQDpu74A6byFAOm8jgDpvI8A6byTAOm8lgDpvKAA6by7AOm9gwDpvYoA6b2SAOm+jQDpvo4A6b6cAOm+nwDpvqAA6pynAOqdrwDqrLcA6q2SAOqwgADqsIEA6rCCAOqwgwDqsIQA6rCFAOqwhgDqsIcA6rCIAOqwiQDqsIoA6rCLAOqwjADqsI0A6rCOAOqwjwDqsJAA6rCRAOqwkgDqsJMA6rCUAOqwlQDqsJYA6rCXAOqwmADqsJkA6rCaAOqwmwDqsJwA6rCdAOqwngDqsJ8A6rCgAOqwoQDqsKIA6rCjAOqwpADqsKUA6rCmAOqwpwDqsKgA6rCpAOqwqgDqsKsA6rCsAOqwrQDqsK4A6rCvAOqwsADqsLEA6rCyAOqwswDqsLQA6rC1AOqwtgDqsLcA6rC4AOqwuQDqsLoA6rC7AOqwvADqsL0A6rC+AOqwvwDqsYAA6rGBAOqxggDqsYMA6rGEAOqxhQDqsYYA6rGHAOqxiADqsYkA6rGKAOqxiwDqsYwA6rGNAOqxjgDqsY8A6rGQAOqxkQDqsZIA6rGTAOqxlADqsZUA6rGWAOqxlwDqsZgA6rGZAOqxmgDqsZsA6rGcAOqxnQDqsZ4A6rGfAOqxoADqsaEA6rGiAOqxowDqsaQA6rGlAOqxpgDqsacA6rGoAOqxqQDqsaoA6rGrAOqxrADqsa0A6rGuAOqxrwDqsbAA6rGxAOqxsgDqsbMA6rG0AOqxtQDqsbYA6rG3AOqxuADqsbkA6rG6AOqxuwDqsbwA6rG9AOqxvgDqsb8A6rKAAOqygQDqsoIA6rKDAOqyhADqsoUA6rKGAOqyhwDqsogA6rKJAOqyigDqsosA6rKMAOqyjQDqso4A6rKPAOqykADqspEA6rKSAOqykwDqspQA6rKVAOqylgDqspcA6rKYAOqymQDqspoA6rKbAOqynADqsp0A6rKeAOqynwDqsqAA6rKhAOqyogDqsqMA6rKkAOqypQDqsqYA6rKnAOqyqADqsqkA6rKqAOqyqwDqsqwA6rKtAOqyrgDqsq8A6rKwAOqysQDqsrIA6rKzAOqytADqsrUA6rK2AOqytwDqsrgA6rK5AOqyugDqsrsA6rK8AOqyvQDqsr4A6rK/AOqzgADqs4EA6rOCAOqzgwDqs4QA6rOFAOqzhgDqs4cA6rOIAOqziQDqs4oA6rOLAOqzjADqs40A6rOOAOqzjwDqs5AA6rORAOqzkgDqs5MA6rOUAOqzlQDqs5YA6rOXAOqzmADqs5kA6rOaAOqzmwDqs5wA6rOdAOqzngDqs58A6rOgAOqzoQDqs6IA6rOjAOqzpADqs6UA6rOmAOqzpwDqs6gA6rOpAOqzqgDqs6sA6rOsAOqzrQDqs64A6rOvAOqzsADqs7EA6rOyAOqzswDqs7QA6rO1AOqztgDqs7cA6rO4AOqzuQDqs7oA6rO7AOqzvADqs70A6rO+AOqzvwDqtIAA6rSBAOq0ggDqtIMA6rSEAOq0hQDqtIYA6rSHAOq0iADqtIkA6rSKAOq0iwDqtIwA6rSNAOq0jgDqtI8A6rSQAOq0kQDqtJIA6rSTAOq0lADqtJUA6rSWAOq0lwDqtJgA6rSZAOq0mgDqtJsA6rScAOq0nQDqtJ4A6rSfAOq0oADqtKEA6rSiAOq0owDqtKQA6rSlAOq0pgDqtKcA6rSoAOq0qQDqtKoA6rSrAOq0rADqtK0A6rSuAOq0rwDqtLAA6rSxAOq0sgDqtLMA6rS0AOq0tQDqtLYA6rS3AOq0uADqtLkA6rS6AOq0uwDqtLwA6rS9AOq0vgDqtL8A6rWAAOq1gQDqtYIA6rWDAOq1hADqtYUA6rWGAOq1hwDqtYgA6rWJAOq1igDqtYsA6rWMAOq1jQDqtY4A6rWPAOq1kADqtZEA6rWSAOq1kwDqtZQA6rWVAOq1lgDqtZcA6rWYAOq1mQDqtZoA6rWbAOq1nADqtZ0A6rWeAOq1nwDqtaAA6rWhAOq1ogDqtaMA6rWkAOq1pQDqtaYA6rWnAOq1qADqtakA6rWqAOq1qwDqtawA6rWtAOq1rgDqta8A6rWwAOq1sQDqtbIA6rWzAOq1tADqtbUA6rW2AOq1twDqtbgA6rW5AOq1ugDqtbsA6rW8AOq1vQDqtb4A6rW/AOq2gADqtoEA6raCAOq2gwDqtoQA6raFAOq2hgDqtocA6raIAOq2iQDqtooA6raLAOq2jADqto0A6raOAOq2jwDqtpAA6raRAOq2kgDqtpMA6raUAOq2lQDqtpYA6raXAOq2mADqtpkA6raaAOq2mwDqtpwA6radAOq2ngDqtp8A6ragAOq2oQDqtqIA6rajAOq2pADqtqUA6ramAOq2pwDqtqgA6rapAOq2qgDqtqsA6rasAOq2rQDqtq4A6ravAOq2sADqtrEA6rayAOq2swDqtrQA6ra1AOq2tgDqtrcA6ra4AOq2uQDqtroA6ra7AOq2vADqtr0A6ra+AOq2vwDqt4AA6reBAOq3ggDqt4MA6reEAOq3hQDqt4YA6reHAOq3iADqt4kA6reKAOq3iwDqt4wA6reNAOq3jgDqt48A6reQAOq3kQDqt5IA6reTAOq3lADqt5UA6reWAOq3lwDqt5gA6reZAOq3mgDqt5sA6recAOq3nQDqt54A6refAOq3oADqt6EA6reiAOq3owDqt6QA6relAOq3pgDqt6cA6reoAOq3qQDqt6oA6rerAOq3rADqt60A6reuAOq3rwDqt7AA6rexAOq3sgDqt7MA6re0AOq3tQDqt7YA6re3AOq3uADqt7kA6re6AOq3uwDqt7wA6re9AOq3vgDqt78A6riAAOq4gQDquIIA6riDAOq4hADquIUA6riGAOq4hwDquIgA6riJAOq4igDquIsA6riMAOq4jQDquI4A6riPAOq4kADquJEA6riSAOq4kwDquJQA6riVAOq4lgDquJcA6riYAOq4mQDquJoA6ribAOq4nADquJ0A6rieAOq4nwDquKAA6rihAOq4ogDquKMA6rikAOq4pQDquKYA6rinAOq4qADquKkA6riqAOq4qwDquKwA6ritAOq4rgDquK8A6riwAOq4sQDquLIA6rizAOq4tADquLUA6ri2AOq4twDquLgA6ri5AOq4ugDquLsA6ri8AOq4vQDquL4A6ri/AOq5gADquYEA6rmCAOq5gwDquYQA6rmFAOq5hgDquYcA6rmIAOq5iQDquYoA6rmLAOq5jADquY0A6rmOAOq5jwDquZAA6rmRAOq5kgDquZMA6rmUAOq5lQDquZYA6rmXAOq5mADquZkA6rmaAOq5mwDquZwA6rmdAOq5ngDquZ8A6rmgAOq5oQDquaIA6rmjAOq5pADquaUA6rmmAOq5pwDquagA6rmpAOq5qgDquasA6rmsAOq5rQDqua4A6rmvAOq5sADqubEA6rmyAOq5swDqubQA6rm1AOq5tgDqubcA6rm4AOq5uQDquboA6rm7AOq5vADqub0A6rm+AOq5vwDquoAA6rqBAOq6ggDquoMA6rqEAOq6hQDquoYA6rqHAOq6iADquokA6rqKAOq6iwDquowA6rqNAOq6jgDquo8A6rqQAOq6kQDqupIA6rqTAOq6lADqupUA6rqWAOq6lwDqupgA6rqZAOq6mgDqupsA6rqcAOq6nQDqup4A6rqfAOq6oADquqEA6rqiAOq6owDquqQA6rqlAOq6pgDquqcA6rqoAOq6qQDquqoA6rqrAOq6rADquq0A6rquAOq6rwDqurAA6rqxAOq6sgDqurMA6rq0AOq6tQDqurYA6rq3AOq6uADqurkA6rq6AOq6uwDqurwA6rq9AOq6vgDqur8A6ruAAOq7gQDqu4IA6ruDAOq7hADqu4UA6ruGAOq7hwDqu4gA6ruJAOq7igDqu4sA6ruMAOq7jQDqu44A6ruPAOq7kADqu5EA6ruSAOq7kwDqu5QA6ruVAOq7lgDqu5cA6ruYAOq7mQDqu5oA6rubAOq7nADqu50A6rueAOq7nwDqu6AA6ruhAOq7ogDqu6MA6rukAOq7pQDqu6YA6runAOq7qADqu6kA6ruqAOq7qwDqu6wA6rutAOq7rgDqu68A6ruwAOq7sQDqu7IA6ruzAOq7tADqu7UA6ru2AOq7twDqu7gA6ru5AOq7ugDqu7sA6ru8AOq7vQDqu74A6ru/AOq8gADqvIEA6ryCAOq8gwDqvIQA6ryFAOq8hgDqvIcA6ryIAOq8iQDqvIoA6ryLAOq8jADqvI0A6ryOAOq8jwDqvJAA6ryRAOq8kgDqvJMA6ryUAOq8lQDqvJYA6ryXAOq8mADqvJkA6ryaAOq8mwDqvJwA6rydAOq8ngDqvJ8A6rygAOq8oQDqvKIA6ryjAOq8pADqvKUA6rymAOq8pwDqvKgA6rypAOq8qgDqvKsA6rysAOq8rQDqvK4A6ryvAOq8sADqvLEA6ryyAOq8swDqvLQA6ry1AOq8tgDqvLcA6ry4AOq8uQDqvLoA6ry7AOq8vADqvL0A6ry+AOq8vwDqvYAA6r2BAOq9ggDqvYMA6r2EAOq9hQDqvYYA6r2HAOq9iADqvYkA6r2KAOq9iwDqvYwA6r2NAOq9jgDqvY8A6r2QAOq9kQDqvZIA6r2TAOq9lADqvZUA6r2WAOq9lwDqvZgA6r2ZAOq9mgDqvZsA6r2cAOq9nQDqvZ4A6r2fAOq9oADqvaEA6r2iAOq9owDqvaQA6r2lAOq9pgDqvacA6r2oAOq9qQDqvaoA6r2rAOq9rADqva0A6r2uAOq9rwDqvbAA6r2xAOq9sgDqvbMA6r20AOq9tQDqvbYA6r23AOq9uADqvbkA6r26AOq9uwDqvbwA6r29AOq9vgDqvb8A6r6AAOq+gQDqvoIA6r6DAOq+hADqvoUA6r6GAOq+hwDqvogA6r6JAOq+igDqvosA6r6MAOq+jQDqvo4A6r6PAOq+kADqvpEA6r6SAOq+kwDqvpQA6r6VAOq+lgDqvpcA6r6YAOq+mQDqvpoA6r6bAOq+nADqvp0A6r6eAOq+nwDqvqAA6r6hAOq+ogDqvqMA6r6kAOq+pQDqvqYA6r6nAOq+qADqvqkA6r6qAOq+qwDqvqwA6r6tAOq+rgDqvq8A6r6wAOq+sQDqvrIA6r6zAOq+tADqvrUA6r62AOq+twDqvrgA6r65AOq+ugDqvrsA6r68AOq+vQDqvr4A6r6/AOq/gADqv4EA6r+CAOq/gwDqv4QA6r+FAOq/hgDqv4cA6r+IAOq/iQDqv4oA6r+LAOq/jADqv40A6r+OAOq/jwDqv5AA6r+RAOq/kgDqv5MA6r+UAOq/lQDqv5YA6r+XAOq/mADqv5kA6r+aAOq/mwDqv5wA6r+dAOq/ngDqv58A6r+gAOq/oQDqv6IA6r+jAOq/pADqv6UA6r+mAOq/pwDqv6gA6r+pAOq/qgDqv6sA6r+sAOq/rQDqv64A6r+vAOq/sADqv7EA6r+yAOq/swDqv7QA6r+1AOq/tgDqv7cA6r+4AOq/uQDqv7oA6r+7AOq/vADqv70A6r++AOq/vwDrgIAA64CBAOuAggDrgIMA64CEAOuAhQDrgIYA64CHAOuAiADrgIkA64CKAOuAiwDrgIwA64CNAOuAjgDrgI8A64CQAOuAkQDrgJIA64CTAOuAlADrgJUA64CWAOuAlwDrgJgA64CZAOuAmgDrgJsA64CcAOuAnQDrgJ4A64CfAOuAoADrgKEA64CiAOuAowDrgKQA64ClAOuApgDrgKcA64CoAOuAqQDrgKoA64CrAOuArADrgK0A64CuAOuArwDrgLAA64CxAOuAsgDrgLMA64C0AOuAtQDrgLYA64C3AOuAuADrgLkA64C6AOuAuwDrgLwA64C9AOuAvgDrgL8A64GAAOuBgQDrgYIA64GDAOuBhADrgYUA64GGAOuBhwDrgYgA64GJAOuBigDrgYsA64GMAOuBjQDrgY4A64GPAOuBkADrgZEA64GSAOuBkwDrgZQA64GVAOuBlgDrgZcA64GYAOuBmQDrgZoA64GbAOuBnADrgZ0A64GeAOuBnwDrgaAA64GhAOuBogDrgaMA64GkAOuBpQDrgaYA64GnAOuBqADrgakA64GqAOuBqwDrgawA64GtAOuBrgDrga8A64GwAOuBsQDrgbIA64GzAOuBtADrgbUA64G2AOuBtwDrgbgA64G5AOuBugDrgbsA64G8AOuBvQDrgb4A64G/AOuCgADrgoEA64KCAOuCgwDrgoQA64KFAOuChgDrgocA64KIAOuCiQDrgooA64KLAOuCjADrgo0A64KOAOuCjwDrgpAA64KRAOuCkgDrgpMA64KUAOuClQDrgpYA64KXAOuCmADrgpkA64KaAOuCmwDrgpwA64KdAOuCngDrgp8A64KgAOuCoQDrgqIA64KjAOuCpADrgqUA64KmAOuCpwDrgqgA64KpAOuCqgDrgqsA64KsAOuCrQDrgq4A64KvAOuCsADrgrEA64KyAOuCswDrgrQA64K1AOuCtgDrgrcA64K4AOuCuQDrgroA64K7AOuCvADrgr0A64K+AOuCvwDrg4AA64OBAOuDggDrg4MA64OEAOuDhQDrg4YA64OHAOuDiADrg4kA64OKAOuDiwDrg4wA64ONAOuDjgDrg48A64OQAOuDkQDrg5IA64OTAOuDlADrg5UA64OWAOuDlwDrg5gA64OZAOuDmgDrg5sA64OcAOuDnQDrg54A64OfAOuDoADrg6EA64OiAOuDowDrg6QA64OlAOuDpgDrg6cA64OoAOuDqQDrg6oA64OrAOuDrADrg60A64OuAOuDrwDrg7AA64OxAOuDsgDrg7MA64O0AOuDtQDrg7YA64O3AOuDuADrg7kA64O6AOuDuwDrg7wA64O9AOuDvgDrg78A64SAAOuEgQDrhIIA64SDAOuEhADrhIUA64SGAOuEhwDrhIgA64SJAOuEigDrhIsA64SMAOuEjQDrhI4A64SPAOuEkADrhJEA64SSAOuEkwDrhJQA64SVAOuElgDrhJcA64SYAOuEmQDrhJoA64SbAOuEnADrhJ0A64SeAOuEnwDrhKAA64ShAOuEogDrhKMA64SkAOuEpQDrhKYA64SnAOuEqADrhKkA64SqAOuEqwDrhKwA64StAOuErgDrhK8A64SwAOuEsQDrhLIA64SzAOuEtADrhLUA64S2AOuEtwDrhLgA64S5AOuEugDrhLsA64S8AOuEvQDrhL4A64S/AOuFgADrhYEA64WCAOuFgwDrhYQA64WFAOuFhgDrhYcA64WIAOuFiQDrhYoA64WLAOuFjADrhY0A64WOAOuFjwDrhZAA64WRAOuFkgDrhZMA64WUAOuFlQDrhZYA64WXAOuFmADrhZkA64WaAOuFmwDrhZwA64WdAOuFngDrhZ8A64WgAOuFoQDrhaIA64WjAOuFpADrhaUA64WmAOuFpwDrhagA64WpAOuFqgDrhasA64WsAOuFrQDrha4A64WvAOuFsADrhbEA64WyAOuFswDrhbQA64W1AOuFtgDrhbcA64W4AOuFuQDrhboA64W7AOuFvADrhb0A64W+AOuFvwDrhoAA64aBAOuGggDrhoMA64aEAOuGhQDrhoYA64aHAOuGiADrhokA64aKAOuGiwDrhowA64aNAOuGjgDrho8A64aQAOuGkQDrhpIA64aTAOuGlADrhpUA64aWAOuGlwDrhpgA64aZAOuGmgDrhpsA64acAOuGnQDrhp4A64afAOuGoADrhqEA64aiAOuGowDrhqQA64alAOuGpgDrhqcA64aoAOuGqQDrhqoA64arAOuGrADrhq0A64auAOuGrwDrhrAA64axAOuGsgDrhrMA64a0AOuGtQDrhrYA64a3AOuGuADrhrkA64a6AOuGuwDrhrwA64a9AOuGvgDrhr8A64eAAOuHgQDrh4IA64eDAOuHhADrh4UA64eGAOuHhwDrh4gA64eJAOuHigDrh4sA64eMAOuHjQDrh44A64ePAOuHkADrh5EA64eSAOuHkwDrh5QA64eVAOuHlgDrh5cA64eYAOuHmQDrh5oA64ebAOuHnADrh50A64eeAOuHnwDrh6AA64ehAOuHogDrh6MA64ekAOuHpQDrh6YA64enAOuHqADrh6kA64eqAOuHqwDrh6wA64etAOuHrgDrh68A64ewAOuHsQDrh7IA64ezAOuHtADrh7UA64e2AOuHtwDrh7gA64e5AOuHugDrh7sA64e8AOuHvQDrh74A64e/AOuIgADriIEA64iCAOuIgwDriIQA64iFAOuIhgDriIcA64iIAOuIiQDriIoA64iLAOuIjADriI0A64iOAOuIjwDriJAA64iRAOuIkgDriJMA64iUAOuIlQDriJYA64iXAOuImADriJkA64iaAOuImwDriJwA64idAOuIngDriJ8A64igAOuIoQDriKIA64ijAOuIpADriKUA64imAOuIpwDriKgA64ipAOuIqgDriKsA64isAOuIrQDriK4A64ivAOuIsADriLEA64iyAOuIswDriLQA64i1AOuItgDriLcA64i4AOuIuQDriLoA64i7AOuIvADriL0A64i+AOuIvwDriYAA64mBAOuJggDriYMA64mEAOuJhQDriYYA64mHAOuJiADriYkA64mKAOuJiwDriYwA64mNAOuJjgDriY8A64mQAOuJkQDriZIA64mTAOuJlADriZUA64mWAOuJlwDriZgA64mZAOuJmgDriZsA64mcAOuJnQDriZ4A64mfAOuJoADriaEA64miAOuJowDriaQA64mlAOuJpgDriacA64moAOuJqQDriaoA64mrAOuJrADria0A64muAOuJrwDribAA64mxAOuJsgDribMA64m0AOuJtQDribYA64m3AOuJuADribkA64m6AOuJuwDribwA64m9AOuJvgDrib8A64qAAOuKgQDrioIA64qDAOuKhADrioUA64qGAOuKhwDriogA64qJAOuKigDriosA64qMAOuKjQDrio4A64qPAOuKkADripEA64qSAOuKkwDripQA64qVAOuKlgDripcA64qYAOuKmQDripoA64qbAOuKnADrip0A64qeAOuKnwDriqAA64qhAOuKogDriqMA64qkAOuKpQDriqYA64qnAOuKqADriqkA64qqAOuKqwDriqwA64qtAOuKrgDriq8A64qwAOuKsQDrirIA64qzAOuKtADrirUA64q2AOuKtwDrirgA64q5AOuKugDrirsA64q8AOuKvQDrir4A64q/AOuLgADri4EA64uCAOuLgwDri4QA64uFAOuLhgDri4cA64uIAOuLiQDri4oA64uLAOuLjADri40A64uOAOuLjwDri5AA64uRAOuLkgDri5MA64uUAOuLlQDri5YA64uXAOuLmADri5kA64uaAOuLmwDri5wA64udAOuLngDri58A64ugAOuLoQDri6IA64ujAOuLpADri6UA64umAOuLpwDri6gA64upAOuLqgDri6sA64usAOuLrQDri64A64uvAOuLsADri7EA64uyAOuLswDri7QA64u1AOuLtgDri7cA64u4AOuLuQDri7oA64u7AOuLvADri70A64u+AOuLvwDrjIAA64yBAOuMggDrjIMA64yEAOuMhQDrjIYA64yHAOuMiADrjIkA64yKAOuMiwDrjIwA64yNAOuMjgDrjI8A64yQAOuMkQDrjJIA64yTAOuMlADrjJUA64yWAOuMlwDrjJgA64yZAOuMmgDrjJsA64ycAOuMnQDrjJ4A64yfAOuMoADrjKEA64yiAOuMowDrjKQA64ylAOuMpgDrjKcA64yoAOuMqQDrjKoA64yrAOuMrADrjK0A64yuAOuMrwDrjLAA64yxAOuMsgDrjLMA64y0AOuMtQDrjLYA64y3AOuMuADrjLkA64y6AOuMuwDrjLwA64y9AOuMvgDrjL8A642AAOuNgQDrjYIA642DAOuNhADrjYUA642GAOuNhwDrjYgA642JAOuNigDrjYsA642MAOuNjQDrjY4A642PAOuNkADrjZEA642SAOuNkwDrjZQA642VAOuNlgDrjZcA642YAOuNmQDrjZoA642bAOuNnADrjZ0A642eAOuNnwDrjaAA642hAOuNogDrjaMA642kAOuNpQDrjaYA642nAOuNqADrjakA642qAOuNqwDrjawA642tAOuNrgDrja8A642wAOuNsQDrjbIA642zAOuNtADrjbUA6422AOuNtwDrjbgA6425AOuNugDrjbsA6428AOuNvQDrjb4A642/AOuOgADrjoEA646CAOuOgwDrjoQA646FAOuOhgDrjocA646IAOuOiQDrjooA646LAOuOjADrjo0A646OAOuOjwDrjpAA646RAOuOkgDrjpMA646UAOuOlQDrjpYA646XAOuOmADrjpkA646aAOuOmwDrjpwA646dAOuOngDrjp8A646gAOuOoQDrjqIA646jAOuOpADrjqUA646mAOuOpwDrjqgA646pAOuOqgDrjqsA646sAOuOrQDrjq4A646vAOuOsADrjrEA646yAOuOswDrjrQA6461AOuOtgDrjrcA6464AOuOuQDrjroA6467AOuOvADrjr0A646+AOuOvwDrj4AA64+BAOuPggDrj4MA64+EAOuPhQDrj4YA64+HAOuPiADrj4kA64+KAOuPiwDrj4wA64+NAOuPjgDrj48A64+QAOuPkQDrj5IA64+TAOuPlADrj5UA64+WAOuPlwDrj5gA64+ZAOuPmgDrj5sA64+cAOuPnQDrj54A64+fAOuPoADrj6EA64+iAOuPowDrj6QA64+lAOuPpgDrj6cA64+oAOuPqQDrj6oA64+rAOuPrADrj60A64+uAOuPrwDrj7AA64+xAOuPsgDrj7MA64+0AOuPtQDrj7YA64+3AOuPuADrj7kA64+6AOuPuwDrj7wA64+9AOuPvgDrj78A65CAAOuQgQDrkIIA65CDAOuQhADrkIUA65CGAOuQhwDrkIgA65CJAOuQigDrkIsA65CMAOuQjQDrkI4A65CPAOuQkADrkJEA65CSAOuQkwDrkJQA65CVAOuQlgDrkJcA65CYAOuQmQDrkJoA65CbAOuQnADrkJ0A65CeAOuQnwDrkKAA65ChAOuQogDrkKMA65CkAOuQpQDrkKYA65CnAOuQqADrkKkA65CqAOuQqwDrkKwA65CtAOuQrgDrkK8A65CwAOuQsQDrkLIA65CzAOuQtADrkLUA65C2AOuQtwDrkLgA65C5AOuQugDrkLsA65C8AOuQvQDrkL4A65C/AOuRgADrkYEA65GCAOuRgwDrkYQA65GFAOuRhgDrkYcA65GIAOuRiQDrkYoA65GLAOuRjADrkY0A65GOAOuRjwDrkZAA65GRAOuRkgDrkZMA65GUAOuRlQDrkZYA65GXAOuRmADrkZkA65GaAOuRmwDrkZwA65GdAOuRngDrkZ8A65GgAOuRoQDrkaIA65GjAOuRpADrkaUA65GmAOuRpwDrkagA65GpAOuRqgDrkasA65GsAOuRrQDrka4A65GvAOuRsADrkbEA65GyAOuRswDrkbQA65G1AOuRtgDrkbcA65G4AOuRuQDrkboA65G7AOuRvADrkb0A65G+AOuRvwDrkoAA65KBAOuSggDrkoMA65KEAOuShQDrkoYA65KHAOuSiADrkokA65KKAOuSiwDrkowA65KNAOuSjgDrko8A65KQAOuSkQDrkpIA65KTAOuSlADrkpUA65KWAOuSlwDrkpgA65KZAOuSmgDrkpsA65KcAOuSnQDrkp4A65KfAOuSoADrkqEA65KiAOuSowDrkqQA65KlAOuSpgDrkqcA65KoAOuSqQDrkqoA65KrAOuSrADrkq0A65KuAOuSrwDrkrAA65KxAOuSsgDrkrMA65K0AOuStQDrkrYA65K3AOuSuADrkrkA65K6AOuSuwDrkrwA65K9AOuSvgDrkr8A65OAAOuTgQDrk4IA65ODAOuThADrk4UA65OGAOuThwDrk4gA65OJAOuTigDrk4sA65OMAOuTjQDrk44A65OPAOuTkADrk5EA65OSAOuTkwDrk5QA65OVAOuTlgDrk5cA65OYAOuTmQDrk5oA65ObAOuTnADrk50A65OeAOuTnwDrk6AA65OhAOuTogDrk6MA65OkAOuTpQDrk6YA65OnAOuTqADrk6kA65OqAOuTqwDrk6wA65OtAOuTrgDrk68A65OwAOuTsQDrk7IA65OzAOuTtADrk7UA65O2AOuTtwDrk7gA65O5AOuTugDrk7sA65O8AOuTvQDrk74A65O/AOuUgADrlIEA65SCAOuUgwDrlIQA65SFAOuUhgDrlIcA65SIAOuUiQDrlIoA65SLAOuUjADrlI0A65SOAOuUjwDrlJAA65SRAOuUkgDrlJMA65SUAOuUlQDrlJYA65SXAOuUmADrlJkA65SaAOuUmwDrlJwA65SdAOuUngDrlJ8A65SgAOuUoQDrlKIA65SjAOuUpADrlKUA65SmAOuUpwDrlKgA65SpAOuUqgDrlKsA65SsAOuUrQDrlK4A65SvAOuUsADrlLEA65SyAOuUswDrlLQA65S1AOuUtgDrlLcA65S4AOuUuQDrlLoA65S7AOuUvADrlL0A65S+AOuUvwDrlYAA65WBAOuVggDrlYMA65WEAOuVhQDrlYYA65WHAOuViADrlYkA65WKAOuViwDrlYwA65WNAOuVjgDrlY8A65WQAOuVkQDrlZIA65WTAOuVlADrlZUA65WWAOuVlwDrlZgA65WZAOuVmgDrlZsA65WcAOuVnQDrlZ4A65WfAOuVoADrlaEA65WiAOuVowDrlaQA65WlAOuVpgDrlacA65WoAOuVqQDrlaoA65WrAOuVrADrla0A65WuAOuVrwDrlbAA65WxAOuVsgDrlbMA65W0AOuVtQDrlbYA65W3AOuVuADrlbkA65W6AOuVuwDrlbwA65W9AOuVvgDrlb8A65aAAOuWgQDrloIA65aDAOuWhADrloUA65aGAOuWhwDrlogA65aJAOuWigDrlosA65aMAOuWjQDrlo4A65aPAOuWkADrlpEA65aSAOuWkwDrlpQA65aVAOuWlgDrlpcA65aYAOuWmQDrlpoA65abAOuWnADrlp0A65aeAOuWnwDrlqAA65ahAOuWogDrlqMA65akAOuWpQDrlqYA65anAOuWqADrlqkA65aqAOuWqwDrlqwA65atAOuWrgDrlq8A65awAOuWsQDrlrIA65azAOuWtADrlrUA65a2AOuWtwDrlrgA65a5AOuWugDrlrsA65a8AOuWvQDrlr4A65a/AOuXgADrl4EA65eCAOuXgwDrl4QA65eFAOuXhgDrl4cA65eIAOuXiQDrl4oA65eLAOuXjADrl40A65eOAOuXjwDrl5AA65eRAOuXkgDrl5MA65eUAOuXlQDrl5YA65eXAOuXmADrl5kA65eaAOuXmwDrl5wA65edAOuXngDrl58A65egAOuXoQDrl6IA65ejAOuXpADrl6UA65emAOuXpwDrl6gA65epAOuXqgDrl6sA65esAOuXrQDrl64A65evAOuXsADrl7EA65eyAOuXswDrl7QA65e1AOuXtgDrl7cA65e4AOuXuQDrl7oA65e7AOuXvADrl70A65e+AOuXvwDrmIAA65iBAOuYggDrmIMA65iEAOuYhQDrmIYA65iHAOuYiADrmIkA65iKAOuYiwDrmIwA65iNAOuYjgDrmI8A65iQAOuYkQDrmJIA65iTAOuYlADrmJUA65iWAOuYlwDrmJgA65iZAOuYmgDrmJsA65icAOuYnQDrmJ4A65ifAOuYoADrmKEA65iiAOuYowDrmKQA65ilAOuYpgDrmKcA65ioAOuYqQDrmKoA65irAOuYrADrmK0A65iuAOuYrwDrmLAA65ixAOuYsgDrmLMA65i0AOuYtQDrmLYA65i3AOuYuADrmLkA65i6AOuYuwDrmLwA65i9AOuYvgDrmL8A65mAAOuZgQDrmYIA65mDAOuZhADrmYUA65mGAOuZhwDrmYgA65mJAOuZigDrmYsA65mMAOuZjQDrmY4A65mPAOuZkADrmZEA65mSAOuZkwDrmZQA65mVAOuZlgDrmZcA65mYAOuZmQDrmZoA65mbAOuZnADrmZ0A65meAOuZnwDrmaAA65mhAOuZogDrmaMA65mkAOuZpQDrmaYA65mnAOuZqADrmakA65mqAOuZqwDrmawA65mtAOuZrgDrma8A65mwAOuZsQDrmbIA65mzAOuZtADrmbUA65m2AOuZtwDrmbgA65m5AOuZugDrmbsA65m8AOuZvQDrmb4A65m/AOuagADrmoEA65qCAOuagwDrmoQA65qFAOuahgDrmocA65qIAOuaiQDrmooA65qLAOuajADrmo0A65qOAOuajwDrmpAA65qRAOuakgDrmpMA65qUAOualQDrmpYA65qXAOuamADrmpkA65qaAOuamwDrmpwA65qdAOuangDrmp8A65qgAOuaoQDrmqIA65qjAOuapADrmqUA65qmAOuapwDrmqgA65qpAOuaqgDrmqsA65qsAOuarQDrmq4A65qvAOuasADrmrEA65qyAOuaswDrmrQA65q1AOuatgDrmrcA65q4AOuauQDrmroA65q7AOuavADrmr0A65q+AOuavwDrm4AA65uBAOubggDrm4MA65uEAOubhQDrm4YA65uHAOubiADrm4kA65uKAOubiwDrm4wA65uNAOubjgDrm48A65uQAOubkQDrm5IA65uTAOublADrm5UA65uWAOublwDrm5gA65uZAOubmgDrm5sA65ucAOubnQDrm54A65ufAOuboADrm6EA65uiAOubowDrm6QA65ulAOubpgDrm6cA65uoAOubqQDrm6oA65urAOubrADrm60A65uuAOubrwDrm7AA65uxAOubsgDrm7MA65u0AOubtQDrm7YA65u3AOubuADrm7kA65u6AOubuwDrm7wA65u9AOubvgDrm78A65yAAOucgQDrnIIA65yDAOuchADrnIUA65yGAOuchwDrnIgA65yJAOucigDrnIsA65yMAOucjQDrnI4A65yPAOuckADrnJEA65ySAOuckwDrnJQA65yVAOuclgDrnJcA65yYAOucmQDrnJoA65ybAOucnADrnJ0A65yeAOucnwDrnKAA65yhAOucogDrnKMA65ykAOucpQDrnKYA65ynAOucqADrnKkA65yqAOucqwDrnKwA65ytAOucrgDrnK8A65ywAOucsQDrnLIA65yzAOuctADrnLUA65y2AOuctwDrnLgA65y5AOucugDrnLsA65y8AOucvQDrnL4A65y/AOudgADrnYEA652CAOudgwDrnYQA652FAOudhgDrnYcA652IAOudiQDrnYoA652LAOudjADrnY0A652OAOudjwDrnZAA652RAOudkgDrnZMA652UAOudlQDrnZYA652XAOudmADrnZkA652aAOudmwDrnZwA652dAOudngDrnZ8A652gAOudoQDrnaIA652jAOudpADrnaUA652mAOudpwDrnagA652pAOudqgDrnasA652sAOudrQDrna4A652vAOudsADrnbEA652yAOudswDrnbQA6521AOudtgDrnbcA6524AOuduQDrnboA6527AOudvADrnb0A652+AOudvwDrnoAA656BAOueggDrnoMA656EAOuehQDrnoYA656HAOueiADrnokA656KAOueiwDrnowA656NAOuejgDrno8A656QAOuekQDrnpIA656TAOuelADrnpUA656WAOuelwDrnpgA656ZAOuemgDrnpsA656cAOuenQDrnp4A656fAOueoADrnqEA656iAOueowDrnqQA656lAOuepgDrnqcA656oAOueqQDrnqoA656rAOuerADrnq0A656uAOuerwDrnrAA656xAOuesgDrnrMA6560AOuetQDrnrYA6563AOueuADrnrkA6566AOueuwDrnrwA6569AOuevgDrnr8A65+AAOufgQDrn4IA65+DAOufhADrn4UA65+GAOufhwDrn4gA65+JAOufigDrn4sA65+MAOufjQDrn44A65+PAOufkADrn5EA65+SAOufkwDrn5QA65+VAOuflgDrn5cA65+YAOufmQDrn5oA65+bAOufnADrn50A65+eAOufnwDrn6AA65+hAOufogDrn6MA65+kAOufpQDrn6YA65+nAOufqADrn6kA65+qAOufqwDrn6wA65+tAOufrgDrn68A65+wAOufsQDrn7IA65+zAOuftADrn7UA65+2AOuftwDrn7gA65+5AOufugDrn7sA65+8AOufvQDrn74A65+/AOuggADroIEA66CCAOuggwDroIQA66CFAOughgDroIcA66CIAOugiQDroIoA66CLAOugjADroI0A66COAOugjwDroJAA66CRAOugkgDroJMA66CUAOuglQDroJYA66CXAOugmADroJkA66CaAOugmwDroJwA66CdAOugngDroJ8A66CgAOugoQDroKIA66CjAOugpADroKUA66CmAOugpwDroKgA66CpAOugqgDroKsA66CsAOugrQDroK4A66CvAOugsADroLEA66CyAOugswDroLQA66C1AOugtgDroLcA66C4AOuguQDroLoA66C7AOugvADroL0A66C+AOugvwDroYAA66GBAOuhggDroYMA66GEAOuhhQDroYYA66GHAOuhiADroYkA66GKAOuhiwDroYwA66GNAOuhjgDroY8A66GQAOuhkQDroZIA66GTAOuhlADroZUA66GWAOuhlwDroZgA66GZAOuhmgDroZsA66GcAOuhnQDroZ4A66GfAOuhoADroaEA66GiAOuhowDroaQA66GlAOuhpgDroacA66GoAOuhqQDroaoA66GrAOuhrADroa0A66GuAOuhrwDrobAA66GxAOuhsgDrobMA66G0AOuhtQDrobYA66G3AOuhuADrobkA66G6AOuhuwDrobwA66G9AOuhvgDrob8A66KAAOuigQDrooIA66KDAOuihADrooUA66KGAOuihwDroogA66KJAOuiigDroosA66KMAOuijQDroo4A66KPAOuikADropEA66KSAOuikwDropQA66KVAOuilgDropcA66KYAOuimQDropoA66KbAOuinADrop0A66KeAOuinwDroqAA66KhAOuiogDroqMA66KkAOuipQDroqYA66KnAOuiqADroqkA66KqAOuiqwDroqwA66KtAOuirgDroq8A66KwAOuisQDrorIA66KzAOuitADrorUA66K2AOuitwDrorgA66K5AOuiugDrorsA66K8AOuivQDror4A66K/AOujgADro4EA66OCAOujgwDro4QA66OFAOujhgDro4cA66OIAOujiQDro4oA66OLAOujjADro40A66OOAOujjwDro5AA66ORAOujkgDro5MA66OUAOujlQDro5YA66OXAOujmADro5kA66OaAOujmwDro5wA66OdAOujngDro58A66OgAOujoQDro6IA66OjAOujpADro6UA66OmAOujpwDro6gA66OpAOujqgDro6sA66OsAOujrQDro64A66OvAOujsADro7EA66OyAOujswDro7QA66O1AOujtgDro7cA66O4AOujuQDro7oA66O7AOujvADro70A66O+AOujvwDrpIAA66SBAOukggDrpIMA66SEAOukhQDrpIYA66SHAOukiADrpIkA66SKAOukiwDrpIwA66SNAOukjgDrpI8A66SQAOukkQDrpJIA66STAOuklADrpJUA66SWAOuklwDrpJgA66SZAOukmgDrpJsA66ScAOuknQDrpJ4A66SfAOukoADrpKEA66SiAOukowDrpKQA66SlAOukpgDrpKcA66SoAOukqQDrpKoA66SrAOukrADrpK0A66SuAOukrwDrpLAA66SxAOuksgDrpLMA66S0AOuktQDrpLYA66S3AOukuADrpLkA66S6AOukuwDrpLwA66S9AOukvgDrpL8A66WAAOulgQDrpYIA66WDAOulhADrpYUA66WGAOulhwDrpYgA66WJAOuligDrpYsA66WMAOuljQDrpY4A66WPAOulkADrpZEA66WSAOulkwDrpZQA66WVAOullgDrpZcA66WYAOulmQDrpZoA66WbAOulnADrpZ0A66WeAOulnwDrpaAA66WhAOulogDrpaMA66WkAOulpQDrpaYA66WnAOulqADrpakA66WqAOulqwDrpawA66WtAOulrgDrpa8A66WwAOulsQDrpbIA66WzAOultADrpbUA66W2AOultwDrpbgA66W5AOulugDrpbsA66W8AOulvQDrpb4A66W/AOumgADrpoEA66aCAOumgwDrpoQA66aFAOumhgDrpocA66aIAOumiQDrpooA66aLAOumjADrpo0A66aOAOumjwDrppAA66aRAOumkgDrppMA66aUAOumlQDrppYA66aXAOummADrppkA66aaAOummwDrppwA66adAOumngDrpp8A66agAOumoQDrpqIA66ajAOumpADrpqUA66amAOumpwDrpqgA66apAOumqgDrpqsA66asAOumrQDrpq4A66avAOumsADrprEA66ayAOumswDrprQA66a1AOumtgDrprcA66a4AOumuQDrproA66a7AOumvADrpr0A66a+AOumvwDrp4AA66eBAOunggDrp4MA66eEAOunhQDrp4YA66eHAOuniADrp4kA66eKAOuniwDrp4wA66eNAOunjgDrp48A66eQAOunkQDrp5IA66eTAOunlADrp5UA66eWAOunlwDrp5gA66eZAOunmgDrp5sA66ecAOunnQDrp54A66efAOunoADrp6EA66eiAOunowDrp6QA66elAOunpgDrp6cA66eoAOunqQDrp6oA66erAOunrADrp60A66euAOunrwDrp7AA66exAOunsgDrp7MA66e0AOuntQDrp7YA66e3AOunuADrp7kA66e6AOunuwDrp7wA66e9AOunvgDrp78A66iAAOuogQDrqIIA66iDAOuohADrqIUA66iGAOuohwDrqIgA66iJAOuoigDrqIsA66iMAOuojQDrqI4A66iPAOuokADrqJEA66iSAOuokwDrqJQA66iVAOuolgDrqJcA66iYAOuomQDrqJoA66ibAOuonADrqJ0A66ieAOuonwDrqKAA66ihAOuoogDrqKMA66ikAOuopQDrqKYA66inAOuoqADrqKkA66iqAOuoqwDrqKwA66itAOuorgDrqK8A66iwAOuosQDrqLIA66izAOuotADrqLUA66i2AOuotwDrqLgA66i5AOuougDrqLsA66i8AOuovQDrqL4A66i/AOupgADrqYEA66mCAOupgwDrqYQA66mFAOuphgDrqYcA66mIAOupiQDrqYoA66mLAOupjADrqY0A66mOAOupjwDrqZAA66mRAOupkgDrqZMA66mUAOuplQDrqZYA66mXAOupmADrqZkA66maAOupmwDrqZwA66mdAOupngDrqZ8A66mgAOupoQDrqaIA66mjAOuppADrqaUA66mmAOuppwDrqagA66mpAOupqgDrqasA66msAOuprQDrqa4A66mvAOupsADrqbEA66myAOupswDrqbQA66m1AOuptgDrqbcA66m4AOupuQDrqboA66m7AOupvADrqb0A66m+AOupvwDrqoAA66qBAOuqggDrqoMA66qEAOuqhQDrqoYA66qHAOuqiADrqokA66qKAOuqiwDrqowA66qNAOuqjgDrqo8A66qQAOuqkQDrqpIA66qTAOuqlADrqpUA66qWAOuqlwDrqpgA66qZAOuqmgDrqpsA66qcAOuqnQDrqp4A66qfAOuqoADrqqEA66qiAOuqowDrqqQA66qlAOuqpgDrqqcA66qoAOuqqQDrqqoA66qrAOuqrADrqq0A66quAOuqrwDrqrAA66qxAOuqsgDrqrMA66q0AOuqtQDrqrYA66q3AOuquADrqrkA66q6AOuquwDrqrwA66q9AOuqvgDrqr8A66uAAOurgQDrq4IA66uDAOurhADrq4UA66uGAOurhwDrq4gA66uJAOurigDrq4sA66uMAOurjQDrq44A66uPAOurkADrq5EA66uSAOurkwDrq5QA66uVAOurlgDrq5cA66uYAOurmQDrq5oA66ubAOurnADrq50A66ueAOurnwDrq6AA66uhAOurogDrq6MA66ukAOurpQDrq6YA66unAOurqADrq6kA66uqAOurqwDrq6wA66utAOurrgDrq68A66uwAOursQDrq7IA66uzAOurtADrq7UA66u2AOurtwDrq7gA66u5AOurugDrq7sA66u8AOurvQDrq74A66u/AOusgADrrIEA66yCAOusgwDrrIQA66yFAOushgDrrIcA66yIAOusiQDrrIoA66yLAOusjADrrI0A66yOAOusjwDrrJAA66yRAOuskgDrrJMA66yUAOuslQDrrJYA66yXAOusmADrrJkA66yaAOusmwDrrJwA66ydAOusngDrrJ8A66ygAOusoQDrrKIA66yjAOuspADrrKUA66ymAOuspwDrrKgA66ypAOusqgDrrKsA66ysAOusrQDrrK4A66yvAOussADrrLEA66yyAOusswDrrLQA66y1AOustgDrrLcA66y4AOusuQDrrLoA66y7AOusvADrrL0A66y+AOusvwDrrYAA662BAOutggDrrYMA662EAOuthQDrrYYA662HAOutiADrrYkA662KAOutiwDrrYwA662NAOutjgDrrY8A662QAOutkQDrrZIA662TAOutlADrrZUA662WAOutlwDrrZgA662ZAOutmgDrrZsA662cAOutnQDrrZ4A662fAOutoADrraEA662iAOutowDrraQA662lAOutpgDrracA662oAOutqQDrraoA662rAOutrADrra0A662uAOutrwDrrbAA662xAOutsgDrrbMA6620AOuttQDrrbYA6623AOutuADrrbkA6626AOutuwDrrbwA6629AOutvgDrrb8A666AAOuugQDrroIA666DAOuuhADrroUA666GAOuuhwDrrogA666JAOuuigDrrosA666MAOuujQDrro4A666PAOuukADrrpEA666SAOuukwDrrpQA666VAOuulgDrrpcA666YAOuumQDrrpoA666bAOuunADrrp0A666eAOuunwDrrqAA666hAOuuogDrrqMA666kAOuupQDrrqYA666nAOuuqADrrqkA666qAOuuqwDrrqwA666tAOuurgDrrq8A666wAOuusQDrrrIA666zAOuutADrrrUA6662AOuutwDrrrgA6665AOuuugDrrrsA6668AOuuvQDrrr4A666/AOuvgADrr4EA66+CAOuvgwDrr4QA66+FAOuvhgDrr4cA66+IAOuviQDrr4oA66+LAOuvjADrr40A66+OAOuvjwDrr5AA66+RAOuvkgDrr5MA66+UAOuvlQDrr5YA66+XAOuvmADrr5kA66+aAOuvmwDrr5wA66+dAOuvngDrr58A66+gAOuvoQDrr6IA66+jAOuvpADrr6UA66+mAOuvpwDrr6gA66+pAOuvqgDrr6sA66+sAOuvrQDrr64A66+vAOuvsADrr7EA66+yAOuvswDrr7QA66+1AOuvtgDrr7cA66+4AOuvuQDrr7oA66+7AOuvvADrr70A66++AOuvvwDrsIAA67CBAOuwggDrsIMA67CEAOuwhQDrsIYA67CHAOuwiADrsIkA67CKAOuwiwDrsIwA67CNAOuwjgDrsI8A67CQAOuwkQDrsJIA67CTAOuwlADrsJUA67CWAOuwlwDrsJgA67CZAOuwmgDrsJsA67CcAOuwnQDrsJ4A67CfAOuwoADrsKEA67CiAOuwowDrsKQA67ClAOuwpgDrsKcA67CoAOuwqQDrsKoA67CrAOuwrADrsK0A67CuAOuwrwDrsLAA67CxAOuwsgDrsLMA67C0AOuwtQDrsLYA67C3AOuwuADrsLkA67C6AOuwuwDrsLwA67C9AOuwvgDrsL8A67GAAOuxgQDrsYIA67GDAOuxhADrsYUA67GGAOuxhwDrsYgA67GJAOuxigDrsYsA67GMAOuxjQDrsY4A67GPAOuxkADrsZEA67GSAOuxkwDrsZQA67GVAOuxlgDrsZcA67GYAOuxmQDrsZoA67GbAOuxnADrsZ0A67GeAOuxnwDrsaAA67GhAOuxogDrsaMA67GkAOuxpQDrsaYA67GnAOuxqADrsakA67GqAOuxqwDrsawA67GtAOuxrgDrsa8A67GwAOuxsQDrsbIA67GzAOuxtADrsbUA67G2AOuxtwDrsbgA67G5AOuxugDrsbsA67G8AOuxvQDrsb4A67G/AOuygADrsoEA67KCAOuygwDrsoQA67KFAOuyhgDrsocA67KIAOuyiQDrsooA67KLAOuyjADrso0A67KOAOuyjwDrspAA67KRAOuykgDrspMA67KUAOuylQDrspYA67KXAOuymADrspkA67KaAOuymwDrspwA67KdAOuyngDrsp8A67KgAOuyoQDrsqIA67KjAOuypADrsqUA67KmAOuypwDrsqgA67KpAOuyqgDrsqsA67KsAOuyrQDrsq4A67KvAOuysADrsrEA67KyAOuyswDrsrQA67K1AOuytgDrsrcA67K4AOuyuQDrsroA67K7AOuyvADrsr0A67K+AOuyvwDrs4AA67OBAOuzggDrs4MA67OEAOuzhQDrs4YA67OHAOuziADrs4kA67OKAOuziwDrs4wA67ONAOuzjgDrs48A67OQAOuzkQDrs5IA67OTAOuzlADrs5UA67OWAOuzlwDrs5gA67OZAOuzmgDrs5sA67OcAOuznQDrs54A67OfAOuzoADrs6EA67OiAOuzowDrs6QA67OlAOuzpgDrs6cA67OoAOuzqQDrs6oA67OrAOuzrADrs60A67OuAOuzrwDrs7AA67OxAOuzsgDrs7MA67O0AOuztQDrs7YA67O3AOuzuADrs7kA67O6AOuzuwDrs7wA67O9AOuzvgDrs78A67SAAOu0gQDrtIIA67SDAOu0hADrtIUA67SGAOu0hwDrtIgA67SJAOu0igDrtIsA67SMAOu0jQDrtI4A67SPAOu0kADrtJEA67SSAOu0kwDrtJQA67SVAOu0lgDrtJcA67SYAOu0mQDrtJoA67SbAOu0nADrtJ0A67SeAOu0nwDrtKAA67ShAOu0ogDrtKMA67SkAOu0pQDrtKYA67SnAOu0qADrtKkA67SqAOu0qwDrtKwA67StAOu0rgDrtK8A67SwAOu0sQDrtLIA67SzAOu0tADrtLUA67S2AOu0twDrtLgA67S5AOu0ugDrtLsA67S8AOu0vQDrtL4A67S/AOu1gADrtYEA67WCAOu1gwDrtYQA67WFAOu1hgDrtYcA67WIAOu1iQDrtYoA67WLAOu1jADrtY0A67WOAOu1jwDrtZAA67WRAOu1kgDrtZMA67WUAOu1lQDrtZYA67WXAOu1mADrtZkA67WaAOu1mwDrtZwA67WdAOu1ngDrtZ8A67WgAOu1oQDrtaIA67WjAOu1pADrtaUA67WmAOu1pwDrtagA67WpAOu1qgDrtasA67WsAOu1rQDrta4A67WvAOu1sADrtbEA67WyAOu1swDrtbQA67W1AOu1tgDrtbcA67W4AOu1uQDrtboA67W7AOu1vADrtb0A67W+AOu1vwDrtoAA67aBAOu2ggDrtoMA67aEAOu2hQDrtoYA67aHAOu2iADrtokA67aKAOu2iwDrtowA67aNAOu2jgDrto8A67aQAOu2kQDrtpIA67aTAOu2lADrtpUA67aWAOu2lwDrtpgA67aZAOu2mgDrtpsA67acAOu2nQDrtp4A67afAOu2oADrtqEA67aiAOu2owDrtqQA67alAOu2pgDrtqcA67aoAOu2qQDrtqoA67arAOu2rADrtq0A67auAOu2rwDrtrAA67axAOu2sgDrtrMA67a0AOu2tQDrtrYA67a3AOu2uADrtrkA67a6AOu2uwDrtrwA67a9AOu2vgDrtr8A67eAAOu3gQDrt4IA67eDAOu3hADrt4UA67eGAOu3hwDrt4gA67eJAOu3igDrt4sA67eMAOu3jQDrt44A67ePAOu3kADrt5EA67eSAOu3kwDrt5QA67eVAOu3lgDrt5cA67eYAOu3mQDrt5oA67ebAOu3nADrt50A67eeAOu3nwDrt6AA67ehAOu3ogDrt6MA67ekAOu3pQDrt6YA67enAOu3qADrt6kA67eqAOu3qwDrt6wA67etAOu3rgDrt68A67ewAOu3sQDrt7IA67ezAOu3tADrt7UA67e2AOu3twDrt7gA67e5AOu3ugDrt7sA67e8AOu3vQDrt74A67e/AOu4gADruIEA67iCAOu4gwDruIQA67iFAOu4hgDruIcA67iIAOu4iQDruIoA67iLAOu4jADruI0A67iOAOu4jwDruJAA67iRAOu4kgDruJMA67iUAOu4lQDruJYA67iXAOu4mADruJkA67iaAOu4mwDruJwA67idAOu4ngDruJ8A67igAOu4oQDruKIA67ijAOu4pADruKUA67imAOu4pwDruKgA67ipAOu4qgDruKsA67isAOu4rQDruK4A67ivAOu4sADruLEA67iyAOu4swDruLQA67i1AOu4tgDruLcA67i4AOu4uQDruLoA67i7AOu4vADruL0A67i+AOu4vwDruYAA67mBAOu5ggDruYMA67mEAOu5hQDruYYA67mHAOu5iADruYkA67mKAOu5iwDruYwA67mNAOu5jgDruY8A67mQAOu5kQDruZIA67mTAOu5lADruZUA67mWAOu5lwDruZgA67mZAOu5mgDruZsA67mcAOu5nQDruZ4A67mfAOu5oADruaEA67miAOu5owDruaQA67mlAOu5pgDruacA67moAOu5qQDruaoA67mrAOu5rADrua0A67muAOu5rwDrubAA67mxAOu5sgDrubMA67m0AOu5tQDrubYA67m3AOu5uADrubkA67m6AOu5uwDrubwA67m9AOu5vgDrub8A67qAAOu6gQDruoIA67qDAOu6hADruoUA67qGAOu6hwDruogA67qJAOu6igDruosA67qMAOu6jQDruo4A67qPAOu6kADrupEA67qSAOu6kwDrupQA67qVAOu6lgDrupcA67qYAOu6mQDrupoA67qbAOu6nADrup0A67qeAOu6nwDruqAA67qhAOu6ogDruqMA67qkAOu6pQDruqYA67qnAOu6qADruqkA67qqAOu6qwDruqwA67qtAOu6rgDruq8A67qwAOu6sQDrurIA67qzAOu6tADrurUA67q2AOu6twDrurgA67q5AOu6ugDrursA67q8AOu6vQDrur4A67q/AOu7gADru4EA67uCAOu7gwDru4QA67uFAOu7hgDru4cA67uIAOu7iQDru4oA67uLAOu7jADru40A67uOAOu7jwDru5AA67uRAOu7kgDru5MA67uUAOu7lQDru5YA67uXAOu7mADru5kA67uaAOu7mwDru5wA67udAOu7ngDru58A67ugAOu7oQDru6IA67ujAOu7pADru6UA67umAOu7pwDru6gA67upAOu7qgDru6sA67usAOu7rQDru64A67uvAOu7sADru7EA67uyAOu7swDru7QA67u1AOu7tgDru7cA67u4AOu7uQDru7oA67u7AOu7vADru70A67u+AOu7vwDrvIAA67yBAOu8ggDrvIMA67yEAOu8hQDrvIYA67yHAOu8iADrvIkA67yKAOu8iwDrvIwA67yNAOu8jgDrvI8A67yQAOu8kQDrvJIA67yTAOu8lADrvJUA67yWAOu8lwDrvJgA67yZAOu8mgDrvJsA67ycAOu8nQDrvJ4A67yfAOu8oADrvKEA67yiAOu8owDrvKQA67ylAOu8pgDrvKcA67yoAOu8qQDrvKoA67yrAOu8rADrvK0A67yuAOu8rwDrvLAA67yxAOu8sgDrvLMA67y0AOu8tQDrvLYA67y3AOu8uADrvLkA67y6AOu8uwDrvLwA67y9AOu8vgDrvL8A672AAOu9gQDrvYIA672DAOu9hADrvYUA672GAOu9hwDrvYgA672JAOu9igDrvYsA672MAOu9jQDrvY4A672PAOu9kADrvZEA672SAOu9kwDrvZQA672VAOu9lgDrvZcA672YAOu9mQDrvZoA672bAOu9nADrvZ0A672eAOu9nwDrvaAA672hAOu9ogDrvaMA672kAOu9pQDrvaYA672nAOu9qADrvakA672qAOu9qwDrvawA672tAOu9rgDrva8A672wAOu9sQDrvbIA672zAOu9tADrvbUA6722AOu9twDrvbgA6725AOu9ugDrvbsA6728AOu9vQDrvb4A672/AOu+gADrvoEA676CAOu+gwDrvoQA676FAOu+hgDrvocA676IAOu+iQDrvooA676LAOu+jADrvo0A676OAOu+jwDrvpAA676RAOu+kgDrvpMA676UAOu+lQDrvpYA676XAOu+mADrvpkA676aAOu+mwDrvpwA676dAOu+ngDrvp8A676gAOu+oQDrvqIA676jAOu+pADrvqUA676mAOu+pwDrvqgA676pAOu+qgDrvqsA676sAOu+rQDrvq4A676vAOu+sADrvrEA676yAOu+swDrvrQA6761AOu+tgDrvrcA6764AOu+uQDrvroA6767AOu+vADrvr0A676+AOu+vwDrv4AA67+BAOu/ggDrv4MA67+EAOu/hQDrv4YA67+HAOu/iADrv4kA67+KAOu/iwDrv4wA67+NAOu/jgDrv48A67+QAOu/kQDrv5IA67+TAOu/lADrv5UA67+WAOu/lwDrv5gA67+ZAOu/mgDrv5sA67+cAOu/nQDrv54A67+fAOu/oADrv6EA67+iAOu/owDrv6QA67+lAOu/pgDrv6cA67+oAOu/qQDrv6oA67+rAOu/rADrv60A67+uAOu/rwDrv7AA67+xAOu/sgDrv7MA67+0AOu/tQDrv7YA67+3AOu/uADrv7kA67+6AOu/uwDrv7wA67+9AOu/vgDrv78A7ICAAOyAgQDsgIIA7ICDAOyAhADsgIUA7ICGAOyAhwDsgIgA7ICJAOyAigDsgIsA7ICMAOyAjQDsgI4A7ICPAOyAkADsgJEA7ICSAOyAkwDsgJQA7ICVAOyAlgDsgJcA7ICYAOyAmQDsgJoA7ICbAOyAnADsgJ0A7ICeAOyAnwDsgKAA7IChAOyAogDsgKMA7ICkAOyApQDsgKYA7ICnAOyAqADsgKkA7ICqAOyAqwDsgKwA7ICtAOyArgDsgK8A7ICwAOyAsQDsgLIA7ICzAOyAtADsgLUA7IC2AOyAtwDsgLgA7IC5AOyAugDsgLsA7IC8AOyAvQDsgL4A7IC/AOyBgADsgYEA7IGCAOyBgwDsgYQA7IGFAOyBhgDsgYcA7IGIAOyBiQDsgYoA7IGLAOyBjADsgY0A7IGOAOyBjwDsgZAA7IGRAOyBkgDsgZMA7IGUAOyBlQDsgZYA7IGXAOyBmADsgZkA7IGaAOyBmwDsgZwA7IGdAOyBngDsgZ8A7IGgAOyBoQDsgaIA7IGjAOyBpADsgaUA7IGmAOyBpwDsgagA7IGpAOyBqgDsgasA7IGsAOyBrQDsga4A7IGvAOyBsADsgbEA7IGyAOyBswDsgbQA7IG1AOyBtgDsgbcA7IG4AOyBuQDsgboA7IG7AOyBvADsgb0A7IG+AOyBvwDsgoAA7IKBAOyCggDsgoMA7IKEAOyChQDsgoYA7IKHAOyCiADsgokA7IKKAOyCiwDsgowA7IKNAOyCjgDsgo8A7IKQAOyCkQDsgpIA7IKTAOyClADsgpUA7IKWAOyClwDsgpgA7IKZAOyCmgDsgpsA7IKcAOyCnQDsgp4A7IKfAOyCoADsgqEA7IKiAOyCowDsgqQA7IKlAOyCpgDsgqcA7IKoAOyCqQDsgqoA7IKrAOyCrADsgq0A7IKuAOyCrwDsgrAA7IKxAOyCsgDsgrMA7IK0AOyCtQDsgrYA7IK3AOyCuADsgrkA7IK6AOyCuwDsgrwA7IK9AOyCvgDsgr8A7IOAAOyDgQDsg4IA7IODAOyDhADsg4UA7IOGAOyDhwDsg4gA7IOJAOyDigDsg4sA7IOMAOyDjQDsg44A7IOPAOyDkADsg5EA7IOSAOyDkwDsg5QA7IOVAOyDlgDsg5cA7IOYAOyDmQDsg5oA7IObAOyDnADsg50A7IOeAOyDnwDsg6AA7IOhAOyDogDsg6MA7IOkAOyDpQDsg6YA7IOnAOyDqADsg6kA7IOqAOyDqwDsg6wA7IOtAOyDrgDsg68A7IOwAOyDsQDsg7IA7IOzAOyDtADsg7UA7IO2AOyDtwDsg7gA7IO5AOyDugDsg7sA7IO8AOyDvQDsg74A7IO/AOyEgADshIEA7ISCAOyEgwDshIQA7ISFAOyEhgDshIcA7ISIAOyEiQDshIoA7ISLAOyEjADshI0A7ISOAOyEjwDshJAA7ISRAOyEkgDshJMA7ISUAOyElQDshJYA7ISXAOyEmADshJkA7ISaAOyEmwDshJwA7ISdAOyEngDshJ8A7ISgAOyEoQDshKIA7ISjAOyEpADshKUA7ISmAOyEpwDshKgA7ISpAOyEqgDshKsA7ISsAOyErQDshK4A7ISvAOyEsADshLEA7ISyAOyEswDshLQA7IS1AOyEtgDshLcA7IS4AOyEuQDshLoA7IS7AOyEvADshL0A7IS+AOyEvwDshYAA7IWBAOyFggDshYMA7IWEAOyFhQDshYYA7IWHAOyFiADshYkA7IWKAOyFiwDshYwA7IWNAOyFjgDshY8A7IWQAOyFkQDshZIA7IWTAOyFlADshZUA7IWWAOyFlwDshZgA7IWZAOyFmgDshZsA7IWcAOyFnQDshZ4A7IWfAOyFoADshaEA7IWiAOyFowDshaQA7IWlAOyFpgDshacA7IWoAOyFqQDshaoA7IWrAOyFrADsha0A7IWuAOyFrwDshbAA7IWxAOyFsgDshbMA7IW0AOyFtQDshbYA7IW3AOyFuADshbkA7IW6AOyFuwDshbwA7IW9AOyFvgDshb8A7IaAAOyGgQDshoIA7IaDAOyGhADshoUA7IaGAOyGhwDshogA7IaJAOyGigDshosA7IaMAOyGjQDsho4A7IaPAOyGkADshpEA7IaSAOyGkwDshpQA7IaVAOyGlgDshpcA7IaYAOyGmQDshpoA7IabAOyGnADshp0A7IaeAOyGnwDshqAA7IahAOyGogDshqMA7IakAOyGpQDshqYA7IanAOyGqADshqkA7IaqAOyGqwDshqwA7IatAOyGrgDshq8A7IawAOyGsQDshrIA7IazAOyGtADshrUA7Ia2AOyGtwDshrgA7Ia5AOyGugDshrsA7Ia8AOyGvQDshr4A7Ia/AOyHgADsh4EA7IeCAOyHgwDsh4QA7IeFAOyHhgDsh4cA7IeIAOyHiQDsh4oA7IeLAOyHjADsh40A7IeOAOyHjwDsh5AA7IeRAOyHkgDsh5MA7IeUAOyHlQDsh5YA7IeXAOyHmADsh5kA7IeaAOyHmwDsh5wA7IedAOyHngDsh58A7IegAOyHoQDsh6IA7IejAOyHpADsh6UA7IemAOyHpwDsh6gA7IepAOyHqgDsh6sA7IesAOyHrQDsh64A7IevAOyHsADsh7EA7IeyAOyHswDsh7QA7Ie1AOyHtgDsh7cA7Ie4AOyHuQDsh7oA7Ie7AOyHvADsh70A7Ie+AOyHvwDsiIAA7IiBAOyIggDsiIMA7IiEAOyIhQDsiIYA7IiHAOyIiADsiIkA7IiKAOyIiwDsiIwA7IiNAOyIjgDsiI8A7IiQAOyIkQDsiJIA7IiTAOyIlADsiJUA7IiWAOyIlwDsiJgA7IiZAOyImgDsiJsA7IicAOyInQDsiJ4A7IifAOyIoADsiKEA7IiiAOyIowDsiKQA7IilAOyIpgDsiKcA7IioAOyIqQDsiKoA7IirAOyIrADsiK0A7IiuAOyIrwDsiLAA7IixAOyIsgDsiLMA7Ii0AOyItQDsiLYA7Ii3AOyIuADsiLkA7Ii6AOyIuwDsiLwA7Ii9AOyIvgDsiL8A7ImAAOyJgQDsiYIA7ImDAOyJhADsiYUA7ImGAOyJhwDsiYgA7ImJAOyJigDsiYsA7ImMAOyJjQDsiY4A7ImPAOyJkADsiZEA7ImSAOyJkwDsiZQA7ImVAOyJlgDsiZcA7ImYAOyJmQDsiZoA7ImbAOyJnADsiZ0A7ImeAOyJnwDsiaAA7ImhAOyJogDsiaMA7ImkAOyJpQDsiaYA7ImnAOyJqADsiakA7ImqAOyJqwDsiawA7ImtAOyJrgDsia8A7ImwAOyJsQDsibIA7ImzAOyJtADsibUA7Im2AOyJtwDsibgA7Im5AOyJugDsibsA7Im8AOyJvQDsib4A7Im/AOyKgADsioEA7IqCAOyKgwDsioQA7IqFAOyKhgDsiocA7IqIAOyKiQDsiooA7IqLAOyKjADsio0A7IqOAOyKjwDsipAA7IqRAOyKkgDsipMA7IqUAOyKlQDsipYA7IqXAOyKmADsipkA7IqaAOyKmwDsipwA7IqdAOyKngDsip8A7IqgAOyKoQDsiqIA7IqjAOyKpADsiqUA7IqmAOyKpwDsiqgA7IqpAOyKqgDsiqsA7IqsAOyKrQDsiq4A7IqvAOyKsADsirEA7IqyAOyKswDsirQA7Iq1AOyKtgDsircA7Iq4AOyKuQDsiroA7Iq7AOyKvADsir0A7Iq+AOyKvwDsi4AA7IuBAOyLggDsi4MA7IuEAOyLhQDsi4YA7IuHAOyLiADsi4kA7IuKAOyLiwDsi4wA7IuNAOyLjgDsi48A7IuQAOyLkQDsi5IA7IuTAOyLlADsi5UA7IuWAOyLlwDsi5gA7IuZAOyLmgDsi5sA7IucAOyLnQDsi54A7IufAOyLoADsi6EA7IuiAOyLowDsi6QA7IulAOyLpgDsi6cA7IuoAOyLqQDsi6oA7IurAOyLrADsi60A7IuuAOyLrwDsi7AA7IuxAOyLsgDsi7MA7Iu0AOyLtQDsi7YA7Iu3AOyLuADsi7kA7Iu6AOyLuwDsi7wA7Iu9AOyLvgDsi78A7IyAAOyMgQDsjIIA7IyDAOyMhADsjIUA7IyGAOyMhwDsjIgA7IyJAOyMigDsjIsA7IyMAOyMjQDsjI4A7IyPAOyMkADsjJEA7IySAOyMkwDsjJQA7IyVAOyMlgDsjJcA7IyYAOyMmQDsjJoA7IybAOyMnADsjJ0A7IyeAOyMnwDsjKAA7IyhAOyMogDsjKMA7IykAOyMpQDsjKYA7IynAOyMqADsjKkA7IyqAOyMqwDsjKwA7IytAOyMrgDsjK8A7IywAOyMsQDsjLIA7IyzAOyMtADsjLUA7Iy2AOyMtwDsjLgA7Iy5AOyMugDsjLsA7Iy8AOyMvQDsjL4A7Iy/AOyNgADsjYEA7I2CAOyNgwDsjYQA7I2FAOyNhgDsjYcA7I2IAOyNiQDsjYoA7I2LAOyNjADsjY0A7I2OAOyNjwDsjZAA7I2RAOyNkgDsjZMA7I2UAOyNlQDsjZYA7I2XAOyNmADsjZkA7I2aAOyNmwDsjZwA7I2dAOyNngDsjZ8A7I2gAOyNoQDsjaIA7I2jAOyNpADsjaUA7I2mAOyNpwDsjagA7I2pAOyNqgDsjasA7I2sAOyNrQDsja4A7I2vAOyNsADsjbEA7I2yAOyNswDsjbQA7I21AOyNtgDsjbcA7I24AOyNuQDsjboA7I27AOyNvADsjb0A7I2+AOyNvwDsjoAA7I6BAOyOggDsjoMA7I6EAOyOhQDsjoYA7I6HAOyOiADsjokA7I6KAOyOiwDsjowA7I6NAOyOjgDsjo8A7I6QAOyOkQDsjpIA7I6TAOyOlADsjpUA7I6WAOyOlwDsjpgA7I6ZAOyOmgDsjpsA7I6cAOyOnQDsjp4A7I6fAOyOoADsjqEA7I6iAOyOowDsjqQA7I6lAOyOpgDsjqcA7I6oAOyOqQDsjqoA7I6rAOyOrADsjq0A7I6uAOyOrwDsjrAA7I6xAOyOsgDsjrMA7I60AOyOtQDsjrYA7I63AOyOuADsjrkA7I66AOyOuwDsjrwA7I69AOyOvgDsjr8A7I+AAOyPgQDsj4IA7I+DAOyPhADsj4UA7I+GAOyPhwDsj4gA7I+JAOyPigDsj4sA7I+MAOyPjQDsj44A7I+PAOyPkADsj5EA7I+SAOyPkwDsj5QA7I+VAOyPlgDsj5cA7I+YAOyPmQDsj5oA7I+bAOyPnADsj50A7I+eAOyPnwDsj6AA7I+hAOyPogDsj6MA7I+kAOyPpQDsj6YA7I+nAOyPqADsj6kA7I+qAOyPqwDsj6wA7I+tAOyPrgDsj68A7I+wAOyPsQDsj7IA7I+zAOyPtADsj7UA7I+2AOyPtwDsj7gA7I+5AOyPugDsj7sA7I+8AOyPvQDsj74A7I+/AOyQgADskIEA7JCCAOyQgwDskIQA7JCFAOyQhgDskIcA7JCIAOyQiQDskIoA7JCLAOyQjADskI0A7JCOAOyQjwDskJAA7JCRAOyQkgDskJMA7JCUAOyQlQDskJYA7JCXAOyQmADskJkA7JCaAOyQmwDskJwA7JCdAOyQngDskJ8A7JCgAOyQoQDskKIA7JCjAOyQpADskKUA7JCmAOyQpwDskKgA7JCpAOyQqgDskKsA7JCsAOyQrQDskK4A7JCvAOyQsADskLEA7JCyAOyQswDskLQA7JC1AOyQtgDskLcA7JC4AOyQuQDskLoA7JC7AOyQvADskL0A7JC+AOyQvwDskYAA7JGBAOyRggDskYMA7JGEAOyRhQDskYYA7JGHAOyRiADskYkA7JGKAOyRiwDskYwA7JGNAOyRjgDskY8A7JGQAOyRkQDskZIA7JGTAOyRlADskZUA7JGWAOyRlwDskZgA7JGZAOyRmgDskZsA7JGcAOyRnQDskZ4A7JGfAOyRoADskaEA7JGiAOyRowDskaQA7JGlAOyRpgDskacA7JGoAOyRqQDskaoA7JGrAOyRrADska0A7JGuAOyRrwDskbAA7JGxAOyRsgDskbMA7JG0AOyRtQDskbYA7JG3AOyRuADskbkA7JG6AOyRuwDskbwA7JG9AOyRvgDskb8A7JKAAOySgQDskoIA7JKDAOyShADskoUA7JKGAOyShwDskogA7JKJAOySigDskosA7JKMAOySjQDsko4A7JKPAOySkADskpEA7JKSAOySkwDskpQA7JKVAOySlgDskpcA7JKYAOySmQDskpoA7JKbAOySnADskp0A7JKeAOySnwDskqAA7JKhAOySogDskqMA7JKkAOySpQDskqYA7JKnAOySqADskqkA7JKqAOySqwDskqwA7JKtAOySrgDskq8A7JKwAOySsQDskrIA7JKzAOyStADskrUA7JK2AOyStwDskrgA7JK5AOySugDskrsA7JK8AOySvQDskr4A7JK/AOyTgADsk4EA7JOCAOyTgwDsk4QA7JOFAOyThgDsk4cA7JOIAOyTiQDsk4oA7JOLAOyTjADsk40A7JOOAOyTjwDsk5AA7JORAOyTkgDsk5MA7JOUAOyTlQDsk5YA7JOXAOyTmADsk5kA7JOaAOyTmwDsk5wA7JOdAOyTngDsk58A7JOgAOyToQDsk6IA7JOjAOyTpADsk6UA7JOmAOyTpwDsk6gA7JOpAOyTqgDsk6sA7JOsAOyTrQDsk64A7JOvAOyTsADsk7EA7JOyAOyTswDsk7QA7JO1AOyTtgDsk7cA7JO4AOyTuQDsk7oA7JO7AOyTvADsk70A7JO+AOyTvwDslIAA7JSBAOyUggDslIMA7JSEAOyUhQDslIYA7JSHAOyUiADslIkA7JSKAOyUiwDslIwA7JSNAOyUjgDslI8A7JSQAOyUkQDslJIA7JSTAOyUlADslJUA7JSWAOyUlwDslJgA7JSZAOyUmgDslJsA7JScAOyUnQDslJ4A7JSfAOyUoADslKEA7JSiAOyUowDslKQA7JSlAOyUpgDslKcA7JSoAOyUqQDslKoA7JSrAOyUrADslK0A7JSuAOyUrwDslLAA7JSxAOyUsgDslLMA7JS0AOyUtQDslLYA7JS3AOyUuADslLkA7JS6AOyUuwDslLwA7JS9AOyUvgDslL8A7JWAAOyVgQDslYIA7JWDAOyVhADslYUA7JWGAOyVhwDslYgA7JWJAOyVigDslYsA7JWMAOyVjQDslY4A7JWPAOyVkADslZEA7JWSAOyVkwDslZQA7JWVAOyVlgDslZcA7JWYAOyVmQDslZoA7JWbAOyVnADslZ0A7JWeAOyVnwDslaAA7JWhAOyVogDslaMA7JWkAOyVpQDslaYA7JWnAOyVqADslakA7JWqAOyVqwDslawA7JWtAOyVrgDsla8A7JWwAOyVsQDslbIA7JWzAOyVtADslbUA7JW2AOyVtwDslbgA7JW5AOyVugDslbsA7JW8AOyVvQDslb4A7JW/AOyWgADsloEA7JaCAOyWgwDsloQA7JaFAOyWhgDslocA7JaIAOyWiQDslooA7JaLAOyWjADslo0A7JaOAOyWjwDslpAA7JaRAOyWkgDslpMA7JaUAOyWlQDslpYA7JaXAOyWmADslpkA7JaaAOyWmwDslpwA7JadAOyWngDslp8A7JagAOyWoQDslqIA7JajAOyWpADslqUA7JamAOyWpwDslqgA7JapAOyWqgDslqsA7JasAOyWrQDslq4A7JavAOyWsADslrEA7JayAOyWswDslrQA7Ja1AOyWtgDslrcA7Ja4AOyWuQDslroA7Ja7AOyWvADslr0A7Ja+AOyWvwDsl4AA7JeBAOyXggDsl4MA7JeEAOyXhQDsl4YA7JeHAOyXiADsl4kA7JeKAOyXiwDsl4wA7JeNAOyXjgDsl48A7JeQAOyXkQDsl5IA7JeTAOyXlADsl5UA7JeWAOyXlwDsl5gA7JeZAOyXmgDsl5sA7JecAOyXnQDsl54A7JefAOyXoADsl6EA7JeiAOyXowDsl6QA7JelAOyXpgDsl6cA7JeoAOyXqQDsl6oA7JerAOyXrADsl60A7JeuAOyXrwDsl7AA7JexAOyXsgDsl7MA7Je0AOyXtQDsl7YA7Je3AOyXuADsl7kA7Je6AOyXuwDsl7wA7Je9AOyXvgDsl78A7JiAAOyYgQDsmIIA7JiDAOyYhADsmIUA7JiGAOyYhwDsmIgA7JiJAOyYigDsmIsA7JiMAOyYjQDsmI4A7JiPAOyYkADsmJEA7JiSAOyYkwDsmJQA7JiVAOyYlgDsmJcA7JiYAOyYmQDsmJoA7JibAOyYnADsmJ0A7JieAOyYnwDsmKAA7JihAOyYogDsmKMA7JikAOyYpQDsmKYA7JinAOyYqADsmKkA7JiqAOyYqwDsmKwA7JitAOyYrgDsmK8A7JiwAOyYsQDsmLIA7JizAOyYtADsmLUA7Ji2AOyYtwDsmLgA7Ji5AOyYugDsmLsA7Ji8AOyYvQDsmL4A7Ji/AOyZgADsmYEA7JmCAOyZgwDsmYQA7JmFAOyZhgDsmYcA7JmIAOyZiQDsmYoA7JmLAOyZjADsmY0A7JmOAOyZjwDsmZAA7JmRAOyZkgDsmZMA7JmUAOyZlQDsmZYA7JmXAOyZmADsmZkA7JmaAOyZmwDsmZwA7JmdAOyZngDsmZ8A7JmgAOyZoQDsmaIA7JmjAOyZpADsmaUA7JmmAOyZpwDsmagA7JmpAOyZqgDsmasA7JmsAOyZrQDsma4A7JmvAOyZsADsmbEA7JmyAOyZswDsmbQA7Jm1AOyZtgDsmbcA7Jm4AOyZuQDsmboA7Jm7AOyZvADsmb0A7Jm+AOyZvwDsmoAA7JqBAOyaggDsmoMA7JqEAOyahQDsmoYA7JqHAOyaiADsmokA7JqKAOyaiwDsmowA7JqNAOyajgDsmo8A7JqQAOyakQDsmpIA7JqTAOyalADsmpUA7JqWAOyalwDsmpgA7JqZAOyamgDsmpsA7JqcAOyanQDsmp4A7JqfAOyaoADsmqEA7JqiAOyaowDsmqQA7JqlAOyapgDsmqcA7JqoAOyaqQDsmqoA7JqrAOyarADsmq0A7JquAOyarwDsmrAA7JqxAOyasgDsmrMA7Jq0AOyatQDsmrYA7Jq3AOyauADsmrkA7Jq6AOyauwDsmrwA7Jq9AOyavgDsmr8A7JuAAOybgQDsm4IA7JuDAOybhADsm4UA7JuGAOybhwDsm4gA7JuJAOybigDsm4sA7JuMAOybjQDsm44A7JuPAOybkADsm5EA7JuSAOybkwDsm5QA7JuVAOyblgDsm5cA7JuYAOybmQDsm5oA7JubAOybnADsm50A7JueAOybnwDsm6AA7JuhAOybogDsm6MA7JukAOybpQDsm6YA7JunAOybqADsm6kA7JuqAOybqwDsm6wA7JutAOybrgDsm68A7JuwAOybsQDsm7IA7JuzAOybtADsm7UA7Ju2AOybtwDsm7gA7Ju5AOybugDsm7sA7Ju8AOybvQDsm74A7Ju/AOycgADsnIEA7JyCAOycgwDsnIQA7JyFAOychgDsnIcA7JyIAOyciQDsnIoA7JyLAOycjADsnI0A7JyOAOycjwDsnJAA7JyRAOyckgDsnJMA7JyUAOyclQDsnJYA7JyXAOycmADsnJkA7JyaAOycmwDsnJwA7JydAOycngDsnJ8A7JygAOycoQDsnKIA7JyjAOycpADsnKUA7JymAOycpwDsnKgA7JypAOycqgDsnKsA7JysAOycrQDsnK4A7JyvAOycsADsnLEA7JyyAOycswDsnLQA7Jy1AOyctgDsnLcA7Jy4AOycuQDsnLoA7Jy7AOycvADsnL0A7Jy+AOycvwDsnYAA7J2BAOydggDsnYMA7J2EAOydhQDsnYYA7J2HAOydiADsnYkA7J2KAOydiwDsnYwA7J2NAOydjgDsnY8A7J2QAOydkQDsnZIA7J2TAOydlADsnZUA7J2WAOydlwDsnZgA7J2ZAOydmgDsnZsA7J2cAOydnQDsnZ4A7J2fAOydoADsnaEA7J2iAOydowDsnaQA7J2lAOydpgDsnacA7J2oAOydqQDsnaoA7J2rAOydrADsna0A7J2uAOydrwDsnbAA7J2xAOydsgDsnbMA7J20AOydtQDsnbYA7J23AOyduADsnbkA7J26AOyduwDsnbwA7J29AOydvgDsnb8A7J6AAOyegQDsnoIA7J6DAOyehADsnoUA7J6GAOyehwDsnogA7J6JAOyeigDsnosA7J6MAOyejQDsno4A7J6PAOyekADsnpEA7J6SAOyekwDsnpQA7J6VAOyelgDsnpcA7J6YAOyemQDsnpoA7J6bAOyenADsnp0A7J6eAOyenwDsnqAA7J6hAOyeogDsnqMA7J6kAOyepQDsnqYA7J6nAOyeqADsnqkA7J6qAOyeqwDsnqwA7J6tAOyergDsnq8A7J6wAOyesQDsnrIA7J6zAOyetADsnrUA7J62AOyetwDsnrgA7J65AOyeugDsnrsA7J68AOyevQDsnr4A7J6/AOyfgADsn4EA7J+CAOyfgwDsn4QA7J+FAOyfhgDsn4cA7J+IAOyfiQDsn4oA7J+LAOyfjADsn40A7J+OAOyfjwDsn5AA7J+RAOyfkgDsn5MA7J+UAOyflQDsn5YA7J+XAOyfmADsn5kA7J+aAOyfmwDsn5wA7J+dAOyfngDsn58A7J+gAOyfoQDsn6IA7J+jAOyfpADsn6UA7J+mAOyfpwDsn6gA7J+pAOyfqgDsn6sA7J+sAOyfrQDsn64A7J+vAOyfsADsn7EA7J+yAOyfswDsn7QA7J+1AOyftgDsn7cA7J+4AOyfuQDsn7oA7J+7AOyfvADsn70A7J++AOyfvwDsoIAA7KCBAOygggDsoIMA7KCEAOyghQDsoIYA7KCHAOygiADsoIkA7KCKAOygiwDsoIwA7KCNAOygjgDsoI8A7KCQAOygkQDsoJIA7KCTAOyglADsoJUA7KCWAOyglwDsoJgA7KCZAOygmgDsoJsA7KCcAOygnQDsoJ4A7KCfAOygoADsoKEA7KCiAOygowDsoKQA7KClAOygpgDsoKcA7KCoAOygqQDsoKoA7KCrAOygrADsoK0A7KCuAOygrwDsoLAA7KCxAOygsgDsoLMA7KC0AOygtQDsoLYA7KC3AOyguADsoLkA7KC6AOyguwDsoLwA7KC9AOygvgDsoL8A7KGAAOyhgQDsoYIA7KGDAOyhhADsoYUA7KGGAOyhhwDsoYgA7KGJAOyhigDsoYsA7KGMAOyhjQDsoY4A7KGPAOyhkADsoZEA7KGSAOyhkwDsoZQA7KGVAOyhlgDsoZcA7KGYAOyhmQDsoZoA7KGbAOyhnADsoZ0A7KGeAOyhnwDsoaAA7KGhAOyhogDsoaMA7KGkAOyhpQDsoaYA7KGnAOyhqADsoakA7KGqAOyhqwDsoawA7KGtAOyhrgDsoa8A7KGwAOyhsQDsobIA7KGzAOyhtADsobUA7KG2AOyhtwDsobgA7KG5AOyhugDsobsA7KG8AOyhvQDsob4A7KG/AOyigADsooEA7KKCAOyigwDsooQA7KKFAOyihgDsoocA7KKIAOyiiQDsoooA7KKLAOyijADsoo0A7KKOAOyijwDsopAA7KKRAOyikgDsopMA7KKUAOyilQDsopYA7KKXAOyimADsopkA7KKaAOyimwDsopwA7KKdAOyingDsop8A7KKgAOyioQDsoqIA7KKjAOyipADsoqUA7KKmAOyipwDsoqgA7KKpAOyiqgDsoqsA7KKsAOyirQDsoq4A7KKvAOyisADsorEA7KKyAOyiswDsorQA7KK1AOyitgDsorcA7KK4AOyiuQDsoroA7KK7AOyivADsor0A7KK+AOyivwDso4AA7KOBAOyjggDso4MA7KOEAOyjhQDso4YA7KOHAOyjiADso4kA7KOKAOyjiwDso4wA7KONAOyjjgDso48A7KOQAOyjkQDso5IA7KOTAOyjlADso5UA7KOWAOyjlwDso5gA7KOZAOyjmgDso5sA7KOcAOyjnQDso54A7KOfAOyjoADso6EA7KOiAOyjowDso6QA7KOlAOyjpgDso6cA7KOoAOyjqQDso6oA7KOrAOyjrADso60A7KOuAOyjrwDso7AA7KOxAOyjsgDso7MA7KO0AOyjtQDso7YA7KO3AOyjuADso7kA7KO6AOyjuwDso7wA7KO87J2YAOyjvQDso74A7KO/AOykgADspIEA7KSCAOykgwDspIQA7KSFAOykhgDspIcA7KSIAOykiQDspIoA7KSLAOykjADspI0A7KSOAOykjwDspJAA7KSRAOykkgDspJMA7KSUAOyklQDspJYA7KSXAOykmADspJkA7KSaAOykmwDspJwA7KSdAOykngDspJ8A7KSgAOykoQDspKIA7KSjAOykpADspKUA7KSmAOykpwDspKgA7KSpAOykqgDspKsA7KSsAOykrQDspK4A7KSvAOyksADspLEA7KSyAOykswDspLQA7KS1AOyktgDspLcA7KS4AOykuQDspLoA7KS7AOykvADspL0A7KS+AOykvwDspYAA7KWBAOylggDspYMA7KWEAOylhQDspYYA7KWHAOyliADspYkA7KWKAOyliwDspYwA7KWNAOyljgDspY8A7KWQAOylkQDspZIA7KWTAOyllADspZUA7KWWAOyllwDspZgA7KWZAOylmgDspZsA7KWcAOylnQDspZ4A7KWfAOyloADspaEA7KWiAOylowDspaQA7KWlAOylpgDspacA7KWoAOylqQDspaoA7KWrAOylrADspa0A7KWuAOylrwDspbAA7KWxAOylsgDspbMA7KW0AOyltQDspbYA7KW3AOyluADspbkA7KW6AOyluwDspbwA7KW9AOylvgDspb8A7KaAAOymgQDspoIA7KaDAOymhADspoUA7KaGAOymhwDspogA7KaJAOymigDsposA7KaMAOymjQDspo4A7KaPAOymkADsppEA7KaSAOymkwDsppQA7KaVAOymlgDsppcA7KaYAOymmQDsppoA7KabAOymnADspp0A7KaeAOymnwDspqAA7KahAOymogDspqMA7KakAOympQDspqYA7KanAOymqADspqkA7KaqAOymqwDspqwA7KatAOymrgDspq8A7KawAOymsQDsprIA7KazAOymtADsprUA7Ka2AOymtwDsprgA7Ka5AOymugDsprsA7Ka8AOymvQDspr4A7Ka/AOyngADsp4EA7KeCAOyngwDsp4QA7KeFAOynhgDsp4cA7KeIAOyniQDsp4oA7KeLAOynjADsp40A7KeOAOynjwDsp5AA7KeRAOynkgDsp5MA7KeUAOynlQDsp5YA7KeXAOynmADsp5kA7KeaAOynmwDsp5wA7KedAOynngDsp58A7KegAOynoQDsp6IA7KejAOynpADsp6UA7KemAOynpwDsp6gA7KepAOynqgDsp6sA7KesAOynrQDsp64A7KevAOynsADsp7EA7KeyAOynswDsp7QA7Ke1AOyntgDsp7cA7Ke4AOynuQDsp7oA7Ke7AOynvADsp70A7Ke+AOynvwDsqIAA7KiBAOyoggDsqIMA7KiEAOyohQDsqIYA7KiHAOyoiADsqIkA7KiKAOyoiwDsqIwA7KiNAOyojgDsqI8A7KiQAOyokQDsqJIA7KiTAOyolADsqJUA7KiWAOyolwDsqJgA7KiZAOyomgDsqJsA7KicAOyonQDsqJ4A7KifAOyooADsqKEA7KiiAOyoowDsqKQA7KilAOyopgDsqKcA7KioAOyoqQDsqKoA7KirAOyorADsqK0A7KiuAOyorwDsqLAA7KixAOyosgDsqLMA7Ki0AOyotQDsqLYA7Ki3AOyouADsqLkA7Ki6AOyouwDsqLwA7Ki9AOyovgDsqL8A7KmAAOypgQDsqYIA7KmDAOyphADsqYUA7KmGAOyphwDsqYgA7KmJAOypigDsqYsA7KmMAOypjQDsqY4A7KmPAOypkADsqZEA7KmSAOypkwDsqZQA7KmVAOyplgDsqZcA7KmYAOypmQDsqZoA7KmbAOypnADsqZ0A7KmeAOypnwDsqaAA7KmhAOypogDsqaMA7KmkAOyppQDsqaYA7KmnAOypqADsqakA7KmqAOypqwDsqawA7KmtAOyprgDsqa8A7KmwAOypsQDsqbIA7KmzAOyptADsqbUA7Km2AOyptwDsqbgA7Km5AOypugDsqbsA7Km8AOypvQDsqb4A7Km/AOyqgADsqoEA7KqCAOyqgwDsqoQA7KqFAOyqhgDsqocA7KqIAOyqiQDsqooA7KqLAOyqjADsqo0A7KqOAOyqjwDsqpAA7KqRAOyqkgDsqpMA7KqUAOyqlQDsqpYA7KqXAOyqmADsqpkA7KqaAOyqmwDsqpwA7KqdAOyqngDsqp8A7KqgAOyqoQDsqqIA7KqjAOyqpADsqqUA7KqmAOyqpwDsqqgA7KqpAOyqqgDsqqsA7KqsAOyqrQDsqq4A7KqvAOyqsADsqrEA7KqyAOyqswDsqrQA7Kq1AOyqtgDsqrcA7Kq4AOyquQDsqroA7Kq7AOyqvADsqr0A7Kq+AOyqvwDsq4AA7KuBAOyrggDsq4MA7KuEAOyrhQDsq4YA7KuHAOyriADsq4kA7KuKAOyriwDsq4wA7KuNAOyrjgDsq48A7KuQAOyrkQDsq5IA7KuTAOyrlADsq5UA7KuWAOyrlwDsq5gA7KuZAOyrmgDsq5sA7KucAOyrnQDsq54A7KufAOyroADsq6EA7KuiAOyrowDsq6QA7KulAOyrpgDsq6cA7KuoAOyrqQDsq6oA7KurAOyrrADsq60A7KuuAOyrrwDsq7AA7KuxAOyrsgDsq7MA7Ku0AOyrtQDsq7YA7Ku3AOyruADsq7kA7Ku6AOyruwDsq7wA7Ku9AOyrvgDsq78A7KyAAOysgQDsrIIA7KyDAOyshADsrIUA7KyGAOyshwDsrIgA7KyJAOysigDsrIsA7KyMAOysjQDsrI4A7KyPAOyskADsrJEA7KySAOyskwDsrJQA7KyVAOyslgDsrJcA7KyYAOysmQDsrJoA7KybAOysnADsrJ0A7KyeAOysnwDsrKAA7KyhAOysogDsrKMA7KykAOyspQDsrKYA7KynAOysqADsrKkA7KyqAOysqwDsrKwA7KytAOysrgDsrK8A7KywAOyssQDsrLIA7KyzAOystADsrLUA7Ky2AOystwDsrLgA7Ky5AOysugDsrLsA7Ky8AOysvQDsrL4A7Ky/AOytgADsrYEA7K2CAOytgwDsrYQA7K2FAOythgDsrYcA7K2IAOytiQDsrYoA7K2LAOytjADsrY0A7K2OAOytjwDsrZAA7K2RAOytkgDsrZMA7K2UAOytlQDsrZYA7K2XAOytmADsrZkA7K2aAOytmwDsrZwA7K2dAOytngDsrZ8A7K2gAOytoQDsraIA7K2jAOytpADsraUA7K2mAOytpwDsragA7K2pAOytqgDsrasA7K2sAOytrQDsra4A7K2vAOytsADsrbEA7K2yAOytswDsrbQA7K21AOyttgDsrbcA7K24AOytuQDsrboA7K27AOytvADsrb0A7K2+AOytvwDsroAA7K6BAOyuggDsroMA7K6EAOyuhQDsroYA7K6HAOyuiADsrokA7K6KAOyuiwDsrowA7K6NAOyujgDsro8A7K6QAOyukQDsrpIA7K6TAOyulADsrpUA7K6WAOyulwDsrpgA7K6ZAOyumgDsrpsA7K6cAOyunQDsrp4A7K6fAOyuoADsrqEA7K6iAOyuowDsrqQA7K6lAOyupgDsrqcA7K6oAOyuqQDsrqoA7K6rAOyurADsrq0A7K6uAOyurwDsrrAA7K6xAOyusgDsrrMA7K60AOyutQDsrrYA7K63AOyuuADsrrkA7K66AOyuuwDsrrwA7K69AOyuvgDsrr8A7K+AAOyvgQDsr4IA7K+DAOyvhADsr4UA7K+GAOyvhwDsr4gA7K+JAOyvigDsr4sA7K+MAOyvjQDsr44A7K+PAOyvkADsr5EA7K+SAOyvkwDsr5QA7K+VAOyvlgDsr5cA7K+YAOyvmQDsr5oA7K+bAOyvnADsr50A7K+eAOyvnwDsr6AA7K+hAOyvogDsr6MA7K+kAOyvpQDsr6YA7K+nAOyvqADsr6kA7K+qAOyvqwDsr6wA7K+tAOyvrgDsr68A7K+wAOyvsQDsr7IA7K+zAOyvtADsr7UA7K+2AOyvtwDsr7gA7K+5AOyvugDsr7sA7K+8AOyvvQDsr74A7K+/AOywgADssIEA7LCCAOywgwDssIQA7LCFAOywhgDssIcA7LCIAOywiQDssIoA7LCLAOywjADssI0A7LCOAOywjwDssJAA7LCRAOywkgDssJMA7LCUAOywlQDssJYA7LCXAOywmADssJkA7LCaAOywmwDssJwA7LCdAOywngDssJ8A7LCgAOywoQDssKIA7LCjAOywpADssKUA7LCmAOywpwDssKgA7LCpAOywqgDssKsA7LCsAOywrQDssK4A7LCvAOywsADssLEA7LCyAOywswDssLQA7LC1AOywtgDssLcA7LC4AOywuOqzoADssLkA7LC6AOywuwDssLwA7LC9AOywvgDssL8A7LGAAOyxgQDssYIA7LGDAOyxhADssYUA7LGGAOyxhwDssYgA7LGJAOyxigDssYsA7LGMAOyxjQDssY4A7LGPAOyxkADssZEA7LGSAOyxkwDssZQA7LGVAOyxlgDssZcA7LGYAOyxmQDssZoA7LGbAOyxnADssZ0A7LGeAOyxnwDssaAA7LGhAOyxogDssaMA7LGkAOyxpQDssaYA7LGnAOyxqADssakA7LGqAOyxqwDssawA7LGtAOyxrgDssa8A7LGwAOyxsQDssbIA7LGzAOyxtADssbUA7LG2AOyxtwDssbgA7LG5AOyxugDssbsA7LG8AOyxvQDssb4A7LG/AOyygADssoEA7LKCAOyygwDssoQA7LKFAOyyhgDssocA7LKIAOyyiQDssooA7LKLAOyyjADsso0A7LKOAOyyjwDsspAA7LKRAOyykgDsspMA7LKUAOyylQDsspYA7LKXAOyymADsspkA7LKaAOyymwDsspwA7LKdAOyyngDssp8A7LKgAOyyoQDssqIA7LKjAOyypADssqUA7LKmAOyypwDssqgA7LKpAOyyqgDssqsA7LKsAOyyrQDssq4A7LKvAOyysADssrEA7LKyAOyyswDssrQA7LK1AOyytgDssrcA7LK4AOyyuQDssroA7LK7AOyyvADssr0A7LK+AOyyvwDss4AA7LOBAOyzggDss4MA7LOEAOyzhQDss4YA7LOHAOyziADss4kA7LOKAOyziwDss4wA7LONAOyzjgDss48A7LOQAOyzkQDss5IA7LOTAOyzlADss5UA7LOWAOyzlwDss5gA7LOZAOyzmgDss5sA7LOcAOyznQDss54A7LOfAOyzoADss6EA7LOiAOyzowDss6QA7LOlAOyzpgDss6cA7LOoAOyzqQDss6oA7LOrAOyzrADss60A7LOuAOyzrwDss7AA7LOxAOyzsgDss7MA7LO0AOyztQDss7YA7LO3AOyzuADss7kA7LO6AOyzuwDss7wA7LO9AOyzvgDss78A7LSAAOy0gQDstIIA7LSDAOy0hADstIUA7LSGAOy0hwDstIgA7LSJAOy0igDstIsA7LSMAOy0jQDstI4A7LSPAOy0kADstJEA7LSSAOy0kwDstJQA7LSVAOy0lgDstJcA7LSYAOy0mQDstJoA7LSbAOy0nADstJ0A7LSeAOy0nwDstKAA7LShAOy0ogDstKMA7LSkAOy0pQDstKYA7LSnAOy0qADstKkA7LSqAOy0qwDstKwA7LStAOy0rgDstK8A7LSwAOy0sQDstLIA7LSzAOy0tADstLUA7LS2AOy0twDstLgA7LS5AOy0ugDstLsA7LS8AOy0vQDstL4A7LS/AOy1gADstYEA7LWCAOy1gwDstYQA7LWFAOy1hgDstYcA7LWIAOy1iQDstYoA7LWLAOy1jADstY0A7LWOAOy1jwDstZAA7LWRAOy1kgDstZMA7LWUAOy1lQDstZYA7LWXAOy1mADstZkA7LWaAOy1mwDstZwA7LWdAOy1ngDstZ8A7LWgAOy1oQDstaIA7LWjAOy1pADstaUA7LWmAOy1pwDstagA7LWpAOy1qgDstasA7LWsAOy1rQDsta4A7LWvAOy1sADstbEA7LWyAOy1swDstbQA7LW1AOy1tgDstbcA7LW4AOy1uQDstboA7LW7AOy1vADstb0A7LW+AOy1vwDstoAA7LaBAOy2ggDstoMA7LaEAOy2hQDstoYA7LaHAOy2iADstokA7LaKAOy2iwDstowA7LaNAOy2jgDsto8A7LaQAOy2kQDstpIA7LaTAOy2lADstpUA7LaWAOy2lwDstpgA7LaZAOy2mgDstpsA7LacAOy2nQDstp4A7LafAOy2oADstqEA7LaiAOy2owDstqQA7LalAOy2pgDstqcA7LaoAOy2qQDstqoA7LarAOy2rADstq0A7LauAOy2rwDstrAA7LaxAOy2sgDstrMA7La0AOy2tQDstrYA7La3AOy2uADstrkA7La6AOy2uwDstrwA7La9AOy2vgDstr8A7LeAAOy3gQDst4IA7LeDAOy3hADst4UA7LeGAOy3hwDst4gA7LeJAOy3igDst4sA7LeMAOy3jQDst44A7LePAOy3kADst5EA7LeSAOy3kwDst5QA7LeVAOy3lgDst5cA7LeYAOy3mQDst5oA7LebAOy3nADst50A7LeeAOy3nwDst6AA7LehAOy3ogDst6MA7LekAOy3pQDst6YA7LenAOy3qADst6kA7LeqAOy3qwDst6wA7LetAOy3rgDst68A7LewAOy3sQDst7IA7LezAOy3tADst7UA7Le2AOy3twDst7gA7Le5AOy3ugDst7sA7Le8AOy3vQDst74A7Le/AOy4gADsuIEA7LiCAOy4gwDsuIQA7LiFAOy4hgDsuIcA7LiIAOy4iQDsuIoA7LiLAOy4jADsuI0A7LiOAOy4jwDsuJAA7LiRAOy4kgDsuJMA7LiUAOy4lQDsuJYA7LiXAOy4mADsuJkA7LiaAOy4mwDsuJwA7LidAOy4ngDsuJ8A7LigAOy4oQDsuKIA7LijAOy4pADsuKUA7LimAOy4pwDsuKgA7LipAOy4qgDsuKsA7LisAOy4rQDsuK4A7LivAOy4sADsuLEA7LiyAOy4swDsuLQA7Li1AOy4tgDsuLcA7Li4AOy4uQDsuLoA7Li7AOy4vADsuL0A7Li+AOy4vwDsuYAA7LmBAOy5ggDsuYMA7LmEAOy5hQDsuYYA7LmHAOy5iADsuYkA7LmKAOy5iwDsuYwA7LmNAOy5jgDsuY8A7LmQAOy5kQDsuZIA7LmTAOy5lADsuZUA7LmWAOy5lwDsuZgA7LmZAOy5mgDsuZsA7LmcAOy5nQDsuZ4A7LmfAOy5oADsuaEA7LmiAOy5owDsuaQA7LmlAOy5pgDsuacA7LmoAOy5qQDsuaoA7LmrAOy5rADsua0A7LmuAOy5rwDsubAA7LmxAOy5sgDsubMA7Lm0AOy5tQDsubYA7Lm3AOy5uADsubkA7Lm6AOy5uwDsubwA7Lm9AOy5vgDsub8A7LqAAOy6gQDsuoIA7LqDAOy6hADsuoUA7LqGAOy6hwDsuogA7LqJAOy6igDsuosA7LqMAOy6jQDsuo4A7LqPAOy6kADsupEA7LqSAOy6kwDsupQA7LqVAOy6lgDsupcA7LqYAOy6mQDsupoA7LqbAOy6nADsup0A7LqeAOy6nwDsuqAA7LqhAOy6ogDsuqMA7LqkAOy6pQDsuqYA7LqnAOy6qADsuqkA7LqqAOy6qwDsuqwA7LqtAOy6rgDsuq8A7LqwAOy6sQDsurIA7LqzAOy6tADsurUA7Lq2AOy6twDsurgA7Lq5AOy6ugDsursA7Lq8AOy6vQDsur4A7Lq/AOy7gADsu4EA7LuCAOy7gwDsu4QA7LuFAOy7hgDsu4cA7LuIAOy7iQDsu4oA7LuLAOy7jADsu40A7LuOAOy7jwDsu5AA7LuRAOy7kgDsu5MA7LuUAOy7lQDsu5YA7LuXAOy7mADsu5kA7LuaAOy7mwDsu5wA7LudAOy7ngDsu58A7LugAOy7oQDsu6IA7LujAOy7pADsu6UA7LumAOy7pwDsu6gA7LupAOy7qgDsu6sA7LusAOy7rQDsu64A7LuvAOy7sADsu7EA7LuyAOy7swDsu7QA7Lu1AOy7tgDsu7cA7Lu4AOy7uQDsu7oA7Lu7AOy7vADsu70A7Lu+AOy7vwDsvIAA7LyBAOy8ggDsvIMA7LyEAOy8hQDsvIYA7LyHAOy8iADsvIkA7LyKAOy8iwDsvIwA7LyNAOy8jgDsvI8A7LyQAOy8kQDsvJIA7LyTAOy8lADsvJUA7LyWAOy8lwDsvJgA7LyZAOy8mgDsvJsA7LycAOy8nQDsvJ4A7LyfAOy8oADsvKEA7LyiAOy8owDsvKQA7LylAOy8pgDsvKcA7LyoAOy8qQDsvKoA7LyrAOy8rADsvK0A7LyuAOy8rwDsvLAA7LyxAOy8sgDsvLMA7Ly0AOy8tQDsvLYA7Ly3AOy8uADsvLkA7Ly6AOy8uwDsvLwA7Ly9AOy8vgDsvL8A7L2AAOy9gQDsvYIA7L2DAOy9hADsvYUA7L2GAOy9hwDsvYgA7L2JAOy9igDsvYsA7L2MAOy9jQDsvY4A7L2PAOy9kADsvZEA7L2SAOy9kwDsvZQA7L2VAOy9lgDsvZcA7L2YAOy9mQDsvZoA7L2bAOy9nADsvZ0A7L2eAOy9nwDsvaAA7L2hAOy9ogDsvaMA7L2kAOy9pQDsvaYA7L2nAOy9qADsvakA7L2qAOy9qwDsvawA7L2tAOy9rgDsva8A7L2wAOy9sQDsvbIA7L2zAOy9tADsvbUA7L22AOy9twDsvbgA7L25AOy9ugDsvbsA7L28AOy9vQDsvb4A7L2/AOy+gADsvoEA7L6CAOy+gwDsvoQA7L6FAOy+hgDsvocA7L6IAOy+iQDsvooA7L6LAOy+jADsvo0A7L6OAOy+jwDsvpAA7L6RAOy+kgDsvpMA7L6UAOy+lQDsvpYA7L6XAOy+mADsvpkA7L6aAOy+mwDsvpwA7L6dAOy+ngDsvp8A7L6gAOy+oQDsvqIA7L6jAOy+pADsvqUA7L6mAOy+pwDsvqgA7L6pAOy+qgDsvqsA7L6sAOy+rQDsvq4A7L6vAOy+sADsvrEA7L6yAOy+swDsvrQA7L61AOy+tgDsvrcA7L64AOy+uQDsvroA7L67AOy+vADsvr0A7L6+AOy+vwDsv4AA7L+BAOy/ggDsv4MA7L+EAOy/hQDsv4YA7L+HAOy/iADsv4kA7L+KAOy/iwDsv4wA7L+NAOy/jgDsv48A7L+QAOy/kQDsv5IA7L+TAOy/lADsv5UA7L+WAOy/lwDsv5gA7L+ZAOy/mgDsv5sA7L+cAOy/nQDsv54A7L+fAOy/oADsv6EA7L+iAOy/owDsv6QA7L+lAOy/pgDsv6cA7L+oAOy/qQDsv6oA7L+rAOy/rADsv60A7L+uAOy/rwDsv7AA7L+xAOy/sgDsv7MA7L+0AOy/tQDsv7YA7L+3AOy/uADsv7kA7L+6AOy/uwDsv7wA7L+9AOy/vgDsv78A7YCAAO2AgQDtgIIA7YCDAO2AhADtgIUA7YCGAO2AhwDtgIgA7YCJAO2AigDtgIsA7YCMAO2AjQDtgI4A7YCPAO2AkADtgJEA7YCSAO2AkwDtgJQA7YCVAO2AlgDtgJcA7YCYAO2AmQDtgJoA7YCbAO2AnADtgJ0A7YCeAO2AnwDtgKAA7YChAO2AogDtgKMA7YCkAO2ApQDtgKYA7YCnAO2AqADtgKkA7YCqAO2AqwDtgKwA7YCtAO2ArgDtgK8A7YCwAO2AsQDtgLIA7YCzAO2AtADtgLUA7YC2AO2AtwDtgLgA7YC5AO2AugDtgLsA7YC8AO2AvQDtgL4A7YC/AO2BgADtgYEA7YGCAO2BgwDtgYQA7YGFAO2BhgDtgYcA7YGIAO2BiQDtgYoA7YGLAO2BjADtgY0A7YGOAO2BjwDtgZAA7YGRAO2BkgDtgZMA7YGUAO2BlQDtgZYA7YGXAO2BmADtgZkA7YGaAO2BmwDtgZwA7YGdAO2BngDtgZ8A7YGgAO2BoQDtgaIA7YGjAO2BpADtgaUA7YGmAO2BpwDtgagA7YGpAO2BqgDtgasA7YGsAO2BrQDtga4A7YGvAO2BsADtgbEA7YGyAO2BswDtgbQA7YG1AO2BtgDtgbcA7YG4AO2BuQDtgboA7YG7AO2BvADtgb0A7YG+AO2BvwDtgoAA7YKBAO2CggDtgoMA7YKEAO2ChQDtgoYA7YKHAO2CiADtgokA7YKKAO2CiwDtgowA7YKNAO2CjgDtgo8A7YKQAO2CkQDtgpIA7YKTAO2ClADtgpUA7YKWAO2ClwDtgpgA7YKZAO2CmgDtgpsA7YKcAO2CnQDtgp4A7YKfAO2CoADtgqEA7YKiAO2CowDtgqQA7YKlAO2CpgDtgqcA7YKoAO2CqQDtgqoA7YKrAO2CrADtgq0A7YKuAO2CrwDtgrAA7YKxAO2CsgDtgrMA7YK0AO2CtQDtgrYA7YK3AO2CuADtgrkA7YK6AO2CuwDtgrwA7YK9AO2CvgDtgr8A7YOAAO2DgQDtg4IA7YODAO2DhADtg4UA7YOGAO2DhwDtg4gA7YOJAO2DigDtg4sA7YOMAO2DjQDtg44A7YOPAO2DkADtg5EA7YOSAO2DkwDtg5QA7YOVAO2DlgDtg5cA7YOYAO2DmQDtg5oA7YObAO2DnADtg50A7YOeAO2DnwDtg6AA7YOhAO2DogDtg6MA7YOkAO2DpQDtg6YA7YOnAO2DqADtg6kA7YOqAO2DqwDtg6wA7YOtAO2DrgDtg68A7YOwAO2DsQDtg7IA7YOzAO2DtADtg7UA7YO2AO2DtwDtg7gA7YO5AO2DugDtg7sA7YO8AO2DvQDtg74A7YO/AO2EgADthIEA7YSCAO2EgwDthIQA7YSFAO2EhgDthIcA7YSIAO2EiQDthIoA7YSLAO2EjADthI0A7YSOAO2EjwDthJAA7YSRAO2EkgDthJMA7YSUAO2ElQDthJYA7YSXAO2EmADthJkA7YSaAO2EmwDthJwA7YSdAO2EngDthJ8A7YSgAO2EoQDthKIA7YSjAO2EpADthKUA7YSmAO2EpwDthKgA7YSpAO2EqgDthKsA7YSsAO2ErQDthK4A7YSvAO2EsADthLEA7YSyAO2EswDthLQA7YS1AO2EtgDthLcA7YS4AO2EuQDthLoA7YS7AO2EvADthL0A7YS+AO2EvwDthYAA7YWBAO2FggDthYMA7YWEAO2FhQDthYYA7YWHAO2FiADthYkA7YWKAO2FiwDthYwA7YWNAO2FjgDthY8A7YWQAO2FkQDthZIA7YWTAO2FlADthZUA7YWWAO2FlwDthZgA7YWZAO2FmgDthZsA7YWcAO2FnQDthZ4A7YWfAO2FoADthaEA7YWiAO2FowDthaQA7YWlAO2FpgDthacA7YWoAO2FqQDthaoA7YWrAO2FrADtha0A7YWuAO2FrwDthbAA7YWxAO2FsgDthbMA7YW0AO2FtQDthbYA7YW3AO2FuADthbkA7YW6AO2FuwDthbwA7YW9AO2FvgDthb8A7YaAAO2GgQDthoIA7YaDAO2GhADthoUA7YaGAO2GhwDthogA7YaJAO2GigDthosA7YaMAO2GjQDtho4A7YaPAO2GkADthpEA7YaSAO2GkwDthpQA7YaVAO2GlgDthpcA7YaYAO2GmQDthpoA7YabAO2GnADthp0A7YaeAO2GnwDthqAA7YahAO2GogDthqMA7YakAO2GpQDthqYA7YanAO2GqADthqkA7YaqAO2GqwDthqwA7YatAO2GrgDthq8A7YawAO2GsQDthrIA7YazAO2GtADthrUA7Ya2AO2GtwDthrgA7Ya5AO2GugDthrsA7Ya8AO2GvQDthr4A7Ya/AO2HgADth4EA7YeCAO2HgwDth4QA7YeFAO2HhgDth4cA7YeIAO2HiQDth4oA7YeLAO2HjADth40A7YeOAO2HjwDth5AA7YeRAO2HkgDth5MA7YeUAO2HlQDth5YA7YeXAO2HmADth5kA7YeaAO2HmwDth5wA7YedAO2HngDth58A7YegAO2HoQDth6IA7YejAO2HpADth6UA7YemAO2HpwDth6gA7YepAO2HqgDth6sA7YesAO2HrQDth64A7YevAO2HsADth7EA7YeyAO2HswDth7QA7Ye1AO2HtgDth7cA7Ye4AO2HuQDth7oA7Ye7AO2HvADth70A7Ye+AO2HvwDtiIAA7YiBAO2IggDtiIMA7YiEAO2IhQDtiIYA7YiHAO2IiADtiIkA7YiKAO2IiwDtiIwA7YiNAO2IjgDtiI8A7YiQAO2IkQDtiJIA7YiTAO2IlADtiJUA7YiWAO2IlwDtiJgA7YiZAO2ImgDtiJsA7YicAO2InQDtiJ4A7YifAO2IoADtiKEA7YiiAO2IowDtiKQA7YilAO2IpgDtiKcA7YioAO2IqQDtiKoA7YirAO2IrADtiK0A7YiuAO2IrwDtiLAA7YixAO2IsgDtiLMA7Yi0AO2ItQDtiLYA7Yi3AO2IuADtiLkA7Yi6AO2IuwDtiLwA7Yi9AO2IvgDtiL8A7YmAAO2JgQDtiYIA7YmDAO2JhADtiYUA7YmGAO2JhwDtiYgA7YmJAO2JigDtiYsA7YmMAO2JjQDtiY4A7YmPAO2JkADtiZEA7YmSAO2JkwDtiZQA7YmVAO2JlgDtiZcA7YmYAO2JmQDtiZoA7YmbAO2JnADtiZ0A7YmeAO2JnwDtiaAA7YmhAO2JogDtiaMA7YmkAO2JpQDtiaYA7YmnAO2JqADtiakA7YmqAO2JqwDtiawA7YmtAO2JrgDtia8A7YmwAO2JsQDtibIA7YmzAO2JtADtibUA7Ym2AO2JtwDtibgA7Ym5AO2JugDtibsA7Ym8AO2JvQDtib4A7Ym/AO2KgADtioEA7YqCAO2KgwDtioQA7YqFAO2KhgDtiocA7YqIAO2KiQDtiooA7YqLAO2KjADtio0A7YqOAO2KjwDtipAA7YqRAO2KkgDtipMA7YqUAO2KlQDtipYA7YqXAO2KmADtipkA7YqaAO2KmwDtipwA7YqdAO2KngDtip8A7YqgAO2KoQDtiqIA7YqjAO2KpADtiqUA7YqmAO2KpwDtiqgA7YqpAO2KqgDtiqsA7YqsAO2KrQDtiq4A7YqvAO2KsADtirEA7YqyAO2KswDtirQA7Yq1AO2KtgDtircA7Yq4AO2KuQDtiroA7Yq7AO2KvADtir0A7Yq+AO2KvwDti4AA7YuBAO2LggDti4MA7YuEAO2LhQDti4YA7YuHAO2LiADti4kA7YuKAO2LiwDti4wA7YuNAO2LjgDti48A7YuQAO2LkQDti5IA7YuTAO2LlADti5UA7YuWAO2LlwDti5gA7YuZAO2LmgDti5sA7YucAO2LnQDti54A7YufAO2LoADti6EA7YuiAO2LowDti6QA7YulAO2LpgDti6cA7YuoAO2LqQDti6oA7YurAO2LrADti60A7YuuAO2LrwDti7AA7YuxAO2LsgDti7MA7Yu0AO2LtQDti7YA7Yu3AO2LuADti7kA7Yu6AO2LuwDti7wA7Yu9AO2LvgDti78A7YyAAO2MgQDtjIIA7YyDAO2MhADtjIUA7YyGAO2MhwDtjIgA7YyJAO2MigDtjIsA7YyMAO2MjQDtjI4A7YyPAO2MkADtjJEA7YySAO2MkwDtjJQA7YyVAO2MlgDtjJcA7YyYAO2MmQDtjJoA7YybAO2MnADtjJ0A7YyeAO2MnwDtjKAA7YyhAO2MogDtjKMA7YykAO2MpQDtjKYA7YynAO2MqADtjKkA7YyqAO2MqwDtjKwA7YytAO2MrgDtjK8A7YywAO2MsQDtjLIA7YyzAO2MtADtjLUA7Yy2AO2MtwDtjLgA7Yy5AO2MugDtjLsA7Yy8AO2MvQDtjL4A7Yy/AO2NgADtjYEA7Y2CAO2NgwDtjYQA7Y2FAO2NhgDtjYcA7Y2IAO2NiQDtjYoA7Y2LAO2NjADtjY0A7Y2OAO2NjwDtjZAA7Y2RAO2NkgDtjZMA7Y2UAO2NlQDtjZYA7Y2XAO2NmADtjZkA7Y2aAO2NmwDtjZwA7Y2dAO2NngDtjZ8A7Y2gAO2NoQDtjaIA7Y2jAO2NpADtjaUA7Y2mAO2NpwDtjagA7Y2pAO2NqgDtjasA7Y2sAO2NrQDtja4A7Y2vAO2NsADtjbEA7Y2yAO2NswDtjbQA7Y21AO2NtgDtjbcA7Y24AO2NuQDtjboA7Y27AO2NvADtjb0A7Y2+AO2NvwDtjoAA7Y6BAO2OggDtjoMA7Y6EAO2OhQDtjoYA7Y6HAO2OiADtjokA7Y6KAO2OiwDtjowA7Y6NAO2OjgDtjo8A7Y6QAO2OkQDtjpIA7Y6TAO2OlADtjpUA7Y6WAO2OlwDtjpgA7Y6ZAO2OmgDtjpsA7Y6cAO2OnQDtjp4A7Y6fAO2OoADtjqEA7Y6iAO2OowDtjqQA7Y6lAO2OpgDtjqcA7Y6oAO2OqQDtjqoA7Y6rAO2OrADtjq0A7Y6uAO2OrwDtjrAA7Y6xAO2OsgDtjrMA7Y60AO2OtQDtjrYA7Y63AO2OuADtjrkA7Y66AO2OuwDtjrwA7Y69AO2OvgDtjr8A7Y+AAO2PgQDtj4IA7Y+DAO2PhADtj4UA7Y+GAO2PhwDtj4gA7Y+JAO2PigDtj4sA7Y+MAO2PjQDtj44A7Y+PAO2PkADtj5EA7Y+SAO2PkwDtj5QA7Y+VAO2PlgDtj5cA7Y+YAO2PmQDtj5oA7Y+bAO2PnADtj50A7Y+eAO2PnwDtj6AA7Y+hAO2PogDtj6MA7Y+kAO2PpQDtj6YA7Y+nAO2PqADtj6kA7Y+qAO2PqwDtj6wA7Y+tAO2PrgDtj68A7Y+wAO2PsQDtj7IA7Y+zAO2PtADtj7UA7Y+2AO2PtwDtj7gA7Y+5AO2PugDtj7sA7Y+8AO2PvQDtj74A7Y+/AO2QgADtkIEA7ZCCAO2QgwDtkIQA7ZCFAO2QhgDtkIcA7ZCIAO2QiQDtkIoA7ZCLAO2QjADtkI0A7ZCOAO2QjwDtkJAA7ZCRAO2QkgDtkJMA7ZCUAO2QlQDtkJYA7ZCXAO2QmADtkJkA7ZCaAO2QmwDtkJwA7ZCdAO2QngDtkJ8A7ZCgAO2QoQDtkKIA7ZCjAO2QpADtkKUA7ZCmAO2QpwDtkKgA7ZCpAO2QqgDtkKsA7ZCsAO2QrQDtkK4A7ZCvAO2QsADtkLEA7ZCyAO2QswDtkLQA7ZC1AO2QtgDtkLcA7ZC4AO2QuQDtkLoA7ZC7AO2QvADtkL0A7ZC+AO2QvwDtkYAA7ZGBAO2RggDtkYMA7ZGEAO2RhQDtkYYA7ZGHAO2RiADtkYkA7ZGKAO2RiwDtkYwA7ZGNAO2RjgDtkY8A7ZGQAO2RkQDtkZIA7ZGTAO2RlADtkZUA7ZGWAO2RlwDtkZgA7ZGZAO2RmgDtkZsA7ZGcAO2RnQDtkZ4A7ZGfAO2RoADtkaEA7ZGiAO2RowDtkaQA7ZGlAO2RpgDtkacA7ZGoAO2RqQDtkaoA7ZGrAO2RrADtka0A7ZGuAO2RrwDtkbAA7ZGxAO2RsgDtkbMA7ZG0AO2RtQDtkbYA7ZG3AO2RuADtkbkA7ZG6AO2RuwDtkbwA7ZG9AO2RvgDtkb8A7ZKAAO2SgQDtkoIA7ZKDAO2ShADtkoUA7ZKGAO2ShwDtkogA7ZKJAO2SigDtkosA7ZKMAO2SjQDtko4A7ZKPAO2SkADtkpEA7ZKSAO2SkwDtkpQA7ZKVAO2SlgDtkpcA7ZKYAO2SmQDtkpoA7ZKbAO2SnADtkp0A7ZKeAO2SnwDtkqAA7ZKhAO2SogDtkqMA7ZKkAO2SpQDtkqYA7ZKnAO2SqADtkqkA7ZKqAO2SqwDtkqwA7ZKtAO2SrgDtkq8A7ZKwAO2SsQDtkrIA7ZKzAO2StADtkrUA7ZK2AO2StwDtkrgA7ZK5AO2SugDtkrsA7ZK8AO2SvQDtkr4A7ZK/AO2TgADtk4EA7ZOCAO2TgwDtk4QA7ZOFAO2ThgDtk4cA7ZOIAO2TiQDtk4oA7ZOLAO2TjADtk40A7ZOOAO2TjwDtk5AA7ZORAO2TkgDtk5MA7ZOUAO2TlQDtk5YA7ZOXAO2TmADtk5kA7ZOaAO2TmwDtk5wA7ZOdAO2TngDtk58A7ZOgAO2ToQDtk6IA7ZOjAO2TpADtk6UA7ZOmAO2TpwDtk6gA7ZOpAO2TqgDtk6sA7ZOsAO2TrQDtk64A7ZOvAO2TsADtk7EA7ZOyAO2TswDtk7QA7ZO1AO2TtgDtk7cA7ZO4AO2TuQDtk7oA7ZO7AO2TvADtk70A7ZO+AO2TvwDtlIAA7ZSBAO2UggDtlIMA7ZSEAO2UhQDtlIYA7ZSHAO2UiADtlIkA7ZSKAO2UiwDtlIwA7ZSNAO2UjgDtlI8A7ZSQAO2UkQDtlJIA7ZSTAO2UlADtlJUA7ZSWAO2UlwDtlJgA7ZSZAO2UmgDtlJsA7ZScAO2UnQDtlJ4A7ZSfAO2UoADtlKEA7ZSiAO2UowDtlKQA7ZSlAO2UpgDtlKcA7ZSoAO2UqQDtlKoA7ZSrAO2UrADtlK0A7ZSuAO2UrwDtlLAA7ZSxAO2UsgDtlLMA7ZS0AO2UtQDtlLYA7ZS3AO2UuADtlLkA7ZS6AO2UuwDtlLwA7ZS9AO2UvgDtlL8A7ZWAAO2VgQDtlYIA7ZWDAO2VhADtlYUA7ZWGAO2VhwDtlYgA7ZWJAO2VigDtlYsA7ZWMAO2VjQDtlY4A7ZWPAO2VkADtlZEA7ZWSAO2VkwDtlZQA7ZWVAO2VlgDtlZcA7ZWYAO2VmQDtlZoA7ZWbAO2VnADtlZ0A7ZWeAO2VnwDtlaAA7ZWhAO2VogDtlaMA7ZWkAO2VpQDtlaYA7ZWnAO2VqADtlakA7ZWqAO2VqwDtlawA7ZWtAO2VrgDtla8A7ZWwAO2VsQDtlbIA7ZWzAO2VtADtlbUA7ZW2AO2VtwDtlbgA7ZW5AO2VugDtlbsA7ZW8AO2VvQDtlb4A7ZW/AO2WgADtloEA7ZaCAO2WgwDtloQA7ZaFAO2WhgDtlocA7ZaIAO2WiQDtlooA7ZaLAO2WjADtlo0A7ZaOAO2WjwDtlpAA7ZaRAO2WkgDtlpMA7ZaUAO2WlQDtlpYA7ZaXAO2WmADtlpkA7ZaaAO2WmwDtlpwA7ZadAO2WngDtlp8A7ZagAO2WoQDtlqIA7ZajAO2WpADtlqUA7ZamAO2WpwDtlqgA7ZapAO2WqgDtlqsA7ZasAO2WrQDtlq4A7ZavAO2WsADtlrEA7ZayAO2WswDtlrQA7Za1AO2WtgDtlrcA7Za4AO2WuQDtlroA7Za7AO2WvADtlr0A7Za+AO2WvwDtl4AA7ZeBAO2XggDtl4MA7ZeEAO2XhQDtl4YA7ZeHAO2XiADtl4kA7ZeKAO2XiwDtl4wA7ZeNAO2XjgDtl48A7ZeQAO2XkQDtl5IA7ZeTAO2XlADtl5UA7ZeWAO2XlwDtl5gA7ZeZAO2XmgDtl5sA7ZecAO2XnQDtl54A7ZefAO2XoADtl6EA7ZeiAO2XowDtl6QA7ZelAO2XpgDtl6cA7ZeoAO2XqQDtl6oA7ZerAO2XrADtl60A7ZeuAO2XrwDtl7AA7ZexAO2XsgDtl7MA7Ze0AO2XtQDtl7YA7Ze3AO2XuADtl7kA7Ze6AO2XuwDtl7wA7Ze9AO2XvgDtl78A7ZiAAO2YgQDtmIIA7ZiDAO2YhADtmIUA7ZiGAO2YhwDtmIgA7ZiJAO2YigDtmIsA7ZiMAO2YjQDtmI4A7ZiPAO2YkADtmJEA7ZiSAO2YkwDtmJQA7ZiVAO2YlgDtmJcA7ZiYAO2YmQDtmJoA7ZibAO2YnADtmJ0A7ZieAO2YnwDtmKAA7ZihAO2YogDtmKMA7ZikAO2YpQDtmKYA7ZinAO2YqADtmKkA7ZiqAO2YqwDtmKwA7ZitAO2YrgDtmK8A7ZiwAO2YsQDtmLIA7ZizAO2YtADtmLUA7Zi2AO2YtwDtmLgA7Zi5AO2YugDtmLsA7Zi8AO2YvQDtmL4A7Zi/AO2ZgADtmYEA7ZmCAO2ZgwDtmYQA7ZmFAO2ZhgDtmYcA7ZmIAO2ZiQDtmYoA7ZmLAO2ZjADtmY0A7ZmOAO2ZjwDtmZAA7ZmRAO2ZkgDtmZMA7ZmUAO2ZlQDtmZYA7ZmXAO2ZmADtmZkA7ZmaAO2ZmwDtmZwA7ZmdAO2ZngDtmZ8A7ZmgAO2ZoQDtmaIA7ZmjAO2ZpADtmaUA7ZmmAO2ZpwDtmagA7ZmpAO2ZqgDtmasA7ZmsAO2ZrQDtma4A7ZmvAO2ZsADtmbEA7ZmyAO2ZswDtmbQA7Zm1AO2ZtgDtmbcA7Zm4AO2ZuQDtmboA7Zm7AO2ZvADtmb0A7Zm+AO2ZvwDtmoAA7ZqBAO2aggDtmoMA7ZqEAO2ahQDtmoYA7ZqHAO2aiADtmokA7ZqKAO2aiwDtmowA7ZqNAO2ajgDtmo8A7ZqQAO2akQDtmpIA7ZqTAO2alADtmpUA7ZqWAO2alwDtmpgA7ZqZAO2amgDtmpsA7ZqcAO2anQDtmp4A7ZqfAO2aoADtmqEA7ZqiAO2aowDtmqQA7ZqlAO2apgDtmqcA7ZqoAO2aqQDtmqoA7ZqrAO2arADtmq0A7ZquAO2arwDtmrAA7ZqxAO2asgDtmrMA7Zq0AO2atQDtmrYA7Zq3AO2auADtmrkA7Zq6AO2auwDtmrwA7Zq9AO2avgDtmr8A7ZuAAO2bgQDtm4IA7ZuDAO2bhADtm4UA7ZuGAO2bhwDtm4gA7ZuJAO2bigDtm4sA7ZuMAO2bjQDtm44A7ZuPAO2bkADtm5EA7ZuSAO2bkwDtm5QA7ZuVAO2blgDtm5cA7ZuYAO2bmQDtm5oA7ZubAO2bnADtm50A7ZueAO2bnwDtm6AA7ZuhAO2bogDtm6MA7ZukAO2bpQDtm6YA7ZunAO2bqADtm6kA7ZuqAO2bqwDtm6wA7ZutAO2brgDtm68A7ZuwAO2bsQDtm7IA7ZuzAO2btADtm7UA7Zu2AO2btwDtm7gA7Zu5AO2bugDtm7sA7Zu8AO2bvQDtm74A7Zu/AO2cgADtnIEA7ZyCAO2cgwDtnIQA7ZyFAO2chgDtnIcA7ZyIAO2ciQDtnIoA7ZyLAO2cjADtnI0A7ZyOAO2cjwDtnJAA7ZyRAO2ckgDtnJMA7ZyUAO2clQDtnJYA7ZyXAO2cmADtnJkA7ZyaAO2cmwDtnJwA7ZydAO2cngDtnJ8A7ZygAO2coQDtnKIA7ZyjAO2cpADtnKUA7ZymAO2cpwDtnKgA7ZypAO2cqgDtnKsA7ZysAO2crQDtnK4A7ZyvAO2csADtnLEA7ZyyAO2cswDtnLQA7Zy1AO2ctgDtnLcA7Zy4AO2cuQDtnLoA7Zy7AO2cvADtnL0A7Zy+AO2cvwDtnYAA7Z2BAO2dggDtnYMA7Z2EAO2dhQDtnYYA7Z2HAO2diADtnYkA7Z2KAO2diwDtnYwA7Z2NAO2djgDtnY8A7Z2QAO2dkQDtnZIA7Z2TAO2dlADtnZUA7Z2WAO2dlwDtnZgA7Z2ZAO2dmgDtnZsA7Z2cAO2dnQDtnZ4A7Z2fAO2doADtnaEA7Z2iAO2dowDtnaQA7Z2lAO2dpgDtnacA7Z2oAO2dqQDtnaoA7Z2rAO2drADtna0A7Z2uAO2drwDtnbAA7Z2xAO2dsgDtnbMA7Z20AO2dtQDtnbYA7Z23AO2duADtnbkA7Z26AO2duwDtnbwA7Z29AO2dvgDtnb8A7Z6AAO2egQDtnoIA7Z6DAO2ehADtnoUA7Z6GAO2ehwDtnogA7Z6JAO2eigDtnosA7Z6MAO2ejQDtno4A7Z6PAO2ekADtnpEA7Z6SAO2ekwDtnpQA7Z6VAO2elgDtnpcA7Z6YAO2emQDtnpoA7Z6bAO2enADtnp0A7Z6eAO2enwDtnqAA7Z6hAO2eogDtnqMA8JGCmgDwkYKcAPCRgqsA8JGErgDwkYSvAPCRjYsA8JGNjADwkZK7APCRkrwA8JGSvgDwkZa6APCRlrsA8JGkuADwnYWX8J2FpQDwnYWY8J2FpQDwnYWY8J2FpfCdha4A8J2FmPCdhaXwnYWvAPCdhZjwnYWl8J2FsADwnYWY8J2FpfCdhbEA8J2FmPCdhaXwnYWyAPCdhrnwnYWlAPCdhrnwnYWl8J2FrgDwnYa58J2FpfCdha8A8J2GuvCdhaUA8J2GuvCdhaXwnYWuAPCdhrrwnYWl8J2FrwDwoISiAPCglJwA8KCUpQDwoJWLAPCgmLoA8KCghADwoKOeAPCgqKwA8KCtowDwoZOkAPChmqgA8KGbqgDwoaeIAPChrJgA8KG0iwDwobekAPCht6YA8KKGgwDwooafAPCijLEA8KKblADwoqGEAPCioYoA8KKsjADwoq+xAPCjgIoA8KOKuADwo42fAPCjjpMA8KOOnADwo4+DAPCjj5UA8KORrQDwo5qjAPCjoqcA8KOqjQDwo6u6APCjsrwA8KO0ngDwo7uRAPCjvZ4A8KO+jgDwpImjAPCki64A8KSOqwDwpJiIAPCknLUA8KSglADwpLC2APCkspIA8KS+oQDwpL64APClgYQA8KWDsgDwpYOzAPClhJkA8KWEswDwpYmJAPClkJ0A8KWYpgDwpZqaAPClm4UA8KWlvADwpaqnAPClrqsA8KWygADwpbOQAPClvoYA8KaHmgDwpoioAPCmiYcA8KaLmQDwpoy+APCmk5oA8KaUowDwppaoAPCmnqcA8KaetQDwpqy8APCmsLYA8KazlQDwprWrAPCmvKwA8Ka+sQDwp4OSAPCnj4oA8KeZpwDwp6KuAPCnpaYA8KeyqADwp7uTAPCnvK8A8KiXkgDwqJetAPConK4A8KivugDwqLW3APCphYUA8KmHnwDwqYiaAPCpkIoA8KmSlgDwqZa2APCprLAA8KqDjgDwqoSFAPCqiI4A8KqKkQDwqo6SAPCqmIAA" - }, - { - "type": "Replace", - "pattern": { - "Regex": " {2,}" - }, - "content": "▁" - } - ] - }, - "pre_tokenizer": { - "type": "Metaspace", - "replacement": "▁", - "prepend_scheme": "always", - "split": true - }, - "post_processor": null, - "decoder": { - "type": "Metaspace", - "replacement": "▁", - "prepend_scheme": "always", - "split": true - }, - "model": { - "type": "BPE", - "dropout": null, - "unk_token": "", - "continuing_subword_prefix": null, - "end_of_word_suffix": null, - "fuse_unk": true, - "byte_fallback": true, - "ignore_merges": false, - "vocab": { - "": 0, - "▁t": 1, - "▁th": 2, - "▁a": 3, - "in": 4, - "re": 5, - "▁the": 6, - "▁w": 7, - "▁s": 8, - "▁o": 9, - "er": 10, - "ou": 11, - "at": 12, - "nd": 13, - "it": 14, - "▁h": 15, - "▁c": 16, - "▁b": 17, - "is": 18, - "en": 19, - "on": 20, - "ing": 21, - "▁f": 22, - "▁to": 23, - "▁m": 24, - "es": 25, - "▁p": 26, - "or": 27, - "an": 28, - "▁d": 29, - "ll": 30, - "▁I": 31, - "ed": 32, - "▁and": 33, - "▁l": 34, - "▁of": 35, - "▁in": 36, - "▁y": 37, - "ar": 38, - "▁g": 39, - "▁you": 40, - "as": 41, - "om": 42, - "▁n": 43, - "ve": 44, - "▁that": 45, - "le": 46, - "ic": 47, - "us": 48, - "ow": 49, - "et": 50, - "al": 51, - "▁e": 52, - "ut": 53, - "▁it": 54, - "ot": 55, - "▁be": 56, - "▁T": 57, - "ion": 58, - "▁is": 59, - "▁wh": 60, - "▁re": 61, - "▁on": 62, - "▁we": 63, - "ent": 64, - "▁A": 65, - "ay": 66, - "▁ha": 67, - "▁Th": 68, - "id": 69, - "▁S": 70, - "ac": 71, - "gh": 72, - "ver": 73, - "ke": 74, - "▁for": 75, - "im": 76, - "ly": 77, - "ur": 78, - "ld": 79, - "▁he": 80, - "▁st": 81, - "all": 82, - "ro": 83, - "st": 84, - "se": 85, - "ct": 86, - "ith": 87, - "ir": 88, - "am": 89, - "▁this": 90, - "if": 91, - "▁W": 92, - "oo": 93, - "ri": 94, - "▁was": 95, - "ght": 96, - "▁u": 97, - "▁with": 98, - "ad": 99, - "ch": 100, - "▁se": 101, - "▁k": 102, - "▁an": 103, - "▁The": 104, - "▁li": 105, - "▁do": 106, - "▁B": 107, - "▁have": 108, - "▁as": 109, - "th": 110, - "▁are": 111, - "▁sh": 112, - "ust": 113, - "ce": 114, - "ally": 115, - "ill": 116, - "▁H": 117, - "▁j": 118, - "ter": 119, - "▁go": 120, - "▁And": 121, - "ation": 122, - "▁C": 123, - "▁so": 124, - "ome": 125, - "▁not": 126, - "op": 127, - "il": 128, - "ore": 129, - "▁ne": 130, - "▁can": 131, - "▁me": 132, - "▁at": 133, - "ould": 134, - "ant": 135, - "▁M": 136, - "▁like": 137, - "ere": 138, - "▁they": 139, - "ra": 140, - "ers": 141, - "▁ab": 142, - "▁de": 143, - "▁kn": 144, - "ge": 145, - "▁Y": 146, - "▁ch": 147, - "ul": 148, - "pp": 149, - "▁or": 150, - "▁al": 151, - "▁con": 152, - "▁com": 153, - "ess": 154, - "▁su": 155, - "out": 156, - "▁your": 157, - "▁So": 158, - "ate": 159, - "▁one": 160, - "▁all": 161, - "▁ex": 162, - "est": 163, - "▁fr": 164, - "▁just": 165, - "▁pro": 166, - "▁know": 167, - "▁O": 168, - "ain": 169, - "▁but": 170, - "ol": 171, - "ive": 172, - "▁v": 173, - "use": 174, - "very": 175, - "art": 176, - "qu": 177, - "▁my": 178, - "el": 179, - "▁N": 180, - "nt": 181, - "▁It": 182, - "▁what": 183, - "ab": 184, - "▁P": 185, - "▁wor": 186, - "▁out": 187, - "▁there": 188, - "▁up": 189, - "um": 190, - "▁from": 191, - "pe": 192, - "▁tw": 193, - "▁r": 194, - "and": 195, - "ight": 196, - "ort": 197, - "un": 198, - "▁L": 199, - "ist": 200, - "▁about": 201, - "ide": 202, - "ig": 203, - "ake": 204, - "▁D": 205, - "em": 206, - "os": 207, - "king": 208, - "rou": 209, - "ind": 210, - "our": 211, - "res": 212, - "▁We": 213, - "▁get": 214, - "▁E": 215, - "▁G": 216, - "ack": 217, - "▁le": 218, - "ity": 219, - "od": 220, - "▁F": 221, - "ard": 222, - "▁pl": 223, - "▁our": 224, - "▁int": 225, - "ment": 226, - "▁will": 227, - "ies": 228, - "▁by": 229, - "ink": 230, - "ca": 231, - "▁if": 232, - "red": 233, - "her": 234, - "ie": 235, - "▁us": 236, - "▁some": 237, - "▁don": 238, - "ven": 239, - "ood": 240, - "ast": 241, - "▁R": 242, - "▁his": 243, - "▁tim": 244, - "▁tr": 245, - "▁more": 246, - "ich": 247, - "ous": 248, - "ame": 249, - "▁going": 250, - "▁had": 251, - "▁them": 252, - "ook": 253, - "▁pe": 254, - "▁Wh": 255, - "▁You": 256, - "▁But": 257, - "ine": 258, - "▁here": 259, - "▁would": 260, - "cause": 261, - "right": 262, - "so": 263, - "ost": 264, - "ure": 265, - "▁has": 266, - "ect": 267, - "▁think": 268, - "▁fe": 269, - "ong": 270, - "▁see": 271, - "▁when": 272, - "▁who": 273, - "▁were": 274, - "▁really": 275, - "▁their": 276, - "▁want": 277, - "one": 278, - "ople": 279, - "▁then": 280, - "▁time": 281, - "▁sa": 282, - "ap": 283, - "▁te": 284, - "▁He": 285, - "▁ye": 286, - "ck": 287, - "▁her": 288, - "▁thing": 289, - "▁right": 290, - "▁which": 291, - "itt": 292, - "ice": 293, - "act": 294, - "▁people": 295, - "ty": 296, - "▁two": 297, - "▁J": 298, - "▁im": 299, - "ther": 300, - "ci": 301, - "ose": 302, - "▁cl": 303, - "▁qu": 304, - "▁man": 305, - "▁also": 306, - "ree": 307, - "▁en": 308, - "ud": 309, - "▁how": 310, - "reat": 311, - "ak": 312, - "hing": 313, - "ag": 314, - "▁any": 315, - "ff": 316, - "ace": 317, - "per": 318, - "▁because": 319, - "▁very": 320, - "own": 321, - "▁ad": 322, - "▁act": 323, - "▁been": 324, - "▁now": 325, - "▁ag": 326, - "▁into": 327, - "▁comp": 328, - "ars": 329, - "ions": 330, - "are": 331, - "ite": 332, - "iv": 333, - "▁these": 334, - "ays": 335, - "ep": 336, - "▁This": 337, - "▁she": 338, - "ans": 339, - "ah": 340, - "een": 341, - "▁over": 342, - "ry": 343, - "▁lo": 344, - "age": 345, - "▁pr": 346, - "▁sp": 347, - "ue": 348, - "▁co": 349, - "ick": 350, - "ber": 351, - "▁did": 352, - "ip": 353, - "ach": 354, - "▁back": 355, - "▁no": 356, - "▁cont": 357, - "▁other": 358, - "▁every": 359, - "pt": 360, - "▁need": 361, - "▁him": 362, - "▁U": 363, - "▁In": 364, - "▁work": 365, - "irst": 366, - "▁part": 367, - "▁look": 368, - "ittle": 369, - "ble": 370, - "iz": 371, - "▁un": 372, - "▁make": 373, - "omet": 374, - "nder": 375, - "ish": 376, - "na": 377, - "▁little": 378, - "▁off": 379, - "▁than": 380, - "▁got": 381, - "ually": 382, - "▁per": 383, - "▁good": 384, - "▁way": 385, - "▁could": 386, - "▁ac": 387, - "▁imp": 388, - "able": 389, - "▁where": 390, - "iff": 391, - "▁That": 392, - "▁res": 393, - "ount": 394, - "pl": 395, - "ance": 396, - "▁first": 397, - "▁ro": 398, - "▁pre": 399, - "ass": 400, - "▁say": 401, - "int": 402, - "ated": 403, - "ire": 404, - "uch": 405, - "ase": 406, - "▁somet": 407, - "ound": 408, - "▁down": 409, - "▁diff": 410, - "sel": 411, - "▁gu": 412, - "▁am": 413, - "ress": 414, - "▁lot": 415, - "ence": 416, - "▁dis": 417, - "orm": 418, - "ix": 419, - "▁po": 420, - "ving": 421, - "enty": 422, - "▁K": 423, - "▁spe": 424, - "und": 425, - "he": 426, - "▁much": 427, - "▁ar": 428, - "round": 429, - "▁app": 430, - "co": 431, - "ark": 432, - "▁new": 433, - "ater": 434, - "ult": 435, - "end": 436, - "▁even": 437, - "▁start": 438, - "ations": 439, - "rough": 440, - "ile": 441, - "fter": 442, - "▁well": 443, - "be": 444, - "▁They": 445, - "▁three": 446, - "ign": 447, - "ild": 448, - "▁said": 449, - "ough": 450, - "ang": 451, - "▁too": 452, - "ade": 453, - "▁bl": 454, - "ens": 455, - "▁inc": 456, - "ia": 457, - "▁those": 458, - "▁mo": 459, - "▁take": 460, - "▁through": 461, - "▁fl": 462, - "▁kind": 463, - "▁things": 464, - "▁bet": 465, - "▁only": 466, - "▁St": 467, - "▁let": 468, - "cess": 469, - "▁Ch": 470, - "ary": 471, - "vel": 472, - "▁If": 473, - "xt": 474, - "other": 475, - "av": 476, - "ical": 477, - "ord": 478, - "▁again": 479, - "▁something": 480, - "onna": 481, - "fore": 482, - "▁may": 483, - "ting": 484, - "▁bu": 485, - "▁differe": 486, - "urn": 487, - "▁gonna": 488, - "▁does": 489, - "uct": 490, - "og": 491, - "▁twenty": 492, - "▁gr": 493, - "▁Ye": 494, - "wn": 495, - "▁should": 496, - "▁comm": 497, - "ition": 498, - "▁under": 499, - "▁hel": 500, - "ory": 501, - "▁fo": 502, - "▁use": 503, - "igh": 504, - "ife": 505, - "▁actually": 506, - "▁tal": 507, - "▁call": 508, - "ents": 509, - "ious": 510, - "ull": 511, - "▁There": 512, - "▁Yeah": 513, - "▁most": 514, - "▁ke": 515, - "ors": 516, - "ved": 517, - "ys": 518, - "▁sc": 519, - "▁happ": 520, - "ope": 521, - "▁help": 522, - "atch": 523, - "▁What": 524, - "▁rem": 525, - "ple": 526, - "▁Now": 527, - "▁br": 528, - "ool": 529, - "oth": 530, - "▁four": 531, - "self": 532, - "▁str": 533, - "ne": 534, - "thing": 535, - "▁put": 536, - "ial": 537, - "▁great": 538, - "ail": 539, - "ub": 540, - "ning": 541, - "▁sm": 542, - "▁feel": 543, - "▁five": 544, - "ody": 545, - "undred": 546, - "iss": 547, - "ank": 548, - "get": 549, - "aking": 550, - "▁many": 551, - "▁hundred": 552, - "▁years": 553, - "▁being": 554, - "▁come": 555, - "▁mean": 556, - "ily": 557, - "▁different": 558, - "▁after": 559, - "▁ser": 560, - "▁show": 561, - "form": 562, - "ful": 563, - "oy": 564, - "▁six": 565, - "▁vide": 566, - "▁V": 567, - "▁its": 568, - "▁point": 569, - "▁day": 570, - "▁des": 571, - "ons": 572, - "▁bit": 573, - "▁bel": 574, - "▁before": 575, - "▁aw": 576, - "▁end": 577, - "▁Oh": 578, - "▁still": 579, - "ath": 580, - "▁long": 581, - "▁'": 582, - "ise": 583, - "ob": 584, - "day": 585, - "▁add": 586, - "ft": 587, - "ves": 588, - "ces": 589, - "ady": 590, - "▁cr": 591, - "▁around": 592, - "▁try": 593, - "les": 594, - "vers": 595, - "kay": 596, - "ian": 597, - "ates": 598, - "▁find": 599, - "ward": 600, - "▁As": 601, - "▁eight": 602, - "lic": 603, - "▁same": 604, - "▁pos": 605, - "▁em": 606, - "▁made": 607, - "▁supp": 608, - "▁life": 609, - "▁Be": 610, - "pect": 611, - "▁dec": 612, - "▁play": 613, - "ange": 614, - "▁att": 615, - "▁pers": 616, - "ways": 617, - "▁high": 618, - "▁hand": 619, - "▁next": 620, - "▁cons": 621, - "▁own": 622, - "▁inv": 623, - "ower": 624, - "▁ind": 625, - "ert": 626, - "ng": 627, - "ave": 628, - "▁year": 629, - "▁big": 630, - "ating": 631, - "▁world": 632, - "▁rel": 633, - "▁sure": 634, - "▁tra": 635, - "ew": 636, - "ered": 637, - "▁fin": 638, - "▁Well": 639, - "▁sl": 640, - "▁doing": 641, - "bs": 642, - "▁set": 643, - "▁rec": 644, - "ual": 645, - "cial": 646, - "▁ph": 647, - "erm": 648, - "▁love": 649, - "ph": 650, - "▁real": 651, - "▁last": 652, - "ict": 653, - "▁bo": 654, - "▁ra": 655, - "ible": 656, - "▁wr": 657, - "mer": 658, - "▁count": 659, - "ities": 660, - "▁always": 661, - "inet": 662, - "ments": 663, - "uc": 664, - "▁might": 665, - "▁inter": 666, - "▁video": 667, - "gin": 668, - "▁tell": 669, - "▁never": 670, - "vent": 671, - "▁import": 672, - "ied": 673, - "▁sy": 674, - "▁How": 675, - "ically": 676, - "ought": 677, - "▁thir": 678, - "▁rep": 679, - "ks": 680, - "ib": 681, - "▁fam": 682, - "ject": 683, - "▁bas": 684, - "▁She": 685, - "▁give": 686, - "akes": 687, - "▁ninet": 688, - "▁reg": 689, - "▁min": 690, - "▁op": 691, - "▁def": 692, - "▁didn": 693, - "te": 694, - "▁cour": 695, - "▁why": 696, - "▁ent": 697, - "▁place": 698, - "▁ins": 699, - "▁car": 700, - "ather": 701, - "▁person": 702, - "ular": 703, - "▁inst": 704, - "▁prod": 705, - "lect": 706, - "▁Al": 707, - "▁today": 708, - "▁bec": 709, - "▁sur": 710, - "▁All": 711, - "▁another": 712, - "▁bus": 713, - "▁keep": 714, - "ell": 715, - "ese": 716, - "riend": 717, - "▁quest": 718, - "▁talk": 719, - "als": 720, - "ings": 721, - "▁mon": 722, - "cond": 723, - "old": 724, - "▁acc": 725, - "▁la": 726, - "▁num": 727, - "ident": 728, - "▁che": 729, - "iness": 730, - "▁turn": 731, - "▁ear": 732, - "▁No": 733, - "ousand": 734, - "▁better": 735, - "ific": 736, - "▁loo": 737, - "▁gl": 738, - "oc": 739, - "▁important": 740, - "ited": 741, - "▁An": 742, - "▁thousand": 743, - "ility": 744, - "llow": 745, - "▁used": 746, - "▁gen": 747, - "▁sim": 748, - "li": 749, - "▁happen": 750, - "▁Un": 751, - "▁Let": 752, - "air": 753, - "ock": 754, - "ably": 755, - "gg": 756, - "▁watch": 757, - "▁For": 758, - "▁sw": 759, - "ren": 760, - "ute": 761, - "ever": 762, - "▁pol": 763, - "▁sch": 764, - "▁When": 765, - "▁such": 766, - "▁fif": 767, - "▁home": 768, - "▁cle": 769, - "▁contin": 770, - "ouse": 771, - "▁friend": 772, - "uring": 773, - "▁Okay": 774, - "gr": 775, - "▁able": 776, - "▁stud": 777, - "▁eff": 778, - "hip": 779, - "body": 780, - "▁top": 781, - "ness": 782, - "▁exper": 783, - "▁pret": 784, - "▁both": 785, - "▁done": 786, - "cri": 787, - "▁mark": 788, - "▁while": 789, - "▁old": 790, - "ros": 791, - "ont": 792, - "▁second": 793, - "ative": 794, - "▁thought": 795, - "▁best": 796, - "▁found": 797, - "iew": 798, - "▁belie": 799, - "▁each": 800, - "erest": 801, - "▁tri": 802, - "▁eas": 803, - "▁ca": 804, - "▁fact": 805, - "▁care": 806, - "▁fun": 807, - "atter": 808, - "ures": 809, - "▁head": 810, - "▁lear": 811, - "▁water": 812, - "▁hard": 813, - "▁few": 814, - "▁side": 815, - "ween": 816, - "▁exp": 817, - "▁away": 818, - "its": 819, - "▁ext": 820, - "lud": 821, - "▁run": 822, - "▁trans": 823, - "ince": 824, - "▁sk": 825, - "▁open": 826, - "cus": 827, - "▁between": 828, - "▁called": 829, - "▁wee": 830, - "▁pretty": 831, - "ason": 832, - "▁far": 833, - "ember": 834, - "omm": 835, - "▁interest": 836, - "any": 837, - "ner": 838, - "uff": 839, - "▁pres": 840, - "▁cur": 841, - "▁child": 842, - "ee": 843, - "▁toget": 844, - "▁together": 845, - "olog": 846, - "▁God": 847, - "ond": 848, - "▁char": 849, - "▁looking": 850, - "stem": 851, - "az": 852, - "cent": 853, - "▁ob": 854, - "▁ass": 855, - "land": 856, - "▁doesn": 857, - "▁business": 858, - "▁course": 859, - "▁ten": 860, - "ps": 861, - "arch": 862, - "ced": 863, - "ms": 864, - "ize": 865, - "nce": 866, - "▁ref": 867, - "▁name": 868, - "ross": 869, - "▁grow": 870, - "oney": 871, - "▁went": 872, - "ics": 873, - "teen": 874, - "▁cou": 875, - "▁prob": 876, - "▁ret": 877, - "▁guys": 878, - "▁came": 879, - "ash": 880, - "led": 881, - "▁Eur": 882, - "ues": 883, - "▁ide": 884, - "gan": 885, - "▁everything": 886, - "▁getting": 887, - "▁ask": 888, - "▁cor": 889, - "▁build": 890, - "▁sign": 891, - "▁small": 892, - "uck": 893, - "▁el": 894, - "▁col": 895, - "▁Is": 896, - "ational": 897, - "stand": 898, - "cy": 899, - "▁conf": 900, - "der": 901, - "▁bre": 902, - "▁cap": 903, - "▁mod": 904, - "ets": 905, - "ike": 906, - "▁number": 907, - "▁comple": 908, - "ertain": 909, - "▁ever": 910, - "▁coll": 911, - "▁hum": 912, - "▁Europe": 913, - "▁cre": 914, - "▁met": 915, - "▁exam": 916, - "▁move": 917, - "▁pass": 918, - "▁left": 919, - "▁system": 920, - "▁includ": 921, - "▁Thank": 922, - "cept": 923, - "▁wom": 924, - "▁product": 925, - "ten": 926, - "▁rest": 927, - "▁probably": 928, - "▁dri": 929, - "▁Do": 930, - "▁gener": 931, - "▁anything": 932, - "▁lar": 933, - "▁My": 934, - "▁school": 935, - "▁lead": 936, - "▁sub": 937, - "▁ty": 938, - "▁plan": 939, - "▁seem": 940, - "▁whole": 941, - "irect": 942, - "▁light": 943, - "▁must": 944, - "▁mom": 945, - "▁opp": 946, - "▁support": 947, - "▁family": 948, - "ices": 949, - "amp": 950, - "▁proble": 951, - "▁dr": 952, - "ready": 953, - "▁using": 954, - "ense": 955, - "▁prov": 956, - "ush": 957, - "ax": 958, - "▁power": 959, - "▁Re": 960, - "alth": 961, - "▁ev": 962, - "▁stand": 963, - "▁war": 964, - "ts": 965, - "▁": 966, - "e": 967, - "t": 968, - "o": 969, - "a": 970, - "n": 971, - "i": 972, - "s": 973, - "r": 974, - "h": 975, - "l": 976, - "d": 977, - "u": 978, - "c": 979, - "m": 980, - "y": 981, - "g": 982, - "w": 983, - "f": 984, - "p": 985, - ".": 986, - "b": 987, - ",": 988, - "v": 989, - "k": 990, - "'": 991, - "I": 992, - "T": 993, - "A": 994, - "S": 995, - "x": 996, - "W": 997, - "j": 998, - "B": 999, - "C": 1000, - "H": 1001, - "?": 1002, - "M": 1003, - "O": 1004, - "Y": 1005, - "N": 1006, - "P": 1007, - "E": 1008, - "q": 1009, - "L": 1010, - "D": 1011, - "z": 1012, - "G": 1013, - "F": 1014, - "R": 1015, - "!": 1016, - "J": 1017, - "U": 1018, - "K": 1019, - "V": 1020, - "Q": 1021, - "Z": 1022, - "X": 1023 - }, - "merges": [ - "▁ t", - "▁t h", - "▁ th", - "▁ a", - "i n", - "r e", - "▁th e", - "▁t he", - "▁ w", - "▁ s", - "▁ o", - "e r", - "o u", - "a t", - "n d", - "i t", - "▁ h", - "▁ c", - "▁ b", - "i s", - "e n", - "o n", - "in g", - "i ng", - "▁ f", - "▁t o", - "▁ m", - "e s", - "▁ p", - "o r", - "a n", - "▁ d", - "l l", - "▁ I", - "e d", - "▁an d", - "▁a nd", - "▁ and", - "▁ l", - "▁o f", - "▁ in", - "▁ y", - "a r", - "▁ g", - "▁y ou", - "a s", - "o m", - "▁ n", - "v e", - "▁th at", - "l e", - "i c", - "u s", - "o w", - "e t", - "a l", - "▁ e", - "u t", - "▁ it", - "o t", - "▁b e", - "▁ be", - "▁ T", - "i on", - "▁ is", - "▁w h", - "▁r e", - "▁ re", - "▁o n", - "▁ on", - "▁w e", - "en t", - "e nt", - "▁ A", - "a y", - "▁h a", - "▁T h", - "i d", - "▁ S", - "a c", - "g h", - "ve r", - "v er", - "k e", - "▁fo r", - "▁f or", - "i m", - "l y", - "u r", - "l d", - "▁h e", - "▁ he", - "▁s t", - "▁ st", - "al l", - "a ll", - "r o", - "s t", - "s e", - "c t", - "it h", - "i th", - "i r", - "a m", - "▁th is", - "i f", - "▁ W", - "o o", - "r i", - "▁w as", - "gh t", - "▁ u", - "▁w ith", - "a d", - "c h", - "▁s e", - "▁ se", - "▁ k", - "▁a n", - "▁ an", - "▁Th e", - "▁T he", - "▁l i", - "▁ li", - "▁d o", - "▁ B", - "▁ha ve", - "▁h ave", - "▁a s", - "▁ as", - "t h", - "▁ar e", - "▁a re", - "▁ are", - "▁s h", - "us t", - "u st", - "c e", - "all y", - "al ly", - "il l", - "i ll", - "▁ H", - "▁ j", - "te r", - "t er", - "▁g o", - "▁An d", - "▁A nd", - "at ion", - "▁ C", - "▁s o", - "▁ so", - "om e", - "▁no t", - "▁n ot", - "o p", - "i l", - "or e", - "o re", - "▁n e", - "▁ ne", - "▁ca n", - "▁c an", - "▁m e", - "▁a t", - "▁ at", - "ou ld", - "an t", - "a nt", - "▁ M", - "▁li ke", - "▁l ike", - "er e", - "e re", - "▁the y", - "r a", - "er s", - "▁a b", - "▁ ab", - "▁d e", - "▁k n", - "g e", - "▁ Y", - "▁c h", - "▁ ch", - "u l", - "p p", - "▁o r", - "▁ or", - "▁a l", - "▁ al", - "▁co n", - "▁c on", - "▁co m", - "▁c om", - "es s", - "▁s u", - "ou t", - "o ut", - "▁you r", - "▁y our", - "▁S o", - "at e", - "a te", - "▁on e", - "▁o ne", - "▁ one", - "▁al l", - "▁a ll", - "▁ all", - "▁e x", - "es t", - "e st", - "▁f r", - "▁j ust", - "▁pr o", - "▁p ro", - "▁kn ow", - "▁ O", - "a in", - "▁bu t", - "▁b ut", - "o l", - "iv e", - "i ve", - "▁ v", - "us e", - "u se", - "ver y", - "ve ry", - "ar t", - "q u", - "▁m y", - "e l", - "▁ N", - "n t", - "▁I t", - "▁wh at", - "a b", - "▁ P", - "▁w or", - "▁o ut", - "▁ out", - "▁the re", - "▁th ere", - "▁u p", - "u m", - "▁fr om", - "p e", - "▁t w", - "▁ r", - "an d", - "a nd", - "igh t", - "i ght", - "or t", - "u n", - "▁ L", - "is t", - "i st", - "▁ab out", - "id e", - "i g", - "ak e", - "a ke", - "▁ D", - "e m", - "o s", - "k ing", - "ro u", - "r ou", - "in d", - "i nd", - "ou r", - "o ur", - "re s", - "r es", - "▁W e", - "▁g et", - "▁ get", - "▁ E", - "▁ G", - "ac k", - "a ck", - "▁l e", - "▁ le", - "it y", - "i ty", - "o d", - "▁ F", - "ar d", - "▁p l", - "▁ pl", - "▁o ur", - "▁ our", - "▁in t", - "▁ int", - "m ent", - "▁w ill", - "ie s", - "i es", - "▁b y", - "in k", - "c a", - "▁ if", - "re d", - "r ed", - "he r", - "h er", - "i e", - "▁u s", - "▁ us", - "▁s ome", - "▁do n", - "▁d on", - "ve n", - "v en", - "oo d", - "o od", - "as t", - "a st", - "▁ R", - "▁h is", - "▁t im", - "▁t r", - "▁mo re", - "▁m ore", - "ic h", - "i ch", - "ou s", - "o us", - "am e", - "▁go ing", - "▁ha d", - "▁h ad", - "▁the m", - "▁th em", - "oo k", - "▁p e", - "▁ pe", - "▁W h", - "▁Y ou", - "▁B ut", - "in e", - "i ne", - "▁her e", - "▁he re", - "▁h ere", - "▁w ould", - "ca use", - "ri ght", - "r ight", - "s o", - "os t", - "o st", - "ur e", - "u re", - "▁ha s", - "▁h as", - "e ct", - "▁th ink", - "▁f e", - "on g", - "o ng", - "▁se e", - "▁s ee", - "▁wh en", - "▁wh o", - "▁we re", - "▁w ere", - "▁real ly", - "▁re ally", - "▁the ir", - "▁w ant", - "on e", - "o ne", - "op le", - "o ple", - "▁the n", - "▁th en", - "▁tim e", - "▁s a", - "a p", - "▁t e", - "▁ te", - "▁H e", - "▁y e", - "c k", - "▁he r", - "▁h er", - "▁ her", - "▁th ing", - "▁t hing", - "▁ thing", - "▁r ight", - "▁ right", - "▁wh ich", - "it t", - "ic e", - "i ce", - "ac t", - "a ct", - "▁pe ople", - "t y", - "▁tw o", - "▁ J", - "▁ im", - "th er", - "t her", - "c i", - "os e", - "o se", - "▁c l", - "▁ qu", - "▁m an", - "▁al so", - "re e", - "r ee", - "▁e n", - "▁ en", - "u d", - "▁h ow", - "re at", - "a k", - "h ing", - "a g", - "▁an y", - "▁ any", - "f f", - "ac e", - "a ce", - "pe r", - "p er", - "▁be cause", - "▁ very", - "ow n", - "o wn", - "▁a d", - "▁ ad", - "▁ac t", - "▁a ct", - "▁ act", - "▁be en", - "▁b een", - "▁no w", - "▁n ow", - "▁a g", - "▁ ag", - "▁int o", - "▁com p", - "ar s", - "ion s", - "i ons", - "ar e", - "a re", - "it e", - "i te", - "i v", - "▁the se", - "▁th ese", - "ay s", - "a ys", - "e p", - "▁Th is", - "▁sh e", - "▁s he", - "an s", - "a h", - "ee n", - "e en", - "▁o ver", - "r y", - "▁l o", - "ag e", - "a ge", - "▁p r", - "▁s p", - "u e", - "▁c o", - "▁ co", - "ic k", - "i ck", - "be r", - "b er", - "▁d id", - "i p", - "ac h", - "a ch", - "▁b ack", - "▁n o", - "▁con t", - "▁co nt", - "▁c ont", - "▁o ther", - "▁ other", - "▁ever y", - "▁e very", - "p t", - "▁ne ed", - "▁h im", - "▁ U", - "▁I n", - "▁wor k", - "ir st", - "▁p art", - "▁loo k", - "▁l ook", - "itt le", - "b le", - "i z", - "▁u n", - "▁ un", - "▁m ake", - "ome t", - "om et", - "nd er", - "n der", - "is h", - "n a", - "▁l ittle", - "▁of f", - "▁o ff", - "▁th an", - "▁go t", - "▁g ot", - "ual ly", - "u ally", - "▁pe r", - "▁p er", - "▁ per", - "▁go od", - "▁g ood", - "▁w ay", - "▁cou ld", - "▁c ould", - "▁a c", - "▁ ac", - "▁im p", - "ab le", - "a ble", - "▁wh ere", - "if f", - "i ff", - "▁Th at", - "▁re s", - "▁r es", - "▁ res", - "ou nt", - "p l", - "an ce", - "a nce", - "▁f irst", - "▁r o", - "▁ ro", - "▁pr e", - "▁p re", - "as s", - "▁sa y", - "▁s ay", - "in t", - "i nt", - "ate d", - "at ed", - "ir e", - "i re", - "uc h", - "u ch", - "as e", - "a se", - "▁some t", - "▁s omet", - "ou nd", - "o und", - "▁do wn", - "▁d own", - "▁d iff", - "se l", - "s el", - "▁g u", - "▁a m", - "▁ am", - "res s", - "r ess", - "▁lo t", - "▁l ot", - "en ce", - "e nce", - "▁d is", - "or m", - "i x", - "▁p o", - "v ing", - "ent y", - "en ty", - "▁ K", - "▁sp e", - "▁s pe", - "un d", - "u nd", - "h e", - "▁m uch", - "▁a r", - "▁ ar", - "rou nd", - "ro und", - "r ound", - "▁a pp", - "c o", - "ar k", - "▁ne w", - "▁n ew", - "ate r", - "at er", - "a ter", - "ul t", - "en d", - "e nd", - "▁ev en", - "▁e ven", - "▁st art", - "ation s", - "at ions", - "rou gh", - "r ough", - "il e", - "i le", - "ft er", - "f ter", - "▁we ll", - "▁w ell", - "b e", - "▁The y", - "▁th ree", - "ig n", - "il d", - "i ld", - "▁sa id", - "ou gh", - "an g", - "a ng", - "▁to o", - "▁t oo", - "ad e", - "▁b l", - "en s", - "▁in c", - "i a", - "▁th ose", - "▁m o", - "▁t ake", - "▁th rough", - "▁f l", - "▁k ind", - "▁thing s", - "▁th ings", - "▁be t", - "▁b et", - "▁on ly", - "▁S t", - "▁le t", - "▁l et", - "ces s", - "c ess", - "▁C h", - "ar y", - "a ry", - "ve l", - "v el", - "▁I f", - "x t", - "oth er", - "ot her", - "o ther", - "a v", - "ic al", - "or d", - "▁ag ain", - "▁somet hing", - "▁some thing", - "on na", - "f ore", - "▁m ay", - "t ing", - "▁b u", - "▁diff ere", - "ur n", - "▁g onna", - "▁do es", - "uc t", - "u ct", - "o g", - "▁tw enty", - "▁g r", - "▁ gr", - "▁Y e", - "w n", - "▁sh ould", - "▁com m", - "▁c omm", - "it ion", - "▁un der", - "▁u nder", - "▁he l", - "▁h el", - "or y", - "o ry", - "▁f o", - "▁us e", - "▁u se", - "▁ use", - "ig h", - "i gh", - "if e", - "▁act ually", - "▁t al", - "▁ca ll", - "▁c all", - "ent s", - "en ts", - "i ous", - "ul l", - "u ll", - "▁The re", - "▁Th ere", - "▁Ye ah", - "▁mo st", - "▁m ost", - "▁k e", - "▁ ke", - "or s", - "ve d", - "v ed", - "y s", - "▁s c", - "▁ha pp", - "op e", - "o pe", - "▁hel p", - "at ch", - "▁Wh at", - "▁re m", - "▁r em", - "pl e", - "p le", - "▁No w", - "▁N ow", - "▁b r", - "oo l", - "o ol", - "ot h", - "o th", - "▁fo ur", - "▁f our", - "sel f", - "▁st r", - "n e", - "th ing", - "t hing", - "▁p ut", - "ia l", - "i al", - "▁g reat", - "a il", - "u b", - "n ing", - "▁s m", - "▁fe el", - "▁f ive", - "od y", - "und red", - "is s", - "an k", - "ge t", - "g et", - "ak ing", - "a king", - "▁man y", - "▁m any", - "▁h undred", - "▁year s", - "▁ye ars", - "▁be ing", - "▁com e", - "▁c ome", - "▁me an", - "il y", - "i ly", - "▁differe nt", - "▁a fter", - "▁se r", - "▁s er", - "▁sh ow", - "f orm", - "f ul", - "o y", - "▁s ix", - "▁v ide", - "▁ V", - "▁it s", - "▁ its", - "▁po int", - "▁d ay", - "▁ day", - "▁de s", - "▁d es", - "on s", - "▁b it", - "▁be l", - "▁b el", - "▁be fore", - "▁a w", - "▁en d", - "▁e nd", - "▁ end", - "▁O h", - "▁st ill", - "at h", - "a th", - "▁lo ng", - "▁l ong", - "▁ '", - "is e", - "i se", - "o b", - "d ay", - "▁ad d", - "f t", - "ve s", - "v es", - "ce s", - "c es", - "ad y", - "▁c r", - "▁ar ound", - "▁a round", - "▁tr y", - "▁t ry", - "le s", - "l es", - "ver s", - "v ers", - "k ay", - "ia n", - "i an", - "ate s", - "at es", - "▁fin d", - "▁f ind", - "w ard", - "▁A s", - "▁e ight", - "li c", - "l ic", - "▁s ame", - "▁po s", - "▁p os", - "▁e m", - "▁ em", - "▁m ade", - "▁su pp", - "▁l ife", - "▁B e", - "pe ct", - "p ect", - "▁de c", - "▁pl ay", - "ang e", - "an ge", - "▁at t", - "▁per s", - "▁p ers", - "w ays", - "▁h igh", - "▁ha nd", - "▁h and", - "▁ne xt", - "▁con s", - "▁c ons", - "▁o wn", - "▁ own", - "▁in v", - "ow er", - "▁in d", - "▁ ind", - "er t", - "n g", - "av e", - "a ve", - "▁ye ar", - "▁b ig", - "at ing", - "a ting", - "▁wor ld", - "▁re l", - "▁r el", - "▁sur e", - "▁su re", - "▁s ure", - "▁tr a", - "▁t ra", - "e w", - "ere d", - "er ed", - "e red", - "▁f in", - "▁We ll", - "▁W ell", - "▁s l", - "▁do ing", - "b s", - "▁se t", - "▁s et", - "▁re c", - "u al", - "ci al", - "c ial", - "▁p h", - "▁ ph", - "er m", - "▁lo ve", - "p h", - "▁re al", - "▁la st", - "▁l ast", - "ic t", - "i ct", - "▁b o", - "▁r a", - "▁ ra", - "ib le", - "i ble", - "▁w r", - "m er", - "▁cou nt", - "▁c ount", - "it ies", - "▁al ways", - "ine t", - "in et", - "ment s", - "m ents", - "u c", - "▁m ight", - "▁int er", - "▁in ter", - "▁vide o", - "g in", - "▁te ll", - "▁t ell", - "▁ne ver", - "▁n ever", - "ven t", - "ve nt", - "v ent", - "▁imp ort", - "ie d", - "i ed", - "▁s y", - "▁H ow", - "ical ly", - "ic ally", - "ough t", - "ou ght", - "▁th ir", - "▁re p", - "▁r ep", - "k s", - "i b", - "▁f am", - "j ect", - "▁b as", - "▁S he", - "▁g ive", - "ake s", - "ak es", - "▁n inet", - "▁re g", - "▁m in", - "▁o p", - "▁ op", - "▁de f", - "▁did n", - "t e", - "▁cou r", - "▁co ur", - "▁c our", - "▁wh y", - "▁en t", - "▁e nt", - "▁ ent", - "▁pl ace", - "▁in s", - "▁ca r", - "▁c ar", - "ath er", - "at her", - "a ther", - "▁pers on", - "ul ar", - "▁ins t", - "▁in st", - "▁pro d", - "▁pr od", - "le ct", - "l ect", - "▁A l", - "▁to day", - "▁be c", - "▁su r", - "▁s ur", - "▁Al l", - "▁A ll", - "▁an other", - "▁bu s", - "▁b us", - "▁ke ep", - "el l", - "e ll", - "es e", - "e se", - "ri end", - "▁qu est", - "▁tal k", - "al s", - "ing s", - "▁mo n", - "▁m on", - "co nd", - "c ond", - "ol d", - "o ld", - "▁ac c", - "▁l a", - "▁n um", - "ide nt", - "id ent", - "▁ch e", - "▁c he", - "in ess", - "i ness", - "▁t urn", - "▁e ar", - "▁N o", - "ous and", - "▁bet ter", - "if ic", - "▁lo o", - "▁l oo", - "▁g l", - "o c", - "▁import ant", - "ite d", - "it ed", - "▁A n", - "▁th ousand", - "il ity", - "ll ow", - "▁use d", - "▁us ed", - "▁g en", - "▁s im", - "l i", - "▁happ en", - "▁U n", - "▁L et", - "a ir", - "oc k", - "o ck", - "ab ly", - "g g", - "▁w atch", - "▁F or", - "▁s w", - "re n", - "r en", - "ut e", - "u te", - "e ver", - "▁po l", - "▁p ol", - "▁sc h", - "▁s ch", - "▁Wh en", - "▁su ch", - "▁s uch", - "▁f if", - "▁h ome", - "▁cl e", - "▁c le", - "▁cont in", - "ous e", - "ou se", - "o use", - "▁f riend", - "ur ing", - "▁O kay", - "g r", - "▁ab le", - "▁a ble", - "▁ able", - "▁st ud", - "▁e ff", - "h ip", - "b ody", - "▁to p", - "▁t op", - "n ess", - "▁exp er", - "▁ex per", - "▁pre t", - "▁pr et", - "▁bo th", - "▁b oth", - "▁don e", - "▁do ne", - "▁d one", - "c ri", - "▁m ark", - "▁wh ile", - "▁o ld", - "▁ old", - "ro s", - "r os", - "on t", - "o nt", - "▁se cond", - "at ive", - "▁th ought", - "▁be st", - "▁b est", - "▁fo und", - "▁f ound", - "ie w", - "i ew", - "▁bel ie", - "▁e ach", - "ere st", - "er est", - "▁tr i", - "▁t ri", - "▁e as", - "▁c a", - "▁ ca", - "▁f act", - "▁car e", - "▁ca re", - "▁c are", - "▁f un", - "at ter", - "ure s", - "ur es", - "u res", - "▁he ad", - "▁le ar", - "▁w ater", - "▁h ard", - "▁fe w", - "▁f ew", - "▁s ide", - "w een", - "▁ex p", - "▁aw ay", - "it s", - "i ts", - "▁ex t", - "▁e xt", - "l ud", - "▁r un", - "▁tr ans", - "in ce", - "i nce", - "▁s k", - "▁op en", - "c us", - "▁bet ween", - "▁call ed", - "▁we e", - "▁w ee", - "▁pret ty", - "as on", - "▁f ar", - "em ber", - "om m", - "▁inter est", - "▁int erest", - "an y", - "ne r", - "n er", - "u ff", - "▁pre s", - "▁pr es", - "▁p res", - "▁c ur", - "▁ch ild", - "e e", - "▁to get", - "▁toget her", - "ol og", - "▁G od", - "on d", - "o nd", - "▁ch ar", - "▁look ing", - "▁loo king", - "st em", - "a z", - "ce nt", - "c ent", - "▁o b", - "▁ ob", - "▁as s", - "▁ ass", - "l and", - "▁does n", - "▁bus iness", - "▁cour se", - "▁te n", - "▁t en", - "▁ ten", - "p s", - "ar ch", - "ce d", - "c ed", - "m s", - "iz e", - "n ce", - "▁re f", - "▁n ame", - "ros s", - "▁gr ow", - "one y", - "▁we nt", - "▁w ent", - "ic s", - "te en", - "t een", - "▁co u", - "▁c ou", - "▁pro b", - "▁pr ob", - "▁re t", - "▁r et", - "▁gu ys", - "▁c ame", - "as h", - "le d", - "l ed", - "▁E ur", - "ue s", - "u es", - "▁ ide", - "g an", - "▁every thing", - "▁get ting", - "▁as k", - "▁co r", - "▁c or", - "▁bu ild", - "▁s ign", - "▁sm all", - "uc k", - "u ck", - "▁e l", - "▁ el", - "▁co l", - "▁c ol", - "▁I s", - "ation al", - "st and", - "c y", - "▁con f", - "d er", - "▁br e", - "▁b re", - "▁ca p", - "▁c ap", - "▁mo d", - "▁m od", - "et s", - "e ts", - "i ke", - "▁num ber", - "▁comp le", - "▁com ple", - "ert ain", - "▁ev er", - "▁e ver", - "▁ ever", - "▁col l", - "▁co ll", - "▁h um", - "▁Eur ope", - "▁cr e", - "▁c re", - "▁me t", - "▁m et", - "▁ex am", - "▁mo ve", - "▁p ass", - "▁le ft", - "▁sy stem", - "▁inc lud", - "▁Th ank", - "ce pt", - "▁w om", - "▁prod uct", - "te n", - "t en", - "▁res t", - "▁re st", - "▁r est", - "▁prob ably", - "▁dr i", - "▁d ri", - "▁D o", - "▁gen er", - "▁any thing", - "▁la r", - "▁l ar", - "▁M y", - "▁sch ool", - "▁le ad", - "▁su b", - "▁s ub", - "▁t y", - "▁ ty", - "▁pl an", - "▁see m", - "▁se em", - "▁who le", - "ire ct", - "ir ect", - "▁li ght", - "▁l ight", - "▁m ust", - "▁mo m", - "▁m om", - "▁op p", - "▁o pp", - "▁supp ort", - "▁fam ily", - "ice s", - "ic es", - "i ces", - "am p", - "▁prob le", - "▁pro ble", - "▁d r", - "re ady", - "▁us ing", - "ens e", - "en se", - "▁pro v", - "us h", - "a x", - "▁p ower", - "▁R e", - "al th", - "▁e v", - "▁st and", - "▁ stand", - "▁w ar", - "t s" - ] - } -} \ No newline at end of file diff --git a/parakeet-ctc-110m-coreml/tokenizer_config.json b/parakeet-ctc-110m-coreml/tokenizer_config.json deleted file mode 100644 index d9eaa0fb393ab6e3ac6890fe99efb6191cded97c..0000000000000000000000000000000000000000 --- a/parakeet-ctc-110m-coreml/tokenizer_config.json +++ /dev/null @@ -1,27 +0,0 @@ -{ - "added_tokens_decoder": { - "0": { - "content": "", - "lstrip": false, - "normalized": false, - "rstrip": false, - "single_word": false, - "special": true - }, - "1024": { - "content": "", - "lstrip": false, - "normalized": false, - "rstrip": false, - "single_word": false, - "special": true - } - }, - "clean_up_tokenization_spaces": false, - "extra_special_tokens": {}, - "model_max_length": 1000000000000000019884624838656, - "pad_token": "", - "processor_class": "ParakeetProcessor", - "tokenizer_class": "PreTrainedTokenizer", - "unk_token": "" -} diff --git a/parakeet-ctc-110m-coreml/vocab.json b/parakeet-ctc-110m-coreml/vocab.json deleted file mode 100644 index 33b95831722563db5362641673b37885e25fb47a..0000000000000000000000000000000000000000 --- a/parakeet-ctc-110m-coreml/vocab.json +++ /dev/null @@ -1 +0,0 @@ -{"0": "", "1": "▁t", "2": "▁th", "3": "▁a", "4": "in", "5": "re", "6": "▁the", "7": "▁w", "8": "▁s", "9": "▁o", "10": "er", "11": "ou", "12": "at", "13": "nd", "14": "it", "15": "▁h", "16": "▁c", "17": "▁b", "18": "is", "19": "en", "20": "on", "21": "ing", "22": "▁f", "23": "▁to", "24": "▁m", "25": "es", "26": "▁p", "27": "or", "28": "an", "29": "▁d", "30": "ll", "31": "▁I", "32": "ed", "33": "▁and", "34": "▁l", "35": "▁of", "36": "▁in", "37": "▁y", "38": "ar", "39": "▁g", "40": "▁you", "41": "as", "42": "om", "43": "▁n", "44": "ve", "45": "▁that", "46": "le", "47": "ic", "48": "us", "49": "ow", "50": "et", "51": "al", "52": "▁e", "53": "ut", "54": "▁it", "55": "ot", "56": "▁be", "57": "▁T", "58": "ion", "59": "▁is", "60": "▁wh", "61": "▁re", "62": "▁on", "63": "▁we", "64": "ent", "65": "▁A", "66": "ay", "67": "▁ha", "68": "▁Th", "69": "id", "70": "▁S", "71": "ac", "72": "gh", "73": "ver", "74": "ke", "75": "▁for", "76": "im", "77": "ly", "78": "ur", "79": "ld", "80": "▁he", "81": "▁st", "82": "all", "83": "ro", "84": "st", "85": "se", "86": "ct", "87": "ith", "88": "ir", "89": "am", "90": "▁this", "91": "if", "92": "▁W", "93": "oo", "94": "ri", "95": "▁was", "96": "ght", "97": "▁u", "98": "▁with", "99": "ad", "100": "ch", "101": "▁se", "102": "▁k", "103": "▁an", "104": "▁The", "105": "▁li", "106": "▁do", "107": "▁B", "108": "▁have", "109": "▁as", "110": "th", "111": "▁are", "112": "▁sh", "113": "ust", "114": "ce", "115": "ally", "116": "ill", "117": "▁H", "118": "▁j", "119": "ter", "120": "▁go", "121": "▁And", "122": "ation", "123": "▁C", "124": "▁so", "125": "ome", "126": "▁not", "127": "op", "128": "il", "129": "ore", "130": "▁ne", "131": "▁can", "132": "▁me", "133": "▁at", "134": "ould", "135": "ant", "136": "▁M", "137": "▁like", "138": "ere", "139": "▁they", "140": "ra", "141": "ers", "142": "▁ab", "143": "▁de", "144": "▁kn", "145": "ge", "146": "▁Y", "147": "▁ch", "148": "ul", "149": "pp", "150": "▁or", "151": "▁al", "152": "▁con", "153": "▁com", "154": "ess", "155": "▁su", "156": "out", "157": "▁your", "158": "▁So", "159": "ate", "160": "▁one", "161": "▁all", "162": "▁ex", "163": "est", "164": "▁fr", "165": "▁just", "166": "▁pro", "167": "▁know", "168": "▁O", "169": "ain", "170": "▁but", "171": "ol", "172": "ive", "173": "▁v", "174": "use", "175": "very", "176": "art", "177": "qu", "178": "▁my", "179": "el", "180": "▁N", "181": "nt", "182": "▁It", "183": "▁what", "184": "ab", "185": "▁P", "186": "▁wor", "187": "▁out", "188": "▁there", "189": "▁up", "190": "um", "191": "▁from", "192": "pe", "193": "▁tw", "194": "▁r", "195": "and", "196": "ight", "197": "ort", "198": "un", "199": "▁L", "200": "ist", "201": "▁about", "202": "ide", "203": "ig", "204": "ake", "205": "▁D", "206": "em", "207": "os", "208": "king", "209": "rou", "210": "ind", "211": "our", "212": "res", "213": "▁We", "214": "▁get", "215": "▁E", "216": "▁G", "217": "ack", "218": "▁le", "219": "ity", "220": "od", "221": "▁F", "222": "ard", "223": "▁pl", "224": "▁our", "225": "▁int", "226": "ment", "227": "▁will", "228": "ies", "229": "▁by", "230": "ink", "231": "ca", "232": "▁if", "233": "red", "234": "her", "235": "ie", "236": "▁us", "237": "▁some", "238": "▁don", "239": "ven", "240": "ood", "241": "ast", "242": "▁R", "243": "▁his", "244": "▁tim", "245": "▁tr", "246": "▁more", "247": "ich", "248": "ous", "249": "ame", "250": "▁going", "251": "▁had", "252": "▁them", "253": "ook", "254": "▁pe", "255": "▁Wh", "256": "▁You", "257": "▁But", "258": "ine", "259": "▁here", "260": "▁would", "261": "cause", "262": "right", "263": "so", "264": "ost", "265": "ure", "266": "▁has", "267": "ect", "268": "▁think", "269": "▁fe", "270": "ong", "271": "▁see", "272": "▁when", "273": "▁who", "274": "▁were", "275": "▁really", "276": "▁their", "277": "▁want", "278": "one", "279": "ople", "280": "▁then", "281": "▁time", "282": "▁sa", "283": "ap", "284": "▁te", "285": "▁He", "286": "▁ye", "287": "ck", "288": "▁her", "289": "▁thing", "290": "▁right", "291": "▁which", "292": "itt", "293": "ice", "294": "act", "295": "▁people", "296": "ty", "297": "▁two", "298": "▁J", "299": "▁im", "300": "ther", "301": "ci", "302": "ose", "303": "▁cl", "304": "▁qu", "305": "▁man", "306": "▁also", "307": "ree", "308": "▁en", "309": "ud", "310": "▁how", "311": "reat", "312": "ak", "313": "hing", "314": "ag", "315": "▁any", "316": "ff", "317": "ace", "318": "per", "319": "▁because", "320": "▁very", "321": "own", "322": "▁ad", "323": "▁act", "324": "▁been", "325": "▁now", "326": "▁ag", "327": "▁into", "328": "▁comp", "329": "ars", "330": "ions", "331": "are", "332": "ite", "333": "iv", "334": "▁these", "335": "ays", "336": "ep", "337": "▁This", "338": "▁she", "339": "ans", "340": "ah", "341": "een", "342": "▁over", "343": "ry", "344": "▁lo", "345": "age", "346": "▁pr", "347": "▁sp", "348": "ue", "349": "▁co", "350": "ick", "351": "ber", "352": "▁did", "353": "ip", "354": "ach", "355": "▁back", "356": "▁no", "357": "▁cont", "358": "▁other", "359": "▁every", "360": "pt", "361": "▁need", "362": "▁him", "363": "▁U", "364": "▁In", "365": "▁work", "366": "irst", "367": "▁part", "368": "▁look", "369": "ittle", "370": "ble", "371": "iz", "372": "▁un", "373": "▁make", "374": "omet", "375": "nder", "376": "ish", "377": "na", "378": "▁little", "379": "▁off", "380": "▁than", "381": "▁got", "382": "ually", "383": "▁per", "384": "▁good", "385": "▁way", "386": "▁could", "387": "▁ac", "388": "▁imp", "389": "able", "390": "▁where", "391": "iff", "392": "▁That", "393": "▁res", "394": "ount", "395": "pl", "396": "ance", "397": "▁first", "398": "▁ro", "399": "▁pre", "400": "ass", "401": "▁say", "402": "int", "403": "ated", "404": "ire", "405": "uch", "406": "ase", "407": "▁somet", "408": "ound", "409": "▁down", "410": "▁diff", "411": "sel", "412": "▁gu", "413": "▁am", "414": "ress", "415": "▁lot", "416": "ence", "417": "▁dis", "418": "orm", "419": "ix", "420": "▁po", "421": "ving", "422": "enty", "423": "▁K", "424": "▁spe", "425": "und", "426": "he", "427": "▁much", "428": "▁ar", "429": "round", "430": "▁app", "431": "co", "432": "ark", "433": "▁new", "434": "ater", "435": "ult", "436": "end", "437": "▁even", "438": "▁start", "439": "ations", "440": "rough", "441": "ile", "442": "fter", "443": "▁well", "444": "be", "445": "▁They", "446": "▁three", "447": "ign", "448": "ild", "449": "▁said", "450": "ough", "451": "ang", "452": "▁too", "453": "ade", "454": "▁bl", "455": "ens", "456": "▁inc", "457": "ia", "458": "▁those", "459": "▁mo", "460": "▁take", "461": "▁through", "462": "▁fl", "463": "▁kind", "464": "▁things", "465": "▁bet", "466": "▁only", "467": "▁St", "468": "▁let", "469": "cess", "470": "▁Ch", "471": "ary", "472": "vel", "473": "▁If", "474": "xt", "475": "other", "476": "av", "477": "ical", "478": "ord", "479": "▁again", "480": "▁something", "481": "onna", "482": "fore", "483": "▁may", "484": "ting", "485": "▁bu", "486": "▁differe", "487": "urn", "488": "▁gonna", "489": "▁does", "490": "uct", "491": "og", "492": "▁twenty", "493": "▁gr", "494": "▁Ye", "495": "wn", "496": "▁should", "497": "▁comm", "498": "ition", "499": "▁under", "500": "▁hel", "501": "ory", "502": "▁fo", "503": "▁use", "504": "igh", "505": "ife", "506": "▁actually", "507": "▁tal", "508": "▁call", "509": "ents", "510": "ious", "511": "ull", "512": "▁There", "513": "▁Yeah", "514": "▁most", "515": "▁ke", "516": "ors", "517": "ved", "518": "ys", "519": "▁sc", "520": "▁happ", "521": "ope", "522": "▁help", "523": "atch", "524": "▁What", "525": "▁rem", "526": "ple", "527": "▁Now", "528": "▁br", "529": "ool", "530": "oth", "531": "▁four", "532": "self", "533": "▁str", "534": "ne", "535": "thing", "536": "▁put", "537": "ial", "538": "▁great", "539": "ail", "540": "ub", "541": "ning", "542": "▁sm", "543": "▁feel", "544": "▁five", "545": "ody", "546": "undred", "547": "iss", "548": "ank", "549": "get", "550": "aking", "551": "▁many", "552": "▁hundred", "553": "▁years", "554": "▁being", "555": "▁come", "556": "▁mean", "557": "ily", "558": "▁different", "559": "▁after", "560": "▁ser", "561": "▁show", "562": "form", "563": "ful", "564": "oy", "565": "▁six", "566": "▁vide", "567": "▁V", "568": "▁its", "569": "▁point", "570": "▁day", "571": "▁des", "572": "ons", "573": "▁bit", "574": "▁bel", "575": "▁before", "576": "▁aw", "577": "▁end", "578": "▁Oh", "579": "▁still", "580": "ath", "581": "▁long", "582": "▁'", "583": "ise", "584": "ob", "585": "day", "586": "▁add", "587": "ft", "588": "ves", "589": "ces", "590": "ady", "591": "▁cr", "592": "▁around", "593": "▁try", "594": "les", "595": "vers", "596": "kay", "597": "ian", "598": "ates", "599": "▁find", "600": "ward", "601": "▁As", "602": "▁eight", "603": "lic", "604": "▁same", "605": "▁pos", "606": "▁em", "607": "▁made", "608": "▁supp", "609": "▁life", "610": "▁Be", "611": "pect", "612": "▁dec", "613": "▁play", "614": "ange", "615": "▁att", "616": "▁pers", "617": "ways", "618": "▁high", "619": "▁hand", "620": "▁next", "621": "▁cons", "622": "▁own", "623": "▁inv", "624": "ower", "625": "▁ind", "626": "ert", "627": "ng", "628": "ave", "629": "▁year", "630": "▁big", "631": "ating", "632": "▁world", "633": "▁rel", "634": "▁sure", "635": "▁tra", "636": "ew", "637": "ered", "638": "▁fin", "639": "▁Well", "640": "▁sl", "641": "▁doing", "642": "bs", "643": "▁set", "644": "▁rec", "645": "ual", "646": "cial", "647": "▁ph", "648": "erm", "649": "▁love", "650": "ph", "651": "▁real", "652": "▁last", "653": "ict", "654": "▁bo", "655": "▁ra", "656": "ible", "657": "▁wr", "658": "mer", "659": "▁count", "660": "ities", "661": "▁always", "662": "inet", "663": "ments", "664": "uc", "665": "▁might", "666": "▁inter", "667": "▁video", "668": "gin", "669": "▁tell", "670": "▁never", "671": "vent", "672": "▁import", "673": "ied", "674": "▁sy", "675": "▁How", "676": "ically", "677": "ought", "678": "▁thir", "679": "▁rep", "680": "ks", "681": "ib", "682": "▁fam", "683": "ject", "684": "▁bas", "685": "▁She", "686": "▁give", "687": "akes", "688": "▁ninet", "689": "▁reg", "690": "▁min", "691": "▁op", "692": "▁def", "693": "▁didn", "694": "te", "695": "▁cour", "696": "▁why", "697": "▁ent", "698": "▁place", "699": "▁ins", "700": "▁car", "701": "ather", "702": "▁person", "703": "ular", "704": "▁inst", "705": "▁prod", "706": "lect", "707": "▁Al", "708": "▁today", "709": "▁bec", "710": "▁sur", "711": "▁All", "712": "▁another", "713": "▁bus", "714": "▁keep", "715": "ell", "716": "ese", "717": "riend", "718": "▁quest", "719": "▁talk", "720": "als", "721": "ings", "722": "▁mon", "723": "cond", "724": "old", "725": "▁acc", "726": "▁la", "727": "▁num", "728": "ident", "729": "▁che", "730": "iness", "731": "▁turn", "732": "▁ear", "733": "▁No", "734": "ousand", "735": "▁better", "736": "ific", "737": "▁loo", "738": "▁gl", "739": "oc", "740": "▁important", "741": "ited", "742": "▁An", "743": "▁thousand", "744": "ility", "745": "llow", "746": "▁used", "747": "▁gen", "748": "▁sim", "749": "li", "750": "▁happen", "751": "▁Un", "752": "▁Let", "753": "air", "754": "ock", "755": "ably", "756": "gg", "757": "▁watch", "758": "▁For", "759": "▁sw", "760": "ren", "761": "ute", "762": "ever", "763": "▁pol", "764": "▁sch", "765": "▁When", "766": "▁such", "767": "▁fif", "768": "▁home", "769": "▁cle", "770": "▁contin", "771": "ouse", "772": "▁friend", "773": "uring", "774": "▁Okay", "775": "gr", "776": "▁able", "777": "▁stud", "778": "▁eff", "779": "hip", "780": "body", "781": "▁top", "782": "ness", "783": "▁exper", "784": "▁pret", "785": "▁both", "786": "▁done", "787": "cri", "788": "▁mark", "789": "▁while", "790": "▁old", "791": "ros", "792": "ont", "793": "▁second", "794": "ative", "795": "▁thought", "796": "▁best", "797": "▁found", "798": "iew", "799": "▁belie", "800": "▁each", "801": "erest", "802": "▁tri", "803": "▁eas", "804": "▁ca", "805": "▁fact", "806": "▁care", "807": "▁fun", "808": "atter", "809": "ures", "810": "▁head", "811": "▁lear", "812": "▁water", "813": "▁hard", "814": "▁few", "815": "▁side", "816": "ween", "817": "▁exp", "818": "▁away", "819": "its", "820": "▁ext", "821": "lud", "822": "▁run", "823": "▁trans", "824": "ince", "825": "▁sk", "826": "▁open", "827": "cus", "828": "▁between", "829": "▁called", "830": "▁wee", "831": "▁pretty", "832": "ason", "833": "▁far", "834": "ember", "835": "omm", "836": "▁interest", "837": "any", "838": "ner", "839": "uff", "840": "▁pres", "841": "▁cur", "842": "▁child", "843": "ee", "844": "▁toget", "845": "▁together", "846": "olog", "847": "▁God", "848": "ond", "849": "▁char", "850": "▁looking", "851": "stem", "852": "az", "853": "cent", "854": "▁ob", "855": "▁ass", "856": "land", "857": "▁doesn", "858": "▁business", "859": "▁course", "860": "▁ten", "861": "ps", "862": "arch", "863": "ced", "864": "ms", "865": "ize", "866": "nce", "867": "▁ref", "868": "▁name", "869": "ross", "870": "▁grow", "871": "oney", "872": "▁went", "873": "ics", "874": "teen", "875": "▁cou", "876": "▁prob", "877": "▁ret", "878": "▁guys", "879": "▁came", "880": "ash", "881": "led", "882": "▁Eur", "883": "ues", "884": "▁ide", "885": "gan", "886": "▁everything", "887": "▁getting", "888": "▁ask", "889": "▁cor", "890": "▁build", "891": "▁sign", "892": "▁small", "893": "uck", "894": "▁el", "895": "▁col", "896": "▁Is", "897": "ational", "898": "stand", "899": "cy", "900": "▁conf", "901": "der", "902": "▁bre", "903": "▁cap", "904": "▁mod", "905": "ets", "906": "ike", "907": "▁number", "908": "▁comple", "909": "ertain", "910": "▁ever", "911": "▁coll", "912": "▁hum", "913": "▁Europe", "914": "▁cre", "915": "▁met", "916": "▁exam", "917": "▁move", "918": "▁pass", "919": "▁left", "920": "▁system", "921": "▁includ", "922": "▁Thank", "923": "cept", "924": "▁wom", "925": "▁product", "926": "ten", "927": "▁rest", "928": "▁probably", "929": "▁dri", "930": "▁Do", "931": "▁gener", "932": "▁anything", "933": "▁lar", "934": "▁My", "935": "▁school", "936": "▁lead", "937": "▁sub", "938": "▁ty", "939": "▁plan", "940": "▁seem", "941": "▁whole", "942": "irect", "943": "▁light", "944": "▁must", "945": "▁mom", "946": "▁opp", "947": "▁support", "948": "▁family", "949": "ices", "950": "amp", "951": "▁proble", "952": "▁dr", "953": "ready", "954": "▁using", "955": "ense", "956": "▁prov", "957": "ush", "958": "ax", "959": "▁power", "960": "▁Re", "961": "alth", "962": "▁ev", "963": "▁stand", "964": "▁war", "965": "ts", "966": "▁", "967": "e", "968": "t", "969": "o", "970": "a", "971": "n", "972": "i", "973": "s", "974": "r", "975": "h", "976": "l", "977": "d", "978": "u", "979": "c", "980": "m", "981": "y", "982": "g", "983": "w", "984": "f", "985": "p", "986": ".", "987": "b", "988": ",", "989": "v", "990": "k", "991": "'", "992": "I", "993": "T", "994": "A", "995": "S", "996": "x", "997": "W", "998": "j", "999": "B", "1000": "C", "1001": "H", "1002": "?", "1003": "M", "1004": "O", "1005": "Y", "1006": "N", "1007": "P", "1008": "E", "1009": "q", "1010": "L", "1011": "D", "1012": "z", "1013": "G", "1014": "F", "1015": "R", "1016": "!", "1017": "J", "1018": "U", "1019": "K", "1020": "V", "1021": "Q", "1022": "Z", "1023": "X"} \ No newline at end of file diff --git a/parakeet-tdt-0.6b-v2-coreml/.DS_Store b/parakeet-tdt-0.6b-v2-coreml/.DS_Store deleted file mode 100644 index ad6e41b077109ad8b0188aa879c7d5c2f9afdae2..0000000000000000000000000000000000000000 Binary files a/parakeet-tdt-0.6b-v2-coreml/.DS_Store and /dev/null differ diff --git a/parakeet-tdt-0.6b-v2-coreml/Decoder.mlmodelc/analytics/coremldata.bin b/parakeet-tdt-0.6b-v2-coreml/Decoder.mlmodelc/analytics/coremldata.bin deleted file mode 100644 index 7e6fa4eb79131e62089f252a08b860e723f5114f..0000000000000000000000000000000000000000 --- a/parakeet-tdt-0.6b-v2-coreml/Decoder.mlmodelc/analytics/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:46de1a6fe2e49d19a2125bc91acf020df7f2aea84ba821532aade8427a440b05 -size 243 diff --git a/parakeet-tdt-0.6b-v2-coreml/Decoder.mlmodelc/coremldata.bin b/parakeet-tdt-0.6b-v2-coreml/Decoder.mlmodelc/coremldata.bin deleted file mode 100644 index 79b160cb7e3eb3213fb22cde5fbf6ac7f324c91b..0000000000000000000000000000000000000000 --- a/parakeet-tdt-0.6b-v2-coreml/Decoder.mlmodelc/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:d200ca07694a347f6d02a3886a062ae839831e094e443222f2e48a14945966a8 -size 554 diff --git a/parakeet-tdt-0.6b-v2-coreml/Decoder.mlmodelc/metadata.json b/parakeet-tdt-0.6b-v2-coreml/Decoder.mlmodelc/metadata.json deleted file mode 100644 index e3a08d21a5d1577db2bedd1749ea0ddae5d21b73..0000000000000000000000000000000000000000 --- a/parakeet-tdt-0.6b-v2-coreml/Decoder.mlmodelc/metadata.json +++ /dev/null @@ -1,123 +0,0 @@ -[ - { - "metadataOutputVersion" : "3.0", - "shortDescription" : "Parakeet decoder (RNNT prediction network)", - "outputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 640 × 1)", - "shortDescription" : "", - "shape" : "[1, 640, 1]", - "name" : "decoder", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 2 × 1 × 640)", - "shortDescription" : "", - "shape" : "[2, 1, 640]", - "name" : "h_out", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 2 × 1 × 640)", - "shortDescription" : "", - "shape" : "[2, 1, 640]", - "name" : "c_out", - "type" : "MultiArray" - } - ], - "storagePrecision" : "Float16", - "modelParameters" : [ - - ], - "author" : "Fluid Inference", - "specificationVersion" : 8, - "mlProgramOperationTypeHistogram" : { - "Select" : 1, - "Ios17.squeeze" : 4, - "Ios17.gather" : 1, - "Ios17.cast" : 8, - "Ios17.lstm" : 2, - "Split" : 2, - "Ios17.add" : 1, - "Ios17.transpose" : 2, - "Ios17.greaterEqual" : 1, - "Identity" : 1, - "Stack" : 2 - }, - "computePrecision" : "Mixed (Float16, Float32, Int16, Int32)", - "isUpdatable" : "0", - "stateSchema" : [ - - ], - "availability" : { - "macOS" : "14.0", - "tvOS" : "17.0", - "visionOS" : "1.0", - "watchOS" : "10.0", - "iOS" : "17.0", - "macCatalyst" : "17.0" - }, - "modelType" : { - "name" : "MLModelType_mlProgram" - }, - "inputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1 × 1)", - "shortDescription" : "", - "shape" : "[1, 1]", - "name" : "targets", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1)", - "shortDescription" : "", - "shape" : "[1]", - "name" : "target_length", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 2 × 1 × 640)", - "shortDescription" : "", - "shape" : "[2, 1, 640]", - "name" : "h_in", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 2 × 1 × 640)", - "shortDescription" : "", - "shape" : "[2, 1, 640]", - "name" : "c_in", - "type" : "MultiArray" - } - ], - "userDefinedMetadata" : { - "com.github.apple.coremltools.conversion_date" : "2025-09-25", - "com.github.apple.coremltools.source" : "torch==2.7.0", - "com.github.apple.coremltools.version" : "9.0b1", - "com.github.apple.coremltools.source_dialect" : "TorchScript" - }, - "generatedClassName" : "parakeet_decoder", - "method" : "predict" - } -] \ No newline at end of file diff --git a/parakeet-tdt-0.6b-v2-coreml/Decoder.mlmodelc/model.mil b/parakeet-tdt-0.6b-v2-coreml/Decoder.mlmodelc/model.mil deleted file mode 100644 index 167cc9db34dd7828e0e61eab40d6a09f8a234d30..0000000000000000000000000000000000000000 --- a/parakeet-tdt-0.6b-v2-coreml/Decoder.mlmodelc/model.mil +++ /dev/null @@ -1,73 +0,0 @@ -program(1.0) -[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3500.14.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.7.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "9.0b1"}})] -{ - func main(tensor c_in, tensor h_in, tensor target_length, tensor targets) { - tensor y_batch_dims_0 = const()[name = tensor("y_batch_dims_0"), val = tensor(0)]; - tensor y_validate_indices_0 = const()[name = tensor("y_validate_indices_0"), val = tensor(false)]; - tensor module_prediction_embed_weight_to_fp16 = const()[name = tensor("module_prediction_embed_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; - tensor targets_to_int16_dtype_0 = const()[name = tensor("targets_to_int16_dtype_0"), val = tensor("int16")]; - tensor cast_1_dtype_0 = const()[name = tensor("cast_1_dtype_0"), val = tensor("int32")]; - tensor greater_equal_0_y_0 = const()[name = tensor("greater_equal_0_y_0"), val = tensor(0)]; - tensor targets_to_int16 = cast(dtype = targets_to_int16_dtype_0, x = targets)[name = tensor("cast_9")]; - tensor cast_1 = cast(dtype = cast_1_dtype_0, x = targets_to_int16)[name = tensor("cast_8")]; - tensor greater_equal_0 = greater_equal(x = cast_1, y = greater_equal_0_y_0)[name = tensor("greater_equal_0")]; - tensor slice_by_index_0 = const()[name = tensor("slice_by_index_0"), val = tensor(1025)]; - tensor add_2 = add(x = cast_1, y = slice_by_index_0)[name = tensor("add_2")]; - tensor select_0 = select(a = cast_1, b = add_2, cond = greater_equal_0)[name = tensor("select_0")]; - tensor y_cast_fp16_cast_uint16_axis_0 = const()[name = tensor("y_cast_fp16_cast_uint16_axis_0"), val = tensor(0)]; - tensor select_0_to_int16_dtype_0 = const()[name = tensor("select_0_to_int16_dtype_0"), val = tensor("int16")]; - tensor select_0_to_int16 = cast(dtype = select_0_to_int16_dtype_0, x = select_0)[name = tensor("cast_7")]; - tensor y_cast_fp16_cast_uint16_cast_uint16 = gather(axis = y_cast_fp16_cast_uint16_axis_0, batch_dims = y_batch_dims_0, indices = select_0_to_int16, validate_indices = y_validate_indices_0, x = module_prediction_embed_weight_to_fp16)[name = tensor("y_cast_fp16_cast_uint16_cast_uint16")]; - tensor input_3_perm_0 = const()[name = tensor("input_3_perm_0"), val = tensor([1, 0, 2])]; - tensor split_0_num_splits_0 = const()[name = tensor("split_0_num_splits_0"), val = tensor(2)]; - tensor split_0_axis_0 = const()[name = tensor("split_0_axis_0"), val = tensor(0)]; - tensor h_in_to_fp16_dtype_0 = const()[name = tensor("h_in_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor h_in_to_fp16 = cast(dtype = h_in_to_fp16_dtype_0, x = h_in)[name = tensor("cast_6")]; - tensor split_0_cast_fp16_0, tensor split_0_cast_fp16_1 = split(axis = split_0_axis_0, num_splits = split_0_num_splits_0, x = h_in_to_fp16)[name = tensor("split_0_cast_fp16")]; - tensor split_1_num_splits_0 = const()[name = tensor("split_1_num_splits_0"), val = tensor(2)]; - tensor split_1_axis_0 = const()[name = tensor("split_1_axis_0"), val = tensor(0)]; - tensor c_in_to_fp16_dtype_0 = const()[name = tensor("c_in_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor c_in_to_fp16 = cast(dtype = c_in_to_fp16_dtype_0, x = c_in)[name = tensor("cast_5")]; - tensor split_1_cast_fp16_0, tensor split_1_cast_fp16_1 = split(axis = split_1_axis_0, num_splits = split_1_num_splits_0, x = c_in_to_fp16)[name = tensor("split_1_cast_fp16")]; - tensor input_lstm_layer_0_lstm_h0_squeeze_axes_0 = const()[name = tensor("input_lstm_layer_0_lstm_h0_squeeze_axes_0"), val = tensor([0])]; - tensor input_lstm_layer_0_lstm_h0_squeeze_cast_fp16 = squeeze(axes = input_lstm_layer_0_lstm_h0_squeeze_axes_0, x = split_0_cast_fp16_0)[name = tensor("input_lstm_layer_0_lstm_h0_squeeze_cast_fp16")]; - tensor input_lstm_layer_0_lstm_c0_squeeze_axes_0 = const()[name = tensor("input_lstm_layer_0_lstm_c0_squeeze_axes_0"), val = tensor([0])]; - tensor input_lstm_layer_0_lstm_c0_squeeze_cast_fp16 = squeeze(axes = input_lstm_layer_0_lstm_c0_squeeze_axes_0, x = split_1_cast_fp16_0)[name = tensor("input_lstm_layer_0_lstm_c0_squeeze_cast_fp16")]; - tensor input_lstm_layer_0_direction_0 = const()[name = tensor("input_lstm_layer_0_direction_0"), val = tensor("forward")]; - tensor input_lstm_layer_0_output_sequence_0 = const()[name = tensor("input_lstm_layer_0_output_sequence_0"), val = tensor(true)]; - tensor input_lstm_layer_0_recurrent_activation_0 = const()[name = tensor("input_lstm_layer_0_recurrent_activation_0"), val = tensor("sigmoid")]; - tensor input_lstm_layer_0_cell_activation_0 = const()[name = tensor("input_lstm_layer_0_cell_activation_0"), val = tensor("tanh")]; - tensor input_lstm_layer_0_activation_0 = const()[name = tensor("input_lstm_layer_0_activation_0"), val = tensor("tanh")]; - tensor concat_1_to_fp16 = const()[name = tensor("concat_1_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1312128)))]; - tensor concat_2_to_fp16 = const()[name = tensor("concat_2_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4588992)))]; - tensor concat_0_to_fp16 = const()[name = tensor("concat_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7865856)))]; - tensor input_3_cast_fp16 = transpose(perm = input_3_perm_0, x = y_cast_fp16_cast_uint16_cast_uint16)[name = tensor("transpose_2")]; - tensor input_lstm_layer_0_cast_fp16_0, tensor input_lstm_layer_0_cast_fp16_1, tensor input_lstm_layer_0_cast_fp16_2 = lstm(activation = input_lstm_layer_0_activation_0, bias = concat_0_to_fp16, cell_activation = input_lstm_layer_0_cell_activation_0, direction = input_lstm_layer_0_direction_0, initial_c = input_lstm_layer_0_lstm_c0_squeeze_cast_fp16, initial_h = input_lstm_layer_0_lstm_h0_squeeze_cast_fp16, output_sequence = input_lstm_layer_0_output_sequence_0, recurrent_activation = input_lstm_layer_0_recurrent_activation_0, weight_hh = concat_2_to_fp16, weight_ih = concat_1_to_fp16, x = input_3_cast_fp16)[name = tensor("input_lstm_layer_0_cast_fp16")]; - tensor input_lstm_h0_squeeze_axes_0 = const()[name = tensor("input_lstm_h0_squeeze_axes_0"), val = tensor([0])]; - tensor input_lstm_h0_squeeze_cast_fp16 = squeeze(axes = input_lstm_h0_squeeze_axes_0, x = split_0_cast_fp16_1)[name = tensor("input_lstm_h0_squeeze_cast_fp16")]; - tensor input_lstm_c0_squeeze_axes_0 = const()[name = tensor("input_lstm_c0_squeeze_axes_0"), val = tensor([0])]; - tensor input_lstm_c0_squeeze_cast_fp16 = squeeze(axes = input_lstm_c0_squeeze_axes_0, x = split_1_cast_fp16_1)[name = tensor("input_lstm_c0_squeeze_cast_fp16")]; - tensor input_direction_0 = const()[name = tensor("input_direction_0"), val = tensor("forward")]; - tensor input_output_sequence_0 = const()[name = tensor("input_output_sequence_0"), val = tensor(true)]; - tensor input_recurrent_activation_0 = const()[name = tensor("input_recurrent_activation_0"), val = tensor("sigmoid")]; - tensor input_cell_activation_0 = const()[name = tensor("input_cell_activation_0"), val = tensor("tanh")]; - tensor input_activation_0 = const()[name = tensor("input_activation_0"), val = tensor("tanh")]; - tensor concat_4_to_fp16 = const()[name = tensor("concat_4_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7871040)))]; - tensor concat_5_to_fp16 = const()[name = tensor("concat_5_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11147904)))]; - tensor concat_3_to_fp16 = const()[name = tensor("concat_3_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(14424768)))]; - tensor input_cast_fp16_0, tensor input_cast_fp16_1, tensor input_cast_fp16_2 = lstm(activation = input_activation_0, bias = concat_3_to_fp16, cell_activation = input_cell_activation_0, direction = input_direction_0, initial_c = input_lstm_c0_squeeze_cast_fp16, initial_h = input_lstm_h0_squeeze_cast_fp16, output_sequence = input_output_sequence_0, recurrent_activation = input_recurrent_activation_0, weight_hh = concat_5_to_fp16, weight_ih = concat_4_to_fp16, x = input_lstm_layer_0_cast_fp16_0)[name = tensor("input_cast_fp16")]; - tensor obj_3_axis_0 = const()[name = tensor("obj_3_axis_0"), val = tensor(0)]; - tensor obj_3_cast_fp16 = stack(axis = obj_3_axis_0, values = (input_lstm_layer_0_cast_fp16_1, input_cast_fp16_1))[name = tensor("obj_3_cast_fp16")]; - tensor obj_3_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("obj_3_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; - tensor obj_axis_0 = const()[name = tensor("obj_axis_0"), val = tensor(0)]; - tensor obj_cast_fp16 = stack(axis = obj_axis_0, values = (input_lstm_layer_0_cast_fp16_2, input_cast_fp16_2))[name = tensor("obj_cast_fp16")]; - tensor obj_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("obj_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; - tensor transpose_0_perm_0 = const()[name = tensor("transpose_0_perm_0"), val = tensor([1, 2, 0])]; - tensor transpose_0_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("transpose_0_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; - tensor transpose_0_cast_fp16 = transpose(perm = transpose_0_perm_0, x = input_cast_fp16_0)[name = tensor("transpose_1")]; - tensor decoder = cast(dtype = transpose_0_cast_fp16_to_fp32_dtype_0, x = transpose_0_cast_fp16)[name = tensor("cast_2")]; - tensor c_out = cast(dtype = obj_cast_fp16_to_fp32_dtype_0, x = obj_cast_fp16)[name = tensor("cast_3")]; - tensor h_out = cast(dtype = obj_3_cast_fp16_to_fp32_dtype_0, x = obj_3_cast_fp16)[name = tensor("cast_4")]; - tensor target_length_tmp = identity(x = target_length)[name = tensor("target_length_tmp")]; - } -> (decoder, h_out, c_out); -} \ No newline at end of file diff --git a/parakeet-tdt-0.6b-v2-coreml/Decoder.mlmodelc/weights/weight.bin b/parakeet-tdt-0.6b-v2-coreml/Decoder.mlmodelc/weights/weight.bin deleted file mode 100644 index a895491b78de520abc21d90321ffd90c95483662..0000000000000000000000000000000000000000 --- a/parakeet-tdt-0.6b-v2-coreml/Decoder.mlmodelc/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:27d26890221d82322c1092fd99d7b40578e435d5cf4b83c887c42603caf97aba -size 14429952 diff --git a/parakeet-tdt-0.6b-v2-coreml/Encoder.mlmodelc/analytics/coremldata.bin b/parakeet-tdt-0.6b-v2-coreml/Encoder.mlmodelc/analytics/coremldata.bin deleted file mode 100644 index 4738782a8e36389621ff4b3b90dd2a661b35146c..0000000000000000000000000000000000000000 --- a/parakeet-tdt-0.6b-v2-coreml/Encoder.mlmodelc/analytics/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:42e638870d73f26b332918a3496ce36793fbb413a81cbd3d16ba01328637a105 -size 243 diff --git a/parakeet-tdt-0.6b-v2-coreml/Encoder.mlmodelc/coremldata.bin b/parakeet-tdt-0.6b-v2-coreml/Encoder.mlmodelc/coremldata.bin deleted file mode 100644 index e016ed32c20e1563399f15adad451a442644b3d7..0000000000000000000000000000000000000000 --- a/parakeet-tdt-0.6b-v2-coreml/Encoder.mlmodelc/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:4def7aa848599ad0e17a8b9a982edcdbf33cf92e1f4b798de32e2ca0bc74b030 -size 485 diff --git a/parakeet-tdt-0.6b-v2-coreml/Encoder.mlmodelc/metadata.json b/parakeet-tdt-0.6b-v2-coreml/Encoder.mlmodelc/metadata.json deleted file mode 100644 index c87f9eb24e76b8e518dc1fbead46c8c0029d3925..0000000000000000000000000000000000000000 --- a/parakeet-tdt-0.6b-v2-coreml/Encoder.mlmodelc/metadata.json +++ /dev/null @@ -1,106 +0,0 @@ -[ - { - "metadataOutputVersion" : "3.0", - "shortDescription" : "enc6bit-palettize quantized - encoder", - "outputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 1024 × 188)", - "shortDescription" : "", - "shape" : "[1, 1024, 188]", - "name" : "encoder", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1)", - "shortDescription" : "", - "shape" : "[1]", - "name" : "encoder_length", - "type" : "MultiArray" - } - ], - "storagePrecision" : "Mixed (Float16, Palettized (6 bits))", - "modelParameters" : [ - - ], - "author" : "Fluid Inference", - "specificationVersion" : 8, - "mlProgramOperationTypeHistogram" : { - "Ios17.floor" : 3, - "Ios17.logicalAnd" : 2, - "Ios17.reshape" : 145, - "Ios16.softmax" : 24, - "Ios17.matmul" : 72, - "Ios17.transpose" : 172, - "Split" : 24, - "Ios17.expandDims" : 5, - "Select" : 72, - "Ios17.add" : 174, - "Tile" : 1, - "Ios17.sliceByIndex" : 48, - "Ios16.sigmoid" : 24, - "Pad" : 48, - "Ios17.logicalNot" : 2, - "Ios17.layerNorm" : 120, - "Ios16.silu" : 72, - "Ios17.less" : 1, - "Ios16.constexprLutToDense" : 294, - "Ios17.conv" : 77, - "Ios16.relu" : 3, - "Ios17.linear" : 193, - "Ios17.cast" : 4, - "Ios17.mul" : 99 - }, - "computePrecision" : "Mixed (Float16, Float32, Int32)", - "isUpdatable" : "0", - "stateSchema" : [ - - ], - "availability" : { - "macOS" : "14.0", - "tvOS" : "17.0", - "visionOS" : "1.0", - "watchOS" : "10.0", - "iOS" : "17.0", - "macCatalyst" : "17.0" - }, - "modelType" : { - "name" : "MLModelType_mlProgram" - }, - "inputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 128 × 1501)", - "shortDescription" : "", - "shape" : "[1, 128, 1501]", - "name" : "mel", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1)", - "shortDescription" : "", - "shape" : "[1]", - "name" : "mel_length", - "type" : "MultiArray" - } - ], - "userDefinedMetadata" : { - "com.github.apple.coremltools.conversion_date" : "2025-09-25", - "com.github.apple.coremltools.source" : "torch==2.7.0", - "com.github.apple.coremltools.version" : "9.0b1", - "com.github.apple.coremltools.source_dialect" : "TorchScript" - }, - "generatedClassName" : "parakeet_encoder", - "method" : "predict" - } -] \ No newline at end of file diff --git a/parakeet-tdt-0.6b-v2-coreml/Encoder.mlmodelc/model.mil b/parakeet-tdt-0.6b-v2-coreml/Encoder.mlmodelc/model.mil deleted file mode 100644 index 93d1301ccde7c2ca447af6c01ab05bfda5a5dd2c..0000000000000000000000000000000000000000 --- a/parakeet-tdt-0.6b-v2-coreml/Encoder.mlmodelc/model.mil +++ /dev/null @@ -1,3404 +0,0 @@ -program(1.0) -[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3500.14.1"}, {"coremlc-version", "3500.32.1"}})] -{ - func main(tensor mel, tensor mel_length) { - tensor var_30 = const()[name = tensor("op_30"), val = tensor(-1)]; - tensor x_1_perm_0 = const()[name = tensor("x_1_perm_0"), val = tensor([0, 2, 1])]; - tensor mel_to_fp16_dtype_0 = const()[name = tensor("mel_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor var_86_to_fp16_dtype_0 = const()[name = tensor("op_86_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor var_87_promoted_to_fp16 = const()[name = tensor("op_87_promoted_to_fp16"), val = tensor(-0x1p+0)]; - tensor mel_length_to_fp16 = cast(dtype = var_86_to_fp16_dtype_0, x = mel_length)[name = tensor("cast_3")]; - tensor var_88_cast_fp16 = add(x = mel_length_to_fp16, y = var_87_promoted_to_fp16)[name = tensor("op_88_cast_fp16")]; - tensor _inversed_90_y_0_to_fp16 = const()[name = tensor("_inversed_90_y_0_to_fp16"), val = tensor(0x1p-1)]; - tensor _inversed_90_cast_fp16 = mul(x = var_88_cast_fp16, y = _inversed_90_y_0_to_fp16)[name = tensor("_inversed_90_cast_fp16")]; - tensor var_91_to_fp16 = const()[name = tensor("op_91_to_fp16"), val = tensor(0x1p+0)]; - tensor lengths_1_cast_fp16 = add(x = _inversed_90_cast_fp16, y = var_91_to_fp16)[name = tensor("lengths_1_cast_fp16")]; - tensor lengths_3_cast_fp16 = floor(x = lengths_1_cast_fp16)[name = tensor("lengths_3_cast_fp16")]; - tensor var_95_promoted_to_fp16 = const()[name = tensor("op_95_promoted_to_fp16"), val = tensor(-0x1p+0)]; - tensor var_96_cast_fp16 = add(x = lengths_3_cast_fp16, y = var_95_promoted_to_fp16)[name = tensor("op_96_cast_fp16")]; - tensor _inversed_98_y_0_to_fp16 = const()[name = tensor("_inversed_98_y_0_to_fp16"), val = tensor(0x1p-1)]; - tensor _inversed_98_cast_fp16 = mul(x = var_96_cast_fp16, y = _inversed_98_y_0_to_fp16)[name = tensor("_inversed_98_cast_fp16")]; - tensor var_99_to_fp16 = const()[name = tensor("op_99_to_fp16"), val = tensor(0x1p+0)]; - tensor lengths_7_cast_fp16 = add(x = _inversed_98_cast_fp16, y = var_99_to_fp16)[name = tensor("lengths_7_cast_fp16")]; - tensor lengths_9_cast_fp16 = floor(x = lengths_7_cast_fp16)[name = tensor("lengths_9_cast_fp16")]; - tensor var_103_promoted_to_fp16 = const()[name = tensor("op_103_promoted_to_fp16"), val = tensor(-0x1p+0)]; - tensor var_104_cast_fp16 = add(x = lengths_9_cast_fp16, y = var_103_promoted_to_fp16)[name = tensor("op_104_cast_fp16")]; - tensor _inversed_106_y_0_to_fp16 = const()[name = tensor("_inversed_106_y_0_to_fp16"), val = tensor(0x1p-1)]; - tensor _inversed_106_cast_fp16 = mul(x = var_104_cast_fp16, y = _inversed_106_y_0_to_fp16)[name = tensor("_inversed_106_cast_fp16")]; - tensor var_107_to_fp16 = const()[name = tensor("op_107_to_fp16"), val = tensor(0x1p+0)]; - tensor lengths_13_cast_fp16 = add(x = _inversed_106_cast_fp16, y = var_107_to_fp16)[name = tensor("lengths_13_cast_fp16")]; - tensor lengths_cast_fp16 = floor(x = lengths_13_cast_fp16)[name = tensor("lengths_cast_fp16")]; - tensor input_1_axes_0 = const()[name = tensor("input_1_axes_0"), val = tensor([1])]; - tensor mel_to_fp16 = cast(dtype = mel_to_fp16_dtype_0, x = mel)[name = tensor("cast_2")]; - tensor x_1_cast_fp16 = transpose(perm = x_1_perm_0, x = mel_to_fp16)[name = tensor("transpose_315")]; - tensor input_1_cast_fp16 = expand_dims(axes = input_1_axes_0, x = x_1_cast_fp16)[name = tensor("input_1_cast_fp16")]; - tensor input_3_pad_type_0 = const()[name = tensor("input_3_pad_type_0"), val = tensor("custom")]; - tensor input_3_pad_0 = const()[name = tensor("input_3_pad_0"), val = tensor([1, 1, 1, 1])]; - tensor input_3_strides_0 = const()[name = tensor("input_3_strides_0"), val = tensor([2, 2])]; - tensor input_3_dilations_0 = const()[name = tensor("input_3_dilations_0"), val = tensor([1, 1])]; - tensor input_3_groups_0 = const()[name = tensor("input_3_groups_0"), val = tensor(1)]; - tensor module_pre_encode_conv_0_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1856))), name = tensor("module_pre_encode_conv_0_weight_to_fp16_palettized"), shape = tensor([256, 1, 3, 3])]; - tensor module_pre_encode_conv_0_bias_to_fp16 = const()[name = tensor("module_pre_encode_conv_0_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2048)))]; - tensor input_3_cast_fp16 = conv(bias = module_pre_encode_conv_0_bias_to_fp16, dilations = input_3_dilations_0, groups = input_3_groups_0, pad = input_3_pad_0, pad_type = input_3_pad_type_0, strides = input_3_strides_0, weight = module_pre_encode_conv_0_weight_to_fp16_palettized, x = input_1_cast_fp16)[name = tensor("input_3_cast_fp16")]; - tensor input_5_cast_fp16 = relu(x = input_3_cast_fp16)[name = tensor("input_5_cast_fp16")]; - tensor input_7_pad_type_0 = const()[name = tensor("input_7_pad_type_0"), val = tensor("custom")]; - tensor input_7_pad_0 = const()[name = tensor("input_7_pad_0"), val = tensor([1, 1, 1, 1])]; - tensor input_7_strides_0 = const()[name = tensor("input_7_strides_0"), val = tensor([2, 2])]; - tensor input_7_groups_0 = const()[name = tensor("input_7_groups_0"), val = tensor(256)]; - tensor input_7_dilations_0 = const()[name = tensor("input_7_dilations_0"), val = tensor([1, 1])]; - tensor module_pre_encode_conv_2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2624))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4416))), name = tensor("module_pre_encode_conv_2_weight_to_fp16_palettized"), shape = tensor([256, 1, 3, 3])]; - tensor module_pre_encode_conv_2_bias_to_fp16 = const()[name = tensor("module_pre_encode_conv_2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4608)))]; - tensor input_7_cast_fp16 = conv(bias = module_pre_encode_conv_2_bias_to_fp16, dilations = input_7_dilations_0, groups = input_7_groups_0, pad = input_7_pad_0, pad_type = input_7_pad_type_0, strides = input_7_strides_0, weight = module_pre_encode_conv_2_weight_to_fp16_palettized, x = input_5_cast_fp16)[name = tensor("input_7_cast_fp16")]; - tensor input_9_pad_type_0 = const()[name = tensor("input_9_pad_type_0"), val = tensor("valid")]; - tensor input_9_strides_0 = const()[name = tensor("input_9_strides_0"), val = tensor([1, 1])]; - tensor input_9_pad_0 = const()[name = tensor("input_9_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor input_9_dilations_0 = const()[name = tensor("input_9_dilations_0"), val = tensor([1, 1])]; - tensor input_9_groups_0 = const()[name = tensor("input_9_groups_0"), val = tensor(1)]; - tensor module_pre_encode_conv_3_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(5184))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(54400))), name = tensor("module_pre_encode_conv_3_weight_to_fp16_palettized"), shape = tensor([256, 256, 1, 1])]; - tensor module_pre_encode_conv_3_bias_to_fp16 = const()[name = tensor("module_pre_encode_conv_3_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(54592)))]; - tensor input_9_cast_fp16 = conv(bias = module_pre_encode_conv_3_bias_to_fp16, dilations = input_9_dilations_0, groups = input_9_groups_0, pad = input_9_pad_0, pad_type = input_9_pad_type_0, strides = input_9_strides_0, weight = module_pre_encode_conv_3_weight_to_fp16_palettized, x = input_7_cast_fp16)[name = tensor("input_9_cast_fp16")]; - tensor input_11_cast_fp16 = relu(x = input_9_cast_fp16)[name = tensor("input_11_cast_fp16")]; - tensor input_13_pad_type_0 = const()[name = tensor("input_13_pad_type_0"), val = tensor("custom")]; - tensor input_13_pad_0 = const()[name = tensor("input_13_pad_0"), val = tensor([1, 1, 1, 1])]; - tensor input_13_strides_0 = const()[name = tensor("input_13_strides_0"), val = tensor([2, 2])]; - tensor input_13_groups_0 = const()[name = tensor("input_13_groups_0"), val = tensor(256)]; - tensor input_13_dilations_0 = const()[name = tensor("input_13_dilations_0"), val = tensor([1, 1])]; - tensor module_pre_encode_conv_5_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(55168))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(56960))), name = tensor("module_pre_encode_conv_5_weight_to_fp16_palettized"), shape = tensor([256, 1, 3, 3])]; - tensor module_pre_encode_conv_5_bias_to_fp16 = const()[name = tensor("module_pre_encode_conv_5_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57152)))]; - tensor input_13_cast_fp16 = conv(bias = module_pre_encode_conv_5_bias_to_fp16, dilations = input_13_dilations_0, groups = input_13_groups_0, pad = input_13_pad_0, pad_type = input_13_pad_type_0, strides = input_13_strides_0, weight = module_pre_encode_conv_5_weight_to_fp16_palettized, x = input_11_cast_fp16)[name = tensor("input_13_cast_fp16")]; - tensor input_15_pad_type_0 = const()[name = tensor("input_15_pad_type_0"), val = tensor("valid")]; - tensor input_15_strides_0 = const()[name = tensor("input_15_strides_0"), val = tensor([1, 1])]; - tensor input_15_pad_0 = const()[name = tensor("input_15_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor input_15_dilations_0 = const()[name = tensor("input_15_dilations_0"), val = tensor([1, 1])]; - tensor input_15_groups_0 = const()[name = tensor("input_15_groups_0"), val = tensor(1)]; - tensor module_pre_encode_conv_6_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57728))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(106944))), name = tensor("module_pre_encode_conv_6_weight_to_fp16_palettized"), shape = tensor([256, 256, 1, 1])]; - tensor module_pre_encode_conv_6_bias_to_fp16 = const()[name = tensor("module_pre_encode_conv_6_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(107136)))]; - tensor input_15_cast_fp16 = conv(bias = module_pre_encode_conv_6_bias_to_fp16, dilations = input_15_dilations_0, groups = input_15_groups_0, pad = input_15_pad_0, pad_type = input_15_pad_type_0, strides = input_15_strides_0, weight = module_pre_encode_conv_6_weight_to_fp16_palettized, x = input_13_cast_fp16)[name = tensor("input_15_cast_fp16")]; - tensor x_3_cast_fp16 = relu(x = input_15_cast_fp16)[name = tensor("x_3_cast_fp16")]; - tensor var_157_perm_0 = const()[name = tensor("op_157_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_158 = const()[name = tensor("op_158"), val = tensor([1, 188, -1])]; - tensor var_157_cast_fp16 = transpose(perm = var_157_perm_0, x = x_3_cast_fp16)[name = tensor("transpose_314")]; - tensor input_17_cast_fp16 = reshape(shape = var_158, x = var_157_cast_fp16)[name = tensor("input_17_cast_fp16")]; - tensor module_pre_encode_out_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(107712))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3253504))), name = tensor("module_pre_encode_out_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor module_pre_encode_out_bias_to_fp16 = const()[name = tensor("module_pre_encode_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3253696)))]; - tensor linear_0_cast_fp16 = linear(bias = module_pre_encode_out_bias_to_fp16, weight = module_pre_encode_out_weight_to_fp16_palettized, x = input_17_cast_fp16)[name = tensor("linear_0_cast_fp16")]; - tensor padding_length_dtype_0 = const()[name = tensor("padding_length_dtype_0"), val = tensor("int32")]; - tensor expand_dims_0 = const()[name = tensor("expand_dims_0"), val = tensor([[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187]])]; - tensor var_196_axes_0 = const()[name = tensor("op_196_axes_0"), val = tensor([-1])]; - tensor encoder_length = cast(dtype = padding_length_dtype_0, x = lengths_cast_fp16)[name = tensor("cast_1")]; - tensor var_196 = expand_dims(axes = var_196_axes_0, x = encoder_length)[name = tensor("op_196")]; - tensor pad_mask_1 = less(x = expand_dims_0, y = var_196)[name = tensor("pad_mask_1")]; - tensor var_198_axes_0 = const()[name = tensor("op_198_axes_0"), val = tensor([1])]; - tensor var_198 = expand_dims(axes = var_198_axes_0, x = pad_mask_1)[name = tensor("op_198")]; - tensor var_199 = const()[name = tensor("op_199"), val = tensor([1, 188, 1])]; - tensor pad_mask_for_att_mask_1 = tile(reps = var_199, x = var_198)[name = tensor("pad_mask_for_att_mask_1")]; - tensor var_201_perm_0 = const()[name = tensor("op_201_perm_0"), val = tensor([0, 2, 1])]; - tensor var_201 = transpose(perm = var_201_perm_0, x = pad_mask_for_att_mask_1)[name = tensor("transpose_313")]; - tensor pad_mask_for_att_mask = logical_and(x = pad_mask_for_att_mask_1, y = var_201)[name = tensor("pad_mask_for_att_mask")]; - tensor const_7 = const()[name = tensor("const_7"), val = tensor([[[true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true]]])]; - tensor att_mask = logical_and(x = pad_mask_for_att_mask, y = const_7)[name = tensor("att_mask")]; - tensor mask_1 = logical_not(x = att_mask)[name = tensor("mask_1")]; - tensor pad_mask = logical_not(x = pad_mask_1)[name = tensor("pad_mask")]; - tensor input_21_axes_0 = const()[name = tensor("input_21_axes_0"), val = tensor([-1])]; - tensor module_layers_0_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_0_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3255808)))]; - tensor module_layers_0_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_0_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3257920)))]; - tensor var_9_to_fp16 = const()[name = tensor("op_9_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_21_cast_fp16 = layer_norm(axes = input_21_axes_0, beta = module_layers_0_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_0_norm_feed_forward1_weight_to_fp16, x = linear_0_cast_fp16)[name = tensor("input_21_cast_fp16")]; - tensor module_layers_0_feed_forward1_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3260032))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(6405824))), name = tensor("module_layers_0_feed_forward1_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_1_bias_0_to_fp16 = const()[name = tensor("linear_1_bias_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(6406016)))]; - tensor linear_1_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_0_feed_forward1_linear1_weight_to_fp16_palettized, x = input_21_cast_fp16)[name = tensor("linear_1_cast_fp16")]; - tensor input_25_cast_fp16 = silu(x = linear_1_cast_fp16)[name = tensor("input_25_cast_fp16")]; - tensor module_layers_0_feed_forward1_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(6414272))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9560064))), name = tensor("module_layers_0_feed_forward1_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_2_bias_0_to_fp16 = const()[name = tensor("linear_2_bias_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9560256)))]; - tensor linear_2_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_0_feed_forward1_linear2_weight_to_fp16_palettized, x = input_25_cast_fp16)[name = tensor("linear_2_cast_fp16")]; - tensor var_232_to_fp16 = const()[name = tensor("op_232_to_fp16"), val = tensor(0x1p-1)]; - tensor var_233_cast_fp16 = mul(x = linear_2_cast_fp16, y = var_232_to_fp16)[name = tensor("op_233_cast_fp16")]; - tensor input_31_cast_fp16 = add(x = linear_0_cast_fp16, y = var_233_cast_fp16)[name = tensor("input_31_cast_fp16")]; - tensor query_1_axes_0 = const()[name = tensor("query_1_axes_0"), val = tensor([-1])]; - tensor module_layers_0_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_0_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9562368)))]; - tensor module_layers_0_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_0_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9564480)))]; - tensor query_1_cast_fp16 = layer_norm(axes = query_1_axes_0, beta = module_layers_0_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_0_norm_self_att_weight_to_fp16, x = input_31_cast_fp16)[name = tensor("query_1_cast_fp16")]; - tensor module_layers_0_self_attn_linear_q_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9566592))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10353088))), name = tensor("module_layers_0_self_attn_linear_q_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_3_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_0_self_attn_linear_q_weight_to_fp16_palettized, x = query_1_cast_fp16)[name = tensor("linear_3_cast_fp16")]; - tensor var_249 = const()[name = tensor("op_249"), val = tensor([1, -1, 8, 128])]; - tensor q_1_cast_fp16 = reshape(shape = var_249, x = linear_3_cast_fp16)[name = tensor("q_1_cast_fp16")]; - tensor module_layers_0_self_attn_linear_k_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10353280))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11139776))), name = tensor("module_layers_0_self_attn_linear_k_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_4_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_0_self_attn_linear_k_weight_to_fp16_palettized, x = query_1_cast_fp16)[name = tensor("linear_4_cast_fp16")]; - tensor var_253 = const()[name = tensor("op_253"), val = tensor([1, -1, 8, 128])]; - tensor k_1_cast_fp16 = reshape(shape = var_253, x = linear_4_cast_fp16)[name = tensor("k_1_cast_fp16")]; - tensor module_layers_0_self_attn_linear_v_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11139968))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11926464))), name = tensor("module_layers_0_self_attn_linear_v_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_5_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_0_self_attn_linear_v_weight_to_fp16_palettized, x = query_1_cast_fp16)[name = tensor("linear_5_cast_fp16")]; - tensor var_257 = const()[name = tensor("op_257"), val = tensor([1, -1, 8, 128])]; - tensor v_1_cast_fp16 = reshape(shape = var_257, x = linear_5_cast_fp16)[name = tensor("v_1_cast_fp16")]; - tensor value_3_perm_0 = const()[name = tensor("value_3_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor module_layers_0_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_0_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11926656)))]; - tensor var_269_cast_fp16 = add(x = q_1_cast_fp16, y = module_layers_0_self_attn_pos_bias_u_to_fp16)[name = tensor("op_269_cast_fp16")]; - tensor module_layers_0_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_0_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11928768)))]; - tensor var_271_cast_fp16 = add(x = q_1_cast_fp16, y = module_layers_0_self_attn_pos_bias_v_to_fp16)[name = tensor("op_271_cast_fp16")]; - tensor q_with_bias_v_1_perm_0 = const()[name = tensor("q_with_bias_v_1_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor x_7_transpose_x_0 = const()[name = tensor("x_7_transpose_x_0"), val = tensor(false)]; - tensor x_7_transpose_y_0 = const()[name = tensor("x_7_transpose_y_0"), val = tensor(false)]; - tensor op_273_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11930880))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12218944))), name = tensor("op_273_to_fp16_palettized"), shape = tensor([1, 8, 128, 375])]; - tensor q_with_bias_v_1_cast_fp16 = transpose(perm = q_with_bias_v_1_perm_0, x = var_271_cast_fp16)[name = tensor("transpose_312")]; - tensor x_7_cast_fp16 = matmul(transpose_x = x_7_transpose_x_0, transpose_y = x_7_transpose_y_0, x = q_with_bias_v_1_cast_fp16, y = op_273_to_fp16_palettized)[name = tensor("x_7_cast_fp16")]; - tensor x_9_pad_0 = const()[name = tensor("x_9_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_9_mode_0 = const()[name = tensor("x_9_mode_0"), val = tensor("constant")]; - tensor const_14_to_fp16 = const()[name = tensor("const_14_to_fp16"), val = tensor(0x0p+0)]; - tensor x_9_cast_fp16 = pad(constant_val = const_14_to_fp16, mode = x_9_mode_0, pad = x_9_pad_0, x = x_7_cast_fp16)[name = tensor("x_9_cast_fp16")]; - tensor var_281 = const()[name = tensor("op_281"), val = tensor([1, 8, -1, 188])]; - tensor x_11_cast_fp16 = reshape(shape = var_281, x = x_9_cast_fp16)[name = tensor("x_11_cast_fp16")]; - tensor var_285_begin_0 = const()[name = tensor("op_285_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_285_end_0 = const()[name = tensor("op_285_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_285_end_mask_0 = const()[name = tensor("op_285_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_285_cast_fp16 = slice_by_index(begin = var_285_begin_0, end = var_285_end_0, end_mask = var_285_end_mask_0, x = x_11_cast_fp16)[name = tensor("op_285_cast_fp16")]; - tensor var_286 = const()[name = tensor("op_286"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_1_cast_fp16 = reshape(shape = var_286, x = var_285_cast_fp16)[name = tensor("matrix_bd_1_cast_fp16")]; - tensor matrix_ac_1_transpose_x_0 = const()[name = tensor("matrix_ac_1_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_1_transpose_y_0 = const()[name = tensor("matrix_ac_1_transpose_y_0"), val = tensor(false)]; - tensor transpose_96_perm_0 = const()[name = tensor("transpose_96_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_97_perm_0 = const()[name = tensor("transpose_97_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_97 = transpose(perm = transpose_97_perm_0, x = k_1_cast_fp16)[name = tensor("transpose_310")]; - tensor transpose_96 = transpose(perm = transpose_96_perm_0, x = var_269_cast_fp16)[name = tensor("transpose_311")]; - tensor matrix_ac_1_cast_fp16 = matmul(transpose_x = matrix_ac_1_transpose_x_0, transpose_y = matrix_ac_1_transpose_y_0, x = transpose_96, y = transpose_97)[name = tensor("matrix_ac_1_cast_fp16")]; - tensor matrix_bd_3_begin_0 = const()[name = tensor("matrix_bd_3_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_3_end_0 = const()[name = tensor("matrix_bd_3_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_3_end_mask_0 = const()[name = tensor("matrix_bd_3_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_3_cast_fp16 = slice_by_index(begin = matrix_bd_3_begin_0, end = matrix_bd_3_end_0, end_mask = matrix_bd_3_end_mask_0, x = matrix_bd_1_cast_fp16)[name = tensor("matrix_bd_3_cast_fp16")]; - tensor var_295_cast_fp16 = add(x = matrix_ac_1_cast_fp16, y = matrix_bd_3_cast_fp16)[name = tensor("op_295_cast_fp16")]; - tensor _inversed_scores_1_y_0_to_fp16 = const()[name = tensor("_inversed_scores_1_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; - tensor _inversed_scores_1_cast_fp16 = mul(x = var_295_cast_fp16, y = _inversed_scores_1_y_0_to_fp16)[name = tensor("_inversed_scores_1_cast_fp16")]; - tensor mask_3_axes_0 = const()[name = tensor("mask_3_axes_0"), val = tensor([1])]; - tensor mask_3 = expand_dims(axes = mask_3_axes_0, x = mask_1)[name = tensor("mask_3")]; - tensor var_12_to_fp16 = const()[name = tensor("op_12_to_fp16"), val = tensor(-0x1.388p+13)]; - tensor scores_3_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_1_cast_fp16, cond = mask_3)[name = tensor("scores_3_cast_fp16")]; - tensor var_301_cast_fp16 = softmax(axis = var_30, x = scores_3_cast_fp16)[name = tensor("op_301_cast_fp16")]; - tensor var_11_to_fp16 = const()[name = tensor("op_11_to_fp16"), val = tensor(0x0p+0)]; - tensor input_33_cast_fp16 = select(a = var_11_to_fp16, b = var_301_cast_fp16, cond = mask_3)[name = tensor("input_33_cast_fp16")]; - tensor x_13_transpose_x_0 = const()[name = tensor("x_13_transpose_x_0"), val = tensor(false)]; - tensor x_13_transpose_y_0 = const()[name = tensor("x_13_transpose_y_0"), val = tensor(false)]; - tensor value_3_cast_fp16 = transpose(perm = value_3_perm_0, x = v_1_cast_fp16)[name = tensor("transpose_309")]; - tensor x_13_cast_fp16 = matmul(transpose_x = x_13_transpose_x_0, transpose_y = x_13_transpose_y_0, x = input_33_cast_fp16, y = value_3_cast_fp16)[name = tensor("x_13_cast_fp16")]; - tensor var_305_perm_0 = const()[name = tensor("op_305_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_306 = const()[name = tensor("op_306"), val = tensor([1, -1, 1024])]; - tensor var_305_cast_fp16 = transpose(perm = var_305_perm_0, x = x_13_cast_fp16)[name = tensor("transpose_308")]; - tensor input_35_cast_fp16 = reshape(shape = var_306, x = var_305_cast_fp16)[name = tensor("input_35_cast_fp16")]; - tensor module_layers_0_self_attn_linear_out_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12219136))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(13005632))), name = tensor("module_layers_0_self_attn_linear_out_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_7_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_0_self_attn_linear_out_weight_to_fp16_palettized, x = input_35_cast_fp16)[name = tensor("linear_7_cast_fp16")]; - tensor input_39_cast_fp16 = add(x = input_31_cast_fp16, y = linear_7_cast_fp16)[name = tensor("input_39_cast_fp16")]; - tensor x_17_axes_0 = const()[name = tensor("x_17_axes_0"), val = tensor([-1])]; - tensor module_layers_0_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_0_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(13005824)))]; - tensor module_layers_0_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_0_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(13007936)))]; - tensor x_17_cast_fp16 = layer_norm(axes = x_17_axes_0, beta = module_layers_0_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_0_norm_conv_weight_to_fp16, x = input_39_cast_fp16)[name = tensor("x_17_cast_fp16")]; - tensor input_41_perm_0 = const()[name = tensor("input_41_perm_0"), val = tensor([0, 2, 1])]; - tensor input_43_pad_type_0 = const()[name = tensor("input_43_pad_type_0"), val = tensor("valid")]; - tensor input_43_strides_0 = const()[name = tensor("input_43_strides_0"), val = tensor([1])]; - tensor input_43_pad_0 = const()[name = tensor("input_43_pad_0"), val = tensor([0, 0])]; - tensor input_43_dilations_0 = const()[name = tensor("input_43_dilations_0"), val = tensor([1])]; - tensor input_43_groups_0 = const()[name = tensor("input_43_groups_0"), val = tensor(1)]; - tensor module_layers_0_conv_pointwise_conv1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(13010048))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(14582976))), name = tensor("module_layers_0_conv_pointwise_conv1_weight_to_fp16_palettized"), shape = tensor([2048, 1024, 1])]; - tensor input_41_cast_fp16 = transpose(perm = input_41_perm_0, x = x_17_cast_fp16)[name = tensor("transpose_307")]; - tensor input_43_cast_fp16 = conv(dilations = input_43_dilations_0, groups = input_43_groups_0, pad = input_43_pad_0, pad_type = input_43_pad_type_0, strides = input_43_strides_0, weight = module_layers_0_conv_pointwise_conv1_weight_to_fp16_palettized, x = input_41_cast_fp16)[name = tensor("input_43_cast_fp16")]; - tensor x_19_split_num_splits_0 = const()[name = tensor("x_19_split_num_splits_0"), val = tensor(2)]; - tensor x_19_split_axis_0 = const()[name = tensor("x_19_split_axis_0"), val = tensor(1)]; - tensor x_19_split_cast_fp16_0, tensor x_19_split_cast_fp16_1 = split(axis = x_19_split_axis_0, num_splits = x_19_split_num_splits_0, x = input_43_cast_fp16)[name = tensor("x_19_split_cast_fp16")]; - tensor x_19_split_1_sigmoid_cast_fp16 = sigmoid(x = x_19_split_cast_fp16_1)[name = tensor("x_19_split_1_sigmoid_cast_fp16")]; - tensor x_19_cast_fp16 = mul(x = x_19_split_cast_fp16_0, y = x_19_split_1_sigmoid_cast_fp16)[name = tensor("x_19_cast_fp16")]; - tensor var_328_axes_0 = const()[name = tensor("op_328_axes_0"), val = tensor([1])]; - tensor var_328 = expand_dims(axes = var_328_axes_0, x = pad_mask)[name = tensor("op_328")]; - tensor input_45_cast_fp16 = select(a = var_11_to_fp16, b = x_19_cast_fp16, cond = var_328)[name = tensor("input_45_cast_fp16")]; - tensor input_47_pad_0 = const()[name = tensor("input_47_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_47_mode_0 = const()[name = tensor("input_47_mode_0"), val = tensor("constant")]; - tensor const_17_to_fp16 = const()[name = tensor("const_17_to_fp16"), val = tensor(0x0p+0)]; - tensor input_47_cast_fp16 = pad(constant_val = const_17_to_fp16, mode = input_47_mode_0, pad = input_47_pad_0, x = input_45_cast_fp16)[name = tensor("input_47_cast_fp16")]; - tensor input_49_pad_type_0 = const()[name = tensor("input_49_pad_type_0"), val = tensor("valid")]; - tensor input_49_groups_0 = const()[name = tensor("input_49_groups_0"), val = tensor(1024)]; - tensor input_49_strides_0 = const()[name = tensor("input_49_strides_0"), val = tensor([1])]; - tensor input_49_pad_0 = const()[name = tensor("input_49_pad_0"), val = tensor([0, 0])]; - tensor input_49_dilations_0 = const()[name = tensor("input_49_dilations_0"), val = tensor([1])]; - tensor const_248_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(14583168))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(14590144))), name = tensor("const_248_to_fp16_palettized"), shape = tensor([1024, 1, 9])]; - tensor const_249_to_fp16 = const()[name = tensor("const_249_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(14590336)))]; - tensor input_51_cast_fp16 = conv(bias = const_249_to_fp16, dilations = input_49_dilations_0, groups = input_49_groups_0, pad = input_49_pad_0, pad_type = input_49_pad_type_0, strides = input_49_strides_0, weight = const_248_to_fp16_palettized, x = input_47_cast_fp16)[name = tensor("input_51_cast_fp16")]; - tensor input_53_cast_fp16 = silu(x = input_51_cast_fp16)[name = tensor("input_53_cast_fp16")]; - tensor x_21_pad_type_0 = const()[name = tensor("x_21_pad_type_0"), val = tensor("valid")]; - tensor x_21_strides_0 = const()[name = tensor("x_21_strides_0"), val = tensor([1])]; - tensor x_21_pad_0 = const()[name = tensor("x_21_pad_0"), val = tensor([0, 0])]; - tensor x_21_dilations_0 = const()[name = tensor("x_21_dilations_0"), val = tensor([1])]; - tensor x_21_groups_0 = const()[name = tensor("x_21_groups_0"), val = tensor(1)]; - tensor module_layers_0_conv_pointwise_conv2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(14592448))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15378944))), name = tensor("module_layers_0_conv_pointwise_conv2_weight_to_fp16_palettized"), shape = tensor([1024, 1024, 1])]; - tensor x_21_cast_fp16 = conv(dilations = x_21_dilations_0, groups = x_21_groups_0, pad = x_21_pad_0, pad_type = x_21_pad_type_0, strides = x_21_strides_0, weight = module_layers_0_conv_pointwise_conv2_weight_to_fp16_palettized, x = input_53_cast_fp16)[name = tensor("x_21_cast_fp16")]; - tensor input_55_perm_0 = const()[name = tensor("input_55_perm_0"), val = tensor([0, 2, 1])]; - tensor input_55_cast_fp16 = transpose(perm = input_55_perm_0, x = x_21_cast_fp16)[name = tensor("transpose_306")]; - tensor input_57_cast_fp16 = add(x = input_39_cast_fp16, y = input_55_cast_fp16)[name = tensor("input_57_cast_fp16")]; - tensor input_59_axes_0 = const()[name = tensor("input_59_axes_0"), val = tensor([-1])]; - tensor module_layers_0_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_0_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15379136)))]; - tensor module_layers_0_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_0_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15381248)))]; - tensor input_59_cast_fp16 = layer_norm(axes = input_59_axes_0, beta = module_layers_0_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_0_norm_feed_forward2_weight_to_fp16, x = input_57_cast_fp16)[name = tensor("input_59_cast_fp16")]; - tensor module_layers_0_feed_forward2_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15383360))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(18529152))), name = tensor("module_layers_0_feed_forward2_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_8_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_0_feed_forward2_linear1_weight_to_fp16_palettized, x = input_59_cast_fp16)[name = tensor("linear_8_cast_fp16")]; - tensor input_63_cast_fp16 = silu(x = linear_8_cast_fp16)[name = tensor("input_63_cast_fp16")]; - tensor module_layers_0_feed_forward2_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(18529344))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21675136))), name = tensor("module_layers_0_feed_forward2_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_9_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_0_feed_forward2_linear2_weight_to_fp16_palettized, x = input_63_cast_fp16)[name = tensor("linear_9_cast_fp16")]; - tensor var_366_to_fp16 = const()[name = tensor("op_366_to_fp16"), val = tensor(0x1p-1)]; - tensor var_367_cast_fp16 = mul(x = linear_9_cast_fp16, y = var_366_to_fp16)[name = tensor("op_367_cast_fp16")]; - tensor input_69_cast_fp16 = add(x = input_57_cast_fp16, y = var_367_cast_fp16)[name = tensor("input_69_cast_fp16")]; - tensor input_71_axes_0 = const()[name = tensor("input_71_axes_0"), val = tensor([-1])]; - tensor module_layers_0_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_0_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21675328)))]; - tensor module_layers_0_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_0_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21677440)))]; - tensor input_71_cast_fp16 = layer_norm(axes = input_71_axes_0, beta = module_layers_0_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_0_norm_out_weight_to_fp16, x = input_69_cast_fp16)[name = tensor("input_71_cast_fp16")]; - tensor input_73_axes_0 = const()[name = tensor("input_73_axes_0"), val = tensor([-1])]; - tensor module_layers_1_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_1_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21679552)))]; - tensor module_layers_1_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_1_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21681664)))]; - tensor input_73_cast_fp16 = layer_norm(axes = input_73_axes_0, beta = module_layers_1_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_1_norm_feed_forward1_weight_to_fp16, x = input_71_cast_fp16)[name = tensor("input_73_cast_fp16")]; - tensor module_layers_1_feed_forward1_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21683776))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(24829568))), name = tensor("module_layers_1_feed_forward1_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_10_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_1_feed_forward1_linear1_weight_to_fp16_palettized, x = input_73_cast_fp16)[name = tensor("linear_10_cast_fp16")]; - tensor input_77_cast_fp16 = silu(x = linear_10_cast_fp16)[name = tensor("input_77_cast_fp16")]; - tensor module_layers_1_feed_forward1_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(24829760))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27975552))), name = tensor("module_layers_1_feed_forward1_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_11_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_1_feed_forward1_linear2_weight_to_fp16_palettized, x = input_77_cast_fp16)[name = tensor("linear_11_cast_fp16")]; - tensor var_395_to_fp16 = const()[name = tensor("op_395_to_fp16"), val = tensor(0x1p-1)]; - tensor var_396_cast_fp16 = mul(x = linear_11_cast_fp16, y = var_395_to_fp16)[name = tensor("op_396_cast_fp16")]; - tensor input_83_cast_fp16 = add(x = input_71_cast_fp16, y = var_396_cast_fp16)[name = tensor("input_83_cast_fp16")]; - tensor query_3_axes_0 = const()[name = tensor("query_3_axes_0"), val = tensor([-1])]; - tensor module_layers_1_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_1_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27975744)))]; - tensor module_layers_1_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_1_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27977856)))]; - tensor query_3_cast_fp16 = layer_norm(axes = query_3_axes_0, beta = module_layers_1_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_1_norm_self_att_weight_to_fp16, x = input_83_cast_fp16)[name = tensor("query_3_cast_fp16")]; - tensor module_layers_1_self_attn_linear_q_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27979968))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(28766464))), name = tensor("module_layers_1_self_attn_linear_q_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_12_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_1_self_attn_linear_q_weight_to_fp16_palettized, x = query_3_cast_fp16)[name = tensor("linear_12_cast_fp16")]; - tensor var_412 = const()[name = tensor("op_412"), val = tensor([1, -1, 8, 128])]; - tensor q_7_cast_fp16 = reshape(shape = var_412, x = linear_12_cast_fp16)[name = tensor("q_7_cast_fp16")]; - tensor module_layers_1_self_attn_linear_k_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(28766656))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(29553152))), name = tensor("module_layers_1_self_attn_linear_k_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_13_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_1_self_attn_linear_k_weight_to_fp16_palettized, x = query_3_cast_fp16)[name = tensor("linear_13_cast_fp16")]; - tensor var_416 = const()[name = tensor("op_416"), val = tensor([1, -1, 8, 128])]; - tensor k_5_cast_fp16 = reshape(shape = var_416, x = linear_13_cast_fp16)[name = tensor("k_5_cast_fp16")]; - tensor module_layers_1_self_attn_linear_v_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(29553344))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(30339840))), name = tensor("module_layers_1_self_attn_linear_v_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_14_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_1_self_attn_linear_v_weight_to_fp16_palettized, x = query_3_cast_fp16)[name = tensor("linear_14_cast_fp16")]; - tensor var_420 = const()[name = tensor("op_420"), val = tensor([1, -1, 8, 128])]; - tensor v_3_cast_fp16 = reshape(shape = var_420, x = linear_14_cast_fp16)[name = tensor("v_3_cast_fp16")]; - tensor value_5_perm_0 = const()[name = tensor("value_5_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor module_layers_1_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_1_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(30340032)))]; - tensor var_432_cast_fp16 = add(x = q_7_cast_fp16, y = module_layers_1_self_attn_pos_bias_u_to_fp16)[name = tensor("op_432_cast_fp16")]; - tensor module_layers_1_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_1_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(30342144)))]; - tensor var_434_cast_fp16 = add(x = q_7_cast_fp16, y = module_layers_1_self_attn_pos_bias_v_to_fp16)[name = tensor("op_434_cast_fp16")]; - tensor q_with_bias_v_3_perm_0 = const()[name = tensor("q_with_bias_v_3_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor x_29_transpose_x_0 = const()[name = tensor("x_29_transpose_x_0"), val = tensor(false)]; - tensor x_29_transpose_y_0 = const()[name = tensor("x_29_transpose_y_0"), val = tensor(false)]; - tensor op_436_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(30344256))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(30632320))), name = tensor("op_436_to_fp16_palettized"), shape = tensor([1, 8, 128, 375])]; - tensor q_with_bias_v_3_cast_fp16 = transpose(perm = q_with_bias_v_3_perm_0, x = var_434_cast_fp16)[name = tensor("transpose_305")]; - tensor x_29_cast_fp16 = matmul(transpose_x = x_29_transpose_x_0, transpose_y = x_29_transpose_y_0, x = q_with_bias_v_3_cast_fp16, y = op_436_to_fp16_palettized)[name = tensor("x_29_cast_fp16")]; - tensor x_31_pad_0 = const()[name = tensor("x_31_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_31_mode_0 = const()[name = tensor("x_31_mode_0"), val = tensor("constant")]; - tensor const_24_to_fp16 = const()[name = tensor("const_24_to_fp16"), val = tensor(0x0p+0)]; - tensor x_31_cast_fp16 = pad(constant_val = const_24_to_fp16, mode = x_31_mode_0, pad = x_31_pad_0, x = x_29_cast_fp16)[name = tensor("x_31_cast_fp16")]; - tensor var_444 = const()[name = tensor("op_444"), val = tensor([1, 8, -1, 188])]; - tensor x_33_cast_fp16 = reshape(shape = var_444, x = x_31_cast_fp16)[name = tensor("x_33_cast_fp16")]; - tensor var_448_begin_0 = const()[name = tensor("op_448_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_448_end_0 = const()[name = tensor("op_448_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_448_end_mask_0 = const()[name = tensor("op_448_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_448_cast_fp16 = slice_by_index(begin = var_448_begin_0, end = var_448_end_0, end_mask = var_448_end_mask_0, x = x_33_cast_fp16)[name = tensor("op_448_cast_fp16")]; - tensor var_449 = const()[name = tensor("op_449"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_5_cast_fp16 = reshape(shape = var_449, x = var_448_cast_fp16)[name = tensor("matrix_bd_5_cast_fp16")]; - tensor matrix_ac_3_transpose_x_0 = const()[name = tensor("matrix_ac_3_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_3_transpose_y_0 = const()[name = tensor("matrix_ac_3_transpose_y_0"), val = tensor(false)]; - tensor transpose_98_perm_0 = const()[name = tensor("transpose_98_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_99_perm_0 = const()[name = tensor("transpose_99_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_99 = transpose(perm = transpose_99_perm_0, x = k_5_cast_fp16)[name = tensor("transpose_303")]; - tensor transpose_98 = transpose(perm = transpose_98_perm_0, x = var_432_cast_fp16)[name = tensor("transpose_304")]; - tensor matrix_ac_3_cast_fp16 = matmul(transpose_x = matrix_ac_3_transpose_x_0, transpose_y = matrix_ac_3_transpose_y_0, x = transpose_98, y = transpose_99)[name = tensor("matrix_ac_3_cast_fp16")]; - tensor matrix_bd_7_begin_0 = const()[name = tensor("matrix_bd_7_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_7_end_0 = const()[name = tensor("matrix_bd_7_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_7_end_mask_0 = const()[name = tensor("matrix_bd_7_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_7_cast_fp16 = slice_by_index(begin = matrix_bd_7_begin_0, end = matrix_bd_7_end_0, end_mask = matrix_bd_7_end_mask_0, x = matrix_bd_5_cast_fp16)[name = tensor("matrix_bd_7_cast_fp16")]; - tensor var_458_cast_fp16 = add(x = matrix_ac_3_cast_fp16, y = matrix_bd_7_cast_fp16)[name = tensor("op_458_cast_fp16")]; - tensor _inversed_scores_5_y_0_to_fp16 = const()[name = tensor("_inversed_scores_5_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; - tensor _inversed_scores_5_cast_fp16 = mul(x = var_458_cast_fp16, y = _inversed_scores_5_y_0_to_fp16)[name = tensor("_inversed_scores_5_cast_fp16")]; - tensor scores_7_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_5_cast_fp16, cond = mask_3)[name = tensor("scores_7_cast_fp16")]; - tensor var_464_cast_fp16 = softmax(axis = var_30, x = scores_7_cast_fp16)[name = tensor("op_464_cast_fp16")]; - tensor input_85_cast_fp16 = select(a = var_11_to_fp16, b = var_464_cast_fp16, cond = mask_3)[name = tensor("input_85_cast_fp16")]; - tensor x_35_transpose_x_0 = const()[name = tensor("x_35_transpose_x_0"), val = tensor(false)]; - tensor x_35_transpose_y_0 = const()[name = tensor("x_35_transpose_y_0"), val = tensor(false)]; - tensor value_5_cast_fp16 = transpose(perm = value_5_perm_0, x = v_3_cast_fp16)[name = tensor("transpose_302")]; - tensor x_35_cast_fp16 = matmul(transpose_x = x_35_transpose_x_0, transpose_y = x_35_transpose_y_0, x = input_85_cast_fp16, y = value_5_cast_fp16)[name = tensor("x_35_cast_fp16")]; - tensor var_468_perm_0 = const()[name = tensor("op_468_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_469 = const()[name = tensor("op_469"), val = tensor([1, -1, 1024])]; - tensor var_468_cast_fp16 = transpose(perm = var_468_perm_0, x = x_35_cast_fp16)[name = tensor("transpose_301")]; - tensor input_87_cast_fp16 = reshape(shape = var_469, x = var_468_cast_fp16)[name = tensor("input_87_cast_fp16")]; - tensor module_layers_1_self_attn_linear_out_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(30632512))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(31419008))), name = tensor("module_layers_1_self_attn_linear_out_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_16_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_1_self_attn_linear_out_weight_to_fp16_palettized, x = input_87_cast_fp16)[name = tensor("linear_16_cast_fp16")]; - tensor input_91_cast_fp16 = add(x = input_83_cast_fp16, y = linear_16_cast_fp16)[name = tensor("input_91_cast_fp16")]; - tensor x_39_axes_0 = const()[name = tensor("x_39_axes_0"), val = tensor([-1])]; - tensor module_layers_1_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_1_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(31419200)))]; - tensor module_layers_1_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_1_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(31421312)))]; - tensor x_39_cast_fp16 = layer_norm(axes = x_39_axes_0, beta = module_layers_1_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_1_norm_conv_weight_to_fp16, x = input_91_cast_fp16)[name = tensor("x_39_cast_fp16")]; - tensor input_93_perm_0 = const()[name = tensor("input_93_perm_0"), val = tensor([0, 2, 1])]; - tensor input_95_pad_type_0 = const()[name = tensor("input_95_pad_type_0"), val = tensor("valid")]; - tensor input_95_strides_0 = const()[name = tensor("input_95_strides_0"), val = tensor([1])]; - tensor input_95_pad_0 = const()[name = tensor("input_95_pad_0"), val = tensor([0, 0])]; - tensor input_95_dilations_0 = const()[name = tensor("input_95_dilations_0"), val = tensor([1])]; - tensor input_95_groups_0 = const()[name = tensor("input_95_groups_0"), val = tensor(1)]; - tensor module_layers_1_conv_pointwise_conv1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(31423424))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32996352))), name = tensor("module_layers_1_conv_pointwise_conv1_weight_to_fp16_palettized"), shape = tensor([2048, 1024, 1])]; - tensor input_93_cast_fp16 = transpose(perm = input_93_perm_0, x = x_39_cast_fp16)[name = tensor("transpose_300")]; - tensor input_95_cast_fp16 = conv(dilations = input_95_dilations_0, groups = input_95_groups_0, pad = input_95_pad_0, pad_type = input_95_pad_type_0, strides = input_95_strides_0, weight = module_layers_1_conv_pointwise_conv1_weight_to_fp16_palettized, x = input_93_cast_fp16)[name = tensor("input_95_cast_fp16")]; - tensor x_41_split_num_splits_0 = const()[name = tensor("x_41_split_num_splits_0"), val = tensor(2)]; - tensor x_41_split_axis_0 = const()[name = tensor("x_41_split_axis_0"), val = tensor(1)]; - tensor x_41_split_cast_fp16_0, tensor x_41_split_cast_fp16_1 = split(axis = x_41_split_axis_0, num_splits = x_41_split_num_splits_0, x = input_95_cast_fp16)[name = tensor("x_41_split_cast_fp16")]; - tensor x_41_split_1_sigmoid_cast_fp16 = sigmoid(x = x_41_split_cast_fp16_1)[name = tensor("x_41_split_1_sigmoid_cast_fp16")]; - tensor x_41_cast_fp16 = mul(x = x_41_split_cast_fp16_0, y = x_41_split_1_sigmoid_cast_fp16)[name = tensor("x_41_cast_fp16")]; - tensor input_97_cast_fp16 = select(a = var_11_to_fp16, b = x_41_cast_fp16, cond = var_328)[name = tensor("input_97_cast_fp16")]; - tensor input_99_pad_0 = const()[name = tensor("input_99_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_99_mode_0 = const()[name = tensor("input_99_mode_0"), val = tensor("constant")]; - tensor const_27_to_fp16 = const()[name = tensor("const_27_to_fp16"), val = tensor(0x0p+0)]; - tensor input_99_cast_fp16 = pad(constant_val = const_27_to_fp16, mode = input_99_mode_0, pad = input_99_pad_0, x = input_97_cast_fp16)[name = tensor("input_99_cast_fp16")]; - tensor input_101_pad_type_0 = const()[name = tensor("input_101_pad_type_0"), val = tensor("valid")]; - tensor input_101_groups_0 = const()[name = tensor("input_101_groups_0"), val = tensor(1024)]; - tensor input_101_strides_0 = const()[name = tensor("input_101_strides_0"), val = tensor([1])]; - tensor input_101_pad_0 = const()[name = tensor("input_101_pad_0"), val = tensor([0, 0])]; - tensor input_101_dilations_0 = const()[name = tensor("input_101_dilations_0"), val = tensor([1])]; - tensor const_250_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32996544))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33003520))), name = tensor("const_250_to_fp16_palettized"), shape = tensor([1024, 1, 9])]; - tensor const_251_to_fp16 = const()[name = tensor("const_251_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33003712)))]; - tensor input_103_cast_fp16 = conv(bias = const_251_to_fp16, dilations = input_101_dilations_0, groups = input_101_groups_0, pad = input_101_pad_0, pad_type = input_101_pad_type_0, strides = input_101_strides_0, weight = const_250_to_fp16_palettized, x = input_99_cast_fp16)[name = tensor("input_103_cast_fp16")]; - tensor input_105_cast_fp16 = silu(x = input_103_cast_fp16)[name = tensor("input_105_cast_fp16")]; - tensor x_43_pad_type_0 = const()[name = tensor("x_43_pad_type_0"), val = tensor("valid")]; - tensor x_43_strides_0 = const()[name = tensor("x_43_strides_0"), val = tensor([1])]; - tensor x_43_pad_0 = const()[name = tensor("x_43_pad_0"), val = tensor([0, 0])]; - tensor x_43_dilations_0 = const()[name = tensor("x_43_dilations_0"), val = tensor([1])]; - tensor x_43_groups_0 = const()[name = tensor("x_43_groups_0"), val = tensor(1)]; - tensor module_layers_1_conv_pointwise_conv2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33005824))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33792320))), name = tensor("module_layers_1_conv_pointwise_conv2_weight_to_fp16_palettized"), shape = tensor([1024, 1024, 1])]; - tensor x_43_cast_fp16 = conv(dilations = x_43_dilations_0, groups = x_43_groups_0, pad = x_43_pad_0, pad_type = x_43_pad_type_0, strides = x_43_strides_0, weight = module_layers_1_conv_pointwise_conv2_weight_to_fp16_palettized, x = input_105_cast_fp16)[name = tensor("x_43_cast_fp16")]; - tensor input_107_perm_0 = const()[name = tensor("input_107_perm_0"), val = tensor([0, 2, 1])]; - tensor input_107_cast_fp16 = transpose(perm = input_107_perm_0, x = x_43_cast_fp16)[name = tensor("transpose_299")]; - tensor input_109_cast_fp16 = add(x = input_91_cast_fp16, y = input_107_cast_fp16)[name = tensor("input_109_cast_fp16")]; - tensor input_111_axes_0 = const()[name = tensor("input_111_axes_0"), val = tensor([-1])]; - tensor module_layers_1_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_1_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33792512)))]; - tensor module_layers_1_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_1_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33794624)))]; - tensor input_111_cast_fp16 = layer_norm(axes = input_111_axes_0, beta = module_layers_1_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_1_norm_feed_forward2_weight_to_fp16, x = input_109_cast_fp16)[name = tensor("input_111_cast_fp16")]; - tensor module_layers_1_feed_forward2_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33796736))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36942528))), name = tensor("module_layers_1_feed_forward2_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_17_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_1_feed_forward2_linear1_weight_to_fp16_palettized, x = input_111_cast_fp16)[name = tensor("linear_17_cast_fp16")]; - tensor input_115_cast_fp16 = silu(x = linear_17_cast_fp16)[name = tensor("input_115_cast_fp16")]; - tensor module_layers_1_feed_forward2_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36942720))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40088512))), name = tensor("module_layers_1_feed_forward2_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_18_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_1_feed_forward2_linear2_weight_to_fp16_palettized, x = input_115_cast_fp16)[name = tensor("linear_18_cast_fp16")]; - tensor var_529_to_fp16 = const()[name = tensor("op_529_to_fp16"), val = tensor(0x1p-1)]; - tensor var_530_cast_fp16 = mul(x = linear_18_cast_fp16, y = var_529_to_fp16)[name = tensor("op_530_cast_fp16")]; - tensor input_121_cast_fp16 = add(x = input_109_cast_fp16, y = var_530_cast_fp16)[name = tensor("input_121_cast_fp16")]; - tensor input_123_axes_0 = const()[name = tensor("input_123_axes_0"), val = tensor([-1])]; - tensor module_layers_1_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_1_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40088704)))]; - tensor module_layers_1_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_1_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40090816)))]; - tensor input_123_cast_fp16 = layer_norm(axes = input_123_axes_0, beta = module_layers_1_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_1_norm_out_weight_to_fp16, x = input_121_cast_fp16)[name = tensor("input_123_cast_fp16")]; - tensor input_125_axes_0 = const()[name = tensor("input_125_axes_0"), val = tensor([-1])]; - tensor module_layers_2_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_2_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40092928)))]; - tensor module_layers_2_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_2_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40095040)))]; - tensor input_125_cast_fp16 = layer_norm(axes = input_125_axes_0, beta = module_layers_2_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_2_norm_feed_forward1_weight_to_fp16, x = input_123_cast_fp16)[name = tensor("input_125_cast_fp16")]; - tensor module_layers_2_feed_forward1_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40097152))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(43242944))), name = tensor("module_layers_2_feed_forward1_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_19_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_2_feed_forward1_linear1_weight_to_fp16_palettized, x = input_125_cast_fp16)[name = tensor("linear_19_cast_fp16")]; - tensor input_129_cast_fp16 = silu(x = linear_19_cast_fp16)[name = tensor("input_129_cast_fp16")]; - tensor module_layers_2_feed_forward1_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(43243136))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46388928))), name = tensor("module_layers_2_feed_forward1_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_20_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_2_feed_forward1_linear2_weight_to_fp16_palettized, x = input_129_cast_fp16)[name = tensor("linear_20_cast_fp16")]; - tensor var_558_to_fp16 = const()[name = tensor("op_558_to_fp16"), val = tensor(0x1p-1)]; - tensor var_559_cast_fp16 = mul(x = linear_20_cast_fp16, y = var_558_to_fp16)[name = tensor("op_559_cast_fp16")]; - tensor input_135_cast_fp16 = add(x = input_123_cast_fp16, y = var_559_cast_fp16)[name = tensor("input_135_cast_fp16")]; - tensor query_5_axes_0 = const()[name = tensor("query_5_axes_0"), val = tensor([-1])]; - tensor module_layers_2_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_2_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46389120)))]; - tensor module_layers_2_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_2_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46391232)))]; - tensor query_5_cast_fp16 = layer_norm(axes = query_5_axes_0, beta = module_layers_2_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_2_norm_self_att_weight_to_fp16, x = input_135_cast_fp16)[name = tensor("query_5_cast_fp16")]; - tensor module_layers_2_self_attn_linear_q_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46393344))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47179840))), name = tensor("module_layers_2_self_attn_linear_q_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_21_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_2_self_attn_linear_q_weight_to_fp16_palettized, x = query_5_cast_fp16)[name = tensor("linear_21_cast_fp16")]; - tensor var_575 = const()[name = tensor("op_575"), val = tensor([1, -1, 8, 128])]; - tensor q_13_cast_fp16 = reshape(shape = var_575, x = linear_21_cast_fp16)[name = tensor("q_13_cast_fp16")]; - tensor module_layers_2_self_attn_linear_k_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47180032))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47966528))), name = tensor("module_layers_2_self_attn_linear_k_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_22_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_2_self_attn_linear_k_weight_to_fp16_palettized, x = query_5_cast_fp16)[name = tensor("linear_22_cast_fp16")]; - tensor var_579 = const()[name = tensor("op_579"), val = tensor([1, -1, 8, 128])]; - tensor k_9_cast_fp16 = reshape(shape = var_579, x = linear_22_cast_fp16)[name = tensor("k_9_cast_fp16")]; - tensor module_layers_2_self_attn_linear_v_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47966720))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48753216))), name = tensor("module_layers_2_self_attn_linear_v_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_23_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_2_self_attn_linear_v_weight_to_fp16_palettized, x = query_5_cast_fp16)[name = tensor("linear_23_cast_fp16")]; - tensor var_583 = const()[name = tensor("op_583"), val = tensor([1, -1, 8, 128])]; - tensor v_5_cast_fp16 = reshape(shape = var_583, x = linear_23_cast_fp16)[name = tensor("v_5_cast_fp16")]; - tensor value_7_perm_0 = const()[name = tensor("value_7_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor module_layers_2_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_2_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48753408)))]; - tensor var_595_cast_fp16 = add(x = q_13_cast_fp16, y = module_layers_2_self_attn_pos_bias_u_to_fp16)[name = tensor("op_595_cast_fp16")]; - tensor module_layers_2_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_2_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48755520)))]; - tensor var_597_cast_fp16 = add(x = q_13_cast_fp16, y = module_layers_2_self_attn_pos_bias_v_to_fp16)[name = tensor("op_597_cast_fp16")]; - tensor q_with_bias_v_5_perm_0 = const()[name = tensor("q_with_bias_v_5_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor x_51_transpose_x_0 = const()[name = tensor("x_51_transpose_x_0"), val = tensor(false)]; - tensor x_51_transpose_y_0 = const()[name = tensor("x_51_transpose_y_0"), val = tensor(false)]; - tensor op_599_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48757632))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(49045696))), name = tensor("op_599_to_fp16_palettized"), shape = tensor([1, 8, 128, 375])]; - tensor q_with_bias_v_5_cast_fp16 = transpose(perm = q_with_bias_v_5_perm_0, x = var_597_cast_fp16)[name = tensor("transpose_298")]; - tensor x_51_cast_fp16 = matmul(transpose_x = x_51_transpose_x_0, transpose_y = x_51_transpose_y_0, x = q_with_bias_v_5_cast_fp16, y = op_599_to_fp16_palettized)[name = tensor("x_51_cast_fp16")]; - tensor x_53_pad_0 = const()[name = tensor("x_53_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_53_mode_0 = const()[name = tensor("x_53_mode_0"), val = tensor("constant")]; - tensor const_34_to_fp16 = const()[name = tensor("const_34_to_fp16"), val = tensor(0x0p+0)]; - tensor x_53_cast_fp16 = pad(constant_val = const_34_to_fp16, mode = x_53_mode_0, pad = x_53_pad_0, x = x_51_cast_fp16)[name = tensor("x_53_cast_fp16")]; - tensor var_607 = const()[name = tensor("op_607"), val = tensor([1, 8, -1, 188])]; - tensor x_55_cast_fp16 = reshape(shape = var_607, x = x_53_cast_fp16)[name = tensor("x_55_cast_fp16")]; - tensor var_611_begin_0 = const()[name = tensor("op_611_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_611_end_0 = const()[name = tensor("op_611_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_611_end_mask_0 = const()[name = tensor("op_611_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_611_cast_fp16 = slice_by_index(begin = var_611_begin_0, end = var_611_end_0, end_mask = var_611_end_mask_0, x = x_55_cast_fp16)[name = tensor("op_611_cast_fp16")]; - tensor var_612 = const()[name = tensor("op_612"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_9_cast_fp16 = reshape(shape = var_612, x = var_611_cast_fp16)[name = tensor("matrix_bd_9_cast_fp16")]; - tensor matrix_ac_5_transpose_x_0 = const()[name = tensor("matrix_ac_5_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_5_transpose_y_0 = const()[name = tensor("matrix_ac_5_transpose_y_0"), val = tensor(false)]; - tensor transpose_100_perm_0 = const()[name = tensor("transpose_100_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_101_perm_0 = const()[name = tensor("transpose_101_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_101 = transpose(perm = transpose_101_perm_0, x = k_9_cast_fp16)[name = tensor("transpose_296")]; - tensor transpose_100 = transpose(perm = transpose_100_perm_0, x = var_595_cast_fp16)[name = tensor("transpose_297")]; - tensor matrix_ac_5_cast_fp16 = matmul(transpose_x = matrix_ac_5_transpose_x_0, transpose_y = matrix_ac_5_transpose_y_0, x = transpose_100, y = transpose_101)[name = tensor("matrix_ac_5_cast_fp16")]; - tensor matrix_bd_11_begin_0 = const()[name = tensor("matrix_bd_11_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_11_end_0 = const()[name = tensor("matrix_bd_11_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_11_end_mask_0 = const()[name = tensor("matrix_bd_11_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_11_cast_fp16 = slice_by_index(begin = matrix_bd_11_begin_0, end = matrix_bd_11_end_0, end_mask = matrix_bd_11_end_mask_0, x = matrix_bd_9_cast_fp16)[name = tensor("matrix_bd_11_cast_fp16")]; - tensor var_621_cast_fp16 = add(x = matrix_ac_5_cast_fp16, y = matrix_bd_11_cast_fp16)[name = tensor("op_621_cast_fp16")]; - tensor _inversed_scores_9_y_0_to_fp16 = const()[name = tensor("_inversed_scores_9_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; - tensor _inversed_scores_9_cast_fp16 = mul(x = var_621_cast_fp16, y = _inversed_scores_9_y_0_to_fp16)[name = tensor("_inversed_scores_9_cast_fp16")]; - tensor scores_11_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_9_cast_fp16, cond = mask_3)[name = tensor("scores_11_cast_fp16")]; - tensor var_627_cast_fp16 = softmax(axis = var_30, x = scores_11_cast_fp16)[name = tensor("op_627_cast_fp16")]; - tensor input_137_cast_fp16 = select(a = var_11_to_fp16, b = var_627_cast_fp16, cond = mask_3)[name = tensor("input_137_cast_fp16")]; - tensor x_57_transpose_x_0 = const()[name = tensor("x_57_transpose_x_0"), val = tensor(false)]; - tensor x_57_transpose_y_0 = const()[name = tensor("x_57_transpose_y_0"), val = tensor(false)]; - tensor value_7_cast_fp16 = transpose(perm = value_7_perm_0, x = v_5_cast_fp16)[name = tensor("transpose_295")]; - tensor x_57_cast_fp16 = matmul(transpose_x = x_57_transpose_x_0, transpose_y = x_57_transpose_y_0, x = input_137_cast_fp16, y = value_7_cast_fp16)[name = tensor("x_57_cast_fp16")]; - tensor var_631_perm_0 = const()[name = tensor("op_631_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_632 = const()[name = tensor("op_632"), val = tensor([1, -1, 1024])]; - tensor var_631_cast_fp16 = transpose(perm = var_631_perm_0, x = x_57_cast_fp16)[name = tensor("transpose_294")]; - tensor input_139_cast_fp16 = reshape(shape = var_632, x = var_631_cast_fp16)[name = tensor("input_139_cast_fp16")]; - tensor module_layers_2_self_attn_linear_out_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(49045888))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(49832384))), name = tensor("module_layers_2_self_attn_linear_out_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_25_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_2_self_attn_linear_out_weight_to_fp16_palettized, x = input_139_cast_fp16)[name = tensor("linear_25_cast_fp16")]; - tensor input_143_cast_fp16 = add(x = input_135_cast_fp16, y = linear_25_cast_fp16)[name = tensor("input_143_cast_fp16")]; - tensor x_61_axes_0 = const()[name = tensor("x_61_axes_0"), val = tensor([-1])]; - tensor module_layers_2_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_2_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(49832576)))]; - tensor module_layers_2_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_2_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(49834688)))]; - tensor x_61_cast_fp16 = layer_norm(axes = x_61_axes_0, beta = module_layers_2_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_2_norm_conv_weight_to_fp16, x = input_143_cast_fp16)[name = tensor("x_61_cast_fp16")]; - tensor input_145_perm_0 = const()[name = tensor("input_145_perm_0"), val = tensor([0, 2, 1])]; - tensor input_147_pad_type_0 = const()[name = tensor("input_147_pad_type_0"), val = tensor("valid")]; - tensor input_147_strides_0 = const()[name = tensor("input_147_strides_0"), val = tensor([1])]; - tensor input_147_pad_0 = const()[name = tensor("input_147_pad_0"), val = tensor([0, 0])]; - tensor input_147_dilations_0 = const()[name = tensor("input_147_dilations_0"), val = tensor([1])]; - tensor input_147_groups_0 = const()[name = tensor("input_147_groups_0"), val = tensor(1)]; - tensor module_layers_2_conv_pointwise_conv1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(49836800))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(51409728))), name = tensor("module_layers_2_conv_pointwise_conv1_weight_to_fp16_palettized"), shape = tensor([2048, 1024, 1])]; - tensor input_145_cast_fp16 = transpose(perm = input_145_perm_0, x = x_61_cast_fp16)[name = tensor("transpose_293")]; - tensor input_147_cast_fp16 = conv(dilations = input_147_dilations_0, groups = input_147_groups_0, pad = input_147_pad_0, pad_type = input_147_pad_type_0, strides = input_147_strides_0, weight = module_layers_2_conv_pointwise_conv1_weight_to_fp16_palettized, x = input_145_cast_fp16)[name = tensor("input_147_cast_fp16")]; - tensor x_63_split_num_splits_0 = const()[name = tensor("x_63_split_num_splits_0"), val = tensor(2)]; - tensor x_63_split_axis_0 = const()[name = tensor("x_63_split_axis_0"), val = tensor(1)]; - tensor x_63_split_cast_fp16_0, tensor x_63_split_cast_fp16_1 = split(axis = x_63_split_axis_0, num_splits = x_63_split_num_splits_0, x = input_147_cast_fp16)[name = tensor("x_63_split_cast_fp16")]; - tensor x_63_split_1_sigmoid_cast_fp16 = sigmoid(x = x_63_split_cast_fp16_1)[name = tensor("x_63_split_1_sigmoid_cast_fp16")]; - tensor x_63_cast_fp16 = mul(x = x_63_split_cast_fp16_0, y = x_63_split_1_sigmoid_cast_fp16)[name = tensor("x_63_cast_fp16")]; - tensor input_149_cast_fp16 = select(a = var_11_to_fp16, b = x_63_cast_fp16, cond = var_328)[name = tensor("input_149_cast_fp16")]; - tensor input_151_pad_0 = const()[name = tensor("input_151_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_151_mode_0 = const()[name = tensor("input_151_mode_0"), val = tensor("constant")]; - tensor const_37_to_fp16 = const()[name = tensor("const_37_to_fp16"), val = tensor(0x0p+0)]; - tensor input_151_cast_fp16 = pad(constant_val = const_37_to_fp16, mode = input_151_mode_0, pad = input_151_pad_0, x = input_149_cast_fp16)[name = tensor("input_151_cast_fp16")]; - tensor input_153_pad_type_0 = const()[name = tensor("input_153_pad_type_0"), val = tensor("valid")]; - tensor input_153_groups_0 = const()[name = tensor("input_153_groups_0"), val = tensor(1024)]; - tensor input_153_strides_0 = const()[name = tensor("input_153_strides_0"), val = tensor([1])]; - tensor input_153_pad_0 = const()[name = tensor("input_153_pad_0"), val = tensor([0, 0])]; - tensor input_153_dilations_0 = const()[name = tensor("input_153_dilations_0"), val = tensor([1])]; - tensor const_252_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(51409920))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(51416896))), name = tensor("const_252_to_fp16_palettized"), shape = tensor([1024, 1, 9])]; - tensor const_253_to_fp16 = const()[name = tensor("const_253_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(51417088)))]; - tensor input_155_cast_fp16 = conv(bias = const_253_to_fp16, dilations = input_153_dilations_0, groups = input_153_groups_0, pad = input_153_pad_0, pad_type = input_153_pad_type_0, strides = input_153_strides_0, weight = const_252_to_fp16_palettized, x = input_151_cast_fp16)[name = tensor("input_155_cast_fp16")]; - tensor input_157_cast_fp16 = silu(x = input_155_cast_fp16)[name = tensor("input_157_cast_fp16")]; - tensor x_65_pad_type_0 = const()[name = tensor("x_65_pad_type_0"), val = tensor("valid")]; - tensor x_65_strides_0 = const()[name = tensor("x_65_strides_0"), val = tensor([1])]; - tensor x_65_pad_0 = const()[name = tensor("x_65_pad_0"), val = tensor([0, 0])]; - tensor x_65_dilations_0 = const()[name = tensor("x_65_dilations_0"), val = tensor([1])]; - tensor x_65_groups_0 = const()[name = tensor("x_65_groups_0"), val = tensor(1)]; - tensor module_layers_2_conv_pointwise_conv2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(51419200))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52205696))), name = tensor("module_layers_2_conv_pointwise_conv2_weight_to_fp16_palettized"), shape = tensor([1024, 1024, 1])]; - tensor x_65_cast_fp16 = conv(dilations = x_65_dilations_0, groups = x_65_groups_0, pad = x_65_pad_0, pad_type = x_65_pad_type_0, strides = x_65_strides_0, weight = module_layers_2_conv_pointwise_conv2_weight_to_fp16_palettized, x = input_157_cast_fp16)[name = tensor("x_65_cast_fp16")]; - tensor input_159_perm_0 = const()[name = tensor("input_159_perm_0"), val = tensor([0, 2, 1])]; - tensor input_159_cast_fp16 = transpose(perm = input_159_perm_0, x = x_65_cast_fp16)[name = tensor("transpose_292")]; - tensor input_161_cast_fp16 = add(x = input_143_cast_fp16, y = input_159_cast_fp16)[name = tensor("input_161_cast_fp16")]; - tensor input_163_axes_0 = const()[name = tensor("input_163_axes_0"), val = tensor([-1])]; - tensor module_layers_2_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_2_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52205888)))]; - tensor module_layers_2_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_2_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52208000)))]; - tensor input_163_cast_fp16 = layer_norm(axes = input_163_axes_0, beta = module_layers_2_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_2_norm_feed_forward2_weight_to_fp16, x = input_161_cast_fp16)[name = tensor("input_163_cast_fp16")]; - tensor module_layers_2_feed_forward2_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52210112))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(55355904))), name = tensor("module_layers_2_feed_forward2_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_26_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_2_feed_forward2_linear1_weight_to_fp16_palettized, x = input_163_cast_fp16)[name = tensor("linear_26_cast_fp16")]; - tensor input_167_cast_fp16 = silu(x = linear_26_cast_fp16)[name = tensor("input_167_cast_fp16")]; - tensor module_layers_2_feed_forward2_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(55356096))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58501888))), name = tensor("module_layers_2_feed_forward2_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_27_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_2_feed_forward2_linear2_weight_to_fp16_palettized, x = input_167_cast_fp16)[name = tensor("linear_27_cast_fp16")]; - tensor var_692_to_fp16 = const()[name = tensor("op_692_to_fp16"), val = tensor(0x1p-1)]; - tensor var_693_cast_fp16 = mul(x = linear_27_cast_fp16, y = var_692_to_fp16)[name = tensor("op_693_cast_fp16")]; - tensor input_173_cast_fp16 = add(x = input_161_cast_fp16, y = var_693_cast_fp16)[name = tensor("input_173_cast_fp16")]; - tensor input_175_axes_0 = const()[name = tensor("input_175_axes_0"), val = tensor([-1])]; - tensor module_layers_2_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_2_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58502080)))]; - tensor module_layers_2_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_2_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58504192)))]; - tensor input_175_cast_fp16 = layer_norm(axes = input_175_axes_0, beta = module_layers_2_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_2_norm_out_weight_to_fp16, x = input_173_cast_fp16)[name = tensor("input_175_cast_fp16")]; - tensor input_177_axes_0 = const()[name = tensor("input_177_axes_0"), val = tensor([-1])]; - tensor module_layers_3_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_3_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58506304)))]; - tensor module_layers_3_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_3_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58508416)))]; - tensor input_177_cast_fp16 = layer_norm(axes = input_177_axes_0, beta = module_layers_3_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_3_norm_feed_forward1_weight_to_fp16, x = input_175_cast_fp16)[name = tensor("input_177_cast_fp16")]; - tensor module_layers_3_feed_forward1_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58510528))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61656320))), name = tensor("module_layers_3_feed_forward1_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_28_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_3_feed_forward1_linear1_weight_to_fp16_palettized, x = input_177_cast_fp16)[name = tensor("linear_28_cast_fp16")]; - tensor input_181_cast_fp16 = silu(x = linear_28_cast_fp16)[name = tensor("input_181_cast_fp16")]; - tensor module_layers_3_feed_forward1_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61656512))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64802304))), name = tensor("module_layers_3_feed_forward1_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_29_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_3_feed_forward1_linear2_weight_to_fp16_palettized, x = input_181_cast_fp16)[name = tensor("linear_29_cast_fp16")]; - tensor var_721_to_fp16 = const()[name = tensor("op_721_to_fp16"), val = tensor(0x1p-1)]; - tensor var_722_cast_fp16 = mul(x = linear_29_cast_fp16, y = var_721_to_fp16)[name = tensor("op_722_cast_fp16")]; - tensor input_187_cast_fp16 = add(x = input_175_cast_fp16, y = var_722_cast_fp16)[name = tensor("input_187_cast_fp16")]; - tensor query_7_axes_0 = const()[name = tensor("query_7_axes_0"), val = tensor([-1])]; - tensor module_layers_3_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_3_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64802496)))]; - tensor module_layers_3_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_3_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64804608)))]; - tensor query_7_cast_fp16 = layer_norm(axes = query_7_axes_0, beta = module_layers_3_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_3_norm_self_att_weight_to_fp16, x = input_187_cast_fp16)[name = tensor("query_7_cast_fp16")]; - tensor module_layers_3_self_attn_linear_q_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64806720))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65593216))), name = tensor("module_layers_3_self_attn_linear_q_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_30_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_3_self_attn_linear_q_weight_to_fp16_palettized, x = query_7_cast_fp16)[name = tensor("linear_30_cast_fp16")]; - tensor var_738 = const()[name = tensor("op_738"), val = tensor([1, -1, 8, 128])]; - tensor q_19_cast_fp16 = reshape(shape = var_738, x = linear_30_cast_fp16)[name = tensor("q_19_cast_fp16")]; - tensor module_layers_3_self_attn_linear_k_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65593408))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(66379904))), name = tensor("module_layers_3_self_attn_linear_k_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_31_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_3_self_attn_linear_k_weight_to_fp16_palettized, x = query_7_cast_fp16)[name = tensor("linear_31_cast_fp16")]; - tensor var_742 = const()[name = tensor("op_742"), val = tensor([1, -1, 8, 128])]; - tensor k_13_cast_fp16 = reshape(shape = var_742, x = linear_31_cast_fp16)[name = tensor("k_13_cast_fp16")]; - tensor module_layers_3_self_attn_linear_v_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(66380096))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67166592))), name = tensor("module_layers_3_self_attn_linear_v_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_32_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_3_self_attn_linear_v_weight_to_fp16_palettized, x = query_7_cast_fp16)[name = tensor("linear_32_cast_fp16")]; - tensor var_746 = const()[name = tensor("op_746"), val = tensor([1, -1, 8, 128])]; - tensor v_7_cast_fp16 = reshape(shape = var_746, x = linear_32_cast_fp16)[name = tensor("v_7_cast_fp16")]; - tensor value_9_perm_0 = const()[name = tensor("value_9_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor module_layers_3_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_3_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67166784)))]; - tensor var_758_cast_fp16 = add(x = q_19_cast_fp16, y = module_layers_3_self_attn_pos_bias_u_to_fp16)[name = tensor("op_758_cast_fp16")]; - tensor module_layers_3_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_3_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67168896)))]; - tensor var_760_cast_fp16 = add(x = q_19_cast_fp16, y = module_layers_3_self_attn_pos_bias_v_to_fp16)[name = tensor("op_760_cast_fp16")]; - tensor q_with_bias_v_7_perm_0 = const()[name = tensor("q_with_bias_v_7_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor x_73_transpose_x_0 = const()[name = tensor("x_73_transpose_x_0"), val = tensor(false)]; - tensor x_73_transpose_y_0 = const()[name = tensor("x_73_transpose_y_0"), val = tensor(false)]; - tensor op_762_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67171008))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67459072))), name = tensor("op_762_to_fp16_palettized"), shape = tensor([1, 8, 128, 375])]; - tensor q_with_bias_v_7_cast_fp16 = transpose(perm = q_with_bias_v_7_perm_0, x = var_760_cast_fp16)[name = tensor("transpose_291")]; - tensor x_73_cast_fp16 = matmul(transpose_x = x_73_transpose_x_0, transpose_y = x_73_transpose_y_0, x = q_with_bias_v_7_cast_fp16, y = op_762_to_fp16_palettized)[name = tensor("x_73_cast_fp16")]; - tensor x_75_pad_0 = const()[name = tensor("x_75_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_75_mode_0 = const()[name = tensor("x_75_mode_0"), val = tensor("constant")]; - tensor const_44_to_fp16 = const()[name = tensor("const_44_to_fp16"), val = tensor(0x0p+0)]; - tensor x_75_cast_fp16 = pad(constant_val = const_44_to_fp16, mode = x_75_mode_0, pad = x_75_pad_0, x = x_73_cast_fp16)[name = tensor("x_75_cast_fp16")]; - tensor var_770 = const()[name = tensor("op_770"), val = tensor([1, 8, -1, 188])]; - tensor x_77_cast_fp16 = reshape(shape = var_770, x = x_75_cast_fp16)[name = tensor("x_77_cast_fp16")]; - tensor var_774_begin_0 = const()[name = tensor("op_774_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_774_end_0 = const()[name = tensor("op_774_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_774_end_mask_0 = const()[name = tensor("op_774_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_774_cast_fp16 = slice_by_index(begin = var_774_begin_0, end = var_774_end_0, end_mask = var_774_end_mask_0, x = x_77_cast_fp16)[name = tensor("op_774_cast_fp16")]; - tensor var_775 = const()[name = tensor("op_775"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_13_cast_fp16 = reshape(shape = var_775, x = var_774_cast_fp16)[name = tensor("matrix_bd_13_cast_fp16")]; - tensor matrix_ac_7_transpose_x_0 = const()[name = tensor("matrix_ac_7_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_7_transpose_y_0 = const()[name = tensor("matrix_ac_7_transpose_y_0"), val = tensor(false)]; - tensor transpose_102_perm_0 = const()[name = tensor("transpose_102_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_103_perm_0 = const()[name = tensor("transpose_103_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_103 = transpose(perm = transpose_103_perm_0, x = k_13_cast_fp16)[name = tensor("transpose_289")]; - tensor transpose_102 = transpose(perm = transpose_102_perm_0, x = var_758_cast_fp16)[name = tensor("transpose_290")]; - tensor matrix_ac_7_cast_fp16 = matmul(transpose_x = matrix_ac_7_transpose_x_0, transpose_y = matrix_ac_7_transpose_y_0, x = transpose_102, y = transpose_103)[name = tensor("matrix_ac_7_cast_fp16")]; - tensor matrix_bd_15_begin_0 = const()[name = tensor("matrix_bd_15_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_15_end_0 = const()[name = tensor("matrix_bd_15_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_15_end_mask_0 = const()[name = tensor("matrix_bd_15_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_15_cast_fp16 = slice_by_index(begin = matrix_bd_15_begin_0, end = matrix_bd_15_end_0, end_mask = matrix_bd_15_end_mask_0, x = matrix_bd_13_cast_fp16)[name = tensor("matrix_bd_15_cast_fp16")]; - tensor var_784_cast_fp16 = add(x = matrix_ac_7_cast_fp16, y = matrix_bd_15_cast_fp16)[name = tensor("op_784_cast_fp16")]; - tensor _inversed_scores_13_y_0_to_fp16 = const()[name = tensor("_inversed_scores_13_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; - tensor _inversed_scores_13_cast_fp16 = mul(x = var_784_cast_fp16, y = _inversed_scores_13_y_0_to_fp16)[name = tensor("_inversed_scores_13_cast_fp16")]; - tensor scores_15_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_13_cast_fp16, cond = mask_3)[name = tensor("scores_15_cast_fp16")]; - tensor var_790_cast_fp16 = softmax(axis = var_30, x = scores_15_cast_fp16)[name = tensor("op_790_cast_fp16")]; - tensor input_189_cast_fp16 = select(a = var_11_to_fp16, b = var_790_cast_fp16, cond = mask_3)[name = tensor("input_189_cast_fp16")]; - tensor x_79_transpose_x_0 = const()[name = tensor("x_79_transpose_x_0"), val = tensor(false)]; - tensor x_79_transpose_y_0 = const()[name = tensor("x_79_transpose_y_0"), val = tensor(false)]; - tensor value_9_cast_fp16 = transpose(perm = value_9_perm_0, x = v_7_cast_fp16)[name = tensor("transpose_288")]; - tensor x_79_cast_fp16 = matmul(transpose_x = x_79_transpose_x_0, transpose_y = x_79_transpose_y_0, x = input_189_cast_fp16, y = value_9_cast_fp16)[name = tensor("x_79_cast_fp16")]; - tensor var_794_perm_0 = const()[name = tensor("op_794_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_795 = const()[name = tensor("op_795"), val = tensor([1, -1, 1024])]; - tensor var_794_cast_fp16 = transpose(perm = var_794_perm_0, x = x_79_cast_fp16)[name = tensor("transpose_287")]; - tensor input_191_cast_fp16 = reshape(shape = var_795, x = var_794_cast_fp16)[name = tensor("input_191_cast_fp16")]; - tensor module_layers_3_self_attn_linear_out_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67459264))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(68245760))), name = tensor("module_layers_3_self_attn_linear_out_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_34_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_3_self_attn_linear_out_weight_to_fp16_palettized, x = input_191_cast_fp16)[name = tensor("linear_34_cast_fp16")]; - tensor input_195_cast_fp16 = add(x = input_187_cast_fp16, y = linear_34_cast_fp16)[name = tensor("input_195_cast_fp16")]; - tensor x_83_axes_0 = const()[name = tensor("x_83_axes_0"), val = tensor([-1])]; - tensor module_layers_3_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_3_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(68245952)))]; - tensor module_layers_3_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_3_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(68248064)))]; - tensor x_83_cast_fp16 = layer_norm(axes = x_83_axes_0, beta = module_layers_3_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_3_norm_conv_weight_to_fp16, x = input_195_cast_fp16)[name = tensor("x_83_cast_fp16")]; - tensor input_197_perm_0 = const()[name = tensor("input_197_perm_0"), val = tensor([0, 2, 1])]; - tensor input_199_pad_type_0 = const()[name = tensor("input_199_pad_type_0"), val = tensor("valid")]; - tensor input_199_strides_0 = const()[name = tensor("input_199_strides_0"), val = tensor([1])]; - tensor input_199_pad_0 = const()[name = tensor("input_199_pad_0"), val = tensor([0, 0])]; - tensor input_199_dilations_0 = const()[name = tensor("input_199_dilations_0"), val = tensor([1])]; - tensor input_199_groups_0 = const()[name = tensor("input_199_groups_0"), val = tensor(1)]; - tensor module_layers_3_conv_pointwise_conv1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(68250176))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(69823104))), name = tensor("module_layers_3_conv_pointwise_conv1_weight_to_fp16_palettized"), shape = tensor([2048, 1024, 1])]; - tensor input_197_cast_fp16 = transpose(perm = input_197_perm_0, x = x_83_cast_fp16)[name = tensor("transpose_286")]; - tensor input_199_cast_fp16 = conv(dilations = input_199_dilations_0, groups = input_199_groups_0, pad = input_199_pad_0, pad_type = input_199_pad_type_0, strides = input_199_strides_0, weight = module_layers_3_conv_pointwise_conv1_weight_to_fp16_palettized, x = input_197_cast_fp16)[name = tensor("input_199_cast_fp16")]; - tensor x_85_split_num_splits_0 = const()[name = tensor("x_85_split_num_splits_0"), val = tensor(2)]; - tensor x_85_split_axis_0 = const()[name = tensor("x_85_split_axis_0"), val = tensor(1)]; - tensor x_85_split_cast_fp16_0, tensor x_85_split_cast_fp16_1 = split(axis = x_85_split_axis_0, num_splits = x_85_split_num_splits_0, x = input_199_cast_fp16)[name = tensor("x_85_split_cast_fp16")]; - tensor x_85_split_1_sigmoid_cast_fp16 = sigmoid(x = x_85_split_cast_fp16_1)[name = tensor("x_85_split_1_sigmoid_cast_fp16")]; - tensor x_85_cast_fp16 = mul(x = x_85_split_cast_fp16_0, y = x_85_split_1_sigmoid_cast_fp16)[name = tensor("x_85_cast_fp16")]; - tensor input_201_cast_fp16 = select(a = var_11_to_fp16, b = x_85_cast_fp16, cond = var_328)[name = tensor("input_201_cast_fp16")]; - tensor input_203_pad_0 = const()[name = tensor("input_203_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_203_mode_0 = const()[name = tensor("input_203_mode_0"), val = tensor("constant")]; - tensor const_47_to_fp16 = const()[name = tensor("const_47_to_fp16"), val = tensor(0x0p+0)]; - tensor input_203_cast_fp16 = pad(constant_val = const_47_to_fp16, mode = input_203_mode_0, pad = input_203_pad_0, x = input_201_cast_fp16)[name = tensor("input_203_cast_fp16")]; - tensor input_205_pad_type_0 = const()[name = tensor("input_205_pad_type_0"), val = tensor("valid")]; - tensor input_205_groups_0 = const()[name = tensor("input_205_groups_0"), val = tensor(1024)]; - tensor input_205_strides_0 = const()[name = tensor("input_205_strides_0"), val = tensor([1])]; - tensor input_205_pad_0 = const()[name = tensor("input_205_pad_0"), val = tensor([0, 0])]; - tensor input_205_dilations_0 = const()[name = tensor("input_205_dilations_0"), val = tensor([1])]; - tensor const_254_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(69823296))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(69830272))), name = tensor("const_254_to_fp16_palettized"), shape = tensor([1024, 1, 9])]; - tensor const_255_to_fp16 = const()[name = tensor("const_255_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(69830464)))]; - tensor input_207_cast_fp16 = conv(bias = const_255_to_fp16, dilations = input_205_dilations_0, groups = input_205_groups_0, pad = input_205_pad_0, pad_type = input_205_pad_type_0, strides = input_205_strides_0, weight = const_254_to_fp16_palettized, x = input_203_cast_fp16)[name = tensor("input_207_cast_fp16")]; - tensor input_209_cast_fp16 = silu(x = input_207_cast_fp16)[name = tensor("input_209_cast_fp16")]; - tensor x_87_pad_type_0 = const()[name = tensor("x_87_pad_type_0"), val = tensor("valid")]; - tensor x_87_strides_0 = const()[name = tensor("x_87_strides_0"), val = tensor([1])]; - tensor x_87_pad_0 = const()[name = tensor("x_87_pad_0"), val = tensor([0, 0])]; - tensor x_87_dilations_0 = const()[name = tensor("x_87_dilations_0"), val = tensor([1])]; - tensor x_87_groups_0 = const()[name = tensor("x_87_groups_0"), val = tensor(1)]; - tensor module_layers_3_conv_pointwise_conv2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(69832576))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70619072))), name = tensor("module_layers_3_conv_pointwise_conv2_weight_to_fp16_palettized"), shape = tensor([1024, 1024, 1])]; - tensor x_87_cast_fp16 = conv(dilations = x_87_dilations_0, groups = x_87_groups_0, pad = x_87_pad_0, pad_type = x_87_pad_type_0, strides = x_87_strides_0, weight = module_layers_3_conv_pointwise_conv2_weight_to_fp16_palettized, x = input_209_cast_fp16)[name = tensor("x_87_cast_fp16")]; - tensor input_211_perm_0 = const()[name = tensor("input_211_perm_0"), val = tensor([0, 2, 1])]; - tensor input_211_cast_fp16 = transpose(perm = input_211_perm_0, x = x_87_cast_fp16)[name = tensor("transpose_285")]; - tensor input_213_cast_fp16 = add(x = input_195_cast_fp16, y = input_211_cast_fp16)[name = tensor("input_213_cast_fp16")]; - tensor input_215_axes_0 = const()[name = tensor("input_215_axes_0"), val = tensor([-1])]; - tensor module_layers_3_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_3_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70619264)))]; - tensor module_layers_3_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_3_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70621376)))]; - tensor input_215_cast_fp16 = layer_norm(axes = input_215_axes_0, beta = module_layers_3_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_3_norm_feed_forward2_weight_to_fp16, x = input_213_cast_fp16)[name = tensor("input_215_cast_fp16")]; - tensor module_layers_3_feed_forward2_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70623488))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73769280))), name = tensor("module_layers_3_feed_forward2_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_35_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_3_feed_forward2_linear1_weight_to_fp16_palettized, x = input_215_cast_fp16)[name = tensor("linear_35_cast_fp16")]; - tensor input_219_cast_fp16 = silu(x = linear_35_cast_fp16)[name = tensor("input_219_cast_fp16")]; - tensor module_layers_3_feed_forward2_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73769472))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(76915264))), name = tensor("module_layers_3_feed_forward2_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_36_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_3_feed_forward2_linear2_weight_to_fp16_palettized, x = input_219_cast_fp16)[name = tensor("linear_36_cast_fp16")]; - tensor var_855_to_fp16 = const()[name = tensor("op_855_to_fp16"), val = tensor(0x1p-1)]; - tensor var_856_cast_fp16 = mul(x = linear_36_cast_fp16, y = var_855_to_fp16)[name = tensor("op_856_cast_fp16")]; - tensor input_225_cast_fp16 = add(x = input_213_cast_fp16, y = var_856_cast_fp16)[name = tensor("input_225_cast_fp16")]; - tensor input_227_axes_0 = const()[name = tensor("input_227_axes_0"), val = tensor([-1])]; - tensor module_layers_3_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_3_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(76915456)))]; - tensor module_layers_3_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_3_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(76917568)))]; - tensor input_227_cast_fp16 = layer_norm(axes = input_227_axes_0, beta = module_layers_3_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_3_norm_out_weight_to_fp16, x = input_225_cast_fp16)[name = tensor("input_227_cast_fp16")]; - tensor input_229_axes_0 = const()[name = tensor("input_229_axes_0"), val = tensor([-1])]; - tensor module_layers_4_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_4_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(76919680)))]; - tensor module_layers_4_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_4_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(76921792)))]; - tensor input_229_cast_fp16 = layer_norm(axes = input_229_axes_0, beta = module_layers_4_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_4_norm_feed_forward1_weight_to_fp16, x = input_227_cast_fp16)[name = tensor("input_229_cast_fp16")]; - tensor module_layers_4_feed_forward1_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(76923904))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(80069696))), name = tensor("module_layers_4_feed_forward1_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_37_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_4_feed_forward1_linear1_weight_to_fp16_palettized, x = input_229_cast_fp16)[name = tensor("linear_37_cast_fp16")]; - tensor input_233_cast_fp16 = silu(x = linear_37_cast_fp16)[name = tensor("input_233_cast_fp16")]; - tensor module_layers_4_feed_forward1_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(80069888))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83215680))), name = tensor("module_layers_4_feed_forward1_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_38_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_4_feed_forward1_linear2_weight_to_fp16_palettized, x = input_233_cast_fp16)[name = tensor("linear_38_cast_fp16")]; - tensor var_884_to_fp16 = const()[name = tensor("op_884_to_fp16"), val = tensor(0x1p-1)]; - tensor var_885_cast_fp16 = mul(x = linear_38_cast_fp16, y = var_884_to_fp16)[name = tensor("op_885_cast_fp16")]; - tensor input_239_cast_fp16 = add(x = input_227_cast_fp16, y = var_885_cast_fp16)[name = tensor("input_239_cast_fp16")]; - tensor query_9_axes_0 = const()[name = tensor("query_9_axes_0"), val = tensor([-1])]; - tensor module_layers_4_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_4_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83215872)))]; - tensor module_layers_4_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_4_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83217984)))]; - tensor query_9_cast_fp16 = layer_norm(axes = query_9_axes_0, beta = module_layers_4_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_4_norm_self_att_weight_to_fp16, x = input_239_cast_fp16)[name = tensor("query_9_cast_fp16")]; - tensor module_layers_4_self_attn_linear_q_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83220096))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84006592))), name = tensor("module_layers_4_self_attn_linear_q_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_39_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_4_self_attn_linear_q_weight_to_fp16_palettized, x = query_9_cast_fp16)[name = tensor("linear_39_cast_fp16")]; - tensor var_901 = const()[name = tensor("op_901"), val = tensor([1, -1, 8, 128])]; - tensor q_25_cast_fp16 = reshape(shape = var_901, x = linear_39_cast_fp16)[name = tensor("q_25_cast_fp16")]; - tensor module_layers_4_self_attn_linear_k_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84006784))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84793280))), name = tensor("module_layers_4_self_attn_linear_k_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_40_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_4_self_attn_linear_k_weight_to_fp16_palettized, x = query_9_cast_fp16)[name = tensor("linear_40_cast_fp16")]; - tensor var_905 = const()[name = tensor("op_905"), val = tensor([1, -1, 8, 128])]; - tensor k_17_cast_fp16 = reshape(shape = var_905, x = linear_40_cast_fp16)[name = tensor("k_17_cast_fp16")]; - tensor module_layers_4_self_attn_linear_v_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84793472))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85579968))), name = tensor("module_layers_4_self_attn_linear_v_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_41_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_4_self_attn_linear_v_weight_to_fp16_palettized, x = query_9_cast_fp16)[name = tensor("linear_41_cast_fp16")]; - tensor var_909 = const()[name = tensor("op_909"), val = tensor([1, -1, 8, 128])]; - tensor v_9_cast_fp16 = reshape(shape = var_909, x = linear_41_cast_fp16)[name = tensor("v_9_cast_fp16")]; - tensor value_11_perm_0 = const()[name = tensor("value_11_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor module_layers_4_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_4_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85580160)))]; - tensor var_921_cast_fp16 = add(x = q_25_cast_fp16, y = module_layers_4_self_attn_pos_bias_u_to_fp16)[name = tensor("op_921_cast_fp16")]; - tensor module_layers_4_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_4_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85582272)))]; - tensor var_923_cast_fp16 = add(x = q_25_cast_fp16, y = module_layers_4_self_attn_pos_bias_v_to_fp16)[name = tensor("op_923_cast_fp16")]; - tensor q_with_bias_v_9_perm_0 = const()[name = tensor("q_with_bias_v_9_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor x_95_transpose_x_0 = const()[name = tensor("x_95_transpose_x_0"), val = tensor(false)]; - tensor x_95_transpose_y_0 = const()[name = tensor("x_95_transpose_y_0"), val = tensor(false)]; - tensor op_925_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85584384))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85872448))), name = tensor("op_925_to_fp16_palettized"), shape = tensor([1, 8, 128, 375])]; - tensor q_with_bias_v_9_cast_fp16 = transpose(perm = q_with_bias_v_9_perm_0, x = var_923_cast_fp16)[name = tensor("transpose_284")]; - tensor x_95_cast_fp16 = matmul(transpose_x = x_95_transpose_x_0, transpose_y = x_95_transpose_y_0, x = q_with_bias_v_9_cast_fp16, y = op_925_to_fp16_palettized)[name = tensor("x_95_cast_fp16")]; - tensor x_97_pad_0 = const()[name = tensor("x_97_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_97_mode_0 = const()[name = tensor("x_97_mode_0"), val = tensor("constant")]; - tensor const_54_to_fp16 = const()[name = tensor("const_54_to_fp16"), val = tensor(0x0p+0)]; - tensor x_97_cast_fp16 = pad(constant_val = const_54_to_fp16, mode = x_97_mode_0, pad = x_97_pad_0, x = x_95_cast_fp16)[name = tensor("x_97_cast_fp16")]; - tensor var_933 = const()[name = tensor("op_933"), val = tensor([1, 8, -1, 188])]; - tensor x_99_cast_fp16 = reshape(shape = var_933, x = x_97_cast_fp16)[name = tensor("x_99_cast_fp16")]; - tensor var_937_begin_0 = const()[name = tensor("op_937_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_937_end_0 = const()[name = tensor("op_937_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_937_end_mask_0 = const()[name = tensor("op_937_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_937_cast_fp16 = slice_by_index(begin = var_937_begin_0, end = var_937_end_0, end_mask = var_937_end_mask_0, x = x_99_cast_fp16)[name = tensor("op_937_cast_fp16")]; - tensor var_938 = const()[name = tensor("op_938"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_17_cast_fp16 = reshape(shape = var_938, x = var_937_cast_fp16)[name = tensor("matrix_bd_17_cast_fp16")]; - tensor matrix_ac_9_transpose_x_0 = const()[name = tensor("matrix_ac_9_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_9_transpose_y_0 = const()[name = tensor("matrix_ac_9_transpose_y_0"), val = tensor(false)]; - tensor transpose_104_perm_0 = const()[name = tensor("transpose_104_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_105_perm_0 = const()[name = tensor("transpose_105_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_105 = transpose(perm = transpose_105_perm_0, x = k_17_cast_fp16)[name = tensor("transpose_282")]; - tensor transpose_104 = transpose(perm = transpose_104_perm_0, x = var_921_cast_fp16)[name = tensor("transpose_283")]; - tensor matrix_ac_9_cast_fp16 = matmul(transpose_x = matrix_ac_9_transpose_x_0, transpose_y = matrix_ac_9_transpose_y_0, x = transpose_104, y = transpose_105)[name = tensor("matrix_ac_9_cast_fp16")]; - tensor matrix_bd_19_begin_0 = const()[name = tensor("matrix_bd_19_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_19_end_0 = const()[name = tensor("matrix_bd_19_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_19_end_mask_0 = const()[name = tensor("matrix_bd_19_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_19_cast_fp16 = slice_by_index(begin = matrix_bd_19_begin_0, end = matrix_bd_19_end_0, end_mask = matrix_bd_19_end_mask_0, x = matrix_bd_17_cast_fp16)[name = tensor("matrix_bd_19_cast_fp16")]; - tensor var_947_cast_fp16 = add(x = matrix_ac_9_cast_fp16, y = matrix_bd_19_cast_fp16)[name = tensor("op_947_cast_fp16")]; - tensor _inversed_scores_17_y_0_to_fp16 = const()[name = tensor("_inversed_scores_17_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; - tensor _inversed_scores_17_cast_fp16 = mul(x = var_947_cast_fp16, y = _inversed_scores_17_y_0_to_fp16)[name = tensor("_inversed_scores_17_cast_fp16")]; - tensor scores_19_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_17_cast_fp16, cond = mask_3)[name = tensor("scores_19_cast_fp16")]; - tensor var_953_cast_fp16 = softmax(axis = var_30, x = scores_19_cast_fp16)[name = tensor("op_953_cast_fp16")]; - tensor input_241_cast_fp16 = select(a = var_11_to_fp16, b = var_953_cast_fp16, cond = mask_3)[name = tensor("input_241_cast_fp16")]; - tensor x_101_transpose_x_0 = const()[name = tensor("x_101_transpose_x_0"), val = tensor(false)]; - tensor x_101_transpose_y_0 = const()[name = tensor("x_101_transpose_y_0"), val = tensor(false)]; - tensor value_11_cast_fp16 = transpose(perm = value_11_perm_0, x = v_9_cast_fp16)[name = tensor("transpose_281")]; - tensor x_101_cast_fp16 = matmul(transpose_x = x_101_transpose_x_0, transpose_y = x_101_transpose_y_0, x = input_241_cast_fp16, y = value_11_cast_fp16)[name = tensor("x_101_cast_fp16")]; - tensor var_957_perm_0 = const()[name = tensor("op_957_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_958 = const()[name = tensor("op_958"), val = tensor([1, -1, 1024])]; - tensor var_957_cast_fp16 = transpose(perm = var_957_perm_0, x = x_101_cast_fp16)[name = tensor("transpose_280")]; - tensor input_243_cast_fp16 = reshape(shape = var_958, x = var_957_cast_fp16)[name = tensor("input_243_cast_fp16")]; - tensor module_layers_4_self_attn_linear_out_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85872640))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86659136))), name = tensor("module_layers_4_self_attn_linear_out_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_43_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_4_self_attn_linear_out_weight_to_fp16_palettized, x = input_243_cast_fp16)[name = tensor("linear_43_cast_fp16")]; - tensor input_247_cast_fp16 = add(x = input_239_cast_fp16, y = linear_43_cast_fp16)[name = tensor("input_247_cast_fp16")]; - tensor x_105_axes_0 = const()[name = tensor("x_105_axes_0"), val = tensor([-1])]; - tensor module_layers_4_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_4_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86659328)))]; - tensor module_layers_4_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_4_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86661440)))]; - tensor x_105_cast_fp16 = layer_norm(axes = x_105_axes_0, beta = module_layers_4_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_4_norm_conv_weight_to_fp16, x = input_247_cast_fp16)[name = tensor("x_105_cast_fp16")]; - tensor input_249_perm_0 = const()[name = tensor("input_249_perm_0"), val = tensor([0, 2, 1])]; - tensor input_251_pad_type_0 = const()[name = tensor("input_251_pad_type_0"), val = tensor("valid")]; - tensor input_251_strides_0 = const()[name = tensor("input_251_strides_0"), val = tensor([1])]; - tensor input_251_pad_0 = const()[name = tensor("input_251_pad_0"), val = tensor([0, 0])]; - tensor input_251_dilations_0 = const()[name = tensor("input_251_dilations_0"), val = tensor([1])]; - tensor input_251_groups_0 = const()[name = tensor("input_251_groups_0"), val = tensor(1)]; - tensor module_layers_4_conv_pointwise_conv1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86663552))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(88236480))), name = tensor("module_layers_4_conv_pointwise_conv1_weight_to_fp16_palettized"), shape = tensor([2048, 1024, 1])]; - tensor input_249_cast_fp16 = transpose(perm = input_249_perm_0, x = x_105_cast_fp16)[name = tensor("transpose_279")]; - tensor input_251_cast_fp16 = conv(dilations = input_251_dilations_0, groups = input_251_groups_0, pad = input_251_pad_0, pad_type = input_251_pad_type_0, strides = input_251_strides_0, weight = module_layers_4_conv_pointwise_conv1_weight_to_fp16_palettized, x = input_249_cast_fp16)[name = tensor("input_251_cast_fp16")]; - tensor x_107_split_num_splits_0 = const()[name = tensor("x_107_split_num_splits_0"), val = tensor(2)]; - tensor x_107_split_axis_0 = const()[name = tensor("x_107_split_axis_0"), val = tensor(1)]; - tensor x_107_split_cast_fp16_0, tensor x_107_split_cast_fp16_1 = split(axis = x_107_split_axis_0, num_splits = x_107_split_num_splits_0, x = input_251_cast_fp16)[name = tensor("x_107_split_cast_fp16")]; - tensor x_107_split_1_sigmoid_cast_fp16 = sigmoid(x = x_107_split_cast_fp16_1)[name = tensor("x_107_split_1_sigmoid_cast_fp16")]; - tensor x_107_cast_fp16 = mul(x = x_107_split_cast_fp16_0, y = x_107_split_1_sigmoid_cast_fp16)[name = tensor("x_107_cast_fp16")]; - tensor input_253_cast_fp16 = select(a = var_11_to_fp16, b = x_107_cast_fp16, cond = var_328)[name = tensor("input_253_cast_fp16")]; - tensor input_255_pad_0 = const()[name = tensor("input_255_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_255_mode_0 = const()[name = tensor("input_255_mode_0"), val = tensor("constant")]; - tensor const_57_to_fp16 = const()[name = tensor("const_57_to_fp16"), val = tensor(0x0p+0)]; - tensor input_255_cast_fp16 = pad(constant_val = const_57_to_fp16, mode = input_255_mode_0, pad = input_255_pad_0, x = input_253_cast_fp16)[name = tensor("input_255_cast_fp16")]; - tensor input_257_pad_type_0 = const()[name = tensor("input_257_pad_type_0"), val = tensor("valid")]; - tensor input_257_groups_0 = const()[name = tensor("input_257_groups_0"), val = tensor(1024)]; - tensor input_257_strides_0 = const()[name = tensor("input_257_strides_0"), val = tensor([1])]; - tensor input_257_pad_0 = const()[name = tensor("input_257_pad_0"), val = tensor([0, 0])]; - tensor input_257_dilations_0 = const()[name = tensor("input_257_dilations_0"), val = tensor([1])]; - tensor const_256_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(88236672))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(88243648))), name = tensor("const_256_to_fp16_palettized"), shape = tensor([1024, 1, 9])]; - tensor const_257_to_fp16 = const()[name = tensor("const_257_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(88243840)))]; - tensor input_259_cast_fp16 = conv(bias = const_257_to_fp16, dilations = input_257_dilations_0, groups = input_257_groups_0, pad = input_257_pad_0, pad_type = input_257_pad_type_0, strides = input_257_strides_0, weight = const_256_to_fp16_palettized, x = input_255_cast_fp16)[name = tensor("input_259_cast_fp16")]; - tensor input_261_cast_fp16 = silu(x = input_259_cast_fp16)[name = tensor("input_261_cast_fp16")]; - tensor x_109_pad_type_0 = const()[name = tensor("x_109_pad_type_0"), val = tensor("valid")]; - tensor x_109_strides_0 = const()[name = tensor("x_109_strides_0"), val = tensor([1])]; - tensor x_109_pad_0 = const()[name = tensor("x_109_pad_0"), val = tensor([0, 0])]; - tensor x_109_dilations_0 = const()[name = tensor("x_109_dilations_0"), val = tensor([1])]; - tensor x_109_groups_0 = const()[name = tensor("x_109_groups_0"), val = tensor(1)]; - tensor module_layers_4_conv_pointwise_conv2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(88245952))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(89032448))), name = tensor("module_layers_4_conv_pointwise_conv2_weight_to_fp16_palettized"), shape = tensor([1024, 1024, 1])]; - tensor x_109_cast_fp16 = conv(dilations = x_109_dilations_0, groups = x_109_groups_0, pad = x_109_pad_0, pad_type = x_109_pad_type_0, strides = x_109_strides_0, weight = module_layers_4_conv_pointwise_conv2_weight_to_fp16_palettized, x = input_261_cast_fp16)[name = tensor("x_109_cast_fp16")]; - tensor input_263_perm_0 = const()[name = tensor("input_263_perm_0"), val = tensor([0, 2, 1])]; - tensor input_263_cast_fp16 = transpose(perm = input_263_perm_0, x = x_109_cast_fp16)[name = tensor("transpose_278")]; - tensor input_265_cast_fp16 = add(x = input_247_cast_fp16, y = input_263_cast_fp16)[name = tensor("input_265_cast_fp16")]; - tensor input_267_axes_0 = const()[name = tensor("input_267_axes_0"), val = tensor([-1])]; - tensor module_layers_4_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_4_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(89032640)))]; - tensor module_layers_4_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_4_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(89034752)))]; - tensor input_267_cast_fp16 = layer_norm(axes = input_267_axes_0, beta = module_layers_4_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_4_norm_feed_forward2_weight_to_fp16, x = input_265_cast_fp16)[name = tensor("input_267_cast_fp16")]; - tensor module_layers_4_feed_forward2_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(89036864))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92182656))), name = tensor("module_layers_4_feed_forward2_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_44_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_4_feed_forward2_linear1_weight_to_fp16_palettized, x = input_267_cast_fp16)[name = tensor("linear_44_cast_fp16")]; - tensor input_271_cast_fp16 = silu(x = linear_44_cast_fp16)[name = tensor("input_271_cast_fp16")]; - tensor module_layers_4_feed_forward2_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92182848))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95328640))), name = tensor("module_layers_4_feed_forward2_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_45_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_4_feed_forward2_linear2_weight_to_fp16_palettized, x = input_271_cast_fp16)[name = tensor("linear_45_cast_fp16")]; - tensor var_1018_to_fp16 = const()[name = tensor("op_1018_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1019_cast_fp16 = mul(x = linear_45_cast_fp16, y = var_1018_to_fp16)[name = tensor("op_1019_cast_fp16")]; - tensor input_277_cast_fp16 = add(x = input_265_cast_fp16, y = var_1019_cast_fp16)[name = tensor("input_277_cast_fp16")]; - tensor input_279_axes_0 = const()[name = tensor("input_279_axes_0"), val = tensor([-1])]; - tensor module_layers_4_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_4_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95328832)))]; - tensor module_layers_4_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_4_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95330944)))]; - tensor input_279_cast_fp16 = layer_norm(axes = input_279_axes_0, beta = module_layers_4_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_4_norm_out_weight_to_fp16, x = input_277_cast_fp16)[name = tensor("input_279_cast_fp16")]; - tensor input_281_axes_0 = const()[name = tensor("input_281_axes_0"), val = tensor([-1])]; - tensor module_layers_5_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_5_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95333056)))]; - tensor module_layers_5_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_5_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95335168)))]; - tensor input_281_cast_fp16 = layer_norm(axes = input_281_axes_0, beta = module_layers_5_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_5_norm_feed_forward1_weight_to_fp16, x = input_279_cast_fp16)[name = tensor("input_281_cast_fp16")]; - tensor module_layers_5_feed_forward1_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95337280))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98483072))), name = tensor("module_layers_5_feed_forward1_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_46_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_5_feed_forward1_linear1_weight_to_fp16_palettized, x = input_281_cast_fp16)[name = tensor("linear_46_cast_fp16")]; - tensor input_285_cast_fp16 = silu(x = linear_46_cast_fp16)[name = tensor("input_285_cast_fp16")]; - tensor module_layers_5_feed_forward1_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98483264))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(101629056))), name = tensor("module_layers_5_feed_forward1_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_47_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_5_feed_forward1_linear2_weight_to_fp16_palettized, x = input_285_cast_fp16)[name = tensor("linear_47_cast_fp16")]; - tensor var_1047_to_fp16 = const()[name = tensor("op_1047_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1048_cast_fp16 = mul(x = linear_47_cast_fp16, y = var_1047_to_fp16)[name = tensor("op_1048_cast_fp16")]; - tensor input_291_cast_fp16 = add(x = input_279_cast_fp16, y = var_1048_cast_fp16)[name = tensor("input_291_cast_fp16")]; - tensor query_11_axes_0 = const()[name = tensor("query_11_axes_0"), val = tensor([-1])]; - tensor module_layers_5_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_5_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(101629248)))]; - tensor module_layers_5_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_5_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(101631360)))]; - tensor query_11_cast_fp16 = layer_norm(axes = query_11_axes_0, beta = module_layers_5_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_5_norm_self_att_weight_to_fp16, x = input_291_cast_fp16)[name = tensor("query_11_cast_fp16")]; - tensor module_layers_5_self_attn_linear_q_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(101633472))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102419968))), name = tensor("module_layers_5_self_attn_linear_q_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_48_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_5_self_attn_linear_q_weight_to_fp16_palettized, x = query_11_cast_fp16)[name = tensor("linear_48_cast_fp16")]; - tensor var_1064 = const()[name = tensor("op_1064"), val = tensor([1, -1, 8, 128])]; - tensor q_31_cast_fp16 = reshape(shape = var_1064, x = linear_48_cast_fp16)[name = tensor("q_31_cast_fp16")]; - tensor module_layers_5_self_attn_linear_k_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102420160))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(103206656))), name = tensor("module_layers_5_self_attn_linear_k_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_49_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_5_self_attn_linear_k_weight_to_fp16_palettized, x = query_11_cast_fp16)[name = tensor("linear_49_cast_fp16")]; - tensor var_1068 = const()[name = tensor("op_1068"), val = tensor([1, -1, 8, 128])]; - tensor k_21_cast_fp16 = reshape(shape = var_1068, x = linear_49_cast_fp16)[name = tensor("k_21_cast_fp16")]; - tensor module_layers_5_self_attn_linear_v_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(103206848))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(103993344))), name = tensor("module_layers_5_self_attn_linear_v_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_50_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_5_self_attn_linear_v_weight_to_fp16_palettized, x = query_11_cast_fp16)[name = tensor("linear_50_cast_fp16")]; - tensor var_1072 = const()[name = tensor("op_1072"), val = tensor([1, -1, 8, 128])]; - tensor v_11_cast_fp16 = reshape(shape = var_1072, x = linear_50_cast_fp16)[name = tensor("v_11_cast_fp16")]; - tensor value_13_perm_0 = const()[name = tensor("value_13_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor module_layers_5_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_5_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(103993536)))]; - tensor var_1084_cast_fp16 = add(x = q_31_cast_fp16, y = module_layers_5_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1084_cast_fp16")]; - tensor module_layers_5_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_5_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(103995648)))]; - tensor var_1086_cast_fp16 = add(x = q_31_cast_fp16, y = module_layers_5_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1086_cast_fp16")]; - tensor q_with_bias_v_11_perm_0 = const()[name = tensor("q_with_bias_v_11_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor x_117_transpose_x_0 = const()[name = tensor("x_117_transpose_x_0"), val = tensor(false)]; - tensor x_117_transpose_y_0 = const()[name = tensor("x_117_transpose_y_0"), val = tensor(false)]; - tensor op_1088_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(103997760))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(104285824))), name = tensor("op_1088_to_fp16_palettized"), shape = tensor([1, 8, 128, 375])]; - tensor q_with_bias_v_11_cast_fp16 = transpose(perm = q_with_bias_v_11_perm_0, x = var_1086_cast_fp16)[name = tensor("transpose_277")]; - tensor x_117_cast_fp16 = matmul(transpose_x = x_117_transpose_x_0, transpose_y = x_117_transpose_y_0, x = q_with_bias_v_11_cast_fp16, y = op_1088_to_fp16_palettized)[name = tensor("x_117_cast_fp16")]; - tensor x_119_pad_0 = const()[name = tensor("x_119_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_119_mode_0 = const()[name = tensor("x_119_mode_0"), val = tensor("constant")]; - tensor const_64_to_fp16 = const()[name = tensor("const_64_to_fp16"), val = tensor(0x0p+0)]; - tensor x_119_cast_fp16 = pad(constant_val = const_64_to_fp16, mode = x_119_mode_0, pad = x_119_pad_0, x = x_117_cast_fp16)[name = tensor("x_119_cast_fp16")]; - tensor var_1096 = const()[name = tensor("op_1096"), val = tensor([1, 8, -1, 188])]; - tensor x_121_cast_fp16 = reshape(shape = var_1096, x = x_119_cast_fp16)[name = tensor("x_121_cast_fp16")]; - tensor var_1100_begin_0 = const()[name = tensor("op_1100_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_1100_end_0 = const()[name = tensor("op_1100_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_1100_end_mask_0 = const()[name = tensor("op_1100_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_1100_cast_fp16 = slice_by_index(begin = var_1100_begin_0, end = var_1100_end_0, end_mask = var_1100_end_mask_0, x = x_121_cast_fp16)[name = tensor("op_1100_cast_fp16")]; - tensor var_1101 = const()[name = tensor("op_1101"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_21_cast_fp16 = reshape(shape = var_1101, x = var_1100_cast_fp16)[name = tensor("matrix_bd_21_cast_fp16")]; - tensor matrix_ac_11_transpose_x_0 = const()[name = tensor("matrix_ac_11_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_11_transpose_y_0 = const()[name = tensor("matrix_ac_11_transpose_y_0"), val = tensor(false)]; - tensor transpose_106_perm_0 = const()[name = tensor("transpose_106_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_107_perm_0 = const()[name = tensor("transpose_107_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_107 = transpose(perm = transpose_107_perm_0, x = k_21_cast_fp16)[name = tensor("transpose_275")]; - tensor transpose_106 = transpose(perm = transpose_106_perm_0, x = var_1084_cast_fp16)[name = tensor("transpose_276")]; - tensor matrix_ac_11_cast_fp16 = matmul(transpose_x = matrix_ac_11_transpose_x_0, transpose_y = matrix_ac_11_transpose_y_0, x = transpose_106, y = transpose_107)[name = tensor("matrix_ac_11_cast_fp16")]; - tensor matrix_bd_23_begin_0 = const()[name = tensor("matrix_bd_23_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_23_end_0 = const()[name = tensor("matrix_bd_23_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_23_end_mask_0 = const()[name = tensor("matrix_bd_23_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_23_cast_fp16 = slice_by_index(begin = matrix_bd_23_begin_0, end = matrix_bd_23_end_0, end_mask = matrix_bd_23_end_mask_0, x = matrix_bd_21_cast_fp16)[name = tensor("matrix_bd_23_cast_fp16")]; - tensor var_1110_cast_fp16 = add(x = matrix_ac_11_cast_fp16, y = matrix_bd_23_cast_fp16)[name = tensor("op_1110_cast_fp16")]; - tensor _inversed_scores_21_y_0_to_fp16 = const()[name = tensor("_inversed_scores_21_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; - tensor _inversed_scores_21_cast_fp16 = mul(x = var_1110_cast_fp16, y = _inversed_scores_21_y_0_to_fp16)[name = tensor("_inversed_scores_21_cast_fp16")]; - tensor scores_23_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_21_cast_fp16, cond = mask_3)[name = tensor("scores_23_cast_fp16")]; - tensor var_1116_cast_fp16 = softmax(axis = var_30, x = scores_23_cast_fp16)[name = tensor("op_1116_cast_fp16")]; - tensor input_293_cast_fp16 = select(a = var_11_to_fp16, b = var_1116_cast_fp16, cond = mask_3)[name = tensor("input_293_cast_fp16")]; - tensor x_123_transpose_x_0 = const()[name = tensor("x_123_transpose_x_0"), val = tensor(false)]; - tensor x_123_transpose_y_0 = const()[name = tensor("x_123_transpose_y_0"), val = tensor(false)]; - tensor value_13_cast_fp16 = transpose(perm = value_13_perm_0, x = v_11_cast_fp16)[name = tensor("transpose_274")]; - tensor x_123_cast_fp16 = matmul(transpose_x = x_123_transpose_x_0, transpose_y = x_123_transpose_y_0, x = input_293_cast_fp16, y = value_13_cast_fp16)[name = tensor("x_123_cast_fp16")]; - tensor var_1120_perm_0 = const()[name = tensor("op_1120_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_1121 = const()[name = tensor("op_1121"), val = tensor([1, -1, 1024])]; - tensor var_1120_cast_fp16 = transpose(perm = var_1120_perm_0, x = x_123_cast_fp16)[name = tensor("transpose_273")]; - tensor input_295_cast_fp16 = reshape(shape = var_1121, x = var_1120_cast_fp16)[name = tensor("input_295_cast_fp16")]; - tensor module_layers_5_self_attn_linear_out_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(104286016))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(105072512))), name = tensor("module_layers_5_self_attn_linear_out_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_52_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_5_self_attn_linear_out_weight_to_fp16_palettized, x = input_295_cast_fp16)[name = tensor("linear_52_cast_fp16")]; - tensor input_299_cast_fp16 = add(x = input_291_cast_fp16, y = linear_52_cast_fp16)[name = tensor("input_299_cast_fp16")]; - tensor x_127_axes_0 = const()[name = tensor("x_127_axes_0"), val = tensor([-1])]; - tensor module_layers_5_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_5_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(105072704)))]; - tensor module_layers_5_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_5_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(105074816)))]; - tensor x_127_cast_fp16 = layer_norm(axes = x_127_axes_0, beta = module_layers_5_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_5_norm_conv_weight_to_fp16, x = input_299_cast_fp16)[name = tensor("x_127_cast_fp16")]; - tensor input_301_perm_0 = const()[name = tensor("input_301_perm_0"), val = tensor([0, 2, 1])]; - tensor input_303_pad_type_0 = const()[name = tensor("input_303_pad_type_0"), val = tensor("valid")]; - tensor input_303_strides_0 = const()[name = tensor("input_303_strides_0"), val = tensor([1])]; - tensor input_303_pad_0 = const()[name = tensor("input_303_pad_0"), val = tensor([0, 0])]; - tensor input_303_dilations_0 = const()[name = tensor("input_303_dilations_0"), val = tensor([1])]; - tensor input_303_groups_0 = const()[name = tensor("input_303_groups_0"), val = tensor(1)]; - tensor module_layers_5_conv_pointwise_conv1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(105076928))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(106649856))), name = tensor("module_layers_5_conv_pointwise_conv1_weight_to_fp16_palettized"), shape = tensor([2048, 1024, 1])]; - tensor input_301_cast_fp16 = transpose(perm = input_301_perm_0, x = x_127_cast_fp16)[name = tensor("transpose_272")]; - tensor input_303_cast_fp16 = conv(dilations = input_303_dilations_0, groups = input_303_groups_0, pad = input_303_pad_0, pad_type = input_303_pad_type_0, strides = input_303_strides_0, weight = module_layers_5_conv_pointwise_conv1_weight_to_fp16_palettized, x = input_301_cast_fp16)[name = tensor("input_303_cast_fp16")]; - tensor x_129_split_num_splits_0 = const()[name = tensor("x_129_split_num_splits_0"), val = tensor(2)]; - tensor x_129_split_axis_0 = const()[name = tensor("x_129_split_axis_0"), val = tensor(1)]; - tensor x_129_split_cast_fp16_0, tensor x_129_split_cast_fp16_1 = split(axis = x_129_split_axis_0, num_splits = x_129_split_num_splits_0, x = input_303_cast_fp16)[name = tensor("x_129_split_cast_fp16")]; - tensor x_129_split_1_sigmoid_cast_fp16 = sigmoid(x = x_129_split_cast_fp16_1)[name = tensor("x_129_split_1_sigmoid_cast_fp16")]; - tensor x_129_cast_fp16 = mul(x = x_129_split_cast_fp16_0, y = x_129_split_1_sigmoid_cast_fp16)[name = tensor("x_129_cast_fp16")]; - tensor input_305_cast_fp16 = select(a = var_11_to_fp16, b = x_129_cast_fp16, cond = var_328)[name = tensor("input_305_cast_fp16")]; - tensor input_307_pad_0 = const()[name = tensor("input_307_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_307_mode_0 = const()[name = tensor("input_307_mode_0"), val = tensor("constant")]; - tensor const_67_to_fp16 = const()[name = tensor("const_67_to_fp16"), val = tensor(0x0p+0)]; - tensor input_307_cast_fp16 = pad(constant_val = const_67_to_fp16, mode = input_307_mode_0, pad = input_307_pad_0, x = input_305_cast_fp16)[name = tensor("input_307_cast_fp16")]; - tensor input_309_pad_type_0 = const()[name = tensor("input_309_pad_type_0"), val = tensor("valid")]; - tensor input_309_groups_0 = const()[name = tensor("input_309_groups_0"), val = tensor(1024)]; - tensor input_309_strides_0 = const()[name = tensor("input_309_strides_0"), val = tensor([1])]; - tensor input_309_pad_0 = const()[name = tensor("input_309_pad_0"), val = tensor([0, 0])]; - tensor input_309_dilations_0 = const()[name = tensor("input_309_dilations_0"), val = tensor([1])]; - tensor const_258_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(106650048))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(106657024))), name = tensor("const_258_to_fp16_palettized"), shape = tensor([1024, 1, 9])]; - tensor const_259_to_fp16 = const()[name = tensor("const_259_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(106657216)))]; - tensor input_311_cast_fp16 = conv(bias = const_259_to_fp16, dilations = input_309_dilations_0, groups = input_309_groups_0, pad = input_309_pad_0, pad_type = input_309_pad_type_0, strides = input_309_strides_0, weight = const_258_to_fp16_palettized, x = input_307_cast_fp16)[name = tensor("input_311_cast_fp16")]; - tensor input_313_cast_fp16 = silu(x = input_311_cast_fp16)[name = tensor("input_313_cast_fp16")]; - tensor x_131_pad_type_0 = const()[name = tensor("x_131_pad_type_0"), val = tensor("valid")]; - tensor x_131_strides_0 = const()[name = tensor("x_131_strides_0"), val = tensor([1])]; - tensor x_131_pad_0 = const()[name = tensor("x_131_pad_0"), val = tensor([0, 0])]; - tensor x_131_dilations_0 = const()[name = tensor("x_131_dilations_0"), val = tensor([1])]; - tensor x_131_groups_0 = const()[name = tensor("x_131_groups_0"), val = tensor(1)]; - tensor module_layers_5_conv_pointwise_conv2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(106659328))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(107445824))), name = tensor("module_layers_5_conv_pointwise_conv2_weight_to_fp16_palettized"), shape = tensor([1024, 1024, 1])]; - tensor x_131_cast_fp16 = conv(dilations = x_131_dilations_0, groups = x_131_groups_0, pad = x_131_pad_0, pad_type = x_131_pad_type_0, strides = x_131_strides_0, weight = module_layers_5_conv_pointwise_conv2_weight_to_fp16_palettized, x = input_313_cast_fp16)[name = tensor("x_131_cast_fp16")]; - tensor input_315_perm_0 = const()[name = tensor("input_315_perm_0"), val = tensor([0, 2, 1])]; - tensor input_315_cast_fp16 = transpose(perm = input_315_perm_0, x = x_131_cast_fp16)[name = tensor("transpose_271")]; - tensor input_317_cast_fp16 = add(x = input_299_cast_fp16, y = input_315_cast_fp16)[name = tensor("input_317_cast_fp16")]; - tensor input_319_axes_0 = const()[name = tensor("input_319_axes_0"), val = tensor([-1])]; - tensor module_layers_5_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_5_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(107446016)))]; - tensor module_layers_5_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_5_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(107448128)))]; - tensor input_319_cast_fp16 = layer_norm(axes = input_319_axes_0, beta = module_layers_5_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_5_norm_feed_forward2_weight_to_fp16, x = input_317_cast_fp16)[name = tensor("input_319_cast_fp16")]; - tensor module_layers_5_feed_forward2_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(107450240))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110596032))), name = tensor("module_layers_5_feed_forward2_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_53_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_5_feed_forward2_linear1_weight_to_fp16_palettized, x = input_319_cast_fp16)[name = tensor("linear_53_cast_fp16")]; - tensor input_323_cast_fp16 = silu(x = linear_53_cast_fp16)[name = tensor("input_323_cast_fp16")]; - tensor module_layers_5_feed_forward2_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110596224))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(113742016))), name = tensor("module_layers_5_feed_forward2_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_54_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_5_feed_forward2_linear2_weight_to_fp16_palettized, x = input_323_cast_fp16)[name = tensor("linear_54_cast_fp16")]; - tensor var_1181_to_fp16 = const()[name = tensor("op_1181_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1182_cast_fp16 = mul(x = linear_54_cast_fp16, y = var_1181_to_fp16)[name = tensor("op_1182_cast_fp16")]; - tensor input_329_cast_fp16 = add(x = input_317_cast_fp16, y = var_1182_cast_fp16)[name = tensor("input_329_cast_fp16")]; - tensor input_331_axes_0 = const()[name = tensor("input_331_axes_0"), val = tensor([-1])]; - tensor module_layers_5_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_5_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(113742208)))]; - tensor module_layers_5_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_5_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(113744320)))]; - tensor input_331_cast_fp16 = layer_norm(axes = input_331_axes_0, beta = module_layers_5_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_5_norm_out_weight_to_fp16, x = input_329_cast_fp16)[name = tensor("input_331_cast_fp16")]; - tensor input_333_axes_0 = const()[name = tensor("input_333_axes_0"), val = tensor([-1])]; - tensor module_layers_6_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_6_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(113746432)))]; - tensor module_layers_6_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_6_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(113748544)))]; - tensor input_333_cast_fp16 = layer_norm(axes = input_333_axes_0, beta = module_layers_6_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_6_norm_feed_forward1_weight_to_fp16, x = input_331_cast_fp16)[name = tensor("input_333_cast_fp16")]; - tensor module_layers_6_feed_forward1_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(113750656))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(116896448))), name = tensor("module_layers_6_feed_forward1_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_55_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_6_feed_forward1_linear1_weight_to_fp16_palettized, x = input_333_cast_fp16)[name = tensor("linear_55_cast_fp16")]; - tensor input_337_cast_fp16 = silu(x = linear_55_cast_fp16)[name = tensor("input_337_cast_fp16")]; - tensor module_layers_6_feed_forward1_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(116896640))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(120042432))), name = tensor("module_layers_6_feed_forward1_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_56_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_6_feed_forward1_linear2_weight_to_fp16_palettized, x = input_337_cast_fp16)[name = tensor("linear_56_cast_fp16")]; - tensor var_1210_to_fp16 = const()[name = tensor("op_1210_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1211_cast_fp16 = mul(x = linear_56_cast_fp16, y = var_1210_to_fp16)[name = tensor("op_1211_cast_fp16")]; - tensor input_343_cast_fp16 = add(x = input_331_cast_fp16, y = var_1211_cast_fp16)[name = tensor("input_343_cast_fp16")]; - tensor query_13_axes_0 = const()[name = tensor("query_13_axes_0"), val = tensor([-1])]; - tensor module_layers_6_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_6_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(120042624)))]; - tensor module_layers_6_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_6_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(120044736)))]; - tensor query_13_cast_fp16 = layer_norm(axes = query_13_axes_0, beta = module_layers_6_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_6_norm_self_att_weight_to_fp16, x = input_343_cast_fp16)[name = tensor("query_13_cast_fp16")]; - tensor module_layers_6_self_attn_linear_q_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(120046848))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(120833344))), name = tensor("module_layers_6_self_attn_linear_q_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_57_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_6_self_attn_linear_q_weight_to_fp16_palettized, x = query_13_cast_fp16)[name = tensor("linear_57_cast_fp16")]; - tensor var_1227 = const()[name = tensor("op_1227"), val = tensor([1, -1, 8, 128])]; - tensor q_37_cast_fp16 = reshape(shape = var_1227, x = linear_57_cast_fp16)[name = tensor("q_37_cast_fp16")]; - tensor module_layers_6_self_attn_linear_k_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(120833536))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121620032))), name = tensor("module_layers_6_self_attn_linear_k_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_58_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_6_self_attn_linear_k_weight_to_fp16_palettized, x = query_13_cast_fp16)[name = tensor("linear_58_cast_fp16")]; - tensor var_1231 = const()[name = tensor("op_1231"), val = tensor([1, -1, 8, 128])]; - tensor k_25_cast_fp16 = reshape(shape = var_1231, x = linear_58_cast_fp16)[name = tensor("k_25_cast_fp16")]; - tensor module_layers_6_self_attn_linear_v_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121620224))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(122406720))), name = tensor("module_layers_6_self_attn_linear_v_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_59_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_6_self_attn_linear_v_weight_to_fp16_palettized, x = query_13_cast_fp16)[name = tensor("linear_59_cast_fp16")]; - tensor var_1235 = const()[name = tensor("op_1235"), val = tensor([1, -1, 8, 128])]; - tensor v_13_cast_fp16 = reshape(shape = var_1235, x = linear_59_cast_fp16)[name = tensor("v_13_cast_fp16")]; - tensor value_15_perm_0 = const()[name = tensor("value_15_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor module_layers_6_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_6_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(122406912)))]; - tensor var_1247_cast_fp16 = add(x = q_37_cast_fp16, y = module_layers_6_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1247_cast_fp16")]; - tensor module_layers_6_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_6_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(122409024)))]; - tensor var_1249_cast_fp16 = add(x = q_37_cast_fp16, y = module_layers_6_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1249_cast_fp16")]; - tensor q_with_bias_v_13_perm_0 = const()[name = tensor("q_with_bias_v_13_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor x_139_transpose_x_0 = const()[name = tensor("x_139_transpose_x_0"), val = tensor(false)]; - tensor x_139_transpose_y_0 = const()[name = tensor("x_139_transpose_y_0"), val = tensor(false)]; - tensor op_1251_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(122411136))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(122699200))), name = tensor("op_1251_to_fp16_palettized"), shape = tensor([1, 8, 128, 375])]; - tensor q_with_bias_v_13_cast_fp16 = transpose(perm = q_with_bias_v_13_perm_0, x = var_1249_cast_fp16)[name = tensor("transpose_270")]; - tensor x_139_cast_fp16 = matmul(transpose_x = x_139_transpose_x_0, transpose_y = x_139_transpose_y_0, x = q_with_bias_v_13_cast_fp16, y = op_1251_to_fp16_palettized)[name = tensor("x_139_cast_fp16")]; - tensor x_141_pad_0 = const()[name = tensor("x_141_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_141_mode_0 = const()[name = tensor("x_141_mode_0"), val = tensor("constant")]; - tensor const_74_to_fp16 = const()[name = tensor("const_74_to_fp16"), val = tensor(0x0p+0)]; - tensor x_141_cast_fp16 = pad(constant_val = const_74_to_fp16, mode = x_141_mode_0, pad = x_141_pad_0, x = x_139_cast_fp16)[name = tensor("x_141_cast_fp16")]; - tensor var_1259 = const()[name = tensor("op_1259"), val = tensor([1, 8, -1, 188])]; - tensor x_143_cast_fp16 = reshape(shape = var_1259, x = x_141_cast_fp16)[name = tensor("x_143_cast_fp16")]; - tensor var_1263_begin_0 = const()[name = tensor("op_1263_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_1263_end_0 = const()[name = tensor("op_1263_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_1263_end_mask_0 = const()[name = tensor("op_1263_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_1263_cast_fp16 = slice_by_index(begin = var_1263_begin_0, end = var_1263_end_0, end_mask = var_1263_end_mask_0, x = x_143_cast_fp16)[name = tensor("op_1263_cast_fp16")]; - tensor var_1264 = const()[name = tensor("op_1264"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_25_cast_fp16 = reshape(shape = var_1264, x = var_1263_cast_fp16)[name = tensor("matrix_bd_25_cast_fp16")]; - tensor matrix_ac_13_transpose_x_0 = const()[name = tensor("matrix_ac_13_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_13_transpose_y_0 = const()[name = tensor("matrix_ac_13_transpose_y_0"), val = tensor(false)]; - tensor transpose_108_perm_0 = const()[name = tensor("transpose_108_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_109_perm_0 = const()[name = tensor("transpose_109_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_109 = transpose(perm = transpose_109_perm_0, x = k_25_cast_fp16)[name = tensor("transpose_268")]; - tensor transpose_108 = transpose(perm = transpose_108_perm_0, x = var_1247_cast_fp16)[name = tensor("transpose_269")]; - tensor matrix_ac_13_cast_fp16 = matmul(transpose_x = matrix_ac_13_transpose_x_0, transpose_y = matrix_ac_13_transpose_y_0, x = transpose_108, y = transpose_109)[name = tensor("matrix_ac_13_cast_fp16")]; - tensor matrix_bd_27_begin_0 = const()[name = tensor("matrix_bd_27_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_27_end_0 = const()[name = tensor("matrix_bd_27_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_27_end_mask_0 = const()[name = tensor("matrix_bd_27_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_27_cast_fp16 = slice_by_index(begin = matrix_bd_27_begin_0, end = matrix_bd_27_end_0, end_mask = matrix_bd_27_end_mask_0, x = matrix_bd_25_cast_fp16)[name = tensor("matrix_bd_27_cast_fp16")]; - tensor var_1273_cast_fp16 = add(x = matrix_ac_13_cast_fp16, y = matrix_bd_27_cast_fp16)[name = tensor("op_1273_cast_fp16")]; - tensor _inversed_scores_25_y_0_to_fp16 = const()[name = tensor("_inversed_scores_25_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; - tensor _inversed_scores_25_cast_fp16 = mul(x = var_1273_cast_fp16, y = _inversed_scores_25_y_0_to_fp16)[name = tensor("_inversed_scores_25_cast_fp16")]; - tensor scores_27_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_25_cast_fp16, cond = mask_3)[name = tensor("scores_27_cast_fp16")]; - tensor var_1279_cast_fp16 = softmax(axis = var_30, x = scores_27_cast_fp16)[name = tensor("op_1279_cast_fp16")]; - tensor input_345_cast_fp16 = select(a = var_11_to_fp16, b = var_1279_cast_fp16, cond = mask_3)[name = tensor("input_345_cast_fp16")]; - tensor x_145_transpose_x_0 = const()[name = tensor("x_145_transpose_x_0"), val = tensor(false)]; - tensor x_145_transpose_y_0 = const()[name = tensor("x_145_transpose_y_0"), val = tensor(false)]; - tensor value_15_cast_fp16 = transpose(perm = value_15_perm_0, x = v_13_cast_fp16)[name = tensor("transpose_267")]; - tensor x_145_cast_fp16 = matmul(transpose_x = x_145_transpose_x_0, transpose_y = x_145_transpose_y_0, x = input_345_cast_fp16, y = value_15_cast_fp16)[name = tensor("x_145_cast_fp16")]; - tensor var_1283_perm_0 = const()[name = tensor("op_1283_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_1284 = const()[name = tensor("op_1284"), val = tensor([1, -1, 1024])]; - tensor var_1283_cast_fp16 = transpose(perm = var_1283_perm_0, x = x_145_cast_fp16)[name = tensor("transpose_266")]; - tensor input_347_cast_fp16 = reshape(shape = var_1284, x = var_1283_cast_fp16)[name = tensor("input_347_cast_fp16")]; - tensor module_layers_6_self_attn_linear_out_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(122699392))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123485888))), name = tensor("module_layers_6_self_attn_linear_out_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_61_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_6_self_attn_linear_out_weight_to_fp16_palettized, x = input_347_cast_fp16)[name = tensor("linear_61_cast_fp16")]; - tensor input_351_cast_fp16 = add(x = input_343_cast_fp16, y = linear_61_cast_fp16)[name = tensor("input_351_cast_fp16")]; - tensor x_149_axes_0 = const()[name = tensor("x_149_axes_0"), val = tensor([-1])]; - tensor module_layers_6_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_6_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123486080)))]; - tensor module_layers_6_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_6_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123488192)))]; - tensor x_149_cast_fp16 = layer_norm(axes = x_149_axes_0, beta = module_layers_6_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_6_norm_conv_weight_to_fp16, x = input_351_cast_fp16)[name = tensor("x_149_cast_fp16")]; - tensor input_353_perm_0 = const()[name = tensor("input_353_perm_0"), val = tensor([0, 2, 1])]; - tensor input_355_pad_type_0 = const()[name = tensor("input_355_pad_type_0"), val = tensor("valid")]; - tensor input_355_strides_0 = const()[name = tensor("input_355_strides_0"), val = tensor([1])]; - tensor input_355_pad_0 = const()[name = tensor("input_355_pad_0"), val = tensor([0, 0])]; - tensor input_355_dilations_0 = const()[name = tensor("input_355_dilations_0"), val = tensor([1])]; - tensor input_355_groups_0 = const()[name = tensor("input_355_groups_0"), val = tensor(1)]; - tensor module_layers_6_conv_pointwise_conv1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123490304))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(125063232))), name = tensor("module_layers_6_conv_pointwise_conv1_weight_to_fp16_palettized"), shape = tensor([2048, 1024, 1])]; - tensor input_353_cast_fp16 = transpose(perm = input_353_perm_0, x = x_149_cast_fp16)[name = tensor("transpose_265")]; - tensor input_355_cast_fp16 = conv(dilations = input_355_dilations_0, groups = input_355_groups_0, pad = input_355_pad_0, pad_type = input_355_pad_type_0, strides = input_355_strides_0, weight = module_layers_6_conv_pointwise_conv1_weight_to_fp16_palettized, x = input_353_cast_fp16)[name = tensor("input_355_cast_fp16")]; - tensor x_151_split_num_splits_0 = const()[name = tensor("x_151_split_num_splits_0"), val = tensor(2)]; - tensor x_151_split_axis_0 = const()[name = tensor("x_151_split_axis_0"), val = tensor(1)]; - tensor x_151_split_cast_fp16_0, tensor x_151_split_cast_fp16_1 = split(axis = x_151_split_axis_0, num_splits = x_151_split_num_splits_0, x = input_355_cast_fp16)[name = tensor("x_151_split_cast_fp16")]; - tensor x_151_split_1_sigmoid_cast_fp16 = sigmoid(x = x_151_split_cast_fp16_1)[name = tensor("x_151_split_1_sigmoid_cast_fp16")]; - tensor x_151_cast_fp16 = mul(x = x_151_split_cast_fp16_0, y = x_151_split_1_sigmoid_cast_fp16)[name = tensor("x_151_cast_fp16")]; - tensor input_357_cast_fp16 = select(a = var_11_to_fp16, b = x_151_cast_fp16, cond = var_328)[name = tensor("input_357_cast_fp16")]; - tensor input_359_pad_0 = const()[name = tensor("input_359_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_359_mode_0 = const()[name = tensor("input_359_mode_0"), val = tensor("constant")]; - tensor const_77_to_fp16 = const()[name = tensor("const_77_to_fp16"), val = tensor(0x0p+0)]; - tensor input_359_cast_fp16 = pad(constant_val = const_77_to_fp16, mode = input_359_mode_0, pad = input_359_pad_0, x = input_357_cast_fp16)[name = tensor("input_359_cast_fp16")]; - tensor input_361_pad_type_0 = const()[name = tensor("input_361_pad_type_0"), val = tensor("valid")]; - tensor input_361_groups_0 = const()[name = tensor("input_361_groups_0"), val = tensor(1024)]; - tensor input_361_strides_0 = const()[name = tensor("input_361_strides_0"), val = tensor([1])]; - tensor input_361_pad_0 = const()[name = tensor("input_361_pad_0"), val = tensor([0, 0])]; - tensor input_361_dilations_0 = const()[name = tensor("input_361_dilations_0"), val = tensor([1])]; - tensor const_260_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(125063424))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(125070400))), name = tensor("const_260_to_fp16_palettized"), shape = tensor([1024, 1, 9])]; - tensor const_261_to_fp16 = const()[name = tensor("const_261_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(125070592)))]; - tensor input_363_cast_fp16 = conv(bias = const_261_to_fp16, dilations = input_361_dilations_0, groups = input_361_groups_0, pad = input_361_pad_0, pad_type = input_361_pad_type_0, strides = input_361_strides_0, weight = const_260_to_fp16_palettized, x = input_359_cast_fp16)[name = tensor("input_363_cast_fp16")]; - tensor input_365_cast_fp16 = silu(x = input_363_cast_fp16)[name = tensor("input_365_cast_fp16")]; - tensor x_153_pad_type_0 = const()[name = tensor("x_153_pad_type_0"), val = tensor("valid")]; - tensor x_153_strides_0 = const()[name = tensor("x_153_strides_0"), val = tensor([1])]; - tensor x_153_pad_0 = const()[name = tensor("x_153_pad_0"), val = tensor([0, 0])]; - tensor x_153_dilations_0 = const()[name = tensor("x_153_dilations_0"), val = tensor([1])]; - tensor x_153_groups_0 = const()[name = tensor("x_153_groups_0"), val = tensor(1)]; - tensor module_layers_6_conv_pointwise_conv2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(125072704))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(125859200))), name = tensor("module_layers_6_conv_pointwise_conv2_weight_to_fp16_palettized"), shape = tensor([1024, 1024, 1])]; - tensor x_153_cast_fp16 = conv(dilations = x_153_dilations_0, groups = x_153_groups_0, pad = x_153_pad_0, pad_type = x_153_pad_type_0, strides = x_153_strides_0, weight = module_layers_6_conv_pointwise_conv2_weight_to_fp16_palettized, x = input_365_cast_fp16)[name = tensor("x_153_cast_fp16")]; - tensor input_367_perm_0 = const()[name = tensor("input_367_perm_0"), val = tensor([0, 2, 1])]; - tensor input_367_cast_fp16 = transpose(perm = input_367_perm_0, x = x_153_cast_fp16)[name = tensor("transpose_264")]; - tensor input_369_cast_fp16 = add(x = input_351_cast_fp16, y = input_367_cast_fp16)[name = tensor("input_369_cast_fp16")]; - tensor input_371_axes_0 = const()[name = tensor("input_371_axes_0"), val = tensor([-1])]; - tensor module_layers_6_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_6_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(125859392)))]; - tensor module_layers_6_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_6_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(125861504)))]; - tensor input_371_cast_fp16 = layer_norm(axes = input_371_axes_0, beta = module_layers_6_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_6_norm_feed_forward2_weight_to_fp16, x = input_369_cast_fp16)[name = tensor("input_371_cast_fp16")]; - tensor module_layers_6_feed_forward2_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(125863616))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(129009408))), name = tensor("module_layers_6_feed_forward2_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_62_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_6_feed_forward2_linear1_weight_to_fp16_palettized, x = input_371_cast_fp16)[name = tensor("linear_62_cast_fp16")]; - tensor input_375_cast_fp16 = silu(x = linear_62_cast_fp16)[name = tensor("input_375_cast_fp16")]; - tensor module_layers_6_feed_forward2_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(129009600))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132155392))), name = tensor("module_layers_6_feed_forward2_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_63_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_6_feed_forward2_linear2_weight_to_fp16_palettized, x = input_375_cast_fp16)[name = tensor("linear_63_cast_fp16")]; - tensor var_1344_to_fp16 = const()[name = tensor("op_1344_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1345_cast_fp16 = mul(x = linear_63_cast_fp16, y = var_1344_to_fp16)[name = tensor("op_1345_cast_fp16")]; - tensor input_381_cast_fp16 = add(x = input_369_cast_fp16, y = var_1345_cast_fp16)[name = tensor("input_381_cast_fp16")]; - tensor input_383_axes_0 = const()[name = tensor("input_383_axes_0"), val = tensor([-1])]; - tensor module_layers_6_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_6_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132155584)))]; - tensor module_layers_6_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_6_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132157696)))]; - tensor input_383_cast_fp16 = layer_norm(axes = input_383_axes_0, beta = module_layers_6_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_6_norm_out_weight_to_fp16, x = input_381_cast_fp16)[name = tensor("input_383_cast_fp16")]; - tensor input_385_axes_0 = const()[name = tensor("input_385_axes_0"), val = tensor([-1])]; - tensor module_layers_7_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_7_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132159808)))]; - tensor module_layers_7_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_7_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132161920)))]; - tensor input_385_cast_fp16 = layer_norm(axes = input_385_axes_0, beta = module_layers_7_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_7_norm_feed_forward1_weight_to_fp16, x = input_383_cast_fp16)[name = tensor("input_385_cast_fp16")]; - tensor module_layers_7_feed_forward1_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132164032))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(135309824))), name = tensor("module_layers_7_feed_forward1_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_64_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_7_feed_forward1_linear1_weight_to_fp16_palettized, x = input_385_cast_fp16)[name = tensor("linear_64_cast_fp16")]; - tensor input_389_cast_fp16 = silu(x = linear_64_cast_fp16)[name = tensor("input_389_cast_fp16")]; - tensor module_layers_7_feed_forward1_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(135310016))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(138455808))), name = tensor("module_layers_7_feed_forward1_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_65_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_7_feed_forward1_linear2_weight_to_fp16_palettized, x = input_389_cast_fp16)[name = tensor("linear_65_cast_fp16")]; - tensor var_1373_to_fp16 = const()[name = tensor("op_1373_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1374_cast_fp16 = mul(x = linear_65_cast_fp16, y = var_1373_to_fp16)[name = tensor("op_1374_cast_fp16")]; - tensor input_395_cast_fp16 = add(x = input_383_cast_fp16, y = var_1374_cast_fp16)[name = tensor("input_395_cast_fp16")]; - tensor query_15_axes_0 = const()[name = tensor("query_15_axes_0"), val = tensor([-1])]; - tensor module_layers_7_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_7_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(138456000)))]; - tensor module_layers_7_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_7_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(138458112)))]; - tensor query_15_cast_fp16 = layer_norm(axes = query_15_axes_0, beta = module_layers_7_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_7_norm_self_att_weight_to_fp16, x = input_395_cast_fp16)[name = tensor("query_15_cast_fp16")]; - tensor module_layers_7_self_attn_linear_q_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(138460224))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(139246720))), name = tensor("module_layers_7_self_attn_linear_q_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_66_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_7_self_attn_linear_q_weight_to_fp16_palettized, x = query_15_cast_fp16)[name = tensor("linear_66_cast_fp16")]; - tensor var_1390 = const()[name = tensor("op_1390"), val = tensor([1, -1, 8, 128])]; - tensor q_43_cast_fp16 = reshape(shape = var_1390, x = linear_66_cast_fp16)[name = tensor("q_43_cast_fp16")]; - tensor module_layers_7_self_attn_linear_k_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(139246912))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140033408))), name = tensor("module_layers_7_self_attn_linear_k_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_67_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_7_self_attn_linear_k_weight_to_fp16_palettized, x = query_15_cast_fp16)[name = tensor("linear_67_cast_fp16")]; - tensor var_1394 = const()[name = tensor("op_1394"), val = tensor([1, -1, 8, 128])]; - tensor k_29_cast_fp16 = reshape(shape = var_1394, x = linear_67_cast_fp16)[name = tensor("k_29_cast_fp16")]; - tensor module_layers_7_self_attn_linear_v_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140033600))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140820096))), name = tensor("module_layers_7_self_attn_linear_v_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_68_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_7_self_attn_linear_v_weight_to_fp16_palettized, x = query_15_cast_fp16)[name = tensor("linear_68_cast_fp16")]; - tensor var_1398 = const()[name = tensor("op_1398"), val = tensor([1, -1, 8, 128])]; - tensor v_15_cast_fp16 = reshape(shape = var_1398, x = linear_68_cast_fp16)[name = tensor("v_15_cast_fp16")]; - tensor value_17_perm_0 = const()[name = tensor("value_17_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor module_layers_7_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_7_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140820288)))]; - tensor var_1410_cast_fp16 = add(x = q_43_cast_fp16, y = module_layers_7_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1410_cast_fp16")]; - tensor module_layers_7_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_7_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140822400)))]; - tensor var_1412_cast_fp16 = add(x = q_43_cast_fp16, y = module_layers_7_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1412_cast_fp16")]; - tensor q_with_bias_v_15_perm_0 = const()[name = tensor("q_with_bias_v_15_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor x_161_transpose_x_0 = const()[name = tensor("x_161_transpose_x_0"), val = tensor(false)]; - tensor x_161_transpose_y_0 = const()[name = tensor("x_161_transpose_y_0"), val = tensor(false)]; - tensor op_1414_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140824512))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141112576))), name = tensor("op_1414_to_fp16_palettized"), shape = tensor([1, 8, 128, 375])]; - tensor q_with_bias_v_15_cast_fp16 = transpose(perm = q_with_bias_v_15_perm_0, x = var_1412_cast_fp16)[name = tensor("transpose_263")]; - tensor x_161_cast_fp16 = matmul(transpose_x = x_161_transpose_x_0, transpose_y = x_161_transpose_y_0, x = q_with_bias_v_15_cast_fp16, y = op_1414_to_fp16_palettized)[name = tensor("x_161_cast_fp16")]; - tensor x_163_pad_0 = const()[name = tensor("x_163_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_163_mode_0 = const()[name = tensor("x_163_mode_0"), val = tensor("constant")]; - tensor const_84_to_fp16 = const()[name = tensor("const_84_to_fp16"), val = tensor(0x0p+0)]; - tensor x_163_cast_fp16 = pad(constant_val = const_84_to_fp16, mode = x_163_mode_0, pad = x_163_pad_0, x = x_161_cast_fp16)[name = tensor("x_163_cast_fp16")]; - tensor var_1422 = const()[name = tensor("op_1422"), val = tensor([1, 8, -1, 188])]; - tensor x_165_cast_fp16 = reshape(shape = var_1422, x = x_163_cast_fp16)[name = tensor("x_165_cast_fp16")]; - tensor var_1426_begin_0 = const()[name = tensor("op_1426_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_1426_end_0 = const()[name = tensor("op_1426_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_1426_end_mask_0 = const()[name = tensor("op_1426_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_1426_cast_fp16 = slice_by_index(begin = var_1426_begin_0, end = var_1426_end_0, end_mask = var_1426_end_mask_0, x = x_165_cast_fp16)[name = tensor("op_1426_cast_fp16")]; - tensor var_1427 = const()[name = tensor("op_1427"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_29_cast_fp16 = reshape(shape = var_1427, x = var_1426_cast_fp16)[name = tensor("matrix_bd_29_cast_fp16")]; - tensor matrix_ac_15_transpose_x_0 = const()[name = tensor("matrix_ac_15_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_15_transpose_y_0 = const()[name = tensor("matrix_ac_15_transpose_y_0"), val = tensor(false)]; - tensor transpose_110_perm_0 = const()[name = tensor("transpose_110_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_111_perm_0 = const()[name = tensor("transpose_111_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_111 = transpose(perm = transpose_111_perm_0, x = k_29_cast_fp16)[name = tensor("transpose_261")]; - tensor transpose_110 = transpose(perm = transpose_110_perm_0, x = var_1410_cast_fp16)[name = tensor("transpose_262")]; - tensor matrix_ac_15_cast_fp16 = matmul(transpose_x = matrix_ac_15_transpose_x_0, transpose_y = matrix_ac_15_transpose_y_0, x = transpose_110, y = transpose_111)[name = tensor("matrix_ac_15_cast_fp16")]; - tensor matrix_bd_31_begin_0 = const()[name = tensor("matrix_bd_31_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_31_end_0 = const()[name = tensor("matrix_bd_31_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_31_end_mask_0 = const()[name = tensor("matrix_bd_31_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_31_cast_fp16 = slice_by_index(begin = matrix_bd_31_begin_0, end = matrix_bd_31_end_0, end_mask = matrix_bd_31_end_mask_0, x = matrix_bd_29_cast_fp16)[name = tensor("matrix_bd_31_cast_fp16")]; - tensor var_1436_cast_fp16 = add(x = matrix_ac_15_cast_fp16, y = matrix_bd_31_cast_fp16)[name = tensor("op_1436_cast_fp16")]; - tensor _inversed_scores_29_y_0_to_fp16 = const()[name = tensor("_inversed_scores_29_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; - tensor _inversed_scores_29_cast_fp16 = mul(x = var_1436_cast_fp16, y = _inversed_scores_29_y_0_to_fp16)[name = tensor("_inversed_scores_29_cast_fp16")]; - tensor scores_31_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_29_cast_fp16, cond = mask_3)[name = tensor("scores_31_cast_fp16")]; - tensor var_1442_cast_fp16 = softmax(axis = var_30, x = scores_31_cast_fp16)[name = tensor("op_1442_cast_fp16")]; - tensor input_397_cast_fp16 = select(a = var_11_to_fp16, b = var_1442_cast_fp16, cond = mask_3)[name = tensor("input_397_cast_fp16")]; - tensor x_167_transpose_x_0 = const()[name = tensor("x_167_transpose_x_0"), val = tensor(false)]; - tensor x_167_transpose_y_0 = const()[name = tensor("x_167_transpose_y_0"), val = tensor(false)]; - tensor value_17_cast_fp16 = transpose(perm = value_17_perm_0, x = v_15_cast_fp16)[name = tensor("transpose_260")]; - tensor x_167_cast_fp16 = matmul(transpose_x = x_167_transpose_x_0, transpose_y = x_167_transpose_y_0, x = input_397_cast_fp16, y = value_17_cast_fp16)[name = tensor("x_167_cast_fp16")]; - tensor var_1446_perm_0 = const()[name = tensor("op_1446_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_1447 = const()[name = tensor("op_1447"), val = tensor([1, -1, 1024])]; - tensor var_1446_cast_fp16 = transpose(perm = var_1446_perm_0, x = x_167_cast_fp16)[name = tensor("transpose_259")]; - tensor input_399_cast_fp16 = reshape(shape = var_1447, x = var_1446_cast_fp16)[name = tensor("input_399_cast_fp16")]; - tensor module_layers_7_self_attn_linear_out_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141112768))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141899264))), name = tensor("module_layers_7_self_attn_linear_out_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_70_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_7_self_attn_linear_out_weight_to_fp16_palettized, x = input_399_cast_fp16)[name = tensor("linear_70_cast_fp16")]; - tensor input_403_cast_fp16 = add(x = input_395_cast_fp16, y = linear_70_cast_fp16)[name = tensor("input_403_cast_fp16")]; - tensor x_171_axes_0 = const()[name = tensor("x_171_axes_0"), val = tensor([-1])]; - tensor module_layers_7_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_7_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141899456)))]; - tensor module_layers_7_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_7_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141901568)))]; - tensor x_171_cast_fp16 = layer_norm(axes = x_171_axes_0, beta = module_layers_7_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_7_norm_conv_weight_to_fp16, x = input_403_cast_fp16)[name = tensor("x_171_cast_fp16")]; - tensor input_405_perm_0 = const()[name = tensor("input_405_perm_0"), val = tensor([0, 2, 1])]; - tensor input_407_pad_type_0 = const()[name = tensor("input_407_pad_type_0"), val = tensor("valid")]; - tensor input_407_strides_0 = const()[name = tensor("input_407_strides_0"), val = tensor([1])]; - tensor input_407_pad_0 = const()[name = tensor("input_407_pad_0"), val = tensor([0, 0])]; - tensor input_407_dilations_0 = const()[name = tensor("input_407_dilations_0"), val = tensor([1])]; - tensor input_407_groups_0 = const()[name = tensor("input_407_groups_0"), val = tensor(1)]; - tensor module_layers_7_conv_pointwise_conv1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141903680))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(143476608))), name = tensor("module_layers_7_conv_pointwise_conv1_weight_to_fp16_palettized"), shape = tensor([2048, 1024, 1])]; - tensor input_405_cast_fp16 = transpose(perm = input_405_perm_0, x = x_171_cast_fp16)[name = tensor("transpose_258")]; - tensor input_407_cast_fp16 = conv(dilations = input_407_dilations_0, groups = input_407_groups_0, pad = input_407_pad_0, pad_type = input_407_pad_type_0, strides = input_407_strides_0, weight = module_layers_7_conv_pointwise_conv1_weight_to_fp16_palettized, x = input_405_cast_fp16)[name = tensor("input_407_cast_fp16")]; - tensor x_173_split_num_splits_0 = const()[name = tensor("x_173_split_num_splits_0"), val = tensor(2)]; - tensor x_173_split_axis_0 = const()[name = tensor("x_173_split_axis_0"), val = tensor(1)]; - tensor x_173_split_cast_fp16_0, tensor x_173_split_cast_fp16_1 = split(axis = x_173_split_axis_0, num_splits = x_173_split_num_splits_0, x = input_407_cast_fp16)[name = tensor("x_173_split_cast_fp16")]; - tensor x_173_split_1_sigmoid_cast_fp16 = sigmoid(x = x_173_split_cast_fp16_1)[name = tensor("x_173_split_1_sigmoid_cast_fp16")]; - tensor x_173_cast_fp16 = mul(x = x_173_split_cast_fp16_0, y = x_173_split_1_sigmoid_cast_fp16)[name = tensor("x_173_cast_fp16")]; - tensor input_409_cast_fp16 = select(a = var_11_to_fp16, b = x_173_cast_fp16, cond = var_328)[name = tensor("input_409_cast_fp16")]; - tensor input_411_pad_0 = const()[name = tensor("input_411_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_411_mode_0 = const()[name = tensor("input_411_mode_0"), val = tensor("constant")]; - tensor const_87_to_fp16 = const()[name = tensor("const_87_to_fp16"), val = tensor(0x0p+0)]; - tensor input_411_cast_fp16 = pad(constant_val = const_87_to_fp16, mode = input_411_mode_0, pad = input_411_pad_0, x = input_409_cast_fp16)[name = tensor("input_411_cast_fp16")]; - tensor input_413_pad_type_0 = const()[name = tensor("input_413_pad_type_0"), val = tensor("valid")]; - tensor input_413_groups_0 = const()[name = tensor("input_413_groups_0"), val = tensor(1024)]; - tensor input_413_strides_0 = const()[name = tensor("input_413_strides_0"), val = tensor([1])]; - tensor input_413_pad_0 = const()[name = tensor("input_413_pad_0"), val = tensor([0, 0])]; - tensor input_413_dilations_0 = const()[name = tensor("input_413_dilations_0"), val = tensor([1])]; - tensor const_262_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(143476800))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(143483776))), name = tensor("const_262_to_fp16_palettized"), shape = tensor([1024, 1, 9])]; - tensor const_263_to_fp16 = const()[name = tensor("const_263_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(143483968)))]; - tensor input_415_cast_fp16 = conv(bias = const_263_to_fp16, dilations = input_413_dilations_0, groups = input_413_groups_0, pad = input_413_pad_0, pad_type = input_413_pad_type_0, strides = input_413_strides_0, weight = const_262_to_fp16_palettized, x = input_411_cast_fp16)[name = tensor("input_415_cast_fp16")]; - tensor input_417_cast_fp16 = silu(x = input_415_cast_fp16)[name = tensor("input_417_cast_fp16")]; - tensor x_175_pad_type_0 = const()[name = tensor("x_175_pad_type_0"), val = tensor("valid")]; - tensor x_175_strides_0 = const()[name = tensor("x_175_strides_0"), val = tensor([1])]; - tensor x_175_pad_0 = const()[name = tensor("x_175_pad_0"), val = tensor([0, 0])]; - tensor x_175_dilations_0 = const()[name = tensor("x_175_dilations_0"), val = tensor([1])]; - tensor x_175_groups_0 = const()[name = tensor("x_175_groups_0"), val = tensor(1)]; - tensor module_layers_7_conv_pointwise_conv2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(143486080))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(144272576))), name = tensor("module_layers_7_conv_pointwise_conv2_weight_to_fp16_palettized"), shape = tensor([1024, 1024, 1])]; - tensor x_175_cast_fp16 = conv(dilations = x_175_dilations_0, groups = x_175_groups_0, pad = x_175_pad_0, pad_type = x_175_pad_type_0, strides = x_175_strides_0, weight = module_layers_7_conv_pointwise_conv2_weight_to_fp16_palettized, x = input_417_cast_fp16)[name = tensor("x_175_cast_fp16")]; - tensor input_419_perm_0 = const()[name = tensor("input_419_perm_0"), val = tensor([0, 2, 1])]; - tensor input_419_cast_fp16 = transpose(perm = input_419_perm_0, x = x_175_cast_fp16)[name = tensor("transpose_257")]; - tensor input_421_cast_fp16 = add(x = input_403_cast_fp16, y = input_419_cast_fp16)[name = tensor("input_421_cast_fp16")]; - tensor input_423_axes_0 = const()[name = tensor("input_423_axes_0"), val = tensor([-1])]; - tensor module_layers_7_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_7_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(144272768)))]; - tensor module_layers_7_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_7_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(144274880)))]; - tensor input_423_cast_fp16 = layer_norm(axes = input_423_axes_0, beta = module_layers_7_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_7_norm_feed_forward2_weight_to_fp16, x = input_421_cast_fp16)[name = tensor("input_423_cast_fp16")]; - tensor module_layers_7_feed_forward2_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(144276992))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147422784))), name = tensor("module_layers_7_feed_forward2_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_71_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_7_feed_forward2_linear1_weight_to_fp16_palettized, x = input_423_cast_fp16)[name = tensor("linear_71_cast_fp16")]; - tensor input_427_cast_fp16 = silu(x = linear_71_cast_fp16)[name = tensor("input_427_cast_fp16")]; - tensor module_layers_7_feed_forward2_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147422976))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(150568768))), name = tensor("module_layers_7_feed_forward2_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_72_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_7_feed_forward2_linear2_weight_to_fp16_palettized, x = input_427_cast_fp16)[name = tensor("linear_72_cast_fp16")]; - tensor var_1507_to_fp16 = const()[name = tensor("op_1507_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1508_cast_fp16 = mul(x = linear_72_cast_fp16, y = var_1507_to_fp16)[name = tensor("op_1508_cast_fp16")]; - tensor input_433_cast_fp16 = add(x = input_421_cast_fp16, y = var_1508_cast_fp16)[name = tensor("input_433_cast_fp16")]; - tensor input_435_axes_0 = const()[name = tensor("input_435_axes_0"), val = tensor([-1])]; - tensor module_layers_7_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_7_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(150568960)))]; - tensor module_layers_7_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_7_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(150571072)))]; - tensor input_435_cast_fp16 = layer_norm(axes = input_435_axes_0, beta = module_layers_7_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_7_norm_out_weight_to_fp16, x = input_433_cast_fp16)[name = tensor("input_435_cast_fp16")]; - tensor input_437_axes_0 = const()[name = tensor("input_437_axes_0"), val = tensor([-1])]; - tensor module_layers_8_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_8_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(150573184)))]; - tensor module_layers_8_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_8_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(150575296)))]; - tensor input_437_cast_fp16 = layer_norm(axes = input_437_axes_0, beta = module_layers_8_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_8_norm_feed_forward1_weight_to_fp16, x = input_435_cast_fp16)[name = tensor("input_437_cast_fp16")]; - tensor module_layers_8_feed_forward1_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(150577408))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(153723200))), name = tensor("module_layers_8_feed_forward1_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_73_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_8_feed_forward1_linear1_weight_to_fp16_palettized, x = input_437_cast_fp16)[name = tensor("linear_73_cast_fp16")]; - tensor input_441_cast_fp16 = silu(x = linear_73_cast_fp16)[name = tensor("input_441_cast_fp16")]; - tensor module_layers_8_feed_forward1_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(153723392))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(156869184))), name = tensor("module_layers_8_feed_forward1_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_74_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_8_feed_forward1_linear2_weight_to_fp16_palettized, x = input_441_cast_fp16)[name = tensor("linear_74_cast_fp16")]; - tensor var_1536_to_fp16 = const()[name = tensor("op_1536_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1537_cast_fp16 = mul(x = linear_74_cast_fp16, y = var_1536_to_fp16)[name = tensor("op_1537_cast_fp16")]; - tensor input_447_cast_fp16 = add(x = input_435_cast_fp16, y = var_1537_cast_fp16)[name = tensor("input_447_cast_fp16")]; - tensor query_17_axes_0 = const()[name = tensor("query_17_axes_0"), val = tensor([-1])]; - tensor module_layers_8_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_8_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(156869376)))]; - tensor module_layers_8_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_8_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(156871488)))]; - tensor query_17_cast_fp16 = layer_norm(axes = query_17_axes_0, beta = module_layers_8_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_8_norm_self_att_weight_to_fp16, x = input_447_cast_fp16)[name = tensor("query_17_cast_fp16")]; - tensor module_layers_8_self_attn_linear_q_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(156873600))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157660096))), name = tensor("module_layers_8_self_attn_linear_q_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_75_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_8_self_attn_linear_q_weight_to_fp16_palettized, x = query_17_cast_fp16)[name = tensor("linear_75_cast_fp16")]; - tensor var_1553 = const()[name = tensor("op_1553"), val = tensor([1, -1, 8, 128])]; - tensor q_49_cast_fp16 = reshape(shape = var_1553, x = linear_75_cast_fp16)[name = tensor("q_49_cast_fp16")]; - tensor module_layers_8_self_attn_linear_k_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157660288))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158446784))), name = tensor("module_layers_8_self_attn_linear_k_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_76_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_8_self_attn_linear_k_weight_to_fp16_palettized, x = query_17_cast_fp16)[name = tensor("linear_76_cast_fp16")]; - tensor var_1557 = const()[name = tensor("op_1557"), val = tensor([1, -1, 8, 128])]; - tensor k_33_cast_fp16 = reshape(shape = var_1557, x = linear_76_cast_fp16)[name = tensor("k_33_cast_fp16")]; - tensor module_layers_8_self_attn_linear_v_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158446976))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159233472))), name = tensor("module_layers_8_self_attn_linear_v_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_77_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_8_self_attn_linear_v_weight_to_fp16_palettized, x = query_17_cast_fp16)[name = tensor("linear_77_cast_fp16")]; - tensor var_1561 = const()[name = tensor("op_1561"), val = tensor([1, -1, 8, 128])]; - tensor v_17_cast_fp16 = reshape(shape = var_1561, x = linear_77_cast_fp16)[name = tensor("v_17_cast_fp16")]; - tensor value_19_perm_0 = const()[name = tensor("value_19_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor module_layers_8_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_8_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159233664)))]; - tensor var_1573_cast_fp16 = add(x = q_49_cast_fp16, y = module_layers_8_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1573_cast_fp16")]; - tensor module_layers_8_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_8_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159235776)))]; - tensor var_1575_cast_fp16 = add(x = q_49_cast_fp16, y = module_layers_8_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1575_cast_fp16")]; - tensor q_with_bias_v_17_perm_0 = const()[name = tensor("q_with_bias_v_17_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor x_183_transpose_x_0 = const()[name = tensor("x_183_transpose_x_0"), val = tensor(false)]; - tensor x_183_transpose_y_0 = const()[name = tensor("x_183_transpose_y_0"), val = tensor(false)]; - tensor op_1577_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159237888))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159525952))), name = tensor("op_1577_to_fp16_palettized"), shape = tensor([1, 8, 128, 375])]; - tensor q_with_bias_v_17_cast_fp16 = transpose(perm = q_with_bias_v_17_perm_0, x = var_1575_cast_fp16)[name = tensor("transpose_256")]; - tensor x_183_cast_fp16 = matmul(transpose_x = x_183_transpose_x_0, transpose_y = x_183_transpose_y_0, x = q_with_bias_v_17_cast_fp16, y = op_1577_to_fp16_palettized)[name = tensor("x_183_cast_fp16")]; - tensor x_185_pad_0 = const()[name = tensor("x_185_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_185_mode_0 = const()[name = tensor("x_185_mode_0"), val = tensor("constant")]; - tensor const_94_to_fp16 = const()[name = tensor("const_94_to_fp16"), val = tensor(0x0p+0)]; - tensor x_185_cast_fp16 = pad(constant_val = const_94_to_fp16, mode = x_185_mode_0, pad = x_185_pad_0, x = x_183_cast_fp16)[name = tensor("x_185_cast_fp16")]; - tensor var_1585 = const()[name = tensor("op_1585"), val = tensor([1, 8, -1, 188])]; - tensor x_187_cast_fp16 = reshape(shape = var_1585, x = x_185_cast_fp16)[name = tensor("x_187_cast_fp16")]; - tensor var_1589_begin_0 = const()[name = tensor("op_1589_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_1589_end_0 = const()[name = tensor("op_1589_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_1589_end_mask_0 = const()[name = tensor("op_1589_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_1589_cast_fp16 = slice_by_index(begin = var_1589_begin_0, end = var_1589_end_0, end_mask = var_1589_end_mask_0, x = x_187_cast_fp16)[name = tensor("op_1589_cast_fp16")]; - tensor var_1590 = const()[name = tensor("op_1590"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_33_cast_fp16 = reshape(shape = var_1590, x = var_1589_cast_fp16)[name = tensor("matrix_bd_33_cast_fp16")]; - tensor matrix_ac_17_transpose_x_0 = const()[name = tensor("matrix_ac_17_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_17_transpose_y_0 = const()[name = tensor("matrix_ac_17_transpose_y_0"), val = tensor(false)]; - tensor transpose_112_perm_0 = const()[name = tensor("transpose_112_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_113_perm_0 = const()[name = tensor("transpose_113_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_113 = transpose(perm = transpose_113_perm_0, x = k_33_cast_fp16)[name = tensor("transpose_254")]; - tensor transpose_112 = transpose(perm = transpose_112_perm_0, x = var_1573_cast_fp16)[name = tensor("transpose_255")]; - tensor matrix_ac_17_cast_fp16 = matmul(transpose_x = matrix_ac_17_transpose_x_0, transpose_y = matrix_ac_17_transpose_y_0, x = transpose_112, y = transpose_113)[name = tensor("matrix_ac_17_cast_fp16")]; - tensor matrix_bd_35_begin_0 = const()[name = tensor("matrix_bd_35_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_35_end_0 = const()[name = tensor("matrix_bd_35_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_35_end_mask_0 = const()[name = tensor("matrix_bd_35_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_35_cast_fp16 = slice_by_index(begin = matrix_bd_35_begin_0, end = matrix_bd_35_end_0, end_mask = matrix_bd_35_end_mask_0, x = matrix_bd_33_cast_fp16)[name = tensor("matrix_bd_35_cast_fp16")]; - tensor var_1599_cast_fp16 = add(x = matrix_ac_17_cast_fp16, y = matrix_bd_35_cast_fp16)[name = tensor("op_1599_cast_fp16")]; - tensor _inversed_scores_33_y_0_to_fp16 = const()[name = tensor("_inversed_scores_33_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; - tensor _inversed_scores_33_cast_fp16 = mul(x = var_1599_cast_fp16, y = _inversed_scores_33_y_0_to_fp16)[name = tensor("_inversed_scores_33_cast_fp16")]; - tensor scores_35_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_33_cast_fp16, cond = mask_3)[name = tensor("scores_35_cast_fp16")]; - tensor var_1605_cast_fp16 = softmax(axis = var_30, x = scores_35_cast_fp16)[name = tensor("op_1605_cast_fp16")]; - tensor input_449_cast_fp16 = select(a = var_11_to_fp16, b = var_1605_cast_fp16, cond = mask_3)[name = tensor("input_449_cast_fp16")]; - tensor x_189_transpose_x_0 = const()[name = tensor("x_189_transpose_x_0"), val = tensor(false)]; - tensor x_189_transpose_y_0 = const()[name = tensor("x_189_transpose_y_0"), val = tensor(false)]; - tensor value_19_cast_fp16 = transpose(perm = value_19_perm_0, x = v_17_cast_fp16)[name = tensor("transpose_253")]; - tensor x_189_cast_fp16 = matmul(transpose_x = x_189_transpose_x_0, transpose_y = x_189_transpose_y_0, x = input_449_cast_fp16, y = value_19_cast_fp16)[name = tensor("x_189_cast_fp16")]; - tensor var_1609_perm_0 = const()[name = tensor("op_1609_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_1610 = const()[name = tensor("op_1610"), val = tensor([1, -1, 1024])]; - tensor var_1609_cast_fp16 = transpose(perm = var_1609_perm_0, x = x_189_cast_fp16)[name = tensor("transpose_252")]; - tensor input_451_cast_fp16 = reshape(shape = var_1610, x = var_1609_cast_fp16)[name = tensor("input_451_cast_fp16")]; - tensor module_layers_8_self_attn_linear_out_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159526144))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160312640))), name = tensor("module_layers_8_self_attn_linear_out_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_79_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_8_self_attn_linear_out_weight_to_fp16_palettized, x = input_451_cast_fp16)[name = tensor("linear_79_cast_fp16")]; - tensor input_455_cast_fp16 = add(x = input_447_cast_fp16, y = linear_79_cast_fp16)[name = tensor("input_455_cast_fp16")]; - tensor x_193_axes_0 = const()[name = tensor("x_193_axes_0"), val = tensor([-1])]; - tensor module_layers_8_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_8_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160312832)))]; - tensor module_layers_8_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_8_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160314944)))]; - tensor x_193_cast_fp16 = layer_norm(axes = x_193_axes_0, beta = module_layers_8_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_8_norm_conv_weight_to_fp16, x = input_455_cast_fp16)[name = tensor("x_193_cast_fp16")]; - tensor input_457_perm_0 = const()[name = tensor("input_457_perm_0"), val = tensor([0, 2, 1])]; - tensor input_459_pad_type_0 = const()[name = tensor("input_459_pad_type_0"), val = tensor("valid")]; - tensor input_459_strides_0 = const()[name = tensor("input_459_strides_0"), val = tensor([1])]; - tensor input_459_pad_0 = const()[name = tensor("input_459_pad_0"), val = tensor([0, 0])]; - tensor input_459_dilations_0 = const()[name = tensor("input_459_dilations_0"), val = tensor([1])]; - tensor input_459_groups_0 = const()[name = tensor("input_459_groups_0"), val = tensor(1)]; - tensor module_layers_8_conv_pointwise_conv1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160317056))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161889984))), name = tensor("module_layers_8_conv_pointwise_conv1_weight_to_fp16_palettized"), shape = tensor([2048, 1024, 1])]; - tensor input_457_cast_fp16 = transpose(perm = input_457_perm_0, x = x_193_cast_fp16)[name = tensor("transpose_251")]; - tensor input_459_cast_fp16 = conv(dilations = input_459_dilations_0, groups = input_459_groups_0, pad = input_459_pad_0, pad_type = input_459_pad_type_0, strides = input_459_strides_0, weight = module_layers_8_conv_pointwise_conv1_weight_to_fp16_palettized, x = input_457_cast_fp16)[name = tensor("input_459_cast_fp16")]; - tensor x_195_split_num_splits_0 = const()[name = tensor("x_195_split_num_splits_0"), val = tensor(2)]; - tensor x_195_split_axis_0 = const()[name = tensor("x_195_split_axis_0"), val = tensor(1)]; - tensor x_195_split_cast_fp16_0, tensor x_195_split_cast_fp16_1 = split(axis = x_195_split_axis_0, num_splits = x_195_split_num_splits_0, x = input_459_cast_fp16)[name = tensor("x_195_split_cast_fp16")]; - tensor x_195_split_1_sigmoid_cast_fp16 = sigmoid(x = x_195_split_cast_fp16_1)[name = tensor("x_195_split_1_sigmoid_cast_fp16")]; - tensor x_195_cast_fp16 = mul(x = x_195_split_cast_fp16_0, y = x_195_split_1_sigmoid_cast_fp16)[name = tensor("x_195_cast_fp16")]; - tensor input_461_cast_fp16 = select(a = var_11_to_fp16, b = x_195_cast_fp16, cond = var_328)[name = tensor("input_461_cast_fp16")]; - tensor input_463_pad_0 = const()[name = tensor("input_463_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_463_mode_0 = const()[name = tensor("input_463_mode_0"), val = tensor("constant")]; - tensor const_97_to_fp16 = const()[name = tensor("const_97_to_fp16"), val = tensor(0x0p+0)]; - tensor input_463_cast_fp16 = pad(constant_val = const_97_to_fp16, mode = input_463_mode_0, pad = input_463_pad_0, x = input_461_cast_fp16)[name = tensor("input_463_cast_fp16")]; - tensor input_465_pad_type_0 = const()[name = tensor("input_465_pad_type_0"), val = tensor("valid")]; - tensor input_465_groups_0 = const()[name = tensor("input_465_groups_0"), val = tensor(1024)]; - tensor input_465_strides_0 = const()[name = tensor("input_465_strides_0"), val = tensor([1])]; - tensor input_465_pad_0 = const()[name = tensor("input_465_pad_0"), val = tensor([0, 0])]; - tensor input_465_dilations_0 = const()[name = tensor("input_465_dilations_0"), val = tensor([1])]; - tensor const_264_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161890176))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161897152))), name = tensor("const_264_to_fp16_palettized"), shape = tensor([1024, 1, 9])]; - tensor const_265_to_fp16 = const()[name = tensor("const_265_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161897344)))]; - tensor input_467_cast_fp16 = conv(bias = const_265_to_fp16, dilations = input_465_dilations_0, groups = input_465_groups_0, pad = input_465_pad_0, pad_type = input_465_pad_type_0, strides = input_465_strides_0, weight = const_264_to_fp16_palettized, x = input_463_cast_fp16)[name = tensor("input_467_cast_fp16")]; - tensor input_469_cast_fp16 = silu(x = input_467_cast_fp16)[name = tensor("input_469_cast_fp16")]; - tensor x_197_pad_type_0 = const()[name = tensor("x_197_pad_type_0"), val = tensor("valid")]; - tensor x_197_strides_0 = const()[name = tensor("x_197_strides_0"), val = tensor([1])]; - tensor x_197_pad_0 = const()[name = tensor("x_197_pad_0"), val = tensor([0, 0])]; - tensor x_197_dilations_0 = const()[name = tensor("x_197_dilations_0"), val = tensor([1])]; - tensor x_197_groups_0 = const()[name = tensor("x_197_groups_0"), val = tensor(1)]; - tensor module_layers_8_conv_pointwise_conv2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161899456))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(162685952))), name = tensor("module_layers_8_conv_pointwise_conv2_weight_to_fp16_palettized"), shape = tensor([1024, 1024, 1])]; - tensor x_197_cast_fp16 = conv(dilations = x_197_dilations_0, groups = x_197_groups_0, pad = x_197_pad_0, pad_type = x_197_pad_type_0, strides = x_197_strides_0, weight = module_layers_8_conv_pointwise_conv2_weight_to_fp16_palettized, x = input_469_cast_fp16)[name = tensor("x_197_cast_fp16")]; - tensor input_471_perm_0 = const()[name = tensor("input_471_perm_0"), val = tensor([0, 2, 1])]; - tensor input_471_cast_fp16 = transpose(perm = input_471_perm_0, x = x_197_cast_fp16)[name = tensor("transpose_250")]; - tensor input_473_cast_fp16 = add(x = input_455_cast_fp16, y = input_471_cast_fp16)[name = tensor("input_473_cast_fp16")]; - tensor input_475_axes_0 = const()[name = tensor("input_475_axes_0"), val = tensor([-1])]; - tensor module_layers_8_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_8_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(162686144)))]; - tensor module_layers_8_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_8_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(162688256)))]; - tensor input_475_cast_fp16 = layer_norm(axes = input_475_axes_0, beta = module_layers_8_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_8_norm_feed_forward2_weight_to_fp16, x = input_473_cast_fp16)[name = tensor("input_475_cast_fp16")]; - tensor module_layers_8_feed_forward2_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(162690368))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165836160))), name = tensor("module_layers_8_feed_forward2_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_80_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_8_feed_forward2_linear1_weight_to_fp16_palettized, x = input_475_cast_fp16)[name = tensor("linear_80_cast_fp16")]; - tensor input_479_cast_fp16 = silu(x = linear_80_cast_fp16)[name = tensor("input_479_cast_fp16")]; - tensor module_layers_8_feed_forward2_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165836352))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(168982144))), name = tensor("module_layers_8_feed_forward2_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_81_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_8_feed_forward2_linear2_weight_to_fp16_palettized, x = input_479_cast_fp16)[name = tensor("linear_81_cast_fp16")]; - tensor var_1670_to_fp16 = const()[name = tensor("op_1670_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1671_cast_fp16 = mul(x = linear_81_cast_fp16, y = var_1670_to_fp16)[name = tensor("op_1671_cast_fp16")]; - tensor input_485_cast_fp16 = add(x = input_473_cast_fp16, y = var_1671_cast_fp16)[name = tensor("input_485_cast_fp16")]; - tensor input_487_axes_0 = const()[name = tensor("input_487_axes_0"), val = tensor([-1])]; - tensor module_layers_8_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_8_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(168982336)))]; - tensor module_layers_8_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_8_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(168984448)))]; - tensor input_487_cast_fp16 = layer_norm(axes = input_487_axes_0, beta = module_layers_8_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_8_norm_out_weight_to_fp16, x = input_485_cast_fp16)[name = tensor("input_487_cast_fp16")]; - tensor input_489_axes_0 = const()[name = tensor("input_489_axes_0"), val = tensor([-1])]; - tensor module_layers_9_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_9_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(168986560)))]; - tensor module_layers_9_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_9_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(168988672)))]; - tensor input_489_cast_fp16 = layer_norm(axes = input_489_axes_0, beta = module_layers_9_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_9_norm_feed_forward1_weight_to_fp16, x = input_487_cast_fp16)[name = tensor("input_489_cast_fp16")]; - tensor module_layers_9_feed_forward1_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(168990784))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(172136576))), name = tensor("module_layers_9_feed_forward1_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_82_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_9_feed_forward1_linear1_weight_to_fp16_palettized, x = input_489_cast_fp16)[name = tensor("linear_82_cast_fp16")]; - tensor input_493_cast_fp16 = silu(x = linear_82_cast_fp16)[name = tensor("input_493_cast_fp16")]; - tensor module_layers_9_feed_forward1_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(172136768))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(175282560))), name = tensor("module_layers_9_feed_forward1_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_83_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_9_feed_forward1_linear2_weight_to_fp16_palettized, x = input_493_cast_fp16)[name = tensor("linear_83_cast_fp16")]; - tensor var_1699_to_fp16 = const()[name = tensor("op_1699_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1700_cast_fp16 = mul(x = linear_83_cast_fp16, y = var_1699_to_fp16)[name = tensor("op_1700_cast_fp16")]; - tensor input_499_cast_fp16 = add(x = input_487_cast_fp16, y = var_1700_cast_fp16)[name = tensor("input_499_cast_fp16")]; - tensor query_19_axes_0 = const()[name = tensor("query_19_axes_0"), val = tensor([-1])]; - tensor module_layers_9_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_9_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(175282752)))]; - tensor module_layers_9_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_9_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(175284864)))]; - tensor query_19_cast_fp16 = layer_norm(axes = query_19_axes_0, beta = module_layers_9_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_9_norm_self_att_weight_to_fp16, x = input_499_cast_fp16)[name = tensor("query_19_cast_fp16")]; - tensor module_layers_9_self_attn_linear_q_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(175286976))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(176073472))), name = tensor("module_layers_9_self_attn_linear_q_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_84_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_9_self_attn_linear_q_weight_to_fp16_palettized, x = query_19_cast_fp16)[name = tensor("linear_84_cast_fp16")]; - tensor var_1716 = const()[name = tensor("op_1716"), val = tensor([1, -1, 8, 128])]; - tensor q_55_cast_fp16 = reshape(shape = var_1716, x = linear_84_cast_fp16)[name = tensor("q_55_cast_fp16")]; - tensor module_layers_9_self_attn_linear_k_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(176073664))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(176860160))), name = tensor("module_layers_9_self_attn_linear_k_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_85_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_9_self_attn_linear_k_weight_to_fp16_palettized, x = query_19_cast_fp16)[name = tensor("linear_85_cast_fp16")]; - tensor var_1720 = const()[name = tensor("op_1720"), val = tensor([1, -1, 8, 128])]; - tensor k_37_cast_fp16 = reshape(shape = var_1720, x = linear_85_cast_fp16)[name = tensor("k_37_cast_fp16")]; - tensor module_layers_9_self_attn_linear_v_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(176860352))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(177646848))), name = tensor("module_layers_9_self_attn_linear_v_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_86_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_9_self_attn_linear_v_weight_to_fp16_palettized, x = query_19_cast_fp16)[name = tensor("linear_86_cast_fp16")]; - tensor var_1724 = const()[name = tensor("op_1724"), val = tensor([1, -1, 8, 128])]; - tensor v_19_cast_fp16 = reshape(shape = var_1724, x = linear_86_cast_fp16)[name = tensor("v_19_cast_fp16")]; - tensor value_21_perm_0 = const()[name = tensor("value_21_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor module_layers_9_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_9_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(177647040)))]; - tensor var_1736_cast_fp16 = add(x = q_55_cast_fp16, y = module_layers_9_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1736_cast_fp16")]; - tensor module_layers_9_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_9_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(177649152)))]; - tensor var_1738_cast_fp16 = add(x = q_55_cast_fp16, y = module_layers_9_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1738_cast_fp16")]; - tensor q_with_bias_v_19_perm_0 = const()[name = tensor("q_with_bias_v_19_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor x_205_transpose_x_0 = const()[name = tensor("x_205_transpose_x_0"), val = tensor(false)]; - tensor x_205_transpose_y_0 = const()[name = tensor("x_205_transpose_y_0"), val = tensor(false)]; - tensor op_1740_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(177651264))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(177939328))), name = tensor("op_1740_to_fp16_palettized"), shape = tensor([1, 8, 128, 375])]; - tensor q_with_bias_v_19_cast_fp16 = transpose(perm = q_with_bias_v_19_perm_0, x = var_1738_cast_fp16)[name = tensor("transpose_249")]; - tensor x_205_cast_fp16 = matmul(transpose_x = x_205_transpose_x_0, transpose_y = x_205_transpose_y_0, x = q_with_bias_v_19_cast_fp16, y = op_1740_to_fp16_palettized)[name = tensor("x_205_cast_fp16")]; - tensor x_207_pad_0 = const()[name = tensor("x_207_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_207_mode_0 = const()[name = tensor("x_207_mode_0"), val = tensor("constant")]; - tensor const_104_to_fp16 = const()[name = tensor("const_104_to_fp16"), val = tensor(0x0p+0)]; - tensor x_207_cast_fp16 = pad(constant_val = const_104_to_fp16, mode = x_207_mode_0, pad = x_207_pad_0, x = x_205_cast_fp16)[name = tensor("x_207_cast_fp16")]; - tensor var_1748 = const()[name = tensor("op_1748"), val = tensor([1, 8, -1, 188])]; - tensor x_209_cast_fp16 = reshape(shape = var_1748, x = x_207_cast_fp16)[name = tensor("x_209_cast_fp16")]; - tensor var_1752_begin_0 = const()[name = tensor("op_1752_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_1752_end_0 = const()[name = tensor("op_1752_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_1752_end_mask_0 = const()[name = tensor("op_1752_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_1752_cast_fp16 = slice_by_index(begin = var_1752_begin_0, end = var_1752_end_0, end_mask = var_1752_end_mask_0, x = x_209_cast_fp16)[name = tensor("op_1752_cast_fp16")]; - tensor var_1753 = const()[name = tensor("op_1753"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_37_cast_fp16 = reshape(shape = var_1753, x = var_1752_cast_fp16)[name = tensor("matrix_bd_37_cast_fp16")]; - tensor matrix_ac_19_transpose_x_0 = const()[name = tensor("matrix_ac_19_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_19_transpose_y_0 = const()[name = tensor("matrix_ac_19_transpose_y_0"), val = tensor(false)]; - tensor transpose_114_perm_0 = const()[name = tensor("transpose_114_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_115_perm_0 = const()[name = tensor("transpose_115_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_115 = transpose(perm = transpose_115_perm_0, x = k_37_cast_fp16)[name = tensor("transpose_247")]; - tensor transpose_114 = transpose(perm = transpose_114_perm_0, x = var_1736_cast_fp16)[name = tensor("transpose_248")]; - tensor matrix_ac_19_cast_fp16 = matmul(transpose_x = matrix_ac_19_transpose_x_0, transpose_y = matrix_ac_19_transpose_y_0, x = transpose_114, y = transpose_115)[name = tensor("matrix_ac_19_cast_fp16")]; - tensor matrix_bd_39_begin_0 = const()[name = tensor("matrix_bd_39_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_39_end_0 = const()[name = tensor("matrix_bd_39_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_39_end_mask_0 = const()[name = tensor("matrix_bd_39_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_39_cast_fp16 = slice_by_index(begin = matrix_bd_39_begin_0, end = matrix_bd_39_end_0, end_mask = matrix_bd_39_end_mask_0, x = matrix_bd_37_cast_fp16)[name = tensor("matrix_bd_39_cast_fp16")]; - tensor var_1762_cast_fp16 = add(x = matrix_ac_19_cast_fp16, y = matrix_bd_39_cast_fp16)[name = tensor("op_1762_cast_fp16")]; - tensor _inversed_scores_37_y_0_to_fp16 = const()[name = tensor("_inversed_scores_37_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; - tensor _inversed_scores_37_cast_fp16 = mul(x = var_1762_cast_fp16, y = _inversed_scores_37_y_0_to_fp16)[name = tensor("_inversed_scores_37_cast_fp16")]; - tensor scores_39_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_37_cast_fp16, cond = mask_3)[name = tensor("scores_39_cast_fp16")]; - tensor var_1768_cast_fp16 = softmax(axis = var_30, x = scores_39_cast_fp16)[name = tensor("op_1768_cast_fp16")]; - tensor input_501_cast_fp16 = select(a = var_11_to_fp16, b = var_1768_cast_fp16, cond = mask_3)[name = tensor("input_501_cast_fp16")]; - tensor x_211_transpose_x_0 = const()[name = tensor("x_211_transpose_x_0"), val = tensor(false)]; - tensor x_211_transpose_y_0 = const()[name = tensor("x_211_transpose_y_0"), val = tensor(false)]; - tensor value_21_cast_fp16 = transpose(perm = value_21_perm_0, x = v_19_cast_fp16)[name = tensor("transpose_246")]; - tensor x_211_cast_fp16 = matmul(transpose_x = x_211_transpose_x_0, transpose_y = x_211_transpose_y_0, x = input_501_cast_fp16, y = value_21_cast_fp16)[name = tensor("x_211_cast_fp16")]; - tensor var_1772_perm_0 = const()[name = tensor("op_1772_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_1773 = const()[name = tensor("op_1773"), val = tensor([1, -1, 1024])]; - tensor var_1772_cast_fp16 = transpose(perm = var_1772_perm_0, x = x_211_cast_fp16)[name = tensor("transpose_245")]; - tensor input_503_cast_fp16 = reshape(shape = var_1773, x = var_1772_cast_fp16)[name = tensor("input_503_cast_fp16")]; - tensor module_layers_9_self_attn_linear_out_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(177939520))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(178726016))), name = tensor("module_layers_9_self_attn_linear_out_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_88_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_9_self_attn_linear_out_weight_to_fp16_palettized, x = input_503_cast_fp16)[name = tensor("linear_88_cast_fp16")]; - tensor input_507_cast_fp16 = add(x = input_499_cast_fp16, y = linear_88_cast_fp16)[name = tensor("input_507_cast_fp16")]; - tensor x_215_axes_0 = const()[name = tensor("x_215_axes_0"), val = tensor([-1])]; - tensor module_layers_9_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_9_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(178726208)))]; - tensor module_layers_9_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_9_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(178728320)))]; - tensor x_215_cast_fp16 = layer_norm(axes = x_215_axes_0, beta = module_layers_9_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_9_norm_conv_weight_to_fp16, x = input_507_cast_fp16)[name = tensor("x_215_cast_fp16")]; - tensor input_509_perm_0 = const()[name = tensor("input_509_perm_0"), val = tensor([0, 2, 1])]; - tensor input_511_pad_type_0 = const()[name = tensor("input_511_pad_type_0"), val = tensor("valid")]; - tensor input_511_strides_0 = const()[name = tensor("input_511_strides_0"), val = tensor([1])]; - tensor input_511_pad_0 = const()[name = tensor("input_511_pad_0"), val = tensor([0, 0])]; - tensor input_511_dilations_0 = const()[name = tensor("input_511_dilations_0"), val = tensor([1])]; - tensor input_511_groups_0 = const()[name = tensor("input_511_groups_0"), val = tensor(1)]; - tensor module_layers_9_conv_pointwise_conv1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(178730432))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(180303360))), name = tensor("module_layers_9_conv_pointwise_conv1_weight_to_fp16_palettized"), shape = tensor([2048, 1024, 1])]; - tensor input_509_cast_fp16 = transpose(perm = input_509_perm_0, x = x_215_cast_fp16)[name = tensor("transpose_244")]; - tensor input_511_cast_fp16 = conv(dilations = input_511_dilations_0, groups = input_511_groups_0, pad = input_511_pad_0, pad_type = input_511_pad_type_0, strides = input_511_strides_0, weight = module_layers_9_conv_pointwise_conv1_weight_to_fp16_palettized, x = input_509_cast_fp16)[name = tensor("input_511_cast_fp16")]; - tensor x_217_split_num_splits_0 = const()[name = tensor("x_217_split_num_splits_0"), val = tensor(2)]; - tensor x_217_split_axis_0 = const()[name = tensor("x_217_split_axis_0"), val = tensor(1)]; - tensor x_217_split_cast_fp16_0, tensor x_217_split_cast_fp16_1 = split(axis = x_217_split_axis_0, num_splits = x_217_split_num_splits_0, x = input_511_cast_fp16)[name = tensor("x_217_split_cast_fp16")]; - tensor x_217_split_1_sigmoid_cast_fp16 = sigmoid(x = x_217_split_cast_fp16_1)[name = tensor("x_217_split_1_sigmoid_cast_fp16")]; - tensor x_217_cast_fp16 = mul(x = x_217_split_cast_fp16_0, y = x_217_split_1_sigmoid_cast_fp16)[name = tensor("x_217_cast_fp16")]; - tensor input_513_cast_fp16 = select(a = var_11_to_fp16, b = x_217_cast_fp16, cond = var_328)[name = tensor("input_513_cast_fp16")]; - tensor input_515_pad_0 = const()[name = tensor("input_515_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_515_mode_0 = const()[name = tensor("input_515_mode_0"), val = tensor("constant")]; - tensor const_107_to_fp16 = const()[name = tensor("const_107_to_fp16"), val = tensor(0x0p+0)]; - tensor input_515_cast_fp16 = pad(constant_val = const_107_to_fp16, mode = input_515_mode_0, pad = input_515_pad_0, x = input_513_cast_fp16)[name = tensor("input_515_cast_fp16")]; - tensor input_517_pad_type_0 = const()[name = tensor("input_517_pad_type_0"), val = tensor("valid")]; - tensor input_517_groups_0 = const()[name = tensor("input_517_groups_0"), val = tensor(1024)]; - tensor input_517_strides_0 = const()[name = tensor("input_517_strides_0"), val = tensor([1])]; - tensor input_517_pad_0 = const()[name = tensor("input_517_pad_0"), val = tensor([0, 0])]; - tensor input_517_dilations_0 = const()[name = tensor("input_517_dilations_0"), val = tensor([1])]; - tensor const_266_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(180303552))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(180310528))), name = tensor("const_266_to_fp16_palettized"), shape = tensor([1024, 1, 9])]; - tensor const_267_to_fp16 = const()[name = tensor("const_267_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(180310720)))]; - tensor input_519_cast_fp16 = conv(bias = const_267_to_fp16, dilations = input_517_dilations_0, groups = input_517_groups_0, pad = input_517_pad_0, pad_type = input_517_pad_type_0, strides = input_517_strides_0, weight = const_266_to_fp16_palettized, x = input_515_cast_fp16)[name = tensor("input_519_cast_fp16")]; - tensor input_521_cast_fp16 = silu(x = input_519_cast_fp16)[name = tensor("input_521_cast_fp16")]; - tensor x_219_pad_type_0 = const()[name = tensor("x_219_pad_type_0"), val = tensor("valid")]; - tensor x_219_strides_0 = const()[name = tensor("x_219_strides_0"), val = tensor([1])]; - tensor x_219_pad_0 = const()[name = tensor("x_219_pad_0"), val = tensor([0, 0])]; - tensor x_219_dilations_0 = const()[name = tensor("x_219_dilations_0"), val = tensor([1])]; - tensor x_219_groups_0 = const()[name = tensor("x_219_groups_0"), val = tensor(1)]; - tensor module_layers_9_conv_pointwise_conv2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(180312832))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(181099328))), name = tensor("module_layers_9_conv_pointwise_conv2_weight_to_fp16_palettized"), shape = tensor([1024, 1024, 1])]; - tensor x_219_cast_fp16 = conv(dilations = x_219_dilations_0, groups = x_219_groups_0, pad = x_219_pad_0, pad_type = x_219_pad_type_0, strides = x_219_strides_0, weight = module_layers_9_conv_pointwise_conv2_weight_to_fp16_palettized, x = input_521_cast_fp16)[name = tensor("x_219_cast_fp16")]; - tensor input_523_perm_0 = const()[name = tensor("input_523_perm_0"), val = tensor([0, 2, 1])]; - tensor input_523_cast_fp16 = transpose(perm = input_523_perm_0, x = x_219_cast_fp16)[name = tensor("transpose_243")]; - tensor input_525_cast_fp16 = add(x = input_507_cast_fp16, y = input_523_cast_fp16)[name = tensor("input_525_cast_fp16")]; - tensor input_527_axes_0 = const()[name = tensor("input_527_axes_0"), val = tensor([-1])]; - tensor module_layers_9_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_9_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(181099520)))]; - tensor module_layers_9_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_9_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(181101632)))]; - tensor input_527_cast_fp16 = layer_norm(axes = input_527_axes_0, beta = module_layers_9_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_9_norm_feed_forward2_weight_to_fp16, x = input_525_cast_fp16)[name = tensor("input_527_cast_fp16")]; - tensor module_layers_9_feed_forward2_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(181103744))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184249536))), name = tensor("module_layers_9_feed_forward2_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_89_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_9_feed_forward2_linear1_weight_to_fp16_palettized, x = input_527_cast_fp16)[name = tensor("linear_89_cast_fp16")]; - tensor input_531_cast_fp16 = silu(x = linear_89_cast_fp16)[name = tensor("input_531_cast_fp16")]; - tensor module_layers_9_feed_forward2_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184249728))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(187395520))), name = tensor("module_layers_9_feed_forward2_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_90_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_9_feed_forward2_linear2_weight_to_fp16_palettized, x = input_531_cast_fp16)[name = tensor("linear_90_cast_fp16")]; - tensor var_1833_to_fp16 = const()[name = tensor("op_1833_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1834_cast_fp16 = mul(x = linear_90_cast_fp16, y = var_1833_to_fp16)[name = tensor("op_1834_cast_fp16")]; - tensor input_537_cast_fp16 = add(x = input_525_cast_fp16, y = var_1834_cast_fp16)[name = tensor("input_537_cast_fp16")]; - tensor input_539_axes_0 = const()[name = tensor("input_539_axes_0"), val = tensor([-1])]; - tensor module_layers_9_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_9_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(187395712)))]; - tensor module_layers_9_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_9_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(187397824)))]; - tensor input_539_cast_fp16 = layer_norm(axes = input_539_axes_0, beta = module_layers_9_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_9_norm_out_weight_to_fp16, x = input_537_cast_fp16)[name = tensor("input_539_cast_fp16")]; - tensor input_541_axes_0 = const()[name = tensor("input_541_axes_0"), val = tensor([-1])]; - tensor module_layers_10_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_10_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(187399936)))]; - tensor module_layers_10_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_10_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(187402048)))]; - tensor input_541_cast_fp16 = layer_norm(axes = input_541_axes_0, beta = module_layers_10_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_10_norm_feed_forward1_weight_to_fp16, x = input_539_cast_fp16)[name = tensor("input_541_cast_fp16")]; - tensor module_layers_10_feed_forward1_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(187404160))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190549952))), name = tensor("module_layers_10_feed_forward1_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_91_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_10_feed_forward1_linear1_weight_to_fp16_palettized, x = input_541_cast_fp16)[name = tensor("linear_91_cast_fp16")]; - tensor input_545_cast_fp16 = silu(x = linear_91_cast_fp16)[name = tensor("input_545_cast_fp16")]; - tensor module_layers_10_feed_forward1_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190550144))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(193695936))), name = tensor("module_layers_10_feed_forward1_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_92_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_10_feed_forward1_linear2_weight_to_fp16_palettized, x = input_545_cast_fp16)[name = tensor("linear_92_cast_fp16")]; - tensor var_1862_to_fp16 = const()[name = tensor("op_1862_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1863_cast_fp16 = mul(x = linear_92_cast_fp16, y = var_1862_to_fp16)[name = tensor("op_1863_cast_fp16")]; - tensor input_551_cast_fp16 = add(x = input_539_cast_fp16, y = var_1863_cast_fp16)[name = tensor("input_551_cast_fp16")]; - tensor query_21_axes_0 = const()[name = tensor("query_21_axes_0"), val = tensor([-1])]; - tensor module_layers_10_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_10_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(193696128)))]; - tensor module_layers_10_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_10_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(193698240)))]; - tensor query_21_cast_fp16 = layer_norm(axes = query_21_axes_0, beta = module_layers_10_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_10_norm_self_att_weight_to_fp16, x = input_551_cast_fp16)[name = tensor("query_21_cast_fp16")]; - tensor module_layers_10_self_attn_linear_q_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(193700352))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194486848))), name = tensor("module_layers_10_self_attn_linear_q_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_93_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_10_self_attn_linear_q_weight_to_fp16_palettized, x = query_21_cast_fp16)[name = tensor("linear_93_cast_fp16")]; - tensor var_1879 = const()[name = tensor("op_1879"), val = tensor([1, -1, 8, 128])]; - tensor q_61_cast_fp16 = reshape(shape = var_1879, x = linear_93_cast_fp16)[name = tensor("q_61_cast_fp16")]; - tensor module_layers_10_self_attn_linear_k_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194487040))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195273536))), name = tensor("module_layers_10_self_attn_linear_k_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_94_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_10_self_attn_linear_k_weight_to_fp16_palettized, x = query_21_cast_fp16)[name = tensor("linear_94_cast_fp16")]; - tensor var_1883 = const()[name = tensor("op_1883"), val = tensor([1, -1, 8, 128])]; - tensor k_41_cast_fp16 = reshape(shape = var_1883, x = linear_94_cast_fp16)[name = tensor("k_41_cast_fp16")]; - tensor module_layers_10_self_attn_linear_v_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195273728))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196060224))), name = tensor("module_layers_10_self_attn_linear_v_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_95_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_10_self_attn_linear_v_weight_to_fp16_palettized, x = query_21_cast_fp16)[name = tensor("linear_95_cast_fp16")]; - tensor var_1887 = const()[name = tensor("op_1887"), val = tensor([1, -1, 8, 128])]; - tensor v_21_cast_fp16 = reshape(shape = var_1887, x = linear_95_cast_fp16)[name = tensor("v_21_cast_fp16")]; - tensor value_23_perm_0 = const()[name = tensor("value_23_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor module_layers_10_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_10_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196060416)))]; - tensor var_1899_cast_fp16 = add(x = q_61_cast_fp16, y = module_layers_10_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1899_cast_fp16")]; - tensor module_layers_10_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_10_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196062528)))]; - tensor var_1901_cast_fp16 = add(x = q_61_cast_fp16, y = module_layers_10_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1901_cast_fp16")]; - tensor q_with_bias_v_21_perm_0 = const()[name = tensor("q_with_bias_v_21_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor x_227_transpose_x_0 = const()[name = tensor("x_227_transpose_x_0"), val = tensor(false)]; - tensor x_227_transpose_y_0 = const()[name = tensor("x_227_transpose_y_0"), val = tensor(false)]; - tensor op_1903_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196064640))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196352704))), name = tensor("op_1903_to_fp16_palettized"), shape = tensor([1, 8, 128, 375])]; - tensor q_with_bias_v_21_cast_fp16 = transpose(perm = q_with_bias_v_21_perm_0, x = var_1901_cast_fp16)[name = tensor("transpose_242")]; - tensor x_227_cast_fp16 = matmul(transpose_x = x_227_transpose_x_0, transpose_y = x_227_transpose_y_0, x = q_with_bias_v_21_cast_fp16, y = op_1903_to_fp16_palettized)[name = tensor("x_227_cast_fp16")]; - tensor x_229_pad_0 = const()[name = tensor("x_229_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_229_mode_0 = const()[name = tensor("x_229_mode_0"), val = tensor("constant")]; - tensor const_114_to_fp16 = const()[name = tensor("const_114_to_fp16"), val = tensor(0x0p+0)]; - tensor x_229_cast_fp16 = pad(constant_val = const_114_to_fp16, mode = x_229_mode_0, pad = x_229_pad_0, x = x_227_cast_fp16)[name = tensor("x_229_cast_fp16")]; - tensor var_1911 = const()[name = tensor("op_1911"), val = tensor([1, 8, -1, 188])]; - tensor x_231_cast_fp16 = reshape(shape = var_1911, x = x_229_cast_fp16)[name = tensor("x_231_cast_fp16")]; - tensor var_1915_begin_0 = const()[name = tensor("op_1915_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_1915_end_0 = const()[name = tensor("op_1915_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_1915_end_mask_0 = const()[name = tensor("op_1915_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_1915_cast_fp16 = slice_by_index(begin = var_1915_begin_0, end = var_1915_end_0, end_mask = var_1915_end_mask_0, x = x_231_cast_fp16)[name = tensor("op_1915_cast_fp16")]; - tensor var_1916 = const()[name = tensor("op_1916"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_41_cast_fp16 = reshape(shape = var_1916, x = var_1915_cast_fp16)[name = tensor("matrix_bd_41_cast_fp16")]; - tensor matrix_ac_21_transpose_x_0 = const()[name = tensor("matrix_ac_21_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_21_transpose_y_0 = const()[name = tensor("matrix_ac_21_transpose_y_0"), val = tensor(false)]; - tensor transpose_116_perm_0 = const()[name = tensor("transpose_116_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_117_perm_0 = const()[name = tensor("transpose_117_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_117 = transpose(perm = transpose_117_perm_0, x = k_41_cast_fp16)[name = tensor("transpose_240")]; - tensor transpose_116 = transpose(perm = transpose_116_perm_0, x = var_1899_cast_fp16)[name = tensor("transpose_241")]; - tensor matrix_ac_21_cast_fp16 = matmul(transpose_x = matrix_ac_21_transpose_x_0, transpose_y = matrix_ac_21_transpose_y_0, x = transpose_116, y = transpose_117)[name = tensor("matrix_ac_21_cast_fp16")]; - tensor matrix_bd_43_begin_0 = const()[name = tensor("matrix_bd_43_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_43_end_0 = const()[name = tensor("matrix_bd_43_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_43_end_mask_0 = const()[name = tensor("matrix_bd_43_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_43_cast_fp16 = slice_by_index(begin = matrix_bd_43_begin_0, end = matrix_bd_43_end_0, end_mask = matrix_bd_43_end_mask_0, x = matrix_bd_41_cast_fp16)[name = tensor("matrix_bd_43_cast_fp16")]; - tensor var_1925_cast_fp16 = add(x = matrix_ac_21_cast_fp16, y = matrix_bd_43_cast_fp16)[name = tensor("op_1925_cast_fp16")]; - tensor _inversed_scores_41_y_0_to_fp16 = const()[name = tensor("_inversed_scores_41_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; - tensor _inversed_scores_41_cast_fp16 = mul(x = var_1925_cast_fp16, y = _inversed_scores_41_y_0_to_fp16)[name = tensor("_inversed_scores_41_cast_fp16")]; - tensor scores_43_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_41_cast_fp16, cond = mask_3)[name = tensor("scores_43_cast_fp16")]; - tensor var_1931_cast_fp16 = softmax(axis = var_30, x = scores_43_cast_fp16)[name = tensor("op_1931_cast_fp16")]; - tensor input_553_cast_fp16 = select(a = var_11_to_fp16, b = var_1931_cast_fp16, cond = mask_3)[name = tensor("input_553_cast_fp16")]; - tensor x_233_transpose_x_0 = const()[name = tensor("x_233_transpose_x_0"), val = tensor(false)]; - tensor x_233_transpose_y_0 = const()[name = tensor("x_233_transpose_y_0"), val = tensor(false)]; - tensor value_23_cast_fp16 = transpose(perm = value_23_perm_0, x = v_21_cast_fp16)[name = tensor("transpose_239")]; - tensor x_233_cast_fp16 = matmul(transpose_x = x_233_transpose_x_0, transpose_y = x_233_transpose_y_0, x = input_553_cast_fp16, y = value_23_cast_fp16)[name = tensor("x_233_cast_fp16")]; - tensor var_1935_perm_0 = const()[name = tensor("op_1935_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_1936 = const()[name = tensor("op_1936"), val = tensor([1, -1, 1024])]; - tensor var_1935_cast_fp16 = transpose(perm = var_1935_perm_0, x = x_233_cast_fp16)[name = tensor("transpose_238")]; - tensor input_555_cast_fp16 = reshape(shape = var_1936, x = var_1935_cast_fp16)[name = tensor("input_555_cast_fp16")]; - tensor module_layers_10_self_attn_linear_out_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196352896))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(197139392))), name = tensor("module_layers_10_self_attn_linear_out_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_97_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_10_self_attn_linear_out_weight_to_fp16_palettized, x = input_555_cast_fp16)[name = tensor("linear_97_cast_fp16")]; - tensor input_559_cast_fp16 = add(x = input_551_cast_fp16, y = linear_97_cast_fp16)[name = tensor("input_559_cast_fp16")]; - tensor x_237_axes_0 = const()[name = tensor("x_237_axes_0"), val = tensor([-1])]; - tensor module_layers_10_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_10_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(197139584)))]; - tensor module_layers_10_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_10_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(197141696)))]; - tensor x_237_cast_fp16 = layer_norm(axes = x_237_axes_0, beta = module_layers_10_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_10_norm_conv_weight_to_fp16, x = input_559_cast_fp16)[name = tensor("x_237_cast_fp16")]; - tensor input_561_perm_0 = const()[name = tensor("input_561_perm_0"), val = tensor([0, 2, 1])]; - tensor input_563_pad_type_0 = const()[name = tensor("input_563_pad_type_0"), val = tensor("valid")]; - tensor input_563_strides_0 = const()[name = tensor("input_563_strides_0"), val = tensor([1])]; - tensor input_563_pad_0 = const()[name = tensor("input_563_pad_0"), val = tensor([0, 0])]; - tensor input_563_dilations_0 = const()[name = tensor("input_563_dilations_0"), val = tensor([1])]; - tensor input_563_groups_0 = const()[name = tensor("input_563_groups_0"), val = tensor(1)]; - tensor module_layers_10_conv_pointwise_conv1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(197143808))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198716736))), name = tensor("module_layers_10_conv_pointwise_conv1_weight_to_fp16_palettized"), shape = tensor([2048, 1024, 1])]; - tensor input_561_cast_fp16 = transpose(perm = input_561_perm_0, x = x_237_cast_fp16)[name = tensor("transpose_237")]; - tensor input_563_cast_fp16 = conv(dilations = input_563_dilations_0, groups = input_563_groups_0, pad = input_563_pad_0, pad_type = input_563_pad_type_0, strides = input_563_strides_0, weight = module_layers_10_conv_pointwise_conv1_weight_to_fp16_palettized, x = input_561_cast_fp16)[name = tensor("input_563_cast_fp16")]; - tensor x_239_split_num_splits_0 = const()[name = tensor("x_239_split_num_splits_0"), val = tensor(2)]; - tensor x_239_split_axis_0 = const()[name = tensor("x_239_split_axis_0"), val = tensor(1)]; - tensor x_239_split_cast_fp16_0, tensor x_239_split_cast_fp16_1 = split(axis = x_239_split_axis_0, num_splits = x_239_split_num_splits_0, x = input_563_cast_fp16)[name = tensor("x_239_split_cast_fp16")]; - tensor x_239_split_1_sigmoid_cast_fp16 = sigmoid(x = x_239_split_cast_fp16_1)[name = tensor("x_239_split_1_sigmoid_cast_fp16")]; - tensor x_239_cast_fp16 = mul(x = x_239_split_cast_fp16_0, y = x_239_split_1_sigmoid_cast_fp16)[name = tensor("x_239_cast_fp16")]; - tensor input_565_cast_fp16 = select(a = var_11_to_fp16, b = x_239_cast_fp16, cond = var_328)[name = tensor("input_565_cast_fp16")]; - tensor input_567_pad_0 = const()[name = tensor("input_567_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_567_mode_0 = const()[name = tensor("input_567_mode_0"), val = tensor("constant")]; - tensor const_117_to_fp16 = const()[name = tensor("const_117_to_fp16"), val = tensor(0x0p+0)]; - tensor input_567_cast_fp16 = pad(constant_val = const_117_to_fp16, mode = input_567_mode_0, pad = input_567_pad_0, x = input_565_cast_fp16)[name = tensor("input_567_cast_fp16")]; - tensor input_569_pad_type_0 = const()[name = tensor("input_569_pad_type_0"), val = tensor("valid")]; - tensor input_569_groups_0 = const()[name = tensor("input_569_groups_0"), val = tensor(1024)]; - tensor input_569_strides_0 = const()[name = tensor("input_569_strides_0"), val = tensor([1])]; - tensor input_569_pad_0 = const()[name = tensor("input_569_pad_0"), val = tensor([0, 0])]; - tensor input_569_dilations_0 = const()[name = tensor("input_569_dilations_0"), val = tensor([1])]; - tensor const_268_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198716928))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198723904))), name = tensor("const_268_to_fp16_palettized"), shape = tensor([1024, 1, 9])]; - tensor const_269_to_fp16 = const()[name = tensor("const_269_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198724096)))]; - tensor input_571_cast_fp16 = conv(bias = const_269_to_fp16, dilations = input_569_dilations_0, groups = input_569_groups_0, pad = input_569_pad_0, pad_type = input_569_pad_type_0, strides = input_569_strides_0, weight = const_268_to_fp16_palettized, x = input_567_cast_fp16)[name = tensor("input_571_cast_fp16")]; - tensor input_573_cast_fp16 = silu(x = input_571_cast_fp16)[name = tensor("input_573_cast_fp16")]; - tensor x_241_pad_type_0 = const()[name = tensor("x_241_pad_type_0"), val = tensor("valid")]; - tensor x_241_strides_0 = const()[name = tensor("x_241_strides_0"), val = tensor([1])]; - tensor x_241_pad_0 = const()[name = tensor("x_241_pad_0"), val = tensor([0, 0])]; - tensor x_241_dilations_0 = const()[name = tensor("x_241_dilations_0"), val = tensor([1])]; - tensor x_241_groups_0 = const()[name = tensor("x_241_groups_0"), val = tensor(1)]; - tensor module_layers_10_conv_pointwise_conv2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198726208))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(199512704))), name = tensor("module_layers_10_conv_pointwise_conv2_weight_to_fp16_palettized"), shape = tensor([1024, 1024, 1])]; - tensor x_241_cast_fp16 = conv(dilations = x_241_dilations_0, groups = x_241_groups_0, pad = x_241_pad_0, pad_type = x_241_pad_type_0, strides = x_241_strides_0, weight = module_layers_10_conv_pointwise_conv2_weight_to_fp16_palettized, x = input_573_cast_fp16)[name = tensor("x_241_cast_fp16")]; - tensor input_575_perm_0 = const()[name = tensor("input_575_perm_0"), val = tensor([0, 2, 1])]; - tensor input_575_cast_fp16 = transpose(perm = input_575_perm_0, x = x_241_cast_fp16)[name = tensor("transpose_236")]; - tensor input_577_cast_fp16 = add(x = input_559_cast_fp16, y = input_575_cast_fp16)[name = tensor("input_577_cast_fp16")]; - tensor input_579_axes_0 = const()[name = tensor("input_579_axes_0"), val = tensor([-1])]; - tensor module_layers_10_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_10_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(199512896)))]; - tensor module_layers_10_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_10_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(199515008)))]; - tensor input_579_cast_fp16 = layer_norm(axes = input_579_axes_0, beta = module_layers_10_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_10_norm_feed_forward2_weight_to_fp16, x = input_577_cast_fp16)[name = tensor("input_579_cast_fp16")]; - tensor module_layers_10_feed_forward2_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(199517120))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202662912))), name = tensor("module_layers_10_feed_forward2_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_98_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_10_feed_forward2_linear1_weight_to_fp16_palettized, x = input_579_cast_fp16)[name = tensor("linear_98_cast_fp16")]; - tensor input_583_cast_fp16 = silu(x = linear_98_cast_fp16)[name = tensor("input_583_cast_fp16")]; - tensor module_layers_10_feed_forward2_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202663104))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(205808896))), name = tensor("module_layers_10_feed_forward2_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_99_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_10_feed_forward2_linear2_weight_to_fp16_palettized, x = input_583_cast_fp16)[name = tensor("linear_99_cast_fp16")]; - tensor var_1996_to_fp16 = const()[name = tensor("op_1996_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1997_cast_fp16 = mul(x = linear_99_cast_fp16, y = var_1996_to_fp16)[name = tensor("op_1997_cast_fp16")]; - tensor input_589_cast_fp16 = add(x = input_577_cast_fp16, y = var_1997_cast_fp16)[name = tensor("input_589_cast_fp16")]; - tensor input_591_axes_0 = const()[name = tensor("input_591_axes_0"), val = tensor([-1])]; - tensor module_layers_10_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_10_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(205809088)))]; - tensor module_layers_10_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_10_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(205811200)))]; - tensor input_591_cast_fp16 = layer_norm(axes = input_591_axes_0, beta = module_layers_10_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_10_norm_out_weight_to_fp16, x = input_589_cast_fp16)[name = tensor("input_591_cast_fp16")]; - tensor input_593_axes_0 = const()[name = tensor("input_593_axes_0"), val = tensor([-1])]; - tensor module_layers_11_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_11_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(205813312)))]; - tensor module_layers_11_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_11_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(205815424)))]; - tensor input_593_cast_fp16 = layer_norm(axes = input_593_axes_0, beta = module_layers_11_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_11_norm_feed_forward1_weight_to_fp16, x = input_591_cast_fp16)[name = tensor("input_593_cast_fp16")]; - tensor module_layers_11_feed_forward1_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(205817536))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208963328))), name = tensor("module_layers_11_feed_forward1_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_100_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_11_feed_forward1_linear1_weight_to_fp16_palettized, x = input_593_cast_fp16)[name = tensor("linear_100_cast_fp16")]; - tensor input_597_cast_fp16 = silu(x = linear_100_cast_fp16)[name = tensor("input_597_cast_fp16")]; - tensor module_layers_11_feed_forward1_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208963520))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(212109312))), name = tensor("module_layers_11_feed_forward1_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_101_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_11_feed_forward1_linear2_weight_to_fp16_palettized, x = input_597_cast_fp16)[name = tensor("linear_101_cast_fp16")]; - tensor var_2025_to_fp16 = const()[name = tensor("op_2025_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2026_cast_fp16 = mul(x = linear_101_cast_fp16, y = var_2025_to_fp16)[name = tensor("op_2026_cast_fp16")]; - tensor input_603_cast_fp16 = add(x = input_591_cast_fp16, y = var_2026_cast_fp16)[name = tensor("input_603_cast_fp16")]; - tensor query_23_axes_0 = const()[name = tensor("query_23_axes_0"), val = tensor([-1])]; - tensor module_layers_11_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_11_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(212109504)))]; - tensor module_layers_11_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_11_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(212111616)))]; - tensor query_23_cast_fp16 = layer_norm(axes = query_23_axes_0, beta = module_layers_11_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_11_norm_self_att_weight_to_fp16, x = input_603_cast_fp16)[name = tensor("query_23_cast_fp16")]; - tensor module_layers_11_self_attn_linear_q_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(212113728))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(212900224))), name = tensor("module_layers_11_self_attn_linear_q_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_102_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_11_self_attn_linear_q_weight_to_fp16_palettized, x = query_23_cast_fp16)[name = tensor("linear_102_cast_fp16")]; - tensor var_2042 = const()[name = tensor("op_2042"), val = tensor([1, -1, 8, 128])]; - tensor q_67_cast_fp16 = reshape(shape = var_2042, x = linear_102_cast_fp16)[name = tensor("q_67_cast_fp16")]; - tensor module_layers_11_self_attn_linear_k_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(212900416))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(213686912))), name = tensor("module_layers_11_self_attn_linear_k_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_103_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_11_self_attn_linear_k_weight_to_fp16_palettized, x = query_23_cast_fp16)[name = tensor("linear_103_cast_fp16")]; - tensor var_2046 = const()[name = tensor("op_2046"), val = tensor([1, -1, 8, 128])]; - tensor k_45_cast_fp16 = reshape(shape = var_2046, x = linear_103_cast_fp16)[name = tensor("k_45_cast_fp16")]; - tensor module_layers_11_self_attn_linear_v_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(213687104))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(214473600))), name = tensor("module_layers_11_self_attn_linear_v_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_104_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_11_self_attn_linear_v_weight_to_fp16_palettized, x = query_23_cast_fp16)[name = tensor("linear_104_cast_fp16")]; - tensor var_2050 = const()[name = tensor("op_2050"), val = tensor([1, -1, 8, 128])]; - tensor v_23_cast_fp16 = reshape(shape = var_2050, x = linear_104_cast_fp16)[name = tensor("v_23_cast_fp16")]; - tensor value_25_perm_0 = const()[name = tensor("value_25_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor module_layers_11_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_11_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(214473792)))]; - tensor var_2062_cast_fp16 = add(x = q_67_cast_fp16, y = module_layers_11_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2062_cast_fp16")]; - tensor module_layers_11_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_11_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(214475904)))]; - tensor var_2064_cast_fp16 = add(x = q_67_cast_fp16, y = module_layers_11_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2064_cast_fp16")]; - tensor q_with_bias_v_23_perm_0 = const()[name = tensor("q_with_bias_v_23_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor x_249_transpose_x_0 = const()[name = tensor("x_249_transpose_x_0"), val = tensor(false)]; - tensor x_249_transpose_y_0 = const()[name = tensor("x_249_transpose_y_0"), val = tensor(false)]; - tensor op_2066_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(214478016))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(214766080))), name = tensor("op_2066_to_fp16_palettized"), shape = tensor([1, 8, 128, 375])]; - tensor q_with_bias_v_23_cast_fp16 = transpose(perm = q_with_bias_v_23_perm_0, x = var_2064_cast_fp16)[name = tensor("transpose_235")]; - tensor x_249_cast_fp16 = matmul(transpose_x = x_249_transpose_x_0, transpose_y = x_249_transpose_y_0, x = q_with_bias_v_23_cast_fp16, y = op_2066_to_fp16_palettized)[name = tensor("x_249_cast_fp16")]; - tensor x_251_pad_0 = const()[name = tensor("x_251_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_251_mode_0 = const()[name = tensor("x_251_mode_0"), val = tensor("constant")]; - tensor const_124_to_fp16 = const()[name = tensor("const_124_to_fp16"), val = tensor(0x0p+0)]; - tensor x_251_cast_fp16 = pad(constant_val = const_124_to_fp16, mode = x_251_mode_0, pad = x_251_pad_0, x = x_249_cast_fp16)[name = tensor("x_251_cast_fp16")]; - tensor var_2074 = const()[name = tensor("op_2074"), val = tensor([1, 8, -1, 188])]; - tensor x_253_cast_fp16 = reshape(shape = var_2074, x = x_251_cast_fp16)[name = tensor("x_253_cast_fp16")]; - tensor var_2078_begin_0 = const()[name = tensor("op_2078_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_2078_end_0 = const()[name = tensor("op_2078_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_2078_end_mask_0 = const()[name = tensor("op_2078_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_2078_cast_fp16 = slice_by_index(begin = var_2078_begin_0, end = var_2078_end_0, end_mask = var_2078_end_mask_0, x = x_253_cast_fp16)[name = tensor("op_2078_cast_fp16")]; - tensor var_2079 = const()[name = tensor("op_2079"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_45_cast_fp16 = reshape(shape = var_2079, x = var_2078_cast_fp16)[name = tensor("matrix_bd_45_cast_fp16")]; - tensor matrix_ac_23_transpose_x_0 = const()[name = tensor("matrix_ac_23_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_23_transpose_y_0 = const()[name = tensor("matrix_ac_23_transpose_y_0"), val = tensor(false)]; - tensor transpose_118_perm_0 = const()[name = tensor("transpose_118_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_119_perm_0 = const()[name = tensor("transpose_119_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_119 = transpose(perm = transpose_119_perm_0, x = k_45_cast_fp16)[name = tensor("transpose_233")]; - tensor transpose_118 = transpose(perm = transpose_118_perm_0, x = var_2062_cast_fp16)[name = tensor("transpose_234")]; - tensor matrix_ac_23_cast_fp16 = matmul(transpose_x = matrix_ac_23_transpose_x_0, transpose_y = matrix_ac_23_transpose_y_0, x = transpose_118, y = transpose_119)[name = tensor("matrix_ac_23_cast_fp16")]; - tensor matrix_bd_47_begin_0 = const()[name = tensor("matrix_bd_47_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_47_end_0 = const()[name = tensor("matrix_bd_47_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_47_end_mask_0 = const()[name = tensor("matrix_bd_47_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_47_cast_fp16 = slice_by_index(begin = matrix_bd_47_begin_0, end = matrix_bd_47_end_0, end_mask = matrix_bd_47_end_mask_0, x = matrix_bd_45_cast_fp16)[name = tensor("matrix_bd_47_cast_fp16")]; - tensor var_2088_cast_fp16 = add(x = matrix_ac_23_cast_fp16, y = matrix_bd_47_cast_fp16)[name = tensor("op_2088_cast_fp16")]; - tensor _inversed_scores_45_y_0_to_fp16 = const()[name = tensor("_inversed_scores_45_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; - tensor _inversed_scores_45_cast_fp16 = mul(x = var_2088_cast_fp16, y = _inversed_scores_45_y_0_to_fp16)[name = tensor("_inversed_scores_45_cast_fp16")]; - tensor scores_47_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_45_cast_fp16, cond = mask_3)[name = tensor("scores_47_cast_fp16")]; - tensor var_2094_cast_fp16 = softmax(axis = var_30, x = scores_47_cast_fp16)[name = tensor("op_2094_cast_fp16")]; - tensor input_605_cast_fp16 = select(a = var_11_to_fp16, b = var_2094_cast_fp16, cond = mask_3)[name = tensor("input_605_cast_fp16")]; - tensor x_255_transpose_x_0 = const()[name = tensor("x_255_transpose_x_0"), val = tensor(false)]; - tensor x_255_transpose_y_0 = const()[name = tensor("x_255_transpose_y_0"), val = tensor(false)]; - tensor value_25_cast_fp16 = transpose(perm = value_25_perm_0, x = v_23_cast_fp16)[name = tensor("transpose_232")]; - tensor x_255_cast_fp16 = matmul(transpose_x = x_255_transpose_x_0, transpose_y = x_255_transpose_y_0, x = input_605_cast_fp16, y = value_25_cast_fp16)[name = tensor("x_255_cast_fp16")]; - tensor var_2098_perm_0 = const()[name = tensor("op_2098_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_2099 = const()[name = tensor("op_2099"), val = tensor([1, -1, 1024])]; - tensor var_2098_cast_fp16 = transpose(perm = var_2098_perm_0, x = x_255_cast_fp16)[name = tensor("transpose_231")]; - tensor input_607_cast_fp16 = reshape(shape = var_2099, x = var_2098_cast_fp16)[name = tensor("input_607_cast_fp16")]; - tensor module_layers_11_self_attn_linear_out_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(214766272))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(215552768))), name = tensor("module_layers_11_self_attn_linear_out_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_106_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_11_self_attn_linear_out_weight_to_fp16_palettized, x = input_607_cast_fp16)[name = tensor("linear_106_cast_fp16")]; - tensor input_611_cast_fp16 = add(x = input_603_cast_fp16, y = linear_106_cast_fp16)[name = tensor("input_611_cast_fp16")]; - tensor x_259_axes_0 = const()[name = tensor("x_259_axes_0"), val = tensor([-1])]; - tensor module_layers_11_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_11_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(215552960)))]; - tensor module_layers_11_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_11_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(215555072)))]; - tensor x_259_cast_fp16 = layer_norm(axes = x_259_axes_0, beta = module_layers_11_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_11_norm_conv_weight_to_fp16, x = input_611_cast_fp16)[name = tensor("x_259_cast_fp16")]; - tensor input_613_perm_0 = const()[name = tensor("input_613_perm_0"), val = tensor([0, 2, 1])]; - tensor input_615_pad_type_0 = const()[name = tensor("input_615_pad_type_0"), val = tensor("valid")]; - tensor input_615_strides_0 = const()[name = tensor("input_615_strides_0"), val = tensor([1])]; - tensor input_615_pad_0 = const()[name = tensor("input_615_pad_0"), val = tensor([0, 0])]; - tensor input_615_dilations_0 = const()[name = tensor("input_615_dilations_0"), val = tensor([1])]; - tensor input_615_groups_0 = const()[name = tensor("input_615_groups_0"), val = tensor(1)]; - tensor module_layers_11_conv_pointwise_conv1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(215557184))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(217130112))), name = tensor("module_layers_11_conv_pointwise_conv1_weight_to_fp16_palettized"), shape = tensor([2048, 1024, 1])]; - tensor input_613_cast_fp16 = transpose(perm = input_613_perm_0, x = x_259_cast_fp16)[name = tensor("transpose_230")]; - tensor input_615_cast_fp16 = conv(dilations = input_615_dilations_0, groups = input_615_groups_0, pad = input_615_pad_0, pad_type = input_615_pad_type_0, strides = input_615_strides_0, weight = module_layers_11_conv_pointwise_conv1_weight_to_fp16_palettized, x = input_613_cast_fp16)[name = tensor("input_615_cast_fp16")]; - tensor x_261_split_num_splits_0 = const()[name = tensor("x_261_split_num_splits_0"), val = tensor(2)]; - tensor x_261_split_axis_0 = const()[name = tensor("x_261_split_axis_0"), val = tensor(1)]; - tensor x_261_split_cast_fp16_0, tensor x_261_split_cast_fp16_1 = split(axis = x_261_split_axis_0, num_splits = x_261_split_num_splits_0, x = input_615_cast_fp16)[name = tensor("x_261_split_cast_fp16")]; - tensor x_261_split_1_sigmoid_cast_fp16 = sigmoid(x = x_261_split_cast_fp16_1)[name = tensor("x_261_split_1_sigmoid_cast_fp16")]; - tensor x_261_cast_fp16 = mul(x = x_261_split_cast_fp16_0, y = x_261_split_1_sigmoid_cast_fp16)[name = tensor("x_261_cast_fp16")]; - tensor input_617_cast_fp16 = select(a = var_11_to_fp16, b = x_261_cast_fp16, cond = var_328)[name = tensor("input_617_cast_fp16")]; - tensor input_619_pad_0 = const()[name = tensor("input_619_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_619_mode_0 = const()[name = tensor("input_619_mode_0"), val = tensor("constant")]; - tensor const_127_to_fp16 = const()[name = tensor("const_127_to_fp16"), val = tensor(0x0p+0)]; - tensor input_619_cast_fp16 = pad(constant_val = const_127_to_fp16, mode = input_619_mode_0, pad = input_619_pad_0, x = input_617_cast_fp16)[name = tensor("input_619_cast_fp16")]; - tensor input_621_pad_type_0 = const()[name = tensor("input_621_pad_type_0"), val = tensor("valid")]; - tensor input_621_groups_0 = const()[name = tensor("input_621_groups_0"), val = tensor(1024)]; - tensor input_621_strides_0 = const()[name = tensor("input_621_strides_0"), val = tensor([1])]; - tensor input_621_pad_0 = const()[name = tensor("input_621_pad_0"), val = tensor([0, 0])]; - tensor input_621_dilations_0 = const()[name = tensor("input_621_dilations_0"), val = tensor([1])]; - tensor const_270_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(217130304))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(217137280))), name = tensor("const_270_to_fp16_palettized"), shape = tensor([1024, 1, 9])]; - tensor const_271_to_fp16 = const()[name = tensor("const_271_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(217137472)))]; - tensor input_623_cast_fp16 = conv(bias = const_271_to_fp16, dilations = input_621_dilations_0, groups = input_621_groups_0, pad = input_621_pad_0, pad_type = input_621_pad_type_0, strides = input_621_strides_0, weight = const_270_to_fp16_palettized, x = input_619_cast_fp16)[name = tensor("input_623_cast_fp16")]; - tensor input_625_cast_fp16 = silu(x = input_623_cast_fp16)[name = tensor("input_625_cast_fp16")]; - tensor x_263_pad_type_0 = const()[name = tensor("x_263_pad_type_0"), val = tensor("valid")]; - tensor x_263_strides_0 = const()[name = tensor("x_263_strides_0"), val = tensor([1])]; - tensor x_263_pad_0 = const()[name = tensor("x_263_pad_0"), val = tensor([0, 0])]; - tensor x_263_dilations_0 = const()[name = tensor("x_263_dilations_0"), val = tensor([1])]; - tensor x_263_groups_0 = const()[name = tensor("x_263_groups_0"), val = tensor(1)]; - tensor module_layers_11_conv_pointwise_conv2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(217139584))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(217926080))), name = tensor("module_layers_11_conv_pointwise_conv2_weight_to_fp16_palettized"), shape = tensor([1024, 1024, 1])]; - tensor x_263_cast_fp16 = conv(dilations = x_263_dilations_0, groups = x_263_groups_0, pad = x_263_pad_0, pad_type = x_263_pad_type_0, strides = x_263_strides_0, weight = module_layers_11_conv_pointwise_conv2_weight_to_fp16_palettized, x = input_625_cast_fp16)[name = tensor("x_263_cast_fp16")]; - tensor input_627_perm_0 = const()[name = tensor("input_627_perm_0"), val = tensor([0, 2, 1])]; - tensor input_627_cast_fp16 = transpose(perm = input_627_perm_0, x = x_263_cast_fp16)[name = tensor("transpose_229")]; - tensor input_629_cast_fp16 = add(x = input_611_cast_fp16, y = input_627_cast_fp16)[name = tensor("input_629_cast_fp16")]; - tensor input_631_axes_0 = const()[name = tensor("input_631_axes_0"), val = tensor([-1])]; - tensor module_layers_11_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_11_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(217926272)))]; - tensor module_layers_11_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_11_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(217928384)))]; - tensor input_631_cast_fp16 = layer_norm(axes = input_631_axes_0, beta = module_layers_11_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_11_norm_feed_forward2_weight_to_fp16, x = input_629_cast_fp16)[name = tensor("input_631_cast_fp16")]; - tensor module_layers_11_feed_forward2_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(217930496))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(221076288))), name = tensor("module_layers_11_feed_forward2_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_107_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_11_feed_forward2_linear1_weight_to_fp16_palettized, x = input_631_cast_fp16)[name = tensor("linear_107_cast_fp16")]; - tensor input_635_cast_fp16 = silu(x = linear_107_cast_fp16)[name = tensor("input_635_cast_fp16")]; - tensor module_layers_11_feed_forward2_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(221076480))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(224222272))), name = tensor("module_layers_11_feed_forward2_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_108_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_11_feed_forward2_linear2_weight_to_fp16_palettized, x = input_635_cast_fp16)[name = tensor("linear_108_cast_fp16")]; - tensor var_2159_to_fp16 = const()[name = tensor("op_2159_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2160_cast_fp16 = mul(x = linear_108_cast_fp16, y = var_2159_to_fp16)[name = tensor("op_2160_cast_fp16")]; - tensor input_641_cast_fp16 = add(x = input_629_cast_fp16, y = var_2160_cast_fp16)[name = tensor("input_641_cast_fp16")]; - tensor input_643_axes_0 = const()[name = tensor("input_643_axes_0"), val = tensor([-1])]; - tensor module_layers_11_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_11_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(224222464)))]; - tensor module_layers_11_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_11_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(224224576)))]; - tensor input_643_cast_fp16 = layer_norm(axes = input_643_axes_0, beta = module_layers_11_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_11_norm_out_weight_to_fp16, x = input_641_cast_fp16)[name = tensor("input_643_cast_fp16")]; - tensor input_645_axes_0 = const()[name = tensor("input_645_axes_0"), val = tensor([-1])]; - tensor module_layers_12_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_12_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(224226688)))]; - tensor module_layers_12_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_12_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(224228800)))]; - tensor input_645_cast_fp16 = layer_norm(axes = input_645_axes_0, beta = module_layers_12_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_12_norm_feed_forward1_weight_to_fp16, x = input_643_cast_fp16)[name = tensor("input_645_cast_fp16")]; - tensor module_layers_12_feed_forward1_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(224230912))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(227376704))), name = tensor("module_layers_12_feed_forward1_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_109_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_12_feed_forward1_linear1_weight_to_fp16_palettized, x = input_645_cast_fp16)[name = tensor("linear_109_cast_fp16")]; - tensor input_649_cast_fp16 = silu(x = linear_109_cast_fp16)[name = tensor("input_649_cast_fp16")]; - tensor module_layers_12_feed_forward1_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(227376896))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(230522688))), name = tensor("module_layers_12_feed_forward1_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_110_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_12_feed_forward1_linear2_weight_to_fp16_palettized, x = input_649_cast_fp16)[name = tensor("linear_110_cast_fp16")]; - tensor var_2188_to_fp16 = const()[name = tensor("op_2188_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2189_cast_fp16 = mul(x = linear_110_cast_fp16, y = var_2188_to_fp16)[name = tensor("op_2189_cast_fp16")]; - tensor input_655_cast_fp16 = add(x = input_643_cast_fp16, y = var_2189_cast_fp16)[name = tensor("input_655_cast_fp16")]; - tensor query_25_axes_0 = const()[name = tensor("query_25_axes_0"), val = tensor([-1])]; - tensor module_layers_12_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_12_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(230522880)))]; - tensor module_layers_12_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_12_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(230524992)))]; - tensor query_25_cast_fp16 = layer_norm(axes = query_25_axes_0, beta = module_layers_12_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_12_norm_self_att_weight_to_fp16, x = input_655_cast_fp16)[name = tensor("query_25_cast_fp16")]; - tensor module_layers_12_self_attn_linear_q_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(230527104))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(231313600))), name = tensor("module_layers_12_self_attn_linear_q_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_111_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_12_self_attn_linear_q_weight_to_fp16_palettized, x = query_25_cast_fp16)[name = tensor("linear_111_cast_fp16")]; - tensor var_2205 = const()[name = tensor("op_2205"), val = tensor([1, -1, 8, 128])]; - tensor q_73_cast_fp16 = reshape(shape = var_2205, x = linear_111_cast_fp16)[name = tensor("q_73_cast_fp16")]; - tensor module_layers_12_self_attn_linear_k_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(231313792))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(232100288))), name = tensor("module_layers_12_self_attn_linear_k_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_112_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_12_self_attn_linear_k_weight_to_fp16_palettized, x = query_25_cast_fp16)[name = tensor("linear_112_cast_fp16")]; - tensor var_2209 = const()[name = tensor("op_2209"), val = tensor([1, -1, 8, 128])]; - tensor k_49_cast_fp16 = reshape(shape = var_2209, x = linear_112_cast_fp16)[name = tensor("k_49_cast_fp16")]; - tensor module_layers_12_self_attn_linear_v_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(232100480))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(232886976))), name = tensor("module_layers_12_self_attn_linear_v_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_113_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_12_self_attn_linear_v_weight_to_fp16_palettized, x = query_25_cast_fp16)[name = tensor("linear_113_cast_fp16")]; - tensor var_2213 = const()[name = tensor("op_2213"), val = tensor([1, -1, 8, 128])]; - tensor v_25_cast_fp16 = reshape(shape = var_2213, x = linear_113_cast_fp16)[name = tensor("v_25_cast_fp16")]; - tensor value_27_perm_0 = const()[name = tensor("value_27_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor module_layers_12_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_12_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(232887168)))]; - tensor var_2225_cast_fp16 = add(x = q_73_cast_fp16, y = module_layers_12_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2225_cast_fp16")]; - tensor module_layers_12_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_12_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(232889280)))]; - tensor var_2227_cast_fp16 = add(x = q_73_cast_fp16, y = module_layers_12_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2227_cast_fp16")]; - tensor q_with_bias_v_25_perm_0 = const()[name = tensor("q_with_bias_v_25_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor x_271_transpose_x_0 = const()[name = tensor("x_271_transpose_x_0"), val = tensor(false)]; - tensor x_271_transpose_y_0 = const()[name = tensor("x_271_transpose_y_0"), val = tensor(false)]; - tensor op_2229_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(232891392))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(233179456))), name = tensor("op_2229_to_fp16_palettized"), shape = tensor([1, 8, 128, 375])]; - tensor q_with_bias_v_25_cast_fp16 = transpose(perm = q_with_bias_v_25_perm_0, x = var_2227_cast_fp16)[name = tensor("transpose_228")]; - tensor x_271_cast_fp16 = matmul(transpose_x = x_271_transpose_x_0, transpose_y = x_271_transpose_y_0, x = q_with_bias_v_25_cast_fp16, y = op_2229_to_fp16_palettized)[name = tensor("x_271_cast_fp16")]; - tensor x_273_pad_0 = const()[name = tensor("x_273_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_273_mode_0 = const()[name = tensor("x_273_mode_0"), val = tensor("constant")]; - tensor const_134_to_fp16 = const()[name = tensor("const_134_to_fp16"), val = tensor(0x0p+0)]; - tensor x_273_cast_fp16 = pad(constant_val = const_134_to_fp16, mode = x_273_mode_0, pad = x_273_pad_0, x = x_271_cast_fp16)[name = tensor("x_273_cast_fp16")]; - tensor var_2237 = const()[name = tensor("op_2237"), val = tensor([1, 8, -1, 188])]; - tensor x_275_cast_fp16 = reshape(shape = var_2237, x = x_273_cast_fp16)[name = tensor("x_275_cast_fp16")]; - tensor var_2241_begin_0 = const()[name = tensor("op_2241_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_2241_end_0 = const()[name = tensor("op_2241_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_2241_end_mask_0 = const()[name = tensor("op_2241_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_2241_cast_fp16 = slice_by_index(begin = var_2241_begin_0, end = var_2241_end_0, end_mask = var_2241_end_mask_0, x = x_275_cast_fp16)[name = tensor("op_2241_cast_fp16")]; - tensor var_2242 = const()[name = tensor("op_2242"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_49_cast_fp16 = reshape(shape = var_2242, x = var_2241_cast_fp16)[name = tensor("matrix_bd_49_cast_fp16")]; - tensor matrix_ac_25_transpose_x_0 = const()[name = tensor("matrix_ac_25_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_25_transpose_y_0 = const()[name = tensor("matrix_ac_25_transpose_y_0"), val = tensor(false)]; - tensor transpose_120_perm_0 = const()[name = tensor("transpose_120_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_121_perm_0 = const()[name = tensor("transpose_121_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_121 = transpose(perm = transpose_121_perm_0, x = k_49_cast_fp16)[name = tensor("transpose_226")]; - tensor transpose_120 = transpose(perm = transpose_120_perm_0, x = var_2225_cast_fp16)[name = tensor("transpose_227")]; - tensor matrix_ac_25_cast_fp16 = matmul(transpose_x = matrix_ac_25_transpose_x_0, transpose_y = matrix_ac_25_transpose_y_0, x = transpose_120, y = transpose_121)[name = tensor("matrix_ac_25_cast_fp16")]; - tensor matrix_bd_51_begin_0 = const()[name = tensor("matrix_bd_51_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_51_end_0 = const()[name = tensor("matrix_bd_51_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_51_end_mask_0 = const()[name = tensor("matrix_bd_51_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_51_cast_fp16 = slice_by_index(begin = matrix_bd_51_begin_0, end = matrix_bd_51_end_0, end_mask = matrix_bd_51_end_mask_0, x = matrix_bd_49_cast_fp16)[name = tensor("matrix_bd_51_cast_fp16")]; - tensor var_2251_cast_fp16 = add(x = matrix_ac_25_cast_fp16, y = matrix_bd_51_cast_fp16)[name = tensor("op_2251_cast_fp16")]; - tensor _inversed_scores_49_y_0_to_fp16 = const()[name = tensor("_inversed_scores_49_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; - tensor _inversed_scores_49_cast_fp16 = mul(x = var_2251_cast_fp16, y = _inversed_scores_49_y_0_to_fp16)[name = tensor("_inversed_scores_49_cast_fp16")]; - tensor scores_51_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_49_cast_fp16, cond = mask_3)[name = tensor("scores_51_cast_fp16")]; - tensor var_2257_cast_fp16 = softmax(axis = var_30, x = scores_51_cast_fp16)[name = tensor("op_2257_cast_fp16")]; - tensor input_657_cast_fp16 = select(a = var_11_to_fp16, b = var_2257_cast_fp16, cond = mask_3)[name = tensor("input_657_cast_fp16")]; - tensor x_277_transpose_x_0 = const()[name = tensor("x_277_transpose_x_0"), val = tensor(false)]; - tensor x_277_transpose_y_0 = const()[name = tensor("x_277_transpose_y_0"), val = tensor(false)]; - tensor value_27_cast_fp16 = transpose(perm = value_27_perm_0, x = v_25_cast_fp16)[name = tensor("transpose_225")]; - tensor x_277_cast_fp16 = matmul(transpose_x = x_277_transpose_x_0, transpose_y = x_277_transpose_y_0, x = input_657_cast_fp16, y = value_27_cast_fp16)[name = tensor("x_277_cast_fp16")]; - tensor var_2261_perm_0 = const()[name = tensor("op_2261_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_2262 = const()[name = tensor("op_2262"), val = tensor([1, -1, 1024])]; - tensor var_2261_cast_fp16 = transpose(perm = var_2261_perm_0, x = x_277_cast_fp16)[name = tensor("transpose_224")]; - tensor input_659_cast_fp16 = reshape(shape = var_2262, x = var_2261_cast_fp16)[name = tensor("input_659_cast_fp16")]; - tensor module_layers_12_self_attn_linear_out_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(233179648))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(233966144))), name = tensor("module_layers_12_self_attn_linear_out_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_115_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_12_self_attn_linear_out_weight_to_fp16_palettized, x = input_659_cast_fp16)[name = tensor("linear_115_cast_fp16")]; - tensor input_663_cast_fp16 = add(x = input_655_cast_fp16, y = linear_115_cast_fp16)[name = tensor("input_663_cast_fp16")]; - tensor x_281_axes_0 = const()[name = tensor("x_281_axes_0"), val = tensor([-1])]; - tensor module_layers_12_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_12_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(233966336)))]; - tensor module_layers_12_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_12_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(233968448)))]; - tensor x_281_cast_fp16 = layer_norm(axes = x_281_axes_0, beta = module_layers_12_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_12_norm_conv_weight_to_fp16, x = input_663_cast_fp16)[name = tensor("x_281_cast_fp16")]; - tensor input_665_perm_0 = const()[name = tensor("input_665_perm_0"), val = tensor([0, 2, 1])]; - tensor input_667_pad_type_0 = const()[name = tensor("input_667_pad_type_0"), val = tensor("valid")]; - tensor input_667_strides_0 = const()[name = tensor("input_667_strides_0"), val = tensor([1])]; - tensor input_667_pad_0 = const()[name = tensor("input_667_pad_0"), val = tensor([0, 0])]; - tensor input_667_dilations_0 = const()[name = tensor("input_667_dilations_0"), val = tensor([1])]; - tensor input_667_groups_0 = const()[name = tensor("input_667_groups_0"), val = tensor(1)]; - tensor module_layers_12_conv_pointwise_conv1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(233970560))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(235543488))), name = tensor("module_layers_12_conv_pointwise_conv1_weight_to_fp16_palettized"), shape = tensor([2048, 1024, 1])]; - tensor input_665_cast_fp16 = transpose(perm = input_665_perm_0, x = x_281_cast_fp16)[name = tensor("transpose_223")]; - tensor input_667_cast_fp16 = conv(dilations = input_667_dilations_0, groups = input_667_groups_0, pad = input_667_pad_0, pad_type = input_667_pad_type_0, strides = input_667_strides_0, weight = module_layers_12_conv_pointwise_conv1_weight_to_fp16_palettized, x = input_665_cast_fp16)[name = tensor("input_667_cast_fp16")]; - tensor x_283_split_num_splits_0 = const()[name = tensor("x_283_split_num_splits_0"), val = tensor(2)]; - tensor x_283_split_axis_0 = const()[name = tensor("x_283_split_axis_0"), val = tensor(1)]; - tensor x_283_split_cast_fp16_0, tensor x_283_split_cast_fp16_1 = split(axis = x_283_split_axis_0, num_splits = x_283_split_num_splits_0, x = input_667_cast_fp16)[name = tensor("x_283_split_cast_fp16")]; - tensor x_283_split_1_sigmoid_cast_fp16 = sigmoid(x = x_283_split_cast_fp16_1)[name = tensor("x_283_split_1_sigmoid_cast_fp16")]; - tensor x_283_cast_fp16 = mul(x = x_283_split_cast_fp16_0, y = x_283_split_1_sigmoid_cast_fp16)[name = tensor("x_283_cast_fp16")]; - tensor input_669_cast_fp16 = select(a = var_11_to_fp16, b = x_283_cast_fp16, cond = var_328)[name = tensor("input_669_cast_fp16")]; - tensor input_671_pad_0 = const()[name = tensor("input_671_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_671_mode_0 = const()[name = tensor("input_671_mode_0"), val = tensor("constant")]; - tensor const_137_to_fp16 = const()[name = tensor("const_137_to_fp16"), val = tensor(0x0p+0)]; - tensor input_671_cast_fp16 = pad(constant_val = const_137_to_fp16, mode = input_671_mode_0, pad = input_671_pad_0, x = input_669_cast_fp16)[name = tensor("input_671_cast_fp16")]; - tensor input_673_pad_type_0 = const()[name = tensor("input_673_pad_type_0"), val = tensor("valid")]; - tensor input_673_groups_0 = const()[name = tensor("input_673_groups_0"), val = tensor(1024)]; - tensor input_673_strides_0 = const()[name = tensor("input_673_strides_0"), val = tensor([1])]; - tensor input_673_pad_0 = const()[name = tensor("input_673_pad_0"), val = tensor([0, 0])]; - tensor input_673_dilations_0 = const()[name = tensor("input_673_dilations_0"), val = tensor([1])]; - tensor const_272_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(235543680))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(235550656))), name = tensor("const_272_to_fp16_palettized"), shape = tensor([1024, 1, 9])]; - tensor const_273_to_fp16 = const()[name = tensor("const_273_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(235550848)))]; - tensor input_675_cast_fp16 = conv(bias = const_273_to_fp16, dilations = input_673_dilations_0, groups = input_673_groups_0, pad = input_673_pad_0, pad_type = input_673_pad_type_0, strides = input_673_strides_0, weight = const_272_to_fp16_palettized, x = input_671_cast_fp16)[name = tensor("input_675_cast_fp16")]; - tensor input_677_cast_fp16 = silu(x = input_675_cast_fp16)[name = tensor("input_677_cast_fp16")]; - tensor x_285_pad_type_0 = const()[name = tensor("x_285_pad_type_0"), val = tensor("valid")]; - tensor x_285_strides_0 = const()[name = tensor("x_285_strides_0"), val = tensor([1])]; - tensor x_285_pad_0 = const()[name = tensor("x_285_pad_0"), val = tensor([0, 0])]; - tensor x_285_dilations_0 = const()[name = tensor("x_285_dilations_0"), val = tensor([1])]; - tensor x_285_groups_0 = const()[name = tensor("x_285_groups_0"), val = tensor(1)]; - tensor module_layers_12_conv_pointwise_conv2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(235552960))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(236339456))), name = tensor("module_layers_12_conv_pointwise_conv2_weight_to_fp16_palettized"), shape = tensor([1024, 1024, 1])]; - tensor x_285_cast_fp16 = conv(dilations = x_285_dilations_0, groups = x_285_groups_0, pad = x_285_pad_0, pad_type = x_285_pad_type_0, strides = x_285_strides_0, weight = module_layers_12_conv_pointwise_conv2_weight_to_fp16_palettized, x = input_677_cast_fp16)[name = tensor("x_285_cast_fp16")]; - tensor input_679_perm_0 = const()[name = tensor("input_679_perm_0"), val = tensor([0, 2, 1])]; - tensor input_679_cast_fp16 = transpose(perm = input_679_perm_0, x = x_285_cast_fp16)[name = tensor("transpose_222")]; - tensor input_681_cast_fp16 = add(x = input_663_cast_fp16, y = input_679_cast_fp16)[name = tensor("input_681_cast_fp16")]; - tensor input_683_axes_0 = const()[name = tensor("input_683_axes_0"), val = tensor([-1])]; - tensor module_layers_12_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_12_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(236339648)))]; - tensor module_layers_12_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_12_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(236341760)))]; - tensor input_683_cast_fp16 = layer_norm(axes = input_683_axes_0, beta = module_layers_12_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_12_norm_feed_forward2_weight_to_fp16, x = input_681_cast_fp16)[name = tensor("input_683_cast_fp16")]; - tensor module_layers_12_feed_forward2_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(236343872))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(239489664))), name = tensor("module_layers_12_feed_forward2_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_116_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_12_feed_forward2_linear1_weight_to_fp16_palettized, x = input_683_cast_fp16)[name = tensor("linear_116_cast_fp16")]; - tensor input_687_cast_fp16 = silu(x = linear_116_cast_fp16)[name = tensor("input_687_cast_fp16")]; - tensor module_layers_12_feed_forward2_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(239489856))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(242635648))), name = tensor("module_layers_12_feed_forward2_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_117_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_12_feed_forward2_linear2_weight_to_fp16_palettized, x = input_687_cast_fp16)[name = tensor("linear_117_cast_fp16")]; - tensor var_2322_to_fp16 = const()[name = tensor("op_2322_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2323_cast_fp16 = mul(x = linear_117_cast_fp16, y = var_2322_to_fp16)[name = tensor("op_2323_cast_fp16")]; - tensor input_693_cast_fp16 = add(x = input_681_cast_fp16, y = var_2323_cast_fp16)[name = tensor("input_693_cast_fp16")]; - tensor input_695_axes_0 = const()[name = tensor("input_695_axes_0"), val = tensor([-1])]; - tensor module_layers_12_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_12_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(242635840)))]; - tensor module_layers_12_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_12_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(242637952)))]; - tensor input_695_cast_fp16 = layer_norm(axes = input_695_axes_0, beta = module_layers_12_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_12_norm_out_weight_to_fp16, x = input_693_cast_fp16)[name = tensor("input_695_cast_fp16")]; - tensor input_697_axes_0 = const()[name = tensor("input_697_axes_0"), val = tensor([-1])]; - tensor module_layers_13_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_13_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(242640064)))]; - tensor module_layers_13_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_13_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(242642176)))]; - tensor input_697_cast_fp16 = layer_norm(axes = input_697_axes_0, beta = module_layers_13_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_13_norm_feed_forward1_weight_to_fp16, x = input_695_cast_fp16)[name = tensor("input_697_cast_fp16")]; - tensor module_layers_13_feed_forward1_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(242644288))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(245790080))), name = tensor("module_layers_13_feed_forward1_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_118_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_13_feed_forward1_linear1_weight_to_fp16_palettized, x = input_697_cast_fp16)[name = tensor("linear_118_cast_fp16")]; - tensor input_701_cast_fp16 = silu(x = linear_118_cast_fp16)[name = tensor("input_701_cast_fp16")]; - tensor module_layers_13_feed_forward1_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(245790272))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(248936064))), name = tensor("module_layers_13_feed_forward1_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_119_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_13_feed_forward1_linear2_weight_to_fp16_palettized, x = input_701_cast_fp16)[name = tensor("linear_119_cast_fp16")]; - tensor var_2351_to_fp16 = const()[name = tensor("op_2351_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2352_cast_fp16 = mul(x = linear_119_cast_fp16, y = var_2351_to_fp16)[name = tensor("op_2352_cast_fp16")]; - tensor input_707_cast_fp16 = add(x = input_695_cast_fp16, y = var_2352_cast_fp16)[name = tensor("input_707_cast_fp16")]; - tensor query_27_axes_0 = const()[name = tensor("query_27_axes_0"), val = tensor([-1])]; - tensor module_layers_13_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_13_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(248936256)))]; - tensor module_layers_13_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_13_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(248938368)))]; - tensor query_27_cast_fp16 = layer_norm(axes = query_27_axes_0, beta = module_layers_13_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_13_norm_self_att_weight_to_fp16, x = input_707_cast_fp16)[name = tensor("query_27_cast_fp16")]; - tensor module_layers_13_self_attn_linear_q_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(248940480))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(249726976))), name = tensor("module_layers_13_self_attn_linear_q_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_120_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_13_self_attn_linear_q_weight_to_fp16_palettized, x = query_27_cast_fp16)[name = tensor("linear_120_cast_fp16")]; - tensor var_2368 = const()[name = tensor("op_2368"), val = tensor([1, -1, 8, 128])]; - tensor q_79_cast_fp16 = reshape(shape = var_2368, x = linear_120_cast_fp16)[name = tensor("q_79_cast_fp16")]; - tensor module_layers_13_self_attn_linear_k_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(249727168))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(250513664))), name = tensor("module_layers_13_self_attn_linear_k_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_121_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_13_self_attn_linear_k_weight_to_fp16_palettized, x = query_27_cast_fp16)[name = tensor("linear_121_cast_fp16")]; - tensor var_2372 = const()[name = tensor("op_2372"), val = tensor([1, -1, 8, 128])]; - tensor k_53_cast_fp16 = reshape(shape = var_2372, x = linear_121_cast_fp16)[name = tensor("k_53_cast_fp16")]; - tensor module_layers_13_self_attn_linear_v_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(250513856))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(251300352))), name = tensor("module_layers_13_self_attn_linear_v_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_122_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_13_self_attn_linear_v_weight_to_fp16_palettized, x = query_27_cast_fp16)[name = tensor("linear_122_cast_fp16")]; - tensor var_2376 = const()[name = tensor("op_2376"), val = tensor([1, -1, 8, 128])]; - tensor v_27_cast_fp16 = reshape(shape = var_2376, x = linear_122_cast_fp16)[name = tensor("v_27_cast_fp16")]; - tensor value_29_perm_0 = const()[name = tensor("value_29_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor module_layers_13_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_13_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(251300544)))]; - tensor var_2388_cast_fp16 = add(x = q_79_cast_fp16, y = module_layers_13_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2388_cast_fp16")]; - tensor module_layers_13_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_13_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(251302656)))]; - tensor var_2390_cast_fp16 = add(x = q_79_cast_fp16, y = module_layers_13_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2390_cast_fp16")]; - tensor q_with_bias_v_27_perm_0 = const()[name = tensor("q_with_bias_v_27_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor x_293_transpose_x_0 = const()[name = tensor("x_293_transpose_x_0"), val = tensor(false)]; - tensor x_293_transpose_y_0 = const()[name = tensor("x_293_transpose_y_0"), val = tensor(false)]; - tensor op_2392_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(251304768))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(251592832))), name = tensor("op_2392_to_fp16_palettized"), shape = tensor([1, 8, 128, 375])]; - tensor q_with_bias_v_27_cast_fp16 = transpose(perm = q_with_bias_v_27_perm_0, x = var_2390_cast_fp16)[name = tensor("transpose_221")]; - tensor x_293_cast_fp16 = matmul(transpose_x = x_293_transpose_x_0, transpose_y = x_293_transpose_y_0, x = q_with_bias_v_27_cast_fp16, y = op_2392_to_fp16_palettized)[name = tensor("x_293_cast_fp16")]; - tensor x_295_pad_0 = const()[name = tensor("x_295_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_295_mode_0 = const()[name = tensor("x_295_mode_0"), val = tensor("constant")]; - tensor const_144_to_fp16 = const()[name = tensor("const_144_to_fp16"), val = tensor(0x0p+0)]; - tensor x_295_cast_fp16 = pad(constant_val = const_144_to_fp16, mode = x_295_mode_0, pad = x_295_pad_0, x = x_293_cast_fp16)[name = tensor("x_295_cast_fp16")]; - tensor var_2400 = const()[name = tensor("op_2400"), val = tensor([1, 8, -1, 188])]; - tensor x_297_cast_fp16 = reshape(shape = var_2400, x = x_295_cast_fp16)[name = tensor("x_297_cast_fp16")]; - tensor var_2404_begin_0 = const()[name = tensor("op_2404_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_2404_end_0 = const()[name = tensor("op_2404_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_2404_end_mask_0 = const()[name = tensor("op_2404_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_2404_cast_fp16 = slice_by_index(begin = var_2404_begin_0, end = var_2404_end_0, end_mask = var_2404_end_mask_0, x = x_297_cast_fp16)[name = tensor("op_2404_cast_fp16")]; - tensor var_2405 = const()[name = tensor("op_2405"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_53_cast_fp16 = reshape(shape = var_2405, x = var_2404_cast_fp16)[name = tensor("matrix_bd_53_cast_fp16")]; - tensor matrix_ac_27_transpose_x_0 = const()[name = tensor("matrix_ac_27_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_27_transpose_y_0 = const()[name = tensor("matrix_ac_27_transpose_y_0"), val = tensor(false)]; - tensor transpose_122_perm_0 = const()[name = tensor("transpose_122_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_123_perm_0 = const()[name = tensor("transpose_123_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_123 = transpose(perm = transpose_123_perm_0, x = k_53_cast_fp16)[name = tensor("transpose_219")]; - tensor transpose_122 = transpose(perm = transpose_122_perm_0, x = var_2388_cast_fp16)[name = tensor("transpose_220")]; - tensor matrix_ac_27_cast_fp16 = matmul(transpose_x = matrix_ac_27_transpose_x_0, transpose_y = matrix_ac_27_transpose_y_0, x = transpose_122, y = transpose_123)[name = tensor("matrix_ac_27_cast_fp16")]; - tensor matrix_bd_55_begin_0 = const()[name = tensor("matrix_bd_55_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_55_end_0 = const()[name = tensor("matrix_bd_55_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_55_end_mask_0 = const()[name = tensor("matrix_bd_55_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_55_cast_fp16 = slice_by_index(begin = matrix_bd_55_begin_0, end = matrix_bd_55_end_0, end_mask = matrix_bd_55_end_mask_0, x = matrix_bd_53_cast_fp16)[name = tensor("matrix_bd_55_cast_fp16")]; - tensor var_2414_cast_fp16 = add(x = matrix_ac_27_cast_fp16, y = matrix_bd_55_cast_fp16)[name = tensor("op_2414_cast_fp16")]; - tensor _inversed_scores_53_y_0_to_fp16 = const()[name = tensor("_inversed_scores_53_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; - tensor _inversed_scores_53_cast_fp16 = mul(x = var_2414_cast_fp16, y = _inversed_scores_53_y_0_to_fp16)[name = tensor("_inversed_scores_53_cast_fp16")]; - tensor scores_55_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_53_cast_fp16, cond = mask_3)[name = tensor("scores_55_cast_fp16")]; - tensor var_2420_cast_fp16 = softmax(axis = var_30, x = scores_55_cast_fp16)[name = tensor("op_2420_cast_fp16")]; - tensor input_709_cast_fp16 = select(a = var_11_to_fp16, b = var_2420_cast_fp16, cond = mask_3)[name = tensor("input_709_cast_fp16")]; - tensor x_299_transpose_x_0 = const()[name = tensor("x_299_transpose_x_0"), val = tensor(false)]; - tensor x_299_transpose_y_0 = const()[name = tensor("x_299_transpose_y_0"), val = tensor(false)]; - tensor value_29_cast_fp16 = transpose(perm = value_29_perm_0, x = v_27_cast_fp16)[name = tensor("transpose_218")]; - tensor x_299_cast_fp16 = matmul(transpose_x = x_299_transpose_x_0, transpose_y = x_299_transpose_y_0, x = input_709_cast_fp16, y = value_29_cast_fp16)[name = tensor("x_299_cast_fp16")]; - tensor var_2424_perm_0 = const()[name = tensor("op_2424_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_2425 = const()[name = tensor("op_2425"), val = tensor([1, -1, 1024])]; - tensor var_2424_cast_fp16 = transpose(perm = var_2424_perm_0, x = x_299_cast_fp16)[name = tensor("transpose_217")]; - tensor input_711_cast_fp16 = reshape(shape = var_2425, x = var_2424_cast_fp16)[name = tensor("input_711_cast_fp16")]; - tensor module_layers_13_self_attn_linear_out_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(251593024))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(252379520))), name = tensor("module_layers_13_self_attn_linear_out_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_124_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_13_self_attn_linear_out_weight_to_fp16_palettized, x = input_711_cast_fp16)[name = tensor("linear_124_cast_fp16")]; - tensor input_715_cast_fp16 = add(x = input_707_cast_fp16, y = linear_124_cast_fp16)[name = tensor("input_715_cast_fp16")]; - tensor x_303_axes_0 = const()[name = tensor("x_303_axes_0"), val = tensor([-1])]; - tensor module_layers_13_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_13_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(252379712)))]; - tensor module_layers_13_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_13_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(252381824)))]; - tensor x_303_cast_fp16 = layer_norm(axes = x_303_axes_0, beta = module_layers_13_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_13_norm_conv_weight_to_fp16, x = input_715_cast_fp16)[name = tensor("x_303_cast_fp16")]; - tensor input_717_perm_0 = const()[name = tensor("input_717_perm_0"), val = tensor([0, 2, 1])]; - tensor input_719_pad_type_0 = const()[name = tensor("input_719_pad_type_0"), val = tensor("valid")]; - tensor input_719_strides_0 = const()[name = tensor("input_719_strides_0"), val = tensor([1])]; - tensor input_719_pad_0 = const()[name = tensor("input_719_pad_0"), val = tensor([0, 0])]; - tensor input_719_dilations_0 = const()[name = tensor("input_719_dilations_0"), val = tensor([1])]; - tensor input_719_groups_0 = const()[name = tensor("input_719_groups_0"), val = tensor(1)]; - tensor module_layers_13_conv_pointwise_conv1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(252383936))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(253956864))), name = tensor("module_layers_13_conv_pointwise_conv1_weight_to_fp16_palettized"), shape = tensor([2048, 1024, 1])]; - tensor input_717_cast_fp16 = transpose(perm = input_717_perm_0, x = x_303_cast_fp16)[name = tensor("transpose_216")]; - tensor input_719_cast_fp16 = conv(dilations = input_719_dilations_0, groups = input_719_groups_0, pad = input_719_pad_0, pad_type = input_719_pad_type_0, strides = input_719_strides_0, weight = module_layers_13_conv_pointwise_conv1_weight_to_fp16_palettized, x = input_717_cast_fp16)[name = tensor("input_719_cast_fp16")]; - tensor x_305_split_num_splits_0 = const()[name = tensor("x_305_split_num_splits_0"), val = tensor(2)]; - tensor x_305_split_axis_0 = const()[name = tensor("x_305_split_axis_0"), val = tensor(1)]; - tensor x_305_split_cast_fp16_0, tensor x_305_split_cast_fp16_1 = split(axis = x_305_split_axis_0, num_splits = x_305_split_num_splits_0, x = input_719_cast_fp16)[name = tensor("x_305_split_cast_fp16")]; - tensor x_305_split_1_sigmoid_cast_fp16 = sigmoid(x = x_305_split_cast_fp16_1)[name = tensor("x_305_split_1_sigmoid_cast_fp16")]; - tensor x_305_cast_fp16 = mul(x = x_305_split_cast_fp16_0, y = x_305_split_1_sigmoid_cast_fp16)[name = tensor("x_305_cast_fp16")]; - tensor input_721_cast_fp16 = select(a = var_11_to_fp16, b = x_305_cast_fp16, cond = var_328)[name = tensor("input_721_cast_fp16")]; - tensor input_723_pad_0 = const()[name = tensor("input_723_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_723_mode_0 = const()[name = tensor("input_723_mode_0"), val = tensor("constant")]; - tensor const_147_to_fp16 = const()[name = tensor("const_147_to_fp16"), val = tensor(0x0p+0)]; - tensor input_723_cast_fp16 = pad(constant_val = const_147_to_fp16, mode = input_723_mode_0, pad = input_723_pad_0, x = input_721_cast_fp16)[name = tensor("input_723_cast_fp16")]; - tensor input_725_pad_type_0 = const()[name = tensor("input_725_pad_type_0"), val = tensor("valid")]; - tensor input_725_groups_0 = const()[name = tensor("input_725_groups_0"), val = tensor(1024)]; - tensor input_725_strides_0 = const()[name = tensor("input_725_strides_0"), val = tensor([1])]; - tensor input_725_pad_0 = const()[name = tensor("input_725_pad_0"), val = tensor([0, 0])]; - tensor input_725_dilations_0 = const()[name = tensor("input_725_dilations_0"), val = tensor([1])]; - tensor const_274_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(253957056))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(253964032))), name = tensor("const_274_to_fp16_palettized"), shape = tensor([1024, 1, 9])]; - tensor const_275_to_fp16 = const()[name = tensor("const_275_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(253964224)))]; - tensor input_727_cast_fp16 = conv(bias = const_275_to_fp16, dilations = input_725_dilations_0, groups = input_725_groups_0, pad = input_725_pad_0, pad_type = input_725_pad_type_0, strides = input_725_strides_0, weight = const_274_to_fp16_palettized, x = input_723_cast_fp16)[name = tensor("input_727_cast_fp16")]; - tensor input_729_cast_fp16 = silu(x = input_727_cast_fp16)[name = tensor("input_729_cast_fp16")]; - tensor x_307_pad_type_0 = const()[name = tensor("x_307_pad_type_0"), val = tensor("valid")]; - tensor x_307_strides_0 = const()[name = tensor("x_307_strides_0"), val = tensor([1])]; - tensor x_307_pad_0 = const()[name = tensor("x_307_pad_0"), val = tensor([0, 0])]; - tensor x_307_dilations_0 = const()[name = tensor("x_307_dilations_0"), val = tensor([1])]; - tensor x_307_groups_0 = const()[name = tensor("x_307_groups_0"), val = tensor(1)]; - tensor module_layers_13_conv_pointwise_conv2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(253966336))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(254752832))), name = tensor("module_layers_13_conv_pointwise_conv2_weight_to_fp16_palettized"), shape = tensor([1024, 1024, 1])]; - tensor x_307_cast_fp16 = conv(dilations = x_307_dilations_0, groups = x_307_groups_0, pad = x_307_pad_0, pad_type = x_307_pad_type_0, strides = x_307_strides_0, weight = module_layers_13_conv_pointwise_conv2_weight_to_fp16_palettized, x = input_729_cast_fp16)[name = tensor("x_307_cast_fp16")]; - tensor input_731_perm_0 = const()[name = tensor("input_731_perm_0"), val = tensor([0, 2, 1])]; - tensor input_731_cast_fp16 = transpose(perm = input_731_perm_0, x = x_307_cast_fp16)[name = tensor("transpose_215")]; - tensor input_733_cast_fp16 = add(x = input_715_cast_fp16, y = input_731_cast_fp16)[name = tensor("input_733_cast_fp16")]; - tensor input_735_axes_0 = const()[name = tensor("input_735_axes_0"), val = tensor([-1])]; - tensor module_layers_13_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_13_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(254753024)))]; - tensor module_layers_13_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_13_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(254755136)))]; - tensor input_735_cast_fp16 = layer_norm(axes = input_735_axes_0, beta = module_layers_13_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_13_norm_feed_forward2_weight_to_fp16, x = input_733_cast_fp16)[name = tensor("input_735_cast_fp16")]; - tensor module_layers_13_feed_forward2_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(254757248))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(257903040))), name = tensor("module_layers_13_feed_forward2_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_125_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_13_feed_forward2_linear1_weight_to_fp16_palettized, x = input_735_cast_fp16)[name = tensor("linear_125_cast_fp16")]; - tensor input_739_cast_fp16 = silu(x = linear_125_cast_fp16)[name = tensor("input_739_cast_fp16")]; - tensor module_layers_13_feed_forward2_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(257903232))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(261049024))), name = tensor("module_layers_13_feed_forward2_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_126_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_13_feed_forward2_linear2_weight_to_fp16_palettized, x = input_739_cast_fp16)[name = tensor("linear_126_cast_fp16")]; - tensor var_2485_to_fp16 = const()[name = tensor("op_2485_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2486_cast_fp16 = mul(x = linear_126_cast_fp16, y = var_2485_to_fp16)[name = tensor("op_2486_cast_fp16")]; - tensor input_745_cast_fp16 = add(x = input_733_cast_fp16, y = var_2486_cast_fp16)[name = tensor("input_745_cast_fp16")]; - tensor input_747_axes_0 = const()[name = tensor("input_747_axes_0"), val = tensor([-1])]; - tensor module_layers_13_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_13_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(261049216)))]; - tensor module_layers_13_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_13_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(261051328)))]; - tensor input_747_cast_fp16 = layer_norm(axes = input_747_axes_0, beta = module_layers_13_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_13_norm_out_weight_to_fp16, x = input_745_cast_fp16)[name = tensor("input_747_cast_fp16")]; - tensor input_749_axes_0 = const()[name = tensor("input_749_axes_0"), val = tensor([-1])]; - tensor module_layers_14_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_14_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(261053440)))]; - tensor module_layers_14_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_14_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(261055552)))]; - tensor input_749_cast_fp16 = layer_norm(axes = input_749_axes_0, beta = module_layers_14_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_14_norm_feed_forward1_weight_to_fp16, x = input_747_cast_fp16)[name = tensor("input_749_cast_fp16")]; - tensor module_layers_14_feed_forward1_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(261057664))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(264203456))), name = tensor("module_layers_14_feed_forward1_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_127_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_14_feed_forward1_linear1_weight_to_fp16_palettized, x = input_749_cast_fp16)[name = tensor("linear_127_cast_fp16")]; - tensor input_753_cast_fp16 = silu(x = linear_127_cast_fp16)[name = tensor("input_753_cast_fp16")]; - tensor module_layers_14_feed_forward1_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(264203648))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(267349440))), name = tensor("module_layers_14_feed_forward1_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_128_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_14_feed_forward1_linear2_weight_to_fp16_palettized, x = input_753_cast_fp16)[name = tensor("linear_128_cast_fp16")]; - tensor var_2514_to_fp16 = const()[name = tensor("op_2514_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2515_cast_fp16 = mul(x = linear_128_cast_fp16, y = var_2514_to_fp16)[name = tensor("op_2515_cast_fp16")]; - tensor input_759_cast_fp16 = add(x = input_747_cast_fp16, y = var_2515_cast_fp16)[name = tensor("input_759_cast_fp16")]; - tensor query_29_axes_0 = const()[name = tensor("query_29_axes_0"), val = tensor([-1])]; - tensor module_layers_14_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_14_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(267349632)))]; - tensor module_layers_14_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_14_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(267351744)))]; - tensor query_29_cast_fp16 = layer_norm(axes = query_29_axes_0, beta = module_layers_14_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_14_norm_self_att_weight_to_fp16, x = input_759_cast_fp16)[name = tensor("query_29_cast_fp16")]; - tensor module_layers_14_self_attn_linear_q_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(267353856))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(268140352))), name = tensor("module_layers_14_self_attn_linear_q_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_129_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_14_self_attn_linear_q_weight_to_fp16_palettized, x = query_29_cast_fp16)[name = tensor("linear_129_cast_fp16")]; - tensor var_2531 = const()[name = tensor("op_2531"), val = tensor([1, -1, 8, 128])]; - tensor q_85_cast_fp16 = reshape(shape = var_2531, x = linear_129_cast_fp16)[name = tensor("q_85_cast_fp16")]; - tensor module_layers_14_self_attn_linear_k_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(268140544))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(268927040))), name = tensor("module_layers_14_self_attn_linear_k_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_130_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_14_self_attn_linear_k_weight_to_fp16_palettized, x = query_29_cast_fp16)[name = tensor("linear_130_cast_fp16")]; - tensor var_2535 = const()[name = tensor("op_2535"), val = tensor([1, -1, 8, 128])]; - tensor k_57_cast_fp16 = reshape(shape = var_2535, x = linear_130_cast_fp16)[name = tensor("k_57_cast_fp16")]; - tensor module_layers_14_self_attn_linear_v_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(268927232))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(269713728))), name = tensor("module_layers_14_self_attn_linear_v_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_131_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_14_self_attn_linear_v_weight_to_fp16_palettized, x = query_29_cast_fp16)[name = tensor("linear_131_cast_fp16")]; - tensor var_2539 = const()[name = tensor("op_2539"), val = tensor([1, -1, 8, 128])]; - tensor v_29_cast_fp16 = reshape(shape = var_2539, x = linear_131_cast_fp16)[name = tensor("v_29_cast_fp16")]; - tensor value_31_perm_0 = const()[name = tensor("value_31_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor module_layers_14_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_14_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(269713920)))]; - tensor var_2551_cast_fp16 = add(x = q_85_cast_fp16, y = module_layers_14_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2551_cast_fp16")]; - tensor module_layers_14_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_14_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(269716032)))]; - tensor var_2553_cast_fp16 = add(x = q_85_cast_fp16, y = module_layers_14_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2553_cast_fp16")]; - tensor q_with_bias_v_29_perm_0 = const()[name = tensor("q_with_bias_v_29_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor x_315_transpose_x_0 = const()[name = tensor("x_315_transpose_x_0"), val = tensor(false)]; - tensor x_315_transpose_y_0 = const()[name = tensor("x_315_transpose_y_0"), val = tensor(false)]; - tensor op_2555_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(269718144))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(270006208))), name = tensor("op_2555_to_fp16_palettized"), shape = tensor([1, 8, 128, 375])]; - tensor q_with_bias_v_29_cast_fp16 = transpose(perm = q_with_bias_v_29_perm_0, x = var_2553_cast_fp16)[name = tensor("transpose_214")]; - tensor x_315_cast_fp16 = matmul(transpose_x = x_315_transpose_x_0, transpose_y = x_315_transpose_y_0, x = q_with_bias_v_29_cast_fp16, y = op_2555_to_fp16_palettized)[name = tensor("x_315_cast_fp16")]; - tensor x_317_pad_0 = const()[name = tensor("x_317_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_317_mode_0 = const()[name = tensor("x_317_mode_0"), val = tensor("constant")]; - tensor const_154_to_fp16 = const()[name = tensor("const_154_to_fp16"), val = tensor(0x0p+0)]; - tensor x_317_cast_fp16 = pad(constant_val = const_154_to_fp16, mode = x_317_mode_0, pad = x_317_pad_0, x = x_315_cast_fp16)[name = tensor("x_317_cast_fp16")]; - tensor var_2563 = const()[name = tensor("op_2563"), val = tensor([1, 8, -1, 188])]; - tensor x_319_cast_fp16 = reshape(shape = var_2563, x = x_317_cast_fp16)[name = tensor("x_319_cast_fp16")]; - tensor var_2567_begin_0 = const()[name = tensor("op_2567_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_2567_end_0 = const()[name = tensor("op_2567_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_2567_end_mask_0 = const()[name = tensor("op_2567_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_2567_cast_fp16 = slice_by_index(begin = var_2567_begin_0, end = var_2567_end_0, end_mask = var_2567_end_mask_0, x = x_319_cast_fp16)[name = tensor("op_2567_cast_fp16")]; - tensor var_2568 = const()[name = tensor("op_2568"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_57_cast_fp16 = reshape(shape = var_2568, x = var_2567_cast_fp16)[name = tensor("matrix_bd_57_cast_fp16")]; - tensor matrix_ac_29_transpose_x_0 = const()[name = tensor("matrix_ac_29_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_29_transpose_y_0 = const()[name = tensor("matrix_ac_29_transpose_y_0"), val = tensor(false)]; - tensor transpose_124_perm_0 = const()[name = tensor("transpose_124_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_125_perm_0 = const()[name = tensor("transpose_125_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_125 = transpose(perm = transpose_125_perm_0, x = k_57_cast_fp16)[name = tensor("transpose_212")]; - tensor transpose_124 = transpose(perm = transpose_124_perm_0, x = var_2551_cast_fp16)[name = tensor("transpose_213")]; - tensor matrix_ac_29_cast_fp16 = matmul(transpose_x = matrix_ac_29_transpose_x_0, transpose_y = matrix_ac_29_transpose_y_0, x = transpose_124, y = transpose_125)[name = tensor("matrix_ac_29_cast_fp16")]; - tensor matrix_bd_59_begin_0 = const()[name = tensor("matrix_bd_59_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_59_end_0 = const()[name = tensor("matrix_bd_59_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_59_end_mask_0 = const()[name = tensor("matrix_bd_59_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_59_cast_fp16 = slice_by_index(begin = matrix_bd_59_begin_0, end = matrix_bd_59_end_0, end_mask = matrix_bd_59_end_mask_0, x = matrix_bd_57_cast_fp16)[name = tensor("matrix_bd_59_cast_fp16")]; - tensor var_2577_cast_fp16 = add(x = matrix_ac_29_cast_fp16, y = matrix_bd_59_cast_fp16)[name = tensor("op_2577_cast_fp16")]; - tensor _inversed_scores_57_y_0_to_fp16 = const()[name = tensor("_inversed_scores_57_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; - tensor _inversed_scores_57_cast_fp16 = mul(x = var_2577_cast_fp16, y = _inversed_scores_57_y_0_to_fp16)[name = tensor("_inversed_scores_57_cast_fp16")]; - tensor scores_59_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_57_cast_fp16, cond = mask_3)[name = tensor("scores_59_cast_fp16")]; - tensor var_2583_cast_fp16 = softmax(axis = var_30, x = scores_59_cast_fp16)[name = tensor("op_2583_cast_fp16")]; - tensor input_761_cast_fp16 = select(a = var_11_to_fp16, b = var_2583_cast_fp16, cond = mask_3)[name = tensor("input_761_cast_fp16")]; - tensor x_321_transpose_x_0 = const()[name = tensor("x_321_transpose_x_0"), val = tensor(false)]; - tensor x_321_transpose_y_0 = const()[name = tensor("x_321_transpose_y_0"), val = tensor(false)]; - tensor value_31_cast_fp16 = transpose(perm = value_31_perm_0, x = v_29_cast_fp16)[name = tensor("transpose_211")]; - tensor x_321_cast_fp16 = matmul(transpose_x = x_321_transpose_x_0, transpose_y = x_321_transpose_y_0, x = input_761_cast_fp16, y = value_31_cast_fp16)[name = tensor("x_321_cast_fp16")]; - tensor var_2587_perm_0 = const()[name = tensor("op_2587_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_2588 = const()[name = tensor("op_2588"), val = tensor([1, -1, 1024])]; - tensor var_2587_cast_fp16 = transpose(perm = var_2587_perm_0, x = x_321_cast_fp16)[name = tensor("transpose_210")]; - tensor input_763_cast_fp16 = reshape(shape = var_2588, x = var_2587_cast_fp16)[name = tensor("input_763_cast_fp16")]; - tensor module_layers_14_self_attn_linear_out_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(270006400))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(270792896))), name = tensor("module_layers_14_self_attn_linear_out_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_133_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_14_self_attn_linear_out_weight_to_fp16_palettized, x = input_763_cast_fp16)[name = tensor("linear_133_cast_fp16")]; - tensor input_767_cast_fp16 = add(x = input_759_cast_fp16, y = linear_133_cast_fp16)[name = tensor("input_767_cast_fp16")]; - tensor x_325_axes_0 = const()[name = tensor("x_325_axes_0"), val = tensor([-1])]; - tensor module_layers_14_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_14_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(270793088)))]; - tensor module_layers_14_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_14_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(270795200)))]; - tensor x_325_cast_fp16 = layer_norm(axes = x_325_axes_0, beta = module_layers_14_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_14_norm_conv_weight_to_fp16, x = input_767_cast_fp16)[name = tensor("x_325_cast_fp16")]; - tensor input_769_perm_0 = const()[name = tensor("input_769_perm_0"), val = tensor([0, 2, 1])]; - tensor input_771_pad_type_0 = const()[name = tensor("input_771_pad_type_0"), val = tensor("valid")]; - tensor input_771_strides_0 = const()[name = tensor("input_771_strides_0"), val = tensor([1])]; - tensor input_771_pad_0 = const()[name = tensor("input_771_pad_0"), val = tensor([0, 0])]; - tensor input_771_dilations_0 = const()[name = tensor("input_771_dilations_0"), val = tensor([1])]; - tensor input_771_groups_0 = const()[name = tensor("input_771_groups_0"), val = tensor(1)]; - tensor module_layers_14_conv_pointwise_conv1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(270797312))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(272370240))), name = tensor("module_layers_14_conv_pointwise_conv1_weight_to_fp16_palettized"), shape = tensor([2048, 1024, 1])]; - tensor input_769_cast_fp16 = transpose(perm = input_769_perm_0, x = x_325_cast_fp16)[name = tensor("transpose_209")]; - tensor input_771_cast_fp16 = conv(dilations = input_771_dilations_0, groups = input_771_groups_0, pad = input_771_pad_0, pad_type = input_771_pad_type_0, strides = input_771_strides_0, weight = module_layers_14_conv_pointwise_conv1_weight_to_fp16_palettized, x = input_769_cast_fp16)[name = tensor("input_771_cast_fp16")]; - tensor x_327_split_num_splits_0 = const()[name = tensor("x_327_split_num_splits_0"), val = tensor(2)]; - tensor x_327_split_axis_0 = const()[name = tensor("x_327_split_axis_0"), val = tensor(1)]; - tensor x_327_split_cast_fp16_0, tensor x_327_split_cast_fp16_1 = split(axis = x_327_split_axis_0, num_splits = x_327_split_num_splits_0, x = input_771_cast_fp16)[name = tensor("x_327_split_cast_fp16")]; - tensor x_327_split_1_sigmoid_cast_fp16 = sigmoid(x = x_327_split_cast_fp16_1)[name = tensor("x_327_split_1_sigmoid_cast_fp16")]; - tensor x_327_cast_fp16 = mul(x = x_327_split_cast_fp16_0, y = x_327_split_1_sigmoid_cast_fp16)[name = tensor("x_327_cast_fp16")]; - tensor input_773_cast_fp16 = select(a = var_11_to_fp16, b = x_327_cast_fp16, cond = var_328)[name = tensor("input_773_cast_fp16")]; - tensor input_775_pad_0 = const()[name = tensor("input_775_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_775_mode_0 = const()[name = tensor("input_775_mode_0"), val = tensor("constant")]; - tensor const_157_to_fp16 = const()[name = tensor("const_157_to_fp16"), val = tensor(0x0p+0)]; - tensor input_775_cast_fp16 = pad(constant_val = const_157_to_fp16, mode = input_775_mode_0, pad = input_775_pad_0, x = input_773_cast_fp16)[name = tensor("input_775_cast_fp16")]; - tensor input_777_pad_type_0 = const()[name = tensor("input_777_pad_type_0"), val = tensor("valid")]; - tensor input_777_groups_0 = const()[name = tensor("input_777_groups_0"), val = tensor(1024)]; - tensor input_777_strides_0 = const()[name = tensor("input_777_strides_0"), val = tensor([1])]; - tensor input_777_pad_0 = const()[name = tensor("input_777_pad_0"), val = tensor([0, 0])]; - tensor input_777_dilations_0 = const()[name = tensor("input_777_dilations_0"), val = tensor([1])]; - tensor const_276_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(272370432))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(272377408))), name = tensor("const_276_to_fp16_palettized"), shape = tensor([1024, 1, 9])]; - tensor const_277_to_fp16 = const()[name = tensor("const_277_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(272377600)))]; - tensor input_779_cast_fp16 = conv(bias = const_277_to_fp16, dilations = input_777_dilations_0, groups = input_777_groups_0, pad = input_777_pad_0, pad_type = input_777_pad_type_0, strides = input_777_strides_0, weight = const_276_to_fp16_palettized, x = input_775_cast_fp16)[name = tensor("input_779_cast_fp16")]; - tensor input_781_cast_fp16 = silu(x = input_779_cast_fp16)[name = tensor("input_781_cast_fp16")]; - tensor x_329_pad_type_0 = const()[name = tensor("x_329_pad_type_0"), val = tensor("valid")]; - tensor x_329_strides_0 = const()[name = tensor("x_329_strides_0"), val = tensor([1])]; - tensor x_329_pad_0 = const()[name = tensor("x_329_pad_0"), val = tensor([0, 0])]; - tensor x_329_dilations_0 = const()[name = tensor("x_329_dilations_0"), val = tensor([1])]; - tensor x_329_groups_0 = const()[name = tensor("x_329_groups_0"), val = tensor(1)]; - tensor module_layers_14_conv_pointwise_conv2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(272379712))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(273166208))), name = tensor("module_layers_14_conv_pointwise_conv2_weight_to_fp16_palettized"), shape = tensor([1024, 1024, 1])]; - tensor x_329_cast_fp16 = conv(dilations = x_329_dilations_0, groups = x_329_groups_0, pad = x_329_pad_0, pad_type = x_329_pad_type_0, strides = x_329_strides_0, weight = module_layers_14_conv_pointwise_conv2_weight_to_fp16_palettized, x = input_781_cast_fp16)[name = tensor("x_329_cast_fp16")]; - tensor input_783_perm_0 = const()[name = tensor("input_783_perm_0"), val = tensor([0, 2, 1])]; - tensor input_783_cast_fp16 = transpose(perm = input_783_perm_0, x = x_329_cast_fp16)[name = tensor("transpose_208")]; - tensor input_785_cast_fp16 = add(x = input_767_cast_fp16, y = input_783_cast_fp16)[name = tensor("input_785_cast_fp16")]; - tensor input_787_axes_0 = const()[name = tensor("input_787_axes_0"), val = tensor([-1])]; - tensor module_layers_14_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_14_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(273166400)))]; - tensor module_layers_14_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_14_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(273168512)))]; - tensor input_787_cast_fp16 = layer_norm(axes = input_787_axes_0, beta = module_layers_14_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_14_norm_feed_forward2_weight_to_fp16, x = input_785_cast_fp16)[name = tensor("input_787_cast_fp16")]; - tensor module_layers_14_feed_forward2_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(273170624))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(276316416))), name = tensor("module_layers_14_feed_forward2_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_134_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_14_feed_forward2_linear1_weight_to_fp16_palettized, x = input_787_cast_fp16)[name = tensor("linear_134_cast_fp16")]; - tensor input_791_cast_fp16 = silu(x = linear_134_cast_fp16)[name = tensor("input_791_cast_fp16")]; - tensor module_layers_14_feed_forward2_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(276316608))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(279462400))), name = tensor("module_layers_14_feed_forward2_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_135_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_14_feed_forward2_linear2_weight_to_fp16_palettized, x = input_791_cast_fp16)[name = tensor("linear_135_cast_fp16")]; - tensor var_2648_to_fp16 = const()[name = tensor("op_2648_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2649_cast_fp16 = mul(x = linear_135_cast_fp16, y = var_2648_to_fp16)[name = tensor("op_2649_cast_fp16")]; - tensor input_797_cast_fp16 = add(x = input_785_cast_fp16, y = var_2649_cast_fp16)[name = tensor("input_797_cast_fp16")]; - tensor input_799_axes_0 = const()[name = tensor("input_799_axes_0"), val = tensor([-1])]; - tensor module_layers_14_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_14_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(279462592)))]; - tensor module_layers_14_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_14_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(279464704)))]; - tensor input_799_cast_fp16 = layer_norm(axes = input_799_axes_0, beta = module_layers_14_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_14_norm_out_weight_to_fp16, x = input_797_cast_fp16)[name = tensor("input_799_cast_fp16")]; - tensor input_801_axes_0 = const()[name = tensor("input_801_axes_0"), val = tensor([-1])]; - tensor module_layers_15_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_15_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(279466816)))]; - tensor module_layers_15_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_15_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(279468928)))]; - tensor input_801_cast_fp16 = layer_norm(axes = input_801_axes_0, beta = module_layers_15_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_15_norm_feed_forward1_weight_to_fp16, x = input_799_cast_fp16)[name = tensor("input_801_cast_fp16")]; - tensor module_layers_15_feed_forward1_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(279471040))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(282616832))), name = tensor("module_layers_15_feed_forward1_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_136_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_15_feed_forward1_linear1_weight_to_fp16_palettized, x = input_801_cast_fp16)[name = tensor("linear_136_cast_fp16")]; - tensor input_805_cast_fp16 = silu(x = linear_136_cast_fp16)[name = tensor("input_805_cast_fp16")]; - tensor module_layers_15_feed_forward1_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(282617024))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(285762816))), name = tensor("module_layers_15_feed_forward1_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_137_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_15_feed_forward1_linear2_weight_to_fp16_palettized, x = input_805_cast_fp16)[name = tensor("linear_137_cast_fp16")]; - tensor var_2677_to_fp16 = const()[name = tensor("op_2677_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2678_cast_fp16 = mul(x = linear_137_cast_fp16, y = var_2677_to_fp16)[name = tensor("op_2678_cast_fp16")]; - tensor input_811_cast_fp16 = add(x = input_799_cast_fp16, y = var_2678_cast_fp16)[name = tensor("input_811_cast_fp16")]; - tensor query_31_axes_0 = const()[name = tensor("query_31_axes_0"), val = tensor([-1])]; - tensor module_layers_15_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_15_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(285763008)))]; - tensor module_layers_15_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_15_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(285765120)))]; - tensor query_31_cast_fp16 = layer_norm(axes = query_31_axes_0, beta = module_layers_15_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_15_norm_self_att_weight_to_fp16, x = input_811_cast_fp16)[name = tensor("query_31_cast_fp16")]; - tensor module_layers_15_self_attn_linear_q_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(285767232))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(286553728))), name = tensor("module_layers_15_self_attn_linear_q_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_138_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_15_self_attn_linear_q_weight_to_fp16_palettized, x = query_31_cast_fp16)[name = tensor("linear_138_cast_fp16")]; - tensor var_2694 = const()[name = tensor("op_2694"), val = tensor([1, -1, 8, 128])]; - tensor q_91_cast_fp16 = reshape(shape = var_2694, x = linear_138_cast_fp16)[name = tensor("q_91_cast_fp16")]; - tensor module_layers_15_self_attn_linear_k_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(286553920))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(287340416))), name = tensor("module_layers_15_self_attn_linear_k_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_139_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_15_self_attn_linear_k_weight_to_fp16_palettized, x = query_31_cast_fp16)[name = tensor("linear_139_cast_fp16")]; - tensor var_2698 = const()[name = tensor("op_2698"), val = tensor([1, -1, 8, 128])]; - tensor k_61_cast_fp16 = reshape(shape = var_2698, x = linear_139_cast_fp16)[name = tensor("k_61_cast_fp16")]; - tensor module_layers_15_self_attn_linear_v_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(287340608))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(288127104))), name = tensor("module_layers_15_self_attn_linear_v_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_140_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_15_self_attn_linear_v_weight_to_fp16_palettized, x = query_31_cast_fp16)[name = tensor("linear_140_cast_fp16")]; - tensor var_2702 = const()[name = tensor("op_2702"), val = tensor([1, -1, 8, 128])]; - tensor v_31_cast_fp16 = reshape(shape = var_2702, x = linear_140_cast_fp16)[name = tensor("v_31_cast_fp16")]; - tensor value_33_perm_0 = const()[name = tensor("value_33_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor module_layers_15_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_15_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(288127296)))]; - tensor var_2714_cast_fp16 = add(x = q_91_cast_fp16, y = module_layers_15_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2714_cast_fp16")]; - tensor module_layers_15_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_15_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(288129408)))]; - tensor var_2716_cast_fp16 = add(x = q_91_cast_fp16, y = module_layers_15_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2716_cast_fp16")]; - tensor q_with_bias_v_31_perm_0 = const()[name = tensor("q_with_bias_v_31_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor x_337_transpose_x_0 = const()[name = tensor("x_337_transpose_x_0"), val = tensor(false)]; - tensor x_337_transpose_y_0 = const()[name = tensor("x_337_transpose_y_0"), val = tensor(false)]; - tensor op_2718_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(288131520))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(288419584))), name = tensor("op_2718_to_fp16_palettized"), shape = tensor([1, 8, 128, 375])]; - tensor q_with_bias_v_31_cast_fp16 = transpose(perm = q_with_bias_v_31_perm_0, x = var_2716_cast_fp16)[name = tensor("transpose_207")]; - tensor x_337_cast_fp16 = matmul(transpose_x = x_337_transpose_x_0, transpose_y = x_337_transpose_y_0, x = q_with_bias_v_31_cast_fp16, y = op_2718_to_fp16_palettized)[name = tensor("x_337_cast_fp16")]; - tensor x_339_pad_0 = const()[name = tensor("x_339_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_339_mode_0 = const()[name = tensor("x_339_mode_0"), val = tensor("constant")]; - tensor const_164_to_fp16 = const()[name = tensor("const_164_to_fp16"), val = tensor(0x0p+0)]; - tensor x_339_cast_fp16 = pad(constant_val = const_164_to_fp16, mode = x_339_mode_0, pad = x_339_pad_0, x = x_337_cast_fp16)[name = tensor("x_339_cast_fp16")]; - tensor var_2726 = const()[name = tensor("op_2726"), val = tensor([1, 8, -1, 188])]; - tensor x_341_cast_fp16 = reshape(shape = var_2726, x = x_339_cast_fp16)[name = tensor("x_341_cast_fp16")]; - tensor var_2730_begin_0 = const()[name = tensor("op_2730_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_2730_end_0 = const()[name = tensor("op_2730_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_2730_end_mask_0 = const()[name = tensor("op_2730_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_2730_cast_fp16 = slice_by_index(begin = var_2730_begin_0, end = var_2730_end_0, end_mask = var_2730_end_mask_0, x = x_341_cast_fp16)[name = tensor("op_2730_cast_fp16")]; - tensor var_2731 = const()[name = tensor("op_2731"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_61_cast_fp16 = reshape(shape = var_2731, x = var_2730_cast_fp16)[name = tensor("matrix_bd_61_cast_fp16")]; - tensor matrix_ac_31_transpose_x_0 = const()[name = tensor("matrix_ac_31_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_31_transpose_y_0 = const()[name = tensor("matrix_ac_31_transpose_y_0"), val = tensor(false)]; - tensor transpose_126_perm_0 = const()[name = tensor("transpose_126_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_127_perm_0 = const()[name = tensor("transpose_127_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_127 = transpose(perm = transpose_127_perm_0, x = k_61_cast_fp16)[name = tensor("transpose_205")]; - tensor transpose_126 = transpose(perm = transpose_126_perm_0, x = var_2714_cast_fp16)[name = tensor("transpose_206")]; - tensor matrix_ac_31_cast_fp16 = matmul(transpose_x = matrix_ac_31_transpose_x_0, transpose_y = matrix_ac_31_transpose_y_0, x = transpose_126, y = transpose_127)[name = tensor("matrix_ac_31_cast_fp16")]; - tensor matrix_bd_63_begin_0 = const()[name = tensor("matrix_bd_63_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_63_end_0 = const()[name = tensor("matrix_bd_63_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_63_end_mask_0 = const()[name = tensor("matrix_bd_63_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_63_cast_fp16 = slice_by_index(begin = matrix_bd_63_begin_0, end = matrix_bd_63_end_0, end_mask = matrix_bd_63_end_mask_0, x = matrix_bd_61_cast_fp16)[name = tensor("matrix_bd_63_cast_fp16")]; - tensor var_2740_cast_fp16 = add(x = matrix_ac_31_cast_fp16, y = matrix_bd_63_cast_fp16)[name = tensor("op_2740_cast_fp16")]; - tensor _inversed_scores_61_y_0_to_fp16 = const()[name = tensor("_inversed_scores_61_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; - tensor _inversed_scores_61_cast_fp16 = mul(x = var_2740_cast_fp16, y = _inversed_scores_61_y_0_to_fp16)[name = tensor("_inversed_scores_61_cast_fp16")]; - tensor scores_63_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_61_cast_fp16, cond = mask_3)[name = tensor("scores_63_cast_fp16")]; - tensor var_2746_cast_fp16 = softmax(axis = var_30, x = scores_63_cast_fp16)[name = tensor("op_2746_cast_fp16")]; - tensor input_813_cast_fp16 = select(a = var_11_to_fp16, b = var_2746_cast_fp16, cond = mask_3)[name = tensor("input_813_cast_fp16")]; - tensor x_343_transpose_x_0 = const()[name = tensor("x_343_transpose_x_0"), val = tensor(false)]; - tensor x_343_transpose_y_0 = const()[name = tensor("x_343_transpose_y_0"), val = tensor(false)]; - tensor value_33_cast_fp16 = transpose(perm = value_33_perm_0, x = v_31_cast_fp16)[name = tensor("transpose_204")]; - tensor x_343_cast_fp16 = matmul(transpose_x = x_343_transpose_x_0, transpose_y = x_343_transpose_y_0, x = input_813_cast_fp16, y = value_33_cast_fp16)[name = tensor("x_343_cast_fp16")]; - tensor var_2750_perm_0 = const()[name = tensor("op_2750_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_2751 = const()[name = tensor("op_2751"), val = tensor([1, -1, 1024])]; - tensor var_2750_cast_fp16 = transpose(perm = var_2750_perm_0, x = x_343_cast_fp16)[name = tensor("transpose_203")]; - tensor input_815_cast_fp16 = reshape(shape = var_2751, x = var_2750_cast_fp16)[name = tensor("input_815_cast_fp16")]; - tensor module_layers_15_self_attn_linear_out_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(288419776))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(289206272))), name = tensor("module_layers_15_self_attn_linear_out_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_142_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_15_self_attn_linear_out_weight_to_fp16_palettized, x = input_815_cast_fp16)[name = tensor("linear_142_cast_fp16")]; - tensor input_819_cast_fp16 = add(x = input_811_cast_fp16, y = linear_142_cast_fp16)[name = tensor("input_819_cast_fp16")]; - tensor x_347_axes_0 = const()[name = tensor("x_347_axes_0"), val = tensor([-1])]; - tensor module_layers_15_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_15_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(289206464)))]; - tensor module_layers_15_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_15_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(289208576)))]; - tensor x_347_cast_fp16 = layer_norm(axes = x_347_axes_0, beta = module_layers_15_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_15_norm_conv_weight_to_fp16, x = input_819_cast_fp16)[name = tensor("x_347_cast_fp16")]; - tensor input_821_perm_0 = const()[name = tensor("input_821_perm_0"), val = tensor([0, 2, 1])]; - tensor input_823_pad_type_0 = const()[name = tensor("input_823_pad_type_0"), val = tensor("valid")]; - tensor input_823_strides_0 = const()[name = tensor("input_823_strides_0"), val = tensor([1])]; - tensor input_823_pad_0 = const()[name = tensor("input_823_pad_0"), val = tensor([0, 0])]; - tensor input_823_dilations_0 = const()[name = tensor("input_823_dilations_0"), val = tensor([1])]; - tensor input_823_groups_0 = const()[name = tensor("input_823_groups_0"), val = tensor(1)]; - tensor module_layers_15_conv_pointwise_conv1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(289210688))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(290783616))), name = tensor("module_layers_15_conv_pointwise_conv1_weight_to_fp16_palettized"), shape = tensor([2048, 1024, 1])]; - tensor input_821_cast_fp16 = transpose(perm = input_821_perm_0, x = x_347_cast_fp16)[name = tensor("transpose_202")]; - tensor input_823_cast_fp16 = conv(dilations = input_823_dilations_0, groups = input_823_groups_0, pad = input_823_pad_0, pad_type = input_823_pad_type_0, strides = input_823_strides_0, weight = module_layers_15_conv_pointwise_conv1_weight_to_fp16_palettized, x = input_821_cast_fp16)[name = tensor("input_823_cast_fp16")]; - tensor x_349_split_num_splits_0 = const()[name = tensor("x_349_split_num_splits_0"), val = tensor(2)]; - tensor x_349_split_axis_0 = const()[name = tensor("x_349_split_axis_0"), val = tensor(1)]; - tensor x_349_split_cast_fp16_0, tensor x_349_split_cast_fp16_1 = split(axis = x_349_split_axis_0, num_splits = x_349_split_num_splits_0, x = input_823_cast_fp16)[name = tensor("x_349_split_cast_fp16")]; - tensor x_349_split_1_sigmoid_cast_fp16 = sigmoid(x = x_349_split_cast_fp16_1)[name = tensor("x_349_split_1_sigmoid_cast_fp16")]; - tensor x_349_cast_fp16 = mul(x = x_349_split_cast_fp16_0, y = x_349_split_1_sigmoid_cast_fp16)[name = tensor("x_349_cast_fp16")]; - tensor input_825_cast_fp16 = select(a = var_11_to_fp16, b = x_349_cast_fp16, cond = var_328)[name = tensor("input_825_cast_fp16")]; - tensor input_827_pad_0 = const()[name = tensor("input_827_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_827_mode_0 = const()[name = tensor("input_827_mode_0"), val = tensor("constant")]; - tensor const_167_to_fp16 = const()[name = tensor("const_167_to_fp16"), val = tensor(0x0p+0)]; - tensor input_827_cast_fp16 = pad(constant_val = const_167_to_fp16, mode = input_827_mode_0, pad = input_827_pad_0, x = input_825_cast_fp16)[name = tensor("input_827_cast_fp16")]; - tensor input_829_pad_type_0 = const()[name = tensor("input_829_pad_type_0"), val = tensor("valid")]; - tensor input_829_groups_0 = const()[name = tensor("input_829_groups_0"), val = tensor(1024)]; - tensor input_829_strides_0 = const()[name = tensor("input_829_strides_0"), val = tensor([1])]; - tensor input_829_pad_0 = const()[name = tensor("input_829_pad_0"), val = tensor([0, 0])]; - tensor input_829_dilations_0 = const()[name = tensor("input_829_dilations_0"), val = tensor([1])]; - tensor const_278_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(290783808))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(290790784))), name = tensor("const_278_to_fp16_palettized"), shape = tensor([1024, 1, 9])]; - tensor const_279_to_fp16 = const()[name = tensor("const_279_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(290790976)))]; - tensor input_831_cast_fp16 = conv(bias = const_279_to_fp16, dilations = input_829_dilations_0, groups = input_829_groups_0, pad = input_829_pad_0, pad_type = input_829_pad_type_0, strides = input_829_strides_0, weight = const_278_to_fp16_palettized, x = input_827_cast_fp16)[name = tensor("input_831_cast_fp16")]; - tensor input_833_cast_fp16 = silu(x = input_831_cast_fp16)[name = tensor("input_833_cast_fp16")]; - tensor x_351_pad_type_0 = const()[name = tensor("x_351_pad_type_0"), val = tensor("valid")]; - tensor x_351_strides_0 = const()[name = tensor("x_351_strides_0"), val = tensor([1])]; - tensor x_351_pad_0 = const()[name = tensor("x_351_pad_0"), val = tensor([0, 0])]; - tensor x_351_dilations_0 = const()[name = tensor("x_351_dilations_0"), val = tensor([1])]; - tensor x_351_groups_0 = const()[name = tensor("x_351_groups_0"), val = tensor(1)]; - tensor module_layers_15_conv_pointwise_conv2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(290793088))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(291579584))), name = tensor("module_layers_15_conv_pointwise_conv2_weight_to_fp16_palettized"), shape = tensor([1024, 1024, 1])]; - tensor x_351_cast_fp16 = conv(dilations = x_351_dilations_0, groups = x_351_groups_0, pad = x_351_pad_0, pad_type = x_351_pad_type_0, strides = x_351_strides_0, weight = module_layers_15_conv_pointwise_conv2_weight_to_fp16_palettized, x = input_833_cast_fp16)[name = tensor("x_351_cast_fp16")]; - tensor input_835_perm_0 = const()[name = tensor("input_835_perm_0"), val = tensor([0, 2, 1])]; - tensor input_835_cast_fp16 = transpose(perm = input_835_perm_0, x = x_351_cast_fp16)[name = tensor("transpose_201")]; - tensor input_837_cast_fp16 = add(x = input_819_cast_fp16, y = input_835_cast_fp16)[name = tensor("input_837_cast_fp16")]; - tensor input_839_axes_0 = const()[name = tensor("input_839_axes_0"), val = tensor([-1])]; - tensor module_layers_15_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_15_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(291579776)))]; - tensor module_layers_15_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_15_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(291581888)))]; - tensor input_839_cast_fp16 = layer_norm(axes = input_839_axes_0, beta = module_layers_15_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_15_norm_feed_forward2_weight_to_fp16, x = input_837_cast_fp16)[name = tensor("input_839_cast_fp16")]; - tensor module_layers_15_feed_forward2_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(291584000))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(294729792))), name = tensor("module_layers_15_feed_forward2_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_143_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_15_feed_forward2_linear1_weight_to_fp16_palettized, x = input_839_cast_fp16)[name = tensor("linear_143_cast_fp16")]; - tensor input_843_cast_fp16 = silu(x = linear_143_cast_fp16)[name = tensor("input_843_cast_fp16")]; - tensor module_layers_15_feed_forward2_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(294729984))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(297875776))), name = tensor("module_layers_15_feed_forward2_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_144_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_15_feed_forward2_linear2_weight_to_fp16_palettized, x = input_843_cast_fp16)[name = tensor("linear_144_cast_fp16")]; - tensor var_2811_to_fp16 = const()[name = tensor("op_2811_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2812_cast_fp16 = mul(x = linear_144_cast_fp16, y = var_2811_to_fp16)[name = tensor("op_2812_cast_fp16")]; - tensor input_849_cast_fp16 = add(x = input_837_cast_fp16, y = var_2812_cast_fp16)[name = tensor("input_849_cast_fp16")]; - tensor input_851_axes_0 = const()[name = tensor("input_851_axes_0"), val = tensor([-1])]; - tensor module_layers_15_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_15_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(297875968)))]; - tensor module_layers_15_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_15_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(297878080)))]; - tensor input_851_cast_fp16 = layer_norm(axes = input_851_axes_0, beta = module_layers_15_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_15_norm_out_weight_to_fp16, x = input_849_cast_fp16)[name = tensor("input_851_cast_fp16")]; - tensor input_853_axes_0 = const()[name = tensor("input_853_axes_0"), val = tensor([-1])]; - tensor module_layers_16_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_16_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(297880192)))]; - tensor module_layers_16_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_16_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(297882304)))]; - tensor input_853_cast_fp16 = layer_norm(axes = input_853_axes_0, beta = module_layers_16_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_16_norm_feed_forward1_weight_to_fp16, x = input_851_cast_fp16)[name = tensor("input_853_cast_fp16")]; - tensor module_layers_16_feed_forward1_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(297884416))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(301030208))), name = tensor("module_layers_16_feed_forward1_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_145_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_16_feed_forward1_linear1_weight_to_fp16_palettized, x = input_853_cast_fp16)[name = tensor("linear_145_cast_fp16")]; - tensor input_857_cast_fp16 = silu(x = linear_145_cast_fp16)[name = tensor("input_857_cast_fp16")]; - tensor module_layers_16_feed_forward1_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(301030400))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(304176192))), name = tensor("module_layers_16_feed_forward1_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_146_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_16_feed_forward1_linear2_weight_to_fp16_palettized, x = input_857_cast_fp16)[name = tensor("linear_146_cast_fp16")]; - tensor var_2840_to_fp16 = const()[name = tensor("op_2840_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2841_cast_fp16 = mul(x = linear_146_cast_fp16, y = var_2840_to_fp16)[name = tensor("op_2841_cast_fp16")]; - tensor input_863_cast_fp16 = add(x = input_851_cast_fp16, y = var_2841_cast_fp16)[name = tensor("input_863_cast_fp16")]; - tensor query_33_axes_0 = const()[name = tensor("query_33_axes_0"), val = tensor([-1])]; - tensor module_layers_16_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_16_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(304176384)))]; - tensor module_layers_16_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_16_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(304178496)))]; - tensor query_33_cast_fp16 = layer_norm(axes = query_33_axes_0, beta = module_layers_16_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_16_norm_self_att_weight_to_fp16, x = input_863_cast_fp16)[name = tensor("query_33_cast_fp16")]; - tensor module_layers_16_self_attn_linear_q_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(304180608))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(304967104))), name = tensor("module_layers_16_self_attn_linear_q_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_147_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_16_self_attn_linear_q_weight_to_fp16_palettized, x = query_33_cast_fp16)[name = tensor("linear_147_cast_fp16")]; - tensor var_2857 = const()[name = tensor("op_2857"), val = tensor([1, -1, 8, 128])]; - tensor q_97_cast_fp16 = reshape(shape = var_2857, x = linear_147_cast_fp16)[name = tensor("q_97_cast_fp16")]; - tensor module_layers_16_self_attn_linear_k_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(304967296))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(305753792))), name = tensor("module_layers_16_self_attn_linear_k_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_148_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_16_self_attn_linear_k_weight_to_fp16_palettized, x = query_33_cast_fp16)[name = tensor("linear_148_cast_fp16")]; - tensor var_2861 = const()[name = tensor("op_2861"), val = tensor([1, -1, 8, 128])]; - tensor k_65_cast_fp16 = reshape(shape = var_2861, x = linear_148_cast_fp16)[name = tensor("k_65_cast_fp16")]; - tensor module_layers_16_self_attn_linear_v_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(305753984))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(306540480))), name = tensor("module_layers_16_self_attn_linear_v_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_149_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_16_self_attn_linear_v_weight_to_fp16_palettized, x = query_33_cast_fp16)[name = tensor("linear_149_cast_fp16")]; - tensor var_2865 = const()[name = tensor("op_2865"), val = tensor([1, -1, 8, 128])]; - tensor v_33_cast_fp16 = reshape(shape = var_2865, x = linear_149_cast_fp16)[name = tensor("v_33_cast_fp16")]; - tensor value_35_perm_0 = const()[name = tensor("value_35_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor module_layers_16_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_16_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(306540672)))]; - tensor var_2877_cast_fp16 = add(x = q_97_cast_fp16, y = module_layers_16_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2877_cast_fp16")]; - tensor module_layers_16_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_16_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(306542784)))]; - tensor var_2879_cast_fp16 = add(x = q_97_cast_fp16, y = module_layers_16_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2879_cast_fp16")]; - tensor q_with_bias_v_33_perm_0 = const()[name = tensor("q_with_bias_v_33_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor x_359_transpose_x_0 = const()[name = tensor("x_359_transpose_x_0"), val = tensor(false)]; - tensor x_359_transpose_y_0 = const()[name = tensor("x_359_transpose_y_0"), val = tensor(false)]; - tensor op_2881_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(306544896))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(306832960))), name = tensor("op_2881_to_fp16_palettized"), shape = tensor([1, 8, 128, 375])]; - tensor q_with_bias_v_33_cast_fp16 = transpose(perm = q_with_bias_v_33_perm_0, x = var_2879_cast_fp16)[name = tensor("transpose_200")]; - tensor x_359_cast_fp16 = matmul(transpose_x = x_359_transpose_x_0, transpose_y = x_359_transpose_y_0, x = q_with_bias_v_33_cast_fp16, y = op_2881_to_fp16_palettized)[name = tensor("x_359_cast_fp16")]; - tensor x_361_pad_0 = const()[name = tensor("x_361_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_361_mode_0 = const()[name = tensor("x_361_mode_0"), val = tensor("constant")]; - tensor const_174_to_fp16 = const()[name = tensor("const_174_to_fp16"), val = tensor(0x0p+0)]; - tensor x_361_cast_fp16 = pad(constant_val = const_174_to_fp16, mode = x_361_mode_0, pad = x_361_pad_0, x = x_359_cast_fp16)[name = tensor("x_361_cast_fp16")]; - tensor var_2889 = const()[name = tensor("op_2889"), val = tensor([1, 8, -1, 188])]; - tensor x_363_cast_fp16 = reshape(shape = var_2889, x = x_361_cast_fp16)[name = tensor("x_363_cast_fp16")]; - tensor var_2893_begin_0 = const()[name = tensor("op_2893_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_2893_end_0 = const()[name = tensor("op_2893_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_2893_end_mask_0 = const()[name = tensor("op_2893_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_2893_cast_fp16 = slice_by_index(begin = var_2893_begin_0, end = var_2893_end_0, end_mask = var_2893_end_mask_0, x = x_363_cast_fp16)[name = tensor("op_2893_cast_fp16")]; - tensor var_2894 = const()[name = tensor("op_2894"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_65_cast_fp16 = reshape(shape = var_2894, x = var_2893_cast_fp16)[name = tensor("matrix_bd_65_cast_fp16")]; - tensor matrix_ac_33_transpose_x_0 = const()[name = tensor("matrix_ac_33_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_33_transpose_y_0 = const()[name = tensor("matrix_ac_33_transpose_y_0"), val = tensor(false)]; - tensor transpose_128_perm_0 = const()[name = tensor("transpose_128_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_129_perm_0 = const()[name = tensor("transpose_129_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_129 = transpose(perm = transpose_129_perm_0, x = k_65_cast_fp16)[name = tensor("transpose_198")]; - tensor transpose_128 = transpose(perm = transpose_128_perm_0, x = var_2877_cast_fp16)[name = tensor("transpose_199")]; - tensor matrix_ac_33_cast_fp16 = matmul(transpose_x = matrix_ac_33_transpose_x_0, transpose_y = matrix_ac_33_transpose_y_0, x = transpose_128, y = transpose_129)[name = tensor("matrix_ac_33_cast_fp16")]; - tensor matrix_bd_67_begin_0 = const()[name = tensor("matrix_bd_67_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_67_end_0 = const()[name = tensor("matrix_bd_67_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_67_end_mask_0 = const()[name = tensor("matrix_bd_67_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_67_cast_fp16 = slice_by_index(begin = matrix_bd_67_begin_0, end = matrix_bd_67_end_0, end_mask = matrix_bd_67_end_mask_0, x = matrix_bd_65_cast_fp16)[name = tensor("matrix_bd_67_cast_fp16")]; - tensor var_2903_cast_fp16 = add(x = matrix_ac_33_cast_fp16, y = matrix_bd_67_cast_fp16)[name = tensor("op_2903_cast_fp16")]; - tensor _inversed_scores_65_y_0_to_fp16 = const()[name = tensor("_inversed_scores_65_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; - tensor _inversed_scores_65_cast_fp16 = mul(x = var_2903_cast_fp16, y = _inversed_scores_65_y_0_to_fp16)[name = tensor("_inversed_scores_65_cast_fp16")]; - tensor scores_67_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_65_cast_fp16, cond = mask_3)[name = tensor("scores_67_cast_fp16")]; - tensor var_2909_cast_fp16 = softmax(axis = var_30, x = scores_67_cast_fp16)[name = tensor("op_2909_cast_fp16")]; - tensor input_865_cast_fp16 = select(a = var_11_to_fp16, b = var_2909_cast_fp16, cond = mask_3)[name = tensor("input_865_cast_fp16")]; - tensor x_365_transpose_x_0 = const()[name = tensor("x_365_transpose_x_0"), val = tensor(false)]; - tensor x_365_transpose_y_0 = const()[name = tensor("x_365_transpose_y_0"), val = tensor(false)]; - tensor value_35_cast_fp16 = transpose(perm = value_35_perm_0, x = v_33_cast_fp16)[name = tensor("transpose_197")]; - tensor x_365_cast_fp16 = matmul(transpose_x = x_365_transpose_x_0, transpose_y = x_365_transpose_y_0, x = input_865_cast_fp16, y = value_35_cast_fp16)[name = tensor("x_365_cast_fp16")]; - tensor var_2913_perm_0 = const()[name = tensor("op_2913_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_2914 = const()[name = tensor("op_2914"), val = tensor([1, -1, 1024])]; - tensor var_2913_cast_fp16 = transpose(perm = var_2913_perm_0, x = x_365_cast_fp16)[name = tensor("transpose_196")]; - tensor input_867_cast_fp16 = reshape(shape = var_2914, x = var_2913_cast_fp16)[name = tensor("input_867_cast_fp16")]; - tensor module_layers_16_self_attn_linear_out_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(306833152))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(307619648))), name = tensor("module_layers_16_self_attn_linear_out_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_151_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_16_self_attn_linear_out_weight_to_fp16_palettized, x = input_867_cast_fp16)[name = tensor("linear_151_cast_fp16")]; - tensor input_871_cast_fp16 = add(x = input_863_cast_fp16, y = linear_151_cast_fp16)[name = tensor("input_871_cast_fp16")]; - tensor x_369_axes_0 = const()[name = tensor("x_369_axes_0"), val = tensor([-1])]; - tensor module_layers_16_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_16_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(307619840)))]; - tensor module_layers_16_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_16_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(307621952)))]; - tensor x_369_cast_fp16 = layer_norm(axes = x_369_axes_0, beta = module_layers_16_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_16_norm_conv_weight_to_fp16, x = input_871_cast_fp16)[name = tensor("x_369_cast_fp16")]; - tensor input_873_perm_0 = const()[name = tensor("input_873_perm_0"), val = tensor([0, 2, 1])]; - tensor input_875_pad_type_0 = const()[name = tensor("input_875_pad_type_0"), val = tensor("valid")]; - tensor input_875_strides_0 = const()[name = tensor("input_875_strides_0"), val = tensor([1])]; - tensor input_875_pad_0 = const()[name = tensor("input_875_pad_0"), val = tensor([0, 0])]; - tensor input_875_dilations_0 = const()[name = tensor("input_875_dilations_0"), val = tensor([1])]; - tensor input_875_groups_0 = const()[name = tensor("input_875_groups_0"), val = tensor(1)]; - tensor module_layers_16_conv_pointwise_conv1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(307624064))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(309196992))), name = tensor("module_layers_16_conv_pointwise_conv1_weight_to_fp16_palettized"), shape = tensor([2048, 1024, 1])]; - tensor input_873_cast_fp16 = transpose(perm = input_873_perm_0, x = x_369_cast_fp16)[name = tensor("transpose_195")]; - tensor input_875_cast_fp16 = conv(dilations = input_875_dilations_0, groups = input_875_groups_0, pad = input_875_pad_0, pad_type = input_875_pad_type_0, strides = input_875_strides_0, weight = module_layers_16_conv_pointwise_conv1_weight_to_fp16_palettized, x = input_873_cast_fp16)[name = tensor("input_875_cast_fp16")]; - tensor x_371_split_num_splits_0 = const()[name = tensor("x_371_split_num_splits_0"), val = tensor(2)]; - tensor x_371_split_axis_0 = const()[name = tensor("x_371_split_axis_0"), val = tensor(1)]; - tensor x_371_split_cast_fp16_0, tensor x_371_split_cast_fp16_1 = split(axis = x_371_split_axis_0, num_splits = x_371_split_num_splits_0, x = input_875_cast_fp16)[name = tensor("x_371_split_cast_fp16")]; - tensor x_371_split_1_sigmoid_cast_fp16 = sigmoid(x = x_371_split_cast_fp16_1)[name = tensor("x_371_split_1_sigmoid_cast_fp16")]; - tensor x_371_cast_fp16 = mul(x = x_371_split_cast_fp16_0, y = x_371_split_1_sigmoid_cast_fp16)[name = tensor("x_371_cast_fp16")]; - tensor input_877_cast_fp16 = select(a = var_11_to_fp16, b = x_371_cast_fp16, cond = var_328)[name = tensor("input_877_cast_fp16")]; - tensor input_879_pad_0 = const()[name = tensor("input_879_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_879_mode_0 = const()[name = tensor("input_879_mode_0"), val = tensor("constant")]; - tensor const_177_to_fp16 = const()[name = tensor("const_177_to_fp16"), val = tensor(0x0p+0)]; - tensor input_879_cast_fp16 = pad(constant_val = const_177_to_fp16, mode = input_879_mode_0, pad = input_879_pad_0, x = input_877_cast_fp16)[name = tensor("input_879_cast_fp16")]; - tensor input_881_pad_type_0 = const()[name = tensor("input_881_pad_type_0"), val = tensor("valid")]; - tensor input_881_groups_0 = const()[name = tensor("input_881_groups_0"), val = tensor(1024)]; - tensor input_881_strides_0 = const()[name = tensor("input_881_strides_0"), val = tensor([1])]; - tensor input_881_pad_0 = const()[name = tensor("input_881_pad_0"), val = tensor([0, 0])]; - tensor input_881_dilations_0 = const()[name = tensor("input_881_dilations_0"), val = tensor([1])]; - tensor const_280_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(309197184))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(309204160))), name = tensor("const_280_to_fp16_palettized"), shape = tensor([1024, 1, 9])]; - tensor const_281_to_fp16 = const()[name = tensor("const_281_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(309204352)))]; - tensor input_883_cast_fp16 = conv(bias = const_281_to_fp16, dilations = input_881_dilations_0, groups = input_881_groups_0, pad = input_881_pad_0, pad_type = input_881_pad_type_0, strides = input_881_strides_0, weight = const_280_to_fp16_palettized, x = input_879_cast_fp16)[name = tensor("input_883_cast_fp16")]; - tensor input_885_cast_fp16 = silu(x = input_883_cast_fp16)[name = tensor("input_885_cast_fp16")]; - tensor x_373_pad_type_0 = const()[name = tensor("x_373_pad_type_0"), val = tensor("valid")]; - tensor x_373_strides_0 = const()[name = tensor("x_373_strides_0"), val = tensor([1])]; - tensor x_373_pad_0 = const()[name = tensor("x_373_pad_0"), val = tensor([0, 0])]; - tensor x_373_dilations_0 = const()[name = tensor("x_373_dilations_0"), val = tensor([1])]; - tensor x_373_groups_0 = const()[name = tensor("x_373_groups_0"), val = tensor(1)]; - tensor module_layers_16_conv_pointwise_conv2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(309206464))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(309992960))), name = tensor("module_layers_16_conv_pointwise_conv2_weight_to_fp16_palettized"), shape = tensor([1024, 1024, 1])]; - tensor x_373_cast_fp16 = conv(dilations = x_373_dilations_0, groups = x_373_groups_0, pad = x_373_pad_0, pad_type = x_373_pad_type_0, strides = x_373_strides_0, weight = module_layers_16_conv_pointwise_conv2_weight_to_fp16_palettized, x = input_885_cast_fp16)[name = tensor("x_373_cast_fp16")]; - tensor input_887_perm_0 = const()[name = tensor("input_887_perm_0"), val = tensor([0, 2, 1])]; - tensor input_887_cast_fp16 = transpose(perm = input_887_perm_0, x = x_373_cast_fp16)[name = tensor("transpose_194")]; - tensor input_889_cast_fp16 = add(x = input_871_cast_fp16, y = input_887_cast_fp16)[name = tensor("input_889_cast_fp16")]; - tensor input_891_axes_0 = const()[name = tensor("input_891_axes_0"), val = tensor([-1])]; - tensor module_layers_16_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_16_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(309993152)))]; - tensor module_layers_16_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_16_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(309995264)))]; - tensor input_891_cast_fp16 = layer_norm(axes = input_891_axes_0, beta = module_layers_16_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_16_norm_feed_forward2_weight_to_fp16, x = input_889_cast_fp16)[name = tensor("input_891_cast_fp16")]; - tensor module_layers_16_feed_forward2_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(309997376))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(313143168))), name = tensor("module_layers_16_feed_forward2_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_152_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_16_feed_forward2_linear1_weight_to_fp16_palettized, x = input_891_cast_fp16)[name = tensor("linear_152_cast_fp16")]; - tensor input_895_cast_fp16 = silu(x = linear_152_cast_fp16)[name = tensor("input_895_cast_fp16")]; - tensor module_layers_16_feed_forward2_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(313143360))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(316289152))), name = tensor("module_layers_16_feed_forward2_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_153_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_16_feed_forward2_linear2_weight_to_fp16_palettized, x = input_895_cast_fp16)[name = tensor("linear_153_cast_fp16")]; - tensor var_2974_to_fp16 = const()[name = tensor("op_2974_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2975_cast_fp16 = mul(x = linear_153_cast_fp16, y = var_2974_to_fp16)[name = tensor("op_2975_cast_fp16")]; - tensor input_901_cast_fp16 = add(x = input_889_cast_fp16, y = var_2975_cast_fp16)[name = tensor("input_901_cast_fp16")]; - tensor input_903_axes_0 = const()[name = tensor("input_903_axes_0"), val = tensor([-1])]; - tensor module_layers_16_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_16_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(316289344)))]; - tensor module_layers_16_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_16_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(316291456)))]; - tensor input_903_cast_fp16 = layer_norm(axes = input_903_axes_0, beta = module_layers_16_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_16_norm_out_weight_to_fp16, x = input_901_cast_fp16)[name = tensor("input_903_cast_fp16")]; - tensor input_905_axes_0 = const()[name = tensor("input_905_axes_0"), val = tensor([-1])]; - tensor module_layers_17_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_17_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(316293568)))]; - tensor module_layers_17_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_17_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(316295680)))]; - tensor input_905_cast_fp16 = layer_norm(axes = input_905_axes_0, beta = module_layers_17_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_17_norm_feed_forward1_weight_to_fp16, x = input_903_cast_fp16)[name = tensor("input_905_cast_fp16")]; - tensor module_layers_17_feed_forward1_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(316297792))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(319443584))), name = tensor("module_layers_17_feed_forward1_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_154_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_17_feed_forward1_linear1_weight_to_fp16_palettized, x = input_905_cast_fp16)[name = tensor("linear_154_cast_fp16")]; - tensor input_909_cast_fp16 = silu(x = linear_154_cast_fp16)[name = tensor("input_909_cast_fp16")]; - tensor module_layers_17_feed_forward1_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(319443776))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(322589568))), name = tensor("module_layers_17_feed_forward1_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_155_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_17_feed_forward1_linear2_weight_to_fp16_palettized, x = input_909_cast_fp16)[name = tensor("linear_155_cast_fp16")]; - tensor var_3003_to_fp16 = const()[name = tensor("op_3003_to_fp16"), val = tensor(0x1p-1)]; - tensor var_3004_cast_fp16 = mul(x = linear_155_cast_fp16, y = var_3003_to_fp16)[name = tensor("op_3004_cast_fp16")]; - tensor input_915_cast_fp16 = add(x = input_903_cast_fp16, y = var_3004_cast_fp16)[name = tensor("input_915_cast_fp16")]; - tensor query_35_axes_0 = const()[name = tensor("query_35_axes_0"), val = tensor([-1])]; - tensor module_layers_17_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_17_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(322589760)))]; - tensor module_layers_17_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_17_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(322591872)))]; - tensor query_35_cast_fp16 = layer_norm(axes = query_35_axes_0, beta = module_layers_17_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_17_norm_self_att_weight_to_fp16, x = input_915_cast_fp16)[name = tensor("query_35_cast_fp16")]; - tensor module_layers_17_self_attn_linear_q_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(322593984))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(323380480))), name = tensor("module_layers_17_self_attn_linear_q_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_156_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_17_self_attn_linear_q_weight_to_fp16_palettized, x = query_35_cast_fp16)[name = tensor("linear_156_cast_fp16")]; - tensor var_3020 = const()[name = tensor("op_3020"), val = tensor([1, -1, 8, 128])]; - tensor q_103_cast_fp16 = reshape(shape = var_3020, x = linear_156_cast_fp16)[name = tensor("q_103_cast_fp16")]; - tensor module_layers_17_self_attn_linear_k_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(323380672))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(324167168))), name = tensor("module_layers_17_self_attn_linear_k_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_157_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_17_self_attn_linear_k_weight_to_fp16_palettized, x = query_35_cast_fp16)[name = tensor("linear_157_cast_fp16")]; - tensor var_3024 = const()[name = tensor("op_3024"), val = tensor([1, -1, 8, 128])]; - tensor k_69_cast_fp16 = reshape(shape = var_3024, x = linear_157_cast_fp16)[name = tensor("k_69_cast_fp16")]; - tensor module_layers_17_self_attn_linear_v_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(324167360))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(324953856))), name = tensor("module_layers_17_self_attn_linear_v_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_158_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_17_self_attn_linear_v_weight_to_fp16_palettized, x = query_35_cast_fp16)[name = tensor("linear_158_cast_fp16")]; - tensor var_3028 = const()[name = tensor("op_3028"), val = tensor([1, -1, 8, 128])]; - tensor v_35_cast_fp16 = reshape(shape = var_3028, x = linear_158_cast_fp16)[name = tensor("v_35_cast_fp16")]; - tensor value_37_perm_0 = const()[name = tensor("value_37_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor module_layers_17_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_17_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(324954048)))]; - tensor var_3040_cast_fp16 = add(x = q_103_cast_fp16, y = module_layers_17_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3040_cast_fp16")]; - tensor module_layers_17_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_17_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(324956160)))]; - tensor var_3042_cast_fp16 = add(x = q_103_cast_fp16, y = module_layers_17_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3042_cast_fp16")]; - tensor q_with_bias_v_35_perm_0 = const()[name = tensor("q_with_bias_v_35_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor x_381_transpose_x_0 = const()[name = tensor("x_381_transpose_x_0"), val = tensor(false)]; - tensor x_381_transpose_y_0 = const()[name = tensor("x_381_transpose_y_0"), val = tensor(false)]; - tensor op_3044_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(324958272))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(325246336))), name = tensor("op_3044_to_fp16_palettized"), shape = tensor([1, 8, 128, 375])]; - tensor q_with_bias_v_35_cast_fp16 = transpose(perm = q_with_bias_v_35_perm_0, x = var_3042_cast_fp16)[name = tensor("transpose_193")]; - tensor x_381_cast_fp16 = matmul(transpose_x = x_381_transpose_x_0, transpose_y = x_381_transpose_y_0, x = q_with_bias_v_35_cast_fp16, y = op_3044_to_fp16_palettized)[name = tensor("x_381_cast_fp16")]; - tensor x_383_pad_0 = const()[name = tensor("x_383_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_383_mode_0 = const()[name = tensor("x_383_mode_0"), val = tensor("constant")]; - tensor const_184_to_fp16 = const()[name = tensor("const_184_to_fp16"), val = tensor(0x0p+0)]; - tensor x_383_cast_fp16 = pad(constant_val = const_184_to_fp16, mode = x_383_mode_0, pad = x_383_pad_0, x = x_381_cast_fp16)[name = tensor("x_383_cast_fp16")]; - tensor var_3052 = const()[name = tensor("op_3052"), val = tensor([1, 8, -1, 188])]; - tensor x_385_cast_fp16 = reshape(shape = var_3052, x = x_383_cast_fp16)[name = tensor("x_385_cast_fp16")]; - tensor var_3056_begin_0 = const()[name = tensor("op_3056_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_3056_end_0 = const()[name = tensor("op_3056_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_3056_end_mask_0 = const()[name = tensor("op_3056_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_3056_cast_fp16 = slice_by_index(begin = var_3056_begin_0, end = var_3056_end_0, end_mask = var_3056_end_mask_0, x = x_385_cast_fp16)[name = tensor("op_3056_cast_fp16")]; - tensor var_3057 = const()[name = tensor("op_3057"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_69_cast_fp16 = reshape(shape = var_3057, x = var_3056_cast_fp16)[name = tensor("matrix_bd_69_cast_fp16")]; - tensor matrix_ac_35_transpose_x_0 = const()[name = tensor("matrix_ac_35_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_35_transpose_y_0 = const()[name = tensor("matrix_ac_35_transpose_y_0"), val = tensor(false)]; - tensor transpose_130_perm_0 = const()[name = tensor("transpose_130_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_131_perm_0 = const()[name = tensor("transpose_131_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_131 = transpose(perm = transpose_131_perm_0, x = k_69_cast_fp16)[name = tensor("transpose_191")]; - tensor transpose_130 = transpose(perm = transpose_130_perm_0, x = var_3040_cast_fp16)[name = tensor("transpose_192")]; - tensor matrix_ac_35_cast_fp16 = matmul(transpose_x = matrix_ac_35_transpose_x_0, transpose_y = matrix_ac_35_transpose_y_0, x = transpose_130, y = transpose_131)[name = tensor("matrix_ac_35_cast_fp16")]; - tensor matrix_bd_71_begin_0 = const()[name = tensor("matrix_bd_71_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_71_end_0 = const()[name = tensor("matrix_bd_71_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_71_end_mask_0 = const()[name = tensor("matrix_bd_71_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_71_cast_fp16 = slice_by_index(begin = matrix_bd_71_begin_0, end = matrix_bd_71_end_0, end_mask = matrix_bd_71_end_mask_0, x = matrix_bd_69_cast_fp16)[name = tensor("matrix_bd_71_cast_fp16")]; - tensor var_3066_cast_fp16 = add(x = matrix_ac_35_cast_fp16, y = matrix_bd_71_cast_fp16)[name = tensor("op_3066_cast_fp16")]; - tensor _inversed_scores_69_y_0_to_fp16 = const()[name = tensor("_inversed_scores_69_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; - tensor _inversed_scores_69_cast_fp16 = mul(x = var_3066_cast_fp16, y = _inversed_scores_69_y_0_to_fp16)[name = tensor("_inversed_scores_69_cast_fp16")]; - tensor scores_71_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_69_cast_fp16, cond = mask_3)[name = tensor("scores_71_cast_fp16")]; - tensor var_3072_cast_fp16 = softmax(axis = var_30, x = scores_71_cast_fp16)[name = tensor("op_3072_cast_fp16")]; - tensor input_917_cast_fp16 = select(a = var_11_to_fp16, b = var_3072_cast_fp16, cond = mask_3)[name = tensor("input_917_cast_fp16")]; - tensor x_387_transpose_x_0 = const()[name = tensor("x_387_transpose_x_0"), val = tensor(false)]; - tensor x_387_transpose_y_0 = const()[name = tensor("x_387_transpose_y_0"), val = tensor(false)]; - tensor value_37_cast_fp16 = transpose(perm = value_37_perm_0, x = v_35_cast_fp16)[name = tensor("transpose_190")]; - tensor x_387_cast_fp16 = matmul(transpose_x = x_387_transpose_x_0, transpose_y = x_387_transpose_y_0, x = input_917_cast_fp16, y = value_37_cast_fp16)[name = tensor("x_387_cast_fp16")]; - tensor var_3076_perm_0 = const()[name = tensor("op_3076_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_3077 = const()[name = tensor("op_3077"), val = tensor([1, -1, 1024])]; - tensor var_3076_cast_fp16 = transpose(perm = var_3076_perm_0, x = x_387_cast_fp16)[name = tensor("transpose_189")]; - tensor input_919_cast_fp16 = reshape(shape = var_3077, x = var_3076_cast_fp16)[name = tensor("input_919_cast_fp16")]; - tensor module_layers_17_self_attn_linear_out_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(325246528))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(326033024))), name = tensor("module_layers_17_self_attn_linear_out_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_160_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_17_self_attn_linear_out_weight_to_fp16_palettized, x = input_919_cast_fp16)[name = tensor("linear_160_cast_fp16")]; - tensor input_923_cast_fp16 = add(x = input_915_cast_fp16, y = linear_160_cast_fp16)[name = tensor("input_923_cast_fp16")]; - tensor x_391_axes_0 = const()[name = tensor("x_391_axes_0"), val = tensor([-1])]; - tensor module_layers_17_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_17_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(326033216)))]; - tensor module_layers_17_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_17_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(326035328)))]; - tensor x_391_cast_fp16 = layer_norm(axes = x_391_axes_0, beta = module_layers_17_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_17_norm_conv_weight_to_fp16, x = input_923_cast_fp16)[name = tensor("x_391_cast_fp16")]; - tensor input_925_perm_0 = const()[name = tensor("input_925_perm_0"), val = tensor([0, 2, 1])]; - tensor input_927_pad_type_0 = const()[name = tensor("input_927_pad_type_0"), val = tensor("valid")]; - tensor input_927_strides_0 = const()[name = tensor("input_927_strides_0"), val = tensor([1])]; - tensor input_927_pad_0 = const()[name = tensor("input_927_pad_0"), val = tensor([0, 0])]; - tensor input_927_dilations_0 = const()[name = tensor("input_927_dilations_0"), val = tensor([1])]; - tensor input_927_groups_0 = const()[name = tensor("input_927_groups_0"), val = tensor(1)]; - tensor module_layers_17_conv_pointwise_conv1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(326037440))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(327610368))), name = tensor("module_layers_17_conv_pointwise_conv1_weight_to_fp16_palettized"), shape = tensor([2048, 1024, 1])]; - tensor input_925_cast_fp16 = transpose(perm = input_925_perm_0, x = x_391_cast_fp16)[name = tensor("transpose_188")]; - tensor input_927_cast_fp16 = conv(dilations = input_927_dilations_0, groups = input_927_groups_0, pad = input_927_pad_0, pad_type = input_927_pad_type_0, strides = input_927_strides_0, weight = module_layers_17_conv_pointwise_conv1_weight_to_fp16_palettized, x = input_925_cast_fp16)[name = tensor("input_927_cast_fp16")]; - tensor x_393_split_num_splits_0 = const()[name = tensor("x_393_split_num_splits_0"), val = tensor(2)]; - tensor x_393_split_axis_0 = const()[name = tensor("x_393_split_axis_0"), val = tensor(1)]; - tensor x_393_split_cast_fp16_0, tensor x_393_split_cast_fp16_1 = split(axis = x_393_split_axis_0, num_splits = x_393_split_num_splits_0, x = input_927_cast_fp16)[name = tensor("x_393_split_cast_fp16")]; - tensor x_393_split_1_sigmoid_cast_fp16 = sigmoid(x = x_393_split_cast_fp16_1)[name = tensor("x_393_split_1_sigmoid_cast_fp16")]; - tensor x_393_cast_fp16 = mul(x = x_393_split_cast_fp16_0, y = x_393_split_1_sigmoid_cast_fp16)[name = tensor("x_393_cast_fp16")]; - tensor input_929_cast_fp16 = select(a = var_11_to_fp16, b = x_393_cast_fp16, cond = var_328)[name = tensor("input_929_cast_fp16")]; - tensor input_931_pad_0 = const()[name = tensor("input_931_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_931_mode_0 = const()[name = tensor("input_931_mode_0"), val = tensor("constant")]; - tensor const_187_to_fp16 = const()[name = tensor("const_187_to_fp16"), val = tensor(0x0p+0)]; - tensor input_931_cast_fp16 = pad(constant_val = const_187_to_fp16, mode = input_931_mode_0, pad = input_931_pad_0, x = input_929_cast_fp16)[name = tensor("input_931_cast_fp16")]; - tensor input_933_pad_type_0 = const()[name = tensor("input_933_pad_type_0"), val = tensor("valid")]; - tensor input_933_groups_0 = const()[name = tensor("input_933_groups_0"), val = tensor(1024)]; - tensor input_933_strides_0 = const()[name = tensor("input_933_strides_0"), val = tensor([1])]; - tensor input_933_pad_0 = const()[name = tensor("input_933_pad_0"), val = tensor([0, 0])]; - tensor input_933_dilations_0 = const()[name = tensor("input_933_dilations_0"), val = tensor([1])]; - tensor const_282_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(327610560))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(327617536))), name = tensor("const_282_to_fp16_palettized"), shape = tensor([1024, 1, 9])]; - tensor const_283_to_fp16 = const()[name = tensor("const_283_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(327617728)))]; - tensor input_935_cast_fp16 = conv(bias = const_283_to_fp16, dilations = input_933_dilations_0, groups = input_933_groups_0, pad = input_933_pad_0, pad_type = input_933_pad_type_0, strides = input_933_strides_0, weight = const_282_to_fp16_palettized, x = input_931_cast_fp16)[name = tensor("input_935_cast_fp16")]; - tensor input_937_cast_fp16 = silu(x = input_935_cast_fp16)[name = tensor("input_937_cast_fp16")]; - tensor x_395_pad_type_0 = const()[name = tensor("x_395_pad_type_0"), val = tensor("valid")]; - tensor x_395_strides_0 = const()[name = tensor("x_395_strides_0"), val = tensor([1])]; - tensor x_395_pad_0 = const()[name = tensor("x_395_pad_0"), val = tensor([0, 0])]; - tensor x_395_dilations_0 = const()[name = tensor("x_395_dilations_0"), val = tensor([1])]; - tensor x_395_groups_0 = const()[name = tensor("x_395_groups_0"), val = tensor(1)]; - tensor module_layers_17_conv_pointwise_conv2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(327619840))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(328406336))), name = tensor("module_layers_17_conv_pointwise_conv2_weight_to_fp16_palettized"), shape = tensor([1024, 1024, 1])]; - tensor x_395_cast_fp16 = conv(dilations = x_395_dilations_0, groups = x_395_groups_0, pad = x_395_pad_0, pad_type = x_395_pad_type_0, strides = x_395_strides_0, weight = module_layers_17_conv_pointwise_conv2_weight_to_fp16_palettized, x = input_937_cast_fp16)[name = tensor("x_395_cast_fp16")]; - tensor input_939_perm_0 = const()[name = tensor("input_939_perm_0"), val = tensor([0, 2, 1])]; - tensor input_939_cast_fp16 = transpose(perm = input_939_perm_0, x = x_395_cast_fp16)[name = tensor("transpose_187")]; - tensor input_941_cast_fp16 = add(x = input_923_cast_fp16, y = input_939_cast_fp16)[name = tensor("input_941_cast_fp16")]; - tensor input_943_axes_0 = const()[name = tensor("input_943_axes_0"), val = tensor([-1])]; - tensor module_layers_17_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_17_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(328406528)))]; - tensor module_layers_17_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_17_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(328408640)))]; - tensor input_943_cast_fp16 = layer_norm(axes = input_943_axes_0, beta = module_layers_17_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_17_norm_feed_forward2_weight_to_fp16, x = input_941_cast_fp16)[name = tensor("input_943_cast_fp16")]; - tensor module_layers_17_feed_forward2_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(328410752))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(331556544))), name = tensor("module_layers_17_feed_forward2_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_161_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_17_feed_forward2_linear1_weight_to_fp16_palettized, x = input_943_cast_fp16)[name = tensor("linear_161_cast_fp16")]; - tensor input_947_cast_fp16 = silu(x = linear_161_cast_fp16)[name = tensor("input_947_cast_fp16")]; - tensor module_layers_17_feed_forward2_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(331556736))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(334702528))), name = tensor("module_layers_17_feed_forward2_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_162_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_17_feed_forward2_linear2_weight_to_fp16_palettized, x = input_947_cast_fp16)[name = tensor("linear_162_cast_fp16")]; - tensor var_3137_to_fp16 = const()[name = tensor("op_3137_to_fp16"), val = tensor(0x1p-1)]; - tensor var_3138_cast_fp16 = mul(x = linear_162_cast_fp16, y = var_3137_to_fp16)[name = tensor("op_3138_cast_fp16")]; - tensor input_953_cast_fp16 = add(x = input_941_cast_fp16, y = var_3138_cast_fp16)[name = tensor("input_953_cast_fp16")]; - tensor input_955_axes_0 = const()[name = tensor("input_955_axes_0"), val = tensor([-1])]; - tensor module_layers_17_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_17_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(334702720)))]; - tensor module_layers_17_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_17_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(334704832)))]; - tensor input_955_cast_fp16 = layer_norm(axes = input_955_axes_0, beta = module_layers_17_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_17_norm_out_weight_to_fp16, x = input_953_cast_fp16)[name = tensor("input_955_cast_fp16")]; - tensor input_957_axes_0 = const()[name = tensor("input_957_axes_0"), val = tensor([-1])]; - tensor module_layers_18_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_18_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(334706944)))]; - tensor module_layers_18_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_18_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(334709056)))]; - tensor input_957_cast_fp16 = layer_norm(axes = input_957_axes_0, beta = module_layers_18_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_18_norm_feed_forward1_weight_to_fp16, x = input_955_cast_fp16)[name = tensor("input_957_cast_fp16")]; - tensor module_layers_18_feed_forward1_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(334711168))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(337856960))), name = tensor("module_layers_18_feed_forward1_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_163_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_18_feed_forward1_linear1_weight_to_fp16_palettized, x = input_957_cast_fp16)[name = tensor("linear_163_cast_fp16")]; - tensor input_961_cast_fp16 = silu(x = linear_163_cast_fp16)[name = tensor("input_961_cast_fp16")]; - tensor module_layers_18_feed_forward1_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(337857152))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(341002944))), name = tensor("module_layers_18_feed_forward1_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_164_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_18_feed_forward1_linear2_weight_to_fp16_palettized, x = input_961_cast_fp16)[name = tensor("linear_164_cast_fp16")]; - tensor var_3166_to_fp16 = const()[name = tensor("op_3166_to_fp16"), val = tensor(0x1p-1)]; - tensor var_3167_cast_fp16 = mul(x = linear_164_cast_fp16, y = var_3166_to_fp16)[name = tensor("op_3167_cast_fp16")]; - tensor input_967_cast_fp16 = add(x = input_955_cast_fp16, y = var_3167_cast_fp16)[name = tensor("input_967_cast_fp16")]; - tensor query_37_axes_0 = const()[name = tensor("query_37_axes_0"), val = tensor([-1])]; - tensor module_layers_18_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_18_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(341003136)))]; - tensor module_layers_18_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_18_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(341005248)))]; - tensor query_37_cast_fp16 = layer_norm(axes = query_37_axes_0, beta = module_layers_18_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_18_norm_self_att_weight_to_fp16, x = input_967_cast_fp16)[name = tensor("query_37_cast_fp16")]; - tensor module_layers_18_self_attn_linear_q_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(341007360))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(341793856))), name = tensor("module_layers_18_self_attn_linear_q_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_165_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_18_self_attn_linear_q_weight_to_fp16_palettized, x = query_37_cast_fp16)[name = tensor("linear_165_cast_fp16")]; - tensor var_3183 = const()[name = tensor("op_3183"), val = tensor([1, -1, 8, 128])]; - tensor q_109_cast_fp16 = reshape(shape = var_3183, x = linear_165_cast_fp16)[name = tensor("q_109_cast_fp16")]; - tensor module_layers_18_self_attn_linear_k_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(341794048))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(342580544))), name = tensor("module_layers_18_self_attn_linear_k_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_166_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_18_self_attn_linear_k_weight_to_fp16_palettized, x = query_37_cast_fp16)[name = tensor("linear_166_cast_fp16")]; - tensor var_3187 = const()[name = tensor("op_3187"), val = tensor([1, -1, 8, 128])]; - tensor k_73_cast_fp16 = reshape(shape = var_3187, x = linear_166_cast_fp16)[name = tensor("k_73_cast_fp16")]; - tensor module_layers_18_self_attn_linear_v_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(342580736))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(343367232))), name = tensor("module_layers_18_self_attn_linear_v_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_167_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_18_self_attn_linear_v_weight_to_fp16_palettized, x = query_37_cast_fp16)[name = tensor("linear_167_cast_fp16")]; - tensor var_3191 = const()[name = tensor("op_3191"), val = tensor([1, -1, 8, 128])]; - tensor v_37_cast_fp16 = reshape(shape = var_3191, x = linear_167_cast_fp16)[name = tensor("v_37_cast_fp16")]; - tensor value_39_perm_0 = const()[name = tensor("value_39_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor module_layers_18_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_18_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(343367424)))]; - tensor var_3203_cast_fp16 = add(x = q_109_cast_fp16, y = module_layers_18_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3203_cast_fp16")]; - tensor module_layers_18_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_18_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(343369536)))]; - tensor var_3205_cast_fp16 = add(x = q_109_cast_fp16, y = module_layers_18_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3205_cast_fp16")]; - tensor q_with_bias_v_37_perm_0 = const()[name = tensor("q_with_bias_v_37_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor x_403_transpose_x_0 = const()[name = tensor("x_403_transpose_x_0"), val = tensor(false)]; - tensor x_403_transpose_y_0 = const()[name = tensor("x_403_transpose_y_0"), val = tensor(false)]; - tensor op_3207_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(343371648))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(343659712))), name = tensor("op_3207_to_fp16_palettized"), shape = tensor([1, 8, 128, 375])]; - tensor q_with_bias_v_37_cast_fp16 = transpose(perm = q_with_bias_v_37_perm_0, x = var_3205_cast_fp16)[name = tensor("transpose_186")]; - tensor x_403_cast_fp16 = matmul(transpose_x = x_403_transpose_x_0, transpose_y = x_403_transpose_y_0, x = q_with_bias_v_37_cast_fp16, y = op_3207_to_fp16_palettized)[name = tensor("x_403_cast_fp16")]; - tensor x_405_pad_0 = const()[name = tensor("x_405_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_405_mode_0 = const()[name = tensor("x_405_mode_0"), val = tensor("constant")]; - tensor const_194_to_fp16 = const()[name = tensor("const_194_to_fp16"), val = tensor(0x0p+0)]; - tensor x_405_cast_fp16 = pad(constant_val = const_194_to_fp16, mode = x_405_mode_0, pad = x_405_pad_0, x = x_403_cast_fp16)[name = tensor("x_405_cast_fp16")]; - tensor var_3215 = const()[name = tensor("op_3215"), val = tensor([1, 8, -1, 188])]; - tensor x_407_cast_fp16 = reshape(shape = var_3215, x = x_405_cast_fp16)[name = tensor("x_407_cast_fp16")]; - tensor var_3219_begin_0 = const()[name = tensor("op_3219_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_3219_end_0 = const()[name = tensor("op_3219_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_3219_end_mask_0 = const()[name = tensor("op_3219_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_3219_cast_fp16 = slice_by_index(begin = var_3219_begin_0, end = var_3219_end_0, end_mask = var_3219_end_mask_0, x = x_407_cast_fp16)[name = tensor("op_3219_cast_fp16")]; - tensor var_3220 = const()[name = tensor("op_3220"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_73_cast_fp16 = reshape(shape = var_3220, x = var_3219_cast_fp16)[name = tensor("matrix_bd_73_cast_fp16")]; - tensor matrix_ac_37_transpose_x_0 = const()[name = tensor("matrix_ac_37_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_37_transpose_y_0 = const()[name = tensor("matrix_ac_37_transpose_y_0"), val = tensor(false)]; - tensor transpose_132_perm_0 = const()[name = tensor("transpose_132_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_133_perm_0 = const()[name = tensor("transpose_133_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_133 = transpose(perm = transpose_133_perm_0, x = k_73_cast_fp16)[name = tensor("transpose_184")]; - tensor transpose_132 = transpose(perm = transpose_132_perm_0, x = var_3203_cast_fp16)[name = tensor("transpose_185")]; - tensor matrix_ac_37_cast_fp16 = matmul(transpose_x = matrix_ac_37_transpose_x_0, transpose_y = matrix_ac_37_transpose_y_0, x = transpose_132, y = transpose_133)[name = tensor("matrix_ac_37_cast_fp16")]; - tensor matrix_bd_75_begin_0 = const()[name = tensor("matrix_bd_75_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_75_end_0 = const()[name = tensor("matrix_bd_75_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_75_end_mask_0 = const()[name = tensor("matrix_bd_75_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_75_cast_fp16 = slice_by_index(begin = matrix_bd_75_begin_0, end = matrix_bd_75_end_0, end_mask = matrix_bd_75_end_mask_0, x = matrix_bd_73_cast_fp16)[name = tensor("matrix_bd_75_cast_fp16")]; - tensor var_3229_cast_fp16 = add(x = matrix_ac_37_cast_fp16, y = matrix_bd_75_cast_fp16)[name = tensor("op_3229_cast_fp16")]; - tensor _inversed_scores_73_y_0_to_fp16 = const()[name = tensor("_inversed_scores_73_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; - tensor _inversed_scores_73_cast_fp16 = mul(x = var_3229_cast_fp16, y = _inversed_scores_73_y_0_to_fp16)[name = tensor("_inversed_scores_73_cast_fp16")]; - tensor scores_75_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_73_cast_fp16, cond = mask_3)[name = tensor("scores_75_cast_fp16")]; - tensor var_3235_cast_fp16 = softmax(axis = var_30, x = scores_75_cast_fp16)[name = tensor("op_3235_cast_fp16")]; - tensor input_969_cast_fp16 = select(a = var_11_to_fp16, b = var_3235_cast_fp16, cond = mask_3)[name = tensor("input_969_cast_fp16")]; - tensor x_409_transpose_x_0 = const()[name = tensor("x_409_transpose_x_0"), val = tensor(false)]; - tensor x_409_transpose_y_0 = const()[name = tensor("x_409_transpose_y_0"), val = tensor(false)]; - tensor value_39_cast_fp16 = transpose(perm = value_39_perm_0, x = v_37_cast_fp16)[name = tensor("transpose_183")]; - tensor x_409_cast_fp16 = matmul(transpose_x = x_409_transpose_x_0, transpose_y = x_409_transpose_y_0, x = input_969_cast_fp16, y = value_39_cast_fp16)[name = tensor("x_409_cast_fp16")]; - tensor var_3239_perm_0 = const()[name = tensor("op_3239_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_3240 = const()[name = tensor("op_3240"), val = tensor([1, -1, 1024])]; - tensor var_3239_cast_fp16 = transpose(perm = var_3239_perm_0, x = x_409_cast_fp16)[name = tensor("transpose_182")]; - tensor input_971_cast_fp16 = reshape(shape = var_3240, x = var_3239_cast_fp16)[name = tensor("input_971_cast_fp16")]; - tensor module_layers_18_self_attn_linear_out_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(343659904))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(344446400))), name = tensor("module_layers_18_self_attn_linear_out_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_169_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_18_self_attn_linear_out_weight_to_fp16_palettized, x = input_971_cast_fp16)[name = tensor("linear_169_cast_fp16")]; - tensor input_975_cast_fp16 = add(x = input_967_cast_fp16, y = linear_169_cast_fp16)[name = tensor("input_975_cast_fp16")]; - tensor x_413_axes_0 = const()[name = tensor("x_413_axes_0"), val = tensor([-1])]; - tensor module_layers_18_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_18_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(344446592)))]; - tensor module_layers_18_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_18_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(344448704)))]; - tensor x_413_cast_fp16 = layer_norm(axes = x_413_axes_0, beta = module_layers_18_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_18_norm_conv_weight_to_fp16, x = input_975_cast_fp16)[name = tensor("x_413_cast_fp16")]; - tensor input_977_perm_0 = const()[name = tensor("input_977_perm_0"), val = tensor([0, 2, 1])]; - tensor input_979_pad_type_0 = const()[name = tensor("input_979_pad_type_0"), val = tensor("valid")]; - tensor input_979_strides_0 = const()[name = tensor("input_979_strides_0"), val = tensor([1])]; - tensor input_979_pad_0 = const()[name = tensor("input_979_pad_0"), val = tensor([0, 0])]; - tensor input_979_dilations_0 = const()[name = tensor("input_979_dilations_0"), val = tensor([1])]; - tensor input_979_groups_0 = const()[name = tensor("input_979_groups_0"), val = tensor(1)]; - tensor module_layers_18_conv_pointwise_conv1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(344450816))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(346023744))), name = tensor("module_layers_18_conv_pointwise_conv1_weight_to_fp16_palettized"), shape = tensor([2048, 1024, 1])]; - tensor input_977_cast_fp16 = transpose(perm = input_977_perm_0, x = x_413_cast_fp16)[name = tensor("transpose_181")]; - tensor input_979_cast_fp16 = conv(dilations = input_979_dilations_0, groups = input_979_groups_0, pad = input_979_pad_0, pad_type = input_979_pad_type_0, strides = input_979_strides_0, weight = module_layers_18_conv_pointwise_conv1_weight_to_fp16_palettized, x = input_977_cast_fp16)[name = tensor("input_979_cast_fp16")]; - tensor x_415_split_num_splits_0 = const()[name = tensor("x_415_split_num_splits_0"), val = tensor(2)]; - tensor x_415_split_axis_0 = const()[name = tensor("x_415_split_axis_0"), val = tensor(1)]; - tensor x_415_split_cast_fp16_0, tensor x_415_split_cast_fp16_1 = split(axis = x_415_split_axis_0, num_splits = x_415_split_num_splits_0, x = input_979_cast_fp16)[name = tensor("x_415_split_cast_fp16")]; - tensor x_415_split_1_sigmoid_cast_fp16 = sigmoid(x = x_415_split_cast_fp16_1)[name = tensor("x_415_split_1_sigmoid_cast_fp16")]; - tensor x_415_cast_fp16 = mul(x = x_415_split_cast_fp16_0, y = x_415_split_1_sigmoid_cast_fp16)[name = tensor("x_415_cast_fp16")]; - tensor input_981_cast_fp16 = select(a = var_11_to_fp16, b = x_415_cast_fp16, cond = var_328)[name = tensor("input_981_cast_fp16")]; - tensor input_983_pad_0 = const()[name = tensor("input_983_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_983_mode_0 = const()[name = tensor("input_983_mode_0"), val = tensor("constant")]; - tensor const_197_to_fp16 = const()[name = tensor("const_197_to_fp16"), val = tensor(0x0p+0)]; - tensor input_983_cast_fp16 = pad(constant_val = const_197_to_fp16, mode = input_983_mode_0, pad = input_983_pad_0, x = input_981_cast_fp16)[name = tensor("input_983_cast_fp16")]; - tensor input_985_pad_type_0 = const()[name = tensor("input_985_pad_type_0"), val = tensor("valid")]; - tensor input_985_groups_0 = const()[name = tensor("input_985_groups_0"), val = tensor(1024)]; - tensor input_985_strides_0 = const()[name = tensor("input_985_strides_0"), val = tensor([1])]; - tensor input_985_pad_0 = const()[name = tensor("input_985_pad_0"), val = tensor([0, 0])]; - tensor input_985_dilations_0 = const()[name = tensor("input_985_dilations_0"), val = tensor([1])]; - tensor const_284_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(346023936))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(346030912))), name = tensor("const_284_to_fp16_palettized"), shape = tensor([1024, 1, 9])]; - tensor const_285_to_fp16 = const()[name = tensor("const_285_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(346031104)))]; - tensor input_987_cast_fp16 = conv(bias = const_285_to_fp16, dilations = input_985_dilations_0, groups = input_985_groups_0, pad = input_985_pad_0, pad_type = input_985_pad_type_0, strides = input_985_strides_0, weight = const_284_to_fp16_palettized, x = input_983_cast_fp16)[name = tensor("input_987_cast_fp16")]; - tensor input_989_cast_fp16 = silu(x = input_987_cast_fp16)[name = tensor("input_989_cast_fp16")]; - tensor x_417_pad_type_0 = const()[name = tensor("x_417_pad_type_0"), val = tensor("valid")]; - tensor x_417_strides_0 = const()[name = tensor("x_417_strides_0"), val = tensor([1])]; - tensor x_417_pad_0 = const()[name = tensor("x_417_pad_0"), val = tensor([0, 0])]; - tensor x_417_dilations_0 = const()[name = tensor("x_417_dilations_0"), val = tensor([1])]; - tensor x_417_groups_0 = const()[name = tensor("x_417_groups_0"), val = tensor(1)]; - tensor module_layers_18_conv_pointwise_conv2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(346033216))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(346819712))), name = tensor("module_layers_18_conv_pointwise_conv2_weight_to_fp16_palettized"), shape = tensor([1024, 1024, 1])]; - tensor x_417_cast_fp16 = conv(dilations = x_417_dilations_0, groups = x_417_groups_0, pad = x_417_pad_0, pad_type = x_417_pad_type_0, strides = x_417_strides_0, weight = module_layers_18_conv_pointwise_conv2_weight_to_fp16_palettized, x = input_989_cast_fp16)[name = tensor("x_417_cast_fp16")]; - tensor input_991_perm_0 = const()[name = tensor("input_991_perm_0"), val = tensor([0, 2, 1])]; - tensor input_991_cast_fp16 = transpose(perm = input_991_perm_0, x = x_417_cast_fp16)[name = tensor("transpose_180")]; - tensor input_993_cast_fp16 = add(x = input_975_cast_fp16, y = input_991_cast_fp16)[name = tensor("input_993_cast_fp16")]; - tensor input_995_axes_0 = const()[name = tensor("input_995_axes_0"), val = tensor([-1])]; - tensor module_layers_18_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_18_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(346819904)))]; - tensor module_layers_18_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_18_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(346822016)))]; - tensor input_995_cast_fp16 = layer_norm(axes = input_995_axes_0, beta = module_layers_18_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_18_norm_feed_forward2_weight_to_fp16, x = input_993_cast_fp16)[name = tensor("input_995_cast_fp16")]; - tensor module_layers_18_feed_forward2_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(346824128))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(349969920))), name = tensor("module_layers_18_feed_forward2_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_170_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_18_feed_forward2_linear1_weight_to_fp16_palettized, x = input_995_cast_fp16)[name = tensor("linear_170_cast_fp16")]; - tensor input_999_cast_fp16 = silu(x = linear_170_cast_fp16)[name = tensor("input_999_cast_fp16")]; - tensor module_layers_18_feed_forward2_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(349970112))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(353115904))), name = tensor("module_layers_18_feed_forward2_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_171_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_18_feed_forward2_linear2_weight_to_fp16_palettized, x = input_999_cast_fp16)[name = tensor("linear_171_cast_fp16")]; - tensor var_3300_to_fp16 = const()[name = tensor("op_3300_to_fp16"), val = tensor(0x1p-1)]; - tensor var_3301_cast_fp16 = mul(x = linear_171_cast_fp16, y = var_3300_to_fp16)[name = tensor("op_3301_cast_fp16")]; - tensor input_1005_cast_fp16 = add(x = input_993_cast_fp16, y = var_3301_cast_fp16)[name = tensor("input_1005_cast_fp16")]; - tensor input_1007_axes_0 = const()[name = tensor("input_1007_axes_0"), val = tensor([-1])]; - tensor module_layers_18_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_18_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(353116096)))]; - tensor module_layers_18_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_18_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(353118208)))]; - tensor input_1007_cast_fp16 = layer_norm(axes = input_1007_axes_0, beta = module_layers_18_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_18_norm_out_weight_to_fp16, x = input_1005_cast_fp16)[name = tensor("input_1007_cast_fp16")]; - tensor input_1009_axes_0 = const()[name = tensor("input_1009_axes_0"), val = tensor([-1])]; - tensor module_layers_19_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_19_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(353120320)))]; - tensor module_layers_19_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_19_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(353122432)))]; - tensor input_1009_cast_fp16 = layer_norm(axes = input_1009_axes_0, beta = module_layers_19_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_19_norm_feed_forward1_weight_to_fp16, x = input_1007_cast_fp16)[name = tensor("input_1009_cast_fp16")]; - tensor module_layers_19_feed_forward1_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(353124544))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(356270336))), name = tensor("module_layers_19_feed_forward1_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_172_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_19_feed_forward1_linear1_weight_to_fp16_palettized, x = input_1009_cast_fp16)[name = tensor("linear_172_cast_fp16")]; - tensor input_1013_cast_fp16 = silu(x = linear_172_cast_fp16)[name = tensor("input_1013_cast_fp16")]; - tensor module_layers_19_feed_forward1_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(356270528))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(359416320))), name = tensor("module_layers_19_feed_forward1_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_173_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_19_feed_forward1_linear2_weight_to_fp16_palettized, x = input_1013_cast_fp16)[name = tensor("linear_173_cast_fp16")]; - tensor var_3329_to_fp16 = const()[name = tensor("op_3329_to_fp16"), val = tensor(0x1p-1)]; - tensor var_3330_cast_fp16 = mul(x = linear_173_cast_fp16, y = var_3329_to_fp16)[name = tensor("op_3330_cast_fp16")]; - tensor input_1019_cast_fp16 = add(x = input_1007_cast_fp16, y = var_3330_cast_fp16)[name = tensor("input_1019_cast_fp16")]; - tensor query_39_axes_0 = const()[name = tensor("query_39_axes_0"), val = tensor([-1])]; - tensor module_layers_19_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_19_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(359416512)))]; - tensor module_layers_19_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_19_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(359418624)))]; - tensor query_39_cast_fp16 = layer_norm(axes = query_39_axes_0, beta = module_layers_19_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_19_norm_self_att_weight_to_fp16, x = input_1019_cast_fp16)[name = tensor("query_39_cast_fp16")]; - tensor module_layers_19_self_attn_linear_q_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(359420736))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(360207232))), name = tensor("module_layers_19_self_attn_linear_q_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_174_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_19_self_attn_linear_q_weight_to_fp16_palettized, x = query_39_cast_fp16)[name = tensor("linear_174_cast_fp16")]; - tensor var_3346 = const()[name = tensor("op_3346"), val = tensor([1, -1, 8, 128])]; - tensor q_115_cast_fp16 = reshape(shape = var_3346, x = linear_174_cast_fp16)[name = tensor("q_115_cast_fp16")]; - tensor module_layers_19_self_attn_linear_k_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(360207424))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(360993920))), name = tensor("module_layers_19_self_attn_linear_k_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_175_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_19_self_attn_linear_k_weight_to_fp16_palettized, x = query_39_cast_fp16)[name = tensor("linear_175_cast_fp16")]; - tensor var_3350 = const()[name = tensor("op_3350"), val = tensor([1, -1, 8, 128])]; - tensor k_77_cast_fp16 = reshape(shape = var_3350, x = linear_175_cast_fp16)[name = tensor("k_77_cast_fp16")]; - tensor module_layers_19_self_attn_linear_v_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(360994112))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(361780608))), name = tensor("module_layers_19_self_attn_linear_v_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_176_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_19_self_attn_linear_v_weight_to_fp16_palettized, x = query_39_cast_fp16)[name = tensor("linear_176_cast_fp16")]; - tensor var_3354 = const()[name = tensor("op_3354"), val = tensor([1, -1, 8, 128])]; - tensor v_39_cast_fp16 = reshape(shape = var_3354, x = linear_176_cast_fp16)[name = tensor("v_39_cast_fp16")]; - tensor value_41_perm_0 = const()[name = tensor("value_41_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor module_layers_19_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_19_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(361780800)))]; - tensor var_3366_cast_fp16 = add(x = q_115_cast_fp16, y = module_layers_19_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3366_cast_fp16")]; - tensor module_layers_19_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_19_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(361782912)))]; - tensor var_3368_cast_fp16 = add(x = q_115_cast_fp16, y = module_layers_19_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3368_cast_fp16")]; - tensor q_with_bias_v_39_perm_0 = const()[name = tensor("q_with_bias_v_39_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor x_425_transpose_x_0 = const()[name = tensor("x_425_transpose_x_0"), val = tensor(false)]; - tensor x_425_transpose_y_0 = const()[name = tensor("x_425_transpose_y_0"), val = tensor(false)]; - tensor op_3370_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(361785024))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(362073088))), name = tensor("op_3370_to_fp16_palettized"), shape = tensor([1, 8, 128, 375])]; - tensor q_with_bias_v_39_cast_fp16 = transpose(perm = q_with_bias_v_39_perm_0, x = var_3368_cast_fp16)[name = tensor("transpose_179")]; - tensor x_425_cast_fp16 = matmul(transpose_x = x_425_transpose_x_0, transpose_y = x_425_transpose_y_0, x = q_with_bias_v_39_cast_fp16, y = op_3370_to_fp16_palettized)[name = tensor("x_425_cast_fp16")]; - tensor x_427_pad_0 = const()[name = tensor("x_427_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_427_mode_0 = const()[name = tensor("x_427_mode_0"), val = tensor("constant")]; - tensor const_204_to_fp16 = const()[name = tensor("const_204_to_fp16"), val = tensor(0x0p+0)]; - tensor x_427_cast_fp16 = pad(constant_val = const_204_to_fp16, mode = x_427_mode_0, pad = x_427_pad_0, x = x_425_cast_fp16)[name = tensor("x_427_cast_fp16")]; - tensor var_3378 = const()[name = tensor("op_3378"), val = tensor([1, 8, -1, 188])]; - tensor x_429_cast_fp16 = reshape(shape = var_3378, x = x_427_cast_fp16)[name = tensor("x_429_cast_fp16")]; - tensor var_3382_begin_0 = const()[name = tensor("op_3382_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_3382_end_0 = const()[name = tensor("op_3382_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_3382_end_mask_0 = const()[name = tensor("op_3382_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_3382_cast_fp16 = slice_by_index(begin = var_3382_begin_0, end = var_3382_end_0, end_mask = var_3382_end_mask_0, x = x_429_cast_fp16)[name = tensor("op_3382_cast_fp16")]; - tensor var_3383 = const()[name = tensor("op_3383"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_77_cast_fp16 = reshape(shape = var_3383, x = var_3382_cast_fp16)[name = tensor("matrix_bd_77_cast_fp16")]; - tensor matrix_ac_39_transpose_x_0 = const()[name = tensor("matrix_ac_39_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_39_transpose_y_0 = const()[name = tensor("matrix_ac_39_transpose_y_0"), val = tensor(false)]; - tensor transpose_134_perm_0 = const()[name = tensor("transpose_134_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_135_perm_0 = const()[name = tensor("transpose_135_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_135 = transpose(perm = transpose_135_perm_0, x = k_77_cast_fp16)[name = tensor("transpose_177")]; - tensor transpose_134 = transpose(perm = transpose_134_perm_0, x = var_3366_cast_fp16)[name = tensor("transpose_178")]; - tensor matrix_ac_39_cast_fp16 = matmul(transpose_x = matrix_ac_39_transpose_x_0, transpose_y = matrix_ac_39_transpose_y_0, x = transpose_134, y = transpose_135)[name = tensor("matrix_ac_39_cast_fp16")]; - tensor matrix_bd_79_begin_0 = const()[name = tensor("matrix_bd_79_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_79_end_0 = const()[name = tensor("matrix_bd_79_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_79_end_mask_0 = const()[name = tensor("matrix_bd_79_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_79_cast_fp16 = slice_by_index(begin = matrix_bd_79_begin_0, end = matrix_bd_79_end_0, end_mask = matrix_bd_79_end_mask_0, x = matrix_bd_77_cast_fp16)[name = tensor("matrix_bd_79_cast_fp16")]; - tensor var_3392_cast_fp16 = add(x = matrix_ac_39_cast_fp16, y = matrix_bd_79_cast_fp16)[name = tensor("op_3392_cast_fp16")]; - tensor _inversed_scores_77_y_0_to_fp16 = const()[name = tensor("_inversed_scores_77_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; - tensor _inversed_scores_77_cast_fp16 = mul(x = var_3392_cast_fp16, y = _inversed_scores_77_y_0_to_fp16)[name = tensor("_inversed_scores_77_cast_fp16")]; - tensor scores_79_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_77_cast_fp16, cond = mask_3)[name = tensor("scores_79_cast_fp16")]; - tensor var_3398_cast_fp16 = softmax(axis = var_30, x = scores_79_cast_fp16)[name = tensor("op_3398_cast_fp16")]; - tensor input_1021_cast_fp16 = select(a = var_11_to_fp16, b = var_3398_cast_fp16, cond = mask_3)[name = tensor("input_1021_cast_fp16")]; - tensor x_431_transpose_x_0 = const()[name = tensor("x_431_transpose_x_0"), val = tensor(false)]; - tensor x_431_transpose_y_0 = const()[name = tensor("x_431_transpose_y_0"), val = tensor(false)]; - tensor value_41_cast_fp16 = transpose(perm = value_41_perm_0, x = v_39_cast_fp16)[name = tensor("transpose_176")]; - tensor x_431_cast_fp16 = matmul(transpose_x = x_431_transpose_x_0, transpose_y = x_431_transpose_y_0, x = input_1021_cast_fp16, y = value_41_cast_fp16)[name = tensor("x_431_cast_fp16")]; - tensor var_3402_perm_0 = const()[name = tensor("op_3402_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_3403 = const()[name = tensor("op_3403"), val = tensor([1, -1, 1024])]; - tensor var_3402_cast_fp16 = transpose(perm = var_3402_perm_0, x = x_431_cast_fp16)[name = tensor("transpose_175")]; - tensor input_1023_cast_fp16 = reshape(shape = var_3403, x = var_3402_cast_fp16)[name = tensor("input_1023_cast_fp16")]; - tensor module_layers_19_self_attn_linear_out_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(362073280))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(362859776))), name = tensor("module_layers_19_self_attn_linear_out_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_178_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_19_self_attn_linear_out_weight_to_fp16_palettized, x = input_1023_cast_fp16)[name = tensor("linear_178_cast_fp16")]; - tensor input_1027_cast_fp16 = add(x = input_1019_cast_fp16, y = linear_178_cast_fp16)[name = tensor("input_1027_cast_fp16")]; - tensor x_435_axes_0 = const()[name = tensor("x_435_axes_0"), val = tensor([-1])]; - tensor module_layers_19_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_19_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(362859968)))]; - tensor module_layers_19_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_19_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(362862080)))]; - tensor x_435_cast_fp16 = layer_norm(axes = x_435_axes_0, beta = module_layers_19_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_19_norm_conv_weight_to_fp16, x = input_1027_cast_fp16)[name = tensor("x_435_cast_fp16")]; - tensor input_1029_perm_0 = const()[name = tensor("input_1029_perm_0"), val = tensor([0, 2, 1])]; - tensor input_1031_pad_type_0 = const()[name = tensor("input_1031_pad_type_0"), val = tensor("valid")]; - tensor input_1031_strides_0 = const()[name = tensor("input_1031_strides_0"), val = tensor([1])]; - tensor input_1031_pad_0 = const()[name = tensor("input_1031_pad_0"), val = tensor([0, 0])]; - tensor input_1031_dilations_0 = const()[name = tensor("input_1031_dilations_0"), val = tensor([1])]; - tensor input_1031_groups_0 = const()[name = tensor("input_1031_groups_0"), val = tensor(1)]; - tensor module_layers_19_conv_pointwise_conv1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(362864192))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(364437120))), name = tensor("module_layers_19_conv_pointwise_conv1_weight_to_fp16_palettized"), shape = tensor([2048, 1024, 1])]; - tensor input_1029_cast_fp16 = transpose(perm = input_1029_perm_0, x = x_435_cast_fp16)[name = tensor("transpose_174")]; - tensor input_1031_cast_fp16 = conv(dilations = input_1031_dilations_0, groups = input_1031_groups_0, pad = input_1031_pad_0, pad_type = input_1031_pad_type_0, strides = input_1031_strides_0, weight = module_layers_19_conv_pointwise_conv1_weight_to_fp16_palettized, x = input_1029_cast_fp16)[name = tensor("input_1031_cast_fp16")]; - tensor x_437_split_num_splits_0 = const()[name = tensor("x_437_split_num_splits_0"), val = tensor(2)]; - tensor x_437_split_axis_0 = const()[name = tensor("x_437_split_axis_0"), val = tensor(1)]; - tensor x_437_split_cast_fp16_0, tensor x_437_split_cast_fp16_1 = split(axis = x_437_split_axis_0, num_splits = x_437_split_num_splits_0, x = input_1031_cast_fp16)[name = tensor("x_437_split_cast_fp16")]; - tensor x_437_split_1_sigmoid_cast_fp16 = sigmoid(x = x_437_split_cast_fp16_1)[name = tensor("x_437_split_1_sigmoid_cast_fp16")]; - tensor x_437_cast_fp16 = mul(x = x_437_split_cast_fp16_0, y = x_437_split_1_sigmoid_cast_fp16)[name = tensor("x_437_cast_fp16")]; - tensor input_1033_cast_fp16 = select(a = var_11_to_fp16, b = x_437_cast_fp16, cond = var_328)[name = tensor("input_1033_cast_fp16")]; - tensor input_1035_pad_0 = const()[name = tensor("input_1035_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_1035_mode_0 = const()[name = tensor("input_1035_mode_0"), val = tensor("constant")]; - tensor const_207_to_fp16 = const()[name = tensor("const_207_to_fp16"), val = tensor(0x0p+0)]; - tensor input_1035_cast_fp16 = pad(constant_val = const_207_to_fp16, mode = input_1035_mode_0, pad = input_1035_pad_0, x = input_1033_cast_fp16)[name = tensor("input_1035_cast_fp16")]; - tensor input_1037_pad_type_0 = const()[name = tensor("input_1037_pad_type_0"), val = tensor("valid")]; - tensor input_1037_groups_0 = const()[name = tensor("input_1037_groups_0"), val = tensor(1024)]; - tensor input_1037_strides_0 = const()[name = tensor("input_1037_strides_0"), val = tensor([1])]; - tensor input_1037_pad_0 = const()[name = tensor("input_1037_pad_0"), val = tensor([0, 0])]; - tensor input_1037_dilations_0 = const()[name = tensor("input_1037_dilations_0"), val = tensor([1])]; - tensor const_286_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(364437312))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(364444288))), name = tensor("const_286_to_fp16_palettized"), shape = tensor([1024, 1, 9])]; - tensor const_287_to_fp16 = const()[name = tensor("const_287_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(364444480)))]; - tensor input_1039_cast_fp16 = conv(bias = const_287_to_fp16, dilations = input_1037_dilations_0, groups = input_1037_groups_0, pad = input_1037_pad_0, pad_type = input_1037_pad_type_0, strides = input_1037_strides_0, weight = const_286_to_fp16_palettized, x = input_1035_cast_fp16)[name = tensor("input_1039_cast_fp16")]; - tensor input_1041_cast_fp16 = silu(x = input_1039_cast_fp16)[name = tensor("input_1041_cast_fp16")]; - tensor x_439_pad_type_0 = const()[name = tensor("x_439_pad_type_0"), val = tensor("valid")]; - tensor x_439_strides_0 = const()[name = tensor("x_439_strides_0"), val = tensor([1])]; - tensor x_439_pad_0 = const()[name = tensor("x_439_pad_0"), val = tensor([0, 0])]; - tensor x_439_dilations_0 = const()[name = tensor("x_439_dilations_0"), val = tensor([1])]; - tensor x_439_groups_0 = const()[name = tensor("x_439_groups_0"), val = tensor(1)]; - tensor module_layers_19_conv_pointwise_conv2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(364446592))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(365233088))), name = tensor("module_layers_19_conv_pointwise_conv2_weight_to_fp16_palettized"), shape = tensor([1024, 1024, 1])]; - tensor x_439_cast_fp16 = conv(dilations = x_439_dilations_0, groups = x_439_groups_0, pad = x_439_pad_0, pad_type = x_439_pad_type_0, strides = x_439_strides_0, weight = module_layers_19_conv_pointwise_conv2_weight_to_fp16_palettized, x = input_1041_cast_fp16)[name = tensor("x_439_cast_fp16")]; - tensor input_1043_perm_0 = const()[name = tensor("input_1043_perm_0"), val = tensor([0, 2, 1])]; - tensor input_1043_cast_fp16 = transpose(perm = input_1043_perm_0, x = x_439_cast_fp16)[name = tensor("transpose_173")]; - tensor input_1045_cast_fp16 = add(x = input_1027_cast_fp16, y = input_1043_cast_fp16)[name = tensor("input_1045_cast_fp16")]; - tensor input_1047_axes_0 = const()[name = tensor("input_1047_axes_0"), val = tensor([-1])]; - tensor module_layers_19_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_19_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(365233280)))]; - tensor module_layers_19_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_19_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(365235392)))]; - tensor input_1047_cast_fp16 = layer_norm(axes = input_1047_axes_0, beta = module_layers_19_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_19_norm_feed_forward2_weight_to_fp16, x = input_1045_cast_fp16)[name = tensor("input_1047_cast_fp16")]; - tensor module_layers_19_feed_forward2_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(365237504))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(368383296))), name = tensor("module_layers_19_feed_forward2_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_179_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_19_feed_forward2_linear1_weight_to_fp16_palettized, x = input_1047_cast_fp16)[name = tensor("linear_179_cast_fp16")]; - tensor input_1051_cast_fp16 = silu(x = linear_179_cast_fp16)[name = tensor("input_1051_cast_fp16")]; - tensor module_layers_19_feed_forward2_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(368383488))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(371529280))), name = tensor("module_layers_19_feed_forward2_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_180_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_19_feed_forward2_linear2_weight_to_fp16_palettized, x = input_1051_cast_fp16)[name = tensor("linear_180_cast_fp16")]; - tensor var_3463_to_fp16 = const()[name = tensor("op_3463_to_fp16"), val = tensor(0x1p-1)]; - tensor var_3464_cast_fp16 = mul(x = linear_180_cast_fp16, y = var_3463_to_fp16)[name = tensor("op_3464_cast_fp16")]; - tensor input_1057_cast_fp16 = add(x = input_1045_cast_fp16, y = var_3464_cast_fp16)[name = tensor("input_1057_cast_fp16")]; - tensor input_1059_axes_0 = const()[name = tensor("input_1059_axes_0"), val = tensor([-1])]; - tensor module_layers_19_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_19_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(371529472)))]; - tensor module_layers_19_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_19_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(371531584)))]; - tensor input_1059_cast_fp16 = layer_norm(axes = input_1059_axes_0, beta = module_layers_19_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_19_norm_out_weight_to_fp16, x = input_1057_cast_fp16)[name = tensor("input_1059_cast_fp16")]; - tensor input_1061_axes_0 = const()[name = tensor("input_1061_axes_0"), val = tensor([-1])]; - tensor module_layers_20_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_20_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(371533696)))]; - tensor module_layers_20_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_20_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(371535808)))]; - tensor input_1061_cast_fp16 = layer_norm(axes = input_1061_axes_0, beta = module_layers_20_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_20_norm_feed_forward1_weight_to_fp16, x = input_1059_cast_fp16)[name = tensor("input_1061_cast_fp16")]; - tensor module_layers_20_feed_forward1_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(371537920))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(374683712))), name = tensor("module_layers_20_feed_forward1_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_181_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_20_feed_forward1_linear1_weight_to_fp16_palettized, x = input_1061_cast_fp16)[name = tensor("linear_181_cast_fp16")]; - tensor input_1065_cast_fp16 = silu(x = linear_181_cast_fp16)[name = tensor("input_1065_cast_fp16")]; - tensor module_layers_20_feed_forward1_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(374683904))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(377829696))), name = tensor("module_layers_20_feed_forward1_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_182_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_20_feed_forward1_linear2_weight_to_fp16_palettized, x = input_1065_cast_fp16)[name = tensor("linear_182_cast_fp16")]; - tensor var_3492_to_fp16 = const()[name = tensor("op_3492_to_fp16"), val = tensor(0x1p-1)]; - tensor var_3493_cast_fp16 = mul(x = linear_182_cast_fp16, y = var_3492_to_fp16)[name = tensor("op_3493_cast_fp16")]; - tensor input_1071_cast_fp16 = add(x = input_1059_cast_fp16, y = var_3493_cast_fp16)[name = tensor("input_1071_cast_fp16")]; - tensor query_41_axes_0 = const()[name = tensor("query_41_axes_0"), val = tensor([-1])]; - tensor module_layers_20_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_20_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(377829888)))]; - tensor module_layers_20_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_20_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(377832000)))]; - tensor query_41_cast_fp16 = layer_norm(axes = query_41_axes_0, beta = module_layers_20_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_20_norm_self_att_weight_to_fp16, x = input_1071_cast_fp16)[name = tensor("query_41_cast_fp16")]; - tensor module_layers_20_self_attn_linear_q_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(377834112))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(378620608))), name = tensor("module_layers_20_self_attn_linear_q_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_183_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_20_self_attn_linear_q_weight_to_fp16_palettized, x = query_41_cast_fp16)[name = tensor("linear_183_cast_fp16")]; - tensor var_3509 = const()[name = tensor("op_3509"), val = tensor([1, -1, 8, 128])]; - tensor q_121_cast_fp16 = reshape(shape = var_3509, x = linear_183_cast_fp16)[name = tensor("q_121_cast_fp16")]; - tensor module_layers_20_self_attn_linear_k_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(378620800))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(379407296))), name = tensor("module_layers_20_self_attn_linear_k_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_184_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_20_self_attn_linear_k_weight_to_fp16_palettized, x = query_41_cast_fp16)[name = tensor("linear_184_cast_fp16")]; - tensor var_3513 = const()[name = tensor("op_3513"), val = tensor([1, -1, 8, 128])]; - tensor k_81_cast_fp16 = reshape(shape = var_3513, x = linear_184_cast_fp16)[name = tensor("k_81_cast_fp16")]; - tensor module_layers_20_self_attn_linear_v_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(379407488))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(380193984))), name = tensor("module_layers_20_self_attn_linear_v_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_185_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_20_self_attn_linear_v_weight_to_fp16_palettized, x = query_41_cast_fp16)[name = tensor("linear_185_cast_fp16")]; - tensor var_3517 = const()[name = tensor("op_3517"), val = tensor([1, -1, 8, 128])]; - tensor v_41_cast_fp16 = reshape(shape = var_3517, x = linear_185_cast_fp16)[name = tensor("v_41_cast_fp16")]; - tensor value_43_perm_0 = const()[name = tensor("value_43_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor module_layers_20_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_20_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(380194176)))]; - tensor var_3529_cast_fp16 = add(x = q_121_cast_fp16, y = module_layers_20_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3529_cast_fp16")]; - tensor module_layers_20_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_20_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(380196288)))]; - tensor var_3531_cast_fp16 = add(x = q_121_cast_fp16, y = module_layers_20_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3531_cast_fp16")]; - tensor q_with_bias_v_41_perm_0 = const()[name = tensor("q_with_bias_v_41_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor x_447_transpose_x_0 = const()[name = tensor("x_447_transpose_x_0"), val = tensor(false)]; - tensor x_447_transpose_y_0 = const()[name = tensor("x_447_transpose_y_0"), val = tensor(false)]; - tensor op_3533_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(380198400))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(380486464))), name = tensor("op_3533_to_fp16_palettized"), shape = tensor([1, 8, 128, 375])]; - tensor q_with_bias_v_41_cast_fp16 = transpose(perm = q_with_bias_v_41_perm_0, x = var_3531_cast_fp16)[name = tensor("transpose_172")]; - tensor x_447_cast_fp16 = matmul(transpose_x = x_447_transpose_x_0, transpose_y = x_447_transpose_y_0, x = q_with_bias_v_41_cast_fp16, y = op_3533_to_fp16_palettized)[name = tensor("x_447_cast_fp16")]; - tensor x_449_pad_0 = const()[name = tensor("x_449_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_449_mode_0 = const()[name = tensor("x_449_mode_0"), val = tensor("constant")]; - tensor const_214_to_fp16 = const()[name = tensor("const_214_to_fp16"), val = tensor(0x0p+0)]; - tensor x_449_cast_fp16 = pad(constant_val = const_214_to_fp16, mode = x_449_mode_0, pad = x_449_pad_0, x = x_447_cast_fp16)[name = tensor("x_449_cast_fp16")]; - tensor var_3541 = const()[name = tensor("op_3541"), val = tensor([1, 8, -1, 188])]; - tensor x_451_cast_fp16 = reshape(shape = var_3541, x = x_449_cast_fp16)[name = tensor("x_451_cast_fp16")]; - tensor var_3545_begin_0 = const()[name = tensor("op_3545_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_3545_end_0 = const()[name = tensor("op_3545_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_3545_end_mask_0 = const()[name = tensor("op_3545_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_3545_cast_fp16 = slice_by_index(begin = var_3545_begin_0, end = var_3545_end_0, end_mask = var_3545_end_mask_0, x = x_451_cast_fp16)[name = tensor("op_3545_cast_fp16")]; - tensor var_3546 = const()[name = tensor("op_3546"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_81_cast_fp16 = reshape(shape = var_3546, x = var_3545_cast_fp16)[name = tensor("matrix_bd_81_cast_fp16")]; - tensor matrix_ac_41_transpose_x_0 = const()[name = tensor("matrix_ac_41_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_41_transpose_y_0 = const()[name = tensor("matrix_ac_41_transpose_y_0"), val = tensor(false)]; - tensor transpose_136_perm_0 = const()[name = tensor("transpose_136_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_137_perm_0 = const()[name = tensor("transpose_137_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_137 = transpose(perm = transpose_137_perm_0, x = k_81_cast_fp16)[name = tensor("transpose_170")]; - tensor transpose_136 = transpose(perm = transpose_136_perm_0, x = var_3529_cast_fp16)[name = tensor("transpose_171")]; - tensor matrix_ac_41_cast_fp16 = matmul(transpose_x = matrix_ac_41_transpose_x_0, transpose_y = matrix_ac_41_transpose_y_0, x = transpose_136, y = transpose_137)[name = tensor("matrix_ac_41_cast_fp16")]; - tensor matrix_bd_83_begin_0 = const()[name = tensor("matrix_bd_83_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_83_end_0 = const()[name = tensor("matrix_bd_83_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_83_end_mask_0 = const()[name = tensor("matrix_bd_83_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_83_cast_fp16 = slice_by_index(begin = matrix_bd_83_begin_0, end = matrix_bd_83_end_0, end_mask = matrix_bd_83_end_mask_0, x = matrix_bd_81_cast_fp16)[name = tensor("matrix_bd_83_cast_fp16")]; - tensor var_3555_cast_fp16 = add(x = matrix_ac_41_cast_fp16, y = matrix_bd_83_cast_fp16)[name = tensor("op_3555_cast_fp16")]; - tensor _inversed_scores_81_y_0_to_fp16 = const()[name = tensor("_inversed_scores_81_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; - tensor _inversed_scores_81_cast_fp16 = mul(x = var_3555_cast_fp16, y = _inversed_scores_81_y_0_to_fp16)[name = tensor("_inversed_scores_81_cast_fp16")]; - tensor scores_83_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_81_cast_fp16, cond = mask_3)[name = tensor("scores_83_cast_fp16")]; - tensor var_3561_cast_fp16 = softmax(axis = var_30, x = scores_83_cast_fp16)[name = tensor("op_3561_cast_fp16")]; - tensor input_1073_cast_fp16 = select(a = var_11_to_fp16, b = var_3561_cast_fp16, cond = mask_3)[name = tensor("input_1073_cast_fp16")]; - tensor x_453_transpose_x_0 = const()[name = tensor("x_453_transpose_x_0"), val = tensor(false)]; - tensor x_453_transpose_y_0 = const()[name = tensor("x_453_transpose_y_0"), val = tensor(false)]; - tensor value_43_cast_fp16 = transpose(perm = value_43_perm_0, x = v_41_cast_fp16)[name = tensor("transpose_169")]; - tensor x_453_cast_fp16 = matmul(transpose_x = x_453_transpose_x_0, transpose_y = x_453_transpose_y_0, x = input_1073_cast_fp16, y = value_43_cast_fp16)[name = tensor("x_453_cast_fp16")]; - tensor var_3565_perm_0 = const()[name = tensor("op_3565_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_3566 = const()[name = tensor("op_3566"), val = tensor([1, -1, 1024])]; - tensor var_3565_cast_fp16 = transpose(perm = var_3565_perm_0, x = x_453_cast_fp16)[name = tensor("transpose_168")]; - tensor input_1075_cast_fp16 = reshape(shape = var_3566, x = var_3565_cast_fp16)[name = tensor("input_1075_cast_fp16")]; - tensor module_layers_20_self_attn_linear_out_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(380486656))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(381273152))), name = tensor("module_layers_20_self_attn_linear_out_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_187_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_20_self_attn_linear_out_weight_to_fp16_palettized, x = input_1075_cast_fp16)[name = tensor("linear_187_cast_fp16")]; - tensor input_1079_cast_fp16 = add(x = input_1071_cast_fp16, y = linear_187_cast_fp16)[name = tensor("input_1079_cast_fp16")]; - tensor x_457_axes_0 = const()[name = tensor("x_457_axes_0"), val = tensor([-1])]; - tensor module_layers_20_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_20_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(381273344)))]; - tensor module_layers_20_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_20_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(381275456)))]; - tensor x_457_cast_fp16 = layer_norm(axes = x_457_axes_0, beta = module_layers_20_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_20_norm_conv_weight_to_fp16, x = input_1079_cast_fp16)[name = tensor("x_457_cast_fp16")]; - tensor input_1081_perm_0 = const()[name = tensor("input_1081_perm_0"), val = tensor([0, 2, 1])]; - tensor input_1083_pad_type_0 = const()[name = tensor("input_1083_pad_type_0"), val = tensor("valid")]; - tensor input_1083_strides_0 = const()[name = tensor("input_1083_strides_0"), val = tensor([1])]; - tensor input_1083_pad_0 = const()[name = tensor("input_1083_pad_0"), val = tensor([0, 0])]; - tensor input_1083_dilations_0 = const()[name = tensor("input_1083_dilations_0"), val = tensor([1])]; - tensor input_1083_groups_0 = const()[name = tensor("input_1083_groups_0"), val = tensor(1)]; - tensor module_layers_20_conv_pointwise_conv1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(381277568))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(382850496))), name = tensor("module_layers_20_conv_pointwise_conv1_weight_to_fp16_palettized"), shape = tensor([2048, 1024, 1])]; - tensor input_1081_cast_fp16 = transpose(perm = input_1081_perm_0, x = x_457_cast_fp16)[name = tensor("transpose_167")]; - tensor input_1083_cast_fp16 = conv(dilations = input_1083_dilations_0, groups = input_1083_groups_0, pad = input_1083_pad_0, pad_type = input_1083_pad_type_0, strides = input_1083_strides_0, weight = module_layers_20_conv_pointwise_conv1_weight_to_fp16_palettized, x = input_1081_cast_fp16)[name = tensor("input_1083_cast_fp16")]; - tensor x_459_split_num_splits_0 = const()[name = tensor("x_459_split_num_splits_0"), val = tensor(2)]; - tensor x_459_split_axis_0 = const()[name = tensor("x_459_split_axis_0"), val = tensor(1)]; - tensor x_459_split_cast_fp16_0, tensor x_459_split_cast_fp16_1 = split(axis = x_459_split_axis_0, num_splits = x_459_split_num_splits_0, x = input_1083_cast_fp16)[name = tensor("x_459_split_cast_fp16")]; - tensor x_459_split_1_sigmoid_cast_fp16 = sigmoid(x = x_459_split_cast_fp16_1)[name = tensor("x_459_split_1_sigmoid_cast_fp16")]; - tensor x_459_cast_fp16 = mul(x = x_459_split_cast_fp16_0, y = x_459_split_1_sigmoid_cast_fp16)[name = tensor("x_459_cast_fp16")]; - tensor input_1085_cast_fp16 = select(a = var_11_to_fp16, b = x_459_cast_fp16, cond = var_328)[name = tensor("input_1085_cast_fp16")]; - tensor input_1087_pad_0 = const()[name = tensor("input_1087_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_1087_mode_0 = const()[name = tensor("input_1087_mode_0"), val = tensor("constant")]; - tensor const_217_to_fp16 = const()[name = tensor("const_217_to_fp16"), val = tensor(0x0p+0)]; - tensor input_1087_cast_fp16 = pad(constant_val = const_217_to_fp16, mode = input_1087_mode_0, pad = input_1087_pad_0, x = input_1085_cast_fp16)[name = tensor("input_1087_cast_fp16")]; - tensor input_1089_pad_type_0 = const()[name = tensor("input_1089_pad_type_0"), val = tensor("valid")]; - tensor input_1089_groups_0 = const()[name = tensor("input_1089_groups_0"), val = tensor(1024)]; - tensor input_1089_strides_0 = const()[name = tensor("input_1089_strides_0"), val = tensor([1])]; - tensor input_1089_pad_0 = const()[name = tensor("input_1089_pad_0"), val = tensor([0, 0])]; - tensor input_1089_dilations_0 = const()[name = tensor("input_1089_dilations_0"), val = tensor([1])]; - tensor const_288_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(382850688))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(382857664))), name = tensor("const_288_to_fp16_palettized"), shape = tensor([1024, 1, 9])]; - tensor const_289_to_fp16 = const()[name = tensor("const_289_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(382857856)))]; - tensor input_1091_cast_fp16 = conv(bias = const_289_to_fp16, dilations = input_1089_dilations_0, groups = input_1089_groups_0, pad = input_1089_pad_0, pad_type = input_1089_pad_type_0, strides = input_1089_strides_0, weight = const_288_to_fp16_palettized, x = input_1087_cast_fp16)[name = tensor("input_1091_cast_fp16")]; - tensor input_1093_cast_fp16 = silu(x = input_1091_cast_fp16)[name = tensor("input_1093_cast_fp16")]; - tensor x_461_pad_type_0 = const()[name = tensor("x_461_pad_type_0"), val = tensor("valid")]; - tensor x_461_strides_0 = const()[name = tensor("x_461_strides_0"), val = tensor([1])]; - tensor x_461_pad_0 = const()[name = tensor("x_461_pad_0"), val = tensor([0, 0])]; - tensor x_461_dilations_0 = const()[name = tensor("x_461_dilations_0"), val = tensor([1])]; - tensor x_461_groups_0 = const()[name = tensor("x_461_groups_0"), val = tensor(1)]; - tensor module_layers_20_conv_pointwise_conv2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(382859968))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(383646464))), name = tensor("module_layers_20_conv_pointwise_conv2_weight_to_fp16_palettized"), shape = tensor([1024, 1024, 1])]; - tensor x_461_cast_fp16 = conv(dilations = x_461_dilations_0, groups = x_461_groups_0, pad = x_461_pad_0, pad_type = x_461_pad_type_0, strides = x_461_strides_0, weight = module_layers_20_conv_pointwise_conv2_weight_to_fp16_palettized, x = input_1093_cast_fp16)[name = tensor("x_461_cast_fp16")]; - tensor input_1095_perm_0 = const()[name = tensor("input_1095_perm_0"), val = tensor([0, 2, 1])]; - tensor input_1095_cast_fp16 = transpose(perm = input_1095_perm_0, x = x_461_cast_fp16)[name = tensor("transpose_166")]; - tensor input_1097_cast_fp16 = add(x = input_1079_cast_fp16, y = input_1095_cast_fp16)[name = tensor("input_1097_cast_fp16")]; - tensor input_1099_axes_0 = const()[name = tensor("input_1099_axes_0"), val = tensor([-1])]; - tensor module_layers_20_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_20_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(383646656)))]; - tensor module_layers_20_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_20_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(383648768)))]; - tensor input_1099_cast_fp16 = layer_norm(axes = input_1099_axes_0, beta = module_layers_20_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_20_norm_feed_forward2_weight_to_fp16, x = input_1097_cast_fp16)[name = tensor("input_1099_cast_fp16")]; - tensor module_layers_20_feed_forward2_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(383650880))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(386796672))), name = tensor("module_layers_20_feed_forward2_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_188_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_20_feed_forward2_linear1_weight_to_fp16_palettized, x = input_1099_cast_fp16)[name = tensor("linear_188_cast_fp16")]; - tensor input_1103_cast_fp16 = silu(x = linear_188_cast_fp16)[name = tensor("input_1103_cast_fp16")]; - tensor module_layers_20_feed_forward2_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(386796864))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(389942656))), name = tensor("module_layers_20_feed_forward2_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_189_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_20_feed_forward2_linear2_weight_to_fp16_palettized, x = input_1103_cast_fp16)[name = tensor("linear_189_cast_fp16")]; - tensor var_3626_to_fp16 = const()[name = tensor("op_3626_to_fp16"), val = tensor(0x1p-1)]; - tensor var_3627_cast_fp16 = mul(x = linear_189_cast_fp16, y = var_3626_to_fp16)[name = tensor("op_3627_cast_fp16")]; - tensor input_1109_cast_fp16 = add(x = input_1097_cast_fp16, y = var_3627_cast_fp16)[name = tensor("input_1109_cast_fp16")]; - tensor input_1111_axes_0 = const()[name = tensor("input_1111_axes_0"), val = tensor([-1])]; - tensor module_layers_20_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_20_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(389942848)))]; - tensor module_layers_20_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_20_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(389944960)))]; - tensor input_1111_cast_fp16 = layer_norm(axes = input_1111_axes_0, beta = module_layers_20_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_20_norm_out_weight_to_fp16, x = input_1109_cast_fp16)[name = tensor("input_1111_cast_fp16")]; - tensor input_1113_axes_0 = const()[name = tensor("input_1113_axes_0"), val = tensor([-1])]; - tensor module_layers_21_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_21_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(389947072)))]; - tensor module_layers_21_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_21_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(389949184)))]; - tensor input_1113_cast_fp16 = layer_norm(axes = input_1113_axes_0, beta = module_layers_21_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_21_norm_feed_forward1_weight_to_fp16, x = input_1111_cast_fp16)[name = tensor("input_1113_cast_fp16")]; - tensor module_layers_21_feed_forward1_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(389951296))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(393097088))), name = tensor("module_layers_21_feed_forward1_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_190_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_21_feed_forward1_linear1_weight_to_fp16_palettized, x = input_1113_cast_fp16)[name = tensor("linear_190_cast_fp16")]; - tensor input_1117_cast_fp16 = silu(x = linear_190_cast_fp16)[name = tensor("input_1117_cast_fp16")]; - tensor module_layers_21_feed_forward1_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(393097280))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(396243072))), name = tensor("module_layers_21_feed_forward1_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_191_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_21_feed_forward1_linear2_weight_to_fp16_palettized, x = input_1117_cast_fp16)[name = tensor("linear_191_cast_fp16")]; - tensor var_3655_to_fp16 = const()[name = tensor("op_3655_to_fp16"), val = tensor(0x1p-1)]; - tensor var_3656_cast_fp16 = mul(x = linear_191_cast_fp16, y = var_3655_to_fp16)[name = tensor("op_3656_cast_fp16")]; - tensor input_1123_cast_fp16 = add(x = input_1111_cast_fp16, y = var_3656_cast_fp16)[name = tensor("input_1123_cast_fp16")]; - tensor query_43_axes_0 = const()[name = tensor("query_43_axes_0"), val = tensor([-1])]; - tensor module_layers_21_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_21_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(396243264)))]; - tensor module_layers_21_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_21_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(396245376)))]; - tensor query_43_cast_fp16 = layer_norm(axes = query_43_axes_0, beta = module_layers_21_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_21_norm_self_att_weight_to_fp16, x = input_1123_cast_fp16)[name = tensor("query_43_cast_fp16")]; - tensor module_layers_21_self_attn_linear_q_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(396247488))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(397033984))), name = tensor("module_layers_21_self_attn_linear_q_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_192_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_21_self_attn_linear_q_weight_to_fp16_palettized, x = query_43_cast_fp16)[name = tensor("linear_192_cast_fp16")]; - tensor var_3672 = const()[name = tensor("op_3672"), val = tensor([1, -1, 8, 128])]; - tensor q_127_cast_fp16 = reshape(shape = var_3672, x = linear_192_cast_fp16)[name = tensor("q_127_cast_fp16")]; - tensor module_layers_21_self_attn_linear_k_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(397034176))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(397820672))), name = tensor("module_layers_21_self_attn_linear_k_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_193_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_21_self_attn_linear_k_weight_to_fp16_palettized, x = query_43_cast_fp16)[name = tensor("linear_193_cast_fp16")]; - tensor var_3676 = const()[name = tensor("op_3676"), val = tensor([1, -1, 8, 128])]; - tensor k_85_cast_fp16 = reshape(shape = var_3676, x = linear_193_cast_fp16)[name = tensor("k_85_cast_fp16")]; - tensor module_layers_21_self_attn_linear_v_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(397820864))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(398607360))), name = tensor("module_layers_21_self_attn_linear_v_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_194_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_21_self_attn_linear_v_weight_to_fp16_palettized, x = query_43_cast_fp16)[name = tensor("linear_194_cast_fp16")]; - tensor var_3680 = const()[name = tensor("op_3680"), val = tensor([1, -1, 8, 128])]; - tensor v_43_cast_fp16 = reshape(shape = var_3680, x = linear_194_cast_fp16)[name = tensor("v_43_cast_fp16")]; - tensor value_45_perm_0 = const()[name = tensor("value_45_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor module_layers_21_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_21_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(398607552)))]; - tensor var_3692_cast_fp16 = add(x = q_127_cast_fp16, y = module_layers_21_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3692_cast_fp16")]; - tensor module_layers_21_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_21_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(398609664)))]; - tensor var_3694_cast_fp16 = add(x = q_127_cast_fp16, y = module_layers_21_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3694_cast_fp16")]; - tensor q_with_bias_v_43_perm_0 = const()[name = tensor("q_with_bias_v_43_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor x_469_transpose_x_0 = const()[name = tensor("x_469_transpose_x_0"), val = tensor(false)]; - tensor x_469_transpose_y_0 = const()[name = tensor("x_469_transpose_y_0"), val = tensor(false)]; - tensor op_3696_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(398611776))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(398899840))), name = tensor("op_3696_to_fp16_palettized"), shape = tensor([1, 8, 128, 375])]; - tensor q_with_bias_v_43_cast_fp16 = transpose(perm = q_with_bias_v_43_perm_0, x = var_3694_cast_fp16)[name = tensor("transpose_165")]; - tensor x_469_cast_fp16 = matmul(transpose_x = x_469_transpose_x_0, transpose_y = x_469_transpose_y_0, x = q_with_bias_v_43_cast_fp16, y = op_3696_to_fp16_palettized)[name = tensor("x_469_cast_fp16")]; - tensor x_471_pad_0 = const()[name = tensor("x_471_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_471_mode_0 = const()[name = tensor("x_471_mode_0"), val = tensor("constant")]; - tensor const_224_to_fp16 = const()[name = tensor("const_224_to_fp16"), val = tensor(0x0p+0)]; - tensor x_471_cast_fp16 = pad(constant_val = const_224_to_fp16, mode = x_471_mode_0, pad = x_471_pad_0, x = x_469_cast_fp16)[name = tensor("x_471_cast_fp16")]; - tensor var_3704 = const()[name = tensor("op_3704"), val = tensor([1, 8, -1, 188])]; - tensor x_473_cast_fp16 = reshape(shape = var_3704, x = x_471_cast_fp16)[name = tensor("x_473_cast_fp16")]; - tensor var_3708_begin_0 = const()[name = tensor("op_3708_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_3708_end_0 = const()[name = tensor("op_3708_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_3708_end_mask_0 = const()[name = tensor("op_3708_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_3708_cast_fp16 = slice_by_index(begin = var_3708_begin_0, end = var_3708_end_0, end_mask = var_3708_end_mask_0, x = x_473_cast_fp16)[name = tensor("op_3708_cast_fp16")]; - tensor var_3709 = const()[name = tensor("op_3709"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_85_cast_fp16 = reshape(shape = var_3709, x = var_3708_cast_fp16)[name = tensor("matrix_bd_85_cast_fp16")]; - tensor matrix_ac_43_transpose_x_0 = const()[name = tensor("matrix_ac_43_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_43_transpose_y_0 = const()[name = tensor("matrix_ac_43_transpose_y_0"), val = tensor(false)]; - tensor transpose_138_perm_0 = const()[name = tensor("transpose_138_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_139_perm_0 = const()[name = tensor("transpose_139_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_139 = transpose(perm = transpose_139_perm_0, x = k_85_cast_fp16)[name = tensor("transpose_163")]; - tensor transpose_138 = transpose(perm = transpose_138_perm_0, x = var_3692_cast_fp16)[name = tensor("transpose_164")]; - tensor matrix_ac_43_cast_fp16 = matmul(transpose_x = matrix_ac_43_transpose_x_0, transpose_y = matrix_ac_43_transpose_y_0, x = transpose_138, y = transpose_139)[name = tensor("matrix_ac_43_cast_fp16")]; - tensor matrix_bd_87_begin_0 = const()[name = tensor("matrix_bd_87_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_87_end_0 = const()[name = tensor("matrix_bd_87_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_87_end_mask_0 = const()[name = tensor("matrix_bd_87_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_87_cast_fp16 = slice_by_index(begin = matrix_bd_87_begin_0, end = matrix_bd_87_end_0, end_mask = matrix_bd_87_end_mask_0, x = matrix_bd_85_cast_fp16)[name = tensor("matrix_bd_87_cast_fp16")]; - tensor var_3718_cast_fp16 = add(x = matrix_ac_43_cast_fp16, y = matrix_bd_87_cast_fp16)[name = tensor("op_3718_cast_fp16")]; - tensor _inversed_scores_85_y_0_to_fp16 = const()[name = tensor("_inversed_scores_85_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; - tensor _inversed_scores_85_cast_fp16 = mul(x = var_3718_cast_fp16, y = _inversed_scores_85_y_0_to_fp16)[name = tensor("_inversed_scores_85_cast_fp16")]; - tensor scores_87_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_85_cast_fp16, cond = mask_3)[name = tensor("scores_87_cast_fp16")]; - tensor var_3724_cast_fp16 = softmax(axis = var_30, x = scores_87_cast_fp16)[name = tensor("op_3724_cast_fp16")]; - tensor input_1125_cast_fp16 = select(a = var_11_to_fp16, b = var_3724_cast_fp16, cond = mask_3)[name = tensor("input_1125_cast_fp16")]; - tensor x_475_transpose_x_0 = const()[name = tensor("x_475_transpose_x_0"), val = tensor(false)]; - tensor x_475_transpose_y_0 = const()[name = tensor("x_475_transpose_y_0"), val = tensor(false)]; - tensor value_45_cast_fp16 = transpose(perm = value_45_perm_0, x = v_43_cast_fp16)[name = tensor("transpose_162")]; - tensor x_475_cast_fp16 = matmul(transpose_x = x_475_transpose_x_0, transpose_y = x_475_transpose_y_0, x = input_1125_cast_fp16, y = value_45_cast_fp16)[name = tensor("x_475_cast_fp16")]; - tensor var_3728_perm_0 = const()[name = tensor("op_3728_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_3729 = const()[name = tensor("op_3729"), val = tensor([1, -1, 1024])]; - tensor var_3728_cast_fp16 = transpose(perm = var_3728_perm_0, x = x_475_cast_fp16)[name = tensor("transpose_161")]; - tensor input_1127_cast_fp16 = reshape(shape = var_3729, x = var_3728_cast_fp16)[name = tensor("input_1127_cast_fp16")]; - tensor module_layers_21_self_attn_linear_out_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(398900032))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(399686528))), name = tensor("module_layers_21_self_attn_linear_out_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_196_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_21_self_attn_linear_out_weight_to_fp16_palettized, x = input_1127_cast_fp16)[name = tensor("linear_196_cast_fp16")]; - tensor input_1131_cast_fp16 = add(x = input_1123_cast_fp16, y = linear_196_cast_fp16)[name = tensor("input_1131_cast_fp16")]; - tensor x_479_axes_0 = const()[name = tensor("x_479_axes_0"), val = tensor([-1])]; - tensor module_layers_21_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_21_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(399686720)))]; - tensor module_layers_21_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_21_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(399688832)))]; - tensor x_479_cast_fp16 = layer_norm(axes = x_479_axes_0, beta = module_layers_21_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_21_norm_conv_weight_to_fp16, x = input_1131_cast_fp16)[name = tensor("x_479_cast_fp16")]; - tensor input_1133_perm_0 = const()[name = tensor("input_1133_perm_0"), val = tensor([0, 2, 1])]; - tensor input_1135_pad_type_0 = const()[name = tensor("input_1135_pad_type_0"), val = tensor("valid")]; - tensor input_1135_strides_0 = const()[name = tensor("input_1135_strides_0"), val = tensor([1])]; - tensor input_1135_pad_0 = const()[name = tensor("input_1135_pad_0"), val = tensor([0, 0])]; - tensor input_1135_dilations_0 = const()[name = tensor("input_1135_dilations_0"), val = tensor([1])]; - tensor input_1135_groups_0 = const()[name = tensor("input_1135_groups_0"), val = tensor(1)]; - tensor module_layers_21_conv_pointwise_conv1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(399690944))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(401263872))), name = tensor("module_layers_21_conv_pointwise_conv1_weight_to_fp16_palettized"), shape = tensor([2048, 1024, 1])]; - tensor input_1133_cast_fp16 = transpose(perm = input_1133_perm_0, x = x_479_cast_fp16)[name = tensor("transpose_160")]; - tensor input_1135_cast_fp16 = conv(dilations = input_1135_dilations_0, groups = input_1135_groups_0, pad = input_1135_pad_0, pad_type = input_1135_pad_type_0, strides = input_1135_strides_0, weight = module_layers_21_conv_pointwise_conv1_weight_to_fp16_palettized, x = input_1133_cast_fp16)[name = tensor("input_1135_cast_fp16")]; - tensor x_481_split_num_splits_0 = const()[name = tensor("x_481_split_num_splits_0"), val = tensor(2)]; - tensor x_481_split_axis_0 = const()[name = tensor("x_481_split_axis_0"), val = tensor(1)]; - tensor x_481_split_cast_fp16_0, tensor x_481_split_cast_fp16_1 = split(axis = x_481_split_axis_0, num_splits = x_481_split_num_splits_0, x = input_1135_cast_fp16)[name = tensor("x_481_split_cast_fp16")]; - tensor x_481_split_1_sigmoid_cast_fp16 = sigmoid(x = x_481_split_cast_fp16_1)[name = tensor("x_481_split_1_sigmoid_cast_fp16")]; - tensor x_481_cast_fp16 = mul(x = x_481_split_cast_fp16_0, y = x_481_split_1_sigmoid_cast_fp16)[name = tensor("x_481_cast_fp16")]; - tensor input_1137_cast_fp16 = select(a = var_11_to_fp16, b = x_481_cast_fp16, cond = var_328)[name = tensor("input_1137_cast_fp16")]; - tensor input_1139_pad_0 = const()[name = tensor("input_1139_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_1139_mode_0 = const()[name = tensor("input_1139_mode_0"), val = tensor("constant")]; - tensor const_227_to_fp16 = const()[name = tensor("const_227_to_fp16"), val = tensor(0x0p+0)]; - tensor input_1139_cast_fp16 = pad(constant_val = const_227_to_fp16, mode = input_1139_mode_0, pad = input_1139_pad_0, x = input_1137_cast_fp16)[name = tensor("input_1139_cast_fp16")]; - tensor input_1141_pad_type_0 = const()[name = tensor("input_1141_pad_type_0"), val = tensor("valid")]; - tensor input_1141_groups_0 = const()[name = tensor("input_1141_groups_0"), val = tensor(1024)]; - tensor input_1141_strides_0 = const()[name = tensor("input_1141_strides_0"), val = tensor([1])]; - tensor input_1141_pad_0 = const()[name = tensor("input_1141_pad_0"), val = tensor([0, 0])]; - tensor input_1141_dilations_0 = const()[name = tensor("input_1141_dilations_0"), val = tensor([1])]; - tensor const_290_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(401264064))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(401271040))), name = tensor("const_290_to_fp16_palettized"), shape = tensor([1024, 1, 9])]; - tensor const_291_to_fp16 = const()[name = tensor("const_291_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(401271232)))]; - tensor input_1143_cast_fp16 = conv(bias = const_291_to_fp16, dilations = input_1141_dilations_0, groups = input_1141_groups_0, pad = input_1141_pad_0, pad_type = input_1141_pad_type_0, strides = input_1141_strides_0, weight = const_290_to_fp16_palettized, x = input_1139_cast_fp16)[name = tensor("input_1143_cast_fp16")]; - tensor input_1145_cast_fp16 = silu(x = input_1143_cast_fp16)[name = tensor("input_1145_cast_fp16")]; - tensor x_483_pad_type_0 = const()[name = tensor("x_483_pad_type_0"), val = tensor("valid")]; - tensor x_483_strides_0 = const()[name = tensor("x_483_strides_0"), val = tensor([1])]; - tensor x_483_pad_0 = const()[name = tensor("x_483_pad_0"), val = tensor([0, 0])]; - tensor x_483_dilations_0 = const()[name = tensor("x_483_dilations_0"), val = tensor([1])]; - tensor x_483_groups_0 = const()[name = tensor("x_483_groups_0"), val = tensor(1)]; - tensor module_layers_21_conv_pointwise_conv2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(401273344))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(402059840))), name = tensor("module_layers_21_conv_pointwise_conv2_weight_to_fp16_palettized"), shape = tensor([1024, 1024, 1])]; - tensor x_483_cast_fp16 = conv(dilations = x_483_dilations_0, groups = x_483_groups_0, pad = x_483_pad_0, pad_type = x_483_pad_type_0, strides = x_483_strides_0, weight = module_layers_21_conv_pointwise_conv2_weight_to_fp16_palettized, x = input_1145_cast_fp16)[name = tensor("x_483_cast_fp16")]; - tensor input_1147_perm_0 = const()[name = tensor("input_1147_perm_0"), val = tensor([0, 2, 1])]; - tensor input_1147_cast_fp16 = transpose(perm = input_1147_perm_0, x = x_483_cast_fp16)[name = tensor("transpose_159")]; - tensor input_1149_cast_fp16 = add(x = input_1131_cast_fp16, y = input_1147_cast_fp16)[name = tensor("input_1149_cast_fp16")]; - tensor input_1151_axes_0 = const()[name = tensor("input_1151_axes_0"), val = tensor([-1])]; - tensor module_layers_21_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_21_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(402060032)))]; - tensor module_layers_21_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_21_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(402062144)))]; - tensor input_1151_cast_fp16 = layer_norm(axes = input_1151_axes_0, beta = module_layers_21_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_21_norm_feed_forward2_weight_to_fp16, x = input_1149_cast_fp16)[name = tensor("input_1151_cast_fp16")]; - tensor module_layers_21_feed_forward2_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(402064256))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(405210048))), name = tensor("module_layers_21_feed_forward2_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_197_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_21_feed_forward2_linear1_weight_to_fp16_palettized, x = input_1151_cast_fp16)[name = tensor("linear_197_cast_fp16")]; - tensor input_1155_cast_fp16 = silu(x = linear_197_cast_fp16)[name = tensor("input_1155_cast_fp16")]; - tensor module_layers_21_feed_forward2_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(405210240))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(408356032))), name = tensor("module_layers_21_feed_forward2_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_198_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_21_feed_forward2_linear2_weight_to_fp16_palettized, x = input_1155_cast_fp16)[name = tensor("linear_198_cast_fp16")]; - tensor var_3789_to_fp16 = const()[name = tensor("op_3789_to_fp16"), val = tensor(0x1p-1)]; - tensor var_3790_cast_fp16 = mul(x = linear_198_cast_fp16, y = var_3789_to_fp16)[name = tensor("op_3790_cast_fp16")]; - tensor input_1161_cast_fp16 = add(x = input_1149_cast_fp16, y = var_3790_cast_fp16)[name = tensor("input_1161_cast_fp16")]; - tensor input_1163_axes_0 = const()[name = tensor("input_1163_axes_0"), val = tensor([-1])]; - tensor module_layers_21_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_21_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(408356224)))]; - tensor module_layers_21_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_21_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(408358336)))]; - tensor input_1163_cast_fp16 = layer_norm(axes = input_1163_axes_0, beta = module_layers_21_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_21_norm_out_weight_to_fp16, x = input_1161_cast_fp16)[name = tensor("input_1163_cast_fp16")]; - tensor input_1165_axes_0 = const()[name = tensor("input_1165_axes_0"), val = tensor([-1])]; - tensor module_layers_22_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_22_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(408360448)))]; - tensor module_layers_22_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_22_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(408362560)))]; - tensor input_1165_cast_fp16 = layer_norm(axes = input_1165_axes_0, beta = module_layers_22_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_22_norm_feed_forward1_weight_to_fp16, x = input_1163_cast_fp16)[name = tensor("input_1165_cast_fp16")]; - tensor module_layers_22_feed_forward1_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(408364672))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(411510464))), name = tensor("module_layers_22_feed_forward1_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_199_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_22_feed_forward1_linear1_weight_to_fp16_palettized, x = input_1165_cast_fp16)[name = tensor("linear_199_cast_fp16")]; - tensor input_1169_cast_fp16 = silu(x = linear_199_cast_fp16)[name = tensor("input_1169_cast_fp16")]; - tensor module_layers_22_feed_forward1_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(411510656))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(414656448))), name = tensor("module_layers_22_feed_forward1_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_200_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_22_feed_forward1_linear2_weight_to_fp16_palettized, x = input_1169_cast_fp16)[name = tensor("linear_200_cast_fp16")]; - tensor var_3818_to_fp16 = const()[name = tensor("op_3818_to_fp16"), val = tensor(0x1p-1)]; - tensor var_3819_cast_fp16 = mul(x = linear_200_cast_fp16, y = var_3818_to_fp16)[name = tensor("op_3819_cast_fp16")]; - tensor input_1175_cast_fp16 = add(x = input_1163_cast_fp16, y = var_3819_cast_fp16)[name = tensor("input_1175_cast_fp16")]; - tensor query_45_axes_0 = const()[name = tensor("query_45_axes_0"), val = tensor([-1])]; - tensor module_layers_22_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_22_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(414656640)))]; - tensor module_layers_22_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_22_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(414658752)))]; - tensor query_45_cast_fp16 = layer_norm(axes = query_45_axes_0, beta = module_layers_22_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_22_norm_self_att_weight_to_fp16, x = input_1175_cast_fp16)[name = tensor("query_45_cast_fp16")]; - tensor module_layers_22_self_attn_linear_q_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(414660864))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(415447360))), name = tensor("module_layers_22_self_attn_linear_q_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_201_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_22_self_attn_linear_q_weight_to_fp16_palettized, x = query_45_cast_fp16)[name = tensor("linear_201_cast_fp16")]; - tensor var_3835 = const()[name = tensor("op_3835"), val = tensor([1, -1, 8, 128])]; - tensor q_133_cast_fp16 = reshape(shape = var_3835, x = linear_201_cast_fp16)[name = tensor("q_133_cast_fp16")]; - tensor module_layers_22_self_attn_linear_k_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(415447552))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(416234048))), name = tensor("module_layers_22_self_attn_linear_k_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_202_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_22_self_attn_linear_k_weight_to_fp16_palettized, x = query_45_cast_fp16)[name = tensor("linear_202_cast_fp16")]; - tensor var_3839 = const()[name = tensor("op_3839"), val = tensor([1, -1, 8, 128])]; - tensor k_89_cast_fp16 = reshape(shape = var_3839, x = linear_202_cast_fp16)[name = tensor("k_89_cast_fp16")]; - tensor module_layers_22_self_attn_linear_v_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(416234240))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(417020736))), name = tensor("module_layers_22_self_attn_linear_v_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_203_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_22_self_attn_linear_v_weight_to_fp16_palettized, x = query_45_cast_fp16)[name = tensor("linear_203_cast_fp16")]; - tensor var_3843 = const()[name = tensor("op_3843"), val = tensor([1, -1, 8, 128])]; - tensor v_45_cast_fp16 = reshape(shape = var_3843, x = linear_203_cast_fp16)[name = tensor("v_45_cast_fp16")]; - tensor value_47_perm_0 = const()[name = tensor("value_47_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor module_layers_22_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_22_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(417020928)))]; - tensor var_3855_cast_fp16 = add(x = q_133_cast_fp16, y = module_layers_22_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3855_cast_fp16")]; - tensor module_layers_22_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_22_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(417023040)))]; - tensor var_3857_cast_fp16 = add(x = q_133_cast_fp16, y = module_layers_22_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3857_cast_fp16")]; - tensor q_with_bias_v_45_perm_0 = const()[name = tensor("q_with_bias_v_45_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor x_491_transpose_x_0 = const()[name = tensor("x_491_transpose_x_0"), val = tensor(false)]; - tensor x_491_transpose_y_0 = const()[name = tensor("x_491_transpose_y_0"), val = tensor(false)]; - tensor op_3859_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(417025152))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(417313216))), name = tensor("op_3859_to_fp16_palettized"), shape = tensor([1, 8, 128, 375])]; - tensor q_with_bias_v_45_cast_fp16 = transpose(perm = q_with_bias_v_45_perm_0, x = var_3857_cast_fp16)[name = tensor("transpose_158")]; - tensor x_491_cast_fp16 = matmul(transpose_x = x_491_transpose_x_0, transpose_y = x_491_transpose_y_0, x = q_with_bias_v_45_cast_fp16, y = op_3859_to_fp16_palettized)[name = tensor("x_491_cast_fp16")]; - tensor x_493_pad_0 = const()[name = tensor("x_493_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_493_mode_0 = const()[name = tensor("x_493_mode_0"), val = tensor("constant")]; - tensor const_234_to_fp16 = const()[name = tensor("const_234_to_fp16"), val = tensor(0x0p+0)]; - tensor x_493_cast_fp16 = pad(constant_val = const_234_to_fp16, mode = x_493_mode_0, pad = x_493_pad_0, x = x_491_cast_fp16)[name = tensor("x_493_cast_fp16")]; - tensor var_3867 = const()[name = tensor("op_3867"), val = tensor([1, 8, -1, 188])]; - tensor x_495_cast_fp16 = reshape(shape = var_3867, x = x_493_cast_fp16)[name = tensor("x_495_cast_fp16")]; - tensor var_3871_begin_0 = const()[name = tensor("op_3871_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_3871_end_0 = const()[name = tensor("op_3871_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_3871_end_mask_0 = const()[name = tensor("op_3871_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_3871_cast_fp16 = slice_by_index(begin = var_3871_begin_0, end = var_3871_end_0, end_mask = var_3871_end_mask_0, x = x_495_cast_fp16)[name = tensor("op_3871_cast_fp16")]; - tensor var_3872 = const()[name = tensor("op_3872"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_89_cast_fp16 = reshape(shape = var_3872, x = var_3871_cast_fp16)[name = tensor("matrix_bd_89_cast_fp16")]; - tensor matrix_ac_45_transpose_x_0 = const()[name = tensor("matrix_ac_45_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_45_transpose_y_0 = const()[name = tensor("matrix_ac_45_transpose_y_0"), val = tensor(false)]; - tensor transpose_140_perm_0 = const()[name = tensor("transpose_140_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_141_perm_0 = const()[name = tensor("transpose_141_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_141 = transpose(perm = transpose_141_perm_0, x = k_89_cast_fp16)[name = tensor("transpose_156")]; - tensor transpose_140 = transpose(perm = transpose_140_perm_0, x = var_3855_cast_fp16)[name = tensor("transpose_157")]; - tensor matrix_ac_45_cast_fp16 = matmul(transpose_x = matrix_ac_45_transpose_x_0, transpose_y = matrix_ac_45_transpose_y_0, x = transpose_140, y = transpose_141)[name = tensor("matrix_ac_45_cast_fp16")]; - tensor matrix_bd_91_begin_0 = const()[name = tensor("matrix_bd_91_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_91_end_0 = const()[name = tensor("matrix_bd_91_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_91_end_mask_0 = const()[name = tensor("matrix_bd_91_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_91_cast_fp16 = slice_by_index(begin = matrix_bd_91_begin_0, end = matrix_bd_91_end_0, end_mask = matrix_bd_91_end_mask_0, x = matrix_bd_89_cast_fp16)[name = tensor("matrix_bd_91_cast_fp16")]; - tensor var_3881_cast_fp16 = add(x = matrix_ac_45_cast_fp16, y = matrix_bd_91_cast_fp16)[name = tensor("op_3881_cast_fp16")]; - tensor _inversed_scores_89_y_0_to_fp16 = const()[name = tensor("_inversed_scores_89_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; - tensor _inversed_scores_89_cast_fp16 = mul(x = var_3881_cast_fp16, y = _inversed_scores_89_y_0_to_fp16)[name = tensor("_inversed_scores_89_cast_fp16")]; - tensor scores_91_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_89_cast_fp16, cond = mask_3)[name = tensor("scores_91_cast_fp16")]; - tensor var_3887_cast_fp16 = softmax(axis = var_30, x = scores_91_cast_fp16)[name = tensor("op_3887_cast_fp16")]; - tensor input_1177_cast_fp16 = select(a = var_11_to_fp16, b = var_3887_cast_fp16, cond = mask_3)[name = tensor("input_1177_cast_fp16")]; - tensor x_497_transpose_x_0 = const()[name = tensor("x_497_transpose_x_0"), val = tensor(false)]; - tensor x_497_transpose_y_0 = const()[name = tensor("x_497_transpose_y_0"), val = tensor(false)]; - tensor value_47_cast_fp16 = transpose(perm = value_47_perm_0, x = v_45_cast_fp16)[name = tensor("transpose_155")]; - tensor x_497_cast_fp16 = matmul(transpose_x = x_497_transpose_x_0, transpose_y = x_497_transpose_y_0, x = input_1177_cast_fp16, y = value_47_cast_fp16)[name = tensor("x_497_cast_fp16")]; - tensor var_3891_perm_0 = const()[name = tensor("op_3891_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_3892 = const()[name = tensor("op_3892"), val = tensor([1, -1, 1024])]; - tensor var_3891_cast_fp16 = transpose(perm = var_3891_perm_0, x = x_497_cast_fp16)[name = tensor("transpose_154")]; - tensor input_1179_cast_fp16 = reshape(shape = var_3892, x = var_3891_cast_fp16)[name = tensor("input_1179_cast_fp16")]; - tensor module_layers_22_self_attn_linear_out_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(417313408))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(418099904))), name = tensor("module_layers_22_self_attn_linear_out_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_205_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_22_self_attn_linear_out_weight_to_fp16_palettized, x = input_1179_cast_fp16)[name = tensor("linear_205_cast_fp16")]; - tensor input_1183_cast_fp16 = add(x = input_1175_cast_fp16, y = linear_205_cast_fp16)[name = tensor("input_1183_cast_fp16")]; - tensor x_501_axes_0 = const()[name = tensor("x_501_axes_0"), val = tensor([-1])]; - tensor module_layers_22_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_22_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(418100096)))]; - tensor module_layers_22_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_22_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(418102208)))]; - tensor x_501_cast_fp16 = layer_norm(axes = x_501_axes_0, beta = module_layers_22_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_22_norm_conv_weight_to_fp16, x = input_1183_cast_fp16)[name = tensor("x_501_cast_fp16")]; - tensor input_1185_perm_0 = const()[name = tensor("input_1185_perm_0"), val = tensor([0, 2, 1])]; - tensor input_1187_pad_type_0 = const()[name = tensor("input_1187_pad_type_0"), val = tensor("valid")]; - tensor input_1187_strides_0 = const()[name = tensor("input_1187_strides_0"), val = tensor([1])]; - tensor input_1187_pad_0 = const()[name = tensor("input_1187_pad_0"), val = tensor([0, 0])]; - tensor input_1187_dilations_0 = const()[name = tensor("input_1187_dilations_0"), val = tensor([1])]; - tensor input_1187_groups_0 = const()[name = tensor("input_1187_groups_0"), val = tensor(1)]; - tensor module_layers_22_conv_pointwise_conv1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(418104320))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(419677248))), name = tensor("module_layers_22_conv_pointwise_conv1_weight_to_fp16_palettized"), shape = tensor([2048, 1024, 1])]; - tensor input_1185_cast_fp16 = transpose(perm = input_1185_perm_0, x = x_501_cast_fp16)[name = tensor("transpose_153")]; - tensor input_1187_cast_fp16 = conv(dilations = input_1187_dilations_0, groups = input_1187_groups_0, pad = input_1187_pad_0, pad_type = input_1187_pad_type_0, strides = input_1187_strides_0, weight = module_layers_22_conv_pointwise_conv1_weight_to_fp16_palettized, x = input_1185_cast_fp16)[name = tensor("input_1187_cast_fp16")]; - tensor x_503_split_num_splits_0 = const()[name = tensor("x_503_split_num_splits_0"), val = tensor(2)]; - tensor x_503_split_axis_0 = const()[name = tensor("x_503_split_axis_0"), val = tensor(1)]; - tensor x_503_split_cast_fp16_0, tensor x_503_split_cast_fp16_1 = split(axis = x_503_split_axis_0, num_splits = x_503_split_num_splits_0, x = input_1187_cast_fp16)[name = tensor("x_503_split_cast_fp16")]; - tensor x_503_split_1_sigmoid_cast_fp16 = sigmoid(x = x_503_split_cast_fp16_1)[name = tensor("x_503_split_1_sigmoid_cast_fp16")]; - tensor x_503_cast_fp16 = mul(x = x_503_split_cast_fp16_0, y = x_503_split_1_sigmoid_cast_fp16)[name = tensor("x_503_cast_fp16")]; - tensor input_1189_cast_fp16 = select(a = var_11_to_fp16, b = x_503_cast_fp16, cond = var_328)[name = tensor("input_1189_cast_fp16")]; - tensor input_1191_pad_0 = const()[name = tensor("input_1191_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_1191_mode_0 = const()[name = tensor("input_1191_mode_0"), val = tensor("constant")]; - tensor const_237_to_fp16 = const()[name = tensor("const_237_to_fp16"), val = tensor(0x0p+0)]; - tensor input_1191_cast_fp16 = pad(constant_val = const_237_to_fp16, mode = input_1191_mode_0, pad = input_1191_pad_0, x = input_1189_cast_fp16)[name = tensor("input_1191_cast_fp16")]; - tensor input_1193_pad_type_0 = const()[name = tensor("input_1193_pad_type_0"), val = tensor("valid")]; - tensor input_1193_groups_0 = const()[name = tensor("input_1193_groups_0"), val = tensor(1024)]; - tensor input_1193_strides_0 = const()[name = tensor("input_1193_strides_0"), val = tensor([1])]; - tensor input_1193_pad_0 = const()[name = tensor("input_1193_pad_0"), val = tensor([0, 0])]; - tensor input_1193_dilations_0 = const()[name = tensor("input_1193_dilations_0"), val = tensor([1])]; - tensor const_292_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(419677440))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(419684416))), name = tensor("const_292_to_fp16_palettized"), shape = tensor([1024, 1, 9])]; - tensor const_293_to_fp16 = const()[name = tensor("const_293_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(419684608)))]; - tensor input_1195_cast_fp16 = conv(bias = const_293_to_fp16, dilations = input_1193_dilations_0, groups = input_1193_groups_0, pad = input_1193_pad_0, pad_type = input_1193_pad_type_0, strides = input_1193_strides_0, weight = const_292_to_fp16_palettized, x = input_1191_cast_fp16)[name = tensor("input_1195_cast_fp16")]; - tensor input_1197_cast_fp16 = silu(x = input_1195_cast_fp16)[name = tensor("input_1197_cast_fp16")]; - tensor x_505_pad_type_0 = const()[name = tensor("x_505_pad_type_0"), val = tensor("valid")]; - tensor x_505_strides_0 = const()[name = tensor("x_505_strides_0"), val = tensor([1])]; - tensor x_505_pad_0 = const()[name = tensor("x_505_pad_0"), val = tensor([0, 0])]; - tensor x_505_dilations_0 = const()[name = tensor("x_505_dilations_0"), val = tensor([1])]; - tensor x_505_groups_0 = const()[name = tensor("x_505_groups_0"), val = tensor(1)]; - tensor module_layers_22_conv_pointwise_conv2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(419686720))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(420473216))), name = tensor("module_layers_22_conv_pointwise_conv2_weight_to_fp16_palettized"), shape = tensor([1024, 1024, 1])]; - tensor x_505_cast_fp16 = conv(dilations = x_505_dilations_0, groups = x_505_groups_0, pad = x_505_pad_0, pad_type = x_505_pad_type_0, strides = x_505_strides_0, weight = module_layers_22_conv_pointwise_conv2_weight_to_fp16_palettized, x = input_1197_cast_fp16)[name = tensor("x_505_cast_fp16")]; - tensor input_1199_perm_0 = const()[name = tensor("input_1199_perm_0"), val = tensor([0, 2, 1])]; - tensor input_1199_cast_fp16 = transpose(perm = input_1199_perm_0, x = x_505_cast_fp16)[name = tensor("transpose_152")]; - tensor input_1201_cast_fp16 = add(x = input_1183_cast_fp16, y = input_1199_cast_fp16)[name = tensor("input_1201_cast_fp16")]; - tensor input_1203_axes_0 = const()[name = tensor("input_1203_axes_0"), val = tensor([-1])]; - tensor module_layers_22_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_22_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(420473408)))]; - tensor module_layers_22_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_22_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(420475520)))]; - tensor input_1203_cast_fp16 = layer_norm(axes = input_1203_axes_0, beta = module_layers_22_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_22_norm_feed_forward2_weight_to_fp16, x = input_1201_cast_fp16)[name = tensor("input_1203_cast_fp16")]; - tensor module_layers_22_feed_forward2_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(420477632))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(423623424))), name = tensor("module_layers_22_feed_forward2_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_206_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_22_feed_forward2_linear1_weight_to_fp16_palettized, x = input_1203_cast_fp16)[name = tensor("linear_206_cast_fp16")]; - tensor input_1207_cast_fp16 = silu(x = linear_206_cast_fp16)[name = tensor("input_1207_cast_fp16")]; - tensor module_layers_22_feed_forward2_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(423623616))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(426769408))), name = tensor("module_layers_22_feed_forward2_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_207_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_22_feed_forward2_linear2_weight_to_fp16_palettized, x = input_1207_cast_fp16)[name = tensor("linear_207_cast_fp16")]; - tensor var_3952_to_fp16 = const()[name = tensor("op_3952_to_fp16"), val = tensor(0x1p-1)]; - tensor var_3953_cast_fp16 = mul(x = linear_207_cast_fp16, y = var_3952_to_fp16)[name = tensor("op_3953_cast_fp16")]; - tensor input_1213_cast_fp16 = add(x = input_1201_cast_fp16, y = var_3953_cast_fp16)[name = tensor("input_1213_cast_fp16")]; - tensor input_1215_axes_0 = const()[name = tensor("input_1215_axes_0"), val = tensor([-1])]; - tensor module_layers_22_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_22_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(426769600)))]; - tensor module_layers_22_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_22_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(426771712)))]; - tensor input_1215_cast_fp16 = layer_norm(axes = input_1215_axes_0, beta = module_layers_22_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_22_norm_out_weight_to_fp16, x = input_1213_cast_fp16)[name = tensor("input_1215_cast_fp16")]; - tensor input_1217_axes_0 = const()[name = tensor("input_1217_axes_0"), val = tensor([-1])]; - tensor module_layers_23_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_23_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(426773824)))]; - tensor module_layers_23_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_23_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(426775936)))]; - tensor input_1217_cast_fp16 = layer_norm(axes = input_1217_axes_0, beta = module_layers_23_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_23_norm_feed_forward1_weight_to_fp16, x = input_1215_cast_fp16)[name = tensor("input_1217_cast_fp16")]; - tensor module_layers_23_feed_forward1_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(426778048))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(429923840))), name = tensor("module_layers_23_feed_forward1_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_208_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_23_feed_forward1_linear1_weight_to_fp16_palettized, x = input_1217_cast_fp16)[name = tensor("linear_208_cast_fp16")]; - tensor input_1221_cast_fp16 = silu(x = linear_208_cast_fp16)[name = tensor("input_1221_cast_fp16")]; - tensor module_layers_23_feed_forward1_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(429924032))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(433069824))), name = tensor("module_layers_23_feed_forward1_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_209_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_23_feed_forward1_linear2_weight_to_fp16_palettized, x = input_1221_cast_fp16)[name = tensor("linear_209_cast_fp16")]; - tensor var_3981_to_fp16 = const()[name = tensor("op_3981_to_fp16"), val = tensor(0x1p-1)]; - tensor var_3982_cast_fp16 = mul(x = linear_209_cast_fp16, y = var_3981_to_fp16)[name = tensor("op_3982_cast_fp16")]; - tensor input_1227_cast_fp16 = add(x = input_1215_cast_fp16, y = var_3982_cast_fp16)[name = tensor("input_1227_cast_fp16")]; - tensor query_axes_0 = const()[name = tensor("query_axes_0"), val = tensor([-1])]; - tensor module_layers_23_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_23_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(433070016)))]; - tensor module_layers_23_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_23_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(433072128)))]; - tensor query_cast_fp16 = layer_norm(axes = query_axes_0, beta = module_layers_23_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_23_norm_self_att_weight_to_fp16, x = input_1227_cast_fp16)[name = tensor("query_cast_fp16")]; - tensor module_layers_23_self_attn_linear_q_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(433074240))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(433860736))), name = tensor("module_layers_23_self_attn_linear_q_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_210_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_23_self_attn_linear_q_weight_to_fp16_palettized, x = query_cast_fp16)[name = tensor("linear_210_cast_fp16")]; - tensor var_3998 = const()[name = tensor("op_3998"), val = tensor([1, -1, 8, 128])]; - tensor q_139_cast_fp16 = reshape(shape = var_3998, x = linear_210_cast_fp16)[name = tensor("q_139_cast_fp16")]; - tensor module_layers_23_self_attn_linear_k_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(433860928))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(434647424))), name = tensor("module_layers_23_self_attn_linear_k_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_211_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_23_self_attn_linear_k_weight_to_fp16_palettized, x = query_cast_fp16)[name = tensor("linear_211_cast_fp16")]; - tensor var_4002 = const()[name = tensor("op_4002"), val = tensor([1, -1, 8, 128])]; - tensor k_93_cast_fp16 = reshape(shape = var_4002, x = linear_211_cast_fp16)[name = tensor("k_93_cast_fp16")]; - tensor module_layers_23_self_attn_linear_v_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(434647616))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(435434112))), name = tensor("module_layers_23_self_attn_linear_v_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_212_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_23_self_attn_linear_v_weight_to_fp16_palettized, x = query_cast_fp16)[name = tensor("linear_212_cast_fp16")]; - tensor var_4006 = const()[name = tensor("op_4006"), val = tensor([1, -1, 8, 128])]; - tensor v_cast_fp16 = reshape(shape = var_4006, x = linear_212_cast_fp16)[name = tensor("v_cast_fp16")]; - tensor value_perm_0 = const()[name = tensor("value_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor module_layers_23_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_23_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(435434304)))]; - tensor var_4018_cast_fp16 = add(x = q_139_cast_fp16, y = module_layers_23_self_attn_pos_bias_u_to_fp16)[name = tensor("op_4018_cast_fp16")]; - tensor module_layers_23_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_23_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(435436416)))]; - tensor var_4020_cast_fp16 = add(x = q_139_cast_fp16, y = module_layers_23_self_attn_pos_bias_v_to_fp16)[name = tensor("op_4020_cast_fp16")]; - tensor q_with_bias_v_perm_0 = const()[name = tensor("q_with_bias_v_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor x_513_transpose_x_0 = const()[name = tensor("x_513_transpose_x_0"), val = tensor(false)]; - tensor x_513_transpose_y_0 = const()[name = tensor("x_513_transpose_y_0"), val = tensor(false)]; - tensor op_4022_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(435438528))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(435726592))), name = tensor("op_4022_to_fp16_palettized"), shape = tensor([1, 8, 128, 375])]; - tensor q_with_bias_v_cast_fp16 = transpose(perm = q_with_bias_v_perm_0, x = var_4020_cast_fp16)[name = tensor("transpose_151")]; - tensor x_513_cast_fp16 = matmul(transpose_x = x_513_transpose_x_0, transpose_y = x_513_transpose_y_0, x = q_with_bias_v_cast_fp16, y = op_4022_to_fp16_palettized)[name = tensor("x_513_cast_fp16")]; - tensor x_515_pad_0 = const()[name = tensor("x_515_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_515_mode_0 = const()[name = tensor("x_515_mode_0"), val = tensor("constant")]; - tensor const_244_to_fp16 = const()[name = tensor("const_244_to_fp16"), val = tensor(0x0p+0)]; - tensor x_515_cast_fp16 = pad(constant_val = const_244_to_fp16, mode = x_515_mode_0, pad = x_515_pad_0, x = x_513_cast_fp16)[name = tensor("x_515_cast_fp16")]; - tensor var_4030 = const()[name = tensor("op_4030"), val = tensor([1, 8, -1, 188])]; - tensor x_517_cast_fp16 = reshape(shape = var_4030, x = x_515_cast_fp16)[name = tensor("x_517_cast_fp16")]; - tensor var_4034_begin_0 = const()[name = tensor("op_4034_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_4034_end_0 = const()[name = tensor("op_4034_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_4034_end_mask_0 = const()[name = tensor("op_4034_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_4034_cast_fp16 = slice_by_index(begin = var_4034_begin_0, end = var_4034_end_0, end_mask = var_4034_end_mask_0, x = x_517_cast_fp16)[name = tensor("op_4034_cast_fp16")]; - tensor var_4035 = const()[name = tensor("op_4035"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_93_cast_fp16 = reshape(shape = var_4035, x = var_4034_cast_fp16)[name = tensor("matrix_bd_93_cast_fp16")]; - tensor matrix_ac_transpose_x_0 = const()[name = tensor("matrix_ac_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_transpose_y_0 = const()[name = tensor("matrix_ac_transpose_y_0"), val = tensor(false)]; - tensor transpose_142_perm_0 = const()[name = tensor("transpose_142_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_143_perm_0 = const()[name = tensor("transpose_143_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_143 = transpose(perm = transpose_143_perm_0, x = k_93_cast_fp16)[name = tensor("transpose_149")]; - tensor transpose_142 = transpose(perm = transpose_142_perm_0, x = var_4018_cast_fp16)[name = tensor("transpose_150")]; - tensor matrix_ac_cast_fp16 = matmul(transpose_x = matrix_ac_transpose_x_0, transpose_y = matrix_ac_transpose_y_0, x = transpose_142, y = transpose_143)[name = tensor("matrix_ac_cast_fp16")]; - tensor matrix_bd_begin_0 = const()[name = tensor("matrix_bd_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_end_0 = const()[name = tensor("matrix_bd_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_end_mask_0 = const()[name = tensor("matrix_bd_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_cast_fp16 = slice_by_index(begin = matrix_bd_begin_0, end = matrix_bd_end_0, end_mask = matrix_bd_end_mask_0, x = matrix_bd_93_cast_fp16)[name = tensor("matrix_bd_cast_fp16")]; - tensor var_4044_cast_fp16 = add(x = matrix_ac_cast_fp16, y = matrix_bd_cast_fp16)[name = tensor("op_4044_cast_fp16")]; - tensor _inversed_scores_93_y_0_to_fp16 = const()[name = tensor("_inversed_scores_93_y_0_to_fp16"), val = tensor(0x1.6ap-4)]; - tensor _inversed_scores_93_cast_fp16 = mul(x = var_4044_cast_fp16, y = _inversed_scores_93_y_0_to_fp16)[name = tensor("_inversed_scores_93_cast_fp16")]; - tensor scores_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_93_cast_fp16, cond = mask_3)[name = tensor("scores_cast_fp16")]; - tensor var_4050_cast_fp16 = softmax(axis = var_30, x = scores_cast_fp16)[name = tensor("op_4050_cast_fp16")]; - tensor input_1229_cast_fp16 = select(a = var_11_to_fp16, b = var_4050_cast_fp16, cond = mask_3)[name = tensor("input_1229_cast_fp16")]; - tensor x_519_transpose_x_0 = const()[name = tensor("x_519_transpose_x_0"), val = tensor(false)]; - tensor x_519_transpose_y_0 = const()[name = tensor("x_519_transpose_y_0"), val = tensor(false)]; - tensor value_cast_fp16 = transpose(perm = value_perm_0, x = v_cast_fp16)[name = tensor("transpose_148")]; - tensor x_519_cast_fp16 = matmul(transpose_x = x_519_transpose_x_0, transpose_y = x_519_transpose_y_0, x = input_1229_cast_fp16, y = value_cast_fp16)[name = tensor("x_519_cast_fp16")]; - tensor var_4054_perm_0 = const()[name = tensor("op_4054_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_4055 = const()[name = tensor("op_4055"), val = tensor([1, -1, 1024])]; - tensor var_4054_cast_fp16 = transpose(perm = var_4054_perm_0, x = x_519_cast_fp16)[name = tensor("transpose_147")]; - tensor input_1231_cast_fp16 = reshape(shape = var_4055, x = var_4054_cast_fp16)[name = tensor("input_1231_cast_fp16")]; - tensor module_layers_23_self_attn_linear_out_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(435726784))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(436513280))), name = tensor("module_layers_23_self_attn_linear_out_weight_to_fp16_palettized"), shape = tensor([1024, 1024])]; - tensor linear_214_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_23_self_attn_linear_out_weight_to_fp16_palettized, x = input_1231_cast_fp16)[name = tensor("linear_214_cast_fp16")]; - tensor input_1235_cast_fp16 = add(x = input_1227_cast_fp16, y = linear_214_cast_fp16)[name = tensor("input_1235_cast_fp16")]; - tensor x_523_axes_0 = const()[name = tensor("x_523_axes_0"), val = tensor([-1])]; - tensor module_layers_23_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_23_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(436513472)))]; - tensor module_layers_23_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_23_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(436515584)))]; - tensor x_523_cast_fp16 = layer_norm(axes = x_523_axes_0, beta = module_layers_23_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_23_norm_conv_weight_to_fp16, x = input_1235_cast_fp16)[name = tensor("x_523_cast_fp16")]; - tensor input_1237_perm_0 = const()[name = tensor("input_1237_perm_0"), val = tensor([0, 2, 1])]; - tensor input_1239_pad_type_0 = const()[name = tensor("input_1239_pad_type_0"), val = tensor("valid")]; - tensor input_1239_strides_0 = const()[name = tensor("input_1239_strides_0"), val = tensor([1])]; - tensor input_1239_pad_0 = const()[name = tensor("input_1239_pad_0"), val = tensor([0, 0])]; - tensor input_1239_dilations_0 = const()[name = tensor("input_1239_dilations_0"), val = tensor([1])]; - tensor input_1239_groups_0 = const()[name = tensor("input_1239_groups_0"), val = tensor(1)]; - tensor module_layers_23_conv_pointwise_conv1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(436517696))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(438090624))), name = tensor("module_layers_23_conv_pointwise_conv1_weight_to_fp16_palettized"), shape = tensor([2048, 1024, 1])]; - tensor input_1237_cast_fp16 = transpose(perm = input_1237_perm_0, x = x_523_cast_fp16)[name = tensor("transpose_146")]; - tensor input_1239_cast_fp16 = conv(dilations = input_1239_dilations_0, groups = input_1239_groups_0, pad = input_1239_pad_0, pad_type = input_1239_pad_type_0, strides = input_1239_strides_0, weight = module_layers_23_conv_pointwise_conv1_weight_to_fp16_palettized, x = input_1237_cast_fp16)[name = tensor("input_1239_cast_fp16")]; - tensor x_525_split_num_splits_0 = const()[name = tensor("x_525_split_num_splits_0"), val = tensor(2)]; - tensor x_525_split_axis_0 = const()[name = tensor("x_525_split_axis_0"), val = tensor(1)]; - tensor x_525_split_cast_fp16_0, tensor x_525_split_cast_fp16_1 = split(axis = x_525_split_axis_0, num_splits = x_525_split_num_splits_0, x = input_1239_cast_fp16)[name = tensor("x_525_split_cast_fp16")]; - tensor x_525_split_1_sigmoid_cast_fp16 = sigmoid(x = x_525_split_cast_fp16_1)[name = tensor("x_525_split_1_sigmoid_cast_fp16")]; - tensor x_525_cast_fp16 = mul(x = x_525_split_cast_fp16_0, y = x_525_split_1_sigmoid_cast_fp16)[name = tensor("x_525_cast_fp16")]; - tensor input_1241_cast_fp16 = select(a = var_11_to_fp16, b = x_525_cast_fp16, cond = var_328)[name = tensor("input_1241_cast_fp16")]; - tensor input_1243_pad_0 = const()[name = tensor("input_1243_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_1243_mode_0 = const()[name = tensor("input_1243_mode_0"), val = tensor("constant")]; - tensor const_247_to_fp16 = const()[name = tensor("const_247_to_fp16"), val = tensor(0x0p+0)]; - tensor input_1243_cast_fp16 = pad(constant_val = const_247_to_fp16, mode = input_1243_mode_0, pad = input_1243_pad_0, x = input_1241_cast_fp16)[name = tensor("input_1243_cast_fp16")]; - tensor input_1245_pad_type_0 = const()[name = tensor("input_1245_pad_type_0"), val = tensor("valid")]; - tensor input_1245_groups_0 = const()[name = tensor("input_1245_groups_0"), val = tensor(1024)]; - tensor input_1245_strides_0 = const()[name = tensor("input_1245_strides_0"), val = tensor([1])]; - tensor input_1245_pad_0 = const()[name = tensor("input_1245_pad_0"), val = tensor([0, 0])]; - tensor input_1245_dilations_0 = const()[name = tensor("input_1245_dilations_0"), val = tensor([1])]; - tensor const_294_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(438090816))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(438097792))), name = tensor("const_294_to_fp16_palettized"), shape = tensor([1024, 1, 9])]; - tensor const_295_to_fp16 = const()[name = tensor("const_295_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(438097984)))]; - tensor input_1247_cast_fp16 = conv(bias = const_295_to_fp16, dilations = input_1245_dilations_0, groups = input_1245_groups_0, pad = input_1245_pad_0, pad_type = input_1245_pad_type_0, strides = input_1245_strides_0, weight = const_294_to_fp16_palettized, x = input_1243_cast_fp16)[name = tensor("input_1247_cast_fp16")]; - tensor input_1249_cast_fp16 = silu(x = input_1247_cast_fp16)[name = tensor("input_1249_cast_fp16")]; - tensor x_527_pad_type_0 = const()[name = tensor("x_527_pad_type_0"), val = tensor("valid")]; - tensor x_527_strides_0 = const()[name = tensor("x_527_strides_0"), val = tensor([1])]; - tensor x_527_pad_0 = const()[name = tensor("x_527_pad_0"), val = tensor([0, 0])]; - tensor x_527_dilations_0 = const()[name = tensor("x_527_dilations_0"), val = tensor([1])]; - tensor x_527_groups_0 = const()[name = tensor("x_527_groups_0"), val = tensor(1)]; - tensor module_layers_23_conv_pointwise_conv2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(438100096))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(438886592))), name = tensor("module_layers_23_conv_pointwise_conv2_weight_to_fp16_palettized"), shape = tensor([1024, 1024, 1])]; - tensor x_527_cast_fp16 = conv(dilations = x_527_dilations_0, groups = x_527_groups_0, pad = x_527_pad_0, pad_type = x_527_pad_type_0, strides = x_527_strides_0, weight = module_layers_23_conv_pointwise_conv2_weight_to_fp16_palettized, x = input_1249_cast_fp16)[name = tensor("x_527_cast_fp16")]; - tensor input_1251_perm_0 = const()[name = tensor("input_1251_perm_0"), val = tensor([0, 2, 1])]; - tensor input_1251_cast_fp16 = transpose(perm = input_1251_perm_0, x = x_527_cast_fp16)[name = tensor("transpose_145")]; - tensor input_1253_cast_fp16 = add(x = input_1235_cast_fp16, y = input_1251_cast_fp16)[name = tensor("input_1253_cast_fp16")]; - tensor input_1255_axes_0 = const()[name = tensor("input_1255_axes_0"), val = tensor([-1])]; - tensor module_layers_23_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_23_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(438886784)))]; - tensor module_layers_23_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_23_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(438888896)))]; - tensor input_1255_cast_fp16 = layer_norm(axes = input_1255_axes_0, beta = module_layers_23_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_23_norm_feed_forward2_weight_to_fp16, x = input_1253_cast_fp16)[name = tensor("input_1255_cast_fp16")]; - tensor module_layers_23_feed_forward2_linear1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(438891008))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(442036800))), name = tensor("module_layers_23_feed_forward2_linear1_weight_to_fp16_palettized"), shape = tensor([4096, 1024])]; - tensor linear_215_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = module_layers_23_feed_forward2_linear1_weight_to_fp16_palettized, x = input_1255_cast_fp16)[name = tensor("linear_215_cast_fp16")]; - tensor input_1259_cast_fp16 = silu(x = linear_215_cast_fp16)[name = tensor("input_1259_cast_fp16")]; - tensor module_layers_23_feed_forward2_linear2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(442036992))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(445182784))), name = tensor("module_layers_23_feed_forward2_linear2_weight_to_fp16_palettized"), shape = tensor([1024, 4096])]; - tensor linear_216_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = module_layers_23_feed_forward2_linear2_weight_to_fp16_palettized, x = input_1259_cast_fp16)[name = tensor("linear_216_cast_fp16")]; - tensor var_4115_to_fp16 = const()[name = tensor("op_4115_to_fp16"), val = tensor(0x1p-1)]; - tensor var_4116_cast_fp16 = mul(x = linear_216_cast_fp16, y = var_4115_to_fp16)[name = tensor("op_4116_cast_fp16")]; - tensor input_cast_fp16 = add(x = input_1253_cast_fp16, y = var_4116_cast_fp16)[name = tensor("input_cast_fp16")]; - tensor audio_signal_axes_0 = const()[name = tensor("audio_signal_axes_0"), val = tensor([-1])]; - tensor module_layers_23_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_23_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(445182976)))]; - tensor module_layers_23_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_23_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(445185088)))]; - tensor audio_signal_cast_fp16 = layer_norm(axes = audio_signal_axes_0, beta = module_layers_23_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_23_norm_out_weight_to_fp16, x = input_cast_fp16)[name = tensor("audio_signal_cast_fp16")]; - tensor obj_1_perm_0 = const()[name = tensor("obj_1_perm_0"), val = tensor([0, 2, 1])]; - tensor obj_1_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("obj_1_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; - tensor obj_1_cast_fp16 = transpose(perm = obj_1_perm_0, x = audio_signal_cast_fp16)[name = tensor("transpose_144")]; - tensor encoder = cast(dtype = obj_1_cast_fp16_to_fp32_dtype_0, x = obj_1_cast_fp16)[name = tensor("cast_0")]; - } -> (encoder, encoder_length); -} \ No newline at end of file diff --git a/parakeet-tdt-0.6b-v2-coreml/Encoder.mlmodelc/weights/weight.bin b/parakeet-tdt-0.6b-v2-coreml/Encoder.mlmodelc/weights/weight.bin deleted file mode 100644 index e6313f81d9d94af46347f28087f51a60eb28164c..0000000000000000000000000000000000000000 --- a/parakeet-tdt-0.6b-v2-coreml/Encoder.mlmodelc/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:4adc7ad44f9d05e1bffeb2b06d3bb02861a5c7602dff63a6b494aed3bf8a6c3e -size 445187200 diff --git a/parakeet-tdt-0.6b-v2-coreml/JointDecision.mlmodelc/analytics/coremldata.bin b/parakeet-tdt-0.6b-v2-coreml/JointDecision.mlmodelc/analytics/coremldata.bin deleted file mode 100644 index 5007841fb604e70990e535a2a002ee18f93e83e4..0000000000000000000000000000000000000000 --- a/parakeet-tdt-0.6b-v2-coreml/JointDecision.mlmodelc/analytics/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:f1183ba213bb94a918c8d2cad19ab045320618f97f6ca662245b3936d7b090f7 -size 243 diff --git a/parakeet-tdt-0.6b-v2-coreml/JointDecision.mlmodelc/coremldata.bin b/parakeet-tdt-0.6b-v2-coreml/JointDecision.mlmodelc/coremldata.bin deleted file mode 100644 index 72290f46bfb535192ed1ed5ca1fec941874efbca..0000000000000000000000000000000000000000 --- a/parakeet-tdt-0.6b-v2-coreml/JointDecision.mlmodelc/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:e2c6752f1c8cf2d3f6f26ec93195c9bfa759ad59edf9f806696a138154f96f11 -size 534 diff --git a/parakeet-tdt-0.6b-v2-coreml/JointDecision.mlmodelc/metadata.json b/parakeet-tdt-0.6b-v2-coreml/JointDecision.mlmodelc/metadata.json deleted file mode 100644 index db3b2bc33cfda6f7cd1073d21182446180e53f94..0000000000000000000000000000000000000000 --- a/parakeet-tdt-0.6b-v2-coreml/JointDecision.mlmodelc/metadata.json +++ /dev/null @@ -1,103 +0,0 @@ -[ - { - "metadataOutputVersion" : "3.0", - "shortDescription" : "Parakeet single-step joint decision (current frame)", - "outputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1 × 1 × 1)", - "shortDescription" : "", - "shape" : "[1, 1, 1]", - "name" : "token_id", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 1 × 1)", - "shortDescription" : "", - "shape" : "[1, 1, 1]", - "name" : "token_prob", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1 × 1 × 1)", - "shortDescription" : "", - "shape" : "[1, 1, 1]", - "name" : "duration", - "type" : "MultiArray" - } - ], - "storagePrecision" : "Float16", - "modelParameters" : [ - - ], - "author" : "Fluid Inference", - "specificationVersion" : 8, - "mlProgramOperationTypeHistogram" : { - "Ios17.reduceArgmax" : 2, - "Ios17.squeeze" : 1, - "Ios17.cast" : 4, - "Ios17.linear" : 3, - "Ios17.transpose" : 2, - "Ios17.sliceByIndex" : 2, - "Ios17.add" : 1, - "Ios16.relu" : 1, - "Ios16.softmax" : 1, - "Ios17.gatherAlongAxis" : 1, - "Ios17.expandDims" : 3 - }, - "computePrecision" : "Mixed (Float16, Float32, Int16, Int32)", - "isUpdatable" : "0", - "stateSchema" : [ - - ], - "availability" : { - "macOS" : "14.0", - "tvOS" : "17.0", - "visionOS" : "1.0", - "watchOS" : "10.0", - "iOS" : "17.0", - "macCatalyst" : "17.0" - }, - "modelType" : { - "name" : "MLModelType_mlProgram" - }, - "inputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 1024 × 1)", - "shortDescription" : "", - "shape" : "[1, 1024, 1]", - "name" : "encoder_step", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 640 × 1)", - "shortDescription" : "", - "shape" : "[1, 640, 1]", - "name" : "decoder_step", - "type" : "MultiArray" - } - ], - "userDefinedMetadata" : { - "com.github.apple.coremltools.conversion_date" : "2025-09-25", - "com.github.apple.coremltools.source" : "torch==2.7.0", - "com.github.apple.coremltools.version" : "9.0b1", - "com.github.apple.coremltools.source_dialect" : "TorchScript" - }, - "generatedClassName" : "parakeet_joint_decision_single_step", - "method" : "predict" - } -] \ No newline at end of file diff --git a/parakeet-tdt-0.6b-v2-coreml/JointDecision.mlmodelc/model.mil b/parakeet-tdt-0.6b-v2-coreml/JointDecision.mlmodelc/model.mil deleted file mode 100644 index 339c35b303b074de7fe047bb099311a4cc08cf53..0000000000000000000000000000000000000000 --- a/parakeet-tdt-0.6b-v2-coreml/JointDecision.mlmodelc/model.mil +++ /dev/null @@ -1,58 +0,0 @@ -program(1.0) -[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3500.14.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.7.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "9.0b1"}})] -{ - func main(tensor decoder_step, tensor encoder_step) { - tensor input_1_perm_0 = const()[name = tensor("input_1_perm_0"), val = tensor([0, 2, 1])]; - tensor encoder_step_to_fp16_dtype_0 = const()[name = tensor("encoder_step_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor input_3_perm_0 = const()[name = tensor("input_3_perm_0"), val = tensor([0, 2, 1])]; - tensor decoder_step_to_fp16_dtype_0 = const()[name = tensor("decoder_step_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor joint_module_enc_weight_to_fp16 = const()[name = tensor("joint_module_enc_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; - tensor joint_module_enc_bias_to_fp16 = const()[name = tensor("joint_module_enc_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1310848)))]; - tensor encoder_step_to_fp16 = cast(dtype = encoder_step_to_fp16_dtype_0, x = encoder_step)[name = tensor("cast_3")]; - tensor input_1_cast_fp16 = transpose(perm = input_1_perm_0, x = encoder_step_to_fp16)[name = tensor("transpose_1")]; - tensor linear_0_cast_fp16 = linear(bias = joint_module_enc_bias_to_fp16, weight = joint_module_enc_weight_to_fp16, x = input_1_cast_fp16)[name = tensor("linear_0_cast_fp16")]; - tensor joint_module_pred_weight_to_fp16 = const()[name = tensor("joint_module_pred_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1312192)))]; - tensor joint_module_pred_bias_to_fp16 = const()[name = tensor("joint_module_pred_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2131456)))]; - tensor decoder_step_to_fp16 = cast(dtype = decoder_step_to_fp16_dtype_0, x = decoder_step)[name = tensor("cast_2")]; - tensor input_3_cast_fp16 = transpose(perm = input_3_perm_0, x = decoder_step_to_fp16)[name = tensor("transpose_0")]; - tensor linear_1_cast_fp16 = linear(bias = joint_module_pred_bias_to_fp16, weight = joint_module_pred_weight_to_fp16, x = input_3_cast_fp16)[name = tensor("linear_1_cast_fp16")]; - tensor var_23_axes_0 = const()[name = tensor("op_23_axes_0"), val = tensor([2])]; - tensor var_23_cast_fp16 = expand_dims(axes = var_23_axes_0, x = linear_0_cast_fp16)[name = tensor("op_23_cast_fp16")]; - tensor var_24_axes_0 = const()[name = tensor("op_24_axes_0"), val = tensor([1])]; - tensor var_24_cast_fp16 = expand_dims(axes = var_24_axes_0, x = linear_1_cast_fp16)[name = tensor("op_24_cast_fp16")]; - tensor input_5_cast_fp16 = add(x = var_23_cast_fp16, y = var_24_cast_fp16)[name = tensor("input_5_cast_fp16")]; - tensor input_7_cast_fp16 = relu(x = input_5_cast_fp16)[name = tensor("input_7_cast_fp16")]; - tensor joint_module_joint_net_2_weight_to_fp16 = const()[name = tensor("joint_module_joint_net_2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2132800)))]; - tensor joint_module_joint_net_2_bias_to_fp16 = const()[name = tensor("joint_module_joint_net_2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3451264)))]; - tensor linear_2_cast_fp16 = linear(bias = joint_module_joint_net_2_bias_to_fp16, weight = joint_module_joint_net_2_weight_to_fp16, x = input_7_cast_fp16)[name = tensor("linear_2_cast_fp16")]; - tensor token_logits_begin_0 = const()[name = tensor("token_logits_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor token_logits_end_0 = const()[name = tensor("token_logits_end_0"), val = tensor([1, 1, 1, 1025])]; - tensor token_logits_end_mask_0 = const()[name = tensor("token_logits_end_mask_0"), val = tensor([true, true, true, false])]; - tensor token_logits_cast_fp16 = slice_by_index(begin = token_logits_begin_0, end = token_logits_end_0, end_mask = token_logits_end_mask_0, x = linear_2_cast_fp16)[name = tensor("token_logits_cast_fp16")]; - tensor duration_logits_begin_0 = const()[name = tensor("duration_logits_begin_0"), val = tensor([0, 0, 0, 1025])]; - tensor duration_logits_end_0 = const()[name = tensor("duration_logits_end_0"), val = tensor([1, 1, 1, 1030])]; - tensor duration_logits_end_mask_0 = const()[name = tensor("duration_logits_end_mask_0"), val = tensor([true, true, true, true])]; - tensor duration_logits_cast_fp16 = slice_by_index(begin = duration_logits_begin_0, end = duration_logits_end_0, end_mask = duration_logits_end_mask_0, x = linear_2_cast_fp16)[name = tensor("duration_logits_cast_fp16")]; - tensor var_43_axis_0 = const()[name = tensor("op_43_axis_0"), val = tensor(-1)]; - tensor var_43_keep_dims_0 = const()[name = tensor("op_43_keep_dims_0"), val = tensor(false)]; - tensor var_43_output_dtype_0 = const()[name = tensor("op_43_output_dtype_0"), val = tensor("int32")]; - tensor token_id = reduce_argmax(axis = var_43_axis_0, keep_dims = var_43_keep_dims_0, output_dtype = var_43_output_dtype_0, x = token_logits_cast_fp16)[name = tensor("op_43_cast_fp16")]; - tensor var_49 = const()[name = tensor("op_49"), val = tensor(-1)]; - tensor token_probs_all_cast_fp16 = softmax(axis = var_49, x = token_logits_cast_fp16)[name = tensor("token_probs_all_cast_fp16")]; - tensor var_58_axes_0 = const()[name = tensor("op_58_axes_0"), val = tensor([-1])]; - tensor var_58 = expand_dims(axes = var_58_axes_0, x = token_id)[name = tensor("op_58")]; - tensor var_59 = const()[name = tensor("op_59"), val = tensor(-1)]; - tensor var_61_validate_indices_0 = const()[name = tensor("op_61_validate_indices_0"), val = tensor(false)]; - tensor var_58_to_int16_dtype_0 = const()[name = tensor("op_58_to_int16_dtype_0"), val = tensor("int16")]; - tensor var_58_to_int16 = cast(dtype = var_58_to_int16_dtype_0, x = var_58)[name = tensor("cast_1")]; - tensor var_61_cast_fp16_cast_int16 = gather_along_axis(axis = var_59, indices = var_58_to_int16, validate_indices = var_61_validate_indices_0, x = token_probs_all_cast_fp16)[name = tensor("op_61_cast_fp16_cast_int16")]; - tensor var_63_axes_0 = const()[name = tensor("op_63_axes_0"), val = tensor([-1])]; - tensor var_63_cast_fp16 = squeeze(axes = var_63_axes_0, x = var_61_cast_fp16_cast_int16)[name = tensor("op_63_cast_fp16")]; - tensor var_63_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("op_63_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; - tensor var_66_axis_0 = const()[name = tensor("op_66_axis_0"), val = tensor(-1)]; - tensor var_66_keep_dims_0 = const()[name = tensor("op_66_keep_dims_0"), val = tensor(false)]; - tensor var_66_output_dtype_0 = const()[name = tensor("op_66_output_dtype_0"), val = tensor("int32")]; - tensor duration = reduce_argmax(axis = var_66_axis_0, keep_dims = var_66_keep_dims_0, output_dtype = var_66_output_dtype_0, x = duration_logits_cast_fp16)[name = tensor("op_66_cast_fp16")]; - tensor token_prob = cast(dtype = var_63_cast_fp16_to_fp32_dtype_0, x = var_63_cast_fp16)[name = tensor("cast_0")]; - } -> (token_id, token_prob, duration); -} \ No newline at end of file diff --git a/parakeet-tdt-0.6b-v2-coreml/JointDecision.mlmodelc/weights/weight.bin b/parakeet-tdt-0.6b-v2-coreml/JointDecision.mlmodelc/weights/weight.bin deleted file mode 100644 index 8764a151a786b38079c36a409fe5144ddf8b95a9..0000000000000000000000000000000000000000 --- a/parakeet-tdt-0.6b-v2-coreml/JointDecision.mlmodelc/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:ca22a65903a05e64137677da608077578a8606090a598abf4875fa6199aaa19d -size 3453388 diff --git a/parakeet-tdt-0.6b-v2-coreml/Preprocessor.mlmodelc/analytics/coremldata.bin b/parakeet-tdt-0.6b-v2-coreml/Preprocessor.mlmodelc/analytics/coremldata.bin deleted file mode 100644 index 84e7c818c15d2650b6c55b20e618fba3e6289763..0000000000000000000000000000000000000000 --- a/parakeet-tdt-0.6b-v2-coreml/Preprocessor.mlmodelc/analytics/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:03ab3c1327a054c54c07a40325db967ec574f2c91dcc8192bfa44aa561bcf2d8 -size 243 diff --git a/parakeet-tdt-0.6b-v2-coreml/Preprocessor.mlmodelc/coremldata.bin b/parakeet-tdt-0.6b-v2-coreml/Preprocessor.mlmodelc/coremldata.bin deleted file mode 100644 index b53fff2bc93fbe4b468811db15aa17767f3cb33c..0000000000000000000000000000000000000000 --- a/parakeet-tdt-0.6b-v2-coreml/Preprocessor.mlmodelc/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:d88ea1fc349459c9e100d6a96688c5b29a1f0d865f544be103001724b986b6d6 -size 494 diff --git a/parakeet-tdt-0.6b-v2-coreml/Preprocessor.mlmodelc/metadata.json b/parakeet-tdt-0.6b-v2-coreml/Preprocessor.mlmodelc/metadata.json deleted file mode 100644 index 887f7d4adc6d7f933686f7b984e23a47b45f57e3..0000000000000000000000000000000000000000 --- a/parakeet-tdt-0.6b-v2-coreml/Preprocessor.mlmodelc/metadata.json +++ /dev/null @@ -1,110 +0,0 @@ -[ - { - "metadataOutputVersion" : "3.0", - "shortDescription" : "int8-linear quantized - preprocessor", - "outputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32)", - "shortDescription" : "", - "shape" : "[]", - "name" : "mel", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1)", - "shortDescription" : "", - "shape" : "[1]", - "name" : "mel_length", - "type" : "MultiArray" - } - ], - "storagePrecision" : "Int8", - "modelParameters" : [ - - ], - "author" : "Fluid Inference", - "specificationVersion" : 8, - "mlProgramOperationTypeHistogram" : { - "Range1d" : 2, - "Ios17.reshape" : 2, - "Ios17.matmul" : 1, - "Ios17.expandDims" : 10, - "Select" : 3, - "Ios17.add" : 4, - "Tile" : 2, - "Ios17.sliceByIndex" : 3, - "Ios16.reduceSum" : 4, - "Shape" : 3, - "Ios17.gather" : 3, - "Pad" : 1, - "Ios17.log" : 1, - "Ios16.constexprAffineDequantize" : 3, - "Ios17.conv" : 2, - "Ios17.sub" : 4, - "Ios17.pow" : 2, - "Ios17.cast" : 10, - "Ios17.realDiv" : 4, - "Stack" : 1, - "Ios17.concat" : 3, - "Ios17.floorDiv" : 1, - "Ios17.less" : 1, - "Ios17.sqrt" : 1, - "Ios17.greaterEqual" : 1, - "Ios17.mul" : 1 - }, - "computePrecision" : "Mixed (Float16, Float32, Int32, UInt16)", - "isUpdatable" : "0", - "stateSchema" : [ - - ], - "availability" : { - "macOS" : "14.0", - "tvOS" : "17.0", - "visionOS" : "1.0", - "watchOS" : "10.0", - "iOS" : "17.0", - "macCatalyst" : "17.0" - }, - "modelType" : { - "name" : "MLModelType_mlProgram" - }, - "inputSchema" : [ - { - "dataType" : "Float32", - "hasShapeFlexibility" : "1", - "isOptional" : "0", - "shapeFlexibility" : "1 × 1...240000", - "shapeRange" : "[[1, 1], [1, 240000]]", - "formattedType" : "MultiArray (Float32 1 × 1)", - "type" : "MultiArray", - "shape" : "[1, 1]", - "name" : "audio_signal", - "shortDescription" : "" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1)", - "shortDescription" : "", - "shape" : "[1]", - "name" : "audio_length", - "type" : "MultiArray" - } - ], - "userDefinedMetadata" : { - "com.github.apple.coremltools.conversion_date" : "2025-09-25", - "com.github.apple.coremltools.source" : "torch==2.7.0", - "com.github.apple.coremltools.version" : "9.0b1", - "com.github.apple.coremltools.source_dialect" : "TorchScript" - }, - "generatedClassName" : "parakeet_preprocessor", - "method" : "predict" - } -] \ No newline at end of file diff --git a/parakeet-tdt-0.6b-v2-coreml/Preprocessor.mlmodelc/model.mil b/parakeet-tdt-0.6b-v2-coreml/Preprocessor.mlmodelc/model.mil deleted file mode 100644 index e94cce4bc7fbf55e55788449bf067ffe09a5ca91..0000000000000000000000000000000000000000 --- a/parakeet-tdt-0.6b-v2-coreml/Preprocessor.mlmodelc/model.mil +++ /dev/null @@ -1,169 +0,0 @@ -program(1.0) -[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3500.14.1"}, {"coremlc-version", "3500.32.1"}})] -{ - func main(tensor audio_length, tensor audio_signal) [FlexibleShapeInformation = tuple, dict, tensor>>, tuple, dict, list, ?>>>>((("DefaultShapes", {{"audio_signal", [1, 1]}}), ("RangeDims", {{"audio_signal", [[1, 1], [1, 240000]]}})))] { - tensor var_9 = const()[name = tensor("op_9"), val = tensor(1)]; - tensor var_10 = const()[name = tensor("op_10"), val = tensor(160)]; - tensor var_34 = const()[name = tensor("op_34"), val = tensor(512)]; - tensor var_35 = add(x = audio_length, y = var_34)[name = tensor("op_35")]; - tensor var_36 = const()[name = tensor("op_36"), val = tensor(512)]; - tensor var_37 = sub(x = var_35, y = var_36)[name = tensor("op_37")]; - tensor floor_div_0 = floor_div(x = var_37, y = var_10)[name = tensor("floor_div_0")]; - tensor var_38_to_fp16_dtype_0 = const()[name = tensor("op_38_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor var_39_promoted_to_fp16 = const()[name = tensor("op_39_promoted_to_fp16"), val = tensor(0x1p+0)]; - tensor floor_div_0_to_fp16 = cast(dtype = var_38_to_fp16_dtype_0, x = floor_div_0)[name = tensor("cast_9")]; - tensor seq_len_1_cast_fp16 = add(x = floor_div_0_to_fp16, y = var_39_promoted_to_fp16)[name = tensor("seq_len_1_cast_fp16")]; - tensor seq_len_dtype_0 = const()[name = tensor("seq_len_dtype_0"), val = tensor("int32")]; - tensor var_43_begin_0 = const()[name = tensor("op_43_begin_0"), val = tensor([0, 0])]; - tensor var_43_end_0 = const()[name = tensor("op_43_end_0"), val = tensor([1, 1])]; - tensor var_43_end_mask_0 = const()[name = tensor("op_43_end_mask_0"), val = tensor([true, false])]; - tensor var_43_squeeze_mask_0 = const()[name = tensor("op_43_squeeze_mask_0"), val = tensor([false, true])]; - tensor audio_signal_to_fp16_dtype_0 = const()[name = tensor("audio_signal_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor audio_signal_to_fp16 = cast(dtype = audio_signal_to_fp16_dtype_0, x = audio_signal)[name = tensor("cast_8")]; - tensor var_43_cast_fp16 = slice_by_index(begin = var_43_begin_0, end = var_43_end_0, end_mask = var_43_end_mask_0, squeeze_mask = var_43_squeeze_mask_0, x = audio_signal_to_fp16)[name = tensor("op_43_cast_fp16")]; - tensor var_44_axes_0 = const()[name = tensor("op_44_axes_0"), val = tensor([1])]; - tensor var_44_cast_fp16 = expand_dims(axes = var_44_axes_0, x = var_43_cast_fp16)[name = tensor("op_44_cast_fp16")]; - tensor var_46_begin_0 = const()[name = tensor("op_46_begin_0"), val = tensor([0, 1])]; - tensor var_46_end_0 = const()[name = tensor("op_46_end_0"), val = tensor([1, 0])]; - tensor var_46_end_mask_0 = const()[name = tensor("op_46_end_mask_0"), val = tensor([true, true])]; - tensor var_46_cast_fp16 = slice_by_index(begin = var_46_begin_0, end = var_46_end_0, end_mask = var_46_end_mask_0, x = audio_signal_to_fp16)[name = tensor("op_46_cast_fp16")]; - tensor var_48_begin_0 = const()[name = tensor("op_48_begin_0"), val = tensor([0, 0])]; - tensor var_48_end_0 = const()[name = tensor("op_48_end_0"), val = tensor([1, -1])]; - tensor var_48_end_mask_0 = const()[name = tensor("op_48_end_mask_0"), val = tensor([true, false])]; - tensor var_48_cast_fp16 = slice_by_index(begin = var_48_begin_0, end = var_48_end_0, end_mask = var_48_end_mask_0, x = audio_signal_to_fp16)[name = tensor("op_48_cast_fp16")]; - tensor var_49_to_fp16 = const()[name = tensor("op_49_to_fp16"), val = tensor(0x1.f0cp-1)]; - tensor var_50_cast_fp16 = mul(x = var_48_cast_fp16, y = var_49_to_fp16)[name = tensor("op_50_cast_fp16")]; - tensor var_51_cast_fp16 = sub(x = var_46_cast_fp16, y = var_50_cast_fp16)[name = tensor("op_51_cast_fp16")]; - tensor input_1_interleave_0 = const()[name = tensor("input_1_interleave_0"), val = tensor(false)]; - tensor input_1_cast_fp16 = concat(axis = var_9, interleave = input_1_interleave_0, values = (var_44_cast_fp16, var_51_cast_fp16))[name = tensor("input_1_cast_fp16")]; - tensor concat_0x = const()[name = tensor("concat_0x"), val = tensor([1, 1, -1])]; - tensor input_3_cast_fp16 = reshape(shape = concat_0x, x = input_1_cast_fp16)[name = tensor("input_3_cast_fp16")]; - tensor input_5_pad_0 = const()[name = tensor("input_5_pad_0"), val = tensor([0, 0, 0, 0, 256, 256])]; - tensor input_5_mode_0 = const()[name = tensor("input_5_mode_0"), val = tensor("reflect")]; - tensor const_1_to_fp16 = const()[name = tensor("const_1_to_fp16"), val = tensor(0x0p+0)]; - tensor input_5_cast_fp16 = pad(constant_val = const_1_to_fp16, mode = input_5_mode_0, pad = input_5_pad_0, x = input_3_cast_fp16)[name = tensor("input_5_cast_fp16")]; - tensor concat_1x = const()[name = tensor("concat_1x"), val = tensor([1, -1])]; - tensor input_cast_fp16 = reshape(shape = concat_1x, x = input_5_cast_fp16)[name = tensor("input_cast_fp16")]; - tensor expand_dims_3 = const()[name = tensor("expand_dims_3"), val = tensor([160])]; - tensor expand_dims_4_axes_0 = const()[name = tensor("expand_dims_4_axes_0"), val = tensor([1])]; - tensor expand_dims_4_cast_fp16 = expand_dims(axes = expand_dims_4_axes_0, x = input_cast_fp16)[name = tensor("expand_dims_4_cast_fp16")]; - tensor conv_0_pad_type_0 = const()[name = tensor("conv_0_pad_type_0"), val = tensor("valid")]; - tensor conv_0_pad_0 = const()[name = tensor("conv_0_pad_0"), val = tensor([0, 0])]; - tensor conv_0_dilations_0 = const()[name = tensor("conv_0_dilations_0"), val = tensor([1])]; - tensor conv_0_groups_0 = const()[name = tensor("conv_0_groups_0"), val = tensor(1)]; - tensor expand_dims_1_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("expand_dims_1_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132096))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(131712)))]; - tensor conv_0_cast_fp16 = conv(dilations = conv_0_dilations_0, groups = conv_0_groups_0, pad = conv_0_pad_0, pad_type = conv_0_pad_type_0, strides = expand_dims_3, weight = expand_dims_1_to_fp16_quantized, x = expand_dims_4_cast_fp16)[name = tensor("conv_0_cast_fp16")]; - tensor conv_1_pad_type_0 = const()[name = tensor("conv_1_pad_type_0"), val = tensor("valid")]; - tensor conv_1_pad_0 = const()[name = tensor("conv_1_pad_0"), val = tensor([0, 0])]; - tensor conv_1_dilations_0 = const()[name = tensor("conv_1_dilations_0"), val = tensor([1])]; - tensor conv_1_groups_0 = const()[name = tensor("conv_1_groups_0"), val = tensor(1)]; - tensor expand_dims_2_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("expand_dims_2_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132736))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(264768))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(264384)))]; - tensor conv_1_cast_fp16 = conv(dilations = conv_1_dilations_0, groups = conv_1_groups_0, pad = conv_1_pad_0, pad_type = conv_1_pad_type_0, strides = expand_dims_3, weight = expand_dims_2_to_fp16_quantized, x = expand_dims_4_cast_fp16)[name = tensor("conv_1_cast_fp16")]; - tensor stack_0_axis_0 = const()[name = tensor("stack_0_axis_0"), val = tensor(-1)]; - tensor stack_0_cast_fp16 = stack(axis = stack_0_axis_0, values = (conv_0_cast_fp16, conv_1_cast_fp16))[name = tensor("stack_0_cast_fp16")]; - tensor var_17_promoted_to_fp16 = const()[name = tensor("op_17_promoted_to_fp16"), val = tensor(0x1p+1)]; - tensor var_67_cast_fp16 = pow(x = stack_0_cast_fp16, y = var_17_promoted_to_fp16)[name = tensor("op_67_cast_fp16")]; - tensor var_69_axes_0 = const()[name = tensor("op_69_axes_0"), val = tensor([-1])]; - tensor var_69_keep_dims_0 = const()[name = tensor("op_69_keep_dims_0"), val = tensor(false)]; - tensor var_69_cast_fp16 = reduce_sum(axes = var_69_axes_0, keep_dims = var_69_keep_dims_0, x = var_67_cast_fp16)[name = tensor("op_69_cast_fp16")]; - tensor x_11_transpose_x_0 = const()[name = tensor("x_11_transpose_x_0"), val = tensor(false)]; - tensor x_11_transpose_y_0 = const()[name = tensor("x_11_transpose_y_0"), val = tensor(false)]; - tensor const_2_to_fp16_quantized = constexpr_affine_dequantize()[axis = tensor(1), name = tensor("const_2_to_fp16_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(265408))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(298560))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(298368)))]; - tensor x_11_cast_fp16 = matmul(transpose_x = x_11_transpose_x_0, transpose_y = x_11_transpose_y_0, x = const_2_to_fp16_quantized, y = var_69_cast_fp16)[name = tensor("x_11_cast_fp16")]; - tensor var_76_to_fp16 = const()[name = tensor("op_76_to_fp16"), val = tensor(0x1p-24)]; - tensor var_77_cast_fp16 = add(x = x_11_cast_fp16, y = var_76_to_fp16)[name = tensor("op_77_cast_fp16")]; - tensor x_13_epsilon_0 = const()[name = tensor("x_13_epsilon_0"), val = tensor(0x1p-149)]; - tensor x_13_cast_fp16 = log(epsilon = x_13_epsilon_0, x = var_77_cast_fp16)[name = tensor("x_13_cast_fp16")]; - tensor var_79_shape_cast_fp16 = shape(x = x_13_cast_fp16)[name = tensor("op_79_shape_cast_fp16")]; - tensor gather_4 = const()[name = tensor("gather_4"), val = tensor(1)]; - tensor gather_5_axis_0 = const()[name = tensor("gather_5_axis_0"), val = tensor(0)]; - tensor gather_5_batch_dims_0 = const()[name = tensor("gather_5_batch_dims_0"), val = tensor(0)]; - tensor gather_5_validate_indices_0 = const()[name = tensor("gather_5_validate_indices_0"), val = tensor(false)]; - tensor var_79_shape_cast_fp16_to_uint16_dtype_0 = const()[name = tensor("op_79_shape_cast_fp16_to_uint16_dtype_0"), val = tensor("uint16")]; - tensor gather_5_indices_0_to_uint16 = const()[name = tensor("gather_5_indices_0_to_uint16"), val = tensor(2)]; - tensor var_79_shape_cast_fp16_to_uint16 = cast(dtype = var_79_shape_cast_fp16_to_uint16_dtype_0, x = var_79_shape_cast_fp16)[name = tensor("cast_7")]; - tensor gather_5_cast_uint16 = gather(axis = gather_5_axis_0, batch_dims = gather_5_batch_dims_0, indices = gather_5_indices_0_to_uint16, validate_indices = gather_5_validate_indices_0, x = var_79_shape_cast_fp16_to_uint16)[name = tensor("gather_5_cast_uint16")]; - tensor gather_5_cast_uint16_to_int32_dtype_0 = const()[name = tensor("gather_5_cast_uint16_to_int32_dtype_0"), val = tensor("int32")]; - tensor const_3 = const()[name = tensor("const_3"), val = tensor(0)]; - tensor const_4 = const()[name = tensor("const_4"), val = tensor(1)]; - tensor gather_5_cast_uint16_to_int32 = cast(dtype = gather_5_cast_uint16_to_int32_dtype_0, x = gather_5_cast_uint16)[name = tensor("cast_6")]; - tensor var_81 = range_1d(end = gather_5_cast_uint16_to_int32, start = const_3, step = const_4)[name = tensor("op_81")]; - tensor var_82_axes_0 = const()[name = tensor("op_82_axes_0"), val = tensor([0])]; - tensor var_82 = expand_dims(axes = var_82_axes_0, x = var_81)[name = tensor("op_82")]; - tensor concat_2_axis_0 = const()[name = tensor("concat_2_axis_0"), val = tensor(0)]; - tensor concat_2_interleave_0 = const()[name = tensor("concat_2_interleave_0"), val = tensor(false)]; - tensor concat_2 = concat(axis = concat_2_axis_0, interleave = concat_2_interleave_0, values = (gather_4, gather_5_cast_uint16_to_int32))[name = tensor("concat_2")]; - tensor shape_0 = shape(x = var_82)[name = tensor("shape_0")]; - tensor real_div_0 = real_div(x = concat_2, y = shape_0)[name = tensor("real_div_0")]; - tensor time_steps = tile(reps = real_div_0, x = var_82)[name = tensor("time_steps")]; - tensor var_85_axes_0 = const()[name = tensor("op_85_axes_0"), val = tensor([1])]; - tensor mel_length = cast(dtype = seq_len_dtype_0, x = seq_len_1_cast_fp16)[name = tensor("cast_5")]; - tensor var_85 = expand_dims(axes = var_85_axes_0, x = mel_length)[name = tensor("op_85")]; - tensor valid_mask = less(x = time_steps, y = var_85)[name = tensor("valid_mask")]; - tensor var_87_axes_0 = const()[name = tensor("op_87_axes_0"), val = tensor([1])]; - tensor var_87 = expand_dims(axes = var_87_axes_0, x = valid_mask)[name = tensor("op_87")]; - tensor var_24_to_fp16 = const()[name = tensor("op_24_to_fp16"), val = tensor(0x0p+0)]; - tensor var_88_cast_fp16 = select(a = x_13_cast_fp16, b = var_24_to_fp16, cond = var_87)[name = tensor("op_88_cast_fp16")]; - tensor x_mean_numerator_axes_0 = const()[name = tensor("x_mean_numerator_axes_0"), val = tensor([2])]; - tensor x_mean_numerator_keep_dims_0 = const()[name = tensor("x_mean_numerator_keep_dims_0"), val = tensor(false)]; - tensor x_mean_numerator_cast_fp16 = reduce_sum(axes = x_mean_numerator_axes_0, keep_dims = x_mean_numerator_keep_dims_0, x = var_88_cast_fp16)[name = tensor("x_mean_numerator_cast_fp16")]; - tensor x_mean_denominator_axes_0 = const()[name = tensor("x_mean_denominator_axes_0"), val = tensor([1])]; - tensor x_mean_denominator_keep_dims_0 = const()[name = tensor("x_mean_denominator_keep_dims_0"), val = tensor(false)]; - tensor cast_3_to_fp16_dtype_0 = const()[name = tensor("cast_3_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor valid_mask_to_fp16 = cast(dtype = cast_3_to_fp16_dtype_0, x = valid_mask)[name = tensor("cast_4")]; - tensor x_mean_denominator_cast_fp16 = reduce_sum(axes = x_mean_denominator_axes_0, keep_dims = x_mean_denominator_keep_dims_0, x = valid_mask_to_fp16)[name = tensor("x_mean_denominator_cast_fp16")]; - tensor var_93_axes_0 = const()[name = tensor("op_93_axes_0"), val = tensor([1])]; - tensor var_93_cast_fp16 = expand_dims(axes = var_93_axes_0, x = x_mean_denominator_cast_fp16)[name = tensor("op_93_cast_fp16")]; - tensor x_mean_cast_fp16 = real_div(x = x_mean_numerator_cast_fp16, y = var_93_cast_fp16)[name = tensor("x_mean_cast_fp16")]; - tensor var_96_axes_0 = const()[name = tensor("op_96_axes_0"), val = tensor([2])]; - tensor var_96_cast_fp16 = expand_dims(axes = var_96_axes_0, x = x_mean_cast_fp16)[name = tensor("op_96_cast_fp16")]; - tensor var_97_cast_fp16 = sub(x = x_13_cast_fp16, y = var_96_cast_fp16)[name = tensor("op_97_cast_fp16")]; - tensor var_98_cast_fp16 = select(a = var_97_cast_fp16, b = var_24_to_fp16, cond = var_87)[name = tensor("op_98_cast_fp16")]; - tensor var_17_promoted_1_to_fp16 = const()[name = tensor("op_17_promoted_1_to_fp16"), val = tensor(0x1p+1)]; - tensor var_99_cast_fp16 = pow(x = var_98_cast_fp16, y = var_17_promoted_1_to_fp16)[name = tensor("op_99_cast_fp16")]; - tensor var_101_axes_0 = const()[name = tensor("op_101_axes_0"), val = tensor([2])]; - tensor var_101_keep_dims_0 = const()[name = tensor("op_101_keep_dims_0"), val = tensor(false)]; - tensor var_101_cast_fp16 = reduce_sum(axes = var_101_axes_0, keep_dims = var_101_keep_dims_0, x = var_99_cast_fp16)[name = tensor("op_101_cast_fp16")]; - tensor var_103_to_fp16 = const()[name = tensor("op_103_to_fp16"), val = tensor(0x1p+0)]; - tensor var_104_cast_fp16 = sub(x = var_93_cast_fp16, y = var_103_to_fp16)[name = tensor("op_104_cast_fp16")]; - tensor var_105_cast_fp16 = real_div(x = var_101_cast_fp16, y = var_104_cast_fp16)[name = tensor("op_105_cast_fp16")]; - tensor x_std_1_cast_fp16 = sqrt(x = var_105_cast_fp16)[name = tensor("x_std_1_cast_fp16")]; - tensor var_25_to_fp16 = const()[name = tensor("op_25_to_fp16"), val = tensor(0x1.5p-17)]; - tensor x_std_cast_fp16 = add(x = x_std_1_cast_fp16, y = var_25_to_fp16)[name = tensor("x_std_cast_fp16")]; - tensor var_110_axes_0 = const()[name = tensor("op_110_axes_0"), val = tensor([2])]; - tensor var_110_cast_fp16 = expand_dims(axes = var_110_axes_0, x = x_std_cast_fp16)[name = tensor("op_110_cast_fp16")]; - tensor x_cast_fp16 = real_div(x = var_97_cast_fp16, y = var_110_cast_fp16)[name = tensor("x_cast_fp16")]; - tensor var_112_shape_cast_fp16 = shape(x = x_cast_fp16)[name = tensor("op_112_shape_cast_fp16")]; - tensor gather_6_batch_dims_0 = const()[name = tensor("gather_6_batch_dims_0"), val = tensor(0)]; - tensor gather_6_validate_indices_0 = const()[name = tensor("gather_6_validate_indices_0"), val = tensor(false)]; - tensor var_112_shape_cast_fp16_to_uint16_dtype_0 = const()[name = tensor("op_112_shape_cast_fp16_to_uint16_dtype_0"), val = tensor("uint16")]; - tensor gather_6_cast_uint16_axis_0 = const()[name = tensor("gather_6_cast_uint16_axis_0"), val = tensor(0)]; - tensor select_0_to_uint16 = const()[name = tensor("select_0_to_uint16"), val = tensor(2)]; - tensor var_112_shape_cast_fp16_to_uint16 = cast(dtype = var_112_shape_cast_fp16_to_uint16_dtype_0, x = var_112_shape_cast_fp16)[name = tensor("cast_3")]; - tensor gather_6_cast_uint16_cast_uint16 = gather(axis = gather_6_cast_uint16_axis_0, batch_dims = gather_6_batch_dims_0, indices = select_0_to_uint16, validate_indices = gather_6_validate_indices_0, x = var_112_shape_cast_fp16_to_uint16)[name = tensor("gather_6_cast_uint16_cast_uint16")]; - tensor gather_6_cast_uint16_to_int32_dtype_0 = const()[name = tensor("gather_6_cast_uint16_to_int32_dtype_0"), val = tensor("int32")]; - tensor const_5 = const()[name = tensor("const_5"), val = tensor(0)]; - tensor const_6 = const()[name = tensor("const_6"), val = tensor(1)]; - tensor gather_6_cast_uint16_to_int32 = cast(dtype = gather_6_cast_uint16_to_int32_dtype_0, x = gather_6_cast_uint16_cast_uint16)[name = tensor("cast_2")]; - tensor mask_1 = range_1d(end = gather_6_cast_uint16_to_int32, start = const_5, step = const_6)[name = tensor("mask_1")]; - tensor gather_7_axis_0 = const()[name = tensor("gather_7_axis_0"), val = tensor(0)]; - tensor gather_7_batch_dims_0 = const()[name = tensor("gather_7_batch_dims_0"), val = tensor(0)]; - tensor gather_7_validate_indices_0 = const()[name = tensor("gather_7_validate_indices_0"), val = tensor(false)]; - tensor gather_7_indices_0_to_uint16 = const()[name = tensor("gather_7_indices_0_to_uint16"), val = tensor(0)]; - tensor gather_7_cast_uint16 = gather(axis = gather_7_axis_0, batch_dims = gather_7_batch_dims_0, indices = gather_7_indices_0_to_uint16, validate_indices = gather_7_validate_indices_0, x = var_112_shape_cast_fp16_to_uint16)[name = tensor("gather_7_cast_uint16")]; - tensor gather_7_cast_uint16_to_int32_dtype_0 = const()[name = tensor("gather_7_cast_uint16_to_int32_dtype_0"), val = tensor("int32")]; - tensor concat_3_axis_0 = const()[name = tensor("concat_3_axis_0"), val = tensor(0)]; - tensor concat_3_interleave_0 = const()[name = tensor("concat_3_interleave_0"), val = tensor(false)]; - tensor gather_7_cast_uint16_to_int32 = cast(dtype = gather_7_cast_uint16_to_int32_dtype_0, x = gather_7_cast_uint16)[name = tensor("cast_1")]; - tensor concat_3 = concat(axis = concat_3_axis_0, interleave = concat_3_interleave_0, values = (gather_7_cast_uint16_to_int32, var_9))[name = tensor("concat_3")]; - tensor expand_dims_0_axes_0 = const()[name = tensor("expand_dims_0_axes_0"), val = tensor([0])]; - tensor expand_dims_0 = expand_dims(axes = expand_dims_0_axes_0, x = mask_1)[name = tensor("expand_dims_0")]; - tensor var_116 = tile(reps = concat_3, x = expand_dims_0)[name = tensor("op_116")]; - tensor mask = greater_equal(x = var_116, y = var_85)[name = tensor("mask")]; - tensor var_119_axes_0 = const()[name = tensor("op_119_axes_0"), val = tensor([1])]; - tensor var_119 = expand_dims(axes = var_119_axes_0, x = mask)[name = tensor("op_119")]; - tensor processed_signal_cast_fp16 = select(a = var_24_to_fp16, b = x_cast_fp16, cond = var_119)[name = tensor("processed_signal_cast_fp16")]; - tensor processed_signal_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("processed_signal_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; - tensor mel = cast(dtype = processed_signal_cast_fp16_to_fp32_dtype_0, x = processed_signal_cast_fp16)[name = tensor("cast_0")]; - } -> (mel, mel_length); -} \ No newline at end of file diff --git a/parakeet-tdt-0.6b-v2-coreml/Preprocessor.mlmodelc/weights/weight.bin b/parakeet-tdt-0.6b-v2-coreml/Preprocessor.mlmodelc/weights/weight.bin deleted file mode 100644 index a9af1ac1e407ef336eebc2fd4da9de57e6d172ad..0000000000000000000000000000000000000000 --- a/parakeet-tdt-0.6b-v2-coreml/Preprocessor.mlmodelc/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:a5f7df6c7f47147ae9486fe18cc7792f9a44d093ec3c6a11e91ef2dc363c48dc -size 298880 diff --git a/parakeet-tdt-0.6b-v2-coreml/config.json b/parakeet-tdt-0.6b-v2-coreml/config.json deleted file mode 100644 index 0967ef424bce6791893e9a57bb952f80fd536e93..0000000000000000000000000000000000000000 --- a/parakeet-tdt-0.6b-v2-coreml/config.json +++ /dev/null @@ -1 +0,0 @@ -{} diff --git a/parakeet-tdt-0.6b-v2-coreml/parakeet_vocab.json b/parakeet-tdt-0.6b-v2-coreml/parakeet_vocab.json deleted file mode 100644 index a111d3bd29d44781c4588447c3eac1ea549cb1f1..0000000000000000000000000000000000000000 --- a/parakeet-tdt-0.6b-v2-coreml/parakeet_vocab.json +++ /dev/null @@ -1,1033 +0,0 @@ -{ - "479" : "▁happ", - "493" : "ial", - "386" : "▁diff", - "398" : "▁little", - "168" : "▁ch", - "872" : "2", - "227" : "ie", - "251" : "▁R", - "488" : "▁day", - "38" : "ed", - "610" : "▁might", - "127" : "▁can", - "70" : "ct", - "903" : "Ă ", - "329" : "▁been", - "694" : "ren", - "1005" : "Ćœ", - "661" : "▁reg", - "597" : "▁around", - "442" : "ys", - "764" : "▁number", - "705" : "▁trans", - "275" : "▁had", - "271" : "ak", - "1024" : "▁warming", - "564" : "▁play", - "32" : "▁y", - "463" : "▁under", - "684" : "ute", - "529" : "▁quest", - "719" : "ock", - "235" : "her", - "861" : "q", - "987" : "Č", - "593" : "urn", - "555" : "ath", - "216" : "▁think", - "211" : "el", - "500" : "▁bit", - "353" : "ry", - "490" : "▁Yeah", - "750" : "▁sim", - "283" : "▁these", - "320" : "▁sp", - "945" : "ÎČ", - "179" : "ca", - "934" : "т", - "258" : "ear", - "870" : "R", - "974" : "Ί", - "157" : "▁H", - "280" : "ine", - "381" : "▁well", - "21" : "▁m", - "668" : "▁la", - "429" : "▁bo", - "812" : "▁ty", - "537" : "ual", - "743" : "ss", - "119" : "▁kn", - "863" : "P", - "855" : "0", - "333" : "ple", - "850" : "x", - "959" : "я", - "467" : "▁bu", - "262" : "▁would", - "835" : "g", - "254" : "▁who", - "233" : "▁here", - "993" : "ĆŒ", - "889" : "Z", - "231" : "ack", - "788" : "ution", - "466" : "ning", - "332" : "▁did", - "388" : "cc", - "260" : "▁E", - "839" : ",", - "582" : "▁three", - "1000" : "Ă”", - "73" : "ver", - "486" : "ign", - "541" : "▁ser", - "800" : "▁exper", - "192" : "um", - "199" : "▁from", - "223" : "▁because", - "68" : "ent", - "608" : "▁tra", - "383" : "▁pre", - "774" : "▁place", - "194" : "▁your", - "148" : "ra", - "955" : "Ü", - "158" : "▁fr", - "670" : "▁sub", - "906" : "π", - "829" : "l", - "202" : "pe", - "948" : "ÎŽ", - "480" : "ater", - "342" : "ree", - "958" : "ƛ", - "967" : "Á", - "171" : "ate", - "91" : "am", - "114" : "ation", - "390" : "int", - "1013" : "À", - "794" : "▁read", - "1011" : "ф", - "355" : "ble", - "921" : "Ć«", - "234" : "▁pe", - "573" : "▁id", - "785" : "gan", - "335" : "▁other", - "963" : "Ö", - "570" : "get", - "2" : "▁th", - "273" : "ide", - "574" : "▁Oh", - "807" : "▁open", - "155" : "▁The", - "506" : "vers", - "729" : "▁sort", - "318" : "▁got", - "550" : "ank", - "896" : "Ăł", - "563" : "ict", - "1010" : "Đ·", - "520" : "▁gonna", - "1020" : "Κ", - "548" : "ved", - "848" : "S", - "578" : "▁rel", - "828" : "r", - "361" : "▁cont", - "379" : "ish", - "259" : "▁tim", - "441" : "▁imp", - "44" : "as", - "620" : "ways", - "34" : "▁I", - "87" : "st", - "13" : "nd", - "350" : "▁every", - "866" : "D", - "545" : "▁V", - "182" : "if", - "615" : "cial", - "69" : "ke", - "276" : "so", - "732" : "ics", - "749" : "als", - "605" : "aking", - "857" : "-", - "446" : "▁down", - "41" : "ar", - "203" : "un", - "816" : "▁met", - "543" : "▁ind", - "926" : "ĂŠ", - "64" : "▁on", - "246" : "▁them", - "172" : "qu", - "229" : "▁tr", - "775" : "▁gener", - "592" : "ower", - "617" : "▁give", - "399" : "ical", - "319" : "ag", - "669" : "▁last", - "400" : "▁gr", - "389" : "ittle", - "533" : "▁fu", - "460" : "uct", - "825" : "i", - "992" : "Ç", - "432" : "▁much", - "638" : "ction", - "700" : "▁love", - "988" : "έ", - "658" : "▁inv", - "549" : "▁still", - "322" : "act", - "143" : "il", - "966" : "ĐČ", - "245" : "▁It", - "445" : "▁yeah", - "524" : "self", - "742" : "ata", - "380" : "kay", - "176" : "▁su", - "624" : "▁gen", - "766" : "▁saying", - "42" : "▁that", - "136" : "▁ab", - "224" : "▁by", - "811" : "▁getting", - "76" : "id", - "430" : "be", - "124" : "▁know", - "88" : "ch", - "420" : "▁she", - "822" : "o", - "991" : "Å", - "623" : "ily", - "230" : "▁then", - "65" : "▁T", - "86" : "▁st", - "873" : "J", - "464" : "▁br", - "165" : "pp", - "325" : "iz", - "237" : "▁F", - "939" : "č", - "250" : "▁say", - "423" : "ord", - "516" : "▁fl", - "1030" : "▁urge", - "977" : "Îč", - "358" : "▁kind", - "471" : "co", - "7" : "▁w", - "252" : "▁people", - "12" : "er", - "932" : "ψ", - "632" : "▁cour", - "440" : "▁comm", - "664" : "▁cr", - "637" : "▁num", - "588" : "▁To", - "286" : "▁sa", - "292" : "pt", - "772" : "▁better", - "396" : "▁him", - "1007" : "Γ", - "26" : "▁d", - "450" : "ens", - "780" : "cept", - "936" : "ω", - "123" : "▁And", - "349" : "▁need", - "536" : "we", - "882" : "8", - "546" : "he", - "854" : "?", - "875" : ":", - "712" : "▁exam", - "437" : "ary", - "394" : "ip", - "344" : "own", - "709" : "▁fin", - "457" : "▁K", - "996" : "χ", - "990" : "ы", - "89" : "▁li", - "878" : "3", - "503" : "fe", - "61" : "et", - "782" : "▁understand", - "126" : "op", - "135" : "▁at", - "831" : "u", - "456" : "ody", - "594" : "▁okay", - "695" : "erest", - "324" : "▁also", - "51" : "▁it", - "542" : "▁rem", - "174" : "▁ex", - "1006" : "ț", - "128" : "▁or", - "357" : "ue", - "641" : "▁um", - "802" : "▁ele", - "213" : "▁some", - "606" : "▁pos", - "929" : "Μ", - "885" : "!", - "784" : "▁thought", - "716" : "▁stud", - "232" : "▁pl", - "522" : "ces", - "184" : "▁if", - "37" : "is", - "867" : "N", - "653" : "uc", - "413" : "ult", - "167" : "ess", - "22" : "en", - "439" : "ving", - "104" : "▁r", - "755" : "oc", - "63" : "▁re", - "622" : "ward", - "1003" : "Ɓ", - "468" : "▁use", - "879" : "K", - "838" : "p", - "11" : "ou", - "54" : "le", - "118" : "▁not", - "558" : "ft", - "858" : "M", - "579" : "▁before", - "652" : "les", - "98" : "▁do", - "1026" : "▁issue", - "360" : "▁back", - "790" : "ason", - "731" : "▁today", - "706" : "▁count", - "650" : "▁didn", - "348" : "▁where", - "371" : "ep", - "898" : "ĂŒ", - "384" : "▁two", - "121" : "▁B", - "720" : "▁used", - "120" : "ight", - "1022" : "Ο", - "345" : "▁tw", - "307" : "▁work", - "655" : "ating", - "745" : "ween", - "596" : "▁bel", - "852" : "B", - "986" : "ĂŹ", - "197" : "▁get", - "79" : "▁he", - "504" : "▁doing", - "644" : "▁own", - "791" : "▁problem", - "931" : "Îł", - "33" : "▁l", - "187" : "ab", - "771" : "▁between", - "783" : "▁fun", - "823" : "a", - "723" : "▁No", - "303" : "▁mo", - "482" : "ition", - "164" : "▁M", - "346" : "▁part", - "103" : "ad", - "902" : "ç", - "781" : "ull", - "411" : "▁actually", - "598" : "ful", - "1002" : "ħ", - "351" : "pl", - "343" : "▁into", - "908" : "Ăș", - "337" : "ite", - "894" : "ĂĄ", - "912" : "Ä«", - "864" : "z", - "925" : "τ", - "153" : "▁con", - "144" : "▁but", - "433" : "▁per", - "711" : "▁pol", - "631" : "▁sm", - "431" : "ount", - "999" : "Í", - "946" : "Ăž", - "141" : "use", - "77" : "▁for", - "733" : "▁vide", - "730" : "▁For", - "418" : "ations", - "693" : "▁always", - "642" : "ood", - "1015" : "Ā", - "154" : "▁all", - "659" : "ably", - "1027" : "▁stay", - "777" : "▁ins", - "740" : "▁keep", - "96" : "▁so", - "806" : "▁partic", - "313" : "▁im", - "507" : "av", - "880" : "4", - "792" : "▁doesn", - "426" : "▁am", - "1001" : "ě", - "526" : "▁If", - "465" : "▁take", - "436" : "vel", - "116" : "ere", - "576" : "ever", - "568" : "oth", - "166" : "▁com", - "151" : "ul", - "798" : "ah", - "779" : "cond", - "556" : "▁end", - "161" : "ea", - "39" : "▁g", - "765" : "ention", - "100" : "th", - "461" : "▁only", - "590" : "▁hel", - "505" : "▁St", - "957" : "Ƅ", - "876" : "5", - "580" : "▁feel", - "321" : "ans", - "972" : "ь", - "718" : "▁car", - "851" : "W", - "960" : "đ", - "469" : "▁Ch", - "291" : "one", - "83" : "ly", - "295" : "▁has", - "84" : "▁go", - "981" : "Ƒ", - "923" : "λ", - "52" : "▁be", - "821" : "t", - "323" : "▁te", - "50" : "al", - "760" : "ense", - "149" : "ore", - "306" : "▁le", - "403" : "▁thr", - "628" : "ob", - "299" : "▁look", - "406" : "▁This", - "0" : "", - "75" : "all", - "475" : "▁call", - "341" : "reat", - "519" : "ents", - "66" : "▁A", - "186" : "nt", - "928" : "ĐŸ", - "244" : "ople", - "393" : "ence", - "834" : "m", - "833" : "y", - "414" : "able", - "297" : "▁very", - "364" : "▁pr", - "865" : "L", - "35" : "it", - "339" : "omet", - "27" : "es", - "150" : "▁there", - "715" : "ell", - "677" : "▁import", - "681" : "▁ear", - "820" : "e", - "139" : "▁So", - "449" : "na", - "302" : "▁time", - "532" : "▁What", - "133" : "ck", - "970" : "Îż", - "773" : "cus", - "228" : "▁us", - "552" : "▁wr", - "680" : "▁made", - "871" : "E", - "874" : "U", - "895" : "ÂŁ", - "846" : "T", - "55" : "ion", - "492" : "ile", - "787" : "cy", - "105" : "ir", - "662" : "lic", - "629" : "▁tell", - "367" : "▁good", - "514" : "form", - "656" : "olog", - "334" : "ually", - "294" : "ong", - "485" : "ade", - "682" : "▁ac", - "106" : "▁was", - "1028" : "▁together", - "910" : "ĂŁ", - "142" : "ter", - "922" : "Δ", - "175" : "very", - "314" : "▁ag", - "327" : "▁That", - "826" : "s", - "183" : "ive", - "979" : "Đł", - "751" : "vern", - "997" : "э", - "278" : "eah", - "517" : "fter", - "311" : "per", - "535" : "▁show", - "918" : "^", - "954" : "ĆĄ", - "722" : "stand", - "6" : "re", - "93" : "ce", - "435" : "▁differe", - "746" : "▁stuff", - "915" : "ρ", - "602" : "▁supp", - "209" : "▁L", - "767" : "▁commun", - "769" : "akes", - "375" : "▁lot", - "859" : "H", - "397" : "▁make", - "340" : "ber", - "886" : "%", - "736" : "▁Al", - "600" : "ise", - "938" : "ć", - "478" : "ting", - "962" : "ĐŒ", - "138" : "ol", - "125" : "ome", - "309" : "are", - "673" : "▁inst", - "97" : "▁have", - "562" : "ject", - "678" : "ific", - "257" : "ect", - "17" : "on", - "953" : "с", - "933" : "ē", - "844" : "'", - "949" : "η", - "190" : "▁v", - "980" : "Đș", - "452" : "▁fo", - "247" : "ame", - "612" : "▁help", - "501" : "▁spe", - "604" : "ange", - "654" : "ib", - "842" : "k", - "815" : "ave", - "687" : "▁form", - "222" : "res", - "421" : "sel", - "477" : "other", - "308" : "▁their", - "212" : "▁N", - "737" : "▁important", - "90" : "▁u", - "630" : "▁Now", - "425" : "ia", - "14" : "▁i", - "315" : "▁J", - "331" : "▁fe", - "304" : "▁ar", - "29" : "ll", - "499" : "▁sc", - "919" : "€", - "301" : "itt", - "201" : "ri", - "137" : "ould", - "289" : "▁man", - "81" : "▁this", - "1008" : "П", - "312" : "ions", - "725" : "ks", - "837" : "f", - "458" : "▁through", - "714" : "▁maybe", - "487" : "thing", - "424" : "▁may", - "31" : "▁and", - "78" : "ro", - "961" : "Đ»", - "956" : "Ă„", - "198" : "cause", - "95" : "im", - "899" : "ñ", - "111" : "ally", - "523" : "▁There", - "538" : "ons", - "797" : "▁el", - "726" : "▁interest", - "53" : "▁wh", - "296" : "▁any", - "489" : "fore", - "911" : "φ", - "248" : "▁We", - "639" : "▁add", - "108" : "▁W", - "521" : "▁point", - "416" : "▁dis", - "739" : "▁run", - "747" : "ract", - "3" : "▁a", - "530" : "▁most", - "734" : "▁bec", - "338" : "age", - "544" : "▁pers", - "113" : "▁se", - "691" : "▁able", - "847" : "A", - "651" : "stem", - "115" : "od", - "527" : "▁same", - "575" : "ves", - "696" : "▁As", - "277" : "▁qu", - "728" : "ited", - "640" : "▁set", - "502" : "ub", - "497" : "▁try", - "881" : "V", - "219" : "▁G", - "561" : "ph", - "759" : "▁All", - "177" : "ain", - "515" : "ors", - "71" : "▁S", - "509" : "ian", - "803" : "▁cou", - "25" : "an", - "583" : "iss", - "417" : "▁first", - "840" : "b", - "768" : "▁An", - "419" : "▁something", - "569" : "▁acc", - "607" : "atch", - "534" : "ug", - "195" : "▁my", - "832" : "c", - "634" : "cess", - "809" : "▁everything", - "1" : "▁t", - "557" : "▁bas", - "481" : "▁inc", - "57" : "ot", - "518" : "ail", - "924" : "α", - "173" : "▁lo", - "905" : "ÎŒ", - "762" : "▁probably", - "626" : "▁dec", - "647" : "▁its", - "415" : "orm", - "917" : "ĂŽ", - "560" : "body", - "474" : "▁put", - "572" : "▁em", - "689" : "▁system", - "909" : "Ξ", - "408" : "▁res", - "862" : "1", - "830" : "d", - "748" : "▁question", - "285" : "▁now", - "717" : "▁prod", - "60" : "▁e", - "818" : "oney", - "814" : "▁Because", - "893" : "Ă­", - "587" : "▁uh", - "377" : "▁things", - "454" : "▁ro", - "205" : "▁up", - "849" : "j", - "152" : "out", - "994" : "ÎŻ", - "789" : "ope", - "1018" : "Ćș", - "973" : "ĆŸ", - "950" : "Đż", - "901" : "Ăš", - "288" : "▁Wh", - "710" : "▁prob", - "581" : "igh", - "1009" : "ĐŽ", - "47" : "us", - "756" : "ness", - "648" : "▁God", - "43" : "om", - "952" : "Đœ", - "744" : "▁never", - "688" : "▁guys", - "272" : "▁co", - "36" : "▁in", - "405" : "ated", - "741" : "▁fact", - "56" : "ut", - "290" : "ous", - "770" : "▁belie", - "943" : "ĂČ", - "284" : "▁how", - "697" : "▁mod", - "671" : "▁att", - "453" : "▁comp", - "690" : "ew", - "101" : "▁an", - "264" : "ven", - "200" : "▁don", - "279" : "▁were", - "99" : "ht", - "305" : "hing", - "1014" : "Î", - "525" : "▁many", - "738" : "▁such", - "940" : "Δ", - "995" : "ζ", - "491" : "ark", - "18" : "▁h", - "947" : "Ƃ", - "702" : "▁ask", - "49" : "ow", - "808" : "▁gl", - "484" : "▁should", - "907" : "Ă€", - "916" : "Ăą", - "447" : "ang", - "107" : "▁as", - "843" : "v", - "1025" : "▁global", - "267" : "▁really", - "892" : "\/", - "703" : "old", - "679" : "ix", - "601" : "▁ob", - "498" : "ious", - "298" : "▁But", - "300" : "iv", - "565" : "▁Is", - "282" : "ther", - "249" : "our", - "539" : "▁Be", - "356" : "ap", - "528" : "▁sy", - "470" : "xt", - "373" : "ick", - "853" : "C", - "215" : "and", - "869" : "F", - "10" : "at", - "1004" : "Ɠ", - "686" : "ative", - "553" : "ought", - "976" : "ę", - "473" : "ild", - "4" : "in", - "352" : "▁ad", - "951" : "Ă«", - "265" : "▁our", - "427" : "▁her", - "676" : "▁rep", - "982" : "Ú", - "243" : "ies", - "48" : "ic", - "978" : "б", - "28" : "or", - "566" : "ates", - "965" : "ș", - "9" : "▁s", - "92" : "ur", - "16" : "▁c", - "30" : "▁of", - "15" : "▁b", - "547" : "▁str", - "645" : "▁life", - "24" : "▁p", - "310" : "▁his", - "472" : "ory", - "189" : "▁going", - "699" : "ings", - "1029" : "▁bipartisan", - "178" : "▁one", - "67" : "▁ha", - "824" : "n", - "422" : "▁let", - "888" : "$", - "649" : "pect", - "635" : "nds", - "540" : "ically", - "511" : "red", - "827" : "h", - "19" : "ing", - "969" : "Ăč", - "985" : "υ", - "225" : "ake", - "239" : "ard", - "618" : "ike", - "240" : "▁right", - "585" : "ne", - "378" : "▁In", - "616" : "▁world", - "130" : "▁me", - "434" : "▁even", - "599" : "te", - "897" : "ā", - "206" : "▁P", - "336" : "▁U", - "82" : "ld", - "877" : "9", - "368" : "▁than", - "1016" : "ė", - "181" : "ist", - "438" : "▁app", - "611" : "ert", - "567" : "▁ph", - "374" : "way", - "395" : "ase", - "621" : "▁min", - "188" : "▁about", - "663" : "▁stu", - "571" : "▁years", - "94" : "ith", - "758" : "ize", - "180" : "art", - "40" : "▁you", - "904" : "Âż", - "845" : "I", - "268" : "▁more", - "253" : "▁see", - "683" : "▁def", - "856" : "O", - "370" : "▁gu", - "551" : "▁rec", - "8" : "▁o", - "428" : "▁said", - "675" : "▁happen", - "102" : "▁with", - "776" : "▁ca", - "363" : "▁somet", - "757" : "arch", - "914" : "ĂȘ", - "208" : "ort", - "660" : "▁sure", - "207" : "▁out", - "613" : "ost", - "942" : "Đž", - "761" : "blem", - "117" : "▁like", - "129" : "▁sh", - "293" : "ff", - "407" : "▁off", - "595" : "▁long", - "366" : "ire", - "614" : "▁too", - "256" : "ure", - "328" : "▁cl", - "793" : "ational", - "724" : "▁mon", - "989" : "х", - "134" : "▁what", - "45" : "▁n", - "707" : "ility", - "62" : "ay", - "801" : "▁four", - "392" : "▁those", - "160" : "ge", - "369" : "ace", - "236" : "▁will", - "80" : "se", - "708" : "▁high", - "217" : "em", - "944" : "р", - "672" : "▁op", - "754" : "▁Of", - "58" : "▁we", - "131" : "ill", - "810" : "▁eff", - "146" : "▁ne", - "998" : "Æ", - "496" : "▁being", - "975" : "у", - "625" : "▁find", - "132" : "ant", - "884" : "7", - "382" : "▁could", - "459" : "▁start", - "147" : "▁de", - "448" : "▁mean", - "1019" : "Κ", - "241" : "▁thing", - "819" : "▁", - "698" : "▁done", - "444" : "ress", - "46" : "ve", - "971" : "ч", - "646" : "ities", - "713" : "▁pres", - "483" : "▁different", - "577" : "▁inter", - "372" : "og", - "786" : "iew", - "692" : "ied", - "362" : "iff", - "1021" : "ÎŹ", - "586" : "▁why", - "636" : "▁big", - "1023" : "ό", - "162" : "▁Y", - "685" : "▁next", - "193" : "ok", - "221" : "▁D", - "665" : "▁ev", - "159" : "▁pro", - "59" : "▁is", - "1017" : "Ć ", - "74" : "▁Th", - "813" : "▁Am", - "5" : "▁the", - "218" : "oug", - "376" : "▁un", - "270" : "ose", - "140" : "▁C", - "122" : "▁they", - "412" : "▁talk", - "913" : "σ", - "326" : "ice", - "451" : "▁does", - "559" : "erm", - "23" : "▁f", - "210" : "ment", - "778" : "▁ass", - "984" : "Îș", - "316" : "▁no", - "666" : "ments", - "494" : "▁come", - "887" : "Q", - "891" : "Ă©", - "404" : "uch", - "920" : "É", - "455" : "▁bl", - "170" : "est", - "72" : "ig", - "443" : "▁again", - "727" : "▁ent", - "983" : "Ω", - "410" : "ance", - "281" : "▁act", - "763" : "hip", - "220" : "os", - "261" : "▁You", - "214" : "ich", - "930" : "Ăź", - "805" : "▁called", - "156" : "ers", - "753" : "▁course", - "868" : "G", - "238" : "▁which", - "836" : "w", - "704" : "ered", - "513" : "▁ke", - "964" : "Ă»", - "589" : "▁cons", - "409" : "ac", - "609" : "gr", - "476" : "▁new", - "619" : "▁Okay", - "287" : "ud", - "603" : "ady", - "633" : "▁real", - "508" : "ty", - "1012" : "ÂĄ", - "401" : "▁year", - "110" : "ust", - "196" : "ind", - "242" : "▁want", - "817" : "▁Like", - "890" : "X", - "191" : "▁wor", - "402" : "ass", - "266" : "ci", - "347" : "alk", - "387" : "ach", - "804" : "ont", - "701" : "ism", - "841" : ".", - "317" : "▁en", - "385" : "irst", - "255" : "ast", - "510" : "onna", - "20" : "▁to", - "667" : "▁another", - "591" : "▁after", - "354" : "▁over", - "112" : "▁j", - "674" : "▁sl", - "935" : "ß", - "204" : "ity", - "735" : "▁Well", - "900" : "ö", - "163" : "▁O", - "169" : "▁al", - "531" : "▁great", - "185" : "ink", - "752" : "ather", - "795" : "▁trying", - "796" : "▁sch", - "721" : "oy", - "657" : "▁person", - "799" : "atter", - "554" : "day", - "274" : "ough", - "109" : "▁are", - "643" : "ible", - "968" : "Ø", - "269" : "ound", - "145" : "▁just", - "263" : "▁when", - "85" : "▁k", - "365" : "nder", - "359" : "▁po", - "937" : "ĂŻ", - "226" : "▁int", - "584" : "▁des", - "512" : "wn", - "495" : "▁They", - "391" : "▁He", - "883" : "6", - "627" : "ular", - "927" : "а", - "941" : "Đ”", - "330" : "▁way", - "860" : "Y", - "462" : "▁bet" -} \ No newline at end of file diff --git a/parakeet-tdt-ctc-110m/convert_tdt_decoder.py b/parakeet-tdt-ctc-110m/convert_tdt_decoder.py deleted file mode 100644 index 4ed15a38318bb788204bfdbd2bc466cff9ff6e5c..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/convert_tdt_decoder.py +++ /dev/null @@ -1,323 +0,0 @@ -#!/usr/bin/env python3 -""" -Convert Parakeet TDT-CTC 110M decoder components to CoreML. - -This script exports the TDT decoder (prediction network) and joint network -with the SAME format as the working 0.6B model: -- JointDecision outputs token_id, token_prob, duration (argmax done inside) -- Uses shape [1, dim, 1] for encoder/decoder steps -- Matches the interface expected by TdtDecoderV3 -""" - -import argparse -import os -import torch -import torch.nn.functional as F -import coremltools as ct -import numpy as np -from pathlib import Path - -# NeMo imports -import nemo.collections.asr as nemo_asr - - -def get_model_config(model): - """Extract model configuration.""" - encoder_dim = None - pred_hidden = 640 # Default for parakeet models - num_layers = 1 - vocab_size = 1024 - num_durations = 5 - - # Get encoder dimension - if hasattr(model, 'encoder'): - encoder = model.encoder - if hasattr(encoder, 'd_model'): - encoder_dim = encoder.d_model - elif hasattr(encoder, '_feat_out'): - encoder_dim = encoder._feat_out - - # Get decoder config - if hasattr(model, 'decoder'): - decoder = model.decoder - if hasattr(decoder, 'pred_hidden'): - pred_hidden = decoder.pred_hidden - if hasattr(decoder, 'pred_rnn_layers'): - num_layers = decoder.pred_rnn_layers - - # Get joint config - if hasattr(model, 'joint'): - joint = model.joint - if hasattr(joint, 'num_extra_outputs'): - num_durations = joint.num_extra_outputs - if hasattr(joint, 'num_classes'): - vocab_size = joint.num_classes - num_durations - - return { - 'encoder_dim': encoder_dim, - 'pred_hidden': pred_hidden, - 'num_layers': num_layers, - 'vocab_size': vocab_size, - 'num_durations': num_durations, - } - - -class DecoderWrapper(torch.nn.Module): - """ - Wrapper for the RNNT/TDT decoder (prediction network). - - Matches 0.6B format: - - Input: targets[1,1], target_lengths[1], h_in[num_layers,1,pred_hidden], c_in[...] - - Output: decoder_output[1,pred_hidden,2], h_out[...], c_out[...] - """ - - def __init__(self, decoder, pred_hidden): - super().__init__() - self.decoder = decoder - self.pred_hidden = pred_hidden - - def forward(self, targets, target_lengths, h_in, c_in): - """ - Args: - targets: [1, 1] - previous token ID - target_lengths: [1] - always 1 - h_in: [num_layers, 1, pred_hidden] - c_in: [num_layers, 1, pred_hidden] - Returns: - decoder_output: [1, pred_hidden, 2] - prediction network output (transposed) - h_out: [num_layers, 1, pred_hidden] - c_out: [num_layers, 1, pred_hidden] - """ - state = (h_in, c_in) - # pred_output shape: [batch, time, pred_hidden] = [1, 1, pred_hidden] - pred_output, new_state = self.decoder.predict(targets, state=state, add_sos=False) - h_out, c_out = new_state - - # Transpose to [batch, pred_hidden, time] and concat two time steps - # (0.6B outputs [1, 640, 2] - we match this by duplicating) - pred_transposed = pred_output.transpose(1, 2) # [1, pred_hidden, 1] - decoder_output = torch.cat([pred_transposed, pred_transposed], dim=2) # [1, pred_hidden, 2] - - return decoder_output, h_out, c_out - - -class JointWrapper(torch.nn.Module): - """ - Wrapper for the TDT joint network with internal argmax. - - Matches 0.6B format: - - Input: encoder_step[1,encoder_dim,1], decoder_step[1,pred_hidden,1] - - Output: token_id[1,1,1], token_prob[1,1,1], duration[1,1,1] - """ - - def __init__(self, joint, vocab_size, num_durations=5): - super().__init__() - self.joint = joint - self.vocab_size = vocab_size - self.num_durations = num_durations - - def forward(self, encoder_step, decoder_step): - """ - Args: - encoder_step: [1, encoder_dim, 1] - decoder_step: [1, pred_hidden, 1] - Returns: - token_id: [1, 1, 1] - argmax token ID - token_prob: [1, 1, 1] - probability of selected token - duration: [1, 1, 1] - argmax duration bin - """ - # Transpose to [batch, 1, dim] for joint network - enc = encoder_step.transpose(1, 2) # [1, 1, encoder_dim] - dec = decoder_step.transpose(1, 2) # [1, 1, pred_hidden] - - # Run joint network - # Joint output: [1, 1, 1, vocab_size + 1 (blank) + num_durations] - joint_out = self.joint.joint(enc, dec) - - # Debug: print shape on first call - if not hasattr(self, '_debug_printed'): - self._debug_printed = True - print(f" Joint output shape: {joint_out.shape}") - print(f" Expected: vocab={self.vocab_size} + blank=1 + durations={self.num_durations} = {self.vocab_size + 1 + self.num_durations}") - - # Split: token logits include vocab + blank, durations are separate - # vocab_size = 1024 tokens (0-1023), blank = index 1024, durations = indices 1025+ - num_tokens = self.vocab_size + 1 # Include blank at vocab_size - logits = joint_out[..., :num_tokens] # [1, 1, 1, vocab_size + 1] - duration_logits = joint_out[..., num_tokens:] # [1, 1, 1, num_durations] - - # Apply softmax and get probabilities - probs = F.softmax(logits, dim=-1) - - # Argmax for token - token_id = torch.argmax(logits, dim=-1, keepdim=True) # [1, 1, 1, 1] - token_id = token_id.squeeze(-1) # [1, 1, 1] - - # Get probability of selected token - token_prob = torch.gather(probs, -1, token_id.unsqueeze(-1)) # [1, 1, 1, 1] - token_prob = token_prob.squeeze(-1) # [1, 1, 1] - - # Argmax for duration - duration = torch.argmax(duration_logits, dim=-1, keepdim=False) # [1, 1, 1] - - return token_id.int(), token_prob, duration.int() - - -def convert_decoder(model, config, output_dir: Path): - """Convert decoder to CoreML.""" - print(f"Converting Decoder...") - print(f" pred_hidden={config['pred_hidden']}, num_layers={config['num_layers']}") - - wrapper = DecoderWrapper(model.decoder, config['pred_hidden']) - wrapper.eval() - - # Create example inputs - targets = torch.zeros(1, 1, dtype=torch.long) - target_lengths = torch.ones(1, dtype=torch.long) - h_in = torch.zeros(config['num_layers'], 1, config['pred_hidden']) - c_in = torch.zeros(config['num_layers'], 1, config['pred_hidden']) - - # Trace the model - with torch.no_grad(): - traced = torch.jit.trace(wrapper, (targets, target_lengths, h_in, c_in)) - - # Convert to CoreML - mlmodel = ct.convert( - traced, - inputs=[ - ct.TensorType(name="targets", shape=(1, 1), dtype=np.int32), - ct.TensorType(name="target_lengths", shape=(1,), dtype=np.int32), - ct.TensorType(name="h_in", shape=(config['num_layers'], 1, config['pred_hidden']), dtype=np.float32), - ct.TensorType(name="c_in", shape=(config['num_layers'], 1, config['pred_hidden']), dtype=np.float32), - ], - outputs=[ - ct.TensorType(name="decoder_output"), - ct.TensorType(name="h_out"), - ct.TensorType(name="c_out"), - ], - minimum_deployment_target=ct.target.iOS17, - compute_precision=ct.precision.FLOAT16, - ) - - # Add metadata - mlmodel.author = "Fluid Inference" - mlmodel.short_description = "Hybrid TDT Decoder (110M)" - - # Save - output_path = output_dir / "Decoder.mlpackage" - mlmodel.save(str(output_path)) - print(f" Saved to {output_path}") - - return mlmodel - - -def convert_joint(model, config, output_dir: Path): - """Convert joint network to CoreML.""" - print(f"Converting JointDecision...") - print(f" encoder_dim={config['encoder_dim']}, pred_hidden={config['pred_hidden']}") - print(f" vocab_size={config['vocab_size']}, num_durations={config['num_durations']}") - - wrapper = JointWrapper( - model.joint, - vocab_size=config['vocab_size'], - num_durations=config['num_durations'] - ) - wrapper.eval() - - # Create example inputs - shape [1, dim, 1] - encoder_step = torch.randn(1, config['encoder_dim'], 1) - decoder_step = torch.randn(1, config['pred_hidden'], 1) - - # Trace the model - with torch.no_grad(): - traced = torch.jit.trace(wrapper, (encoder_step, decoder_step)) - - # Convert to CoreML - mlmodel = ct.convert( - traced, - inputs=[ - ct.TensorType(name="encoder_step", shape=(1, config['encoder_dim'], 1), dtype=np.float32), - ct.TensorType(name="decoder_step", shape=(1, config['pred_hidden'], 1), dtype=np.float32), - ], - outputs=[ - ct.TensorType(name="token_id"), - ct.TensorType(name="token_prob"), - ct.TensorType(name="duration"), - ], - minimum_deployment_target=ct.target.iOS17, - compute_precision=ct.precision.FLOAT16, - ) - - # Add metadata - mlmodel.author = "Fluid Inference" - mlmodel.short_description = "Hybrid Joint Decision (110M)" - - # Save - output_path = output_dir / "JointDecision.mlpackage" - mlmodel.save(str(output_path)) - print(f" Saved to {output_path}") - - return mlmodel - - -def main(): - parser = argparse.ArgumentParser(description="Convert TDT decoder to CoreML (0.6B format)") - parser.add_argument( - "--model-name", - default="nvidia/parakeet-tdt_ctc-110m", - help="NeMo model name or path" - ) - parser.add_argument( - "--output-dir", - type=Path, - default=Path("./output"), - help="Output directory for CoreML models" - ) - args = parser.parse_args() - - # Create output directory - args.output_dir.mkdir(parents=True, exist_ok=True) - - # Load model - print(f"Loading model: {args.model_name}") - model = nemo_asr.models.EncDecRNNTBPEModel.from_pretrained(args.model_name) - model.eval() - - # Get model configuration - config = get_model_config(model) - - # Auto-detect encoder dim if not found - if config['encoder_dim'] is None: - print("Auto-detecting encoder dimension...") - dummy_audio = torch.randn(1, 16000) - dummy_length = torch.tensor([16000]) - with torch.no_grad(): - enc_out, enc_len = model.encoder( - audio_signal=dummy_audio, - length=dummy_length - ) - config['encoder_dim'] = enc_out.shape[-1] - - print(f"\nModel config:") - for k, v in config.items(): - print(f" {k}: {v}") - - # Convert components - print() - convert_decoder(model, config, args.output_dir) - convert_joint(model, config, args.output_dir) - - print("\nConversion complete!") - print(f"Models saved to: {args.output_dir}") - print("\nNext steps:") - print("1. Compile to .mlmodelc:") - print(f" cd {args.output_dir}") - print(" xcrun coremlcompiler compile Decoder.mlpackage .") - print(" xcrun coremlcompiler compile JointDecision.mlpackage .") - print("2. Copy to model cache:") - print(" cp -r Decoder.mlmodelc JointDecision.mlmodelc ~/Library/Application\\ Support/FluidAudio/Models/parakeet-ctc-110m-coreml/") - print("3. Test with: swift run fluidaudio hybrid-earnings-benchmark --max-files 1") - - -if __name__ == "__main__": - main() diff --git a/parakeet-tdt-ctc-110m/coreml/audio/yc_first_minute_16k_15s.wav b/parakeet-tdt-ctc-110m/coreml/audio/yc_first_minute_16k_15s.wav deleted file mode 100644 index 0b8040f5f124d20a1d7e812c576c831a7573eddf..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/audio/yc_first_minute_16k_15s.wav +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:c79c8bc763b4efccb3e12f199ec0a59aa2edc5e9e4d21ca70fde8f36762d4147 -size 480078 diff --git a/parakeet-tdt-ctc-110m/coreml/compiled_models/CTCHead.mlmodelc/analytics/coremldata.bin b/parakeet-tdt-ctc-110m/coreml/compiled_models/CTCHead.mlmodelc/analytics/coremldata.bin deleted file mode 100644 index 4093e2df453ec508a533d293a574198b2372845b..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/compiled_models/CTCHead.mlmodelc/analytics/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:fc681823d92eca3dbece3a30c975afa7251eedae0e718b07ffbf1a8b4313b87e -size 243 diff --git a/parakeet-tdt-ctc-110m/coreml/compiled_models/CTCHead.mlmodelc/coremldata.bin b/parakeet-tdt-ctc-110m/coreml/compiled_models/CTCHead.mlmodelc/coremldata.bin deleted file mode 100644 index db4ae8de920867977480465fffece41bc606c0a2..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/compiled_models/CTCHead.mlmodelc/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:2ebec8fc38c063de4b2159e21b1f981309fa5947c24d7e4883aca20f7c15fbb9 -size 377 diff --git a/parakeet-tdt-ctc-110m/coreml/compiled_models/CTCHead.mlmodelc/metadata.json b/parakeet-tdt-ctc-110m/coreml/compiled_models/CTCHead.mlmodelc/metadata.json deleted file mode 100644 index 4abe30bf0b35b72d2c9c013b61cc78b4f26a85e1..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/compiled_models/CTCHead.mlmodelc/metadata.json +++ /dev/null @@ -1,66 +0,0 @@ -[ - { - "metadataOutputVersion" : "3.0", - "shortDescription" : "Parakeet 110M CTC decoder head", - "outputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 188 × 1025)", - "shortDescription" : "", - "shape" : "[1, 188, 1025]", - "name" : "ctc_logits", - "type" : "MultiArray" - } - ], - "storagePrecision" : "Float16", - "modelParameters" : [ - - ], - "author" : "Fluid Inference", - "specificationVersion" : 8, - "mlProgramOperationTypeHistogram" : { - "Ios17.cast" : 2, - "Ios17.conv" : 1, - "Ios17.transpose" : 1, - "Ios16.softmax" : 1, - "Ios17.log" : 1 - }, - "computePrecision" : "Mixed (Float16, Float32, Int32)", - "isUpdatable" : "0", - "stateSchema" : [ - - ], - "availability" : { - "macOS" : "14.0", - "tvOS" : "17.0", - "visionOS" : "1.0", - "watchOS" : "10.0", - "iOS" : "17.0", - "macCatalyst" : "17.0" - }, - "modelType" : { - "name" : "MLModelType_mlProgram" - }, - "inputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 512 × 188)", - "shortDescription" : "", - "shape" : "[1, 512, 188]", - "name" : "encoder_output", - "type" : "MultiArray" - } - ], - "userDefinedMetadata" : { - "com.github.apple.coremltools.source_dialect" : "TorchScript", - "com.github.apple.coremltools.source" : "torch==2.9.0", - "com.github.apple.coremltools.version" : "8.3.0" - }, - "generatedClassName" : "parakeet_ctc_head", - "method" : "predict" - } -] \ No newline at end of file diff --git a/parakeet-tdt-ctc-110m/coreml/compiled_models/CTCHead.mlmodelc/model.mil b/parakeet-tdt-ctc-110m/coreml/compiled_models/CTCHead.mlmodelc/model.mil deleted file mode 100644 index 67b3b5f87ce33caed71f9233828f90775bb8ac9d..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/compiled_models/CTCHead.mlmodelc/model.mil +++ /dev/null @@ -1,24 +0,0 @@ -program(1.0) -[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3510.2.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.9.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "8.3.0"}})] -{ - func main(tensor encoder_output) { - tensor var_4 = const()[name = tensor("op_4"), val = tensor(-1)]; - tensor var_18_pad_type_0 = const()[name = tensor("op_18_pad_type_0"), val = tensor("valid")]; - tensor var_18_strides_0 = const()[name = tensor("op_18_strides_0"), val = tensor([1])]; - tensor var_18_pad_0 = const()[name = tensor("op_18_pad_0"), val = tensor([0, 0])]; - tensor var_18_dilations_0 = const()[name = tensor("op_18_dilations_0"), val = tensor([1])]; - tensor var_18_groups_0 = const()[name = tensor("op_18_groups_0"), val = tensor(1)]; - tensor encoder_output_to_fp16_dtype_0 = const()[name = tensor("encoder_output_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor module_decoder_layers_0_weight_to_fp16 = const()[name = tensor("module_decoder_layers_0_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; - tensor module_decoder_layers_0_bias_to_fp16 = const()[name = tensor("module_decoder_layers_0_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1049728)))]; - tensor encoder_output_to_fp16 = cast(dtype = encoder_output_to_fp16_dtype_0, x = encoder_output)[name = tensor("cast_1")]; - tensor var_18_cast_fp16 = conv(bias = module_decoder_layers_0_bias_to_fp16, dilations = var_18_dilations_0, groups = var_18_groups_0, pad = var_18_pad_0, pad_type = var_18_pad_type_0, strides = var_18_strides_0, weight = module_decoder_layers_0_weight_to_fp16, x = encoder_output_to_fp16)[name = tensor("op_18_cast_fp16")]; - tensor input_perm_0 = const()[name = tensor("input_perm_0"), val = tensor([0, 2, 1])]; - tensor input_cast_fp16 = transpose(perm = input_perm_0, x = var_18_cast_fp16)[name = tensor("transpose_0")]; - tensor out_objects_softmax_cast_fp16 = softmax(axis = var_4, x = input_cast_fp16)[name = tensor("out_objects_softmax_cast_fp16")]; - tensor out_objects_epsilon_0 = const()[name = tensor("out_objects_epsilon_0"), val = tensor(0x1p-149)]; - tensor out_objects_cast_fp16 = log(epsilon = out_objects_epsilon_0, x = out_objects_softmax_cast_fp16)[name = tensor("out_objects_cast_fp16")]; - tensor out_objects_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("out_objects_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; - tensor ctc_logits = cast(dtype = out_objects_cast_fp16_to_fp32_dtype_0, x = out_objects_cast_fp16)[name = tensor("cast_0")]; - } -> (ctc_logits); -} \ No newline at end of file diff --git a/parakeet-tdt-ctc-110m/coreml/compiled_models/CTCHead.mlmodelc/weights/weight.bin b/parakeet-tdt-ctc-110m/coreml/compiled_models/CTCHead.mlmodelc/weights/weight.bin deleted file mode 100644 index 23cedd570125191064c6111997f08c48898245f7..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/compiled_models/CTCHead.mlmodelc/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:fb9bead064427ffcb7529c0e3f378e421b4dde8e6d81447b6d1ca3352ca850e1 -size 1051842 diff --git a/parakeet-tdt-ctc-110m/coreml/compiled_models/Decoder.mlmodelc/analytics/coremldata.bin b/parakeet-tdt-ctc-110m/coreml/compiled_models/Decoder.mlmodelc/analytics/coremldata.bin deleted file mode 100644 index dd015548a03cc453e6477a24374e8fee955e75ef..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/compiled_models/Decoder.mlmodelc/analytics/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:990455f6431342750254f66edf27bfb41be62a7ba17a18e1dd6afd4f5f56e9eb -size 243 diff --git a/parakeet-tdt-ctc-110m/coreml/compiled_models/Decoder.mlmodelc/coremldata.bin b/parakeet-tdt-ctc-110m/coreml/compiled_models/Decoder.mlmodelc/coremldata.bin deleted file mode 100644 index cbea0affef0c3bb63be750e134dc0c9ed70aadb1..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/compiled_models/Decoder.mlmodelc/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:29009727821ad8551ab5fe9271e93c597d92a9714f64b94aa533a9ceb6e22b93 -size 498 diff --git a/parakeet-tdt-ctc-110m/coreml/compiled_models/Decoder.mlmodelc/metadata.json b/parakeet-tdt-ctc-110m/coreml/compiled_models/Decoder.mlmodelc/metadata.json deleted file mode 100644 index 786f4bb2823b1aa7ebb55fe335c9385ee0065b7c..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/compiled_models/Decoder.mlmodelc/metadata.json +++ /dev/null @@ -1,118 +0,0 @@ -[ - { - "metadataOutputVersion" : "3.0", - "shortDescription" : "Parakeet 110M decoder (RNNT prediction network)", - "outputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 640 × 1)", - "shortDescription" : "", - "shape" : "[1, 640, 1]", - "name" : "decoder", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 1 × 640)", - "shortDescription" : "", - "shape" : "[1, 1, 640]", - "name" : "h_out", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 1 × 640)", - "shortDescription" : "", - "shape" : "[1, 1, 640]", - "name" : "c_out", - "type" : "MultiArray" - } - ], - "storagePrecision" : "Float16", - "modelParameters" : [ - - ], - "author" : "Fluid Inference", - "specificationVersion" : 8, - "mlProgramOperationTypeHistogram" : { - "Ios17.squeeze" : 2, - "Ios17.gather" : 1, - "Ios17.cast" : 6, - "Ios17.lstm" : 1, - "Ios17.transpose" : 2, - "Identity" : 1, - "Ios17.expandDims" : 2 - }, - "computePrecision" : "Mixed (Float16, Float32, Int16, Int32)", - "isUpdatable" : "0", - "stateSchema" : [ - - ], - "availability" : { - "macOS" : "14.0", - "tvOS" : "17.0", - "visionOS" : "1.0", - "watchOS" : "10.0", - "iOS" : "17.0", - "macCatalyst" : "17.0" - }, - "modelType" : { - "name" : "MLModelType_mlProgram" - }, - "inputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1 × 1)", - "shortDescription" : "", - "shape" : "[1, 1]", - "name" : "targets", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1)", - "shortDescription" : "", - "shape" : "[1]", - "name" : "target_length", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 1 × 640)", - "shortDescription" : "", - "shape" : "[1, 1, 640]", - "name" : "h_in", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 1 × 640)", - "shortDescription" : "", - "shape" : "[1, 1, 640]", - "name" : "c_in", - "type" : "MultiArray" - } - ], - "userDefinedMetadata" : { - "com.github.apple.coremltools.version" : "8.3.0", - "com.github.apple.coremltools.source_dialect" : "TorchScript", - "com.github.apple.coremltools.source" : "torch==2.9.0" - }, - "generatedClassName" : "parakeet_decoder", - "method" : "predict" - } -] \ No newline at end of file diff --git a/parakeet-tdt-ctc-110m/coreml/compiled_models/Decoder.mlmodelc/model.mil b/parakeet-tdt-ctc-110m/coreml/compiled_models/Decoder.mlmodelc/model.mil deleted file mode 100644 index f69c7247525bc9d95f9341acc4f11eca709c7d00..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/compiled_models/Decoder.mlmodelc/model.mil +++ /dev/null @@ -1,45 +0,0 @@ -program(1.0) -[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3510.2.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.9.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "8.3.0"}})] -{ - func main(tensor c_in, tensor h_in, tensor target_length, tensor targets) { - tensor y_axis_0 = const()[name = tensor("y_axis_0"), val = tensor(0)]; - tensor y_batch_dims_0 = const()[name = tensor("y_batch_dims_0"), val = tensor(0)]; - tensor y_validate_indices_0 = const()[name = tensor("y_validate_indices_0"), val = tensor(false)]; - tensor module_prediction_embed_weight_to_fp16 = const()[name = tensor("module_prediction_embed_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; - tensor targets_to_int16_dtype_0 = const()[name = tensor("targets_to_int16_dtype_0"), val = tensor("int16")]; - tensor targets_to_int16 = cast(dtype = targets_to_int16_dtype_0, x = targets)[name = tensor("cast_8")]; - tensor y_cast_fp16_cast_uint16 = gather(axis = y_axis_0, batch_dims = y_batch_dims_0, indices = targets_to_int16, validate_indices = y_validate_indices_0, x = module_prediction_embed_weight_to_fp16)[name = tensor("y_cast_fp16_cast_uint16")]; - tensor input_3_perm_0 = const()[name = tensor("input_3_perm_0"), val = tensor([1, 0, 2])]; - tensor input_lstm_h0_squeeze_axes_0 = const()[name = tensor("input_lstm_h0_squeeze_axes_0"), val = tensor([0])]; - tensor h_in_to_fp16_dtype_0 = const()[name = tensor("h_in_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor h_in_to_fp16 = cast(dtype = h_in_to_fp16_dtype_0, x = h_in)[name = tensor("cast_7")]; - tensor input_lstm_h0_squeeze_cast_fp16 = squeeze(axes = input_lstm_h0_squeeze_axes_0, x = h_in_to_fp16)[name = tensor("input_lstm_h0_squeeze_cast_fp16")]; - tensor input_lstm_c0_squeeze_axes_0 = const()[name = tensor("input_lstm_c0_squeeze_axes_0"), val = tensor([0])]; - tensor c_in_to_fp16_dtype_0 = const()[name = tensor("c_in_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor c_in_to_fp16 = cast(dtype = c_in_to_fp16_dtype_0, x = c_in)[name = tensor("cast_6")]; - tensor input_lstm_c0_squeeze_cast_fp16 = squeeze(axes = input_lstm_c0_squeeze_axes_0, x = c_in_to_fp16)[name = tensor("input_lstm_c0_squeeze_cast_fp16")]; - tensor input_direction_0 = const()[name = tensor("input_direction_0"), val = tensor("forward")]; - tensor input_output_sequence_0 = const()[name = tensor("input_output_sequence_0"), val = tensor(true)]; - tensor input_recurrent_activation_0 = const()[name = tensor("input_recurrent_activation_0"), val = tensor("sigmoid")]; - tensor input_cell_activation_0 = const()[name = tensor("input_cell_activation_0"), val = tensor("tanh")]; - tensor input_activation_0 = const()[name = tensor("input_activation_0"), val = tensor("tanh")]; - tensor concat_1_to_fp16 = const()[name = tensor("concat_1_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1312128)))]; - tensor concat_2_to_fp16 = const()[name = tensor("concat_2_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4588992)))]; - tensor concat_0_to_fp16 = const()[name = tensor("concat_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7865856)))]; - tensor input_3_cast_fp16 = transpose(perm = input_3_perm_0, x = y_cast_fp16_cast_uint16)[name = tensor("transpose_2")]; - tensor input_cast_fp16_0, tensor input_cast_fp16_1, tensor input_cast_fp16_2 = lstm(activation = input_activation_0, bias = concat_0_to_fp16, cell_activation = input_cell_activation_0, direction = input_direction_0, initial_c = input_lstm_c0_squeeze_cast_fp16, initial_h = input_lstm_h0_squeeze_cast_fp16, output_sequence = input_output_sequence_0, recurrent_activation = input_recurrent_activation_0, weight_hh = concat_2_to_fp16, weight_ih = concat_1_to_fp16, x = input_3_cast_fp16)[name = tensor("input_cast_fp16")]; - tensor obj_3_axes_0 = const()[name = tensor("obj_3_axes_0"), val = tensor([0])]; - tensor obj_3_cast_fp16 = expand_dims(axes = obj_3_axes_0, x = input_cast_fp16_1)[name = tensor("obj_3_cast_fp16")]; - tensor obj_3_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("obj_3_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; - tensor obj_axes_0 = const()[name = tensor("obj_axes_0"), val = tensor([0])]; - tensor obj_cast_fp16 = expand_dims(axes = obj_axes_0, x = input_cast_fp16_2)[name = tensor("obj_cast_fp16")]; - tensor obj_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("obj_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; - tensor transpose_0_perm_0 = const()[name = tensor("transpose_0_perm_0"), val = tensor([1, 2, 0])]; - tensor transpose_0_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("transpose_0_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; - tensor transpose_0_cast_fp16 = transpose(perm = transpose_0_perm_0, x = input_cast_fp16_0)[name = tensor("transpose_1")]; - tensor decoder = cast(dtype = transpose_0_cast_fp16_to_fp32_dtype_0, x = transpose_0_cast_fp16)[name = tensor("cast_3")]; - tensor c_out = cast(dtype = obj_cast_fp16_to_fp32_dtype_0, x = obj_cast_fp16)[name = tensor("cast_4")]; - tensor h_out = cast(dtype = obj_3_cast_fp16_to_fp32_dtype_0, x = obj_3_cast_fp16)[name = tensor("cast_5")]; - tensor target_length_tmp = identity(x = target_length)[name = tensor("target_length_tmp")]; - } -> (decoder, h_out, c_out); -} \ No newline at end of file diff --git a/parakeet-tdt-ctc-110m/coreml/compiled_models/Decoder.mlmodelc/weights/weight.bin b/parakeet-tdt-ctc-110m/coreml/compiled_models/Decoder.mlmodelc/weights/weight.bin deleted file mode 100644 index 4543ad4ea89b21ebed0de0d5ec3975d07eace7fb..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/compiled_models/Decoder.mlmodelc/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:dd90b58597ee2c172c672dffe13b1110898ba07394c1a15efc96cc8c6b18411b -size 7871040 diff --git a/parakeet-tdt-ctc-110m/coreml/compiled_models/Encoder.mlmodelc/analytics/coremldata.bin b/parakeet-tdt-ctc-110m/coreml/compiled_models/Encoder.mlmodelc/analytics/coremldata.bin deleted file mode 100644 index d2787d4e5376d7f894d2b0a5a09f497a9b486c73..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/compiled_models/Encoder.mlmodelc/analytics/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:b7ae65e2af616df46066b7efca2d7c19941666ac0685f4ed005666890a052b0d -size 243 diff --git a/parakeet-tdt-ctc-110m/coreml/compiled_models/Encoder.mlmodelc/coremldata.bin b/parakeet-tdt-ctc-110m/coreml/compiled_models/Encoder.mlmodelc/coremldata.bin deleted file mode 100644 index 9d9cdbfd392e8f7391b47c1c96f0fcbaab440389..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/compiled_models/Encoder.mlmodelc/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:0713c2d6ac5f8f6fb9582be250351ebd8efc925f71f4261191165f1406f2ee5d -size 437 diff --git a/parakeet-tdt-ctc-110m/coreml/compiled_models/Encoder.mlmodelc/metadata.json b/parakeet-tdt-ctc-110m/coreml/compiled_models/Encoder.mlmodelc/metadata.json deleted file mode 100644 index 8e3f6cc9ddca206205987b2c828d2e595e19bfc2..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/compiled_models/Encoder.mlmodelc/metadata.json +++ /dev/null @@ -1,105 +0,0 @@ -[ - { - "metadataOutputVersion" : "3.0", - "shortDescription" : "Parakeet 110M encoder (15 s window)", - "outputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 512 × 188)", - "shortDescription" : "", - "shape" : "[1, 512, 188]", - "name" : "encoder_output", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1)", - "shortDescription" : "", - "shape" : "[1]", - "name" : "encoder_length", - "type" : "MultiArray" - } - ], - "storagePrecision" : "Float16", - "modelParameters" : [ - - ], - "author" : "Fluid Inference", - "specificationVersion" : 8, - "mlProgramOperationTypeHistogram" : { - "Ios17.logicalAnd" : 2, - "Ios17.reshape" : 103, - "Ios16.softmax" : 17, - "Ios17.matmul" : 51, - "Ios17.transpose" : 123, - "Split" : 17, - "Ios17.expandDims" : 17, - "Select" : 51, - "Ios17.add" : 128, - "Tile" : 8, - "Ios17.sliceByIndex" : 34, - "Ios16.sigmoid" : 17, - "Pad" : 34, - "Ios17.logicalNot" : 2, - "Ios17.layerNorm" : 85, - "Ios16.silu" : 51, - "Ios17.less" : 5, - "Ios17.sub" : 3, - "Ios17.conv" : 56, - "Ios16.relu" : 3, - "Ios17.linear" : 137, - "Ios17.cast" : 11, - "Ios17.floorDiv" : 3, - "Ios17.mul" : 77 - }, - "computePrecision" : "Mixed (Float16, Float32, Int32)", - "isUpdatable" : "0", - "stateSchema" : [ - - ], - "availability" : { - "macOS" : "14.0", - "tvOS" : "17.0", - "visionOS" : "1.0", - "watchOS" : "10.0", - "iOS" : "17.0", - "macCatalyst" : "17.0" - }, - "modelType" : { - "name" : "MLModelType_mlProgram" - }, - "inputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 80 × 1501)", - "shortDescription" : "", - "shape" : "[1, 80, 1501]", - "name" : "mel_features", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1)", - "shortDescription" : "", - "shape" : "[1]", - "name" : "mel_length", - "type" : "MultiArray" - } - ], - "userDefinedMetadata" : { - "com.github.apple.coremltools.source_dialect" : "TorchScript", - "com.github.apple.coremltools.source" : "torch==2.9.0", - "com.github.apple.coremltools.version" : "8.3.0" - }, - "generatedClassName" : "parakeet_encoder", - "method" : "predict" - } -] \ No newline at end of file diff --git a/parakeet-tdt-ctc-110m/coreml/compiled_models/Encoder.mlmodelc/model.mil b/parakeet-tdt-ctc-110m/coreml/compiled_models/Encoder.mlmodelc/model.mil deleted file mode 100644 index 344bd1abf313c03c9ed9287c00893739f5d361bd..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/compiled_models/Encoder.mlmodelc/model.mil +++ /dev/null @@ -1,2690 +0,0 @@ -program(1.0) -[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3510.2.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.9.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "8.3.0"}})] -{ - func main(tensor mel_features, tensor mel_length) { - tensor var_23 = const()[name = tensor("op_23"), val = tensor(-1)]; - tensor x_1_perm_0 = const()[name = tensor("x_1_perm_0"), val = tensor([0, 2, 1])]; - tensor mel_features_to_fp16_dtype_0 = const()[name = tensor("mel_features_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor tensor_1_axes_0 = const()[name = tensor("tensor_1_axes_0"), val = tensor([1])]; - tensor mel_features_to_fp16 = cast(dtype = mel_features_to_fp16_dtype_0, x = mel_features)[name = tensor("cast_182")]; - tensor x_1_cast_fp16 = transpose(perm = x_1_perm_0, x = mel_features_to_fp16)[name = tensor("transpose_207")]; - tensor tensor_1_cast_fp16 = expand_dims(axes = tensor_1_axes_0, x = x_1_cast_fp16)[name = tensor("tensor_1_cast_fp16")]; - tensor expand_dims_0 = const()[name = tensor("expand_dims_0"), val = tensor([[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390, 391, 392, 393, 394, 395, 396, 397, 398, 399, 400, 401, 402, 403, 404, 405, 406, 407, 408, 409, 410, 411, 412, 413, 414, 415, 416, 417, 418, 419, 420, 421, 422, 423, 424, 425, 426, 427, 428, 429, 430, 431, 432, 433, 434, 435, 436, 437, 438, 439, 440, 441, 442, 443, 444, 445, 446, 447, 448, 449, 450, 451, 452, 453, 454, 455, 456, 457, 458, 459, 460, 461, 462, 463, 464, 465, 466, 467, 468, 469, 470, 471, 472, 473, 474, 475, 476, 477, 478, 479, 480, 481, 482, 483, 484, 485, 486, 487, 488, 489, 490, 491, 492, 493, 494, 495, 496, 497, 498, 499, 500, 501, 502, 503, 504, 505, 506, 507, 508, 509, 510, 511, 512, 513, 514, 515, 516, 517, 518, 519, 520, 521, 522, 523, 524, 525, 526, 527, 528, 529, 530, 531, 532, 533, 534, 535, 536, 537, 538, 539, 540, 541, 542, 543, 544, 545, 546, 547, 548, 549, 550, 551, 552, 553, 554, 555, 556, 557, 558, 559, 560, 561, 562, 563, 564, 565, 566, 567, 568, 569, 570, 571, 572, 573, 574, 575, 576, 577, 578, 579, 580, 581, 582, 583, 584, 585, 586, 587, 588, 589, 590, 591, 592, 593, 594, 595, 596, 597, 598, 599, 600, 601, 602, 603, 604, 605, 606, 607, 608, 609, 610, 611, 612, 613, 614, 615, 616, 617, 618, 619, 620, 621, 622, 623, 624, 625, 626, 627, 628, 629, 630, 631, 632, 633, 634, 635, 636, 637, 638, 639, 640, 641, 642, 643, 644, 645, 646, 647, 648, 649, 650, 651, 652, 653, 654, 655, 656, 657, 658, 659, 660, 661, 662, 663, 664, 665, 666, 667, 668, 669, 670, 671, 672, 673, 674, 675, 676, 677, 678, 679, 680, 681, 682, 683, 684, 685, 686, 687, 688, 689, 690, 691, 692, 693, 694, 695, 696, 697, 698, 699, 700, 701, 702, 703, 704, 705, 706, 707, 708, 709, 710, 711, 712, 713, 714, 715, 716, 717, 718, 719, 720, 721, 722, 723, 724, 725, 726, 727, 728, 729, 730, 731, 732, 733, 734, 735, 736, 737, 738, 739, 740, 741, 742, 743, 744, 745, 746, 747, 748, 749, 750, 751, 752, 753, 754, 755, 756, 757, 758, 759, 760, 761, 762, 763, 764, 765, 766, 767, 768, 769, 770, 771, 772, 773, 774, 775, 776, 777, 778, 779, 780, 781, 782, 783, 784, 785, 786, 787, 788, 789, 790, 791, 792, 793, 794, 795, 796, 797, 798, 799, 800, 801, 802, 803, 804, 805, 806, 807, 808, 809, 810, 811, 812, 813, 814, 815, 816, 817, 818, 819, 820, 821, 822, 823, 824, 825, 826, 827, 828, 829, 830, 831, 832, 833, 834, 835, 836, 837, 838, 839, 840, 841, 842, 843, 844, 845, 846, 847, 848, 849, 850, 851, 852, 853, 854, 855, 856, 857, 858, 859, 860, 861, 862, 863, 864, 865, 866, 867, 868, 869, 870, 871, 872, 873, 874, 875, 876, 877, 878, 879, 880, 881, 882, 883, 884, 885, 886, 887, 888, 889, 890, 891, 892, 893, 894, 895, 896, 897, 898, 899, 900, 901, 902, 903, 904, 905, 906, 907, 908, 909, 910, 911, 912, 913, 914, 915, 916, 917, 918, 919, 920, 921, 922, 923, 924, 925, 926, 927, 928, 929, 930, 931, 932, 933, 934, 935, 936, 937, 938, 939, 940, 941, 942, 943, 944, 945, 946, 947, 948, 949, 950, 951, 952, 953, 954, 955, 956, 957, 958, 959, 960, 961, 962, 963, 964, 965, 966, 967, 968, 969, 970, 971, 972, 973, 974, 975, 976, 977, 978, 979, 980, 981, 982, 983, 984, 985, 986, 987, 988, 989, 990, 991, 992, 993, 994, 995, 996, 997, 998, 999, 1000, 1001, 1002, 1003, 1004, 1005, 1006, 1007, 1008, 1009, 1010, 1011, 1012, 1013, 1014, 1015, 1016, 1017, 1018, 1019, 1020, 1021, 1022, 1023, 1024, 1025, 1026, 1027, 1028, 1029, 1030, 1031, 1032, 1033, 1034, 1035, 1036, 1037, 1038, 1039, 1040, 1041, 1042, 1043, 1044, 1045, 1046, 1047, 1048, 1049, 1050, 1051, 1052, 1053, 1054, 1055, 1056, 1057, 1058, 1059, 1060, 1061, 1062, 1063, 1064, 1065, 1066, 1067, 1068, 1069, 1070, 1071, 1072, 1073, 1074, 1075, 1076, 1077, 1078, 1079, 1080, 1081, 1082, 1083, 1084, 1085, 1086, 1087, 1088, 1089, 1090, 1091, 1092, 1093, 1094, 1095, 1096, 1097, 1098, 1099, 1100, 1101, 1102, 1103, 1104, 1105, 1106, 1107, 1108, 1109, 1110, 1111, 1112, 1113, 1114, 1115, 1116, 1117, 1118, 1119, 1120, 1121, 1122, 1123, 1124, 1125, 1126, 1127, 1128, 1129, 1130, 1131, 1132, 1133, 1134, 1135, 1136, 1137, 1138, 1139, 1140, 1141, 1142, 1143, 1144, 1145, 1146, 1147, 1148, 1149, 1150, 1151, 1152, 1153, 1154, 1155, 1156, 1157, 1158, 1159, 1160, 1161, 1162, 1163, 1164, 1165, 1166, 1167, 1168, 1169, 1170, 1171, 1172, 1173, 1174, 1175, 1176, 1177, 1178, 1179, 1180, 1181, 1182, 1183, 1184, 1185, 1186, 1187, 1188, 1189, 1190, 1191, 1192, 1193, 1194, 1195, 1196, 1197, 1198, 1199, 1200, 1201, 1202, 1203, 1204, 1205, 1206, 1207, 1208, 1209, 1210, 1211, 1212, 1213, 1214, 1215, 1216, 1217, 1218, 1219, 1220, 1221, 1222, 1223, 1224, 1225, 1226, 1227, 1228, 1229, 1230, 1231, 1232, 1233, 1234, 1235, 1236, 1237, 1238, 1239, 1240, 1241, 1242, 1243, 1244, 1245, 1246, 1247, 1248, 1249, 1250, 1251, 1252, 1253, 1254, 1255, 1256, 1257, 1258, 1259, 1260, 1261, 1262, 1263, 1264, 1265, 1266, 1267, 1268, 1269, 1270, 1271, 1272, 1273, 1274, 1275, 1276, 1277, 1278, 1279, 1280, 1281, 1282, 1283, 1284, 1285, 1286, 1287, 1288, 1289, 1290, 1291, 1292, 1293, 1294, 1295, 1296, 1297, 1298, 1299, 1300, 1301, 1302, 1303, 1304, 1305, 1306, 1307, 1308, 1309, 1310, 1311, 1312, 1313, 1314, 1315, 1316, 1317, 1318, 1319, 1320, 1321, 1322, 1323, 1324, 1325, 1326, 1327, 1328, 1329, 1330, 1331, 1332, 1333, 1334, 1335, 1336, 1337, 1338, 1339, 1340, 1341, 1342, 1343, 1344, 1345, 1346, 1347, 1348, 1349, 1350, 1351, 1352, 1353, 1354, 1355, 1356, 1357, 1358, 1359, 1360, 1361, 1362, 1363, 1364, 1365, 1366, 1367, 1368, 1369, 1370, 1371, 1372, 1373, 1374, 1375, 1376, 1377, 1378, 1379, 1380, 1381, 1382, 1383, 1384, 1385, 1386, 1387, 1388, 1389, 1390, 1391, 1392, 1393, 1394, 1395, 1396, 1397, 1398, 1399, 1400, 1401, 1402, 1403, 1404, 1405, 1406, 1407, 1408, 1409, 1410, 1411, 1412, 1413, 1414, 1415, 1416, 1417, 1418, 1419, 1420, 1421, 1422, 1423, 1424, 1425, 1426, 1427, 1428, 1429, 1430, 1431, 1432, 1433, 1434, 1435, 1436, 1437, 1438, 1439, 1440, 1441, 1442, 1443, 1444, 1445, 1446, 1447, 1448, 1449, 1450, 1451, 1452, 1453, 1454, 1455, 1456, 1457, 1458, 1459, 1460, 1461, 1462, 1463, 1464, 1465, 1466, 1467, 1468, 1469, 1470, 1471, 1472, 1473, 1474, 1475, 1476, 1477, 1478, 1479, 1480, 1481, 1482, 1483, 1484, 1485, 1486, 1487, 1488, 1489, 1490, 1491, 1492, 1493, 1494, 1495, 1496, 1497, 1498, 1499, 1500]])]; - tensor var_87_axes_0 = const()[name = tensor("op_87_axes_0"), val = tensor([1])]; - tensor var_87 = expand_dims(axes = var_87_axes_0, x = mel_length)[name = tensor("op_87")]; - tensor time_mask_1 = less(x = expand_dims_0, y = var_87)[name = tensor("time_mask_1")]; - tensor var_89_axes_0 = const()[name = tensor("op_89_axes_0"), val = tensor([-1])]; - tensor var_89 = expand_dims(axes = var_89_axes_0, x = time_mask_1)[name = tensor("op_89")]; - tensor var_91_reps_0 = const()[name = tensor("op_91_reps_0"), val = tensor([1, 1, 80])]; - tensor var_91 = tile(reps = var_91_reps_0, x = var_89)[name = tensor("op_91")]; - tensor var_97_axes_0 = const()[name = tensor("op_97_axes_0"), val = tensor([1])]; - tensor cast_3_to_fp16_dtype_0 = const()[name = tensor("cast_3_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor var_91_to_fp16 = cast(dtype = cast_3_to_fp16_dtype_0, x = var_91)[name = tensor("cast_181")]; - tensor var_97_cast_fp16 = expand_dims(axes = var_97_axes_0, x = var_91_to_fp16)[name = tensor("op_97_cast_fp16")]; - tensor input_1_cast_fp16 = mul(x = tensor_1_cast_fp16, y = var_97_cast_fp16)[name = tensor("input_1_cast_fp16")]; - tensor tensor_3_pad_type_0 = const()[name = tensor("tensor_3_pad_type_0"), val = tensor("custom")]; - tensor tensor_3_pad_0 = const()[name = tensor("tensor_3_pad_0"), val = tensor([1, 1, 1, 1])]; - tensor tensor_3_strides_0 = const()[name = tensor("tensor_3_strides_0"), val = tensor([2, 2])]; - tensor tensor_3_dilations_0 = const()[name = tensor("tensor_3_dilations_0"), val = tensor([1, 1])]; - tensor tensor_3_groups_0 = const()[name = tensor("tensor_3_groups_0"), val = tensor(1)]; - tensor module_pre_encode_conv_0_weight_to_fp16 = const()[name = tensor("module_pre_encode_conv_0_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; - tensor module_pre_encode_conv_0_bias_to_fp16 = const()[name = tensor("module_pre_encode_conv_0_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4736)))]; - tensor tensor_3_cast_fp16 = conv(bias = module_pre_encode_conv_0_bias_to_fp16, dilations = tensor_3_dilations_0, groups = tensor_3_groups_0, pad = tensor_3_pad_0, pad_type = tensor_3_pad_type_0, strides = tensor_3_strides_0, weight = module_pre_encode_conv_0_weight_to_fp16, x = input_1_cast_fp16)[name = tensor("tensor_3_cast_fp16")]; - tensor cast_1_to_fp16_dtype_0 = const()[name = tensor("cast_1_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor var_108_promoted_to_fp16 = const()[name = tensor("op_108_promoted_to_fp16"), val = tensor(0x1p+0)]; - tensor mel_length_to_fp16 = cast(dtype = cast_1_to_fp16_dtype_0, x = mel_length)[name = tensor("cast_180")]; - tensor var_109_cast_fp16 = add(x = mel_length_to_fp16, y = var_108_promoted_to_fp16)[name = tensor("op_109_cast_fp16")]; - tensor var_110_promoted_to_fp16 = const()[name = tensor("op_110_promoted_to_fp16"), val = tensor(0x1p+0)]; - tensor var_111_cast_fp16 = add(x = var_109_cast_fp16, y = var_110_promoted_to_fp16)[name = tensor("op_111_cast_fp16")]; - tensor var_112_promoted_to_fp16 = const()[name = tensor("op_112_promoted_to_fp16"), val = tensor(0x1.8p+1)]; - tensor var_113_cast_fp16 = sub(x = var_111_cast_fp16, y = var_112_promoted_to_fp16)[name = tensor("op_113_cast_fp16")]; - tensor var_21_promoted_to_fp16 = const()[name = tensor("op_21_promoted_to_fp16"), val = tensor(0x1p+1)]; - tensor floor_div_0_cast_fp16 = floor_div(x = var_113_cast_fp16, y = var_21_promoted_to_fp16)[name = tensor("floor_div_0_cast_fp16")]; - tensor var_115_promoted_to_fp16 = const()[name = tensor("op_115_promoted_to_fp16"), val = tensor(0x1p+0)]; - tensor current_lengths_3_cast_fp16 = add(x = floor_div_0_cast_fp16, y = var_115_promoted_to_fp16)[name = tensor("current_lengths_3_cast_fp16")]; - tensor cast_4_dtype_0 = const()[name = tensor("cast_4_dtype_0"), val = tensor("int32")]; - tensor expand_dims_1 = const()[name = tensor("expand_dims_1"), val = tensor([[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390, 391, 392, 393, 394, 395, 396, 397, 398, 399, 400, 401, 402, 403, 404, 405, 406, 407, 408, 409, 410, 411, 412, 413, 414, 415, 416, 417, 418, 419, 420, 421, 422, 423, 424, 425, 426, 427, 428, 429, 430, 431, 432, 433, 434, 435, 436, 437, 438, 439, 440, 441, 442, 443, 444, 445, 446, 447, 448, 449, 450, 451, 452, 453, 454, 455, 456, 457, 458, 459, 460, 461, 462, 463, 464, 465, 466, 467, 468, 469, 470, 471, 472, 473, 474, 475, 476, 477, 478, 479, 480, 481, 482, 483, 484, 485, 486, 487, 488, 489, 490, 491, 492, 493, 494, 495, 496, 497, 498, 499, 500, 501, 502, 503, 504, 505, 506, 507, 508, 509, 510, 511, 512, 513, 514, 515, 516, 517, 518, 519, 520, 521, 522, 523, 524, 525, 526, 527, 528, 529, 530, 531, 532, 533, 534, 535, 536, 537, 538, 539, 540, 541, 542, 543, 544, 545, 546, 547, 548, 549, 550, 551, 552, 553, 554, 555, 556, 557, 558, 559, 560, 561, 562, 563, 564, 565, 566, 567, 568, 569, 570, 571, 572, 573, 574, 575, 576, 577, 578, 579, 580, 581, 582, 583, 584, 585, 586, 587, 588, 589, 590, 591, 592, 593, 594, 595, 596, 597, 598, 599, 600, 601, 602, 603, 604, 605, 606, 607, 608, 609, 610, 611, 612, 613, 614, 615, 616, 617, 618, 619, 620, 621, 622, 623, 624, 625, 626, 627, 628, 629, 630, 631, 632, 633, 634, 635, 636, 637, 638, 639, 640, 641, 642, 643, 644, 645, 646, 647, 648, 649, 650, 651, 652, 653, 654, 655, 656, 657, 658, 659, 660, 661, 662, 663, 664, 665, 666, 667, 668, 669, 670, 671, 672, 673, 674, 675, 676, 677, 678, 679, 680, 681, 682, 683, 684, 685, 686, 687, 688, 689, 690, 691, 692, 693, 694, 695, 696, 697, 698, 699, 700, 701, 702, 703, 704, 705, 706, 707, 708, 709, 710, 711, 712, 713, 714, 715, 716, 717, 718, 719, 720, 721, 722, 723, 724, 725, 726, 727, 728, 729, 730, 731, 732, 733, 734, 735, 736, 737, 738, 739, 740, 741, 742, 743, 744, 745, 746, 747, 748, 749, 750]])]; - tensor var_124_axes_0 = const()[name = tensor("op_124_axes_0"), val = tensor([1])]; - tensor current_lengths_3_cast_fp16_to_int32 = cast(dtype = cast_4_dtype_0, x = current_lengths_3_cast_fp16)[name = tensor("cast_179")]; - tensor var_124 = expand_dims(axes = var_124_axes_0, x = current_lengths_3_cast_fp16_to_int32)[name = tensor("op_124")]; - tensor time_mask_3 = less(x = expand_dims_1, y = var_124)[name = tensor("time_mask_3")]; - tensor var_126_axes_0 = const()[name = tensor("op_126_axes_0"), val = tensor([-1])]; - tensor var_126 = expand_dims(axes = var_126_axes_0, x = time_mask_3)[name = tensor("op_126")]; - tensor var_128_reps_0 = const()[name = tensor("op_128_reps_0"), val = tensor([1, 1, 40])]; - tensor var_128 = tile(reps = var_128_reps_0, x = var_126)[name = tensor("op_128")]; - tensor var_134_axes_0 = const()[name = tensor("op_134_axes_0"), val = tensor([1])]; - tensor cast_5_to_fp16_dtype_0 = const()[name = tensor("cast_5_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor var_128_to_fp16 = cast(dtype = cast_5_to_fp16_dtype_0, x = var_128)[name = tensor("cast_178")]; - tensor var_134_cast_fp16 = expand_dims(axes = var_134_axes_0, x = var_128_to_fp16)[name = tensor("op_134_cast_fp16")]; - tensor expanded_mask_3_reps_0 = const()[name = tensor("expanded_mask_3_reps_0"), val = tensor([1, 256, 1, 1])]; - tensor expanded_mask_3_cast_fp16 = tile(reps = expanded_mask_3_reps_0, x = var_134_cast_fp16)[name = tensor("expanded_mask_3_cast_fp16")]; - tensor input_3_cast_fp16 = mul(x = tensor_3_cast_fp16, y = expanded_mask_3_cast_fp16)[name = tensor("input_3_cast_fp16")]; - tensor tensor_5_cast_fp16 = relu(x = input_3_cast_fp16)[name = tensor("tensor_5_cast_fp16")]; - tensor input_5_cast_fp16 = mul(x = tensor_5_cast_fp16, y = expanded_mask_3_cast_fp16)[name = tensor("input_5_cast_fp16")]; - tensor tensor_7_pad_type_0 = const()[name = tensor("tensor_7_pad_type_0"), val = tensor("custom")]; - tensor tensor_7_pad_0 = const()[name = tensor("tensor_7_pad_0"), val = tensor([1, 1, 1, 1])]; - tensor tensor_7_strides_0 = const()[name = tensor("tensor_7_strides_0"), val = tensor([2, 2])]; - tensor tensor_7_groups_0 = const()[name = tensor("tensor_7_groups_0"), val = tensor(256)]; - tensor tensor_7_dilations_0 = const()[name = tensor("tensor_7_dilations_0"), val = tensor([1, 1])]; - tensor module_pre_encode_conv_2_weight_to_fp16 = const()[name = tensor("module_pre_encode_conv_2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(5312)))]; - tensor module_pre_encode_conv_2_bias_to_fp16 = const()[name = tensor("module_pre_encode_conv_2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9984)))]; - tensor tensor_7_cast_fp16 = conv(bias = module_pre_encode_conv_2_bias_to_fp16, dilations = tensor_7_dilations_0, groups = tensor_7_groups_0, pad = tensor_7_pad_0, pad_type = tensor_7_pad_type_0, strides = tensor_7_strides_0, weight = module_pre_encode_conv_2_weight_to_fp16, x = input_5_cast_fp16)[name = tensor("tensor_7_cast_fp16")]; - tensor var_154_promoted_to_fp16 = const()[name = tensor("op_154_promoted_to_fp16"), val = tensor(0x1p+0)]; - tensor var_155_cast_fp16 = add(x = current_lengths_3_cast_fp16, y = var_154_promoted_to_fp16)[name = tensor("op_155_cast_fp16")]; - tensor var_156_promoted_to_fp16 = const()[name = tensor("op_156_promoted_to_fp16"), val = tensor(0x1p+0)]; - tensor var_157_cast_fp16 = add(x = var_155_cast_fp16, y = var_156_promoted_to_fp16)[name = tensor("op_157_cast_fp16")]; - tensor var_158_promoted_to_fp16 = const()[name = tensor("op_158_promoted_to_fp16"), val = tensor(0x1.8p+1)]; - tensor var_159_cast_fp16 = sub(x = var_157_cast_fp16, y = var_158_promoted_to_fp16)[name = tensor("op_159_cast_fp16")]; - tensor var_21_promoted_1_to_fp16 = const()[name = tensor("op_21_promoted_1_to_fp16"), val = tensor(0x1p+1)]; - tensor floor_div_1_cast_fp16 = floor_div(x = var_159_cast_fp16, y = var_21_promoted_1_to_fp16)[name = tensor("floor_div_1_cast_fp16")]; - tensor var_161_promoted_to_fp16 = const()[name = tensor("op_161_promoted_to_fp16"), val = tensor(0x1p+0)]; - tensor current_lengths_5_cast_fp16 = add(x = floor_div_1_cast_fp16, y = var_161_promoted_to_fp16)[name = tensor("current_lengths_5_cast_fp16")]; - tensor cast_6_dtype_0 = const()[name = tensor("cast_6_dtype_0"), val = tensor("int32")]; - tensor expand_dims_2 = const()[name = tensor("expand_dims_2"), val = tensor([[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375]])]; - tensor var_170_axes_0 = const()[name = tensor("op_170_axes_0"), val = tensor([1])]; - tensor current_lengths_5_cast_fp16_to_int32 = cast(dtype = cast_6_dtype_0, x = current_lengths_5_cast_fp16)[name = tensor("cast_177")]; - tensor var_170 = expand_dims(axes = var_170_axes_0, x = current_lengths_5_cast_fp16_to_int32)[name = tensor("op_170")]; - tensor time_mask_5 = less(x = expand_dims_2, y = var_170)[name = tensor("time_mask_5")]; - tensor var_172_axes_0 = const()[name = tensor("op_172_axes_0"), val = tensor([-1])]; - tensor var_172 = expand_dims(axes = var_172_axes_0, x = time_mask_5)[name = tensor("op_172")]; - tensor var_174_reps_0 = const()[name = tensor("op_174_reps_0"), val = tensor([1, 1, 20])]; - tensor var_174 = tile(reps = var_174_reps_0, x = var_172)[name = tensor("op_174")]; - tensor var_180_axes_0 = const()[name = tensor("op_180_axes_0"), val = tensor([1])]; - tensor cast_7_to_fp16_dtype_0 = const()[name = tensor("cast_7_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor var_174_to_fp16 = cast(dtype = cast_7_to_fp16_dtype_0, x = var_174)[name = tensor("cast_176")]; - tensor var_180_cast_fp16 = expand_dims(axes = var_180_axes_0, x = var_174_to_fp16)[name = tensor("op_180_cast_fp16")]; - tensor expanded_mask_7_reps_0 = const()[name = tensor("expanded_mask_7_reps_0"), val = tensor([1, 256, 1, 1])]; - tensor expanded_mask_7_cast_fp16 = tile(reps = expanded_mask_7_reps_0, x = var_180_cast_fp16)[name = tensor("expanded_mask_7_cast_fp16")]; - tensor input_7_cast_fp16 = mul(x = tensor_7_cast_fp16, y = expanded_mask_7_cast_fp16)[name = tensor("input_7_cast_fp16")]; - tensor tensor_9_pad_type_0 = const()[name = tensor("tensor_9_pad_type_0"), val = tensor("valid")]; - tensor tensor_9_strides_0 = const()[name = tensor("tensor_9_strides_0"), val = tensor([1, 1])]; - tensor tensor_9_pad_0 = const()[name = tensor("tensor_9_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor tensor_9_dilations_0 = const()[name = tensor("tensor_9_dilations_0"), val = tensor([1, 1])]; - tensor tensor_9_groups_0 = const()[name = tensor("tensor_9_groups_0"), val = tensor(1)]; - tensor module_pre_encode_conv_3_weight_to_fp16 = const()[name = tensor("module_pre_encode_conv_3_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10560)))]; - tensor module_pre_encode_conv_3_bias_to_fp16 = const()[name = tensor("module_pre_encode_conv_3_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141696)))]; - tensor tensor_9_cast_fp16 = conv(bias = module_pre_encode_conv_3_bias_to_fp16, dilations = tensor_9_dilations_0, groups = tensor_9_groups_0, pad = tensor_9_pad_0, pad_type = tensor_9_pad_type_0, strides = tensor_9_strides_0, weight = module_pre_encode_conv_3_weight_to_fp16, x = input_7_cast_fp16)[name = tensor("tensor_9_cast_fp16")]; - tensor input_9_cast_fp16 = mul(x = tensor_9_cast_fp16, y = expanded_mask_7_cast_fp16)[name = tensor("input_9_cast_fp16")]; - tensor tensor_11_cast_fp16 = relu(x = input_9_cast_fp16)[name = tensor("tensor_11_cast_fp16")]; - tensor input_11_cast_fp16 = mul(x = tensor_11_cast_fp16, y = expanded_mask_7_cast_fp16)[name = tensor("input_11_cast_fp16")]; - tensor tensor_13_pad_type_0 = const()[name = tensor("tensor_13_pad_type_0"), val = tensor("custom")]; - tensor tensor_13_pad_0 = const()[name = tensor("tensor_13_pad_0"), val = tensor([1, 1, 1, 1])]; - tensor tensor_13_strides_0 = const()[name = tensor("tensor_13_strides_0"), val = tensor([2, 2])]; - tensor tensor_13_groups_0 = const()[name = tensor("tensor_13_groups_0"), val = tensor(256)]; - tensor tensor_13_dilations_0 = const()[name = tensor("tensor_13_dilations_0"), val = tensor([1, 1])]; - tensor module_pre_encode_conv_5_weight_to_fp16 = const()[name = tensor("module_pre_encode_conv_5_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(142272)))]; - tensor module_pre_encode_conv_5_bias_to_fp16 = const()[name = tensor("module_pre_encode_conv_5_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146944)))]; - tensor tensor_13_cast_fp16 = conv(bias = module_pre_encode_conv_5_bias_to_fp16, dilations = tensor_13_dilations_0, groups = tensor_13_groups_0, pad = tensor_13_pad_0, pad_type = tensor_13_pad_type_0, strides = tensor_13_strides_0, weight = module_pre_encode_conv_5_weight_to_fp16, x = input_11_cast_fp16)[name = tensor("tensor_13_cast_fp16")]; - tensor var_215_promoted_to_fp16 = const()[name = tensor("op_215_promoted_to_fp16"), val = tensor(0x1p+0)]; - tensor var_216_cast_fp16 = add(x = current_lengths_5_cast_fp16, y = var_215_promoted_to_fp16)[name = tensor("op_216_cast_fp16")]; - tensor var_217_promoted_to_fp16 = const()[name = tensor("op_217_promoted_to_fp16"), val = tensor(0x1p+0)]; - tensor var_218_cast_fp16 = add(x = var_216_cast_fp16, y = var_217_promoted_to_fp16)[name = tensor("op_218_cast_fp16")]; - tensor var_219_promoted_to_fp16 = const()[name = tensor("op_219_promoted_to_fp16"), val = tensor(0x1.8p+1)]; - tensor var_220_cast_fp16 = sub(x = var_218_cast_fp16, y = var_219_promoted_to_fp16)[name = tensor("op_220_cast_fp16")]; - tensor var_21_promoted_2_to_fp16 = const()[name = tensor("op_21_promoted_2_to_fp16"), val = tensor(0x1p+1)]; - tensor floor_div_2_cast_fp16 = floor_div(x = var_220_cast_fp16, y = var_21_promoted_2_to_fp16)[name = tensor("floor_div_2_cast_fp16")]; - tensor var_222_promoted_to_fp16 = const()[name = tensor("op_222_promoted_to_fp16"), val = tensor(0x1p+0)]; - tensor current_lengths_cast_fp16 = add(x = floor_div_2_cast_fp16, y = var_222_promoted_to_fp16)[name = tensor("current_lengths_cast_fp16")]; - tensor cast_8_dtype_0 = const()[name = tensor("cast_8_dtype_0"), val = tensor("int32")]; - tensor expand_dims_3 = const()[name = tensor("expand_dims_3"), val = tensor([[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187]])]; - tensor var_231_axes_0 = const()[name = tensor("op_231_axes_0"), val = tensor([1])]; - tensor current_lengths_cast_fp16_to_int32 = cast(dtype = cast_8_dtype_0, x = current_lengths_cast_fp16)[name = tensor("cast_175")]; - tensor var_231 = expand_dims(axes = var_231_axes_0, x = current_lengths_cast_fp16_to_int32)[name = tensor("op_231")]; - tensor time_mask = less(x = expand_dims_3, y = var_231)[name = tensor("time_mask")]; - tensor var_233_axes_0 = const()[name = tensor("op_233_axes_0"), val = tensor([-1])]; - tensor var_233 = expand_dims(axes = var_233_axes_0, x = time_mask)[name = tensor("op_233")]; - tensor var_235_reps_0 = const()[name = tensor("op_235_reps_0"), val = tensor([1, 1, 10])]; - tensor var_235 = tile(reps = var_235_reps_0, x = var_233)[name = tensor("op_235")]; - tensor var_241_axes_0 = const()[name = tensor("op_241_axes_0"), val = tensor([1])]; - tensor cast_9_to_fp16_dtype_0 = const()[name = tensor("cast_9_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor var_235_to_fp16 = cast(dtype = cast_9_to_fp16_dtype_0, x = var_235)[name = tensor("cast_174")]; - tensor var_241_cast_fp16 = expand_dims(axes = var_241_axes_0, x = var_235_to_fp16)[name = tensor("op_241_cast_fp16")]; - tensor expanded_mask_13_reps_0 = const()[name = tensor("expanded_mask_13_reps_0"), val = tensor([1, 256, 1, 1])]; - tensor expanded_mask_13_cast_fp16 = tile(reps = expanded_mask_13_reps_0, x = var_241_cast_fp16)[name = tensor("expanded_mask_13_cast_fp16")]; - tensor input_13_cast_fp16 = mul(x = tensor_13_cast_fp16, y = expanded_mask_13_cast_fp16)[name = tensor("input_13_cast_fp16")]; - tensor tensor_15_pad_type_0 = const()[name = tensor("tensor_15_pad_type_0"), val = tensor("valid")]; - tensor tensor_15_strides_0 = const()[name = tensor("tensor_15_strides_0"), val = tensor([1, 1])]; - tensor tensor_15_pad_0 = const()[name = tensor("tensor_15_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor tensor_15_dilations_0 = const()[name = tensor("tensor_15_dilations_0"), val = tensor([1, 1])]; - tensor tensor_15_groups_0 = const()[name = tensor("tensor_15_groups_0"), val = tensor(1)]; - tensor module_pre_encode_conv_6_weight_to_fp16 = const()[name = tensor("module_pre_encode_conv_6_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147520)))]; - tensor module_pre_encode_conv_6_bias_to_fp16 = const()[name = tensor("module_pre_encode_conv_6_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(278656)))]; - tensor tensor_15_cast_fp16 = conv(bias = module_pre_encode_conv_6_bias_to_fp16, dilations = tensor_15_dilations_0, groups = tensor_15_groups_0, pad = tensor_15_pad_0, pad_type = tensor_15_pad_type_0, strides = tensor_15_strides_0, weight = module_pre_encode_conv_6_weight_to_fp16, x = input_13_cast_fp16)[name = tensor("tensor_15_cast_fp16")]; - tensor input_15_cast_fp16 = mul(x = tensor_15_cast_fp16, y = expanded_mask_13_cast_fp16)[name = tensor("input_15_cast_fp16")]; - tensor tensor_cast_fp16 = relu(x = input_15_cast_fp16)[name = tensor("tensor_cast_fp16")]; - tensor x_3_cast_fp16 = mul(x = tensor_cast_fp16, y = expanded_mask_13_cast_fp16)[name = tensor("x_3_cast_fp16")]; - tensor var_275_perm_0 = const()[name = tensor("op_275_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_276 = const()[name = tensor("op_276"), val = tensor([1, 188, -1])]; - tensor var_275_cast_fp16 = transpose(perm = var_275_perm_0, x = x_3_cast_fp16)[name = tensor("transpose_206")]; - tensor input_17_cast_fp16 = reshape(shape = var_276, x = var_275_cast_fp16)[name = tensor("input_17_cast_fp16")]; - tensor module_pre_encode_out_weight_to_fp16 = const()[name = tensor("module_pre_encode_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(279232)))]; - tensor module_pre_encode_out_bias_to_fp16 = const()[name = tensor("module_pre_encode_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2900736)))]; - tensor linear_0_cast_fp16 = linear(bias = module_pre_encode_out_bias_to_fp16, weight = module_pre_encode_out_weight_to_fp16, x = input_17_cast_fp16)[name = tensor("linear_0_cast_fp16")]; - tensor cast_12_dtype_0 = const()[name = tensor("cast_12_dtype_0"), val = tensor("int32")]; - tensor var_314_axes_0 = const()[name = tensor("op_314_axes_0"), val = tensor([-1])]; - tensor encoder_length = cast(dtype = cast_12_dtype_0, x = current_lengths_cast_fp16)[name = tensor("cast_173")]; - tensor var_314 = expand_dims(axes = var_314_axes_0, x = encoder_length)[name = tensor("op_314")]; - tensor pad_mask_1 = less(x = expand_dims_3, y = var_314)[name = tensor("pad_mask_1")]; - tensor var_316_axes_0 = const()[name = tensor("op_316_axes_0"), val = tensor([1])]; - tensor var_316 = expand_dims(axes = var_316_axes_0, x = pad_mask_1)[name = tensor("op_316")]; - tensor var_317 = const()[name = tensor("op_317"), val = tensor([1, 188, 1])]; - tensor pad_mask_for_att_mask_1 = tile(reps = var_317, x = var_316)[name = tensor("pad_mask_for_att_mask_1")]; - tensor var_319_perm_0 = const()[name = tensor("op_319_perm_0"), val = tensor([0, 2, 1])]; - tensor var_319 = transpose(perm = var_319_perm_0, x = pad_mask_for_att_mask_1)[name = tensor("transpose_205")]; - tensor pad_mask_for_att_mask = logical_and(x = pad_mask_for_att_mask_1, y = var_319)[name = tensor("pad_mask_for_att_mask")]; - tensor const_63 = const()[name = tensor("const_63"), val = tensor([[[true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true]]])]; - tensor att_mask = logical_and(x = pad_mask_for_att_mask, y = const_63)[name = tensor("att_mask")]; - tensor mask_9 = logical_not(x = att_mask)[name = tensor("mask_9")]; - tensor pad_mask = logical_not(x = pad_mask_1)[name = tensor("pad_mask")]; - tensor input_21_axes_0 = const()[name = tensor("input_21_axes_0"), val = tensor([-1])]; - tensor module_layers_0_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_0_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2901824)))]; - tensor module_layers_0_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_0_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2902912)))]; - tensor var_9_to_fp16 = const()[name = tensor("op_9_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_21_cast_fp16 = layer_norm(axes = input_21_axes_0, beta = module_layers_0_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_0_norm_feed_forward1_weight_to_fp16, x = linear_0_cast_fp16)[name = tensor("input_21_cast_fp16")]; - tensor module_layers_0_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_0_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2904000)))]; - tensor module_layers_0_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_0_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(5001216)))]; - tensor linear_1_cast_fp16 = linear(bias = module_layers_0_feed_forward1_linear1_bias_to_fp16, weight = module_layers_0_feed_forward1_linear1_weight_to_fp16, x = input_21_cast_fp16)[name = tensor("linear_1_cast_fp16")]; - tensor input_25_cast_fp16 = silu(x = linear_1_cast_fp16)[name = tensor("input_25_cast_fp16")]; - tensor module_layers_0_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_0_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(5005376)))]; - tensor module_layers_0_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_0_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7102592)))]; - tensor linear_2_cast_fp16 = linear(bias = module_layers_0_feed_forward1_linear2_bias_to_fp16, weight = module_layers_0_feed_forward1_linear2_weight_to_fp16, x = input_25_cast_fp16)[name = tensor("linear_2_cast_fp16")]; - tensor var_352_to_fp16 = const()[name = tensor("op_352_to_fp16"), val = tensor(0x1p-1)]; - tensor var_353_cast_fp16 = mul(x = linear_2_cast_fp16, y = var_352_to_fp16)[name = tensor("op_353_cast_fp16")]; - tensor input_31_cast_fp16 = add(x = linear_0_cast_fp16, y = var_353_cast_fp16)[name = tensor("input_31_cast_fp16")]; - tensor query_1_axes_0 = const()[name = tensor("query_1_axes_0"), val = tensor([-1])]; - tensor module_layers_0_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_0_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7103680)))]; - tensor module_layers_0_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_0_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7104768)))]; - tensor query_1_cast_fp16 = layer_norm(axes = query_1_axes_0, beta = module_layers_0_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_0_norm_self_att_weight_to_fp16, x = input_31_cast_fp16)[name = tensor("query_1_cast_fp16")]; - tensor module_layers_0_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_0_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7105856)))]; - tensor module_layers_0_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_0_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7630208)))]; - tensor linear_3_cast_fp16 = linear(bias = module_layers_0_self_attn_linear_q_bias_to_fp16, weight = module_layers_0_self_attn_linear_q_weight_to_fp16, x = query_1_cast_fp16)[name = tensor("linear_3_cast_fp16")]; - tensor var_370 = const()[name = tensor("op_370"), val = tensor([1, -1, 8, 64])]; - tensor q_1_cast_fp16 = reshape(shape = var_370, x = linear_3_cast_fp16)[name = tensor("q_1_cast_fp16")]; - tensor module_layers_0_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_0_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7631296)))]; - tensor module_layers_0_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_0_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8155648)))]; - tensor linear_4_cast_fp16 = linear(bias = module_layers_0_self_attn_linear_k_bias_to_fp16, weight = module_layers_0_self_attn_linear_k_weight_to_fp16, x = query_1_cast_fp16)[name = tensor("linear_4_cast_fp16")]; - tensor var_375 = const()[name = tensor("op_375"), val = tensor([1, -1, 8, 64])]; - tensor k_1_cast_fp16 = reshape(shape = var_375, x = linear_4_cast_fp16)[name = tensor("k_1_cast_fp16")]; - tensor module_layers_0_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_0_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8156736)))]; - tensor module_layers_0_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_0_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8681088)))]; - tensor linear_5_cast_fp16 = linear(bias = module_layers_0_self_attn_linear_v_bias_to_fp16, weight = module_layers_0_self_attn_linear_v_weight_to_fp16, x = query_1_cast_fp16)[name = tensor("linear_5_cast_fp16")]; - tensor var_380 = const()[name = tensor("op_380"), val = tensor([1, -1, 8, 64])]; - tensor v_1_cast_fp16 = reshape(shape = var_380, x = linear_5_cast_fp16)[name = tensor("v_1_cast_fp16")]; - tensor value_3_perm_0 = const()[name = tensor("value_3_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_0_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_0_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8682176)))]; - tensor var_392_cast_fp16 = add(x = q_1_cast_fp16, y = module_layers_0_self_attn_pos_bias_u_to_fp16)[name = tensor("op_392_cast_fp16")]; - tensor module_layers_0_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_0_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8683264)))]; - tensor var_394_cast_fp16 = add(x = q_1_cast_fp16, y = module_layers_0_self_attn_pos_bias_v_to_fp16)[name = tensor("op_394_cast_fp16")]; - tensor q_with_bias_v_1_perm_0 = const()[name = tensor("q_with_bias_v_1_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_7_transpose_x_0 = const()[name = tensor("x_7_transpose_x_0"), val = tensor(false)]; - tensor x_7_transpose_y_0 = const()[name = tensor("x_7_transpose_y_0"), val = tensor(false)]; - tensor var_396_to_fp16 = const()[name = tensor("op_396_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8684352)))]; - tensor q_with_bias_v_1_cast_fp16 = transpose(perm = q_with_bias_v_1_perm_0, x = var_394_cast_fp16)[name = tensor("transpose_203")]; - tensor x_7_cast_fp16 = matmul(transpose_x = x_7_transpose_x_0, transpose_y = x_7_transpose_y_0, x = q_with_bias_v_1_cast_fp16, y = var_396_to_fp16)[name = tensor("x_7_cast_fp16")]; - tensor x_9_pad_0 = const()[name = tensor("x_9_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_9_mode_0 = const()[name = tensor("x_9_mode_0"), val = tensor("constant")]; - tensor const_70_to_fp16 = const()[name = tensor("const_70_to_fp16"), val = tensor(0x0p+0)]; - tensor x_9_cast_fp16 = pad(constant_val = const_70_to_fp16, mode = x_9_mode_0, pad = x_9_pad_0, x = x_7_cast_fp16)[name = tensor("x_9_cast_fp16")]; - tensor var_404 = const()[name = tensor("op_404"), val = tensor([1, 8, -1, 188])]; - tensor x_11_cast_fp16 = reshape(shape = var_404, x = x_9_cast_fp16)[name = tensor("x_11_cast_fp16")]; - tensor var_408_begin_0 = const()[name = tensor("op_408_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_408_end_0 = const()[name = tensor("op_408_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_408_end_mask_0 = const()[name = tensor("op_408_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_408_cast_fp16 = slice_by_index(begin = var_408_begin_0, end = var_408_end_0, end_mask = var_408_end_mask_0, x = x_11_cast_fp16)[name = tensor("op_408_cast_fp16")]; - tensor var_409 = const()[name = tensor("op_409"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_1_cast_fp16 = reshape(shape = var_409, x = var_408_cast_fp16)[name = tensor("matrix_bd_1_cast_fp16")]; - tensor matrix_ac_1_transpose_x_0 = const()[name = tensor("matrix_ac_1_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_1_transpose_y_0 = const()[name = tensor("matrix_ac_1_transpose_y_0"), val = tensor(false)]; - tensor transpose_51_perm_0 = const()[name = tensor("transpose_51_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_52_perm_0 = const()[name = tensor("transpose_52_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_52 = transpose(perm = transpose_52_perm_0, x = k_1_cast_fp16)[name = tensor("transpose_201")]; - tensor transpose_51 = transpose(perm = transpose_51_perm_0, x = var_392_cast_fp16)[name = tensor("transpose_202")]; - tensor matrix_ac_1_cast_fp16 = matmul(transpose_x = matrix_ac_1_transpose_x_0, transpose_y = matrix_ac_1_transpose_y_0, x = transpose_51, y = transpose_52)[name = tensor("matrix_ac_1_cast_fp16")]; - tensor matrix_bd_3_begin_0 = const()[name = tensor("matrix_bd_3_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_3_end_0 = const()[name = tensor("matrix_bd_3_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_3_end_mask_0 = const()[name = tensor("matrix_bd_3_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_3_cast_fp16 = slice_by_index(begin = matrix_bd_3_begin_0, end = matrix_bd_3_end_0, end_mask = matrix_bd_3_end_mask_0, x = matrix_bd_1_cast_fp16)[name = tensor("matrix_bd_3_cast_fp16")]; - tensor var_418_cast_fp16 = add(x = matrix_ac_1_cast_fp16, y = matrix_bd_3_cast_fp16)[name = tensor("op_418_cast_fp16")]; - tensor _inversed_scores_1_y_0_to_fp16 = const()[name = tensor("_inversed_scores_1_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_1_cast_fp16 = mul(x = var_418_cast_fp16, y = _inversed_scores_1_y_0_to_fp16)[name = tensor("_inversed_scores_1_cast_fp16")]; - tensor mask_11_axes_0 = const()[name = tensor("mask_11_axes_0"), val = tensor([1])]; - tensor mask_11 = expand_dims(axes = mask_11_axes_0, x = mask_9)[name = tensor("mask_11")]; - tensor var_12_to_fp16 = const()[name = tensor("op_12_to_fp16"), val = tensor(-0x1.388p+13)]; - tensor scores_3_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_1_cast_fp16, cond = mask_11)[name = tensor("scores_3_cast_fp16")]; - tensor var_424_cast_fp16 = softmax(axis = var_23, x = scores_3_cast_fp16)[name = tensor("op_424_cast_fp16")]; - tensor var_11_to_fp16 = const()[name = tensor("op_11_to_fp16"), val = tensor(0x0p+0)]; - tensor input_33_cast_fp16 = select(a = var_11_to_fp16, b = var_424_cast_fp16, cond = mask_11)[name = tensor("input_33_cast_fp16")]; - tensor x_13_transpose_x_0 = const()[name = tensor("x_13_transpose_x_0"), val = tensor(false)]; - tensor x_13_transpose_y_0 = const()[name = tensor("x_13_transpose_y_0"), val = tensor(false)]; - tensor value_3_cast_fp16 = transpose(perm = value_3_perm_0, x = v_1_cast_fp16)[name = tensor("transpose_204")]; - tensor x_13_cast_fp16 = matmul(transpose_x = x_13_transpose_x_0, transpose_y = x_13_transpose_y_0, x = input_33_cast_fp16, y = value_3_cast_fp16)[name = tensor("x_13_cast_fp16")]; - tensor var_428_perm_0 = const()[name = tensor("op_428_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_429 = const()[name = tensor("op_429"), val = tensor([1, -1, 512])]; - tensor var_428_cast_fp16 = transpose(perm = var_428_perm_0, x = x_13_cast_fp16)[name = tensor("transpose_200")]; - tensor input_35_cast_fp16 = reshape(shape = var_429, x = var_428_cast_fp16)[name = tensor("input_35_cast_fp16")]; - tensor module_layers_0_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_0_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9068416)))]; - tensor module_layers_0_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_0_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9592768)))]; - tensor linear_7_cast_fp16 = linear(bias = module_layers_0_self_attn_linear_out_bias_to_fp16, weight = module_layers_0_self_attn_linear_out_weight_to_fp16, x = input_35_cast_fp16)[name = tensor("linear_7_cast_fp16")]; - tensor input_39_cast_fp16 = add(x = input_31_cast_fp16, y = linear_7_cast_fp16)[name = tensor("input_39_cast_fp16")]; - tensor x_17_axes_0 = const()[name = tensor("x_17_axes_0"), val = tensor([-1])]; - tensor module_layers_0_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_0_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9593856)))]; - tensor module_layers_0_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_0_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9594944)))]; - tensor x_17_cast_fp16 = layer_norm(axes = x_17_axes_0, beta = module_layers_0_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_0_norm_conv_weight_to_fp16, x = input_39_cast_fp16)[name = tensor("x_17_cast_fp16")]; - tensor input_41_perm_0 = const()[name = tensor("input_41_perm_0"), val = tensor([0, 2, 1])]; - tensor input_43_pad_type_0 = const()[name = tensor("input_43_pad_type_0"), val = tensor("valid")]; - tensor input_43_strides_0 = const()[name = tensor("input_43_strides_0"), val = tensor([1])]; - tensor input_43_pad_0 = const()[name = tensor("input_43_pad_0"), val = tensor([0, 0])]; - tensor input_43_dilations_0 = const()[name = tensor("input_43_dilations_0"), val = tensor([1])]; - tensor input_43_groups_0 = const()[name = tensor("input_43_groups_0"), val = tensor(1)]; - tensor module_layers_0_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_0_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9596032)))]; - tensor module_layers_0_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_0_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10644672)))]; - tensor input_41_cast_fp16 = transpose(perm = input_41_perm_0, x = x_17_cast_fp16)[name = tensor("transpose_199")]; - tensor input_43_cast_fp16 = conv(bias = module_layers_0_conv_pointwise_conv1_bias_to_fp16, dilations = input_43_dilations_0, groups = input_43_groups_0, pad = input_43_pad_0, pad_type = input_43_pad_type_0, strides = input_43_strides_0, weight = module_layers_0_conv_pointwise_conv1_weight_to_fp16, x = input_41_cast_fp16)[name = tensor("input_43_cast_fp16")]; - tensor x_19_split_num_splits_0 = const()[name = tensor("x_19_split_num_splits_0"), val = tensor(2)]; - tensor x_19_split_axis_0 = const()[name = tensor("x_19_split_axis_0"), val = tensor(1)]; - tensor x_19_split_cast_fp16_0, tensor x_19_split_cast_fp16_1 = split(axis = x_19_split_axis_0, num_splits = x_19_split_num_splits_0, x = input_43_cast_fp16)[name = tensor("x_19_split_cast_fp16")]; - tensor x_19_split_1_sigmoid_cast_fp16 = sigmoid(x = x_19_split_cast_fp16_1)[name = tensor("x_19_split_1_sigmoid_cast_fp16")]; - tensor x_19_cast_fp16 = mul(x = x_19_split_cast_fp16_0, y = x_19_split_1_sigmoid_cast_fp16)[name = tensor("x_19_cast_fp16")]; - tensor var_453_axes_0 = const()[name = tensor("op_453_axes_0"), val = tensor([1])]; - tensor var_453 = expand_dims(axes = var_453_axes_0, x = pad_mask)[name = tensor("op_453")]; - tensor input_45_cast_fp16 = select(a = var_11_to_fp16, b = x_19_cast_fp16, cond = var_453)[name = tensor("input_45_cast_fp16")]; - tensor input_47_pad_0 = const()[name = tensor("input_47_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_47_mode_0 = const()[name = tensor("input_47_mode_0"), val = tensor("constant")]; - tensor const_73_to_fp16 = const()[name = tensor("const_73_to_fp16"), val = tensor(0x0p+0)]; - tensor input_47_cast_fp16 = pad(constant_val = const_73_to_fp16, mode = input_47_mode_0, pad = input_47_pad_0, x = input_45_cast_fp16)[name = tensor("input_47_cast_fp16")]; - tensor input_49_pad_type_0 = const()[name = tensor("input_49_pad_type_0"), val = tensor("valid")]; - tensor input_49_groups_0 = const()[name = tensor("input_49_groups_0"), val = tensor(512)]; - tensor input_49_strides_0 = const()[name = tensor("input_49_strides_0"), val = tensor([1])]; - tensor input_49_pad_0 = const()[name = tensor("input_49_pad_0"), val = tensor([0, 0])]; - tensor input_49_dilations_0 = const()[name = tensor("input_49_dilations_0"), val = tensor([1])]; - tensor const_234_to_fp16 = const()[name = tensor("const_234_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10646784)))]; - tensor const_235_to_fp16 = const()[name = tensor("const_235_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10656064)))]; - tensor input_51_cast_fp16 = conv(bias = const_235_to_fp16, dilations = input_49_dilations_0, groups = input_49_groups_0, pad = input_49_pad_0, pad_type = input_49_pad_type_0, strides = input_49_strides_0, weight = const_234_to_fp16, x = input_47_cast_fp16)[name = tensor("input_51_cast_fp16")]; - tensor input_53_cast_fp16 = silu(x = input_51_cast_fp16)[name = tensor("input_53_cast_fp16")]; - tensor x_21_pad_type_0 = const()[name = tensor("x_21_pad_type_0"), val = tensor("valid")]; - tensor x_21_strides_0 = const()[name = tensor("x_21_strides_0"), val = tensor([1])]; - tensor x_21_pad_0 = const()[name = tensor("x_21_pad_0"), val = tensor([0, 0])]; - tensor x_21_dilations_0 = const()[name = tensor("x_21_dilations_0"), val = tensor([1])]; - tensor x_21_groups_0 = const()[name = tensor("x_21_groups_0"), val = tensor(1)]; - tensor module_layers_0_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_0_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10657152)))]; - tensor module_layers_0_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_0_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11181504)))]; - tensor x_21_cast_fp16 = conv(bias = module_layers_0_conv_pointwise_conv2_bias_to_fp16, dilations = x_21_dilations_0, groups = x_21_groups_0, pad = x_21_pad_0, pad_type = x_21_pad_type_0, strides = x_21_strides_0, weight = module_layers_0_conv_pointwise_conv2_weight_to_fp16, x = input_53_cast_fp16)[name = tensor("x_21_cast_fp16")]; - tensor input_55_perm_0 = const()[name = tensor("input_55_perm_0"), val = tensor([0, 2, 1])]; - tensor input_55_cast_fp16 = transpose(perm = input_55_perm_0, x = x_21_cast_fp16)[name = tensor("transpose_198")]; - tensor input_57_cast_fp16 = add(x = input_39_cast_fp16, y = input_55_cast_fp16)[name = tensor("input_57_cast_fp16")]; - tensor input_59_axes_0 = const()[name = tensor("input_59_axes_0"), val = tensor([-1])]; - tensor module_layers_0_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_0_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11182592)))]; - tensor module_layers_0_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_0_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11183680)))]; - tensor input_59_cast_fp16 = layer_norm(axes = input_59_axes_0, beta = module_layers_0_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_0_norm_feed_forward2_weight_to_fp16, x = input_57_cast_fp16)[name = tensor("input_59_cast_fp16")]; - tensor module_layers_0_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_0_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11184768)))]; - tensor module_layers_0_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_0_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(13281984)))]; - tensor linear_8_cast_fp16 = linear(bias = module_layers_0_feed_forward2_linear1_bias_to_fp16, weight = module_layers_0_feed_forward2_linear1_weight_to_fp16, x = input_59_cast_fp16)[name = tensor("linear_8_cast_fp16")]; - tensor input_63_cast_fp16 = silu(x = linear_8_cast_fp16)[name = tensor("input_63_cast_fp16")]; - tensor module_layers_0_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_0_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(13286144)))]; - tensor module_layers_0_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_0_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15383360)))]; - tensor linear_9_cast_fp16 = linear(bias = module_layers_0_feed_forward2_linear2_bias_to_fp16, weight = module_layers_0_feed_forward2_linear2_weight_to_fp16, x = input_63_cast_fp16)[name = tensor("linear_9_cast_fp16")]; - tensor var_495_to_fp16 = const()[name = tensor("op_495_to_fp16"), val = tensor(0x1p-1)]; - tensor var_496_cast_fp16 = mul(x = linear_9_cast_fp16, y = var_495_to_fp16)[name = tensor("op_496_cast_fp16")]; - tensor input_69_cast_fp16 = add(x = input_57_cast_fp16, y = var_496_cast_fp16)[name = tensor("input_69_cast_fp16")]; - tensor input_71_axes_0 = const()[name = tensor("input_71_axes_0"), val = tensor([-1])]; - tensor module_layers_0_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_0_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15384448)))]; - tensor module_layers_0_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_0_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15385536)))]; - tensor input_71_cast_fp16 = layer_norm(axes = input_71_axes_0, beta = module_layers_0_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_0_norm_out_weight_to_fp16, x = input_69_cast_fp16)[name = tensor("input_71_cast_fp16")]; - tensor input_73_axes_0 = const()[name = tensor("input_73_axes_0"), val = tensor([-1])]; - tensor module_layers_1_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_1_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15386624)))]; - tensor module_layers_1_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_1_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15387712)))]; - tensor input_73_cast_fp16 = layer_norm(axes = input_73_axes_0, beta = module_layers_1_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_1_norm_feed_forward1_weight_to_fp16, x = input_71_cast_fp16)[name = tensor("input_73_cast_fp16")]; - tensor module_layers_1_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_1_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15388800)))]; - tensor module_layers_1_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_1_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(17486016)))]; - tensor linear_10_cast_fp16 = linear(bias = module_layers_1_feed_forward1_linear1_bias_to_fp16, weight = module_layers_1_feed_forward1_linear1_weight_to_fp16, x = input_73_cast_fp16)[name = tensor("linear_10_cast_fp16")]; - tensor input_77_cast_fp16 = silu(x = linear_10_cast_fp16)[name = tensor("input_77_cast_fp16")]; - tensor module_layers_1_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_1_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(17490176)))]; - tensor module_layers_1_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_1_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19587392)))]; - tensor linear_11_cast_fp16 = linear(bias = module_layers_1_feed_forward1_linear2_bias_to_fp16, weight = module_layers_1_feed_forward1_linear2_weight_to_fp16, x = input_77_cast_fp16)[name = tensor("linear_11_cast_fp16")]; - tensor var_526_to_fp16 = const()[name = tensor("op_526_to_fp16"), val = tensor(0x1p-1)]; - tensor var_527_cast_fp16 = mul(x = linear_11_cast_fp16, y = var_526_to_fp16)[name = tensor("op_527_cast_fp16")]; - tensor input_83_cast_fp16 = add(x = input_71_cast_fp16, y = var_527_cast_fp16)[name = tensor("input_83_cast_fp16")]; - tensor query_3_axes_0 = const()[name = tensor("query_3_axes_0"), val = tensor([-1])]; - tensor module_layers_1_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_1_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19588480)))]; - tensor module_layers_1_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_1_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19589568)))]; - tensor query_3_cast_fp16 = layer_norm(axes = query_3_axes_0, beta = module_layers_1_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_1_norm_self_att_weight_to_fp16, x = input_83_cast_fp16)[name = tensor("query_3_cast_fp16")]; - tensor module_layers_1_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_1_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19590656)))]; - tensor module_layers_1_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_1_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(20115008)))]; - tensor linear_12_cast_fp16 = linear(bias = module_layers_1_self_attn_linear_q_bias_to_fp16, weight = module_layers_1_self_attn_linear_q_weight_to_fp16, x = query_3_cast_fp16)[name = tensor("linear_12_cast_fp16")]; - tensor var_544 = const()[name = tensor("op_544"), val = tensor([1, -1, 8, 64])]; - tensor q_7_cast_fp16 = reshape(shape = var_544, x = linear_12_cast_fp16)[name = tensor("q_7_cast_fp16")]; - tensor module_layers_1_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_1_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(20116096)))]; - tensor module_layers_1_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_1_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(20640448)))]; - tensor linear_13_cast_fp16 = linear(bias = module_layers_1_self_attn_linear_k_bias_to_fp16, weight = module_layers_1_self_attn_linear_k_weight_to_fp16, x = query_3_cast_fp16)[name = tensor("linear_13_cast_fp16")]; - tensor var_549 = const()[name = tensor("op_549"), val = tensor([1, -1, 8, 64])]; - tensor k_5_cast_fp16 = reshape(shape = var_549, x = linear_13_cast_fp16)[name = tensor("k_5_cast_fp16")]; - tensor module_layers_1_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_1_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(20641536)))]; - tensor module_layers_1_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_1_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21165888)))]; - tensor linear_14_cast_fp16 = linear(bias = module_layers_1_self_attn_linear_v_bias_to_fp16, weight = module_layers_1_self_attn_linear_v_weight_to_fp16, x = query_3_cast_fp16)[name = tensor("linear_14_cast_fp16")]; - tensor var_554 = const()[name = tensor("op_554"), val = tensor([1, -1, 8, 64])]; - tensor v_3_cast_fp16 = reshape(shape = var_554, x = linear_14_cast_fp16)[name = tensor("v_3_cast_fp16")]; - tensor value_5_perm_0 = const()[name = tensor("value_5_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_1_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_1_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21166976)))]; - tensor var_566_cast_fp16 = add(x = q_7_cast_fp16, y = module_layers_1_self_attn_pos_bias_u_to_fp16)[name = tensor("op_566_cast_fp16")]; - tensor module_layers_1_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_1_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21168064)))]; - tensor var_568_cast_fp16 = add(x = q_7_cast_fp16, y = module_layers_1_self_attn_pos_bias_v_to_fp16)[name = tensor("op_568_cast_fp16")]; - tensor q_with_bias_v_3_perm_0 = const()[name = tensor("q_with_bias_v_3_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_29_transpose_x_0 = const()[name = tensor("x_29_transpose_x_0"), val = tensor(false)]; - tensor x_29_transpose_y_0 = const()[name = tensor("x_29_transpose_y_0"), val = tensor(false)]; - tensor var_570_to_fp16 = const()[name = tensor("op_570_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21169152)))]; - tensor q_with_bias_v_3_cast_fp16 = transpose(perm = q_with_bias_v_3_perm_0, x = var_568_cast_fp16)[name = tensor("transpose_196")]; - tensor x_29_cast_fp16 = matmul(transpose_x = x_29_transpose_x_0, transpose_y = x_29_transpose_y_0, x = q_with_bias_v_3_cast_fp16, y = var_570_to_fp16)[name = tensor("x_29_cast_fp16")]; - tensor x_31_pad_0 = const()[name = tensor("x_31_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_31_mode_0 = const()[name = tensor("x_31_mode_0"), val = tensor("constant")]; - tensor const_80_to_fp16 = const()[name = tensor("const_80_to_fp16"), val = tensor(0x0p+0)]; - tensor x_31_cast_fp16 = pad(constant_val = const_80_to_fp16, mode = x_31_mode_0, pad = x_31_pad_0, x = x_29_cast_fp16)[name = tensor("x_31_cast_fp16")]; - tensor var_578 = const()[name = tensor("op_578"), val = tensor([1, 8, -1, 188])]; - tensor x_33_cast_fp16 = reshape(shape = var_578, x = x_31_cast_fp16)[name = tensor("x_33_cast_fp16")]; - tensor var_582_begin_0 = const()[name = tensor("op_582_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_582_end_0 = const()[name = tensor("op_582_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_582_end_mask_0 = const()[name = tensor("op_582_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_582_cast_fp16 = slice_by_index(begin = var_582_begin_0, end = var_582_end_0, end_mask = var_582_end_mask_0, x = x_33_cast_fp16)[name = tensor("op_582_cast_fp16")]; - tensor var_583 = const()[name = tensor("op_583"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_5_cast_fp16 = reshape(shape = var_583, x = var_582_cast_fp16)[name = tensor("matrix_bd_5_cast_fp16")]; - tensor matrix_ac_3_transpose_x_0 = const()[name = tensor("matrix_ac_3_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_3_transpose_y_0 = const()[name = tensor("matrix_ac_3_transpose_y_0"), val = tensor(false)]; - tensor transpose_53_perm_0 = const()[name = tensor("transpose_53_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_54_perm_0 = const()[name = tensor("transpose_54_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_54 = transpose(perm = transpose_54_perm_0, x = k_5_cast_fp16)[name = tensor("transpose_194")]; - tensor transpose_53 = transpose(perm = transpose_53_perm_0, x = var_566_cast_fp16)[name = tensor("transpose_195")]; - tensor matrix_ac_3_cast_fp16 = matmul(transpose_x = matrix_ac_3_transpose_x_0, transpose_y = matrix_ac_3_transpose_y_0, x = transpose_53, y = transpose_54)[name = tensor("matrix_ac_3_cast_fp16")]; - tensor matrix_bd_7_begin_0 = const()[name = tensor("matrix_bd_7_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_7_end_0 = const()[name = tensor("matrix_bd_7_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_7_end_mask_0 = const()[name = tensor("matrix_bd_7_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_7_cast_fp16 = slice_by_index(begin = matrix_bd_7_begin_0, end = matrix_bd_7_end_0, end_mask = matrix_bd_7_end_mask_0, x = matrix_bd_5_cast_fp16)[name = tensor("matrix_bd_7_cast_fp16")]; - tensor var_592_cast_fp16 = add(x = matrix_ac_3_cast_fp16, y = matrix_bd_7_cast_fp16)[name = tensor("op_592_cast_fp16")]; - tensor _inversed_scores_5_y_0_to_fp16 = const()[name = tensor("_inversed_scores_5_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_5_cast_fp16 = mul(x = var_592_cast_fp16, y = _inversed_scores_5_y_0_to_fp16)[name = tensor("_inversed_scores_5_cast_fp16")]; - tensor scores_7_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_5_cast_fp16, cond = mask_11)[name = tensor("scores_7_cast_fp16")]; - tensor var_598_cast_fp16 = softmax(axis = var_23, x = scores_7_cast_fp16)[name = tensor("op_598_cast_fp16")]; - tensor input_85_cast_fp16 = select(a = var_11_to_fp16, b = var_598_cast_fp16, cond = mask_11)[name = tensor("input_85_cast_fp16")]; - tensor x_35_transpose_x_0 = const()[name = tensor("x_35_transpose_x_0"), val = tensor(false)]; - tensor x_35_transpose_y_0 = const()[name = tensor("x_35_transpose_y_0"), val = tensor(false)]; - tensor value_5_cast_fp16 = transpose(perm = value_5_perm_0, x = v_3_cast_fp16)[name = tensor("transpose_197")]; - tensor x_35_cast_fp16 = matmul(transpose_x = x_35_transpose_x_0, transpose_y = x_35_transpose_y_0, x = input_85_cast_fp16, y = value_5_cast_fp16)[name = tensor("x_35_cast_fp16")]; - tensor var_602_perm_0 = const()[name = tensor("op_602_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_603 = const()[name = tensor("op_603"), val = tensor([1, -1, 512])]; - tensor var_602_cast_fp16 = transpose(perm = var_602_perm_0, x = x_35_cast_fp16)[name = tensor("transpose_193")]; - tensor input_87_cast_fp16 = reshape(shape = var_603, x = var_602_cast_fp16)[name = tensor("input_87_cast_fp16")]; - tensor module_layers_1_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_1_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21553216)))]; - tensor module_layers_1_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_1_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22077568)))]; - tensor linear_16_cast_fp16 = linear(bias = module_layers_1_self_attn_linear_out_bias_to_fp16, weight = module_layers_1_self_attn_linear_out_weight_to_fp16, x = input_87_cast_fp16)[name = tensor("linear_16_cast_fp16")]; - tensor input_91_cast_fp16 = add(x = input_83_cast_fp16, y = linear_16_cast_fp16)[name = tensor("input_91_cast_fp16")]; - tensor x_39_axes_0 = const()[name = tensor("x_39_axes_0"), val = tensor([-1])]; - tensor module_layers_1_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_1_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22078656)))]; - tensor module_layers_1_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_1_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22079744)))]; - tensor x_39_cast_fp16 = layer_norm(axes = x_39_axes_0, beta = module_layers_1_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_1_norm_conv_weight_to_fp16, x = input_91_cast_fp16)[name = tensor("x_39_cast_fp16")]; - tensor input_93_perm_0 = const()[name = tensor("input_93_perm_0"), val = tensor([0, 2, 1])]; - tensor input_95_pad_type_0 = const()[name = tensor("input_95_pad_type_0"), val = tensor("valid")]; - tensor input_95_strides_0 = const()[name = tensor("input_95_strides_0"), val = tensor([1])]; - tensor input_95_pad_0 = const()[name = tensor("input_95_pad_0"), val = tensor([0, 0])]; - tensor input_95_dilations_0 = const()[name = tensor("input_95_dilations_0"), val = tensor([1])]; - tensor input_95_groups_0 = const()[name = tensor("input_95_groups_0"), val = tensor(1)]; - tensor module_layers_1_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_1_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22080832)))]; - tensor module_layers_1_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_1_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23129472)))]; - tensor input_93_cast_fp16 = transpose(perm = input_93_perm_0, x = x_39_cast_fp16)[name = tensor("transpose_192")]; - tensor input_95_cast_fp16 = conv(bias = module_layers_1_conv_pointwise_conv1_bias_to_fp16, dilations = input_95_dilations_0, groups = input_95_groups_0, pad = input_95_pad_0, pad_type = input_95_pad_type_0, strides = input_95_strides_0, weight = module_layers_1_conv_pointwise_conv1_weight_to_fp16, x = input_93_cast_fp16)[name = tensor("input_95_cast_fp16")]; - tensor x_41_split_num_splits_0 = const()[name = tensor("x_41_split_num_splits_0"), val = tensor(2)]; - tensor x_41_split_axis_0 = const()[name = tensor("x_41_split_axis_0"), val = tensor(1)]; - tensor x_41_split_cast_fp16_0, tensor x_41_split_cast_fp16_1 = split(axis = x_41_split_axis_0, num_splits = x_41_split_num_splits_0, x = input_95_cast_fp16)[name = tensor("x_41_split_cast_fp16")]; - tensor x_41_split_1_sigmoid_cast_fp16 = sigmoid(x = x_41_split_cast_fp16_1)[name = tensor("x_41_split_1_sigmoid_cast_fp16")]; - tensor x_41_cast_fp16 = mul(x = x_41_split_cast_fp16_0, y = x_41_split_1_sigmoid_cast_fp16)[name = tensor("x_41_cast_fp16")]; - tensor input_97_cast_fp16 = select(a = var_11_to_fp16, b = x_41_cast_fp16, cond = var_453)[name = tensor("input_97_cast_fp16")]; - tensor input_99_pad_0 = const()[name = tensor("input_99_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_99_mode_0 = const()[name = tensor("input_99_mode_0"), val = tensor("constant")]; - tensor const_83_to_fp16 = const()[name = tensor("const_83_to_fp16"), val = tensor(0x0p+0)]; - tensor input_99_cast_fp16 = pad(constant_val = const_83_to_fp16, mode = input_99_mode_0, pad = input_99_pad_0, x = input_97_cast_fp16)[name = tensor("input_99_cast_fp16")]; - tensor input_101_pad_type_0 = const()[name = tensor("input_101_pad_type_0"), val = tensor("valid")]; - tensor input_101_groups_0 = const()[name = tensor("input_101_groups_0"), val = tensor(512)]; - tensor input_101_strides_0 = const()[name = tensor("input_101_strides_0"), val = tensor([1])]; - tensor input_101_pad_0 = const()[name = tensor("input_101_pad_0"), val = tensor([0, 0])]; - tensor input_101_dilations_0 = const()[name = tensor("input_101_dilations_0"), val = tensor([1])]; - tensor const_236_to_fp16 = const()[name = tensor("const_236_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23131584)))]; - tensor const_237_to_fp16 = const()[name = tensor("const_237_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23140864)))]; - tensor input_103_cast_fp16 = conv(bias = const_237_to_fp16, dilations = input_101_dilations_0, groups = input_101_groups_0, pad = input_101_pad_0, pad_type = input_101_pad_type_0, strides = input_101_strides_0, weight = const_236_to_fp16, x = input_99_cast_fp16)[name = tensor("input_103_cast_fp16")]; - tensor input_105_cast_fp16 = silu(x = input_103_cast_fp16)[name = tensor("input_105_cast_fp16")]; - tensor x_43_pad_type_0 = const()[name = tensor("x_43_pad_type_0"), val = tensor("valid")]; - tensor x_43_strides_0 = const()[name = tensor("x_43_strides_0"), val = tensor([1])]; - tensor x_43_pad_0 = const()[name = tensor("x_43_pad_0"), val = tensor([0, 0])]; - tensor x_43_dilations_0 = const()[name = tensor("x_43_dilations_0"), val = tensor([1])]; - tensor x_43_groups_0 = const()[name = tensor("x_43_groups_0"), val = tensor(1)]; - tensor module_layers_1_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_1_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23141952)))]; - tensor module_layers_1_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_1_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23666304)))]; - tensor x_43_cast_fp16 = conv(bias = module_layers_1_conv_pointwise_conv2_bias_to_fp16, dilations = x_43_dilations_0, groups = x_43_groups_0, pad = x_43_pad_0, pad_type = x_43_pad_type_0, strides = x_43_strides_0, weight = module_layers_1_conv_pointwise_conv2_weight_to_fp16, x = input_105_cast_fp16)[name = tensor("x_43_cast_fp16")]; - tensor input_107_perm_0 = const()[name = tensor("input_107_perm_0"), val = tensor([0, 2, 1])]; - tensor input_107_cast_fp16 = transpose(perm = input_107_perm_0, x = x_43_cast_fp16)[name = tensor("transpose_191")]; - tensor input_109_cast_fp16 = add(x = input_91_cast_fp16, y = input_107_cast_fp16)[name = tensor("input_109_cast_fp16")]; - tensor input_111_axes_0 = const()[name = tensor("input_111_axes_0"), val = tensor([-1])]; - tensor module_layers_1_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_1_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23667392)))]; - tensor module_layers_1_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_1_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23668480)))]; - tensor input_111_cast_fp16 = layer_norm(axes = input_111_axes_0, beta = module_layers_1_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_1_norm_feed_forward2_weight_to_fp16, x = input_109_cast_fp16)[name = tensor("input_111_cast_fp16")]; - tensor module_layers_1_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_1_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23669568)))]; - tensor module_layers_1_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_1_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(25766784)))]; - tensor linear_17_cast_fp16 = linear(bias = module_layers_1_feed_forward2_linear1_bias_to_fp16, weight = module_layers_1_feed_forward2_linear1_weight_to_fp16, x = input_111_cast_fp16)[name = tensor("linear_17_cast_fp16")]; - tensor input_115_cast_fp16 = silu(x = linear_17_cast_fp16)[name = tensor("input_115_cast_fp16")]; - tensor module_layers_1_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_1_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(25770944)))]; - tensor module_layers_1_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_1_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27868160)))]; - tensor linear_18_cast_fp16 = linear(bias = module_layers_1_feed_forward2_linear2_bias_to_fp16, weight = module_layers_1_feed_forward2_linear2_weight_to_fp16, x = input_115_cast_fp16)[name = tensor("linear_18_cast_fp16")]; - tensor var_669_to_fp16 = const()[name = tensor("op_669_to_fp16"), val = tensor(0x1p-1)]; - tensor var_670_cast_fp16 = mul(x = linear_18_cast_fp16, y = var_669_to_fp16)[name = tensor("op_670_cast_fp16")]; - tensor input_121_cast_fp16 = add(x = input_109_cast_fp16, y = var_670_cast_fp16)[name = tensor("input_121_cast_fp16")]; - tensor input_123_axes_0 = const()[name = tensor("input_123_axes_0"), val = tensor([-1])]; - tensor module_layers_1_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_1_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27869248)))]; - tensor module_layers_1_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_1_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27870336)))]; - tensor input_123_cast_fp16 = layer_norm(axes = input_123_axes_0, beta = module_layers_1_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_1_norm_out_weight_to_fp16, x = input_121_cast_fp16)[name = tensor("input_123_cast_fp16")]; - tensor input_125_axes_0 = const()[name = tensor("input_125_axes_0"), val = tensor([-1])]; - tensor module_layers_2_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_2_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27871424)))]; - tensor module_layers_2_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_2_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27872512)))]; - tensor input_125_cast_fp16 = layer_norm(axes = input_125_axes_0, beta = module_layers_2_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_2_norm_feed_forward1_weight_to_fp16, x = input_123_cast_fp16)[name = tensor("input_125_cast_fp16")]; - tensor module_layers_2_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_2_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27873600)))]; - tensor module_layers_2_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_2_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(29970816)))]; - tensor linear_19_cast_fp16 = linear(bias = module_layers_2_feed_forward1_linear1_bias_to_fp16, weight = module_layers_2_feed_forward1_linear1_weight_to_fp16, x = input_125_cast_fp16)[name = tensor("linear_19_cast_fp16")]; - tensor input_129_cast_fp16 = silu(x = linear_19_cast_fp16)[name = tensor("input_129_cast_fp16")]; - tensor module_layers_2_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_2_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(29974976)))]; - tensor module_layers_2_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_2_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32072192)))]; - tensor linear_20_cast_fp16 = linear(bias = module_layers_2_feed_forward1_linear2_bias_to_fp16, weight = module_layers_2_feed_forward1_linear2_weight_to_fp16, x = input_129_cast_fp16)[name = tensor("linear_20_cast_fp16")]; - tensor var_700_to_fp16 = const()[name = tensor("op_700_to_fp16"), val = tensor(0x1p-1)]; - tensor var_701_cast_fp16 = mul(x = linear_20_cast_fp16, y = var_700_to_fp16)[name = tensor("op_701_cast_fp16")]; - tensor input_135_cast_fp16 = add(x = input_123_cast_fp16, y = var_701_cast_fp16)[name = tensor("input_135_cast_fp16")]; - tensor query_5_axes_0 = const()[name = tensor("query_5_axes_0"), val = tensor([-1])]; - tensor module_layers_2_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_2_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32073280)))]; - tensor module_layers_2_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_2_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32074368)))]; - tensor query_5_cast_fp16 = layer_norm(axes = query_5_axes_0, beta = module_layers_2_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_2_norm_self_att_weight_to_fp16, x = input_135_cast_fp16)[name = tensor("query_5_cast_fp16")]; - tensor module_layers_2_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_2_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32075456)))]; - tensor module_layers_2_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_2_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32599808)))]; - tensor linear_21_cast_fp16 = linear(bias = module_layers_2_self_attn_linear_q_bias_to_fp16, weight = module_layers_2_self_attn_linear_q_weight_to_fp16, x = query_5_cast_fp16)[name = tensor("linear_21_cast_fp16")]; - tensor var_718 = const()[name = tensor("op_718"), val = tensor([1, -1, 8, 64])]; - tensor q_13_cast_fp16 = reshape(shape = var_718, x = linear_21_cast_fp16)[name = tensor("q_13_cast_fp16")]; - tensor module_layers_2_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_2_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32600896)))]; - tensor module_layers_2_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_2_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33125248)))]; - tensor linear_22_cast_fp16 = linear(bias = module_layers_2_self_attn_linear_k_bias_to_fp16, weight = module_layers_2_self_attn_linear_k_weight_to_fp16, x = query_5_cast_fp16)[name = tensor("linear_22_cast_fp16")]; - tensor var_723 = const()[name = tensor("op_723"), val = tensor([1, -1, 8, 64])]; - tensor k_9_cast_fp16 = reshape(shape = var_723, x = linear_22_cast_fp16)[name = tensor("k_9_cast_fp16")]; - tensor module_layers_2_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_2_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33126336)))]; - tensor module_layers_2_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_2_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33650688)))]; - tensor linear_23_cast_fp16 = linear(bias = module_layers_2_self_attn_linear_v_bias_to_fp16, weight = module_layers_2_self_attn_linear_v_weight_to_fp16, x = query_5_cast_fp16)[name = tensor("linear_23_cast_fp16")]; - tensor var_728 = const()[name = tensor("op_728"), val = tensor([1, -1, 8, 64])]; - tensor v_5_cast_fp16 = reshape(shape = var_728, x = linear_23_cast_fp16)[name = tensor("v_5_cast_fp16")]; - tensor value_7_perm_0 = const()[name = tensor("value_7_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_2_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_2_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33651776)))]; - tensor var_740_cast_fp16 = add(x = q_13_cast_fp16, y = module_layers_2_self_attn_pos_bias_u_to_fp16)[name = tensor("op_740_cast_fp16")]; - tensor module_layers_2_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_2_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33652864)))]; - tensor var_742_cast_fp16 = add(x = q_13_cast_fp16, y = module_layers_2_self_attn_pos_bias_v_to_fp16)[name = tensor("op_742_cast_fp16")]; - tensor q_with_bias_v_5_perm_0 = const()[name = tensor("q_with_bias_v_5_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_51_transpose_x_0 = const()[name = tensor("x_51_transpose_x_0"), val = tensor(false)]; - tensor x_51_transpose_y_0 = const()[name = tensor("x_51_transpose_y_0"), val = tensor(false)]; - tensor var_744_to_fp16 = const()[name = tensor("op_744_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33653952)))]; - tensor q_with_bias_v_5_cast_fp16 = transpose(perm = q_with_bias_v_5_perm_0, x = var_742_cast_fp16)[name = tensor("transpose_189")]; - tensor x_51_cast_fp16 = matmul(transpose_x = x_51_transpose_x_0, transpose_y = x_51_transpose_y_0, x = q_with_bias_v_5_cast_fp16, y = var_744_to_fp16)[name = tensor("x_51_cast_fp16")]; - tensor x_53_pad_0 = const()[name = tensor("x_53_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_53_mode_0 = const()[name = tensor("x_53_mode_0"), val = tensor("constant")]; - tensor const_90_to_fp16 = const()[name = tensor("const_90_to_fp16"), val = tensor(0x0p+0)]; - tensor x_53_cast_fp16 = pad(constant_val = const_90_to_fp16, mode = x_53_mode_0, pad = x_53_pad_0, x = x_51_cast_fp16)[name = tensor("x_53_cast_fp16")]; - tensor var_752 = const()[name = tensor("op_752"), val = tensor([1, 8, -1, 188])]; - tensor x_55_cast_fp16 = reshape(shape = var_752, x = x_53_cast_fp16)[name = tensor("x_55_cast_fp16")]; - tensor var_756_begin_0 = const()[name = tensor("op_756_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_756_end_0 = const()[name = tensor("op_756_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_756_end_mask_0 = const()[name = tensor("op_756_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_756_cast_fp16 = slice_by_index(begin = var_756_begin_0, end = var_756_end_0, end_mask = var_756_end_mask_0, x = x_55_cast_fp16)[name = tensor("op_756_cast_fp16")]; - tensor var_757 = const()[name = tensor("op_757"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_9_cast_fp16 = reshape(shape = var_757, x = var_756_cast_fp16)[name = tensor("matrix_bd_9_cast_fp16")]; - tensor matrix_ac_5_transpose_x_0 = const()[name = tensor("matrix_ac_5_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_5_transpose_y_0 = const()[name = tensor("matrix_ac_5_transpose_y_0"), val = tensor(false)]; - tensor transpose_55_perm_0 = const()[name = tensor("transpose_55_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_56_perm_0 = const()[name = tensor("transpose_56_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_56 = transpose(perm = transpose_56_perm_0, x = k_9_cast_fp16)[name = tensor("transpose_187")]; - tensor transpose_55 = transpose(perm = transpose_55_perm_0, x = var_740_cast_fp16)[name = tensor("transpose_188")]; - tensor matrix_ac_5_cast_fp16 = matmul(transpose_x = matrix_ac_5_transpose_x_0, transpose_y = matrix_ac_5_transpose_y_0, x = transpose_55, y = transpose_56)[name = tensor("matrix_ac_5_cast_fp16")]; - tensor matrix_bd_11_begin_0 = const()[name = tensor("matrix_bd_11_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_11_end_0 = const()[name = tensor("matrix_bd_11_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_11_end_mask_0 = const()[name = tensor("matrix_bd_11_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_11_cast_fp16 = slice_by_index(begin = matrix_bd_11_begin_0, end = matrix_bd_11_end_0, end_mask = matrix_bd_11_end_mask_0, x = matrix_bd_9_cast_fp16)[name = tensor("matrix_bd_11_cast_fp16")]; - tensor var_766_cast_fp16 = add(x = matrix_ac_5_cast_fp16, y = matrix_bd_11_cast_fp16)[name = tensor("op_766_cast_fp16")]; - tensor _inversed_scores_9_y_0_to_fp16 = const()[name = tensor("_inversed_scores_9_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_9_cast_fp16 = mul(x = var_766_cast_fp16, y = _inversed_scores_9_y_0_to_fp16)[name = tensor("_inversed_scores_9_cast_fp16")]; - tensor scores_11_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_9_cast_fp16, cond = mask_11)[name = tensor("scores_11_cast_fp16")]; - tensor var_772_cast_fp16 = softmax(axis = var_23, x = scores_11_cast_fp16)[name = tensor("op_772_cast_fp16")]; - tensor input_137_cast_fp16 = select(a = var_11_to_fp16, b = var_772_cast_fp16, cond = mask_11)[name = tensor("input_137_cast_fp16")]; - tensor x_57_transpose_x_0 = const()[name = tensor("x_57_transpose_x_0"), val = tensor(false)]; - tensor x_57_transpose_y_0 = const()[name = tensor("x_57_transpose_y_0"), val = tensor(false)]; - tensor value_7_cast_fp16 = transpose(perm = value_7_perm_0, x = v_5_cast_fp16)[name = tensor("transpose_190")]; - tensor x_57_cast_fp16 = matmul(transpose_x = x_57_transpose_x_0, transpose_y = x_57_transpose_y_0, x = input_137_cast_fp16, y = value_7_cast_fp16)[name = tensor("x_57_cast_fp16")]; - tensor var_776_perm_0 = const()[name = tensor("op_776_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_777 = const()[name = tensor("op_777"), val = tensor([1, -1, 512])]; - tensor var_776_cast_fp16 = transpose(perm = var_776_perm_0, x = x_57_cast_fp16)[name = tensor("transpose_186")]; - tensor input_139_cast_fp16 = reshape(shape = var_777, x = var_776_cast_fp16)[name = tensor("input_139_cast_fp16")]; - tensor module_layers_2_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_2_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34038016)))]; - tensor module_layers_2_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_2_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34562368)))]; - tensor linear_25_cast_fp16 = linear(bias = module_layers_2_self_attn_linear_out_bias_to_fp16, weight = module_layers_2_self_attn_linear_out_weight_to_fp16, x = input_139_cast_fp16)[name = tensor("linear_25_cast_fp16")]; - tensor input_143_cast_fp16 = add(x = input_135_cast_fp16, y = linear_25_cast_fp16)[name = tensor("input_143_cast_fp16")]; - tensor x_61_axes_0 = const()[name = tensor("x_61_axes_0"), val = tensor([-1])]; - tensor module_layers_2_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_2_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34563456)))]; - tensor module_layers_2_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_2_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34564544)))]; - tensor x_61_cast_fp16 = layer_norm(axes = x_61_axes_0, beta = module_layers_2_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_2_norm_conv_weight_to_fp16, x = input_143_cast_fp16)[name = tensor("x_61_cast_fp16")]; - tensor input_145_perm_0 = const()[name = tensor("input_145_perm_0"), val = tensor([0, 2, 1])]; - tensor input_147_pad_type_0 = const()[name = tensor("input_147_pad_type_0"), val = tensor("valid")]; - tensor input_147_strides_0 = const()[name = tensor("input_147_strides_0"), val = tensor([1])]; - tensor input_147_pad_0 = const()[name = tensor("input_147_pad_0"), val = tensor([0, 0])]; - tensor input_147_dilations_0 = const()[name = tensor("input_147_dilations_0"), val = tensor([1])]; - tensor input_147_groups_0 = const()[name = tensor("input_147_groups_0"), val = tensor(1)]; - tensor module_layers_2_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_2_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34565632)))]; - tensor module_layers_2_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_2_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35614272)))]; - tensor input_145_cast_fp16 = transpose(perm = input_145_perm_0, x = x_61_cast_fp16)[name = tensor("transpose_185")]; - tensor input_147_cast_fp16 = conv(bias = module_layers_2_conv_pointwise_conv1_bias_to_fp16, dilations = input_147_dilations_0, groups = input_147_groups_0, pad = input_147_pad_0, pad_type = input_147_pad_type_0, strides = input_147_strides_0, weight = module_layers_2_conv_pointwise_conv1_weight_to_fp16, x = input_145_cast_fp16)[name = tensor("input_147_cast_fp16")]; - tensor x_63_split_num_splits_0 = const()[name = tensor("x_63_split_num_splits_0"), val = tensor(2)]; - tensor x_63_split_axis_0 = const()[name = tensor("x_63_split_axis_0"), val = tensor(1)]; - tensor x_63_split_cast_fp16_0, tensor x_63_split_cast_fp16_1 = split(axis = x_63_split_axis_0, num_splits = x_63_split_num_splits_0, x = input_147_cast_fp16)[name = tensor("x_63_split_cast_fp16")]; - tensor x_63_split_1_sigmoid_cast_fp16 = sigmoid(x = x_63_split_cast_fp16_1)[name = tensor("x_63_split_1_sigmoid_cast_fp16")]; - tensor x_63_cast_fp16 = mul(x = x_63_split_cast_fp16_0, y = x_63_split_1_sigmoid_cast_fp16)[name = tensor("x_63_cast_fp16")]; - tensor input_149_cast_fp16 = select(a = var_11_to_fp16, b = x_63_cast_fp16, cond = var_453)[name = tensor("input_149_cast_fp16")]; - tensor input_151_pad_0 = const()[name = tensor("input_151_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_151_mode_0 = const()[name = tensor("input_151_mode_0"), val = tensor("constant")]; - tensor const_93_to_fp16 = const()[name = tensor("const_93_to_fp16"), val = tensor(0x0p+0)]; - tensor input_151_cast_fp16 = pad(constant_val = const_93_to_fp16, mode = input_151_mode_0, pad = input_151_pad_0, x = input_149_cast_fp16)[name = tensor("input_151_cast_fp16")]; - tensor input_153_pad_type_0 = const()[name = tensor("input_153_pad_type_0"), val = tensor("valid")]; - tensor input_153_groups_0 = const()[name = tensor("input_153_groups_0"), val = tensor(512)]; - tensor input_153_strides_0 = const()[name = tensor("input_153_strides_0"), val = tensor([1])]; - tensor input_153_pad_0 = const()[name = tensor("input_153_pad_0"), val = tensor([0, 0])]; - tensor input_153_dilations_0 = const()[name = tensor("input_153_dilations_0"), val = tensor([1])]; - tensor const_238_to_fp16 = const()[name = tensor("const_238_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35616384)))]; - tensor const_239_to_fp16 = const()[name = tensor("const_239_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35625664)))]; - tensor input_155_cast_fp16 = conv(bias = const_239_to_fp16, dilations = input_153_dilations_0, groups = input_153_groups_0, pad = input_153_pad_0, pad_type = input_153_pad_type_0, strides = input_153_strides_0, weight = const_238_to_fp16, x = input_151_cast_fp16)[name = tensor("input_155_cast_fp16")]; - tensor input_157_cast_fp16 = silu(x = input_155_cast_fp16)[name = tensor("input_157_cast_fp16")]; - tensor x_65_pad_type_0 = const()[name = tensor("x_65_pad_type_0"), val = tensor("valid")]; - tensor x_65_strides_0 = const()[name = tensor("x_65_strides_0"), val = tensor([1])]; - tensor x_65_pad_0 = const()[name = tensor("x_65_pad_0"), val = tensor([0, 0])]; - tensor x_65_dilations_0 = const()[name = tensor("x_65_dilations_0"), val = tensor([1])]; - tensor x_65_groups_0 = const()[name = tensor("x_65_groups_0"), val = tensor(1)]; - tensor module_layers_2_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_2_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35626752)))]; - tensor module_layers_2_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_2_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36151104)))]; - tensor x_65_cast_fp16 = conv(bias = module_layers_2_conv_pointwise_conv2_bias_to_fp16, dilations = x_65_dilations_0, groups = x_65_groups_0, pad = x_65_pad_0, pad_type = x_65_pad_type_0, strides = x_65_strides_0, weight = module_layers_2_conv_pointwise_conv2_weight_to_fp16, x = input_157_cast_fp16)[name = tensor("x_65_cast_fp16")]; - tensor input_159_perm_0 = const()[name = tensor("input_159_perm_0"), val = tensor([0, 2, 1])]; - tensor input_159_cast_fp16 = transpose(perm = input_159_perm_0, x = x_65_cast_fp16)[name = tensor("transpose_184")]; - tensor input_161_cast_fp16 = add(x = input_143_cast_fp16, y = input_159_cast_fp16)[name = tensor("input_161_cast_fp16")]; - tensor input_163_axes_0 = const()[name = tensor("input_163_axes_0"), val = tensor([-1])]; - tensor module_layers_2_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_2_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36152192)))]; - tensor module_layers_2_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_2_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36153280)))]; - tensor input_163_cast_fp16 = layer_norm(axes = input_163_axes_0, beta = module_layers_2_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_2_norm_feed_forward2_weight_to_fp16, x = input_161_cast_fp16)[name = tensor("input_163_cast_fp16")]; - tensor module_layers_2_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_2_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36154368)))]; - tensor module_layers_2_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_2_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(38251584)))]; - tensor linear_26_cast_fp16 = linear(bias = module_layers_2_feed_forward2_linear1_bias_to_fp16, weight = module_layers_2_feed_forward2_linear1_weight_to_fp16, x = input_163_cast_fp16)[name = tensor("linear_26_cast_fp16")]; - tensor input_167_cast_fp16 = silu(x = linear_26_cast_fp16)[name = tensor("input_167_cast_fp16")]; - tensor module_layers_2_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_2_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(38255744)))]; - tensor module_layers_2_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_2_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40352960)))]; - tensor linear_27_cast_fp16 = linear(bias = module_layers_2_feed_forward2_linear2_bias_to_fp16, weight = module_layers_2_feed_forward2_linear2_weight_to_fp16, x = input_167_cast_fp16)[name = tensor("linear_27_cast_fp16")]; - tensor var_843_to_fp16 = const()[name = tensor("op_843_to_fp16"), val = tensor(0x1p-1)]; - tensor var_844_cast_fp16 = mul(x = linear_27_cast_fp16, y = var_843_to_fp16)[name = tensor("op_844_cast_fp16")]; - tensor input_173_cast_fp16 = add(x = input_161_cast_fp16, y = var_844_cast_fp16)[name = tensor("input_173_cast_fp16")]; - tensor input_175_axes_0 = const()[name = tensor("input_175_axes_0"), val = tensor([-1])]; - tensor module_layers_2_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_2_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40354048)))]; - tensor module_layers_2_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_2_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40355136)))]; - tensor input_175_cast_fp16 = layer_norm(axes = input_175_axes_0, beta = module_layers_2_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_2_norm_out_weight_to_fp16, x = input_173_cast_fp16)[name = tensor("input_175_cast_fp16")]; - tensor input_177_axes_0 = const()[name = tensor("input_177_axes_0"), val = tensor([-1])]; - tensor module_layers_3_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_3_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40356224)))]; - tensor module_layers_3_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_3_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40357312)))]; - tensor input_177_cast_fp16 = layer_norm(axes = input_177_axes_0, beta = module_layers_3_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_3_norm_feed_forward1_weight_to_fp16, x = input_175_cast_fp16)[name = tensor("input_177_cast_fp16")]; - tensor module_layers_3_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_3_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40358400)))]; - tensor module_layers_3_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_3_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(42455616)))]; - tensor linear_28_cast_fp16 = linear(bias = module_layers_3_feed_forward1_linear1_bias_to_fp16, weight = module_layers_3_feed_forward1_linear1_weight_to_fp16, x = input_177_cast_fp16)[name = tensor("linear_28_cast_fp16")]; - tensor input_181_cast_fp16 = silu(x = linear_28_cast_fp16)[name = tensor("input_181_cast_fp16")]; - tensor module_layers_3_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_3_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(42459776)))]; - tensor module_layers_3_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_3_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44556992)))]; - tensor linear_29_cast_fp16 = linear(bias = module_layers_3_feed_forward1_linear2_bias_to_fp16, weight = module_layers_3_feed_forward1_linear2_weight_to_fp16, x = input_181_cast_fp16)[name = tensor("linear_29_cast_fp16")]; - tensor var_874_to_fp16 = const()[name = tensor("op_874_to_fp16"), val = tensor(0x1p-1)]; - tensor var_875_cast_fp16 = mul(x = linear_29_cast_fp16, y = var_874_to_fp16)[name = tensor("op_875_cast_fp16")]; - tensor input_187_cast_fp16 = add(x = input_175_cast_fp16, y = var_875_cast_fp16)[name = tensor("input_187_cast_fp16")]; - tensor query_7_axes_0 = const()[name = tensor("query_7_axes_0"), val = tensor([-1])]; - tensor module_layers_3_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_3_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44558080)))]; - tensor module_layers_3_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_3_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44559168)))]; - tensor query_7_cast_fp16 = layer_norm(axes = query_7_axes_0, beta = module_layers_3_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_3_norm_self_att_weight_to_fp16, x = input_187_cast_fp16)[name = tensor("query_7_cast_fp16")]; - tensor module_layers_3_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_3_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44560256)))]; - tensor module_layers_3_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_3_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45084608)))]; - tensor linear_30_cast_fp16 = linear(bias = module_layers_3_self_attn_linear_q_bias_to_fp16, weight = module_layers_3_self_attn_linear_q_weight_to_fp16, x = query_7_cast_fp16)[name = tensor("linear_30_cast_fp16")]; - tensor var_892 = const()[name = tensor("op_892"), val = tensor([1, -1, 8, 64])]; - tensor q_19_cast_fp16 = reshape(shape = var_892, x = linear_30_cast_fp16)[name = tensor("q_19_cast_fp16")]; - tensor module_layers_3_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_3_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45085696)))]; - tensor module_layers_3_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_3_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45610048)))]; - tensor linear_31_cast_fp16 = linear(bias = module_layers_3_self_attn_linear_k_bias_to_fp16, weight = module_layers_3_self_attn_linear_k_weight_to_fp16, x = query_7_cast_fp16)[name = tensor("linear_31_cast_fp16")]; - tensor var_897 = const()[name = tensor("op_897"), val = tensor([1, -1, 8, 64])]; - tensor k_13_cast_fp16 = reshape(shape = var_897, x = linear_31_cast_fp16)[name = tensor("k_13_cast_fp16")]; - tensor module_layers_3_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_3_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45611136)))]; - tensor module_layers_3_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_3_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46135488)))]; - tensor linear_32_cast_fp16 = linear(bias = module_layers_3_self_attn_linear_v_bias_to_fp16, weight = module_layers_3_self_attn_linear_v_weight_to_fp16, x = query_7_cast_fp16)[name = tensor("linear_32_cast_fp16")]; - tensor var_902 = const()[name = tensor("op_902"), val = tensor([1, -1, 8, 64])]; - tensor v_7_cast_fp16 = reshape(shape = var_902, x = linear_32_cast_fp16)[name = tensor("v_7_cast_fp16")]; - tensor value_9_perm_0 = const()[name = tensor("value_9_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_3_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_3_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46136576)))]; - tensor var_914_cast_fp16 = add(x = q_19_cast_fp16, y = module_layers_3_self_attn_pos_bias_u_to_fp16)[name = tensor("op_914_cast_fp16")]; - tensor module_layers_3_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_3_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46137664)))]; - tensor var_916_cast_fp16 = add(x = q_19_cast_fp16, y = module_layers_3_self_attn_pos_bias_v_to_fp16)[name = tensor("op_916_cast_fp16")]; - tensor q_with_bias_v_7_perm_0 = const()[name = tensor("q_with_bias_v_7_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_73_transpose_x_0 = const()[name = tensor("x_73_transpose_x_0"), val = tensor(false)]; - tensor x_73_transpose_y_0 = const()[name = tensor("x_73_transpose_y_0"), val = tensor(false)]; - tensor var_918_to_fp16 = const()[name = tensor("op_918_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46138752)))]; - tensor q_with_bias_v_7_cast_fp16 = transpose(perm = q_with_bias_v_7_perm_0, x = var_916_cast_fp16)[name = tensor("transpose_182")]; - tensor x_73_cast_fp16 = matmul(transpose_x = x_73_transpose_x_0, transpose_y = x_73_transpose_y_0, x = q_with_bias_v_7_cast_fp16, y = var_918_to_fp16)[name = tensor("x_73_cast_fp16")]; - tensor x_75_pad_0 = const()[name = tensor("x_75_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_75_mode_0 = const()[name = tensor("x_75_mode_0"), val = tensor("constant")]; - tensor const_100_to_fp16 = const()[name = tensor("const_100_to_fp16"), val = tensor(0x0p+0)]; - tensor x_75_cast_fp16 = pad(constant_val = const_100_to_fp16, mode = x_75_mode_0, pad = x_75_pad_0, x = x_73_cast_fp16)[name = tensor("x_75_cast_fp16")]; - tensor var_926 = const()[name = tensor("op_926"), val = tensor([1, 8, -1, 188])]; - tensor x_77_cast_fp16 = reshape(shape = var_926, x = x_75_cast_fp16)[name = tensor("x_77_cast_fp16")]; - tensor var_930_begin_0 = const()[name = tensor("op_930_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_930_end_0 = const()[name = tensor("op_930_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_930_end_mask_0 = const()[name = tensor("op_930_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_930_cast_fp16 = slice_by_index(begin = var_930_begin_0, end = var_930_end_0, end_mask = var_930_end_mask_0, x = x_77_cast_fp16)[name = tensor("op_930_cast_fp16")]; - tensor var_931 = const()[name = tensor("op_931"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_13_cast_fp16 = reshape(shape = var_931, x = var_930_cast_fp16)[name = tensor("matrix_bd_13_cast_fp16")]; - tensor matrix_ac_7_transpose_x_0 = const()[name = tensor("matrix_ac_7_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_7_transpose_y_0 = const()[name = tensor("matrix_ac_7_transpose_y_0"), val = tensor(false)]; - tensor transpose_57_perm_0 = const()[name = tensor("transpose_57_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_58_perm_0 = const()[name = tensor("transpose_58_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_58 = transpose(perm = transpose_58_perm_0, x = k_13_cast_fp16)[name = tensor("transpose_180")]; - tensor transpose_57 = transpose(perm = transpose_57_perm_0, x = var_914_cast_fp16)[name = tensor("transpose_181")]; - tensor matrix_ac_7_cast_fp16 = matmul(transpose_x = matrix_ac_7_transpose_x_0, transpose_y = matrix_ac_7_transpose_y_0, x = transpose_57, y = transpose_58)[name = tensor("matrix_ac_7_cast_fp16")]; - tensor matrix_bd_15_begin_0 = const()[name = tensor("matrix_bd_15_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_15_end_0 = const()[name = tensor("matrix_bd_15_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_15_end_mask_0 = const()[name = tensor("matrix_bd_15_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_15_cast_fp16 = slice_by_index(begin = matrix_bd_15_begin_0, end = matrix_bd_15_end_0, end_mask = matrix_bd_15_end_mask_0, x = matrix_bd_13_cast_fp16)[name = tensor("matrix_bd_15_cast_fp16")]; - tensor var_940_cast_fp16 = add(x = matrix_ac_7_cast_fp16, y = matrix_bd_15_cast_fp16)[name = tensor("op_940_cast_fp16")]; - tensor _inversed_scores_13_y_0_to_fp16 = const()[name = tensor("_inversed_scores_13_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_13_cast_fp16 = mul(x = var_940_cast_fp16, y = _inversed_scores_13_y_0_to_fp16)[name = tensor("_inversed_scores_13_cast_fp16")]; - tensor scores_15_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_13_cast_fp16, cond = mask_11)[name = tensor("scores_15_cast_fp16")]; - tensor var_946_cast_fp16 = softmax(axis = var_23, x = scores_15_cast_fp16)[name = tensor("op_946_cast_fp16")]; - tensor input_189_cast_fp16 = select(a = var_11_to_fp16, b = var_946_cast_fp16, cond = mask_11)[name = tensor("input_189_cast_fp16")]; - tensor x_79_transpose_x_0 = const()[name = tensor("x_79_transpose_x_0"), val = tensor(false)]; - tensor x_79_transpose_y_0 = const()[name = tensor("x_79_transpose_y_0"), val = tensor(false)]; - tensor value_9_cast_fp16 = transpose(perm = value_9_perm_0, x = v_7_cast_fp16)[name = tensor("transpose_183")]; - tensor x_79_cast_fp16 = matmul(transpose_x = x_79_transpose_x_0, transpose_y = x_79_transpose_y_0, x = input_189_cast_fp16, y = value_9_cast_fp16)[name = tensor("x_79_cast_fp16")]; - tensor var_950_perm_0 = const()[name = tensor("op_950_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_951 = const()[name = tensor("op_951"), val = tensor([1, -1, 512])]; - tensor var_950_cast_fp16 = transpose(perm = var_950_perm_0, x = x_79_cast_fp16)[name = tensor("transpose_179")]; - tensor input_191_cast_fp16 = reshape(shape = var_951, x = var_950_cast_fp16)[name = tensor("input_191_cast_fp16")]; - tensor module_layers_3_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_3_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46522816)))]; - tensor module_layers_3_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_3_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47047168)))]; - tensor linear_34_cast_fp16 = linear(bias = module_layers_3_self_attn_linear_out_bias_to_fp16, weight = module_layers_3_self_attn_linear_out_weight_to_fp16, x = input_191_cast_fp16)[name = tensor("linear_34_cast_fp16")]; - tensor input_195_cast_fp16 = add(x = input_187_cast_fp16, y = linear_34_cast_fp16)[name = tensor("input_195_cast_fp16")]; - tensor x_83_axes_0 = const()[name = tensor("x_83_axes_0"), val = tensor([-1])]; - tensor module_layers_3_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_3_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47048256)))]; - tensor module_layers_3_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_3_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47049344)))]; - tensor x_83_cast_fp16 = layer_norm(axes = x_83_axes_0, beta = module_layers_3_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_3_norm_conv_weight_to_fp16, x = input_195_cast_fp16)[name = tensor("x_83_cast_fp16")]; - tensor input_197_perm_0 = const()[name = tensor("input_197_perm_0"), val = tensor([0, 2, 1])]; - tensor input_199_pad_type_0 = const()[name = tensor("input_199_pad_type_0"), val = tensor("valid")]; - tensor input_199_strides_0 = const()[name = tensor("input_199_strides_0"), val = tensor([1])]; - tensor input_199_pad_0 = const()[name = tensor("input_199_pad_0"), val = tensor([0, 0])]; - tensor input_199_dilations_0 = const()[name = tensor("input_199_dilations_0"), val = tensor([1])]; - tensor input_199_groups_0 = const()[name = tensor("input_199_groups_0"), val = tensor(1)]; - tensor module_layers_3_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_3_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47050432)))]; - tensor module_layers_3_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_3_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48099072)))]; - tensor input_197_cast_fp16 = transpose(perm = input_197_perm_0, x = x_83_cast_fp16)[name = tensor("transpose_178")]; - tensor input_199_cast_fp16 = conv(bias = module_layers_3_conv_pointwise_conv1_bias_to_fp16, dilations = input_199_dilations_0, groups = input_199_groups_0, pad = input_199_pad_0, pad_type = input_199_pad_type_0, strides = input_199_strides_0, weight = module_layers_3_conv_pointwise_conv1_weight_to_fp16, x = input_197_cast_fp16)[name = tensor("input_199_cast_fp16")]; - tensor x_85_split_num_splits_0 = const()[name = tensor("x_85_split_num_splits_0"), val = tensor(2)]; - tensor x_85_split_axis_0 = const()[name = tensor("x_85_split_axis_0"), val = tensor(1)]; - tensor x_85_split_cast_fp16_0, tensor x_85_split_cast_fp16_1 = split(axis = x_85_split_axis_0, num_splits = x_85_split_num_splits_0, x = input_199_cast_fp16)[name = tensor("x_85_split_cast_fp16")]; - tensor x_85_split_1_sigmoid_cast_fp16 = sigmoid(x = x_85_split_cast_fp16_1)[name = tensor("x_85_split_1_sigmoid_cast_fp16")]; - tensor x_85_cast_fp16 = mul(x = x_85_split_cast_fp16_0, y = x_85_split_1_sigmoid_cast_fp16)[name = tensor("x_85_cast_fp16")]; - tensor input_201_cast_fp16 = select(a = var_11_to_fp16, b = x_85_cast_fp16, cond = var_453)[name = tensor("input_201_cast_fp16")]; - tensor input_203_pad_0 = const()[name = tensor("input_203_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_203_mode_0 = const()[name = tensor("input_203_mode_0"), val = tensor("constant")]; - tensor const_103_to_fp16 = const()[name = tensor("const_103_to_fp16"), val = tensor(0x0p+0)]; - tensor input_203_cast_fp16 = pad(constant_val = const_103_to_fp16, mode = input_203_mode_0, pad = input_203_pad_0, x = input_201_cast_fp16)[name = tensor("input_203_cast_fp16")]; - tensor input_205_pad_type_0 = const()[name = tensor("input_205_pad_type_0"), val = tensor("valid")]; - tensor input_205_groups_0 = const()[name = tensor("input_205_groups_0"), val = tensor(512)]; - tensor input_205_strides_0 = const()[name = tensor("input_205_strides_0"), val = tensor([1])]; - tensor input_205_pad_0 = const()[name = tensor("input_205_pad_0"), val = tensor([0, 0])]; - tensor input_205_dilations_0 = const()[name = tensor("input_205_dilations_0"), val = tensor([1])]; - tensor const_240_to_fp16 = const()[name = tensor("const_240_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48101184)))]; - tensor const_241_to_fp16 = const()[name = tensor("const_241_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48110464)))]; - tensor input_207_cast_fp16 = conv(bias = const_241_to_fp16, dilations = input_205_dilations_0, groups = input_205_groups_0, pad = input_205_pad_0, pad_type = input_205_pad_type_0, strides = input_205_strides_0, weight = const_240_to_fp16, x = input_203_cast_fp16)[name = tensor("input_207_cast_fp16")]; - tensor input_209_cast_fp16 = silu(x = input_207_cast_fp16)[name = tensor("input_209_cast_fp16")]; - tensor x_87_pad_type_0 = const()[name = tensor("x_87_pad_type_0"), val = tensor("valid")]; - tensor x_87_strides_0 = const()[name = tensor("x_87_strides_0"), val = tensor([1])]; - tensor x_87_pad_0 = const()[name = tensor("x_87_pad_0"), val = tensor([0, 0])]; - tensor x_87_dilations_0 = const()[name = tensor("x_87_dilations_0"), val = tensor([1])]; - tensor x_87_groups_0 = const()[name = tensor("x_87_groups_0"), val = tensor(1)]; - tensor module_layers_3_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_3_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48111552)))]; - tensor module_layers_3_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_3_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48635904)))]; - tensor x_87_cast_fp16 = conv(bias = module_layers_3_conv_pointwise_conv2_bias_to_fp16, dilations = x_87_dilations_0, groups = x_87_groups_0, pad = x_87_pad_0, pad_type = x_87_pad_type_0, strides = x_87_strides_0, weight = module_layers_3_conv_pointwise_conv2_weight_to_fp16, x = input_209_cast_fp16)[name = tensor("x_87_cast_fp16")]; - tensor input_211_perm_0 = const()[name = tensor("input_211_perm_0"), val = tensor([0, 2, 1])]; - tensor input_211_cast_fp16 = transpose(perm = input_211_perm_0, x = x_87_cast_fp16)[name = tensor("transpose_177")]; - tensor input_213_cast_fp16 = add(x = input_195_cast_fp16, y = input_211_cast_fp16)[name = tensor("input_213_cast_fp16")]; - tensor input_215_axes_0 = const()[name = tensor("input_215_axes_0"), val = tensor([-1])]; - tensor module_layers_3_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_3_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48636992)))]; - tensor module_layers_3_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_3_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48638080)))]; - tensor input_215_cast_fp16 = layer_norm(axes = input_215_axes_0, beta = module_layers_3_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_3_norm_feed_forward2_weight_to_fp16, x = input_213_cast_fp16)[name = tensor("input_215_cast_fp16")]; - tensor module_layers_3_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_3_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48639168)))]; - tensor module_layers_3_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_3_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(50736384)))]; - tensor linear_35_cast_fp16 = linear(bias = module_layers_3_feed_forward2_linear1_bias_to_fp16, weight = module_layers_3_feed_forward2_linear1_weight_to_fp16, x = input_215_cast_fp16)[name = tensor("linear_35_cast_fp16")]; - tensor input_219_cast_fp16 = silu(x = linear_35_cast_fp16)[name = tensor("input_219_cast_fp16")]; - tensor module_layers_3_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_3_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(50740544)))]; - tensor module_layers_3_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_3_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52837760)))]; - tensor linear_36_cast_fp16 = linear(bias = module_layers_3_feed_forward2_linear2_bias_to_fp16, weight = module_layers_3_feed_forward2_linear2_weight_to_fp16, x = input_219_cast_fp16)[name = tensor("linear_36_cast_fp16")]; - tensor var_1017_to_fp16 = const()[name = tensor("op_1017_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1018_cast_fp16 = mul(x = linear_36_cast_fp16, y = var_1017_to_fp16)[name = tensor("op_1018_cast_fp16")]; - tensor input_225_cast_fp16 = add(x = input_213_cast_fp16, y = var_1018_cast_fp16)[name = tensor("input_225_cast_fp16")]; - tensor input_227_axes_0 = const()[name = tensor("input_227_axes_0"), val = tensor([-1])]; - tensor module_layers_3_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_3_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52838848)))]; - tensor module_layers_3_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_3_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52839936)))]; - tensor input_227_cast_fp16 = layer_norm(axes = input_227_axes_0, beta = module_layers_3_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_3_norm_out_weight_to_fp16, x = input_225_cast_fp16)[name = tensor("input_227_cast_fp16")]; - tensor input_229_axes_0 = const()[name = tensor("input_229_axes_0"), val = tensor([-1])]; - tensor module_layers_4_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_4_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52841024)))]; - tensor module_layers_4_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_4_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52842112)))]; - tensor input_229_cast_fp16 = layer_norm(axes = input_229_axes_0, beta = module_layers_4_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_4_norm_feed_forward1_weight_to_fp16, x = input_227_cast_fp16)[name = tensor("input_229_cast_fp16")]; - tensor module_layers_4_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_4_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52843200)))]; - tensor module_layers_4_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_4_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(54940416)))]; - tensor linear_37_cast_fp16 = linear(bias = module_layers_4_feed_forward1_linear1_bias_to_fp16, weight = module_layers_4_feed_forward1_linear1_weight_to_fp16, x = input_229_cast_fp16)[name = tensor("linear_37_cast_fp16")]; - tensor input_233_cast_fp16 = silu(x = linear_37_cast_fp16)[name = tensor("input_233_cast_fp16")]; - tensor module_layers_4_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_4_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(54944576)))]; - tensor module_layers_4_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_4_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57041792)))]; - tensor linear_38_cast_fp16 = linear(bias = module_layers_4_feed_forward1_linear2_bias_to_fp16, weight = module_layers_4_feed_forward1_linear2_weight_to_fp16, x = input_233_cast_fp16)[name = tensor("linear_38_cast_fp16")]; - tensor var_1048_to_fp16 = const()[name = tensor("op_1048_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1049_cast_fp16 = mul(x = linear_38_cast_fp16, y = var_1048_to_fp16)[name = tensor("op_1049_cast_fp16")]; - tensor input_239_cast_fp16 = add(x = input_227_cast_fp16, y = var_1049_cast_fp16)[name = tensor("input_239_cast_fp16")]; - tensor query_9_axes_0 = const()[name = tensor("query_9_axes_0"), val = tensor([-1])]; - tensor module_layers_4_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_4_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57042880)))]; - tensor module_layers_4_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_4_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57043968)))]; - tensor query_9_cast_fp16 = layer_norm(axes = query_9_axes_0, beta = module_layers_4_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_4_norm_self_att_weight_to_fp16, x = input_239_cast_fp16)[name = tensor("query_9_cast_fp16")]; - tensor module_layers_4_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_4_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57045056)))]; - tensor module_layers_4_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_4_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57569408)))]; - tensor linear_39_cast_fp16 = linear(bias = module_layers_4_self_attn_linear_q_bias_to_fp16, weight = module_layers_4_self_attn_linear_q_weight_to_fp16, x = query_9_cast_fp16)[name = tensor("linear_39_cast_fp16")]; - tensor var_1066 = const()[name = tensor("op_1066"), val = tensor([1, -1, 8, 64])]; - tensor q_25_cast_fp16 = reshape(shape = var_1066, x = linear_39_cast_fp16)[name = tensor("q_25_cast_fp16")]; - tensor module_layers_4_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_4_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57570496)))]; - tensor module_layers_4_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_4_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58094848)))]; - tensor linear_40_cast_fp16 = linear(bias = module_layers_4_self_attn_linear_k_bias_to_fp16, weight = module_layers_4_self_attn_linear_k_weight_to_fp16, x = query_9_cast_fp16)[name = tensor("linear_40_cast_fp16")]; - tensor var_1071 = const()[name = tensor("op_1071"), val = tensor([1, -1, 8, 64])]; - tensor k_17_cast_fp16 = reshape(shape = var_1071, x = linear_40_cast_fp16)[name = tensor("k_17_cast_fp16")]; - tensor module_layers_4_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_4_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58095936)))]; - tensor module_layers_4_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_4_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58620288)))]; - tensor linear_41_cast_fp16 = linear(bias = module_layers_4_self_attn_linear_v_bias_to_fp16, weight = module_layers_4_self_attn_linear_v_weight_to_fp16, x = query_9_cast_fp16)[name = tensor("linear_41_cast_fp16")]; - tensor var_1076 = const()[name = tensor("op_1076"), val = tensor([1, -1, 8, 64])]; - tensor v_9_cast_fp16 = reshape(shape = var_1076, x = linear_41_cast_fp16)[name = tensor("v_9_cast_fp16")]; - tensor value_11_perm_0 = const()[name = tensor("value_11_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_4_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_4_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58621376)))]; - tensor var_1088_cast_fp16 = add(x = q_25_cast_fp16, y = module_layers_4_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1088_cast_fp16")]; - tensor module_layers_4_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_4_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58622464)))]; - tensor var_1090_cast_fp16 = add(x = q_25_cast_fp16, y = module_layers_4_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1090_cast_fp16")]; - tensor q_with_bias_v_9_perm_0 = const()[name = tensor("q_with_bias_v_9_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_95_transpose_x_0 = const()[name = tensor("x_95_transpose_x_0"), val = tensor(false)]; - tensor x_95_transpose_y_0 = const()[name = tensor("x_95_transpose_y_0"), val = tensor(false)]; - tensor var_1092_to_fp16 = const()[name = tensor("op_1092_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58623552)))]; - tensor q_with_bias_v_9_cast_fp16 = transpose(perm = q_with_bias_v_9_perm_0, x = var_1090_cast_fp16)[name = tensor("transpose_175")]; - tensor x_95_cast_fp16 = matmul(transpose_x = x_95_transpose_x_0, transpose_y = x_95_transpose_y_0, x = q_with_bias_v_9_cast_fp16, y = var_1092_to_fp16)[name = tensor("x_95_cast_fp16")]; - tensor x_97_pad_0 = const()[name = tensor("x_97_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_97_mode_0 = const()[name = tensor("x_97_mode_0"), val = tensor("constant")]; - tensor const_110_to_fp16 = const()[name = tensor("const_110_to_fp16"), val = tensor(0x0p+0)]; - tensor x_97_cast_fp16 = pad(constant_val = const_110_to_fp16, mode = x_97_mode_0, pad = x_97_pad_0, x = x_95_cast_fp16)[name = tensor("x_97_cast_fp16")]; - tensor var_1100 = const()[name = tensor("op_1100"), val = tensor([1, 8, -1, 188])]; - tensor x_99_cast_fp16 = reshape(shape = var_1100, x = x_97_cast_fp16)[name = tensor("x_99_cast_fp16")]; - tensor var_1104_begin_0 = const()[name = tensor("op_1104_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_1104_end_0 = const()[name = tensor("op_1104_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_1104_end_mask_0 = const()[name = tensor("op_1104_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_1104_cast_fp16 = slice_by_index(begin = var_1104_begin_0, end = var_1104_end_0, end_mask = var_1104_end_mask_0, x = x_99_cast_fp16)[name = tensor("op_1104_cast_fp16")]; - tensor var_1105 = const()[name = tensor("op_1105"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_17_cast_fp16 = reshape(shape = var_1105, x = var_1104_cast_fp16)[name = tensor("matrix_bd_17_cast_fp16")]; - tensor matrix_ac_9_transpose_x_0 = const()[name = tensor("matrix_ac_9_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_9_transpose_y_0 = const()[name = tensor("matrix_ac_9_transpose_y_0"), val = tensor(false)]; - tensor transpose_59_perm_0 = const()[name = tensor("transpose_59_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_60_perm_0 = const()[name = tensor("transpose_60_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_60 = transpose(perm = transpose_60_perm_0, x = k_17_cast_fp16)[name = tensor("transpose_173")]; - tensor transpose_59 = transpose(perm = transpose_59_perm_0, x = var_1088_cast_fp16)[name = tensor("transpose_174")]; - tensor matrix_ac_9_cast_fp16 = matmul(transpose_x = matrix_ac_9_transpose_x_0, transpose_y = matrix_ac_9_transpose_y_0, x = transpose_59, y = transpose_60)[name = tensor("matrix_ac_9_cast_fp16")]; - tensor matrix_bd_19_begin_0 = const()[name = tensor("matrix_bd_19_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_19_end_0 = const()[name = tensor("matrix_bd_19_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_19_end_mask_0 = const()[name = tensor("matrix_bd_19_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_19_cast_fp16 = slice_by_index(begin = matrix_bd_19_begin_0, end = matrix_bd_19_end_0, end_mask = matrix_bd_19_end_mask_0, x = matrix_bd_17_cast_fp16)[name = tensor("matrix_bd_19_cast_fp16")]; - tensor var_1114_cast_fp16 = add(x = matrix_ac_9_cast_fp16, y = matrix_bd_19_cast_fp16)[name = tensor("op_1114_cast_fp16")]; - tensor _inversed_scores_17_y_0_to_fp16 = const()[name = tensor("_inversed_scores_17_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_17_cast_fp16 = mul(x = var_1114_cast_fp16, y = _inversed_scores_17_y_0_to_fp16)[name = tensor("_inversed_scores_17_cast_fp16")]; - tensor scores_19_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_17_cast_fp16, cond = mask_11)[name = tensor("scores_19_cast_fp16")]; - tensor var_1120_cast_fp16 = softmax(axis = var_23, x = scores_19_cast_fp16)[name = tensor("op_1120_cast_fp16")]; - tensor input_241_cast_fp16 = select(a = var_11_to_fp16, b = var_1120_cast_fp16, cond = mask_11)[name = tensor("input_241_cast_fp16")]; - tensor x_101_transpose_x_0 = const()[name = tensor("x_101_transpose_x_0"), val = tensor(false)]; - tensor x_101_transpose_y_0 = const()[name = tensor("x_101_transpose_y_0"), val = tensor(false)]; - tensor value_11_cast_fp16 = transpose(perm = value_11_perm_0, x = v_9_cast_fp16)[name = tensor("transpose_176")]; - tensor x_101_cast_fp16 = matmul(transpose_x = x_101_transpose_x_0, transpose_y = x_101_transpose_y_0, x = input_241_cast_fp16, y = value_11_cast_fp16)[name = tensor("x_101_cast_fp16")]; - tensor var_1124_perm_0 = const()[name = tensor("op_1124_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_1125 = const()[name = tensor("op_1125"), val = tensor([1, -1, 512])]; - tensor var_1124_cast_fp16 = transpose(perm = var_1124_perm_0, x = x_101_cast_fp16)[name = tensor("transpose_172")]; - tensor input_243_cast_fp16 = reshape(shape = var_1125, x = var_1124_cast_fp16)[name = tensor("input_243_cast_fp16")]; - tensor module_layers_4_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_4_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59007616)))]; - tensor module_layers_4_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_4_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59531968)))]; - tensor linear_43_cast_fp16 = linear(bias = module_layers_4_self_attn_linear_out_bias_to_fp16, weight = module_layers_4_self_attn_linear_out_weight_to_fp16, x = input_243_cast_fp16)[name = tensor("linear_43_cast_fp16")]; - tensor input_247_cast_fp16 = add(x = input_239_cast_fp16, y = linear_43_cast_fp16)[name = tensor("input_247_cast_fp16")]; - tensor x_105_axes_0 = const()[name = tensor("x_105_axes_0"), val = tensor([-1])]; - tensor module_layers_4_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_4_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59533056)))]; - tensor module_layers_4_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_4_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59534144)))]; - tensor x_105_cast_fp16 = layer_norm(axes = x_105_axes_0, beta = module_layers_4_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_4_norm_conv_weight_to_fp16, x = input_247_cast_fp16)[name = tensor("x_105_cast_fp16")]; - tensor input_249_perm_0 = const()[name = tensor("input_249_perm_0"), val = tensor([0, 2, 1])]; - tensor input_251_pad_type_0 = const()[name = tensor("input_251_pad_type_0"), val = tensor("valid")]; - tensor input_251_strides_0 = const()[name = tensor("input_251_strides_0"), val = tensor([1])]; - tensor input_251_pad_0 = const()[name = tensor("input_251_pad_0"), val = tensor([0, 0])]; - tensor input_251_dilations_0 = const()[name = tensor("input_251_dilations_0"), val = tensor([1])]; - tensor input_251_groups_0 = const()[name = tensor("input_251_groups_0"), val = tensor(1)]; - tensor module_layers_4_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_4_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59535232)))]; - tensor module_layers_4_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_4_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60583872)))]; - tensor input_249_cast_fp16 = transpose(perm = input_249_perm_0, x = x_105_cast_fp16)[name = tensor("transpose_171")]; - tensor input_251_cast_fp16 = conv(bias = module_layers_4_conv_pointwise_conv1_bias_to_fp16, dilations = input_251_dilations_0, groups = input_251_groups_0, pad = input_251_pad_0, pad_type = input_251_pad_type_0, strides = input_251_strides_0, weight = module_layers_4_conv_pointwise_conv1_weight_to_fp16, x = input_249_cast_fp16)[name = tensor("input_251_cast_fp16")]; - tensor x_107_split_num_splits_0 = const()[name = tensor("x_107_split_num_splits_0"), val = tensor(2)]; - tensor x_107_split_axis_0 = const()[name = tensor("x_107_split_axis_0"), val = tensor(1)]; - tensor x_107_split_cast_fp16_0, tensor x_107_split_cast_fp16_1 = split(axis = x_107_split_axis_0, num_splits = x_107_split_num_splits_0, x = input_251_cast_fp16)[name = tensor("x_107_split_cast_fp16")]; - tensor x_107_split_1_sigmoid_cast_fp16 = sigmoid(x = x_107_split_cast_fp16_1)[name = tensor("x_107_split_1_sigmoid_cast_fp16")]; - tensor x_107_cast_fp16 = mul(x = x_107_split_cast_fp16_0, y = x_107_split_1_sigmoid_cast_fp16)[name = tensor("x_107_cast_fp16")]; - tensor input_253_cast_fp16 = select(a = var_11_to_fp16, b = x_107_cast_fp16, cond = var_453)[name = tensor("input_253_cast_fp16")]; - tensor input_255_pad_0 = const()[name = tensor("input_255_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_255_mode_0 = const()[name = tensor("input_255_mode_0"), val = tensor("constant")]; - tensor const_113_to_fp16 = const()[name = tensor("const_113_to_fp16"), val = tensor(0x0p+0)]; - tensor input_255_cast_fp16 = pad(constant_val = const_113_to_fp16, mode = input_255_mode_0, pad = input_255_pad_0, x = input_253_cast_fp16)[name = tensor("input_255_cast_fp16")]; - tensor input_257_pad_type_0 = const()[name = tensor("input_257_pad_type_0"), val = tensor("valid")]; - tensor input_257_groups_0 = const()[name = tensor("input_257_groups_0"), val = tensor(512)]; - tensor input_257_strides_0 = const()[name = tensor("input_257_strides_0"), val = tensor([1])]; - tensor input_257_pad_0 = const()[name = tensor("input_257_pad_0"), val = tensor([0, 0])]; - tensor input_257_dilations_0 = const()[name = tensor("input_257_dilations_0"), val = tensor([1])]; - tensor const_242_to_fp16 = const()[name = tensor("const_242_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60585984)))]; - tensor const_243_to_fp16 = const()[name = tensor("const_243_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60595264)))]; - tensor input_259_cast_fp16 = conv(bias = const_243_to_fp16, dilations = input_257_dilations_0, groups = input_257_groups_0, pad = input_257_pad_0, pad_type = input_257_pad_type_0, strides = input_257_strides_0, weight = const_242_to_fp16, x = input_255_cast_fp16)[name = tensor("input_259_cast_fp16")]; - tensor input_261_cast_fp16 = silu(x = input_259_cast_fp16)[name = tensor("input_261_cast_fp16")]; - tensor x_109_pad_type_0 = const()[name = tensor("x_109_pad_type_0"), val = tensor("valid")]; - tensor x_109_strides_0 = const()[name = tensor("x_109_strides_0"), val = tensor([1])]; - tensor x_109_pad_0 = const()[name = tensor("x_109_pad_0"), val = tensor([0, 0])]; - tensor x_109_dilations_0 = const()[name = tensor("x_109_dilations_0"), val = tensor([1])]; - tensor x_109_groups_0 = const()[name = tensor("x_109_groups_0"), val = tensor(1)]; - tensor module_layers_4_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_4_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60596352)))]; - tensor module_layers_4_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_4_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61120704)))]; - tensor x_109_cast_fp16 = conv(bias = module_layers_4_conv_pointwise_conv2_bias_to_fp16, dilations = x_109_dilations_0, groups = x_109_groups_0, pad = x_109_pad_0, pad_type = x_109_pad_type_0, strides = x_109_strides_0, weight = module_layers_4_conv_pointwise_conv2_weight_to_fp16, x = input_261_cast_fp16)[name = tensor("x_109_cast_fp16")]; - tensor input_263_perm_0 = const()[name = tensor("input_263_perm_0"), val = tensor([0, 2, 1])]; - tensor input_263_cast_fp16 = transpose(perm = input_263_perm_0, x = x_109_cast_fp16)[name = tensor("transpose_170")]; - tensor input_265_cast_fp16 = add(x = input_247_cast_fp16, y = input_263_cast_fp16)[name = tensor("input_265_cast_fp16")]; - tensor input_267_axes_0 = const()[name = tensor("input_267_axes_0"), val = tensor([-1])]; - tensor module_layers_4_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_4_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61121792)))]; - tensor module_layers_4_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_4_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61122880)))]; - tensor input_267_cast_fp16 = layer_norm(axes = input_267_axes_0, beta = module_layers_4_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_4_norm_feed_forward2_weight_to_fp16, x = input_265_cast_fp16)[name = tensor("input_267_cast_fp16")]; - tensor module_layers_4_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_4_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61123968)))]; - tensor module_layers_4_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_4_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(63221184)))]; - tensor linear_44_cast_fp16 = linear(bias = module_layers_4_feed_forward2_linear1_bias_to_fp16, weight = module_layers_4_feed_forward2_linear1_weight_to_fp16, x = input_267_cast_fp16)[name = tensor("linear_44_cast_fp16")]; - tensor input_271_cast_fp16 = silu(x = linear_44_cast_fp16)[name = tensor("input_271_cast_fp16")]; - tensor module_layers_4_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_4_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(63225344)))]; - tensor module_layers_4_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_4_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65322560)))]; - tensor linear_45_cast_fp16 = linear(bias = module_layers_4_feed_forward2_linear2_bias_to_fp16, weight = module_layers_4_feed_forward2_linear2_weight_to_fp16, x = input_271_cast_fp16)[name = tensor("linear_45_cast_fp16")]; - tensor var_1191_to_fp16 = const()[name = tensor("op_1191_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1192_cast_fp16 = mul(x = linear_45_cast_fp16, y = var_1191_to_fp16)[name = tensor("op_1192_cast_fp16")]; - tensor input_277_cast_fp16 = add(x = input_265_cast_fp16, y = var_1192_cast_fp16)[name = tensor("input_277_cast_fp16")]; - tensor input_279_axes_0 = const()[name = tensor("input_279_axes_0"), val = tensor([-1])]; - tensor module_layers_4_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_4_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65323648)))]; - tensor module_layers_4_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_4_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65324736)))]; - tensor input_279_cast_fp16 = layer_norm(axes = input_279_axes_0, beta = module_layers_4_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_4_norm_out_weight_to_fp16, x = input_277_cast_fp16)[name = tensor("input_279_cast_fp16")]; - tensor input_281_axes_0 = const()[name = tensor("input_281_axes_0"), val = tensor([-1])]; - tensor module_layers_5_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_5_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65325824)))]; - tensor module_layers_5_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_5_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65326912)))]; - tensor input_281_cast_fp16 = layer_norm(axes = input_281_axes_0, beta = module_layers_5_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_5_norm_feed_forward1_weight_to_fp16, x = input_279_cast_fp16)[name = tensor("input_281_cast_fp16")]; - tensor module_layers_5_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_5_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65328000)))]; - tensor module_layers_5_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_5_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67425216)))]; - tensor linear_46_cast_fp16 = linear(bias = module_layers_5_feed_forward1_linear1_bias_to_fp16, weight = module_layers_5_feed_forward1_linear1_weight_to_fp16, x = input_281_cast_fp16)[name = tensor("linear_46_cast_fp16")]; - tensor input_285_cast_fp16 = silu(x = linear_46_cast_fp16)[name = tensor("input_285_cast_fp16")]; - tensor module_layers_5_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_5_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67429376)))]; - tensor module_layers_5_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_5_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(69526592)))]; - tensor linear_47_cast_fp16 = linear(bias = module_layers_5_feed_forward1_linear2_bias_to_fp16, weight = module_layers_5_feed_forward1_linear2_weight_to_fp16, x = input_285_cast_fp16)[name = tensor("linear_47_cast_fp16")]; - tensor var_1222_to_fp16 = const()[name = tensor("op_1222_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1223_cast_fp16 = mul(x = linear_47_cast_fp16, y = var_1222_to_fp16)[name = tensor("op_1223_cast_fp16")]; - tensor input_291_cast_fp16 = add(x = input_279_cast_fp16, y = var_1223_cast_fp16)[name = tensor("input_291_cast_fp16")]; - tensor query_11_axes_0 = const()[name = tensor("query_11_axes_0"), val = tensor([-1])]; - tensor module_layers_5_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_5_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(69527680)))]; - tensor module_layers_5_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_5_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(69528768)))]; - tensor query_11_cast_fp16 = layer_norm(axes = query_11_axes_0, beta = module_layers_5_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_5_norm_self_att_weight_to_fp16, x = input_291_cast_fp16)[name = tensor("query_11_cast_fp16")]; - tensor module_layers_5_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_5_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(69529856)))]; - tensor module_layers_5_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_5_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70054208)))]; - tensor linear_48_cast_fp16 = linear(bias = module_layers_5_self_attn_linear_q_bias_to_fp16, weight = module_layers_5_self_attn_linear_q_weight_to_fp16, x = query_11_cast_fp16)[name = tensor("linear_48_cast_fp16")]; - tensor var_1240 = const()[name = tensor("op_1240"), val = tensor([1, -1, 8, 64])]; - tensor q_31_cast_fp16 = reshape(shape = var_1240, x = linear_48_cast_fp16)[name = tensor("q_31_cast_fp16")]; - tensor module_layers_5_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_5_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70055296)))]; - tensor module_layers_5_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_5_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70579648)))]; - tensor linear_49_cast_fp16 = linear(bias = module_layers_5_self_attn_linear_k_bias_to_fp16, weight = module_layers_5_self_attn_linear_k_weight_to_fp16, x = query_11_cast_fp16)[name = tensor("linear_49_cast_fp16")]; - tensor var_1245 = const()[name = tensor("op_1245"), val = tensor([1, -1, 8, 64])]; - tensor k_21_cast_fp16 = reshape(shape = var_1245, x = linear_49_cast_fp16)[name = tensor("k_21_cast_fp16")]; - tensor module_layers_5_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_5_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70580736)))]; - tensor module_layers_5_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_5_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71105088)))]; - tensor linear_50_cast_fp16 = linear(bias = module_layers_5_self_attn_linear_v_bias_to_fp16, weight = module_layers_5_self_attn_linear_v_weight_to_fp16, x = query_11_cast_fp16)[name = tensor("linear_50_cast_fp16")]; - tensor var_1250 = const()[name = tensor("op_1250"), val = tensor([1, -1, 8, 64])]; - tensor v_11_cast_fp16 = reshape(shape = var_1250, x = linear_50_cast_fp16)[name = tensor("v_11_cast_fp16")]; - tensor value_13_perm_0 = const()[name = tensor("value_13_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_5_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_5_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71106176)))]; - tensor var_1262_cast_fp16 = add(x = q_31_cast_fp16, y = module_layers_5_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1262_cast_fp16")]; - tensor module_layers_5_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_5_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71107264)))]; - tensor var_1264_cast_fp16 = add(x = q_31_cast_fp16, y = module_layers_5_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1264_cast_fp16")]; - tensor q_with_bias_v_11_perm_0 = const()[name = tensor("q_with_bias_v_11_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_117_transpose_x_0 = const()[name = tensor("x_117_transpose_x_0"), val = tensor(false)]; - tensor x_117_transpose_y_0 = const()[name = tensor("x_117_transpose_y_0"), val = tensor(false)]; - tensor var_1266_to_fp16 = const()[name = tensor("op_1266_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71108352)))]; - tensor q_with_bias_v_11_cast_fp16 = transpose(perm = q_with_bias_v_11_perm_0, x = var_1264_cast_fp16)[name = tensor("transpose_168")]; - tensor x_117_cast_fp16 = matmul(transpose_x = x_117_transpose_x_0, transpose_y = x_117_transpose_y_0, x = q_with_bias_v_11_cast_fp16, y = var_1266_to_fp16)[name = tensor("x_117_cast_fp16")]; - tensor x_119_pad_0 = const()[name = tensor("x_119_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_119_mode_0 = const()[name = tensor("x_119_mode_0"), val = tensor("constant")]; - tensor const_120_to_fp16 = const()[name = tensor("const_120_to_fp16"), val = tensor(0x0p+0)]; - tensor x_119_cast_fp16 = pad(constant_val = const_120_to_fp16, mode = x_119_mode_0, pad = x_119_pad_0, x = x_117_cast_fp16)[name = tensor("x_119_cast_fp16")]; - tensor var_1274 = const()[name = tensor("op_1274"), val = tensor([1, 8, -1, 188])]; - tensor x_121_cast_fp16 = reshape(shape = var_1274, x = x_119_cast_fp16)[name = tensor("x_121_cast_fp16")]; - tensor var_1278_begin_0 = const()[name = tensor("op_1278_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_1278_end_0 = const()[name = tensor("op_1278_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_1278_end_mask_0 = const()[name = tensor("op_1278_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_1278_cast_fp16 = slice_by_index(begin = var_1278_begin_0, end = var_1278_end_0, end_mask = var_1278_end_mask_0, x = x_121_cast_fp16)[name = tensor("op_1278_cast_fp16")]; - tensor var_1279 = const()[name = tensor("op_1279"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_21_cast_fp16 = reshape(shape = var_1279, x = var_1278_cast_fp16)[name = tensor("matrix_bd_21_cast_fp16")]; - tensor matrix_ac_11_transpose_x_0 = const()[name = tensor("matrix_ac_11_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_11_transpose_y_0 = const()[name = tensor("matrix_ac_11_transpose_y_0"), val = tensor(false)]; - tensor transpose_61_perm_0 = const()[name = tensor("transpose_61_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_62_perm_0 = const()[name = tensor("transpose_62_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_62 = transpose(perm = transpose_62_perm_0, x = k_21_cast_fp16)[name = tensor("transpose_166")]; - tensor transpose_61 = transpose(perm = transpose_61_perm_0, x = var_1262_cast_fp16)[name = tensor("transpose_167")]; - tensor matrix_ac_11_cast_fp16 = matmul(transpose_x = matrix_ac_11_transpose_x_0, transpose_y = matrix_ac_11_transpose_y_0, x = transpose_61, y = transpose_62)[name = tensor("matrix_ac_11_cast_fp16")]; - tensor matrix_bd_23_begin_0 = const()[name = tensor("matrix_bd_23_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_23_end_0 = const()[name = tensor("matrix_bd_23_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_23_end_mask_0 = const()[name = tensor("matrix_bd_23_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_23_cast_fp16 = slice_by_index(begin = matrix_bd_23_begin_0, end = matrix_bd_23_end_0, end_mask = matrix_bd_23_end_mask_0, x = matrix_bd_21_cast_fp16)[name = tensor("matrix_bd_23_cast_fp16")]; - tensor var_1288_cast_fp16 = add(x = matrix_ac_11_cast_fp16, y = matrix_bd_23_cast_fp16)[name = tensor("op_1288_cast_fp16")]; - tensor _inversed_scores_21_y_0_to_fp16 = const()[name = tensor("_inversed_scores_21_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_21_cast_fp16 = mul(x = var_1288_cast_fp16, y = _inversed_scores_21_y_0_to_fp16)[name = tensor("_inversed_scores_21_cast_fp16")]; - tensor scores_23_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_21_cast_fp16, cond = mask_11)[name = tensor("scores_23_cast_fp16")]; - tensor var_1294_cast_fp16 = softmax(axis = var_23, x = scores_23_cast_fp16)[name = tensor("op_1294_cast_fp16")]; - tensor input_293_cast_fp16 = select(a = var_11_to_fp16, b = var_1294_cast_fp16, cond = mask_11)[name = tensor("input_293_cast_fp16")]; - tensor x_123_transpose_x_0 = const()[name = tensor("x_123_transpose_x_0"), val = tensor(false)]; - tensor x_123_transpose_y_0 = const()[name = tensor("x_123_transpose_y_0"), val = tensor(false)]; - tensor value_13_cast_fp16 = transpose(perm = value_13_perm_0, x = v_11_cast_fp16)[name = tensor("transpose_169")]; - tensor x_123_cast_fp16 = matmul(transpose_x = x_123_transpose_x_0, transpose_y = x_123_transpose_y_0, x = input_293_cast_fp16, y = value_13_cast_fp16)[name = tensor("x_123_cast_fp16")]; - tensor var_1298_perm_0 = const()[name = tensor("op_1298_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_1299 = const()[name = tensor("op_1299"), val = tensor([1, -1, 512])]; - tensor var_1298_cast_fp16 = transpose(perm = var_1298_perm_0, x = x_123_cast_fp16)[name = tensor("transpose_165")]; - tensor input_295_cast_fp16 = reshape(shape = var_1299, x = var_1298_cast_fp16)[name = tensor("input_295_cast_fp16")]; - tensor module_layers_5_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_5_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71492416)))]; - tensor module_layers_5_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_5_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(72016768)))]; - tensor linear_52_cast_fp16 = linear(bias = module_layers_5_self_attn_linear_out_bias_to_fp16, weight = module_layers_5_self_attn_linear_out_weight_to_fp16, x = input_295_cast_fp16)[name = tensor("linear_52_cast_fp16")]; - tensor input_299_cast_fp16 = add(x = input_291_cast_fp16, y = linear_52_cast_fp16)[name = tensor("input_299_cast_fp16")]; - tensor x_127_axes_0 = const()[name = tensor("x_127_axes_0"), val = tensor([-1])]; - tensor module_layers_5_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_5_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(72017856)))]; - tensor module_layers_5_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_5_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(72018944)))]; - tensor x_127_cast_fp16 = layer_norm(axes = x_127_axes_0, beta = module_layers_5_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_5_norm_conv_weight_to_fp16, x = input_299_cast_fp16)[name = tensor("x_127_cast_fp16")]; - tensor input_301_perm_0 = const()[name = tensor("input_301_perm_0"), val = tensor([0, 2, 1])]; - tensor input_303_pad_type_0 = const()[name = tensor("input_303_pad_type_0"), val = tensor("valid")]; - tensor input_303_strides_0 = const()[name = tensor("input_303_strides_0"), val = tensor([1])]; - tensor input_303_pad_0 = const()[name = tensor("input_303_pad_0"), val = tensor([0, 0])]; - tensor input_303_dilations_0 = const()[name = tensor("input_303_dilations_0"), val = tensor([1])]; - tensor input_303_groups_0 = const()[name = tensor("input_303_groups_0"), val = tensor(1)]; - tensor module_layers_5_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_5_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(72020032)))]; - tensor module_layers_5_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_5_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73068672)))]; - tensor input_301_cast_fp16 = transpose(perm = input_301_perm_0, x = x_127_cast_fp16)[name = tensor("transpose_164")]; - tensor input_303_cast_fp16 = conv(bias = module_layers_5_conv_pointwise_conv1_bias_to_fp16, dilations = input_303_dilations_0, groups = input_303_groups_0, pad = input_303_pad_0, pad_type = input_303_pad_type_0, strides = input_303_strides_0, weight = module_layers_5_conv_pointwise_conv1_weight_to_fp16, x = input_301_cast_fp16)[name = tensor("input_303_cast_fp16")]; - tensor x_129_split_num_splits_0 = const()[name = tensor("x_129_split_num_splits_0"), val = tensor(2)]; - tensor x_129_split_axis_0 = const()[name = tensor("x_129_split_axis_0"), val = tensor(1)]; - tensor x_129_split_cast_fp16_0, tensor x_129_split_cast_fp16_1 = split(axis = x_129_split_axis_0, num_splits = x_129_split_num_splits_0, x = input_303_cast_fp16)[name = tensor("x_129_split_cast_fp16")]; - tensor x_129_split_1_sigmoid_cast_fp16 = sigmoid(x = x_129_split_cast_fp16_1)[name = tensor("x_129_split_1_sigmoid_cast_fp16")]; - tensor x_129_cast_fp16 = mul(x = x_129_split_cast_fp16_0, y = x_129_split_1_sigmoid_cast_fp16)[name = tensor("x_129_cast_fp16")]; - tensor input_305_cast_fp16 = select(a = var_11_to_fp16, b = x_129_cast_fp16, cond = var_453)[name = tensor("input_305_cast_fp16")]; - tensor input_307_pad_0 = const()[name = tensor("input_307_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_307_mode_0 = const()[name = tensor("input_307_mode_0"), val = tensor("constant")]; - tensor const_123_to_fp16 = const()[name = tensor("const_123_to_fp16"), val = tensor(0x0p+0)]; - tensor input_307_cast_fp16 = pad(constant_val = const_123_to_fp16, mode = input_307_mode_0, pad = input_307_pad_0, x = input_305_cast_fp16)[name = tensor("input_307_cast_fp16")]; - tensor input_309_pad_type_0 = const()[name = tensor("input_309_pad_type_0"), val = tensor("valid")]; - tensor input_309_groups_0 = const()[name = tensor("input_309_groups_0"), val = tensor(512)]; - tensor input_309_strides_0 = const()[name = tensor("input_309_strides_0"), val = tensor([1])]; - tensor input_309_pad_0 = const()[name = tensor("input_309_pad_0"), val = tensor([0, 0])]; - tensor input_309_dilations_0 = const()[name = tensor("input_309_dilations_0"), val = tensor([1])]; - tensor const_244_to_fp16 = const()[name = tensor("const_244_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73070784)))]; - tensor const_245_to_fp16 = const()[name = tensor("const_245_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73080064)))]; - tensor input_311_cast_fp16 = conv(bias = const_245_to_fp16, dilations = input_309_dilations_0, groups = input_309_groups_0, pad = input_309_pad_0, pad_type = input_309_pad_type_0, strides = input_309_strides_0, weight = const_244_to_fp16, x = input_307_cast_fp16)[name = tensor("input_311_cast_fp16")]; - tensor input_313_cast_fp16 = silu(x = input_311_cast_fp16)[name = tensor("input_313_cast_fp16")]; - tensor x_131_pad_type_0 = const()[name = tensor("x_131_pad_type_0"), val = tensor("valid")]; - tensor x_131_strides_0 = const()[name = tensor("x_131_strides_0"), val = tensor([1])]; - tensor x_131_pad_0 = const()[name = tensor("x_131_pad_0"), val = tensor([0, 0])]; - tensor x_131_dilations_0 = const()[name = tensor("x_131_dilations_0"), val = tensor([1])]; - tensor x_131_groups_0 = const()[name = tensor("x_131_groups_0"), val = tensor(1)]; - tensor module_layers_5_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_5_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73081152)))]; - tensor module_layers_5_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_5_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73605504)))]; - tensor x_131_cast_fp16 = conv(bias = module_layers_5_conv_pointwise_conv2_bias_to_fp16, dilations = x_131_dilations_0, groups = x_131_groups_0, pad = x_131_pad_0, pad_type = x_131_pad_type_0, strides = x_131_strides_0, weight = module_layers_5_conv_pointwise_conv2_weight_to_fp16, x = input_313_cast_fp16)[name = tensor("x_131_cast_fp16")]; - tensor input_315_perm_0 = const()[name = tensor("input_315_perm_0"), val = tensor([0, 2, 1])]; - tensor input_315_cast_fp16 = transpose(perm = input_315_perm_0, x = x_131_cast_fp16)[name = tensor("transpose_163")]; - tensor input_317_cast_fp16 = add(x = input_299_cast_fp16, y = input_315_cast_fp16)[name = tensor("input_317_cast_fp16")]; - tensor input_319_axes_0 = const()[name = tensor("input_319_axes_0"), val = tensor([-1])]; - tensor module_layers_5_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_5_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73606592)))]; - tensor module_layers_5_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_5_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73607680)))]; - tensor input_319_cast_fp16 = layer_norm(axes = input_319_axes_0, beta = module_layers_5_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_5_norm_feed_forward2_weight_to_fp16, x = input_317_cast_fp16)[name = tensor("input_319_cast_fp16")]; - tensor module_layers_5_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_5_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73608768)))]; - tensor module_layers_5_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_5_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(75705984)))]; - tensor linear_53_cast_fp16 = linear(bias = module_layers_5_feed_forward2_linear1_bias_to_fp16, weight = module_layers_5_feed_forward2_linear1_weight_to_fp16, x = input_319_cast_fp16)[name = tensor("linear_53_cast_fp16")]; - tensor input_323_cast_fp16 = silu(x = linear_53_cast_fp16)[name = tensor("input_323_cast_fp16")]; - tensor module_layers_5_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_5_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(75710144)))]; - tensor module_layers_5_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_5_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(77807360)))]; - tensor linear_54_cast_fp16 = linear(bias = module_layers_5_feed_forward2_linear2_bias_to_fp16, weight = module_layers_5_feed_forward2_linear2_weight_to_fp16, x = input_323_cast_fp16)[name = tensor("linear_54_cast_fp16")]; - tensor var_1365_to_fp16 = const()[name = tensor("op_1365_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1366_cast_fp16 = mul(x = linear_54_cast_fp16, y = var_1365_to_fp16)[name = tensor("op_1366_cast_fp16")]; - tensor input_329_cast_fp16 = add(x = input_317_cast_fp16, y = var_1366_cast_fp16)[name = tensor("input_329_cast_fp16")]; - tensor input_331_axes_0 = const()[name = tensor("input_331_axes_0"), val = tensor([-1])]; - tensor module_layers_5_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_5_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(77808448)))]; - tensor module_layers_5_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_5_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(77809536)))]; - tensor input_331_cast_fp16 = layer_norm(axes = input_331_axes_0, beta = module_layers_5_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_5_norm_out_weight_to_fp16, x = input_329_cast_fp16)[name = tensor("input_331_cast_fp16")]; - tensor input_333_axes_0 = const()[name = tensor("input_333_axes_0"), val = tensor([-1])]; - tensor module_layers_6_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_6_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(77810624)))]; - tensor module_layers_6_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_6_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(77811712)))]; - tensor input_333_cast_fp16 = layer_norm(axes = input_333_axes_0, beta = module_layers_6_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_6_norm_feed_forward1_weight_to_fp16, x = input_331_cast_fp16)[name = tensor("input_333_cast_fp16")]; - tensor module_layers_6_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_6_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(77812800)))]; - tensor module_layers_6_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_6_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(79910016)))]; - tensor linear_55_cast_fp16 = linear(bias = module_layers_6_feed_forward1_linear1_bias_to_fp16, weight = module_layers_6_feed_forward1_linear1_weight_to_fp16, x = input_333_cast_fp16)[name = tensor("linear_55_cast_fp16")]; - tensor input_337_cast_fp16 = silu(x = linear_55_cast_fp16)[name = tensor("input_337_cast_fp16")]; - tensor module_layers_6_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_6_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(79914176)))]; - tensor module_layers_6_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_6_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82011392)))]; - tensor linear_56_cast_fp16 = linear(bias = module_layers_6_feed_forward1_linear2_bias_to_fp16, weight = module_layers_6_feed_forward1_linear2_weight_to_fp16, x = input_337_cast_fp16)[name = tensor("linear_56_cast_fp16")]; - tensor var_1396_to_fp16 = const()[name = tensor("op_1396_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1397_cast_fp16 = mul(x = linear_56_cast_fp16, y = var_1396_to_fp16)[name = tensor("op_1397_cast_fp16")]; - tensor input_343_cast_fp16 = add(x = input_331_cast_fp16, y = var_1397_cast_fp16)[name = tensor("input_343_cast_fp16")]; - tensor query_13_axes_0 = const()[name = tensor("query_13_axes_0"), val = tensor([-1])]; - tensor module_layers_6_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_6_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82012480)))]; - tensor module_layers_6_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_6_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82013568)))]; - tensor query_13_cast_fp16 = layer_norm(axes = query_13_axes_0, beta = module_layers_6_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_6_norm_self_att_weight_to_fp16, x = input_343_cast_fp16)[name = tensor("query_13_cast_fp16")]; - tensor module_layers_6_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_6_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82014656)))]; - tensor module_layers_6_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_6_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82539008)))]; - tensor linear_57_cast_fp16 = linear(bias = module_layers_6_self_attn_linear_q_bias_to_fp16, weight = module_layers_6_self_attn_linear_q_weight_to_fp16, x = query_13_cast_fp16)[name = tensor("linear_57_cast_fp16")]; - tensor var_1414 = const()[name = tensor("op_1414"), val = tensor([1, -1, 8, 64])]; - tensor q_37_cast_fp16 = reshape(shape = var_1414, x = linear_57_cast_fp16)[name = tensor("q_37_cast_fp16")]; - tensor module_layers_6_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_6_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82540096)))]; - tensor module_layers_6_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_6_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83064448)))]; - tensor linear_58_cast_fp16 = linear(bias = module_layers_6_self_attn_linear_k_bias_to_fp16, weight = module_layers_6_self_attn_linear_k_weight_to_fp16, x = query_13_cast_fp16)[name = tensor("linear_58_cast_fp16")]; - tensor var_1419 = const()[name = tensor("op_1419"), val = tensor([1, -1, 8, 64])]; - tensor k_25_cast_fp16 = reshape(shape = var_1419, x = linear_58_cast_fp16)[name = tensor("k_25_cast_fp16")]; - tensor module_layers_6_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_6_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83065536)))]; - tensor module_layers_6_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_6_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83589888)))]; - tensor linear_59_cast_fp16 = linear(bias = module_layers_6_self_attn_linear_v_bias_to_fp16, weight = module_layers_6_self_attn_linear_v_weight_to_fp16, x = query_13_cast_fp16)[name = tensor("linear_59_cast_fp16")]; - tensor var_1424 = const()[name = tensor("op_1424"), val = tensor([1, -1, 8, 64])]; - tensor v_13_cast_fp16 = reshape(shape = var_1424, x = linear_59_cast_fp16)[name = tensor("v_13_cast_fp16")]; - tensor value_15_perm_0 = const()[name = tensor("value_15_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_6_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_6_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83590976)))]; - tensor var_1436_cast_fp16 = add(x = q_37_cast_fp16, y = module_layers_6_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1436_cast_fp16")]; - tensor module_layers_6_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_6_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83592064)))]; - tensor var_1438_cast_fp16 = add(x = q_37_cast_fp16, y = module_layers_6_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1438_cast_fp16")]; - tensor q_with_bias_v_13_perm_0 = const()[name = tensor("q_with_bias_v_13_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_139_transpose_x_0 = const()[name = tensor("x_139_transpose_x_0"), val = tensor(false)]; - tensor x_139_transpose_y_0 = const()[name = tensor("x_139_transpose_y_0"), val = tensor(false)]; - tensor var_1440_to_fp16 = const()[name = tensor("op_1440_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83593152)))]; - tensor q_with_bias_v_13_cast_fp16 = transpose(perm = q_with_bias_v_13_perm_0, x = var_1438_cast_fp16)[name = tensor("transpose_161")]; - tensor x_139_cast_fp16 = matmul(transpose_x = x_139_transpose_x_0, transpose_y = x_139_transpose_y_0, x = q_with_bias_v_13_cast_fp16, y = var_1440_to_fp16)[name = tensor("x_139_cast_fp16")]; - tensor x_141_pad_0 = const()[name = tensor("x_141_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_141_mode_0 = const()[name = tensor("x_141_mode_0"), val = tensor("constant")]; - tensor const_130_to_fp16 = const()[name = tensor("const_130_to_fp16"), val = tensor(0x0p+0)]; - tensor x_141_cast_fp16 = pad(constant_val = const_130_to_fp16, mode = x_141_mode_0, pad = x_141_pad_0, x = x_139_cast_fp16)[name = tensor("x_141_cast_fp16")]; - tensor var_1448 = const()[name = tensor("op_1448"), val = tensor([1, 8, -1, 188])]; - tensor x_143_cast_fp16 = reshape(shape = var_1448, x = x_141_cast_fp16)[name = tensor("x_143_cast_fp16")]; - tensor var_1452_begin_0 = const()[name = tensor("op_1452_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_1452_end_0 = const()[name = tensor("op_1452_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_1452_end_mask_0 = const()[name = tensor("op_1452_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_1452_cast_fp16 = slice_by_index(begin = var_1452_begin_0, end = var_1452_end_0, end_mask = var_1452_end_mask_0, x = x_143_cast_fp16)[name = tensor("op_1452_cast_fp16")]; - tensor var_1453 = const()[name = tensor("op_1453"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_25_cast_fp16 = reshape(shape = var_1453, x = var_1452_cast_fp16)[name = tensor("matrix_bd_25_cast_fp16")]; - tensor matrix_ac_13_transpose_x_0 = const()[name = tensor("matrix_ac_13_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_13_transpose_y_0 = const()[name = tensor("matrix_ac_13_transpose_y_0"), val = tensor(false)]; - tensor transpose_63_perm_0 = const()[name = tensor("transpose_63_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_64_perm_0 = const()[name = tensor("transpose_64_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_64 = transpose(perm = transpose_64_perm_0, x = k_25_cast_fp16)[name = tensor("transpose_159")]; - tensor transpose_63 = transpose(perm = transpose_63_perm_0, x = var_1436_cast_fp16)[name = tensor("transpose_160")]; - tensor matrix_ac_13_cast_fp16 = matmul(transpose_x = matrix_ac_13_transpose_x_0, transpose_y = matrix_ac_13_transpose_y_0, x = transpose_63, y = transpose_64)[name = tensor("matrix_ac_13_cast_fp16")]; - tensor matrix_bd_27_begin_0 = const()[name = tensor("matrix_bd_27_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_27_end_0 = const()[name = tensor("matrix_bd_27_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_27_end_mask_0 = const()[name = tensor("matrix_bd_27_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_27_cast_fp16 = slice_by_index(begin = matrix_bd_27_begin_0, end = matrix_bd_27_end_0, end_mask = matrix_bd_27_end_mask_0, x = matrix_bd_25_cast_fp16)[name = tensor("matrix_bd_27_cast_fp16")]; - tensor var_1462_cast_fp16 = add(x = matrix_ac_13_cast_fp16, y = matrix_bd_27_cast_fp16)[name = tensor("op_1462_cast_fp16")]; - tensor _inversed_scores_25_y_0_to_fp16 = const()[name = tensor("_inversed_scores_25_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_25_cast_fp16 = mul(x = var_1462_cast_fp16, y = _inversed_scores_25_y_0_to_fp16)[name = tensor("_inversed_scores_25_cast_fp16")]; - tensor scores_27_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_25_cast_fp16, cond = mask_11)[name = tensor("scores_27_cast_fp16")]; - tensor var_1468_cast_fp16 = softmax(axis = var_23, x = scores_27_cast_fp16)[name = tensor("op_1468_cast_fp16")]; - tensor input_345_cast_fp16 = select(a = var_11_to_fp16, b = var_1468_cast_fp16, cond = mask_11)[name = tensor("input_345_cast_fp16")]; - tensor x_145_transpose_x_0 = const()[name = tensor("x_145_transpose_x_0"), val = tensor(false)]; - tensor x_145_transpose_y_0 = const()[name = tensor("x_145_transpose_y_0"), val = tensor(false)]; - tensor value_15_cast_fp16 = transpose(perm = value_15_perm_0, x = v_13_cast_fp16)[name = tensor("transpose_162")]; - tensor x_145_cast_fp16 = matmul(transpose_x = x_145_transpose_x_0, transpose_y = x_145_transpose_y_0, x = input_345_cast_fp16, y = value_15_cast_fp16)[name = tensor("x_145_cast_fp16")]; - tensor var_1472_perm_0 = const()[name = tensor("op_1472_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_1473 = const()[name = tensor("op_1473"), val = tensor([1, -1, 512])]; - tensor var_1472_cast_fp16 = transpose(perm = var_1472_perm_0, x = x_145_cast_fp16)[name = tensor("transpose_158")]; - tensor input_347_cast_fp16 = reshape(shape = var_1473, x = var_1472_cast_fp16)[name = tensor("input_347_cast_fp16")]; - tensor module_layers_6_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_6_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83977216)))]; - tensor module_layers_6_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_6_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84501568)))]; - tensor linear_61_cast_fp16 = linear(bias = module_layers_6_self_attn_linear_out_bias_to_fp16, weight = module_layers_6_self_attn_linear_out_weight_to_fp16, x = input_347_cast_fp16)[name = tensor("linear_61_cast_fp16")]; - tensor input_351_cast_fp16 = add(x = input_343_cast_fp16, y = linear_61_cast_fp16)[name = tensor("input_351_cast_fp16")]; - tensor x_149_axes_0 = const()[name = tensor("x_149_axes_0"), val = tensor([-1])]; - tensor module_layers_6_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_6_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84502656)))]; - tensor module_layers_6_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_6_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84503744)))]; - tensor x_149_cast_fp16 = layer_norm(axes = x_149_axes_0, beta = module_layers_6_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_6_norm_conv_weight_to_fp16, x = input_351_cast_fp16)[name = tensor("x_149_cast_fp16")]; - tensor input_353_perm_0 = const()[name = tensor("input_353_perm_0"), val = tensor([0, 2, 1])]; - tensor input_355_pad_type_0 = const()[name = tensor("input_355_pad_type_0"), val = tensor("valid")]; - tensor input_355_strides_0 = const()[name = tensor("input_355_strides_0"), val = tensor([1])]; - tensor input_355_pad_0 = const()[name = tensor("input_355_pad_0"), val = tensor([0, 0])]; - tensor input_355_dilations_0 = const()[name = tensor("input_355_dilations_0"), val = tensor([1])]; - tensor input_355_groups_0 = const()[name = tensor("input_355_groups_0"), val = tensor(1)]; - tensor module_layers_6_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_6_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84504832)))]; - tensor module_layers_6_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_6_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85553472)))]; - tensor input_353_cast_fp16 = transpose(perm = input_353_perm_0, x = x_149_cast_fp16)[name = tensor("transpose_157")]; - tensor input_355_cast_fp16 = conv(bias = module_layers_6_conv_pointwise_conv1_bias_to_fp16, dilations = input_355_dilations_0, groups = input_355_groups_0, pad = input_355_pad_0, pad_type = input_355_pad_type_0, strides = input_355_strides_0, weight = module_layers_6_conv_pointwise_conv1_weight_to_fp16, x = input_353_cast_fp16)[name = tensor("input_355_cast_fp16")]; - tensor x_151_split_num_splits_0 = const()[name = tensor("x_151_split_num_splits_0"), val = tensor(2)]; - tensor x_151_split_axis_0 = const()[name = tensor("x_151_split_axis_0"), val = tensor(1)]; - tensor x_151_split_cast_fp16_0, tensor x_151_split_cast_fp16_1 = split(axis = x_151_split_axis_0, num_splits = x_151_split_num_splits_0, x = input_355_cast_fp16)[name = tensor("x_151_split_cast_fp16")]; - tensor x_151_split_1_sigmoid_cast_fp16 = sigmoid(x = x_151_split_cast_fp16_1)[name = tensor("x_151_split_1_sigmoid_cast_fp16")]; - tensor x_151_cast_fp16 = mul(x = x_151_split_cast_fp16_0, y = x_151_split_1_sigmoid_cast_fp16)[name = tensor("x_151_cast_fp16")]; - tensor input_357_cast_fp16 = select(a = var_11_to_fp16, b = x_151_cast_fp16, cond = var_453)[name = tensor("input_357_cast_fp16")]; - tensor input_359_pad_0 = const()[name = tensor("input_359_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_359_mode_0 = const()[name = tensor("input_359_mode_0"), val = tensor("constant")]; - tensor const_133_to_fp16 = const()[name = tensor("const_133_to_fp16"), val = tensor(0x0p+0)]; - tensor input_359_cast_fp16 = pad(constant_val = const_133_to_fp16, mode = input_359_mode_0, pad = input_359_pad_0, x = input_357_cast_fp16)[name = tensor("input_359_cast_fp16")]; - tensor input_361_pad_type_0 = const()[name = tensor("input_361_pad_type_0"), val = tensor("valid")]; - tensor input_361_groups_0 = const()[name = tensor("input_361_groups_0"), val = tensor(512)]; - tensor input_361_strides_0 = const()[name = tensor("input_361_strides_0"), val = tensor([1])]; - tensor input_361_pad_0 = const()[name = tensor("input_361_pad_0"), val = tensor([0, 0])]; - tensor input_361_dilations_0 = const()[name = tensor("input_361_dilations_0"), val = tensor([1])]; - tensor const_246_to_fp16 = const()[name = tensor("const_246_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85555584)))]; - tensor const_247_to_fp16 = const()[name = tensor("const_247_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85564864)))]; - tensor input_363_cast_fp16 = conv(bias = const_247_to_fp16, dilations = input_361_dilations_0, groups = input_361_groups_0, pad = input_361_pad_0, pad_type = input_361_pad_type_0, strides = input_361_strides_0, weight = const_246_to_fp16, x = input_359_cast_fp16)[name = tensor("input_363_cast_fp16")]; - tensor input_365_cast_fp16 = silu(x = input_363_cast_fp16)[name = tensor("input_365_cast_fp16")]; - tensor x_153_pad_type_0 = const()[name = tensor("x_153_pad_type_0"), val = tensor("valid")]; - tensor x_153_strides_0 = const()[name = tensor("x_153_strides_0"), val = tensor([1])]; - tensor x_153_pad_0 = const()[name = tensor("x_153_pad_0"), val = tensor([0, 0])]; - tensor x_153_dilations_0 = const()[name = tensor("x_153_dilations_0"), val = tensor([1])]; - tensor x_153_groups_0 = const()[name = tensor("x_153_groups_0"), val = tensor(1)]; - tensor module_layers_6_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_6_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85565952)))]; - tensor module_layers_6_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_6_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86090304)))]; - tensor x_153_cast_fp16 = conv(bias = module_layers_6_conv_pointwise_conv2_bias_to_fp16, dilations = x_153_dilations_0, groups = x_153_groups_0, pad = x_153_pad_0, pad_type = x_153_pad_type_0, strides = x_153_strides_0, weight = module_layers_6_conv_pointwise_conv2_weight_to_fp16, x = input_365_cast_fp16)[name = tensor("x_153_cast_fp16")]; - tensor input_367_perm_0 = const()[name = tensor("input_367_perm_0"), val = tensor([0, 2, 1])]; - tensor input_367_cast_fp16 = transpose(perm = input_367_perm_0, x = x_153_cast_fp16)[name = tensor("transpose_156")]; - tensor input_369_cast_fp16 = add(x = input_351_cast_fp16, y = input_367_cast_fp16)[name = tensor("input_369_cast_fp16")]; - tensor input_371_axes_0 = const()[name = tensor("input_371_axes_0"), val = tensor([-1])]; - tensor module_layers_6_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_6_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86091392)))]; - tensor module_layers_6_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_6_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86092480)))]; - tensor input_371_cast_fp16 = layer_norm(axes = input_371_axes_0, beta = module_layers_6_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_6_norm_feed_forward2_weight_to_fp16, x = input_369_cast_fp16)[name = tensor("input_371_cast_fp16")]; - tensor module_layers_6_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_6_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86093568)))]; - tensor module_layers_6_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_6_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(88190784)))]; - tensor linear_62_cast_fp16 = linear(bias = module_layers_6_feed_forward2_linear1_bias_to_fp16, weight = module_layers_6_feed_forward2_linear1_weight_to_fp16, x = input_371_cast_fp16)[name = tensor("linear_62_cast_fp16")]; - tensor input_375_cast_fp16 = silu(x = linear_62_cast_fp16)[name = tensor("input_375_cast_fp16")]; - tensor module_layers_6_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_6_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(88194944)))]; - tensor module_layers_6_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_6_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90292160)))]; - tensor linear_63_cast_fp16 = linear(bias = module_layers_6_feed_forward2_linear2_bias_to_fp16, weight = module_layers_6_feed_forward2_linear2_weight_to_fp16, x = input_375_cast_fp16)[name = tensor("linear_63_cast_fp16")]; - tensor var_1539_to_fp16 = const()[name = tensor("op_1539_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1540_cast_fp16 = mul(x = linear_63_cast_fp16, y = var_1539_to_fp16)[name = tensor("op_1540_cast_fp16")]; - tensor input_381_cast_fp16 = add(x = input_369_cast_fp16, y = var_1540_cast_fp16)[name = tensor("input_381_cast_fp16")]; - tensor input_383_axes_0 = const()[name = tensor("input_383_axes_0"), val = tensor([-1])]; - tensor module_layers_6_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_6_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90293248)))]; - tensor module_layers_6_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_6_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90294336)))]; - tensor input_383_cast_fp16 = layer_norm(axes = input_383_axes_0, beta = module_layers_6_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_6_norm_out_weight_to_fp16, x = input_381_cast_fp16)[name = tensor("input_383_cast_fp16")]; - tensor input_385_axes_0 = const()[name = tensor("input_385_axes_0"), val = tensor([-1])]; - tensor module_layers_7_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_7_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90295424)))]; - tensor module_layers_7_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_7_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90296512)))]; - tensor input_385_cast_fp16 = layer_norm(axes = input_385_axes_0, beta = module_layers_7_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_7_norm_feed_forward1_weight_to_fp16, x = input_383_cast_fp16)[name = tensor("input_385_cast_fp16")]; - tensor module_layers_7_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_7_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90297600)))]; - tensor module_layers_7_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_7_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92394816)))]; - tensor linear_64_cast_fp16 = linear(bias = module_layers_7_feed_forward1_linear1_bias_to_fp16, weight = module_layers_7_feed_forward1_linear1_weight_to_fp16, x = input_385_cast_fp16)[name = tensor("linear_64_cast_fp16")]; - tensor input_389_cast_fp16 = silu(x = linear_64_cast_fp16)[name = tensor("input_389_cast_fp16")]; - tensor module_layers_7_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_7_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92398976)))]; - tensor module_layers_7_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_7_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(94496192)))]; - tensor linear_65_cast_fp16 = linear(bias = module_layers_7_feed_forward1_linear2_bias_to_fp16, weight = module_layers_7_feed_forward1_linear2_weight_to_fp16, x = input_389_cast_fp16)[name = tensor("linear_65_cast_fp16")]; - tensor var_1570_to_fp16 = const()[name = tensor("op_1570_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1571_cast_fp16 = mul(x = linear_65_cast_fp16, y = var_1570_to_fp16)[name = tensor("op_1571_cast_fp16")]; - tensor input_395_cast_fp16 = add(x = input_383_cast_fp16, y = var_1571_cast_fp16)[name = tensor("input_395_cast_fp16")]; - tensor query_15_axes_0 = const()[name = tensor("query_15_axes_0"), val = tensor([-1])]; - tensor module_layers_7_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_7_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(94497280)))]; - tensor module_layers_7_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_7_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(94498368)))]; - tensor query_15_cast_fp16 = layer_norm(axes = query_15_axes_0, beta = module_layers_7_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_7_norm_self_att_weight_to_fp16, x = input_395_cast_fp16)[name = tensor("query_15_cast_fp16")]; - tensor module_layers_7_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_7_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(94499456)))]; - tensor module_layers_7_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_7_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95023808)))]; - tensor linear_66_cast_fp16 = linear(bias = module_layers_7_self_attn_linear_q_bias_to_fp16, weight = module_layers_7_self_attn_linear_q_weight_to_fp16, x = query_15_cast_fp16)[name = tensor("linear_66_cast_fp16")]; - tensor var_1588 = const()[name = tensor("op_1588"), val = tensor([1, -1, 8, 64])]; - tensor q_43_cast_fp16 = reshape(shape = var_1588, x = linear_66_cast_fp16)[name = tensor("q_43_cast_fp16")]; - tensor module_layers_7_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_7_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95024896)))]; - tensor module_layers_7_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_7_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95549248)))]; - tensor linear_67_cast_fp16 = linear(bias = module_layers_7_self_attn_linear_k_bias_to_fp16, weight = module_layers_7_self_attn_linear_k_weight_to_fp16, x = query_15_cast_fp16)[name = tensor("linear_67_cast_fp16")]; - tensor var_1593 = const()[name = tensor("op_1593"), val = tensor([1, -1, 8, 64])]; - tensor k_29_cast_fp16 = reshape(shape = var_1593, x = linear_67_cast_fp16)[name = tensor("k_29_cast_fp16")]; - tensor module_layers_7_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_7_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95550336)))]; - tensor module_layers_7_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_7_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96074688)))]; - tensor linear_68_cast_fp16 = linear(bias = module_layers_7_self_attn_linear_v_bias_to_fp16, weight = module_layers_7_self_attn_linear_v_weight_to_fp16, x = query_15_cast_fp16)[name = tensor("linear_68_cast_fp16")]; - tensor var_1598 = const()[name = tensor("op_1598"), val = tensor([1, -1, 8, 64])]; - tensor v_15_cast_fp16 = reshape(shape = var_1598, x = linear_68_cast_fp16)[name = tensor("v_15_cast_fp16")]; - tensor value_17_perm_0 = const()[name = tensor("value_17_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_7_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_7_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96075776)))]; - tensor var_1610_cast_fp16 = add(x = q_43_cast_fp16, y = module_layers_7_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1610_cast_fp16")]; - tensor module_layers_7_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_7_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96076864)))]; - tensor var_1612_cast_fp16 = add(x = q_43_cast_fp16, y = module_layers_7_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1612_cast_fp16")]; - tensor q_with_bias_v_15_perm_0 = const()[name = tensor("q_with_bias_v_15_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_161_transpose_x_0 = const()[name = tensor("x_161_transpose_x_0"), val = tensor(false)]; - tensor x_161_transpose_y_0 = const()[name = tensor("x_161_transpose_y_0"), val = tensor(false)]; - tensor var_1614_to_fp16 = const()[name = tensor("op_1614_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96077952)))]; - tensor q_with_bias_v_15_cast_fp16 = transpose(perm = q_with_bias_v_15_perm_0, x = var_1612_cast_fp16)[name = tensor("transpose_154")]; - tensor x_161_cast_fp16 = matmul(transpose_x = x_161_transpose_x_0, transpose_y = x_161_transpose_y_0, x = q_with_bias_v_15_cast_fp16, y = var_1614_to_fp16)[name = tensor("x_161_cast_fp16")]; - tensor x_163_pad_0 = const()[name = tensor("x_163_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_163_mode_0 = const()[name = tensor("x_163_mode_0"), val = tensor("constant")]; - tensor const_140_to_fp16 = const()[name = tensor("const_140_to_fp16"), val = tensor(0x0p+0)]; - tensor x_163_cast_fp16 = pad(constant_val = const_140_to_fp16, mode = x_163_mode_0, pad = x_163_pad_0, x = x_161_cast_fp16)[name = tensor("x_163_cast_fp16")]; - tensor var_1622 = const()[name = tensor("op_1622"), val = tensor([1, 8, -1, 188])]; - tensor x_165_cast_fp16 = reshape(shape = var_1622, x = x_163_cast_fp16)[name = tensor("x_165_cast_fp16")]; - tensor var_1626_begin_0 = const()[name = tensor("op_1626_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_1626_end_0 = const()[name = tensor("op_1626_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_1626_end_mask_0 = const()[name = tensor("op_1626_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_1626_cast_fp16 = slice_by_index(begin = var_1626_begin_0, end = var_1626_end_0, end_mask = var_1626_end_mask_0, x = x_165_cast_fp16)[name = tensor("op_1626_cast_fp16")]; - tensor var_1627 = const()[name = tensor("op_1627"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_29_cast_fp16 = reshape(shape = var_1627, x = var_1626_cast_fp16)[name = tensor("matrix_bd_29_cast_fp16")]; - tensor matrix_ac_15_transpose_x_0 = const()[name = tensor("matrix_ac_15_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_15_transpose_y_0 = const()[name = tensor("matrix_ac_15_transpose_y_0"), val = tensor(false)]; - tensor transpose_65_perm_0 = const()[name = tensor("transpose_65_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_66_perm_0 = const()[name = tensor("transpose_66_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_66 = transpose(perm = transpose_66_perm_0, x = k_29_cast_fp16)[name = tensor("transpose_152")]; - tensor transpose_65 = transpose(perm = transpose_65_perm_0, x = var_1610_cast_fp16)[name = tensor("transpose_153")]; - tensor matrix_ac_15_cast_fp16 = matmul(transpose_x = matrix_ac_15_transpose_x_0, transpose_y = matrix_ac_15_transpose_y_0, x = transpose_65, y = transpose_66)[name = tensor("matrix_ac_15_cast_fp16")]; - tensor matrix_bd_31_begin_0 = const()[name = tensor("matrix_bd_31_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_31_end_0 = const()[name = tensor("matrix_bd_31_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_31_end_mask_0 = const()[name = tensor("matrix_bd_31_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_31_cast_fp16 = slice_by_index(begin = matrix_bd_31_begin_0, end = matrix_bd_31_end_0, end_mask = matrix_bd_31_end_mask_0, x = matrix_bd_29_cast_fp16)[name = tensor("matrix_bd_31_cast_fp16")]; - tensor var_1636_cast_fp16 = add(x = matrix_ac_15_cast_fp16, y = matrix_bd_31_cast_fp16)[name = tensor("op_1636_cast_fp16")]; - tensor _inversed_scores_29_y_0_to_fp16 = const()[name = tensor("_inversed_scores_29_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_29_cast_fp16 = mul(x = var_1636_cast_fp16, y = _inversed_scores_29_y_0_to_fp16)[name = tensor("_inversed_scores_29_cast_fp16")]; - tensor scores_31_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_29_cast_fp16, cond = mask_11)[name = tensor("scores_31_cast_fp16")]; - tensor var_1642_cast_fp16 = softmax(axis = var_23, x = scores_31_cast_fp16)[name = tensor("op_1642_cast_fp16")]; - tensor input_397_cast_fp16 = select(a = var_11_to_fp16, b = var_1642_cast_fp16, cond = mask_11)[name = tensor("input_397_cast_fp16")]; - tensor x_167_transpose_x_0 = const()[name = tensor("x_167_transpose_x_0"), val = tensor(false)]; - tensor x_167_transpose_y_0 = const()[name = tensor("x_167_transpose_y_0"), val = tensor(false)]; - tensor value_17_cast_fp16 = transpose(perm = value_17_perm_0, x = v_15_cast_fp16)[name = tensor("transpose_155")]; - tensor x_167_cast_fp16 = matmul(transpose_x = x_167_transpose_x_0, transpose_y = x_167_transpose_y_0, x = input_397_cast_fp16, y = value_17_cast_fp16)[name = tensor("x_167_cast_fp16")]; - tensor var_1646_perm_0 = const()[name = tensor("op_1646_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_1647 = const()[name = tensor("op_1647"), val = tensor([1, -1, 512])]; - tensor var_1646_cast_fp16 = transpose(perm = var_1646_perm_0, x = x_167_cast_fp16)[name = tensor("transpose_151")]; - tensor input_399_cast_fp16 = reshape(shape = var_1647, x = var_1646_cast_fp16)[name = tensor("input_399_cast_fp16")]; - tensor module_layers_7_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_7_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96462016)))]; - tensor module_layers_7_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_7_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96986368)))]; - tensor linear_70_cast_fp16 = linear(bias = module_layers_7_self_attn_linear_out_bias_to_fp16, weight = module_layers_7_self_attn_linear_out_weight_to_fp16, x = input_399_cast_fp16)[name = tensor("linear_70_cast_fp16")]; - tensor input_403_cast_fp16 = add(x = input_395_cast_fp16, y = linear_70_cast_fp16)[name = tensor("input_403_cast_fp16")]; - tensor x_171_axes_0 = const()[name = tensor("x_171_axes_0"), val = tensor([-1])]; - tensor module_layers_7_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_7_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96987456)))]; - tensor module_layers_7_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_7_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96988544)))]; - tensor x_171_cast_fp16 = layer_norm(axes = x_171_axes_0, beta = module_layers_7_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_7_norm_conv_weight_to_fp16, x = input_403_cast_fp16)[name = tensor("x_171_cast_fp16")]; - tensor input_405_perm_0 = const()[name = tensor("input_405_perm_0"), val = tensor([0, 2, 1])]; - tensor input_407_pad_type_0 = const()[name = tensor("input_407_pad_type_0"), val = tensor("valid")]; - tensor input_407_strides_0 = const()[name = tensor("input_407_strides_0"), val = tensor([1])]; - tensor input_407_pad_0 = const()[name = tensor("input_407_pad_0"), val = tensor([0, 0])]; - tensor input_407_dilations_0 = const()[name = tensor("input_407_dilations_0"), val = tensor([1])]; - tensor input_407_groups_0 = const()[name = tensor("input_407_groups_0"), val = tensor(1)]; - tensor module_layers_7_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_7_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96989632)))]; - tensor module_layers_7_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_7_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98038272)))]; - tensor input_405_cast_fp16 = transpose(perm = input_405_perm_0, x = x_171_cast_fp16)[name = tensor("transpose_150")]; - tensor input_407_cast_fp16 = conv(bias = module_layers_7_conv_pointwise_conv1_bias_to_fp16, dilations = input_407_dilations_0, groups = input_407_groups_0, pad = input_407_pad_0, pad_type = input_407_pad_type_0, strides = input_407_strides_0, weight = module_layers_7_conv_pointwise_conv1_weight_to_fp16, x = input_405_cast_fp16)[name = tensor("input_407_cast_fp16")]; - tensor x_173_split_num_splits_0 = const()[name = tensor("x_173_split_num_splits_0"), val = tensor(2)]; - tensor x_173_split_axis_0 = const()[name = tensor("x_173_split_axis_0"), val = tensor(1)]; - tensor x_173_split_cast_fp16_0, tensor x_173_split_cast_fp16_1 = split(axis = x_173_split_axis_0, num_splits = x_173_split_num_splits_0, x = input_407_cast_fp16)[name = tensor("x_173_split_cast_fp16")]; - tensor x_173_split_1_sigmoid_cast_fp16 = sigmoid(x = x_173_split_cast_fp16_1)[name = tensor("x_173_split_1_sigmoid_cast_fp16")]; - tensor x_173_cast_fp16 = mul(x = x_173_split_cast_fp16_0, y = x_173_split_1_sigmoid_cast_fp16)[name = tensor("x_173_cast_fp16")]; - tensor input_409_cast_fp16 = select(a = var_11_to_fp16, b = x_173_cast_fp16, cond = var_453)[name = tensor("input_409_cast_fp16")]; - tensor input_411_pad_0 = const()[name = tensor("input_411_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_411_mode_0 = const()[name = tensor("input_411_mode_0"), val = tensor("constant")]; - tensor const_143_to_fp16 = const()[name = tensor("const_143_to_fp16"), val = tensor(0x0p+0)]; - tensor input_411_cast_fp16 = pad(constant_val = const_143_to_fp16, mode = input_411_mode_0, pad = input_411_pad_0, x = input_409_cast_fp16)[name = tensor("input_411_cast_fp16")]; - tensor input_413_pad_type_0 = const()[name = tensor("input_413_pad_type_0"), val = tensor("valid")]; - tensor input_413_groups_0 = const()[name = tensor("input_413_groups_0"), val = tensor(512)]; - tensor input_413_strides_0 = const()[name = tensor("input_413_strides_0"), val = tensor([1])]; - tensor input_413_pad_0 = const()[name = tensor("input_413_pad_0"), val = tensor([0, 0])]; - tensor input_413_dilations_0 = const()[name = tensor("input_413_dilations_0"), val = tensor([1])]; - tensor const_248_to_fp16 = const()[name = tensor("const_248_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98040384)))]; - tensor const_249_to_fp16 = const()[name = tensor("const_249_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98049664)))]; - tensor input_415_cast_fp16 = conv(bias = const_249_to_fp16, dilations = input_413_dilations_0, groups = input_413_groups_0, pad = input_413_pad_0, pad_type = input_413_pad_type_0, strides = input_413_strides_0, weight = const_248_to_fp16, x = input_411_cast_fp16)[name = tensor("input_415_cast_fp16")]; - tensor input_417_cast_fp16 = silu(x = input_415_cast_fp16)[name = tensor("input_417_cast_fp16")]; - tensor x_175_pad_type_0 = const()[name = tensor("x_175_pad_type_0"), val = tensor("valid")]; - tensor x_175_strides_0 = const()[name = tensor("x_175_strides_0"), val = tensor([1])]; - tensor x_175_pad_0 = const()[name = tensor("x_175_pad_0"), val = tensor([0, 0])]; - tensor x_175_dilations_0 = const()[name = tensor("x_175_dilations_0"), val = tensor([1])]; - tensor x_175_groups_0 = const()[name = tensor("x_175_groups_0"), val = tensor(1)]; - tensor module_layers_7_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_7_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98050752)))]; - tensor module_layers_7_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_7_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98575104)))]; - tensor x_175_cast_fp16 = conv(bias = module_layers_7_conv_pointwise_conv2_bias_to_fp16, dilations = x_175_dilations_0, groups = x_175_groups_0, pad = x_175_pad_0, pad_type = x_175_pad_type_0, strides = x_175_strides_0, weight = module_layers_7_conv_pointwise_conv2_weight_to_fp16, x = input_417_cast_fp16)[name = tensor("x_175_cast_fp16")]; - tensor input_419_perm_0 = const()[name = tensor("input_419_perm_0"), val = tensor([0, 2, 1])]; - tensor input_419_cast_fp16 = transpose(perm = input_419_perm_0, x = x_175_cast_fp16)[name = tensor("transpose_149")]; - tensor input_421_cast_fp16 = add(x = input_403_cast_fp16, y = input_419_cast_fp16)[name = tensor("input_421_cast_fp16")]; - tensor input_423_axes_0 = const()[name = tensor("input_423_axes_0"), val = tensor([-1])]; - tensor module_layers_7_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_7_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98576192)))]; - tensor module_layers_7_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_7_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98577280)))]; - tensor input_423_cast_fp16 = layer_norm(axes = input_423_axes_0, beta = module_layers_7_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_7_norm_feed_forward2_weight_to_fp16, x = input_421_cast_fp16)[name = tensor("input_423_cast_fp16")]; - tensor module_layers_7_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_7_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98578368)))]; - tensor module_layers_7_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_7_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(100675584)))]; - tensor linear_71_cast_fp16 = linear(bias = module_layers_7_feed_forward2_linear1_bias_to_fp16, weight = module_layers_7_feed_forward2_linear1_weight_to_fp16, x = input_423_cast_fp16)[name = tensor("linear_71_cast_fp16")]; - tensor input_427_cast_fp16 = silu(x = linear_71_cast_fp16)[name = tensor("input_427_cast_fp16")]; - tensor module_layers_7_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_7_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(100679744)))]; - tensor module_layers_7_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_7_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102776960)))]; - tensor linear_72_cast_fp16 = linear(bias = module_layers_7_feed_forward2_linear2_bias_to_fp16, weight = module_layers_7_feed_forward2_linear2_weight_to_fp16, x = input_427_cast_fp16)[name = tensor("linear_72_cast_fp16")]; - tensor var_1713_to_fp16 = const()[name = tensor("op_1713_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1714_cast_fp16 = mul(x = linear_72_cast_fp16, y = var_1713_to_fp16)[name = tensor("op_1714_cast_fp16")]; - tensor input_433_cast_fp16 = add(x = input_421_cast_fp16, y = var_1714_cast_fp16)[name = tensor("input_433_cast_fp16")]; - tensor input_435_axes_0 = const()[name = tensor("input_435_axes_0"), val = tensor([-1])]; - tensor module_layers_7_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_7_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102778048)))]; - tensor module_layers_7_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_7_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102779136)))]; - tensor input_435_cast_fp16 = layer_norm(axes = input_435_axes_0, beta = module_layers_7_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_7_norm_out_weight_to_fp16, x = input_433_cast_fp16)[name = tensor("input_435_cast_fp16")]; - tensor input_437_axes_0 = const()[name = tensor("input_437_axes_0"), val = tensor([-1])]; - tensor module_layers_8_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_8_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102780224)))]; - tensor module_layers_8_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_8_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102781312)))]; - tensor input_437_cast_fp16 = layer_norm(axes = input_437_axes_0, beta = module_layers_8_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_8_norm_feed_forward1_weight_to_fp16, x = input_435_cast_fp16)[name = tensor("input_437_cast_fp16")]; - tensor module_layers_8_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_8_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102782400)))]; - tensor module_layers_8_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_8_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(104879616)))]; - tensor linear_73_cast_fp16 = linear(bias = module_layers_8_feed_forward1_linear1_bias_to_fp16, weight = module_layers_8_feed_forward1_linear1_weight_to_fp16, x = input_437_cast_fp16)[name = tensor("linear_73_cast_fp16")]; - tensor input_441_cast_fp16 = silu(x = linear_73_cast_fp16)[name = tensor("input_441_cast_fp16")]; - tensor module_layers_8_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_8_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(104883776)))]; - tensor module_layers_8_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_8_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(106980992)))]; - tensor linear_74_cast_fp16 = linear(bias = module_layers_8_feed_forward1_linear2_bias_to_fp16, weight = module_layers_8_feed_forward1_linear2_weight_to_fp16, x = input_441_cast_fp16)[name = tensor("linear_74_cast_fp16")]; - tensor var_1744_to_fp16 = const()[name = tensor("op_1744_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1745_cast_fp16 = mul(x = linear_74_cast_fp16, y = var_1744_to_fp16)[name = tensor("op_1745_cast_fp16")]; - tensor input_447_cast_fp16 = add(x = input_435_cast_fp16, y = var_1745_cast_fp16)[name = tensor("input_447_cast_fp16")]; - tensor query_17_axes_0 = const()[name = tensor("query_17_axes_0"), val = tensor([-1])]; - tensor module_layers_8_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_8_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(106982080)))]; - tensor module_layers_8_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_8_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(106983168)))]; - tensor query_17_cast_fp16 = layer_norm(axes = query_17_axes_0, beta = module_layers_8_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_8_norm_self_att_weight_to_fp16, x = input_447_cast_fp16)[name = tensor("query_17_cast_fp16")]; - tensor module_layers_8_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_8_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(106984256)))]; - tensor module_layers_8_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_8_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(107508608)))]; - tensor linear_75_cast_fp16 = linear(bias = module_layers_8_self_attn_linear_q_bias_to_fp16, weight = module_layers_8_self_attn_linear_q_weight_to_fp16, x = query_17_cast_fp16)[name = tensor("linear_75_cast_fp16")]; - tensor var_1762 = const()[name = tensor("op_1762"), val = tensor([1, -1, 8, 64])]; - tensor q_49_cast_fp16 = reshape(shape = var_1762, x = linear_75_cast_fp16)[name = tensor("q_49_cast_fp16")]; - tensor module_layers_8_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_8_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(107509696)))]; - tensor module_layers_8_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_8_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108034048)))]; - tensor linear_76_cast_fp16 = linear(bias = module_layers_8_self_attn_linear_k_bias_to_fp16, weight = module_layers_8_self_attn_linear_k_weight_to_fp16, x = query_17_cast_fp16)[name = tensor("linear_76_cast_fp16")]; - tensor var_1767 = const()[name = tensor("op_1767"), val = tensor([1, -1, 8, 64])]; - tensor k_33_cast_fp16 = reshape(shape = var_1767, x = linear_76_cast_fp16)[name = tensor("k_33_cast_fp16")]; - tensor module_layers_8_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_8_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108035136)))]; - tensor module_layers_8_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_8_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108559488)))]; - tensor linear_77_cast_fp16 = linear(bias = module_layers_8_self_attn_linear_v_bias_to_fp16, weight = module_layers_8_self_attn_linear_v_weight_to_fp16, x = query_17_cast_fp16)[name = tensor("linear_77_cast_fp16")]; - tensor var_1772 = const()[name = tensor("op_1772"), val = tensor([1, -1, 8, 64])]; - tensor v_17_cast_fp16 = reshape(shape = var_1772, x = linear_77_cast_fp16)[name = tensor("v_17_cast_fp16")]; - tensor value_19_perm_0 = const()[name = tensor("value_19_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_8_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_8_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108560576)))]; - tensor var_1784_cast_fp16 = add(x = q_49_cast_fp16, y = module_layers_8_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1784_cast_fp16")]; - tensor module_layers_8_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_8_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108561664)))]; - tensor var_1786_cast_fp16 = add(x = q_49_cast_fp16, y = module_layers_8_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1786_cast_fp16")]; - tensor q_with_bias_v_17_perm_0 = const()[name = tensor("q_with_bias_v_17_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_183_transpose_x_0 = const()[name = tensor("x_183_transpose_x_0"), val = tensor(false)]; - tensor x_183_transpose_y_0 = const()[name = tensor("x_183_transpose_y_0"), val = tensor(false)]; - tensor var_1788_to_fp16 = const()[name = tensor("op_1788_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108562752)))]; - tensor q_with_bias_v_17_cast_fp16 = transpose(perm = q_with_bias_v_17_perm_0, x = var_1786_cast_fp16)[name = tensor("transpose_147")]; - tensor x_183_cast_fp16 = matmul(transpose_x = x_183_transpose_x_0, transpose_y = x_183_transpose_y_0, x = q_with_bias_v_17_cast_fp16, y = var_1788_to_fp16)[name = tensor("x_183_cast_fp16")]; - tensor x_185_pad_0 = const()[name = tensor("x_185_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_185_mode_0 = const()[name = tensor("x_185_mode_0"), val = tensor("constant")]; - tensor const_150_to_fp16 = const()[name = tensor("const_150_to_fp16"), val = tensor(0x0p+0)]; - tensor x_185_cast_fp16 = pad(constant_val = const_150_to_fp16, mode = x_185_mode_0, pad = x_185_pad_0, x = x_183_cast_fp16)[name = tensor("x_185_cast_fp16")]; - tensor var_1796 = const()[name = tensor("op_1796"), val = tensor([1, 8, -1, 188])]; - tensor x_187_cast_fp16 = reshape(shape = var_1796, x = x_185_cast_fp16)[name = tensor("x_187_cast_fp16")]; - tensor var_1800_begin_0 = const()[name = tensor("op_1800_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_1800_end_0 = const()[name = tensor("op_1800_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_1800_end_mask_0 = const()[name = tensor("op_1800_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_1800_cast_fp16 = slice_by_index(begin = var_1800_begin_0, end = var_1800_end_0, end_mask = var_1800_end_mask_0, x = x_187_cast_fp16)[name = tensor("op_1800_cast_fp16")]; - tensor var_1801 = const()[name = tensor("op_1801"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_33_cast_fp16 = reshape(shape = var_1801, x = var_1800_cast_fp16)[name = tensor("matrix_bd_33_cast_fp16")]; - tensor matrix_ac_17_transpose_x_0 = const()[name = tensor("matrix_ac_17_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_17_transpose_y_0 = const()[name = tensor("matrix_ac_17_transpose_y_0"), val = tensor(false)]; - tensor transpose_67_perm_0 = const()[name = tensor("transpose_67_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_68_perm_0 = const()[name = tensor("transpose_68_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_68 = transpose(perm = transpose_68_perm_0, x = k_33_cast_fp16)[name = tensor("transpose_145")]; - tensor transpose_67 = transpose(perm = transpose_67_perm_0, x = var_1784_cast_fp16)[name = tensor("transpose_146")]; - tensor matrix_ac_17_cast_fp16 = matmul(transpose_x = matrix_ac_17_transpose_x_0, transpose_y = matrix_ac_17_transpose_y_0, x = transpose_67, y = transpose_68)[name = tensor("matrix_ac_17_cast_fp16")]; - tensor matrix_bd_35_begin_0 = const()[name = tensor("matrix_bd_35_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_35_end_0 = const()[name = tensor("matrix_bd_35_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_35_end_mask_0 = const()[name = tensor("matrix_bd_35_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_35_cast_fp16 = slice_by_index(begin = matrix_bd_35_begin_0, end = matrix_bd_35_end_0, end_mask = matrix_bd_35_end_mask_0, x = matrix_bd_33_cast_fp16)[name = tensor("matrix_bd_35_cast_fp16")]; - tensor var_1810_cast_fp16 = add(x = matrix_ac_17_cast_fp16, y = matrix_bd_35_cast_fp16)[name = tensor("op_1810_cast_fp16")]; - tensor _inversed_scores_33_y_0_to_fp16 = const()[name = tensor("_inversed_scores_33_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_33_cast_fp16 = mul(x = var_1810_cast_fp16, y = _inversed_scores_33_y_0_to_fp16)[name = tensor("_inversed_scores_33_cast_fp16")]; - tensor scores_35_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_33_cast_fp16, cond = mask_11)[name = tensor("scores_35_cast_fp16")]; - tensor var_1816_cast_fp16 = softmax(axis = var_23, x = scores_35_cast_fp16)[name = tensor("op_1816_cast_fp16")]; - tensor input_449_cast_fp16 = select(a = var_11_to_fp16, b = var_1816_cast_fp16, cond = mask_11)[name = tensor("input_449_cast_fp16")]; - tensor x_189_transpose_x_0 = const()[name = tensor("x_189_transpose_x_0"), val = tensor(false)]; - tensor x_189_transpose_y_0 = const()[name = tensor("x_189_transpose_y_0"), val = tensor(false)]; - tensor value_19_cast_fp16 = transpose(perm = value_19_perm_0, x = v_17_cast_fp16)[name = tensor("transpose_148")]; - tensor x_189_cast_fp16 = matmul(transpose_x = x_189_transpose_x_0, transpose_y = x_189_transpose_y_0, x = input_449_cast_fp16, y = value_19_cast_fp16)[name = tensor("x_189_cast_fp16")]; - tensor var_1820_perm_0 = const()[name = tensor("op_1820_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_1821 = const()[name = tensor("op_1821"), val = tensor([1, -1, 512])]; - tensor var_1820_cast_fp16 = transpose(perm = var_1820_perm_0, x = x_189_cast_fp16)[name = tensor("transpose_144")]; - tensor input_451_cast_fp16 = reshape(shape = var_1821, x = var_1820_cast_fp16)[name = tensor("input_451_cast_fp16")]; - tensor module_layers_8_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_8_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108946816)))]; - tensor module_layers_8_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_8_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109471168)))]; - tensor linear_79_cast_fp16 = linear(bias = module_layers_8_self_attn_linear_out_bias_to_fp16, weight = module_layers_8_self_attn_linear_out_weight_to_fp16, x = input_451_cast_fp16)[name = tensor("linear_79_cast_fp16")]; - tensor input_455_cast_fp16 = add(x = input_447_cast_fp16, y = linear_79_cast_fp16)[name = tensor("input_455_cast_fp16")]; - tensor x_193_axes_0 = const()[name = tensor("x_193_axes_0"), val = tensor([-1])]; - tensor module_layers_8_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_8_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109472256)))]; - tensor module_layers_8_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_8_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109473344)))]; - tensor x_193_cast_fp16 = layer_norm(axes = x_193_axes_0, beta = module_layers_8_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_8_norm_conv_weight_to_fp16, x = input_455_cast_fp16)[name = tensor("x_193_cast_fp16")]; - tensor input_457_perm_0 = const()[name = tensor("input_457_perm_0"), val = tensor([0, 2, 1])]; - tensor input_459_pad_type_0 = const()[name = tensor("input_459_pad_type_0"), val = tensor("valid")]; - tensor input_459_strides_0 = const()[name = tensor("input_459_strides_0"), val = tensor([1])]; - tensor input_459_pad_0 = const()[name = tensor("input_459_pad_0"), val = tensor([0, 0])]; - tensor input_459_dilations_0 = const()[name = tensor("input_459_dilations_0"), val = tensor([1])]; - tensor input_459_groups_0 = const()[name = tensor("input_459_groups_0"), val = tensor(1)]; - tensor module_layers_8_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_8_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109474432)))]; - tensor module_layers_8_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_8_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110523072)))]; - tensor input_457_cast_fp16 = transpose(perm = input_457_perm_0, x = x_193_cast_fp16)[name = tensor("transpose_143")]; - tensor input_459_cast_fp16 = conv(bias = module_layers_8_conv_pointwise_conv1_bias_to_fp16, dilations = input_459_dilations_0, groups = input_459_groups_0, pad = input_459_pad_0, pad_type = input_459_pad_type_0, strides = input_459_strides_0, weight = module_layers_8_conv_pointwise_conv1_weight_to_fp16, x = input_457_cast_fp16)[name = tensor("input_459_cast_fp16")]; - tensor x_195_split_num_splits_0 = const()[name = tensor("x_195_split_num_splits_0"), val = tensor(2)]; - tensor x_195_split_axis_0 = const()[name = tensor("x_195_split_axis_0"), val = tensor(1)]; - tensor x_195_split_cast_fp16_0, tensor x_195_split_cast_fp16_1 = split(axis = x_195_split_axis_0, num_splits = x_195_split_num_splits_0, x = input_459_cast_fp16)[name = tensor("x_195_split_cast_fp16")]; - tensor x_195_split_1_sigmoid_cast_fp16 = sigmoid(x = x_195_split_cast_fp16_1)[name = tensor("x_195_split_1_sigmoid_cast_fp16")]; - tensor x_195_cast_fp16 = mul(x = x_195_split_cast_fp16_0, y = x_195_split_1_sigmoid_cast_fp16)[name = tensor("x_195_cast_fp16")]; - tensor input_461_cast_fp16 = select(a = var_11_to_fp16, b = x_195_cast_fp16, cond = var_453)[name = tensor("input_461_cast_fp16")]; - tensor input_463_pad_0 = const()[name = tensor("input_463_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_463_mode_0 = const()[name = tensor("input_463_mode_0"), val = tensor("constant")]; - tensor const_153_to_fp16 = const()[name = tensor("const_153_to_fp16"), val = tensor(0x0p+0)]; - tensor input_463_cast_fp16 = pad(constant_val = const_153_to_fp16, mode = input_463_mode_0, pad = input_463_pad_0, x = input_461_cast_fp16)[name = tensor("input_463_cast_fp16")]; - tensor input_465_pad_type_0 = const()[name = tensor("input_465_pad_type_0"), val = tensor("valid")]; - tensor input_465_groups_0 = const()[name = tensor("input_465_groups_0"), val = tensor(512)]; - tensor input_465_strides_0 = const()[name = tensor("input_465_strides_0"), val = tensor([1])]; - tensor input_465_pad_0 = const()[name = tensor("input_465_pad_0"), val = tensor([0, 0])]; - tensor input_465_dilations_0 = const()[name = tensor("input_465_dilations_0"), val = tensor([1])]; - tensor const_250_to_fp16 = const()[name = tensor("const_250_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110525184)))]; - tensor const_251_to_fp16 = const()[name = tensor("const_251_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110534464)))]; - tensor input_467_cast_fp16 = conv(bias = const_251_to_fp16, dilations = input_465_dilations_0, groups = input_465_groups_0, pad = input_465_pad_0, pad_type = input_465_pad_type_0, strides = input_465_strides_0, weight = const_250_to_fp16, x = input_463_cast_fp16)[name = tensor("input_467_cast_fp16")]; - tensor input_469_cast_fp16 = silu(x = input_467_cast_fp16)[name = tensor("input_469_cast_fp16")]; - tensor x_197_pad_type_0 = const()[name = tensor("x_197_pad_type_0"), val = tensor("valid")]; - tensor x_197_strides_0 = const()[name = tensor("x_197_strides_0"), val = tensor([1])]; - tensor x_197_pad_0 = const()[name = tensor("x_197_pad_0"), val = tensor([0, 0])]; - tensor x_197_dilations_0 = const()[name = tensor("x_197_dilations_0"), val = tensor([1])]; - tensor x_197_groups_0 = const()[name = tensor("x_197_groups_0"), val = tensor(1)]; - tensor module_layers_8_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_8_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110535552)))]; - tensor module_layers_8_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_8_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(111059904)))]; - tensor x_197_cast_fp16 = conv(bias = module_layers_8_conv_pointwise_conv2_bias_to_fp16, dilations = x_197_dilations_0, groups = x_197_groups_0, pad = x_197_pad_0, pad_type = x_197_pad_type_0, strides = x_197_strides_0, weight = module_layers_8_conv_pointwise_conv2_weight_to_fp16, x = input_469_cast_fp16)[name = tensor("x_197_cast_fp16")]; - tensor input_471_perm_0 = const()[name = tensor("input_471_perm_0"), val = tensor([0, 2, 1])]; - tensor input_471_cast_fp16 = transpose(perm = input_471_perm_0, x = x_197_cast_fp16)[name = tensor("transpose_142")]; - tensor input_473_cast_fp16 = add(x = input_455_cast_fp16, y = input_471_cast_fp16)[name = tensor("input_473_cast_fp16")]; - tensor input_475_axes_0 = const()[name = tensor("input_475_axes_0"), val = tensor([-1])]; - tensor module_layers_8_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_8_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(111060992)))]; - tensor module_layers_8_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_8_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(111062080)))]; - tensor input_475_cast_fp16 = layer_norm(axes = input_475_axes_0, beta = module_layers_8_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_8_norm_feed_forward2_weight_to_fp16, x = input_473_cast_fp16)[name = tensor("input_475_cast_fp16")]; - tensor module_layers_8_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_8_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(111063168)))]; - tensor module_layers_8_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_8_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(113160384)))]; - tensor linear_80_cast_fp16 = linear(bias = module_layers_8_feed_forward2_linear1_bias_to_fp16, weight = module_layers_8_feed_forward2_linear1_weight_to_fp16, x = input_475_cast_fp16)[name = tensor("linear_80_cast_fp16")]; - tensor input_479_cast_fp16 = silu(x = linear_80_cast_fp16)[name = tensor("input_479_cast_fp16")]; - tensor module_layers_8_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_8_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(113164544)))]; - tensor module_layers_8_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_8_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(115261760)))]; - tensor linear_81_cast_fp16 = linear(bias = module_layers_8_feed_forward2_linear2_bias_to_fp16, weight = module_layers_8_feed_forward2_linear2_weight_to_fp16, x = input_479_cast_fp16)[name = tensor("linear_81_cast_fp16")]; - tensor var_1887_to_fp16 = const()[name = tensor("op_1887_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1888_cast_fp16 = mul(x = linear_81_cast_fp16, y = var_1887_to_fp16)[name = tensor("op_1888_cast_fp16")]; - tensor input_485_cast_fp16 = add(x = input_473_cast_fp16, y = var_1888_cast_fp16)[name = tensor("input_485_cast_fp16")]; - tensor input_487_axes_0 = const()[name = tensor("input_487_axes_0"), val = tensor([-1])]; - tensor module_layers_8_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_8_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(115262848)))]; - tensor module_layers_8_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_8_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(115263936)))]; - tensor input_487_cast_fp16 = layer_norm(axes = input_487_axes_0, beta = module_layers_8_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_8_norm_out_weight_to_fp16, x = input_485_cast_fp16)[name = tensor("input_487_cast_fp16")]; - tensor input_489_axes_0 = const()[name = tensor("input_489_axes_0"), val = tensor([-1])]; - tensor module_layers_9_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_9_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(115265024)))]; - tensor module_layers_9_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_9_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(115266112)))]; - tensor input_489_cast_fp16 = layer_norm(axes = input_489_axes_0, beta = module_layers_9_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_9_norm_feed_forward1_weight_to_fp16, x = input_487_cast_fp16)[name = tensor("input_489_cast_fp16")]; - tensor module_layers_9_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_9_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(115267200)))]; - tensor module_layers_9_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_9_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(117364416)))]; - tensor linear_82_cast_fp16 = linear(bias = module_layers_9_feed_forward1_linear1_bias_to_fp16, weight = module_layers_9_feed_forward1_linear1_weight_to_fp16, x = input_489_cast_fp16)[name = tensor("linear_82_cast_fp16")]; - tensor input_493_cast_fp16 = silu(x = linear_82_cast_fp16)[name = tensor("input_493_cast_fp16")]; - tensor module_layers_9_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_9_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(117368576)))]; - tensor module_layers_9_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_9_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(119465792)))]; - tensor linear_83_cast_fp16 = linear(bias = module_layers_9_feed_forward1_linear2_bias_to_fp16, weight = module_layers_9_feed_forward1_linear2_weight_to_fp16, x = input_493_cast_fp16)[name = tensor("linear_83_cast_fp16")]; - tensor var_1918_to_fp16 = const()[name = tensor("op_1918_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1919_cast_fp16 = mul(x = linear_83_cast_fp16, y = var_1918_to_fp16)[name = tensor("op_1919_cast_fp16")]; - tensor input_499_cast_fp16 = add(x = input_487_cast_fp16, y = var_1919_cast_fp16)[name = tensor("input_499_cast_fp16")]; - tensor query_19_axes_0 = const()[name = tensor("query_19_axes_0"), val = tensor([-1])]; - tensor module_layers_9_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_9_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(119466880)))]; - tensor module_layers_9_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_9_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(119467968)))]; - tensor query_19_cast_fp16 = layer_norm(axes = query_19_axes_0, beta = module_layers_9_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_9_norm_self_att_weight_to_fp16, x = input_499_cast_fp16)[name = tensor("query_19_cast_fp16")]; - tensor module_layers_9_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_9_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(119469056)))]; - tensor module_layers_9_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_9_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(119993408)))]; - tensor linear_84_cast_fp16 = linear(bias = module_layers_9_self_attn_linear_q_bias_to_fp16, weight = module_layers_9_self_attn_linear_q_weight_to_fp16, x = query_19_cast_fp16)[name = tensor("linear_84_cast_fp16")]; - tensor var_1936 = const()[name = tensor("op_1936"), val = tensor([1, -1, 8, 64])]; - tensor q_55_cast_fp16 = reshape(shape = var_1936, x = linear_84_cast_fp16)[name = tensor("q_55_cast_fp16")]; - tensor module_layers_9_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_9_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(119994496)))]; - tensor module_layers_9_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_9_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(120518848)))]; - tensor linear_85_cast_fp16 = linear(bias = module_layers_9_self_attn_linear_k_bias_to_fp16, weight = module_layers_9_self_attn_linear_k_weight_to_fp16, x = query_19_cast_fp16)[name = tensor("linear_85_cast_fp16")]; - tensor var_1941 = const()[name = tensor("op_1941"), val = tensor([1, -1, 8, 64])]; - tensor k_37_cast_fp16 = reshape(shape = var_1941, x = linear_85_cast_fp16)[name = tensor("k_37_cast_fp16")]; - tensor module_layers_9_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_9_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(120519936)))]; - tensor module_layers_9_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_9_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121044288)))]; - tensor linear_86_cast_fp16 = linear(bias = module_layers_9_self_attn_linear_v_bias_to_fp16, weight = module_layers_9_self_attn_linear_v_weight_to_fp16, x = query_19_cast_fp16)[name = tensor("linear_86_cast_fp16")]; - tensor var_1946 = const()[name = tensor("op_1946"), val = tensor([1, -1, 8, 64])]; - tensor v_19_cast_fp16 = reshape(shape = var_1946, x = linear_86_cast_fp16)[name = tensor("v_19_cast_fp16")]; - tensor value_21_perm_0 = const()[name = tensor("value_21_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_9_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_9_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121045376)))]; - tensor var_1958_cast_fp16 = add(x = q_55_cast_fp16, y = module_layers_9_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1958_cast_fp16")]; - tensor module_layers_9_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_9_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121046464)))]; - tensor var_1960_cast_fp16 = add(x = q_55_cast_fp16, y = module_layers_9_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1960_cast_fp16")]; - tensor q_with_bias_v_19_perm_0 = const()[name = tensor("q_with_bias_v_19_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_205_transpose_x_0 = const()[name = tensor("x_205_transpose_x_0"), val = tensor(false)]; - tensor x_205_transpose_y_0 = const()[name = tensor("x_205_transpose_y_0"), val = tensor(false)]; - tensor var_1962_to_fp16 = const()[name = tensor("op_1962_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121047552)))]; - tensor q_with_bias_v_19_cast_fp16 = transpose(perm = q_with_bias_v_19_perm_0, x = var_1960_cast_fp16)[name = tensor("transpose_140")]; - tensor x_205_cast_fp16 = matmul(transpose_x = x_205_transpose_x_0, transpose_y = x_205_transpose_y_0, x = q_with_bias_v_19_cast_fp16, y = var_1962_to_fp16)[name = tensor("x_205_cast_fp16")]; - tensor x_207_pad_0 = const()[name = tensor("x_207_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_207_mode_0 = const()[name = tensor("x_207_mode_0"), val = tensor("constant")]; - tensor const_160_to_fp16 = const()[name = tensor("const_160_to_fp16"), val = tensor(0x0p+0)]; - tensor x_207_cast_fp16 = pad(constant_val = const_160_to_fp16, mode = x_207_mode_0, pad = x_207_pad_0, x = x_205_cast_fp16)[name = tensor("x_207_cast_fp16")]; - tensor var_1970 = const()[name = tensor("op_1970"), val = tensor([1, 8, -1, 188])]; - tensor x_209_cast_fp16 = reshape(shape = var_1970, x = x_207_cast_fp16)[name = tensor("x_209_cast_fp16")]; - tensor var_1974_begin_0 = const()[name = tensor("op_1974_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_1974_end_0 = const()[name = tensor("op_1974_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_1974_end_mask_0 = const()[name = tensor("op_1974_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_1974_cast_fp16 = slice_by_index(begin = var_1974_begin_0, end = var_1974_end_0, end_mask = var_1974_end_mask_0, x = x_209_cast_fp16)[name = tensor("op_1974_cast_fp16")]; - tensor var_1975 = const()[name = tensor("op_1975"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_37_cast_fp16 = reshape(shape = var_1975, x = var_1974_cast_fp16)[name = tensor("matrix_bd_37_cast_fp16")]; - tensor matrix_ac_19_transpose_x_0 = const()[name = tensor("matrix_ac_19_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_19_transpose_y_0 = const()[name = tensor("matrix_ac_19_transpose_y_0"), val = tensor(false)]; - tensor transpose_69_perm_0 = const()[name = tensor("transpose_69_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_70_perm_0 = const()[name = tensor("transpose_70_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_70 = transpose(perm = transpose_70_perm_0, x = k_37_cast_fp16)[name = tensor("transpose_138")]; - tensor transpose_69 = transpose(perm = transpose_69_perm_0, x = var_1958_cast_fp16)[name = tensor("transpose_139")]; - tensor matrix_ac_19_cast_fp16 = matmul(transpose_x = matrix_ac_19_transpose_x_0, transpose_y = matrix_ac_19_transpose_y_0, x = transpose_69, y = transpose_70)[name = tensor("matrix_ac_19_cast_fp16")]; - tensor matrix_bd_39_begin_0 = const()[name = tensor("matrix_bd_39_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_39_end_0 = const()[name = tensor("matrix_bd_39_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_39_end_mask_0 = const()[name = tensor("matrix_bd_39_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_39_cast_fp16 = slice_by_index(begin = matrix_bd_39_begin_0, end = matrix_bd_39_end_0, end_mask = matrix_bd_39_end_mask_0, x = matrix_bd_37_cast_fp16)[name = tensor("matrix_bd_39_cast_fp16")]; - tensor var_1984_cast_fp16 = add(x = matrix_ac_19_cast_fp16, y = matrix_bd_39_cast_fp16)[name = tensor("op_1984_cast_fp16")]; - tensor _inversed_scores_37_y_0_to_fp16 = const()[name = tensor("_inversed_scores_37_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_37_cast_fp16 = mul(x = var_1984_cast_fp16, y = _inversed_scores_37_y_0_to_fp16)[name = tensor("_inversed_scores_37_cast_fp16")]; - tensor scores_39_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_37_cast_fp16, cond = mask_11)[name = tensor("scores_39_cast_fp16")]; - tensor var_1990_cast_fp16 = softmax(axis = var_23, x = scores_39_cast_fp16)[name = tensor("op_1990_cast_fp16")]; - tensor input_501_cast_fp16 = select(a = var_11_to_fp16, b = var_1990_cast_fp16, cond = mask_11)[name = tensor("input_501_cast_fp16")]; - tensor x_211_transpose_x_0 = const()[name = tensor("x_211_transpose_x_0"), val = tensor(false)]; - tensor x_211_transpose_y_0 = const()[name = tensor("x_211_transpose_y_0"), val = tensor(false)]; - tensor value_21_cast_fp16 = transpose(perm = value_21_perm_0, x = v_19_cast_fp16)[name = tensor("transpose_141")]; - tensor x_211_cast_fp16 = matmul(transpose_x = x_211_transpose_x_0, transpose_y = x_211_transpose_y_0, x = input_501_cast_fp16, y = value_21_cast_fp16)[name = tensor("x_211_cast_fp16")]; - tensor var_1994_perm_0 = const()[name = tensor("op_1994_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_1995 = const()[name = tensor("op_1995"), val = tensor([1, -1, 512])]; - tensor var_1994_cast_fp16 = transpose(perm = var_1994_perm_0, x = x_211_cast_fp16)[name = tensor("transpose_137")]; - tensor input_503_cast_fp16 = reshape(shape = var_1995, x = var_1994_cast_fp16)[name = tensor("input_503_cast_fp16")]; - tensor module_layers_9_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_9_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121431616)))]; - tensor module_layers_9_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_9_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121955968)))]; - tensor linear_88_cast_fp16 = linear(bias = module_layers_9_self_attn_linear_out_bias_to_fp16, weight = module_layers_9_self_attn_linear_out_weight_to_fp16, x = input_503_cast_fp16)[name = tensor("linear_88_cast_fp16")]; - tensor input_507_cast_fp16 = add(x = input_499_cast_fp16, y = linear_88_cast_fp16)[name = tensor("input_507_cast_fp16")]; - tensor x_215_axes_0 = const()[name = tensor("x_215_axes_0"), val = tensor([-1])]; - tensor module_layers_9_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_9_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121957056)))]; - tensor module_layers_9_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_9_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121958144)))]; - tensor x_215_cast_fp16 = layer_norm(axes = x_215_axes_0, beta = module_layers_9_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_9_norm_conv_weight_to_fp16, x = input_507_cast_fp16)[name = tensor("x_215_cast_fp16")]; - tensor input_509_perm_0 = const()[name = tensor("input_509_perm_0"), val = tensor([0, 2, 1])]; - tensor input_511_pad_type_0 = const()[name = tensor("input_511_pad_type_0"), val = tensor("valid")]; - tensor input_511_strides_0 = const()[name = tensor("input_511_strides_0"), val = tensor([1])]; - tensor input_511_pad_0 = const()[name = tensor("input_511_pad_0"), val = tensor([0, 0])]; - tensor input_511_dilations_0 = const()[name = tensor("input_511_dilations_0"), val = tensor([1])]; - tensor input_511_groups_0 = const()[name = tensor("input_511_groups_0"), val = tensor(1)]; - tensor module_layers_9_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_9_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121959232)))]; - tensor module_layers_9_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_9_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123007872)))]; - tensor input_509_cast_fp16 = transpose(perm = input_509_perm_0, x = x_215_cast_fp16)[name = tensor("transpose_136")]; - tensor input_511_cast_fp16 = conv(bias = module_layers_9_conv_pointwise_conv1_bias_to_fp16, dilations = input_511_dilations_0, groups = input_511_groups_0, pad = input_511_pad_0, pad_type = input_511_pad_type_0, strides = input_511_strides_0, weight = module_layers_9_conv_pointwise_conv1_weight_to_fp16, x = input_509_cast_fp16)[name = tensor("input_511_cast_fp16")]; - tensor x_217_split_num_splits_0 = const()[name = tensor("x_217_split_num_splits_0"), val = tensor(2)]; - tensor x_217_split_axis_0 = const()[name = tensor("x_217_split_axis_0"), val = tensor(1)]; - tensor x_217_split_cast_fp16_0, tensor x_217_split_cast_fp16_1 = split(axis = x_217_split_axis_0, num_splits = x_217_split_num_splits_0, x = input_511_cast_fp16)[name = tensor("x_217_split_cast_fp16")]; - tensor x_217_split_1_sigmoid_cast_fp16 = sigmoid(x = x_217_split_cast_fp16_1)[name = tensor("x_217_split_1_sigmoid_cast_fp16")]; - tensor x_217_cast_fp16 = mul(x = x_217_split_cast_fp16_0, y = x_217_split_1_sigmoid_cast_fp16)[name = tensor("x_217_cast_fp16")]; - tensor input_513_cast_fp16 = select(a = var_11_to_fp16, b = x_217_cast_fp16, cond = var_453)[name = tensor("input_513_cast_fp16")]; - tensor input_515_pad_0 = const()[name = tensor("input_515_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_515_mode_0 = const()[name = tensor("input_515_mode_0"), val = tensor("constant")]; - tensor const_163_to_fp16 = const()[name = tensor("const_163_to_fp16"), val = tensor(0x0p+0)]; - tensor input_515_cast_fp16 = pad(constant_val = const_163_to_fp16, mode = input_515_mode_0, pad = input_515_pad_0, x = input_513_cast_fp16)[name = tensor("input_515_cast_fp16")]; - tensor input_517_pad_type_0 = const()[name = tensor("input_517_pad_type_0"), val = tensor("valid")]; - tensor input_517_groups_0 = const()[name = tensor("input_517_groups_0"), val = tensor(512)]; - tensor input_517_strides_0 = const()[name = tensor("input_517_strides_0"), val = tensor([1])]; - tensor input_517_pad_0 = const()[name = tensor("input_517_pad_0"), val = tensor([0, 0])]; - tensor input_517_dilations_0 = const()[name = tensor("input_517_dilations_0"), val = tensor([1])]; - tensor const_252_to_fp16 = const()[name = tensor("const_252_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123009984)))]; - tensor const_253_to_fp16 = const()[name = tensor("const_253_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123019264)))]; - tensor input_519_cast_fp16 = conv(bias = const_253_to_fp16, dilations = input_517_dilations_0, groups = input_517_groups_0, pad = input_517_pad_0, pad_type = input_517_pad_type_0, strides = input_517_strides_0, weight = const_252_to_fp16, x = input_515_cast_fp16)[name = tensor("input_519_cast_fp16")]; - tensor input_521_cast_fp16 = silu(x = input_519_cast_fp16)[name = tensor("input_521_cast_fp16")]; - tensor x_219_pad_type_0 = const()[name = tensor("x_219_pad_type_0"), val = tensor("valid")]; - tensor x_219_strides_0 = const()[name = tensor("x_219_strides_0"), val = tensor([1])]; - tensor x_219_pad_0 = const()[name = tensor("x_219_pad_0"), val = tensor([0, 0])]; - tensor x_219_dilations_0 = const()[name = tensor("x_219_dilations_0"), val = tensor([1])]; - tensor x_219_groups_0 = const()[name = tensor("x_219_groups_0"), val = tensor(1)]; - tensor module_layers_9_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_9_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123020352)))]; - tensor module_layers_9_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_9_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123544704)))]; - tensor x_219_cast_fp16 = conv(bias = module_layers_9_conv_pointwise_conv2_bias_to_fp16, dilations = x_219_dilations_0, groups = x_219_groups_0, pad = x_219_pad_0, pad_type = x_219_pad_type_0, strides = x_219_strides_0, weight = module_layers_9_conv_pointwise_conv2_weight_to_fp16, x = input_521_cast_fp16)[name = tensor("x_219_cast_fp16")]; - tensor input_523_perm_0 = const()[name = tensor("input_523_perm_0"), val = tensor([0, 2, 1])]; - tensor input_523_cast_fp16 = transpose(perm = input_523_perm_0, x = x_219_cast_fp16)[name = tensor("transpose_135")]; - tensor input_525_cast_fp16 = add(x = input_507_cast_fp16, y = input_523_cast_fp16)[name = tensor("input_525_cast_fp16")]; - tensor input_527_axes_0 = const()[name = tensor("input_527_axes_0"), val = tensor([-1])]; - tensor module_layers_9_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_9_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123545792)))]; - tensor module_layers_9_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_9_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123546880)))]; - tensor input_527_cast_fp16 = layer_norm(axes = input_527_axes_0, beta = module_layers_9_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_9_norm_feed_forward2_weight_to_fp16, x = input_525_cast_fp16)[name = tensor("input_527_cast_fp16")]; - tensor module_layers_9_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_9_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123547968)))]; - tensor module_layers_9_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_9_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(125645184)))]; - tensor linear_89_cast_fp16 = linear(bias = module_layers_9_feed_forward2_linear1_bias_to_fp16, weight = module_layers_9_feed_forward2_linear1_weight_to_fp16, x = input_527_cast_fp16)[name = tensor("linear_89_cast_fp16")]; - tensor input_531_cast_fp16 = silu(x = linear_89_cast_fp16)[name = tensor("input_531_cast_fp16")]; - tensor module_layers_9_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_9_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(125649344)))]; - tensor module_layers_9_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_9_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127746560)))]; - tensor linear_90_cast_fp16 = linear(bias = module_layers_9_feed_forward2_linear2_bias_to_fp16, weight = module_layers_9_feed_forward2_linear2_weight_to_fp16, x = input_531_cast_fp16)[name = tensor("linear_90_cast_fp16")]; - tensor var_2061_to_fp16 = const()[name = tensor("op_2061_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2062_cast_fp16 = mul(x = linear_90_cast_fp16, y = var_2061_to_fp16)[name = tensor("op_2062_cast_fp16")]; - tensor input_537_cast_fp16 = add(x = input_525_cast_fp16, y = var_2062_cast_fp16)[name = tensor("input_537_cast_fp16")]; - tensor input_539_axes_0 = const()[name = tensor("input_539_axes_0"), val = tensor([-1])]; - tensor module_layers_9_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_9_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127747648)))]; - tensor module_layers_9_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_9_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127748736)))]; - tensor input_539_cast_fp16 = layer_norm(axes = input_539_axes_0, beta = module_layers_9_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_9_norm_out_weight_to_fp16, x = input_537_cast_fp16)[name = tensor("input_539_cast_fp16")]; - tensor input_541_axes_0 = const()[name = tensor("input_541_axes_0"), val = tensor([-1])]; - tensor module_layers_10_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_10_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127749824)))]; - tensor module_layers_10_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_10_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127750912)))]; - tensor input_541_cast_fp16 = layer_norm(axes = input_541_axes_0, beta = module_layers_10_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_10_norm_feed_forward1_weight_to_fp16, x = input_539_cast_fp16)[name = tensor("input_541_cast_fp16")]; - tensor module_layers_10_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_10_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127752000)))]; - tensor module_layers_10_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_10_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(129849216)))]; - tensor linear_91_cast_fp16 = linear(bias = module_layers_10_feed_forward1_linear1_bias_to_fp16, weight = module_layers_10_feed_forward1_linear1_weight_to_fp16, x = input_541_cast_fp16)[name = tensor("linear_91_cast_fp16")]; - tensor input_545_cast_fp16 = silu(x = linear_91_cast_fp16)[name = tensor("input_545_cast_fp16")]; - tensor module_layers_10_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_10_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(129853376)))]; - tensor module_layers_10_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_10_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(131950592)))]; - tensor linear_92_cast_fp16 = linear(bias = module_layers_10_feed_forward1_linear2_bias_to_fp16, weight = module_layers_10_feed_forward1_linear2_weight_to_fp16, x = input_545_cast_fp16)[name = tensor("linear_92_cast_fp16")]; - tensor var_2092_to_fp16 = const()[name = tensor("op_2092_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2093_cast_fp16 = mul(x = linear_92_cast_fp16, y = var_2092_to_fp16)[name = tensor("op_2093_cast_fp16")]; - tensor input_551_cast_fp16 = add(x = input_539_cast_fp16, y = var_2093_cast_fp16)[name = tensor("input_551_cast_fp16")]; - tensor query_21_axes_0 = const()[name = tensor("query_21_axes_0"), val = tensor([-1])]; - tensor module_layers_10_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_10_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(131951680)))]; - tensor module_layers_10_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_10_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(131952768)))]; - tensor query_21_cast_fp16 = layer_norm(axes = query_21_axes_0, beta = module_layers_10_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_10_norm_self_att_weight_to_fp16, x = input_551_cast_fp16)[name = tensor("query_21_cast_fp16")]; - tensor module_layers_10_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_10_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(131953856)))]; - tensor module_layers_10_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_10_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132478208)))]; - tensor linear_93_cast_fp16 = linear(bias = module_layers_10_self_attn_linear_q_bias_to_fp16, weight = module_layers_10_self_attn_linear_q_weight_to_fp16, x = query_21_cast_fp16)[name = tensor("linear_93_cast_fp16")]; - tensor var_2110 = const()[name = tensor("op_2110"), val = tensor([1, -1, 8, 64])]; - tensor q_61_cast_fp16 = reshape(shape = var_2110, x = linear_93_cast_fp16)[name = tensor("q_61_cast_fp16")]; - tensor module_layers_10_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_10_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132479296)))]; - tensor module_layers_10_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_10_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133003648)))]; - tensor linear_94_cast_fp16 = linear(bias = module_layers_10_self_attn_linear_k_bias_to_fp16, weight = module_layers_10_self_attn_linear_k_weight_to_fp16, x = query_21_cast_fp16)[name = tensor("linear_94_cast_fp16")]; - tensor var_2115 = const()[name = tensor("op_2115"), val = tensor([1, -1, 8, 64])]; - tensor k_41_cast_fp16 = reshape(shape = var_2115, x = linear_94_cast_fp16)[name = tensor("k_41_cast_fp16")]; - tensor module_layers_10_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_10_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133004736)))]; - tensor module_layers_10_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_10_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133529088)))]; - tensor linear_95_cast_fp16 = linear(bias = module_layers_10_self_attn_linear_v_bias_to_fp16, weight = module_layers_10_self_attn_linear_v_weight_to_fp16, x = query_21_cast_fp16)[name = tensor("linear_95_cast_fp16")]; - tensor var_2120 = const()[name = tensor("op_2120"), val = tensor([1, -1, 8, 64])]; - tensor v_21_cast_fp16 = reshape(shape = var_2120, x = linear_95_cast_fp16)[name = tensor("v_21_cast_fp16")]; - tensor value_23_perm_0 = const()[name = tensor("value_23_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_10_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_10_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133530176)))]; - tensor var_2132_cast_fp16 = add(x = q_61_cast_fp16, y = module_layers_10_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2132_cast_fp16")]; - tensor module_layers_10_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_10_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133531264)))]; - tensor var_2134_cast_fp16 = add(x = q_61_cast_fp16, y = module_layers_10_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2134_cast_fp16")]; - tensor q_with_bias_v_21_perm_0 = const()[name = tensor("q_with_bias_v_21_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_227_transpose_x_0 = const()[name = tensor("x_227_transpose_x_0"), val = tensor(false)]; - tensor x_227_transpose_y_0 = const()[name = tensor("x_227_transpose_y_0"), val = tensor(false)]; - tensor var_2136_to_fp16 = const()[name = tensor("op_2136_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133532352)))]; - tensor q_with_bias_v_21_cast_fp16 = transpose(perm = q_with_bias_v_21_perm_0, x = var_2134_cast_fp16)[name = tensor("transpose_133")]; - tensor x_227_cast_fp16 = matmul(transpose_x = x_227_transpose_x_0, transpose_y = x_227_transpose_y_0, x = q_with_bias_v_21_cast_fp16, y = var_2136_to_fp16)[name = tensor("x_227_cast_fp16")]; - tensor x_229_pad_0 = const()[name = tensor("x_229_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_229_mode_0 = const()[name = tensor("x_229_mode_0"), val = tensor("constant")]; - tensor const_170_to_fp16 = const()[name = tensor("const_170_to_fp16"), val = tensor(0x0p+0)]; - tensor x_229_cast_fp16 = pad(constant_val = const_170_to_fp16, mode = x_229_mode_0, pad = x_229_pad_0, x = x_227_cast_fp16)[name = tensor("x_229_cast_fp16")]; - tensor var_2144 = const()[name = tensor("op_2144"), val = tensor([1, 8, -1, 188])]; - tensor x_231_cast_fp16 = reshape(shape = var_2144, x = x_229_cast_fp16)[name = tensor("x_231_cast_fp16")]; - tensor var_2148_begin_0 = const()[name = tensor("op_2148_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_2148_end_0 = const()[name = tensor("op_2148_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_2148_end_mask_0 = const()[name = tensor("op_2148_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_2148_cast_fp16 = slice_by_index(begin = var_2148_begin_0, end = var_2148_end_0, end_mask = var_2148_end_mask_0, x = x_231_cast_fp16)[name = tensor("op_2148_cast_fp16")]; - tensor var_2149 = const()[name = tensor("op_2149"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_41_cast_fp16 = reshape(shape = var_2149, x = var_2148_cast_fp16)[name = tensor("matrix_bd_41_cast_fp16")]; - tensor matrix_ac_21_transpose_x_0 = const()[name = tensor("matrix_ac_21_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_21_transpose_y_0 = const()[name = tensor("matrix_ac_21_transpose_y_0"), val = tensor(false)]; - tensor transpose_71_perm_0 = const()[name = tensor("transpose_71_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_72_perm_0 = const()[name = tensor("transpose_72_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_72 = transpose(perm = transpose_72_perm_0, x = k_41_cast_fp16)[name = tensor("transpose_131")]; - tensor transpose_71 = transpose(perm = transpose_71_perm_0, x = var_2132_cast_fp16)[name = tensor("transpose_132")]; - tensor matrix_ac_21_cast_fp16 = matmul(transpose_x = matrix_ac_21_transpose_x_0, transpose_y = matrix_ac_21_transpose_y_0, x = transpose_71, y = transpose_72)[name = tensor("matrix_ac_21_cast_fp16")]; - tensor matrix_bd_43_begin_0 = const()[name = tensor("matrix_bd_43_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_43_end_0 = const()[name = tensor("matrix_bd_43_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_43_end_mask_0 = const()[name = tensor("matrix_bd_43_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_43_cast_fp16 = slice_by_index(begin = matrix_bd_43_begin_0, end = matrix_bd_43_end_0, end_mask = matrix_bd_43_end_mask_0, x = matrix_bd_41_cast_fp16)[name = tensor("matrix_bd_43_cast_fp16")]; - tensor var_2158_cast_fp16 = add(x = matrix_ac_21_cast_fp16, y = matrix_bd_43_cast_fp16)[name = tensor("op_2158_cast_fp16")]; - tensor _inversed_scores_41_y_0_to_fp16 = const()[name = tensor("_inversed_scores_41_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_41_cast_fp16 = mul(x = var_2158_cast_fp16, y = _inversed_scores_41_y_0_to_fp16)[name = tensor("_inversed_scores_41_cast_fp16")]; - tensor scores_43_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_41_cast_fp16, cond = mask_11)[name = tensor("scores_43_cast_fp16")]; - tensor var_2164_cast_fp16 = softmax(axis = var_23, x = scores_43_cast_fp16)[name = tensor("op_2164_cast_fp16")]; - tensor input_553_cast_fp16 = select(a = var_11_to_fp16, b = var_2164_cast_fp16, cond = mask_11)[name = tensor("input_553_cast_fp16")]; - tensor x_233_transpose_x_0 = const()[name = tensor("x_233_transpose_x_0"), val = tensor(false)]; - tensor x_233_transpose_y_0 = const()[name = tensor("x_233_transpose_y_0"), val = tensor(false)]; - tensor value_23_cast_fp16 = transpose(perm = value_23_perm_0, x = v_21_cast_fp16)[name = tensor("transpose_134")]; - tensor x_233_cast_fp16 = matmul(transpose_x = x_233_transpose_x_0, transpose_y = x_233_transpose_y_0, x = input_553_cast_fp16, y = value_23_cast_fp16)[name = tensor("x_233_cast_fp16")]; - tensor var_2168_perm_0 = const()[name = tensor("op_2168_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_2169 = const()[name = tensor("op_2169"), val = tensor([1, -1, 512])]; - tensor var_2168_cast_fp16 = transpose(perm = var_2168_perm_0, x = x_233_cast_fp16)[name = tensor("transpose_130")]; - tensor input_555_cast_fp16 = reshape(shape = var_2169, x = var_2168_cast_fp16)[name = tensor("input_555_cast_fp16")]; - tensor module_layers_10_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_10_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133916416)))]; - tensor module_layers_10_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_10_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134440768)))]; - tensor linear_97_cast_fp16 = linear(bias = module_layers_10_self_attn_linear_out_bias_to_fp16, weight = module_layers_10_self_attn_linear_out_weight_to_fp16, x = input_555_cast_fp16)[name = tensor("linear_97_cast_fp16")]; - tensor input_559_cast_fp16 = add(x = input_551_cast_fp16, y = linear_97_cast_fp16)[name = tensor("input_559_cast_fp16")]; - tensor x_237_axes_0 = const()[name = tensor("x_237_axes_0"), val = tensor([-1])]; - tensor module_layers_10_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_10_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134441856)))]; - tensor module_layers_10_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_10_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134442944)))]; - tensor x_237_cast_fp16 = layer_norm(axes = x_237_axes_0, beta = module_layers_10_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_10_norm_conv_weight_to_fp16, x = input_559_cast_fp16)[name = tensor("x_237_cast_fp16")]; - tensor input_561_perm_0 = const()[name = tensor("input_561_perm_0"), val = tensor([0, 2, 1])]; - tensor input_563_pad_type_0 = const()[name = tensor("input_563_pad_type_0"), val = tensor("valid")]; - tensor input_563_strides_0 = const()[name = tensor("input_563_strides_0"), val = tensor([1])]; - tensor input_563_pad_0 = const()[name = tensor("input_563_pad_0"), val = tensor([0, 0])]; - tensor input_563_dilations_0 = const()[name = tensor("input_563_dilations_0"), val = tensor([1])]; - tensor input_563_groups_0 = const()[name = tensor("input_563_groups_0"), val = tensor(1)]; - tensor module_layers_10_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_10_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134444032)))]; - tensor module_layers_10_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_10_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(135492672)))]; - tensor input_561_cast_fp16 = transpose(perm = input_561_perm_0, x = x_237_cast_fp16)[name = tensor("transpose_129")]; - tensor input_563_cast_fp16 = conv(bias = module_layers_10_conv_pointwise_conv1_bias_to_fp16, dilations = input_563_dilations_0, groups = input_563_groups_0, pad = input_563_pad_0, pad_type = input_563_pad_type_0, strides = input_563_strides_0, weight = module_layers_10_conv_pointwise_conv1_weight_to_fp16, x = input_561_cast_fp16)[name = tensor("input_563_cast_fp16")]; - tensor x_239_split_num_splits_0 = const()[name = tensor("x_239_split_num_splits_0"), val = tensor(2)]; - tensor x_239_split_axis_0 = const()[name = tensor("x_239_split_axis_0"), val = tensor(1)]; - tensor x_239_split_cast_fp16_0, tensor x_239_split_cast_fp16_1 = split(axis = x_239_split_axis_0, num_splits = x_239_split_num_splits_0, x = input_563_cast_fp16)[name = tensor("x_239_split_cast_fp16")]; - tensor x_239_split_1_sigmoid_cast_fp16 = sigmoid(x = x_239_split_cast_fp16_1)[name = tensor("x_239_split_1_sigmoid_cast_fp16")]; - tensor x_239_cast_fp16 = mul(x = x_239_split_cast_fp16_0, y = x_239_split_1_sigmoid_cast_fp16)[name = tensor("x_239_cast_fp16")]; - tensor input_565_cast_fp16 = select(a = var_11_to_fp16, b = x_239_cast_fp16, cond = var_453)[name = tensor("input_565_cast_fp16")]; - tensor input_567_pad_0 = const()[name = tensor("input_567_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_567_mode_0 = const()[name = tensor("input_567_mode_0"), val = tensor("constant")]; - tensor const_173_to_fp16 = const()[name = tensor("const_173_to_fp16"), val = tensor(0x0p+0)]; - tensor input_567_cast_fp16 = pad(constant_val = const_173_to_fp16, mode = input_567_mode_0, pad = input_567_pad_0, x = input_565_cast_fp16)[name = tensor("input_567_cast_fp16")]; - tensor input_569_pad_type_0 = const()[name = tensor("input_569_pad_type_0"), val = tensor("valid")]; - tensor input_569_groups_0 = const()[name = tensor("input_569_groups_0"), val = tensor(512)]; - tensor input_569_strides_0 = const()[name = tensor("input_569_strides_0"), val = tensor([1])]; - tensor input_569_pad_0 = const()[name = tensor("input_569_pad_0"), val = tensor([0, 0])]; - tensor input_569_dilations_0 = const()[name = tensor("input_569_dilations_0"), val = tensor([1])]; - tensor const_254_to_fp16 = const()[name = tensor("const_254_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(135494784)))]; - tensor const_255_to_fp16 = const()[name = tensor("const_255_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(135504064)))]; - tensor input_571_cast_fp16 = conv(bias = const_255_to_fp16, dilations = input_569_dilations_0, groups = input_569_groups_0, pad = input_569_pad_0, pad_type = input_569_pad_type_0, strides = input_569_strides_0, weight = const_254_to_fp16, x = input_567_cast_fp16)[name = tensor("input_571_cast_fp16")]; - tensor input_573_cast_fp16 = silu(x = input_571_cast_fp16)[name = tensor("input_573_cast_fp16")]; - tensor x_241_pad_type_0 = const()[name = tensor("x_241_pad_type_0"), val = tensor("valid")]; - tensor x_241_strides_0 = const()[name = tensor("x_241_strides_0"), val = tensor([1])]; - tensor x_241_pad_0 = const()[name = tensor("x_241_pad_0"), val = tensor([0, 0])]; - tensor x_241_dilations_0 = const()[name = tensor("x_241_dilations_0"), val = tensor([1])]; - tensor x_241_groups_0 = const()[name = tensor("x_241_groups_0"), val = tensor(1)]; - tensor module_layers_10_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_10_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(135505152)))]; - tensor module_layers_10_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_10_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136029504)))]; - tensor x_241_cast_fp16 = conv(bias = module_layers_10_conv_pointwise_conv2_bias_to_fp16, dilations = x_241_dilations_0, groups = x_241_groups_0, pad = x_241_pad_0, pad_type = x_241_pad_type_0, strides = x_241_strides_0, weight = module_layers_10_conv_pointwise_conv2_weight_to_fp16, x = input_573_cast_fp16)[name = tensor("x_241_cast_fp16")]; - tensor input_575_perm_0 = const()[name = tensor("input_575_perm_0"), val = tensor([0, 2, 1])]; - tensor input_575_cast_fp16 = transpose(perm = input_575_perm_0, x = x_241_cast_fp16)[name = tensor("transpose_128")]; - tensor input_577_cast_fp16 = add(x = input_559_cast_fp16, y = input_575_cast_fp16)[name = tensor("input_577_cast_fp16")]; - tensor input_579_axes_0 = const()[name = tensor("input_579_axes_0"), val = tensor([-1])]; - tensor module_layers_10_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_10_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136030592)))]; - tensor module_layers_10_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_10_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136031680)))]; - tensor input_579_cast_fp16 = layer_norm(axes = input_579_axes_0, beta = module_layers_10_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_10_norm_feed_forward2_weight_to_fp16, x = input_577_cast_fp16)[name = tensor("input_579_cast_fp16")]; - tensor module_layers_10_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_10_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136032768)))]; - tensor module_layers_10_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_10_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(138129984)))]; - tensor linear_98_cast_fp16 = linear(bias = module_layers_10_feed_forward2_linear1_bias_to_fp16, weight = module_layers_10_feed_forward2_linear1_weight_to_fp16, x = input_579_cast_fp16)[name = tensor("linear_98_cast_fp16")]; - tensor input_583_cast_fp16 = silu(x = linear_98_cast_fp16)[name = tensor("input_583_cast_fp16")]; - tensor module_layers_10_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_10_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(138134144)))]; - tensor module_layers_10_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_10_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140231360)))]; - tensor linear_99_cast_fp16 = linear(bias = module_layers_10_feed_forward2_linear2_bias_to_fp16, weight = module_layers_10_feed_forward2_linear2_weight_to_fp16, x = input_583_cast_fp16)[name = tensor("linear_99_cast_fp16")]; - tensor var_2235_to_fp16 = const()[name = tensor("op_2235_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2236_cast_fp16 = mul(x = linear_99_cast_fp16, y = var_2235_to_fp16)[name = tensor("op_2236_cast_fp16")]; - tensor input_589_cast_fp16 = add(x = input_577_cast_fp16, y = var_2236_cast_fp16)[name = tensor("input_589_cast_fp16")]; - tensor input_591_axes_0 = const()[name = tensor("input_591_axes_0"), val = tensor([-1])]; - tensor module_layers_10_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_10_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140232448)))]; - tensor module_layers_10_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_10_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140233536)))]; - tensor input_591_cast_fp16 = layer_norm(axes = input_591_axes_0, beta = module_layers_10_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_10_norm_out_weight_to_fp16, x = input_589_cast_fp16)[name = tensor("input_591_cast_fp16")]; - tensor input_593_axes_0 = const()[name = tensor("input_593_axes_0"), val = tensor([-1])]; - tensor module_layers_11_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_11_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140234624)))]; - tensor module_layers_11_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_11_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140235712)))]; - tensor input_593_cast_fp16 = layer_norm(axes = input_593_axes_0, beta = module_layers_11_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_11_norm_feed_forward1_weight_to_fp16, x = input_591_cast_fp16)[name = tensor("input_593_cast_fp16")]; - tensor module_layers_11_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_11_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140236800)))]; - tensor module_layers_11_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_11_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(142334016)))]; - tensor linear_100_cast_fp16 = linear(bias = module_layers_11_feed_forward1_linear1_bias_to_fp16, weight = module_layers_11_feed_forward1_linear1_weight_to_fp16, x = input_593_cast_fp16)[name = tensor("linear_100_cast_fp16")]; - tensor input_597_cast_fp16 = silu(x = linear_100_cast_fp16)[name = tensor("input_597_cast_fp16")]; - tensor module_layers_11_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_11_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(142338176)))]; - tensor module_layers_11_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_11_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(144435392)))]; - tensor linear_101_cast_fp16 = linear(bias = module_layers_11_feed_forward1_linear2_bias_to_fp16, weight = module_layers_11_feed_forward1_linear2_weight_to_fp16, x = input_597_cast_fp16)[name = tensor("linear_101_cast_fp16")]; - tensor var_2266_to_fp16 = const()[name = tensor("op_2266_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2267_cast_fp16 = mul(x = linear_101_cast_fp16, y = var_2266_to_fp16)[name = tensor("op_2267_cast_fp16")]; - tensor input_603_cast_fp16 = add(x = input_591_cast_fp16, y = var_2267_cast_fp16)[name = tensor("input_603_cast_fp16")]; - tensor query_23_axes_0 = const()[name = tensor("query_23_axes_0"), val = tensor([-1])]; - tensor module_layers_11_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_11_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(144436480)))]; - tensor module_layers_11_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_11_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(144437568)))]; - tensor query_23_cast_fp16 = layer_norm(axes = query_23_axes_0, beta = module_layers_11_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_11_norm_self_att_weight_to_fp16, x = input_603_cast_fp16)[name = tensor("query_23_cast_fp16")]; - tensor module_layers_11_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_11_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(144438656)))]; - tensor module_layers_11_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_11_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(144963008)))]; - tensor linear_102_cast_fp16 = linear(bias = module_layers_11_self_attn_linear_q_bias_to_fp16, weight = module_layers_11_self_attn_linear_q_weight_to_fp16, x = query_23_cast_fp16)[name = tensor("linear_102_cast_fp16")]; - tensor var_2284 = const()[name = tensor("op_2284"), val = tensor([1, -1, 8, 64])]; - tensor q_67_cast_fp16 = reshape(shape = var_2284, x = linear_102_cast_fp16)[name = tensor("q_67_cast_fp16")]; - tensor module_layers_11_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_11_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(144964096)))]; - tensor module_layers_11_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_11_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145488448)))]; - tensor linear_103_cast_fp16 = linear(bias = module_layers_11_self_attn_linear_k_bias_to_fp16, weight = module_layers_11_self_attn_linear_k_weight_to_fp16, x = query_23_cast_fp16)[name = tensor("linear_103_cast_fp16")]; - tensor var_2289 = const()[name = tensor("op_2289"), val = tensor([1, -1, 8, 64])]; - tensor k_45_cast_fp16 = reshape(shape = var_2289, x = linear_103_cast_fp16)[name = tensor("k_45_cast_fp16")]; - tensor module_layers_11_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_11_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145489536)))]; - tensor module_layers_11_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_11_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146013888)))]; - tensor linear_104_cast_fp16 = linear(bias = module_layers_11_self_attn_linear_v_bias_to_fp16, weight = module_layers_11_self_attn_linear_v_weight_to_fp16, x = query_23_cast_fp16)[name = tensor("linear_104_cast_fp16")]; - tensor var_2294 = const()[name = tensor("op_2294"), val = tensor([1, -1, 8, 64])]; - tensor v_23_cast_fp16 = reshape(shape = var_2294, x = linear_104_cast_fp16)[name = tensor("v_23_cast_fp16")]; - tensor value_25_perm_0 = const()[name = tensor("value_25_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_11_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_11_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146014976)))]; - tensor var_2306_cast_fp16 = add(x = q_67_cast_fp16, y = module_layers_11_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2306_cast_fp16")]; - tensor module_layers_11_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_11_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146016064)))]; - tensor var_2308_cast_fp16 = add(x = q_67_cast_fp16, y = module_layers_11_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2308_cast_fp16")]; - tensor q_with_bias_v_23_perm_0 = const()[name = tensor("q_with_bias_v_23_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_249_transpose_x_0 = const()[name = tensor("x_249_transpose_x_0"), val = tensor(false)]; - tensor x_249_transpose_y_0 = const()[name = tensor("x_249_transpose_y_0"), val = tensor(false)]; - tensor var_2310_to_fp16 = const()[name = tensor("op_2310_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146017152)))]; - tensor q_with_bias_v_23_cast_fp16 = transpose(perm = q_with_bias_v_23_perm_0, x = var_2308_cast_fp16)[name = tensor("transpose_126")]; - tensor x_249_cast_fp16 = matmul(transpose_x = x_249_transpose_x_0, transpose_y = x_249_transpose_y_0, x = q_with_bias_v_23_cast_fp16, y = var_2310_to_fp16)[name = tensor("x_249_cast_fp16")]; - tensor x_251_pad_0 = const()[name = tensor("x_251_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_251_mode_0 = const()[name = tensor("x_251_mode_0"), val = tensor("constant")]; - tensor const_180_to_fp16 = const()[name = tensor("const_180_to_fp16"), val = tensor(0x0p+0)]; - tensor x_251_cast_fp16 = pad(constant_val = const_180_to_fp16, mode = x_251_mode_0, pad = x_251_pad_0, x = x_249_cast_fp16)[name = tensor("x_251_cast_fp16")]; - tensor var_2318 = const()[name = tensor("op_2318"), val = tensor([1, 8, -1, 188])]; - tensor x_253_cast_fp16 = reshape(shape = var_2318, x = x_251_cast_fp16)[name = tensor("x_253_cast_fp16")]; - tensor var_2322_begin_0 = const()[name = tensor("op_2322_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_2322_end_0 = const()[name = tensor("op_2322_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_2322_end_mask_0 = const()[name = tensor("op_2322_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_2322_cast_fp16 = slice_by_index(begin = var_2322_begin_0, end = var_2322_end_0, end_mask = var_2322_end_mask_0, x = x_253_cast_fp16)[name = tensor("op_2322_cast_fp16")]; - tensor var_2323 = const()[name = tensor("op_2323"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_45_cast_fp16 = reshape(shape = var_2323, x = var_2322_cast_fp16)[name = tensor("matrix_bd_45_cast_fp16")]; - tensor matrix_ac_23_transpose_x_0 = const()[name = tensor("matrix_ac_23_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_23_transpose_y_0 = const()[name = tensor("matrix_ac_23_transpose_y_0"), val = tensor(false)]; - tensor transpose_73_perm_0 = const()[name = tensor("transpose_73_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_74_perm_0 = const()[name = tensor("transpose_74_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_74 = transpose(perm = transpose_74_perm_0, x = k_45_cast_fp16)[name = tensor("transpose_124")]; - tensor transpose_73 = transpose(perm = transpose_73_perm_0, x = var_2306_cast_fp16)[name = tensor("transpose_125")]; - tensor matrix_ac_23_cast_fp16 = matmul(transpose_x = matrix_ac_23_transpose_x_0, transpose_y = matrix_ac_23_transpose_y_0, x = transpose_73, y = transpose_74)[name = tensor("matrix_ac_23_cast_fp16")]; - tensor matrix_bd_47_begin_0 = const()[name = tensor("matrix_bd_47_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_47_end_0 = const()[name = tensor("matrix_bd_47_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_47_end_mask_0 = const()[name = tensor("matrix_bd_47_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_47_cast_fp16 = slice_by_index(begin = matrix_bd_47_begin_0, end = matrix_bd_47_end_0, end_mask = matrix_bd_47_end_mask_0, x = matrix_bd_45_cast_fp16)[name = tensor("matrix_bd_47_cast_fp16")]; - tensor var_2332_cast_fp16 = add(x = matrix_ac_23_cast_fp16, y = matrix_bd_47_cast_fp16)[name = tensor("op_2332_cast_fp16")]; - tensor _inversed_scores_45_y_0_to_fp16 = const()[name = tensor("_inversed_scores_45_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_45_cast_fp16 = mul(x = var_2332_cast_fp16, y = _inversed_scores_45_y_0_to_fp16)[name = tensor("_inversed_scores_45_cast_fp16")]; - tensor scores_47_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_45_cast_fp16, cond = mask_11)[name = tensor("scores_47_cast_fp16")]; - tensor var_2338_cast_fp16 = softmax(axis = var_23, x = scores_47_cast_fp16)[name = tensor("op_2338_cast_fp16")]; - tensor input_605_cast_fp16 = select(a = var_11_to_fp16, b = var_2338_cast_fp16, cond = mask_11)[name = tensor("input_605_cast_fp16")]; - tensor x_255_transpose_x_0 = const()[name = tensor("x_255_transpose_x_0"), val = tensor(false)]; - tensor x_255_transpose_y_0 = const()[name = tensor("x_255_transpose_y_0"), val = tensor(false)]; - tensor value_25_cast_fp16 = transpose(perm = value_25_perm_0, x = v_23_cast_fp16)[name = tensor("transpose_127")]; - tensor x_255_cast_fp16 = matmul(transpose_x = x_255_transpose_x_0, transpose_y = x_255_transpose_y_0, x = input_605_cast_fp16, y = value_25_cast_fp16)[name = tensor("x_255_cast_fp16")]; - tensor var_2342_perm_0 = const()[name = tensor("op_2342_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_2343 = const()[name = tensor("op_2343"), val = tensor([1, -1, 512])]; - tensor var_2342_cast_fp16 = transpose(perm = var_2342_perm_0, x = x_255_cast_fp16)[name = tensor("transpose_123")]; - tensor input_607_cast_fp16 = reshape(shape = var_2343, x = var_2342_cast_fp16)[name = tensor("input_607_cast_fp16")]; - tensor module_layers_11_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_11_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146401216)))]; - tensor module_layers_11_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_11_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146925568)))]; - tensor linear_106_cast_fp16 = linear(bias = module_layers_11_self_attn_linear_out_bias_to_fp16, weight = module_layers_11_self_attn_linear_out_weight_to_fp16, x = input_607_cast_fp16)[name = tensor("linear_106_cast_fp16")]; - tensor input_611_cast_fp16 = add(x = input_603_cast_fp16, y = linear_106_cast_fp16)[name = tensor("input_611_cast_fp16")]; - tensor x_259_axes_0 = const()[name = tensor("x_259_axes_0"), val = tensor([-1])]; - tensor module_layers_11_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_11_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146926656)))]; - tensor module_layers_11_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_11_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146927744)))]; - tensor x_259_cast_fp16 = layer_norm(axes = x_259_axes_0, beta = module_layers_11_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_11_norm_conv_weight_to_fp16, x = input_611_cast_fp16)[name = tensor("x_259_cast_fp16")]; - tensor input_613_perm_0 = const()[name = tensor("input_613_perm_0"), val = tensor([0, 2, 1])]; - tensor input_615_pad_type_0 = const()[name = tensor("input_615_pad_type_0"), val = tensor("valid")]; - tensor input_615_strides_0 = const()[name = tensor("input_615_strides_0"), val = tensor([1])]; - tensor input_615_pad_0 = const()[name = tensor("input_615_pad_0"), val = tensor([0, 0])]; - tensor input_615_dilations_0 = const()[name = tensor("input_615_dilations_0"), val = tensor([1])]; - tensor input_615_groups_0 = const()[name = tensor("input_615_groups_0"), val = tensor(1)]; - tensor module_layers_11_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_11_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146928832)))]; - tensor module_layers_11_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_11_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147977472)))]; - tensor input_613_cast_fp16 = transpose(perm = input_613_perm_0, x = x_259_cast_fp16)[name = tensor("transpose_122")]; - tensor input_615_cast_fp16 = conv(bias = module_layers_11_conv_pointwise_conv1_bias_to_fp16, dilations = input_615_dilations_0, groups = input_615_groups_0, pad = input_615_pad_0, pad_type = input_615_pad_type_0, strides = input_615_strides_0, weight = module_layers_11_conv_pointwise_conv1_weight_to_fp16, x = input_613_cast_fp16)[name = tensor("input_615_cast_fp16")]; - tensor x_261_split_num_splits_0 = const()[name = tensor("x_261_split_num_splits_0"), val = tensor(2)]; - tensor x_261_split_axis_0 = const()[name = tensor("x_261_split_axis_0"), val = tensor(1)]; - tensor x_261_split_cast_fp16_0, tensor x_261_split_cast_fp16_1 = split(axis = x_261_split_axis_0, num_splits = x_261_split_num_splits_0, x = input_615_cast_fp16)[name = tensor("x_261_split_cast_fp16")]; - tensor x_261_split_1_sigmoid_cast_fp16 = sigmoid(x = x_261_split_cast_fp16_1)[name = tensor("x_261_split_1_sigmoid_cast_fp16")]; - tensor x_261_cast_fp16 = mul(x = x_261_split_cast_fp16_0, y = x_261_split_1_sigmoid_cast_fp16)[name = tensor("x_261_cast_fp16")]; - tensor input_617_cast_fp16 = select(a = var_11_to_fp16, b = x_261_cast_fp16, cond = var_453)[name = tensor("input_617_cast_fp16")]; - tensor input_619_pad_0 = const()[name = tensor("input_619_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_619_mode_0 = const()[name = tensor("input_619_mode_0"), val = tensor("constant")]; - tensor const_183_to_fp16 = const()[name = tensor("const_183_to_fp16"), val = tensor(0x0p+0)]; - tensor input_619_cast_fp16 = pad(constant_val = const_183_to_fp16, mode = input_619_mode_0, pad = input_619_pad_0, x = input_617_cast_fp16)[name = tensor("input_619_cast_fp16")]; - tensor input_621_pad_type_0 = const()[name = tensor("input_621_pad_type_0"), val = tensor("valid")]; - tensor input_621_groups_0 = const()[name = tensor("input_621_groups_0"), val = tensor(512)]; - tensor input_621_strides_0 = const()[name = tensor("input_621_strides_0"), val = tensor([1])]; - tensor input_621_pad_0 = const()[name = tensor("input_621_pad_0"), val = tensor([0, 0])]; - tensor input_621_dilations_0 = const()[name = tensor("input_621_dilations_0"), val = tensor([1])]; - tensor const_256_to_fp16 = const()[name = tensor("const_256_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147979584)))]; - tensor const_257_to_fp16 = const()[name = tensor("const_257_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147988864)))]; - tensor input_623_cast_fp16 = conv(bias = const_257_to_fp16, dilations = input_621_dilations_0, groups = input_621_groups_0, pad = input_621_pad_0, pad_type = input_621_pad_type_0, strides = input_621_strides_0, weight = const_256_to_fp16, x = input_619_cast_fp16)[name = tensor("input_623_cast_fp16")]; - tensor input_625_cast_fp16 = silu(x = input_623_cast_fp16)[name = tensor("input_625_cast_fp16")]; - tensor x_263_pad_type_0 = const()[name = tensor("x_263_pad_type_0"), val = tensor("valid")]; - tensor x_263_strides_0 = const()[name = tensor("x_263_strides_0"), val = tensor([1])]; - tensor x_263_pad_0 = const()[name = tensor("x_263_pad_0"), val = tensor([0, 0])]; - tensor x_263_dilations_0 = const()[name = tensor("x_263_dilations_0"), val = tensor([1])]; - tensor x_263_groups_0 = const()[name = tensor("x_263_groups_0"), val = tensor(1)]; - tensor module_layers_11_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_11_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147989952)))]; - tensor module_layers_11_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_11_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(148514304)))]; - tensor x_263_cast_fp16 = conv(bias = module_layers_11_conv_pointwise_conv2_bias_to_fp16, dilations = x_263_dilations_0, groups = x_263_groups_0, pad = x_263_pad_0, pad_type = x_263_pad_type_0, strides = x_263_strides_0, weight = module_layers_11_conv_pointwise_conv2_weight_to_fp16, x = input_625_cast_fp16)[name = tensor("x_263_cast_fp16")]; - tensor input_627_perm_0 = const()[name = tensor("input_627_perm_0"), val = tensor([0, 2, 1])]; - tensor input_627_cast_fp16 = transpose(perm = input_627_perm_0, x = x_263_cast_fp16)[name = tensor("transpose_121")]; - tensor input_629_cast_fp16 = add(x = input_611_cast_fp16, y = input_627_cast_fp16)[name = tensor("input_629_cast_fp16")]; - tensor input_631_axes_0 = const()[name = tensor("input_631_axes_0"), val = tensor([-1])]; - tensor module_layers_11_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_11_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(148515392)))]; - tensor module_layers_11_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_11_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(148516480)))]; - tensor input_631_cast_fp16 = layer_norm(axes = input_631_axes_0, beta = module_layers_11_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_11_norm_feed_forward2_weight_to_fp16, x = input_629_cast_fp16)[name = tensor("input_631_cast_fp16")]; - tensor module_layers_11_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_11_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(148517568)))]; - tensor module_layers_11_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_11_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(150614784)))]; - tensor linear_107_cast_fp16 = linear(bias = module_layers_11_feed_forward2_linear1_bias_to_fp16, weight = module_layers_11_feed_forward2_linear1_weight_to_fp16, x = input_631_cast_fp16)[name = tensor("linear_107_cast_fp16")]; - tensor input_635_cast_fp16 = silu(x = linear_107_cast_fp16)[name = tensor("input_635_cast_fp16")]; - tensor module_layers_11_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_11_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(150618944)))]; - tensor module_layers_11_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_11_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(152716160)))]; - tensor linear_108_cast_fp16 = linear(bias = module_layers_11_feed_forward2_linear2_bias_to_fp16, weight = module_layers_11_feed_forward2_linear2_weight_to_fp16, x = input_635_cast_fp16)[name = tensor("linear_108_cast_fp16")]; - tensor var_2409_to_fp16 = const()[name = tensor("op_2409_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2410_cast_fp16 = mul(x = linear_108_cast_fp16, y = var_2409_to_fp16)[name = tensor("op_2410_cast_fp16")]; - tensor input_641_cast_fp16 = add(x = input_629_cast_fp16, y = var_2410_cast_fp16)[name = tensor("input_641_cast_fp16")]; - tensor input_643_axes_0 = const()[name = tensor("input_643_axes_0"), val = tensor([-1])]; - tensor module_layers_11_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_11_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(152717248)))]; - tensor module_layers_11_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_11_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(152718336)))]; - tensor input_643_cast_fp16 = layer_norm(axes = input_643_axes_0, beta = module_layers_11_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_11_norm_out_weight_to_fp16, x = input_641_cast_fp16)[name = tensor("input_643_cast_fp16")]; - tensor input_645_axes_0 = const()[name = tensor("input_645_axes_0"), val = tensor([-1])]; - tensor module_layers_12_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_12_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(152719424)))]; - tensor module_layers_12_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_12_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(152720512)))]; - tensor input_645_cast_fp16 = layer_norm(axes = input_645_axes_0, beta = module_layers_12_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_12_norm_feed_forward1_weight_to_fp16, x = input_643_cast_fp16)[name = tensor("input_645_cast_fp16")]; - tensor module_layers_12_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_12_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(152721600)))]; - tensor module_layers_12_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_12_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(154818816)))]; - tensor linear_109_cast_fp16 = linear(bias = module_layers_12_feed_forward1_linear1_bias_to_fp16, weight = module_layers_12_feed_forward1_linear1_weight_to_fp16, x = input_645_cast_fp16)[name = tensor("linear_109_cast_fp16")]; - tensor input_649_cast_fp16 = silu(x = linear_109_cast_fp16)[name = tensor("input_649_cast_fp16")]; - tensor module_layers_12_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_12_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(154822976)))]; - tensor module_layers_12_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_12_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(156920192)))]; - tensor linear_110_cast_fp16 = linear(bias = module_layers_12_feed_forward1_linear2_bias_to_fp16, weight = module_layers_12_feed_forward1_linear2_weight_to_fp16, x = input_649_cast_fp16)[name = tensor("linear_110_cast_fp16")]; - tensor var_2440_to_fp16 = const()[name = tensor("op_2440_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2441_cast_fp16 = mul(x = linear_110_cast_fp16, y = var_2440_to_fp16)[name = tensor("op_2441_cast_fp16")]; - tensor input_655_cast_fp16 = add(x = input_643_cast_fp16, y = var_2441_cast_fp16)[name = tensor("input_655_cast_fp16")]; - tensor query_25_axes_0 = const()[name = tensor("query_25_axes_0"), val = tensor([-1])]; - tensor module_layers_12_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_12_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(156921280)))]; - tensor module_layers_12_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_12_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(156922368)))]; - tensor query_25_cast_fp16 = layer_norm(axes = query_25_axes_0, beta = module_layers_12_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_12_norm_self_att_weight_to_fp16, x = input_655_cast_fp16)[name = tensor("query_25_cast_fp16")]; - tensor module_layers_12_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_12_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(156923456)))]; - tensor module_layers_12_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_12_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157447808)))]; - tensor linear_111_cast_fp16 = linear(bias = module_layers_12_self_attn_linear_q_bias_to_fp16, weight = module_layers_12_self_attn_linear_q_weight_to_fp16, x = query_25_cast_fp16)[name = tensor("linear_111_cast_fp16")]; - tensor var_2458 = const()[name = tensor("op_2458"), val = tensor([1, -1, 8, 64])]; - tensor q_73_cast_fp16 = reshape(shape = var_2458, x = linear_111_cast_fp16)[name = tensor("q_73_cast_fp16")]; - tensor module_layers_12_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_12_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157448896)))]; - tensor module_layers_12_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_12_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157973248)))]; - tensor linear_112_cast_fp16 = linear(bias = module_layers_12_self_attn_linear_k_bias_to_fp16, weight = module_layers_12_self_attn_linear_k_weight_to_fp16, x = query_25_cast_fp16)[name = tensor("linear_112_cast_fp16")]; - tensor var_2463 = const()[name = tensor("op_2463"), val = tensor([1, -1, 8, 64])]; - tensor k_49_cast_fp16 = reshape(shape = var_2463, x = linear_112_cast_fp16)[name = tensor("k_49_cast_fp16")]; - tensor module_layers_12_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_12_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157974336)))]; - tensor module_layers_12_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_12_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158498688)))]; - tensor linear_113_cast_fp16 = linear(bias = module_layers_12_self_attn_linear_v_bias_to_fp16, weight = module_layers_12_self_attn_linear_v_weight_to_fp16, x = query_25_cast_fp16)[name = tensor("linear_113_cast_fp16")]; - tensor var_2468 = const()[name = tensor("op_2468"), val = tensor([1, -1, 8, 64])]; - tensor v_25_cast_fp16 = reshape(shape = var_2468, x = linear_113_cast_fp16)[name = tensor("v_25_cast_fp16")]; - tensor value_27_perm_0 = const()[name = tensor("value_27_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_12_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_12_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158499776)))]; - tensor var_2480_cast_fp16 = add(x = q_73_cast_fp16, y = module_layers_12_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2480_cast_fp16")]; - tensor module_layers_12_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_12_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158500864)))]; - tensor var_2482_cast_fp16 = add(x = q_73_cast_fp16, y = module_layers_12_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2482_cast_fp16")]; - tensor q_with_bias_v_25_perm_0 = const()[name = tensor("q_with_bias_v_25_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_271_transpose_x_0 = const()[name = tensor("x_271_transpose_x_0"), val = tensor(false)]; - tensor x_271_transpose_y_0 = const()[name = tensor("x_271_transpose_y_0"), val = tensor(false)]; - tensor var_2484_to_fp16 = const()[name = tensor("op_2484_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158501952)))]; - tensor q_with_bias_v_25_cast_fp16 = transpose(perm = q_with_bias_v_25_perm_0, x = var_2482_cast_fp16)[name = tensor("transpose_119")]; - tensor x_271_cast_fp16 = matmul(transpose_x = x_271_transpose_x_0, transpose_y = x_271_transpose_y_0, x = q_with_bias_v_25_cast_fp16, y = var_2484_to_fp16)[name = tensor("x_271_cast_fp16")]; - tensor x_273_pad_0 = const()[name = tensor("x_273_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_273_mode_0 = const()[name = tensor("x_273_mode_0"), val = tensor("constant")]; - tensor const_190_to_fp16 = const()[name = tensor("const_190_to_fp16"), val = tensor(0x0p+0)]; - tensor x_273_cast_fp16 = pad(constant_val = const_190_to_fp16, mode = x_273_mode_0, pad = x_273_pad_0, x = x_271_cast_fp16)[name = tensor("x_273_cast_fp16")]; - tensor var_2492 = const()[name = tensor("op_2492"), val = tensor([1, 8, -1, 188])]; - tensor x_275_cast_fp16 = reshape(shape = var_2492, x = x_273_cast_fp16)[name = tensor("x_275_cast_fp16")]; - tensor var_2496_begin_0 = const()[name = tensor("op_2496_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_2496_end_0 = const()[name = tensor("op_2496_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_2496_end_mask_0 = const()[name = tensor("op_2496_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_2496_cast_fp16 = slice_by_index(begin = var_2496_begin_0, end = var_2496_end_0, end_mask = var_2496_end_mask_0, x = x_275_cast_fp16)[name = tensor("op_2496_cast_fp16")]; - tensor var_2497 = const()[name = tensor("op_2497"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_49_cast_fp16 = reshape(shape = var_2497, x = var_2496_cast_fp16)[name = tensor("matrix_bd_49_cast_fp16")]; - tensor matrix_ac_25_transpose_x_0 = const()[name = tensor("matrix_ac_25_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_25_transpose_y_0 = const()[name = tensor("matrix_ac_25_transpose_y_0"), val = tensor(false)]; - tensor transpose_75_perm_0 = const()[name = tensor("transpose_75_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_76_perm_0 = const()[name = tensor("transpose_76_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_76 = transpose(perm = transpose_76_perm_0, x = k_49_cast_fp16)[name = tensor("transpose_117")]; - tensor transpose_75 = transpose(perm = transpose_75_perm_0, x = var_2480_cast_fp16)[name = tensor("transpose_118")]; - tensor matrix_ac_25_cast_fp16 = matmul(transpose_x = matrix_ac_25_transpose_x_0, transpose_y = matrix_ac_25_transpose_y_0, x = transpose_75, y = transpose_76)[name = tensor("matrix_ac_25_cast_fp16")]; - tensor matrix_bd_51_begin_0 = const()[name = tensor("matrix_bd_51_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_51_end_0 = const()[name = tensor("matrix_bd_51_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_51_end_mask_0 = const()[name = tensor("matrix_bd_51_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_51_cast_fp16 = slice_by_index(begin = matrix_bd_51_begin_0, end = matrix_bd_51_end_0, end_mask = matrix_bd_51_end_mask_0, x = matrix_bd_49_cast_fp16)[name = tensor("matrix_bd_51_cast_fp16")]; - tensor var_2506_cast_fp16 = add(x = matrix_ac_25_cast_fp16, y = matrix_bd_51_cast_fp16)[name = tensor("op_2506_cast_fp16")]; - tensor _inversed_scores_49_y_0_to_fp16 = const()[name = tensor("_inversed_scores_49_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_49_cast_fp16 = mul(x = var_2506_cast_fp16, y = _inversed_scores_49_y_0_to_fp16)[name = tensor("_inversed_scores_49_cast_fp16")]; - tensor scores_51_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_49_cast_fp16, cond = mask_11)[name = tensor("scores_51_cast_fp16")]; - tensor var_2512_cast_fp16 = softmax(axis = var_23, x = scores_51_cast_fp16)[name = tensor("op_2512_cast_fp16")]; - tensor input_657_cast_fp16 = select(a = var_11_to_fp16, b = var_2512_cast_fp16, cond = mask_11)[name = tensor("input_657_cast_fp16")]; - tensor x_277_transpose_x_0 = const()[name = tensor("x_277_transpose_x_0"), val = tensor(false)]; - tensor x_277_transpose_y_0 = const()[name = tensor("x_277_transpose_y_0"), val = tensor(false)]; - tensor value_27_cast_fp16 = transpose(perm = value_27_perm_0, x = v_25_cast_fp16)[name = tensor("transpose_120")]; - tensor x_277_cast_fp16 = matmul(transpose_x = x_277_transpose_x_0, transpose_y = x_277_transpose_y_0, x = input_657_cast_fp16, y = value_27_cast_fp16)[name = tensor("x_277_cast_fp16")]; - tensor var_2516_perm_0 = const()[name = tensor("op_2516_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_2517 = const()[name = tensor("op_2517"), val = tensor([1, -1, 512])]; - tensor var_2516_cast_fp16 = transpose(perm = var_2516_perm_0, x = x_277_cast_fp16)[name = tensor("transpose_116")]; - tensor input_659_cast_fp16 = reshape(shape = var_2517, x = var_2516_cast_fp16)[name = tensor("input_659_cast_fp16")]; - tensor module_layers_12_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_12_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158886016)))]; - tensor module_layers_12_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_12_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159410368)))]; - tensor linear_115_cast_fp16 = linear(bias = module_layers_12_self_attn_linear_out_bias_to_fp16, weight = module_layers_12_self_attn_linear_out_weight_to_fp16, x = input_659_cast_fp16)[name = tensor("linear_115_cast_fp16")]; - tensor input_663_cast_fp16 = add(x = input_655_cast_fp16, y = linear_115_cast_fp16)[name = tensor("input_663_cast_fp16")]; - tensor x_281_axes_0 = const()[name = tensor("x_281_axes_0"), val = tensor([-1])]; - tensor module_layers_12_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_12_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159411456)))]; - tensor module_layers_12_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_12_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159412544)))]; - tensor x_281_cast_fp16 = layer_norm(axes = x_281_axes_0, beta = module_layers_12_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_12_norm_conv_weight_to_fp16, x = input_663_cast_fp16)[name = tensor("x_281_cast_fp16")]; - tensor input_665_perm_0 = const()[name = tensor("input_665_perm_0"), val = tensor([0, 2, 1])]; - tensor input_667_pad_type_0 = const()[name = tensor("input_667_pad_type_0"), val = tensor("valid")]; - tensor input_667_strides_0 = const()[name = tensor("input_667_strides_0"), val = tensor([1])]; - tensor input_667_pad_0 = const()[name = tensor("input_667_pad_0"), val = tensor([0, 0])]; - tensor input_667_dilations_0 = const()[name = tensor("input_667_dilations_0"), val = tensor([1])]; - tensor input_667_groups_0 = const()[name = tensor("input_667_groups_0"), val = tensor(1)]; - tensor module_layers_12_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_12_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159413632)))]; - tensor module_layers_12_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_12_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160462272)))]; - tensor input_665_cast_fp16 = transpose(perm = input_665_perm_0, x = x_281_cast_fp16)[name = tensor("transpose_115")]; - tensor input_667_cast_fp16 = conv(bias = module_layers_12_conv_pointwise_conv1_bias_to_fp16, dilations = input_667_dilations_0, groups = input_667_groups_0, pad = input_667_pad_0, pad_type = input_667_pad_type_0, strides = input_667_strides_0, weight = module_layers_12_conv_pointwise_conv1_weight_to_fp16, x = input_665_cast_fp16)[name = tensor("input_667_cast_fp16")]; - tensor x_283_split_num_splits_0 = const()[name = tensor("x_283_split_num_splits_0"), val = tensor(2)]; - tensor x_283_split_axis_0 = const()[name = tensor("x_283_split_axis_0"), val = tensor(1)]; - tensor x_283_split_cast_fp16_0, tensor x_283_split_cast_fp16_1 = split(axis = x_283_split_axis_0, num_splits = x_283_split_num_splits_0, x = input_667_cast_fp16)[name = tensor("x_283_split_cast_fp16")]; - tensor x_283_split_1_sigmoid_cast_fp16 = sigmoid(x = x_283_split_cast_fp16_1)[name = tensor("x_283_split_1_sigmoid_cast_fp16")]; - tensor x_283_cast_fp16 = mul(x = x_283_split_cast_fp16_0, y = x_283_split_1_sigmoid_cast_fp16)[name = tensor("x_283_cast_fp16")]; - tensor input_669_cast_fp16 = select(a = var_11_to_fp16, b = x_283_cast_fp16, cond = var_453)[name = tensor("input_669_cast_fp16")]; - tensor input_671_pad_0 = const()[name = tensor("input_671_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_671_mode_0 = const()[name = tensor("input_671_mode_0"), val = tensor("constant")]; - tensor const_193_to_fp16 = const()[name = tensor("const_193_to_fp16"), val = tensor(0x0p+0)]; - tensor input_671_cast_fp16 = pad(constant_val = const_193_to_fp16, mode = input_671_mode_0, pad = input_671_pad_0, x = input_669_cast_fp16)[name = tensor("input_671_cast_fp16")]; - tensor input_673_pad_type_0 = const()[name = tensor("input_673_pad_type_0"), val = tensor("valid")]; - tensor input_673_groups_0 = const()[name = tensor("input_673_groups_0"), val = tensor(512)]; - tensor input_673_strides_0 = const()[name = tensor("input_673_strides_0"), val = tensor([1])]; - tensor input_673_pad_0 = const()[name = tensor("input_673_pad_0"), val = tensor([0, 0])]; - tensor input_673_dilations_0 = const()[name = tensor("input_673_dilations_0"), val = tensor([1])]; - tensor const_258_to_fp16 = const()[name = tensor("const_258_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160464384)))]; - tensor const_259_to_fp16 = const()[name = tensor("const_259_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160473664)))]; - tensor input_675_cast_fp16 = conv(bias = const_259_to_fp16, dilations = input_673_dilations_0, groups = input_673_groups_0, pad = input_673_pad_0, pad_type = input_673_pad_type_0, strides = input_673_strides_0, weight = const_258_to_fp16, x = input_671_cast_fp16)[name = tensor("input_675_cast_fp16")]; - tensor input_677_cast_fp16 = silu(x = input_675_cast_fp16)[name = tensor("input_677_cast_fp16")]; - tensor x_285_pad_type_0 = const()[name = tensor("x_285_pad_type_0"), val = tensor("valid")]; - tensor x_285_strides_0 = const()[name = tensor("x_285_strides_0"), val = tensor([1])]; - tensor x_285_pad_0 = const()[name = tensor("x_285_pad_0"), val = tensor([0, 0])]; - tensor x_285_dilations_0 = const()[name = tensor("x_285_dilations_0"), val = tensor([1])]; - tensor x_285_groups_0 = const()[name = tensor("x_285_groups_0"), val = tensor(1)]; - tensor module_layers_12_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_12_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160474752)))]; - tensor module_layers_12_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_12_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160999104)))]; - tensor x_285_cast_fp16 = conv(bias = module_layers_12_conv_pointwise_conv2_bias_to_fp16, dilations = x_285_dilations_0, groups = x_285_groups_0, pad = x_285_pad_0, pad_type = x_285_pad_type_0, strides = x_285_strides_0, weight = module_layers_12_conv_pointwise_conv2_weight_to_fp16, x = input_677_cast_fp16)[name = tensor("x_285_cast_fp16")]; - tensor input_679_perm_0 = const()[name = tensor("input_679_perm_0"), val = tensor([0, 2, 1])]; - tensor input_679_cast_fp16 = transpose(perm = input_679_perm_0, x = x_285_cast_fp16)[name = tensor("transpose_114")]; - tensor input_681_cast_fp16 = add(x = input_663_cast_fp16, y = input_679_cast_fp16)[name = tensor("input_681_cast_fp16")]; - tensor input_683_axes_0 = const()[name = tensor("input_683_axes_0"), val = tensor([-1])]; - tensor module_layers_12_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_12_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161000192)))]; - tensor module_layers_12_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_12_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161001280)))]; - tensor input_683_cast_fp16 = layer_norm(axes = input_683_axes_0, beta = module_layers_12_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_12_norm_feed_forward2_weight_to_fp16, x = input_681_cast_fp16)[name = tensor("input_683_cast_fp16")]; - tensor module_layers_12_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_12_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161002368)))]; - tensor module_layers_12_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_12_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(163099584)))]; - tensor linear_116_cast_fp16 = linear(bias = module_layers_12_feed_forward2_linear1_bias_to_fp16, weight = module_layers_12_feed_forward2_linear1_weight_to_fp16, x = input_683_cast_fp16)[name = tensor("linear_116_cast_fp16")]; - tensor input_687_cast_fp16 = silu(x = linear_116_cast_fp16)[name = tensor("input_687_cast_fp16")]; - tensor module_layers_12_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_12_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(163103744)))]; - tensor module_layers_12_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_12_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165200960)))]; - tensor linear_117_cast_fp16 = linear(bias = module_layers_12_feed_forward2_linear2_bias_to_fp16, weight = module_layers_12_feed_forward2_linear2_weight_to_fp16, x = input_687_cast_fp16)[name = tensor("linear_117_cast_fp16")]; - tensor var_2583_to_fp16 = const()[name = tensor("op_2583_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2584_cast_fp16 = mul(x = linear_117_cast_fp16, y = var_2583_to_fp16)[name = tensor("op_2584_cast_fp16")]; - tensor input_693_cast_fp16 = add(x = input_681_cast_fp16, y = var_2584_cast_fp16)[name = tensor("input_693_cast_fp16")]; - tensor input_695_axes_0 = const()[name = tensor("input_695_axes_0"), val = tensor([-1])]; - tensor module_layers_12_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_12_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165202048)))]; - tensor module_layers_12_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_12_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165203136)))]; - tensor input_695_cast_fp16 = layer_norm(axes = input_695_axes_0, beta = module_layers_12_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_12_norm_out_weight_to_fp16, x = input_693_cast_fp16)[name = tensor("input_695_cast_fp16")]; - tensor input_697_axes_0 = const()[name = tensor("input_697_axes_0"), val = tensor([-1])]; - tensor module_layers_13_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_13_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165204224)))]; - tensor module_layers_13_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_13_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165205312)))]; - tensor input_697_cast_fp16 = layer_norm(axes = input_697_axes_0, beta = module_layers_13_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_13_norm_feed_forward1_weight_to_fp16, x = input_695_cast_fp16)[name = tensor("input_697_cast_fp16")]; - tensor module_layers_13_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_13_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165206400)))]; - tensor module_layers_13_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_13_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(167303616)))]; - tensor linear_118_cast_fp16 = linear(bias = module_layers_13_feed_forward1_linear1_bias_to_fp16, weight = module_layers_13_feed_forward1_linear1_weight_to_fp16, x = input_697_cast_fp16)[name = tensor("linear_118_cast_fp16")]; - tensor input_701_cast_fp16 = silu(x = linear_118_cast_fp16)[name = tensor("input_701_cast_fp16")]; - tensor module_layers_13_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_13_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(167307776)))]; - tensor module_layers_13_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_13_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169404992)))]; - tensor linear_119_cast_fp16 = linear(bias = module_layers_13_feed_forward1_linear2_bias_to_fp16, weight = module_layers_13_feed_forward1_linear2_weight_to_fp16, x = input_701_cast_fp16)[name = tensor("linear_119_cast_fp16")]; - tensor var_2614_to_fp16 = const()[name = tensor("op_2614_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2615_cast_fp16 = mul(x = linear_119_cast_fp16, y = var_2614_to_fp16)[name = tensor("op_2615_cast_fp16")]; - tensor input_707_cast_fp16 = add(x = input_695_cast_fp16, y = var_2615_cast_fp16)[name = tensor("input_707_cast_fp16")]; - tensor query_27_axes_0 = const()[name = tensor("query_27_axes_0"), val = tensor([-1])]; - tensor module_layers_13_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_13_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169406080)))]; - tensor module_layers_13_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_13_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169407168)))]; - tensor query_27_cast_fp16 = layer_norm(axes = query_27_axes_0, beta = module_layers_13_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_13_norm_self_att_weight_to_fp16, x = input_707_cast_fp16)[name = tensor("query_27_cast_fp16")]; - tensor module_layers_13_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_13_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169408256)))]; - tensor module_layers_13_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_13_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169932608)))]; - tensor linear_120_cast_fp16 = linear(bias = module_layers_13_self_attn_linear_q_bias_to_fp16, weight = module_layers_13_self_attn_linear_q_weight_to_fp16, x = query_27_cast_fp16)[name = tensor("linear_120_cast_fp16")]; - tensor var_2632 = const()[name = tensor("op_2632"), val = tensor([1, -1, 8, 64])]; - tensor q_79_cast_fp16 = reshape(shape = var_2632, x = linear_120_cast_fp16)[name = tensor("q_79_cast_fp16")]; - tensor module_layers_13_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_13_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169933696)))]; - tensor module_layers_13_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_13_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170458048)))]; - tensor linear_121_cast_fp16 = linear(bias = module_layers_13_self_attn_linear_k_bias_to_fp16, weight = module_layers_13_self_attn_linear_k_weight_to_fp16, x = query_27_cast_fp16)[name = tensor("linear_121_cast_fp16")]; - tensor var_2637 = const()[name = tensor("op_2637"), val = tensor([1, -1, 8, 64])]; - tensor k_53_cast_fp16 = reshape(shape = var_2637, x = linear_121_cast_fp16)[name = tensor("k_53_cast_fp16")]; - tensor module_layers_13_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_13_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170459136)))]; - tensor module_layers_13_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_13_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170983488)))]; - tensor linear_122_cast_fp16 = linear(bias = module_layers_13_self_attn_linear_v_bias_to_fp16, weight = module_layers_13_self_attn_linear_v_weight_to_fp16, x = query_27_cast_fp16)[name = tensor("linear_122_cast_fp16")]; - tensor var_2642 = const()[name = tensor("op_2642"), val = tensor([1, -1, 8, 64])]; - tensor v_27_cast_fp16 = reshape(shape = var_2642, x = linear_122_cast_fp16)[name = tensor("v_27_cast_fp16")]; - tensor value_29_perm_0 = const()[name = tensor("value_29_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_13_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_13_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170984576)))]; - tensor var_2654_cast_fp16 = add(x = q_79_cast_fp16, y = module_layers_13_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2654_cast_fp16")]; - tensor module_layers_13_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_13_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170985664)))]; - tensor var_2656_cast_fp16 = add(x = q_79_cast_fp16, y = module_layers_13_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2656_cast_fp16")]; - tensor q_with_bias_v_27_perm_0 = const()[name = tensor("q_with_bias_v_27_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_293_transpose_x_0 = const()[name = tensor("x_293_transpose_x_0"), val = tensor(false)]; - tensor x_293_transpose_y_0 = const()[name = tensor("x_293_transpose_y_0"), val = tensor(false)]; - tensor var_2658_to_fp16 = const()[name = tensor("op_2658_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170986752)))]; - tensor q_with_bias_v_27_cast_fp16 = transpose(perm = q_with_bias_v_27_perm_0, x = var_2656_cast_fp16)[name = tensor("transpose_112")]; - tensor x_293_cast_fp16 = matmul(transpose_x = x_293_transpose_x_0, transpose_y = x_293_transpose_y_0, x = q_with_bias_v_27_cast_fp16, y = var_2658_to_fp16)[name = tensor("x_293_cast_fp16")]; - tensor x_295_pad_0 = const()[name = tensor("x_295_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_295_mode_0 = const()[name = tensor("x_295_mode_0"), val = tensor("constant")]; - tensor const_200_to_fp16 = const()[name = tensor("const_200_to_fp16"), val = tensor(0x0p+0)]; - tensor x_295_cast_fp16 = pad(constant_val = const_200_to_fp16, mode = x_295_mode_0, pad = x_295_pad_0, x = x_293_cast_fp16)[name = tensor("x_295_cast_fp16")]; - tensor var_2666 = const()[name = tensor("op_2666"), val = tensor([1, 8, -1, 188])]; - tensor x_297_cast_fp16 = reshape(shape = var_2666, x = x_295_cast_fp16)[name = tensor("x_297_cast_fp16")]; - tensor var_2670_begin_0 = const()[name = tensor("op_2670_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_2670_end_0 = const()[name = tensor("op_2670_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_2670_end_mask_0 = const()[name = tensor("op_2670_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_2670_cast_fp16 = slice_by_index(begin = var_2670_begin_0, end = var_2670_end_0, end_mask = var_2670_end_mask_0, x = x_297_cast_fp16)[name = tensor("op_2670_cast_fp16")]; - tensor var_2671 = const()[name = tensor("op_2671"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_53_cast_fp16 = reshape(shape = var_2671, x = var_2670_cast_fp16)[name = tensor("matrix_bd_53_cast_fp16")]; - tensor matrix_ac_27_transpose_x_0 = const()[name = tensor("matrix_ac_27_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_27_transpose_y_0 = const()[name = tensor("matrix_ac_27_transpose_y_0"), val = tensor(false)]; - tensor transpose_77_perm_0 = const()[name = tensor("transpose_77_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_78_perm_0 = const()[name = tensor("transpose_78_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_78 = transpose(perm = transpose_78_perm_0, x = k_53_cast_fp16)[name = tensor("transpose_110")]; - tensor transpose_77 = transpose(perm = transpose_77_perm_0, x = var_2654_cast_fp16)[name = tensor("transpose_111")]; - tensor matrix_ac_27_cast_fp16 = matmul(transpose_x = matrix_ac_27_transpose_x_0, transpose_y = matrix_ac_27_transpose_y_0, x = transpose_77, y = transpose_78)[name = tensor("matrix_ac_27_cast_fp16")]; - tensor matrix_bd_55_begin_0 = const()[name = tensor("matrix_bd_55_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_55_end_0 = const()[name = tensor("matrix_bd_55_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_55_end_mask_0 = const()[name = tensor("matrix_bd_55_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_55_cast_fp16 = slice_by_index(begin = matrix_bd_55_begin_0, end = matrix_bd_55_end_0, end_mask = matrix_bd_55_end_mask_0, x = matrix_bd_53_cast_fp16)[name = tensor("matrix_bd_55_cast_fp16")]; - tensor var_2680_cast_fp16 = add(x = matrix_ac_27_cast_fp16, y = matrix_bd_55_cast_fp16)[name = tensor("op_2680_cast_fp16")]; - tensor _inversed_scores_53_y_0_to_fp16 = const()[name = tensor("_inversed_scores_53_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_53_cast_fp16 = mul(x = var_2680_cast_fp16, y = _inversed_scores_53_y_0_to_fp16)[name = tensor("_inversed_scores_53_cast_fp16")]; - tensor scores_55_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_53_cast_fp16, cond = mask_11)[name = tensor("scores_55_cast_fp16")]; - tensor var_2686_cast_fp16 = softmax(axis = var_23, x = scores_55_cast_fp16)[name = tensor("op_2686_cast_fp16")]; - tensor input_709_cast_fp16 = select(a = var_11_to_fp16, b = var_2686_cast_fp16, cond = mask_11)[name = tensor("input_709_cast_fp16")]; - tensor x_299_transpose_x_0 = const()[name = tensor("x_299_transpose_x_0"), val = tensor(false)]; - tensor x_299_transpose_y_0 = const()[name = tensor("x_299_transpose_y_0"), val = tensor(false)]; - tensor value_29_cast_fp16 = transpose(perm = value_29_perm_0, x = v_27_cast_fp16)[name = tensor("transpose_113")]; - tensor x_299_cast_fp16 = matmul(transpose_x = x_299_transpose_x_0, transpose_y = x_299_transpose_y_0, x = input_709_cast_fp16, y = value_29_cast_fp16)[name = tensor("x_299_cast_fp16")]; - tensor var_2690_perm_0 = const()[name = tensor("op_2690_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_2691 = const()[name = tensor("op_2691"), val = tensor([1, -1, 512])]; - tensor var_2690_cast_fp16 = transpose(perm = var_2690_perm_0, x = x_299_cast_fp16)[name = tensor("transpose_109")]; - tensor input_711_cast_fp16 = reshape(shape = var_2691, x = var_2690_cast_fp16)[name = tensor("input_711_cast_fp16")]; - tensor module_layers_13_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_13_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171370816)))]; - tensor module_layers_13_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_13_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171895168)))]; - tensor linear_124_cast_fp16 = linear(bias = module_layers_13_self_attn_linear_out_bias_to_fp16, weight = module_layers_13_self_attn_linear_out_weight_to_fp16, x = input_711_cast_fp16)[name = tensor("linear_124_cast_fp16")]; - tensor input_715_cast_fp16 = add(x = input_707_cast_fp16, y = linear_124_cast_fp16)[name = tensor("input_715_cast_fp16")]; - tensor x_303_axes_0 = const()[name = tensor("x_303_axes_0"), val = tensor([-1])]; - tensor module_layers_13_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_13_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171896256)))]; - tensor module_layers_13_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_13_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171897344)))]; - tensor x_303_cast_fp16 = layer_norm(axes = x_303_axes_0, beta = module_layers_13_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_13_norm_conv_weight_to_fp16, x = input_715_cast_fp16)[name = tensor("x_303_cast_fp16")]; - tensor input_717_perm_0 = const()[name = tensor("input_717_perm_0"), val = tensor([0, 2, 1])]; - tensor input_719_pad_type_0 = const()[name = tensor("input_719_pad_type_0"), val = tensor("valid")]; - tensor input_719_strides_0 = const()[name = tensor("input_719_strides_0"), val = tensor([1])]; - tensor input_719_pad_0 = const()[name = tensor("input_719_pad_0"), val = tensor([0, 0])]; - tensor input_719_dilations_0 = const()[name = tensor("input_719_dilations_0"), val = tensor([1])]; - tensor input_719_groups_0 = const()[name = tensor("input_719_groups_0"), val = tensor(1)]; - tensor module_layers_13_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_13_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171898432)))]; - tensor module_layers_13_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_13_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(172947072)))]; - tensor input_717_cast_fp16 = transpose(perm = input_717_perm_0, x = x_303_cast_fp16)[name = tensor("transpose_108")]; - tensor input_719_cast_fp16 = conv(bias = module_layers_13_conv_pointwise_conv1_bias_to_fp16, dilations = input_719_dilations_0, groups = input_719_groups_0, pad = input_719_pad_0, pad_type = input_719_pad_type_0, strides = input_719_strides_0, weight = module_layers_13_conv_pointwise_conv1_weight_to_fp16, x = input_717_cast_fp16)[name = tensor("input_719_cast_fp16")]; - tensor x_305_split_num_splits_0 = const()[name = tensor("x_305_split_num_splits_0"), val = tensor(2)]; - tensor x_305_split_axis_0 = const()[name = tensor("x_305_split_axis_0"), val = tensor(1)]; - tensor x_305_split_cast_fp16_0, tensor x_305_split_cast_fp16_1 = split(axis = x_305_split_axis_0, num_splits = x_305_split_num_splits_0, x = input_719_cast_fp16)[name = tensor("x_305_split_cast_fp16")]; - tensor x_305_split_1_sigmoid_cast_fp16 = sigmoid(x = x_305_split_cast_fp16_1)[name = tensor("x_305_split_1_sigmoid_cast_fp16")]; - tensor x_305_cast_fp16 = mul(x = x_305_split_cast_fp16_0, y = x_305_split_1_sigmoid_cast_fp16)[name = tensor("x_305_cast_fp16")]; - tensor input_721_cast_fp16 = select(a = var_11_to_fp16, b = x_305_cast_fp16, cond = var_453)[name = tensor("input_721_cast_fp16")]; - tensor input_723_pad_0 = const()[name = tensor("input_723_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_723_mode_0 = const()[name = tensor("input_723_mode_0"), val = tensor("constant")]; - tensor const_203_to_fp16 = const()[name = tensor("const_203_to_fp16"), val = tensor(0x0p+0)]; - tensor input_723_cast_fp16 = pad(constant_val = const_203_to_fp16, mode = input_723_mode_0, pad = input_723_pad_0, x = input_721_cast_fp16)[name = tensor("input_723_cast_fp16")]; - tensor input_725_pad_type_0 = const()[name = tensor("input_725_pad_type_0"), val = tensor("valid")]; - tensor input_725_groups_0 = const()[name = tensor("input_725_groups_0"), val = tensor(512)]; - tensor input_725_strides_0 = const()[name = tensor("input_725_strides_0"), val = tensor([1])]; - tensor input_725_pad_0 = const()[name = tensor("input_725_pad_0"), val = tensor([0, 0])]; - tensor input_725_dilations_0 = const()[name = tensor("input_725_dilations_0"), val = tensor([1])]; - tensor const_260_to_fp16 = const()[name = tensor("const_260_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(172949184)))]; - tensor const_261_to_fp16 = const()[name = tensor("const_261_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(172958464)))]; - tensor input_727_cast_fp16 = conv(bias = const_261_to_fp16, dilations = input_725_dilations_0, groups = input_725_groups_0, pad = input_725_pad_0, pad_type = input_725_pad_type_0, strides = input_725_strides_0, weight = const_260_to_fp16, x = input_723_cast_fp16)[name = tensor("input_727_cast_fp16")]; - tensor input_729_cast_fp16 = silu(x = input_727_cast_fp16)[name = tensor("input_729_cast_fp16")]; - tensor x_307_pad_type_0 = const()[name = tensor("x_307_pad_type_0"), val = tensor("valid")]; - tensor x_307_strides_0 = const()[name = tensor("x_307_strides_0"), val = tensor([1])]; - tensor x_307_pad_0 = const()[name = tensor("x_307_pad_0"), val = tensor([0, 0])]; - tensor x_307_dilations_0 = const()[name = tensor("x_307_dilations_0"), val = tensor([1])]; - tensor x_307_groups_0 = const()[name = tensor("x_307_groups_0"), val = tensor(1)]; - tensor module_layers_13_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_13_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(172959552)))]; - tensor module_layers_13_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_13_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(173483904)))]; - tensor x_307_cast_fp16 = conv(bias = module_layers_13_conv_pointwise_conv2_bias_to_fp16, dilations = x_307_dilations_0, groups = x_307_groups_0, pad = x_307_pad_0, pad_type = x_307_pad_type_0, strides = x_307_strides_0, weight = module_layers_13_conv_pointwise_conv2_weight_to_fp16, x = input_729_cast_fp16)[name = tensor("x_307_cast_fp16")]; - tensor input_731_perm_0 = const()[name = tensor("input_731_perm_0"), val = tensor([0, 2, 1])]; - tensor input_731_cast_fp16 = transpose(perm = input_731_perm_0, x = x_307_cast_fp16)[name = tensor("transpose_107")]; - tensor input_733_cast_fp16 = add(x = input_715_cast_fp16, y = input_731_cast_fp16)[name = tensor("input_733_cast_fp16")]; - tensor input_735_axes_0 = const()[name = tensor("input_735_axes_0"), val = tensor([-1])]; - tensor module_layers_13_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_13_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(173484992)))]; - tensor module_layers_13_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_13_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(173486080)))]; - tensor input_735_cast_fp16 = layer_norm(axes = input_735_axes_0, beta = module_layers_13_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_13_norm_feed_forward2_weight_to_fp16, x = input_733_cast_fp16)[name = tensor("input_735_cast_fp16")]; - tensor module_layers_13_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_13_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(173487168)))]; - tensor module_layers_13_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_13_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(175584384)))]; - tensor linear_125_cast_fp16 = linear(bias = module_layers_13_feed_forward2_linear1_bias_to_fp16, weight = module_layers_13_feed_forward2_linear1_weight_to_fp16, x = input_735_cast_fp16)[name = tensor("linear_125_cast_fp16")]; - tensor input_739_cast_fp16 = silu(x = linear_125_cast_fp16)[name = tensor("input_739_cast_fp16")]; - tensor module_layers_13_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_13_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(175588544)))]; - tensor module_layers_13_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_13_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(177685760)))]; - tensor linear_126_cast_fp16 = linear(bias = module_layers_13_feed_forward2_linear2_bias_to_fp16, weight = module_layers_13_feed_forward2_linear2_weight_to_fp16, x = input_739_cast_fp16)[name = tensor("linear_126_cast_fp16")]; - tensor var_2757_to_fp16 = const()[name = tensor("op_2757_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2758_cast_fp16 = mul(x = linear_126_cast_fp16, y = var_2757_to_fp16)[name = tensor("op_2758_cast_fp16")]; - tensor input_745_cast_fp16 = add(x = input_733_cast_fp16, y = var_2758_cast_fp16)[name = tensor("input_745_cast_fp16")]; - tensor input_747_axes_0 = const()[name = tensor("input_747_axes_0"), val = tensor([-1])]; - tensor module_layers_13_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_13_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(177686848)))]; - tensor module_layers_13_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_13_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(177687936)))]; - tensor input_747_cast_fp16 = layer_norm(axes = input_747_axes_0, beta = module_layers_13_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_13_norm_out_weight_to_fp16, x = input_745_cast_fp16)[name = tensor("input_747_cast_fp16")]; - tensor input_749_axes_0 = const()[name = tensor("input_749_axes_0"), val = tensor([-1])]; - tensor module_layers_14_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_14_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(177689024)))]; - tensor module_layers_14_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_14_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(177690112)))]; - tensor input_749_cast_fp16 = layer_norm(axes = input_749_axes_0, beta = module_layers_14_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_14_norm_feed_forward1_weight_to_fp16, x = input_747_cast_fp16)[name = tensor("input_749_cast_fp16")]; - tensor module_layers_14_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_14_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(177691200)))]; - tensor module_layers_14_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_14_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(179788416)))]; - tensor linear_127_cast_fp16 = linear(bias = module_layers_14_feed_forward1_linear1_bias_to_fp16, weight = module_layers_14_feed_forward1_linear1_weight_to_fp16, x = input_749_cast_fp16)[name = tensor("linear_127_cast_fp16")]; - tensor input_753_cast_fp16 = silu(x = linear_127_cast_fp16)[name = tensor("input_753_cast_fp16")]; - tensor module_layers_14_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_14_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(179792576)))]; - tensor module_layers_14_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_14_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(181889792)))]; - tensor linear_128_cast_fp16 = linear(bias = module_layers_14_feed_forward1_linear2_bias_to_fp16, weight = module_layers_14_feed_forward1_linear2_weight_to_fp16, x = input_753_cast_fp16)[name = tensor("linear_128_cast_fp16")]; - tensor var_2788_to_fp16 = const()[name = tensor("op_2788_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2789_cast_fp16 = mul(x = linear_128_cast_fp16, y = var_2788_to_fp16)[name = tensor("op_2789_cast_fp16")]; - tensor input_759_cast_fp16 = add(x = input_747_cast_fp16, y = var_2789_cast_fp16)[name = tensor("input_759_cast_fp16")]; - tensor query_29_axes_0 = const()[name = tensor("query_29_axes_0"), val = tensor([-1])]; - tensor module_layers_14_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_14_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(181890880)))]; - tensor module_layers_14_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_14_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(181891968)))]; - tensor query_29_cast_fp16 = layer_norm(axes = query_29_axes_0, beta = module_layers_14_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_14_norm_self_att_weight_to_fp16, x = input_759_cast_fp16)[name = tensor("query_29_cast_fp16")]; - tensor module_layers_14_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_14_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(181893056)))]; - tensor module_layers_14_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_14_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182417408)))]; - tensor linear_129_cast_fp16 = linear(bias = module_layers_14_self_attn_linear_q_bias_to_fp16, weight = module_layers_14_self_attn_linear_q_weight_to_fp16, x = query_29_cast_fp16)[name = tensor("linear_129_cast_fp16")]; - tensor var_2806 = const()[name = tensor("op_2806"), val = tensor([1, -1, 8, 64])]; - tensor q_85_cast_fp16 = reshape(shape = var_2806, x = linear_129_cast_fp16)[name = tensor("q_85_cast_fp16")]; - tensor module_layers_14_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_14_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182418496)))]; - tensor module_layers_14_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_14_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182942848)))]; - tensor linear_130_cast_fp16 = linear(bias = module_layers_14_self_attn_linear_k_bias_to_fp16, weight = module_layers_14_self_attn_linear_k_weight_to_fp16, x = query_29_cast_fp16)[name = tensor("linear_130_cast_fp16")]; - tensor var_2811 = const()[name = tensor("op_2811"), val = tensor([1, -1, 8, 64])]; - tensor k_57_cast_fp16 = reshape(shape = var_2811, x = linear_130_cast_fp16)[name = tensor("k_57_cast_fp16")]; - tensor module_layers_14_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_14_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182943936)))]; - tensor module_layers_14_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_14_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183468288)))]; - tensor linear_131_cast_fp16 = linear(bias = module_layers_14_self_attn_linear_v_bias_to_fp16, weight = module_layers_14_self_attn_linear_v_weight_to_fp16, x = query_29_cast_fp16)[name = tensor("linear_131_cast_fp16")]; - tensor var_2816 = const()[name = tensor("op_2816"), val = tensor([1, -1, 8, 64])]; - tensor v_29_cast_fp16 = reshape(shape = var_2816, x = linear_131_cast_fp16)[name = tensor("v_29_cast_fp16")]; - tensor value_31_perm_0 = const()[name = tensor("value_31_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_14_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_14_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183469376)))]; - tensor var_2828_cast_fp16 = add(x = q_85_cast_fp16, y = module_layers_14_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2828_cast_fp16")]; - tensor module_layers_14_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_14_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183470464)))]; - tensor var_2830_cast_fp16 = add(x = q_85_cast_fp16, y = module_layers_14_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2830_cast_fp16")]; - tensor q_with_bias_v_29_perm_0 = const()[name = tensor("q_with_bias_v_29_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_315_transpose_x_0 = const()[name = tensor("x_315_transpose_x_0"), val = tensor(false)]; - tensor x_315_transpose_y_0 = const()[name = tensor("x_315_transpose_y_0"), val = tensor(false)]; - tensor var_2832_to_fp16 = const()[name = tensor("op_2832_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183471552)))]; - tensor q_with_bias_v_29_cast_fp16 = transpose(perm = q_with_bias_v_29_perm_0, x = var_2830_cast_fp16)[name = tensor("transpose_105")]; - tensor x_315_cast_fp16 = matmul(transpose_x = x_315_transpose_x_0, transpose_y = x_315_transpose_y_0, x = q_with_bias_v_29_cast_fp16, y = var_2832_to_fp16)[name = tensor("x_315_cast_fp16")]; - tensor x_317_pad_0 = const()[name = tensor("x_317_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_317_mode_0 = const()[name = tensor("x_317_mode_0"), val = tensor("constant")]; - tensor const_210_to_fp16 = const()[name = tensor("const_210_to_fp16"), val = tensor(0x0p+0)]; - tensor x_317_cast_fp16 = pad(constant_val = const_210_to_fp16, mode = x_317_mode_0, pad = x_317_pad_0, x = x_315_cast_fp16)[name = tensor("x_317_cast_fp16")]; - tensor var_2840 = const()[name = tensor("op_2840"), val = tensor([1, 8, -1, 188])]; - tensor x_319_cast_fp16 = reshape(shape = var_2840, x = x_317_cast_fp16)[name = tensor("x_319_cast_fp16")]; - tensor var_2844_begin_0 = const()[name = tensor("op_2844_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_2844_end_0 = const()[name = tensor("op_2844_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_2844_end_mask_0 = const()[name = tensor("op_2844_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_2844_cast_fp16 = slice_by_index(begin = var_2844_begin_0, end = var_2844_end_0, end_mask = var_2844_end_mask_0, x = x_319_cast_fp16)[name = tensor("op_2844_cast_fp16")]; - tensor var_2845 = const()[name = tensor("op_2845"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_57_cast_fp16 = reshape(shape = var_2845, x = var_2844_cast_fp16)[name = tensor("matrix_bd_57_cast_fp16")]; - tensor matrix_ac_29_transpose_x_0 = const()[name = tensor("matrix_ac_29_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_29_transpose_y_0 = const()[name = tensor("matrix_ac_29_transpose_y_0"), val = tensor(false)]; - tensor transpose_79_perm_0 = const()[name = tensor("transpose_79_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_80_perm_0 = const()[name = tensor("transpose_80_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_80 = transpose(perm = transpose_80_perm_0, x = k_57_cast_fp16)[name = tensor("transpose_103")]; - tensor transpose_79 = transpose(perm = transpose_79_perm_0, x = var_2828_cast_fp16)[name = tensor("transpose_104")]; - tensor matrix_ac_29_cast_fp16 = matmul(transpose_x = matrix_ac_29_transpose_x_0, transpose_y = matrix_ac_29_transpose_y_0, x = transpose_79, y = transpose_80)[name = tensor("matrix_ac_29_cast_fp16")]; - tensor matrix_bd_59_begin_0 = const()[name = tensor("matrix_bd_59_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_59_end_0 = const()[name = tensor("matrix_bd_59_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_59_end_mask_0 = const()[name = tensor("matrix_bd_59_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_59_cast_fp16 = slice_by_index(begin = matrix_bd_59_begin_0, end = matrix_bd_59_end_0, end_mask = matrix_bd_59_end_mask_0, x = matrix_bd_57_cast_fp16)[name = tensor("matrix_bd_59_cast_fp16")]; - tensor var_2854_cast_fp16 = add(x = matrix_ac_29_cast_fp16, y = matrix_bd_59_cast_fp16)[name = tensor("op_2854_cast_fp16")]; - tensor _inversed_scores_57_y_0_to_fp16 = const()[name = tensor("_inversed_scores_57_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_57_cast_fp16 = mul(x = var_2854_cast_fp16, y = _inversed_scores_57_y_0_to_fp16)[name = tensor("_inversed_scores_57_cast_fp16")]; - tensor scores_59_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_57_cast_fp16, cond = mask_11)[name = tensor("scores_59_cast_fp16")]; - tensor var_2860_cast_fp16 = softmax(axis = var_23, x = scores_59_cast_fp16)[name = tensor("op_2860_cast_fp16")]; - tensor input_761_cast_fp16 = select(a = var_11_to_fp16, b = var_2860_cast_fp16, cond = mask_11)[name = tensor("input_761_cast_fp16")]; - tensor x_321_transpose_x_0 = const()[name = tensor("x_321_transpose_x_0"), val = tensor(false)]; - tensor x_321_transpose_y_0 = const()[name = tensor("x_321_transpose_y_0"), val = tensor(false)]; - tensor value_31_cast_fp16 = transpose(perm = value_31_perm_0, x = v_29_cast_fp16)[name = tensor("transpose_106")]; - tensor x_321_cast_fp16 = matmul(transpose_x = x_321_transpose_x_0, transpose_y = x_321_transpose_y_0, x = input_761_cast_fp16, y = value_31_cast_fp16)[name = tensor("x_321_cast_fp16")]; - tensor var_2864_perm_0 = const()[name = tensor("op_2864_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_2865 = const()[name = tensor("op_2865"), val = tensor([1, -1, 512])]; - tensor var_2864_cast_fp16 = transpose(perm = var_2864_perm_0, x = x_321_cast_fp16)[name = tensor("transpose_102")]; - tensor input_763_cast_fp16 = reshape(shape = var_2865, x = var_2864_cast_fp16)[name = tensor("input_763_cast_fp16")]; - tensor module_layers_14_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_14_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183855616)))]; - tensor module_layers_14_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_14_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184379968)))]; - tensor linear_133_cast_fp16 = linear(bias = module_layers_14_self_attn_linear_out_bias_to_fp16, weight = module_layers_14_self_attn_linear_out_weight_to_fp16, x = input_763_cast_fp16)[name = tensor("linear_133_cast_fp16")]; - tensor input_767_cast_fp16 = add(x = input_759_cast_fp16, y = linear_133_cast_fp16)[name = tensor("input_767_cast_fp16")]; - tensor x_325_axes_0 = const()[name = tensor("x_325_axes_0"), val = tensor([-1])]; - tensor module_layers_14_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_14_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184381056)))]; - tensor module_layers_14_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_14_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184382144)))]; - tensor x_325_cast_fp16 = layer_norm(axes = x_325_axes_0, beta = module_layers_14_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_14_norm_conv_weight_to_fp16, x = input_767_cast_fp16)[name = tensor("x_325_cast_fp16")]; - tensor input_769_perm_0 = const()[name = tensor("input_769_perm_0"), val = tensor([0, 2, 1])]; - tensor input_771_pad_type_0 = const()[name = tensor("input_771_pad_type_0"), val = tensor("valid")]; - tensor input_771_strides_0 = const()[name = tensor("input_771_strides_0"), val = tensor([1])]; - tensor input_771_pad_0 = const()[name = tensor("input_771_pad_0"), val = tensor([0, 0])]; - tensor input_771_dilations_0 = const()[name = tensor("input_771_dilations_0"), val = tensor([1])]; - tensor input_771_groups_0 = const()[name = tensor("input_771_groups_0"), val = tensor(1)]; - tensor module_layers_14_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_14_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184383232)))]; - tensor module_layers_14_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_14_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185431872)))]; - tensor input_769_cast_fp16 = transpose(perm = input_769_perm_0, x = x_325_cast_fp16)[name = tensor("transpose_101")]; - tensor input_771_cast_fp16 = conv(bias = module_layers_14_conv_pointwise_conv1_bias_to_fp16, dilations = input_771_dilations_0, groups = input_771_groups_0, pad = input_771_pad_0, pad_type = input_771_pad_type_0, strides = input_771_strides_0, weight = module_layers_14_conv_pointwise_conv1_weight_to_fp16, x = input_769_cast_fp16)[name = tensor("input_771_cast_fp16")]; - tensor x_327_split_num_splits_0 = const()[name = tensor("x_327_split_num_splits_0"), val = tensor(2)]; - tensor x_327_split_axis_0 = const()[name = tensor("x_327_split_axis_0"), val = tensor(1)]; - tensor x_327_split_cast_fp16_0, tensor x_327_split_cast_fp16_1 = split(axis = x_327_split_axis_0, num_splits = x_327_split_num_splits_0, x = input_771_cast_fp16)[name = tensor("x_327_split_cast_fp16")]; - tensor x_327_split_1_sigmoid_cast_fp16 = sigmoid(x = x_327_split_cast_fp16_1)[name = tensor("x_327_split_1_sigmoid_cast_fp16")]; - tensor x_327_cast_fp16 = mul(x = x_327_split_cast_fp16_0, y = x_327_split_1_sigmoid_cast_fp16)[name = tensor("x_327_cast_fp16")]; - tensor input_773_cast_fp16 = select(a = var_11_to_fp16, b = x_327_cast_fp16, cond = var_453)[name = tensor("input_773_cast_fp16")]; - tensor input_775_pad_0 = const()[name = tensor("input_775_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_775_mode_0 = const()[name = tensor("input_775_mode_0"), val = tensor("constant")]; - tensor const_213_to_fp16 = const()[name = tensor("const_213_to_fp16"), val = tensor(0x0p+0)]; - tensor input_775_cast_fp16 = pad(constant_val = const_213_to_fp16, mode = input_775_mode_0, pad = input_775_pad_0, x = input_773_cast_fp16)[name = tensor("input_775_cast_fp16")]; - tensor input_777_pad_type_0 = const()[name = tensor("input_777_pad_type_0"), val = tensor("valid")]; - tensor input_777_groups_0 = const()[name = tensor("input_777_groups_0"), val = tensor(512)]; - tensor input_777_strides_0 = const()[name = tensor("input_777_strides_0"), val = tensor([1])]; - tensor input_777_pad_0 = const()[name = tensor("input_777_pad_0"), val = tensor([0, 0])]; - tensor input_777_dilations_0 = const()[name = tensor("input_777_dilations_0"), val = tensor([1])]; - tensor const_262_to_fp16 = const()[name = tensor("const_262_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185433984)))]; - tensor const_263_to_fp16 = const()[name = tensor("const_263_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185443264)))]; - tensor input_779_cast_fp16 = conv(bias = const_263_to_fp16, dilations = input_777_dilations_0, groups = input_777_groups_0, pad = input_777_pad_0, pad_type = input_777_pad_type_0, strides = input_777_strides_0, weight = const_262_to_fp16, x = input_775_cast_fp16)[name = tensor("input_779_cast_fp16")]; - tensor input_781_cast_fp16 = silu(x = input_779_cast_fp16)[name = tensor("input_781_cast_fp16")]; - tensor x_329_pad_type_0 = const()[name = tensor("x_329_pad_type_0"), val = tensor("valid")]; - tensor x_329_strides_0 = const()[name = tensor("x_329_strides_0"), val = tensor([1])]; - tensor x_329_pad_0 = const()[name = tensor("x_329_pad_0"), val = tensor([0, 0])]; - tensor x_329_dilations_0 = const()[name = tensor("x_329_dilations_0"), val = tensor([1])]; - tensor x_329_groups_0 = const()[name = tensor("x_329_groups_0"), val = tensor(1)]; - tensor module_layers_14_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_14_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185444352)))]; - tensor module_layers_14_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_14_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185968704)))]; - tensor x_329_cast_fp16 = conv(bias = module_layers_14_conv_pointwise_conv2_bias_to_fp16, dilations = x_329_dilations_0, groups = x_329_groups_0, pad = x_329_pad_0, pad_type = x_329_pad_type_0, strides = x_329_strides_0, weight = module_layers_14_conv_pointwise_conv2_weight_to_fp16, x = input_781_cast_fp16)[name = tensor("x_329_cast_fp16")]; - tensor input_783_perm_0 = const()[name = tensor("input_783_perm_0"), val = tensor([0, 2, 1])]; - tensor input_783_cast_fp16 = transpose(perm = input_783_perm_0, x = x_329_cast_fp16)[name = tensor("transpose_100")]; - tensor input_785_cast_fp16 = add(x = input_767_cast_fp16, y = input_783_cast_fp16)[name = tensor("input_785_cast_fp16")]; - tensor input_787_axes_0 = const()[name = tensor("input_787_axes_0"), val = tensor([-1])]; - tensor module_layers_14_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_14_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185969792)))]; - tensor module_layers_14_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_14_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185970880)))]; - tensor input_787_cast_fp16 = layer_norm(axes = input_787_axes_0, beta = module_layers_14_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_14_norm_feed_forward2_weight_to_fp16, x = input_785_cast_fp16)[name = tensor("input_787_cast_fp16")]; - tensor module_layers_14_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_14_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185971968)))]; - tensor module_layers_14_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_14_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188069184)))]; - tensor linear_134_cast_fp16 = linear(bias = module_layers_14_feed_forward2_linear1_bias_to_fp16, weight = module_layers_14_feed_forward2_linear1_weight_to_fp16, x = input_787_cast_fp16)[name = tensor("linear_134_cast_fp16")]; - tensor input_791_cast_fp16 = silu(x = linear_134_cast_fp16)[name = tensor("input_791_cast_fp16")]; - tensor module_layers_14_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_14_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188073344)))]; - tensor module_layers_14_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_14_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190170560)))]; - tensor linear_135_cast_fp16 = linear(bias = module_layers_14_feed_forward2_linear2_bias_to_fp16, weight = module_layers_14_feed_forward2_linear2_weight_to_fp16, x = input_791_cast_fp16)[name = tensor("linear_135_cast_fp16")]; - tensor var_2931_to_fp16 = const()[name = tensor("op_2931_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2932_cast_fp16 = mul(x = linear_135_cast_fp16, y = var_2931_to_fp16)[name = tensor("op_2932_cast_fp16")]; - tensor input_797_cast_fp16 = add(x = input_785_cast_fp16, y = var_2932_cast_fp16)[name = tensor("input_797_cast_fp16")]; - tensor input_799_axes_0 = const()[name = tensor("input_799_axes_0"), val = tensor([-1])]; - tensor module_layers_14_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_14_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190171648)))]; - tensor module_layers_14_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_14_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190172736)))]; - tensor input_799_cast_fp16 = layer_norm(axes = input_799_axes_0, beta = module_layers_14_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_14_norm_out_weight_to_fp16, x = input_797_cast_fp16)[name = tensor("input_799_cast_fp16")]; - tensor input_801_axes_0 = const()[name = tensor("input_801_axes_0"), val = tensor([-1])]; - tensor module_layers_15_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_15_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190173824)))]; - tensor module_layers_15_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_15_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190174912)))]; - tensor input_801_cast_fp16 = layer_norm(axes = input_801_axes_0, beta = module_layers_15_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_15_norm_feed_forward1_weight_to_fp16, x = input_799_cast_fp16)[name = tensor("input_801_cast_fp16")]; - tensor module_layers_15_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_15_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190176000)))]; - tensor module_layers_15_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_15_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(192273216)))]; - tensor linear_136_cast_fp16 = linear(bias = module_layers_15_feed_forward1_linear1_bias_to_fp16, weight = module_layers_15_feed_forward1_linear1_weight_to_fp16, x = input_801_cast_fp16)[name = tensor("linear_136_cast_fp16")]; - tensor input_805_cast_fp16 = silu(x = linear_136_cast_fp16)[name = tensor("input_805_cast_fp16")]; - tensor module_layers_15_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_15_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(192277376)))]; - tensor module_layers_15_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_15_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194374592)))]; - tensor linear_137_cast_fp16 = linear(bias = module_layers_15_feed_forward1_linear2_bias_to_fp16, weight = module_layers_15_feed_forward1_linear2_weight_to_fp16, x = input_805_cast_fp16)[name = tensor("linear_137_cast_fp16")]; - tensor var_2962_to_fp16 = const()[name = tensor("op_2962_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2963_cast_fp16 = mul(x = linear_137_cast_fp16, y = var_2962_to_fp16)[name = tensor("op_2963_cast_fp16")]; - tensor input_811_cast_fp16 = add(x = input_799_cast_fp16, y = var_2963_cast_fp16)[name = tensor("input_811_cast_fp16")]; - tensor query_31_axes_0 = const()[name = tensor("query_31_axes_0"), val = tensor([-1])]; - tensor module_layers_15_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_15_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194375680)))]; - tensor module_layers_15_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_15_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194376768)))]; - tensor query_31_cast_fp16 = layer_norm(axes = query_31_axes_0, beta = module_layers_15_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_15_norm_self_att_weight_to_fp16, x = input_811_cast_fp16)[name = tensor("query_31_cast_fp16")]; - tensor module_layers_15_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_15_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194377856)))]; - tensor module_layers_15_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_15_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194902208)))]; - tensor linear_138_cast_fp16 = linear(bias = module_layers_15_self_attn_linear_q_bias_to_fp16, weight = module_layers_15_self_attn_linear_q_weight_to_fp16, x = query_31_cast_fp16)[name = tensor("linear_138_cast_fp16")]; - tensor var_2980 = const()[name = tensor("op_2980"), val = tensor([1, -1, 8, 64])]; - tensor q_91_cast_fp16 = reshape(shape = var_2980, x = linear_138_cast_fp16)[name = tensor("q_91_cast_fp16")]; - tensor module_layers_15_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_15_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194903296)))]; - tensor module_layers_15_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_15_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195427648)))]; - tensor linear_139_cast_fp16 = linear(bias = module_layers_15_self_attn_linear_k_bias_to_fp16, weight = module_layers_15_self_attn_linear_k_weight_to_fp16, x = query_31_cast_fp16)[name = tensor("linear_139_cast_fp16")]; - tensor var_2985 = const()[name = tensor("op_2985"), val = tensor([1, -1, 8, 64])]; - tensor k_61_cast_fp16 = reshape(shape = var_2985, x = linear_139_cast_fp16)[name = tensor("k_61_cast_fp16")]; - tensor module_layers_15_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_15_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195428736)))]; - tensor module_layers_15_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_15_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195953088)))]; - tensor linear_140_cast_fp16 = linear(bias = module_layers_15_self_attn_linear_v_bias_to_fp16, weight = module_layers_15_self_attn_linear_v_weight_to_fp16, x = query_31_cast_fp16)[name = tensor("linear_140_cast_fp16")]; - tensor var_2990 = const()[name = tensor("op_2990"), val = tensor([1, -1, 8, 64])]; - tensor v_31_cast_fp16 = reshape(shape = var_2990, x = linear_140_cast_fp16)[name = tensor("v_31_cast_fp16")]; - tensor value_33_perm_0 = const()[name = tensor("value_33_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_15_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_15_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195954176)))]; - tensor var_3002_cast_fp16 = add(x = q_91_cast_fp16, y = module_layers_15_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3002_cast_fp16")]; - tensor module_layers_15_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_15_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195955264)))]; - tensor var_3004_cast_fp16 = add(x = q_91_cast_fp16, y = module_layers_15_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3004_cast_fp16")]; - tensor q_with_bias_v_31_perm_0 = const()[name = tensor("q_with_bias_v_31_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_337_transpose_x_0 = const()[name = tensor("x_337_transpose_x_0"), val = tensor(false)]; - tensor x_337_transpose_y_0 = const()[name = tensor("x_337_transpose_y_0"), val = tensor(false)]; - tensor var_3006_to_fp16 = const()[name = tensor("op_3006_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195956352)))]; - tensor q_with_bias_v_31_cast_fp16 = transpose(perm = q_with_bias_v_31_perm_0, x = var_3004_cast_fp16)[name = tensor("transpose_98")]; - tensor x_337_cast_fp16 = matmul(transpose_x = x_337_transpose_x_0, transpose_y = x_337_transpose_y_0, x = q_with_bias_v_31_cast_fp16, y = var_3006_to_fp16)[name = tensor("x_337_cast_fp16")]; - tensor x_339_pad_0 = const()[name = tensor("x_339_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_339_mode_0 = const()[name = tensor("x_339_mode_0"), val = tensor("constant")]; - tensor const_220_to_fp16 = const()[name = tensor("const_220_to_fp16"), val = tensor(0x0p+0)]; - tensor x_339_cast_fp16 = pad(constant_val = const_220_to_fp16, mode = x_339_mode_0, pad = x_339_pad_0, x = x_337_cast_fp16)[name = tensor("x_339_cast_fp16")]; - tensor var_3014 = const()[name = tensor("op_3014"), val = tensor([1, 8, -1, 188])]; - tensor x_341_cast_fp16 = reshape(shape = var_3014, x = x_339_cast_fp16)[name = tensor("x_341_cast_fp16")]; - tensor var_3018_begin_0 = const()[name = tensor("op_3018_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_3018_end_0 = const()[name = tensor("op_3018_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_3018_end_mask_0 = const()[name = tensor("op_3018_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_3018_cast_fp16 = slice_by_index(begin = var_3018_begin_0, end = var_3018_end_0, end_mask = var_3018_end_mask_0, x = x_341_cast_fp16)[name = tensor("op_3018_cast_fp16")]; - tensor var_3019 = const()[name = tensor("op_3019"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_61_cast_fp16 = reshape(shape = var_3019, x = var_3018_cast_fp16)[name = tensor("matrix_bd_61_cast_fp16")]; - tensor matrix_ac_31_transpose_x_0 = const()[name = tensor("matrix_ac_31_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_31_transpose_y_0 = const()[name = tensor("matrix_ac_31_transpose_y_0"), val = tensor(false)]; - tensor transpose_81_perm_0 = const()[name = tensor("transpose_81_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_82_perm_0 = const()[name = tensor("transpose_82_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_82 = transpose(perm = transpose_82_perm_0, x = k_61_cast_fp16)[name = tensor("transpose_96")]; - tensor transpose_81 = transpose(perm = transpose_81_perm_0, x = var_3002_cast_fp16)[name = tensor("transpose_97")]; - tensor matrix_ac_31_cast_fp16 = matmul(transpose_x = matrix_ac_31_transpose_x_0, transpose_y = matrix_ac_31_transpose_y_0, x = transpose_81, y = transpose_82)[name = tensor("matrix_ac_31_cast_fp16")]; - tensor matrix_bd_63_begin_0 = const()[name = tensor("matrix_bd_63_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_63_end_0 = const()[name = tensor("matrix_bd_63_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_63_end_mask_0 = const()[name = tensor("matrix_bd_63_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_63_cast_fp16 = slice_by_index(begin = matrix_bd_63_begin_0, end = matrix_bd_63_end_0, end_mask = matrix_bd_63_end_mask_0, x = matrix_bd_61_cast_fp16)[name = tensor("matrix_bd_63_cast_fp16")]; - tensor var_3028_cast_fp16 = add(x = matrix_ac_31_cast_fp16, y = matrix_bd_63_cast_fp16)[name = tensor("op_3028_cast_fp16")]; - tensor _inversed_scores_61_y_0_to_fp16 = const()[name = tensor("_inversed_scores_61_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_61_cast_fp16 = mul(x = var_3028_cast_fp16, y = _inversed_scores_61_y_0_to_fp16)[name = tensor("_inversed_scores_61_cast_fp16")]; - tensor scores_63_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_61_cast_fp16, cond = mask_11)[name = tensor("scores_63_cast_fp16")]; - tensor var_3034_cast_fp16 = softmax(axis = var_23, x = scores_63_cast_fp16)[name = tensor("op_3034_cast_fp16")]; - tensor input_813_cast_fp16 = select(a = var_11_to_fp16, b = var_3034_cast_fp16, cond = mask_11)[name = tensor("input_813_cast_fp16")]; - tensor x_343_transpose_x_0 = const()[name = tensor("x_343_transpose_x_0"), val = tensor(false)]; - tensor x_343_transpose_y_0 = const()[name = tensor("x_343_transpose_y_0"), val = tensor(false)]; - tensor value_33_cast_fp16 = transpose(perm = value_33_perm_0, x = v_31_cast_fp16)[name = tensor("transpose_99")]; - tensor x_343_cast_fp16 = matmul(transpose_x = x_343_transpose_x_0, transpose_y = x_343_transpose_y_0, x = input_813_cast_fp16, y = value_33_cast_fp16)[name = tensor("x_343_cast_fp16")]; - tensor var_3038_perm_0 = const()[name = tensor("op_3038_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_3039 = const()[name = tensor("op_3039"), val = tensor([1, -1, 512])]; - tensor var_3038_cast_fp16 = transpose(perm = var_3038_perm_0, x = x_343_cast_fp16)[name = tensor("transpose_95")]; - tensor input_815_cast_fp16 = reshape(shape = var_3039, x = var_3038_cast_fp16)[name = tensor("input_815_cast_fp16")]; - tensor module_layers_15_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_15_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196340416)))]; - tensor module_layers_15_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_15_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196864768)))]; - tensor linear_142_cast_fp16 = linear(bias = module_layers_15_self_attn_linear_out_bias_to_fp16, weight = module_layers_15_self_attn_linear_out_weight_to_fp16, x = input_815_cast_fp16)[name = tensor("linear_142_cast_fp16")]; - tensor input_819_cast_fp16 = add(x = input_811_cast_fp16, y = linear_142_cast_fp16)[name = tensor("input_819_cast_fp16")]; - tensor x_347_axes_0 = const()[name = tensor("x_347_axes_0"), val = tensor([-1])]; - tensor module_layers_15_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_15_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196865856)))]; - tensor module_layers_15_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_15_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196866944)))]; - tensor x_347_cast_fp16 = layer_norm(axes = x_347_axes_0, beta = module_layers_15_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_15_norm_conv_weight_to_fp16, x = input_819_cast_fp16)[name = tensor("x_347_cast_fp16")]; - tensor input_821_perm_0 = const()[name = tensor("input_821_perm_0"), val = tensor([0, 2, 1])]; - tensor input_823_pad_type_0 = const()[name = tensor("input_823_pad_type_0"), val = tensor("valid")]; - tensor input_823_strides_0 = const()[name = tensor("input_823_strides_0"), val = tensor([1])]; - tensor input_823_pad_0 = const()[name = tensor("input_823_pad_0"), val = tensor([0, 0])]; - tensor input_823_dilations_0 = const()[name = tensor("input_823_dilations_0"), val = tensor([1])]; - tensor input_823_groups_0 = const()[name = tensor("input_823_groups_0"), val = tensor(1)]; - tensor module_layers_15_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_15_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196868032)))]; - tensor module_layers_15_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_15_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(197916672)))]; - tensor input_821_cast_fp16 = transpose(perm = input_821_perm_0, x = x_347_cast_fp16)[name = tensor("transpose_94")]; - tensor input_823_cast_fp16 = conv(bias = module_layers_15_conv_pointwise_conv1_bias_to_fp16, dilations = input_823_dilations_0, groups = input_823_groups_0, pad = input_823_pad_0, pad_type = input_823_pad_type_0, strides = input_823_strides_0, weight = module_layers_15_conv_pointwise_conv1_weight_to_fp16, x = input_821_cast_fp16)[name = tensor("input_823_cast_fp16")]; - tensor x_349_split_num_splits_0 = const()[name = tensor("x_349_split_num_splits_0"), val = tensor(2)]; - tensor x_349_split_axis_0 = const()[name = tensor("x_349_split_axis_0"), val = tensor(1)]; - tensor x_349_split_cast_fp16_0, tensor x_349_split_cast_fp16_1 = split(axis = x_349_split_axis_0, num_splits = x_349_split_num_splits_0, x = input_823_cast_fp16)[name = tensor("x_349_split_cast_fp16")]; - tensor x_349_split_1_sigmoid_cast_fp16 = sigmoid(x = x_349_split_cast_fp16_1)[name = tensor("x_349_split_1_sigmoid_cast_fp16")]; - tensor x_349_cast_fp16 = mul(x = x_349_split_cast_fp16_0, y = x_349_split_1_sigmoid_cast_fp16)[name = tensor("x_349_cast_fp16")]; - tensor input_825_cast_fp16 = select(a = var_11_to_fp16, b = x_349_cast_fp16, cond = var_453)[name = tensor("input_825_cast_fp16")]; - tensor input_827_pad_0 = const()[name = tensor("input_827_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_827_mode_0 = const()[name = tensor("input_827_mode_0"), val = tensor("constant")]; - tensor const_223_to_fp16 = const()[name = tensor("const_223_to_fp16"), val = tensor(0x0p+0)]; - tensor input_827_cast_fp16 = pad(constant_val = const_223_to_fp16, mode = input_827_mode_0, pad = input_827_pad_0, x = input_825_cast_fp16)[name = tensor("input_827_cast_fp16")]; - tensor input_829_pad_type_0 = const()[name = tensor("input_829_pad_type_0"), val = tensor("valid")]; - tensor input_829_groups_0 = const()[name = tensor("input_829_groups_0"), val = tensor(512)]; - tensor input_829_strides_0 = const()[name = tensor("input_829_strides_0"), val = tensor([1])]; - tensor input_829_pad_0 = const()[name = tensor("input_829_pad_0"), val = tensor([0, 0])]; - tensor input_829_dilations_0 = const()[name = tensor("input_829_dilations_0"), val = tensor([1])]; - tensor const_264_to_fp16 = const()[name = tensor("const_264_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(197918784)))]; - tensor const_265_to_fp16 = const()[name = tensor("const_265_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(197928064)))]; - tensor input_831_cast_fp16 = conv(bias = const_265_to_fp16, dilations = input_829_dilations_0, groups = input_829_groups_0, pad = input_829_pad_0, pad_type = input_829_pad_type_0, strides = input_829_strides_0, weight = const_264_to_fp16, x = input_827_cast_fp16)[name = tensor("input_831_cast_fp16")]; - tensor input_833_cast_fp16 = silu(x = input_831_cast_fp16)[name = tensor("input_833_cast_fp16")]; - tensor x_351_pad_type_0 = const()[name = tensor("x_351_pad_type_0"), val = tensor("valid")]; - tensor x_351_strides_0 = const()[name = tensor("x_351_strides_0"), val = tensor([1])]; - tensor x_351_pad_0 = const()[name = tensor("x_351_pad_0"), val = tensor([0, 0])]; - tensor x_351_dilations_0 = const()[name = tensor("x_351_dilations_0"), val = tensor([1])]; - tensor x_351_groups_0 = const()[name = tensor("x_351_groups_0"), val = tensor(1)]; - tensor module_layers_15_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_15_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(197929152)))]; - tensor module_layers_15_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_15_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198453504)))]; - tensor x_351_cast_fp16 = conv(bias = module_layers_15_conv_pointwise_conv2_bias_to_fp16, dilations = x_351_dilations_0, groups = x_351_groups_0, pad = x_351_pad_0, pad_type = x_351_pad_type_0, strides = x_351_strides_0, weight = module_layers_15_conv_pointwise_conv2_weight_to_fp16, x = input_833_cast_fp16)[name = tensor("x_351_cast_fp16")]; - tensor input_835_perm_0 = const()[name = tensor("input_835_perm_0"), val = tensor([0, 2, 1])]; - tensor input_835_cast_fp16 = transpose(perm = input_835_perm_0, x = x_351_cast_fp16)[name = tensor("transpose_93")]; - tensor input_837_cast_fp16 = add(x = input_819_cast_fp16, y = input_835_cast_fp16)[name = tensor("input_837_cast_fp16")]; - tensor input_839_axes_0 = const()[name = tensor("input_839_axes_0"), val = tensor([-1])]; - tensor module_layers_15_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_15_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198454592)))]; - tensor module_layers_15_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_15_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198455680)))]; - tensor input_839_cast_fp16 = layer_norm(axes = input_839_axes_0, beta = module_layers_15_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_15_norm_feed_forward2_weight_to_fp16, x = input_837_cast_fp16)[name = tensor("input_839_cast_fp16")]; - tensor module_layers_15_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_15_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198456768)))]; - tensor module_layers_15_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_15_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(200553984)))]; - tensor linear_143_cast_fp16 = linear(bias = module_layers_15_feed_forward2_linear1_bias_to_fp16, weight = module_layers_15_feed_forward2_linear1_weight_to_fp16, x = input_839_cast_fp16)[name = tensor("linear_143_cast_fp16")]; - tensor input_843_cast_fp16 = silu(x = linear_143_cast_fp16)[name = tensor("input_843_cast_fp16")]; - tensor module_layers_15_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_15_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(200558144)))]; - tensor module_layers_15_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_15_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202655360)))]; - tensor linear_144_cast_fp16 = linear(bias = module_layers_15_feed_forward2_linear2_bias_to_fp16, weight = module_layers_15_feed_forward2_linear2_weight_to_fp16, x = input_843_cast_fp16)[name = tensor("linear_144_cast_fp16")]; - tensor var_3105_to_fp16 = const()[name = tensor("op_3105_to_fp16"), val = tensor(0x1p-1)]; - tensor var_3106_cast_fp16 = mul(x = linear_144_cast_fp16, y = var_3105_to_fp16)[name = tensor("op_3106_cast_fp16")]; - tensor input_849_cast_fp16 = add(x = input_837_cast_fp16, y = var_3106_cast_fp16)[name = tensor("input_849_cast_fp16")]; - tensor input_851_axes_0 = const()[name = tensor("input_851_axes_0"), val = tensor([-1])]; - tensor module_layers_15_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_15_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202656448)))]; - tensor module_layers_15_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_15_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202657536)))]; - tensor input_851_cast_fp16 = layer_norm(axes = input_851_axes_0, beta = module_layers_15_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_15_norm_out_weight_to_fp16, x = input_849_cast_fp16)[name = tensor("input_851_cast_fp16")]; - tensor input_853_axes_0 = const()[name = tensor("input_853_axes_0"), val = tensor([-1])]; - tensor module_layers_16_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_16_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202658624)))]; - tensor module_layers_16_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_16_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202659712)))]; - tensor input_853_cast_fp16 = layer_norm(axes = input_853_axes_0, beta = module_layers_16_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_16_norm_feed_forward1_weight_to_fp16, x = input_851_cast_fp16)[name = tensor("input_853_cast_fp16")]; - tensor module_layers_16_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_16_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202660800)))]; - tensor module_layers_16_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_16_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(204758016)))]; - tensor linear_145_cast_fp16 = linear(bias = module_layers_16_feed_forward1_linear1_bias_to_fp16, weight = module_layers_16_feed_forward1_linear1_weight_to_fp16, x = input_853_cast_fp16)[name = tensor("linear_145_cast_fp16")]; - tensor input_857_cast_fp16 = silu(x = linear_145_cast_fp16)[name = tensor("input_857_cast_fp16")]; - tensor module_layers_16_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_16_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(204762176)))]; - tensor module_layers_16_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_16_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(206859392)))]; - tensor linear_146_cast_fp16 = linear(bias = module_layers_16_feed_forward1_linear2_bias_to_fp16, weight = module_layers_16_feed_forward1_linear2_weight_to_fp16, x = input_857_cast_fp16)[name = tensor("linear_146_cast_fp16")]; - tensor var_3136_to_fp16 = const()[name = tensor("op_3136_to_fp16"), val = tensor(0x1p-1)]; - tensor var_3137_cast_fp16 = mul(x = linear_146_cast_fp16, y = var_3136_to_fp16)[name = tensor("op_3137_cast_fp16")]; - tensor input_863_cast_fp16 = add(x = input_851_cast_fp16, y = var_3137_cast_fp16)[name = tensor("input_863_cast_fp16")]; - tensor query_axes_0 = const()[name = tensor("query_axes_0"), val = tensor([-1])]; - tensor module_layers_16_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_16_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(206860480)))]; - tensor module_layers_16_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_16_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(206861568)))]; - tensor query_cast_fp16 = layer_norm(axes = query_axes_0, beta = module_layers_16_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_16_norm_self_att_weight_to_fp16, x = input_863_cast_fp16)[name = tensor("query_cast_fp16")]; - tensor module_layers_16_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_16_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(206862656)))]; - tensor module_layers_16_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_16_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(207387008)))]; - tensor linear_147_cast_fp16 = linear(bias = module_layers_16_self_attn_linear_q_bias_to_fp16, weight = module_layers_16_self_attn_linear_q_weight_to_fp16, x = query_cast_fp16)[name = tensor("linear_147_cast_fp16")]; - tensor var_3154 = const()[name = tensor("op_3154"), val = tensor([1, -1, 8, 64])]; - tensor q_97_cast_fp16 = reshape(shape = var_3154, x = linear_147_cast_fp16)[name = tensor("q_97_cast_fp16")]; - tensor module_layers_16_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_16_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(207388096)))]; - tensor module_layers_16_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_16_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(207912448)))]; - tensor linear_148_cast_fp16 = linear(bias = module_layers_16_self_attn_linear_k_bias_to_fp16, weight = module_layers_16_self_attn_linear_k_weight_to_fp16, x = query_cast_fp16)[name = tensor("linear_148_cast_fp16")]; - tensor var_3159 = const()[name = tensor("op_3159"), val = tensor([1, -1, 8, 64])]; - tensor k_65_cast_fp16 = reshape(shape = var_3159, x = linear_148_cast_fp16)[name = tensor("k_65_cast_fp16")]; - tensor module_layers_16_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_16_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(207913536)))]; - tensor module_layers_16_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_16_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208437888)))]; - tensor linear_149_cast_fp16 = linear(bias = module_layers_16_self_attn_linear_v_bias_to_fp16, weight = module_layers_16_self_attn_linear_v_weight_to_fp16, x = query_cast_fp16)[name = tensor("linear_149_cast_fp16")]; - tensor var_3164 = const()[name = tensor("op_3164"), val = tensor([1, -1, 8, 64])]; - tensor v_cast_fp16 = reshape(shape = var_3164, x = linear_149_cast_fp16)[name = tensor("v_cast_fp16")]; - tensor value_perm_0 = const()[name = tensor("value_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_16_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_16_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208438976)))]; - tensor var_3176_cast_fp16 = add(x = q_97_cast_fp16, y = module_layers_16_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3176_cast_fp16")]; - tensor module_layers_16_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_16_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208440064)))]; - tensor var_3178_cast_fp16 = add(x = q_97_cast_fp16, y = module_layers_16_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3178_cast_fp16")]; - tensor q_with_bias_v_perm_0 = const()[name = tensor("q_with_bias_v_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_359_transpose_x_0 = const()[name = tensor("x_359_transpose_x_0"), val = tensor(false)]; - tensor x_359_transpose_y_0 = const()[name = tensor("x_359_transpose_y_0"), val = tensor(false)]; - tensor var_3180_to_fp16 = const()[name = tensor("op_3180_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208441152)))]; - tensor q_with_bias_v_cast_fp16 = transpose(perm = q_with_bias_v_perm_0, x = var_3178_cast_fp16)[name = tensor("transpose_91")]; - tensor x_359_cast_fp16 = matmul(transpose_x = x_359_transpose_x_0, transpose_y = x_359_transpose_y_0, x = q_with_bias_v_cast_fp16, y = var_3180_to_fp16)[name = tensor("x_359_cast_fp16")]; - tensor x_361_pad_0 = const()[name = tensor("x_361_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_361_mode_0 = const()[name = tensor("x_361_mode_0"), val = tensor("constant")]; - tensor const_230_to_fp16 = const()[name = tensor("const_230_to_fp16"), val = tensor(0x0p+0)]; - tensor x_361_cast_fp16 = pad(constant_val = const_230_to_fp16, mode = x_361_mode_0, pad = x_361_pad_0, x = x_359_cast_fp16)[name = tensor("x_361_cast_fp16")]; - tensor var_3188 = const()[name = tensor("op_3188"), val = tensor([1, 8, -1, 188])]; - tensor x_363_cast_fp16 = reshape(shape = var_3188, x = x_361_cast_fp16)[name = tensor("x_363_cast_fp16")]; - tensor var_3192_begin_0 = const()[name = tensor("op_3192_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_3192_end_0 = const()[name = tensor("op_3192_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_3192_end_mask_0 = const()[name = tensor("op_3192_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_3192_cast_fp16 = slice_by_index(begin = var_3192_begin_0, end = var_3192_end_0, end_mask = var_3192_end_mask_0, x = x_363_cast_fp16)[name = tensor("op_3192_cast_fp16")]; - tensor var_3193 = const()[name = tensor("op_3193"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_65_cast_fp16 = reshape(shape = var_3193, x = var_3192_cast_fp16)[name = tensor("matrix_bd_65_cast_fp16")]; - tensor matrix_ac_transpose_x_0 = const()[name = tensor("matrix_ac_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_transpose_y_0 = const()[name = tensor("matrix_ac_transpose_y_0"), val = tensor(false)]; - tensor transpose_83_perm_0 = const()[name = tensor("transpose_83_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_84_perm_0 = const()[name = tensor("transpose_84_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_84 = transpose(perm = transpose_84_perm_0, x = k_65_cast_fp16)[name = tensor("transpose_89")]; - tensor transpose_83 = transpose(perm = transpose_83_perm_0, x = var_3176_cast_fp16)[name = tensor("transpose_90")]; - tensor matrix_ac_cast_fp16 = matmul(transpose_x = matrix_ac_transpose_x_0, transpose_y = matrix_ac_transpose_y_0, x = transpose_83, y = transpose_84)[name = tensor("matrix_ac_cast_fp16")]; - tensor matrix_bd_begin_0 = const()[name = tensor("matrix_bd_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_end_0 = const()[name = tensor("matrix_bd_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_end_mask_0 = const()[name = tensor("matrix_bd_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_cast_fp16 = slice_by_index(begin = matrix_bd_begin_0, end = matrix_bd_end_0, end_mask = matrix_bd_end_mask_0, x = matrix_bd_65_cast_fp16)[name = tensor("matrix_bd_cast_fp16")]; - tensor var_3202_cast_fp16 = add(x = matrix_ac_cast_fp16, y = matrix_bd_cast_fp16)[name = tensor("op_3202_cast_fp16")]; - tensor _inversed_scores_65_y_0_to_fp16 = const()[name = tensor("_inversed_scores_65_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_65_cast_fp16 = mul(x = var_3202_cast_fp16, y = _inversed_scores_65_y_0_to_fp16)[name = tensor("_inversed_scores_65_cast_fp16")]; - tensor scores_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_65_cast_fp16, cond = mask_11)[name = tensor("scores_cast_fp16")]; - tensor var_3208_cast_fp16 = softmax(axis = var_23, x = scores_cast_fp16)[name = tensor("op_3208_cast_fp16")]; - tensor input_865_cast_fp16 = select(a = var_11_to_fp16, b = var_3208_cast_fp16, cond = mask_11)[name = tensor("input_865_cast_fp16")]; - tensor x_365_transpose_x_0 = const()[name = tensor("x_365_transpose_x_0"), val = tensor(false)]; - tensor x_365_transpose_y_0 = const()[name = tensor("x_365_transpose_y_0"), val = tensor(false)]; - tensor value_cast_fp16 = transpose(perm = value_perm_0, x = v_cast_fp16)[name = tensor("transpose_92")]; - tensor x_365_cast_fp16 = matmul(transpose_x = x_365_transpose_x_0, transpose_y = x_365_transpose_y_0, x = input_865_cast_fp16, y = value_cast_fp16)[name = tensor("x_365_cast_fp16")]; - tensor var_3212_perm_0 = const()[name = tensor("op_3212_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_3213 = const()[name = tensor("op_3213"), val = tensor([1, -1, 512])]; - tensor var_3212_cast_fp16 = transpose(perm = var_3212_perm_0, x = x_365_cast_fp16)[name = tensor("transpose_88")]; - tensor input_867_cast_fp16 = reshape(shape = var_3213, x = var_3212_cast_fp16)[name = tensor("input_867_cast_fp16")]; - tensor module_layers_16_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_16_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208825216)))]; - tensor module_layers_16_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_16_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(209349568)))]; - tensor linear_151_cast_fp16 = linear(bias = module_layers_16_self_attn_linear_out_bias_to_fp16, weight = module_layers_16_self_attn_linear_out_weight_to_fp16, x = input_867_cast_fp16)[name = tensor("linear_151_cast_fp16")]; - tensor input_871_cast_fp16 = add(x = input_863_cast_fp16, y = linear_151_cast_fp16)[name = tensor("input_871_cast_fp16")]; - tensor x_369_axes_0 = const()[name = tensor("x_369_axes_0"), val = tensor([-1])]; - tensor module_layers_16_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_16_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(209350656)))]; - tensor module_layers_16_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_16_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(209351744)))]; - tensor x_369_cast_fp16 = layer_norm(axes = x_369_axes_0, beta = module_layers_16_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_16_norm_conv_weight_to_fp16, x = input_871_cast_fp16)[name = tensor("x_369_cast_fp16")]; - tensor input_873_perm_0 = const()[name = tensor("input_873_perm_0"), val = tensor([0, 2, 1])]; - tensor input_875_pad_type_0 = const()[name = tensor("input_875_pad_type_0"), val = tensor("valid")]; - tensor input_875_strides_0 = const()[name = tensor("input_875_strides_0"), val = tensor([1])]; - tensor input_875_pad_0 = const()[name = tensor("input_875_pad_0"), val = tensor([0, 0])]; - tensor input_875_dilations_0 = const()[name = tensor("input_875_dilations_0"), val = tensor([1])]; - tensor input_875_groups_0 = const()[name = tensor("input_875_groups_0"), val = tensor(1)]; - tensor module_layers_16_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_16_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(209352832)))]; - tensor module_layers_16_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_16_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210401472)))]; - tensor input_873_cast_fp16 = transpose(perm = input_873_perm_0, x = x_369_cast_fp16)[name = tensor("transpose_87")]; - tensor input_875_cast_fp16 = conv(bias = module_layers_16_conv_pointwise_conv1_bias_to_fp16, dilations = input_875_dilations_0, groups = input_875_groups_0, pad = input_875_pad_0, pad_type = input_875_pad_type_0, strides = input_875_strides_0, weight = module_layers_16_conv_pointwise_conv1_weight_to_fp16, x = input_873_cast_fp16)[name = tensor("input_875_cast_fp16")]; - tensor x_371_split_num_splits_0 = const()[name = tensor("x_371_split_num_splits_0"), val = tensor(2)]; - tensor x_371_split_axis_0 = const()[name = tensor("x_371_split_axis_0"), val = tensor(1)]; - tensor x_371_split_cast_fp16_0, tensor x_371_split_cast_fp16_1 = split(axis = x_371_split_axis_0, num_splits = x_371_split_num_splits_0, x = input_875_cast_fp16)[name = tensor("x_371_split_cast_fp16")]; - tensor x_371_split_1_sigmoid_cast_fp16 = sigmoid(x = x_371_split_cast_fp16_1)[name = tensor("x_371_split_1_sigmoid_cast_fp16")]; - tensor x_371_cast_fp16 = mul(x = x_371_split_cast_fp16_0, y = x_371_split_1_sigmoid_cast_fp16)[name = tensor("x_371_cast_fp16")]; - tensor input_877_cast_fp16 = select(a = var_11_to_fp16, b = x_371_cast_fp16, cond = var_453)[name = tensor("input_877_cast_fp16")]; - tensor input_879_pad_0 = const()[name = tensor("input_879_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_879_mode_0 = const()[name = tensor("input_879_mode_0"), val = tensor("constant")]; - tensor const_233_to_fp16 = const()[name = tensor("const_233_to_fp16"), val = tensor(0x0p+0)]; - tensor input_879_cast_fp16 = pad(constant_val = const_233_to_fp16, mode = input_879_mode_0, pad = input_879_pad_0, x = input_877_cast_fp16)[name = tensor("input_879_cast_fp16")]; - tensor input_881_pad_type_0 = const()[name = tensor("input_881_pad_type_0"), val = tensor("valid")]; - tensor input_881_groups_0 = const()[name = tensor("input_881_groups_0"), val = tensor(512)]; - tensor input_881_strides_0 = const()[name = tensor("input_881_strides_0"), val = tensor([1])]; - tensor input_881_pad_0 = const()[name = tensor("input_881_pad_0"), val = tensor([0, 0])]; - tensor input_881_dilations_0 = const()[name = tensor("input_881_dilations_0"), val = tensor([1])]; - tensor const_266_to_fp16 = const()[name = tensor("const_266_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210403584)))]; - tensor const_267_to_fp16 = const()[name = tensor("const_267_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210412864)))]; - tensor input_883_cast_fp16 = conv(bias = const_267_to_fp16, dilations = input_881_dilations_0, groups = input_881_groups_0, pad = input_881_pad_0, pad_type = input_881_pad_type_0, strides = input_881_strides_0, weight = const_266_to_fp16, x = input_879_cast_fp16)[name = tensor("input_883_cast_fp16")]; - tensor input_885_cast_fp16 = silu(x = input_883_cast_fp16)[name = tensor("input_885_cast_fp16")]; - tensor x_373_pad_type_0 = const()[name = tensor("x_373_pad_type_0"), val = tensor("valid")]; - tensor x_373_strides_0 = const()[name = tensor("x_373_strides_0"), val = tensor([1])]; - tensor x_373_pad_0 = const()[name = tensor("x_373_pad_0"), val = tensor([0, 0])]; - tensor x_373_dilations_0 = const()[name = tensor("x_373_dilations_0"), val = tensor([1])]; - tensor x_373_groups_0 = const()[name = tensor("x_373_groups_0"), val = tensor(1)]; - tensor module_layers_16_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_16_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210413952)))]; - tensor module_layers_16_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_16_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210938304)))]; - tensor x_373_cast_fp16 = conv(bias = module_layers_16_conv_pointwise_conv2_bias_to_fp16, dilations = x_373_dilations_0, groups = x_373_groups_0, pad = x_373_pad_0, pad_type = x_373_pad_type_0, strides = x_373_strides_0, weight = module_layers_16_conv_pointwise_conv2_weight_to_fp16, x = input_885_cast_fp16)[name = tensor("x_373_cast_fp16")]; - tensor input_887_perm_0 = const()[name = tensor("input_887_perm_0"), val = tensor([0, 2, 1])]; - tensor input_887_cast_fp16 = transpose(perm = input_887_perm_0, x = x_373_cast_fp16)[name = tensor("transpose_86")]; - tensor input_889_cast_fp16 = add(x = input_871_cast_fp16, y = input_887_cast_fp16)[name = tensor("input_889_cast_fp16")]; - tensor input_891_axes_0 = const()[name = tensor("input_891_axes_0"), val = tensor([-1])]; - tensor module_layers_16_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_16_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210939392)))]; - tensor module_layers_16_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_16_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210940480)))]; - tensor input_891_cast_fp16 = layer_norm(axes = input_891_axes_0, beta = module_layers_16_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_16_norm_feed_forward2_weight_to_fp16, x = input_889_cast_fp16)[name = tensor("input_891_cast_fp16")]; - tensor module_layers_16_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_16_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210941568)))]; - tensor module_layers_16_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_16_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(213038784)))]; - tensor linear_152_cast_fp16 = linear(bias = module_layers_16_feed_forward2_linear1_bias_to_fp16, weight = module_layers_16_feed_forward2_linear1_weight_to_fp16, x = input_891_cast_fp16)[name = tensor("linear_152_cast_fp16")]; - tensor input_895_cast_fp16 = silu(x = linear_152_cast_fp16)[name = tensor("input_895_cast_fp16")]; - tensor module_layers_16_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_16_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(213042944)))]; - tensor module_layers_16_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_16_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(215140160)))]; - tensor linear_153_cast_fp16 = linear(bias = module_layers_16_feed_forward2_linear2_bias_to_fp16, weight = module_layers_16_feed_forward2_linear2_weight_to_fp16, x = input_895_cast_fp16)[name = tensor("linear_153_cast_fp16")]; - tensor var_3279_to_fp16 = const()[name = tensor("op_3279_to_fp16"), val = tensor(0x1p-1)]; - tensor var_3280_cast_fp16 = mul(x = linear_153_cast_fp16, y = var_3279_to_fp16)[name = tensor("op_3280_cast_fp16")]; - tensor input_cast_fp16 = add(x = input_889_cast_fp16, y = var_3280_cast_fp16)[name = tensor("input_cast_fp16")]; - tensor audio_signal_axes_0 = const()[name = tensor("audio_signal_axes_0"), val = tensor([-1])]; - tensor module_layers_16_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_16_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(215141248)))]; - tensor module_layers_16_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_16_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(215142336)))]; - tensor audio_signal_cast_fp16 = layer_norm(axes = audio_signal_axes_0, beta = module_layers_16_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_16_norm_out_weight_to_fp16, x = input_cast_fp16)[name = tensor("audio_signal_cast_fp16")]; - tensor obj_1_perm_0 = const()[name = tensor("obj_1_perm_0"), val = tensor([0, 2, 1])]; - tensor obj_1_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("obj_1_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; - tensor obj_1_cast_fp16 = transpose(perm = obj_1_perm_0, x = audio_signal_cast_fp16)[name = tensor("transpose_85")]; - tensor encoder_output = cast(dtype = obj_1_cast_fp16_to_fp32_dtype_0, x = obj_1_cast_fp16)[name = tensor("cast_172")]; - } -> (encoder_output, encoder_length); -} \ No newline at end of file diff --git a/parakeet-tdt-ctc-110m/coreml/compiled_models/Encoder.mlmodelc/weights/weight.bin b/parakeet-tdt-ctc-110m/coreml/compiled_models/Encoder.mlmodelc/weights/weight.bin deleted file mode 100644 index c5bffa9667270bb21c093e55898f4fb2e299e426..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/compiled_models/Encoder.mlmodelc/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:cecf7994b2758397d992802a4f6e5d656e3a1aeb7bbedc2aa430b1316d62474c -size 215143424 diff --git a/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecision.mlmodelc/analytics/coremldata.bin b/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecision.mlmodelc/analytics/coremldata.bin deleted file mode 100644 index 4b4b92e5e8e2b45c67e95237edd74d11f323a65a..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecision.mlmodelc/analytics/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:983ba26dd9276b8d2d4f75f3475aefb1817c542df87dbd0fdac95bd63647494f -size 243 diff --git a/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecision.mlmodelc/coremldata.bin b/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecision.mlmodelc/coremldata.bin deleted file mode 100644 index 74e1d92e032200815d325fa2c7c84ee8562d6aad..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecision.mlmodelc/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:0800e3bdf4ecb1bd46fd27e1826d33125cd574f9ae1e15dd9ff70ea42944ca2d -size 476 diff --git a/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecision.mlmodelc/metadata.json b/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecision.mlmodelc/metadata.json deleted file mode 100644 index fe49b70e7e98f7ec9524130b0027886d862f6f8a..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecision.mlmodelc/metadata.json +++ /dev/null @@ -1,102 +0,0 @@ -[ - { - "metadataOutputVersion" : "3.0", - "shortDescription" : "Parakeet 110M joint + decision head (split, softmax, argmax)", - "outputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1 × 188 × 1)", - "shortDescription" : "", - "shape" : "[1, 188, 1]", - "name" : "token_id", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 188 × 1)", - "shortDescription" : "", - "shape" : "[1, 188, 1]", - "name" : "token_prob", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1 × 188 × 1)", - "shortDescription" : "", - "shape" : "[1, 188, 1]", - "name" : "duration", - "type" : "MultiArray" - } - ], - "storagePrecision" : "Float16", - "modelParameters" : [ - - ], - "author" : "Fluid Inference", - "specificationVersion" : 8, - "mlProgramOperationTypeHistogram" : { - "Ios17.reduceArgmax" : 2, - "Ios17.squeeze" : 1, - "Ios17.cast" : 4, - "Ios17.linear" : 3, - "Ios17.transpose" : 2, - "Ios17.sliceByIndex" : 2, - "Ios17.add" : 1, - "Ios16.relu" : 1, - "Ios16.softmax" : 1, - "Ios17.gatherAlongAxis" : 1, - "Ios17.expandDims" : 3 - }, - "computePrecision" : "Mixed (Float16, Float32, Int16, Int32)", - "isUpdatable" : "0", - "stateSchema" : [ - - ], - "availability" : { - "macOS" : "14.0", - "tvOS" : "17.0", - "visionOS" : "1.0", - "watchOS" : "10.0", - "iOS" : "17.0", - "macCatalyst" : "17.0" - }, - "modelType" : { - "name" : "MLModelType_mlProgram" - }, - "inputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 512 × 188)", - "shortDescription" : "", - "shape" : "[1, 512, 188]", - "name" : "encoder", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 640 × 1)", - "shortDescription" : "", - "shape" : "[1, 640, 1]", - "name" : "decoder", - "type" : "MultiArray" - } - ], - "userDefinedMetadata" : { - "com.github.apple.coremltools.version" : "8.3.0", - "com.github.apple.coremltools.source_dialect" : "TorchScript", - "com.github.apple.coremltools.source" : "torch==2.9.0" - }, - "generatedClassName" : "parakeet_joint_decision", - "method" : "predict" - } -] \ No newline at end of file diff --git a/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecision.mlmodelc/model.mil b/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecision.mlmodelc/model.mil deleted file mode 100644 index edfc5050f47ffbe1f5344799cef5dc0b21c392e0..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecision.mlmodelc/model.mil +++ /dev/null @@ -1,58 +0,0 @@ -program(1.0) -[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3510.2.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.9.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "8.3.0"}})] -{ - func main(tensor decoder, tensor encoder) { - tensor input_1_perm_0 = const()[name = tensor("input_1_perm_0"), val = tensor([0, 2, 1])]; - tensor encoder_to_fp16_dtype_0 = const()[name = tensor("encoder_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor input_3_perm_0 = const()[name = tensor("input_3_perm_0"), val = tensor([0, 2, 1])]; - tensor decoder_to_fp16_dtype_0 = const()[name = tensor("decoder_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor joint_module_enc_weight_to_fp16 = const()[name = tensor("joint_module_enc_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; - tensor joint_module_enc_bias_to_fp16 = const()[name = tensor("joint_module_enc_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(655488)))]; - tensor encoder_to_fp16 = cast(dtype = encoder_to_fp16_dtype_0, x = encoder)[name = tensor("cast_6")]; - tensor input_1_cast_fp16 = transpose(perm = input_1_perm_0, x = encoder_to_fp16)[name = tensor("transpose_1")]; - tensor linear_0_cast_fp16 = linear(bias = joint_module_enc_bias_to_fp16, weight = joint_module_enc_weight_to_fp16, x = input_1_cast_fp16)[name = tensor("linear_0_cast_fp16")]; - tensor joint_module_pred_weight_to_fp16 = const()[name = tensor("joint_module_pred_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(656832)))]; - tensor joint_module_pred_bias_to_fp16 = const()[name = tensor("joint_module_pred_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1476096)))]; - tensor decoder_to_fp16 = cast(dtype = decoder_to_fp16_dtype_0, x = decoder)[name = tensor("cast_5")]; - tensor input_3_cast_fp16 = transpose(perm = input_3_perm_0, x = decoder_to_fp16)[name = tensor("transpose_0")]; - tensor linear_1_cast_fp16 = linear(bias = joint_module_pred_bias_to_fp16, weight = joint_module_pred_weight_to_fp16, x = input_3_cast_fp16)[name = tensor("linear_1_cast_fp16")]; - tensor var_23_axes_0 = const()[name = tensor("op_23_axes_0"), val = tensor([2])]; - tensor var_23_cast_fp16 = expand_dims(axes = var_23_axes_0, x = linear_0_cast_fp16)[name = tensor("op_23_cast_fp16")]; - tensor var_24_axes_0 = const()[name = tensor("op_24_axes_0"), val = tensor([1])]; - tensor var_24_cast_fp16 = expand_dims(axes = var_24_axes_0, x = linear_1_cast_fp16)[name = tensor("op_24_cast_fp16")]; - tensor input_5_cast_fp16 = add(x = var_23_cast_fp16, y = var_24_cast_fp16)[name = tensor("input_5_cast_fp16")]; - tensor input_7_cast_fp16 = relu(x = input_5_cast_fp16)[name = tensor("input_7_cast_fp16")]; - tensor joint_module_joint_net_2_weight_to_fp16 = const()[name = tensor("joint_module_joint_net_2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1477440)))]; - tensor joint_module_joint_net_2_bias_to_fp16 = const()[name = tensor("joint_module_joint_net_2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2795904)))]; - tensor linear_2_cast_fp16 = linear(bias = joint_module_joint_net_2_bias_to_fp16, weight = joint_module_joint_net_2_weight_to_fp16, x = input_7_cast_fp16)[name = tensor("linear_2_cast_fp16")]; - tensor token_logits_begin_0 = const()[name = tensor("token_logits_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor token_logits_end_0 = const()[name = tensor("token_logits_end_0"), val = tensor([1, 188, 1, 1025])]; - tensor token_logits_end_mask_0 = const()[name = tensor("token_logits_end_mask_0"), val = tensor([true, true, true, false])]; - tensor token_logits_cast_fp16 = slice_by_index(begin = token_logits_begin_0, end = token_logits_end_0, end_mask = token_logits_end_mask_0, x = linear_2_cast_fp16)[name = tensor("token_logits_cast_fp16")]; - tensor duration_logits_begin_0 = const()[name = tensor("duration_logits_begin_0"), val = tensor([0, 0, 0, 1025])]; - tensor duration_logits_end_0 = const()[name = tensor("duration_logits_end_0"), val = tensor([1, 188, 1, 1030])]; - tensor duration_logits_end_mask_0 = const()[name = tensor("duration_logits_end_mask_0"), val = tensor([true, true, true, true])]; - tensor duration_logits_cast_fp16 = slice_by_index(begin = duration_logits_begin_0, end = duration_logits_end_0, end_mask = duration_logits_end_mask_0, x = linear_2_cast_fp16)[name = tensor("duration_logits_cast_fp16")]; - tensor var_43_axis_0 = const()[name = tensor("op_43_axis_0"), val = tensor(-1)]; - tensor var_43_keep_dims_0 = const()[name = tensor("op_43_keep_dims_0"), val = tensor(false)]; - tensor var_43_output_dtype_0 = const()[name = tensor("op_43_output_dtype_0"), val = tensor("int32")]; - tensor token_id = reduce_argmax(axis = var_43_axis_0, keep_dims = var_43_keep_dims_0, output_dtype = var_43_output_dtype_0, x = token_logits_cast_fp16)[name = tensor("op_43_cast_fp16")]; - tensor var_49 = const()[name = tensor("op_49"), val = tensor(-1)]; - tensor token_probs_all_cast_fp16 = softmax(axis = var_49, x = token_logits_cast_fp16)[name = tensor("token_probs_all_cast_fp16")]; - tensor var_58_axes_0 = const()[name = tensor("op_58_axes_0"), val = tensor([-1])]; - tensor var_58 = expand_dims(axes = var_58_axes_0, x = token_id)[name = tensor("op_58")]; - tensor var_59 = const()[name = tensor("op_59"), val = tensor(-1)]; - tensor var_61_validate_indices_0 = const()[name = tensor("op_61_validate_indices_0"), val = tensor(false)]; - tensor var_58_to_int16_dtype_0 = const()[name = tensor("op_58_to_int16_dtype_0"), val = tensor("int16")]; - tensor var_58_to_int16 = cast(dtype = var_58_to_int16_dtype_0, x = var_58)[name = tensor("cast_4")]; - tensor var_61_cast_fp16_cast_int16 = gather_along_axis(axis = var_59, indices = var_58_to_int16, validate_indices = var_61_validate_indices_0, x = token_probs_all_cast_fp16)[name = tensor("op_61_cast_fp16_cast_int16")]; - tensor var_63_axes_0 = const()[name = tensor("op_63_axes_0"), val = tensor([-1])]; - tensor var_63_cast_fp16 = squeeze(axes = var_63_axes_0, x = var_61_cast_fp16_cast_int16)[name = tensor("op_63_cast_fp16")]; - tensor var_63_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("op_63_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; - tensor var_66_axis_0 = const()[name = tensor("op_66_axis_0"), val = tensor(-1)]; - tensor var_66_keep_dims_0 = const()[name = tensor("op_66_keep_dims_0"), val = tensor(false)]; - tensor var_66_output_dtype_0 = const()[name = tensor("op_66_output_dtype_0"), val = tensor("int32")]; - tensor duration = reduce_argmax(axis = var_66_axis_0, keep_dims = var_66_keep_dims_0, output_dtype = var_66_output_dtype_0, x = duration_logits_cast_fp16)[name = tensor("op_66_cast_fp16")]; - tensor token_prob = cast(dtype = var_63_cast_fp16_to_fp32_dtype_0, x = var_63_cast_fp16)[name = tensor("cast_3")]; - } -> (token_id, token_prob, duration); -} \ No newline at end of file diff --git a/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecision.mlmodelc/weights/weight.bin b/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecision.mlmodelc/weights/weight.bin deleted file mode 100644 index 747ec9d24e2d3e008bc869d2c404bb3d2653325c..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecision.mlmodelc/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:b3f771cb65b190f1873e39629676ed79b65a8361522f451b37bdba8b1106e6ff -size 2798028 diff --git a/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecisionSingleStep.mlmodelc/analytics/coremldata.bin b/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecisionSingleStep.mlmodelc/analytics/coremldata.bin deleted file mode 100644 index ab3e67b973cf846c0af5e1f075f1b4b7ffb77a00..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecisionSingleStep.mlmodelc/analytics/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:c7c11c6bb985fab7f835ba687a575f1eb04f4c93b0783155d634adbc49f0e797 -size 243 diff --git a/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecisionSingleStep.mlmodelc/coremldata.bin b/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecisionSingleStep.mlmodelc/coremldata.bin deleted file mode 100644 index bc71cc6aa73de6959f4c6718500197ebbf364633..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecisionSingleStep.mlmodelc/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:1af2cb9bcc13eec83ce006e4f1c2cf158393745cd9187428333fbcb6917da244 -size 535 diff --git a/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecisionSingleStep.mlmodelc/metadata.json b/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecisionSingleStep.mlmodelc/metadata.json deleted file mode 100644 index b96db3615aa1d5e110c61c1f4694def827cb0a1f..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecisionSingleStep.mlmodelc/metadata.json +++ /dev/null @@ -1,123 +0,0 @@ -[ - { - "metadataOutputVersion" : "3.0", - "shortDescription" : "Parakeet 110M single-step joint decision (current frame)", - "outputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1 × 1 × 1)", - "shortDescription" : "", - "shape" : "[1, 1, 1]", - "name" : "token_id", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 1 × 1)", - "shortDescription" : "", - "shape" : "[1, 1, 1]", - "name" : "token_prob", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1 × 1 × 1)", - "shortDescription" : "", - "shape" : "[1, 1, 1]", - "name" : "duration", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1 × 1 × 1 × 64)", - "shortDescription" : "", - "shape" : "[1, 1, 1, 64]", - "name" : "top_k_ids", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 1 × 1 × 64)", - "shortDescription" : "", - "shape" : "[1, 1, 1, 64]", - "name" : "top_k_logits", - "type" : "MultiArray" - } - ], - "storagePrecision" : "Float16", - "modelParameters" : [ - - ], - "author" : "Fluid Inference", - "specificationVersion" : 8, - "mlProgramOperationTypeHistogram" : { - "Ios17.reduceArgmax" : 2, - "Ios17.linear" : 3, - "Ios17.transpose" : 2, - "Ios17.sliceByIndex" : 2, - "Ios17.add" : 1, - "Ios17.topk" : 1, - "Ios16.relu" : 1, - "Ios16.softmax" : 1, - "Ios17.expandDims" : 3, - "Ios17.squeeze" : 1, - "Ios17.cast" : 6, - "Ios17.gatherAlongAxis" : 1 - }, - "computePrecision" : "Mixed (Float16, Float32, Int16, Int32, UInt16)", - "isUpdatable" : "0", - "stateSchema" : [ - - ], - "availability" : { - "macOS" : "14.0", - "tvOS" : "17.0", - "visionOS" : "1.0", - "watchOS" : "10.0", - "iOS" : "17.0", - "macCatalyst" : "17.0" - }, - "modelType" : { - "name" : "MLModelType_mlProgram" - }, - "inputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 512 × 1)", - "shortDescription" : "", - "shape" : "[1, 512, 1]", - "name" : "encoder_step", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 640 × 1)", - "shortDescription" : "", - "shape" : "[1, 640, 1]", - "name" : "decoder_step", - "type" : "MultiArray" - } - ], - "userDefinedMetadata" : { - "com.github.apple.coremltools.source_dialect" : "TorchScript", - "com.github.apple.coremltools.source" : "torch==2.9.0", - "com.github.apple.coremltools.version" : "8.3.0" - }, - "generatedClassName" : "parakeet_joint_decision_single_step", - "method" : "predict" - } -] \ No newline at end of file diff --git a/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecisionSingleStep.mlmodelc/model.mil b/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecisionSingleStep.mlmodelc/model.mil deleted file mode 100644 index 42bd4ac5def601f37ece871ff09bd6242de025aa..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecisionSingleStep.mlmodelc/model.mil +++ /dev/null @@ -1,69 +0,0 @@ -program(1.0) -[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3510.2.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.9.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "8.3.0"}})] -{ - func main(tensor decoder_step, tensor encoder_step) { - tensor input_1_perm_0 = const()[name = tensor("input_1_perm_0"), val = tensor([0, 2, 1])]; - tensor encoder_step_to_fp16_dtype_0 = const()[name = tensor("encoder_step_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor input_3_perm_0 = const()[name = tensor("input_3_perm_0"), val = tensor([0, 2, 1])]; - tensor decoder_step_to_fp16_dtype_0 = const()[name = tensor("decoder_step_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor joint_module_enc_weight_to_fp16 = const()[name = tensor("joint_module_enc_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; - tensor joint_module_enc_bias_to_fp16 = const()[name = tensor("joint_module_enc_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(655488)))]; - tensor encoder_step_to_fp16 = cast(dtype = encoder_step_to_fp16_dtype_0, x = encoder_step)[name = tensor("cast_9")]; - tensor input_1_cast_fp16 = transpose(perm = input_1_perm_0, x = encoder_step_to_fp16)[name = tensor("transpose_1")]; - tensor linear_0_cast_fp16 = linear(bias = joint_module_enc_bias_to_fp16, weight = joint_module_enc_weight_to_fp16, x = input_1_cast_fp16)[name = tensor("linear_0_cast_fp16")]; - tensor joint_module_pred_weight_to_fp16 = const()[name = tensor("joint_module_pred_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(656832)))]; - tensor joint_module_pred_bias_to_fp16 = const()[name = tensor("joint_module_pred_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1476096)))]; - tensor decoder_step_to_fp16 = cast(dtype = decoder_step_to_fp16_dtype_0, x = decoder_step)[name = tensor("cast_8")]; - tensor input_3_cast_fp16 = transpose(perm = input_3_perm_0, x = decoder_step_to_fp16)[name = tensor("transpose_0")]; - tensor linear_1_cast_fp16 = linear(bias = joint_module_pred_bias_to_fp16, weight = joint_module_pred_weight_to_fp16, x = input_3_cast_fp16)[name = tensor("linear_1_cast_fp16")]; - tensor var_23_axes_0 = const()[name = tensor("op_23_axes_0"), val = tensor([2])]; - tensor var_23_cast_fp16 = expand_dims(axes = var_23_axes_0, x = linear_0_cast_fp16)[name = tensor("op_23_cast_fp16")]; - tensor var_24_axes_0 = const()[name = tensor("op_24_axes_0"), val = tensor([1])]; - tensor var_24_cast_fp16 = expand_dims(axes = var_24_axes_0, x = linear_1_cast_fp16)[name = tensor("op_24_cast_fp16")]; - tensor input_5_cast_fp16 = add(x = var_23_cast_fp16, y = var_24_cast_fp16)[name = tensor("input_5_cast_fp16")]; - tensor input_7_cast_fp16 = relu(x = input_5_cast_fp16)[name = tensor("input_7_cast_fp16")]; - tensor joint_module_joint_net_2_weight_to_fp16 = const()[name = tensor("joint_module_joint_net_2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1477440)))]; - tensor joint_module_joint_net_2_bias_to_fp16 = const()[name = tensor("joint_module_joint_net_2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2795904)))]; - tensor linear_2_cast_fp16 = linear(bias = joint_module_joint_net_2_bias_to_fp16, weight = joint_module_joint_net_2_weight_to_fp16, x = input_7_cast_fp16)[name = tensor("linear_2_cast_fp16")]; - tensor token_logits_begin_0 = const()[name = tensor("token_logits_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor token_logits_end_0 = const()[name = tensor("token_logits_end_0"), val = tensor([1, 1, 1, 1025])]; - tensor token_logits_end_mask_0 = const()[name = tensor("token_logits_end_mask_0"), val = tensor([true, true, true, false])]; - tensor token_logits_cast_fp16 = slice_by_index(begin = token_logits_begin_0, end = token_logits_end_0, end_mask = token_logits_end_mask_0, x = linear_2_cast_fp16)[name = tensor("token_logits_cast_fp16")]; - tensor duration_logits_begin_0 = const()[name = tensor("duration_logits_begin_0"), val = tensor([0, 0, 0, 1025])]; - tensor duration_logits_end_0 = const()[name = tensor("duration_logits_end_0"), val = tensor([1, 1, 1, 1030])]; - tensor duration_logits_end_mask_0 = const()[name = tensor("duration_logits_end_mask_0"), val = tensor([true, true, true, true])]; - tensor duration_logits_cast_fp16 = slice_by_index(begin = duration_logits_begin_0, end = duration_logits_end_0, end_mask = duration_logits_end_mask_0, x = linear_2_cast_fp16)[name = tensor("duration_logits_cast_fp16")]; - tensor var_43_axis_0 = const()[name = tensor("op_43_axis_0"), val = tensor(-1)]; - tensor var_43_keep_dims_0 = const()[name = tensor("op_43_keep_dims_0"), val = tensor(false)]; - tensor var_43_output_dtype_0 = const()[name = tensor("op_43_output_dtype_0"), val = tensor("int32")]; - tensor token_id = reduce_argmax(axis = var_43_axis_0, keep_dims = var_43_keep_dims_0, output_dtype = var_43_output_dtype_0, x = token_logits_cast_fp16)[name = tensor("op_43_cast_fp16")]; - tensor var_49 = const()[name = tensor("op_49"), val = tensor(-1)]; - tensor token_probs_all_cast_fp16 = softmax(axis = var_49, x = token_logits_cast_fp16)[name = tensor("token_probs_all_cast_fp16")]; - tensor var_58_axes_0 = const()[name = tensor("op_58_axes_0"), val = tensor([-1])]; - tensor var_58 = expand_dims(axes = var_58_axes_0, x = token_id)[name = tensor("op_58")]; - tensor var_59 = const()[name = tensor("op_59"), val = tensor(-1)]; - tensor var_61_validate_indices_0 = const()[name = tensor("op_61_validate_indices_0"), val = tensor(false)]; - tensor var_58_to_int16_dtype_0 = const()[name = tensor("op_58_to_int16_dtype_0"), val = tensor("int16")]; - tensor var_58_to_int16 = cast(dtype = var_58_to_int16_dtype_0, x = var_58)[name = tensor("cast_7")]; - tensor var_61_cast_fp16_cast_int16 = gather_along_axis(axis = var_59, indices = var_58_to_int16, validate_indices = var_61_validate_indices_0, x = token_probs_all_cast_fp16)[name = tensor("op_61_cast_fp16_cast_int16")]; - tensor var_63_axes_0 = const()[name = tensor("op_63_axes_0"), val = tensor([-1])]; - tensor var_63_cast_fp16 = squeeze(axes = var_63_axes_0, x = var_61_cast_fp16_cast_int16)[name = tensor("op_63_cast_fp16")]; - tensor var_63_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("op_63_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; - tensor var_66_axis_0 = const()[name = tensor("op_66_axis_0"), val = tensor(-1)]; - tensor var_66_keep_dims_0 = const()[name = tensor("op_66_keep_dims_0"), val = tensor(false)]; - tensor var_66_output_dtype_0 = const()[name = tensor("op_66_output_dtype_0"), val = tensor("int32")]; - tensor duration = reduce_argmax(axis = var_66_axis_0, keep_dims = var_66_keep_dims_0, output_dtype = var_66_output_dtype_0, x = duration_logits_cast_fp16)[name = tensor("op_66_cast_fp16")]; - tensor var_72 = const()[name = tensor("op_72"), val = tensor(64)]; - tensor var_76_axis_0 = const()[name = tensor("op_76_axis_0"), val = tensor(-1)]; - tensor var_76_ascending_0 = const()[name = tensor("op_76_ascending_0"), val = tensor(false)]; - tensor var_76_sort_0 = const()[name = tensor("op_76_sort_0"), val = tensor(true)]; - tensor var_76_return_indices_0 = const()[name = tensor("op_76_return_indices_0"), val = tensor(true)]; - tensor var_76_cast_fp16_cast_int16_output_indices_dtype_0 = const()[name = tensor("op_76_cast_fp16_cast_int16_output_indices_dtype_0"), val = tensor("uint16")]; - tensor var_76_cast_fp16_cast_int16_0, tensor var_76_cast_fp16_cast_int16_1 = topk(ascending = var_76_ascending_0, axis = var_76_axis_0, k = var_72, output_indices_dtype = var_76_cast_fp16_cast_int16_output_indices_dtype_0, return_indices = var_76_return_indices_0, sort = var_76_sort_0, x = token_logits_cast_fp16)[name = tensor("op_76_cast_fp16_cast_int16")]; - tensor var_76_cast_fp16_cast_int16_1_to_int32_dtype_0 = const()[name = tensor("op_76_cast_fp16_cast_int16_1_to_int32_dtype_0"), val = tensor("int32")]; - tensor var_76_cast_fp16_0_to_fp32_dtype_0 = const()[name = tensor("op_76_cast_fp16_0_to_fp32_dtype_0"), val = tensor("fp32")]; - tensor top_k_logits = cast(dtype = var_76_cast_fp16_0_to_fp32_dtype_0, x = var_76_cast_fp16_cast_int16_0)[name = tensor("cast_4")]; - tensor top_k_ids = cast(dtype = var_76_cast_fp16_cast_int16_1_to_int32_dtype_0, x = var_76_cast_fp16_cast_int16_1)[name = tensor("cast_5")]; - tensor token_prob = cast(dtype = var_63_cast_fp16_to_fp32_dtype_0, x = var_63_cast_fp16)[name = tensor("cast_6")]; - } -> (token_id, token_prob, duration, top_k_ids, top_k_logits); -} \ No newline at end of file diff --git a/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecisionSingleStep.mlmodelc/weights/weight.bin b/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecisionSingleStep.mlmodelc/weights/weight.bin deleted file mode 100644 index 747ec9d24e2d3e008bc869d2c404bb3d2653325c..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/compiled_models/JointDecisionSingleStep.mlmodelc/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:b3f771cb65b190f1873e39629676ed79b65a8361522f451b37bdba8b1106e6ff -size 2798028 diff --git a/parakeet-tdt-ctc-110m/coreml/compiled_models/Preprocessor.mlmodelc/analytics/coremldata.bin b/parakeet-tdt-ctc-110m/coreml/compiled_models/Preprocessor.mlmodelc/analytics/coremldata.bin deleted file mode 100644 index d34457003aea6648633a14e7cd6134edc9c30f17..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/compiled_models/Preprocessor.mlmodelc/analytics/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:a1ac15543fbb9301fba5f018b147e44d767479dec352aaa91dfe7bcf65949693 -size 243 diff --git a/parakeet-tdt-ctc-110m/coreml/compiled_models/Preprocessor.mlmodelc/coremldata.bin b/parakeet-tdt-ctc-110m/coreml/compiled_models/Preprocessor.mlmodelc/coremldata.bin deleted file mode 100644 index cd81f47cc6672a441cbe9556757a30ffc0ff0bc7..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/compiled_models/Preprocessor.mlmodelc/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:4940877938cc1b6d8830bbdd68ac8a49377cc57d75b61308883da5235b6a1914 -size 439 diff --git a/parakeet-tdt-ctc-110m/coreml/compiled_models/Preprocessor.mlmodelc/metadata.json b/parakeet-tdt-ctc-110m/coreml/compiled_models/Preprocessor.mlmodelc/metadata.json deleted file mode 100644 index 6087d5e76f7dd060cf4f135f1f3ae0bbbeb2f30a..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/compiled_models/Preprocessor.mlmodelc/metadata.json +++ /dev/null @@ -1,112 +0,0 @@ -[ - { - "metadataOutputVersion" : "3.0", - "shortDescription" : "Parakeet 110M preprocessor (15 s window)", - "outputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32)", - "shortDescription" : "", - "shape" : "[]", - "name" : "mel_features", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1)", - "shortDescription" : "", - "shape" : "[1]", - "name" : "mel_length", - "type" : "MultiArray" - } - ], - "storagePrecision" : "Float16", - "modelParameters" : [ - - ], - "author" : "Fluid Inference", - "specificationVersion" : 8, - "mlProgramOperationTypeHistogram" : { - "Range1d" : 3, - "Ios17.equal" : 1, - "Ios17.notEqual" : 1, - "Ios17.reshape" : 2, - "Identity" : 1, - "Ios17.matmul" : 1, - "Select" : 6, - "Ios17.expandDims" : 12, - "Ios17.add" : 3, - "Tile" : 2, - "Ios17.sliceByIndex" : 3, - "Ios16.reduceSum" : 4, - "Shape" : 4, - "Ios17.gather" : 4, - "Ios17.logicalNot" : 1, - "Pad" : 1, - "Ios17.log" : 1, - "Ios17.less" : 2, - "Ios17.sub" : 4, - "Ios17.conv" : 2, - "Ios17.pow" : 2, - "Ios17.cast" : 10, - "Ios17.concat" : 3, - "Stack" : 1, - "Ios17.floorDiv" : 1, - "Ios17.realDiv" : 4, - "Ios17.sqrt" : 1, - "Ios17.greaterEqual" : 1, - "Ios17.mul" : 1 - }, - "computePrecision" : "Mixed (Float16, Float32, Int16, Int32, UInt16)", - "isUpdatable" : "0", - "stateSchema" : [ - - ], - "availability" : { - "macOS" : "14.0", - "tvOS" : "17.0", - "visionOS" : "1.0", - "watchOS" : "10.0", - "iOS" : "17.0", - "macCatalyst" : "17.0" - }, - "modelType" : { - "name" : "MLModelType_mlProgram" - }, - "inputSchema" : [ - { - "dataType" : "Float32", - "hasShapeFlexibility" : "1", - "isOptional" : "0", - "shapeFlexibility" : "1 × 1...240000", - "shapeRange" : "[[1, 1], [1, 240000]]", - "formattedType" : "MultiArray (Float32 1 × 1)", - "type" : "MultiArray", - "shape" : "[1, 1]", - "name" : "audio", - "shortDescription" : "" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1)", - "shortDescription" : "", - "shape" : "[1]", - "name" : "audio_length", - "type" : "MultiArray" - } - ], - "userDefinedMetadata" : { - "com.github.apple.coremltools.source_dialect" : "TorchScript", - "com.github.apple.coremltools.source" : "torch==2.9.0", - "com.github.apple.coremltools.version" : "8.3.0" - }, - "generatedClassName" : "parakeet_preprocessor", - "method" : "predict" - } -] \ No newline at end of file diff --git a/parakeet-tdt-ctc-110m/coreml/compiled_models/Preprocessor.mlmodelc/model.mil b/parakeet-tdt-ctc-110m/coreml/compiled_models/Preprocessor.mlmodelc/model.mil deleted file mode 100644 index ae325b6071a8ee5dad74a2e516e60889d89771d8..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/compiled_models/Preprocessor.mlmodelc/model.mil +++ /dev/null @@ -1,191 +0,0 @@ -program(1.0) -[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3510.2.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.9.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "8.3.0"}})] -{ - func main(tensor audio, tensor audio_length) [FlexibleShapeInformation = tuple, dict, tensor>>, tuple, dict, list, ?>>>>((("DefaultShapes", {{"audio", [1, 1]}}), ("RangeDims", {{"audio", [[1, 1], [1, 240000]]}})))] { - tensor var_9 = const()[name = tensor("op_9"), val = tensor(1)]; - tensor var_10 = const()[name = tensor("op_10"), val = tensor(160)]; - tensor var_12 = const()[name = tensor("op_12"), val = tensor(0)]; - tensor var_34 = const()[name = tensor("op_34"), val = tensor(512)]; - tensor var_35 = add(x = audio_length, y = var_34)[name = tensor("op_35")]; - tensor var_36 = const()[name = tensor("op_36"), val = tensor(512)]; - tensor var_37 = sub(x = var_35, y = var_36)[name = tensor("op_37")]; - tensor floor_div_0 = floor_div(x = var_37, y = var_10)[name = tensor("floor_div_0")]; - tensor var_40 = equal(x = audio_length, y = var_12)[name = tensor("op_40")]; - tensor var_41 = const()[name = tensor("op_41"), val = tensor([0])]; - tensor mel_length = select(a = var_41, b = floor_div_0, cond = var_40)[name = tensor("seq_len")]; - tensor audio_to_fp16_dtype_0 = const()[name = tensor("audio_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor audio_to_fp16 = cast(dtype = audio_to_fp16_dtype_0, x = audio)[name = tensor("cast_27")]; - tensor var_43_shape_cast_fp16 = shape(x = audio_to_fp16)[name = tensor("op_43_shape_cast_fp16")]; - tensor gather_0_axis_0 = const()[name = tensor("gather_0_axis_0"), val = tensor(0)]; - tensor gather_0_batch_dims_0 = const()[name = tensor("gather_0_batch_dims_0"), val = tensor(0)]; - tensor gather_0_validate_indices_0 = const()[name = tensor("gather_0_validate_indices_0"), val = tensor(false)]; - tensor var_43_shape_cast_fp16_to_int16_dtype_0 = const()[name = tensor("op_43_shape_cast_fp16_to_int16_dtype_0"), val = tensor("int16")]; - tensor select_0_to_uint16 = const()[name = tensor("select_0_to_uint16"), val = tensor(1)]; - tensor var_43_shape_cast_fp16_to_int16 = cast(dtype = var_43_shape_cast_fp16_to_int16_dtype_0, x = var_43_shape_cast_fp16)[name = tensor("cast_26")]; - tensor gather_0_cast_uint16 = gather(axis = gather_0_axis_0, batch_dims = gather_0_batch_dims_0, indices = select_0_to_uint16, validate_indices = gather_0_validate_indices_0, x = var_43_shape_cast_fp16_to_int16)[name = tensor("gather_0_cast_uint16")]; - tensor gather_0_cast_uint16_to_int32_dtype_0 = const()[name = tensor("gather_0_cast_uint16_to_int32_dtype_0"), val = tensor("int32")]; - tensor const_0 = const()[name = tensor("const_0"), val = tensor(0)]; - tensor const_1 = const()[name = tensor("const_1"), val = tensor(1)]; - tensor gather_0_cast_uint16_to_int32 = cast(dtype = gather_0_cast_uint16_to_int32_dtype_0, x = gather_0_cast_uint16)[name = tensor("cast_25")]; - tensor var_44 = range_1d(end = gather_0_cast_uint16_to_int32, start = const_0, step = const_1)[name = tensor("op_44")]; - tensor var_45_axes_0 = const()[name = tensor("op_45_axes_0"), val = tensor([0])]; - tensor var_45 = expand_dims(axes = var_45_axes_0, x = var_44)[name = tensor("op_45")]; - tensor var_46_axes_0 = const()[name = tensor("op_46_axes_0"), val = tensor([1])]; - tensor var_46 = expand_dims(axes = var_46_axes_0, x = audio_length)[name = tensor("op_46")]; - tensor timemask = less(x = var_45, y = var_46)[name = tensor("timemask")]; - tensor var_49_begin_0 = const()[name = tensor("op_49_begin_0"), val = tensor([0, 0])]; - tensor var_49_end_0 = const()[name = tensor("op_49_end_0"), val = tensor([1, 1])]; - tensor var_49_end_mask_0 = const()[name = tensor("op_49_end_mask_0"), val = tensor([true, false])]; - tensor var_49_squeeze_mask_0 = const()[name = tensor("op_49_squeeze_mask_0"), val = tensor([false, true])]; - tensor var_49_cast_fp16 = slice_by_index(begin = var_49_begin_0, end = var_49_end_0, end_mask = var_49_end_mask_0, squeeze_mask = var_49_squeeze_mask_0, x = audio_to_fp16)[name = tensor("op_49_cast_fp16")]; - tensor var_50_axes_0 = const()[name = tensor("op_50_axes_0"), val = tensor([1])]; - tensor var_50_cast_fp16 = expand_dims(axes = var_50_axes_0, x = var_49_cast_fp16)[name = tensor("op_50_cast_fp16")]; - tensor var_52_begin_0 = const()[name = tensor("op_52_begin_0"), val = tensor([0, 1])]; - tensor var_52_end_0 = const()[name = tensor("op_52_end_0"), val = tensor([1, 0])]; - tensor var_52_end_mask_0 = const()[name = tensor("op_52_end_mask_0"), val = tensor([true, true])]; - tensor var_52_cast_fp16 = slice_by_index(begin = var_52_begin_0, end = var_52_end_0, end_mask = var_52_end_mask_0, x = audio_to_fp16)[name = tensor("op_52_cast_fp16")]; - tensor var_54_begin_0 = const()[name = tensor("op_54_begin_0"), val = tensor([0, 0])]; - tensor var_54_end_0 = const()[name = tensor("op_54_end_0"), val = tensor([1, -1])]; - tensor var_54_end_mask_0 = const()[name = tensor("op_54_end_mask_0"), val = tensor([true, false])]; - tensor var_54_cast_fp16 = slice_by_index(begin = var_54_begin_0, end = var_54_end_0, end_mask = var_54_end_mask_0, x = audio_to_fp16)[name = tensor("op_54_cast_fp16")]; - tensor var_55_to_fp16 = const()[name = tensor("op_55_to_fp16"), val = tensor(0x1.f0cp-1)]; - tensor var_56_cast_fp16 = mul(x = var_54_cast_fp16, y = var_55_to_fp16)[name = tensor("op_56_cast_fp16")]; - tensor var_57_cast_fp16 = sub(x = var_52_cast_fp16, y = var_56_cast_fp16)[name = tensor("op_57_cast_fp16")]; - tensor x_3_interleave_0 = const()[name = tensor("x_3_interleave_0"), val = tensor(false)]; - tensor x_3_cast_fp16 = concat(axis = var_9, interleave = x_3_interleave_0, values = (var_50_cast_fp16, var_57_cast_fp16))[name = tensor("x_3_cast_fp16")]; - tensor var_60 = logical_not(x = timemask)[name = tensor("op_60")]; - tensor var_16_to_fp16 = const()[name = tensor("op_16_to_fp16"), val = tensor(0x0p+0)]; - tensor input_1_cast_fp16 = select(a = var_16_to_fp16, b = x_3_cast_fp16, cond = var_60)[name = tensor("input_1_cast_fp16")]; - tensor concat_1x = const()[name = tensor("concat_1x"), val = tensor([1, 1, -1])]; - tensor input_3_cast_fp16 = reshape(shape = concat_1x, x = input_1_cast_fp16)[name = tensor("input_3_cast_fp16")]; - tensor input_5_pad_0 = const()[name = tensor("input_5_pad_0"), val = tensor([0, 0, 0, 0, 256, 256])]; - tensor input_5_mode_0 = const()[name = tensor("input_5_mode_0"), val = tensor("constant")]; - tensor const_3_to_fp16 = const()[name = tensor("const_3_to_fp16"), val = tensor(0x0p+0)]; - tensor input_5_cast_fp16 = pad(constant_val = const_3_to_fp16, mode = input_5_mode_0, pad = input_5_pad_0, x = input_3_cast_fp16)[name = tensor("input_5_cast_fp16")]; - tensor concat_2x = const()[name = tensor("concat_2x"), val = tensor([1, -1])]; - tensor input_cast_fp16 = reshape(shape = concat_2x, x = input_5_cast_fp16)[name = tensor("input_cast_fp16")]; - tensor expand_dims_3 = const()[name = tensor("expand_dims_3"), val = tensor([160])]; - tensor expand_dims_4_axes_0 = const()[name = tensor("expand_dims_4_axes_0"), val = tensor([1])]; - tensor expand_dims_4_cast_fp16 = expand_dims(axes = expand_dims_4_axes_0, x = input_cast_fp16)[name = tensor("expand_dims_4_cast_fp16")]; - tensor conv_0_pad_type_0 = const()[name = tensor("conv_0_pad_type_0"), val = tensor("valid")]; - tensor conv_0_pad_0 = const()[name = tensor("conv_0_pad_0"), val = tensor([0, 0])]; - tensor conv_0_dilations_0 = const()[name = tensor("conv_0_dilations_0"), val = tensor([1])]; - tensor conv_0_groups_0 = const()[name = tensor("conv_0_groups_0"), val = tensor(1)]; - tensor expand_dims_1_to_fp16 = const()[name = tensor("expand_dims_1_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; - tensor conv_0_cast_fp16 = conv(dilations = conv_0_dilations_0, groups = conv_0_groups_0, pad = conv_0_pad_0, pad_type = conv_0_pad_type_0, strides = expand_dims_3, weight = expand_dims_1_to_fp16, x = expand_dims_4_cast_fp16)[name = tensor("conv_0_cast_fp16")]; - tensor conv_1_pad_type_0 = const()[name = tensor("conv_1_pad_type_0"), val = tensor("valid")]; - tensor conv_1_pad_0 = const()[name = tensor("conv_1_pad_0"), val = tensor([0, 0])]; - tensor conv_1_dilations_0 = const()[name = tensor("conv_1_dilations_0"), val = tensor([1])]; - tensor conv_1_groups_0 = const()[name = tensor("conv_1_groups_0"), val = tensor(1)]; - tensor expand_dims_2_to_fp16 = const()[name = tensor("expand_dims_2_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(263296)))]; - tensor conv_1_cast_fp16 = conv(dilations = conv_1_dilations_0, groups = conv_1_groups_0, pad = conv_1_pad_0, pad_type = conv_1_pad_type_0, strides = expand_dims_3, weight = expand_dims_2_to_fp16, x = expand_dims_4_cast_fp16)[name = tensor("conv_1_cast_fp16")]; - tensor stack_0_axis_0 = const()[name = tensor("stack_0_axis_0"), val = tensor(-1)]; - tensor stack_0_cast_fp16 = stack(axis = stack_0_axis_0, values = (conv_0_cast_fp16, conv_1_cast_fp16))[name = tensor("stack_0_cast_fp16")]; - tensor var_19_promoted_to_fp16 = const()[name = tensor("op_19_promoted_to_fp16"), val = tensor(0x1p+1)]; - tensor var_75_cast_fp16 = pow(x = stack_0_cast_fp16, y = var_19_promoted_to_fp16)[name = tensor("op_75_cast_fp16")]; - tensor var_77_axes_0 = const()[name = tensor("op_77_axes_0"), val = tensor([-1])]; - tensor var_77_keep_dims_0 = const()[name = tensor("op_77_keep_dims_0"), val = tensor(false)]; - tensor var_77_cast_fp16 = reduce_sum(axes = var_77_axes_0, keep_dims = var_77_keep_dims_0, x = var_75_cast_fp16)[name = tensor("op_77_cast_fp16")]; - tensor x_11_cast_fp16 = identity(x = var_77_cast_fp16)[name = tensor("x_11_cast_fp16")]; - tensor x_13_transpose_x_0 = const()[name = tensor("x_13_transpose_x_0"), val = tensor(false)]; - tensor x_13_transpose_y_0 = const()[name = tensor("x_13_transpose_y_0"), val = tensor(false)]; - tensor const_4_to_fp16 = const()[name = tensor("const_4_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(526528)))]; - tensor x_13_cast_fp16 = matmul(transpose_x = x_13_transpose_x_0, transpose_y = x_13_transpose_y_0, x = const_4_to_fp16, y = x_11_cast_fp16)[name = tensor("x_13_cast_fp16")]; - tensor var_84_to_fp16 = const()[name = tensor("op_84_to_fp16"), val = tensor(0x1p-24)]; - tensor var_85_cast_fp16 = add(x = x_13_cast_fp16, y = var_84_to_fp16)[name = tensor("op_85_cast_fp16")]; - tensor x_15_epsilon_0 = const()[name = tensor("x_15_epsilon_0"), val = tensor(0x1p-149)]; - tensor x_15_cast_fp16 = log(epsilon = x_15_epsilon_0, x = var_85_cast_fp16)[name = tensor("x_15_cast_fp16")]; - tensor var_87_shape_cast_fp16 = shape(x = x_15_cast_fp16)[name = tensor("op_87_shape_cast_fp16")]; - tensor gather_5 = const()[name = tensor("gather_5"), val = tensor(1)]; - tensor gather_6_axis_0 = const()[name = tensor("gather_6_axis_0"), val = tensor(0)]; - tensor gather_6_batch_dims_0 = const()[name = tensor("gather_6_batch_dims_0"), val = tensor(0)]; - tensor gather_6_validate_indices_0 = const()[name = tensor("gather_6_validate_indices_0"), val = tensor(false)]; - tensor var_87_shape_cast_fp16_to_uint16_dtype_0 = const()[name = tensor("op_87_shape_cast_fp16_to_uint16_dtype_0"), val = tensor("uint16")]; - tensor select_6_to_uint16 = const()[name = tensor("select_6_to_uint16"), val = tensor(2)]; - tensor var_87_shape_cast_fp16_to_uint16 = cast(dtype = var_87_shape_cast_fp16_to_uint16_dtype_0, x = var_87_shape_cast_fp16)[name = tensor("cast_24")]; - tensor gather_6_cast_uint16 = gather(axis = gather_6_axis_0, batch_dims = gather_6_batch_dims_0, indices = select_6_to_uint16, validate_indices = gather_6_validate_indices_0, x = var_87_shape_cast_fp16_to_uint16)[name = tensor("gather_6_cast_uint16")]; - tensor gather_6_cast_uint16_to_int32_dtype_0 = const()[name = tensor("gather_6_cast_uint16_to_int32_dtype_0"), val = tensor("int32")]; - tensor const_5 = const()[name = tensor("const_5"), val = tensor(0)]; - tensor const_6 = const()[name = tensor("const_6"), val = tensor(1)]; - tensor gather_6_cast_uint16_to_int32 = cast(dtype = gather_6_cast_uint16_to_int32_dtype_0, x = gather_6_cast_uint16)[name = tensor("cast_23")]; - tensor var_89 = range_1d(end = gather_6_cast_uint16_to_int32, start = const_5, step = const_6)[name = tensor("op_89")]; - tensor var_90_axes_0 = const()[name = tensor("op_90_axes_0"), val = tensor([0])]; - tensor var_90 = expand_dims(axes = var_90_axes_0, x = var_89)[name = tensor("op_90")]; - tensor concat_3_axis_0 = const()[name = tensor("concat_3_axis_0"), val = tensor(0)]; - tensor concat_3_interleave_0 = const()[name = tensor("concat_3_interleave_0"), val = tensor(false)]; - tensor concat_3 = concat(axis = concat_3_axis_0, interleave = concat_3_interleave_0, values = (gather_5, gather_6_cast_uint16_to_int32))[name = tensor("concat_3")]; - tensor shape_8 = shape(x = var_90)[name = tensor("shape_8")]; - tensor real_div_0 = real_div(x = concat_3, y = shape_8)[name = tensor("real_div_0")]; - tensor time_steps = tile(reps = real_div_0, x = var_90)[name = tensor("time_steps")]; - tensor var_93_axes_0 = const()[name = tensor("op_93_axes_0"), val = tensor([1])]; - tensor var_93 = expand_dims(axes = var_93_axes_0, x = mel_length)[name = tensor("op_93")]; - tensor valid_mask = less(x = time_steps, y = var_93)[name = tensor("valid_mask")]; - tensor var_95_axes_0 = const()[name = tensor("op_95_axes_0"), val = tensor([1])]; - tensor var_95 = expand_dims(axes = var_95_axes_0, x = valid_mask)[name = tensor("op_95")]; - tensor var_96_cast_fp16 = select(a = x_15_cast_fp16, b = var_16_to_fp16, cond = var_95)[name = tensor("op_96_cast_fp16")]; - tensor x_mean_numerator_axes_0 = const()[name = tensor("x_mean_numerator_axes_0"), val = tensor([2])]; - tensor x_mean_numerator_keep_dims_0 = const()[name = tensor("x_mean_numerator_keep_dims_0"), val = tensor(false)]; - tensor x_mean_numerator_cast_fp16 = reduce_sum(axes = x_mean_numerator_axes_0, keep_dims = x_mean_numerator_keep_dims_0, x = var_96_cast_fp16)[name = tensor("x_mean_numerator_cast_fp16")]; - tensor x_mean_denominator_axes_0 = const()[name = tensor("x_mean_denominator_axes_0"), val = tensor([1])]; - tensor x_mean_denominator_keep_dims_0 = const()[name = tensor("x_mean_denominator_keep_dims_0"), val = tensor(false)]; - tensor cast_6_to_fp16_dtype_0 = const()[name = tensor("cast_6_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor valid_mask_to_fp16 = cast(dtype = cast_6_to_fp16_dtype_0, x = valid_mask)[name = tensor("cast_22")]; - tensor x_mean_denominator_cast_fp16 = reduce_sum(axes = x_mean_denominator_axes_0, keep_dims = x_mean_denominator_keep_dims_0, x = valid_mask_to_fp16)[name = tensor("x_mean_denominator_cast_fp16")]; - tensor var_101_axes_0 = const()[name = tensor("op_101_axes_0"), val = tensor([1])]; - tensor var_101_cast_fp16 = expand_dims(axes = var_101_axes_0, x = x_mean_denominator_cast_fp16)[name = tensor("op_101_cast_fp16")]; - tensor x_mean_cast_fp16 = real_div(x = x_mean_numerator_cast_fp16, y = var_101_cast_fp16)[name = tensor("x_mean_cast_fp16")]; - tensor var_104_axes_0 = const()[name = tensor("op_104_axes_0"), val = tensor([2])]; - tensor var_104_cast_fp16 = expand_dims(axes = var_104_axes_0, x = x_mean_cast_fp16)[name = tensor("op_104_cast_fp16")]; - tensor var_105_cast_fp16 = sub(x = x_15_cast_fp16, y = var_104_cast_fp16)[name = tensor("op_105_cast_fp16")]; - tensor var_106_cast_fp16 = select(a = var_105_cast_fp16, b = var_16_to_fp16, cond = var_95)[name = tensor("op_106_cast_fp16")]; - tensor var_19_promoted_1_to_fp16 = const()[name = tensor("op_19_promoted_1_to_fp16"), val = tensor(0x1p+1)]; - tensor var_107_cast_fp16 = pow(x = var_106_cast_fp16, y = var_19_promoted_1_to_fp16)[name = tensor("op_107_cast_fp16")]; - tensor var_109_axes_0 = const()[name = tensor("op_109_axes_0"), val = tensor([2])]; - tensor var_109_keep_dims_0 = const()[name = tensor("op_109_keep_dims_0"), val = tensor(false)]; - tensor var_109_cast_fp16 = reduce_sum(axes = var_109_axes_0, keep_dims = var_109_keep_dims_0, x = var_107_cast_fp16)[name = tensor("op_109_cast_fp16")]; - tensor var_111_to_fp16 = const()[name = tensor("op_111_to_fp16"), val = tensor(0x1p+0)]; - tensor var_112_cast_fp16 = sub(x = var_101_cast_fp16, y = var_111_to_fp16)[name = tensor("op_112_cast_fp16")]; - tensor var_113_cast_fp16 = real_div(x = var_109_cast_fp16, y = var_112_cast_fp16)[name = tensor("op_113_cast_fp16")]; - tensor x_std_1_cast_fp16 = sqrt(x = var_113_cast_fp16)[name = tensor("x_std_1_cast_fp16")]; - tensor var_115_cast_fp16 = not_equal(x = x_std_1_cast_fp16, y = x_std_1_cast_fp16)[name = tensor("op_115_cast_fp16")]; - tensor x_std_3_cast_fp16 = select(a = var_16_to_fp16, b = x_std_1_cast_fp16, cond = var_115_cast_fp16)[name = tensor("x_std_3_cast_fp16")]; - tensor var_25_to_fp16 = const()[name = tensor("op_25_to_fp16"), val = tensor(0x1.5p-17)]; - tensor x_std_cast_fp16 = add(x = x_std_3_cast_fp16, y = var_25_to_fp16)[name = tensor("x_std_cast_fp16")]; - tensor var_120_axes_0 = const()[name = tensor("op_120_axes_0"), val = tensor([2])]; - tensor var_120_cast_fp16 = expand_dims(axes = var_120_axes_0, x = x_std_cast_fp16)[name = tensor("op_120_cast_fp16")]; - tensor x_cast_fp16 = real_div(x = var_105_cast_fp16, y = var_120_cast_fp16)[name = tensor("x_cast_fp16")]; - tensor var_122_shape_cast_fp16 = shape(x = x_cast_fp16)[name = tensor("op_122_shape_cast_fp16")]; - tensor gather_7_axis_0 = const()[name = tensor("gather_7_axis_0"), val = tensor(0)]; - tensor gather_7_batch_dims_0 = const()[name = tensor("gather_7_batch_dims_0"), val = tensor(0)]; - tensor gather_7_validate_indices_0 = const()[name = tensor("gather_7_validate_indices_0"), val = tensor(false)]; - tensor var_122_shape_cast_fp16_to_uint16_dtype_0 = const()[name = tensor("op_122_shape_cast_fp16_to_uint16_dtype_0"), val = tensor("uint16")]; - tensor select_7_to_uint16 = const()[name = tensor("select_7_to_uint16"), val = tensor(2)]; - tensor var_122_shape_cast_fp16_to_uint16 = cast(dtype = var_122_shape_cast_fp16_to_uint16_dtype_0, x = var_122_shape_cast_fp16)[name = tensor("cast_21")]; - tensor gather_7_cast_uint16 = gather(axis = gather_7_axis_0, batch_dims = gather_7_batch_dims_0, indices = select_7_to_uint16, validate_indices = gather_7_validate_indices_0, x = var_122_shape_cast_fp16_to_uint16)[name = tensor("gather_7_cast_uint16")]; - tensor gather_7_cast_uint16_to_int32_dtype_0 = const()[name = tensor("gather_7_cast_uint16_to_int32_dtype_0"), val = tensor("int32")]; - tensor const_7 = const()[name = tensor("const_7"), val = tensor(0)]; - tensor const_8 = const()[name = tensor("const_8"), val = tensor(1)]; - tensor gather_7_cast_uint16_to_int32 = cast(dtype = gather_7_cast_uint16_to_int32_dtype_0, x = gather_7_cast_uint16)[name = tensor("cast_20")]; - tensor mask_1 = range_1d(end = gather_7_cast_uint16_to_int32, start = const_7, step = const_8)[name = tensor("mask_1")]; - tensor gather_8_axis_0 = const()[name = tensor("gather_8_axis_0"), val = tensor(0)]; - tensor gather_8_batch_dims_0 = const()[name = tensor("gather_8_batch_dims_0"), val = tensor(0)]; - tensor gather_8_validate_indices_0 = const()[name = tensor("gather_8_validate_indices_0"), val = tensor(false)]; - tensor select_8_to_uint16 = const()[name = tensor("select_8_to_uint16"), val = tensor(0)]; - tensor gather_8_cast_uint16 = gather(axis = gather_8_axis_0, batch_dims = gather_8_batch_dims_0, indices = select_8_to_uint16, validate_indices = gather_8_validate_indices_0, x = var_122_shape_cast_fp16_to_uint16)[name = tensor("gather_8_cast_uint16")]; - tensor gather_8_cast_uint16_to_int32_dtype_0 = const()[name = tensor("gather_8_cast_uint16_to_int32_dtype_0"), val = tensor("int32")]; - tensor concat_4_axis_0 = const()[name = tensor("concat_4_axis_0"), val = tensor(0)]; - tensor concat_4_interleave_0 = const()[name = tensor("concat_4_interleave_0"), val = tensor(false)]; - tensor gather_8_cast_uint16_to_int32 = cast(dtype = gather_8_cast_uint16_to_int32_dtype_0, x = gather_8_cast_uint16)[name = tensor("cast_19")]; - tensor concat_4 = concat(axis = concat_4_axis_0, interleave = concat_4_interleave_0, values = (gather_8_cast_uint16_to_int32, var_9))[name = tensor("concat_4")]; - tensor expand_dims_0_axes_0 = const()[name = tensor("expand_dims_0_axes_0"), val = tensor([0])]; - tensor expand_dims_0 = expand_dims(axes = expand_dims_0_axes_0, x = mask_1)[name = tensor("expand_dims_0")]; - tensor var_126 = tile(reps = concat_4, x = expand_dims_0)[name = tensor("op_126")]; - tensor mask = greater_equal(x = var_126, y = var_93)[name = tensor("mask")]; - tensor var_129_axes_0 = const()[name = tensor("op_129_axes_0"), val = tensor([1])]; - tensor var_129 = expand_dims(axes = var_129_axes_0, x = mask)[name = tensor("op_129")]; - tensor cast_15_to_fp16 = const()[name = tensor("cast_15_to_fp16"), val = tensor(0x0p+0)]; - tensor processed_signal_cast_fp16 = select(a = cast_15_to_fp16, b = x_cast_fp16, cond = var_129)[name = tensor("processed_signal_cast_fp16")]; - tensor processed_signal_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("processed_signal_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; - tensor mel_features = cast(dtype = processed_signal_cast_fp16_to_fp32_dtype_0, x = processed_signal_cast_fp16)[name = tensor("cast_18")]; - } -> (mel_features, mel_length); -} \ No newline at end of file diff --git a/parakeet-tdt-ctc-110m/coreml/compiled_models/Preprocessor.mlmodelc/weights/weight.bin b/parakeet-tdt-ctc-110m/coreml/compiled_models/Preprocessor.mlmodelc/weights/weight.bin deleted file mode 100644 index d0c6ac87e5f78315662149af9ab4ea3658497dbe..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/compiled_models/Preprocessor.mlmodelc/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:c062338de852a26607ce4101f74e6895de3a4134a57b07232bd72bfc6f1d7f1a -size 567712 diff --git a/parakeet-tdt-ctc-110m/coreml/compiled_models/metadata.json b/parakeet-tdt-ctc-110m/coreml/compiled_models/metadata.json deleted file mode 100644 index 6db872ae8a7d463c95dc180210ace0fce019f4a6..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/compiled_models/metadata.json +++ /dev/null @@ -1,247 +0,0 @@ -{ - "model_id": "nvidia/parakeet-tdt_ctc-110m", - "model_type": "hybrid_rnnt_ctc", - "sample_rate": 16000, - "max_audio_seconds": 15.0, - "max_audio_samples": 240000, - "max_symbol_steps": 1, - "vocab_size": 1024, - "joint_extra_outputs": 5, - "encoder_dim": 512, - "decoder_dim": 640, - "decoder_hidden": 640, - "decoder_layers": 1, - "blank_id": 1024, - "checkpoint": { - "type": "pretrained", - "model_id": "nvidia/parakeet-tdt_ctc-110m" - }, - "coreml": { - "compute_units": "CPU_ONLY", - "compute_precision": "FLOAT32" - }, - "components": { - "preprocessor": { - "inputs": { - "audio_signal": [ - 1, - 240000 - ], - "audio_length": [ - 1 - ] - }, - "outputs": { - "mel": [ - 1, - 80, - 1501 - ], - "mel_length": [ - 1 - ] - }, - "path": "parakeet_preprocessor.mlpackage" - }, - "encoder": { - "inputs": { - "mel": [ - 1, - 80, - 1501 - ], - "mel_length": [ - 1 - ] - }, - "outputs": { - "encoder": [ - 1, - 512, - 188 - ], - "encoder_length": [ - 1 - ] - }, - "path": "parakeet_encoder.mlpackage" - }, - "ctc_head": { - "inputs": { - "encoder": [ - 1, - 512, - 188 - ] - }, - "outputs": { - "log_probs": [ - 1, - 188, - 1025 - ] - }, - "path": "parakeet_ctc_head.mlpackage" - }, - "mel_encoder": { - "inputs": { - "audio_signal": [ - 1, - 240000 - ], - "audio_length": [ - 1 - ] - }, - "outputs": { - "encoder": [ - 1, - 512, - 188 - ], - "encoder_length": [ - 1 - ] - }, - "path": "parakeet_mel_encoder.mlpackage" - }, - "decoder": { - "inputs": { - "targets": [ - 1, - 1 - ], - "target_length": [ - 1 - ], - "h_in": [ - 1, - 1, - 640 - ], - "c_in": [ - 1, - 1, - 640 - ] - }, - "outputs": { - "decoder": [ - 1, - 640, - 1 - ], - "h_out": [ - 1, - 1, - 640 - ], - "c_out": [ - 1, - 1, - 640 - ] - }, - "path": "parakeet_decoder.mlpackage" - }, - "joint": { - "inputs": { - "encoder": [ - 1, - 512, - 188 - ], - "decoder": [ - 1, - 640, - 1 - ] - }, - "outputs": { - "logits": [ - 1, - 188, - 1, - 1030 - ] - }, - "path": "parakeet_joint.mlpackage" - }, - "joint_decision": { - "inputs": { - "encoder": [ - 1, - 512, - 188 - ], - "decoder": [ - 1, - 640, - 1 - ] - }, - "outputs": { - "token_id": [ - 1, - 188, - 1 - ], - "token_prob": [ - 1, - 188, - 1 - ], - "duration": [ - 1, - 188, - 1 - ] - }, - "path": "parakeet_joint_decision.mlpackage" - }, - "joint_decision_single_step": { - "inputs": { - "encoder_step": [ - 1, - 512, - 1 - ], - "decoder_step": [ - 1, - 640, - 1 - ] - }, - "outputs": { - "token_id": [ - 1, - 1, - 1 - ], - "token_prob": [ - 1, - 1, - 1 - ], - "duration": [ - 1, - 1, - 1 - ], - "top_k_ids": [ - 1, - 1, - 1, - 64 - ], - "top_k_logits": [ - 1, - 1, - 1, - 64 - ] - }, - "path": "parakeet_joint_decision_single_step.mlpackage" - } - } -} \ No newline at end of file diff --git a/parakeet-tdt-ctc-110m/coreml/compiled_models/vocab.json b/parakeet-tdt-ctc-110m/coreml/compiled_models/vocab.json deleted file mode 100644 index 33b95831722563db5362641673b37885e25fb47a..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/compiled_models/vocab.json +++ /dev/null @@ -1 +0,0 @@ -{"0": "", "1": "▁t", "2": "▁th", "3": "▁a", "4": "in", "5": "re", "6": "▁the", "7": "▁w", "8": "▁s", "9": "▁o", "10": "er", "11": "ou", "12": "at", "13": "nd", "14": "it", "15": "▁h", "16": "▁c", "17": "▁b", "18": "is", "19": "en", "20": "on", "21": "ing", "22": "▁f", "23": "▁to", "24": "▁m", "25": "es", "26": "▁p", "27": "or", "28": "an", "29": "▁d", "30": "ll", "31": "▁I", "32": "ed", "33": "▁and", "34": "▁l", "35": "▁of", "36": "▁in", "37": "▁y", "38": "ar", "39": "▁g", "40": "▁you", "41": "as", "42": "om", "43": "▁n", "44": "ve", "45": "▁that", "46": "le", "47": "ic", "48": "us", "49": "ow", "50": "et", "51": "al", "52": "▁e", "53": "ut", "54": "▁it", "55": "ot", "56": "▁be", "57": "▁T", "58": "ion", "59": "▁is", "60": "▁wh", "61": "▁re", "62": "▁on", "63": "▁we", "64": "ent", "65": "▁A", "66": "ay", "67": "▁ha", "68": "▁Th", "69": "id", "70": "▁S", "71": "ac", "72": "gh", "73": "ver", "74": "ke", "75": "▁for", "76": "im", "77": "ly", "78": "ur", "79": "ld", "80": "▁he", "81": "▁st", "82": "all", "83": "ro", "84": "st", "85": "se", "86": "ct", "87": "ith", "88": "ir", "89": "am", "90": "▁this", "91": "if", "92": "▁W", "93": "oo", "94": "ri", "95": "▁was", "96": "ght", "97": "▁u", "98": "▁with", "99": "ad", "100": "ch", "101": "▁se", "102": "▁k", "103": "▁an", "104": "▁The", "105": "▁li", "106": "▁do", "107": "▁B", "108": "▁have", "109": "▁as", "110": "th", "111": "▁are", "112": "▁sh", "113": "ust", "114": "ce", "115": "ally", "116": "ill", "117": "▁H", "118": "▁j", "119": "ter", "120": "▁go", "121": "▁And", "122": "ation", "123": "▁C", "124": "▁so", "125": "ome", "126": "▁not", "127": "op", "128": "il", "129": "ore", "130": "▁ne", "131": "▁can", "132": "▁me", "133": "▁at", "134": "ould", "135": "ant", "136": "▁M", "137": "▁like", "138": "ere", "139": "▁they", "140": "ra", "141": "ers", "142": "▁ab", "143": "▁de", "144": "▁kn", "145": "ge", "146": "▁Y", "147": "▁ch", "148": "ul", "149": "pp", "150": "▁or", "151": "▁al", "152": "▁con", "153": "▁com", "154": "ess", "155": "▁su", "156": "out", "157": "▁your", "158": "▁So", "159": "ate", "160": "▁one", "161": "▁all", "162": "▁ex", "163": "est", "164": "▁fr", "165": "▁just", "166": "▁pro", "167": "▁know", "168": "▁O", "169": "ain", "170": "▁but", "171": "ol", "172": "ive", "173": "▁v", "174": "use", "175": "very", "176": "art", "177": "qu", "178": "▁my", "179": "el", "180": "▁N", "181": "nt", "182": "▁It", "183": "▁what", "184": "ab", "185": "▁P", "186": "▁wor", "187": "▁out", "188": "▁there", "189": "▁up", "190": "um", "191": "▁from", "192": "pe", "193": "▁tw", "194": "▁r", "195": "and", "196": "ight", "197": "ort", "198": "un", "199": "▁L", "200": "ist", "201": "▁about", "202": "ide", "203": "ig", "204": "ake", "205": "▁D", "206": "em", "207": "os", "208": "king", "209": "rou", "210": "ind", "211": "our", "212": "res", "213": "▁We", "214": "▁get", "215": "▁E", "216": "▁G", "217": "ack", "218": "▁le", "219": "ity", "220": "od", "221": "▁F", "222": "ard", "223": "▁pl", "224": "▁our", "225": "▁int", "226": "ment", "227": "▁will", "228": "ies", "229": "▁by", "230": "ink", "231": "ca", "232": "▁if", "233": "red", "234": "her", "235": "ie", "236": "▁us", "237": "▁some", "238": "▁don", "239": "ven", "240": "ood", "241": "ast", "242": "▁R", "243": "▁his", "244": "▁tim", "245": "▁tr", "246": "▁more", "247": "ich", "248": "ous", "249": "ame", "250": "▁going", "251": "▁had", "252": "▁them", "253": "ook", "254": "▁pe", "255": "▁Wh", "256": "▁You", "257": "▁But", "258": "ine", "259": "▁here", "260": "▁would", "261": "cause", "262": "right", "263": "so", "264": "ost", "265": "ure", "266": "▁has", "267": "ect", "268": "▁think", "269": "▁fe", "270": "ong", "271": "▁see", "272": "▁when", "273": "▁who", "274": "▁were", "275": "▁really", "276": "▁their", "277": "▁want", "278": "one", "279": "ople", "280": "▁then", "281": "▁time", "282": "▁sa", "283": "ap", "284": "▁te", "285": "▁He", "286": "▁ye", "287": "ck", "288": "▁her", "289": "▁thing", "290": "▁right", "291": "▁which", "292": "itt", "293": "ice", "294": "act", "295": "▁people", "296": "ty", "297": "▁two", "298": "▁J", "299": "▁im", "300": "ther", "301": "ci", "302": "ose", "303": "▁cl", "304": "▁qu", "305": "▁man", "306": "▁also", "307": "ree", "308": "▁en", "309": "ud", "310": "▁how", "311": "reat", "312": "ak", "313": "hing", "314": "ag", "315": "▁any", "316": "ff", "317": "ace", "318": "per", "319": "▁because", "320": "▁very", "321": "own", "322": "▁ad", "323": "▁act", "324": "▁been", "325": "▁now", "326": "▁ag", "327": "▁into", "328": "▁comp", "329": "ars", "330": "ions", "331": "are", "332": "ite", "333": "iv", "334": "▁these", "335": "ays", "336": "ep", "337": "▁This", "338": "▁she", "339": "ans", "340": "ah", "341": "een", "342": "▁over", "343": "ry", "344": "▁lo", "345": "age", "346": "▁pr", "347": "▁sp", "348": "ue", "349": "▁co", "350": "ick", "351": "ber", "352": "▁did", "353": "ip", "354": "ach", "355": "▁back", "356": "▁no", "357": "▁cont", "358": "▁other", "359": "▁every", "360": "pt", "361": "▁need", "362": "▁him", "363": "▁U", "364": "▁In", "365": "▁work", "366": "irst", "367": "▁part", "368": "▁look", "369": "ittle", "370": "ble", "371": "iz", "372": "▁un", "373": "▁make", "374": "omet", "375": "nder", "376": "ish", "377": "na", "378": "▁little", "379": "▁off", "380": "▁than", "381": "▁got", "382": "ually", "383": "▁per", "384": "▁good", "385": "▁way", "386": "▁could", "387": "▁ac", "388": "▁imp", "389": "able", "390": "▁where", "391": "iff", "392": "▁That", "393": "▁res", "394": "ount", "395": "pl", "396": "ance", "397": "▁first", "398": "▁ro", "399": "▁pre", "400": "ass", "401": "▁say", "402": "int", "403": "ated", "404": "ire", "405": "uch", "406": "ase", "407": "▁somet", "408": "ound", "409": "▁down", "410": "▁diff", "411": "sel", "412": "▁gu", "413": "▁am", "414": "ress", "415": "▁lot", "416": "ence", "417": "▁dis", "418": "orm", "419": "ix", "420": "▁po", "421": "ving", "422": "enty", "423": "▁K", "424": "▁spe", "425": "und", "426": "he", "427": "▁much", "428": "▁ar", "429": "round", "430": "▁app", "431": "co", "432": "ark", "433": "▁new", "434": "ater", "435": "ult", "436": "end", "437": "▁even", "438": "▁start", "439": "ations", "440": "rough", "441": "ile", "442": "fter", "443": "▁well", "444": "be", "445": "▁They", "446": "▁three", "447": "ign", "448": "ild", "449": "▁said", "450": "ough", "451": "ang", "452": "▁too", "453": "ade", "454": "▁bl", "455": "ens", "456": "▁inc", "457": "ia", "458": "▁those", "459": "▁mo", "460": "▁take", "461": "▁through", "462": "▁fl", "463": "▁kind", "464": "▁things", "465": "▁bet", "466": "▁only", "467": "▁St", "468": "▁let", "469": "cess", "470": "▁Ch", "471": "ary", "472": "vel", "473": "▁If", "474": "xt", "475": "other", "476": "av", "477": "ical", "478": "ord", "479": "▁again", "480": "▁something", "481": "onna", "482": "fore", "483": "▁may", "484": "ting", "485": "▁bu", "486": "▁differe", "487": "urn", "488": "▁gonna", "489": "▁does", "490": "uct", "491": "og", "492": "▁twenty", "493": "▁gr", "494": "▁Ye", "495": "wn", "496": "▁should", "497": "▁comm", "498": "ition", "499": "▁under", "500": "▁hel", "501": "ory", "502": "▁fo", "503": "▁use", "504": "igh", "505": "ife", "506": "▁actually", "507": "▁tal", "508": "▁call", "509": "ents", "510": "ious", "511": "ull", "512": "▁There", "513": "▁Yeah", "514": "▁most", "515": "▁ke", "516": "ors", "517": "ved", "518": "ys", "519": "▁sc", "520": "▁happ", "521": "ope", "522": "▁help", "523": "atch", "524": "▁What", "525": "▁rem", "526": "ple", "527": "▁Now", "528": "▁br", "529": "ool", "530": "oth", "531": "▁four", "532": "self", "533": "▁str", "534": "ne", "535": "thing", "536": "▁put", "537": "ial", "538": "▁great", "539": "ail", "540": "ub", "541": "ning", "542": "▁sm", "543": "▁feel", "544": "▁five", "545": "ody", "546": "undred", "547": "iss", "548": "ank", "549": "get", "550": "aking", "551": "▁many", "552": "▁hundred", "553": "▁years", "554": "▁being", "555": "▁come", "556": "▁mean", "557": "ily", "558": "▁different", "559": "▁after", "560": "▁ser", "561": "▁show", "562": "form", "563": "ful", "564": "oy", "565": "▁six", "566": "▁vide", "567": "▁V", "568": "▁its", "569": "▁point", "570": "▁day", "571": "▁des", "572": "ons", "573": "▁bit", "574": "▁bel", "575": "▁before", "576": "▁aw", "577": "▁end", "578": "▁Oh", "579": "▁still", "580": "ath", "581": "▁long", "582": "▁'", "583": "ise", "584": "ob", "585": "day", "586": "▁add", "587": "ft", "588": "ves", "589": "ces", "590": "ady", "591": "▁cr", "592": "▁around", "593": "▁try", "594": "les", "595": "vers", "596": "kay", "597": "ian", "598": "ates", "599": "▁find", "600": "ward", "601": "▁As", "602": "▁eight", "603": "lic", "604": "▁same", "605": "▁pos", "606": "▁em", "607": "▁made", "608": "▁supp", "609": "▁life", "610": "▁Be", "611": "pect", "612": "▁dec", "613": "▁play", "614": "ange", "615": "▁att", "616": "▁pers", "617": "ways", "618": "▁high", "619": "▁hand", "620": "▁next", "621": "▁cons", "622": "▁own", "623": "▁inv", "624": "ower", "625": "▁ind", "626": "ert", "627": "ng", "628": "ave", "629": "▁year", "630": "▁big", "631": "ating", "632": "▁world", "633": "▁rel", "634": "▁sure", "635": "▁tra", "636": "ew", "637": "ered", "638": "▁fin", "639": "▁Well", "640": "▁sl", "641": "▁doing", "642": "bs", "643": "▁set", "644": "▁rec", "645": "ual", "646": "cial", "647": "▁ph", "648": "erm", "649": "▁love", "650": "ph", "651": "▁real", "652": "▁last", "653": "ict", "654": "▁bo", "655": "▁ra", "656": "ible", "657": "▁wr", "658": "mer", "659": "▁count", "660": "ities", "661": "▁always", "662": "inet", "663": "ments", "664": "uc", "665": "▁might", "666": "▁inter", "667": "▁video", "668": "gin", "669": "▁tell", "670": "▁never", "671": "vent", "672": "▁import", "673": "ied", "674": "▁sy", "675": "▁How", "676": "ically", "677": "ought", "678": "▁thir", "679": "▁rep", "680": "ks", "681": "ib", "682": "▁fam", "683": "ject", "684": "▁bas", "685": "▁She", "686": "▁give", "687": "akes", "688": "▁ninet", "689": "▁reg", "690": "▁min", "691": "▁op", "692": "▁def", "693": "▁didn", "694": "te", "695": "▁cour", "696": "▁why", "697": "▁ent", "698": "▁place", "699": "▁ins", "700": "▁car", "701": "ather", "702": "▁person", "703": "ular", "704": "▁inst", "705": "▁prod", "706": "lect", "707": "▁Al", "708": "▁today", "709": "▁bec", "710": "▁sur", "711": "▁All", "712": "▁another", "713": "▁bus", "714": "▁keep", "715": "ell", "716": "ese", "717": "riend", "718": "▁quest", "719": "▁talk", "720": "als", "721": "ings", "722": "▁mon", "723": "cond", "724": "old", "725": "▁acc", "726": "▁la", "727": "▁num", "728": "ident", "729": "▁che", "730": "iness", "731": "▁turn", "732": "▁ear", "733": "▁No", "734": "ousand", "735": "▁better", "736": "ific", "737": "▁loo", "738": "▁gl", "739": "oc", "740": "▁important", "741": "ited", "742": "▁An", "743": "▁thousand", "744": "ility", "745": "llow", "746": "▁used", "747": "▁gen", "748": "▁sim", "749": "li", "750": "▁happen", "751": "▁Un", "752": "▁Let", "753": "air", "754": "ock", "755": "ably", "756": "gg", "757": "▁watch", "758": "▁For", "759": "▁sw", "760": "ren", "761": "ute", "762": "ever", "763": "▁pol", "764": "▁sch", "765": "▁When", "766": "▁such", "767": "▁fif", "768": "▁home", "769": "▁cle", "770": "▁contin", "771": "ouse", "772": "▁friend", "773": "uring", "774": "▁Okay", "775": "gr", "776": "▁able", "777": "▁stud", "778": "▁eff", "779": "hip", "780": "body", "781": "▁top", "782": "ness", "783": "▁exper", "784": "▁pret", "785": "▁both", "786": "▁done", "787": "cri", "788": "▁mark", "789": "▁while", "790": "▁old", "791": "ros", "792": "ont", "793": "▁second", "794": "ative", "795": "▁thought", "796": "▁best", "797": "▁found", "798": "iew", "799": "▁belie", "800": "▁each", "801": "erest", "802": "▁tri", "803": "▁eas", "804": "▁ca", "805": "▁fact", "806": "▁care", "807": "▁fun", "808": "atter", "809": "ures", "810": "▁head", "811": "▁lear", "812": "▁water", "813": "▁hard", "814": "▁few", "815": "▁side", "816": "ween", "817": "▁exp", "818": "▁away", "819": "its", "820": "▁ext", "821": "lud", "822": "▁run", "823": "▁trans", "824": "ince", "825": "▁sk", "826": "▁open", "827": "cus", "828": "▁between", "829": "▁called", "830": "▁wee", "831": "▁pretty", "832": "ason", "833": "▁far", "834": "ember", "835": "omm", "836": "▁interest", "837": "any", "838": "ner", "839": "uff", "840": "▁pres", "841": "▁cur", "842": "▁child", "843": "ee", "844": "▁toget", "845": "▁together", "846": "olog", "847": "▁God", "848": "ond", "849": "▁char", "850": "▁looking", "851": "stem", "852": "az", "853": "cent", "854": "▁ob", "855": "▁ass", "856": "land", "857": "▁doesn", "858": "▁business", "859": "▁course", "860": "▁ten", "861": "ps", "862": "arch", "863": "ced", "864": "ms", "865": "ize", "866": "nce", "867": "▁ref", "868": "▁name", "869": "ross", "870": "▁grow", "871": "oney", "872": "▁went", "873": "ics", "874": "teen", "875": "▁cou", "876": "▁prob", "877": "▁ret", "878": "▁guys", "879": "▁came", "880": "ash", "881": "led", "882": "▁Eur", "883": "ues", "884": "▁ide", "885": "gan", "886": "▁everything", "887": "▁getting", "888": "▁ask", "889": "▁cor", "890": "▁build", "891": "▁sign", "892": "▁small", "893": "uck", "894": "▁el", "895": "▁col", "896": "▁Is", "897": "ational", "898": "stand", "899": "cy", "900": "▁conf", "901": "der", "902": "▁bre", "903": "▁cap", "904": "▁mod", "905": "ets", "906": "ike", "907": "▁number", "908": "▁comple", "909": "ertain", "910": "▁ever", "911": "▁coll", "912": "▁hum", "913": "▁Europe", "914": "▁cre", "915": "▁met", "916": "▁exam", "917": "▁move", "918": "▁pass", "919": "▁left", "920": "▁system", "921": "▁includ", "922": "▁Thank", "923": "cept", "924": "▁wom", "925": "▁product", "926": "ten", "927": "▁rest", "928": "▁probably", "929": "▁dri", "930": "▁Do", "931": "▁gener", "932": "▁anything", "933": "▁lar", "934": "▁My", "935": "▁school", "936": "▁lead", "937": "▁sub", "938": "▁ty", "939": "▁plan", "940": "▁seem", "941": "▁whole", "942": "irect", "943": "▁light", "944": "▁must", "945": "▁mom", "946": "▁opp", "947": "▁support", "948": "▁family", "949": "ices", "950": "amp", "951": "▁proble", "952": "▁dr", "953": "ready", "954": "▁using", "955": "ense", "956": "▁prov", "957": "ush", "958": "ax", "959": "▁power", "960": "▁Re", "961": "alth", "962": "▁ev", "963": "▁stand", "964": "▁war", "965": "ts", "966": "▁", "967": "e", "968": "t", "969": "o", "970": "a", "971": "n", "972": "i", "973": "s", "974": "r", "975": "h", "976": "l", "977": "d", "978": "u", "979": "c", "980": "m", "981": "y", "982": "g", "983": "w", "984": "f", "985": "p", "986": ".", "987": "b", "988": ",", "989": "v", "990": "k", "991": "'", "992": "I", "993": "T", "994": "A", "995": "S", "996": "x", "997": "W", "998": "j", "999": "B", "1000": "C", "1001": "H", "1002": "?", "1003": "M", "1004": "O", "1005": "Y", "1006": "N", "1007": "P", "1008": "E", "1009": "q", "1010": "L", "1011": "D", "1012": "z", "1013": "G", "1014": "F", "1015": "R", "1016": "!", "1017": "J", "1018": "U", "1019": "K", "1020": "V", "1021": "Q", "1022": "Z", "1023": "X"} \ No newline at end of file diff --git a/parakeet-tdt-ctc-110m/coreml/convert-parakeet.py b/parakeet-tdt-ctc-110m/coreml/convert-parakeet.py deleted file mode 100644 index 6cd1c871c81f26b9449bf6313077afe8e27527d9..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/convert-parakeet.py +++ /dev/null @@ -1,697 +0,0 @@ -#!/usr/bin/env python3 -"""CLI for exporting Parakeet TDT-CTC 110M Hybrid components to CoreML.""" -from __future__ import annotations - -import json -from dataclasses import asdict -from pathlib import Path -from typing import Dict, Optional, Tuple - -import coremltools as ct -import numpy as np -import soundfile as sf -import torch -import typer - -import nemo.collections.asr as nemo_asr - -from individual_components import ( - CTCHeadWrapper, - DecoderWrapper, - EncoderWrapper, - ExportSettings, - JointWrapper, - JointDecisionWrapper, - JointDecisionSingleStep, - PreprocessorWrapper, - MelEncoderWrapper, - _coreml_convert, -) - -DEFAULT_MODEL_ID = "nvidia/parakeet-tdt_ctc-110m" -AUTHOR = "Fluid Inference" - - -def _compute_length(seconds: float, sample_rate: int) -> int: - return int(round(seconds * sample_rate)) - - -def _prepare_audio( - validation_audio: Optional[Path], - sample_rate: int, - max_samples: int, - seed: Optional[int], -) -> torch.Tensor: - if validation_audio is None: - if seed is not None: - torch.manual_seed(seed) - audio = torch.randn(1, max_samples, dtype=torch.float32) - return audio - - data, sr = sf.read(str(validation_audio), dtype="float32") - if sr != sample_rate: - raise typer.BadParameter( - f"Validation audio sample rate {sr} does not match model rate {sample_rate}" - ) - - if data.ndim > 1: - data = data[:, 0] - - if data.size == 0: - raise typer.BadParameter("Validation audio is empty") - - if data.size < max_samples: - pad_width = max_samples - data.size - data = np.pad(data, (0, pad_width)) - elif data.size > max_samples: - data = data[:max_samples] - - audio = torch.from_numpy(data).unsqueeze(0).to(dtype=torch.float32) - return audio - - -def _save_mlpackage(model: ct.models.MLModel, path: Path, description: str) -> None: - # Ensure iOS 17+ target for MLProgram ops and ANE readiness - try: - model.minimum_deployment_target = ct.target.iOS17 - except Exception: - pass - model.short_description = description - model.author = AUTHOR - path.parent.mkdir(parents=True, exist_ok=True) - model.save(str(path)) - - -def _tensor_shape(tensor: torch.Tensor) -> Tuple[int, ...]: - return tuple(int(dim) for dim in tensor.shape) - - -def _parse_compute_units(name: str) -> ct.ComputeUnit: - """Parse a human-friendly compute units string into ct.ComputeUnit. - - Accepted (case-insensitive): ALL, CPU_ONLY, CPU_AND_GPU, CPU_AND_NE. - """ - normalized = str(name).strip().upper() - mapping = { - "ALL": ct.ComputeUnit.ALL, - "CPU_ONLY": ct.ComputeUnit.CPU_ONLY, - "CPU_AND_GPU": ct.ComputeUnit.CPU_AND_GPU, - "CPU_AND_NE": ct.ComputeUnit.CPU_AND_NE, - "CPU_AND_NEURALENGINE": ct.ComputeUnit.CPU_AND_NE, - } - if normalized not in mapping: - raise typer.BadParameter( - f"Unknown compute units '{name}'. Choose from: " + ", ".join(mapping.keys()) - ) - return mapping[normalized] - - -def _parse_compute_precision(name: Optional[str]) -> Optional[ct.precision]: - """Parse compute precision string into ct.precision or None. - - Accepted (case-insensitive): FLOAT32, FLOAT16. If None/empty, returns None (tool default). - """ - if name is None: - return None - normalized = str(name).strip().upper() - if normalized == "": - return None - mapping = { - "FLOAT32": ct.precision.FLOAT32, - "FLOAT16": ct.precision.FLOAT16, - } - if normalized not in mapping: - raise typer.BadParameter( - f"Unknown compute precision '{name}'. Choose from: " + ", ".join(mapping.keys()) - ) - return mapping[normalized] - - -# Fixed export choices: CPU_ONLY + FP32, min target iOS17 - - -app = typer.Typer(add_completion=False, pretty_exceptions_show_locals=False) - - -@app.command() -def convert( - nemo_path: Optional[Path] = typer.Option( - None, - "--nemo-path", - exists=True, - resolve_path=True, - help="Path to parakeet-tdt_ctc-110m .nemo checkpoint (skip to auto-download)", - ), - model_id: str = typer.Option( - DEFAULT_MODEL_ID, - "--model-id", - help="Model identifier to download when --nemo-path is omitted", - ), - output_dir: Path = typer.Option(Path("parakeet_110m_coreml"), help="Directory where mlpackages and metadata will be written"), - preprocessor_cu: str = typer.Option( - "CPU_ONLY", - "--preprocessor-cu", - help="Compute units for preprocessor (default CPU_ONLY)", - ), - mel_encoder_cu: str = typer.Option( - "CPU_ONLY", - "--mel-encoder-cu", - help="Compute units for fused mel+encoder (default CPU_ONLY)", - ), - compute_precision: Optional[str] = typer.Option( - None, - "--compute-precision", - help="Export precision: FLOAT32 (default) or FLOAT16 to shrink non-quantized weights.", - ), -) -> None: - """Export all Parakeet TDT-CTC 110M Hybrid sub-modules to CoreML with a fixed 15-second window. - - This exports both CTC and TDT components from the hybrid model. - """ - # Runtime CoreML contract keeps U=1 so the prediction net matches the streaming decoder. - export_settings = ExportSettings( - output_dir=output_dir, - compute_units=ct.ComputeUnit.CPU_ONLY, # Default: CPU-only for all components - deployment_target=ct.target.iOS17, # iOS 17+ features and kernels - compute_precision=_parse_compute_precision(compute_precision), - max_audio_seconds=15.0, - max_symbol_steps=1, - ) - - typer.echo("Export configuration:") - typer.echo(asdict(export_settings)) - - output_dir.mkdir(parents=True, exist_ok=True) - pre_cu = _parse_compute_units(preprocessor_cu) - melenc_cu = _parse_compute_units(mel_encoder_cu) - - if nemo_path is not None: - typer.echo(f"Loading NeMo model from {nemo_path}
") - # 110M is a hybrid model: EncDecHybridRNNTCTCBPEModel - asr_model = nemo_asr.models.EncDecHybridRNNTCTCBPEModel.restore_from( - str(nemo_path), map_location="cpu" - ) - checkpoint_meta = { - "type": "file", - "path": str(nemo_path), - } - else: - typer.echo(f"Downloading NeMo model via {model_id}
") - # 110M is a hybrid model: EncDecHybridRNNTCTCBPEModel - asr_model = nemo_asr.models.EncDecHybridRNNTCTCBPEModel.from_pretrained( - model_id, map_location="cpu" - ) - checkpoint_meta = { - "type": "pretrained", - "model_id": model_id, - } - asr_model.eval() - - sample_rate = int(asr_model.cfg.preprocessor.sample_rate) - max_samples = _compute_length(export_settings.max_audio_seconds, sample_rate) - - # Look for a bundled 15s 16kHz audio file - default_audio = (Path(__file__).parent / "audio" / "yc_first_minute_16k_15s.wav").resolve() - if default_audio.exists(): - typer.echo(f"Using trace audio: {default_audio}") - audio_tensor = _prepare_audio(default_audio, sample_rate, max_samples, seed=None) - else: - typer.echo("No trace audio found, using random noise for tracing") - audio_tensor = _prepare_audio(None, sample_rate, max_samples, seed=42) - audio_length = torch.tensor([max_samples], dtype=torch.int32) - - preprocessor = PreprocessorWrapper(asr_model.preprocessor.eval()) - encoder = EncoderWrapper(asr_model.encoder.eval()) - decoder = DecoderWrapper(asr_model.decoder.eval()) - joint = JointWrapper(asr_model.joint.eval()) - # CTC head for hybrid model - ctc_head = CTCHeadWrapper(asr_model.ctc_decoder.eval()) - - decoder_export_flag = getattr(asr_model.decoder, "_rnnt_export", False) - asr_model.decoder._rnnt_export = True - - try: - with torch.inference_mode(): - mel_ref, mel_length_ref = preprocessor(audio_tensor, audio_length) - mel_length_ref = mel_length_ref.to(dtype=torch.int32) - encoder_ref, encoder_length_ref = encoder(mel_ref, mel_length_ref) - encoder_length_ref = encoder_length_ref.to(dtype=torch.int32) - # CTC log probs - ctc_log_probs_ref = ctc_head(encoder_ref) - - # Clone Tensors to drop the inference tensor flag before tracing - mel_ref = mel_ref.clone() - mel_length_ref = mel_length_ref.clone() - encoder_ref = encoder_ref.clone() - encoder_length_ref = encoder_length_ref.clone() - ctc_log_probs_ref = ctc_log_probs_ref.clone() - - vocab_size = int(asr_model.tokenizer.vocab_size) - num_extra = int(asr_model.joint.num_extra_outputs) - decoder_hidden = int(asr_model.decoder.pred_hidden) - decoder_layers = int(asr_model.decoder.pred_rnn_layers) - - typer.echo(f"Model info:") - typer.echo(f" Vocab size: {vocab_size}") - typer.echo(f" Num extra (duration bins): {num_extra}") - typer.echo(f" Decoder hidden: {decoder_hidden}") - typer.echo(f" Decoder layers: {decoder_layers}") - typer.echo(f" Encoder output shape: {_tensor_shape(encoder_ref)}") - - targets = torch.full( - (1, export_settings.max_symbol_steps), - fill_value=asr_model.decoder.blank_idx, - dtype=torch.int32, - ) - target_lengths = torch.tensor( - [export_settings.max_symbol_steps], dtype=torch.int32 - ) - zero_state = torch.zeros( - decoder_layers, - 1, - decoder_hidden, - dtype=torch.float32, - ) - - with torch.inference_mode(): - decoder_ref, h_ref, c_ref = decoder(targets, target_lengths, zero_state, zero_state) - joint_ref = joint(encoder_ref, decoder_ref) - - decoder_ref = decoder_ref.clone() - h_ref = h_ref.clone() - c_ref = c_ref.clone() - joint_ref = joint_ref.clone() - - typer.echo(f" Decoder output shape: {_tensor_shape(decoder_ref)}") - typer.echo(f" Joint output shape: {_tensor_shape(joint_ref)}") - typer.echo(f" CTC log probs shape: {_tensor_shape(ctc_log_probs_ref)}") - - typer.echo("Tracing and converting preprocessor
") - # Ensure tracing happens on CPU explicitly - preprocessor = preprocessor.cpu() - audio_tensor = audio_tensor.cpu() - audio_length = audio_length.cpu() - traced_preprocessor = torch.jit.trace( - preprocessor, (audio_tensor, audio_length), strict=False - ) - traced_preprocessor.eval() - preprocessor_inputs = [ - # Allow variable-length audio up to the fixed 15s window using RangeDim - ct.TensorType( - name="audio", - shape=(1, ct.RangeDim(1, max_samples)), - dtype=np.float32, - ), - ct.TensorType(name="audio_length", shape=(1,), dtype=np.int32), - ] - preprocessor_outputs = [ - ct.TensorType(name="mel_features", dtype=np.float32), - ct.TensorType(name="mel_length", dtype=np.int32), - ] - # Preprocessor compute units (parametrized; default CPU_ONLY) - preprocessor_model = _coreml_convert( - traced_preprocessor, - preprocessor_inputs, - preprocessor_outputs, - export_settings, - compute_units_override=pre_cu, - ) - preprocessor_path = output_dir / "parakeet_preprocessor.mlpackage" - _save_mlpackage( - preprocessor_model, - preprocessor_path, - "Parakeet 110M preprocessor (15 s window)", - ) - - typer.echo("Tracing and converting encoder
") - traced_encoder = torch.jit.trace( - encoder, (mel_ref, mel_length_ref), strict=False - ) - traced_encoder.eval() - encoder_inputs = [ - ct.TensorType(name="mel_features", shape=_tensor_shape(mel_ref), dtype=np.float32), - ct.TensorType(name="mel_length", shape=(1,), dtype=np.int32), - ] - encoder_outputs = [ - ct.TensorType(name="encoder_output", dtype=np.float32), - ct.TensorType(name="encoder_length", dtype=np.int32), - ] - # Encoder: CPU only - encoder_model = _coreml_convert( - traced_encoder, - encoder_inputs, - encoder_outputs, - export_settings, - compute_units_override=ct.ComputeUnit.CPU_ONLY, - ) - encoder_path = output_dir / "parakeet_encoder.mlpackage" - _save_mlpackage( - encoder_model, - encoder_path, - "Parakeet 110M encoder (15 s window)", - ) - - # CTC Head for hybrid model - typer.echo("Tracing and converting CTC head
") - traced_ctc_head = torch.jit.trace( - ctc_head, (encoder_ref,), strict=False - ) - traced_ctc_head.eval() - ctc_head_inputs = [ - ct.TensorType(name="encoder_output", shape=_tensor_shape(encoder_ref), dtype=np.float32), - ] - ctc_head_outputs = [ - ct.TensorType(name="ctc_logits", dtype=np.float32), - ] - ctc_head_model = _coreml_convert( - traced_ctc_head, - ctc_head_inputs, - ctc_head_outputs, - export_settings, - compute_units_override=ct.ComputeUnit.CPU_ONLY, - ) - ctc_head_path = output_dir / "parakeet_ctc_head.mlpackage" - _save_mlpackage( - ctc_head_model, - ctc_head_path, - "Parakeet 110M CTC decoder head", - ) - - # Optional fused export: Preprocessor + Encoder - typer.echo("Tracing and converting fused mel+encoder
") - mel_encoder = MelEncoderWrapper(preprocessor, encoder) - traced_mel_encoder = torch.jit.trace( - mel_encoder, (audio_tensor, audio_length), strict=False - ) - traced_mel_encoder.eval() - mel_encoder_inputs = [ - # Keep fixed 15s window for fused Mel+Encoder - ct.TensorType(name="audio", shape=(1, max_samples), dtype=np.float32), - ct.TensorType(name="audio_length", shape=(1,), dtype=np.int32), - ] - mel_encoder_outputs = [ - ct.TensorType(name="encoder_output", dtype=np.float32), - ct.TensorType(name="encoder_length", dtype=np.int32), - ] - # Fused mel+encoder compute units (parametrized; default CPU_ONLY) - mel_encoder_model = _coreml_convert( - traced_mel_encoder, - mel_encoder_inputs, - mel_encoder_outputs, - export_settings, - compute_units_override=melenc_cu, - ) - mel_encoder_path = output_dir / "parakeet_mel_encoder.mlpackage" - _save_mlpackage( - mel_encoder_model, - mel_encoder_path, - "Parakeet 110M fused Mel+Encoder (15 s window)", - ) - - typer.echo("Tracing and converting decoder
") - traced_decoder = torch.jit.trace( - decoder, - (targets, target_lengths, zero_state, zero_state), - strict=False, - ) - traced_decoder.eval() - decoder_inputs = [ - ct.TensorType(name="targets", shape=_tensor_shape(targets), dtype=np.int32), - ct.TensorType(name="target_length", shape=(1,), dtype=np.int32), - ct.TensorType(name="h_in", shape=_tensor_shape(zero_state), dtype=np.float32), - ct.TensorType(name="c_in", shape=_tensor_shape(zero_state), dtype=np.float32), - ] - decoder_outputs = [ - ct.TensorType(name="decoder", dtype=np.float32), - ct.TensorType(name="h_out", dtype=np.float32), - ct.TensorType(name="c_out", dtype=np.float32), - ] - # Decoder: CPU only - decoder_model = _coreml_convert( - traced_decoder, - decoder_inputs, - decoder_outputs, - export_settings, - compute_units_override=ct.ComputeUnit.CPU_ONLY, - ) - decoder_path = output_dir / "parakeet_decoder.mlpackage" - _save_mlpackage( - decoder_model, - decoder_path, - "Parakeet 110M decoder (RNNT prediction network)", - ) - - typer.echo("Tracing and converting joint
") - traced_joint = torch.jit.trace( - joint, - (encoder_ref, decoder_ref), - strict=False, - ) - traced_joint.eval() - joint_inputs = [ - ct.TensorType(name="encoder", shape=_tensor_shape(encoder_ref), dtype=np.float32), - ct.TensorType(name="decoder", shape=_tensor_shape(decoder_ref), dtype=np.float32), - ] - joint_outputs = [ - ct.TensorType(name="logits", dtype=np.float32), - ] - # Joint: CPU only - joint_model = _coreml_convert( - traced_joint, - joint_inputs, - joint_outputs, - export_settings, - compute_units_override=ct.ComputeUnit.CPU_ONLY, - ) - joint_path = output_dir / "parakeet_joint.mlpackage" - _save_mlpackage( - joint_model, - joint_path, - "Parakeet 110M joint network (RNNT)", - ) - - # Joint + decision head (split logits, softmax, argmax) - typer.echo("Tracing and converting joint decision head
") - vocab_size = int(asr_model.tokenizer.vocab_size) - num_extra = int(asr_model.joint.num_extra_outputs) - joint_decision = JointDecisionWrapper(joint, vocab_size=vocab_size, num_extra=num_extra) - traced_joint_decision = torch.jit.trace( - joint_decision, - (encoder_ref, decoder_ref), - strict=False, - ) - traced_joint_decision.eval() - joint_decision_inputs = [ - ct.TensorType(name="encoder", shape=_tensor_shape(encoder_ref), dtype=np.float32), - ct.TensorType(name="decoder", shape=_tensor_shape(decoder_ref), dtype=np.float32), - ] - joint_decision_outputs = [ - ct.TensorType(name="token_id", dtype=np.int32), - ct.TensorType(name="token_prob", dtype=np.float32), - ct.TensorType(name="duration", dtype=np.int32), - ] - # JointDecision: CPU only - joint_decision_model = _coreml_convert( - traced_joint_decision, - joint_decision_inputs, - joint_decision_outputs, - export_settings, - compute_units_override=ct.ComputeUnit.CPU_ONLY, - ) - joint_decision_path = output_dir / "parakeet_joint_decision.mlpackage" - _save_mlpackage( - joint_decision_model, - joint_decision_path, - "Parakeet 110M joint + decision head (split, softmax, argmax)", - ) - - # Single-step JointDecision for [1,512,1] x [1,640,1] -> [1,1,1] - # Note: 110M encoder dim is 512 (not 1024 like 0.6B) - typer.echo("Tracing and converting single-step joint decision
") - jd_single = JointDecisionSingleStep(joint, vocab_size=vocab_size, num_extra=num_extra) - # Create single-step slices from refs - enc_step = encoder_ref[:, :, :1].contiguous() - dec_step = decoder_ref[:, :, :1].contiguous() - traced_jd_single = torch.jit.trace( - jd_single, - (enc_step, dec_step), - strict=False, - ) - traced_jd_single.eval() - jd_single_inputs = [ - ct.TensorType(name="encoder_step", shape=(1, enc_step.shape[1], 1), dtype=np.float32), - ct.TensorType(name="decoder_step", shape=(1, dec_step.shape[1], 1), dtype=np.float32), - ] - jd_single_outputs = [ - ct.TensorType(name="token_id", dtype=np.int32), - ct.TensorType(name="token_prob", dtype=np.float32), - ct.TensorType(name="duration", dtype=np.int32), - ct.TensorType(name="top_k_ids", dtype=np.int32), - ct.TensorType(name="top_k_logits", dtype=np.float32), - ] - # Single-step JointDecision: CPU only - jd_single_model = _coreml_convert( - traced_jd_single, - jd_single_inputs, - jd_single_outputs, - export_settings, - compute_units_override=ct.ComputeUnit.CPU_ONLY, - ) - jd_single_path = output_dir / "parakeet_joint_decision_single_step.mlpackage" - _save_mlpackage( - jd_single_model, - jd_single_path, - "Parakeet 110M single-step joint decision (current frame)", - ) - - # Export vocabulary - typer.echo("Exporting vocabulary
") - vocab_path = output_dir / "vocab.json" - vocab_dict = { - "vocab_size": vocab_size, - "blank_id": int(asr_model.decoder.blank_idx), - "tokens": asr_model.tokenizer.vocab, - } - vocab_path.write_text(json.dumps(vocab_dict, indent=2, ensure_ascii=False)) - - metadata: Dict[str, object] = { - "model_id": model_id, - "model_type": "hybrid_rnnt_ctc", - "sample_rate": sample_rate, - "max_audio_seconds": export_settings.max_audio_seconds, - "max_audio_samples": max_samples, - "max_symbol_steps": export_settings.max_symbol_steps, - "vocab_size": vocab_size, - "joint_extra_outputs": num_extra, - "encoder_dim": int(encoder_ref.shape[1]), # 512 for 110M - "decoder_dim": int(decoder_ref.shape[1]), # 640 - "decoder_hidden": decoder_hidden, - "decoder_layers": decoder_layers, - "blank_id": int(asr_model.decoder.blank_idx), - "checkpoint": checkpoint_meta, - "coreml": { - "compute_units": export_settings.compute_units.name, - "compute_precision": ( - export_settings.compute_precision.name - if export_settings.compute_precision is not None - else "FLOAT32" - ), - }, - "components": { - "preprocessor": { - "inputs": { - "audio_signal": list(_tensor_shape(audio_tensor)), - "audio_length": [1], - }, - "outputs": { - "mel": list(_tensor_shape(mel_ref)), - "mel_length": [1], - }, - "path": preprocessor_path.name, - }, - "encoder": { - "inputs": { - "mel": list(_tensor_shape(mel_ref)), - "mel_length": [1], - }, - "outputs": { - "encoder": list(_tensor_shape(encoder_ref)), - "encoder_length": [1], - }, - "path": encoder_path.name, - }, - "ctc_head": { - "inputs": { - "encoder": list(_tensor_shape(encoder_ref)), - }, - "outputs": { - "log_probs": list(_tensor_shape(ctc_log_probs_ref)), - }, - "path": ctc_head_path.name, - }, - "mel_encoder": { - "inputs": { - "audio_signal": [1, max_samples], - "audio_length": [1], - }, - "outputs": { - "encoder": list(_tensor_shape(encoder_ref)), - "encoder_length": [1], - }, - "path": mel_encoder_path.name, - }, - "decoder": { - "inputs": { - "targets": list(_tensor_shape(targets)), - "target_length": [1], - "h_in": list(_tensor_shape(zero_state)), - "c_in": list(_tensor_shape(zero_state)), - }, - "outputs": { - "decoder": list(_tensor_shape(decoder_ref)), - "h_out": list(_tensor_shape(h_ref)), - "c_out": list(_tensor_shape(c_ref)), - }, - "path": decoder_path.name, - }, - "joint": { - "inputs": { - "encoder": list(_tensor_shape(encoder_ref)), - "decoder": list(_tensor_shape(decoder_ref)), - }, - "outputs": { - "logits": list(_tensor_shape(joint_ref)), - }, - "path": joint_path.name, - }, - "joint_decision": { - "inputs": { - "encoder": list(_tensor_shape(encoder_ref)), - "decoder": list(_tensor_shape(decoder_ref)), - }, - "outputs": { - "token_id": [ - _tensor_shape(encoder_ref)[0], - _tensor_shape(encoder_ref)[2], - _tensor_shape(decoder_ref)[2], - ], - "token_prob": [ - _tensor_shape(encoder_ref)[0], - _tensor_shape(encoder_ref)[2], - _tensor_shape(decoder_ref)[2], - ], - "duration": [ - _tensor_shape(encoder_ref)[0], - _tensor_shape(encoder_ref)[2], - _tensor_shape(decoder_ref)[2], - ], - }, - "path": joint_decision_path.name, - }, - "joint_decision_single_step": { - "inputs": { - "encoder_step": [1, int(encoder_ref.shape[1]), 1], - "decoder_step": [1, int(decoder_ref.shape[1]), 1], - }, - "outputs": { - "token_id": [1, 1, 1], - "token_prob": [1, 1, 1], - "duration": [1, 1, 1], - "top_k_ids": [1, 1, 1, 64], - "top_k_logits": [1, 1, 1, 64], - }, - "path": jd_single_path.name, - }, - }, - } - - metadata_path = output_dir / "metadata.json" - metadata_path.write_text(json.dumps(metadata, indent=2)) - typer.echo(f"Export complete. Metadata written to {metadata_path}") - - finally: - asr_model.decoder._rnnt_export = decoder_export_flag - - -if __name__ == "__main__": - app() diff --git a/parakeet-tdt-ctc-110m/coreml/hybrid_earnings_benchmark.json b/parakeet-tdt-ctc-110m/coreml/hybrid_earnings_benchmark.json deleted file mode 100644 index 3dd6bc7551d07ef8a966ba06590d2454e4324983..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/hybrid_earnings_benchmark.json +++ /dev/null @@ -1,35 +0,0 @@ -{ - "approach" : "single-encoder", - "model" : "parakeet-tdt-ctc-110m-hybrid", - "results" : [ - { - "audioLength" : 15, - "ctcDetections" : [ - { - "endTime" : 6.0800000000000001, - "inReference" : true, - "score" : -8.3699999999999992, - "source" : "ctc", - "startTime" : 4.96, - "word" : "LATAM Airlines" - } - ], - "dictFound" : 1, - "dictTotal" : 1, - "fileId" : "4329526_chunk0", - "hypothesis" : "goodday everyone and welcome to latam airlines group earnings release confonference call just as a reminder this conference is being recorded lat tam airlines group eararnings releaseed for the", - "processingTime" : 0.070000000000000007, - "reference" : "good day everyone and welcome to latam airlines group earnings release conference call just as a reminder this conference is being recorded latam airlines group earnings released for the", - "wer" : 24.140000000000001 - } - ], - "summary" : { - "avgWer" : 24.140000000000001, - "dictPass" : 1, - "dictRate" : 100, - "dictTotal" : 1, - "totalAudioDuration" : 15, - "totalProcessingTime" : 0.070000000000000007, - "totalTests" : 1 - } -} \ No newline at end of file diff --git a/parakeet-tdt-ctc-110m/coreml/individual_components.py b/parakeet-tdt-ctc-110m/coreml/individual_components.py deleted file mode 100644 index 0f6ae75f46ab359edc5c6dd6799dccf0b04a7ed7..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/individual_components.py +++ /dev/null @@ -1,265 +0,0 @@ -#!/usr/bin/env python3 -"""Export Parakeet TDT-CTC 110M Hybrid RNNT components into CoreML and validate outputs.""" -from __future__ import annotations - -from dataclasses import dataclass -from pathlib import Path -from typing import Optional, Tuple - -import coremltools as ct -import torch - - -@dataclass -class ExportSettings: - output_dir: Path - compute_units: ct.ComputeUnit - deployment_target: Optional[ct.target.iOS17] - compute_precision: Optional[ct.precision] - max_audio_seconds: float - max_symbol_steps: int - - -@dataclass -class ValidationSettings: - audio_path: Optional[Path] - seconds: float - seed: Optional[int] - rtol: float - atol: float - skip: bool - - -@dataclass -class ValidationDiff: - name: str - max_abs_diff: float - max_rel_diff: float - - -@dataclass -class ValidationResult: - source: str - audio_num_samples: int - audio_seconds: float - token_length: int - atol: float - rtol: float - diffs: Tuple[ValidationDiff, ...] - - -class PreprocessorWrapper(torch.nn.Module): - def __init__(self, module: torch.nn.Module) -> None: - super().__init__() - self.module = module - - def forward(self, audio_signal: torch.Tensor, length: torch.Tensor) -> Tuple[torch.Tensor, torch.Tensor]: - mel, mel_length = self.module(input_signal=audio_signal, length=length.to(dtype=torch.long)) - return mel, mel_length - - -class EncoderWrapper(torch.nn.Module): - def __init__(self, module: torch.nn.Module) -> None: - super().__init__() - self.module = module - - def forward(self, features: torch.Tensor, length: torch.Tensor) -> Tuple[torch.Tensor, torch.Tensor]: - encoded, encoded_lengths = self.module(audio_signal=features, length=length.to(dtype=torch.long)) - return encoded, encoded_lengths - - -class DecoderWrapper(torch.nn.Module): - def __init__(self, module: torch.nn.Module) -> None: - super().__init__() - self.module = module - - def forward( - self, - targets: torch.Tensor, - target_lengths: torch.Tensor, - h_in: torch.Tensor, - c_in: torch.Tensor, - ) -> Tuple[torch.Tensor, torch.Tensor, torch.Tensor]: - state = [h_in, c_in] - decoder_output, _, new_state = self.module( - targets=targets.to(dtype=torch.long), - target_length=target_lengths.to(dtype=torch.long), - states=state, - ) - return decoder_output, new_state[0], new_state[1] - - -class JointWrapper(torch.nn.Module): - """Joint network for 110M hybrid model. - - Note: The 110M model has encoder_dim=512 and decoder_dim=640. - The joint network projects both to 640, then combines them. - """ - def __init__(self, module: torch.nn.Module) -> None: - super().__init__() - self.module = module - - def forward(self, encoder_outputs: torch.Tensor, decoder_outputs: torch.Tensor) -> torch.Tensor: - # Input: encoder_outputs [B, D_enc, T], decoder_outputs [B, D_dec, U] - # For 110M: D_enc=512, D_dec=640 - # Transpose to match what projection layers expect - encoder_outputs = encoder_outputs.transpose(1, 2) # [B, T, D_enc] - decoder_outputs = decoder_outputs.transpose(1, 2) # [B, U, D_dec] - - # Apply projections - enc_proj = self.module.enc(encoder_outputs) # [B, T, 640] - dec_proj = self.module.pred(decoder_outputs) # [B, U, 640] - - # Explicit broadcasting along T and U to avoid converter ambiguity - x = enc_proj.unsqueeze(2) + dec_proj.unsqueeze(1) # [B, T, U, 640] - x = self.module.joint_net[0](x) # ReLU - x = self.module.joint_net[1](x) # Dropout (no-op in eval) - out = self.module.joint_net[2](x) # Linear -> logits [B, T, U, vocab+1+durations] - return out - - -class CTCHeadWrapper(torch.nn.Module): - """CTC decoder head for 110M hybrid model. - - Takes encoder output and produces log probabilities over vocabulary. - The NeMo CTC decoder (ConvASRDecoder) uses Conv1d so it expects [B, D, T] format. - """ - def __init__(self, module: torch.nn.Module) -> None: - super().__init__() - self.module = module - - def forward(self, encoder_outputs: torch.Tensor) -> torch.Tensor: - # Input: encoder_outputs [B, D_enc, T] - already in the format CTC decoder expects - # The NeMo CTC decoder uses Conv1d internally, so it expects [B, D, T] - # Output: log probabilities [B, T, vocab+1] - log_probs = self.module(encoder_output=encoder_outputs) - return log_probs - - -class MelEncoderWrapper(torch.nn.Module): - """Fused wrapper: waveform -> mel -> encoder. - - Inputs: - - audio_signal: [B, S] - - audio_length: [B] - - Outputs: - - encoder: [B, D, T_enc] - - encoder_length: [B] - """ - def __init__(self, preprocessor: PreprocessorWrapper, encoder: EncoderWrapper) -> None: - super().__init__() - self.preprocessor = preprocessor - self.encoder = encoder - - def forward(self, audio_signal: torch.Tensor, audio_length: torch.Tensor) -> Tuple[torch.Tensor, torch.Tensor]: - mel, mel_length = self.preprocessor(audio_signal, audio_length) - encoded, enc_len = self.encoder(mel, mel_length.to(dtype=torch.int32)) - return encoded, enc_len - - -class JointDecisionWrapper(torch.nn.Module): - """Joint + decision head: outputs label id, label prob, duration frames. - - Splits joint logits into token logits and duration logits, applies softmax - over tokens, argmax for both heads, and gathers probability of the chosen token. - - Inputs: - - encoder_outputs: [B, D, T] - - decoder_outputs: [B, D, U] - - Returns: - - token_id: [B, T, U] int32 - - token_prob: [B, T, U] float32 - - duration: [B, T, U] int32 (frames; for v3 bins=[0,1,2,3,4]) - """ - def __init__(self, joint: JointWrapper, vocab_size: int, num_extra: int) -> None: - super().__init__() - self.joint = joint - self.vocab_with_blank = int(vocab_size) + 1 - self.num_extra = int(num_extra) - - def forward(self, encoder_outputs: torch.Tensor, decoder_outputs: torch.Tensor): - logits = self.joint(encoder_outputs, decoder_outputs) - token_logits = logits[..., : self.vocab_with_blank] - duration_logits = logits[..., -self.num_extra :] - - # Token selection - token_ids = torch.argmax(token_logits, dim=-1).to(dtype=torch.int32) - token_probs_all = torch.softmax(token_logits, dim=-1) - # gather expects int64 (long) indices; cast only for gather - token_prob = torch.gather( - token_probs_all, dim=-1, index=token_ids.long().unsqueeze(-1) - ).squeeze(-1) - - # Duration prediction (bins are identity mapping to frames for v3) - duration = torch.argmax(duration_logits, dim=-1).to(dtype=torch.int32) - return token_ids, token_prob, duration - - -class JointDecisionSingleStep(torch.nn.Module): - """Single-step variant for streaming: encoder_step [1, 512, 1] -> [1,1,1]. - - Note: For 110M model, encoder_dim is 512 (not 1024 like 0.6B). - - Inputs: - - encoder_step: [B=1, D=512, T=1] - - decoder_step: [B=1, D=640, U=1] - - Returns: - - token_id: [1, 1, 1] int32 - - token_prob: [1, 1, 1] float32 - - duration: [1, 1, 1] int32 - - top_k_ids: [1, 1, 1, K] int32 - - top_k_logits: [1, 1, 1, K] float32 - """ - def __init__(self, joint: JointWrapper, vocab_size: int, num_extra: int, top_k: int = 64) -> None: - super().__init__() - self.joint = joint - self.vocab_with_blank = int(vocab_size) + 1 - self.num_extra = int(num_extra) - # Emit top-K candidates to enable host-side re-ranking with contextual biasing - self.top_k = int(top_k) - - def forward(self, encoder_step: torch.Tensor, decoder_step: torch.Tensor): - # Reuse JointWrapper which expects [B, D, T] and [B, D, U] - logits = self.joint(encoder_step, decoder_step) # [1, 1, 1, V+extra] - token_logits = logits[..., : self.vocab_with_blank] - duration_logits = logits[..., -self.num_extra :] - - token_ids = torch.argmax(token_logits, dim=-1, keepdim=False).to(dtype=torch.int32) - token_probs_all = torch.softmax(token_logits, dim=-1) - token_prob = torch.gather( - token_probs_all, dim=-1, index=token_ids.long().unsqueeze(-1) - ).squeeze(-1) - duration = torch.argmax(duration_logits, dim=-1, keepdim=False).to(dtype=torch.int32) - - # Also expose top-K candidates for host-side re-ranking. - # Shapes preserved as [1, 1, 1, K] to match CoreML broadcasting expectations. - # Note: topk expects last dimension; original shape is [1, 1, 1, V]. - topk_logits, topk_ids_long = torch.topk(token_logits, k=min(self.top_k, token_logits.shape[-1]), dim=-1) - topk_ids = topk_ids_long.to(dtype=torch.int32) - return token_ids, token_prob, duration, topk_ids, topk_logits - - -def _coreml_convert( - traced: torch.jit.ScriptModule, - inputs, - outputs, - settings: ExportSettings, - compute_units_override: Optional[ct.ComputeUnit] = None, -) -> ct.models.MLModel: - cu = compute_units_override if compute_units_override is not None else settings.compute_units - kwargs = { - "convert_to": "mlprogram", - "inputs": inputs, - "outputs": outputs, - "compute_units": cu, - } - print("Converting:", traced.__class__.__name__) - print("Conversion kwargs:", kwargs) - if settings.deployment_target is not None: - kwargs["minimum_deployment_target"] = settings.deployment_target - if settings.compute_precision is not None: - kwargs["compute_precision"] = settings.compute_precision - return ct.convert(traced, **kwargs) diff --git a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/metadata.json b/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/metadata.json deleted file mode 100644 index 6db872ae8a7d463c95dc180210ace0fce019f4a6..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/metadata.json +++ /dev/null @@ -1,247 +0,0 @@ -{ - "model_id": "nvidia/parakeet-tdt_ctc-110m", - "model_type": "hybrid_rnnt_ctc", - "sample_rate": 16000, - "max_audio_seconds": 15.0, - "max_audio_samples": 240000, - "max_symbol_steps": 1, - "vocab_size": 1024, - "joint_extra_outputs": 5, - "encoder_dim": 512, - "decoder_dim": 640, - "decoder_hidden": 640, - "decoder_layers": 1, - "blank_id": 1024, - "checkpoint": { - "type": "pretrained", - "model_id": "nvidia/parakeet-tdt_ctc-110m" - }, - "coreml": { - "compute_units": "CPU_ONLY", - "compute_precision": "FLOAT32" - }, - "components": { - "preprocessor": { - "inputs": { - "audio_signal": [ - 1, - 240000 - ], - "audio_length": [ - 1 - ] - }, - "outputs": { - "mel": [ - 1, - 80, - 1501 - ], - "mel_length": [ - 1 - ] - }, - "path": "parakeet_preprocessor.mlpackage" - }, - "encoder": { - "inputs": { - "mel": [ - 1, - 80, - 1501 - ], - "mel_length": [ - 1 - ] - }, - "outputs": { - "encoder": [ - 1, - 512, - 188 - ], - "encoder_length": [ - 1 - ] - }, - "path": "parakeet_encoder.mlpackage" - }, - "ctc_head": { - "inputs": { - "encoder": [ - 1, - 512, - 188 - ] - }, - "outputs": { - "log_probs": [ - 1, - 188, - 1025 - ] - }, - "path": "parakeet_ctc_head.mlpackage" - }, - "mel_encoder": { - "inputs": { - "audio_signal": [ - 1, - 240000 - ], - "audio_length": [ - 1 - ] - }, - "outputs": { - "encoder": [ - 1, - 512, - 188 - ], - "encoder_length": [ - 1 - ] - }, - "path": "parakeet_mel_encoder.mlpackage" - }, - "decoder": { - "inputs": { - "targets": [ - 1, - 1 - ], - "target_length": [ - 1 - ], - "h_in": [ - 1, - 1, - 640 - ], - "c_in": [ - 1, - 1, - 640 - ] - }, - "outputs": { - "decoder": [ - 1, - 640, - 1 - ], - "h_out": [ - 1, - 1, - 640 - ], - "c_out": [ - 1, - 1, - 640 - ] - }, - "path": "parakeet_decoder.mlpackage" - }, - "joint": { - "inputs": { - "encoder": [ - 1, - 512, - 188 - ], - "decoder": [ - 1, - 640, - 1 - ] - }, - "outputs": { - "logits": [ - 1, - 188, - 1, - 1030 - ] - }, - "path": "parakeet_joint.mlpackage" - }, - "joint_decision": { - "inputs": { - "encoder": [ - 1, - 512, - 188 - ], - "decoder": [ - 1, - 640, - 1 - ] - }, - "outputs": { - "token_id": [ - 1, - 188, - 1 - ], - "token_prob": [ - 1, - 188, - 1 - ], - "duration": [ - 1, - 188, - 1 - ] - }, - "path": "parakeet_joint_decision.mlpackage" - }, - "joint_decision_single_step": { - "inputs": { - "encoder_step": [ - 1, - 512, - 1 - ], - "decoder_step": [ - 1, - 640, - 1 - ] - }, - "outputs": { - "token_id": [ - 1, - 1, - 1 - ], - "token_prob": [ - 1, - 1, - 1 - ], - "duration": [ - 1, - 1, - 1 - ], - "top_k_ids": [ - 1, - 1, - 1, - 64 - ], - "top_k_logits": [ - 1, - 1, - 1, - 64 - ] - }, - "path": "parakeet_joint_decision_single_step.mlpackage" - } - } -} \ No newline at end of file diff --git a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_ctc_head.mlpackage/Data/com.apple.CoreML/model.mlmodel b/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_ctc_head.mlpackage/Data/com.apple.CoreML/model.mlmodel deleted file mode 100644 index b807d0d0234624c649b82a7bd3e36bd9095fa961..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_ctc_head.mlpackage/Data/com.apple.CoreML/model.mlmodel +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:6459b9564e0630f2eec300eb732fceccbc1d2d16f12cb0694ce310d84bfbecf2 -size 3366 diff --git a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_ctc_head.mlpackage/Data/com.apple.CoreML/weights/weight.bin b/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_ctc_head.mlpackage/Data/com.apple.CoreML/weights/weight.bin deleted file mode 100644 index 23cedd570125191064c6111997f08c48898245f7..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_ctc_head.mlpackage/Data/com.apple.CoreML/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:fb9bead064427ffcb7529c0e3f378e421b4dde8e6d81447b6d1ca3352ca850e1 -size 1051842 diff --git a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_ctc_head.mlpackage/Manifest.json b/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_ctc_head.mlpackage/Manifest.json deleted file mode 100644 index 6fa9ee67551a75a4b78fabbd02c58c0b39b247ed..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_ctc_head.mlpackage/Manifest.json +++ /dev/null @@ -1,18 +0,0 @@ -{ - "fileFormatVersion": "1.0.0", - "itemInfoEntries": { - "6651E3CE-C3ED-4267-AAC3-5A772FC3515A": { - "author": "com.apple.CoreML", - "description": "CoreML Model Specification", - "name": "model.mlmodel", - "path": "com.apple.CoreML/model.mlmodel" - }, - "A3F7798B-67CA-418C-B8BB-58731D3A413F": { - "author": "com.apple.CoreML", - "description": "CoreML Model Weights", - "name": "weights", - "path": "com.apple.CoreML/weights" - } - }, - "rootModelIdentifier": "6651E3CE-C3ED-4267-AAC3-5A772FC3515A" -} diff --git a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_decoder.mlpackage/Data/com.apple.CoreML/model.mlmodel b/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_decoder.mlpackage/Data/com.apple.CoreML/model.mlmodel deleted file mode 100644 index 44a413afd31f449824ec8abc34caed9f15d411c3..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_decoder.mlpackage/Data/com.apple.CoreML/model.mlmodel +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:a05548eb455c5cd564782b125a5f9279a789be1f4141e5f044453ea79cd68b47 -size 6729 diff --git a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_decoder.mlpackage/Data/com.apple.CoreML/weights/weight.bin b/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_decoder.mlpackage/Data/com.apple.CoreML/weights/weight.bin deleted file mode 100644 index 4543ad4ea89b21ebed0de0d5ec3975d07eace7fb..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_decoder.mlpackage/Data/com.apple.CoreML/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:dd90b58597ee2c172c672dffe13b1110898ba07394c1a15efc96cc8c6b18411b -size 7871040 diff --git a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_decoder.mlpackage/Manifest.json b/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_decoder.mlpackage/Manifest.json deleted file mode 100644 index f969ba20230983a15c01e6a386db029d1a892816..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_decoder.mlpackage/Manifest.json +++ /dev/null @@ -1,18 +0,0 @@ -{ - "fileFormatVersion": "1.0.0", - "itemInfoEntries": { - "83E7B87A-4EBE-48BF-BF3C-EE74DEA4C7AF": { - "author": "com.apple.CoreML", - "description": "CoreML Model Specification", - "name": "model.mlmodel", - "path": "com.apple.CoreML/model.mlmodel" - }, - "98BF03AC-26AF-410B-95AC-C9B99B3B240C": { - "author": "com.apple.CoreML", - "description": "CoreML Model Weights", - "name": "weights", - "path": "com.apple.CoreML/weights" - } - }, - "rootModelIdentifier": "83E7B87A-4EBE-48BF-BF3C-EE74DEA4C7AF" -} diff --git a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_encoder.mlpackage/Data/com.apple.CoreML/model.mlmodel b/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_encoder.mlpackage/Data/com.apple.CoreML/model.mlmodel deleted file mode 100644 index ecd864b1e23052bed6937545ee41e528f892980c..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_encoder.mlpackage/Data/com.apple.CoreML/model.mlmodel +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:70d7747b57beba0248fabb6cbfa5d276e3604d0d7e234f4ccb578ea0a4d25110 -size 508107 diff --git a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_encoder.mlpackage/Data/com.apple.CoreML/weights/weight.bin b/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_encoder.mlpackage/Data/com.apple.CoreML/weights/weight.bin deleted file mode 100644 index c5bffa9667270bb21c093e55898f4fb2e299e426..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_encoder.mlpackage/Data/com.apple.CoreML/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:cecf7994b2758397d992802a4f6e5d656e3a1aeb7bbedc2aa430b1316d62474c -size 215143424 diff --git a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_encoder.mlpackage/Manifest.json b/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_encoder.mlpackage/Manifest.json deleted file mode 100644 index 5da111c5105a7bcf84de61da5c96b86c946acf57..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_encoder.mlpackage/Manifest.json +++ /dev/null @@ -1,18 +0,0 @@ -{ - "fileFormatVersion": "1.0.0", - "itemInfoEntries": { - "82FE91E6-784B-4D60-921C-8CAE58385192": { - "author": "com.apple.CoreML", - "description": "CoreML Model Weights", - "name": "weights", - "path": "com.apple.CoreML/weights" - }, - "DCA53FEB-DFA1-45B9-9E10-E5D3F5C04877": { - "author": "com.apple.CoreML", - "description": "CoreML Model Specification", - "name": "model.mlmodel", - "path": "com.apple.CoreML/model.mlmodel" - } - }, - "rootModelIdentifier": "DCA53FEB-DFA1-45B9-9E10-E5D3F5C04877" -} diff --git a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint.mlpackage/Data/com.apple.CoreML/model.mlmodel b/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint.mlpackage/Data/com.apple.CoreML/model.mlmodel deleted file mode 100644 index 40cb45ad4abce1672fdfcd80332f638fe690d4bc..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint.mlpackage/Data/com.apple.CoreML/model.mlmodel +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:7d115951cfb00e12b26f4b1a153e39dceffad34f5c940ede2ddb92d7a21859aa -size 4548 diff --git a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint.mlpackage/Data/com.apple.CoreML/weights/weight.bin b/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint.mlpackage/Data/com.apple.CoreML/weights/weight.bin deleted file mode 100644 index 747ec9d24e2d3e008bc869d2c404bb3d2653325c..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint.mlpackage/Data/com.apple.CoreML/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:b3f771cb65b190f1873e39629676ed79b65a8361522f451b37bdba8b1106e6ff -size 2798028 diff --git a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint.mlpackage/Manifest.json b/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint.mlpackage/Manifest.json deleted file mode 100644 index 51666c08cb1568e77793f3f8b7df68a1ab931cd8..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint.mlpackage/Manifest.json +++ /dev/null @@ -1,18 +0,0 @@ -{ - "fileFormatVersion": "1.0.0", - "itemInfoEntries": { - "16695A4B-7F7A-41C7-9F44-8306983C0AA4": { - "author": "com.apple.CoreML", - "description": "CoreML Model Specification", - "name": "model.mlmodel", - "path": "com.apple.CoreML/model.mlmodel" - }, - "E3F73234-AB5F-41AB-9362-57ADB67197EA": { - "author": "com.apple.CoreML", - "description": "CoreML Model Weights", - "name": "weights", - "path": "com.apple.CoreML/weights" - } - }, - "rootModelIdentifier": "16695A4B-7F7A-41C7-9F44-8306983C0AA4" -} diff --git a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint_decision.mlpackage/Data/com.apple.CoreML/model.mlmodel b/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint_decision.mlpackage/Data/com.apple.CoreML/model.mlmodel deleted file mode 100644 index 49e6f595a1328fbcf9ffafd61b2de9c7b27f5b17..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint_decision.mlpackage/Data/com.apple.CoreML/model.mlmodel +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:02bb6ce108de2737a945f3d7aa71db5bd710ff75282565ba20741d82000347d6 -size 8711 diff --git a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint_decision.mlpackage/Data/com.apple.CoreML/weights/weight.bin b/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint_decision.mlpackage/Data/com.apple.CoreML/weights/weight.bin deleted file mode 100644 index 747ec9d24e2d3e008bc869d2c404bb3d2653325c..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint_decision.mlpackage/Data/com.apple.CoreML/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:b3f771cb65b190f1873e39629676ed79b65a8361522f451b37bdba8b1106e6ff -size 2798028 diff --git a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint_decision.mlpackage/Manifest.json b/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint_decision.mlpackage/Manifest.json deleted file mode 100644 index c00fe89cce0df115b26086dfee3c450663c06e4f..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint_decision.mlpackage/Manifest.json +++ /dev/null @@ -1,18 +0,0 @@ -{ - "fileFormatVersion": "1.0.0", - "itemInfoEntries": { - "0A2C8DEE-CF15-4176-9EF2-786EF0484092": { - "author": "com.apple.CoreML", - "description": "CoreML Model Weights", - "name": "weights", - "path": "com.apple.CoreML/weights" - }, - "2EF6A0AB-12F9-4ED0-84E2-94A0D744E07D": { - "author": "com.apple.CoreML", - "description": "CoreML Model Specification", - "name": "model.mlmodel", - "path": "com.apple.CoreML/model.mlmodel" - } - }, - "rootModelIdentifier": "2EF6A0AB-12F9-4ED0-84E2-94A0D744E07D" -} diff --git a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint_decision_single_step.mlpackage/Data/com.apple.CoreML/model.mlmodel b/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint_decision_single_step.mlpackage/Data/com.apple.CoreML/model.mlmodel deleted file mode 100644 index 24daa1ae01d93329a3e31b992186474b6521c4e6..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint_decision_single_step.mlpackage/Data/com.apple.CoreML/model.mlmodel +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:6e47db8f5e51f21028004c1fe76f4b2f1430ccf82b9613fe7dbe9d087478aeba -size 10620 diff --git a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint_decision_single_step.mlpackage/Data/com.apple.CoreML/weights/weight.bin b/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint_decision_single_step.mlpackage/Data/com.apple.CoreML/weights/weight.bin deleted file mode 100644 index 747ec9d24e2d3e008bc869d2c404bb3d2653325c..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint_decision_single_step.mlpackage/Data/com.apple.CoreML/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:b3f771cb65b190f1873e39629676ed79b65a8361522f451b37bdba8b1106e6ff -size 2798028 diff --git a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint_decision_single_step.mlpackage/Manifest.json b/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint_decision_single_step.mlpackage/Manifest.json deleted file mode 100644 index a3794d1952292c241c090386207789e999e901f5..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_joint_decision_single_step.mlpackage/Manifest.json +++ /dev/null @@ -1,18 +0,0 @@ -{ - "fileFormatVersion": "1.0.0", - "itemInfoEntries": { - "BDFCDB9F-5819-45ED-B191-3F50652482FC": { - "author": "com.apple.CoreML", - "description": "CoreML Model Weights", - "name": "weights", - "path": "com.apple.CoreML/weights" - }, - "CAC241D9-FB01-43BD-AFA9-1285CC5E9E24": { - "author": "com.apple.CoreML", - "description": "CoreML Model Specification", - "name": "model.mlmodel", - "path": "com.apple.CoreML/model.mlmodel" - } - }, - "rootModelIdentifier": "CAC241D9-FB01-43BD-AFA9-1285CC5E9E24" -} diff --git a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_mel_encoder.mlpackage/Data/com.apple.CoreML/model.mlmodel b/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_mel_encoder.mlpackage/Data/com.apple.CoreML/model.mlmodel deleted file mode 100644 index 64a97d13c860d2e7021389f6e3108ad2d16d8a1f..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_mel_encoder.mlpackage/Data/com.apple.CoreML/model.mlmodel +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:aea604d3edc6e879e525d8ec2e0eb8413c5d3d4a9fae86233cbe4ba23d89a258 -size 1241478 diff --git a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_mel_encoder.mlpackage/Data/com.apple.CoreML/weights/weight.bin b/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_mel_encoder.mlpackage/Data/com.apple.CoreML/weights/weight.bin deleted file mode 100644 index 6b1f64e4316b232d108b8ca7402ffb8071a98a93..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_mel_encoder.mlpackage/Data/com.apple.CoreML/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:a1c90b88b52667ee4f349b39994003cc2a9fb1b12c310752399d90c880286a7b -size 215951360 diff --git a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_mel_encoder.mlpackage/Manifest.json b/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_mel_encoder.mlpackage/Manifest.json deleted file mode 100644 index adbaa38b4e0eda5372704a0880329c85f2729354..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_mel_encoder.mlpackage/Manifest.json +++ /dev/null @@ -1,18 +0,0 @@ -{ - "fileFormatVersion": "1.0.0", - "itemInfoEntries": { - "BCB3EEA3-98D8-45A1-9DF8-D4B55B9897A5": { - "author": "com.apple.CoreML", - "description": "CoreML Model Weights", - "name": "weights", - "path": "com.apple.CoreML/weights" - }, - "E71B04CC-8B97-4E78-A14F-E65CA1517841": { - "author": "com.apple.CoreML", - "description": "CoreML Model Specification", - "name": "model.mlmodel", - "path": "com.apple.CoreML/model.mlmodel" - } - }, - "rootModelIdentifier": "E71B04CC-8B97-4E78-A14F-E65CA1517841" -} diff --git a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_preprocessor.mlpackage/Data/com.apple.CoreML/model.mlmodel b/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_preprocessor.mlpackage/Data/com.apple.CoreML/model.mlmodel deleted file mode 100644 index 888c5ddc608dc69393977dc0334df8848b29537f..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_preprocessor.mlpackage/Data/com.apple.CoreML/model.mlmodel +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:04c6e8c1d678969c477d45d0bf8b231aa83c3a7ca0892d7820f701d18392e777 -size 25277 diff --git a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_preprocessor.mlpackage/Data/com.apple.CoreML/weights/weight.bin b/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_preprocessor.mlpackage/Data/com.apple.CoreML/weights/weight.bin deleted file mode 100644 index d0c6ac87e5f78315662149af9ab4ea3658497dbe..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_preprocessor.mlpackage/Data/com.apple.CoreML/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:c062338de852a26607ce4101f74e6895de3a4134a57b07232bd72bfc6f1d7f1a -size 567712 diff --git a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_preprocessor.mlpackage/Manifest.json b/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_preprocessor.mlpackage/Manifest.json deleted file mode 100644 index f490745b3939f47130ed35b5ed897f1ec76acb45..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/parakeet_preprocessor.mlpackage/Manifest.json +++ /dev/null @@ -1,18 +0,0 @@ -{ - "fileFormatVersion": "1.0.0", - "itemInfoEntries": { - "4EC5BFFD-7A9E-42E4-86D0-365412CFD8E5": { - "author": "com.apple.CoreML", - "description": "CoreML Model Weights", - "name": "weights", - "path": "com.apple.CoreML/weights" - }, - "7A405622-5F02-4930-AB77-D20FA3B49A18": { - "author": "com.apple.CoreML", - "description": "CoreML Model Specification", - "name": "model.mlmodel", - "path": "com.apple.CoreML/model.mlmodel" - } - }, - "rootModelIdentifier": "7A405622-5F02-4930-AB77-D20FA3B49A18" -} diff --git a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/vocab.json b/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/vocab.json deleted file mode 100644 index b3436e3ca1609c34360bbdf9fa4dd50cd3eb1a3d..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/coreml/parakeet_110m_coreml/vocab.json +++ /dev/null @@ -1,1030 +0,0 @@ -{ - "vocab_size": 1024, - "blank_id": 1024, - "tokens": [ - "", - "▁t", - "▁th", - "▁a", - "in", - "re", - "▁the", - "▁w", - "▁s", - "▁o", - "er", - "ou", - "at", - "nd", - "it", - "▁h", - "▁c", - "▁b", - "is", - "en", - "on", - "ing", - "▁f", - "▁to", - "▁m", - "es", - "▁p", - "or", - "an", - "▁d", - "ll", - "▁I", - "ed", - "▁and", - "▁l", - "▁of", - "▁in", - "▁y", - "ar", - "▁g", - "▁you", - "as", - "om", - "▁n", - "ve", - "▁that", - "le", - "ic", - "us", - "ow", - "et", - "al", - "▁e", - "ut", - "▁it", - "ot", - "▁be", - "▁T", - "ion", - "▁is", - "▁wh", - "▁re", - "▁on", - "▁we", - "ent", - "▁A", - "ay", - "▁ha", - "▁Th", - "id", - "▁S", - "ac", - "gh", - "ver", - "ke", - "▁for", - "im", - "ly", - "ur", - "ld", - "▁he", - "▁st", - "all", - "ro", - "st", - "se", - "ct", - "ith", - "ir", - "am", - "▁this", - "if", - "▁W", - "oo", - "ri", - "▁was", - "ght", - "▁u", - "▁with", - "ad", - "ch", - "▁se", - "▁k", - "▁an", - "▁The", - "▁li", - "▁do", - "▁B", - "▁have", - "▁as", - "th", - "▁are", - "▁sh", - "ust", - "ce", - "ally", - "ill", - "▁H", - "▁j", - "ter", - "▁go", - "▁And", - "ation", - "▁C", - "▁so", - "ome", - "▁not", - "op", - "il", - "ore", - "▁ne", - "▁can", - "▁me", - "▁at", - "ould", - "ant", - "▁M", - "▁like", - "ere", - "▁they", - "ra", - "ers", - "▁ab", - "▁de", - "▁kn", - "ge", - "▁Y", - "▁ch", - "ul", - "pp", - "▁or", - "▁al", - "▁con", - "▁com", - "ess", - "▁su", - "out", - "▁your", - "▁So", - "ate", - "▁one", - "▁all", - "▁ex", - "est", - "▁fr", - "▁just", - "▁pro", - "▁know", - "▁O", - "ain", - "▁but", - "ol", - "ive", - "▁v", - "use", - "very", - "art", - "qu", - "▁my", - "el", - "▁N", - "nt", - "▁It", - "▁what", - "ab", - "▁P", - "▁wor", - "▁out", - "▁there", - "▁up", - "um", - "▁from", - "pe", - "▁tw", - "▁r", - "and", - "ight", - "ort", - "un", - "▁L", - "ist", - "▁about", - "ide", - "ig", - "ake", - "▁D", - "em", - "os", - "king", - "rou", - "ind", - "our", - "res", - "▁We", - "▁get", - "▁E", - "▁G", - "ack", - "▁le", - "ity", - "od", - "▁F", - "ard", - "▁pl", - "▁our", - "▁int", - "ment", - "▁will", - "ies", - "▁by", - "ink", - "ca", - "▁if", - "red", - "her", - "ie", - "▁us", - "▁some", - "▁don", - "ven", - "ood", - "ast", - "▁R", - "▁his", - "▁tim", - "▁tr", - "▁more", - "ich", - "ous", - "ame", - "▁going", - "▁had", - "▁them", - "ook", - "▁pe", - "▁Wh", - "▁You", - "▁But", - "ine", - "▁here", - "▁would", - "cause", - "right", - "so", - "ost", - "ure", - "▁has", - "ect", - "▁think", - "▁fe", - "ong", - "▁see", - "▁when", - "▁who", - "▁were", - "▁really", - "▁their", - "▁want", - "one", - "ople", - "▁then", - "▁time", - "▁sa", - "ap", - "▁te", - "▁He", - "▁ye", - "ck", - "▁her", - "▁thing", - "▁right", - "▁which", - "itt", - "ice", - "act", - "▁people", - "ty", - "▁two", - "▁J", - "▁im", - "ther", - "ci", - "ose", - "▁cl", - "▁qu", - "▁man", - "▁also", - "ree", - "▁en", - "ud", - "▁how", - "reat", - "ak", - "hing", - "ag", - "▁any", - "ff", - "ace", - "per", - "▁because", - "▁very", - "own", - "▁ad", - "▁act", - "▁been", - "▁now", - "▁ag", - "▁into", - "▁comp", - "ars", - "ions", - "are", - "ite", - "iv", - "▁these", - "ays", - "ep", - "▁This", - "▁she", - "ans", - "ah", - "een", - "▁over", - "ry", - "▁lo", - "age", - "▁pr", - "▁sp", - "ue", - "▁co", - "ick", - "ber", - "▁did", - "ip", - "ach", - "▁back", - "▁no", - "▁cont", - "▁other", - "▁every", - "pt", - "▁need", - "▁him", - "▁U", - "▁In", - "▁work", - "irst", - "▁part", - "▁look", - "ittle", - "ble", - "iz", - "▁un", - "▁make", - "omet", - "nder", - "ish", - "na", - "▁little", - "▁off", - "▁than", - "▁got", - "ually", - "▁per", - "▁good", - "▁way", - "▁could", - "▁ac", - "▁imp", - "able", - "▁where", - "iff", - "▁That", - "▁res", - "ount", - "pl", - "ance", - "▁first", - "▁ro", - "▁pre", - "ass", - "▁say", - "int", - "ated", - "ire", - "uch", - "ase", - "▁somet", - "ound", - "▁down", - "▁diff", - "sel", - "▁gu", - "▁am", - "ress", - "▁lot", - "ence", - "▁dis", - "orm", - "ix", - "▁po", - "ving", - "enty", - "▁K", - "▁spe", - "und", - "he", - "▁much", - "▁ar", - "round", - "▁app", - "co", - "ark", - "▁new", - "ater", - "ult", - "end", - "▁even", - "▁start", - "ations", - "rough", - "ile", - "fter", - "▁well", - "be", - "▁They", - "▁three", - "ign", - "ild", - "▁said", - "ough", - "ang", - "▁too", - "ade", - "▁bl", - "ens", - "▁inc", - "ia", - "▁those", - "▁mo", - "▁take", - "▁through", - "▁fl", - "▁kind", - "▁things", - "▁bet", - "▁only", - "▁St", - "▁let", - "cess", - "▁Ch", - "ary", - "vel", - "▁If", - "xt", - "other", - "av", - "ical", - "ord", - "▁again", - "▁something", - "onna", - "fore", - "▁may", - "ting", - "▁bu", - "▁differe", - "urn", - "▁gonna", - "▁does", - "uct", - "og", - "▁twenty", - "▁gr", - "▁Ye", - "wn", - "▁should", - "▁comm", - "ition", - "▁under", - "▁hel", - "ory", - "▁fo", - "▁use", - "igh", - "ife", - "▁actually", - "▁tal", - "▁call", - "ents", - "ious", - "ull", - "▁There", - "▁Yeah", - "▁most", - "▁ke", - "ors", - "ved", - "ys", - "▁sc", - "▁happ", - "ope", - "▁help", - "atch", - "▁What", - "▁rem", - "ple", - "▁Now", - "▁br", - "ool", - "oth", - "▁four", - "self", - "▁str", - "ne", - "thing", - "▁put", - "ial", - "▁great", - "ail", - "ub", - "ning", - "▁sm", - "▁feel", - "▁five", - "ody", - "undred", - "iss", - "ank", - "get", - "aking", - "▁many", - "▁hundred", - "▁years", - "▁being", - "▁come", - "▁mean", - "ily", - "▁different", - "▁after", - "▁ser", - "▁show", - "form", - "ful", - "oy", - "▁six", - "▁vide", - "▁V", - "▁its", - "▁point", - "▁day", - "▁des", - "ons", - "▁bit", - "▁bel", - "▁before", - "▁aw", - "▁end", - "▁Oh", - "▁still", - "ath", - "▁long", - "▁'", - "ise", - "ob", - "day", - "▁add", - "ft", - "ves", - "ces", - "ady", - "▁cr", - "▁around", - "▁try", - "les", - "vers", - "kay", - "ian", - "ates", - "▁find", - "ward", - "▁As", - "▁eight", - "lic", - "▁same", - "▁pos", - "▁em", - "▁made", - "▁supp", - "▁life", - "▁Be", - "pect", - "▁dec", - "▁play", - "ange", - "▁att", - "▁pers", - "ways", - "▁high", - "▁hand", - "▁next", - "▁cons", - "▁own", - "▁inv", - "ower", - "▁ind", - "ert", - "ng", - "ave", - "▁year", - "▁big", - "ating", - "▁world", - "▁rel", - "▁sure", - "▁tra", - "ew", - "ered", - "▁fin", - "▁Well", - "▁sl", - "▁doing", - "bs", - "▁set", - "▁rec", - "ual", - "cial", - "▁ph", - "erm", - "▁love", - "ph", - "▁real", - "▁last", - "ict", - "▁bo", - "▁ra", - "ible", - "▁wr", - "mer", - "▁count", - "ities", - "▁always", - "inet", - "ments", - "uc", - "▁might", - "▁inter", - "▁video", - "gin", - "▁tell", - "▁never", - "vent", - "▁import", - "ied", - "▁sy", - "▁How", - "ically", - "ought", - "▁thir", - "▁rep", - "ks", - "ib", - "▁fam", - "ject", - "▁bas", - "▁She", - "▁give", - "akes", - "▁ninet", - "▁reg", - "▁min", - "▁op", - "▁def", - "▁didn", - "te", - "▁cour", - "▁why", - "▁ent", - "▁place", - "▁ins", - "▁car", - "ather", - "▁person", - "ular", - "▁inst", - "▁prod", - "lect", - "▁Al", - "▁today", - "▁bec", - "▁sur", - "▁All", - "▁another", - "▁bus", - "▁keep", - "ell", - "ese", - "riend", - "▁quest", - "▁talk", - "als", - "ings", - "▁mon", - "cond", - "old", - "▁acc", - "▁la", - "▁num", - "ident", - "▁che", - "iness", - "▁turn", - "▁ear", - "▁No", - "ousand", - "▁better", - "ific", - "▁loo", - "▁gl", - "oc", - "▁important", - "ited", - "▁An", - "▁thousand", - "ility", - "llow", - "▁used", - "▁gen", - "▁sim", - "li", - "▁happen", - "▁Un", - "▁Let", - "air", - "ock", - "ably", - "gg", - "▁watch", - "▁For", - "▁sw", - "ren", - "ute", - "ever", - "▁pol", - "▁sch", - "▁When", - "▁such", - "▁fif", - "▁home", - "▁cle", - "▁contin", - "ouse", - "▁friend", - "uring", - "▁Okay", - "gr", - "▁able", - "▁stud", - "▁eff", - "hip", - "body", - "▁top", - "ness", - "▁exper", - "▁pret", - "▁both", - "▁done", - "cri", - "▁mark", - "▁while", - "▁old", - "ros", - "ont", - "▁second", - "ative", - "▁thought", - "▁best", - "▁found", - "iew", - "▁belie", - "▁each", - "erest", - "▁tri", - "▁eas", - "▁ca", - "▁fact", - "▁care", - "▁fun", - "atter", - "ures", - "▁head", - "▁lear", - "▁water", - "▁hard", - "▁few", - "▁side", - "ween", - "▁exp", - "▁away", - "its", - "▁ext", - "lud", - "▁run", - "▁trans", - "ince", - "▁sk", - "▁open", - "cus", - "▁between", - "▁called", - "▁wee", - "▁pretty", - "ason", - "▁far", - "ember", - "omm", - "▁interest", - "any", - "ner", - "uff", - "▁pres", - "▁cur", - "▁child", - "ee", - "▁toget", - "▁together", - "olog", - "▁God", - "ond", - "▁char", - "▁looking", - "stem", - "az", - "cent", - "▁ob", - "▁ass", - "land", - "▁doesn", - "▁business", - "▁course", - "▁ten", - "ps", - "arch", - "ced", - "ms", - "ize", - "nce", - "▁ref", - "▁name", - "ross", - "▁grow", - "oney", - "▁went", - "ics", - "teen", - "▁cou", - "▁prob", - "▁ret", - "▁guys", - "▁came", - "ash", - "led", - "▁Eur", - "ues", - "▁ide", - "gan", - "▁everything", - "▁getting", - "▁ask", - "▁cor", - "▁build", - "▁sign", - "▁small", - "uck", - "▁el", - "▁col", - "▁Is", - "ational", - "stand", - "cy", - "▁conf", - "der", - "▁bre", - "▁cap", - "▁mod", - "ets", - "ike", - "▁number", - "▁comple", - "ertain", - "▁ever", - "▁coll", - "▁hum", - "▁Europe", - "▁cre", - "▁met", - "▁exam", - "▁move", - "▁pass", - "▁left", - "▁system", - "▁includ", - "▁Thank", - "cept", - "▁wom", - "▁product", - "ten", - "▁rest", - "▁probably", - "▁dri", - "▁Do", - "▁gener", - "▁anything", - "▁lar", - "▁My", - "▁school", - "▁lead", - "▁sub", - "▁ty", - "▁plan", - "▁seem", - "▁whole", - "irect", - "▁light", - "▁must", - "▁mom", - "▁opp", - "▁support", - "▁family", - "ices", - "amp", - "▁proble", - "▁dr", - "ready", - "▁using", - "ense", - "▁prov", - "ush", - "ax", - "▁power", - "▁Re", - "alth", - "▁ev", - "▁stand", - "▁war", - "ts", - "▁", - "e", - "t", - "o", - "a", - "n", - "i", - "s", - "r", - "h", - "l", - "d", - "u", - "c", - "m", - "y", - "g", - "w", - "f", - "p", - ".", - "b", - ",", - "v", - "k", - "'", - "I", - "T", - "A", - "S", - "x", - "W", - "j", - "B", - "C", - "H", - "?", - "M", - "O", - "Y", - "N", - "P", - "E", - "q", - "L", - "D", - "z", - "G", - "F", - "R", - "!", - "J", - "U", - "K", - "V", - "Q", - "Z", - "X" - ] -} \ No newline at end of file diff --git a/parakeet-tdt-ctc-110m/coreml/parakeet_ctc_coreml/.DS_Store b/parakeet-tdt-ctc-110m/coreml/parakeet_ctc_coreml/.DS_Store deleted file mode 100644 index 7b37950bb8ce3a439220b30b68fcd5a01d75cb8d..0000000000000000000000000000000000000000 Binary files a/parakeet-tdt-ctc-110m/coreml/parakeet_ctc_coreml/.DS_Store and /dev/null differ diff --git a/parakeet-tdt-ctc-110m/get-pip.py b/parakeet-tdt-ctc-110m/get-pip.py deleted file mode 100644 index 22f29f881487fe04461440ad6bc7b6a2e5e781de..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/get-pip.py +++ /dev/null @@ -1,27368 +0,0 @@ -#!/usr/bin/env python -# -# Hi There! -# -# You may be wondering what this giant blob of binary data here is, you might -# even be worried that we're up to something nefarious (good for you for being -# paranoid!). This is a base85 encoding of a zip file, this zip file contains -# an entire copy of pip (version 25.3). -# -# Pip is a thing that installs packages, pip itself is a package that someone -# might want to install, especially if they're looking to run this get-pip.py -# script. Pip has a lot of code to deal with the security of installing -# packages, various edge cases on various platforms, and other such sort of -# "tribal knowledge" that has been encoded in its code base. Because of this -# we basically include an entire copy of pip inside this blob. We do this -# because the alternatives are attempt to implement a "minipip" that probably -# doesn't do things correctly and has weird edge cases, or compress pip itself -# down into a single file. -# -# If you're wondering how this is created, it is generated using -# `scripts/generate.py` in https://github.com/pypa/get-pip. - -import sys - -this_python = sys.version_info[:2] -min_version = (3, 9) -if this_python < min_version: - message_parts = [ - "This script does not work on Python {}.{}.".format(*this_python), - "The minimum supported Python version is {}.{}.".format(*min_version), - "Please use https://bootstrap.pypa.io/pip/{}.{}/get-pip.py instead.".format(*this_python), - ] - print("ERROR: " + " ".join(message_parts)) - sys.exit(1) - - -import os.path -import pkgutil -import shutil -import tempfile -import argparse -import importlib -from base64 import b85decode - - -def include_setuptools(args): - """ - Install setuptools only if absent, not excluded and when using Python <3.12. - """ - cli = not args.no_setuptools - env = not os.environ.get("PIP_NO_SETUPTOOLS") - absent = not importlib.util.find_spec("setuptools") - python_lt_3_12 = this_python < (3, 12) - return cli and env and absent and python_lt_3_12 - - -def include_wheel(args): - """ - Install wheel only if absent, not excluded and when using Python <3.12. - """ - cli = not args.no_wheel - env = not os.environ.get("PIP_NO_WHEEL") - absent = not importlib.util.find_spec("wheel") - python_lt_3_12 = this_python < (3, 12) - return cli and env and absent and python_lt_3_12 - - -def determine_pip_install_arguments(): - pre_parser = argparse.ArgumentParser() - pre_parser.add_argument("--no-setuptools", action="store_true") - pre_parser.add_argument("--no-wheel", action="store_true") - pre, args = pre_parser.parse_known_args() - - args.append("pip") - - if include_setuptools(pre): - args.append("setuptools") - - if include_wheel(pre): - args.append("wheel") - - return ["install", "--upgrade", "--force-reinstall"] + args - - -def monkeypatch_for_cert(tmpdir): - """Patches `pip install` to provide default certificate with the lowest priority. - - This ensures that the bundled certificates are used unless the user specifies a - custom cert via any of pip's option passing mechanisms (config, env-var, CLI). - - A monkeypatch is the easiest way to achieve this, without messing too much with - the rest of pip's internals. - """ - from pip._internal.commands.install import InstallCommand - - # We want to be using the internal certificates. - cert_path = os.path.join(tmpdir, "cacert.pem") - with open(cert_path, "wb") as cert: - cert.write(pkgutil.get_data("pip._vendor.certifi", "cacert.pem")) - - install_parse_args = InstallCommand.parse_args - - def cert_parse_args(self, args): - if not self.parser.get_default_values().cert: - # There are no user provided cert -- force use of bundled cert - self.parser.defaults["cert"] = cert_path # calculated above - return install_parse_args(self, args) - - InstallCommand.parse_args = cert_parse_args - - -def bootstrap(tmpdir): - monkeypatch_for_cert(tmpdir) - - # Execute the included pip and use it to install the latest pip and - # any user-requested packages from PyPI. - from pip._internal.cli.main import main as pip_entry_point - args = determine_pip_install_arguments() - sys.exit(pip_entry_point(args)) - - -def main(): - tmpdir = None - try: - # Create a temporary working directory - tmpdir = tempfile.mkdtemp() - - # Unpack the zipfile into the temporary directory - pip_zip = os.path.join(tmpdir, "pip.zip") - with open(pip_zip, "wb") as fp: - fp.write(b85decode(DATA.replace(b"\n", b""))) - - # Add the zipfile to sys.path so that we can import it - sys.path.insert(0, pip_zip) - - # Run the bootstrap - bootstrap(tmpdir=tmpdir) - finally: - # Clean up our temporary working directory - if tmpdir: - shutil.rmtree(tmpdir, ignore_errors=True) - - -DATA = b""" -P)h>@6aWAK2ms3fSz9=Nlw0@!003bD000jF003}la4%n9X>MtBUtcb8c|DL%OT<77#qaYeLNB_YQ}7R -JLBWgQMLc*D8D`sbJJ4o^By}nH;Y}-R7|8JQc>H)h=LtgSXPp^CfHalN3Xv#l)Rak_3*j4C>~Hr+sIG -4Pb>*Dvu!kuoI*)vi2F4`%Dav2)18n<}T4!jWSs$bTq|)*=0iTP-{H3s6e~1QY-O00;of09jja_pjHy0RRA2 -0{{RI0001RX>c!JUu|J&ZeL$6aCu!)OK;mS48HqU5b43r;JP^vOMxACEp{6QLy+m1h%E`C9MAjpBNe- -8r;{H19{ebpf{zJ27j)n8%0=-6Z#elILRo@w9oRWWbO{z8ujDS!QAC@3T%nJCf;1rX6ghzu#Z}R@K&*?Hgj1WFD91+adaM4G`4Xs@*hA^t@nbDYdL)-aOjsW~3}QVVby(8=@7U$ -Fzj5Y{w!2hUUH`?e9j7WDA;>-1aos>7j{2$~BfyL8p@__Y98dsP#Bs7^lWF -=e_gr;(4^?am?Cp93+7b-!?~nb}-$cPSR1zckA*zNp!)$;YjlZrfn&RWNM}=QA7*cb8A{(9@{5!vBfq -rEMoeu5FvJZngI@N#4#(2v$WnMGCAVD?b9t8W^qDfcFBe5ZZF%dPAPaq#>ikclG~yPvCg`JUGb_W2#PdCXxx}7!|T*xc9qdnTILbO-nAJaF2 -~0snMFDU<%E01X4*yW9@|}F2;vY~;0|XQR000O8%K%whB^=tD)dBzjss#W56#xJLaA|NaUte%(a4 -m9mZf<3AUtcb8d3{vDPTN2bz56Q$bEqvD7eOzLnyL~CDi@L_;ZRYu+Sp^V)ZR6_YZ?pj@13#Z1R`1=l -J)M)n>TOXIt;_f2D8Q^;6`S?Y{9RUgUr+|m;!25C-6tno(2iIDhjlyJ)nM4*651XX%H+qrBEdT{cBla -4$^`0^qPP-6zv*|ge-jzUzxn2=uGMl9#)iA)y8^Cdr~rxdixH}OOJhxFbsp>7(O2Tg09*VTBnRAqE#) -uTB%a`7P2*FzrkVV`K)SOhdyilnqJR#!6l}Q6a+(^)-m{nsTFZ3tf`=GYik||DD|c)gW1pJ_vy8mPk! -87%_j>OLv)_N=Qs$09E*XCaNb7Sbvyz%2H(~=0(GyA#Q^BB=o_mcOvCiSC>?bfF%-ta6OhP5HUX=GiK -PR!(uKJlo!!9~IAAmCk)?77i`J23la2CGx64oXMzMaXkQ<~~8EU?%I}z$$rRNujEIu~M()ri%^Gi%ri -C!gNA@cLO=l6KV$(xV^&hZYbU&TCtOIChO4g;gfcAY_>ak~kgLxGa?L$cMXVJ{&`q`lnqv$j}Cr3vW0 -iSMRu8%^e>;b`+HM=<$xdKPpu@6SuMN-LR>$cFaIP$0f`iClb~O`=>COC -NvJms>bV(-KMn=QG5PXY-h9vs;@fOIZ_lmTi^Fg`mulO!YZVO^KOIvSWtj)HD-~+vPht4%90 -Btz&yv-M$^(udyl?*`G;BgAr}tWa5RRHyc3Sz7-4H^#tC)@T$~!*>z3|0?b+NOYH~Nw+WUjLL%ySwnL -u=yutnX)PsolsJX_Jx&=d9p7_`6i5S -Mvz$qHBvD4Gc2vqMK2J#u@ySRoi8HJ74pBUZpQaDYr)B{xbde@6aWAK2ms3fSz8Tl@oYE&00931000>P003}la4%nJZggdGZ -eeUMUtei%X>?y-E^v8uQNc>YKn%UlSA_Ml1)W|5bwQ7w#FICXGTWwYU^+7-sY}6+H?6MX!CdkPkC&I1 -p7R7v)6Y6HHVx2JGAo3XvIeD`#JPUu6e_-52zR!@h!P7KI~18)C%HP@fr -1dD$kQ8NRuGKz%ZZysu2=G*Ual7)rq;5cIqz_b_FXoN_lu6z|qu{_j%fT!+RBl=guKIY1=QS5bb04|v -WA;eKlsTs@d!Jvgx1?RGCC*+Bw@QIOxwu-SziJ7_J091)~tDk`9(d78okVj;x!LdG5$Q)?DBIa2R7@M -sdD>u3!!MCee1<#q{z2%~C|LtPJ~<9zgOp6arcP+QP7iOeYV&Gp@_KO5Zof3NVEl$Vli`umm>uNm@}6 --N7T`WbHNRPGZ{O9KQH000080LuVbTZC7+>nRWb0C_h602%-Q0B~t=FJEbHbY*gGVQepAb!lv5UuAA~ -E^v9>8(nkUw((uR0-+yFrX-X2q0_m^wd0Sot*I+%BspzHqhW9)PGZd?Spt-FQT5-uiw_b2d1s}4N^i! -#Be+;B_Inpl5Cm`fvMR``zAL+?-m+Sdp0k2%nvRsbsi-KMniPFk);EL~B^P9kGvF}@f}^8N*KA3aZF< -pnEXzo_ZJSOITGx$`bNSJc9;=$08<=Ju8*YBJRNPkO+C1`7u;KS^fD-IM+;_B9OXf{gv0N@-);#SB*0 -JJUnTrWbO4qr8I~J^?>xwBLv1{3Y;Zw_TGnJ}@t*Rh5my`=<)FZL^~625o}pcd%eCnr;^pd<} -22FJ)bz*=$^OTO1Mi%peDF_MmsfKy%=6SmI2LzL$gh5Wv3i9}I8-dl?KxJ)T=!kr?ud!sb^GqNCbxgo -FCSF2L}s<$GFj7AcbP!w}kNz=9M2dc{Q-6Zr4?=;#RL2UIVOmq0cx(1t+%uVZ;W$zvt^(sYKMB~${b-kW -x-&tzv2jRGzH^>eNbWH%qa&}-UAD1jVzMy!IIE_RJ(M5)H)8L3KfOXk*)g;PdB_^c~da`o_t4w3}q}u -!n!O=+g&i2n^edh2k(?&Xx4s)YlUFaj;|f9P|z}v69cYj_{@6{+fMBo-=m$|SYrvc?giNZAckTBy7Y? -`lEUaaD_VuoFUZ=yrgyZgWLQT8As|kHt-w-{zJNtR4M^TGj3Q*cvqpb>^}S}$xnvLr*>+o6mA?WcY>z -Y>pzlfc_cV5x+~Ks3dP6*Q%$OBoOmrOe-uwRv1qb?qHR73dDXhX-l$ClBLUT`s_xY9I7AQej%Cbo?)z8!yJ@}VN@Fa*2 -8)Mzo25I~&8(5GO>NJ!E&9r$@M`0l-&_4%pNkQw%Y<>s2XGF?UkA7jG|+lHvUhi@udTDW=A_*#;X<&Y -^u`cg{XR!{?(90NBCS~;>{ZcerVuY=VQlfe|YyKb7wBmh3^+~}>7xuvGleVI+D1HkDR=;YP)S7!P&&L -HwZEFN1qyFCH&iu{XAmu9wSgpHrK2q=xm&i?Odhxq}>=J+godwOzldH(wNKwu;!=&kIY+^mzy1#E~7HFB$=oWFm8rU+}WGY$DiI=`%B4X19=)p!K=`!ReK2BGl* -zCM-Ap9Ndn1VqV;L{Ji1;YU7s&j+56S&%g^VlW#G>XULNA$4cIPt(q-`KkLjzxeFv#Q&z7zajDDwq%X -5{?LOZ0nIOd;w96{(xPP~7&$) -u2qN<}jf~C#nEVO!+j$tc@VST0{5363I*t^aXj%$1=|L@aThCiv?>(tC*DHC0Y}tahtw^MAgFrM1zWN -TdYwTEVz>?Y)qK6cUg&0?bGqCE?qq&+otGYzI0M+CX3VWG|~dS(6-@1rN8M6V>IMsgO{*7#Z&EkO00S -B8&H%Q5*@mRXaR)F4EXrB$|@P9t4s%al+FPEslWv!NZ#BkQD|QlAQ#e-4B!FApK`k) -=CU!fESO!qDGe%JVi5;U9v4#CYy;_-&K6bI@58?R*P|TMnswiS3y$321d}U@2{^TpYOUmD4|ej -7WfZ~>&1k4Z-qooPvoowX2F6M!fRXU?9zI`G^i8WtDNJh6`14~q{}D-d}TOxc;gAeB@zxO!tZOF&l`g -dikFAoVN4PPkk~V>T^euE3yn02t_P=BWH|g4GBs#|ps)shT}Lx?<^a}gvz&F}FhYbFiM6K9qpeDYG4< -!<<1(@KEe$9)8dkY -_4Hw#8%I9b~vkbBhz9ZvFFL=rcXFOhh%h6?kS0V!~)*<@#(*)Gm+u4S^Yc_A8#=4f~uIpzq*+zL9qKb -re^ykZI#!C7J+s6JyTkvQWKEzO<&RO03u!Jxg9e*Cv@N)79!IEXCzp_NQTwW@rf?fjz657p0BKW?E-* -;AA+Z+Q81o`TMoh&xLwQG23EiTRG0SoB2eb01Dimq?xU_*Qatl|3Z(y0BT|y~TXkqQqB%`maM#nM0&?9gPzhxd&zGB^^jCbGhCGQx8FKabpfKit= -yr{_*R|-aLH338GF-CSJtjLL}O~$u_MXnn~dsc(9~SrVt+4rU&%Sl1qB&?47IcBEhohMphBo6 -a^b9lXBnoe3)>Qt=)G3;1>?5HB^pm4+O+#+Z(&E=H^>~+nZ&JC2h=PmmW|9t -u2TdkUhv?4&bX;acnVGke&9=a$!_5r+%R39f5PBBmJ)T|;B}{E(2Gs-C( -GTU``ol%71^fpaBk`{&HbgcVCV(+6obLEcQ@RZiitXB#7j{Vs2Ait=#Yp>(KW3WW1_?eX|Hr@xO=ZhA -^FmkP{7j)<(%nT3)qvwpI3|P3A{WMlz{RilZkMe2HB@+adF8s)|Oj;E9jlxyxVy(5beoq(F -YuHrHE;D~BI1B2X6e@mHpj8uwMKfHQ;+*62)$GdhrM>)Q>g+VIr9Fq9Xkz1er)1Iu!*mN~(`eEY0%jE -Of70WV(>ab-8MnDtqp&0ayOF0Go*194fm12YS!kqtM4_ -$eOZx%qe?{)njRmuP7?o2nJXY38#oi7zr+1j^-(jZfpu#Y_@DGPn^hP{~;_}(+%a0h9&wjrAczSYj`u -SpYesO%c8pkBWh{8&gjsI{$r0qo-s6QMavx6(OuG4umz^;IGq>wSxK%~5_N?{Tb&f3U|P_1n18k2Kf| -MduFF{^J2>O1krsV+#07i3&j1sAC~)1-CEP+4c}1>A>y*p0WwSRT51^l<0^oIT_j-ho}no3K}Jz-g== -a@Jx4|JpevG0>0}L1Jse&4P*b!o5%XwIdf)!aLe29Ywjy{A(=UjiRI=xh?k(C$+ZdIo!iNBjUDc!(q9 -pNlbf%?))n~7t_mO;gRj~RKqTFxF-NS`0@|DM$YKVi%eX3*BGzk;mVcO7zL-KE-#T+_}Z%K6M05ZmFm -FUJ#5pLo5JCYC?@w5oK6u*;uI0wLWQ@0_M*WlHbKL0Gt18vY{L9^rNZY5K)JbTT=(u4hlf0@Mm}8JlA -;I(;9v?ZtL#1ZnnT}x+fS&^c&ip;c?eUW#f*7-}9B<~SM4S -b0G~bwY^LOUxYW+`bj*BDJqj=+2M8Kzs)UJGPK^WXtkfT9dm=Tu2!0f@k)y6mUdt8MzT6- -lXqrJS89Bn>V2?KcWlr(22as^Y{z4gvhLaBrThzpC2lB{GAr>W!@5T=6`ID$tVSR;2tB`^)?rK!_6z4 -AVR&G|rKN!&LR`uGH9&0s4q(q-2lLE~HZ92}cpREU%jKhu?rEBz%=@zE;r=Re%{!c84l=tG*+2ogw^6 -GHuX+%BrE5Hc6Rw7w>2EfpInPwp-p^5Yi5XJ2H2-Xre6q$Q%ub?JyFy-H_b?83+EJ)}ZSOF~k(RztWb<=ufSF%TSJm8yk -0pPW_+M=Jvee&W`|qSVZT${h=nssSIbQ?02^7y*K_}8<57&mIGxuli?*Y!S9tipk+5H6kcap@pbmw5} -HU7!Lb2j<^W^5fTA#*rs#PJsLD4fWORo68gWDWlTP)h>@6aWAK2ms3fSzE;-+kFQN003zy000&M003} -la4%nJZggdGZeeUMV_{=xWiD`e-CAvr+r|<8e!pU?zz9{yJlAmCR)C5+&N+7Bm%w)54~GM>5?9jtC@# -Y#PgdIg_dYYbB$py}J~vHTR4~rszRb@1JhRkfGI=UjTP0q}TyE4(<<(>|IXhdIW+#$lT~~Ffl0@iTVa -iI#JU5lBw8`z+nIGFqyUp~ndHiTi-h9u@W~1{>^JuK2TgZxbG(>;EqnoG>1(rACPx6Cjq|im2+^9S?W -n9SBwIr%>B{#NN`(AElLg$q#i&EillFOaykKCxzg7MoZ)|Jj$k}H{;T(4xNe^yK`WQGanGKv -&Au1;4fdoTwn}Bsbf$Rg$j+TfRc7NcjV)%;<4Wy -{1RS+$#j|6_l!uw1Y0M_qI#2CsDv+hs2H85P49RzPM*g5mv1lA4-l*y&k3|WqI7y~wXK&uV`2NM~eWI7Wnw-hkau -C(TrQ>ItC<9(>a?)1pUDqq1!(S+Aerqg(y(~NuEH8B}9`exG2xx7e#4qbgC?T -mSQ>e7SkA6n^L5*l7g*8ZiC1rQgh;e=XQ|E=i)uKmc~@1l?vZ^DsIkoyIAqCx2}>TvMO88LJDAuHUHY -=o?%vIUBJN8xZO8xr@+2~lOU;dWCS=iHYUf3wS}xvJJoHZqvLMNqQ9Na5BPax -zUV|L^h&M92NVrtF7E5&k{9fCkFb*8I>pwb)%1h!RG*!lVS1`^fF7>mz~Lm|&mJYdJH8PKz8RhOg}B -x>ZsiV`!7Oys_KJcB7KM<3;Bi#h|psQ|MwFi;hki=h9|J+`$KnL_4gx@sDW=Vq^NXj{)N)*EdnuMuA; -x|G^O2k9(GWrD;dF;)_PkS>CVr=WuWEyAMPDtdY%!R07po};cFS%bL8B-DRHP7B8-pOYBNx$u}Gpq61 -#*4w`)#wba-B|$=TWejXnay+Vmu%N4>W5ku{$El?`~Ib|p#nGfkqeWHRhbMRqL!j*m>1_BlV<9ziF&$ydjJj#8iX^#672D}m -fop!j_mr6uB7di)RU4sgP0M}<|J^r&mH2|K!fo}z3ahSH-zxDN{;@n5D>{Yl}<_AFtiqQeXMc)sEr?t -cgT(BZR4Q9?~T;1_NH5}bb=Fx4(=bq*=W3DMt*F}(%h|E@3DD~|i^1sQm16f2dP{t+-7#!R76y#iZI5 -=opWDrhMQ`U6o4XwQ1EipvRm*c6!rY5f>uoNYSjx5(uf1+Psc3w*9ug>*PwTi_>yop78?qe$i%BlGoJ -sf#i<5P5`E0DUw%A~Uffi-vvdmh9O;y9*BhzM?JNyUr3jioajLx`@X1ExW-D88g&lE+fYc8xogNv|;I -p)OIZ31lo4hEZBDzMo_NU_tgGny&?li_?y_N1lZS)pxNRsk^&TG}1(7r;%tdZX>PT -{QqHFrG~r)pVmqQNYECNDPV-zZAMoGtFTM2!CsnJAl+@y)4GuC3k4isMf~+5%Y?geW#kjZT0v2{V2V) -oeCOC@|aTeAaX3@Ds{ETrs_~en|+Y<2DeYen?D*>itT5%juA7R?xS)w)9+f;bnVpE;1>G=oaN2gKdODPpqTo_C>ZR0<|A78<+M%? -I0;nMXX~(E7#B0ZMt|qTa#Ys?CU!%!rpunBUsA9Lg4l{KwZzhx%v -Yf!^)MJ%^m{;_=(c0Jd+0B6SMc-h^@J~q*uwX%&0@DM>E3H3zHOxUJWtbhN}>G_Q`RYMd@dds*fn -{<(srUBUE0q{yc{=YljH)Trp5ot~&H2S2hG`@ej)%#G&VvuSWMGsO6Hx^P5aE6RL=3?< -Zfm^>Z%go+MzbWNPt!qRwLJbcozssBj$Xn5=zO1tw?p}*t$v -NSD5u_vBrbC|H${WK`s6+qz1?@5>3?Ah?90!9!;|+&z0(|Pzv#?Bx;vfu-|rX5OfHUtE4N_S1M(iGJ3 -6R$9T!rB^A6vwgizyXEZ7@&P^2+W$kcDHwyuWy?I-TJf_46~{qr|pNQf;lA{}#69i~{I2Xqz;Xk3@gH -1^W=6T?VRMLJ@iRHe+qcdefB&;rQpZSQ>)JN>di5-o=CKUDPMbPv$i_M-n7qMg90Ja;}BW<3`U{HIfX -KBihXyhY+nR5O+glUkbx+MPMP#TM*W-2dJK!I!oEKbo?=dYTwwp=qD%@pKA9%(AAo(2>SE;Fmt2IeW~ -evMs=#S55)fhNYz^a#~SOs-@p0p&79xwb!t~X)f%4ZUmoWpyKZ}NTL4F$|KY(mj~`JmAdhY6F7lX0>n -HT<0J@FoIVTMcxuwrtj76Wx8)qT*nvp=%k;GZWR&!4C$X$If^RU~I -S9zIc3#JCvQN1K6J?hO{h+zCRw7q>0b78;3D=-cr^(Z6x*O4xO9mzwUpgJj5;3#x!m>}NGA@MqXHdr- -(Z2vtO9KQH000080LuVbTkxXRe@YPm0O&XX03HAU0B~t=FJEbHbY*gGVQepBZ*FF3XLWL6bZKvHE^v9 -pTmNs{IF|q2e+5^?pf=V~ckUL48w0-G%p~mulO{#dc4iP5TE(JcZDmOiWvAY3?r*=(dn6@Nl9Oz)zy( -EPOXTDG{*5m|5d2i+tMqPL#dTWbViV`_o!rR07E0FJax}UwHd55G+N61r6?gmD=t$)8MvjCyR^q-&>s -sa_&Bc-diB+0O6=KmY&bGAbGyN8^QpGZjYnh1qv|fu?13403d0y1KUyVlQnFd?Nm6DZxmKDo5llEJo>?gHX8uYTrGEIvLU~KRn0Lq -dekv~H{l|SN4T5D_hEbYhQyyob{JNASgS1=5V2lu+Y)`AIsHXkrZy~Usifxu$6!nktyoeK-Oh=QUbGC -JwHAlo{nWU9ExGsb^%ec17e?7Z6x%~C|@Ny{EokR4Utk3ZToISW>ld6V)GFK!oU^K<&?PfH(itja@A6 -P(Q?#V0cz80^<^%{`Yah{BRN%I6749~{_eGjHW)zxG)`swoG?b*BN$K&hMi{tmFSQij>X`TimU0Fhf% -#_q`=-gm*dht1;_1DqayNk=K(-Ydbla+!D06SRAtc3o(5+9+lUvQuj#mZM*McNFVjw>0C^pZJKtHu54 -`t)iZY(Wj+fu2u9*L5kE=_+0Fig)LiKOUdYgS)J_jWfOKdar7^5x=I1sD&o3^OWM|WT8O-;HgT5zLy{jN2ym8(T#2Z8!End)-;6a*9Q_ado-q&Mq -u$3~r177m9x8?Wz_v=>#Z>~fvEHa+TP>v+ONX&V#ol>BevV)c>H-lAm3?#*g+6+@|PCEr1@0EDR%Vz@z|qa(q&kv$}FwNpnod9pBVN*1tZTbL>zO+%) -BIB);Hv?rY#>0BQ?5atKI{a-848BDulmYhkq~9d`zFFqT}i&Vg((dS0oA(077yV7Aid$lhZSbvl#I&c -L^PgbG_3L9Yd{A{WBRr2IXGvp90Y%Lw+qGRPu)7;D3Y3v{}ID^*wb+kK0upH1rE&4f6geOe|mMXooL` -Ee8lRLVsw^cYV=pNOsfCQgT!?anoh>qPvb<9oF?ZI_(l>wmw4dmKhAz*Wbn;{R}pfE;+elH*keo4L(% -G4q-;Y+62e)63j~dQWiNL0hl>3 -ok1@JceOb0wXQ+%$fi%N2g)yZowX-$#!09XrTnF?v8kYmYk;YEAjFK&U5VpzIHiK}hV5>pnb&iCC(g!zM1rh_T30JLpa -(y`BD(_8T!iw0s9=?MHt>i}1WfzUA#J@uOVE6vyr<6_KT=t9bc^sL_QJ9^ZizjdpN}8;pwn5-373=d -L5B(m8f;ysRAVUZ>*ohz~gz%9QrOfh-EB^+V$7A=1Q3q?y1A>i(akaN6_?p1Tn2-VF+H4;*coe*o2(s -4iaxbJCzCsrX=)@h>5LC%=CrSV4C*7I~PbY_E%0bz8B@LI!XFPhVSEicMlqpin9w+px@x(u;dsetD;b -7(1fi@XZ4|mTENa?8;%AEn_dL`O&i)9Utp?-y%5F5mMDrHvNOqWu^mQvIPLDs%0vwa6xa8V#;FW1E4p -h+)cE_q`eE{OT5%3s&^HD@BrVuQpa33XNMLG2ZUo}icohr;*D#RsAJ!79XK;Ao&apn+4ZXfu-Wu;YEt -*4$|1P&(!5;|ml|tBGk?t>c}H8dv2eKZCxkdAlOGQ&HO9)>EFyOQ_5a$0v5ZJFD;1*8Ib==_rNa60yd ->w~6Jj8!@;L2Bsaij#Xriw!Pi~voVX~BJ%n@JJn5Wy;a=DnkWr0345zwUYpTR@D|lbhFnVD1 -(?8flP2@;7oqroAPHc|ZupLZqTKEFE)060dOqxqIdB3AeYhC7fZqq*+-_tH~x!buLB1vg{VfPqX -il|IWn?TwWq~`8LZIgYH(VX}=kJ{Ut4V7J|q>a3>&=En7n0rUR!z1@GCpl!iSJ#1=Ilz)%rBDF3NJqv -=d(KG;TxPj3^KXgRMNN}tYF`l_EsV64GgZKLcSd;iv$>ECqMU3htQ3ybUW2G!j3bPIy7`pMX!r}7%_7 -LPy7J1d6+i0_9Qd9wR@{hfL(#!dhAJ2kN}c5095x5ep)E_2g-D@VPzokD_{My0Fw&w@E?$wp+Gp;HyQQ;>oraOMI_obBCxwHG -P`>eY~HN`&<`Udl?uyOHT?No^fcZY`^e+E%P?<(D$eD&RVvbwzo)(ngh-zt_}B))fn2{kka1cQ!Z`vt -{XdO*bbT4vS51W=1F)bYdwpC)X}~jzeV7GNrv3iWpPtsSs3%qiG%6zr#!`5c_Qimf!axP;aW?fw?#3zKq|LY54|O{#L3tfBxaW{_`d2^w1RFaCQy4RFThDffAuF?TX+XGh-pg|6-QyJKvg9}H%IGz* -^SP5OIp%!bAr4!F;*Yi!Qo*VLRd*qd9w_Fy;0b)k6JO>NXLjd*stv%}Y1ytz|n&STnTOB%L?W`j0$U_s7A>L*I1UGo6NZ*HYR={>`7KDNkN!I+m0g?XIvgtmVo8;<1<=7EJcW}^`#_xnl+cF@v9~?jEO-*&}s-2 -&yIWR4Bp>8Et3B0fc!nR0*os{#$s}&R)1{(9wH~=6UuF?d7pZ3xOI{V+cqRIOc*iu -O?a48iTV*NZJYIBNFkim5WgUVF0k22PTyGK=>*r=fK;_ZNb*uX!Zx7VO<4QBpO6++C3u8}4ti_$% -SR&~Hd`sk>2V$=`)8XdG9XrAh}%E6Ym2&rxzZ_cY6LG6q)6NQq+U@1Mu -j@!n^D0vhU!fmkXbV`ENQ~o!biJ_Rnoz_<>1}T12UxsTiunpG!DsTEz^2m`*0oV!9GII5N|zQJutT@b -;$<-O -|G}1dyj1I_o)7bLC%90h5zx=i^7R(2cOKJb@V3Ey2YzkI;fkY?-$`zd -(jWR`i38V%5JrEwcRBv_1={c$+Y{@7b}NR6E(FYIvHy234qsQngFYhS0q`ga?VsX2jvu-+SbwMc!JX>N37a&BR4FJ*XRWpH$9Z*FrgaCzN5+j8 -T`b?^F$HYTYdZ4$G)zLblJwyf1?cBXbNilkZFrBx^pXp*o%07iq-a2Zym@(oVqH}Vhp6Msq0rMuC%QP -g;=QjThOL;~HXPoF-0?%nb{@9a}3^D@p-CDuwtqFff)da=ybrO1QuE?7wa=;&%0E3wL=bt1*PkC#{C{ -f8@278_A!B3|WLQHptytwfM+%4M7`#6yg~#cdwv{xnWYS)@U73(b$RToyqoeL*ncKlv$_VTmJWVkR}X -U`8bN3ijo{f@jf4WslSrglDI%H6G!hv#U-?I#=N%mp%;|K#F4u*eP7U@ -!3OxcM!_kp`7G~)@UgSZcWZ^t3<)<=*fg?Q -w+nuZvrsuqM>e~gt>Uj_8AL%Dn4aGRO1_L$S;Vt7QlhV4D$9qgxbvgzAx*L%vdF%I{qs^k?CT6z33#*h)JaO}m4G%KVx6u;#%Nf3B(U&txXuDB1jY{&ooTdktAqA6va4BkQg;H0y -v=VP(4L2pj&{9F2#5(9+n`zL{LbBA?Q!!cG8T58q81+rm=*v001%sh!N-kJiJMyOBnbMWKNE-Sd#$p7 -z9`Y`I>u>ZUN1%a=`4<7*(qr&(?7Q{B`S_l`>4LCr=xOpV9t=O&OQdX~m05Ci9W~_VeoF9Wdy`sI6;{ -&;e%k9)cnTRJm+~gzYz?k@}{=&tmLZ;rv?3rW3CH2>uJ6L3wWH -;2_+ck$4_H(_7Wgci^KR3W2H=v4zU(TB~Ec0OK+f6k0{&`Eyr&E%p)2VM%@NOcu^*Qgp_J;081k14i7 -}$u|3s`UB>?2Y-gW7`b8*bn=n%Z%X8+iPz*ukmDzHS%z$JNv*&NwPeIw~?sEojtROCHPH -q?|RE{c3_Jo(aqMv*j>N4nTTuWX~k6V%4yG=ZiP -(beUMr=Oq*k_eAC0#cBx|^)(5Px?X?)2KQ3pEI$MOz^>lxDm}7^`H~;frKt17gw`Wvp?f?3h|LEy3{7 -EKfY>%I0a)15L|NiS={{8Ex%@Cd*G{d?L{kP2dZ(dOE8ur{lOX-RI6KhY3C*l){8->WatOo)%l=xt%Gb+yLUzHYOa9FVD_^NL`Wq3Iu@>4Rs3l*7`EnlFp&>r^9(oO -iO?hw%x6xrHA}Kv9_NDg!@2jC4B)K=v-5Nm;Up~A;2DuesNe~^ji(f&TNmV@^|_7T$`ri^A&pmx4ov9 -wbUjfW?RX-@XkD;pRCRWNz6#<*MpqdqxE#E;Q*h|wC@AHiHCi>uMG#{^_F*Yg@gT(Bgxc+`0Wr_ltg$r|Mtjc|gQQ;M4E=W|*R*f%m~=ag8v?O=6S+ -q{+(u12?4Jm0$G?w<4Px6TMYm|6&B=M8H}`))@t+}y8CpQSMk@?9Fw>Rp`ip|SJLGUdBi!3rDvR)1`!q;CCoya0{Z6=OHH@W*cVvFYB -$8BK<$;){!9@e!k5{-9)o_%(`IxUVJ2ZCZY{;ZZQl5Gn8W%EDzC}2p&>w0skO9mCOd@{ixOZ6k_02BRh;_`vQ -*9$5yrceQkU9+_Qb0-SC9$F2aAP&Htict4X1r3U%9g=Bs~nOUCi#To2aF4H-+>(wRdIWZsigN=`6^qd5zXf -I*68#h3;URp9AsCJ-3CSm2`92&vS^LL$C5+wusGU0SOpt2E}}KJlLg5!o$DgYLLF`o*l!U`c+$$BxnW -?QtYbCw83UeUv`)YX3sxHXNu8|Oj2ODvwlNY)a8mr)yiWC*Bo-iMkWr0A2behk8uhWm(CSK_Ok!a3qrWj;&FXU -q(|G^Y63;n^DWTNABsiJ;j~F;K;+Yxg%!*jRS(uc9jti6D(YNU8|pcHvEYgkI{}5Hl?bB&J9Z{O@QT; ->3v*eDM~4lU08ZSeNK;N%?DUOiL{pc}*pH8Db{qG+WK$lxEU9t;Z3W$vG| -_=){rmIDv?n$6!d54v;;FzCp22*OT#=H^cVgol;2#v@6Ay;q7R>w^dK*xC2SCkiZj#n@X%!(uk=;{j8 -wI2k2oT60qTzQ9vvy0HNj}&RqN3b@iMyxk5>Kjs+wv{SQWBVf;(ZzL5_luo8uJL#;xCWHh+SvX)Q%1~ -^O?T%07JQKO9>Tt<&5AreJkHi>mUF0(QDmuAC(qzbHBMV-r_2~2F@*YOA<;3+gh9~l<*04fUD41#zh< -X(nxiT*BN4fPPUitHh^iNF*k|DXDgk#YjqX-(90M*v^_KF-#Ptjd>haqv*j6{X|2Hw7A!I4p7L*kQE0 -(jDzR7jB$pY0C9LeZ#%Jx;jUox*5;tNg9!QOjl~*Xgl6}T$D1gN-e@Brl94(Y^AG`-T;a*2%aB -h&(ztDW3NS`QbBIGH8|c{K&))l9>r(+kk72^ZW;}E7RP2{_1BhO(7=9ngI;wjxjv>_o#M8_*|31C5NI -_pNmv7o60y|Oi6jgsUBjCh)v##|)bb(h{6?Z0RQDP_Q4uMfKCTGZa@~HvlrljniF$dTeAMl6mq~Vuu} -4kRd`w|BSg<4+>^^W#TLxK!@RXM?iHusB?cu=eqKtMQ)s_(`HwbBDH+U`C4B?uJ7)3U;;l7a#Q;w0F5-_X4J9;J;i(~MJ2$K; -hv(%b@j^-xpo)F)JZhxM1)Z&c*ADKiZt~rHq_2-rS+#B7*B}Tx&L-?x^L8O@cF~e`XLeTgMe>c1d{fd -4+ktv&%1{c;;-da?56m+OD%(g;zH)-YcZGrK&LYr -3LZHwv_O3EAAyYU{}oVjuZg@)`0OkBP?Fbth9U*E1c0V$-G_pBuH4FRm-4EC4Xfu_T-#F+I_^s_|HU0 -!JzuqdhTZYw5E}(X_=l1?My@;=GdIm0<+~O~Wh#VIHSWK8~{|Q#3Wz1Z;IxUPHf|TB%sa(#G>z&+c+9 --2&#i%2D5Q=<0nYf_oqj1zos_@9j%6H^H<6e+bw|DLCc&>cO0o-q%}@h65MT;Z88ZnVG=N^^2PyQm^ZCirZ`+5$POwDb32o~`?%Ku?XN$2Xrw*Sa`RX|_4RO4%A -jd2PKTo8Ygd<0_)&A(`VcW(#1I@g~69qsZW}=5BhA0fm;09NLd7e%)5O0{fo~STb4*S@g0`{DYridOl -Zvy|xQwL9Vi|i_jY88cDMa?^SwuEaP_SzWhq>OVQu5B*N0R%9X@n{id+sjO#b3UgET6B7_aJttL>9a9 -70I$`T9S`fTuWu?)L5_#6v`M247u$G0XkYiGZwC{VH4P!r1yjLv2X}IFoCK>`6bKNp($9)W7L|HH!i0 -&^_^N3(pN4p5Cp8DG94{7HxsYrPY{LmBtjvS}n7_p2NJkT3bDFb1%c9MI1Ef@{U&EzVNW^AD4VK|7P% -o@VS%bDXueRoR*2X1kb@s;Tf)Cy458E8Nk3s4832`@8sRJj1D;87u)E?ejpyO7GyEJ=9$BQhZu*Xs=# -&r_j2dIZ&#U|neJ*iSC=3#2*)Zh?azznkuDB$@#4q&!}Mt$rzjotWE8umOHPbNQes6ui58{P7vG;=S{Qsl5oCLeuYop^1kVX&kS|W -~bjTe5Fq)WmE^%exKqxUwmc~@ILqzf&sv*;1a!Tn?A%4dh>wHI^1RE_c^Ni|&C{YLMZsVA=_IN0zsX$ -uViuy^6gOJ;TxvaM9!mf4+HpGI|K~*ATkAFI<$fiW*Fe%O-RRvp03pT(JnvI-f(sK&1SvYw7G311!HG -pMMfOBpR0XhCcr^8hbW>QNe9O_@@XU&YriN|96Wwobh4|j_xj1%*YLK6y1iUK(k^h<3q)>zj~Q`bx7R -|XLg3`PlntMO!BLddfAQ1~5BE5#`tK633P&MsE;;nhGkaDoxeI67JR=(6H@kL<3ts$8aBlT4WGqjEr< -DgaK)*9pa#V4LDh3#qaMb|tLi)VxFw;BGK32V*{DsAG4OVN -Y$5*_)9<@kNW1Fnxfy>T_1A#MG&F(Q8OG4R>biJC%LhU84joku_p)KshFaw_-Sf?bfjo_;cm+wP`n!y -RM4l4ccIP)aCAA;(PlFkk4P~8}9K(Js;^$eynBbR6tz9DcA5~MZCYOG-Z2w+|CKRBh#yl|vSZijtnj=ONf|g#}F5B+~TkNUE#21UCy*x;}B~Br?*)l}cHW%H5V+$pRV@t5n-F{iyl17mjXD5ia9 -1Gq -SppIK-sG(5wM$!kQ}m6P|gjma211lG4tIz?sgsgDjP2;$;>Ak;1qD0~!oGz6GGluXbDW+sBKyZOH}pv$3J5T%E&K2{vv(Q&Og;KB-+t -0p9qMEU+N0#Q~~xj{_-lOXGE67KSWKUP!5q}7V(Nwh? -vLU93*e~P~G-7UysFwhdud$032>HqXk;eQSp&!Qs*r69U-WPRR%HA6355rq@i%b&<5i*bd?}E -!!9bL7TknQ`Qf1Gz%ro>bAT~25#rI(#ZMZ6%`E>k`&cL5tot=nxOvd_BOfJwah93wL+tcs~2l^jZLr8 -EZMJmmh9^F+2xryKRNyGh!gMcLJFbp1->I@Q%J;oc+tkHy3B`&fZ^Lx{LPe&D%3 -yXx&Ncy3QkVtXvwIr5y?I9B|ijBNtF*4*22&iob^#i$3J1V1?d-3r*MSV*ag=l+pMn7m|@ESjfwqMq1 -vvn_}fZrex`Q5gAQDl7^Hf8&wIj8`8nbeMgt@+#IHfE$9;hK-Fx?0)%S0ToHr820+u2tlNx+RosaB)u -Ybx698G@9X}lkUmZGz-m(o%Zw;L^t2|`q1{Xb*8IcL|r}?h^Z)>=yybn<4KArZ>8yS_WVJ`Zr+lhmkz= -pO&6;e?Uxp`PDCf_3K0b=>+yIaDo280q2MQQ+M#4py8Y-GzF$gr&%$eFI5c^t^%DWqdnKO#$LgcK%cU -L;S&)dD?Q?5+V;O0L;r%=ox-XQlI$Z;{tajWI>#g&FbWF}D#a%i$Q!I_rAuE_-I6dDAt8c;Ym54gxdd -(CG2OkPlT2)7-fQ^+2BWVcacsr$yR*_aPc0U%8%#X?nIroS>-5-iU1kmEfM#N--gTFplrmyeR`}xeG{ -f;McR`u9>?h}a=u8RU8*G5qeDGYTUDLy?I4vYh9-%wLI;Ugq@-?!A?nsV*v_cHbH>89MEG>X#G45~tg*o~cGgw6tV`pz-yZ;96h~2}ILaqB9tj)d4#V0Pjr9uw*0T9z1l#D)EP!f)@>iVfx12*+PZ`8%&GSN9hnY|%}J%itYj-`i -nUR4F`@>5ntH$>L4~I5LmzwV -az6fTqX^m*0v+qV*Bz?byJ@k2zP0URZB -41ExtJrJ5}VA1-yp?Z-66g58%A|JBcGD|5~VZoBa7&4Epm%-+5J{n^DgFj_s_1ZYh0tdYJt8-|J18D% -`9S6MO8eH#8fH`I6D4xAf^$V&T16)>(Hl4?5HZG{=n7 -rS8_ng#E+cz&~=ra5C>Ec4)U|^d1_Ju3=NT=BzS>bjElwsk#`g%ruNs_m%!7hRazA)hUxc)*%B^uIl$ -rWLAMgbkk6)b9pjfQ5?it=kSX>Y~QrHRIY0-1du2c4I(4d#fyJyA9u!wwB4P`-Ks3lU-E;1>SHtGzus -w%|34pMAEkGnKA@(v_1FcnDN=_8Tp-ttIxZ*DTaGfsr>u7pHRq)$)>?b!eJ!Pd&&%ldu;GR72F3bl0S -gvB5jfM_AmRt12ggu_2voY#u^ms->B@I6M7t@sd4%BwTg>L=Wd3YQHXP2ac+({Z=l5aD$tR_TAeQI}dpd_v9=oVV4+q9aIO+I`qxV-8h1 -EU#rd?WbJ_3;qJt5C&G;6#H2L9f(?OGIQ-{=Q+T+SCw=(uNB+6grF3oLw>#s=8K!d&dYV|{mU}o;{@h ->n)vEh{+A=CrXQahT)G5%9X~~?Oo#3@ZSk1uVO!O^%D9uMthPsy^p*E^^CQgeO^q4S1cZ8=Ms18Z3S7?cr6bS-VX2RE0<-f+u7#NO7;#FMk -&BWju>OWyb*v3^Pg9}8XR$kbhlGlwWUa61qN*%rG!daJ}M%C!hBIlT&g<1W5h_-B}XiEzvbVz6zn)Na`N0|fYUm`dop@!SZ>x%ctn?ZjaRtTzOnhyqMLPuE4#yfd=T5XG -NI4_xZMS!FLZx{u3E-$_e*$(G6zvJbAZ?4^|XXNWj+`f6j3piqne6aRd1>Zf?M#)B|;`xf2tEc0^NBg -qPj(74Rd{gU}p4p<(ROnBKqnPc#bG@cvL7!#g7DTNOG2c$CgtR_S)z2%-#1M%0S-J%HcWoNt|cT*-uNt|Khp5v^tw}A+~Aoyy@dMu*a=)>@CYOSCbbnkfHNU9{6$gLZ&YiUD^Ah%co>Gl8!8p*zeT*&{W6X2~_CtBmwLwm=_a>h~*Y?x#y_+BL#mm3p6h9YgCPZj&KG@lA -+xHi8aMbX`kOzxS7G%Jn04kTF?ZZ&>h6OmxH=4Iw7Oa74w@U=&US|53FA{j>iCP)h>@6aWAK2ms3fSz -GL$&@br#002?}000#L003}la4%nJZggdGZeeUMZDDC{E^v80kU>iXF${(8{S{$OE@*oZ)B!zu^yE#Xb -f@V|pwpHlqYU`t?T)M=knjjEA1`rT!TYr7#^^oJ+A1rAQmO`}TErL=F~J(B+ytIwgiN9zqWlRz@Ky`L -^D$0v_X1ROG|+1phWCX8dN)Qv{=x{UGSnPl>b^B2$i&oCK$oEX!w#a9Gn^3^6Ec>G%6!<7T5~L~-6F( -D7q*Zgq6W%x3xWuXYLn5qnRMCCcB-pudkd~nb0o5%xQF|vM|%Z4X1XNMFRbtE$lcU#okA=Y<-OIe$~6 -}M&7n2qKNQ9uTKN}`bWY|MP)h>@6aWAK2ms3fSzC|3|9W-?007Go000^Q003}la4%nJZggdGZeeUMaC -vZYZ)#;@bS`jtl~-Mh+%^<_pI@Om4?BiglR`_EfeExU4}r8SN#4R@G`5^k)|OmJ-i-$$tYGq;kS2-WK9ZCHhq;+YQ6t8)Y4DTc5# -eW3<^#uf5gV@z&_i$l;Gcw$XLBq$yK(Dn4tep4Z>8Y<5*p=`sq#*Z%!z5 -!|1`aa4^R60C)z&Mu|09s8I~rKE@KNM2PA8!vUs4a@;Ilh$P*?3%sRl?OW0O)wWeU)iS&&t_ui-!taM#wy`^ifsY4*+}vys6(MjZ2-AE@s`OLtC@K3h%WQpLj -KbnQHjMdG46TYx5}aowGnp(aLre!HRoAudZsgTS>=EZ(rZCUw{6MO&HoeZ*qy%1CMc3LozM@H&ppe|Oei4am2L+B0N#MfAsbwCeUpL -Kf!O^ju)YtRaERllAUd<#1~Zn|+dQA8J*Z0Sc?sf^Vu2a7&HT6P@GT1}qP1fGoHF(0oY(4`wEp9w`2w -fovBi%T3HfH?YGmR|-^xn58L+k&m270-r5)NRF~O}Jy3qK({0>vl&YJFoA>3)}Sk-DPS{UyJpaBOBjl -_&Iyq091w^J7_bSJ)uY6z|gZ&ORN$CD5(y-A~C51{9XJ(GBR9Cqlb7pJEw#EILVsN~nxX)Ysdk9oG3e -s%ceWpPuy%)LA_wFmBlI;0-ot-LId*iiyy5DXS1c!3hdt5 -$qE`NkmOAf4ZrtCI?^Dxu8s7=HKJ171?dsNj9 -wfHY&{CJXpFR^he!ys6^yw~gTLKhzxLTmU1!5S)m>f! -6Qyi8vw5)FZ-}{}pkQ|EMY*Ia}#o;+;&i$q>TDfOwT6DT=MVhj5-^f-oUe{7{UCLT57Va$PTIg~w{Bz -OOnU=Cr{$R(|t}5^R<9j7*|GkpiZCP*qqil4;Th)`2DsH&m`A2nBi-H)#l{~AsQbMtT3wvZC^sxb5|M -KhlxMnYSRq?w@EG$G*HgWnO>Rh(CB)jLtNc*U>N=xSN2Q#@En1*%yw}GJ<3XW+<&&RY -n}n!^Y*@LJfu(=Z^Sd8T0CkG_!i&C0++QGZOyB=taI@Q2C7QHTCInFRo3?}{NeL>R(z2b^w=Z6v8Ugf -!#U)=(3}kA5w=37xhQy7=~UDYWh?8wsP!zW%tfVQC5Sa>G;LZ@^wz)j$8T^dT2W_0%~7d^KFIbSWYpM -bZ_4J{p3EV2y2Q-A#w{Pjgy351Il`|x(2zvmfvOZC1<3CLcT?L>B;Qmm!&P>a@`%I28nWow+_m^Px_mqtsbs^hri0W};F_`aAH^%j>^bu9nN?3yX(j&0b!=W^kK -592a<$5w&F8;KmCB{~fP$xC+xd;UOgwG7cXj7aH{WgRC7qeF!YJez0MAnduwAt*MaQ!mJ;j<^YYuPz(Wrs4aS+GxoMUl -IknTbE9hUSSn50G?+bLS&UfgPd{>xtB~<~WSS^IFoEQ0L03e;1QrxG=j7%BIOr;%5A09LUCO{QZw$OJnGmdpAx%?`ZQFy#y_LANm3X)+@j*{GipyA&tLp*iUX!3WsB+Zp9noDCUrLU%Q -G78^3)chB$P7U)voA1J7h3tg4*m7e~T>1hWR&G5APxD;p5>A~Ol85C$cr98ZtQW4!?N;S2UtW#+C2&J -mnR*uN=9YPvyL_QS=O&RiB+Y>F(KybXwkOPn#28#^6+N?Yi3s7YZB!pMgS=Hsjh=5$klw_s;OfkWgDX -3b272ArS`(%e!Y6CnWL-&S>cL{3C4f9oR9ljg|=O*yxcxXvh6PzGYHbGbnkWYiiW=wO%E${Y?vMNHsK -_``hCzZTyJ9-wFB{>wPl4!Z6zGq`v9#6;=;FHFF-d(Bo7NHvz@u4k|ors^li4E)P^Mk;gS=X!>ad*-) -oUvLP}9e@K&xe+s*M@rN#xirvWO|L*RvU -8(17bMTyrrM<{#+cnJC~CI=sY|^~G1iX2Q0pKQ-GHCG!pX+eTYz=$K9Bw4;MK*oRT`N^`wyUQO~en}W -?ahq!?>8HQHnf;@D!$81CS`No&r@MP_RDomP(PWM!Bho2;IC?2-Ks={+`c0T6Dg -ppcl2b5*oXLXEs3u)ZGfit*e;bBcRx+4g~|OhnfY&+gQ_v^u4HyyI-COXk7sfhrWWYM5<)9(QvTsV+i$0RI!i&CTt46mz -$08-+pT58^Q-loD!Uxu-M}jhS6}<<8*DD0qhMDd*A9?0h0&s~y^)#<{rbw$btj=@5ftQOK^&mti+Kwt -DZKMphasPj3(B$<<;;V~{HNy=8{VYH`Z*kCtIi#FqUH`=dgW9a5OXensAKf^4w^&v7hMBA-@IFg{yUZ -CK4*`i_2DqN+D3qJOvd&fU>_3>Bg|fZ2egB=x#Bd$DmEHH(nG?1ntJm2x0^icZijgcCN(Sz{3Pw~4Zc -B#MF=fW~)xFf#i?GC>Spj}jVKo@eSr@*qSbz$BE5LTj$6S4`$SY%$THOID=w(|i^EraC^^ne+=fwA-} -kx1FGQ8VvEO+?TeD7p?erhw61CH_dEYhLR&`4#PBunknvMX3`@WR-tmhs-iS~TWM8EbkJm#7}?MU>5f -5H#le0P<8^YTX;+CfPcb|dOGl|)Xu6S$Jf+OpAR+Y|O+gQNcTZfq1m&h -Bhp3{hW1jhk15^gt}|$K=NzFtj~{0SjeJ&z3-c-xIY;+@8jf|oc_*I%R?0Wc -J?1QS={MtL?ZD2B-Ep?%megkeM$QN6A;6vb;+x6viX&^cinf7K0JoWWF3b2M7N{vdoz;8B-8_MYcz$? -(j?9haSiM0mM0&UGY-D9}E^v9>SzC|W#ua|ouNcz^N+DV{HsW3cC_rnk3j;|U*iQ1WFo;=dB(Y18G-t -T$jhp}8@64SSX%_+dQUZJ7%$)mm4wEE#+Z6{PRmWA+u^oP{`?A`W^`6}wgs3Fjb`OjDw?Vig@(-db>R@cQ -KIA5l|7`yvE7xGaJwG_{pH>LD{xmpl6%WpTeew#E9!02fm(6HiG!?HB?Pmu1+SaBEO=G^Ci0pe#cH^8 -D>?~o&f&91Jg@aFK)1h`)fKb<#4MFH;>HiOOAf -*GlI5@!?TcrBeLRj_Y)eJXG>6RX&7teWC;GvHj)q`{ad6B5Yp2U>5IE4X-r{(U)zcuyQu;6uur1dP-&W$Pg7;lb75s6< -PVIbLEEZd_V|mfkio;Z%$5gHve0j+(zGK%-Q>_;a57O8x0VN#ET2RhY2Z)Vdbd)SrHO-9#wHtw~#vHC -}p@G*e@A#1c@rMAR8YsJyB4d}?K0{J%E}m!495Ma36Udjf<|2)Cm`TBrkW;x_Ix*y->EWF^QKT05(|X -3~Rtql~`w#oEsl|Fg^PQpTMC97-=e(ElAG;=Tnoextqgov;Wiez3csF4)(WtpvF)n3))S@3myu(V#_$ -@Chil}VBO}QgEG9QYlJSZ7@O*Va3XeT!U(DRhl0%_Kw^-x_@ko;Mz+}FI4Vi6P5If~a?|B{t5hquRMe -tG$xfkbiB^eSVo8qJla8Xf9uq)WsLqN+gzNisR7{D15XUx0hkcbZipbrcKv5qRbjZU -NpAC)#BhrvaGpjWa(U7)Q2H1Vh0`!c0U -|Pi#|n}$HBzfCN#waiz`K?qk48Z}4vT2v=4WwPvw9@Xv`8rB?mkDROi!8AQX}l4c!b>QJ4Ia*!Q7;q) -U4=D7&b`^Lw7i-T_d$!a7K=+j|X~7@91|1!IRC1UgYQ;9zEL+rh}N=C3=$*6Ox$t+`_NmH@cVB7@1%D -GNkpS`+dcd5wbnP=ZoPE`^n008`i0HgrxJmWIF0OW64(|K?6G7jw^Bl{7z}bb>GY2q0{a?jbN7pMaGB -jJg)z(W+P_Ksns$wr&-s(JJxfs*OUKgjf{7lW==q%NC%G&GLP>OY4SrUB~|$!ES?K{v$0;Sy8D1*Obf -i()9E}LHth8!XLN;%u*y9U?Gn2`;bhIe%wzvOc}O~-UY=4E`z~;wJjAk2tf_5Z==x@pl1V7;UlZKJwz3%&$zagKzl?D|Iobv7a%dP`Ii0~*_%ta}?<6Vu|7+TXvPa*S_)Wa3}qj_KB-zchM0fwutd<<8`I3bv -$y3CQfT(@J5Y}K|EEt=LET|($ELT{0hkG**IH(R1gbN0`Gn`>C(nAbx=0l1eOL{ZqW -_ori(`H`ODd)U=N_SLMo?92lQ{;?$bDn1=K^} -Nwx}n&{p@3Fu0>~1V>x(LNI;#yA}oSr5I^q-}8>w%3K0x(BWEW=Qxt|vh%$ -+g?5lILqP_%yAm}6cT(g$9H?jTan&5l0-TwIcXucYV-ZmZDXFl~NaVIbkA_=%(RVmm+~b7og`B|$X=> -IsZC`Qa21(JyJbqvttB^>At6;yWHf>JV6N1SmJ;}>XaB9uXUD~AjwQo!|$omo(h0Rx6CJ+4ii@$6p)! -RL%gNvTvKlA&>P0qpUk?9#bjk8c9(E(9>R!4J6f;ba -!=@(VpE)MngW04)IPI(qHWT3SZ7jScm05I?Y~0JPNdIIe`s^qMeBkb&Adk~v-$is*DurLn|kG$$*nqY -#bjR;NCb8sq>=Hco~lQnF#7iJlhLi6VFuNVwR!b%g&)_38j1XHSQ>n@VL5UD(|6gAg5bz}bQ5yq3h>^ -sSHzIIZ(*NuUn3S4#vMXG6h6wN(e2HYSFKyLEOKQRNK?2sNThZvvjaH6z%VXOHBa -=!L$A>oGHQVJ;xJ&%h8{g^CYQhXAvBXkPi4Ag9C*4unB=x7t59)8MpG|*$elns6LaQM0T2n2qHFs7! -Hk7b=y)eO?7*9ZzV)BLS_$c&(Sy!bwj&JnghLb1$UPWE$*0RoYyq@^ecg3Y%8y8}0WBUBr|D -{|#oNKM|+VC&@GzemPft0NBtpkPf&m&3_v3ixa@n5Su=1XhPG@cNpCnZ(DS|dX_ybb(0=EK)X~h0Ja$ -ro`%R_mcXCwQI}3f)F;D*Y3Gu>H-~q9Ro&Z($)q?YE-7BkWyKNEk0&xCJbYEXCaCs$+F%APE32D -@HWGjLU%^u!RzkgMlSf}&AV?Q?!R4AolHwi=L*~k~%|<45B(=K+?vKA2en_MwZ<6{UYm=PsncobVEX& -@@Bd>`$QA1k!#FQqqGqRTFwqb@#LAcO{HVxa8hYQ*5THcajceg9EEL$uNN}fqo9Xit~R#k+bTd548LP -$gLpk6Henrux=m1g~K9lq&{c8qEX@@h*~(uX>5duN!U4-IpNIQ%kFHacwAC(4D}GnbZ&W7v2jE?z@Wz -O1+aNVX!mRO$8RO|=o{&%(KwOIxxeszw)nHjpLh6V;lW~x0q(5^8R6OUX; -ac06S77JL7gTnHH-2O^_mx8Sp1U?XfJ7Ls?+)Ln24pK!CBJx1c6o|#1=g-&*R||ajY&ZDsfySic*9D|0SL1DclXtsKR$n`-hTM^{rkJmcNu{UfE!L#wj#BL{TVU9 -8cC8P4`Az;iITyFkSSwET-(GLip(=;Sy%`O*v(bUr&^Q83J*C2c%#a1B^+*wrXX<$T{D88}i2-mr@bvl9o3bPa_zQSne -<3EMbktZ;x{U-GZ?iN?EV(`ZRM(fcOl8;P`s8C6WaEi61YYn9fy|WW+T#X|S`z1md`P;gm?#fS -p+~0q^`}C>$=iU4JsZF7U{5V&K_{rf!1+}!a>;+XvT?938-Od!SQ#Dc?EBFSy+yaC}8M`uJyj?_~3X= -=eE|Hxb8y>mrG~lU)+5=~Ohah%#qmPu^=uWU5X`Os9<=OZBY@HRGpSM=w=eQ=xiVPimsK}_J=Fdy==u -Gurouzgky!{-?ycfKkG0LBZkv)t7T(q5;75JqWh~GxyzQo1ic@?olK<0SdgwdOkBY$MUW?Ni?`>wQL{ -0v+w%f+ES=6yBVN-NvRSx(o|rLQLTa;D5D_Tjv%JpwDZq)5e?b<}P*QgXM0DI#LD+x>#XzwUNR{0`*s -6(>;mfkwiw7+pr((h{=J0S6YZ@BEc%lY&%8>cOwkct1-(8g3KDS_6MK@MoovdEd$@ -(tZ#;V*C&{-|J!lfkB)i>nv%}=L -qm;W0%kHV&b|+nh~^ddF9{fao%Hl86T&nozFVeJW+U)d+9Q7AyV#HGZ*O*Kj9i?)4Wwb*yWZ#U^BoB7 -LD>WG$WO(%yk7*nu`~R|^t-SxDsPwD8Oh2&B^UYTdFaxy6l}G#y|3_B<#_tCZ!|A=D}TMUx^G~B*?Ql -s@JYZMdF^ek(THifJOn1HJG=D)GO$bpX~cUYFVhuuzp$E^Zu1j>Jo+eEDF -%agR>Y2aSL8o?kKl?(8OO_cj8TCyJZkDO%$c;~h1;EW71Ambt|BkJX2N&A%Ma7$KxO@I58)039z*e`o2BB%kEp+aMBD}+Bq|pcmB!jU7ctzWx)nXh9qbeo1P>-7_zfo7PMdn^F -cdyIzd{~!PDxnA*-*r{v_5`ma@h&gy>j^%-Qng%55rA)j#FHwDWB0z&R^aj|I^!^lQS!3I`ZIjXz|ir -aJ@6aWAK2ms3fSzFom@aG2$000ak001E -X003}la4%nJZggdGZeeUMV{BO4q%gVWiWQs#kTEXTn%n|F+ee-Pfd_IVw!_=r&~w#&h1RuQx-!@>H12pM3Ln -7XPIR0wAqRrXpW(^&}?DBS1fIXd^}=!uJ7+}@2Bqd?*97y{PyXAUp&okuK3mbp0{?N7QZZRKP=iy@8| -cAPvOZ;KlSxaQ*&bZfUXP4}9whF#fHaebsWnYMyuR#}ZRYk4Yt1)dn=^hht%r^kwJ -{ZrOoVp-HZASleml6CTqx>nyJ?F8GIPXGLikMX|&30eXeX~C6dk*`zQyyeWf11-LmEBopH@Zk-x)m$4 -3(GajH_Dd3*&69PtB+#QE$~v!kaVr}FcO){T{s}Adkxy4%QVs$iO%9|m4-tP`4izCw`^hBr8a3_;2aF -+8;MYaC@7MO`5FT26`3Y+hFkGjmWtyJao4h1kR@IZ)zwg&!=G;gRuND{%ESlkE;$2Q~gu@ia^nUZ$p&UqpTyJMsn=WOi8wbr2T%qv-$!IWU=9|&UhB_Npu2==)5*U8XHC(LnhT~)nc(`P%ppLuMYmP*kLUBK1Zef|e4*w -?Ore22>l=nD=9pvnJ=-DAx|{|4Q7f24@7o_=>wJ+Bq{^i2LQL^62)c7yX>T-N6Ihq8IBTj}z8!Y=ld{ -+KRI$&s5o?7+Ol+a5HwLQ(n7w$f6qNYzabhs{yg8@7~@C=j--BZ^D`j45FxRok$0C^PVTT$@-wWAJ53 -MN@fl2eB?4{}2Fma5J=Up{LRnBEnxg=(8W1p5f~H;_0Wu0NVsY#{!4A#s?(6Tgt!4{Zb1PD8MV5c -V6mF5MoN8G_K`=>VOeFxMW`aM~JcN}s(r+-`zdc&|R&}l&QdhGl-V}EuM0l;b%j9QVTo=lI?=R`WuW>(>@78+ygCmC-3(x`pal80BIlOeh?R8StTWUJ@lqS~F*$Gri1vn{zvODu%& -|KPZ^#0a5?;ol+sPchL!eHqjurNBr^lRAP3R%q7^HR2?rqntGAX9h?3I}HmZr+m> -y>ho<;$uMiTiil*cTbPOCknfN2?kL?RyMa>M2D4VAK&%g=MRq${QUB9e!K8UE8iW&HasV3&-~q8$(~# -U9{MQIYjx%bT+E$>6idLknXQ!x7H&j!#*$)Fk;WYyfFgp&K8g(0jRzI^WnFHeWJDl=)8N9-PKmAWFXM -{AN!aTyI)*y*6>wN|5#orX&qc$Dw%(?X7oE>q;{%X`K8|kF|28*p?5K2(<_LP6%U4y~p1plz-!e1~3f -7M*ih^^q3B?jG!~g=(FgiNx!KhOftx`8CjKfFyIAlcD@qP>uMbo8}Dd=05Z$ftIjLAhFX|~E*_esGOK -O_?>D~}*^xu?EJNX9B(WK_5yL1E0dDn+cPY_E$dM;yWCT6w^}we0j~wo1f097vz^9BAnf?9lv%jHdgI -rPD68r@ieGYsz|K*19?JQ&g&_vu5%itH||7)B68Q_&CsQP8}!huCqOB2SQQ;B?9-UrzFID(6O#1tq1( -2$d{+(?jXkUreD5}{sT};0|XQR000O8%K%wh4L++KLLmSEF?j$09{>OVaA|NaUukZ1WpZv|Y%gPMX)j -}KWN&bEX>V?GE^v9>J^gdr#&W;wueb+KhvX}X#PLhhVWUhHMX}ZA*cwU6Yev-|5P2jKfdC%>O6GI@zj -wdx3j}4Eil?1sCKd_Y?d|RD?bq!sqbNG47ip!Zs$4Z$S=1^mrfOc5>s5VlaJkTGRb|B_Tjg3Ubfr~Rt -EN;_JLx9kEN!%E7J7Lw&&!)M&o{Ap3*9Q6)@7k?bXLsaecD(+_gS-0 -=>)$V_R`f#7t@}aqy_w6RT~9R8+~x2n|rMbwMwg6*RfhzMOfz=jBIUS9lsc6NSpagn?_d -3`oiZwXoR_a!os`ErKcs-gNN&DXkSXl7&sZK|h0FnPCGAra-xQ2{G=3~Q0zlns=P0#uXEp3tv4h{}LE~zTbYMt=LNi?oSpO<&s;ze#Uwrsns=d|wzYV@6YQx~G;0voy&jA79ZEai+GB+~4CX=_W}EkaM2&!$<0PgUHtRAgNH)WG>$<^xoTm*+Bxgf_p^%~?)Kl35W -1voI8;nU;1{*-$#D41#&A%-)usgJ@6lgnKV0w>uj7HoDtrNB?)-?|8uUwldNe%4}VZAqfOaCkT$*o2m -v+M(11N$B1M0Xn@LEZg2y~%dG0a2H0G!#fPxHi}HX~5LG$%?evBxR9raG`uL--l`R2SfEf8Uw}0(Ff* -{nx;F|XhCVP-*lbGf&hU2K^u#fKKavOFeGjoB)P?TUX);d_|Sc9pQ%%nRiLfKIt0Ec?~A-lr?tpQ#7= -pOQL81e5)A@&6O63|QeCdAiKg~d1_q`zQSr&BUg=3T%O>i6p+Q-JxmKe{+zE-gG-AFS8 -NP&FV4&e=V7VcZel%*aSR0VeXj_d&MLB}$$x;B7VCDh=xnRT`aY$@t+q7UAVi7Rz;8vDFsy=}9hQK9h -K^Ye)fCjO9ES#0iLYZLS!wa85r@*<@y0DDWxuAXrv2vzaH4aeu*n$$e$DlJ1*qY3PjVBH(1*@Pemm2Z -7E^-ZiN3;zpH@pR0cEP%q^^=0j-sRu=Rc?3vkP=c(~PYTLfUPA>p;xPay^DfwY7F`2>cxfMjx0 -O01cea6|gRFVb!M|-dN}gv%C8Tjk)DUyCEkz22g2vc^Xg3)uzAQ`YHr>4ZU|0qq8?4CB0h91rvSd{yTVP1^Kh!`N2Zxt7L2g6<2W_oA?I;mQ -V_d8d(Jvi6bC~d>SDT`8EZubr^XMBr@hd!jFg1#iG2Yk1;$RAS>Lg=Lx#-d!aL;{s9&aVhb1W0KgV(s -)eKxo+xgO5sEBFJXt)>l^TNKp|=0f+3BmQ_v7I -6B?k1BMOqbnS!e-vUosrF=FHiq%j}@>I88A00NzZ@q%H^jGSI8{1Ez3k@-!LhvF`)%7Vg-cWIU3RM=Z^g9ztPQY{k(K6kSS~rzWZ%6e8cy~DxDOCrgd0@zoQHn)EKn -!4xh|R(7k=r_;$?ORLJC#^rPgZQ#5QIJyKy?r8hFnq4|tSetj9I2%nQ|I -OI^#{iW01f9HVsm&>Rd=>K)9z%1fx(j^o}^t3GP3qT_?KWc>a4{t;j@phJ83}aL8Qj -bEjWUZ!2r@882;U1NI(Mo{3nm)6;E0b*N`K)2iJJAof$}h8Ct-mIYH7qQ_$AY!lBzR!xL#$)J= -M};N(;Ln2*=z`o=>4_o3+7{mZwEg4VU-?9b!&rL!b9K4|1v`7#D4-7z1N5r#eHAuCIzG7P^@)teY^*} -_s&w3-X*G87NClcZjlcGpp$J`0erp(8$DA$knfn*x8%SP98c=e_qNpYAAY*j`ys0lMO@3}iozgvV42U -Y-k=qhK#Mz2)O&6Zfif~sf?2eO*tsXBjO`RZvsaIwz+@Ij`BV1DNkJ3Gtud+SNEUF1bu4c|_Fz;B-w0 -+Z&u%IV@?|r&If+0d!6f|ebHCet*<=9>HUSSHtI1Z!*qLk&Gnb*nCx*(+<2r!nHqJlEr$I}0z1!?9wB -Bvn(#k&JZ2)d^i;1F8Bgt%2MSbl-KSl;?@GF;k{KXenNVPGhA>2>@{hAua@Z=63!rX<`!2GE)UGHm|Q -|Ff9?LXuBWLt{sYkxXFrSG!kda2R|Y}QpnX&?X#u*qsLpjDyo)vL?PGmNejn)X}Vxk54{Jf=&YI5C8= -V}3sPr%schVF(uE$SEaUz;fd0K@NtgJ>7`K<%)Jr8&!3=mk5OUN@pj+1$OI&72LI+1Vjm&F*Rm@cbgv -11A<&&h#oETa&n6w8Hin3J3|G-f$f3z15K5}QDB7CO`Wj#?R$s3L4ZkuM|NjFGnNp=2vOaDI027ox?z -A$1dK*J!iYzR-1!+#IT+u2+hJ&gg0&_|ivQf>KR5VKYNV)7tSj%C(K4+4Jh*mAwh$3zBP8V4Jngh&;; ->OX`PZ)p;5SVcL)Fg)bR-9#?g!Zl|K1G1tHk`9emb}jNkXC#qjtJ_EDLU%71%joEAPRUeb8k@wDpa#9 -?njF8fn;6Q?PQ6j}(H%QxgmZ-K;TE6#J7uohi|ZpeT&9Os9WlX@ce#TC3j0#cOr!f&E2S{1N8YL-~xZ -=(L=y?O~Y@0`)&Rd0D~7rR`A4B+sB<@-*raj1D3@*t=+LI1l}45R<{+Q5Wb5ORAVwA`~d|C|OteUJ@{KqK^QL0FZmM+N?6~i3H}%fNh5S`TTW@ZkP -xa*K}Bl9ei?qpMrS;Juu3M)0Ei683%5_Nx*!nEr~d=`rs0?1;w@CtUt1rE4?b~3`|l!iO#dMka*PD$( -j24%NJIkkN_ss{7y$_rvo}<kDcw|QVoCLAIZfghx^At8qkju@- -6a&=QamU`yCg0z&j}8K>r>+bNjOc>pt?5JC58dXtA>ROj+}jNiq!?Ll1=jzqE%}Ot=7BxdD4|!4|u+J -42ODedo%9gr+YJ}q#i=js6K^A(Xhvzoi^an&fv``&x+fc$tltGK&NAygxY)pqs+X8jQnx7IL3kBmnke -`g$yZ_n5mSva>WW3cVeLy%`#Vfde}eas_>=l`HhNhvAt=zf -t82y)%lUuN498qfn9flENx;hAX<7|!kK;M9X~#FZlyEjqHzJvjhqVYZ~-_qCbOFUJ?CLotbtS6i}XD3 -}2z1A1?ct#Rpr)oMKDSd?}1DN;iEjj-R=g8M)pwdQtxU%kD!d~@{E$)kkCPib|_Y1=pu(i{JdCCNFfD -0hfa87h5;htnVm<_ncIKJ}(0&>4xeNbj_qapMvc8}fcG+9dNcFdk2_Nw{4ckS*%@xe`y^FJ6Gi(e2m_ -V4|xg9|<{ao*|*y2>}7Sk?xb^M?a9c1T^ScfsKFrWZO<(`2==r=WT2eZ9C265u|H-{dRa*d_yGUzFsuN^Mi6lt)l_2fYosV)#$pEv=3TNyNoxkEz%(s%uU8);32Q$p%(xf*v{xw-G*aT)tnQ4;oekXa;j7E%WB%>m-fZ*Q#9`tn{3 -@5v`zmGtjcRgbjAJ~a_cf_^c?Tx!OhIhi7C;dTaXkxbP4cLw*{w!Y)y(J1`=f?i9>0IcGSauR-KzpSr -%|`59I@diov)9EP29%?3F-w3L=({hFY_aNXcSLiDFPoY3?3z1I)>VaWE=*mz_nb+aP)=8n1I^sv<;;g -HNGx?!Y;5x6OUx=&tvqVJp%9ytM-dBiVXh9YYV$rA{+vb2{PVPJaM0QZ6jDMJbJ$6MPTuG35oDmpg6n -dPtCi??lY9+xaRP=QIPqhKjqT(*-l1y}ylqp)Yw-V(He2(5h;oKDj$5`gQj0NnTq2-9!6lOp>&j+!Z@ -X^6ARVf!B}Y1CNe5RvWO;>SZUt6Lo0%=y=qA}FR>!R9(b>tfEoINgzQQm9`A>BD%J2y-W -UP)Rn9EixNvfAi8Q9wCxt^SG(OGc|t)a<&r#^TBvEd*DCeFiD@H(r56j98r%)u}mP^x}e&@si*clI&) -QEt}aozxVs$nXW)(4k!700vqNdx0+R+RLhfAUTPcOMlDKy+$sC`IorHL}7!_9JWT2Lv~kqeTeUI@8n -Jo%*Pqk-uv3~oLRbuN6iN2)*(vRy3--pJ_Tr>x4gxuA94?0n({6|F_FJD!+aMjIheAn;nBw3a^aYGwJ -zv{fRh?y^>eL*-5#=HE1*`(S-EXwY2boU>W10IT=nx419?5zpu>y!Fzh`r$}$DNB?U&q8!Cm!KC8{A? -QWnK6(Dnr$GQRjGqD>LKh`)(nL}VgE2O0!Pjo!bxcf&Gmtd7)IBW}0bAPzfbBU#gfWkU?%N=bWVEo*2}zlT_no)y@}Nj!@>rl;5rOx+Of;)qNIm5^8lU1(1X=`+1t2oJPrO -V7m76rUrE1t(+@3EfN3mQX0A#>t=u?}4`cAzVZvjNz5P@b}VkEFxR9f2=VbIudf-=rrNkZK0n}9Z -{(6b@DR;oK>Gkc(cve}Osrp&`0V=7^{5>rB>mx2&R*liG2~EL)WEcFZMx53-5b7rMFVos -gN~9_`&OmTlBU;=J%+BrLvq7%r|XgN7I28+f9BqPR&aE=8>O*&Pk;YZeSnV^ot~#Dso -_e4Srd%=tF0%g?yyyxAX^9P#8SX~3=dg}N{Rzoihn-aa;H(RHnRS7ujBn{uTRPp3xFf67MbcQ!@@E)<|(?E8P3y_grDgpflilk&+%qXFlM1REt -`>1o>L1AHg&D~I>uuqkZm}d7x9_|tsc*yY-+_EWdi7ST-ca>^-{8L!{OspZ -0v&tVsu!zM*5j5X-9nZ(b6t5=T0TWE;I94j&dd#gDo8&Lgd+o#FkFA4Gd<5F0b4PU#)W^;6E!!LErKV=18o|O$*rt4GOHF64WmeZCLjwORB9>8#93_xU`G})5b5vV% --uGR_JULD*p#shD3O%C9Vm+_X_R=GwZzF@0V%fIfd1n4ucksne8maLg*}S0&MG>0MD&A8k50vgf@7Bp{t#Nqup^J?zk%w6oBmr1a0Nw_+rD$$LWwAH2xXEZ-)aZvvE6~Lg@+Yo^4g& -2-|W_hrYK?vTPBd<-r6jcujRL_Rd&D2qXVQiP91b+?D`sRXhuY@B>Fzl8cN)7Uyb$r#@+uqU-W4p6d8 -HN?tJ8VzSebc%l<=##a-J`|5lNv`-LMzL;moWR80d#9++mfXXX76~8qH(QL4O433WryWETi>PvSPxd7e#`JFmxw}qJwrcFjCbk@SA~ehXSavD2d7GM;knd7# -yfFtm%fA!cjZkU}mX6>ALJZ>cli9j!e1}J1f&aj_?eLy8T`!sfz;ceJ9;I8>DO(mfUc0EkSAd0CgP^; -&pgdlZy>Abhab17MxP$SHYo3!VYpE`Q9k@JceTd@0eKAei$CA9R;avhO!iEb)I&FL4>eKp|aVFtwJW4h5oQ(NL5$8 -b@rHOJOLpBj^S#CiV0)vb}1Mlmpo0(^}#M>Njw4IN>91%4PnWAH^`8VUM(%vB6JTf<-q41_aCB#=n_j -Oz5X!;iSS5ek1iD>Evk0C5S}Lf~1*guII+-^sI6}T^YAmvF_BzI60KUD=7aWMNCb0C{Wji2g&o -(E2k5PLWe2}3B&_{s~PAs&bC1$dtvh)WxerPnyFRFb%0H0Tt9vkpRFs6S48^wU=S15f7-?%R9hUkjF| -M;k-X6d@CYr}KPFtcwJb6F@%=t8glrCP1`6F}5AkTP3XDs`JNuKQYa4pYjK0FDj<0>JNM -7|^k{?pQ%QoGJB#@N7_Bu<#q@-I><~s2tY!%v+6rLcM<}i@14cQujEpQf+AI+2+hwYKGFzR8IzGQESG -?(noI(bUP5{6~cSL?0{Cf&n*On(GdFmN`lfNyxA%;>{*0lR%wW2-b>FrU^KofBb;FJxabs5;cTg~Y;} -!1Ku-=7fUV-BHfdUXuELI|JIqroj`SJ5kU7)g3X(%}CL|Hc}Olowb!_B9!&3Bkpf<5ELQpPE(^j)ZTW -+F$s0#ob4=eV`0D)la*e5{rsCT<|Tdve`r18)&a&2$YXJBaY9%Se#`XXGE|9pCH3bx+Pl>oQMl04s#; -Ie9y(16%}A{sM!vH3YFy8vBjeVhjCwDqeXu-e#2w~A{PpD5!(WbG3wR_b3_t&9BC*AWdH}PAE~>zJ|G -_?aOm1=j7%ym&c~plLQ?uIC6z}7^D`Lo`AVGUPsP;fZx9bk8#cvKxFe(-D$ -zFT(*!cqtZe+lPn=~9MdWv3d!D^}H))4k~jVsJ%-1S<#XbHz@Pv -WYj!&1iUbF-I11%pT{Ievd4MRUmA&035iF~KHl}#5qw)|r%TXstD*4YU`NE$~N8pK* -ja~Rcm{?Vkbz$gk-{YNdN6svqM*g0$+Rq-8rHtvPM63`!wu+w3^NWc2}U&gI&>CI%Sfy@2s@Ip4GWl8 -C5!SBsHW-44U6VM@q5mW_xZ&}mEeE)$?(%yfFBl)p}?TXnURePFhn&5pQnV63ICpm`4dYGexJ2M23G#Q6D6G^>S)YX|* -o_zabzkyzq`Waf_{8|K+xD%zEi7s13mq!)vhhQe6r{zg^?Yd_lvguZP&Csf6%bCqyfd!$VoLJI>f~LD -LBny2L5U-0$dW9eOj%pmvrnq08;~cgn%haa$Llg#MzcfhRXs}T^zO^5bePz+_yKHjzZSR$n{=wpz`*aVjmv_EuVlQxz9$UXTUW!pHp!B!w#piTF -BB%eBxtd~itp_lh3QWt0T)bv=b-z_gx2r*fh>Phj>;9*gIbjNbVS)GmH-mH{A%>w3otm+qR#jLgfC(I -2=ZZFnPVa@Epl&;MltMnqy$?&3S|q4&q(H63JgNN_6@zX9AXr9uRwr^a-e%me&x8 -4i=3K#A)N-%!=z+(}%M*kgWJH2f*ih%+IBAFhz2O -8<8&4i4l;QPWUSBXBT3~&-2*9(=^h02C%txF$s_j^*}?nfD*ky7!=wtqpPqVT-WB! -;J7G2V;7PYGaZOr6FnqDSy?i?P)mSaRe=zj`AIW%OxxGiQpDjD*FGx_2uEhrM)ID*9$UKby2T)4`1QY --O00;of09jiI3LIfQ0RR9o0{{Rd0001RX>c!JX>N37a&BR4FJo+JFJo_QZDDR?Ut@1>bY*ySE^v8;Qb -A6`Fc7@^6;@A;RPumIMW6~02e=S%sVd7&Hf4olN9%2Aevi`-hlbMb!IF1p$1@|BP8(AH#ORAPm(n?ylWCdWpdHq -rW2J-+U>n2)Rq?+?I00CZoONLR -{NJCpyIvhXRgQ`buJd98yNc0@&j>x-_^CfN%7b+XiCM?zm5QrI+885)=t@2Wyd!(YYkDeZEF}@&P) -h>@6aWAK2ms3fSzFfEZ*Hsy0031M001HY003}la4%nJZggdGZeeUMV{BMd?cwb|0ZEaz0WG--dt -ypVs<2Dxku3tfTK8%gEm1&D@vB7%LB9l&oO=kv^Y=L4h5EzNJxsgRJNyTv=|Gnq(p_k?9VzEMi$l`tR -^4u3^vsrSbD$8YMSg!VaUKK1axiP}9!%pTqw&zDCD=Re5EjviNW4ywLLOinzW_N;B!XA`7!ft3S?iOkBq2wG*eW$$HuCybG*V&saAmI5ZE$qbH4Hd^gjmT -el_Xpv=s4TP1l+=9m>N%&T(?H2OlPi5kpIT|l#(^R=tsx(ndshce4`A+zCRoj|t6OCg(@v;$t*B)y!H -muzL{Qf$-`r-QOpEqyc1S@MK{G(*l+Bus)XMLHiv^vpM)j@VCzP)zq9#QOZLU9g(soM+m#mLq1n -!7uchG+B^_j^kv)i -`Qk!wy_mu;+|vMpn6Shh>`DiKaM~@OicUuZ=o~&Q8z&$Jpq{y5x4F^gehnST2gs8Jyc^V5bw7cBeka_ -S5$t-rc{udUu=M-`-_cw>Q^s?-!9VEfp`YH!U~EtW>$r3-b_s>0u|dm@hnePz)}B(Pdp0ARHnHp(CsL -L^uUZQC6b=-}#)4!vHXJ3Ogx^ZDq*O>wybL+e*+jYiqp|waAP&Q3K6uA+UlG(M)GL -zc0SIL1Rjt*NET~3!`Dt{g3!c^n82#~6_kF{V?HSDE*6Mj+U9wrFyxp@U`<|CZ^unGVNP?hKwpI`LXd -jCK7qnq%C~0V2%Ig||G{<|#8q7Y`2K;E=^Y4jLuh0xN_OwAj&}KTO7wkh9LQHd1W{vX9sfIGKukFzt# -5mx3E(T&dU=R)n2KdjL(zakFbWMYy1KI!vZv=`n3u0D?2foE=2XVvG-VCuznQe%w6L`bF?zf=2)D{U( -G7Pd{@nU$iFQ|dh;w55JQ9&2IK7!#7Efcn9O&WH#Wax>ceQ8Egcw^NbTsbexBa2DG?NxvLt~!2ljm*$ -=BuBMJR>1P|)ToN>tgX%E#l;rwUh_a@{-SpE=wg4Yj~7V!Fzw=-umAQndzASjT9|4ehH?&7p0P_Qa@; -wOkT4+ZWul=U2Wh6WdjyDOUpU$~ETDB5CZLXn+CTJ)XKBvD>wjxZBDZTgvO-di60=YHt)&WXB(HT02o -+`08bOv)-9w6uON*qB=UFMXJU_0e@|c+ug4z(36{6OtkSK)N22Bt8S(AwhWmFU+%y%6SZ2hwqsbv+&f -S(i8MY})H7*d_H<^yiG$#vZLzZJ@2yKm$*0_^JP%h`f<-ua5G7#(;I80+w#c-q$vq@FuEED#I@IXA$C -I|Xm7BkADl@=PXOd(4l@T}&0rbFN$W-aDe+$-iJfAl@-qp5Hmf~f;g)kb -a+Sp`F<@eJ@xSw~C9X@de64Ept?0dR<)T@J-8r^(Fc>3*N~27wqU;R)(pKQnEjtTQLFcT(3<8ePwdN`~S?mCV;bO)ekjh?J;O&rS{P+MP?(?V7Na});>d@TlRwTUTHoo8rE4xG3W*aOZ!o;%SWKq%qEayT66Tk_Ev^1VJ7nBY!g-@rQ-{$a -ju^|S`C~k5JfQ2QyCAwR2y;ho;xPV>hz{20wJqss{p=0AryC%EiY|FHN>8nb(JIH*3PJ~o4HI-cP_Xx -Q9(G%t7+!WR_T2G|Pv%_ncrrb-@2}Og-Z%slZtxUE+M&hmOo3VhMaHp*{u52tjz38%-+^1Xk(GAbZDw -r33-%*_#L$g2a?YoWV-3~ZAa*(-p?BDEb8b!Yh-+ -!7TX)P)h>@6aWAK2ms3fSzF!Fyn+-30093A000>P003}la4%nJZggdGZeeUMV{B_ZR;v~-fV(4;_8cD$GWzUN5Eiqmufl1SveKKI<6@;v|K)Ly -!iki(!=ErYT~kO}TISY<+)=XsX3&hAB3wPP3^Rh3Zt!MY%%F*ZaEpJmgTwzbr>ymg&2-TZ3({NzvP$D -?!xGJbM1sG+RjnVgZjthFlT7;J6#gQghV?0h+;Z)xtOGSwiRCpW9bw+f1TX_{B{(v_bTYDY5suLeEUG -(;3;olk1`;GA`@NsmG4vtUDmqXc@EF<1z6JEK;16z<|#QB_9nf&41VvMcf43IM-Up%J4e!O0dY(I{99 -`t+Un-NT4b;d~^N7qSlOK>7qTjq@+M4!FwsFw$2ovm&;(gGwN4EhIuWCDa2jN(>@}HdA49wb)~ -@PC)zJ0T$PjNJx8X`}PJokx=$zb_F!ZID$kx{5o>P$;FyQt8RYnPw^@%ot<$P~P-H4-gkD>**hNO#i0 -&rYlH_1syqJ|c%xCm&Z0Y4=s@Y4$}MBsk0M*@;-aU}6*>@iuZ$#J%fbFh!Z2*bvL -aVc(FfqVfh;s7)z;Vf%z2G}MJY@R3{RGC7X4MF;%Na!xZ(Xr+Z!0&7;9~htP3TcBV2)_7d=Ds!ruY{N6x^(NG(0xWLJ3NA564Ty6% -Afz0V<-uNfXBghRg;J?Q)}iAJ*}{#JY2jGBr`o=>>lws}>>(N*QEX#b-Pl*$#eiB+N}m0+GpFI!BQxg -*xVRx5-oClbVXp@wAGbSd6?ZpGArf_sCn8bE7Kv&qP*Xmx5{7-8jg3S%NH>+gumsk|SF0U8{SnVNug) -=-=1$T-{xt$eum!99q7kd`69#s5}f^Aj|~O~wGJHloJS@nt&XpLn`GV}1YT<{^uhSK>D?$uSNQlgF;Y -3-CuC&!_CXZ6>7%?1npn%PmE!2h*nf@zMqlGmG|_6xoTfZZCDDfhgjTKY)zZl^2?;*cOnVG#g4%pnTG -lV#h6gsT76h%~-ZdqYCqDvA_{$?lxB)Vj(aHNm6IsY5RUj`H!Two|4<f4gmm*78n8ZL#n35z#p>^>4Ts4GwJ^c}=29AIFVq26R_% -d@KZ8Jhr%2aG)ICtYrvuRj#ZSt))6RSn^dzQ$?t<|(A7Vmi2cSG~QvTZ=rufM~j1U|K9)iL()~$?z1G -yoj-_YbSlry*XVoYe!TSyI5-V7d6kq;VPF6?Z&RiZ<(EO&PIRVV}SK!*iYPA$r{K@DaNU=DD>s0SFw; -md8>JGZ;O1hnO|~-aHh#>s)-|fvYOdu{2(mOGTDDnO9KQH000080LuVbTc;hRFJJ}$0O=3_03ZMW0B~ -t=FJEbHbY*gGVQepBY-ulTVQFq(aA9(DWpXZXd5u?5Z`(Eye)q2+R1CH;S!)Lj1B|)AmNi>|A`O}bd+ --c_mMDjdL<%I8#7OhscSn&T<*3;Vw2^qnV%;Cn&qH7wW+ANS1+1Uw!ItEidlabuYV>*#SfqCmMbes$xT~Q494%QqFrtDpR8` -rrD~`FGm~az(?H<#;Dme6OJ11NUP1GQ*HqfHs`d~EQl+g*v+eHU -o68GK^r*zP|H27ql3_wjSsiiUq-jeO0F%8ZT@r%|5%w*bTk9WxekM=O9)!P=neO049iNjJ2{yGUi4_ -TWJF4f58`faPaL;a(-jOc=&m{Xta*%_#%y -tARg<02KkE?V8)H_sLynxgJ6vsc;RBz0~Lr-*fRpFS0J&2>zIoYa3kc1aL;* -*8bw6^w0|yzNjPFOLJ?}N+C27j)q#oXAMZZ1K1r84_suUjc5=J!>&aVyg<@?d3w}r?NAmQ(5QPJz -CXm-L_S?6?Hv41q<4!pQqdJGYRb4aC3~De0rMov&7gH%xYB^V2*fa}EO-_=k_H71XQ3r -;E#rFH+WezVo|1Njt~JhQQJ;J|?#)0G0n^{KRq^EI^7JyY>0DvBv*~EUi`!Rt~o(wlY_*7^x(`y12?e --&|f^T>gO)boSx;IrX#`Wa2-!MyHm -5kM25Lij}Co9t{F54Qq}a}OR%_wDV~&)X)!lYkUn|bXjWIyV5v=EWP{?>D=bKcEv60h8vwC=4L>_I0x}ox0h3mxYku_~fV4LhgL82Dpadr7^MaW^z -1=FmZ@NMu2ul4n-{bPqnXd$hF3)=DB4P4+S7ERzMVA;j+wtWjQ036_k#8OVpL=Z1g_%7-BZ*jt9L6ll5_HQQhxIIve|PQo}u(CYLN+^ -%<{WvH`8-m}vnxS7PSri1?MUrc!JX>N37a&BR4F -Jo+JFK}UUb7gWaaCyyIZExJT5&pivf{;;Ix?3yp9r|T)IW*17UCw2qU0sSr3TX*e9t8=n~}Ns_-($%tqb7Y%7?Cs{`%Q+=Bx$zrkX#DNsWwpV?}ih}S%D>_AJ(+EX -zh+Hh}vs&zSyxF;1;HbH|d%5owul?Oox}S=DR<|8(&0M9ZYgTH`IlU^~guk$cbySJYQ8hTM=2wnD_HE -lSDb0IzY(W(F?s;=ul6T@Gtg375np)oG1qUu&L+iY(c`ji=y(~&mG3iErdiS??@BjI(c=hq)`;W5&MY -H9*zSFvJqh5p$m+|eZu5*isIiYrBe&EutxnmVARY7~TFK|H7clDA;)hTf$M7>y`#=%<8oa*x(Q*ZUNv -?v;S0E1h`th4phOksk2f3rl}YRQbn(`)q=nUe(;>MIGVHWNewzCH>|r7;5B<7 -CtF94zr5LV?LIiZVq5Sa-f&eEsbuwbNxq}qPOgkyzWeU_*5792;%Bf_!_3rpotar$A-UG^p_htWF+yv -;Ygk2Yx!QaFxt<$jkU3bf?6`Jo&&$lTP7d^Qfm{p6HQ9g`e|$P2&RK6mTP3oS32Mh*&dqjK7|tf|r|Ex2Je#|@YKNM@dN%&TS -%63LAGOfF~m@CG&)m0mgz$mZ6i=rxnY4Xt|!nwSZsF=kAJgb?70a7kb%(kUY%b!K?xsNDmTqTKkr7Pl -CMC)aH4nj3SbmWvyf1uL-el_dG^f?GXILU_+;1^!@T*WT$sm^weG#1fEa5G5w4WD={y&pMXF4gH=ZtK -XVrB1tdlj#)YM?-Npl{`jhqfC7lxGx!Ud7fpSHKa4c&w&w8v6;rp2HHM*%K%gaC-hrj8K;iN{qehS*U -FWltALI^`#c$0gBkkQiEX&zNXEmFwQ4*OtQ~_qjr>PbR;yD4XSvJ-dFrR2BkjUyOcw&;*f3)LGm39W9 -KyP5>eO+s;(utY~*fm2IL~yO^bG&Q-)9T>i7j<~1^GQ?PH6J!;M^?yY`C?QF88kp| -p`*>uT+4viyi`7zReR03EXj%LM7I|5aRPd2uQ -P08WB$HS|cTPufW}SQ$q*-)PR|f7n(YNCwkXN5qK&Kuk?uR08G~X}CFoK+K^n_qJ5>OC|Hs0|Y*o|G` -o`(=o3+AGNR#hzUe0}X!kcYG~Z$$_PW(aI81Iu2Td8ew^kImD+w4H!>Q%A2-%MJahfaI(5gw_o{S&tO -7iInpTbH0vsD-3qo6?eR{?rmBZW}C2Ou@yUsDg~cH0%_K(VT(hr<;Y~JfbrHWA(JSqn0Ux+suZ9cJaL -j()2VY#cEq?&U=zm7atNAu-_jOA?!W;5i(i;!!mFtFn7XM0v)J9ov-_6*l-Ba&w -(upnW=9LpAYlznKlp!Goa&CV(9do=WR1=A^ ->;d&ur0TKgh#d#2PvCGr0`KyOK>$GU$FPNltw7OLQ0Z1h#H8Sg9mdVH!Q8&%z1A(NwN@EuX|vRxVE%4 -GEFRff=Shj$mKea0N&Ci%L7*Z8-Z^=X_ZsaDB{!n?j#tdcj5}GftHl+qljYzfP`a!jVpd78%bk%A=+3 -yg+H{250TlLpyMG+Oiqo2z$`x=m=UH#$q{NWg3tiSOy}7g0_tw9E9&r# -54l+*l-A;c6Q32VId5uRgxMeELH)F>-LiJNH9Qnx*G4+1Jkq-!&S0A6j)Z)N(>)Aqd9CT8}j^Jxdq~o%PCup ->PvIpGB)_U&LU4++q;cx@eL^?3%&VYoB@Jl6m}PbD@^UfJ3l3feJpT`F9~+kymf^%FI$j5<_e0X-{iu -$!8p;WBYqTNhfhIGXpofOF7-i*1+eB24QT8};y216MZ-(GY3Agmz_$>%O4R_qTUKH>XbWIppC(kGmE&{vjuPb`ldjuw!|YKXVVjefrml-_~K%0t^`!34=3t5F2wD#LVOPFlW{{zr-83=bL3p-L|)mick?;u$9tG`8`o4=FJ3#Ysd&-6%c1> -0u@p7YX?uMa(8elu1l(jhKM&Fcp)w_IgMVw#X!}@?8!D&4t+&;6<-S++fJsyxkxMgLQQ1aDd)#X`_61 -0m2Vffjlq$gnY%3_xVb*;1jx9 -{dYn7UO(O*3E{kLa=yEE(fuS{OIlUollRKE=-Yu4*0KBsKA3gUD{NmoO*B!{sT?vFBelVbee0u~=C_k -fi`-2>C&B&3-CyjWJ4Jz=7)*BwlTt0b0sL+&s$9A!T0C@_I(DIKer> -lO098{2quslljh}yWnMEPVFt3-Cyjd0VlQK#1YL6Hj+%~X2*@nH+P+p*BXS-YgU-0PuB`%Xu#aAwa|} -2r8Cl)(sgExeSv{+Key_T*xN5$z>)-?wuDu3Y|goJH-|pH;3uf0xo)sr2PWGXr;N)TMA+b6-p;FeHb9 -2M!KA}Uqw9NW)Zj+MNGtLwe-dSQ4mcJaCGLC%yi0pwz6X;A3=;AZscqFv`(W8lj#mJ_;SU4T{&t{>b9Eu*5oL2sVX#jOf=^nfLDP3z5W{Dzfem91QY --O00;of09jk>>>*=*1poj%5&!@q0001RX>c!JX>N37a&BR4FJo+JFK}{iXL4n8b6;X%a&suZ` -?KzzUx;I>P71)LK+}F7*G#&n>0Wg6!xN*%|f81k(UjZR7lF(jpP5`p-AevuAKlu*YH{zlEZm?-*D(!$ -vr8Gx;3qWq9APFNM#5WLK`uShR+J%;cTz;Wthj4b@zH346)}#$hs`sL(a2s|J -P@t8pBdz!(Y2vy!t+I}7?iTv7w?*xrWm)Di`s>@Nzm!Jvm#Z{6!9d}^73JqUAZacdaY8J9OVqukwqDM -|!@wCS@iOjNkO!%!5hNgh&}#Ort!g(zNPGC_gD@sRz^CZJ`r-fjoP*tZB%YxUrC50OnQMVyjY*0uek= -po)lQB*-q3JZ~<^4Tu}R*GvhqKh%}`<(aq+*}_G=&eh^nfrs&3&bhT@|!!cM)aH$3<~)|-b>-7m`B4X -?>8IIK%?En6?y*izz(-(>c+b7?Rlq*Y7H3Kus_Lj;+o)#X%Sht88JH`=sj(@A$wU -t7Dvvqnjoj0*`2_C46>Y~T{T9?Bry{;<|Yi59qP~p(;_D}GvJdf%eJ^Q>E};jpTwIGgscKxD%MymZ0` -)!AIFmWc$H3cOxTpfoC!duwIxFSWAIPb=HE_& -&1E*-je2hP5B$?HrWJR8lm#freFz>$cW#(!ggQ0OWWFMI`r!_(TSvFa=4#4OQYMTn{ -Z{kf&39W9aC?TJ#arzI;?E1BXwpuDu?ppEDvbrx-lN=mvvfQMJ{82+Vh@@M#T0L(4GC)`)#wFIDB2TN -Z%QFUR{F%>5S1w(B?92X%1uhDwDEAOV!?mA9`u4$;2R;h1>%V~C!N{GZaGEAN -k3W690XB}cDL=MUs}R@pPzo&h~n?M~DL2(!-}6)d_+(1eElNO>&LRwIcl%T{@Pt-!OsQ`c!}8lVboq_ -@&-bPMWe4*vUN15o^N;(J2`tS -=I4}&7JLr1l_~^HXe2z!a8LePG0EuwCXk(s6h|k+;e>Y=E9nIP-soBEy1)ajDM$f%*RgSw9;YgpJ`Lc -)?}TMq>clW$(`48Ph?H*IyLtn#HgI|;sKCTl8J;AW>nQfrKB*TR9ao|8`S+#6{m2bKe54j -U9?gD`UUrzy{!{EHd_c(X1$8r1y>^+mn>%Z~th&2XlREh7QmBmWr>~tNicValDTuv_mFRsGYv9iBBcB -%~hXv^D>MXNAADB8qXuppf7^A3B+j_TA8SKq9m%6dQT@&4nYk^IUV}aNQTaT}WZ@6a<=>0#z -`*7DbIiVlPPxLQPO9KQH000080LuVbTbOBx{`3t10OvIT03ZMW0B~t=FJEbHbY*gGVQepBY-ulYWpQ6 -)Z*6U1Ze%WSdBt1ZkK?uxfA?QO=pJG(Y@vNMpf;C1C%5Q*Z1>y*!C@iL5^Xapiz+Ft&rS2ccZQ@sC{b -Sfk|KUtiJTdc!{2;x$PI$vR^%n+xgzDBkq7qgic5B2d3h@i2b!m3Llq-QMwJRDll7h}^FrW(=Y$f)3M -y&I$U&r4#)uMHQexOqEGw9|E45Wp0k80MJ3Npg -dYkaOB6hD`3^MM^PI^O!LiaTBfM1PndPaF@R5`i@W}FnGc_lyOu|VJ1e3{DiUWz`ZBu~hg~(LR3tA{S%}lQ-N_Ypw+0GwnRx!(2o(hmx_q@(eb8-sEO-&dJTIBdX%=oOo~1F%pKIPv)uxaFwe%f!t@39sMUIPim#Y$FGU=0(fXA+E@m*e-m{c885WNKq9QP*-t(g1NybE$e&!1tF(=%|5a0OTx^MYk8JqC8dyaaxd947n -RZ88!mOTPmms4m`TLW3k?BLkP4EDq_6)J}EVG?m6*;cihvCS}vEt3Fo= -3N*orKEDl%D6a{dy&^N*1GXtxA}N;Rvv|Xf+1*iHzVhTT>u}GD3~-RWB_YnTS}O3JA1A2wP|}n3QQYp -sR3-qMctbouK>yZ4GRWlAS+y*MsL#zfgd;a5d&3>%rQ%OsV&gooq>-T>C$$d;(I-;5kh=`;ehSWC$q6 -3BT?hDPVo*6r{iH18}xcty%x3Q8FSe?<2=i0hu@m9e(fmV&#XjN;j)D~Y}AVl&kxx3>|ZP3K!~zK5vN -?bE47AJFimY@r}?l69p(euanP*oMmZoRegexM&$9HR^Vz1_3qiSRA4B1lJs{Rr8Bg7kOrF`<=Wn)S$th -C2pV4oZXA3Iu2`#Cm%j>r6{WtY+RaiXjc7Gs_XsNLWJf9n!;xfUtmA&eE_Mync`jgyazGLS|wLYHS+B -_jmX4-G}(Qf8D)%6W{&t;p4+9zWw;&)B54{-G}w3U`C>d1b+o7lYgvM_piTuw~E)R_xEq^9^&=S_p48 -^2A_wE6dHl64?lWVFkIcgUA>Qgdb?V^i*H}wzFql<9#;P}!NAE>hZDuNtb3Q7w5Ge%p% -f^vKM(B_8t(#*{a&CF&Eu}NtBd5bIZ7160W8V3<(y)&Lb7DvI -XlcbX5xM#V})WPpa;|9>Nz&8q@N0R$?fseS-7Oxp}R93%nL+3YF-HK>tKGk&XZAZ0ebl*-_doarm!Di2sJ*lO1LLk5|o$FmDAk`pp$6*84*k2t|tyu;c+tK8-s6_?n -bvk+hfTC-RZ6s8&je}m3v%-Q_`*nvEpYe_D{8mGyr=J6Ta#E_;CaZGV2uw0HZ%rR_9Fq_OUYEY?27i4 -A*K|0!^wPCi8(;)FA@D6C2=Ud6Ju0VYpCcTDDkBK}DXs{M>K-8(wE7M&7NZ!GVMZMM#Tl=BI$LI?j22 -hQ;fN&kwx&U(x2`B&};mejb>srB3O5|YHMSDj7bBd3+(}%AGTdK+hSc-B?r*AKg0Qh1n!ts4SQ -kg>>m=5~&DnQPU8mknB3Xq0%of=VK`&x9}{dT>+CxKs%r1_2!G&-gri-iy(Hdw)Gi#-t1bAZJ`Zy>`l -IXXZ>v}2MVbo}up&*dK6>ux_LW3d2ZSy;ZPQLIkTd07JlaGahrP)Ec-UY}NaHy;iuI^iFH)LYmA+Sam -S<7;oQ*zHD-dKn?x{)usFXUg)OrD-?$nLMMayxSoy`V8$m93iFE#!s^0=qDL#kw=3F~3lcbBN^;G^!#-1 -AwpHw9sh%wGFhrnY?2w5fPQ+r{rTy5vT!kx#oe&56N`_pF_S+w8J9*+k^L%}1P6)IPc-O->OtPxEuWFzUVc -C!?wZ#@Z5h$eqb_o9BJoJeLESuZ-kXLB?df$E#8_&|Rp3LPIjF89?)|C2j7@h++zUZsSe^=|-6NK3$w$z!JaVikL*k%!j -)^!F8D@QWCfLdf9-3{T+1hUy=93aW55`p&I<@YuO`_{E`dG!gWvn0Nq~~*VRKG?*_%QG*4dAW!h>C$5 -sefmjIsQ6cBkDX43H!K>`VB*ti#@;Atnne;!`c?n!@f}gjjxLO4WleQzZLnLJ*ssbS=4*?D5U4oqJG1 -WrRUP3-g__SXo39#PEB0&VDu$?_83R7`2sG_VDch7UWvtvarnD1cqRUR2llSQ-IbWT7;l{#Jsv7|j}% -)h=-OyoMR6&5X_dO{{JnHwz-oKzw5)I^ka(*9`QrC -5%R2v5WwTN{L^gcXB}k0EEvK0FAX`Ra(=*>rX|hue5$YEQ9|ZFp4d-`v-q8u~8$wI7Y#oY56=Hb&vkP -NFtPKhfDu5dLPviQ^Abt5U=?#eMt*hx>j-;IF0ij}U*CCT#tmaavUY=@l^dvSnwMW0P#bibo%LI&CN3 -DLfz)EC#g2_BViMRE1_URE}30U}@mqkSlUGCSR*2S3Q2PnKPtVu@cX!R5UCE^(cb5-YU$-;_Ow5?VKk -L&TfV#?3m&3tY=h`2jk}#t=(AoaXkv=uL1HUt2D&#*I>6ERDG-GDj?TP7lQecSzW8GKNqX7$cK;X)sp -xtJdlB%@Z1hQji_EMZ1=);!*0*1L63zbCp-=;;QDyf#G+nE8HA61sBl$wtp!~gAzq3O|1X2WUms4F#I -Ud)hrapKhnqjl=y_~CLPn|aM8iEl>;7hcsDF4x?zV*Hr!I#+lR)6MH!e&?*EBX2h^`+85%_@+D6~n~( -{<{}2ZSi&f!$eJw2ZJv81VzvdNkPN+6rzi(){gX}{S-j -ynm8ULQ+xmx;;fC(a!CkFo*9e`3iQL!zX|ALDsQ2y?JC+qWZ`}d^4^T@31QY-O00;of09jkI0%$xQ2> -<}X8~^|v0001RX>c!JX>N37a&BR4FJo+JFLQ8dZf<3Ab1ras%~@@a+qe<_u3teYIK<9*wMnl%1dDFkL -$YZWO?QKJf_z900xi)tv9zd?(mK9g|9hVysTa%EZgM>o-4AO^91e%``pkG~^oFJBvaV_)(v-q%Yt6bTdEm2ErZ_pM3uHGV3E7u^e6O_Yy??e!NGQ7gHyF&`C$|_AK?HNv*_Gu%mfSD -;!mG>cPk8$?`_)MarO`)6MCwViH8=h`Sl_coNBD?cY)T2|;!5pi>^ZB;oL9o0S^@j#u&}5+EkdhHN(n -)@DfNs_R5DC1V<)f&e??fgFVti9jg@A2?xNX?My2UqTUkKu@l?HWek;?w9#u!UG^WVOYPEqHcoOP*0? -sWhwBw1|vR;KhuQBvcMMZbU0Ak?K2PoTP4(Rym{GH$&yg=}oc6nytq-))Bh$>f7?y^vh=whReA`^&&5Hc-D8tpqIhp=HZ*yg=`7op6&rB0Lr1^3*wi3 -$Ddd1x+F#j!>dCx=IM_}y6~kiR&8_;P*8C@4n!`QXHXoKf`Gt~E}=*9|WaIVK|Rx)g=0eU$q83C$hWH -ol*RS#jnuzx)5Qm*-Kx17U$DTWQpVTpG01s0jcl9cD&X12knQO9N5~0#gRBLymZa_jp{b9SCb$L|+tU -Y*AB-PqiZy2s}`RSEx6NU0%Juo&_#45eZfwqdFCkAeXjOda&Bk79Mk?3IyYJmDe&yJCSC!v1nhrD-M$BP72`eaF&h;l`&4Mb|B`|K{<2edxUMo^=DoX0W6wEJ?zM%?x1={s58$ZQ?- -_}EEBLHy -Whmc5@YeOKYl`8T>rkL+Gd`0FmGXD;jEnqA|<;q;Mrg-=;s%tr2Z -JNQ&tEV60b4RffQlp_q#~Z8yI#%1am@s4ao}qtQK7bci?iv&vE9AM -XOAq1-g{ObFL*7D`O|elY8K>Z}=u%#Iv3hWYMCdx)QTrAr^`Zsbsqp{cWRpsm4hA@L^hfj18>JTe;?0 -=%z*4$iF_{LlsSp-1fx3kT?;ha!Rz8|ZH@3Nm#58nvQd*m~9C5qqVJ$FTbvAgsxqU~_Wv&sqvsfLBOt -p{%x{GBB^VOf#!LZ@&Q!^ybmQo4XP?Na|h%p2$@@8o*Bi`3y%b0q53f30ruecYZLZY)(E8pS_#z71P6 -uJX|1qvfo!EomSc+*JFL8SRP|JESN|5?G?^rJlf*<=JM4=`trlg{QAa?)xJDieEQ2D{-ok){^t7Qk(G -7)u1uunGv~rKAQkAXV0V(!FH -qX>4NurygB$s^TZ8FaSQvg!IXmLdZZhMyZ8-`Q@OGwZLbHC9=_e)5Gf68(qpYtv0XRU277+(lVMOYti -f)uWqiB#Ge$c!JX>N37a&BR4FJo_QZDDR?b1z?CX>MtBUtcb8d7W2Fk -K;BBzWY}Y?xD#dllB~73T&s_c7Q$>=(ar!79Pc>V_{^;mE_F2kN;j$vJ*S;Z0pU%5ucw(ilWjqP2P(9 -NgNPzqrH$?>H`R+ppgxc>RRZ+B`L`zr8PBVS=sv5VwQod8)H2Pt&JC68tsy#r^Szw`>2!~r7#a7gO(O -G7QSs1MiJjN?78Q@)?M$9U~Ds~Jz6c)Hdk`H7Y?%?*&u%xmr2t9x!0wE+mVx1I^v`oNNZEH3bRJ|Y6H -0v&SfmU0oRh8?b3CUd7bASIuW5T=s>`})p-S`gbKy610_q(;7enPq8c!*Z(6@212^zSi@jp4qm -e}jZc)f!Pr)iT&Z}rZU{}x5N1mf?LwXVCKJh8pB0Q!1unDqd%ay@aYVkyy^5_Ck6i%Xrm1{7UrOc^-I(V)K1bLQa -M>1sTaWI>4NFPD!U;rFaj^GPg2wRwEz1}7h$z4dngY$L+qcpy!4F8n?{ZgrES4yS3F@F_!fj_^y?Vz? -fOWsFPZ)S!=+Yjf(ZF}ZAeo|%fr=&OkZwUc^#b#6NKDxe=f$(!U=+NW9b5yvtE&kC;i!~kQa8BvSUIHXVIx -+qe>NcoNtCLCdb<3P^ew&V(f(Fz_X^m!NS5++CVmLu(T*b7W0m9^&P+cN$m3dx@An1%-{X***1WYCbF -&ET=LF~dlja42NS%TAI!t^tZw -*s$>!@jPxetW1(W@~5_wC{)5tYVM02BUCgC%4%GX`wI$blxot3=Xf7AKN*#)+i$OL)R~Jc`aZdH*CyQ -eX+WMVhaMXT^%20y_u;=c`vw&%z$ujvpm!xcO=Tx{K)PWyIgRV1Q11Y~wmdE*rjc?}nW4BjY@FA0*G{ ->}WeN^j{AX8hR(g@dI<;k5ozE^!(b!e_7 -w_XO9KQH000080LuVbTYDi$q9_Rf045*+03QGV0B~t=FJEbHbY*gGVQepBZ*6U1Ze(*WV_{=xWiD`e% -^7QN<2Lepeg&b$B2r^3+3ghv81)_wd%0eLMH(dOeQ^y%mS|h8EUF@9Z!XAxznLK=k&>Lp^+PW}Y>S*3 -&ig@SyHZq&33-y~Uy6TKhbk^l)|J>XQK~YoBxh##-*H}ZsrS@FnH8I6=X0Fr@hayhdb7;RFv<$Wt0K< -BB+tTCEP0fO-7YTD7D+!ILy;<0wFIP;%Le80=H{oH3A_FI^XJRk+c8Z3HQ}YoL?L^}TvnB+ChR&Z>DL -%;t+KpDTW2|!4^r{nQP40Kn~kM~4Od@ag&!v#?lPIQCx~hnt4L=50lnpf$mSgpKZ=TY$D=4XIXOx4SV -~ZHvgJk|-+TwN69#{t=h63%g)DiZ80s0e;|)hXP9-~RInP;w6v7h%y{%WqB66lzrYy~>8N1zz15;bBp ->Yi;is~UWKv}UC1`2sM1z@B6hXmjw2Vu)Muo!IjkyP?_xEh{c1$_FRM?Xf6E*lM1 -BWSasQiz@zm4H}E~yx4<1@2Z;9b}bQa-dYO(-j|iwpn`iVsMg*Rs*x>PMwe_cooe0B7uy=FH~|Al*xo -LeJyfkl*5dZINms5YDPOZVO(UU0$(KA|2WY4kXxRYbMKbt7&7P<;K60xgeHr3qS(wBQDMG0GayGW^ -g43EiK`aMoK=_=@TrVmqD>|0(!ZFvnhQFmZy_3UUPZgO?3A8niCRbn*{jL3i-pd`6LJgOTP`PLoaM4J -PsLST73|Tq26&q{(+ohai6{^vSadbxp$iV-aJ2`B95jGrFawG4)&td!*Z}LNXTZrZ8)$maz#Z6Ar@5* -wNMjRboOZ?<9mUlKo?18$ -I5f`G}aNSb_`NwBy?GpC^!IE0NwsSI{Zy<`925ZJp6(xVQlKEJ2|a$ERbTr6n~@~tK0GXgj}WSA=s`5 -HvfhcW-X444wE*$~tl&=xloIGzNvXw#gUCV~Wc^j7{hjw7gD?Vu?$ -5_`iriK$4wj@;b;Htx7cA>GBzHtHVU5gL1WyiMeLE2V>#gg%7^vgpd#A^gMt4#;gy1x_3ZwUl9ql3Q? -kjqJ&K@({AQNDL41ojsX2wv}ap`>SpiIESpg+GXuV2o=5{j=FF>Rq%|s*I@l1Mh-%g_sIqrAV#4Myeo -iklUVSfJ*I`?7YVN#tW>IiL%y2E|SL7bpO#x4TyP}8S}mgecWs(aa47y2TG -35YGFoH#SZwwmzYV`V3uf$bMEa-yl`?1(1{9z#k@6AUzL7er(Pa}?#)(|jyA?11lf%F6?yj|U0N0J#a -nvusU7kO%Rh{L^*t(SrXp9>24Wx`111EY(kbHy~&@ovAM&G(nLmJ>6Smp}xdpu*8-(7rorg|FB9lF?v -?dBtE!%BX{@be?JZsc{oGFJFC0Xo3|5|NY(l7I88(;Q_GtYyS;VT -E{Vw|t)lIBZ2pF(W?86@|prduzjGb1q*6e_?_%T>0kR9Eo9zRo14Q62T(zG7d9LeTDT91Y9$wb=ZdH9 -M5#;1dWXBh6InZd9@Y;886tZinp*b6qIQGp=Gx=8X8?@ElO6P;*}3=(!*DPgol6)!sT0JH6t4K~dqEa -+RMITJEOvFV-X#ua9CP0ynM+Z>lM940RD-6_#RwIHsyj+-Qa?tpAY5X-KlMF3 -Pj&B5EuOLdOUX=H-g%g4ka)VZ}FqA0^B*)0(GO>;)n41y(=axDgy6;3Fi-yU5gVNJDfrv;v3yWl+KpE -ke3p5v@TjNou+2Mq7G=nz84|=s32y_7X`Gka7&?7kK@#7Sf|?sbn*#_VZqX@Fpd7+I#Z+9wU|Y#VK*z -xR2=c*>#|2$JK9pO$e54CszQ?gIN$ScXwBl=gops*{cpr{({sYz(U@KPkTQ^&4Q}*N9DG~>M3Z$tZQU -4QXz)EbGV*V5uMetezUzv3Dugfxu0Ow()*6&9Qk0IE&2J;^{N%m_E+qyMl_uruI?;T>wlwTqNI#M|NW --5Fw@2TLjI8}`dLlKds(A)J9YP(?<~U<3OA(m)^IROj4OWn83vfp~3oH1z -6^+BssF9Gf^Yg>u5Na@6aWAK2ms3fSzG*jYEi%f006`Y0018V003}la4%nJZggdGZeeUMV{dJ3VQyq|FJowBV{0 -yOd9_wsYuhjse)q391PV4}X3v4`CG@Rq&~AMRK`8dIqg9qXT~hMzcjT+mIPFU(0+DsD-*-Ma5lyR&1F -7p;s5+^PZa}xL<;Eh2-uTR=N1ODvSZW^x2F=X36-U_jfJ8vndL_3DuKR*6^T -*K4v?e_lZX(_Tn$4l<`7*V5RC05)^$55Sw(W*h^mO^!lhC88^&5I)zFJp={0b1P=YkBd)Yh)JPIy&~s -`ZhRYIUF104r6x~DpE0mC1r5SmU{3`NGh+JmO42IDcqSkUfUSXRO)tuM%ZEuc{PGM(cZP*B}o_}8u&I -&GpCFX-C4?*;tevEB}r09ZY^vBxM57`@SSZsfUDK&8I7pUFhPO=t%V~$A&ej-LWwP^5>+A4<`kK9DMY -QbL1yr;7xaCi@lapCjK|LGE#;zv;1AupYmKg9N74&J+E0mSzF@pcEy`*QgLyMZat++n^dyCKJL}9ogD ->ByX}1}&0>ndwplafUQLTfe@*hqz&89>|dx%{Vc2!AHIC=;v-byEp_H%S3qA(xoH=S&YQoA&VbwllGm -Xij|+%qdFyAcQ9QI%%v#Y}UdHx6^pSu$-aZGcGogN5saL>!NVY%ZctNDOO57$X{Xap*oDd@kZbfk{c{ -DE(#Z<_b%E1Fks>YjtOVK7$m)ehsfjQ5->7P}@h%EHqN+qU-#E8AJ81h?Nh1l(R91 -9`@qqglAlKZkxwHCs_mvg5{0w&{VSz&`%e>~!Z%0rPR^K{g|UshZfgAq0(;RVzU1qDt_Ge4P0nu%ny@ -?4X)C#-O3ea8L-E}3*BfX?k3}>y5>buU4hU2HqU#{WFHbtS3DltTM)g1FlL4j0clcQS&~PyN1yw8)OX -PY*7c*NbF#c8e2T)4`1QY-O00;of09jiI(q8hL1polD5dZ)r0001RX>c!JX>N37a&BR4FJo_QZDDR?b -1!3WZE$R5bZKvHE^v9JSZ#0HHW2>qUvW^3h3YJ&>9@I1vDR~e0BvK~?ux=OG)1BvF0!POl=I^Gzwby% -)~nmj(}zSN?;hU!9nCgL%q55?(PP=Gq>u7}dr5D_;|;`#Ogt_qqs93b1QwXpVcQg$5fWjpql`M_`Pv&wbe9jWNKT1VWbOqBZf!*0?dNrBP@PJ(p@G(&DJ7cJigZyz3&n -KyYh4KV!7WNaC1}RZakWNr0AV~`Thf@$<@St`+0s{* -lz4xLjMw8y}1bJ_99Bd5KJiL-|r8vxqcL^7y`QT(oll+zNtIxlz%cVG|?b?d2X5POtBygSQF)q#Y^iq -dwEt`RKXInZa>B11wqSl4>1n}n(Vm2s-_}j^m#fRf0|pDrg@a3(@xnzm&>}jx{SqEuvkDJ^>*pj3(-K -4O16cC%#H(L>;+yb<5e2@uw1?dc}Js)AZMU<6&2{bSS6mk8-VZOd5T9{I$)goGJLA=X|sXR|GhLc4F# -9~+;TYFdyG0051zdB9pc -rT2W5wlNP9wP@A&W3VEkEU)M$*Dy^r;ZIBIA1+vFB`4TAoxiBIcbTa0E?bv;dAvnBs&w;KOC^2h`S)pJn#~q5rIELx)|I52W9LE#P*)K;q*#@i^i@y~pdWfde!+s)l0f -gg3uFHto8AKJ0)1Q#hco4tg86*$&+Fy6smhr>=WH`m?Uk6#AS5M}tx3ceChv%mIRlKg_WJyBv>jFod< -pU$(!97=&eQUm_6HiR(G6B3gAUvsg}z-5e#zsiAr&;xt*lrsD~hdWJ;lz}QOv2!>9-V{s=KAp7GrO0CvjxpY_YNZA>l4~ -n9u+0Ci+jEC%&!?)^20K}^=l`pi9%7oBtxnaf1MOth8dxXjR_TOI4UH*|AKFsgG1Mv77zd|g03mj3FI`8q0TBbdgkqM42Qm8S$C3w8j%WH;J1 -v23)lZIXpSZvx?PWzrdiAt-{q(uo*tr`7hz^~c5womrP@EE)-(|}~)=CV511Y$*SP05YJQ+Yg*c+E&^aPp8+@A$*Nf*rY;2qI%SBe-jyncm|+94EyhxMWp7q?`qtT@n29&0|X -QR000O8%K%whX6zHCjbBdaA|NaUukZ1WpZv|Y%gPPZEaz0WOFZLZ*FF3XLWL6bZKvHE^ -v9h8f$Ocy7jw%1*f8j49L3eJ{s2-=+d+c*4q|MbN9hB1lgi&t+J$%QW7J}f8RNWq(thGG`;8ow6;W^_ -nUMo$|H-Sw5c1xqljh4suVSgi=wRKIx7n~o9TCXx!-5S-aMAle3Q+#66KCdIZF{;S=Ci6Bsa+Zi1UW) -wfa;64zu#6I0fuH&hvPi^O@RFWmOnuMa@MK=ix5T!fh;hv@4ItxJV4-xq6%ek-&T-0Vv_pfLwg}^65* -!uCC9|FRrdeAmum4FY=f`!F@pSV$1BpZq4qheE`% -t>*zTszav5qf7y!V|i#me0(;&x8?#zDV6|A)n*saKF9+gd9HT7(UXt`i3W9ASxexP@L6czCioMS$l&3 -1V#rK=5X3SWQDmn635dGz^vT*P}0Sa)&G(182A+`?B!U*U}QUNu{7X1P>v7A((Xy<{KYe^wtji+6Q)! -(C|Rmx3wpf*q)OCM&+n(u^l_jfN)~pt*qmX_E_2AXx$I(LT_=0s<8**?X=)ZqSPvNMG)X~NiumTn>=A -#&XN*>n?o!UFD3%`>R0T>kQlh6+ou|HWgs#IsR?QZTYZa17y6pahx5kd^`4HuZp9Qr;GM87E<#9XZ4Xd; -LoJagP9tVFe)IMZ(P9vdB+Fhp-w{UQ1HD`DMzAvf!*t*<#U1E&}+}VS&nBEF?rGi-Af7zQB_JsNH0uZ -sME}@B*I7;s}91>dBG;R!Eg8Lo)EFqXv_=;svWqf<=B6 -neB;ifM>-`UWaYe)0b>tlu%Fv|4);_=MAt7x#g?SzTZep3ad6t;NMG3OKklsB3~#6`)xz<|HPK{Vxhv -8{i;3}Ra_rn0iR}107HzA1=uuz900H`rl-Lu)}Kd-Lflm=D4WGe5~)4Xmpo7BZ1EfW2$o*9`0#bOJJL -QGb^w06gRmOCd4cmDYG%FiMwif$Ac=VrmiPdS9ASm5xGM*qS1Xq`(Z5hu1J44l!VhlpGb$ov*Ag<Cck%Y}>!&Zw8zQm;;wBauwi6QlAG#RGMCDDE@Fa8r=N*sp+4D80^V*o^)DraokEt)JAsQxh_ -%6=nm=e8fR1~J!ldSo1aSMWtr2x!C-k%c?1ASCHT===dQI7dJQekEtl?oSru2j252z1aVviB20QfDoo -9>V{%mNMFlvZ?%6fvbmkmopBKbaF7;F|N9tkz~7iEo;%W;On*O)dNbqxVG5@lv_5NF1rX0!$QICUELy -tw^SV9_egafu0|-Um|Z256=+o>m1%k#AupSK*s^w=M!Q3HJEY#BX!iC|<~2L%VGJ83TZeyTn}HY)omN -FtkTwI|56LAn`isRLo?0!EzM(7clxSH5>L5|Ijbxeq1$)b(oj*eVroFLU(iskkVS>cZ^{WlNgLc-!w{ -(P6y?)4X&=^U}A34V^PZMMQOF75nTyX)t%`xoxKY;|y3;$hapbnTZc;Z{6=W}=e+jUTax-3=S14{}AEJ5(PF{VW$iN5_bnDa&G^eO5kVbIY -yQ3TorGRwcc@UT*$BrqOTBX-1oq`d&DZpE?&@vTa%Bl%ViLgZ;PHOClE2Z-Qyp4sS3pI?p)#z`H{fAF~d-I_KvD&O@psUNT7ad4j3?r`Dx-DHo->?VA-l=j;^j9{yIdQNvSG66=^IYqeJ$NvJ|r9t?wURkMf7`3{LkEB -@;<_Jdb*>(~GJx>7PpM41FdoShcB@7)FDoZb>4g7JX0NS=#4^@LF0c#|%AqBneLPH6)8ml^phL(LmZm -F~{mp7EhZCd}7RDsLlh;XtMj;(3067sK)mQ-mA=L*pIn2c|@n;3jc$gbc1S!PoQE8~V6kx4N$X0q=QX -E+%AxLBU&(dn%|eFSC}%8FH-2|TkCC8U06L{l!hSy6Sk_PN=VV$?*_k99Rf3YbMA7(_t`VFc;a_JUp3 -Dz%k2?IdoUDHAwWI;cx0b%+?4!$k{8U!qEIiGrxM>)0YlN+%Q10ZsuJ6~U*nk+2KZFv;`-aY>O#6y}2 -zmZBFSMZz_bt?-n)yK(VH5*nexI-?3MIBu^36OCcn9pg372eeB@S&<+E8~ZpK(5tm)oW-6iSSmE`(3w -Fo+{VBpgo*?Ww^a9I3L3`hT8xJPUa#~W#l6cZQ#}4wSZ%*YgDH)4Mdy(OE%B&9tr^+orZDwf=4mTqmx -rzbt%eeC+z(0>Jc>*F)^h69#c2}_yq8fa89_QQP%pYCPKx^$kiA&Nk!9>~)i%Ufhg^t0{dg77Hy6V*1dZ5qwN|&LJa5X`DI)8V -;H{jFNkL^d+2a8JdJV95PPbh+$Ryp_Q&ou)d@D8+h-F*q#Lv7l%ALc2c)`9NGIO(7oYi_3Z|9p^4$|P -A8urY@WuAS*Q`5K(dx$PmdTplp0ASA&;*HatDIlquJZE_G#3fE@cR2jlKtFs;@ -raRcazu!31@j6-dSVS8-3`Kb-aE%jq+P?$;CH?!b57qFtQletLGV$UQJmqx8-K&>+wA6<(taP8r?-Lm -$ifUD*`L)Nz9zPI-vjkLRs!4P4B+4R_Smr)|lAN69?-te&R$=;=N+H$hh|9;|%r1vq|Z_*)Kk&qFoK; -nRa>#iQK-sbK8IUSRH-sTe_9w8R5hipV#`$;pU+=Xu5HUg{Ww8%n(IwND85rMjgO+5P~deskQyZY)z` --(Z?_vZEeRk()XD%L?c^syNC$D0;;Gy!Sw*DfnH*cc_UGGANQ>y)PcZk(;sf%x|Uk?#*vMApe=B?p`m -0F8eO~h{1sAxqBuuJ#?ER+}33VL%y}qvgb%6K2gJ7uz#YFcIY=L#Qq7`1$5mhZufHon&rW+h`iUW_I{ -9J7KZUUGA98xe-Mxt=EptbM}B5?bl%di3Mz(-pTH3B)&f0Xg1nw~9jFcQ@b$7d4P&{?_>tR)aE?QUoaWw7n+CXi3$kg@mbCGGNNB#FZi -YK?Px>48v15ir?1QY-O00;of09jikZx9Ik2mk<;8UO$v0001RX>c!JX>N37a&BR4FJo_QZDDR?b1!6N -Vs&ROaCx0s-EZ5v6@T|%!Rex?1h_(r`!u62&?PB$fv&f^Nw$Y9AZUrUxk{o&Qb~<0|NGA2heXPD8!Qj -8OrG=O`96p8R>_u7x*g1*IHg3iy;O#tl`tQ-A;77@Kx&Yqz`(rMO4|Fo+-@(=08 -|7aKC2`J=bC5Z?Ya5egjM6CLo91kKJJup%;BgA9z`(yT^+bC+}Cp~6KG&)**8uNQ#)>Ge=vIx>F_;bjm=(aM?Ug;HirL!eLPS2F!Lb_0pvEgVqYw-L>ctBpJLlEXq?jF!LD;cv -*poxt-+ug?t?#8At~~8}eB9yb{|321z42uFHt}W25DgDBj8k$nov~#H`|^S9~kJ6RC(Ew(tZglZ&ouh -8j^YX+d9IOh^_O%z8JjgjGm9B|zy?_l+=lMnHpi#ogO`vITgg08^7~38tzo3fM<%Nw!{RxPeO}=KW&f -#J*9i)4<;XiIF60ZKcf6ut94yu#-=5sNrrQcM>^C-~<#Uk;6(5+=sd0kSXR;ct#KB(X)+v>(r@WlmAA -6`itMNu}&)1kqsw9-*Z(lZI7KiYNgQ<=)fQvqOx>KMCXB-qi;|E9w9diFks&>jFvYTBh*Jf1jNUii@$ -=5V1?dD*+BKvgjHr>O>-c!gMahAg`4k~Ng9 -f`sZde1+uyhU9J@E(Ltlt(RocAqwiW#Hw2yhvT<{0?*iuZ>6W(BxXN~$<<{s{@tfHBEyqXuoXk&3@25 -M5g)8vOJXvmSWF1?wA*NilTn0SMU!{0!mu)ixvYYx!q_{f#r#2q;z~=L!2x$7YrXOO%ghibamRC;6`s -v2UY#6xjgdZL$j!nnz*mOHhD|!ufP0!)Fr0jtD&b!klmh54pG=9pp$bK1W$7I2#Wt7HQyF8^E*`P-0) -*sDt$k6zTC0NTZr>hOTaS9U#zblg*jd9A^f*SHf^78avu)drV?KXxd)csD0;mQp22N_GK+pr11z<+)t -5N3vB=b*s+6B{Hnxeu#N1h6rXus=TS2XYpYo|n)TT%%U}Q?X3 -u@u7^OlF^$JQbq?S?4nj7^nIAVfbEjuGPOEqu;_7k@|2dd=IIA?FdSj_7(oNhm{E=G1+!jaI|fQ&`~a -6-9K7~bl*hg(@`%e7`ZTePw%LW0erxP4*nB#uN)QmoXia}n}n*Q!nO+s+mfh;uA}W;B|QZ4bO8gNFK( -oj{1XikkrhAJe5s8H(iFks0Q;o^08Gw#mgHPniqyr}d0r4p#ghwz>HI%csx3UEHh^B?lM?p@dG8&mQC -I*fiS=%Q#bY`CAm~nY&bEBFe+XO828x0RvpJe!X -47=;mY-hr3xYe3Y0K73n}>^>QH@f#)QUYQ7!NMTt`560*vFW>qh3Out0{)VIq46Vu;DyZ<41uJw~SM! -+K|FFYJ9$*!lqK0!stu<74ss!Yk_)j?YYagH>VJ{#SSRNhoODPowxT6l-(4 -(jYO>T*9#ooZ|6JEibwjlFt>7+0g=|0x-%(2I6JHl*Ch_Eo7xJu>JrMoESfwG84FtkAVZz;Om5u?AW2 -BYN?#?qaCwUztNVw|n)XQ)y*!S3uFg#`fLN{nRoVHjtY5NofXVC~3zpUoe_hS+t`HQ+AxpFt?w1|CAT -+I@^fR@c5W^7)Jg42c=6pFF6gOg+GX4NF$r+=_8fGl3Hu9SZZ(gV*mo+J2WUkyq>UU?cpqCUMzp&AQTb`hQT`FWe#6w45!|sE>2>;#vck^`v&6P -$O81NT^*<98593;_4nJW>ziwO@sa-J_RHnxD*%K0v%gQH-Y1F9z+K$_@%e{PGlDZPCw> -j=d71k?wCtpI=Oyi;`zis+mA=*XwSEn@H@r-08mQ<1QY-O00;of09jjP4BpJJ1polk6951r0001RX>c -!JX>N37a&BR4FJo_QZDDR?b1!6XcW!KNVPr0Fd8JrwZ`(Ey{_bBvcqk?Vwzhp31_V_wAl-lg#ZW9=zX -SnGOJ|D>C8{LlRQvTiQj|yX)ddZ)y2U44}5O9f(mML*YB*! -GdCs81+cnp&WX~7nO38;0!Bsd|Gm5WmR@TonB$HHzg=paucff@ot-RBO109&zD`lmw{9(=v}P4hx>_ -t2xnNq8H}bg_k`*4#WQy?WWF+D4lxd4|n&ZQSuX7LRuxp^|vx-V%#absT@uSq(b?XG(fa$RStxyyyOQ<=I9Vh!i1JYw~Yl(l -foNgRdaAJ0|n7CE)xc_goALAxan-UAzl@n)n?)AWp_jf#Zz23{lR^&!`xr*2=O6_NLBpub4&oo2=UG!l< -+6|f_$D5$4TG>F}bY;MWlvW6QklMuUXi;tz#>5(Bh7Nj)C56oOlE}g7a7qRwf{~ZECFI>u7OIkp44cn -I2Om&ws%35?+vNj5Kpsr-rD!<~qD?LF=rc1aOR2#esw5D68u5pNqBd(=z?dDPmt;WUy&9A@wb}46GLj -L5Bzb2UY9$Obs51#g&+)DYnn32N1mYZU^K3Umdmp1dO7)Vye27AC)3Hp9`LS|{QrROm)H17=sV-6A%TK$4rY`UkpKvTb7A -Lti=xjwUw99qGnGHLG{9k3#U%#0UPo4zycc!Zv5@@6n*={XtY)J80I#n{(r5*EM$kJw6Gp%>$O{N7p6 -L7sn=X63nQJVM0KGK)zQkjC4N`D7ax;VcR(9`bUFv22yrzm_=pL2(rWJVJCfgR}+Ix6C_$H8o!vj{OBQP#r0#S4zh>dl(G2cnK%0`I -R-ySCs;92x^K|uAEW!}KpAwgPYbGa6i+qnYYifX)$W`E+H31Wf+sIA1!E^|nYk7vAt^UYy`Yrhy{IUt --3Kfc+*7xDeh404)v6t%UWW1rZ`k=J)=zjwY(T6Y3YiJIq5K++o#2k{tb8oNO@0+;$9iH-9$>5H_#!$ -LuyOTg~pvSTKiQ~!esUOypqm>aZ0yOWcHqq|9ZaLqI>0^?OS=&pSm8Xe`X;-9hIo|-i;{U&zT&T^$1C -pnE*-GXjA!u81wZ+J0nCkV%2|J$&Hw(eNs3GIl9W`{XIEEKB=S8#P_O#^k242Okd*e=8+k`ex?lPvGs -Jjd+awRV0uvXEEykA@{#rk^CQljFuJ3K7$)di>|=e^5&U1QY-O00;of09jl9R$of+0{{RU3;+Ni0001 -RX>c!JX>N37a&BR4FJo_QZDDR?b1!CcWo3G0E^v9RR!xuFHW0o0R}caU$fzx(?QK!@vQ2li-aYHxMNv~h-81SXQ;H!eNkDhZ7+MJ1F2tG -Fy`DF&^&YsgS&g;x}d9Y?5Y-E~|wgVk;Nyk>%qJC9`401hU1A8&78-`!m@(FQD!ITU*f@Yi^S^mWO^J -0Y>4tPFR9Ho8WNU1c6yy7j7oJ*)7by=*G<##lhCc3#luj@OT%8qC*h`wQD?EpJwg{wz3eRY0^jf<;P* -0`4{YndLui*bhIh2e%CzW0>Rs!QG>?56MY5XC?cRQNr2RRXh{VNF`a19{vt4$g?+&^22PyvPMu8vGWe -X*XIL(90K&2_gAkaM`Ujzczvbv{$$+z(xUwa0pinq*1cp=?5^CFkEymCZlvZgARUf4~f;tjzAFPKc_DB)vU`Um*?~K(gyYK1xFg -gk);Z0NZf*e8aV`1KD1t(a;aY1wsdRUwc@=+tEx3$@3ak8Btu@Dl$U71ahS}5>3#3(av4AuF}de6Od< -^99_yaUIN+?Ls1NLbh$K$>!%bIkH1y?l3o1a+uJ_-LsH%R-V%7N0VGWaoauOOJ{~FvVI>@olP>>ByCW -eK-hY2|B8`Vmr9x#v_%zqss|H@to)Yi@DnRbg%U{7fGHMN$Q#}k5hRrNrt+h3nX;Jx+}GL!zIG=O7M1 -{UcpgV=eZ1ob_eQD2ac?jddDS&t?BV^SfiQJuE+Eeeo0?GZO`B4|B2(4>a-+$0_}p}#h7L9m3$g5J!v -(=4GFyvc1z6zitL;2b(&*EXh;>!g+spn4}ef67x755s5%$_D=OFM@@wJ%J_-{aE4|-o_l<8kB6V|t9_ --84tVlgM57VIdLkpEp2Giu~5y2Oc1pch0Z(s|JaC~>xk`h3@Qgub?c`#+KlNUC488BgK$4NctWz^^s4 -eXd3$Mwk;VSZWHTUV{s-sG<~bN}np=&@eKk+bQWBHr|?*M`PpOJ5;qToB?=q7`N`lS1vwutDUL?d5bQ -S{!1VFw^ZQaPVdZU2s!)({Kkf4y>k*1}FXX@w$tp=ETng%lZKIV>Ldu0qtHqu_2lg1_}bJ(+%bXn~bM -&a6TPbbDG3vUTJD1qjEf&YlSv4R|+R{oscz-^6)j63)GU}XhxOcpz3I`-JHQ!hJW+p<(nuM(2jx$p%- -30`}c~zWYAkA=*wGM;dM$DGresmQvTXQ1!I5zim_*-4flojvEVPG~9F9gN+7bCpyZ6 -a$ppfw2*=xxCK_|w&kjvLavA9dm+|w*INuK!A>EY?e=b)k=?iJ>&J(VX^rSMk{PFAsFB#E1F(C*^Imx;ehGhb-jxsPY@WAPfXDo+}E&|Fs675jHOW&6_Q&^ZZN -O0>|}hW>xkeJ$8mV$F2VxgK?tHN{F1uhl61Uy;x$@K(THcDQSjemoj?k%$N@-ip(HQDHDt94;X+}!fu -FrF(1LQ7R6+uAnv@#%JK&BVBm^X(ka3BHAM*(xz4eFYB);g3uE_%mFSN3c#Dc6+vbrR^2@Sn!C~6GGjf}F|=)OxnWJ}J|SuWG7AX7$0(bi+1XLt*1TuEWwgLURH2#F0e>nJLsv>W -^uNuK7@8H{1!iX}K&u;)uKW?7U0*Bdpoesr|VB<53c^tgBQLN>8)+|UTGGC6JJHyFRL>+fo4q#VC5EN -bAvvib0gFmGnMV!j)aGxM1`$JJQUG{&OHll>ZLO_16XxUddPc!JX>N37a&BR4FJo_QZDDR?b1!IRY;Z1cd394wkDD+ -Mz4I$ZtSUgtqMoDfUZRz{=d`^<5prMvw_v+tv)QQXfA8>N7EtPov1Z^p_|3R!1uCoj((0a)6`{D3%Jl=!gdyd9Qt?Y*Xg-BBn1b -HK@g|FC{DG*P0vJZey7>37MU!mpAI$4 -EshMjzk5PK%I&$Hs!9jvX{M|JluhxieHVi?qnUE6L@?E?sGpC1-R$fB-S?{&&mDq(X5cU+pR-{!p*Z; -$0~H0m91<9`2_oskp%OnquYLke1M4c4PHiYm97S*hrU2ssF|n10xq|)3GPC=EK7%lVmIPst*|XJCEDK -kp5gTQ-1EB{eOquvP{@RSbTfQh3V2loz7Kn_R!QhOQu9{LFaMK;MD3oKy0mXD14J0K0zstWQ4&f2P^r -x08X3k^*@cN!M3$ti&lRA!*QbW9?SwEV1!MCygA5cpJ1QY-O00;of09jkAW6N7N1^@ti6aWAp0001RX ->c!JX>N37a&BR4FJo_QZDDR?b1!LbWMz0RaCwzjZHwDD5dNNDp>rwN;F|VpI2X#cJql%S+3g)1AsDT# -B&1@)o+85P5VGS2WkLx -?EPA~3>kZ`IOXM1lUFi>QEMS<$sRC4Dl|l=!vx -`bOm(gyqpATNX~z&*cz*qZ?iq(d0U0tI6wQ=iwp;Cp0qRiZ!b0q9jPE8XG9B=rqe-Bn9H|r7WtvzLbaY>XtGrj -xS~8tUU&-o>x0vGlCNX`VQX_E2i43@Jc<)iC2sSO_N77S}olI#I$kh9J3EX;CqMh^oW=6mAWzhYb|JsReB|rdQ>tg{r9bw%c5X<#sy-rXbKk8G4LtPF&O{Yomn@D8aaI+_~FfKrS*p*Gbt5V2>#R>ghmJGzd|1gRHu$Ig0L={Q%vj-3c$ZtY1QpW+op&z<$}CMx16_2JUR=a -$zj4RY)cn{-~l}78=jp#z}14oNToI*AeEPO-9dInD;A|enA}y<^i#%zPf_c`Qj3jG<XSRk|PTWS%yC>n&y#Th8h#`#J4!ajT}nF7RNu%ZkaT};00u5x*Ku@Tttj9lmdj58 -p@yo_IG%_jt=Jb90aqaZD#x0%8!Z=Nsezfrv5#Ei5+ry$Ds*T(yWSvzb^%+ -~>-Y`<$fIynooEN_kq8UgUT0oaH!wevCYAQqijmn60|qTkm>804zJKP>G(ZCL2Z~3`O$yLihw}t@9*?%;W(OlbPeLtKze% -YsXxTQdo+=CmkTe$`rLj^TsLuZh;X)~_Tj8Z9MicnPTK(D7*y}FHew-nYs9P=!&06%rC>YGubM+VKjj -=#&e#8_{vtZCRQ>lX@uE}T3*d!A=R*^*It#XqR=aTfqx8cP*!9pfPUVNUGy+X}eO}Biv5|$+M -`Q0(mSpGYlMntMfT;#OJ|%^LHk}6VIo4Sg3x4jusMLb5;mub615ir?1QY-O00;of09jk3koovT1ONbO -3;+Nj0001RX>c!JX>N37a&BR4FJo_QZDDR?b1!Lbb8uy2bS`jtrB>f>+cpq>_g`^PF-!s+VS6>;6^+7_C}V?$X+b>KF|=Ba~xbtX5N)sA -FVctk8m`>5WWDlI@W>%}BP{__?EjeI!6hS^OPj -Ib({q$XK!2Y%0mEg%8ShSh=w=`S@AvHb8$_mW9s|1n!7XX-Y#m&`0M`eBOJ78xhs?N@I^FA8)OE*_JTf%GB1?;C(~n4_m($Ef|v}zPvpOWhjIRo$q`WzcwMtmkz1690&a -eTk6PhwGzReElKA6swww&|gb3-~+(=z0M%P1kzhl6hr2W -q10x%BnBBeCJgEi_U;aYLtrnhJ=>b7p5t -tkF?$Fe&mPoR6(^)8hIQ-#z!ANIxyhh_Fdbxm8O^etun9p%NDVc46dq=0DLXvrF*`!4@wyYyX5L^Qn$ -1l+X2cnbF7{%E4v( -W;6BsT`F-k;7cJ-G`@RHu$r59qMYp%;^4V5fW#2gAJ$zfzZ9ECt4zgd*sFTrRWRre&KF)BdWnSwSQz|s{@yyMvv*c*y^yS6U;thPddv^zpH})-98fJ@c6X1LLHr>SBAC&#jim5dQU2 -jynMTum~A_ls7(eC0F1(JD`r`*;b`N!SNJ@Ij=yc|QkcD`_OoUUJpyBJkt8 -M-z8l$Er&9S*zj6>0aWpZ!vN@n2t+)VyZiALQJxgK(Y3IS2PMpTij{zzRiYNzj08*rOCln|jzG(?W^a -m$TOubE3Y4LMBN9iNhQ>Mp&MnvDpvV1k4_T%T#k5<}`eywYWQD)%&)mB5@wjlUHKHs}o=?9DkCj9kyJ -UM>V56`o){afyCmj0v2wl-s_&OmVd1U-!4#rNnX91|9_KTjd}L;k_!qHctU>Gg9KD`ODO-Z$GWfYm*P -2fBZbVwi71J!U4b0VG%IvTy3ndiAD-b~>9#nmullXotDr81jp5s|5+*qrNlDh1e#c!z2Alr?P)h>@6a -WAK2ms3fSzA?UNT#bE000Pg001EX003}la4%nJZggdGZeeUMV{dJ3VQyq|FKKRbbYX04E^v9xJ!_NOx -RKxIS0Hq`hDs#PI(563nu@m0Yflp2Irfwv*-On%aVZiSaZQn0f}F9=@%{Ge2S5TOsCn3xi)w2P94GSv5(MSEV{S($8dFmzDYQi>gZVceRss(o}W#U75esrfu?~`<6H -IXpI9_yJnZvN}A!`zCS;IbN2Ra{Fk$r=X3G5q-dq$rn{us6#2@uJcqAG`c1xz;%ixERUPe;^h0u$msb -%Es7&iLDXTJ1lOq2g8JEdcj(gVfR~V~Nk^Gp-9nIMEeoNN*2UhuRI*sdRoi4qBeOW&$n=9~Dv?}Qc_z|2Q5&fu@w(Z_`X*O$4iK76uEe&=+Cqw(sx30HGUSq(_ --(RoWc^#jM%*My)N)rOsm$i^muia!C;>3?(f}9v8X!th${Q~-W*`8Smua;Hs^_ah0*sry*-%#i46I$# -kj*A(j+ID?S|-`PLu)Ro8Xgwa4U8`m?7fk)z^TfQGHn|vl6sXlbyDv|T4hpHC5)#U=p$BpJqyO&gJom -{iFg3ymUWpFk#9cYktnNQlT-1hfBWW%#kj6hnm|Lto*xm*XZXb1w8%So+GZ8Y$V9`aZ(;gulMga(lKM -(E@owL2KtLw?f+)s7`)yKYrgL~C;PVD#G_U1Wmd!JMKd0x&n>a0GQntJK(QFDX(6?z-j0EqnI(zfx^_ -w{?yl2l~-T8Ws*8?M?F*vL1s%A7^Fb7Fpj2{zme`xc+ZN -~Dsv3rxRIQEnj{gTP_$ZoT)f-q0H#9M*P-YP7;x@&98G>hy!GZG@Z`w}4`?)t#SJ<+{kz<~hs~wAozH -+gPIxjO5o|Ap{{ZU`+d~9P|~;_1CEmBza~XI?UE2u>^2}f24LJ)*RikdhzP{>mR+s#m(1wA=O?ra%;L -Ypk(5kI&TutK5>?N$;O?F>IxOUMevvK#7}7MvCUQLPLJs-KXS8TAA@9r=GMzgmI^mTz=FTi3eI<+{c8 -nGx3uG=zBR9sh8qhG+BC2FxI -G!CH9i>I0lo2(pUc#}?o1GF)K@ixYDR}D#&4x&w9?`pnINzVY5Fu*P3U^gl3mI%dY9?BRLR0t9h1$s! -I0j6ObmzZ4dy9Pi1&Auzy6B$21SM5;Z!%5wTn-w*CJ3NGEQG4sI>k*d-M|`zW-KzWgC>Qlglcz4CS_J -CVn(D5Ec{?%xJ-gMY^G>$!0^|3CJ9nf3Yn-qh}9l9S)R#WB+#^i?cCCc_!ukRh -%Rk71lEaiKS}OYui9$-WmCn*Ix-@SC#X!P=24NrVCWp@8EgK&UT%BAsq{>lmpR6m~5fQ8pnjYsp_`-`tySu6r@QPsk5i -koqM$QL~c#b}vHc{@xJZpU+=NwnX(Q6%t@3Lx#dS%S5u({wT&0V$`Woz7qf+$o%Z>JY38H8}d$Gh9G49r1O>=Z -2C1Gu`u7*X5&f>LXG0QO^Le?-?euzLXVo~r#(}b;ZoqzOBlwW<_sZOsa7EvA(V>`7JMZTH~*mOj1Yd6 -(yQ2?0=zvQwyOrS5vtRk?L8c8uF6KLWHmau(wu>~8~0}l)rSOKL_X`KTZEi`aPL-Si{!x^v!gv*o{p9 -hp)kL1$89O3^u-ZdwvsN)Gqc=+?rhATh+?A;}BjgyU!V^M1Y{`1d*d29l5I9SOK6pE1tRPNoNl9Fc`? -kc776{saZbg{0bocVxQuE3!^;a=95s?)n!a07XMKeht&&ga)S-ggh)8s6(o7D;OSn5-wR-Z5Yk#ds&I -pu9kf(??Ui$->qE9X22$EiXpRKnO7^9$enhP{swg%}XHRrn>2A3gjU?k=ANsBI87Uzjb?|5kD1=^wTvJM1Rr)>@Trs=JG0H9??C{Su@jqoRg!tszRRu&x^T3 -SOj^Uq{JdYFsr{xj -IQ1j4NcCY-7T7GfjI|vbHDc#UyQZk&}Z|`nj6JajUH|~6I{JH8L>6=0E-tS7=x9#{GiR&yZgOEV<<)6 -6o~bKjcXI=Wx*NxHQtSn#4G_bYE3(L)2$w`emPZjz^Nr)$nI=L;uvrJo3FJEYDVbgp|t1pfG7$_Vsh} -q$>O;$1hmX|a`)|%f4_J1h{E7J{6j)HC=+!KJI6SZe(w+uCO7h{Sjh&(Lp{jASYRFkobQK)dijY0L5PSjl=}pm+1=h%H`qKHiPar+ksxo%cS1lT -d9DJ^1!Y;a`hhAS?lq5LxmIfAG*4NO+dw~q`vpa7_^|+3UlAleUufU-ePHo?5MnezrK~oSwiGugD -utZqw$WnhX@oG>p!&G)?DXSI9(3Ju}(YBP|wylx$2q#Xk?Ou-fJy$K3ceL!?OuRieMz~&ylnc^AQ;591=qz^Ig+K{^r;XeGOBAgzlfhY$$?IGc -}_^;q4#v6-Z2N_j|up($-W#1*4ld{o>47|?aA7SiOBZ`0l$u -MB*(ctq~Fjbmzk0h4FO@mI=ARB2pNWVc(gIS5PInj_lR)7%7Co*_OTeAh9Y`@65=n6fho -m3gT|HB!K%uoXh5<4xm}shn4w5X;VhOzKn)|r}dVmeucdvR3CY&}qd#~Sn^LyL}mQNF7IcJ@-ouOZwI -y2@k*^AcXxriqM%C2Btgu#0%qfI%dy+E -TM5h2+Hf1Pt1dIF`I22@=c(7A(gT}Pu2Z*&OZ^a8zPGK -l_(5QS;DAPytQ7tmjvKl9X0*`p^GK|>&8eu?>(MI%pQ1&|2OgLHs2iyQ-8EOx>{i5VJWAIDI2|?gY_y -{`_zB+8XjEdv0WT$1V07pmJn8d;^S+7z^I3V>%4lrRSQ5neKC30$oII^TN}DU)I(tdCVM_RxMps$5bm -cMp0M@;!#t3GIy0W5Q4j&UI#x94%L>ua^2Gnji=;L?;ilTl??nVq`3q97adZjm9E*y-5uH!Q7v8G(38 -&jKcv~Ja5!Tk3zk>WDE}I^PC5#v5H+4wmlVwa;01%aoTs7L1CP9S^jZuAzVooToxI -a%peJWyKb=J2Mt5f6$&wP5WJV5WKMYT<-^Fu>4ioqT-a*p*7wa>2uBSxnb2{fpJMIpA=YHhXgupLazq -Gil0Z~b(dhV4P}MS&FZ3%x94M4mobsjs->$KehloE5T^G6cEfKXCXNO*E(Edg()N=7%_*T`iloF1tg4LxW_^;WY|>m3y(pG6CK> -IN%KSOjha8nbaljP7U82-VcB(Og`OU2~|!Y72Pf6*!XFs6VE| -1Bt{I0b#ec~?xu|34bWjcBt{A(4F=%1CAiO$>B!Q#Aec3x5_$~Vr!F}aOz8j0*BddRh28eKUE0UTdk* -S}`;?>?8g}m#zywDk8>3kdSh@>vaS_E7!#$s@9or4;Ms2JsnH%bTSrHes4*M{pl$qP`)s6pw0e=F{g{ -Q|oTbSx0WlRO@80WNA&aTEWs;LzqTBea1K{gHSCbCHvac=&v;o_O6nvLea(#DCL5LM$_r0F7Xsn1QAm9l!du$#5=&WJgajcl@_`J5q1@0 -(3UaAppG~!uOJC&9<8sg}Z;!4|O?v5j_?4W$-CkEJbsh9~-!%HF3~Ms|!-J}=$X9VMo^*;iwx -7jUu)Qnk;TB?l*Xi7YD|cn|eYj_`7PqA_m_GJ5JX5|L-oOw#L>p^^T!GbqCzVyZ-Qsm1)!JKbO3sh`2 -res_dsvn_G$6AMl2P7Nb&yc3jju+b3tj-~itCt>2a$gfM(B=&ALM>%?zD-Lra_i~t_#@*8Tf7ne=r9w -{XmUlj-vDF^C-54*|<7iwVYWlk(^PEQCU<4_DirJ$C_q5%7xu0>TbZH^Y)jRd9z>ZP9vlYE%UM;VLZ5 -|J%L8sj5|ei(ax}~DcxOqP{y4+9NpCu+Tg(@#qHGVk!dh~FAM|QFz7n?WP|&>_@sM%9_Lx6U2zS}hBO -WqvOq-14tMGhyNpdG%kr2i=O9?L)B1L3zeCx*_IrHBa*D5$I#0m2C;0tFK78M0NXFoXQ|fq%MNgO)H? -FEC$H@9{dw90cTfOJtZJVZ`c<~&h?Qu-H(C(1FIhnt-@AUdD*`}yW1?<3>X6kpCMc1?^42pq45R4hSmkZC1p#lcj0N00~gc{S1Z -Oe{9e!C%a%VlU|G(F}ttUrPcV0OSN#I{M`;kh4E}dqOn}8elq_jvrHY)7L9J?yo9j?7FH`PuZ{o%DIz+a(d=w4n9de9jnLBfCkf -z{oKdTB+Kw)Q9V}}tiB`F@$;1_O4L6}qaFU3e}TfPGy`HI!3-Zr(Aqki6Q)BX*9Dm(&%)Eoozn4vT{wRSXy)=FAk%!dS=SvHU_jy -SY}x}cY?q#s{Pc-h-J{&;LZVNb;nnP(H1EQY@Puyzp6OiB1)Z_>uJdFwea!j8qw1{2AtF>IWebR^DMw -kg@1VAZ*5<;3MOu%3>UZlTLqPQ|8acIx!xc`1 -QD3#t$QuPbxauZ2M`zf*!WRzKPg^wLm21cegZpNxm;*Im!hBp;-{RV7WU7Do2WovxGWb3}^of=O!oQp -JA~B`tU6)$4a>r+5Jg=j9bRB_eBgg-&z|e4(SL`f6cy!j~!UdKZTxetaJy$MV3g)Jw%>R3F5%*5ifU_ -448yHw;Q4&VEuoWVf3qJ>3OM{Km;T_+&%)EV_uE$-OG^sUP{`$>rzyggFrK1wlF#*R#`Cf<1iV$f3r0 -;C~EBJE?UZPF{LBr@u@Ym?_${M*C8d2cM=XZDn0Q&#BD))4pmyb0x+E**_V~KNF|o4s6E{`!aGLJFw0 -UOh-Su2zhBw*>AV$QJ3EOIrFimw}ij_jm&hG9Y-dcnL%xic^d-}mtg?^4<@%6OGT1$71;c^C<1TXcKG -BL0oJxmiU%{nKo5hTN*&pyj#wamgjT`fq!@mN~euD{vYMs{ib=wIm^9sk3RnMd08x^Tlr^Ff5@Qj;I!yXL5Z>Tb{uh -a{X2=tktgKThp2ox@n8E4jmL=BPJ1&IyVtoD_kcB4bMEbea*P$#Y>&BwI86AC^4W`S>ZB8AMa3WbQB! -Coejbz5zXhpufhdKTu0$a*2Bz(t}x2ysHO00RrrSu*S^`%wll$DI!jQI;5tcvXtR0%q^+CIW)Km%}N(DX*;#woqyLUTRsQ@+-td_&Jd -UWrK1vL_|REh5%Lu>e3N!leFXQ@#WLkH^H}6V7wOR@d{=sck3^m5bI(}MM$9x=&C)dGBQzy;{#bVaEraJ< -#t7)o1VKQZEt9@IOZ-#=Wac;&d`rBEl1aK<=b81b$$A~pPZW{yRw8S4+V3|o}2H7`FE$Dt^f6l12fvm -8Vwd8csD1F8C?O)H@_}A55wn$=&H%<94~#8T(|~;)#E$qcRdo%P&(;;Gq7Zm90PxP&9HoLsuGtet~k1 -a_UzTuA6}ltbQA5{7w^t2P*l?jn^B7Z6;@vqcnfEDg{DRBP)1jH>d3-z7TL7M`FO|s*C4mPZ6MQ?%|L -9P!JeGGP!%s`oGC5ne%KKyy -Pl?Fb$s}h_4E*Ltf;3z#9Vi-%x_4IZH!KDAFvgdeKpzGR3nS$9y^q#gi8df>%j32QbB{J(@DK4{+ule -eL;*p7;73cVu?;+F}b<}y%G1QjLmLdgJhB(tkvMU5K}2K9~5SsYV&~(Aq>q|HWtCk>}OtX -QXOEZYJ2OGK(xk}Tvre6=pDe_aBlM?)cTS%44+%(K=(P1Fhh9Y0i%D6ugf7RC#6D-0e?^F>I@2W -EQ^x*BRiRk9s!!u50@QjYQSh1Bph^7~PKRYcJW-YqpbmOrd{8M2Glk*Cw8xr=~&2VFnE|kdJg7Zkc-Q -`7NPrb}cK#i=*gf4=TJ%>Wh)Xw_JUo4(nir;~E4*qfnwgstHq=46Lq3?L>t{f&GS+^A76Z2Cw9z_Beg -} -aLN_v(8I)!4~4TXLO)zNo)|FR+2V@}Lsh10y1%ux|oLD!`JdQib>5^J>`D^!^J0J -HNr?$3kDI4S3-$i_<*}pD@{(XfxOT7|B{y~bf7q9;I^d&O$?A7bDSMPeAsXOAjRu7WoqN02YG}5PHfxPH0YQrSbQ2E;2s$^f@c=A;0f -p#{V>Ie_qeYCkZIsdRzuc-7bW=Dpj`X|_^ldTPMG|O2X5QXMOIS`d*R(qrk62$TEuqXu{QwVS-z9%Aa -*)aX=9Q{gX==PsQ951|AGVCueUcEbIvgjQ#zuKU+D`=yq_`o6dq5OHQ`TtEzBS3q} -6ob(h7JAWGHXdTxQ2{)h(lJL84z6NQ?Y$O4%fjoREn)i}HOcwEt%w0l^`Wgbf#MvSrz1b@*#b(0d?^R -1-6+B5U&}u_NCZGH*6f_y|M?tB({^<;F(#@o>@wetMcOB)`i7r`DC)|C3d|41*eky%r*A`vigsmeig5 -SYL^E=+5{|``00|XQR000O8%K%whX){rSd<_5q$}|7~9smFUaA|NaUukZ1WpZv|Y%gPPZEaz0WOFZUX ->)WgaCzlg-EZSI5`Xt!L8vGq9UP&34d6aBm)$NdXcxD8Ee=7Tr7hZ4BTFhtC2{xmf4>=iiIgZOyV>5$ -`602%;cz%#{LM&P-X2I%Y_DqJ?Q7cQ^uSK7yk}g%0qxp1%{{Cuf7PQ)V!rsBSB -+%6p>@8gE4SsLgu^MV#tHMPDcNJOX%7e5l*4W?_?owUcc~}lRtvCMFJO7e+%ez%>+Epow(*AyLR+d*DC{1!$TN4JS; -OW869O?6b-wL+G4wJy55-H9cKN;n^tblulfvL$)om1ITR%dQ6>s= -f;2eX7$%mHTfi{HiU~jALogP`Jm7AIhK?>CMZ@< -?Du$Lv7wz%Nl64$|YU_#M5K9Qw^34#_CwjNCuKAWE&yT$o6_a4|QU-&Qq)EWs|o8@OIwyyY!;kzge%N -|2RHNH-CU-Yymk%x_@Gl}&Ux9P?U@rG?5BMUerSw#qfIFoBa!Le-c9PlErEmQ)P{D&)+P7{?Jy7~9l+ -i8^Iv1#Ve0YX;*+f>6##4eMcD)u8df^5?*dHKLMeX91oAlPxc;5}g98UDHB>=Wy-{{zHDdhts|1JyX- -YfE{4Kib=?~9dCE!#-yuT$EajmLd&u+6DdUTHY4Xhke}Oz4X%g7Iix$JgImEZNF9X|@4QWxqmgs#O`{ -pAlENwx5)UFl@(}?@M*CW>UeQ{xaB9!$Zj~tATkcNv9+&L1#qX!$_nmBUhrdRA%l+$!Kb8T+x8C>Hh; -O+Q@GeY2B7c$m^;G`4g&X~A3m@*Qc6zhbK?paU>}iehB=>L=?WrKUx?R&6w#B5gBxOqgO{vf&$&{0(u -8-%_w2+77*-LbF-UGmzoLk6M5}A0pxqoTuBhlBXJC2w5KzF$9Ib<&OnLaRps2o*V!2Q2ZvoBBxz>fcg -qo~A=9?#A>U>iu<1hHMo?pfV-2$G1o(?m{wJlZ{$F<$}fKw8u+A`JmIoC+ql{DixFzIwJTKTvr}Ls4& -1Hi@pa%1fYN=nst;O>g#X1p%{4hk250Ib(k^{D$R6l08r78bc~{VO&rPMChSyb{E8KKvE-8wj_F^C0w -4W(8*)k2QWm=G~t=$xKX7Nf+O(f5|EPIt9^%ijA;FSUE{!>=jJm1f{POE1RRPP4qn1`p{tS`A`m=n$v -P;I06$pTT_@Yc~jOHg}UC-h|K$jjpmI8)Nu|v4*Gt-OI@KW!tn+0P>(xW*-U@k*$-z~kZv(H%; -oHiRk?;etN%?ilCwgf3Qd$ze#2zY6NTwIuLMFbK|sMqldCuIw11S&&?giMbkk8=BlT+=^~u1b -e|x@YwMsBKCPZQ0jY5DEAy46>@4dw_nroAJ;#XOc7wG5OAm5D46Un%m3iW+^C?k0fJRm;2eF=%Y>I~zLNJ^hF8X3XG@QV@3RIFS9L~2 -l&omaq9YEt1*j=Vx7ch%y#avTjlxcwhmmNpit)^JEm -~`Zf0iRBvfpm~G>MAy@h&;^LI;pWt=ozVSs#^gv3Z916H__HjmKtN)u3ZUQJCpM6ufCUhYV96bMgzgJ -Ac6Vakrtz1bpjE?OnlAS#2u}PeF`msvKOi4Fw$&(H;=KaU8Rr<#THLx&E55kBb=Q_)gJ43292?{lqZE&)EW;vsDx>zYIdhYGIy1IsN(L9;7B8^HJcP -B7tMgUMwAz00m5X0u|7>18Wri2s^nk=DNL)pf^;KJoz_?vIptydo&2Cj??+Ae+fXIW6#ow@fy?wcItk -+Huha|Vr{azsF*ZJq?w;}W2BbNDiCjEJWt05l2v~>nsXDsSsh=0q3~L;}BK+hhW>+=9y^Tp15@37JNZ -EHa$mg}WmkPq*rAaD9NfjFNSi#sG16lOrCFK$L7znatBk-XKD -d-U}lsm;Lt*TZDP+2>C^04L!=lcpS;U2xmo^LHxU8?`y4&I`aqXNSekfs1}SPO3{TP22Gpo^YWuzFN} -8YZsJ^fIwkiKm1qTtV-;h&}%t-Tx{erc~rVqhj2CuvP)9l@;0?BSp)bv^^j$}#9$*CAXV;liIZB)3@=AE6UUd2GVlvF`!C$GaeT}|VKgh8!9L3 -ZV@ME4HDHE(BW|`u}siD^BY=bUVgH$;_M=iqnG*nd7?jw$snDIIMn5=>XQ~h8Tlx(peT)yQ;9o>VN`LB>pS`mVv`f9LGUU>1dL(nFq -t(#U=Ph@{S0ucRF?G%0$ghyeLpgl3tDg0T;o&aJU=*&NRZS1;lA0`fGGL+=^SEVTAJC{v*&Qp)=v2}r -cJ1`lpcf;L!YqCYWSs!$j)lL;Oxf_PZbuC0Osy~ZaW0tuD6rUWc6id{qf#)Z4@$M4R;!ObEc_G+GW8z -T=%unR+iEq8%M}i^afT8 -DfNE-3D?5P-bt~QEg^gPoLbX~z)>xy2da4wR+A^P!y?QLU}s4m0>d0@J8jexdk9~6ZEq>c~Tv@`^bE& -M>OS)+m>gSzayX&{|2Fs9B?si7&h|XXi%r5r5IL$PM*s(cMmI}FdRAJ5hnF4tmTDiVol7uAr? -`_qj$$3KlMjlm6KB{63Wy(Aua0Vhp1?#};gdYGgzO98;j^v?*5$ZX4Sy@uzJ99S^+YjHvSH3RW=H6P` -J{-~SP&Qy`h{t==i|gXUpRTU16EAJ(jQR -?we{SWgwl_`eQIp#PVWTnqK60ArhlL0ZR@f@f+Be4 -cX&f^-?ugrqnAyiI{)MV8Ttfu(X2upQ9%7wdNdZcE`Z}`7^J2#B=OOJRyACxPI+2f8CrR}bOdZ&aVd$ -V|Y{!f{k_pB^azO$(J9@X%Faj#`ZJ7V`c~(RbZW5O^ZmqbXYebefa}R6WP)vk)#G37c7*!0N>PPWAzu -E?^cZP6cFy3rbnT4|i{=sI6Vkev>R?0|XQR000O8%K%whn7mPi; -syW!r4|4H9smFUaA|NaUukZ1WpZv|Y%gPPZEaz0WOFZUZ)0mNaCxOzZExE)5dQ98L1-u-1G3tEST_V! -u_4QV0NENeX*VD!2u3=|Tqsc`DX04PJ5rP^I&x5F0uoE&-SJ)?-ivtMN@a-d^{P~|CbBgxQyK^#7Ctl -4fnRHx*$RF?E?wi}%CGP?rMxkq8Ya@5<=aWX-&+82Zr?1D3O*9%b%C36+2|vLse-RGmvzmWVq}U}1h0 -?qwc`qEXw3Prp1Ajq+-zwsfHhs4tk%cmYQs#YQFH-%6nyjP{QTnPW{IUnI-%3H7czgCoZ+4l!ZqD7!3 -#Pp(~=9I4XYQemn|sfOiu4RE{fqlHHzgAphnRZ7#b`riRuhTorJ6^-c&QrD}3`DN6mE}I@4~4lynD5q -iRR|HgFrLTUzi*)mwc29V3w&sruDQxXnqnS0tv+8I-P9t5q(T*5nG^Kd=fzXK`V+fa(;`LQ_DKb9`;JKU+EW4aVvElEju$A}K14WZyuqNvEyj)LsFiTgZ9 -I(G96*{(OEzS|$HQ`$(t6^d8k8+r#*o -4CA;C4f76WNHKR2@?cfq;vH0Ec)|qJ8L7l6Oj>@WG_sY-K|{mRy8d!3>ht=@eLZ(`k$$LS>9JA+OM=i -ro36@P*uM;0t*t0$q-B~4+WHj6$SP5 -u?A7D$>u%zOEz%kWfEdvrYGf@I$hIKbEV`NyYI6&TN9GytIwha5sBIibwJA{b^g!Nv_4MwjmuH -ThgCbs^#+&0mOzQjvtV2ck9|9UlF-@M#)+(?pC-FeR4Z9boGisHe%)_fjDxf|rnhq)G1>PAv%(WD*Mx -QF7(?DBjfBNy~lTon+>>P&e#}J(gtCCk#gE3e=PUpAA^MX~s~q2k>%erDWG0&ANjGM-yfnr)-uUtEF0B($$jL+h7f;lGy};^>oWuU^@X<)Fa~Jnq>ZnQD+ebZcb -$qoL>4W0Qg1h;4bkVI$3L2FBLFPY5ZALhegLYV3%xU9nBr4vN?LmNgYNF&rBcQg)_Ob`70?mEwTe`Uh -(U_8lAxh*vL-mLhlLAkQK#e?pAK#DAVKMZe7m@kjLqxiLGXEPcE+ -^GYKAUsrz#i|y3CF7rcB-%*GVsmBPv)HfA!B#lBXclrgr?$%1_rW58Zygj4&k9?*k2b}y9 -^dv90fjy-vRl=`;&DaG^*ne%Q(zCvbcPZiwq$hpOvMr1a~dUXGAZO^BP1)<8i -)c?v0M5X@uQU%OtgV1~H?z{%8m4N5l`eYMBXy74Rs0lrBGBj_lO4gzfaF&ukGU%r-qk-Ri*IW^i!u>z<~-Ug~Zi;Uc8>o5i-v03=&hq3rj@;cnuUeC1!;}yNqHtE^DnExUng2C``;5%m7Q>yhb&mFS -b)@%IqVJ~m1oj|R>hI!O}%o$M|7K+2nH-|@6aWAK2ms3fSzC4K2GVK=005R2001BW00 -3}la4%nJZggdGZeeUMV{dJ3VQyq|FLPyKa${&NaCxm*TW{Mo6n@XIAXF5PR$J(>?Zul5bZLqfNYbEAw -gRr9vKDP~l_`~^l6aZ_`_AD-5+$YG!xSKq$aCj+F3d)jdy-|FTGx_g8R7e?l$uZ>N=-E{g*rKLYel); -aU}sEl)pFAX=CEJr%hZacV -ywW0CA@pJTxkYH^Gg{GN-)xD|xIRv;i}O^hzzWfw26Pv4v^ve)l!E?-T^;^y@9Y_S*x)E;^g^%hxb=pX~{p=Xviod_*bZui=5qizN^F4+c8?f^ -j#^xq!v&OHLpIHg= -n9NQHJc91oj6e{~m#X0iHtUv%9p?RJ;{#dbMlaT2Pa#;#Cb2=c9u$->RytLs -^lc}+7v7aDFeT?b35D0vaaIilQ1wz8&g>e6Uzy4k)64UIw4eUyZAHFhAe1<%{K(vfu#%FlazyJ-MSJ- -7aVX0Ip$6{N46U41_AvBbp{Yg8YPV^vjQ+c4b7s7=#Faz8^$Zk9vkAujaG;w)jl5$Z`?@fgb@;a&mqO -LJD%$u@F2Fj-a?!^?i<9N^Rjav@5Lf`ef-_bznl-lFLRz2$z>Xd&U^X;ERK+CjkpCk7T~JBR9qG!v|1t5pXl2xd$I2#DjhmA#8#QlbExKTUt;*$VGz!l6O>4Hh9yH~QjKXjQJHAzSFA&j*H8H)9Ie2B1IMl<4o8M&(ISB|h$$ -`0CXXYSeb+T=gXBPJNLUJX*N%!cI+pH1}P;{5&9smV8}V!+P}$hsv?4m0v##HHr -`#eq<}1x46*-IV;v$Q^~+)a)5CZ9C+U-P#N7aDr7O7B?;mE?5eRK@YS^*av_Bhmq4aMd*@1sUdfatf_ -$J$V2s)#3Y=et^9AXCwNLU_BBcJ2q(xP_clV~j|YgB?tK-K9osZ46#JHq%U0SdrBwwNirB4mY7;(BF -cKOOz|Gl1z_vTgMMpjX1XcxIvnJ?xCs1oZP#}kI+Mv(95Pvm^AGfX->M>WEnZWClRTQhV<4{ -{5VNmn%cIN15IWUyeQ1w5I#0#4DVp38EGd3X=a1TpZsH5Y;TSOcaa=}h6oA=Jd=6tfZm6b<4rs9a&`< -t(Vg1k= -xoZT41jpkKA%yWp(KR8+KOds|qv34|ayN)69A`eA3(u;eKD#dtqn288!%Kp&LpqWjGR!k_}S}Qz~o*) -|Y0#uO9O-%)uc^e*)Rl`ym>aqvu;N5qJ7fLY}#$u*v}mMmlcG+k`BmDKI2MG~v_heFK_RuoB96Mt<&J -v;NOkidX2U%{Pl-GOX<1@VVP-dzX&UvWbs|YZPGBpIDD=I&!R2@=L$E20~ttAD<7i_6fm%^O;^|3w09 -+7;`pUH<|~WjxNL5$=uo8zYJ6S75nxc*f2c&HtKQoQABW<5&Zl>;-P=f$WKX<{#9}j2dX=L+%PpKw22 -;Y;Yi{KyLL=E=g3XsIU2__dl;xa9wD#~PGM)n9$YGd(udMRXHAowf3N6;JpZ9%+a@)*mfQ%ASBEoUe& -e{`ACi=x-upNn`Dd1?OYk!5R#U7Ro_0cDTAT;nFgf`bP)h>@6aWAK2ms3fSzAyCrI@@4004p?0015U0 -03}la4%nJZggdGZeeUMV{dJ3VQyq|FLP*bcP?;wof&Iy+c@&Oe+8jp5eX;jcK2R@aVpTurhCQq(jsYh -KLmk5OSGenENLk!iMMzE{bq(DMM|=JIHR>Ca$cPGFgw{CSeEU2)k~3OtUR=hRE*bkqqr)Yx?8P=wW`_ -g%X)8KJ0?45$AZ6$^$SkstIt6DBQVNibNRWCZdr;aUfFgw@$Ac{ -|YTZxt4Zp${wo-d8FU^;0REdU9w>j1HR|%FY_?S(ZvLzy0M;Ec``O)%vk+o@@3|lVqQ -7-h--Y_OC$+Q9^X`2XLZHH5r(s~(4iM4Rr3tGLLKE5?>ct-&^!-r5e~nR)(DtscDpo=Gv1aGkg}9= -Qnc(?@F;n8OiB|r(DGka5NC>HhgJ*l$^lkWU?5idkwJQ6Xv>CHOx0@k_#*E87yrF1!;M4b{Z;1O*-6h -7X(s14pbsov_Qp`)H-j*AFu+nv`uKJ^QGoe -}}S<)oY-Lx}tW#UhaEf`CKY@+DhbqbnE_trNft)`S=mfO=$KdTX8l}MPIY*eR5CRhxK-rUOE;^?H$iQyQInzu$xK{`ik<_l~>te&c>99+If61@l=ixvl -a>_)8s@FD$9qZF!-b)DDS5KN6Chy=OR0Gj5m1gR>%=pB(3*Xjo4H6vLt}hsS;=<_(ZUQM>bw-zZM0cdM#|Ba)2Co0@Vjpq69UD-kcPQWS;9S_lId26OqOLvyzQdqy$oz;RSn)^Ny2LAva8XFNOJTPPRTje3iD -X~CfaA3?W)`asmuOAC+^93&Wb{9(cpofmlEV4Ecp3y!jYIGTLq<9Lgm!&5)+jSz0 -D15gG4XJ{-~kj^#08m3xE-*@-p;Y{B&K?{R_!p0hRXfU01!?xe@j`Iq#_(jL?cA~*5&J~t&93-IuV1Y -pjm?Kckpa?L_ZbDCDPX_hBhrUZe{^(kfmvE*d;~fp~fBpVDRy3lkFBEizatFh37i#*IrK)ngn2;u`k -MRT^VKGbxSNxgVAt4k+!MxAsXzL$1aN3V}N%tEOjzqfi-NAKijVQ`StJowvgXr%7TOw~D0hOFYlvsh;!Qo{omQgBcN{c{efzbJfm~a#6$(zHmqgr%t-x#2D{Ls8o9mvlo(yznQ -h2-C40@j_0yVrPaFp8^*UInjV87-g;yUyn0~_pZ$n@ev&BF(89fQ2$f4;oyrzMWCLba~-TevJpUn-%2 -tKeng>h@{F-D9boKIocLh0g|Kt6@^X1MPdF?2YdhhGo(AyX_J#9m8q!Va%f7JeVIKSy*~p!5Cw4&{uw -=Yi^qGpFS=inCs8k@5n$Q%ZSn@4+0DVa-#<`Qn7V5Oz@WF4AK+tJlHRK+z;<=k9J4=p@?{laGJI8+)T -S8TE%wU++d=K`++Z<5>z=U)_iUotd0=#$x7AG3$jeoBCgLhR(b34ss-5EBZOiOXoszGTUG-9E-S}QQa -%dnMHl#OiMFjoRj1VtI>M6ZCntAs+E<(yPN#q#2^&YJD!uo>EstBaT*9K&-vziQ62}yi#4Z0o)Pjb0R -*$F9+F1SV2g9zoYE!z#af4+U^iEnA8vlYhHl{|b2|Ob5|U97PP|q(3uig27)18%G~Ad?itXyZP)h>@6 -aWAK2ms3fSz9p97vS9l000~h001KZ003}la4%nJZggdGZeeUMV{dJ3VQyq|FLiEdZgX^DY-}!Yd7W2T -Z`(K!e)q2!2qb}IAAT}Mn>jXv|Mc`B=9~^1X|O`&W}`v(EGI3_G`Z7e(6(FvJ+js5rEt&8OQ~=*e--og3ZuB(a?>a- -EH_#QYZe}uXdvqm%68~}4QERpu*Er1E%H)W3tv>{WS8G%_qm$Od%6yalOzeB1243wM#~nhAF>8%K|Gf -TPD3-$Roc2fs#u)a;1{Kj3Pyks2G^~yTlk|f2!*t+Et#HcBrO6!K -W5ZkJe_4{hUGvJ%)d3s>f+>fWWdEGH3XL>V*@V;MDBHDiI1f!BJQndN|!x8P0xMe4#H%KclLs3k7b=r -Z4OqRLF<3r9?y6x31XBzslxG2Qp<)(Op{og*USL^+!Nm0IQd>q;G(>etcuK*wJ!r%5zeOq(zC_so5O# -K!+*ldsJ3B`&--XwYKO?FFV03U`KZ;617Tlas|`>K-AYr1W(0aWD(eDE447icn_M~%G3L7$r;r)m+O| -yJ#;*@(j4W`bw??8u7dJHR=&$Q!@tx64y?no#YD>?hLv1#c%&N%u_ -XmC~UC^hKH@WrtnHRf`#71sCM$6B#=T4u<33{0b!P3r%ZKoddLy=})?4y5@&+0&0ig8u1;hh1wQ&b2A -+{P6i_(DTC=b_(_@U^CKZA1KuYpmlA!jd@5{+UleI?1)~G=08@su{_ihQxE6w-wJ5?@DoydL0AC``oX -32xr~rH2bqzp2J_fK+R|BOt<<04!;w%{`oE(J7fPpBuTo`E|zgok3!P7gJ{2Pep7ysk;cPgMlO`IH+s(%TrZjfrnx<$zYl~S(Mf)P%3GU$i4 -;pG@RSe~bg!x7#-7*~~_u9mTjCN4pK9i~BnXefmfri#wqlRyJ{{c`-0|XQR000O8%K%whIZn|;+6Djs -eHQ=#9{>OVaA|NaUukZ1WpZv|Y%gPPZEaz0WOFZfXk}$=E^v9RSYL13HV}XJr{G)^l>uAs9)edTsWuzYY-YlLD$~E2ZD)m?Ra( -_dwYKHj01LD~xa`2pJh2sRnv2GQZn#VfUTg=xKeqt#nnosBrRW)px=;)YQ+mC!;{LQ@DZ>C9>HvI6POUAt;AKzswFrRH|&DXHisoERbj5!*4 -oxwt=Dc0(}S1DMWj@yH8g9y8^VcD-1)zUzoV8)AHK_QWabdBD6@Jl6ezyj&!dF<^hNjD$JKM5B!M%SuNP=E&D|^0dIO}IK{F -^d?q~zBm;+ocl9`59*wMAuvfTHk4P9=o6Y^AJf+r{bvv%J`IKPAJ83gz(xTH;Ga~X9WShl)IZymLboL -A-nPP8H_U)2}w9D*XGc7gxoS!N|gPIcJQRxR-G-Y3K(wKR1@M-aF5D)uJ+Nk_qG+ghK#btu3k^5(gwr -@tHxmX^5aJR!`xLDL1{SOMD6s}-hDmSh$m&zJlT -q+_qxF16vteU*t+wJ8c#w2-8e5jP6QIp?uO|lo4qgA6;xUGq5KUsXxtw|~`NLe;Fmga7C2WLvqM=lLa -l%M%~)x5IID`%qe`*zqhvAXQA!70=ExJX=@xRu!C2CT&@WS)l&F8lv`-VQ{|VM2P$;9wKk<;e#<|IKT(69v&Ql*H(Q9&MZ56T$zt~~5~=wE1WoxytFD -cq7_TU&!STchI?vp}MxXcG(@_cj+SuUh<;c6mC1MdK!QxNR*(C&@>8^1t!jgsQT!qsYHlcchUs4!*i# -74iH4#`C4?V=0!ac5`fgVlTP;3!w=Omp{LW1OhG$BOU;ktPz6YPYhIrqd~YS~*DiT7R(y$NcHPbJ9-% -K`7hS!RlBV2cl7!BwE%mk_b=)g~uLvM3XW)&qs3vrlrgWF=V0@EE{G)!4-~PCx8t%FvG;t@@2r_jo4Q -=Xkn71k3C*R))zG#f4+Wp^7#uN{}IMmLW(}_}2 -f*#K&cKro#3&PT*3L|OE8yujT|v+zGAdN{IxN*qG<*hs~_hbK7|^HMW~e>yh;qdrlmu~^F?=jPKlQKXxO+H^^DR_WAhY>|Cno1?!>p0JcyuwWw;pst1!% -IR-Vaa@I+iK8MxUgS2k=%G5F-qYOS1a9vR< -x{zP39R176P>nq6$hgtsZ@2))AuQ%?Bldgp(vcccrrH@=U*b<*MqV) -JX06Mq(PMiRKX&A*FaCN*iM;l-NkkfPtI8B`xUYv}?85U)0&w>j_LX*Ut3-(>cUdPkLW(k|<@wIO<|L -EpG@$nh28eshmp3nYyR2FQfo8g-IA5cpJ1QY-O00;of09jkpbqy{%0RRA60{{Rg0001RX>c!JX>N37a -&BR4FJx(RbaH88b#!TOZgVeRUukY>bYEXCaCwzd&1%Ci48Hp*M0Z#VT^^u^!C*V|H0TC91}9NkQO9m& -XQ6LDC22N~7)lK}$ny82PpR28aH3O|lMTpLo2()qn2I)#vpy@dN1*29IKWFG9bJak=!L3pG5EfmC_Y@ -vCEiES9T3e#@YNT$@QmmFlM5fT`NeL>a);-Z7#p-fDOF(&jXq)?i{`cC37zxb`=tUoQDc)JU8@y5Jtx -!J4EV)jvBS`^^`IecgUCWSAnfTIZGu{m1H%KnRTOk#5F&RTI+AI%>GZ`zf#8R<;eVX)dB6*_vQ|;LvK -id#@&d-~sf5-BGSDHEj5?rrY+NP5g1E7j=4!IpWMdq4oJfXOXXC8>;bGXs?w1j6rLJDKLKhzUXokhFr -)s|WggE(pK+VKw=%n_$OvN2}T#I1hU%g|RA5cpJ1QY-O00;of09jk6tDgJS0ssIe2LJ#g0001RX>c!J -X>N37a&BR4FJx(RbaH88b#!TOZgVeUVRL0JaCxm&O>f&U488kT5WOr068``L9oBZi77Qzfp_iQ;*`gC -IvgAr~vjF??qom0BsN2&51c)v2y~jsVX^rlns@lN~2CE80*K6Z|DWx5ALMxkP>0RAqtq*sHUZ_n9efs -rdSv`DRK73m}e#v4)FZ!Yq%ArwA79BdqnPWxGET*)3`u$m0XGA@Co~cBG&nbMRtYcFC#OFa6>`*z<)n -1a98PI0}ryhjl6{?q!+`m|=7h$yWSVwd;S(Y`DSqpbHNn?#WmpJoVtzu8-tCLhbTK{DyIoFFp{CHNuRp_xzV*q}lal?+Ne4WbNQyM5Q6R^bHeV>=s%mCYwsVhAPOMrEi;wWA6_!Z>*r7)=%Zm{pKOGTNSlwHX7r8$9g<;2?>W!vkc -NkT~{XD9y1P^KgDJ8()E9$Um$$~9&M{{U!k8V(Z=|I>EL&&TNLq^b58(8K(NX2YU-KSl!v;?tnv4_*` -qm>!4*6X4-*weaJ{A_}%G$6Bly{6%ol8{5g8-Mrk!gnTN7$)c{dDL#z^)gg|d})j(&nFxuA2Zqv#Vkd?Y=!fpP1

    i!2%O9K -QH000080LuVbTlPUWvYY_`0HFf_04M+e0B~t=FJEbHbY*gGVQepCX>)XPX<~JBX>V?GFKKRbbYX04Wn -?aJd3933irX*{z3VFmpR%y^2jo!NZo_sdET!zFCm~zoWQZ)OBYDBlkMG#3?Iff+*?Mo@ym_OJLl0VaL -mn8l2I^zr495GAO{U+?TGq*!b68QB|GTS}R!8D@>TF$tI2&y66a -&0jAV_Y``}#+q!Fx-X(}Rimi8HnLYB=Z;h}kh2>7`>Hiv9Q-?R@ICtc;!`%xAuFX?XHtUY#$LAE-Shr -X+E!Kd6C-01xg-GsEea4#dp1X`)l7UlQ34w}-f!E-%%rhNEc@}=YRI9`hvtD+iFW -01mrl!BXgwTC7Ix54cR+p{hti!oi|b8gq^M+r;)1yD-^1QY-O00;of09jjIsG<(11^@u#82|tz -0001RX>c!JX>N37a&BR4FJx(RbaH88b#!TOZgVelWNCABE^v9(SX*z~HWYsMuOL(mumM+FAKDEDGHmO -DF4$HSYk^@X3W1i69X1lBl2lS7%YWbDRiq@lZMqIvUs9Xr#_xRRLhnS|kfPZ2suxfcgf(3&6rr5AiYn -G}xmdXG>UO_pe81Rfm9njCD2+f)x24y+P$1~G1}mfv9V+;$)byWbr_9{Q=FV6Wj6D@Cmt2w2LNhR2}x3O%g1h4($7EfrYevTyOEWrB`I-?%0!KI>gZ2&kIq7`#$4 -JcYsMMJRY`a`bppq9Dpx1DH9kintoIR{bPFm+i-1WkWv@SN@e@!zG`Z59Oo-3LE+FA$4`&L#+Q=F^(* -LH%aFTNVYU4U%ECSS-q#Vw5l2UX;n@mUG~)Y-CxM{Q;c-i5?}PL>g@pCn{eU$yb*^P_U!IKux5Aj_g` -N)Fo;Y4p2!FaCQfzqvaI>BZ+Fs7L1`iAYxQOmHP;+gC7v*4um?GPZjKl#Yl&Kg&4DHDWTr2$iwf^xgd -X&=Pd_&4QJ@0!fUYI0bMgm1ib5+01YPTIeXVbeIVNdVR+M;@vB_B_Sy&?`X<^#j*7V!2!{f&l*H6bR>Ck{(?Mt@=cEAjbN2#mS}(~OxVaEP*F2%Lt7|mj}?Doi(Xf@G}uht;L$ -Y*U64$Ttc3{NRVt?RFj^Cioukw+hd79aTdZTpH>(D>&l7Wota?)IqfQJ6yfR{KYLzL6gMAv(8sF#RGALWVs~Ij+ -e7`0QlLr6DVMt|e!;aaoOf1a@pNOQS(@+fFHxpx)UV<_Sg%^~|%w*W*HikG`$L1!k+}%#QaYB~G?!#) -@$bAnB(-*j`{5T3uF&*~Dv<HHnYe_lhd0#3Hyk*!DEz`czRJDJb0+A3yr`>&K72^UV^{wt -uX&j+K?&@ADffIG#=0ME1MGWM-ZrlhZxQtlnSAJX_C~E^4n&beRvvV&4JAA5LemB$XYMdd$;wn5<1nW -m6;g7V^u+PFI_j4~vAt-BA*^S|7PeW-aG$TgI0S?UvX@a7B@7AH)ek`L*NOiS{VAtTr^APsov5DwVR{1cafW7=)GsxW-MppzK;S*T^=G -!h=oRCj()R!v58wy(1=I9$csH0SF9CZoJ~k0-Y)8`mDlz}^e2{wew)_S4%Mo!i4-oxADZ`%N5RM!Dk- -6m2PI;i$?tw3&VJf7>mXACx@AviSr95#mBFB(bPt=nodFn0lai)9=@LHcW@CTQlHO4=6FTU5>+a*Hp-QZVzQ?_Hinhy+&ZUwwUB$JmlsbNeZcH;U+NekmFUayI@V6^~* -(!!)|RWYUp$*==Yhce>v*Uban#uXe062Z8=+15c+*lg2V=uZfk@JOZDcXdF1H+{0FX+nz(JA -$kXP+gV-)jx2G?Tt0hw!;LIJ|ZL`=Z?N?xN(LSUT^I^gmEb0|XQR000O8%K%whdaIqnIsyOyR0RM4Bm -e*aaA|NaUukZ1WpZv|Y%gSKb98cPVs&(BZ*FrhcW7m0Y%XwlbyLBP(=ZUd=PO1%rBcZUR06>+1y-OHf -|XXBEGM3{Rvp{ePFF#EJY##4CZ)|O9?yGzGjFDKrUS;>KJ*S51KC;Q0&uNO;6WPgi=vG=VX#u~$NQ&W --*@c(%kKW`!{g`I3@dHLPN;=(mF3Nm?+H-#K`KAO8m^6&4Oj9HGR-@jL!}PrG*^|7J~&zT>G`VW9!LC -#nr?U#KF%_(F?1NX;DL|0vH<$-h-r@8r06H9&<{Rf_k2K9Ik?BbUgD971z=qivRzicO@J#R>Ru|rP@h -H>cQLTjC+UptP=^QY16K-NQ522h-a}F_msnn_c%7eItyVwKI`l;4fkV;@bG=oO5}yvzgD~iUOa+Z7$l -n17YPm}$1NRy3qeW07(Ku{^8lczFOBB`kX6Afvow4Y`kWK!6dXWT%I5U*OOg%j8O1w!)xy{Goc -ut_a+sD`)<>2g7B+^oXiJvA{JUCl*K|718d5MpT+C0EMOVz5ZVEDwhbtbSsM&X7w1OuH~l^e!q8=0z* -2`ZjCaGNO9U?sA7T04H?hYrL5Oa}{qd_o2L{Z++?I^_?kGX!HN-mkWvO^?3|(ow%Tx=AxXQjrh%Unzo -G=@O7=Ww>Rz*C%sXrv0JE(QRnOCfX$KRB!Sk?@6$jo$J$c+i+@l{0|XQR000O8%K%whaD#$%9{>OV9R -L6T9{>OVaA|NaUukZ1WpZv|Y%ghUWMz0SUtei%X>?y-E^v8MQd08FOG&Lz$jmEAElNx-$;{7FNX}15R -Z>#o0sv4;0|XQR000O8%K%wh8&`BAc@qEtIX?gZAOHXWaA|NaUukZ1WpZv|Y%ghUWMz0SV{dG1Wn*-2 -axQRr#N=MgXd!@60%rRq)L#M)hzq%dtLw#ASv5z_ieg -dEP=p-2haNp7zDxWVk5;iEh@3=nq94=sMn(16iRICyep-EZ*48)9X)$rR5vf{vXog{H*v<_5>8g~G_|UcZF*`KD!Osv_g~oqHiB(sWxi$~88EPen`_iaqd@bAeG!`aF1@f9CAe%m6s#2x1E~BP%45uj}^9vv;6St*q -Mcwb8ga10%wMG!$);gi4m{2y7xvO;84jypgdNNiJ2^Q@p1&epI)kGg)XZ93RBIUj#eG?nDp+_oqM6g# -i=YpGHPxrO+3pyCT_;V1HQz>kXCv@CM->;#|r5;vfn!765)6)IjBxTgss5qwwORQ0|J=Ccuuq#V%S -Na{=ruRqz9U(2b`6VjlSckL;9!65PjbpK<4GytvL0`taRWZ`&}Bhq;eUKb?d?dvaJ8No-H)TS&VyYY -a4sz(bbp2~_QfVUEP3n3fYXtH%aHfSgb&~FfNF&1%Sg~6$obru0_+f5{3$L0`;b(3Ciaoy63_+t2>Kv -Y_!w3S_LnsHmOryqk&+wM>w@UIF!IdWoUgSe~p1StypXpWJJA$XvC2}%OY&5H&?7PY)+Iyvu!%<=hTi -!`_+mTGR~22?j;hW!$`F+8=h^JNBZ9GG8e^F*At!hxDAzIuJZ;-lS%D?VB_E-+639;pIDLrNSPDS;QD -{R^J*A{Os@c`3p3DsTyES7b%qsq!Go;zlAKI+VdP^b)-;8j@$jgKXLs0~Q!Pqf73O6QLfYw>5}S4tA} -q(;Q>w)W9m!<{AVRBBz*zd_u-O*~}<`5i1Ga5u&7RQOVp&_YnavS|id1!PG&KMf&qte3`W!$jd>1uLQ -Z+YGSi}(byqk!ikI_?9>sZScCe3&?tuYhxlQiwM~KQWEJ*OT$9F42!b`5BnCZ=+(XtF{Ok)g&`&(OAD -@O88ItBvD2oA~V}y?xL@U!v{a5P=2a|CwSKW0Od?ni!nY)7!R6JEdG#Tg8=F@769?{~0+u9%6XD&bbJ -_i+O*9HrLrXv2aE-K%_;pcKRLQ8*{yZR+d(K-HA{Iq-?zc>cL`dyy)d}`vAH!}ApJQ3f60@T&BR_xP? -oXrZ7tsI1x4F!bBSxV6|=Dg?;2J255;Zw*DWR`XyRLG?oV*P7bL4u8XAn$gV=K)LGP1+6sw0lrJ0l6) -xwD{fcFmnOZ4Ysr_!F0iYR4o7+VFm&UxE`+G$iuT+ItA?QE!SB9`0-Z!=f!gYO5QfjPA5*vuEZc$AEkrhy64cpbjGjs~K+opFXAo9!SbO4Z4G{ -RMVLhvH0-2nLvA_EYrMdTnp*OeH^<&bD;Y3l%>eElhC`Tzuwr`Q%hLu}Ec>vfT#@g_H#r5H5qF)yNXk -Ya991wI#~w%GO<;K6_xdzUtKRUQV*>nf9I -cSXUP=5J`7Ql&JS;i5E`{qt*=8B=u2P`t*PyEA5Kc6phbWZr&xMN&R+bi9aL!&OIR3zY!@f$ey$swOIJkl} -8qlS`MeotT+(0IPnZXuGoItFaZo@zi2R(4iYI$(FywWzIem*~1WFghZya-EU8lj{06%?F>cvt1}PEVb -vreIAOj5W@U$;c;I=)_Z?Kk&;UB#BF!J99RbKiMrQc3X&g^&>c2+HFO#jo*$-e^g33k+mA#6b@1*MeB-(FB@6^9)(BMu4Bf^!W~gLC4}3FNRlH+V2|(Pr+7Z2A_nWHTU^WouwrmIwsGG8-zHW?M-ZpOO8ZwJ4y|#x%nVU5P3@P0h -1gTPF5;)^KeOp1w@8>hoO&xb<&#h_BWh4=-Tzkp92S4yp}Tes03jSERdmDXT+7vyge=rtsodq%s4@_B -e0p7MKLRj5cZsTA`dv;e?(&~ZXwTPLL=xUl?cp6OY_BZ8$$Lqk- -8+JnOOF|W6F!s1~HhpwDxqL^Pf!=+Rg4+R#(aQB{%+SDD;b|rPP8;i38gzaR}aJ|wq9)d&hC+&@49FT ->aR78R$?@}@blQw7pzS5-fqh6M$EDI+NHqzeDnxYbhO(3r*A8xDzOWaTgevTa7WLiL>j~Dd}l)S$yF! -$|w|A2#z8!2}j18^x76-2!K_VpXG>M6M*O5h;lX5C^+XzTtA(sD?zSA9}!sX7tCdVSZ2yMU0KwF$=1j -a-}ZxF5k9fLzD-pBYtmnFaGK&+J5FT-e#?Mb_FDvR-zDiNobpf36J>2SN_=J#Biej|bm$nH`B@0JeC? -VhnMD2AGhi6jud^e9%YE)=*AbsSVJFN|i7ZQx^Ok5(*;+7yTpt!I;rv -ysi{$!_vEukGTk0atzq(sAJl5B6i3()>c~YixZ%lh5}}P72jg>Vbt5Q49n?7+Nwj9Y(`LcqI_LSQ^9y -oHk&f9?h0PX5+;;o;Ayz9E_EOFzq!k?(6PM3DJ^3RHip#E -DQh1>K#1a0bs1Y0T&S4T>8%NO$xBhduJ>LG#BU=g2uafyh0@iD9SOORZg?p~FMHL}MO}pB$+P^fL`IO -Y^0Fgi(bB)0cmvNpyQd_q#Rvtli@CiSJONI>m|s4>GDX9ewqWSqc_9~( -*CaNk4vfXXZiiw0UarW$Ys=qXut6S9UAnO`d_U}jbS$^~w5fnDgWCWvxde!NVqFxp?7o+6f-io3etW(c=?>!#45hG|qMrbnIdzlq*7eVO -V_qg}%-*=5u!Mo4j?V6V`>zKCEx$ERoyL;Ny#(w8S9&2y~P7bU8qg3w|^#uKQj16Qp00}O0v5yi330} -;U1fw3;IxUM9%*j?|2U{DCL+{_u9BQeS>)`$0>AcG~ynMju@*Vqw*tmYBcvD3 -O##$CcgN~3kMK?RC_4<*xTOha(U)8S@hMd2{=IQw8=R+ly{(eblHzmn1@r*F+A+0Ji~qvWLYRGS-r(J -7AFIJeY@rwD#YlqYlbEShT4vl0R_E;IyFqcNy-#3_cn~29wPMD)V!j`DTkf7x@`^%&zKt9J#qhFqD_m -dxLs}UOgz#WoNISwYw3;^nCw+>jY9Vk0w35)cT2cdmU0{2?Me)7#*}Sv&e8Y2(el89-+CyN&8N!tNU$ -02`YCkb@$gXC1EN+>uMO@R1%#20k_OajtYzcz*>r9tMmP(HD>8V&24b@cOx}>ma -Jbq6@hu;g3`MG}Op}%IY9UVBwoXfEb`%=JVe<#{xJnY~JORRm|KiBN;y?x>5{W?A`%pE-Q0t1MaTb_IY6w^sJY$V&CdQ`CH^YoXura8`kV2l --V=RkKvvY>Vq&{UIl$!<7*H=sKUp1X-w2!FYtck-i&dwRE(2jL`SGMhI(*>#EjQK2A=m`h?4Sd^*qW? ->Ah)1bfz*i$2n2pBCA8@*utDVC6;=FSess;xj|h82nU|CjA08U^R2hQOL8I6vIEH@8l-009ZudArZJU -4of9v;S<-af=FK+J9;&MRGI2Rm2>35g{(ml$fL|vUr>7@KG-{dy^*5J6kG*d$u+VLFDdJSvli!h&jV- -S(o)YE1a#d!2TJB}Hy)~r4}IBxjgA~|*?;w?7?v%E+?`7J$n0lEYjn%OYLSoLAi|9X5BsAo-5>X)nf~ -V5u0U|GgCO8wjY?07zLHF}G<*kU6rJKSPxDAdi0i4Z*_+eftzfjdHtrzA -@Ivz6uQ$c1zji8a3o5Rri7yUbl4pd)8-8ZhrmI-glj|Qg>y_^~xFQqOV-PBJ( -fIYs%g-Vpn@ER4^!t0_Mua8Pbm{mqAGJ})c4Vf@)TR65K3j>_6-a{Cb*whMAfky$Y>*VX(0h3 -Em`cmaxt`Q3-v0>~vHtrgCQeWMpYYJJb`z{|oxjP3)awc>^!XdG_vX#SSFiu^-B;tjKXGt;t9u>%UVP -+NPCo_Xr+A5xJh*vfInFj<7{jrAL}SIDkKkch(CzVX8!0wjZZqtyyL$9VKH>|8G-3Y9dW2BTF&<$35E -DlX@Q}B%QT%4YT6}c*_i5oT!SIZ>I3835|EhFw#JD)D)1G3KiX96j__oc6*7Ku*1uYme!gb{T1yD-^1 -QY-O00;of09ji+J1c!JX>N37a&BR4FKKRMWq2=eVPk7yXJubzX>Md?axQRr -dv9Y;=F>o~4h+1!=a1w-Ue!kPjY0F9j^(m -QV9?Xk)6>(hX-p=QAM0ILR;o=JwJAEaO1e6Ear5HoEGes%dZ*gSWO8)0ZtBY<&)2(d*Qh*C%FAutbV* -TFbysv{UA0F?a!pmcOZ|PltCn3|Z(4n!>{R1j*wp9eWp%E9Hj2Ai)|-u565y;jTk8Iv!sEr+MjEY(u2 -^o0wgv2^fw{yjx|=Nk*EiFn1U{$#_2MLd{N2gpzde2a?Tjw{u*IH=4gdP7f*-S^sf>DCZnOMKRjazmw -#D*oaqfU@w`y6gOVzX^jkB&eHyEB4=VMJ_GB&2|^SZe#Hs!Bs^}M)Lvt(IRbyY6uQeMHOF-TW{6b!_W -uDXKfR)6QKvn2nzXqEiB1(Traar>v(M}9NxiG&jtt`G3#B$s%8f!A=t*mLTkJ&fU7WuNPQ=rha97=E#GaMa+sFFiOvbi>i){ -?ZWxZ0HHiOw%mB4Vyl9+};>*2A!IRapf)7$0Cy6T#GqoF*ZtB?8O$kT|Pi~mT|Ru#?iB5#-V)*$tjWQr)Ir4nNV8Wy=e-e495+w^n9r`1xIh5S-^<{`BI-4?musd}9gn5_W~PFLCF -zTpj70t9PBMT31e-0*$}j70v4CD9>RoVZ&J@uO{Bgn9Pz1N)ofsepKyl)8XY8G&1!B>DEkM9~~WSXkr -0~ndCGl0h;F^=D$XIfn0|BB>P`=yX{83N~f=934ev5&k_(yuU{XT{$2v6fd1dq4Ls*Gv;Vmwc4jsiYz -2UIr-=6mC+r;b+Eo|ABTM&J)J15s=caao?@%@(J$#QIu`N2s7f_$g4Bf2{W -fm>V#BYKp4c}$~9>NO4rH53p+Ei46OIz&Z;MZ>; -b(=Cl80X`q0PIbR?2$Xp=?wXsq0amuTL&!yZQKqK3W{fOoG4oougk88lY_lnS<1O3KYBJezU+OCsQ(j -ZqN}cV_eS{`Yk!&Ox_~|5|U6Ov%OcU7tgsDGiC$n1)X=B!bgFSGvL&w<_)3CyL>+PBbW3I;(O``)ZdMinZJ;Fenf&LbG60@tgYz8z-<&X -!p~wgdj0;Xb=mCKeQ``k$agMcUjH)mtD)|Ar`xJFDvsH_}V5-(U+GDmKbNXMmQ=i%6S*QbSE$QkHTFZrfM`CC@Y&29z8OO=35QD5gmXV`rU&>TdLgVet)-m0VpLpMoaLjR -0!Feg!!W!vW2rHQS;90t;FdXp9!{D0QJ$S#o*-nn%B;ZmdXie$O5!*8jNxcAuD7!qj3lUWNwdHjoIw?hSRCVkvV|B{S -z$!KEnDN>;;7==+HCy6G^Zld_VLg&w4Oi3_pRdZQ(;^{k6cDG_tT(bGx!{%^p8RV={xI9dRDnpbY`P<@z7@D? -O@(oMp)`J9{k;0Kh*EJ|;acq$deQ`j|<82qAPJ;q7%e}2*;Z0J|(iaV=%X=4s)`M8YZv;5$u4~FQng7Agr|Hc7RX#P)Ms5PyU61?)Cuc2rrCxN|z{l*6CoWXmk(bJ -oz%Y-FP`ZO9)m18h4`FsHb6g5s5Y%GW<7*cE%LU2gDN ->*g)*(`aw_C{8=d}Ee;%J_Z?G9BdI85y6xl++J_z|qnS{iW6UUy$>|u4W+jQDx~+V)cMAzJxl=U5GRHkrvdA4t}Rma}#d#ZOh*1`xe -41nc(9EZ5)#xXzZWutEm5OxXY-~IEF{`f%U{SX3!%Ib1X6prVFt95V7)lqnNPjGtoowg(7kx(RmR%Y- -7U{PFzo(!}f?^OW2g`ritwIz9%O?Ozt)}++BY>m&LMf-#z$zf}UDqQ>3h?`2Ug-8A90)$7-HJ&Dh%Ev$f=zeYOQ>3D{6@NAZj%U;NwWpW9hQk;)eFo-N`Jd -!(MfJut{y?+&-x8A5-y5NgPijnTBGfPS^F3?pquTj_J3(M0;K@IVAfY=B9K+oJRXMQB4c? -KE<@hl1JkC4<#g?$`x-U^w9Hh`TebC!#ou+??OI+LHuF6<{T}In -a`Y}B;<(md3B&vj2vys`>F;>7)fS^>=u7uA!CQ8dMwBgPIxa~)FU -jK*En1~NZf4mx9TRnBFC&_TT`Sr%d*#_dJ?rwp>5i7{DijBE4DY%Nj8}xUM-kytMVMKOe5mDKi0`i_Q -Pn`tRM_;b{J&y&~ZYJCf4{;b!pJ?bULiW0f`ndx*e*-O^3wRdx7|&7o&O~ptnzC4TV)A!pn=EO(+M7P -4rJ%6P@HB7_Xtkh_a)>?j3)HZB5w+pkH*VfgMQv&p9Wo@B~^9;T-yEyI0XfCvo)igLBDo5wASoio_U* -nMYvyv7{%2Wm3aV?|3b*FHobA3>#36FxDf(n9;^AbSRQaT@xJa5EU&8L_A68mFW_U;&kLlu?`vlZvsc -K&tq>hR7U8MhfUdC5G+-)C(C~b-5FKQUp3FKNTp*eF=CYbfKOhRtBJHQ9njB2y;{c(4a}#Xu*ieR -Z9i~4;i0dskUrn+NXZXhioshR^n^k4<39=^DmTCt;uKs3E;$0wPN*2GY{;7$*OFZMMKt}?X9XZ^xmi+x0&#_l(?#LPWq`bH@~X~Zqf#w8&R((NbuW?ZO@a; -@>#=4S$`_Dyy6vvnf~XK{uR{4NtqdT#kS{^nBJ|Jc8iH^s#;l!``aLTj>~GN0ARpM?-DPXU%D!t>d7CKylGDd5kg$=sF;UOrR9*i{P?iX@a>u(%~$jJfDQ;KTK~L6iRV{#64vUi&<~x -vu}72JBTqNs3-9Uf;p6Q120!d^Ac!Sf3?RFIA6vnI8Oewuj-ss%VOK^Hnbj&P5GDO}Z?ijw*6>Q#^%uuaN;G4F8kZyG91z;Q8OeT{guf9Pq*{)MTPwDhca;ZR%U- -@aQNF`TgegtOyFFTwu92u%R#Z^&m@EnUDe8<3`2M7~Tme|gNhgll-mZ!uz5Bs?CQ7GUcgI -#CA@f%nK@j7t=u_@w98F!wBGuB3Bqit;b9CO#@fd>VQ+hYissAb*HcN>~SA?#zK9@u -UDcCt6PSrs2KP0XZh1^HoMbJIXlwwWl~X9|Y9e?Z8E#AuoX3Czdrn%$E8?{uyNqeQ!k^g@2Tm(Ao%!= -uXPZ3$!y9g#_g5Z*ACzwuCDgYk_$CO6=}dNW0OFa$@^#T=e(4@o`YE{6U#>3SC2C{~aS6`a3{$(}Gl$ -R&R2`9;l0e?Y$9}lc4>pG8BCLL2D|RKoBCa6zJ6`xdXYnB|%(u`kcErTO`{g%`ii$<_ -s{f#8_w|(X- -EJV^s3Yh5;AN5q$=n-&#N!lV`k*Mn!ps9Fexg1g>CP%5AEyVrI&>(6f)UEyJyUM_)m!7=M!mPj=uHJ# -1XW)8H{uD}o&d2c15*a3tNT0pz38_F1lv5MMGp3SXn4F&l!sjENg#%vEbpb#?l&Y@{CGH_ONGpmS577 -UphhY-^zu5m{UGvW;rU@4Ou5V0 -Y%jL9)ou;`H6zXwA9OLjR!8w_U>##fTAi^Ss+;ar8KtCLZioHZ=d1a?1Z0zNA-Ah#8j&f6);qRGBzPM -`I(}VAAuj5!H^H09a`vHsaEKY=2V7mx1~}Ee0*@GJ;^Ix82+Gyit<1FnreEKHb{ -$N)8x00xgOC*ToH@4=CXXdC;^HVewr82=oZBdbqqOs&f}>c4?6Ucnve!)mwm}o$ns1D2=X4wA}mh<{s -M~7>1X};hChqu4V`om@HW}09j%R9c8|BP3lNw%ZNFA)8RvM3a3j-TTYjnG*j`L`EFnv{dXQl+c2(1WU -LSNqoe5+LCiqpM?q$Nmg0qHvhA_v6rUD#BfG~r<|1zaxcWHO`5H{RU)U>u7 -v6EQ1ghhO9LdZLR5(Q3jIUC&;#x^W@~Xw0<1MD+no<`V^4mToS0bQCt^1*$9_r!iE7XFZbW= -=Em12HIxs^O0v|AQ)?AZNF?C5ZuHyfFEQ)MbL2g^gFE{_vq3WnE+>KfX`J~aGJEvl+8&w%`Y)*6VavJ -@Z6V~q#)|d`PLncul--cGJq|~Nm>v@9NH9Tz`HuD(czEhH43M)BOsb9*pE)<==h}Kramv1to$I&p1A= -~nb!VdZaHz2kg|u~+Z8H_T!6<1VWR6T+9_rlIM7cFDW}>jmI}|ZL}ewp9Q|zP0Xs#{4JhdeM+WV)c1F -1H3aR)`O%ca^!gJvm2qwskpS>|;9TKLQ^W84@NY!Zy&@Q8y^j_nG)AOSV{$Mu2v&Zq|u2GX|I3LVXb2 -es9MW5f7$tC9t{*EExlcsU~q58vg3u{Ql(Opm8>p?QfKIrpSh2uVug9PRjucTmtw%ZM4Y)@bUNyN!&% -=*oG*`pCNzdyJb(>?H78NM~?*yf69N8tuatjDz(=u^xP&LXjxV!DH?+u@1Es!OBU4X(=OB+{e6&-rvd -bn_tj)Pum=!4y%0dq{EFAVS^EpUx03Uq!TC`qz83^l#gx8!bcugqp<(A9d-Go*2X&Kbn0%7BAat;3HC -wRkq2W16rcFsy;)CP|Qvw!}4BnA$jD-;zKR%e--B*t6q<2pDFXyi4~f=>b3=hMk##fbd -6~oWW3eNSDXj#pR|szKJ~k#!OC}z(e*K5op-m6peZIC{I9Q+#HXr_N=bRc!mtbz~3H&e~1m~$whC((?4o?QI)^!l;pA16hMRT1Ve+Ip+>4pe{Asz@ -q^PG3DMLbesDhcV7*51@CX7W^PAF%^Pp&=Y4)3|eE4M%=YB!Nq2Q>Q94=HikEYlRJ^YdDig4K -GXU)`LG=Y=&dm6kVn?$?8YHYe<)jq@jDz27sgm~-Sh4MgXhWhs3dqK;$r<;E$)i$V(q(!%TV=zGVq9o -GOa3VPpS{AxIAEx-Bz4}A3yyvIh2*aHL!7;7`#~y(I#Fb%47m+F_#*-A*sPT6u(x;sGi1-{mr|bDWG3kIPEnFYQH$KfL#4qzFWh58>xuXOd+A9ZJ#lA9K^ofD0(;fXYYApHyLFyR2w+d%%K^D -Q!Jrnw?|#WpyUG^+&)G0h7v8VfKv^ISj4yP}2fJo-cqbm|?DIoCz=d}=7C46dL%-4Z+_jWtur0Saff{7RX4j%8{-Oqx$M&aR0uc-d&D;xeI -djeQ{he$0ph=#C&F#z8n^Ixj38$2L`l$gwxO$cVvA)cgaB-YKpxg*1SE|7>B_Q1OiBYmU!AeY0+^Mv_ -#K|Fltm@?D%V!Q8dcItex;Ra!n5xu9p^`xy=q0q0$BzVx?1C9)v0xO;0ka6CVOtQfOCym{p1lz? -pb-7i@4SJ-&qpkKvrF?P5C)?4<>}2dm=({Pgn9Y9rVvBuPVjJs`u*vrhcGyIUz{93L!IP&Hk=2C(k{2 -i63sTf&I|&C(oeYd7x&hj8*>dc?B^7LR@vYSHB%=dwytA0ZJE#}bIhx(zjCoqZoyO*8EqY@b0Dj0=!r -VwZ1_T}+)NkLG+pVcdw@>C19=X&Bemj`53A$U>^O@BRD612*yt60 -c|o~HcC5JWZ&%N5fd3A1)yma4wk^wZ?H*Sc-H@J>sdxT1niHN*p0MI{jU1Dqle+0GRFNL7XHtee%CJK -@0fBK(VfOs%)mwTQOlp0`R9;`RaLmi1Xy!s^F?~Two8D1j>U~&d)vq{`FbsG8=XQuqa&|aW>*f8Vsg$ -bF&CVr%!w*smcu5dHsvKOUvdqucI4A{oh>d70IJ1aa{NNay=VH!H^WsGyMSmM#@XY{rxYZugb?9!q`R -iHl#0~q!K+!ZT2bIKS~P7>8A^97`oMJ@R{v<2{s;x)kIPEYRQim1a(Dv7kr9N^vaec_%^mL<%QPz~eA -x_xBn}Jj1%Fp`NX;!GhXEK&B)cm4P#Zd`B?~KnXJCnLMfvADS -csGhbdzlA%E<`Fk6*y$CDKhhEb(}=)NbLCG7H~G-kd$L-UpkitY$(BB1DQrPw35v%_)HnS-!3f1 -(YNq8_+Y}KfX3c`1niW3*wzIwv9D>k#+QyA^kQtK_3Qf3NN^!Jc$ -VlnGjHn6mKQ}?WrTbwdu?#AiO7=$V1mNXROwO3jYo;XIT&b6Toh?hn(P#3|BA@r04Xa7!y&u*99V3bm -g8MXVIllJs6hDi^ZLew_81%4Q&5&psOO@*Qe5*V%4lfg@@`zVClyj}^#qow)Mv}SIr;jhZ>J&={cnGf2u*m3Srb;Ts$x@tIlKk}JvmS`S68MB!Ngfip-$xeawS@E<*%@!?2lbN?M&0>`BqJOY+qh*Nko -{FUA_dESmt40m0iE<=x5lO|y{K&|6yx^Db3ae$cUghQVrkzeb(eH{*y%j&(@DOY$sA0rf*hvt4Qt3j% -U65)aJ1(B?jN(2;?bChtn}3GQCS-D@>ortJ1;N;;Z9WQy_X|0plsmGhkI_p;uGJV?lyZyZaB4uLmvv< -GN;FXw9RJ8ZhPJoE5*6&c6}f0-92q?c>T}|&(Mmr=a`{WJJ6GNTXccp#hGK7I>?Yj5PW7f%fUdfr^n; -i!J6Vah4u+Nw$~=#lr6tQ5t=`*JFb-uEQ;$^e36ehE%>Ra_}s6>;Vxf^R>D{^@9?Bf5+L)|YTTe8^(J -UanWT6UP)cCa3x$pXK`fm}(F00I$thd0Qr&`zAUTIeJ?_bM0N7c~LmP`8($H!{6z;m>dWU_YXlAjKRQ -NG9xXnv0k{mZOU4q2hpj`1vI0J{YSWrAayBYRl_PP+FyHEGuKSL2Oi9754ffhSievaEY?mg0pDPdfS^ -yNlvQES3i%c4K-$`)(k+9VPp(Qn~EGlUG00MXx|$r-urc2{4y>yrys^&E#b}8-QFpIAAFr5zUaw6-Rm34eU25EuLLOfml74rZV5I -RrYJxu~=bahu0_|R*km1wCI;-9pU^UITz7)`GeG#saqSH|Gp(7V`MWN`r!R9GGK4^uyMoIA -EHr1ahH5P2ekBtzHZnBOjrS#gHt@?IXG^@SN=zT0Q_>v$>a0;Lt!b -%PZTMs(8;lt63EP2>Rv-dtLQ3!~$qk|d{>SGh)qZ4`4+;B`tef7J3*4CUE3|mK*475lt9?J|CQlD$qjXzX&{0zE;gF9+Jtt# -|HkuJTLtd|xlX{LSxP3d7+NLJYi`+Xwf}sskmX$4(PmUKpsEd9SDpD_}Y^k<(B5!k+_$b^oP{{=FOVw -Kh9wlGNBM(3ON^pE=UW5wvjkKKLtGm=CCYl>6V4j}7@716W2Xwk0F+MS`YnWN(w>4!#-4xOln_{b4@j -g*0442V4Yl6lFfpoPX7|S`9&tg7BtnvME=q1Oy7>Uc!7p2}6=zW??ZK;P3ADYpD<#2yp*Li+cG{+CJ& -+nTSy-ZhZcMQ#JlRZoj&OdY6zw3!V;$`^4eLE~JJH2Qw!E~hJG0Ycyt{hcHluBqNfCZN&I*u*S>g96R -=)-kuYNOh}=-OjC)vZD^q9Qp~cE;=H&4s{YM@AR+p?E+tRcRdm#^t%F9EB@WMg?`m%J?|1RCXg4ef*FL@wv>-UmxMB@2Y{*K{5z7{Y@Oz$P -!yO7V)4$}CFt)v;C1Fd{2GSmALAa%wq-N)I-8zT^a*w73$!$R6yI43UMc-+96~6Jt4!2h6#|q%8m4T3 -$rXEHtVQy14ZkOg@uRNh3w1!s0X5%V4<7Lu1Zi%b|CshT!bs8QM|*2T)4`1QY-O00;of09jk#z}3a-2 -LJ%SApig#0001RX>c!JX>N37a&BR4FKKRMWq2=hZ*_8GWpgfYdF>h7Zre8W-Cset7-|=`wr6iHR&2`_ -ARC4@-5vr%Q6$>tLW=@P=d#Pc@0>$P6eZeCnzqA+1w~<#=YBu*TB{vP)Aio&HBVC}c9qhWWo4;sW`!! -vXjGp`wb_Vr(|p~D9k=%tH_d@Eqcu{?m6SXu+=g7i*V$g$LgaP;lVn$U19xh<&aNbPKtLV%HFw`+Qqr -BftVEThH@qyAPO2=w&QP{wZ-s2RPDTv=PRpB2isDTnj6D&Om)VZbx;MXWIhWy0o|UQ;Io$rvQ-54#+V -IpMrF0!2xSLy~)VY~l3vvi&h5%;EiW5qFSyT25J!iFZ&7t73NsYNVC?Fi0vN^0N|xz-D{eSQXe=K6aQ@>N!=Qe=c!-svrhUoC(}7UV_p -Lmp;<;FprD}N$D};ucA@3h9#|VW;A$GTR=Mr;5B0(rGn6a7!Lbp3nG`Y%Gwq&#C#-#YU#hW1>)UoABC -Jmone6eE7B^z6QXFo?S<2HcO6NtfYNoJ+*&(&MVLdOT>e*;cyG=f{OdjVA^--4Q&8>p2Am23@}LQ)Ew -reww$C9zi^3^{~ID5W9BDYN@GLC7q~&lf-o6k!*uKD1W9fru`(UD*H>EaZqX^T>$ZDf1J&nD&SnU|aG -c^=dB}nX#iN!gv55HiNwCTTqobCo@*6+DnS=D+NaA>D!VZ(1>;{vy`g_&X|Qz=rV0Tuc3rMtN~w9Lgz -~ianKWQ{i&B0|A2JukDAeh&=Ud;AVoi90k}u(pf{7Os-Pzl$^au*B_#B|ws!OPjA8F%yS=e&n?c@ZjM -^fjZpJ`QYIw{yJCCWbc->+G!nnY0p;2s+ixE)fD0RV_%ynlTpn~Q@jT&z>4V3IP?5nlm@FS*QK!)Gty -t2{GKI9h_;;0~D>pv#QP?%r -l|O^BMb|Tc}C(R3~ss#+C`O5VIoYg{a|8H(6-15M^i1@Scg)E#L8Qw2xq2;5T1AVYT58AQY2%f-uWrM -?J_i5Xk3fyvRIY+^_Q=51%rl<8XoVLSW@ayre(C1*+Z)E(=Dx+~Gi@LPI^`J!bWH2p4Np!@ym|q;941 -Rq{+hWMk7jQX!*-Q0X~PfX-t__Nig~(*k>%~{GuU{U*;kc>ay!FJpaJO$$0Jw -E7C;-|5{Kck!lN*&h#ZzRRpBa1L;wP6@D0Qjfz+(lLK~a7V*Or47`8UQr5&rd=UW`r%6ShuzH5B2dWn -v4LVvWXNRJ^);`H?V^(3E?#r0qj&hnmgG92q2`}XKzx$G&vM=}p~jmOK^HCB_hulY_{{>Z$2XTAn``k -ze}(%`0fr&_S!+3J(#R4558LEVFYO$Nra>88y;orH(;=oEwhAzQ++4(E0PN1~TM@XuCc*jwc+l{$A5wMszwdmksQ&~JTg%%%v0m+k -EK;|nI`>H!+;HmY>_RWv-&P*s#s^eGepdl9F60Y(%Wc@?Tc=A}zioGKBkLL|YxkApxLVdOnst5#sE_K -=wcP>hV+aBo`|+-MUXENrNm7O${SvoYO&Uv07r3LH;)@v`N#c;SCLg{Zwur%AN}#pPA=~BQ -~YdyjZYt2m$1amo))sG61t4dN0xO9KQH000080LuVbTb}pZ$L< -gS0BJV>03!eZ0B~t=FJEbHbY*gGVQepKZ)0I}X>V?GFJE72ZfSI1UoLQY&07C++qM<|U4I2ew>N2%sW -{2f##d!?61VZqNqup$w3qoPN`xfV6rmBM9j)vCzV`(HL5h-{uDdO_)iww`y!Y^VUx9g{Rw9b#Raq4>i -bT3vtD+Qfo~tr0QdEGSi`+E#*qu$(Ju)oxVPe{(Ac3?bCnZnYUS^-aX>4&p7FTekAxqEs -d*M*#s_;!sQlF9C6H)&B;aVGN{yQ--290ZK2Jds6Ie>Y-=c3#xdOE`_SleFIx@bCTQ>D%bV2}m8iJbO -E&v4t$7Rs54GqE(v1e{W==f#>?8E@lz%Jx@RJSgIqXKSlE*UddZkT=S?-OBt=>*>${-nujV4{NnhqH| -X`ED9$nf$5?!%uR)WUfY$`L<2239j_pw->5p|V?tflqaS0e|3;{%2LyU>Wh!fe4HpH`oPCyvA>h*evW -m$lP2Gk+YI;XLI6y>0gOAnKaZ=d93rgo>~(FZ0?1u#$|d}G6dpbJW5{c&(vBu6=pmg}r7E -0G^aD^dx&7^014aj0Y18MF(9C|b+H2o;GO3Je07T!tN1{i8Se#WL_SJn@jugS -BT6OW8Q&;XZdjQ_RZOw?|Kl^Wg3p>(Ku79Sh$guL9fjuD?NOh-xqNqWejafMzvKsTD -tB=<223JHwewxX3DgG)qr$wQPny@j`+M}3(jt$2m_yR{LNOT-QA(IHq03u5XK`5~cZItSBXgUUU%*v1 -8B8M0+u3E=ci^tR3f&lh{(HC`483kx?=I>gpnv6ZgLc(t|*>9~Lig_^l*^6Soj37X{onYFN$FK*2x@( -^;^;NkMzjJFtZpf7FPEO@!X}y40#ngw^e8ImB@F<*iObUP@`!6mb4QDZY-+4lK8rx>_7MOk&Q?d$+wb -Nm0(7;Iqy9WGwsaBHR2x`YL5;lVi23wVbZ@G;(;*6r9215c9EG6JS7G~2!t2K_HznKvk)xxyCJF{z)s -6E?$MCm>U)NLcYz8W}M7*1)ke&8t(PtxhnAYgEx(@W$ujdPRXb)4Tz`Yz%m0lnWU{3Ye=z@SBkznw+{ -;0D8PDOW*x{e^lUl!T2 -DA+oR1OOVdi+LUvY2m<%Rma#za^QB%{OFr?i=X2{`vBRCJJir5B4%#mRZM!{cWf4(MKEATfv?WN6*VO -Ktd*(y24QWrd?q0#pz^#=nxEu5zya{5mhVX6YzXush=YZJ%0p32c-&V!$o#v@#YP+(#1wCuNVG5E!tM -Z+5)t-Hx|QMi{?Yw+~fkZ8)?7{9br9Lnor1l{jE#`FwMkUo1E8?ukl+NLm$jzzGe1DHk38%Ke(;7>#+ -MJxa(*9HK1Y-3P_1-x07Yl~kE*VhWQ+J}I}}XLCCU)pw{ba{9mNs3bld+U~}|(?>^74gbtGeFwO+t -x##7fndcfc7R8?n_>rh1h`qHD+r?%IBlB91tF>rpu2+XpX4AJG28+fJU_<;I`;o!ojDx?k;WHyy7d=U828pf_|b1wxj -?kT-TPRz{MX6nHmk#r!&ZP~v$EI;gpDA{Z#g~OYh^%f6B8yvtHod5auX=X^1ppR=KY<7B6TgEY>1-&t -EW>?K=y^E(>T97Wf!5)a`dt4 -uw{KfU4jy|B2-Lu)T3=Bg?%ZC1#m)85@Kjbyd(dpQcgGL+E%Uh^xe}M@MM?^svQ&=+kadO11zRAQR$Y -}~zdX!C~_FcZic`*@AlvCQLW6l3##S}icM^ -2ph+k4I@(^j@IgPW3QY1fq(ljW3ptQ?C?3^!1 -Ly`j9g}6Fs%*JiWFQy!7-`M*Df5P?dgkx(i;Q;Ef0(N(BFn{lujO>L_XM)q@oQUP`AkM@(=7U)E?17K -VZmr#oPrnkIOd5BRRH*a_BVBlza?M%-?X(1F$wC|Qq4*$w4<&sjOOM -zKd$9!q!M3+c3FHgYF1Hv2Z!iaB!a?~^d-}MIsBO{q;|phO9ZdV$w!eLL*MvYL-Zlt*+Z8Tua2R68)UR30^~_Dq5A*E?K6P(fgCKs-TMuhuqq_DiY7d -@q6AUFy?64jX1rm+7`Gp=DXOrQ=16OgC*Ptb6<0Xm@uSjTm6r(hbenf{vOQ?%IR2GF9cW<^uhwX?=75 -j{9(679KsDE2Hh?Ek3O)B(=AU#ocP%&n%i}v+d4Q{nVrV&AR~+zZ)|%+mpI4-1Z>e&AJ;3b$6<(v*3Z -j^eG6qK5k)P=+Uzm6YlI_9UlUK_e`hHjHg8){mZ`z5qemu)qYEJvb|kg;teFW!D2@9f$>SD4|W3( -R|(=bNFYH1`(!;|?BCpYR4IPbmvJL@w0ew_OjJ`hz_=N~u2{9Y4B4EA2zg$S-23U%b5ti1l4Gg7Qv06 -GuA|6hoQl64|YJ{Vi(Wh({=fPvGGdTmBA)`g`wXfYzS7s(rBCc_ep3x7pFFn~$as-p-d0FRcccM+}u< -+kMpEcbAy;lK|8My%?BMJCnPW$VCfgepeQgqc*sV?ce1$w){a?w|9?fTWs8|P?PC_GB3WrJo~}k$3Xl -o(lsjbMqQ&bPS>J;lg2{k8hC|22R0nHN*j6*2v!6CWtZz3yh5w0m{C0gi76hFcb&$!7IXl9Qd^|h4uO -+qlwf-#ogqE1iM^!zJZgIuQe#?Y{0kS^9=Xo`p^=J0Ug;6az#;)1t1r7rwoALsX4DbW<4A1_#4>AqT?*dG_IS}%#X1`A!Z -^+&SUs#xHfS<2_32Ys2UTL*SZ1qck6fXdDl7n{-oxn!x#0nFy7~FM&&+6*b*PvpaW?0<3TPoASJm4WV -=3D`ia(wNGdAU$Iz!EBL%yUF9tpu3caC5uQOgt}HgE6z84kNJ<+6ee1&oX&2Lm;cwVg`1`Txw2OIxH*Xbf*0m`nw~?mS77y;oD7 -Su9S1FfZF2PR((+LI@z&8B*Obc`s@aN**|WLWsH%%5=BivN#Nrhk#k7#O7zn8xqd(Z44k6p2s|0-d0x -ZueO%Fi3NS0l0$r2we%`b-#kfgoj`0B}%heu6XgJDa{lx|^%W?VOB?Y{+#L}}joGT38?7fL(J)uzQnR -^iN1E51+02S4rt1#-qI1m4iK51Ven+%)a>2wnFM&Gq%&L*rthZkc+mQm;kY?0*g#4_!!X*FLXbbMrMT -RNQ;I)BbF0PXAGRxRc=4Jy4^I3Umg;W@8mEkyu@U_d4CP)`s#1`?Y~;ZRx_y?QYcv0bf*bfD1XZ&>hb -S>jy@p1ovRQPraP-bybxhuc>^rD@PM@7(4*x^Lana)_iB2R7as-A^ -OEc8QREj$`)WxaBbc@e0^#J*D6bmR>IL#`J24>D8|+>J(L{t2B!X`|JX9PIPGK>Izlmlyprd-Z22r_}VJ6(<2;WzO!+tlu3nCNOc^S9r^Q@Nt(^WBswt?s9N^L+kLFCn}qpK}6sm(M+ -c-Ov7apU>7_|NHYX_3ORQ=ig9E0|XQR000O8%K%wh@MJ@%-3R~xR~G;PB>(^baA|NaUukZ1WpZv|Y%g -qYV_|e@Z*FrhUu0=>baixTY;!JfdDU2LZ`-;R{_bBvC>SDVj*Mc!-Qd9+F529!Sl0qehjs`8ftDzntt -@IJ701ik-@fPYMK4Z#@33z(f>O2)l_n87ECtA^15cmc -2%`t!WK!A%w`wt<4%CGuGK*{ThQdv8eWfwMt@NHq6q&&URX4V6#8A#}&A1 -S=ejUl0Lj%|HNw}NXpxawX(f5Y^z9&Rjf0Dl@jV$S(ARf%tl#~+Zjg)A%$v>cpwFY?OLl;ZAhZr?kwB -!2LWxqh{8fisHWT*(L6}48nPvWiPA9qm8ubJbMz)f?(r-*nw{K3lM2c})3^cauy-p2+gaP1)pEI&cGq -n(_-ff6TfT&4OKD6e%<^}?{o$2!Jh4__iN`{;mZfjqi5snzUa%i6BDm?Sa -KGoCYs(FU1i`i1!l-ThsZ4w_N8^?4sL&7u<@rdr427$=CrW86K_JfMqh8r<9V$Z)Qo$qCX{|`b_HqK@gO`+L3hBX7_JGYkRdBg -X}xcO{0D@8SOp*J*jxk0Or&=vo}t)aB_ifvG*VC*2kukTbKWB+FFDbLsL%WrC>?Ac9eRLxO~FL7qgR2$7Iur=$2PBIEzBviIlKM?>?!OWSwj5oef|?t()Zs0KN=-GxMF}+AMG*qNn+7c3D~^- -i=&uq0~eJSX5(+e&63i)u)7fova4GfAi*5gOA+8gy~&k3N2gzBOyL_$@O)kt|?O2MYT=tgBsWxE0^{x -=1~m6^GnX`rS@RdtC6)_n<;rZ?@0m<{k@R|>gbK2{|}K1NF+Rl1+Jz8GU?CX9upSKGMma3(#A -q{wRu}OHx<+6|kZc)SyHIO{hxuhqLQ|A2kkBKbgIv%$GysG^JVo652s -li>740Q#{sxP~px0Bu14jicx*6t80#69jvHj%72O6PZ2S?7Yu?xd|Y6_xh!4Tg -BK5_$BG&T`!Ru3Yt&dXlZ{25y(0!XAMw_S~1-qo=U9|D}l~-q~U3M)id!UAqq -Z~naF6;wEU|u!c|THbFypZ>{dzo$NtKM$O -~`#7N4!osg>%4>`Dwhet}THp=~EIsgWfZ70OA)sbh`5`1QIHon^n@0JdjP(o<4n=_sL5i=;3SS4Dp}_ -5fR*sj+%2KCiOGQ*kmU-)b@=t0=+~17x+w39)2zAt`sM|BQesRu2qSjGM~=OYU%vj$%%dh7?y*_qt~g -tN7()P9$mx -biKl@LI$3FoZs5TRCN`fNQT9{TfWw;Uh+i2Tact^uL;E2iG{k(71-hNedbg^}96T09oTM82k--<`u-$ -3~Tmo?wup6R8AF)dzX!3m*pX_qOzUH)bn>IUDXc+7ToUtjRrT|}(79=B^Bu9eet91)UAPc#eAa|>AFT -Q^W4wps$9{n+5Uc14#?1g{+O*1uT@`4j0VU&a!ByrHEmpqYs;6r6);FBwnd;f&9Pm(S2%JQJGw$gL9u -!90VBLg{1gWy7a}d|tWgbAX2nzASc3Q5WiuOYr8J-mhE}`*O@+$UUIFd?2WFy<#%6QzJqns|*Ge>iFN6&o&#?Tn7Reb^?Nah8^n|Yv5&n`rWf0?YX;A@Mk<4t0(YTKfi?zDCm --6;ljLN9%iB=7jZ}%gUsy+Q(RU1Hnxrs7TenWfUltk4kV!`jz~h{uAc$nV&qb~j145=_m~7dm-qYhBW -}(|ptzAu?(Iq)>;R5dVQTpN`SXgsZJ5GvGP_< -q=8|V;Bbjai+w_Zssm-G`s27cQqe@_;t@$?QZIBqZIi#C3ARUCgv|11g71D(x){ClTy{n;YXu)JLFO_ -ZsM?fVmu%0A54Wr$@INU1HIB6=aQd-E9B{iOR^ONJBmzu0jjZUVPCF4^2Io}QW9IELgnFEXId$^R8&% ->pl@b8%?tbFbM53KwJ`ae)h0|XQR000O8%K%whf~Afk;|Tx&Bpv_&B>(^baA|NaUukZ1WpZv|Y%gqYV -_|e@Z*FrhUvqhLV{dL|X=g5Qd9_(>Z`-;R{_bBvs2^f?in4CNJ_N{s+%DbCZq@`%yKD#wnU?67jZ6xn -6vycL-}fAnk|B$jAZ)d3DP8kgYGo!9|5>3Q9;A9`-)hs -oGTZQFrqxnDOeRZ0E@hddQq`hTyhy}zCQ1w1?sKW0cp>xKY$ev>tzIvrvb)CYlM&UWGBqy>`-xvU2D{ -d|C`@83NQ->?_UDVgU!=cW{OdMm4u4FxsU<|8WK}o35bCKFuNtMGY}%+?RB8X*>y<3H>9)#NTs?@i(z -=e>gQ(Lr|DvmOEfs!0iON`oq)D}TPL~y5i;b=xClfNLsMuL+*#xpz_Oof4DlALW*<><##op^$7^W4&x -tBQ3Yqo4uW(}F(@h7#!F&bD-6k9f*(-Cs9ZC1R79<-PprDY|S^7)8)yW@CRY*-0@0ec-K?Bhxp0leCXNxU+d$X-qT0sx5e@{qqR)K -RWN9%-LE=dodIoO@&EZ4%pFPj2Z?Vk3CG24ivqF_r$oXYBaTY@u~=I$`*UqUh~PKr1r+c?tDl*7fPb8I2V#px7J1?q+<`(MxpZlSXoR@ -8TqiHq0?ELt?oVy;I=_4JS@q|0WQ`Vw9UnU}Xg%cxf@dv2Xkr64gQx`yHXYk?78A!aFyDHGw4+_Lk?e-0lg%`1ny3NQZuxD69_d|9-ZV4)lRmRhfjGq3ZtER2`vkvULb#V{*KsY(Xftym(E -P(+Nov~%~75OyT=|jkh-_ozzxL2Kyhj2zicJG7-q+9R<=lDVJ1JbUt4D@Za@}}&R`<2@49^byJ_sVqBd>T6MH`&|3F6Xw+cMD!*{! -lU5Bwtj7o&Q@0b7Yp7-vskXDn!6!;Wa94`F!V$IfjPp$n}^}__ -|*~Rt%r)z4x`wy%WzOOS_LgPh2JpI$ZZMed3i1CJQxdyJMQBoMHN7C%i0$%ERTJAVH&MWvIun)7>di# -hr+RPVLME_1|HZ5bkp_j$Ow7ZFj{}`r{tbJ;T9=(rB1Kk=XwIt8n!`@-roRUGUeG -TIIoLBNZkOR&q)awWy4*_^0-hP;**rQ@3vdN|b#5V0(O!3N74|@0q)d={k&#j#Am@EA+hBr(IX_dKK> -I?272a)Ie7!fN#5Tmn6ga(#trTkIPt!kH|Pz&~Ii_sUL;39RYG#iLBPtm)bG3630Z4JOt?7B_EW=>~3 -d1#7^nPCNiD9qJrJm&{8gBA?q!hZfF?ntYT*ckhiXAm~Ll~Hu8>*L{ygB4lK!*4@7G8YqaMqJW)%MaY -OjvY3)w+q^$Il%!Qr)x&ztkvDN_FF8P_Zx&}$l<^~63cizd*hi(X`(ebf$Wz3>>An$9RO}YOasP**_b -bz{-3~a{jFK-Z2gWBK=rPuU3BCe$vtF2?J+XmGg@D;laro~fsM>ERq-A9*yove*67WOzE%FmS)r=!ho -!fu85f!^7}qUJ=y)~U2K~Ob`=>IqQHrmR$9hNjUhEKz#CQA#1W`>~FDhZ5iW|&>; -21ct*D~+&m^8@J+&QWXX_&DBHPv8Zd&2eLID_d~G!#-A4vx)1v*My_)T(a+-Ht!|x3JsO*(VpB(LON1 -xvR^8Whn{ZYd1ga^ZqRyase0^qCn7m#CACISB#zDkXnafUVtd=be}$NZBOameLe79$8_e$*p9UKXxXx&$wl-T@@Pou)8q@R -G^IUb);-5ds#f^h?HFWGYs#*h==SJ7qOkWY@Dc9|)tPT@`YSQ&6p=#C{Q3Ob=yTk!Xetq|aBssEz6&` -UD733^+#hci2f3<}Wo=q;|270iE6v--^tyf4)i~RBjHVuB2e<-lMQWRHn?c%VHvfn%kv{l}Y{-tG{5^ -h=+k&sTrrmd4?wbf}+n8`?vp~h6>0iGd;tnP(-&^&qGfMLKUEwSCPFI<4KB;=0G^RQ!G{W%&LRLZ4Je -Bh*XKS8a-v(a~_;T0>qUE|D#qLl4^w|~~h;F2cToLMm;r{fEe|~anukuf>TnRdf_HT%WhZ@6aWAK2ms3fSz8W~Ai#J8008z00018V003}la4%nJZggdG -ZeeUMY;R*>bZKvHb1!0Hb7d}Yd7W0edkvUs0UH4OeM%e0Rb**I0|AEjsjnh7I8qV$fdLqwae} -gu-UO0FS*67UBMorR)~eO8Du;s -ce7#Nl}54;4M)@^*%c=*bmoPq5h9iQtF+RZ1N`TMUE!1&Vont}921Y -1ajo)Q{7Xm>#^UXpFg~%PYdhiNQ~dZ8#$aK?Z1fWxiW^%9TdrCs(iM|So@k|PB>B!SJPBiRUQ4#&e*4 -U+#-CyPYDha_S2S#u45`$-*cvW_%##DJ+!{`>>|HxkkI_xDkde -Ow>_weAml`f8()};f3ML38U(I;IEzeF7f*-yYHVW^cuXczIsv@R2fbbL*4aFErbpuF!%&Aa_xP}FYATV**8U;=GFDyFsqO!b~1)a85ga4|yy}Fyh!{55EtJr@|`05fSQD&dy8cE=?q^_?8UQPSlxr6BR>! -4mHTZV6v&wt*1oPYXob8$WUGQasf9F2m108mQ<1QY-O00;of09jk1nmd{v2LJ%T761Su0001RX>c!JX ->N37a&BR4FKuOXVPs)+VJ}}_X>MtBUtcb8d9_$=Z`(Ey{;pp^cpvI?ssJxtHyBN^q;Zs`pWWN~^FmqKo6!T)CYCEc(nN>f@C5SHSr`PatoJ=O)#!xM}Q8F|l5aE_9mXHq>B5gMw$K-0w6(oE~mdxxIcT7s2=L{q}LZR -a7s$fK{$d>NZ)FORM&Q~O$o!n8$>9RoY)A{*4{qXB7J-?X0|K;l4`B{1~yE=V!dUXn5z?fp-AJ;|^Av -w>W$idOAqw;{muo3vkW3mD!od?gsg{O%_^uEJLjQ_twR-B*iU6T$%TDL2zv$ex=z%)=s^BlOiDa4W%* -o2g(rEYfJ<1&=XYg#C_SR|I}4ZQ_tU3eP}{g{pG5~@X+go1SvNi$z4xX+MsNN1LZF>tz?NC4``SF&c0 -k6e0n2zm-BW3%-@P9izbe<3Xl?}^WN^9h{=N;!d4QGf#)-Aa%bCu9sfPo5r49TY;o35>otL9=8Fxl|T -AqUy@DnIYpzvt2Sc5lVty7}=KVb#MeEr~{y@@msc=NhxGZexpUr%<};lCTC*w0Nvl4!#+n5>XPjsGb! -MNrVRx!0Eg{h+_dwtP>IudJ-R0?Mjwxk4l?=N5;?iYgNZ>$3#>`gc{)*yNsvCn6hU>3vs-Gs9r$IkHt -O5nTUvl5TeIxeliDj2$EUU64wqrnZH|VdLJy}}%kgk+Oz!12HW>p1yQa=IT&Ep^(%=YB`6?|L%UN#w` -y~|C^z&kKS`?Pz+9u%8H+|q!gJGNdcAOs+T -Ga2ac*hu@`6xE)j_ktE=vh%TG~3zO>8xF|oprun9+HxrO^bjAU~hA%XDt*?sBV5Xh3-uEv!UH>V~7*YvgX)EQQKmHJszgOHOxNPc{g-t3J?< -nipZ#gnpA5lS)dDtRzfo=6wZ-Fp}G^2qr0D5NdPQyfX5;6A>jy0O-TW{ehumVk^Ey0r%%R#_SqN1VMk -Ta4Obl6A#@BF0?`!As9zrof;ZV5?J-SAcun8fhG9BC+<|gHvW%QG -mJdmS?UKXdav}^=|IzHcI_L_90ol`LxJGVkpkX5MU=R#ei>JyeiNia%is!57}Wz{HnE&}<%W_eCO%q-SbRFq>E -AT32f5cgvmj^YDrn%=!ZRtf=p#dGYWH`(&t_;RtOLkLtgOu{~z<_Pv?A&45vo#ej%KH@JID~wR>ugb;GP_xWmS!lCREDRR5(Z+VF<(8} -izjd%$Zu-Fk;xNf*5zdH)4aO9KQH000080LuVbTk>*lf%*dg0GA2?03QGV0B~t=FJEbHbY*gGVQepLW -prU=VRT_HUutu2ZZ2?njaE@_+c*q<_pcD#huMJb;@*k?Js?<8Yc^CaF&(#gEhtS-u%XGXps(l&DE*jMTH++Z{7yvrRHl-^!vELcy)v@6s9!&8eZDGfvZ@F^z& -WC9IZ08A3T1h8iZg_BtD=DXWqaXbx(y`sBVMl0znPNX!1Sru*UJbZ-q>(;W<(SKVi4VOq)fZcxp68TUs+>;lN;JP7SXbllrHI-qqg7{g3Qw091k1hTSl*S1Bh&<3rLm2f2cIQNR&3 -MsrTC)u0YtMQoia2)nI#I4D;|Bj=y6>pEviFf?!1dJ0ULG&sq{6znl%v7qfmJZh`NxXl -rh++_&NScQ?}i~{ykOc~ReozR-szF=dL$nxPOC|kDDeX_)2DWMv>VU!awy0 -TEqIDm3js!4^^ix0V;cC50jYutOfWK!0a~)*PGV_ -ACktOY>L-&0ji}V1tj||2{Y;7h8-oHTUA5UMu#MqQ1_Lyx24hAn3d6D35&%xNUHRa4OJ+zacob^qI{5 -y@xE>;kx8P|q#45f^E^YCb4S5`KBuWd8goXL@|WZPn9BLWQa1@ba!%AN?e~?L|{~%Rn0P%$bEWcAr)X7NB{W;LkeX4FS*ripZe8ZTSBD-o(I9`ro!`>`x?jU2u=o4&|#d|p4gS4*V9A -eSV2$z`OkgZvT$6%*=FaBt%=cKI;%iuYcpXDEoIJ&4;wh+4ZI8Y(BfGm)`q8m?$mRX1Fafx)J7^)*as -{NH}CWDbd5=L2wg;n-n(Q6CZ$!?pKI?C$1!5F9mn^=8QhUBfQ|5iN9e%#xc-P9{^$4@tIGx!!G0KJ(} -KGp^UPJ}R+tj=i32s(AgQAE{X=<9w8Tz!lwCluqZy0GF&+%IZ!)waRx#6(S9~}Tf!AoC63HdP`;l)m* -j>cMx*js}a!BxUm<&!5?{tZlcvyf);bwv`&F~oa6yvm@ZS}rj{x*Y6SJslYcF?2HzO4e+wTJufbwN)I -46m2W#<+fd@=&oS^o9d<9Gf~@?|)EB0|XQR000O8%K%whwVq`iB^m$#Ok)559smFUaA|NaUukZ1WpZv -|Y%gtPbYWy+bYU-IVRL0JaCy}{YjfL1lHdI+rr=AN^pS9~x5?$0(%oreIj23pPL`9@s&wFx7*dEp0E+ -=la~@61=R;a -MJUgp_rl^IR+miUu31d&tgTJ$*G}FUogvCIt3b+T@J6V -(lcp#3YKM@6{-bvs|qCa?k8=QyMOA-YfkLFP -f5}6xIa;S$lQ%l%B_G`sY`|Kv*H~p@fvRkkm$k%FjW}Kso?M15FVf0B#cRGXG$dArC!ovAmBz432WlMq-4&U;y+M?x;T!|ZeXKRST7XS1&7SQt7+vl+s5(H4Hdh(UwuC8zQAeH1=5xM~#5UvkoPlmeVDmX*4Q5Q -1`8$DcCTy{7SW{=v6>!{@^|gXs+3C#`B45SqCt9%iobUsvKrEa-wV%)97{P%hM1@H7`0JMEjBm -A_j5v_8na`16^Z9oyTiG9=(!l+w8ekC`f$^#W!-W=JjY8`rc&e!QhnW2tGzSgY>yPq3gUyEA(_LcL~44!1bC!q2>>Y18t -L~~W9Rtb_7vqCUh=D?x$C&4F`)Q0Wo1&j8)a|cz!`ViA9UTGpq|_UbL>0KZucOuS -JV{!*`Fh~qDE~#_W@(sF>BMTZyWPjMkrrnG!_KombAQUu%L`(M=tkEy%D79YQi`b3QOoRu@FlWZqA}l -<*pQ8?f|3LesNbSr}iM>8~&A$2iTbKn~ivk^Ylx5B$c?H_G+(k4lD;J18j0pN4yjTl(Arwq74UANmLJ -!qW3Cb%icSQ~pK$xg3w^>f#5p%w5PzMqCy=b~z(nIU>n`Ybs;TGc=su=uAs2X}I{YVt{h6QF -)I3v*6;|f78OtdvHR?7y1No0SMo|_|*M@;&Wx!*d7{K!V?nF; -wmjw62<0Q)&ciH2NnGi)#f`<24ffJ1-65UtCiKzC#WR5ZY>9`_t0Amk5<6|f!p2ZdN-a-`03a41L#K- -|`d4S7!Bxn#vM96&m4&kN)aZ!}f{IyIB!wq9aXOg4$viwq1jI0wK7@N0Pm(j(DU;gsfc&tPzfulpthV -#Lv-%!=s-9J9bqJ}qe;e!f6y88uDPTG}@H9*FNf#=yB2UZeFtN56DY=Gl^=cM7TmfN?Mb=mVqXsu{cJjs>Ng<6_3yEjYEmXz*b!q^dkr& -g%1w?YO@ee(akW=0y2KTAvq!AZXYE^l<;Uun28n19!)U9Yf+nv)%V~&(z|3+1zb{J(4mK_*MkU)uOc= -5kau6)re^^*v;^v5vB)ccRzwMs3d}jYuY%(2*nZ;-vHup4%m)971AmHC2AN?VFBT&w05kW{XTQj9>4Y -DEnFFT%JqbdMdkCmWviF4j@*@Befx-0}hqK>#U3CzIZ9tRzQJP<|$!+D`mpd-g^iUmEp1TyQp7EKTHq -ArpG3~-9VO9s_utSD>S1*iT-S4~sN*^?)0&>PzYX3m}fC-Wy&Rz1n2+zR>Rt1tiX$A1|QEm>eY+Li&i -d`N&@vgA&Z>kG>x$s!9Tvh-+6tOSQMPqh;ZW*45ZG- -WCw%q2to`eYtoh3-epG}El+;;ore{OKr9$$O`^ZgYmc-hY2z+$xGoHyAfgDENfH=p|}$Zz%(f*8*87@ -S%EM01XbB5ty1E|@jaq32s5%xj%6((JrIF2QgBcW(^1JiA6l_+Xv6Cm*a=HvM3nu`p=<2Xify&-L3o7 -*0Pe_6d$aX?Vw^i?Yny4l1E$V?Lh;oGu_T8ckuq>I7=wzJ?&e{t>Kd^3pLoDziFyKcORY48x0d9?UW -w(kGM3xbDtnL0fTIe@$W`8PFTCz!2geDzKV^ymy}?k`F$BtB;)D#jBSH51OxnS-<~q0DQ4gQSHp-lLl ->O2P|? -D1;>AX^OLtY5+w~R7m10e^(>0zjidXmK!htaB^tXf5fWO1*lst1vkh28@60q7Thq&SpB8)Ii|Mge@I^ -dN5{W;PS_4i^)>%S&p%|TdlbHDNfr3M{rU4nBjpkBK4^Hnx@M#}7X3;+)*3`~lF((}mjjr2fG&91igB -e7Lanv6sITJ14fgMi_BouDnJOljPd4x6zuZ{be)fEVwj986dP(MOeoAdI4EG+|K^4U)nolpFU6qh&mGgTWfoZmx_r+^65OK)MB&fMxHmm|P!>{U-oQhK!2&`%~TlL#bD?~57n2hTl -C8ZhCbp;9a_+orM`xK#pt_pIC&isO1knbJLjSm;KUm@%cta7rVMU}xA-0N^c1HR$)J0{^e8!2I_JD*p -4ES1$)!3|Etx7GrcITKPp)7ZvY#4YAES#EEE6FfeaID*OwI=HO!n|Ds8!`QiFE|BqHHRA=2rtw-a*g} -K3k*8_T?vyQd{@lcnkqSKde&X1ozKRJuK3;ODfA~^dIa|-%fFKsM2PU(#umngY2 -G9hcO=VfpR%Ksnm2v1iB{6L-KS#IkZb=bSCrs4omz_+q -MBj{&Ma)|pRJFmB-eRtv!1R2Ub!R=Yq7rG8OODp3awNulGQfMDM()4;TX4_*l0P%oFonUn>CZ~r;b~B -!6J8X^a-H{)F9;?zZa_qvg)sY=+d}v}p74iLM7l8UbW*7`O_O{}UupvlQS36^-=C+^Tzo*p2PRBu*5~ -5-<&FQUFc4h+G_|8QFfN|}D)y<;yFk8r`GP!Wp)~1#bJ%XVAqU(z~dg -B9|Tq3rVRG?KnYGBr4uxF_F!CTX-#S9g7-*BPg;TH%h0325`Wdei}a#uO}5&jkmcFjLJn=SAKbi;D_o -}Y-JeMu{rR;xtREc`n*eEWzW!nKZaTMq@&)<$hRI8$Y+)aW2C?WqAzKROelL{i`js$U;pjAHvge_2-&va?*LJxeoygLKPUA9vS3t34tX}% -1@ZmZzyr#3s6OR*p=caoYlz7enUC4wh!i#w{xTQf2zb#Xi)Rl!^J~8;WG}3^%wi7@2})%+QWdTH2NY+ -MUdMtSsX4kGHIb{_c#5@{!z1-LxWoS(3g0o?;B7u@$tbNz`D9Fcdi(e4x$Tc8+TSAx|6ru)A&Y -7Qhc)G?hB1v>1wcW1U2)9IlOIpcP;+lGY4VFys5)cULaBP3Zr0FV@>RL>;Ne0~*b0mCH}P|B#s!AZ-HI^Z}YmuJ-Dn#=a72<4x4xsX -UmjisbCSz&8MIR+WQ?RZ_S?4D$-Gz#|fb__&iKXdkicNy!+ScEqCB`hdNSOvfBb>}a5qF -u^!sqBo~9KvN?C@V4!U3p@`aie;wV?}<9fA!pvfS@ObRTBC937J27YL!Z*%NgIX0MCB4Ii#-=PEb=Gf1?hREQBdALywk%Pw( -stNH?7``S$CH+wrqiYHEaqK+AQ-ieQZ-o9>TSHuJ5u(kCxR%F(9o7SuW@-5g;YtW3Rnk1=# -C@O(Xs<+Q*~0an=}1{-1|f2HWE+LIjSit^Cbrfub&%Bpew@k2w*_TbIcRz -Kpms3-YOF?|w6F2EtYmtju{>RM!TsO5}V-yP>KpL1Q;I0q@l8Tq&ygJPZ -|M8b%j3dPNXD!&q1lDZ@DSOmRnSF;A$2luoA*7kdx!ARZ=0nL#v9YlVxUw~ -3@V)SY1*1doRHbi-tsi4qEtQF_;-7hpd$&x4_I$i*t#akF^}S(Of?+|W_0<8dD&t9~{j1Edddb8xJ=l -_bH;7M0%{Sc(9&LsRse#_uaEDA?)C(`hEJ%o?k5dcVKEiokaVVs`AMjFhuN1rFc}EnzW*v6!mF*xO;! -hpeT!{eUvdl*n|ti_mh1zjR6YNEM?&L^)N?OvST$?9xE$NUg=v!J~RM@BFt1&|otT-7-^72OJ1(x64N -5+|`vJsMw6S0^%1yd??7waw>% -;8S>zK4huQy3`{bTiLiG@2)!%QF6kPiKC@Azvu!Wk+G!1zoTdtb -q^2qBD>E-p0=(;@Fd^}2hlO}k;XwGaHKvu+RA2w2ZOUt*!Y8JbX4vR()Z{jpPwlqriTDhe2Ait0khc6 -Jx~Ak>d_YK7>4a!4K-3XjnMnNq1?EQ?+(s=#B|W-hcV6XL{SJH)hY(>6$}Jh#pI0bTL*TdG7s?GN#JoVSP(AqcFSo9IH -kxydC!1se>7Q={5Ar_miQqY+BQLQi`<1HDnCc;YSzrY#TCCS)(YW;`26b$S{IU)W`3&ULdEe&ZOre0f -ZG6@GJy1wZ4>ESZ!9)(xGS1;wk$g$x>u~ZFuh4Ae*;mx-8aI%!29WIo>=0J~U3>udVGjQ-3efwb$Il` -ev7o>{_k6dNlgrH~pv-yB-}Z+z0ylF?tkYmz%*3!oD4t+hOWAYXNDW<_-EqQJY9ZLn*zqiZnqy7X8S+ -V;?kuqFW5EHC!8vr|tT88dXZy$l9!b`&U;Yx1zk`G -sIJ{qv_ng``qg}K{!2~w+y&1VM5!r{GHD`VJX|sM1v(^xZnl@g+MRwNJx7V#57eNG|<2qD@K}4$#1T= -Mt8UKdXgdp@@*8>0eHgE8*7JDU2XKeht$89i_icq(~(AOD-vVMZ$o$88q`BzHp{5zwrZIEf -YpMYS%@_zQF}=2X1E2SIKPnp6Oy0mn(?ry}j8;Nz-O4!?UiAUG-`1vb^3__fH}FDEGw5Z7N+J%>Uem6 -n-+^?LB4=Q#vBLT@VXtAA@exQBk1_0nL?c?5(!c)pmiWc<1xZ%&3fCU+|`#0Y^_WGPq%BN>KsNaq|s8 -;YEe~8qe_cTr+@i!`f -r!NnlJh@PNtWS&9m`8^_26|4GuRI>92CWmPdfX=?^boot=Dt{O06P*}gWI-Fb8^-A#uR@Gm;Wqyc&Hd -TTsjoi8LCeIds^bG!BYY(sllMW@M~k^l&+CBZId9}0<=(gPD@f2liX$|B#-r~nr@0IR}X19_cQGj~n3 -*?9ME3wCX@VwrDKVXuLa12FCTaA4XCa;>OVdZ@u&RMMHQ2O|rQ8?m>v4)EWCV^+K@j4to$M^xzQT(Hy -GR+m|LrMg@?_Ky5&h(qgT;F5l<*IP0Sej4>Lu>NCx4JqB#F^r(OunUWq^sOM?jzF=>*4tVi(-iM%3GG -!Zbi0po$GUgg`OI^=QpX8Kx-}l#gcr8X-#-i*a-VV7TSphbtE3vU!6OM83B{mhUjk#=auql%p@LMNC} -jt~3nWV8Efq06{`Pkr>?OG!Mu+(;_S9L?TBsYaO^XhUwh(9nTC&~H-II>qp5f_RUBZl=oMhvfho0K#A -YQWR<{|F#qZf}KLOqR8M(U1B+9=DYoGv9bmyRQZ^taojPU`to*x)oRuD6~X7Tx3xrrTCn#| -9#K%kd#PCc6;4|qLC=_d4FeiS{+zX?E0?Pq9~X=v`Tl3H;w9eCskusE4L;NT-NzP7{m7>UTnEGLi@E^ -?f0_T`&VWB)3a=MvKFfyA}v*2i;_^}d|P_VKMl)kcHIfhx3zFjoziP|!`l{t6O8VJ&;Dq-V({~hA4K( -`Z{gS~TiJcBZ1dttG?mhRbNje2v@ojIr7*Kyz)|_gk&9f5fA><0gJ?SA5C3gkyL7bgWIddDr}Tl>@>x -_!;hL4aQH?BVsc2y75~L@g4T!|U`iFh}6(>kTj2rtlmh<8QP^rTo1yeT!R#-LwRjgprH=zNlsM=t2AUqUGH~fyatkm1KG6fgGe6?4^>Xt@(-0uUX@TO -z-C*9FfXMHpS6$rvT$H?G*gq7FQbXsnEvfE+xP1EVqdj+_|wI``ngQjB_ubZ8E -kHpiPl`$G_QSiD3kGx{H$-7rKZ;~}jLNxdu088#xD|9@e;e_+Rd@s7c&{|rMi4_F^Mqscfi98@#tyX2 -tjbS6v>F7$?hNwdr8wNDc8yH0&yzaH?Tl?x^T{k!4p!CU`i-nRT$+cg3uho;RfavalcO-nnyxbYe-7u -0n2Srm<;QExW2+W5EX&A4KVh7#^rz%mZ3d2HR$em)qC-YV?e!>tUa!-~8T!|foMA^u$C{iQp-Fg7N@$ -u&t*j%&isT1aI&8)o)(#hD%KaiVPe(?vvi1~gU>^c0<$Vo7gLWTf -_dv@^OR#pPcJziulgHF?2|hcGQE9XXe|Cu_V=$o65MlMxR6(CN?34J6QvF1)ggJ?zX3G-Bh$qaVoPEk -e`?{N{F>=}bBt=~Xu<}TY2saA|7?4X-N&WSv^A=o8EP$yY7M%lx-kOr|-@X3>@?Z4pQ+NMaXc+kYWQRQr -Ngvx5fh@}f+To1T?q5*t(I)_q9cB%h5hU_43#cmQ(lg{Bs9K03~+YS(|C`^}1Bh!)3I(~wj>}<)y%c0 -6O828fzE6JA%6I%c$nnC2&zBmuUNMxx4Ye}6aTF^G#XYOoSvOVSy&xaP?HdTOyW{6A6cRG -Yj^h=<4uT?j)m0v$p`KxGhws{z?q3SIe!Z0bn}2M=Ula-XQcE=mQ}88lgAVn(N$+2F8mW6Q5`2SOQh$ -Cx4lMuJaYaL)l9)}6Cxl-#`e@apxe53iEVoUjq~=c$~(tmC*W@x9QM=WQK*9&|A+w9|OZWyLPEkvVoBvH?Fyv)C%WSFe@{u;v-l&RKaeVkY-pnKJVhK_=E0dZ5g|$pvz02UB!~k&RkaeAh$7 -CCf`FN_Ct32@`t<4+2)ZD;-NW0{K33M(Hqj%^Dalqu!2gLcSAh1_&csU8(((GC(rEX@00nl(GFel?PQ -n#$g6+K_ANXv6yLUP4x9hf3g26$6QWH&Vv1v0aG%96tIUBNTTw%^!*~N?ArhEt(V^-Nc>G^IYr@jE`B -x0&l9&BT@0!(R9>-Xb0f(-nn^oIE4Kz$u&cKxo*i&7Ls`xq#!yxDC(hSG;C7K_L*xkT}{i)5OS^}^^! -u4gP-*xtEwCh5~c5PJTl)Zs{xrbl(}4XdpB<=wZM-Sg^5~(6-#Dg3U1cJU@I0AjX5QPN -$^CV88*DuO5A)2!;uPqxg%2$#jB7DK?Jd!~!)*fR=JT72Ri9i)7_W_JM;iCZZpmW0{O#DSe{nf -|tEjeN&O~rg5xwWvP#f7W8~n6w=|3EVk+BXF7zE9;nuRyfmB2}J(c-9isv(?!#S`>ESr`gv*t}*4 -LJBkK_mMdZ9Qyb)%K_WLq!ZyG6x-1RbRZxHbNQ6lAk>s=I{KAJ$3A1Ig7wIcBXb7_<+SL2mp -1%)!2*|Hnn*`ilM!EO4wf<)N)dNk$iT`HlTOvV-$(|GXiknz>dtXbv=c4RePUSgsj%A6aO2A94CYgK# -jU-$%xl-8-wW*z!esBnAvhTum8<#PwIyc@vEd!f}o_wr&^dg4v<>fBw_U>vwPeIv$$konS~}tDYe+M9 -^lLXXd{#*vmB!*?_$QgH%bSif!kIOR+RM9UJ4E@nv|RZ+!`ly?DxxTsJr@Ox~$*vnZH)o -CE_~*a;|m>plhk3mnCT@a@YRV-*TAIZH -EVw0(N23nJq{gVl6K^^@fHY)8owNLCCYiC{(x-Xr3bAfhPI!7_K~Jo}W4v8JcS~<~@oaTM?A=tbK`eI -H=O*=Z!~*susDreS2I2y0extBJbN(n!26Z`HX%5$<>EsUf{H?&zOz#9fD*Xu1e-_=@8-XK=@n|lLM%* -11&P9gtM?lZKK!h?sCgW(yD8LB5ID@$s$xu7h!sO6Hr!h*6kyZ2&m_T9TR -)`r((GIy_HQxSJ1(>1MkWQ+nxw^e~&=#?~!7B7o#xqao9>XL -zXUoul*f#X5UTjc=!{9IJEr*Xn2V<&M1)t+OZW6{7Hh?)DCEgRfX9%l0tTWznpdB9tJFV-kro5Y^EsE -JpSE6to>}3(kM1e@m)&Rp&r}YJ~OcH -$0Ws?Vx$jT3L|Ik4r-$MShxg+1_I3{b@N%el70X}8_VhIo*WeM3zLA@7JWgH=Wo3|&>?31WfOGFXg=> -`WE#|}rr7QoVdgVhO582ZLj;P4__o)oYh0|Gy&C*gX+cz_Ye$9w2bYS8fjbk3es4rGtT{FTwrIuCMFUNSd%vjDVi3qEQMEs4k3CUa*g- -@v8YD -+Ca&&BIVlQ7`X>MtBUtcb8c}pwG&sETiPsuDUDOSkLEyyn_QE&lriZYW*OEUBGxY8gBQuE5-@~(MhnM -L_|xv6<2TwL+-i8(p(@d~yI(Mo=fzOG6-3QEWZfcQunlw!HKK+1sXl_19DWF{$b0RT`-0|XQR000O8% -K%wheqW~rdjtRg^a=m~D*ylhaA|NaUukZ1WpZv|Y%gtPbYWy+bYU-PZE$aLbZlv2FJEJCZE#_9E^v93 -R@-jeHV}Q+R}A*W+CUZ(^v!^aI=!VRV)&t#haix-66Ix0kpfAry-D-$Jwr-bt-V29!?1UGIWu$S%nVy -+JF2R72!oSVMXKwq3nYxOK?G%u&$4kX{nKhwb~1=Y1TpPcpS6q=_PsK96ULR<7j)xnu(j0%)xw7?o$F -O!Ry%1L>&jl#_u>xM${{H2r_0yYbwaC8vbh$WH2o*tZ5p7UA2oHK7YxJJU2|^h_!?Qb|XzZQz(j>Q#rAR}}1wdcL6JwOG$|)1c_76x$FJ0Kr!eCKv}Ca7?e#^dA|M{S}z=q9q=Y;ybv^Fj5`@nI<2j=tz$39n{O7*kumyrQ8Nhr7lpKktxT!|* -ZjFz+&=RM(KpILl-BLXY`wes@%M|AGbgnI!O+EXZvfFdwG)ALQT3w`Loy*t4miRI$Y*JKlsbcA`Mw}; -PpMsTl{of{{ht&mAA1WjIh@%P33xJWw5m&b3lX8|o$=*HCZPk#>LykK%MYh(4G9(z5XK>#h8c=tCb6@ -sp{>|)gT;3zADn^_)_Et)epF>i?;wPB&=4#D&^42PT6M}WClSxv4&Y*c(Bz``AnP1M-W0To{H=MgYxN@xqHRKuORVRo0Qn6ZwjI;sjZ`4g1_kaarSJ~_aT>Cqhkgg)T1tPu) -iGA`{Of`s@{!C)s6-obo#_k8i&S6NNeZ0#6l&whsb>J0E2jrq|kGEG-{(3^P5(E?3IzV1Ds9!r(%IpE -1_isMv3K`yWtC0|XQR000O8%K%who`0Lp5DEYQx#^*R?X-NLA~zv;Iu%xva9i(3;-~cW6H=kD=~Y__rzXIms#sHzeLq0jiHa8$n^j#Ku|h?Z|E(SI|lP1$?iG_t-+*+0d{PSiz@O=* -YV3TXaa-uLI`h|um`e8%9f;M$~T$(e;{WRvBeL|v*jYj|IZ^|zw+f}R;!up76&2Uo+QBzqJdgU!f1j*~S`< -Z>HM_li2Kx4v6v0Sl&PX5G8DY4r%uaQ8yJaLO^aK$v(33U5_xXl^Zf~vl(#CT|cBbscq-I7b7FGPqF= -AWQq6uGWRtr&T79Rvg?MuV%g=m=B3HEQ<{D3Axd^=ELC75bNO;_}ksm-sT< -)jz#dM9UW2}rSRbtJ`TAFmUh4VGfyZVEvN*_hwi-4&ZZc;d%Rwo&Py=KwO`cmP -!#^#mZ8|)mGWE)(;F~_XXID@*u9TLa$phKu2z1utEpn<#6(rS)E#D&Ogq{BxGG;AS0RtHYgk~3(f$&{ -`9YBI-E81Vkg+XgORSvR}zxNI@%3D-=Wc7O^HS_z@p+%2_v`o7gJ;}Yn9$*x~ZgzqgPSNUMDx_uh&TuU_h@$Kwt($&hTpm%*G*<+%k*J@Yd*q1n -k@B!SVn2DjLC)&|!e(hX7S8Dr9$cGWU`2+pZ$8IkWCf>M$Yaq@4luaoPJw$4iQG0O??nk!&lVFRIZsV -v;8faVBH@NXYqcHKwRxRz_FO$P-)}vm83>(Jx-GpK1twURCNqXn_<|E3<1U6=E`&XMt=KdV4}n@WGgr -nxM|S;95;AL?R>M*aPx(!X;FAjEE1){GhtJeA~9F$2}^93~a<_)2jrr41q%$flv0|Nz6k(5VAvJUij= -TD=>EWll5@l_BHHBaU>2Gdf~1PoB&7RDOnH3-8|$6w$-xON97rFL1GtzMw -B@vHS3$1z~~I7g>D70s6vVe+-ESLi)I358@Fm``|S3NIH)~smXZLhAt*bGq*?SM}rcl!Un1D<<^X@^} -d0}HJNuX6GD03P4&)ZWL7GnvApj2y#S@uz#EFBI%u-$^S+q%-4f(aa{kM!cOJIBbJMKDN{)=m8q%4`C -jJV(`p(r@VLf(eaDv`u+O77@%-7UNEb^;`ktmyYzxX#Pj;@dEW5S1n#0!&A`5f@c8j6ge8^g!s&C?}x -3OnV4E=&)fE4)L_im~F3DQrv(QxmL~Rpw`O%m4$XB(MlF`LszgZsJK^pF{!{0K_%P|C7f)hLh(s5EML -U^1&lv)I8CN48RUR|9=E1l^;vAAoP<@cFZ7~n6#0&5e&zN -I%48mb{n%FEzPQgGG-^;qZh#WEKVD2il0arWbdtO6-TwYwaok%ho_Z*63t~cruE=CSsvFQEV>&rKn*O -$>UbVlOjq~>_0|HRYP_b@Y?wkwSFv<636D%RAI!xV9VN?t%Igq*WftiRqrLk69ropjSB@RV}Hv)*ZYP ->N0Zy6OjU7FO@mE;Os6*fI8V9oIwSonU>0K05&Qlrn3bv!HKxO>)m=HjmxDFuOJ$2Isz$UntDBNcd}y -I7HibwaW$oVVjmQbjloHZ{d#C5)h{8Rh;+AuoKC9v_w~ -yspShdY>lr5K5Jw2N_Y9ztCDR0p(76AcrHkuhBbLSmN)X=i&0~@_d0N#Mk+cb9ERg~?ob0wIb{m~kcF -Pa>OyS@uGlMQjK03Lcx97nSx>2fn3Zpl)dDEf>4>s1bgL#+k(2_IPMJpTIYv_^xVdll5!0$0G*>KfG< -`7!EI`E@r(H4#(79(kBJVbVvF>!Z?^T9qznN}V;iAy%P>TlUzYRXD1h}xg784nr|7UNif$USua?hr^x -gTV*I9P~XjLcBhDGFb6-FohH2nxv}>v6QURfR$t^pNvf&IDN5`RXGmtVTS${dJjr3QO``2WNs-jL4A} -0@DMgy`J5iw7hEXt$-XBV2K19ove|QMoa^8o57mcvY_Ig -monFgint%SKVM3+4gTs~jk;&o`%a0fi{8~ogsy|q?_GpIPu$5-^bhj -8f&dmy(ow)ou85BV>%LpmfWqY71_r8@9|)^s&=vbdZ=-{%@s3T}R&5nwO1_7=1+0LI|18@ft#H%*IYb -`Z2$$HA$nI$CBxg~DDCP++ZU__0#FjFXK05^@J*l^sAP%W^`D=XS47IRvbPqXIC|YYzELJ$xNB>J{r6 -Fsa~s3UEN%s!w)|aiV7|Zb-$i3j6uuyFYyQ$KTC85tI=xnk@BQz?o&TILwi-ETIp8sA>6<*M?<&Jxk& -@K@`n?YK#f-e4k3kLrNK@53~g6FOIG(WBU9p0>Th>vmTMGr{zF2qAo>U$i4`q7C(%v<4?O(3$Q^*LIJ -&BSl5#T^BIR;_nYrcDGP&0QVhp6qPlsAh)%R(xb#A?sALfBb&x_^@iE65fj|G96uM_cpX2mlWe$_n3n -VQYaNCV2cn1k<(ewkxgzfp!S&Fj?5+ri)pL&8-k|PNOQSg2Ta1db!2-QQ&n?{t`q(poK=_kKm#-oT?& -*glZ52XpfvORC_MLVCsY=HoXa2;9oSFK_xWJfHSfYzsj^bn8Mh@!{jo3KMwjh9gNZjjhAXdg&hy?Z=K -<$q910|XQR000O8%K%wh4G2><{{{d6)f4~#DF6TfaA|NaUukZ1WpZv|Y%gtPbYWy+bYU-PZE$aLbZlv -2FJEPDc5^Opd8JtGZre5#{qLvXJQS4ymk-b(!!|6%3T)}H7U-}b5NPSj=0b~_NF{Z#hu^uRzNKav79f -c%k%#x=+;gv7)b=dWv}k-&Q<@5Wu59gvDog8?*S2(Yf;TJnHh;eYCE-s+2!rL- -)~L`X`@tRcJ$-g3r&aF_eYVq&j4s5pO-!%G-J`-qT+Vm07#oUn-o|t$J7SMrAhfVWlgXK2Vw4T2?B%S -8SPVyf&`ek!%Eh-j)x_==|e}$h?6mr8?6-Jh`#;8IL|we$(H}RB20{h1=<{v{JRB^pP>rq6gWHZP6v% -fPH<`{lddFOKBX9EPO5r_s&uK$nG0jm-ihI#^!apFQ?rvtdF|Ev`CdNPzqt?j>n}O;f~ZlT# -r8YwulZ^nDx0q&w%e0OU@*`c@pQHLpXwbs<1Z(Jhh4vwPJ3l+1kJ#pV1{6R2H^H_A8~V^}Ado>DswWP -j>YJAHi_p01@n=|DNf;YJU;3mfTEMwMUX-V^7xH`dJztwh;cLorw6j!2UCxEh<42wkh)|%uB)ExV>;T;Mvk#@Mn4$!$hwr2S_bQ -WhW8ql>&ei-*cWl9}o6JBk~vw7|{Sf+LRn9Qw^7sXGjOg%7ZQLr4t7_sRwQA-5zq5hn(k@g~P)`had+DG(=>5picr`$h|R!zzAe*0sUQ;WH7E3TD#v#)2 -6vO)2)6*QQ`lw#@?ID`X$Se%kU_ALug3KDKE7{UlEXx9{sV^fice0I%bU%P`Hk@6f%v9O&f~KC -zez$%~k4MTlh|=ib4_4fl_7NJVGSAFK9T0Pw(=)56J>N%p22TruK?pV~WCX|csyEb_W+Xc~wjx`2=_EBAR2i+xh%dBRsy8!nnm< -m}AoK?Im&|6cy{aqB;7z@}f#T{3h4?jNW#!VtnT+869)FD!rl@Mx9cNzr!XllK)MKhe6~yVAwk`1hY`hBkVGj;~XR_-tSk?chh92hK?$`q?BuqPD$? -nX8$+VmT~y$czk7ac`ta>a$$&QC$R-bIQg_jk5-KjG^WNk!t#bkjY%VnF%dSjg%gj<*stA)AqHLEi4( -+im~b%6bW;#hC=U7hppEty84M_N4aI|~yrOgEhaLf+s5N@AB?6>0H*oHathri}?XCoBv%Z?O5 -X}!eMD#(CXS`te0lM*W39BdZXE2H8hI7IA?Q2W-0wC~gT>QX5>Z%ue>I!3&dO(}s4o*w#-^+m==dEu1 -TQ2GoqLwNg{Dm2K<4&d%{=HgFUS)_t+jq1>ALv`P)h>@6aWAK2ms3fSzDh{i=#RK002G!001BW003}l -a4%nJZggdGZeeUMZEs{{Y;!MPUukY>bYEXCaCuWwQgT!%NKDR7OixuP$w(|wNY2kINzBYER>;jyNzEy -S2o|Ll6r~oY=9MS_ab`(oYO$V@k`fmHP)h>@6aWAK2ms3fSzF8BYdc2)008j<001EX003}la4%nJZgg -dGZeeUMZEs{{Y;!MTVQyq;WMOn=E^v93QOi!lFc7@^D;A#`srUmT6(Mm#LL3m{f{?W&o66L&BYWEesQ --?gI7zt`U%Zc*-5I+JzK0eW8;dbg1k(>b@LCe0%aFkgN}tHJKBxhk1MZF6s}qH2yqigHQ~D3afFTlKo -C00CmphGP3J$I6eM>efYusUGKc!d*aleS&B;Gf<-D@~fCB_mwy2xlPavJ=At~CiwwSaV_lrJ;{isU!~ -!r(bY1}(^DMoQ>>0CO=!C7KxdT1V?SY7IMhmkxW9Tad*>4zpPKF@bZthR9*e3aDP*EEhApF0OLkQkQl -ZXISe^B-?G*Dcg0#n62CjZtvjPyJ9jQ|M^7?tE870nZq(}7M-u^3i#^n&xdWwU0pL{e3VIWb?GxMpwU -4}R|A3g7#wu+C-+sBdlSyE{1VIxU;>ov49KOD*PSaVegRNR0|XQR000O8%K%whn~;%qlm`F+n;8HAA^ --pYaA|NaUukZ1WpZv|Y%gtZWMyn~FJx(QWn*+-b#iPjaCx;@ZI9bF68^4V!BY#I)q0KO%Kchbj9RcW=0wvgKhXXz;cHIZ2Y|yskn01zRNye(6i8K!>-gPN$=V8(zzjOuJI5Rl) -wNrfS{vQWYDy7va)Pq3v{?Bufxwm6)W0`c&Q|BhB=C-WrboHBqfs#C`*g@QvwpVahlM -w5wT$93XZJZaXPAHqdVRdVnReC8?%5dms*^6si(tVl|B%+zoO#U4PFM>g>r50T)SO_-_wNm49B)=8+F6LCO(Z;^rA;&co)6)Ird{z?I-V^3Ar_9k_2V7&n~K^PUhNdC}E}9O -UF}HIZ;D_1pu@3U4XiO>5tp40DS{F*<^1fdh&V4IsD74wz#WG$jtPi86^qepR}pTd6f78@j%&7&l0hB -TI0qRCcs4G>yio0UFezF-B+XNA~<{)o7Jq-G(<`s|bR(vaVP?TZg_1Bj@&A@G4nDpz;RDCf|K`|ADJJ -2p(iFg(=Qw4;`=hy`VkF#G>v?^dIFe<%RY*VP!tt&)J^BdJp@9qPC&E(Iq6(7bJfxx=mY}i%yOM(9F>XlcL -tsS;PW{l!3=<7cbH_KIjb+=%^&%5)=WOx -PB8Av%3N+)RUD4(!hJziw%69ooX%YYx7)d;hHZCSEXbm#3yW8Q0^w#18bUG~LgF6@3-b+O6(s1?`DE%Esr8v48bmU>!0UeK_Aq*L|LCA4-QSw;0IfQ -lZT-k}CAKXf76AFiM6wZG-4?dOmA~XCdt0UNKkY5@G^rMP;4fi=NEB^RwEakYcwG9IlY`AJ) -(wV5C@MJ*o5Y%+mMNKAfPmAgsBP@!a;h>9+z?Q^tnv$|`u-n)nH0J7H$CcuD=(|B)HjXj))*?u!U2hC>QK! -=-AVZV9C|z04$&vj((l9`0o~8Nz+h|q_6pHzRAR%e@iDx_i}44Vqn2cKvS$krcotN!Q3;P8I@w6P`&s -<{1RGys&#|yz^$8aK1HdU3>uBZltMqr2IE3KyZ&D1ceQFM9^NDy8ZMZT$ne%!z0JCcvhZSF;kFE)v{9 -Jb9XK<6Bv-AGWSRb38GAH7(rUIdJ#&H+Anox4tA5 -P>btk&M&j&n9IyDtV4)@_$fE0|XQR000O8%K%wh|0GbUQUd@0rwIT6CIA2caA|NaUukZ1WpZv|Y%gtZ -WMyn~FJ^CYZDDj@V{dMBa&K%daCxm&OK;RL5WeSE4Cay*yFpwd6(J!IT)1+mD$7k~yQYqv+D^Bn^uII -q=3$4W2ZRs1@jSkHkN4Je!1;dggGJ6kcD=D4gx1CjFO7CZu@5r6>??kcx;3`yMRO}|q`s*JFO{2=8lj -DrjZpF@axFScRi*Z5wNMp4G}uSvN$;cSI-%Q7)*4$BMWcjs@Y&c-_>V?=YgAnT9md!fR6SbQ8w;coeF -ztlbP#PF53)IcfRl?De1dBPsht-}VOtSM5z#nz%6LaHU%(~PhF?o9>?7L(Gg>`P*Q=9si+j?LT6)gQN -D0TG+HdFEj8PNAUpjBE;1_%|8p%ZXTSOBJFAHyA_dZ_hSy%2COCa5fn>j*I1YT}Z-Nf{H?NngW&myi; -{P{Z<$w87^nohiHMS>foX4i&e3OS_k=j~QjQ(raE9{V -(uq{&1QGIHKR~IKw7SdV3&q6s|a7HRMSb2jqHr|d=-EQFG0w&t%F%d)m-cnPp%x}`x9w%cj58@uc+>~2BhZp-kNk(xQWRK)w#41uB*yE -z_a#NXp;jR=P&m7VsYJqd$c0DI7#BXbcf)C;JSJipA=HLlMUTJQH;+S8T| -c92z_{;#v7H%YjFuF=77+_(H@6aWAK2ms3fS -zF(ism;Ox000IA0012T003}la4%nJZggdGZeeUMZEs{{Y;!MZZe(S6E^v8;Q%!H9Fc7`-D@NW6NGZ8T -%HDgaqE&CJmF2=v+&Z@98QSLGcMKs3TdE=iqw%~qA8%}rF(if|IcJ~hXcA#j6pb@Pa4^lQdB%I+;>WH -4d6ZIrFeW7QOkm~+7(>AL7Ger-wnYLoWHEth8WbuHfWD2v3QkChkxgrIfk|uF!(SyOs0}EO>_U?*vd8 -F9pZ)O^?QC~e{KO$iZrz5Xv0nX~I9u$5*LurZmxQj{3#nu94i%cU6AIcABw_A1!XlWpCFq@&@!JP$+tSH(P(Z$SB2AQ#Lpch2cMAbhg$x(OTWGbrS6HrS71QY-O00; -of09ji@+@@_r1ONaR3jhEs0001RX>c!JX>N37a&BR4FKusRWo&aVX>N0LVQg$+bZKvHUvgz|Z*p`laC -wDRZExE)5dQ98ad5uGfF&m@nzaV@25bce3~P~XpMpT3rBlS2CRviQYZUqKy`wByV%%1MVTn9G_x9W!Y -iD*)8l|ud(rBNHO$l;mjSKJ(er!>fIICGG9Iez({oZf-IP2D}w0W^dT^W~KQEtT(iRC-tw&;8>USTa- -6$;dQ>5Sf?4(H>&P(d)xlNCiVf|fd<(?aDAKk|=L$aj@d55`Z{ -UA#bUYRs9Khhe`@ixjbS_-k$USx+jqnY@?!+;H-5`jKC#T@hppKxUxWPgAKAu=XM@teRSyWJpvcUvT= -qgIYJs=2oj3Cz7mUuxRn)7mYtC-&{8$Swb+UTF&GO9_f!QVh%~PK|NJ#-fzX~JsbW! -y!8~Aco_2#m+@0@!mkqt-f;uHltOCzoOVUZIXh?KbT~^^C-iYjQ|j0c=a3hj==_noxTg__)%D1d#i-A -R)@eRG8-`KerFg)7bIhC$Pp8EE)JrLh^V|6}%X2EZ;=Df^{fXR(tt9g>+2H5J+#Locd$@>{;WwiEH)i -w%_Aw>nH#z_M=FRGAbf?Kq*PaC9-}k@2UqjhCYSBR*m#13yMv>QC)4m&@-6aV|fZU33FH}pjDghl!oz -yNV!QJ4ZzH;FI1M7*`q=x20PPd1ssctDNw4}z^ct-@^mJK(LEb9W2^ -+VIR1dk+W%yLs*f<@N|>!H;=Yoi(5Y`N*^u{VVwI43LGm#C^=zLJn98Jc#JQ!%gX(TWujY9+lV(ywlB -Z{NO|(rAQFRJw2-9Z%xt*%wet0|XQR000O8%K%whK}Y*`-x&Y^AyohX8~^|SaA|NaUukZ1WpZv|Y%gt -ZWMyn~FKlUUYc6nk)ja)o+_sUw>#yKZmxRo%XiJWhRM&0ox=QV9V*6Rn8+xfE*I4|`6(w)b8YD)Se;7DaPYD9OugWG=E -c60;&*W{3q}e((K(c7!UZ-tM&~c0LJMI5ovGBLym?oG2T2#5bO*Avqje- -0Eo1fLtt7`Z5kmF<%%nj=$m$Lnq+M+bAU`x{!Hgj)3c-7?4ZzWoQu4aK0O<6#LChk&P%RmIUjWU;2CW -HDKo|dquw$B1^xLJx8fZfR8rYK85IJ*kkI_Q*4_R`buU%!6!=IzO!)8|i5pP!rpR#B*y;_=~8I1BLGq -p!cv-;N%CX}%qqwnO}Pl|K#<<0qD#BcRPsvYNkbIvMX#Q}At3T_Gnh@xm~~28F0(Oc_||hF2w57c$Rf -PRj_cMcGM}1aFsx3K$ki9*kYqxC#Pzj+8qOR-I~t3m{Cj304TF3F<{~K3eFx9_9eeM1>;(8=xZy2BId -aocH2i54B7*3blh^s^D@7Jt7b7bXGUO3~eP9a0@;HAXShpWp;%FZoph+(-2 -uB2QuOhA+LY~_%CNHMS!4p?!CyJA!slCmNa9$4xEoJxWq=9*FAe<>8oD-br97u_Y!T{$q7S1p#&nmjV!SV5Fkbd#wfC#S~Vt5+&}b79do_71I%UW?hO5gvN|$!gu?ho<$$BWnE --a-Np0h-QPYv`#cU4gdy5~qljI&+dTX^On$73DuNxAfcsiTwB^VaF>#~)iL;O7FSbx)R@X5zPA(T<+Q -Ig=pjAP)r;(E7A`TAz5WK7_d2ENF9R<~jPm`^)Fw7^k!7JeYI_C!K79^Q~9keqqK!J~0H_A#y7J(R?a -52g(_<4X9$j2!t>THR^#B7Gd5X@AnTLq%LtQ(N8RWHJzy=zo=b=R)DmW7EV%|L~10ElzoRFzDMR-$Oq -2%3A!g4yDMz>7YXY*oR1s_21jM~r4#qs)15LQj{YlYfMb7oPbc#R^3FLXP!Dxk!pyL>elEy%G@yg&V@_}m#3TdftObGH9JOK -Wq~svv6s&=2=Lg7!xpfj3$~D@1h)(qt6}-WC1}T?Hrc^#!>+pDJor!@2iHs%&?iV1!+5P+fd)dL)g&$ -N%$&F#%Yi-^&gU_;CU%;^5-ZZG6b68^)kgTISNZU-Wl{R~84dxq3$b5_Zl(ynfAcNH5kXWWG6}>2opF -jpS~LZ~(czO2^p9x3^n+I)CmzWC;0Fo2^S-OW{aH6fP2L0k#L@6|C9c4XL)eQeU24B0uRHWEX;hxLYj -gwEgq}xUlb(uJ8pl2@P{=GnK8_a1JBy-G7)xO+?NNk0FLGPdx@J!q9z(b$N)D%BKD7a_ZRJvM0P;5@= -zj3DRP_ue`~jSV`i5O}!jv#u6_?A_`0uEU8u08ua4>jdvtwjuS$AbloO)sB2wqj_Q*^iJS^_feXr&!v -Jzm}x=ns-Ds7t^{19X7Fnj7>gYs&$Rd=D{{0f8QAO%NKwniq}}TkZjI^oJ(6R8&_-&>jt*!OtV}+%S` -tNPv}q@oid#S~?0JJa`bsy>NPZ33&yBdGwDq==Gx}pONy2XaOYz2T}2&xW(DR*cA -x)heGSMY%;K-2io&G4#uAcd*D}r8ove&jQpiW5lk38`T3KhZ{iTuU-IHFFJHYmdG_@5B%W3(5&j_=nx -Tt8oS$+VdI0wc7PA^?U9^ivw0v$a76ZVRK3b%~?S&;w2{`qXVEcEE@f75qW~WmWM`SkQC`~g*CVRmnA -h?yPCVij*H1fJoxGeUVDPH-{XQ#M8aBv!w+!}Tac7JWaCnhFt(Q=G;Dj^*~B-fr9%?&-wn0IHab_g?G --mJ$SFfd7|$mR?afDR!KL6`;a!tgAP2lO%%Bw#{il}9}hBQr}JJKX;gU7fP{Nd}+X-Ub{=sUVUGSl>^ -M7RVFGV!J1?L;n*KYhFhO}k6omS6(Uq__uDbjNs48#83R9 -EFmY>0Ybq2nxjYVN^;YrI$sSD&A3QV;#7k+Z|Fyz1x10-xiN01fsfzjVJ$OrtT13OJB=Uu -0m2a%}?e`K7F6BT8Iumtk62LpY~|h&+DqfK!VPkfHtvO0U`MoAB%0C{5Lsj^5N3wkeqtZDh93UKJP|f -=G#i)wW0>)n(Vp6orK3>Ov^ex7q~k7PF -_pLMU_{Fn1-9&4f|>!v&*on-Bsv~>$SS7YKE&rSNwv8B!%Nj9KDDds$I&U%QV7ZJP`*_F7IoOAz7aq^G&&_;x5s ->-@mX^K{$kCjt~hNx-P!_C9r$TVI9M-pxmcnDI_k%Zfd-ckX`r|S)&xn{mcz`b*yU#P@cm6vYgu69=j@f0P&@Ju%W|Epeg<{b6CM!wHZ3u*CWHuo@2T -kd)b6-!+1UCZDg1Rd2*M}d_5A1Oz!%0S)cFMU-t*Xg#0`T)Uy5GTT>UZveck(r|%z=7qnH -BJzyemBYc=W|%oh`?sA8-Qt9@v+xSA1!K*9XF50M%{AQ-C{E0ylX<37^;zroOFT&CH~|N6YF4vHNHpOgVNVM}M=MGA -UJVGT6(X3$S`}c?Ha*c25d%~x5j~?t>Of+&M5CggnuH|^h -CLi5-_%hO8rE!1XwAp2*$7yXWA=d}?7Y6I%39=pM%@YD-clCEDCF5tZ}O&&nbWclg|`koN^#J> -Tx+T5+5;AqElEE31Px<5!_EKkYd#%;ktf;_&axJD7`H9rP*4e%mZJqYVSM8QG$~HW}`zrFJBy_hIqg@m8cm@ivB@m#kW7h_y}%C%Gx7J_9z -Z5rjx0S&vY>E@42kJH_;81t+e6e@dQt~{b7Ic#yhFuZ$4I;eo6;Z7Z2S7*!o}b-_WTw~SNcsv5ATeWV0lggK=OY3T_>b}KA{qZD;lY9T-m76ALKgx(48sjHb(K56LC8H -M(X|u`1cV39(>y$j%n==nx0mhv%bq3BCpgB9PDxV)KR(p{)~%mFaxmOtcw8A0PUNRk~E>+*mJ@8IIkm -Uj_H$6ljsDJ*fn6$RTN@q=drf!dhe0B8z*6EWt>NwNC6o7%4;gVVP&yCPNM; -#w#Z)sJHCy7)?UrEPB$92&?AW>$r -yz=HsS*$h}rpa!PlLp@m&u)=kPT?y%rPo6+4y7e9Ww`PbQE#w5__ex`c2o6d2zn&f*7GUd4DgLP^<1d -n{xO(h-B`VTh6eHI#lZ!yDD7)~Xkb{OQXAgpQrW!9F){&psJ>vI0F#&_#9lV-SI`#n~e5=pFw!?x-09 -Ehy3ZmSm#C^8UOs!hHTRXXAqfER9G+MC^%4e_Axri`uz!HmSM|g3A`WK}}r&9B%3uJ_;?BBu>NDxmW^26*Ud| -0qZDy;FuM1m*)L#&;%VvUFT>jW1OuZQ!y}!eaRGEa>Bzti2AOm)X_WZD%`@L0g$4~$}X4aw}C&{*t;W -Z2r+KzW?($@MM&zJVi#QJgnx_PJw5oJhX=nq``LWa -|0e0d+2^J${?}m*H0K)}o8}!uxO5)nBFtsC@$rRo(aq|Skp`yWACWBB2b&rI^nM4+y1v45>?=&yYI+5 -lQd*0Nvc{XVg)W$8I3Ta)$R0Vg0q5QnvNQ==ZmzGVqa-ZlU>Cg?Wp^|DYj@G%XnC5+{QXO4Oysr>tbc -#se;h|>&se*EcYZylOCVijSk={mArDTKOvDnp>Jo>|L%wdVOKIK*Pbg8NHZaRA$A?_B&!r;HgcEQtVG -kiTnTJPTCJ!f`7fN+U~(}Q7u`5l_9ia$MS!mWH#Z3uleHP)-+;NA5ZW&W=8Wjpr(-dwr6+se>3tlG-I)_(*`j8`ZkZvuX+Llno -ez)28H&y$Xv_+K{JeRHT~eRF>dvdw*A2aqgzxDR&lFn;&&%ryMbG;E!`P_(oC+&7|QfoRLEb5vLHy -^0I(74YqWl&;Xdp^F_+|Hh=)-Y78VU3`lGkGtM?1*j)kWgC7~fcvf(Q#V=I>uz;<&2RbO_lDTltlZRC ->v+)DH0vB5)%_7bbT#o(8RjTdqhYfG+0z_84(p7rElo9XjKn@=J` -5sP1xR-iW~P+q$N+kc>|P_!?R$=7*nN%#Xt0S_L6|`4k$$n6nQY!P8r0$^s9NyNZvJb;TVfAe5%xGk$ -deUE-u$w)6_0MoPYb3dV}M9VWq_L0X6!plNXtoxWw;zgwl1*|I7!svV`wdY}p{gvTn|2KOWD-?c(5K- -0xS(yJ59`-V`B=Xh@s;q!?MmS406Gttg8%VT+=frffUsY1Wc=du%x0C3Xmx3ZVj2x~mt>9=aeSQ=1N* -{B#gH`cgDf?JkS&wzHaI``Jo@uTsqW -8&A5E>idHiu94H9#ccFdQVK!Xj%ri>&c7d}Zrq*RsZwecU5DpP~|4l!l7vu39b0SObf}9dwe-qU`EUu -~H-^G|;SmpsLv$iSH4r6&DI%$$N2G=~CZSe>07~a60AKPoB`lPFCsEiTKGalqdl~{C38deBGVxUs^T& -LJj}&Hbp!5=a_TqXh&U(a_k_9uXMwjhhYjx-)5ZaZaOWOG}=Xc!Y<;$;h9d?JauoU;+4m*AAbG2Szk% -CL|%ryq}!1Y-3SQP>9Qt1(APp2X0oBrXe){si{2z#EJF*6O}Ufd@-~6~rtejxXEOcl)qiB@kPy{R`pL -Y5&F}^8KBY)N0{Z!2B%V_`*PvS1J}2?Qw^j5J`U-ZVK$dkk+sdrTKk&OGz!ymzIQj|_1iq3*@j`-3 -Z$P&G$1d`shjUhwLv)SlCN-;?4C`(_68W^%4G_EcZ>G>JZG&>kP(B;HJ{lVz09$5lQktJFb~5%CksE~ -J`2Rpcd{^bafWyn!TmA3C+Jm;4eqq^E5=Z#AR|MdX59f&EQhRFn9YwIHPgH6hByP2bQq&p4_lAoz5fG -HO9KQH000080LuVbTh=>&D%1r407Do603HAU0B~t=FJEbHbY*gGVQepLZ)9a`b1!gtY;R+0E^v9}7;A -6aHt@TD1>x?4gt@}@)qsZvZMFr?+M*pc6sb$l(m5tVi3&+Maku^V-BC|lbed#cF<=A|SmeEU@8jC&oN ->OcTxEcBCi7Ao$3#(RC!ExUolKg&OgND$VJ+Bp%fMt!1X8UOq#isGtF)zZ=D>)RYJdj#?2u#)Q3|(Fa -@C%`!O!}@?Mhi(w)?-ne{;&8|9txV<=LyBX6&TcO&WYzCVUG;rcDd$bgm@783M=x#VUkKa-D&)NhS@- -&MTu@q!W{Fe}_N~T+<(JSSM7+@~r-R3)dBjlSAP`$d!}G1|{t -@h`+hhoHU1qTJP&-K4G{98su>6ps-}8mcTnu>&fKJ`}5b&U-BQ%&QJNPlV48x+f%kc^CqdzOQ|3-(~s -p&>GbMD5?y>umfyx764LEA(|9tOWUyw^@~u!Z<24*#BN{J64iO1A_tuQr(;wJMYgJ2T0KKm8W={x`Ef -#EAqkH00hK3~Y2qnpdOE*zK&{Cq;j@t*Zg@BgD>(kpQZ=w3BbG!@Jl -O5(rJ!2W~%=_&ZVCoCe0wkpYlkj1Q4@#~U8P$;^_vbn%@1r|ZLfvqINkBFg#bLqIudY-mg)j@%J+k-z -Y{fiFXQ>f)>=|LtP<$<|3=8*X#I5ojJ{{q?>Lt6O>29Owd#~{&#w8*%V`49nR2HL8zHo|Ve&Ihj5v3| -tG-(~2#^|i^@d-Xm}xUZ$UBT6`a&)+uy>VI(0&=t2}4ucf)@EvmJ!d!v5pX8=4E*3Mi+YpocAZ1WuL4 -_ifSoXb$n5PCgDt9_yL+pB&J7S*>X-byVKlZi1V>e~40LXAD2C=Yj5G##X`tR!B3b -%+w3iUP-G3Kqa(4u?SkNwh+Y~lvnM-^GPnjarzz}>zl&IsKnN(h1=WgK&dmNo-lY=*49A^c&&(nWZ>U?=Vcqt~XDYfWk(i?7s*f(V_r}2Ay^^vQ -HczMGnMF07PLL*#a_%+E}{-Twn^V8Ha7uc42p1G+_7^>nILd>@jGVc%e_O_q@VabhE(>7;rVjr -~Dnbc+&=i<5ft1X?K?6077XPvQuNU2^g|<=B7OMuG|O~HsS-4M=fY5*6Wjh*#Z{EI(owFMpr6h1z;nD -5|-F|jG*qL#WDRZYu6nX>JzJL-hpSan-5MeJi#xJn-!L6 -dDYSgK_RQ%~fWE@c)vgD4al0Pp&9uTfbUE4$yDY6%XSV5uhF#}z#XLkQ4^+tGl<{>Vw0ryUx(bd-{vo -KA!M`Z&4Mz(${XY3_dZe3%Vk?a?VF@lpeVI=0oWyKO1UY4 -f(?Rdz795c1%K6`cxH7U6&(D%CxLEeIq22NMVl9N2x(>sn-STJU270?iNY+Pu3aS(TSI8VV*XO~@X5# -)dC+i}9l!z2_0p2Ms=hM!X(@0wDA5ZTez78!ZoqBTp|dAvJfUi6m5O?8180j&ndbD>Rl&j!twm(e$!~ -6BoTk5*)olhHgGC%;Riu1?VXvUuA-ob5E^v8eQASGIH9W^Am87?4fs3pb#aYwxKw1 --#&?$8HZ2Ri=^*jAv9JcEs`(m`h+M8>IhGB1iMTFJYky4dECH^p5sYNoZKL+-Q*R~VGIO}1_?4do_F= -1tA9feICENcNMWZ@S1jLB^~epO>%8@@4t4f(3kLT_Om6oy{QYwoK#b2rFXgN3rZX9mbv}`!dGd{}Em% -$scML@4TfOGeEE!;?51EFWGt+0dO%?0v)Oo*gJ0>AQNmQHH^|118Wq5$}?Gw?34dG@<4;EzDBG6p9SG -c!JX>N37a&BR4FKusRWo&aVb7f(2V`yJjHW2>4ze05==c -OB^O$mjAElEQQfukI`lzvDLM{8@>%Gs79$$Pgc&2R5B()!)k7HaacmPVs_dFGk1Dr;>VQ%)MEbzxU3E -EM{1P}QM*)!jy|t$L`1%iD#<6ouOHS{N%=d!tL12`AD*SSu|%3w2AYAX2NECr4Rj+Tdh*C+ITXI8}5< -Qc>wjrJ_)uWm1V!4v(q|23MlsrOsqwxvDbx(4M?LUB6k9_3C-5OPJrv=E5cpZa*qsDx0>#l{RIthXSx -+MrI;)iD+D&kb(rUSgj~@(y(nCLw=C%%CE(VcO*&HtJUgd7n2xgKg(+S&NOnp@=NSS3X|qHsjlVais6 -Z($p8Oym5SOng>ZvY@wY^5}5iv+9QNXp{hF@ZC -K>vN$&W!N+*6Hx(z3L*-KU^>)u;uZv9SwZo1RzxMVxvEa-XZeJ!QSb>eCmUVF2c7-oj)-A9=^CiZ|#hYqCNC+0(M><7N6^9lQ@QAIZJinBL+ -CMqV(p($nMn03N9%kGuS*_RFKl$GMdc5;wbjcz%2-~+ke6(RC0iX<_)a`o=?Y#&Ok*Dk8=u5O73wbojdw -(8RjENzp)%8psinZk=z(aK7R>OjFdRpaYsgcaoWiIL(bb7ve5L4jk9vINo1(6*x@$Tw|$x7@BnQbVw7 -5n|w+h55#1iLOoDskY#5V7dHF+uKIQK}2I!s1o~p|^iJRKc##R=4Pu(c;K8>&h_#Bmd=+QXa+c6dZK1dlK>=O-6*S^9b8Z5ZQFi -=QZyqyCbtCX$D;W^IcR>7#HaU#awK|Qz66Lw9Q_CWdV?bWdZ(s2Cf=o*b`dd&XihFUgfgd{eU<(zsp{@z9edn^dZqTTWP*yJJkq@b?L(LN -k&tXToV6!9_}Wt(fIt<-!=VE+3N;)KPoXL% -0Bw9bD7?SbJ*~A(cD4rB_;)yZV^u&j@dPNrf$<7If8$Fn=Zw87K&j0k4`Efrt8Ef&Ez^mjn0Sr4CVkCgK*V$}iB8 -EP-MDa!G_OJWlwoQVA&(PHzUk5z(HTPLPYD`Zs8Vt66voqxU@~;}!cJ7}B<5hH4IizaNf`;`k9(_0N@7z04D$d0B~t=F -JEbHbY*gGVQepLZ)9a`b1!pcY-M9~X>V>{aB^j4b1rasbyZ7m+b|5i`&SU|p)JsQFYK@Y9WeB9T!XE{ -4#Qw5Hr;5EB~O&o7TAv;CE0Q0COO23#K%W|AL+)L7Rs{eU2m~0LAIT-4usam2`7!FENi??C%dAQ+M(4 -#6|JdJQPCLN3RhM}J8RVFd>2ly<6)L%C-888bH9LxwItApHNo=5t&N5c#k;qsBcQ$uaX3 -2E!ih==A|#IE46G11-4?B5d&XlxOh!c(9`GF%Rv!J*VuKcSg|x_kKY5;q=dG?pN4*jb3EY~43SCPMgg -~P+bdz@kW1h4!qL%tG1YsA3Ku##dN6<)x3}H+Yew1aXi934*zrsnlW)JgJz6ia+I=`5o^LNR)Edvo;_ -`X5f@%@_!wp>Xq?92XzeAI@ywfMJ}7OCt)?+1(D5#|Zp7=7xvno9LGa%Hd&5o&BiuN-WJ>e18CiVb&) -H3KNBHFr;`ArMIyC?Z*-C3spRx7WfDXsihYyM`247f~5Pk>NnSr#gmmYvDL`CCRf9$!`V2!{<)PN_xB -ld69vNWqH9VB|}=H>c&7Fnogm|q8ULTF#555B#aAWuB4W(EYEhmV<|LKC-*zDfR!<7em5g78Zq;CTpL -ctDT#}8VBjBoH5%E=KPIPy-(!&vb3Ire2z(BgpTpn9q?bIu<{mML;@}*CS^eR|BHMW^I$=dSYtPSt_b -?R9n)U1p8g+0sFh{5fcai+ZG_br~!pBeV&x1%hw5@c0+K1^M2Q|**Fn`4y&Ai87W?ha|07vlU7x05(h -QT3$6oHOgfdP`~ZN<<;wGodDd%2^D-x8+n`1P<-DR=p|!TB;S_&hEQ@^c#NgqyfY$cGE!E%uc -+3As6f2Y6GYf<3PLaXzwSpo?U)T_w6bWH=k@oyY7vb)CH#uV#8ojT5rjHu;~L9xe3nMfM+1O9KQH000 -080LuVbTPB)h_Z|fR0Fw{^03`qb0B~t=FJEbHbY*gGVQepLZ)9a`b1!sZa%W|9UvPPJXm4&VaCyB~L2 -v6e48G@AaOq(Q;3r`3ZI=KU&_f68vK3ozMd7E|a-vn13`uU5y}kcFQgUs}cKUjn4~Z?46iIy_Nmtg?k -mptFTPyM$WZf9+fop9%_tI#$T6Jsgg7+24H?m>*i_oR9tl`BoKT3UMo*&(-{sTXbWu^AQYOYx8rE;t= -b%VwaQb~Ww$9hNMbJsR>FUkyFgmvfx>Q^TBCfBAk{<_^-y5iN6Hn!$U{v-0Sk=1HdDDE6OXmPLmi$58 -?TLE6{^*a3D>4G<|Ron{){v@EWrV-Xld^8nsh)NLlZmhsxFj{~_>0lId5e!wvqIAp;zShojr7|eZSu4 -XaI^he}Zyj#(+$rN7p2HpuiwV2kFMom;@ziq2n2j-!vE)dR8amgXQJb*?##8=C -%w+cV#(u!gsHpVsNFXFhHrL&+|>M^-Rw{H-M+UrlkFw6;R#G!~K>x(DbF>f -YstZ|q#XG7IBT7kOAKI8yw0@bpg|%n!L74)1b78ZTPLKaqgK5YtR}l)7V?=`kE0CuSlY;4{EtAbUI(N -2!Y+>afJKznpPQXU6BZJSFP&VH2arf!z^OhyX9!^*WD1|5fB0vd!^e6ZsgO3@V@+UZ-grA~7GckR -meQ07x(|`(yAX*?eABS#$Z`rA2kKg{sv>h>w%8nHN*w-*$fdUv7-)a|3^RgD58yh9WpU_JFC7)iXjLf -im5OC6$IPmr`W(WMwtffoirVhMqD$pnqm^2{65iQ0|)smV4M{BDh^Gltk2Cvo-t6NWH0fv%w?GraL_D -K0a+n+95%NK~$!{^hF8_6iF9S;h|#ehw3pA8T&dsTod}{Io72bg)ae7CFvU8>#|fWYg- -c3m}OYo#15XWL>Yok!I2>u4+*4!-gRimHJ*7>r4dXhD?N3f(NzSNq!090~0cYhyD;Ob014;Y#3BxG3S -?MMs+Voxlcktw7b|vYVqpZ>XK~5HnYv_NqH8lAbS9YGW) -snL%9+M86OY5zK$n@oboyE%&`4=03LQ?@u#~!L{RIR5y8jYF{?2|~hy3AM)*&0SI7$f~Ww$&WT5oVW; -4l=-fd~T>d)O}n@{i~&0dg2Z?(gp5$B#cj80v|GCs908ffkDTVkyfoSMwvMftU?v`9M@9B^qjC9>So3 -U8kf88>_GtmB7)az>ZuS?2vK6(qJL2^V?;>_lO?&xm>_+;GVt=x>I-iKrPhPW>%Wd*s@|fauWTSm`N> -QHFDq@QAnIiFyFilL(NeZ>FBV~Ki;ANM^o=kJMx`vp2?lNvLAn!#ru;#y5^L&V|Rb^!mBvXJ6;PKktQ_Fa6I@6aWAK2ms3fSzDBMO32y+0 -03wU0012T003}la4%nJZggdGZeeUMZEs{{Y;!MnXk}$=E^v9pR$Z&yHV}RHuMqnZFKkQ*EtG`>`jEhV -DVH>~50}Mg?eVS}TQZWo+5G#Sk^CLsq%DOQF8IpxbT5+B+e|=POYyDpiw^x)iN7H&qh -86a_gayb=CN;wG&!ofECJ@G=!2CR#Rx7wN@S-bgegwbGX?+3j{oQd!f8Jg++6S;}*vn$}p49NKv4mC> -%ZDNJ2c5l%9BE(XH~Pgb65>aAMUX8DbDX>8WY;*-2^Dlac?aQG-M-!kpIsz=*25qKW-P1dUXlNopq0s -Lf{uTB0lA^cX;Z>|bY14BYd1Nfqc6%O^L8&@z7@Y2O)(5VY?Td0;*FRp9xS~5GMV!ep($M%*l(dthnM~67qNbaW$ptlb6Kez -ZSs;*Chf&Ny*lkhAeV6^OI)FK0{yZMG#}#Oxl?tSV~P$F_&f~Nn;EoapgsaDz+yTC+)u1u$mf-@jr5+ -%d<+?y5LmC3^G|R-P55@KN+?AKwT09#AU>`{vPvCBg392@>Vq?Qa~XGgq$m1E%am(_JNPh`;|Qf!5N8 -#;uH~5n*plvfqjWzoY<7(0LS3g&JVy4pvTO61P+y1=_X66`FN+b=kRQ5l&Pw;IUDZ){0+sVC*HqA|#h6Qm=$*C39>qf?^fP3~ -N-&MZwFkb)$mVNaXYGAn`aKz%?mksv%oJ#1SfslK%iB{L<}5U1YyMUX;dgJ^u=RN)wxmA+PG`egVcv4 -4EFUT3_~}BmU(Iqj$i0D+GFJI7M^UOGeB;*mQ*i&@#1GB`@2k%7n14$L6S*rr{3{)Z&S4o$X7i|XgqV -T0h2?>3uu-mdfS~xww~WBj%MrnB>ONi}48_Ti7FD&iENE8Dv^+wq4OVy|T|^M(oW&q-MwOG;F)H_#v_ -sXvrb$}i3Y;RU6g5^ljgVJ@|F5#kHX)@dg+>q{?#dZ$Q6vBRplr~HpRg-t -GUZC;{xaAVap@(015ir?1QY-O00;of09jjO5aR(c0000n0000X0001RX>c!JX>N37a&BR4FK%UYcW-i -QFJE72ZfSI1UoLQYQ&LiL&d)1J%*-oRC@3vT&8bw#OD!qSFUr;hauQ2YQxr-|GIKIZGEc!JX>N37a&BR4FK%UYcW-iQFJX0bXfAMh#XM_s+s -KjM^(zp#u9n=RVLRT9W7RRWUiq=Ab+kU)n^elBfFMcr_ -cWNM=^5*uFcf`uy+6rUDMR7Y+r -PONqmY`sS$bdGjYTE9QfRoIilmDOET6LjU1mTRyy-u!X4I4&)C(%@MLzBz^4!w3w;5j_d{>L?m=8Z+A4XcV>^Dy -2&Wtu$bsYXiP$%FFCPHj4KJPN^B6LhrN}NpeLyf#wmaUqL_clU{OG8mtt3mkty!M8z{NbDA07St>>sa -B&tBwDPv1QG@#V9pFzz{M3^4boYyj=##H!KE@_v3y2m@xv2?(S7A`qz^vv>5qytX&{Q#i83m -iRWPwxK3P*UAHPO>b!ej!n4_TIeG!Y2}~=q%%t>}+TnL$U}N$0TfFh9O~$aF*X+nM?#viaXfSmTn$>T -@uBY@JaUS0IX?0k7(8dU2W@fi`WP-_hr#E2c=)&cit-szMmlL+iq|+oWn*zw{zNRWD81d)$}8CABSfa -sQ3I|!L2*Vn_=&w@{OQ1tT!EzN5}~@u1Dl7EY}BL)xe1Xe5t`IOWtm4u+}2sEl{?&Eb69MHPVq=APSM -cEFXNI{V}_rS(>ep)=XQVikrpwl*lO6!(iakI4_7j^n(BZ}@?Rn)y3Rd7I0cyEE}r ->I&(L<84g0eJeF6achv*{ULm=8XFw4Ny?^{dpvDL&x#7eTC`EfAdGf3v-U-KQEY(oFbp1p#1##1H -v%AII6)w;7)>>ExDJ?iI}Eo0rq=eW8iui!^b}?%J=e}(4Pck#tZ6M7;{#qAE=ZrF=5;4}7)k(MJ$d_6 -rZ@hX_7oIicPoAY7#R7lw4U0cA@yK|0x}r1?G4+3K?una7z+o?w(s}i_};w@7|mgo0Uz%DRzRr1@2RQ -xRt;a>d;I86e}4S%kKaLtu)hCbRXuq0_+j%icmNL?>VhQ5O!kJ -i+vOAZ4(T>vJG84&d!pfkYkcVj1;`Wv%*IyZA!dkJ$P1%->c<1NVY{xF-n&)=o5{_*ywU(QZ{d3JX4% -Tq{4(sP)T4*mM>pHqh=0)@`x(^)sPGv6QYREniPqOa<9=}dY0>~CkkzI-_v1%Zh>?A7V3XZ{#Ka|!?K -&ACfgKrNG$!$QO~MT2`dO)d^@7iZ4Yp6778tM@Y?8=79h_O5t|&lj(o1^%{jCceI$eSJ~6PDnyA7T1Z -fF9ejKIu-u&`pRO{!X;kC!$ZBqFRcSZOdz0~__MR8 -MoV|KwdNlO(_@1v}5va|@*J4!(S$00YE$7U(;s8)*U(a;y@xbRLspT+B-8hm!640q(J7ANH)oOqA)_!7K>s1f*O|m0T5_gn)wWjoJweiqcaP8=}s+5gRgA -CU15f6pfK@H*8!g_KZEdTnzvIcLH1lrctcQAri!zu33|PVCB$Lj0p570Eec*$R&;O4fh62ag?CBNCigR1N(!l>}D;`;fm#Cj=T|QZ7?R!Aq -baHbAp;z4ATD{+Pu8(XF6^fGn-1zo;ioz&QgR4M5S1IG>2kDtQ^N&hW?CnJ-#ZE7!aGKCvS3Ebi5pXl -m^b8^x$zW{}tO>DBJI3(XU7o@y~2;hJRU76=2y@ylrBon*D=cOQQL==<-Flc&7>rl)Mn -&OoxEa?`?!*vR(evgm5kqFO|tf>}e|16#s&`~HAj^h!_qql_Hk=EqG^W&wLVF}Ojjj{+muZ*k(gbMH3 -R737ZETM1uA74w-o;J`2`B8M7395R-HNofM2A_w=GfIwgk-fgBfbB#O@7Fn_1Lyk4Eo|3oc0Pu|W&-u -_+5hq~*g0mP=PRfhga`qdqTUty6M83{WEeOz3;uzShXkF+A_h>Dq_7%p2Qyc+cqpmYj5}~2Di&ki96E -w&Le867G+*V>@CyO(EP^B2`l^MKpf(_XK6VM%V0HMkh_Br@9_5lxY#BtKpqJIZUcCKxhf!C_!0> -kitT%5hT#4uc4cH`o3UiwlggU!sMQ+=^X5lRM{=K^a!7`uL?$EsFISxE_Pw0HfQm0fv;yvs7gAw -Xc0%6W1Ci4bIv7#x6L@&Ip)cA++J|>a95~N9lZ0<3YvSPK3LEOWbjfY~`QLZAh63DVU>KZ^J=F6v-6`K*NeoB}Q73wZQ@rDJ~ez$3qp4A`1S%1qKHBzz2xmT -arzAbS&i-G`Q{$h#?;pmV2SUs_iu`8*Fwc`mo79Dy! -0Dh+fVM2g6cDW5DQr)m0H?g;HjkB9DLU?@f^l1&Y{Ag<~i%w%pI<_-jlxXUeBPmE@0{1qkcLmTo`6QtLDnJ>9Hi+R^bc)w2*80^zD -kC@6ya`YTY37MsDLBEK9tOx;~$d!Y%5&iqK2riN-FRrDyY`zrA|hQ(ZM{x;z@H=9Pog7g9@_6)8!T*s -lE=(#VPmK;GziWBzP1=-V#Sy?HSF+vnzu#zeYI=s3rN6S3UV8s}t0XVLi4KLa3j -HK9##25?=8-BC5()tkDdb8di>{QCN3vMc&>tB=*JwSq^P)ru2DwL*ha9YDBi*)FgFUP_PS#uU5akWl+ -svB91(OJAh}u4G?PJIN2iPV1V9GYc{F63OlwKPi%_{6hYV%9`s0&lNEF@ctY5*qB@ueD!> -VWBfua|YTbPtj1R2avKbI3)rO~7M}e1vzT|<-{FS{jgqO>r=~`pPv_LY$5ao8@nSp|1l$Sw=axej4=J -GpY_z)ys2C?!j>>*BGM%%j3+|9F27W314=^P7R_w4G_HOiG8e47(D3t50em<2P?H?ZLJ06$@|FjJ;b|-!h@kN)(FO~!NYQ5j9=U -VMZh{Q$(noU5k_z!`G;0Az|!R1{qyj+wPjasb~F{HX*-sigJJXNTYCwXn_tP+(TTiXj_{ffbthdUxbs -}ibEpBS%MOu9NZi($aLNY -R8AxT*$f3nMlu~*MX5{zHi14^rfHFj%J&4Ikg&f)(x(U)NV_kXlQNGV*E2^@cj*3t^|X$XIx?xa9Xy) -sF+8lOp0L#=JD1?D74}fDJ^iw)-D&clxz=eUop#(8QAlEcTk5!P9(aNXdv|tSSPuHtPPfJ>Ya-d0I@$ -vTijSFq|RwV22PRYr+KA%c`y>yO~)ai)C=}uuWU0bU0TN7C=M1$RpW639Ugs0#bEjMKGrv6jsyr*0hveyZ#Regh5a%#bI{M)V;*l^|op?+IcH+SOax*ufc8pxY$hA` -(qTem>qK=Kw>(W$Yai)1zQK0c6DG5<{g&pyvwk4@g1&bIAlWH+%Bx9h~+o~ZfJLkkSN(yTLs-pSeLf% -!X?g?J7G!!uaY0(l@sRVpua=$I-hkA->>fHmYS@!c!~P*<9;*t2O-2(r-d#>iprEg{Zt$2K{h5D*=1e -7PYccRV(_lbrAtvgzJuzaLnU-8g&$WKx!t`)58{evFjM%9pM#gS@4Z?5Xo2YE`3rf~poPei>AmA!`^q -)QI4fNr)Smz{o+KWm`Uw8mcG6`3fr=9Z1NhBMt8nBd=_R+BH@w74xYea@M=2DIdvk|l -oWG9M}PIW5Z?7O4mGLFHb{Q7q~KI9ce2-m -qS$uljIApTMi*dphRQubY9q#WFetSymWswkEerwGcTNh&ZfJY2}HXDuhJHq7VTjJUx#IIg-L7!;`)Tt -p=;T<@^G74nchdrvAJ853sSE-jw#Y+p$e8ddSpp#w^u?sQVe1Am`^We_WG;>AC^OqqADXY2G=boPkWM+yDYagxqWbGkh -tW{qHeAG$sh&?mQ3<=Tb%D_@{>X44AJiMkI#rA=RyHC^J9w{4q8vJ8Kfb+Ue`cZ;(Mi~_S5?;^rxdny4jINn6kd^|AUw8N@{{llOU=T*g$S_Ug>AYY#Qs5Olo+L8dEcE^wr)d59I+=hW|~&swq6B$kXa4T{f1L-YPe -gD?f7TxA!hRwSSbN9=O_1==`xZV<4~+DOLP<1yA>Mwdo||AY`j`gv};gz2s)en9br!pt$W+LIcWEJXW -JjWoyJUXEnoU1B-M+WH~5E0>-KddI=N0;iR)dn^mZFRC)L|u#?xy3(HJlN--~fSo{&K5Du$i~W8UMEUqeu`GtR^h+2W!z_J^8mdbY`W+mn-G`1ym(x1)MIK#3 -m0lqbgZ}|3xbI6r^hGHB1|_!cVn!1FP>%J?k=N_uR11EmA?gVUbCMQ)>HYMLi2wGD6e#yw*j7BbLDL9 -3OP|X(nIj?tTl{}eO9KQH000080LuVbTY+Npkb(vP0R9pH03HAU0B~t=FJEbHbY*gGVQepMWpsCMa%( -SRVPj}zE^vA6SWS=HHW0o0R}jvja;wz>{kXVD(I#lx1&TIkQuMG`2$VEhF_A=-q}MUfAKy15^|kh*pU -3*Jw#DIa=FOWo^lrD?@87tJR#oA^xI@ -sn*1PPJigsoR!HZ<)!W936Z|JsIh=8LfIpA6%=7(I7tQWH0Y2T2<1KQyq;qJQ(`zK-Sr -;(-pPT?w`pHz-<+@0xYJNZN?DWtcFv2vW~iOQ?e0w$?BafQOewPL(WF2e_ZO}lvm22BHEntflQ~Y<>G -{OLqm&w>1Oxzk=QYnk$TI#fXB^dQ&BCH&3R+)VMt6&U_Yg@6IWk}LhE)P&jYV|aZezVGn0KPwzHRQDq -hJJTXDqKk7&V1+QTqlqopd?=VXjFHt|PIJiGl-_wYOay$cb5?kcer=BTTW#hsA-I-H}gkL8`ZRNkwhmr;-etPpQXER$25=U(| -Ws7(5M%FaZD!EA4-F|i2H-gC^3M^MK!(*e$289zA-h|WQU`fyer_H%D6iynHNKC{LojiGWPyta;L`@+ -%p^Z2i`Jj9ME^%ULWsTr8XN+j$ME`@>u#?UQ=bx|)q{6 -Bm2N2SVCOFF~u7|GuK{i6z#@<{355<+Dj$ZwfUw-}}jTF*{`^LqrLIZ(FY#h>s -mVgD`SEY}At7U~00q5(}>l-|R3Db%31XVgJ8`&${u|s>51KguMnKRU~*MK>8G#olVP5oi?LGm1f-XS9-srcAC -+|7ss@QCtzTJWPdIS~K%M?87Psu}e;C>e<2Tk}D-fCIC|3kfopz9$*10$GCoNMMFyH1JhMdOpm@#YaF -h5bR~NUhn=yY%^Y+Nr*6*oymxVKinp(Z95;^Ba7*f-EK~L0WG63JwHO2pPusT{oYm&)USg;Z<%7dHB-S9QGsZ0N;*2}zoMBGxL)%{< -M(EEq?E-;Y*%cLr?;bo2uG!P>oUQ)mtn~TG77ARF_qm^b8vVlxq=}vrAcy^!R@=&F%j0CewIn;Mp`C66eyYRW8*e^~pai8-Ir31%%FffY&dC@<>qDA^o|UU_8B{An -~3CD?two^8ad(VH841Ah@fc_V1FL!k?4$LGLzy08mQ<1QY-O00;of09jjik6iFU4*&p)F#rG|0001RX ->c!JX>N37a&BR4FK%UYcW-iQFJy0bZftL1WG--dom%U2+qe<`uD=3RCR3`V=_Gx8&{Q|gCAo{Id5!Js -eaLt?5D7`BDUwH!wv{yhdv_NP0wksQs+l+v2`sR?zkLGp`TTWtUltY5*+v#Zv3uF>Slv|HMkvK(nahl -~Ril#md_J3PnrhF|bknt6Bhr+~eO)yz<7HX3yp>g{W;63l?71wGJy79WVILONb}P%R{kfNW(H?4{?1M -__C9g@rBcC}tl-dtWuTud1@teCwS~5Y@U!0NLWfY4Y&t;>6fOmbZ5 -3JNP~|wA33&GM?>(9el+%;wLbxSNJeyQl&UEr7@?70FV{%V9C&QV6APc;$_Zj7#*`uvcA&8PEqPw*#q -zGo0qTOTwcYj;ZlioQ#Gj~Lx3j+Q)#yA%3C!>l1L?cshl|ifjw_iCBF!ykn^lfdDreVmjI}ksk)G@cx -Y!cG=0&qmGuD$;NcxTi_)~@dvGU<*=&}J4NDO!XtJr>7OoNDucb4e1v~oz98}Bx%RW@4Sk4%H+UB6Y; -HMGbrb-+`Q%Rm{K8t6h0vljHuMe$IH(C!MN8AG1R8eJYwIY9!V7vrkRR -i%2xPWk1_Q(M6Bh+!4+Ew>x3>ssF5|g9gYl^WKoGh3R-|hjQGtE94NuKJlQ&twizs1aMH4uSY310Zzy -XT_?qTm&NHuJCjh;JY07y5nX&c4_^5jS$tc}1HudUE0oA+KPOioEn$Tacv`YG&uXfT!%zG1I+5pW`m*bPP`L$qKFpk2j^NZKK%5 -tUxWbYQo0rUiNhXx)=tXmJ98oJI|8`j#OtQ>=J|7A-_hTmzhHl2&!4^c>xs2J621(lw@9ybFe9@G!kg -yO$rQN=sEj5K))o=$0cQvyeoVxGJ5T^9sL!0^f!1*8b2Y$g1*16LmKX1(D!7i?D2`?G#kai -sDExYcnkMOTF(HnLxq7;J3N38T%GnM{FZZYRu5sGz)nvsWg7zLp?YB{AVT6)Gx~)V$p#Yp%k5*d0S2t -@`ceJ=JL98ErmhFpqad&)f-~J|X%4lc5;mVeT`7WX#&kQX|JpVO8N{si6F++e0XfJs>8iijw#G|41u! --zO#rmiY~gpJ1QHx`kJ4kIomr+(d8t$H59}?12*C-c5&8X9$DiiI^iiNHV+xV+5a{1y; -9A=7Jafvbt^rmbc)3ODe6Li;3BWLnQ@1e9g04kfcsbU@(;tK4Q@CfT7<3^*n@%Cm+fV+;0#$l1I`rt% -0|o6aa8hRkw_{w9s)dW6BPL*P?|)hi#!cRoYRGVP?E$n8;atdgFSeO_}PH#($Tw$E^1iH(*JUg68ZPSBgUjF0`n2pPT) -w`hH)W6utB?ERh%3=9X1(-(?hg|#A>+y(%INmF4EZ3Ce05aG1mp6Dl}v~{N -qblLY;@C^n)KNF{ilzM~JW&(^W*RLT%H*9b~F1S|JKc^RH@{q_#*B^+KC^lry>#8b70UaOGkOcRD4jc -7Dh(ZO?n{XrSAIyzk^o!m%pdFo75WSdMPfN`Ot=|`P#PS4X&=aU7!ikG;peMS7Rx*R!@}L``-}elM3x -ZZ&3zP^jb^vdS3nCHXlcx9WK5+YAL{lZ!4NdMt32_@eJF(HW>BRW(CydYVZm7T{K1T7cJ#he -Ef(BsNi7j@$>O&T;)@agK=!Ha2CXd1gZDnK=w7{_gOP(>uLf-Nuv+sr+YJOyW(~4(H8L9b4OS@`cWGH -KL^rJMYBM-y`a7j>@aN{{Ccr7gEeNMll$l`QpLZ2yD6XE`LztJfaXTsiDTX!Or7O)gM;;NKY3yPYI&% -_ev55-Wnq^}V%sfa@Xn!5EXqt>?jVTmSMFEiQ1!fu-p_E8K16hO?YkB6HE%Hei=w-_~%=sO?Knxfn3| -EOxelUU|g`O){uaM(goou!l)_f6qk?B8`EF-&*_`4aGmWLNS>8PQq!SnTko!_tL3+#3K*{wM6GM=ytU4}Xn@spd|5}wBRzNET^b+sFa<%% -uglB1Bf)Iz2Af|F1Pidpc&w5#rUQ^FdAT5~|dT3{x}QylU>4neXKSV_+VO{rFp6)>@c~jN3@X}%Yge)d~9&6LeVQfb8d7Q5?`N>F2HldvczPAHNKcAmc+ -DOw@HiQ~09@g)$mrnFloj+AEGi$M@`C=ZA==m+(BPO?a@7xQpJ2z5D*B~uBY{HiAjWH7nS+&3sC@D48H;2hlGxcQ+i^8P5Xv)1q1FN{-JZH&mPxL -SxnpeJfDb5DSL0zGh!}@JjW9v4qC=#v6c0539f*0yhGUa(-I2A}$`T=zjA&vCoZ8tdU$lv)on}=ous- -gO>yu9}FRw3NzDr+y{BZr|!}W+LNYAG`*5CjNu5=nJB^derOv!G3nozI(kk#2}^kZ_8>9y?P<=_N79> -tBAeY_f;W7-eq39{@wAN~Yf2wJujJ{a3lwz2Rx8!Wl?y3p@+ev6Di0w8?mqMJQMdz& -Au;WKROB>xhLR=EPM_jXx@X(br}kM(-p;m+1Yu>)1Bl-4l&eqCDd#>V(}dG5A0be6%7+w*f^fN&zkayi06mXBhwxmMyp1GRrA83^HI^%eSaT{jcYg7?s!YVGth$ -FgYYG9tLVhbm624Y??=ESp~^qDhW?9Ze`nDra0U|UYm+iBoJLe^TqS6^_j~9UF?c!!(#@!ZUGD(XrY? -2b=XtJ-J1$L&81)QK!tEg@@N{>(!y1nAgq}M*bhkY<l@^P}ikI9Q_l>9aooT((63L{%^)YIYD^ -DLbJNESr6Bpu4(}7MI%scPISp)tR$p6;5082n>!A#s<(P7NOpN>n&F7lwWiaL-W6>#RDqlnv^FB1d@U -c)3Xxmb5o>C4rvq}yW_QO>^2hL*6j(V9V3VOr;OENjp5?PuVCsQ4NSC%6lQEUyJ_0WVz)TTLF*VWy{< -)pv~IkYdn%Qg|^9o3-mIx0XiIB*gr|%S+Xv1Rb?9OAW3-?Sp%5ej9H%I;r*8NSnoK{_~q_Euj@%|1lN -o>g%_~^qMA)nsID3xTfZ8DzAlga;D_7Xw4oAO=?RLE3E`I(+x*W6 -ut*up0nC;jaBmwJ2~a_`XkR4JqnI5B!=AUD~5LV{jIlxYv=pZh;@WWV$Y@_4G_II|vXUX>c0dzE+_GN -3f44)DW|^=XDoDcd9oB3$>BkW*ly9+@lL}olViHT{P7`JDM5+m>bhi+boZ}65LgaFwqx5WaV#h+(YsV?m{$!^>f0YkM)F -ISN3p~+O+7!2RS6spj)jdPMN_*g(kVrOGrf5Ff{X8MNjDRvRa;>=XlQt6y7BJ>I-DnJgJ)PxhT(mxjM -%==f8^eId1&?=T7EAozoRQi`l+T^qj4}5DJ!1A3S6^-$Vq+C2i(ZkFUNSsn5Ti`BH$4x~?E=psU)Lqh -K~%=5(M7NzzEJ{H~G^eXFMF0F}8%$+daF_UV4&`d}_v?5ox~j2+OUJfb`U;on2q8&nz~3n*)#2yWPNv -WlC&q3Ppi5K5eY>^(`6QNPZx6FX?ghmA{1Z3^=f2MK2M(q4h@PY`;A9>F&h#mM~kZXLSe!zM`3`S8-h -s3+Wqd>clLe_!G4t%*|2yM-KuRjsLLr*C#obU#hct-G4 -nuj2PBw5Coahsg2}9n=c=A5cpJ1QY-O00;of09jj8A5$OQ2><}z9RL6$0001RX>c!JX>N37a&BR4FK% -UYcW-iQFKl6Yd0%&EWo2wGaCwzkUvJwu5`WL9AhZuF@5w^4kG}Qda%pl6uGbW4iUn!}0xeNC8(PwdQW -Ixyzx&M$DUp)pB=tiqi}R1encvJXK@fc4&wKXu`V%X^i<-T^yZaOb!O2NpmphgudE2zLND?M@RarNT7 -e(3dMwW#>IY|<(6s#@TDnQ0tUTnmz_-`w8^QF*jg)AZqA{Jy)Hwns29=AfMq^(tOZ?&wY7HQLIFuxKd -@+Rtj)N-?J3~^d21+UP6n6J~0>tCV}HE+tgqbeHl)F`>`2<)33FZf1Ssz&UpTq@DgeB?Wk-3k1`>-`n -1Sh{^Bi%my$UhK{5XF1^hC(pjh3b`U?AM6GzS;fhBQDkKuS3G^-8xR+_ja0f`oh(os8cvZ -){G82yN;^yPs<;T0^;{BJ8|4u$%f4z*@t9#E%A -OcTP#kFQH1Ydahl!^+io-Ixod@5i7#!TdFA}B+mDc}qiqce#ZEH4;%A!2ru3x9hvHvj#=okk@j)UAL# -Di~+JA6dEnhLLZAY%ADC!f`V)hJCsH5T6kH-BxPWf==2exCJ)Kx`1F!$;v|QSuUD%OHy#-Bt55}B#s1 -O4scmuEO&7bv2U$zm?+A2vt{6GVzG9C=(RvDFu6^^nK6mE&Y5mOS8oEZ1p`=Mz0;$YPK0E@j_7!;NQ4 -~xew59YAtFHn-2fh;m=OHlyL`sDX3x3tD}H`rS73B4cs3_DAd+EL>OKjLdkP^kjoaRq@Qw75CqoPI5Y -b=01`b|i01gV&;YP$N&um_=QhW+^b;q1&f|h&<~g6Cd_WKgD7h)hT6k*Db4Y{^nwZ5VC -Y~ffABiO?9s{_ILv9t{Auco8CtL3s=2L-3?u@EUyiMnC_-L{lh0ppV8o9j>UG(8x$%j5HcD$-YhS9-$ -cf0EG)}U4VSq{P$#r&*|H;Trb&f`|XDoCQ}(MBu8WNx193CG7{lBD#Zht5J(TGU{!MVMD(HEU~v|3Wq -r@qjy5Q>i2xfRiLak|YWHQ_w=?ui!?+2+I-;IJ3Z$gh)C>b&^O3^RX<1i~cRdF)&!UG5$2^f={B>OAD -qm7Z@W@4Uy6*?uWV{xLpEtsD+Pw~<1$yrB^@V7gXrbMxU>|df2$g0?GVi -jDNn3#VclHh(eVxBu8mv&Vu#65jr6JkOlL%6 -GKvl!Fh^hX-;vcu|q20V1t=E+WxR!tZu;|u`p8xKpc$smTpD*kU-}Go}s0G0H9THA3d%)Fo+(EgALP5O_j%h;-IRt8m>!!ZFib_isOTfghOx}CHG-#g{00_1X=4VDP#{I~#fKhFQ$4G!c6t3s7MFmT))eUcVwu<2A9HI-W0<=a_6e -&a<Zo!DxYPGj=5&$s@dQ9(N}0_#SIbM~`YzhNSiZ5 --E%QZT8Pu`BA)UgYiJHoh(8U%KkPq -s8cf1ZS6dbjNti@X^bM>VrcEQJZG~P!pMXT2^)mckENv_y&d-tCnTbQjo>oRwP+}YspW$B>H_+d41(u -jvTg}*0jLCrai-LgdhKEOLbv(I_*$2$ZA4Gx#q?*efI4?l51a~`itmmT)MgS36}2qZuAciBcDi7n>DZ -vhx2A-s7k0;6)yyEaK@`*5vQ-%d@$ayozk34C7VM&}YbdHt=bU(K)E8nyCizUiF4!gP)9ExEt!^RL_C -3pVS{z0M9fR>e=X)-wp5up&R|qw( -8N3Z>uLC^t+aA_S3Uckw>l0(sf(W*Xt6LxGf(C1Ff=KHP}^IN$~kG6z>l?T-j={u9AIhcXlTK2q6OX# -Ko{MMoE%K>;~D7KCNXFr|B0G(bKJj(r*wf=IuoUNDBf@(eOH4=S4zDG0L8ttzQ37Tq4bEkzxPvq;s@j -4gAjE6WJdb~C^eH_@^?ul!?xj#!|0YM_YD7YED9b -^v7d2iN`a)NPG=!^NwXt>FF9DpHNf^e^L2RKnmczLM%9FDeS_M<~AU$gVfrY&jwbEDufqSA@&n7l^Pz -ix`HXx%L4LdGjT#Tely0-iL(A~xn34>>p8s#x^6q*g_NHrUxS?wCw?v=z -$M+q-*~hpuiNSkuD4HX4kNCewiXNuX|popcT?k}W@lnIQ+GF}Q~79aFD!w%^B9g?)X17(f>khL5Qx?a -(CQl%r)9PGlB*wdY0L}I$lOej%Rfo{i><-LfYSUIqtcVcj@HfEgrsG -Apq?>4f5Z&J^HKwt5@LF&5DDmm6LtveGT^Qyd&InfEfCJIU(%b&&C2HJzM}7@6?-MfQvW8g0OE$$TDh -D0%fEc{U9o`$D&xSHMT{SBHMk#&_E3G7sT{V4sxSBUBIw(?iANhA9{(`#~LEdiKfB|JiKki{fF)CJzG -w*80U^q@CD|kdIW58;Dw(!Uh%fChJ0ihf#%Fx;o+`^pnjX@)K2tHKKn2ccEuN0lx51`(?1u>HSwxobP -EQ%-vRwizh-sCR4OEQj$8139kAy;mWCeL&=eb?1cZ~FE@3pi8NYUo?legI&v)GrFwrT54^JEGF=omRzRTz>|dKGCx$g -h~!d6qg0gJtaxCB_(vsFiDQ>?RTaFn;1*)dA7xeqBhT|jqh*n8SQsv=vMNLvGP%LUU}Tz=G)av{rcG? -POoFPEiLy=0Dvio4YZ_!$f6I9s7qAc8@Vm-V`#s54E5P4AZsbOkyIj;`v-(?}@N$_I8~d!*h(P^RE%G -9ZoK@7W{#q4DA{Rl<3ngsZmKR7krS~0WNg^WJWxyAa-TZ5j3ZPh4OxPQq=P-W4Zjhj9B=kf;7RhhM`y -&Q_xBq>85uX3;;{2ES+t(9nI!kx@%Tk-OAetGfg?EU3!cztzs`Saz~`7aZOK))T(-)?WOyQn#B -pMkUi=1*!k*M)$C74bXWJz?*J%HbTvXCSDus3PQZ)euaHIZR5GEajtkDWkGK-yA^z#7;H0AWXiHM*};cEO%aJ;Xnj&yfq`cLthoxyMC~8Gf>WzC0)EF+*gt$DRXuu9$s`Wp; -6of9!uFFanlMsC=4n45m{xNLf+({P&|ZTUUba=tlynKfL7G{z-dv)Pi>Gqb^bheP^6-YId?jkFgADve -H+^Alkv;9Yld;Xt9*t1LL+8w@`T*GSlG^+*O!)?E4d}lc5`e2h!Vnl+R(T@sKujl$KCIfjBaL|nl%kI -N`}+}b%qqfo;c7o}B(QIL7~uejMtaP -EU?bPP<(|MPZkF3nFMb4uBs)M=0=4wtT0|MglVl`r0V!v%FmkB^YRQ=IAW;$P39A2`KCwv^KWT=@wN= -Ctjs7@E8utDg{8!=V<>+u~7kIufUuNP%bqP;2of_GYAw5D1-{uMGa$z0USg_rm;kBhyfHJ=lBZ1aEe7 -hNx+1(698G-#o*vLORLQSq~HLb@+_|sUf4A!3`Qq{;~9g}f&sYp@ScH$fLt69d8FYC)5yvUnd?!;LT} -69i})c9;OQmV7O1nxfJ%s;rcYV$B{+&W{d#vjr_EJ>l_bd71K%;Jak8s{BS^Zq5NouPj{@;Q;%-p*7C -d@b6}SeF*K()=7G245T`f@S9_727AAue@l1hOv9liMBr=LdW^YG&Buk&|TZ{J+Jy$ydodpAG(`SRihg -vU2*boK`d0Jy^fEEG70*y{cKet!8fq|f({vfj;8{&P9ghtSooT7r` -np3R5Z$P5yXnAT3115lB~Xz+nd-a_86N8XJ}(xiDbN`f(-{qGkCOh(c!l5xEzge9gXK3WTKH$Fk5Or* -~Lr+VFJeqb4QmuL5%lv?JTRkGb$D{;LjJf~2TIa!ULSe^T-fF*_O4QNf;z5G+|*>T?34Gk8oKd&u?(q -oD8sNsbU0L1>>2dTLYHSX62|2O$F5i;vVv1A&c8i0ee*P1#2ecLZNNIC|8W#uJ1V$1D6aX=dj;AT{>e -z$wrzcC>yFRcnI^+&K$_qT1DDMV5qxfaQaJJ1{i{tW%~vhZ+Jb+Zzl42r7K|coYdO3vAC5XWe~k_uB5J&+o -s9OkUsxzh^S_TWcZf`A=gKnJ1K&-a^eP-$u~Cch3@xPKFao!CqGwr~ms8c@Em4$7BB@ye!`n*Vs`Y&q -b-WpwoO0ELrvBjT=Pvr2t|uzeP5x$I--)TIHf5ML9Er0^BT!p?z#b-L_!jNDpfENj-GfYCs`jLz`)|& -)jY!#Q0_rbM`c4Pj@HxlV1NHGWX4SI0_8D%GkS~4D4&%hA|%Xzfo-V&qor#x;PScI`9LN_+_l%+;Mzc -u3b9Q3K6joesL$Zj-w5L7*h9p_w}Lcpk%7nwsm)W-&MTMQjPtA30F653p5C_Bw$QrAbpjP(VWf{Krq@Mt1uk`oN8D{RZe}Mtgx8On(K~-5T7_?h1Ealviya^~f5C$uhWeZBh -qgc8DY3pHT;HmV=+IlF-#S@ElxSU%Z9zKK%9K-Oc>!t)2$^ta9>ZNQPkOp8StL4yeZwG@wj9+Z)%M)P -ZP101WWD6mW*NkOk|~fns%%qai7g8*JAy0xrccJpv{jTnr-prFK};9e7$?f$sZ2Ga$k@fIqM$mk9(i3 -WCzKOm@aHg9ly1DT`fOv1&fagD;3fjR0y31w2<2GY8Ji?aQn8w-Z*%jR0t~ -}HKOx_5L}ovVaGe*j)@Y9676oFr@I$d0W37G%(XcUVKK7Ro;OAZeWFS=H0#6SnI3r(yf_U`nE%r -v`Z4<9bg^Kk$EPQU$1e_#e;WT!`s~BNWQb106NI5r$$keQIzW+eoCxM?cwh4}h};LU@d|^IQfDg*a~b0VG#H3Iq}f)PxEW_H1O -)t$_PoFp}@@$SEEt_WOW)P7&VlQB$4yB0AR8ot|HeF?KCYZM54U&%!$HK$M_WezQ4J67oNSw@Qc2_ -Ji9$>luR@GGwP)tM0nLtBR3fIKGN7aXu7{c>v8|-sMl^-=Kd4=h*wdcyg)O)-&%EsAC0NUku&@V;%bv -CpL#(>0)yfOk(&A5gc4-l7}HBd89s=enjsC1!+;@Ylz{lMOH9DMx-mB$zH@(KL=z)Vh#}(!BM2In4Kn -5;W7@m?XvUYI9ye3UD*(xev726_gr;nnWWW)BntUrgO?2Zuw32+x=$~dTP){y*2J)Y{kYSZ(NF)k?jX -i&MGP!%9Vj@`z@ITEx+vjIcC~oh0oTv?Cf&%}Un=tncwJ>Qt9yv5rISjN=9ETWtP;r)(3Q=jmsMF))b -~nI6h0{Q80QZawf>K&I@T+gp?5X^#n~S2z=sFBq+CUDhcBU+$wGB4a%~7SuEckg>3N^oC+2VH*m6Hwx -Py;q|0&-vC&6}VFH=x+~s$)t9$neAQ5B(v=4$mkK_5rnVU=#ZTgZUOlHkkTL@1q0d)0BM#2q2Oo)%w~ -YQqS0JvRE!9Y=7onNgHVf0Tn)x9ySqfkmV<+I#!!fOJxfb-k*ZE5aJ#HTzwB6tS}~|=H}eMy9L$FJiJ -^v0;dnW>i$mGg2#B>IUBlL?0XU)doB3qool=fZvg@Dv|Tj&u}f>lrASxh+MBYP2AJ2{GwG76^6-rzC( -N;w?GOLd@mlsBG5W4G#Xv6>3vb*X8zq?%BuILqZg7Q^*cH1+wPvI-K59hL18svr{*hI*VRH#)6zm)MB --1*BM^7jgJdR1QK72rbJJM4=d_a2ZyJD2eVpUOw7slg7HM2L`fkq8ID1amNO;s}->>6<7mKQN_s!n8r -Q(EfVOu9fe7ifV_?t_&=f1=WE9*m23Ktl;81Ezr+g9+0u8gkei94@-HFbI>g&<2)Vb2_=x?j1p$$k1S -6?aW-N!LI3cj9|g0F_RC*r%6}Vln+Bp(}v+byY!u7fEqeZox8|=`v1G{9)LnCFaUL_#K7U=YC-6L)&Y -@yYQ|MLMxjBCuLVDD_Ch_*-RHhNHJAN>99}%uJtkG$XqLyp0SXMdO( -Y2o^mY*3O!g*yfa9+4Sz`7G*3Et1;nnwj3kQdIGv|9-t=N>i`M2Fk -5PePpX!TcfI_33Q_%ExF0tMCliIDoVzM7=4OCB*(FFvyKqJhusz3p7fY_r02OecN3%=&A?Bx0Bk3arj -S;5)kKZO_iq@|WFJ;!@k#Oi0K-pAjgt0%q6Ambyq6#YO=(-!~^%oqY}Z~t9P`?!!*0;f0B&hh6|$+S} -^)bxTAshRRKuJ*V1*!#7pIYbv@M2#er3e9Hl|HLdm)3F0BIKz)|k2pir-%sEFr^ITCIsLj0%`_Oo)KM -AOleV4w1`FIVCPsD(imedtj2%04!9%iy8MMo$Y$n8V5rANeAMc2CGv;3T$ut&ESnQ-sDa_~My?=>&V6 -Rs3Z;FBWri;I>QLB?XVk4VMP)aOPRSEXeHF!&jqS)UAgNyb34Khls<96WR0ecG$vC6@0kk#O&&ccFE2 -fClGcSBF-!@xC-xDvX+Lp)m7O+(2uV4X!GsIWJIkV0ty2y=n6i0gtDn*<2lQpA)KnX_`FqT#-KDqysuZ|^{~c8qdMAQi^x-pVA=UOdC56;sLtTlfWL930R&J3(>GU1H9IDT&O2X -u9bRYX(|ts|9<;Hhddj;|oXgjSGQT^rYd);Nc3l_Becy(lWM8Zn9ie-0)EzTEou%djpT%G*SEIrjy8+h3!NI+G|D; -}ZRG4sD!NM&h6b+YIUw1P1)g47fod1P@1z~|TPO7-0JM}(E$t+i#;}VP(V^bLpYJ#GDcldbka6 -P&Uz9rTdDRDKHfujmS&uI{x835oqS72iY9O@x0(`m4eoROW<7I7V-Z?M{|2m4QauoKk@ -$*b(C>7gqFpT-SV52+zt=Q_Qh$uo?+Pu~>yyd>Owq{*fbm;d49M+o+n)QGO+6XlJ?UAV+JrcQ1LQH{s -8x6*UJ*u{E7r!AW)6319LUX5to$IqF`-S(@=m20*f3l;i>K2_~TOtm+(CrS+(k8>L+f0q_gjmdM0xe1 -*YD3cTgUI8E@JbQ&?(kWFzQusCN1b+#Q>}ndOAjvy(1|f-L{s7;{_~gOQ~MSPQ+yijy?H$g`FjM-e6k -H6^`uYZHZ=VjPJp#yK1^}sjowDmd@->PGal5eZC7;Yy)Ubqz}{@cj8v;PX5TaKs7}2Os -dgu`J+SQlmA(S>xXZ)NHV3FRu+eYd?9>}@3ie|7Ilth5y`caD9OF3RMNF?#X*ZT)1e99+w%4AWe1Ue; -FVL>mt`Z(YV;v0_HMf)XE5xlg!zMHJ-chj`Xw$sAHI?h6BsSMy0V9 -u+}WLK@6DJ>e2I*z)96Lf&O0pncwP!1%tod{zet34kwv2KBe{L4je=7B>bHo%?n$-jD5g3EqTk=fS_} -H(V{xIy!YLEbScc8}cBVn^+V*Pw)@EkCUZ8n)GSpSSx;NB`FnjkIS;>UKqZ+rl}R|r9YMev|EBgm*#oj!{R67yTg^t;m&o#wVjoUF)xLUs%hgnb1TqQ__&H^yA7Nhs#bdy;Vz2gB`I>Vz&Dv2@7X& -b4jU-(r)Z!$VGy+-_V=rh=lN9K$P=oL?Y?3}LKycdKs)|_~6=s>kF{?&ZoHufW7O`+d}rgN9Xh=M~~S -0UQp4Cw;{0-GN`!FRkNM3Op-Vt-s#y>RUzOb#1t8uPRx2?0uY<2auS+RetuIX`vFdHavb_LQ5 -A3B=uFegrE`%Paf#*~26S;JLUmLjDRg$eF9ZLzl5wN6A;MIm~b=(eUdNins~swB3>1Q?XJ;{~UEv>&I -^k2_F#D(H`bO#2TZujHhXxK?<5Q;-5m!=Mk1TD3fk01E)&Y$JW -6iz+^~rbv_eGLW&?2%+hr9w3wk8#H^F0YBCr~Li{Ty)z>!3X1ona%;WegRx*{aDt*E0C6V%@6aWA -K2ms3fSzBtT^GMJI008?B0015U003}la4%nJZggdGZeeUMZe?_LZ*prdb#!TLb1rasrB~Z-+eQ$5*H; -Xb2U~?C$nsr?3%Icx2Q8c;b`YRJAS}tDv=zl=c9&L^!2jNvU0y^=j+;~uGQG>0IdeaXwOT@`k_1C5Rc -eMJ2th70r7gV16t%EQcRDe@mdR{5eZVwQdZ_Vpj>ekdQbm}UQsf<)jI`jUgiI~!R3tfzt)<~y@2~ULQJsKN~SXcxLjgt4Tv;?oBR8}V2&a}ZD55OL1{UYWFc9Dh$4(2Sd)q?MMn|b -**uOd$P{MMI!*LvJPwaf#z*6GadHp@r>6&}y%LrT5`RNVz( -w52CMX`TmJ*3dvADH!dH`HzgqCt?DJoTzehdkzr3ggp%@~mlhvZ2Vea<>ZMQmB0!0fBcifY*bMX~Dz7 -Ll&)+^L~xkg3XZD;9KHhlD}z)ldV(w1v0&7pT_6OvtqFjt$yE13nA&w>7`+3dVA2SXEXQWuGdYbyNk2 -g%awAJu1by%6#Za2ot%a=CD3$q)HP!VQ>g=qcV*1#497QNTK0q0Dh5%OTx#A^sMGGoYOaQ7tshP!qLb -jx%y-RUCc<&TIGz7Z3weqz_HRBwMTFJyjHPilM!yjle7d#Qk+HK3}xfVxJ(Uu%09vy@`+Q-iCvKJ9;r -lbmw;25siN8`OiNC(3p7upVdv)h^6L8C-2@^T+K8 -qRfw|`9U-u!({V&mfz_yrC|y$%Dv^5&=2UWaTkhdA!04ROSV(y?r1cEm&KJYgQ-i(OE?~lY8_ -kI-*3Q+yWiRx9^q9YsV2dD1K+&iM;F>$C+>tq=<%^zDeH9#2MF5nlh987FSI+qvwj8|0c%>^2sAg;6P -#kdDb%kv?*%{f%9bn+r{O$L7sh*19@eed2X{X;)BwZ+D99JEqv<=Yg)0{I(e-2B^|b%~Im#(Bkx -d>2q+-8#GYT~P5i3DFAu6mH3{1O=`@=O!@EcSk1JRl?&37>m!aE$rxJO@G|PSz%*Sq=v($;+AA(kvGe -V_e9F;h?L0KZ`E*);lkw0mSE}5<|}?l1#HHjS9`3Dm$up_#}l(+dAeP$!Z6o0opm)S@%np82(1p -HV&@E9<7tIc`xDuOeanh3eQRdhZhixYLeHOVT{~uW8Pb1zowwti*K)XY<=Dyxo#*^ILAhnv{s#s-+;u-LJY;+b|JaY?50VlT>ZXjR|5%Q(AD{h -WM_DRS&mbvmMqF=|`lSf_^o6!fU#Ka<@J&_$Gl9HiH}^Wn$@Z)*h(a}# -Idl8<6z(-7Xk>8Q+;vkIf83;b(wlz)P)h>@6aWAK2ms3fSzFV*bqmP?001Tj0018V003}la4%nJZggd -GZeeUMZe?_LZ*prdcx`NQaAPiTd3{w)kJB&^z2{e0kyc8ithrYz2UZJ -eRB5&@#p=!&>=jYXtO*nZjEv%81;yASUl-$fy#7Byu;IWUS#3a1=uf3)pu9>Kk6^W#n{1-ak3H%6Olw -JKy%3)W&@53mvSQ;7VpsBuCLXBsXnuu>alk%-133FW-`RYdS<@n%!^2<$;yw>{U85%vi=I&vxbUP3h1 -G2{m95sJR%N&JH1O;eU&qR7(?hOB$u=>gowHkL4^cm<55Q55BpTkA*s+;%YtnasKx1x`cbDD1}QXHfv -*WOu~~y9+0fXTg2yNan5u();hI3hhC|cmcs2mdCZ06e;nd+%crj%tQWOfr|t#;GTu^wr^B21%#G%fm=?96#=ucZih0mh&I -JH`vcp?hyz7LgeTC-(}S9{-t@7(l-(3$0cY?P#s1Doy&!7D4nGUr&WtrgWXHyYFQJvIx9dZ%qU9&2-- -G#m$uT{m;|dQ99LevhM9-aMu_WGFx_84>e#?!(T?+a5ewqK1O%=gt;7@V=zv8$k~3l<7jS#6FHYlw&R -EGc0-^&;Kq`dqmV%GB6R*g<;KQK8mDK`KoQGz0n5yVl2?e1{sK@-0|XQR000O8%K%wh000000ssI200 -000Bme*aaA|NaUukZ1WpZv|Y%gzcWpZJ3X>V?GFJE72ZfSI1UoLQY0{~D<0|XQR000O8%K%whixY~tN -e2J`1{VMTApigXaA|NaUukZ1WpZv|Y%gzcWpZJ3X>V?GFJowBV{0yOd7W6>ZrnByec!Jj)(9d2iax-F -`_Nq4pt%%DkUqEzfh$qFtR+$*V3U#bFYuW!JIVZYb%#_Hw^6!4x%8738 -IXzxEyT`-NejOvm?rs~(3zF|DWftf#yy2dXqve#am7X?ov|qBc@?+P_q+8EdE9+z;TGf0Jzp{5+*yhv -p0W^oI7V`7jRnoV!yGd?>412#r+b6cbQVD)!IF73-NW!PdKZ?IBS^4?*MuJru+qr-J2&q(;N%Q@xR@h -va=`4SHOVnhGg+ZI>zmOAl$fWlCetovx45V|-c;Nl=1zWy7^md_UR*dV!wauqu92wNinMHW157oeVZ4kVGtSgV=68w8**Rsk -6wyFRZo4TV~1MqBxJOjTVH`xrH{L-Sy1^P0C7#HI*yCfaBejRzR1LYUw{So{-Qqky*v+YH~=14@FMk^O=!RL -{0tSc23XPY{~5V%!0Sq*ja?RyppgG@ug!<)OoxEGJKXKA~$5q_dmS4ulETA^E@z)SkxDYg$UeNO(DGD -3NASIx-DTfB8BsDhV9b<1t>)PBXg*y< -Wx2yM{$LH*#A#Co)|n{BZ!zkA-oJ!AmKMlnR)O^kAVZaqX}WwR>wsb&K({@eG -@mBOY2oo{6J@0?IgSnBTuc_zhU17M)~NPsCv0kCMWO%RCZUdbCKNT8!cb|R4Sjw%6I%(m8uWq2&HWsi -LX6!1uI3mevgu*K=;tbb7B6kx@m1B*B?McC&tjIfR&oGpQMNn+^gqa3@Y;s+rWy9^!i`WmpYbI2zDLe -CMHdHl7pMG%$G418Q8yajAKdjx^ZHFkClsj@7Kn_sr6VcatStFopb+_-mS&p=b{JR-n}h8So+P+0MqC -R74ny6|A#q1%j#(t9>p?E^Ww%Bm9rgvC%cK3-F5{8M9LhvgK-6M@_VcxM=0KzC`O!uN7C=O(MPYMf_S -4gd-SyeCfZ3zIlwBl0=g9y4YT2{L>cZj4s4lErT2X+_{z#sxqytYUPT~*r2V -|&7uwioVaXYvO8JXhJ`Fs%~($)K(>;eXS@_HVc*P7I+2vNA_h4L#(3POF- -V&{FJFl{Jd(P+mbd-mQ5H0OZi&aIyZ~hnIlu|G1i0!PO{UFgI8;|^}M=@xHR_IdSEqgIh-7wpwL&cz{ -D`LD+U#)nZCk!=j*dBaU7pD=$Q^3<2b35M5%jo9V6M+0*q78x?{nD+_3vAV7NS1y{fEb75KO3Oe_^dW -@7qwo)odlguVSV?ceLnNWbW&#`XBRn`k+md;&GUfQq%s?H$O%VrHg$a!_eR@9I+>#QBB(1QWOO7798(~`!Q^>xD -MY8R!Ry>)6@TM2Dc47Y6wv!-*IycUlpqc_By9OaozB^LIzax~hd;NRTDYffatC$hLp}_Wb#)Z -GK|+bOu1~XC{zK@esCA;HQXeMT22uISM?@x^K}T3*QI|_q`*_r_Ue|hHQ2+#|vu(5X!b-6$Z=eU7oKs -29w$1Wxk~P`9vkxFl+xU;p&p5`8q-E_nZNb3=SXXYtRaQ549EN=GO@v80PdI8DGZwVKgv$Aq)9#7k%J -z1Q$`;>`*z=+3$=EaJK=WMA*)xcMJzuU%W9>S`Lb4u?zBAHoU#Z8JFtVXvC!U@;6a&1|#v<<9i_qLxA -e}W{1+1o5G?XhHTKiMu&~P=Yb#J{pBX}nrA#X6^TjV?GFJ^LOWqM^UaCyxdU2o*J@qK>PuGa_G5V#WMRZc6aBxQTg^?%O -{KO{;kY4>_9P~vdbQp4eVb7rVq@I6VAyitu{NkWQ!&4nVgs<@)6;FVk~>|DyrlBH(OPL=%Op{O3*Pc9 -cZ;`1P5x=kI@D_$vDR7@<%-x&O%io@eO1y9?OnWY(3G%cx=OuA+5)RN={D>DPA4mIrTU~cH1WgnUv(3 -sbCQOC(6t1>R)nx>y9(iu0ZC|kTYTwmwhy|J`F_${ -8hA!*>lQjEfcrsFEqR2`#r5PkU}LDElVx(YazH8L$POyqBJ@LlKr)iN)+1$7cR*IQwggc3&E>BtJD}0 -!TuGeK!p-b3JPec2^R!LVdzKzD#67G%L;|+S&qLhL7&S-G_{e2D@xXyz^}jja((k5xxF4+q16*-g#hD -88c`-YR@ST|(R)K>QnUK|@4rp#q+mG{tV+QE&?=cA8)u#|Nabfe7baS9;~u#GS;ej3`Tw`>UHP6W}giSBTzh9{n$A(zFC@$_5b&4-)KxYS -%Vf+Wgy=Q{+U&`otzNu%j-g~%77&azXv8AYL2700GmlHW|k$H=LXWqzc1NM_qc3dBWD}r!%eiZ-n38V -Wu*VlxzKb0ONPF6yArgGQX6s_#!`&Jw)9M@x-k%T4-Z+l`5)+btrtf-f6#fAxxV&N)u* -{IAK5EAX_9CMC};iHZrLl1vlf>r9!zAI20U=0icD3X -ACNg&N`2&Rdm+&85vFm`v?JX$5e2m;Yxa`MDT#Q-bN^NF#VC;<_f@SHp$;X5i}YuL$d?o-DNZ-BtCha -d$0(FZVUB&|@Fl1N;wr`oj^S7Jp3QGhAu0`M3{-Y!F$igag)C)DT2vhC3Xiu2Gs6pv^Ogr;c&j>X6m6 -_q~i@&O`R*Nj4dbP^=0*fSaXR;lA}&+Xb^fh_oEHAFB9h!8R?d)A7E?{FpPfk^+TJ>hsN0D#x53VUlJ -0{U??{S|HxJWed?2`FpJ;pht-kho|xF$RR%Q_pbN?HnBZr9w%hJq#{_bJn%+ldw4E-{J^XB3PaQwelo ->)`I*zw0TiwE8yvqJPW=!4=W&L&nq~r&jYUoI}Z>N7YPoz+HrZF6aY0(nUt%HflX$RBpkkg+ZB5TkN# -3_;P_bOjAB-OV!^bbjlYS;eFBH}o{h0iamxI`i8MU=OHmz#xmlVAz%k9T5RpyqhU0+QyWVsXtUcqV{C -?-&?Dg+c0cKE@K_WRHH*Z4sB|A=u0WB*c=K{9`hGpASMNi -=TsVIpoGYOoNXTIxmws7VzCX`CU~J%Vrw}!BQiPoD~0ytWDXaJ9Uuk7?ZCVUUexF7g%r15*GXh>8Y-A -y<@RG`gK`#_eS1B%9B!vHzb(HrV|{bGc6%!Fc}jhKuaXv??|5q^^eytwHLkqY)Qc>3o(d -Xu!o1udtIb6Eafmy#FX>zw?E?}edI-ULP`J+*_lnaqG{l-vGSy!r(+g2>^+1%V>%hd$FSPK-mku9nZF -I{h(^4>+9p+uqI8>Xx-u0w$KhFgYYdKbpr!z1d!R@GbrK -awc;2BHJ)4Tz-*i5g8kv9T`))VF#DFlu$@H98RE)u%YllJK)^QZpl!w|=ms$;Bhs%%h5e8tIYk`NF_* -e{r~=m!0Q(-f -?zE|q8a?VNPpsy&`Pc&3E9xXRZ}u`AWL+4_AQdu*$dBZ^V@00h);%|VBPZEL=bmuHX|u;vGya_yRVcZ -DR3XF;fT1`VaV}3o#_pi9ydWRoM9QJ;=1+LG_oS#51t%GhyFFlc)OHltLDzut&CIkBG$lRjiyvv(Fc9 -&Qd}4=9N%z}~lIJyf*7g*`-b{pkri>jB4?{eDnCMnukIiv^^1Q}0Mw1-7?F#$0x&v$b=ejAjK5As^wj -%sATPbnq_QVBlH5{Dy(>3E=LzC6RhP>bID%k@~55!-CLojT1kvw?OC2rssEIJzCm1f4ZMQno)`N+n?QYFAE3GHHJz6^99RD5L -?buCKo(d)b^m4q<5o#Js+Xc5Mxr{Nu#T6J$3gQUe1l>yf8%|=1tjmquE1|d8oTR74o3Il=g3vU(N3|co{M1(CFZPV3X2YCRt#e2qFm-GM#=lr8Pd2Ia4*tQu&zsn -m=K3@F>Dh7tMqT8;rUo-@e-h=uf%@>#baAis9PdawG*=T6*{ri*NYH#DsF(w?&G&gs7a6ud<7l=$BvE -t5Gbt)!!Nx~OP?Q0a=`?R&1vT3zC)GYKg5XJ5SXmK1_O!?_SYYj`0CLSZLhL@!0=8+$KM$GF!*ZKcY& -M!_>uSX&mnuTdek8#`i4K{k&^ZyIa$DQUALh!gY0+Z^UJq!w?Nl}pMn|z=fg5Vyz2+o&y-bDz>8%^4<4TU@F$nbs?uOb -DebczYsan^KsJ#SK5Y+*-Ep#XKKf@cDF9cN=s>uk%^>lLR7v?hWVGD&yQat_iPX&@^~KFx#n?A?XAA+ -*R$Oh8ji4TW?1ZXZ;lJS3j}z7{hs{r-M6Is>_w7VFbP{PO&+U`7Mer+tPe%P4r=r;pXT3Ch>)9&i>G4 -bQoK*p{Ly!0T%`SADQW8({5?eQmq9*WU0y2oD#z4!xAO9KQH000080LuVbTa4r$p_Lr~0P=7E03!eZ0B~t=FJEbHbY -*gGVQepNaAk5~bZKvHb1!gmWpH6~WiD`e-CcQe8%K8kpP!-wmP$YZj@ElsmRrb1*0EAWl2tObPRiwi8 -$g2?YcMmO9*}|+fA_qj@96<4dpDH|t1J_kzTfqGM-L~H$=gn~vQt_}u_<-mm8(NvHnrF`ohWPF%c_$2 -#bh!$J3ALwI|Yxbs<|!eYq4(jdsX*B@0vrk5i2D+wQp|JMo9Rw*(=dDB|I-{(PPI$f9O;J!|adk@lxP -)>;9r?aRON_p36$Bvu)Sx#d67wmrGIZ+otPq=@3XJHz+uee1k`(0J8?9;dKksrVk@HEW#SO4|)+49MYXHWj)<(t2rnXYA9EN -@i3X}Y47>ksl8Iuy7YK7_YNW+WrkUz;y -+aDSiTCn@T82-SK>MYk{jyU(9!j7p41MQ_h3N$Jw?}Nhcp9|JjiWxURU0j9dM4oKDa^g@o9_6&t6oUG -dtd8Ai_g_&P7UzO*QMUee!cVGysv>GwQch<;Nxdq*L3{#Yt>66@ymJ*^TGmQkNo6~>TjFw!;=QaSSP* -yw$_7%%yL^Q>y7$osI8V;{MdQkp>Lnlm$}Y+vo=R&sa1G}b8)TurEL|FFFRbbQI!U2;b|@pU1i68+^u -)zje1#cn-l%2vi@Luyn^3PG`-y^p<6{{M0>ecf8BK)a!IVx~=4Ki3*)JfolR0wR_r?8^z3lpE|XXutRz1cf7?wC!XACul6m -7rlb57e0BRTW$MU&N97eheA{B^{u!~%2+)_`c0f%>Q3p<71XEcO4l4ZGp{?T}`Qh^549%vqPD -YXEwB;GW&;!WVsZQ1YQ125?RIO#m-J@Z`#n<)&TtlrkZ+#SXFi}xG|bF!5um|3Sg^0{(z2iaih9bqsty}DF_HLgg-wVV1)-%BR${~&27FRYHHZTUeN>v|IQ0XK*@s -X1_-E=4sK+BYS%UwpoJDz9r5V!%Pd&#!u^Ju{&$aA>0j|2lm_nswH@u1C84Pi?<^5M)3@TkX+^L-S{N -%zl<#HRfYmJPvR^KzT2Sl4(9SmEaBaNh6n;Kcl+?Gg+(i2z6s#9`vbj`s*fM=ch3yTOpUHYeFnJp(8{)&#lo%ML -%C2N(i8_CZ+8fk7$Zgytp85z+}Nf5wHGSb+!O&HAW2as|g7w4i4`xFq6K{;uHDlGhRyHBB>h6Pm*@IvDwAIe%Wo9tCe^tHVGtjO!nou3tu8 -6YNG@vujzl?W@VQGdN@XLbiv44^%uH1bZfH_0AAq5fsqO(C@jfRiM*FIo)AW4y9E41dc=Y$TOy+@;>G!mF!I6g=6rzVjc3Sd~wvHVN=Eh6Fr*iM6eQYOgRzHB1n5@Z@Gb6Sk=z}sN0*62g5d05H{O!rtxN -z{3Z+GR|+m~C4yEYx1h3raDd9)~z1gK`j`x2wJcueMm4e!>#%hsbuD2j+vTlOSzFEs*a&!ag|JrC%Ug -{FS{#J3(}8i0(5*v_Ut0+t8rlH!|LB}3Xcx45md1*EO8*&?<;sIPGG&$!g-|!D28Mxunz}lMa$|-wV`$RFN0^<*1Lo$5NXG( ->?SQAiF;f_AS(n(}U|$NOIf8^Vq$dNO*^;_~e+pJto@waI1WE+_La`4MaL8E4laSh}rxOZ^AKl8XUV= -kQ;h?<{c#aVuRtEHIG?v$OPGLPEuo2El0RnKCZ_4JNs|fuPE0IIKZaPHYv?>l(BQ~+$Qaey7+vWhu!k -h&H8ZyVoo*rmek#xURFw6$nNG+Q+)ABu86N1#Jl@@G%#03PzHz&n8IVMu89dbd@rW>Mo?QlhpxGd{2GJV^Nr-rF9iLpJ -eDZ+KOspiEsY1L{7lAAnlX-T?Xk)`Bv*?=!!;YlQN6sT9cJMjs0Yb`whalEx;%8kNP{X8S(2P=yv~mh -Nn1}=0s2}B}c>C(zZ~pYf@5Hok(6rV^L8e{-=F)zl=&$?T{$kVtfnTtuy^jgc&oBKO= -sLKVyN)c=Qv2Q$7kxRa^#X3;!?-{9f9KuYb(>s>?W2H4o&T$hNe~G+y07gWUzQz69H^(uY2KlRn8|Oz -NSo<@&gU)i%`)E-LN@-yil09`{{|YjJ%ufhxd63NS_C*vn`=%Q=OiHRw(ATm0BD5j1eW=bG50&mtyO+ -;nNvOAAY@hB{&8Ul>vw&QSvqu$dcd_8#_C&O5M=gmyRrx441j^HLAf%Y&+XK>+y#iw>w;;B{r8wi_FL_=UAf*cLL@2!lb -a%wMRk@gGZk;V>4F7;_&Gs>EfLUU7Y6bBhU_Ox&$Nw5l_FrJxBLAo#>{&0Ap2t7s8N~ompfpt_kue+8YR55Ubl|ilPV8epl*XD2`58Bu^2rb#*|NFQZ9J4MBCBrG&$WvgVbn8#ox++yID|ni -N;PR@1m83C3&)$%C=txK-Re>Ng}*Jkp?d;o@@BS-Rq}&7_q57ckfpK@mRj$#?pBHaQE=YHE099(JCIe -_Grvz483_687w5aU>-0>GG{pO5Z7oeddTG$#G*W}3 -%4<%p2*_&1mfIlWF%$|a>KV;v#E4){vo$I?tBxbJ1~wH$De6seZopMqv66Hly)c?sj48$>TElaEgO(Z --thM4mkcmFAAqUZFFvdv_*`Dxnl2JA`yBH^B*>`kbnJ6HxDc+QGHZ!Y^gg%*nxXwcfMB{(k%!0qi$_5xGggJ3(Fo5+HZ?V*r5RWt9KPI)zUqlCiDKPC$?FNC?R -TWDJZ15BTzEPj;xwo-Gu3LbG-NNU6ggZ=N{y$VL@B<0O$8WWL1?N&tgk{qug?7z$~E>YN!sqYR{-ocq -Glj4sVdmXqmun>)z(N6c`w+PEX&v4mptr4jCq*HQ?=o7-ct?H=*p1tplML|7|UE>)MZV0!ps45Lmvc3 -S!usu|K@D?d)Iyj$7iMo37VyYVGAj_J8z3e%u`w;7KI4s#R+}w;R$|w;+2RcH_=>W^zM{DtDCyJh@5p -8Ulv#>52Cwesbg#?wm=72_kk^mBPHF-=?py3>4?9j4HQ3!FxVuyMkG`d45J -h_IA=nxfNqI(O$mWCc=>hJI$8qR -Q#G~G@cW4c|xiK4L>?13&Rlo_pUJubqg04aXG+r~Bt<{ -XOF^2Ft?>wfm0gM~9I%MMV6MI46f{+!hpK{JzzenDSVeBMAkVNE9N72L8o>JYvezDRdUtH)YD5m)AC3 -7L)+h=dZAOG;L!rfOO%P?n#++rwcMG4D>VmQJS?PG0%a~lV2#y&{GE)X-zNwOcH*%Nvf`<6T -tLY==Oh92XEAD%~;yJ|hAnp1eR5a^V`_if}mevbLgaKnC$-W1nIA#Y|wS6frD9CzCo3 -Xjhp5j2Xl#8A>jj>QpzFI~p1%fXmT?*apUZvc;h}J_|ephG%i{Y#Ro_aV$}XDw_7q{4V>*XTAoSqPrB7)hN@?qM!8H|cQOls3Q&g72@dji$h5Vh!{Gva@d -H-rn1gDxsnfH=WDd)e;g6S#pI+FIBcdommSKwcNwlNi&xG;*k0xV3E&*w#nzyd%&zxJ@~YC|}qaDMmJI98xh0(Q|hPhenx#rxQF66Mk}>A@a)^vB5?!Mt$x -ySKh>pn5wj1=9gM0PLz22e#!=2paXZYRT3S6#a4nPqr4Xt-mO&L0dt;hoEW;>b|gh&2ntTrDAAdm23f -BH5EBEd2OU?q5VzB%XDZ^x`_R4OE}1%wp>iGvUS8jttQg&{oF)-X+la(z!EK9(wZbUr=bB2kkwY^%)P -p>+aW->`5HF#cdSdvs&}~)r)5$pq$`{d1i$N)mtvVUr<(#r;;l;f;zdkJM8NVrdf*&Riv?sFZ0QvxiA -*Bf8V3-;2=bR=YSFxc-t9lGBc6}+MdtiD@WKG*DD%*yR`qJHI|8@vIzl#PKhN*!L -FoQHGqkhtR^7b840HtZq2xd6Ti(88~~6jS(ecwX6S*$O4$U_zuFnohE2xMr}21xH<|b^+rm*j}qX818 -$59T>}Ax`_1maNb-B+Jf3_JFKjTrZ6FANgjC^sDV?pPJqT&H8hw^pepMx5o-8S6MsC^cG)+}?IC18lY -{1a9U0E!oU|PT=|mvDKEcV6>;&i^R}jl)eo5cj#^14H;;FL2&4o2JF%mB5`E+F1;)6OO0CtQri?+oj# -#GzNxOX|jZN^ic(k)MYMIZCb17ZWpT=C#`>{Pq_X!L0AEa)7f_UGcAS~negC26T4m5gGzoJuEVI$;gTSoUUeke_gA@nnr -@d4=O6v`^oK%T4B>3jmy}fU}}>4~6=Zv{`8dGN|2jaqncvFd22*m`jeB+B>u`Twhl9LJREqGq!jx9&Z -MOPYTI%FyC*2c94wesyz78CKm?a{j;cD9Er_9FUu!*Y07!4V?`he#**>`2;LGXQ&^=T`I -0GTu}I;0e9D2^mdiO*m}t<4HGW4jWMdz{$e4V-qH~G&F;bE*k(lowW_6-xj3quZg>A9HxQCF`>)5OE3 -ID7xxQ*~q88_ESSBMS9)IFb0rliPSn8%+xN(9aY*XWm9V~i_xBWqt|ZwpmiWOAtBLecSXC2AKcxX2kX -LNKgv3_@D2*yV(afX>^IfTi^#<8r>Jiz?OC#LaLvv}ANoR^Xc_>Y*|TrN@Y}fxV8%cVP#SVCJ2v4V4jGJeFAOm(C&`zV# -mmCw2TPJY;!skpL_7Aqtiv@JD~*L_b0d_J%W915doOp%(a}s_bAUPd^xHe3N#`ahFB*ok)4j8jiw3AD1GHcFLc}c%aaq*c)> -L7Nr#LGVjZu_@z|>-*|_To_=-dGqL2;^w=keA -JNWbz_s1v1wH&c);2NHZno;2^PfZ*YWo?VJyYgq>X*f{>ZQb{-gIax3w*B8Ealw^9Ys%h(YjZ^?o$x -_^BdfAE-|Wt13GLW^a)_%6fpJ;=!^n--~8QUsIGji+t25c9l*9r$&mMqBia-=szULl^__-Lg!yIjUfh -TkO_`c<;mb>*7;N0PM%Pq3=?S>(*Fb>es*FHa)v4K5h2)4xZM5|EbC_ifA8PD5}%MTKcA9YL!}hi{8Q#c6FJ{TwUdsw&bgAls(4Te7CTF9EHhoa!8Z@!+q}VOer2!gz{6xjuC7-RO_Mlknpb=N@;_fc -yQJ4g>B~otuFHORSm9mJNA0ndkJ_?*RBC-t`q39({OR|<``>|Hmav}(zr2v#1+0j43;*)$sqgCEm|E= -A$4z;yFejWrB;Vklh-eQLI$UmHq++lT5pcqWr~%q;RaT&!hY-Z1|A_i{5e`e080-!QR(ja~!~tOhU;+ -^Ew}1HK;uC)H^CVOWc(n_aConHg)7g|F=5>Q+RQuT7D4)wF1!!tFvbCD!HlRXU{glh3CkO9WX~MmCy` -(!BJyRf@yp?FqeM~mI#~8Xw3l2<2>Xt_^i^(|u;@rF=pC< -gvlAqQ3y+K-hrxIdmDCTs -eWMlR1H<5zKFd=)(b^Et3Q(c$=-UHguM0Jh#y5OJSdXLM4Z=Xjw1o;wobbC_+fT&WR9A?h&?oJX*CN- --K11S>j|8x?4ED`Q_lg^Y~F?}5IQ08&I&02NS%LeZ#?LLG9pAuaH2x5-C3(>%Sxs{ -+SrSEM=%a@4^B}n-h2-X*q%kT&r&eX9UZ&?J{T)le+YNxiT^U$lhwr1$0TU2b+??Rpv@(FVQ_agKp=jse`u&FM -}ymtjJ?5e;bG6hMo|yq3QUY+sWvu6*6(*eEt{g>RfzE#OM7wD7OAoCmLx!z%w#?M^St7SOLhNWd*O_t -4HHuGzmVM`@pve-`F{08mQ<1QY-O00;of09jiA00002000000000g0001RX>c!JX>N37a&BR4FK=*Va -$$67Z*FrhVs&Y3WG`P|X>MtBUtcb8c>@4YO9KQH000080LuVbTdQvEu#N@*0HYEB051Rl0B~t=FJEbH -bY*gGVQepNaAk5~bZKvHb1!0bX>4RKVs&Y3WM6c0VPk7$axQRrtytS`+c*$?_g4@uij@mj55+z#3}ms -rq>G?Mu}D+wLr^HRMB8j+QYEP*-fsVW&kS{ql{OD+AW1FG^>F6Qs2gYZBFi>)sGZ6(q4$+_K}ch4kU? -AHXS3#6ZcR{kq15a4$xiy+;A?5OTWz-OSLWJ-cu+S8JJHU* -PF13RJpEGmQ|Ok!9-CSNg#NA7oj!if8fbx%ygbr}oN(Ddjro(x*Y~tE|wjbBgb8pp_%0!)!*`C?`(Zw -9>5#@A+AhWk&9S)?zlB{n5jcHvim9BOx{`)JD`*A%n`4xyhQ-tT|4eJ`LgHTv{4{AWIeX@v|a5nDHf4T4Pk}Q$&m_z -*9NvpeDRYAN{hg0XQdxV*Zbo)RCE=aVEK4Lox^Pp8(#B4QcsT>?V5$DR|@|3VnZ|StX4}j3$M%pR#c( -n>WXu}+ik)w~^9I5%KEXP{(DO5R)DkEEZb*v)2SLhM#x#3K%lCke)tYKh4N$n5S;mE{2kjMHz? -spsZ7Qhb3E4kM)ujue1o=volvv%A#)X>TNud~0hSD((_zsWAnU%q~KdOnY(f2+f8KzZruoem!ZjKi(; -A`f>xNVo~M4*bdC7`?_X6yO*H)bcCBtrH?%2)`DCQ`+6N+ZB7+H)&mlkS -WZPyKw!87*d=4Mu7Re!>ETQj^FS5bmJ%2=|_sC@#_Ho)rldQq-8w=bw -!JTC4IPbb+k(=0Jf!g8o#_-rzB8S#GaExX1^}o$+CGo#|8KG4r4)i6Dh9&^03G$tN&c(091e;8nSyRH -X1K)D@pJqa}?P{Jm4AH-H+boT{W##S-dm+?rS*_+uk2Ik2hb{)F9#tE*wqDQA@|uC7vXhIob+gx}d)F -W78eJFjnm*Ah0hkA*L_hj&C6NfU#(iTna00Mk~Uy7pSE?SAcr+OJhXup+QMDk0alMpE!EiNzV$$RZnT -4jT26WeLF)E2E!g<6a<%G;6wn{@go3CQf%_rYnc##7rNIoZ|w1?G)Of7+1q8R&F1|Vwgkhr%zz|A|uMB~%imzP -SKx=M!V(Y+lq;;E(xuGL}Hw=W5EM@C(l -1tsB-Nt4lz6Gy$jEDCC$V|8MlF-i9AL7yiWPl35#3qjc -a3YowsDvu=hMPumzn4`;K2;l^23n=x??oncXdWH-1UF0QE?S&I2BIE-5n+9&fmY@Yu%pS&pC3|BM_?F -1eMtHmg|OGHhq4UyRceEO1S&QpJ{r|?qp)Xf>N2lyK~*&FqL^!YhkdZ5+0n||EvZ&axQzha5JnL^z6) -NK6FI&@EE4PP)L-d_b+tu|DF$z+rFz|!`94DLg)#+(?~`x*Qk+LrIT!8nGDR(<%6)r@*9K`qHBwD!&A --m44CH222d$22yb!Tmw&rUv>CULxA0;wtr+a(VP?c~d{bjq(Z>uG)37`(KJjtdsOvH9OmYR93_w(de! -08LHcX0lyTqb{cYbWAiv&?B%{{T=+0|XQR000O8%K%wh)&*O!T>=0AjRgPzDgXcgaA|NaUukZ1WpZv|Y%gzcWpZJ3X> -V?GFJg6RY-BHOWprU=VRT_GaCwE4O>f&U42JLi6@+(~4Y&q&7zPB$4&Azs!-it(-B2thW@}4=B&W^4A -0@?3+%}yqwm$fhdPpkI^IzzN6kfm%6W39GTJTNi%J|p-hs7G04Ef)n!AO*KKxA5`9r#g!Qf6$MX)8IRtgO?{MIo+?G+{N#--}u1T;JId%WSEc1fRL6jqqJM7_xRoG_Hm54t{IK;5%`tfJ5eb4AdP6*z#In -AKK?k#8@=Ko-)|*c9Yyac8Gn&tp<zG|;1jo -OJqx7LWB_l}Xp!H>AXA!;sVe4PuDu@mGB0P%L@d!b>xL$|DAK1qkqpDz!*U^v@Vb!N>SK;3$^RiM&U0 -HXRMSiFR^x+kfZ^w5I98Cui*G_4RKZDn*}WMOn+U -u9%zbYWs_WiD`eeN;_v+b|5h`&SU&VK(4;JM|DCJ9O(l4jYPXcSE+ASgkAxlAJdGev}ls@!HfETNe39 -zDKew%b(~OXP&_x6`|3!6dy() -o+9M4q?Wp&biVy-XRM1gW&->#L~EEIlVOnWjM4ZnJn!I_CK~+6E_HH{P-0-+0fQa=3*3h}Is-8 -l4YSV*c68mPaE~LR{E#(nX!zifW>2g49!7Sc*!m1;pgtGGK;+hF=R_~0@W+&)ABBT3edLcWAm9@Xak~ -JdW+f)x=-O(rj_TaS)Pwjr$7dz0W#(J2Cs~(^YrT9WW)o{?RzC -mqHxH@lxdIg-e$6Lz+m;($SFEg@CIJ~$=Dnl6(pWSKHwx@RuaV4RKcW7m0Y%XwleNx+w(=ZTy_g9P}t&~W`5)TNeQh8W*5l9GD`vaESOj7SUcCg2k3ix+ -?X|ibx_{Fz5Gv}O{&e;Kk=mG_YLVy}Z>j -@KKKm)%ZBV+smAmH{^)Y_Dms+D+;*}wEMrs~NY0rJw#zZ^xa|#!XK_?E1_QH=UUxxTEdt@w8LkF9QBk -z$u=eJr2BZsK0VT`rd!8TD5hdrVeBaUxgzpWX>m7><*$*uBrTga~eLA?YGNV1hAMXOxYY~68@p_I7%Y -2eknbb?>-#Ts1a(5k9(8Hg|N$#{x29`Rd<*Z~be@KF=c9^v8s0V3t10!|ijma{x!?4j{6O4=B|d!OlXvMoM(eL-g7OdtBtvo}MS=N{R_O+BaU1%&xd`cCjwlC&S6asbGn)O-D!avoBOeg -GO!p<4~Whyoo_th2P+WgJ+(>bhY!XG;v3GVY2&A;joI%i#7eUM6HTPEti -CmYroNQ4aH(N`08mQ<1QY-O00;of09jj}hMlxT0ssKU1pojr0001RX>c!JX>N37a&BR4FK=*Va$$67Z -*FrhVs&Y3WG{DUWo2w%Wn^h|VPb4$E^v8$RNHQxFc5v`D@I6>K-yY;%ELO%zTVPTu%Pd#YWZ7sFs_VGyrlayVZHONDKZ!539t(=E-h-w1aX!NLyW6DYK>;b(@sj%^vLajnt{ -hd@9$ZMSr+O<4L4K4q(l2P+6qb~>wTMLmSqjLP=&ha@qj2OHr%pGV3C3Ug5MsD)w_7CQtkVEfz`*PQtV#s{Z^5`axV27*Q;f6(et0x}o$QK@Op_M(bM_{<0gD* -9;Cs_bw&OWjpZ^*3l?l#nS6K2QJKvcY2bdQ|CxL-dxx)090sPQ$wI3%{0@`-3&pHdVpT87d=4(>T}U! -bej!4dYADiD?$6Smbv^^dzCvQ(!33DJcdokaz?qPni}P+XXg!~swP_t%o>}l$je}i2P(Oaab#LQzQER -2Qo5=6t-((c&Jd7&?71qodj8aaw`8LAH88ix$(hlq;g?W_lr^&ae2$XN%Y?#ZHB|U{#S&yXL$HfbHrS -Zf<{)uscO=@W+z^L?4*lWJaJmXmz5*{ykQd}07qhE-s%>DsTO9KQH000080LuVbTg?!!`ZNFl05Sjo0 -4x9i0B~t=FJEbHbY*gGVQepNaAk5~bZKvHb1!Lbb97;BY%gD5X>MtBUtcb8c~eqSa?3AL$jwhF%}Fg* -C`!#qEJ;mKD9KmI%quQQ%*n~jOIIjJOwLYBPc7C{Qc~gq08mQ<1QY-O00;of09jj>u5tpqApih0Z2$l -%0001RX>c!JX>N37a&BR4FK=*Va$$67Z*FrhX>N0LVQg$KcW7m0Y%Xwl%{^;#+eVV#^(!XO6z*7>Gw>cC_35Hb=7ogRW -&Lv+b%1LygXG|xm5E_UM%rPccxWzd8Ty{sd-*z%~sXf;ygRm*lnG4NfbqUd#k2et2AA0x=o|gROM@eo -RwwOWnEsC?cSa|TU14%7yQgVTGx4@VSM+b?DV^?$mjM+dttuvAMagJou0yE&CjY$>a066KO1enwP%|y -w_~(hI{_T)vdL=m`!Z`v0NC=n=2@%ne___NX^J*m>2!Yo3tcX%rRGN7vXX4RFs;AWrEapWYNqOO2jAv -}mT&azM8Dkt6oH}FSzaV--L{Bdf$$^$>nqRH#duI_|>{+ONXW_KT?sp)X4yHvl!ucbiW8OE!CPJz|Uu#X-;KYsWkefj) -{^vB~D4<0>u@gRNj;K$>sdPHqsHpLC_+Qpe(TYfls_?P1!kJG;%|NR7Z0KFRPOY})rc1>2OEPu@wK;5 -)k*XfEFtI>6lEpC9a&Re@BPr62D>z`n|*BZ81E!Xp4ts2>QE%9sDWQ$HOY2>u4(xqNyn*vw_u$R`wriC_ovFY5VgdLWGELf*`_->S_J -X_)cGgDYytC=nK_VyM<*0#hwvDx608PBg-q%h&CS*9pbXX-iqc!uH?-!FCB$s6D;UjA8Dp6caH%|TA> -NvoA=vkM)!x>!xr!8htjRcamqKQ#;w(?Bt<8+Bylc>=8a13imV+LLKI+1qn7fR-Yh;>bOXPyjZQJ=vL -h`S9d#&#Oyl+->SYzj7_9;gqG=E5H{0e*GF|f)j2uazktvH&t`k$LB2wh#9EM4glU&Uw^Ic-JumRaB7 -E8cyXW`pKDNwrHWZsEP(+|0cTTH=<>8Xn~?IwG1R6-PQo!s5T0fCzPuk#P`e_eGt&cxwOmS*^X4^SY| -{wPE@B`@k@{??wc@jc8^)7Kvefh_LOE96jV4KhlA?~IqljTPvI;YMcpb|MCW0?e&}m+!=b$@J+Zb7w& -_>8v=8KMT11e{Ln{&Or1d=G|xLL!_X;oLLx~X7$7Fz{>N|h;~B5FC!`>7p+(FW=}##?SE2lyR)iPxW(buA@(`!szZK -E&1I!8nhi+76)gmejMi+HaQge|=bwL8Aa9>Oe?-LAEx;}cRcgIN+|0Gw)SwjflG?-W(a!8i^49Xkg!d -m!4Dl!o6E*W_0^JlZ(rv_5-k?q`HE<8VOF&jG@ku>dQHb%hASc5%o6gZ|Cq6|bL06BQv9z0_p$nvnWi|P{O=|mlY -eCS}@NCqa$2~DQD0^!!iXe`VIc7`YF+@3~xH&x+p+2mlyw-qQgW7i$2t82DOHu>-&%0ik$N|V5Cj?+l -c>P@{DZS3(DaHIDsIc=&RirsY4o$cl}}x5dt3y_NNb^xk~La6;SOe5XD%A)z -=9X*RDmoq3nSR!yriefHqRUt)PSvwjB>>@^V;zNr85^fCyeM;h?H#(f2g&C0DZ1j`yt5&&c#f+dfUy+ -IJv;yWdt+-u!qWxbmfbk_Vzzff3;FfuFi_B5dlJM4*6{hnxKMw)L)hmh`tF -IxrsLK{CA=%{8`GQ4;Mk9;?;EX7TU&tJ> ->Dd=3sR&Qg9zE3^9l~|Ia>Kn&vgI=N;7ts^e$r}LmAB9xhRq0i2WiMF5i2$a&D5PC*#UJzGA0jl9X7` -iaEHIL6ESlIzh@0zqmKOt_$AIK)&PNbAhqF?UB^6lt`30I-99O}1BiKsNF) -5evM_{baFX2iiqANwG;W)6EU95aH!jdz?uK^+4`>EYkK$?)#P-WyvU{F8c1G-H{?d4>RZSFm;~XbaF^gj%#O=w0l=3a|q9bR2dkEqyms -B4lf&7_b6}tB6~wtFYs%JFl<7>5M#m)#@AIL}qPQ*SM9Vgr#{L^uP~pvE*0KP`#PHnjPE)J@T1~NEx6 -7)@uOo185>;0{jm+pGXnCfW30%=6f9oX?5$$5$gID7*P4*4A$;zk;EZFh6RVWqcPLf8WG@zj2^;{&6n -VZf>^|z3Mc{5M|ut;c0j0YwE_1;ub_Tw -oE>XMh2wf^`_GMKrvPG(@ufLmv2GP+&Rx@*bhdW_qv$pAGSB4Fo$TvzN%#-vhamc6qTxr3D=v+H+u8G -M|2r{y8C!Q^-S9h9kPA)6Z1MlL8_F1|VWSdHUjbrjYR@n!^l$Vd)IGkF16m$V1CSE}@0K`Tfn*V+n%9 -{J74x#E{^*>32135#U`gK!AxApufmPQv{5TK93uxJA`Y{P6RI#N2&|(LZrhrA2 -#TG|FV;&JyYmR44Nly~*XuR5=P@41iMVfsueVDWJrl3i -$>#87?F%)I3|9qk&c~(dyF&2h!}U>*{uPcz6nOY%@=Q28Z>w&JOFmKFr&8quawTzqu(a=RhIt432b;HbGuRY0pwcWVnbR>Cjc_s-5pYn1Es{zxoBFTRS{0`f_(HLXBV863PEk(ZPTo)Bnb(ycgu+WyHis}^8QWyw1Z;fU -%Dr=s+O^wdi>{HT(vAY5)`|yp3~KN&3>{;FuotHHQI6p!j36 -Tikj}O?o~WE+Y2S-gZB6Ufl4LPG#Ix~v!+n>=HcNr7&*1GTXKnGl9fjZc#P^k4Ujfbn={W-vHCy{Ld; -n^wR+OyCn#CD%glGj0NK~dVOCWw15ULjjF`#`6^(PhjH>A$W5VL)M1RhS*H|q1yci==#A{Gcxm1j;S2 -5f3pryXDz{%JtH838|2pSD1E1nQhy7~b1x7)u#)3iJSmb)8Cm+3RH2pVU8Px;nHjf*4=#kmH4X*a?>mCX2QV$h{=B1kp^tspXb8pZmWD{dh4(Hp;Qn?t`iadO~J^G? -!L~^b?(eXZ1t|srB=?&8pK@)ZaN7RL$V(*>5+K8PBzYr#Df{S(>MmD1fX}tCYSP-L4+gdN$SRT%L0dg -lgf^;GM{hq8~G#fxjQ27RzhuHAEqPQOjiG?ZEFsmmM5%$}Ku}Nfj3D78-Q{S|EMApxG(*aEhq;77t$` -mE|P>#$uun_8PEXu%ynt}3$HCWUL{1X^72OUO+AXpS+`+?xy)LhE~3$guVRIqF>IYZA%Z(l&6OGfqxT -H+LXY|&h-o9b8aJXGczwvuH&N11nDgZ0{dnd&M-K%w(S;cb%y60cc)zJQKwHVPvZrq6Z4zJxT(l{31e6sKqa -OG!PjoW%VPLK{6Uv}dB*>NtOp<`DztxKj!$jVuKnfH4B7fUU>eD$rK|@J -6V})q8?B}rfSYJ@dT{ffx+x1Uuua@kK)+4TC_5pf!({^fKe(8TZ9|9LU)89fRq9Vz=vnxlvwAbL -Ur^p2ZIlKBauQ(!w?kk#M6la~+nJDGRMj~#D3~B`-W)$UdHMYK+2d#F(`PRpKYeoY=8clL(%Ko?7g>` -N#o1{kI2i9~szM%_onwNfAo4w}5Ug@AS5y#3Qa~@xN|wARm_MBR=dw;nLuZv*^1I{xm#%O75NOF-E#$u -*cNZ0{WPb?X_+xH_{$vtWOz-k4+PW*S0A(cqrCL#r}^H()7XKEcKqUE{!Zjs}a5Od7)lHG2X1p%OJ#4 --Sk`dN5|(cwPXVJ;ESVIvk%(oKFCZLv{4QQ&`~~t}w48y8!UEswmHJq^{hot|M<=?}fLc4lf7<3hYK6nrDN5;G!i;myO4#8SR`63vG86= -qWd%4925FY@?2-`5zSXd+6(;NCmHWI=0y{BpId;;Yx<}atlteW0sr6>AbhjfyP@i@97++Ujn;M8)dBs -xv<$AkcfAZYR~#UTTm>lFri034;k6jryH+J;hJ@*xhi=DAV=lN3};ge9A -8=K-P+{#0Q^9V6px3TJ?TQ&#q@XDPH)RN0aY>-G!`V}CZ)9Q^|o1Ev!8m+LJqvuz6uLsgqoaN|34Yw; -Waz}PfJk?g-Nw_-5K{=WgdWbyPku+h~a{cZ7NDW^;eXrmt~Qc=LTB=|L -kL05JMLRpVcZrCt9OG6l!DufOoW@#B=4N6*}bQMjJx`h=mR3dQJApY+UERQFwD^k${+H51jI&RC+m?`UpNJYab0kJT-*bcCA{c2%9>61@6Od={=7&?M`kBMB*GExu8A!d4c1Z}ZY<(w_!c4g2B8mU08~Kr6$ -Au!0WR^9qbhLHwj~ZD%fnY3#M0%z6*~(<}F4=%)zab%?dijgR@DKomlotPhDQCH{VT-*`j2q!11$+9+ -wwckuOc4(X&csV9h!B*Wk%Luz|~dxkO7-$4%Vmx0}2{*YiYco1b*Yx46nYshZQnJ`58Aq?(FD8vB6lq -r-g=ga@Fu4vZ7OG8L#jAjtS=0=W_${J6YimbY3GRkUuecsdRyFosV?&e7nwTG-Qh0MuM+nh*+mulO`^ -2n)$Y6WqZVn3Sd76URc&^4dFX7kl3)HmhlZLvr%&R%_jHAv!9?0_{^4u_a}JY;L^LEykJU2-2HIzrm9 -k8;o9CWoF0%q8nAcT4imgo5V9x) -Qp0txe~w^DFi@-j_t4Ri#0$XRoGf%#Q1H26@&EiA^oH5rP0ueBV3kYBH62VaEsa$M4xck!n&mU>Lk7~ -eA?e9!*F>Xw0RJ>bWwgl*bc?`)3LfaB3BESn_YBbf^Pe8e8`k6N&8E5d|k8#J{uvs=x$NaV@H$Vp}OH -k+x3f4p;ds-FGu{LUZmerfApI3R@FQ=ZbONH$5=;!i$c0h-PF4*qlJV+hCL;l!C7Qsig#R*oz42{*tL -9^VikSaBw7Y6|>Dnz`eZGk1J4#Qap;thXDG_U1I5Q81Et4vO_f*qe~E)GbNknQ`z0lSb9b7>s1uy4ac -D-a`_33W6yTmK^t;YdZHB1R|xgK^mSL;WM^!{FADytcPISa<8_~hXPAaIbl#U7EieFWz({6%V)1-HKH -ppnmffwT6Qxh!#%(r{lFWdfB~i{z+gO(7hV=cg7$R&C_vCCp*?u8;>EcnsJx>MA;L&zlvLK^Dkr68l} -Dp3)`GVnfp5K}lf7;e-nOQ%2jUE@e5&-~6RLVCL?fw;xezR}&gY)Fq!htD_?wsW>w%~9)(2$R;~@o<= -A@4(5Yh`ZxT=Iz#Z@*oC3NJUJ&i4+O_g`_Ab;OEC{L71Zkdj6A(ti6mi26qWFIyKHG)k=L2<;v6x|u# -BHDwZy0GDPFF7q{C*~e%47%`1=%MSn2Xv -QaCp{n>pfp?n*PQUz&6h+=EjFr(YQlfyDDkOa_+>e9V@8uQp9}TmTd7}-499OX2D)SKP1FYNq2!Jr?6 -KjD0hX0fIolC>aoh3c1P(Azf2;pmI6?`yXwJAZ>0FauJDyC3^5_Ve2c>_DX;^~iBUYXBb_|$4|@fV7k -DLW4rjeHgZL3`%+9~DR;Y5zLgd4xX;~4_6x=>eCejM+U_yls%opuN>PAp6r!_`S#D@?AU_H~xoFVI%Z -4+r<*xh&aNpnp>ql{w<^cxHmsh>3A_8|z2inS&S%G20&-UH`rI8|AOskL_&NU(N^ej2vgHu4G6GZ*0AP~l+`E?>PD4!{?`sJ~(57w -zWZa5o7h{H&svCHVRpqh`t+1HE#a{(7dayaj@gNxqK|$q?8uoVZJr$yz-g-$@0J%dFwWJom~`=k*w}|Ejk;P|eP!>yGT(KbaQ$LrsQD`w()Sm(^vd -9P+I5(IcCi$+a4%(JG3mdXm&oNae|Z%#NPBfH^v2_1b3cpT}j&lc}F66OyA9?1F|Xi?0Sqy`KE+hWh5>wUZp|S5VtR -XbhsB=#$o2W+75Hk49#2(&kuAG+d}&fVf`zF!tSn$uB~mr>s4e4pi-A>5F3-$ftM=+UF{&}8Jvzj6xS --I8HWj=dN5h)VJRS{*cyAE@D(Fg)29Nk+nKsq>;m=WGd#~^EAxU$;1kahO9(vRnE+C|jv&Y94J3?JQ& -N#XQFk3!u!gyrkmy_YN*=zMk)`vA6S*AyQ+y!_QUQEBHBfuxFh@um(9MjRuVO@xSPnrbjhZm?h4EAna -^Xzypb2+$QE~a?Lbk9juDYc-RJxp$(cOi?th9zKD*D4iXKJI1BLN?u3$IqtlA4a8#w~n7lIs|K=Gw(J -I7eDZWYWqIQpfhG_UE!*?8%WIWNSy7uxng)c1D=|rO8(w}$@i}qqEM6o#y?>co$hSe0nM%d17P00^I< ->%TK2J1>>mp0xA%zf__U;(Q!M5W9J%_}HCWYUpn3I7m;m#i1N-@9{lO;MA!PU5bVs-Dh&1vTe`K7YS{ -gEWAFS;1-4Jo6F5`Sd_fd&5+M1(`QY@45MFedS`LbqTE@5PUA9^Mq7ClJ%qR4*R(!R`AfC28!*f61N? -oCzqSOM)$u-aNTI>?I`W&Fwej#B6r*3NoBHf67bnYZo@m(he9I@qgTKdC_M<0j;8clV-y<44aK2mw&w -Ypn5DCVdc79LxMqdm>7HOQ6jYwe@z9w$-Lth;}1Bj!D4UYhF2%Mj0IX_q9)Myii9UEH`}MUx#rPG+e&&@_2R*l3yRcS%@JG-WZRpjEYWS#C?|Cu4LxjoLnvlF$3@V -1F>K572=FpvAZgsZYChTE~!Dew@yie^Na>xoy`FjhI@{3p7M>irja-Y|Cq3Ny -$6&8r%`@e&7jjg*7UF&yTwJJtRB_tn3v--GMTO-=WIcv@G!uCo%>&{&hi;>M{S$r|6j&6h^cumY`TPT -T)8!MIJ$VebBg@S@iY+_8ddv7ziUkoSu-SZchYVCi%W3%*~$TEa9Avhd(Wn7Wr63{Y<1-P}dBCv>Q{4 -a^?Kp#8No?U}k_o9OYAlNS$u_~H0@4!6f9T9-?|<@c97uEJRgrQXanC(*y*?uPvT*vo}) -n{o`JmYJkuFEJhIe{P;g#9-q88e&k)0zVd1aJ@%DHAH+-+Y}CUO9cl<+$q|YWA9TI{&V_{l?D;ajQiZ -Rl^h2L?JvY5D_s-aRhNuf&`XP#D(OytAm -B-Qlo+f$xluN?0@MZ$*1FpCAmM>oO;YGBwnuv;g{JKU2t5sx_NAxFUGw)IQ6L)gFOux}-0CwRYG2cNX -vx=tvg1k-2jinFXjyQH>BK|{l_BZHAu77)*(h5^~84pTejsNb{qf?53Y78>*r0OJ4Ns0AAvKG98()3|s$yE4FEGNg+>A`9& -uo;Kltf*b7J6{MIvg-peXI0i2GY+5aHQkI0f4hAT|&?&!B&FLB;S-Vz4Qm|91zM=bwi$#Z1CjSMwHr= -zil!>*2YNAByenWfy3{T4Qw?EN=TO9KQH000080LuVbThGT>edq%K0O1P&0384T0B~t=FJEbHbY*gGV -QepQWpOWGUukY>bYEXCaCxm(ZL8Zh6#njCAyNi5)xLfl|V+gVL5TDV1fGD`|=%*jgn7w398EggT$OJQuz5lC9V5)J9j -ofdtL0(i5si_sX=z2L2t -PnEf=N^5V!xsKaZlbV3Dfs+ -;?yhcVqwOYA^oEJxa!2*;9~9g?@5p1dE&&XhZ8&mDWS_99LV4J_U|saxpNR*<^`XE3hPw-KLtEiDeIL -b|4c{p){6>yFRX)xo_}83pO}`^w3-jD@LQj~134)rru7j^3I8-T+(N!E+*@@Wus0UQtV<86J*RZ#XUU -2rWMsFSR0{wCH_&%`@!ctjy#Vn_c`tS0p`}&i1f-qPzzd({6}#zw*|WUn$-ErK%=!yw67G^uD8JUP*RB=xe3RI1k>uef_TTQVPfeR^%o5F@&&ib#Cxi=af*g -5ge%~%OdWqOq|x};@N4wlnv{QTCA}que?1xMrfnr${#juo&clM -78}zR4eSie^qhNuXhs^}%hTz@nn{sZd)wcWDBE!MD0vl!A;_E6!u2vqN(IEdMhtS -c1*^9g0mWNJ=l8u;?R>_|0Yq&siS0D_mgKP)z3N5f!-`AUS9+Ac7G2B3H+}D0{*-h5QeQhsKBet1-wj -o>7w%IanZ(j`KX7TnFN^wXx>1hr8LjTV#43@m|KkhPE;a|8yjKi|S8_^Z38QT`1!&<`Ob;7abpApigXaA|NaUukZ1WpZv|Y%g+UaW7+UZgX^Ubz^j -Ia&s9h?v{Z_NTQb_7=qljvfcmv-g^K@fTX -l?Qs12>$x0yb@bJDrzyv|?pFF+I%Pq_IRU!7GYs+>qr4pXjZQ -96M&1Ua+g5{auY$s~?rdYG%4g5zafuQyxW1Q{z2Y{OH+VYmk%`|wG|14%oF4$h)AzA=p&v=qW}SfOIxP}S8BEsMFptV0_G|@(^@o*s99TPqK56ujIZ+|Zw~YjAhBAgMne&_ -S4@Y4V{kawUhuNPJ=cP@3bq%868CiG`@GrZC9CoZpX)7Ijz=t0*jsL96LThrlWGah7qy!-5>(s)1-3{w$8E-g|3(z!Ex*>xcoGMQ4M48 -Dtu0UWoZXEA?Li6(73Ux_AML`ohqspt_8w#fznxfB&n_6&}x(9+u3kE~8S&{&bppPqdHDmAz4B83NAK)Zlh8jFXusAXS*xx;@kW228o7oI -?7|hYiDsv3y`;NxKBq{kG%t17hm8dz|S(w*Jc${>@@vS(hFq+L~nb@!d%yM2O;$u^D72?sC42B|h@dw -tlVC1jh*8+#wKiHpTDQ=d8-yWdX0+hN*cVQj;J&a#O;q}$k->+|OUR>Wq{~9cakP*$+;9&_uppksT_8 -Jm?h~%YAx_SF}Tg$cze;dU$VHif;;;--*P8mZ1g$#@PfsioJ7?a4u0yzsGNMa>G3>y_>cuv;=p|dImHSY}YiETE2 -W)i3+UB0gx(VS#R<8#TWna#qVD}OC?4?jkws>d=IXz=XMZYUEg0fEWWtuRpL3Q#2>@>d1TZZKY$LhE9 -yr!1>dR_AolgU*O!-iT#L*zmK%$V;Sy)$1wC3SX#$ch_#L&AZ2>`4anBD*(?L@T+&0*){fRJCx`2GtfpcKWR+N!6T~_3zO%_LYtvE$$Tg -A*fRaB8b3b6Vy;7_e;;_T7?x0t{xt5f*1o>4%)m3;tw(}Lj|5fWMQv{QAA?g{U*Z&LVMCUY=Q?k?c;- -9^7w$qd<;4(uDt~~M>rMwq|R}K_De#RVPx73!O6leI{RP@T2mifGxif4(lHtyXlv&&W}1l7T)V8)gt- -QapnwDF{g@W~{{kYCXzKtHY)}d@QmI3EhF%X~)=kXbBd_jr2v*iKM5d~32H+rk10b}XyQg7%%kwh!W* -emR#y(#7p}kiiStv6x{(59zkpP(`kAb(wKk)#H1>rq6!Vuzvg(Z;wVr0p9G@gS9{f!r`(1EEZ77)KA> -HxXg$A6Ak01kk#3t?uA6j*!CDAq*hifU-1?o9g#>ubrULz<&AC*v$E@Uc^d?fgImW|^Y{k0eHZ%`VOr -C_v}|hxFq##Jh&#FFk#|g&kiwf*gEnUg1)*@NRBW-S}lX -fRSyhdMPj4^lYjyASZuctF7ho!w2r6_;|^DVMpEbQvlu<~nJ{|%%OV#ajuX#~)S?T}92|#&p~ER!h9( -?LvjTITK-|{U67rQ2qjJx_0DvhTfcg#=KYASLC4OS&iwAf#Q1dki(ybS1S}yv$qa$?(A%*pW{$Hh}5C -l=82a;WOy6Sm4w1NFGFFT&k!==R=r`Uelh!3QICI$j -9F=o074C`K~UiY=*8WmQ2dj)#u;{RyFI~xX1Nq5YG0gTJW266&0$3OxCM(w9479io5zCzxE~!-tO0&j#-$v9ms0URU^#0^L|N6rvM1@@C#UXaN;y -H&;W6PO@eyt;20GmN_lQv_MfW~(2^JBgkPzd}1U@wgGKVXii33ZTdop`5sg#>nKM#vtq|HB*|y>zn4Y -vrfHc}ex95dVW~d((;LvNMmbaXFG!7snJ?Uy>sX2RxfaQ)10e3NzuKo#1!Cpgt%5@yF_R3y3HE$SO$k;A8?ugFGN4U! -4**M~ygruh5d2w5h#)jQ~?v9gEkv50$MZ5RRZ=bu$zqY3_@W=f3BALX -q&vuLR=Ny3{cX7q|<9|#K+;sf(kOgzXj2XQsPYKW1yi%S|=zQqmk64yuKa=StBGbzP#zU_JSZ2Cn5N= -$+G-P2Z7LG%_keaJd$kR{2n+spR>Bs;%wz_Wk;W<7w~>ipnH0(t0dv*ph!?w91J$# -t48QnZv0>3p^*NNMiN+1se24a+l3z5cHf{p;24T#po-rRFX1o-h1>CfDQ0NG0mY!iuI%Z7=v`~aNaXY -nkGjFtbOjt@>>cvJ++3HUzA?PxAoQZY24TIO3wi>_rZ`2$u&j_V}vgWBmF%sd$(4K4pqzSLLEqOlhTj -;^>a*&*HG(aSdNDBZ7)9~ZNX%l49YQ`Jg0u`ebC}az49g=FzFX}m!2yq9~I~Xl#h?V3whsifcZQjI3d -@WD`l;c&`$_pVT`B{{!_jL=JL_@BGzD1k3Gyoh25iI!dTDxr?xSJB+RKkd4V5E_ -(loxKXd;o2QPkKDUY$z7oB$aE7pKJ7%aSMZhB#A1ft`jX3xT>ljw4#Ys8pT~b~zH -h}Ema#7F1RsXer?2&Bc{pu%Kq{yK4!_(AUO%M0WP_$PXW0G6fCQ8^Qf`lL&ETXXlkX7l}W&!Zv$A>oC3Y{g}W}Jj@lgY+ee80=n9riLIFC@oLH-%bA`@#5hMW#XH(y -3|VZs53yKf@RgMR_=1qciG?4IiHY5hr-bPaKTH%x;1CP3Ff68M-*Ebdh-=G3)LZESN8Gy%F|X(QhzdT -wtV``eMlYc+9pMAbE6diuA;#Fe`72IbRr5l?j!AC|ZDE(uRy@AK=;<2D-VtqJ#5LwZK@a6vX$t!f+71 -*`_UZYtf)Q1k}fVWfQsjv%<*k^UYCejuax%Qt^n5~;=@S%Wv~DD~Xu+g-y<-c*WID8=i814n;Kh!S)VO?sC}lW9 -wyCDp;y-2jz6kWtD$Ej_a$_Kg*zE^M;Odxl2d#vAzJ_Q -nC4~IYHA;YGnULVkx5O9SCmXz`@;&3146kP8U5-=?pPX@GL$`O_OBa^BRk2y;KBl^!S^a^*+#Ow$3ls -ZPPizLGtZf0qMJMgTm*KLzC$bRylglO#!z^boFs`Q)xOE(prt%kQs22V#>%8^c5=>d}U;LkU0BuIuql -JkNk`#j!FCy$Oud*;R#$1pp(0PKY!I|%odDX!-bR|P4umny`1XYP#YiEpQw}8a0Ga=#xMqN8ob8K_Bb -PV-4k{TNJ&epwfgWGnCDd6MFg*{(%OLZzn(6NEXL7Y69a1 -Zh8Q`2Xl`g7Iwbx;o%7!xcoc*Com1~io>98>ZqU4O0x`ULCj!D=WMAL_a2W)@q4@Kc4!_+JRw&}xTua -2;RAFM~zbp^1*;?P~m!942@~%_wsHs=g9TS8Ip30%z%|kAYsW($|MISZl$k&W-hW>n5p=Koj4XDZ@gC -$NEk1K+yHInAOVsGQC@;V?I_fyn@%5%WOZHEBj1rx6C%qW5T@D3`UutQnkLUHGpKU17X^TA=SJs5ua@ -C#!V;BK}7LxTo~!{yXD>G^665_$kNRmeT1RO+orBbq3P%XFcdd7K5Ss#MA6#t-8L|+N0^Gz5W>&{0_S -k}k%g3SJvcBPqoBtm>AjK0GV~5QS%y1VDVx0a=E3}>^|5h2<)i9N#t+Z3#1zLb8{fqfl;G%!7po)Pg{ -f0$S;5ZDDi~qaMO2Wct@iC~;q{q9<-rl;2ev5O^)v&E@fyC?^_=P2K>?oqW`{OzH+8ogSbUo+_^ -07Llcf3)-(&4VzYm9_p$tJii|3nqH2QzhWXXjWl05ki#uTSkse{{Uw`-HljQ-eLD+G+v~>y4vhrBezW -yEx=Xxm1UclSR}2FXhrBA+HDF$&%{lDwJaJ%)Cs-r;4AtqQy;6tSd!2>Bat$jRyOh{9Q%E=1M4reZsk -L(hsl%)FCDyy3Od*>T5I-$O?#!j6i2XrFFihTP>7;quq^qwV5Tc7H+1C;^~W -zi94Q=kX^fY_{F5ukXVh+u?wDs*lX{o>zf-$?qJn(1v^$dYsY>`!d*GO*E(>w|_?fMgk?Nf?2Pf6?q+ -U&2OP>lGJi~f+yzr^Lw0Reg7I}Cih~Lz*-olPU+*F|T$jU+`-4L!{R)8IN8uX%GRf# -l;2A~@efmi=cETL}axq8j$a@Bm8z}}_<1~VzHv~;B-nQJhW6u$-m?8Tz$X2T&#aX*eTs -$V0WK3NSg?o3Xdp9Yo_Oz0J{Dt1Rb-l0~L$%8M&o|6F -zKz&dNFnNim|Ls_4x94!&ER*!bi-wxLm~lwzxKm2eVz0kZ;#LEa<;A2g+0#)jq@|5%C`*1UZw72Prsr -=mce}sw`;JIZSeXA(NTH*L?O+MP58f%RBOzvkRHoLu5LcBfaB-LJsnzd>axz1a$uYZ*358O#|cMtWHI -Z{xpe+lC=`l3o&D3_{%v~v-?5nNICnJ3lY6>suc%y|>8mQB81DO#-}772N43)-wFQdWg2!NTP5tG8UZ -}QcsL)&;V*C$DcjP>byJGaib1jkPMZ4$bm1B3VjsSW!dRm3NO}yUDluGp?=Bd`eVy1neSm=$ShUR|Kz3VXgS{ArLrYZ5mL^$}N-6~T_njd{*@|T>P_#b8v@{&PdGqE -mtr2gd3r(eMVI1Ovb4@Zt7m2 -|@vbhc&|7uQp-F1Grv!_M^%|+B+s3cjr|jA1yG|`#m+{QDe-r)#|aDfU*O(4&K3b*9qBd$43Pn3az#f -R!i8H1Si5ShwZ+zDpXvgja!6~$QpAk)3$P;cP@I0Xa$X^eI)mSA~q@Qg?412A4!=rJ-O?UUMb_D9dWl -rr7KMBi8Z7TIFSiKLJt>UJRmJzhlEU>cVfyo@fqn1&=ouIBdZbJDGk_{k&&Q+%^2cse~hP3FSu+X)B< -Zo_P>|mJY3KjDN!{niV;Csj)~$xLGdq>9S``IaNnD>yw&MHZjuJ!<8^m)uGcT``rV`6B* -Lb{W$2Ui5Pr#wzP;DL0#;2x7Gx&;jOH(N|YqsN)0JS{Lo}clUgVH`gNxOb&lRQxb&)EEfaE|6=LVq>W -q5K0=glS*^7FRLb0(3C%>P@?jW5AxzJs4RTD_hT$zd5BRP=P41SXJ|4HsLb#x%^!$F^T#(pqQ~Gk)&V -3(V15a27&ZkFsk|LCssb@N~1C(zZJRb#qmf+cfS`7_FNh?F-qd_A{+BmblH-B+<*S-cDTuIeB|8Cx+w -LMZbfa#&67j8sku%p@mb*5VYnKuuiWnP;90GC)9G0HaQ{~XwrEYHWw66*G^4+@zt){<0-O`4zMzvfkG -&2GY-wW!|J&lpp$QOi^pV%Pi6;EUb}yeFqv;=V!88&YFwa{{v7<0|XQR000O8% -K%whMrTO|ofiNAqD}w+9RL6TaA|NaUukZ1WpZv|Y%g+UaW8UZabIR>Y-KKRdA&UQd)vs3zw57Toz(Dao~icBq)rxDA_ -Nz@ttpZ%^!#-7FG<#LRafsyz(m&!dJLY6zn@Lx7k`?*`18rzpGQ>mEZ@l=FOn?7vE|#_;vEN4=A+(F5 -w}dsC{A;jVV-2sB1<*s#bycfAQA@}@=pscYjyjRT+}!g7EY@Q@pc#Y6|C(UB1Lr}kQ-RQOhkew)nXls -MNt|8oKx937QuS2e1yk@SpZ;~FZm}}sl~_Sn#U`Atu(|lS^A0!_W{Uxy)CMbFeM=XE7LbE&t*xk7sr4 -zFy@5K`m|LG_C?yu7v)_Y=jVOBRK{0n1^n5e5OXo*dTL>08t>o%KV?TE?d@Oz}#0ehSS4c< -)ScT?z4@n}(ON*HQT_5U$g*LC?T{v&=xu^Nv1BW!Xq?7cXB^Je?WyFoYck=Cl@PLVXz1B1AsGVaK(T-Dn5+Z#X1GG0RmWT1vNT5JwN#y -5H!h`Nws844$ID$IK7kUbjr@_B|Coj@NslJ0$7h89zSN&Wr{2t^`>us!P$rC^6zJVfcgJ)`Q7C${zep -jnms%G&xeOk5Tq69>$_9ir=i7M26Wm=!y(?sSRm#R?R_V5h4u -;ADW(l4>2obPY<+G&^VYj^xaH9CH>w}lVtX_QmN-*z>zedfo>>z@oG9iIYZ&EET?vCsx!C;T_hCl|1w! -yXke+O(IBwIfZE19oxx@^lQ-EVE<*Jb;@OZHXc)YC#~_cFl9PtcsG%y0cxq21Da<^ppwM5P?A>0u{&9 -g$xa8^WoEXlv00Wi&GJ6rR^32t+%9Fln#YuVBLFoSHcEAz-@heyfGX*0@XCJ{n87?qDtv#(9CilImLEUzisCb%> -^jp!-?8WbZL3sG8mphOI?NV{O?7$v~xlEP*w6RqyBx%o&%k1HrS^kRAR+9mZys_1$RE_2`Prt)>J;=MJLF&`!vCDex_3w)(b>bRH -cb?VAhum|eb8M=$>_=f7(1r_NVP?4bnLqmid?`>_-Ewvv)Pa3(udl%IRPqPC7FO)S`gz{&N_N$B%#p1 -SCY8kMpHA`&8m*C4OTMSr5v`X4k%z-w4xVWM3P;`FzsaHPzRo~1e6(UYG8=sCfOMva(fygQcpujGf?C -YBMD2-iE4pkpobWXF^Yj+u#tn&&$Mn0j)i8idO?oF*A63e_yCk%-oUQc3I|{zj>R=9k-*7PkHHWPL9i?* -$HNvJ=!qD_vYjmtv`epZ@ReiOd8LlKvoZfv0yW_6Ce{vnw}+yYEmrHIxW}P|!%6sS?CiZ?rxs~#_nKM -25Ww#@(!j>$fQfdIBSRv3B;Jy0Bkm2pkiNzIb7%yu2eR0$VB4eqmQV+DCa9|~x25BG~YBgS6^#_ -J7}jP>*Hyau#jUW&z7P&I(T2OKMwLmY%d&+-;*-chX{I+eWK(hg3crD{4%f--e}K1t2>+vhUiu -+HUIDSdO{;e`s3;EV-brv6~TaAn}L{J6W!ho@2C4svHakf4l5|FXTo7f1w{cJjC8?-qax6?&8bMXPag -j1_`99UPg{qaf#)Zx;jNIf}JkW!&HHWYq$y$MOIOYaOFziw%TPEF2XqVc>;P6#Hrz0feV1 -X-?N%7>7f(D*;SXxg4UciP`;#4eH?>O_;_y2wyg1zRrkJt9{g11sK9BgKwTsIW#W -R1D_0ezzo?dM?&x7PN%wm`;WRz+TuJ272-OBp3$r0!!gwEyFa*vV06yFec%u -vnV5m6PTgfG=Kti28l?e=BtsO!-#4gK@e~HCwyEIC5fQYgYM56*avD~;kr;IVH=wmF4na@E)mj -&4v-+-@le+lP1@P+x`Idh~3&YqwJFfg-6?h#x~9nsZOHnDKmDjg(Qy)K&TH3RJfNG|AL6EPulB%>JTd -MH-}^fonDf&@i!F(rc=?tU5_acp)pl$kIL#? -zpUdN<0IT-0em_rsiOyZ2o;X=N)&PXh+!{J>o~TE;rYg>*BY*SBB8NSLExa<7^Qk(qUM9WJsArRAMDhfiUP<@kQ -fJU#@>QdOl*a+r1X`n5O#q%HadMsCNwJFK-||f=`=vqqfLlKM{)>ECiKSV9!tOLEhEX=$oj%QYuy2$( -DLJ>X*(L@Y$n4$S^^@)OGW1h|aS))I4O3li1`t^j!n+K+*}gMn#Cjar@AmJCo=_#|OP5WC$<3=yq#6Z -TgWDik%pg`T!FRL!>haipuqHBrl8OneLY;*j8`B_V{?AfxWHVmO3V3IAL^WH$AESt-8w?E_GETpxD)zA@N98mMm -%RbqcR1Zrf>+>ClPWVWuI4r4Wp~%uDZSKdZj9-mcLag#T~du>x+ZgM&9K_lRj;@~pz>F1^EEdhC!9za5WMrJ~99ckK_zNH7%*}BKjhkgOm~_%)wNY^uX2-^=G?uyGf#%M+(Zt_n0 -%vGmN11*>vRYBzn?6+Y$}JChqJy~M3*uAp=szQm7>DiG)o1H*6zhoNDzyWLVRI*1)wNK`xFm%Op -uWnv?Z2WoA`YX0fQ`yx#}i -j{2V*7rgQ?8IUaGz;mLl{l$KWtB8I0j6s_7Opt7N{yn2TK2gbs(r!lx!;yVAK{*|*3h?n%UjQ5if>v^K2 -YqiDCJ=Fw+YS7ynawxRw@CQkTb5NN<|)5$pS2r -lx)CmsQSln5ZJJX2V`gS6fD?r&tF72SErMse3eSTS<5zR6-`>h^PScW_I|0@;XyjxvC9qBIIb^9(_gu -f$ZeIB3JkfJIM)U+*yhY!1daTCu8I*bal)8xqqf;WE}WCJzE5LXv*CcfJ-wKY*=AP)GO0-9y+7d4^g5 ->_By^=Wyj~Ye)j>XV4|05TbPbZ$%<&xBb(X)^n1qB~*Yp}ZVtITB;PN5k938a$jUtYq#2w?0T)@XxwROR;G -i8kJ+nP%%0uel0L4>Ba9%_7OR$qr9q@G30O4eYBbXpCY)Dg5_i*S1DnG+}~{Az`Uhc-Pj)N2QB~$bXk -RuQI5+y+GS5Wv|B%yQ~VanePpL$@lYC_>QehE%JWigxSq`^Lm6x8KId`O3pEA1U5CvTG~($?6S9arM{ -B(Uqy2@A$d_uo@;?wh_}U(6)+mge56#${zUHk>J|EXj`VPOJYTBJc*Zwny>kz@b)T=2T*Weu?1^SR== -Wu>VY{8PNL~(Cmm&7?mn6>@QrGWGqj&1E9-fhguD$DXb1*XJbifW9^$qBSn>1k`3V|LHbo-zVp)d5oG -Hq2DMkY^2x^O}0&W-iJDY=Qde`sH}|A3pHRA8*$Vkj7I@{Z*+jw4!jDDY4HVanfqKrM%NuE2k!m@!Os -#|E&QVZ<=o4Hac|a>b*FP317^5tWrir27=G~h-;RI7{t(>Z%jBp`%OmV_gw1~X5UUydLh+5oI;p^r0tW%f?R}v@e^o#Z?tf -K)Akc^~VA;$?RDim3g%kXW5j?pHqRsJdEqWPabFke1U?h -pGAfABwi@a@v6I1gn6Q}`;nV+@>}-|5l*%xlrU@{(>y4 -^INI>Q*_-s9;m0tNZPbnj5Gget(E>eV(q}2H8wq%HgaiXRy~Z5cSLe@`3ApT);(bgEltTjR&iY7Tbe~ -BL;12}R_rho|c?L@*Z~db|z2?cyj;T+)uuZXSQ2jQ*m>C|a6ll88dSFXZ$bV24gZ2^J!WNAg$AKh@fvcm^KLM%_54OtHB>x|C>m;V<$jP=yT$S -ihQskqt|;YKOG8u??)RKtz+J+Ol;t`rL-YJ}ug=UFQQqUMvj9?;gjG*)2f&tS?#H;bY)@5DrWbk_QXU ->_&`;>H21bn5tGPq5}ahjgOi70E5;S`U>>eXxe$P8G#0s5{TSK-*9j(m-Z(iphruKs##&f8E6e^!Nym -SoYFu;cW!uUl`dhEOwIe3d4D(?Kf3Zp$Ck5)<705VCT!k6=(mTaZFa)CUsTX1)?IIaPC;c}=f1m|#0H -uw6V*JTO;HSnvxf$`ej*lW+V8p>4pH_H!sFUN^Y(`xJm!2a)#*!WO8n>E<-xCo#m@X?2}-8SO|JM-*A -8SP2z4gLPoRm4tC$9kjjO&Y5xynV&@~ja03-s0j8re*t1m_9kd2wz_rRR~*-L)~QiWA>Jk%sBP_-pU5 -|c4>jM^RX*0a(ODTxM1CaB_F(J-+l$t3N(+R?sZz|zE$HZb6)LtrQ-KMSSY3y_rPI${s}T?>$d*9Wa@ -PUdSy`e`p`D`n4LvwK!{l&H%;tzN_4zW`860|XQR000O8%K%whCl=OM&maH*DS7|^AOHXWaA|NaUukZ -1WpZv|Y%g+UaW8UZabIa}b97;BY%Xwl%{^>cg??$E0lAZY;;TYln*q`D -{D2)!22-pdK{cZdbQE^|jM4hW#|wjsAAhFQ%#9sLwyw$8Ft+i!IKe>Y7GWG+tJ&D>aR$`td4xIf}kqs -}9|9JeK_+RLj4V%_IhXcPz(!Q?J#l*YK5JjHe^aS3Uo}JzdBS^)V~1MZ4+x>{wPGOI%iVDEp71A5`-{ -(${kxdhz$E?!`g0qwe*_JUiEQIEt#?*4A)8swY2+vpo^hx>HrQU0YXWQ~xZAwmgV)J+EOjz^uj~fBgk -Lm;mL#AByW)0H3|7i@A_*97_F`9z)vK>Yknz;RW-*^2gdEJZax|;?9_^j^faSU41Riy)LC&pr^_QFBEZhjDBxh -cm|wO(8#@b^PGh-dXM_Vs!qUZim+^n*7X*Sdg*qg9(M -{)Qo!io0lHli6ao<}kDjq3Z1p1keP^=j(&qiX*OKF6ANN7(W#6o>-?{Ov@GEkvPS{30lY*wiD6lmMK1 -@ld!mvrtATSJ31`rb7&X-WQ1$sO6?GyJhSS%_=#RABFYssHK?;41^E6OFBiwm?fM4#l!;K%@#ujtvLAT3uE%ZrPPswn|wQkq+O_lgLZrYZgV203~l6+LvY -z4y4babJ$G(N4))sBJ4Y38(}Z7HtoFwJ)#n<+ke&v_+UJ*WENGu)}+ij9s!7W3>lp0{Ep7*JTR}HxmJ -?YLk@B(1E$rjwK9)E2^smU^091&^#lN3}wBs`)7I~%5jul&v -rgD!{=&9DS95qJZ}y0)tLUJYFShum^!KFJM*Hdz9^0B-LeCXyQVUw$+uis^*&2Lx)7z%nkH!<|9g>qG -Qc|1%jX&-@$rTFqeToy?;M9NJa9e7{oXN=AF)rkq3c*_dwBcZl>ZBya(9Eq9usXaIrM4rU5p`FddIgJ --=vBI|rb-)={KihFI%ma`_$2^h$SzEh2j2Ny-3r4Z -YUw7_~!1J?^s>q&j|Bs0ZbGpzj0_p^;6oT-uz}^);xYvMJb>9V{FtZ{NIPWZj^D) -OXilhinubahs#i$1fH<+vrFHD0^td;DCP68u46OduB**V%deae!a0#-QZT0|7KvaxHx%~yng;V`TEP> -(sXU(#BN0}a}wtZ31;aMRkondE>d!)lC?P^VLJ;d430+8&AzVo7F;AA5Nm^tH{DH(20EdW90}8Z!()5A&ab2&CfROIHQF -e8X^aeMU??d?j{pPsz3j(&3pS}?Gc`k)PL}-SN(5j*5RD^l>|{5UJ(wnf*99gI?0T?0fB`_=f)|e;!0 -Q~mH_#nhozQ -J7xjEMPSbNxl(NnBXS3y8zhU6qnNJ`>tYAcMBC9}sQ$N=9S7x_~^0rj`&x(r -i=SlCXDgB^D?6{euZXro9=a&{PLo+i@RcA?p3IP`VC0&h6qk;jwUyTCrQt#LLI?I#ai@sTVfOJusl^X -y4Br#2Cqy&VXyT(slP0egr~@Vu%~L*pwkpq+%OJ&N#0js<}F{OPNg0B1LG(75aE0a(9-LW3RO9giI=q -Ll?rWV;MwGy=TpwyLqMln?;PG88;cb=(YCvh)4IV-M+qK{NHyZJ=YbS3y*km?sR&E*vZB1x%I9kSD)c -MXl5up-@Y&Rm@C><_xHvK(8LUq80vEK{huN9=PO)(E-$VG+$E6;NY`(GC+i0ws^kh6yO -<3gcb)M8_i+gJ6#VSEAf`DO8|M-3&{Q2093-mmAY=j_}pSh91@tSaP18@a=WHBXf3+kTGPy>KCX@=%Y -7BCwP3XSoY>R;Q3J{1)amy3y%3X&WC4Fns}$|3I45i!{^8SAsmdYBOqQ45*e^peWZO`UIS4=ztZknj9 -oz;?4k4I-Fttfl6w^w;b>Ctz{b)nXEU$w%pRe8A-iAA9E_a9wnua1`yn}?`m&{<(r9!}gmxd9&F);NY -k^%0yQ>n5_m9dJrE&+uG3P9RqU@`EeJ!M+e8t9NWX&A*0%8=!0((-B1HKwdN11}5)kTKSe5iM!x-Fe< -wrLr0DwEIiWSg>|avO^iFN{%%zSCXm8izBmqmvDxowFDLLi%m!W__0isIu)Z6+LhRM{}+95!ie;{L`& -+CrH1Thv(8DT%89s{q`V$ZBN$XZ0dk}bE{h4h3mJ%p#F-j*qGOp;)Ba@R$)w4GO^eH0Sk0OYF_t@ID) -Vk@ax=qY?QDVzBNtnmIg7s7^l8X27Aw;Dw*AMvctj|Zkv94%e%Xv=zLLRDq?>pJ%+0!^}5^wTpr|3NB -{C$Cm!^x{?4NF(oB*a=VBw)({5&>X_|iDf`Gal*;w0)<|xB4J2o*%b%_-{B1m-Tpbil5uN+wk3fyAXW -k2v(`aLd@ytJ-|Ar_{&W%SLl&0s2>f#C;~tQWZwzK{Z*OUoyE{Nv -#<+0_gzfVg3a#9Ygn;sNm8992EeIhz;LVMiSw7;jW|E^*bnwmH -dhl+A5MMb^xd|if@sBwH4#oMxDMu`$U1)bxff;YjO&`bs@4AzmX;z~1BbfAKmdqmgkCU~^xo73((=m- -#1!u(m_BG~HbQ?y?w0=<=1h%!>tI;1_U*5GbNSFV^zx~s5GEDN6Ag8zxgM#ZY$sKne6$c3CHP+}`e%u -yPY_Yks+PNFe{u(N351d*JhG}kPXXy9|B*Ml_W=hy%Qh>2HtwiAZ$)Ce2&_ev{}?1&zFUO@m_uVQ|V# -4?*Rb9}=C+{meddy40?*oi}V;9+9aQD*yOMxdc12k#EYMy%*;^->1)^888m6;6-A{LXp#oqPvV&+KXZ -B4SzTQ+bSJ9`YIi@=JGFCV8Gbp@3Nn-t(d!k%P?ZrOfmor#`t&bCQBN{Bvg%4|#=!jF}O3l`CH8xWJX -pVT-si5n{le#gztNoK5v8dCjK_Od4p4WE#q?XikX$;X9QhyQV6j=k~@;jUi@Q0G~^p7>;Y6c#}ZIm8L -5v1nC42XRSd=0~jWhS2l!9O@48k(&*_G&*E-L+Jx6yx53nABv@!*$(dmIXl}#o0k9pjdicIUG=6NmoA -z?ob(=&c$)G}X_PK;>Z6oj~F{`c`>uaRk<)xGpml6c#PyJ58b6|nDHlk`Zl*414O~-pHy166@a50M@e -rMTLJe=g~v7d%frfUdu;*EGD)Nl$=3^^qMx&)Jx4lmhAz&I7~FYwIaB=K=cyAfu8i79D@qH@*jV;byQ -i~EC2MryJ5aCr&rHzeY)78^Wys9Tww;lAyF5KZEDs9^+AUftzeQBY*hTQS%Lx|C6Nu+)MJ($Hed5W$iDpQt6vlAK^r -4io2I*AoX`~I>6Z#!w9{rMU}YVMb3ph_*>4i!Zpk6>kamSsyhmb0mSy7AEfgQZ9+IJok#oloQJXr#S0aIuA4OdbZe!4W&2+#?uo4 -bHUeBU0z!+JpeK0b!1>+k7GHQ&*(UQBm%@8=de0OG?QdCfKRN4*!s+L7XGKSvRh -9U*I7gL-=RLCjPbJ5$WZ3Y4Uk2YbcCFa9myCB|R(H@=rsA0?YJFA~-_nMcE|5(kNkx*JsCt6)^*Z=I)7TXh&j{A$Kxdvp3KGv-sid% -U91<$*$?vm_P#NdJ7)NCMOHf?r*7t2ccfKmSJ(3eA2aaEO;pEmU0}LvJ&P{3r3%C8hAp4S!*qu=?-=t -sWc(V*X5&QlUfD-1SMkOFq~h;ma+7z2Oi4;R*!0GXj==FN^*HQ^p!NdGJ7$UtfTgcqBb3Swm85Bw$%n -Kg5>Snm(Nr}4Tl-@BrRy9Hij-afIg_vtdIbu@j{AoOUOY>N|eA;9U1|KBaoa3g=JzM*1Kp)OFf2p$ixo>5{_od|WwsZe?o#LM^5uX1@j1A4gj2K!!7d%~9GXcD7@>4OVGx5n^ZR76-* -Fnz&-O{7KCX1;O_r(5O5@!3g@)t_cuthgIQp4BRhfq7K=rdBwIrh}rC|&#HGt>#07e2glQE -{o&%*;$l@4eC|xjgPn^clj`n*~#}iSbc0)H8)(vklGd!g%hY77_C(&KzBWWK%)%I+{!~xfka*Ej&&UD -|UvPdp{zg($;Id2bkM?8$mbNi6bx1>~?+!7{MVzpN(;<8vOV9osh1luxwzlt{Pj`8PnC< -t4>p9#HKMmzYCZPB&Gp&nGksc^&m(R}r@(LbF^Atm^XR`9XlNZI=|3C0MWaeR~NIdII0t4V4MDzw5iW -LS7OX`X-$@ffrQs*V+FkjH3&Dcoeqt8Rx0v4As*%uz0AZN*8Yz?Tg?M){9$!;*O*1FxVAyrZ}D*ms+$ -vJ*_Gy-c_v?0yEW|BL!la)OQNcYh>yyZkeBp2{7?Rk4qor=v3$suaX?z^NBJB@y_Hp -DGN1(pZo#g0t=xmSWPDL$1y5_$2!$Cq>M&d7U~QYa8cmFXY#TrRzmhJ$l6v-OeCK6o25E-iMfcsYuR* -vqr+5tgfzGQHXyV*R&JW7xYEi)?Qb|M#2WJd0j%8bbqf^k&qy#P~SUA>P4HCxoGgV9Fa=$Z>DqMDX>( -~pd(kj8b@pY#^?QkxU++2rhdDLHb87+{XlwWMmyKbp@ytNrPOgs%d8o~*sBF_}!EtbTC!T30KAw|Xzp -G05QU@HY>)S-$}f7|r0n2p6oKTlaY50n>)LxLe)m6$$-0a0w6o`D9&_D6JQ$>c%0f(#Qa8{oI@iu&?~ -P5ZL|@tnu)7s(y5{~nU8@*rA;UC-CTJQ%S0d&E|;mDa4?MT&xx1_KnD)O;Bb7T4B6nt>)m+B1{7XG@$ -zl-t#zhL>t!G7Mz37ru=ie$xB7GaR|sXVS5x@gMaFoKgcW_r7NioXtf!26NMq-Zx1m+yw_wwu6=AIaf -Zz)oh%rif0@0(8Gf37t%Q8Gg)+TbEPbyRUVz;_tt{{t -Ch*gM6N`*XK|JZ7`bHOj4%8OE50Z!=hjMc<(+Q%J-Nu0Qz!z)I(PvqIuuhiTPAM|Ob_x#K}E^@VGT^? -(jX*`{&~qF6tG%Q-md~S05BWbrO1Tjn#$D-?4cpQ4s2rJxMNH1@Ua6d<6`SY!yQ;}L2V_2 -L8y!LX_hM#oYIrV&qK{rvClv{r-FHz%h|7pGGr#|wwhAS&hlH2h{t~3&(J_gtfgZ}X9#4#wGptUzVlr ->9)xF(HuB|23(K;hS3Qz%F{h)4kcA7|Em -vN~CURl?Kd?qavQ#In}=Q_y~?%97qN*kJ;6ynG)RdtNK@4jAizJUtx+9?UPpAfY~ -30#8LHPij8W9%!+KFlRXtQ+gA#sp?py4A2{5L+1c3e(hHOG{!kiJ7Ub!W{jA*`3a#(kQ{exBVnQ=bh6 -Ynj(KN>5`mNgEDjv32YDS6J(soZHTvZjG9Qu6mT}0?ITqh!#J5bC}J9{Vs<--qi>|MzB<#&c%Eb{AQ( -Xa?I25bJdyTuHcWn?hF*k8*x<&@7SXUv^EcykPI+zZ8kcdyO+xlHoRXW5VKa|ia_@U|15h~~ -)pqLuge#M-J5XsQQ1%JsOH5|Nlj4ABK$xRlyY!5}befpg-k%7)qwZULK42 -TC#F`xIoZ8b*j?(Vz4+9dD~Wb7=>=aHcAV5+z&RsbfS?I~0$|E}xZ8)iD0*M!0gWlk$^G%R -U#z!&(Dqsf1H`Xko&Na2JbN=q6^^jhmf8~<&qD}z;cE0~5iV_4??zTRv^OE;9~`bOxvdo-+pGQNb+m$ -a({I^+iE3T5RQdGVBgY3ks+<7FeRMT3DCck`%g$o!fGRQd(wzV-*t2$Vy3CZsV9{ae0cWiVDyRHt3+T -}wErrXKB+M%SvvFFw?oZcP!Ls51CdByRgT34jA8_yDVYdbRjo{h? -ynT^HmxC2rD -ow(q_+tCcs&!b&)Amy%pm4fHTmo#}-j^z(mqR#upd<@u-Sk00wJ{dNX;dsv=0Ds#tN=aa5XI4~wmT`N -isfHC()`l6ex)FVgbTyf4i&d`biYHA8T|H90>qJW2vG95&KelR(bbVCd -C51Vi$;Fc-+_)2p9HpV2FTmBdCFmZs$BwhUH9Twtqvn34sB)NJ&*A(bOH5ChMZmT5yB8{?|xzWQ~ -|Jg{MJZg6GB{Y?ocJB2q(AcL5pNdOb20Y_zL0B!!9sX}-N`WxPC38#Iq(q-d*)OljG^g%-DyFU+)|E= -uG20+3g@%%ZwN*2Vy3|$vrmp-0nyjkrhhK73W(yprE^amxRX^i!h1ySkR-3)eFsJ_Z?(+671B&SDZnR -8V&Bpqc0k{}9SOqTjsFqZM#d(D~HoT?3PBQPwA$EV-BIfsC426=pZk1OHC?cv-TjmpibpV9AS{Wq -pKyBfi06y`*2yfJWtVjX$;!F_{->jh9{A;Q!Y6zcQT-HA@tjg#kTk&%%vT)Xk~GKY3}c1?VOM$X8O&W -RZ_n88e7l29SVkO8(mo@0rs&GN+9gMsoFGUFO$G&VY?9BzTX1=@0TQ`RRyC7`Avcl1g)!UHJpr>jgv_ -hcoA2`pN>-tqdaW!(ti1B3g35%;VzniMzj0ZoB+*UV5Nx1{?7>p?x5KaV-1q`iQ46E}os)J8zKtod+7 -u%h$!T=PzD;|I>4?HjD07fMLG*^=xK2sY)}wpd8V=oF0ErAh?|Vo}4!q^K_e8-An4!JzXNUBB{dHmDQ -9Pc~dqY7hyQbg22`Pt%ea#3dj02z^N3>9Pqg);E7H4?){mn0m9_76%eO7Q+Jxa-I -y{gUK7mJPT@BB`)vr`Us*Eh~^Wou#!}oqa!-D?Q`e7TH?@F2O+(o8G;zbVJ#g31tA1PdTv@ZwXwb;>u -H55*tZqQKk$5J%RT)|)-(u1y4(^az(PfDlBmHN=klqV*Crl>a&Ho3t|BGhe7@iQb&%pBz{kZn>{75MS -(sE*_DW-vlL?-@FLBd`%zQ@iOZ`e^~Y5dSAm{PQ4~V}AMLFKsLdIc>UjCw~WlGKELFrH)qRA0zo2km5 -EtUY_z7l?8466c|~bCk&47^{-dsa-m&Xj%Cl$b}^hZ3r4z2ewUaNV*FhHUw#Blu^0{ -ysQIa0({B;XZME>jAcT!&Oe&lJ9yr34JAlS>`;Ox^A>Kh0{g3e7STEZrEw4hnPA2c&+ -|9fG$458P!NilxTQ2QAc`5u){s3>v#DHG$mz=(K$w&)l@gfgUT+{KpRUM_lF$TkR~UxhG%yQ<4H6}I6 -m)hjhhx_oVBj1zTo|io6BHjXqbGF~_XAC1MW5+|0gR1w66LfHM{r{o(?}K%fu|ZmK8jZBfww0){E+Bq -Y7|j-?5+rHNTIZ|C0ve`Nw!REz(g_O1(C`+!z5EB80xwvp7p{#!EP>Cxy+ffOUJ$D@+IyE!JJ{}xyK% -CyU~(_BupL^ml1M`j_nxxSbdnWuZf20zeG7N!aK6AdA93hElYOQ((jI -*Hk%?2X#|P$c6%N(wMNOK!KFco2yyi=3{eaS>9ojzGI{#dNjC#8FjU1SHID1p&MxwYPvOK2~lX2V*$E -8GT)yGpZu^EsV;QDe6>^ax9xjc>ut14Wnk>@#p`tCqSAUlIfE7uWqO1EQ1JvVfX&#%3OAI$9*&-X-d& -CCQf=aRb7s!6!LoW34ogKOBc{r>C)nOaRfYbNu$x)(U&>OzW|;d^eX?eNWYk?OdN|4tlt>P4FAV8Uvt -F3|L(W!wPy3A^uFD;*W#jcQYp7jDHpY>IcRHhWi5`CsvUqw4(xfauWOg_WBmRO0@;F=Z3Uc(Fa>=jbB -Nf<^onM+o82lVhi})VYtkoDF?jtYTg+UCB^P2rFk_xkL0i%i(Vdo`u*o_b9-l9mk#@(AFLQ{+tciI1QY-O00;of09jkubWlL>8UO$zUH||h0001RX>c!JX>N37a&BR4FLGsZF -LGsZUv+M2ZgX^DY-}!Yd96HaliN0u-{)6w>2jHJBu#`ZZx>=nq8D->wUlPM3zP6u5G$L%Imu6^S*3qd3L6rt@rh+Z<iDvhJuW9-*Rlry^XTWg{}EbBtlJ%HFu&)${Nh(IoH=26{rySys@EeMHu -)QLLZ2?<|zeJ8{;q+aaWtSCFf{NmSq1?&F=&tH}uh+NZwI5>Y!L%S`zexFyOzSZ!$eGMY8$@Z|TE*pM -7JF95F(Tid4NtPkjSvEU6J1fLG%H*mm+n%1v@N&ujH5dmZ8k@dTJQp#+zFE64cDA?I7TKw0bB9} -Da`=%EB@6`oCjN|z0PUO9a`mF%vtg2)58z5P(OHo7!I%?MZTUiMit!Qu&Esr$)9<)?8m6*e_4+>!!t( -%UfY|2|vN7vT{2;%j1N;BRFun{2V_uv2hed5pZG1^<#o9hvXAu|it+~hc;14sP21x^ZLgBzaT8)K5&OFQZ7*0B -*7b$4QWQ8Ye%Q4X0=6M -Vk7KTY(eO2ynV(URN>EZT3$R~aiCA%K62}^gdP(>HmKb5&)MJ#9bw#Rfs6KQnZAHiU^*+dG+-VZ(c4AWxs9q{Q@NBa@Aa&Sp -a!v-x$cO&8423M&AHb0S+>P*{qP}$9V)qBVafaeqVuSFYrr)gYcn5H7&G>OnGVJ{(gkj--@1SKa2YO1 -`SEx8Pk$w0F#_^7AfQvh}Ua{11K0Vcin(31;d_~5`g4&B)Sg#KMltHM{NVHc1h^3(Jc_+*w4dCufKo$ -(^ubq`|1sCQ7y9V`&M3@pSStyCP%*nH~=NZhRQvkcjUrh%tGx=6bqgwsJf6R -DolsJTVM0P9C14xuD-$d(;&3WF;Gjll5JOXrSqNodrEKFiFd7gq70BT)Z%gDAW`vkwbnv6;T!`AiX#p -3T*&QBpk5iU*nB*JA^u?sDL?js>J(~#JhQ{1)UEr0>l0W8{wD>4~q74eV3QS*!_=k?)t)M)U)eXi7SS -Jov&LQ%cnEg`ySn6m6uYq1hi=DwO;PL4^K*QDj+#!>^@F~FZ=^i^9iReAyleoOVfs9( -M1BjpT|+1V&uF&XLO28%5xG!9_^D`Zj*4~|n<=trA~>!A<&j8$rXB$oOvQP0BaSa1z_Nq~pzW(7lQ$) -WF*mz`z}tskDt@(`MkzRuQxb)?Q -rgux68y9n?ZOF&q3KejNxz2B>5lF#iEj)KOecQmCW}BYG491|%^(PBRgLaUruRe*0SLb7jSFT^_| -tMfdk(4V=~cYb^ujdtqTtmpa6GVxu*(UaNKg)M-nIwNQ}WoGNoVMd@12*+&2v*k~i`zBn7wmC>r1(`* -01W5tDZ$ZZ@lofj#N?BQBC}uRVsxk^>ChUYBBlk;K?9?cZsXCNUNGo{wdGj565e6$#l^Qdi}r$xD5iw^BUrf6m6yEHXYNS2Xg{ -1ap-8fPcaM~MK<*>cDVqJw+_DuTH~W+Ol -Qrd?EEdMitvRaCDgvvx<%osG!?yMWI8LdTBGEHpEq?WS-3hQEfK-sA3AsHb0480u4I|!-R%Plp?I`wcK!qaG -nZqEIb^0JdT-He$Z?z=Jd2I1EKD-j6!bkwGg{v;J?7^Db}vgg{sMg>ZBclvP&FzSZ0-M*$|P?51IBJM -P~0DRnPY?$^;KiN)t~9&+3UGKRFgO)m&i0>K|wW8aZXcQP@z5caZo*;;!#<4OKxP(b<)oEtMr%WK+Ww -9!o}-XAPuLE`%$P8JX&l|KLe8pPm4h>Re@QW{s;Cz&jfC0iKU|Kl&BnXQS*zW4<1$Y2NBda2*Y$ln6H -=;N$7|K}8q^5IkPjZG6!f!I{|XljxlwBSAXOacIfLAdS8%u&RTM*9lG#hEZvfSGA9TC19rhJ_M0OKjt -jt?hyWmCFQ|!HWtkJurzoR$ml^b-}brOio&m;7{~CH?X*>B9W^bMnCKJ%O)7%W{(lSqaRmqLDU1PweH -^MwDsbNonHg0Zm1%%9rdpH3a!UzMqVIAhHbnz`q@oZ)FAF$a>Vo=}TS6Z|5UB`9s*$t$1%d>`A!+78b%L4jPzrco{msZY18o-hRhKnh$O_1(&CkK_@FKN;q5D|@r5A+C>O*{Sr;#2Bid -FK#uK5-aT18DQ*NNf3_H^yqWU4_I*!9yvKW7ct5zivKd&(yW(~r~i9(mQYi>($M8F-0EYMt_Nb0kgQE -WZ<`#QulB||zfCic4pEte!_)D|){K%pu!{AP`J+&|Z~2eJ>GqUYuZGD-w9`r8Ww)n)TYqCAK`PZ7e^X -u=W*b>lo$qV<2`7?3R{h;+Tr2ltMTVek0gAi>TN4}Ks*EDNq=mzXcrN5}-r9uouN`Do24D*sh&plKoM -O~`-ciHLfVuxX{u+hL}jfKFe3|JyKoweQd@c)6WZu^_kVDBN5_nmtl4|F||fY7{Boq&C5+SPAI}IudO -}2aoW-RJ4fXafj#OpxGuHX`=kL0k-cN$XbuOOr{1#ecplnkg1K2qnKGK6?2DR3;t#~AmmV#(fRHvxjFZm1EW$Hb%Qq07Y_&Wv)V)azMR|lF;Wjv3bx0YHVP|dLUa6~e-LeRUaWec&~Z<6?pXsdjMOFxy?s_r7ebkR -!;jewyBdA4dofAt4O?`wD9Xq>2=B~Mrd{o&rA6CgjNAhfI#qN2$ccatQgqv|C7CI@<^l8 -zB^Ir2)kpIz`gap&7rrJ(BZmk>NcV%Q%C17;hG_YN!&(d`h7@!lM!$W<=i^=?%G0QAgHP}S2C85jVZcQC -pV&oy)wYFkNX4y%M7G}IjA(iW25P89MS`${DET`CL16L_B2*iczjk}GP -?s#u+}6KLHa!tFa;x8CkMa1kJw)TL3cMya(6Gou}1G(7YqRo$K+C+=A5pe$>#8r$Q!Tqq25W!^T=TzQ -Xv>h=PS+0f0#Mpbem#ziNyh8;?KusvVpMr#1^!H;0PA#u -+}kz?m7x*8rDtk$s?`{5mkxj$p_V}aLBn_GI;X=3KEygC@Nf^N}V}vIYBd(Cyr8W#As&iT5(j5qY#pv -QIMDRJY}Pd=m#cQHqq-Cz|T0ox2>aJ&}XXJOZRF7ICNK?FVRpiz -^fS;aIuYHu{IlIWi<^(1sj0Z$dcX<|MR(QD{>jHnp+LwHTrQIR>{_f1Ca{@spymjO{SnDgjXbjt=Evi -&$5kH6uxjeTf|eaDuNUte>3pjw7Qjs(I=VbyuAuWeLd3P>iJ$eCN_)ZI8%STPy6kJu7_o!8HL%f>u=D -xamW8*^0O>BD;46s7Teg2U7f*yRUm8g>WGLUzJD!-?E-4n&sRskn~s15h6p{!=yMVQ^hew(_%mvhY-1 -#PjiLmhm9Y44wbJv%=;Ri!Yc!Fgmb6bAYf3euE%)1<8D(wBfu`!c(sVOxN+Y%3gl5RTzqszwNoGnH0(T@KvD2acfai3qk?aCDupQLPopPHvPdnC$6rtrC8ewF -uY={P;@g6A_m5+gVQPis5Ql{#axuSZlBA35?o=roM+!=qS_I9Iq+H(I5>eR4#vA>s7IQD(0ro#qDt7# -t|@`n?}`1Muj=u;4+20O7()6nW5vJ5BwMjiQkc+sdkNcZo{8V4yVYSaY(rJ03;QXB8c5VbEpMjW`H3E -j{^lHrwSzYOLCBoKy3HyL?J=uTJW*j2{Ob;>HU4kI?n7f1j;XRd6UFHvQkIMzs~E~e@6f)iH|Y1zT-d -fY=(yg@W&r2$h}cRNGw*{yxxFd0Cu+}WUq*G4#d!9NHAYcJ|n>LZA9nF8X -gzp>jrjTnrlwkWD1FXzN%ukWOs5>!GD6r`!kRbD+(P$M8>V%;vm&rsaRd_uW-RO0+R>X0{7t^O(*z@0n= -UTYgF>f5qwYAi~U=KOI|oZ4o<1enUopCf3wo~9|N>b}uhCfl -l;*fxntqr+|r9fynJ;EJPo4GfDTywy!DpLYGl9942xdE$wW0X+_1uG|s>;|oh_}6zXnS;{izMj?Yc|*xszbLZr1nHLo6ZnAt2@mK8>>OY5h~-? -Xk7`6#AdysftZT=|8t{)_&fKhs~J-DgImz(FhI43&+#ff9mGqw_DaC^9oBUrb}JxsUt!{=l0~&PH&W -yfp~+sgqoFR~nd}lk-FTCl6);w^Jz6ZVJ-ZSM#n^?o6z<78xHu(A|qaGt0lN{K#o&<*k~F(G;WU%OgcS^i~t^tck9Vhu?QId3}6pnVcN -8U+~e_hK9HWQg)Hcqqghb_;=`YYP!>r_p7bC{MplsYN&*}6svsC9b0rFRbhn%T*_^3b;>fVG8L)iP8O -f0U#vd+Y|$P+PrpEb(XI4;?wsE}aKHTW3!;aMPX^-}JNN1CsIN46Ai>504=4;i*9X&u7xCrMzxk%q7^ -8%Xq9%khLEg-bB|;K3wFR#u#^MyRcXYTr3Fh*%t7%B>SY=0|@At`0t~8GKV_ZzX(jkS#;!}}+fI4()= -wm#O;`BU@J{b|^1ByZ6ETN3y#`w07F78ggge2hIrp=XyikFA@>iOfVAEPh6_*4h~CLRTmQnz+>BnUyS -$qcTqDKPYfc}X>6*>Jo-hw!Tg0|9+iM{AnwkFdr#X}}MK5e+!7#K(O`J-C6hQO>FwdVf=>?R~u^Cxwn -E!RyU@D`}i0PnsqzedC;7)U(@JCDV5K53U+>7NHKf34$JkTg)LdhZkdAxa ->Yzqglqm_^hsIT-3V0CF(;#uP$yckiA?MhT?<_YS%bv-pqc>sVGkuI(9tp9gHrM=>#;Ew;4Tj|`nY!f ->5M?jiHUmJK@1-O{5eubY-mQMUaW~$z9Cm~^S@mBBq;+LhFx8NnG&W*-XxR{nWSuI_`F -p-XWotS)~znGDu%z(O-X^;tOInC3s+;_J<(j1?9$yvbq0NqA!HT!c*;Hf{Y@{lg`-hSVSN>-9lwt|;@ -`mfAI6ME*C`|%2mPi^(SIPn$dQK+<6<%4BjQ2bGj4x6Rf;B9OJ+p#Ya~ -u%ahBubMQ>Q3U)eN1G5kPhvD5c8)ljUOz4Tc3NBxql5A_k_K^D1WbBlC&?6G4g_NyjB1NucP?=dx@_`dM~$mC*m~vCxm8b9sQzfo;!YX;!uvM#Q`VSC^GI^E9kGw#b;N;nT%Ql9;_gpPU3y -PUi?Sw_h*@jnhf;)lW1bL<894r8y&agfIoE$Y^*TK@9@Lx)})B)SsXtOTaIpF*!Hjq5#YIa -zH4-*m$GI!{}sj$$ogq&(Ia(fF+r?xGbr@AiGg)gnVrV~i_CedxaIcYHW{bGmGhm@CNJKwVD>w1kjBI -i!1#xy|T;dnm*{jX4QM3T`JQo3dG|4>T<1QY-O00;of09jiA00002000000000a0001RX>c!JX>N37a -&BR4FLGsbZ)|mRX>V>XUtei%X>?y-E^v7R08mQ<1QY-O00;of09jjk5xxlg0000%0ssIY0001RX>c!J -X>N37a&BR4FLGsbZ)|mRX>V>XVqtS-E^v8`kil-kFbsz8c?usv>Ue-kS+Ao_&_yg-sqj_;wof#oES3$u`*Hqrp@NgJWuX{<;IxOTM>A&f?E{w7v!|>Sz!%4D2 -ra`O3>E{Fp(CKUQ+N*^xnJ#$^(|iLe8NJhgESsAo}Ro5&6E~5+2q!q0?y)TPFywq?;!A{u!$mX5z)tO -n7_Lar?H5UJ%Fa%xH%-r?(597=Fv}KbkL4O9KQH000080LuVbTL1t600IC20000004o3h0B~t=FJEbH -bY*gGVQepQWpi(Ab#!TOZZB+QXJKP`FJE72ZfSI1UoLQY0{~D<0|XQR000O8%K%whjl0M*3K{?a{9OP -5D*ylhaA|NaUukZ1WpZv|Y%g+Ub8l>QbZKvHFKlIJVPknOa%FRGY<6XGE^v9RJZqENMsnZxSB$x(ELp -{cD|g9!-!D6xaR-R;A$+ -kR;cT#oDIMiKxb@l!rMP*<0(<@Q8;#8kRt;9+8`>sFAN)$@$yQb+L)#mC7o{6D5-8S+;HZb$|vF_zjw -gbEr@GCOWwapp+BZWH2QWQ87pa48^>bu`$IjmvcebHBqR7!MvQFiUVsmlTW=tbKN89{+RU$ozb4W0ku?SGmCxQs{sfww$)lZ)~#B -IjzP^u*U+iR##X0S<~I$*X_OeQ=J -vhE`jhsF&t~AS8`vB%}@csqA^*qE6upqgX{~S7!Ru9uc7Oj2~2q?>XzRP=M$`LUi?S|EgFqKh(ne?0A -0F1I~C=};@%mH9Ip4Z?A3+AnM-Xm^PPR>E>yRGCoGQ7m+^X?} -c**MRc{p^fMrwVvN^PeG5&Wd7C-Xi2*_(~M<)^RcVQTVGFK;INy-?EKW^ztf$wp;O-F`IFzK4Hb82X6 -(@R*Yt+=0rom}jS6o{C;3nE#gFBsT+cga3mEL~ftEng_U|L}FAn8L6QZ8pmZWd -BS8^xX)T@?QRH;AA -Bl}2YvPVK%16E4_R8642Z87TuV{8)ppD+qam@(ugUg{FeKWdI8SnmEerir#KO*^6Vav6tenx3oD725N -K?U=#oX&=FaSC^0M(7l>o=Ti4$n>z4jo_T7Z$(6J0;?s9YV`NeFy(?e*jXb01>v=dS9FHxzz85}~HZB -uNbW?ccH(mnxpLc;~`a`IZdKLDHRs_xXeEz$UvAeLtgumBR!6&Lu)_-wFixO3njgd~+{l?dlTa -9H$c>L7@@2cwG2>O6p%`1=S@0SyhS=0zI__)Me;s0h`G*EdJ;+o(|K!Ay0;=?+AeIMRd0l|$E~t`el7 -KW`=igs5$C6H64vEO@n|IUFH~S9K$epC|u}p{;blB9me-PJq|6gj}s+Y-*J|nz2ojZ#fN5Y+&uZbd_W -4{$YoC+pC)M71|;Kd_Z&^0a;Ajn -&V=(lAx0L}4;wMmXLLcbM@Y?1v2lCeA%rzOyT?a@Z_0)eLWJKS|8cVoEc*EkhrZ(o3PL_1|>3d|5r6L -oO=B(ZE)RyNKu!t7NBsxuVxsh~$9DQRzSAkG_y_CYfs;Xq^5OLqY?5p0 -zzAVR_R;tcF9iWFQ?zHZwzZ-3k6b6LImV5Hyypnu^FqQ<;I&v_dV5>%ZxLfz;gg}eRj{_&pKtNy;tN9 -}_Yaa<4|#4GqhzzUOxA%hh*Tmbw!=6J*!c0keNFLrPASvINtVR_k}Si5qG{BY&4u{X&Ki}>;eU -OuAc^Fl!oJk1=cdjxLnc!>8EQptgc_ -WC~a+OVNwMq)I4=y&s1GNTOo+ae!%93CoVu+5@Upui4bbM!(xsN_<{(3h}fn4oL0O^o>84O%qc>v$`{ZI%Oi5{z}l5%P$xT9d1pT-p+&YIsuj!X=brGJ0I#v28jKP?$ -ZoI2S9fFi~qsMX^Wi6r3`O@&gZ2m?9`5EOuw{>#vwE$0e9wf6eTEer!ujt=_(-cu!$bvUCf>Q#Gl0fa -nN|iEE7gAG@)sfI1e3PALk0foMk%0j7h47!ecVJ}l&dyFLM%=^btla{CETvM8OoJPUIGFO#3`0shsdmaVek^g8T>Q(X#lBTvJ_~Gn1ddtS?`OE&H<<0L>ZqmFk=IcHa&dn&?YNRsDw+Hu< -{}lXN(V?a}wwEm7}GYx`fK=#K$v2KNBtoa5jO?J2BCLw6C%BQ4bGo7(eF ->p497bYJM+VZLnT3Z-tl|)YnHwE`$D)nD{o*r&HCy{K83QXYnk>X4>^ -T9bSmDbS5lAoMrDKj3luztxG7w;=F#rb0MGG`?kAU!}YcoEgp=NW6Y%Dbud_9+GDl)b}yA}&J_MU+ki -X0RX?hbxVxJAkb(ayFQYt#9P&LBN|_(;ZK=f>S+0y!RVGN`c~>PBlBft?Ohw3LJtpuk|x -4s-Ym(VF(?Zq(+=bJ~))20t_^QH(4hwauQ3O~I(Glhl=BlSK6kWHsn;;U^q+rG5Q7GxrGL{4 -QU&#+kg#kWy=86U-N#AgS)FCsQg`^C~kig-lg?)1x=V%hEIQ0~6!zpNI4@Hq-ir@?JEQO!e(9%61^#l -`_`fCI;|2E-(=9(B`TKpWT@G!l_*@bvIWyYL6B0&Gy;fblOlX@L{~YXM;#Ca{vwKf|0PiJbkxSk4~x_z~%a^ -$tf#|Z8|o=#h0_T>p~0z;l=?iLo}S;q?Pglet%jizoD3sVoU}qEKp|0O{7|1ZZ-)eUAQGw!3{;YNdPh -SjZ^GOOr6$b=_$-XZm7Ukav(|up#pc_LX4_cMl~)R{MQuMa&W52>_F&t*eR+*H*mmN15KH@mpM0@ufy_{E8K%#GDlb4)`kecTQ2w{_8^5z) -x}*ICC=@_|S1v#Ow~|-)J0CC;!*t9h(J@^sbFW0V#CaZfYnTSDjNNfDO4=YQ*Fkb)4 -%GB{=2D;7q(e2Tf8l}5#ubp6v$}13l%BMazHH0pB1As7Pr~tKEIKRIS;IW&pc5dqL^81mqBKig#o;R! -L8k-AXVBL_Cw2=U*y0t{2;CNoxw^@>lw?f(7bI3D;$ma(7!NZ@XDbdEJH@1a?t~~oqtK7o50uJxXEV) -F!h?>ytgowH!l6PW=iq%(L}okA>Wo$*H^pMkk -jP!NH;@he*f?KaPA%!{?(kFtv8~g$HWEfSdfCVajOU2Koj -zl7j!r2JEw$G%p`N=lLY(f7P}=O#T~Vvz1dNNX)9zd-wK;CmyQ#jO_5jK!hcXi{TN~tYw;}g+n$VcrS -^;p`^_6KLD#yO((m0b~YDDh||P;?Q@PnhiP-V(c`fzjo9{(?=1 -(owJCk&@2Y5-aurdFd73wJRZ)Jo)pRmHdTt>cvW%NMj4|Ybr+?~u%Dy*BBP6A^*-msY8hLWRcL$y2(G -L?V|{~q*wW;OGISc4^a%dy$%0F?;AApWE<6Aj9d|$51tw~SjL*Vk$P=m7R;o4Rqyvp7dUdpKruIw9!Q4_=*a6{{;2tGgA5UF+DP1KX4Y2a>fWpzHd-ZyaP0sZIndbrKp&~sNXMz(Y+ -JQJBn$I{!cG{-s|GeDW2e~$c*rI<+o*-YiEfp_~C0O0BzJq!0C4od}F~`#JbPMrb1=6%gVU)U_l|toB -=iNgvclCh!XcT`R3w7Z5nT{lcc+FLi1SBy2e2Gw7LORNPL`!}P9MBJe@fF9}Q5a(({c83eO*cLcG -VNVrdnZm{tZNz7vcVGU|re?jfA;?Z55j88Z)g9k_nJe<(FStzsd|O4WFC;QnC;bkoKatyg*Q&AC5hO@ -p!IFygQ?OoS`m{T&)mo3DV2)lTSh1gv8$+7uJQu_ruxV5&u*p6aPtEpi)Zd;?r`Xmyo>J7MQC3a)o;I(YJK(*$ez@9Mc<$ -!#xm+?~9ozv6Fg+}%jmHY(K-Q(>i4)#X`(07kwkM#_{6Yp3nI{ud`6R*Y|vy>xPsID90EAi -zTSpv2L+?W3S_&PEveRjTGA;p5oZShU!vxXx2xmW+UxMM3P9%%j=&0^q(|Ew< -ln1bYomzrH+0+o$Vv)uT2q!;rV$+;8M4&RxS@?pI!<7DPcIqHcXOGSh-+bdDTZ58c3N$yp>7Xdx!eI~ -Y@nn7*9R8;Yhd^ByU55tA4i&C(E`G)~2v5Waaedv(0Z7|ObsbD3$A5QyeP6dllTQZ?E(pA23{o1M*%^ -sqG`LkESKWxYl9n>FP>6sFo)@3rxmm_2tGUJy!|-F%s7#1I2srqHBg-)pos27o#}fe4FNB1S?ESCle8hk>eauu<52Wc -(7;q#_-a#=O^D6AL)i$qJ01*k#(H=`Kk5iQUT4h0akbuz|`E5wMzGlhtgH*f}H$6rYp~?(s~tIT4D-+*(D?krIAM -nLH_}2O;F>!!okNsWd}{g(*bOq*I}8i!J+trT$CIkK?;vecniM7ya4KeGS_Z0yD4C* -*M#@j>n&5!TDbN2} -mj|k!2oYfP}6mpZabG@_WW_P>8aT1-jJnM6Gv%KZ-Nm^UmDO>MoE;oay$7Ze-#+GPKX1-+k{)I(@vS? -*_=id<9~v=pnMxQN}G~C1IueS}d>s`Cp>E^I7%9-=9@Ke7=zDGYJaL&dD?|U5)!Hae5X_HG$RKCAV@78C;M5wRfDlJXWkHNDFC|VG{JHy_DQ?XcL4pca(kEYRzQy0d^9f;@Q7kcgxHYmuGNkJhiIZOQiw=7 -Iy=x1G*1&Ol?v)SEr2L7C4^N*uITsV=t|WMlRNqFnp3*Cra+7ST!jHhI3TvCnl6}5$BTte& -=pCoR>uqmbBD=%+3wB??InH=p=i -jB6wlUyHH8qN~KSAB=H8+V91P;F{6(|Ic@0%n#@m+;dJ}fX2^^XdqV!Yc{7Hv+ -Xp#Du9uDf*%LQh?YGrzzi#9k-ZaXcW{z&7|ho)C)JHhggDP6>YEAfOn^kw=8NFkt4&`G)}Dow>T`0rk -{j%jMRGwexR#@17pidNPthG6)(s7hNkPqNs=@!IRe}-z{dOA8yJfn9&`0{LXu~ -6iaO*5JU`moXK-92ayrcVwDWa2K!wYbq20d9g5RCEp)@$;nw%pjjX8EHWwKiw6l;we|rrn>v`m;41pGgS%_IzSRaZ4}h%Sd>gGB-k$^MK_lUOj$-ai_)!liCQqi}X% -@g3BY?3x)lqQvp$-PRjhC3ixGc_TbCFj{^OV9%uRrU&WhI+^c845KVc|+H|&{i+6x| -pA(|`E4e>e^i9oSK~WxRNo`4%+L}iO46oyT@PZyqAISh|naMR!tyX@zfwi1!Z0Hjd&peyPU!uAt>?bu -md=ddIo#Q3IUHZ~WVwSrM-b{)MMqDL5*>7{r1(@bI?k*9T8VbRLRlL3lMg2mq^DUoT6Gbbs_mS+3&cs -z1b52VpFIH)bLaSf>A5cpJ1QY-O00;of09jiA00002000000000l0001RX>c!JX>N37a&BR4FLGsbZ) -|mRX>V>Xa%FRGY<6XAX<{#5UukY>bYEXCaCrj&P)h>@6aWAK2ms3fSz8}v(j6oP006fW001ih003}la -4%nJZggdGZeeUMa%FRGY;|;LZ*DJgWpi(Ac4cg7VlQH0b7d}YdDU2LZ`(Ey{_bBvsK6ito_>IV3fl1l1!{qMUYDT=m}ELj~43li8O@AbKdcT^g!Hmoe0-u4>Gl8H^Hv}L?)mE~5 -b)=VZ1$y7>8sN71yuc|@tqXo^cBzU2kTV6?S446S6E_n&N?L>P$Y<}v=y-L*uTi -!z9aN$bprg%;VdVnsQ<}XMGAf-Zz2Q=Ph5C$d;3>5`bcJCJihuaWG)B>4+8B4SQ8h?YNYUO5;-~n1#% -RTqaNCPJE}^v#DQmc;KU^@DOBo^ckLRSUVe^Qx%dJ>*vmUYj)jdrn?|54a1RhS6y6HQ1!ffA3SPXD4v -M7-)XR!*4L(j|QWHPCtVGW^Y%NC)_JFPw=73O=8&KP{LnwvSpaKA!pz%F(rmQ(igFU(jypD_Fp4MUa< -T5gGcG$0pS7QDD+FH*vnPGd6JEGs@M(dGt|hC27V)9ED9ZnAH|!1u+sME(0R!;IcWR^~p|lS|#hbmEr -SJLEVjAk;pnNWl3n^Jvc(U_L-kORn@O2*e1)k@ns?4;!#+Rl5bk7B-#T=9M%Ko_+60QsY*h29i%_!9R -VAese_Rai>J7603Ymj%SR%%^#rNqanF0u^6Mf1=(;%0qcyljgJhE2&8d6k(X17fcv-t%X6!HW5Wpmd- -8#!*=aaxAyvw(EU%Q3Wtkh0jfJ45;B3Z}T@y`SfMYlW!niKbR**Xzv -xjVqT1g&f%6{D7=_U!NqeM&!YqXms`^kqnu0*BKPJ%{w8pNNf!T~KttkO)qAzI!|Wzce;Mc-mO$`_;X -@+rn7WbqLmGD4Wr;4m{NRhL)WP^)U_fpZqy;RAGe!_I;F2*pnk0ps-7ZrMf{?1|MHYWDm2DNe{CCc|) -g*tIas4;3g<5JqUwO6{tTbN1MYypQZVRjz0-KKRIR|)VDZ3I?#3-Srq1ppMkNr>`u6`A -X%1OVj5Y{Vx*NZ)i926M_+aUal!pWn+>^!0XZ9K*0TGxsMZK9q8n}a -nglHe^J=vyieia-9oKe7S`wV21m=eBURF5;vP5b4!n@pxH0e=|3v|8Po!)D$be>)GR~~hJa+>v{}6j$S}J{EW_*gvkv;pTw#X{u -j2Sp`Pj?D5uHITEdB4~)qN$MWb{QHjs1n@Tlis@E+M?)l{IvLm&Udd_@>||%)PMdGE7W-Ev6Km|KUGJ -i{$DR;#nRgAev60o@FwzM+6Svu#SnxtB$F^0Z>Z=1QY-O00;of09jkf?KIa@6953_PXGWh0001RX>c! -JX>N37a&BR4FLGsbZ)|mRX>V>Xa%FRGY<6XAX<{#9VQyq;WMOn=b1ras?L2F5+enh%^(%U43@jZq+?i -P%aPXZLo1J)ejGIe*8Rr&@AYf>*WpP508a8D|>&t(?di0~3q!cG{4ttB!2iszIS65fpyNX`5%|>O}YS --;ron@-rY@4=Id0jVM-jz*lCKLOtYHn`I`o?`Vdox+#>Sa?^dP#HB{Cep?Ty(n4uPe=yy8RZ`y0NRjU -Y=(^{B-`qzh6B2^ISct_jC2K1_*Pt%uP4pHQREVX1BU7nl|0$%eOghpSJq9UD@i5t~=xRs&DhEEMA16 -V-R*-S^1^Unsx(g|E7y)`9|aZbyJs1dX&}h=oqA1-5S8!!TK-$D$y@%AVZy3seZTATOyL1J)J1{{Fs- -OF0Pu(R&VoG|B%;3S>&Cb)96pRS)aFU)AHAg+H`qU(Z&3-yWY0VQX3OZAVS!Ay0|1%EwKO4T@Il0i+T -xYl%@kZ^+#y=d@>t5!p6>W2l!KN^iu%Sme)JPXmm|e=*py3S-*A5euls23ZIw^FATPfFRZ`#Zmo6YR{ -sURj7-Z(leqH16$=t;k|?RBY&Mt -P($ceEu5-IZ*=!FJ)33(Hp^y{39^kx0Vb2+zasF-ld9EuvG@Gz=Ouhk$+^9@!2oiVeE51YIZ;n;o3a1 -}Rj;ZNG!!VHyRuuWZmrej{_;YxBx{vi?z?qUC+XzH`G3B=cyazqcKO$qs>_0 -z24H7U%a`SFg~vdWFo22R34fuDmEpZh5Tnuno?>il@d0<}`5AqlCM;;8cjD9z41$X0_Xs+FHI5f70Pm -Y$pZtA_SyNeT*q;lf;aHma|t(aLa1J2*pTcH3=L>SEz{jUI>D%Vwhu0S!~bvoNnT4UP$b?umqj`Zx! -zC1JT|Vcr_dT-LC83^V1^2UV&6P&JKtGrfk-VEWrd6Q0P;nf<9 -J_K#x>`^G1dv6jWRBgK8Yj$}Ho-nE3YmF)~XQ#`C9yCYSw`JSZ?1=+9s@A$Ns%v10TIp`N2DI>b={&V -mp)ZICV^ofy#_?5zr`cl62MPz>e6?u+KX4my=SqWH+_Z216~2|M0hpTl-L^q@G2cRaK*)3uJ+i<~;y75sJ1bgdR#IVy4-(OP%%bBP -f9l66+Im+9c%Ir+nX*KCJRqE8PMz?D)TVgK!YBwdNb;nAZTkcXVy^kYm|p4%eBABf7QCt}4{F0V#8&5 -)0<=rL63UX_B$CJS5%=h}3izfvHY!QbuvR-*BEf|_Ts)39j5yKUEoYjhvmOl8+h+<{FQSLMZr~2MZJbW@fv< -zO4a_2eW24G{2U{p$q3V1nu#DrxEK6M;;r;X;J%Dg0v#;9svuFddLt=x{xfaB0D{*R+QdG|IeG5Fo)X -50ZfKd2rI*oCi3yzu1`bX=7WSqe-V2tOg0l99^FgHrrUpGw^p)bN~n(DT5+s{T%(oeey3fe6#!hL*&= ->HMT0R$U+kGNp=KyMW8D0?wDy_Asykm13vm=MET51?Z?572FX#HPZ~;E>2zLuX;p-$B*4dWHG7u5n3| -h&^aMu*Z-!B9Zh`)^i8P&&?UO@dL*P9TN(IA?b1mloh)G&FZ@mDe#Hx!zT?YCLFe}Y@$=TL648kFT$@ -(W6JvYF{cp@$JB~SdQ+!l?(PF?rz^0V*)|7%%paMagMGsx^6W(3 -N17vS`gP16Ns?Wf6eI7M3)=vgVVyC&gH7%*V)9}pOMfC3K6hm#HBXL -K_~Zjr9~sB}2xNh`(BSM0C%grcXIh&E%jiI@WdC!%GX55^J%oeZA`bcSOi$itZ47Kv$B&!0ZW1PDnwSf(@fRW~}Xjk*I?!J1&F -*+Gwhbr6Q+;G}eSViW2bl@OH2dOYkK?lYCGr94!HwI#G3&L1 -roDP}JX2NoF32&s>sB#FFj$bh?nTpFK>g(Bvc6o?G8JuyUFf#jQz; -kS)NTzox6F5V8)e5qQV)j?0a;1=*tl_5QBj7h0dR>i7(T--n{{&+#Wf77frtytHmsl=jE#^mA( -BQjDIw&k5joQSZicsdu(V3_dAE!y6nkJ{Cz*5^iDi0T8KJ}J0#NjMSw~n4SSh88(1MAJB=n=))JMUb^ -bZuT8ubOnfG1qEtoCk17s+Hq8V;4Zhf2?8Qb-{$1n%t3wW{a(4^^9J5@0kJ5z@m2E5Z!|9KBt@T_cr1 -gB74G%(5*hnbp^hy-47QoklqvIMkHL^1?sVOl;gShm_gVm@|q3IBjaXD_%)8#Lo8XGoe1uHzhl~y~3o -VkXn_kA?=D(fFGC}ss)f;jrqo=g^LZ2vmvXOQ-zh15f#XbDw?~xYVyMTj+Iaz@-O(=m%4JM(nh~2<(A -zMvn`>Q3G}AH-|WqjH06QFI~~$=hFl(zls!KoQ>dLQLCT8FLnir$4UDfD+K(#e;=QQEhxm+l0SxbfeJ2t^EA$`E_l}^$yB -xoA+R7PH>f-8&Favpj8+f9A);Bps_#8IzC6LcvG@4OZ=eOumA;9!kBW-m6661mlzbV1Cpj(l*%XN9HX -@n|&0Ct5oatLSiSY!Qe<4harRs5MzaYAa~#rVY!*;@E+JD4|}r9+#BAKYe9MTV@0%0i+QUpyRlF^t=- -@mQ?-77jn;?+&vNR@N5wcNlz(+>~9{`Y6GfH#gAtGTd`FfzA$6F>u*zpo?B$Rlyxy@5-t$i>r30XWl^ ->G?#s4XKSG=ck>^jJpQ&H4kp~keIoMumj -x7}S5;w>&t#&=;=(_B-H*H>MEW!d|Z^Wt0+j5&(5DS&q?vj95(&vMrGseuVP(k}ZQQK}LiqRb4Na!x({JEu -8d8o0%=YdWBLCc#%?1Xa?FpSx5Kti|RQ%j~ti=Q(bT7K9jle=?yz8g= -Y*QOF1?Lo+S;PUT!BmF&SK9AP6W1J9z2%%3wRy)avDu1aA5l(lqOo#5EXajs$7<^u197Izs2X|(GrTs2kI+|e6O`5t7ApnR8tywauRYFC%P?X+TpU>jmkyF^62;x5P?c74j -$PRB@CaRZCkp>IV{4d?i=<)-kscv9^Zcr4pBJ1XC&n9Zw!wx_K|t|rEbdT~xo%mZ%Qe1-%^NiA1Oe?w1wovD(J-okvNp-J=Ylq-Z-76v&2s#n7(xh{uW^cc>4-L@hwHiC^lblh+lbf1a1^6GKoX=k<35>cTkzyM}Vjx$J6v -)LdTo|C?N?Nf~3rQZ_Xe(d|#1j~(opMIOFpIC6Zbt4gT%)tI|vk;&bux!GL4Y{H#Z#6j4oERLV_o}W@!JEF0BuC1_$7ggNHwMBchTy`^6FKe -_OhiJgv{yo$j`^_02i35D4RtI1%%W(JxK+}>c+wNNGm+z%lM6MNr(!AdLhWm8%2FX1j5nV)zfd~AHJD -$0wgE!WvdPb-biFrXk3=jN)spaqmv(@68XmJW1a)^|xG6s1$76!7>X*IW~xL^?B%u)j?tD7b&nXB&xH -S*;C{j2-;uhsqgBxP54%CM4olK$G1?EG7z%y^4OjV}?$#l1%tg3P(4sC~*`KCbt87^WfYxxC=0FX+p6 -1q+^Yu_t}=p{Zj#H`)`zhx7=4dU#0;_mGJ2NO1XRoxzT2!n^X*aU|pW{PZApK#9-ns|A@-C{$jehQTm -xGk`inu~Sa&s2gM&v#bMh@U${P`jDuY5Rr=>DcW{$ltc1`BGCK#7R&xQ%bG8MUd@iaTDz7l5AeM%(Y#2?tDD@QFe|+avdj2O -a0~0qNa!8Sj;>cLK%JnaD`{ILZaO{#0^&e7%V|BdUNB=_yLMRBV^(%b$S*2L1Q+oe@KTN3 -H1#xTj0f5iKPi1aF-HLse*#^r5>(G;$8@?Y|c&Q=JkMuJL_lo}OB0^e*wPth#St8JB+22q*u{q6X2Ks -{r1K6jS=qFF@K3+xv*dkNxtage}64NAkN5{~M%FFYJ6g%#U-9Klg>R-o=gF+m~_Q-1}}{rYD#b-Twek -O9KQH000080LuVbTiBJewG|-%09=Cr04@Lk0B~t=FJEbHbY*gGVQepQWpi(Ab#!TOZZC3Wb8l>RWo&6 -;FJ@t5bZ>HbE^vA6J!^C0#*yFkD-dk?L^>4Kb|rN=Uh#69%{tzSv)QsYaa@WDhRC4=3j}BkC@rJtzo) -zB)$;(!wUd0fI#$^$0y8~5J>5P1n&GX-52CNG8!=OV4P8IW{!neV&~KJ~Eq>UEb}1BuVUJHu$ -H`d+pLgHBeO12v;?>LFzkc((89jPgH4QQ=@I` -ZBgEfcGdO8wp!j*IF+IYjOt!&L_0`3h4#K`>eW~7)gcISE0*=T7QOBNwR!TjIM{PH)Q#=*rt3E_>fgi -)XE+1o+pevb^r&p%(IH6pFgh$mFo8d-r?X7|5ku&O?3#N4Yi_!Jg->?4s;}XJlr<2Z6R-v*>D#I)meu -lBm|idO%d2+i_p|JqTOk^Hkm@GhFU6La#$Wf})^h0UYa01YH+>Duqw79f>7zV&5alI!q*0SfLj*4?S>o)(q3AYfaXhRpzVc8{xYA -Q(WV6tSAgCnascq=6@W8{lf5HWUv(WPxSTkXL{Eb+*gR`Mk6*G5SBe^z>!Asa2w%A{sAfQ?u?a36X^P-fHC#x;p_UQ?RtwFQZD&kuaqpe#9(2o?Y`8e6+l{d^K8IXz -9dyu6vdTB8FPtUEQq8byYXu6Un(UUN0jv^9leyJ_UWr>ae6WJggv -idq%Ri&^yo=glLCKLa{s@gK^qnbbCxAMp}LZo^6XCCIG+-LTwTq=P1S-(j&6+_GIuPCd-l -#`>Xs4q7jR!=5%&;9yJRtDHBY8q+mNm+(AW|&^;+Qv-xSt0qrGHK52KdA!t?_4G~!YK+xNi`#<-W0rb -B35+w2#gSB(@tdX8RTtBd(*s6Q!s{qMF%vwhgxogt8|2|^&9g{qHl!^5s`(-L~-wiy(ovR*5}s3q8bZ -Z%0q5&T((@8@JzCG?|SP;65G4C?1kQ(@!pATUeXJSNm{U4y9!%Q59psa@fih%4e^sCr}%r6kZqK;#`C -PNvg?IBet%YO4Iv8~!P)B~7Nv?64YLP*KAzyxoG50fKId+!4E~!88sX%?S)KWwa)h8;a`K29M_GJY@z -G9}$o_6Jg-SEBF3m_b2!xph;2lr#AP|4KPpuj*=wZ<(=3=pGm!%Af8jEOeRs_V|em2(s`7-|_QQ7xM8MZh8s9BtUB62`a$-_ -7G_!^w;?hz+u{Nbv^ah6Eo(>qg$;;@BQ{nb5!|uAd%YbZ49(=$V}o@T4-T7o;H%YTZ+igrO0GQ0Zw3C -p3-U+|$QtO)$!YW=fGV2{;8c*%+rd1^&UpbN{RH4&bjG(>;pCXAN -1m}{L!nGPd+j5#-=M40i3WLZ1$$#s*AD`**#s9Z1Zffxy?Ve@t{`lKJf)$OSyR1?Io>}p9nsS!GU+9Z -u^?(aAfT2lYQh3au89O~&H`NVzI>Yj|@XSvYT5Zz~g}0>7vTuaN2ox+}Lq>qxOs)lNZW=SdYSjs4KoH -a-0QTFiJ%?8vAl{=2#lT&4Q`K#eVkD%Jhqqa?!pzV|*368o>-Lz_5}CY_O6SZp4(JlcMD-ZCKa(Uh{4 -^dbMu-^U9Bsi>`*6j|z1a=)R9rD~HW}S@kRAm -dopACo&4r0uja4NAmLd)jYf0i>B!w@GJ1J7_Rd1IUgqe0b-GgT0G{Ps=wQ9^Lh5J--)MFX@;Maz*+tg -$u3D*U@uN#&I1luQf882f)ACK6c9UCFWB&ZUQSYC^8Z3%K0i??63>q@AhRbZt`X)q#_R~p)q%Ze$^b5 -7%Yj*sHnBeUGU~2@%5=?jj>`*PuYt`D1bLwCWk3xxOAbjC}N@JlcFGxSAlr~c1F -T^u0j0hFN!*vw1ETzB^$czroJaTrDA|`BeHpseIa_0J&5f4ogA{|ZP$TWg9gpI?%@H;K_IYRY>|ao;^ -bL|vDS`^piRLZf6;Blde=02g7&uWz8A|Oqf9ZHEnr~wJw72jW>ZVDk5Ko~;8bbiz!`ioI*Y`*{KP5KN -=nfW&M3h|ZkD^0SVnAxaTCg|XB(X8HK6qBZT8Dge&xa^sCtyA97GN{=t4z@jyNwE(y^fkCvF&jK>Tr< -iAC2lsdd8n$IZgy`RqRy7Mv?F@!IM{uW6{hPKF-f>s+lEQsG4HiTs_v%heLy51cM=olEjuh-ed1%)9mT%rK@50Ias5Ij_^8pxEws0t%2Q9-H)&=3 -GvvlXdKFk+v&l~y^Jd -`uENb-?iweTI+1|?l-B~(ZXzmgK3lbqmIrKQ%j*mKSKLUc`NeHu^Tg`}8)1u?Acb>LZLw6ucX8xdu(G -r#{nuyn?Axln18Fz%lu@l0<=a+ZCO1M#66f^B^bGh-)un(9u?5})1~F=Ra$i^3`C@U7kudmoaXtk|)j -*cTwFuzIt`Pu3*P6iC*L#4q_mK1X3bWDr{Q|#CkPL;%Ea^A2o8y4UbV#N^9*JpW38=T2G)mi{iJY1FA -bhzcchL4t>LT1mwz{qk`YaY%4kXFbjq#%y9o5?@0DndR55PmETDJ#*nLQcwwS1_D+eu!0oR8?%qaY9A -uiufy;;)BV2s9z%MVUrC%qvJLPE~L -(=EO7=M=6K}*=V4ey6j?G^+R1XX(xAlCmHuu@y)^_nr?GsGr`Z14aIzp#0^Z$fbba&9K=5Hz7~xpw>1 -usW{y(~tNA5(J{iY%4IhVb1KDpEd8PjAaPxC-r=hoPv@28OYF#m`e<&Z!3p>H{u^w*P|2~x6@%PP?hY -Th>xQ?scQrO+V?a$Jg=adaWR~Kw8o!-3w;n8)}EbKl9X0JA>dGKO9ZHa)#i@!2qV~Y7K~MCLt0VZb~}*Ps(Gm -PG5`hG!*=iIe$YXiF&m^8=V)JC?`|kK`+!e;Ty3%oErPI4cf#iG(?$FAuD34ipv+X3x1E_9j_A9qN&y -{9+3BV7ylyAtp=`5y(b4FdCG8yP&uchL-#M?gj)F8tviLt)-q{?OV85e$STCquz6N^bsbE=mjlw0BwT -Kc6G{j&J*oVso*2FPShlM}TC|A6dW3&ZX&*f~vf6dUA0VA}&soScdEfIqWD8{r$k4mG-{kh4|UKvQEb -`XwMN8y`kS|j|J3qfBM2u4x}TZ4cO`CmtH{s -}>NhHEZB;ubqm!S6wW8wfUP-DSZq<1Ls+>>c2%!Dtl_!*Xi{uzwy5$bMc%^qKockfWWVi{$44Rl~QLR -*%BLb(w$%nuZ70_`9*NY;wUQXfl+KZ^U*=8nEW=lc<2bWyHSr|kaxKLLEw7#}q$Sk8R*Nx+rm{h4dKS -~9cFqE9cCn*Eh39Kpbs<>67i+0%)u3ADO#{DVGkB%>s_LCNBw;fNV>M-mF7~kQ?lMR^QWCZhBHkl?As -QI{veWJg%Bt}w)r+7{yqI4MH%d3%SK>49UWVTyD3Zc(EO9|h$i=}*iUJcSP+e+x6ka6RtfJ=Pz$D$GG -{ujiVi91u)hr&21Qu!FT#zNX1$Rk}``M*@ -^;9d6@T5HJtbuLJiiZ-Jgm%)8Ezj~JIf)BqPlS7-v0DvsT_WWxGuK!TFX>M#bP=VkHaEDZMBm06Ni;O -{Ir{H9uo4A^#ghv@7x>fIxIlh`TWa56DcCq$to>A}Z -(n99jYx5bP&{Ue8kiBf^CeiH=Ta!8rKx3(G`E9-kOZr}I8s(&VUx~Cx#q?tK3ni2*{jyqyzI)G8dE$^b#9rMEihlN% -(>5FIfBxq?OP}X|K;9SC2=XSH_t=qr9(joFb;cQn@cFVe=Lo8Yo>0n)q9qeuFbWDduDU>^+-IO#DXn` -t|Ao8Ht)ClB>lMHoOX@%N=7`AG3XYn3SD0mBXK^+ni0$}0{tEliQ4vG@2R3HyiUwb5qPTv7Kp#AlcsF -R>#!7ZTfm?@j+NWfIv)3_LsXm~?P;A-O6;h{^>dY++DvQ&0sMPU80y!$l~h4hLZx1wNyx|3FqV0sfG>ZIiAY50BOsbt@bi!4IV?El -n^i&5zT5Q~JM)*Q-K$Mm(8wzddn{rh8a_d)kxw~!+Adc)AYIbKO;!fJsR6xpTo!mi#vE(fAKKv1+sed-|qSlt6MRGtq_z6)YM->IWw(e`Zu7x=bn_W?)@{DCcc+8^$0B^Th|azi9D$HXhPL``Qy -uB+u8tn>0NoA9$O4&tWTUn|oUrU9#{j+A%;2wNx+oG5}QPO)S|RX$Y-fT^3K(D!KZ78syI%+gGG(lMc -=2c`y0mN;9PxEfKFW4Dwf&X_vp4*a`FA*$JGLOIoCqo9b;2tG86&>YmxuT2fb!9qqDt86@5*Yo~FY=%$U}f&Xs-Fc`?E;=ATi6BVlA`Esx&q!p<$5CZNH`uY$R}nq -`kq_+L;vJhSD)qTOxKHWw3z<1}qJu=GHJ1z*7)tSZ3Cfv%Pmi6AeZN463^T?gW@!Gh^SS6wB{2#uQta -7bRG+cYM+Qxf2akvI}o94(LGRoE&tnY*Nl)#0Ro^QiM^4Y|_{Nz}0$Wdk)!utggY(_~Xs(`$;k<1kxN -4IP8FU`eGyIi2Wj#tWMUkDQ5@C?)Go@x#RYME^FTVA23#2zXw?BCM$~pLl6T`7DlWn$IX-u`>*PaD;& -YCq7YE+-L)V410S7fjC8nEZm|6_d6X~V!!pJ2OF2WPQ-XjUoW5!^u%yE=^hr%dbDyh9gG1(@(K5QPHF0a)DytIP2oD~bp77!#}cX_#%d-6Q4F4&=M4Nx1%uQ9 -s`jVfNU#k9Fj?}x-CTT01WQkipUJ`MpP8KVYoJpDqRY^b;9z#mUyCwycG1^`MDo;I~1%93Ven)p-|Xj -P`A0_Ksbe7AjJWE*V6@;GP}cV2RAtrG(vI(AySFVe;xJvnrd^ssj9m~@wGjg>F3I^>|b=Y$(t{9=a#Aw;~EWtuAb`Bw -!J-3brCp2ZC<*nds@(4DmUCxi`sU~#n2#RwBI!9+y;gJXnIBrvuhPK^wXL;@4m!B2?{s=V;Ad*)o3H^ -0{+68w|x5HdF+2^bMebAOyyi*Q^vn1YLjP*v -U^g;JJr5ylQUq9yzGsUtG8tzs)dN%uja{0=>$iD|aowjxTM9*o$!ZRb8WZ;k-~`pdRUtH3u#g*9x)HC -aXP+^J{zg%C9Y?Lq||4$!?JxG0R{f)eVS2Nx4fVV>H)J-8%%oI^jYasx|NJ=~QcA-!$Dd7yx+FXrT@( -ni@Gi5XJ)-D2(p5K4u^iSUpo?6cp5i1`Apg-7o?_vtrG++-I(VKO`ClAV5p8B8R6#;Z#s#{Ttt+n&f@ -Dd-5nnEItKj=nagkV&h3Hi*Wslx+Nt3SvbtB(dLd7^O_g#iz)d-4sr{i=J5^raj138s*&-g4i!AmnSw -^}1;zne@u2YVmIaW@3U`8Q}Iw@$#!DHylThFj}F8Pp9X2;x}SI+uGqTn45jV^@IoNI3S@L#gk&Z -1p4DxsaAmJ}^HDGZq@uvx>|*C+$>60oyR>)yO&{TE~tYOq1U<;$1t=BEMrl{ov@XqlyaREXxDSNw2j+ -&UiOeg$3<%Cl=-a>|(_HF-O(>d&WdNQ3Fr$aOkH&YSwxk3$oW#%gax$xJzBIlai;il0HP4Jvx}`qIL4 -cf7U;-dSYg%q*&>#zFGTGPw>eE!5%GU0Pi5*D0UcID#t&zmz$iOigimpgthe>AvgEWd}yg@4PT{<1PU -I*TSF*SAdeSBXcnCwU%U@lxS!+X{@Qv}%qBKFeZ9{fHD*uKJQS@HY02eI`Q2VHWvPWKO1@zEl2^AB-O -{h^7AJ_QAAs>ZBn4v35c~n{lC7rYk&*`1dYwPQH=r9?&X0`8k!UW&0%*(Rv48X|S%DtLfYZj1yrc7Q< -@G*sak&RKRB)yzICT;)i=IYGMV3+w>QIGV$M{HvI&fTmKo47xN3#zyRMMTYUxvr)F`KV`eTGT{KKJ|$ -y_yG(KqoL&rznhyJSBH=sCPs@JWqmrbFZrW?&;*dx-pu=nO==i^8kRvY^GP9k`G#h_eXwR(Jha;*+Bk -+BB1(!xgyax!2-lys2t!9Q;T5<%k401<^1{c8`zw7*Z3PV&$s(+^?X}zpVw0Egna(#r~mnjUp}KR@EW -D|;?rMz^4YIG`Q@*E@tN-{sp2e*$77;wtU!xyy5xa5%^h~0*+()_la<33TYTTuZAdvPKgh9iHy&iCB3TKB=qBP{q-8O -&PF+lR!g0|k#1El7|=a348JF+hx{rO&ACGK*J4lOJm6RliKe!=qi6u(U+!lPP!x&}Ii^gTdx}-$4 -}zj>+P%pBQhdu)+cTwXF{y?B(zz@gxnR8NTYw5{G%~!Ep~v16&sQ(qJEhn0%a>(5SlJ3-ePvxMxSBU>@CKs>o(~@C%uAFvRe5DjY(&J?$kfK> -pqj6WqUDpQz^8mN470oCBp>sr5bN6Ek%DrTIOrN!vY`b)kBTZtZO&oN8)yQj|-s&I_0a{z?;qR6z)`a -BfAjh3&yML^X~Fwo5X`@Fpo!Dv4zPq*39I__AxE8>g0l(TUn%sZ>fxoDry+%`ZNM5r3ZLq*rqIOFf6j -{#AY}A;Ge}7A4pu^m~3$U||Gtm2cMmJ=?v!Mh15zs|i-!eU{*nlMv-Kq&`<%o0T -%iONXBo?t2m020NEpFN;6?Vhoa^QEiEA1U;ykg0vDpdFqi4jbe{JK~0=#&!TGi+;sTvhJHDtpQs*F -ZD`IXAL0lFq(CF2XBXCEHnoF_r0M9Y6wa3$*zDzFKshqY%&uq3fisIcPYy{fi=IiNk(tnmJfZ0P2R_M -aAbJW&c&)zeq)4H<8d8U|Q&x1JV}RRYv!1#l*1_@mxfdW>OA{55a-qM~2;r}A7QV2BMXATvwLhh-Gwk -vI3th4?uXZSr@PRG45+ynyDbE?*9JBOTnvA1TMEc2RfFMf)5J@6aWAK2ms3fSzA8nw}0gZ004p)001`t003}la4%nJZggdGZeeUMa%FRGY;|; -LZ*DJgWpi(Ac4cg7VlQTIb#7!|V_|M&X=Gt^WpgfYdF5E&Z`(Eye%D_?xF}{9u7C}DGL`~uumJ(qZos -;|WPw6UruC$}lDOq6SwpK;P2kdoL7i@;ndtr5LQd$QS$f*1XktkEv8MUqX9I+=SVT}w2!pR*O%C#XlS*K7A -hm3$(_}a@dB@V+qQ56we?$|8BdpI^n0b^K&c#ubt>4$O+8^i+%UBXDG>vT7amD@>&w~^dhUDS>sQxav -t;8HuWU>${)Sx~A?kuAi-j>=9a%j1dIKg$m3JhB>Iv0-tvSS+H*$0AAc+SP`VM5ukGjl&^0!bjUiQL| -R)yE~4N-;(O0pwtV*a+|jMt{h*IHf!+#EE(cDTAK=z3^j@9p*VP4Dn^xAW%aCW@kra*-%FkT7Q5QbC{HbV0qIpLVkn -FJJWny}nH(@WBts-Z-3wv?8NNtEz*%C-|ql$)ym?zwg8WZ6T>UqgU?IGo-B~T*%4W29UP-9?n7t -2J|A1f_5vz<&H(7f0&7zSN{}&y&?ZCbF&G*eTMp;w0=*)2<(UH`JKdv!@KLK-7pb2ROjdcWY#ShYy@P0XZ=Ob}-hX`@Tv}QlI$X -@enwfja6sFbBwt-#JD`V@AAK&-AphpS_@18 -Tt@&uAF!|URsU*dnVpWbKm(co65l!$%BT!xD*bKW=0;8=7VxOk5`LBCn`IPga|_Fh>BprYwDiabtexr -3n-ucKt&Ocl#Kxt%Pd;6?Wn-L12u1Eftdl2KSxDH*6)6@F*8a@I6_Hfr~wRi&gM4yR$3|TJrsI{oS@Xaqa9Q3lY-fQY3%yshJPwzd9t2-kzig6D0-i~jhEAC#6EKsFwh&v>6Wvm -OOci7IHa1IN57^r#Ai53@$+A^tMwT;`U2R%8HMUpzQ^EKt$| -YWbJj!8{XJ`J+L@wQb44*L6!QF0frZJbFeQ;?+fK#Q5}6>~tncXNp%}5^+2wdI3+E*2^u!e_Asq@J+@YkN3>yNLXJ;p4OWRv+50}{{e!PEC@ab(N#D|2D)PS9#Sob^MP)P*o -~tl7{q|E;O8n*V-wQPbA{hv8wJwHQ(5_Suk2z0+=B_xHwVvs` -4l{K|<;`{Z4$M!@C%D8%D@OA{az&@jQrn4{UR2{mCcLyhmcJ=BM6nXcNJa2oof>q65qye(w!{|>R7;=>sD#2*5%=Y!VEd8g1^OD`JDHOzU-$ -Ko8c2S;Cs_hvHBH;4N56YPa8vOEUIruZ2c`VUY`0|XQR000O8%K%wh1+IQqp$-55u`B=pF8}}laA|Na -UukZ1WpZv|Y%g+Ub8l>QbZKvHFLGsbZ)|pDY-wUIaB^>UX=G(`E^v9xT5XTpwjF=(Pr*4aA_qF{343# -^yIzt@d+3_pl54i&aFA(>_HieRDkYucqUdMe?@v-9DciYpz~W%u3EK9l} -&$k#aH+4S-Ynb{ext}@VWW%`Z9m@&E>0auik#0+NFIhK5(5*W)pj^mUWgtimFgL)1p!3qbTJvtF?NRh -0p`4i)GVlzG|=CgYdYlKzYSW0QiaiVWDYxSq?1PKi`2Fsl`@Qtvj79c_aE=uXt6+g13;wE651g05@gf -DSK$uZybnwclnPWuHId~&#!;H`R4n#`PJ*ow>MXBt}frXBWqBu^uFK!#{TIZYL8)nc=@p|>NZoO;3ESy*|m(#fW ->6PI4iJk)slouv){^1tsOnRGAx)3F4{1yNQm@1B+AQS3%P?h~Z6$tx%##$)4Cf9?&KYz`YDSmlP>L3bfMh6!B6B4(@Io -6JS43u|Yp*nW^Q!i6h3f0QYNIiGoS1Ch}$?&Z0}%vTK>ZPauT*?-2`YC{Jqqv~tG}$dHZ_x<_6Tx?Vw -mv=!~WDoE}XtGRCFsw)w^*aB#a9JY+`*DMTy3)$oiwAZYqDDvfgK?DLsXbTRUHJrspw0R+F_#NCv5*% -X@-jwWiZ#hE{X$;9G>S0X~50LruhV_9EM^{j?Y_`T`=ma7lR19gPfJXSS2JK+6V9R}IZrVJ)7iB@((~8kJK_DC_YpbMnNGizcb! -wi{2a&1M!}_h?eP^~jW8Yl7`u3b0PhqQEnFz8IthrYmTmzoJdLVhbkP4PUgpFlI77p0-?ZtPOx0nCAd -3W*tZpv&Xbk@%`nK@z=F=Q-)VvQBKRTfD=X$PuNTk_Qng(8aYFq8(u&DcwTXe0u&bCAJ;R0YzTTEdGg -8ZCqa_5dep$g@51%{i-xyi)!Q4P9Sev%mb|&xS6zYw*OZu)6FT2g~ZlGUYtKIlJD&jjNO-2LmIqAwhA -rlr2Jx7Fl412s2f_ME-dY5KRuR4YX^*FBwrXy@d-q9fwB@UL(HGP&e(figpj210yz=^uexW-PZ|8R7l -JQr@l5)_W3-{;S>miYRbCqIoRM^lNw!t329#pZVEjFPoOffBJPI>-8B?;zZLst;EU2A1gk2tP@@+RL= -n^KCyUU^4fH3XtaHx(IoPm85TIQw&dw@zhW*WA5rgXLjpl{OiP%*JCh5FYl{k->$@0ExnD^|Ao`(iE4;dTnJHiHe=;7+<+?Y2 -U8ZV7cMC{W3Xu<05tG+B5v>A||5FjIwqq^}TA}{Wi=(O0Q6g#rLkp%}poeR?k-I;1c6O}{HBd?CKIR# -zFMs?bR@&OD2rY+Yfxlp48t3~OAqzVuyb2Lg)YAM=QOVqNxH}`hj|8^1VK%L&uT^!vTxujuDnk$_T73 -XIo8H6u!ZJqV|#AZ6sN9)HEPR=Ygh4NB}V;*|4638^1akSYV$GV!%DE!(JA^F3wYv_s!M+T0I5Z -vCmye|_J2>fF)pH-xc745QBK0V0Bhfq;Ipa^h^@)DiEo>1tJUNcSA^Z|P!ZAT}E-UA=;`j!c6rmGX%I -Dij?Q_(Q{f)oGAIr5^=Rt2S%qYgLwNM;1M3&eKAl7l%0?37V=CGA1gB${pV~&T?!{Lq!6ekHEa_WyP$ -lY>M-=S8y9JU>*`_dsA2}gIchNAd&`Ii9QkKQY%p&P~AaAFTPMp_K>I>s0UpnGt(fEUyh+sim^^Qh#I -V-UhC!jQ^yO~1%e8k*w%@32BeyrwJ&=7=S`<+-A)%)Rt4Tzr1m~2~lbKJcN*jkFog6qtg!(AF3~fU~=fk0y2{yCnF -qQwbi@HKG3lWSr86~k0n-1`IrYVe0vvj8?_-}REsEiHg+E~LNS#)s2&ZNjGY+60%uF!*zRNm -Z;dKHMYlkuxAVP1k)f`oZd%iNJt|okgQCHJ9SM=-TAGrSO*R5(`p)J-*48m7792PYaZHm}YkZ+V5A@^ -ZGM|59f_kVhmwTd?ZA9uTE1dQ5#%$zTpP($V(v!OTltE-j7$}Vre|UCyR?$eT6P^oDaSC9J+R5V7W)I -PA40!PAd8EZB6Y{7${eQ^9lpp`6S=etlm4azR{(x*mF0hf0BC3weY_gz1vln!bcd2(mdUaix)0G*8z+ -!D)*iuZlhbaph!`^7s)yVomH@NBCJ;Ld<-sjSBVu!ISjrDVX@CCc5_7NYZrU2@x9TH7bCbCVIvRaw4m -!e8VQwei>!G3u6_v~|kL7ONE08C~Az}zg!5=dn}*#)}M`Fy@u1orP17mK0H=J-_>E3S*D@osGq4|>r( -LiBXkL0PesTiG5XCB*EB`eLz2>1%Gke>LARWg%JYr5HyYQVNQFrDcnb*K%EwhqOISX|H+G(;!_-_L6d -U6M&;~NOz^3DZy`0<(ScIaBtNpBo&?6kWN#3lOJhytyFbJ8FLyQ=#79fu( -6rJ8t}V>Tw8i{#4M0^J5-1ub+P!8tiagA@81d%m>8MeBNYd`|v}PM3t3a0#&&V%cprd?`!W? -q|pAuU7Bl)IfD|BEfGG#Lh5j&U_{kAL%-Wy>C@*!|*|QZ>}J0C0MC28oSyyO0%ZyFqPwTv;Cn`yJ`f^ -TVq@rpzO(Ibcgtcmw~(Z`@PdAZuBUbl68-^#W37G+!&cT4ue3dXuBFY=I|;ReKOURDa*3#4gi8;_mI? -Q>}wJX!uICgUE+AU$v1}PBDBJ>?+n8r3F8XdVWVxeeX(qvK0 -5*(8jN^g>PsW1Jn-<}t@4L2F{^!MfGL;d&&-`0|@vfY$J~FNXZvSlkN~ry{N7)e3o?AGk=R)R#S+Xz; -4UG@y&_7RoewuHT^Zu&Ws7-;2vF)1ot`aJqbu}OW7fGlyis)+OJh{gf^OW(xOMOf&(Vw(88jiLpPU`N@=R1+UyR -G=Lm)$!!Td$ZD32)jA#%oVwmX91KoSAn!Hn+Mgu~Wleu^+$b9@%V2X`Y+$GMOze|MCLZQbsZSr`vTlt$NXnTe6U)=f -^}ou48h43qQlk9eDBZSH@@TWkl}^e;z>VxEw|p9*%f*+^F7;5$E986AjUjG7zyj1~{l%UATil3dwWpz -Ml$+0?0J3^fTv6J2AaEV$l^FSO#|eR!iaCNdcy&EnZvc>C$D*iIrHl24f1JQc=@9Su}No0xR!{&hI@j -9xC%FFl%pP?vKMfwR?A0;M$?RW8}sTQ-y`j}vUINxWn*C^wvsfplE2$K#1&CJqu?0%^K9}TP)h>@ -6aWAK2ms3fSzGmN;aS84002b~001ul003}la4%nJZggdGZeeUMa%FRGY;|;LZ*DJgWpi(Ac4cg7VlQ% -KaBp&SWpXZXdEHlCYvV=~efO`pxCA8DRap8`1SzB`g@txowuMsCAjZ;_J&mMMXGX~yH~+n7^f|JX-Rz -dWRD{?X%{}+ruQPJ4RY^&bH@4B7BqYmPX-lH2loeL0%B)to+mouI;Iv&CHf5ZPrm&ez?b)74Y|~-ikD -{(+wd?j2YPZA9@NCES2duE}p4vn0J-n_Cz4uzyQS!i5rgWs)sN#VOxy4tvkM&6DTVeQJzu22bTdWPwZ -$y>JOjzE~4Om(&WMwz>j{j<;=8`MBS`{rHx|+BJNm7Xt=?Pb>RayvR=x13Eegfa=tp~O7-NDyg7_SHy -@`jYlO4=j|3>WzZnH$KMc*xf+R(zUw^f$d#6@!`k2q%RVH2MUY>(qscfF1&m@T+#jY5+&=n{4B8bMep!A;f$(*x -?>DDgnr}jUrcz -S===k`YHw7o16{^>rr3;KtODr2w9g3c-$aBObgvuUC=AKc|NcGlsVz(W(v-ebP$(8L$pN9sCevz6Zga -Oc+#N-(O9q7op_7>yBdXD_RdC}BLS@{qBsk67Kd|L -~=(*t`2v-az6D60|~B9u(>e&EG)E+A!^)aqhW&qIx+7z;&ET21Rn`&t#sg3hDY-LT7*}jpzW8_410F-!HrDKsmvwo^E^$v63g$MM^vIM!Ep)T>0DyR0(O -O8NO&+Ig>*de2*eSxAEJ!6%`O1$>>ckrx9Z+kKl^15+H_6Bduk_UnPMk%)XoXqw}ZoIx`X@gB*dZwY$ -B&?qwpImwA1}o?H6=>CI$n;XtWB~;W<{KAu3saen_09?_+8c_KD~{&V>%wJ&xxL6ZE_*m#g-O!jbi2w -=={G-D7yZ7i;?gdyn`4cFMnJ{@HB%O|0001RX>c! -JX>N37a&BR4FLGsbZ)|mRX>V>Xa%FRGY<6XAX<{#OWpQhBvKWrq;8u3zH>-jNZE4Y#u>N|7o0V!-|?m&dBgO+ovr5Ou -<5mIDq8X{EPqWa7L$w$DR^dvQUOD*5L{^12wEl$`;=rLxo+A_%TWXlL$oYYm{0RY)+zg}8>;3Ya9Ovd -M|fp_{>eVIykQj+x)UH-QN{XI&#B0HPBjQTeEmbpi?)u5vo$6KUQc%yS!5+uioEz-mptRzSr|Isqve9 -YU(O1!8W&urX{cCP7~R;TmgA`rIr$lsD45-af4Ddq){@h4-@-u#-F3hQ0)CDN4Fg{hI2QU~k&Do<`c{ -Y{xHbrBYVMw<@C}ubrZ?fWkNv2*i2@gYP1V}P2h(lgslDwp)Vp|%VnBvq*kxVS|OAWrt8oe1Lhk=hQvyL-rzz>#DjK?rT1YMD#`TjDK`_Mmg~0E~wxmWlbZSTfEqZ+S=^+%Q@5Oure)8K#$;)NWbNv*hJ-s%~h&^ -ZWRz@o~|~xg|zRU-3gv2`Ns|HGO+N$^a2Y$Eq4&$4vK+;7}710XvxI|+|I0sN$?h|i9ad9iP(e{B}29 -de40uT0w@YUCNm%aEx{bw8k7TuC11@Eu-GWpUyMAkM26VA%4P$Y14jSCgkn0}$~PhCADn{#*9%9+#O{ -seqQQzgBM|PTk4fv6xp7TxlR{BW|?QIk(Xc&jnT+=7^)_f^Gb*gs@HKa)mT?2%ql?kK -Kg+TIyF-U78gv-=oiw*6r^*=di&oaMd0C0hdl4!E8Tua-o%D#s@}(`=*_a1$?ss1q?bqSwYbUS-=E2a -zc4UH;SCSI3qtk`6<{zZzOVmdutBRf|R?`NEIVWzV -EFd8U&tmdxOp+vt$xEoibMjl@z!(Cp=mk?IR}Q;@MOR#tN;ZrH>$ai-?nQIDDp?SdE3VhF)ecU>)dkk -`lw@lv&JCYcj0!BlN%%WxuV~A1^9UWCU+>FMfQ@5oHY&-Ez#jO*o0YHwh=O-kW!xWEH)Os -l5LqAMDAYJnCC8^f3~_Qg0U8_ZEgFXj?0z+cTCn55&Jt&2;FC1gk$f(6nw*KrA -IokP*AaGsBtt+_C@>mW@T$DcR;-d*NXm+a(mWbIE=T-ERMF^Y}h(J -DD~@SGs0obqNn7ey)@3C-s`=vZ(Z))2K4Sdm}9KkrUK8`ZhJyu`qnsEWDA)izC!MAJSOUg5(*p3j);a -ix6}RCb_TAMb{PMllKn5y84Mf!12d~$Q)4a^bOk!>Ex$LJqVblKFrLessr3Nj$J-rq)b$8)L&8?jOUg -^UW%NL4$_zcsv*MhUzP%>wW2@uksfG{x&#`px%h^4M{1~+TeUSA-k*1f{{{~P?0|XQR000O8%K%wh!o -psF$`AkmeKY_7F8}}laA|NaUukZ1WpZv|Y%g+Ub8l>QbZKvHFLGsbZ)|pDY-wUIa%FRGY<6XGE^v9ZT -ldKaIWG68+{ ->VKmVG<*zTErBD9+?8y($fBJqYGHIJDH?PuFlD<`CZtA3x*@GlvN!#echQ*mIO{p_k=zplRlm~UbvR2 -jv|2_@x`Eg|y=3kKxy2D-+w)ZPs+>d8%ubH7 -ftUmeoEMgNCr{QWi-jv%Lz_U+tADp3unDRP`=X6)SDPV0@})b#3aYwYtpJJ1=#WZs}(osJw8a-|^toV -pFNwSxZvYs*<(p8FRo9Nsd808*{aMSt_xHAzAC@) -+*_zkt5dT#<#^WC)EQ0kWX6x%YFEo#1+AS{WEhd0PbrxlJP>?r%}O+FRj4~n#*`$seQ47Bi*p^>|=7J+h*8^nEyKU<~ -WQui?azzSg7P!^vY4S&s)Mr;jXVJv}@U+0YQyqu>4H#ZXBTb`FJ2j1HHbnZ^3dQicn_J;B{7AmG3*dt -C8AI)W8Z8^cf!A1t(O?#|u&9vR+VWe+J#nqx_xvaY??I>Q$?Uyq_~?+TfLJ}5;3PqFzX{y`de2&?V9< -7yG%o)4hwIChQ<-w|HNy9|}-QWOad>sm1aUiClEv#ZdTCBYAbqbHb@EpWl~Ss|f3Vv&Kg2Z|Z>ErpW8}2L@Y5YOZCc$GrsKuAoQ%l%V`VLS+h6b_O`48&EHoTbL%8B<4!xWoGGAa(|(~ -K`3qsDY(p7YFnK<(l(7Tl$7!sE;QuFG**#wYo*v0a>xFH#aIP0J3s^wp)I?P3a}OXx=J(|T5ZUj(;PN61@JVUHYV-J -_3}G+fFZSuu5c}W2{I^6Z_WcD+}z+9LBwuwnB?5As0U>CrY-WmZr~~jX~{5%72t7jz;U%W%_(@{i8no -YmO4QpbO27f2OrPgFLgV?-B5ywuQ`TD;Ri1F4Obrd+zv5#IJVBBcW;@A -~b!3*uWeNgR!SOJI^;eA(LU8!*KXaM~CS+K%ar|Ekh*?yOI!3Y@q**eU1yV%M7vAuBVM?jPIR=2MasN -@H9-_}b1VxZ^9Zn1Zh1oDj*xmoz{zHRDydJkLq$-2J>hZZuH --T2C~KquyE)Xn7#i^7nI-wmpsmofbs5(5jk2>+mdlm1zL{IL8s9N7=hO|m)J^L>WHy_9uNqekv98B}6 -xaF`v4>kF{ZX8U(Q#DH8wV#NfCkHyA0MEz?}b=f*LDb%)!a<3(|3XaOB7hH!>lxB4%BVT6sy -u0r?a6%ro&k`Q&1;!etMmxE2WHG6KGWTYIBh06cl02xBoi7#0!???6`;+-QO6At8C)E>RF53JLsBNk%z3b;0*@l9+);W(o4arRz?^Cy<*9>2vLA*%& -xqsC@Z?e*ixaG-r3{ZxlQ2jZ##F+>Oe|p*C2`$hAxv_19LMLYFa~T-J9TKg$5he=)=;Kf4n)aN5#wp6 -Tt$}AC9Qq+(C^^A8hU6PMAPF!l{?HN+Yd3KFt7(~%^5b%N`hu1;|4BWJ1yO%Pf#UIv@Z08#8nDvb$XQ -M#mx~GBj!3$iH{45kxrU@ZQ30o1Rc29-*}@^vd;G5VP_xyFfQm&@01UO6!{^PDS>@f6*%ycdNL7{Jl? -xV62QOoo~2yt@U5oX*xc_Z3u~C1drYM#ToSo4Z<~ncL|#)H+QN++AJZhadZ+S+qmvZm@F~Ap0{dyH{Y -8>wX$4;+Jxc}x+SeK7(rJ>^=IbIS=0esxHeXWlg5+G)YO9b!XP15@ng5P)H%d;SyfHV+jXPvBNe -)XyB3}y^#NkAmogD8_DoVkwl@ZorQnMR7Q0}v@Df0T$|&8MD9ijXwOcI?MAQ38C@&0@SkF~{5XTk@3po5f(37oBC(Eo_mAXRUU# -Cc=QXM*IB#N4$0niQD#|tUtewANAdU?yB_A;P~Jh2vbcA*FB(t@B%HpWIOQ^@%{35(q+waI17&6#HId1*F^Xse&?jM&XjEQuMmbE -)p5(!*A3r>b_3r2uc!JX>N37a&BR4FLiWjY;!MPUukY>bYEXCaCrj&P) -h>@6aWAK2ms3fSzCl@dC}Je000&Z001KZ003}la4%nJZggdGZeeUMb#!TLb1z?NVRB((Z(np}cyumsd -Bs-SQsYJteb-lXnFs8OB2sxIm8}R{EZP{`C5t5#P%xIp*6hj})r<@{U!T(>+t}EcNu9)3@pA -PN#Fes|?z)r7_Wz9U_p?}=te`Iu7iCHP>NmjWws#;JbpQZ6KT`Orx&%yV3@nEFYE0an4=*- -Zoh6?e->VRnXzJV51PO7oMZxl?+WeqJ^WoycGC4VEUmy-PfLLp$`%Iw)-WwWYDSr?0JWeiYC>=d;#R? -ocx_$5>k!=cJ#g$Ni~8=bE*gvz-JsD(m^s5KpNEcNrL=k;~H#uCEvL0DZ;5>Q|3<;;5CM4F|t7O`XUL -K-<+)4~V@JNJPqnVYggn1%G&by2MeA!ispYlN!UdI1;fwFgZ&XhA=3o;k4`5@D^*lmH&(I$JH_nBZK| -T$PA-w<#obrkfp|9+Mz;Z~~dG4yQJKxkR$;_R~@C)#*cmLyh7}2ZQ -Aqn#!WuJ7ZFXz=jnBrL=;bHl8isbgJ?jVaEkAaPuFpJF}_L(8c8@xZ)ki@;pm3`j7I~XqQ53dG@a5o@ -#4$LFpeOLNB!Z|ARfJ^Gwd6UQyRvXaSG6MOiZ|eilZrmT}Da&0-xboJdD#D-#d@f5o4Z@6AEb(CTZNi -8iol?u9C@k8i9TQ*ik$>PrwpgMx!(UD`XUX#0O0;!r_pqyzmOvCu|?}$CH~Ret(hD#dtV~AUum;S$H; -#S}HizABORzPlNC>d>^@8W1u7+tF_kA^+m)Y<_q!PPvh~3-RO@;X@XB5UMA^o>vcShd -P38;*|7GJ;Gzqw@gJ4xaxDAyfFa^$uyRaWVdtF=>g!69eiL>)1|~&t3{I)4tlL*nLnVtN!w*wZhBh=w -vk_MScjfOsd93fT%3aG_>7={w;NHUtpD!aJLuV_yAOVR-0Qj3+xu@T-5$@iKKAK&Cd~2ucEf@2I}A{} -t61s3ZoI49)+6XKz~6L~=_U8euORI${+lDMG3_q*gK*td{P%U=4mCIM+azrAy(^6XSA(hozJ4CI@$W8 -0y>EuU%S_Kz!roF~Zn%#BjOM^}{AqU0& -QEZyg$7oi^)suW}Cdad_y6TvDBei&v{0VlJ@7?)pJN)&dQ8Db{Vfh2$z3eiON9n)>wQM81e+jnaku1l -UJ=RNv9~T58zGpvLiYH%7NZIDfal+;771NvjIv=03u*BLe?=;ZUZk1uIke{Es%^IRx)3yT59=FR1Dch~8?s_@g8^`U?t2kF?Bx2H -e@6aWAK2ms3fSzAhYIX~6`00 -8#`000{R003}la4%nJZggdGZeeUMb#!TLb1z?PZ)YxWd2N!xPUA2ThVOogQTCElsU;3vD(y-O6oiC|1 -X`|X+Dubx9XqleP23;XaZ+@pcLtgVHEHsFp3 -+jz_0(d@LvoO51mp+k4n4QEU!;i^$EFQ6kO*dgp|DY2m)$sp9~1e(MQbh;TNWpu~dg+~d7x9FZ?CYEI -x33t6VZel$c0(7UI_AvhXtxhFrSjX6pvl2k!JIcVnT`ouLE$G4ZgQJ%_bRUQc?$$Sd9tf?0$IbmSdt1 -NUJio5tCtO`K#-a&tF*<_f2{j&z6$4sXau(us35|EvbJj3s_gKq-#X?NVCG14w83Ihe;z>HbC&eU{Ta -0-tjK$*b$8;=U`3p8$VdV77pc{+=FTn{ZA31{DsPrXLkPANX)R|{Ly_mhMo94+AZm9aNW@h6qE7(6 -!O9KQH000080LuVbTT_zpU629*0Fed&03HAU0B~t=FJEbHbY*gGVQepTbZKmJFJW+SWNC79E^v9BR84 -Q&FbuuxR}k){!BQJr4+Dawz%D!Ww(e9E24mB)R@o9nxk)bj?W6pW=4)NIT)VLxvSHrAkr~= -r+iawk=3{;q|ofI=o^!#R0QjHKwB-cr#)I`T?_4T4`WNyH8C=ek@;A9@`^~NDkcs0na^QUp+Bu|uTPHW6TwU!4tj#n) -`?l8Qm4S#k%j($rqY5qIGv0NOKG!2ztL@Srsmr=+Gjg%<;1fWV1J*uQdj#cqSRb;%+Y(E*67Q);Vc)y -0XG;s*pmgwMDV2!!SN!?);iSPN~YLkZs#ygKGV!{C1oE!-`?K!nx3%n`C=Ne^`rcG`Ueq(NiM&`4e0Z -{V*?uTq>e*h_VYU(%Rrp6EM>SJox|kDG5jIzy9RM7@ugm+(c4ewet8^nKkYZL)!Un0ga>L;xVQFMZgt -2g6!CK0kTI!xM`Q2-9+QYlp44wpO9KQH000080LuVbTk(Rx3eN-p0MZEn0384T0B~t=FJEbHbY*gGVQ -epTbZKmJFJo_QaA9;VaCv=JZExE)5dN-TaZo;31{^tQ)^+oi0v(bL9t=B(19(LcD7194iBO_SQg(%|| -9wYNlI<**znE|L_S|!K3d3+|>O2RtWd<^y=cqLBvRz_n;I1)SS;CGkxGXiWGJ{GmlS@?u8P`Vfb#3fb -9EM>K@S>8+fd3PG9V^tbRw?T4O32NImz!=Y`wQLZAXl;g(^T}J)dAB*5F`l`B1zyH9>Wcqq+({9)I1A -E5NeZ=ouqo%XuCDC5VW_;)pUNdSS`b+APB5gRHUFaACsDyJByN}WCbS4ke+3j!Y~SJncRI@+}+NX%b6Ezjh$Y&@Q(3|>|z3xMHehDjd#JO(*ECjE%wzw0|~@l` -;A0RK^w@W5OPncD+DiJ^dW$)F_oT-#~W_8^*T;vF|L}5jVY3GRSPlx_3Hf@@h7ddC03%|UtN5-xVZBA -?*vORBU*9|d0nQKV3wDprm{?}5sd6vO5_5%|1|yE($gvU%h7h^iLJM?78xJ`Xlf!-0~-l!&zd%xiDF! -nj>Xbv0ajb|K?B)mU34H7q^ukmSR>)f2#jN_xML18MnzH+`xeQ3@p&ec2c?c>g{4D(WOs -|?{$??seSv@cu9F-dW#m -RcJl6hPpBdhsQPA1w-|D1`k_&t9d%UBnH2kLp|)7h72o(DKk_oR32CH<)sfp*VNiH9g|9hl`W^jwsXo -JH6_X^I!j> -7eYPTh;VQupBU#PL8#a{Qvb_S)!7olro)5Rm5JxyMB-smmAE&4iY9T~#7_hRU4^d$SC3J-o=cZ<%WYy -a0v*n#pE#2>FuF>af9JGp`bX?G0X-r_^wPAXvP_yx_+8EQEH1B8dL=@NodVz`cYpdRrWg}7Bcl_qpErLVe(9{ -kcphTQS2rXjoin9cYE;kJhCx0?|z<~KJ?3E$CQo+v*{lcw5_J$WWchc^VoJWWwfo}C?uWX(9WFb_z7GS<5R&yr;)mq{dRGOrJ=U+FHlPZ1QY-O00;of09jiI95 -r?I1^@u&82|t&0001RX>c!JX>N37a&BR4FLiWjY;!MTZ*6d4bZKH~Y-x0PUvyz-b1ras%~?%z+cprr> -sO%gq2xrSEhljj%_yC=)9K{WX4*_Iu7-h0NJ2~!90HWAasA)B3xXdYDcMf?5glX-1RmZl-Yym}rSxZ> -@q%eiSQe9=EN;k7=nZ-Q(|hvb;wAY}ra9B%S|md6$p^Mp0i|>@SrsxRVYn)FS@19$#)_!ZA~Ma3yx{O1W-R5y)@_%Ekg7j*%A#(QKbeFq&w&UlqD=^_$O2ptAVZOGub^J=%n -!XksDrh-;We?SR9dDX4ahqUF5 -HuAPS#S!D4v3+a@%JE5Q;EGN(kksOgThLOu)351HlNBiO|T$D9>bwz(JsQ_u<_=4bX2)d)0E|Vq7IW^ -DL{7`eeTJoq^~-qW}PRmDpK%M=S04rFr6s;;qn5Dl}LHi1duaBF?_mR=D1G-;5~*pO)R=sb1Cb;UtLv -AwJSF(c*ZJAjPjRNPyN?7$LywYYPutzXxK&RbgkrhdTMM17xs)KQ2hXN2@YPs@N!KjndgbD;FIK8J3G -M;ai?Sw`*}YI4oym;SH%`d^AYDfwGbEnQpVwS*KYSS;3_&+b1F`Zy)Qcu{Ef&iV2F!L80P#{Zy_nRu< -+LRAhw~Sh;x=cG2BCV@%bC-DEG) --nd0-D1+jJhi1@qe+j0jnh$Q|z*(mc)E6(@g@Uu4Ez8Q8{oy`XUDC+*<=?uK0_rcrT@HK*d*s|Bf`5J -Aq%ss~mvIl65G4kj|M5XDV*5S$vk)TRR#$3}L>9g#teaOk+$b04uIV}`|FHc8)kLM=<3D}_4;G|kp$7X -7^&`RL<(mYEOE`HkKu2{0iMS{&1J&KyNP>q0S3Bo1nTHkXT#|Lrhk9qHA$vE;tVj_v(LyikLM)H6}@w -3yuY*k?y2<-I!Bc`USILJ^2uR?s~e*uVdnJ1sSkoc>3M8?8?PVO+GZF^>2Ee4m{+(_Hq?o^hr>YO464 --HNh;@&SP6J9G6V9_zEK~8O5$_K@hokPX2=+~+~&vu)0m?U9sud{=M>LE^!WvyHcJ8xNIM`OpkF=pP9 -zELyi6%LM4983WlLmPknewZ7jy~KUY%Rc`;L{3$H`unWXCd1wk#2ComU$_ZK=FlS^bvPk&z@af&*q%f -i#E?1-oE&`l^1#(`rw(U(BJSTZ8S8sDv@}0k;3MQv+g93#&@+wV39u^g6lS&3SV-$MrJ9Fo-cZyMgTD -;jVc&v)X*zrN20s+0ds=-m)WG2`wJ8gVyt@(B49us}{gT4-b=;Z_17+r%fVk2DbJ#?yYwlcMt5`cDrw -ezFN3Hr-SY&mIZzdUo|E!dKcmKr`;RJcAMiybT8c!JX>N37a&BR4FLiWjY;!MUVRU75X>DaLaC -u#h!485j5Jd0$6_cJI(S87l9Q_SaVYQ9WlCnzpd&^2ZxTneN+nF`STvNvCIbseoRu^S~B=ny9C&V)bC -yc^KnAjR)h{80dEV|fDA{zK!js~}NVd=)==T5wf)y6J22vJU$QWSsa0;;NUZgER#YzSpE0heE}-Gx=R -s+D+zcX{-W(3Yb>vujc~v%GRVQrqt>eE?8P0|XQR000O8%K%whzrM=1QUw42a1H4;ZWMy!2Wn*DV>WaCxm)U2oeq6n)pPxKJNv4~_!t(U=PaO|}K-(xM5_4Mic)5-o -G3NsXkO8lM0A?jwO%ZqLD@_w9q&moiNd3FOEU3Zd$TQdh@(@MVK%2{w)OQ9P41g_W;EjY6~cx -Fc7vIVx(or9XUhF8o9t!$Cy=rCIuy@9e^b*?kKEXk)f4wzEf1;Z?hYf?82I_rP-zJm+5Ybk!*yHUFtJ -n^sY{#^{winb`XT-DkXEvvpTa#d`(u^6L2)-q%Hc%WY{%V>xBF$+~^3>Mul0=$BUcTab7aNHskEpycx -0TLDz+wdaGZtvcIxO;Q+bpP(5{QdFQ5`X=2{|77}Pj=>jXk?XS*1~(y-i}ax424eO=X9O{{#I>NUI#IKsVhyfHpRR`V649wt7&0Ek1|Mm5`i@pO5x99$e`;2<1kbd2A9ord(sn;{kvEbaM(=1bq!ah?RE*gtvCdM%=uJ&}D1-yJ -$X%NX&+^XG4v%SXx*Kzoye(FXJi2ahLi62|r2w6M~~^I@=yJ{LX3iiW$ler=L_;$YLbK}?F2!TI#WL2-gskAh7g{-5B@{KN`n -9hO~2H)Tj;axx&V;4Vm#?hGWb53KN3y_^)*%NhoAR6d4bopkvNJ(~e*TU3?Blj$fe^3+JZWb$&QSwFH -bPUZ1LUN&1K*(SdGLzg!3%yAwbz`mWl4utg`hT6pO)>8a?%cI5eSns{o5L{fCLI -O@Jh|AXE)HJv=zdOl(zV#A3!^iRVyALpVIIIc8%BP`c4P9HDf=oJj<|-uV&0#)S^Yp7fQLO`u&Ilk{#3$|3`cuPItZsH_zA;#BN7cQ7KDeVuurrgk3w=+WGZ$Biy= -M7MS^UyKC7sHuc)Ove)~=IjEnHhMeB;kQaYxp~hrLjF>C&zAM7l$(G5EPg4W%OS8n5$azGa^bMPuosi -QzN>}~4X|EGrpS>a_3Vp%_{E^&b{R<^_Xw*RCU0wd)WO(r%zoqEQ`QvW!!cz(#bJZ8gw_GlMnr8n3P) -h>@6aWAK2ms3fSz9J3(lhD<004jt001cf003}la4%nJZggdGZeeUMb#!TLb1!6Ra%E$5Uv+Y9Uub1)a -Ak6HE^v9RR^5x-HV}W`zhdw~y5<~{hdeE~B%}lmN?)4hr4*sH$7l6y$(7{Ymh}Gjj^r=NyL+J|hGpgP -eEu}^lWMIxV5}bOpph{UU9YqSE~T>E3MI{MS38|v^f?pKqL#ePJ5^z8@=9ndEgN(@8@|_NBTo1rYn8! -I{_}6T^~+KAxioI^g40&Wr`g@lw7#Y~SV5W;*UXwdw+*w(<#Ns7R4$sd-0gN1*083}xM4l^Z{NOQlTc -%g8fA(5RqGH|BzcDiFjlAV`bYPF*a1CBl6+9|gbu`29ju~PN(nUnHITGJX`PXfk2#rm(=||j;;pE-#e -E8J2ciZcy(k0Q2&%+P-j2hgHx0cOJ~g<91m3~uxz_yAeKYj&_dmSI`7}Zr%i|hz5pGjD7PSp=U%-C(- -~%LIHZC0E&-bB}e$VeN)JoNL!ZJ}kCyU3_udw|}jzO>8&NEJ6(GubAkC`;wG>5pV8Md7FJ<4jEf;X9y -HMkmcV_XBgl>9Uh>Ho+(&e;sd@lEM}*2otrC!t=EP+Q#akC}tFB{H< -R9Jg9&hLy1V*y7%(L6_(TZ=I)Qat@)-MY?t$c`Jh151gghy4!?wF(JO6;})OvfRj>sB}F0;=D28#!03GhV^|(&A%@{Zd!OQ$cp|MyI=bfl!mCS&Rab^P_&w)Z3j8jZL*d}HGAKIy$g>BSe+1wSIs -dmpDd{i1ffB4l~nGDVd60veDSqmzX?CA)qfL9eP1d)fsdjsPrr8{`2i7#=jaEVD?wMNq`S(%ZWy?wQ@ -_pUP#7CPk8Y@E$-!I={u%Z;t*y21C^Q~@}$$Uu<~eHI3t*RvYh3h1b%k0;zXI3mg^tsEXC1?QID$&%o -zl>A||qTJVh+~!Yv;gfcClcVy9kj*2dsXH#4Aktbkf%@VF)tCjNJ@8O^q+GdTsw}{kin;;4ua5geKOg;2t^~x#wnIuQh3wT;6?rBlALxb?6m-+8@@wqwlgy$6w*?_s<%kZJ&xGid`3b~WY -gIktGkDy7hwkO9*|vVE%_G9>Qd=)lZz`n5t*g&Yjp06#1u#tddN2X;7|RT1@UiTEM1WtZu{tr{~B^P> -F5fc^TF`F3J7Pw4gGX5L^9^@XvJ9RA_y675yk^GIiw5MqVQaNS=aM{|sK!v~id^2MI0Z>Z= -1QY-O00;of09jiFSX{=40{{S<2><{e0001RX>c!JX>N37a&BR4FLiWjY;!MVXJ=n*X>MySaCxOw+m72 -d5PjEI43vjRh%INcL4p`)fucx(06`vdL4Yha1T9fA*EA{cvX0aC-#bIytcAB}wtBE7^337z%&oR&4~* -5p2P+r@xo?g2z_m7>dug;wl77v&;mwNS%^h*SCXQCN1(VtftGOzaDLd?%gY^agj+yx!ZKPrXB -@m#uDF%f2d;#^9XPTC@UcN_sT&q_g&Y%V5qGC=rVVS9*>JT;@+4u5D}@*Q1+F466-~pG)I0RbZyCA4@ -HX))UEB*9E{2fnBuOe!0~=wQ?}f`OKg72+IB#=U{t8Mue}x~{>y9@~>T*>df%|?`*cBF?bl5?nJT?;F;2W1*r7D*)gEPyL#aLw$XHa5D4azy7LJ4nAZ!Gkt-Pvl -X^oV<0yht?aT9K3i1<(BJ4RO1B?{75H9bx#zjrNZxQ>wW9it5t*SVN;Z5ziN*yUrn4V>0A(Q_59++%e -U`Ip2ylRNkie -_@!km?zKML1TVLlDODfb0bPWWzB)B9VQJGY(Xc@Wte9$*2sN1okw^)ft<$z-hv|FbAu|qw%}FQl -sS9|x>0P`rQfK2ii@A9|vzGN}B^{}d4d8r!%t_D(#Azas&i99DDE^-a)BBE>(wKP0s<;y~^dB=^e8Wb -8-O(Vz9;Cz25o2FnA8}D&DPMvsh2~b8Uc!qUsB}?ae$xMjYkPa7UQ$b4#ZaM^#itLi-?I;IKE7U%GkV -^gqv5l>+f^tBii^yJmxi{Q*Oa@=mlrQ13?zzAa4aa_%7jxOca(`=y0b -e;6ceBY4EPSQCN6gUoquCkeIQ*9s>rN+6bnB`HGaT3q#WJW$>I%1*R2^x=h0LPCm@X)qIG(#G>|JRPr -bt9tv`9}1^K6JJX{R2=-0|XQR000O8%K%whw$^TKNCf}@{0sm9A^-pYaA|NaUukZ1WpZv|Y%g_mX>4; -ZWo~qGd2nxOZgg`laCx0qO>f*b5WVlO7$^s80YyIL5CaL|q(&2qVOX-Z! -MQAKq=6Hb7umntAhP=FO0cJ0I76qxUH`bHT+IZoW(QY;y-zhJxH>P#d3$uB0&EQqL`E3oSnt -^-O=Y`U6)S}JHR^^>gdJ4kN$Ltu(<)!I*;V0YoZ!drN<^A<+R$P3zc>V6?{MAnv#pU_W7uWQHZfAsl2 -~d(OJl6dD1w8Vlv<_JSyfz5_$=I*oW0N2GpmcP;o2h!@a~7Zq(Y9tKt3p0_soBMW)t4m6D9Po6Ouk05 -f21DhC;2C%blMB+-F)-9HkQIG;=)ZW$a`z5q4eniW_FBCgu9!~W|gd|SXt5c(x$@RyEAH)^S931JNk< -*jh0}=Kj!rGC1U!RVv;1UrS?{|G6I+S;3S!vlyif00ZUlrl)k5+m8fC -qHgbfB|jSP}Rs?L5fJDkCtkj@(e!NR#QV$*WERWQR4~OCRZVW(n>w?Hj{2kxM1KL=ffBJf&lN^lYgrz -)&a4Mrh@_P{ft6nt8;qunb6*vJ&j^K5+Gh!=v)n^fsXPtte#xXk^=Gw34LFYOH1KOdZo5?Bga8nlbj# -1qj5TaGx~H;h-rA9~05s+9staZ*y_eecdrPw$m)h6Qhq(|+BHQ}~=l8J#|R_WXHNY_( -J<8$fphtB0}BG-%}6u0s~=x*wF=yELjATWigtMQ_$-pc0A3v_~N#JKxaB)3Du6g1jW;k5qQ>?4U#Ixq -7kT?Rpb2POX(Fg$~Oy4D5WIV#u#pNaqWVT83R+Ai;u}6~Ifw0AsE<1AN0eHdL#7uB{n-NSHAo!SHEJA -dbVHCZt{}YjlU_e3HdNX<^qzD+MW?95tyk(6 -;p(`Q86fZ{b9IZ(chT#Jsna^|Ax5}rZi<5oqEC?3hfCd&J<`K}nTFtpBTM_{S**=0f+dDe=BWJ*_2WQ -3BG@3*E#kOkT6VLU0)u^&bC#d=@*`o1;#?e0x`fb(PM)Q -ruO^cP*IF>3{E--Q^M&=4P*;3amYBf=9jIhzBaYk?}^@UDzL@gwX?z>poS6tRSly!Zf@e-3IXTE<*8GVbd -g9>{ZaQf`FBrGB&FQvQYSM@w8LsyQFYgUdEVPzk<9PFc8oS~mw3dOT0KMulD1hKI-I^Ncp%xMw@V@Gn -D954bJt#EmUoj5kK`F3r$Zfw}@EXF3U~;&65N8=}RfW40e;_BT*V0|XQR000O8%K%wh&)r~7RR{n8c@ -+QvApigXaA|NaUukZ1WpZv|Y%g_mX>4;ZW@&6?b9r-gWo<5Sd5u_YkK4Er{_bDFIzVIJc@<}`*}E>_q -IYT11}(COydVLBgFs7^%`7ddB(+{0?#K6;q2Bzmo9>6il4pj)d3l~07tZcPnih4aok~-oca?QP$g;FS -25n0}nY8bUawkK!?H;WkJ|&e5{TEJ_x!rY7zVVZS4#})BDr1w0TxZ=$|5T-NGFaD1l|kJHqt~54`(`I -gxlyfDP`j$oMs+f8Pkt@DR_n=;O8U!Og|R-^AKlb3 -GCGdm`QxVy{GQ%cntX<((2z>}4Z@jYdS+Dxr#kVvjH3zpw>fYA-5hGv)n)M9S9MUOhENe62l-!25cq@ -Fl{46W|Bi%C&zagBbU`^zDkB43S*%uKy4U3oms4S#KpDX)FGFxMi<71Zq=FCAd32xtaPQic;)AQz -gkR!dHPlt2yB`z_T235h?gdWqo>xY^k=>R$WMS{x+e{axKh^%%_l|j_-FF8<*=G)87<`jkHq3LM;eWfxj4PsK}hH4Wr6VBRbQ|J;w%kcr(V-ZD3s+2R$8Dp -qm1qZjvDOUINukdwl#a@T4t%JDumne+fBq!P^wNw8%xNVgqxZ|R4<9{rz?lV;dar6F_bJm@=*yi`OG+ -vzuv|kRVB5(&n3RmxM-h!S#r8oZ`{5;j}N@Ec-yb`@87YC}l1e%PlJ9eL<-nK>Gb6sCB2 -2Lh_KLRZ3W|@DL844#2yDBB%8rpdwzWeTWFWtsp6F_ly_n7!)Kr1Tr`0sD=KGW&+UGz)}GI?S2Et+)w -70oFs3!tI$LLEtgQ&~im@QYsYO=B;8faeiPy~v!BLFK|L6uL4EXdtKes`PlyNpr2ZA%ZIDDcAlsc5D% --Xkn|pJ)$N7KVIlS`gwa=yeX;^7_+93q0`xJhv>;ElrxgSsEDo2z1 -05qL}r?kt>7tU2&cl!&~iEXKP~+K}8k%zL#2~E6<7fx*)M>f^or$6k}}kwF#(FVyoFu4# -iO7C!O4AW%5?D|Du4ZF-MY7CMitqx3k0M#%&7Ul-aD4L99wUn*)L;P66#R1+SR0WYwv>K%R1WEHD?O-)UP`S?1lh-L|L`X+9RcVw7v*>H1!vaP)om2v{2GY5wUdWA8YDe%) -MQ=RHoV|q5PSho#DkvC&Mxz(KO1L#oPJNJRQAj;G$bkjN9MM+!M}p-6n&!fm;@6w^Z$5)*omOq5G5RB -_p1%I1F*NP{VaYPLA5bnJM|lDrrmzKNHIHp)EDlBEOC~7hY?4hzvR>a$PiBY*`C -yFX=PRunFt{on5jj7E^#}JE+FN%!%Ivk+hEQEUd$z7Oh;Dkag)CL$ESDVkNElJyH`Jd=pJ5ue)*r#3=wA>VvAQiIi8tyh{=u>epL<&fyc -&kM&SvVmnfH81eCQ_ndEF!lMtGXn!d+6Pcfv+CP%h@p`Hp*+v?30#9fy(4VdPvxeP==tP6}-9N7VA#s -#1qa@gj`wXX3AwBbD=%7A?jUJsPESCN??w*}Bl1^KGIq_K$Ah4e!;`>tVo4G-_Z()rOKK*#KG0mo5)6M%ua;Av9%n{f*zZrX%@pA#nS-lkhTw -*24)cL<^BFT8}gX`(B#cW5(ih6)l48uM; -$<;3Q6*?*lU*@dQsZm{2j4&JLg?9$G{QD(KMG`oS8|j#XVYKn}FwA29=lZ$L?5tPGo|haJp*eZLwduQzj$Y5wKrAe4{Z|FL+UQS5_?;kFrBu6ok&gIe3>plVvl<1o~_h4J*A%_-``T7$TNgtMtU(t_q48YF_et-&tBZLjTdT;m-(q1m3Hm|`zzVim -+YSVVY%Plecq5Tq6rHjZ`kqm^xZpolzjj?-z#=dpN4U=tgm9~m}UC|F|J}vU=@%>q5Hi}-q%`7-x(KB -Tpv°tpolW=BenS>4iAbjQiz$@;^{Z0|XQR000O8%K%whBZAX&P5}S_u>t@9AOHXWaA|NaUukZ1Wp -Zv|Y%g_mX>4;ZW@&6?ba`-Pb1rasjZ(pC0x=N1=PQPsTxtXLD70Y9ibAnPr4>cisM#9GCLzgIyIP!Y!T9!y1gwtHHJV@nPCjACvW{nHrt&%|;0wJG+ -IDSpy*)$$ahHoQy1j_b1MalELb4gFS5OS-;!Uxv|#@G^pEy@CSY+ZC{8>_;)pb?qPZ&eY%Qi;z1FQ7A43MaO}G1hS6gp%{PDJ -2K4017{E!?9Jgcvv!J_QJKIBL;~}nAC9Xx>8aecacJBai1$M$S+V!0|XQR000O8%K%wh_fo6j2nGNEj -t&3-8~^|SaA|NaUukZ1WpZv|Y%g_mX>4;ZXKZO=V=i!ctyk}B<3RMZ2m9EX -vQCQPvcrra!k8)Yirc`OR{w`aCt!rajRd1pho%~FAroZW0*Ye_s@HhD^VWKwv2!$i$kHn-~x0h5wuuP -r7s^6MS-1*iWKWK2B>AJup33EqU||ArL1J_qeJ>ISh#>TNk9#^I36vLGW -1uctRz9HvQ!9TWlMG|YV!Uj?^mmYuI$v1D2=948bhVJm1Q5$tfo{_+Ui?b^W`gL*l!Qv@n*%Ftaj5{-l)@Iet*Fv@SB8rp)7IL?T+HViWw%`l@2h#!+pVtWS@@rpI-M6;vB*zP -&(h+@;&gF(n!U+S-z*nJah9K-%Ee-No{2?zzL>96Iv-3WO@F*voWD9*Enc5FdmsSlmF#`Fyrjua)o06 -z_h-Zf&W>BNZGZoAQ+MleM(A=B^Jwz!X8rN{>w0zd_TBp1htD@xpROmf{b5*tavs6E=fQI>JWlR2hQc -}IGcez!lP10>mdWCLbj@{4_jh>gR$`h`*~q&t?=sm~ihi*GO*<<;w5@Jq`g9W>zKOE)i?fTv{<_*Z;I -k`7@m;7=q@|@QbAv%JX35{VK{b71`~7E4@%blcT2$2`c7yjN^CZUbrwFC -R?b8fc*~T8S(lIg@w|s;b>18J`LriZe$%@e_gF+G2}ESVO;I`7Xhdebt5(?NYR^^CRYU&s+6^;q+=Zp -$9%|Ft#<@f41vf?w*wWzul8uD@21t8R#cN90XxytPQs}g>R}yDsp?zkOs?Ixoo;VgMhdk-2O!LTs)e;iKzE~)B_^?DT?b&}J_Oo2bPL(A|UofQ6r3fzu^t)6|cI?K6DD}l ->{NUox7YVYh+nC<|*TyKp>=xF|*vo+zu{nxf^6Um&6-gm_*&Xo5W~Ls$_1kb>!ndV3ib*vfaQxq+@D5 -QDhzy2mt1#IL!BC4hd^z)!5`|Nd|VUnfj&?nS|m!sj_jYhHJl2rcau}M%u@X;fdMg(8jUdhYCB^(wn| -YHG-q1;V{TM1x7&6_MwkC93rG3P&y>QBlvOWCEU!bC`SKu2=tP(z`X$ZrdB&7%P-Ce=qYAkQ7Q!-wi#>%rVy=RFwUM=9;8Zdk@LP%sq-Jx}OHmmTZ=1u+G&%1Z4!#}Koak2}F~is( -!aJr8KyF-_uznfu_nO7B3TKL#TI8OE2%;}W*M7hwI6o-N~?Vza?#cLwriY@n`H3H>%k*IJp6lZDBaD=b8#rZXfnYp6NYM8nnODr61F`+oE`lOP -)h>@6aWAK2ms3fSzApLULxfN004#)0012T003}la4%nJZggdGZeeUMb#!TLb1!INb7*CAE^v9RSZ#0H -HW2>qUqQGZDg%yybuSn^WLTD^tuZty(yYS}7&0wWHfxDgNh+?<^}p}#NXe3=bRAZINc2KJeD3bK<58K ->1xu2wa+MKD!emiu;}};;JMN@bb~Fmlw%l%Wxf*3SCDnN@QW_BRRa(!RJ7M@L7k-dClu~W#cXxlizDh -1%US0k^zkV@gze>f;VSY1Z*ZNk#@@nK~m9mTzsQ^jwJQg2PQPMg!&P(9&O4@?EbUUr@EG#r-b7dXRb0 -U6aj5b4vm6N%R3uznHM(}l#ZY#AzYqKuX*@(f9aDoTwX$h`98jYT?yR86_R2PL%Yq6$*jOQE8bjBRKE -Hu1vF}vJywGr@0TDIQ`B_4#K0l?b{7!6=}%K<*6vGyk>9ouX56e3W;;}uVLdv4Ycf}#XVS2CCG5RaZ; -yuG=d-(4jyFW$Ul=WJ}Z{Mq+EjEUeiFN8%hwFyYhNVhF4T;_TYq*n)8`s|F@YjQdWVa=9Ha3IZGE&=P -ZG6syv4^tSgEs_;9Q!7}5JC+M>K|+>9A}zq28DS-usOYgL7LVTC-QHZkAih^`fbo4zc}nEJ{qehS_5C --`#tX0`&AGJ{abbN}#^dn?+Z$e%o|aarAbZlSNOu+~7mRPX1kn)KPU*cm+i -1OJ4?M2~h#{=uFV-T1RFO)$B#ErZvuQw|v9(Oyy>({Fz+k#sEZEjzN?8#MfF#RmWZ0XaLkVjcE#_VH8mP-eBmK%h_pP8d@rXoIAtY$pyd -Ooa)tbfU1)q&paYPuMHI1BF9@34SHT@O;k!z(OVgiwMjC*$5`yR~!W$qBQBj1>X0-VS%FA#dvR`2{GF -tBaay&!O5Zlz=*}k(d^o3t=NJ(n$T*{hnWSnQS{ov3(i6pfx@xufVv`@Gp*-zU16xkfkWNA0AXT1s2B -HMnrBQ%1gMLKhhcj44k|4M1g1zM@&gk^2_A}MEMlNQ5oiI~Qlv7IU?iv>tei~oZyi`)nK51cS{~g;gd -7r|8t6SSWDI)hbTmF298556gKzM`URtaiU{v5oSzJQF0BbRrE0`Bp+F_bX?7J#%BpT!+ -k8w_S1x^3ryA3+5}g;oq|xOA(Re_pnnvbtJIFNtDAM37ANmOJCf;Meltx31bB9K)KDU!pSIUuONuy94 -W~X4bDxzGdC{zQVDB-5ZC-!8c$>f+A(k2+3dPuAGC0LYZj4B@#oW!Hj1ERCiI`0%bztiL1N9sczZsdb -dt#XA*ugSCvRpGzxaF7#?cT;G6l@@L*BaV(%0@#Bd*W|=(036Q#Z%!sbLa~kk?$KIlyvAFVo ->GL%iRL)yEPWQMG;X4Hf48*(q9hj4Th}lwQW&0{zK7r>{}_XvX9N8pT?6>$6#_V8XH2{+p!1+-lJ8G# -QS`3SX`A58xaxI$vQ1@ZG4#^z{ko7yo^DwUaqi1^^*VS-vQtFrntWK6Uku_F4~fIcbBs*URGNtZC~1YwB>Cf0j#V6nw4!-ziOwEvJF(;ZqKeDaq3$n2;SG^!?W}D#o~XWCeLv -UwL5#?})6HWLY^XGN-gl<8$A)Lmd2U6!Z$qO)*YE(Buh -BCOiRjqZo}ctOZKCA+B3q&Z-Q&k3Oi>jcK1H?`SuDpj~@=+Y|-&c`D?+l!}=ON9r4%0^ocSqLtEUqJ_tar9xC -J37W!jv8Ocz|_I<$4?1a+L}YBS1Q?%uk9}%gw=d0fdjVQ(}x#_r$ht+b|(I}h6X07&{*Z$)oYE%{31=Ap!6b -b8u^Jr01U3x*Z8%brhmOYsTa|b`j2l}$|MQQXFD5hS7C(LcJr$!v+x2~CwMDxvn2i1bP)h>@6aWAK2m -s3fSzAxdZ1)BZ002xc0015U003}la4%nJZggdGZeeUMb#!TLb1!UfXJ=_{XD)Dgy;^N^+r|<8&R=nbM -iaoM;MnPOQcjqOY|Dq5-z2(PMA+J{0e7NT5q9^++N<^`{m9hdH#+PSawvf^o -utLMm1CX0e+w0oS+GRxF^Uh*ofWu<53X`N<8suWjth8xl|gtRZNfbrAw>ndHZxmj))v@_m`qPDN<-P( -I|T<*r~4O0C%twwsrTCC&bj+eQt;!0%87?h)A!R?xG#NhKr503qZ^EJuU8oK~OAlb}LZJw`o>(SD -T_XBe3tARKZg;z|>MKmIeF|f?3ASHbwDcGd!@cqZ{1mM}3Y13lv|hcv;&8m-=yNa=mLbb%_tYRs~pDR -i;I}sYRjU1bDG?e?GZ*eSUc|W(!__02W*9bk=2ZwN4#F*QaN1&#y0sr{{cK@r*3qu6cX%@#5so@zv@1 -S@P56`vjieo&IB(#7Zb*iR)QPpADLa!Tr(y7~?y%a-)|Je --Qb-mw056<2!T|YdrpPLYMxzmV1bbb{PrSUO%|7V!J*g^Lg(sggPDU9`$gYAQ(Ds!I#d9X+X6Z+<=Cp -#zEZbCy&Eadkb9F1~5XNgeBjHo5RDo;3Zz&~dRF*!lk<2SEU#T{3-V+QXhtT -h<1qwm>F%0jaYJUSXos4l~L5`HQUajs-n)dp-12M*YI$*u($=T?nb$~1F$zMN%g$<$JA*=EgB#ZFGoe -m?%dAObBp3Jz~iKAwI&nHWOLx?Zcvix;!aLdEM{y_97vtHleU)P}$O>t9~JddkFazW(#eZ>>-egstiB -$?z^Mlo7&=2qj9evohnT(J=#8LFpri<~iyorg+Un$wK-TYl+6pJfSdXFF!PqJ~m%hX_=zC;TeLmz_-! --!g8~kz;g{o_==9ih3Ff!)-h}Ty=@!Uu}vq;kVd=7mTW7E0u;YdoJ_h^!h1D9=}1%3ZnMV&>KgPu@sA%SrFC6}4Z$%Byj8)NecgnEQ+fd|3hG$FI*1?*e>pl_9UbOZhd)jZf0`U#1|vHi5QQj3og|?XQ}JRxc2CdJY60Gksnw08c}`mZ>#OyI -yc@P5Gxrug`?T$>C1s$AH9ED2tu1X&E_7Z%A~U{IWM!U3M_s#&M~>2DomLRwI(!~$PJxEN{82iCEkp_ -C2mxeMB(VCLd8&mBGdY(>399z+Gln8_;t3Rtql<6ePKlW~g)P*dx<0r;o7@4rT5>J16)P<;(CpX}5=d%9i -5a`*ESKe1HQT0$nlQR`C>2tE0wF(f7FgDOL^5G9e1gZN+1JF$#l`tWzaHF9L1XpyonFQ~C)bKHcA~9! -;&uUuaX~{58_638hRWJVnWt6m^bf{112>|Kpgk0{Sy(IbY{wDP9yi;82UZLqN=1;=3#z1yexqrua_Zd -12&hGi+|-tv;B<7zegAP3t#V>>7E$k}ZK}_kccZQ+Ogl%s9}wE7>z2)fUumu;eFd|vZ1*ha?gFVSfd8 -HiB)~5FK@k5-iZZ;h%jq;IP16@Z9iC&Y3sKW)g$Tu>+o)@9p7f(1nzGK%T|95{CY_3|$whvtHu1<&O{73qhGwk!B?pQd-rWvS$FQfiuG0ETv;r;N*0DI9) -10g4R#eGg%kRK-v~{PF0DM4`sTOOnnK{t#*_i3 -FtqosDIHLwvu$YMr{85U7<9F*msAe!TY9Jlxms)Ud29%zF{TF`rr -!7fsF*v?f*`uyVU<;1mJZVI958^|KJ>_57k0Nvi)_5jX)ukA6fJ#5snZ9`z1`7 -Io`6pBI$w$lzebuTH>558SP3EB73>(@*>jhZlOrbe!Gr8K8%Fi3d^$Bgw3WI0qQz)0O67$ag@rJvA?; -H(te%IXuAvs=ze4xt;IYXx&+r%EiKo|M!(`ou=gME7&dA=##d=ZdFu+&HZ`TL^GPz(l!77o}9S$QX44 -kYDgq1%5iVUl1V?QugvtbJ=k{u_mo^~tqupdr(EaJ(Pagu)_M_XJQ|%F9CJR|jgNo-tv^Sj>4{p+1yF9qfju -(a822pe7t#$;hL$z5XvWDD;s`fAl1e7_@m<9sev*O@PStZvdZ0TL6^sjfMl$x3A()DbpQ?AaX_GDe;@ -n3#<00K0R)dYmJEmmQ!hppJABvXOFf>nc(D{gA^m^zal`?`fl9Y?gRor-Z#lyezIH(jBbD^;X>Rzj5Z~h}qWID -aZC_B;Lhjl&8@My2_wK$>Ia&2#&W4(GoS7znn9f)ATP4^ls?wRBv<3z4`*#P3!wN)cWbGr~q6(gF6;d -_~`G?tEMsp3c5+7OH4RMQ6PfB|HKW`G;fU$1ST=|9E6n-0Ine -fe=(|a;~pM`UGnF(AN0%Yuy^;dbu)tDGmBl1F2XDzJSF>29BOX(2Trmj{jRqY9cW_Ld%jC -DaDAcAa)eHhuQKdbr|%Ni0RI?IW(({Q6v8f=B*RMa0tm@CY7jnbl_uHo=~RIbA-9*$KqKf}}!ftUeHoS8r2kb2eAy- -%y(UJkA{*O>$Bj)oU@aKR9P$N -47o(zpCOSpLnU%imm_UY))<{?Joa4y-smdv~r6#03ZQ1}`yj=nSKnzuHwh=oUdGWgS?9H@$U0i5WG~S -zxkAV8}n)lm#HbXH;_(kPorqG2{#yz_v1(FFm388Y3tv1SHn7fz_6g`E?rE~Zt88eFD#oDn}pz1CATF-|B -;-^}Uje)Lo^tI|qvrL}6uiDf6F*(z}TYJ@2_0U8UyPThG9Zb^Qp5VPs!QdW2so9pXn2$cZ0^o^cLxrz -Oa$dvusU_}kJf(Z-yuJWx>_Vp`*rVtNM>5tJE#0G}Yz0WLidE=~j>=EmudW{f7c?z~_P`u-Oy-bUVjwcA -iUD4ugIf};I%C4;ZZE163E^v9}JZpE{ww2%YD-e3misMq$!*002db*_l*nAt8yu#Xx`LKC8J1W%T-y`A})%uj_ -a%})Xt86CabC_&CQjpSFuv&b`h&Zp3Tf{US3^g#g(}&ZM#*evfEW$FU$kAXzDDtcQB^C-YDlXDT{e#0 -X9_*qlK%uQgUaGQ%K4@mkA*p#axu3dqAbm;+)@RrFs0mt^oe&nSKrfcY4jU93IW%Tmf3Ogg<}FR#9FSS5np72Z;r9H9g) -U?dV0lS+(eWhxowc&*5Pp&Z{_)vpBiNTiY5CG0puQ0smeWn}9C9&5F3%oSyOBt2obzP;_^;Lb!3xcjx -l9Hti7jU*d{yl33Nj&SXa;ovp&?MiyyV!6GHsF>*s#0g1TQh2Eb+s|A>1KOdz`SnaToyOB -LsJy6Dp6CU5^%HccXpx(CJnmaSX|KBdeyS7r1W^Dy);v+Jl;gendTWXTeFJNtTInxb!{%8IdJ}MZSLn -~wT$aXWq-3UGX)5P^fTwVa0+akL(|T9l`P_=ROS{b4Xkoy?q#7+ojG8D#mXBQ#ZA44kVp}z5c4is$mB -YbcDZ5EXuZw*SEPFYU^>ou(C%_)2X&sTfM{Czg@Dv+dgezFVjD#u9_P?N3y$xddy}1$+vK#Em+-i5R= -K=DvA*~=R`SHK=Vc%|_dWT(-1!UhRR;_#e;~}C%bho}2KGG1SGI|&fyw#O2su83;X1dMmpf+#3=aK_X -1KWM>Kjx_JtP4Ns2A#5RiEstiC{YIe0Tc#BzpCyAKv^qdVl&KC(!@!;OMD%Og}q2JE@$LbOEOusq!ac -?>QKdYD$ZSZ&ue=$O6#2R3UOt_;Z;RzPW){xGbduDf~n+r9;^Ja05D+oxve -PY*hs;AR+Uts@6dL{;w|H=M-+7_%7Pk$-nat0grMpLsEUxuF$Jp_gFfmb`KyLuA;f=WCEnnGFs@q3B% -5!5ET&l~&KFX{P2vJPk{bzb%tqs8Fya;Jpui0nagoi%haSEkM5bV%Vh;xoEjDNMLRRYxEw^6wVMk!~M -GVF~6@C`VkkFrm9nxonKz|`8;v4Z~YRBrm6~8$a{u8k`+H(S8;vWRez%t(k|M^4kkNZz9rcW+AEROcx -_Z-awa?>9}vu>-5eH5E13Mug=Qp5705w2x?EupCr>kPP{sm003Ur*nkhlXBY` -)$VIG#Y`Z)XKW5-utQB!d6~h7oH#JpOHcFhlIXgf3{_Xis=D%mkg8;_*1Ce_7 -ZQ>L&d1AI|L?gWzy8BtnbAhc@v&(Ex}VA?Te$i#P?s;;QG#*T54|!)xEwv1&`yFt{prmns5MmH15vMB -XAMX&5^)b8<6UJE|i35AQj+7h-QV_;E8^uc3zMh`(LKJ70#s|(J -egvg>1u9lXyzWym12-M0{T;jJz^J#$RInZVg|*NzOLLczf$m*EEvq63=oG;PzKXb=>BX->(X_BedsUM -8248||Nx}UfyvWGvAjId%F$%hop2IQyotRN^xjTaY92H9-4v!|T#Gnu?>UyQ7`}=8`00)wW4$JCl|4D -dww7*`&wJPMA?!raA%pY-J*Gi4PZ|QC@(g;p)Aas7F!l6u;-^O9!EVn~p0qarSe;k3xz%qlo(OW^{!( -Jy%^apye9)Yoo@?$&|Z(pAre0A^?H1xM|C54SNM5DkK0LN;G@Fr7a^=uP8_Q(=)(1wxcv>Wefyiv81J -7wfhTU+B{{?c00Dy${kAw-A=P@eG8qj`CtRkA!6^BAMA4$V4RU4iFUrji -vqra#z@h5Gr3J#z(<=zDX_xSQbo%4F)AJJ-exm(C;@CLdrz -NoOI};%;f)I8AFli2f9pIPl0IikUHZd!Qw&B3+JVH_GF6Laz1Y^6BdTA_!bovi)H?rmU5$-=tKljl?* -xMPn1kV)3xJ4p;Bb~jd$#AWt$78_|2HZyJ58^1KxV1_23dmH|3SB^K4ek+)1J2Q -JEHEWk9p`~Y@0f8PWQV(4*3?^crVl7NL!D}XMQAq}(hp`!uE*P+$T1_W^Mr(2 -jyP%Txq@+ToKOyze%Cfl5}^%kC5RWQTuY7ixUDx0c>;NN{w-A`z75SY%BkgH(U%2Hq|ydk)|doScfsc -#yB7+`XWlUSN(yOzo^t=j@M1I!WTEo|OjWRzOX*zOWr-hObs^ys}eu0m02T~_Jmp%Dwy{=b2X-ogs?Z -yzfdw=OXZ^uZ#w=)iauoi=TzXrHG|WU)tzD_%9o|0P}kR~$|-i$xng!ap#+?>W9By9*ivjLPn2H-JA< -;ZIF}pauSx1^h*~clU7jcNz|NBtX;|o9q@P{+-hwY(<^tbJP1^Hxz{5tv;7xl8J+fIyG -J69CcmVxb90)eY@?w1NaegfA%VK*@#;Ch3J&(;t`w{+0#)J?tnuUXsH$X=TAs;{x+b-kkp6bO_9x?hi -)b9E=2ytWVp?`g1QUV3aJ?i^Pu49sGdkdWH?0%?f|EsZmx -H?0v>WQ2`gXY7t|SeMHjbtSkk%_>}Ne$ -Oj-I`6k(FEjw#3SbHH+mpBA+2L2hG;34!Vl72pmSjX&VTxQ{#mR<;;k|hXg3zmE1m@ -gu}apoeQgHijq$@OcHPQ8K?0GnvE~p*BhfO(vaf;2^4$;D~SHjHpMDH0~TfG5#<&@QsoMY4pc#VV5th%4wGtPSA|66de>D8E(CkR+ -l34$8&J*>_>r(ZzYRlOv=7DeNBNG<-z02^8*dnp3&jX|x0+*D2)h8^=Rs{P2{0 -03g-ZBAjtj`iAiQV=d%4U!hzLc4~#prNPhoNaZqYL;}bi5*PhLiY{C3{wX&zM9WPSjU(bC=Fq7dxz2x -nASH%Q(#pmYrbnj_L@mpl9oKf|KU*@Uir9!T^#K&g -xrq6^Y&Y5DxplI7SsUS2Isl3TRtgNs+5~paum)57AmEgQVNX-K00%m=;0O0{+06NuKW=)fB8GnT05Z^W9GP~!vHU9M(H#`p)%`>+)f#N9Q_3GQ)AakURk -tQ9hb9YI3w#6IID$}Tp4|rcb}e_oHtC8Y^c`Z+78Hfna&t2sT)RyIOq9?V8~}kwpuqT|f%6F6&|@I4; -)XLBWbJLn7v|S2bsm$5S!#?I5PyR=SF~!7)0LE!=&Ht1bGfZZ%Kx-s>gO?ginyiV#Hba@u1DY`P{gXP -Xrn0*kGE|)am7(cg&@#ew?t@EsffjgGG6{z#j6#l22+ZtF~_udq^QMIm-R7;Ur^_z1b2mLMNtrzOmS; -J$-yp;bSLaOV|Hi+5&%7YH#t*>4j7b>_O4G|_}S440juP|OajM*`jY(43CNvH6|i&)Bm_N}l$2csBSk -42DjT798V+s~>Rm$QY~!)L!{{Q-lG=wD!fzpfFMw4l7;iXFCzFYVYHTA~a -WoTm6m3k!wcL=~3crC4pfg5Qg+Q+*toMC8`h}+kCPI7Umqwk!LouQGH~7t&**xdDZA%aINK8Y%8B&qSaX;PPH%$yJsmgqWgk7Ic^IYKY-{q)9B^%Y@wui_GO -Nd_C1Iwu^6%ol4n!2U&A>7b~4;LNxmYau&2g-L~9w1Jxlj{`BZXtkES=ycMTZBmaW4Oe6{~srhAO#vN -Oy{8W?oHnz7MMLfAgBQh3k`lo5LXMY4echzw7kbTGd_qZ=;cLi4Fhid?r`IXJTWsG*0Sm(mGJ>@8l|v6 -BbN`+q|@~PZhf9^ -TVRZLw|nS4LK)Qognm%N*|nlY-Ddl_tzA~TDb;VTIfE(aBqMZtcr%uCALFM!H}&C6+aX%xhRyHqOtbq -JeG?be-~cMr$Y^AIV|lcv(7rV6)K>Fz`kt*IxuIBfe!0`nhd5TEwTkBkuC@@&iJIex`v3>cV0|1o)7z -H(aB&J%Vzhn*6M(V{3hi`4V3@E5qdLT@?6OP3I6R2zLgui`6Q3LZ#}&_2)V@ -W`u_|h6enw3bs;l*#Yg3^_m2ep*N-{V@@&kq;(*A&M_a?O>M6)FUK4=58+2*^-8w*bA8{p*VlJ3Ko#CoT*Ls*!gknWJ1Tf*&FAo(m~V(XFrjvjGAUOZ?U*q=cFp -XjW^ylfHSp9)Zkks6*h|3qQ)Tw>d@d*^a -45U`VpT^h}Mm8N;7N;p`{C2Cp-U#km`uTLwGr-Nm*cO~kVo-iWjWlowu!s)cb@h^d7 -HSjc>(0SAk@Z_j3%;=Dj(vFlH(4Ch=8R78{V7XRSzyA7@lY~>A!&v5_y}Hdl77dgs-g(%RHCSxs_M^V -?tRn37k4-bYX3!b|J-wd!JALPqBNJm#spjEdcziEVBX}sk1DAeK6IeVj~aL&O5Db6^IL7sTCWRbjTjH -@UDFflHh$toZyuVbkH|I9wLCM4f2m-fHpK~hl)|)>3b%iuZ^_5ZKuRH$O%{Oz2{+j}>WJ4tYttEcI7% -R1K5=}^Gu8K+?-yArBMM!7oqIEK*yjQbvVF`Q0?Q7^?1oNU$OGvfLgmhsup3tmlP$wLtPX>l3p4zx0* -5x?vz!ENS*F=M#-LS$`ws3IIy$w;C#KL1?s{Y}JKCqRr>f~K9&&2AhR%)N^$+EQbTst-bK8X@FYr{U0dQ%P -9nPtqE6{Vz45O$0)n*m%1N{9=sYa^(!=taiCJCSzJKP -)Y;LhX{H&c;8Pyx(KECNOb(s)e&vv8_AJD>=hDcEsP-&W>nglO9lc7G@#uc&Hey^yG52*ISODl~Yq{k -{Pz&=#7|(Au3m?DpZq$zRV;-o1JG`t?tH-@knG@{cEf`0zpSmpbf0o_X06xCc5u4zHwgH~fafva1;v4 -}Ous9RH!kN(Xtt&)>+S+htr`%fZQU`yG;y4s7B!o~P;*ViXS#j}OV=*(^f0_J%57c#QG!P4V~k`G%w` -865+v-Qz`imbr|a>R5lY9*e?W-P*0C{e90zL&1QlW`&!$qII!vU|w-uZ+vT4Vzr>7^={G047|B(w+!b -{eDjSsI&h*EN+Wd6sB1UdM~Hzwug#@uw8_i4Km7@sP%Yx4XJ7hsh=KrMy+{8gz0e20k9D~y`Ff!DZ5d -Jp%Zd^9yj>z3lcqoga!$F?oQsQ&?fiNUhaT%U$8;K-T6JZ55jLw7%MzV;W`oJh*=|T=RL$mRy0X$Kw# -5BDqNVxB&4oi$D!?4wlt7Rudz=q>Fsr8!x~~pk!iT{1zB+=dCv^1`uAWVTf$=kVVxc}md~sojjJv-Uj1szUF!O-G$y2E^i*ZBK_d=bglq^EhzZ^eFTL1;aQH`VQpmStKwYRArdL#yFYFvCSh& ->0tGQe!$ijOT$`+vz-mdY?-EJBlhrM0*C^p2jWOP6<>F#Q@4bc|hkd|$LUv-V&wl3PUXg!Cm12}T8niZ8~4Jpx`W6+&e(3we~(G7+;I9`BD;;=(mEW -2>;+sKUv1xG;U+t!F)Mz&J7M{o^S$r^(;H8r2p;=B_on{DJ=$~wl#wg-BYo`aY&K_7ZJ(XfZkAjHzEF -*@y$$p-!T9VZYQ3^6|*K62f&F0~c-&-FHH)60J_wf{-v<9YND9;4WR0-#2TdH?^h+;Haqe6B-3Rp+!+ -1kOveUFd%`ZChHIuje^g?SBB(9>o1YPTSIRO!)p)X3e<&J+s~-aQ`Y3x9C4OcW_|9my(f{iidgqQ(EC -4p>QHP`y0*mCnh@8_5UP2^Z8U5pZkbX2|8UybW`VjZ_dZP`(R8(J%IG5Y{mbH)zc9ECl+zsg8VP9BbP -46RS{+L2z=EjR#8=!^+@XJj7pm!n!@H7vtu}|nzW>SFgy<>ekihAlV);i#NC`xz+EUCOW#z{#2kG&QENF!kI= -Fm)W#`H?%moOLn;ANAg!;`Oc`=~Q-O!G;-MJPPAQCKfvzHn*2wZ|W6g*gM&={O%1O_)8#E|qSwn7^{h -WfDnQA&&u#cn1`-7<6EcM{*WUQ}mIB`XknHl_h?E4Qt`^xTabVJlH1O@1ce!S4_0s3Eyv48)3R ->n3M+vIX8LjCA6XQRDdacxhgC7WwEbFjP^0PiaBy#Zv1SOA$SL)<745iv*O8-M+Xx7uaeCby!8f4VPN -rRNMN}@BEC2lgM+x-P#5!SPHQFHd#6uTXpTja?;^|Xa(NJa#ZIawr<%uEg@sZAG|4ifEIsET+L9v=ML -*Aynzy(X@lA%`EwEo{(5t5TLGuaKR?~MIHZG`2A_F6!4Z_e}o1#%j-Oz1%T647wQXG5H&UQnYl@SRSc -%~E<&HXTfXOPuwsNz57>;!X=Sdt1Y)kRIQLIAyq8`ZU{LhnPd%#o!ZU<3=+Yb0x;fCArct_KtPUW(>bcmVu8ifqEF -WYzoAF$arK%NLcMLmv6Yx8Kaf`nOWwAr;-3)QLk1-ZfqlMF%18>lOkFit;EjubDV~l-SIP=yHqLk$p( -@s7)^&3x5K4q-J8Yosih$*EL{j=X1oid>3Aee*m7XH8wqj9N2C|#Qz{Lr^}qqIK;&iO6NVDlO7D0Lo@ -k|c@Wvyco_4StKzD5@MTBK8aio%F=WfrMl#*=@!v$zDdzB3j-{E4u?GU*TjE*?_4*1qChkghIlFj>su -t1PD(j?~UDbO2>C;%|gYSQ5;B-HyjH0?D#s@O@4A)WHT3w)@ye>(d|~yIPVA -#{B=iUj#W&eHES7YV4jr89oorQgH7k?Dc@4!u;|kbHQdE;1@ix@PlFs}^Yl!f{Z=hW4b-u7Vf@>l?8k -N4*>{LZ0w1J>Uz*H+`C*nbmx9lXAc9N0z6T+8TuF$6k|72tJ6TqzJ>8vcxkad_QJmL37Y4UoXl4oC%rgfp4_|&6zWXWT4Llq+ADHFq`$0|5}D|@;#!Nnw!w#*|`Dk|ik!&p_SuF(c?X`5l -(Q>mnth|+mV-jWe^7~QRh?v9ij@yfY6tTIksA_5(f#udn_3Ns(|3^eTPic5>pUvxad(a@@ngbx!=XH- -Bx)o;e>bF>Pl`r|Y&;s$gYg|JLVsMiWRzIMswo4w4~Urc!JX>N37a&BR4FLiWjY;!MiWpr|RE^v8mRAFn|FcAIjUvXq#?BehJ!ke)#_M@ZD`gCzOe;$j|PW -dBQ;4VuHbILOLx^S3&A}H*Sd7K=Wvp=)PW&qPHCt)qHra2QlsM;oE~fJc%Hf?iKjV{WN>@GHvaOme(i -aP*Vpq*b0GG=+wC5y(FUD1;56Xm^uX*kTzrZx07ia!FyJ2~#OfU!hEf^h!S$uNn&BQ6bUN-=OIKw;BUTfm>5)O9`}*~yG(Ew!-|iPizVH#DA`Y$I0uo -~J2CFT+fBwIDMoJ_3Ppc3&g_4%BHwks%g82cf_+ld84FN3?6Ts%_5^_t=Z34`_lc(x&R7lNI*sdUpyBVgp3BrdK1@@rX9PVbyF2`@;X{aC7tTxJ2StIIk*K2*19 -Bsr_BlWt%vaLOFlfHfT8YNni~EF8qVt7J1Q(*A3oJb7$v*@w^vK&MX6``FqZzVVPC;hZvsRwxV#Q -ChVuNHI8dnD`?luoanr-S>97#}_iMCVu#nfB}{faIQ*C29s<|@0|5Zwt+i%CC+@6aWAK2ms3fSzD1m{w8n>000*w001EX003}la4%nJZ -ggdGZeeUMb#!TLb1!psVsLVAV`X!5E^v9RT3c`9HWq&Oui$Dhh~0Ixi+v301?H03256f>CeuE+fj~=i -%tj_%QnDxL_P_5rhonSGN;2JD0~uS=;kjQvQnNLB&x&H(Hmwmw!Q}p+O~bfSy5Wt~%Fbr~Uajx%rMeH -7+J;|tSBo#*MY}l|T?%VwTjX8px)vpI%K4@Y0aUB2h$p*;i^!j}9XVp^$yi{f=&)-tysP6=ZJ`6qXMg*01Uf>3=Bs_)=@IrS07NovbRMZpD{oHhZV{d -#OjK1v~#|e0?!v@X4}_zTX+a -8^L%?!Sv==NQ*on^fC)4GSz#1Mds!}9(q>Y$pRX?% -^rB8}gt-**uN{aImdBauSk;*MBL>V0T%|VwXiCy#>4K0FHX&_ILcF_@u+KP0^D|X%NggHqI*+-^Y@;= -zryFY(g3Y1ePg{X{x+ExjX4>b5{&zmBCGIzeGECH!^6LE&yX?#Brn;-ynZudepmSRy!d#L$i8t7NJCe -EgA%`z-t-~>>;;$;1+i`?^iQ(L`cC&4PM&Vbk905K#A*62u`X2(&IK%z)CWrSzd? -7-V4SFAA-s|}JCYNgbbcrWl@s$xvOkB`kxEA|h}1JpD~@qBjNQsS`It!C&LkW;snu)tJlb%zaGB>qSqDl~T3A$CC<|TM$?F?5ngq@_2I`6GqOQRND{Q`ql%992M)5_MUc%dI6HV}<4n%6L2?xy6rWkMaL(dr@$a5qk1WKXVXl@(3Dr -jy~_GNShr_hK}Wr-)%P7F#ZBQ5EEHa$`q#Yudn+++1Lf9}-0_|7MvCBk)N{Az03)+uxgL4kgjV)_}5E -3z>8EIYd(d`Mtk<`}tpR`@Pumw#M$Joe^O)Oa~WQoAkc -N$%qgF+gvvzuswx7+3CVd#5fLwVb)ib#Lj~v|DGorP)yiq2SAMN1&8Lez6%NsQedEozmJ8?j@(GPulQ -aNUs~Jgb1)KW5N>zlpmS#a5dmA64dc{+m~1K3%-s!Pc3l*DD#2V)Egce|YN-OVB!9Z-dfySfqgc9T;U -8>tv&-4fJE7cy&WwQhzfq!`q{kNvaO10h=V^v{d<0X6lSBNfS)*0b~+pJTPpwtSBoVYuTW$3A|MdoV42 -nA7Vxhj;A3GgQ8OzDX7tAQ=_6fCjJ`KmZ?V#Y6ct}2bcK_I)CR@^p%cNh&;o}Z>ZF0EE^ltGZ@i~)Fg -KX+gzjo|`6Js1Fvnv>@JZv$fJj?i_j7RNyTzlGqxc?`fbG -t=t4OJ!$~6uq?4P8IFMDvoNu2qS=Os?Z!gImC2z9aJN6~i=r!epiCnTV0#4t92`f3nvi3$^IHXA#Cp5 -3%cp+aFx`Pu@EjRBe4alq&Rff{z!ZieKgK0MAB=@biqqaR!9n74-t#^pSrSab&r2sL0TpIQFk(8Ll(CRdSm_QcfIMxe`-qwFlVt$sGv(967Y6^Ofj{Gk1Ic~7AS;1d=6Fj?uzrjr$9Cnp$c==#!Zg;*ZW<8lt2Re#lm*GSXM=ad -yV*-i}@L+1-sK!bIKo>w}8a4c6?j+zff-`I};)^KTW|*_1WuVUZW}MjR(|%)3K$NaV9&!QTTTY->MAE*oS*#?4Onu%uf8j(-y$?WMxUKFgRyN0RkI -y0%B9m#=$fFxr^f2O1r=LpX2877NfM4&{aDqll{>rjnROV08dKY{`5JfWviR6ASDhUmxeoX6g(ASW7G -2rx>I#Jx4^P!zle;u3ZV;D}ThmQkK{EjS6S`~iC<0megh-eFYin<;Uf<94n#`?se8>UW|wwXVQ>Ycur -LSZ8=1hn;`nA-uE>*wN;VZw28$=)9P9Zs!(jzr<>G_a+U2EV^vT3G -m;G@zLXJa(U1NRwj6tAe_FaB07hq$m&XVZo!W6Z_6?%Q;mmPt=G{jR#dcT%haWn$^<-hs40Ptw2Rq1OJanU0f*_>?#^eVIGZPAOVB^< -Su+hz5Tg4O|BlzYVyuJBdF;2nv^K3-!ptKm|f7!`D5`FtUS;kI}<5ugXgB+UOGcC -$_@h1*B>Kev2yHZD9%ZeTG=0z%J^?P#p7xPD!yj1dAME`=P#qL+&Ax7xI#U4Bki8S}$1n}J3;_(?ncd -Y)==%cJ?pV^1aad(H)*jQ|Ih~>HGE@SXr0gP9A?L&i~j#cpDZ7b{QE$k!U`FC1Bpc44z?RMANrF-M(q -=eH|uGpK8S57DtYq=8$_<3`NJBx#$mrF3_pWr*vzCUQCAHCE1EKJ`R^%?Qpiy#X1fALa=0+F7{lB?qX -^AIAppi|^26gv6hD?kc$&6MMB+{(D$vaYDOr%>DyVO9KQH000080LuVbTT;B -q;jIe*09zyg03QGV0B~t=FJEbHbY*gGVQepTbZKmJFLY&Xa9?C;axQRr#aeBX+qMz@u3v%3v<~Rq^*#q`hsS3+%DnAtPXsy(P|T(!pEhlfOTE{v_Lc@#T8HZDdufgu0n~%WPgg-F%*X^>nj&o;|*N{B -*OPi+V2(tJ}R0)hg%tUIfxto4;kmX+1aMyA^SolSyY;dv@^LMekJh1N#DG2YK)h(0+X4bK%YSHEQ%wHv{zy$ZFlf|KL4C{SPH5xo{nrQ_WVX_0~9kNMWbhs0McdI!iuUSQ=?IP0r7E -~*}ZN+q(W4}i3{MFyK6WSV_R2mK;ZA+^(?J$fs&Xe3}f+v5&aFX8eo9iCJqD1o+)6?onR1pZV$%bvBH -!rd%pgue=D55(Dn{sdbIaP#2Q1S7&!t9nt_*i8T03I4ohM+(8{KhO)4-;ycP?9^ -{WzgIX3!RXCeBm19??5z|T-0i|9|>4LRsfpR9aJ<3kghRgX#`bp_}%I`E$&vKv0UeM~L(KoVysO8M&x -(3UIln&>x_8edTa<0JRtg%4oG({vtjZgOwIfF{t#|VqUqP6B}^=QwUoz{jm^y7r+^%?JPkdB)7BxJ=&}c(#-K{T{)Agov*Hcl|U}H~09?`c7oCo=sSAxUWrUqAz``R(}sdcr -NHf$gIl4AgbXk%?b#Ee1S?8OTRgW3oyl*cq!4!Aes*^@4vGjiKFVSB?7q8S$9=Q{TcA(*hf*9^z}AWFy7jdf%mYqB!EeXPf*4iFEBZCFzy|t;kYA3#X5;xN98xJQpt&1e3NNue&`|_gFM80zojebu -sTPH=lxoXu|gv*0^>FNlvj3etlTLLz&p4i1B|VW?TE=%M7ta`*QQ2*lXfZx{}Q{hwK-U;l{f|tK!V0V -0>Jnuy|n#2V9!c+3;k4>wrg?>y}FXQbiGpm=nAW=o6gm`#(bFIr}S+)dK@+dt3=BJ9!$yg0yE`Sf(=P -N4$?FVKzJL?R7|WuD8Svs{!#6-{Ally%=fZ>{}n{1Mu6?)0qw(VEpk~NJEKZP%~}-F;Vr^q)`AhG8!( -YRy$Cb3*krY`CJ**YhYoESk3B^z>(I)5?VH=ErjBrx3XpvCVnvx)1m<(y4oO -~XI?kH&@ZK3GHopTI)2Z4MqLQVx<+Okk)JaqNPe_R{D4#pk3a)OrX!7ao|FG>&tkBm&1uyPJf1I~0FU -GO{?wyRm80)5{**c#}CbE6cy!$9Bn%*Ry`31|a=z#2M -#2?qAeRm!FPoOO17j%7YfjWJ6Tlz8q*ku`o*V#hUh2Ma>BM=p^8N<_Q1|z!k1fQoqZF=UX#iwJ1C3&L -W#?-{JZ?C=KLXu8h2_m;bZ+lFn(`d*SIrxTk#fAzI;s<{W@%{t$l>j&Lim0d-ZAH$Zc{8h_gxRBo6D) -&kp1?u`pt(hMj*K?oWRHhRs<$D&!D~ard`tsAa{{LbK!Fz;JcsW4LpHgQwX<*y)>uIS$!P`mw@UttB= -Gudk2oc|5ym1ws5NkPl2>DCEGgBO%uwvY1S)M+U$ir*^zgR8?+bv0rMz30^IF!c@6j9YR`X>a1%7q|f -wygTi{Ot*YcZ0>SFEer3!*Kc}C)E0C@`bEUDY!sA>cF1;}rfGt8;!P4!`D%$1yZAZ -OnhdJU1&a>vX(o&s4EtQ~AenBtQ5aADKK+d!(|zOh^7jyE4^lOeMM-RJ4uE4t0ktnj!1uM9r{8L+V(s -wL4L4K233qroGeiyxysF(rv^%gggS#cVM6#AE-TbiiCb{iNoRy0Krdj)SU@B+3e -zaTpzz!M*8@b|aty(utQf75u|I{weT@j>42u}1Aba%CK4JxcAE0??Q?O++$*LX&z@$mWB(eRb!DO7@4 -=pRb+%)+EsUW4%OwdWcYk1|mX%ZiNatvih%Ig{`VCUheTbxHov3#RmhMxyuJa9HHdLciC{4~cuGx|gK -C=GR76<}c+#R|i#Ak~PO5B)*vCoMwr8`cB}m#5hWfy`dK-XO`8fB-%mO$Y+B_bW`h>gPxS$$OWtL+x0 -e2JV!x2F?um{Zj*zz@T!3+NV1$Mk>bVG{b-Dh6WLqmjdntfvcpOaDz}bJrmdih3|>oaompV)sI{*i*K -RsC>M=h7SXTeJ(C#US?oE$5tvisCpdwAx@T6RjojhjDUzNpPhnxe9HGJx?F^P6+v3h1@w+S?DglPZ;1 -A$xiIX1!w(!uu7~)d-8&}y>u|xzG!t`cx7}OH}-^MoGn8V*a^F4?50?Gxq9r$rL(!?s&DRVM?A|%MGT -e$Etz&=={;Lsnb?Kzah;5$FuZgjKT<1+wnbh{4wIh=!C+aCi^uf(@hX;`h_?lK$)GzKm{3_6gda4KY3 -0zgwPbXjb@X#oIYB~_{6XH|q(E)RTFgR)qlKOe0gk!(^y8I=V)2*(TV{N(QW=JLt2FFwm2KfQeQ8vF6 -;$8`Ph*B^ZNo8SKK_kZ}~pU%ClNV)(FFBhf;KmOW(?!`tI!RwA!+k@xtfBp5m&)?Cg#viu!f7;t8CE< -albG^(vKN1&oi?fTo0=SG{`wK%qG$%bq&AdZ0@nji1i-Z8;WyL(q97H#jBJ6CvxQ3sdwB$7DGCn2?r1QyFTOqo- -3H}?b}y&jgCH}B@*A#pV(~!&>Zt{uv^+sQNC`#0kOgK3?C4&8r@20C!T)r0hI(hgU(M%K;Ck97rkL>ZkHT^GV_FvZIKWkKAw06EU3@*L4uB0Dvb=7JaIU+KuM-w -L%hcaPu-q;>Ortru22Kmn@7Pi9;T$)9EU&c?2FR>|#yD60UI9$@&_Y9=Hzso}H0>c%@DUXP|l>QS7dW -62Do6SF-y}p|E3^lNyaE&-OZy2kGpDnESZ~n&|AFxlUx0C>C@D9ne3a?W9Os0JwpWf%0pFi{XFAP2Y4 -Nyx11QY-O00;of09jjAqaGw~4gdhIG5`P|0001RX>c!JX>N37a&BR4FLiWjY;!MlZg62^YiVw0E^vA6 -TJLh)wh{k+Pk~b7hOEz&Wv6kfR>`=IoY6F~J+?fPv7(`HJlq{(9?1}-q_g8rzjpc@eX%@AcNYK&ki3( -ew9|elXKWowV6oW0zg;j0g4eaoWG#hamB@H4=8RYAN?r;T2SG5IEUIG7l4Q};O(l|q$#q#&HRE|+)V! -8Ot|k-nEGw2vnJ?{aq3kcUYG9T24(7GLH|?OBR|}a5d;3n7bW7`}MV5(_R*U&8wM?AUqT(|<$jq=W-^ -lVLUle8sr7Yt_<}e}WSuCzoQ4+b9R1R!q8SUt}s)}lC8ty{HxJbPzMLRc~Fu4Bw=0S4&-RbeOS0~S(y -_(Xq)5k9d&%S#zc=YDvh4(DlzW7?IcKae%sFS&@CKHsssMwKJLcA39H}ot_lANzWb~4?o$)v6}hs3H`diL%CV|y$we&iE+K1%rRQBNupEqHc9R7lR6tY$x+i&?V_1Ejt#<_#) -S!6IO8mt1D3>L7BM*<;W1Y{M2s#a?XcRgtrY@x!|=pV_-Lza68VT5icFQQyJnWG)sgZK_J-b<(W4dK* -UU-WLpXs>y+eeJ#NBSHccsG>l!XWV(WJmqiAi$X5Ikbjs?2DHtwchCt4EC9@5=DH5ZJN8om`Ug~~CkK -23vPd};zICD-D43nyaAdqB&&tV<3sIcVq>JWUjB91eu>NB`G*L>qkSPK8d89bI{80-g8`}_Ov1HF=5F -n9qQ(NlOnH{OC5jq~0S!G+A{M2`K*eSkD02CAl6C~Gs5=tYMuS+hHuO8Y`$Q+A*^5*hO3EWC3fjB{ox -F&KP?qo<(Ab0f(WR|7e!!!}fr5J{x7qR3h+cuCY0tE)z!&>$3%RnTlyzo1(Mj?4^oEj4s)A`;42#8W0 -NMFp;a@4+Hri$xC8ni;SH0ko(#_>AW^$Vb+A(qg?X@}#W9LSBPvpTRtOfzCxM8F8sj1bXlkM4@V!GHy -drI~=x!TGO*SDvmIpau9L*f;8px1zSSK{K+pV%uvwWtCtf{reS+OuwaL0N<8r3f4k%PG%J30c84a0N%+c6lm!wvP}Sp6c>v9cAZ9 -i@e*ABSh&x>JYlxff&c=0&Q!r9lleIn+Vt7NWCB2(VEiA!#JidoKYX9a^N+Ggc%Bs>d -OKmox}TcwJZQi!a#0XC?uI`O-j)-j!H38LzZ(@Bz;^tvp~~fL#<}E|86x-j~pqu8xtHu%(W$F@BYtyn -1>3H9d{!F}?q-R=VA7ljDAojS&CC7-@VnD^h}Z=5_&quNH8JOTlK*)3>PoHPR#}r0s?1m0Il=c8!Y&G -c@@rL`0d>O6AbLE5P0`9w0F&iW$8s+N1^r6yiNW0s$gt;wlL9Qp_XB -0vf*5D=6#T7jJ?9(H%m3Ed*%0~kL05fA;5Fr#eBJhyNXPo5-H10}k; -Ci3Q*ctlkx#CjF-+Y^BRV@cA5you;D -sYhA#w48=M^D_u4DTf{;TwNTpVB{l$mq5#Q3X`I(Xi#GlnlfRQ855{;N?_Lxbb)Wzh1XmCu7?=$P^hG -&71uxy<0DM$#kh@~`Lx|x*Pa7@|6%B}E3k!=fPP0Nm_8xsLcV_bu8)$j_cr0YCV!$tj!E}L@E}Dn%MwgivNKOi!sS4DTV|osxayzR1h!U1ys538dKw-& -cF=%45~wT-WBfKDfo@1QA)FS4{k-uYXyx`y1pGxuq$4T?#fPggRJQ-UGXXhE%{k3ctQ3re56p9Fz;u8ksm`a+w# -#s=OGx9Uq-(ry-0)o6z?Ng@6g{V;ZjtqjOW2;jvdK6PrF`*4m^*c3X?N%Y^O76uSADov_F6?hJd5N4pzaisntto5W~d3D?n8S4yxUncOIwoI4r=$modaicmOP$ -u7;YwgJH;atYw~*9Ve3z)yuvQWl=%&w-s>SNMJ1Ps36q8uyt_T5-1?&19FTV0-LURd0Hf0_3Nah_2IT -Wx7b;(M@Mt7LxEE*y08xAiRtkd?M385P56{tt-|tjP_;AYpFl7giqM_gQ#~s{MCmR<4;iW`}i&A7XD^ukqLm+EkTe+jXL%XOaPb-^r0DJ7IwkR!ZLsY?!oazuHWXKX4+(P7 -8!E!-p0f7si<)PNbM{jEz?1&~_KVheQhp2Gu0-Ch0pR8`y -vm%2WK_F2oH#8W4X{%rdhY3?y6-kTWv7qRJ~2k(H?9N&$W8%E`B6G(5tJASd{=1{qMuqhHqnbXOzZv=c~Sbw=6fEY%`swk^Jh{4 -ZKj}O76(S@6wzAIuW&!NnNEv_L7V8bO9h$i8|kB!7~?}=IA=fD4xJ&X??#6SP-ABLV^iH(?wwxI(a -__{zd?a#<^MYHQ0?@|Jlo9c5g?=7+-uroD -DUR#!RO3g+R9;fC;BN#>3FAyOYjj_Q)mog839NY{AJm>NTyV5hx?Ui5#xy4Qb9uQay{o_+jA{^yecv@ -EI`HEllkY(JB;S9gABJ`B9+1o0zI4!0Z(?y=V;U=3YE)DnVM&DSLzF#%p66p(6!fM*7<##|dsFFb+aH -aL4n{MoaLrDLicU-xYJuE$EcP(Py-}tsXADv5^V_H$Qv)I7xlvlPUMCz6H_;UDS&6i -MGVx_Imj=CWx%BGd9F)ELr`kCVg8)3tFag}EUpW6!Hj+$CI-_urmM@4_^VW}YPju6w@7PU-coEuLJU0 -?B^x(XgUVcdqi7Q-`5pC>%+es9c!JX ->N37a&BR4FLiWjY;!Mla%^)haCwbXO>f&U488kTaOx#SQ>W{8=nSO5kZwSMq64~i#jq8D>d3P;MRX;( -K``vUkCf$nbUn<6grdYp^6^P}*`f=8KGVJBgVXlMuwkTD1Xl2vP?O1QngEC6mA)$gpdG4P;ZDHj -UlgGbW^GVB5x`J^J&V~0U4`s36=@Wo!L=$}&S1zUwh#~n1I!9k#L%|Ce_L#P3Y*d?W)#ZT0BBx045i5 -Nd4D1uqJR3}6G3CyX`Im@RLO=K}kF)7*q#nOhGo~l^*-cjkj>W5bNz$C0R!N;w5ECTdL^D)%oD|S#x3 -uw(6Vu5`M0`Q7u)FvdHLO%r6A^o -RHB0@pPPK|;#vb2Dr?_qzG}@O^e5X&qVe+JWUTyO@sl9?P(sb&BbLh6C;jb1J?NRJM@=lY!@Q{*%?^? -cXq;a$m8ttJqGSnpGau|!)z$KoO>Yr(pmM4QwN2&V5cs7L{z$XiL+AeVAqr&6-Adb^h?sO}OcaKX&wa -{3M{-r=nX659TH0UQaq9_vD4Od$^B?ExVto&#u|T4JL^}1zvk7jU;79=r6 -;u`C3&3w!8;deZMum7IeYzMl>@CN`J2k91dPjzYd_^l7b+LDZEMq -G?rKV8}<60E?9bi3+XGZ=SGD<)G^Gt(@)$GJ6;GrDc{EMktpd266HkX95q8j+BgJOEY=Bp%=fEGJDaz -2`*!48c_z8^^_BP8?#so_>@QGD0|XQR000O8%K%whchf_09R&aYe+>WtApigXaA|NaUukZ1WpZv|Y%g -_mX>4;Zc4=~Sbzy8}Zgwtkd5u?XZ`(Ey{_bCKQ$Iu=Y_%IUv?%5dLFN|4x-Mu^Y{NFFv~;q$(xgCAc8 -zuaeRrhwVpmBUAc!R&@9lZ+Ii|8^3wd5OzOk6+Ah)%#9)#A$3ong!lgaU&GV8U}>-N#O_LIf-las#xe -4wLz1}!Yw!`T}BtMD$(b1k+MayrRhzP@_(_NQfj^T*BY@|XPPRa}GYQMe$y&(f#GVUeIs)r8OAxx4aR~szeJCGIEl&yrrytY`$57eeIygNQ_g)cVQB%Hh5NCz7-Q0U-#c@5t^`X0 -CdC585|lh!vv0Y#M7=q+mRvNm9>4+#=IbpO=?sC~kFo6_4EB!g6v*}(FS@)WK`ambIMzob@+LbbJ0IS -&0;bz`+aL1#Udu$O)V^`RCQogvwsmiizVJnX-Qf+`z!VI$n2?m=2fTk9E;_tkhyRNxBtcwwAd|LKEi9Gu`A>aiqgTqJ@EfM -j2<%b5Xc$RBRm-qqkVuzxZn~@1tCg=^v%(fBh0ux($kSfl|WKJjzsxT6a_j5)C8z#4o^WG@Bmny?2|qsU?IBhb>2Rt1#x(0qH#nsso^c+o#zIzySx92lNag2J1ABw5DI1C_)LxME^v? -E{guz83zP)`uf098O_*TpM!IQ_|Jf3#O!N%06)3a)3#D>C@1mq%Vd^LZPOu3C!pY(tX70*qXPBTZt?p -q)!6R|Ur0P|Kz#gx;yj{xvt{7_)+;n$m`wZ`_2k4_numMYrhYh_k61fSO}cy|Xs-a$qDjM@n&+6h5e9 -f^NrMhimhy+me!@UY*Y9zQQ=tO8g~T^mOt;0>(eJny8FD?+K_y>rK`y&EREZJZC(6@+tOo>9u|c@eQc -?%&f%*R1e^f3&_VOUe69C`D|d^+eh1g^f~BL8LlAv}~utQk`jZH>l+MjsXsKqQ3< -t;+9d70um=QCV|x?>xP)-nk{lBWF>hJj(s1vI71*#YDyIp7~s0-wh(?`q;pGl9xa9KqMI`k(8Qmae;Le$|zzeb(#LIe<>G>eB`bP>d{} -~c)dIwJrOW5WN?B}m1S+GsnK_8Dmk1mcCFz7*)<27MzO;e{yxSf!iEEbRidM2Zq60#mVKaWfv*Ml!!^ -qga&%{-ny9hc01@Cm6BZ3;)z8_>>(%p&r4|si-(^0j+wU4V&(X68p_Cnu+#lBBG&gXIBU}x-DMp>!??5V?Y$k1X44^-t?$_D^uA2 -m+d|*q9{drI+np}LYZQ9c^}oON#O)bYUneZ5m}LD$?ffQ7>Y@>Osj%icP!A!9bQJp>GLi}Z+f2$@c -&Rt0|XQR000O8%K%whX2+rgJO%&&bP)gm8~^|SaA|NaUukZ1WpZv|Y%g_mX>4;ZcW7m0Y%XwlomXv-( ->M_R-e2JhI*~|O!}75wR)V|b4jl^!IFN3&s@%kF4T&Au&X!X3zu$}UGdN%ub`R@J8H?J4Jzqxw*?8D3FuP(_`lAh<}aWz9KI1g~Kv+|1TinIq7r6({p01&dAvXv29cFYxtB05O3x)0zXiI%db|m!V)2&R5il{h -|A7yIBB%H6(uiy)pEiiBaNjLmeibELzc{l!r3yG$&%^P(3$)Ej@woP7-$E>A!_CRrlm0U4Be4ZXkJ*Q -cLc6_hnQDMKK*rh`3gK(Rx-00gW%hw*xdDk*vQLO!xP~7vn3z%y8|NYl8r -VTixe0-1a9u6OPJHm?3-!T{B_$(V=ve>@m~>*}!ooA&6TbUSEMdKO~t4xY+RG%c4+q165m!O4wc61r* -YC=d>$CpH1E;-*M>_a7MMd;CB)Ve)j~rtImazmI(>fJXP^P5(L4JX9gM%j~3 -0lO8KX>T_p7v%h2wz40U;zG4m>B@o0f&xIcSS54{w05P)A{p)H$PvVmzz$e<1 -9>8+PQKzPh`I{%fctqp%w(W!qX5KZ)k)kNxU092QGQ2au@XA@dkW;-M-d(#Irl%QmL3F@ekdqv-aQob -%j(yx_d_D<(->+)|diZsp8|w)I| -^+>v1)q~rrs(>D(c@B`l^}+^x={aZk3qvsScP12ol+h_z -IjBY$c-mIfbbn)#g>*q06n#}p1rO}7_7#pZ#fqiquQ-keZMhfVTN0c-Sxw0UqpUhdHD*+=i?e^IuJFl -Gn3yG|nG$xRUT=!P)lwUYN!Bk{Sw@t{(%^^O_CQaJVkQ$vd;Jw@#R3F4U6tVtb~-Q}=j&?L(uf`4RH> -xumXoDN%K6fiKlb28Sg@2rvYX|1Pw-~b~km*3@FI~m;JfU(kTI54>?6f7KOZlw+0Ku;!Y@R9^#V>V8E -HD5um5v$-9jKm7I==cgSi`j>;DnloI#_fvZ=Hbk+4T`oxUNfjVo*FrJ(WIQN$rJh*cOQD5{pQ9F#*L( -s-M-=cV)R!Lz^v9~Nx?7VR -b?zOp)iJQ;e~A#ABy>;2EQug#PxbC>4#Yl8e=V{1n(IP?^*!58HC%@)4@$!&6ffC^23*dH@6aWAK2ms3fSz9ly-o-Bg002}10012T003}la4%nJZggdGZeeUMc4KodUtei% -X>?y-E^v8uQO!;QK@7g{Q#9cqB)HMUlUD2*F3 -M)Iuf&UV#gkGiAWk!m1>vo>PX*;4lP|3|ut(mW*2sKsN+?m!U(q#Wg|8=|&-u`6+{s9g9Ls)R7n$6D& -I4)s0U|k}%Y!UDW0?d^01ZY-BIwl;r>s50KS|aPGGfj#m6TH<#t@DZ3t-&-0OCReLycTY?Q^TA{VimI -{P!1M{aD_|(^*yF4phFXQFAW!Kdy=b~!8#7{)Ij<3C52P26q;xSyO*f@LJv8^0n0Eu__2-AnV`>G%Ll -7AvQPAVLJ_A@*agS}UZtn|V<<&xTo2+R0CFvsBk0zNlI!toz*ON|9KLc3xIij?#hHDFwgsxMGW0|XQR -000O8%K%whcdOf9%me@cmJR>_8vponJ~3vkF;jCxW3yohg -z$Y8`Ie+*%wYzRhF#EuJ3EK?029CY83_9(uPxN;uF^f@=8@@wUXV&RY2hLrm@BERj$@<^@*3t@(;?sQ -%#lqs_f#cVId6RFwNiPFe>mUjOM2%~*?cj@#+`f0R6{Qs`0d&v`Qpq6iU)Wd1^lEA^N}NH| -1vBB{s}7wY6S|Ydy-@R7(Oy-vDlPm2lNP+>2E;y9{GhGd!rD4qU&_o}XiRY-$}jgWh~%Y`ki3;a_W1M -KAm7rG+d^eNKb{i+iyoNaRPz&M`e;`0Cn3h_h-`i)QW(2~#XjlHYYt4V$S&OIa1E=|KIpCG+@=Ms_<4 -Z9aG3)^xv|dbngDZ>33<`8@~54QPo4mg4$27om=GYC&jo|}pejC&=R{ln1NI|hG0?Pqu!?$hPnTE#35 -OsZY;OymegOgz)mZY}JU7HD_#GjO9|(Uc82O(*7#HOV0f>g@Lp&lM@+3cP$*W{yB$w=noC3Hv_l@8-- -1pD9oNV51$=#LB<2{C7QOd*3-#4e57rn)QN50ZBbMyek@ue|?FhvLfJ|G6fE5)&}XO7<>c;Ota=O^Si -7dcvZnk-<19Ex>Q^Y8KXoBJ>twgO}720dQDH8C;84l6hW&`hDu)YuG+Y~L}UH;CMds!Bd?KkJ -4M$SxP4_aDOrsVyE!u?qjN;DkxYnuO6blV{|k%%J16ea;Y-V;3Zi;Tc{TspfEG5JYdmw+U|InH}gqV8 -jwzEU2R?M!Je3hewh!De->73!WTAm4#D-*J?C$xnv_KXzD=HkJf$Dg0xw=JEp9Htvcrc@_{AKNQ&vrpT! -lb0{y>4s$)5q4aN4uLfiLqiuZh?r?=7)+vgG5E>B054%ZA*J^<=ObvfW6NzwZgDy9)eE6Kfva|t8Aj+ -4l`3_^>HsVJCgWT&IZ(Y`y*ppA`yZ&-&g(+e=1RkG0>vU+dNAD9QdpD##isZ;wXORBG_|N#Q}&~&2eh -;5`xln9ufl|2FRPj6a3%JlnTPhcS`YnsJuImCQ>j%S5QYt!ySx3OE|^7BsW24LK*H2`#VY-{!oOEsj1 -GS^vHKlYVsgr;FtCRS!wV-@@IN=U2tjoc+RWog*oEb(rng);zOKpY`3}Fe -**_NFHA#NDl*FN8`c)mXb{Gih-7ppIcyF}$pM@&4ByBzdo%<*G;kH+&D8guOjS4d4u>N|7|x@1XuyjP -zLIHvgh3uI9{>Im1^)q1O9KQH000080LuVbTbuu)p28LY0LV%J02lxO0B~t=FJEbHbY*gGVQepUV{=}ddCgq^bK5qu|J{EDj%miEBT;dB@7+7Sr)youX*|~?9@}l-jO{QG2}x`yQYA<^n!fz+Z+8KZ;5X -UnMXiy&a}dXwb^<7t`|yoj?@9vtXrNw!?Z>C*hnq`%?C%KXIdNj -x_{b78)fc>>R&Z7$o-a`Ag9WO1;-NkqKhVZx;pl2t@D2@i#OU+gw8x_Nby?he%RO}z1gI4wk;^28UPL -b0KFo0dQD=rmiec^Zw_c`6H@B($)zJkRnz5@iu5(qG3iG$dZfQ6$o9@u?WGD3&m(U4ZKyu~i(209G?h -N0#$Vo`phI6I%0+A_&!#`uc4{>2F0YVUA&z7I~JKp6&sI&q-OV&eKIUqFV{!7I7NU5`!WOVE(^}un6)jD*~Kw02`ChPt_V`JUZ|W4uBEBoVZ}K#;*J3IW{ -QMp(gsx!NCDan#iGv7CC<@il6A28w4p|gP_7_2S1))2fv(MUY);t8(g0Kc=o^0BNslfb`vK;F)Mf8FJ -UyzPv$W8gk9uf5r6Xad-pba=F#&v{I@JOk8l0@#YmPkXO -|HrX3_D_R8YX3gvdq&R+vH*lUm~{2WHHV~BDfSxisf3Qg=F^vD5i%63wbUWpJ!#^BdcGZT)pmFwDbOk -j~6H7SI@_Pzx(5_&p&_TKx-KM^n$rpA_OgGVYb0>vmy(##A9*=vbBOK_)PdqpB)0jz89bPdXo^{=_1P -}bDsO~{O|w>?P0d#yr*v#LrN*~kwy8C8o0Oj&%XD*(~Yl0zHPKxm2)J0sCek!yq`?(o_SN<{Cb5P!Y* -S83P`rcOdf}uXo;m6`|M3A3kHf08gvh&f!mr6Pd2dbaxL;WWZ{bEJOmwNGGkmaFo3Mc*c|zlE=tMgiO -~Joub>{;y`+JqCYHn>1v@;QPBgm?S(fAPTljs8zkRdtoA=YZBaeQ40~}L21iS=U{^8x#8L84iBo-k4* -+&^5;b6t(%9SfVg@xoE8~=mNvn-kFu3)@Vdf?V_;jd8$v5$xJ$h5K_^6J?O(wCF;Hf6Z$KO7u)?lkd5$J(%*$x -_`m*<844H(4ZRJ0Y7y)I4umbHu)lCvDElasy7v^KA|jb8EhO2h~2rGKkmfvOoc)O$Id -zKZ1#OrAQVd^pLdOfI0vs^{GC7f)T^MnT93sT7w^YeXn=|D6Z -9!#)Gxsg5x5HqX2LnPh6bzs7K<&i%u -v!%xdqIVp;rTIt*N33eRxez8jrzX&;)bzz9D*M917+)&&Uz{YNrqXOAX7zYI*%GHh#6W9M!h-O{;hiiK_+z!JilVQ+A}&F_NEE0)5w-UG-r=4g|Hg8+o@YrM0=!;;{f-e=qw^Q*O|jFPj -l(LJjOxuR>J}-QRSu93FrqdshB9z_;WLOY8IPhESu&zEj(=yQ#S(v6q_mu(0^@wtbPxoT65au%J%%lk -h2Rm(7VP-Ne<{U=nYFBhCIWs30trdOfK8N=F9kg>B9Z4cZ_ojO%-%RQOAxb+f5-pR`tPdNz?Mykp&D% -VuriW+>+!(Z$3V#-E=bPYUAEYwgxLU&T1DlOIMnQo@=DQC3mXj55Tu# -~`MynFExJg9XYRaZ}2W_Q8IetK$8Wxw7sq*n;dux?a*0NTt5!vp$z50c)A5C&f?ZHGV4Fb)-2kC|Ase -Y}*z0Z*WlgRh!nNo~^ZF9;jXr)B81^W-`i+XV%gUP~4`ZZ5Z5KLP9RO?ifA?S^ZEZi?6SlT^{n);M!Ox8#Wa7G-qM~*b^3c>tAXuRfNL*lH -I$qol3TNJA#8{-{vHK;QiT5rk%WIGR5SRwmvtuA7YyOFT&0e$v8B9vDUwBdqO2l=^KLOu?X(2)KngO3uN)gc#(U8Y$pu -E_@&>~MEa3luNan7<3oUQb`i{hxS^}O7LQ>Nn622>ID(#XH!9s2|S`Ah!UvFxbkj -Jz>ixudiYv8ryREJdJ9@bDFqB -O}?_1BZeyJ4U*S$@;cUDbtZM+2cm!{sB^(W6b3&Hnfzve*N2~Arpk$JmHkF$!EW2OT&cN -WjU_(&jVb&=46|oN=5c*P*>B9$_o_a#;W8*PS_QQt(+^1O6_3v#9@`RnJSt_d9v(c9Ki3tiw14NKzX= -Um9g4`dgi7UIi0G{Y%;(U(AkfXw>jq4v;MXC1K=)GV-9d_4ITW9c>Bmio7=m#P}!ax{YiAp6G4-eeoM -1O`!$Sm243gy(UwJrCQG(chG*BRcS#x?AFOb%a3B_+oTXTdA1RGu{)&7QX(+lNpP5aDN3A3Wtmg -#!HZS^f&}zdZqfXcHY!yBP&if)&NRoO9F^m?wjpg8N(*FaU^SdpXyWh(%(I=?vUx*QG}ux~aPRFyHAL -5XIx-Fk9oEdHWi71vQz8ZD;&nWojsq%E2Np|TD%^!$6EncsZqb?hYyN7EM$nXdi(R -t8Ol2UJG}3usSHLXO46Kxc -0w8T=s@XMjUY%W^z78mCx1)o{DNWm^clg|E+ItecZJUmIHMLSfkHa(&nd;O^Cn1b-&1fu@;aJ97VS>T -05NE3Sa-1=W`EZb_C79Z`oGe)86tGSP$^@ZNnWh43{v3w1P5_{n{Q@W3yYn8K&1eR7qLt)^_O%urDj@ -s`KTG+=IkIav#WZ2{wkR90>y^KZSjA~*cX^;c{9`}&L-gX};1{%Q-ikKj4GKSq>Vsg%Dh9xHVZl}mS_ -75QNV#JRB){7xI>|aFLK0FVCz6ApFA_fFaKpBzx3;+0Ut)5n8b1cI(l{TX+loQ=~Or2Yut~NqMl9d}G -joUTfu|=7NxXY$~iiuMSIDpTEK3R}0EQM>PKlUc|@zb!#owxYWW#_Kvp_gNw^_k6+Y(7ngTOQGF^}j< -&?0wPA_u!|AKK4mSXZ=kFq*rF_OuDzl-(EtcAYGc-4g(37i=o_y-OiPjWA#^)mW1VZC -5EHXeOvgL8X-5JWL-r>E>a+LLt?B!gUnrEdRsGgdfK&b3jw54ML*9x4kH-zq!}@w4)bh5u*xSwGwKHZ -9rR6KqM>g?OGYX$QkP(R=;+{OCkiDou+wx(tKG85dHAu8krs|geOdGT>~y+W8+tQf -5%+JPGoBB;u?8Nrr!?z<{`QKxE$%u`zG$C&%qPL&PR-iw%EuQY;08R=3519#bfVS02~uoDw1Sc;xjxX -m9i^6ej7LJ|>nti0QAc>RgmM${LIm||t>zUDnN{DcBO{p10vT`Oyy{nt_o_#2+}9T+Ogm>n@zcbCz{l -sKAuzb#vA4WjmM-EYA|sV;^OYLl2~fr0N5ogPkWi)3Ag`OdhM(LA5Uq|r-)scV42+ybddQGk>c);NGD -FegFq8VE|7*yI>hsn)iyXy*4E89@1^1AN1$Z}z5Ge!l$4q?{I{F}mO((l9IWXsoc)+Lu@ -zlzlf0>Z1X5Vg;Zu;UKCl@V2tSq?xk(Wo`!_%@Y2vf~7(gkTzJ&Kc*13?>I?S$`-$Hyd$WRFum6z!F^>G0!(b -1`fQ}sg&BcUsza`U{#d%E6zo<`zRPrX}JoqNQp7QF>!m}!JKHA6J1P#{bIqFLo@VG7nR`3)|i(`udz{ -;?s!vqsOl^zYpq-&t&}l6T}+wA$iDns{T5-M(2AEfDd*Est%@Q ->H$z^Jj9KYmJPVB)a6^2h$T;`Fd^B|5s*7NBBWb@<(@VxCC5HHVT%&0y%@nld<9yq<}Ft$6J=j9nOIj -$8gyPD9jZ<;>rMyEA&EpjWfM#dP73NPCbc!{(D~M<8!**X7ZNZku^dm?p8GEKs$NlJ4EFT#Auh+U^ws -khY>jtJ)Ljwey0T$EU!F5P*c$kUZ7O>0wrVmK=dXk@ZxG$eO{UI+fEpC^KU+5NoVSt>058g7(rTn#Pn -Yxv)CndpCu$LLPxV|MW6ho0`)9CY8k={MO~4VV9fGn%J1i!~k+05T(HVrST&^^)TK<@{1n&u{g6OX*C`@D2&A2akwVx4>GH(svpC3-DDJn0&^0_*?>`W6Lti -a1)qQz}2-V%{J`OK$@m)XE_lS43t{vNs-X#!X?ST;q=g4Jvy8m_RhH)0C=}SSCu=8=I!zAi^=V`j*b1 -^nX`H$U}Ky#^A#=gTwfnliyFo#)?tJ$K6&&Bms1R{8fRPV^DMD_phO>bSMp$Ke=psRAKUBO@pA*-WIL -u~*!2-v%I(LC9e$T9{@!8FdQJxt{a*KLK1sF=@DzxM0sxqO&fS9k#-FxsuRh|6MmK7B+yRb8f@R-s>J -lHQKvn5$-xtyw*<0POFG5vp)7d=r@&N#5Z%}s^{>A7?x%qm!&11%cY=uyuszQE+WAtdO(1Bb)_z8iVc -9V>5=fWgWT(-j0oGCknulzPCbM&4_S~PIy%q5s^=O%oo69kQx)l2kNu<9?{QAE{oF@}C^`24Tf*=qaD -6P=YeJ}H)bE?vElmRqWqXIlT3#i?73Vj9@uqpwg~eSkjIIvX@}PvC~{J6=cr9(z;UnoITUAxSm&$M8l -CFAIHWu>n7=3afoW-P}BKS+0sQ`JjA{KF%m>1>?h_~-~9!c%)X*=icO3`Gz-@C}THg(<8M&&#FN) -RllDP5Bp#y&+)SCFmW7NXv=(hUHuH7YD}rPYxi*C&z#N&;NM-+y*?j&JirT0BNFY?yC_1e39=)5L$4$ -v#xITtI(*6f{=hO(nyDEbfLGUOK}=wQ96(yl`9HQah#he8{_2*QjA0&_MU)3G|LRBxF>@+>yUZY`l}N -d9KOJ|C~8F3^UYzgj<jZEmnhN{X)4s!>y2=b7WdRE`_1RcNW2-AwDgcWTo54kWp}xA;N8r4dXR$ZslP1tUBAGO2%d^tLhXp|i)tc2k3WK&zW#?@6xne7RNkx(qQ~vl>7-Qd9@^EyuqCA -psOPkI%s){ek3r%S-CWZNLrE-xAQP1>NmKrZo_J;L_Rlad_<^jBBC>3}D?z)@u2p|_%wfgI+U`8)u!sL5WSg`d`?CJHWl%E^Bs9IF4YPw#10b$+C4$<_U76xm3 -C*<8m8_dcaG26$AFQ*O-s>78yt-Z21UsjjLcsVXMzU0NT5d#~@lwU1$aQPkJysh=4g#5)#*D|73X6yB -!Vn|#zKba@m7+fb4LRM`oa@M8l7cB1+7l!=&5a}xN#1HYTRq&cFVlHp>5omAA5OnJAa!2cZpX8mH}BrKFk@(Hq|AyydzWMSRc862u#LZAWjie7Qnq_F|n% -E`3kC@HD8iCdRkja&J+=gb&)i$)6Y`?{#11&zDjWBZ_>Pb=O)vW&KBW4f6weX^%@eY#_zS=D)Ux!ZLq#wWbi*`Y=ysH;txoaBifE%Ua`P2wxHfmVxL-|56ks_Ba#tlcH%Pnw4gdU++%G1P6_}! -?0`^BIH8d3>5wvllmjofQico6ROP~E!hy;R&*+|XJVMP#0-XUw*bBtU8-BF!dp4~U -$LN*;}L5xm80Awo#KB~SDdO^n3X3h_V`6ud#QClJH9_!!>gm0pe&Cp8$Va9SdOEIxU0<(Q^ec5yXXo -v3|O|R2?W-F*@et?mke0jmn>0_(JO30yTdpxo~J*`e+UsYzO#%T%9by)*Iak2SH1H;yBhp~f#CoVLqA*>oZ?JZ34)9%pjV)%wT-HRQ8KGq4+$R}fdXP -$O^Nwu_RPGfwkOq`N(YS-3q%U5(gMAqX@wt4BN85PnJAKFy>lwP(o+;b1(&Se5O-XP#8DMxk%3Pu`=u -#bgM_akB<_ierK2*v@Te?EPG+(NoAa9pyC=z*Aqoa=HR9!($c|9L|P5pZyKMT)Y()CSq66yOeSpQM79 --pXZZg`p*E)=xVfSn1`qCvh!7H*n7&KUTA73R+N6PzqB_VBwz*K+Tx#TK%w=zJ{a`?y@@%c*VUGQZrrqOpq_0-XNJ`1zxiz88y6TB^12HR9e^Y8hRy9)Q -bGuc@naO4<%8~lLr!oRcf8%hVqhP!r0K6g6oMLcid0%yKDO{VH+u%C}UqYdv?GSXx-yg&*p2m-fB`x@ -v*%^<0%}*dob1o;|i4ULXLA6&H%QFT!zqK;8i5qc4LqJmj~1&$Z&Ng?yihx*)I$`umoc_aIf-HQC632 -cTQv=1A~ek;SzIb(?{+OA5+HjsFQaUh?WWgHv_VBU`1ehxR2ZEX`F+(W*ePOHl4Kb(AJRy>D^T?nxrQ -bA6c?9{66->#;mlfU1@CYUc_HO6H=jjy6@~9pU(*!;v$;%L7@azK_Azr7BN6AYH`bGos&xgXa^;IRd8^)vB(tCI? -Ds_lTdpw@>2&qU7JmXzO9KQH000080LuVbTO86j_-+mW01GYv03QGV0B~t=FJEbHbY*gGVQepUV{J(uleCtTwvey%irl+n#3L_U@+6_{fA2H -nxr1XRwT)UjzE;`%$`NK -7S%g7=0{P^w>Uw}isU|zZnjxemswN_OfN+UL3U{-{6h?$dH%)Uz2(EeI}G&fQtSsa%R>@Attyes@E(g -kCeD%+qN!cR*M5{=;YCWK+_`d$ztVe-;TL}RRq*@~J$aC55@#HzbY5Q?h5KDC1q-Dn(g3J=*u*oYRob -zz87#%Fi0p8@5%E<~SNfI2&#jPn;S-UEQf|DdqW5r9RXg`t+8;#MJgy)Bu}h@>s94^LYEvYd7qMb;Q` -A`!@g#|Gq)Pr>DOTx?A7`+*QrQR%FPN;#fX!a8%c97%beebgY0f^BwJ_UAa?`lC56mLEcDPks>97 -pTO8TTf_PDW{Xjx2MnGi7OaBLf($vU#6G(U+LNJi;!v#D%J>|q&BkYJK50gY3)0?{><7XjD&`1x;Wf{ -$e9K1AN~gloia`;fH8@01T14x+^Ndx>i^UR~UMvSMDU_snEl8%kfE2kDjCO{*Ln9{AWFlb*UqorrCmo -U*?S=2~6LcmCkVu7$RwOs%qQD#eQlz=BrFIs=u+3~Q=*vqo+STF)CMErraxfJ8fW3M9*Y961j2EzLlB ->+Wxl;v`zG!d^ZE6_~4!jei5g8*z6&+0WN`_$4XO>3U*;#9-gI!_f0IrLiM{3svl~WiX?F3OkAk<%_IA#I+`hKpO?$=c=KU -{vxGmCWts?zF2%J}iXmngg@badOoCMfyA;qJ3?ZQ(q_lm?X>vntHmjx3J`REKlcrx6Wx@)b#2KLz!4p -!yS#5yGEQ8BrU7=g-V!LEeZjVmF|hb+4XU+(}mf^ -h6$4dgN+IX-CDjzToE>~mIR+`HXPSx+ZbI)>t!h{a)!eH^e&Q6cSGQSB-46|kL3!+Q$1q&AaGv;r589 -^vf_`P+3N6_&u;l+<5FbUN@PS5pJ#F%Y;`(2D_{-plL70g4+M-aPaWcia$TR6N4UeK}?KLm4G(a8&T?HA&a_<1?3xf9ydu{#8pw=PANrhR8}cKySQRCLU-=aBLa3)fCE%Dv}BbcN^r -L4S(-3ZIKP0LN=mIq7iGGRnw4imaCN-_1k*5PP;lLafuN=8UZQGJ``&^FgEd;R{p8)QIqPmR1~UF44o -PDrr&0mY5ITeUn!IeXqC%Lrh}2s~9)}8GgcK=~;s9rVy2YH_Dl;6+)rDC|cm@mn!F&4Wy;87;bgE9MO -H-lF=IGf2$$Y|6iNeF)I6;$1w+N+~Jx-EP=}2$q^$JzqpBo&GNd^e``F027Am}EKDZe$JxzEd=_g%V|QR|8tXYASu!wz$h(s -OZP{uBQSyp|u@&kCp7`ySB9lk~gc_Q1o4q(=N8wYJ6hhPGT!(^WVCiu>xO2b{qSd=)k0aczw*B2Np{OLMywE$h~$F(gZ})X{8z6s)GAWM!*yBIJO3zX -G%I$6OjUpb>bg^N+h(EG-nBkOKxS{T3FSb>MOt$I4U|xxxmT$sEx9yf&JP -sazXt>4go2{kI=gmF$aWUpQySZ;Qfh~UJ9aJ)_?$OPU#joM-^XfN*E>`%5`u_g8sSCB?|xzw&s_}X9V -sPh^>J%oHi=k7ReZ64Hva3e~_5AJDry-R)T9;{+po@Qbf*%oxfp=lUcv2OYk -|l%6cEO2atkzLTe0KG^}-`vQpA|e$Vb|UzZ}nqa#LJKvPz(-vo{ra`4}uv -o8E&B)Wrhzw4*eMV0PR9_XPH5S`hsLXru{?3jSqw&2J?gSiRMW&;$}`98#T)Tk0Dom8nfyumb&qo5+_ -UTdCT{z42J6jSVA4m5QG(deU{2yKo#~^(UR(ana&n0b(m0uIr7oM{%IrEnBg@glFT)29ubh!q&)JKr< -M$k=JC}xGo9II+Q+9i2U>k{-5GS=)5`Ki5NkJDx6x79!ZI$e7DwOS_2fonV^}RQrE5MP-5&gV%)6;gH -)Pflihhb?CbFR#=5=xI0)=lAXUG(0+?UB5m#K6*UAd)M!I$&O -EH*V8GRpDs?GEsh>LyYAu{FmQb{Tj%w?QXYr%*?E>;odbsW=HZbZzi;r*;=;t{ -Xj`r^;enO_SYSXi)M1${ns(OZDZ~ohrV%%!W?y%x2#_Z@`si-AR-xqCi4yHXSNud~`ij>L}Dp*v`PAg -qQjbC5%<#L?b5!OVWL0;|wqo}*aCEMF~uw$=# -RyI$D}`O1axO9QUOQ#D~;*=RK=j6TS2#HbZ*w5P(ho}tf5|OL$oPuMxpSqdmTGt_N$BDFh%2d(;KXJD -DuUq^QjmLH#00S>i&TDH86QzNo^8R{WbjFY42o7V_3Iu>a$s1sfX`V7n>*+h((=%cAO#v^@6pp!iHz4jp9Tw22# ->efaRVkew^dD09RMWN+%Qg518i5|W+&{KK?+!J+)3{Z7jdrfR{qmX$u+*$l=Riw-L|Z;?~m!y!cGOE* -V!RjMMVYHpZ*1aG<-Q(c_?TBoHO9SH}4ImQ!eVN~i)cDS$h9#Rg?g|dvi4jB=yL-zi=*Y6&^f{Ffk&Y -n)`sznu)18I7$~m4NpAg1LcT%->-3mx;@KkHVd{5DQ1VC5Unv|Uvl79o16s1s0IvS>ORq~ZUz -@>{#lFmwE5-Er{X+Ybm553j-RbE`@LD)$hjdi7*M27ycS?A5 -DeQb@NKFA7To>;5b|lp$H6(8r4YDQ)7x6R2x<@r&QPBzGr+Bm$Ej(W~gtB9jh8MO1DyQjFVbk)4 -MG6%Z@+HBG(F_C+VT8mSab!%Kpj73wT8jEwq8hS=rQ?1Yj+Y&%k@5?q6Y>O;Pe#K -mMWQ1lsocb5UK0-46Hy^J`@foFUS_=sMBGKp;6~wyR|sPUPFVs(AaBf9TZ|3i&SI7yUNpA8Of4^>AI3 -?}mL|)t$PpYPoNU12t4nwspmG;Ei{8TQmb#_q7B>6q3LZ2dCTh9p^}`S&Zed-glzpQigcz*+1`YG>#= -e&s}Tva3i{1(dVUVoENtBu6W1IW8zkxWB;73*ky8C%H-jaI>8$>0XKR`)dMB2$eNdo~P~rtYQ?AnvHaUlxs2?eqf;zHWh3+rHl?O -GdvEFE(|<=@*_~u_p{@nGt8Glho5E^24jVO;5mT-HkuYkAzfTf;WvMoXJtU;XU -hW(Sh7A?*je@=*`@HkFUpd`=@1q3LFI?cl*J5g&Kq{UFfp@co5ndmunLz!oHb-xAuc3>^%6|d;UzT=z -v150^!ld7Mh&zq2VU^XeR -OL8C>3edIaEBAIC61B}OJRX+3#(&z>Nl|59(Dhv*KgkLnKk`9cDXHLo`{tUOXFy|5#7Li9N<`kau$Tg -CfEl%&Wk@AX@{w>;IX4q5vW1|nkOzaI|$77Uag?+eKg2wNYR&zE7l1sgSgU+?|`z#rVV@ZaRtpPJJv{fhpkBFQB4Anq%^-XTDCY7Zw ->E+!P=16ONs)6#{o-s@eUGe}Ryp -ji!zAzq(9&8T{ryYJSntG-KYK|OaweIsgmAu(3nXw2S{@`|kFV>rg_6v|bYZRb^Vvl;ygc>AbWf+*t -oRB%kfs~Vdp?qvq$Mh=mxOaQ*d;+~TA{e--ya8E`+B0DW3p~?5;d6RtQxXV70V3?U@WcAK*8Y}e%uKv -({`dU-q8LW1Yvi@Mlg!#azx51EsMbeAW07ESwzVuKFaWMqZL!68!Vn8|qfKf&II%GULy8(?Xi?v5 -Hlk<#wXVFj>^T -z1BbAlHTC}nL{;{jQi9G+BkX%aKt(^qCQ?xai)nP-?VZ;)-64)uo`jA8I_+)`9|dw6siE_29!g0I0yS -#_Lt-xj6#?+lo-qv77WP-PjH;g4qD3{%lM_gd)<3uA~txxi -i~+F{chOI{>J8DhSp1*i=l$!A9w5Fv=OobTK=Y_AIcx>=cyJ$#;<7O5UrU&7;r`tL$oLak-&5ZGmNVe8`+d>w7-=)wMRtM59S%f%_T<7FM1ubT1ihv(u -LI6NyE?MG(Q0V(ZV_RYrMJi)S91$sacWJsJ(cwB>#IUXSdH`I_;?;q=rti7Kml7n7o$z5r>M=!@ZL^IXVt>k3PUJDA^FPhPY#2dg32vSqf$@YEuauS{-MII+kqbur*Ak&#; -$I;q7+1dsJ2nGv`D$&$;VO&nGe>d(51!cDxBTPc&+-qqF##gm+|iryVwdLxpDuH+IUGx=+m8$y~jxQr -UrL{!dXX94-}GC%LLoPf?{_RFt0h|Us0LTu)u*0*f+jLf9*=dSjQ%XeIYkSak8$a1e?!EjHne=!v7`$>7*+qTN`nQKG!|&6Cj>ydvXG^h0e -~iO+9xvz(pah2h+nFC&oS?p$Z>8V&CbIUhl_&i8Icr^5bhe6ua@9Sm}=vn)g1qi6WYZm3blpu -Md-(PETHJ}x~c$BfdWI3xHSZONoM|2?!4xT!%L8FCa0&!z#M5eav^--^l#j8YzXQB}0h0B&`zGV;o)Q -5k%@$uQb}nGu=0Ehnvpsswo%OYs}61v(JM2KoaX3QtCi*1?Rl+C{(w6+mv$Y|xS`!o7$XLR!|8@iH?p -5}831*!2>lk#v)MG#<^T^JH;^V7S8kyuPB-)YpjRI%cv9CS5+tnyPH1pEg$k=U#aEY!B42r8Y?SQ^qz=3oaM+(g{Vf(;B_x=yV9RsWKxm+wFk`8Lbg*CI6X -$$kV@ROjVLOYy|+Mg@}fC}-lXS)dXD964&xy^YPQ>3LR=Xlv}+jm04s3Qn8|uqH$`Wpb0Zp9T`;Zl6K -_<;K6;L(-B}~X=zf{*Kh`6%wJXm`D)>b+ZPP-Sz2&?tRic};b(d(C -F1o9w92%-+Mb-<5jMyzElQ!Bnl0d!`dLg;aMdrp7dpZE -kW^d|&F3?AVK)iqot3VrCBE-p?ZCn!GF)x8IJXM719!Wjr_Del0QdprJ2CX5Ete&O-2b~8);Ce@>!1^8@JB<+NO;C<#+B}jqcf~;l#R*F@62>!?;3~hYT_9#Bs1dP{5FsuDeE=3kiH8 -G280!Hb!hvXojA8?BjDm+wR=F)OaPcAuHML@c$q+--B?JAkulJtLLyglaMOhX)X^1xBs0~mR8T0CXl6 -7^;Rw@0MFSVTAteI}>d{&~``q|#ji%B5=)&=18`0OpM8Dcu@j7bbL)b0{PgUhR9b~cCgM8e_aADp56o -aKYPKCogfMKkougOZH=Ih9?u`@4MaLSL{;|MMkA*Fn>&L+>i3BBAp9?|%mqR&h`7`omIN2*HcG@wW7- -T!>zg7M_GU_|SJYJ>(WhESzRa{qJ*5RE|5R8m$2!g#xSInt~q}8ar9^VhDSw<#y?b`qhjU|A7+r#*x% -+$4@_g2Tr;g8&)pm6Rfhw-4OLk`#CNr=57AMR6t}^5gJI?5xF;0MA%pWP`5=hLPodoW#7;uin)%+b%D -evXc=ogRD4C8(H70Dt=Pf -syeyc1nz7OMwM_8eSBN& -DRti{)r|Z0Ai62}U0T9Xu_=2$@b0i=o36N-EpQBqK)`Y*PZGPOFsgAR29cliGe!Wa7G;vH+9Yp8hr@m -Z(0}MY;c_Wp!89mAdK&?~lvvTvTZ$e~(6HB6umrz;@)azLUX-Hw`uyzdE9m+4GS-#y3i7sng4LDKa(6 -&>oxzi{?+?%Sdx|jVD(vhg|A=aBbpF&|{UMQY!=5gb%1?n{OZIee`P6M+87?6Ia!T+x2Y;-+O;lPc*cn-#0y>e5idYYGZm#&`Sv;GB_ExpvPM5 -ylbo?zS~3ojUMEd6fYkWiwRVcWgAMIY^-DI>p6jOckVS1vPE&h+@m0g7q=9iiGLW#g!zBiLG}_*h;{$ -sLHya6j7N)e;z0)qX5pmB6R)o!G$cIkC`an&jw(L$wYvVi8Z)5ZF*c&Uyic$MhzQp+6#>w;+{74O6on -l?hhbQ=uo9j{N>05JSB{(hz53g)}T74S$Wy1d5u|WDtDsR2Q{4P>WP_E=NhOr=POT8OzR^Yk0^#Q#M! -zbZ%lXcARTpgIC~zwo^D)PJrm`tFhkr)=eZh}puA-yYv$ZU>UgpUSa@uawP;qF`WFQu@C#;hcCSqzb# -Q>(J={(!x>|xNS|1>;eLo1=GOK&-2!QgTYTDLq{=t;$1|LdrUbaWqQ>cUpIwuiz0GD5WPGEJRfUy2fsBAGzUgc(h18raTp!ez9jFFo-VN9 -Ir{2LufSl#dn|S2cd55TQODS%%S7D-GIH*zj3w3@t|f&ezCj`@RLM;%d -u4-!Pv~PF+nV;xW`V2pJ3N=a0+J;qDv4&~1|d@2;;AvFmHvtP=ASyka()l=47DQCNO`!HJuCF~b{{0kuOw*s -&l_Ip!7&I+KHK02l0I%*HR%}m`x{lX6jdnp?0qB*S(<--6J}$Oxb^@uWFT(-ieCq&C>E+al>8F -3dEIQjZw{7P|=D7kBsMF;)>lp)WhLbaN-iG4!LX$%sccJqzWkSSKCxqG`iFv9;$Z7ZfwjeHO>~r`o7n -QE}}_9sxz!9(qpt2M(~{7#TLuSxE`Zd_9|M$80a5eSPocwurjaWJV;IuCbje{*4%Dc1aqTz^iG~7jJI -bCw{GeL{BQ=@Og;`5T-=(2!#o;IGu=AWKD4G0J$DZEE1e;`?*uazmSCU?K~lkH=2^jg -4d?{S1%a+4m{gaMbg>sus@vDQVv|6J$?Fr|2s?Cdk7lR634 -`P>2noOdd0WLNJ)7|kc*arDrcF|^gfop^_|G$44nee@lzOb4{H@OE~Ae=IbPQ?>Dl@^(v8zBbjNYS0_fT^sAk@LJeQgdoO_xYOJdv_0zBUDaW^bv -~xh|zEKsJ>g7E>twNPjvL6&?Uxxc!uXu5&fmA%YQk&jMV+0>7FY-+XlmdXz^IRb$u#0s1Rls;c36dzc+Se@g3o@(r!I4q!y`O#xC+ -<-~+2SryH5(KZL$LX{a(lY7K)X!ZAmur5Urz`itk--?2l5`rHPMq)%f7Iqnm85SQqBXZj3}5q?|3j#V0`;A`CC=VaIq&)spb$c>UKGuixSkFQKZ&&}K0IGe4tluUcJ)hWo_&6eF%Z{(+ae?6{Wg7ky+UMyEO<@n_Aw66_!b -Wv{1IAzrmV6)&mC5?uv##t*u7M7BT8)e&hT{(uKvr8E^ZGbDU|aq;2}kkGXT!`Rlm8Z{?%$jJv0B^Yh -|>LU6h_A3^2D#}!l>qLS@egsUI+Qp(({J>rOjJ*-dz$d_`KrLhl5^Drx?=aTK*Hl5K1?WonAIiknVgy -|Y2%k!^3j4uvi7}F@i(!GWn?|hh3S4ga4vfUEZWrqFVN|bfSsJTJ+azx|$5LZiqkK}WG3D8d_@E|%g5 -PJ{ANEC7i!gS>;b*vXsA6vaAt3gcz**cMXM|_s -fBONy1mVyZZ}z7YbS*qm>TjJiW*>_R5Iyg-W7jyywFyAH4Jv8MHH+3S8hAw|lsgDgDMz*DrXacE>U4n -d(wW>r1c@u1_v+KI7RHX_S&drm6l?6qkMKK{qnces46Bkom02Jc!Jc4cm4Z*nhEML|SOMJ{r4bgf!la~n6ZeV<=}+80YKir#fL_rbeKmDkq2QI-5rwz8@FB!V*_i -D+gpJD8Eg{Q7%NH!vJZ_9nM#Q+7orfkvbI^yv@ihaXP=EB@))D)heHnNOki<)QfD>55`ubdjq}?Ki#c -y^ChCc>nX4o5jMkVK5tKno#+z#}fBgx^6Hwn%@0u^xQRs!k9RgrHgSp*7d>Mi$F8%U9FN)NJnkCX%Js%#UFq8Bpp%96bsZie#8E@&&n>-Ew+?0va6J|6W{nb>RIU)^NMe?+UY3{7WS+VQ~;dvksD_VUA*OSAR0ixhAJnfI=bKD1BlaGyo{)< -Cf*2He)JRLyZI(fsAD?s%#HTMduHdi{GkZpXG9aJFa{Mi{C+n9bxqsg?!z}0PNQd@%i%2yZ4u?KEe<@iVpNtzVBV^LR&FPIy;p6Q;87d+p-=jh+VIT5bBtWiVZK*hHu( -7!j3TC-rl<>N3GY{r)THT$hCuP_=@JYx4!km?d_^N@SNha7{b=Bn1%6AJ`PYx%}s>t;U`A#-d5tr%5@ -Ou+LEWFFy6v@>veR)*vZA92R!8J?XoEMA<#`w2)?%rbK{$?b`3NYff+5f_QL`DOzT`Fd;r(GdoJTAkW -tv<>4%m1DLmloehE<7e&87r4OIFstM3tU4rbrkmXdnc)wXm3DGlBS8;eO1adq_g4#1-Tk<_~}O7tK7u -A -W3#ct0EGq@2Vpt}NJ*b#-EULqY{nhP;L*k6tfOhlv9>+@8C^N{lovrYe(?9CLl@7Na=^w7gX>?)Ia_B -31W**$f>ok<2>l&EGZ`HUW1Yp{$PaLDALED}Vmbhow>kj}D2KO%u9a7Z7Z3Je&hEVhdT1Ab?l$EEVCY -;9hc{?z2hvRk;LO*^H;e!7?atFYiv_T5e%U+aG(8mNZ*i?aJ(KwlM8ZT)1K0bl9P$7}@D^~1jglc0Bz -d3nBF4w|#bA=Aus?PJUI=UQ@!$!SM7t#VX`PBo+La{#IM*aW$qQgga2fMUKRVG20uqZ+itV~O6WG%1) -fcjkKOE-Yn}r7FB3n@*27@2MNOVvhqKzOv;NBPg*tWu5H2BBOy$@ro4@_KHbIaRf*-}L4KSwAl!kN7( -mj6OX2CjJlbkuG;WH{i`jcrHa7PvUn0N@Ekl)AI!Qd~4taF4H6^!j^y$Is_}<<7B9n;5VPh>vX!=NHi -c_&urQ6p>MkvR)H(2q_X*a``ghf=L?^C@&)uWx&?@Uhcryn|w~58R1TIwVeTxIB8GM@Cjij!9dfrDYB -3{-HR&2PFP=Thmrz+O~#v`UjmNu6aps)O_~wRWOBiad-OsdXN-FJywTsy;Zmel51qwYVH$-XKRd#qrmNdvj*RsULWvVym%TR;iXgUeI%##y4@4)R!w+o4sMPfD -;Z101wo*pT{6OyTs23jlW!N((ggehUovX$ofm608Dh4a$6E@ewSt)g~c>2N-B8+voD>0Hh%I-x0E3zFbq%J+ft$(4`@cum)DNE{>=#ku -RA=r=)ez|0@*184R)!1l&AqG^fP2Fz -fxV?In>uVd?k|$h@5hKeGux@EA4=YLCi?YAniOs!vbn86c3;TZls;J7SsWN0Zrb`AS&esPl^f}In0u?7U9)mWfOVl`9b*~?dsHH2mMC~(-jLT5dY}Yi^8zG9?t*P -wDQcty)s0FV3Bp=rNCB&6*|2MAD-Cmv+U}523SzO65JC!-HZLCH3Z%Tx#3C&29$}wU-@(|k{i}&o@TU -rbr>QWUB0^3!e`-*;PvqAB$(%{xRAVXeT8$G~o+}mF8osR#=B`D5$}(YoWt-!qLr~bAOFW1`x8kK)oq -3V!Oeo(9JwMZ=l(s~2NEc#62?U#=i7nnsuYh2NPn~V3E{LG_n-OdfHKlc&B}m%eS<2M*4(YwpB34Z=` -W2IPIHYU5b<*5M2Mr4q=dG2{JF?Dkua#k2FZ7h%_m}W>}IQ~?rOZMI8y>?Z2BT| -SZ)C&m1#TUhx3*B#le01R5Zad=fE0mh+b!lFkwZOS`g6O0`==+x<;jJB>9k~0z5C*q&T|)nM@=`3Yurf_%J5H5zM+Qbsab6#_f(k*QClw6{#9^NdL!l}?o#`Z#insg&pC~nK-#ucNFDFFhcl!J041>7>lOPMxkW@NMO -XzBt*??!sXU)9){H>v)I~@$UQ%0I_K+h5yJDz7i#6I|s=uI#d+e|{=RRmOM&*BLu^Z;C!HIFPRI)J2L -exsO^|2}P*yQ7v3_nt5@qRwmD4vbwo;K!Diw`$22pBD0r|=9dK41c2k#x7aWrfbcN}5VM`%s*}aj@|; -*3;%5V3064Eojg==aIxJjU*V&3uqYG<2+gQ#0S -5tY#cZEW~}qXGGzCHfT=^e>t3qs}kO!lY=|Ltt=+JZOsBv>7K9QBQdwlphqaZlS%o -#t%Ck3al*^=kEMy@p3_s5Erl`rX(o!{jPL<>JzM+Z9Yyi;ZtWY9mg=EKfLc{*$qpBr=QYDN6T6z*iV_dKeqFGj0JlAwi#ThEE-vsVZMh&AFD~L8I5Abu`JYGL0z|aw79e -GBTgOm5J9tjI%jxh3r!<=@kY}W-)eh?5QpOy)r$3=^=7XK|W+Tb~zL(I0tj<}*b6Bc*!L~w#<~ndJ`A -K2;wZM|tx{*o9e*4IeGVHz5-2*NU!b6mR(xkZup1;FkvEAyGQ(-E^y=$P`7?WVp?2FwYS2m}P@l5hqKvQgle -i?JYGkSco|Hg^FP)2eudlg8)5QW=@UONOF0*FjqFzSd|u2@Re!j}G(qB;mHTjr)<2_gyMT1fR5wN>c| -=WBEd7M$iCa0RlKGy^`SK2YQbkPC5`g45|qIL5$|zgKU&3(e#)xtI_j=%QB6nL4yY~`CBGiAwMR`=V$ -<@V4KQoN?37HMwQp=otNs4sM{aClqM2Kv+~hq8b8j@PPcn_edIkJh`aa0AjMJ%o|CX;6j8Ygs8-_t1^&dzun-zE?+t -d`6t>4B*8#N)? -!mr8izm7bLgdrpZo^*qR~B#}chGeaxP1B_#UJ32V;Ndi%sS2-!sr2YGDJvDi0X9@ -CWU+mFj&)c?r=Dnmw{K|0jn$9euqAuNMz61}pE7Zv6!&W#Sr#1@*C=D#N}$mD05qUBMLbb9h~YvlE?l -1%m>4_|3|AW!ZGWSvIyS9&Mm8$BP*&Q9(r-B+Ic=kc_Sacl&4=fdspmDB#60VT+EDuH$HF8}fKhj-VP -?>O&={?hvDIqUUJKf1(ZKXVQ~_LNnvZ&UlRbT4Q?I1~9jS0}mSgmMk&?#?&KJ}@(^y*ayOjz>-$W;Mo`a -LB$)uq9e_37y{qoc0<@N2mtIu=eOJxxA1Y}a;nXb*M_^&5d`uSL*iypVU`E5#`w}5SV$2Y35L$7q~n$ -y!@A5aW)yME80p@Dl2?a~u-ZIyhdUu>D>Wtkl6C>o-gP637^`ibxJxMQkw8y25ClEsHGq+yvnIiujHC -MsvCx>$Vp$lrvRC~jr`7o^B)H9hrr9N?(?iv0BX{7q!~1)-VzRq=mNO9KQH000080LuVbTLO|0dT9m# -04oy!02crN0B~t=FJE?LZe(wAFJE72ZfSI1UoLQYtyoQu+cprr>sJuYp?2zMlLl>ppgAM<+QPKLnTCALBD(WrTdC6|d95_OsMe{1250-7U2S?@hz*I{1O-clP)*Z&xz -Qvf+zu+H)jQ}VeOE{iC4_TJ)+v~Y(wau}o%5DtlO_usFDzwQ+56s7mceIdtQT_QjP5;U@i|F`^-7z~_ -}jR&lfE7w-OeS+BXMcAZQMX0Muneh(|co`*w!kqMJ;a$xNZ8j&v#KhxPy`l5p*)T&b+iV`a0I(t_Krk=-3hI3NGzPXTMRDctyDUK=>qi0<<+adf4_W -tbtzto*NBJsc&~N{x0Jp1YoU#~LB3*mg|d|6oV!2~DR4kUjx?s*T8lE|HX^X45Hy@2RNkGV-cY65mKg -yPKpCLLDx4uA@2TlLOQAldsNIFLL>%CagD`D?$n{;&DW=JIaxWVJfJ@pa*#`Q&Ch$xmM23BEbF-5OqG>j__pivVP04l@MUX_;tk!X!a9Ri$zQYbIUX$X1UVgoO4 -s14NvFSYlbdvS45DPQ-SG&ju!q%G>%lJfsvSo)_Y=iNouYkl#8EF(7b$RVqz5i6$+ESW*e1c&+g)_RpKK`CESqey4n+(8E>}$1eEfwLrC&Z>1!KC)sq^AZu)e -a^*n`tY;vKpUSn{!VsB$Ez^=L83h}n)@EKxh?9|84FI)_|Y6zH20WS=;y0CdO;aP5pjhg4oPB+;6DEx -owD9{f#OAIb+5L<_sVo_OL7k-^aS3~H}9*dFv#@uzBYR2zMVB8uWrLtHNhUwwznax7d3W33(7746(Q( -uA>LAAG=A#s^0MAU+P4ctt2t(68#Xv?ETz91&8>5r(<&SmwS5n7hXlPbkv}=Gbamii9L|unEJtp(%8t -*jPqSQ^eIRG<(MliEuJIgwT+4vf|2)(QZT24ZYc57mEh)MdCcQ>3)TC9EpRrc%TaRts%Pi_elK#tey% -F?EJ7d-be?o)+7rk-?L7w>$LLAMY>pvXRE_@0me9@c&GUnm%g70CtjxNii1UgajFVtG?BX2bjy -U$H+!W(vl%VMogX_kwQ+Xf*I+oc6x)obvLEohm&NJ0OYgPrnKOK&^uPTsWLQrz>?Dm=|Bt&^b>&Dx+f!D9Fvs$7qMj9`aena~_q< -L#-ie@(G}7K`VE+8uHPb9NT2uawA?(e-mo{2G$fz?Xxb{^;F8TQuDJFjNu+7qLhyL0+H^+8PI`>B4t1 -jjlXzGmfGfd`l=&)dHcx%fFl|N+5l;2ezV&-(#@Ou!7%XyxE|Bsh+v7f`?v*5d@rUnB;&LlK5tC20%> -8u#@>b@w0hEJS#lIvB^FC8;X~Zv -L`5c$%XRE`l{ouu$Mju~Nyv2Ks!#(jhpu?d24^T@31QY-O00;of09jjp|AEHb000120RR9N0001RX>c -!Jc4cm4Z*nhoWo~3|axQdubTyDca>F1DMfaSd7cgV6(@qwwauSLan+78yq!_ -0)qxlx;0Pil#kQ~Yc^>#-xDb4Fi<_PQ1y9u1AQs@?q>^XiA|JfXc#FPgY;lo16f|@g7AR#m|iYa$K@^ZW5v6K}iBm-RWIh@rq#+n!9nUDCfFE{xdgM2})2geM6tC -{cltSJ_;iyDf$@oB%O?ViapN@hM!klrbEIq67)sMZqxz=@&Y-W!%BFRKAMD7zfem91QY-O00;of09jk -c+(2Dr0RR9l0ssIc0001RX>c!Jc4cm4Z*nhVVPj}zV{dMBa&K%eOi4pUPE$oLba-@SkwH(xFbsw7`4v -w)fi$|&Bb>?@qGIZ#b$~c#t=n1Zl8~g7{d?RoBoMhomY?kJJuj>~dT#d->HIpqO6Tb{u=HGErG2nQsT -Egb0@dhEO&)Pi$ega^J9nmdAs^Tpq{$o?qhNAwO)hWM(e^m#BaDsc1IYjmXt*QLXGJc84JvzZ9cxn&2 -9EpFlJkf}y`2ZxUDTr0qIOa>zg5stDahN$=x%1SVHgs9qD-uBXWh3d%vSkgS(M8wAyw+xbQA(Sz4z?J -;O+?Rgrm|sq3iSj>!Dju(u;QF96U!h?G(W_anPPr%^`Z;_3=+YKUCuMw?HD&29vBrUQTeAm3cW;ule? -I{k+9%w%KIGHeZ%lZ?ISw_xUzo7g9YSE8g&w7xz<;#3M`S)A^BI0C|j#>O_@9f0j4)8zwlaxRHx?+O) -lH39bDgZ^Q^3`2&aXctOx=)$xI&j;H#cokV>BP)h>@6aWAK2ms3fSz8Nt7P&wH005-|001Na003}la4 -%nWWo~3|axY_HV`yb#Z*FvQZ)`7LUukY>bYEXCaCwDMO-lnY5WVMD4C$o{HndeQ1;J7k3snR?2qGb5r -`ynM5@u6t|GnAnuI*u=xn%O*e7t!X!28?7=jh2;T=;MxQ*7b~{>B+h$CF!lB%@)asPc{t*q}*X8I7*O -(K1ID&DtQ%U>=m-;AkqwjB(Bv^=si>ByYi#!Bd2#i%4swQ84zRwg@Hys_FpIcPVS~CFr_Vx3zl{H+~n -vdRKr`C01z=R-zuln3QUlc#;zCfS-xwiNaDD+c21zd*dB>y%nB{n*Kqm64WG&#pFg@r;bMj!32?bmwO -WDYJ|#FNK15b&$pa6)P)A^ZToxJi}ia_A@mm?-LTEVUkuXWCwur+hqSJofPWu4sAsD_%J#zv`(~`wXM -Ir^S1{HFqQJOHA<|IF`ZuM_lBxN#a(U`E?9yI$)^=Eyu|H5t0|XQR000O8%K%wh`EO56xdH$H$p!!bA -OHXWaA|NaUv_0~WN&gWV_{=xWn*t{baHQOFJEJAWG--dl~rA9+b|S;_pi{L7aJ0{E9}MKL6>H2Sr*Eg -!p0av*f)-tEoqY6uKo4XXX2!RGQT+1Jv#T?d#;oTy1%>nKKo2Nm~-(gX}i^H_^V5FasK`TT1Z->C6S7 -=WRgrgcu8wu3TC$rFikfUWQi^Xskbn@D9)3OQ%MrmI^MRFwSEhctuU -D&Zn%`~(3=c*c2RnaeijBF5$hj!3p~GB%)i{Ot`m(&{y$sIBUnGc9>nkcJ2iyK -QqDoBPg`i=07ryOYLmxq=5}mpa+42M -_G<$AJK|4@G`u}!eEy02wfIn#ns}|!*%W;Tf!Qfq>4Tl_gVt!Z~S_ekBLar#PzT*4>DTNSb -XL0XWy`bn`b;`KRLF}`$CCjJCwFQGFt*X{?fT^ijQqs{Ekd>`%mm`zLP{|zUsJCX8RB$&=k?* -LRFq`b~B!&uo)cR%_Ts%t$*`>B;ov6~aISUWpmPm7?<+C~ejqi7wkHrKU_X<*iww_RSJarae1np&ytv -DDI-@Xa@N`3<((IzX#z}$io<|=}qMF~UTnr*_Fvg#M}*p1O-S2&N}CqAnqBPJ# -`5^w843GAJ?;G4Cg&$}1*TYaW`7i+DO2RX$kVQf6@PMtgmzKw*3E@VP*dcf|-v?RdC)n|EOX -pi}S)vf|n&X;7J^b{0%ljF}lt^IuX`D8<-L2F9u1>&GPd2T)4`1QY-O00;of09jjg1FwCy2LJ%a82|t -y0001RX>c!Jc4cm4Z*nhVVPj}zV{dMBa&K%eVPs)&bY*fbaCx;@ZExc?68`RAK_n0bSj+sZ#KWQ3!g -H}?KVH50C)+3yu}@rSnawU|msa?bh#=Fj_!vW0h(v%|vg@2j2Uz?VT+I?Ck1Pz6vM3dVArr@3s)F$>l -Laq?%=BzljU{Ck6;h_U9w^Q^xF09rj|!6d=kKZ5TjMAtvWMF6_CGh9aQ$Jk{@2IP?|pWioqe{JX$lcd -=CfaUPU17Gn~OXMpCF5+3Kaae1YPKW$2>1U>DuIn?e^x{j*X~AwA4sug_5cLQR}@EQL(0h*Ocu46jZH -wv$_5C`t#=VHoV=u`*^!qZ^OIo_4do%<_=dwx@)KNitz-oJSkjsV|~rjbkCzlg84W52{A~zpg80^s`|K8Uj-bN0dZ|*x?QJ`5D;}U!LKh;^ZS4>B=r>rgnBV1fA$a_p*YA`vlaW};^xct!erG8thP4*~I}396Gt#wclPkiVnx5V&2934W`q4HU??Jk!aPKbbC0HYM)@b4m4!78Zp^qX4h -rEopYzp@1JU%A@A&r -bI3-F5LSa;>36jNx;5fwilTG)c_J;BGu7s{;ZG)Bv6Tn!Wn>PVYZmesM*@#Xw$w~Q`P}eJORr9Vlrig -vS8FLx8A9e880xWk5ko4vY*I7C%)_?exc6Y0p~y*I^S3QiVB>BnZ~RB&pVdjIfC5wKLIEu)wfJU_;ti -;PHLovW)xu0mRTqFrAQmu+B3gz#5cQWXhv6*?JfSH=aSKu`p7T=m??h)8INP+J06>!D`A+xO6JG}Mc3 -y#`ju!<=D3N1HwdE4VnNBp)36eCk!-m`E)xV5w;OVC%d%wQ*RW=H-=t-9j|s=89skSp|{iRt!$8|F!*D -I-EDa}>bt5rwrO82qzQwqy%%*x> -Bs~LrKxH7#F_DyYAlp(R>9rVR<=*8bCmLWurf0AIJ}7DP**Fi+xloqan|1X4sOK{X -YF4=FujwNQvfjV>bcZfQ0X=#BN0+Bk_e4dgCl~%NJ7W)Uv|s74e)H?uA>6VTz2lJ|hiA!#Z6 -*#?lT~qj=tJ5b6HIrWR-QJ&3J$+F0D0M{KpCcshXXMQvYze4s6v-~x1cg!lz`!-U$5kvo@RKY-^id^= -(S>?Z2BvFQ)B)x&}Y#&zB0T7aDdZIj~^@j5hqLnGCd9A;zUjpI~5bANVtTf)W< -HV(Q^+0=<5eq0_COnk!IEF;p>v7!)!6tJa>#w{eEAd6s2Hzw`AY&VR8iNx^6VX$JH0T<_H^ -TxwmMbSuCmkrZ4>9O259t<<@RHRMU{i+i|t&$Y9&0oFjjM2c<4iGu1>E@)L;$?>1y5I$NrQ<#~=#OW| -unGoq*p6*sYjM1m=oy>fPgO+sB(ikfbn8qqb?6%D#4hpFU7v>a!SqA!o)OyJbqk?kk7^JYy`Qu_teghwK0t5j*-9t_T$mPbfp;^a%;yq2YIfSf-PIHdQLmM+1U^&PfcZZ -T%nxX?o(CJ=UKL(_W<`zgBB7=DM%5l(UwJDsDJH}>pRrU|AAE&A!lZ!H=k$6ETq}}ZQP)h>@6aWAK2m -s3fSzC@Wmhh!jhM<7tMa6}dni!%4-{6^@UiBgPI64%lpkkh(2Ui(i; -pIJucZt~7|El32LJav35_1TH$moY&HVFu=jCH1doYo+F5ITwtMe1DQyzP$kz*@xb$BY5Dl#yG%_03R!%(ZnROrD7fbP!~S*bLY1nVY{Sktw^(sa98Xo&bk=!gA~$ -{9ds@4(F$QfscwwX5TXW@#taJaNWVnNQ+GiHTJ9K-y=ph~SZm=BH{fmz`imJ$ao>*7p* -!QmKaivWDkWG2^Eusf;tCIj^LZ2`ovpY-4y5REDc^%z)Z7nJs%L3YQcxVTP9;B;BhwnnAoV82AIy@L^ -ze0TC?2GA2e+aSl55#)X<}BYmKiJ%q~aNIY5R~3JP1x*EPGwFN^hGv9KSTO+g$s2-`TRco9x=9E6nMCb^HFJor!Pp`v*`<0|XQR000O8%K%wh!`*UJdK3Tvok{=zCIA2caA|NaUv_0 -~WN&gWV_{=xWn*t{baHQOFJo_RbaHQOY-MsTaCy}{YjYb%a^LYQCd*P0uoo+e(lHKFv?Ebej&+o*BI& -!N7#cO$9RO=Cc9DIM#5w-&>F$}?nb{X1`O2xFN+z(F>FMeB3qD}ipWpw-^h2D8vuwM|BmWOLGJr0gWBhT|jH*3MxWw}k_#WCZ_D$C%S#clWtOw55bP12vRwRkC5o~LXy<5|XB#%3FRQZ3#eB}<4Q^sQO|z1hah4XNk$M*LM5GbV_2(p8 -t>Sd0KjuQeE8|W5{!#GzTrA}|V#DJksLD7g^g&y=u84Rk=H=hzqH?zdsQUYx|NMMDKl}6f*K+m*@%lJXOkHfM9K#24Wv$V{!M1Iiwal; -ut@Y%Hh-`9VYbPQ^;{gA90Ppkp)yKD=&OhDEug>3JT%DiY%&%|W --u!fZehuf0Umv~l+3Tb4VC5(hODTi%Dv!tTZ^l4qCT#jWE30iHZsC*9+W)ze%mE$xjJpfQ8Q!8QkWD4 -fA{Q(1#8x0g{ftPO}1ugWzl@6HFL%C*mcV>x_S=6tmQ{ht -8(4z?4G1^wOdGF-zFz$BpeW(hAq^h&iNE*Pjc`;?`EzFg&5wJl&HeI3Y;W4SRGpg$=a%ZInG?tHcU_O -*HY#=Jc-Z@;@k3Wo`t%}5^3q_iYp!!+J+kvbsFOCAbVW~>k(MoVp_Sh3EkBw{395d%90cn4&-a6d9{K -9AG5oX^Mf1Ydyx|8TnScLAAr#2hDs1z*nE+QxqU|jV1EP -<)4((6k|yyOn05XXiY*Gl*!wX}c8o8IEU8Mozy(;59B@yP>` -_Dk-X2bVQF!bHqf-qZb+C>$gz*g_2XBHoT5ThfO)KGC+)#pl7#pJF-pX|AH3j~ -EKbL_7U%9J$O(KPzEGeDjSILb0VyO!m6N_dc$Vpm?S{ -uX~$&U7${&Qaxg3^2&&!G9i0<_&5)14+9+dlXHLFPv&VEE#W|Awfpqs*hyMo@P8L9(5-ed6#G+gVS-y -I?E;q@`d>Q`p@#{ASh2+}wPrQ~IPj+XBKa&qMYX90kj1p-e*>Ct3?ur&n?R1hdn^$+0$BKAQA2xG~US2a)CH&Ud$NF(voLrjHH|DE*9p?dEse0d+^f|3frx)M^}9i -Dj1Yp`VyUcI(|S?|^ViYN?ReZO0I7bzK{-WwTJfxEo`ABu((!inL_?Qt>SbD-1C?p2eP&zJIIvot@_5*Q|0|~8?Qo -}Y0WPr>s#)=`nMK0e!a|o&+4H5H^=P8JiiXId#A=Ol;%KT#qq!p2`zKzkHL0mRzYc(Ecv|4K1Tp|1wI -;~1_w-iv_b!+#PJiDdeQlG)d##FUSjh=w!!~6UZ636BZ=%8+$S2Ay@!+_LF)v<~On7LS64e1=2KD^Ab -dkPe`Xrny`?L(h+2v!*CgSzsXhU!o3#Pc06m1R}gG)Vj$zf0VO?D%+*G1%?oS -0x)59Jwn*LRuv2r1CywgYsgOLMa@H`stsg`5amtRq6fM|2C!p}*=&MkJWPUBed-k{%C}3#4L!L^TEh( -%-)3w`n@>*;)J6kSM@uyWL!A+euy@lPy5>Z<72VvA@EAkCbL9t<%boRd9584xNHt-b8VX_Or=UK3%TS -^6h0Jl^kF|#VSt_a?h)J*3nZ1FNAkc9;WQji?wE#yhEV_sbZBPP&d_E{rn)bhZBa4jtQ9l*cmbL>}jj -(rlmv9fF%>{${cO^|(P^5BZOklKK3lQGs}2M@P`m!>*xdI1s)HHVZWLAZqrH96pdRZ&5dFpdS&Dlu_< -I8bMd1Px>jaDw7KP=$AeU{kEHEZ+*qJZUfym0l4rZ1mdu$FWXlA}R`k@?jAaki*oT9LB`mHtE+yz_L@ -b5CJ2)bz!EX(S*4yrW+k!N#fPE&!Q~qjb;Q0sz>=dJp~eX#z|;)@S4M`B-j52Tt~x7^a1d=&o}yR`6! -gHR6Kx|Kp;?=eVGr^Gx2iT&>;{qIrS>cs=sm_s -fKO)7u9MvF%<5*k13)N({k?%}>|D#q2`W3MUFCFp!|$s -cQ{wquVNVUm7bT<($R^$)=(7qA8(R!)cW#>1=Lo -V6tNLvEC=1V!TTP!He_;SS>Pb?h;gpYW(D)ng3ySNeID1GMF=S?^bK& -B2veLf!!icC*`xTYJ&n7g>le9~T_97H{z`1vJCo^+L+3DN&~()_QpOs*!&0VPYCI^Sg;OX;y&M -Ij2pUk6U?^9dL9@`2p4+^tq&&3n6Jo8i+-BR4vr83Ph3$S!kUE-%vo~Rz(d|tZYSbB!@--g-TBQg-M3 -kfE42bG=QjrSb&xHmDN~9^Q%^#V&Gq-tf;~@VduAdsWOZ2PIe3-(LD~Ra6hH(pz8z1;rhn}GxnbFkmn -yCog5Z*#RXG3{BD&|?IdfJPU@G8*+~9bPK{$t6W2TMGgCS+f!YkSh)i;Uj&Hj+4ureyT<%%|#EtAl4V -b238i)El#|?NE#lyB0`r06|Qvh_-I5J2$DS<38$7%a`2V8%dv$?Nb@5I-BHBWqfVcoMW`U=2NtHpHqf -OSOh+;UK#ld1OI{t&j@I6IRQ64OhOu5i*63z~Kv&iNW|B4=FmYy_GjS&K45?sZf}cv!$JjA$ZX$fy7u -AXs8FSl~yc$&tBllnJVP0h}bExXc1p+{at&3`s53^W&w{zH#NG8BDlb!>L;}E;3E%wwmd#e7@Fn|gBr&9(W#%?sz@;I)Ng(3F+zo<8DwhnJ|5w_Iq4{6c=q~cA*bM^BtyRnO%d$( -Edl<+V&NBMG_z^bCBLS7Ilsx-hdnyrmwCDvra*_@uk4P=sTiJck9?P||kJ8=MYpe9krDvjQ~gnc*}Iu -m*S_2+hu40Ih^1En5W--yvr;YIh!XiUU#Ro1|e3t5W&!LIZ1wO6mqLUdSXzHC7JP4LR@e7|LX<$$xI8 -l?G2G?jho4EXbn$9VF{cLmMz6nHiHvm^@w=E6tI6X24&E9JQqw~gkf<_EK%jMYUu~tWQG<2z -XBY@gw><&kgNS_FxMn2vl-!m&-lSfW-S~{$-D1d)Nv+nBR -PFjI}x7lI>tGQK+UUC&qw}(h;-9c-MX9n9s@fqHC6ytn2=aO=Dnh!H_0U4yM*qHHruOM6;;Xx_ -Naj*M|Gqi#*Dq~2iAI-G!=Va{#;{{1rfq6k-O4U0?KFD>MLm4w&7zt6WwTChzsOoVYiipo8rD&F{>`v -@-R1YuA#Vgv3zT~Mcf{~T14vz$#@UcfqqfqX4j7hte0BIV%NGs$nBV2o)Y*IcwI;#8a_~TsshM;S9O6 -!q+pj%;LSpMjwrmV`1cm@9`_;I^Ku-M)|K7&Nw0B`Q<-4EmrD^zjM%(Z;sBv<^kk;X=U-R%atbO>hYV -AGU2~*P^?6?{)=oZ&ab=!XGRJnh|O3mw|H>lY(NCx&8o}}UtXG~Baw3nnn0awXFY7Io9E+-vi7$p-B% -#nRr4oC6QGsqyW&62sYC{$L~`2o%b1E@#oYBAU@XuKh1*h59*UBg&qWhwMEc+pea{R5`@8{B)o&Fv$kYpoOoIxc20&QO=j;6rW9T)gmsY+c#5 -dm|%K*Vkfl*ss$e#5R@q9Y8s7xc2&qs667Ij2q%frhTq%an>Bzdi*fQ7a{+r-l|c~5)~jk}_^RxO`Dx -_@fp{MNn2~-u0Wwn52g{ix# -}9mUt7lKe#RbJ4>9?<_t}H(mTS0iDK^EpGxNLG|1$(#z;9l~+4NgirTn0_E1Vr49g!>2_CnEsJ;dDbv -X>63M{$Z^Z1_zh@6aWAK2ms3fSz9cDx5@4X006@f001Wd003}la4%nW -Wo~3|axY_HV`yb#Z*FvQZ)`7SX>4V8a$#_AWpXZXdA(OnZ{s);z57=X$ssnzRiYn-chH~; -vjy7wkmpR^CojxI-+$@r7W$p)HzEXYPv8uD{>*XVphvv=9esPwK@%VpWd*3)vSOx-vimA2MJ9gVYLH0 -qw!wQiDqq`E0VVsHIf*HkyrKmRmUcXl^b3td%u1^baHhX(VJ! -be~>F*aFV891=V`uXYU;R}TZ{IGVES2_d%Tj~6n3H<7H)4*3jxwyEH@5dSzYMWju?OkgNrB&G|xB8Rv -sJ0uEZ`9VnDiSGrA-{_o_Q8`bl(Eq%El{hrEXhw1r~<3Vs<_cnrJw|rS1u6jW5S|W%G*fTD(@)bdE8( -P30UrIgmXc;GT|ijY$R{*fS_aaW(AA@+N2uqUJW(?9bhQ9LB85CPhpU}>6Sljgb)X)4Om}Oy&@kH1FZ -_zBK`>?FusNwhN6kSA5$GI5>>BrCNr -YQSv*K{UjE_jaXd}AF;C@h*vulCVLHr%A4yrI7{yO|N{3ATiR+{5K4z8wXYM7XK6}qEC0AM=fW1wR8{ -dx*7%Z%9%R|E=_zke*cqeW4g4{@ -R9=qeosQ^guX2bhXtx2%aLu@Z%1r%&Bx>BME|9%e8~gxAEphN9u!s=GU8VtOK`Y{kvyIXdAIVV1RAal -Hho4WD{+BFn4>yd+#AhsZ&U@!ck!T(W?nyC4wpQgnML#+IxTf^Ly!t(Sc`*nu;m^J%Zu9Eryp -`Mx{HP;GQb$_En;&(Wx7&YpA45A*Bm!sQ`rP=uY$oL^tRyGB6*I)r^quabM*q_Rz{t4keSJu`KSrphnQQR?3O?<6kOucZEL1Ivw%TuK(J4=S-)KkI&Z{|iQvqEK9|q4o8e1; -!G}vh4H>KIke?FYdzg81G@(L7_e=6GO(w_oz%66kd%d>h@OM49ybH`?`s#;k)9jWk7&p>%fogdZdLGlERF3k_>?%|^p2cF!W^Y#Z4kZvhk8h^xX7cwzPf$#I{Kkks>~&9sIK*uzLLUJpHcZ13D+T&S~;IZI(=S89)%Hqks_E{E0;g_JJf#HRIa=B$-|4QL~>dkN- -Xo08A@hgnT~rvDCRI;F44Bp&DhEnnR=uc*?Qpb3iO}TAe8U&s^1(~pk1Hi*C$OEZHvb)mj<}R!}4Tt?>b{H6z&`Y&Iok*-s$45>e(%odtl4L@qCrYEF3Yea#U$8#P6-3*Z8Y<@iBd<)FzQY?@EVoprdL4r= -6x~pwYuw;)Y~2vmu}}Sjhpb}(;aV|I4+nKzzyLGpmp(Bb=vCxc>m3h@kVYkk%GZF^)fH%{PNzEVCjq` -|xEJ!LFmf1N?=ihnw$79;S#DdpdV$_6rnB2$f_gzXmpuG@>oH!8neXIG?H+X2$@~n+Ah`VLngt{FdPU -GBEziXLzln4YgyYS@z%3UuB|?Ym%~u`}X4uVGD!%odfi<_5cOoyT`M2CI{_ua;i_`G7zN*Y>e{BpYmE -;D>wb+G}DT^Zgg^QY%geKb#iHQbZKLAE^v9RSZ#0HHW2>qU%@#rYJ0Z)5~sz4 -Wk8lDO|Ycx8V6kw1OhElF&9eINGgug^}p{(k$Q1ni`8E&N_WTSUU<(*Tj=cL+rNf~l#zX&tp&YWsyVq -;KAMgvJLo{r5S?HldD_})Z56^NG$g4c!{dmgipG=(AMIs0T$16`8@FP?6BGn-u5v+w0MR7lLLr={Twz -6dDqAgkFT{+b5f*N(XhI782rHsjPTQ>qMkXRm;o|a+r&fg|m@+R{lu4%*bCF=Bbq6Q|FUZDapzaD?wa -#dI<<#b1J{|=7M+f_V9KU}zKzr$WfKG9yMF4u-eD1dp{EU^1W?pau9&zF2f-$<-b|#PJ^N*j1%)oonY -R!*N4}!zfIod_-4mi?@P-lTUUr^`B@0|LbGxY9s-fp#85s4AG%_syC$BKxq@!YQmLHprX@%k6Nhlqe- -{ntao-^xb?P`!eu=8uc=p_lSi*ToZp81{^`Adtkgkp3B&9W5%S7$_br363;2RpBr@OC|ienJOJ!mw8V -MUnQVxu2h~e((UznuuI6Wl<0RX$q~r~m5PSe1=xOtMGEQYO2Xm+*e@dMH{$Ft*q9W``qyr^%j-`CG>m -RAqY;|i-a6Z;a7#bP32Z7uK^OYIS`w&SDEHC`j{eT2O5iw1uT|cigPcQ#s6?6IH#8yd#A}B#W5%I6zs=z}3Ve9APS-(+5v&c&1(gF=Y8}c@i<;? -b~K+G|j>NS*X)!TXr6>_*tu}!WEp2b=zWUo -9NQ?UMQQvw%k(EYj;!pqOEVYFPlkqj_)?9?p_{C((6KiY|Z_Vv|;|N0U}g*%$Ye7C`H -1BA9T^FyfMpmRYkXd02*K|*DC*)ZM2TRP*Sb7={6gAeWFN0d$b7=}Dmf-_^Ho#hMY)c3l8RVi!kps>4 -VqV#;x_4~jz{XSUxCnr)#O~b%g*3L_pD&RdkLz(nBTJCVDbQ3t -%q8xSP%ylbu=)%4#{mQl0Wh&*{;V)pBeIO4w%Vs?3oiW&$o?5RL&RTHx>+@zfX}I_t(tF5QTp=Rbyy{S=x5e7mN=1g}N3=U&>8^H@kipk8f{vm*k`7Pj-97P;}Yl>G(jm8XDN)H22=OqvcNFs7W -7>B!F!tW?!*Z?Bw@ce8P@lcLFlf_8qbfnLcCWynmKkKfr+or0$9ty$$CmBhYtU@wMiCPhHJ%Wa>c*xm -EtB+E71b;5YM-X2x%(AWRU*4E4&Qk;2n$)8mqX|`oClHnCY`7E~XV#RD4oy|^Vpi9gjdyB`$+*6#iS0N8-(JujzMsx~_s&+yKmfEYuk5H!rNMQyY` -L4WPGcx@dd+T8(*{zH9`E-*RSo|o+ZfzMA{`lA5j@EP`JWX}LbAQHszK0zcwoeb#<5D!F{yEPc+cGiu1YJ0I$5Sf%7D?h_##m5lF+pHet<>@4yAta>!a=0<8iO -y@dGA)-L2z(*87T^?T3T=CZ8f_6E--mu?d~~R#JTK@vahc*Rgf)NltztA6pEj%KigTO9KQH000080Lu -VbTL1t600IC20000003iSX0B~t=FJE?LZe(wAFJob2Xk}w>Zgg^QY%g$mE_8WtWn=>YP)h>@6aWAK2m -s3fSzE2?V88DM001i#001Qb003}la4%nWWo~3|axY_HV`yb#Z*FvQZ)`7fWpZg@Y-xIBE^v9xSZ#0HH -W2>qUvX1D#0nIlufQO{1G+Tbf^}`s2JHqTaA}E(*~pYeQg*#A|9y9)UMS1yioF2K7u)3V?s)gyJ;xlu -+so&_oW2kRKP#)95;rSz#Xp$@9zA&Y6wZ}M;U!bLl#`>$krRF?QZ6+=y~wyUA{SgG@T_9#il07;A53y -pu3@prYf~$}Sb$hpr82;zEDbZFlzK90)k2Q#@kaU3=VSu-cdrkvl1s%arg-+otxqPCv|w7pTdo8v#P1ket4MQ|aa5u-1`uWtJP9UtBjY(_^ -}5ni`_su5dDQ)CEGGM=dVVHRaLPoIV$3=lwSO#Xrdvoxf@2O#hE}^b -6X81r9w!*AXzvNrUv6f)PWL9tmriA6b5{<|MffbP=4wyzoWqq@vl`Z~~9J6eli3jJb>$BcDl9*;ZEfE -)KDGk=*>;?MA!^5##f-g36o%vvRv+eG+{jY#IYtG)tx?J`Z#Rt&$+w+p)W(0(a;mBEQbpmDxFD8D;xJpJQ#%*$knMkj7*byRo_kEMyDMPjABn=yMZ7j -?LmWf(<5rgsoT@zSh(5E*fjp*uQF0*d1~!2Sl4n^<|r-1Hza}ZDq+nKf$;f9*`eq-M2B^86Lfx^xwL# -M{tG*Z9jyvRr(<*)5k -iTO)NVyBNym=5NKpheVqq!dg8I?yN?{Z{+eOK;f536hRcCxaYzu1F3C2B?fC_M6@dv8cyb_!XCHsKWy -OX>z@fB5Pn_;jq{|nKr3+V2t{ddeEsHC>CwQ`THnJX~@lzb`tGT|6gD#2!eN*vwSyd%zz0?aoms@q+R -!BG&2;?45^)#fr)bYk*6m1P52P5c#06AI(5^{@yRi~BfG9m;c$Y`UUPbi?mm_mTuY9S6u$=BIMhfx{+SiC)#l>c=GDbw -950t|gn_5130Jq$(__@+>t09O8E5D)Uhgl755Ac>sezsbmpgn#Nq7{)v+lc{swg_(-VY(Z(k!V#EuF3md)odBuRZ>lPQkR6!giKM1 -wIy60&CG~N?F}>O14CbM0fX;lm_>YVBZ4_Q$BvWYkSl&Xd7B^(R -nE0|XQR000O8%K%wh*gSt_sX*jo^)N>u^&IhvgMC>Tl`|_-gA!by*l=cZ -XR!*R$sY*Sd|}|?+V<)JBIY~;?rlk);y*6OdBQ1jLaPPo~IxUtnM<9$a4-lq}L@&3s_wS7bMqePm?6C -u+oqul1+vo`j1~g}a|MD$t+JGj6WRMT4n%9a3j0ydaf{oqUyIt0~T0HCqL4VT+C4olutA|Md+0M$mDk -W@otE(pKE3#JXikZ^!-G{@rC(ra_O;W)Q^;^rr%V^|3rl^f6|H^BeImxYCWv>E~Jz$|S%kRF17lX&F1 -zlYo%{J`BEYSOUzM$6FjRTF-0b#&}j*GY4;oe$)>Kp43#}wxPU(Is3b(b3T)*Ya`*Qi@|X@8lKU-iUC -w-=2AqdNK%kB%Y^yzpKQ_NtONKPwQ+@cjI2*|t0lF(XV3(qMm)vc{4>P)h>@6aWAK2ms3fSz9I0PW-a -~001um001ih003}la4%nWWo~3|axY_HV`yb#Z*FvQZ)`7PVPj}zb1z?CX>MtBUtcb8d5wb}e}FnYyAa9cQNVjI)tl(=Q>bo{8!;1GV09gja`77 -{DIAW*u?v%KKaXkIPt6_y+GfsYeKtSq?z{jO>ihw>*mnP)h>@6aWAK2ms3fSzEyRfRL#L -000#b001oj003}la4%nWWo~3|axY_HV`yb#Z*FvQZ)`7PVPj}zb1!CTY-L|#VPj}zE^v9RSKn{kHVl5 -hze2mb*xSX)vb6&a$gnj_yI^TrG#R!c2n?M~pS9SQA<4-d&_8~pB>%iinuq$umMBu>N9q%Oz^*_3@$c -bVQNWq3w@R#Ya|0W*WGBZzzGCM}B{&*gr{ -ds=&*TvO4x2WcZ_Gwpr#No_KL>}-b2b0sT4J$gBYrKB2fh=&W0C2G3E5$u?!qz%VcA8yWQ)U;$}O3-6L%wxsqf)vq_{xO -2gn2tjBqw8$*8MDLaF^8FY(og3;fqa?5mOF{&QywFZW+#@hHXp!1?)`rlXgXU~-vu#|x?i%>)~JdTep -xyT?-jvTW57X{2+0a`JPqiM(1EMmcO!u?B5|aByg|DzL|tRlI~JY#;aWFTK=szhOBz>7I*!R}p0Jk+5kJ~+Ld)QW61KP7mk01Hgx -=VngO87LXJR%@WB7Qce1)HnYJS23pe3x3h^I0A#;At?Zbk7rXhOeY~E1N3a!xmnj{OyC&U?wEqMY%xo -FB(jrjdX##)QFOHwELyMqDF-J6rVnei<^B^!hi?D?mVnrCv*-{?!1TS1osZ^*>p%D*F1w_%d8#Cc%(` -UNZ)ILSGcts?NvTt@A*A2jm?eI?N|jiSay1NHTg)}8uK#&4f=>tTB1c1>cFp9jM!x1EXVTx^hPy9TC+ -7XVHRs03x00r=HWQ;2kajkwQLmQiD`IYnQXKs$Wq4(A^P5SVo`ao>6=_U^=^)7c&DO$a4ubBlVBd}WT -L`5V832cNKJZU4HD#tN-)5w<>H-{3hII`#MUYfUc1GO4m)6i2>h!y?7CAr_`;&4uL3gc&9911 -eU@;9;UQL&&$p=J+-}2eACz6<-w%=Tl7tB*tLT8eWQ3Ps0YN_=ya$*OSjbPXsmT=7c0CHO3RX3x?me& -_^JTgIu=~5T~(#>u_`1_b=2+BrC@Ic_JHot_N#Fnvg17LVjQ;Q&&0d8bDzZi-iCAtn9M|A8}&Q@9368 -u?IBtqt)@+aza!RG4gB6?gsl}{!yW+eK8eZhJ_s&8+*&fbd(p*WPdu6GnQEuT+AD8z$$X=`V_zT(i`v -ty{K80fF9ni^->Q(Fkvd1H)@YkX6|s>qB|9cI32?&4*>$eVzGlWP6(;_ZgO08ZTJL&a%S&v96=^#s68 -`<^GIGJ$-Yb>ZUCvHI@D`EFFg=r@6aWAK2ms3fS -zEmI_qRp@003$Q001rk003}la4%nWWo~3|axY_HV`yb#Z*FvQZ)`7PVPj}zb1!mbWNC9>V_{=xWiD`e -eUwpe(=ZUn-}5Ps_9c;8YfR!L(vYAOHZe^T3<-o#>lv^IlSj?Z` -#qummBIWi+;Ck{6P^$%LL^+Yf@GkrkUG$B+~?rQwepq(f1Y6+a<|QaCW4xE*@5SIXO#8;p236ETg)ST -2DodKL!W~(#HogJidord_-a{4D&p4iNe_%+e^jmT(O;K28=_FNUU_9aNmF9lVP=DgY)mxvvDi|{p85d -6TWNoR1pk+f1dQ*Pf}WJAEoXaH!T;>;mvGo<$*!XEBFM({3N=X$H^%;z7}c50| -*|X_g&nlE)Vs+Y^W0b@JgVa2k#}cDqhoPSGjGH>%VqT}wHlolGW|s2g-pTNgva5KFCrcqEy`LuNpRbC -dTxO#;4oghxf|fl!_U8-8~re*sWS0|XQR000O8%K%wh6ND==J^}y$-2(ss8vpU62+TEpr??z~^IeMIyHj>XiJ#aEPa#w+br#UGBVspLt(IUa%fzDmC4VC#_KFn -W+t5KS?7HRY&{0mkC%{+JX>YjK9Fr$Pe2fpw=HC`~uw10`yVV?K%OB|tbdkM*R(sh~mm^#=>*%P^X1| -#tkn8Fb`c3DfxH*g;?)RQrTAoF~q1Z||mtzTwf(>b0^Lo5+zl0qlt=TNO4diXs(wlHcE<^a`+Rh&?i= -z>)W_|?G2<(Ht3Amu*R!Y7xoZnsjm*3aZlQS4Rq^sBJKc2~6bWIG$Xwdc!Jc4cm4Z*nhVWpZ?BW@#^9 -UukY>bYEXCaCtS#!3uyN3;@u(UomzTF~MVB5-}nc3F=}>{eB^Ie(z8z;C4}lR4&mDyVTCLHw(|n%El0 -FITOJQ3;u`}a4@c)->fZ39S4IC(M{g@h~@=QO9KQH000080LuVbTQ4B3;hq2h0RI3003QGV0B~t=FJE -?LZe(wAFJonLbZKU3FJE76VQFq(UoLQYO^(40!!QU%@1DZ)DV7osP^pJ*& -)eoBZ{0AQK%HU|EcxnK3c)C}f%8Tatn`D?+D>8R8l3m598Si?SglH^2hLsRy@4mgzt&#gg%40m0|XQR -000O8%K%whCe=C#5~CCJ}?gwAZEcE^C&<<0>Q6u(9X=t%-VUX&bjwme`UG%Rs}&2pupGi-N9D$ZhA0n?CmJ+4 -@orU&=P_?tOO;A8GO>n!nuBEWM{~obLZu{QrG{K@k7)(@?<}Y|G&x|&AxREMdH}&g3>}c)*hW|Vwfou~ -X@B@k*(DmxkA;u30b_mxY&;u47z;rvB5(jwwM;?gvddVThwnhlw&Q0;{AA}yD@Bl{#I6g%2fmrV-J)k -l0f42V^UX-;jEN|MfAhWd1zgq!SmigLdSsJYJQ9Dh>jTB -IR|X9|@CTvvSw740y&<}!3=bB_B=Bn65QW03jg>Cfm(w`t_#_+YD2f+RE6M6Ed|7YzhXSFA}`igN@t4 -=?MFhgA`xOzTgPh;A*53MNZR6I|UnSdRrZpQ#M^o&oz(J>V#+z0Apz%K5qJVPM?4OTtdMwRS#L{6l4j -cW3&g)_)vltO%w+nb#U1Zp}WpHkE%d`v^vJf3bV+N$YK0_}6IVI27h629H{iDi`Nsyzeyv}Z&Z4tE_A^8%EJ#TZ{3Fx}RTB~P5;E5NY}kXlQapK3+r8=!3r#(d92s`9Y^Fq{ -c*kRg0_drsV>=UOFxH#7?`o$$OsTKm+ai7@EjVT(bbF@tPOGO1Wjswd~le*D@~4e@|#NrbNiU>hxLLf*4sg5C|Z-sp3`}hy{^4b4?u^G%keR`kZh -^`-apT?&(tn6Py*?oy#`K{mON*U!n3=nR2hR|oMcUpY%MTVmvAw!QE-=hTqG)qYt<%4rYre5M^8Z{>4b!8AcaH}D}npK$&(bm=nY -CUgbBF}cKCxQzR$yCZy4xP!efu)TLT$PYl_HSW)F0{ld3jA!SD$w~k|(3}1OucYC7p -Uz7d2!_y36wz$En<5x~*|LLG=`qW~Fsh{8j;R5FB?57HK*B?|qVy$Qx2M -<(oh~I}$Y$FBP$~y1^L2S%~L=Gf%$kL7MtTZ7;4^(;}@xLMh!ed$OHMaXk1X#wX-^Tq1+b*oOeKPHzN -q|v^E%Qq6h3j!2DK@n`2#lQN8TB!2Qcf?NlcDklvfT2bB8Cgx~LLVxQuin!ZF7l$1)q;d-N(1HXk1QwB8ZZT^wFqkEn_$?2BX{+T -k9<6u|43tGe^+W{mgGX1GgI2ThNl-QONpM}9nxB7G;QB%jFC9%T@&Gi|@q29qbPI6)!?xZh(o>;(Mgc -pREvu1_~&8ZJFsl2FOd=@J;JTh?P%kwM5ly$OY=G<=D7?tB$+(C&d9Z7$@q-CdMYr7rpy66S76%$P18 -s-B8-6F3i}%Pvr)8wQaK{9a0tEaG=^y$3HB2&f0$g#4QfD%^mGNLt2969?gmGtN|dq9WloTk!6_OE2> -S$M=XZjL~8_cg)-q*M46DPCVhx^pde0r{>6nndWrmOoU@5UMq)OG)`S|jI+nkc)`d(O7dza11o -&FT!8aK?DofVC`Ft(Sek=?CLE;@1FPaJ-ve9HNJvWNBAL4v#v%}wUZPp&_L$FISXj@2{h}8f(G7xmxi -!Su>h9i->}F3YZfD7xYu`?>qE(mmOs_39D!-X(o*oM@>e~zENI*BoCQ0UHmKb^5Rzto^ZM0t+*Q&;*qrYX;BQ@As!qI(4-NV577x -1=J|~!1PjyTpLG7-7Z;SCkKYCc_|?Mvs|E)6)xvx?Fx_uW+e%wk)5oDlp@uVnX`bD^6_V5BjKR*gP5; -bm>tdT&U*#fK3IqgwNGHX-Ih@EQZh0xF@ -B73Y-lb8Lup{7k=VKKr9u&soDEQg3vHT80wjXS57ixeInc-VBUdSqpK%kwdg{vXy5g8>c~6PLmz*?2v* -h?fR6KLD{CaKl;Sx$-$=<864f={eg_=^=p*DYP9zu)m^ZV!Qg;jO_hE%VPik1u -a{Cp=EC>ru|IIY#P=$^S_;zss92k<5a)}WzSWX>s343c1rAsIL(ZuIR??W6S^FPCn%^d_PM_2ym16sY -Z33G44a{r=UmY-09y2fT_g$G-HZlu!m#SVE8FjDOoDwfGpC#E|(RDuPOuH4+fdqVJ1$GHO% -hT*PWR!7_IdRjPhxbf6>WHj#dGk+}R8&q5{$P#C@T|BBA;0D&CYOcu%WN7hIn{YIG8! -khauu=2lerT-bsWq_Z%deX%=D0)Y~UJ9IB|W1A)#xJmDsiFaYcq<6z*aga)TJ*7}f5plu)O>$I41^)R -EW=ner56J1bHkdSrnhqJauuTOGBOnlx?K)C=cQ(7RKmbWeZJ5wI0BEW8HsNo3AB(p2w6U9#C3=hL($z1}A8H -17m!?%6VnjeJGMiy7*(`q0PUQB3xRTo(lDj#;IL=-mJF^^zBPsE~;h2OBQ`9nyUN5VPs0eJtSDwsP6+QZ&b?%&NZ7z( -U8pFhXm^G1~6c(^;K2VwXL&Ppr!&#^6zNf3^w%9zvP2A=9NM-78+O_p#AK;rlYgqKM+B)G?eWKUXI=e -m_+Q#Ah#4q@CRz?aA+}o^z0P&1_c+jH9o#o(!GgiV$kTBco5R^M}T3+|QESL9o#F-fjraG*|S1oldj2 -dQXrB@(n>rU0S^pi@2yJDT36A#3d4&A^mEpkXE`5xtMwv}4exl6UT!^ -EceOJn;6Cbg?3O3=6@LMo&+t=7*z%HOPt+xknEb>gyeZemj7DS|%}{b{pVB_RC)R$}DJ!a@mSLmnJg5 -Oi05vJqvsZ`bn7P}f@*r*x-;N{0i1|BSlU4x{cQYvlD#%$7=X}@i#H)ENJ1H|&1s{M#x;*o}H~Ce2j) -bhVwWG;S7<6Qw;kXAwV=o<({XkT(h=x|~8M&FpY!}N(c}63!0=yXl4doR0uUd;Ks2@c|8g&Ld5~q6>V -T?#zXmjuid3a>@z<2cn^`8&tc03Jc`ASz@P@23ErfPw@wt>*o>fYJf3nfB3y59nRBq#MG0J3 -Zi`TY)#NG85qWhm7>>C6ACjiZ&h^l5g1&{*)|(*qdRK+ix)0Xm|2vu2FP|SB@KPp-wb9B-tUt-Xl?)wz-m86! -{HVM$!WXzWZSZaGgndAhS&X#p7*mQG7_lO}d1jZRjnexzgcqlUosNyYR~K*FGypUN4DL2Mpi3A;}gIM -c@PWm%&W#tzq`xHB5ukvm7%s-yHEX=5eIBr>B*seiGbV-R)oOwCgOqt1{|0Gp(9p(}vCljK({>_yeUq -t&|o4-od-AllSiQh8)#0COQAcX&-#(CA|T=^J>1C7}22k>ee#f@ISKcNh)GZ<+gYWKdBu3{_(c9h5xy -&VwxXk4g5RP{&Cj8zccM`v(9tA1b#!ynGagdZ?t@i&pf!|qOXjOzMMpfB$9l0;jKO-VnFFF`8L|CwWP -75P~_aXq~TlbHnvinPu}Y&TgUP-XKX%m0&+Xo$N7jMjeX%n{y6$);L6ExGAt=#&t3PVZo65UovW0!Gk -1}8#}}NTexS*g2kkMB$)S2>9_L4IB(kW<3$-r0E}J#;fou693nX@vJ$>k5ki-8+|( -Z|EMNop5C>1va0g5EJ6dOG0Dr&VuzV@CE{&n;m@X_bl=X^L=AR#vT_vd@3?gO3rkyG5}4BmA|ZRAdP< -LbRT*izr7blMV~*Se(Fbr(5AVKRQ{;x0SsdbIDQ-{<#fs`{^9t~wx7j>gb|rr{uKH9Q(xeR&Yx40buw -L!lgRG$Aa!kyk-lzDdnCR_zWOv-An}gRVmGcB+NSPZ&lVMRj;C5{U?%44myc;Gm4l7oDKQk9x6a)4-m -AUd`A5;_zMTQtRFjP!qDndCqYW=dxkWB}vPAoomv&7Bw8= -mF1v0=Er)}8-n11x$pR7M`P2p1b5QrO!G-1Wahu<6xelOq=n9$?Mv)W-gD8j)D1|{J0#i7$LMM -u$2#S9Ongn%7(L+eRGj`sAHe|&ke1p9h+_G;{_??CmZ%72EpMvIwgXAUyt*7Ge2-`=HEqbG8f*Qfdnyu2(0>L%hN|z5(b8(q!^R()ZB_-E^&Ws_V43NM*8hOV-ZVr -bVrNko5-9?Ji&vfhIjk-_a-e5q+eu{@nK* -6~M56kfv<(iTG5n!T75P8q=m4teO#C-#YI^;6wJk~KhtR`d^B)OJ_14FK_rWS4g2C77v8UfuGyz; -aXSjL$v@)Lz5v&x^A2~9uVI=9T1DhY;MiQ;Yg1?;7+tIc=D7`tyOEVqrl1R_G`KXbarUQXH|4st -iBEG0i88xw?C+0ke_G&>KNKk6Hp|NRM7uKQt7v)$Ayz5hQ3_=pO)zu>0i2?)S6=7s+O&7+<|Lkf>eE? -v6=lncOI)TXuXc`|{{q2m>!pyW&>&YGxB>z=o(HY4F@yQnP(bC3Z!0fw}47Fz%K}bpj@ny!o1{_x((a -cbf%I&W%VI_0g%2zDI#P@rcUJ_YZUIEUPSIt$Q&Vf(%78+jOrpc=fO&=h89Ze#-X>2W2G=Tdx)Ttmnd=sJ`1F+jyI`6P7o_s*a71wM~SZIgM%8s5wNOFr%%5a0 -5YHJt72p4(7*spW1q{5?P1ffeu{&*UQQN1KtL>hu)hIBo6r%C5i|U8?xV7K<~>*Fab63s>GIG9G)q-@ -|7Yqmvu}aed#qgF+^rjzVX!x}Nbfo<$Ea%KU5OfgZ+GnY?0Ik68U~?Iqv6n4#qz_&Can0G!*#smst`D -W81fvXRsR%x9wXuIKvV;e*M<=m~5?)Nyvldks4zi92+JoSqf$g3SQ<7Vkrb^+>sC&^zYd@1CPz+C40{ -%+oFgv=@eQr0VA-Zrp{V#o4ORWa$wwAs>-Z1MK=W5iCrx1SO0Ri`}yG{YvxJJuc-}pMw_^Jf794hXT) -y67=5yN~z3?563QSYI3`0;51PL64z||3@Ac}xd0s#pGMFBi*oLSpKCWIV8GZ_+Shi)iq$|R$-W{?*FGg)h7A9Nr&5D9lxaGC3qp -m(Aw1O?PL0S^8Q1`o3+7)D>4URK(OY@DP%|%A>%_Sv>4Kr7Fc^xGMH#kP$`8i0m(Sa3vS4T4ew+)nvExz4n{rGd|2Q!+#f$4%6{lL}gZfgf@Bo -+{wJK#E3il>wrLDF@jtE08@WtzrMbCoT{*8zpAfa+NOy6C1(H547O?eduA~IBPKRafB8WJeW}I@s|aw -qzB^T0r_)s%Bb&dg>~|5g{OK8NpTB#fA1X9py8d@!VEmAv`j`2sYNTS9^A*@p7J*sdL%l^A=JWeG1{* -61UdQ3=Hk{dZt7A6^RNvR^>vc-!ieml#*f5@bw-LqmfV1`h^`m)Qc(BRGn>(^L -iomx;PhSzY%vk^)kRqX`xfpunMmNFScOZY&>~g?PjgiX%VHxc}88sASD!)33jJ1a=h99_zJVhtEO^J{li!H_A!I+V`F`atZiiPnQ --7z2RYBom!)BUof3xQPN)$KSXC^>Y;8&}@htIdvZkTkx{be5AA5Ewf2kOTbiVxBO1a_kbd$ZrX%TD7v -TiOWASNj?B|%O(-UZ)qprvQB_+X(vC)@+W>b!8Nx#Q4*b4zEX5D{iLdZ9XXZx7(kv3Uu8lX_@|veNgM -sh8l__~4=;$8#z4Ewa*+G7J4=!5Wa#ml-o{rxCpB%jqa0Vpg=SI=RtntnCwSs9rAG -81Gd;xx5W`?OrdlTVk*$(Y28a713``gqlN_kxku*iJCuFC}Fj%{|8YcXqHYp@H<3exi9HMQ6#kYWFf* -V(X7=$9faRLoEG*_ZA+rK~jrs3XJ@4VXYlFzc4OTYe^v`>vGO+a@2HuKauTem+oaIzQSjX;M1F8C2hx -L3oR>@tED}#)9WXXYY5Iz$G?rT)>ICIRp#v%v1^hI0RPVyaZ3aJ;{YxGf}(%-oNp-kdHd&okiI6K$ZitA0TM@|mZ$ -qDT_%{k}X`a~I|*9mdIP=jl*~H!ulKAl -wX|OAQwlsbk*%MbjaYX9lsq(rWHSa144y9J!pwPLzj9tefIi&k};skzqn;~0yj$xX)8dW -E9s7ZsB~C8u0}(`i?%p$(*jYsdWrC>erF4?U~Vq*ABrj`Ln@aqDyKppis~+@mdKY7z%eZTze`?0{=`OA6wzO$cMH`H9AJZ~=j5Nd -;{dnKOlMN%jV%g;A5a8uUcu#2KXMmamT2fQZ%0$=R{L#~L@*-b}Q{Lb$@nypMS7JJmFj>MB{%E#V~=h -NiI8Fy70jshvS>lgTj?SLq{F_MI_6-z^~qiMzY#>0?2kQ3S_)8jtA_%UwF?DWWYSdk-M+0VPAFKgxK& -c!YU@JJ<(Sg| -NY3x08fhFkIz5i7i^p@q+}u?7A>XrO^&D?2dfqy{Pe{SJC+=%j5Mx!k2P{sFN)${aT#^U^e%bj{Ifz?l^s+1%V?*d-Jm -r2YPeZTcy1Mji>1qpq7=O@eD)R7-oY!n8etVFR?y1I_d^-BeEnL<*S4W-8pChI)q?!E4cu#f6uDc%&s -QTM!$od)xDUB9^Cz6|{Lf|NVQGQjO$D@ggtv-W0}0GgZGI2Rn%vooEFL5g$tmekQddDdI!&-wyeXEXn -rCBc!?*`^65Z}Q?|FrLA0D$Dxa(F2p`Th_$tb;m4i`gtgsV9a$Ra6*(O7O5q?nCqIapYg~0zijXSo2F -Xi@6$=y6bDeuB*~DVks`ZdN##2C6)P$scJyYuxcS;$QYz1fCu+3NpzS+6Vd9x}FoE@jkXHq2A{ii2cs9K2I -r{k7YN06`q5v$B&SM!p9Z+I1>ER8(qIEzOY7*SVU{T*3>LBzT>mfjd+XjAhS9s^IJOeZ7SuDYp;pZpNA3hlOjPfr#KSMO!iuaC#=aSUe|D=VpuUo_;rK#>#pnATI3;+KYOx4kZ>hClh+fjRoJ+9j|Sq+f)h}U<~99v6x?j|5md8U1~$4z -cx(%b>u*^?##CEgfb{^TaWtg4-DLezvAoF=e@OM6EkJ*|zzD=GGsf4CJT%peC*d2nn=m`@-obC@YaQ{ -YDg*Fanc?_cA=Go1>;eli!P?~ALEF;xlox{g%K5%nF!IwFip}%J8OPPDwPf=yP -ACJu2dTU5V`rrp6BM4I4)r9qh%2r2}BDCiISaa9$j{R4m@(yG&)Meuw?WqAibw;ObdK9Lw)bbIhH~`6 -OR*F^Do)UgWR?CXaj>*Ozs^RwwdRg9!1IL?NV=!8Q<;84(K4r=IsZ(J(F0BBe2Ech_?o++J`mi4{O_p -%Y3y(x0T3D|bx#JVJ37jotNmBI$huLiK*NWTeSjF^cF5g;SRkC|}#K0iQWz0c1UD4)T)NB^0=8*t1IP -X{(6}#*{{P(B!JAjk0JbxL4_U|Dz1HwOQs`(Khfs3(}VnftJ_rALVIYr49R+1i;_u{J^htWTnlpFn0V -t!x9U*bNeh&4F(DgoSZcJJwV2=;Iv{w-?3>9D865PV-Y77ycZtVTXFM*`kU5>$B=OB2Lj*$asx3cMtQeMC*?q -z7)&mDTxn4>+1^eK)DFA7G{Vv_t2jvkI!yPX8+DK+D*o>W2R_j0Th#hrM#8^&;?G$4$4Nh;Aw_JlaC} -Q6Lm&`?Fbste7=kF2gis8p2oj?{74=ul%Fs<$*w6;ucF9=b5yv*+0Hro~+@K~rMA=)J_&hp}gd4eLOgn@4TVsgW`6$Ybd+g4+R)U6u!T~dL8n -|1KC9^n^<2%YDL3d#T@_D83Dge>F>;F%LMzT^8kO52{vj -L;orrV8ydk2Gqr9t3TIfI3Ugn*e6Cc0(m3~H+NWltDu}Q^-m^kXt7Gv+D4@AsY?Y8Qx5>ATQa -gchg0%aq{d+*U}>ZJA(OB3K@P6Oaeprlx&|7VTxQ>y=Z6W;mMXf2r{NyPEuw;u%$8vbyA85XI={G8EG -Nx(wG^<&5_L(qwF#uuQ-)q}z`5L>c~`tiJ)?Gyb~bzHUD9k*E)8(z=kIKB5fjHeG8FS5{>(c@?|s->n -Fh{=JaTRAOGxS!kxt1R=0=drA{{0vyxlh*v?*x;}I2bSmduWJam#YOE}A6XLq#l*6m~|(x~J4DAYR&Z`R&B$u#;`tgEBM6mmWjQ3gd}(%uulOK$G4dEim5AxW$ -z&M`XRzMFK^of3g2SkQFy?^t!!}xxwTKUx~18H_#I%bC&2NRh_=FY_7TR;4k5B-Wv`H)zCBwA+UgNEU -oqOk0R;8lGU6Su-|8Ky?JP3hARELsw-Z{&iQdJRf4PqYFKr&t^u1uwuseQz`bvY+!%C@8h_0jN{3&3v -43Pg)#b|=^4+VDDf5SJ^04dD|&GS1 -r@Iw`(ve3=+t_tdjlq|vu>}A%l6}7n%#t|EHb~KuVhqzRQa%sYMF -@A09&b3Tr%1gln=$#OAy^zHR+GgsbNU`l-R!ax<-gnt7bQe-!_V7t!gL%@8yvp{Z;z@YSSvonZ|c43OrHAFy8`^UncCWUjy_Y3z -Gj1;`!c=m{RmHD?17X|=_31pmFN=2>nqpWE4R_>9x3H!tY8IM>vE|uP#mt-h;E+p6L06qDqEu=teQR6*ez7n+?WUPVx;W!XOHJmC&Z#WNB80wUYZAj)a3M`6i)q7^;+SgbXGoBc)Cl{8$x(% -o^YH4niby&E#Wab3!EeLlQ9pAC@9Wxgvz>;LONcfIL^{asDn93aUm#w-?RBJmj7Lw=Lw>pH>n6!G8d^ -c8%pqhk~o-p6H%Ryhgb|5hRg>>+*$jP(pe;&L4^+A5Y?|(ADj~I?Z7(v1$LgEBS;y4B42)W`S1w$A~5 -gh_Z9CMZ+LrWXHJ0AAsC?Lom?@E?3{?TEeiLs&sNOtz#pINKuh$a|ZE>^=HQIJWDB?=1- -4MP2JWU}&2Mgl%bc>zr0{v7Ju*6@;_p2I2W1WmFQFj<>5!#{%Qnw3uFDJvdhMk;(J*Rj%~&#>f|zDSq -9#K^~<7J*;yF{xI!D38B>T{6Su&OJ8_TF$kSUVT0j8O9{OvVwmhl -GHK*;bGnVzJ<<>d}0QSp{3mgHM;uDfDEIf6xe}EGc9?N+V~DhkH^vBe13q-mtP4tPyyo1JW_07JkGc* -iReCJ$&q!NsCS2xinZmmz{<|+27)-*%=22kQRkN$Bq$E!z+$ds8z48s$+AW@$ZwNy5ebK(xUV>BPrd_546hmgewYkecau+E07tLD0V3&9*bqPqN7jG$M -=gVZ5Y0BbtO(tsYZk-;)s#xs-PHL?Y`nbgd%uHHjGZSQ2tiCMT}*T?ftyVcgyULW -S{TQ60IgNJwV7-bc$k|ixHxmhXRFL4lxdHrgw`}eI!t!n@YD1_TCTXH`a2sc&h>20(8+fbn>o@T^%zL --R$IQLbvQH_<+9iha+O!4rVo{t5HRxO)5m+I{-u3w(BqS&YD{9E5o+T}`!4_EG;gqe-};#Z>pd<})I4uaW*9)v}R*~bfSepP^YdYa^~=B16UKZ|?YZ2`X*_c%E`OA7P|7f -o`mc|i=16LGuf{FN{T$xp>HV*(T~SgH0EES$x2@j8iKJ(7Op)fH+zN{T{Ct=yRm(a6h03~A&K*kYbyI -&ZpVmc&;(1Gy2Ob%hSVm$WPH_6)jcD)o_;i;J!=qWu_;Ie5Ifd{z$7b1eUeN?BOg6aC -zVGsz&n855!CoW6S5;xGUc@cLTl^)X({-ly#*8#5I2JGT;waIQqh$rc1^^-8yfX9CNxbc=igb9t&bc; -r9yaW-a)u}-;SN`tCPeafWZmbnLVf=gH|GW8WhR0FWyqU*?j9d{Qk~AaDm2>eavn@EuNd=s%;bCZnmV -XQ&K>U2QTAQji>4`$>IwKe`847a*%|iLc(7(L=F;puT!Csj3}N^Q2g`uW(E5XKXUv#jT;`=kr? -T-zl-CXPu?n2I|vMZX|@_hsivH&6%Z=CB{=f`S1KKc<~wy4udnwMcGU&qk5ustNn6k>CQRKIA^pU%*_ -9P@=Cu0k6FSsY><+o+wbmpB~bfp6Q*^Yln+cd7{Ij*N8$)y?zFSkS8{A*tbv$)|2HZ?OLGd;*eJFg@O -1a6!K@lq^yY$6@lPNh8h>Nk_;Jymi)3j7u{Jfou(VMn06+as!{6$-*b#0UFndjp#e%Ol>u~TG<86*^H -%ee?FbEJ$QyCgOOa`(g{nYi(6=Prwnv=Cq_T+jS28(iR1$w9BXMzG5C^XyGs+Zp2zNQPcXVXcXn|4+X -bd8%`cv9UQFQ^J8jaqyMAH2fBf$+>mzw@8lRqz@;}|*{YP~BuNdqb4*q4|9|003DGUcm5+`8-!61^tF -$_dN45mPAZGZ_JM2Sy>YS0^#wq54K4IROEH42ivw|7t9xQA_sx215b9Q|_#11C0N4otS_6_IW(#QJ-M -VDi21>pDD6Z5b&U_#4QK!41XN)3yj7vSH_jg~@jQ3XN&DZ7G4Tz6bfUyC|b*>epGtpDb==X46ZH9wA@)ry|p0k -$f6mz)0DrV8jU<)-FEwbTiajjtw>>A6sHp8;GTzQs6!_iDEI$Cie2DyQ6*%=%REHA#DIl%k%JHVZQ0KxPFAU2$mW1 -#&2o?>|SSozUeHj95rX7MTxW5dqAnMqL?rNc6hQU7SS|@HR>Ci-#TVYtME)aZw>PuA}DeJ=B>j;|?1< -028w!0+g0@)7_43iF}_1ZZzQ1i%y(=F>eur@&^7!esAs$kIey8v`S-7{|YFw1CQ$W0+_4Elb@Z0de#( -JUvsMY3)gS~yo*SbdEtF=DY{EXIxf7NfVPt?RZ;%8DmmLOyw<8)16ytpbGr#-{8il -2Bm+IrBU-~gk6v4M~0n222Uflp@WkJv^FRUp<~b%yr^y1H3q!8F?WJ_F(~iti;7HfZg#?iy*H9lFxP} -Eie2e4*X}sLy#yrJ3#<>jvtPtPI(6M*-mUzF7O#2Uk0Vutr=%5$2WtC>9_%jlZHg&`2gSnGdfB2=cG`($-F -{mdAV4Hv>ng_}$7fck+OkXTM|2wI3ebLxR=NzbJ&b_NVGm!zFV*|^q4k<7LLgFpZ1Z<#%bI|UO*TFI* -Q|xZ9o2RGDsWYj@7hme_(q5x>>TUt>U>*RR?P -Z%SvK;9~F|O9$b9}l*^-YB-3r_SCF|KWdug!>k -Mb48bb|-#w?t%_dK$B%QL!HP0zwgZm={ARQR6Ip6G++yX1M=MV -BqH!hGg29{D?w$koZr!eiGKJ0zF_HYOsZ3M+5)soh=Ft~h`_nGdLOKG -9}f5a5~jUQluw~ZRDeX-AVDz81&O&fDTp>*?I9V?YQ2iN5nIzQYSo2+>G@1mf3U7Tb2(&h1R+Ij8Q;$k?{EC0N?c8gtAw<`}2_At5KZLha}MU1le-QT!`{zb^j1KI~2 -B{Z1xQmn>N^MEV*#H_L%`l}+Vwb=T7wdnCmK^QXBmFPU{K5%ne|C1G^3A{+Ll-jg_a38*VkosRcAwum -PfTOj$4Ui%IcsJoNl{>`2!JSA>;Wu|9^K2-1FNI6P*(dE|?GP|N3hXz*S(-3O>j$PRv6tD^2-_!#h*- -?zI^v3jp_$g75Kof_avF_qqJK)#WZ4M$Y?A4IK-^H!9gB><6HY`G=Wy7j^#+?f3+avN2irS(&h1iGFZ -&!R^F4<9K3zTHv{(d?f+G^krBXr$8eX+srre81yLC%dABbkmTmzs5IZsSm^{@a;=ltUmeqC5s($HS4jv -kFh)e60E}{ADw`3YDnvgtSLAk(oNi$PDQMx6>pk#o#_VH0ng&;(C~!3*0}T`c`*$4wZP(%eO0es_voYI4h{Ug#m@k@XoJs$!Y8kWz8Avyz|PLV<%7#` -i**(={0a|b3Z2pmvA6*?42s>VlKyx)Z_c063WdYjkjJ$aEGd#xi%wij-4e;Wu%fuvIT;$G)Xjzt91T~ -<95+b(u=8LN6jwylVj0`zoLal2v?Nh%J+gCrPCwj<_e2av)ufd=KBaPcL0YHU#M7#4uG>!%ji3D&u|F -90v%MQIOW@oJ%)z^=ko&wGD_v=HMf0598GlU6fH2>%TQFBL|BDXBpZmPYABLUK2J{EPY}lc8;^UmFaW -LF;MPT{9v)bEVO#8w}`PLw#Iy10xPRW+7ugBBdHE#pDXa|Nx!pzIqsv6_l#M+KyyU}dR$&=9#N-haR* -uS&n0-3GNI5Kf;k(bBLw0k4Mo3Eu+Qj5Ny>?l<;h3@_mRnj_0wU2cQPsRySDh%^7;wgj(qL^Xn78UC^ -v8F;~#>9QV%;&1t(A{YegmesJTz8{9ne8-E4+g9ZJrZPj^B@gMg|rkcwYQm!?q1==2$Ams+bO5nQ|s} -?yOl^pPbM1Rnc}oXw^Sq(9h#UAvZ8lly2Xr$JNqw!2T$eSa|XCJs^jURPyyrme;83O{ -72EG_w~YLI>u`XVRR7Q{ZODY%i^NjhE5h%WwiKikSiqf2i8;`*q4|Lgw|>S!3Yq1^xRFX{h05tKy7sQ -o{49u5Cz&eHL+Ej^8!IjXcMS^MZG}%TTl34=*s>plvk+eBNvDwLmCCT|9g-~oTpLF->Mzj -p8kxA75IW~=DU^S~}6bGXjN<}$&}FC)sg;xq&&DE-r2za5ay=4P?wO$Sk(KNQB`F))hr<*`cOq5JaVL -4zZC-?e@dQ`l>0C`(g#wvg{0qI8X_*01%a@FoF{vBAR(Va>omRjD-q;-lpfd)pLY~`+g*hSMVEf}xa5NDR98Io`x8OyGWRUs=kD=f^ -0_(5kZPn;zATwzJp+$U$P~6CVn~8giO#_rrsY#0!Ez(9VbeZx6}eLlb}x=hOk7xBnK6jh#vbo`KP=EF-ze=YPieEA*>!{x66mJU$&)AJy8f> -7BU`X;7|B|@~bX&{BkE}&-|b<3noENq -l#aBKiKM7smL>|#)uX!%&qQ;LFY;)^w*fNt&*RFD4P!# ->lWM3Gz>T|xErP(O_zu4GU4m845un}+3-aFgV7uFn)Ka}CI8C)m$O`pw)-Ou4nf!zoxeQ+oD=*6`TzL -EUlIDRU-)C*34#(diqSNMvLwwA#OHG2oB?w-IRVfzP5>DN0VOyP{sIeSb{lqKfiK63%JDCTQCWc6E99 --pL$IQ-kg4bT}rsswIdL5x -#N;TT+8r}nI>ci*rU0GZnU_t!YA5I6l+>@AR*Lw6i2U04$JF&#vvAp8d`uAlwCQuT$5q+e)yxZN38GA -pJK*5ZxB;D7s@$NFBnqi&@<-UCTvDYJ*nbodqK1;Xe(lg;fxym_>Br0~05uM7Hz8(&4xhqcMAd|KZt+owD(L|4>M|JJLn@BnK@=(&Dn%NK@>lCkny6 -_iaYD5PA6L37@ZB@Xt@*k&oPq56!V_h)kT@x1)JXo~R=nO1XVBdOskM__H_kAJ2t8d2;*#+WWmEa!zs -L?{O2p?DZ!v`SOcpG$s0{Q8Ms*{z2kJ4YA407|-T_Pw;y;vAYTB>d@T3ja{^y8^R{V>2A<3=E;;$V=N -vjAu{_U1zO~Rk=gzIl3&+Xw-a#daeQ`9TOsm(Ej6=jcnV^ER(A)prHORNQ -qtMFejyT&MQ*;hK=2gcKqvEb*H=~@`0vFXngNa%leh}TcO=+Y9=oZuRO3s9?bpmfC!iLl -JPGqKXmz9-$_yldu(qhicdIfM9~R^9UDx>v`JVH?ClwEb4;mA7CYF2AK!YbHS7`xd-ohK*H?~LZbuo{=CIA)JYZIHKI6(mi#=SRLGMne -vFnvUwX0U6SnfLMjC)iSNxsPBrxgnDs$d1-07<^gK+!en=DxXfr#%uY5wrZ^>bhN>1{HXJ@M7<`O5_C -_89c-!HPG(GnY2@bNXd)b^dVk`VQs{hj-Q&yFhsaeG%vCKfHzNJGl!=GOGR;uOYbq5g_lmTh2D_tg&MkfEGop*De8>U7q2 -XN^0^EyR4AD#93%~C?pI@GGx#2MIB{!h_nPx#0>6`?32b!1dr_sF%b}RhGL1eBq&OAtThl=LF}xpkvG -EqJCo;tJi=W`VVtcsa7KbPKHr`YMvjY9vc`SEWK?A-DIz_tostXcgNg>-Nio+nROi#HFH`re3XhHe8@ -w6Q$lB@b4ghd+bdAfs7(gA*|`Z@zay}tP1@yhicf?i}UmzesjxNGdFOhU&6Qas9=ilf^@2(F8D14VMh@|0-l3{8$y0JTkVbuu^`6 -<^Ouri@!l`m)bZlc;%C(3N3dN1>dC-SX~zn>gOi>&D7hXD6mhY$pD!%Lml2{r@V=W!}~2~W?dSyL(${t$1J?33r}-nq{9^BfMA-v(!u_A&qOnAmNB_2 -P$07TZ^M(iWbHizKFKS(sNvo^?*HZd2lq1({EMf3g_=K}@`K+3N|O|hP$a=3IKyHDO0p!01KS}Mr1-3 -0^J#)lL;(Z=CwrT$YUN|(HYx%P{BdBZ%p%|bafKd={x|)(G7RXOw_tGK{Im6JSH4AT!2}@TeDgk7x!T -IoKpG{viOq6w2#T`502<&jiGaaXCIhFYpb<=h+{_jDQg(BnjDc1f;9$JFhfEBB$4>~!oO99F@P(u%^x1hhJXiUsGjLv`Q8NvI=4;!CuP_maUw_jtPGR9W$&51`c)ZbIR0I7Y -hF?%bKP@(z!X!s)f5y4b@8 -uZuG^+=;l88vg+ei5j0=ZLZT6-b*jDhqF4f<>NIu>P+p=ES|wB=`P8`9!}OZyWI!x@K9P!<8+~p4$zz -yO?^;8=dd4-W2Yx0RyC({R7ASMD&tdYGcC~ZfSTTK`+Q&G29LA0m(I{LI_Qmg91e>s)t+?E+p};KMClUWPbZ2iZr -Y)|?O%)AzNM4zM2PJ*b}XRo30Yo)EkbX>XQs0lA9l88_;%~)(~C%);W$*r?KZmyH>qgLvyr!EZ*>?PC4NGfA4GesVT5yONY+*ndF>Tjxkoeo}XC%dwI8UhTdB~O?Ru;p?g1nB8GUn^d8C|u5B~tN*z^IF7!11~*<{h@fTKFoG=~ -1*y3uf~ZdDh^SXq5O4eMTakJ$Wy#nWzZpeRh++&9f?)sVa-9%F|aGSVYBCL;){i^%DvkmW$5!I&MLTy -Y-4AhrhYPN}wgh+W_#iJFS8Lh^Kg@}P~I1z><{W5twXs+){6n(^g4d?61d={sc&Jhmb|XLu3)V<*PBD8zz3U -cWV1q?mzH1ADV>r>lF+6)=IYavMIUF52N4zsOh5JW6-R*}6#u` -5ZdS)Ipl{M5J@@6falGXph{>zbfQPeJPZFDiwEak^x5&yNp$3cr{2vmdnk)3fYMkh61f8$Pmv#zg&>f -L}*ylI~ul#YyPxO&Jwjo3ETW!k_&acU1x(CaH=sb_ucOJ&Zt{%JOmf=->(UzG^X3LVjN|G#1Do;2voGR~m6S^78PEvN0Z;(+QVv8f23~BmM%k9%(k?mQtFI031rigW9{BpQh6D6 -J}K9wJ@!!dqy#v?ys0ybzd|RRm%EeUvYI`|b1S?r@K!fjFTs{q -odeyyVOdvRhNC_eZ~~<^$(;;r+zQZkyZZEa26vtnAhYO6%(OY1rAw^^L*fq8+kRjykuo<6x+bOvY7=b -!&AB)SL*7p?&^NT^*~vw+tmDV&pc}E3dCg6~u{&-S;?iM8lvzgm1YXAp3pct-vb$%Ubuf=gaffhyHQ* -#5?g~q}zDD#S&{dZ+q;tI7@#rO$Huu~b!- -Yi$#jLl1s86&LeQkyhWig{b=}<=2Z-F&G#;nbZEGK31Upba&UKBdKEDF2P*?WF?}+$f*qF&9~t -dJJS|AVgLbv@Nv)pt=QU~FsAtOb@+6AB6NNjLk7^Sm|uD$KgS*YLVm~2m!g{C`i8LcOOy2@UhL}&<>( -@`r(#LnoBV-rXR`CP_MJc2&W_!OfMB?AnHQBGY!BuU`pF*eM+3c0>bhgde%KQor$D=-8Kf7E^vGOx98V&GX`;d_7Du1ey3?DS)~G(>_0VUcHOI7;@ -xqJNJ@AA}BMOaOCEpcAoQ^t;VFx*T{+~e3Cdm<~>s^9HHg@~uNf7J_4l?(d&u9aid=hyW|?KiOQkLS8 -sQnp>)71wO|6hZ$RFZ<2$@^>%!w{Qxau~=}Rolx8STDChmG?O_hyMpq~ -y724Wy12WQ(1=+{J>E__A1d^?9<@0MX`qs&8?6Z=JBI6}S3t5V7NR8a&~Xl -v0@<{bam@9j>pidD5$}Q~#=9=soy-Dijgyux>fL`~B&{x~`ntbyAuVf|#)8fr;<6#H&XSbU%TJGEe`W -&he&U_07y7JGqL6Y3cckPzon7fCbt1@laZdcUSFVMUu2Ah5K4Dt-Yi2W;IFgiS606c2DD)Na4DG-I=!X&=rY0omx;@cUf8!$aJ^{{=LzM9 -lyJq7>kSQoUeSFkeCxJKOc_2oSM)*bx*4brWzal@Je_i0x4naX``+3QfoIHaUe&UncykE9rwcTsNMe@ -L-}(DK$%Rlgj>wzce8?a9j13(LiId4v;c0Q&+(Q!1$kGyll7{#8{L$!P`a&`7J+wu?zNg(>n4)+USt -1UUV8+uPws%(tehxvQZT2sv$!)Z1&4GbJ8zQ6?l$vc!5{MiV%nYeDEggUA~+qobtv5&RY?eBQQ?Csp> -=ztA{CzB43W>-LbVEg%zI_#-`HC9>~!dE?Cg5-_1B@h=$vl9{1YKK(u}YYtE9JJ(ChM;aylm%%#6PbA -vj6GGIB+pvbkuWP-EUNRf|06J0B*POIqzI(;@{83-6gPH~-=Ew5L7~WzZu33Ap{L`ToC!yZ?RH|1$vp -?zR6GiBWoslp?_KG`(dNtT46`NdTKH(4DXJFxy1lIS63*+`gU+ -~&TNd60cRx?aC`X$5|b3T)>;rCTlB(8X&D;u)`<9bi~~F-H&g3uo0eZcf`I3bLl*?x48=iVu_7_X0>B -43&fA<8o^9sYe~-j8IBFlWzeVEalW2qg?~wSgaO77cUfIYGNc^4X`k$jR^jEC+8!CszFHsr#E7tovs0 -@7%%b%n20A__hPLumstEX{w+h<{HlV0q?Yuph8TU)h)lYZN|QrNgW&J7Xc4Vviwr=r9z=+ -+FYpzYrJphF$b)e6A<1Fw|FY-eKkLR@+3}MyJtf?qZo_$BSX3hjX19I;>~@v{&?XMb9Cr-Wc#XbU!p>u@>sZ-j2qZ -$UziheVqp@$^L1<+M?;gF`!mm=@>FWM93HymK0e~}(-J63v=4;IvI<2@edV0PhM{=G^JViMP9;kS;;k -0TI2;msyg^hW@x9r4z6C1=lU#>uY_@3aMcR&eUAxyTc}eQ=g+zFQ?>0t0c|WNI{*#PY(&n7A;0MH|$) -qn=wd>2#xi9a1DE}#b@_V&_^@#Q<)~1KK$-}p)Lh(eF?;P9X((Yn)iiPnbZ-i`tCrK2okQ8@uc@>Vwm -CK3?tRlnYq{_SXU%N7UTW216Wt0;y8p7OV!Jsk0ZXcuGCq?dIOjdi&RGuHNJbly-Ii%X|s2J{hw&ki! -`Sf{KTUK+smHY1;V;{k@I}b(m(E+5Dp6@%Ii^#bz4==TUH9TK_b%y -frWGQgcPjgm;STlP$4wxSkgnGIiWyFi?bY!1J_{{`JTP28^CVR*(d8ZFyWonOhk1 -hWKp!!*2x&EY8|@v_%NN&;t>__*k1LUSTzIDzENF19)NeyHm*>~M`I^1ZG+Cdl0Jg&9cc1tPb$|0EUx -D=Vr~fdzMIj7Mk}OTL6oKLlLw`EI#cUnbI0C}2Y?EIzTRvp8$*fl%3VfUiFqxc!!K=^bnm6;dnA*CyB -!Emd0aOzhS<{>F8v!0M84G&0E4~5;H4LAe!h6!7TOtqycKaAZM{v>@1=8?=YQPzK58T)gQ@l)8y@A^F0czhK+g9@D4CQ~g66o -tv{$8bve?sP`^t6?aU{3dWq;%_I-Fh)%~xxq3V@W*>T&bO7BN;$QEq_OzakB~A`vK -(cP(ZR^`gz3`*maOybZNx)zsq**WkNPp`t;FBm?Vj=|v&vvcQ$sVZAIznP;s}wLWA6(L9FaM1-_?OWQ -C=0`cM2BG-SYs4HEb4Bdg-nS?7?&`(V4|bv*M*kZv7%=YOghq*p;$pFL3bl=2LurU>oe__bqx*<>THu -UnqVuG;4<|Cz$S-^CeB@^Sh=T6nZOn67%$x^RmfqAR3pln!*iUC*$rZ>cIuO27%NHn|78k2v1z_eZaWi8 -rK_iBL~XO3Ple+*x|im%FrhIVvCMQ^60RRY9+KAubmoTIu`2ZpG{4jL6y~)?X1KOUC<${=-tejwqx9N+p7R6q`$v`E1Avu{Q`g_t4fMAm<}Y6I8N~eY^dBIGC25?&7>ve%a*d -|RPuCI&M>pu4(RY_25j!<#cjH* -<~@5UYeiU0zx?2Fzc(d*0B*RM=0#{qVbDUj2#AhZC0+=@Hv&6o`+G;61;UmN -Cx5!fTackPA<{E7eoIHG5Ma^A5M$cILK!yUH|E|nBr|oKSRv+*ndC7{J4^T0x_fN -n&zB5Bl0vvnB;VaXSLvs=NPjUsM+Jjh#l|Nx{qE><)K -!Jb$axUPBr%a83$+Mbv-q{b21=%@60lrms5A<@RHjdGmKcv=UGLM$qNh4G24>1B3pTk`KU?3okX-MOS -AWpYs7-PL)0?MjEEMmg|(4!YHVN5Bys(c>el2_MO?zj2)D0i#r4%Ah%`AFc{oM)R-+t=HkDV5!ZB9Ud -FzG8)=_%rFVmEI43YiM4F&p0(G-=^(=}$(Ni_-}mar234ti2WtlH)D+J<=Z4nIELkzh7_ -nYj$CZ7eEHS$KLeL9&;1@*XokgUfGG%#6TqAa{j~Q7vfom$_X4>v1;ig11H>Xk2K*-D&7cWqz{n`G -@{rHTOUi6q2M5F?L%`VV$`WD}2&z}o1N^0PFakquCi*LmVH?d^?}Pn{ywKagD!#rf2Lr2A4A=*Z?c#i+gzuBI_GcI&nM6bdQ;z!CB97Fsh1n!N&J -i~CO{V3{l~~MO>*&dKezsbEOyfPV|4b%KcR}tEg!_i{sUD$9{aWm==V|O^Ho4!Q00$R{KCHSGxy4wH* -)CV?_fK{Q#OQkViygBI}=U6To{F=I?;UKDpCX#F~)G_Vs1D^O?9klxhnm!<#o=6G?`6x)!=J1zc|9p9 -iK9M^YFa1XIIv`Ow+YvwG*zT{W4gg-XV|Eg}#Q?J9=~(*@qBZRoO{Od@{e9Z>v^`0tvJk$=>Qg(&CJI -)IPs(=GA#)r)q_?Ys5A)-0pPv#`IS2Km&6uBkirhO2_#ljrgHI!1Vow2fBezU{4&cix5Q({d`oRxu=dv3o3*%7n?No9x>WYagBT8JCrTFav9;4s58f0z1${3FUUL)yIdOO?Dy%x6Am{@?~U$x9=#g= -Qx|3meZ2vwI;pmY$TPD;u=Z~M>j3)K;#o4sOQ8h)vWrr+xiWda89FCao`ssU>oYJDW*u+q&a?zh7Oa -2)HkEx4(c<4dx>uK;Ysgc>hb7cYr%%L(Z}Y;`J^9ZH5$HtXD_BY-Wt{m`ukdV?AI#j6wTG|%d^`br_W -)Aft^UFI!sKcdd7sqQNR1zKD5Gv7Y7|R$A%3(iAKJ`*Z6HS)E$2iUCtAySI=1QU48A`1>X^SERw -oTU~WXw`1~07@8|wjKSzeWn-cVd(S>sGQ|h*7^6kdhp3H3m&Jf-vU#LGhT75QSC!ueT3rAii1>73zFfpi39j3x7ij8D4t6ui70?J38)jY2A*G`hA1DG3g*b9&o41_NM$99iY=Fr;^nx5^PI#x9l~;iOKtqRgNwp@B5{al)~mN%Q3_{N|?7Z% -h|6d1o+89bOKOTbuiM1TSZSaZbSd=w~(ex>w{Q!q&e5#@p95(-F`jVO!kqy8gA~qP -L);oY|((vMr8&i!o#q5Dvl+o0}yDiYV-F0_A^5lbN=LWnMrR-8~rO{3nWmn?y?v_hEh3ts! -jM}-cPL%nS83||j&G`x&k)%1KRX8K}V%z=f50U5|!ewCkL^%Rk(fmeH%eIX)s>RicxN2Hg!WhvC>%Cw -gEiMWro{{AD<=R4~NC+godHSv4Ya>A3ziDL93IW(e!qxj(w_SM3PcR3N@*wjcN -X%4a+Ua~XVx!J&noAgKR{6z7qgHeq*%J>!Ddyo(!Sk@%UUOPR^Bl4$oU6l-CL96c|EDP+36g -Mm_^yiV9wVufhuDMoyriu;ia0{9so?-%kiWCV7A$BSaclWu1JC~z(7$92cp_Kz(6+OPXa0{7bMl8xOD -%8YF)6?h|jdEsy5Cj4K$%lE-TpWfkzAO)JGiBEkO)}1cfWSJn3XNw{I&}0%6pm)r=kC0%;L2{7CM*qT -%klqXeG4K&XfbwyDDX=AE!B&WD;Y1O(c@;+B!RSMiN&TDsl`IpCCAuZCfjx1vJ0c>mOENgPO}Y*B0s{#e2sMErAp$&?{;t*d1Y(fK;p^Tf@_YBw++$PuIcb15A=6PO^_ -!x<_T#3e!t?31^33{zmCqQPUm;clnTBiJ_D|DECe8lpj6jMI&@H3^@^=Ah^GeZ*uVN6{+q52Fm-@ozW -dEiDw>>3&CC%_%WO=b}r0N@J9IeH~zebjSB$j_X1_c!l|1{NWyW;+8t`++iPlnJB$a~iKNc{ejT%~}# -H_02-ANi*O$Ug;dM*&;<--IWCAqDlT{Bo|vuV0Uk9r;v;-!(vik;P{XQ0P~Isl;w%J6xxxm#9fHDam=#E8)aX@Z;^muB>xN+ -QY!&6?t^uS(5iW~>Yu@|5bg_NA_$RfzmxkNGYb&hQYl^nOI*-h^(RhP%R9yNTrs+leMj5Lu1hU^L(lW -Z`wRSCf0>NjcW=g3pF~NP1Xd3nY03h`rtg;*#Tlm-|sdk^Cu%r*qk~QA#N0yV5=FOx{t~@YT>hMJ?(4 -*Z|!?1^N@J{1~FDrUP#9JLbj?XN!)X_UoYUEhI$QFAY%WvkvH|4NyJmddE0$F4lBA_KNJU=@lj4(4dr -if}N6nYFQPjsF&04pT4&rPeN5Wc_t8Ag&gQi6jRdnhwIP_k?p^mNYJu|#`mvW9^)_M}vCXPwF!UuNf~7|qR0oJ3b_(K<4V>_jzXqs -X{W=ednLjf|>r#LNDr@d5L=Kz?T}$-$B3QJw9$d%doR8?@oeeU^v9CU(JOUOQfg!@4**!ywLv{m`#iA -i*5z8?y=__@!-EipM`@znDNyB=KaGov<^w@Cp;K4og1vk -_E+gXE9|+P8gSPC0ubo^?=Rj#KcVJ-;WA%g@oz8jgH_W@M2qaT{EBX`A)xiWTHQQMs-vLjmxi+i_r&0i7 -d*fMLY7`d>*4NIfM%hjayZ;30tm0u!@fsbH2N1L_iT(`^DDBzmJJ!29rTkud|v*c|?lNq!y|lz);Ik7Wya4{cN?+ua^6BzTdCn{Z?QTro7dOJ9qHN$GU;J*OOC9TuG%&5ZGQud#AmH!!P#CJ<<4D -I%mW>t4CN`EA~vZTTnPI7lAjmK10eb2+8=ldN4)4K)g&zN&5v{+9~v?C9Zjh%{fY*6NF#a^AYY` -G_A`1n$HX#a-2<8>)4y6Bl*0bknn+YjNv?1Xt^6ftonPVx8+v#zThNem>B -Y0`Ytae2CR8@i)G|w+}3iN3fx8>usi>|zV>#50hF;Z(BQMm3p+Z4{akO`2v@3q^GeUQ|J -%ZXZO#>AcyVG8#oR}z{O`Xtvi7yE`C=@0_by>1v?|FxdJL`wr*pVcywD)(=9hNNEy~g8W3lIO|B5@ -95C>iAXUaj?~$Sia&zFH_6?#d;xDA+ozKRhs@s%~l3b}%Dib|CHxZtY~jTXNd!cc^3uUDCcUC2}p6_j -%55A#_g)w`V-rT^{b)!~0Td*NVy@Tq?}@0CXfBa9w6 -*p$)WXW^GAfO_AwqJn-U=7IA`5|cKzkIE4V*r2o#J`pNQ_0p6NP2?=~#{aHI!TOPk%_%KM_NDYm}B6#;YVZ?P-|SXRS-Bl+jkpTI%)#C;8!g87 -vSB_867F7lP2*oBAJfPalE!xP;%F}Tuc|1%qcJj}1s^M@MU#@I1F`_J#dVMCRbw0$=>=e-Xuhnga{9Q -+PM(5gQ_IGulL#5pzh!=Qm8SrSOPUb*NS0*yFTvPi(G;+3`82=iPSsCW9oo_27R+Nv+OzmYQ#krxLjx -{}%o`pVum!94)L-yC`S(9?51bnx}r>kriIH^MBTPG8-{@WLDOP7hwzfYof&gzCFHzr;!?7aS)O)}d(Z -oK5BE$vw2hHJsjI7`iT&L-A)hWeKIGe3aOB1cYzHG%ZR -JqORw7U*DY?J$IVopZgM8j@#WN0W$x>@NNJ@pI6z(1=Cx5y%DdQIcQbcx4&qTYAM(BK7nTg8>czx!d{ -%a-zCVnIqP$SOkeJ?19#O}cplGt}$w3%tME91EL}@!nGDe3xYucMIsO&k{ -kc8(Qxx%cf_n4tu=c%COGA>m`A`_b&B((;m(X=R2*yJ@W4-R&gDmS9WlrxBlyurGR=u;xa69Vz>bbkz -T4p%!x_7e-hJ!J0Yo{m-_ip)ELY5W&vCw^L@mylr@ti>2t;l5;KHT8f443(H@#u=@H6MMRZ+Jey{CT- -5J}uwr_u0h?ip_3M;da1!G)*o|$w1={tFE0>^s)}?Q=?c%{=vo1Sobe$)xhD)+Zq^;dW`K!-Z$?PUAa -E3Jx(Ozz&BBPL#THX*utn;e%I$#6ZlcPqt%{lynJW*o$x4J<61wJ#A!~i$U(9KW1E+MRwqtTPst8)s` -kWtH{(uU@FMclO7E6)x9@U2Wi@!<`VPT={{SW#qxw2*bSG=a@A2ZJ`S|q-&VSthgZqyivi@o@B!;jQ{ -qwQRf9Xo!4r~7MD&ON5h0!SRYNs&*TR$Qs!hDKo?B?dWMROq=m0G#a3OxA-_u1yr;_b*~rEk&aKi*#{i!?Eh_NS_&sF%*@{^Fo@z@ -q9A-sp)V_ABF_|)&xQ~3)HhS=%_cz-1V$s1Q@g3QH44=W&<;tjKwKj4m=rpyj39@V>A$jkuzKjfvV5?5=$}Yu?o7PDJI&2F>}uHzCQr9skmo*gq -3*TmRuICVCLLeyZgHSIJ#L4e&`z1IBD!uLS6w73G&;uNO}O2;0QDI)R@*Z~L~lkl^_ftu)3pJPB`box -cic&#D)H%@7QcGp$CskqdTI?Rj2 -2%I;I?Yx%V;D&#oF!ig`V~*6(V}blQ*?QK~4D+jmRj@OL;w=$%X5Si9w0xbBkdD4(%zF4nKw#fC`Kh9 -BY`)Alg0Cg)0=4B|$f6xDdaC`Y@I(WbNV*cVUB^_4co$i-obEH?UxIX*E7iZiC(=lWT`X(fw1&XoF`_ -gqee7x}#c_K{tdoqT-BjQ=o#Oci?N=^%T;bJ1F2i#daI37#mCl@|mn#ehNp6kf0Z3$}Mw6m(0%5Ig!Ql!%rp8iuis-;qe-3=tyWmF_E_7Pl!kMhy>H -uoE%|;7~&Sn*V4>m0?v}SE;w?Ir|i$B(KnfVU|2&RoA-0&f_t)(C`xz>9^O91=W$p09ft3-8x2)QeV% -+Qy>p|fv&~yR255_Xt1T*XErx%nEhOV|6??VC%Z6TQHk2*pRGT+}t77|?cE2C$*A=XbN$NPy8?WkNxa -OH0kI*%@=)?a@-g`AYif!A1@BE7Mt~f2c*>89wyoWbDNJv5$VSass(o|*c%w1LYp4$=K9Z^-EgfI&gT -4T()<{U#}z15OgeWZV_k^k$)zMW^%Do`L7X%^YLRy*(^}>*kzb4Fj5R@@`Qda5T^s(g%I5mqVbt`?b8SrT8TaKGLRwR;f0cM99Ggm!Z6%Hj>YQCFawTb8w5lD<6HYUefTz|gQQ5RTE|^uW3JS} -{sngLB4g@*Q^D62!`Mc%kQnMb|Y1LJ-|!Z5KZ`>lK!wWa8n~;7{qPhU7@P-O;y6y>d3eCDwGpKEP7&! -i~d7qVO5tDH%{^Z>bHXC;ZaMB~#H{k$Dl_gMU_WEj1a=w*}tLlcxIt5nxV(-f9f~qQE0~a); -I`}G272glR0!nGUwM99NWY&^2qtU8z8mmc_Oq2E$w4DlCMy~PwycI2&jvf%Ro1zfDdV65)nbJlyZH>2 -dx>=5BSz&QE?DOj}86_yjZ>CsZ`2#;EOH$#)PP-Mw1Jre}mDNKEezB9D+cu%j1yguAOn9EF0v;mJyta -#F_Zsz{#mFU;*ypp8F74Oy6Bke)<|}S$bhAP4jkBh@;;8y0RzdDHJoU*Fyyn~`tlR1ztYI_L+8qvcUw -rD|KasL;pU}>KQzQc62m=31?Edk=-?-dw4*X%g`IlWS_-9EL{y4~C2kY6n&rZ5_9)x{lm#OqfCZOo?0 -{(SE9i@&sW#S`Wkbg4k`&*Rw&_42LZQR*j_NgE)(j#lI(<}%+7JaGTDKAsW$1VUo3TbwxM`TAveCKOe -ei+;p-VgfM< -eLKyDG=YYlx-%`|`_3*Jw6t;t+cq@n*oVM(f^rl_2Mds+@o9?jyDo_zJ9}7C*n-`t-Cl2$y2HB-o{ZL -uns#$}+)gb?{Z3f)yj}*+kk&EPhF&H@te+2%It69D_8g8{|ZnSBej2uNkz@Mss*gS0K!wF9Q>If%u^7 -m%`!#$R6{ynza=r53z~k -gW1@z2GyVBso6l)$N(E7s<%{Dm^=}J9F$4ani8e=7l8Q!ev6VIWA2fP*DReC-XtYcpsS$myBU2m;0Nf -Btw95-B?;njqCPV%TLeXWOzvFwWQ!{n0yjh3Id=DEIQfb_mCx~dL{BGl%7xgUGyKcqQ00cyj!-wJQc+ -*Csngar)8oLErLAcdI -GFiUI4MH{qkKo^3dcW7`Rohxed6Izv!k((_RoJ_O;tsuX3J%zes%uY72*nY;O#Z^qD^G4_WyfP; -tB7*|vzy0VlAcJoW3XDbd|&a)lO1HCZg7{fo_(YKcFQ -h`BM>S1sMngN8h6+X2k8+ia#%-zbTS;dki0b`ZFXw1qRF?&~Nim}un5Qmmz6ijA5;w#`Ql3&Skp?zJdo0=C*WM>`1Y;x!=N2N-j5wK1EMcsGs&*3e1Wx79WO+*s8zjC7 -hwm{u<51E`UgtMaV|ZUFm!{78{EZL3e7mVUc^>=A!+oyLTU-*rC5wrWg|QV*;1RD=z7>VhYERLlp3kr -L6{%6qJ#N}L>tBk@tTUFK$w$#!t9)5b6AW;&!>TgVThUXVjj2tjQ_K(31@iIzE0rpxs$fS8af9KZ^BQ -n!so-y_r><|jT|sXEB+`*R97DMAp5#Mds0@6a#&FPH1pRt#2pH}U8D&Z6?TW}xeC`zLwLS55xO3ohRV -_e1Ob7iEuk0^sI}n4h|z#JT`d7qR?SDcOJ3|>bui_s!!e0yM3=NkOjN%R05(!&+KB8NRG6DOQ -6R?6#=op^FmYvemcsVId$yi>hbxpjo#9YycQ5mWU9@v|rfL8&g~Sc!RR+zezg#Ec_He6q#!{JKZgk5@|C3 -_Q_nGWs&SIkq$?WCXF$G$OO4# -AEo!2?=kHpOly-3MmoXQfuUTJ~@Zp8$Mv34z@xCy~_^x3y1VB!CK`k_JrnsfoN0V&)WLWe=&{1?Y<~V -*v4z56}Q0q5uC^~)FsOrj%HB;r(1rr9Gy;8O1qq}D(2Gt|FUJbA1zn^vypRidl9uy~dT5;;+ -lG);V}G=37Mc=Z(NZkEB+*+FFj7m8cpdfVs8>%|F_an+e?G(r^zHLoi!l1$>?p97iZ{e_QJK{I@n*3^ -E!@&h}l3G&a$5?`QOd2Z7afezFB2Hc9kl$140@y-2HvP84=6|h{7ziA2`-MKe<^)+Za4d~Ueri<{--_ -JmJ*|NTF-Az+h)2uEJs$sBnOyR~GcJTFq%&}H#s8q6FUMIyEhN@&k?3`{|b9wdHD?JGcugwdDvcyDQq -i2QP{L`zUvP-Agp##WbKb%>axo9VGlIh}=&35N*tI+GxquiW|mH3rnq>LBq^{K@yyNqNDArJZK@stGv -lA_L-6(Ptaz2B65^w83aU?N(eQkoSnd61pQA%ZjyqrSIxBCRbauq9htC#aEz1TZiMkplKH#@pi(v|-R -qZgJCQaW=zfsOj$2ld)l6AWU%Sq^cSb&qC1I$ra1!#_|S01W6p>+>Z)Ns>v;5@i~JMZ+ -ALqIOig7xOU@#TPoW+!=Lbx;Y6F{ZWyhdA)fNjHj_@0`c#R24NN`Fw%dtIUWu~LBt9Hwyrqv73IT}0Y -hgN7UhJx6o%XzQy54S~>`GGn*6y*_=AsI5=G*lXtwD`YO9s96SG9>-m55O5e*t|HD -=OvJx~)KDiw5$W);pVQC6J772F>Qx?aPAY2};1u*fc2mQ4i%b`z=9}bW^4a&ZBr)M7x7vw1SB+!2m;E -{f!@~^9e5Oml^@4PIBe#6De{8+u*IbM!@bg0o|DYS$RdWC$H$-vJAj@B{!=p+L_RuMV%>Fh6#9yFpj< -XH9#6CdrCLv1WSbgto@-~A;QkK9`0pDJdA^kKKzgISYMZ)M@= -6*wdu{hjFKW3+>zw9)z3_O3w_emwZQ2G_Ab7l-nnieEHM60Fe|{UP@C7Km2nn+N4LGWU%}{i#plFo33 -4!}Z`Vm806%CbVOz@KZYa>-+mVw+H;E`};e$2mGh|`%ky`t<)6wt6EdqVyQef=crNM%7v(kM>3Qt5-W -Ouj^|AWNM5aQq>=p=Pw%ajyViQ4M$^np+Zpr5Wei=c9|Gev@A=X!K262)e4;VKK-Q3(&w#D2t{I5EJA -nH8O;-w38MZ|7;1>QNyl@D2*puMlq50fS%K=IakUt9wMQy{u7$z -J$|uhF|o+x*LMVLSuq)i3F0goECBJHbs~)ATH>^*Xjy{qtHA_n$PLm&64r%K@s`&-bh#i8y6f}6&;jf -92*V4dN@=77F-{;qXzH~T1$DWdM#2!F%KB*7Dqpo_&CUm0rsAt+{b08nsk09P=yMJmu*fF$9;AI1sdo8yx(3*LS;~mB;ezrkXY8Cav~v>9dln>c{DOpo&^ensWt -~I~{Yp*6IC<6i%_PViARvHO_cvi05H<%zSK}8Cl7bs5X}My;Q~tUumJb$wcd|H~b=Jt%M4~o9NwQ}eg -38bZD9Ach?oK6wS2CPP%Pi79r6gItmFlLEj25f0XlTv5x)tITONPnrIl;zDPd6s{-U0)kdmqcTd8U8Z^Zviwm5lV9*Y5JCo}K%M1QVqcr#5uI)MQfBk@ZEd5{ -bf59L~?fAFv@IkZ#>3%}ae|qdUEd0;M{aA~G;2`o#NX$RYejfwnLobpz+6Sml!yxjJ13R(~pI(E*CG$ -6sNPc7|)1!GA!H*Ra1o?O@eU3g1FVh2%VCq1)@*_2kl}E#03I5I;i8}lq67*=X*vC_X2O#6{XPM~JDu -|(^62pg#-=fb^5kptp~=GbdMvYlGo_TBrBFM@5Y{8=MB| -Zhf>)uRm`Cel@IrGAiI#!}|58eqARD{JK!oSe=2w?GpvvR}(jybKvoTqPxAAQ125wGtZYnbG*Gjsvg~ -xuGc+VF4#Ou^ddZ~{MWhg!s~FkBAdh4mRa?s-3F=xBOK3z`ps83ST3v3Lt$Waq)WK9G$T;^)VE%jX6h -`4i1nu%IHuu9S6h@>m+1u2aO(c;aa0%D?b*&uiH;~6SRM13y>Rd?L`u1+s-%kVOH5x@Kh!wN0L2NCbi -cb0@vgsrA>c0~X%k5V%_%h< -t=~4r8s*wYtFEA?$p5BooMpf-g&rj%OR6SIvL=i4Q^KoKj`38q>x_H8RoL#byMF}^W5>e5Zff}DX;bf -5o3j+>UZLX&bbMd;6jz?VJY?RgePuVtHf{CR{ZZDL&zt6`=aP@gvBBM-?~m{CnGHjY!!DfZ*@6lGD3y -Nb55Ot>wZQAXbQxmJ7S1qF?j)K)4dxP!t1E@%NH)bR1b!iA*c6A>J3^EJP+@D#zcj3t*)dFkx_*8O5V -?i>iNj4PKfeVByJq))FV6$McQ#6=5Z*W$!JVvh1iX?sLtvDyVH`i=Ev8r1PUBChrzv+>!gWpcXRUZNK -$4YPrVo`6r50DRr=Av|TEk8c5I&DEO*~BVA{9!J=*{&Jm#rmoj^0QoP<2p&>OkO$oIQkKBl@+x6L=UD -$6lX7I4OSLVpsJ+j;G!!`|T#)uJG)}7_>xZG4xu3(9SG+0XEiq>6VK|3h(DG#GF^H6tjB3QN?90>CI& -o{Ti+K){Yw?+MXvO_CpAXB-7v-T}i-{gL^VR?_9yUcafdSA-c6_SBIEr#GS0vTdFsQH_1D;A%XLDET* -cns$Hbft?x(#D9?I^PA_|Nacmt}D{Xr*+zB`YZKvl+{+|G9-v8gi>favz*Kqyy(0@bhUpD&hn3my34F -miU1A+5T^obsA8X0&XA^Is}0O8|B`s;$(J&??(j|24{OMVOoA@H!P-m$aXXgU7~{` -Us&Qzgti!4%LAjxAF8i5&6-)Ch-HJ_apbQfz-d?PiqPMDZSVuzp*17m6)b?iJ*!ZiV{oRoPzZ=@`NA^! -F6O_kNbM`znCq%oj#ygz(jli0>@}R1&bc(8b0PVF>sG2caN7?St4lEAxV9Ghp)(dj&D*DaMRDThM6BT -2MH;tsW+~8+*x}=rK<{Ahft_ek>RB!g9BkJ08g=Ru)BH)C<55@|gG%Jc3;;IoBs0nj}Q_pcL4sHvR&h -w1Nz{_$qp;)SfOGj3k5Gp;nL=Y_%x`~L0iLK6H>$n9A3VIfuhtsJ;+c+G8T`1Kj9!}`vi9i@*&0!aGu_;n7PWf1E-a{D`SJ3De -qPg&&mf0_D04Y4T9fYFfeSZvog(7_Z@F!lZo{;Vmz`=AQyhWW>X)@s);YiI`$AM0vBhG_dn`UBw>$RD -rQ4@&#d=e5Ssn97rs&3;n}I$;0vI1kp2?11f3*B=}|LZ!{`)K`fKY)B5-AwH?>3&Yko=9%xo<1y-d$M -9pczz@TP^)+Rw?*t@7jy7Jqj6O?%dZ|JCiG6YnVIK~+55Jgz1NVRLGeLtyI40)B@fw3n_|TJ7!aSMgv -+$g>ug_5jAQQJhT6ANJ2eUz5I&xgR)%}TS#k-JLBS9$Yd!?U>w^NmCjOa#vL!isTXZcnPAO<{!%XpRe -nryqir%2LmgZJ>XTS2~KIXS$vlrWz#Vy0l@X}9=|A)@xYB)sTm5?Sd2z&ewdhg%uKZAk0m1YhD+zKzx -%jz|7@LPV3|EQp96+vF2M8Cem|s>L5zSA48%bgp%4TiVd$3`W*9%(LiX7A08rwiFF<@Wz -p*1^Jp@^k1G>Ra!zB3Y{>(kF#i3(uzW6jk=EV`)?l_93N07dsV8_=4JAzsCv*v#QKXK%PiQi?I3;2lM -(c~lXntzt}!2_R5?1TQ$N67#@;Qr{tqz+AleCG~(;9h>B>yO<8jvOiPoqdqWk=;IwDafN^2mDI~G9fv -@2Kv6*uh7b6pM9sf8HDSvL0bNfguoX$9$o#(hKSAX0=Q2Fd$N6|AlrsP``dyQDYZOIv#vj^3L^rJc@AkX8TNWCj<*eA(%tUG;}=n`NCPKpOjvJ7)-60g(n96-9k -Bj0|ISMl@;sZQH&ASWkg1YRB2?^4yJXIi)}r8pomu)yoR&$Xo^wkK8ycWB&0SHfoE>}K^9Dnw*LC->W -Y&eiE=kh(&h$~H7X)CrT6NejSoD-eIaNwBDRLF7j(#D@jRjTI8S6NEb=Gw{LvDz2G}kBd>=)y%#SW#E -$w`KZf1pkKpNeM4_*8)*d(7qor3U+|2H1B{Gnn6LHoYu=;!&UyJ2sk^+jY?3vtcE{P)wKo7u#QnZJi| -qxPTGTPh;8ZkET6Y3yLm!hP4dynNX0KNM8@Au3jzs*b+6l%KkClR8K%(e4pQ?dL&EUbgQ9CYmjqw7;T -e3z<$z;_nbir4>Li*|2;G7Ypk7?555so)1Q2?uq_}lZ@T>_{#>g|z-Ae&jlHX8-CE|2^XuT5^K2#E1aAOxX=hqn|u#09ym%E?=BWpR$0~+_K4tc2I1?$&BK`~9tn(iI;G_*xV8u{xEi{kT+z -^8|PEs13+tSWm%uM!ZSo!gBK0lU6G*pl$FjH8(gs}-#Y9|S|yiqo8)LriQZ{kF4D?Q9R=-7zSw61S}B-aBXR -+-rzi4D(Qn`JM;cibG#dGfT1^UJ$9BuqE$$@2F1sryLj>jR-WHk9lWa+VNd05;8N@qoNsyRn^HRu~hb -@)lL|qr4_japL*7etT9uX)={-)Jfq&T%t4MzJbj;kqe-r>%jNO+hVWeB`C_F*3Gzad2e}iQgic(Q#=k -5ImTyXOPLX}UuNsnW!LGM`u97CS=P5{IfjlF-~IKm%x+%fk?z{*kP!g -wb68yDyGo8Xh&rkwcdE7Xo}Bo2wS4M1Vff&AI{|_|! --l_zFMj%ibLi+=KFTTgq5LI1lE95+ne=<^LJkK5>2CVDGw5IXp;*^HG>rgyo)yc}IlomH$A2zAX&PA$ -*fJKV2bu-W?DtH9{dL=>OV-{&2M4NH0`Q)G|)`wDEdw_F8nxhgZp=zO)9pRnh-pz170w$?TLzF@aP15m>Xc!+aE1&&x+5D7nVJFIk-)~(&Q -a_31eT#5)DLp1QuqX2o(^{7k@AX-@@PFltZ7PCQp^`;4We`!z479736O=|Eg5%^7BzjJ&=e{H%bR>#g ->$`G#MKEpLyYpY&aM7c1Toz3)Pm1e3sqfGbY=@o-Vj>!X&L(mgxt;prSAL2R1=JYwb0OgCm6*ELzh5bZ^(|wuI)Zgrz5Cmbl?$#@6Ydq+vd0 -YyVZfhW4CbcVl|j#M_cv^g`I6(>`Cgx;jn}TgTsd-*5lR^E6=1`1BA6-VZT!GIit2;VIT57^u%KB$4K@A>SJVy#cjul_Y(=n2;yU;`+oZ^eEA^ -x44tP{`pbzNixEG!V9_AWMYKlK3(fJgZqYE!d-TPkKeu??f<+73apv~x6`seLbDEiH?n$dN{idai&Sw -UpS@i37q|te`pWyQ)ZuHDlRnJrhY-RiZf#{YuV4N4 -p`HX4}|*3$ycLn>&Yu2ocGSbzyF_^NGagw}&pSOz^GgfD -5R7pZ^f~-7 -1S=lb;-LpEz(CfbF(f%+xp6*3d-)w4k$2^KJ+IO5h`QFkG`xkzXA~%`v8oT69cYxN;{3X=(AOIId7te -|Mei5l)CL)x2uszM?6%|h!$)$+j3R{D|GDw^)nz~ZSwdgqD?Vl7Wc`(CH!_gZ^vtzF)(EGt-Ooo}>_H -6wTF=)kgPK(kGkKCe_MO2WBRhN(UJ^pI8L_=xp_*QJ@fhcoA!-JQUNch*8o(h2p$dNz8Ki`<()6is;v3*YfXV%TV)h87wgR^NoeO3w+7g+O}En`06B7s6zsW%PJgWpLb!|3yKU(GS9IkRs+dnD5tk -WaJy22=gxY?rCw1Pc5t$kjdcqRWnSNj`NvxXA&F?L65_x3Jl&E5+s>LNTh>(siQ=0=vU+1G&gGR(ZlZ -*XQko%{Os=q4``*#ogK@#@IgZ>Z>w-L*p!U~?O$2xtPL$ZEpS=7Rj{Qz)h&uFQzg)fih?^h2ULTi53_J>+_MnC; -k38*R5s4ka6P6w$M6n}aB0tNd_($0YhmIzCiu@P;IqEQG{7W-N;nQKX-oFJOET{gPRn|vg$ool<&KI% -;;6D}$UlW_Yy$2vU<2Ddy{h9tRn)F9I6Q60!^NSkK9%TT&kRSQjuNAKIpZcD@1|nA))~?yCUu9I=^gj -n8i|%?Dpu2mpA@Dzmx7+)CCC~gf?e@9#_MjxRJan``emQ3EZ<2Tb1AWzX-9L~xI{BW9h3s(OR@w0Kj_ -$(nUH=v2e+W*1zlcs|CV$H~y7B4rSraIxKP3JlLCQ=7F#AwS0f4~8u)T+B@fV1{+Jl%WV8;n%ZCyg*C -`D^c=OM4gNYog-zWwBVG2+3UE;BFPMrQ*+S}i>@_Io%o))ON|EZJ+<8{}P|&+1$g!VpE1s0Q{L*xQ?q -`5Qb&=g>6~zr2)e4QR~#dMeFbNRjDm-(b?X*cN)m*DN9_i4XLGXTXe&s*>;=S~nBVQc -Sdy>zx0ui*mm*)r!mIo=o8woNFwT)5MO#;?`vG8VdK%WZ^#^)fVYsF_`9C-n1nb}_W6CcMc7?PdDfBe -9|3dXZTbFydvkSy-pKxAHs&XEa&@ -H-{e((DY?82wZ5J;af9}Iyj>Ql#t;?7E3A_Z<~kaNd63X%% -h`lTW8W%ENZdZ)}8G!JOb#CWsFTTqHfKQ~+8lstKj|b=4WXw$hZ{xtF-GVu+WCKUCIuYQVAQOC?vYDf -ir;^QjAN{M~vv-D`LNYTfBodg398E_$y#oxNACow#+%Y=LF@3w&?843YKs)jhTHG|meOLHLCuP9D18f -dEMO+tiKud@)gt^b64)TAx}r55D*m^XT4CPFe^m7;+FS>Oju)6Of -Qty*(I?=nf=AZfG|T0&s^s%WE;SwRgzn3Xh~y2uZ!>0iMeW|Ynq*(*JvJ)zszTQwH%hJq8^fR%nVX{x -%AduVR=AQnwQ0N|T_<@aZ)Eya;IowG`)W!#%k{2mN=xzb}A%654l=J*^>Mb|Vk!IR#^@h{}wpl$#>2D -Pu2rcC-FRYZErv~Aew_7>_bRu-o{R3;Re)NVu_ID-q@?wW}2Al?+le03M>3d{*a)wD+PQX`7UiQ&hxHH?2{Fd2Mwfq!kD!v2N+x=$TN4$h -Mv?CD5}9xd+pQU9GmN4pjNahpDLeLhnB2>PixCyrDv^1IaRp%nb_zb=a-NnDagEjU3OXM~X>XPkUggO -9iHzhLxWQ1FL-9kRm3!S@p4xQoN12KiK}!^b-M;evzwi=_UN9T`VTVP(IuBb!!&lknmY^_DKhc4xEqO -WopMx>?MCzVV*M$lsEk>e14Z1cIBwdz^ab|I{C(X+mqaso2&YOzn}RQXg~!_{uk`{QAxEC0lm(gC6!S -M|$>4E0g@=wkwT%v&jO!`eb==uym2Hztg+FMlN4|&EFh~h`~>z`h$Ww3=C+$=|BHv`?+>G;cGd|iMb_ -C?3W!P0KT2rp^3cZlyWQSIf9+2MVmAetUJ$m=wy%CN=pkFPvO0k9Z53jzO8|sN -bcRYl?kd=vUJl>4Xw#iONyiqWKWI3dmC?mJ@KkhBe_c#LH!MG1rvqy?%r#Nw!PvPHWpNSyyC(B(15gSAY&tu90}T --GO|ZpUy&8lCGbHM>Z0{iKf6ivH#yD!r&@FS`7);lM+1Nhya$JYf4TN-cGIyM=wz-0HY?HlYnxlCMlA -GvFQH_UvF?v#!XcwF -2mjAOlnILzUyeMYTnN3lmT+a?P1Roq5_uF8qZi(p1h^TG|K)|rXnqJ?_lZxUf=L2*v-nrRE9w@U1^jq -DRK5LUfI6LulW0T^1|X(6o(MtA0GE_3G#!FTdI8m)8U=_Hq@WFUD=qd<{xdl%DAQ|(z)n0pV0#Ki)VU -SqrJBEJb;Dw{rr%I;ZkzW%L!%iL({ccpj7!U=G+Z~eM;%bl;#4W(5Rx?-<@)5HGsm|nuEa```XZUH~I -1ni9HJ=xg~VK-`@UCw9D<;3gl2HqMIv>qujj>uELZ?CJyl^b-g(iX@U$Hp;;DX{&FMtDB65)+o(O@ay -}PwsU%%l9)?N)$*G?)i#Hd`cJl=~Y&SNanr(pg2BMm3^Z7ZRuMG*TRyaIK>tw{#Y%@!Yi21-?2@gy1P -93GuOt7`lN8E+U^SEUIEwYs_d@nH>(V~^eo?l_qZ^R`wKksbN*W^l*!4ebN;07^cr<2m%>!TxagA4I# -c981Xo3J+Bi2#P=2G|p%>}<_v(vy7ju1^U9DK|Lq`h9lGOHyl0T`EY~cmUpgNi{1Yn_owM1RC+`UU7l -w@KT+jY@{91zy3k;DYo3t5+3P(4+qDF -cJNd^ -Rx=Sglzn!JN06VcypMN7kshQeO^&z3#~Gr4j_O4+`FOM^M`{X$j{Fn#SzSM*89%B(Ao?MXpB_CJ`xgf -P|F`vA{<-!1@HYkjKht_H|D)FPS4ch6@KkTNB1}TlX|P+_l5p)uw3h?)%h~SGW0Rh*w|+rLE>mNNc}z -)kR%COoUsVDczZ-v0n`*jF5KsJY!`S(GLuX=8!Z7<{h7_QeQI# -d!Or<0sADh^vYY8)|11Zo8bBgVFf}T-AObny%3|3e~zK_Kh=7E1_FnMY3aUGd8NPV;Tq)^RMjTd`46q -`IQ*(Kukq~SoOgv7n{$}6M9`8uOL$>i{=v}#!@0-(gr>brf@I;Gp(3ms}Oz) -+ZPtkVY7(ChHrocg7M(rpr)bxV(55?DzsMX82{$!%&-gHgAVAp3;4#HcC{TYz?=RuHP=Ov)Av=581PV -;aUil&vy8~fC39oamZ7zIHp8q`L?%tkus4$5!e7gHRlpzu{{a@W}JE+tF@tpa{^$WtNvO**vY9M;!dx -E^N6+TDHhoi-<9*d|vy$Y93O0qXITd@^gQ*&9t|Xq77%4S36y+EsqRj>)}@^MjdkO@AjKbDLh*Z8>cQ -olmab+pP&OPGRu4FwOr&JeVGz!#CU54<^AN^eN%mxBpz>^`CeBUgGtu-F_(d!cl}I2$Uj72tgnm1%J8 -Xk`y23AoM_`J?c)f!vg`!k5s~rcstzf*p#Nn>dXEb`}K;;ppZ?p`&B$5S$~9%mR -uZS2%XSkJh*SQt)r}R*)mRkrtm^0D2ssd`jL^H-cjTwl4*4v@9BGJ;#>#=!*sq6ds0wncyKy!Vq^ksNZS8Z}EfZ2uN#e^y>hjxRK3{$B3oTj5c{@pU2>{o9v_{phUWm!HeLa*^8pE -TRYSv$7BBdzlw<

    zyL$!_h^Xkg}uZMjWSGakRu5|XjH(c-pFh#UKOkBS%jA+*b=zc2iI!5}lx5Ex@ -f7>K)M-P9Gvcq(J66*DBc5kf%^!>*vUgpzle4uVZpTt{%y_LQs--Kq~%4gA~tRkZjb-8J0S*6lUSfi$ -(@cX;R&|U+44!1n*L_qM>h*V6=Luc9+6)ciJ`s0Mqp4<+v -XBP&$Tb>+c~(kv=L7sD~v@f<4?#s>yKg+Nr@Xl705bxpWY*>b;&2UNBy}+xzNK&M%J3H)y3YBRJbpgp -DPP0r_!vJy!s$jzpbm16s!g)p7VG3C|!p52@xKhl`QHdQX+$}+fzID{Y;3#mdfsLjq`pg>A4gw)AK`D -wG}anAs({Rn;YU(A8L)MzBHH1v}NzObv*a#r}T*yXVud9{n`HoZi(W5a4Vwk`R1lA--wr+n5!Hr-b9Y -SUS`!-t8xn8F0(yurz=JbSI8h@@(R4;a!a7?O}rtx8&CBO=P&hZd&@e5(J6#d%Og^D-nNhGKGi`ouVn -`Fq`5Km^^Dz2fa`=lx6@iKw@4XsX~a`Akd&qD``V3t$sq6VG!enMSQV}gHugT$b^@n4$lQA#L0I;AM*Xo2A(%||2>kURg{XPCOY3Wq|59P`F&x%?_ZrV4Q9-SDjn{md&{r*qi -Q#b*3!9Vg#dML_2T$QAJLz;aT{1*O?#y{Ab<m|5=8V;|c*G#%#5#vb(T;BEFY#-Sh<4~GlA~XX!jCfHBi^GviX_ -S5mIo%sBL2RS__!El;E}1`<3Rj(Iq@S7!~cbsN11U7AJqdXb_9Uf7grQa4ygXq1l`kX;Jqdmts_2T&73Zgz@2(!ceh -zQmV{8Jm$)8z#LN};%p4`UyG>C&JD`_hH|=K|nlSWYvZ%`z;ZXo -c26iJL>wql}R;b`x!w2@UZDJhP^|Q6aB@kLqz@^ -=El?mc>Q_jD{IhOwB6Ay%xJJNkpGtVF!5i=!~(9WtFBxJ-lN_Jar@3&Xs;gN{t6Su5q7}NQvjR!s$mX -p!{40*4w42#W+2IB%RSa?%di@XyE&UrsHHzPxoOZiK@#_Ijw^t!Uw&^~7;-$pbSDXF -c%)MEcqh_-;c+ao!W%r`b5%b_X5FiEtVjk`o1qcvg5W}zkKxO;2%T*rF>F!#&vNDl)ct}YRJ0f=M0MN -n5ALrd^^X`I%moOGz`OQXxTp*Em5k8*+F`d^N<=w%T)CS`k@rIb$wt3c#Qa)V(Wr}?|d~F_!y;bQUU0 -|ym!+S+C1uDvJ8tj;(443@*c$v*a6|cv`J1Zp}SodZ1tq~~lI9%e(70Mh#eZYCSGQ^F48GAe -V5wXCC(v1)a$`Eu6VzB`?K%oVSm3 -%Ho^(Lb3VMF!~5==jz=#`R2@jy@Tu5q)OoH4z4GjdlXERyngA;kfH>{pvb#0x9T?spK-va^2Anzhx*j -5?L{BW#dSR0d`h(zo=ss+c}M?6XzTiDy!11m+L5$nEx9n3XGsIQU=X{(!^U>|YuX -yA4WjJWHxTEEYJigdwawJJm|qf7^o5%uK*Fi^X%cQ-siBwKGph8G4wLqFJK3^8Z~a1U412~#v!uIlP~ -+-E4!7~9)K?lgmyQO!{NXGrt)auzb(VPI=Gn!YNJhJdwMDykODP4A>J7%9Atjc=d}Vddj7S0K6&OADpTbGh(v>t5zJ5F)Nojs|ibZv+?ewv+5RXNq&B00?#2yhOwc(yw8q5%JHnv5W!<1WnwQf9&@9%bYJ11Q7p}WsOWEGnx24f@jg1 -#~h-7o`G?u-WRsrE%Vcv#A=unoB4Aa;*PjX`LDWO!rlFXG1hUV$ncHs2AFQ2l3pj(DibJT;=$SXSF1! -Dv$E_;NLkcC5JLfk-tbK+{h3t*}q2yexywXXc$7^pKulWtdIN*x=jEH!~J0&PKElxns{X}C3n^q`bTg -WLXYk>R6mbhl7tWi#hVmXa;deeR#5D&(J`s_&onT~vj9g{n{*f^h;OX$+@md`DFzh7q3>0KLPk{rC>I -AcUY0j70G7G=W7QMbiVqWC!jza@gsS7}zn?9^-$RUK~fq4+pUNc2)qQ5Ar7QAwN>_@s%b=rW(hOn4H9 -q<|_#KNh~gdk92g7|CmreqQv)5RsNBXiw`_@=b8vN&e~YSqKK -Q?ks-(Y-s*1k!R>rIn7rUR#>5cc`wUYNzyD;V!x>d57=DTInaQ!15*X6M{)R6tkiM1a$I+!5G7 -Wgx`DQP4FA$rAFts2Hx*_2R_wu-e1K>3gLwA$BWwNq;P2pk`(lYFbx=Z(kJaM@2>MUB=qfavigdEgJr -RXB=2*aZF-&abA@*&uy&!Fu0)9Xg1ysn8)GJgDnN>ku9qEbUY<5--YUMF81#}ib$r#tQl6A|!NRwzA~ -2a&){eC_Ej&r#^9pO(peZjvN-hCjaa9QWVuKLFB`VzR;as4zN+94^@95FwbB463ZlpPzPcr+&2tDE}S -QSX_22hz{43;;+p+fQ$uVP#qjk^^eqhSxCr9#J47~jM^D}`y|Bq=wc^mi=xOBSa#l??I_`>Y!(y&p|lb_#2YW?=_Z|=jk3-1?{Y84Ul2E)!q(aPY;;ZKxy$9_%X0lWG+Q -*#s)!S*M7rzo_A|cV6YylRxb`bs+zS2bG-n`8Do2UcH#diR7Ga#JFQ=*Kwo5!it3acSE0Sc3kUrHs*= -V0&e7c2$Cvq?_XAemF!!>)IT!}5B3-X**q~$jn)biF(WAumYPU4psy;ygOw@~exI#WOL-MckJ?gl5-C -v8~-CW6hoRo3C*!J9U*uyL9LbaIFv$%+_o2WD?Xr6)oSyl#I;Q5PGI#<hb!yvGRUOWc`;(v8*zmf -*D4}418af$4L2G$7CV$%gnAL#+LRfw3;3Q2DS5 -OY9%0Dv3>f4uTQ~ST*S!1n4t)XUNgB^s_EC_625biR_6YgpK!ImDxwP{L$u_Vnw(D#VkS!w*SK1%Zq` -qC-G5FennLenv_s|XHo0sCawPfApJs1>+T3Vk3Z${5c}6A(mWsn~F@IUX4SrPOKm;XitH&Tyu|+51!n -*Gh9XO$+$Vw23<;&oiDM_~DRZw4@e6$i+^fjiGgUUv2m?H;N6ZZh%QN~c -vP(uvjuiF6psA0PfC-`@l7L|n+TqlZnj%no$5%K -4<0Oh4BeY?ZgeV$8X`K2d!Qg;B=}~Tp(xX^Ue5p*p2OOe4%C`{uBTXLSLoxJYN`9NaNgb#wPLCwg4zD -8WQ@f#$no36gh(aF~-%lDV`}Dpf`G!}&)L@FPcb^!lS8ogw-t`=r#7Gb8AO!ioL)9PO`B)%ESgOj))F+MnP^@5mnFo+@QU#qSN_{|l&U{ -T-@0L{5>fQPnR8|Cdpf`nOTlw$7LlV&qGZvsNviid%)Jh~LBkp$KWQOyIQCr!IL7WUnEu-*@FQ_0cBA -;C-}hG2=j3 -T?1DF7P1T(4l=$VRwPN(NR@==>?M0o0Nt*SAogoSJ1IGrWb<^`>+=G#5d7{bk~dA=1O$<}R127cQr -Zm$`JS;yNPy2I$QL4(`UJkK*S3rZencIYYkK*mZba*`NQr73v`5DbExv^?i6fj7nC@(ccc~bh9m3$9a -=+GYMSA6x;ye|}qkOR1kuiA!%1OK~aeCfnAX}$pe*<(g#Fkfkl%jO5ZhX~ebB(_|G>ddG`7!fFDt~D@NuTo9%hP~xXsuv4-=*A?gCl|`K2wng38vE23`J2ftpmGbLQ%!P<8c1v -hxZ6Er+(~n@N<=A6O;;7Ne3zoTyGg8QMmqc?FwQv#IZ)CE7+)@fG~oU%C1v#j832R%8kK{p(EIFA5mJ -0%PY)h+=~FQF29L9|O{4)!zoFV7vy;4tb3$M&h>)>?nl}(wQ$2x6#ee+`wV^*mRZEc@Z>4%MkK1~gqt -Le|L%Oy{%Hj`D6=xQR8^QcDb?p?XG+Ve#gdryVk_X=P+2zAk+kxxY-v}jyPt4nUtWjx?Xtw~$rMV@bU -_G?neOG-wqT$PZG6f2w4Pn1B+$ND{(~8RdHDCBVkwlhuUMlGM8#@iH1PBgWtnJlHl+7AXOoB^6N2%mH -y4;M>IxB*AiSlO@jB0yM-*0uQ7;jUwVrdN<_ug;dx%e;UgxMuxKOjo92+1a`ls9a@ZU~~MX$e}ymT~% -Y%fdz5Es9&buX*_J)8!HfcVIfP!EPm!^C`PYwbRc0z{d6za;cFvh}K?dZsL-zdCTUld!ge=$E#jQPNC -)BhG%pGAf`G^V&KFhSk4varldkOQc~*^?W!7= -9mfOixj4?psdL)y|$7doGoZ=H`;*8tx(L;+yjC?0J7jta96M#VxJ@N)=3gV6Nibs5rr@?K3SePH&D4e -&@V56=9PgUo5mVh+cNm_@>gJ6{8|y^cUovf&f<}5(Je#Bwj3vq3~O^+awVUs8X3Gq=i({YAnGwxg$fZ -ugG~t+QRm=liNxyEo#86es^zQCT#7v0 -659+leU;9So0{NL$*EZL9!ck*n$7F+(ei~S=wuKdR0h3WKI0yd -qcn)opD+AqDULpaA&)6Z`;Fj#%5;3Ii6p}m|>9 -N@xczI^3YcW7ez>X2SB8o>(SOJ-9{7^R3r_0EtVwquyLv1t4a6UQt`fVh-h) -${y!s`^lYmlXYe#ll&=VTM`{AiB=C``RFdz|V1+q9b!TZl=+)RkSxBftvFVw%i_jqGZ9QcQ0co(l~GgZ+q?0kI>?x^_yOnMh$iqo|?^-% -y8MGVY#6VN^bv$E07@I5%vkjJ*9Vo++y(sxM^I%4Lnwt`Er5wD4(InEk$TWc -W)zF#^UHBZw0#U0mikm%Hh1g>=nDTB~UlBZe9Yzl)#bA-dGl`%>}%pnhe{kcX>T)*Z2W>RK;%hmL4>qZ{%}&IW>-mIB%lU3q$0E2V^dcl-QV;W_ZlFzdNyI>ZS##gzX4s -N-)}_+6^m(KsD0vB8AH3&a!rC8UE|Jx#5D7?hgP{dqu`4@XC#DXSTGKLF3(RuDzq(DsfG@o9sak51AW -*1p#PIj@h>j+=T7mrlk{*5KU5WI3WgvG$59f6X@o*4jK+7|iDL*se$+C4=M>2g-*fkZDEz7Td`U*6M{ -x(uex&JxL9>VbQF>H@@^8~;a{Op0h(5y%4jpCc5d2F%es(_yK}U1v?$pxclMtW|hr2r|6grwazvC3YR -B%2;BP=PJnA_;j)2N5v|JKm8lXT%$8}Rf``jU#Z(R3Stj(CZM{4V-S9YP4kM{Cb15Rhs+)X?1?2%;-lLlZ~%+2!y1Kwjt>WpUETO -Jg<`O8iM_?DO0ox~NoA{`dq1t$_TjZnBOAvDqIAH`R1D1e!@c86IHref+&5(K6hEDLoZo4QvR%B`t?& -2p)oH(XeQ5YKp74&A=tmW69|y$%4m3TZeqprU%5RdGg~6EowUT|yFH8ky5$ -m{<7i6ycD(0ip}5sBwiy=ddI5Pps*$C4l}_c=1mb`kfX1?IORnpcq1yDDCdp?ATz*6g5%wJex|_~@#r^xiyoV0sp5aGe0;5N8v!9=$AMF72xMZ3g7BNUt6QEm -5szBr~4)0yD`{@v0pK}vHU{5|6)*`UyiAlua%8l`!~~S_<8yD)#SKQf0dpdY+n56>^+X>eAMsGcm>1= -v`8G2{gR%B%q~P~?ijiu@#zpFauJ_@P7EjOA)zS!zLA;#31(6s@x}99`9RKT%Qa0~$cSpzNIs8N>z~a -2S)A{sMw_*>=92^3_rPDf14Mc>}8j -=3yf<=0v5LJvOmz7*im%d^ -z+cT<7GI?;?NW9mwsZk}(kA?VBEbG(kw)NG+vA&g%g+d)3)?EP`Q{!)MQmb9G)VyL-jKej-qmVP-FqB -*fd*$E)lj99Y&{TK<7rgX!)>)}mENcFHe7gBQ>;ni*WGEqF+j@%1WCch=-;Brl$U%9AzA37n1LM6_Gs -$SrekrEhZz=?3JI^|m`tSaNctu8VeJA4>G*6;+@@*{sAh^*$YfQHeNXSNXk&$_%rLmC)F`C|8CywSAje0Vc$yluH6M&!GKp_O8<%G?Ue;EduO$#5D{XV$8XPdwtvsjqr*G;^t&A+UeKd>dN9VKqvpA982V_NAJGo|;~pI$pH6@*KS*Z#%ITl7#Ll;)cP#OK@q -Mi8dheT@b({mwgmHtTQAn!a1rNplR`BrG{oYaM1pbZRJKP@dZ~Wfj_JF_P_cm1w4h=!^u!ma(9kOQvp -kO@DoFy)}GdvpQ;e@@g^J%!|t7uVI0!&Wo=9@1{)B9}cCJK^$gi<%-Om>9_3l|{ou(1URJ(y>oWYU>V -K>qqT@38uQGCYXvQ~e5eWUqPOY2W|3i}sU^B=ABJG1%KD!;Lt7!KnQgyR%Vk|+t0BuS&Y-8`l%KrlhT7=nH~=|X+_e -3D~Q)1g6{AL;rSJCgiI*U=YY2lPYQIOa8eJ7!1^dwYy!Opo0EZkqN}_c-CmxO@a5j!Xyj6GwLc_2}OH -Qfl~K%z%H&2jqvQn}0?uMEPMWiR7@v=Um<2Pb5|{2r$K#0>ewMKt{ -*Mg1ucY7htdYmLeXq?&n96;3PbeP-||KwgT}M@;CBt#=8~lpjSFYFE0j9GohQ^L -k{NO1X>G?8BT(6f$y&>jWu;%XKdyR(anaaa0z+o;8Lh;z7x~e920Eh}K$TXGMi>+8BN81*ATJB^3--> -2ef5#q+?0ke8GkNp&7%5Z;8GXH!tp_Y(3N -5oq!k?cj*Jvo<41lcQTn+A_+o6D)>hERkl{04^;7JSxMJxYMxzY(FfC6WyNF0R{!Q#Muj_%Zgm|n4_zlLm@;zTci+Kp-`2;xn?e%HLf^=jfPKvD$dIi)FEbWQ;{7d~a&9dCRO(R3-H -(;0rad24ah(VJqN*Reo+q>HbZzsP(%6PJ>a&?>y#Z|*>&eaqpeuv412l`~;C4Z`Vx7e?pa2L!NOm=D17*vdf=7DL -6^Fiv=I57#yZRU2zS=Sg%q*mc1V;_Q7*P`v(4)8!g8}|4M@)V -h2#6=o?cb8vv|(9%k{F{W*zS((U_r$!8P_QNT$P{^}!RMCXP39 -v8EL>PfjvA-g0+Ed=+^b=%jnFfxGbW~imgs5%9@rk0GnW?Bo^Bi-O -?q6CchjQI0be3mfTXbuh#QCI%L@SbNVKDl?q2}R4y3Bj%DULV2P9-9KSD>vwtM6P9%F^er|xf!g*q^Q -`}73snxuQLPK$lQ18}dWEWQCBSgFzxtYe>jY2~3uTS{YRh`p+FfHbX11ptznM7mxY=2;;v3U07xbC3q -rzf46?~rx>@gl#>1pU(memf-;rXhL{SLOl@d2~qPeEyiJ?KX -v<&f1P4ys4=Wx{uM)lsm|cY&)v!E2(y(;?X5n%`$dJQ=~I+QDZ0#T*`(V7jJw-0&tSKQy9mAj6i+_v=4Zni`3hJm#lh`$@ZnnAMv -|Bg$RO?~qjqjHyukDp*3b3i&AIPlz8I1f97!+zEc`?FgF{_GzA>=uDPyT?DfMc`Za_&0`onE)YJ7Ig2 -tA(}N=Lxl}8+f;t^8llk=$}ma1*eGQaN}=2=lrr2N$rDjQ1OFsjGvx!_hE@Y(7UY{UYbu9|3RA!ePr& -fjXks4S(=MEatahn5c~10~waMxf;R4A86zf-%W=C4I~B`ZVTCyN9ea$8cVmjr~=-jTL@R1Z -4#AOq)(x`NU=EfjqQ6=nEG(Wj*`8tVnrM^(yQi4YG7iUSJl-3UGuLoV^IaZC>k|Z{L@kTT23o!0xh8$ -^Cidb0AD{>3n@X>{q+qp#UxP_jB7nEe4Y6N%@NXE397DQt{6+IV$_BoE>7TH}33&dTbLR0CtX#wDaad -JXN{kcqYT7Uzh4K(*DYU6yc(LL(oE#DV0}R(DbUkTSC8IZ(8=QHwI(XTKn0PDJ&LIQ+2OIBz0!jeKj(xYBt^@I+0qjWltc_8hxdMq -^+^4!&DA!d4A!^P<$S2_a|o#)P_bIl&SKi8tj_qWYVD6i}J@-ivx*)pM!_Tc$)&LB7m!BqkL`S! -8l!6)yVu&wBbyVkz~7qo0Gm;6ZH76n&E$edccA8CV~c;3MINXPZbg|l2sXy+GABmRSGTQOr#w<>4+e$s=-8Y^YjFe_Ho3T&OChrUxRmeoZqlToDeIey -3FX1w-mcdqj|`|D4|)O}dCJ(E6-u=5l$lLk!et5w(v+siUWPgr=~n@hITnH99b3)!o8&PX&Y|@RSWsA -bQ`N3dCH^(~XG7wv@QE84 -Frx#al>J+4qLPbxD++yh5QQ>5su_J$TYEeNDMu(bvt^^AxyM%bNI!U#T`DMiU|&pSog4?KaimpYk`UZ -9F>g@xT0aL;SIC8TtXLNi}Q7SLQ>U>{xD5(s(OVbfm*hK{vA{!wWCV_0Vw*$>pp@! -+*Qh&oJ>Xt@GQCFAU$|AxU5)1w%MRQY22p6bbLg2?8gn{m1X_CtpLTS&1H+g+Fdm8?a`q(hfnE7W -t`$GRumn -4fb1hignrcdsgKO<{siiSO9||#xRBV7baIqusLzoYK8_IS0Vfmm;07e|4ru=rC!z->fch0BAcsfwwpM -F>d0xon!Bq}-{T@y%ziRPvUG-Pv-0$}Jf!FEEDxoiLq5oI}?ofdIY1BK^9V}g~&McOTd=vjwqm{xht` -l9HyX$I#Q=8WQza#PcVq*GEnqj$C%K -f2Gqf)|0G03*XUj<4NG4w=ghQWl@{<8iZc+g4!Xj)>)}WVN#eKhEi6eN8soxV;csLO9ipFK%tkDh&racT{nBRC7{ux^@5A++YbFjp*a+#P(WCmf9B?F0xN6ujMwu0>HB4#gVMLlO3()1Y^5f3p&Vy3Ocy&FmZTX2S -p$;WQdgCmpmA^m>gEz|~L*UnIhgaJB5yr0FeG;!q`DC8%#PGP(q;sVA)c2<>Lm?lf+J&XySoblcQHvj -LKxUZ>X7qa^LC%h^T;TjaD;p}t7K-=F=uyVXq=68b5LZ*jK9aJ5h%{-{qXb2 -e2nE49Y%WW4@UB0HF}tfn%ySGwf?43oqB`!wUCBYYiPTF}y_(0tsUvYHZy9?7ET_|C&Q( -gmt*p@+_o;3Wq`xdzN_59lW$g6QZPcJZ9HC@jHOe@8&jKbORy|q!#hB*>N7ix1ejfODqQ?xbu%6i?{o -!Ebvcb;?pox_SA*yN)~xmTt}8^l+H#Vx8>KmKe -1bQ-{cQ|4=)OvlQZazo(0+n>zWg90^oF~E2x4Y+fp&|P~8^Cc7u_5!cDMl=S-~6)hKRHRwHEOBxuGgS -@F`0`t+O&`%sp0J;kJuhl^ZS38Spn`Gme-dVO`R=S8}}ZgzfzWKGRJKRkGM57p#u^2=vW?yqHwo=||X -OGl919pAKMptO^y3|6I_Bmb2JQC*MaMoA!5nPx-04^2n?U -C5%8I|%B{p^q$eqjP4K14q9$JGp=*V5Nh{-yN-E_gCQuq{!b8idX&?prV*YBy6pQPj;%3YD{OIi_*0_ -gwC(zGqFEE`wqU@*HA9uX@2(}Ecax&HYODR~HBdT}><1bnepf-_+m3a)DFb6AtKTp2v2w=iC({k?5mAk=({t(Y)3noHnprs3O*yiuzFH}na3S!g~P` --dycqS5xF8Oi1XdUcA7Rn?nLWaM28{ApuBjJb|xSWu&7g#L@>bDwNX7I6Jl^9;pBJtdv{cJj&sDikjs -aZcozJg0*Pl5$s1Q7`kryzec)aiSJe7}CHd6tTdSP|hz}t4xvOtDG8D@-3d`iIWJ7V&5dvL4Q57lc3D%q -03y;*moF`Os|6KeL(QgS~>K2j0w~gUETTnj9o~a=#Ar9mI&`2g9C#jq8*+dJ^YH1HbcMv&M`lg{ -AQ`qq)E;70z9Vq?*-c`6HR+w?LD;qjCaM>ixha~D#~yT_Dap3&>4a**Uj- -yVYlomGSDltvIMK~^jX3zA#x0j3sCbf6QlThg*tdzRk67e&6@gW{z>V)GS_+UkiBnB?7%!XJhpoE@?9(QQN?(J)FK5%bZYgnXv0e7p8@@O-eNI)g$-T_?g0#$lEpqAOo!8b=?anWl -%t${t-Gl*IUWD2l#gq2W)xDE?HT_WhvgkspBaPkkvnnuJm8(74A@Y%G7lnT~h2M!V(?{KwYHeny`d&!=esew{ci -~9P%CS)|)8;bI;){hPI)IXrB_>ygd^ywIlOwgbtQdj|?dEx7j`#tH3Rh<6eT*M0x8rwjHJg`wW$z0-n -ER-hM-zxMe|S$?q3SJqwe6HfTd;(DP=r@ZAOQ*E-tUS-UJS|SE1IxTp9U^p8o -{*T;UqS1=g&l+bWjiRPS^LvOs%O(=eNz5vR{Al#(q_Be&NPJ{P4>g7Cqbijg=ybPw$ae>^rw5aKzb+jgN}%Znd{Xfta#nFUD0Fgxod#bzv4oFgv@^ekj8-(5!JpgY9(!7@S`yUzd2a3% -HQ1)@b#Ad%8vdH!?1POHoM^u8s@YM< -DQvY6!Jb0q{Ox`=df@8`9vf^7VhgI+RTf(ty<>FmQg8yJ3MG`rh+^e$l1J|?cgNaGlmi-Sa40B2PJvcjL^)iO_J>j0v}Ub -x@!VWs-_2_0;J;RdO?1^L73_bSLl|f*8MHMp2TjRZ6i`{+8*9KZE~uz6aGkjvcD26xlS&7pE=P663~C -YU)-q9BlIqr`*KMKin()H?J+qN>h&v!#ELBF@h)Sw_E3Y_ZTTt48p*~T1AC3SEuT?u1b&Fqk{>2_jMPC5vupK#6bsE)UlpLfOwWf6 -O2Z!W#_v_7iZ&a_s0H_7)X|?GAFyTd3)8%ge-4tsA5|Hfy1>TXBAwiX&!;A`Bq8g6%9mpzZb)(Q#yZh -Go>zwB#kRuRpI+t(hUg19m;{5S%4u1I#=Y=1?PfBftHD%bCd+WtGHy+rgDgn;Hw><+Y^F>nnWhppf>M -BW`kqfq8QxlhzhaL)q#e=}TX%nv@20O!8oO6b)3`yk)o#EbFl?Mprwi^Uq$4 -4xzzj?U@DTr98e1wT$$eeQT8<43G{pHiULi&;pz6iHED5iX<;G*Bx! -iuDwF|Nh>bOn?-y24x}-fySwIQ1KXKb#CHCE6d@rg~&kh(lL3Ncy4B-S2-hU=q0CKZ_3DhwghKJgCQA -r2nHZeuf^x?m(Cm&ES<5dy(O;&G%lm7v^r@fmrW(#rVp|uQCtY|9{{V{6xU_4Bb#J=_J0#w{k3y{53j -y+-k&2Zm?numq@*yEM$tXM{LV0Uh&Y0;7&>xkd%#71WH*kKTYND7;Li}7Jy1;d=X|?)ZhtyLAMI_#XW -}|deOl$PqseWLD|6(d>VY2s2>HynIjXklp%ayVPYiSiLzbv%jHiIR(M2|5b;ffwLu~E2MA01n+WSUD?h7kuEK8- -mLs}le5tErY-fO)@1*=Al$F5hrmhFd0uGq- -`h(iuXHn-LKdb7d6;II7fOx&gbjRz|RKE;$SkWwFSiKAg)gk_D*Mxf==Q!@})kRTu -Ct&MMwj0va}2S>f1Y{QupVk@AmoJB@+c}=;W9Ja0=|JR-ReULZTRfCF>$8wnp1fL|bp(%3mwp-_h$k4 -Y-J{b_AAGQLx99TeLK4f|>wY5hBmU#T1zYiLk~nyp^tw?-~L28mr=RcE`iX3bNPJYUCvhp8?qCmwC6> -7Y=>T0PhUwlQ+kVnx2&?LU8Pv+OW++67*=z`$L$j4u>qYTT`qLjZ*3`?^$S{HBNT*wgU8eT^W+5mpu9LFS#_Rg2NwUENDKI0N&-X0dpnk`F}86M3Vki#`kAvwC!=LBlJWit#aT> -_DrlqDON19L*t?<&a3(kA&XuV^l4QarWy;oe#fVKaCY|v!Er;@$ESX5Y(|{ -duigEF+1l{10ekiql3T~@5UPk^Uvu#+!s!yWD^l~dxqHiJ(!mI4l>ELTKBvKcko*P~X*h0*^?1bvWQ` -hB*T^1pUGu5XoR2RmEIJJaK!x_K!W;I1yPYD+R)&zU@fTpMoR(`EbL)RrN=QrWbD`%ex8-B_64{;hlq -L(CkNs(=fp)Q&!V3~5g6ZTonF)jx@jiRP9^ROoa#hD&Jh|%_Z0V-meETvn>c`0xgpIA8Q-Gz4nzum9l -5_ro~!PKI~0Whv{ls#$0{_*8TC`KUvkxj`N>)iBoX=i5-TWjCN&oJ)d)jg?|I#h8|N_-Q~;foB6dq6p -HoD@|R7k6Z#E%A%>WYt?_k&H5EfD3Pd+AUw?=h+q(>VjhSBPZdfKP_&-0GG%}yHkBNwmgZ1?zVyNb12 -%`$oe5In9il75*SYxS3c{AbxxgiJiV|}NmOUb;qo^?IT@V$^T6uMvyDyJWH~5#3`wZx6y8h*>eY0VLy -Y_K(Eb{YBlQHC3{3tq(&F3k3pwtCa=ueF{jV?5J3l@5|2g_-|NF?(TUqwQ(M+3tdEmd>Y&_N3J*xj}% -=%_H{a?xd{R09)^bfy1^S^$`cU%>HS$@mXAEqJu;n~*fS4i&vZ+rV0DgNbs{kBYu?TC;@@o$df#o~jI -m(U;aP~s3~gOUS4LD>h9p@-~y1bw81vL9pQo2ZcboPm%Z^`Vb$8X6sljUYay#$ErJ -|BX?Byy&e+`J!IkVPpxBzt4C%qKTgI!19_nISmKzD_$dJA@MmCfcEDH+ITDEbhOq1apF0fP!Rn3~4+) -3pQ;s7YEY+IbvS`0)!AqAm%&fK@0*h2>tp4RF~~;SHEt*@$IuuL|0j+7#~TQD>GN -F1T@q7HralgYQJwp_HE_g7woI`0{+c{eYIY|zge)a)(iMI3-;A|0sm>ier47%@NL*JnVZ~i5?VtI;5u6}lc4?g&skS{0{BAnpsQ*z}f2mvevjINXJ -o-h8R7x2ZDP8ivzEfGH9iL1~3<7E(33=a0?r7q_xjF|b#^(6innxVbIKp8lbcYkTPSUml$wg(kx=^N! -qF92iH>eY7{cN8oOCkbKeJ^eAC ->C>!m9rQl@X5BL(HVLTp{?dPxQ -0GB&MJMea9k3d!1n1;--&0kV#v7pdcaGZh{&2d$fAYG&;&g%kHg^2e+8wy*MmQuZU_7) -ulrM{YpJhKUIy?nlu$s_(t2xEyKYpm&uuG8*|yDDNPi(W<;^0yaS(Ri1tNycM;h(?AvZOuw`@LTpOPt -z={r^yuQ=sg|FO7ZSg+vXjJ_%#6`yC -iW^laU!k|kygkN1st^7>@VDfd6F{wI}W6SxW-|KXwH`d?<0g+ERPCmYQ~un(%qp^Y+=W -ZjnI?V!LQJLjr-CSax3@wqu>$grKcIszxaJm{{;Z#SQD)}@$F6`4u+p8ZU-cyo1|7yLs0pJ$BX@difS -ds5pFJ^hD4cX`_g@O*B5+!xyVy#1(n50gP`?8HgS9pk!GKp3oN_7OdH>Jvo#;dEd{--v{zp~zs`o8Z4 -;5bhZ&v8k-`nXK)@}qJNj|;UG>7=r)#Pc7|zAof}MI%;~bY7K20=?KgU@&ylR~KKT{ONea^64d -6uzr!_pG0U?vkuBs(9<;`an?phOF&gCNQ --Wxx64RSL-=J>K>9VEmhs;HE9OYwT_z^*%;p4M~w`r&TAL -`{wIqHe6V-_(3?FR*w?|hZ&CoZBa5f2iF^uE$(t}0t{&7aFnsOtV0nK5YPs6t&yD5ZZ}Fe*5csEC{HH -qvwp;vl!GVM8I+G?G=#;Y?o3mgHc+Hd{fG5u2kB7WAb!#E-ra$2mtagc -U$IEYcOuzLYy$-pp7C3UKui9?r?_iZkzL!IZ%>(gdh@q@G0Nj2i*i;w2?gkD);^K!8l1>CLv@Y!Qc)Ob7dZ&404Yn@8|JFO3JAR9;3&RNmBp -lwH5!(s=?Vl4h8)rxzs56*U*NiO4P2)g9kHc3>4rSRh9(@=zh|>_$Y@Qw5AgD8le-4qop;v`K%me{=2 -&E;N?~sygRQpR*&0QOy0ApUi3ev$w3LG`R1p!ZF&Bnh-I-RYZ>KD%wG1#DPpSoy$w)YMsY6SF0pC`?ZMw}?Z8CzhYmz^FQ}=q265*qbUy2iD0rvxEwaQ8e} -p3}J`1&l|*b?HZ0>sq_2SRjarkr6X|&U*C*U=P8FAy0b9NLC1NE2`+zYHv6N{0kFI)jrG*$^ps>}?Oa -n%c%Y*Y&DW&44cD>Ypy^mFZz7o`;{8Uj`^Z=Y3QteBxP&HYT@pQ6c`55-xxyi@7UsskYK -LI(>FY(;1NoJvBg%C+k`0h`1~^>Gi#T5RFtb9FIOFf!Mm61aaU&t#G`z;`&ydAMq2^Ex?sH -X73CR={FnnAziq}WKN4Enu&Ci8qjavFX2x2mM&D!Oio?y&>z%FMh;boKVpOdUh-2D48~ku{8 -s6jb=LDO#FaKFh5;RxM3Zuw~(JTfT)-FEa%Jhz;oO~D9EZF(EL8iPeogS{t=e~I$E?;c8oF;N~Ar^-q -kndn;^vc`H=w7L^zk=D8CUhksL!3wAQ_ck(epVx%N*CL7qKsexfHZmJ^3pW+pP-Yrg0jqbxM&11#IJ{ -7bUY$`@HYQyMhY5x0JIGQx&B#wYg!xrUy1gE*fps-~QXxI!^&Tl3Mwz{v-QCY)gYLS4v;crq5RMChp{ -yz8xW8}oi<;|M;!LGMXo&+(Vx*G0aW!f%dzOO#~!UKW|#_#K!3{rR^wX8+h(Y{w@4fKoyexBc^PDpvn -vC;fzH|8TaIaE^7TwKINi|H8qIGCp!7x(UXd6X>@8#T7GMYTb|xm-ao0YbHrPQ|@-q?&xi_1^^j4@K_l -P{Dwje&*TO>$gBgJfYNdAr@TLvQ8^ALEj?UU|^a>aj}kSVe!29TY5LhVCo*Ny);k@hfbc<%A%_?NgBG -OID?+>50rl$FvH^7EJFhc%4)5k1`wf66%Vxq7tW%Ji+FTEpl<<=M|h1K(V?+k%w=M$I5$ZG84kiR(ufOLaXxLXWX5B9`MHrZy$=zz+WYu^@2O-=EpPx?XYFbg_pK9n0<_|eP7JaZ7armi-9V^FW~#}wVJF -N-vYWl%#PigtyMSFCe)ie%j{=(U0v7hmT}`GAjIs^mp%j8?KQJRsD1xSJicC&`8?o{6MiZ@1Amf;V}7 -Hz{5am=KTS74xda1_LVH(+z=P3v -fn*KuWn(K+#-bj_4kT6>BX}4!7qU5eMDXm#HH{9U?gA1~3s*ge&d^CSPxPk`FF`h=>kRj|Sy0{*lfgo -S~o)8L+1_jH9~37VxO(mUmIyqdwlQ^x}!BImHu3fkN+UcVMJqzN{M*IeKyWK$}ra6w8Shgt@ij$+q;P -dSJ#LCLz4ln+R%dYWVCuCgb#@$rD$AVJt<=P{zNjSXyX@WcBm;WUETtUFr_j9{TNAaHNQXm5KP1T+CX -TT&7I29^<$$J)ok7cPEj@gwr8~uNe4~fPo0J`DV}G2pAva3-9kVSKur1g{DKYR%;2~U7w&i)M7*!)_5 -`M65d*2PRi_Y94qnK%i@3!S_k{{kp}{&f_Spjo}{M_Onjc%`t9YE@>aC4&vGEL3pWjLOjKx)fRX@jFf -{TDpHm)(_*j%UT;Y;-*89YgV$C0<27)kuK09rT_ZPsQtgB9=je^9K2m4)D5I6K{4!DC3iqGMzs2`6h5M0H|jEEJ8GNvEY6@8Xm?#S9kcJUBRf$Z}wbB@R@WFF3`K3k|Zo>Yg9X#UK@s2l-Wgs6d -GThfE~u<@C=^aFd*~B6R%Em`3f@ilXywV3{I|skX?=V -=s*1IaAchVHwrrZBw|JM)B^S?R%(@*|z>~{@+1O;!CL2+uuTr|lbIEqjxMPWEj&=f|K6pmpOMG!Q~kS -OzIJsv@J5+<@yA~)fjUH@j2V?;LDL~0*fuzI%DJ2Kn2w4WtBHj&1CC>fS+E-4PTGm-GtAiX*`^kXc+? -i5!aw>qzIA9@h~Ov@+O2Ny8$_Io`T34Vw7Z1(C1iG5hV?g$H?|;`-nO4Wy9s!9>5Fg&d2^7zwgNS_bJQ!$ad=oQAA6a|B~Hq)gwG -Cp7WSuRLE}bqwsF4KfoT(z|ly+cZf%b#knl1qkh9*TL|?nx2hfxEzm9hWEI@agc)mh9^a<<>^{!C&w% -#U-|G8gtik8T&uyys<5&Yu{fpaD4jgj^qFU_sI}j>zjcGP-OK!}}q_)S^=V9yZn&-PPszx@8*=N#&F` -c_8-E27ntUGhHx86qc*J{t-O3&DIAoNa{G(bPQ7cWVirN8B9HW$xpha{NC#`t`GV`!T{=h{m2?CM!6Q -k5kE>78`K0yw~HY8*@eyWVsp*ObGFrDwglGE9`SUf#xG}nh -ml?n^Ht7z=aU}X4%AVF_1_GMF*#tepg}ghm0u$Z^pW-lVD&kFCCj>MdMmvL`d~kXaz*2z;3%+VO1vXu -tcn|S%!E5t@^TLsjPtm3jYY3Cm9@xI3xh>_S7nvo3H2Y8QWciSZK^<{q05!^$q -rpBI*ncy2z|H7e#W&eQgbv4kF>13$@&j8sJJTyBub`tcvOT+Vii?00s7}_II^seH6P&kf|-La^=gjk52n~ID-SbyFWR5&4-A -p?h@s7;9rgx`hv@h;3;RgA5*2HbE3XTK;zzh3QDNM!O>^DAIup--rU)L -97w9Y&RJzM*Hb|oGOk@>J-SZBWDFek5LgpjYW@ne_XY1GD8I9i)@S|Ebw|w7&X$+m0qjZ=dL1TkSW2WDDOmpJX+~I~m&F%p}U2wraB?dG -6k53+cWc`ir^y|AW{0BfbCgYkZgOUtuZ3z%Yz#G9?&|!OYgjK#}N(A!7#RY`SqMJyhV@jKdT| -bn`fl{@Qk_~hx&>js{?&sG-&DET+ztCsocRIDtFo8?_P0@da}=&uw<|9#eGz!&{#Bw) -z8$s)y_iTucjf@@%zxEUayHgM&9;4l1Fo&{dTs_J`#6@}sgVDT3Niz_B~AKSAE;n%kE5huQo -5NRJPY3{MHz@)_^ucGorW{8$nh;lN=wt=^eK%^4Mu11_h@0N)~1f?SG+0(8J}=jI$-u7y1?Xlp@vb}( -}=d>jsqo;a$SHAcfx^q9W!uu1^Z`%s?bS~{0Hq*$+8v2@LRQ4B?{iCLV8W85?+5`*eYap%a_?~f;cKB -JG(GiP8Db!z5lBpB?l$Dg?_wmc?PY-c25)+HS15f5XD>N$?Hj3q7QtZYL -duH=yT+~d)yCr01b>{KdC!gk;@xdXd)a}E7ANFDC1#P#w7{y90J{r}7QpjG>0H&v%;S2sF~?`pPLk>v -e^Q_Wm1Z<0@y#BcH{se80!s!Lzv3Xd*az!EK=Q=_F1}5LG8dRHZFvnbqqq-YtdzSl*qo8wGpJCqp?c0 -`xSS4}>qgh&XiZo^kuT$%C2zo-D4mZVGewl7J^BMm+3t;bS7+X`mlHlF0)?aN8<-bw#fH>?h>BbB>iD -YZ50l8=fFRB;&OJC^#%B!{*VEAR+;o@O$F*D^fgNAWNn%Fl5~tYmJ~w)fOe&Xi*@&yifDjFU-P31nm9 -K{)*ieV6UW*g9MH6CQzdQnihJToj0sNg@cSBoSC)npDyW*rWU3&}UmKrwO3uOMiB?H)dGQN)hv_Cp3j -#71aQgSIFs{oTtuv>7c$0-93B5)k2>-AVRMyS0nxA9}<*~>N}p*>-W%ME1YPr$+aT_LC5=nG6hqy^E*_y0vmNI?>xpRe~lNWee(#&Bpi6a -l@D4H0BanHKZdgCj{d#)M0tX{i#1nbl#MJ%bu09mQS}V^%v6 -(^rO4Hl*kVWf)`FXLOh(acYY7+ch4=JlwFH2r^>D()kKn(LRWye;E;eGN0e!YJUwXr16?Gx5@$lmqdc -VzD#-Zd{d#G{Az<^)eNS^-U#U0iJs;1o01mxFOvUZ`%W(3TR8`Ux(Qj-N#7Vur2%7jeBY7r4ALYNv&2 -uP*aYJy=J+r@IGZ-OO2>5S~+<@Ny`YF`>zjY6vF!aB?{$C*~{1y7#@)Y}|j1^pN1IL(6{BCQWU^e}uO -$Qm@`5WOr)iL?HaIs4^!Q_^g3CNwe$ZYJv^~B-c+Ypf(Vhr~klD*Y-J^eQPjQm+%W|zWa(5-bAM%U|Z -B$IHT3cYpZqWh5Lk1Iu6=fuXIOgDWMVsDSw8gr8^m4gPGRtdU~{~-445tjar{bS)m;y1-S=<|uNDIm8{1#Zwibb{^GQyO9 -Z3RP|V)*^V}&w90c&Ww0^t8Vz1y9k -N$Rnh#ATApQp1RMYxKd9l9rK -+!2lc6HeY9#rm-gg5l{Fbl4U>RP{gz2;8)RX>I1ox2$wJkShe-ZXO{)^ohnPEouU?F%e`FXgi@MDCqK -0p!9rXO<2Js-9h<$?041qSb1%wHzosFUW- -!V3;coiFGmte&Www*8?hJVwj%iT<@z$bI=3&Qc3@M;7qory<6bw#KtKo`+G!Pl6`;LmYA4Vv<9HJ3f~2Ovu`cOhidjinMjS -|RQ|8?t%GdWwjW9tsa+h2n3dzRl_ao6CF3phymVxqk -!fuXt^%XQ7A)mHIBfY`(B2~VZ@Xw&nIzY<>A*pyqZExV2c`ZmB#o|4^8>p9Ep_}sxzzgek)d73l86Ea -=rXCB;C396EXX)Mo=Tz1LQR;Bg#$Me`KXd|CyoJ)Bg%f`nMg}WF1V7U`&-*u{MumIu0%6-0ySns)B?7 -n^>H=m?&zZkaa)6Lx+gd}~B7{{@rSO(m|elBvt3_KyxaJ##NGA-r}@SizNOS{jXEmBU>Nl9R#+2+Ts@5`(A>_(XbtMovGWZAof33%~2 -F_cYGZOJQPm*LNFnv%9OEYeIU-gC*^o+8?RC0#O?96#B^iNX1j?hnIlQpjYIWn0C8M>JFDF~U^kbVVJ=pzr*k -yRG|*eMR$XOpomq<}$YFY0AS@v-!|X_p20~WXdm_qXHyUA)&o;txY@@X80x~l+{6phv<4daXF-2!6pt -WiN3%;?x#rW?d#3S(c=ngUEfh)V5*t!0a;IdgqJ?_@}?*0|07WLQ-kyWHk|#RzK{O~(*AG{{|IYoWS< -FwZ@Qw}G+=bgfF{U>Y~gt8om~O#CVz)+KruCHfArrFyO{GP&$kh(iC}B(U2!-o4@NXfUap3q*1b0~x9Pr;lTEo%~+pJz7 -#Eg%O>((Rl(KLgohiHYi8)+wcse8|iz9n0?&YaI}a~8CFEfe^Y%oE>KeW!kIdC;RU`&=xoAhS{ba@*~ -G+kd@l;NRT#$GG9|ldphk>9hWseSHtI@4TK{9oJhp&O^n!?;zXjJgm2#82jxz;^Lv%Hu(6xo)`w=J*7 -*$w>}ts`Ctd>u&{UKX#b%vW^ao*DH)(A`8ivW0zPFg!LKD>z)zpPa>VjJ|ZH`ik*3&D -w%N`pZS+&`Lk8Zmh%GX1WjNqnXY^KAF7rAT{yjFmyb-7B*-pnP3mTky^#B9(16O>qqu}504aQbQ%xre -KB7s!9nk@wM;(yVm{hp^a9{<#Xqsar8!Uc!-PU05oLGc-u=)Q!~fYAlU~GH#@p^y7(wy?s+-A5MG~py -v<5*A*(E8?37VNF#7&OLn7KKJ>x1xa3*`DQ^&;!?%q1ei#zafq+7wvhh!f9Z&l`!{#ieZGh!A~O&PDo -J=sKwiM26RuG1PQlq?eX=c*Xd2v9FsUeNY_TU7M>JlDF~jjex{^uOzwgDx*CC!o-`zGs^5Irhj6`8FDlH$)58-cOBKOR)+2KoIV7^g -00Gi?wy^%&bQnuajTVwH`BlRAtcy(aaKd0BwL*>+8Hcyx|RN-e@st^!Dh#iKhhHhE6e(uI^6V&<{Q#A -8_^42%wMv-X25*^U8X=W+2+$`ZRUCJcdTvrLm|ul=r;Tog##ZpJYGV~p8wj7BFwLog`AF`W6_;I1iq| -UD_K}Ig;_hS816$r;roqF$z8t?_(^yy(X8wRhAVg%OEGqX^;jNRQ+d}rtk>@>wL`w~so{PZ3Np)# -#|^hxIQ0Cgj#7V?h^#C;1xuI#2yo3;ENk|!aYgID0_#!x^>R(i*F2H%L -(alCp&sf?p^6(0Vzosr5LZXos@ziM;9+Pd_)uuctj)49Aau#nU<-~|{3utj -?f@pOTm)+X-}BT$l#QLI+mgV2W^!_MIPx01)ERE8cZ$3ER9-{ESyRdbcv|BE1M8L}z!4_wVaMlrJWUQ -!aJW}R+Oc}ZBTe0<6G05)y)Nkb6|*|7QweJD7jYUfhAgocXz}X_imR3tHUY|5mBDE2Q`DhRUPphVS*S -X3alv7_11p2ik-^Nx4@)^PWr~FG?fAAPK;hyi{3yAOt?n^O95t%H^kS6+Cj^Ry;PldcPB97pI>0q^Gf -zSg2QBr%~#IJ~r8KaS58HHa^Hz-6jde8!WzqFrkKeeRfS*2A_{4|;HPxKs$|KAeH>xJVZ`Xm5NcW5*zf -dpZp|i;+C2zg)FD_RRW&-nI~whInrC2!Ga32$ru7$0IOa26~ICQSV8<2Dq^t3YsJnQDN=+D;f}A$M)Z -AKoEo2)Cm<<1S`I?^-HsG504q^F#JX%2L7f2kqNAAFkTlfOD7)m!};h*^*uW+ce$5D#FZZ-D4(m~2H8 -yIpMf#V--6hfvLAy4{Ca`Z`Q|Aw(Xmq}E~1mCqJF5F{ylO&8TKbJ0j0(Rh3p#MNng?dj7!!)Y!GQFu; -J!4&0|YZFX*~3!FYnC$5Smv$VI?qgjspR32ry(cbsr$Y+X902hOO%lJvXohEA9*_$H9-}@dvGu9X*Soe`Gq%U -K8v;RlE>w}ARF?%}(o7#y+31v{NZumU0Qsj%`m4KX!>+AASg2cF}^s5Dx -eV;(5At57}Ob{$$BezJm0C`R*eKfa%@<3m(zh3asS#BM60F@%9=VeI73DCR@ZXHRcC0@|g(I2>zA^P& -^WB?S>aN}QL3gu4AC{4rgZv0Aa4RlpEZO*uH1e1666opL#$-Kzs|BkP_)e(>zcJTfy?y5Vw83N4OlS4 -C&Wf(B}V#S~(e0+s|E_XE5a(X+JgLhYFm=HxEu*?a!eIjh*5`{eUiNr_0 -NN~FRZ*lc$|Bu0|8jBVYn2X>kEq -bl2??Vzr)PNNnOL#7~Am$vPoVcdqHOfAX|V*Z7sUn_@S>`QdWpV?CmQXdk5Wu&UN=S7$wnN?tLG^xnd -n+7pP*GO|x{3S=Zm_&t@TQ!T3(DU6I|2vf!Nryy7^L+M?QRk|w&vvm0v(-3N2-bsrMhC{Wvc4r*(x*= -8s0C)wms_ZXkq)wj}roP{K28+1a)Pv~@Y{fnkb5xTHjPfBd%(*5{jcAuXuRy{m_?C02Y6NHaIWaG24x -n$lE>ujG6rYZO;_YJtrv#wieX5GaT;1t|ncG}sr^JNP(%ljI9Sv$LDBh#vfZN{bWF=6At1>4ns2tfb+ -)whcSf9~`z5YYMY%8m~;uL%fPIBl3!pEWYI?YgIkX1o-Ha(OIv>=?&TD@e> -TqO{Yq6yUh%?kFvG%gdB78_R|;cfKtK>mS!IqJEggl8)i@S-uHar@E&mBiiNYH)Vg2!#Dc|oTC^GSGx -}fowLrFnbt3*g&RZWnsHLuha@=-Hc=C%ce(Ub -Q-~h7p;3LVCZk{;pg6d55oAJ5RT6X~Xxe2+0ug -K_09#P>0j~eLF}s+)9_wZXsSFsmf6mnCfWiSVHH(uLqk48&Ts`@OZ2ruQYAU;Zj$TCsM$CCyrI;qH*h -#iVQt6iR3KO9$YGy!RSnAoj2Z65U-w!%2AiFSl<=$^^C9rnej(*15`yi;nd6S_|&<9sNQ=M=`FT@4l) -FbI#;FM)jLXJIC)QQMksgIt347WP}3SX>VX3OZF$-_!!T@OA-KR-6lPn|RG?SU^EzRk002)!OYGN+h693LvX~8F*xE3LXHLfI -edK9D=fh8809PVICG&6B@SVyGP|keyG}|?k= -L_sleIRU>mg(Bs|A5E>I80D^F;2DnSwJ?pMvX1(~u%4;sBqm-v_~x=p(Q^|+At3vHwTMal0f7rcS4R# -Z{uEonwzUJ@`hoAtXQwR3Ok=|rk++YQSo@o0UxoGGm+@N=Y~8gP1?$Kzyi%G~?NBMHNiZzW?UW$&pT@ -BF}mY3w%?GTvB?1q%oCP9kz>vXm!z -ZN6S3CtN}>_9*#PJ^-?%4ATlU+MMpevm?Rfg4Iyxv$ra6#CE`;#tL=3?jv#3&hrAYvk2BsiN2p$wwxa -TGQ1WQLSnk$nYf%wv?#IlJJ36hX=5U@q{V@Ldo^;GV$Ew>$n4;FH8x-}>8&_09KhU*^=YE3<_`W{XIywTu_>9bQv_= -i8AExG`-aXuw9crqJK}wuO&v~UA&Xo68jt2O7`UcF0%!vTk@YwHu3<(H!U%GhgK_`T47qS>x)L%cA0e}wkxwBge4*QEkn=O}ZkS_a8-?^?k*D@;8r5BYG(^ffkocBZcNEjP7TcvG{Z`1VzR -xsb9zp;ks8iLKb9A>wJn3|vUYS{K!Ci#S8w49FrYg~wjc&Poji(9?L -;SL=`+F-19E>=GXb=K)%Iqn^UCI_61;*QFJ)b6y9z;RGaLU&eAKj>JuSLvWJKXBQsg^|O!Q*_|OM!Es -0`!XuY1Pm*%Z4`3g~-Lr$~2Yo*j$QfWTF>>n$A|DpNybGnTSUhAV>g>qjD~b~PB4W6d$_HXO%8Uo;JO -{0xJs}fj@UrMNP-d&CIQ4Ru!TjpxSvL44r}c|v17K^z{SE`jy7!hXm^#mDb<#&(tfLnt>t@!J)kkZ87jP0WCD9`pF(t+obj3dNgvn?Bo0OcH{%pwectx7yWG@VHNvxc8X3l`*< -&^FSewtldc62>Qe?6)zHEGptz@^4XW79O6#tWII7~-Ja9JMhUn|09b(ka=X2BQapEQtdkF831?W51NW -)fdTK_5G9pRYQHSWFX(5Od^3}Txv -5~W)b01sEU@Uz*N;XYgi-HX=I=BRLZuV0|sS=Io4^>^uB{azi|>NrSr^LlGYPh;DIMEF)h-{dHn&F^e -_6niiq;2Tpj{O3-f-JvYsI~4fgA^C$=4xl3?*L?k^kBas;%sl_kJCx<;4rMDFf4A*_+M$&Hre=XU&4N -a^`%3{bw=(KvQ5JN_EaO6(4lF@uRxV@M7o`4p%4P&rxk;ASaP%yNxqdx3%%G5JiQkW*4>&MXnwLz{?w -Gn_O6G84`}?4;aW?W;8MgN#%HEbhL=)K2| -Os_GEXx|0;N+~3d`%Dj@mTQg>m*yZ$pAWT!mQX?+Pd16wt4W*U^vsh-m7=jbjl@s-3LdjGD1E~y;@ko -21b9PyVY>)@@*dNfGQ~Hpj|v29mcczigbpWe0AB%4)vWNkPxRS_Zo#v!Tr!xZgBuzEeWBoY8#K5?JqD -@HF|ZU!TM+m)a7MW9!02u?Enwyo_JWQVCtUBRWJ -3mhA$E1BMVaxdt5|IGtT1yKPDRqHcZ>Y2gI;)uxXHosPi(qL3Rc(P()9m^E;FQlDRYJsFZXiluY%Jt`OmjSRpg#8^IMUE9LqQ**M8Q0d$V0bk|{pvIS~HVNFk` -o|4{vsc_RM>%p{_iKxspLKn=XSchbHRM_t&9i|LPNLfydkYG>ELjbNLtQOqFYAVuF$k}U0l9=mYfDv( -6A!sV9Y@bf=E-*8D3onem7ajS&?1vAj~Y5wz}OqKc;+rkD5SAS<4kqzYN&m-`O)fba3kHFS5KWTk&dy -B5ITV#o-KMCHaT!_O9X+`ny=~cL?l(Cx%c+cpT4R-CNb90r;dn -TMBF*!VKW8IHwNHV#KHy^yPUZ#de0Ah -ufa9UtU)%t{+VE~FG(NUgj{^XFe|1=sST*a^v>2JHXO8t+It-${TRRf^~~Ij<%e%g(JN%ZqHkm1)_(- -OwTY1UcIZcs3z}?a-Fj89tt3MJOeZ6GS7HwH(t$Ng*Qj3HaM^QHtA?sG -$ejoOD0`-pO4t83sNrHy#IaPW^fh#NNlZ;#*MAjG}7TJ1rXc?{5`NaF0e+{zzGeJt_SM7LkYKpO-C{( -^zt{}u!NO`rSUVxSEI0XqhIbL&p(cLcRh!qytf9a&grIlguuw9ln9tJPq}Nk}Nsm9ko$0$qf$yq$-XT -q4)&YB-JX_1jG3{l%T%kI25{?P|BtyfEyTF~lr>zSGK|%ibTz68JZg7GHR!9vFtn -MoI8A-2#JA}S?JU}u*d_0&Euh+%vwJI8ysHtf``|l$H~PBlnUb4k0*!BF&zRhzoAoUIt~G4Y4oyCSZxdu7WI6b+R;^ho}~i5tW^D&rP8j?*ZWb0&v%YCL10f~MC1NogAN -D$UG+Uh~m%!`!B4KX>qd2D!(Z~J -^pr6f4kkzzbWl~vm%?E%RP7&zx_X}UFyerFASJte=nBgH?F9P`k7R}P04R&a^F;8%HsYT+Xv5~=t-vlxevf{Zv%+^_{Rr -Ej*oP2)ymwIGyKvl!w^s9x2fGwaypP-6^ICMc<ezNqXB=@r0>@g -5Svjqo$abN8SFDPhI~7b1;ek6)C+tPZ_l=2>-Wmit>=3C;t=0B(_DWnc-|hT|KY@s?J+IPUVu$dM*Sw -Ys=kGnT3YkvAS0S-Zv+`w-8we@vQ3qbx4*h6VD(d$RSKTUeF(jBFZF7mc9GQk^PWRg%l&n+R7;+9_M0 -T@yRz0b=>xZ_09iOaLa7=e)PBlGzmi3x3ea6As0*BDrn{$n>tcy&BwwE&i5h!-Ja$w1*&i8LpZRq*+5 -S^Q==8ZrTU&=z)qyjtC)lNFHGjRT?>-(MZ}Z1G0p>RU_@I9}wZRr`{$?B8ZQ(xO6>a@7O@4cv=l3T@? --{T@r@6n!TOsUU2*382HLGMt0ReOI&zy|qo6abaurNwB5HaYLN$QpD^T?~BPwh>+Pya7(@73idwss5O -^DFi}-*-d~J;r-M5+D&Il7KsM5J@0J`1J>5m)mxix4YlH$LZm=T?j2Ll~vE0^9ggdtYozptiEDTY@rC -b#EWGEyvpMOYoNiDc~&K!221Pf5T489T!q+EN6yAs??Yjwq?nILq@Y*Oe5^0lp$Qp+WvvG^bc&`I0W; -w2uKfZ%BlpQLw6z^98ZA(2b{NH#dV!@t(D|Wmisk4wQlBzNa#Bu+3wS|v9gnwbI$_V#nnNg<(#Rb%F$ -i+xOp!62r+5N6!&GJ4>O}HXyB%lN)pkO@to#E|nDVNC!IA5c>cI{O`n*Jn2$A!w(>GHc@bPx1=}aM$$ -Mq_hZ556d9CM7O_+r9fDZrdG5lhGpOy{nO5732>#gUsvWthfW+OLcpKigPCnapIux>eMEaOAiZYYFto}^jxGlz(@&bsu}!{#S~YvVR0#a&`yd -5qmMDR6}L2LT2i|SgzBR|gWOR+-(>V^E@e7}mlTjM7aRlC4ZSZ>rRBOsY5kE*(P%85eyPJJNAIyqu0D -K}rZR-+`^J9^eh>OO**``MsMd)W`;MQx05@D28;B&YO2kM|l93%v%|UeGVAHYviz7d5t$2wRU*b03^W -I{U5kQr~;*8xLOdfD!`1Q3>)|A#~)xJ=jjJebeDj&E}gJ+cFr+hpW@ZmT`r=eX*Tjf%qYSGK{1u>1&+ -I3HH#>Hi)jw;-Td})(^(BB~P8|Qh;$uYJi{mZ~VtY<`<(Zql8`u5v*OZ;oUOE+Z8`xmGGpZ`y>_@7l@ -|0@fApo7mB`yK|+5VX-Vn!q6vgJ1-LKXMI~S<;(9BMm&>bhB@AM1x9-K_*{S^peUvGQO-COln2CQMaXL>d;u`#gSZU=qOPsTetq=KD_LGM -Hn@=-9O)17i{>~F&}yB7|!XN -W~E5VjkSP9FlVXFGPTHZIA6ak1`Ec7+_%5h~HNHFg&@zdDX>3beT$9R7BuX&0rKRY$mPWbLbA)Gn#Y` -jW9rTzpljdq0U|P0jow{`!@(L4axrJ3)h0OPw|6D`~Ujb-TVD!*^JB+8?(3^NxXky5*mD4E*ga%QE-b -dG|#rGQw@EF~eis+vCepp!1JC(}rtpSaZV>q}d3qz9SUWx(-ZrG^V+sZ;mM -}>9HT6W!j>W=vt-NKB7@@5+ciQMI+PC;Z&Wum(mw5TLMCgJ^2m`6Oii&Pj_jDD`c(h<={Wxbx>Zf&adQ6seJ<;5P-)iqMuu0hDIMUc%4XC!1U -JkD#J5u4Z!2+WUQD0O^Y`uFj7&ozDIkA(N8i{~GKGgO_Mow99on40YMHm76Q8W+)lO|DQlwKYcY@Rw= -ATNq;;39#{{Bxq}#jK`AGf>EzE;`v{hg(5l(z9wWz?u!JL8R5CFX%K%LM>qPYVFwANYyj$qvuY}rDyH -L1=|EFrM=?AO`57lJt}I1ig!RkAv05AOBMXYgiW7E(o+|ei>!(qd4dd-k~wg0NhpD`dZa7=Bwu;6FVK -dcMKR%kQk0TWSq%kWNuFkU(s&|ca>bpr#99Nv9xcOIveFCnmY%4-`Mg)7C=`{+drJPjz!SA>xD1rY -Bp;*$SdP}&;a~4f%!>e%Pw=XJ#3gxngJa(w>PC_`DO4|51BmsHV6q^+G2xACOY>so{2gs*h0M$+c(LWSvIN50l5P<(!F2v#b(H4?n(1^3JLIjAelgEpvBFW#D{8ge -^EVi(<}Y8eB#yeLu(UZ&->CL*qkdd)x)YeQ76GI5(Bd}8AGP1@b_tN={^Z -GeDPBaI`>EV2po_>Imv3Sq}5gc9_u(1nbe1r<^u8H@!%>_s%Ssk3MvYn(AWEKKy#%xYtnwy=WL!PQ%B0&_y<4BfsAO3Ofas?} -3fz?%;L)K^1>el>ofMleLgI`Z9yoYIboA);DQXzu9NZJ!I^GVJajoF6e_R3Az=U;8Q|Xb{gSXbQ*uM@ -!uE(dG>zC@K)V&0vsNt7FvC1Bk)&b883iPr}bDH}HU{M_y-618Cm)caux;3m_m`|t;b9`i@6$DIovLD -s5yyBwvq+^pg4l0D?#gX!ho@^&mOu#O6?-P?cq^KIUmzF;l==C;9Xb4ex7rmf=;-71p%KwSl`rkXj8s -_duvu9Aaw(r{h{rJzK{Qt>9-_-g4<|5w%HHso}0)r5G?`}n52*N(i;84V#xPg(q&1kpjgZFGm@ctFOR -WHJ5Pekv{qwofdJ_TwaxqA<8Xles>@!NKG16L%u>tVpy9~AWNJ`HyG7sh`CZykN?Z66qhziMN}cR;tn -GMw0JXbJR3nCu8Ld5a`R@Lm$!fEV;$bi--^wP$L!>%070yhA!XeGAP6J09G7kJ3HmzU>1IxA*&7poZ> -UN7OefLOEHM5Sjx%HXi}d#r_h(?&xZiUmSA< -?7kzT&)@Xo5mygn5A&u{9Fmmvau>rnd8CI&utKV?lD`k2Y~CUv7ksuiuerQ%-%-Gpz~fcM3cr9R31>n -bySNj@>t(GeZYbJbbkzhtLbCQ?}_e*rrv+qe8@HmluD1@>$OAEKve`~QnGPbthQM}5h1Y^S6RYpPOfF -CaAkrVH-zWL-a*W&(cM9Cf*0&_M0e=yg56wt$G;YJ9=Mu}O36aC>$7N|+U50P``Hf;mS6G+Yk;_N?oH -J9e1j58ACdVQz-Gx-%Y@<3;<-nn$sz -+&f%Z3}B6BO30#OtRSym0M2 -Q?*^wNf4p(G&|}+acXgmm5*DEz#OHtsakrK{5l43cyKQm~(q%L;ngcI)BQx&q-4ehm^AA`Yd~jaAJl- -_0qXmv05Iwct;77bBf4U{pmUpV4GNUMyVhw%7#%~BUVEfKY>qn16uZK!v{XSPX|9f_E;X#gvq^%%tIb -j)tw%fS3v>3a|_Xztreyx95_4Z3a4AB({pLo;^u&EufJZ(P*Cvv^e@~?rSe9;BSYiO&E=rqYXDTcxkf -?jP#R4x5I%8O)`YpQkpE%PsWYAV?!tq^C1EwE_@V#3L?@+bCkDk;pK32)HmywL*8FtpgdhYl?|Aq25skmQqXQk+m&kGJ9pG&rs4zN->1icpU21Z&um%0H -d4v#20gfjJatec<@n^Kq60*oHlNOLdxr~MeNf?mGW_<<>|^jK5c1hg3@@(@151-^qNDs*QBq-0L{zOk6gx#$w#f8VAbc^{p0*x{a?eXOV8TWG6G>6mn%Hc(nk5OG -xSm=M~S|ZuoIQ8t9baV#+<^j^!xJDi9%jUE1_+zFdQeld- -4=Byu|v*5unbGL)HnfE*zcTe;+ErLDWfbu5d!C-VMZhK9Rh*?$EZ{%WCr1P!-Eet?D;jgtsU!6-tL2u -xterxQ*a@Y=l}u)QgPO!wYxD&41(Hi);k9i)4?3?_HeCN%hi`i1%qx}z^L+Sfqd4$?UG9`6qK9ip*>9g01J0i2>d6oa7Tr}e*z1 -4R0wB?Ho -|5;&i{zGg0QK$|-sgnI=)*Yn)}DQdJeb;=3|H`k1UmSwmp8YQtskwLM53hfA;?O%D>tB0>jJ2{bWQA( -mdqIr9(G{pOJB^YGv<8Nekq3%)c7BD7N7zJ))|8YxEZsuRB20)%1Q{q?FkO%mj!)IRbzz|U15=}XVZ# -^<*WOv~8WQ#?j1-)L9MDzmc*g10*qU2{T(-XApJugRWYiJu)B0=^ra$X~pWyCooW1Wn2U?(zD>)A<_A -N73ljNWQYgNRm7bmKuvUIDdV{QArAqOr9iiAvjQ;PM0SYdaT2BcM!-JXvJ?qIS@M8#W$s#?n76>AogM -;PRyTBGk0NiNby@&{o~*)IKOAy2J*K<^nbA2&w={yE%kl4CJ>B55E7#?6i2q-Q2f)r+2lP~gTi~0D;( -^NX38yv6bq+O^W@a4(af$gUcaVtee|>n-u#oVyYGjmB>)ZGjzm5B -bo)3fIU-yDafeC~Ri{^%h0j0(kOesEF9(Hu)AAA>h5zvh93VxlDGhj-)&1kKdJ@o!cjSOYn9C(C}W3! -K2?Yd1CLypudTw$xiQet7THj9*I{>(By6mxQ3siGTqbmJ -}-Bz?YYZ(xg0IspTKs7k(+pZ2^gyMeSD+=$xL7Oak#$VSFaO?OQxYL)Oo@U?ZEDQr*tW}do7Xa4*7^F -I%bab#V;n1CC(*ldIk{q=_3!LOO-v9Mc{K=X~$xCp*-FeI0^>Ze9@>oWCe?4OzKonkE2L$VZTwkT)Rd -L5Ym3mvC?DgON0ou47U-SJW;kkRPIfCX*tqhMeCUilpOqALwYHuHO@R{6zV>@P -dvHW)nPam~a!x+ZoXqCV}zd$uzgTR&}7uZjKqYZGjcmP+lYCY&c;bMUY1L&US_7lngYh`ZP|NEv2EEj -hEMGbDqSq?$aCro4*%FE>wo{xyDCFV$qt_gKQEYt77HHSmkPU%|P6y%s_6A3MwE?_bsQW -#+lQ#5*{iJT#{x;SDfn&xR#L8aDjNH?mNH1oJP;#Ua~vynfuOl;U%dbmGr6(H${*x-#zkjX31W&G5cM -4%*xjJ`aQI-$COsfok_ic^;JU<{pjIWv)^`NH# -vjV|ldTq4xKIVDBA7Ky>V9HwcB@Z0>Tc*~{4pJ9BFymX2w8{p-X}&s*`yyOS*S>wmn;08eU+yOgI*$y#=UR_rbBcT2maQhMq#osdavuNXR -Jx(v5+HEo}^w@hUN~Ppuu^idP&xfoG~E{F45WcGYFWppxc}qFO#YsshVe)T-R*mT!84rbVjWR#$mzjr -pl8iXK}-`5!|pHTsoglFb9!9eRgIDzFNqQUupYUTZeuo3&t!%b8yJx(Idnmel#5TPbMAyE=>C;1pj}# -)<1=gKhb;?-e4oP{d*H6j^Q|tQ21xK5vO;Qh`e1JH-xu0xxhPgjG&zqZ(oM=TcteOC39)$Q{0F}d$TM -K@ANu`_X#KD4IN=@2Xqm!tG;btBIsLW7yct6_paP+Z6f{^Zrs2jvR8EW5wLjg=iS?dqWwxj?mGv|CPW3JMFH|3RUo6?f|BDI`r1J6y)$hNr03Ds;Z&ramq5HsJCOpoqY2chE9$v2rV -x(+gb;{znHVX}^JBwbbAo8s|GpZ#Kca;GgIgAd*A!U@D^{O?BSZ_gd(1+Fa$WoasA;OJ?1v+&m$p%$e -m)CUl55AJY=VyK&fm_1{?1+HWS)7(9b2+MB_rFeqT)8uT=S~{<>R{vI1IyFsPZxc~$ebm#lUChm+IF&~cdC6)^^)n#H)w(8SSnB&5y{?9&lElecYO&q8z(F;6WF394KjlP;vmjg4?H -rl^q*O0fZ5UK$ku`hM<9A=!-+%QnZubBG%OHO4p#0w~@U3I=^E-YVYJ*9VAPAVGNfg5{dRGJ4f-;0cF -agszN`9&@mF^*RBgF}}Q|>)JzcaBL$A{h`QM*7$ut(dsMAPT`Qg5sJ?RikL2gNP&1@PNz0^TDc{Co!4)HdJDGU+sR@N?$93P(YLrdP3>iP9C-_kzNb94d%h^ -9@53E$CC?y!CkN8MQ<(orz*9GpfMhUroiL{|_3&Ug!q2o`-D$nW3)zOgu@A#r*Fff%S04u(shaT5D8s -&I5K3t_4Gvn^@5F)1>2#;FlZ!zu92i0#6etqm~biK}M`61xROkX6=uQTw&#gbK3lbNLK>$t@nOQ -A>?-fgymTl{T$Z697!OxDy&!P?$}@G<#u{^2cb4}Ltx%s)T&XL%g>t#-)CmI&!`_Fv6#R6Fu=i8SsaL -dJC%vfhw^d3=VKGr^C-7CdJ29-@ctJgnj%=lvW8CnyS*!))ocam`7gk43l_;WiyDt995Bz5{+Xnwi=O -XrtTKOK%F7z9gb5FYM_W+NWB95Iiw+&CVZZvE@+F6ZCX2&o9wXWn<=nK#S;$+#iQJbnd4Is$Rq6Ydk& -GEU2YdjT6M992F4EYVEp3J+FulcOGqE;vuK!=V=0z=GDATIvF~b!ju+OU?S{*1WSw@K}>(Bm-7SaF5z -0)t;r%NT`K3(N^hsWo60xIoB)_p2^xlo<)MVlwMkt2`n+F6c6AaHYzWF`j8fC+{yrCAum+2sn8S$=P) -Y|8iPc=-PKL?wDvLaf*S62#$K7496NI!k+Adzl#i_BeP#xoXzdDspl(SXbY6#bF*;Aq>4*^U!nqHEC; -z2mb*Bn0Rets-z6pm}7!mj5~yKpN^3bWg*IdNbElI>ypC}-osDO&Dl)nmANguneaT^RTf -HNjbi0>O(n2=Zv_#*UlvA+lF4K`{PyN}GuFu?D!=Gom}M2N>LPL{=|QdPr4P!t;n)&C7Ox-LAGGb)H3 -Q3MVwj>UIsh3CAz#Z7tQ#1A2ev47VEjHuBx0Is0L7C{M3Q0Pk8zW+N)beQ>%ELX5%n&3iawdy!*`X?r8=qgw>}>2x$b*fqf^LJPqcpMX4PG@2Y2; -}k%ZLox|Q_zCOt0_?j3ab9)OK~aHS`tzl)R~sCzS>7h8a9v|u?Fksb@oBJ>nLqoru4?)qCwwWTy~nN^ -;%va_+*g{+;ZvaXm3>%vSJPwaMGL%O)YMZcftT$uVhwR);lyPi%ytahMm}o@Uf;;Mjo9N-9}^nGqy}{ -30eswN=is-_<95iq*DD1gp6@0Q=Zq2M?$hlAHgK7&hil{=DMl~6KEp9hRV$F7&dtL1s5MySd{(i$5SU -W80<9Ex^F~%vOp&D*?ukp*Kg8wf)+ocqC-Mo7%%QN+Lv{{>^US%?Wu&7uqK+&Kl^a-Ybf083dSvSwL9 -_N`9&3p}2iudn`=hRS5|<=PpFH%ozU&22>*SbFjSQ}^CBJUz0CWtk8tK6rEv96lwZtS?Clb>*fBEm1989RQb~NQ*?i^ -IE^pD*j%?C4IxV#(g3JCCEXlQQEbg5AWyAy3HI+X9@CC}o5AOaPZh)I^j0wBMJuK{vYni15cpo5Ch>* -=c;3z7gKH^IX$=b>;7oD3I^kKTKXokg_u-f>o0jmen`|cMWck+@c=QDtiQJBk^#A#IY{OSyZa%hS6Mr -GILKCj!M~8F?#%(r -^5eQhProLGL(>eyfn|=3(T!}7E}-3Wq9aKtCPLt6NkI73GsG8gWt6uWLG%Y0VPH3MI7usN}HlP%f|72EC#0kpyJ;$ ->tHXSC*TK!yUJ#F}hRnGeWPwB@KzAb37Uy<4LP0c{ME-ZijI-{zs46O3^Nw=;W7B#-isFo -ks?CqKss3P>Uh`T0#mk>A$Q|y|KR+)&i#Rz$lnGal8UAF7CVX|U~IwRnVAKcy<{fz)eh;vuXY&v}ygU -$5E6TU4d3tv$x_pN{)^E=*EC2tZKPx_kA*wwF(XG{fP=n(x1m5n`kG3L(2jfa^~v!Y8d -gb`wBS|<}%4GiepDjF?(o$T^Q^#kBTsQVYm@*ncwrX1P5e&L&r&Gct_E)voLdT)Vj@%`RU=gDJPM&z{ -FX1^@V{FLKS`G8?2|0++XA?`LF>i{^ZGz+t+rFw}IZwfuyJ=vY1TibfkOf1>A(mmqvRi=rn7~K$Mt-(;MG9MCXD`XmXI*%Z;HGd#&le9gXcRID -U<8BJw4`>HlLl#e2r@ry%>gdp~05U+?>Y?IIXS(geK=J&_o^MMxT^VQdSMB!xq}9ux{A-)QRl%m;e0K}N(HEnyWY0cg8ZooyC>$KN5 -^W;WHf&sgOzSw{YA$Saw>*!<0-;vMpMVR?_{fAg(?es;~!^q+hyoUpF!sBzYq7ebZ2f8+s8$K -J^J7PWw{VM|iKvh;sa4xSUQ2SRjxoie*=G2xEX2Zc#%zkb}8sW?lXEgCxY;sbUNXM5?+)Ui1(a(o6+{ -Or|M+Y~f#ug9#NkRlx()q?A~#OqG3x1q=#;u*N@?X6%hZ3lGa$1CI2tC$4oU -wbllyB~QBK2D&_NIlohL5Za_)HKHJs5z~W-qaPXgWD^Lc&VhyEk*uDt!#HT*TS6 -5U#fQO!ElzS0jKd<)z9cGwsuYTx0T>OdX_qfC!@aPogcz>AZX}%pizUMA%a=u}q#4InM#Pw4jk93P;` -O@BfQUl>>?)&3w;|KTk7Css9I4{>DGSuViNhl;`NL$2LT$L{gz5=v>4Y2zRORkjbR&FH%uV(ngGMf9w4kwird}==U+fYi(EC?udm>c -%$J?o;cn0K0#;&Hdxd1wGg^Do7g@TsmBBX?^1N6qT*vbVtRhW~%;Ba?(uU16 -N(58ZXz&fUoY{N)HS@y67DS>k0Z-&EkT7@yl&rX0ws#Rr`+c}PeOZ3!r7|J0t3?Nf~+ct=()TwInlTG -}YDo(~MWSax{nC-RwD*Vm;w(f$j~BO7@RBL!~XZU9bdG5i39V5XyG%X)kf9qiX=w?misQFF_+2d=r&dK%bE3RuIy4I7Mi_b32-Uuvij7839qWvVY1gw$r6sUXRO5RXd?aIc;d%e+jQCLr+bI -KWHAw7QogW10=BI;weK1m6=BI*?OZDl}VnuNwAc!6&mcHmNr}h#bPIIm{4_FoZvrOfgUxnlm9L!3nA3 -9;)?rq>5T-*VltnA;z`H=yk+4;ntyf#oBW#od>NPDXS^Uk4roW?v`?v*9pKrRLyvf|BiNI&~x;lu^-3ULeH+@jQ$ -f(RIb5R%04jaCvkLX+sOo{3ZB78daB3y6fEPv!0ccqd?Z`o|YH3iK -_aw9~nC*SsX(?xMToLcCMM1o;#(1?XNC2%(+pA#bxS|pxC2xb*UEGd-w`F5{wPU*olY9MxM)uA{Bo#Xz@dwt3re$@i;;Uu@P8p>B%wJ&Hp@Ms++}Rli_&AI~h`Ps&F{6olYg%;E#f7S}IG&b2(+4U*0kSa!UYJ#0=oZU{glF^<~R_l?}u|tuO8G(IDt=x6&0kB5?#X_RgBag -~$U*rmNOI?|-39E@4hr;Tqn6*a6)B9aGm~~@G&7F19uu|?f8-qn)m=VWw;#*1_repQYZz6@*te_z0sj -mz#;B!h|Ef4Yyw(y@f#+|R3yFh9yW#>Ti1^@=D*O@+@S^q*cHQoCMP^N663hd6*CigMtfd&nx4|#I0l -Pw&lv0*#qx-e;sD_0&!|BdDn;$8`SV@Z4(nW%&ZtWPD{sYfy6`Yo{l{MKUJ`?Qgme;3pEKm#gj=8m^( -r{P1bXd&6>PN#@NAim=C`MOKiv3YEGKpq299SfX5RURHVl+g(=jTqYif{v=JN$4W|Qt`UiSY4apgu|>$On4q<^9nd@ql{6=vdL{(PYIuO>f;O1@=FI+Xw(J#%Q%}LJeC#C>bE00h?i1%&ScEJq@V;4(ZeD#Y%l1FtT)y7tMPFL#?+?Ox(w!48j4kL5p^xN20oDHHHpm@`P>erk;<@Ip(dPWQP1PlMfMerGsIE6PrB1MH^4|Pyt -`{3`F&_TbHxIA@Zfy?mlxei@02E#qXc&345}&eAQq%%3V1;fv!52(x8EJdGiNFazhnJ%X3g{98RmwB- -;3Ce6LQu3(~&HElVWyah@X$9L#?kNxd7S}>5#3y@Z-Up -!NJX4Sse{oSojzf2S~VJ-aW+VV9;4}3{X3EPv^MJ?kmMyo$vQ;+SB!x4AiR1JAg67#^9gjFR3V)aaJG -!$$-Ulkrz;cX@+XBxjx%6S{M+wj~Dr$*$_PcuF2z5DE2)P1O9|!ye9HhC06xJ -)Ylu|F1vmwmC=`@ZiW%BQ#S`gUF5u!*9PR45!G(=paahZeHz;dJ3Q-kp*`1bQPzHavW_ho?sJ+oDz!mJXSv2CQhJm7AV?|rmv&LA|eb$+KojU@T;G=zPIYd}4pciU+{4jo0DHk22ZSfTcW(?o -wH|+91l1EMAKI3}cCiyOyYtSwwo -l=c)tr-@9b9{bb7*GiU6P;Uj~HYum>EQYUYA|Hg5<90Yp4_ZwxHZw5>UJ+LUx|uDZG|M=@H7@n13yng<-pN_PaIk1{JV7G;5%@-jkSEM!{7 --+)@^i?yGTFg6u+~vT&^Lm#O@IRiAQcbEI2@f5guE-^wk33iFqs9!ouV1Ptl#|BlodF#h2)Iv7=zedvt~WQ+Y}z9B5|?)FY*ST+ -zWiWS-5^UxY8;=O#1bkgX?<)^#=#nLVG*7Mr->56m&OnjYjX!s5APF{PJ7}de1P4&Z?~J^Gn+aCHN(K -S2-|K0;bQ4ss%p0U_JCSe6fQ45)+i#UlioeucJMbiQ0BuTf%Ku?Y@ZT^*|sQP{g;d?dB!q@cVa>={;@ -rZx(rTU*I$OhDs?C>%f=QH7jrmZm($+Ik7`Ay8P7JDh~MUWt*~Mwx|LS;WNJU&*%~g?O`-5$GBWtp9z -^9PW`)gQkOF4P1D`j?NtK1yMqU$8~4pD&lgKHL_l&D34@#D`V_;5MASu$TqN~AWg|p3D#{Q>lLgO=i> -;)m*K2z|Nejz#7cJ4UQzj7r#EL3t93SL^q9=Ic$1t@f)*?P6?Rb}`?Ae^wl3JqB}#go4B1T4+qUfG_iDMUvEQ9y~hR6B9SJZqzEMZ -p0@KmS~S1R&dp1(9<)%$lw)|{je8+^aaWf7h)R#Fko}liU(I2;!$5YK?^oQav3Y*I9*};nb^lCNV4nI -jv7J0sN+u2$U(t7=xaJf -U}>~=ccsnQI*XEma$N}zx?rS$1pd%0;HZl=oDBN8`3d3VcFhhbR@Y55R_!ZCmvz37f&bg&TRz_k$?>3 -*=}z=ko+z{R(Z_9{tb)MUg>ohRjfRN)g7vg>d_#XH}=S25B_B(QNGKr -XGJpnODXE&nZ!E`{rarSS^f#X{uh?}0JEPg^#h(oaTdgh3O3rZ>Si;J{<+7qrE$LZyF)@bY1l -Jj0nhmR+oPF+lIuRH{272yVBR^y~DvirLjvVqi1v(J^q#vQ@f}SDZ!5s`ozRQYeLiG+h)VvTqF> -3fLwmDIkiOk-h`o$U2k(aDcvn!x@g1LTuR*^x;5U>VeX+)U-weGm@3sO#>@GUgyT3T1_FwV4-IySM>r -ML}>1dV@0DD$b_wHr$aMRHn@_9P2`N+QkUlVXA>$^rz|d^^~YiWXMgfck#A-Bn|n#BWU}+Mf&lY54J -!h3K35IPiDm$M4Wc61U(?c5nz=rUkJ>GJ#1liq{oL5e*9zOH-O74C6yK-=XfM_^l`RsO)z)&FR;;-$0 -h%AMUWPbk1FDeHA*#&`pyZ9iUnfnsHz6jk{ytd+;K$aHJxyGch=IsZx(#R~`e*A5!-bT@KH67RXA)bu -|wy8DbtlOzN)-BMhOA_=^1Ut-AJxS8}sdO%!@NO;N-0u)CHRL4&TK#yq-lQlNhz;cbwq9^1nhTIj0 -;S^l*JNXLZR(X}(+~QV6MEf|VBDvp|`+~oK&r)F^NNIUTOc;%HV=8Q?6a^kGEP9^Xzs!#S|3rTLoBH^ -%IZ*8i=O{b_mq1TN_+=ok(rUy3dHG}KC9mXI%+SRr>p1S+@sUv1JZ&z%)Z;E1U%GW(@$(Q0r3y@ida -yAE{8$cXwImNOzBtqaRQuv!)u=%>efp#XnK%#)<4FLKKFd4CGSA2#xKe1>Y)MyAEj5p2%iAhw=*TJqK;DgLxIJ)#zM0EkNI4@K2ZN58LBCA}kI2V^=KqOF?Lr&+H==Q>1 -?xL&h%rj7E1hKLpliu5ZeEjVbhuk?Fg0@@>F`zfX;X+RI!awVQts@ANe@;+3fcl!{!i(78z(sr)m)LU$l+70 -2*bQd}b!|k~IRSzbyAAHHXqY0w+BX5`2h8G;KcuzrR-gxczSXAMPI+mGRtOmpIxtvoM{Ea -T+8mdV2Yu@=ZiiMPqbk2Gt-n+YT8J;m^K95Tx%F7FTyEy>A_-Dz$5F(k^K-J3n48WUOgYrcv<6!P}nC -Y$|Nnq9(K7*76i|W38b%}@+bv%F>JI5!4t1?uVxRG(x -`J{5fa;1;2lPz<_U`uu1DSuvPD^k*}AQuakV0Zl=SMnHRRjku^b{W0UKWgJ(<6G!cEweSBj9&Q4!Cv3 --W6IjTeHX-uiPFVcjab2wLLLPwu?Fq84 -#K=3w1$&1sd_~DWLU>PAZ;WIMfm=9*eib2i$zWn%B?$I -R{1yPWfEyrt&t&*+>!jXhdD}w7J3?+dC(^g+=UcmY3-#NhIJUnY_HG^60(bJ}3k0-Ln%|OzXdg7Jz6} -~G%$d9}N-wOdf?dhK_;A0o+)nw4Vcnmzf`i4?&hzWQ;njrh^O0xFZ$pNNqs2F)jcOqS%o|?*J#?+kU* -&%=m6tdCzx}!H>GCh0dh7QWcTYX%XUnNy>=5|f7Js)x;D;^#1N)&B95)0i`T!fP-kCd|DO81Q&$tPfT -V^eji5*(2F4F6IsBp0{QpvsJhvN&2WHP%~3W{?s?#LG(km8GcYwd$OBnS&);G_~kPt3P#(c&kUzr9F@ -)E>w{HDaI}H%(Ej -#>ybL2yL& -_o+doXKd-tK46J&DI8FCKsu+W#f{A@y{2Uacd$L)(r`C8aM+iLQ8cKdSl?4YN3TM~@f#Grh%rs^&ly8 -FY#S4tL!(t+OxWPKeGzcOae?sJz&O0f*+OzZ?>7OCIjj16Q`3R_ch+$anJzyOQD*9R}PV+C?Y@L>qUF -7hY%)=5ivCY>*$Y)^uhYLJiF^jW_06#a%6&X$8VsdS%FY&hACPyIPfL86ZSiGJQQTkFh44Be{CZ^J2l -gedsH){=d|{*OsHm)-`(1Q|x=4?-$;DWBeC*BaE;T+~I{02!Vuo`U6yDbyarN?&{wAT%6I>IWkcYBGR -mwGuB#j`Xe>6I7t_8XX3f4$m2 -v0suz;*8eK?udbpLSTSh!{h52W4cpbWXP%EwnWU{BMG6JJ`B!K*S#Yv2_pnxqAJ6CWY?qs?_c@v`$GR -+ga6S>l*C>BiZ(E%dTQu_x*MhwQeN%uWW|m?R;!&zK^yWAaYN{l3U7lePC#JrOtL;5@HiN-RuR|U2-J -^_>S&FY}-kZ{V;7Eyxs+-+qMSoc7)Mi+6=L$ZSRW2Z|FWnL^!WQlyBy^`!l)^Kco9i$CoFb0qzgb*P-Uw9@|6Et_&6E{DMHhZDAsb!ZNctjC&orD? -Pz_7@=)vQM4tjk?cA33D};K-3}wp5RdIzE$l1hoXo2%KRccXV0WrMwtryL@uZzft@vwM|~F|NraU_QM -Y@_dmSyDd_$FT_1t&7kB*7Sc*dgxYhfTBt<|7iD4uOlh9VQg2NO+At*r-2tq(G34`QM;~xZVaGTnVH3 -(z_${@T+A7{yS2(gt0lN(%)cb({<$m@WJNXC1kEWG$YqljzE-dpUpn)rcx8+}2|C$)|6R4p`kUzGc7RXoyJ(F05%2YVdGLNy?uyjbxAnux$M!n#!vOZXbwiGImYpueEg#WCs3q-18DpP+Ddsq15ObM{UNC?py+e<~vBsTCf -N5ilfkRe0+6sYn2M$95a!{vyRoM-^4=xLt^Q}10UQzG}dwHRE8Sq*rKB!=SHU!N>3j!qW9znh(YG|i~ -E$&!tE0oJm^UJ-9JrAzb}-n2lOU^)Wd-DO-jDnFjooD5BYy*1clj4N?|dAN1IKE@$1&6Q&f2WJ2xJv}NpZ?!l7Vc -&a@d2l^Xo!|rffWdMZNZlWhl=p~}5vlW{oQroAI`3m#TqeB@B49$OX92m9-MhK(z^@tNZY3n}5j#q{+3@K~nxM{)^wBDuU^-j+A&>{T7bh21DAq&$c -vAT=%Jo`#Zn#okIhV9QBeoe&OWRdN~WLB~g`TRs^+rpJ6@9`rYs1#9E_XLklZ%Z_~6P=C7neo`1QuT2 -c5KtcJ~3E`ZpZs#xNFw*BmSZDr$zX=$kOYQq&*uw}yi`%Ms*3%q0UHB3BWqGm?c_~cXaS -aMRMvr(tVXBP>_F&_x2oQ`yCb}nsb>-e;?)gfjC&lOE1&13CF@i2R*8|1BF1F<(V*XO^svo|rk)6TZW -!VbQ9#@N2)<2>nf@p6=dG-O0^64S!3AIIJT)ouEOqQ+XH145N8Qz_cYoHS0oANE=`ULU-f49ngcy|>O -^gSv$OJ+_-(7E8?>z0T|388MNfQ31@aL4w2U~DW`^~4FgE1ESGAZk|>atci;7m6&(f~X%#!7f?qz((_ -+$+s(G-*IDt(%Y*UIfb -5+{b+Q}p=#Dh+o{SOFqZNN2j)mMC;q*a%F#zLp2tfVXwva12t@QMyh}I2Pwzw*ZgWdgWIzL6XzM;_`} -J=DzsJxO^BgToEsKuvS4A|V?hr2wr%m~i>P|a8F)?X#=RBVXn~@VYXF<9_Z{iJ)0VQ*mvxlbOZNd5(h -S%5}Pkn!Lo*Z&_WB(TO&sTkUF|Vh|5Ng~}EL%5M1FX5jJ0!@D?fgEZ0VAvXAj(o^Y3- -VDPr7US(h*oB~82t+#hf7x3Qj9#Lu>Gv*A0w`As@A{Z2R&1|ed_a(|5fwWjBPYt9dx__yZ!p=%7mNrV -Ddj*KD%1>!J)g9J|DI0B+~1&$=WVdrntU^MRN==TevPomwI|0ROC%?vMJ{>SB(cAw9h4^9ualL1XS*^yg~n?vLu1oV>DTC;pl -ftEN5wRcfPNn$VY=m^|vser;>q2QTH9$Nv7Qfq(MY-#<0*Pab=F>W@?y@bg@l+CYA#+Dw(*gxB|Z=YnvfIl3zez540XRMxMLDmjNc<6KMda|Eq^TrLjv -#q^)GC98@a~S~yliz2iKmVqUdP{RmO_z*MR!GO4>+Jf%@>AU7|tr?t=#RF9GG%bqpa@AhZ`bZ?M -grYx88xsqG>EsAp`TEM+1KbKh?)|vfx?O2ZXr=n#mZHl5#Vav`UY>^Kj=z|;U_gQlA@d@i$>)$J8pp$ -T2l*S8=YNE_-P_{79ESh1IsY7ne`CHMNIXap7)lT;g@vb?&qv_B6I*{$O@Jkp+qTSpa+4M!WiTBXA*=9i7iNH#~*I>9Z_%PiaGJ -j=}z9}>BqRZQxMqAcz4azG`U-yR1$^+Vt1=ayDym4_1**>ldNu#|UBKvrnJubt0b>PbBNo+frSf4co| -B}0RTah02%_7~D2IITgGn}J#o^WYI4wh4sUfSTR}csiPK&HYtFxMT@}2uMiPQ$Q8w(b9HW&FNyUPTr`+5t-JZ#-aCxf=Q -1W*|5C(ldQ;p+}cHJS+u_@Xb&8|le&2xRe*|SPO@_l&qm{fy_m!gy~i*b8X+!zQs?oEJAu3v`=!-}!i -ANtMd4_<7xPbgcO+c~m|G?O|dwPd4u#7~bWboER+<@gFkzz;23Q)^u2sW-RWoO#hSetw;yM@OCdeCVx --vLWeYQG%=>N^(2AJu405Ab{WQgT;XhF8hHO+YMY`5aT^-&3ed>64VadaPUOK^f*Q5DY!ZhJ(Rwb<_KLQEgP -?H6+c|lT+xMrS889ym(?YM$kb)BB|hw!Z{UOuP(!2~W?j!kqfj^T9VN|1hhE -rI-*q}aD!-tQ__fL}?9vGEXOXgJ3*8IDU+j(YQW6NBCt+7F6i!1t13KhY)rLc>q<`&=U|AB0NTottV1 -uR@eNZ5ObBy0Tay_X34%b6o9EoO$+6SX{jht4%^Y)5>5k-c@&(dhn}R<-CDZEBlyF?3siN7RWXIZ1l; -CcaC>-3a#6D7;X{?CMQ^VkT?cWX?7EBdxZkGr&&81g}m6;cvk4vqqG2bue^Ka9}iwwbO%9nb7xM2<%KQ*4zNQan}Igo-sX&>7x*c#!V_B<$8J8aF^uQ)08OKHB -v_}42&{e->K@WKy=WA(%L0~0Vfo*g9xTla>j}C8?@FRPZS@KOX?O1B0q0${rXzCLE(`&^FpE}&ZsWkl -`-XJ`Ys{B<2{~V(|BcZjDfE+rQl~uEuU6}*;71$Re@8Hy#6H)aI*4T7oWQDq8%}(>ThCii54(D>SIYZ -m!RZx(_$snTrOisf_Lpv<;ULfD2XAt2`^7q<`}rQR+5+sPA2}%?kqIAoofYAo0)_JLD`?qpxINmh!Qglz0P+yyTL3~GYyKemY*jwS -%UirFFa%wAh#dctr?lSV5d*)xNue0U`&eCtT=o;-=GBHb34OKd2ny1gM0-D5sGE=K=-gg3U5=YE?VwLRt2aff7KXwAK;s#;Z=fb3+}LDDahbV&~V~OD(U#6&sMm&gNr=2_P6(jf` -D&{l# -_7cp8s5NEqz6l3rylZGz=3g!1co=EdHHZLPc<%qj3-N7VLl_sX@_TOJlMN0P_6d7h-Pw-fXh4)Q-+c0 -(0x_g{zW0s&x(dtRt$Wnwk`3ND}&*aN$Gidl7u)1-21)I>v#4~YTMt&0zVo2BUR3IWu*fmUGVhaWeJ2 -nJa&n0x5~jGy>5XrDq|V#PF!NXv-GhLDy2uKoRv*ZeMF{xc%=}oQHScdEV@sSXlqY3}MF01j -yuYrma!HvaGSsx6E9zgBK(xc@g^ndSbaWf;yDtVH8D)* -b(+r-EBi3bgfhZJl4cc;?FID0Dl2Tj?OaTl$d17H+8Z<1bVXb{$L#*+OOdMXZ3_4(Fm>X8_%s&{AqIi -b?jGon))cNjGz#Uxd8%eK+T!bjv8OH*apfE7c{po-lY99nO&5qY~co$4Izc&*I%8Xs3pY -lFs*cM<7tAfSRk4MZbhBFyp>Y -C`Sa&OPEijR@16RH-h@LYmgMay$)Ufc;FviD2Y!(*94{X^+*K12%aZ0;gG7!^2c?8}y*A4mMQRw+gy; -L`h;t`R@s48(US9pkMO~%X-gm2ynb7A)^xoTaKqyQd>m}8y+@RA2qzGqb!G413mRU|w_j6|fv>&O%Qi -wZltKgMJ0rU!dq@6T?zgpT3xtmfaB^X&mN2RkO&#<`vwQ7A5{OltY)famjaqx&C1; -4Jg6f=^U4xff%7KtwMFG-wUWwR`Fa2} -p83-3T#0oLiAYh4A4r}-|cR|?`&`1?QX1WL-%P&bCkJ0muggigtYR -ewqEqlgPF(hMB0jir)39J*Q2OGSSy-_Xn-2CjLC@6c^dc^3%l{faCBK2#a} -NAqa+0Du_stVfg6(f<^Dun7Km^Hg-YqD#Pw;`(o>h*GaI}|yq5-5rCG?+Umn_WEiV=YX&`xGXYl@rIw -!mGmS)yKSFkk9)*f|Wm?Q1qk3qyJ&D2yTva*2Z!ox4Ks_X2AP$q0%VgJH7x|=f#x;l{fjitEqbuEAx42!X7FE_ -?u=)}Xr?|gpGj@@3ca{y{F+LQ$9}W^;7JkdYJnz|e+V_#}s;kJMgN*91a>QO9HTHZ8J6N5XgacjSIJo -wdl>mrTRCF1X+0G5MvCQ|l1XkIaxBlLo5BHpqj?c$6@x!Zh&&L)a+dv7Q7{)3bOy7bEIKQr;BcYnEHQ -1F1my;kikNMTvS*HhfZJ==PJTSXdJz+@ciJz%s%8Ey$h3zMZU3dUeYul3)qAsr5TWm^OM0HS0S0dhZ2 -ga}wi^zP8ggVQI33*2Iqe&J1Q6HEjg$V5mm==>LYb?Kci6ZCm?`H#cmCP}Grym;Pzj8A)z8fmFb+V#@ -{{PvEi$8P4|IUmbaO!tw`ym%gqASKC5tx7p4B6Zk*Uw=Jh9MmO8Pv+iO{Qs6CEtx9qA#sAUrK=EttWr -$ypO&bC4xWq_FC^FvkmmD)H2;ozgJj>?oHlP+_+v)vEIX9CoYGZ^eI@`+VewBb-@5gqOx5BfnTsq -m16tX)flH~5{g>QHe#kOJSZn^m7OAy?3J)!**d{gvXKd!5Q!vBS%T@@ht4a-FV-k=5fO}0~Gve>u{4} -}DMNPTwmPvMp<%0ElM{vndN&SqbxKD=P>@J~L4db=Hm2QWDSvh#&6 -VKybJiJo4T^r`B@Kn1gZ~`%6A^tU}zq4BS=EwL`J4WYw2C7Yf;b@@dDBn>~ZTh(OvCeWmr_;OM!K0CeeLtYT%T>EGA!+NvUGk%x4CKNQ~4Pu -Itpu;W(X7;Jg)4gnR@BgibOz33oMvn$vmL1W1td64M$IHoH9T}hHIH3226tV!G#g`kW1SW#{tj}~!O& -d*zizOpvIGO5`PXF%z$2cwts==ny4me9G=9TedS<_*D5UH{-|INb@cqR&-yw~RBVl^2y+2qw+t+JS1o -8kW~`X$r?bNGBcz*@ZX@`kg?cz`S=50x``6zIF1t#6a|oyD9gzz<>m&*jegQz^=AO8rQf={mP@oGBKjgD}l+HyHMP9*q --1-+}YXW4LF=FIs;739dCLz4I!uy(~z-z7$lXj&Mv+Ra^R7l20Cu$Pe8_jANbP8!u3KZ@xz1^-2i}jZ -T!P+rbd2^Z0GiU;}cs!t`E#*{n#;hKS5}89ITJKJ5XB8={cd<9*#oer}>+q;UG#{`^Pp*8;L!y6UfRM -eq`jqlQ)-&-idq^$cWhK9(O#~_M5vRG~&@x%Y@h*1^bdR1TXvS09wf2k4u>>Gjqnh+r_wN9GHc{J%Dlj8PY0xl -N>ZWT3knCTpT~1&#uOD3zV(Gg3T1@UKsse?5U9{1IP!>`EOxTlv+S%0%tWlH4ZYXM!FI@?!7 -vr>afD^4(z=>~X3EQYB!o4(`>xu~=(iN{TIFiQ4lriXt0qqEDRoX;>NJS4+80dva2jnNqKRZ1g{yga1pKR}FIrQCUgFaLe$_@ -7PvOAPr})Bc#L*iPs`5QR{RKrs+UNRT2RXvLF1_wSB2*q82yX;`u$LUQ*5OUbQc48*qn?G+J*#FjF^( -4Y6fp<96n32&ee-CNu?m$5Ipr_fe@mmpiBAlwbN*SX`}k$Wo>%zgzX(XXPZ>!et^fl&9&ydW5L0ngoNlhlk-^8y?=De;1kx$^VW1QUjj#cU1AM2Snp8kBsHW_R3*aolwc -2QWdgjeHMHA5n(afuYjcU8DIe)U1z@mByo+)Kbj||4}b-HLRep#=QtVCi{As4zz3>a0VJe3%x6%!U;B -Q|pO*~$-8Fw+GVpiT{CUa1@2~miIS}B7B#1;1k>QspqRp)K$g$dpUyZGcDuY(SiAJ1~wr_~lX|wZCV$ -slA9S|<6Smt~e1aOBewxzZbgK>tf#SX$HlD7f_ECH -h45$pV|ywQzNUgHq$=NP?O#D5r-5fc6XDk}fussDXc{x{S9pgBZM{5D -CAH3nl_ZW$-{q;Ilkd4JiNhi!2lA-4@7-Z}j`V;XI(b;LePyw?cA*d{2nh2CsC3D|N%(bi(P-nb6llG -^ZAqP_J)r(5bc!S-(Fzm(u(o1F{V-=Z?B!a@ztfr8s~$j8LzM$`XaL}d-tf0OEeAC+O9(?3V${o4O8q -Vn&o`R6^;z_+Gp@f~tk7h*XY%+Mdu&N!%U5 -1MxXwa!nVB5w-vDG=eJd=(P@R~ec3b-&Q^ae;BIJ+spyhM06I8bElWAHXYKH`Dm)$#&{Z^!e_}jXQOr -Z0OcfXjcj4DbA$&VQRp4fxP_Gx+aAJ -W4L2o$~PTGajbc--`0MO>yB+DG-M3C`Ixt76lT{?U3fk -OFCTL9uJun>;{8M@4-iaMLpE|K2c~VNc`Z&UrT#WH5&zAW$NP@LtdHVVkjAnY42sn_Uf2o(IUz#J2+S -fX1j;#&T3k0|L!`Jsind1A92K?J2TPGh4&P8=@Ke1iDUYoVW#dN0*HL?0^emU5h?XRTztNA4ZeLMy0d -Aj0Mp5xAp>F|PURg~@63GHreD`ks}XKWBaFC7QbE={Ahot5@*UV5TssgOQ)#_|jv| -G=!G|0YiL3rZpKBn#XQ}O-WAJjgAMxVO@Pp-4LAU^w_e_otW;vbv(RyA_ES*X5RaJsP1M6d`QOxaxCN -y=aE^rvekHdU63O%yx_M&Y{8ym`BWHPF2V7 -V1Z&5e;nzyjfCX5%Jzn{;@G4Y*8pjPi3w>(f*rwPw%&LDdku<-?8}0gn=+^w_Di-I7Xr=N*T}u$l^(n -;Vv!ncqj}(+hsYz1!OiSsDFiIN^#eieLh_2d=>OD*6#A7K->@(7MI=1!A&d^N%@Jki}%71h6Yb~tv|9 -Z;!X<>o(M2`&@)*aj(od3sl*j8Z6(4>qSEq&&n*C^{h>KeiP&nd6ZwKj&`yjlzWPB}#ECxEH%}kHX(4 -WPJbC@W%Mhj_a% -nDc^%S$?y@=m<}t-$b6Ga43QJg~adIB6$aYlqKl`O&r5RE>p4(4&$Y1Ssu8HsYaV?0_T(&{p?M3s0i> -e;NwZW0|(Weu(t)xQSIEXo+Q_;E5r#Z5@CXy3o0Vfbn~6GOs~Ru&YTE_<`$M1&go6>4(QONGU9Yq6!c)FkM7lBkHewA0+^g4al+mpB=#U -)y^?Qjx_fYhSix{Uw4!{xiEXj|;EqvpfBD^w5Pk2=7Zf}Q=;34l$qf$6*c3}AM1-@Ec;X!?%f)+$u#- -pWKs})Nl?s@fCbX^Xu>inqO0683AuNt7P*R;rz`L{73&Hl@x<=|g}t~vi>vD&|Ee%*vP!{Y7rQ_ --~ydqrvAUNzj!+bQioA-;b;%Ll&ujTwGylZG%9UfKGJ=nxEnHccG@CO2zxf`mW}gVz7RpEu}H+hHmd? -B51F#+|_{BCy9vtVz!;!Lc5SY7;a#WOtw$Ly%BGPbQrl6Q{uapX0!_kM|B#&qvv+Fx_Jz+XTUhE7x%Q)mMpe^*X+Li)Q -5`Anf&v){R?t>lNb(?hIL4na_H;@+-cbRF3XG3a4dI*!qwe+iUZW$yfTi(kalhy{hI{e7H`)IF_4uZb -doY)W{rp16-K4#ES-f6(jq44fuL-Ou0{vR+}1`^My3-j#hkNIG#DUm9A?hI&xoNp9*lax1r2BUU*+){ -Es^D^HvAGPBOTTjB2GnYMgb_>~TEhAE;GlHYs7ibn8(5ZgyGF1-A|qi>K^$f^B({=P76R7pi4{Fz|DT -d(c^)yLYBn4_i*{E7tQka{g`CJF(hFpM%V9OGA%gOG0CYgq4os0_mx$HxKZF=O|Ji@4G{1xkxGl0dRRe=ALL|pG|dbCjHh -D*-ZACRs$~_eiT>98yz<|d_B1seyh0p$>` -D>y08YjQI+Pwd#4?hLl-@fT1w*KUXA7Uwn!8k-^?FU#1AWMOmP4;qi&*B~64bM@+9sQ&lJo3C?K -X(NqaJ!S0BaJ`-Mjlal#XGN%h6FRJuc0w54Bu@y%=t@?R5W+z`RQ@9t;P)B$fWs>=UxC+(GS&sfxC{| -p#<`5o@XO@XVm$@rB>yS5zTB;Fe4PQ9&bf|)}bG^j}c97B0=7 -AYP;6dz`Ae-2x=I`>%sb|Z<@)W9toH>hk`x@lsxs=}JiDZ+5<;>zyxP -O8E=K@HgdSHV4U_R;u$@bDCsr%Hk;79@E#~n5BYo&Fp-nA4NX`Gn#-nvEHpK8!O;~-&+rs_g~nT2 -QWAEQZVaHLRaMp??P?&0c?Q}1hcYs>-e&9~HK$oYQC#qphYEiQ7c8rN~>vH<#{Li1ntCx32F20rL64b -9IO6{Sv_$H#DS(K5W)87bUXJn>5cD@`asoPAbRhbVi<&y||R?;%gxN?x3-c|X{eu*52T>b1~6yy -=y4L7M&5!0V{g?PgZC`Px-Q#tiQsSd#Na`TRl0tXi-rL_7sj<`rwFWxs$Jtf@vBHb;|o*9$SK?&(PB| -sAh7u8n`v1)EtK+x)X;auS8l*4vY-BMOH&F5m7oM^TY$}i-jA+qb3>v}%>S|{gMlUWA)Jlzt -5a=bWkwQ2+wwfvLq*|U0k1L}e9Ny*O?N8S$s=<+)h%pZ6NE%}lTyaRi1?__)A4h`Y71XDsf`cgnI`LG -Zq{0B^n3?$bQK#|#_-rH6KD`GDO%;+VG7tCh1m;Ln0et+-M`b88{&5R;3-rTX -_xsa|pHKq-fV%O|Xp$l6zfu8f_D-6c-IQe!EVLAiLv2w&_ntJHsS@sTs46SOL?fdc9N0&88;Zgkdmn1 -4an6)%Q@@_Eadf(dC5LQz7v7t}(tn(8`6^=U@L(JIsQ1H8E@>tba*K*{w$}5PyVFrT&7{<*taO!1-o! -H*VaYvdb%a0{#>Dw4o>9SMbSl{u(}gu{Qn>;M0blfZu>m!Umsy?KGWyez#5X-Ho< -<1ME-;g=)~Iri@TDmhn8ie`0b94vg6gdft!rr-eT=;Vwepw$KDsT@l6R2t)Nl1oF*PuCJZ)~ZHV^|O8152liF0UjwI3iLTy$Xc>eJ)7r!WvKF*E8M1$~%32XP&;jM!$yk -StMcduO$5=`5AA?Ps80utV6A^SU#W4{TJu`c3StVIe*B$Qfm|h5p?r+-y$IT(>g~O-+~B=Zmer1Ses` -5-W43dTd;`m?Qm@)hUu1VrG8#fgGIaJ`feH-$6NMuQ~%%f7@*zVa{VOS^{umbBbV^*UxR1A*2~JGEt# -6e`*9w$HQH>YKHzrqrV;^dYFy|SGcR(h)PeR29%S!iUSlx0li)3mCtHn8nr)?#yBNbh;a}57VvBv$w> -n3eOU~t6T(FS%xf6Tto8PpaH+dU}bz@jtWK>_=N&Cr}ma@?GM+A~8!LSO3Nv_tG53Kvw=U<;db3}Ka)d?;-6yM$4-6p$x^Y`B9U-h&`!=XXV*#fDCZhBgP -wp8Qpdiim&WSJZHog}xVDc0gITA3rd!;haTYlKbqu+doeLu@Pk@~B(Rmzu+ibqiRg3hW;e!{si#^|hN -C_+qCyFiz#~vJCEBc%&s?tZ@CKnkLvM74+Iiym_6>Sto;otNfi)&fjQ~q4oDOoZt6Au)X}jmO$K@0Uf -|Giwao`-th;%EDVnkb^h-8l!+d<#Skf3&!_SEQqUx;+}Wo|3BM+O_Mkdo76oG{BF2%CqhpX@f1_$xtx -+n3A92*Y9Vky82?okadLu434SC2{o0y+R_!2hnK>}benaS~fdo}^j!+5N^CiXrzl7Ub -?_0t8Q8PwQ}UWmR7;#qw|)j!K!PI92*_uYI?MU(LxJ9E(n;j)nNsb8+$#z+R3P^pY{Ss{$J0Z*08 -eeim@iO^G|yE#o~oG}HNcq2;3&2HfQKDSoFrbh?#bE3F(LJ!ekR7v=(7hyrRsM|nyVa%RAikWY`tOO= -^wWRKkx--Ak3C-$^H{1NgKkn?KA%Azs$M(mV?N5C!&mGVf&r%qBl7mz(Zp~LCLsuH#+j-0Ttb`nsYJ` -w?ZeWhUUawj=@d$}OhwxBeCsw=m`jCGF0*o0*G7@K(a#6>*a4z)Ei&x09bmYI@Hz|f}r&aOs-RI26}c -ZKsJ1ZXhHrOBY5R}u>|OKvK+BzWBYTttWCC<#kHh!Q@sQ+2*SyAyxbJVPp>VYVy{oiPS@kuCOKeDZX= -ccz~hj~G+6rb#NkK7KUG?4lu`9om&C1I4KXTIk@+55s3quy(`U{;KiFr?%?&N-@2U>sdv1&YdoXQci1 -CSQvew4d8E3=ASmyZ04r4bHlNs5A!=my+*j9!+KKOZc$GVgc_RUd9s~VM^+LYun*ei$+;@IoEhlm98` -dknIui!Ye0Zd{!-_&j8T^*QZFq{pEA`r&C68ckAVsFKJJqXPEn>jm$hZl6fUME@EkFZ!)V2SQR*dBVf -QG}dTPE$+on>RGGtaJj#iuZC(}?RZb-NT%44qR#$|h(qZ9BDhE6C%8@cLvbEAEvoEiL;`y(Mz8Z5UK}Ys!#sqvyzQj1%!5UV7DNOi)n8`Kinb=)!~xT~I6yfaB6 -a=gxEOS_y2Ji-lAFFc#*k!84%Gk>H&~h7eu5K7x>eLrWBkiG$ZhPn7cf(nW8pp-Fs}SLlYzrI3c}-PYo0HUg3dm%#ncdjWZv{-(V9zoWk -=pyy|ctQCkR1&t^u`f;ug&z0-In!D<_>PR6|IwNi!5@cp&VFYia?N&IX@7GxRoGtasLcviod^bEh)*A -Zd{&cIeJ&+XBNtc$$}A^7`GqrW&FfEu5=7lmo+LTim&w63Uo0ziRJ!8IHdXAUEvxy8nx3|7S0L-tk@E -r*Zo)E*!#t*>gXn_x@k@|DpfS*9`17{1;_(pFre)d+rY)`FY+SNHq>_PEi=X8N(6~Mj;r55}UNwN~I} -;!axc}@t>L>eua(53K+MfC`D|UfG^gp=q@ph>^e0oeB1O{p`U`vO@U^w#6x!L!DLs%4&e=8u3zGNSM` -buF??gaDYhZPb>voRm;6ffCfahrD>EmtEklbT8+Z)EjSeT!CZxJP%oo6zgjnR@ayPrGW|?Y-l9?W`1a=L1aFz -aCc?j@vL($_{}o*yKBIu{IZsRx)k7JE&p^Wz&ASoi>3T~`&fquPBQ)!*L~o9$0{8@S4%oSjd!NCQEuc -_bG;S!tPPwQg&FiN3n)_|GoiGG2eVv5ICnsp$*_k&_JUUBS(7cs?wW$pd=g@LhyX#AO%*RYuAw~39yh -X`&Wt+}H90`WfRm>Z_$;6D%U)O+6obaBlJdC}SxQ1rNXqO33Gf1?{=z8Z7gV0+JR81ax026{qXM`h!6 -@S)zTwmhvU*umVfz$I?^+R1uxw+`uM5oT73uk?!lM8Wy183gnXMfh@pl6>@0V52NrI?Mq)C!-E%PuOs -v0{y5kF8a2jdy){YeS}{}mVRJEM*qPyUV8#;;Lr?@URP@A>wBIUfI)-{DhK{xtiK`q(4^A|wc{{2oRj -3M0t%D`MS9Q5=F`1jjH4{b~8l%AMC38N-{HSC(!OZ4+Y5w$>GJC;n@c#n6qMZ>d-6r=dCC4@fuu4elW -{-3}GwUFd9$jS;bNcX%)AiT5)}6x$Cl!!=C*itnde=gYcVV8q@>gKa71HT92O*iR~O)l$j1L}T9B({m3H;e5un;HHm_f7 -ZwT+u<4br!h_XNKoyjfYrB{=EjY;y+ie9uD^w89@ZRrJC#^I!Zh&1#gf7Q?vj9?Z8?8uNxd^84W}WPkPS%mMRh^ -<~x|bPE|r=2F22JyPOUB*sWZ5Lix~PX4#9C{OX()2TL2J-X9P%S`;tK@B&aTlD#G{q -g^0?#;U0#Ih#9cYZ~^tGk(}TB{#G%#)a>H)aU|5@Hm;zCljAY^PnBRo~aWI#;Ii8VPjnz0xM)?AQ^kj -K#yXtmnQZN;q8=#terW|Hvy0(f8jm?1>4`_BYH(ANplhc=3IfY07Kh4HdBs=g_D;t(O<FTKCcq>ss+pXniebXBeRUTN)XN>5=TI8Wl_1cYeGgr+miWq3#I -4dsG=Kr-WWL{j0*7XV9!ne2AgCP)txEz3W@2CkZu~BCPk%d=T;a@MLJ{WW@FanLwYEMJ0|&UloX$O_U -E_LOFcsgEKVV-`MC -I92zm~rfl0iO*q{&s^cM$<+HoY~DEeOr;5=Zn7V_=aatOkcc$?37a-z`0$xN##GU@n*Pdg8{+XvJ|49 -779;m(3`_bK=_Rs7_Bp!M#h~0;@u&9a#&`?CO5?9gPIL{AY(_dYZzojmdAsFp!2raO_X(I3 -mwWs#VX -%8#0(fDvjL<$#MODr6eckExp$!8DOO~dD-L2(%}^7k?#&Edq>l5dfNk#m(fO7Jc7V}D -HJ#>lvFEPAkqYW((lMp1hsuBZdRaQxS62__*uTTYP-sMnT#9B0VYY!DCgWy@8P-yh3SB9O^gRk6D(ahfs6vxVR;4=*gZ#Myl>DFPJa`8lZ-A%782zyA -<v0iY)lHs-@-M8&4Fv92|H*t0P_b08V?gOyC2X9r;WyL>?4>C)d;)jmR2d -(kVjsRbe?$%Gm&*Rnq3)lNLQ~lGmexT*heh0xI{E!45K##;w3V~<>MhFbw56}cg!TT3S!YPE1zij$Yp -N>gNA7jfqYC?&_4PqzlJJ3q0qZ3q8hsYWIK&kZC<`!9bc+Qd^lQO70p1w2a{R{e&@qA3Wm4|c>OAljM -1o}+sVTbPftET%;AP^}IObJ7Wbn&n^%RY<2KG9kDVeNJ#H`AXkRd$SU6U6ZXJ9bS!aV+~Q5y`>ocYFv -Xhaz(4>WAhLIr{Jz{FfkbIGR5AH-7H%i{@i@R=?j9QE2mg)FyIhhZD%^zYaK!ej2OudF&W!gTIQGwY$ -*80ba)rdx=8(nIiZ#6}?Zd2lxdt5>nBxb7==WZL{XV_kC+!=Y<^@HDBhBzxUz(*3>b$gL_x7v3bZBpl -AJftlro7Q%H*E);y|n-dFkfWCeMXwniiK_?K0+!muy@=ibd{P8pDSf7nqWl(qDfgaRD9e5|wC*AY9*m -;C9=1hn<9%p~BqSJLl=1A;4svOO`CT*?W-Z-nQ -Qjg@^`1)Gm04rq5&|+BC)uCNd0=f@ozvcjmg>q*ztnD;!RO3+i0K6$-T4 -Ft(^@uD8=oJR{5n80DpPPp7|s!`gtOSDdAcL0jEBfBL7XYB54#=sgOHn$O_V7)7&3R-R-&Slx!&+=`k -QaLc3AQ@ZB5D8Pu_prA}xWcN=vF>F6=oYageJXMj1ASG7c`*}?=$vA>@gM&d34(A+@H%2bV)UGD5Sib -Q%Qk~gTm-ke5cvlYK{rb(hrlGC0lGE3ibFaH+!j+9=T&I4XlLZOi`vGiC)Gdb`gnIwz^RqOOLdnJ|W{ -JxH0_UO>aA^S7bNLc)$M0>Jv^%fyO5mBZ4yywG%(z)vfzI~JdW5#;Ec$A_bLJ -r;XK#X`N`uB=s&`u)KVu)xCertNxSN=4PkpE6k8?fFuguNHb^{F?9{ -`}1_U<3M&f?d~l=+XrByd&^DOASKIwDcmI^*LQrnIc;g+M)@t_`AgUU2$8_i(~CP2IoPCEOX{TVwH2v -He(%MFy(a;92E?TR7{H#77?m;i;X)z0EA)m`NdcR?Lz=!b4#8W9b=Yl4zC!7Y=?5$7#q7M7#na<7fT8TH4kH@@?CIVkqcI$v#Xd-fXD;qYHd8{WpMt -23g95UH -wYD!DWjbyV}U4#R1-mMUG-QZnM{GsN4#ExD;!ao{2_9jd}J2I3RZ1SuY#{Aly>f<-1;)lV)0!HkF_&y -xj-U)Q{g)`bAxO5&fynv;Z` -9*I2C{V2RZf3PP&+XxSI1_BHa;dg1|@We9<&Tcgz04WJzAM`xi-<|HCeZB6(%kf5_c1-p1(wZo_`o&t -dleyO;i!c>nv$evb-Cf`UovGmA&TC{Cj!L|{8iM2H4hzbd30JN&#McnH{j09B6Y4?DEm8JND`1&nJf`s49+6yV4HRs&>+f6ZFsA@j -|Yq4qPE{13v)-Undf@A_H(3aV0v7aYesL_oIgy;;YDj_tUw=*l?~4CM8TV)EGElh=#OFNW)FmoJJ*`1 -APo<(L6KmXzD!><@Y+SpEsdUYk^0cU;)iGB4CoU^i*7e=Q1Ad}f_EG#VfZl{z`IO4=Dv658+wJa+nlP -*;UxZeL-n2whq4-W(|io-63z!u7t7m3iy`iH=8}&p1IhK)d)|Gu#bD>nyHyscn+);`z;&Xl}ok9mBn@ -1sUx3ES$V6RvX5YAT7hm7V7wBx)l##zIX-a`U;JAIw++`PNaQuw30YobLeY=Ybh(H1Q^@d2A{y$PjhGHR(@X -j4@AX)8!1)Ybd-Ys~euTNY -WVM62Bj};tVm0<(etpTnR8C;%r9~938pWj-T_kqwWTp0$M=Y)N_W#b}G^}(2JSXRF%11Wagf7$^8P#% -e^!(o&s5O3tp5B*)r*K@)Kk4uPg$X$+MwIO9hM5Be!JL?ww%ynHb<53_sMNb0zHbHgC+r%9*MwN_Y1? -gxlYp@9}c!S%49BZ*q+SwdVA}dy-h|%85sy=c2Yr>7D~CkmJY?UiVqU42&urD0yExQX-72Ge$kY6ENz -skEq3YGcYRA*Msn+K0sOS+iMb|be~evMVr#6wW$xox-zfE#Zt`UeTSdVc3~>Cv~#^T| -C0k-vfTx;A5V$OvKewkyZWFy8-2g#LYKmTbq7{eZU`L-{0as^U3Z%4g4lJep@V?U(1uy<>lyB<=h#J2 -ZS(g&A~0WtCq(n&qr$A;fu#8$ZRPZ=QIV}@{ZGnxn4=q{)VLnEc4E3-PJJTZ6FGic#QD6FFtdoa$pWJ -+##X`zqjmBV%v(Emb)B{P3e1T{}L|3OYsAvNnSV;=xb6Z!2bJu&c1H0CYZe^<3!!x^7WpJ;;b`tbhaP -k%2`h*l}*mxQthP)=GjB9s=GLz7565fY#NJu_jm@S_wtw?=lZ^?noUP_G`Pi5>STP%5%}bp=aL&UyQZ -jUE%Xk&r866&Ca`U#07-LlA(`OHEX18oP69ocW`)HC(@DLhyyg3h8fZe~nIkapvxZXvW-}F^V`l_vc4 -y1DLDNIH^_^?7a^STL2VrxkDq;X)qB_6zkQj!x@b7g}lt_?YgVO`UgH_WZ0Mu}a-b-e0ZoV#GALKAi@ -4)2q7cpGe`WiNy`Kli`ubAx{X>RJKOTJtZqb9or`E&tRl-aJ7823b8FQcQ$EJrX+*^(5mjW{cI?EP6A ->-^xK&m?xxeN`1mEZ|m2r0kG1keIiW>O)g^ubGg%+GgQ)SlTzwt?h}(22DYHT*lT5us7%~*A>sFozb| -?bH2Nd@8`?@Xv9F`FiulAj!_VV!xV{;7(`+ifj}^Z<2Z?;G(r3d_$9@$L55|Ac! -nU3T;%){3%^4_>f;SdqyHqa!=n~U|CvI+cDT(Bad`Hb+}&~Aj=yN+53kvyjr)nEqVhjd{tnG|aqI>{@ -INu?cX{>Y0oAhX0HxT+?J@gAE0@K`50N}z9`V_!+@G*t8vejn5Cg9mWphW&Jkxq;{c^#I~1$m|N?1Dj|g4@qhiH#k87U}%QpzpwZz<&w)zE~!HQ$7D0^c~IJRsRw -69mkf7{{{3Nm=E}$fW8Cs0sjr?`##$q_!sE=wb8;ip9QXx!IxMyqe8q?TDgoIgXua6sOoFYTDW2OLcW -^sL}{nqA)f=w^m2GD-+ok5y2e0QZAbpv^WN#cAW8RbI+yeI`vKs%lHT%o#YZ^|p{jT1Pbq))`d+HW`N -^_@!uVNVT*u$|EC9a>w|(3D-PG1jtli_8iiD+7dwB@dyP5)#=hFvhE$5hk -<#%NO-RdBwaJjim{4C4oaL=ON@xqqWG8c27!{48)`Tcy0aC^1M^Yti;;nN`b$5|Hs!4L?s6W5Qh@P|2 -?|G{;>&CUF`*Z862(&S!LNfeyMd?Gdn5Rd@#-r -3=HfA{o6rZ;D2(7@Lk=L7wfAZBG-Hy>G?X*1N>>E=f;XjsEkp_zEdRL?*d$4rn>%ZR>Ni&0n -qS&n?uP8VXXNh*;2+gq<_D!YpyC!jJKc9z%S;}Vb8ccnD|Ji&+6i6pl9 -Qv7U~_3A(o{BSu)UGh?y)+*Q~!iXWX6_r#&y4n|LY5n%XoTBPs7$(2NSZ_aw+&rIPLi4jPJ51T?WZpR2~66MSyhyl_)M6`0>=iwA}w+?=&iP_y=J00TFA=KhF=ij -zSQ{fF!PM4A@%v^+%VY)V-DUgvXluHbuy=gQ@vng( -ycNz#w*3#_`biOoh&v*@_2sB41yXnHR->ZlHRXu@Ssd)8PPlbKOD#R|DS ->5fwW?;jEFJm!XjTDU?7IeK2P{upknLe#FSy@<9U(AR1CLM_kHyM!|{yq8&9WS6x2UFDI7^<rbUO;8k+BU?pZJ@Ve90 -AwQeyI>-${&Btz=W8Tmke1UI`jNvBWml$$6G5Sb7dY$L1QRa--ydrUTdAwTkS{e_IP^=2SsWrr&tk7X -y;T(rZ-0>45M(t+LCjzg+l!0ZN+V^=|*+Y4KVC9F -OG8e(=oQkyZRhCM*i~L{>#QMQfCl9WBU#^FoX90?)}K@t5?>E^Hp#f)E_r6HfwfK5Q0Wn8sDPOEeEiP -LnOE}Ju|2L>@{Ebgr?wHYwEBM=a?Xc%<_Fq*`M+<+vGeuTCI2`-{;&VzkD&YS0p{O7>noi6(;45xL;{ -5l^J@~uAcV#V2qF-S#&;}<;qWmxa|~tED21RDzCVe^euW`-sCGDV!N;%(jvOFwM_d^CnV=~Sl{NkefP -p_un8JTONAodTJnSYuJ)J|FgM8FK#Bs?Z_3?3iLk4Yc;0@&M6b`TKuol(}q<&oy2HMWrJGi~|GLa0gOS13MXt5`G9zh+3n`13@ -W_qoGBFhywS7{>l)cgSU%?lS0(Eg?hr!|rfD27GsL`EOPOe7~H(T#XbhJQGQO{0xrUf&i!(&ogJ4iz< -eXeO7%vR=m0Iz=^532ca2Y(*#yn(#q35Pb+Gijd2^A5vHiq&N?*+n8E8+dH4hT@YAg0955emqd~hj5? -O8Gdb6rdI(2l)U&@nZq70(zFMZ0F=c{Ptx_B4y=$|0_a|{wRGKVRDU~Sd! -?uCHYisz^@1INdduD9O3*h5u?$45hom$NC*;k8nTzcBL(h2WA?zxsSI?y(67TZ$Fkd5ql}qAS&+VKP+kgc{%& -fO4%>U7)p`&k(YJUESf?wbT4A$Ue{<6qo_|nD@MWuMnfNyLh6@t!<0Hk>rnDD#P_im+-n2%;g|d8zW8 -_pe(@?ieDSkiyZB>=|0+QrrDv~(JMBDjQ1L@kckCKzn547Kc --*yYD9ee`-$w6^MxNfm=tte2+EI^b^?F4GC!^p34OH1_v1Ss-+42QeT;%WWo&;{Vy}pQnc5=whehE=d -@Ezt$yj;%wgtAGj})PUZb+x(BGjJuYq_EIKPY8@P&3}w;=z3_9={kDF*beuGQr2dU&lp`W5BnAHOu|f -@aEk5yWIP$?k4_6r^qE@GetW=Pv6)8xRQO>YCh0e>#L)Kk$(h@`Y+XiKR14XKXraJ*e^xNomhyEs-Sw -cZ#2_vmbedGcxAx5@W^i4w8W2baMUOR -&c>Or7&K@soGIG*Tgim1Ra(xX#P`g;QhNcHCgLnLwGCv=?ntH3nh5TAhZ)G(erp0zAbh9jd$Yibs%>y -}Shvoi+@2M#)6xs?VJ7GC?0^4|S$&cp;PxCSmfN>bV4zP48Wu*tn3dmwUM_=S*pqCz7Az@iNAl@G?E| -VP*&$lF(gKUaRKf@#XattC!3r^#~|b=i%f9?=ksyKWDT$U8(!!M7bb(JKvlRZN2+bK>KCC6G3QhofwF -3aSLHWQ)Qc53+z=kJ5{#G!-4wtY-tjBT)dGFAnfRLfLjQWvF)l}EWqpgjd}Xj$y)WDUr(@|87 -D~y$NgTQ7I+gTpiS0Gc!r10T2SBm`O+@Bf1yh!&s(!)h -Un%{qHn_NT@&19oEb{Z=#MxM_}2OTk@9*`$JSIB@D_Z#Kz39OSu%*P4Y~a!4g|ib1OG$8pPKfLlN2LB -nZmN1oRCsZPbK$XUG1LmomrNBcc8{?ut|d%niYrP>~EYLJv}<)jO;xs75Ol@g9?dvO`@J(fAIIew&(6(bb-%A$(sj -f>7E#bXm(RqeSUl?1ntG7nD6#QqJ{c|KUBVew(m?PiY{f;nHh-IGWGlL)yZN2{k*(sM?h?Z)Amq4JnL -jbT)dv2-nTTo=Lv{+~`)t9jr81z?-M1)~P}FF6+KC=o9;><8B~(Bv|xH*x$Jr{kT)Qx&^m&Y&yR7mMx -4X;dP0-J8S~(4I&NRytWl)JjEoOMmclt%b-3ct@7wq?7KXaM)ItC4Cm$yhOVS9Q1jK_LB(8jz`(AdKL -~F^MQ$$)+MM)J-o~xXrM0@YC#gY)zu)+k9b5bYYeHlPL2@@1udW!sxMv*^NZ(qqJ2Rc0KVhi<{OqrI> -iaXlhf@@Z(T6OLrYVYqC>n;L!#N+?u@y}0e-J1E!@uT&=*R6IJH&W9yUnYlym%Oic+6s((Ds$x}S=FB8zsiNbd*9k0;p26+ii1An6XsNaWKpE{<+SLLJS79fa+ -abSHv1`5Dh3kH5m7;QJl&B_9{*qs3evT@mUtH@wrt{aZjk!T0D#r?;<$g#MBXiU$`=zL^wSR<7nysP$ -xDFpIMegmaR-KeubmU!~hW^>g+WvK;AcP#h{#99oBUR-K2R`P3bm_*)zZ?+B?Xz($oh51CbP8TT8Q;v -3hyXI~*qq}QLG!WZj%&iZ0j$bE?p1wIGeY(0$jkApf-{n%4zd{^H;J<3<>#P`u$wVSNPoBiT0@WtJ$f -ePneE!gDDN?A+88OQrag6ALLN_Tw2vA=Rdp?V+O5cruKe$NfJc=kK&%$=u2ab`VmYCPM}VnBB^!r72C -j!OWX;9e7q`IZG7KivbLkJ<}Sr6v7-1d{6DYv#Q%#;l}=99`r&wDU?$TC!!Nl4Kfz2IR3))8{kIX>d8 -+rQwpV@Vl)fjs^SQTJ~E1@*e_z<=Oj{NS3@ByhdDexS6=mx1!eR@kTUn1NXjhL*Qp}SfI_tZqEF2cO? -5d3YkeqZB9a9;yQj3Ra)o5XCP9HZgK%+=@N -N9uaxi$abbmZ|%L@j;9%GVAp*z+IN#QXSO0E0*;+C7K%HNpY|M?H`5T2Z`Q&(M=1@#Hy{3NU$gZ|~w -<~*`7i(|j&mkA-sp+qULqaqcbG|l2;9B`Bu^5BXHeN?BTqnaPbzr=&=b&2?$qkRb-1uiQ-$`)tP@-UHB4Q?cZjMcet|kstDrCLfv=HO{Uu#Y4!Q_`bF{a#92v -G&4I5S92#quH_E|*SFBQq=&x&Nv-}qWT^Iyd&$xadBw9Owm!dJfdnc|w$KZu&VQ%2yk-?TCDx9PL{_3 -81W@*QQYG>4V+KeF6*RP@$sdAWQnS?vV)66VIg57CL?{t@Q(r)p=1x}z_$^)<`w$t99M#mE%}cHZd4! -P-O!Js$CM+HI-+ppdD~-oHv?97k*{2fA0^Nuc-VXY1#7A^!;q^M62LiurinR_X^9(<}~po&M$|^FG@K -^q{Q;=#BH_&^H7{eJP%83g7O$)IJy4P?G&l(D|j)5X%1GO}tWKL2PH?G2d->#!ws^ml^Q03N_c$Ue4b -eF6G&SPft1cNFH3>lkxM7-X^*cGGdxm9EWMzYAw9D?cea*U0~-#08L!kagT&X9>Yw;RWH!tqi8XthNi -A~Tv!SCx}IKGSn|`2@lNhpC6O!DqrM`>T(<$%K*nA5-Ug`%V`XD#Ri7l>lLoddbz@*eCDtF2prMvMdW -24cOfkT^hgQ>kZ>%Z?R9SPZx~~zz9Ezx^9Fdc=r(&VZ#jLi|+iLOzL0(TjF7vB%qjdiL!YFmMcVZ`<& -MF{OMF8D+)1$bt4@u9R-s!n`&sM)I@C&qgKCGVCy@A1k6Crwn_57tJsXW -fOE4Y_X|wgZK&9w@x3J^EhMx_!V0H_oF#rZ=20u4ul#pgHAkc45*Hf9^`Fl;X9p(6B@{}PuWhk@CoJT -!^#WJ`P)oN(as+A5TU~=KxqNQF+gfp_jZ|~={o*}4a1*+IkhLy^=tTPod{CQ2nl>vQC6#2eWMR1A;pd -4hdEJ#XQtrPwTb5A^>PO;y>7oK5~e;fYtqoY05W#OqI9^f3&S?)ikp%FQ*j$S15i~P1jlk}6@#(DW7K -VU3&zuB@dDlt3;p87nu*T;T`s)^fk2eKDP2ZU=EFoMer33|)(1}ep-y}k{ -(F+B`!Ti#&*VK|wGkFW%bxQnVp2hdnXMn@w;n|1gm+3U4l5K$z>{z_hL$DQ2jV9FP%@!^L!P%wd4GM` --%2wZFjIFDI5Viz}2N!ypY#_S&GiZaoyyT+OPyDiC-NUBr`}5xAv2M;>L^g0?D$Srz_+r8$wy?CD0pS -)u;=1LUTh`u&n!l7pA_t(KJdk!r<&O#iJ9U8cF;W@&% -w!WgGoTlMjF8f2o=qI;#)Ry?}L7x@#Plx%vvZ9R)Edx1JLlqd`*ffDNDey#(a@WWTGl4#AVJ4o8>uz8 -XY%Ak-?W)C`t>$U)pVmPU3T`wU+=G1%)Sfet*x2kH?3Cwyn0TE_ -=n}!;ohWbA@&fL6KTU6bb@fsMO8O;l6F0|`Q$=Ffc(X4`HS1*f48)0h219&*c`L#2S6(9&o1%{C; -XqE^`j&HfxZrd02rZYjG{0Er3sY8>HX9AOMbb7l^ql9?C;=_@F#(%JSM<)1V)yJ;%rAYJ63|qc|C1HOHNjNs%D{1zYI -sw0{m`)J>iV_Xa(kENFR@zBmnj}89(FoGsNJx}T*!b6c`O~~OU_*bZlMUVaE=vywE!Bl2b&K`INk>>1 -(@8{3lvSzIv#V_(V594+G(Z~L$2AzC{j!2-D-x&Mr#{K#J#cdw=!*2e&2la0c{_)r!8};(4yZrXm4FA -}u*SBf`^e-;?hoB7&zFXSgT^aDZrTv|i1v@JSezep3`h(J0b}`Nj<6vvoxizvh+>@Fv_B~?=XKE)8XP -0~MZZ9h^MJaY`O4g_F^QpUBX~aK+E6FmNGGz-1u5YOkRcA*Nc5cK -zGyILomZX?%E$%Ds~$sJ?OoT?&A}G%ZlM2ns)xVCjIT -Rz7?T=o$=?QgCY$e~ZD7<`bmV=s$1I%tOn=O;Hirw*Y)ek6bWvgi -~as}MN*GM{?rQ&H-;z!z2qe}cDndbii0Bj`Ilw5>V)PYV8h807x&F@1o3j7ajsDesGPm;NXykk3`{&! -vwk)Q_d;zRtr|k36bU@i9dC;%~Z_QS!r9kF7BBsCilW*PX#*Q)@f@P;~AHwGOPI+{Ec7{K;i`IcT2rt -4*zMhUw?49@_{#$H>~q!dKsr!(iHt?(HZQfEfX}Vl~Qn5!S?M{^OK^Sw2)Xv!vj`v -9s-2t}8Opu#u*iDKf)JffIMtianN^jrFM6+^G5K9;Pfg4sdJfx=+Nk -5GoYO{~WA1=2PsP!tN9sDJgIAaiI_Ib$H?Lj*H`C6=rSwJr?0%hdU;v%j8o%wlWJdW -^g4LQ3H2VGc2k@l+E0ID+4D}Hk~1%(k1kWyoB6=v$BP0Y@XnsXnl9iK -J#~q6nznMVaJ=ygXm-*Wt|XgSBXz~OJFPGxQN3Q)s+!Y>RnREZFwdb?n=K>?q?q?E5H>%yKm|7 -8Gvz1htY;BROiFW!``9O(GAF)11sWaSH5(w9!!GYmci7Gchj>AvJw*GI-$35_Ll!2cw;?6{i&$L$PPc -dFU(i3(NS)uEmEw#}4$cQY6JeQ=p_>aAOv0%v?Wfh!T*Z`kS4`fBpL5%Mt7i~EYI>LIRk(x3x-xgs(C -DBCtr-u)>J+vrkOgvJ+ub?3M9meJdzS%j>I6}2`KGIXr77;SGE+O}DcA8xf;gjx!SjC0<7dk!bDv5It -@-oTY8DGKaa-?ne(rhUqlx0(+vAjdH@v}J2@Jrqw>x8IRS_ZB5Z^SqCvI2fQrKTI71NJ(4xFv#DV?bJ -n#Bx@<6Sn}E_rlO*LTiML@IXzCaiZvj*_~=0BiVLby$Eft1Qh>NBWpmvq$`}RRpl>%_ggG4LI_Omdej -}|>b~C2r{5(PGRLIgphRNdopt4-MF*{QM^s3ColYQPn8Dj?`Gg?CxI1M(JjsHeW!Q?k+oWUL1Q)_~^g -{BKHPjy55T6uHifj~@*Vn=YG)n`NrZ(5kLi{uOto!iLS+Fk{}#%!r?)?i!(n@MvBy~vNvyOxSjrnw6c -6U_uqOAXI#>E78qx+w?>f|n>obQ!H3W6`R5ByHX;6#4~@j+NhDPi?T+eioQ61|XhzKjz{@PGkhqOQPi -jW^~eRM>}Y94sAy=(uW&%1KxCb#>bc2D7m9?u4ia{vvCU`qpX|8wADj<A6E-B3trb5RIU0ZP0w(BvJ!NEC#EK$&!(~%)|R#Q9Y&2(wp-ITI&t#B39VW*`!BE(@?T-4zdGyBSP -A(vR>CkEgASth*$k&4Y$t03LjTfK9WIVd=M4Hx{1BfuAbl8aK%dN3{F9e`IJxBq*2sTN(CFlt;=$+xW -9>93PY?amj%*GDmL1ZnofsVk8RW+}qd45vA^M<7^zV{-sQgf29h>N%5v3hm9Z(KBhLebokvdi!M(|m7 -P&Vx2DOVsLBX#J5l0FU096kI)cJ#DEwe({ayu&%N*zw?hg_T}IV6bN|=~uq#|Fq-#{|#1>|9^#*J9U0G?uG~nSoYPUNp-A=CpLWu4{ftY!FqbLgvxH@2v&^ -<=wPJ?Vfvr{-T6QG}oGidhI$Y!c@3FMXv(u5|K_sila;|1NdYZIG!=Bns^gNlnbFR-O1Wr)4xuY$Y-b -~q`1S*Q}CI9FEPmisxxut#a*Vw(#u=cEa>%AT=dX)(Xcna3@OiYz^$?{G4wpaM$TzH+djncPW7CSl2q -`kbf|zF6DyQ7GU9dsJSfQ*tNu`=eUixToD^l-q{`x2ov_C)8wpM9NhmhIxXJ>M7Gb>u%-IRi9-PFSgB --?^UUe+bbbQe#nL`a~1JLz$z7O4W>)w@o$n<$aetF+!HEI84Z)-?#{nY%X>^hDplrp}w7`*!YuDkF1oTFSQWwL)XXgrTX(Ye*xM&3AWMzxaOs^s~2=QuiZ+pX+hp&6VA -J*rf;7`ibL6tZUpUz%G531a!Y-^9s{6qSVZT1qD)!v@?-6Js~FS~I1&KUAGsnUZ_-;oKGH)IFJ!k5K! -kh~9Y0+o -C8Ue?LHOg=y!8iVo1wnf8B&@UwfXmRY?> -|g|^j&TwTsCcrFPa+dIgPu|D1wDSfz*Q6K$P{u!SksAH}L`UnXU -=zwJy@w>qp?AXjoKFJ1$UYt5MqSDWB4wN5~D*`{9a_P_B)sCBXxU>Vkea$<{!|4MY?F&9)6VYQehg3c -Nq`M{`<@(=9bUoD65%B|7nq=)In^mJA5FP(nDD@4P0RI!5bO00J*T_WwTR7>1mHsC<=~xZ$*EmW4w-G -P^!&s%eT_pY8s*lJm-c)uCcu^smBv+wHSg)|IkSnM(!aF^wCET0p{(7GMm_`yMM4-OB*YXB}eJJLlYz -r(e_~~u#mF`5|8L4*GmcRuku&s85gSTy~2+w0gQZop{aT)?(=iCF$>wSlOr3;b^3w@KV-0;W -Tc#J{?YIc2i@A!1Rgi53)J`CIOc*<-|-cdqe*fv_m{$cHzRX6cIO}0iX$@mGp&Ohd+RXr*= -MJZj@nUZ@lSBeC*2E1kAJgIRotr=RUY;1$m;q?P7CT-ADkXfCjZig@s|~A5$icx`z(91)-R-;|F2gp> -;G33%lfN|!11u>b`z*n1 -y6SOUwUf?#4xx;9@3wqAYd3`P=-PVbBeENIl(HMuNf~w1cP!Ve>TCm2)c;Q_mczJ4`pQjOF?!QS%r|T -feorY@UjLhlMerP0(HIBN%4w3qzSCaq?1BlmL;dk;@i~+?{SGV7FDftm2G3x7jpSRU(c2l(Na3U6q(q -(qo}g1zlYWIGE6=bb(-KfMlzHvQdu>cR|6(qArCkz#sIS;-@QUnZ$hqT$UJc6RRCExn_^Rd13<< -S0tY!1_~%e($La^UZs@-0>1r>FnmAB#}b4qYJ}p-B=VU8G1l9?}Gg{tP5zN6|t)nN^2J&*wRNG2B-J!-tj5p?p9;%7OA2Pu#DTQvXE$TB(^ -IngRHm$BoGZd8yFpshG_BOT(0435uV$Vh$aZ&RZ++qoYi$==Ng~nXh_C;gCQ+BZn1}03SZqVLXi|^}l -@ffsfI?=nqN2f2V*a6(tM9Asda&WzA07I~3G9z!z;W3-h -*)UqyQ)xSZFy6~fQ9L%Y(zUJ!4Vg7&mDaIc`IhKFFb1`81nv?(*VUI-IGpV~QKyQ+E3!m8fs -XXpfA58}6^r-8=KJw&?ll(nKltW_@JH9pV`e2q5Oa4#Itp~2O -=<=2ZT#AbwBm#G5_UJwhp7rvRN3y`9@wIQk}wkj^ya$#ZuylR -{64MTcw)5}vZn@0-r$kGM)qX|^1u1kZgL9APzEpYL?*pclWYc33__w8E|>C$+$!xYa=Y4-cZZY^y`$Xe_g=}2Y^ -66Nt*b_kDLOdG=`Ts-Qn{By?uFInLJjJ}%K1cM;yrY#sqVKt*l|V=!f#{z8@MEXTcBNO@bFMwk7_M?! -DG-^d4B~4dLL+7UTw$*11Mc|6Hjzl&t36~nI9pG|gGj^(scXZRW<`pI^Z_|2 -t5I=XFMkt1_(DW38<v|Zm!U$vCPL)M+h46y!l6FK)Yi^`O~~Rpu9=^BmkVDmu@b3+)CGd!n -`;%N4j*l_LrOP{LiY0pS|dQsQ_$*K5EDH4qn9~{J;l2_WgR_Sg&Gw(3q$|^y{Qt51II)9s?C6-Q(aSY -Q+r^@$8CvH$O)aEt(@q$W{(NM-)Hw6+r@HUxhQm+2+`CG))h$!%5X8h|JIZq17+g?Hx9ax#NPAt1 -FD^dDeIwWj?>k0}|_`8k-hn9#uOp0sQ7}gh6}8OdC<|@3ZDLrl4@CGVSJ(la)hm(Gd$eoX -1lJ3m`N`0cRd(H!95>tH~2@jhnyJ`j{*gE%kCwaICXFjr&g5$@g7Z%(Kl|-x}qae%o%;WM?umkG!Q21sVzrVIA{6CK6+Rn -c?>l^h!fQ>4=vZeKNTXA=|C&!<-)gK)8!LNRK$d9SY1Q7nABty{{gJKj(;xtJyG)>STU>U`6kni-VjR -FC97M^T%R0hmj^R2Q%Hx@>rpb=DHz(Zl>DIY1wKe4Tq%B(DBo@Qk>AUZh%^;!ahAtJRo> -Zh9vAk2UOWb#)fVom^I4j_OC$a)5iFck2wVtXZ&07fxuRl1&Wy*8KxFN|;gFl!mQUd?(`wna{XO-}|~ -0kI7q*RKd(pNaotiHL#$;GO@biz(E*al)Md+cn1Q`!NjnTK)GjQT}-VxSsnjx!hWqsTPlXNuA`mW8y2 -aYRjYi)!R>;YVqt@HspDlD~pQ#y$(kG;=ek6^`7TscNG&D74m7xoQVCexEn<~7vt`j9=EX;cmlf -@pG(uQyC)8fO?Xy%UkVgSYuyZ%HORUM`~YTN%ek-4aj39`X_`F4$X*Bqx$Taj)mZ^q(u?d2ae5Ybe%= -dZ)?=eyZwU%un-H(_E0S12TRsn^mRHz=k{oK# -@^@##UG{^jsh0&56tZOPX!x^>O^zK!_t}?fYZznq6i5kjyMsO-*dsl%sm9oBZ%v!G3JMH$P-2M%V>%Z -9ve7yZg`N=TcFYgOzy-e`r?X7wiRc3ty?ZY23#k@HRqTwU0D>>9l_Fef5{aioS3fqhO*EQTujuf -f&5MKgX|F1y3+7@J(=n3w8w#;M6@Bp(`2uaL4Tu{%+)F*Msr=(#GVzI)w^H|8dnTSMAzer@(dm6R-!u -6KOleCY~b;*K7a`C1|=etH~_2C8{;JAT-x`~l93-hYJizj4^lIRCdpe!zK@TuW+W#kST8d|+X)3_|9gZ{j0hssPS}Y@-7PwB49(nRdn -9wWJ1Xt0bs<(=DjX2Vl}4j||z65yt?R_Zl0U}zU{p8+dmamM6dH2&t76#0KevPamvP?n{~71SFADF!;k_)l@3Hd!mgApL1Jge2gtQFQTgLig|c -FS6|aS8?7$|L@|w<%R!0;{04KhtnhAgNIk@PRNGR^E^g+v}ljSbUNxv+yb2mREKU#hU8gj-rwQQ>U}0 -S+k5U|fs@>dbi6IJX_fudYAj_b1A8)Msk=#-TWEyelDc%_bhB3y=u)U~d@&w!bv6AR967M+|31!_ml9 -Zae~U&iFdF{?#0LU^_Set -yq|@IfIUxSouG_5j)D!Z8WpJO7YI02$Ucll~RGrvS^&a*&U(-a6n-E(J0eaxes1={E3dB*7#Zg+Y@Q$ -fa^{C&_c3vrDAU -%YNn%WwNQlCo16F#ISYb{Yy1cBrpRA03qaYQVceo&fzaK0%elCZ(+!;bW}SdCe>|Sna|`cAg3>MVw{G>Z -4*qR4P+DeIawS4`qFhUfqL^&l>X1I)zd%4&zb>{)*gHai)%NdNK~8X_htJnin1C&n*NmUbD1)^}uc9? -jm6$Q;4VM-uzXm`~%dhl4W^Z^jxUNS$dH&FWV4@^K!vS&p#R=eV3j9eTh%FuhXWxRJs%V9IXpoZ+02)!C}yd)OJiN)^VyI -mRAM0^U!@{eX7}1B`A-l%j -BgW+)g;jc|CyIRax)f?gRfL44ZI03ps>1PvH6Zv27e0B>-F?LmYrVRoAy?A=f}Ao@?ko3*rR -3j)|d`3gz`H@8s0AJpwxMPakyB*k22oiVsN?at>{4b%ja~DpT^t`mtcq#sPlk9dGc2R@QRZa~TsZe`B9nzsms%q(#)n$xG8hfFu< -^zi3mmpsBp&OSYjF4&T$%0`?y_zh1PSZ69ezBkCmGK~aVPI?dsJZ#P3)1io@6_G*M1NnRQbj8 -?{eWp_QJ1h<269UfDL)#n3cS3$X_tjpPqu38QSjb|+Vbu5pA&L+NJGlf8DS!;a@R+B9!oB%S%}TS6dwBh&zA6XIm;^%!DP@2|7)`~ia-Bx#5XgB$-;ev}`s=HNfpZ5gp_M|Gt&iLVIT!p;1T3(0WskQOdBdTK$;NsUF}l>)bZ(qVST2rnC -!g*_=+~8*xqS4mQshp=Qv|wY=agK|20cc7oZrTv1EAI4T@omYqUVxs-VuC;EBxFD5hYO80HP?y!J!bJ -ZxJ^2X4rpBodh#Qwl7Rz*1ng444@@Kzl=b5NvK>b{OFylI)(k -BSu3z3opU*%lP(|?5wI9Jb5xZc<)h^@vvoeqGkDtQ}GUJ$$p@Sec`Ay_dY03dBGmMi?Ro-8(Y=LIE;p -er+M#&h$GoshR=NXtTJQ2f+JqmBc!_NKGRrN8k%c)IeX>h?q)MJ`*wLQB;Mxl5w6P`GM&Bn$sG?wddQ -+WFkIRm5X_gLrz<<&d%R}&!SOH4v+iqUeu^^Prc@Gj;(@os>M?S3Kh~&39_}xZK=;0--d|Sx;7?F96n -QK9lJdOdl?1mweLbT}`(zPfq}N}xMC_Y0-MUP(#Ogo^F|O`)HQN0)$pU02E+6&$ofiJvV?U9qe|qi*i -Tc~~eqc^CO=B2_kO;7RLAt3oJ1Vw)9rEf{|OXhJXEM -zS0@=bLs?xT^o?bx3U%*aH3-4TPsCc=@+>9)mp~F8v|O -o9EluY0ea6IDaq9#BHH1cF7$vR=LZ^MxF&~F(^AoZE~N8zscp-zcpcf+IKVwu% -sUzt61hU@@9hy?ki1FzC=fTd6DmSan_1Z+L#d(`t>_Geseyv|0p$CV^Ij<_D?JxZg7&^ck`LWELDpiT -`5j;U>vShRGm_m^0ihz2%J=T}+UbO9I?(X68b@pRn2X@ybejYAVp(c3-#nNoILr~LW1NQ|(HE;Xar12n)!oU*8{%K8m>R@(>&I)rZ55&Dy8kI! -g7yr8N6Bt5xs!E3&I=wrupM6yy##%_!GMy!O7N33?ossj{g>7N?_?zGcQzCY4tX5Xat7nIRUf3nB&L+ -AN$0HXACLN7l=%KHS1ir=4enT_qEj_Cf1uzt1I?$2vJO{j1gT~!ivGVxJv;OhAZ`Ex;-+#!D -m2G{Rg#2dZ6LgRrfkLlhxQfMW%c8(!DZa*A)HKhekCqFPX1N(MiRW`ASJ=PZ+)GN13beMs -Av-I9UWAARd-cR!^j|J5CUzVp8R#|=}Cb1u({uB(Vy3lotfdGZZq>)XY+rz-QArQ+AsnYWh@Z|)p>xN -gL3w;WB)@e12Zb@K%o2}_XsrxFe;#$PYuc@|-;kQ!R^eZQyp%gNxXG5RXOg72Xob!ba~hLg}-w21A81>A#|{Ccd15ge{zN0ql; -13%|q=aR>Oln(KQ9tsr%fMPX)u}Nx;r~fiD97O8BhA=7J({^(zP^*{9kO_Q&FKRJb>Jjm4NV@9ewvDr -k0Iof0woNAkQ`lb={eVTx~quS)Suw%O79~VcaQY}=?-^apTyypG5XXhE3`Ej`s@;Jey8ab*NaRsArxetdJuiu*& -Iq7H1kgqh{#66Z3$PSg;M_P_jadyTywsekLMHp`+cXM=!^IVl<05I{RJfwf1pHbQAHCNg2N<*QV7PZ%m~E^j3O -x*1zy<%Owgb9l7Po*4rK6Y1~?HA!^413tZb;*)KoDB_FH0V6VEfB2PJKN_Lca6&JO|xlR%kH0D=Yh3I -ez@px+0wdmN%r}f7DG-k2QPY+a%Pbn8(ZA&?}>Eo#eMfg0(jv2aKXJS2z*>iF{_&iHGYe~o5P{` -E6_B!6^Tkt{$POUJn);WnXW>!uQlEJB270Y@pR=I_Krd)=8wbuEWX&F6V!Wp#?s=PhP9r~7$3l3fLL6 -$G+uY!;!_I%L&QV_apT{;;#It=rM*Mf_Vm|ff~cj(=dLY%g3N8Y49ifsumc=-J?bD9cGucu!Zd9-@!I -U{IH31>dC;?WVfR5gs)dFTg->npu9Em>O%9CI}55%5Ef1R-k^%!*0+^;KHG4N2de_xa|2e?Q7J2HEM~ -tPpt?;EpZwy85Bpn!7%~cmLw`|M&l#9pUHp`2WWFKA`QN&-J53497{5!f*zKDG&lj63A!JM&+Q40-lP -S)->PR((5B@dBxCRxnji*Y+C}P0IYnjv7muBBDyu%laJJ2ECKr?c>y$FDYKP$6aXABD9pg32ac?Nezo -;S0WZcR1MPHT+eidqPB;)ufaQ>63k!|`-~w?64d$5i=FUjrUpwe)%?U&p1pvXE0N00ap>ipRgZjs@pb -l!69nxz2R6#JR)?6kQV5HHfQ5GABM8WI4+E=BQ2e_Xb_*oV|C6% -JzO_EH{satRaE>rM<>?8xke41XhL^n&RNoU9Te&qP_t1^jLaEb@JImsZe>{q;hBA)n#Phi!^v95d?tF -LcDHt#{y9%LCq2)9@eDiGuSwEX&<+nC|p~YL)2L_`!Q+tBk^Jk$s_Ti`(yOorm%CXv=xFb0{wfpYCMd -#(u!1ZTkS;MY`ldV`UaAjtHPTml+yeI8l=1gZDPU3BWrK@ndtL-A9Q-6GxZxhB_50Os;cJ3Zp*5DU;8 -b@AT&-a{R9ZGUIZ2I0I?I7{Y>f)+PVI4&O^6ZuFeB!91e$3C7+TGm#xxyIanpT=9lh!{bl6kvE1g3}nYq9Fs*D-)y<9K$FYLunH -JRDya0TKO4_6FK2;g5!6@X(ho6?QiVmpxnJQ -;w03L0xGsIJ!<=YJ|eGd-WFamtBqhbG}OsWUpRlRpm&aDQ$@fi$eF -p78h7yTA4}8ujQZ5@frr;@r~p?4Iq3^}%*E&J*2>>3Oa#GPdnmk!r!?Vf>oa&RM*D(>PiDaNgy=%h3F -jKAIuj`7;(lljU7!+!W2&Yc(v*#)PAGorc*8>!f$bdLL&~i{LO1tEwJF+QN^s@H~Y=gzvQ-8$n$w%+_ -_bJEy%a!xOVwT)6h>lDsj@3MW)!h+eqJ?-$ihKQX?b@u5{jZvSo1#`2MeRt^6(!_|j+JlmvpA8d%eXGjf_4{5{zIw*) -2L|?DxbuV_YH?yr>Pgpdd&QYHUGBScX$_Wr=%&nS6*CrIK`UQZgcQsePmZ;}KWPuOIU*hzoFETYTGAV -DVZJX_?y6hL)l0A!e`79D-xKS1y|J8IeMINdWiM~nE*C=VIfE3WU2=Txp|;v3x&!mlQw=*b@BD@CE_4 -6TLn(a+q>K1Y_Gp;*>wTqt$_e6ICLJR6tWa@2BC95CZg05aB`snR+$lwc78_MByN4!^!mc; -tQdy~xFAm0>4yrCT`pnLCeI9~!5U++rh?cR0FaNB!%N9q#Ymt*$Gj~9a6!}>v!PITzdG! -#%nl)WrxQ1WHpw5nL*FPbTIJU;KSpdWY3`|8A2#SSoD6Yic1_)E+%<;B&{y>!zcyP4dF>Qk3!uN84H@ -UxrirH~wrvA7>np6{*2;*7)%SI99nA7{B;#wL4RwdP}{H;{o|reOV@pJr0^QR6~tV-CNk>v1^eM7~Gw -L3BA=WIQpv;~F2C)>c3U%d}DRvr|7Ar -NPY?z`wv8qS2>SF~$eS4S#my1N>5mpm(vHXju`i|!<y0dQ__?5rF-tqLw7OSS5pw5%_+#pz&E2Ag%?seHuGIHjvUoPJKw4?)pul46% -#%(eQ(GUs}AZFG0u5 -UU)(@9$I~?lGeL5N2Xieb4#KetZ~N#xMDJrY?rWsh{pYgCrq*5=M6a0f54KqC70b#?^DpC_)MdIED&? -<#vG4wxKl+bm*_WXJPJ*u6*XN(hd)a^4cmJ_*_g|jvySCl$pXrBb1HVFW_qD881j6wr8)6w -DxNIZHC1?{BOif^Ko;3A$~lreLaokejCx@Q(v49Kbgh2Fs`L=c})vVvRRU1IY+rJs-`1%4#i$DFW<@N8j!@ -e=>;gda+d6FVQ+fUEYN0>5+|RBS`rXU@RbLMRjXjpN&|Q0PlTzq+S%%?bN{#c8@I&`cfgMij^N766a;&4V6utoJJN9HgZyWtxJv__THDC-P@dS< -c1wYP7mEA_0I|sOYU3MFQjuYhxoI_eCpk4LiN7+u*Xjxp -%^H;qy7r}lXmx^?9kF!}|{`^1)ryltR?AW;I0IuZn6EQ9?j9XZan-Wvl(#yJkAyep8dkW6lHN^jx>Kt -R#WAAO^$1p%Ik{VKg`<(kRHj8|HW#$eelPQNB=pCqGEq#hGrqglq5lOqd-1w|AD7@Fa50zxQPjhpWA?cjH=MDaOyW;$_6k6eN!3y -ho}nu=}qA0tA+mLazArehkl_xaE9JR-3J(g>X)OK@pevB{VTohQ`mtOY9Yk9U~U$M7(Rg64mOdeJqeIaMxWU(Z^>h+x@sZO=EEor|DVw>Z` -cENufWXK#GN!;#E!D$`w=$l -HV#$6_-g>Cux5_6R-!*xqCWd=u7pkHS470^S!?#+3oJqc7fkwUm(nQ2}hE}bk>HvkiY9_4C`Uq&xT;a -meUd}q6tHkdDvOfr|IamScw#GE(}4dnKw1*jUkhTVA@55Y-P`mb}9GzEuEkFM0+uo!?o(??nR1Aku4v -ju2}8;b~{uwT;!G+h?iN&d%WnP;tlPh3_qP5*;3AJVTV#%U*|h^b5KqyUY(1FmDO}jI95Yxh7mp|hI` -aHv~&(sS5`?6(eF4~%e-;KX8IvK*j|@jE5EO0q=>{usO^G7gxyXfr#bQX5|*8XF4)(e!unTM?9R|JW+ -N^h$Q}RY%=ZOf??=>gfEnB{P8duE_5A4R>FSfQ@4wENa8t?|H|`~JijF2W_0aM^s1Lk{7Dh@y+^)}2E -E_R=+^;6*%d&g>R@Xp(sy_HZe84j)bYH(Ft|rx*Cq}YgF(E`@S~_|I`&uG}J_dcjrRepRKTm}4u81IX -`=^je&$5E+i)TXNsaA@1U*4Ut71#TB+#;SVEa81G3hL_>n=|B;eL=E>(j&39sy>wt&2sNE#Tbu&%jooo+5c?9` -iVw&lSBi>(qii-WRL;k+t{`ao#6bbmWPjsts6{b@3UTg5?2sHc3j%R5tD#B1|-8EKjDj>Wo1+-HrQ;$ -O;@d&I_Z>K+eIrt#wx-LtnNpIPeF29R*AhI9ME#PuE>?`~2XAJ=D>v?I{=@fI|0$5+E7Kl0?Es?H)XR -Z~hA{W6Ely_Mc25_#x$$;Q?oNJM|&j=f}Ri8=B!0q^eu#3k%1IB%t$fvMlLndNizRnkUQ -2J`wlQU2!!Vez4{Gi&IS8gvckPO7H4H81cu_D^DSr5#4&V5FA7#>@7C}u9dUrVzI%-dVASTQ`&~ -hJC7;>B*zfjD*q!i?raE|&u%#-FZPt&oHEnpvZ^_qelYDDE|>#XVT{~+=Hd4p)@}Rz|A -L(LuWHM0#v&+;f3{Lfs|DZGl|B7T-Tw68Z%plfIPeEz20W<Yf27zvJTnsTACSYjO*VmBC;jg%{#B{FK84=$V<;d%07dFblt74R0X9cgLIh)g -USS9jMyzycJsFezJN;FuOK-_DE5k##8aBZLwHwzXF}%MWHNVbcYz(`p)`vKC6Ew6!?pnT}H!2G!?9Oqz`UL?_a -i2zFT(@Ac3147areRl=@$K%I_o7fx*Ss-XP#!msta=D~ma+b+cl~mg;_)^p6LBp=>`0tpCv)6cEjE){ -`fgGD7SBt}Blkl1i*qUj48OO~Iv~?_2i#rfhd>jr+lJ1-N0Xq!#*pT?2pP17)ppFLvnQh)hs#=HVF78 -S6tM#c9D}n;X9i7<`h^lMpJ2q;{O7N|N<{KeT;}HNPN)`@A)a@KUCdd@V8p$=@}& -BqaeT_j$z9(_XNIAOYuKHOc~2QSynEpx)AkCY<5JS!V47GaianfRTQ~?Lj|^O<+sED-houb;Hw?1Sc! -`aRcQSYaCQWaVTH{?SV>2ldmd2zr3_NV)jaz1Y^0*_FF;SH&`i@Sr1x1Ze_98fUl9yNA;O27(goWU$A -UvwQ|0;_}kC^*pKHMr949$Fy%QXpoO$|wxM0f-)OxWb-crulXJKoY2#s_XE-(tw9y}NTANz$ -rCcVT%)m*)NyNt3ztp!buei{LGLm$)W0x1Sa%Xk>dD%Trw-csoH%o)PV>@vBCB6jAcr**es#({ko|jr -;G2%_|d~<-<-mQ+D;gi{rOfQa{vF>r-4;RD@Z-Rwhu<&g8GyHP7jcw+3U0=Ii!jciQOQ+K#HO$L -4YF;0V$ZaVdp@1yB -oL>%gQSkh1hhy<{voI%;x#ZyR^)jsX@^p9o;1JH`6k|7A-S;QoNqv(gDQxf+NHC-h-e*>f;0TiqtnzN -`6einS#NQ=a>&{E%HrUBE=c&Y9B2_DtagK0gk!Qiqb00GBT?{Y7P0Y$O%PeSr=e)xy=G?-#J~{76?$| -TUnG^E5tCsU@M6DX%cJX59o}@-^hU!-!Rz$Q+`qfW~nT|cF8g%Ay6xBGu(&#u<_>p(poqLZ?No*VONz -f?qa*NAE)$1peoXzTny5e&>F2WUY6tQ-WvABP@r7h+Pvwl|*ghI=;IKm$mUYz#$^*uFC{}P+XKu3tZ2 -**U`%fX6FnPujSw^wUpN8R(7ys5);mxk*2q@+WNFOA=)Z;YTdDGgo)yMJ}q6M7Mk?krimO2yw|9m&-C -t7^b)9%|<_6Jok`wQ3%jq!|^TkMP@BKO8RMvU^Txx9h4BQm+lU$6a|2^@6^&vM5=YwnNK7!>$MaRYX@(0iv_3WgDbA0zdbQ{EB@TT4u=M++-h0h6quRn>clHmOY~N9&?*r -HWn3(<6!JiScpAY<=n4u&I69~e91VIMI8H6HH0;UL*q!0#&ag4-K^wYgL;K@#X?b#EX3?WT7tDYQqwI -j^7P)q??D6xL(^M)ByfRS&u1+o)3n4J~`X!ipx=_W0uw*2Ik72%r`2fghF=jgvN=2s&K44BCk+2(skf -?nKuVHpBCdBBwv1B-s3$;WIP&+ckydj{6@_YuUWZbN^>$p9sTe&S?b+=l*CsSdtEB7ay3{W!YcR1N!Tf*OSvVU? -L65Un{403#3s^w}i&$Fa8QioDy)48;E;T?Zg4&fb&Q+}zlD7)R*Na*5P!~~d{6ze2mkL8RJh8%DUVVk -o(=(n2^`hLOGd}>DWWN_WKlE{<4VmuboQOXKlMJ!kRF=d<<)SOo9{Irdh<~cEtm5{V}Iei@QazQ6jXAZv=BnbH(#XlT96Ham -C#!d -+Wi*9h_zH*!Ou{6AqA*S&D1k8uz1FpBJqr~0p9Zdy+YB`WII@<9z`ut9aFb_XJ7ldHS6~8eY@35+{CT -I6Q$@<8z(k<4{6){juU^!T|SwFDncKm@iZTjx)*0BpL -!_$O_z0c9HroZ>uQYT@;QzZ05O@~As9s#e;U -l#BH-NOjV$4=S-Lj?=kfyN^8?#g*U$Q!iiqH?2DQkn;Z(ULipMls$2BWZ?W+iId>4=MOyP#O(w>MQ2# -j_u@0?Gxb^BMBzzDPZ7gku+oeU`SHe?X;JU6^^g<6bk^jmUbvLl`LniPMn-Nx!n1hyCJfP*d`9KmGYn?(@u8#V9N0K-nz42ySl$2mbOsn#0P$nbNBnBQM -!9WRh1EFy=FAk=fRsli_f61>a)=&ZY@oWQ{U9DM2nehC(#5V?cTymYnG@ejFiq6>gJx%hG3LP;A34C` -ZvV$FTO=yGVq?f=c-DOhVWuE<$Q~I8I7{pWL|fL;n%(zCY|oDFw2!KA2$$3SN03LSi(AQaFR*2#T%PN0B5&Q=j793PKqMXjei5VwG;H18Pf -@USSN20X}_6>%owJXP>)@AT*edVq2i$ibWU&=A$cSu19ADSbACi`9u*BvVm?A{F -ju%Wdeh^-W`75y*TR}j^(p#KcAth5CFYD!Mrd1uD>u30-*OFVBVxUU&OW0_TU%Hn^d9xW6XnA(A%1We -~fwSy+W>74f^pLBYy21(Cn7)b~$ZbguY?*>3C0Fg(S~PWu`bHQwe`Db$^XVQO7 -h_FqBjhI_MCR`AMM5If%@<0BKv-lpUgwT@NExIE_pk&4?3}4QGdIV^eYM<(?OJ$Jag*@^thdg{l -jk?65D5&_F?LV#NgQR8nw~mev0EaB3+5=qwBqsq2;hXvf3!q1?~#s>por~OsscB_na#qMF&D18gJZH^ -sB$fM2Kg7u;4DLpOvX)@9M3&$)#9L=0Od+`mOqJLxA8Ze?F&sb>@Z}k|t&Jcm*^4FB(Vh@|<3Xn~)m%km=Y#sFKH!FC3JTx$A -dtd%dIHSydVvAMZ}N6neWUE|mJE#I~t70xT5OO*2|@&(fIwHB^rs2v&g5FJe>p-ucK@@wtDvR -_?He@CGWW#T=8iyfTbuTqw3yIxz=rCRSym&-3og`f5b1oq250c*0AiKw@*p8CB*@eAUziR$^~`KDves -{|5#yDd%^`SCt=NWIfKA7}42gmcv~S4xWBO}sn2+g3gYPF9}wx~1>qw>b${L~ZP1PwC@nlozdI&D(*_ -9;f{^j*Jd!V0Vy|J>u|fZDO*@&<0+IYY~ULm#{}%mw`KsE4|p=2^O|O406T9%UqI(JRNU~0zFkqqG-c -<<=l%MSy*{luI%^$n@z=G5AOc~dG@bZ&S>qTsd`McT6S~%@#@FQdmf>w@F_TG;qks`Q{v^!ZMpk1%Au -5w>201h -lrxZhc&&6K6Q{E*=PRIj#gMNOi-NE2W#`^qyIA@KO!gZ^%-5>P#OSCl`LSFZQhM*a!c(U|8bYlm!hSpFe2XtuXFHRDl9F|{d>`OuWStA47ZH*ZF5)SPC%L9jp(a!JJYYOG1=?2xtnP{lSCp{6R)DiKC)zsPbx@4MMdjBtH55gnxq>K{Y3usS{&Ts< -Oa+dp?iF`Iu-TW+j}?aKHk;ljTTAWr(P&pLT}eKd7*Df8}A3*7(eMBl;_e|?f4gcQ_DxJV4a2^hvvn! -qT8WEhfwHxUJmVE9^(BINo2O01uvKHX*8N<}33ce1UStpo_{p>3OHD`Ub7uuhw8q1IWxa-GlfA28rvl -8`{+u-1WS0@`@%3FwVSAt|6@E5}=(0MdgHP&A@AAb9C7bADjJ4p9{7R`UdqC-54ZvB_Gx77Q4HuDp+c -L1mn7Vq9d4*G{(l#jg%WBpD2dHW1^yCCv9hs$tUIL^)SdV@GZQy5LOlwyYy9k}UVlp0^Lr|j>vS9 -czC{87i#NaeM7;2qrd~aMJ-C_G!VS|3$VG~$2RIazT?H|dw6WasYGn+G}Iy7)O5hEEKWS==N3xXkDFQ4`l2i$ITA7>wQfY)w0(>H@cVh289$BF%a5CNu);Jfo%u`|uJ)b-2d8@E#-Xmoq&aA4ml{O6j%FNgE57Ou4Exa?Dbq=eThWt{jgPx2h*yS(oPeR&X2|^ch#NoSz}Su4! -R!q)BfBq|$)hGvct*rX$u`v+noS7W9xf+z3lYj6<{gSM>9`FL~>1;JbbXt_R90e;b#Xf@uUD -otnEon`x=EsJK%yAs_Mn8#{`pBdOKsE#F;J*Bs&Tq90OAM60#JPbGP9?MVnXtu+Pf3G+#VT=|zPADh7 -{2d=Z$_vF;&FM%VIKrmJ_EooSSh~c$bD)BeZpMP9shHGF<59kBJ}}j0je8#7yF6A+e5CX+FqI&(a+s^ -1&;9?$+F8sCfg;!*XzzAH88LI}CoKP-(f_}N*c1BETaarRJfksQX?^_i;5UwKdbvf=GQx|yJ#4 -Q)B8HWbW49gMc?(DwZj85r=?sFg9*?n$&T2cBFWO;J6>Ca)Iwcot9Ut=Jyq!)Q@RV6i!h@#_%_UX*t4 -nnmalAP{uHgF69c{7_x6(#!BN(n05h-tK&7o*2>9lh*0f;T1(wkg4!pY!9n_MdEmZZ5WW12V|mUY(E` -7rqRY|E(ax-)kx%TSAxSvM=#S^)31TdOSKD%;@W{~;SqvL^d(UOl2gH{)HKm3;fH%^Y>05j>Cgw4v#r -*Q+Q+V_fhHCZ+lOJC>O_p&W9u{adZ9Hi^8inDYouQ{IVabIr^0~&VYMmuk;4fh(*-Pn-MZt7zuXf#pO -@T$3g%$)yq7HFVH3o_OY}XqXF=!!-O>2mb)q7Ib)z=-4iYBF?|i8v*4{^@`5p;(hd?vyoR`5uH%0)r_ -l*g{X~fpIeUXt?)5Rk3xTeTuh?KA*5O={1gfALz>L0Bl=X9K%tz}HLNR+Qu>}0724nKcMRb@xmn|pvBT; -5&ARogV9BOT9CU -aHQ*Uw{P733$1%#77%FhY9Jrwc3!0xdB7Q6ck2l_p`!@gyA7!H&B${GwoBn*=fOzaj;G=$LTM)Wo%ia -{ts@8K`>UUK*_<^_MpZ0#%j@0IwJ+86S(cXq)>v7mS-WXY#A#*IG3AF?}~d>bv{|0dv_Oi}M(gN)co7 -P`-;?hz~mwHu(I@V@br!w9{P4g9LA@U~ytOTBLjx4(e)DYT93;cqwK9NTNDdpdWz58tB1UO?Cmy3tYW -i@V)Enn3P@4*9#EZl9r`_vs1jx0%^TZ)bPu*Ht#xS@dd_h~R>gCl?dhv?5jIZhG+ala<*M_$SP6Cw9P -p%ItLR)5;t09kVOtpPAiWZTJPV+ld|UPng}l5Af&A&iqp^Cx&rez(_$De)Kpz#lmAevsom0!OJU|lsx -t{vUb<$ZevGo!0+n7=COjuN-oGAb{(#U7e4{@b&(ENpK8cS#7l7Q&w^GOfulbvEQqFOQzbgq3M@NM@* -!qA)GtOkzfRzBLMf)lg89XI&*Ww0$ -)P&cCt3zakOAdY=wYeo`jC@t8n+smZl>iub9`Ve1ml)AL!>;2^h%XXFYY)+raL4!mPHxQZLuV&#mwg$ -<`eb4ErJ>}1ptniqmpA=r==pbd{_61g+a14!RFvAZF^r)QMxrE6q69)xC{B?uf*~XdL)501i49qaea! -Y3!MJF&cLwjSO3AJ#sdvEP21XnDAo*T;h>*R5Bz|{_z#m)WL2m9Va`ir_pXYu$zHqyQkjXo07lViQm$9yUy+C#Jgg$KLFDG1Ss~l%S+xnL -c8_Awik--7qmSgHWf|%J}^;ydk~zzLaK=#k10j#f#bZE9$VBkM`@9UTN(QqQa|b(`Nj@iJGyWCIOl@U -&(Prq?O1(j7`d4UJpr+&moduL-wb|z@z(pMuLXL3=L}jFFMZWuozo-Wt2OA?jeEfs`yN7na0ca=`t8H -#BKiYJPVTq%cbZgvbICSqroKEz{+G&!`E6gzwEK|)e-4?Js?W{r!9<(rxmlvtEv57<>xX*k)mLUfM;gb^o=UOO0BsBtdXJ -=TN21Xk8tM1gM*ems$lpyc8$xXdgZDMVuOqcTq2S)m7w^t?)H@6?j(6Bg-Ua6PlTi$@yBY6Eg~U#Y;r -FUwM7`r}lXnX4dzq0y_Bt+%?Umr&-vHkGaj@w3D9C+}f{KBJd!lGp@WgVpNbvyT5p&~$0Q$d)g5{4WX -hOPU{fL75M*V*i1*PAjAf`!TdwG#HfeID}CL&9tpP(638GAL0`_aRkIt7H*~;DK4(sL#AzXywUI`Zpl%s|Bag5P1%6%|^W17`2WhI>RA&nlw$M!?g;!1Yta&sn#c6W3_bkre~XmkgA-cztgGr3q -1NV_4OZd!C&wE8x;Iz$M3ixzUgO(L{Ws?ppV+H1f@4gjnEX0Y~&B8Hd&1m2!z7eo|yBY9~G3n0|fEC? -QWw@1p2Nvq&xU+K$2qn=-5WNHg&xJ@#ChwP3~?)aU*G)n%xeDyuH`A7t|-K6Y}3Qxp$y!Z1a<)d;ZP| -%#nY;f%(G}Rb+33&gpmZYP6HTjbqaB+b|`6z}CPC<+i57_e`p~rZdL5 -+)6LkHZ|=9O_RzkVRGa(iuoQoWv;RqPk$x6Ng^N#7;D`iB*IdMI5aV{diby?u?#G -z`yYb0p-Mr&BM8gj}>7Xpnz0E{xYM^A0J0^gI$&G2sL0(C+(K_z9zTI3Oeq ->In&)fP^EJ_xd$S+6U}nL&7d^FB)oN$TMO`lH)n!_lLwBGfFdCqfFHU$s@~M&yrQQlV0g~#1pC~;Flj -lL2ENN2L4gcWpvihwo7}nOtev&>N_K#ogd=Nj&x-==PA7DHgwGSKq%tj-LWDe@R5o&O*3!{EJ}$4ewO -tqt&m5u9-CAHG2vp{ouGPoBHkD|Nq2I}=E#edJ6BP15^LcxcpmsRfq%NcW9kK2%dOX``HMc4jgYp>=3 -;K+gblBNOVqCIT5sDVd`B*i~KEP_0>$()FXfXb8tZaSJ3G@_m~!Uc}(Qy#pyMut#ss-eHUuFn1VoS)y%_y5w`6Zko -8{rfH#9Y&hwZZ6MJ#b3n5f*&wX5KU9+Vcca -}ADIRxDev|0StQV5&ccz5xz@=?qIF(31Ds$@KDa?S*B8T317^K(V_h7yDi_+eL9^tZ -AI7?CdwM=X2ru_9pGP7f;_1RW2Z(h&pKr=SNv>oMNvPZcBz~sFM1&WPR%D1%pNtrTzP@<(NmSGU^J(J -NazII~q`{k2F-vS`PT^GAx>k%fmyP%517=}}p_1ydNwKw`3-HBu+Dzk1OP`oS3-TBcw3zdZx7!L!z-s -KU-dY}z%Ea=OI3(v6K|ALQQLmu#9C@cQ(hG{6I|#=Z+&#b>)C0A}o63;@TShv7Fdjlr_gs6lV`cPkr9 -El2qNZ+*D@tl7isD^_$VX9{(n@CZq!#bMC9mA2KX4^^c#LKj2I(t-54JyeQz54PT$2)xm@%9djB_|$H -7RD`m;Um`@y?oFn{2%7sKuyVM60e(Cf|Y)GTp3%?7E(RoM(*~t(UCa2qmwi@s|3!Bj-eUC$(_!PRaWG -@6oUSK!pB!=f6z9z7wGs1wq87F=-0Lw%>7tB)9)zf+T1dB1sJ0v+Z$;+}l|wjQ(&bZsRcs{U$od+i8> -7$GzycuV((HU+G&|M%mli{1@Z?(XKJK{qT1e3`Oq>6iePMCvWqk^erDZ?u8)x_KhY{(XLpdUxnD``99 -jVdqI)!W4+_0v3y^g+ep=(MW5^;guBUSx_8%XtnpK95QXe%KzqL8+i#T4_lwvW+uI2sdwZPbf1iGNYy -_TB46lQe4n-}~T`szU(FE&a|F`MaPePPOfPFpgdm+jre`h)Fzd^sGzd^sa>!BdYh+*u7N4?>w@_=sg< -jZqvKM2b%0EcH{^^FQGrIj%RywMt8qGekPb<62;tht{MiGpkHuFEc`4JQbw7K!c-%8SfG(zy|0^Muya -9J0O)?)mH@jiStQNOq_u6^ -hlNflHGpo$hnHOXnY++~CWqe!JnGJOq$ed9749M6iY>rde><3p#>)f%kX^_3q7Bk!>BoQ%L7yd3TbSC -CNIRlJ+Ao`9WhPg@n9?Z6;6Y141zGich2!M608W+6RCDDc#?M#+a5eVmvW& -P{30Losi=<2KNd33ymql~;(N@8`y~2o>79)9jC35h%1nr;DFDdLs7kUsOZAhKoma}b9Oq;C>wb}>i|T~?+$umfTe-M>v-&RHQ}(^%|Q++;jMgIwht7U@x*S$Cvp}~u*p|!0G1g{36!WV|4-4l4z=k{?I^a!zBx*+9felFC}sgV?@t9oHnWk|iJ5eZ`&CUUNA5{LsC^ -7`ct^`H41eXK-1Q`zBhJn|ZUejqLJ_^=$m6EMhbk4K7h2 -%Q*XWf-UEkg6?M~c){cu7z#ALu4Qe&54w5XubL|miPDzQQ@mR4s@EDu-f!sr#z;5%M7GO$^$p{Uk5h_ -f{fqqT-6mO`Rc-UHuMu6}4b~$N0&NV6%jRiJuGznSdiSY`#UI;F{wqiP+JExfqkT)%NMv6egQ-0ZkwQ ->}L~xvhP>RNo?T;`)peTiZIO3JS`>5!?tRL+YsmP~#Y5MjqdebQsdDlJhcQrNrke+RSpHn-h+elDEzh -}MReLV_+_lA=++DEQ-hDGg1-Dns6_Hut`?EKoToq9KoY@8~6pM5ut&-R(Fz3{qGwdmg)Dco3P67Mgz3 -#PGseU5EDNmneVCp=Ie-;UiN;e)yY_wI>ZR+EAxU6Mq0}sr~bEN -^Y@L5eVO{ttdj?)qFzQPckkvHLx0-G=)cVk6M@*5r7y{1pXS75H}qe1^mYRHBeLsfSAh47*c13{3Gn= -63BaKG_Ph4U5&-_RO*s?)$)FUWNxi&h<c$+YQ!tVvEK~R6i -TJ?Q{j4wRiSQaXd*OqAb#zhf4`*Uci5&Q=w(cLhx+e^voy?;qdiIUXDxN>g_VuWatreTI*eK=Yi1h4y -a;W-2ccIpj)dA;{3USUuY_S%?-GG(IRk4tsI&{*)iDn}@;uv0}Rd_&M)w)CW(k5xjE!}f_QlB`RL5(# -}-68_V{9xSyguI#L1l9ixR=*m_ZLo#@S77yTAMsCN^}C~er+=|c>LLh@qq|W63B%|ng(;Y%ahRka3dT -u<#OaSHN4!1M>{msHEPwm%CHq_Bct_Ly1+kZRGit97N622=*c9z{O8lLg7L&W;Mn6q#Ks%UI@ -$L?=*O1Bg2rac2KPckg*zetk$31=)|HCOj3SX|8l?W*c=%FIpC&~u(`r+hk`J>6*--gyb{8jv$(E5!M -W|m!ZuAiZG)i`ZAbNnWYFOdEz90S|8>Fckd71*)$pFk_{MZy0Ww0`T5_lDN>kFYY}x9)gn2yQPYhN31 -V@*c+neBy>%Ok+N|7M4r{iOYdk)H^D+I2E2WqA!&_UF_X0Mo{#?vym4MS8L4Tq2dLDqPG&S_qQ56=7^ -n(9*}f|b+8RT5oeMPA1o;w#wqLSyuChK(V#Y^0DFykO^_C9IN~@fi|2wS+a41!=`{v$39cIZnL7)~Du -;IdsEV{XpUSzzlhw3dZUDoX@>u<7;IY&3P-410i*95(!SpaR0=OKSPT+$8uT?9a)a-FK;N$veFwLe-# -B{om+|Hbix6W^C>MC*|oAJ`)E?VnhbmjtJ9ld4U)A6}!PbMkvC4>xw8X&7b@l%$kx5L`YvB^kvs*mbv -(GuQYA0RDAHfn+W0#E^$A=otP^{vWP&4Lmg-DAGxjL~1IvL50ameM1d1cHbm%*9!3Bz013j`Z6=qB4- -Q*ke`a89A7qd#ywbocd{j)#`XVCLngXt-g*B2_Fq)|8ybb{JJ2OqdunUC|9mg093CG?2&A!>UkDQvly -hDRRdBEtL>cZsiMH_Dof3Pu!Gck%c}*mP4*S+;({`%ITSDm{0b#E4oaF_Of6AAR)>XT&WBc06``z^^a -S#hK*RVodV(_ZiXV~dP`Hl_E+{6-&;nMu5*9&NZWq$_{WDmoN3B8)R -ga>uw&iPR>IHCE4vr1>(OqK5@7?i!_N)6*5l?*<07VHy*YJTF^1*334^KTsi{0e6en6k?fDoX=v!fB% -LIWn$9Y05PP{VHfnyil??6W&wuf49_fW!BzX9C^nLG~gmqy@B!>p83N9JJi0vdUkw(|*v!(Cku0FC@YE%cb@bj%&H)V07Bse$bt$+MIGv&^5{zX4p3 -bnJx1Qd`=VnA=uwloI4ryr096Y5*Q=3~c4{O5SnFIZ9Ru@akh68ZytN7G--H07#VM9L~jutg<>gD_-F -TF8EG<$|!M()_d1z!72qE-d2vh?arYox*x@KPj)Ym#u5;A7rAwWc8o`G|U!^+;Zf|@dQ9i(X6t(>&w;mBs<{P;j&EtAcB~@A$QrpI%$>!JljM?BaT+Z>mA48O -i6hOsZC>|#Q$!&0^5~mT1sHc%9_@Rr3LDeS5KiM>QGsyUIU@(TP(UQjC3rHPQ{ZN-cfRN}h;F0H5PCZ -bSCJkN5Le2T~)z4nWi(ajl8YyjuI8I9x@vQX%yYq4zZVbg>CUBq5t;;;o;s*AZ)TfyJ#6i-Bwwt-4p< -9y>3ph?V%1Rm3Qb?@ZCVJ?O4vz`PS64K^hAxe19mg5;U5t`m3W44wIM#IAw@3JIxZ??ur7?ze7I<4a> -x}w#COMy-|6%^+^Xx{|gxka2_3zQ-GH+gRH6g2?u+XM*0zW0vgQbP%#n4R3U9H(WvYGv(wi&Of)9WV> -$PzL$+@Zc2U=&cu73WDlYJOPzJuu9}NKyz;a#xsuUAHt!T{Ub1Tlhg$6Xq?+d+RkPr(e4?@K7i#BN8y -v*G@Jt%I34;w`5PS~>b4=$CPd(sbJ`JxDT+7oY)443$CkE2E%l1@Hp7IyE$Kog3lXcT$Z3hYk7pe`X5 -_Tg#7T8FW;lSo(+H7FC((_Z;#HOA$HKuXZQ*$)rQ`kiE)K=~FbzAxW+i&4Ft$Lz+Xh%M>(AP5b6dmS3 -Mr(ZKt;9{c+~$jBUhWyMEojGJrHgR}N(RYjH(ZaO4pSUEm?QNS&VmY2QJ8%rjzF?J#IPHcl0PQ7G$2% -27vn^#q*_+-p^t+&bt<-4yu>Xg-2TbW7WlInkLn|GzU=6lzss4QJ0)OztMBKHK96lx-{4$}!w;O -aG<*gMVkulQ;iW)JtKvBzy4UJ)+d4ZfZYR%7p%e2%%A#uA6T$?9qhHbaFpmY#sXj?C^72y@W8Mw_({^ -KT%q<8Od>ys*6Wb!P4Se;h{g8U26u*Y5xMy!l6={WtdeDyM(i=R2c493lw{qbUMINgSmxj6w(s!B7ms -AQ+}GibfFR!=Al(@1NT+X~QZU*`3{YeM|2?pqp+@-$^?W`gVERWbDUrXq!^r^L^fFXPYiYseS2@eDA% --_E5WW-A=M;+zi?qj(2OTw~svWs|-mr-vK^;$A@oly}|8Q2^_!e4Yot%Z*@nCUCw|q8#l+UiW^7g*s_Bjnl_O=g=2M -K4@)mhE@>{Dr)D*sI1rc}T?e9;=8r_VsdzqW%wdrG4onw4jNGhg-vI$zq`m-GVboP-0PvGdbL$Ndu2W -B%1E1AcXBfAz|M{nGwM(fD-G+6*zPBEQ0;GZF@XLCI>#tGM`0f37EtymNN-DEJYUZP8B`MG^BGjCjY5 -*RL6hDhas>GWSOMArA1nbsG5v2KfyihM*J8}`x39&a{Qc?H -Qa7;h|#bU*V4Ia55iddogmdmRvmpeV(O`DJ_aGgUlP+3eo29#}+3S~jka2lgUT4Y_vlyN&iYlAPh2xQ|xBjGF -y2m{eW7)URyY8K`&cTa6TP(psVao0pao}ta48ghKX5g2gndITzw3+(HS%8S!;Jdd7OnWr_f41!s$=fa -qnXVv#1v>>)aB1n#W@d6P;f}SF5PV>eG@I(k{+zJ%i*U!~5IAO;^0@W~IipHP(41P9WG(V-_0e?Z^rd -OYLb4|F#c-rDTg_`Hngl+OsUzrs6R!`qA-c=A+_R<`J=xiA}#9fDs(Y-SHrJn)7oNQyeD{kE2ORHLn* -Mhj(*H}-O8SOAv9nlEkh#Wjhqbetd7M65a8E*bfjY)KYCSW?%j^ua&+}3(R(p`xj%S*CE$cu=uBZtN6 -97`RXUb$1GN<$j;VxXMv3V08_tP1Y{LA|&e8{HT>!UeDh;}cXeckmU^jH1{Y2}P|!MpYo!rs5<@?c}8 -Z#y56d#xYa$JK$`$W%Drd;4xA3<*ui4dgV{8s9-__wFHVUoF$#Gs+Y`r>t$nKdl!{n!{yPntV~XTb#( -BQXGm#14+Y#_cIBM -hcQqew6vp9V_5Hnqr|hzfTu)K$f9^BFYz!yVQjOL)Z+U|u1&-$)Ep9ZxM!l}i^2Uc;Q$IQ5w5!q%tY5 -ErgI4|dhh>cdwFT885biRv#OGFO1=x`i0n%xH+ttsi4@Lf|V#p+UK5ca5wl4bwtM)YeDD^bn~zLJU8B -A*@LA&`muC=r1K#O0~R^#dXu#LujrKX(A-WnKT$24%0ypAKmj{v$BdEM0=+0SXazUjyKk+1%3 -~hPh=Uw0=&B7-P%X7q3)b+#|7;uF{=@Oj`8U9@Z39MQFT7X|o&6up~o#(iw1Vhyrh-6ey$lfr%J~!jklzqyW;FdcwT(0FcE47)FZ2 -dmo;=5~d66W3JUcEgFN3<%nJG2`g!X?xX53relj+7>N`HZ3SVvUo{?1+ecju3ljUq+eaINjrG3dpdAjzoKT)>T6aV-C$>ou3cRBfVY%=D}c -y_t=<~;%CMot`B2Dl5}8uXXRFra4`>nu*z6><_|R=9WD(KqgB+7R(!wuBy{9ifsc-v$J;zjZ{NtnGcE -c#YVu=#8om!bFd4XRDC|eLF|?S~h>!IMEB)mb^-_8`%Z`rMuTU;?_W-+DuCIcup_1U33*ueYmxmn`;7 -XsSm&I@2+;RG{xcB*?#>~xx+%*3DA -jt7aTouv`INXe~}P0etgm@fkjd#vw16;9;C$QsW>)_gFeZqd3VYi`wHxS&n@nkG7MBYa@n*_&@{)B_xPB!q{ZDCjR$-mQ&!%xz5cf;A6p~&56;*%-$9-4x^#`niy9PM@a2;QgvkhifTeevG;<6*Wk$X8cI=kvqt^S1 -x(hne|z79IrX-e2Li2LYiJ%|_>m+IU6t3A;Ug744qworF0=9Zp%qs?YdnmFHZ1qA%`*4yi5ap6x?Qcz ->J=;89C?K)vp9l+3Q8vNeVv2RVR8bx7{n@vzV!dNyQ{-hiX(-2PJ6+rw0O=~-Hj&1nIMPJTt1*-`v~E -liUddoPAHzxDl;^sJZMqkXy|%<5#IeoR!}d#UlJU(3v_#519`04NPcNwFtceq!?B$la~Ec%X&HpXZZP -bPn>v#Z`qPNp6xf@6QQP46_rDcD9+-c}^eIbViRv_Y7HFP5$xF^c5<1&f;d^b;r_{>+ -bV+9z#o-8(HX7AZ0j0r$2OU0CTv19ugcSDqi0lez%aUhE^4;tf98Lxv~w53^d2c}FT>ZYnk`&D|9_y} -$>uuxq)_D;Ipe)vU;LMSv|8rzBxvHe8m;Y?jjc|9|3e5b&CQjXVXO?oRlUn90yz$gJ=>b90MdTQ`1%y -~f@9DMKgRH~uUI)yz|_*Hr`?gU3KytxO}qYEw%(LF+OQq!4PjsWEMNMQF%=WoD{1yCx)9G#`Gdn?8+r -yGk1T<}3;o=VqwA^MthsP%5Sx$HWhC58oTEf!P5ISrc5*j9K=kISPxQ7{>d7e*L2n)EWcq_?#f_mZo6 -LQ%KUG3f>$Za!#vD!_D?1B)tN-jB3%B!lI_<(N=_s!j1tc`F(7&%4!uFYW5TtIf0Bxjfcb8#!5jvCx4 -4pHsXT6+jWpugwwVh)Ig-ly|FEJxzZm7@`d2^Yq%d(p`A*8pH!baHuVhF~bU1%t+!8OE9uB!|j85Dp$ -El?ODd88BrSxS-D=Wq+vmRXwhkxxuXpf>|V>{#bED&_mPj$vs2cJ_S%RW -4Y5YM+_pWNDYSodfK11thO7)!`g=TtZ#R>kR?9QCt&HhiPsVPU(~%gXfS5QB$ZEOV(~iIBPY{*HN?q? -r4Sjwa-=gWMHo{u4Pvt!&A?7_oS|k@^Udw!qArUw5IuW0|-b#5U&+bxy#}7fB{IA(nwjx19$htW*U`i -rd<-`6$cXgq_l!L$?XO@UfTB?K_dG57~_K#X_IlB1_Itb -*A%QU)u+^-!0q%3(Q$5(VNoOFD(e9v0tpr#_Q4SLd~@g{2C68!RB4|DPy#Py!p)o(O_bmo9*pKHN -y}ZX*txi?dZwyUxjwVxkh3BQ8f)+N4L$K8rxY_vCxC3}(i9x*f5N};f5^XnyZ@i?FZ{cJ5rlv!6o)aK -#>kC>5i~~A1cAdaMiK-}VtdylM4%K5lN1hpSWt)2U2bpuCBgP(goxM)%cjsbrjxvf{GzvEVT$ZD2O~d -jLw{2u>TM>o?Xl5U!47)ONj}6{Q8npANJlg~7P;wt=Oy1 -4qpKAXwxi47Y=w1|x-u4XiJ7;b?&7L{}?|wIX=+fJMd*gu=^^ULmeL-E?Tlb=`ZRoCW4a6sXl{*seIH -{M&oP8cr1U1?ZJP`j0_u8o!@VB^^^F8-EbJ_yvPWT=7+Bfdo{tEZnsTc51xYr*wE64p!9o$r1(bl&QX -V}DsePM2x)Z(Y~_`-naapF}U&c&peZCEOb6~TlP>|Bv!wola~&x>1UoTwOwbo)H_RD4$B3^r8gA={m8 -0ihTim~3T{^F-9*)j#oi9$}(N4{J0r3!;RY6kQI#6xYBnJI24atQpMYHiNpKxo!zRE)VFs939hsH~{_ -QoWeK8U5HJ{d5xbtmKPu1{m9b9pp{ogslx4^Nr|-o&>t|3e@nH5ejLhWch}_L`C -an7r&A1XR3ySpY=muh?N*Jfk^;ybts4ebSP>7)K|Ph+(&Wn|4d!e0iQKevHPPl@>gZ8Z<5o@INU?n= -CDe-~$Ydymz8)B5c`&klp>&Sll!t=}59Zv(z=6xyfF{XAA9IE?cJeL3EozB+7uYVp1S+L8H!XW@RiZt -XeVkv+Pf@oa}PiIaY6>;A%sfnT#s>wUZSIuLyVQ0B;b_b_PMr_w_RkRpiEz!`$6Zl%9w>sz8fa)8{VN4mxCY*)c22q`N}MsM -Q;>e5;+?x7xS1Bi_TwRn>qtIB%7Mo=D@qEeU?iry1-fAoA;ouyI$uNocDw458Oh-ZnJ(aCd^F7AQPBn -~EXY}VT<@`v*(=+JM+dOF>wUrWz7?3VTOUZ_+zT>4;twQvA=*S74)n@U7(Z;~qWi~SDAr1PIgBuxpaDl{jdi2gwRA?OTog -U$`71gk*>A{r8XN(aSXJp9sNCz==r|pQmxa0XKJzjM2NUwqcfc`5X4F4=+DR^?*tA#UZc7YNfEDgI2D -l*pNGA8-(cumd3ng{M0Ig1XZN@YL6*al{6yZ4`&g>WN!ZnP&miGoyK+RIJbR939j#OfuOUUfr_NUCR3;ktCidgwXXy6!1<*o4tAq!K0Y}X;0Z6zUx=B6lfGB2M|Y-l@9IQSG!|)xao{n -+?j?0#mbh{BISn9%H(ZxvUa#<`4GEw&eC78}FF&W1A8>no4u*EkZ(^H7Lgz`bHx!UY$<-}vs`OS~Ork -F?0!RqT5pvZ#fU(f!(A$DOXD4jaf)Cx1;qt{%9xTUqf^ConzWIuA6VwX~V)v(Q2+J}aXHv+)$HW0wQ$ -Iaw^t9ag2BeeY{cb6!Q8+?;V!bt1LvOW2M?eWX;GU$k=ydREn0pobXXdDkC!lIYmk5qVw;mSq5@|)#! -V51)#U$9L$~A<@WhmJqLpu%5u3_XN*gPD6%|JeMbc_J%5$w_#<8Vsm-15~IS;_j_$Dj8QCjuWL-k9Kw -R0)wn^=+dNi)S&*I6M4vm<~u}1AH>TPh+?gu*>DTAxKpSC#MOwnM+@EJ&4bv&)sB!iPu!`OsIF<<~C) -)oMBQjP!Levk6fXQSYOvI>K?0;d|W8n<(6BlkaQqo677_a?S>@E=?*qidSHeNbjY3u44u(J2Cg!J$B( -gVLar_Jb1wg#;=wnT`#+?Y|A)T*fBP^$yZZldi0@1XVQ6D?5WKOu_W(5op#(-@5CowZvj5r_iO6j;wo -yI&F5U;E_ZZ$B-B&6u=H6a!>NuhI=1FSft-l|u)^>guXJ7TKes``;uyHS{8Mk}8ZaLh}Z?qX{A9?z7f9mP){AEvnsbq@ -Jo-@wL`+lbn;>K3x;JFAe5wif7=fde$_5W~aX6trbthGp*hY|M_5{uxDx -i{461Zoi_z_x(`v0LL*#}^qeVNT})G_x4I;46BQeF)e%*D3ec2y3pZ2v)Dos?6l+_Rm!cuUQNA6MPiQqi*1v2of97i% -Ez{?#jfMhrX-om1-|uZ9dHa7OM3@qq9~|!#VI8;}ojRS}v}GPp)=9@+bo26V7Mxyp-%3LC`d+Q_N**s -=<*{A=9Fi1qpK0xw#xK?Yc(bp-vQ17(v2e{ETv9;H6tt%W)*0Lus<96y-v*zK$`2GtfrxjCz$1C?10{ -p{U{3y%K1ZLi#eSh(wKA763OoHt)1?FmzJDz$a0lM_#fZ`PqxqycE~a(~GQQr@FkxASeb|Jxvf*KW(= -sL}xGoS8o@Iy;rvt9AN9x=<7saEr-pg1y9xBRpq6a9_uaXMM;Bh*>+#cc_Am-8OmE6!XPggJ5(j0=*hdwyT%O -+wF^n`TFd1U+$whQZJi|%Z!V%<+lDAd>JWl(@kEp}HNDmSLizRZP1tKnUJSPu04i-Ra?Gf -~Q9YA7}R4wyj@!k$h)nq84Hm)Cv2rPGB&FOWihAlnM%Xb%0+~ZhtxFEQM3g+H87aqdVU_C}6FizLfJ^ -j>qFVy(Bp(g80l8iRSqAbPwr)ChqK+baCGXse_Ffh9IwA&~x+clsz!u)@wZ4H0q}BsYm8NAXbX;OFbm -`^pH(it+t<_sVURkb%iYStYf;D_8_fU(5fPCOaRH%3TJ#Sl`Of6Y{@-e@%FMd=BamLo3=LDgmby?1g$fp!x@s!fwwUOErvNN(IYmC7XG#b9>7jMp_K0yOY^BqSnZksqt!n`mKN}p)3+vW!X{4D-&ZD<8QQ>|uu0t|BoWKby2&h+VVa -$6-)1!8dI7V7J@T6-ecOGs{Yiz=>2$R%?*O(IVjc!e@d1vb6VQl5~MY#eff)>r|)=n09eFfJAn3TsBW4ah&Q>GY=dg5gkhvq=2|aG_up2`UtzHHn#%0-u{~;d{H~#V8*$L -wQR`RDh?%%Z{VkcCxj?r9|ZA&RHJeZDezw>$czi&nO*T2w;z;*Y$>G1chh_Au=|F#t&`2Xuxgi)9s)| -^>V3B1?Zd(eVT0bL4jYwqE(ydpI`DwPTdl&HOpsl)W;J&oGB`Dp(FAL)P$d>v+iO*K9u%e==S3~Pavb -u-`acwNEpp&os%LY8*uA7YTUEY^Dl_j@DwKH~`{X!l4lJ2p#|!@^wC4FBfLgL#QLI#FJ}OlUyay- -Pk!X^f9{p{L^!h4_h}9GcHcUftx{^(4SRT|}=2}oU`3Bgvca5Ksre+OU`!RUHr4;GuB3@UC5V)a^@0} -)E@mobh5yn{RsRkx=uO9^0SA`Fh?o`T&GhOxYJa!Hw6V5d+uNb0`;lSBSBbxp@dA%GyE2shsrPulWJj -|V~HYv470+h|8oL=1K8JiV_^k$Ztlv$+uC=_e%@WMuJ^%iCYVR)9PfHfm2nq4jzs&d&#h~7JS3c6OWn -pF4cYs#zFg?@oO4)5iRD3(2ioFRVG5!Al=#2B8ip|fRcFbFq(aM+}rfU_M%F_UvT$vLQYjxSp<_35*> ->vhUIfjN6Uq)cnZ3?R$6X_W?#KDJPSzRxEcY%3tYaru1{)4rwGj%<02CH2NnO!@K_&XHXch^M?8yZmz#gxPj823D;A7=a@FlUBrYUTvP4B%t!C>vW?n2N&*VW_|13|8pxM3Zj3~ieTN9@Uq61#^w -QM8gq5u<>|!BFa}&^W~%s9DxapKckohz(XG4@%2p*zoR1G#u5YU>Uv^5@9_AeH;`s=-s~vYr{Ke2}ns{vHUUFI$*SR)U_tLf;lF$n;tpl}3Qvp8~j18wY}{IM$TeTnYzwrC -7NH*`%5(MEs)n&1j -aqT98;KU|lXHFbpXS1y#}{DKq(CID|DYyOz?1RD7;_FkXD1F*gn}@J#WSqxV6K3*lPXt?jE30NxSx9& -Li#)h0_HtmrKlRgg`KQx%Kn|uBWEeb@k`3)oWJ(9zHurDLH_k$|2lGue!J)_{-~q-2lIc`(f#S%?_)4 -2h@^HVP9pnqOX2X5gGmwy3Q;IbK-6JGh*1Pepa{N?;6GK>5y$z1J9e-!k~jj_j=-Y$*iWNB`&2s@g7H -HUMwMi#9`WO{KZ?PmKO#QtpoaU4J9OAO9C -8)x;BWAU>>Y>mkWbd|(OCEdT~Z&5{x7)| -e^eeWG>`X@ER6k{D7<|QejEF}vd)lo+gteM7iHb@c7YVBw*)r&2*LJ6D*1O0epjx)yY5f+9%ZOp8yRDt -_mFp+@9x>|jG1lw&R8AN^`y>6{1cC`K@QZqmtoC#Q01qUp*w+bJn3 -7abUPJD+*PvY4q_4Ra)M#^VQMJ}nFdT%3;}8A&-MiEisRS0^c5=yEZf=F|0V^ik-T`nD;**_XdzB8Gs -!?a6$BX5H8_dKk7glHD$<=uLwGBMvIg0!Ozy@hgays0DAfI#g{p|Ad)pvu9axV?!Y^GROr7YPpN<2+> -Avv5@@0g~>2&l>>_68u0yQul<6df3f(ObefOrH|Xm@8gCskX$VGb3zg%xv0ZON?Fyk(gUFlNSeboeLa -LV;w2Ncw#BW3du^m9!kbh=*MN)QWTroaNQKVbea&V`}av4iO5U1lFEFvb1_{9pvbEE$@Y+vTnBf!Tjg -7z*oVaoZI0k6*?8AGni2l>a+y;Xbce2q^5^2c)Pg^ac?y{CS6ESePz!;~o=o;0IvKUFykU}wo?b0#?y -b67#AjQQXyX(z1!dNF)n_jCxZxE8Ngd3qFu}>3W9XJVT;XmfxUvmQj$SbN5#P6dU~3V-yv!4qZbcNUl -8EdKoD5&Hs{)G!OO%RfSC$i22117ZzE@p(qp&U7x}(KhzOz&KTc;X+uv4(-vQf1PG8b$wS$fK%smloi -xVMM(?)Sv;hT%nl8{B9LXZo3OV8&b1net85`za#*l9#lMswdE=dV|H@3+Hw1i)U5&QUFgpJFLW`@|-c -W5VPM|)ph3+CM4Of-3mw)o>dxf3%- -@^2$x-bTmE$eyMM67+-c~n4l+0vd4MA;D*bh0(m2eYYAHjH|#2qax{!B$FD)yiJt$-j6t>D$0{jBuBHp8G -RWtYVZ;YL`>W>JI-ZpY~DEe(!#nrdvWoT|^;F-IG`dQqcMZt1OC@+SehDRA!SeQi?T6PI>%%?*-UJ@naL#>vuar;kP?22Xou3k^n9|NXCH`=!OiNwtu#TW%d9@RPMO*Zjn -VTs)E`PA2`96FRr$fFW9(?QTEuov*G=3Fb`vYhDdo#bWwLhBmoj){5!8C;)CmV1ACI}eCAn5S4BB3L) -7e!$T$A791!H}aZnST8E;`?FDx6_@S1AP%Sh&l++vBQNQdvhuF^Jw1N#|`@SiO2nPLqv^(-g=N;~coa4}bKp&OKCw!8~j_&6EW&1aw&_Vi;+wpyee-(UaKgf?efAX;rq`! -Kl@S!7yZig0lXO=tZ+&>DV|5P7heDopD-~6G$mV?XdT&}scSIHjH5<&ZW13~Ra_dU+x_(*4xZ|-{+7p -h&K_a1gb!?(uHQ?Z_JC7Xdhio;hXidw*m(DHMl<(i -?jWnBZ}@h_-6gg^R(i@wjlG-oL=!N4D24bA=q1XzZQ<8>nMJ^NHFT8jq2pr%{QOBtc;|)f99IcB=ad( -sevgscD8xAjo5auI*Cm2r(Yl^F|f@-u%6~r&d9rc(eFgW?^RqXGv>8L3G85aJhncuMRUw<_1yA}kD?wE(eI~XE664DS3As9_SA5=U{^N-^Y -*iTPGAn1XH^!=#(Na#2sb)%2NRD1EaL(E=a?f|r-7fv5FCh>DCnT$L-1beB9e6(@YC(-vPFR^2gA`gr -rKXK(0ah%OR0?f0c;X(XSol1Xjl=wuT#Q5QpdQA9~syk|<>G2vNK1i1w&Z#*1!CK@fOvz8~N%C>(Kb& -Z8hg$BClh7Zr-F{nt?w0>J>Qv-6w|qYiq6oiq?+MaRMV4L{VdrB%wfob4Y(e~OL_88a!aqX9kKp4ch^ -TML+o-<~5jY_6K*Z0W|Mw9QI3V#aBjQ&TE$QyxF}qiv6Rl3{2F|sy%q<+nugc6dHGTV`Y3abMnUof(I -IbE56^c;XD(i)&z~?eHFB2!lrX^K_Sp&X*#ZkM2Z%l^<3Ec4?f=u!T419qT!`Ule59#SRA$qTr`S~bB -xWXalE}_h3oUCahPK|pvA!{pBN@>%nOCO``Z52>dpu}(Q=}miXO{yKT+WiB*xp6YF{4k{=s9}bshU)q ->88JgJ?4C<8JZ)tJaXSzLCNO*z&v#49pSUxpmBF@2-@z9wL^S8*%OZkJ59SQw3U! -NI$_8F$AH>lKI-8TAC{;x}VpZVEMXC#eIP_UTf0PDV(o#T&1`Z?}j3_BX1B_f;zI)Gd_MHPi1#{>&F# -N-gKvF-96mB!s-32{>&-rBOnSku`ALBoe>FGS4&rZfWo9}yV1h<<}zg-qWfbmhBi> -3>EJVK`AQ#`z1-|+B#{dyO(rG9w^KK$6CUa(|3x^>GrI7`BezE7@=7(HSp}zcfr{Nwv9kPlX8P~V?*B -m7@q=?NK#eWKEpU0(=Fz_3szhGj<90K+4*FgMVrj~9p^R{0Ns{+WCqKdCO-F$3LKr_Qbo2ip0x`rCAI -m|+T@(dmRK?l*$r+p^>v&8Z=O7lAwa8QDRMiqKT!SDGUzIQss-1NDWs7(ttwFG@2+1r(0O#N#r7t;eS -_ri==JQ4<)O$^y>50<0E`VHxgVd**@6aWAK2ms3fSzDd_1J{uQ0043g000{R003}la4% -nWWo~3|axY_La&&2CX)j}Ma%C=XdF@qAYve`{z3W#L+QVvKELd_gc)^&3>?JJ0kU$8cN2<}Z)6*@wOJ -ggC{Pt9-C5@z=4fdR}u)++os{7T)tE!i#X;Pvmz1MluCI3uMk~|)CfTOEggW!=`Zv)T)LEAve7)9DyF -s_U*ZV#mrq`bi+I1icvl+F@b3VD*UAAM|`CkR~_UlrktrX;Cw5As$-h@zyIYy&qR;TLD|HUWN$hx><5 -zdjZZkH6m4gy8(Cy)HQeuy0*^aAFbu!ZWNLKGf5J?7w`EAec;c9wtEL3?;!hW3m-mgZj;{ -A-n%2j0FB<|9)oLri2=r-L$TL}J<4xL*3rRTAIjqMm;1+`$M3yC5`I3Q$CzdyAG4I_9rTswzhHEIAkz -V%_j)HjcP+s)!n3jjgm85Om6C`oX(6nV{0>|GwzE8VoafzGV`MCL8z3f;#5n+=>6e)047)b=x>47aHjP%jdo}0Xo-mfmU!wWmVm03Z+41+eWHF -;v_a)=*v+#3{vHZia4CJf<(AR(P1kW#^F_Q9$#qil4PP*t^_Zfjvw4#m?i5GTfkS6|XW -$%63!8(@FmSBxL3A*{;y~6y5RYVrxp8))835N}5=4!%O)TH%H7|4`DvFfmT>U_G5#|fb;G1Vv;z;I4z -BhP^Q@7jcf8H*b4eL1ZBAB2T6qc`QS{&-NH(AkD@`3w&!8naAvxvqwVmi7BAOB>zf#R+ZYd_G7+y2+oMw$c|_dS3JCW -^)mCSa-bDvenv_<`D{(_o58&pwybV^Vv!YHpz5)AbT;i(?T89=E6=L?p=`~N2RJl7|zg^PWtmMRiV^; -B}vc(e5Gs%w6MV}7jZNveJW$O>q -#C~wq|3rKTCd4qve -&<&l@zhEgtExRs(=?NqVl5<*W6F38i9%Vdt7hr9Y7 -9v>dRd>SY1^x&cNdhGN7(qakUoSEpy!;j;sGX0{qr|+{_p{J8^&e#ErgHBKS@B+uF8oX{IXeX_~worG -~bkq@e3)S!fqn;cd*n_XkV1^@9P@@dm^c})D+xkm2Xj*B!clPW7-TBGXgc0FV-JcuM@H+0z$=n#|JY?et^z6_9>5SlH8`ILa4z|Ygp*OCJpf -Ln}ID2e5YKD*_^6}YD;B?;;fzlf;Lo6c&>TeMR-b-gsADv)C?RgrIQiq$yVq}5wx2`>PXuY+4Z -!cu6wnJmck^4HEDI8S?`$q>a*QXz>N2yl{^Z?`AtKLSgt5Ewv-vUEkjJJAcXCiv%@!<_-KiVnq{cm%D -{A&(nzEV3TIfxPpnLmq!CopFv*7FE4h|5Zp{mXJ#jn<5n)dm@X|Z6dN -KSmIrl~gky8bQ5jpN{7QwU;*^O^1PR-)CmybHM3G_brf#?9p@ -n=5Kd};dMmFL+QxeMvT}t)ubSFw40h6MnKvptGc#0D*@&JN&8NZ@X)bQa&aE0n^$=|djoTZ3W(((J{# -q2LoO9KQH000080LuVbTY4cwo_YWP0Qvv`04V?f0B~t=FJE?LZe(wAFJxtKWo~3;Zew|0XL4_KaC0wT -UukY>bYEXCaCvo&K?=e!5JmSn#gLUk@djO~h+8i(lyro^WG0NM(A!(v6gTnj{(pSl(=@>eW6Dg4qzom -hz*5x;92_h;&w*pOP$n5k!hAJT>sIs|Z>CeeOWImCe(5&G)$O`+e8zODR2cj= -C+)_=VpP)h>@6aWAK2ms3fSzApcg9o4j006B9001cf003}la4%nWWo~3|axY|MaAj^}Wo~16UuSY}b# -QYpUteuuX>MO%E^v93m0xeuFc84s^C?b#N=FlE>;+8Vs|X@%cugexT3`;N -;*1g{n_Emm)`HC%%4AxR%L80c(S7h}dTsg)zobuJFOoqxq{~9FPhGyUn=YhjJxC7d6}r -;OHYBf@1++RJMGor_y#v!xX?CfuV9;2C#%CopW)&*gNRL&{X5@<&tezyEWm(&|);wbAvrvJFj2cCB$`@>}k%u}?G!AsfDnl)Ec -t08lWnonG}K}w`!A>J*r}3k>Qakl=by`z*1Qfb}S~_wKKY`4pvFV?s--00vX}HbzCEy|6B{2KI79bqd?PQR5N@O@ -rd#;SA(z=aS4@`Vy;#NEdjD3YS=u>twD%7mM|3n+KN(Q8mr1=uLIO`z)fjtE_8FloX)jMn2_|f8FPUd -!Ld=J8~!Q%3%2yC`m{;^Q5Vg#kpJ~U_n?7L?9qoI8nM=d=Yz!ENIfQwl&Vl&3w8kXVox*$&(n+Fv!Mq -pHKvXW=jAI(tJpDA8qGZl?n!8%`OZ%*6b#6!%{&;B!2-=O9KQH000080 -LuVbThD+9Wo>0{bYXO9 -Z*DGddF@$SZ`(E$e)q57yoXwwqHPZYh5#8>ti@1ZD~c`HOJLZHMAvL(Q6r_KS(pF5b9fUe(oWNL*vkw -gwnd&RzjLE^QXI%;vm1v|vdxC@Loeh&XxE8>4qS9_)fta@6%rLSSFlI)L>OLk!0pd8(=)3e25(XbuabV43z%l~AXjvm+w|D6*x$by`GkM -HLT0x!wHj1Jx^HD0ge^6G3;t-oBBWO)Yvkf^9yam(bYNEyGDV(c}0wbI-g)`PaJtMAO3I(HUbfy5HUl -E0Ei>vg%%T7A%7qqmS8oIzV7{hTzs9?YheEB$G`CKu%Wz25CrZmeErNw4jMB$FW}xuNaI2=7v3Dn77- -S`|hE$Jg=H;S*v8V_n7=Fb)NC<(Scn?tiLlC!>|Gd!u$ulqNC*1#{tJ|?*Mo(_O7*0CafSV!v@va74Ct -_p6WR7*}a=Fy+PdnoPP_vgzL0Cf&kYSIe4km99u_Kh!z?W5?)4tZF*lbpcwY!pF&R}YAR%+c!5 -`Io~*CF{*bZj~VWLC|10gh%*;_YrzW+{>hTcYoa5M{HSmRpQA$*2OiTo3@7UcPr^faL)NF8zMiDt4SY -l6H@7unbH0K|hPPP0ZEUW4^5X8AO7X1_~)=Ejx0%=Pd($lxT18)Rz%Cs5f)>J>cAEALvkc-T@x!iL_4C_7ZnijlU(|Qk@n#AElN*BY@AasIl36-)T0-!84yPisgR)$rr0wG -K?Lg7zz?Hw6z1?k6YblZUqd(kMZCVSVnkfK~(Xusch`moh8%IlA)TSg^ -}OomF;;sARgXP7`Ab?1rdg{EWcq-JW^z}aX{sqlb2Qo^I&MxrT0v*Pkt=T8`Qc?(}ZYYAZNoi|bxDjgg4 -wWw&WTE7am}!U<4saFB1lJJ;H3B{j)nNhh10;1V@_t=IuTCZMN)c=O-VwOpHiw9omM_LY0VO3_pIhS4 -rRxxqxWTxIn8ETBb@{vk-cIwya$53<+_nX)xDxU$G=scmp<`Y9oGtvbCVw8YerVQj>qW{}A6s{%o5`JA06iR1=ZE*iju1~Y=bae4#?7 -RL$E}}-@-m&O$D!qUnA?yC@O4s0k1}Kw>d9IY9p~o4?m(Rc0%Ud=`KqGGHrKJE -aL4s*I;v&@eO8K3`xDKddu$Aswl{37i6`xJ&l=LmSu)`LQHh`Vy(+eu^K3Gk_mJ)U5|VsjSGno|u$HaY2xe5|L)-}F4b)3|g&LH4Ze?X7?w0m1< -SYqaSekGC0@540K8oGvlo(o@~;dA%pMFqG>V3i1YWbeEKRvXW;zT8Hhd@vLT9Wlc>BNIdc!FCfvrq=4 -9&NOi&8v -{w?R65Jw$Yu$Q*CJTjpespgMwUft+etgFB953f@|u0|7;ltW7G^zJN)?{r2J$xIChj03wUcC&%fDDL$ -4)YcA-qsJl8#@B;RRk`Af#Dy;;0@-)`ae$)R;i>i92KasY{S~CXfAcy2YO5ibXgd+U)+3mw@D}l6>Jv -upJMQzP}-gFMS^xCpGbWp44}uane72lN#K;)B;Wpf3dW6;K;qHxmKLAAN~xw-TG_SY7mEZyA8#LvXVl -xQbjV-{_}D6Q;xFJ`1Yf*1P18}l}?!*7Yiuv1K1LHHPTYt;2il$IV6@L -RzO9KQH000080LuVbTQ8A0QPlzf0ImiA05|{u0B~t=FJE?LZe(wAFJxtKWo~3;Zew|0XL4_KaC0wTY- -w(EUu0!)Wo~3;Zew|0XL4_KaC0tjd6iXNZ__Xoeb2AB@>2q>Md7VdkPu80(u5`$yiAkjB{!{E5*yoD7 -nS~ZuJh5P-H;$JNqs+j&bh~KVRZ#U6pe2z3IVDz+Io<6t-bV0*Di|2HM=p=IvjtymSRPy=Yp0>tg!BD -#x9@GV4Y)~wU<^m<|{f~KB66qysk=->B>kyR)nQe-A7TBV*!;^b^K}(c79w#shq!c-rm7)_@Zlsw0kA -6eBXVnBcMkWV-rwrwjLg1Rqf;B&+xGCPHqs;T5Z>m9|12n0kMORSD(S+uSQvPU1J5>!gcGnMDrbAtzN -UmNy9>wn36Qy`JB2=!Q}X5rE&=0TIaF4cv3!on?zl(lh+cjQEL9dZkizLbr+eS=jbx4Ob|OgBXmU^XU -fB6PoyB&kFr#`@+}8;T641{t#+y`0Z9#8woq!BlM8alwwOJDq>wJ4)s21Lo3k^L$&3w`c`oEYV}i<-# -Y{N1T$Eun&(rkiac{o4VZmEMt9(qO~zL*#a~z8eO+Q+^AYWDBK^A7P4sZ~@HTvjij!$;`| -;CJ}_e(L78gE=&(^GOOs#Dy-ZEyhjl>=wtTG(`*6z!vL)Fy_=t{6mIXj5J@$GvBS}&@Lvvf{34tlDL! -RrJj6d_+2BRkasfB3=|4N-vqRcL;?jh6al!VUGGk1*5Vfr6jykMij|VoiCdH@PVDfJBkQt^t;^xoLtx -rzC3E6YfeE5iut&@}7k=~Wo4~-f4{S(*!=RDqV!{E-YHr6i#@zhQ2(P9p}EqRPE%1YJ2`;!tdD8Z5w0 -V~%Xknw)QFqj_0UAB5@E(1FK@KVJaViggkkt?JrFd@R_F9hqGbpHfN^aoH&0|XQR000O8 -%K%wh)d5NWBm)2dNe2J`EdT%jaA|NaUv_0~WN&gWWMyz=Ze(R{V|ia^a&L8Tb1z?TX>eb6a$#_AWpXZ -Xd6iY)Yuhjoe$QWV=*jLHL$|j;2BQsQP`c7}gHei5?Q^2lwlqm@8@m4QyOZtM$$qf;AyTB%-FM%2Cza -0BAZ6Kx)}fT3YinEp)ifrkpiSdvvtiAxt#TfR|GvGm&gAI*Ze>f>C40;yEmc@!(|H;Eo}tC2z$V|mcc -!&h=uP#2Za5T7UCG?kR)ygNFID<(Hk%b#g4E4}xkqWWO$xMr0hRXQ#s_zs!u%~TF;nm`2?;-9Kv>4%0DvwB?4bQ*oj>bz -6O!Fu{={dm)(GgFaSJGRJk8fkl!w#v58&{7s~!62Sc7qJxeXg)_f2+k@=6@2Bt -f_6VQ0K486i(jj?G2}gXav!lEQ&9irI-LjgOafi6~xzm=VU5XDKYZGxHs_Qe6PTXaxD5UEBjLD0$)#d -8qY;|$^RldLa`1z99Z^+Cq1*HTVtn5-;Z*am1^qd3(y<(Y|jj8=T+`OC+CFXOxr8@bBd6+X}lRXk%wN -+S(okhmHPb@}k`zDe5jQEHSBF^Bpx>*xk2xOkO&JijoNX_jvk>BSL2hPMNTz_0x-ROauw -5vd@NJti&cWzD=x4n>$I&8oYZ+z5eJD8FYQo*}Bk%ID2+DUN^Mg*a+;H)yLNJ`tm8-}{L%x9bv|uhI)8H;w_JBA6u;dh9!I3TB^T-+E|%oCSi3avDMuS1`J$>8SB&QvuD7c!Jc4cm4Z*nhWWpHI~WMyt+d0%I8Z*_2UFJE+TZERm-Z*6d4bS`jtWsc1X -!Y~ko?|q7;UKCm%px{x#gM9&!5SwiYB-yY@ihX?v9xKqDBgUtrl2Kt>KyoRdDi{AXlh^78sbsLGc%$UhGGPLFmYt -OqID&?H;g48+;z$j#>=9MyVgtyX^ar#>oH$3d2!0VP)h>@6aWAK2ms3fSz7=A000620000000 -1Tc003}la4%nWWo~3|axY|MaAj^}Wo~16UuSY}b#QYpaCt6td2nT90{~D<0|XQR000O8%K%whtq;_lt -PlVI!#Dr{9{>OVaA|NaUv_0~WN&gWWNCABY-wUIOi4pUPE$oLba-^#TU&GD$dZ1qUs1tJ+i_&t#&*wm -PHb$bLADjg23Eq|r}s@j+32*8I4WU=`Str{mTt&kpl4?$c6q?oC9hSL`DIpNx8=R-j|Sd1Ik=Uhdq?_ -%+tF9sbHsn_{0K2zXR{=ht3b<1s)9^SWf;qBrlbneXe}>pZqBtFujSn`oF?)$h^3cky`0M+lYSOXW?2 -}&h{pgV^B@cLWG1^R&QiQ_(*5dMQY$5AS+@A)Q*%S|_PYo^ -gpMFU#OX$w?4JP|p6`ZOdGl6cWQ&OJl3PhFV&^suZ&Dt=Sy74oV4 -~JZGyQRmP;pU@bf6=jVl4@HwlV8!}12bqY>VvY2B8Wu}6;oF}gkkR{@?8Z1=Yp1@C8LMG0c*xb%}JOp -FSWEdg{989ugtkPT?se`!^YdFKIi7iP=NwOEdgc+>q(tKodPU2)G|2tVIQFHSf_x*+!NWW7f9j0E?B} -qT`hPX|Z@ze~}${l!%k~Y>_6sRo-|B^9I&KEWG2s8Vpu2XsppWUY!Z`0??cSO)EH%9taN5cEFKvr85UV&#^3u!>c2L6U -{s^5HeV=ToWtpYr75)M#|^q9=$JC)WO#D|a;& -Vtwm?Zw{;e*0-RN43uD5V@V*w4Hslmkrv^!L*%*y_FV+&~ED7IYQ@7L+4Ii=gaqL>y&S3J7+nA9kdO5 -D_)1tZrXdfVedZL%LeT+I$t$tFDu&e(JtES2JKZvd;LCbUFsWO?U8%c@bgujcAwmapRekE-b;Jgpso4 -&(>~g(2JPj+v~@4Pv3DPB!&ldJ+Mo8(Zu;uF?yH~o(Ox!aI|tKlCY6TH`)C{XemaVF)80>YdvEs9UN& -euN6=n1XfKbTy>8G}Yw+UyJ=)7g20q4DmyJAF=e1Ag%SQUG^E!aGA-B3ZgtoJjcAu{{^Y5XwK=w%nM5 -^>JLIAEcP^X+GO2;2Epy+_APn2#e^?|A$fLl`w0%GS-8YO~CSsTKmU<43cL)uX|0fJWA!i!P(T@`OxG -E$jbC(DYdVlDwd^Au$P&}v?$$yG?jDbxwRlVA{~axz1~3-m+)UFHZ~%F)a)Y1mUJK^zp7xC@l#{3N5Q -bv_Q`vic!m%!8|3S(Jvx4WX`zNjAJN#{hLx7{h&V9*{VYowO&Y7pJ5)M0p|9PN&jG*%qO?wsO!2wp1T -%YHhI=V>Jnunv~Xs87LX3+L_D}TvO2)1uHp;=r!_vO(1hHCG;#=QSob)vOj2|Tf+ommnG4Z->h!4!K@ -#%fsQRwGK?z~TkNfJ;Ds1rRK628zQ_Fo44e+9s8o_@iE57gvzjHQ?4cGx=}9iOKWH%xQxtqO?qI_M23 -}fKHCr)wkS9L3<;d|KeQEc5vODPa+>txz`|@_+NxR#1eBbTgNn8v@_m20~^&R=dw}#(F_k;dBEuVuvL -Kr?}4+(;w`aPSJkPmLx=|kWb$MfmBBhHTa2yp@1sF70evF|*`d35?C8jG>YR#Y!;soP~$w~Lb&)_&{M -Dw`N?JPlvNDb>?SDhKtt{G|0@CYWlS%7D8|G>t$hcw(^p7){SsG*62~mQmkqktWmS1WPf9gJ}I5-bBoVpoK~ct3JX|gs -?>vtmR@k200b}3(yAKCqm)P%`Bem+7eZ`GlNohZ8?-d1ZByE@Fy*z8k;;QS^~OZAgR?-u5DhjaKeWz- -eFz}IX`$L -M049OeO1#Iw_;M21&GDmTWz>P>~uxp>?yThivh=cx@=Q>i(9m>6I{&1{bX&NO!w1*tfx+i*c!xXk&OdG3bu9V$X#{Q~guh4Eu(CGgAmodl> -vtb)mlMFZhUA)H=KSb*-FsZvpWO$SZh7+79rdA$;B)~7L)#m<-KPiJlfx%(I6&0JlhM6B;+=vgy1vbG -hy3b3JdlRM?caLHDGsw3-Me0o89^WS?tbLM$JfTk{=IET|qOe6=%GS|X -RlHgIaG>wz-p=2JGd>l0?M;YUg%NK_B(y`4_l1!HkVZ0u#| -6zjOj6tmgGh{16=2RIicwzKmMS3Yr?%{4jnTB`jg+lo^$enrQL$8UgL~_yM@=hBa4(wJ&B6umwY;c!L -LnY&KgQz4ZHwZ;zmgEVDAFubRpw_JcFKKaT#~ztC(z9>kW&|>n0csmj)OQi__rcUv3#+t0kR;HuVfs| -O&E~nQ>lv_rs8RmYR)ByFmqsjVUk+L94Pog6V9rIidGTb;4XkwvU3er{@j+I*PJonO$!1wAxZKq?Ljp -Ff{7tk$x~E;3w|C`E}&?AXp$GDn;7&tgQ3%JcL$GKun{eje1Ue_=f^;yw7lg}zgAK?M$@4R1Wdhxw&ZQq9-y|=Av@xTW7!x0pB9Pq(_bu7A<<0f^^+k#6#1G-R=CFxPT= -yQBlbc{pY=Ak4q>(Lhf9v#kbhoiz_I4Ur#S3{}+zR*3D!eC-!+UZ&-V?{;J=wr}ViWJx-SoRq-)g60DrqtQQWHF?eASyHmH{i^N05%FkdFKzq$4&H9GiSB*b#5RPR?x?-q8)fBi;c#@=( -Ad>VQW|o5r7fuNLMU1&vQ@BgnR7l)NOm7VCmcyg4C(Cz=xfM}E0I%jOZL-I~T>No8hG?-5dbuH2!GgS -ZwgI}0I -ng3HU2~MvK~vhN_z?R-; -uUrE2(4XYuJuv-h(hJrC1SG6~8cOv|0TqOw|OElm^4IQ%yiYzy+N&PV_zQ=HMBRw_@v$8^q(cW`DWS3aS-tvQe}N)NEu1kb69>qIIfX62|nrzm6+9H2~e4(8t7Cku> -tXiNaObXptBVoYU7AQ|4H7dn0+*4L{<*<$=&eblmo5U?V{>bzL^hZGV -G`&5D}qBU;?z(OFpL^axya-C28RpuA`}p9l|!?)5T27D@$IRH1gAKWc#*U-h%7i-56+=%i_3=0S;PQi -sdb8?A;k)UK^#d_@S+VdZjFm-x|QI;GCDQ!xkOD7>-(W{+)^{^LvaC^PIsH^BPz6t2kswA&-%@8l+ag -ilggSS+8G<6H-QPs%O9MDQC?A_tCvKc)Gi%BkD(;@F=R>%`JC27%$KA-_?{q9FkjqIAqU(FVkr-x5Q} -kH$wR#tki6>r+JwGBZtk&WAOl3=G4{NteYwZG){WtC1<1C|6Jg>CE|@XZxv^@o)veX>lGx1zHDPx&Ff -%{kWh=+c_D=uOutt!jWc>2DxH_zL52b6<40iemXP0p=-P5=@)UZxK!jumZ%0>*a543g9{ -nketXaHZpbN3SSc;r1j${M-@V5(3gK-cIV0OrmH$Qlg<(3W58)&dkgJoShxk>x0}PQC>U5}3@3Nk|h{(`#sllx}$ed8G|TuDZT=2`U42}a)?G!Z#~*UB<#k`YZ%$HM>dfTy#e%w(qHc -#28>YN?hCtXAn+{-9eKF4WZ$`aaPORc*C~-5ya&WPc!Gz!24mZvYYw74jRdF%%n&^Kj^W6zFTpV6J05 -I)w3#bmyxLH-#s2_MO9KQH000080LuVbTX4hMon-+40C55U03QGV0B~t=FJE?LZe(wAFJx(RbZlv2FJ -E72ZfSI1UoLQYl~KV?!axwc?^jIH1Few;2?s;sMXZs8fT8HMS=wp4+U_>fLE-P6mSTuE=d|1*nm`?7dxRrMl%DkO -!y9+rJ2B62hb@Sq`G)Orrq$%bM`_07&6uH^pEIUI5^(G$DKFo&{-{T8eI76?;E+2HdMXTR4PtPTH9kr -Hagl}kT+-5-7jg_RWLoNH%MuG6`e-+QWtU~I7Y_pU15ir?1QY-O00;of0 -9jjIN#!pyD*yodp#T6K0001RX>c!Jc4cm4Z*nhWX>)XJX<{#9Z*6d4bS`jt?S1Wb+c=Wwe?J9Qo~$Uh -%*4)2e|Sf|lXjeG@1)~w>~#08+oPsQ%4VA))g)y{lkIc%HTL!Hlid0MKmY_KB|US`%suNLi6jDrLRF! -@Pylzu{@%Wb%Oopqj>NK=?SGEH?CkFBisN#zsYbgYS{>~O_)|aD`A+&SP8W?hrC&~}s;rKL*cFQ^x|v5uq9|c9_i07lpuL!gS7njH$TNf)!lN2y -KFx|*dHLi@{9Zim?@zNj2g(#Tr)LTY^_@qcFVkkVNNX7CI;vCtX-tos^j$*_;n}jtfURi@m^(?RFS0z -nz=lyRX7HJMnRA~`ua{ZgWCc(c+Q=`p+2pt^X4%c_sH%~vHF6ccFX($a?6S(~C)G=RALFNG(SH7^%nJ -HYFY>IBkLj}oEYGeLr~tlJDa=2;2M|9m%ZBDNrB48$P#>!Na1k}PMKn*DP`*Fa;t-mTadHg^xN%JKl}H;wK=0f6 -ewY7teaY{d-pYEeFnMOKJ7!nN0kB2AMTq^e$4$Or5KPwR4?iic<=flZe%-$hm4C_TnMqnMNEck@8!_~MI^cU12%+jFp;a_NbK)_DeMNru^5_dLI5Vg-sI(Vl(+cEpFu=g6k)!Iqr#-s@ -ZGsoX&9ID1t_{;_~#(p8~p+Q`OR>os5p&qYusnG&w!{u8I~5EpJ^-?^Er)svMS6XsPpHgFbh3alewFKcDuuw62Rw#d^- -fhSi;A8cc1tg(Xmbwngx&#EmdMQe_&wf1T!ZDb!J*X4@34g4t49jMjGe%%ZD@Lr}WV1_jQemL0U9k!j -2<-E9E!DsxANqi$y5+(LQ-tPPt-eX4n_8J#v4Wo{{oXp3)F0hM*^@NL6jcOCq&No1p!4IXNt -v|Ux}O9+&Oy*RFoJKRe90dT2Lx4c>%&ZwqDBm+to4$D^%f&FkKRomK;+`6O6tjh0R;5*H`oD5lG+f^qctF_AIM(qjo?iKYy!zPX|X=EkI2qXsEh_-B}g -@dd~n2Shh{SwCQUjzAH}a8XT70gU`PB{K=7s?#nCzjujb9MU=4Y)9N1LFLAvBG-t~i=|3-}4Q!z_w^1 -R$9O-cdT7(%&5k_7@JQLs-+@~vq3Cl9N&a?V9O&lbAJ93#W<(njpDl+f~xWXB0+FBKn -SkEWpMyhX;ha5Ku<+{TLR*t&8H{>%hD0xj+z{R+yL2%l|e^zb=r>N`Kxn*;(H_p;u9EmQGryCMlG{;! -K{EwB$f+ULUWiAdrA+F%ND3BS3O*k;stV)!}P0QBqkHRe$;Q~$UVygzAmppyUfKwXtch7pjv9e^RLg| -zI-lTon45ifrbkrG&?U5Xoa+fR>*(^)!8gVO<0JB+bq5{0AiG`1tGqTswhT%g&tA`3;>lV^5{CvMF}k -{79Q+_GQE#N&5uKqJWAei>LBc{7=L%07Wg|lZIKo{6x*z+(|k5YnW1^m@CB@UoZU(!Ws|3M?SRvgYyF -P;AVg=UNP^+zpTnz9j768`29_a+I4s|Xtho(_?~46>@K3?=X`Xc-yA#^T!l$9r<5#iKf%O1Fd;`-x-S5L -x#R0iiKu+3!S(&3kP0$Snzvwmc5Y|gAAETBAPXv4>#?c(4CAw>2aL -JJENx(>oS(N2VAhI}v_;?AxYLEs*c(yEPsXHXEaa8nyLX&>3?LZ1RW`nF<>zxGyR(TY>o5pPH@J+_@k -r-I;I)4K+(KYb_f#NsJyh2n+dO@Ta>Gw@gj!)iPOwV3loSwaU_7VZn4{uKX{q5&qFVl$Axk0#%&?PZHGPWLQCNGZZ3`ygEmk2v9)P$^nF=uqbCeN4mjdx3GKl_T*8_Qy=@FOvwud?;zv)m|0u4jza -H+Zzs*tr5oDlRjliux@U8ZaZ#&S7`q9Rf$bW;ROzUDO-6ge%!FoTc2jbB-5P~+gL(9lUW*#j9bd$$=s --w}U<35>F1lIgNn&YGD!)K=OgkWv0E7FHEEJt85@kA?wr~$W|guPw`_;ER??T3`i*s@4Kb-K9$ -Ym?@j!Yk}sFTj;SV~hD0a5mD9%M7T_K}m?&248=ECT=jb1nW3f9Y|M@x|%QZCIhTkY1CwjWU!|k0Sjn -r95Br=1uv_aP6KxN_xA1{qUxr0_eJ1gIGrY03{Au3A_4jbvSo*0yp5+Q#*xm4?Ah&NGb4s&sB#bUsZ2 -Z|?gToB=iqtF!2_C_9yXQ0zQ2YpT?9goM6`1ChQ-xdmvmyA_U$I_aBp^6N%T)F22^Du5{s0p& -~3NaKof5cwM#HXuTo02n@Xz+Ij_GW0X%s0(FvUT)>rSVPPiNaEN|HC_mynBj7KM(3xK-enh0~R_s03@ -KwVCO;qBx)_qo$1sS3N;x!l%W-)jJAS+1!DuWQEU{?gF0~P`KnMRd?Cyr>~X-U+sI7UOtE@~vnffG}M -MC7~4D2Px6`-Q1gBAOXuKTDE~Lmp+G8S0M+?nfpdj=`DOvI3(Sp)nWVf#Y93)CMTk1XGKcWTSC`AdCz -H5A9XKkEZk}m=!D_BX5o&07yiW^j%grHMAa)An5_x{`5Pdt-542;D$Y{80Zv4I~`U^oFXZC%oIIeN_U -M#UY2(lsV^7u5s<5jP|B?ILWDul3^~lc0tvZ6m=N5e3{y67zD!a{6l`0iIjq9_REm}XG{$W!2%QZMP+ -S;3uH!0OwBnb|^n{@;hdJ|P)bSy0s_S*_@e(BIp^1+pyxngACi+=>Q>HM3E9_kzWBeJU0o!z(OZ?F*4 -tkHy08dlrgr+1!oi0WeiddNd(Q$+71kL#1AuFC94t&0Ahwh?@?ohPhYYNLZPK{*{xl@X;50WI4x)0Po6r+V#-f_T -m$Q-yNp~5lWBoLd@b_qF4dBT^htq1(6p0-iyD@IdO4eA?_>r7m=iBCfez%ZO -Ysv1I%Ee*QXybB_A`gHHZj`3IZZKgu0c@k02$YlcP$k#D||T7DiD9cymmmV$>FDJF4rcG*QGrE8b(PH -$kwASZ%WJ&r$ma37e(!lwI~K?k)yZ+K#o=SE^LXZt*qs?lwVt`KdB)LGX}L(_Rg;O&usC6u&!lM8p;# -+XweNP&JV-SIL{<0bQpfR!?kCB$re0c_B$0Ia0i10*)QJ -Uji&BCvb|>!|2SNFZ86UL226iJL^Ey8s%h0y*;jx|mapZ!O1R(0hEa@U)TTCRR$PRo>tAgcp_36FzFxgv4N2>^Ria*z;FgYi%ZhX& -hCfC3(a_)T-v|$@r!khb#dQh}5u@#HbSYp)3s;*CkJW=6qneg0hzz#H;NsB6??iu8{J{*zhA^F&9+!E -}n{5#7T*tZ&Re{rNR0Ck1!`)Gud3JZ}Jguf~ZIQO-<%m!==reW~8_Fzxd(^eFL -h*1Hv^aI3@wbHah)+I}@X^?1oGP=atMXH7C?S`ht<(kl@9R0pJ7g7J#@`X_Rx;Mnfva4CW5&731BK%f -CbMf)Y_6$&#^X2Ek2=YZ;z#O!v0P-r)$Fu{x4j0wTW5HP>Q*KVr4r|M~C#p>P7?MGPk-JS*4SW*N}KF -7jy}6*s5_gX6R3C+8;@BRA_Iq4=88{!W^NydI~LwK&pbEp1@{S){uomy^{O7UTRnu8qZrk5LBDwHuJk|v`vfO~i%Beyu(#8Ih!%v?g#b?l;650rcC!l&&Y -P_bUI)?m7R@2cmppiIVfg|>gOGSvVxZ6ebA`IM`h6Ic`g~H$?I6fIyY1JvRnpCv~8eqM`BTOV*nLTxY -Mq7q0j_ix<*Vzpf)v3)llg^moY4OSo`cTG&SuS^Tkk-<@X{V@R`c{y(0_sm03sT7}b6_?w=$lSM+dO( -*@Pu$J6)75dO_;gg8*^50IW`HI+&G$ -kE-`&t|PC%5A{z#*RJM@cl286Mg?r)i1-43*`cMG}e&1c}`XbhBCko&i04eXqdF#Ply5oNrHtq;BZ`uVa3qQM&my^%fm*T1^)uZKAud=)C60A6I>eXg++ixX1r8(3&Umg)pm?~V%<)tW6JY9BfAdz|EUIa>EDD*KrCsn&E2 -_-n1J1r}HT6f8#3D<$&6sjJv~LsiZaHHYI^5Bsv2n9m)68!H3{bSA4nz2d_c~H7n3{F0MM0jKw-aa{t -ZQ2b(LfC1UOej3kDQKwsOrz-rX5FEQ32SoOwkcfrgaCLWG)1f -^DC72WJpx4qWHCz&vEU1gHQx3jFqF -fJ3NA6A4aWN{eK67>nCYX7hjQAYRd6Zu1E#*>Y!PR)LTny2oZSL{bg{h7vzRhQv60Qug{6Qk51a@-dr -RkKn>?BXn>{95vwYhAZlal(}M3=iqH*ssIn%l6 -`5*I{3U6gvRR{Av2ua=3SCT04upm#F$gWT+IGn)?nWwfjnOHz$>?MuAg+bsdW8y}yb?)0yT(~g-%cEG -e#_z~`0w|i*ORf93?Yw^W+b^-E+|vG;f{-%^_>bT5EaTt)J;IsL55$x|K*ByDD#JC1;c{0q)|-;c8H8 -i6q#<&bV0r9+qL;qy5g{a6t4pW762Y9T>}J-v?K-Ba!K^NTEEx0=vz$)%4M(Cfx(FYvDWiLX)I;eA(s -zJUld8-4RhGbQIqx~M0XT;FqFkbEoD2skv_C5y2Qv9T>mU!`KLL3o37Sr|@&6@wfAGtw -KR3Yh_nojZ=x5?d1I)qJm5GHR1`|N3I*4;TmyBs;Dlgsh)Lspbm1We*Ro3!LeTd6Zn$M)&a#27-Z@K} --4X|IRF>#S_xD}U+m5-joSLs8G7@t{WKSCl|z;o=3z9A9_@66*ps#(nu%YrW5@S-X8g5*;x3jib}jKK -46Wf#F1Xrt&ZR^)}@2YWt%)-=A&av4BSz({4n&wq-Fn{vo!tFR1uv8->Mpu(QCF;QZLI?#+ZFrwnS6q -9K+P1Vm1XP)nWF{fn;6c+M9u)nhAvWT#tzur6(D+|({2+58j)L-8FLZY^tx=Q8Jd5?u6U95 -W&IOU^czR5W028uv%?2Uye=pnztc5=8PC++cLxLQkL|{-b6tB9kKzJaEji=_9mT|_o=zZ8<94jgA8Pq -!9Z9a;*EtddP!>i(R~u&L^(WOw>f4lJ%yq-<-auvhp$THU%iG};jl{_QQxId@th{TJyLc5P==(@wG`&6YOr85hu~`>M--bXB@ -@1=@c7GWswD*65_pwFDK$@_}dRJ0`9=X3pj4k3%p+5Oy$YN)O7~YyR7)AR(mjJ7%o=!#ZNX6|DK1_W< -r__f136~ax^5JNWnaO+>*Mx71A~q{9tQPV}sTa5YVyqr3lMCve}tf3GeC!`-J4Bm{glj -*#PW{utc1Y?0QFfA`s^pMLT5u?VU!)K!aLMe!Y9@jEjS2u3nJLEe6~q!3r=a@OCD;h*2(Kc9}Nx)qoD -VzpR>66|Y^6}SU50JBIZsp@6cutLp>xU3eK8NqWuV7F~CZmcEvN?RlH3xp^>l)MPWMG2JjF$;BoKIa(c)-L2ZWCV-=B&tuTgDCD6(R*nC -XqByS`dO6>MQEbP!7g)c#avA@eHVTmQcyJK7MOkL#Kj#};$p#C9g$|n6QW`S%0!J!8ttj-O%m*=7ZJ= -lyC5wR*I_gi=zlqubp-a+|)Y%d+sbIwXRvUyq#VpI1KNjgb=845P6^ZrqWsd9q-A4E;3xn>ned3iFB% -e#>P!_gKMoyw-z5L?IP9>QKw#)%za=zX{)iq+P?h0Aqtwq%C@~|uYasYaKaOKXhLfc-a$+}q@ZRAYVF -TZNGPyW`keR8#JObpYML9N*Tl8gi1cPMi+=y}tTKXS&Q=@yBDnG}AVFK3BG31RQotz&NSWJ7j&)ScTF -n(!mdoet^E$6K1_&K3}1W((7aa6NtM}#lahB%8Nvo`OO(4KkxpM@43(~vdT(u;)@u -4C&C#CbbAMH*iQGZYzy)6fL&$mB$iBV0-Gajyb|%24ghn`Y>mzg5H?W?wCo^Fox009MQ}V4`ni1jtKB -x3(U%?Ddlag<$Oc&&k71G;Mu_MT4Oo7V&zP+eu!>32F(KA{>D&V3(6Ba<3c((*-FWs|prQ+%tMyWM(A -homEJJe;nDJ!qKR0`8k-K2o~hUSOX$Q^~S77K2zaT#q!rdmgF{hUzZggzJaVMdL3x`739KexjI!+xw8 -TavwR?0&A$wEI1Aw!BOx4@oxhh#T~xWwDH#ek_V;DkPc?0lndy&-BQKC)ntP=@q=C2Y{DM?|~k8thQR -k37MrAj-fc6!2`O$hB8~~$nlu&f;#d|?&Ju){5#ZLCrIlKv|GxcsI|@N!-dyElkL_3JtHf3B99^I`IK -_ywV(+ou*t#F*8Qi2y7LszCyn{VM^yf#H&TyU6eYA5Y8{(UKP2^WxJBwivUMb{NK*-rLoB_J*9!{&DF -q9WHqR40O1J`QG!&2UfThG|ES#(eoj0c|3gPibVahd3ZLUTHPt?%%P*l}2p@CfMzAL`Lf?|AqQl*R39 -#Z%XtDWoRe8C$Ad^CDMg>d$XE3S>32*+8pcrlsyn4x+SUXEe6&*D={4w(BdMJDZl4gHREYgs1T%(F$D -8<#)HF|kgnb}O`fqpCIa@q>FWQR`gzyyXG0aHS$RkWq|!ek4~0ig_ -CS0G*JlQSE;@&*>C(}9Q5JL(gtw7hr7K -G~^~NmVW1i|o5vJ8>Ug2l%)yvb!8&#U&clb7>Wo5bMR;P*bVEs?+(r{;JbvP(95=Kbqjp0n6`|GW$tJ -6mQzGVR)veoF+au%%d?{X~pwOL0 -F_C#D>w!ZR}?Oi(-8jLgD3rOanl8(MC+Vi{U)XduQ_m~|Gmo+!%sybq(TlI>S!3LR8zu%*@Z|9Q6pS9K{1MT%5WV3dcZ8E_v;^T6@h -^nl{j7$kcU|06+IlV80$_Z+`Ke%zv-pQP{-COS?>%#yF2yO#QFq{2^uD@(7wf;M@b9u`ndFlAYt2CKH -yqja_HSELQul&JFdEVOU(A#B5k*;Nkx*b+3cRlo0i8QM7z!Nic-%n-~7Jb%9p5bAnzI4t2Gh6DpTHtB -3PxZL$w8}~q-gU*cn74yP&jlR=-<(thk2lK(zxStj6h3CQ4eJ^RZU7I_3ABFDbJ`Ix$ORY1N06yPZBid3AC6;`HQ=`(EP9X#X$MEByD#{u -k4$z2C@peHnL`8G~AH9G?TM7O;H(S8Wv@(k&pqgGa^-sD`iTEeCv6in_r;zk}{bJgU%3G2lv?(lr4)y -lSw>=U5*y$JYj_AP$#}JZY%kRE;_&!PnQx(e#ynOv11%12b+B#EZ(;i0?ABl(VI7RUCmp|~ycL~W!;aRX+woqpRMZ~LtIXm^Wa=)5dO^3YeiHy8 -|_nRf}8CSIOy^}PLM^D#DS2gKjOS=2_z8k$oKC3fzcnwS{!(k{)3_FmE?o6$|Acy|I4nl@W;(+G?9Tr -;VAaWxj=vOf3W35_Kvx^IymkLVgcn*n2kAuq9BW@8=^|JptmDAK(zKEqK%J{Tftxvn9HaPY5Bzq;qbK -kPW2#eaXd{}q6L906F9XLFd}zxC;1{D(7r8&>@-c(9rNThYFk?jD-E>KhRa(6+5n-?X#&S5e$lM*P2u -`yYJW;-9^Cv*g`m|6m={t+sy+W^c=;zjGdJu|x@e&^V*Gi^r8T8-e=2PC&la=TJ@LXvsd5ZyLw+gySO-c^UAp^d-&(zUtj+oZ{E+|eTi -Y`)hDB)$B*$xKK1owxHtNZd8YEavp3Isf#H4IKRkRgg8z*z1OIzFGY0l!ctroHW%m?ASUm^b2igVkSO|IT;m9r~Jh8LLKWQ3(J=GtudVFn79$2?9tc3QU+@ESQx;PDD(SNBH&iSn8J^L -Z@9G%r12G$v|ws^wT?~|$dn9SBRqrLX^O?FW)kd}Zq8ETD;{CutLE)FSe~OUqW8Gff(Nc}g9pDfRTQWARu#HVf>7zsj* -@TAp#VeA$YtZZ{{Uq9x!g?Gx=X!<4MBhs$INI`#$_J5f^Y5+sTJxwXJRO{k3|JQY(S-Myt^!OwTMiEB -eR;=spD@G -V|L|>(S+U`kVe=0V+^Q{I4h3l7xzl2aN37+I4*?Y@;*Lt*bK9f)$CN47(*^&elGa6s+mT=c;n&N{cI% -RmTg$lazDQSPY_8+ff5S(ZpO>2_fjp^6>!SOrzS_KPpAK&Lly~A54{V{Pt?$y+Q@bX#?QFg79>cP*cV -h%1CI>BgQ4O8s;~f(tY>jd{VL(?pgyLMfceD?JZ}X%#rZ`TfdW+Y}*z0ijQPBG{_fmA1q+IX$NeyA_a -7&ih{x9!O8?C$yH_qL9)QSqX9nachXmo3@TWz+|rM~MXfT8jSO~AfP -V&&JP+r@#y6nZC^8`thR -N@4mHRcpC6?AnH}_7RaKgw|6;arg3wNWdo-AV=_+JZ)*H_{!dd|Yqdogq+I<@RCMaREd3KkDMHX(RL# -g`k8>M1Fr^=9{NljT{nYOhz9uhvlo2y9!i(tu2BM?$>HXM?tG`m~-U`2QDBO9KQH000080LuVbTb;et -fUpYy07NPP03ZMW0B~t=FJE?LZe(wAFJx(RbZlv2FLGsbZ*_8GWpgfYd97M|kK4Er|KFd2jd8Gi^{Ts -<>$NbtZ7*q31W1!ZHYv~=hC*Ak&8#d5qP)A#;l6ukh7XZ?c)fQuf~+mh3}=S(_>t6@U3`1N)_6Z^N!`H|i5FS3k*?5$jjT#1y`1*><0{dBB%Mb2)DO? -}TR!Cn`KJms~7XYfT-jKh3+UCGsGnI -H&P!<;eLFTrT#Y3D#ccle@Inla`ozAElM=q45Lg-vgYfZNRzTEN>SCv8QY3Fp@+$eD=}j)@%w0mI)c} -(t+FxDy`^VSlH`03_D@Hn5pe?5yCN41K4Z471g~LT?sp<*xe%#HkuKv(u@6EOhiWZoj)?xB&PT|XiVb -*H=CV$bNQrDSW2nwG$~7eTksErpgWnZ01rBu(ia|Jh7eCCI5-w!M5M -JM3#(^lBn4iQYEV9nJ_M}v -S$p{@ci)X*)#Y3jiLn5D?!|B0}+nTw&01}N%`Bz@yttJ@x2?FvyVVMr$w57wi#b4r@x<$GujLbzp+nJDy3C*}%ftj#nvzcwS{<&*m46|MgTfQ{!P -+#g=mi-1<)eg}^t*6dF=?cEXiE_Bc}JU|PQJb%b#s=G2)^^wlgYv5^ltKphU|V3_>T$TLqOJyA#vHsnR^NTovQhG -81&i65Zb^IHJ~%2Az?Cmx`zDm!ymp3b>z9ccbQ*G)b-dBi1rE7jZj5K&jwnB7;dVHo0F-ri?4L>CTFuO-wU8yOKD?~e7=H#T;fJ# -IYGcUjx07t?`b7Y6v!ad$Qhb1ZtA0P3Auwai#sqbF$tN3JX_t#;(+Gx8h -`81L1DSLHv*f|1A)wHxgRKpChWWzp%Um@qgDHR3kW&eaUm1T!M5$&tXT0(XOdq&UH9r%Pm7L1+|x-CU -EXbHdQ3q|w6<~sc6+psNLBKyI8$gt+68tt==##E6_kVk&Irgsm2Kdi>x>qCM|quOVOg@VWob;kA%umI -f*8FPP5W0(9*#_W07YD8yatwi3{{}26_(p*@`c~=iC@@mWm$ru;-QAE4U}mGUWz7hd0bDRAj98+{*u- -c9YtY_IY)wrHe2;}U!;euQ)%^6kfXnZgIU3K83zCI8sM-dg1M4Jk0lArOej9&Y$0sh-h1RwLZXxjg6l -?ko_;x~x=GazNH$?$r+^BO5wUKtjUcQ;z8JffNvrXU0cyi0f{p)pIj5Zs7BT7+6>KhXu2^|U5R_fWE+}b5b?g%u2<;Z~J_FA{xgCj*3(_pH@ma9 -pXrW1eHNDn;Y9%eWZ59#=O0}A_Vxwu8y}|;V+!56Sm(A=OtUhVfn#?uQlZu;o1_q)JCT12YbV7?}j5h3d*I0R>ZBVKt6FL%u*c{jjh&lJ!884q$;W75W?*P+;OhHmbf$QDH;2V(Z6A|8+yk=+q2$yf}kK`*_>U)F0}il8-37GG|?~> -eqdjb_?yHby8hBOAPHKi?rw`pcWnHB+4ZXZtYLxDAQ-H$VIJIWAln)KwjWwxf(U6uyu;ETH&;gliE;4 -xzp`v0=i`#+R?NZS4-@j&?x{YcT>sJ5CTV2g#r<;)2&gg|3r;L`%VVy>UaWrHM|SC^tYZ}L4vM~*c&8% -RaHe#QE_7FG^q_*^=QbY?~vkDtPa~~LMz8M9NeAG**6vZn`oP6L4=xP;S+A}lY3uOw@;IM0#lx|Lolc -2rz=MLRGxyw77e427n?TXy)*qM!BZdX7>$q%4=7LK#xpUrEaDfaz?=6vEE|m+&yx^tuXlzsH-K`KwF~ -4}E}k!+41=FI^-~gEXZLkGy-VFOXyJqO^U}5U&IX@S3N+S}>@Os~p__KfmD^tqmW6KOuOK*X!yeHtE{ -ihsS7TSqQz+Nw^gz30#;&G|%jJ||o1`NQx!;Qv&&D;hs6==}A<>;zk0ox1?*9)R;x{bzQx>mN!=GBeA -F15~b-R5*lW05!$+Lxx80~(3CSp9;iu65v>!@@_;|)ac@XWjm3rD|Yi&d!1&+Pk#(wiuilCcUX}?=a9aeGBj3z=y>4uD*M*?!q33U;G96`YcRerjJiyPn8*{0^&OGzQJI7L7P3uDJi{KyI?;{0lYWmooYqym-UELe={fFqf -V=no0uJS|^WIW4`H$WSp*dT65hGYiH|fdEs`*T)#RM@hX|(%QHyD1ndJAR4o38Tw5D%OYFh9vDpK`vU -GaY@AAH+9+UiJ;^tf>nof7Z_Sw?EP}KEm#xCAIoRdf4xei3=KfLoNOn{rrfw)7w#rAxr4)#>?RZlunx -Km?vg3VtaiTZf-){mA!e-j}Au6Ep+euO1~wuEw@#-dn7-G;LS4F>E}ls{JWyKRrIo*Vzx)>cZ!<(Wb^ -hP!D>t6wR#t+bfJj3W{HyK0=Xe~Z%CXOoh~xx-GUITb!#}bH+1bw9&<20X&0EdNi2H3ko)g%wBj|0)G -z}72T)4`1QY-O00;of09jih+_>ik7XSdXNB{sH0001RX>c!Jc4cm4Z*nhWX>)XJX<{#PV{&P5baO6nd -96HccH20T|8)vP9zRlvM8!!@Hsft|kCVvT+Hq_@+v(11lGURmNJ31J91^tb?8i}d@3DXPBzu#sDgY!v -QjU{pb-!+j1PX=vu0p|kZ18x%=E*|Dt3y`E<=`bg>Fjm(*y|*{%fxCeng6=So;^KyK6v)@*>m=*h{HQ -}8GaUN0K-nioX0s|Fqts9=Is1Vu9KKuCQG>uGtQ2aB3^`2z%%&5GZw;nvCPD*kO}-+Wt?w#ED_cv=j> -$idVG304&<$5VZ30k&rYu{Ca>OIon2hwLm1atX32($go#a>WRktQlRTfCb@b0DS*>6{=4XSZubHfbnX1pxN66sta4aLMO#%40>;K2AIaIC@^j^Q6e;J -U3I90*RvE7E(m!g^#n&Ql5oCraoJ70D6mW0JhBT`bfna4nF~pwRz9yg$!pASHIIJl*=UB^qGjKH$1-S -bvj5KV8X~yDS&||^vs`5<8Z^LQy4XUGd@0gcXBm-dvrQE9$#L;1kdw2KmL5Xi5M)K13yRIgWzeGnXQh -x@2-vqFS|fo;9D3aF&}mBc;5Z_r_PUIp7YHtx??au&PUxMix2bpns35xMarvR)~z`HV@cffOvsO)h%E-S>*9QQryF2gA2xFz=Ea0l3-z4Wle9+)*WB>o9_9)PZh)CZy1 -W3ZX|H4{stVgepVUtBV&-R_Kgj1mF!?b^@3#N`F#x4RSERgtYEoGV?4u>mXOEE*L57WCe9Ht^2iaZAa86G@)@!|ze|832|Sf^;U;}wGueYWMWL)<*!l0*x9 -3}%IajhS;c<6vjNR>q+FQG~N#7(j~o)~IK|CQAeZwhQr68w*R^ABRiHv&S$mlzOSY7AZ22dteYim17R -mw*vM*Ba$o&bUk5CL(_169j-X;JWjyC#c1imn@~hVCIGCkiW05Q+cih5CEJFE)c_4;AOHu3*hWdXV4! -Cih$U`I#DM#TZxTR5!~iQ?+=TI*2OX5GBoBy2C|I%d82C5ee)buV@yM&e)2ThQr^IiMO^_Gt`N1JuB( -W#iI?NUz+JJq`!dad~g{(HJs05BwPe$jHZBY9J5)sV5xZ%bzgR%o%rI}1q-b^m9PA0FW=SQ#qa`aEAP -)0d|EC;(MeJ=p9J?MeJc4)?L7f0vkXce}lE -77$Va40ycA(T@0CR3km)KRTVxf-k}E{H;hNx4fQTebw{?P#*jP7*wwg%Y{&;t~txWHz2PG99XhQ8GcK -_VE#PJIjSCbQENVKkl%)z6b228JR8nLJcJ>vJ6woN*}xnK4%&rpK9h0lS^!XEIEWZP?m7S-I6nbQv#k -K!w(BI~HVrHM!GwpIRFw3bRgB!34?OELi=x0~8;p_&MiSx7tLSV`MnR7gX+LvN>DHGg6zs*l{hXnv_S -io2>&R6oaBy~qb%b!`?_b!gT0Nx+8HY&1mZlC`>c4QFmWGkkI?OQaWj+c>oNRl2ZKE=>PGE^59)hQY7 -ztyQ&k&%nA}0eQA9bcp(MCoz9>Uu52z)iWB#-kJggNt*8A8+h>vol}95FQzOu-z&dNfS)jE5lZ%30;g -@u}Ffe1@aIoR^!=Q$80<9Snh^u~$%YrjlO_2N&RUrdXGf>FeXT|Ln$N;`m8)T>-_xmxg`7_PoJY& -P#%&5}0e%5|D8#q)9UR;=3c+^*ar4ntR)9rZWOvi7h)2g51@#%m@ku(uU(Cvd}LL_P(#HTrwJYYk-;k -$`+bus)&~%zk8vAaXj4z)8iS>uHos+6h;Zl2%v(h5>29H$l^2$SDWwh$e365r(H9jXI_SMSO=oTBOw~LfCNvC6Qi8a0 -UO_p$p;y?{B1yR(XaLr;HyJJ3Ys?cXirdS&*i05oa4z^dDDrdHj(KX}&__?C6_6%DO~u^dL*!WkGvXZ<~)_mKf4@f8P)Ga*J-|7(u*JhcN1I^2-MH>3`ms8JKxhb#Z%m%rM!FjSBnu!zC~^n7A0Ho&%G+D^%OlZwrvYY_Sn0; -O738gO%nL`HwdbeZN7_%JFwDk-8c{g!;N(+6F4hGw3ur5++z2@DM&Dzpna*e2U}lWAq3Pr1%tE;vO!f -b$OQ$@6khw)rut~z-63QdTgU^{9NvOB1Z;N=(>xRkDm>Us*rL46eoR`3d;j2DJ~Sh0ijK8&v^nE -NHnLZIzwk}Xz{+>wA#?jNUx3@X{j!Yb%=a!7T6!{Knd-3B5a6wxMF@Zo=M!klV5L-giJkq`AIhIUIm-v3uNSBkYuZ&{yvyvl`@Qm=MgIT-G&qzp -^T8(cSr?0K`BasbfMi|cc3rT0D1^X596TroV-h5?B*x-i-JhBf914C;EGhpTz4`?Ymh_YY6I@Si_- -?}dj)5Ff4bhjwLZOf71!Ijr(H##(wb!(2s&==JPSXWMmKV*;ACauQ9mqv+bmTBXInCsjg?fdsvv0xOs -gQ{t46?P2jL!{U=`@lb-wgQoxjEEU%)hwax3J2qbKS&`p!;p(#@e@D2!6RMLlREW_*REW`zE-QGwG6z -v%hdWk%oq{9sz4808oko&Keif-^9kwmBLc^rJE)mSFfY^ZuETRWi+s()I6Sxqwqio>0S0k5=B;Rb?{-SWzl!>=9(h;gf3ET^-Py@O{z1nek+jTWmlLmV*REzw7 -fTk#kch#*cw%m$DX@HiKE8q&AmLGW~k!||LYjJ6^0gdJBE{XS?jLN|E@A|A5;`=9?}ewl)V7Pyn6tH~ -*QeRMv#Iyzw|Q_gn_JScax{kOo5ZZx>Py@mZG*Mch*G)ST;1JE<>WOaLXYXqL)QxROHmraPk>= -?;8vjJ=~Atjhdb=FP?)UtKmO8PLQu^`aaDlu%-FijPN0JO-VE=0bI1-^enqbQgBI-fE7-OpkTtzY_g?>7NouWME+msAa4nX8f#nUg -AD8Z}#pD_@43K)!{<}Hyy!1xM^&%Sh`M^SJGeg5)g&DCV_i4UGL^LB@lToipzQ149oHW&_mqVSN>JdP63o+2A@`TgaX7?1dcF1Kb1-W_)K+O_a7$ -9gZkx!f#-<}r&{^}s}t|cd|Z-vCoJoG(TKQC)dfxZHye(07jEac7r1cSIQ% -NC=dPp~N+Nt5|n368F4s$-x;d$xp>}2 -@L+an4sjV@&ChMgHJ+Z{z04pa_UkR2tAmSqBbTNnINfkh91q8R7L1})YlgClZ2l){QMeVi7Mxue%Xtc -_g+mVCRY&imR_YybWSMR)iDvhIa(GMGgTiq;iz-muWoSx1^3|S04GrRbz{3u^MrvE`|+rHKw(`T}{4r -P>fE(LiVCR5aPPe%vHjOO_!es&-A(hgjpuyqxgJj8v21IJJaB9DlJ=foz+m6*3!3H&UU76Z(T`?!D~R -iYzV;kP1}xv#+NkO2I^_>6Br`pi&O#E|NMO>GfjfZXSqb*-faaZ5^W -r>R>}YAm1!^>=+JRNG$n?u-ee639T`2C2DpHS37m;T6=jd(v%Jz5hQ-0ODEK*vD13u2WwIyE}HWC`^c -)Hz3Ci#fZ@!YQw5skf6&W)sx>E>kTSsDhImmq3L2(c%BDGoj2&CUkWB;6jhi+8B`7PFb&>&nW`(_JP9 -4dkc2%c^t?jrABU;~O15SSb4YR$nQmrXp??J=c05$NzU%z#MN{sDQ>Xd9>z3u72r)EmlIBCOXL^qevm -y@idVaEZRUHQSnk7pC0wEz!s3kFC|Pn`mgVe{7%--FtG_|%AK@J&;KH9Svuzj$CuDN4fUO^tBz#W6eU -b@h6ceIBU?JEf|MM&(&S2}&i<4kRlBR=MwPL})1S!GLI09hJZ~q+Da8S}J#SGS&;Rh%6*n+xt$E%TZR -1kM`##rxo^%9uZLL-=L@ZFeHum+&;gmfAct3~%K24M^s^w`w1De2% -ffF~=4&vt46VUE;eTh&nMyDLmZInO~#>nG7(u>XqqDuV#4A7T6~RM|5_gS`5{dok(3vF&-ZSvm)bkOt -Z-D2H3em{m6rXQ%ei+Ce9t+Hu4~D0P?VKJ})w?+$=d?%tCehY)b`FRlRkKtP0sNyx76H|xkbSx>hMCAFm3Jxx@pB!T -?2>~vS-{UGlvHUKwWQ`ap -8vC+TUQe~kSFK^`Wwjz=ZqJXQejCJ^53A29x5X+2c%A6JpbpO6*i~`yjMH+1)20y9r)NHI~%~}Q2*~L -ypsk(GfLeOQ^Zd3b}A~a<*Y6?Oey)4A42aj#C$t3Ri@ow{_(n(O5Jc2w(Y0-_=5Unu-i6SfYogzuZ%U2gIUb<#mO>= -ep*-Oy6VUDw8h@diq2*u!fyfrSeQ4c$vQ1nUJ$$e$6Gd)!yJAT?lggqYFdb?`Pl+YmNF&#aU73TEvT% -&V2eK2*Esz=DGL+;p4e#d4EZpd-1~nP-yU(t{DLlUi?tIq|hqAX7uxC?r3ueJ=I-OnBRrj#$&3JxBSZ -mzUsl>iWpzRnyMMrbD>q4Jf%Up4Cv#o`-DQjzF^ZxAqahfZLYYsbbRL2!RftIC+%2rf;E`=qYh{tsh> -jK&ZV?sSB*qxY4sbiEiSfxMh8f=1!N`KAIKXhq1a>R#VitYc62^r%Y=H-t4JBMHxnV-SNnG6YZC@Id(`+VB)-FeTAvP10rsK{)lT=wI6G;YD;iavc2BgTpx70N-L;2CUTtn(nW3& -yOgBX&MH;ng$L^AbO~@-HIfflB1oA2B?zw&a7Sv$VzgLlRo5gL{alcm@`rjfAbYIs|JuFqZcv8^VmJrzN+9uME -QiBbd#_sUExl6et=9HiY~^JICk*DHqKL1wphin|k3(xHAsI->`K`6inMne+x4+--^Zow*`SBs=vCrCT -@4fcgYp=EUUTZ3Df1I;$9A||;P2;$Ioc_;3lx5MQ#c9yiA=F{N?l?7!WeOzY&#rV*wrj_ZJzpX+HcCt -U3C*JV%vpUoVX#+m4Q^3EMK5z%JuGy(lwjRc6Wse9^1|5oRkxViXnr-|$KegiQhCT?FQ$31(R<0k$Sn -m;X=hW8EB^A6KLhQCI>$*k<<5t~(8BIFQ -EYrnsk!$MK*4+co)K+LScNhIOGfCe%mfPCX&e}$ndAI{Z3cmi{?7I3U&6r@F59=0t9KiYyOEsyJRacV -`s;|QCZ&2jBfy(H^DE2(dINuO6IOX|atHegBbj5JF6_xx`t>6pzZNv+`h4lP{P?-35BMXp)Ed6uzdqb -WxyB2L(kNouB(m&|G+Ty9d`d{J|FQE`9#9ksQl3Ngs1X88Ppr=iTSyWgP&I-P(-sqa_nyVx9PkfYt%D -m>8~cora*u_>CS1SybLK6rm|630EGxLI2NfTqcZ3zg-k6nDbeT5nP669Ab&pzr9w0V{BP(Bx(R -HJ0PNY8NZfFi(NARUE5!6-6!XL@@1{%5eSrXPyI;>fhs=4#H5e`dAT+7scJJLOC#Q`zkLj;>*ja2p>> -N>SpuC;#)Tq@gY=8?Q%q%y~2}`z^l*t0POoL)80r;80a@h^pNW>Xx#Ysy=XOO$Hf@u -=`@0UN}SiN4-YuFe+w&vY2M5ne1k%OIpzqiBZVP`LBTCF-_Uq2)ORcfIR(v+&M_-(2|s)x*uNRabj)o -17k30DYksR^iV>x2j)`t=1OYk1KVEuK@W+V>;ITIWbHqud!?`?t77ad#n-=U@&{-CdP6)$Mw}+I+!lG ->LcgC{lvGJc8pYD}g5_`2va31v=~b;ImK5i&z$a&>!GNASZNtfEzsQSC{KlBm&=0e3k1QkZuaQ -6Q0-X0EwIsFBpQ-GZdwii-3L~%Ns{5>1>hSk6Ro)Wr*$rtWk~HVpO-DH;Xp7TGw3(*Yx$5?+5@VLbb* ->PHTWC?Ni{qHGwto7|ir&c{V7)5pfX@Lz!CM61^4U<%C&E?{V`MNeG&d8Q@zke5;BQIKnTh#e(X-58& -Sw_AclB>Kyf*@3yDO%~t2Tb$+iekeTb^!eWM%9RofJ&pjBfwgJn*px~XKt}RmgiUw2*A$*Z -24G&M8-(9`}}NAHj0ej!?=O0FCdpj3OsBlzIQ3W_)ldUg5z--S~lPgS<^(jb3kaq`OBFQn#^#Q2QaZ$ -XgUkh(b*+5{T1%m96OBnrcO+35t?uoRcE_^eBNM>_F?d>+5y*YAcgiJ3_&)XYI7B^37eFx%f!nuPdVh -NtaSE?X(d(NQikI=AyD1@&IBa6yQQok7d)$dDBs%;Y?~OG+Ks6RWM*enGCPht-Tjgk;hgRSKES8a71- -Oy#T>a$6EnTqN01HBJC2uB^^117zF!lwfb`l&UM!=eDuIa!N=#rP6tU76D2=C9WepY9AS1m#5UPH^5@ -Kh8h7e$+I9?+fCrlVCpVPos7VQ1jKe@m3~DDX@JsAj -$I}@n_huwZ#eSVr}**EGZ9p&L7l_g7fmJe$@NPq(bU9%6hh!vhF@bEQC|{TPS&ewbarjr(N71)&Q6nr -PD`hSExsv7cS$0`-Ht*EpW{gIWbinALKpkXCrm#Albac&*qe?Y&fAR-P$6?|NxJl<@!>P$gUCA~wMsgC^|L#HgS@>Rf!yo`+ypJit!VrbXAhX1JuK2gU0W+Vr%G -2q7J1%Oukb`9P@avo`Z;{n@j%sD%yQHW6$}c2$Fs53-_1y>JyUAshbK=J^OFc*+TTS8?jrOys2sNG3a -sJ$WDWBPbDSD_RWfk|6r)2+U#f1KQVd#C^-4o*6lV+-8_=;@o)x;eR-Q9ev<3^}X~C&-#9Fz`!)~x9` ->-|iKCo0S@Nv8_yv7_KiN*av_=e(?-3Y1c7zaB^v;c!mrvggxeX7VSyH5a2Ji9f7_s=0wW@hzCd~vjg -w|Bz5R8 -=9?X_c*~VhFN`E-ijRUA9ohv%y^9K1q>;jIlVpLQ?f^uK -bFEJM6BS^AQ?@$$eLKmxqrVgK2Uxr62!y|e0*8L}+BNIaM`7-hLE@N6Ea78g%9A1+8XwlYgJy3m}^7r(%GUz -?hc9;eatyvSznf4(M-wP8PkiH7wM`{%C5-v><+zT~(q;)`&Y(FgUdQ{`BCK?ZvE~XS|Nf5 -C2jqu9$Hg5Ab7}`&BjnpFs!(EJD57PB~Gb;nN -sM217UZUCxDeZKiOtn(X&V!nyw!{kq4IC?!2GGuan_V=)|%dq71WMh}1rsq -TR14>7%9oHc=iuN8Uzr@`yj%Ldt3yxs61Tj_gD?Hf|Nn-ib(=BB-EbJZ@D}~yZ@oM>MB}Xj}C`+wsjx -#$@w15*wso4crf{jtWM#^AMI{_|TEweh)rcZ~DV^qkL`p&an2ja0pYL9D;S<02m*O;LYQ#?!M0kg}BA -3;+;WaV@MoU+4w>MwEIg{HUk2^xmac=nTmRA+n*7SKK*zx4^q{+M$mOI3 -MV|_Ltxh?ype|cZ+`b-$cY(0C3&e64&ghUl0+Z7zIRcZSR$&(x)q;{Rzt-6$M&omkp@qE`P$%YtmbkY -|f6wr+$L141>vJh&gZyF^=1jw!ZNgso-%`}vP~Xp0iy3M~LcVLwg_@NHt>x0&pldmw^bqh9wzLr&LOe -f~?V6HmXvdF&QoRlpHOEH>jX{K*BGU?lJ; -_O4)R6i**MHyVmrIccIVqoASO_C=GFyBj#|1R3n}0sd2ggOL8U^DNOs7GXU4s)@BBV)I1Z~wNX|K)!# ->n -(vOk+Et4%%!ii;YGVdr;6(`1?&EfBx)k-$};GgU%TW-EvEkRv>ZGqnSxayGM0NDkHbqtvWr -SWUmSCYPnIu?>z-Zg1%6DB^XjkS$PV?lpRI -;a#YkRc$C37`#%8LP`*}HrUa~!1!Wrvwm;|!g3)tfcRtD}DR)rvd1$#ZzbaJ};0N(%18cM14#ZWe`eA -YzFW2XCVlHS7W7(+jDd~J)8cHZKY*Z;9WWNnm1jn)7-V%52y325d#pVry5&873x`ovvHV*=<-!1&MKD -W2bX2WHG`Y!cTkyp#|6?b<$qt)GwraY+8O@D%d3?Ok4%(6hvQen&Aak}VpcGQ{GgtNY1_`x3_!LMD9m -addw|GkXkuH0lb%`?>}iarI&er0{XTHh7JMGj|GcikOweK#jg*^e`eoV2mUa4Dcj%TP*e25ku -ijSa)$Jw#otVW&g{`D(?=m{VDPmsKoU=4VZR^!7c2v?chiK31x-Tm=GA6sz~uH`}Mtul0vI2Afab0@Z -#pm{fIpmeyqLE#NuK@)Xu>{Xn6>TQ~ceS+P?v-_42a`xZ(NNdoNxY!R7YN@S2Z7l-P-7DE+O?rwwehQ -Sh$5DxkhWo)XnR8XCG>mQ9$wX*i@x5Eu?bBUZ;7mq>8l8s -CYuD=l$%pb6CTw{+`7)m-Kxicrh;;jVIRiJO-|gKvk)-lV^-1(nJTn&hdg!DQTPT@asys%_8|cJm(G)Y+(O#g@x(?FauVY?n -8;4-w&@z}JpR)%_19q9zPYydF;tRS-5;{SNo$#uJPvG^*4!mxEAFsQQ;`P8=Xvc4RGC{ZSDY|WMqg&`O-JXlk?KiK&? -WaGaWjEDr4Odu|Fk1huRGtZFri2)NyQyCZpTvg>n-a!)sTFoep}L&v#}&0LxMT;)nU>c?>u^|Yn6(>p -a@s*>vl)$k{dQbaujsFF99hYpKAjTCK5aIVP0M>6*A68fEs|~UNW`Q?cUAKRO$WzXxri@lW)^9l?9{k5z^?_vYj_S7KW(nJro9gHHG+b_IkAh-lqs6`o9>n -$}l_%l7d=T(g1Q>gaL%n{aA64XQguU}1TP;ZZQEF?g^U%Bb;7w5vnC+ -wXAh9>6B6ZT$*nOgw%5Wt#^%-`lD%#4|&_v5s#XM%TV0!)JL{-WmIjtP;=-TgJysd>-lg;@xcCTAZ=h -M$Zcw11GGAQ|ojp7!}!CiZt1$g4z2jkVsQ)L4TiFPNj1B0atrnLb?%Dc{^Ywc=R!?&{})505_AJ(0PJSgC$32HQk8y-! -^Hhj(gJ}Ozy*bm?W0?H&by-cNBfol9^%e!OGkzJQir9@G)V9Z;n)M6<0p`aOOgaq -5OLf;5;_e%xpVo*SMOItj%VNRB0(cHWXblEe#i*0yPA@0H)Gz~VcFy^R-($IFFLskQbmi2b^7F$m+3sP~8#b#XowjHOydLzG=NlG3ZXIQa#JjpvA`uO7e<-q{(wW8qMb<7O5e=3+5NR$+XjK3T$#ZiwX|`O7_KFF07HP7~IMH4% -^5xQKZ?HSgdxK}=7TiC|!(AmcLz$Qk7FB!RP?53ze(h8`Uka;ZRs%b;(4sqo~p$=xJ{i;l~_5{vKau*|Gk}R$ev -DtlOP8Q@j8X-#P>9=^S`h*|z)TKRPqhOY5ng1nrEINWe!xBw^&df>pK;OeL~sNf>(tF{9)-nSc!FRAd)=uo -S1I)!vgQ)PP80&O4j|f4kI=Lqs;fwys+CgGIG*-6<)?5Ymsw5HN_i8{Npqlv8xfJ?_(GrrrA@e$6lqG -;iMR!bfun#;$BwH%!AWDo)i$r*Zp0~0@ahebTU>Ejt?y8!PQbtL+eiV*@I?!N$vIk)UeD%%bRcGDnxz -X=J-Pm*COxz`HUke>xPk@jmg~{nQrw*|o!k|+Q4Wx_#3%*o$i_p|rbBDTwz#aE%xf*Kv@H2f9`}ABQ~ -cxRFVmFzlR787qbpEGf3+av_!-X6NL4G6`r`5YNDg!&kU3N`xT1He!0UG$4}$eq8LmI6RCS(tv7{;&7 -`rt)B{?RZ(?$BLw4i0j*~{4{P6Wd;pFKMWb~d4dJpnJR?o&5XZ!zewMyP}~e_bYUD=inLhpqEL+-TCo -u-lkXJ=$`fR$z0TXGEVfy75M>;^c-b)!n%zUxSv91W1pU-GJ;(s^nDsok(s%AaD@vjCJ>G%ZK=AK=yk7(f~>*@@_w)&?V%pYULx6~Q7{*y$qI9J2Ce8xcJ|9u=9r2Yer! -XNdZOMOXWzQ3)Hw^izQD#UE{^?(l#G0o7&oy808pMYRjOjE;7OSizfH~e8-RfU+~mRCJY~tI0fZkX*M -qJxmd!D8eT!X#x$@(aVdj|RXs=ar5X{cT+LGo7Tir{I8I)E~sf$d@vwDrH1{ZqNGTZYq;CW|6vRyb?+ -$VQ_4K*++Wi~0t$2cI9$!(MGwR4p|mN^^vq0Cm<6nTi3JNp49KIVCoIO=&`%y|LlE-9n6EFb4JyShLd$S36F2-4n($HJ{P -SDaq3kmrQNzM;KGvqkm0JZE|VBVqEuBXRn#LfJ2QxzTOWka81eV03?X_v4XoeZs#e`3E8*r#I@>>E -9gyqaDs&)rZejr0SAhdouDv{VJT#j2(Qs$*!iL&ty^-C -&M)16fueSsJoHL}sv`qXa=avR^mO2GHv>GQ9u4;GH$D7u#DNmc2fL25Ftp9PQ%5iPt^~Lj^Z{6u87iRm}t30`Wx=)N$=x -A(kRS7l~!)zuMQR=zoS7v=CU(p6Er21{$HzEVItOL3`k7q(^%O?UKlS -1JVTcXoju`;jVs4S&CsLL&rg1nq=rpH3rzC6c~F3IwXEfjY1WYD&85r5^9EU5UL?}n#2|_YwIz{r)KE -tpUulYFx1<)!2X_|DtO5JXft^AZYVDbmG!cM-XT(<3kJMdOk+2KHfNPo( -1-PswZ-$PsCT59~~^D}MoW+*ffsE*zxcabPhG2j}%W=E8I1K1{*m-O^Y4N6_NJmV8K&PRRR2xaX{U&( -U78i@_tvNu>hj@G&HB_~1l_%cYg1QrMNrc%p$7VY4gIM@>2BT07vrHxerG -Pe(>h9Png>8}MPmnk+-Zfk*0coBG1JOLTGRTkSwJW3X`M!2blI_m9#5o}H+GgoCLpR6SKjgT2d;DkUO -kP6lyIFxE=2&af&r!z*D3>ud)V^C;M((!0_OtGamf4IwA)0s|;fC -tfZm){A4=DN9NiE%=%i8t#HQ`VS^Zj%PcL%N6~mc;+86QTkQ%oAp%&NBaVQt@Du3^dTgNAKLs19&8V_ -3r!PExJD6?6JFtm|AAX4n`KrYH1*&ZVpn1q|nn$!k`(q8H#wR#QOyMnvKOnQ)#M@<-Bf8 -6h4p6B)nWYP%aRAU8|r+^S_9qX<61hkXdG`S0G(S)>%>*IihiZshKqa~tjZi^1FvkDsFd4jD*~p44wRI?!=p*Z=>Q1 -u`H`!2s}=Q{PA6@P3gI_!AmNr|*t)AEUtEd1QmCSqzkpBAH8@wad -rW{TXLCdRMKxdR#Eq+dMvYgp$kRKT;D=f%rYaIcyCIHZ_~JTdglSqvBFB8IxJ{-9z1Uq%>(X6u`dt1Q -6+av=ev5C!_$;eAKhf-328mHE1x6X}piNK9w4)Qo}1#l$we1E%wwbGy*xgm3_?NBt&HSxM--I?Pzyek -30_2L>#?9K6iE9>2Ud9<#Qv~=LEA?fR2s7#%4|>9$de?S&r~kQMOGU+E&rEO3B8@${O5K`%BfbUAGYJxK(@ -NnYI(dotB-ND!X94E6mP$}5`-jEDo%kmBG!a&0?t~gE_pOJyy#J|Ji -Ac2O@K<7tA50=C@9nvifnLgUaYzkY6qHGZ#UFyoLH!)53fSLQ3W4W;oFR-Ds=0ui{geEok{|-HkRD8_ -IsVum2&AEv^B(n0N$Cr3WpNzp9$)6ko>o5d6+IA=3y8?3VXsto987fY~rYZPL%i5HujXBtX3p dUY -NB%UtR&{@(!zXGv4-LEua=wCnQ{uG=O}D^sUh5RaIn!y#t>Ixp9XE?^GYi#nsfanAqv*hywqSg8Lb=QPvtqb4rd%HRUYAio&<)pU)CTzt*p51 -yc$L+^H#+Oy_65qAzhp6@9E+`N)3TUmp3#_1Jar+QXbU=ALG|17$p7^5%+!M&x*4VT)cHHBgXY8*M3^ -YSA@55dE=yE{At5<3dPShUE*B+eC@(jR3R&Ke^?^VCTLynGPjvj}my}OcZkgwF!YBLPU->yRA6u!2@c -TM{O52X0~LGISTQ;S^OGPddG97lHda&>hff9o<$EX-xQ2Mqi?mr~h%?CZavjnS~mIJmfxz4S4FMGBl< -WO@p02fe%C{Z2>3rOQ@hw^dr|^@AQf5wtFOKKBQlfTRL1d+~CVqZu%+M~BODgJz|SSJGD~W%){(HJ+` -`S2pp=Jf%8cskSyQwV6lK4zX(`>g>toRkB{ch57^(g?PnrK2`vI+gNMU{)T2xJ||zn$+bLJnal%n)g` -9O8zAR+bQNJHYSez(JaaZnX>yKM71h#Hds!Q-mz>Rq0v*TwYCWz|gu~DG(TQ`s9_YmD_E+hepzBe(cG -I -L*?)q%xF$nZ&yOkLLN}6eoz3pw8Bv3GIW2D!`L*a|I^e8YFmif9HQ2ShzcQWZS2cT)V`WQ`6>UEEmBJgwAGA?#q67dzo5qBGiM-pP9LM)9?c3*@N-(z3#dg -Bu}iGDc|5Xa*BRKG_U=?8m9+f|yF^L8x9YO{tYPn -P*A)a*Nf--jK?^3bP~}I?PNR<~fAH6G#S(i(pm>O*hcP{X)|<^zeP5X&OC9ph(lh2BB#*J**I#vhd+K -U$J`Y$MnevClFLI;;|Ycr9r1LQ@1c!%-ujfJZYn$!vjW2d1cmKfWj`ggynSDmNF -j5{n@sowN0pyJYGJF1#IG7inKg+1^DwheJUK|2w3#QWu~kj-7(nUSi5(q*$np8>AsB^Ml`{3?p5o(JS -|9QW~Cu*$iyZLr?`nxB-h4ntH8x)bUHCtZ99dm_s`?C -jl!cr}%kX;l3Z8qXV##*X8$ZU$sN3-t%0km^_j7ETX{Yb4SuRwQc%)XmXTmaN|H_XmYi=octLZrl=T> -%-l3EY+jZttEsTJe;=c*Ql}3caPx-%4Xh`TvZa3-8i`88MO+1(;@UWD!Z|ovFok44VoGy(2#GdY1WLgUk^-jfXUzd!L5fmw0?&kc1vq0ET~M>?ndH)(AgDYe{J8M4K;hudwYPrQat -!7QJ6zXv!BrlsnB#!S~6RKXscC2$jP{<$Lo -(udMk-Ks(H6H$wDr@h%D3!D519*$DYviOMQdbZAnPMtXCBn6DG0rx_Fn;kCv6ajUR*asM;%(LhT~v^)dT<*`Aa387<~$8H1#dF{ -yF4*3|@`=`Oy86)9w=$Z3shh%nW2KDM~L-=64OhUO=-)A4j6KZrstCW5j&933yeRyJl(lwfbLXpks2q -9jrQsbYw)#LZF)3|WoT&!&0M*N|km(P`KO1FfO2wQTN#t(!;h^VNg%%~npQ4&B3i7l}Pm@GqWX_!+i- -loy;Rf*$tDb>Oi$4g1J#> -82T`El(!vt|s0;8n%dv0dCSL&V~p>5Q&n+mnvzaY|5^{6;D9m=vCK!D27@y)y9*T`WiI6h4uzIux^>c -(WdDB~kH&<3Qg5N>oB4HB<18gvQCUhBS?Wby$NKVac&yYYcN`ZmsXlvb(g8s=VLTGB#+f38)Ze3mLib -7ouYyXYgC)0I+m?>P=b_Q1cSTfQdQ$gs-;Gb;XiRbIp=F~Y2n5HHeXL+4&+ggFG{Ti#LpPjFY3csxYEB|z;M>3Ux);rL --ezY&h4?U$1hRjU46c6ZgxJOZws~;ISMq5H+dJ+|wHWhX#EWOo)31F67{R^0*#Hli~ieqJ#E!xUOiz5 -=dHjl^eF14Flq>`;umFwyT}jo(7d!rnHuS=nSo!)cSP`7`^7=I|!_i!Z*opDSEIeq -Uu1i{5?>Gj7c_ZLzyq*m(|~a$QBs@{X+o)w8M))D+s6%>9);{#q>0r#GHm -D>67eRx1}N4_{g4SvzUiT1(*bgwropBg2gko{FENBt*^yz)yW9+4WRiiJPy}>B|WQ8;x*iqgP$!&!?z -d6edE?=H7v=QWDhSZh-b^^$E~;FYe9Hf=N5b1(!y~IobS}NBu@(l+mXkEwmvvojKJO<^teoFq=&rhr3<4&9pwCNGF}M$-C(KP>(p?r{(R_SyyNrKrRt8e -C*&GxK4;b%Hn`Z(2xy4Uk1=&`+JY6(n%=b^RYvYXxP2bGkC`QG@wx?O-HMLU_hGOnud=ks@#vBz{~T=$A&}GLXwU_%VJM^Jb{-1j&S!)Io_;1;fsq+Gw|tt)f5$jvkDhFd*f -qQT9qd?IYmQJ~Ye`DcM&HZ4{b*geALA%8}wyjLTE^WuwL}pz!$diu$%mT^ZU( -$Zg94LofFv5B;e``@9`+gf4<5S_x0@ceDr;>e%bD;fj+ut~zX8hkQ0hxxZ_R0H7n)j7UD;CxBc(m|0y -hlJPcW=O#YsT(-S%Qd>I}c+()%V}w6)9xgX~IS&pa?C;y|*_tqn8aBfrQw39y+KI}_`^QtQ#9wP7A79 -fgk5+?{d@FV}Zk0YhC(ZH2nEF67BlyB|e=y0Xv~UT6<1vjEZUwd!@XO0nFGd)eQSPOE3wN2oHrjIIMq -ER(D*XY;xqM865=ar7MvFPyx802|5kuA8KD>Ow0UiSzP8o8sl;7G*7O$Shq{J&d9%{>zft*v~@fHvMR -3O4#4kLIWW;kUt6r!j!MzYelEVwx#1>Q0h-G+`v+60?_B|K|B-e=A8?zLX#i6q5uXx`!+g{qql|7R0f -IraR&p2gxYuJIr(yOgl8|4TQD7Y5<=KPs&qnI7f6kgk!o3>R39fE_Ealw)m8_*YAKEk?W3NuHD)376) -K=A1%j)-~?SotVnwrrXHE&TR{Mx!=W?a?%vQx*TX1uH%Y)4bY8$q{YFp;Zi(9gQ1JovJ2=*!DPUm}Aq;Tlh?n@3B&@!>=v&{)i*FevyG{S -12V~7kNuxd!G|Ieyby^_KqCFvyb>Z7Z4dSNT@E!mu^=d-KW2eg}2cjmi{_);~ -vb#+VQ}7rTXa0!vOwrI0Z1D0JJ@bj~y9IKXe!efcj`e&$tLOy3|Jx!`sz3ls#+*0H}}RX6S2Z{1vr&* -khD*)Oc$kWSE~xhshaIeH0s7IvvR}q2y_-6V}G<18KYGeOlwBQp1~4MBJuCd;iv(^eld;p?zIMxR!v?us_crxrVV#p!-!4hz!Z!ya*hkNpjo4d!u`I0-&}l|BisB~Pl ->YC0mKpZ7w!^UJ@$NXaJ;d3^#dTLYTgCYXRHG7~1RS#8Y1g}goszw#HHg>CGqk%kD+W~1w`+_KuoDm7 -i)oE$a=4+1ov!mR`$9Yw7bT>_$7UK`xVx7Vl5UQ>j_`z?xc00mCaW|z-Qp8i+6QUFKswWT_`V>K3;(8 -Mud$9qzc+EFTgw6{KR~KHWly#r<>^pWVjjv!8BG=MDt{J0%vDG86Co7Q;zU|W?wb@IC&9LO5h -ArR*1&6hb=~AYS|3^twevDrN0&FZ*%pxN&1^he-re#Mf#gbe=FDD&P3?#j#BIz#4if*IxqkLS6cy -T#68!Iva<@P8O%+z1 -)agItL$=2sg1|JI->yqv$)vy83JF`QiyP -_&{)+MKEc`lWWGsZgEN_5;&H7Ohy#AEjf%2;dk2K;AD7t*9IbhhEwRhzft5dSool5OfmKM_n`nUk -z*^O>m8o*FsP#;XHWagVdP2G4J`gj!{wAn#31K_;P=MOx$9a7My-FF9kJ9sBtD%FIuWv!rQ02(uh@m8 -yIpzgF>It$CGsTL;$@QHQkVFE;3kR$(}JWQo%9R-2sQ5b#n2`KtJJC|jM1K@;`v@NKBQ{c6oP4KwYZT -~s?-plni(954CHlL2iMQC9eG@%Zq>S*Y=7|>=H;#rmtQ6GO_sXBq45no&3+1XcjGn;u%(@HYLzrxEr? -FP)ora+RCp}nW$!=xmNlByF@T8Xh64;06L#A06GT&Itf6RM+Lb1Q*Y;_D_g16DAj-*xTS~8!VaKj0N+JCa@dD6dX&XJhIz+uj@F%Qe>*V -LHyTng4%_b6?cd{-$^k2s>Fg7lR^!mLu+w*ttE)+9%Fq)!@6gw*vJD3$q3Lg^jSckE68 -iOMN%d>MqPZ4O{~n)-q0l}%W*pEGLa2>WprM0@;FHf=yYr1%3yg?Sf9NyQ5uPc5w*Z0H$6iS06KzS4M -0+caw|Ftfxu8~a*IngfpTJAl>P88swiUe{w7h#Ra2OC2{n6`T_pUUKbV&ySuR`KMtd`HuI&L=(sY%ai -yRpfTb~u&xqkz4@#0Azs20vS%d-lbvz4$No5S~Wr*797WMLExiWf-Gyfg`fjws_07G=2bsMNM2R@~%= -mb5qB>;Sd(!35W2TXG>KdlW=e+6MR}8(?h1}M(j`bk`ZfR>sIQd7xUod{*^k}rzHKp#srY!Y}x=uX~m -}rq<#!K+wPl_a<CeHtViU0=$PG$9=k;>yzt-W53q{-!dT5YRK$4fJNq_1<~XjNK@>zbJPW0XBe3lBhNa -eq_=7EF*2d6rDJ|HuBAyjqsG&!z969STSTlj~G;ajmYo0$2i*_AxEiHAEgcMmz-xj4S#Vjm*MM`(tc*!AYrTCLacq*B`&KC?KEH=Hv|t`~6qPb^W(c)vQY2$b -$kGyEPO>w1k`p~luE-Dm{J*A#h3kUKA+CshUpl#t)lY9OOrYlY!ZtwvW4ao7T52<#<1hgTWD*2gw#qz -j7VO*{n;Z@Hk@Tp#Jw=^tw3lX()|GEMA8vu-jDJBo|TIjN(#bHr+uB -ZB&%8M*sVj5?0ps4VYK(o!wA^)$=kwI$1WX{3jJM>B{%8HgSqS5nVQxws1e{c)}FNwQ?RLAU8=xGh~4 -Z*qV#+&OrRcM=s6O?JR2QZaZFbb@G6-6z-!ALu=SdwrGofR10f#p>vJtqpido_qGyBJ%$YThH!88+Z< -lV69kg%})(s8%$1x8{Y#dUsD~G3keYSmrkpiSDl2+;X8xXjuD_R-Ka;rFp?7GsdS<0X@DTTIGlqcC*rWO -heyaw1zzVGw~K!n3HtC6h&~+8X~}FrWP=8j7IS_DKTkeQ$8@E$#_su@#)*Dm+;#B3k%UzOiHd%=SApC -=aDF!C;GzQ0B$K{5<{A5c4xR7_N99yC92QsSR@#59=+H=h6DqK1#A|)e}jwZ!8gH{IWXvAx|oeXO>oXssgjxCNtq -M(kc%k+KA{7)ESjfrLmf;%dRNagjU9m}B|knf6nSVuOFE~!m+pHP>q73O*$4hd_tJC2+)IH{p~$8Y7; -KS^=V+CbTH#4$(DD+4alFv^)3$CDDCiiELd1DVCg#8+@L!-l@A&p!ffKYkP&M{Z##bb+_7g?@hCC1$h8(PX>hfA%_#KW`GpDAm^I -?YjN0F6HIU0VM|nHXuV^*)}sg0|X4L7wDw!4ED#DjB~xBjU9O@d8RiTqeMM2M^{rbeC&+?3g|K$&K7> -5se#t9cy(U5+@$2FOY@5oHz`Z4&c02P{1jm7lNHGb8u1FbtpG|zW`9Vnu*v5z(#|=Iw9_niPEp@&KA% -Pbbe6B4fV-RVp -%Ubxz+()Dg3~~pCZQ~~LW|s1>8N;2NF-AjDxF&+2tufwqd_66qlx9BvNRa|DtFZTbC -4DK>?R?8K-Cd;z%-cJkW?o0yeK-s#JA5-_uV~!VE2hE?6-@giV~iby0DlIdU|y#XTS#jFbnhmEyQ^DM -vljdlZL+{P}>g+^$yH@W&ATO;44^v8RquNgI!(7~x%@bNqVs(7Q~jusfUUEQ%WgsIk-N?%M&=Q5Ois$ -?Q8vlJgF`l^%J67!O!+1uVHr@U2zQSV38oGacpjZ8#0N%1O%-J*u#ENzOC`b`$XYExanY^c)Ck)|S7Zgba0Oc -i&GM)j+N^Xfuivm*Hdt`CnT6VgE#E^mx56&`U=^lim@2x%i_)zPp#_0$S+Z)3Z|-VXNMra%aN1IR@2<^Lep>IWp1`;+hSR`0B9Wh5lztD<<7RX`QZg7j7o!{e& -AD!B`Yog{o~)$-1Z>tw_^k=ZFsGduE6jF)L}p!bgfOdmW#O!^;(v=zr~WRW^Ap{wk%0ci%)Y_b%CvfA -D)GN()0*poA0JmSCni8_=Qj*j4b*{86K<3Q#16oyRebwc-mGIR9=l1xx26fppio8D5sx_!Htlcw~lCP --;lmA_9!VcE@4l5T#ysi&7-04x}2MZJtJHV`#Wdg@*vwfTG;c3a8SY5P`i}PzJ3Gw34;;F|80S@O|n` -#;x1L_S#%vVY1+aqw_TS8_TI+Kt)o?8%WXN+7NW_Qe7j;{QAd1KZ^tu1)2-y`@SJ{_7c`)vs;+k#ib} -60bp^R0L9;QWTtu3cjEwsPo&}LBPN6RjOl!D>@L$iPp5{IU(n$!ZWZ8GeWc{v=b+kL9!jsLG2q03Hj4 -6({08fmlmH#9ulwB1`;TFndL?674jdxZ -89)Lrx|0a3r)TO&p#Dh)8jlP#4#u>YW9>;t2FfwVGiy~&2ZK%lkrw-RR8^^d!L({;3(WYjwv^;N-G<( -6u^_4}iSG%a+wtMvQ(Wwh0oobCi46?TV${g{05;{#zOPLe*W25Mit9bxMc(W`GQidi7WwAy{}pd`J;b -5gJ|W_Q+4MUGnrquqrv>(US;yXs{M`*D%*+1$Tr&1Wd@MWPnkw6@NHvxv5$ceiB7ttL5U -l6us(=1$(yj5j;n-jF}HG{?+h`k5C2ah4AE6~aLQHBq=)jSy}<3vDi++NEZI+9PJm4@2w09I%6p;;SQf`Ebo40OG2&p!ZUhvzJ%p}SH+IfNN2`N}E^)MKPB~gt -DxqzV!`Ym2wED%PE=~?3*?_iNh{8m--9*pbM9JMm;oV5dSPb2)4j#J5&8ocIaI<=Y{FM-%j~z@sTVua -VJ)2_xO>S2BjNBXhaq5#frowZ>`J|f_MMy;d7p#kOKnSlM#Mny#7-P4SPZU0)Pt?ITbe|~tG(Qs^ldx -s6Sq9oGhNEQ$+B`x_G>ByqLUs{C!tysBvD_psZ#W+i7d4z;E#^0zUoV6{;_e|zlN)N?`X33%RWnR~r7x#3|GI#)%_M&5>0-ZcO-*Fg|7b%^H>LPOUo9bzd$tP+~a@P5D0G -!O6J7n-ieyQDi>&D|h0U4i!%Lepei?B0$1a2fS&ID{k2BkVckVXfDYIkq)~d{-$>Rl_*$!o+WEd<9l_ -pzOMnm<2nAUMow4?iQ_YH`GuA86@!}G4#sbzUEXQDY66|{!7T+k2R>wF* -ckzW-ySNE&>88?jLKjemfma+yvsnn}r21;rB4eu%)(2i{%b<$BS!lx6Vg2LiqeNK0p&+ps9Xs?h!p0AaghM#QJ!i>tPo7b$jh;czs5|8lezmG}Q`hvf%~$D~?cI1>nlR|@QOHq- -dlbpV^g$ZM=B1@5PN-daX}XgXCM2DtI)^$*(XWsU%@w8}X)zr9C>U($S(2~FKIlAycRV~64@x5&&fh2 -9qbE^6`Ag^-eQ5CBNOFbzs!CNtko2n>^Cd`k;%Jb=`z|ECtEjx)Sf1`(mCQcOyQ<-W;a$}hcym}gbP0jy`ru__60iDMP@r|wvV?eNB$smqzsP-bN}4fngrtLp8Bl48-Xkpxib+F~kp#pmBt3+d!8)h?YWA{Kl%5uZN9y>w9 -e#WQYdm;=hCi+*uOXICmoK7iv82#z5L^x=jyoDKbe8$FqNUtJv^kK3KbTT$GGJ02tTw*tKf>AV-s+C_ -x8$k5=PC5$#{nwp4D8al!3D!7F`Q3m8WfNvVP41KB*Y1IR$;i=mRm5_wODI*mM -%(@P#WkfoR85t9p%52WStR^K9$K*;iZQ@@b@%)lc)8g6*IVY>SI57!PphEvjwjVu`S~jen_KzUXPZ?oi?_uarqo=Thx1T6U)P{dj%0dh6$ -GjZ3|=U;(`mQ4OI`?wU@s##IEh!uUne0?;mQ{K5?;KyWar&SqhAu`Nh8$rETg9|Hx|X85m-wNhk4?9QNz-jMm~O&ewB0;vl+8QvHhhHX_Px2byj(&Se=+}LNK^*je4&))Mhr1xhJUwIrGXn>)`TLTgEXaN<-j}l9LOYQS -BT0|GSNjI-x4&j#AUfxO_>Qo5tU~UFx+k+jeuc_z3b64cx((JkKNOmNLD4KKE$5U2=ulWfIfQK+bPFp -P%;J!+r(suS%EnR&*LNeC!k(t$qmUlto0+v*ccI4L8O5kFaw0?cLAGvnO?=`T!dvVbLWdJWC__ObIw` -yb*8Tf8+qw -8a4mZ83kfG}3ZzFtkPNzlQ|1_{or<7LW0IP=mzv(zsfAvw4;@3Y3Pypcb+38R0DG>rDM?Dx5{^mSiA{ -vSc8O(qtfuS;;^aMae)GQ?fbQ)@q}^aFt!{ -hd&;SkkDrs$>Pw^NEXD^Dv{U-ArL+=7|7zlU?7W7yP!NaJe0*J7lpED=LRWvfcxf97UXM=VmR -$JGPa_rP!`nn>dJoI`(9`|hnCjye-g?<6CN9x3T2T&p)62*=%FmGD!3$+#Q`If#fhYO_9=$40MqPAtT -)RLnz}KTMRPKi1v2E&SQbZzNTuPiEMC@KsPXpCVp-Hq9ums}NAmv|%VKa${7bPclKf%BvN+oJjj=3ln -=~wz#b3zmK5|Jci!U$eu`C`yyRGozSQel`V!3w@iDThQRTg@YeK1kC)HaZ%c5xsJODd3srImylRl9d+ -APdBwmEt1ZLVE_iSU*u0CRA^K>;v5N90k(jKFL2MbivKewoSYds-Ug$oF2R){E}{7?foPA#3E`iub%J -whTs*6VZkdbhItiD2CrC{3SL2-o(x`b$Ov3P%^w!H!avl!+MWzt@#my@l?q%z==>K2uE2_-eRX^Q -*x{;7oa&c(|Wfyy>S&`|54-8%C&Mfh(|N-Ll$vQQ!&$9TK=AjMGrekE -T^9DNltEz7LLVaiad<(&>X{)kH^s6TB8V@JWf9B&2eB-udU`C2JAU_nCzgf -YF3Pyau&^$TWkC(Pgk)!4x-^yrN+U9{@PWOe;&ex8l`5&3Dvb)zMjSm#NY_PMfQ^pE9z2UWDC!HKx>H -Du3?XDX`;F@EMFA|r3R+#^o4DVrF7Ee=i~9@wxwg8vZSeOr{Cxs{AHcViZ@pS&2j0wZL}KYCI#3)r$KkXnLYLehcUwOr+Wb6mxB=fKTbkm2wOj?0nSg?rUPJT$=3KK&?{6u7^4eK*2Comjt^+1JdY1N;8IC{FMArMF -{)2RSeFlB~TjO>)GJszOe+jhu;-3N)LU@e8pv`YyOts3KYIzgwB;&7uetHlSN+kQbT=vKy=tob{R$&c -TNB>)XCUnN|p9!U3*B$)DWHy_RH7Vf7$4YP*g~hVu?>l|r4O$obnf(}@lBrZC(T@!rm*}wXy*jS=Y#%d0J`5UM8Qx$Lr -v}Y__VK$^dGwrK;PZP})3eFwzc;~b7M`^Ue>qBeT_Y514Af7ahL48JZ>^5}au -oQLoR>o|&R=BY!dchbqL$&-3DVZ5_|Y1{#?3w95$4o%FZ&v9pJDPgJ34CQe36?|`wZHv3ELAgUs^@P -0g%>TIq1y0Mv_n%7$OpL3=e(!+<2k*GO~!|8mQWW!%~aYS6Playf6eZ*iOdDZGWm8Exrf-Ilo!PJG+U -}2@Y=?-~S2@jLCt??Y<-}qORQ#Y+DeT?HGQfBM_#ooIBL{)A7<9lF$QPIKFLem@-6~h$67mnm0i~@n8 -C}=(qWe^a7h8Z8-6m*~%M@(1Htr -!&e^%eu;QfZEtvQSFy^8k$ddH3tz9chO$3L3sT$$+(0D)8RGI!oAKm6zO -?6UF_1Mxxn%erFh3<~MTqGb}R|dwG1p5zcjs20Kbfy%cvM*!*O&iqHWm8;9K`9ZHsm=h`Pp;{Kd$1P= -cYE2|VUf+O(>lA5I4QZMuN{}vm>XZ|tG};u0=9CFTr+_(pzKONx7{r6JA1a|g2Coja-p65aY{+Mdpu`YX(KZAr8Nwb%drNVc5FMl{{;8hN$z -ij$3aE96M|+L!62@Q_&f$0%_l^c+%nsC9@q}R2EyLr-ps&?37V48eeGITt-B*M;u0A|1krG)hB|?3;z -i~@v&R!b`|M{x3)P@)y1-FKDBL=>oo!;-=e8*Q_R3@m+)o;J(EI6Bzw>$EqzrZ-{-R7y(4iaM!1ZY40|P)ii$I<6=I8>-IpqVcYVs_%U&w0dKeb!4) -n-!Cr0GoG~#Qlk4_N=roo`!1>p#BIVo3QlpSLPVw8SKejqg}N$Qk3^g^ektEjlNA}Eh}LjLw1(Sj(_B -$OQ1+gh8djZpsKJz+lA7u?*U8=5ZrkW&spNfeDMll`-I$t+UX9!9C=_}%bf}F|`_=M2qbZ6=Chlyf*w -uSwdFtz?plG%oKWVKVpB0L)QeTflsJJVKQf@TBTB^g+Aec5KZO5eVOmDqKB840G*gBR~+uGG1+=MF?+ -?G_7A@61YZuZ6qdjs#*6S&pN$FN}ri+6D)Z?~?3hPnuIZA7iH-X5@hGNtSeDm!ZrbnWigc_L#~$Adc5 -I)%#d`cIrAkc-!T{!_z=an6~8N>W|5&N;`*=Uj5mY1m`lFCUILXlv)3^YLyo52DGo`D5`Fe-5uOA4`^ -gh>=y(mwoBa*z1_}mzjg?*VLPXm??naNbWJ4wj_y{xXeM0$ugIGaIAdJ-FH9i6 -xo{-nD1>aGe5sA|3xTlDJO*As3eV2~`MWKYRj$4zcu=kz+4xod9F%&;okD1sXadJzOUB$NHA$mw;W*` -9LWu*ZyxA(*=xS%lV%K(dIrnXhvW^`t>>Qo^j?&Nr~@IZY@kGgcKx*g~msgi!Td7lnDDZCVtgWb8NXG -0VC)rjbP}mNCd8G8iAl{3zh?C`KhhQFqz~NsBhB6i(WxQn-1od!~<=5ywtJ?2_0Ih5_O -UF+&NgIY}ep?*ohrTdlzD3?Xh6!Dr>Nl-hbU6oXaYYrSqr?k%srPjXfR@81W*xJedI(LUHLAC(-B}j@ -VLGE~;J28&yBOQ7|e0p%+AieqoMzg^f%|1(37ps{k%NbWx$QhH^lU%z;R<>oAD$M(-Nu8org#juxyyK -KwA;|`u>8YUedsS|D;$3v?l}vY#B~TgQu+e?=|EZ%=Jm3^glnL9BVXQuy>b?gwv|g812}S$@2`& -E^4^gxp{c6($wRvhVJ@8a6?&{X)Gkas5%=k64n$wx|eo)BTl<;8PmPWgvY{yWe_-w;= -wtO&B}UFKu`_SpT}B?Y;C+nIxi3H$9aR@q{sz+ARLG=e0fZxk-v5N|&6o0f^urIy)m*loU~aR*Zu^*U -}_5nRxo+a4|=JuC&u;Q4>-4xZg!T4427bGf(uP?jGvm# -a#vPO4v(xsBGi^Fwog}}c3psJdn%C7@TzA0}1Udf7!tXyaBBc%EDh}FvNn%WYbp_wFEDJ01!M_UwY&&8 -VBcg%`U{3<$aGR$4hY`TClMmchK#zcagFmQKkI$yTro4Op^UH7+#?hA#7hpuIOo_2G_NJF&1KvTwiIj -z2{uf=zYu#Jm}gj&2CUGK0ofQ%*aJ@a}N#vs7G1zh)`=p`A4oeVGE61aa=B0&)wn*JxvrFT%Iq#$hSO}|Y8{j43cj6QZ96#VV%lZOIi2XSzE! -J6i(A1?M|`B*3llPo@!3M>t!RABujRj(luT1(h%`(vgbAzr$wSdhbup`&a_HFIj##+fp%Lu?*DakWt&Y3xDr< -U=9&2BW-OCLUBExdXY{t0c%RX8`?#vtm_kB|$NK0?C9KR!kCiB={V1UgUI~@8NXXtsLK=PP<07GH -UJ1o{B{a50LW3_$=$uJgVw)pmQ(dkf+DWoBW0< -4s0gF#Qj?{Gcf}Gsp|N+OaYgN&2eYV-+y=)4?WJWMU7fJP>C`y{Uf#kv(!oSNjceAb~ -f6YHxf?h`Pm3;+P=gTpSah%tz>>rsFbf+QWqsTebKkb}K6+9(1w;eCk%16A)Dk(HI!lWRhh+;L2o~t4 -0TY?wLOkbKMq?}k@uel9<)WZAJb70C|o-1KE*@Db-tMLSZ{D)>L-Rq&?PC-j -TeO53d1MnFA9QmxC0Lv;FW+UWk6j5b^O-&gHPQ75z<;daZW{vj3OvO}dklqv#9Q(Ly -XINF7apww+$!QLzIeg@t5PJu)PV6#io#qhNh;Xss8#yO5C51YEss*r<_Fafos*ey^8&@q&L$H!Nm7G_ -`$ZF2(X;5n&RS`Q*7lk?U(k{|%RaEV}i1B*vVf*(}+wYt#jVOX@ziL~%J?x7wh`lOxPIDdR*maR}>TE -BhiJ2e98yI?o$cpcdqjimF5ML(B+t@K1CVcuxGV9VYU0L&eGbnYI>(lTpHg*;)Hfw1a*n~SO-XV4b -Gt#D{BKB00cG8=Mnx2)Nd~)^bk<4N*~MDe%nHj1^w~*`5g$lzE|%q(%FsG6^mhf&g%C>Z7AYy|HVx8} -rITXYcSvjOJ!Ohlp9Dq`T+q{-3wkzjfdO35Rd7L26BqOZ7Z3|z4X$|Q*j2n12dIa8*Ztfbm)Cvd?Uxn -bk+RQiZ&ql>CFimfsr^1N13~$K5^z9IQH-6TFCj$?;O`V?N*hw7ABxn%I^`*>io8||>y3OAR){F97RB -Ti0;6YZ#gwY3)Lh_;C4H@QelhgM>!dTn43uhw86{Sb -9#F;nI`B94czm&#Fv<)rNSvzDiCIzv{0CdQo1mcr8HqqlO_vuhBQ%_v!!vuo -F|PEW~MY$nAy@mVdhD_g;^+FCrq297v?f4K$v$*JWod9H7qzuXXuT;h7T#ycfty-*V5<0x)#=h!ivu` -(!0Wny|dDr!ulYre;3wAVcjgOPr&++ux^HRov?0!)gi1xlkhTOt%P;4u%fe)<_jxix|AlY&;ltXswiV -=U%EwjV7@3t2rDMzQmC*VgSD5iLdr(xh4j!O1!d@dkX*te2-Xr|?Fy@1Sc75B6;^Bul+uOO0P8ei4S_XYSVLjGMO -X*J8X>IVu!aija9Dc@D=z$&t`XL;um%XL3D(A3SYu&5Bdl?-ewVADoQSt0!Xp{h4}^6ZtUHBu2CUnJb -vCRo23?HF3mI2mqNR`YeJyWFII^cB&QT)^ovtmF69W9u13E^#pRUj5^VHKR$PwCEUUz*-nKc*g;fjk_>f$8BmH+N{dXKyKJId_@=L>boW{Mb6*^ay>yYnm{C4CFSV~@8^D -|Jbku0AfsW(8zI%B{nU>snNCkVt#MYBpO0Vm?(vOKClc3+BHU8xRYbI#QCu-JA@S$YOC;B*c`#ScB#P -@p=r#kh76`?ea-9C;zJ_z-+C)9d53{^T1+@pbSxpx}hlkqXZeu7;DZxg&q@B+bR&=9-sgfsS@6J$zcN -@SiQ^Awr2WY&^dM`j(F4P-Wu*+^z%g%+RInQyIcgIXFKMC%B;5`?~i?^(LJ>HDhJ14D1SOuga)p7t&g -Dh@Slg>6`4#hsCjN*#nW)j`LLXP;`MeAm8-eCt1>k{#>jKH7$;&z&nL0AHWNMq&ja){L((h)1R|)pM=~Xuqdx) -ATp>%rTtS>3hJcXf60%Eg7C3RDHvnTiQqmQ=aA>=&^pJNz!1fIjski6zrA8e3i~bjTlc({#quF3^{xX=R{sxAoMb -eSyA7y1B-xB0qThV6BOd7;9AZu0oyiLEu_m=Fv}hL>y7-P<*Sco2xq42Ol*1%I5%VWVgpiGdJQB5e3B -YO-!ZH0dEJ28}Z|w&X>-19hT;Z-BpD;m%n?8tJ)#yZ8tk4e_Nk8+cna`QqPXH{)IHfeq%du+K39 -RH>@58B7q@tkyRrHoc0icI(A0Ac{h2SsgVLTsP7bs@8Z`zZ!d#T$jQwR7})58njk>igf=#aydfpGmS@ -z;$9cZZ;urDxT<}tjn!6tGWI7-eqPb%eQiZd5V_tWTq6}%+t4cRlUya&H2gvlk5lKu?dOrPw`$V-luuqv&H)i^?lPkvf+q9cCwrz+-9RG -;K!9D$<=MFvQ3Qah_JSC$*6Q#T@BDjUdC;MY04Wa_D^=NNIEtp7fV!Aay7~PFs^kwXKv8g+pS6pV$8t -Y5gUPv(Ja{pCCdv6{Gii(3UN2iP<-^Vj}D*B;mJn`df*l$Cb}o^WelF;+?wb9qdA_rvaFBMqkFNNw6{ -bZzTJ=~??NDs#w~ML^uf9OFI=BwImR%XKaG+m+v|v`@52t^yN^@bo(4(y{6hn~A0sL$xt$b2T2j(9w0 -lk+Gl~O)vLeWj>_L&!g(67tUB-JXY4jkoVUZY1ej)Ez&T{&=4|-wc5wE**6@8eVd&lh^@OkapdDL#^R4W^;I|}__bA-fnC#AW -=#dU$Hnf#kScB0yQO -Zt^i0xJEy!c+E`8=-%6`77E2U6)cSddequeI4KTuKM18P2%VfpM3|;~*5y|h9;2Yi21^6qTTYkE=I=VC`|4K_V4O29Q4OS_6MA*s@ulk&wCr4K+~>&C3zA1BICMNlo=?X{=TYYmoFb4;5x&*4ab(QeNcNok3t*I*aAlj)bypvK0@Idc;PWoBqB2Pvt@ -%(DCQvjUH$xUZpCyGiWLy4v{GNTkYwbV4z8^sB5=w1_Sg+yEE~;i-PR-zFP~mlp7gV0-e -?BxXLZT$TlL$klBEwMna#$yDv)Q``@Ts*G>S6r)!+4$*icbX7JCvJDWoZqs93Rg;LlpdVz@xWeSNMbM -GllK-Q*Ppnhr76%k`^Bsw3f8QGqa1lKEDUrNow(Q048&QBaM@Dg?F3sld@y2BKVyicVIVj}mW>Q+Oh> -s{-8OTjeYR??i4zME%-xI4$=M=*Jm4=9MP4)ifo#+g#VNS;8-ghzP0CuYD0~hGj^fi_~kdQ)jHkrXA~ -J@1x%q5wXv4iAuT(s^|nVKF+P6&sl+-m4H*ZaFgEZO${f@C<;km7LR;HqCN_qwn4caXCym?zqGDd;GP -1H(E<-(Lv!#*k)M9``HO^>4%!zxt?43%DW9ev?&{=>)sL<@P23c*53D&Fthte1OWpy`FlU9FG*;(iPK -!(;-ns}*em~+|m|%p8vN-xRCIuA@SkomY+W8nNt>l;XB;o`fyIDumgz_NMgjJV#7n&mf!&NIs39ChUh -st#*Tw1Tj7-T2HX$0!V-TJjgxK)Pw{I -2>_RICUKPT;$U3i(Wpw>TF#EP^G=l4WKLlx=WTM(6XA225Yh}c_)YRsl^k}s>I^b*IIK -l-np?Q>VMF>88v?tkaa+x}wiWN*O;x#%r -8v|wOr-_F+(h#|fn&+&9^w;8h_qD1?}_-uOJ!vadkF!TKqO2widpgvt#hTlDdJ9#pzNToG_t;Hsh4t! -7&)#Ky7Y?5DS9fwG)-8i&c8b46lGa$DW}M9@KR3k%o0z^e_zt3ati%YRXN26NLDD41Whk5@zB&on!Uu -EmW|D6xkid=%IVJ;M8nZGY48^`e3&9tS-fcQCmM#ONU?CEN$Th`VmD|9rRs?ot;2GHiHTr$>02yTCY^ -BVq}_H8R;>^=a*0Y<9n!5n7aeG#vG8&ovxV?@?XnK8&gDMT6TDXG4%BbQ`Sm|-R<<+PecSn0bN{Q>fc -V12mKqSZT)dJ7M4>TJ)__<^j3iyO2@M(-6%86c+C1Zwn0Yu-_`0gd<<*N#8Z -DD)<(CA!zEjB7n*T=RXCZp0=X9dU7rj*fjoM+buc!I}|ip5pw?nh_mO3FTik5pAnRM7mwQR3qxzR*m@ -2HdG`2qq%Ct-EFHzeBMhn;^SVb5%0YU)rbYZsv0p#QH^L;R3k8QlnB=7#@h>fvBYuuU^>2A8N8Bc~G)(^$<%nBU<%r-!FXf0S|2@hPFI -uIl8?P`~JK{2E*|xi_KLdBB!fYjVRhF7kW)ESB?1K#owwLQ6zUI)re| -5sX)iAs716XY7u9tYCE=jsYOJ%Hq;`H5NZ(-pl!8?oxRi|UibUeBDSWdqiwy2O6mPZ#fW8Btr)TR2oV -A$9WrBhlysmdeqAx*f+od?GcHq%Xo?rQHx$i?KmC5qh-ioZM$L#1Uc7>4#5;Zl2CZmD#K^BT&4|I9+t -Q4Pe41-U?A0RiD``fAmP4T#Q7A)HH6uc4snCoV@8_i%5w6gTSl$xTCe4VqZ$j2c>vGMAk6vu488K#4T -bdCCQJ0lkY+Z{|a|jiRLTv{ul5ZR)jl%76&4_xr!A6{0bFe1!XOC`7XzhR{Kwnl-*Xj_vg|5U;&5G8J -+YXpm+6Oc(Avj4e>_ZcaB5)9FCBMt>|Ivo+y}q*kV86YtUHw@t4QuyW+q%)40dpSvr)hPI_E>T%WoV#75%ADdK#r(71>VmwW78)Y}eh?UJSHY|M -WTw5@89GMi3>?!%pH$(a4Unk^qN#67IXm5uj3rR;2Z&YIMOL$&L$JRfY@h2KLmDdGBi!bx|~JjZ=VoM -@>SD`pw4T4i2Hvx3B7TF2Kkw@sFwO=~f??b+J+)I7vEZG>(9Dbi@bh_C$+obKITjMOzWwmAuFa0^Aj?b7W|lQ!}SJ -OMkj@N(wGA!gavKuAPo=imRjm+bCtrd6vFdLl7N=9|L~E@ngbIGJfXar?A3_E-ub~uIWa2&e^>pUYlTVa2)rw`ImQ(jY%X05tUP2haK -O*PZOPB|E0=}IFyq1R(r#3HPfkda@+H6V~JxnVegBoKcZ;|J56tcjA!kgjTe@5l=r6lID;c9-G7o_d> -uNzierGfq>E(C@_hf9BEJ8ma(gf~b{>yaH@NG|u-`DtW$yY0u4BqFRxKH6JZsZa@*UDrQl*#vD%LKfT -8p~vTsmpNa*e$E+)MEb#TsC|u+$*EYr$t8N>gb7D?mJ~Q8 -B`q~qQ<2nlN{Vw%pxDD6Bu=2&x<<|k)R&15NJ%>5`?hzS8uuG=!*Vp-p)bRUA>&znX%QD!?B0dVcp&A -w#{?-gv(!&dEiNCsS8#FHrd@7b%g(LrhblH5l^e-|rE=j`cFxuVTOr_->UA%fiEwpa>-s -QkuiY=#XNJ2&L}l4N{X>-%u}ssnkxuAEafhUdJ7IRduAVv#a{24GX_rB(%MJJj*Ij6bA&*a7yTkktr> -eoH{s_`(dUw~#vMSBhKr{FQW!#n+8DMMYsM4Aok<4D7YRc6nOMYvGIsqKeFFR+~<+hXj??wy5!TN4tm -y)Hum}THAw@W(;U&;UJ7=SbU*j1u(A};8|J&$$9-Bb>rEby$0giV$|524MZxPsH1a~h3b+AO(IDN=N_ -{Nc)Y$u`~im2FByWme=;A8~}dVIwFo`q^+%VS8yLRnI2jvfI9FBh+O?+5>G7#@)7|(x)`_+9Wj2oOvD -No#km#hOpQsNs~NyK>Y!3Ux~CgP~*pO^WFCAbG6bgoJZW0Et0k!b%sZZ;5os_c^B$p=Q94J_t>oY3_TM -h2R-7CJ;$`t6;P#=B+9=d|^Xq -&hB5oUb<2G^yA`sZlA|Iw9@bP2$k`Orp+QS927!Owvwtb|Um6{q1Lk)EK9COo9y3rYSmm3f8D-l#Wm< -rY4MRl&qaFU*S{#RmDUf;~xFmGDRpV5ABPwgcE-8(+X?ObdCUG_;(3MoTW_s7sX_{X48^k3s*EBn=($5^EXgKzbs5G^DY(*;$uH- -W5Vcz^0OmIpnu9ehaE5FVuE(MZ)sK+IMk~r7woPG*Rb~AAw9-8on^-#6g?>*d!cb?m#vBz;mP+0Btzv -#<%6F$W{5=?htQ}$cv!cuZcQp^(@D!LVlMb$>!qA5A%d9E>AHV?v92-s}QK#@x~4bB=-!ztr;wlCHi& -sw#vYPXMTce3CjopkYQ^?RUg?4gEL7sREv$z+m)Z9JJcm_Gof53}6>Q_Jj8CGpx4SNn>@VIgZemwdCTnmp!IqcHljO2+(zpl -{+bqPxQW%vb!{KuJ@gk;^1ISlt#Ig^2WKITPeTZqOo0v3kGo~CAKBDxRVlB4?>knu8-sR=~*4BJ(nvE -mQ3Tg=VHp?!tS}4vgTd+rq+N|X&5eQZMvA!!8{jJ`Y0DuhQ05cL~{rAX3E&_l+PrM?n%nUj!~kM{|z@ -&c1yvIfhp2^bxkZn(_zYYjxNg??CyoH@QujMbwN1IX_WS#RP)6Zl^s0A<_>UG3pD}qRbD>Qv0@BS(_F -j7pi*Ck(^IZPj<1WWeh9Ex-2N4h$Ro=Fe}YLgvMflYk!8VLj4T5PCS$OfhQVe!ITd8lV6z~f2Ac&-Xs -}tZvXDF$VgOoz0q9Z;K#L12#MzOG9r#VKZz|hoizo7bz|~gQdFNLgP_ZPaqBtnQe%kT1-WKAj!c|orU -DrCVt5~9~DAvZ?&*IyIudN++^-h==Z?B&*H+@d^n^+zRI~exf6xU8Qz_sPqi2zknvn1l|XJ^*nlcBWc -6Vd7Fa<7%vl6B@ezbB|urs;UDCc7JBrd(fXg;|s;*=}E{>bDY`b-GELXL;EB`~+n)PqU3Vvzsn;o#Kd -kk&?b-_mgnx=SI9haL7kusahE3Sr{nlrjVS86(6Z=-{8orBHdFWK98KE6JHafQ_~;;z|nYu+NqRQL`d&Tl6cpy=K6^+5zoPtO>9UY+ -q^X=h&@r?9q(*-s(T*+oe7}-;LQrwraBN9&_B@0e6fYir6>ixUC&ll)2TX{xetAER3|o1`c2P&nB8p* -@N4oRG@%i6X>s2OTiY1Z$$ygvhs$eUBBY!d16RVhrKwxZEuIG$?Ct}8N^cRzgCW2W`=b2m2F@99-c*B -y#E$TCfkdFaPQ!W-3?>*$kW4U-ppc-1U>(8J1ltKF^whB -71R(_N34(*kP7qEIM=+1TO0bUL1%gU~g9OJ3>Iih*DJ;Qwf*Ayv1XhBT1ospCjo@{HT?C&Hd_z!6aEU -*)FpXdyft6ql!Ji49B={RaCBadGI)e7sYnXvx3_$|HJc2t2$_XAMc$(lfg1rRC2pR~c;s -g~vwM6q#g2wAKmzA@#uZev%!^Bdyo7iJVO)ULElgF>}I};nz&vaQF_3`QW<9uNfBoQ3fQyEb|+(4|KJ -T#gupuZw=pU$SS3G5~|+RHwSjbw?;EYg|ICb4)H!zQy>;SxuGW7uN)8$DMaUCb0sxq=nmL3S}dA)g^_@LUQpAMUxY@kf4a%@g$HvPstb -ddKgW@&0$oQ%_s(7jRs5q#3tNF_LEcfKC=BwtZ=7V%HC_O -v*WhyZc2G};ePa|v*qR7iFjnc^!6gIg|qfp-ONrZF0SDX}qoA-N$pheD4%|}gJO;1fnjgJp-V-_zSQz -V!plb*qbBzg?6ET@MhB@NBY9J*X&l1$YeO%Q8YU|BrhA}V)M!4i*s@&a1{tSOd4o9Y;CUuZ9~p+nCTX -(ZV5MMXvzT607lW>^=9`m@Zp+B2+jn2da3rF@V+(nXp`3+W&YinknTAU@(E4ok_f*{zuwB8>^w9ATT5 -ZMPOFZfTaxe2elj%`V%L3i5?Rid``?Y>10^2#;_G1AlfawW_JaJ!Wcg5$t9*gg7dZ{zQIxMH8$R3(OR -YEj1(0F6lx^}zvy5R0TuJ76FhTeugefx##(EsVUQ@r>4!AdE2bnbLOVc%b33)(~ -`9?J12Kh-s1d%!rQGyHv5vL%a-5qr#n}yEM_qUR`aqPTe7uafd#d{n0Rdg@rISS#6py1QVT%+SOuThh -^H2lyFs*8s|aThbaZ6x6~0cS_LM?@-eF|l6p~7l*u++R&?YAnPezLnt@%$S`41U)w9qA<~fjVwDL -LkvQ)7=+}o2$_c563GT?GKE`~NOQ3Wn@yAyku5_=L|Fns&Oip*Mcc_I8DOB_3_>_>mL*5pMT;W$pe|er5a*9L9Ctgq~r@EM2i^MxZE~0d0+(8 -$xvx=M`y%45=C^v}o@4T240!+EA8|sgM;07D_UI0lit-AW?eYfCV0$yvsAA)so5=ID&Izxo -;J?SewO+6*xhX<_J1l*7wbx`p%~qHo+VDqO=wgtp>`Wkm4>->Q_!Zk4%|bwWegB#a<=aGM{0;!dI8qx -5cZxVriJcjN~y&HSDDjqv1GC~6AnE}`&NF#bIKIuouQZO!lt_o -|<(PZy(IjLp>)i!S5ea4&hCCF($KJ#vpzNbfl!ra|;hg~G)u-c>B*IGN4L0uSay^DddNDU0dI#A@G*7 -fm8WHknw`2_myI+#5~o8n~0$1@2@*73~ujOsqSZ$z=8*^Y||&c0HMeWMa3qfy^7=e%{1-lNn4VMmH}u -m{=b&E6PrMZ-TZ&B)|c@Abzi|8ZLPHvXk7(sKWm@$DX2Z=O7^X++gWcA -0thRT%r`ZIr@Ag$SK_QNwGoYFx+!|3E00XjA+JUlHqfC|-sDr35N+4JaOY+2zldUOTV$ztUFznul2B^{2VYKkxXfz5aZz@=uLx_@6HR{{_$Gb(;O(%>kOGzpH(ka -e$`j&*v)tnmFKJ{rffg@z<69yx+Y2u9{?p?&j}TOK%h1EnUP77_4N~YDa0=U3Zr|*SOYJth;CZhCknX --(T*3;K7F;e&o@|{`&Y68=u^?`KhO$+4Agj&%f|rfBXB3FTMQAt6N{&_WB!dzV-Gy+jmq}?X2GQ?(RK -%_r3T2{trGpaPZJahd=(uC!Zeq?C9smzWDO1<6nRC?TM4$eJ|Dg;6C-wA5Yhw`RQ!k&*$nJ&i`_u@!} -=^FE32Az3^~V3lsm-^#4!i|8HLyYTN&RMfpRQaR)^r_9DAhvET1y$4xxyxM7`_{T?s-dN2D1FZ-Xp?D -u-vaXXJ{-|A(DBA<40dSaKtM2os+mX?#BS+F$4mSM{&$cOv%B8zqM{9MZdTYP?2fpxL)NR#6vW)#_?7 -ua%^Sdt40?S;ZkewUMBNt9y7Ld*DqWoc|WOA#xLE!PVx*lZkzF9(wd@NEG*2PY^qoIKW -i-A`Vl$Ri(lew!_PF@Ee%<6L>nIw8ZBk!P@2tp!#?PQD>8!=Aq&+hQGK=wBoPSjEaY!sX;yG7Yu@LqV -Y>-(Xp0Szxzi%+JFj@$`IH;Fm>(WXQ-jWE4`MObU-xV=BBMUSX`3j7&q8wO}!Y7SUP1KKN!4P1M9udC -nk{F{4tB2?g^F3n&We$dDzF5NnsFi=2llGgLh0WrfsGvaWn1QFkYip>zJ+)wuc*~DmS*SJEJcMG3oM3=EE`cUaHze=I&^+c{!mN)l7VW(7JSvjbI6 -GT=2#2z7hCdeo{XC1Ww9>K${P1M<+G{ -JLdR4zyC`RIVV)&jmB+I)bBb)UZ4oyQDa;fC`Cv3-wK3F}Y-v4jY-)%3v0|$43vV!8(Lba2mHk5pUD< -!ijaT;19C~H{f?>`5_qOmK9%1s(Sa`Fk&2+|%ymCG-kGisd-RS222V3}08`Io>znA~>Mibj%ymI`{Oj -pkT<7sWD5A*36SB`(*r;!Y!UAD1y`omk$Yu&I_Am#qHGqmU9amG| -4O|qzp{V(@+;#p`02LG?ek|$?8j%?_D_Dx#HPK~+`nZxMZI(7dV6#GmGOV(y({~dynki-)_-PVTe@mI -WxXGJWNL}uyEVB525VU28P@9apM6UIzI{sFHQt~4@7B=VBS!AJ5~=4(W4*K~77VD9Gi=#u>~?zdp25( -lSDw|dEi48jW{WxjB1ws`abmdW6@#%~v}qYRwh0B+6zatDEOLDFGIlMCE-K1dn6IQTt)MA!Hdix2B%s -WMDD|&^DNv80A%|zDLRTMhO)AK==UFDu*k)43VuXFlcM3-9t>V4VH_4u7%Nf7iW|>xymXm3T$k>wjzaOgl4BHSHhACc9A7BKA$F>7q}_NEl^opEYGiJrnkqtVc=7JLUgb6|8}sRSn)R7|lr@j&%uK5& --~N7SRDy!(SV5BnkPDCBzR`3MXtn$Wv-Lu0@e5nUhhN*l}wsuLY}=S8;Om#XJ -ygIp88DU3k8z`>jVpSzgg6FlgWsQ3p1=1PdTTdY`w}WHN%>NDNPznk$o~JXUVmY;M26q1uSVaA-4f)J -B)?MwXa~8)_QmTpXYzcGY{8@E2c){pwSbS^OT<+Y=-{%j+?LW`|{~z!ZE~tK~`;J)k9mio*Uh-8; -tnFj32ChuQ2{|Z^!Mb|@lh@sKirG#v+bez@y!@_J;={oo*dKC@0&wr6xQ|fYqm}m<7`_>?6mX603r%zY@Mg!FiQpu2%dVihrq+&ISVPy!fl~zEQzvvy$I4iu*PNzYi -7vPZj^*0J$vGGW+*FL2W*=zoE_N-;UaU%hUG9OxeHfk8S_6c>j^l|8*Wvp8wZbA%w;UK3d&o63FvbFJ -Ik9aen$M+1&V)Y_59zxVX#Kw6a8)TPS?hrlXq-Z6v$ul5Cn3bL&-xk5X)(a&+!d14}=ed-UeZ!yi8KBaw^#mmZc?6jRvk9gVBof3Dj3pROFqj~OpeI3Bf@#Vkrgba%QL*F&*a$IrMnEGtdM6+=JX! -c@2Gz+D;>u-!^TzOvbmF*?%*}wE?TA07py)FOW8on+6|2+Tt^K1Led|Eud%A>|N(a1HN0Q~qI@twi%- -tOM%4?khLDo~#x3WO7{hZ=t!ij`+Lg&%Be8_;$%O4{Ch)N($leoyc`xBB`2AqD+^{Lw1EV4QH8j&ve( -8JQTQttS(6uPyR4#VdS(sH;kHM?D=O6Y*=wguhP1N-$qDkcmnfO(y1K$(okwXvB_?>{i^f=(nU5cjWt -EEADt+^kN`y_sTZh54Pccybbr-Hr&0L)!RL&4R=Eu?%}Ptqfa-Hd-8k@lgK=wY<3%M(JT)bw8Y9EI05y@&o?BhF8&@9XIuH$JHE55BBnw|V6s%H(;jNnow+$Z -P!>GGqvwIB_CNOG{%pIXTR3x62h;QbPYF?7;^gWY0YF4142^H&|t5CHwT#PuZnQmzes=Q3R$JT%Jl3+ -I#kw1Ty`erArTp*8}`?V@Y7+(xrFai2$eh{>B4*e?58cp?4jVz%M)|!M;rAcN$~v$|d2f{caeqW3-YNY?^2bAAiskg>_q=)$e*gXhyJd% -Zk^Wc6Uw)_jId>hvM^(51zs3XSzoNHu@E7>6Zd|j!asN4_f%i*OXU9{3#ym=&-u>R**LBdN`wtKaStu -Xfk^frN<+MB=!8NnGi)A-=PTYThLaaq`=^CGXc-Uw;fOv2C;K1g-FYFg#0pxxvdMt0C=wdB_@c(erwE -Yc@2TCeM_ye3@oW7%>QFtRfm2b_X4P1DO@<$Sq?2Yh3U&j&ITcj^=3sm^HTzO<~*4VqU__%M$IAVd`U -f@fhDPgjpY6CP%IPZAl+JIsLUcxj4z61dTdV<~@d>*9m`mgc%@ZrN*Qc@C2Nl9UM+;K-U9$&wHJ$vrC -=h&;SzAAY9lTSWjCr_UA@OdGYZC<2y`{K3$)Wn|(|HNO*+QEu9+Z&En(ZMcGZ6GKHn4rm -~qcXR=I_1A(gYHDg&-RU3Lx92<*v -GC_4R^R@c0a+{vw$%e_|RPb=vKTZ#qz8??esQz^`FZd>EU>r?BPxcJ?6Oz_#+2*vI^kDo(mo9#nmrZ8%@BiRAs{8|>zc<0*~A6Cftl)D(e{ZYm@yv+DZdl^6Ul~; -TN#UDxW6Dj_DiocBFucP>zDgG-Ie>=tBL-8rk4aX?{Ns3?Vi9eW9xRp}Kp%m_>6rQ3Kc2NpnQwp`6Is -0iCXLVCJJ9j&0^&2=l{}N{x4slQX>nZ+C6n`AWpGxr;Q2ZqnzntPfMDd@d_^(p@ofQ87#Xm;z-Cpr^N -(pqM_}5eXJ`{fd#UD)ZZ=v`zDE{pfe=Ws-n&R)E_(#3sH-FOQ$J;3;W_de|TYm^jg(_$MUAm=nfP^5&T6apR-Y2O~i -L{(bu>{_#ns=!EE)*cgfryoL-83%ez~3CV<{1ofG4eeiYZLx#vHg!JuW=r((*9Ha#_Dn<$@J|vCrQf4jfOHfC{A0{X@d*h@=A_=u0+9c}8%6r#lVTIhN -wG;2Zs^fX2>|~=p*JNmIsF)sg<5LqV-Y|pKau`0V-gi{k~tws5Fv6<%b)O1j~!+L3JH`(j|mgHbm>9? -JpSn;{nJfD&6L3;^1ori1c5?NkAIBbPd~a-xH%y2pmzQmP3Jhm5GW=kj3Y*B@{b-hB0Qph#*RJ#CiusUBX&(-P5vk!VlY#O&YdQw$0Ut260b4loxc3U1$ -B;0il%am_wq*ivBI0&M@7UKp27TzqO`IFSboAb)z27Sr6<1~<{{ -A0#*#jQZAW4RJGuElun%HtGuJU4afR52b|X_@F5dpxu37FM2PW)E6p*sH6}?8S#yvQc|!OhIxk?o-ZE -zmdlb%gDU{{`=V@k37O2d+af`apOj|dGls5R(R!=SHyVm%{Si^V}*ZIKEocN@y3fZR@k*`7yIzT55?H -v#5Z5DZ@&44oj7rVx!rDd>eMOr)0xv^Z1D5XKeLM$FR~5woV`Tvhc0-@*`YZ2GKxAnYM1L5P)E0%I=X -wPqkEcl=C870d^?-M_p;mhVYY!EV=wWO)PL7{#itH2Ihf*8XS;qd#UDxWO%y+Y;?JV^w^RJPDgI*=|7 -D85kK#A=bN?Bq{AZl<|LQoUhav-d^ynczfpr(FsXc=G4;?xbnx=agdh`hH(dWkAL7h6?Kz!D{Z)pGi1 -N#r{9HbxAk=}ds={Im7h3y>FduZ6uFxIp0jiEz(wIAA<0(9!w@rIs#`iBnf&_47A#lc_*y>X~c7Z^%$ -23^ysV~^nLZXDV{r)%HN&o6M0Uf)CO-{Zz!9dzyCFf^!lw*XB*-=TPVwev&z*9QA^p|=kG+P7;LM(wL -l|Lc8wk}pN+Memef;U;K*ojE%)2p05IJE&6M`!XP -|0T=bULhluUjJQ2ijMLByFkGeI)-wXVdT>m?s-}A`>_+l$Yx8A*bqfh}jPC1}myLKd5gic~ADWQUdB! -$?{p_cd`_4q3323}=&lq_)~sNER0>ZZe22stAQJ~8HHVU-E|R8r} -MU(Z@w927aSWK8;W#)EBwfs*2VPlO=_%&}2M+MhKKo4IN%bRehyD2RfBfSgZ+!dhw_j0S=Mj$RyqDVB2*P0<;{I0rX*`9z -JOKPv01YajPE-K?Djlj_g_g9aKsbAV>f8g~?ttHee|2^BTB^ss#1o~UVdu`B0)Ozw(W6I29en-u*Ww* -GqD@d67Uto@hXtT*pgv%t-q3as=l%EJ7x<$Nsr_@*!=+1?xc5Wv7s$O4{?vwV{w?^QJ9my#HzfFf^yt -w8z{ddK|G@_z@VDN2i&LF*8Z(PFjPgf)?Ay0b0C3;CcduwGM~)m3cF+JB0DeO|K^+0N*M8*Oa+LFNA9 -6l?ALm2g#qGqL52@sQ>U*3&carm;&z(C=ZKdaL!GHVq?HfWuLPpRSejUk6)Q6w}_@c~FuAm2X0o-9n- -GBMzmjYl%nWMacFZcko0B@8(>iEf1oX38`x#=U$jR!a%wV(5dy+p$<&Tk|d2E4<0-#0m5cZ~DK#>Pi~ -3;rbQJplYwQ0qd4-_uUe4!|#IUvfV22qO;qlq>22dsik0#C}n~fJwa_0GECib97g9fEN4}nY~ncAOZQXitB_v^B3rDs)Z`1kGG*Te -tR$D^OWtUl0w(I2RFf%brRl?LEXa*tO&kjbB4K9)bdG>Sh-G(1K$;K?BQIh$yhL3m92My5e+&!9nV&y -Y#O_E5Q3b3O<(Q286)p!UC4B=ccw_>;UB{7-Fd9mYdwb1EGwEuaDYL0h~9{2?c|u8QV=`;(DBM>IS|G -;FlV_!ph#e4d+VkT_2}DYs|Hq=e6#+B0ZS+cRiT+cRWR=vyM0`j+^kouPlBah3-_g9sR#2Cw#bMMk=x;G*!1HEFB!6vXB!7u$_%EVi3(@fQLj}Cn;N -2XG1bXvwLL3ijDg!*#~<=S<^M5b#tcARsDSngyn!2VS80LlhrGwU267X975X+j+gFd~Zx9VzSB~K?-f -8475DhKc^Iu3NZCD)5*De~*%X4D5V_^(0&Wz;UB{7?9-1Kz+3CU64YX -b-?!trN(2@9{DEU(g8tuP(oZze6;XXfNTFdpXviJU@6IL~vSRtlOtYW?4`k99 -uQA4mcU$Lw$PDyh7)JoMY}vw}fBt!naghq%b#kS4a`ECte(ncRs -J^`jRSxK@GCDb_t4Oijw{h}@nR$Yy6R#6?%Ei>!!eq_rL^alz1s6;g@#8I8vaatv~E!}cjm}6;1OdC> -Yoz{9{XPOzwfilpYEu0>VG#(nKC6IKR^FH;EMK$Hn4K#NmDLN_1SQPVhh`y}I%$+B4D6ikxZd>5g(*zkYoc$xc3 -S;6R?9p3a|o>M4$Xxu~d!+iW(zcI{fxuD|%=3n6zf7HW+a;ElckvI=7%j7Kpaz=QUD;AuN={DtNlo0f -RmvpU9jNNLaOsV)2({%^niww}g1(^UL#y6L8oSy@^6nVFexlB1#?o_z92K6&zFo|2NnD=I4Z{rBI`@4 -N3lfe+^Qz!x;2FF-$`(t$Y)Xn;HgAEGZr9=lwz{G(@p|2h8N<|R#Ij3x3I1ML|y>EUHbK>t_gUt`CPb -tWYxtpfiLuWz7v|A?(yx4uAan3Er0wrm-vHo@_J>#etnF~Ym=zAN|vxC2+zk;+$UeSkjj9_Be1*P$-Z -9?+&md!~BuZqLuQXwUn%J&tkUBUi&bk&%)2J@n8+VjTo|D{YyVl$79}M&7SqKR#&CAWmazfeY|J+wca -^ph8Pp0N^P#ZOGOS9-GhKE;CX8ERQkJKR>@BhHtxfA^-gS*Z8@ge?Cn0ebws^4?g%HFD)%SyJ5oyzIy -fQpKrVEHlCN4C(51Xm4cT*1KI`X1a6?CB`sKkKtBw=M<0#0g7pZ@BjJWI66y}LwDJ)7=f6Z5{nqs#M5 -ea;L}6j!SAY7`pDqEe`M<2k-^{>Y6X+1~Abc&_>W6TDEue+1_&x% -%9nBTz^B`0{;^)ZCtf#)yHV32?+`O&O7fE^#B@x|M>Ca`HUGeggfd0G^hX?)Oj!FVi;e8?|~;w@DOP8 -<}Hk=5T0bopF=`IM!x#$tM_0I@q6$GZvwzeSZ{_K@WPt2-EJ4{Hzp>An@lD?WXKRv5Aa8vmUOhnTWZ@ --P4E%q0r(2?7~ZyLRo8nEcHr;A-(lzTV>>wOj$1F9 -IK`TTWvO>?#)Rf$EzRzX={dAB8ejNws7~|cXN_iqM!DDfSU@4133d(ga`bCIS~N -3gRaLPe;jL!Px$!w{8rpO_^ac8D(_OzpvuM8=sUcVy&3-A{0|w?vX5;^3rwuB&z?P-Kk&c#FM-$bjy?dmtKUaxT>IF+fV;rod;Uv(&p -PlW`XJDuuJ=QJK-Qr@@}>i2``T--@rNINSn#vjh5?{kU7N?a3Gd(;_4@(I%x8Rkeg75QJ@|X~i{4`x@ -3GC5>ICq_6HkcvpbOza8}J2hL3V-`6?Yn+KK(oS{WAQ43$-u7|I~k%s`IZa(W0(-s^9NWH*iBesNdvJ -XOKnsmVj?ipar;79=}Jpzp-^1@aGkfc^-5C(0OePLw_BL0#WPJ>VTOAF>v5jOG;Ly9UwxN8>K=2k%~q&)ed6Z&? -I?b=?Bx3>v}f#7CECj37~6pFHv{s(zK1_V1rRTx&Sl5arHw3rUmCY`>2b!!)ygK>{$ -SHrF*SvqOPASWByA7XG!KBCb{`062> -=Ed~d{eE_`3XnmE>C|53;J!Jn?`TW@d3>D8-OXWVz+eg9v3R~{clk?$J;Ib>ISin?!c$wO3Fu0E!Z>a -Ol48ii=YEP)slgd`*+kc&*<0TMt#33u>99`}ua5I~My0l|mBD!N>C8_S&gSr7kWiC#&&;Xg_}LL-jcipb6~P@&DT&SvVqHLj6Zx}^ab|k^(loS2Znqwa`H -j@IG#8b#XW(y34Z(_inRgb0Y3tx0@ewfuDr$zzeHPG8&y@MC&4e)MCQs -TSkMYB-%}pW>vx$T)Td;u~6d<0nG@SLRcG(d^h(ZFuw`5&@1c<|sN(*Jz|d;oC)Vmr_Tx(Bp4d-NZXx1 -UcHI05q0Rk8f~yf~Qv&NMcLlMRdY0Y&g5wcCx317BY(|3Q2PIfTz2kF&&Mcc9aiFrz(hE*BUJu)2}dh -oLd7F|PbX|DpY{W5;IBm@y+A*bw}2^dpF?gL#AK2WSC(gx^qr-!2<`K_)K6k8!+qN&OEWd&rO>MWaTI -Dg+%SPMj$89`r73AM`(X0rUVbqV0Ih>BQl^k>iK=kxW$M$AZN3SlQ3D51WEK2XeZgG2$zd#q+uJ>R4D -XHW-VyCZ$GB9oZW>`SHFB{8)B@AEyrOismnA|Ed1#g7(wX(?9ZkU*uhZP3Pz5N5+g9BjP*sDbNLhuEh -9&HfN6RkGwa#Z{*J-62)AAfFI`!dLXj$#gfQpM-D_Hm+^l*&(qqqYc~kbF?8tA;;gK!ZL%y!`u6QB^1 -u@&Oc1$?Aoc;xaR&W^xs9Mlb(@TWI_bLI#9Ke>|Fv%2x(ms@hy!8kfy)3}B0nqg_~Va@=e}I7x%$`yc -08C{N!4TC^UuHFPxMFIE1@JYp%jfX|&^_S8V2(&|CR_C#(ep{FYbM3#$ED~i8&@h#f*e2 -UfNuwSBbN&vL@ph;2IM0iN5B|BFW^4(1+;+PhR#Gk-+1GViPF_!6hi*TjT@Iad-m+~Am+w6qYu4%_l~ -q|*)r0xV@J`qAnzC8L%0rn6X*-fOI*2fC6?A~bQNuj_R%Kv&xjEt`k+4Kizj_WUJ|_WtR9adpbOYw926;t{7if*=hECqIXU~FP*tTdN?U1e!`u|e>2 -z=lbD^^6NPoFMw#n6q&IR%*FuWVbik9Gq43tqzW20l1sFTf?y2k;m$4&)8 -K=)ystE#HNf5(Xyqorv69@<4b4`Z(hiY8(?k$chy|Wwt?(X^fr07w7{gq$v-fqH -W`$_h&$>n|{e!Oqy%$cvvnKS1r(orYo&!7L?f&~jM(mP9+EIvo^l^>Z@ -Wr0~>~MCYjzBe?RcF&6_tP&-y*paftZuEaVWeG5B>E#UN?_jLw7d$-ka@SX8w(xF6Sor{Ti|{!AcK&} -BXT9gUgbpYA5NCbaD$|5$r9q5 -I-V*MXZ2WMW$G33iS(ED`Il=9sG}WUm2P#e4H~BgH1gftpl+*VlTvgi2D$CS`>G}hk;LpenW4eAcldz -xa!4zqK^3T+4s?$4)O?y*+4&x6Kqg02le%5?-%|S{5|-3@HOGuhOq>{RQJVqZmtr3e|h$KpBoRm8uT -N(AoLb6Im{mjY$?i$pxKa3O!F(L(ZTMUm59rhQzTh)#c|8B)`e6KnIZ3Dsb`tHw9wAOZ-O$Tq`-jEf4?T^3q -aM%#SP`BHHXHdQ+>1CHwtLH#Ef?L3GPKUp#D;W8HDibLGe>Yuo!D_BzONzGkrEObiq)R50$UBNY%Ko5 -`fJxoSg#l>2VnKJni5tw#$Q-_t(GLqQSF44*J{VEOc3j?T_3wLL9Dt~N2*&B?q0f-CiS5IvCC(glq~k -$>_O*_(gXAvWkao#KStNVt05G4aQ0)*TZ{fxPq^laUZ+WAQfX|b#|P+)9U_zIuUx5^uDMrSBTE(?jE< -0$B@(9(l4-0V-AC6HN+nVzT|bC=6Wv`NB~9!iUlQAIGNsxch6z#^DJ(vrJR$v_>I&LU5<5_4i$0Zz-R -O(y&a0{OezB8eHho)mey>XuRjLz*Y^9Q7Igk4^s+?X9*NbHAM0E2d_|Y-V>|3xEv1TmK6_F9*>uI# -j~N1=dOx|pIyEmv%Ot^%=~Q3oB{&R8c(@y#=~wVkSnPNiKkoHvRn_8|;c!l2R>7czfrSM{VYg}j(jq& -YotIOX8E##epIurUF3v4$on2gLXNC*g3{aXT6lNCX=jMdV(sN3~`Nc(UQ>BgE)cZ~Sgalqz+CMDz=ef -$-wCH_wXE-Oje`$W%pxEE^O=(WQ{?tlNuhi1~0r>?vy>r6ne~kb6ezEB!btENcKu$qI0seQJW`>iC1{ -C+rDQ%k2KR+=WTd25Ab2AIVIZZv@_A)nLXuR#^Tjy;r>ML)1{-Nn_-Zr48mztQGn3R;Z?w1B@7W*QKidD{Bs+ypx$~y;rQ`69Jd>C5=lNRxA@A -#sb{Dv7T+KV|mHQ+77k#*e!|54sly8$;$)=p4)KF#BQ14eqs$MC`Ux{;1xdwbc@}N3 ->pAk@kYNP1~m()M|#*P$PYz)roCo@3NEZEB2H<#Xe}Cws$%IbdEZ=^2Yog-iklWJ5x`3^KaY&uYo_*p -XAT+xA{<@cq!{TxslvjepsF*ZSpz?Y8`EEXh&$Lo@!(m -`9^50r{8?Vfr|@~aim&Ex@wfR-zK?&-&+zNqH@wqcL%*&6oWI)tyMNdhN=5 -9Ge7oFS?j&c+CGvmEqvR>_4EYQBw~C@9DgBfQ%3)=_x`inEwc1hZuRW!W3!Ml_`UZW0vD(;a9A~mKm~ -ZCC`Hk)oPZFDHG$3QuRq3u|DmhBNQb@J*S4I+T_9+LHPn0Lsb?V<#Ka?686Pg^F9a@F=K9Nu6)A=iWIbYAW@*PB>Z}@Lq%?-PQ+_~;6?sE4n_d~a -)SMJU778B{%)e~C;mCmx2IS)mw2 -VV41R4aHSA@Zx2+@6hhi9@aV&Klj&WY4?V@gnEaTg|>!vhCU0O34I%?qdR&By}Ld^U!s4l-({p2U5y^bXk$5z)q&CIFmJUUurjP?tyiq&)<3Mn)(Pt_)`rxtjq!O|`QdzKYbxnJ!@l6j}LDD;G$r#CiQ8gr; -`Bh3kB8>_&YVQsY}d%NA&8SN~0ZsdO;9`EQw;IW5plKg@4d$m33v6Rp}(n(8=D&sX{ow457WNbBlFzT -CsG?UB$<_xpK3bV0nIa|lixu!qZU*K=YcvMQ#5IU_;8ix`>%|gvXtwJ(6@m45X=SD|kgYlhF&rC5NH8 -V-Wm77z{_f6=rKeO#jvg_Lq5e0hM{p=w&C!XrUAE#%}<-7R@{48(m=DB6=2)BaxXPLX#{o1YNHS})x6 -1-Mk7q7b~&5G_HQ(qnnYNs -wOJF^oaM7|YysOwe6HG0+N~Vbv78Q0n)3|F_;0;Oy;I&MA8U-(BEGZb?($diN5mJk_1p9o`XhR#9@d} -HC+n~1uj}vX2ldl>O{0-#X^`S<|X(HLx05%<4h1HPWiE##xot3)ailG;5Z1%sOSAwZ5^=Sy` --*_+uh_fxXP8u?_4~_B%VlZe};PTiLQ5vMrn2&)VbcO8W)UYMRr>DRKHcBb*9nyz`thha~gwf`eL;RQ -BQh_yCg2r%7|2=HGFNXgJwjNK{*7h6vxB7ScEq(4M`45szKIZrG`}k#kx -nJRr_h0g7`AbOtu#k0za0dTVo}tWFmXMvOqu#7OsAj3B)l}__c5|qv-ioMk#(3A9ZoOw6A>L17AF-Ql -$6ihLAY$u|@4W34^09mcZ{)Ui(_E@=XLP-rzH*jwMm4l7ZG@Jq?>6_59sHMh*!;viMtZ8)8e+X@t+Gy -9w~ms2s?1c5D{I -uZ$x5f`|D(^+m(kem*Q<;*#_PsrW4CduiAB00A9+fFQlgY814-LfDC?C?$`)^jx7R!19q~T*&U)v(T7 -Con7QeCI%x~p~1O-!sR_o;#h+w0Jn1sF8|G<|T))4f|lPly!WaEw!&o);=N-DKbsZ=RDl%2{cr6JkgR -JBAMqP|SFcawTZm9z$0bFG7xr&Vb4v_;w$@-L2RbxGDbg!1rU2{k0Kb*x-hZcbi7D<-oLvlwUXi25B^ -3QJ|3SqAIIda_IR>lUBJ{rzO63>lel`*cH#pba^Y$>Z^ud#J(J=?^#keu!yYkG(sVMp2L>=Zj -omh~LTZY{g6-N0^W-(ufxH~tBywj=Ge*xINOYAbb3cUDvtYY3` -&r8BRB+r_;;HbIP1U?h*H>`?)K5wY<7s1Csh%XiOWEy$r9Lm**7_FPC`(y& ->LkZ=_e@RgZ74m22f%xmK=~Yvq@e{|8V@0|XQR000O8%K%whoCtaWPCfwu0LlUY9{>OVaA|NaUv_0~W -N&gWWNCABY-wUIbT%|CVRCIQWq4)my$yU+SD8P4?wy(B&P)P%nY{4UnFMGj1JnXZNNQ_tLSGW7?UJO` -t-DJCRVRSj6j0JyGXY)ZztcU&uXP1u{Jl~jp8uJ@uWnfDrE=oy@ -^?2gc26jY9iREe8vWgAme1_+v^nfqRMai;cO`}RFHMk$zoC1Xu_XGB#jea{5=6zHiPg2}ne@x#=Lxs0 -@*iY&1KsSES#B+^yKzl5BuF!{o03q>3Ypbjf=51lFOdu{UX)m`JazhOH+_7|CR`W(L=p&5-OUrzxEQO -u_P%@SH`O!toks;s_6=O0lP087@$XuZjFrsCfA^x`6kMz2_;kJ3-YXLFI-;&@8m|8|DV=)leH-svi+f -HJ(1zWG>xM+>KEB~INJumyS}_IJ1C!EKOihXZH~y}Z*o$d03#7?v#L3b|${DNeHLZSm?17&L$4o4se_ -xF2;(U>AF@C3Sm4a0YYiI03FA<#RHg?$bUEh(lOCEpggoywidk@hC!S&qCFfKeViTk~&S1p99R -Q^(iDr2b|mj~ukI+Jh47&SSpFY8N|$I)?J0dTUr(vC9=rt#j>5UBi@O7sEZHdsdv!)KXkajqu|CDtzp -W$--*sy4K8U^IX121(VJkG_l&U6!k5rL~S#!brP#}xsKg755GLs7s;#hMe1Cgx7~wZ1(VO@N29fCE1N=M%@9=Xmykjm=& -U>p%*1YE|nW`M3>j{m%MCZKIyGDt=Ci6u4_JY27OhS3;ZFHSS*V8*k^G>&ouElemi`_=odvGOsJqCJB -2d#nsVWKh78nir2bSL`jd5OkEYoa;PUZ?eApf#1VRx6X>KL(oT`A|pvT)XjZH@$Ne=c#yypX*+{yO-Y -O7U#L}&c)w>7cAv3rCIc#ba?Cm334w({!P#WGxWg%y+{h?F*P!1Vv)0u^>ZdEblxO~-Zq&+$MG{qqdN -yonV~_G1J7BZ6Q=A?4Rq~J=-Nl1Ykv=2`xtcXAE9dux<)WAUexkTS0UrRBWJbecQ@XGGH*uNHvxy`z@ --BETOI^%fy>E{Y*x3dNnwSr0lsQx%l2rx^4LOk$Lij)Gyeg3+Pz7=P3mg=bUwTKs(gtRKB2O}6Q@4E! -|5{pdiSPdNxkk|v~%aIbhL%InjGyhi -*vF3H);zNE0~Ctx=SPb&8bv!MAnk5J#o9BN5LG&;_sVcu1tEM^l5Cj83ONqpdVZZPt}f -bUk~12YLes5&2RYBSfMyZf7*Sop~4_fhfPHyY!?ALDosUCs{kwi|8s -nYD}i&m^;mkw5h9bh6f_uf49{DW{LXS*=Aru6v{-1E#bPJSPXGCtERwZ*Pr7cao0A>*p#P;B5TB$myk -zbP%lNnm>{Dt4%u=C#i?aP)?bgHnIn}()*L3yDLSyU7Ef7b{Rhteyd@te~`p#CH9H$_~E@-8`hL4R%< -JhoG*K_GiJ}oN>81ach#(1M_Nk3)p37u_CamzOi}aprK))>mC74`mRu8_%O%lPiD%ESbCO1*G -lk3k!-E3k==%Ow -zm7R${*ilsmwVp+E~jJ~k!G{n6{)?yE5J{ctRn--oio`y-`xb;^hI^ljUS~O0RC%JdVi-NT^;B#A>B> -tvu^yx(D -W|tHtMtgSC4dm{*{iOZ`$w=E53aEyM3p0yCShi*nY>jzje0eXTkG!7CAm$4!)XW3jWDSG1WdE6P|<3K -1=@KL9_$#17T$&Cdl1!%^!Rf{@@WOYae@Hf7f}t8aZ3@i>?^lpL4i6tn=KrGy$GDbWDTwC%~})-m8Fh -qS<#u_W6zs0_GwYYgd>r!^vP}pYqL&!9UP70dI;0Gg%}a@3%ES_ALC^^C-l0N4bO -WXQ>`(3Cg^XsYXPs_=~PTmtNnSf?nRgfL?8vq1V@)m#5d*1J4DIPnUqR44h58jI(IlqbAqk0f`m8y3G -|h0o(`hK7jI6=Br+u0=V!SBq<|s9Pc;deIH||l@F^C7rry8>}BK+409jhU%Nzq!sK#(%gagA%1Ik9=b -%+kjmnvAsE&L!~D{n*$l_a3&)<<#+~6wvg;)?Zl^c#Q0nJd -q;?6x&ta=l!ua7pi8#(Tczw)qGf(MJKQF>b)12(Mez%liShn}gA?({B;GyG@u0fMEt!ER_*sDayp{qk -lLPRla!}?6QD1`3>%1p3Fp$$Z^J$8+>^xqomhk&DscR$tFouIxeQFa!}+jGDX0A0@v?ChNRPbkZQcP&e@0u@@ie -~0|kF;;W|_gA654WK2#r+Vh0epFss0{BY-pJ0#I1@Ld7y2#nBnHF|98L+4fDvRXgNt6TMt#D%!JKQX1 -k5Hct`PCQjjmnGI!CTm&u?G$sY|Pa1wBO>Zj!tJ0;)Me^{?@|Ui&3}T;49aYbTvYD;a}w?0h7;HEi+* -k$VVXE={2*$LgXVJbJ;A<3Jh*z0g}h7^T0RYfVz1Iw`cKOV`4>gpNX{fTeAYk!LMx4MU`3iZ&2P)h8l -sMhKywz%R_rxM0qGLy-P;kx{J{sqCug?(;uUpPMJ0KAwNs+>_|ilVYh^CI~CUGBY)VzRyi%mhxXGgvD -0jOMWhaNpnDm1suJ>T%J5a^Nxo`pQ$_}-k5!FSrLi#KM)rjCc--zzcaG|1@U(YwT7)RA4{<7$%lnR!) -VCy*$R$$03>=D8FchJelMwfjOI!|c`CT!P^ -0obTNa+wh}=xmJv{{(grG_O7Y7$+d3Br}B9&+$y{E|o<#_osONJ)ZX^v8SjkqU}D}G4kDvns0PIsHS$Hg8?X}%#Jgs^yCBlZqqtv!`+a -gQ%KYd%gG_^8u0~)JkLYFO;(ni;gESwRltwQzho{L#n!6^Y(Q(MeeKX1$FnzP58Su`Tc6N}yDrR>0tm -zva&Q$i4Tj%+Wu*G^^;_&f9&na!CAuaJNKD~~cH(~nBN$GVw6SX^J8_w29cqa>r)pvHoCtbkQ@TLdVHvcq7t2G_|D16Bi@KN2p3tFv^K^J5&6Yp;U-9L*oo7RYSv=ek#1N<^gz9arG -tD_3+oHLz09Ftj4sbVtoKWB7*#^1h*rAo0l!|%-m^F$+SaTb?N!shTiE! -lxE@(eu=TlOFe4)CiKbUrBKXTr~npC#yA%pyF$B;;-AlWN#$$+@q^JhiT!PoQh(47zq!YnOjeb)1`n$Z`z{7+ya5#T7Ie7L~7Ez?!B}1QNVXDW_$5@b;w@Qs9gZ@eQS$I3+ -t0uo!v_m=L?SfBcMLd0xL6;Wuw#38pq8)j4xy;)W!PBkPEJNQA -!)o;|M67oT~jm-!gLOQoajpS**q^R4>z@0h_=wEM=0jE+t1Ewr|u^3(dUc9Haa~#f|nSteFF@JRcGQ( -*;p&d80FMnSiAs;mdc?T_dL11Gqm5o!>31O}&tLD(g)v+H2gC?2G#0?FVfT@4Sjn_D+_*A(?-A~Io&s=%js3z -7VU%2#qos>55dm)k=MDMHMW2@R2F@sXCWt7aew-cNT7g%4e+aceL+%eUHj4LUo%MS0=D?Xvd?bIk0mqCNndHY${CPNn^Jx4fX@~E6lJt_?1 -iBXTooT|G;_(@9aQUzy{uBD}WJ)+vhag`ZV0gW;zg~@UrG}-?y5_MVj6cy{(^Dk5YNaS9<~9On|);^) -9fo!-@_5>HP&;Jczq0)7_;5kEaEG`kuza{dp5&fGv2xF2fvXmTek5aGh5SH?J)8=8w$3Z#dT=uM4EZx_(}Y>Qf#+pZov6fz=O<{1D*xA>Uri&m8X7Q!1*PSr!X#0T3nt -%v~?6ml3gCr?Q;42alN2=%ms{5_|!sYt`0a&;{6UHKweMdOFRppV#lB}(RfH`o`!Qr0AwM -*`88m8C}^>i=l-79Z%KuS8619V>71SXx`lA -zv#k54Ucze6`gF+xUSb4R`692{L)~T~4RHron3h{|$K5NAw!-p#GYz$k&q3cwD#kUC}=N9`*^k)lPbw -!N+05_zV1*2>I4^>_ac<<0|7mxkPQE==L3QSSFEg;j6w;5@}x7(*AZX@k;RTS19ArDD}ThDzBpg#27lgD8t}~*X{fAJluz#!q$|L+0N-th=WPWn=)uosWf3qf!sS=KzI43*K#T8pzB}v -P*rO+?A%_oBLTMre?YRWmn^Jy;f-q8vwvZ$J>;*hln^sW_Vwk687dbe%_5}{u~@C0S^Eo+#$BK`DU!9 --R*&{O755h3k%a}yt`PKawUjbKK4(>`8g4Ex7ZPS=nV4o6G0F(qn<4n2wifFP!^l%9Szk~j`wQ0R)g5KP%`dT!D$@22Km68VlRHGTwpz$e9~_rY8h2j -u*s(?5TEN&+?Mgh>byv{9&#yvXuIIMbEtvxX7pHjNa(y7azwfR{bmw-n?+AQe_%&AQFN>V#F&*kBH9r -ko7ctVA%R&~(z3{6H?_XqlV&h%>xh3{YZ7NBkT67VWzjX~0-2)@z8; -ZenEUWd$e-)zklVBJSa167IKpulN;jmq8nJv>lRkjHw`}1sn737lX`DVDQ5v&3d=}^9Z&11yFJ}jcl& -}5<=!5WbC(r-NV#1}nN!m>zpuk1DJLmbs!H$u9r+esT&bV*+x!}EM?IP-Hd{AxkrTY08lUu9P3e>4mn*LCgH0>eC(jlqrRMWWVb8h=+XeXEC5~Bq0L+^rLM1&Iy=O8%g-jxDV`e}ce!9Pg@! -(%21jU2_bSn+JpwywlVdk?pxc<1hu(dBq4XNn(QHJG -roH+y0gwQx{TVWPodpc5LXmTW4UA-_LFURx;#R%Rkt-Y@R`+U6UmRkJKcYMQ|o^gV;Z?3@?WW38s|8R -_QA!=$_)8X4!u8)G$s>`;bn$i%U2_>W~-4rr13^YJM|#aQN2cxm+IV}#L}tWTS13kh`DRrC#SKbAK-f -?Xxb^Ktlyf1RwNA$n$;to~&Nb+J$`06`& ->R-enc{`M4iWSHoTD!e;J)ybY(|IaNGUpTl|!OK*lxbz&@9I}ALUrIhu9kYTDvALXS!AHe!oMZKTwza -4e(qi)2Hp%pBA2(XAJWZ(0b7qr%Z4pcvaMKmBdow9ZPNt4xcVl-MSe3OG+@6J^t6X&OkxdCKmD|ob`no^G(U6iasN`Ksp%zdXL3$_A{ynOs -j7*Nb-cNQsC&(SAt*-3Q~(`_0srdwP*xf6x1MlGszE26S8~osRcz1Kut^zQ7z77VB$B6dz;)ULElQ@CU@W -(_Tfg#L6dlGX=c(_`u+IoMzIE0{426myZjaYpHgLvF<_}(o{Cw!QsG1POoa19*)N^(Rh|oW{$4ABYYG -#naXpZJi;}r#S+LtY=(GYRq*1mzPjq1U0imM7Uiruv~GHMDPm_G|HS>uq|ZMeA9sER?blMo<&Pkq>63 -Yk97dc-cxF&+rt22*{7#b?bB4aQ4?UnhP5SXm(6Pmk*MWTG$C4aV+S_>^hv#+8pN=2*K8tv<(U3o1l) -d9Nz@Yp|Ggv!~s}a0Pw6Dz>jJH7&mqUl=IP(5C;C%98^e5hIHB9lL(=ziun#)Z2272B7rDT)8pth|e9 -~p7MLMb_5AG@V3HA~&$Z(2V(&FVW6%$v5$3tui>_rcGvU2PVAG^MAsp7UyHh$+Q` -rer==52Ps`bc}x=>@9r=_}0s6@J(*xrni0)bRIBG4`q=rromUNat(ZzqU#Ja=mo9x@8Em!apcj!&#ds%h>jkjA?WP|9qyE@tzB$%*c7E)EpN)^_Us!7b56zr+Huw+IUA`mpaE)&KK>t~7j@u2nLQh7ikBajnJ48M -|;=bTaD<69$Kh9GCzIeu}T9szO@1@{(7WnNTe&4@&^dn>CZQJu_?#i2?M(m&&y^ElXhw!cn@5)Tpk@6 -(XcJ8+X)GU?vBbfOdN*0jW*6P8zqR)bXM}i00Au9U<;>SAk=ud{=Cr05jrNDO>M%snt?$(pI+NsTwMf -*DqcKNFoc9{B~Dj*Y;5@fE3@%S13I``cwEqGVJ8uwEll!XPz&sV^MLx+Y+R=~f2?-pp5az^Oee)w-gD -5DbfXYkLpSbIk_Sq(gYHo7yi{wC0Qc`!ETK=7e*Hp{e*nbJLgdkS<|4gR`WO1}b}U~>Z==wGKKjl8k# -p$@_yao~|R;1~Gkt6QGLcfxOdR%^#TY1#;_1*wZPABE51pghCd{2k;EP`}wQ<;m(TS)t`&4qpeKKdh~ -nc>1ESpO0wQTDMr2k!Ze;1b#EccVwko^fzyW{878^BKoLXk8F~a)@Io;;-fyy11#_n$yc2VTKxC`jX` -9G^!DsV4{LvRt8>&weePd2k9Jl6L&-jQ_LiUI&o24N68tJ=mni9TGPQLz_pJaaj9r4mTqfTLXRcYVAO=W*2e*zYjJbmuy~s?k* -4X9Xjxo1ljq~3~o~!CFlZth(hWYr@o|S8*8LKFPd*cI!SX;_mrG;91pvOL+4qX4XI^J?%)-1)AEZO-0@Xz=`_zy@RZj$~$=t@jT0uNVZ7ke*>962c7Jm&JZW& -jO+%F{KxqiPVqGEBRND|lo#`3={smqJpQdb_N^xqu{U&t)+_qu?0zfgqof|{$eNL9Cp=Strww@8LBCX -dt8CgZ+K`t@ZGW{JIP1^U-?a$5`cJ@te5OAGI{gIqiTqiTw;IbtSvz2mwIcA$fpW7^X4n0jMsrZ^3;8 -*_s?fG>H?iVM$Q<>zPFK>ZuQ=Q`A%VzPhMU_W7ieU}(t-J!=6XKS9D;P)A0AMf!yxUZ(DVjRYq${HJpu -TS}BN|Nu0_xSWMga7vY)vP_2`o%9KN62QsMq_3dZly8L9xt^GTG~2%dm7*LbidcqdBqqItv}%Njxt*q -_0yhSMm7cMXrARJz#+e+l*bFAy{SW6wA(FU1TWsw7Rpw499`!g9lF0^v?_1rt_7fj*9ux&*%`=xZ4+X -(9@xg0^I2fy8Y$3@xTt$9IzBEv(XSHg1PC96@cF!Y)Mh8!VB00=HnH7Vh_>&V0v_W&LUC87hBIQb7yI -76HBz_EJ31d}=)M8>A+$lR{VbqDwrJiJ;qubOXi+b|KMJ}WdpkLjs2w8sy6&=G@|&Eo_vMq{$B&PXuQ -4&no?7tds{B|S@t-2zJQ6>~d-h_qo93Q$U5pmeT%In-M4BPrS?KZZ=_$&8hvG!I6_4$k -zQyz-J>ax`c^#RQRf19(#oSY*Cq79!Mk0p+~pElsOmQ9Yk$77MjDK!M2#)`ry<89EhQKRiaF^`+$dmn -%2@Vq&J-lAOV4XkD;+F4CwK1*>Ade)Xn=0@JnPI_^3dBob(u*D^1+GxDt1Y}_N8rDSdc62Ul8eA#nW1 -R4ac_Tw})h33t#<5(T4}6}{SaKw>zZDy|IuHK84gOy+nLMK=bEsmWSXYqwU*!?<`?ww$`lDYp<%Uimo -;!{)SO`T-~V1%!iDLL7suckI>(Y3}{6Z(8&YU>E59WYaI -?E=hofVlxM6@}(R+E!W@_R-kP3tgzY59K80zs%QibwdVROpQdf{Sd4sF`vhWxS0vwdW^DgIXJItaDFD -o;Jk4$F&tW#WsENg-p132DGz2Gep}XEA{jN=YCwf%cGL+LfGuq5bht{&vVVWyU{&o!uLjMCpACV(8r)@ZD4o26raP(l(Xmok`RqD;ih-8mU6<@=b)IpSke{E359ycT!xOF;AMVolpeEo$`uoNQ&_DUoyg2gL^P*`IFYb45Azu7 -q-j(s9|INRE7fIm7pmWQg#``BHuDev&=^akQ>SU|PPTm6A`!(N-Y~Bn$-cu1-I@)}c{GMhx?aX$h^UD -1~P;jw%Z{1 -h4WsvF@vZ;GiwDvoE~GVU2ZG6C4hY;sS}Cq4^(E!L?pj5dj@~taVZVe(@3Nr&ktMM3k}eqckTf#ljAI -XS2+dd5mm+^{^==q?CwxC9YT0`>AhOv9MK22@6-ikD`7)k8)sU)tBU1gC?8Dy`QDq=djG?*I46MsZTm -9)oz``(ih;JGLN@SrB+-EP>yvT>0UHvyJTHy{W})6;eDkAyqy=@yJG^rE+f7bN8J{sHN|>qdjAzi@t$z?;vR8J6X9>Q3cUZD;z`NM=h_nP+P8pJ!~Jq}PSxm~iuBVs-l?HG;G-LGf2Y@g`y -E=J>H_6II~u+;eHV?fx`CGs_}_+Wr$%po)-;a)LXF-jar7Pn-od|u-h^YSbu#WoyqW1Tc!RDSZ(KVyy -6boo-EY(A?oU8>vRhr3pnD>}3;vepm@L-vE(6}=Ly(St7H*vL+e_&yYX&jyU2%6>W^$W(tyqn4ErVVlHte9P!3fE5o*9d&sE5`R -jtma$Svi7pMuJwNjzL8mKBtDj$YWxoEwTITSzOCVWUdI`>isr_PwddKPqEyySb0M4)%5CyJT>V39t+OWVtl2gJlscp3p6ilaC*R6pavU4ESO1g1Y-Jn#KaW8DV -4AvNi1T$NsTnvSmdW3nlG$I>MK~FANMq7(=mj*rCbevf8>wq6ct1@H(w~% -s8yivvx>g?PWK3PvcTqvxycgAmRNX_)hD-R~{r^iPA2V2q!hNQldGAG=`-{XnbNgDhH!0$saPqAF_;| -q_u|!SyOZ)-W`=nqh;`imdlZ$Tjj{eGp1m)%^W_>gg-Q}ydv@_tzTaY|7E%6znCWaFJg=;%X8>j(T)Q -Dz&BBk^Rx~-ZvA6?eZM?KeYYg2?^pkt`ffu#|NYW+?w4Y9{^%9f`C&tyKbD}*ZkpfuAoYLsmJ)uqNxh -|zvBF`K;u$hoJ!Mh~pQoeD?k%Brv_{5mhF^+CQNP>K!9KVHid+) -o6w*vr<#-9(-Hbu<6hsC@q0&-0P2sbx{8ZyW+w7 -bs1`8$YkqBySu|9*;-fTvB3Ha7TM6U`q>S|Ec_pc^Sn#c@JUn3%5`|QN>?puSi-_Iwm|Jj^Ytt|Q-e= -BT7MjR4F6^-xL&}eHrtDQibC^_QWZPNOM#?Jk{8#zQ7$iDTpQx#Mbh@LaA-2q7l{_uqwF=0q5Xa=J)l7TuNP%Zl9NJ>&L4xEd -?10G9K9IbS@2iN$xKa7X8Z+mB4mV(%g8BBM%Ei-Rp+0H{AwM? -A$9BliZ4C`vep1G5O$wKt+aNm|0Jjseqe5mnAv=Ctt864YKbT-^iuHAA??ekPCnv8nAt%joa%I -Y||SdsFV$;Ecag>F}jvfKf@UJu!lEZgO6$vsoZ$f>daf67RSK}MEZv~`znO -_q^EXALr9uqD!&$+kpFATvMEY>8stU+QcDZDdwU>&_WnpAF^%6lmsqb8b7HxF@I|%hf3iGd{7G#S>BI-)=DhRyl+cqpv{Ui@0Q8>v0fNw>dfY1uVtp1 -~kLUwv@Q6MDKL6K~c>ZsE98B_42i@^PR(TmylzT`k_kP5-H1_7q6a5cYT6TQ7R7cVes>2AuN#b(FW4D -#Rt=r<2(lc})pwFb?aT%H%^jw}MTz(Och_;YLhC%ynGh?=>kdK=QANxw^SFhEhEMYzBC)9)1E$jZrkb -<~20WNh0T+%Ori;ja8dQ&z-;`0e<4WU0JciAxcO$%9=*2Oc(7wmzftW@IjN+^xjp1aH{i8MDE$~!+Q% -Bv-OZ&1T-#A#(gsqs`2tM%bmHXm^zjkPY8cprAzDyeZdV3zHX8fhJnHHG%H>n(NSnbx4VsSh%RMQA3GHPSkmKBW6K?-TY_AH{ -W8TN6B9*QX}db-TtdnvY_=A*KiVx(+FAt4lJ>x49Jm=F6Ao-%B%M{QJ`X^6!71e`rUH{OdCC?<*Sr>f -`w5H}LNs1OKu#{@oeJKP8#f?!FBF2FLk_c1`D>HO9XWY5cpvz`yG?{w*8lpV}1TAK*;MKYTOtFFnS;` -GouV&gx3t57T&92_7o*4^86XAFq&yFKIkHGKq)4<*AAFy%Z1s9R8iZ0{)%Ljq&e^3;)Z%|0({(&mZG4 -2HIkqMOwA`55PH-)o#y(@Ba6pXRR`>as2|_h1f|0n%l(J_&`3+;hyr*z8_XRd-1&?Ke=F3^S5UvX>nL8{2IBuv@e -AfZ2;V%<+K)@HPM`WrORC;D+{Hpl_Ym-AEUjj93BZc^8Oh3I`KXp^?4h7ydLh;Nm+LC>8Q^4 -RIq@6ljWhlRf4k+eii3Os)^@!Tjd%W&lxW-NtAaAwpN#gUx4-m+bxYQzoNS%N&dOdKudNKn4X@Ge8ke>uJLr^AkAOs#h>n__zR89A -nReoIX6o|iAJcZD6ovugm6W67%P3TgSSJNeLoGz_5=#tl{OD5JtYY{hqHpB0BR{NnxwD-gw -KCegMyhxKL8ef7QuA+2S$!hyBWHv}TWD+{$Zt`sD)O4sbUWe=~K;Ku4>yVA>Q1fMUXxoLqvh6Yno%{D -I)VWVyi0Rx8%a!Te-4`b6+|#d39-}7t-loa@wm7+`=PUialOX+c?Rw!7u#Dx$>)kd@Z{qV9^)9}Ali} -SYz0>xW_P#b*@4T03!#*a*^lpg7^e)+;ckR-p^)BUo)Vo%bwpXIohPmzkckA8Tae7xIWgGM^Yh0$^d^ -@^x6RtE*B&JW3;p(2#TO#&Eolkqg^;7?ALv<`poUlgduf-FxPk3h -h&GC$_hoC)FxlMu$(zz1ru8-M}71C(2H>sTal;kUtKY56JN7w{Ux_{=n>q*}k3_mL(zufuHLg#4hvvZ -Eb`vW#14UJ<}ZnmUL52|f-Yox%YbJ3lAZ&J#)gpbwr=<{2>*NXYA*3FhxG)B~b^lnQ<1h(ajoAypXeW -#-7y)CQXs|UQ(nQRPY_m~YZ&43A8(j((tJz%a~V+~leZ^I@fG}m0&nvXP}%vbqbuXA=5{y6INiR-y-i -}cThzLK887LlIPyb0K@qI&SJ?rJr12zokBt2u&Jw6;;tv)z#Pn@GFvYMMjVTiV>i8ufIvj(r{KXO(7h -`qOvse64@-QN%m6*ODyyCqc7i&85$w@cu? -dXUtAqFLPP>fXtpGJA66$K)NfHk5gy=FOd&hOg@%hMn1|(K6F}LMn2O1R`T&>*7Ux}N8GrS(Z(J=OM8 ->65pK8mp0;!r)I>h4ZCSmA_|RgYx6TI{G`@%EP$32UPS);{HQOgkrL=b2$nS;VcOv-SmSZJ;iIobQ8I -uvJmr+h8jkhJ}Z=n5*H^Y{W%T!!hM!UTU^?Drjr}B85PkY>=+;fP}Db`(!dPQ$#rG$fvaA0agw)(RX -Z_6Wlzl*=~-`)G&_<31Vwac)ZqYvc62Ex{pe34v%4z&%YiE_2h;|o6|qy3|{0)ekcL%t-#%HyS2c|_GWBE9a_)^`eeTE#^N -H$KeG)qO?^xFGH8!zU+$hD?SZ#{UcirUOK~MU_k00-2XA@b+CX$C)o(m4{CS=hS4u}|T!Yb_@HssGD~ -v;~7=N^5@%X=aM&RFdh4_C1_@4r9*;CYk=ELoiS>rGHIWgcB;M(E8_|Y~#47rQVXIa0cSxUdpo7C2k3 -BRXP%39xgFn){rT+cgIo&{Q-e?cDad9`$iv0{IEOy){YPo7_+<86ZH(h!_VE3M&TfhsGD@Mq9cLfc?x;v7(~B)&IDvGscY1LZeODBn -7<{P_Oq?+noT94+psE-!Ukgbkzle!n!c+Rd}!eX -&jwuhd=14y}7!gt`;W2j;3xmz&2|?_%xFxs&G?lAeB!=J09 -t?>SK1N;Htb`z-C-`^EQH$M+DPyr0#0Tw+)UWZ-vEB7RRrqnBpIXhiEA2SFcwjv(!)5ntCU<#&kzf8& -~*<&0!4-q}HGG++Hgbm -t4QurXFCtF;O?g+E)hXRG!MnEtV75uJBQ{SUl8@Ev#`G|CCeioJoaX}a?DqwiNe9($+x=+nk}SpJsvN -W9Ng#AFUFHPN~tbUf24x^~41x -)=lRNzU>rqX7|9ibD6|iw0+aLtrlz1f~;{r&EZL~Ho4ERHkn`zB>Sq5I}@*^Cf)7OaLjvGoP9>mlh?2 -%-tQ{$KCpsTwBJsqNt_oM0j(|=X!Wp0t4!2~_J;keMyqAXN^6-ZXQWr7*SoM|wAR{njk;y3HP}@qu?D --;8Jh!5bYmLb8c`ODquUj&=?2~A0cLZ8HQhR$2v;cq-R=Us&uVm=Y7bV?oxt?Bs0W#~9!pkqfI9lv3qW7}ou_*rKE?*5=C0l=)#Yd8(742@MAw!&P?3Nr -Wy%}uc-g*=EADcWbNOjivq3@+fW5(~M#`r=!p>q$`RU@7nTkS7snw?QKrue(Gv)+vScswfNbT9np&Lq -a?xYGN5!_l4O6K1ijR(gIQOYCt#d&2WNXtbv>Yw}sg%i4U^gtBOFJDpDjeBU;!&1uob7vlE*roA$~`7 -EsSPRH|>X<|QdZSSq|^ONG^_&o6X4!+NqwofQyXXfRq@6h_)XAJbY2XaUI9XiM_h5jR56U|c?27OmB) -?eAsu*E%}ZK1gXd1iLFB^vF4Zt-+`QiXmIoNB;P&6CfP0**Z@ot?$hCR$J9!Sxh)MrW$n@H~R+u*vRe -hTM=0ZC|8r88KNs4YWrB&1+bwZt-8FbNke$lfbDGv1ow?gZ5!40E|>TQ=TCXCzZqTb2wsO1{y;mI9|X -Z{G9XH7ALdO85{DE?k5Eioj<$~c2E^-;p^MS(Lxyt*FR?#f_^DxP7wGI<_zB -<(>7K5Kq=Te?q=yC2LE6{-A;yZY*Y@$CIOo>X*uJZ@pO_E6#PQ*&_SX|wCj4N)`LJ78&4Bzb --S4F5`G;d&6#(#on+}*$5rK6%s2Rs-gJTR6uR{5PbR*@Bxma?H}C80?&&xuN~91a&rvj_KoOv^59hcY -@Vug!6R?ciNjL3-}Mqtefg~(qxT0YuDPpFUimM=BF`4Wy|I*& -hx^4i+ztAv_}^|gSFvn>gXGE&V?-EM3J2=~(BztY4VEix_QZ3K -Chxb{^ZV$|(YXCc&(T?=2DtB?6!v+87?kLLnli+3tt*Edd-yMc@=HAUQePK) -%Dsr{9OA&I&=2M? -o^z$|sZ9G4lU*%?PkDT;@U01-&%mw@Judo{PueH$Yh*l+ir`b9W;ELAABrCHQy(INHGXcx>onIuJ1dp -qx6pib%Yu38mP(7X$_W{%f_!*!e_PA;nGLw7cV684BxNPq)gs7YVFmnvii^=6W4p=QjF$4S7q)_QgZ8 -V@Wq9RD(v7_tdK-EyD#qEsgCk}Ze-5*;pXm8<8%g{1P`gUDahm~lp$3~|>@L9G$YSx!AlhJ}1>d!&~w1W&1h{D+nclh>cH-+Rrm(xMmozeWwv^B$i@QJrVU%R)11rgYQ)j7p@jPI9 -PkrKE(XD$gR@yX!FSxiYI-k|->T)e8uCA*<)?j;eX^C)c@yZ2`0H@nyV0IY@jSkd>Qd(jMl?G!3Oc?E -+MGoUI0`xbJ$_->jYXjA%b+X8Q-2&5e$S=OgnNe0EMvYMiuIL_W|K0qU5nFpYj#)6)0`1{{CClvOUI% -U(ssQUofOBwKJ)WJNtdl`Yqh<3)~=JtKb*L4|*dUu9X=$HHWA|0szL!I`Lo=;^u+F*wU -K3u+&0-xo#TI%@@d?R`tp2$_fyo3 -zO_S!?oMWh=?o=RVx>f1qNh!=wVs>~AL4wpXTUTo^kv##Jpu0uL5J_#VmAGL+M_~`Z=~5v^!O&p7-LC -umKft6I^S18JI=uS4O)ErD&+43-v!An#o{|r$IbX<%B%-|;Zr`2yMy -DN1iYJUF+1TIm*)eMbcU}hfg>y#YpeY6G>?-5&3fAo=|6xN+(Q6oLL?#pIps4v`U5qpWz+E#ZtJA6U#> -x_1;RK~|>JOwiJ*(4TmSP*CDv&I(4d)0mxSWqtWxnX61QNY_^`^X+D3-L&mjpcqeL!7DD2f3yFj}KYJ -d7VDQ1KYKG+WWRjyQjU*tlIrKQ?@71g1V&nstfU>c_e+fzLX~XRWGia@S}L664zS%NKgE@wx@~xfe$4 -a%1bhomt-g}DOMis>D-m>tL{!0`}FtW`ck@>qtFGI-D#rFrxNdK@uPkP|M>Z$;=CU4*Tdi=?3T-3{%i -(~$09BSed5zmTTW>{2fBVwJ1;cR*g$}>-ml{KHT=Gg-y`_#z^@IzZkhG=;79U;c5*MpfqHuwD1aZ6w({f84pN#(nLL{sh6a(Vq5-ueu8G8 -YJsVW#OR?MY6ph%F*|IO+4>ul=nB9Lhy^|ocSYeT+bQW_0PfvABEkcGc|H(Z9;Z$3GX|Ed`D&0cp;~A -=ACBVHaQfN7$3I+MjM@9_13Wx*vwKoujo}>Mf;rr0;~YS;T^k=Rp@bcXq{tQjG_A0mdruxBGKi{fu=df8 -U34T_&1O#fqFh)@aVH;(JN2lvyFww+{Iz4du-<(2vTTiiXBADBU4Cn@7Z`lfLi8x4ZD|80g~G-V;5hn -hQ1Y{Lc4w(Ake`WIk@Uqe9!$Sg{`HD8qX?+os_2@~)xjt`N<`D8RETm9^$U-mXp+wwT(0J5t5j8pgOU -c8-T0pO5>Vp3Qbv*p(#q{T>GWPbjf=rTxD}Y+gh21f7V(p$jdDN4C1e9_6nfuJGe|nC9>V`JOR*mxfh -3ucVvi?o@&gZTSA`Hiqz03Q|9c9y=3`)c(BvmGWS|7PmjAwMP#4Yx|W6d#xXtfqGH=PJ1P8otE87Fi- -IK!NJ?F8OVz^gYPLYG+o$-O;0g?h7Ro$eG<==e=mh;O_aX(CE0_Ii9IqsXQ`h9K4>(B$2cMS{$2|2*L -RR_YNfLW$?l+i^Hcr4s*VK^x -7)&cD4wshuAxCzfMnZ-H2XxF+Y#Oh&`(|0=vbHxxKO1L%T>WD9n`8(u`2K79K2v`O*=lp -e&efx{U{W-{?*rmY;2!;a7~VE~i$#22V|*?HmA4~B$owNImpUiLQ2vfs`NZF!Pf_;WiOPNyW$%cUy-_ -Q>F+tf=_)j6SFDD@rBvaI7OUg0u9a4_)b2h=>N(I{mzW~nOM=UWg)L*%gd1Noa|^F()y9kp)AB_h$WOziN=5UzCBduL -XF=`glzNvun(jd&LWKSBNFwWH65e=v#IDrec_G}t$Xj%=(OI_+Cls8PwYoa;eCJ})cawJt<^8_bC@{o ->hSK5cC9bYB4|f&hd?{!MwV&2k<}#Xmm_$~kax94$8wF1{r1Up{9`1#GoFsGBVBt^PU~w4=t%u7Mmiq -13px(iV|4t3{m-N0HOD6RLvlM3P$QYNpDx)D$WAiv@An;<%4TeXE@r`IY=+GcvXd41nqBn4!uRN2!)- ->EM~Cw&XkfG*8=Xem;RxmBi@F5?!;3mAupNwv^_}$nal3v#Q=L`qR_s|;h{T09+9Gn_vTM -@PAbLbbUrG{#Ww(ZGhoAK?e1F>i=!xpqJAw(N3h4oM-uLTt?$(>&L4V48%N8r3%?_o@xA-QrJUyR=Tu -D|XE=}8r7QpD(xD_)M0+z)EF;c!@AOmu1o;Svm%>4r74uR&=l0lm-Ku!})Zam|lA -m+}aTC=|voXA_qu7byv+^Xq_KnVEru7B<-JYbtLo(z>EAJ5KL2#tOl)!deUG*t}hogLaYOj=+D&9${f -v?cJq*xxBFW!}JDfQb_zSophVx{r6;VKi^48{N}m&W$!EE;;hcH^f8VtV-8_&J%!zakf`F$sK>a&hK8 -)6T4uJBO%EfAvbCJUL=c@dY8g`b7b44m6eEt(3CtFJ -AxUc2wB1?yI;1kzCXW?%S&ou=rQ6JhjYxr?f!HJD5aQt!jukd3BZe@Wc(%~GwCa{qD(sesPFl8p{8@H -~`^DCD)w{w7eUQmQ(b9 -OYX3)1vQHpgljR?N>nG@7Mb0bUio)IQ~7cxPLw3e$oeud94fkJ3NU0AJWQck;U2P1lOyr@^vl(SGPM~6p!r)6m ->#c_`;l9B^-?(8%=|FvnuXtLFJu~bm8h&fkV^u`&8<$W95x4{Cw~kOXz -%B(#arTyf8m`SLJK0G?mX)72imvsZXGC2Fsx5B*Qm7&uZlP4BuDw5kCI9p5OcVm{&7w$ac27+Krz}^3 -OEVh0DO6m<({fhnbM6{cPg<=S3f)alD1UueSsyJh#N2C&$C4P*0|?$In<<`##cP+V8pL8t9uNgiNRDJ -WhW;wIjk;IJDKq&;6j7ocfpdAL(dcx}opO1^>zCX&ji;pFsBE?&+*u=~^CMfwTpDZL9OXDrx)7<|MY5c)n2krmMu>v66F%(q-avifMlF3!q8m8EbgJWc9@RFPz?ADi6o%qs`Ybi@t_FLs?B)S&F7Jw)C -Vx<=AS!I%f4B()!xrn|LPi7xpYs`x-w_rOYb)=~-<29LI2-RUhZS;`1QGi!?Y*1~|7*31{*;+Y7LNLo -%IpsYVdfb-x1orI^%6kgsb>c_C;~BdxGScP%3OgJ-K6xm~;eFs{d>)R8?FR@)_I_h*hRZ-X43v8ui!? -v}O=tCaaKHnZ=@oznJ>eRAf=T<}#!3?#;UvqLZC2|u3NzF#P;ozAx1oI!mkh~XvCKhSST^BmGRl3DlH -re(IK@cG}4&{wBrwNgCxrfFKJOPV(BhY?;w@WWtRyVoObI`t(s+AJx&FP`!ap^e{;df8xWSHR9P-nOI -8i(cqhiMsVmEJEq%JEf~anv>MOaqXcFJHGeju>jg){(XU#meLU{D_~IC1xQz^rLkK?8b8uhN(zS+T-4 -X0r?FdEMC-Sh^8E8l&GV*l8ya(`b4%B2XEW2eoxcM=Z$r5a8b2Gr&#wW8I~V;*@H2@WE=y%k)#AA_iP -c6B%U%HeM?u?Nuxsu4>@dwqrExYtw<|M7;49Z!k@s7$C*HN?k)_D1x0}>v(%J$lbAcqvqBT@Dq^JJz4 -l{$-L$_(pZUZZ^^C=z}|8{NV9E2G{+i#6ZuJ%nvI*yZCs2mbsD|-zMTr6f5+`FwcFIT)45B -Eiv<-$;AWhsGG1=796Ww-V!8b6&TC?R8kIeAk;)eRwH$VfEx>)}GiKi2#LW}nb$ms{W85eHbyV!nGjY -C*NdM4#6TcblH9`hU{`-izk(ws)#{nBdo)yeaa1j_~IjWu}^Lydh>GwXsrnG@%*ikO4zDoyI7l)q>UHp -aLy!vW311%RMwuEq7t6n@i7zWB+Y-x*ZlTG^Hk|vV!iGA>ENV!s-+X>sScO8vsj6D@;uej3vz7yx@$1 -;Jk{~@*YkLPW;d1f!90Er`)Q@1JY30?;yU)Bm+FwG2(r}6(w4a-R@{sC3fo&b>&RY^P8P$zZ9@KRl=H@Pw3 -o>&P-yBc{S@rrDcG()=*RY}SdHzs4O?Ddxy3&~yVEW8mc9;~x&dj3hMhlVjYG0?$%PF6OuDDOLmJ!a{ -Go_l7a~s~@=*FAd?Va81D|~rtg#;U`IJcs`75+Dg64~PX8r?g75TvaFUaD(Hvi23Ptqzr?szfw-v39o -s!CsrMg6prIG5?GD6iXtDbmwoR++{RxsEO#7VD3-@e2oSs`n6!Ycp -I>o&zZ{h&>5SR+FbY{6Dz#Npg*5uY#C%NgZM-6cS=g}mBHW3$NH}@Z3gXA^AYIQ+otT5Q_X*h&j;U(j -~gW%N1Q_csPBq=2s-a`v!;I|5^Mjp%*S2A-v%u|BQeJ}$R;Mxzq{a*e+&6(%zCLwZF1WgpKC!dGv}*Y -$mjKg2ee1v2LVe(d(j15gS~Ei7Bc7kAzKxswP})dU5i*t#`960WB1RxkML-5Aoe)G8V^px^(1RtL+b? -~hxbz3e!R1~S+=e}o`P>%XEZ*Zifd?wNV9dus!vk$cDw;>q+%dm#F1`vr|B{ttWa9v@Y8u8*%hlgnI^K<*)^L*SsvAbNlV5_ -`!Gs13x{5~9`id`pmHog{kBiJ&DaCO|I$cz -yaQUgJm=O4g{MWp)7Ol1?#JVtb1lXh(LCmyW9*+C_xFdx=iD74e9o=HIhP%A&i!bdb8c+=jF#>a_0wqocOBO18pXeC)l` -lL+xwQq)Y&`mdq4KO?ro7d>|4-|CslZ-)OdEs(xZLg5Sy_!yY@+cqBv=fZ`&Yqrn%EK;s0|;`sBC+k> -wsY$~6fu%f!l;^w)^;!e|##qc@swKNMNcXje3a%SE66=Uk(FJKa0eh2QYOa5p-X-xTfC>A&bo -6FJe&bXm@QZCqgji38$D(!dg*EeMkU%P2P-WO|$-btDI#>w>Sy7!wy1=A)=o$`Q1%q2W#{#E)DZXERu -`1WE%KRrg@3DPDx%k)unXBfUdd|vifWsgdIV7F2JHxX^z@pq{&+(@0XFb9Iv^qCE1}UzHy -g?N1+yXqRnXqGZ7hw!KRCeB7ET?>J~{B=1br-PVRkpQ;sT;{;l$zUF%&ojpA6X-=7uSjF!2USzf@#hzpUKpN#|#@cSAmj&@-YURcF -An+1wPlh=)dW3{Lo@)r5?Y-k~Me=bEgjc7Mwv{s9Rbi>U(gGbS4OIX_1g|J8LYW!%aJYK?-wh7eJ@Ggh;TD0RPoN52WdommCH;lP)UN(DEC2?)3W~=2L -#=wG}lBESox7{h_4N;)%_9Oir|9M`y=92dzgm1YuT&LQSQ6cen9KG3pcfFHF);lR;y>Z5T+%P -{>_0}7>-o4-_Ag4#JcLUe$no;W=?uYB$Y^-;Ac)c4+F0kJ7qt?4RUadD{JehOjydu{-FpcpwA%CJ@+F -5s@-@s*}fqk(~Ki1EUIt%+14)kB&Q!)d;Yu+(&B54cPKL3O$N#k6`!HGNrC#a8PpGi?Tkrc*>h7}Hl6 -Vpyd;zZJga6;DcSd{J2lm^$53*pAc>8I)2HUu1?UGiCLobQCPy^ua}%Z2i71Ma+31l-90?quES6kOUH<<5;sQ++)!=`*^9^ggqU`bm@{ZP;RKeK7Q>D#Pw)CrkR%6?vlkYCizK^%7_``*Ml&IqEejx>oWnub -fYu*LlCT#Fff>^`^?gS$G-K -YF=v)s2Ce!^vvU(!n$cvpM&|shQD^YdA{D!UdGtWX2ak!-yuqR<-oUQeT -Wc9#7Uc(V#+0JYI`Giao`g7tF)svO<>TBe$GPdzvgJ8hh;uSm$-HvBJAtFj?-vbISB4tT!ykqB_)iTu -Gj||=XBhSeu5J~3KEv=|x$h%I+gO(OZKs8@^>!HdEUeS|k=mb|BHF+DFD9+}-N&5c(fV5zr%KL2Q+6) -Y$I7++h2GFfp6E(Fh%>S~gK;;+w<2nKWp!%+#|qKlisua>}? -h=Y0o6!MbmfmA<8ba4RP7Tlv!a(l@&ECT(e1yL4z_rU --VTJlBW&Y8U>FU~kOu_&BkeV~?PATiS&{hj^ -*?4c>wQZ%o@gk9SSC$C6>XG5e`6NPS3t!`(xvCOv_jfJO+2w66sn4O{~KkmO1r**m44Da?Ji9lTnCzAyHDFvfVN*jKU-6-k#X(k$k_M5{k -%kp`}Aq&9fOnyEMgbGW#C)jw2a?DyG!4;NjTH+%r;oQ8Mw+iT%(@zp>FE(9yu32<6_0ch=kuUtoKxl= -$1N{aO#f)Oje)#aUaVt>K4jDj(OB -pIWD-+>bq#+ISVtIq=3^7Mp*^c?I`40|zI{`{uE>$C3`9AI86NppPlo8?_ks!n}BIK5bz&*SK*k@!_S -z==-oH@6F|SIp@qZ@!o8-g)uGe0<5bmv$t^m`%o6+Tf86Rn-dBZnRpQ^-sHIBDvKW1_ -Ft*IK-jz|%3-2>FsB$Ud}Rcp3Yu)uIP$m2cmp2+f!>&~lkH+ZNPJ(D0C+3@qu1a$z^`_%w~Y0nT{vM5F -U0pl>+d>b^7|7#-b!dPC$QotHHHRtm)jCvmzuOZr=g)p?tN|yh%vSev1zhI*9HbK_DUFmCmb2MW<5q8 -Mb@oh!_wBaB;L9=oq=^FE#OgH?UBk6_gZ=0R}H)wn9UiQVkxY+1R##D*&M>hM2wr#o4dC1r&koTnS#G -*M&9|gDUH_rk2#_C=$1-t^P)^CGlnM1*IjYF*;=0zKokIpNY -rfXwiYWjG~yBis|B{9N=e0)D3+23BHzdi|93CAYR^ZbZkYD9k-Mt|vU`j#Y^{q-e|vY{W(Z$_=3Q}vh -U9O)0ZSNXe7)8DoW^*8$bRWNmq=)o0ionH3agMJ@EzvOd@M)f-qQ$Wh<^mR!z``(*q_!6jZX8#`-7mW -813f|7q_}iz(qC>u2XN={B7-La)1_~cXjc3`&c-$skitUH<3HR5sq|xJBlQgownk-}Y*Nx}@_xD%2xx -dQ(H}+QyeAuhzcxq$~k`z8LCYynWg)#7uc|9>+hMn2Mzwg-fQkN#;(9h{F1nsY{7+Q$4m-1-W1QE>Sg%sdVS7Nce*b5LFT;Y_#Pcrdb8*l>mO -s!9^7>#)K*JBJ;TnoAWn1_m@%R;ZsrWgj~8i_X($ty#^>J2T$y&uO>;dCS*N6*x=_ZZ(<0(a$7{@Q#r -Woo4av9)GJX$r0_T}$L}W8dGLtxYZxO>$n|;3@YgWr;!?Kx*A7dWcSAa4sQt9j? -fZfM?i5?=PRP{s``gO9OtjsJweium9C&Sym}|KdD>ma>9d${u;wTw>ob<0Bi3w}BNSoYn4BOFhWkoExGAh5^CL`tx=R3imMT}p{*p4Ip%y -i0JOVWNCWA1ewQ&u@Ub02>S|ia^^RsDL6PCk@IT2U -vv?Mqxe4APdFws<6v@cZ$Mkb8PFBRhYre^T%>WS!tcryl$n{6N9~_Uie7bnq-?|G9x~dOb}#pdE?bW9 -+rBKku8FAMqYb$-P(Q-&D2AN6!Dj5K0_>^GWy<$7KUo}n0{e^iT#Q}66SQ{;&%Oxw9cEw9nC%i7X9-)?iQ<<$8v2bN4Y?ALLQ`_3FPRlpxxyEb1Fv`ufSoFEv}NcnMHO+S77T -_fWnFg|dK`)%t4+Mg7}$>N0m<2^|+`P{GY1T^N!qAqRG2U*Fvp+nL||l$s&qY -ST35)JJ1!DNwcM0m?ihlZ3*-h%bh(Nzc1sR0{(v^K@?p|yU+yNyjq)>!|`s46LWZ%{{z8xCeOaom`ev -}4|7zz7Hjp4-{xw;e~#~ani`j1lRm7jgPPPsyZW^HKH893vhU;>xx$h)o9*-`cI5JGZA?hJQq^h>EJ95c4t`OC8?S>9zg(X9NBaBN?uXAEYd$N|?+W -%pwsM8MKP>A|&t+}0d>0IyTEA&+^W4(Vte_OmSw~4~EGvPa$=YVT{pu4mSiMH=h{8qB}7AePP_#R0Uo -2Q%)6}99FFKxo4oy5#=AB=yIHyo-ew9EW?W5yuc(^9&auOYtgyuMFQ8 -M#@Z>ZMQTXvYTbrq!{{#4BzKbraBlxnTfOaqh0_oYObx3{nF=Vy~W;I9x+baP^jqCa#8Z^h~0|0kHWD -)dyRWb`oLcbTHQBI_ygv`gN1k=QJ-pP$D#%THYIA9NXK^N9 -)m@VL#x0U@K{wb=Tpgs}4&%*r8 -h%WsACGOK?k@+-UjY6!E)qfVB$Ga-|LA$Wb#}y$KS(`Lp2ipudhjPkJcIa8+1}x^VK_|zoSHBn+egYb -$@Wp$HwkkqlJhB3en~p!^L^B1pMkTZ`y9UaX+_8B^94K$xP-j>2@v=24*!Gnt?iON`f&vkZfh=+c|HQ -*AsN%G|1$c&H&r_Gh5tBx#;z3f{eQ_`Hr>m(Q^$^j|JtN{%D~G88K=@5UqzSI-;dw<*Qob3)Ri@84Xk -JXwpIQI4GjJV_QUZV)In?BzMZPAZPEvBte)TCf4UQu9^ZHUSrxZ^w4ap0N21N%9hR<6 -5>b8=_QVPq8ztTM`_x$RWJb%U9_hC}I^H@M)c3b;obRU$#4~R7IVD>&P6^|CGOq~a%|o3GpNe*`(a~FmByU@6W`M0-E -O+?i^iOcb`C8(r}$enr*xm)n9~nXXQ#1Nk?U0Y!PxjnpJbiVKfyX(8nsUIW31D~#++`8Sf`w*b;^iZr -~O(v745Xkn34 -?@A{^WMG?lJ9wu!QU5*ZmuRi4kg+Fa?y?hafG1~cVa7ts1FRP(iOsA-yv|3xJ -M3Ciw<&nn%NU{)7;h_FpKugt(`<)v=(e5>6_tKC+BWbY^8sM~uD&?K?|dx(#@OrIq+hBD -|KE)G!yTU{8X)tX;2QkQp`YcN&^ExxJ0at<%#(6~jQ=h3=LPQ7f6Y7qj9D@H!!U3E!Xf3iJ43EnoUa$ ->QOhgp{3^hjaX$1-*k-LTkH2*WDb -H8Kg-q6c8!W-4SN4vsO{9rFr0M!4;(lAEz+aos{d{jyjEficga|iqJ*-$%8M-g)T1!Y8*RGY!WnGFdg -D6L7M^`G|3I%W?6E%z6`AuP411Lg&M5lo3TX9v&xXdG$CojW=Qt0;Pi5sH<RJnR~Fomd8X8`uK=+?wk -SSw)b#gg2}i(x&yeHOB`z#;;<6r^}oHzfv(T3pt*jqOEO7>K@|c49KSGqI=&ZA{~5frMqVGl8Q1d)hR -@YD@l}WWWQqQiOzUs-_9}6256r}ROVhR(igfWFPVF>NF8cyI@EyoOAD_1U~&RzC+2Z4%2S8BAKzRr>K -_!RcC=HL&reWoeG}zr8@-abjxc^1ixYdywpz&t`C3m&zE!nhe^KAKFhO)%*tXVth`P!<%-n~Xwtfb_d -!|3zPG7MZdqrw3{x5w)`FxdYSx&}-ET1SA(?-zpIgu`71%Gt)VcH089}dmE#X2W$GT*W+B8?|hSLmlXE-+)M^tkTS!)__lqrfk7ICyo|F#!2;@a20d -xFUmcV;not=I(+xpP}_?6g{=Xc8;{1*7FyJo*G*nhm-gd!JJy5k5cap*g_k&B*K!WnthNKCnmFO!*4) -nj9@Oi6ygAnsFG`kfn{Z+a>I_!VpiHBdG|2fZ^=cI~s8Qzp2Qn}yGV810!hHRg%```E)^Y{t(aeQBk@2*{DXJCkL$UDLGqr -Aw*dO#K4p!UG=#ZmZvP-mFD^c)=!NvPi;*X=&zdNCkXH_e+U8nho=`vM-jBJvwT#wABS&yK@SQRLJ@j -KVV$kMmjsuo>`b#-{I{sUFz<;9tAZYwk#Mxcyex)6-X_YY*q!wM&Gsbcqw%&1Pg+ljO-puq(nLwAQ^?$SHqle!+N-}uHiM#rBD^Ps)bmn~lU12X1W8}Q?6SZ5z#k^?&PP4H>ufQ?Jbs$;CfQER`~-C7 -PDUs^M#I0rP!{KoN{w4XGS@Fi`!+$#CiA`|}PFSrJ}^0BRKW6o{-twbGpFJvJ;rAOgY`T>3?PW2jjH1 -7b6R*^=d&dO)=Tj|Hm-0#h@enwRN3zbwD(Rj>N{Kh;z_)nOd*_ZP3$ngCOWKBs6@%_>fS~v-`FvlHP?sw;v95tT3u*s -tpV2=5z!aIO@H458cFW|CWNcpm)Nej!c{iFkmH!gu3^SJ9GQB3?>9w!bqqYd&iGaFn>zJ$5`JLpNjcB -S`WdrIr&4Ru3|w=!VFn-0g=PCCH;0^bGQ#eSZFvH{eg+-5EYyTPwiOnfh9bYtFn!u)muPepuRFV&2OP*w8NW1@NtQ@QREf#`jGIkJ&uD;_!eq-wk;;ty -*)3{GXA!));jYRo!>6mnffZ -1>InbGtvnc=)^IUmHJcEdjb0#a;9J90sxJ`dr2QIQ}ls)uPLDU7A@1+4EnGJ^x@yL!~6|>Y2UzD;<)G -6nR{O8C6=Op=O>)!hT%|~;_*h?u#CztNO}mI`N~UqY&F4LMre7Rmrw&>Yy`F~Kha4ID#$=fjGkFWooBbBMH~$Ts<-bSD)G5dNrs;^%}^!2k(+E*X?vJTDV9T@NS`lfN+g!i++3tHb6@(R%7x1NcJAx18Nbau#|r`LAEk`8`53FO9BmKtrkOK2K@_s=wMLtg*P#zW*=1hp+1P -)n@lA_2zAd}V>!TgliR;66lcrt7k1$@%R=drr>|`pw+*BFs&Qj4=>Sos-dh~3lt;x(Ua8likaZEPGS- -(?wF-93PQRST${+g-E&yw$%Z;s`I5jmd5v!~&FNry+-wM@55yB4VjIBnRqe0dzZ7WB*e)mZT^)E!Hif -1i;@vHZO-Z4bVvavk#inRyhcb6d(Y_I#*n{qb;4MQcP(#aQ>sr;MCSW*#V>lgv9hGor7*8+}cqAEwb4 -6aRKH$284115~M!#Qi1Gh~xpGVgK3ypN=c&7>8HndB -wY-ETs^1TFXRES!qD_oDpLOVs=izWYC)f2{bqN^fSBa<7p)(@mbiw8eb~^I8U)%JJU&41My_>(b*yYW -mhqxy)H%=)ewz?P$~8Mm}ISbvnTh%`<3hOjuKoYy&;)2i*)z6Afm5RMJt#IU)>ffPF9Zw&01@N}p5lm -~H>1Ugt&nVS7W}D$upLGH+lD@T;b-m~@ylQ1S!xtN4i?Y|%vbbgbP!Duj1L=Q@Qlh28g5Nb#5KgK+J= -Om7H0XTWvBSbcBIc%RDuVb^UD>y-1N>8}{34^g&UVVb~m_IxO8Ul*?}Azh#?Uo5-70mfpI@5QM1(Rnh -L8T9((#VQXypGnhR?+dlvGFA7}*6=Fugf^^cHRv4kR&{EU=dWiBFWN-TNNvs?&klq(Apb?(w9&6C-gW -1iIMruQpJA74QNMrFr_SzPP5s{8r+i`gysOX1oedgapc(nIadsCN`LoNXCv73k>SC;X&F;1J=_Ly%a( -s4&%Wjdfe|U_ON5f<)WoKJG?|LF_3klni-+8tr9;pvqTj>GKWe(3;nfF;NCjE!}?$tGU=M#Lu(3dwE| -6M{GyhYssN}DPE?rl?N>uCNwESH=#;s+J-yxFev&Z@B5p$F%DfOBs<4pLE24-ty&lWx -@|PLqab_MG8dCRjybr8)(AKddm+MHLlm7!a?>#76lOlrD>CoO7JWd17486KZ?-9d*aTy;_6u?XlPgYp~vtV?2LCjdAPQ#>0OaXN+Bzv124J_OA}+$FaMk$JL5)F+M-~*tO -8`2m3Q)kulz7S~!OCn7m-1^JgZE&iQdttjD#+np}2)HTkhIR?_I&Pq8L?#G07nj2_>$G1lg_7;AF|v~`rf=!N4<_^6K^LYpb?g -lX)a14@Ri=f1Pbob!kFVc&)tEnS&BpCI3qTSfPh_@>HJiRy0oUC25y1b}(oo&**iL -IrVJJ1-4n{UY0g?Hs7&z(tcOQ*H{@RiuQe9AMy-`cAfmn{H`ub`hoo61)DNJEAlU<-+j8zKD^*?qn0` -N&bpTNNoX?(e~;oUs{OV;G@h>4rcG*n0BuadUkBsUgO6fPzdq|EpDAw<9#0s#)TWS!LI3>!=ah_A<{o -0q^E1|5UoY1EFBlhd)0GSMXNl_1ascb%U1-fd5Wv1TjydL`f9B@ee6h@rS41A>n{|alo2OztXg6ESZe -24Rs>-s4^R~sG;F!{VotCU{411J4wQhZTw(>{r?)i>BbPi|DnFR11_{JQz(*OT`)7N5rzy4S{pU?l!w -4aHcM^wd>=bW>AK|>cpo4C>Q&yxAw0@gxb2z;C!XN0Niilr|LSTb63q~0{g$9!H}-dBEA;EB7H4r3B{ -pB#DjcQy0GCx{}=nsMMb;CT$=*oyj}$NMRGU$b5B;yoy3oN9}-$=k&oYfWhP+warnPRS>~#533U(OT`(ixpZGU`09CT?`8O^05_M&ko&=o@}>PM1_kq}(#{O~x9T|OKC*u^e9Qy%2 -=6{4+U}%&6>X!Bxdn#oYL{^*G7f|q6T4`u9BQ<6Q6~WU5M*1S<(86QVsrbg;OPR1${xX%ta2gC9GS}v -e$hIbU!00^*D=+s4>1PG)DwOJG7GNFNWb$O8D(n@aqHoxW{Eqt23dZ=cnpjcF^V-fKABal(4{dRNH~&^HmwFY4>|-KVpw~_dXSB3r-{bS6 -BzH^}}#bemU;W4t@Gp%EJjDZ+D#_1^gG}0x)Xc^!eT4z5f$CRhWM3`0Gryt%vg?xhMHRB0IkNbrIDc4 -mND8JG}Czn_xtvI>F+7$w=3r*8XzV^a@k4}`n|39P?(uP;_-W!rKa}aa%Uhx#~MudM(b|JbiDcXZo7X4(1D!x$S`uc;C2sO(*l-?8Unc11bi{{QY6uc;-wQFUN*iW{M_l#y3RT -w7>qY8Wa6z88esdc&?!zr3khdxmYUjyTT$a^5EuP96 -Qx?W4EC!Mw{|LVttmh4S1pZ6xoSn%?BO_DdPmmRrhcK0efu|vND&r3PDRCD+lr;hRXcn=KT(T~0GQhu -IY>QS*@XA$Q{eS_>AX`d_0rHNVFBHAFomkyjAPq$>|T#lvZCGFXCi;U&#BYqGce(9I?vKyR$zX!O~F< -%UIq-ntdeD8Qx3^}y)107kAv1?Wo1J;W@S)!Qh*z%1yuLro|)a(ZwDC?=IVczq_%-um>w`jeq+}}>cY -t;C;zb`sSy-an2?$5{mVs6782Vk`jqAwd^7Xh-Nx^DzbvgR_{NS -nZ^$bAarjN21M|e0`=H;iqyp?wWiz}M?0s~2aNt7{U}s~b@rYD9 -XjcZ_8T(kTtx2536`$87l|cOw+p(r#E924oxZ*4%-bs5wri;O(f#15>rC5j`c8>!nhte*sQCSpIMYvo -ezA__8sItVT=Zn9jk?`E7WzVExHlSdVn>3QwFYyn?hnhYl-syRo_r@P`$WcVkcU-efP%+sJ%Porcd2^EE~Li`0XEy76p?HmZ -`g`q0)w%@TZZlzs4?QF`x=ZSY@oW~KLhDaTR{Y{>gqx>-*Yfke}@5lZgOUD^#! -rB8FiO&UP7klMcv`JpirK0PBlcBa_>K-j^`Fyq4d9Tg&Xim#7iH_}M_BH>cq?3F!KuI&{No{cz{~&Mo$uyLXpr8FjkDi0htd#B~QS=lHf^`{BK-w4{>n2Y~ci3XfM{zlO7#w_Lk*MYGxm%}I;Yn66J(#K+JyePa&o)4KL=fk -C#3-3}_8g0|&K#K#fG+8g{pA33zVH^RyBmlZQL(@yZ57d>**vQU`TFk|Xzs^Oq#dY*$(=M6axya3PBD -Ie4J!o(~XfWlx0$Hx6xOCWB&zNzYiwcUD0|q=FQZa6psYd^%JO~-szn}4EW!~3Cep{lv>r7+3-fzk|W -RJ|@)RXFrODxOtk`{g+@BV=Qf5Q5A&Q91em~3mU#n@|k4~V&NPS|^O)U|&RWM -M?rQL?YMCxp5l&vi?0czPqovoK;j1>6%RpF@7m92>{8Do5TKa&FXb1k%4t+vWN<)4$VR -1NeRc^f1Ky1y-9+joUsl?pSf`WSKQ#UQ)&uN)X+k-3^1-+vJh@txliqb0S^p+h+ExQu;Qhj5A96gt~Y -~t4Grt_VM0J%RFGW3-3I9J8sQ(N6VF_UjuRDm!OSVko}HWNW-&yPko^L!x&?|8|MY>Zp`}hv0!_V^?$ -1F`wG*Ic&Qmu7n1IK8M3JPoBB#Q$8B~)UeEMJ`iRB0D;|CRMZ<_=TnTxV{Z2FFZ`$yFdRS#hSQX(L?^ -N#jK>7TwFxn60QE-?0t;oCSrZgY@91^9Rjy1{hl -}{0yr4Gs}PQ<0(c(yuSR6Pxvy7C8lFgLNMzV|;+dexio%=%RmfVUf^?6=D-Hyz)$ZdB#IglB% -YrKtX1PFDF1j7bT{2|*U&$s-+RD!G~*e(UUy4MW-HFYAmhL?w~PEuxiu8p70yeW;kzO(oO9NNHZ -ss2Z4rnYKfI@ChQvtn1rZ -=5&tqQ7roYn_=H9#YboQ-az}5U-*1%K*krAI=FRjj^(-kJYe$N3(s)mC9dxv%4!zDbq)E}r@QL+p*0Q% -L`_4om`FYLbcD*rn-y{u$W1_|Kaf!8 -OZg|G<6OXWe;8}H8*-{;fzN7b8WZbet11bk>clr19A;z-XCv&s{|M2|Ves$hg#ffg-i|q%F{g8LB(ze -__%C`Il&3<5Of}%l>p)Bz&S -O3NOFg2y3g6}8U2gbYCEs0wcbA0UeUa~`;@#B -nyK=s}6z?t#zgxg}bMbC&_}%CE?kc>yD*Ud5?>>)rpAWw)Fa_&D(1}&$oL2hG*`R> -SyCy6idOS;*d+3$>^-Z0SN9<)n#=e=drlWl;Xpu`yJVBY<_F3k%5QEpzhW6-vzn -(CB_2CCO&J|Z4z`A-N3v@UC_{Z{GDt%AMI@I5FKD|Zy)~}c^yk@NKlW3Q+j~Vav2WUt6VWj)aapuWy8 -Sb;^RQwBhh8ppyj?rJkI@$2wp5!*i9ozx -YJZkwk2sh_RBN7|Nm+l20Kradp@(rVPblexO?9{ufgtejmnv&sr7Bl1Sa%xS!A8SI>W7UDWv<|3glk4L+M=%?9l5-X7;&2|Dx&b-xzEPx-7jSw@d%*ITq>(}O$lzj^J~h -xU;Mu3b9RWRwAp7CnM*N2Utz=Cv{x$l?cwRV>Jd-cqzAS|_nplRAv|ZEI%BdjZDZar{TT*SJeajl4_f -wc6Y&Zs%iZwu1*1k2GwtdVv7JuA0dJ{nh!+T=!1@^?^3+#!-7uXY -vFR&*T|2Ox<;!m(A7GG#j{2F{%G~ZM5j^YVT -KM-EYN?>Gb^3vGJ&QzrZ{Nq}LgO@?`p;X;YpmX|9kq<%hBMD?=d}=V~*~KKeiL?z7p514cQNI#Pa52& -tG&CSN=%zIv3Mw2^03#w}O5PF`teZp3FXh6j#~jz^SX_+_w<|Cg!eYn}_c?$6qye4R6WwV6t0W89Q~O -S|DL?{2q)ujZN2WJzze{95tB(L8&EUT=e~FRa(&97S?GZIVv(CU)eS_cna*!h7<$W8*1|o+Irl8QY6G -%&(yRyXm(||B_o6BRosS2=B3!9-%)G+AZXnN4{t;V9Nirv1dGP>V{WL6$d?_9mKc0(9SC(?WlO~)+Oy${|F}McpMEh|?)06S+*ohPLpS%&{z6<@bzkTEM$1(PukBAu_0bdEn)cD|dapxu)IJ`vBlMS -)*6i}}_MU2EC9(B(aNzeF=dZcE?8pOVxktx#FqyO^MYl3dg_WY8W2ijTY;Wy2=ztMheX57edjKzCmwR -f>U(H6k|XM6uBgwdJgcrMA^mV&H>uymar#NgJ+&?0EAoWb5&3(WkoF`Jj -(z#koY3OcnK) -^ZuJLmmHWAW45FT8Kbr%SLI4$Jll>dgp8xx@p+XGwfRPwAH2R(>95>RU~Ffw{&^_7?Cp`<3cHk#&2!` -CFYJ=G&*1UwSc_C!EBVO$c-mxT)DgY~k}q`(E(_7FwCT`(@Zop$h4mko_JrE#lRXdcrfuII%qa)u=$l -=Q-ygjJo|p0z)_#&EW4aaQg9mg12cEasp1%=me>6dysK)+20-2Tb43O`gx~9@L{pDQn7;|?+?%EByuq -QzT2`~Cp|8_9cHhH4%e>z+EpT=`@D#kNW_~my^-q${6PrgRpmpfAJtrhrP1N<$=vu}b3G9Li%ftJRL; -3CZN9Ol4r1uQcDq8$54jGbdM?HqoIGHjpsbUz(byf*2Nx!<_gmVvf2CU%dJDSU+D>{##kCcTzYkauMkc%nE5eEJGGXPgh`H}SBkzqi*fht;nP=Re$y~nWQMtQmXS1bO+ -3M|xfHBIaH8DQ_Gs@02I)1H5LwbSl=9+Jcs82f)VU1dI%O_j&Mc899?!VPqK$;i8UX%WqkjeU1V9$cK -5GTGF(O>=^(|^zOR~%oie(_hdZMPhx?8}^{Ef|lP(=;37F>_y@!+E)C!PM4sb40Lzs(s5)w#;i03|uL -K9=vB<+=bNp=Y?S{ZMCNffA4Y4`}@7h{yUVYd@rbP%j7vhE|puABB5 -)>GaZ%pT6#)oJGNAF)_tG33sGNlimDk6o}pywA23eO-kGiaQzvY32YX6zjuq1WU{9u~uK0_Bfr$rT-Y=MJ`J(C*dZ^=P&OEH{AXL2I)6@}TfG*9=Izjq*%U{ -4n@n+5j;2%uMWM`p>-0dYL`N3$YDc#6xY!%4{%}Mv;(ti&enlTC(yRRyEG_DzKFKiXK07D -i#bk@riyyncNC{K=UN)|!^e*^zKkaQ{;x`qKK+kZXM#RsZ}M)c#-f*SK3V(wi}x{S;NJYfx`~YSF6u* -nitBQtE&7Mwl-i=&o2GebGtBlbHQMV<126XM+*bNb%e=0D>)@r1Ew`W2OPaMy(%9BI+I-+$E!Jc{eSF -qgymr9c;{r|5Y>t8NT1q^pL`nZTO9}bgS^c3lmMv#YwATWwy0oFWw8gqRL3odX$F?MkD%p%5S6vWF<+x3Hac6@6eUexl^;;eLKST^ -SyRa7?=ZIGMzq^whHD`tepCQZu!6*(O|Y6*{;l8zgJ0%hAqOE5BN88Np -$Tl;9C#SG42p7t!+kvBjUtHTYaGCDcan{h5utO6k7=P1Y;Lv{JSIx=dKQeyY-h$Ra;8HE{==s(9rlnf -a;2>ypvzFlupndWm$wnJ@01q8uJWM5i&(IvLY>V}MMt#PIpv=Q^q)Yu%L6`7-5q`H|-2G^G(CysfIi+ -&RbIlWD(AHk3@Dq-K1@o0%Olg#^DNzqHmk?-?_px7ISl1NGu51hHT$vAf>6c3XM%qdrZE$Xu_`_J&6KmHV8dyH7mEW1$o;lEoQ>8`@Tc;*U^5R4){n6~82QKhr% -*mLkWmWW{-!V90#zR`JY_SL*=JgAFS1;%XjBJ|S8Aue6k2C8QXi7fXaV}I93$A@m#vr5(n;zu+G56N$ -U&gk7;yGo)jWGL?0W;740A_OmFZ$4$usS#fR=Uh*9|5a|JR4wT2dt)xRVJ+ZQJ=7~8TUc4V7C1p*AfM -@te7w>jnAwj+)S7){}eDw2MogS`uKA*^k}Hb0Kch;v|$(T=U~ -2PF~6rZ<`heJ4=!g+tI6JgHP^@T^P}KN-gh$K841hh$dhT#16IIrgeLmdRQ_YyWG~~+nz`#+PfIz(y~ -`pBc^lUI+#Fh6uTaA0Sg)OdtsEfMYQ4n&By1pzYt#D47i4@n&jtB8o4!I$r6Q(@$iqoss7&nTUbW?iP7IL=|`v -du9rRoSmYY~pXc6thH?OtGvN0Irsiy^dZVYL`FuFu5%yNq4a14^nCrb$hf@Q4`DqroH&EQzwdF2ditJhrJRT|jPcTNi$>>eWDy^447QSXYSK3fEtBPZ~j@amb;?~QS}4w2?4SSs3j{ID6P%gEo#c-_(SnIrtneL=g -F0pM7HUH8-9Xd&9-o(?4Hevg&0wse0J{TMSuJ$*JS{}#sUYyL*Pu6q7gz^nkjZ(;n65uDJ$umAI0##V -IIDBa3@lxeoAIz?5$HFxt)r05j*0ki-;F -<3{cy7{c6-|K0?ci7D$BBdU&xdwJ>Qf~>iHKvA6|t}W>*?|182vT$$>JEDGUCKj7Cwdc&;K>t{;!Po8 -IQ3lDt3_>?J>A{@d_>iZM^x(}J%#m>vr4`gHMmM(4*mX^r#koFJr5 -|6s(J*FLRc9z^Fqf*hE3mqqxuo)2~3N{#<9?ZG82Yd2OYrGL($>Kh4)yUV&drsy;`?<22ba$g -%C;3a@-)81Wwlw(-9l=Y{Is!M&;f27r{7J4f=BTHB^9%k^TNV1g%kY7DRr-a*wwF7D{`&9NN7|;If_* -VQX5_hlTCJU1$?`oV18GBZ@5&H4jh4(>(ipA_<>2>^V85nGpBTSg%d9Oka$40`G99ICG+ -s$HPzw-~a&icgyDZc}Gqef9HY_nfqD%B4(4{~5x9`eojix{^Roz>~YQNp-Yy=lm#}hR>R8xBsuwzO36 -4@vi%?Qbv57bH4Jg%3qf;{k!gWE%_ULZa*i|_J=y;eK>VCmqy?!?fLqTs&8B?wp(Vj`%|Oc>fw-tC3( -!qet&3w|0qmG}@yz_9n>59jodTS;JzLm%3-~d%{q(RQ_X -a~%86v|;JzP>q-GO`>`?jC#1U&Ra@!r#WM67t(C#^Og_Z8zhHYbQXnHQFP)~kSz)b+=!bBgmIt$y_{p -|)(mt>r0|2gEM#v~uQl0KclG&-rdSKO1n%BL6>D+2Y7aqhC&`Hc{%8OB*sJUYBZd5=Nz(t)940swLGk -cLMbm)Fm)K5%a;+81%eU%aFSHQY}Z0rBu5}{{OWxrhmnlJiu$ap{wjN?pX>9Uzt7gF&E(Kp0z4J-kxV -we!M*v6<>3Y#UlM-e`>V%Wn+#>(%U2x^;@@_%aSvE=5$wyOuMXF{3sMkF00c5@nsu;Oo@y_}QG(W--P19VfrXi -dP;B=fqAGmzsDNz9$Rw^EDAPmGpE(rwf`&8fvpQRnF*>epI!~z`s|-6`TC#M*m;b{W4>elJ_Y13)cv6 -OSZ_hz49+=|7R&ZQe|kGXizjL-F-LCZJA#KbV%60dJp}Q&0N;kr>MA@0_Rb9kEgqT@)xzo_2KZ|3*QI -F@;}&Z-n%hQt320(IM;h++_`x0UvX~oY8Qhy7{LEirC6uwdKYEDMRuL{E3-V)MOUp|1UbJGe^xk1o~R -jT)|o%6`F%v6#3yBpNST|5wCl~NvTyUZ!B>4C`PZh(fiLL(5M+pb_|9*Ao31SgR4^ad3b$PoX@Ld0zy -E8LZ6Fg)(KThG9cMyGy$Mb|5#{=~7G{&*pB4%?eL(4@;9@Im}t`J#-ykctEJ -mC>j3M8Mlh?nXr=$Yo>BP61H^z?ja(rt!1-O&b;8MMF&y!2>}Ma`!~iOXuQCGWFB%b82qu$^eUm -;rdy^eI1W58$y6e2rc5FgX9gUu=q(`1q2|ST7SFYRUn9ed57~@H%51aXnN5^US+S5EMRSbcB#K94 -N#}%ulXsKzY+;t)OV7xrWZ2z`o(9?L9Gjk+yf}Dl87p1Ld|J -$xGuI}%KeT4~9*h^=4OpXg#uq^QJUcy*b9j#+&*}vAoA(Yf=ZXWerzYyBTHD*16GzI^81q%h#=Qv65z -vNic^+$%>$=DrB#09WvE~l&?fAc-QO@-pvl)$)~rn{ -0-+qyFSFaG+VL`oMh~*;j&G3!~O?6dINLD@csR-kII3yQR;}sS$p~d-bKGD`p102d2hG;aAe**uRlss -^VY$~_A<{j){Q=*n-kR7`M=-DQAd8-l#9Wq>{=>#Z`*X}3ferNFUI2MJ$`BRB$q;+n -_VaQSbtEHcvcqh*|!ZV+~rNgfJe!zGU%Uq;lwv^^O@1&d*7~X^PM5|+Nv08q=q^a_4tHYO{f%61>QcK -O0;=KD3mOKzI4&DyA)}v|Ey_GN@3CBoYjsGtj1KaWKzm5EDZtB){$eg)=-z(tjpX|uJlrbd|#Ae1DG~ -YiW?*v6v4d8VHZKAavDSy%5j5$AyoV`V!50(~#r%uZ>Y?*kU$2j|hg=}lDVT1Ancy4^Bt%F^biIcK~S -%!R~Sw=VPBVzkWj`jPXEX#QN$6pJ_!P)$zlH0H5T56&j@T#X=P~Ov#%b3=sJq3dpf7KDxSrVuCUdC@a -6w$`IC)M5g9DK{do@h$YgLk8iyAou6LK#Pmejy@G*+kn=rZYvMji*xbPP1 -@$3|@Ka=^6dzG)aIYH$MZUHY$UcMG%xlkUPWeF-i%$FYu+gy -x^(+PMozJauTy+jmV#+v#kNuDB*dL3Xk -EU$F8GSf;O>!&i(YDSrTz%NNRTM~{7RNj~zzm4GSnzKaJ_cUikhn8G7U`=ul$~Y@wYwF9nA@E%?_H~kKYX;i#Oh1yFsb#uJcYls?Oivz~_Z; -BykVteM6-jlS%#CrMf?*uS&M`9A!Lu0K!Z-!jeHhz-RRq~)GTQt{qVT^g{iOAvZ~Bq;M=^$@;P3MDUT -(i$v-zyd<7kY*X^bIROP038@1f2DV~yUmKi59csQ-H3GuACfHOHVGu7aoWO^G7Vk`wcy#;j_oQt@a$UKWTv0c@wcWPr^4%dVv2GgeJInrQkY^+BoZkWr(C(p5)b*iV)@N* -DJKC7B%ClSAX`o(y*^E|)HlfvlcCJTXykp`R-v+JShkEI2(p!HHc>e)^x1J2m;}GPJPGdaXq@I*IQAXwe#XCtIr9+xRW3g$?|l-*@M(D$WA -&YW)!N0_;OD@rJmD0Z1Cyzf!nujUvqZlVaBa%{GbuWP#y1!HO%mqz*%<{PK9tFLl-nf)`P{8+k{;2Yx$h}1SuBbdH&4zEK -=X^L1ibr$JelBBMa2-aC2#WyhBEwNJ}5 -!Y(DQ5BlTvXp57$$!B>7@MZL57M%o`RzlqRp`>A7i<>7wtT-50-Ezm#cw~CS^OYb4V-u9xtvETYYJLH -^J&ha|>G-%DaiwZT{qY6ZpI&Z-!e0{4 -PaPCUf}tOi(BxQygGi#WxxxrcY>B!fj$rA=l!8Q9=O|&Hav_ecC%+UZF=5@%>EDSbhjOJusm~NMT=&y -aDWyrWy}m~%U8QJr?d1$tv+duNVO#Z$3V|`elTy*TzotFN39|5YnV8!&&`1&=wKd{{Hf5EQs^&NbF7wzr`u67#u+GiV$uPuNb_mvIsBRr#FXb0>{QFh?6j@%m!n -6+RI=6Dzrct2oQN?wKXGjoeYz-^HMACAX_Ct>0-;K})yH#DcKcLlrYpTf3GIK6GaiE!BW9q^;z30B@8 -zhv$G_bY$9ofEt}e-61fO{B?qOL_Zpn?YL!@cs45#6BttR{I0^X>Sq+3l^mJ)`w5z_ovHfuFEyW^AqZr2kM+w!GW7_#dMDgKte4qUxLA6 -YcU0ii8rKAKT8 -ya-?-`Db|*jvISSrsFYvtoWR`ay<-MD)*d0LovoIg>itkuc7`LeM)_7642XK^kDXc^KhB~gBXYLM&6e -ne^e$>0}rrUP+;r)!8?%4fSoJb9c_`2rR4{foRUD?|FxrE7X@PKygiGj%?Sa8$K-S5rQ{p1e=PhaELh -C@|XT2w#1NvfZ15&d+^a`FBF{4T%wj@@i?A>Lm}z7_S^&XwmwMY2BVAMGZOYH@YjhC)@`>*hWmvxaEz -(Du=5_|VTnZIdF_@cEtLH6$JBr+ueWXZ+IlORTnv;G3bMQ;CoZv2W~mU9!ZT{8i_}366>bR+;;9gN<> -@v2PsM+l~ZLNEwlP|5ZyQUOPS!UKeO#yxwjdjn_OM0+{EKsS)jhPJB^oI@B!NO_uGhA89u$+-@G)1^t -lw#ieMGY!~ljyPzjR!`?vqke##7S~Dxk8#}k;pnf*y{mlq^5o#yBX!}(3V&Ts&OfPD-$EFu1UH!R1FB -r4t#%Owhxx|KRc>XnM5%U+L=bweLq^GUC$Gx6+O{!1sDU2Zle779aY0wLnn`6+6W14|$Xv?G*^P+Igi -Fx1uav0aH1g@RFZDGNZ2wHI@l2(k&C7M=*=klVY6HPlpcPkwR!kT}D^44<;(CKtFn%QaiA5_ -0@w*1Jg7Pk5Rf}(wJx?3-VuC?0rr5&tVh!j8*YMfxcNS@L -Aa|g|Huw0XO(=4d%)D?+{{h-*TI`cbew??72|ehoE<^d{Ms*cwdl!ae@BSKsKYz`-(*2eIai2`SvByA -T{Qcs4*vnX^?gNGuxzX6Qjq7cXEsBF)Id}SZ@TBjxb0NZm-7bgm^j`@3=+taml36s7vA?BSdq28GE<78q-Ze -X3RatJIJ-o``~&p5R1a_?i?FKo{j%eJ837itsFw||CunHOJp;f2_7Tcka#%%x)5$jQ8on%H&O;jpgyw -Z~06wb43r#ye|DTXGgSUS;OPdvPmZ3p#4X_}_-HKgf87Mtkwn&MJLyE#Q-KSn0>v$DYTPKK!kSdlQFE -Y)-VrvIqR_h);X3%mI|{Yx@J_m66(ckg-kvKc=ZZg8DQ-_5ug=lP8A9A?47Z__K{`&k%QyjmjGvsdVMHRJ2qD7!p|kngk -1_xQnYXg$Q9$r)megzuS*a}r%+_*p?Hv-8`}_ucjn((cVm$C0C^iRe_WtnDMc_FuS3JD%XiW+GMgh7-l06)4l@Utycd+VbE4Cr!7mJ2jy0Z``xxM?nr^Y=&@#O@IAUVImR~JZez|qFP4h$PQ -5*-kE4CxW4a34Wd5yY5I@^|&P3=3<6wM_>m1ME8q+Z2(DB1?y_gM};AI@@2Xv3lI5bYd{lfJ*5uGQ@_ -X&V-eT-3hD0|!+;1QJVZ}RHt;m&JwDm)JOSqc$FhhHjR3F^K8qd=Ws!(NOLZUhKNsT#k79A6Lwhj(B^)_^+;%ozd*aY4DIe`Xm`h9T>2e9(>aKr=?r~Dgg65*ID((!2+w!i2kA!kUyXgUr -rnooHl2z7Gv61bq#ED%@qBYUAKh^T_jDBd^sYa`ZMU!R{$HWJt&|YAz8%lLMLQVb*Vg?ET{;sT`!9Yz -F7V=gz7F-`zK`LJ`mcL8VE@$$_aEf@I$Z(X{SNO9yFjMkgU;^;_Por)t-R+i%4zPkhOM{4Z={}G!Zq% -2PbgogV}F2tgDq{J{H3dY=A --{vZsnYlf*ZhaxZm*$H>pozS;5q>UD>8fjm5(18^SN=sAzziVc9#p8p|T@Lsfmf%nlvODyN2N-Ysl@o -gCHTd*zIqwQWyv-D4lXS}in?cA@%91dLA_DPwh17vZg8=|I*->hGNv~b_}nI1trt=UT8Ua?0) -q5BE!JcYYJZ`wh}=fVATkdPbVG>9^zQS6*uvgl8jhnR%QLekSSl@`~^^5FXR<@d(^|ewuM}oi1cAzK8 -6EI5R!k@oy{VL3_Y`G46ur@mmkDdTbBp`M4qYJj$wreEtqP8wxU`8^$SeiSxQk9OQ{-UDreUy>U+@h> -yD1)`cB9KaFjM@Q40@cgkL0H@x{4j2Vices^t<`vi+V3*P&;!8^^i7_oO#GUgZUB6qQIuo`??4{C^8HZZ%JTVB@2kjuP#vTs$VF4CPZd66*s*pIvC!%LZV^p@KcT`X4i+w13FFrTPR$| -zm5N9@o4Tm^*o)g-E`am2E&uwn*1>R3P&l4(rR_ya9=6!JAO?+=r_wV9+x}n|S)&owxn_m)VNQ6T8l! -k__d@cs$KaJmBz;iHcOT(HAfyOJPKE@q5&WGpDAuQJ%pjbaP=={#3-B-4P`&%^U=fT87^~u)x?e$TNk -IvTl>@`BT8`?vw=Os1$d^x{foQovK$79^a$NXE72yQ<_72-PXjk6nf -L@0RXBhE#Mu@}z2#k5`$BF1QFU(+rE_x;B)V-cNWb1%*n{0n2p&(=3ScD(1y8;>22_-V(EZPc+?G{{! -z>;C@YzK1W4!`ernozSu)Vk|brhYcF4OByl+=zF@hIs5c1b#xESvIrWXPHzR`YH>oPvY`mc^r?`?XFI^Y!;f8J}{p&h5zd9WRBW7TKa%B~IK{ -ND0L&gT{HvQ3XyWxrTaE8g|7-H5q2?Y^JeRQqW>e;IB48t(Dg%+DfDVORKfzR&1};(5ea*BkC>#^)m& -hOsLRXOyAWUKZbGUw7te6Mc$2bM@+};=uIUmZ;s= -n@ffkyTlT@CzE-@}hKtigRFS8*Ivk$H%J108hFV)kmB2S1)<-_R#k$TbE&@2X?lF -kBl57hct{2Ipas_vv$DzdnqIamugZ-ryo0j=}I7d+d+S?{nXSR%^>H^djbxac&(tDm=qx(AO#ATtKYT -TfhzX)2r!fEXIBe=l$HzvAlj>>$2ECs2b}6_DvJ^i@h+xA!@UD>Z=s-yA3-+UZwtR)?D(C)MHHW1EOwoB8_jMZMyoen)1w)-ZUXOL{(Zezfa1iaGnl{)KjuMWoS=r*UzgY -y(Zr--$A+dQyx1l$)e(r_q?adh9xf5W<2P!ZrVGCn1^vQ$w -!r1A67w$V9b@L>6)|^+^j(e~#_1>^WK -(1GhRR#x*c5n-C7;NVK=?rTF=N@6kL0?U)l<8;7>M{Ue&|#r|gNC`MbT)9-w*7Votr!pEW -$Z?qfySi)*SM!8_@9o}dTM5RgkZ2fsD03St3%?foJen9q3+7(b%zJic#~D?KMnfczx*uwvvN-+~oa;< -t_rp7DzKcFBVEDQ5JXEbhtA@dxHx+|o)LKl_`2$8rh*a_@f(s^CZt`aa(87q=P(>_mKwPe-5sLXCz|5pIr%W%?9A!3`fx)#>M9pk!R;#)9(CRvaF -|luKieD)idEye3!**SkMnG>fh*&-$+3l#&tAZLql~OK36ujUa8cwem8-e$)hBB7= -%4_#rPBEZw!RL58Puh{I!N^+?(Qus|{~f`LXHv?Z+OVm|hh;JN*5vp{@PXLLGi=M#rdv&;p?A;kOfhk -M^1qTIvBZE=+rLRR>$>FlE-$D~H_`S{z{yo$tFK^sq0R5#?7L+GbK}*jBiIBYApQ8^B+TnjThvt}rYF -;wSqp3@x5;N7${i?hH$VAD)w;*VR;?g7GohkLc&uRQvm|86a0hYvBKgcS&;pTM9z^>aw2JkquDQe(T` -@dN#vP2^rAZKR2{$Ti)S?&De}wMelbv&{ax$=$?B@s_Q^!;e8Tb_Xj?Et~d_&Qd&*0Tx~TOzhjDmFfA -12P#o{0l$C|Pw$Om$yg%dU?82a13PY2;fY;Q?VN-z4f!`A$_D~1V4ii$tZiamQI%-N-8;Emw?6k07J= -qM0@4V18S+|D$7Eu@)GIMI!bj)kU9bw6s*9muq@jb6`{{pzCUTj^7baB0oU2M3cYEa4Wk>93Hfot2y! -H>-ezj -`TN)Wg$3&FirG0a)C?O5mw{jN`hkJ2GvHqZ%*cSM@fG6Llf2l;>S&_|q7!fZZ}&L>K!nl658vRCP(ra -Ot!-LG^6w?}vy<0*}*g*vx1_9-mD*t&UyY>%FvZ3Fb4!K4>Y(uZLE^-@HL8uC$rxB;=Tq&i+=i0fx?`unjJ)u$NR7zo@GXoTWb;a9V&I!^KXL_fO2XPC!^w2{2O^#(a -xft)Qt&K4kN80%m|9s7B(>3YR)sW0o8;^SXZJ%RIEvJPcKnL&qKwm4S8)I8_L -v^(obt>Xu$O0*;RqxUE(;h3dVzBOgWFX1^6weKKpD`X+y(c?Q_Y_*ow4kqpyVN*{gC-aJH-UN*~-Wfx -fku@A;YVLp!H*K0Ykx%xc*>rA!Yd?%Oa_lPbZq8mU=W}7sZ~F0E*z3lpF6SDEusS|wp9g$uK -t_y5``=WEIfu3tV$U>j4k?biR(bOG6}RvHabNshxCi{TreqoU`9GU)DIdyF13D1z2G`4@j`JEh?^VCR -GsPF$Q=HOiKp1Oho1m12WGkU?-_|x8&n;8J;JyvHGq?{T_ZD#9n!D>gefhpqjrBM9%*yc!=6l8DD8YR -HiX5dK#MRqiOf(=3;`8rYLC(5B+>cg28=tHA{n%2aV+zW(0s00h*9Ik2m1~32NtJ7Z($R*z5?Is_### -)J`NO);pp63WNA*mdT$ooo9SVk$Rb?((i)Fe5j)^nf7YXSM1;b8%a!K7i3;{pEXM%j#4 -6C4Zyw6mZony3j<#?wFxUdNa)R%^yW)?^BI6TJLRRfodsPXzdFUo* -9dayp+A89@%~_$%em`e9(s?)?BDQx>-7K@fbU?RvEU(FnQ;hyqyH?7YwYto2H^MRpD}-Mj~ikH_oul# -_9^Jjz&&Y53AnH2?qLDm++7du(}xr?!w?4j#7gcL8n6v8?kmJ}CDjjcw+;a~pDE{0Ty2-G-u1Toyl7W -;rLG6NiZVa9R_sq+tY_0}{MglSV_IlEomYx;JPdR|d&NhHd7@?!>5yK5Zr0Pd2ckG9tZl?aZ+7(^&`} -MH^WO7iT-P{<#Ltfy27Ftyfpz!7__r%tSqo!8arQFq;fb~Y%F3QFhE0A?%md*W=t1ziwA+}6GF*H{mx -ueCUhcql)I(o#DU4l>!EtK_Tbc46KeKr-?kRT@x`i39qTUd9L5%yN0*}FX?`^z_`yUrTp6hxTuU_ej^ -HQ$!Y+`Q*wxN=jubxeO4ZnSq^}ZSJCtS{>;_uz`4iEJJ=ODuHyK&Gxo|*QWhi3)OtKpn7?hAPq+%Jc+ -;E^rZj`yMbaL#Bo;5aYb>xm7B_d<6)cfZ_8x`&ZF#xbSB(7k(BFw7@i1MJNEgE*Zyw$x!UM0X{d=;jw+TCMcqY?cr?Fx$b}?5T#O-@HPapg~+Asw5>T_k -@qwu~?Zo%X>*Tv1B+zMUX^yKDH-AbVz2sgWM6J=q;z4P$f6uw`b599mFL3`upqVVuuvtRQSI;ch*?lt ->$d>6$rYCPY^2Itlx+%MUj@8dAuN}6nZcZ-g_i?&abdm694s)N20_hi$|@jyRb`3U^rUA}T9{FcLS1^ -m?WN<78hgnGR=zorhqW%yXE86xkKa34U}H2m&Nxq7ALeoJ6+#pc%Bfv3k>o# -u$A%9w~RFQJdZZpybN%kaKLByvsQdpz{=Rl@f+F7XT(`j)lHp4I;f#xONoxH#QmJpXq^n7jiumb(&_i -2-|+_C-}CihEHiyKG0>bx#HXL4+Q$HJY2;=p=5sI?R*qI$Vwz^@BJiiL#!Mk5|XKx(0Rt_y~4f<-%0{kv4cy -zWd_~rtoa)iElrGb9f1H6+GI>l5K!;2h3r#J8?IhGUhu=-@;jxbLYbt>bl3H-|$Rr -L~A#2IJ?T<`)+}xCs>3{k-%Kd?3>aVLLRvW>!+2fP5Z7>}6t0T-O!%RB+_HdMIaK;NL>=J=YuR5azu?lXqyJt#k2gMbdkfOTFmzXG3Zqv`>Y$oYtar8SZ7@0`&4K=p`KV5)bJ -{f=>KFCz#L3=1R-v(6{zq9@!W5g!%OajAwc5zS5Aone>JIYgvn>%WxlZ%tsFBtblap5Ap|uzM420bcg -ZdKzEog%oFHit4||khvjC0a_a?po2^t1ow=xUPMWe|9oCbX26hSUquxio2N|Qnl+Jt)1Kd-4aYMs4J< -yqqU2+InTKY5NRnVa-Kc+EKQTVe7_2ApQbeRv+Et`qoE%fJW4ngA$z$1oT@`FAms(Y&GI4ENGLy$dF9P;KVDG|r2-9x-;U5X6IX<8PwXGIC2;`g0rMX2 -!L}h!{FVpP>I!*6osEY)w?Q7Hc-91?_e#@GB*qZvp+ -|IPQW+i;9bIk86@EWP0F#0ovpWaQEP0d@B6?42gP8dFkEDYgGxPiRCZbg=^SwO;`(rH9$U5hVk$$$?4 -2i;#s5FZJgza_DD-{k2Rn(6?Un{qja6~`6Sl$!etdciw~BGw9UZ3o?X*!J+OZ -7t?(q}ygs7Od|q!s3`s4eR2L^-{W9Je=Fd6o|8iUuW_MPF^1%9dyU>rOCnWlAG1JC`NK55#&ycItOa2xU|+AqT4{oot$EcfG -Od8PMXhuQJt#-_uhDJ97bd;XfouQv#zr+RgD=MDG-iZHb|;8PZ8-0{H;WX1CXkK9PsFpIndvHJ -%8c-HOPH}J=C-|IJ$0op(*olCg+M&Tf@aq!Mo-p;t=1uTEuH(wF^=4aD)ZTL-Fb9LYR=I;CELq3adpe -xyqEBKo3c3Wd@nf*U_#m&@ldK*z2(T9kZ6Mci|0ircTFA{B@CTJI;1Bk{Ey^ZK>qKk>HAX-LrJJEeaj -}g5<^mn4m&4S)dw18+P(Nv-{OoGlO`Vi4gMBgD=P4psB&vZcph;|{`m*{Y!xnkc5Hy5nB+*esQ;B91olkTr(Z`8yBKir@&xzI&{fX$GL|bMGY9Ja(bQIA=}0t^=vboi3Z?SsOLVl!-M^?iGb*llfIF5~d4J)aM*c@930n1x!s*2^N~7_TJ`& -+dUKRfSlN;wRFaP;E?!Sd-xY0d)!7;%{Pq>7zH{nu^|60O6G9QF}32)cnD#FbPAJxRKCESegMU6i@?v -DMByUQ1R+&#T;!hVFKH1S6rch~D^!k&baH2kL%mT_hfmZhGi!I_%y*&6>`4KC2&qT}xRzE9&{s^R~T# -(#wduhxWLtHDoca2a7*U}w*2{L3}?MGfAfNq?)xf4j#29gTmb#{UD2f0YLB)8K=e_(wGOs0N?V;2I6C -)!;e}zM#?9MGd~J316?le`+v0p{eg0Y&hXwKf_PB*Ow^5L4=bu{%M2*2p19VNO&#bPK38>{Euq<>oxw -tUkTimu#s?xOrNlUa4F$1!sUc-B79I2|03a5QeP+C_3cU6OX}~Wdw9c1clk!2bnkzXPP*$mjc`lC1t; -C*S#i=m|62%mm--^yP0D-9-9MagFT&GLx$`eN<*u(X!d(bg5e_F@PdGyIuMxNd;kX+2{&+g!)-r!J?) -|5>zbw&&U#sE2y~e$M9jtNJ&qc!Ro8YhQ-?aUcwtZ{+18x7Lt^fX~1#Uxl=FD`H-JB`a2RI|soRJH(z -+}%b`sT7GqX%~hbx|I&T!=0qzmxPm0;#fvnk!8@||bNa?I?{-0=Vv7dUJ~1_8C1 -?GBqI!y(-Zt)^VdEUP(_W9f?-F<8y>q}+x~bFSH8W+UJ}IvU1_!x`%g_anfeGyHD9LxjsOFxyNZFW&c -w7WJnWRLVYm;KxM$yoviouzpO`-{JTJ9%!r;4YW7V@SZRR26`v_7GNBRv;b}*z`qy#`oOOr{JLX2h%_ -7++ywUv;CCnd(&lD4OzFAi9wdL-+;j~_CH7!xj^YAyewK7*X}S5}VX!+)4oik%R-wt3$^Ft$E!J7iaC -5P|K`Y>1hy~O`y7S@^>VgQ`5NVm_EK^~w!;q10l^%wS+b6rVMkQ+3GOlnF?e+&Dn;(5JelVQITByZzxh)Mt(tYQ%b{h3NsxBo7qulv+`m} -8xzOUl2cRT-OE+274Eok##sMly{7svAxq28vsw&ThwK)t6=_Z-iI7_PwEnl>;T9q-&tx^tGH3pkQnlf -lEN9p9(P++%-}U;xL1?VM#a2@uG!2W@;;=x^G2d)6WI-^QPw8)Rn2Rh9SNU8*xz^(wCU;$Zy>aNq^!O -M2-P*e&+h#Upx-@&WKQnacVsOJ5`7=%POcv@!Rrlc|qD6g4g?_~_+HB9TSqfYPG>pQ0#O9jpb{AjJmL -dE){}`LiR7?@cDb51P*fO$s;Jj=T2PPT2uXjz6t6Y-}TbAN4+Mjt9;tLxnkG;LAG;k+TN^j{kYc08}$=fxo*_{UHc2yu=+ -;ujjkK@8~R2lYlM0;_x?ftSQD6O;{Q+k|Ic0F%*qN`F8uWr*jkvI+tcZf>%D$1;X48Lbit2n@WX&z%V -W6)Kcc}aG?EW=I$ma2Y3Aq_?!HYhmE_=a*H4N -x8oOsO9sgCujL2h+qr8NxLvRHzltwhY>VKAbu8k4bH|kAv$|>hp=$xSu?YuKNpD|E~AD{x1$n??!+3HS*^kVv)c<{mv-XK~Pd>H&=?!HYH$C(0bI< ->yd~?MMFTV8hD_dTD?e(qO-gtBSj-9g$jwnhZy7dxMD$3bDLo_8oHZ-kGCLSJyX)r8XBje2^yND@lVtE7ij!TH2(34W0}R8#S&u@nbl-P8lD3G$tf{#3%A365~F{7;% -Ifo4E-(EOpeXHA|u^L&md2ONGFqMHql(71w^eE92L~DrF6 -0IY8f#^k|mxrQ3KH^qDe&4h!znoAzDgw1<|!c%ZQc}-9mIb(MqCKL=O@@O0mSS?CU#6*;p+Bd1K{z`C2x} -F0Ay2Mrb^=zp>_YHqm5*_YfP+G%A9xkoX+a-&I?Q?LW*f-~{PPh(8m`Wmo{D28f}SS|_lXnxWoGs(m -f2*PaP>A!gv)?8#l_?m7k5iu-Yq!8OzHCXQ;5&slko%NLWV+4ga0!*&IT9I6`vEp@c33ORzUqMfg=`k*YvyyYllZ& -_^sA7Mxu|D86C(K+a<0W{IyM&<(m%&I|0FJvzD1oc5b2v;{+)HzK3w;7#gQh2T)1=I>HADdl0T6>`6F@>X#Sci{#%-b -|^vAUl8^utS4+B>?8e&pC92!@^4PqNVtVepKwdUX@px5E+FhrxP)+PnLec#Ao&vxB)o#c2N5nK+=lQL -!fmDg2nS335pFN_M>s_4i*N_2uQt>_NPQ6QB;_R>D)mD+Ov+E#AoWAIv(yjaE>b^)yGs2K?k4k3xVy| -h;hSXs3HOlsC)`u!pKyfCKjB_7|Ac$X{I{k4fp9S4euTpb4!ZC#F3C9xFx1;_>mN(%svb+h8mE}!%oGfp`@v^)LkC)|5c! -Df%!U?jx2`9?(COk=&H{m2%-h`87c@v&0%bV~tS>A+iljR*u{gbTkgmYy63FpfE6J8|qPgtP=ND^UB! -jaUTy$Pq0zYpO8!oGw{2sbCZg0Mf~GQxp`w-62@TuHbs;e&)j2-gtqM7W-CZ^HWaRKExt2uI8O5RN5m -B%DKdI$?zdRM~_*2^SIeCR|F`hwxg$zJ$vOHz&NEa0uZlDKFs~DG%X`QXV-#3n6(3`xEviY#{7IIFhh -0VI$$@gp&w|5KbdJjBo*Ag$8&fGJV1;Wcq~5Wcq}+$n*(U%Jc~zl<5<$k?9kzm+8}hQQv|16AmWqOE{ -cxbHY)CLkPza9!5BgutEdM0?D6niR4dsh2%%LO!6bVUGg&sTqXGtJ}UVUu9NWz*UR`cz|?o7_=JNA`w -|W(JdAJ@VTHcjij(mPPnYotXUp(3z$}vC375+7gxAXOGyzyH!xP>v!xOHO;RjLvWq87MGCbjW8J-5H` -cC9eIGC_9oboUI2}enPBjsQE6P_+H4M?*kj-&iboIv@Pm=qH{TQ(2RtG!W9y=uuRIonZnEbh;(u(elFoGl5;NQFPr8HOrS9lPso!;a+n~7$d`%Y=Tk -fjg|m}?4$)$gV;;#lpX7E>{so^|l#YY2gXGVla>!HlVIpj$c)1iVpQng+Ef^AQQ06jGANoL<$1p~EK| -7>T>U&QHpT8KR&SzwBdyH81JtUL+k5=D%GP$*f)7C&(gN#$>H?p{u$5{0}DvPH-R`t*3@#9&7iVFe9u -_=TtJpFi89t*b;iN_5B`DNwlC93hQkR!zMy{mxxC#n7gJpL#(eH)*TNmj$#c>2j|`q^CGI5j-hKh)m@ -70&@2qw>f4i2hO^4w~PIQ?ZR>$*+*(B$Xf5SIpmN)jx;NFFDK0%;mpTO)rqYCi2;Z_aul?2IO=dd%VSPEhN44$t=lwO!b`b=d?p{X(vXDc -9-2!P9ltpLCvYr(6Z3Z;5B|d^@p2%Wu3I^Ze#FT^^ -e0&3gyuxQf*eH+n?srzMVc1YbIx+ij7>e)o{P9`k+lUTLAi1}5gd}b=Q>4ZgpGDw7S(dGd<6o&nKIkyy;Y4G1ry9kR!!S4-;MTCEDc}w{((J|11hG=OgWEKC -e&5bX|T^)%s)B^rIr3(Q5u(`MQ^@;Oj2`B(+}`e8;6vu@%~{!bDPCoEU;wrkdPjN~s3l;t{})NvB|%kf_t;e}E^gx{3;)2s`YlE18%M>Xq$Ysp{s*X -4vCl={%D8*V3mS&!>9>xfn4ze%P?c%=+aSk}{e!mkt7_Z9dBnVx1HCz$+~5x%Hdw+ttLxh_+$S;veb| -7Qrt5#CRDI^h=yXA|B-xQOthgi8s(CG|=8L#a=~TV;L-e=hY!_@LAmVfkFIBU~x-L-;UZx$Y|0W%d08 -J|y);_$#R&!e2;z5dK8!gYY(~55l{oJ_yUrEQ<&~EA>J6C8-a>r=&gzuaf#8TqE^C_-&~V!Uv>22%nb -vAbecrKT_b6GXI3Xl=&z8kxY;DC)ZJ<$X~Tsp!z4*LF33@+Bei_)>WsIzqE&t>)>*oJe&OGb3?9k_Lc -f1e`!%5*Uja+dMWv@CcKvLbA-zYzf5>LVfh@dBK!v7qlDifEY}_7I%^&ImrD6H>$dgeFV`LAy1rcZ*Y -_7#KBwe5u3Q%mCV#oVhg?^e>+s>^{|ezK!g5?7?FOV>K^*zZ=cZhzmK#P*C;!=m<+{3DhtKXW=gBC2j -h#Xf`O9@%X_p}F7D~zgF~VyJ%l8>+w;-p@%gJ9pr{y}hTu0wd{?fulu5-(8;;P8Mh_GCDzJ>5n@?T6? -+DS+|3c0Q<*U{@J{1U?TgynmKTvwN02ADy#F2j -<;pv3$B`nv`M-a{?{|5-mb@*X~i^%^z!qTol+L4x$|9Z)vu-ve#oNyuG?S$p~gS11Db}Cim|G3l_VL5 -KCBfN&Nv|EvOE%oI8Fkxw@Anhvj0|mZY>PKS-6HNYc94YNyq+Lun`O9%{6ydcpKZGBW{0VO$oK3h~Vv -U_m5&6sab!k`Aj&LdY%W<)^lb~sIwwCpwx)DK~~AD6VF87TEb{_`b&jU7%s`7b3b?PjE%kAASgC4{9NjkNm-CjUh;KN>r$aPq&Ca1>!HVQE( --?Xcp=eFRjML|?$pcf -)4&`5>@Gecu##wwhmobJYG<;9RvlBrbNtc^dih)#r!6b2R?;>##dNhg!b`E_TW9e4c;&Kf`6d#nnC6W -qwD@->0km#rjLSst++=Y*Et_^HJ%1?#7jmSf??o<3_PglcDNa%V1TJ -vNpTK4fwrOy&8@8*@U4aWV{&QXUIk@~zTp$|u%oob6JqTRQFD#rl)0zm3|hE6!Ha7weqaYWiZG&7{^}u?}ca$01@J)S}8G);XN{6 -WF5G2VsXkTkW^S`q?a1Ut+y0M{OTDntn&F`$!jIx1OWQFV@F$)P5_Mc7$-KcrM4Xeu#A+XL}UugL!Iy -B-XK<>5JLoJQr-!_>1=Dj`P&|mq)t^NO{D%uvM);VqM5te+0Iwm#oIvo-!!7m4J0f)mf-^*>ilPqrWT=c=o_>+{e~&+URn!wYx!@N+f#nXA#qJYK(@xY%uE;LzyXLH -&UfXY=Pvwu;@KCk`4{*i`IrD^IcRU#Nx`b^--z`H0$4c-^?e)!j9}VqJf(DxX+?cgiEy3$Q=2Yp}x&T -iwEEtMUk(r`9il?P~uppY$aGs+i06&**dK$2l|)Z9MnAW9-4FdT-$KOr~FUyzt($Gj?C< -nIXpcncJ=o^7^3q!K_EGv>BF5bLakFt}w6eJu3YRWlWcK?QZHZ3Ex?by2reCEt%S_<3E1?a8_QG{Y2h -N=WRH^WxEb!_Sls1N$|5jcE9V;{(>J*<%C6_TR&3_v}>)`cJH6v{GF@;VMF_R=*HaEy}o+R)ba1{KIR -PxJ<_*aANyLxf{@P3jR%(g?04XU`+E4DU$A?1_m|>vBET5()v?qK``$C%dg6r_W_;XywAU}Q+P-rr`I -R^S==ty?X{*I#W9Q~e-){Hp@>iyA-Y~A&kk=;MQ8V*IiEqWVi^eaPmme$%$>|DRLuORMfX6?ccJF7S# -@Cvj_xFe1ZklAZMBEkJ@#`edUW2~s_d}1Qj3?&odA{x0 -@yo;e>f7a8<3p_NQQO7=HyK~@OtOVu`Rvf@o^S8JJ?rtl!!{jX_`CVVl#}zG++^62a^<~`Ug*Ah!o)M -4Y*C-xe_TG;B{?a))tf*66gVv~Xw->?cMSdfiT1bFcZ_3f|pmo@+S&!A(!Dubb7z>eFU@ub&3Y37j0a^T4KW#w6c^Z^Qm?7p?wHSLx{R5WC~zmc -oyYZw}h=<=3Zdt;`7%!soTE`8DtTm9djQ-Sb$FDKC8ed5>?)pL+gPJ_}?VfAGTPi=|!f`7qCSdcNm-v -Agf+Qy-^q{X@rF%YF~PczOTG=lZrUoAdcQ5017hRIFG0*G~;8E`0V8#!kPsV|Zcd;#F-&?$5bp>yncX -KCK@+W@2{p=Z{}qxvJZ#zzf-tJw|VR;)j?;Z|wd=H~rG&>@PQUpS$qkIf+n%GXMDY-rIg^_0qJ}t+$o -1tZh9FpH_eP#lpOi&vl&m=Khlf>yM3yY5Qj4CkN-O4XpiQ#Vy83Uypp@O8&fwH+}nxFW)An|CASes;4 -)9e(z1DsSn=UarUmu4;Id?8{PZUab+=2d+GN?H+!qBuIkh8rd0lZ`nA!@k6Z3GbhCH29a?l{d+p0!ZM -==?!{$%A)Biw}LdPuMs^)Ju;aI`I3_$6=IGwr-&=X|#rFEl%SRGt -HweJhSAFR}M}2zT7bW$@*z)Uhuv4vU1n+y~kS~doQ$ -S#}7WPsQ&Ee+>eJGYCh%Isuk-$URS<%Ligag7hCw0mjCAP54+{h)^Xpwv^?Wh|L^v$sk3}>Q?#vi>jM -=lZtpU5;8TU(^Czr4JFd^XE7o=28U5ORQ2%Ii|BEw!YMr~g`>17C_N-s&XMbfugV+6iVz*cdd-dsk-? -Zobe+)awjPpPKCVhL|?9!wD->ltV_2HBy{!`w#_27}6e@yScc)LgYF$o?+_e^a6BKt7x>&k@UjU{`A9 -qRMy+4?SdcdlC!b=T4M#RqJ2KPuX=5{LUMHuY -9RBI_hOu4RESOe)YvFe{rOcYwF5}H_&(HL|dd4!eZs6cDukr_X^#3)sRaocav9}$4{Fe?dY-&F;8`tY -HuPxl!&vVa&E#Log`dH$|N9@OaYM=UH(ySI6+HC%O+qPd@J|1OX`F{NQeQSEnd~9S?t6yJVX$ZQx=+L -QAhjyLp65VFnsaKZH{N|T=%Rc$;ncI{5d5oC!-6*~;;s4d;WS=p2oF29H<-UVnedRlAHEhpIuTQPY^;^IGF}5r2_7O{Wj=wzQ%jAi5$LD<5x#Ja|MPtU#K3eg`f`rrhr=EMVk -d+Rv6?1yVN00bl^4w|53~s(_$b)~pP!!bjmC1h&U30#Bo5)@Xel0yV4jXqH{<*qDS6qf`5XzHJC0%;y#K*l~8WDB##EL*mOF%?3S+nUu?GDKl-gN13vW#ez4%lTTUd`CBODf;ven2`#nE#=jktA@SXI1`a^ -eJ*m0n&|Bv-@a*IKb!yHdq%I{xjpdx^27lNdtMs&^gC}|dvWf7 -&Ix@7><^6Iy7F4VkS}MvegB%EQCCCujjjJG>f7oEy)Eqy=|Zku+7|ut3p;k_|Df-&%u?} -lf-UjLfu>k}Cf?|9y{`pM4&JB6jq4&1(JSoIS7!b@qVzWC&$UZ*V0Zhmlkx9^*+$ -~te2Z=Zks;Z=o;9d`~{XZ^7Mn=6KVI4ZnqdGX8LPV$AB%$Qy4Nf?S -JpS9OUDg$RUi0IppLl)o?8EQovdlBDpGbVpsxRO8^q5B+RjJGWoc57p-;v$hM_DtzoVM@wZG+EV3(Pf -5{vhP|TaRs9pLl6w(sx%*%(*ASyVr((C)bOKVgDuX-Z^~g&GXJgJ+pVjq^GMF^?v%dXB{73*}dkwjftv{UCr>=SM!gl$qEq9*oy*x8+=*Gud*R&cmeR#XAtACB_@$ -R&n7W6%~tgviF$*{mTj`vv+_i5hUp0(Yckc&;?(m_#rz&Oa$hbf*dnBo<{lxD$9*DRFjyt^}*8~g{O{jz -Q2t=&hf)!a{5TXDIR|ME1Z@kcdy~@(Kb$NY&$tE`Pt_*=eIkR)3R@$ -=F~W|PPj+*|C;l^vzXJGvz440zVJT7sj>c9PTA4#I5qT%_=fvm$>G# -!e2LRC-4C4BjOzQXT#i()<3ATKC%&h9=jdei**JKkA(d-RFtZhgG#p#120%Hg=H-qz?}FF!Hv)PS7m`i?(cuI-r{eQSqPaa -9laM@PoY-8%JhLiE~!Z~WP>EF*fZPfqFc-&Ubs`H`O- -uW&sIjHoM+7SI~w;o$--W(qN@!0ThJYUR@-Z|=nmxG?XGy0}Ei5_)# -=Ga%qeVZ5E>4D$w>hO; -{qSMxFc_4dsbadvATh0}-{OBjw{;@c~dth|+&P%&v&!BoF4RvV))oVogVLLY1Nphq^J{p@mj^v)Uktb1l9m!wBW{bFm^X~}KTE4PL1^q-Uwz3P11!>6ChiyrmC=JpX&GolmpalgFub!zndJykQlc_@kW5E=bm5 -8d0*Tk@jUZh30U?Qfc*yPVqbLficK==6>;PyGD4IeJRUueV0$OpgwD@$|XgammrciwFNW^ml7?ZvL$P -*ngk}A=*OXK8Lu5JeI%H;yi(Y;eHP`zALDKzj+(@8*d@)kBYvxfLnKX3srY5U}H@VQ?9{mv*D2?Rzt3 --(3+8Lw%uarZs&n)G{ZdBWXUyW8XWnC`~tJpU@kIe6!If&h|d%&Ki>xYvnDrlb3sL+!U<{ki*M#Ld_4xdBLk2{_T_6ovTo~2eV$I68yy~PSyw}_K^D1g&HztGWG0J@yvK$kZNHw8+=1UkiCcny3<4+Ed|HZ!pYtU|E=`6olW`}_vnUj&9X*T3%84Bdy<9PLDxbvuU -zmk4DeH+kpE=4~+^FLShA5&Cd&Tt6MqU3>ZULHtcfYL_^K(9UeLVf9hclK;^x(OU}Emkwsv0QtTeCB0 -a9OC#GgDJ}aIqK1`&~EFOZn5?=Tj%zWku~+sSuZg)2xzh8Tk$j`8Qz(W+*K6MzRE9z7=7jx<~z*%_#k -&SjR`&cy0$fwuld+mw7OP}SE@@gg8w{=BOB|I-C?r1IRCw5UCNQ?m%=onKbb$#u3dRur|(QNuL`bV42 -5(&6s`~7xd-GiH#dJCGRD}AiFJLvrghRp{j<#~!~(Y)!p(D`r?O-j% -y|Wl;+}t^TT;O -v{9J>els+FXxsHHI2C|jL6rOXy-^DL_gn>Hom+cVRyV%?zk2<<mFuKTUWIC9v7RMfW~urc<<}$Fc;x$v_2Gm&nW_T%i~kh{<{^v*M@K_!VTV7B~+chmP~l!cj -M+BP?p>-xNl@+Mi%rTRT9T0r*@kj8-rmEYTVHAsJk1l(lu^w@jPyw#tj`WYTVH7ydKpV7@NmGg^pcjMubz-*S{Q2`U({Wxu -c#=-UwK+1niR?0io7-)iU%0DVst6Z3Nhk8b-9$JehJ(i6w@9QLH$%tmIH?S|31R`b1v}liSyCyuxs)| -Gy}@(pr#4gFgt=S+qzR+EzLYKh`M|#ep3zRDfimEY=@(FdEjpHjd~J4HMz=_Af|>N1j#W)o-K1IM-@e -+gv@WL?ojG&n;NpndMfl&DGo4sj1iBsatU^aYAV>(?(*88i2Wu-!lm~0i6uhIb?F|hLF78HG_n|D(|8 -&)TRl56yTNSvy48`=dZdBzGnePFgHdZ?J2zaI$GwQvVk=536n`8ayduv%AX7mpSx6_M`C@4oZO -Y)zA_}+tyyMaUS2+_A!4nTW3(e{uUK=YA`Qk@rEVL_t`jelB4D|s$3GPVMKnZOh2CsWta_bp3f4dtJUG0t-h9*g)y=vGkSd -OJN^^>o(E7Ry=z@=|H>bEdXk(0#B(UoK%#L(dop7b$nWt<^F2%V=&K&6@*L$|>ju!)Co=f-p3*(dund -=oGAxHO41>J-as4~<8Z{d73BTMAMDMQW`4jq^3gy*}hZm(R%d6W&&^i1bqBzgO{q#g92A%r1Xe3_h1k -mdQXSg!B-|m9@sqvikgWCiq1Aq9vNZ}xi-j`(T)JUGHQW!gu;>=SosE5H^ze1i;h(C~RpgYI>GS -R-y|14vppJUAPL*X9{dxcft4 -a`g*-SyqH0c9Y+N4<>hzpA+HXz%Gc;WlFKCGCIcSlf;-|-(bPW;2ER -!3dr;t5NH>~?b=S-4O33prNEhVK-i0>&?s_qTe;ockRQ2Y6`u-|69tS^Utm8*eZWKmao@F0HegD|WL+ -f7$ewSVRTQGk$A8PsRhs+%i_eh&#zYLXUEGq~;eQ&k*A+Dovm1pzgzz=>~iDzOH;XKbk9?v+#1z`R(X -|Fv8kG*rwwmlfb-0QA`nqL?z`W4zL@#(M0<96_i`^{OWharz;?mX@T|3USj3*r&X({Yz&#_#ZK|J|9+ -At>k7F7+My0{gHcZaTTclW3~y3}U#@P|Aw^@o99375 -s$zB0j0JWCI$X&i$-4!&c^U_~OwTqU(*S6TeHv+tmj^2>bqMJEBcU%CU- -QHLpV!GLscn0>9}LW*z9eyXg1a+A8+FFQzJN#uSAWQP>M`&!#$Ykf#1^1-=K9 -oqFh_@t;0TVMpL6`~0T*k`pEEK=_pJ`uv>dfVv}J}#~MjM~22V|{U{&q -v>e_i6BB9}9}(b3N&|8I7eCcxeE4|7Q69pj^OL2nO#G{_v`(z&l6ns9N#tLiY@!mxiF&3B8bGuQ(Y{276CF=9m1r)}`-pBL`WDfBM2`}!C0b9 -kd4`}NL?elgBAQAxo9KL^ONl;CbQ94}h<;AAmgrAJ|0LQnQ&0oZNTQ>NCK632Y9o3d(d9(f5q+NMJ48 -PxdXcCvl}|^akwnLuopl$lP%4kUL`R$4{SWHKnH?!OnOs;9Z_Ub|#?E@Cm>seCR(pP~Io4szjT0*ytg -jci$2*sF#^@cV1PqSp2W3x>*wpkg=#h3_8 -O|vkb3F9T0?2a+GI0rS&YnsX8fKVwG>#SU}NOVjQTf$;<^X+C5a02GcMP47ov4u99+3FZ=u|br4Tk$m -femxdf)6A;)F?hy=>ru3I6oN3snS>Hy&QgmD%xnZAj1ZfjX;#m#oiHXjaZJL%0ev%bbMYyFnFfWXb7i -teI8HX_<xYEMJWcXf_SP^Fm_nuh%wv$V;wlIm`Yf1;tOWq}%WyjCop3jVS_ -My{w~y{bDjRZ9LXdO`D5hT~ULWa#Fqp6@do_UK62sZvad|!dYBeastkyJ)rmsgrej4Jy#~XvuW+bSTj -=!ELL8t*wbFAYB`nRy;@Ei5vk?00TE~Yu17+?DU^%RR=W_0rXfEPT)7bqU@ph2JlF8d0^c8F?>Lw;OiLo`vebShTc)y!#ED8HJcFjw6G^zP&@ -57>a3wExN3bC-IzsO6nF|Qyj8yAnMCYuEp&8D%fI_Hs%sC~vZLX9r8XQSAo3bV4zHai+`Im>ZUp`)PC -fwCT;zDDG5qU>iPdC+W3Sr-Ds@rsXli`hbZ{&BAz2JB3&SM5P^VsZT*P90{I(+nG%@xM_ --z{R+QNP`M>3jAq0j)$jO~)3rfwlxRtu145K=tilYy-3gs3*=D2ZI -PeBY_%#CIO8ES^_i*s38R6qm59guqP@A#l4KqCvd --*L-j%TyirHg(Hw<6Hg1aL%DJ9*}wmDrn3iOJSr90Szv8N#Z_HY<$VYrk_w`&qFG^SVm6Rj;*@BaR1Q?s`u~1wpN&VtsoVGbpYMC_s^!bxd%bJD?|O&zuC?}ATfd+1AkQw -U5Xpe&aNnn!em@LwJo>Ywn@)$hyAW1h1Ns0a_29cFQCevap8Xl6dGzF&n^9VLPrbf`o;(wBtjXcSJ;w -^+)r)6}R)`Y7xxf$V%`-WUYei1TeSj2GTm$~1rXNph^y3+CzynM~`AR>XE*{iL=mXdUa^%ZrcWK8p>j -0h+7o}AW(CM)Ru)iuqVgS!(h&ELN&OrHuK(wcg)hY(UFKegV?1P1P0{2UUc}8c9DLF)^hX-UH!u?9X6 -S&U<4%6A*BSM@(eL+Kc=4gc|2W+WTnmmT_xA4ccV!%6xqn+VAlP5|mAI`J-#%k6P&@bGt96>sdXx2xe -2e`J51gj$;N5DG)d2rex+)u>hx*E`PG|yfA3Wf9~6J`?^b|IFbUau6|>sZYvmGvLftZ6(;4EB^x -RfT>4u1C4Z44zS3p?PFdf8YnQ`0kNOz2CBd@^$ETQVrPTQIqhl_`S35 -e&rP`p_#Ch?>_7Ys+AH+?4f+%G(?*`{5vADyS8B&iKE*uS6!x~6_6T`zrhLb09&ZW}i+X&v2$2A2+e& -*lriHzw`yGc9@?g49+W9u>Kl*>0-d@;twu64$&gV`=8{!Iz1JDVQ9zmL#KQJXW+V;*KQ}>5RW!Gj-l!UVv%D4+BN^>v&n@{kbU!~cYRh4xLS+^leIM?iPjQS=weB_4x*H#pV6r&5St#e|! -DPU(EvPNQGR_i54vdcUDwL4Ur{+ckenc^=nFzSaFf_!;m4ziK;6zcyB@I?J;UqqLwZ+Dn|9-}fvxRxA -3R@*As}f6!sl548Vdn&*$~_v4!5N4>oL9PQzl7Ji=fpUP7q^YQy`aM>Q{f -Xt_4}ap>K~b9D&q9QwTq$5f`0uh`&_B5Ll(g$(rgBMnMQZLs6Aso3=+1GO@4evKqG_TM&U=;5x3nc6W^Qe#bc -p?&Kun$nvRC-i8diD_==-=-S;)d^F$hbG3OT+;0t{9ue*(4CsNUz^|-dzU7{ao_S6O(Xz%-K~kPrtzl -mdo-~G*Rl6%N-iaU=RtSzeVUS|=l#52A*_I3qF-YFs^~X`wa{RHF{TQ@x4^I64E|70X)8@EMSq61LBH -X;s;ws0vpp~L8(?tT(11byy3HQ|Z-$-Oj;cgIYLeKgTli#8SY)s!Ex_0xo%&k^mf2P_Q$J*+QK6Xk%8fg1D~{XGc%4 -u9%5ScBe;)ye>eahwQZJyDuZFyxN%RS)An>bHezD(-L}s)+`pK^GwXS=CU=^SG82hJFTpWq>HMAc?|uE`&Eyz{t7KA9R2t>^z%v71DO0bP51$ -pMrh(uz@RAbiSm(fomGGp(Yic+Vl?`FQx%|<uKevf -EjfTbu`J%f4&{$>(l9Gt~|Ij&XD23=#I_s?kJTa-)u2jmWDS*|Jn7X)~h20!$o4il0{&k@bC0{o)?i~ -gxY%S#%@#~4#J-~*`7mQ4Acbn{%PsW_t;FbMa{)<91Hi(b>j48Yj68vN2&lcyd1ifeP4hWG{kF-_N(O -2But<2My$=yoVF+0JpT6tExdK1&n%@aM%2^f%kh(Nr8Faw#9gC4kYOHyQ9rK+koW@-L2cw4Z2G@*A2s -#rSW%rs50B22B*Ro{iwY!PlmuH#HS!mTrOm;XdC>>4j^)Oa1`jO}-}{&Upr49w6TxO-B5;%nA99abLt -2eCMzm#&N#qM96nl`8>#ZJ;@1^oN&GK`NvLp&Uv5jrSaOw3B#R`?_ZG~zIT)i`SJaj>M{T6Zo&W7)Ig -m>Y7xT5V;K?vxd@je)O3tz0^2Sx-$<>&m*75otQ?T|82W5H?*y*h#&_7^I&EBF@B!p+0^lm@)T60iY{ -u4epXW;3>bbApZLQ}%Q_=4h#Pu5`SGZqxh5Hp(xLO#1ItDqWuYY -=BL;&=a{4zTRa8&V4(4MZ-(i)?Iks_R_WIXC#0E9z0k)_Sj=0E-p^YoHaKp3k0J<^C;@<-5-7+56n9i60YFqhBvz`91q$V)nh~yspuoPs05gUZegRJZD0g1$zDZ^`wic{G -L7g-qmkhRP`Tb`}#H7m-2;udq^7Yk-nOJKOM%^1wL2wFRFQdPtBeStb^CT4SymS1!^Xteq1-)y8F%!t -M}{!3o+~;_wxLOcbwJN0IPZZ>)#)2He}B}lvu{@a<9pLwX2U_KwVew-Iv}fZ;vX=jQijE_m$N^SNo!% -|HHIL_f*&Hvz4gw`=tDJu8=inIafU2E93>p(k3)yH{?dhO_5tj>bZwg*RQ4L-MV!Xp`oE-#E22%>8GE*Ovj -Uwl0;5Uj#$5bz0&bdKmAmkIB}v@&#n9^73QASqD_q`iD#_O$c-`E#e(f;x5rGJDi%OQG25kFu>G3|Xu -`w=-<^-xh(8v*c4CZ_zr@VjePZOuk=tX2bjt;dm>4o`sFWKbC`U!u3e-RuUtom-) -uwi1yDV#Nya$}6vkl`B_@)vH$vyWK7_GBT8Wi;9ZGh7B8(FWb6xt2 -pp>ky!M-6dS6fc;}sW#O~d@#ooPp#s2;K#m66iEDjz#D8BsiOL6u@xj68H6i1F6QF?Le)G2Z9>}hfAC -n>6`s>H>M7nQCsnY=NAhY4{Ll*S!{_RY|nTVU+B!GO*3cF|9E6=UQuF-J}l$#R9*AlHeH<$lATiaUYd -3iuY__XGYA;Ex9W1mGvgX5x9^uLgbr@JoRI5%7<@;I{+5AMhUs{tVzR2Yxp2cLIMu@XLY!74VM%{}k} -g0{?;wzIQu}Z~oBrF&JlNU>sjA#J+5dw{OD;KWZjqc~>F7946%PX+oY}A!OA$AuqV#-wFIyz()X+)F1 -evz(-wGaloGg{O5tc2Ka@*-vRuOfPcaT-%bC1x)s4nQe)uX41D;XDg;B)e7lhSf&T>Xp9cObz~2b`k6 -rM^Z6ZwE4V!I=n5h%YwujgU&9B0YLNwQbX?RlC8pPyGjk1P2EOh6MQf`UggI_G;h0ZJWV^TN^ -b52IDdQ55^yXff1cLu)yHKX6N&e(13vy`^ZA{3-2bqXv>_eY0`Mbz{X-uK2^km?(VP|B<9YYpe`(UB -$$dlE0-uMfg!-FbRv?1iz~=!2LxV#?LI;MnxU2x4x9_Ox9}pTCGB7kSbkJX$-{ma8=N-H{4-tC(0jde -3*CP6|fV2Np{ayQnqKiWZhJ-36R0~G`gZ_xXu6`sT1a&kYH0X{y?m&Us=Mg<_itzIuhz1_P^S=%nq)5 -20_IW_lhE08LvkVLg3`H*wYN438-}o?sZ*PD92i=-BwD?AZMuY|rbW&-2cS!^XbQ=)h-9Wqf5$6N;Rd -Aqx?emCU-X4u_(zIqlppYbnhWJ8}EI`V>kEPgYht60Cc%qw}604&m$rR_z&%KTcd_O@Mk~-Y -Yz^*)P6(&iwx~^tErJ6p9lCtU5(GA8T|tV^Sh%?< -ck5i0zZus!wDju01J-9aZ$V%uh65|GnV_i%{SjvvBD=MSz;C9jUvPfyLRmoAAb0uiVcn*JuHqMJt~eLKQ6xc<{Rc$v_XBO)k`J==PFd-K@u=H}+jJ)5`e*uvws+x`kYyQh^`+qUi7dN=cE+Tm7QH*eXxeS4H`=F!5tlXoX -^U#pH@-uK_^-3$e8yY<$;-q*6Nm-j6;H?V=?+}o()#AcjXS})w0sy%#i1>(sYzUvDQ -*p3Zvsxl{JTjr1R7&4Ky$cJ12nbD!+1+pdmMn>@#i8ROdn34h=F*=L`aOU%C|pMwX9arVo~%F0vV^~gsbee~Ln9Xl2b7%+h66j4t6j2Vx~Zsw$ga=qD^`S-2PXZ(B~KW -%XL(;HCw`tR+c?j7CE?&HNsMTr>T)upH7~*Qu#!n)5y8G_CA6mL}>EjPR_#pev -GcYjFi}l_p{XhQrqeLvTqIK)m-O-23PoF+5OG-*=`+48Kee(0qKUZ`@K8kk!T~SdXzxd({bxod(jF?A -3=-j#QzyH1(diM+Jto-@spXCof{P5!^pM0|U*s)`W(bi8a2R{D=d+QDwmJ|0z=|?=pwrVr!H!^uJGG$ -_9(r@r#{B7jAyciiY*JjAPHoM+u{v-6i{r20-AjekFi6!J==gysqe(J};g9nuijvP6nu1Og&w4)v~lm5MX_ -sT6>wn)faB4$=L%>Ji5cJJP;GHEX@EmgKsUS6*LCJ*!h)Hm7*WklLGekW!0K`DJdl(O4yDZSs3veOPJ -yOc;d{Cz2NPDpwF!i58{mHVzqxeNWZ_56wuf9^5f3wf -o-=vp%Kwd~U`=2sS{Z`7rPo?zxNXkC@r0lsz%7;q9!!9X1f`@i*N!jX6DVHCTvZkhH^|eXAL_d6eZ6^ -IjHe_MsKVm1e1M16TUr9N*T*`n?(EsECGU)Mvln;W3DW6Gsp+?G>{Zd-C)ox8(7yTD6UZgHBd+4EuT0 -s|Je$Ij}P`c`O5Fr -3lHp+(rf#T(+^*~qHWu@-64nNv@`ma`uL!(UrE2%|9sAIfq3*Yu4&Y9*eH4UQp#ZPFc3Txosrj*zDIu -bO7#Eq(@zrP%w?H87}?;XKAF_V$1lJ9BKL1zBQxfP$XBQN%9p45DIy;KTFQ~fq#RZuWhi<0+-c9ynZA -ZS@7C>^KB?U{-JUyd)omIcch8mRhn$yn>(;Fm{P_~nPd%y22lbwMU6&7o7wS6o=)5QLOt`q=og-->OxsTK&fu&vE{D@}T{4J -TPQId*HRf1L=p~lO?Z=m$`HL%FNmQWGZ-A10I;7;6I-L58f0AN}z6O8;SN%NY;R<_tazUdRK-gDdG4>8GFEu+U!?Jkv+!fQJn5kTOH3KmM$gk>7xaN+}; -Zq1!WkQpgvV?3p|m_DmiOd!|qF+M=qtSeJg<8OImIS+$uw7`ZMV_1TEQ3)7|rz2xh2d&`2S`^fAB@Gz -^NJYRKMe)hA@Lj-sja}qqb*mH=}p8XH#KFMdVK0X^hsXcgTy`^5ir07R{zpQ)r?yV4`EFt~$i-gn}^1 -ztK;Dz@%)*5jcpK-ip%)s=ft(V+5zn5GG9`eD113YY9IaQuJd+NHJ%_;CX88Y!P?AaM(wBJ@g{qz@x{ -`c@Sy193T_p -0_5Te0dm3kK>5PV$K_X_l_?%vV~j4&7~@Lx!{00Y2R+M4H|ZiIO{AOlK)MZ?(8s&R#~gpjBlZ97#QyR -v@K6jM*3SbEum##Pc}R8gu+qsx5_x!PfP5|%Jj4Xb`QryF9+>Epdb`9J-QTHS|LHS0hB1y{b~qd|H#b -)@E;6#KOs-}pzyA8GEPE|V?pWMYZe0i-U<+@6heD@4(@U|$?=O?BzVf9he)7dh{_+Lz@GN*(1R -myr2d26)#@`Oq?|(=?bV2Dq^m++-dG^_7WnNyM;^%t!FnA$u<^DJ1&Lur%33%879+W+U2l^z#p6&2St -EcWU*KUmcz7N>{2h9}(Z#F_u}nbg7 -&*YgR1}SL5SqGGU@mT0j3f>=``N<7cjv?qNSAB_-{E@09J^x0ewS5i%nqLvmb>kB^r#X3UVwmMv3u{p -FWmD!;>6s6JjuH^&D0D#k*LM>!v0qCM}+oGEL5#(X1fR;@i7F~&-#Jtx5yu0{XWty`NS)_K&RzjNo#J -z`>Fri>py{u}sFC5P11R5^6$P&s162>FkH{6oI{^2_q2mtIozaDGpE$pgm%juQqSoWqa@`cvv5$3nKT ->xDr1Q5NaHAm2})bt%S}rNsRgCb?JMSocAnl};GBWhakO%pr?s1;OxQ?=*JoEq$=iGS#a? -`&i$S*$FC@-8pe*p5n?)8V{F1(k1dhyC9#WjeOMQg=-KTh -pG1*qiHK#kKjC#_ZTBlcI2g=1onUWihXpW>pw(>?b@wY>)~gfdFD6L`PgHRsZ5(Tc<9)%qntf^wvq$K -3}bD_;Dhv%eq+s-a|1%!18s!%P}kl$X1mTkIDZy5xc)}lqW(jdQWh>;_%ZD?BqT)6ojX^_fjp4@0Rsl -e@bGYTpE4j1MkWu&yq9w^#@Ez)(n&}iB5$s`#h8la;Y({S;k#mSYLO*qrnYzUFX5J%RTyv -f|bEdN2fPet$=jSK8bm^kxz~{uN%SU~>W!Q!RsYmn&)GPX9%FbBd5bDU?>)yXm^n>qZJ$m$Lg+5$Ddw -%P!w-gWTAFdn1J`@j8QBiWpkReL9($muw{p@S@IsYczv;+Dd-e+t|{b8(4JD?5FCRi_hFlhe`^(^u9^ -z?(=Q+Ms!RVnm6pGMb9KihSUe~f-*=6sR#K(-RG1>Y)G_5gWH=$q03j#2Dm1IJiTq#tFU(*M8y`s=cw -pg@g*g@uI*kA01Ps)TQQ5@Y0C(7qG*o)h}M$Mx>tsQ8WX5B)#vaXGJPZ-k_U@}|u~PgN{2dGciCTVd; -Je50?Sej9p7JD^=sCe#zkfIitZ9^&7umolO}IG%%Fo;*Rn+3;FK*QMW8S6y}3RkukS`JxWd7M^?VISH -So#%b3?+Kfyb`WgBnCh8C8M9id}e63x(mTQc5lgV_WwAa#a#Q*5;CFH^Ii}mrr{-zJY*sbOX)G3a&^{ -nYD`Y9VDhTy&q^`5%KKBo;O^JV@U@Au2-ch!ITh`M8JU0w*e#{R?;PsmqZc}3~BYrh+r<)};4bzXA}A -nnHf2;$l`{{iiae%JXg#-8QWOO8S0!&vX9|Ddnqc;w0l`*!2Tjq=r3Usd{S*f2BsHrD1DH}RS}W9%Qm -XJ)y%x&0@!*V6AgF1p4ruCdM4WWsE>+Z8_fVtMjLdZ}CVo#e%!9r0=AAJq5X(N9`nUrPTmelIcRUsvP -BSo1XY?^FyMeqlY2h4}{O>YuJ44j4?MqD^RP -atozGqDQ6v(GU!u%kBw`=+*5e)Z7HAlP|8n#P_$f)2J(7e!-freVs1ALHeF4BhM4m_{Ou|DrElSv4kA -8JMQr;NVqpG;f!`l%g?QuagyZ@?8P`I&&)3JfmizIKQtta+O5a_o{qL^Q{{8UbX=to%pNsn|TodP7>?h}>Ec@ZQ_Ig{;PcK@uDC(t`UK&Lj7&~$P!10lOonsMwH -Dg)EDSXDinZ{IJ+AriD5BEE{rp^5wuE}$c?ON?~4uCo66wDW%GUoT3=dxYK(d-jr&cnFLHF3Y}@ss*~ -ANR`ESJk(4aalTBM;}!CuJ1n7edmFF!tv|OiGwn0@hG*9x!k!1IQEQ`5vTQi(-BAYeP84K0atzG{t)* -W`vQJ)xlXp=n1cCh6yn@buJ2(`{)~Z13&)==D<`WpFs=u4P2Omq`ERH7eNXPUDSa&0_u9C}Lw)2P756 -&1PdNGmg<5K~&v6{{{VDV2&3npqzuSms2^m`#%nwj&iCoX+I-+a)+oxZbPya0CbkNdozrNQ-o8W$@Yx|5V=FFKhWx;|4Pr2^Bk_N7uurHXnzRta- -eCK{2_nEjJN`2%W6ZbY2JNK97j??qYi{cawN*~KE?S0p6zwTae-92E+f5&saYQKwnMO@S5+PI;QTo>p -53il?|9+$KIdE*DEwg$WCW4}F@%bzjX(xpqIRzcD|f9Wv&3-Dh##yJJ6HYtQJw&O)`v!o9jBpkcmquhAv_&i}Ff7 -hZTF>b2Ki8^ygL#^a>Jm{%L`4U!J_1^p49F>(L4;OR>?QJ+4#=vux0pE35_xpSvH_uO-n*$*$i_@eTA -^t&AU=>Mq;><{W9+jiY^`tIn*^32gsVH0)qagu-SUfI>!=a|BE4zB64k2$}BE!M83*WC*S^W41G+UGf2MvZ>#*~_#r8;5dZY=-~S;$KfgKTWB&a4Q^$@S`<}&OkrO9QRO`UYmoHaq7s -lL&eNM>!H{NYzf7Bh5nOG;veFWXQYW&x^bLWw;do>T_SkHYK?kyqCl4HhUa7BV~@np0b>o>&%tw_67YmeJ1JLzJ2?P;`;U|l>IMTwk#?oC1sQ`=cb%V!-NSFWT#G@Ay~KCa1g%%{)c-olelKB@MYX%{TZ#6B?Aco -{EqPla(icvyV{ZP&J6cMZ#RO@iam^y$-;Z{WMC#=0Wq#l9xr^vMSf9LT)EwpIIV2fjx6|N7`KV&HY_* -2y(%)~K~&`bMrf8GDLJs@iVsP4?u`fv)sAJsY;QG_1O`DAO1pjE;wftYt8Yuk?eIMmq -R8&O$JA=MhD)jY}KbrVk6OTWbX`HFO6JZy2 -LXVAxhwQJX|Pfbm|03Y>TMn=XzGcz+U^E-KYd1q0_)SR3gbsuFKq@|?=+wJyk`T6-up2iqPIm4zub;a -j?+KwGNxX$_u>ZpMJoueIcZcKeGz#L@67y5fpzKE|^9aU9rGVocJI?Wi)h-VC&qA!a$rO%ldpMHd}rr -Ud$#m9nUFx#T78FM1nSTO30J_p|P!|#US^T-?Nn`lR*)%bl!+17NqVyYe+j&shPIDh2)m2(BoRVP2c>Y -ujz0dfpa$YALYa`$an{J%d-Pjyv2Bru^wYh#=4wufoFs2y{?>qeqWc``q+l?1ze7DRQH&pPEl -Hc4yqqc?0K8^P&fPy{Nylh92iu>n{tBtwI><0E9#s%fKxF^T=2aK -_lY2Ez(@^r-k)t{V0{XI^#f6kCUb&tNDdP-RiA3j|5Dg88M`PI%;`TC;4>iy3(2@lKn-^?QZ1_OWFwr -wg_Gv3Fc-)783dC;G_(!#$vmb>ad%Nz1H-brFz94Fa6$0N=YSU3GL#{LDa`1I4HoAt0SxK~7;IA(J_i -MX6+bL`%=Yu9DCO!FJ%`FkJ>H)j@C$wrw*&fJvuHPJ|z%^vEsC!WBj@yiwJis!GH1iweYlLL7Anw#L) -D|p4T*BS_mMfXlTd99(dq*bi7NwT}2m_Eedv1_bh@Z;j$lQsPa=#<0KajXN!kWZx{TXpst@1txkuN`p1ES>z5fPK8=5e -U$-~w>&rl;kE2lE^m;Umi|`CL9>r7YQ9PbUT_*R&4T>7{^-Y*O$vi81dVJi}DgD}a?P6(fj-E1Z>iD= -R6Z*9u`RJg|KJCr%Goq%9kD4@fN_4;WbE4zh`}V)}=Dtz!@zImVPMTvz1ykbtwVyeCO7Hk_vC)&G;yX -`{8#jGw{M48koySd`+&e0Ma+g_M+nXmxO^J(%j-N5gwKY^_Hus$|eP+D+Sf{hLR@bQQp;ihHnDNo$W= -@ZrF~@n0JJX}5%|t8F2_Mz+JfeAq6=lzC6&x)R8p2UCqwU3Gqo-%9d#OUem%`@Zt$ -MGw5{o2PwO^T0hZ}z^_i1(FS>Fa$NO?|yHlsj!w80fq9nB}wKBCTwK`R#xu==a -ywWUbVQJxMk!jX6TUt_@J*_CMB&{s1BCRS-q0#-S>DF{xdQ!SQy(qmTy)3;Vy((Q~cx0F -}EE#?oVHuGb)(l%lQieUFD5E5!ETbZ$Dnn#?WSTQAnSPmJnUR^+Oj~ABraiMLvm~=Dvm&!9Q)GE$nX@ -ceepz8zky+L(TUJt*J*z0IB&#f|BC9G31E<66usHl2VU9?L)nRiaIqZ%iM~S1%QQ@d^h-{B+bG9YhFF -PzdGTWMM%TCI+XBTCcWS3=EWLIU29FH7xjwQ!0CoCs2N2r5d#{=-nnv;-Y%Sp^h%1O?#=QwhTb4qebb -INkcb1HHwbERgfMp68M0nP<-P%CqG8vAgmy~AhIB~z*>+{U@J&0NGeD!uop -NAiVBJgN(xE~$_mO0DhetKstT$LM4@}3N1sgoJ3Jj;4j)I5Bis?|NN^-Nk`X)?J4zkpj!H+h!#&$G+bi2AJ19FmJ2pEZJ25*s+mT(IU7B5 -0@PQVD=f>tHVSywmR1xQuTsFgYv -@jV)GL667!NF+hWMI95SuWbB8p&Aju#|F%}X`g!CMcTq&ehnO~jn4taS&RzZ+cEM$}j`8Xh(QplweGI -57IydaAp$RV~cp)j#9xzJHqTv%FIURYUJP1$&uL?VE@-P7)6_pt}r!|k#51bd=A+3v6x+e_``_DXxT- -95!K#Vf@pB`764B{n4?B{3yA#gS41{i#s8;{m;~Kxe|BFIMPE67-}9I#LGxsDf^IKrbxNiLk=RLTjO| -Fsaa9SX5Y2SXNk3SVcO_phJDE(8F%FTkL-JFngrkYPZ>wFnunvm)Ohf74|B-NbyK9r&v<_Qo>RqQ>-b -rl%y1UN>NHlN?A%pN>z$T^++|RT2lQ|!%`ztt*N%uq*QxqQ7U~`1$>uC^ML2Fr1`;f{onUnR`{tT_^2 -Xys4{q`svME)k!yx;^2-g&jm)*?+H#X}?YTv{CGa>E@HRX^nrkKHC|i_^KddGZ1dxbctW0%>&kstAg% -uaUic4U@B0UVY5ee(C!ai)UkR)XzMX-_**hv{Ir6T+P%c%9=^e3hN2T)4`1QY-O00;of09jjHWdj+L% -m4rYrU3vO0001RX>c!Jc4cm4Z*nhWX>)XJX<{#QHZ(3}cxB|hd3Y36);L_fB~2EpSsKD7tpo%j8YV7@ -4Vog|QY{@p0c8{v4MtSdOsHlV5fdvtljhnE&Ww(;>bSh4vv02iP{;yVSOY2sP>iBbO#uePECf>DIp@=#^Vup04*U>$zXg|-Ch@0lT46r(%j1>eUzfgL@F;wFtworE53gB-Gqd4) -twqRB?F+G0SONfK+bqJ!|GF}L|5=1;>v!i~t313yf$u9%ngHS0?H88u2*Sc^OBX$;JSYghT;8T|Gkl% -7uv|X;zm^vhe3HOtY%q3Wxwu^QwI#eDjSXp!1o=OEBEk{r=Z{*L}r -1nh4A_%<)(f2O#J`zU$`msM38+Fh|G2sMhY?tBQrd5`El7P2r_g0A;%)fIa}mNMP@dn1lhI?IIn?dL5 -43n&5)zP%xpR0%k*gbCklclyP-_f2We4xp&WI$rdb3{93TkmaUBfz+OGv+N8OGc-_fU?`poImr~NCG}zGIM2CC$ss`zq-ZqKwWP?8FFM*rf04+N17|$E!`utI-hx9N4Un{AFM}~-ER -a9B$OvSemu=m#AH!5`1oB{K+qFLpnRskZr#%UPY!=VY_bAgd41Ahn#~ -G?CCzI`0U3`4RLYiR=7AvDYyWl0%L{_LW|Ytwn!9^S2kq@gcOVeK!8$7C#>rZ*gNvU(hE*qjyB-XBma -wPfLf3>l{^(T`c9Y28-H4%3qTISX-herA4vZ3(dZvIA@>JU%T4*kS1b>3(T}G=Bl0D#x -L@cO#-^=fEVKir#ZAV$(wH^>mn*=%nP?1cNe3T911JK$h9QoF*(YUXCou^ce2zkHg-s7Aq=XIyd81(4 -PUsy8?la&dFSi+i0G4-2_}4d}Z*;TC|g}gTn_ychS3^$(s+dOPl{l$L2%Bb8s(Y;>P=+A8fpnu<-z}P -`&P;PN^O3HG~@W$gGaf@18h<};)WX{@SF~? -m%_KSel4`Rp&xDS?*j}~)Qh@hJhqiT8@IO1_8ac^2|i_v+G!0r+UsngVZ0>U?qk)g88_>FvN3jCpC&x -_VH>qTE@=MmG24=(>6&2}+Mp3sU1puy^TV+56=|$do8ZCJ7TqqZ+hL`4YsU%QFCYm!WCrAH9#DnFE;& -*Hgkgn;lLhQT8|cCsjoOb4+>4!>72l(Bwicp<7VVTs2@ADFl%UC~VJHJ-Z?{-69blKeonpl;_@H)E`e -$Orb(G#BR*a+cPD*bOD~3{fomkNy9@e)*dxram$Kdf>cnUJja4*5PGxXMKxbHuPu(fGViX6scurag;M -H9eJ#v|^v2fXJMN5K1O$d$9#Z1YB_SIABVyyr0I6UzCPaz;|l5!u_LIArf3$~i+h)?-MPptTc)q7SMb -hc76163A|QVfchHGN7I}lp(Mi4R?D?5E7j=!AoKIODu4X3hV&<C501DZ{Ldh~JwLzX9XRAl1jVvyX#-kg)0A)(ELn^PphJ5uMqoTdCO<(5 -&_c-qJNtr#$_xAc~6Ah&SGd;U2Hy!nq%X;;@#>phC`GBm@hAu`bqrT8z2yX1HfU!4Cg#gd{--J!8)DY -|TUyzz4(a*Gx14j5XXhX(o+t(N>&8TJ9dp`xr<-?baURsn_w;a~+tvT?_J5iKiZcRB^)-zkz)c`nw!? -rb>=108Z$5Xodx_tO6e -OY(NTf_WB}0q-w!T&w7-94fn*OIB!%8fZO4)0KRU?Y0~z#Bh>B>Zbx|_eG7a69v%^dV)dEV?WKy$=Ds -b5_bQLeMK=SNc^#(Gu*=?>(tCk|{7*NIR2gd5b_~I0@6QNEGkaje!RA;xS!UhH=2WksWWXnGRu%R~8q -Azr(`D0M8HeE4Ic2rM2^C5|LXMOXjJJA8V=Tnm$m@Tc(u9>mOYIpeH!45W0v6_$m?hE`q@w)h&-hK^D7{oartLSri+B2dtjg3 -%_-?IsC0z`C$GzJkJXTLd?{^m5>+M%#3!zs;VcHKpH-}OU8Rhb*gkFjAX>v=Q1=Tvf-u8rxV5A4I-wn -M_@I&+upN?FM{Ku1kTpR;sVa>c-mMk&i*xQhkRGlMvIdl^R^Y~sF3%9x^~Wv;U}wQWrq~D#+jyO?pbp -f~0$5zFx0=P%u=pqrn5Eo^KGZRBT*kJt#1ETb2$j%S%_pvb+=}MK)?=Z)536-3I#r>W -7g6Znkl5aQL)gCe)9{Eg?Evv@DBu_s1qA_9U3T(D63}4A -otHI&V%Jp0L5VuKh&Od(RB1fTq(}GDWm~zjiC{|X5+fXBB+fEPP2EbE56_ju;9tG;85dAs_{Fs#`pTx -NWV-Cq#5~0$AB2*YoB$SbM~^$#PwE^|4*T~u_`m4?JA`C_=7ngA3(-NLgi?BsulSdXcZr7YaO(~RsnZ -A;YXqnT?iAi0cS^BI9iA&814pG+UV>-Adk<zSU|K4kgYIk_ZjYI4y@3PSpwAI8@dM --+V*Fbc>Ga0$z1}jl7-D;v!XTJDSJy3{UZ7@4PP;-MYhuOhsy2XYx33k^9FtI@t%7c+bpbzJz@I8MgF -jA7MwWoo>?zr5);$m@h1?m5?H@#*IC_@9}ni^MeM~z{Gtl!khHBGMA53Vw0;a*Zc7gZ7&d-Jnau08v6 -PDpGk26qKwlfNt4M~{4fx0EKo_biqfW_0Kzeh;hFr@IeJ5)JpazR;5qiI3Z>=&wR%>7%Sa2VF4G -AH;%xda#s_VEuFbif?b_OH)8NrC`nSHmZZ*+l}6Q4`Bwqbl?E|&rvYNMggzyjt(eg-l{<}9rh@Nq4!{ -Lc`ImLS_j8sQwI$-A|2pmj1(nm>lQxNV|Y(hpk*=rt-R_mkd(+SW}p7eQ7741dl|7LboYPCi6?ZK!q~ -UnIkQ7o-3Q{`hK`9$P|h;`0bHRz8_y^0n?85=?qT9waDk?fBKC9V)5X7w`njmmhJWu5@7=unP531{B; -IdR7Y;AjKORvOY{j?wCQrk|xg-2FzZX=-Bi(y{!K|A7ld=6})y+O7=&vDBQFvH7X4wN67;~(2Q{xoy^q_)%egINa -(#C2(dK<=Yab{Od^=#BK4`K~jtnyRXS}9K9K^us -;%;D{m(rn6Cdx>&e6+Bc;;5Nj!p(qz}}(1CV$qk%*fW%>`smjN1&Njt7TZggU+_nB(X7S%i&)xonzTg -$8cj0g8OEPA+QPfI@V{@dGHLjW0x`cKBxy7YSuXX90e1kBpjwnW;Q;yqTF?H5@7IGEySH83u`w!4_Gq -%?8Tll#T5mD=RC| -o3t=B8HCSOijDS&%tp|*z(EHFHZ7JeYkIWK;vZ?cOU`-T<)>TNe{TwPTob!B8XVx#GVX$Hu;K9x(d;@pBRg&g`gA|=x_BQ4)U -LtL3{q=c+L1dis&RXw#-FeTu9JIx27<7y^acQDDuL6ue1f>)2T&_(TOXJ@Viy4v`|&K?hc(&I+#+ZWy -G<82lxB#~ckD6D#14(I(|KRGpVD#%D6lp)umwVA0v#=6P^+oikw4coDCNu1C1@dE0Dqxe(4+H%k%xg1 -I1Y?JA$zMF@uShE5NXwElj9B1?!2m(d`{Xl^E1tt%MeK$t};Rj@p4R{j -dgWWYAg%gjqYc>c-MXi8aw`x6CLX#7Dlc%pl_85c75m0$R;+0OsD;bGbG4c`8m2yI@ATN9?pe1NWzAV?0KwWZq7i#s(C(BVu1kD>%;Hox*{=xdSwrSM>ot -)K*vT-&D-pYiA -7S#3HCqc3X;&s4eAI;iEXa%{ihs%cA(es8;wkk7MxEvt|$_<1DHg@dKWbXTf&SBK!xq)cG_&}tfEua= -@g7!C&tfRT+8H#7_-S;H)IFc4(L3Rr5CyN2I3t3ko3WlfU$ehd)#Mj}4fG+OBaC_)#?fB>f6L>2rPWD -!izG#}i(I_ml_cBS;f)(ly>kL9qpyN!?+qi53+3#FIR-dyMO(3nax!DN{zR{;(=kk{V?2~}_x6eAJXQ -YuLrYlK4HZxOClE}z|HlKxZ_@G!Z53(|QU%dXM?V>8GK2#OSARRntKg^_$%V*glZOyz_gF{b=lc9H?@Z7F%X;>)V^ -NdGb603njg)rAnF*VEzH49yAh~5!}OzaVeq9iP&evLSe$w6BGUvAhtH{3S33njW<9Mb!LV~&iWMQAC$ -9>GqLeJ&7~g;u!h+19@IrHv;zbq3q4$u2(VK`ht3As=ZGu|5n1j-tiaU?T_qzQNeQyroW0sNccSbtDu -F>ou+{bJaC3d$Wf*G-H$n0(1U&~|$_=mGE`T#42-P(d9@P^f)c?z5q>*vfE -h#SMZ2C?F=cMot~DrB$~}xn;^A?XF%}QxVo(25i`g2uc1Zq+Eq;hpAG`!^Ah{_o@`%j51Y<(@4Tlr5K -O<7=`6xm`f+|)GBl>5!W3k_BMu4Ve(fY$%MD1AlW;FOi=gNHHqu&vObOsa6zDG`6#Yt`&fuSVx+sD+zg>ssIQry;zNQ|Sv%Jn51lSY~cL~94&OA#aMldK7I@bnKS -Q~T$3t_|H}8nGE#O3?m}J20CL=(2#tnnedqO#cjW62WpVfeDQ$~o=F0Dc$!J>1@{|hJVBa(3*<4+A -n-&-Ti#0)a18#OgN3X)2<(qEWyi1t5^jQmz+hiA6}>MJOK%7;Cw@=TKLVAn*q=8{$?&6{z~`+7B?8nN -pSaFi?1#3FgGe+S4QdHP!cN%2de+7DESIYnqI&Q&EVNrkoAWSy2QCtbqy$7?9p4oH@odr*4>PpyWh92 -ABYg*2b}tHu&z*ypr=2MCJ!4mkoC53!w!;^lX3@46A&m%Euh`3GXLQsej)qdQmO5E|Y&4$Zdx1}!;srXoo>G>Z60#awWzZwMw$f8Tqeif;F<5A!V(*Gm+Vex>P{ -!_siEpp9h!uZBmIx4rl5gr(r7{FJxzIPdfuOJq}6zYV-_dZ!3KK#5}3p=YZZ;#YbV)954!Re -P|4s4|dD8#yAiS1Vl~gsg8mSbdb^u&;eo})YJAnMBKP?AIubKY9NseXfMccDydi)b`|SXZ0zYw&=(eX -EyC^4JN0Z=pxvyFsPBgVb{e@OZkG%89#mgoI4U2KSg@PV(}L;4wW`y|C}D#q17w^tPW=X+$Tntbo|oB -Qs1b#mRy^WO{~74HOCOMny5#7jH*UuL+@T&FEyFCzj~l=leBXkoI3?K#dDW>8i59^U8Vd}4h|#2^g3d -RL4h;1)Iz`kS=DA|(4kQ&=Gcqz6Si=x)s~cg*Gf?dSJgOd~d!--iR -=$Z>6WO+@>;2vgC~GGgzTvxYAxZG8&%{x1oz-%!LEEbR!C9p*CJW!nJd8~WHd(AJNVm&{R`){#bhNAB -fD49&P)zyk8OAt#C3b5PgHNYE<4h+T${9r5R!h~A$Dm|Y-oTM5vccah#~*M{LJatzpYM{lwSP&-pK^2 -I2#pw8NHI~&LOPSOkcAi{@Fgcf0W!~H-LZyHEFz25`xx{M>jLbujKmK@0ef_W_fJ|hbNt$Kr9yFq)MT`PC! -LG}eKqn&7{c5Beq${K1~9X?wts-zx$FzD9vuq_z)$r~*~Y&0OU&wE4+BU>^WVo1?9%!B9X?TwmtGe2U -=Mt-?eF8UaHt%6#Ch`m{Lya-RAX{I&He2$95E -3+{oyMp=uI7L|RJh7&nlL6}!0fB35igyNtJa*(~iU7s|W0Szyb~1O=F7pxm?~2vQa-{CIpR!Gjo*qwY -e1p|dAcq8-O0jNGfl@bft7heWi3#8B1O_L56cx@`eU$~%3`W42d!nfB0>2-3`w2YRs8;i#g*YOdvNhp -yLtB%*@v(IlDOe;N@;RKK_-e9Sg8wGyN@-HHQ>&B32pMlm$pXz-^Dn4b)a4$LSiil@P$WEHzfeq+RkuEVp8~ -^lA?B76o=Yv4SlLDJO*-wkJX`^y<0np!bJ-{wI1U}t@)SfyN^v?oqG56WThE(($}bkPOd6mj|O5x7A@$*f -Y%97+LDe8XHW)%(1tV8;2vcsRgH1PCPWmV}#av4KZZnB#C->*<4YDSj+Wx59^g}8#3ijIo&E996@0}( -bAWc1x)hLPU1R!H}OF*r*7NbTl8{0xLJnVB7`=M@Qs%h%gYeBSX=%Zh%yA4wx8LYjYnn23?po7!~{kC -=~Lsf3p7T-u4f^}-8i}}_#1nMhH!fXrSj02j{fgsO7i-D4&*$?A3E*dYWMHvYp<-$GEN5+QAeKyBnZF -nyn=&Jy9dK|!!uOz7tW(azlcH2=?;=aL%OWfNbN#EB;;@;6$hc4=Zee)qX8v)_}zRVon>ny^0Tyxa`N -|9etg{PmP6mM5(RD1qHi!xMu3yQDj-mNX#D_GZCr(6N;Jde+;(Qr3ZbB`p${Zcg~{ -H#1O#Kmd%ow_l5UFl}2Q$b!@AeZ!mrY$_C}o=oAiE3qpQ`a%Xyk!c7+xl--!0g8M@3YSdV`^64rKiETUv+_+Hd=D078iZ+O=(ch&{@l^B=gM+c1a}Wmgs6i_SofBcQk|_MC%yu3Y4BZ$_7%^u3TK+i -J>lF?1+&czvoBi^>_HmYjoX6`EfAFYA~03I;`qFV#Vl`P!>PXx^_m*7|2s(#)?@sRHX|#3y -7=PN@Ns5rylB8Ix(174M2_Z36fMT0u_ZuSG<_1`bISS=bXGYLMa~~@CK{O!D5LFuoSu&lfamRxnPu^y -PGN7Q`jF{CFifdZeEifaauUH<4YgWuS#*(|L -3nIR_;}bT5Cjii@nUqClzyMUoqQnG~jkj#M(2X3-~sv+txjRPY^v9gTUG6q4jMc5Wk_&mVTAYDl -!Rxh+`4<88u(kY9RcMQ&^ojZYO9Q8(sKTk-3(ae-APn96yD`es-a{!SOsi@xAHzzQDT$S=KSvd@eEEY -u?Ah+jCHkufXX$xnJ62@6!zYX)~_uTzFBZ?sJh9g_qlQCuUTafZR@=^iE25rY8%Y -|{Dw4E&>8f$iO-?QkrkZ7XZ1Iv9N7i2;-~oJwJ5`op9V8QLj5RUYcQJ8HWGon)k^s@LebJET060x^r_sJkr=@#r7oD%w0&uiaO-Vpju!gp}4M5?VA1UO64&rYqwaj2fK=;D$|W7G!!RH{W^1Nsi`3X0Wu_YIuS|h1Ct}}2)vhElJRESOqd@@H8M+aA>96PcVYRdTtWn(1-e1qcw^8(VvcEpc -ESUx+MO_p(98Fv5kk~%m4henSXT`7`n)Rc_VH1iw_THI>J+O5n=)YAq!u=1lU;ltCis2VY=cr5STew* -DP8Gw;c}}7M_w&?%L*O#d{hy$cO{$!3uZDk6qccxU~gnCk -*ie0e!ym`8I3@u+hLs3fpc(IvkO&Bd>l@Nw*iB8Edsg$)hlFcQj$_(%F8A!jfWf{wPrmka*Jo7(>pRQ -?n?D9*kdew%tkcFw`*42zbNO;#Q>C3vUapKB4s=y8OS{ -u1C9yHi+V^8Y!a9@keYblGcgx=UnvuFzrRP>qr8)LT2edrry%JwBi=(YdQ9%W_AmDCN%IU&+WvzhI`C -9m9PgJD*YV%r6Z>21|@@+%Va)!0q`e(MN2<_h%##{2PEj_B1cZkxpX(S -7na#`6bLzUlk4U+W&;Tjm#uGN4_2F&xRD``NSnD9s*>&T=6x14vTDnsD0tPyVmv$6L_9H>bJ~Biul>4 -LMV&;8?@KQ323o9hm)2eM+-d -ULm9DEp&}6}$cz*?^)b`DJlP@$o~iC=PL29!!IpOIl5Uv3x`Q>)OJQ;4baR-+Dh!88}$lNkHS_mpib0Is0J_NG>I}{&lg|GP}BqO?z#D@Z`HsB+}&_02-L5r3n -{wg$fT%&D=1e23i9LBtZzT5?GE2%Q`zox!MzB_29M9nJ5{AK7evE47O8!xfbK0wKQKRYY2y}`)bzZZj -WAy$0R520RAMN~*ykH#Wj^!tsddj})_vT(atiLqudom-2gth!=v&2l%!J8t1=pn^GD)z3xUCb4aDQLS -9mBuCNZz)=q+#7%>(HD~S6?-I5RdH4umB}S82F{U5?-etZ&ZGIP<-^0!C&LsK{PcT7PzX+i8)`&lG0- -Rz+CORT?N@7JlIYShP6*4{e#fpE^!*sFYPxLTbtf(ZpWS&^@Bp&;B-V2<&^FguVWy&uSD~7=X9TWywQ -b4VDqN@w~nk!zJ&2u8b{NCnCI7#A$FH@Ihh?@mqTId`k{4%R{Pp6M{#&C=J#Lc-l)AM2x;tQiDVITug -oNu4Rc&5EU&`6(A>g^PhbaHdc@GGb36oEg5+5=FSPXc&2%uDMsw9=gzt&edC-c!@Q5jO6ARxOLNL<*gef+oy;mZEdMhVdkdvV0yEJ -qhKOVn{}Hq1bSzY?E|{Wg95BRJdYj(jS@PK6CIau=+fv^fa}1aSIdFVR{)EIvOJIq4&W9jmvmf0Uz1g -^d>XNwJ~^64pOurv8pFp+vb~?OZBWtO8MGeaOt(hvioAmv9~AM5Vv!eprC`z;P9dEaa!4ry09Jo^dHo -pug6z5o(jO>ST@yl~5H!8*qw`izFW5ZkI?~m)^&1=CAm2@#WaY`T3izHIRvBVu -_;nc76K!DwnKo0H6$IJ;8D~@&pA&B1V!KH8%M{d%sk(=U^99v`KlPJ-#5^oLt{;QDn`@*eCze0}hdLE -)fT4@`;U{k}WBg7UZxlr7ETcn)=rqtS@>W}I^ -_76|+w~90b{q|33y2r64n`fUgFpP0SaI$g%A{2w^a+YH*CtlzNU)9VuuUp$u|x0@8c^)>#zK;bhrh>b -0K8q&<~L{oXLezelGj*$b$P=Agt2J1R)j%B5&z%u=75S4@*`_u9X{ZIK2+~uLnnchR&F5QF_%mdH=UP -@+M!mp;r{F|sJg=rK~El5Sm|TCO*fwEAV$f3fcWh(0PoM0A1y=BNB9B7hCrh}17*SAQ3G?)bZ9ga=F= -F4n}vo_p_P3L5go|S^@g0ZIH+}^+z)t!vj@0@i2EVn60D>}5v#JWO~A*6u`hX|RT)&}i6=%^e2$bt?* -~JcDcJ|j%nt(FT;b2^@MU#CjkM4}^3!6Cg={MZ4z^80ni#4wy|)14Ap?(X&^6fBN!>V=5%6gbxv9|lY -|^`__CKc~N6mw+7y|BPtbR~t-4g3S%EPHmFmsVfePs4A@cjoL;WW2M3sE>;TGFCn6&WDuhUty&BZvvTs18WYjZDCIdRFVr8IH)f>8+AX -yCU9l4uq#=i66xoa$jnzk7yGrCG{rnG~&t=U=Z(ihuB@M444NSqM%X%HkcSggq3I3u!mv|vi?HMC{HLo}_7 -+Vw%Zn63(Z)g*77a*NtIM)}_QEWT=E)=%OOg9)9zHGHM+g+`tF9Bf(-z^6qoz(q*br@)IEDecCW-*Qz(53{S!H*{C8I?)mSKIkS)lNGcNy_cTbElC*R<(4U(u7JliDv>vZnLa&&S?_78pNA -H$JL_$(3rZ>Muc88nVkG6)o|xx|}9Fanx>nDA!zALn0qKvy$Zz;&~-OUQ*857mWV4g -d%6}7!}N_m)dO>rC(V;*BW2SVZA@RS`}AWKaaG==R90lj+3P)*OzmXyipUt*m+}ns8PQaar2H*Fb}Uv -(R`2EIc&KTnu$*H=xGwZ^@mmfc`s_isKtr?pcGWiXo=}B`_2~bFI&d_WitY7DQtqL#|PNq1$BwDCXe< -G#p`y)qY_V_1Y{m0t5wx#H^A7dSMaD$U$PzzE!utD&-zs!{D?E;9kuxc6r6@=KBCP?1<@etSVdQvpaZ -eDQE_Nv>v8uxUOs`9eoS?lqw<95|UozWQ{ZJsM(8%ms0l07CLL$J^fd9N0P#?`Tt&m -+(3?g@=;a9nj>;M3y1{w+b^p(nHZhMwqtwflT%V1uLcT)dE@Elg&50GRrF*kN8SzcEL$ -?FiV8qbWx1a+Jo(4edN?q^NL|8!>wB5YS+~+i=f*lkdMH0(bBBfMn|)XC`i?#5+c!qc~C)ZPJy1-mC4 -A`_p`Ri!wdpkDeF!G1a)-`I+OCNxAbZ%6ReJT=8Am;_X`Q&e{!a-ZTmTDoe2kY##=)c7(n~jU+cC^no -0`?TYn1&_J2f8L(}kJa&h@JoYBgel(Zcl@V$zT@j5Y#SSXX@G3h7O$@3Uh+k<{b|*#=RtA5MecXX(bX=Xbn8(O12RWZ3wfhvbi;tdRi^jvM;8xljtA#-&J*x4Dq9D{uld6MJ~D@yQQt&YfHg|F8 -Or3`?^~2>#CHdX?@AVL^YUS^!M`_3T4R%JA4*OOrWV?Ec$=3UM_1+CVWHh1oj40^p9Hdwhc?2zm4A&x -Lr9e|>eqJA^1n`3QjOiA;T!qTHu9luKXw4LH-ln-w6=A=_T?Z^`(>Zj}Y4sM^#flJ2*?JG4q1H -*)skP)7Bse}q+~;M!PaCBwD)LK^$s%679F6VJoaeC>jm+>=_ysH+jE+^m5`;FRUub}t6!7lA{a=etA_ -PL4VQc>p1viV~v+&jKcotZTg&*{7LJ9zr@NEX3(iYewj;FftOwSo{JVp;`5y!*$fc}7uuNnsW(5q)1& -$#8HZQ?rL3|FxfsgTRQZvm31f5AS+{PMGi8qW-ZB5QoZMoz1&u#`-v0;pYhK!3vQA)s-ssThVl6HAoI -P@+LU1e>{4L#UlNr1TH9_>k_Z!6A*Rqf8sXEDeRV&Qd##&^VZfr@zF4pC}mis*QuPS|}E`s_Bj#n7ps -?E`tB$-$H)6Q}>H#VI904&nmQjfcFiKI*}jdQt;~dx;DcivvQY(8=q}NP0KwOzGhtbn%(y`D$SFsl4v -KnJMjvoyjg81$hpv;21;0Ebwwx9zFE%Rq-qbwegYDfqX6Azl;JfMb>yk{3le3Uj5eR8UmwWq@b2sbTK -qlZqWA0z-ups#ZydTb6FMxC!of?AgJ-63__oF2^I0nC9t~D=X)Jd6c*@z2ovxIl=?}D#OR4gN%+xsW5 -30FBXuixChrUS}qUH&q0on}o%3usz(2^g;W?06BB_{MKv67sD7&fX6=#%cCXEi?Z>Dbvc$jrrr@)#0V -?ek4k9<7g?i5g_>)^_r=Sr-1fiN9hq+CTX-4j%aNHF#GMu;StBdYlG?>wFQxf(5iN7LGnWZxf$r$0_7 -rdDGjJfie@xLkY*?&=z1h_1R985ZFv-pPT~9Y4D)zxrr{QyTD^DHf9su0g{Mq#+8z={*)ja5cKX=QnrU-hQVej2c_eIvxp#Zn%^LDjQX(3zUvdlOb2Hs_5$};`6_b -LAseATnolb_m}!6?u~Sv3(mH_Ex%-TcZsUf|I|edLkGWG;`yk90ujG{!#l8y*t``<101i{TNJp`;t`k -0D)6Z-E4If{*=TU~lY@iF#)#l-%*eyC@2kqtd57=n@(v3{e&u%5fO@R9}vYZb5x{2Yx|?k*!n#2`ZPG ->IRWqrg9Q%CjJ#V_6AGi(ohra?MoZh2h5Q*V -F^lO@S)Q9A#Y}wF={9YCMWe-7NS8DEL=YAy>Ed9auMGg+AABe@(h{3t>JI;_}eP}wwS*?$=^!&+d}@fg1 -_BGZ;W;#B}hEsHNJy>{cpbE9NGj(-<-~w@2G*aN{ma-C>}%Ak;pHHLY?V{`UuTQTH6L#45*~J;;VRQ0 -9B8#qLPI=X9S~4y;C299j*+ARxG%8y`4&b^;RL&4@;!>uIg`G{0#l)8w{hbP;JgD?hkE=NIowMffD|M -mgnkUp=;29qudQ#rsd^Fw^C|ijVH78uaB5(yziIP!W9^eyIo5|*5y9%%fuS<6TM_QXG-i~)6bEDHRCS -y6t7wAZ4@j1NV8hlsy#q}FKAVt9xQ~e7T3)M-Dq<$YDRP|N7*w#o6rc+W=i_vx>~(NTvzDYz(YN^Ege -hRqIP5}IT|F1>-;W#DwcrVgTz&Q*Y4D>VKWAU*65LFoOC%~5MwGy!+eTvR>O*l9X>DZ6lWIiB$>6K_d -XfAGUD6cp{jB86;9^yI=K&iM>3Y--exA+(TsZ0T{LXJwhlcvw;S#cUM3nsw)@yF5nuRoZGU_=*YRPxN -T3ltG5aU+75~vze%s+O(;xMK`Tb-ueN&GysTl8Cf=QZm36WxS&5#vgt8>cU&q_~Ra^El&N7SFlP`(JJV>Qsi^;ibdr&tcnJ{nVL=PRLDCbe7d{KL+U+xSICy_rSx{j0=3({H&E7@Ub -rL;&^R;jL4bV|7#R*D^Ln;aGX8KNxH9c?AIspnp-jaAyMFT&4Esa6JI8&Js$ibR=mM;YbiI4ayw18!F -HMMykHtvG}W)^c;6#?drC{R2k~yZyyTEo$7@bY9{^5nxOcon8&>mSW+d7809r1gmjB6f;C&vwzy1<(+ -iyaSpjT7g3$kx;*yCmXot@~@ghl@qEXmAY(2`{T0%Og|{l*>SNmDWDO-Nen4}YnMe0E%E7=83*Xk3FM -c7jg|&*(@++v#qt?i^l92z=ns&k!xy{9X^7{5~EyvlZz`hq7CBpK-C!aCig>W_NFNP>G~L#+Bu~3R1UhC$TrCFRXY>^(`x%~ -Jnvcx+M?GxgcisxgWUD|7C7-;7tRIE6Ow(C1=+K|VP)LOWLh6xVxoDHR`Q-8XTaA`i-0M4NfSy4qI-l -X-%0qpk%+vvL0`EOx|JMqH3kD8&2Ce+jh(~0S%S}aKS4i-mlc>g4MF#`ShDUNzo=27TpBD8SqO;oe+x -`m%N4`x`hOmj@3_cWOU;7>MCpw3HV%>&9j;P89kk@}v` ->tnDywyryn64QHJu@!+btQsZ&dUoLE3%MC -8zxz@Fni+Yg;LucE^H2mTi{oVwn-j9B`cZOEAZ`Xc8+E;`FY|Lb1on@jy>dgC6;kFVv6jKlgRJo7ld# -~NWhuCUSV%pez#w1!rj$E{@4*Fde|fsJfuBz+UUd7GE|VRQA9)!RFa;)ZF4km8G!+TiIrA*FuWWPF?M -t1K;m&GZgdP;2MO(c3%X@sW+CTUol&gXffaO6Egi1&`=K&jH+DXjBUpc{U38D+1BQ_CniNc^P6b!@FI -K{+LEE^}Zexdc>uNtb9HeGxuohEm}$4M3QYEMcnBGaLrOlj?Ur$mpyb{*#$snr}P0h`BGq;n7iRTnYx -|c4^~y`No>@*^GN%8!UtEc3aC>u4P`NY^(YTY?S8?c+>2k%Q2zi+%QW2g|AcmsUoQF@i0ms+m7Mrwwp -IPg3|&1jPqtlC6JYjCym3id<3};dRq1yc?v|f$>)@j(B~eP5NCNuOBKc?k7!bBz6Fvl`aqz0=LLjBqis`&vnZ3C^T!tmX0P}z?1O8( -cr$u3Hkq%dsGYWu1D`?b4|?MPtdmVJYtK`_Xqui*#pMROw}&o3sGBDlTjcU?jO`bnmr+nkFFA)*U-$` -5RcQWm|9I*CX_>{abAGu#BISgy5+F>>b94YPUv?ET$$z|n0KCcp81)$vdD -aa;oRhrgLe>8Ti0_7>oy55qpMZ}MfU2T|ws#$BAppe>I25}KcNRhMI}?iU6BOS76lWyw88i(1=c)Kkz -dUph2c~fh4Vys>4?w3n3_mk21?EdV4$SYOeSz7b$AS4)@_rlI*!Nxo)B7~xg?F(}uMwST(H_7HE~5|F -@Ce8|Dw1Xqo_Sa<`UYsxf^&lD&0Epo9rdODI#UyqV3PLO|gb -aq6S{}S_0;DUvKV!T0Q5CtI_J5p$_N#lD%jII~Y4&|AmeUSPaEC?v#7gmu5+1$ND%eudwHmeXV9{Zi+ -z-wQ!tLsmRzX>OI~&~EJ`Gmle(3cNf2R;Q$GKp=UqG|X(5z7xD>2O2H_c1l5a%tTuZOSIES$J`i0%tb&vB$WR5+Hx?{>L(+;pK -?bZkZo-Y%VUFu=~K&pIukadJ^Tr-}7)G_4F&{Uo&DMAI7O2#EIm)(~MGXVl@PacXzwswVTtCN8+}z}# -ziwJ^x7YUUO^^*(sHv*^v4#1coX{TT4BLuOfW_=GY-ecEZkg#^NP(;Chur~eVjFgn=QxzPUiAyg{*81 -rV3dwHtrwS6E1@9P!_S4?F?Rt73GXK1wCu!r~A=FFP_R%>bRMOSUEy{w6VlelEI5HSD2HQcPEuW#y_S+s(be&O -o0Nt4aGxS;ShJ?BL&5tjctN-(z1Q`x)8+2>iniCGd>N*~5y)1Dd!RPUd2G*l9Y0&#QU6AmxIK2j)FfY -2TAmVRHTttu)u3m8^UM}F|7ZD^b7%;587Yy{zThKDCa{f5u|8T#+ckrz2OKumy$cu|^7uZh=9Q%r!@5 -G8Gde|#gR15refkJJ|Zj{IAc7eP{;w#{( -1OEytS6W^6=YHOSPO3>CA%-I%TO`Kg=-ZbwCFJuzHu?;UEj$klHA2*(hV{8INAkwz5v+R^V+_X9sl?& -K|H=$XD(G`$*zZR=Ae^dic!$0=1MidEt7UmT3LKaGP1BO@PDFZ=uz#C9dsl$7!YD-Uk6A#|#6R)8jcL -gJZGx35oXJO94AZT|P!K~&&aoPtD2ieH6=!0A6{b1>kVyXNKp{+3bYXJz6RByx7P+O2eU4~;)x406x4 -0ya4@VLYik>j?vZ7}KD>;C|jIU_?ZYHP1~4yfG3ow&RdYm96fR4UYS+H3Nabglo+1Ri{mW3<;;lwpyl -o$WO)v4T5`7r4~Vd*p~MG9v?cKIEybU#pE);QH@SFvoRT^4p035T(D?^@rZzk5mwWYrI7D=wS1S5&?r -iNb;_TqK!KXC5x=bVQWvbKqssq(`bwoIKwAa5A(6DQ2oY-@XSEJM{V)|1hHFTrSW#hF=|hbas~Z*OFj -PH3FxDEOAF~RO`yZ1)#}Nxf8K0{aY{~)GBjr|+r{t3dP<<*77{DIremQQt~I{Q>`IG1*yuJ^Lva -Q<22eE7Kc`>{dN{a8Ig^k||v7{FtxLk9|db65<+UYEvExx-qu-;<+3y6Q)1IhWNSI8av)z6caq+nK!9 -C69E#zNK@P+M;e)WniPdHf|AUTn&M;^N}9Yw@s4t#RTV#And=IHhucd7Ex~oj5vQz(VjpP~L?gB_C3;eZ0efZMguMPe>LoNEx&G;a(ZUFp24fxO -FBypi0-D>#nZ3&|C1R!K=5L#-bjJH*>0ukWtd@+nhytl2u@Ebi5DIAC#2@{lDKK7ukwk7Mw2oa@Hmy#bu_>0n-vlU>bN15xeLr~&O*QGCJcIRGX0Qqx5|G{O{P;_N?USldPHXpN1HM=M8XngKv-?RmH;NT62 -xMcgz$;Cq%Kq^>lDf^1YzG3-yDk3c!VX`hFT0<=;)L>0yZ|(i2Ms89h7T%*l_`mOwga+J3spL?N}f~( -`opz->fDTVMnRony>)nnE~tuiZjLWYFk1MD4;9a}$z-)R+{S8d>8L~xzM{FGfH+dQ3h(KwlfBL2kAGk -%tLmOrfY7!tp!H5{1_xaxR@{INwPok6V#V`TjAd5RrLtR6(6o~6wCj$tMcK~oK+%=}JA^-oW~-MYevJ -ax;gl5XCjq?0pgz}OnwK@`xh$wp()t1TTGi25yf|6mfvyXmWgOpD=eg^_i}M+ -TzZJRF)CCvR*u$2ekj{0@*8SJ6{(-ZP&l55l;*o;HlWgOZiuWzWn@QAYbJ$HZG+Tf -WDlGU$H}O2Q6|o-g;a0aBu@JOsRC{Zo&7xY_)}k-Ca5}uuRVCoHde*vXz#3b8PB7IsIZr)n3tf_WGq# -k~q7Q=k={R7{P|;qEdcE&?|WF2Hc!Vda;x@oK+@wp4pd}MIS3gZC+23%VS;<4h -UxosJpj^1YGl`WKBdM20SDZJuMajpXs#@c3iT8!5y2ii1&ud=1ZsnZpoAn@Ku`5wOPa#KbSU^OBLhrYVGXV104}S -&A+ub9HzE4=4{(Zw8TY-Go0rdJ`wQ)|QHPR!-^~WT<*t7GAkez*||1aRjPRNQ~a~_ZO_2Js9wZvmq^j -4kx4p&IjE%SOK0PZvI@a5(Qy6dR#_3jd~w_|cqmuetyyuB7ZL?Rj5m5F9!czZ$4LFA3Ug(vUSCyyX+e -34M#XpnUT=z5>2OOK#^=1Y(TzMMCM**joE20e7a&!enPD(|EK$(Q-HN{4GudDGsjNjUD^17gPHzkoVV -Roi_Yr$rgy!=GpU2y*cLKGveUwKX+~(Y{J;zSP{E|E;SCq{7906?LQ6~B{O0L%#(9xU7QhR -JLBl-@>Mg}+-_1p|0nlRn%B!Y7P;XOEyc{TdKPXM7Uu)W@HQx*)+0L&s&&0c#KsVnid%KsWX?v=nKIs -mdD3<;bL5Ays?b9;yLI1)lfqzIr(T1RpHEwsVNSB$l5=f#|e78n-ikq6WG2f857P9xV=~g50Lp|ZWtA -m%bIR#i342&JfUlZe(qUEpIX>8ZRs3qKLDE9EdXOI5(J6Qi8V!ChPG2PcoYx;zA_orVA*@_X}12Lj|0 -MGfyt=bth#I!%V=l>ASec5>K>>diPXy)g_^;PD%aJbRZC!V_mrA7>GE#m>*7z*oS8b)p>V!82}fcfw@ -NAv%}Ff`ZZcr5p8JeJ$oln~mh-ipU^ze2IxYx~4%vD;`b$jAqA>j?$)qwre-cM%!$2A$=PuY)q -)24y@-++%`8)lkr<;IB|LN^0w9yu7RM@#=Ynwu&YsJUZT(4@It2Fybyh_u0?{tbGO69!7V`_q|{>VEc -Iuk*jIBWy&?0+zYnlJHVD$L(GraR^Og?`h?}+mRe0XnI?y6q8g>p@KTi=r_u)b33whA_oyk0y{D)>x`Bmok=Vtm2^Lvf?t(ZaQW -3XrFQ0-hK{$wes(L8iq5-ZU8lLF2^63l?}aTYV+d=x#*Rul?0-)HF?Y@Tj0gU$bpWD$d`&0uppgS##s -XFg8cj6r%|r7j)#%LYInlngM$Jl1aGfpT+V$1!w^joPz?W5F-8p}CJR(zjzP#cVsE{)yB8yF&6BtYJM -5Ew90zpoGhbW9df@tq0>BU|{na-dyYp;gXhmYbsa!#dUXd0;8r-BDs*&X~Q0h$Da>__U3bOOImYVC%m -EGmQd-3!V9Wnxd0=it-v&?3^t;=6T`c!B@Aj8BUe}{+lQ{MvH9`XZX1R9KeBqS+W9CKfhnf@@AU8#_i -4s>?jnlkPDXmv8RC%@YUc_uTn7zRU5zl!G8&70dHC}^R6W>Ne5 -HHql$Kcgp2`&Lii&~2K*SP6x0?=(ZV*__Ov8MLj1k`X`BmL_7iucZaTRtCdoB4B$c57>55TAzSzdhZ* -neZq;YrckSw#%h1nA$hwcab`Er&qX{|8}oW&wS)f;vD(-|B32u3zVL^2{`1rL6O?7%Y_WpE7eG{p%z9 -gPtQ*e}TvFM&M?wXNB?Mlc2^uz4ZlJ2OdR%Xq_IsbBFBQey!fIlOwvKlg_}-v1XAIFkRKr8GVW*$t(b --joTPh)!#xfuNEF}hL2l1-edNf(WqfcfZ;O{Izi7JfHK5s^7UqP|SK8nA7wFVCqy3qa=om)me8w7YfK -6^Rk@9&#m=AA^L*(c(m*>%?>;$G>Nue>>#^Tja6&C>c!#?^8zO-7WP1k2RE)uOoYB`bc9Lr@<8saA!G -cyx_Hl8@pTpxF{%)*@&$>nIe7$7#bDCHCPzJX$CH{<8nIM4a|4$co#*ucJ8atxhEaKJ6D)&DP%UponI -fCS#8BNZ43iB(raP^iy*$Rg*(h;oCfJz}StyF{d3^VD5&{b!_h2pdgQ@zVS|96M!hyqu&3)!>?f;|E! -OfSO*5m=i`NZCbS$-4A#cd>IK2tS0cd5h~Ad@Xj2emKG9}?wjyG1cFIqvGIA*lvn5uB=4;NSfPa3z1P -ZM2APWmm@1i(cVuXcj`^MR#;DAm<|9YOs3we-umv2Duo;735oh0n#x1o&KnW`c9eL$4A6)?bmza -+icsfHmTKnqCQ%!KFV3pB}nKN?OCo(v%xRZmDdA<@!V%_z5vK4+>1l;ThjcWPB@ecuSG!!oapRxkBC4K`t6H)a+){0mwiSOB$y)j}mvrI^B`U`jovt^S=DfQw$Z>87e=!!@Uhrr?}=qLq;E -;pvQa?@^P&l?dg;VJY#_S{MhJ%$iEEkT=n#l$bLM(cM{5pqq6hY`--!5y36gb`n|DmRh(D#eGr%(nx( -J!>@ZRu1>Pgzs@5I7ruzv+d@V|tx&Q(zS|&)Rt -jOi52`a5d`97=U47&yUw2R9UVohi%qDxUJw>Nq1LWT`UD@feIOcJ$*`E734a9)$ety+Q-Z~WVSLuoP8 -mC`rkm5C=`nCC#eixrl>389L>Y=NhPhA;@s^>QHci>Rj84eYjy9j*ll8^HkxpD)@Wa?`jaHtYra|`LE -N4v+NTG$zf>O_)sAxS!njVewB7!1`AdeQmz>e8c1)y=nlWI1dCOn}|6NYMVjW0jS?8b1)*O%Y!uVOfj^>br5wQ -vllh3>w;DZ;NamenM6juGpHWz`4f8|JFQ4*|+;lsMdwEDQ|`?mIhs^C;6x~WLfX=qB+ki`iw -7|&Mmscaqasj?|P)eS>be5xRHo9_XqN|G)&t!?X!OhqvVRCp?>8M~*2?a--&rr)bGnD{miI#sP3ohne -nXEHGD@}N^a!)G&@bT;#{H#!wAmb%fYe(^ytl|Q-!$WoSa%+=bR7zNQeh5a%_Ry?3~lc@V1RP}}C9zqqXJ0IH4*c29AN{V8CI(F>$10A@M?FcQzFGo4NCK& -q&Rrnh6;PkR1)kg9qhQdL5mes&we8`--8s%kg+F?2ck@inQlpARS)kO!(tP@$^K<$nSBr%+X54|9Bsz -RcGRQ)RA*Y5L`&8>i|fj#Kprlwm?EgQ?QbZbjcZpj5s6p?mtHFU$2tsS>I$ujPG&x -qr?t{qkO!@5q1P-Z?UnA(s<>-yt3n^==?Fyl79YLV?+ -WVp+`T~2h>qgZ_SN{beT{gKK8so%Q{$RpmX5+iVy`BYKI<7xUu}^ew9gfzyy&x^jfqXBTl?utQeqxOq}Y;Ts?kt5|-FaEBx}ZshsF(45=lhc)H| -$eXrT=VS+VPWC<@i&E#6?Dy1UbC~D@(M@K0#H>cYD^#E_$DDBfEP(m(Rzaz2fB%m?2}+TrU|9PhUI%A -4P@Sbm6wr<)?4+exIs$z--bhm-=HG^0YAaY;B}m8U6Fc%3N3eUvXCn}W!YVIu&1z#IE#H%%-Y^i5z6!`E}I{&0Y%eo>eCYA7L*dSNa`&o!zf`3w%BC3ca;841t>7mmkG8t -TbV2m0cr_5wgzEyuh>O$Li_DRdhuk=8`_bF5~PY^j!2Dyh$Zw9TchtCZb=?9pEv=rvxvbz8h*p`ay|45|Nbbr+oqE5Iz9FqCiKh=@#FVaz)H;VtM#EHu7Z)R@4nfj$>QE#BQ;G30jniNLZU -iCq9E^tFQAF$jdV@0N7B@~!*Kd0MRXQ&}^ES4XhtzAZ8|i>$(-vCfhbK)ELxrk)s!>&16N>ldeo)m3b -wye*_qEU$pjc;Vl}x&13V*4=A!n$oxZ#UxV^D7 -^imp$(tV-UN~=$+P;vo-UGrVZr27b|mfpneA*a5!lx`Cde(|(q(k3uewM$U@KC1lyY6mlIJXY$ut7Ot -|z4Y)-ohzqgQk9q9J*YRgWYYa!dU%%2bz{k-2fg(0eF@jdl1Xp)c=YhhimOk_q#d9KZ8k!OU@`b*d4Zda7&hNY%iSO#vM&hwQvPJ?TG3XEATX6!tW%a(S3rB^giO}#SVq2ZNi74_k5i5q$K -5CZQMEITrkDR-})07uk15vyZS2l+mU!+Aog4Fs=JJ?+l!CKIc4>G@2&RU1Ryo%>sn0KGPirGkLKX`>V -ju8?zBUw6Yky=fqHk9J8pyrxtH5%T>2Ush8@OEb0kHrZ?ondj+;Pote42WKe6g_Fs+7RV-j+1CW~+93KBC-i -Xofm7F6s*htsycwdiR8EP)=$RCMe$U>8)|8s!goR=YM)~@(!_GnFEn-ihDp6A$Kh#NIs0hNWD{p~xNX -^HaTw%%~4B!7?z(R;3Mj3j07m9~CyS&Azn_>`WOz9yeLWbO+e+T9Y0==^Wsi+ovZomV3G;QdRC({)~x -8;=ms7Lca+23>v76%K5aO+Ls3IY#(Bl=lNqoTm3|F1oMOxH)*HGd}hV-7IUPuzjM-_%FnQ197e)Aq?h -zYRMHPDSI>=)U?8+z79ndTJY8u)>@rPG3PfO<#+mb0>BwU7cgSjZ^C^tbMrEj=#Y|!r7vQLY5T|uAf( -m<_J~cXN8304grCbhQG&Z5F|g|TBU9F_$zqL+?G%vH%w^d)WO@ZF?c_&+FUtE54_9Q^CG3ly2M3SxOZ -bzBGOXn(0~Q2FQYX{Itg8|K@8^)vF%P23@TH8z@dx=KXPRBRMr(k -jlu3O?(AyLU8EAEoe~MBl9o-pk&tn-d8&|8M*2ty!iibew;}>Wf -o)V>ryv(*0*vSxICP)Q)SUIlsDhMqe}i~-Pu -OiDT_`X&WlmX0#~48)77HVr#uLZ(=}O*tf*;ER`ykKysk(jZ0d^O)!8-_iB##nUIEaApM)RXXqXny8w -kW>-IB%R!x}gCGe6`J)LBP#RsnNm8?voMEnXR-4|Y?>#n!ilgzrVfbG^{7+wc-smh#N@$io(Qn@83Nu -N^UBM>lBUgY)`q1SgeIgy)L+~O;+JRv;5M{i7Jl&K~El!$>tFGe6sRWThubGN>E5uP#@f@ibJ(c()Mj -On#1~o+?B7yX+tGDSbJY(t#GQEZO4ls6K?!m-VCfA$Fo5emG{1(4JNK3=eq};8=Ri+ZXxekmVWk#`&e -s5EMlYZ}+(9*+NhzjUqs`PvR5?thO3gY&l+XH(8T!FxciUGh61Os_s8RV-kZeI?OKYv8cIJW(9Wz=M-zlk~BfwPg9TvC&Uzq!*UTB^-j$qLmH~t4XMFCejn -y#C2x*d(HrTceClD$h2%yh;u~BE_{{mD{F{d{tMC-J>j_iBdrp5-@iC*T2SPk(fr -WW#TB8&shZkd~h9RBWtA@mnsY7^4fvp%E25YWZql0ld}`Sflsg_4o3>IzyzSlf(8owLxcHK0(b;==(t26Q|-RQ@jq -QQ$U7-%7<15Bak7W(h@q=(h}z&b460RtM49s(~e6_pt)jsx=V}A4XerevTgM6PA$mf+Um-+b^7wBP(! -Mvk~XYnDoPkTCODP0CU)HE!9_`Wx50r-!nZ3zO{JH#D@Mcre)@7;>({@NoNkuB!|llk4cnfa4zxC@PH -BS?Zcn@ohB*y+f%rezvev7abLe{Yj@|=m7T2+N<>apzNP@l$-;M)G5eJ}*WR|S0`{~Vhi8I7pWo*M_Q -~Y23uB5uDZO}P6`$HMbg2}ji&1QGICtYm%5L1QmaSo_=frbew7Gj&ZxWU5t= -a{ax%aKjN&um4ROjV()*A4`~Bnyx^AAx66Ui1Qo;}GPr0;MPU_W+AoB)WImN>=Vl%JFJ$%ZaClYga%gp9fX*BiZ;W5t;55 -)izJJ#h$EwpAc=J<>!lW;M2>Xe$W{4pjxnYg&Wufhv+7`K+A!wA&4aT+zpVjlF>&2fa5gTXn{FtajY! -_%N^Y(*^V$B+0k@nN4eIWvP4-X?StvEYU*9Hao&KjtE4#cd(lpUriJFlkfs8Fd5lxmwspQnoiZuq7(F -nTqu!U>l;#xl||5{~26MA>nfn?8>mDPMfR(>1=n3cE&nAf#09^@_6iY3Z7%RRpF<=W5nN0b@r?U~B>= -@6v0;c-Y!tmJY+H0>{rd+LBMZ9%@UN^2@6bG70*3wTr`e^k&u3u2A42<H4~!|!w@QBl!j@ra8cUC{43t1mC7>oAAqH$|K`jjrBIk)#jpU@h!Ky -xX}&WEVZ7QSn=mUG(HPJj=v(H$0Pl2Y&AIwMu_BZAgrdT0g~ywRPz5R*|*Q)XQXVFOBWJ9Cu|eHTtr> -gn1djy&Q*%3q{OJd{;>Rv8bJwUw~4Q5tnJ%kLecQQJ=RY;k`Y2C%Y;&DvF|Knf1Glvgb18krU~CmTED -OqF~aOz9ubEQh_Yv5ROtBS1+5hnc -&)S+QfkQ+;3UC=RJRXn2>h^X@&(U5vC**VqZGzQI0uX?ccO3!Q7#hMb%u8F}VHAn!|_=l=-_XMj8#ah7 -K6D49%B6A4!x0jGrLicMEnjP1g3x`1FuFtPjL|rH5Ee>)9%d$`BXOB)d6|k`I#V9z_^7>)9=#WG8`u|?x -zww?=7&066!m@-LUB&_NmkTeBs=u!i#vuMAYYO+7Jr0q3ndXXlsD;ZHe# -{}<^0=js2g`;BQ{*Vsq`xb8LI -44B2F3+L`mjbkNB*UWp^`M4bYcYue*ain`1fm3v}>4j@h%StfNDPlJhu;TRBpwGUnP6m<%t}wIImr;*e -1!iDCVVLR2uL#>f9x>fD5@E4;<+}Md}?uA+bZ55CEIa_0P9l1S^fQ5c6BrtyLPvG@QW`S<(u#(I<462 -mZa -pRU%FqaUBudOp%i8cuJBk;2fn0GEhBmjfvb9&`eg%YA%_CVLJWf`!le=o$b -+B2O4dUP<0W(jaB46BWy>gy5sa1zeU2+!txIEtt@IcmAWUQstH?mN- -={wBZ9|*;da&%#^TfUxZhx9>ypu++(O;eN*~j8O2mSN<{+0h2N}FMKzty(MY=82A{9kiq_64wFy3{{h -;=NsLmiu6qSi4YIdkKbB&jJG&k$cIZ*lrVb&>4abGO?e`=N#mI6);4Dhk$-9aFK6H1axf7^1LliM#OE -TN4^rrbE4ma@UJ--HFJXd)y)ti^cvYw|E@Sbc=`HOFZb(Fj%ZemJZNW;bh-x+4d%`lN?d&jC7Q)nsSN -JO=bA1R$R}4T8S@r98)4Hr{c-%c7kv(lr)|WhH(RwyM5u`Q|=k=ayOdrR%)yK)*U!_dV-cGl;fkk+>K -c7u}m<_C?F%$b}JSzTZv9;qa~cqOE|^rSb!V$H>3SBqo@71LYESATz}7#Y=ckd$ZV+z5U)r0o1hKaKp -6ef2e4Xde4uxbK%-Rl;V-YW)M&*JrlqAiT7+wzHaf18Ao3y4bM -c!Qoa(?@Cp^^xX4qfkY;Dp>Nld -kjQT8MN*(GpTPbSHOO_HDL$H*zQh=0Fj3h&n+0#a&+adk%ZL;{*bu1u3Ii!{mEg+$^$;YdZXaRwq)?G -+y0i}2fsI*Ckc}g=*<0+s3RzSonh^s6gb6_i`xEXxL0)ipuoKVz-j080fZHG -U;4A1I4r?_y%*Q{z=WRvsSaZ!6|+GwR8xhlOB~QLhp8PNH4|#EX^?7T=su)Tc3H1%$;McBS&!i_=Ez5 -5Iw_y9Wo@qLc5sm4PV9cQPhY&Si{2Uc{Jac|K!epEDj?54nNy*uKaIZgD@pAn#^8Hb3$<#$z`lZ(%(4NAhOIW1l2 -fFdkbQc@5+7iDuc(csvUyXEVN$&T|=$XR~B0_pU^$xch>;`@WITNUCY15`T#T$`JU% -WXx27xZhhctTJblrnk@0vfSpJalI9!(N7>~yT<+m7*55maX8IKRt$>cm! -}w<8Pcpt0`Fh6Vwt`&Eczm{7-of}#Pam|}#Q13B%NUQZPRgqoACG)K;}N1;&S1QNd@AD+ -s#_KqpN{+##%CZO%lKU6qZp4*1j><&Uxj=ySGC-$_>D=fJjkY -Z){a$TLBv>5CEJiS=2GHrnbbOgkH0bD<&e=59$be4Uy;SEk(}@S2pP9}Preg%1@0reFS{W2;3!e7f^TaPI6`5hjmZtHN}=7gy~%+!bD3%yFx9 -7`xa38#L24g!h=d&p98)C*air;am+)lbwIr=D~afc2)qd2UMfSo4}Hfcmt$INB|qR36crK9J+C*m7>rA8| -s`2Y$URg&M`P3!!h0w?AO)t#qU)21Johl8607_N>Sj+2dHR?Aa6{ -dSn1O^1Z2*LYmttZe(%Gv7{2!r7vq1)_+NVI_rBu1((kq6n9}e4#XibmaiaX=YWKt)_KZ3%J5<1!)S+r*PEty5gq+7K3l4?yTu8Gt%Y6ahYcz -RK8r!4YAYd&#@tW}x1WNB2C3$dTe=-Ut>Yk{bj?J|^zp0O0r%7UoXIRCsvCFw_Jps6a^S0=wxl=^Z-) -Wsc=cf6n|n@TV0#IVwfL1J*}MXeZMm0&vS8fBHqNsmvDUTNFnPmMQXm6CcQMfPDhk|6mc~q9W1#%kUxD;Q7o0(C_H)wG3)=lqP&Z -z{1Q$(L!^i4S;S)|%t*O%3p&R4k(7u(jiC~{^hru?dy+1aI{m5=BVGQ8YTfbe3$gv%7xe7jS$WyR4CHwjec -{PA1`DbY)In68HijOAvLqbP-}5^vBdwO0OV+AFsIQlG`g;&77}4zibS;`STN2(YN~3RKL%kOC(;5Wd1 -|gxr5r%f=Y(elVl)V0Q&8TWVbfX_TENC_n(}in9LXF}pR1IvUTl&XF)1)CVmIX0z-yO|X;Hzfol!(`u -VHP#~0)3QUeh71Ce}~5@y^p@JcEE*2fcTf^PAYY*b{prVdELy$wU+M--J3pwNrL$dLNCb&Hqk^(8uY> -aG&l6+>?zlfX9di!WSU>ibi4Jp6wM*xE`c9VczQJ7hrLU~b$%K}Z3xuHQAo_X`_YqCqq+GpW*+J`&!O -h&s`($8dBY6V{AOw%ubSUK3QO}O7$n`#HnxX?^-5-4i3nBgEg40#G7>DgV6Nz=zn>lS^7FqRJs1>lKR -dUz*a}j9J9RQ;6s_k!(zi02m55QJnBbJRKp8?nHz*$?8Any!aDOm-$RpbAG}1S(Oj*h+~U6% -BXSr>-YW`;euidDM3FIj)@bLh>v3ke?Gl|uPhA)3%T4vZzbXlZb3({$FXOi@QAY#@jGr3YNpFf7CcM~ -R_Hk?2@f8|)yp@{q=TIl_xYmRczvT6c=2Sm*b_%SJ#geRD92nGCg+Ntp^xD<&(Mb==k%bPmN4Ksivu4 -2oTSPy*AWYhi1fQi&a_Gx6VbaDd{oqgmO{~z+)Gr)MO$H9^osi&Ji?wgAXHjUf&F*FX#fjjF=l-!=rb -Dn{I;xt1P_t?9tBz%^57b}iWpGj?&Yd)Q33a&JHLH;io^Q`;s)&eWatlYeZOovN8sGuWH#=hE;j1Z-DwG{@5H8)%XcW%<&5}@Eq?ga$=>_gT)uR97NfZ -H{JZT7Rg#_sXJSX}|o7%aHeGnHfOo%xxR0RiMg7?d3|A3|kyT+z4bihk!_*PMN#}0r}IQ%SI@Rj^dF6 -Av^xuHLvAWWj*hC|jwnT4!Rllao|w+wdIIat0O;yff?zWz>kf&R3By?$NekB!AetMs){<@fkQ_}cX)K -U3*NR}d($`3 -K0J+us4qKA-?sOg_z=E)p=|`~0l0hpI?ge3H`2DeDm7{$j0)K@zhBTpIQ|L8m{_D>y?*n4Xh);j9?n` -ssQparVJvIWm*JFHiO2^d$boc=^k=+En^X=a$v$`#u0>y726DF{D6YEre4u`FH3COK055H(k8m`aO4s -yFA)>*w!Q9tcVU5aUsGb$(s0>qSHW*>VGF%n}BV6}270C$}D888e7GD@Kw`%nFy`teO)XXVJfvo+SpB -Z`$&6NGFf&7FHNC?`Lzy69_ctFDr=f?9Ry3i;t#G}^iQZ3tcjXQy-H*i}YuNt!*rpY5Xw5q=BW)0g>P -hl4dAaO_NU^BQ5k87vX1Yjm6Ne7_6zLsSoZf{8!O)-~!5!q!@c2v54qaR-OUlU~|xV@di-+fp8Tn&C6 -gm{He>1gLMOPdlgpZk4~P=C@j`i-xlfjP>o)Vl6MnW2*7?7Dic{&Qr-683oFdO+WZ>T^Ez&We>7TPksQt% -1tmAYT|T&I`fY>}H|r4FdvjAHv;q7M&5M4caJ+a$riJGpM2u!;e58Ak+K2I46e|`4PV&WPlbAVo -7_bJp|MGF5sjHNKSt@{*?)O1c^0;Vub!_#G_JH8=&zsVkV!jq!0?NbyC(xy?5M&ZZzf`pO_TOnM>LGK -!kwb>q1hxOTx^*9YekZLGQ@?$ak;!YsgMV%UBR=IBF@Ey2@~dVxuP$}prM^zS -bEXVIZOU!sr#f>;4;LYZQ6YvXXrhscdkQNDgv~@1{#GVI@VJn5{I&qdHoVcT>H%R9P|r#oW5z*d3JaM -OL+iIgtmtE3`Q}?RJ@k-*O!e4EArhmncN6;eL{!B9+%P$S-KO` -U_3p5rtX|kpvCMy2x3mlFr`zu9g#DC#U-#HgLRu&6Ckw@X>fOq*pPtEcXX&myFYKqCX&n3MX{IZ$_E; -RB(jE2_Ju|1KkD+eSRE5fu-OlOFdIUMfyMhT*E2|(yYjI~-Iww$oJ1k40Mcp~IK=TflZow;tXFPg>^sWgqf<=?vvui)Pr5&t%~BmV6Y5B}}V-Q(YWM)7Zty76xZwZp%)Dir@VAMtMww#UC6^y~PygS_x>4_+<)ttZ-= -ZlVoix27nDe@o%-aJ*Y<-Y~Qs{%sOOLB+rQrXBvR3YCZWw{LdFzumb}w|$^D{w)W+suq0frt5c>^=JT -7Q;n>}LV%H0t4?1wot`wSOTzXdiv!(m|3PwgbeS-5wBE7t|#n?u`F!fVlR2^6v)3y$Q2Ye(EEJ>WRJbKL?0Q#R -ma#YdZqs&R~GJGrRzCXHZ1H6ZT! -8`=TlGH6#E8oLp3k3}d7baWIEcOu>j!Id@Wf&!q`Y5g^T+y^JJLFSbJa;H*&+%aANxtkb3Ze{W0d}j! -BofIFpX(hcVbR@K?fH8(98!#JuRFK^8;=2$d7bh4DBX>H($VL6}Zv9{MO=CDj?zJ2u_Zkk7+lNEs_T& -(`{v09~?gApveMZ+$PrCZbCGKG -4hZoY21vY!wEYK|`*Fbq4)0gLatm+dnBd>sutN3lSq2PW>A7+yi%CNsl483s-O1t&$A--1J}Bl@cerQi4mxK10Ids+%_Oop0aL9tww=S`i^{zR -j~?V4X&1N7;guX2c67@(Xe7c8RIr%C!&g}s2#|l^JMTT79VoGPS${nszpN~LT{Q-gx9v}%WNI)kicA3sqm-wU21)? -55pkzWf9bx`zA?~ACB;B(ZnRpf-Fs<_zLlCfy7#s6t6CHGFU1R7eUf{G2oHXwb78`IC|UuGx6z4U)S% -Ih~I|dIPorh`2eclDecmIkoIY}^34xOzy%{qyuR$Yw)QdGOs~NG5YJyO9j)Go>;8S%EPOpj`ra&kiO+ -}j6!2C>o!lx>vBec9dPyjwyqZ5PdQARY!V7z%0n#%Nk1*4t%Lk>8Oj13+ps+6t1g>w)NqVcbN@&vyHG -6#M0u0Onu7EJbahCCYK<_y+K05xRbipc}mHy1$_L+E#yFEH{6U;m=;9j6KBpEO5B0S69seBVz8;hbQ( ->+o=p>9oRWGKKnlX86oX<~XVE>2xG={MI32@~gn%2|jAC1<1>0Wnrwx5D^9r&J4*dOCkms@;%0sU?^{ -t?IFdDxrtG=Pp_lbcb+&5c3oE-{m^2XaN*wN79Z(yK>x*LgiRU^eUlp;-^cn^0cVaYd=Kngrc{gA7=S -Y#7?R7JCF&FgzJFLS?831g -j!DJ&vR1D!6&zpSU0CUZ+qGD|V2l+K`pZWJW{sm*fyGCq)IPtN)gttP#LS)YtJT@6?OGyVkK@F -57iAKsvwHfchxzU*nt|7nwgbJ=AOmOF@15 -c$*?9PlV&j&)N0FGL&!NkpQc)PdS6N)vVT3?F&Ka(4%%F!$UeVKq -YGwYbOGO7-;uTBTvLWUo*A?kE~)GE=5nfK&&MR|ud)>Xm&(#P@_>zLF*4Aanj9hfKMi_^!1s&Ie#oP{ -I5;pmcS&VaUjq~QB8MA)KF`bdc=3}(-|LDwVyl%_VPSgor@v&;-apNl~OQIhsIJ#*Rnsxf^FPyRRox_ --2F3$xbr_+@o0;NE*Q;>IJr3jzJ^dFXoO?W5TeX*Aw%GnK9J9poGgq7*^zym{Yu8+MQboED^ylJv>7p#3u2V}opXl*MG-0M}=7awSX{WcZYz(R3qU1)Ty`&z%$oD8jkZ0#1|^#W`HwJ5LQrJ~oeUM4y163JxI@g$>Waa| -`wS&%daipQG4mO-QJmSaA_;(AkyPR?sCw3tuvvDFwejZG%}8KV83bp+@IZZ7TU1TQ;-)od(nK3ndp2K -0-_f3c4cREI`1BXrw+wG$7Ryr!cM9s?inCl(g?pE6yhIC7>qp8dq!w{d=gsk1G;X8r^zVsA?XJ=9>88 -r1H<40d&>Fr8k{BHK2MpsV2k^BY_2~t><<3zV! -&Klr&aL-2>NY-$8Qlf$PHth(gm8rOOAdUCyo_xZ -a=E>4EF0pH+eSfsT91xpsuRiy0$S&DT)Sz&iPfL@Ee=f>i5@?1c4q5Hg|O4Gif0mOKO|uz@ -j!d1W{!P_)8OY^x6;a)8my_Xz{&iLvL?=_p1}@i#*rN@oej7eLOpiT+B2R7a;6-X2`Um -BZF_5gqXcXyjS>NL44r=@RW+YNIxHmm_(+W4ThZ; -dC3PJ2~CQ>4%(t#c4CA7dQ>Po>2p*qd2{h)4819!Kt0o3Qjk1x`oqioW8^9NlyRDsgCFGHJrwCDsXxS -r>i)Dyu06);RxzRNO&fr87CO6e}U1CKix&9Uwm!X=H5R -|$eG1RaVV1vU*k{>Y3#$xXP$<--;~N|7W@?fO(UsfIvGvkxgV-}9J!uYNs`;|G;%YskVG<@n5nJ+f0M -~__?ry=6uIqX^S@EV0C8Y{2Di>?{zs_(SgJpk>kyJcV-0gkPVr|ZHi)SW!dVReJKXv=lR~nh3;o$(TL -d#zcOFnX=ybGcB2Kadz9KNg-%Q~1c|25g0oW&Q`*v`fNAkHknPh@Gi{x_kbg;V^bevS*+n-GE$D_;Aq -sCv2uNqG^UTS>Qc(CwRyW^pTuZF9JgZ^ZKe@-yV<}TFuVeCD6sStCtRn)=W0`M=Jro6pLD%kecONRIt -c-fdo%yhUjY-VNwRYJ -HJ7W$8Vqo3#>`UUP+qhII`x<@x;UZ&`DWM@);raSUD2lVZC~XNA&3%*{}bAfrG9YY#1_hSk&-qM?_yYGG^51F=J!Lj -gOl!@%kI$CrwV6GWEurjHYRc=A`K}1k21>*5sRKr_7l-rEW;y5^XpMD -Hdj^`05E97c(F%2E9{hRRQB6^vEKtX0bYqJ3a5l)EiQ4MQ&Rl5!Vo4nQq0VTqV}e+*b_ClK18rptv<0 -!zwhlwI0*9%}rg}Nl&*ju^E~KakxJ=qwvepRO8*Ce$w2l-&y*I%BnKbn)s1F{$&^(WtWP-aqau>IUwI -`6)F6Mt8t%Z5uGLy!Zg=v7$oYa3b$9A}ohKgo(_{(>@U_Ook6AMix^L07s7Ev|YTn7HKxMR97b`Ba6w -peJV5K?ZS{^e8qEM=KAWHp~g)3jk;Vw;L>f`LRsUyubV2f2>s9!f%%J5JvDnb~Dd70@`M1hGEALF3|R -=h95$gso>DO=rjQ9`7#i0&pQx*(jj7wFJ^?fDr89Hj9@p^F1FZORZW`%w)M&j(W?UcX;D|MZP+hyUi= -#6TzQUUOo=D3Crn1ntNEc{vGnsG2dd{=}_aO=3iGS$nGXT#=|&7^~Xs|crnk_E9C~}1;*3cPu0I0{=x -EGNmF{|de>EW*Z%|Iq10yogW*kdhi9iHhW@Sq`wlt^KtH;wNow8n_^A3f;U9?qBI>J+tfDa=%O#7cu; -kQI>>U4bWYQJOyAVRN@ffrEUC7O`Ra`-(($-0qy!~%4PZ%pZ9f66w%~55iyMCghe>eP!f6I9*m`7`(g -ZiHBZX=d;8;)jha!i=L@fmnmEw%GJ$1nL3QCH*s|@t~PRYZ?0|(Vd_Y(wsUnqt~PLWf3Dsh%+v!^|G0Xf8h#K{U&G -Z~xq7e~ULaE&)bO}^2v-;Dn0l3(A6%`DXN_Eaml{8=zQom$T&-|*YXDOx)4K1V;}RTiU|GzkeVp3rC6 -XBUGeAkVA5mI?!lAyeo)s}SS>Z7%2I4dw&-w~R`_O8Eaq>aJrbN$y%cg!-S-&bP_#{ -EDWtL#$=jF?E46KOb0s{HH-V@bxGDKB3yAq$EH5O^gM<(4rweB|Sh#ZitOtkP-kHXP_dc%gD`uA7r?F -75sEyg9H$rPKP3>BLQwX{6GFPboz^S|8+L&{3k%kT+-5UP8v;Tj -EIxlQ?}7ihX)fVrv#n*Z+f|96M~Z(ktm-u!<={^NddkXL;e1Zu69`r!_0e3H_8 -EOB24_5B^xn>wf;=%9YEgZj4})c9njxB86^YCKYbW9n-7a3_9-O0J?0%r2A?yR*rpY(VlLCjJQ(xMr7K8dh#5wiR^t -8hFD|~fxTF{cbiVf%-+NDSxSt=4=*r8>A>(6Xi&Kg#V00EeB0gn7F^tz>P?wT|sDg&*;)-H$MPOK$o1 -3fC1+22$SK95J#&ftcb2EIne-lM;-JQ(>aU|@$hRW7=^1+0y3FJ~!9FGo%v9o6!JUvs)%s1E^jzWhau -fUL>=`6^~wK*mmh8Iyw4!Y}r_VV&=*#@!DP-wRm7;LL-Sxzx?alX29l2(8UG|z#IGh`MRGVNe78{FWk -I%MdW4)z?j%xpuBqi{L6VBQQHf_p*slj*KnuP=2)Fqutc@ -f(Y(keieAQKdFC=y7Kz!A`n&>2dLB)dYQbY(=oGuMz<(Zlke64mm)6DVM;4#ei&T%zN3!UeT@hg`&+1 -Hd1e9Jrs7LUU%y;j|Wfc~Pj>3FHZe~GtzRl4oou2YrViTeG@0=BaUq_4qc^QFRQLb%qCTQm871%(=@{79lXJu}lXe+X3X -4wpxIU=NJ#8_vMW9;I*g0Z%O6(dx~9pugMQ)(AoZw5s=3G{KA#!F}m(5TAkf>pC{eOpJRy=3%tk>ZCe2?RbG -z4w%jhRzV6q9GBBnD*oV*bGr=o@Pp^#QDggy*As-lHOdCTqjwnb_?yf8bjNL(mhKVghLdoj6+ANn2ETSt0l2b10tf -=KVV`i79|Ap7TuY^~}U10nChXS~Sl9|rylFLhXdE$KhLk3AwIJQm%ce}|x5{u_L^ayy!Sfux@pLi#}% -{ZfNSzX_mrL{#@19Wrt70X-QY_9T6?Lm>QMqK^x7_%~`Bd~h*i4fKKNK2S>n!3~($H9;B@1S!*m2a@m -!dJ;Z9w7F+va6@3V&K{8ApAv>)9e^-IpxHejEFI}LH=;p}Tiqx^zJyOypDF`o2l0)0r}c>mabV|tK5*$2p=3H!+)%{!!Lb+A2%$GOo@jk7=6wTJ1oL_Z;b=*P2sMSsu65p -oeeMfF$jPxPr^2i%}NmJW!&2=UJjfOPnie&Zt=Rre#WCuA0U_o?MT^SCG2g?xs5g#2yn-4Ir-mZRkcL -hSIp$n7_S+XtYk9#X>}5D4WEN(N*@c}(a@20R$v(5pH$MH3lFBB9(Oq1+;&3?k=JeS28AKMBv)K^X*K -8MxDBnLLH!_xt=(GED+LSA>;Q6$&01^si9y -&g_TGPXa^ce3Ks|nY!JO?=m2Q$p~eq(!(Ys5FmHQ77JH4|PZ*J!qrP%ejdki#&V)&MGpY8+m@jSxTRb -mppYz%p|9#P%DkgD_zZ?(ozn^e0}GuPfB6@C+DZXL$S!LqEs+xZ8yFFs672$4{!cSN%|BU^=w1DezTr -J2v3){^nNic#a2I>aB&gB7pQBuXnrsAqUECiN`Ho$6CC^%WW?$=`|sQ^m;HTMMGtY*Rz3vtPG)S2Rc0 -Er^frUrGzvsBV>R3IK=iF9Rzh{4jGV|0`+A!8K9X-m`j%T2n4Sg7(@n&ksu#^!T&yFpr#jLwNh;pAs? -YU`=z?;cXfAi^O8u&HBOJ*u=)=5Je#+tb0eC)>w69ixn6?rJ8pk8@cZHG%j+S8zR#|Lb`!qws(Ttfgq;mE)h!Rznhx=&?Q1Wv-&Y51d=Jt$6)2Ql-|@Za42sJ8z#e4aGZ6pAu -m*i~h&?zXC^iseI1Jk)XnO`gzcgTc-)5F~dUu{|hndh<@afukoM>AGAH*jT!jBx!>msHL{DAZ{vfj

    Y)s&9a=ARv|jYUyqo6fixzArWCMIDml@SW1(K)<{Ycbb`!`26_Gt*O?qv -_l2u%s|>Mz~t+zs^;J_j$qenWJmSE?TBL{HMoL3^(ZcYlld&imy7(4P!|^gw@*qLvr*+ew4PUSJjmWf -@9(C-!_WuvypW-=M8l`)z^nJ`L>X!H)h!ke&ATypQP%;r5Q}>FDS#*6mIY?s^jaYe614>FGmyjt`7g) -g5Wa{CfKsAvMPsk2I`rAjlw$HJ}~nJGXbUDh;arwBv+)4d0uWz3Z4)dpF+Gr=tCXC!l@$fRInsJiP{F -{=uNe9t{E2es*n!Z;DTB`@Hb<@BOY1BEu)#={sDL?0v*qM()7@f9*g<7?H)+aC?C>aQ^{mUtE#~gs&&|p`; -%Y@Gk7jbrDkw287zj8!P%jZ4?W3XPn)08Mf=f+?MF9pZN#|tiwhz{MVn@5cD82dge=Wa%^e!%8_UneK -#0=~A>>-|61g_}BDr?L1#+$CXW|K!+1nJr(**Ks80%Sr-Wtb71=Zs`dNSy3@oBeA?f(7ed^`{5Kj-6F -JpPxRj~}e@J`P%?nPC$X*_7G}GwKB-Mmx`j3xp||V(tR64piQ&wKEGBvlU#jl_cWopG{rtC1fqy73gx -8mka1M{1#;9iPH-m^YRLosaV~SBvEX=mKPqX&-0o~ --nY-hf0dS1Tm=FH`2_XXcMxa!ztyz_jMo%v$kwAG?5wQxaRwk;d8p_&-uVN?R@+~2oF-N@2FJKxlo~f(tHj;p*iG>A4h55Eb(UH$=Imm5P-|2AJ3 -Pf|B1A;1atX_a89TM{ki)`GpSxIwmPO`?0ACsM*Pl!Y?nsaT-3s=yLm}j?TEg(snd7=aUS-|yVY6FUT -F)d~?Pp~d9rC6x*PkaPhrhU2t=K~*5-h|6799dIi?&rFIRh!ewY8MOJWLsRAJ30_sgG*a^+At}9i(-Hm4 -~Y7^05>^;xqH}?&_Egrh@Ew_Pm0`LMOy{V~6@>BEvCOEb>qXKBKwazK+4X0`qac%~3>EA5l9@`kZE5qh{A(Z%efWaNV(=jKDqvT03$+Cy&DOwTVuFEjI_#Ih7&m-DvGx`;)1C4&6iU=V}fUw@ofx5TtIwMJlBB;5qPdc$7v5v1G!m{msua~7O9^58_M< -Y9Ca`Lq5crAKas0VT%E|(DcsFmu0M~{1>FBzc-Z+|e>uUEg@oSa#52?m9#?;EznI6lgwuPtc_}w9W9Ei#fB*KV{g?c@KU~MmyFdK(-y83L5qqZQ*;CINy79k+uPYlle -d8+an>gJG_K!XD%`^SUx6jDW+}6GS>z=0E9OzKyoVw$bfh;<;^wjvy_CGlJMdNuw#;s#6+Z07f^fDLt -y%RI1Moy=28qetjPGdQZ=G4GxB&Rw~Tm2dRfzw7#Kj-u$ryp|Kz-c|Fb(|jHw3^emINi6{9j#&a6W=_pPOoa#7j_2c)TwC%^|&0JsRw2{-3oHlS;$7waE+d19B=_X -EBahkzt3a9a$#&R0Xse#i-PD44>aY{IC)iT=5smy63rzbhB<8(WxTR7dMZSQ|ESKB$w;8fr=o>K#-I! --V7GWs*8jhyc0bPK0{&m8EV`IXIVKX3b=X8vn=(LA?n_Zi{5yd!xz_2=ct!%Io#_tX?N#?9{2-XD&Js -uXv=TKX^jb&UJ3mHWTi{@?5V-|he3>;6B#|DC^IyFcX9;pfYLs{7VnydCFMt?!F@J+_}G|57@Je`Zcry;0Z5+IYK0X8HO5OA1o`!K}dnpJ9xKh(P}WUnuZ0(AhBI-91V}Qh+| -t8^&UwzZvM-aOls0F9v!DzR!WL2O1Iq{cE&_gacg+-_|G%$p#wT7sfSU2kGqo(3S(A0(2ov+!q3m^yF -X|I{}aMtHI1qVjyHNd?~{;FsmWtGJL7P<6^?RAuKErsNYcV6ZCb6^9f%ugo~J;l|z}GjX*~aV|*;o=} -{0C*cX5{MzOGvK0KVIcgt{!14iv5ec@VWXDiU45llZA=v2>zy`zDw6?NHx%<6QOKCKObo0M4neb%ddwz4Eo! -F?v96X+;9!0CniDsf&UQbkOZiQz()a{mcZg*2HKLq{BH&7m;z}5J0j4rQ(6AU0v&%N@Sr~d=uu67mi3@jwgV` -x$sU(73@7hbda<2o0qp7eYM+eiYER;7bJ_>9|Eu -?|`2Gv{xpK-GL7WIwh0&pTC%pZ?jlg$Utw(W@$76eFMI&VE+Knrfg>a2cS>eUb_Gu_&}R5_*MXm?w-RIpc%%WVSUDPi`mcsE0(~7&>uQ!Rq~q^m^=ksqHFrZ=zz)*K?`HMu3 -82c|Tqf5LGH4A;qXFnz__l!k?Lgb$+YUTgOGwo^h$rw{fNoe1G7S7ipkEZT_%{Oet$?{P=xc!vu7r6s -@CKkSR5CkTfu5;k;Wh&`ZGgTV^27}E(+wcA&@hu(&lK|0X^a0X$Iux0wG1pbxzb^C8gR0`%LrS$&s*j(vym#&-~^8PX1RkiJ#}^ -Elvl0G+uH@(p+^(3kc@dV${tbmRd@Kk!DNVlA|{z$0D!KI9MZ*+3iKXZ;k?Q}s;$bD&d?uzXGdiVNpD -DCKGKAfOb(|2-N*R@K%*M*ev5`u-!GxOfY$>348BN^;m?7-`~#EcZ9q -*wva&D(9dQQA0`#MS-hGCJy9TJS8QKfbN1EEq$`|Q{W+tnbfbRMkred|2aNBRrrTY=8O7i@q}1$yWLYZK~$23&+P0(~9Okr$bLyr+ZrjxK@xA -Qd@}6z|Gxhqegmi(DV+0nX!Hm~S|bRN?lK;=K^mM|uP2kWk|hjycH=te)b3w>O_#XB%5&p_J^^cxNm5dQ~gV>BLh#`91wbRM;I$d$6zdsH1<8pv1Y+%tWYGCHX3Q8eW5x`!V -8H^CmzPJJPA4nS;$rw$OdfsoQS!nIFOWTZ_K@o8YVye^pOCh;Ho}+wim7E#G@~k*++SZDO!T!YS01Cv -W6DpLii0n$T)AcqTKGw+zjREgKL^IOpw@xnCvXeul`By#AFHpgr)oOQ*1>;2DVK7n`N}oTx*x!?a*(T -69@Dj=`<$FhOuZiM*B{HtIabTnmza4z=wCuL_O={KaJrtHH5Dy(y@ -As8t7sCE*(368dNQ4PUC;ir499$>RZqcRJYAtXaNhC^1**l`@MB2{JJOVk3kZ0Fn@F>Un%MR6I)YgMf -K(K@nP2bV_;(=W|!{Li;s_sXBMcs^~ka3hHk5;cJTxKA52q}R&Y-Dz9IfcpG~cAy>zU&n%Y06D3{aTZ -M{T|(LUsF(-W-3m*WclW5(pt;Og~_zQs3z2n_c6eBsj+6V_05^J%y} -Z+AC!^R0(ZDJB~DeBld#PY>T<1?4=H%lfa$`Gg4*NOE#AnKy49x$CaG+R6B)O`FI|FTF%|?ASqN{Ns; -5Cf|Phty|9Rc+3FCi1zKf0@ZrNrOiT=kjg2K!r%oj%lZjX?7Mgc+=FB1U=g%j%+;R)K{ -r20*y6ov>$&w{x>C&ZS`SRt&;c(FQW9{0t36|(S%LJ~hz$gEQenR8kpMewD4rI6dcQ%F{ -$LKgp^kdl%TQdwC^9(dpZ^2j6que~dQuj$(QJBAwC8mfIFrHPnAN@5No(@l^|2wqdAeQ6LxlZeC+Eit -~LH00GYtK*QbK9br-mr0noAp06SLN^zV+Yt?CazTGlHLtbzI_G4bSg ->G$SiE?#NKH)@%a<>g^3Bf97VFoqm%eQ4)~zD502DGDy+iQH=>&YnFh_2S~ii{eJXCGpemBnk@)#qHa-rLLfr@!)t7%4-5vnu;8>E3@7_&&FPb4cLWV -5*?|z=tn)p7#c3-Q;JwmtHlvIr2MIPnc-a--jm@wGJH3Ne~aMkKy|={0N3mX82VMzmwq)F?=4wpJMo*82%!|7cl%S9lS?Fj&J_fJq)a+vy*wP!|Sg?UWd@(>Voo`3p& -|T(D~tlE~N-6TrKF94*q3^cV&1^U}pO;dP|T87`o@ZT^z&zW+9;V&?Jp$-0Z9-%9b5XB=*3=d90{jE~1Lfy#P23vPcXe -&p*>=`Hz!)4H7-%&3`uYckH+5^&s6qYCo$IL)0)zQ9{~ydB0t3U{8{>k`on5r^#t=i0(U*r0GWh!j_= -mra3mP8iRd};dN@%a -^&-Tftz+g!*GW8^#`!Q82)OF8nvwFC64u7T^Pw=+xX$Hyj-J3*N*Z6cX>fw7r+|=an>K_U#?N7tA<*) -CYXWY!+rfjx*Lr_#&9PLP|LAq&A+|y!V9&#VFsKJk&VZ9Dsn+M*nsnfppamrF(fF&xyS{4zEKl7en3c -|F(@Q3q|<9ouV@$Gd}Ft!-Gp^~gPcNfE!-OyX#A7oH}4R_EDi}WhDajh42u7Ze|TVXKcHabF`PPes!^ -i`Utl{Q-sZV*KaU`upgW&`ty3q7LT%f5L)G$Cy{dQy83RL@%blDh5p_Cg>fzho-@irKs^vX>!$ZPDf` -c?HZGRyn*w8A#;8D(@a(C?ncoiJzZ#y5}&ZAnT=NufWcVP^HVu;a~73tx5|F*3?TUIJt@ugrsX$W9lZ -q>?Qcz8ZMJixz4hbk4yx8aY5aEu-t_;CJk11{;&;l(l){BYjj%j#+r56^=?tYCi6SFh3|+z{g1fptv? -{rJ|_rCRlNA^yzAVBKkqA1F`r?`>NeIs_LzZPgzhKadqMIH+}dyVJVynKi-w?d?trW6>63cbX_?{GF; -3e9v$qk@IZwx$(vHVsp6(!q+aJD>+Y5`CPAFy<|R;6xrF9dwiDYBc??KiTUvcv2Ie3$X=Kv+V0_;g8f -{VVfkvbhB`7l3U99qSp>ohK?;ow-;d#kGyGJBU&8Qf7=ACq -Ka6vKPXlar%U-6qb}s#JN6^{kewTZ0CT8hBK%R<-e -q{N1T;y+)1r+UnJuJ={IqMQzt6ZXR_idsOEOs=WB(Yqje(aPxS+vfFFg4=yflO*|@Ae8G)j8oyNKMJL -CaO+22jSg~@Y^5tJ>T(zoGx#ygk)Oo&QW&Ge#&H0rI4i#KI(CSnwkMaNJSf&Pld%j-fN|oH%zUnr3wQ -OxZ%OG|5JJ0W4`{hRU+&w(p8!`Soe-n52`tI&^-6egsx#8a7x^?AOF}^D9#T3>D3CfB`-{|9@y9U?e0 -Rd<0{t#X`@vgkYr9?klil!cAcuv>LYqS1uw5Qns2EKT7;1y?QXD}5Vuaaz7sZu5OEb>q6;Ao0HHG!j5a0s2R@OQ`t4 -Cq*$VnYs8xs@L^3I(*sKC7b1o+oteji{PY=tq}wr$%G_%U7*|9$)R(dNyYiSeUNn>I1lKmYtQ9 -XxoD4jnoquLBI{2>AUx!z^9BdiA)C8#j*Qlr3oT-EXKWN&6+KzPoLhmMT-{T -onv5Npc}@0Qv9!9zfPRXq|~cduQl^9`O>9Jl#`QVeZ`vSnuxgnE8MF@dsVIcJ2E2-+#aPr=Na0&2#;Z>zK|@*xp(*4#|LhQv5le!d$iv{8a}HszW -BK1Aj$_`mDM=Evhrlw$3uQb^Y(?PlNwgUwt)`<><;fF$Fa2+__WY5B)fP{J50C*|TTmci;${U>lal+} -vE*VH=PKj*uH{2Qc4$`>n(uGGzNF$l<{Qb=R7;@%Ov@`#qdz8*cF=_}{v9i#QBP{rB?nY6v|>2mbr_@ -24$Wwh+slIA@kN4E{qNd-v{@9k}n=vq#!WUS6L33>x4Apl`4f$OyP?_>Cy?IFavnM6LD`dF&=~-$B$Y -hp5*+qKpefH*ej_Wm~EJSj=_!Z`-ylrGEYTtvQEJW`7BJNE(1Icn-dT9>@Z?<1=J`>eMOO@fkb^zkx6 -G0JH#a@EL|H(C|wl*Dr{YPY~U^cW?2N;LpC^)`7q3N*1a=jh( -;_pf7KpBI=w+WcZ%>4;okoZN4RH!8F7kBf52uXwV@d&uzBSM5*wP!=GpYF3 -=<3SR5VTJ9rKmL2sb{@VmTblL+--{JZ{Wr6J%5kq_hAo@r>!dJ~st#Xl;SsQy;%e$G^nJ7I6Um*+V6?8+<0UL*&LGNJ~>Il0)>wT7}JJVof{qJ(E(k5lzxH#IFGmhIQSWT3p>v?3K~ul1v3pnOhfh+Dowk}ec_|<|NZyhiQ`O>4jNQ -fbQC9(;^?@0_bwgUyo8pIHPXW2zBG53pG2bX&qQzhMAY*PQ3z=GL9=Jp8E<9Jd#&~ipVV-h)t;Mfwc0 -d$+~WRZx5S_2JhN4+R<7*NrvQKGi9H?Ad+4=29f}s{I`ruDf7a1wqrEA8WJmfm!AH`7_RdA3w=^0;PZ -M=xoVxs=vuDtt*>f}aB=)Hd*eBIx8k{#-eJi(##o_Pj>T1(}j`4`|f07667x6&J0``FKiU#1%evfkI4 -J7jzZ(1?BBbk_nB}@Za1pCjonT9@$$KTIcX;AhI8k9Z5CpG_u`TiAAW6;3-ciGJLzef)9U2*ucznA*Y -ww8>12sWqaP_%#s#Dhoa7Vw9kTtCsDvOefQ8BD`+rXg*F75{_+qIWMd4f#ZGUa;CTe3J3UhxQB_ls$t -6WzXFqx9#zVoguz(o@MKxL3MjNin9?#3);rqy#Tr8EdxyW?JuuQy^J!?5eqix0G4}YQbzkU1m4IvBFVXwd&xB+)X3 -w%HPJ=Qhwn}}72ZD`vjdC_L3VSQ42${yQ+EKGyFJ^!11Qc8?J&3rF_rbQWO@?Zmv9~ek8M)svs#|}yw -^f^W|Eys8i{_O9i{xd$wz#Dks2%La7>;ZTynZU>E^JBzc&6I_*}DR_@uOvKC~p>hvvum(wta7nicI&Gnj@CnTAPB!#Jh^%|6HYM{e={2l%rtNd0HMo& -s7v{P07vSS*sB($Jx30c`%EwX}0eTgqV?wlEFSo|y*tBxTR3?2{J9`_jBPKl+4eNM;&7VjBL*dNeUIP -$CMSokoosWuu`aI!clW8hp~aq*B;3(@>0`c~q -+fKFyvzdk6bYYSgF^g@=dJ^5x43aXBF&fkuoNK{IF0ly-gcgnSh10W{e2ff -Xa^-XFZ)NFQahXO&|t(Cm3O+rs1U-@0{cRnB$(uJCW#v}v0`g9gP8960bY`%x(elgUIqdi0>)y?fKgA -Ad}9=gy@$bLL2Vu)YVrpaHP}aYE67H4JEgKZPD57GjQFGXm-GXTbj!?OQhLVU97%nq$D8;gc37h5-G? -uD`s!y{CtSgiM6~u&$@@y1(`M_3JHc!^Gz(F)@+YCJ5iVcI_&2gx$M$OML+Dz!fr5dZpw6`k;GQ=OC| -xEMO0?X=%?a2faP7va{#^Y+Q;wa8YTLr(L^ta~3REAooEqw`R*UVZsCo3k#!q_3BaM#*K+{Yl#c+fNk -hIXi(jr7If&88XLZK|C0W+^`CwmpRG9t;iviL^XJo)DO0Yeq@>WKNt -15&?c0~4qoXC?d0i=W2{gbiKqqhm9rmRHuRXAS7EieU -2HS%Evo57goH+3a?9^y9(%7+Mr5r#5@DB(Gpgw*2$loCY(4abKQ0rc-#gJb^?|~_&K*Dd5!xSoB -^I{#;DtTskt0V+`!yI0}O*way*ZYkSPN9Ym!0rU -#~7_w9Q8;ldmZKbVWNc@@ZnQhv%ab+G(fjxiu<(HBM@CW;bY#)+_h=>U4)~%b=t!2xWN&LZU@Eo6kH| -zlZ2frgXh5jJdh8@5LU=tV@KA3TTz~fADbaeD%xtn(F+LbS?HJ=Kl#UJzP^B=`mbgUPF56hM~x4>P+( -jHje#QG+605J+YRxoNm5q=arh5ui(W({R!Wyu)$`RAWYIPjYJl+V8H9gdM>8TXz1+f-q#dz7|*Bl9=p -Kk)x-kIDEBd&3d9K;E!f)>D~F#Kgo%-^#Ww;~RVp^jqm6>;QHNnLtk<1NdZpK7`L07czo85YL%jyh#U -svwSIWx8bkXRlN@DbsM;WF6aaK -s+_+i$;3^XAQy`mN`?>bMTN1YO5>!~o!~>LZ-jF8OcZF7el|e>wIfLoX47K!@7zhyQ@DLp;*c0p4!du -z?mXTqyNf*)TfjR(tcvoA4bvqv{9jGe0X^w(NfgcN_ltxTw!z^tnw*GC@yGO_lJV3)h1-;0xV??*uIh -cg{~&Jf*(>34h?i_9gY7>#x6-eiuA|y+YrCKkR_@aoID(UE;6b|71U5+y4O#h@Hq^;Qycph)0N>;4#*m;63 -D^_IDu%e2346uZ17uHHEBeFwM^zcZolAwXb$Z6Qqv182fB -PTqm;?7cJd@Y@(gY4J+eeSSCo0eb-_fJb2;p03J@Th#d@QRS9kio(Mn?y`@pbb~(2hCJ0{`*ak*r4t? -O0YjR?v=BwWG6k^wo~zwIf-N4mcKlZpTGDMZ^6)6|`Pnl#nZ8zmNM~?zimLi};R;iv51zqZ;ONiqPpP#y)w6Xw*5PE=R33HtYqXrqJRmqPM>zI(kjwQW6fJwRZXP<=gV -wZaCZY9r!cOId8JRy~uv)SN2QCIX^dXZab87U_8UnoVzRM8wG>UTkB-l3q_r;gSMA@&fzm!zx;>wLezLr@5G)q>O0tzM~&@q^J5LbYtUF;FAP=dd#rOYF -Y;*cM6G#{H|ZPdWqp6K*85N^TUS`z)TdJ@JrUZ -Y43B^y05x_K(CLe524QJ&ByPH>ZCJb3a?)yIL{5!--iMDBL@a9h(B8v#K=7`><42{Ud@mGk4x6NC+cl -dAM>oWHq>~akEl_h)`>ddTi;61J!*c$abEAojvF^_sJ`B<@>v{_TW2L1 -t8MuS=(YD~~a_b;vfLCuf+M_((oufGEqv}~32X6bI;In4LcQYBkfAnm*dq>E!&pPK -sEtU_qE356lfc#}N(nHteaQzJNX94fEn2TQj`$2T?rZ((#bB)&`qEy;DCw@`^EI#>9>vKYpmb_6i)ZZ -vtMRVSgR9rcABghdLAXL!pnTF`>3GS*tIN9boNqCq+pdq&^;eSo^l0-@X=XUjv5xcTDw_^)A$ku&0T= -aix#g7e{>swFz0{(&isGu#=ptvrZp7{->z?k&{iIK0RXb;>BUGMXdLcpM!4T0v_OaI6X?%4N39vt%t?j -Qb)7b>-Of$4P*?0{HG^vj~@7vr@+%9VWYJCPfM9#0Tv*>Gg;M3!3M!7pbmo*mD>ZULbNm{K@*Z$=HF# -n7hGa^3y^wTiZhLDc~54EmV_YDFM@B;n_=g?5U&3gZ#O%$h(I$bN)|0Bm9J9cdB)TvWrz=v70W=X#Xz -l+!h{|{XNKcI`4TVHef?c5Q%a_%VGgq=P{``c<|CC!hRf_)C`>4L{tU$HIP_R{TZVM;cT#qN|)Dm;IL -3Vu5JP#;G;)W=0hZ(;gk<3EpIM$I1<7IxOx*H`YlqBb2B6-6I?^pRZO0Vm}Pl&^$*z?NwKj6>&$K -0HMTtY1o`^Le+!E*l?T=E_MK-EZ;4yZk3r4Jk@?nOsO4`y9X;2a4Lc4N)|xHXkCc*VS*J!Hs`u -TkT{9{n3{yg`r!>T4XAZS?~1UFo5^M?@VtR{g>J{D8+=%Ioth!urU@_cBifDGUl00&&a3;1pLOyIeF`}SF)bTtZT|1)RKj7UpM3sY-u$Qd{c9z2-b-QB5s_wEumweP3uL%0r -lqx=Q#OI)*NO`5Q-6-v{j|AYUTJbCg^jE6mW#C-T1)E3@(=N(yNhF#!VH1I&}@giSFO$B*7)3Epn=C; -jmzlWvYlR!L*kB^tW0r#q^eMQI%yawIy$+@|?E1qy}IX~uMUnBj0ad@a4c=hVlv}DN=xmOI|h&?A&Q+ -&#~<@}gO#b4+W=v6s5Y){oCfdh05H4f}QZQQs~-6#08bKB@I%^oQH416EtoSmHw{ky`vm@cgQCr{h(t -Vy+}?69xL`L{>5d(IPl&o~r$_7?qd&+O`--L=69Yqx*@$`d@d>+(oEIUYJc>KyB%Vd>JP>r5uoE%s5r -EnmL;Un^FuDB^c47E1w-F)Sk^L;lX!luJ)f4^B-@-Ike|DdnkR805@0eN+#RdfJX1JFw4sm&Z86`ga3 -%h_x~FHH+6Ey??ac2j$E8^`difsErC9*FvX}!>N2m*%W+P_(kiQ3Hj+^PHS3yQzSnYh{2c(wx-rZ`dl -#LiggXV@!Frl@J;R;;G1Aaz*Rlpad7K0N*QL&4F_mzC#)Z_e#Ke=YZXskD=lEWP;14S9JoXOG4HCRV4 -34w=QY@(8`d$f7RTBPYd@^}uLP(TK9CYkAO8B_y;*52B~{cw -|p2R^DX3i$n}tGBG<)w!?Jj~90O~PE{q4{_tsl)$vQWD82E5zR~l`Y&{3`@k-H;r$GQP)r*V;;RN)|L2m#uC(uq&j7z(!Ef8PP3CIqejNC1p!kn@>lmHVJyT+%#V14bppnQU*_09UJnmH4ZJZ9c!63GXhO`!eiC4@&PMFswQE-qShS=H7XND8MO -m$j61uvkg4V0zcLz~HxVTi4Z+pTU*c|Yec;q*{f31vsLp9zUfVZ!e75HmB_=b0{l@p$xR_}y2ua(!Xb -dm2}drrI3MZR^df~Z&)>0X%VEyDTVAkm9I1&bb{JAV)6e>?GKv@gpQJg~k8&p|qHb@bKfEyw@aZ@3PU -SoanqM7)R+v0^Zf7A2xZB##^`28m&OO@O?{Q+RQ2$Gw&4FE>t&92y;jfv<@XaUz1RAH�@V7`gc5fa -DXE2(J=*DM8@-+j*5dI#|zq}z5hbm6`14?SB{&nuJOLGh8N -(4m6!V)4C({Pen2%4wFE5E(4{e+%4L3z!>?3jX6VT64h8rdZ^7sSziqgvr3!h4Vg16^pal8za`WVLJ# -TTPt*n~&A@pnE8Jxbv75k=%)sa$qB?}&ti$e8}oV_XtrqGJ;}x{i#GZJ#h;NMuYzLerS20rA5Uh7B6g -bilBf_7Mp&%|N7v@fJY9WWtnA|AJtBT&g8ZAir5vrL6^y0ULGib3yG4$QjCP5}za3p85`tq#4SO##-qmGfl>Y$yfz^(#gCe -36B3*sFJstu4aB`1F&+hF}1Xphl8`1pT+e7iwC)7XGKQJ)3d*?oVLh&vo2l-HTp6QG!-&AO-m|iVCG` -&x{sOBK0xhFL*H9z%EYQ;3iG`BRbv@U6V(uSlZrcFzmpSCh>Q`(-iytMqZJ82b7jwbxYg(DDPYPHk}t -7d1+&q&Q!nUS5bDI+IiPsYKFyo@s$`5A>7cQS;zqPd#c(d=S&Gkco7%zoxB=1_AV^SkCD<~Vbrd4hSG -dA51JIn}(d6{=I9kaZ$60>G!&Cg2BTA7 -udm6Ijp%SyZPA09lUc)cS(0k_e}Rn?;`p5Zu*e)xb -(#I3F-C`>96)z`@e7h1yD-^1QY-O00;of09jkAOc!Jc4cm4Z*nhWX>)XJX -<{#RbZKlZaCyajZF}1`vgr5x3f4Y3q*9r7lBVr?t9HGv+xRqzeQl@h9@$lCiMF|>NQI>AxG($L@4Nv7 -Ku~tt-RJ0OV~GR?gTY`hGZ+lkAFT(w<>Ic&F0Yy(+8qSXww^xQc((QId2n)-28VaeRapebH8+n&h?Y}hO1uY<+-Yp(T}7K>Si{fAF6m(s&D_S%R+t6%gf8GxKuyOT76sGO{ -V31k>whLN`I@%>2=yXYO1^KM*)1)<2rwozMrOxCfKJZdsS6d+d;4%EUM&ko@@t22_3mfE2;+NgHiCNE -Lyuo(^{iZ-|4SSQe7D6H0gX{|9;=xR>|T~2e0O;N|PD(@>?P3gAqdyG2?LsSEUd7}QtGCd-FGmE2CUVzF$7!7-KHKNtqnWO|iO5a6U -Pm(>)?FRQ#HXXPLp5tEO210z>Ld=ktD^SBojW4_=;}9KHlj&C_Za{5`4D-LkmIE|*o(lB{Pp(e7147T{BO4S28N}l-HRGn9@ribE#585hEXRdgGY~$8`BDyLfA6COq -*BqESgM;WS&kYP--1!+7Zl+tV-vARe-f|Xb>)r&jLUk?i~g%wmyflkM_Ubdvmh?&HmmI0It$l*gC4hp -Q7>U?P$+dv2lV^jLN -77-ZL*ZzsDi-@f^-Q*3-%BQbvcJsLduczye?!_BWg`q$5Y`>)48{rBj%FTOk-pPnI^0L-IDv-E-nDxJ -)e>KX==|4euW40yt>t*z1DLo6610hbB53`x@}ZI)F5uVxtu%cQy^(FjaQMO2ma{h|UfLUNfvqVP8A+- -7;MDhAi-T^+2Ii?v~}mS5Ft)Y4jo-!94u#6q#rq}4ntk{pC{LGT)Y*ESMdR>>Ti9tNdCw~m7yg}F48& -Y*TcY!7fikd(oB%EAVOdj-H6C;}Xu1bGGzl3V}`38O@T10LuDnGm>}q{@=>JOu%>oL(WsvZnbnh&7$* -3yz8m1k{OOLNOs_Jn*v~q#1O!3P=QG7g@@^v@z7h=fEZH^oi$5Grcn4>T(%0ylu${AmuV|p!z(%tjgu -0jt1kcGpC5aAy$;{<9VDGvuLp0mQ8a2k5tlyoTict`0Mx%Ncf(_RunD^7)@A!fuH~_C@%sMyWzk9TBk -OccNQ8Bu7JTi!U4u7JPp^v?T$qBppg|MZw=Pk4j$KG@Hp_$5I-d%LGa@$UF69$jouA|FdRI9a8I{wlw -p20HIx{?v!CBy!G!OlC4MxIXcnDbEsJa7lky$?d57FCv#3cs5P(`HC4 -$Y)^~qplKB>)nRDvsAOtOnRz%=m4;K^v<0PB5772&^!Ebj_od@0kVC4oh_iL8su?ak`h9q_)!ammMd; ->*aGTEU!3n=@oTcRtbjI?hf(O0qeH6)vgZ0bGH~UV)x<~8#Nc7V%NiY^DXbS)X`R*gFyZ~AULk{Fz4m -q0Se>BscJ+Mq%d2v+a-r)di+_$$`46Xr5D?1#8kUQ28_8in_dg^=7PoJ;QyqP0CX4=44(3SIC2g -3>+x=n}fD{v4g -QEm`U@6o6{40O$YPqgTpJISgtrgs;&3lh%z(P$cJh -2IfSyoTV{sD4Z`!+<*A&>5#uZf3fu;-~ -4pFah{svuFZiTmh&govd0O6p9h#bMuQV`sIA{C%pdhZ@^t<94X_*o%xgldYHAK2{8hY*QIM -XhKN8g4g(6`=nlRfQrp7!V6KoD~b6`SF{?4@0$g6)`{S!Ep(4Xpu~>lS`O~SIG_P*0A(tIs4;rsRhQJ -UIn)#0SP4(%f)5A3T!GOK3ISH$>$$Ltv^GOKQiSVT>dyOgBejH<%;G5*tw+u@o9hK_emzP8&t8O-MB5 -?eXc9lJ7%_N8;;hx(9=D?FVm;M4{QbAK+&Vf>(2h$d}{bvq29OD{U9}8eb$2=GDvX^OL8!;)VfET+_L -vN1T`ctZ=sxeTP)}2VyxQGlK4EHqQF4XJ*wNMF2l7-%MP7@jza^Ezr9}c_0a%AHFHjKZSc*dDsf%7)h(;F2-^7-A*fLclTF+a{}~!=RFB#T%rqStbO-x -x1UARe9Nke>D<2-t_A2FYn|d+>(+w91^}8xd^ms2p>2Ljd-@rl;U`dqT+K|$f0*qd&t*v-&B%dsj225 -ariC?YV_yVjj;KE-}Lh67B3K<}-TBNg$&5b$*){!NnxoQUjKq5E7AH7>#)-jfep$ZJfrm=C1kBJH4?f -s?ESBY_B^tSZ^Vq4~%ps~Y3<2)_+Z95|fpisjg+bH;zU9m{N^Q6Z3sO7~)_I^m^kV%O2EtwWo3a_D}! -k2i=9WMmc;)cMJ^Lm0$5hO*=MGVPV5U>~p+BprV>Iy$wUa1C%uDSsQ%V#JHz7ymWS#etSa5|$PdzUyS -Q%x!sg}ecOmp3U#u@p^pM~X=k+`^)Zz6`INHOC}gl#9s1Z%DvDWEoUlaeLNMvA<8MK6PwFsP*$`Tp8? -jU(d!?ug*w#OFJJUH27CxYKkKhgouo+0R5o2iDYDf!dFR+{`siBqkV=n0XW28p*6*CvZ`4oFh6br1Yq -v4$+jC>+a5u?PP*48TQB7;L2axlXZGlgZ&fzTbQE{p -8io+c&!}_l_nLSX_!}O4`|i6}|ipV=nWuz^K0>O=sv1NooNzy-EtS-Y{?kAlA+DJjjz}F~vxed79wPM -nWN?af^laeU&Kj1(FE?7L|kJ;757G<3gRylLkFpWnpsjKjqJu0r}SeYDntg2!5RVF5UJf -_9s6qS83`m6H7ce1Ig+A>(-}zZa=vJW$_$tQD^GW65u-y+jWh2} -;JW6iYd8hz&FTAH&3BB@XO!mQ<}%~R64w_ve^E>(`U4aLcTvO4_0jS7=0-6{*x-KMA%o))6OO>I&RrO -Vd1Ol3$&51TfV^SjA)dKbY2$8gJ5CK|keFNT2`6Kya91!<1xPh2nZtce_UGjs$f6i8gBrS!}5nJ-{w3 -@=gI&*)1phhgY9YK$c)I2gX2CiAv0$nm0M0&859sEB_#3Z7f~8}O%uexgeX$P+M_mbvIBpu*<^FbEo3 -O;KfUE;c?72dZHuZ(;pzP>v=PJ^DY#2XBVK1+W5Y^ykU+8iih-UMAB!BoEk`tYd+$k32=NVr)W4su55 -j=Mfg=cRIKpKmQaEc%mT2<}Jjm(V?Q_5W1Tdq=ScN<8VQn7_pEOpXjfaT1E!xkTjJH-Rf#g;;c#Mb>y -f{!~tLflJU<~`~|76Vw#MUae_5HS_~;kBU$ng#|b_1nd_nni(XaJ?DzV%1iA4w@*Nd-6=g$O%h?5epk -UkK7f^T7>u75jYzx^<&$CKh03tjdRKPI+rPY%R<* -npxRluWZ&tfvPiNi2A;Vf6x@J?67CFKOQJNe`D^7_pZcVpMopog;}s8NpDEmi61Z{;9DF-TOop(5sf? -KT&4|RH8}ozXqUyu;F`?hm?8)Gh!M7E -3PYYrB>oE6!QW{hHJ<|9=9BxS{M!zeeK&~HreE?c;3Id!;V>Q`d!trYLdl>L!27JdH6kXd!tqe^O+{9 -eZm4m1Q(D1O;strZ2u;h|#8rWmlALskH-WGNK6NBfz!kOfsrEdtx-t8W -i$ehS|SXV;9rVwleF{-`aC`v3#8;MFRE;aUvcrPpL{h(g-0s(xmSJND%86U5kp>DeX$#xf&=l_I>HP~*K)d|RxteH$A6l9=kflX> -mDSQ~?Gqhabk1?ak1q6#N&PKE=`Ai!!?qo7AXUkmwnw3-dYKa4eEhalPj9-V^jK{XrGcsoeCdMeo6~f -S?VKB{W)9R5~2e16|Uj(rwx_QMJwLL~!wIjiVizgGo8rr9pb@5nHptAz4V`X@aN46%DakmDp68AK(wh -pnBtQ^29up%_24f1V?SRVkhOvp( -0hbNAlQtomNzAg3Z0v|#6Zm@oghH4(`AMJ5gxt_BX&v<;o{XTt2hYWkhdgIJ7I_bqDjgz`8uF0i=duX -8E8l<07ZNw^>P6y@$T5$0t9~D+@S*xaigtrUJr#>xIMuBlf0ZJInQ5wRp4R+1%av4;1Bz@=r~A*3$*$bQ&5)SpnrRy5T3({X&)@9VV5KgrbVkypF{}i7$<4sDEUQ^$N?G$cmUJ8CR{)eya#jqS -Arn+Uz(n@93#6(iQ_xKs12e8>xPOH9Z(Eh9YBZa2(rv*2Ik& -L3eqy;SaW_)1_IVbaVmm}0L1NH=#lJbH5FPjF!#9x%>4IYJ+K=Mz|2wVKp)tR42vg4(-B+^D&$r}I)T -^&N|jNelpahjaImocc?r{q>e^K?m=n8#>jdaP*xFsWM6b4nMg3Ln$1DMZLkds=p^tpcLeRt#F_K7LXw -pFJP1(qTqytHn=>-z|4y5oDkMF13EOrCMGHPv5!i$F-^(S80Soo`+tU*`SVsk)E9>T`dTxD`bjJIQ4u -!M5cAw+AmrxiY~+-{9WI}y%K`yC#n_F1gZj)_MA!8+0?`0!&gN0l)6QojY}lR_m|q0esS7?)gF!S6ZM|1_8z_*FvtUV`!kUQ&jDG+O88tz*Tu}nPqY#e --34>&uTOme6I0_L3Z1XYMzOYzy(C^UN6q-& -US*gnsY%8N%LN8+19Fp~ZKlSekztOq~|f5_6xprh~YFY`PL>dhz-5tSf#Qi-R3(KY)kH-@rZVcYgp_? -G3_gE0Ly3czf;)kKY?&LJ)tF3n(;btiITwExWjf9Y8mVEPA*uTl{qF;|`h~C+UNkiG6hE8D`rsr6dmZ -kbJGKZnT0WNdNe-(R3!iqXX%VYQ41^W=}eil+6Q}(P;L;{Kkq#mC9cwDam%| -0mh6Kjr$e57Aouxf!%exLTe)1;z4JDO4<@%w56zC%IY8FlEdlGeF+JhRxq@F;_;-?AP`h -h59z@SD{$k-g9iL*$*(9Ut?3=`n@)sM%J_zeP-PG9Y_zA&dHQIC8+50?3a`c=OJsO>IeJoC3@m2snJy -`fw2_*)Umc_GTfc4Wg@L;Af?5gHUIB`=>}u!XssgfI;?Z=`oJ1CW^2MyY|)y)yyAjf8PtFO3B*Nt!T$>%NS3tZ^4*xwhyIeo`Dp1Aa4%0=>TCNJiMFk8wFp3IB=!BK`_fQolt?(KfJ_x0rE&hg6*RGUcS&;(5qFGQEKr(Fw@fq|M?23HOBpbg0&9R^RH* -zHnQ88!|m!sb3*!r^=sj7ZUmt~_Q4vTt?$tJXP8M}6sg{vK?gn4y=+6f>iWJb~XoOb)*560I -BJ$Fa7Cies5oE?%lMuPP4-A=HbI-m+rMt-uXMPds{=HLi2qbZzZY3GrrwvL|7{!+>3O7s}C1XX&{WU0 -2HbeP*)88Z6Zv#4cY9yR+C9%FbrMJ^ioBJjcC1mjd@r8g0$ssq?6dr+)!Nu~A|!P?6z;V~Nm -6~vqbGQ;v2ful=exWiIT&dUzs8N8i-I^j{&-uEUd$0s;6ok;5kPio -NP6n3;L;Z5bN$YqpFOn>0wktMhL|3x4ihixj9JoYc6*zikRs#RcCWZtOII_=P{joi{=EC}M1ng7NO%1LYx70axnS6o*DR?j*^ORM@*i@Is%aa-zQIrNoSl1h)N>+QBdw{ -L480%?Yp3h*O+~D;8S{5Z@W-49(`PNS7*IRMP)Je|~QRdiaGw(md4@qB21BF67B=a- -HtYr?7rHXvFp#Lk&>Z2;ZkAWu=<0X6U}TK#h*98QsI^eUEx&4!x=*gz1i}UL7Y -BG_EFiv|0xa@e|dR>S)9Lu^^!Q0{E{xOq$q4?(8dF%$P_g?mI=% -gP5RJD_<1_sBWJTTRbS$;0M_-OlZ#T|9jz(DTt7r)C;G6seK=}J?@Kp~yz#NyS`1`%1zaJd$4XkrjX2 ->l&IHzQ93P<1;f}%=X!%B%~$E|WoAi6o!6lYj23fy-dha{>^Am||p&$`baw#Tc|`q&`5DQ!HTH4I_Fy -mXVx)V%uvBZpv}gVB+giDMPr2*iUo|9(=z8nrzHKS<+1?jza3icB~@nRW^%9vtpFL?r0!1E -7UD;v;A!7(;lGu>_tAJ|ht!&AQbmb+AS!(|C`uXOAJG#2ig-WDoY)#M>zmRh -n4$ivM)LiaYt>#{hL5UMqIM3}E8w}?la5| -)8h)0Vv9flGIZ)*uNVWt9A;3h*;58>l;YE?tzt0aumbLjmIV?iX8qstiv?`TjGEOSr -!)*>k;fF0JdS@x8vcV0UbI`*mDvNemIg%EHwcg3&=Rg0<9P;zeop{c^qY7G1wRov2=8k32@P1`z5<9e -y>Ux&U)|^J--Xv8Wh5toiv0$y@wrl8Hp)g$za!)`TqMQ{U3@{n;`&!XM_=sWM5(5Mwd0<0|dPozxu^? -nCA9M;877S%?lL9Ao8Q5z3dgbhG+pT3yd$*=uDES>L7K;bZIX9lByck=xlYm0w@wjvI-`%)6Hs<ZftCXO5gyT9jIGQIUs*Z+w4F8%)#Uc%v;~A4i93&I^^Pz}3o2fH6YleQWc9epTq3Fkf>Wn(v%{xmJavvFoy2+J^`Jp5g^r -j~jbj|rR%vNeEVKz)7x^f0)MPwxtMTsc7J*LRM<&bJKqiy&%lI2{>aPNrX#Lv;!L0{rZ6tKT_vMvL;R -0M)=)fXwZsGI$5d&BKloy0 -xIu|uBoP?;dD_gi=!ekq>O=s`a0g8a}-D<24Qw#(V0|NjQ7w@244mo|Irp2P!>ns%$&zrrPq$rdYUZK -PK_!&Jr9S0-Is2E@KaEKC<&l{;6TN#TkM!mt$hx8F%V -}`c8=~VXd9=0XrH4rDKCeAh|VWq^{=KbsDUl+`Jm*dmhK;K8WdYZ#PMdF -DB#J?Q>bXpIv;^4oW`}XL(BruFhRv9gv8M+f-j6~Ml1@jap<&1SH(8ch;j%GJ!rpjd$JTt$+XK8#H!; -3?XULoDi4rK;xoio06RRhe7~5w5HAT?rZsJ0CO2TW -QHmvJIcyJ!>C*DF{SC5lEZc?}judKt8--Bqc@lD#6@g(YZ}LcO7>7@~8)LtX+LrYCT=hW2xlcCvhW8> -Xid?MBTq9V)x@cs>jXqlzX!Q~)s?L>ds9CA^qS|R>d!v=KIkg1keVQkgd*Orv*@>snZRxBveNi5YqFq -p4k8$;H#J=)LaFg;KnFvyG57%dO+kx0fD0>Q?!KNEFQJUc0tHjb1;c3(^^9dG=mR0U@vcx1>hWom@*E -X2beay;xf#D%byePQDcO333I@e1@|y*9#71ZQo9@(^3ReAmGUs810yxiCx^}KJ?G$WvDd1*w;Nn~<8ggYPg*E{uuFt(Ux!qr#V-0f+87J0*(yz3B%6a4f;$ExFtCP`=a>DeDYI2hZ -l^XUf@C>hZbc2Mz9R7q+DMg9>M~G)aPNd#!~Coi#LH@!(MCymTeZ;3I1q2D&JmJIcYep4@e-Va2@$bH -MCJA>O>=qfxg*#KDxt)={^-4ZD3^i!$E(PSK$+Uc*=5lJ5d>>@_GS-a!r@**ONmxGmu)EXkdN8!pYn? -{fQS3W@#7ppW*<2dW@XF?eoSYyorD;m>Z3vQMxPG)Z#rr-z#ZZ(uc-Gp&{;`UzKKPZ-2eeYH}NrrYtS -fk&*YdHKj{HgH3D4NfrUa>R+> -$EjbgkDH?K)wsDKgpqor0Vh;6Nn7GQK}7_!;Cerv-d#19*Iu56Kay^!D4w1<8rlGFV8WF&bboQ_X$Tg -sBWSGRC4l0*F*K1Cro-nTh49Kq3jc^nl?K#x)NU3;p9>U1M3Q0dTiL!<%Ln$Dcb_&3PvNc|2;#!KSG- -vBWQ5<1GmdzI;?V9P-7^Kgt%bhbeiUVZwEhIab_fx0Q3mi73ME}Z8|t5MPI`t`{}@VTY-XcRj`3W0xQ -b`R8c|U=0KX_nmeUVDMA_N03tjYQyEGwi>h`sZb6-`Qnu@N6jrGLxE7$$s?+oOotheLLk~G^qU+`9MP -t>~JS$S5hi!W}^@x+r|6ehy5*PoeA@1aI2;YtnC4!C -6JW}7R*5s?Jb{VhDyGyV5q!I14DjTl81unC5q!8rQ-RcV++$OT)d*bJbMcsklhX_ngd4g}?`W6rkG0a -%jybmCKCfu9M -BscE$+LkS{?FC?H7>tRc4ayAGFm;zkR4%uJ~x-Pn7R%T>F6)#L%7j5WR_E_&tqt;PqU=f}l<*epAY-Z|`RIrZ-;3n0!lq0IAjxpFrfepCir7YSk2JI^4)(1MyF4MYmK%FePS*n?r; -W|sP(Vz(H!pg6J%LGF29@<(kU^%nJ_dAel8!}PwgCKLc+I79jdX0yIvb63r -t1=h22hpTclES$EDELj!vH3JNd24)q_K-HYyMMk-f`e^(6m!G}6`ZM=-&LK}M>ow!BYNwbB1W3o2+It -a{|B6kp)&pZW`~Sg(t2Bq5b06IJ{Z-G{8ba%915&qkA|A%Ch@SW5dqD@_`CFuqvq~)hjbOF$Go|bP-| -&jqp3=F>Cz3IbQc_vVs&5i1ghM9878&Xx=c|2G3;3h6jmUV>|@4YT`p8oMflncn2PXmE2Vh2Pz+d)d3 --vLNp!hdC6rM5B;|tYIghH4bIbU9|F*a;%G*L|yo)nl^Oxyf6R6cEu1st)BG#GJBK -A7j)P%V{w$Hmvl%;G}mdrflOUAUvtLr-avn^~hZF1$XdF6m}148E2lt#Tu!r#@NHa$c3f!g=TiK^(^v -2a97Oy1b`4QH8L@Ym_8qv2vecQhMIka!=`UWeor7Du;?hAB`BgUz*w*n8d!T$|>u#Q?AD@O(hShAMkx -Z@#B$gv{EvyBF||t;u#7-VrlI=yvF0CBipFJ-#P~y>!vD;%L&lGP**%Siq(36`$F5Bo6yhO>^fVdvn+ -|r>T(l*ZJs36Ka=9|v(G3(4q)YX3L0MZQ_dHlje^2I(c!S|G%jZ8`w1OlqHR0#O3%8%9Msml%gbAI4? -0>L0kI#M(^x;etdhl*y)?8hTS^G(QPEBBvW`<;4xlV3s{^*o3{PI{S*kDdnxo?htU8Qmrj1-EYR_(bG -R8XSda)?s-fXJ+{jk#R%i!Zt@YInzKw8C-hH~2Wa~Iek!>v@kM%Z)DpX5)M#WTZ5ZASP5Z6OQHxT|cn -%NF@NRAH}*)s_|ih$5w5P=HCID_kYKFL9S*5^NBBHyN19ssB`|%{sbOznJL^=R7MS`{A%Hq=9+P){|@ -7rUwpfU+37xqD!~g*lDU!LM4hKS9=&nn)fa+HP-_@*8@G*gXnm7_fK@5=Z4vHhd=i(f}gQ02d(qXp=> -U|I_mW4GA$_J7QIP;mb=8d2#w5{*`=gMvN>`_eS9~wnV)tfMlvH&2k=wP(`xHPUCC?~gFxr7TYh}-?y^yMIL;*%_EPLmAW**xGc*2*vq2<&2r)>3z(!zC)SAG-aElP;=Pk{**x*xT`Ff~CpApF(uA{&S -D0^uUqjagHJv(|@gK2mE_jb&CgJ`rkS;aD0Mi9*dF<>KJ?v;D;b{?? -XLbhS3VK^$ka9*ToUAx4LhqMsK{?#vpUYq1pL6)<_A191uAQUZm;2xEP4<2~*?WUgmB*M>AdI1OeEAD -qx`SWOe|e_9(_etdzg3xY{tHlbfa+3kN+bC@Y2(WALRH+f9mAotc?2SBTByeCW% -qnjC#ERO6E}=ccYxswP@V^nQIqjD3>Zs8?uWrO#Xtd@8+{O_gPznFXrn7Wx~9KR+(Q5<3eR@#e=#=fC -^&!Pza;R?H4jc&Pf{{oNm}Q+D*=ww&#JOmOY3ax)=B%%ux_xsn#;j?*nSGB>e4%i*K!0lI%3Bw#_#WY -f*Fl`N}5Gdz(Z64{ehtSbr<<(%*{NLY@8*w}u$@sx=CSBOY_HWqz?V*p^d9GOTtFTITxSY=O5Scq9V! -p<}-HWSHlxa7GF&o% -gc+!3!bOu=VRzP!NOyD!tKT-E{k+ -=m=02ZEmm!>CP{5M!|*P3kPaqrorml6uY`P^l~ebbWGc)CA_)4pfZntenD1#aG0lFXZ_q2Swg(s(E$Q -%=7gY%MEzgm`*I?{$b==ejPEnJBb|egLi-}3q5b)uT&b6_l;#D=Ps9U>Q+FSTcz9^3f#<)VbO@=I2pM -PM=U?Z$Y=^Hes?=T!|3w{h{D%J*6yLtIiTu%L*@DA84QhnKlL=@E(Do_YWiwRp&fmR>o37r0>konRH9^hW -0qslBN#PuidWEMFPJVHwjS4aEWsSqe3MSzXS;0Ci!)z3uZ+n?Ryg4Uo7+74I$*VY{B0$78b`6W9->>H -%lQ&aTj8m_W}Gx-H(8nuxj6btVH&#!8W{v=XG$NRA`30eY1bUmIo?Z-{C*6@t<$;pA-D&Fl-~Y52dyp ->bDdP29~wAZ@znT@WY$1H^zBRIi5#bcr(I#{PuKftGlzt@09wB8oWIDawHI1cPmSmSu45`A3`ecUJyrVB -JjcxjG6?NOBHz*rG`M4nY5I%mn{rL6v<6{v0^yZT8XVaY=c=qd7xgPgUb{@`HIwNZMHN3MeZkWa{A;q{D&X>e%1nby?^|A=LDKO-aB!t|1|# -T*Rv<5zyEgnE4ADvSja8`cXB9C@LOwK=%dm2_}p~4s~qPGoDX&IB6_=hOYeSY^|ziGB#!YL+zah@y99(k8S4?V6aID$XK2@&I5Hf_ad!z*)!qJJ}e1VpCS(?uX485|26(y5+(k9VT -sCMzt_%^ei5$QIXsiU*qhE1AP$t;^Dx$Nr@NJSFB&C2;DIJY^m!?4?UyBf!xrC7J+uiBy#IvHO2md@K -$i3_@acm+iH7U!H~L|@_ku+pcePI1K?{cif~sQ;Fi_AI0fyX9421O!(f_u+?`;i7 -;EjVo*|QEGHXngluSP_YIQy*xQNJbuY}9MB2#_oPnEXK3$(_?cXlb;CZ^U6`+&|Cfb~EcxsFZ$OoIb# -Ry5;Cho*n8>sarfuzdlR(46Svg04w7{W1?6+`$Md%C7v-8+_ZqQ+xOr|N{T0jcCj-kr*!4H*NksOaym -K=>yyhx23L3uxbF%x?afsyl)Bdjjk#>{kQhIU`Tknx4afc)VhZpIH_ -ff6>VUS4Pt`H^eRzjk#&6;h3k|rB(=M>;{G{?BHY#AIL)d%bu}gSad+?N1azbS_jdni@9Wh_OJ$V^wT -`9<;yAL}?#d^VS~M6hcaA5=Z}wNPP21UV2&+9bti=i4BBvSHW*C#(6y~>8vY7Bxi11I}2`HC_t#xnCY -mW)%+gZSmEd?Wu8*W&EX;WbfRvXDgX%Ax{R@Z)hftC1U8lGr@~jvp6rY#-R_@|3YgsIQ(vw){7LI(VZ -jo?rqCB|rKPXyV7+7*pF~vLn_eOiqGU%D1UQKp<5Z>w6d6mCMj}r_k|6s~V33Zo!rl+;-=UJ1}`MEj? -z4Po~Y1piu`qvG55);Joa#UCCR0)Qnsf5eGXpFg7N9sZl*UPp=ZNsi-s_5^z@mP_UM`btAF(Sa?Wkol -b3X@Ez@u;*KfiiIi*1)RkBkJgD~sZMG-6^NjCp8s@pCB-r;wom`}@azce|c%-fn?w~i+wi*C*BpIgzR -*AZlj6UPrJKbAgrdMT&304Ry>LbP>R2h^p>}0MQ5fsxOW_l_`yj- -o7&WUnKPA@oEAT;HhIMl*#d}-S;)X_V}d2;+oMJ8lM#e`jT*_Y9SZE#m -coB7ZMEU$#7LL59al`VZjPQXR^>e6?8OGfWamxi!bs{08YKIPNmm#C08!5kBWToUuafFxk_PI(iT*fE -IG_X%!K)|kZdby95X7aJ#6>4pG&oy}&vclrKKVnPTK#vY{wg|3hl^2sO9=^6k^k%UJa=1>5J5_N4NQ> -~4Q*RNFg3TK#-lOpf$ehDxPmyZmA;An-3il2ZYgQ36y(Fx*n^!Sn$?q#xmnE{D{~&Om;Zea|ljG%{&>EzKMnU8olM#f~i;EIJW~C|C -%^v(6Yy*bzGx=gGpkd;Gn=;Nb!IcdzwNsFhenJd -lMH}1qGn5RW`cH3i|v_iAfkGGT4B?0okew5Q4l!l2!BqE6oC`qziShQ!f}jqXl*CP!E;bq^!r^qXWc_ -?30ukB=vQZ;VjKDkU6cwZL(x=18k_o3_Jkyxl*}N+svRe$asq8rj4)*>2e4R8dsp!1Qh(x)-f>C&f$I -#VO%=vx2V*&KyX-?@|cuMK{_o$5;Vd}JX?M@C2xucjl21HyrE!F`n@q*o~VnbTp3+s;^w&#FRQ70#0i -~+&MVdckN8K6k|C4%=v$)swU&KjRlbwn&Nm&t9JwbBm@#q8YJE8$x!oCTcb!$u_h+!?A|9!w>pHM|A` -Er$cKWM6e9hx186I}vU{@xADmv6)SK#kh11g2;_LYE!qpl -&5ozoB`Q6>cBhC_Z@Lyfa8pg?PG?{zGHjdJBP8SbmuUBF=xRK73HqGe-@yu>QjU~9Dv2cTX)ncJX(!9 -|9I%AbLsL{nY87t6L9p9;i~+Z0MzgSQ15_!@L&-@87Eo^dh)*$PU@H|mUfg|3Kp(>G>kPpkP@OEP?F( -`>~bmD5Igr2$gmrQ$bD}X-6nkYr(t!TF8pxI$8N!axB58DiYYIK@I}4ok4v9Y6P&|&i*=<^oO$ogH)W}&LV@&rUDgkr%MqmbX?0+dJ&Zl8`vyU4udGC`(f$KyKi^D4gN%xyc -|Sh1v=wfbTf)ba56VgWTV{>4v}BzgwcR6{#+$qwCmTULis^1NawN(_d}V^R^ZaY7+;_9CLxm-4#htyNmCMNjKZ?J%VkCBtL13dPahP!INCg6y%j4lTRThBNvC5FEV -YJ3*j34Z7Vk*kzQMb5o2ZYe|U$$7oY+8{EWfkUAgT@Up*aJx7dgLW+WAvQ~RPYCbm+PLajJ1j_!5M)o -ntZDxa2wxAj@FX8BByS$QR>|^wUJBmeR`Hod+ozk=fz&LaKwKbfEu^l~JD*x%OrbFpv2F-ArmMn&<}6 ->+D|6I$fsPB5!Z%mW*?Tf)s%{zLSh~e&J=^Bj*4EVZs#>7yRJ%}pdvg;h-;~obkIjv)&gB8gc^$dX@u -TvM_N9M!5{;5cDvB*>8}y#k%%L{PMry@Hf5wHQ2XyaKx#%U7aw9FzeA{wLWt;(qPPNi&_^S21+fHFpb_s%n9D) -fp)gKHcfXrl>RA$1SybaWJs(1(xk;gK~xeph_Z{;*+r=%NbhNn&7rD?!Ri!qT#^=8H_(Bee9cDJEmG( -LE$K{3ww%|Ki>KDs5}6W6jL7TF^*D07(dI1Zq=S9lJhMA^$EO;j2P}{(XDD~hzX%(Yj4zT{|A15xXu6OdT{&3YE-sI-hwh6-;N+@@0HTf`bY}lfa}S^X2~G9z6$Cy3e~SQ} -Yy8)hir&ye{yBzTp{mSTRP1fyv2@M-gU5nUOa!Lp4b_8H2> -OmY=tkSD?6-eK@!>$#F4>w$T-m2ql+mfmbG(nXE`Y-~Y|&2`=QY=iPsJTefh(iqqZ)V7UDeD*P4yeh| -v)t!lSC8xev4lmBCrkd5E7waV0)`M?VDnT$Y@Y0bwkV0%Ypid3YrNeqd%jV(T;l8$VI6^Q6DEcN!^6) -W$Shqo9T{i%#MwO~aW4u$c%FdSv$ybuKLN1GKAMNfkbnG2)wOJoJy{;%^m5pY`nP+-IC(Z4g-e2#1v- -9>9hDjbC9PghT9Q`9Ot|9}K7nb=L=|}^{mzm1N+-=pJSFf=a7V9@oFA7Q1w5)P>0mif@tULqA)t#z)9rfmP1PGHowt&UW7hoKWM -j4Ywd7urLr;-lo24f;d+?JS;R>FEBnm^pHW6OXpGk$Tcyj6=<&904Q+eO}i*ZNi#v%EQT_?pWYBA5SV --N}DNfTTE#0t8H_Leh?_QAq72as{v!oU)8m$oD7cRjIzRRQJwf8cYCJL)HN~JP1wzJfJd8jbLb=UFX| -Zj-l#K6Lm^CqX(dWcAMxD<$iRN**oDlm)pYlff$MeA?^i=?N-0?P`5kb2NJukoU(gr7FQct94%N-_;h -8<`TpvbU7jQ!8&Chzk;q6Ht(B&7s7g@0Bo!$*K#i4^y6t=TdMICo%BuE%6nB?Sk37wsjL25qSQH_0fv6{7Q#dHGG|6=EtGp^wBSc -x>tU{;RJo4{)U+`eR>!J%7Lh(Mq%}qgm`qwqEoh+2*n8TuiaiM7u3wWdkd`;{@)DHV`<9e4pGi^!kCT -ci??Cj4?EX1qcCvRv$ymjG(#iz!LEc|iJ9xlgwxspKY$fORJ%ZIQxmx+;udiSC)i-FdZS=6Gw6k*b!b -JJI?UBEo@@7VIuM^KobEBv+3{=0YBh4nyTkZ+TcIOeEsN@vUw-Yyl5!R$wEu$ETL_vVOw%Wmhbdx -?JNtDWkHdlgmwzGiJJOAZl+d6 -iPq*S^E8xpfhfM7nW8&r+7}Ef^xlNPnTE`oC;A)zx+u_i2#LiYc9ji7|YAZdRsvfteb|_i~{S*0}x;} -TtH!LpkQY%$sL!6USj{4S#)-rMS`pY#Mp0a#vY#K&jN@sSmVwx{!g5<%mxFZ0Z%EOrO?T`e9Qo=RRDh -6do!XkOb91?XBZHO=~#oZ5H$W8q?a-Ne3cbT2`sw?K1g$oS -B7Vg2n9bmjcjBuKL6BylH`l4B#6G!ld>+Mk*wZ$7Js@? -FN;r>+{O>UnV6oDZX{7bSD5*yovpW7v60MYpMDZVpMG+lHM$P;WastQ_&m8zg7=?)I{EaIVX$8`X^vm -^UVr+@$uLNpc6$#%J%0wpp)HCM!*rW!$Z<%Ab@E5eJoQjq6sZ8MMzy|O{I;UP~Iok;^!DDHP|2TPh@MiMz;P_;6_-g0mn}eg*YmyltEpD>PymUfwn -0__x0U!zfjln^a92JW9l==hLM0F~=hs-|@NkznI(`TVPuSu&1>z0NK}?!@kta7GzG-T|Ei%b-20h{Bf@8g_G*5Y7kB1kN5nw8 -mbifWWkyxrZxjEaChIq$3^2C(j@Qpk{xEUH5Z~1T|h2R4IsQ{}N3Em_hAq%V>A_$}g23^-zYdoJBSwe -*}$O`kpvI&F$H3^!5aS_EsFb_uRTU#u^Vw2FVxS`B{;sI=7y#0bMj0j1@h2~K2P|jXHi9>RClSY-C=h -PhCdrs*Oc#i_QwXtDpd(S`r)EU_h12KOD$A>#d%mrV$T(I`I-gsQEVR9oilxoPhFLF7y)f-bRbZuAJ)VtV3m|Ue=%p&}_P*(6HJ -*N(w7g!CRQWK70DfXP_ -CWz{fc4=gIpz`^C6-XFa&U6|T9eScxbxvtuz!s+o+?hR}CIL9XyrL5&FwwSr-B%5&zFBs005bixTdo2 -ZO1VzS^PEsGouHQ$M0Vzu-sw=})Ggyq6$tYs~EWdP16Rocn2bvol68+b~IuvPl9SCjXE@YXMwI-E**Ewcu@64#WFwSga=(2LAcAwsJA_CiY-d9lx1})uN(P&~Ao5lgY -h%_jA#a}0|pwyUnp%1Faoi>%-2K^VBBFa`?C?riT9Lh7$DYi|uN1lV -Wp`ViyGko;BVUxJ#h_l#RnwWIJW6Dqw(o!bOel&+Byz9#grRW;Gx$yn*?(457(7VaWwylOeJ2n??YPk0 -&4Y%K2_~YB}x#wQK@Lx6-HVF3=-f>T1#m%0=yYJnw@yhJ%%px7fakgpE6L)-lck(aovMqN%4(~?>Tz9 -`j-<$6KHGNmy{TaNU8rOLDFDUFccb}p0P92`t-<#?CmOFmzM|}UN?rs;uEH_w~tn9DB%?qf`Jn8jzS|OJDVvW!lvG-7yb9t0Mvn^G%y36AE@9n2Ij5|s -LFAMDfsU?q5q=3f_gR?{xSHK{KjoOynl39uMoD?d?8FB!`yad!-m`W+Zkrf6?)sRgumA>goAROS88C) -y6Z5;pyB-c_u#@;Hf(PAF?`duLYtXY@HQI0kFR0lUH1au6i%Z7#pnGV+>(FX!vFu@|C4{K -A&0E#!WaR$c@!5-+U|5(O9OSR{`ys`v<}HV$Jz4@k#eXo1VFHZb2HousV^)C;Flwig0gyI;w1?|zqu) -oI7>juO7{rMmLj2f&gSffR%reaG7UN>Z{%6UR5F1%Wm>+W4+^S#=1rtN6>Q_Ok`H{2T<{i -3}sYZiSXleHo+V3^CA~hzY#(wf!D8}C5 -h;ffc{M`Vtl9WI*14f@JB!~lAkMDbDFwB#Zi&zr0rbW`~KFJk##2QUfV;msk2$Z7%`^_-o0tOGccM8L -J}9N6k7wpPmY#L%r+$AJh{5ZRNWn5=2jC`fd2{cBZ -e&2N;-m#@nLIrPo{KgpBg?~s5qISF0Y%U0seY9>)=9&0IJkOA6TT?er1)clj?q>Jse=%SoQuQijd-34 -Y=euR5vy?Oc5PAPstjpn4u~;^|I|{#8~eb$ER00qMN@~Op>b~k=$_^WOVh{tgq|P`qkC80{Xn1$Z)H9RyuB-Iu93kefB{w{8RHb90v*D^@dzM0L4QfpsJ+=>Zj)UwWPp#g!KB?Km+`n6e=(_(I -h$U^=TRJI*yIY+XgS4pwKWiZ;eK(DPr6xqz$@QlfVTG$Z?RC4{5bgS|@~GS&yx0xEH|h#PzU6+!B)+Zs$V>TpKj@^id3|K -M4aLxlFgPLWI2;P$LO0oi;eH85*9eDEPB(dsj$bRVps1?6%`w8YWWuO$#JLungvAkz9O7Ee!j$0X=UT -pkH#WxxLZzh}Bip!^E_lnO6+k{uU{F07I4gI-a{x%8o`i8u9^*1N8P&Wwi<3CN?HiJ`+fde#Jw;^-v1OS5A11zihd{U5{1tR7-k)Ki4 -J`wRZ0Z;Bh_FqYkoKw}BnozJ0fZTyihmrrifef`2lhwQujURKwLkO@*umJlhAXxm50}3~3hTP5BIDD8 -A?Ek=~F?n$W`Cp=NH2@c=2nC-4BtAE}>Q*|^wJ;2Z1pqk|Y=;1o#xPtknCh?#`YYItEjj6tf<5p-+7q -f8GZ(!_OBjj+@9aYWFZm-R_y=foH$gR|F%>auLZxfS>+l2|AgJAbivh7N1i-3t6MS#Y)>G8H+mK)e{}s8YysSg&KTex_Y6z(^ -f0zy>wX~1QD&#%isOT!7)uddIjG9CHk79Uo2)RVv*GE|i^@L+6fq=#2@--mBUhM)f4@#*IlNgnn1E55?YpB|(Q7h^`qf#5j^9!OT@i+b^TMjqlPa|hBeVP!G_Dr -BrLOB24(6BZ;3VC~_0DGgG6)nt+>Zpyk@?MaN9xQph4Mn(5&l5Bp!Bw -W@?v4G&bU%_v5VP{9D_kAH=&elsJfW6h`+ADBLq&rc!%9V7p1L}1aH1(m}#U5YjQ{t-rpF*TajPes0uy=%>#*thvS(mhFYAD%_gQU -wI4xU1i_$f*LJ^*YQ60sQjV2%8yqib{Cu5nG@W%Jgi?=t!H#g6udZxo~aEo2+SY)v%VtC+3E7YYvGIB -H8{iqCjG0oLDp9*5d|Y6aNCAi~dJp5N0v4-IJD*CJphkf8hUU0?PzzEZqou_alC^3rtS3)+pd{9S+qc -)?Asr)39>>$T2$kJB?`nud^gFn17|5N}}2KhFZa%jlB-mRku1$)j>OsfD>=nw|vSjcUxmsv4vtPgewXN!*>kIVfst#zmhQfx -Hk1f)cc$E9E=73-&eFja1sDYsIL`84H}hSAh+3SrKGF~ncZddaiE-=mE6ETz|zf5&O$Jvy13rcbN`m^ -I&^R_7s4WFa+5JmfxzK1Ti%NPf(bjl)SWo{9GU222I6`E?W)w+V-JIvb9@Yiz$B7O{sam~i -~$vxA;ZwA_OS8*KnPZ9R)GET5fH3wk@kIpaI?@{VIgNW*+z4Mmc=L?U|7|cIfO#;;}p8UNXXUi_G4wp -gO}1?Q+9(IZG(Dqg&4L;uAZ0+Z3dNOp!jw5kpF`GQ4r|l5O6_FR8~>zlLZ#!^#qaziGhiAQVDTGRpmC -N$lt`ECTIh5lZ@AOwlrW%GqfX}oP^rS3@CsU4d^EI@@lKX);r$d$3s_e2UzUi{`xW-4{}=v^YDfGl3zF4YG?%qetysPW%UNH?Pxa#$vFIgPL3K2u -OvP9vVW808jqLX9Oz0|>)F25^`-Hi%%wpqUilS=q9xHtAa@l^v!dq;#g7WnY_GIaDz$2GFhWf}!xl5| -;1U=QSmR3PDM8u`W_oUj>Qe&6+jTycj!wol)V{;j%ES>)f0*ln^FdKOpqNH^7T@4>bB~g^so)$n??&5 -P0Pi?2<>Y`ZX+D_&9n5*Ymph0tE&kOL{JPsIAW^_Whdd$k_F<19X6rlT{4$wUc&`kj}b}mGFpwY$%b7 -QpCxF84QxDN(l4dre$ej%M_A$BdsM#=t5fc0s&0#VVtA%CPR57ydLEE;6w$z&{8XaGngEsQ~kAV1ZEG -!OaeYHBHwYf(@3t2oU62o*f0u}>q*@KFQt2!iwj&-h#11+)8HHT2YLgKPF({cb3z(f|OazPYsF~ -3S&}RdE%~$5KaV^I*1rWsZ6!P?plu-N}^uU8av7NooBS51ysPZ3sRUtQ2-A|W|z|s=8kBCZ`}&hM=QA;<1#`A@2jG8Oaipm+!PN2Ta&=t1ebf2%Ydexcq7b^4@e23D)Qc@`#)eRlI*GhEJ=M}g -hTc-;$8xn51kR_$-$ILWE^1LHw5XFiVEA5(BD6(NuM=3WPCiQOb+kC5ObF;Q8aYLz{Ao*ak_R>L;qI; -XdVOL!@*pGRlk2X6;WC{xUjed<@@(ikDz;AXW;v>uK$y5kS@u_`D8+-SaB09=S7s@KNIu&`LbIue5`G -60!CTXSFG@%oeWCiP!Rek>lvo09HlA4TZpL~Whs_L -CZHuZI-}6`IrDH3zTc%M8R`HVGvj4zw%V` -b`kFgp00WDY0f5dOY4im-o^Gwp94QAb90n9hmcs`huVA3tFXhGxE7zA*vNv#wG2$&p?gd|KhtsI;#SE -m7A-LrX^1|2?l%}vU^h0t{iJ0femrW=Qsn3+qo%8WoT&{!}QKGP(kk{ -j_Z;C94@vYU9-mC2P(>p?92rveXiUF<^VqxR3Gsij^A@Qej|nT64iFg@`bOF!Jm(NW -OE`q*p2i-)fjyv82v9Wy$j+)DHXx=vkm8{`-mvAqKWhQe5_dq_UxOzXff9Y8J&X> -X-LGHWaZnH7Oa}k(`-z+4ibBcES$1@=;6)xue3b!b&Syn<@(Mg9NmCeR87>h%415V1Ah*Hs&&XF6gF{ -$oR=A0ln6YLWl{_^KMc>?{E!(O%=S9& -egfSui^aDXjOUa6HU*WR7sJO-%dhWgXYTgK8jE)tc7x{7^L@*VC!wsVr)tum{iQSpyP%op}j8&J!OK8 -zp6~gEw~vp2^T0(<{Z$=511D#(@>)e7(y=AjT>{%YRaJL5?o -TlziR8ONXHnlCFGHYiK%A2iYw#oDwOK9jmrN$`7kNz!REd>rVG}4VDk4NDOkg>fG$xtuM?k#j^Xmi!%xKNE7V6{Pq))o@OAonE -=*s4cnw}pJxUifWtTZrZI(i~s1l>)8GvR|kYL&k2`SWrAJtYVg!PiEZ4g82(wD%6$`^1!6m$|b@Aq_h -6RQoA)n1L9Hc(ZpdZQ;*;Fd*oV!i#XwPZs614i= -in$i#R^=qBfTs?@>w$2-4P{H^y)vDbWGa5As>|rYC#-Z0wVD -M{HBDxzz`A`D@INVF?;{MVv*UcI8D7a9SOnSXLCTGnWA%=sW7q5Y>%~P3L_{EB94?`}UoVz1Qgwn~2* -RE_umH?=%$dg>xEeFJ0qjwLHR_pv#0Z!fGYRi0bh*?3VQ2yfepjM6GSoFSJj<0>Uzd{gRu4#CP!t@)F -ra=rAyxHnkf7~q??&Khua9J4fA_-}jhC%AH=3mN=77QDCGRhxV;eZx5#=5p#=+?WeS?Ke0$aSy+&de- -I&cXV#H$^A_~$`lMsmpD!~oP|6n)P$* -Wv=zTuaA$^=S<%}?abgZC`rlow%jNqmKrUqgRhB`tY_-WZTL -&X7&kjX!uIEkivV2-FmTKk^B?yPPxMeh8awY1}Tmivah=TPe<54~W7Ve2K6*AeW1{3dj+A+!;~x-8Wx`8kx5KAch_Z2g-KcWwv7hq#9B*!4>&T{ -c@W`kV%iq#EXa+#K-b)7lT5(~C%9rR>5vCQt$=hz=*&F^rEktV?j65ZeylYF<;uL&J`Lq&i$13zA!1+ -gPw1(u_6gU;+Kp|G6P{S1RZ#qoVD9R2{)j+?bw6jvOiQO|I-B4#1>im{$W^1uio!F_ru%Y?P;#RLk9@ -6;UG@rIJ=IX%(68ovLI#wtRWRjVAxY|&g!0=@MGaSqSl~#~yVSPf9YJ1d5Hrdf7sCw4+@Y5*nq|HPzh -bk6Wb&CX9!2@-O+D}0>ppc#ksTD?`1E{x9^;A2c5upl$HDzK1^8L4AD~{ntZFP@|OR1*=S!DkN&B|m^ -Slo_YAX#KXF4i7hO}a$Y$%Py!1m+(?w825H1q>oa61rtGjj=E+-9i?Bi6gnvY!KFpJ6J|o05zP7i0r4 -90wrkN)bEqRBGsN&Oh61A1!N9gb~?=nlR|Y?kJfI)$xroY4H4U%iYPaB$wD{a{}gwYqBO(jZ2%_bsKr -`6BO_@*%x;URAJiMMu}2l|U)DPhon)MWLUQm5G%+MsH%uqDhpaRQNLtD$h4Q%OqqI#&H&1GFni-K*+n -kn6@irE>)*+Mf*T@$Zsj)}ngm=VGW<)+o_E#G+j*sB{j8ruvsV^SSkK{ln0-1}(r}|IqEA;s6rvhN{) -Px#)q}uK?&zIHa0%O0Most{{=QNT28ePY+;q2uY<_o}>$|t`c1v?wpz#fM{%+DDZ9E`^Hpg1IDm9}$QfKcr5fu@Oi!oj$;6Y -XL<^{^iVXx|ZBM;av15dLIxG@n-RQy&SP38;`>mFT8Cjq;$1%L+a(Tg0kJrD72j2g==CXwn;|B6mv}6{{6l%#P_>vYVtK@v24(K1jy) -)b-m6q=qLHNmHBj_5FH2Vi~T}FO{x$>3iFmp_^chpNc2xcR*tVtD|+%-fSGb*j1QWu$1XSEvDj;?RWm -Dar?;5n6*Y!^-zKM}jXff^X3N~@6LB@B?s;IfJvZA^`qG|U5jsI=C!gzsm??gYRn6ZabU@q1Z5=XW@F -3F)!Qe4N*`g1uxIcf#D{gCF4*G47+rRqlyMXt7#H#a$~3ial8u&0#lWjq3k{+8lPhjGe=0iX#TzB90h -^zhI->U+a^aS8wHvq%+F*rZzE4pV}V%IyJS4lPOh6>x>iI%yJr_I) -RodbPR~mZAN7P$g&s7@^B`I$aJz8_Ycu_O4!37jU6Mn!PEw7=B&)*s>a>u4I%U>ymTEXf1v9br4k^hV -56Y?`^ddnxSJt&IX>dw_mG>R5}{3ofYD3V^lB7Kx2ujAu-#i3$z=m_Gy|6_fR@TskJn8cjHdL&giElL -7?uLEDM{!FI-#dPwW2=!M?~m+XaV{Vj#*dpjSi|jQkRiz#9jcF+bAKZ9&Nz|nUCd@#htkMgbRUmkC8X -sEPRbyrNVo-crqRH&!npr%4?%Vi0`P-kGVu;Q=C)<__Jb4#bBjXL!CsR>k7Q&Qv7ynUM5iEOl|uDnAd1mz2s+u*c+n@teO&1(*_zA17?NQN8Z3)M{Jf+W8Z+yPmmt( -r_;>=ad@e|EwPJzOLRX92rSV6+C8swCcQiPde@+)v_K^I5b8o6yY|MU?Ha_X*)ti6O)VUC7&Syp|NZH?BA -iW1ki_EztSC~?1N>cb1nl(lH~=96D-Vjh&)(*D=r~e(7Bwq;4r;*_d@t?K8P`RSU~vlPI(;E4ziD&Ap8=9{}*`Zy#!M|c?-x -EC4Q)KFYeuqSsB?=VgpSodmcXX=HPQ$5k5-`@mXE~&-#b+s^}|wA$?8Dr!Pw`ea*Jf*A-THJ@u%?2`| -~j?3qvB<~dHUuhU9-^^V0O3C&(o=;jy>A|2* -`S)9cX}?!MN(-H=n?5bX@Zg%Mk!9Vt~06Vf={uf~AVlwb$JXrY>geCp@U0q0{K<0cmwyT8&rPDJ7UnzoCl`x5Iq>ssW;vZ2Kv(I$=;UUd=#BSe-CxD-QOK#lCvSQLg1A -M2GI*`3fGC2zGES1qNKN2#awK_vpXk*1=AZ6ml(tINnGWe@hj=R3!QQ6(Y&xFA<5E3+^_c(HrN2$Eea -6M~aDW+Ipk#&ia^YoP$v(7yJW_charIn5!SbdY`0;MlSdm9lr_jvk}lj#5avuF=&5p6IJQsjKHVx7Ap -6MWEFp_NOE{r70>}Hl_!!kE;tzghN5^zMQ~;PMOlXHu7aa^V?sTG95{gam8)s&?2OfkWXudb&$#9{aW -6;V4$|e|F5YbKP>fk!sTCS7c$%)B_A^`y?$-wj*K7S!wup}hOkTegI-M+8scQpVuoWxZmNp6mJyN7fF!?@29+cKdVpm!;Y1FXN0CeI(6Y{O(Dqs$DX`{&c-yj1d3u`LIa4a^vRv8!0vcoH+D -n8fxKjNk%)6Wuhige_-!*w3#4tVw5>p@veBh@m?Gw&)cg~ -Af1RQ(Ho)h%rD+x{>NwpFx--ay-oSx`yK*7(3uXEITHKXEb+zhuNOoaysPeoRw{0CCs>E$E3$-k{mdB -I1)cG%X$z~K-H8d=nUl?@4yl-`T$s07fQu#T0GxR7w4lE#V%|N{^o9;N8|H~2OS5WgYnN{0B)67U3Cr -p3>$Y&n|B5j+p;00xB3J^_6E$hM;paf|>8c!3Oe?oWvr1fTLJuA(S))$C3+fp-97IJiq{FT88iIQcLd -92NC?{JvkdJLoqKVh+W?uv$QLsegi=f-W#31qJUjmiAh<3?OMwp5_unNx!Y5oSfiFk2fXIT>`8NIMue -EoPz+7bU~29+l9dbQcIO`BbI*I4dQV{ERA${p2-xOC>PD`%5aE%m%fT!Kzji*`&LQY$5n7?)G?nROns -pO4$GGmNpVqh{N#I+a|fB&S_ih5a9DRq)a({D}M(!>$3Xby3_fL -c!4iWNj|{@#ed`g}GvR;bpk{TNU1KQri>rzD#yFr?*b47`~h6U((_Q~=lt2i>kI{?Rg@w*Y;?D+eZD?{_Qgx0-ile3`4;sa(zIOa -ECE0(h47u5{&GWIbBo`rz&!iGmt1IArIAZ8TMnuivfL0T{e5stS*c`P)@-cGLRQh|H?`RXRCy2*xnoQ -w4RMG6DWkc%4 -OK|(fOJ*;9DOppz5sKJf3On0o)6vbUd>@6!*z9^UM(|wSRvW%?a8I>dM@zA*o&zjA2c`{@^ -%fOekfQlToM2zJpWesnw6i14-ziSUW*VjeGA_|Cphu_MoN1oORK;FX`=aUzKQa(Z40s3#YPkq-%mYsr -9F9ij?jcpz&v@B+}pQe!O(cWEYZuB_Fj7MEX@$$C^JY1w+#Sd#8>dkF*fA_{WCI{edXp&bKqgtk=qs7igQNr7}siJxZ3eNJ;mRwU~6Vz5lNO83n9!)eBa#Lt4wkMIgi7WR|Kep0R;(WavXwl&?*>alnk~$ -3|NK5m|~Jr_OEkC=RZncPi=7DX7CH%RLLqG?|Wt`j1qhI!E`Qo|&+i8QtD{7D9XW;^X)DcwxfgG0q=y -_5gwYEOVPd419a%pw%R>AekL)4SjcZSWVSzog=c@g2~hYQ!Pat+Ex|161HMpYra$Ia%+#&*hOuyterR -Ud2%#htSOH0GvSYgKlm1wk|Ff`g&D95OzB*-%2vUkI*e7^jSpqDx^dgZFm^v1;osgvk8zBEDkxl4 -?avZSXOh2;vJqNn22d$KBilBE3usaNsy~u -hW;mO#F6G?wy82+)!b!&j;Vm-^C(AbeH?by`zU^@_ZY1@N00T(QdRr=k4D3DFoC6c>STl{Js0R+j;z( -eu@D{&xz)R@3i5tVo0kll?wkMON-xr=;*Yh+`)k30_5i9)eZ4?tgpelj4&$kbs?8sOmxT<%ayv4wJ+y -pCBrDY;hobB%aJl_c`eQM0BdILQCVt3j%z$HveHk8uP*sg94(vp0GDNShUARZ3GbyyW%wk#z^5q&|qp -<8{=*c_%$W2j!sI{cOb}>MNkNggRK;>a-J5tLkrEFSg`8U -GdM#9;g|5~@Ra$747Ft)1-GmY2F#^c-_IJrOl^uTcv4JKGXo8q)Rlp!%)Q#=9=?t#Kg%B^nV_wM4wEH%mqrC>-R(ehid(bzQ+0dmigU -mmV=Ir{YS1v^tbbr$cJivL<&*OW~>)3H8{eBu~*8;cE0PWI~VmU%Eaw!grJva6O)nre0ZE2KhZOIdtWZ<#Oal{}t89&@}4@YUiHhsIExz^SlWuelEF?pFg0yJQaC)UgBb2#oAI68c2)aM#4pQX9lBqOq?fPj+1uQH}1<{Db-gOA8lRxtx;k13dF(_NpK3p_Os5+TtYL1@Z>Wfk -HBe;z1!enx6v0fV2ei*MORJDP9JEI!_!2WZD_v%$4teHl^>KC;q|S`Ah(el9G3lvL0a%0025t40Y91Q -9qP}FZ4T^!}#$1;OT+&R7M;;h)@Ur36u|1&;xU*B*&&m58*dJi0Yh%&)3(bHqZ41NCO(`P>ID0O_AmW -blveA`w4yJu_~W?{HT2NLjTaDVp6rx)Xp7Poa^y&2kuL=pYVk$SoEozDmPl?Dy!pMLk4L5rVky%4Vg> -C@Epg<213bx0Xhv|V8ow)26&a)lDkp6cZ+@Or~w%V12SbTN>shVFKvDCa;mDMp|lF>bp>5N)l%Fv1-( -OmBs+^DWX4dZ*lkEIB#yB;HKUBKN7_#` -?vA+(Pkpfg>tiuYZ{yrM91{kq()m7{@Tzau;rbo5-u^E5XNMjEJDi!jnDKh*_l%#oSU=16>adYo{xkdqfWYG=^~rp^HR;s! -)?e@}>bxs~uu!XoTQ2vK84w7WU5}%nYA=2}U^;%zp9=_zKlHeXvr=CP5)S#T5P2A@<@J#lUHaOV@SM6 -2n+$RPk&64d-&Tm1e4N6QwgsBYqEhFzOYO7cN7UYhh5|rSP? -yVEN0d}mXK2+rzu^gTB8Qie>;;d$@oS^$-d|JwQ%9g8bP*_^bfDvH2T=6XQ06$YCW`2@LvUzJ2Z?cD@ -2k{1y=)v*yiPxM&4~{d_gEV02_ftb0FA^#6xk*U@V&)FC%ModZS=zx$JFJmUY*~@e4%_q3KmRf+0eH5 -svY`0wBbaeluA$ZDyjd6uJva?nkD+la#H|p9lESFDufTG_(C~$V%(bCFeSb9Df)qJJ)+#AkuetR0W|-CLaAde{g$7=XQ(9ylTDqWTMNbE -$y+m?z0H+{;>dYn{8 -$h(fGDTLk%TZF*TJ0T(M08S%$qK|;^J+a&dY5nuaUmO3mkB*RFsnqnj{~@TRKLB963r}NzjcH+}nVL* -~4WAn=lyz2 --DFIqPklpYyOVCD>a@bxPvFJ_d}nP`_Mt!#X6RnxfUOGMIj7&@=M$T$Kw`UOC?0Uj9WM`B;?^U&xy0> -Weembdi0$7e{n(VAXzv~ZCrhnjFT(<<=QxLlV8M8wXE}dG1jua}NeXi0`o2vE%I+{}z|F8`!J`LkM>IcaYZN#Bf0+oG^Ixi@*!5R??n}@Yr9k#%1r7m -*aRxUrkTZQqj+7ZuCh=$#{s51c1noOojvjG*q1Vf{oLSRqI$bW8?IPksokDsJ(JD4UukN`ywD4YyOZE7_u0D|`b-trP%$T~Udb%oWylOZA`VATewf%N}TwLo#j -`@NEM0vqW;++4*oQBnTh@&)MNmXHUQYp -xqX6oDYS(s*#vgv2t_D9C)-qN=MyLEBUiWte4rMoa8UW-6aX0t~Zr~Qj++-Jx>eNHqDKgAY?t7gch?F -11hL&$bvP>=e71jxBGsy|kc0VSODq3o2M~a9$H6o6a)1eYsPzhZBK%lbq34Xl1%sd)x#b_}`X&UpA^B -^W_F6d)C*Xq#DqRscpC5|C}j!#-Z0@M=|%px!O8a=NUF*P^|()A9Q1UJ*8RIxg0lF~KxZ76qsi(U^sU -*VdNHUU>|0?lm)Ouz-?EsWlDmp%)7quMO|au3eJRx+nIxn@?1Yb#>gU0zb7YT~Bnf-!g)pf4eqTjV}u(wuSC6`i{rm -!>Q9&T= -+R=dfk#u2a9q#z~U_RTGftqIZV1yO2+^HMN%Bkx(bsZk{AE%k6G})Ed#k5wMzvm^*wcw~lCuvvQyi9c -^9K+imv7eI9APP!0+-Ow!EG#>%X@$8-xv#C#+#Eb5SOSA*&$@L3E{^wtErBt!$d{^{eKy -)t(jt0^8enlM8DiWdnBkNCUY+=PP#+}|!D4u)p1M>VHgOO8(aD(hxteD?j#t60Ut8A^MVGWeZ9Cg%A4 -J|faujFunxHCNR%V$4li-#apnK-xB_6qTZr -N(#-R%%|E%gX-`u|>XzE$QD|yaW+yRRfH+|0%{px;LZE|0@(nle$5ZJtftm2SN-TC8XL!z_ea!>~@r+ -XGY^Hd5HQploh-H+P3ai!MVv-Cv|s36DquF?)EtsdQWsFog&b4;&+F~cup)+c3-@fcJgBsaRzeNXDK_ -tqq{82?MP`BjE&?ld%J+h^Jq#sz*Zs4ArVTo&+kGD8XKg=V}IL0Jx3zO*`#q@`SC9gK+lRb{JLVVs-XsOm(Y4 -n$Wbmo>6NE?St-A6Mx~f+@pw8cHP~iI_^I#SX%~d(=nf5G^E{?11sO_0`B4pc8nL>^e=Bf28%qr=QRr -6a)3J{-;L?TK}5b7(Cvtoe`ieTdAW2lO1P-JaYRe*netZ%ANFRD@);x3Q8|6z*vr-Rzx`yFkPb+5>^e8V7lx!@Q0Hp0|T8$i!kitTt%M -57&u%vl~3nHH#&l%{UUSGTZUa_%UV#y?5Zs)Xm{xfVP?L-)mghPG{`^=HCLFL~#vF3n23B&vOn@I^i^ -*$)BGV!XN|IQcDvfmSt0r}UcjI+1oc%|$<3_%k6L@8u{2oqq>Ci2@odG~>bpDtNvwM_pm&_8S&UIZ7_ -oSh{Uu*p>l315*M}D|K;HUxB=uih#h#8p+ist?mJam`VExf9f1`aEuc!{J#^by2h~j)W{^ --94pFi>1xQ|tYC>^^1@Gp9rui;9$+&BIi5YY{A(yzd%-5zeEBc$aHQ10*2yx!`#-9` -3X9rMoIhij3jTc&lrKI~=YGm}~gSRYHz@bf^=Dg!t)*wAv=uTJbXCpK7tfWb*k3w4`Gnq!=L&s;OKnA -3aW_YMUd{V3J&T;RE@1+@$UT)u11Q;&k%s50dNJkwu&4}ChwNHUCVQ3Be07CU*F2goKwqzphgX -JcBU|!f?kTCQz&R|5){Bp8rY?o0VY&URzkDfa=JP{P0cINm)y3QaB++TJGj=nQFE%caSv_4xO7AHK&T -?kJjgV0E$GeyB^$XG++_`R@R?1HOu^E`niW%1;jmoWkEeE{QY}nX1CB0EJkC(3#9fAkQ@EBtLSSJwZ* -4U#T?{nt{W(>D2C5wGoAX194S1p$k7+|vn<-Nffd0zes>>Fj%Y%6PfO=uAlgBvV3#dbYI(RH1&0NLjH -Z`i5{L5{oY&m^bwc5Ho&7#b9)b@k=jS0^{KdIV{2;WrGLohVj4Dd^#LU;tWp1$IYG~5zrLC*3wtxw-J!97dnQL8@GkYmo5y{Y{ON+ad@k3 -ot1`N00rPL%(YZ)jv#kKv%35rUOD)a}KG51K~6&PV*e)2M}9`r^KtNYM*5~ej=?CN_FEooODNGC2RQ) -6+USJd9}Brs2-+!GN~7!XOzCzdE}z!O>Z;(yN+)Cm%3jCaidTYQKE=v+2Gu3pN*tE|FC==t7#lye5e$ -Osse)A0gDu`E-_rskfThmS1~Y!yyD7=J_+9iy?hT;I62F@Z-ijX^`i=CpNHUx>5hW31Cwvk=S=NasAbavEq -6^bJ0npPfloHZ7DFq9sZx3MAEKgaoMJZoiLmJT1T_;Eh>}9jiY7XqT -%_WvrB(Efu06Q()qnx{LPrR~1>&M>lwazr!BhRZ+yYgoTOC-=T8UkRz5em(m-zdTU@F -@X6=I@5~KX4}BE^)Sw5WwD7pTK7mRtavWN4xEJmwN+%+K3{$^<3@(JXK@^@(SeF3*^*1^`Kz{u`b|7YjkI*?@aPiemKQTV2C>TG; -tft@KdPx?70E0ftTqGIgW^65p%#kQqz)G=L{~jI=yngoDOOypDjKB+y~~E)sUtv6)&SVSMDeU1(Z9$s -MS`t93{U*rwtct6c?#A0w-R7IdshR5Wj<(t^mKY6WzqXc1uX%okmaldD|;ihT?ASSa)o -ZVT;u51&u>;4}Dp`i#?O7kv)U=V|(Ufj&FY)%T(Q1CQ9-7=8-EFq(?V!M9U7k>M3pZ8{W&6kLN3zJ^z -J-?r&IcJEhs5B8vAgnx%SFiF}+XdU|K^L6^%i_a&z;aTrE$+bL-2Y8J&rgPGvXCSVO=HMRQ2GOA<-Su -L8%~p;(aFpBM)YM7?3}ft|QCCWf%N(5 -7FFX>ZpU3dB1Cp!l#0t$mvKZ&?=3^%=df8!?-fbB -voF9qpySPK{bpzK@Nk8u3TFU -9iy+I@C|Q1IGfW)6yeGbZtOb+?(nB2 -!cbD0ZWih!477Wa)+3)X+pXlFCU^Mf#9|rK_Wq)4^pT?t?T8r3F&DplxsO>MxiTY?S3W)&thVv_*@^Q -%03JPRql~-(4JjsEugZd!l`DMo*l-u@z!AbFy7XM#kyts0<96NsJ`%f2JiH0*Y4leeh=uJD|t$tllI -|AR?_Pwnx;W$jnqU>@3nYxVRTGT54kEc=%jXytjNyX5k?AO9)A3j#Yac^qxV}AvQQ@-Vqf5OWTP~6lh -gPcBE6jm65qr<`ZsiAXn70Ijm9@}4?^OhB -(fblU*UYd}CkB@e`x(`PZclfBfZy?1U5ZgyVFo(~rmO3re)3#H78k-`%UeNp*3Rr%cNQVO>L}eP?>@AXFAfQNF~cLvK{@hO28rr9zz09P%*p`!fy}LTj`KneE_6)lnq4zhDXOXU!Z+BHTv*r+>xaT# -mpU8ns`!~q^ALCdDD|HRi4nG^k{x+@WjqZ+s(o7=5a$^;t6KxFQYBii+l`v>N~khFxpj}qK(!u95F?t -B6&!=xQoW`NYgnaniUh$E}?U2Xx}hifkxxikTwPFOhX9%3Yik*$y?k$4|xq2=^`c4(+?wpF8Mq0T-lB -^QwRyunzKSV6bd4u;O1cIfUP84Q~#rJ3mtQVIKJq>S6NJ@6oYvOp}BNMq17$y -O*u0;jhv@&BCQJEk&6UDA91s(VNNKi?M$y9-@LxxKn4(SnuP+KAR#q|#nYwu@$bk^XLkR!wHRU2hP7x!$1fk!;n -##Uv9Cp)_Ln2`be#$@P11x}$M5GF3BWIx0XLu@6c?O&4tgHtOuX?=0$|s4sx_ehA;`T+nd#2i09o=%! -q$Ui^1-r3&4^$XoH~kCQnBfB9RS%m(<|1b_SC?+=)kbm^08ZNQu9_Hb{Sfo?f5!W6PHno_fW4O+N%m- -y!l-qV8DCj!J(qY&Ft5O{TfkE#KVkGJHJtWlU%0Mq-ZwrzwSgwOv*I|S>*3ubPVSF^<@xdY~FF1oh6T -`|h-^yY+%uI+9!-h`KSci^Sn?fjp)1D)J~NbynSI*%8fz5E9Hn^{X2>pl)b0oi#}y9igS&!~%GOtFdK -5lk7qguL$JTgbWK&zTf$N{^;mwc`)TmbcIqA-sV6&+_{!uRo1c3E=4C!|_G7x*T_a7S7e1X5=sxg94VO~tB?iccmo?Ocwi)=D0i%x`{g_rB9kv#GeQv@WN -Zz_|1I%Wj`8)9TPT|igMM3`UKOl8Belp@LCT6KfdAZfBU0)Yf=|xjSpiuZp*D-pM?)uZ8r@v9)RrBT? -r$2+!;rJF8fa;?4WjQ?kg_SZ^i0kiuPjOpljS;4w4?jwJMPx(XBFr0GzqT&w@%qT^+I`T<=NDcU#!uh -N@>A3;EFm!9o9nGZ!?@^HBc(}|tW;?p!v-?bynr5dzg*mAGYLhqGq=3Z@~@Nmyz)!%l~Xpm -CVRHHk3cYFkvAhR=ELp@s%Crk0rx+@=3X;{O8Fqj#H(s{9lt{yzN<8Q65T$aTiQkRQ_l(jBDk}mF2%k -hIe6jdig%19u8qRr~D-hr$ZQ5e$_1-0<3zXM%uAKNyi4R#6aC`Yg2b=;)$U$>Qo5?|M!zX(Neq-2v0s -8zPZ~87SO`Nt!zt}zcQ%uMGI#N8TJmajvkN5=fD_plhsb`mrTGEQr-p7^h&gugtPQ$W71pGLMi|=jI@ -H(NW+UX>CVPNZJh1F-=7NcX$G48OFGCNJ59NR;P^6l$iJMX;|otkUG -D`+uxwiLSjjT5oK0uYBe`9r2uS%;IBp8)=(_BOMi&-HCQ+#V?~ShT|4#k_$V$1|=COypQo8(id9^i$O -ZIquE1!X@elFH5O#o@tv}mS)_Yk@gFV5s14_D%j??N>l#3;Kft<|_hT5a7mo!cLESCX{rHF8c+({Zs2 -rq^ogt~AwUZ8cOv0Z?4aDgGLIrS0#Uh6{O1({P6HAArf+wnSxc=?|L8we&I!?K$4OFhfYIRL-nWKBEb -p`s4u>bZHf|sRsM0rFcIDA>-ck6+e>kuaR$8(+#0GhG%c7ZnB{&Y8!tM5WW3DE~3u4Qr>AZdPUHwO}4wp>96Hu+w=F{Kqn+HN}H?q7`$ -^bvW%z|r`3-Se%BqZe33EU|0ctVW>B$NI$EV@&b<4u;EZ^V6>GpWNl50SlHQV?FZj$yPJ(`$+(B)n)PMWGP08lt-;(VPTB~NPCr+ajX58Rodf2R&g>X%F5@+$rYCNi -4GcVxxq&R1W>A98oZ`0QYm&lu|5gST5L8@4$Fy%!(~|MzG3l+2fWmOr+bE)i2T8UxrRfrC7=sSDn83(Ak47Ta#DrmugS9){aQE@tEtFb}2+Bp -CQ%0*XB}fV1K*b!!g=gtA<>s-DHNwid^ucH2g7ou$+S~n=pRJWn-M_4F1&7Y2Msatv1Ono7inCj^gdJ -;aA|%#4oMl=Tz~91i0-}(C{v+sAQTWS18DoS3t*gzNm@N1yK)0Iyo%v -E_`4kbroo?0Bu?SEveg50sP!q+V=la-i&666gZl4U;e(oMCaD%CSKBLqn3;8y{+y+c9+`4PcG~$lz|c -kbJIdu~D)Ot!89*ak5p!u?z^mZHOjkDGgBGx)oUs7l{BrzFPsnBOjR#=N&D8^V7#j-loY0xx4~@?H3f({+M>`M<5~B -F&GAYFcMH_g~8Gd*fzmwVlVg@V)MPQ0W!b$L -4a%GB%v=I$`GE;pLLmrgHnHWoD%HnmotZ`ydisD+09^DDnF)U0bFY -E(NZ~-bnyc-g41%snkrJJ%{3_2)uVeE^illt$Jdas+Hv;AGe7Suw$qQEgL987>e}8+f{I>DEvNFTI1~;TY9cBZ0R;j@QP9)~1q4DsxYwanP$0! -Crp;DXrk0JbnM0EbYR)-igW2Ga)*&h>P0`5xuV7fiB40WDPy)M`3KKYeyro!@d2!w)QJvhpAu*2Gm^muIhfu$ -S*annF^-!+_r7P) -K2{w&ZKsq9p>~$Jb5nVyGo}RxU8Nse8}c{mHI#w8x3wyy$W~bvPlOEuk+!yRY0l&)>a+vr?bKW{`v-F ->e(IOhp -mmIbNk@UL~j8^G@eB+Lt%y53$h0)`4sDwg^?nq35o79_AV~$y}~UW;%+kUYDe~peiBJ>5LkNZ-Q%ys) -}k9$5NA7>$%+4oetFu{gj{3ku6(^k(YkVF-bs>KKPc2&FtSQ)Mmd;71}N8Icgzqdnn;Tv&8eV6QHe~W}qFU=2wzv=l6NDT!Dkl%RbW6K -FZU7q=(IoN!t-K@_jFA$;%sS7CX|q+oLGhwdj7&F(aeLj+;1cJayLP;T=l$MxOrgNV{3Fl6BlgH#3EB -wl+(1N+AKTq`4?}Ea&n}wuwtfCLR2a!@wSuJY05AYCqz#E8yyMkAOo4_A;*=X{WCoX&x7w|Gg> -yJ5#{bd*f|xWwZMTacMVSP1d5c%sKzG3w7GFZ$H0AdOyDOaSUg?hmZMXuGyZX&$;Z8-i50lqkQ!)et| -wOCDR+`268(|;7Tu@j!-YnebG>-g~OZ{4Rxky7Y=jI-{q3Onxhp(lopZ -s0$gM};fb)hki1!PzW(g_j?CJ>A!h$ILB4IxX;6oou|hD?P_h0Gtx{E^HGGAqcuK;{K9E6J=Rvx>~BV -pj|i2$u?%N}U@LM6V^-K(Gfc)R&6iyJIOZG}=|_H5bCM21ke;y5C}J@2cY2{#9BYcvt5^lRwU{siJhd -KZ$fNeLItS-&y&k)p7I4A@a1eFAWoIMD<2s%B9;+Z_4yp2blWBh# -FsiDl*NL+a1W{JjOc9*XmqSa*=cSVFOmq<)7HM=l@Ph -gZ)=be*l@L8Zuo8nevYIUwnQwZT4)SUFwKaR7O}`?zWX+j-U_W=^EIt(o5qa68Jkz3+i$WrgHBn!b__dB=x%N@5Sk@SUQH6a>EAa#A)w+c;zDU-(g%VdKx)e>k;6L$#_4PRJK|^nY3 -g@jJe|C}!9rh8$#Hb!vm;8)pfYjsCczC0bw7A>tp=LO~F4UvLA6}&*-p&}1S3le5R0jM$cZdP`en38; -bjHofoc)A(Rlx7ol-=&YJb$-Gd7EhE#f7zzSFb=IkAc!wNl8=~QbBo(iSeIkwz5StB&BA##Z_xC>)TGx#*e=$sIJ5#WY4KJoyV1&7zRJ(L#+>2aI_^ -D%ps*CQGfF|bYAG;73Ru}EYg`(eckBTygHcmZ2_3&&CIIBrgENoA)URY$ht>*i2DIvYD=gvzes -?v1wKO*_^5LW^=Z3yN1FhXxK_yxNgE6mEy%#oMWQ6u@&d3Dpzn8_$JJJDd*UVh0Mwswmt>x7i@hN)`M -()0oEOCT@C9yY+VcM8}l7!L%cZOak9gcY{jt<%3`)+?}##wtvE|SnLb}b8Pgg{3V%SqTbamK3~DH&*! -mT$!`O;iMj6OfO!X9Emq){tyl8r94=1M6mKBTgSqBimel1{fw>m!MZ<>`{no6*img -AXKxVT#;1uWGQZpe_0nW>@5A_ -Ui3(>+2P-F56a0}C1x~VQL0^A}smz!0WRsn9&nv1LI(k{SlwC3__F1wfm+{S7yCsh~k0JjO6%R$wpe} -G$-=JJ8+GBCi+s=2(Xx>y3-W@;`^sV*S_ZnHI)hg6ra0Jl8N#mX+zBLn&tXw*;2RjGIMr(?BgOM2{MJ -keipZiLgmw>z0$bM-F&g&0H%1G$qnzo-F!+T~N}3pHVFrzB^YgmKOeU#;TpQf~>WxnIN;hp}tdix0Fr4{);eB&50wNN}=sZgj? -aYpLQjAjsL)njD)ANN}-vQLiF@yKAE}wvX+kk?mO8w&cjk{1T@fm$vhWOshrCL@Pz}JrPKOzPhv>Qy< -pV$#x9=(1(NMx(o4tNbzqWU9fziVf^O?_-B{QI8U8%zZ4IZ2F`JJt1(8W!XZ8rwVPkDf!IS?KZg&qE{ -}>U_vLf{Vrz847V1aY-s|s%Q~#`%SmoC>>e&Og=3hNIeSfTS8t+QC=c-Q<1XqK9ab51!nmgl -%H1Q3ZwhKRCoLC@_m|)iiy*J{nM+@DfG%QDf?C!PG%1>{VU2VHqQQw|3nXKV^Fc?7s%l;zOmUus?+wQ -?zaf|CdbR?d~BL;U|j<`<%u4!2bN}OTYeeyjUT9vjNt -2*xDaH0V^>)iek^Q}$_a44nw8ER>Tv!}Pzj>FPt!{ajoor -pHvvBDMCctl1zJFHTi*w-F>iw@2ec7!P_r;L@u4bzV*T3S7+>cuDAb6}Y`)2Lpk>xl{-Q{A4yKA1QU@ -^`HYxWXx#**&0*FUU9aUpJDGCfurZFiA<-0i=}mU(<`B+i(*k9-IGV)g9LsF@0)ahzbw0?5J&W(f!OD -=8}M4yJ^9nI7uohLxZedRrg@sYS -<=@v|U|-;c=<>TwTT7#rrmfXE_yP7}r0KKYCCbaM#1gQ_^w1ifk0>q-A@`VQ<&jp^L+_7}UKlJHk@B` -_PBE&dXeGnoCxdTdkB&}rD~`HAo!a7J_!${3H-aoJw5mC)3w#D))37LJYNKO;qrDJB*#)sqJ;xIqZ-j -O29N1nAX>lnnDunF#Z$v=T;)_^+z6JDSRk|?5iCL-YNOy%^Xi_8Mw>ThtI++%)gQ{4F_$9|)hKa6$PF -B}}&bAJw&3=A+^RE)8OR-6lk#R2C2>ObOk+SA-lw>#MpqZGQ9KcL66^0`nj?`Jh(|UVald;i7{9RdD! -*e^vqd(&TY^e72=k!eTFH^su%me$9n=Wdu`dK*I)W_bmD8$si>>Tb--wV;~57AsnuX$S_Goq+ijT%as -Drx>vBwK$#lK1;v^kG3LE)1jJfGDq=E{mE62NkVArsZ8}L{%K)u}d_u3|;7D89MK(Y%WChr}hdph))I -h-A&CycBwlXgOMEjb7Is3c9|9jz-@;QR|o;vH=0glmvwfa?rbXjs3y{3*srn%{2<&_%ptN!_Ff<#Y}( -x1gC2`m%stHi*tr!z%Yaaor7~qcm@KRb!nu@5VNW{;ROQ4Wb#9h7F=$KkkV6?Bn%r5Dk4o-ynLakPia#O1Td-DU -Um7YOXx;xRI8XHEFp?$*E511JaGkkJn2>1E%3%j8a};prHZL(3`M_TeNIBTU|`iSj&1M#x+0bjNm&uS -}U!ASk##!52flc2Uc|umj1X-*lkSM#US4FI?F>?y=A!T%*OU+3!SMY*i=&PY1)b%VF~Z@8ht)lslg7} -;7@nI4%T(x`+^=hAkye%$ -%9jBCTJdEEhXYgCNmVYbwcMtz6{Cj|ZSM%>q`c;t8S&I8n}3F&(Uwe$EJeyjYu^qT2b-`Gy-@wEiO{`) -egk2+O4&B)fOWXbyuy2aaS#zv8(nR7F)QFBZX~&S>0EA@B8Y$S}>Hjw%EYYJq8|EICxz4J@(aZt+lT< -EKJ*1yO;OX!h6|P-mVI(YuS2--Kp#)-d77d#%;=;bbN9yvy(mV${n`u`MX`!%(v%O6>R|~h0;`%ULE7 -56>u!OhquA*`9s|XyH(o;%bx;yV=d?sWqw+`U24XwXz{k-c$o!d9vq`mD@MGR{p5jI!mVTSuB2E)@~) -&?TjgDuVJiI8LnnxL)$W3~fOOlA9STWv_YY1bCG_@35Q89wqxCmx29kx8*(P)W_C}6frGqo+ -E4~`5EZPKJHX~tbq@ii;Z^dBMLqeO$DJawe7V7Gff43VcJXuJz&l>6?BF<)>LpSZ>@!FV23S?;zAU@X -w#mIoe9gZsZEm=`AeSUS_2`8 -|EvseWBuZDo?-;C?j@@P_c?%qVceC*GzRbP)j#;R)gm*4>5WbB?P=r6$CfYgHIuf$_nl=xsQqNj1F1+ -vu?*T*Uk5){#XuDyp?RRteb}^%YTAx{vG8qe!D^9#ls ->LP&-$`%u8y6tCmE&fxHPv6XY2FM*nYgHx}&qUGgj(5W8dc0SE<&S*cp2V?~MJHcgAuFS9zhBZm8~&jo>}9TrhZ#EW`Spg1Cz;Dn(J!`7;yc#Nl&9I6sm5+vjxNTy-!^Xx@WggoY4l*&ZX3YcZ8<}oxS0YSF=acM8pfv>4R -`q+yw%o96$BNKg7NxR+pSJ|ChX=0UfpoZLZfD)y5Y9(Rg4e7gX_p$g`c2=FD(3&L!oKHz*t$ctDMoi* -Y=jHT=XuMnysx=hFmPQV2|xE-eWtgYI%8c;v*cj{k0BHb?vRy$0zuiQO6k9@mXcsHCXt(;W;7L-Hc`8!T}aRVFsQdVGdZ9_ -G?dX%@pxxU()Pq`CD`0*`*6fQn%40Y$LMM6Y_BZ8)!4eMGmlVO@m5&8N!~PiLT*GT`x -a<^_7&`z6QMemTckF>>GcZs+4fYet`zweq*LBnjhZxA-x13D7ivcSZT=RnCDjbF+6qhNPP1=RS+Nvla -B~A$3`=xc1B}ZQ7cRt0{w`;wZ!@iCxApU9S8GqXabaytISiwj!$~79F>ajgrSeYs#f(RJ8#Mrx?~_a$ -`wrFi+TumbzEdO0db>u{HC#J=z82T=x1b{8t+uczpG-lwl--Fy+wSx&LUdAWf9W|wh-(kFp+!E+ZLgox9RG$r)|ll!=3incXk -#Pq{{BeZe%YV_n2K2xq9#D!FN}S@-l4}SsRTl3azwT)=KyGv%!XdRNN6+x~fZ1n -;V?8p#_?_$(#N-tV?G*H0I-HjGwQVS>2)Q2mD}QDS3=IKGJ?4}HqV7O1QKCw^5l@g%k`RbI8VQ^FRgJ ->eeJIP9<=RePBw)XQ|!ddz0cpPR#lcts&9|-C`1(?!emP{g6%rmzW28$^S*cgu`O)3sh%2`HZ>(*#(QsY&)CBUMMkBK{95H)MXiE_n??e$k&f}G$ -MnMg#cc5#VM8G05Xa?&8cQ;rosWJ;NMeMh#NMhf2tPNiH4OUSXHOIG6^U9gnHPcD(ppAV_@3s8q^t@Z -e1%z)T8jw`x!Uxn)8ffvHZJniLIMGl5ffO^!xbIFQjgSczcjpVc&5nk=*y#cfoWyQa+6A6UYU}Qi6F=eKW6;+;>UuYX#CuVpRD2{D7s -;`!-XGP@3L<%Ra~v6au;fL5WYCvE6jE=|Fp^4&c5C6%hDg5tsX_Ki^E+bY?b*ZU91fjwhjujT^T=ihy -B>4FY|YJ`5g5t8D&43e;VJMf+B2JO3M+3VsgFPR(Vg^7~*hRBiV8$u4Ke&Z^zDWKF@G&6P#11EMCJ`i -B#A_+i53Mw^C-Z8*ZM$-OO-`HJnpz#gTqD+YkEevALx=k1cQPS5mmww7jG^L^xHi8sPo)<=%08$lU{5 -^{q-?_68&zGjTe=ZCEkbnok*^51-k0`M4oWX{Ga}V~D5vy}UwgOXKX7ekJxph0+@7?{00FU+QYf_ZPN -Hv6eK~XywJ{dEY2@V)Er44V78(XaTUHxuwt5Xk~vxTu6)rUGKt?(?8{R!Dh|sRlPrY;Wld=C4V7c7x4 -kh7qqdu`1{pKQRHBZy;WS%#HxUAtVwBuVwCn1)alxQtJX0AC#)m=N|OC&IP<>4%9WrXz^#gYZH<(Elq -0LyWtVN{O6-gFw|QFq0(Mz@DGzU>Z+vnA&halSnH8tBW{Y*WvLsi>1N&TY8mGUll0*@E6nEL$rnxE!& -k?1oDmZHE@giS;X1}dDeHzN2)Ksn8ihS*S;sPQ6whuO(R2`8;b1IafSDA=_tER<&U`3#xsQ?@Mi~GDU -`za&t_2p87Et2Y1f|RM_rK0g1LI77N>kL5!cMvDy>Zs_nQ-Wr{XxGF7!g#anL -f{by!hTLf^lE?L1c8$Usy6aFhtwIPL5Q&*yYB3i|39Q|3c9iq=M$&vCwVqIj5dz;4sxLW*&a8pi{Cy` -8HT$DS4HZmV6Uf&`>GrmEQUouXz%(_F!x-fEm*;F?A}%5Ge*LHJ&spyTUTQs#42Tisf{6mRTAu24PgQ -5P1CX;CoR=~mlqd}!!RHLBQ7p9h}Lq+gcSAnY-j9*X-3W;jsmw;ZX@Y><1Diu) -lumXub)%MqcTuV00$BsCNs!^(g!`M6{yW;Txbce6EC`pwq2e&!#PROrjM1pJ;mZ8jyFbAL>Av*$;h2> -oz)Q$YxXiChEjSEo$UDQfjP+OXwCp%tCB7mc}qiHzHo&mYvIS57=wy=0`Q0pZ%ob|L>N#fEc5V|YWXf -=KtAHzxH3AHU1;qh8mc%}%i1U)(~S<7uR7-7W_g1$jG{#F;kDay4x&Q3q9yRL=XI@70JVlN_u=$i!~W -E-;;i^){HU!Zs)`+%?bMs5r8>_o8Nbr|0GEb6VJ}ylYL@`FZ=C%ayzY*WB_Z)nD>D&ImF-9JPvw^W;BHMSZ&3V<+SF -z?AQ(G@P=Xyoh&gGvnS-tJ0I6b$C{pO-uil@64=ema5&SOf(#o7=Xdxj1Rw -_O^0Z~Vj^?_t!)=djPdk@oGnhnpAP!X8SM8c~GV&W|g@T&>#HjN~Jzt*4HSD!Hc|AA<6>f!@RzxWD#vE()wS}`OcZrh6*U~pM4`F_{65 -31yfG7N+K}ocQmWyvP*Yj?+S-K;H2qrLf6glJK&hSu0Y7|=VU*{{n7CMQq{YaXqKTfY(E9RrP-_eSiJ -qoH-bN7R5wPF7rO2uLU9~jkS3Y1W#7)OkeHG+LzC-vSD%ujH%zcSTDI2e6ORY;j9%Y5Ng?<{7dbP&H1 -mo2jt(C(wj8|)L{*?`K(k|C{dxqn3jZ>5{n -zs7sxYTU2U^A}$0hWj;+E0#=EAvyc+>@=}TY41oEH*GX_u&vzPzJs?!S6{UeWNOad$hXaxw@CcTVJVb)OZhekJ=jIsP0RrsE-nJ>6F$@qArh% -YCGj}bUb~MoexkP|NH#E@T6Z9;sJui1WyxeAlOdu3Biv9u9t)`6WmDvJia;LI|P>CJx!3Kiu1cwQZ6Z}fxO!`SH0$+lG1Q7(W1osi76U-%8O7IlHT7nG(y9iDaTp -{oxJi8EB2%-s+3GxV*5QP1q^3;`Zt0bCFSE}!8$$X5UfFPUT9fDnz*XG4{oh_m`+9E6yETY$Ti@2-6; -`ny*Ba3*li{;ut^fYer^CXxUcZ?zks~|WdL~v$y=*(1Wbav+Cq?{ZfrivUfS!9ZAkwkwzMWRTjKOwqN -FENi`GX0S^F-D9QLq!kK-(c@8`iV#p%IS>}!$r6V79&ImyWCB`qsT3keE5kwDdrHeMY7+a={ZBQlagS -@Bw1q<(rs)XX0QjPr)N%v-)vzM8Nw=33DYEDCdj5g;A`f%nutbhj;6Rrh#2}a`cc2e5N5;0usXhwTv4 -bf5yU>~`i^H5(dkQ6__1pb@)txHS;_ZQkwvDNAd`M=#3|qy#VJOT4V+D6w`7iYD*H_#YI4YyK)*R0%O -v9YEOJStII`($hL}Qc8RP?$Pv;n_nvLVfcQd7%$ -*JO77X76=LZ~tJq%Yx2r5fv0a?Rm)_`ILuRdEMhz%H9pvXLwNWD@0OjxU}4tCS^h$~vEVFm2#vD)~tw -oXw(}=tZvS6x;30JxIZ12Tn$QCe+EP48{@CGnsQIn{mmm5i*2vilVToOsAgbdVb&O$nOjaVP*NqU~Ww -zTFsO~7KNRx<*yohI+-f9dQPcdrW!a}JDuKtB3IYucankM!4%JEEnl)7UsMifGS^V9{~dBsn+{(G9Xd -`r-|C_uu~zsysczB3+c-xjX7r+DHcWrsSQ{448^KS-Ic>Qu#S9<;aMtYFa#J`zdUq^aboKuK*I{ -DA0P&SUEu9Bmd#p)RK`1T=65EgmZV{rm^CTljXov1Fo~Ey#qAcfPP-C{o-f;Z7#byQ}!zB3h -IA0-3m+)0fP)WR`qp;Wg=5WZnjMGTW1x@~K7DaLDXHhSyOG$sg&tA6uwqCjFJn2guw&CN{OcP9{c57L -yspWh0y0X|%tfEtczky;cYoKJ;fM&ehbeBFQ$1LW0|~nD#~*AA0g1>Bq!3#5{$4P1V|7H8kS7pSd}Me -g|uLQ!|B7M~UdbaYxsPZ2;38Mt=k74evRueG1AE^{+eY4eFoS;WA766YSoL-Fs;+q?5A-EU0x=%@8id -s@{bXif1@sF};@C2u*j;U1G@0(%NU;Er_Mo#`QR!2(A-mIqax6)U>l$vrA_j!QaUOz4M{;eG=;_)jW^ -j*1Ot8#aG8s$4kdY$3ah9PghN6mLqLFT|G@b9mJEsEk-uiCn^p@8Qbc2XbP --PxE`nu6T`R}-xHV?H9b8YJ#IZdJsv$g7M>26I(0w}3r8Y7-OW+-m|>YkkBExum6+IT7AF!-*&aj?k~ -BGK>ZByj+~Jwi9rh8Et(maKBxPB3#~|AjTaFbYw&@&4ge`+JGAJuMmGdwmdot%=(xhx#LbmEBA%m@y4 -&q0=h!b%k9>hW6W+4uQM_7arF$q>%c47j@F*G}sZKG3c**Th9TvB32lJ+v%rrM%1GuR=` -hd=nicj4sh;_BAG-J@Zn#-2@@n!K9b)VxK@n{R2=x=q`5x8By?+@WJ9@6KJicDwx!pYA<+_VVq0r(d7 -G{qE}TKOkV>puvHbpdrB_p+kp-g-6^S88v)F^vF>$qsPR?jU9K-_z4s5jlVBp(&WUX6I931M=u<-ioT~qvZ}6{L0{*Lie?~w4TF{^I&G_wxN!IAD`F^AHR?}VEg^!#UdGqGy7 -Zg6UU}4cB`{Lpy4=-Kz$fJ)v{=}0{Er0r%XIDJ;{0l2zT(x@5OE0f|<<-|-f8))6t$SHkls|8H-I>;L{gg8y#R-D#}Vh_{m6RkJ@~u;XS856! -;BV1L+PUuv)~GuR(7*dI06aodJ&-)OMo^pb{Ze7ss@Sdu<29+#Stm^mZHnqW;O9g5t? -8cfHyf?t4w3l#cAR0J(xrT;@y5Z^TOMM%7HusMn>+T`$^7W02cBkgX%p`i;mvSFHLd?vXNskBl5bnam -gA5P)bJzLMATRHsym@()fXhR956#bO;pRpu6c!dzhWYyDMdvLf-`%?QkB-Zu@RUe&G!B|FcgtItN1+5 -Ilai9+;o&|rD{DqpR*m>EBQTr*a|T$)Xd_1Y-PtoSos{TmzKl+tnv`j?2G!vfhrWS2%8_29|EP~sM$# -8`j4MU!lMPLMNXtrjWGqaL1%txuNdBSmQ>OMb_{mO6NHiyBXHF&G99(qjfc|C>Q5#0MjwMrw0FBSl)aSVtTl8s0a|q0z; -EBkbWS&e>L&nU>j4ZpQR*Rl_nk^}NR%*tS`o8BRS*?c8Vrs(uc+IgU5=T-qh#jd+vhJb!JMp|O55gEn -Qg(Jde(;Q`V>EBFHfm*MH+2kQGWQg7ggSGgkJNM(ohf7vnHUXF?Lth>%&=x>rkhg|G7{61vag}9I!~u -0St(0Xt>(#@iAmRkOW|9P?N<8e^ljN -o=Y&j`OlM=`=JvAd~fVp#ePMvYgNJ+IO%$8b7H(Td#e~H4bn%}a)^%3c!^cd|CwzKWTPVW8116w7&6yBOnL29%wYg -|mb0*}ZPR&YBici<->j{aeIo1g?sI|ySoMcSHH-zfap+X#fq;4Fri3jEVmo5M6(I&y2E&H%DEc-o9gNem_v_ -yZeOehquRHpWl|m>&L$@N&SxJi6l$C@gz>Wemd)NuAj~=k6b_ekvFdY{na+^gA|d7^)iP -k?=DumeMWAHM~Ei|w-%SF?K&dreExT>(|>lYQ)?&Va-;=L)OQM2o32P|x#GmFt}#hayrUDWDRE*Ny%~ -F8r1h%3GBL>sONt^>CuL&(I>X7;4hflhoKkXNsFeP)<(jln -gC~(V5jriF=)fas*mmg1qbY_kr7bXHf$LmWXBq@jb470 -Re_?Vc`QF?lI0mh9YwP-qLO$;ehaU#?y#+pt4;S|>Z_CcaJiN#RFghj>$MTc|nR@M2cQbXVOVb%wkpP -8CL>Z&;89F~-j#mn5pezvPi+I|+H_#R9>r_{*_=`q9}mP3D+$fShnN!Pfis=1g=(h06pN=IGCWT?HM# -L$_Old|BCXh5Q34R~MwXP5AtkfcerDN~ZNqbbWNgTxPnj}h+@S6$1mmg5pGm1>Uaj^8?mXAr6h>8TIY -Mmi`XF(xZDBRJDW@jq0nLSROR79t0HGn9=F4HS<~K`%d?`k3X$(3IVJ=+16&geeP>_=u@frI~ -6wjwc$FiWJy`f}@4F3wb&sGz#`clqZB`Y9iMhB)B58x~5WDDAv2iq$ee1i5@PI!BNCImIk7~#AOs=#> -t9ggo0JiNRhxbVrX@jQ4L3x;c>35>i8gU9rcrVS*^#Sld?G*GbXF(Qqqrd?UXwCe -hvT9YvqU7Ig2HGR*p4kDu#N3a=2z8n|N5O4d4(bK2ha`s&Ge^L*-~pD;^pgxcOruP$2kp6 -x8qH#au*l3FnuNL6B)v2w3KG?LzV08dO9;>N=yKnxi`5iR(mrkhW$~CGv@A;Eh@+&NJ^VnQV{!3RKUuCW*-g#G -5)1sLhZ_s~`W^N+?hh9AMVjFSxMdiiu^~XPVCEG?)gFN#1y~oWW{&?E)JFoSB^w>8iFALFot{M{KduT -%soUn*J1Um^z2tFWqhhPK2>jbL_o+VgHP(U!7Ad4WKAcY{2;68#01Y-$C6GRh45`+g95~LGMAc!OwNYIU-1%WH#| -JouHf|CTt2o4kMAt)huhhQDST7nk{o+4OEkVlYCa38^F&3_odK!WZBtq42_E__At5*#BqOz;80V+45w -c>Y~&skKqiKl<(@uEiDi5b00!DWdT#<@NU#F_XfcY#by;e{13PIj@oU|MvN_^=;0z9vAJ=!-KD*h>!Y -Xc6+q7G5*^9xt<#MQr<+j{d36Go<8J%blds?dF`adZIpw8GAdtX)%fZ21CHlzw+8UWP3P9nvv`aNqjeeYcT-s%B*_o{v`Sw#L_tKYjl?8JLl^!KIoj(cqM_op6l689L=?!8~#Kpyl)YMdAv)R-P&C8?zJn__1Pl=abep$Tx?z^I-q(prF`RC&5)vH3Q0`k~ -1z2GK5>d-!XAkR~ncF&k`h+hxMb5(hsRWoMHo(&J@aNq#Hqe& -6CGwYnJO6Kn~W>Xke^&Ne)T>568 -k@J;LtAB;S$Gx0_m&ol)jYqLzpCm8}O?-bom6mUBq|Be}2`X162nuA`ZM?9W^1GJXEDq{Pgbj!QK|PK -YQR1k&q1jcpQ6UUh{Km-xkwz9)A?x*(~zFA@Z>p?DD93<>}u2RS*1L`O%@(9bZ4degWkEW6(fZNx^w6 -f$%@LYV?81szZ4t?EjFIe~j5yS;e2>pZHt$Y^CJS%zs3osGs4V&m%dee&+ZYH&0ED7iy3CSyZ(z$vx? -k7f2$|-T*EHPI*GDsOkYH&7Uz`U_FpaAoGM1feV2b|&p!K1oIQKiA?H~bwpmBzcHL%od&lI_oMakQw%-vQVPm -*wd?$5u)Zz4C#1 -+B7;{>R@!v}^`J@z+ew1R~ym_Lis7NeZwoE+v$XP*@>yzqipvt|wF@49vC#Kw&qxi0(QgA -c^f9qYvWU!>SrA;pIuekk_t-77x&=p%9D$Pw|$C!dJp$B&Dzzy4aBJ9|tV{Z)!nr%th4l$Di<3+H|i- -(HrYqM|}vx^#(U1&?ztYA>RNXh1^aL5+4JlAFd<*qc%THp`pDAlX}tmm|e2IZZqzmx+z?E%AvwqSvRQ -1%>ZO;e9FmAPOHz;m1+Tx2UGY76#fAU|0IRqMBzU%gcnUkw77-JYzNYre5ly=6NgChD@3@E&F&Ghcb1SN9};rfvqCOgFXUT$g -gkP>5Z+AT`%(Bv3O|X$&!q56DEw*)|2Bo+O5t}?cuI5GR}}s%g|Bdg?@lq?O);cW3=1fRH59{6is2N+ -P|-|^Uwcb&VWbonr%7>XnG~1blH&Ix(h+X4dH7(@#yYs9klP^QI?>Hpx}^T3J<(`cK7ia9AAxOL{x-f%m^PGg0DgN-lF-fcWbCc --hmV-e0*R~)bNOi(1`dph@f?=TW-18)6=tEBvQclD1IpZHZ?rNgAMo|92ylK5fK#{)xL%Y`rh>pj(

    ;QQ^}dPE8}{$Nf*=UV(gc+mLA@%IjhA{Ix5Mno|YoPy4O!aqKww*@FfP#kTB4sG -7NIeBn=kMGwY-qI_S5*SY3ZyP$4QE2D*9&B6AM8W9pjTprq~SW=%)L2S)`4kgN$| -j9-1iyqft(1rZ;^4WAKz2!AH`yZnL!!fSq3<%i-AAwdie?K_~>XG8qNn(&|jwLXi8nom@%&r+7;R+{n -xPm&^0^3UjVxphFDx?G+JtfkLgUr*8XxlyA=u|6_4X_!OzczNbvu`o4MJe3_RHp~wd>zQu|lj|xl*iNy_$7}x8Hu7^}+YvdyjR6PfK1F&yc>cj&y~cJ9mnM2M -@AtaOU(0ar*RWapue!QC?mye*E!A@#`<=SU33Xx8KAcfBYeqU6SH0dOz~Jp`Pu9jVpIiLr3Lu>11l?W ->G`;Fg0{9iDq(x=quH?6wSNmQ*y?jKwj(2qHb!($u&B#O3CQWW@*P(N_UX2@dyG?U2o4ei7%fr -L78-=<3rlw8Ww7TVvUX49G8Z~rt^Ss?;YUA3V%^kNk_Gko$US926x;wdd?1ksnhHi-e=2p(l>8)|6Mh -zSKQ2FZ6xvfh(`bt4=rFTl-r(KJ#oqT%r^1&5Vl>Qw)KHfe)9ekL+cJyHGutNuRhAP)2#6S$VL=cccxc>FP37oY~Q{;mE}h5(ubSRq*jD4mj}Poe582y({`uz*Q$J -d6+qO-9|NZxp+Ee-2XP;q!N}f1zLQ?-*;yckGsh#~=QIs;G_0-27fBal&X=(0|Aw&Fxnon5DbBj=y+5 -i{iYasy -;#X8ubkCi4-U;5d3JD46hIs!f{^!r1m!!)q>(r@JU*h4CAAa~jmXwq@_~O4;i}+_^4g&k`v|?UVA=vr_(c@#0Y`EA9RY{#&!bC1q7r)w6 -#Ef2!*p0Q_~(^FoKeQ%+D0ATM#>NIC47l);}8|3L%gLBIV{-bplMd@kk1Dk+nXNa?%PF*R{R_+Pqo39 -`J{&(E(T$wC2WpnSN77T^Lo0*-aj0ltIh$Ro%NH6@$eEG7ZcBTe^1|4)d>dKS4==lBj-{p~aSI9N9BjnT50_78#7DnQp@1z{_t&}5AN*M -(jzR=1u$;?2#JnvP@GwP%+Th;R1;{&x!Q;mC|KKv=47x(GYrz6$p1;8J2QkxFQJ>3o{42rj?-v}J|Sf!;S~0Tp*({ItvvTcokVqNXR4Dr5Do3$QR`NE=GBFN$BrEx@ -=t9%+WBkp1LYU(fu0vA4|vyU0RB|($&x1%2~>YhAR5LJ9-~jIH0b3SH0b3S -byDx$#P=Oi-VPdw|K@k8{O{pt4%US~)%Ps_RMwV2A3~Ya>CkBb4QLPQlP%zndUE5uAo=El0rFL%VGYr -+(yHQ*YccLCCmIwf!nv{=s%*-L& -Z(3q{q^#!=@?zN){Q^v3tj#P3>eS_d7%T!EAR$xz+I;WbwBER^lMOWqOC&PhG*;i{_cgMvdzOE~X9@5IUNC_Z@J4w -6-g=&(jyLMZXn#Q?aS`bz_^ -0wSwaF;A08r~-wxI7aLsFU>a`~uZU&t?b8C$lW_QKDf9(XfbUC?q+Wm&Es;ah()o&@tXFue<&S{<{1J -1O#*e4#02W!i5eRYSVE&TK@Q>N}eixT7I}VSZ>SjFW=Y7^Lj&hUaisaj7GyFBu7i`50XWxDh+s8$DsB -(l3>O6-2Uzt*FLS0=hXf#8#!`hL`Fu&KH!S-h%zvD?p&t9C|`e;CooZ_YWK0~dOqLzrbEX-c`nwo3)Mvn@J8E!x(d1w^ilK&@Sr>&ddViMu26qt)pSRB) -^&{KT6tbdW#P~8|KNiUOr-0K*75Jrqes8w!u}eDOs&V#Ej;6B8qgi;LwGPdp(Xd+agB -2mO2C3mVWCpqR|I_~0z<~pcqN1Y -aL4HWqmr=jJ@5YTAU#BuG>6@H6bEc#+A@P3q-FLH&@ZpCavOEBH;EFubgp@&e@nWtz -(~<%6+2zf!9_AK3gn^uT9sM0ouD{U2Mtd^wMUAZ@KI%e=fi+`}n5b?PK>%jEp|^MAYNo -_l0^dOGu+`jsq8paJCqbOJZfQJWTwL7*Lm+@p;~S;2S&`jK#hj)c4eEpX7IZ53r@G -`3Z*T8@8#Zis7=4JpgFj>w09nF#Gu(g|#++?78<*eU;9zO7SY*$hJvkrn9bszIQI~A#WkWY1N2m`VSE -!GXcl!7S;e_e78=b#k{E6tp_@W}plhQXpbVf)AYRnLg!@&Br=V4$Nfg}N4iG+N9QX5{YBtIc`He`z^Y;{ -50u}Mj`9W*xFEk#W=T$2mzX+rD%Y)4*13H{T?6^nv- -4lY-GRTZ{}aCpK!aW{)QrqI^bF{aP>BX4va0`oeMd&m-ajxzKV7W4mgzSqFtDF3JJeiy)SjWJb``>{Ri)umjghyVD8$ri|bwR0Ob|(4*XFLNFG=JgSa#P#_><8CmiEHpaE?s^cU2BkO -QU`9-sK=;J!SfnK^FNI{;}6-ro}AYw@5Z_a-}P||@EJ5h)=7@8l8&HIU -ZGuW^AF>W2gjP+{BH)Nfpri6$Pqv?j~bV4GiFo?cX>5%#*C9f1Xidf+5TbEMKhf=Q~X1wQ?2+7>E>x> -$v=UAu4YQrbka;0&2-mHlV-Np%s|b|)l8|HPB3fk+cAs3@nC(Ay9RFJI=Ujp`v~?DT&&e!#5-nGYL5r -j_aLoSr2G0(n;(&slr)m`{Sd12u?!o_&Y*L*6b3m(++SBpIsLShVV|jUY#0m1oWh+uq?~Y2%FljfT&{ --$Xl>`_=5`nL?WR$gu0(xCI_Gax-#VJsB$d;!{6?k -DG^JoJl{fjc?G|{LCp`ApW3j*ix)b^jXdh9pqb)*R4P6#`3ckUPXS`B9Ux+y#%y(i;8}mCDlgAv}pQVpJ -0QEsLsK1b=_wUiqMY_$?gF0^-b6TDE)z=Rgc1wCo*Qdihk^Wu4h&q-{=C0@Dvy -C-JQ!p0disESe^BQ=G2h1WcubvZ!yFIf5pz_S>%=_axcwY-kDfl-aq919%$YMM%{bq!>$5PSTfaFsn8 -y+^o{MorWBQnT#r$zxEqSzPb2OOi!yFUj(PxWVKj`U0|1r*$)}DU{E_l{y^AoG4tK+d3&ix?g{ZbPL$ -YX3Rc?{I%O#L)@tTj)^=~F)`Vfpgqu|_(1Duy51p*2s3_stXP=ElSww -#y`Z?$ZF5m&ok5{GgydmbRFs}NVHn#n|Hot%}f%y^473L*{ICy}^SVv>-)+qldTMs<&KnB(SY5M#C`U -U9QfhVYYz>5pY7joaD(L7Ip@#%Hi`1)!?nZP`gU7IJ%)o{SW@<{1UAo%K1&GN7J&rlAb^Bd<`jB|J3) -1S>59o_yfo@2pWT>;@RSDR}zJlE8JNZ)R^Cp`DubFr8kf<6vB^nSIzZV-5Y7pRZ$4G-pT-+Z9DOw=Wh -23f1K{)dh|d-m*%1q&8T1s@)M_+hU1Q17DcL;Vj~06!p$NZUB)^z-RYe)(>G5U&N$Feao{COmhjkx-uD5|VS@kq+D7oBcAJa`agx7}Nw%)5|Mlq6V+@sh?gyf+$Gi;Ym -PpT%NK0gG1kQ=?ez`Dfs>(}FG&6YQiHm8p?QUB!U=cge)jL -Df!49hqsf%YgnJDckUtX0*=6_H=yHRwj2e -Dvtim;PefoIcW_x`yljy716-;J4m-ORiY4g2#$cH)705pHuvsX>2OQ2WR!BO_~c}d^^8N(b0 -#-BEC-mI?^{JUv8=)aLMP}DQ1`;gD;)~$p5{Y1Pd66*NL-+i#x#OrSY%GaOOZ}mP^?ehXpI@Q>F3;44 -=yT-k{j&MSS+UKwQg?+o`dhw+8uzh{sNFIZqfByLmFTVKVMXIBIUbAM+BQL%5QjL1&_19lNM{#7n`s% -CfPJUchty&fS!V52KedCQcI6w6^4EaoD`ZHsA%%_!>mSUXscZ%aA$=?N(L-dUyuWwQxB<4$X9aJFc*U -y~hs5a~2;TJLu9ZuJ0^fHCIEWS+bGeJN7nADm+yKCsj0&OtTLRr)ML`GdO;U~2Zy!qE_n&ICFHlc1pI -RdWw{*J>Rtd`3%Ro!s1)^|ex5&c*670_4lrM}W~!V7b)=#v9?$UoA3Wo|g@IOnMk_RIw}4)n#*_d?$f -{XXp3SVEfcEZVPZkkUspS#vT^J(T&*W87N3)H< ->5hLR17q^H0pZ2ajDyu4sUzAqbOU5*Hlo1hKX&UEoALpLe1w=zbrBGu{X&`mscW>{&;pDya+#lpRc`h_C~_br7v7_e67=sVx#CB1LF{=!_19C;Vj1}gf>Zj8)PszTsTZ)y -#L~K7^KDsV-U0{Fs2w;(0(RplLcpIITq7FQuxkU9>`0XFJp|+M(6?u<~XA}$U5J5vd;{e?dvoJbjyQ2 -g1<@TV4vYJ8?F{#;Ghf_Vzj?V{usY?=pwn;xY<9DUz`3l1Y;H9<529k5BeOWKUp~A9E?_5Z)l$ -ipL6||VPvKNCSfabLBnFh_=Rg>HuynCKdAo~yqK@s$a8Y?&gpjEzTmIwm=I>eHqw0l%?Ths8#MUsV;2 -76)BeAnzZA`1!ZO+S$1xzU=odUJ{1QIA+!zIS!+sRU|vojR6Hh76v`_Ms#tt8iq+R%6%!+ii>8Hq4w~O6j -h=lDo(u)DrxoQD&pdYySBe5tr$H})328-n)ARBJxq;Ayn-||6?2|!)BB_Dtf&7^K@NX5H6&f0xUN|{W -6dN-wFEKkzAfjS(vhqWLSS{m!<>8+?kN@@FW&F={mGKvhhI1Kz9#5T?n3kBFJaoh@85!^)ym;yaf#M2 -uHR_0ZqQ2-Rlz`IEI5YuGMkVMzG#5RJR-#Sl6zYse;bJ@wKZiHq_iz>>ctx_gcJ@;Q$(QmQSHOsz6m}HCl%@qIc15 -)QnnD8{8H5#tNt4G@OZZ@C@95U&aUUXE=hiBb4kT2S^Lar^R#ut)wrR@rV^_-e<=2v`!K#mr*HTZ?Q}my7sa~$@^h=9=V?^C&x(-YSG*1y>vBQM?a=5^b{2^HVe!(=0@|7+1ZM=?zB$WF-{z-W%cY8_7 -*$Hg!>VnB%Tq^i?>9Jh?A-EWw~D-k>AO-s)OpTdMT=e8lo1dC8}C|t?t$VJy}oHGxR-ru6{@_(U0mH{ -fu6t*XdXF8+wQSKp)Vb>SOw(j`X6u&R!p{zlS{AOZC#d@m~0ki^<@7J@IXL3NFVjxHEA`5-B0`$qVG~ -q>1b%pOKT~M{+w2(#5ovzDJMHB(sMVXG!Z0YmN0j;QE!_&53dP19s`YH!XA)JFA>movq+a2b@Tj%&OR -WwuPMpABuE4x?S8J;78cabaULG`;z;zd)z(g8vF`=4gVeQ$xHbH@T4xHudqatND;XrC}xRLQ6`p%$Ha -26N~{x`z_Z>KyF|12Tzo4|ing+kwGrUr-#(Ua()jQ}N_KtaBB -9ibw{AHp53Zi>Z4SEW#La)JG{XG^q0SEA8JQa5((IlP>Cs`zy>?L0TqJ!yJnoH-&@<9<5@m?kd?D1*&FN!M%=-G;|{kwU&vSTKZyQ -9iIHNGC<5FT172^4e~S0ThvE=;RX3R=(`BxFQMOUT)H0O?cwFiA@`yLt3vUWi3bqVE`KSn0ql4%$`T- -^55qKtE3{mG_xGm{S29X?+PaYvpkk#ZP@+CPz5OwKbu;n4Tls15m?4uFpb>{VEFY~n7)9P;xx2mlh9M -Ac{NpNp*NArB%Mo_U{233g~poM-xZ`PIGvha2?hJ^k#nvK?=T__6rSwR}W55Iujqix -*f#r;$BfB7O3N1t9RN9D?Z;}NJOG`=o%D-1nxxol84C-a*)K)c+hbe9ZjdwXDKtaxzwt)p0$oyXDr7~ -wCC81?4@>{{h}RWb6FW%!j`hfS$nsGdy|{)KH;u*H@hFXN8M9yTYfd~!V`E3|F@VX56Ew2M-`_WHB>E -7E0y7I*4z|c?}8`nMFx(+{c$41sRh_3n@BG@kUm8>({1#7I?SAHR+^8QE6h5x9-{PK^Nbk{GwVicfQ2 -m6Vlc-Ntc6yERcTcLf^F<;?HlcQ+q9W2?XmVedxPC*Z?(7C+wCTMr@hB+w);8*9ORe|gIK=C8O3I>Ic -z@6zsK1cwvKHAZ`ls9yqSH?3vur-h9Z<3$embP&9HI{5WbaaweiR2IsmvX^>L)u}D -&kh(&T&`Wh9#J6K%nT>|er9BQ&(ES0jihKwfH$X(*Lf@eu(rB}v$;@Q)s5u0p@j~l4>sfoPjU2~W;%s -)V1*E=oTiq@^UJMt@Wu0tNyVM8jplVUJ4i9peySWTk;7VMDt8pE!$B~eOI*A@4P7DyHkRn;6ic#V&F+ -t>uVli9Hg(y}jszt4+7i%Gk8UAM78Bj^c`g(jv -g3_u_gFqjA^OaUB90Etq-Vm_cz0eDmaA~k?X9iY+xxU2_c8X>}O6Wc|T*eUi1CR1dFoG43VHH<^OY>; -c^dbvp&I#M&p&j~tNr|49jrbp=veV3l7^C4pw>k^3hrH~uvLtd=amwqR`TrQW(<#M@PE|>pT`43P_0| -XQR000O8%K%whEkq02^(_Ga0FnX#9{>OVaA|NaUv_0~WN&gWWNCABY-wUIcQ!OFVRCIQWq4)my$gJl) -tNtj-gjn_nMnfV%AHAqw39)sKmtjt&70sQ0jo=r+OFOH2%y^upfy}tqSYk0?GC2xGLV)QcQ>H5n@n{V -YiOO`2+-XP(u;uF%eFNEZ4+V@1SEro`G3FXypwq|nW)=sfB(<_^WS_J-rG6PdCv1Z=eeBgNAB1mY(fY --e$l89&j|Hri1+ZP2{Fm@y-A|m_0xi9v|D~!Q0H&hlz;EWPk(CT>bvvTuD<7? -L9ZMFG#fBMe5uF1%7UuA+geanCB`(eAioo|E4I?$Tz&#g-GFlB7PNTbwy11xxO`)DRvqEJmyyPzF!xA#hVsgOsm87UR?~NSrYW=!blNf -f1WPt-hnk!rs|?pv%J`=iD7->^w(^@YwKoQH+)`G2(j+=v1#&!sJ~|8ovSym7UIRAf+=?5`mi=8oq>P -XsAQsQ9$%o~bX;fYiRpT-xmP8Wb-dS1=N*^MxMt&~jcai)X#(1a+i@*UmhP?%pGHE`h_n(;;x{f`<;0 -Zu|JUCXO`M&gi{KR9helet3OTa!n+`0@fR!aeC_wTE`CDmy0ljBimbPFy|u~SHFCeBYh;-g-1JvQ*Orf0bbYqIv -a89ztn2=JZ|GX>H6n&62R=6K3}2hjjx`H?=!jd?9nnO0foO@W$`|KQ -$50`f;2JSyNq&Abqdxy&#v0*Vk}q&CcrTU|3Zo3yGAq2qzseSI&gKwxysov2x`Oy7uTRTf9`5m|1iC?AA&lN_ab!+r%m!Rq9k7X+LL$Y=uiIbEZdW7@s0V{TtR+u?GlzjxNC%WN&a5k@8x~|lKgvde=qN+@je~*>HpKZ^A -EVS6`buDCQWU)-v*jK3b+MR#PPNG<>%J}etyQIfM4LmJN)tm-idrw&fBX>*Szg1ov0kr^|VP}(z)R5B -g3Sx%|4dCJ)m!a&`_Rn3$Iu4dUp44!P#BIYw=v4FK*%WPFzW^M?tU2pf&J6P8yTepyhGWo%E08C5=gI -(wwx9(fU!)n&qrD%Ow0qLGywZ)Dge@y?D2m@A8)v7T_QKxHFO7bCG#V|zx8sL&@ -sHwb^|y*y1-FXMd+*WltuZ&xJfbX3Aj<_b=id`F`4FJ7%qe{#F#_J#_4GVQXXFWvJuax0mg**5z; -{S`2-1ZZ&W!aGLTBdFi)l#~!z9$E<1B3`dJ$E_3z?(M{YdFGpp4I+53kzr-c=AzA0bh)N?O(K?p10cn -4Wy2Ci1u<{K2DczQ)_(R?ps(x>m__A8^CEw)BwTdt4qULev4dTgiAG -a%-kMaoheKf}?t&B!Tc{B(-Cr`>qPjkDRsVNS-4Fwawa$_7HB%T|re6ZlVjeHO`#RntChZ}ZTaVITD` -4H1HCSH%jwpi1ivho2qKN=a!hfy8?5BYsqeQz0#^Wb+0Jcuq6$EEFt%|5qw;jwe6B4Xtaznx7rZTi}4 -k9qXWAvmpd$S2z#+en{nN(dfQ4%1UtjNscFBhlS#qlxWvr5kWIePHP9;vIVkYnka!O!?X7oeNWpo(2MuEs+d85x-{+QC{bK7wAz`Nd7LL48kIgeEe=P{!(a7*>uH|-*?5 -`VIc0<(-9V`t@C(QeXr{>5k=%cfmCbV2bVS$V4<*BhpZx@B|36u*B>=~}1rSiSJX{Uh(8PVZHRIz5_g -NSh_}<9z8^GeB -ykygYzZ@-dsx0#~%07j1`{8%Kj575kWnRiQA}Xf)_s;iHuY>PWuiZ1=OTEUzcWPpX&}_k@LhSdJ84=b -?Xyq@KfZm;7u9UvDQx^xGMr_HnT7{YRBk-G%ZUEmqzpO_Fgg7AIf6|E1XXZDw^%@uKvJ3eN57~o`&$i -as@ougrrqM5_E?~MuJZTrE2-Rnki0=TqzZAYZ67b#8xpOSOOsnlP(^~P}Ig0O&WcUsuU!nG=;X88w54 -uiG*1BGvszYXM66gPfFA}xoABW#x^72FZ!PCIG5ASaSKl}L|*IwY_{HPJhM|ms*>4S%)5A)+r)!#85? -o$%(6cg^0QMjKr;eI|T+$q6tS>T?p{o!cegjMF=1gvSAcy4O;K-n3{*5NF1JX=uCruzD9?$AA-5bYO` -t{rsVCf|cT+1}tR#7qyTh+wDgKIYQ$+9->Knd0~^eb=6I@cl?@Df^@S8ArzAkxgFzaFo}2fg>0N{~<$ -X3tEe0+v|gWHx*?D(?vVsOFmaPf_+mvW_*WnUz&{HYBc=b%na)WF8_nZF4Xhbx=ayXtckX(=! -b6Gd*UACb#2>Qvt4(sU^%2I%Pp4W<@kup&&@J*q*rC;hOQkgGdJ{glqu_W$PuJmH|ydY_2vV3?|+}dV -_lXJW?JGBkhtWUxMa6ogw7fGmm7*$7R&Z(&a_RH-eauq{v`FyMA;Wn-o7;{LD2PF-|mhXYf#n&ylXv? -609`SJqG!y7oy}6?&qMs4WK3Avz`U0AItk%68Mo>qMfiu>jL=8SrN6E}>Rhb|g>kzZtX>`6uorI%FO_Bcgj3;pvPvC`u}K -G<_FZNT=*2=qud?{zIJSOxjEWwq24XwZpGSy_@kE+bMsMTChP^$FYYs9l-t8K#b68SHD7=^HB|HsDP< -vMt8yCBJQv-$-M&(-VD1M3CQLW6p|_n<7#-H&jIMEsO7~M9a=-v(`1xEV?A^D^i0J&RrD|>g$4islim -hO2H5GlzQ8y%B&eJ^X>|1A7bBFc0XmwESoYX)2*(Mwu1e`)RX5^D~89`Q_)hTl{}aBqeT0QF9h0L{x;iV5jw=oT2LPziHgD$?D#< -rn!?fF^@pOWZ%(&%@9aT;xrZG(9I8qY7W?X%ws&oAQnzwmr8MLfl_C_@MJoS|=I)n0B%7j+$Y&vzYq? -$F=iU0)h$px!m(-JMx=myI-&9Qj;CW8>1q%A6|GR704exZx^GDx;@372y+U6m;suksp~_aY -XUFO2Z!2aH{TNjYtn2%B<`*MSO^HqBVO;>s{?2Vlym9gky&qjxcs$l0Bw?&nGDTX&bP@GU-oAy>*UQP -P&b|e52zmrhjwaUt`VkOdMflrLvZ-BrAGNC&fpHf@@FZXN*J-cE=SIym~L&zQo2>yd}r_>quDI@H0Ox@UOpSZ1EOw!gaSlZfo)BpJQE2jTZ;R@D|urE9()Dw -Gm$QAs=i0U8ze(c?9ckp)N3>+?=Ozk>a#mnO4(W=IOxyn<$ki{5rc%hE) -UgG*Dqe6Hkpk0psJf*GJ+i)ep&PH5u+R>BQY^S#k7vL(ZCaeZcd7UE6-MFprw7wU;{JIbWJl6|jNf9h -Gk;wXp-+AZdHc6z1TUC2hrHSZf8qtCsn0T+_R -zMew%N^lP+38NqEDeYTv?gVW(dN*S0$pRkcJE&c4@ydB;B6}0`my3y1F*=JcUlm)-AIA7XU_Q#s?zj| -_z`F@T(PrDHtTw`qQHS_e@oW8eQ@QI?)5U}Z%$fiwzNqLp+Md -CU|>ZF74F(tle%llDw0C_#zMPn;y!?O5|&q`i0W&H3>q>I69$M1k!#UWZjR`8TblQPf*zEW}E!wQ~A2 -JcW-kmc?Kym||bdJ^zvK5KoSF~7&^dr}t>)@^KmK@xojdmgpThQ1SVnVGI+( -SD|g6akq_rgg8LF^A@HC-QZoG6EG5R3!1p~nz&JS<{&1MWzCbmevLIwCcRysGfuKRP9^B>%IzsFFA^Ldr$? -YG9}@g?Nxhs|T$O#M8h^z(CC>(tS9!Fmh<#xP% -3>KihOFDkEBEo0Vs}pTv5P8V+ec(?8WNhy?5qw^kdgM8_vlw4FF3a+czNOlnOK&+L;jOHQJeuNa>j%u -fBRLZ8MO-`eo~A*@x~QjjLC@9!%Al>MX#ry1Gq!2+eUTA&-NQ%u||rjL=yM@%}|3T6cV0C?>Cdj)u~zi%7zwH68)$NurMvXj4pew -rjWCnX+_y_P512l1OE`}Wj$82p?FeeQbk;pb!HT%*3aMq4S`W3iW+KHO(B!q)WGIEAo&1D*A6nrQz$V -*D;m#tXi$&k^m}LP(zl#YeT{hn3vVXJk)-Yq>dN=jADDfgqXXT`D7vsr-j38xE!0taVOSP>&PfJZD+FYl!G3i -03%h%F_e~s@|W?q{an|Gru@Z9%Hl-H=}9YQ;fbRx#?F!6#vU%UzLUWaT*xqm>#o%Dk$%y_w8FMshaOM -d9u?td`zQ+{DvBDaC3@V(o!;mck>~fCd~t%$a8OH1k(aSK>GgdfV@4nQK -$-R~GwlHI>|AD)ftMxT*(%M*{fEc&fs%*i8hwqHn(H-{W~rHW?ric(@$Y9S<1112KaS<=e;`l24jqws+vxtT~4`hk|M3cM<=7@I$s2oYCk3tL(@)TaG)&@% -+nCy`$PksJg9&vH^HOuLl1>bFbGNqVDSJjj->~)op2^Kuwi~SVBu%Ux8SnY9Z1ra5f@Ft4kBEV?{{gE -i#&%nybyd8umN`_l}kmR~I=S385XX($a_MgH~r4;pU2>p~qmCt6WCdrMc4h{1Bd(?}Oh78lo;I)`5qt -D}9fuc}CbThL;H0pQHLo?4~~%^D8NTiF%Ihx)`*jul2M^ODAN+Gt$vZxv=USYd)C{vcUXBrp`j!E}`u -HdXAFekL#+>A+f*43WvP$7TqlSgP>QQVn2`%`H4IGnexHEsyV99x?X2rT^v>n#;a%>GXBi&xMRFr{m> -#;Mr2cp__#;Q9;yJ}r2fzxLH2LDMqcXegD;X#|7Effp`I%0w+9En7cag^zXolWzFCpxd{*W?WDgFesl -54??|q0_M$_yYn`RI(%Mk3%0Q^#StLxc8giYc|)YN%RxPTW -K=#mezNG%mlk||z7UJeD|sAiyNU(@c~$M@gHrc;3mid(i$}e0{a`S$zw?dZ=|sIiM>X57AP;<#P$Oaw=1ewz^Mvy9Qu&K}5eL%O??T)5Xs-1Ix~OwrXN1|7-96s6YP9EawCB}Yygjpzvg- -4wZ<%Q9VjGRH?TwA$#?Xg0f8RBV{uOchTlF4e(_N6yBAb{7n-wN5GUl3r`%KYTfcu6*BV476X7--%U+eOr^hN@N$}?E~h$18r&Trn~#{9lNFrP(+q+c~d -yk56-+VN6z?+4F1qEPh;FaLxzy(j_|_hpx*+-fKU&{~jmu1ftRIg@EJbJ^U9J($XeX6^NCt#@V%9^N# -NEYH4STE73Y9zPSNsd&crM>TWN}X8Iz$^FCml+@4X@mR})`SF=oIOHbT{`=>MXThmd89jJo~_|v{+5& -ti5x~B$Zv;12D>jU@=qAsjQHSk%Ex)854&Hb=fbIUrp%{uj(xU2`w8_ahDW<6IyCg;B$ty{pp%?FHKE -g1nhcIP$gv>bJ+(k4G#txb9uGC9f-L-MSLABVJXY9r3Ti -J?blIM|)6}K$iy4Wj^S#4z_rema)EG>)4~lvA#cc?)cpJc-AGGEi~}AM%NJMzRpp7=tI!YZf*LpPF>k -hSCg@`M4R3=xTvD6DpMT4u$($TG~S^F_Bd0G@X_=Z>WX!ToT6^HP3vO54&AxFa*DDkGg}=&*wVA#sVQ -#jO)VlMLxk=R@94`;188GH+YIEiFV_V^{)5z1Emwn@RswXn7HkMORO?dpLF=_9kfs4l|m!(a-I<>KpKBk**_c7pFc+pin=qlbl<;% -9)mWwuh*gIFqd?2sr$pQ$E0nFHgjXW^p{#{Lhx<>4SD_rzb9+vw{;?4b0uhrx_7$NeIf1#GmUU(rqY@ -FAa8@|c+ODI?AuuH5Sh*J)lQE@>jr^GvzETTA2Q5(^fE8|mH_L=PUAl6|1{JgfVz<%1ItDDC}5E%)bE -AMirRj4KDwLrBP`N@a5{9?`ZG3{@APoAPWe1XI^UXWM8>W|P;*wOf30@u*J7;i%rV`^E$^Tw+b<{an@ -4>yGadA0`)@_u=St=W?L&;!Vi(uF;t7xme7`enHA)qF)oUPxd~s|X -l+O2iWj$ -deg|#2uLtR4?`8i{ig=1`{3~{GK&I1Dg0VD+C3i9%`)t`yH&KQ1U%aJ}$uAEzt9r|9j345RNNx%T1K7I;1K4Qt6gM9R3DaTBEtIU(*`=jYkCyuk9M;zK_$zQ0--gygPF#l(=L_5dv2(JqEwK+ ->j8zkp20Pfq!`v<`J)RpLW{dq>1ai_;IVlVJ!Q+Au?Jl0^sDtCyD@bl|d*_9v7^i1o!Xq1J7b4kA~Rn83n(u^~6M!)y>lsyc5S -L;UC>r%%}Zu=DI+-I8{%B3%6!gsp#TKFs_%PlnM0j*-+5o|0ug*>rx5U;nhADZ}ZeBt$xij#d>YTNjE -2g&@HINv#sFJjX}b3orqpzVO1%Ov)H2(+02+PFX)|NZw3yFnYqpz{mdkD$GssK#$AXyi#1jZ5Iukj8! -)yXS<+GZ*|c($Pm6g5G}6;iHBT)CR&$R`K2G&LEaUVXIWM+8g=3D>e45DRI*x -YIA{F1!1sf?^V|mdLN -Q)UPpRct!sv5>Q`*MK2FFT8&Zi4o!f~1)-U3a0{YIHvo3dm=Tr_7}tloneFYfV9esPaWa~;dqa@yFwC -bIJ*_kVwM{=lWRHt^7{`f1@mOwMmP5yCaP=>x~kn{x-B$W-lQn0;T8A9*4A{D}LyX1L^7Fa0=Q5%}U8 -scv&ND}FBqzjMLw9P<0VEyEuhso1r>aKn`@{d#nhesDM2IGko^;OIR{q -+b+N6cf3E5a5e}}%B@gmKEISkM>g~hd9e|%0h0l}@-(e7GudeX6oxwGYHcMCbcar7^e#s$@vtO$cGEt ->L=1PQ&pW&}d->u4lcSWM{5c{?qB1k{q2_76hI#9YC{snxuV6!%Lh~Eyue;Ys`$jo31Mfcq-wunPS3iu7Yn-~^o;^r8KAXxh+g4?Vnx_#+NH@f!Ssz&vB?l -lV^j*5|hEIjBt<;+miONb^bf968K0_)uUE{Q>sJ4Khz|Pw8?qkH~4+Bj<^m>teorMd{BS=JlQewf-jA -{2~qf7S5IvcNM69>`jnA+VyLmG&O?wN=B`(U0{+L|J*RXj}{2I|0+9o{;c -bUg7`eC#5ps`{3qg{O_{LZeNp-VN(S>_6ML`F|Hzf`%7H>H$azw$<(!x7&|ej(af|W&30AuOC^iq*ch -GG#t_AGh!=LkNN*_PKzDP~=fhOY2aSryuvtFd}LFme9XjfD58$5ht_#k}y`GA*aOAYqkw`sTvd?x>QM -x(oX?pr(z|LZ`zreL%dC|mh^=>GxamHh*U9bwq}Ani15sOxjaF4<>>cI!jCEke6>Hmw^@1JCmd(jNI1 -Xwp!T5>)+;o}}}Q**`Al)7yuaN?3<(3^yYdTZ8uKH|@X@Q!h*f{9fpQY1Daz(;o4my`v30r9pN)*;1! -AYG@1aA&S}G&c38(w`gRa7w5OIO>*vwoPUR_^0{==m&=j99O2V1BK+c95n&AT1q=O~W=Y!`Rz68`xp_ -}?Sv?V56#vPC%lV=mK5iBoOgNizKUIXu&t_e3d>Hl20Uy8m>rKN;Y~nxxc-Zo5@DTpA9N##kcvgML5# -I8*Gio7|zILSP1WnzvsmMDYIC0#--!IBoo?Y@h*OyGTDD%I7%wIs8?4B$TCr%yO3myed$+4j&lcbN7q -vE29xF5^!UQH|+{kA&(?LanSZ?qAvOAY9G$6TO~Gvnx<-09hA#4{aux`AgJ=$C6;eq=ahN}x^JTf12@Vh{3mpBLcuWLfA -b^yBZ#X=F>v_=cIBQEM5iH?qUkM*n6`U~RYBtD-rPP)^m8{Cg5x{b9vH^KH*6v`NPKjW@!6X!RPzxU0 -XBlYY2!{tcB`v%;HV1x1xiJ$@5;`}b+^8A%(Ne{mN8|d=Vo2ijxc1X&m*>*(_{U%TR{p#`W6UW`z_kU -ik#rM^LKczF{am3G6ym=yV?D+hZXgBA$bY6)TbIuUTD)CwJoku&~yG#TNP`}5J?;-HBkn5>zEj7dVb` -ZMg6m*uH%MBg#k*k!S@igAk*Vv0ZKIG-~XXd^DFU8_NOFXXTJ!*(;bQuS_EI82b@6pxzAbm6nKxSbd%mVIR;^@V7Zn&HZ_`P}%VD@mTV>`zIFM){60Q_h>A#Fuj)WIaYKMWxNS`Hk!0OqUMcDd^gH ->IlkA&&|8)3x=z#%L_2CY=Cc_0pl4mVW^ZIayKEOXRzzG)4O{cIY&XXnPD2I;-!GaNZ%5~frv4Rbp2x -^sH6LYQuF)ir);dNU;{)M|k<>_Xe=9z2brJl36a2qqv-yT?_R!zVQ|mRde^U{m-zVFFr9b**+qBSW#B ---$6FE;R-VZq5FCY#s&4>(9_l?{ffsW7O*y`d1&+e%mi5^G>>&&p47sPnRT(fW@Hh-D6jxf^z^N+Z80 -_Hlv+yIzPCwQp2SGcb3<=D)#ov3>Y%9&`sOiTt?{|jilwUC8si!AiX`AJF^tZlagcCPn|@@>yuiS9lH -UMt+QLn}mFZqHs(a&UHnei!r9AMlD74rEp9PF-;x&&}8~|wO>7DMes5 -Guk@4+*hGo0i;`I1%?{Mnunm8gaKCt^`J}7Id#p`&9V;4XMm(kZ-u?34xt_NEf-Nny4za9Da~~^5JR# -$D;LTI$Hal-0fTc3gwe! -#Qa)T?1z}Py3n6v>jYv9n7d0Iah)Bi(QnXgB8M~`G3f78^P7ic$aw$QN6A?;k5rJUzfIa%c>2TbiQL0SFKubf{p6Y=aX^yW9By -R>J%;WjQuPik8SrgqGTHjA=>wW4g0YX=w)*NQIK=itD{^a$s+4o2-=(e)x6J(4mUU6wi=wP{V!U3jk5 -%c8d-O|2ezbFPS-Kkev>K5Y;Gr>^Ebeg%3Qe(Dn7abS~OlyI&K`vf>&^=WN*X~*HkaxKd&;QBF}=*~e -rwF!2X^Mq+@{J_ -nsPb7;mbqqfHC>G;gr=HX1Ip-<;y_EOyeE7H5-kA@-m=@>5JtiNFBzy>Ke=a_N{;BWGiwpmAUNnv4#e -Lqb9v**58lXv4r*X;_Ob3rVY`TpdYgpv=3;by1Hh%9FMJxEFNw?$+)ChpK@+H+ -p}I7K^!`ScXG^ivN_)HHTX6hbK9bDlJntOkpD-h<5yDj$Sz%vEH>wghG)g*X1<4Y|6+Iyp0`ef=YfMYwBLd`7m~Keqc;5Tj3bGfBdIeP%ywh4+xiOq(sTn_m%)U!vX2%} -8_27Mb($&Uuv)Nz+7`3)do)8^Lix9`2+d -%dV7zFii4eZY4G-V;|p?!B65BK|Iy!ut>G@6lXx&OLEAzXe?;?(atD43o|oNIyy9oe{bXF|-Bu%RLs{ -Z!^a{<~x7Br{VLHAK@GbFYs~$|66eFFzFr0oh0$U+N5`S0=)--ch~=b-o!D(H6C{>-oo}Sc!RDIZ(KV -}y2tP)-ET4J9!NrW>aEUqpnJ0ZGWwrfV@~_B6nJyb6SncY(U!P|L+uHX7uw_$?VL-)xdrr5*bmpN=d2 -&PqUN@}3i6#IBCZ>ZNP`>x)?ChGHzKPmMQ|nVIp*kv?dcTGrB2PW$j4X!HfY@%xjrD2IIi#Iy -wnHiPxh4g5ZB$J^_0=)9)zz)f3RFjm;Rq~R!`}m>HnqKVVB`I`pvX|oC!TX8#0iqtt{6v)^Sdpe|AqP -(>6ekeMqz@xNy1?xH(0T&uLQvdpc3ayB+L@7@tm_G{NVhdo&n0UhZp~>;x_+IcuoBc@wzIXn`@EseE`~k6 -pw%3sZqT@9vM%s|vCGvej?bU|z{!WRM9|-`?itp>$$2Xo&NTluMz~0Grup!FMZ@N2)&q8P0d(eD(Yof -@qS23QLztkHpeiG)`N5PT|AVPT&AL|BT%( -m0o4(BpLbQK?W29ZGZ-4a(wXW+Eh>waGduM@0Huo`?!`SK5+-)lgL~wnUh-_$G^~{DPBK#BhtNuks_> -3)m#X3B@v^k3!7Kt#&FlKN)%Lhexx(T0cb@eH_KMj -n^jM3Ea1<}O&qm&ufc$v}8jh7kijqx(`gefyl`=K&VD`*2Bjq6|quTP>b&}W_lYCW&#kQ!HWUdweuvl -nrkTHViC#0>Y5Zx^mbx^R=07auebOwij`&_@guSrSqMCI<%q`;UelGour-Rvy4~a#$2(ou#pig&<_&98w&iP<@mv?=*UG<63;)VA# -lO`F{0mt4ce{mue~c*peKdi8&Qwvi_g(ncKgvI}rx^cSasGYSPm{%Bs>t$p*lc>)jA*a?Y8@CCUPs_S -#PFkug0*me=V}yxkzg62C)@_Xivr*V7g|?a+D4MA>}gy&JSSgZQPB@ylw)FYa~wWPGdezQ&~aBw?lj9 -hn}s&V4KNVn`DugTTd^0^A#lN)N8qk^VBlU6V+tJ;3BDPcSbA9yH`M%D9SdaA0u*oqJj+HQe9L=Ai@_Ug##@j!DmVVe -ruQt3?@(y{9qK=7W&P1NE>+!~p!GTxNCd!(5<#X)E=lA-%oFx5Z^LptWU|Gu-Ny-;V@=O!z=#!iz<=Z -UpvdsL6`&Yd&AlI{p{O|)~ZOQ*GZRr!1w#0QPv9^?IX-nxOIz|=+mcN*anA1s&}Q(hj+y}45yw&Y$^JQo^8!A)!)!wxiESuN1o?fr*@oV|u-j?!w-2bcF{!Z;ESGFVO+xS+z -9o_tn?MRi&Ja@e{wjDJLjNe0yGP~83<*f;_%;)#|yPPEbWb=CI9bj3@Pi#wD&9;!3$J&+>%ePtHWr^h -D%%d@bYi!K0rFGTtMNM>f!&mLbbCzF0QHC|S{-wqbAGhcMt(JB(B*cy?Eid -Rtx@ug#vwKR-y;iuU{22Y=hMKtD95W~Ae~Kq$pLkx0{H5`X&!gd9EK(;y2Dt~wx(~*6$Z`$wTuOzsZM -0dmYe#9rpc8zVff?&Q$o4Lv^hF8%1J6IHHpjWnp4po8k2WI>WBOj^TO?!d1 -F`;c|23+=+_lBAlCfR`(t9115$Kk4UaplueJ7&ny{)U>UJZC>v&9I?K47=Nv;!t|$pIbjRs-hRHLjq; -{5EKFhUVI<+6s~8Q-wws#}zN6iSS*h)8Bqj>bA%$`Dm|fr_e=gr+&QW81`!Lul{{TAeW^5{i87@#<1)f}S>oJo)YBQ0$n;PQwe2R_4a1e>mN{mmM7y4 -$S=#ifH;9Z_gd5|db21a4*A3Z@viP_38!;xSqv}Pf;v*>tn!?L_{56;7Sv~K9S$cMH&41=A+!#-Ub1P -vj{~<=3q+%LvZ$+`!oI`MXy4)H!y*{ajqnoqVlMb_e$<=t>_UATP~RfZwgl}|#>M8llGatFi}B37q!a -VsZsu1sL;mqR+L{YE^JJVo&NgoKB*4qlEbx-p((Huvv9w$(o!GCqrM2Uk6xqK&4$d=Q1>W0{-%I}$dA -}ltzf*+vwq57i-rQr-v2(;4_CF0fQ~I=TG^LrwqfGtm=f

    D-b)8R>aR=)6<0eCTT++?J2{3SafgGhhJZa*CA%#UVY>BR|oMoJoxHhq94z3(1 -(-Y-Ui$Q?wgBTVt>{q)S(qNutk^aY!;Siff>6%)5x33mw`Rv{$2;Ra}T0JxxoOwEyk62?t2h?2XEzCy -O*Q8S-;V=@aJV(T$xVxqah7^4qxCs#v$(+f7r1^{Qt|uzwuQOx;&a|h=N#(C-pBJ6TTVo6B23?lYmd2a`AsuK`(BOvryAik;8U#MgWuUFas -$_u*Tnas;rd{=SwFu|gxTM-2H%-)HOg9Rjt?y{_ei0SSppm{@(c<~e@5aw1PlHp$?%tc71vg9&NyJk{ -HpIv7wvCC7GHwAaUM?pT%##AM#cRcW4IqUyrk{FgMaDh9xr3(BqW;`q|RN^eVp6cZWqNz(Y}0AM%+Vk -ECKxIzLuQt;L>u^*!R!%)%oD>ZnP_pHgza>;qE=0L$v5rboaBmk_DHR+g1(U%V(GQ>@uGLGcXb@;rRi -Kw`6^wD`lS>%IVUbX$^8t#e1LM^yjL_qnDdc3Tr(af600zU&l9WuJ^hydKQ*F6tOu%mu)GbrKzI(0Q4 -4h*YhcAe@BkxZAuZ{b@=hD2$Wg(ZyBQf72AwZ4%+R@w&`4(&vh`NM^n%~pa)mO#&A93A;{@BGPpObn) -|ZPF6#E;Ifwgn0Pa7TaPku1sB^foL$_VvSu<)bHQ`l2j=7f&$C|!ufy*<_9>z2DZkS`hEr`vVVF#Iq< -8KdTsC5CiCvwSIqTI)3t9=-6nXV@(1bZ?({I#kcr?{l=D&l$V87HB}Iwb)z}#PvX27gSV`CilyDScoM -b6ly*?W5%?M_mKXGs&k7;c -GcmSXrV*Tj;6TG|m?QJ>szQ$$X$6~R*vfv=lH@Q3Q`f}E3Px&LJ;cjG86k4)sx_|BtlTyt{~c(5Odb1 -i4HU-Xm_;W{C{t14JV`@59uy(;ATrb^B=TteM_`m7p@G1wn}R_^6J)HlL4VL86v;5i`oOVUTze2WpeW -@XHNQ}Wy-q;F?>=(sLj?TOVot~%KJsx(?YI`^fjle^@-LsIf;Hutey&o)c9VTsH@0? -&&KzM#Mt+_YW|#cUnKdyKm^A;SBSCa5n9&kUbM95dh#Muo~?CYbGpU<2X9W|o^1|(T%=ZrZqP1r;Yzf -d;}u0}{amQnCb|dl?A3ajICcXkGsrO|ua+a{@h*XHQ(>O5HRtu{?!AD~gR=fQJwB&@m%8S}&LsRc%I3 -LoEb}u+Po5J`d!}V>zQrCVTYq@u4EUbJKdyGDxs8kMqC3aVJ=3^vd(N1B*oiym#^wV5L7nGL^)#Kaxf -1qdcO5&J5|DeVuusUjb<0-IMYV2C%{$Hs{eGs}o4eu5#_((0r$|?JFyFBD(l)3#Q?B*As-JuR<_@iu=O~FL -jvw|?kNuMo>Cwc2hjpWAIoFh3;=G46HSggr(a5zJRjneEaofY!7+ZZ&wdS1jJ7V<&@3P2yEx)Da4>Q# -sOP7Qwc@8*LITTJmO;hWbZ$>`uzjRP{4(Xc0fGO%^I=ay9y`DPNDY0HO_hp=TsEc^15)^Z9dXeKO14VX1!|uID*tQD5~;7jWQyg4BKWm -E!oR$!Z<72Y7#?Kl!>L%7bbjBFDNWvxWZh4Y@EL5}1<{4Y(UuA5qa%d|(kdbQ0hadO1v~Hi#fP234J1s2#pHepl7ZU^ -j5BJTI}!`73I8KC3PWFzALC%G^F0mb7yd3MncfQy%NOb|0qJv24A7jcscS$;LHqSHM^;Y!2JMnz{d4?130 -~;~*%rNNqmU&j}Fy#2Z@C!q4{K75TUj$w0huw2lt@nJVJw2c1o|B?wr_=SzJ$(h>Tu?;+;m)u+vxOAveohRd4wQT8-%_ZtqLkFCC&^WnG?S0eDGfL5RyPx@8ba$$>dov{7Uj)v(+{!1Ne^# -Bz)n2*u82944W7*PUE^X?Hs*0s;4d&Ws=)sd*d)QG~H1rt!<}>#j-|o=eLvG0C$6DKGurIdSTo*OqK0 -3qoTi47>;5G4DW>@n(-W{r_mpIm(yJn1^!9JqUKAj$81 -{*2RI?xq{Mm0ry$Jd#*dKC%*T3Or8&qYcr%B*W1!AoJQSw7MR+9FE4bpdA1?b9DPllZN@z+uFFk^Z`@ -a=zWPOuxsTZ-(M`Eoq1l3~w6r0P0rto_jTO&$w4C)VoMVXhtw_G24)TQ=YZb96!-Mr`8$ZGSRYe`(@{RLZ|ufT!2Olhep#*gCyy|_L)V!xosA&nE$(UvpKXF=C*n)^K`>-z`_(eowzzKq{L;I{+6o%rp-uUi*A2kzli6Lobcj$!D81x10BpW%(y4gRL(Wa>nW9e4 -&29B7mb&ucFg#wUD~D`r%mkBo4k$G-|L*NIvyVK;@VXz2 -z5YBu#Z&=A{V2<2FG1JqiHSR^*jGaFRKYr;8upNPvo-)2;dne9z6$>_A1YNjq`0@v|S|9KNjFsST_h- -%eS?f-JzXj#y+c-WZN<1y1(LSwOt~Fkvi(=Nd9{HGtc{?ohW0@1t&{_u59p&C?Dn=dmeLue4j&Co4E? -)CJ=`qpR*4X`*-`>f+d)Mf)Uvg)qx#rdBI=rVG@3}Ww(dYC>1}5i+I37@h=ll%O#&e~gOINy>HXxF&_ -C~VCeer#mV)6N?@9Eo;CW<>#)VlCN(Em%$xLvvGH5Hp9KF+yo4&rdMg;re)@{BvezI_&A5j+p_tVm6+ -$=IX*TiCH;z;}R=HFLz6`FDT4@`TM1AlG3bIe-&6Gu=l#_!hxmu-{ -s+6eQsj308O{hF@M<&|^hdEN`#?Piydd!TV0@fpr2fd18r83GB%z;w+&LOMvI`tR?z -y|yAr3H}uX2d)TMaOTf@GTFB`*wx1$G$TZ;6}az|zDj`Y~xW``2$7uVm89WJ0kx<5f!O|D)YalC5o!6RkYhD0?>bZFKbZbb`+Ql-55>zTf3KUM>^qW`{Upi -`#mk;;mOV2`*%SCrA?lYikO|5Z=l8Ndr^vB=kL>q}&AY^vdr)&FTbevd{F1kPYwNNnz=Uxd`i{-H8u2icGS-%N?y>Fyi%7BBHaZ&r)|LON4lW)_(v?^44q0-{D -<#*fOY;|n&S6w($qeyy#JNu%nS=-I<0rK?cx3mWBbw4&2tMy@QV(2T -R@Zh=tCCLOmVq2t~^Mt3LDaW~Snr%Y{oB?%pO -B+&7TCLJHP(DCp8n{?bK={VL;az155vV|e{D&ZQxT8(;nEG~B^;L?z$`Ya}DKN9Yf?+*8432iy;O-O%r0&dSI!Rzr^3Z@J8?O9x__Bjx1h;! -5f+1mC!4Z2ar^;hhNEP}tW8S+|g`ZV0f;R1Y_KAX$e3SDQY{*NGkZ*WaktD0-+pJT}V3<8kBcRaV=(o -QIgUXgpuHJkfhFFScq*M{DF=;1w-qtBr32n=_0n`bT!B%Ha#@_BZ<7EQiaM6!LK8LM)?XY&24fV~Bw0arNr~d^EgjA$^vdft}S$FqAa -m5-)>XqoB-A|j;Y1YY>S;sqx{StjPCn;>hJZUH8zFMU-JAk`Z|kwF2!0o@4IYZDtyFRbuQfCTw52<3* -|b=!N+Vxr#Fe0hccb?cZP_b2O)$(SWp^-;c|VHEl7C_$Kz$Ly)w^Et!a&|#WJ5CFxO-8`+a7ARjeKSguYQ -Y9tW@Ic|m9oi~(I&AKT+YJp8a(&Zl&BmILAX6VByhANl=`cpQ;6q{ewtOdrbWTCVm%H22v`Jpacz#aQ -XtGG!xudC#>*W1rckoCV&XFx#;(=i#lklrwxrl@o)O*86GsbD<= -FCkdA`W>EV-_SSgu%hv)klE-bf(1C2FC!cnS_Je3!q)$%k`_T?^LJ)%I -xbJVEkosNO=%d@*JZC+^*pdC9hww}rvmjrdIUC{r$a0Ti=-MBwzNt2G8f}} -h&whW7O$3mR>w^dWMKAif+(&Wzd8KT(cT5)T&dz1w%4e+1W;w2YyBJd;(+JC;4NY!~G4oXcoA -;ceZu$E9U|%WZEt@lp6q2lecsx!|jAp0Umt^_w0w=1-gS?@rOqJ+!xEG1f<{si`>jpDby9fHXO&qWjB -}vfGH`n`i@Hnv~ndxZ!o%q)?|eY1BU=UIXxtz>{vCMe^$B#W2rOW}h+h4#4(qN4?z8fy<%)gtWP+^MY -shtU%q4X(GaO{LXaMNOOj9_%%oOq~UvSfe6Az%J1{dv`j}>F2G>g`AAn~rb%A?o;+`zY};l^F^ -1_sc^H-rx-0aRqr}=WNf)|57PI?{U??^bcJRAIOY%Rcec8NM|mFFIeR3wAB!c%l6Cy0imRTZulhn%tq -V?e9^u#C9{bIzzgbqcrpEUM2Gri?OFT;Ve9gS3Usms59Ern8c8(iqaPTzr>nu?_(Rt!0f$zF2H}FjK0 -Q>c-pm$?uo^yR)J-$5;oC1|o~zEmfbqK8tNv^(mSLOZJdMK9bJ54|Bki(T_WvIkw~ut$*nOl2OTD?G)IWY7>9R|De4O>m1Ih -Q1j+K99jylJYWo39|+`ZpfR1vNc&Luo!x*mCMgDmxkDNFM;v7`s@onn7YIbxqKb5B<3|B3xsZ<*s-Z_ -9H~631_TXxdIKMBj+#HGG!q>`=F3!1tFr&tMYaH3M?(OwL!h?|@sZg2lF;vb&)NUxn`KMf=!3N7TB1)v)zhF>T2cushvaPuU-#Q#T+DY1r`{(Kw)c7G25;%-}uyxHtyW -@vkbTyc&6mk%#FA@Qt`_0X_#SMdNDd=T~jcP@vM>OKzU(hX@=NE9sjC9@N!)bDmV-|7a`a_}G>B`@rw -S%IeshX!a*+Ej8Q&;!7yA4P|p3@_LkQ-IIiI7RPCvNV~xyjvoN-#P5&X^Mw2TaG!3Z*$P^cj#EI#4WN -_U0|hWQoAq((Y7GbTv;2>ma|zt87iO&3mpk*%RWGhX;a)x-EqA?+>r1+O{Sl$AZf(VV4cs?SQ)5;y7U -s_6em#_nrP@>JcABTIda-Jghn~-(Qk$X~JZ`SpNKSX!`LYuaf^IMv95lJ@3r2)Q1gHZib9 -6zC)E|hb7@_alWugcu3e!wP*-*0I@pA}*$WGzeX(+T*yHRqD|27kXClYWnBv$#I?V`y7%+VWORG*2fn -AAB=8zL9jiYl8O2J|+4PZ=(HeG27peCfZ-n<@i8&8)*4yP2_x)Ix$K6`w4vVuOmOlRu|ihCU2UMa{>r -6d!DhCK5qa#;NBz!fMvj5bfT>_qn$kine+dvSlJcrkn?m}h4N>Zk9l6YZ{|kg(VBzUmY~s(zc)KsA<++>rbWQ+qUVAk7eK*ny%7po4)c>jOz~1Y&^~HXJ(ptXRhRYhvq%=8t3x;;d$!)o_ -Q-jA>Y5hacvskzu$cS{*@nI -{AK~`F`Qb>*f2!jj(lizu0`gc;!<0zNGPV3g4HQ?@LxvR+}r!8hP$>^SUzizOroPeEI%EjcXlz{~`1J -hgNd_MDw=m8XvQhKG&J_xo)LL`bXO&4UfQ|ta9<*xqchlE#CWe=L(LEyWUo59%C#%qS6qC6Yn`^#Chc -(nO4mSrLN%oOZ}l5p0`nbTJcDx2fv1x-qZ=4bKy@>a|9c{psl10q3&of^@d!xzfG)^dS)AR)NjQ~wgs -*OqMm#aGBnGSH|`+=U%Sce5<#vl^qc$6sIzlhYM6)Vt6btZAZjledrR$rqs~^Jf09eEzR|#^jUvCS~4!-zklAOq89ACb4su3iG6$qwJPy(|j?#c -egx?PVJe!9na_G9#53hY-#5n5=u^ILQa2a+PSY!VCS}*HX<<2&T-B0G~bs;;&$$C5w~+&Vdth=?A+HT -uyf<#Pi*Isq?LB=*+IbF3hR}qH!+ -gsoerz-KhMRy_(u<&v{l~$-Zf)B8+Y!Id#? -JjH4*-GRIX9+4Or7XWu|M9XK{yTJw>)F^#_&CzEyCtn0Uu&3xs -HhQskxr#51h+v&Ih>~aG$mhFYOiab21CfvkRw=+_VdE@BUIPH8?%j*g2EyobLbZaM9eEvak1y&NCQ6k -JIwJpPggALEe69DW}IQJ4Mc&&5NBO*Ur>WC$}V!;nDZjFJb -AW?^x@5Nsx{fCBhQ!Se7;;Qw~cu?_lWuY<~hrwXDru>AjeoZuZZ)FNoUT};rNx4^9;v4u0sddgN6slM$_8E=R(C6j57_y$E8_mCJ^rf{p*>n_W6Xh@eX<90dzva`=N!?%y)Wk|(45HwnLDO#7tHXkcTuK&AsX_e-Y;}oQCl`Wp9Ue8(aPUm9(sr*hK<#)P;->@<7#_trx?-;M -i_ibt9x6wp?_rae)oNncJ6Zv-I7=FjgA-@AAzpLW>ZYp~Repe>&yWXMr?R6-Ar%QgPkKs4Z=HvN(ESK -|vyHM`H`$ZGWVxB?p&xbtKgGvSlZ|f5%S)9PLyS_E)&Q^fhiO0k7SwY2{Mr$mx-oLZY -2|Q-?Yj^`a8*l%;VS99GUzkPwhL&#=G@Ta}Cz8Jonxfx@@z{^YiNEcvwr#TA|ipQFao~qn#9= --lGGYcU62h=YS$j7xJYc58#yC1l(flq}Z>^{h9ro*Q%w6!``>Y#Z_JV?{f~1c|a0icx0j(NE(HS(VGMasFg$F1A?tLfL3eo4Kd -d`fbFlCs6kDDq_&fwZN{XflG~C;A24dGm1w5g2DDltY7J=Az5PA(m@f+5&Qmgdipf9J8Foe5F1 -h1&T6Ycc!ZxnKD(r4kC6>QRFXKVI4BkefML(5wf@q1eR>&>yoR44BzD}xeB;mfhu{?La+YUMEr>3AS^ -Q$6`a^LZE7I>L{!~gM4D%Y?Jz7=`{{eB_17vgTv`44(tqW9H2J8@g)bQVD0(Qm9vl2Ux9l@u@5Dc0zA -nqwtHVvqfMBOft85;rWfcpc&6kYaZA -DxLpuEs?VmQuYz-Yd^kJ$QMZ@Z|@dIq~t4Lo8f2SQ;T4m?d*_4REa}{Rr(bMU-a)$vKy;`OZTIulnQX?BkK=w|h9R;(d?=egMv6`PlwSt#5}+bcGtAJ+jjV918 -Z`@>J4a+bn9`mfvo)gO6@F$nPLI@9KOhw;4P?t$U080L@c_-=q9=N#XInr_;P>{GK`r-37KG4ikF@Iy -K3Dg&o9yHnNOG{__WWaK8or4%hkXx0oJujaH|rk2$lZZ`-(j`Dke>3$#;uj1Sfok5^7>>y|MsZ|uRIs -5<@U$ua0dy5835gUo~ZUqEff*j!2Sn1$MX)wIU)A4AuoHu)cLjqam1EzGiPqqYT{G@q8ovSU1c+#$%l -(-*i7u|g+}*@N_rdj{Z8;7+JYYD|Varb+_$A(OtdHPL6-gF@GFP+5poXwsh>cWv8o-gps9zBi2p9A+` -cMSb=KAJBY6{CR%f#GhTf0C6Ai`vFRiG-y*h>|jscz|TXn>+c)2PiF$NX$q}n>gV97uqT~VUI;n27@# -(-XTmoE?bH6DwZL}tPLf@JP9MZeM4C1h#AJvTR~%;!Hljb~NL9`1QQRN)QCST&xudJN9n`dAOEj%A>i -uto9VzVX{0jVeyIK(+asy(~JtS?*p|anhdbUrwj^k-A1eE(L4Vsjx)jIn -6FBwJ&k^&45W9{Ttyy?QP+8`6K8o(o(=AC(Xbmr>34|6dzQ>P`I$@NbphUUt47w|6pCj40LpJ$#ofmY -E(dXba5mSAFO)x)?qI!S_W1%`SX>HU{yGx7XJlU{Ol^m1WLDP%P{Zx0q#P7{{d(r3A9dVv?DvR2*tb= -G>Ri3^L{qLc))V^i=sC^4HEzg)2(Quz8ZP$v(c1`Y|jxy4VGd+XU<}>SM&vcUMfQbRQWR!Ft*e`U_x@ -wkH9>2U=GQXBN@3mI&xvE53^NDQtK{;cDnIRjnx6IO)c>gSeCeP4ZsiT&DKR76R-o9SlbDZRQl*fJi; -@C4?ajX)ubzc4IQJFQBmx#I1H=x^V^SI{<+fD2F>1>)KrZK@Y9N&<*e1Q6-tSdc-^gQcGuYg@Ef4(C4 -e2q=`4e@fQT_{H$m&<~68F&^X6Ylm;u{E2CKbptddbqzky&I-yC1cBbD#pw`hE1ItcZ?dh#pUHJr1lQ -YE$#qKRuI22XvfQ?^}y%*3Yq9FdCjS}12J8VG7%pN-_;>@VMxz?{%M -S?lJ&K@dcM!Y6^OKF0S)E(*YIzt_!@4*zGNEnG?K1;w>Yo$E|m2CgxCk;jBh*D4;v29lXzA}C|?7<@* -LeSHg&GJgSxO5m+5tJ?2m}Lj5;E1TQ0O80;7ZUJ$`l~OII1b15PoQt|I+LaW2UuS%Jy~& -ey0C>jci%StZ_kIkg3N4ux`PTc-22_^(m}A1O1u|5WVW50wnkdDFdzNuB~`wvLCxaZEFnq64rIFybSO -09%mrYBT47p?Qsbm>&YCdh)$1>iDg=O3<_EDlvZ4FW9J@J1=OO5@q}hU*uSEhD&qR6k^*wx&IBXuUoH -gkkggZu@U_|Kkh#nQlCw)FU1L;k2s^gZc~_Td$`|PM19iLqQ0!F$LpiHSN_L;rM{*M)fc}13YuO$ehw -~Zd$kAkc2T{*q9QZaDA=#a(MfizC6CZ8s&-W>z4EX_t#g7vA(kYZ>+B<^PyYx@#Odzd|Au~V -2A1RFgMCP1b#8vhq-9P_FIr9{JkQ_pY)B*&hA+r$hWYUEh&!i+!)i?w_Gxfz*qCghDTcA^IP^X8-?Dt -LzgEfYd62sl(?VK<`mldu`lM)9=)3O|0>d9h&gxi{T=6iy3W$D-Z5(UO-+VN8;NHM9b<0_$%-U1{)Y6 -*RrG&1_2D`CJe$sZJ%?`MXK5^k>N#~o!2JHI(Nfxbp^tVM-(Wb~6uTppqSN<9y4=hjSVHcD%oEv@=EJ@*rZ5p?yz{iF32bI><-fhSfn&lAK)+An2+j=5C+74&;5y(77_J)H% -1ZD5|Aisad`i+DV()orvF15*zE1J{h9udhxRH{;Bb<1-q5Ua(bODg3-}MyF|V4(tq~`FeHoyDJ)>bb- -a|pP~9PliYU0o;ectyE@GnJ3)i27(s&p($id*s~tAUS_WH1_+0{XXdXQ;J)*TB9tO^zJ|G>8-qqqPDH -`67F2`{{a07?QWR|&2?s$6G1%#)+pj|PKbF5>NpubafUbIgW7=+{X%(HTgcOI{o+PpHA+rskukWfE>Q -2S!=i8k1`M$|W6Ce}6Nb>si^IF269y?R!cnF=2B<6FZGU&{}0vGFNy5LtG=Y(xIUqgEN-0tNe)?vzEKGVN5k7F9;?~+CxHp(C3cNDo -_x8od2tf{whXicTA5I&*ViERIKw7zi9)fLMspC#T~fxRz;KH0j@ZPVYhmRh2Ri)dk%QWjK?S=?To7DJ -0%&W#FNgmh~u9rOapO8OxDOh}L3PP`lPe+Kobef{!L>VK!vSNu10Wk$U7Lt0DE(zD?FO9`8QJkB81u| -*dCA2eS%jzn}hA>4id{*`71TOEGJ1FZf}O%E6QaF)#R$1YRXz~>n8&JmBK=wXT9KNs!OV?q91s>_A>dvoah8gR;JF9i==m_<4eJv)wPRA!=yd8TmN^v3e34A?uukLz0cY@p9|+<#xKA5o?_k0?`Nf0FC? -id5GV10N1n3A~vuvxr5DGg-$odTiPxsSbDKjzSS*-BBvlt+24lKg%p{m3+|x+<~r=r!U}kVY2<69dYm -#voBac|1ZUz0)2loj^$kfyU;k(q8bxhi1u!YVGD7W|3ijyW-Pcu?@Jr;9>m6TER*VhGv?Sx-)P^Xl4z -Gt;=ZL0J$9yJP^ue*4Vj#N5BA7aa@qovGZf!u=kbZ-lJAkY|61|_$VwkmeM{?Y97pp9_${R_KrBbJse -$T_)${PXL^(;6mELBD+_;KW<(u_0lvT2gzXOenC|~=uw$?-`v3cl(me+13*^sfiPRZShyL61V`*L-SL -m%&7!4H<6mzY>?^6B6mO$W`jLE^ix3yHMvfJ`L~UgNog>+v@4x@l}{=6Nk|qn&wR6UOZ%<^}5jPD9q1 -R+($&aXTioftM#0d3NgM{c=Q{`MKmN3reNNzGPiamWqo!!r#m0{pom`sSC__>{C4AE_AznQ=oB8}3)WQMUo2m -XhoHu0Po{(r4^KXsihxvV>joT$JwzI8_*0Y!U9^T{|ap-&Vv8le`y0~YzsYQ?-zwxA!v`}`-iq|UzxXVV`61#nU#A9mThaEAwpNi$P -GM+^XCU@pEU*7ekWQIIr$*|J>74LQGMx+hCXo)1$NN(({E}4a&kras>hzx -puXF6d7X%-N&lk-zzk~6+A3x?EJ)`dcd}}+n?|e)Sr(4sdo}riN9m$Zu-Ws}8#OAeTGv8tOj9tOtlmE -22IL!lmq(g^E{@NmZ%1D+C08i0qUs;FZ8>0W&*NNxVI4i4{>XFa*4S&oIy*T!DmHTJG -c+r#B+XK{^^Vt=lI^6c%LKvq;xrw${g4wceLyGujnV94b7nbJ4C0p1hnbK5gTCP+tWT7VmA`jlbO)@Xef4_fLU{vm>cd& -N`k6eEiK3ptMFQ;$aSz$|gn#YE-%f^?0{ty-z=+JkRQ56 -*^8upH0LQ|03g*@&(4}l8A9y9A%s?*86l<$T($0jFT;5oZdb!@_c=gaf&uZVg8P|yBeJ3@t)y&@mtIz -^M2|0pB&5a1�jyLH@$t!Jq2NPpE(e)BB+GIjc3o&4qu;sZ1$yB?FrFUjP=xeD&?#_#)B*nQt--1i;%BzWI<|FkIgeJuRG@97g4zV9>ko`HC&qnO4wld-^T-Tn~iUpp -4WRuFW(aQdHp7M?y*?lEwF__>9db-rU3cDQbu1J9Z{_F;V)3t~{;{dcVe`-SuOZROdC(BUl{9|rtOku -LQZIQ5{%#TjfXbRyV&WIof#W!vnZssYkRcdrumpTJ(M&*VH`Ei8$fXp27&>y*-g*B(n3alTOhY%|GDF -|6)U?w6?4mT`egj7M4(^v~(Er78f4IZ5Sv~1|$1nyzx2R_@Ca?|A<0Ej -~lto-G;5gMhj*b68^|!#Q2WG{L6G7Sjx9lT53&+W0#X?aF}Ktqgi3 -HI4g8yvq>*L_Btv>z{0KVC*ZbU&4^zc2im-!SxuFu6za@;#gOam0FnjH2)3Pn(XF`poT1eDtgayzbXi -1QyGLIHJ+Y8~gn7XrV3v~&JI4r*_<-ndsxZvlHTGISk);U%l< -b4i!m4N`5Be9BSG9dti6PE(G=F%Gn1u|99!5U&EG1bPD|95V!pBfA8+{J(;$hQ=Ho)emM`q>ri& -;vkV@5^(#Vl#Gp|P)?o-fb2@v|9m#4D*E4^n!ZneL-^jyKfL3Xq^|8`8)+wWGX`l~hli#AX!rw{Krn&UqHRq*P_=e5xW=S>lz5GMr<5eQ1S%Fhn -GL0>RO`!Z`mV!QHox8RlHi0|GwER1i(M5-eo}kq<()W10SmKMdacuw58I9$4Q(N5d-JTBL?Lm#0<`@Nz9wvXEP9e}dbE~Ri*uX(v`;)0%j4 -qKmDJ9ts7-#n)lB1oazOjL&CG*2V3sm^O^T@(7<$ef?pA!eC|~>YI8xus@&~&IDHq}Ubt$FDgE5~@7JY7l1{|0Igc*SJAO -5ABpe&-(z_x*$t>vUUK(g>|F3)NR8sbkBfn30oQzm#a7QoKv%f43ETc4Hh%!g@PR>+NCU$2cSS@OH$d -V$Q?#em%W+>}GW%>20kTo8focTH9sBIJNecNY-ZHr7$U_d565QbxwM49lR&6+>kBPs4y&@qQ(! -pwMyhee;jZ!?co^}8%xg>%=& -rrf=@%u4gU+$1GfAKh)9;j&Qr9LCeP$AqMxg7h*^kF>5Stk7*vus#)QDsv@w6`G_(t1Fom7Fk7na1-k -)baV^*K%DzYG0bzxtCPoQ!TQ)jyd)Yn=>uw@u|36@ywb1-{=Pz -1|muQkfeCB?V%Swng4k@h`n1f+s-|cKJp*g<1dSO8Z@hIaR+HdfF@J!Gbyy-H9WANk|^oKlwF;In1ZF -vjh_D#8~DG$E~k};oBBIZ*{FWzHL4d`*<{{D*AvIabwA@*-}GyJyShykVm^4lWv+w}Qp^U@mkyh(W>Z -*Bpuc*s24B<6|FEsy5=6;l~7<1NxH54^q8ZC}mrE~B@t>ZROfQGSZ|c`AGTI(3V|ubcmL;j!fLS0b>c -Cffh%_YJVuC;oMc$FPXyHXGnyxC1;)vO>{X)**bQQl0Pd^&ju7NZ0w;${-(uuLZ^oavH{TA+6_w#Jkd -I%zI>WZ?`Pg(g|@VmP@>@@FKD2a$CNlBsTv`J8pj9HSY6luH6LN)-*wnhC_~m4;QI{FHzr&x`bbc&HD -|~Ejcfw_m_jkLQKnG?{&{A*1C`AG4E~XPElcz)T|pM3yw3zQc!0_9O!l-3m1~!@{ -HpmR)G1pGKTGUQyGwF=G8kYzDj-jU&K#_q)g8f<|)mW)z^*!Uu9UwK@MU+C-kB||M!XS(t4gtY5kN3d -YVjmOdkh*s}4@GR)6?sasF^>xfFIb<>vStlJ{#@L(WRuKQDOooyCs_;j586w@XRRl>Y?Jx-qKq<$CF -ANu;O>W{Ap0X`uYK(UlCuoRcc!2T>8Y&XK#$SPiwbH%WzMPeU@~MO3bkvl#E4B()xDm@-XZy-PG?DI& -EMF`8vt1W|9?w8HVSLx{T=_Th%|TWIIW}O|Ft03*2K3{SwjT18Uo0nx{Bt_z#-X!0q=>yS6`p()dYU( -Aa~Yr6sl5_kLHZyBo4+47ID?eh=EX*G6rPr|%4@4my>8F6@93oTo?*hkkV!*lRkEgDud2lIZ>ToYrzB -rQN$#s>8kN-k(Ty!221`t3=5gFY+FxwFG^1JMj%*nt@L^h)*1%wEVo3@-0mx9x^VRI8i#Llg0bqqd-Y1K-&XsVy;B^BascFU}#BQ~lO2*w4*lJnpe4(q^Sy$8h8K%^zMHut@ -4}izLg4fxjDeFTJ+d*Q3V|%MF=3z3+$^U%1ag{4nV7%dv*fYn88m5R4_Jgv1hyc29Luk2PY%KfyXgyu -Ilmbv>=uH5-0idR@Q=MP0^OPT)PU2eHTb*~sqDm~rjY9$+O1e?x`G4dCY>Y<>CtcD5B~DPx}m>qA}KX -2HakT`11iMEjGwr{b=Gaic?a&BMHE-l5W -$ANT0XPR@$7~_c;i^jQ(>1tLnq_yE5rrwB)H$eY^8#1txdAvls0i1&n&sooXxY%aXzI2cEBDLRKU%Q6 -pNj{GkwUOvsU0ZLF9-v6>r*QpcFPom -v3==Q|Bw}gz7^POOqf!!ZspkJ2F3^0@(WF|6`>J+-1TxCYl}1FtAPda~AeIGCmroCC#;Tnny;wFx -ZdWepqX{W0vZJ?anukX)Pui)2d|VL%fo9iSMa(z{G)#(cC1reX@I1+^PNvzASpZj#$S%O);X*#fNn}H -(C6@=CFtnjJtjwSK_enEyM3chxOR3G`VvmJwB^VeJn?h&ss7kVH?(L2akzj_Lv6M!qRDIpV{It%Ure& -wsA%{O=dwhHPzwlG}KlHv>pGSWvX!T^M>{1u35~7*ljgD-X~jzcx@yfdsK;^p#~n&&!ZakZwHe`?ht3 -{+y(*v2U-NLie~?6#B78-o=bDawNu3za4zr7umNz*8ih3HEy=v)BKV`Vkq+6KAmSr+X%rZB2bh-I7P`EQ#=`dy -eFucRv=2L{hu{0 -&E}P$0Y|bNhidEA^v`PYB;w2aS8ASci3|@L0orJ+yAM2Du{@`xEI?B?{}TAbsFuytu15z6AX?c~9QJ3bis;m~RIbC-Fuq8}dvrmXZ -lEV7Bs@o7`r_8q;2HuyM~-gT*g -BocOrXzGv=`Mptd350mSU2oH$OsUOs7v9@SaJIKrE+09xvVIE%{Ju0X?46W9Vd3`%pi=|1!najF=q25 -I(Kgy#qAvf2DRICQ%7PeR4#7axaaGr&LMr_0zgIOnuCv`Vo_5&vYIaDGzeX{k6HHTW3*wsN8fZy}4;j -t4vdZ@p|TehBl>m+vT(%CN!lzxnWaly72Go?E0QMdMTwlO?tRS?-08XynohbhVAm?ezMVYsw-bN?xUl -}j}sWp=yRGkpz9j2F$wQ~8jtJaS8}}?$!w$>RvNlwG=0=B+nO`@xkQE+F@XMbO!z5~OxP)Ogy|&vWbm -_f7twDZj^#;;tyjdjPxEf4{CCjvOnP3uQ|-V#gA^-bP?Li{2XGfm66>|md^598A8VsEZbIx3;<3=HVc+F7albLht$q8BfySv`v={Nd`2GJe5qJI-tjz>lbJZ-?V -NOVE#=C0b123Xpk`IW_1Ry(@v#9Pk5zqV6N%Y)zdZEj|5Ei#OfOSmcanZ{p3%+dKZ`gm~T;tj4>{gDC -LUYsfd$x?@i|!V`X{dufg%4gCG1rql1^qrHQD>)^{0KMOlL#%;O-u|m(UGsa*b#ofyHf+k_Nc4&j_G49YuiGDao#o -1mYedg=fW&cTa_gtcO1$Y=^dx1+-mK;h)|K-fxEWf08qNqyKs6|KX7SN2Uwye -@c}8E7bpmif1bPg66_*ht&)fiyQuw49K!7&883L)d6&thHLyeEb4m~$cR;r?C4g)VLr -p6Zoou8X}?&~K3Fhc(V)x184UUYn(Mn29&fCE94RmCrTUS14IY9yKw&d4-hTi#2y=y70N?cUB?g4}Ge -azrw8BaEP+~D;JA2G^PWeiuFiq#Bah*V=zq0tpIaTGPj<VTR#`I!-U_M;m&Es)>Y9;!+XfCxaW}|J%QXq%k -x4pzhEmBHvTN>%u)vF4K*2`RJtN`P<@tZLo7tIx`WbU<4T32;7;@vI-7CwAOBJD8w`nC#LL*ncEqWw6 -Rsfts5*|fe8!>h}}$Ai|qMezA00#}7uSgwiN{gUiX<(oMESWMk6ns-*>fBxIm!jWC9aOigCfsfA-O6Q -DK3&H1+4)+kvrP+EuOKmmf%buqe_|I#3cT)X0qyAu5f=8i#z(*@fmz%f5u&sCDJw0EMTpajtSBytfm@ -kI@!v03OLj3 -$jT*3SFm)l1y4JE~HH)*GzfH1N -=?dWYT(qEMg^2Csij{e~?2vRK&5Q-rl9GgUms38}J%!sPB{WaDGX{19Nss9DKXp#aqjB~;+C@Ch8fk} -b@VSr}ZBylr{EJuxKbIxGSE1vd%z>|53gQ_ur|CMJ`=~yWscQ{eN%$hL>l*udPYQYeIPK{tiGLxFd>z -em%G3D=tp#Vgdu8|**qo2)dSY7~n~%86j|>O(R_JY5BZof@>OP^o1Q^s0d??yvxh^`7Bl_myJ)`#<;q -Ms2-`=2dd5k;QucZk5hScj=U?@fBg~v*LPFmBzPwb}sHZ=Z8yKV!I_$QI}*L5U6-Eu~Q&6lwCr?N^VI -q*i9z1`lh^Ye|~B|f80*ihkK#JyC+irsR(8n~F_EHsuMpf=+_YK#xM|}y!l7pnc^`$x`|Kn>>rQsz+zS&Mu6^$}#Ht_6csMyIC5%)bo@k?D&my@^O7?hTw -1e_VROeZc}!jj?raH<}vX;O>4^H7|GVS#=e`^s{QTjtlP}h_w6Z?Y_+OI$4O4ragvEIw#lYi)vWWyZm -~Z!Uc)ND7aL8aE9+_a`Dq;UH4yJ^u+#UIte*BCUjy;*hBV<%(?)Iij@ZY7JXYcN_}pfBdo0Vnm+ud$< -NL!Os4sqn>}(wc%=_=ZR`7ql;R -^0?@=(C>){XA=*Go|nVZRTmVEDRsbLYG0aDu##wi{UM3kjWktMzo8G34(!_}>`gbDq@_MnJip^j2L9# -!ybtN)eRy=Q*yG}hvpnF1KcHu?()XWf{M#4AZ5v56HP=wvYv3Kv3z(;u-c3;&``$>Vwm -7G31D(LXJL+9)HqK#dKqrrGN1rHc&dYIUQ$T$>Jp=YhhW-(Wt2`d(#c9NhfKS&6MR9()Pw=jqnXJOp? -QAyRsTR`wtHB&XY6d3AM{Di*`*7Y0p4NdrCvvgV^;(dmwij?E!DhT9d)=4AD -27Z}?Nb58J@Hf2Mq|6>+X_AbzL;4_q;MdAo}(y=F7>ETVUZm2_vMUTN3~V{ZJKcq8JkACSSr)4cyYF8nQkS=>qc1@5(t{ -IFR@d7=4#Chq%kQ*``68$Sz4@v@^@i}4@lm16v2bbUS58|rfr-EJ>@|BIv%+PIqZRn$9M*S}#S_pj5+ -#%YyD?{KH`%j3d_W|dwxWYV4d98lcJ|KLU8N78eIpFxE5NyjsfvJXbLL;S<|x$0?J^OR`!*ZiUP(v8W -_;w+lRWJi2sIb?B+&)t6#GSgnzeJ9^?rpm@De@8sE@(DGtCr+*R;ZE;2YJm0_9>WdodP>KFuY7_9w#K -mCd>{U$8hD<^Kw_OY3~DWz$*dD)7}9?1@k(?M|{FkPy4`3~}|3Do*f -qg}`^`?$3*#f-}zfVqe_x?y3eV*IXD6F&bfqO@9Kf>+a^4#>B -(dM121mid^Mcb*o4B>O$huFG5ejslYw0SR{<(^bhdv^fWo28B%j>oyyxyOn~j$9uv;zy#q^pNuMsk}~ -F>u##sC@*a->j1sKM*Dk@+*OY=_wc!#)@vhrHBvo@`5qV%bK+ichf$6=A2=UxvIyBA>rJ9x49m-or+q -5U)Lg=Sa$`IWh1C@eI@Sy-8(TxyC9wE9KN3C=&;x=pcBcl;5dC)o5ex?g=XQ6@h>62wx8W- -A-*+&leM-M=g%u6Mt%8|jO@i*(5x@1$>7#KGhLU|p5+Ih-Hm5g(z7dr&+fvrYv|cE!Dn~i*0Xq?i32RoAx!B@)1!a~ -O2aZx1RksueM)3Xh;=mBT|Fbyo(eOn7>^pSZGRSzPF>F7{be$Ha)?0u_MBkA6u`e5}vzKC$0Sd;( -Y`c%Lbm%7mWyyYPFr+L(^BbA`V6pdzWEP6^xLUp*Y;n5cCNajm&Vnzis;~e{Bh2wzjaK=IMCbMU)aWd)K)EK9s~38IF$>T$G}a?p>l+M -V~W>kXO{3YbDldV{76eea2^i94pNz+`$bK6>hhXXmj}x^extm6PY2a)*r-)4X8Tvsc$VmA`b(9_b5fg -gIqncQFwciCXf0>w;%@uej_^C3SI5_hMb+(>BJdzA-%$gvQT+hb&J1R%f|4a1kS}!ocP_7%;QQsq;5 -K#7?8KQTafb${mKaJl%5Wi2*?<&gk8NMG=0}Dv*L|fB|cK7X!@vJ63^bXE^WYACef;GzF?KwOI`xP}X -NZ;=D|2f(N9=Lw_XrrEn=4jqu=-q)?%(Hbp$JAc-qcMT0^!AY8C6T%a>8qXME_gOS4}>4P?EW5y(SoS|$6U)B9npk$BHSt@Lzar(HyFL-JpnjJ6mT{H}ePRpA0~MsVz>fXlA#n~R0pDV=?6HbC^< -KRD`A~Fhgwer?xDGg11l}+BwT(f)N*OV1Ns~CwW!#pupT>T*rg6N8t+e-%+~vcahg-+ffL?~P5xkGn1 -onX;4~OBDM%YK{@ya;yv~9wNDAhYNS?DnEKmK;| -&1Z0Dvy=})6O7nmJPRPHJvU-RT?03z=-fJ6aqH;pd+oJK4$?R7#C*YTG3$^=Rsu%V3Oi~}(ICwq;dozT-oQ_GcfywY&rs;FILhzH#(e -Qq8o|DPKV_%QYuMmEQcQ5++8hH!%7jsit@+RO&C0!SAN~YH;=JmEB&HH`>KQ9s&-@soAVimQE@}3SW? -_yD2C6z}qx9=6dF}H$vL=F?H>%x6;2CL&S7p@lZ${dx9`?ol575_2E;UB=))Q)Tq%VHi&=>HYWqsK3Q -iw$Or7`^yS??Q1d_+8GD;9U-4p;YVlJo~gDcExzP66;`+)G=9{3K5TJ+*iU0d&?PA9%2hTyMz_C&0=+ -zNw95Wm2DR46MUhQ8OM5Svx}G`z=dnOoN+w5wrfOOG50Mz&SqV^IL~oDM{PSz+c^)F -8v2dIPFJkos0SwecLsTp!)b}Vc$Y_*_o4wp+UBP%r6lQs73BRPE6;oy0|mHV_7_&O~jdc&q?FZC&d(! -u8mIMhxW(^E3PPkVa0#IaFwvq=WD(j0hAHobNWjs3wmcC3ol_W{yb(N90*-dRoMra5ohNyfmvFN_|AEcC^bca~vby_AF{U+GgBu{u;5;seGKzzkXE6+TcIgA#v?j5N`*zWS5S)9Qht&g?H$uYfKy -Or)eyn`YHOi$BQuuy9D?R7%^3FX2s;uMSPBSU+2U5Hhe+TWv9btGkQT$~1`F0HFKcA*c*?c~%z6i2iiq*Wf#JbZWu=w-WN-6J=-e -daYp?<0Q{?TXN)mjk0p^A;>zfH0m_)|aPILNlzdAL7dBIbyh())=gXIH!Sn55A~4v$iYz0SP$p)E^+2 -i$!Pk53xP-zNHzPL2;Z?^S^h$ADu?v;vI`dVyA;6X<2o$|AG;m_Omp68D?xn6zO9@wkUC5I<(eh4n$Ip|fuULEZKLTNA3E -T_f(2ak9QblM9J5#%rzf}3oXva#hb2#Sudw^n;)TsM&f8L -|Tl0B}gf~Z^VfhTS(g@#d@3pwa~n8xZrY1e?UqsYwOM*X|}=MP4>7XQ!mmD)9(=_(Ro>bdEC{*{Jyp% -l@&Zeaxm`x;L|;i)-vqok04)aS3#+iXyy0SQd+HhUrs?S>~V-Ls2lblX35$-e92VYX?m6iSm9TwEYzv -(QaTVT>|heB!@X`nlG`pfsQrfz!*3h5v@{1|IVSZpRjI_EX>G!b4-Vm$||MLtw|e2d9x)CYz0x?n|gTz$s~)tl-~?rK^`BC=sI$VW!v2+)k3#4GZ|%Vz+EprtD!Mj4B -P$GPaz^DtS5l{DOJ*9 -u>Y74I3nA~FV9nsabh -&~A3X%yRt;NtR53FO{hbamgyBRO^YUw9aWfcOgFcTvpd4ix{o`g{jHFvp1PCfsQ*{jcMM;u`c&F$U_;aZym5QAL!^`vRK@Q7e$;k#j62e@)IxeEPejMXHC)UN%APq71@n_H$5-T-@sep+iCDuj -p~o&vCsi9KUOcE)eL?M8GjM+w4Jf6^3+O!v-t##>#_Cg`-WG}Z^nDXh(#<>%&8(h7MxXLg;_By34SPM -;s*|r$q?HpRZ1%xqw-l&1asAZufwF%Agy+_jpdwO#VNhSw7JVzH&;b2CarDKWJsr?}4 -JxZ07@x3PH29s5C2zO|1pp44SR{7ijhr(IQB%&)+v6r}Q^b-e;c+(r&W7oZC3*^ztkG%$Vc(1B1Je=r -=1KwyW&Z4C?P0>fZ|zVsoT8M^*wWVur`B*u6-MW+lD9b_3N_EvJ_a5zWy)#7OLitRA`c -by|b&H|g|V_T1;6C;w04!rwL^;ux+Nm>3t{kiR11!f)g;`7Rt6{?V7C$A`Tkq@F7WE*KY{$A|5HIT#= -Iay@;1NEvUu^7-*$FElTY@22^vNL-#fu47cclg4u4+nub}XCDaah|lQt{Tlu=LhgF{Dx!s5m;YQ3J}{ -a=G&xQ3-ter9ZIvH(6}rzWN!%BSbX?LWDsLW(Uxb*c?JPSV@vn_|tLq2Q*EHvg!2dU~T$D9LWgvE^(= -?iY$FSA{ENb|-cfilXcu#dW<3)P-1K?S<9=9|zRqWZ^<~fe#R>pF?m%7H+7C-PL$pZC@^O|v=FYe??& -YXd{Iei3iT@mB+le?k(m969Eih1wC%@dd~8Z9Xs -T079R=OV45>o)z%(Y!(34y#Hn9QB%3U3-B)bz;k$eb>4co-@w1n?{G@@W3OHkQ#Kr`BK%K;ML -(w)uUUPs!X?+h>+iYl`*Q1=Vt@r6@_QyAt{YKZB-ytXN&y{WDaH -8-)=w2|<4qPvYyy`Nexl=Kp}h^3Y4>E34gfz`f8xI_dM8fxSy+2H>|3G7PoK&%+T-M8_$=UF$HJa2$L -}sKBk$qAvl1x3g=lZ3xoKu~!0m(%V+NKl>4sUPL*mS@gLuvzx^60PsnVSn5Au0eSM^%)1ILvucIX_i> -jo`w{)ziMM*_`vk{KP+^jg?oqJPkxg1j)@S#*%w19i$F&#pYkb(m++=iUDheq_LI?O5li_&a>TzRZ&M -X>I(z7H0}i{v$Y7VShDrP`txfq1%GW$e@?5cP=J5bRcOTWz_}&p2s#b&d+oM -z#7n4PBF033ej|?57D>nrCy7owUdnC0gX9I2opN5+%K}=ZjoGX?CrZ%78U4Rvk8~V1e`-AY6Wva}EbyMq{LUpKzm4Ws759bQ!~18Vxh3=Or^ -`25GLm60UL;N9=blCNseE1+Nim#8MUtrwbD>B|;IVO`S3}nY4n1(caW*}TzKbLqKRYgxGI(2xq>K1>f -!?OKs7~J7h34ws^&$?+-j_ril)bXRRoW}d+)wLEdU-GEeN5nX;7{{80t^Jm -95^!+qLS4s7AKvI5SF;adZYyiRG&qLO%)(I%T}FfYIxHoZr4I#%!nN0!RZzImP -%(|I17q~>!TRJznRq+~mtQktcP6)fTyy~&JpY2x>3q+6r!Qrv=Ss<(Oy_LgZ2KzAqUlQmP&ma2TLsrpSH40{eKL99}TB6+MK)PMS(|%c^ACL3Cf$*Av_iQbo|Vbcq(|P$= -p~zcaYon*Q_A+kcS>C`iHpx1U41+Mw}Dle%z5L^ReCicd`D{#5qT~Hk;K89+cudN9&d2C=(xI= -C3~he>o#&;q6utmyFSV6kOve&e@n_zfX~Hoz5|NP>o`RF+n^HTge-BfrnL;y -_sJp}r#Wf|bd;rL73cr+U2|ARjhO||zx)3#<{7~$V{cSPEI@L%8eKjZR -MWBKq6s!t;sw};;2-Qbq%bNpq9`>@JsmRPcXiRv4=9kMs+EVCsp(_AN+oI-9?bbI0HT*f(DdtJ^4)@y -TcC!+xqpT=rvk?Knz9*`Z2y(FO)@p+C*EZ@kVU(b?J#&yaJC+axmU_O3!#&M!Vxkh=|12)qBZp;*x$*#fj>bR{bcT{CzO -e53+u%O=WG#_3c$TFT!iY%BVZ6)u!a`6&A$P1jjT4#I!iTO1R@c^B0`yb>`jRbcKG@_D&I;`36RwWK% -)>Ax{Qe>=u~XMHH^Uhdy+avsHgyH?%uBWiPptm8LP8?tr$rlBQWg=N&9c6wL+;l#FHM(;MBB)S$?EOp -|HlX!uZ=A}!rEOnn`g$~7B1KDSll#vhGpq$8gF_-93Jt+JKT||!_l55OdhN1nRO3pWD+uLcF1q@5j=N+78GDaqL(rjk$&7aueHDzB`WXH{~pK4ySNlTkn -r+EVta8ws6zsYz>dcMbF$8=Gkh&opDMt(%*bm+kJw@#Vx1x9tWn&Sn(FKZl6LjdOfhh@jT{lVX+P#<7 -XI?j6Hl*?xoL<>GMVVy;J`2c)zhZ!ft-!%Do`Gl+{{qGku$~bU4R&&_&z2XEOr4yn+uG -p13HYj@1uv^xw~=^`WNvzoyLdBC6T@{0BMsz4BaN!x-+z&~%c5TjVv+Bz{%a7QfprVL{aTEr#5##yb< -hh+y4viBt5(#NGeYgJ{08Nzh!Juxa7AtnDdUqaaksmW-lfr+XpB<>_fZ-5#_>2b9N!223M@w0?3yS|d -m1ZP9K)8OKIk}mp|{GTi<*c>9wwRT5b?CfY0rY~6JoP=&{_g*Vc(K3@%_HQ@gdO+eyr5aYBNH%NXI)@J!m63T9u@=BTVg;V!$J8hHTD9 -K{9RD@dd&-)u-mfy)1YYyD%lu|5t&e5ar>1P9J-R=!DX|&(VEg47tLnFIXF1&Gg8Ezr-GcbEs5fy!y$ -vMWxzkwXqms3(O-igCRuY^e+<#x)g1SrS`*Nxuzo*joL1}6&iW7Vb!iztH>I{#CNApIE2%|Ax}ltevv$DBd-hy|Dd3zpI(KxkySJd6 -W3!QpwtzLu+IT?U$G{h$VWdOf=W$iRqJ+n6+m~R(Z+9wj$>8BAnYMv1G?uap$CM5V1dy?jW -T(P5+O_tAP&MpZ*%h0+`pjE?SNlN!ic1?E|}bY!%<5l)n1R>VDuUHN2^QhO-y{T-Ddky7FP0BtJ9XbC -CEQ@Y{=ke@$|)>rH`w&DRp_k0S7|-MTNF>wFOZ+9f#_yc;LZvv5X(b0Cgwlc?J$Cb7CfSzz19SqnVku -4ab!l>@{(*RwpF2NJL6cs8=Mz%#1H1Cj^ldrlg^yXTMRd!F*CjXdsW`DY5X<&63m?#F>^&7`}2W*J-E -KcvTIU7n+UHl(n^1bLtjv^Tw>J~pI0(n@+xGx~TVZ0;queQ6F$u9QsHpApTbb^X^&6MH)O?(GWftOv%i+tdZ2WqMWVIH(A=M=`u_RL*%`xI<*gt?*pHJ@&z -A4qwdaV2nmHB!cTh>N=88LX>w64uO_Jg%{J^aq8-ml&!?95;zr_PuTv^Ytjb|Cq|%M{~7ZpRa -?a@cFutXoq!WBKm=zku)?D?TRSv@TG0`TXdRjq&^t!0aoQcqFoVW73j~1DG@@qr8<4k9)q5siA$#^`d -?D-P7%*?_P~DvWf^ojqSFa<==mPWqa+ioUKd-je%(6Z5BBU-&o6#WdT%mI<~Sf(J$5(omSK8+n)bZX+ -j93{p2=(1_4`K@-_Cf}Ik3v)S^jVF$4FPM9Hex=e^P77$DIoG*`4&b9rFqEtK>p7iT!^}leJ%rChbI% -7wG@`LV?eA2hpPXJ*~w~d!qCQ*t#sHHMZDSax=^LQP$|9_I2&s?6f!E)}U1QrnB5q;?+I0){3@~T-x- -w9Wlh4IH&{)~DI9AtAZ91)(ok*9`p!UzAIU>uXBhlQz9 -D!bh`H}Y|&S06%ps|_yJ?*_9hT^Y&AEfibfR@`pWe+5>%6AikeG2v`Wzzn<4xO6*YcIt5{4}@ldxYvo -zgEuQ)XM+UZ_xP;{r{Kdx7Z=yf0(?{j(G$A^z(R@`^{Uwy(f-zkb@*Ub<;c_I-cg~fxdU!<$L^8{(R~ -WWW}S(6yRx;-x(;Z{V%!gTYFGuDLuaeax -3LWIai$5@_2sYf3WL2D8Bm29-~`bNU%2wXCF}eOm_ -QPp5w0FZjh<;1^me_(jXV#4oOTYMfuZ6^&oq{zPQ#!A>N#_;7HA6koc*O|)uO?mreHXNK31ladA;)>eY1vek{*jhv8s`(YTFc@sWzC1 -Q34CIcBKXAiL9Oz~gIYMB_yY5}R0_`LSyJ@*eCoWGj<%klZ!w=mY#fc&;~|;8N<^Czf;>o>e9ty`+qB3w(lg#hO8 -Wq{5q)SRzJNHM@R8uQ30*A7MQua7{0=oR3|q`t=Idmmi=Z3g3}N$QZFY-nvHEB8*dP0gik%}kZ^)VNZ -q1cuaJ$?*`rb}+heZ#_Tjm~GM1p6pY&Swlhd!cbC4jG5PoY7j)mPDzsd_-Z+AId3Z -2Wbs|l-_3li2CbV``k8-^o0bfd(&E_qJ-w0Iq^noZLGPhrkf=pCTBA}hm48y<&RjK9WthC>AV8?SnMU -aa*)~42!ij=}>GI(06^ebv1&W*u8j>osz{Toz)ZM~9tnqvgcrnccS&(0?lg2Ho1Fk -+_P}3R#O<+%#qB{X9^8M3cWL1&Ka67-PE -SccY^j}&;O36^BnhHc<$bJUhBM#K1C*W3}>6!F>F6*f0oujMs?b@le0~plb^5-oWm}HZ>nPNHJcB#0^ -@CY7CTl#eAJ=XM~0@e%6{2SYb#^qgklFqT&3{k7PjODwp*rBbO3fOt6$pM+ApW@^uQ$?p)np-GQ4f{A -9h@i$Qix3`yf8Fe`EDk$F$L2P5D8K-Lu&Kk*~21O27Xs?GbYiY0v*oRs(Bje$Unn{4&9_h~MwtPM>1oQ(lh#;y#>Y*3Omm -9CzEeqbK=^|F*v!V*l?8?j>0Nh&w$*`_M4)ksBH7Mf_3R-MXts3S2ZU?Zl6P-?Wj&hx_mw&zX2$O?58 -CoejNTN_972?VQziqnsVw&QfZ2&{|@a#`42F4ks%-)gP2QAN7ei&wmTqcmB{Ju%Do(!2aC4R^3w6#Ev -Z{UISU9h~|u=N%(l?FtxB6_~w*u9?rkVmh@+vS#ovt-38vP9(##ocII597PzV}E65>UY}V7yj>`qToU~55;#6>DNP@jZ0$8m>SHysxmTrh7qo*J -$8Vl|e_7rkaGK97;db-R>-aopb?DbwVI#<6GM`f7oQOjnjJKHT{k4L4H^F#|_XG2q=5TtV>ML2L`kWG -CP(TI|cwCUXijy|lMGQvVtE4-Nhqabbk_Fc2xVjr42RAfq3B}~v`1s@HpNGp;+9CY3PEbEq1BX%;emd -XR?K0oiak*fBxK+pHf{n`CR^YKBHl)pKv`gS{r917^wk}J!d=ic=2HckEPTUV|d0*JB-E3yLh4R1W=h -OFIIj#0}lEbg1H61SFq&Xw=Z_xAK8_)89-`u_~oCcrQf3x2I<4eME$!_?^g>lLH71kNvw$ubqJCe8Ulw${q2(k=|ebuCV)CPV*9S-5=U7EYpC# -pegVgLhFP6jrtZ)c^^jlQ_KirgJAv#xUCxVBA(DF&H2-O&Wm{UClLFW=Kj+pGn3BX`-tYrKS@3%T6sx -Ub@~rUju-VL7|%;P)z!ZDR1ZC^t1J_lyAd -Ya%{N&d(^R3Y%~!E>LMMU3F_II3D(d3ZdAiTch7;Sbvfn{CYLSNa6j3+AWc|ItOyU>CeaX1!fzl81$t!1J*Wl#PxmubwKbU&~?7vAM6hj+gy(F`d-&s8sgMrX3{I*-#2T=7P~HMInN%`{Vr(zXVCj -qFJ<=PIqLWpFqSAE(#pHG*gw_N0?#P1Pq#x)6*j#oeZuw^bez@)EDZR{+Pu}&Hu#Uj-uda*1r`w6i~D -Ze1@&v%b0W@_XwQZ|-Ts`%PEB=YJ;%?RJ~gcf;zMMf*YZy957sBwI4+ -H>EG1&&(5qML|EjY!PbfHfH3*18Pm4(1uj4oTs*61dBUoa9QpDA~~@p51NOD5)>iywFyakiy|+V%W9eq<=~(bhjnQZLS -6UL{@TqZsDtiLq0=2D$i!y5q~DfZD#w2AWPM|_8`{5VJbJ2PRk`=uRYqzZ5!$Lefz_08^Qlde-)IkBJ -unN6vK(>tQ#+QRp-CyX|UGkfgk8LbB69caPHG{( -eMOGRuK6l`i|pm4$Bv5R}BBXB%W6Fk-oB(`_++5KTLaoDzQBHue)RsV+-fZL+Ju<7(PLG7K%9>dM*%I -56VQ@Z)==?=J`qH!+%je3uGs;1@Ps15AR8y8kR(#gV<@|QNFT5(5Yv8z(pY{Kr9Bu>KSTenqx9dU=Sz_Oyw(R_wvWfO=X< -Q|R@jQ|qi@vLLwm=xY2TiiG_7jpYAgH3MFroP`5pTHDSiGxpU39b&2%f&#pt)H2CQ|13=?_B_*D&GF_ISa@Fnus^@#%f*=Me!PWSr_mEiim)UX~_Z$tiZDFM -ZmPsZm6YLW@MIRR^%-*E$oVEg?%XzEzwL-F|n+$tgux6&&)h?cJ~O@`+MK_|9$`e-+OTV>}Q^tXP$Xx -&YW{*X6FR#f$wj}B?fc`emp!epy_*iKsQJq=d-L={S91pLtLS*MdP!Z4uNRp -Qc5Ukb`qz3E#0?u`?lnY}_M!^s3@X1c9UZPF^L?al!#nt9^{U*PeRNqHGCxaF`#cEvw3OWI&_e|^4dA -oD;OMt}LC>LUwxffj_S1|v+Rvox9z!8cJ4xd2a|;{3&hBmL<(Hp*o$=*k)^!?V?-lC-x@OM|m<6~9KF -@Wt_ObzXpOz4CHZ`}E!Co_lWP1omSo -QtigEV8ZG9A7kNRQ%;Nxhj?>laXoZP%gRj@G--*uKiex)J#p+;-gQsMwU@l7HY~d%-c#! -0*Wlr8p=*f)q*(S_xd17N?;%1X23~!Tz7z)In#Y0E-vo^SRVZ&X=xD`C)f2maMa$TnG=cTEo?vHtu=5VEvmMwOK)PF~kA2)oqFVA<;_+MqqeOAZe(#+aoZ$^ZPccLn151pb2H5!u3g*cI&JpO3Ai7m??b0+*6G@Bb; -hO8!sg~t%IlJqkfNlU6R!oR$(y*xT^ZT0TIWT7UxmNT5zmLI@S*b}$?X2Hl->U(-E_ake#3$La60$Ha -NklyZh-E#cB>t4u3DZjX$f>_lwd!0Qmbf=zZ7I?(|*q -uD=ta`#kQadm7$s&6WvY|3}L$?4x$nfe-kz!L9bl_jxV6w?w_vJ>DD6r%0*i@Bq@)GEFM%mLc_m?_De -l=|22Y0DSL^;RL>S!ti$Ry#otZyL+?WskhYM&}&wvhd=vmX{O}Qe(#nkb%nHQ3)~Y8PlEJpJv`Xi(2? -%1q+Up`t8~5TDyc_2wQDo<4Z^O?QZL!A%~DUsbm8txNGkaJ&J=hi-GA6)A%xeX2ql8ASy)YdyjqWu)f)WCgWCHTxco7&fP_!-%bF7RF0*LnCE+ -3rs8onUwe_}-3%S9n%UxN1wsIBL1WI5Z!3{o1U$3Ut?X;!;k(x{lq#Z>*co>IQ<G-2I!&ds=UOe{{nDa)a8udMn>=FJDa(>wL(KKwV -O=CVU?a9UCCv1Kj&Ch>tI{2}*|i%&PPk>7E{Z-}^2>jh^Jj`}7^DaG&>)H(|b}V-IXU)Ulwq0~yiqFx -hwEF0!s3?g9Bbgy^0$^fw~Xx`A*H+vgKLen8*l)d@aV4(PI}4UI3Wvgv+3SG$vaP0*KI2_QGZ=zVJnS -$Eeb?A=47=pGdPX;^^n2K5`|eJH2<{D6*8^t-3-2HnqVHk7$xfbK?X?*k^AD_H%i$@JZHt+Z(OZCF3y -w7)ug?#5?$Xa{r*B7pwR1bk250q}>L`!+frL&qZNJ|~wT{Aw@aKcOAj_c7ET9RpntJko>yHk=HP@}S{ -r7Jjv-2oJz;!om~e@Lm}1%fk2c5aB&B+@FPalEZtjaJa9(_9LB7xDk1@l)hJ_)e7ES#g?iU8b`Bw -<%X{i^F_Y;d3b+nSkYwsMdCyQ6<<*aWi*XQFLuU8BDrg43q&hY|T&=<}1xeW}0zGwv1Q44LP1pNTdzP -wD}%M$b%59^6xzD{AiFwEa6EC9naPGN+FQ5(GBevj_6y8!xMx;}pa^uKhEg$2<6(mi@u8j1ZLYB-&*3 -c}}>sdOD1$O1tY=|M&W!ZRj)wwB}Dd-=gL*S^cZ3n}mH&wMZ_xI6!CF34Ei-e^y60zTDU%K$!`u4~Kg -?q}@H<20&tAH;SbA27L|$2aORJ>?KL3i+E5cjip;0quk6GksvYfDswJd7A`1H$uF~4LKdnSI_gMaj9+ -=U(e%%t{cM!k{kE(bs^R2{cZT(JrR6wK|aRH-%rSNdePrgXhu`NdR@zX3ZB<7%pb!RJB9gTnB6H%jbS -!9tPt7(53}+xUKbX+*9Kk7#`d@HB5WTDIN#VlzbwAT!rk872Q`!jdMtd4K8wF&^m0E7rmXp5-gn!DkJ8ir^#fzw6HTGEuAfe#{Ngc5}(CcTli&$Etx#bw}0oBYCab4)>B?(0>~xug|?89eqX? -&j;SibWcx{$hJXF`b0jm{&Bi}bPbMoMd0eFfx7(YcDnpP$$K^39~+(pMtW<=x+v&dCg=wKK-YM%I-vX -B0YA^KwR{~CDIx2ozC_l&%=bpGY3;o_3C+z#1uC+R?oXD4^DFK6Z-Ce7RqB6H**E-Mb8KFb)|tAmZr` -PcGN7#y`M1`}_vNUGF8}qY)qKpJ%|}4FwiWVnfz5QCIGqcO9Hz^s;|>AA>P6yrpw8ph89mrLTK5ftb2 -zXvvuPpt5boFi6(mWkb-l@vijwAM#eVEG9X$Kh^!B9ptrhnab^w31C4N4amp-*n<$t$Xsu-_s+cUl`D -R`=|U**OQ8#OV*y)`jWT6GLO3w%$>Q -MxD4&F9Xt -eFLR4$`+TDfgTRJ`5LOr#w3>vGBeOv^vaJ;33FT@e73002jL=S>qIx^78EK;WEbti&>1R+KlK9we5{M3GjHqMaRuZou1uJU2eiUl=~wP(NE*qaHTX!smfkv)afMzD$2{W -kcXNq-LyNYAGRT4sX`pMRXF=RR$!8iJBaM`#nSs44QSKwxt8>1GAYWRib&xk3=g0GO -freb6>m@#>bMz88L*`J*`7pUlSyW^*NybY45g -$tG$S7HTdHP-7q$m+-a%?bcB&BK5#ClxRZp@P?rlN>53aXSD7d2b{hASbr!FR@t56vwq-+`F)b(SI9LbeX1lKNdWYR&%R#o#1!=^j~NzF7Bd{$6|LVuH?o)Nq-(lpUU$*G)S&H6h4#~d>JshG1L -A?aPcy{pO|Beh1M#Gm%T(gQ%aqivaP=r!{;_4I44N0z#~#pE2;bp3bx#@89lbtZc5An-5xrXrCa$ -S;Ghiy+*M#qrwa>!XQLV_Gjf|I|+BAI-A>@}&FaiF3s{ZlEHr5#nl~oYaP?&_`tUChO=sQ8GF^>XWWd -SVezp54b=gSL)mfH#+(qKF=3O_+I0m?B{y|i9SY8$T1NQ`Z?A=fqwr9bpJ7@IXD^*&f^Nl7=g5{q$|1 -h=EG#2=H}+g(g@u>l^$dy<$*WPRMK}Zyb1T$13^YH2*NITz`dlz$AIaWJq>4Z?}R?I1Ai75VXwOre|2 -vjmtKqIxK81;U5ap{n&v1Zq4Jf6O@H+=fQNA>~SpNuSP-KTYQe5|^4d1-t0zJk3}&(fnj9{nD08=a*-ye`k}YCTSpR1nwpc -KM|G>j{kcwd+#4(5oh-SnG5jP-~B9o -BjpAN(Lh^2^|5Z55)A*Le6A%-IMLUczw7|iiB;#EZNAsqW6jzLUB%tBm(xC(It;#S1ni0>mFL%fXWJr -w9U2q6P6W#7&50h#w=?Af8729?>rZ{eU%M*Vv!vBqy#O>B_|sI)aBcB38`g*kqR&7TzVbn66|8kJqrA%bVg_ -)~`H#r>a=IGDP2S*K}puxIQ)V3gn*1D;4rufc9@yCW}B=r2Yt(riL5YQIFmw<_Y7D&#v9yiCE%x4ZiLph8}$&|jsHS1b52Mf_ -R?Kc(Px$Z_J7)GOo-3VvC^8x{FqRmhtZ^6LsY+2Ja8-{C4(D|lN4_f@3tq~QJv-b2AP3f@=20~LJm4p -)DLDELT4{7?lSr{KCBiuSJHNjqHIXT}cK_L7IZ6Y?U3yac%)@^a)okk=ybiM&Z6_uuI%58df1Pegtf@ -&e@DME=M%$SaWtAa6k4AGz-?*YqLC+Y5i~a`pE_gmuvr1w9D1sCCKkUUZ$|8dY5bY -8}-TwqO`><-=uewpm7x1hg2eHj6RECc -^C|eU>rZWJqOPNNGVdCPS{UTa#+YGT01cJbc$`p}!eNNN@N)9s+v9XWo50UXIyd(SyD8d$5+bpFz+lg -M;Bic>CS3BAW61-vmqx<1e2t!91v)%HoeW5qDi;wEvZbFMBQRacf?yv>l?fR;dWX -<14KBhF77T!pp21}YDm-Dvuv7_9Fq`eQnGTab_=sDiM9p(ZPKHD=j28k_xS*q{96>ua&CTWT4+3@#by -Ujq#CW}EPaY0+hDS3vh`+BPD6&~Z%k3fYtbYp`E5-~O35+jx8^jOk3H3~KBau#eJ(MQvh^l?x*_%NoT`l1YCSvGkHT{<` -fm0AZDP~*n{U!*!>M638Er=BIhGhKnly+e$|>Y}n<39=bGFYZmUBD4ZFaY`*W1T#$&Y`d-{pL_WmpXQ -RHxHk?$0!R`eYm}Xa2fcR@*EeTWhC5#em51 -{+*sYcHf~%auou8dNoY1`$RJVf8>75TZ1*#Qso$XVW&6x|EDSA4U(K&C=7IYBexYrhTi1f~ODa+t&k8 -^rwdz|f3mP1*etsQNgP~8e|sV<$-lQ_!m>GUD3aH2@xnvKfxE93oL`2QY%e*_W}v6C$L8_2oIo|QGwA -*b^_K2GuBxvj$8iEmKwCxJWH#~KA+tKdZnzD~iPQt+o0e2>bN?^ST|oFdLM@m!WotR#ix5DPKD-w=|@ -d_YJet-X#$NOYE@JM<~G2ZuZ{!yiUHk6 -?Oer|gV*5&27K$5zbeH4J|T?fMes$5B4loj)V_dpg=EJxccR`m96Lyvs2Ou?*4oJq$-IL99d6ypM9k6 -2wNt*bguqu@O=IA*M$RLCizcgU5{UL`Tn`V3G^pO@QJ|3WoDYPi*kb29j(Rl0}BW|NQg!>C+%28NOMV -+%+xvhv~DN@*U1Tj^pprXTi4=h?NDBC9Et)2r;?n8}?tKF9*tLh5xgdw_Gg_`N#UrfHVgK$gQ9!gN5s -%9$fu5{2%J0IxH+S-6eg&pF@ -WUm>uM_6^l5jAkA&Trk3^5v?-v2)`Sf_BK$(F$_CHanU6(VvbIuIdKjQYwp;YOT9^u -l8i0EQZO(EJK*tXbp>rk7CzBDx;5OdW#{RjUG+PW7km{J~_)8mIcqAqC8IFlPv~=GoNCx&DLkx4a#`& -&f%UI9wk4PM_Aw+Jln>@)1WcO1bNfL4dX>eOxK%btDM5=z~>1Z1 -Hm&pWpPtiTkdQ(RA(4zpmf7*-Rvro4ZFF -Pqmk~Puq5PwD;}c*RfOQE?xca?AE@d4-yM3-n6cx=YbWU -R$tkIZwDb()!py8}Q;vC&#cH!J&dtkTGH?EU_b*6lUH-t*We+~|?=JtpJOBT*|HHx~CQh0>B{FL2wCL -$EGh%1P#V5>~JtuMQz1;t;%l}9C|9?aeC^`QA>EY4?N{&C(Km2p_fX?OqC*@WCsaq(sAJZQq^G@(5LM -iievi~0!8O!a;|7m$0j9Q_Yn^wRwLFTHXvR_Bvm+1 -kLZio3DF<12cibCFJd6#V8jr_k%*y)I>boC1jP9Y{i%pqh&IFm#3ICE#8SjPh~K)AeJLmAXXw)Bi16;AvPd3A~qqCFVS8^UqpYzKtvs4?3b?klaQ -Mc3lNJCOA*Tvs}buFFC#XI{10P2BgP`e+f8J8j)^3|^T!-RDoMz&lgSn%iPzhxELERRsGK1^AfA;Hh- ->2CFg(L<5fswP0?=c(Ae##3-(*8FnXb3MbBF~npG<}Pr`oefm_41u8_Z-zijBnNEG82T{E-UM*!7ltn -$E($z%iZ9oDupgV=>a_M$o|?o0ngvyWH+_Lk)GRQXX6$`wu_Iu@Nr3O7tGEaIg -Hz@j3w-M4G$x{vBV;ByQh<(;L}w2Ki|ooRQwwZSNiGrr;$VQ@ekS){EdNJvf+O#k((9dNykHGvGH(vE -Cl}qL#bntFJU8f5nw+*GTErxG&YXzkSDNlbUR3uhv;`3zvcadn3rrqF6LdxY0kx5ks7&}@AOA5<{vf4 -#avz>axu3Pf?Uk!gd!K~cy-9d+*KrUG1eZ7ydCmH)7&;i*;)S$U7o0L@uw -ZL@w5m6(jG0yac&F@>1kt-CG%Qd7U3}v5u|+xmY(>iCnC^tVZ4oc`b6Wj{)yU;_n#jf41b^h>#!7=+tV;|;F4jGUARml86nQXm9rB^bBax3r9*g{L0_0)H3z0`4FG4;Mc`@?I$V-qXyv@;c;mkT)Q|7kMLc(X%!oe+YS12mbp3Npt{p*k2&`Mec@NgWMf?5w?<>lWyq_pP^8TXy$On -k>pTqt^lplGJC_nN+qWs7Qi}E86MxKa#DDn*C!;$A9ABnsW`6%SY$VVeDMSeH(a^#`NE0N!WycYQw

    >UMK<$~+NC-!B5CI+T$ff0y&%-k}U$+zv=E{S-8<2ha_yC}kR -rV`6fOMTz>q^UEJb_r;YM{U0`td?PXC*lH)-uYezwc^CPIfqyZ^*Xzup{NA#Vr3 -9zA`emlc}``?#h-~RELk-TjB)}ooTV!7FQG+LTzuY_j#Gp`6#N&hAkN!viN&^LS-ETWHr;)79NM5g)J -90tI=0ihC3TkB~6Mck#B&3sc8sNJN+D5_!mwa**FcrO!1%4fs2xeeIZHLqVLRvCQUn|kNV4}M7*Zkn8 -Yxc&B(SSr1YN~VfhBq4P+FHXHp)z!pul^du&VN}x&; -e?gY`UVDM*m9$a824NRXe2Wv+H=-;orz+$nnSqXorO|KEEESu@ozyjgct7z67?n54~{xedyp~i{i2OcRxm-XS1)_91yQ`|BjMXo|(u7ZP}OXli$3p_tZH06C67j! -#4bqal$`$(mP7f>JH;{ov)65%iM*>^%E+2j&}ig++mOMcbu)Kk7u8Ot3H*!G0-{748hdAi~MnN{3MCYzR-DT&{&)nnBb7`jC_ZI -$?E{kj0`%Nl1~>3Zf(v#zb*&2cCz;UQAB$(Wn$!Deg>(^6#ey_gf<&o_j0@po0NP^73o6mu4FG@G5nk>A#+Sxd^92Zbj}T} -r$yoP&1Y-n|t1rU#D`%`v~v>NxHdBFA2|MdztkWa5~PF?oM)ZE?8%5yu0ca -I7WB5sU)Iqgf~f4Oohudd$Hj-vQfxR`UI-(T@u6H2k;VtkI8SWe-I8Zzj;gXPkw=WWuTVD><~q_iFvkUV3<{Aa(GUcOctK>cLyu>qaS_FRl3ln=@FonYb%j& -G=2ek;R#yh^jj#8WbfL&wzDll)KZwjA7Xh|cqbX_h*yisAju?4=Iy>n^hYJ;gAp7rV8-+uLchmvsaEwaanl>najAnYjkeO%54ZgwSp2+R?bXvnIKcFZpq -U%%ZGK?7Fk*>O)CnQp+uN%LxB9@fH}nX0x032vu1ER6wyvx#O*9_!?_3rmrI+I++TNfoBK+0Z3qS8Xd --ls-J^Fc5zJV7T0=6!*J??`&!jKTi9WKNTsw1YNFHbZ#W^%z!GG#DjJS(Co265s_$rOmuuJ9O@Tq2Q0 -mAIuI4GT0AFCF{II92Hce-hf&WMZgMc`b}-O4ANk_l!C>xI!*y+@x`M!%8msfa2sZ9yiy=>sfd-^^5+ -1t6z8;H7xfK*Uv}ok$LypzgNUg1Zr%1rv -xb%c{N+Q8S3+;4}4OD=yJ|@&s!UI)GY(*(H>^;rbJB@;|rn^f5I3yRJ&tU@RcCq)&|HXAsK}q3+G=WC -$P+Q4d8ALdX8^BS&puJ5=A83fw?iD9f$5XK_yX_HpFzD~j@0~dEa9)f -68^e_Rn=`$-N&ShuSEJHh*aIhqg&(=Hd%cXefTQ*l|WwtftVVPpiqh#T6}Pb@&A4JUyL#TY8ZWZZjS_ -i0Z>Z=1QY-O00;of09jkI-i@1j3jhGhCjbB(0001RX>c!Jc4cm4Z*nhkWpQ<7b98erVRdw9E^vA68f$ -OcIP$xH1yA`9duuC?wrOG1SY+GmHfWm`Nfy^bQ3$j}+iYY}C8;>sAotsEh7XC7?Kr)~VGoCMfJ7GO#d -&j>D2h6geJ`2PDyF5Lcm5fEI-ScoS7ag5GGk;Sa!q-zh@LZoz%18%LNymTnbSPYm{epcBsupKI}y?4%agOqsit6$PRgbvfnD-Rf@VZrbW4TmTChGysnk-)v^VS!_<#w|nJy*RoswS=i5 -vqGu^Nhvg%NItM5kb*T$2(!M3AD?fH3B>fyj2jcNPn97Xq%LMM1=ra5Oj74p-4zWxm5N@VGIX(`znDN -#ILbF2-aG9%qA2zK?)`ca8@cAI97HGG9pv1m#;+IwBoz}AKuexwARq{|5ZAB)p!GD~haZE -=rz98PeF|br(|p1LW7NLr%z+;t2{Ar&tJlU_XCfb11mdYU{<$ET$Y)HVB&saXC9NzzUP;uXWo0CvG6q -s$_+K*PC1H#X7aR+PKf%y^8Y)V=$gyzimXTBNA+7XnHe-)fM_kpLtlm@_KLv;;q6h)u5rPunyKDa -ILFs5pwzovE($)ay4OAzs)_$lK8y27Y3r=kHc$HDIrL@>C#V!bs?KHdx08$e**0r}DpR9erEW%If_&8 -+kI$Mio^TdxHTv$x~1j#23xrYh}B&Q3WIdg9Y~YkFm|J-6#XW@`I?ij$Z8tO?rW>{&536%)Wq&soG}N -h&)-TQMC(IRjxaZhEe=Y@VvV&arb&Lq844N9>+|YEbe(i8JJ{%V3Ij7XH~ZU#kiAMGYOf3cOnv2F)%P ->abqmV%S?il1PfhN1x#s4CV-P9UB$9#pMXt)FDL -HX%gawjR|wzr!q>IxSZg|gOl|;!dKL64;KAtBGZ4OtF6Z!ZYa=}8y4uLb5{VxKiPv~i6JU*NgzVV&ik -Scr8=7<0%d%!l`O^5dMv@VkF|Zf}jUm}5kzb4YV2Qk)JlDv&O2V(LC-F51;9z$-+qOGQvY&RP@`3zArnQGH&C&6{rE?q)MU)uR|rW^chhcA#o -0EUHcpWzaay_#aTQzhN2_P-z0`Dpc2&z@f?n=BjuQrARCa*sb43j9{>tOIfmQpx_re$&E}!Ip;-`< -o%w{>FL`Pc&O#zp{R#Um^n#)ydN9bt8fS2oS!*Nopf_3^}_S0Z3GbIL)Bt;(GpoSQf1U43I@ZA$qrAT -BraGkNt26pT2Bp-??xt5coc>Hk~yI1&%(vd0wlZW$Gc|>N$p18m!t2VlJbLaZ2Ght5gZ;!Ig~j7 -+VeVuAhiuO9$7PX&9YGTke`zd`{&iygNu`ki)h$sNQ>^bNS_No_w=1sfF3_mxp;hU<(1d4- -~fdTx-<##uRiGexrg<^7J72uQCI=gv61MROW;NtD^x=pG%{p&Uf%nu&DOE2lRT@DVPA8ptvto)^$1^z -#>T`%s|9X~%f+^}O<`Ac^U{C{M}Ufde!9nRToN+7CF(qWC#okn-+et4JOM?Fj&x%f_pCoi5JG;&zYVQ -mg`A~#dt4*C4~vz;r{2O)CHbZv=T!-m>bu1L%b5nCERlW0o6M2lef*?8;T8CgVF`aN@B66j5gL20c6_VCM(VpaU7*rOi+DQVWB0*!O?ow58;Lk((IYA<6q-6ky#gku2qsO|&0^^vO>@;r+`@HIHQHOfgNGCq0wIn8d3ePe#Sz@eQ -^nW3DPD+cRj=?86wie_ZR)pFpb%+;c}{{dshUoK -9;#&>&zsdv*g|St`u4x&N)9FBl_ptqV2W+na%3m1Ddo-7C~|&vV@4aD-xN@9B#dhwVi9Y=57;8iA5=w -o$4CgO|J_&-eHL4ONE-^_(i;xM4`(cDdS>?T~Y}E4Cqk8CP57~GR99+QTBsXf+s>r&JDJzytoQWl#gpG@ -B)%q)P_=9$O(l!)cse|auT(5(QyV40oK;CIaXKDnCr@x(jJU@ATd~wp-#?Q@%R<6;zz$sScxEldsyV< -_Y0HcFm>pZYSUW5PRxgY^1)4B!(j3OK~gK$`?FsB^aT2DffdIywqxli -+x5H*kF0Z!-bCd{YI5WQ8%l_ne6+Zd)+x|NEeesIh<9uSu9kx?JTG<1Z?i$7xb%DbcJwFmyMzTQ^)=I -nWtOe%NNdz;UJl)f6?6Mj^FnVvIT?Y>;zXDDj-*IQKTDMS;+A1;86{WMt^9%QuO9u)Eouvzqy8%XA-- -bTt7Ehtg(19F&2kDC>vK;miwC{}R$Uy@m<{o{9g2U8zGU}#N&x02*8+}1mc9(JE46EC -R$c++M&Sr+V;-Aq`a$+3nyIxaOc7mV%ne`;^GCm8jR=bLvO&~D;wxQXS|mSkTC#I&6Zuv?W3|68 -RxbDNq#OU_?TgZX0Z>Z=1QY-O00;of09jjynjak-0RRBG0RR9U0001RX>c!Jc -4cm4Z*nhkWpQ<7b98erV`Xx5b1rasRgf`G#4r#=`#HrBnnhY=d!#K86%rNN-txxdWJEGHV|yVD;t-sW -li*}S!BqB0|M@@P-W*EO!-BS;20w($o5QRbb(3&g1*icz1_6Y{r$8C|a#r;| -yCvw+L^N>qo9Ckz;EvBra(Cr^HhXny**@_GpaQiHyPH5gTJB%VJD(4_GHEOWti2?%)?tO9KQH000080 -LuVbTO!V%e>?*K03HVb03HAU0B~t=FJE?LZe(wAFLGsZb!BsOb1!3WZE#_9E^v8uRZ(x-HV}U2uQ<35 -wE<69ESIP$+79bdto -s%TI`xg9>o8P1Fd^<4M^VbO#4%8jTI8qqCkM?^c!4iaJmsFsZ1J&WHy7?&x3*+yL(xja>l`1Rjr4nE< -WzP|5o1S^6f3C0R72G|&OT;Dz8QUm1KtACxtPSn(aikVNQ$5bxlp{P8M!tyQ&~`C4y};FK%M6Pm{QVt -|bID9eGlEJ22mV00$-dnBw0&IeT=mdSu^VWJCAp)z}j0h>aQ?`{R~r$RY$et8xR<^Vsb5UUnkz=hDAQU7-cEuqF24 -w_sq1di(G25`riR|kQM(DXTcUR9&__#n3!m}*iN{Y;ec-~_$XaK?YJ3e$FjS1il$6c7t0k)r6!r6phS -ELq7v0{=Ch2mvpZ`8{j2Qdc4QlHd%8iy+08hJaKV7Hus9>Vlm0fXQ9Z9wbeRTJ -&eRNf;X5#R`rPcD7BWA#w~o3&ha&P5h-~FJMTUax$DDZzbbr{K8Xu -_76{Qr-2=R;a=7qJoM1;7SVI%xhkCWsWq8(-K+ImvK`GVI2iM=*F0EW?k5vX>ds`oT4g4 -qKYFtIS7x%J8KRp-w!@FZGT%Lt^Y+3u+kR)4Ximyob7alsSkGFfI+t37(aGqptU4z{c2@N0f}x6o=UX(M=zd-TiTM*Ig*O9KQH00 -0080LuVbTh`v<*u)b60G>zy03QGV0B~t=FJE?LZe(wAFLGsZb!BsOb1!3WZ)<5~b1ras-8^fL+sKjM= -T~%S7$a?DZj$?QnLxgpFCnsIgm93o&n61%&)-mh+M -k|Y;O{-c+=(`nu`_odVq|2z0xTzu0U+N>*Y%c|^7B5w*Qx<=eek=<7GwbxQ?@4BvKAn!Akena2>J5yV -6EW115_POv|_FY+NQ6Abxbz1CHa}ak@$(7IqKF^fg^_5xAv(k$MkiIa#U`8eDB2(^n)1K`2t~|(#3j& -&M%DR)P&ZO#QB?slhSR^T@BLoVA+yrk!URH{+Y#G*eOP8&B}WI9-TagkS<)__)i|Hd%VEC#;(pt4qrOl%#8d -Bzr{bSMvHj-!Aip#2*+2qHlM*LTtgx2(@eQ+{Y=-XTgQFcSH?%W_)Os6r~-DQ}d5In{~a9r9tVWg -l62$LjHn;BqK>-vXUtctJbxI8bs#4ZGpxdU_CF=ODQ2?~6U=7Lu!Qwu(Jc7bzK(!VRC2)2xyA3?t@Wa -J&OE?jiS|xEYE<6ZOBGt62%sPoX|Jd{d6dvA!lAXjc10)EXp@kuCkzYGtRsoEunqz7uZ%BnPu>oB#yU -k_+O1oQGf|o`|@Pp6OjS_1+HW&%;hdCcV#lzjHmGIbV{DPsS@L-wh9B7js7sMYI(5wIj;6$%#4=`SF* -XV9$#q_qd?5ZY@ff$KzfW&j~2#niv8G``V1O7r)!?n-E_naKslNPfkf3I+JT8Q_MT1y6XiSON_b>cQT35%K -9=^vh$x3cDH&BfaCNya-~$-I+m^9rWxl7@RuIj_y>pKY;s*C)~JSfr6HL)7KADmAew7(u{szt3EKUyG -*lXvva7kgB&lyr|8ZH!UZQf7z#Yaz=DUHklV`os$2i1sVBCa(s;9Ov{Nv~FqZ*avZ9E0zn7<#c*v??P -jHtmupv*t=0qj;B_s^sjwzWze^2r403oR8pZLs-ghvpo|A;>1!p6Qjt+w!j{RDLd?w$g#di$TO1_)}_?g~*IDHoG=Y)?}>w<6YMMUVNSD67t>(^coyt1L}s}k=-i7by`K)EDk~-0xr5RO -zF1&BQp+5hRsD7?f(ywv{RX2ft4jEN(=DbS=XtBg|HI -I#^(AP`!HE@u48_t;~67i)6Zwb#_??M{1TjXsf`VRX}4f`p^^?zH{l!@0U8&SZ3~;5^KLEVo4__s6a! -tSXb`A2k<<2=0bwx4kpqCe;%&6s#$d5HLwS18d$T{veV~T8ko3>0>8KuwG}P-Q{e7*PqKCo|inyTK5m -Dq5CXeKHg2u{=h_wYQ)H`74m@`&m4%rH^K+jU+^a2m@Ms^<5L -dYMKnqEQT^s2}f4Qfca1!T=P>QQ^%242_FvO8tJZCN$$J1k!XUY!9*2i -ykD$>*4-0=j0?G3rMXRlg#Uesv6~bDDq}wDDIcDdYNj4|;usssqr5f3cSNg=k$nl$|-c@C8x>zQt+b=2#*Mdt&k@IHBvD==ZBfU$6`GaqS&{WVYRp!GCwt6XacTUf@A` -zuF^A;2X_4BI`&RIG1J -ZxCn^n560FKtDY2?S{jRQvbHL(EodSLvD?`((8+9O3zF+)$Nm2z%h`nx@Bh -D7XkHiXw^#?%a1<{4DmW)QFoniq0T0*y)Gg$}DnJ(yGQ$0vT3WG2sWS;I?=k^oj;8=NY>Sj~)ka6W;sCp4bUBg7*O5$$&|HxC&Y99b*iFlFe -fyk&JUWqThxIZF4=cJ?mUrT3-)z6%32`t=D(3>R1dLi(JU^wNz#C$6@6Hn -}tX^o02M(tAVMec`KTZ3LVM)`Yr8nt91t#p%EH4PYyH#C%W5ZpkXh!^=`dlYDaxtj95s~+(KHm;>aJq -H^{KtvMmLKy_+ADuj4COxTnaq@}NheY7#M8#H+iFw)|rF)WQ0cO-qk?Y8Gz|N36Bz2xlmKpA9THRh6% -UF&K^D#-ftHVqOHggyS2I;uCCq+nI#j99AI3#5bgxN&s#jaQ%mj3URQ*t|{Jz -xa5Jyo*U&pHk3WXZKj}lcR_9ucw?EzR%YQ)WWuov8}@6`3BtZ>)2BT9+O;E7#5>WQ9%7a`cl!ErJj(y -Yod%K@9X5~y2z7-zwlxwk-K$P;_EeY!?@IOwP3r0gXQ!X7USQzAeit>j*53dK*OWZD5MQk#jgra#<1; -(>gGR(Pa(@}~>bhy#||R<#WPme&p*@qC8YdXm>ujI#x7Sm;KKDrEJT8p=v`GB_YH_y0215p(jfoxi2Hz -4T1lZl{|LOKcfV3w~;PfNInxvSqp%H$R}P$C`e!UxXenvQ>MC89JJkNOYH}PNND8+8IGeq`9e|xwuR& -ADImCZ({1O(-Eu~xWH6N2tK)7T>ky<$tB>Ki~^|jV_^?thAp0xeOrK4F&8NX8RT~~&eu+{*JX?&tn{* -))&;G6Jgl*{7D_no`OvE1La*1O0bAmbM$5*LWYoTC+Ea@EFcX&Nnnjj%%niebz;w_J#jWHg~x|&WvKEGNY9 -ObU3j&gb4lM}M>YKHKHjBI)>er%U>)Jvd2x?yLV`HNkGRIESZzrRqQoI!jvf9VmK#Yg-XekMb68btNa -HLkB4ysxmOY-wYnnmeF^NANv2H7Ian*VmLUUYWDET88cV9`A3VM>v`|Ub|DheX -TL(p)+RK7b?=MLMQ_%Q`f*5xM0BOx>F}y9__O;1b0MAyYM{%hPS`a;O9I>{OSAd|BP^yUctN$2xXd4Z6Da#NU5<7Klk8CLlIfXU*`n;5kRC -Ek43`$@+d;2 -srn|F3Av>LI85p>HZ*YU0u_Yzn?d9D-a9Gcbj>{Ps1r$5c){1o1FhQ8%J3-IH5AqT`S|PIv!?1acp!N -uZWJcT`?nk%PyfIC~1jg0ipYx#BvYT93Q6N!k;me=*_Np)s+WU(1ERK}X9Yj0*@tP_SGWBPP^G`W@cb -Coh$~t3WE?U%h9n{^TqZ|m|S+r@{*Y1+oiccC!I11&2Q%Q~Y*Vf{rS8V`{H%XE!@hjh9BaaV{h(p0#L_OgH@8CIJ -8G*-?8~dkxG2;rh59i{Y40!fs_)yEdR|uK`6&MyDWuYdL17|Tya@V8k(mv`3G8e>foFX)%y`qG*$PYrCF!tM2`l@kfeE{Sz^DfW -1yuXbbiL;gUIe1ssQ5xsN3XTuhAs^LW=)@!u*n+j9@KCCkcN;vz|IY7%1q@ -0s&QHu$TPg^5zrd(BPtO&Hd87R4gjsv?s`#bx05(yMdCqtxQIjJkM6*dGX@;?ThC>is#>6J%4re{IBB -Audb7eG0pbnf1-JRF!Vz$+7b^&+pgm6;*b#=T+m-4|FO2ftfRSE4X2=sS(E;a<}UqY_F(IQ%j4wQuC( -gE8WO$czkYzSbh9qSR~_8|CFd5PF@<0dOnDu9Lv)7g$Q9bf*ki{7(Bvj?jO-VOgczS5^MU<*$Oeif7V -SY`;n?JGx3TI5i&YL+)>>WXe}j?Q9Kmvmh2K|x0{3w6e&i8L%QZ$2Z`E(y!K%+^OiLMk?U$wx?hpIoH ->iRguf!kxO_4A8LJ~Lz$c@lxkdyG)@E8caj1_cNzVgoz48DaLzU$RB5e!FN1moq7W0y%kuqNOZ{-{@C -503m15!b@?)2Mj*doU&gT -%wU3fVTU5meOq?kDW2rE4ok&K5{muKg%c-E{6WW7Rd+FZMa-$Ydkj;Ko6oIY;%#SY<5Wr&Y1;IcwgM$ -?l^&ZzrerycGUZ+=+knG#fPt-D!gFnO>@nP;D$3!+C_ZG>IO;7)?5xV9<%otc|;uUkWqh66?SWfHsyo -b`EcM6dAE{d!+2&7Y6pC@&cFnFEhM9`|tFfI5Bip=u1@1{}%Z(NYTk@Yzvnz)iPZ3f};Ic@(xP)h>@6 -aWAK2ms3fSzEw8U^;;X006KM001HY003}la4%nWWo~3|axZdaadl;LbaO9dcw=R7bZKvHb1rasrB_RH -+cprs>sRc!hsYTXd+04QJtR&tNheM+w%Z*X@OqAfRxQHNX5##3Tjhl_fKfG^I7A|(ky -@5Q3Urk*LH^r2->9Q$^ -V5;9GF^Ka;KheMJ$AzR9DYPiu_oJ}R3dtc<5X3k_9S_P2!tKxJg1;UvcwLC%2e%SXsM%~qMTH@#HJQvkWY3`hHSq-f6D8|NKn~<2I~3c2=+n&Zi3>mXN -H-*osMF-w${)!TMR`Y(+>C|OkW)IOHBky`lRrvl;$Wnv09CV72+PDFu0e^;p_I>;oJYyYuYp=`iTrc$ -m9jyP@ww)f!((dd5l@cqD|le -Mk*qQO;tabiPLbT`{KjUY+TSUJVFQt1jPRC!5m5aQ8QRDJ$P`WKFZ7GyRCa2yF5 -p_`rY9&=b)b`tb^LKolA2cVrno$v_h+JrzExXlZ!)E>+N{!(Q#9n%~owm0x4ggIv -G42l?Q>NN5lF4>nVL66k6gX90Gqr5*EO(Ei}(|ks5zbS~MzOl6Y@cKL{$5t`n{G`@i9dFISRam%9r|^ -)555zCr?`6l86mo%S(BMyNlZMAm4gQVMQ?DC2LUB8UcvUp^5%oJu1Ax+Yt|jl53K9a7Hct?MEe0$bDI -i{U`y`3bbPhxklF@Ti^J4tUG#K(;nKBhjyEXPJc%E>_tvz~F6)%A_zV$IxciZ|NtH6Z3wVpPzUH!5?WuT8TY8yd?`pJ -6ZJXW~=vAKHrI>xa?;6SP%7_>^g(W7D}}1SWGgNZjfeb{S>n85#vVNKze5c!Jc4cm4Z*nhkWpQ<7b9 -8erXk~10E^v9ZS4(f>HW0q+S4`Q1ZNL@U<`y8AF49GUu5Bc{y%dE&TXf7sCRLJ3q6Yr&9g3txNw(pls -F2v!a30@0I71RbzR0{P7;L04lnXANU|l_dVx?3DcC09wn(A3f2uTvYZ>Y~&itxKEsM$)j5ARwR?ax{# -$yUidl)Ox{S0-|)z%TM#(36XmDhj^7^3rJ_XUL{BmmZXeBcBv4cXg -n%LNlkun~RxZy4RN^NzS&AA+=z8hN&|cfa%QXz+62rQz-!ya=tO}#?21Qj=`!nJ1L+q22|wWW0)q5hL -w`9JZG?$UQp0IYxwMo=LNYvt<#|_wEU!t9S9Ma~Gq}B*E(^k -el*iq1KS+{+Fi>*xe^jWI;C5&8L%D7RoW1Mxsw~_!pV>PtX^I`ucUW9xT#O)2&bKry6enYMYa&sVFandS9^x`h*^C(Q8Zz~Fov9}td-+mYPyv6 -c=)EBX)YpCjpVH3^po$ng#WEY!cEmS#KIBLOJvsW9Llh1yMt3Oh)C>-?QVMIbiRAAnU-9h^^9iip<)wvi93!cjUlGV -!XY&4PpfcLS^j>D>g8~;<`-NZ-xwBe`o0!!%jH!Pp6&LQLYX@CC{_y&C|F~SVQ?KEi+KXkHSNl>=-3-Np4$j8xh1L4ztE%wIz(tN -Fbo28J53~suaF*F=jTb_Ds>m{n1NV^w;y@+;08mQ<1QY-O00;of09jkDfkzl;0RRBq0ssIV0001RX>c -!Jc4cm4Z*nhkWpQ<7b98erXm4+8b1rasZIQuhgFq05@B0*ky_gVu03nA$4uwL6Hn$LTqGMcQcJ1z(5< ->dy9aoJ>+W}qu*ZJn(-yjHtq0g4Avr(n>!HVyP5@KCR3rAhFH9=$aNfu*yE>@q9vAx`TF6x(C0l|02#cHB(vZ@>2i5Fe!MTszb2 -+-WAU1OgAh!^8UqIGIfL8vWO;Zo4*J*SRj+z5nt*dwXY>MXTW9w$y4_S6x{w|b&)G*P!`LgQ_`}8mQB -jdyazW$cUe@>b)&N!hMC=}Wwvd#`mwIBwwsqtQ#b17Qdg?&vMSG-T$ROngTsbdLCZ}2>Ez7}Ah+A)rO -K%6<@#Lbxz5#_+wQWiQgy1ey6n15d-U+(1@yQ*PnY%j;ig?ZH0XO+v@Lvj_~e^kK26Pp&FtE=$g51Le -VFE7vm^ED>qlSht(tnRirUQNUDa+k_)X`pH@eBXqOSJ1(x%v?^J`t@b(3zJ5+?8@)gPC7gLPU1AlXy! -d7+_8P4k{n!H@NlV7>FMc@@@qOioK>OA+zv&S{}mSQYE$Y>-P*M -zZ0Y%WTXw}JYr6TWZq`6E**J)8SCppt1}8X&CxW$fUQ|FeRaVZinjKtM&nv*lwVt*|xhhjea;gd-L0@Etb!qP@V$`Y_iTr|MRRYv-46j>VALv`sZjte#$l*?3R9L=F56>JLkgrU -$SN<ac9a75P{*gKZq1Wn8)hnVk#mHp*vGgk{*aIM|R0i8{+tq7eG)rI!x7d8s)G+!M* -#nF+FB+ijbnorU7bkCDK0lkEo;^D|QO5?82oz-Uwa#ZKIG%vS=*w62wld|9o$?ywawFwWobpcNDuL9S -n~tT@FP%z&f^}BG-U=%{#reN@`RdubAJ67?yg#1&baDpcee!7U`Rkw0UjBSGfBwU}pZ_pFJ^9m1DF5n -_dI0}D{(A4^?B!ek**9N3-rL(-mRW1b-t7OMijPHgM8o?(%hY*};EvTM#V@wBSb;3>=wUw$r~2ff-)Rb5A7 -od8j~(DJbx=hRI(>7&Q-KfLfiwA!$#m=)wRAdZH0blGL64p4+qlCXcG+QWK( -PDnZe$&Tl*2PN)9z$vTiqW6D69uheDMW?k0^gbC`hlmB7c`J6cAJZ|5DEMU@i%R$FdS2J}@YZ&moaOs -DqecjF+HD-#7hC9QI3m=b;%6y=E7}<6EA$p4AyT-9yx5G_gFxR-=*yqo6rH5JYPKgG{g%0 -!$@KtvrrolFC)08RlRQ&=M~My_Ue!n?i$-n9$wdN;i3=+o&CO<83yVGPYwUbp(M$7jt}PaR@w^*Bl1g -Pov^Zxi}|redbj^Y8OIzom80l(hvOqAhSpds{57`!E>l>`Qpd7jl+=AYVb|R^V6&x_6R$8bNd9ec7 -=5=YTSlR4T~|x!?8!7-LVKAXXn>2KkWS4!L|G8m-~h!CZJu^4w!1AjlQqMXGInDNAb-W?&1?3Dj!Hwe(B{UvC1LzV+K^-5f$HO$Us({;} -vVR>tCg#%vc(Mz?jAt?gBovTw!x>N5Z#>*^RY67HI>zDQg2Z)sT03xv-8-5o|H+yryB$L3in!h}zzqi -K=$i;}P*e2*q*x`H#&BTFz-We<49cRS+UonDsUNB@AXZT5`XLj<;PDl!9>*(|F9x&6xI`lEj#rm~nzT -yS*_Gp%Ohi`BQCc3aI=_0%cbb*6rBJ)E(v~2=Ku2AlNe?YH=weWzQT+qDejJM$^|3}@z%By0fqo$}cx -Osq0|dYTl=xcN8?wMbqcW{PtYWP+w1Wr&bMZ=%6-{oZJEZpNz;SyEjCP}!1xy)OCVKY`WLPCN`9wcHP -Y$MIMi}-wvu@acH2ZPACqN&}3em)e?J5R1Dr;kq`eR2Ri6>qnI`DBgOlOQT7hD^lzb{#x7us2tO{1@i -dfPgS1vsU}vTfcH;j{`>dmIs}+cY};5gP!?Uu7sWSRjvBy1!IKuBCi)bMB_E!3~&%l8WjV2MHT4$XwFMs#jaY0qrSH(Zii2tmR33!vD`VtI -*-two{*Fdg{}n`;bT5&R=E3KSbjN9;)CAI4whRs-AY$oR4VWbQo_hIK@yL{V21j9Z!F!6q!kg9e`r8$ -2@R0HEWusJQ^evLyvlxXyft$fb^YyA4BlU6%n8Wb(#bQzt7idQ8WSUFN`^S}fitpjeZ>3?}G@#R4&Et -cIB)ThK-hZ@~KNX%iT=+LYPSMWNg`^tE!9IFrE6NfSI;rTxo -6X=3JQW(Vo4p-$Runw4Ef?H)qM)H*V_8Lhb$fuh|J*VQ|XAO?n1d9Tck_{~eyGyzx90EE-d*VD>Uhz| -!oiZ^z*csV{wtE!59~~K%^!@kWn=ph4$2w1-2qdfxa~sZa@Vl312QzhG;=1_vyr|N;xp;V?yMw6@A{B -mMRhe-2tM>r#!}o^!^gJCTM^$w5d2eJ8SGT`-^BfK_QYtP$Y{Oc8 -$vsl;K`-iX6EF)*%O;lwNPTTUz6m){^`#Je1TME!pC(iv%-kvPZ7 -an{*~93#O`91MN_@V&E@4G)N;PZ*4)&}sWc3Ij`{Cac;thKn(H53-^WEv@KM*WrdnUI4)%n2qI79%% -@7ai-+E1CY^}MfaZ=hU2ciUxS%Xzxe)BVncTM{b$5IECu+_gjYFa;6Ep}W$D0wX84y&4Z*C4m-Sn)e> -x3mVq>+<5bHxB%UNu5^|~l}ciA(&-gXoK{sJ0deavKlPy!Z8dsHX%#|&61ne-fhMil^^pQGyMiC7@)k --pFsrW%39OCkO2ZY_C+6iSfV%X(Yp#@ou4O;-@GbyjUL>iwO{YaD7@lT#b?mVCR^YFp*HEgHlFLDH)g -CXoHJggXO>1Ux;vL)bf2TX!6%ns&1Xt4&*S23Q5+t)wET_m3Yv+Nn7N{AORw23XZ|qveN-ZR+~m-t)8 ->FphcQ>A+nbfMK~1!tixCf+*Y~Xx_^Y9|Xld7FL;DpbKqVjcGd|S8B95&=*JC2i -f&uSzNIh6m&Spvbk+36N-f3{}FYToM1F#&UyvLS~UG8Ar}>vUli9AW;C{+?;&GpCM3xyO8C!oxDF<{p -h%Xvj?@w?CMHrhTuac9$lsA#xavvo76sBoq>NNXWG5tlm7guM57a_?7IT4}$~O21@`WUmu80!x8KRVv -R8@=Va*q1g=jFa8n-v<~)Xf#fn=+B8(n~%N7)I6t`~IhQr)TP-4j2RWoRTFh{nriVa-mGi@?89!BWR~ -BNdHg2M8%Ax$hfMBxt1ZIyA(jn<{hZcV`krmlRFQ3#_aOxl!ic|C69eO;XyWUJbt%vIck_Ons_YN9lE -ei%SPSN3IFSvq1#!=s|GmOfL>>?9PCJTx5~a8X92v`9nYPorth?`3^L{UXsFc(5~eIC&f|2u+2|&j?o -R2#>`Tl+mt>3_NSJKOwsYngrdb=k3mfim{~{C^B-S!D+CQp!9lx3v%T2^$^xoZkC_SupK>K%5t%&xU$87SQQ8r5!MWRC6wZU3kb8dQ6U;{gjh -FN-QKZ#jT2iKq_&4sf)JbcNP!owe*oBdw>XmgP2QrKpd%o5WLe<((fKc*+56iw&J)P*2j2-mvaE(y#& -k`tb4NCr@4Sb%Gd#)}b)Hbs)IQ?yY(h%M<3*HAyH>-cl?-QV&#At0YaTo%NeJ&f{M{(7i%<9kNblgy!4hvm@~xZS8zjwsc6I^`rQ8H -fljmLqV>as -3+r;BWMBXvy@evY-HshOkd*Y-LCje~a8?A0K@f(1#g6z0Ipv9rWwxd9-gbs!Im{Va6Slmg^PM@Eg&;g -cgNg6S0W5Dz3o0fue94&D3V+qaLdlDtXeZ%EWGibmJA}B`P -%E-^K+0*=FkRA%S8jMDV5N0;*?lLwQTnn1n5EWP{o;Vi~;5X!9G0}&Y4Dj~bl9dX-I -*n({Y51>rk!Wf+jQWfd3KRjvji(@tU%_IZ_rKDCkFp!OnL0CAv6L%iJMJF4xP;$5-b=fW2Mr3l -o(*(W{78mKRymRYJkMu}mT=wf*E$6m6Q=NYrAT&oY^~1J_QW^)&FyP0WdRR?9(bv!a{;3D@*=L2rX+H -)+-z+@h;hY0X`jc(Rdn2EC>5Nn&(?66=AEtrz+Y8c2*)*0oE1bq2zn8LU?5)^g&>Ln0r?3)l5lk+Chb -hWHVV4=Ep-4XxLyD0Kx)yC%tHPk}Xtt@qp+Iz;#W{HJOWYE1mLaDe6DS)^iYmDncpp(Nf|!25v^Rtju -LpTEcrhZM2041}w>D&#(s>qdb$~x4Lvw@?fY6*!8yjT)Z&O3v2O1}YFG8FLlKUnmEo)#HPfz@_ -B{{3KLBuLqctC+_UI)8rcHLu@JvJO%m6aq=_g^^N5}%T#C~f;3=f(c3wndqja)mcBasK2|fv4iUwXXvLOgupC-Sj20&ZGqFKv1J;Mjp;ldoF|Vbat#0QN2)zRN$lKRkT8%i?R!X*^TD2BE(EOXp^ -2>%Ig>^i=O4&Dwny5p!6T0i$VQ -tEz&G>D;kt{)_Q-W*7wAdwt7 -3yHxSGVY&b;sTubDeaM3?~Od#94K^%fc+eOuaP6pV=Xqo0qe7bi6QM(`J2+H>_{0o>X9a9DRBjyEy4; -8_0fPvCDVY=<&_EWI>tpv4;CPlbPgK{?^GzdRK3RA3vPe~lDlRlf${+mz?36KaioMv#3JhN;#}nMNMb -S-Bno*Xli$5OLnp`&FQ2^#PKP0$y}%l%1W7oIsYw~^6{C>9odTMI6$V5+A}fzaBVmbUR&i(|iv`v`>E -Ahd!evtcY3NiXMGF7pbYLdPL2xG1c;374IC&H&%WD<9s?Xc -^R~~q>REZX1O|Xypjh@ujjGGAy^D^L2h>5tbxD=XBp#1vqX5S(x?8IY9mv~8JaWl^<+HEeZSYPoA}rR -lR;3n4AAuhP$+D7e8=BA4ErCcHTM~7GYV>*o`hKu`?k##OpPmkJXwI!*0ymNEd(k#m4sXjK_144zSt; -L&XPk5O&|F?fy3}G}n7pBI3x_*0vG{Gvj(WL2DpfDG6p6GYV98@r-US%Lxz!-(cssGnL+1kyY%Z2Uwhu1wWVC9?SrKOSg!)llkIo$y7b7Zb3$!xA>sqv&y}P0%pEg1WO3Gq --hf81l|Iy0}wCCiFipED1E$!6Gnb7HI0uA&pObRJz|+*cyILZytB7WrmC|*HXUv#pf`^boz3%DK^t!2 -UE0-W201Pz=LRp~eqNcM%P~{Dzn%H{5O!xeI6{Rm`@xRW+#wb_Tqy@f+^v*}bJwQh40TO`%b90YT;COO&Sz>!=Gbw&rr=)@xX04XGO4G3WCw8fhZZ!!3VArL+wf&LN)lzWDNyitd}1UI -=b{1-NidxZfW!RDT2uJ=16An22$dsF9#9jB*yNz8qIu;i}VELJB&m=D3(yCON~pED*8aEnM&48;j6^o -|HwNk5;X3DeH!c;usBT*ICxy-R80+-v3*3thQIAcwsg1tyW+x6QV5Lry5P>tAv^#=DHj;$CT@)OIbXGd49R36mz3??U*YN@1%@=V-meNwUq{YL -wV{cWxLaW@RTqge7ELFUFMv5Prud}{0GpuX3@ejiORVA8|`q-0Y#|2FyFdsmrs5MVt9q=OB0Hmzy8A{ -Ty};th+Xd(=yP)n)F#RESroM71DLy267LQjxqLBO!x;H|Y_Z1dQ{uO<1-cuJSNsCXO;U#7%Y~4M6E|y -`0!w1cufdkJHXeEZP*yE3!C++2^&~s|6|w%C@~ ->aFYmOlbQ1G#|s(&EEDGYvO}omOxUMT&GOF&%;)EIUB;k5c@8i6IHCnxIH2ej0WoJm1^*hgJKq7(6CG -uCw0*{(z4A8VpG9j}o`nUFiJN^3f9V93NgZr3)e!gXvMi8tHO1BGPCk(uRDyRU^hUei|IFzS6;$d7I- -EJ#(#~}woPF5uhE_dz089Vi0Spv_O!zR7MInw&dU8$AzJwPb^^-e|a$6-C`MwcZ;VT~bR%@+M+U5s2HYObX=7ULw0v-XiQ*V_P<{5#Wk;3&S*{NC6?51^2~g@Re*EgkJ3gGy -3j{xNg%LyXS#@BJjo+wiu7k!#y+tx^7hzM&l)CJWCQTJJnLm5f$R*M3_9T3%*&4T9bIfK7}nj`2eBPm -PT)+hFf7Cu7(TtsB~MQqP`KnArdzT0f*5Ut;2KH|Jc-;%|K8k-S$!-Ub6^Xd1ZZO@&d%wai88rSQ*?m -rsTn8&)a_~JC<%yOuP$W2&2y&>5oTyMmGBE -Ccmz&HUM+uB((5y5~DPBA8Qk&i{wrSdl*S2X#7aM13$(p)%K4ZU2@Ag9aG1Uiw_%-#v*;K@P&GKT?|v -W52XuTD8auP8!Kepq?2kyp>T -jnepFXBhH8EDa@fjizMY;zgqDZ8U&V29xx3Ww*8B%4kz2~qwkhk`znY!#Hmj=mzW8OD>yt5=6%OKInZ^zJu!^tA0n -5dYU9~}=Kw3t-og*lAvGUJ033>M^}4Y1KT;F5D<`n-)xylVx*R;BHt_d+XW`8E&RNA?J73>vI~OBsJgmG5Ea23}-7l+cjzs!iGiJHEusUUX9oB% -%#$?`y!?gy)M}VqM~o#Gs8fr@dZk5XXDFEDO4V0k2rkR@8jS0wfntLj8mJ}5Nfb)r&e7miwcn9xrIHg -DR>XkunvW#WwAzp$2K4EixjIF*)BL3RERjDAN^~{iVp6M{`)>Qt*5ECQ<77`dZhJr+fF*D -M5_^nm8`hDcb001W3fujIam68qd+xx?>KF}|ohF4C{js+_Ih&%2VjasFfdJ|m?KQ!R6#m0}L(GXCN4c -xgT+*+!QnWC-{6kYnrK{Vf&gj}Buhv=4cKfA*u_=H*^>LxBPj*l?hCy0X|(KEIc@&I{J%9&#@u~lCxd~M&H*`AkI?$CNdlvn`-d4W(OH8X=R}-l=jL^0@ivcNM>nN25$DkX+fh}cUkpx_Hg^X! -y40wv%j$?m_#HXV4kY--mwJ%@LDYFq1|zS$!9%u}*%t3yCs>>WYM0G*lX!8sm^`~-PtGpd6lUg|OEy= --nmMw9Gv!?{joJvkmcp1!7RX`|yIK|tL_9yqb%$P6TmS6Rv8h5Yv$IFg8PcrjElg~AwRNe%V68cM%)& -EwlmbkFZfXEMG0*1c9BDgAqZv59E4lTwKft-Qz2funxAHlQ!d}(ZtuLIBu;$H -<&-DSLTvKx!>cE9a^ZlNe^kn7DKx8y9q}cIS9RremHx`nEr$o&H;_PkBsFWRjo0&t%lpLFvTs}<4L2* -LlfUXG;ko?U+8~X7oe13F%rE -mRQ97`*&&JUT(5c=Vh$0;xm -WL2`mBo7>{((ep%df$BA1kkMe7S9!Ow&IxcOfG_Osd8`t`|6MQJ~>=Kii0hc5i#GvYVnDgbG^(cu@z -M_oFuLG2sk_z*Vb#$05E!`o*teT{A%o_CpS>d@2J@k0Mx%cIWjMaL2qxp^qEx+ixX|=tXhguBi -`y|DDOVpm~iI1fs6kohjG(YE8tkYXAJl@WYgEPWXHGW0Ay%ET(Pu`+}oi9!XpnBqU$zuBHlTW2A-fDJ0j*}VGm -yo0}K)UefXyL&dH$Xu&?AA3qK?_$E+#5v}bu)*L*YwGZ;#V*cFXQB$o^KQX>rccc_BW>3izgp5(S-ux -RVUr%%86)!|p1C<>2=Ee0@*%z&!SMDJr -c?d;IQFspT7C2AQE>NeNI*mJggBK{J?DHZQTsmeAU>NAB$Swg8lOimrlhFIH={%S_8&FWQ|Ry#paGTyEJL~eheK7HU$ZBVVhPc}Km7QGaCLOdf1+f{tVGz@ -aPHEce$GX@XBj#Qn#7WDt$P)h>@6aWAK2ms3fSzDMNRt32M001Eb001BW003}la4%nWWo~3|axZdaad -l;LbaO9oVPk7yXJvCPaCxnh&u-%&5XSF*iV=IT-LAOoEm9s}_pqnlq6!f+Zcs2#z)7vD@4iD!BBxd_) -pBG0d^6t%xEzGDoH%S++J&?|Y0DStvCyzF7^C;U*iT?uS;^W>1$0h -ml?$lEwRj86YtO};;0zwGI+uFmI%#6l?n>~P8ibh53ekL)YcEa3dkXS|AxZ;uk>1OIHkk=5b{L}!6U4 -DQkQ>66Eonu0oitK)z@p>>MhquZ&;)!PkwEe?Eb?NKKjaCw%e+J;GB}lK7{cGslX1F_24i87kM1u;+w -Gdn+|Of@A;mL2!pbeNs(HOo%V)H-aXYb -If?GOYqlOoVO)YFBIQiE4UC=4k*O8*cG)kv7Wupxev}hahkuO8>Q9yU%~0wn#emPKUk))RK1h9y#)dF -Mt0cDi_!m45U}=t%=`yXO9KQH000080LuVbTf8{G|9Bw)04H|<03ZMW0B~t=FJE?LZe(wAFLGsZb!Bs -Ob1!pcb8~5LZgVbhdF?%GbK5wQ-~B7F$_Gm+nNB9T-K=VqD#!8cp{XLy&f~Q`z -5s{Q?Mpl;h0avu;#V6PW}WK)=x4=!OP^!Cs|*ZnUXQVzes=W9~FKShnYMmD8rB-RNR2yBH%+$*0y3UGAXuqgRW88phxG2l(Oq;PvoBC>ufIq{ZqhxOY^UBtnv -Z_^S>^HM@U+ZkG_f}Q8R*N#vVQ8GUl9%tQrLOeyWwxlf%`&Yu_E28$rOxYgkKZS=tf+NWr1`9=v)loy -%UO}u*^Qo=x`Gu*>vWlJpt)=IAAy-xYjd0DKL+RodkBj2@2FGBbe2#TEU%xA#D;$_nP6Z?R!k=%u;>Rv4>F>(86g$lusXU8IZa+q7CLCPj9U< -ypOzF@e){%LhZH9;l-HIi083nAfNxWE3D_%8$JBbA>}4zbnr -!m4MwU|3sT#oCo_sf$(C8aG&Wy%UQ&K7ksJsqlKJyxQc+XpDh0 -1U7)_0-*nhXd{TQo{_^AOWCtH-NWImElUIKM8b$X5OKo^#X8(alr`bvYs*yPzFbIKfXR$qYz -bG=dHV@um0E96g_d7@seRLV594R8a&b-?Si+K#dLR^4WKt}Zkq1R$2cO(3(XG1zC`u6u5mx<_A-U?Gx -lfVV(z#tE|*h>lyZZRh$&&v*8Z#^L@C)C*!#1XPrDurxK)$z`H$bd{~PBRd?n8amh+tPiBWSm$gXEqx3KuDSm?zf%3(csWm@OG-#<&9R**f2|hnnWf+YXb3~s1 -x3;xXCK3482Ql%516992gxJ!@!W`Jj5@XiYN(fQhb(BvjU*b2-sA8ye457f~vR~A=5^yfv#&a93jDOE -Ju7IDA)~vU>^stMLFyr5vY&Q;M3loN00D*pb2Pyb2yXT&Z6IKes}AD_aOGP7ZOOLE=sUUobiRcBmflw -On7c&>A`=6Ucg&K;)PzMjnSY%TW!H}K(+YGKB)2ym<4)&qCS55LUP>lg}SJ83N6q4p%j<}u+8V�ha -E!?Ni!z7=3Pc#(io7!K^XgAvZ-`eX$6$2GN`Bb5>03CQW -;6C9h_6TH$Ps&R1yan2jg;fZJ1`Bxrh6x2Lk61is-c@tFmqKAgTnG|cCSkonwFg$3(nqfzKkzQ8613q -+MGAm(k$P_WA&$G|uiBFZyvwxGewC24jKAn5o-3bs3laIJEkf>4*B)Ks%6dFLpr>@owJTv;Lz%euz(Z -MM)JJm{rU*x^lcVQ@-C-LA|*7y|vl${ZzLN8FgyqA@iXO>4w5@?00N^vod57lI>AU}X26u}{9`br7)cZ2h>Nl!vJ!)kL9}9c8XX~+t|P&YS)G@PEMVfTvARWmI -N;7i0@pw5pO8ml`!(V2!vVJ+jDilLmtv=OXnT*To -{t<~?Q9su`v(uddcs_~`vYy&F>u3v!vF##(j2USU;bnu-1ZM7t?lI>fvn8CPPf@A6dLl@2@c55#7!3; -<6s|;3ku~nx(9IHQk_vD{g;-JknchR+_2i9JW1)*jg+$fgV0YwSQjeflJZJ?bU=4 -=MP0dmd_jlUOXUQQu;-gMayVxGohQ0ZrXRos7FC!0w --~bsz9e{$K7)u=@qoWf}?>@BkCv+cz^|rq!#m+?JL|Ut|WNT;*Rsn4G_2Meli2M0_>eab%O6qQl?aG|TXorF@Nr`t>eA*{5=BR+g6k|$^IV2 -*&o3vKe2PTGRG^i$ --t(Z^dsKPE(1_fcv>dYGI+Hmve@FBI?Un0d=wo4 -|o(o^C)6cCW_Y9V1V9m#o>|v)JepZ7MSgl4&@s}>_FCG&N#?Dii;+GD%4u_pGa)~Wqtp%Tt(5BDPJ$$ -h#%RK2)evy|T0UeOCr#V{;1FB#vfC7>|Oo15)suN9-$Il);1NW6%Jsrj91?j;!H&`?XVK_KE9P|*5wf -n4MsrEz@<``IR_QMSalR<*VCc}>%JJ*9K8`y{rS3Y$X`?jpsX`cO!B%SsIaXo8vFsLpmzfX~xnowW;VFoZw>j@snuLJw`1X$MF+0iAcJ&@)Vr_KZ3OeV8`V -;u3d*yvt%gqzg1cQ8$8iAkwLQ>zwRcH)>37ve=!i?c4FzIOMv22*&NzVO<{LM%+{jy{|BQ3b0Vq?xQ_ -8IXqmJ>ohBpI$w6%k0QXwU@+$y`2B1w!}w4B`0V%Oz-=^ifES(@hs+jJ80{g!MWJh0VHd%U>xq{w(~t -dQ6k?v3ct-Dn755CWA8%j@ck0Mcw0Z&jB&vJCE&5Sz$4MajaeO9^W!ix2E -H6Ve)Q~Vj6%YQcUVR^MXSBDdOAHzYK_NWX|;7KcD$n{&~Tj=K*b#`dsNlESMucPA0KsPsi7dSus53C1 -KH^gUhCqrz5@O4o}6zt8vj18(!#8Cbzt|%L02JSd|Etfz)UzycM-i@U_%$1ax>)beePCoS3JXm#k8>l -`)L=tYq-AVG=%{lZx8GwB7U;R1n7<0Q*wjyC=RKMx5*CZPRF|uEDCdK7u-zL3mXI1`=FNA4XPri{_6jB?M{UJMC`56yX@W|CWuW5%Sr^Q9Q&S6jYBU^IC+Ow?mH_)j9>b -c+$8g^f;Dk1#~*VV6Vj1J9}`u%08dLWg&D}qce$j;Bi;kvfl@bG5d25N;~dH=!AEtNAD5XJ@$U&V=FE -O{k>kjA4D7ck3=5$pZ+E`3?cvaHvyBFmoyW&iW7@-cFZx_}9eMnGQ`EzLEDM1%=|!qMjmRPW{-xj+8U}s-4FlL08a0a@AA -|Rr&dF}kRK`#z9vG0S$HQMayP!-ZdtQtm8@zDm`?*A{7XTSM0!c-w%CRD*?&W!Q$%j{ba?mOcR7$|S8 -!*{1C>4%=OUYoM%uI&r&W+iQ1voce7{)(;_y#*h+@BsByX6R=_#ikH2{a`i)}s)2PG34MJjL9+G04f? -6%nQ5PoNYYhk3L^RFDCL@q6-ZBT32@S_&u ->4IML}{bgs_j)TErYB-Hj(a3$_ECWDgT{VU>*hqJ(V1Fi$UZ8k;+}%+>X9>|*w*ROrV%A{UC1uuZ353 -k&c|0jZ;{l9N!@(RfoM8J~WtR=7x}bB6`yuR2mF)QroTTlM=YEq`8SaB75mvzTo0t_ooW2HNH<)7ra2 -=et?`H2$-~Ibvm_vv)cFZPG@9^O777Q0|4E(kRx5#GhFg~#xC;5=xg9YQjLlA_8#Ir!3pv=-(8;$^^E -*#-$5uW}iX?e=WA1D#D87MTu2z-xYAM_qel9uq8OVVOI(37eLzidKj9RH}UaOXEJNYtk{M)C0t9f&yd -I7+VEZ5fXvQ5b~^sRt{*TAJ(@?Y^F#fOq1@;2u?YvT*{PEChWNOfQh=K%Z++NBpvNq7D&~==tMw!}^j~0F@ZlR#LIfOq9rpOy#I -^ukGB6sNDG(2j(a8_FsLESdAPfbYO;%R@d%i&L5o3PibCHb+)ny4LOQ!2pQdXCTSM@qStX2z*Sw1lAf -F1lkd73=+Y3%lKKgKnkN{q^AJu4}X0_iK0`rT9wempsPISA?Orabu5$@##a;L8%YEv*SrvV5bK;UfirZaN7{yO*#k8;q#`1Jgg#u)!?ZR95fbot@)GEvM_eZBcwSg*iT -!*=DsQH{)#mCDGn+0o-4K@CgfmX#o#0?>VJH0^pWc7{*hN|M2MPH>{L1S-h2EKp*FO&z0_AHE3vgE=( -94_9YIsNufvXq}8Qn^H}d~RS$bIwt5FMw#Htt-~i8!DA_n=huhoZ8PE&V8|nilJWM8b9nYUXx1|VCG6 -}vh5}>97V7rcy-+(d@_N`3ezZY3S0tGKzjQ)Fjq^v%E_vq25=REqmic{<8*za>KB}!E|IbxRX9I&U`= -RgkV-(Nc$Aps9Nn_*&~na%KgyVW+Lmj|swIuO^=e}r1u|67!<|A6@ej5D;7iE@Itrbmy4mEEblg_X$?;-c-ft1ne8l02Y#A -88#oU^!1JItS%SD;v@a{WGGrCbWIF5$5?4o4<_%Qiw?H3Y&{fXLNm1S~~R+A@>zk9YncBg9Om7&Stw| -V;L@pt=QM$s~XjE!u~l*D`QWw$o(FJFRnpWI;3(Pd{!hI>dtPGJvg%3|^bj0bb5wtN(ZMTKcVwr^lSj -9>&zO_5;ZjJ^!jGnkk@G}s1Uqd^ve!PhX~j5;lOQi`M-i+ha_q^!#tL+Hz3G;m}EOlE8;$(&{ -`!?LZ7G4^ypP6M79MZ@_31@-K3a#>S#gQsCAHUvj9aa`soIJcBIrYU%H7}64~ii~R1+R-otm@3o=+8V -5WwM83*LRW&=n&tk0=^(j#_f8hWq#yHOsZ$?vDUcRz)+U? -mg=7P5L=oH)~Zi>kC~4*;O`+fJXAX1@a^H(scHQJX>d2=;j1F&h(ipS0qcuC?3;-1#w;pCXQm%fP)Dal -os2zY#YT?Qg%BCjAgo48ElvfUULj<|n!xvkZjhj(|bb5!LZi}eZ_l`U=QqpPjd`?-{p)_)J2&feH(&l -ql*PXbI^SR5t&b!A{{+WWB_7!pufB7;B+&NjGm+)n+x7bv$_3Kl=RDL*$m^{_^RoC!y1y%VfX#Lq)iDUH;k7~R=j3qx~4LjK|tPs!DMezY%5MIC(MJlN_o>)hv+Uoo0!Si{rH -`jQANwG}f{7TYYvzqt$9dPIBDm6~rych!&zS@(KyUbWy2r9L=rO{%twUNotoSpkYOEL^BS~^nTK?gM@ -C}l4x5HltNLEmpz({h}RjXFt+u+ST4X&K)-z4i&{q`;0*FUk`1L6vS=-O-7bc_p*%n65`6qoa|#rLUX -Q@Ibwxf=9xUt*n%05p`!R>@XuEiU|0xI+|xViqbU3ozX^Nz$8%gKrM>!#kyN?W(&UAR%lTweukwB{^_ -GTQu36mkJZpN!-ImC)?W$n-G5}`2fe+Z2sDfs$X{@BkP_n|dRMFVRXd?l -t`tHbB?Fy~jVo*FTxEaIf)?@U{4n)3AUvXo8>U0AU=sr9p9U-xU!)q-gEzFid;vSH+tc`>sXQ9jO*|= -G$Ey*`p4^w9`;DBAlRWHti{mPFz8)b?7|kol3UgxSP2q*h|KqoFa3ye+Nt;7z%4Ai0I)XOUdx5U2pi5 -nJRons0<(VSAE2Z4`KO7f2{|jG<+uc9=#4B+l5r{2Yqwo(VpR+F?)#DjJ -5HrN`iJr!?vOxKBxwzl~dx7`gew#ftpR)0`<9&zAIbtF&KiqXW4q; -~z&!96jGU$$h#}slP;{iP`kSChP7$@XXse-}VuoV)>PLh&7C2I85`5?ap^T7vqKn%wJDs+29xeL=PE* -bu@4JFdBV&Ht^U?S3;!4Q9~Hxc3N*9}E_;*M3#=lj@1!fuXu2Ryi$&%><@CP=X)04{m!lxDce%XHZpo -AzvfXbft$p-e32m*50|!f(M?bS`lCGQ-;M!(*&RC-8g5iEFZ!y$?773pH5Y=u|P!v`MI4WZ? -+}V>srdAlFxy){Om0V1qRqOerB{>rP`oJBo7RT4ME1={eJ|e(s2#$Ew-&U0+ojMrI1DxGU^i6 -a!OG7#hJ|_^BT$_P8xPcha8Z_dM+5P0u#1Jh79@_gWr=p}2pV^#5;vW%4ULVI1&f)b`i4UMVV<_tSo9 -K96*r=)IJsH^Ij}PW-r|BO*rIe8EnBYvduffA1T4$j0>vQ69gi#*3YjKhbW_4kMI$mQsu^1dFGo$4Ba-;d?MjjMK*prg-i=Q -FK_(;no0%L&^PvuR+9h$3nj)`f5Y2A^rX{!ms%GYWx)im;(1Uy6y2v1ZgXWA*&|0*B4pE}Z8pxqbPd~Eh+k5fs$2tXfct7mv^f)(YGTR*BIQ9kq`fjs3EwkUEc!~@D+Ja -Z|xV)%=_35He1rE!(rfp&wCyfiokBQNC1tyB6AAtA6mBmsLwDABd7;+%n>R{5pNe`mka4?DY$F}{Kz7 -g$3W_HECHyLHsBDdA^uLMtANpwDM{xAIekI9|&X0fmjp+Z2h>irgAK8~W*JLW+cOQ -Rw^7`cbo6YENU$aHpD??VW$^ytw{!*+=XV-)+*ZjmR59 -ab9BWXg$3;|L&(*vd_S`V7}?Q45g|R@GV(P~$K}FGC%P$GSg~zovx|d+YJW|C(kxK^{rr-eC2`hvzVg -^B^p;1%C{rQC8kUDqbVLnll+W3kV%wu#1yWQ{#`R794;BpXjMub!4tm-OIK#Kg#d^z!b6(bXEpVQ0=( -1&;W;m^Q)~>JZ(=$8=nm9(wh55ElM1ZgVof_b?E<#k$QnnYRrv@eHG?Ja$nzk|mDTn|TiU^c<1F2hK*2+Rhx)sC?U!S}i8fbTW%>-VBt4((w%jW`U{x@!|Z5|Vvv1L#j41IM> -@cwK`+__M2Kwc;Dd=_iybDxAxR8!8Q(fOHJ-v5SX*J1bT?{3{AWdtu4QiJa{6p$vHwOfmv8d<4jduST -;PEyw%WEx>iU42|PTZk0q}sBfd|C!@P3gDXBs3_Lww(bY*@9=a--F);FM@5F3xSTFpu2BM3*o)qua7> -)w`QOep~Jk@{yWiq3CI;M18NB>nT_h7B)5>~zesGr~P6I}tvHPB~uxp`%O^rG*YK;vyiwx*a*Mk8v-8 -D}?|-5RQQ91?gA!JP*Kp*`!qA7AOkHUIvMm!ZzJl$dIn$t8TIT)Q8hG2c>{4?N5FzAvc$#wFc?HA{kq -jO7g-FiB{g#j@7TwiE<}3a2P%oMZ2fWZ0ar&j)mVXHRSV>z2BgRuKJ*8}A-~o}XwM@*V90ermsjTtq) -@GR7Y~DSU|b$58q%gyaUE=}bSU65p2e8jto@?Qd|;f|*V~ekov;Vp0Oh+|vEKQ`IZ~5WrC95G;*^W^^d?Ttuq(%fM -qh$+IA=>cZM8!k*&s!1wRcM{Tu_32a$?fmez>dCcmYn9{UyP^32;e{US*%hZltUTt2+9HRPNoy9kPPW -&bv5*;n_F6S11$W1-{RIZ%eXDlufmm92~afBBw9-p}*m3@V4x$PvJXV_(}j=;x91O-&J1*UND4d#{~q4XM|K#SQUw-U -VSrmSkdeqgE1LM{<0v2F)5!MHD8S$*jz(fMl)miUo7d}=+lHu|BrWTMlrak(GHy6#}fUXgu0Qs2`T20 -8M~R_~{=e5fH>!+GJ4V_DPN(F4sxcoCEjfxzND3%XlDM2|5~01N9UDB1W$sDxkT-cbJaJNVlz&DG+6i -Pa6k^oKX$@MCV_&B6j7%k_?#0O^-Y(=$oxHc!Jc4cm4Z*nhkWpQ<7b98erb97;Jb#q^1Z)9b2E^v8$7t3xNH}Jk+L2 -M9IY8}Ov<-{?NT>3zZ7H!eymIM^FOG&JE$t}sXtp@#y9(w85^-DUO;gXc?R0n%GFV1_0wyx{y?Ow9=x --qSE>oqf*-(=%iE4NBZ#{}!eFetraTWfmOn7%ifY0(32Yb^Wr?c1N3^TLl#y2=~o_flD=$6i|1h>oG$ -6DFEQI;S?BWKsc*WdekIZ`Ee>Aj=@2RgG6h3wx-Pm%Uq5)z^2TA3Aw;RaFdrZ{ECNsvnH?%*x+K>AXw -u39!IB-+cxWfDnXnD_dp3tj|6m@%ib?PcSXK6oc`IU7cNiQuKTCJjz<0o>tYbrk7+X+lw2?`mys0v?f -P^^}1ynUN1gWYxnfsc3L^nKP-)YE)USRyX1 -p?1bPp{alZjAOykFq|QpG5dsfV*4en?2+fAFT1FF`X+&mVpEim^XS?pRigA+Z~<(hj(?NJI5>bvbi1% -V6ucIN*Jx;-Wc8)k`W`xkKmDIN&1@WKp31C{)6{8LFg+(cgD2g`#oes`2X)ee`QH%VPmE6vQ0WI$*NE -c!!aDK&_;7H`n|CV;-YROSK!nj`f*HjcfNE*Qh4yVEPSy%y$BcxrSuue2$0!dpv#5zN^~Vz1ZBdPfe@ -3NQw9a9De3Qn@T?wCs4VUhozj7`+EA+@*GlIsQ7_eV=rI}mg#8k$8eFi5b$yx?3)eJzqab|}%Q;1*H# -f57P>8+IU}1*=MPQ8PvpO+gytZc4Z6O7N<)q{$bTUj9{73{b4xaa>Ra+%XW{d<_5|aMzP#|99ns|Uih -d=CR;pz-cCL~bSD7Y9u!>>-Z7m_j-vI5#SWEgaNaHE$Gkr6UK32~TU*w!B54jpGBp@*V^Olfw$IE^v< -vQwZ2``TJ#DTpfy;zqPQ)@m>bj*~GfisGnaLHtp;cNN=)IP_%Q?<1JkQW7I9OF5^|+Kp;ksS6P)dNJ; --ewI^WMFH$3+}-j(6?-(5X+9?U7n3+DdjT`Bxa?(Yb9VE~(V1&3V;JGg6 -|&&qVktHs@`pn%pZ^XXyW((S@34=<-hLo!=LVC3I}o5~Q^XwciTWO}b}ry8P_C)H0ai57du^-O0cf`% -eK(ysMHoHS&q^vCh6aqq5TuCNQ=a-b3aK?hgYo-N#^0$RtV5i=#q{N&u4;dM7a-9K^$I6~+0Jq(`X#; -Y}A~qR6`eid7tWI8j{2HVG=@7DKII`chmJsEKJw9Z@-I1F|T(L_%{^$5|Y7ivc6Y74QduG)0myjC*IA ->r(Y4MwATVtsJC=ifmARaZ1!Nmd3Uc(h%KuuOURAgPZ_@{zP -&Voilxp}!s`U%E+^08F&!==wbjJ!v@%iLQ0WiTnUIajvF0Pb;01aYZO7UQvm77z;2__B8B$o(QW8>ar -9Gz6oTyHMfq#u!n%f1;FA?h8w^=V3IR6;~R(`h8$Wd@8k}MVW2xuF9p?fO=sLA`S*nVDl -Kft<&3zX83F5_w6JY*hZI}}ig4Tg^K)o0QG{YAqBA|Hp|6az7bXl9R`mxT+5w-UCkF;&jrmG1s!g?RSo*u -dDstjvY>fI=IZuVwb0``)hp=Bca4-SX^?Q~dKwf#Re9$_krwWwfuW_pHTL@dCR<5z|9m{%{;aC$&5<7 ->v+!1WUDL`*Z<2amRTaEn8SEg+&*oRCQFyEd^a+E`)rB{WW<T{uK0lWHDd`J*InL(OEyxs%Y)ylNg&cZ5zGmTu#c%K^nr-X#`l -J4Hy{^$VosEAN{(i_>Lz5V-X7%UnX?w+f&q5WVwP4CI5TfM(IFQ5P_r1Pzd -+Md6|c!w^^GNZNSGrFWNBg&_Iuo!RBbN|v1H5Qp41Z@%8layFYK2A?{#&gRycu5_J2n|y!yOOm(337A -0Ow!%THcWB`6fg_m_W3o27hEgkqrIWg`AnR5e2Y;Fh4OVYu>3UFJ>bfPG9`H%Dt!y?KeCnK7DeRw-&a^1<1&Eal~Qxm{Y%=HQlcLO~MCrxqkGf>SOGY%Z+GM`aVkYkflX`ENrEZ*%O(_3Ve74Tv56b%)O$H{rsjZv32tYLeHsRN|wHxzkIQ}nIA&M?d`k2?ihLs3vdB*!wx>OnSS-rqmVXJAWV<`2rfG-Zo=LpM~mtO9h$FZ+o+MgBe-TG`1v9gg~Cri;XaqLK`Ydwp+dPk -_ZTBM2)Z`v1%hXvLVUuNmZJLbC`#qAPO>wO&bnA@7hW@WC?IqPn}TG?ZFSRkyNALZWWjDQF1@6;Uu=C{2rk){KXEH~Fa|+!J`HKF>xPg%wA?c^Q2acCXcQy!XF8#9S4sSG -ef(EYU&JY47wDFC$5P<%Yry4;j=n&RB(6k_Q3j{v|E3gR~VkXMNCF8WBWx(ON&YCFOqPYbV;8p`SbHw -*Jx?8y|KhOZ_;x)MI=?^bZA21zz|}Uz{wC)9jrdWQk8>FEq);MDfxrNIP5+sXxu&@Rl8Z{_xB)4zwd2 -l|34(KDE!b^ik7V~`+iRhUpBApO1=R6t~}n5{${xL=hH#g@(|`D$IoZAp8e -JbHUB#Df1Ce4`}q}f0skV;2i;(6`g7^FMt7SnU1f$gz13l7S>_U*a{RHwi6y6Z)RK$0T;}96Y&>6tDH -?tst?8chxKyuG_C|~3e^5&U1QY-O00;of09jjH$!AdnDF6V;fdBv<0001RX>c!Jc4cm4Z*nhkWpQ<7b -98erb#!TLb1ras-FKx~;g*in1if$Juz3I=0-Xb!=BJ$!TRDnp$EdTD#MB~+*{s_5hxfO!adMeeYF?zvnO2LkxXw~tDXf~ -+cWS!KCv{fjm8vh3S|z1c%Sxx}CaW*ATpbDM4O^PbssgJ_moqgjN~LdWUFOM5O#sMpzCd8nMgVASWb; -K))=-e@N!2{%HME&!=jKgTm|umhRciCADkfLDwoi4roHS2&wn44ui)l8q&2Ex1&+-df{#(Yy)3TTY4z -rn-VN3a&GS#I{Ut|ECD=o4`6kqE+Ey`$F&H&+kBCu&OPp)))sS{{d4b>tkD;+~UX&sqr6I}xN_+1R}$ -_l!TV^i$B+Su4sbrI)Don7k~Mhf^Vn7U<7Gja*|4b}OwmZtKGCgD7pT)~Lb3Whdc0DyCvUDIur^Bl#* -Ag1e3sZEs^za}H~^3lWH4F!Mkj|b0Q92~_zKYMlf;@QcYqvQC8znvTmR}~(g9C2Bvh+ntIGu#*mQjDm -@g=+tlEEZG=AI_6Xn@03}SL=#@Ug#Pa`8Ml5$2z~x$|9fZyyh|-`-Lv!`fg#WXH_f{8t<+`i0AXNt{3 -qvtAO<@74tiMlzCq>6vXE)KEGQemGtVq>7x1TvZyuxghBI2Vf}?=hkH8AuajAp{zwD%UY#Dj;^))6)Hhk4Cg(H#Ly_Jclx0yic(Keb% -LcXQFtvwyrSmF7-e947eQfl(7RTxCMuFFX=>oXKFMo;(D2_`8CKN00VDqtG3I;sQY;6}L51TO@QSa&-Z-rYl`uqvq7 -tS+T5Cs%P0e1A$WK`Z9rcWn}w%_Wb$5+mra<_479`4qyKXdPe-Ck42Uu)i0KsIyI@?p5M(pAwa$;D5isyc -IVU>H7k7%Zt~?Y6hIMTvQtbB-K-&bxVB#5>VcSmqk?%xF-_z@~(LxF%wOk0ef875@tSiNQJ~HeydUgt -K{k>tTTFJ4RBIvx<1x*jd~tP%^gK=7CQfnz6-tT_$TE*{^j6r@$=K8ql4Ec@#*n_C@>F);JBF9piuR@ -pR!3=RPgjpDE037vMfPte`XrKI}}8e!6fkE4ZNpK^%iM)kfW@Ch%RR-u#q}C_)~oRcLkapf4%tS&C!b -?sOEAGa`qc71GNWc;Q$#UEkYUGUzfVP`#G5{^}%h}rV+5B(|a&Dd-%Q=yEnieC9)03K{A^csQzX*$sI -f?*2rCCY#TcdUxbQjM#!{zBz$wsszfmu{5@F1EtWT?K)c!_5>VE@GJSpt-))k7!nh}wdUB<%5f4QR0_ -nE;QY#}53YY@86rMpKC#X)e17_>#te7OT@3w@3u$)nAdW(|8%~g`8^nOwRmsN`*PhE0vs!5U;c?NB`e -q@W!|dH?#oa<1RZM*{(;QSye`x=JWl$qtYvvcgQY}=HG8e(9Bl%D(hyvmj0EhF^t>bV -D>SwQG{bQ(1h`W%_2;y!>ZvKm|@HQ -EAcBc`s!h0v>oo@CQZr;WNsMsheUa*=bDOTW1p_N%5xi~(GRn)OO6s}{Q_$65?#!#e8DI*0mPYg%?JWo2RIsX+3$VIpIu8qJ -R~I#GLEz0reF@D`A<+8?FR1aNKywgmlCgSdu^>rw>~&I*S(W8go#0rDA!;oLvrQ#+)n0V_6=*D-)VPe -ztaeLq_dDHwq7VoeUQSccHDF9-`Sz{~v#QvB^5pApws!}h@nI1_voV^?GQhQAjRLSOBC^at*a&Vg=+L -l0D;9?y$1W9H6k%d|^_os}stO1&iIRw3!Y4uzS&xzxGkoE9|H -X%lENo86&uL(2lzSz?-0SCcYZ)Id#IM5-DLdYU0*<2D>w)kvVzBHyaj6&l7jEWJkFp+CT@Kx+p!hR7( -t=@Pl6ts!R`J*mP7lgnj(h4p3G>&UsMt6)ed^SDTN#l46AZsRrjur<@5@vp!#rA<$(_x_$Y)iJH&hi| -CS59kww-f3E8n-MHAf&{ -tXd!aEV+70nk(S!cw5CF%}Ddz<)8*1dIO=*%YY>cj4!)(O;#K{saH`q}MoCI;m^xkWO2b+$V!XpCX7& -DmWTC(HUW1X~a`FiBSFSR>linD!=J0~w{u>dbn%0OmmewBRJ_BD)6Z2VlIPpbi>infoJgrzKvj{0MLf -19EH1wj1r)rf#o@aXIc!*+Dy2Vbf$d5DTwb-(nj4W9QTKb3oWnQMozJRbfF-2AWYBuia~Wc#95Zalug -OMyuCF(wOqzhMYQWVdk?8yk=@43^2JCAkSd$l;L#H))-AC*!)x>wCvfAdI&+4PK9%jj4Ri(^Xwc4s*@ -6=hPRs~P6}#W7*C@f?W_^-Ue&~`+5j&zK*D(JkqMhI(T;WuH6r|#C-Ze)PG)7OPXC4t8iNSA69gaw;s -RT@c9ViOH}BJ;ve!m?1&DP4leI&p>`hUoo$6E}#vwgypq@Q^PmHK`S9u0PsVL3g4^7Lx_ubt^hu=icb -hZpIBu6cnu+;#&h_)KlRkjdGAo*YBU;WV`Hq*l1&s;~3u)l-_0RF86o*|CoEMK$TqOFK%LC|9rsSISW5Eg+(r_*4!D?3Vef~QxSH$vIX -X+H@EhP0yri650;yPxj5n8;&?y@4f)B9Eq*&Y;>efjKL>wZsYU8#h-ZM_tOB<(82{&LucHD{0W@eqeiZT3Y&j854Lr;uk -8B|O}LqGuv{JWfM@xfIMC{erk+i;Kzx%lLzk7SCjdK@bC0sBp0GMpuaDjW6#W)!fsOF--zo@qB7bhV#~VJj -4q&CBip+IYeoXEpM4vA{@S0$Qu2z(fw9Cj@pqb)va7Y)#8^3o>xg>L+EC6!dv<~Ws?vc=#ujt?Fnu>_#S5qUO5K#m2sSE2ei@Wz -oOxKSP{B=$87IO3A&MN9$(tmvn+2C!R%(6);{isRaNHww4>_R~7yafqY{#hwz)0K#RtpYuOV;(e0QL#_g_xoAZFEiUqwyz0340b&DIZB01%+T`8h3ywixdNNOd5*O8jFQj8(cte+pIORK -4?tz_K@p~QaW;ET3m++MiU8%)G1qh(gJ0s+X|gRL&HG{^RJUZHdG+K(U5h+Xw(?^-+lL8&uGrVEqKPv -6L_^XaI9ko=9*glP`B99+G^^Z0HbU2F@zMh#gL{VeWHo^B1k_0Kj@eb=$x5gX6}HQuZ8XF;{P1eA3C)g#^y{1T&oE2wQSwSb~U@UOV)x}#U}q0kpJtV;IS>D^#3Bi_V -z6)n@(RmlrjI?8zOeNc?0wBUZVGDp4HXzGP^2vZghF|n_gb1&}Je#d-_2{!qLgaY>Nl^5~wSR`0}E64 -)D(J)X~f5YVYCWC+e89fYqKP6ddO4Nu4ZbNhwLam{L1Wwkr)!2?`Lyq+CPJESPBw=vN>py#8KFiV%i! -vV1aI0);7)Gu@ULbfnZX{}tzgQF}Bx60{e%XQ~U^)ore?-1=;DQb8qKAp$YaQdp{? -IKLO695u<##$1$mpD(zFx{MHPmYY(L+!KC?t4t0qTk!QBYM~hwp#HWggZ8w%K|bckRi3)9oZd8gIh#F -xRh#m#yJ=Nn@WQ6WA0ZHu*|`tF%(`~Bw6_-io-_TKx(gbNn0MIEcuKUcl9MyLIfNJ#oU=LW?R#u7zf1 -S)2zA-&|6lH1B^UkI8?ZmRyUaDFdd*cq{j5ISy))1q704D=I02bhg$#<`CMRb4PfcOZtIFAbsF0lkHHw&ZrAf$Q7t3;{IF8^4a8pQ=eE7#e})LXH -zvdXM)_Zj$Pkmt%wE-t36?Nxr@1VtmgU%(Q7Sybe32@dp}G+6>h|$r5XlOLGFoU3^a*34Q)MQoQ6liN -<^cbBS=^x0jE`AKJ^{huOcrDtY1GTRlrQO^i#q|KK;$tfuz0`yi7PYD(?f5*yVKoxmT_d=QSm{gX?Dw -D7pPYiCJD(+i$;ZtNiSNoiHHhjBrPxaRDwe9<}phFJl$vK!Rl)=GH(T0|+>o6hgK+dZr^94 -m;U0)Xf<`5C5yimWv^U*udj_*KgM)hqS80Qnk?AkoFej>v&X1kCtSc-nCv5FUKsMJ$$)ZTne=p<1$wS$#SJ@nO)r${M5Zk{czFj%DLftnXi8MVF4AI&= -@u2J#7ndIVc+(hj5v4NgZSOXnQo#~0}>T&nKSY;vTj;t(4I1N(Lho^WL}ORwA1&AcVohCOT*PCj;%3uwjM;Yo6mZfPA$)N^Y4XW%kVswW>uO_AB -^1WkUyKyU1(MpfFPCUJ(7ld#IepaxWu%teTsS{@ebLAPQO5hk8Wm)Il)yKuA#iWQ2 -y-_%GPHxaa>kvi>1;c5|@qG8^ETF!!tfSKbp8Ahttp73=Z=qql{7uR}ZV5Z5XMSz+HX;*c=ICkqLD~jU+^)qgG -OH}oxrW2nSLn^nP!qSyQ55AWg*Mq& -C&w60~*VA|3LkU2&rwY3``u>?_%A0XV81V&z#PX`G^HnpRb(I*q{u)L^#2A(`Io2bz{rM`q-_t}811YUG}1v&KCqfQchULREe@jv5!Y5B)X0`|5&x449iGdx_S0EO*3122xQ{CAI=%4A -0PJi*PVBufPP~u>Dn5v9r3SZ6BQo!B!Yzg_TSRCaB?=09h{0-4qfm81uiTQ+Lg!0)I($HqFv-*gY2ov -#xLBVj6?)0At&Bf$V_Wj!x=Qy8*1{uLGly6P%b)~V)%u>AXlOT~{TCvO!W-KG^V>XCqI*2ld@ -%YrNho#AA)RaNjju2BOb`KHj%MeSCdptD30~w%!E}_5sF)3C`Lv1c%3OMgl`!^j)S3=g9b&fh|jS}njc%1lxBt9Jax<3L*&_uvdK}bRa4Gptj$CtJf>n`vki*3iwzh -Tn%Qt|&Ii@wH5X*Mn#9ifspI*Ot``4r6lhdDn`Om-oJvpC%ez>^IK3>h{d9nD -ltm@_U&F$T90jQba;qKo4qpu(T$CGciqxc`ehS`u92Xcd6xGb|qm7M0=x|$@kC%8lZGiDK(SH{Cek=c -aC)!eaV0owf~k5^6sg2Y^FKDKNs4iyNWzDQ=0sG4s-JAQt6$UA)r#%SbubGCGeKhbEZ6POkOEQ1ov@+I-sruLd*>uvtMxSa -{U`JIxlw4{cI(?18%|PcdHRHzQOOfCmzmGSN1AgBoZNq4RvU%KL&Fj?qcaA&@+-nqSu8Km$!i2=hv7} -`CX@y#IoIn+7N7rgz#H*>yErTF@qlYgsXOfHa-ZX@RZZZw=b2lp2RC^wf~XJ? -_xTQ`3#KxC8!0ZYM%_yg>J7olc<6biGv}txRAFW|l%f4kQjDV>-yWHlR`*i^*6aX2TB6Sp>RrN0>$5Z -Iy2LVo??_2+JiLR?>G$XgNtR%0@OtMZ!aAI?)QMm06oY4toq5A9snTX|f!TFJWjR%ocPW&dJ5|ge1X> -aVm9JEa@<(xzz^5{xU=` -j*E0qT1v$SI&clZc$&*1$E%SshTOe_u9bfSl;gxigt1uY93WjoVW0V2p7&cQmP;Fpud4iYN6nG`01qq -&=taVd`uFCmB=xI@761TLXMFe3F241H`?JtBVI8uA{2ZN8RWEu^(CH2pNMO}xm*k%}x2J?8|HiZ -2KvyL0FER~ed18#({AxCKuzU_)TUy+;i%n%Up``s>L2_i(*xl_`06`%?bYf2nX+J$(q|;2HuugQD!yXn?vCY -O55#Ypo+bW%f*5#NqgYtE3Frl%4{rrb8YIfl*37cD|)dJKh8tPJ1JnCH^2yzN@cuzH16<9-witA+hQQ -BgCp&}jtEW6i!?3}I>gF4+YrxcX-whLQ@oJK9FxH7SGcmtXaW=VmB6~uV6Y;pYz|XCi`cmqBxd{Zu*| -DCDbbl^9*x+zNVM*WFePJ$EP;6|CKSa-(#}-`209npdkDInY~r{oIh!h1fKtY}X(C5L_%@}TogM6`>R -8*XMq5K_Z8nVU?@xlg2$HOLx1JnvF%|s%hPl1P+#o2Y&wkW)BlN>ryqn}9y -5ct&m}O}QYck$)!x!{9N~LY`w5Z?mkWiSGI3ruQ{dUz!);3a1vz|*bq<+tqriIN=-2i(TWm+2Q2r4s3*nB5EyM!+r#)E6JLm&=TBw#m%01%Fu40kMm?FaVHn2}ON-Tl5XGtJiUevc<+ -rMaJ!+w^tY`#>!nvw19nuo}SSsr^)AFqr)t12JK~(7FM7b(+WG{>qC@)Y5HzeST$)UtN}UIl?4`aG5I -zD2Pg6w*u(C4LU;v0+#D|!;%c$07T=kLJw^9U6;&W%cg?d#K@L6D>m7V` -NXq1Egcq$^tz0??P>>fVM3gS{J(Ao?axG?ztDm!gq{om~VdFSt%O|`6juOdqDK+w>rjj+ee@2VG`zxs -b?ezl$O2GNt4gM$mSWhj2~wdlK#?b+c$T%hGq|1Y5WL;ES=Pc)Nrl#8J(?^>t&~MLo8KBJ<4%>XA8$+ -LMR_`Sfj*nnMvMu0*-EN>2;)@`B_D(0!1}vbItrDuX;guR2700F3}zxjl7IqdrxS?hJeHX+Se1oh)nq -!qrkbB7AF;LD%**!ngZ%6Oe6b{7l;DE*uJUXhws+eiOevdh?%uB -fE}L))+mn|Ed*H7y6PYEdCGR3L-X5gsjpeqrnaK4`+`%iti;co&OJNCPZ`1h~ReEa$vVrz ->>yic2SVv}Ht%fu}qpbSMJf-D|xi8laO(I}zkkVv1Yak0YjyyXJj8B-`#CTs$oLpc!IGsrf`0(JiNzD -jWt+FK|xT9dHy$s ->Po3S--ThGPpxeUxn7brd|~52+lO52=jQiZ8eHwu9=ygPjMh4}75GKIh}0SONVV+prnC)=oVZP-A$=w -}tq?*euJhs`j~5JGtfp2HDADWs_XlGOOtR;a5;TQcuyDGTMovXa}@9%=^cMzWDB2AnRp?Z_8?VXOi$5Y+`gKmOGq0m&~mx-Vr@D2RGNi==>9T8^m*}qb=m9* -`okbgmLMK?*iIr8~WXIZpv(ln!?tmnR|d%u2SJ(AcwC46XK8$t8Cv8i4!F=Yw!UcPnjuq(Bk6bAK&33 -|EG30N3(sfW4mFvfh_cv(+3;vAu1a{TSXhk?DqKdB~J#$U^@SzgeXb8T=Q-twXDeFm_HdB_Cmf!gC&` -_8LtjaR61^be06Rsu#R6<%JcD0XBP2}F}E -9&@NVE1O#k*$7<9Hhr^o1HBeIcYc(!yuTM9l2Fs!x{>;woQ?MBB&!edNk6nF+f{ba0XS4Y9m6Et0mdQbF&1E*Ug& -ht?r86Q&xm#(_lyW$t{ly -@c3ru-zZ6#A<(U<#@le{n}9_&ZzeH`bXs|P(Dz>CJ5+Y(6TXGs&O`rQHfI85L(n=rE*E5r2a&SKMDEy -DC~h5JrwJ`Bb?;+!^ZTb4xF4JE@5r(!kme>+08~XK|pcIpfIj4#-G<#<+TAMfrS#CF0je4wXV94GmO}RdJ=d~H9d~rbFQQQRM8zy -{97Y*aVkJMgqOdsvIrsSsy;ElC723+0Xnwi>z-kYW`J`aaqgm}T%0f0f7NL~V^dlwv-D$48vgQPO3p* -gcI(C|W0`DzN^vta1C?%Sgh-(p0pKb)UkTaMp2fh~IGd7yT5Q7tz!?QqsQ9$P}loFF$JKBP5{ap1_$v -UX!dpOZAj$R4QkO3$v{&x_d+1b;O5lrqzBL36vNFmA_RW;fuCRGB9I7mCZsTPKa1FMfu0f9cW+ -t;-JI4&1}Xh&E@a8&T*8H~q9yaOfgKl2|$N0PnAy#7QQqa=50VGR706Irjf7T^_a7> -hg+|e -0g8^mF0Un^P(DVqB|e(^iTCngC83PQ{g8Ex8IeY2W;-k+s~~dJkwx~Zp(AN4<1i+0@;w$U3#)-EXKh-!_2{F -KO!rUDfIg2@YXUr4@llg<2LFKGHs392?GIfQ&Bw1;j*W$4?v!b1ZtU;ARQlf4guM)XmQas^rtjA%xgK -zZY5C#bTJ`NR%s--1+!uc&pUj!PcIS&%%hvdUV9=bCz>itS=S6xKE+nK7i$k-Sl%H@0rhKrGZYZ)}@4 -@uhQcqyTq1zheiXZc#8YX$+UCx@Me0iCogxxg5dvAPYjL5|FjiBZW3~k&ZtYki_)cPhv)8369v_8_%S -Rif4L=5|_ouU6xb@pwpaj4nOyD0mBzc87T1ibg7@6aWAK2ms3fSzE!Wt;^d1008d-0015U003}la4%nWWo~3|axZd -ab8l>RWo&6;FHA{8MNU&iRglY$+aMH%_dLbXtXj#9+OE2&suCN=NElMU6V1|zNkWwv%f>TGpS}lAr)f -7}@PGL@c^Gb^Kb<4IJG_U_A3uG`!2X&)48A`>ex7mtI*)BXPd~@+r)%pkt3$YcBV_BwtDhzxdWZq$$z -^cBtPHN9^GDpG*CLAwvnjvW5Jq7Ts@+)}ZQQ)xLfWXEjv=o=zN@mOo13)TZk4#!HRSNlsY0W@zx)uydr3sZ!V74shRAh*&S#C4*zkQ)TSJJQ_4Q8PR3R= -@mLeP+`e~x5VQEL=7YJ8PxgiZ9NRy7D^5->D|VF9%?S#fR%YckkX4AFEVrjpRyawU@*ORc7fK(dYC8z -ys>jA~I^rBT@6R2M@ARCALWRb$g|)DnmplD#G_<4!X`ilv`ve0)up)bjuFugGsuO9KQH000080LuVbT -Q=0RVblNs038AV03rYY0B~t=FJE?LZe(wAFLGsbZ)|pDY-wUIUtei%X>?y-E^v8mkUa~+Fcd}i{EE;? -K?=G%Xm#i&#Z4rn?JES*y!u}B_Z#hZHCyhvgqx$3wpJ;*mp389i4Hk?(uKbezDx-eZRU9fjP^C=N6nB -ygXG67bp<*$?7{g&yV8`}Lk}}}JPkn|7$@i|YYGqLNW=3_Z_u5IkbdK|ZjjmJ7Jk_k@dVJ!z+|dY;li -P(REX|EOoAXF+Qz$h8wnvQ!dk{nXMm-tKmU_UHC>Oo-w(a1lA+U189J-)q@~n$=wE#;`Y*3%Gj+>w6t -^Uuwn_Y*vKDVpO9KQH000080LuVbTX~PlW8?_{0Ma4=03!eZ0B~t=FJE?LZe(wAFLGsbZ)|pDY-wUIa -B^>UX=G(`b1ras?OI)r+qe~d_pc!4CHBB$U|(Id!5~etgCNOb6Le7oflbSlO^igkBsCK+`r~^pNr|Lv -Pc}i&eP{z@5y|A`{XF;ZYHOogwp!Jl>x@{fm~0Q)IL4LIjytK9olI(4c!7 -5U%a_{_j>MMy%tIsS%rr?Ck)>-B71V$gr~RsKmx>CvlTUb*0|MbYD -H7e@u1yam?a6Ev75iKjn>VQEf$N(WD+FaZeS6w9K=1ytuWK*sZV54u_2)@pykgQVCWEMtdx@x8)>#+;!+PE3Oe}T__P7`1O6DGQev}>d+r93ZE!*=4?+j^ -}J%}nCcHjmn(+abLY#hv!daMmu!}uU6x8y-`wqDm#ZC7#ObhchEatEY+Tw1~2X?NEi8oym^UmBxLa`; -ZT)gk3jQl9)jJS8_RDY-Mn5X8<`-?A^1Pj7sld`_zd(?*N>I{Spx*>agGqv;A#>s|r?fPJDH2Czrn+{6)w*Wu6vQ+s9Q72uM*;QOwq)KrI_50*WFaYH6&yT6oEgeiS_>dR%(oO-HB9{mYhg45tH02T}gCozdNag3HaW^XP+Znc_>}3~ -!6_y)8hG+juCs;|O3G0GL3Wd0PievqONJZrHQz<@^Np-&s2%&hS9UF@VWje897I|ff?Fhu|C8uvUC@ -w0x!+lGIkeJ2?AoP9umD(W-vgY5wxzTDvUaLkb&O;nrv(DpspKWEcBw$}ynT8qAER}?a{F`_PpdI(lAIXPqo2F>R;HKKsr7qtXE*ofTH&}6_6`h6fZ0!ZAcS|VO#5?e1MtB>A(rM+Qk -Y~e~fnq|8R<`)B#0t;05Fnmq_*3lj8(WgJtAInei?23kMt(~_M9qeNao*==miei_^`p1}{tH?Sw{8R* -&Y~Hr-DZ*XuzM@NJn-10ST1<-lS -GOv(=a&XCYFaDY;PSB%MN2>O9JHU4t=Dl+Jo9<47rJAqt3kR2uTX@|?al1NOtPT?luY;pc~yW**V)7J -?UTH4^EH~B2|e`8{r+rB&-0icpT(01PRObK}$oE}J$xReNf&{VXNFmPhq&Jcr0E{+Hn}wrP -6fr)N(&beTp^OX8V>gn;iO5n7C*&%T614`8%g=Z#(N!FI8IIi1y%)%yt_q+zC$2 -)A$7)kAsk>j?(wVkho#C?{qJ`gu6~vd#-kXp)i3ee@GFl84{`Dhp`HmuMpm1*hJ3Moy(ed)AWUM(ZDewO1% -nMBX^M=_D2MO?z7659d*%G{w@mSP5_v7g81E9*8TTBRD$)BsrS=%=0RKzY@6aWAK2ms3fSz7=A00062000000018V003}la4%nWWo~3|axZdab8l>R -Wo&6;FK~G-ba`-PWCH+DO9KQH000080LuVbTOD4#64L?z0QCm|03!eZ0B~t=FJE?LZe(wAFLGsbZ)|p -DY-wUIa%FIDa&%>Kb1rasrB%yr+b|Hk>nj#HIe_91(4mLAK^g=|i_|?7L159;Dq%~LDoG`E(2wsfsi& -JXZDZJ$Es`_C*;%S~b_bhH-G*1SQkJTV}l6F8lPn~&k6fRncIGR{pI87X7&Dbb#Z=qmczuP=Y{2r9tql8JLJn1s^%^1~WrLZAd4_{(cR1iDQ#Bj^ -%7dxSD=pd8R>>E5>RAWYyur20dxjL_;a=5+aA7Io&l#*nABd%?Z$;EM5K9jICgR3pAskNeGcQMRNMVF%W=!?mLef{()+=cSyCIY;ig#K*e!iW% -X*MwR1dPS4u{vL&1U23dB3%M7alO$Ss75`D4;_NDpY=!(59DR|3XU+M_8)kF*a(9g&`$tDHK1KTN=VL -gyF<$mzprNK1b5*sDZ_Q-R)X}rmjHkmtat>WM33BSb#s)H9Q6~+iGwm->O9KQH000080LuVbTku8q;a -LX&02CPj03iSX0B~t=FJE?LZe(wAFLGsbZ)|pDY-wUIb98cbV{~&aaCxm*TW{Mo6n@vQ;Jg`!tpu`hT$mEQLIK5J(6-_WclwqhZl*I<L(UIAZu&{PpqediML<+2zI6`>}pCORFl -qt-w6JhZ;ngnTHFM*)8cO#f5qLknV`RdAK6c3leRfui@_o>dZbZ4mn%ygb|JGUBNRU%jl~OH-0qeF$P)RV_50j$DyOfAKGl!{1W40W5IOFrkz&kRkCML2qECJ_ -{ey8_aSZJ@qfPVKa9IY_?BjcstP+q1We=pMACUzd`JcF?R!tQ7hT_T!{Dg}Ee)dsOOI~nA6AUpr$P+( -0d>r$=Pv4pBf*sC`>Iyzs|>&Y6!SCttdE`i*%Lc{P^ob>1rq=!`G((|X?Ks%nEDdm2?hr=5Zbd4+eVb -$Y8DEucjzEy~+Zm!CpK -uHzy>{w=ecjz<KMR+QkKV -?_LK{yiXXuW%pyY@9cDR6ZFVd#>sP^Q5zax;?xozv$6ndjvb(Bo)5StOa@A^JVzjFcwl5KJTk=*Mi@X -5d%~a(J9b)`w3NU;rBwrOh2YqlEMIuU+9pT+40Bv2q%ZP6m4QoKJC&s^HX%_91=cmbhjY-FYf1#fDs9 -FI2gexGS7PMXG=JcLfx%}=kf!?)ktIz?te0xm;pE*sDxYpgg0#5;F{66TM8y7DvrL|gT<0w#v#mTf585XF^ -u1S#)@_3DCWrt-dl0Lv^!8#5jtVV*){XCONm2R8}_<4ds3J8-K-3c8SGSBB75cYC{e9`$RQa@XMJNZ$ -u@{i9J&tYt8r-K1PzOE+=`PV`TCiA%LQdld~c?@b`(b`Z8vlY?CvhVHiJoHKXb#xwRyXs9K}CO0lMbI5IwNZjijvG3I2F+=T0wQy9A)zL#ln -5gS1J9xL~|c$pe8<;=C@|%}@=x`EuqsLDTm7*0Z+5y4B}(yW@Y0`B1`F#r^T<;Mknk%;ql-#2&MVI8m -9(4Yxf7h;=j#bNUC30G$9k(6xc{EA0*^)r^yxa%e=?l{)vz5*%ksKK^*p(c$W1~Bu#4$>xXs -BBrx6$CS-L`k?x=IC{oXEjEYAK`0e~*mZWArajO9KQH000080LuVbTd`2$d+`7O0DuAj04x9i0B~t=F -JE?LZe(wAFLGsbZ)|pDY-wUIa%FRGY<6XGb1z?CX>MtBUtcb8d2LcbZiFxlyyq1Wr&dDygB~in2d=B$ -stRR;ks^hJI;QIL>!mTIKyH~C+v765(tqX*W(0N#EdWb_gQ`YQqq5G~{AX!y-0!RvSwqe$Dq -Ar0Bl^Yc5Z-r!2SV2$0J~-w2Qe@MB?m^k2t6Db1%1mNA!zT{I^2t~UhrLKXf}6T{%0CFq4?8}p;z6*1 -x-YN?|n*XbAiv9zrAK~;aFsx=S+PB$0w{B2IyBA -Ja9qeo;(>j-ao9l4y?Kcd*{kYPda|KQuF`UOx+0|XQR000O8%K%whXZ0Csr~&{02L=EDEC2uiaA|NaU -v_0~WN&gWa%FRGY<6XAX<{#OWpi(Ac4cyNFJWSHbaG*1bS`jtbyU%6<1i3>_g5@_N{i1QVCX@w7jmUg -O3BMnf}_~$M8uYSD=EFeef+H~%W+z2Fj&&=?9A@S4e1WbvT@cCmLtFYJ -jG<)u)+Q*dj@0Vry@N4<-^!U8W;9i|GSfN6aRR)h1N$gt8;8}E7uU+3_5>WEoSaOv$5$Iu^!4to){Qi -{0C9z(S=nu@Kxh_*Az|RQNlRn8BX#%?6F|w%a%XlykCiEe*;2*-^>$o`!tVr?(Yo6G}jbw@6aWAK2ms3fSzEFR#o@yO008I)001oj003}la4%nWWo~3 -|axZdab8l>RWo&6;FLGsbZ)|pDa&s?Za%psBa%pdFE^v8$R7;PWFc7};E2f;1h~)<$Z7*q6?WKoRZmS -h>VAxD!Y-*d$M%(`Pj*SE4p+<@XX6Ad$!;R4`lx5Sq-e6e*wVgH&gi_iGM_Sn=X?%=3b)??T-`847to --*3exO1_RR&KEjo3+yPfnYdo9EVLY|M-C55%HHpD=DvWOfnK54yqT -^0%OFHO^D>JIOs -Hu%Lct=O5oEEx7y`T2s*&U^Y+*r)5s8~$35!53fXSmc*@1T%XED!x>9HsxaGcIz8Tf&eL&c@>J@A@8S -exP!o_Md?-ud0jVYT|EJ0vS$4q42NpAl?1Bj&qaG4gc1e$K -@(@?^qCl(Vmsne!dbJ$o&#!55UzuQxB0?#WwTF%3rUkgPZGC-nC&SM>iMntxDB0|XQR000O8%K%wh?r -F~OP67Y`=mr1)E&u=kaA|NaUv_0~WN&gWa%FRGY<6XAX<{#OWpi(Ac4cyNFJ*XRWpH$9Z*FrgaCx;<+ -iKfD5PjEIOx+g=RDFOUg`k+!6hbI|DWM2mM)8LAt~#>{H6j1rSy?wbaoSQMuq|10&di)MJ2lxIWLe$1 -mN3ge?HWrCa$~H^owA05p!P9t-zalGyx)AkU1jN)Rr+;(a~(r!wZ_ti$8e1XNtHb;N!U?aIv$YeCWde -RgFjnExI^RCrna=}vIT*vkI9z-$aUZGovm7pNzlPpQ&=UbA7^m~KHl0^R~==6u+2@SB}SPeF1R%Nj;?5nLVIBwk9 -AuF>N^;Sgv1f%PDpZwToCuk(8#RqQB6**skJHv5WY}jwCB!ykZykTqEuMR6ey!ymW7PEj$ssgImtYt; -L~Tgu?7!le{tI&g;6rHjQD7Z5#pJN7gLU4R)q{qUqqKi9;YmXGN%R>^M0Z18(zZG^CCISQaD07sbW|p -4_29wu%VEm&UNyx9M_ci-;5^qFpA=lDjM}(8%fHForI)fv8u$N|C4)QS$0R{P4(YWeW009oEk8Mcmsy -}ckt+z?4zn=Y^Q`zALtn>iwv?}faS<@y+7;PakcULI@)>Hdec@!mR@?xv?2S}8dkTI7P)h>@6aWAK2ms3fSz9lsaj@eT005L;001rk003}la4%nWWo~3|axZdab -8l>RWo&6;FLGsbZ)|pDa&s?oWpi(Ab#!TOZZ2?nwLIH)+eVV_`idUu!DJ4FwP$u-c$PWq$c{(rP2%Ht -XU5^tp+KN1A_M^}E}}VJ|NB-|ccZTWDJQ{$O#ywWuBxtk_ogoQY_-~SZCCTvie>w%tXmcrMcKw}Ru;| -0g?yKkdCn8@PQJ=oUbkhL_bHuWkXBc%AdVXZ&Ei0bqWJD}+Oi!2h>{&U;?8x5cKc_q -0m!?Ppjr7eCismZvLR-|CSDoJu(3H2@kXt%CEV{QEs`%KVYnGnk?2@>VuaY8(S*Q0;Hj3)z5wN_a&`w -O4*yB=9b4S^)GWF48RRVKbJ{+?2&8&kz;r8-UT_c(*WG(`0}V_sfT}e8qWXyau>mD>!he7um)p*PPlC -#0DDxGLn8>{p(pziW!rrztG*IsH)4yEai3MqVS#kFk|m(zTq`55^i>^IW{*seit|Vy?l9bvHI!Vhua@ -+fBtZLv-zOKUFc(`2f9MR6*JWbBIl3_O(;JZEvk>auKE$}3)^K#7bu&sY@$Q -QAmM3jA5nm~+!bYwp9&{lSVXO+cUChY7T8;Ud|lU -N?T_KDN3N+FUzEMD%37KJCXq_8(>n9=d!;1xiTuv%qB)~;3>> -kZF08c|BB&OLEzVnVl9!gOv)QNJyTKiti8YvO2C0SBu|oliARc-bI@!{Pu8O7Sl(UqeXiI~ -pp*8{HcQw#PIkX`dy==0F)#sWn8q2t?3=HQ=BpYB1Y?h?2p7`$k)H%tNIWHr2Vo(OKsi=`umnZm^rMd -d3CINNsXg{2r^zQc7!kpnUQ@)AGIFDuIA31B&>}{Pms=3+pGd$UpL1u7IBPh2*|v4I1_4aXCYLnWAm`cWD{t~O~&hrw^uXJRo>+u7rkt0=0fR`p`_T4oNS1?!eNTW` -j6kM?90}ERFHJSc0S^h-FgyqMLoB61cM0siyn~;@_=uCOUi-6#-QOFt+}~@DE#qJ+2%{a0I<)3-+WUJ -6!~H#4KgyO>286+s24HjO6YmXleSoIwFEmTK3JD##9eTvw_M~gbQ#x!jSO#>g2M~qJ))+D{>++kLB47 -ww&a3Z?54+(O*Fnd-;Vo!1wOHPlfJ^TUtwJw6_#Htvkz;${CprU)d}sqRM*@lH0a_Nl`v5GSIpGg8Rg -SnMO>nrW_x8?5*@$uAC=5defLdBreI#unHhK~7`5e3l+O0ayrb%wO4prm~GGs5{u&$+`O!HL9`=m}b#(JW9*0Z>Jc|AldhyC?F}d??rLmnGqyW15TSlKdO&n_{YD*#4S8gGESF -%apXrjqw%AuK4v#$AqdO$IC?InNrph9WD#>@cIqjYK(73{>GV$84%;to?gyW6ovSj~)7MS{Q_4>;dIg%Q7`64~irs8^vXfDoLw>vlYGbHXqcl$14tDgB9m;ywL -?RT_Y|AcBouT6V5Ff#K!ys^1p!rsyOa+rEtCMaILK~$QKeqxMC(w1 -(w#_9`mx?HDVa(s*;AlAh6+S=sa$~fX|9(Xc;Q5Se4a;Y%Q`+2Ul%rS9B58TI^LzS!{>gU!ud&i! -qQ~K++jfc#A-rJJN6>FiVp1I5oxrbj=uMQ2q-`7tZI1_anc3@>WnCL7|T<#gEY -SzvXzImeeA+K^t4GqZ?ex@u6L?tYK_RAz^ku_Onx1-PYzb%gIV2 -Eamv8LY}sVkMOLK_Zc8BdP|z;K)E9^diZD9Hn@eu9Tjs-+ae{&>edk*8A3=c -S%`cQXf5>F>RM-q5HV-QWGmqay0*-BMdmwAq29!XV(Gq;;K%bUmvUwIi<6@UqOAZD{gAeaY -C9!T~EEf_Fz0&ryk!+AJ5qNL{Sf7wbAaHSgh6R&W!a+tjl&IlzXq}BhiNnEX^9+j05ztD0Nnj{dx3!l -(a#TDr$PIPtso}&_zj?gs6A34Yw|wedv9>fLVaqYF=Yg3n$uI`p1?R=qHf<&rs9P$N~WHa2GT%@s5+# -H=o<4*HRi8zp%UD`;P$bN|^_V3LrFev>};=h!wkJ -63=>a;dXlR{zN0~z?I7=6hZd&nL+x~es|vJee>Ib+`udT+ov&DyOsk{QimFJC~@7osivJmB)-CB4A+w -F!96=LnuP8kxKTSNrT?+hXMfFcd~6klHuIG6&5MAVre6PQ7uIDJ&Oiaua(c`}|@tM*sRF5gpgQUjrLu -SQ~r>k%QA?9tqcuR6q^X0Rj&E8T<=QOHO_b;FHy~R50{4ME?#Dc@--kpHaa`3T^=Vr{M>LIB_VlEZhs -)J0G9v`TiP5>~lYS3#IRf05AhDcv+a2iXaP}!fIf&`*c&|Y3MirHUdy -@NMIcLw1tvL<3&bYy+JHxA>2`z3&@eWvza^D|yeIRouOp9=EiA)U-jhUSWhm}OfeL{b%!3;l8{Bu~70 -=xMW)RdFWE#xinqb4mLPhi0uOFG}(22LuFa2PLZ3gN!=OT`&l--nlyI_B@mEjeW*;1)OS}wAd;DSFqk -N^u~9G{fue)WQWn+$?<0m8y->7P8e!tRD?c-(JTRV}L=fnWT=6y9h0k3fDVe -!h6F7J=Y;fRfQCWRD2T*9}#HL@GPux%$N)&P6LBTFL}I`7JPkvtY6AQ+{)uE*Avm<%V?Kj~VuXH|*e*=979k -4%^Biv~x(-0YM2UsGL9?USkGwy?OMb-~U**aHoSZTzQEr|v@@Z$zcMS5>3NAeU;KOt@eYMh0$@d114i7hWmdEkkDvqe8{9eH$HLK^qt74`H#D)Oa>%&H`GkEM7$zic -Dfqsx+9R2uu!kx;)(#j8(VT&0Q6!uY>TEM|XA>edB@31mij_mDG%wWv#Tfy -#E*}es`O8rWPnmsL_S^1m8q8KXmmH$&OJ#yu{UQE0QvQV*kzCXTq~b$E3CD=bfc)6b -yqgR2>=AQwgWN4-?Q#bZSL-2e|#tvem9&^rKXmm|=Cj#0%ZqG$4Rb;_6nX8%%*de8NQ2zT~x)< -8#7e=B@%^_lc`aTQj3$GGBRcc77V8h{IMBqoloMhN<*hrPMto{QNBQcPs)_8K>JCaA&GO4lIc-FlwKM -Z|%isV;1#4SmAOf?l<3@PK#$dAKfTn)TYAj1(m)PwRFq!P)2d>IW#K}pA4LvPj4 -E=X?hK(p64!`%!4Xa-4%vI90AfpjO{n|A7|K&& -e*;MfcKdMFXiFhsFe>U!(1XmlvSfK_APJvd7>%YW27MkCL-Kju0&WPQ;k?!P1i^!o8c(MNu{Z(4Xhl0 -*X4_x9S#&R4IcFt1&MPJq8JLi|JWx;x(*tpybUB&l8F09;% -BA+Qg|tWEhPNSeA;P-}z=8xG#}?kQU1w1Entn6(ZOGp&Vz31d&)#Kyc{?C)83v_V;bRpzkT1l&fdaLgBC4S#aKbF!0RA4jV&5709_3>hj#-~vkyoa$s%w$!K_rEOyM-noZSI+oX -Hns>tUmyNS3|gWL8wQckigIcosf`Da3f=&i(n`epTX>-8}93il+;jehu#%xhOu%3t>T?PVEZG`XM`Tk9{13L#czLo1WOpAbTvJbgY2=O@Uf>gmcCHgJjO!8oxSWf -F1iDV&eXiIm9;a9t(olfrNNJsjy~!RyTUGvMlj}D~bb@Qbp{5a_&k?p)eeqIk-{;hnU7b*l5i5-zF7;?q%`R(qQqz&FChCLmFCDfX-!oSgj!^h8BO1?h(N& -~O28wVXg^9frIpDXrGYAdW#lw7;%^2jd|!>YKluY&Xr_vnQU2MLYEOJ8y*q&8!|lc|#l&fbw1K$?L(` -zAHmyc73g%kYJ2sH|nM6Gp(qLqyAbmETLQV_g?~!dv{U#$qPDSbFBEBWe){-?MqDC -*5K9|S2bA|7>(pP`Zi9!10Z?v{JC*K(a`OW3I|wo7aJF2QkDmNMZg~n2ex2Sf%Q>#t($okcUD<urm+zh_QF!V4R_ax_?EKOS|bgHXV$Ye8YEbQIO4RksGyhKUNx#;A1thTTi&_I}P@(%E -EOluFnbj%_6BXGu<=s5D$`i;eNXrF9s2f>%r`|3MbWsJEPD -{Q3uB6CQImQWNbhsLiU~snJ0w1;Q{)B6sr2R7MPU#HSBREQ7kGE`1iudxo5Guq%Tydq$Mn&MWvyGw=P -lfxn%tzWJ{|i(M3I6+Xqn?9a7`mi~Dk%Jw>~>1Z@c|z&t8=Lm5XWaBl4HnmxqAr0kF5c8xeNt_$vCaR -eN;k~NH;TO2AlmL0i?ZGDvI8gK?tJh4XGC#iYdgivy?)do^B?TzfoKHzNDj?NV=j-H{Ccja~!@+jp*SIu`H7FGCh>1}_wzT{vfR -v|+0IxXbLlQ~RRdekpEyeP$4pBnVd}g;b)5KDn3wic67(E&%od0w*s19X&J~y1=*^#)5@V{}$sAqU4| -fCeW61b^dkBPnp`H!T$qLO9KQH000080LuVbTmP2wN^1fD03ZYa02lxO0B~t=FJE?LZe(wAFLG&PXfI -4jLq$$gMO{%#kDD+MzVjX|(^oGbH<}goHdZ-)kmum$%WLx& -RN&BfLDnJj2fRJ><=LI@LoIm1aC!@2wladg#n(4sUQ8>misnhT~`qxFa-OJ)TSq!9hK|!NrUo0qziLJ -J{g_H8dy`;biEL#=B#<)}z7U7HaQZV{7Dt)-}`F451FBceK6n@DMr!i|US8JkBDmsr$$dK)v@CTy5yw -6ks$yjJ6?;80^sWQ%fG++r2&8y8^-I^gcq;$zuZY8N=DN_DFv-$GS|1-ufy+^8+@{*$b{a<(MsroI9E~ugOVvNl5NHmlv`EGilcmrw1p^E3RQT{3;l$uSOflw52!XQ -%Sb9>TXe6;A0)EeE3w{a*vKs9SX^>+#+DhsNnxlYVF~D4c{(;BlO-@{DTlu&NAex2uu;*ho7Q@6aW -AK2ms3fSzBR)sl@;Y007Dt000~S003}la4%nWWo~3|axZdeV`wj5UukY>bYEXCaCxm*U2o$y7Jc`xAU -qFi2O6U-`edvMWYc6OEt&)~Srm(Opwbd$aYhzZl8VRsu)lrJC8;l4_Uyw3h$WGihvdEIevtF|{DVvjX -2n;_c$Kk@;Ek1eQ?gv?9k*6iTP7=J3jra$lNB#l^Z9%>le=1J%aobrTJ4xU)K1j5-akg{`Cm`B$<6!Q -n_urA-bL)XIz;S-mnGknB4UrVl?q@H`)7rhnMYYAScs?$T3V=5@hp+`!4|3-^s+asQh)I^yZiCG?`N| -l0dkVCn7z!m!X~Mzj4H)EVskCy!Su4;O0c{=_FQK-ITlp%On`YLoZQbeMWgc+mpRj=kRRPfIgP%9K|Ly7V!*g=;`_Jf9W>s@rto -X*zbAfMcH{EB8<#HyzrlPj&PL|?9**n!#*{#+}d!RSae#Hu2*8s$f{T{^Ex|FH3<$-OX@`^#3tC9zZ^ujpNwug;o;v -@vY%gsAy()<6pt@^9@7C?+W(%5Z&7PEzp2ZB!%}!L-Q0#Csd0at_-=M;(M-RJcC-_wy_DW~}2lUpQ&N -{WMR|yD`DZLtwnY30V_7lc(dg#@^oE;f;fwJu%_u)F49a2GEYZse8Pd`7SeCU;Fn)N5v}C -jVq}--i8kZGTjuF_)uYbN~1PiBD~)2{L74iZ$CPRigclmj)e2T+&eB6D^7P#VpK{nWl!r!Bt-+XY~XZ -yz0m9O%Vs7NM;a_DWr|uB1D3q%qnNvO(RSD7J0|1z+>FO>^lV+p1qQfW-J$O5Nu$CaO1uaZ(Q#rr=d) -?fLbD^f}TLVkD}(9Tg`N8->2r -5%6cN7Y$9y#7cM{~7od!a@Wdc)l-bg(${VazebOgBa4%PY+(WZ{^ACU=S{4a-&=js9Pee^c)A -kzNGd#?N=^&WmeDn~hk2`N4}7a$A)6x1AS0>n1>$FS1U)h=V!yP|q|2CrG$SrJU+d -3z2?FQmNCXx`8LHGSI^MybVjh#JqztPAsIJG4qOqZ4_uG~LL#u_eXVU4@-7_bsGj@XRfnqjKk{u2$NituPMQVhY9gdSYeWw)i+QbZy`FnvH -g>IgHSGvruuI9_xe*sWS0|XQR000O8%K%wh(rsf?i3k;@uHKV!b3m;3x -?flL!-D3=znLNCC+~PJRp%EXdm8FXXGLzRqQBC$T&`T~+6tK6NhNIaR~GFo+Uj2W}9>qqxDz5mx>r!j(Hzo5ST3jB6oYw$Y_Cvr6VHuC!**#LD_=ex?5#Bj{yD5qEK6`$nIkP -rmn-65%-p^~vW2zMw}Eo%+JuuOUz42pY0ax^g}6Q<3U>BMMEwENmkhGfELU@iy^CMJu5N64HL=b9O%xRAhhLzke0|)t?Mt*foZyD3!1Kk=zF^}#rD`9bB -*)(!=DOV<*lBj9gb1WegVsl_4$&e=4lm~f9M8S(?KlOSnB0-<@8HEBYHo+=IsGRO7spcT!jvi9S;r^f -l#qek`XTUm}QOcSngPHI|vT2tEpyyI6B25TER=cU^>)Nww9ai%LYd4~ylxL7DX`Jx>sH5311dFgPVXL -F{0(dK8j@Fzb@m#S%7*5MvDIIUsc2s&2)u!|Fz-LA=!}dbzJDq2#8&_uIVX$ir8HGJ(E=hAtEEHy{%r -_@hlKMx%ZmE)93sLRco9~h>x${B9nBP5lqQ|qMyS3Qe_O*HBJvU|D(~nB+(E+SYyTS!3L-p1v;DN_o7 -shNJ%{^kS8|-$^A)*Le=5ZfeKAOcIQNMdOO#%bqo#YYWzLS1fm)z{WyHpD_IGEe|Zw|G(ex)MJT)EoLVIatL*J$YU -FMO_<9Qll&s(@hd44^$x^o45)KROhb-$qi_NcIj}>tCmfyeb3|fkPNW0i -YQ7Qb0_psdY5Wz1#bof -$}nn5;*i`j#srao2zR<-KO%sE5Lz80Dwde=q%nyJm11oNUPvPz2fkjGz1G@Pb6+mDCl@E{Xkpi)Y)n^ -yNsuoKN?u6Hpso8UuURd4RTivH>nO7J(PYBwm3`OtS<+{HhGaV$5YNV(`rE#Xo;O8h*!JTz6pecoBG> -vZ9a!j(Eqe&SUnAuEFF*P2n|JTM_vO9o`~SN0#l0K9{_*bhdmns)hwt9G{=>(g;X(YrdHct^``dmg;0 -lGUAYgp(+4$P6@xcx8aPawf|4WGZ?!$k4|M{QBukQ;< -7;1!4?Y?n+!e$9561hSj`!af@BeeW|HtY(;~-6j)(*l;sDA&WugKv?cky+2=lbEtw+}zPefZXUPS -|NHAN?%(*z0L9(IzumzjIsEHy$%FTAss-}km%q9H)~yF`-8uZ^F1i2b*YAJ$y9e)ocKG(2hi~7i!?Q{ --i|tV_lu=bNc)bdX{z+s@W_8i6#auFN@oukY2M+22@1Af`Jdj~sC3$PH*Olj$A`P>;Un@isLdvS*1y} -enqE*~_9BH#g!%(!gYZQGdS?e(xi7z@xaTH}%4@PsuM2gC&&BY<@qSOnaTi<$Sk8CurH=Em7Rnt(mj1 -Otvpo*o+cxOxVk(^LSL+#eOC5^~ie9VpYiB-7b?`a<@NMhus=$uk8r6_PvZqYxHr&Cz2ny;ByL06fbEK0>fttWpJb5wN6!x*g^d7x^VD33QWqB@P-j<%?sdBQWi$r=HrDhjxyJ$Z}yi{VGLfBlgx4RiL*vyR!8t$0>xtaDi)(& -2G0i)k}heN!&*AWCM~2alAqJwn$r>PDMdYiBQ@vC{y&al~QKn0V-O;FQe9+=wl2~2FGOUNXb72 -F-??nIP9hF!b2ED5OEh)K*(74zsU08-f)`yJvOFG<^C*FV?=ycE)#fnmOPXDpOsrS>;Yz*ID8icCP+Q -MTrnWA6&y3@>qmR7uPIucM0&HCYxHw^L`#;(Cm6gOd_r+vIol28kQ&Gv%MaZMFhNEN{_{+!((QjDlnk -mjMcR2bGKd}a>u$L;1_3f(5ZoT5GhcMO80{P8+rOQJ2Z#SMO@8}yFc~|A7_D~>y0;E{TITwg!sq1gcI -@V~MAm?SL5##A;8gpBOvPTnPnl-YpAfRkv!5|Cj(2u8=g&kfN1WN6nA3X2roNZ7>kG2!LD -1Wsw0e0p9kgy#pGS*w`$DQ~|$!7HM#C-|AD3Sr+NQXBLj5+_x+QIC%n@O9vj>bU(nGud+x%%Iu#`sL$ -NumlH9_@mT#R)XA@A%MJHA8WJ9u8wo;wu>OKxi7q(X7MHE+PHUz1x06v`H+C4z3=FHT@X2;-XMNgXra=X;&Q)Pvxq6SF -X}Y)`*zj%R17*w^HhojMmIVo`>+{2(D>^3@4zzUGO`OCo+saFnl_+GESC -Nh-ZQThgq#A_ip?!%fM=)$mrYI53_MLPkbGN|=8o7?Cs<$tX^d$K);S)IIDamr-=R;)TG}h{U0IweUU ->zuNYtESw**nk`e!8k`nvozqNAN>ZV0o;A7JJ05*)w}Y)SB`@#NU}A5cpJ1QY-O00 -;of09jjww)mB53IG7%CjbB-0001RX>c!Jc4cm4Z*nhkX=7+FUt?u#Y+rY2WOQhAE^v8WnoX;xIT6M8{ -V6WE8KIS=@>0oS7NUZ}>=ayxG%eHUpfk$YBJ=I7bLzQ$b29W)kzexUJ(a5b@6TR-_219`dimR@A3pu~ -@$;u|U;g^j%l9Au@#(|oPd|SD-47rC`Qd-xef#-(1$|No+X0p{;t>e#}(1%vqVozM0DW%)j;U< -(iM?=ZcrOq)7aGJo`38kS~(WTks^Vt_5lm^toPgTT#(uVau -BP>Z&wG)Qy){6crOtj*o-o>tex1iBl8vw{bV3>XsF)?n12znE$gZ=2)l7Ln9IKGh`Dyr^@`Oe>1)_(m`GaVk?DkegxZ%sWm{$^&wV^&A<`k@?imG3t=AM!vA_u3K?c)Vb-C(<7+idO9?r5Fz3v8uCvXxo{;`5MNvMxZY0F^Ryu;Pl!5)uIAusQl1b~H--#zaFT~E+ -=Qu2c|!DaYMm3FKp19$Jk(8(cH@oNTvyW<;g-%XmRw>48V_L8V;IRnc$IWQna|DT&;#4{K%HB=w7J2y -OPzaRL|bfKB;)f~=W~a|owER&Kwgerk4&%20M`(RspwViqeynbsStIVY_LcYR{7kp!Mcg@k0eXVt=)M -ShdgH~E==^Zci6@+u*1UlDYtgt*t8+fI>&~s3#kzEGq!Bnv9T3Ho|8Na7A|BH2tzT89PL@rM9KqVf6g -jhnzSAv&*g|~Gnz>G20JY0>|G(oKJ{$Q2_JP%aZaY=Zk@dT4v6`jz7*oaly7jc^(BWZ?pweu*Tmd7GI -L9vUT_b35yRpIa~{$J;=1=DThVzVI4AH88vGwOlL(>V(h0c)qBE{m(PMJE0?z -3Mrq{rwX}GxLD^xL~}qaS;_-Cp(}(7m-2wP&O!>@Scs5Ipmji#I(H+4JRo*m$^&9!r#zsuq*sNFuqj- -G+9eq3lqcMTI-<>f?z8hX))5h7suROqd6ZyFY!~*#TQOtdR0poOW}&nvHJhwglNBrYs5e|0{6S5{*zu -1#G4|)94$Ke->O7ti=EL+q>c9?nB9v?CU5K~I;NwJK@J5^ThGXMn2F#IH0^q|)wF{v2fwb;3zjpkcQFc)#-)e5{ -oI;JCtfC)S92(dF_1Q52`C~xI`L3%!W%sYSJxo}mwv@BJp-6W3EqjfVoWsDRX4qjR0po?Nvl^+D&{-P -w0c2})<0C&CT+4wo9f@R!qTm1Y3{O6R|nhV;F}z&NbjskJNDIgdK==A4Y -JcxIEcYfomDp6W$LnltD{*V0Q%tUd8ud9U@tS-R?Qd~gPR=)^L-hmxLj`goyqS4G+qH^o@uG4$q)G(3 -b*vJA7)Rz>I{cVeV*d>{2zdeHiZ66ZM5xW1rHqZDbZZ7B6y2jlTMYMD|xY%#H2*b=uY*yvTJI_fsuZ* -$P^I!g0#DpW0sjg{!z)0})7eAIzydLMP*Zp2+4o8~Imd{ENCRpAvu*@YLeY@bNeBM#Lsyc28P4gLm{_ -GDuRLx~F=*ARNL)#;nN!IyxN!S9X-k5EIs>rsgHoi9!j{qs;=F-*e@M@E~>J(x8 --)QM&9OV+Tj`U2d!&!|@E-|K&&VDb6#%K6q(@>2YQu~ah7uaEMVg*_DfKE1Pbt(DxbU4hlq?$ -3J+Xujg_i)Loh{f3o1_w)QBl#-EqihWF@=Euln#f%U#O}5?pQ6Ts}h(7D7)}Td&s75^4EW^TcSH%^;4p#jc0BOv3c$Hl+A0MH#G!3;^lMpK|5ZK&MSCbvCT& -O8?QXPgml!SR+_adgle`1&dD7)}TER{r@hB-+k*b>Wdtipknd*Wr*S?sfKD79*MQ?!5w?O;UfKD8GHohI3#Q9(@3N@u^cJfbcm- --pN1k0FCo-AD+z0kIeL`fMl9AR(l{hgOK@e~p(M2K754m5XK9=~DCu`M*sh3+n}%Ki_?vF2Ln3#?#4R -(XJSc?e8T=6$dS4mvn5g~E*qO2Ks?b@6&ZZfBA9>XhRNH!@>Z>X))T@?^{(Ld&mBHXYjONfVzww5rmg -p6rKN%XOuaUi4Lw|<`n3*AK8j*TsfDAqEjGY-?t_=R553PSDu&tJS45-@p<`2Bo@;iEz#q@H)iQh)YQjE56i7b;(&U-kJ2Qu7rg^_kI}<4CO -;o*CCOOE}B<&eHeQ216id`MClhyX#}#U(EGn>?>nERjn%{AHG8S)hnkh3B9kzJvW@MsZUY9_$MOf>v~ -@D{fr}jK0Y%>%?z~+0iJQt`H_tEnsGHwM~M3VVvi?7kHMld1g#IA0R746cLul5;%+%J7>eTwtm<5?mJ -fxY`<;&lW~pQU!Mc@B -{)FOUgaNGoEhV3{GRq_X?M{#M1b{VXjK^?hi=zv>25jl(EG~Bow&!<-JO3+75X}A0eWV*T^VUI@`sNz -gL4Gku5ozJ?)tdDm`_#LQQMxk!FD|#@b-*c!Jc4cm4Z*nhkX=7+FUuA7?YH43%Z)9b2E^v -9Az1wmeIg%*&zF#qC#%ye6Y}HoXJ=5K@X8X7gyJli1#P8i2MFPn_nzm{0Hmo-*4 -^3q@smF+82-EJqo^;RSG2SYy!(#)6i3y1vA3k?oC9O@&`I{*z>qCrRe&tElY#35Zj@=|{kW8xEy5%uC -E!Js1m4dM_(J|ajSga+(2z%oDSX8fxN`UydULGFA1>~;L4llL&vOO5nn?k7Q>3fTG1x^7MejFU&0<}& -~TCJF99+lZMmdk>(zg8`=+kak&=27IuBG~~iTIL|aro@R+38l1lxKm+tegXUS%@q^382?&+teINHVDi<~5$N9HYyU(mdRQ#xe6|S<+*O={E%BrZY~Wk_vtq7h|U*1}$P4S0Vh(DK>Roe=_o`!} -S)=6K-jQ9r65rX`~Z2&O2eyMcsJ;xLnn_(vJ890?Yf{5W~<{y7&dp;_fgA@lSk*01+1G2LL2|A>!7=) -Np}wR>#Q_CfP6F1jPu(W0qu9J1STlgOYUP)ccni#(z^Z_}7Z4o*4rehw-Qr;NE@&*eFh7EMd@@n6F{p -B~MJpd4HBm4h!OC6^&1Wbm*t}D@RD9XNVrY1vp>2y#&_1XL#0kli1I6_g@(xSc#tDg`}G&{%=!1Mlfx -Ex|1>3-e-nR%`-l%n4z|?-QPZD%gYqaXs&9N{riu$ulK;X!8Hf5VI8t_E~-zCNBw>DwSRN_} -~BEAAb0O{qVz|{;dD@2mQDI(0^lfyg&CfOYYKe`yTMw&RwU7DM`G<7K8h0d@kX+t(AnOd@2PPq?c4~g -@Sl&!1OZ6@+2M$Ta`zFwyMvumi77MH4XKK2KnKPkV8f5Utp!^`$^0Q1U+v3;BtW4-#~kW%)$$-gZ)HU -YXtd>IvaI|LV~a^zr$C& -Fef{sx8SJa-dkso4~UjWFZpRx&py#!cgj(jj7mpVb3`Z+>41%!zyGHn`%exR+>E8LSI^ZE9VudUr)9Q -k2Y;a}mg9mY|Q8)gjqfM5e}M*J)F6A!Phukf51<`K7|69m!%3wq{U+|+=KWxkVv)#452cG(}R&uo1+ -8d=IbzuRTw%g$+(W^rcTQ@kd>1s}+b5IbY7MRHrhRv;E1?O4~XB0=-Fr0Y)sNTEJHS8$%W_{ -s%PfW~X0yA^=_EU|5uWOOlj}wlUa+8pzCc)Pl!PrkZ;=tph(5rC&SV!jG9Pdj%P-b-HX&onthe0REm^ -o?hysu#<{9nmz6|b%%DeERIGgM=TKuTtLw@5KvuVYX6ZYwMW#rXCyVsgdUd|SgM$4Z-%%PawF{RSI=% -W=}vbN>xq{b&4ne`bhcT5qvKI>Vv~-Ue!hGXdXY`Px{&#cSJ4oW`@tDL%t?bta$V79P>)HDP9w3?}oH -Zs^4(^!4W99bVx1Oob4_JM18atm8-8D-fV5fZ}rDUxb!4&W15U)+QzlIw*4Y4%^{StnVeKWz;so$ln3 -Lo&JnPX~qKFtJi>#4RaJfe21-Vm|tOBN%CZXpnCp{xKiY`1?@c^Wsbsn$US|JM`gr4ih-YYwWsnP%U( -DKa{+>!V#p*lcz#E)UNEcuwfA^zI1KZ9huSmD>6-%hY( -j$^Ge**##95}eemDBXt_0l}5$Ub>DB#;kyt2Q*E958+d+{vp$`ODf-(dL+K?ZSX{D2P##&wGg5={9VL -%B)fOB;fo05pla#wQN}G%q}8Q>NNBjFR+sUN`iW*!DlwAzqJfRlSt&+C6Tsm_p>nnqcFF!C;t?*xv~3 -5$gr{=#upd-|aN^51m?xNOTx*?$UA<^!}xj^PR%KB6;aikJy~)I^NfifeCn^0sbg}9n1$jHWAm^905M -VXU6JuqyQWPpc)MOfM=iRUe^x|Gw}f{pg@>sKJqa?V091yg2dj4dYj#rWY2^WBf -)(R&Yf%V!q=35bUw6+0_S=SGeYIyq?uN&hSO`&6K4aR(qKgso1gAOdT2Bn>F`Sk^HvmcC@o&!HZ4}5U`<)0CWy9Z&LN-zscxsF -hN8uyPA}o$q7&^m^M9P0^#AXA0qV)kIg?ZQ5$vwlsa;t*8jTaa?7FxtatW!ANVT_lt_ZV?Dh{W+E$1J -15Xm$8ze$vnKB5YG^W;6M_w?MTjQa%CzH*Nk1>njYQ2>Zb>f;P5XeUT^o70 -<(bzQc@J@fBhPZ}?Y=A(Rfsl -4_gt^%CFUZ9rSM?ty@}!5M~uP|beMwI2JD+V;-4c& -etHjGHuaMccf8|naTDZx1#YQw*QQh6&^X`Sg4RIk&f1yLC=QH-=1CxkY}Cm^IOgAlu~au!`*ECgxM -ItR0>5rF9>z+ASK!)Z)E$<$InH-CpmD$wc?41ttpAE&$;9jXstemc*C8JF6NVn=t4!G1UlCFsq(g4O5 -DU2R8%ZTyeTbS^6#2qMl*~5Jt=KC;%r`mkP42s>fa*2~pL=#)ae&Cb8peq_Lf+lO2;v$vA0-`BUWRL{ -#T}ArpW#X@cWjWX&Jo+Z+>pao6w31^fU;_=?1QuE&X65&2uzKH)Y`44@p -jwbMF;4tJg6IggO(aMUTD_z2q-^g6|}zfNQS5cL}@p|4YH2MY8~oaqDRC6iNk7Ot<`-4LiYbtO+s -{+KhUgGM(_o_OYnEv2p)5>ccxMBhz-f0Q -SgcQwj}te6*4XQpW2ANq^E!GPi^43g9ax84MqdP#3Lr)+59sYt^N1VArUqKIHv^#O|1~m?NyNELfau2t669IGKOL?7 -ysoV7j4***DuwE*THcO`o!zBLUd)x-s3W(_ih>n#3Z11sGEk-FezL+;Z4I(kaYGBo;nFh -k~2h7*bi}gSB~sD7zQRM-`)x|as3+Q@{-dMXbe-jvbl-d2 -uY?dX$&}i6H#~%5XdeYT9D46NNjJ!zBEde|S6wg?CUe}k)%=A})#6y3cOR}7_PoX`L;#fUIZP~Er^Sy -&Vqqd6r`d}v%j}f~;(p7ftg$e41z7*ujM-zgE7|*~sH8 -G-~Q$|JB&Gs7W2hS@A1;_{>Zo+tkan!ps^n-+9JN~Y;(dRqY|LbJMhD$H>@5DT9MNY4qvx -O~3|7kg(;FTjRmaMs8GTgA0^4ityTT~rQ9yT}>^+=AF4*Jf&3oFpCgFtqkwwF$hyZlS*o3|eS#1Bcc+ -+>TEREpElgt;ZD>X{t_47V((Q^#iretb!vO=3=>sr5m-eZ2-VLa~pMRfbg2QV?e<4{r$-k^wwEjs#)E -FwvqIV)?p#hzwr>gr;caf&EV#PYP*;D9To9h!$8%z3ymVcRj#|^Slmb&c->fWUV;e}=RfAk^;=(s<~h -$dp9!v@2jO#SMIUPi`QRxA^Uuk35Io}eO7UbsT!D}z;IYOkW{JdT3Ah~^HYc{l+YSvI6T8E0hlWjw8= -`H8h7F0W!L~!gWCe!pcS1{#abucY>NCz&iE;lO4W+N`nr7J~pl{Tg7TKh5UEej0F-d3t;Hi_4B -thcTQQ9ezP0*7%Kg(3gfVH<4Fs)v1P1%_gv_8Bwg=dnu)b6b*J(AG=$x}BaNrUCbxBH%-araIKxE6mx -VAk*Q&7w36rv^Z6mEWUQ(M!YAgMdwOZC$-2I5n4IJ(?0!z8s{RzR6p4_VTjlMvI#PbHOj{ -zC)g^u&v_8`N^=>dS(0}%eRXjK40>L~4KkS2ySD&VTuPN9h1iHX!rfg&OU*@cJLfC&}n0_gD -04Q@3Kuhcy>`#_K%%9$VdW6<=~H0SZwtLZdAvt8=uYAa^Ch~YrJuS5XO21Sh)|)QBO^P`)_co9Mq+ks -4)^plyR)$ZKzLCtPZPDo{fZK2O-*6Z{09}MZi&ZbfFIMdciC=M(S90>pDV-^OS|+1Xx*><)0CB%x3~$ -21ZI{msb!lQ319CBn9+D6f#Gsz|`tH)Pw^=wf1rVGHy9iarN%9az(bk(3o?{q=-bRLtoGp?#@Q?>iY( -6l)vct8fr{3uP0t8`l%k_jh{?-5LMuJZ1A#U`d;5U^nA#~W5hYuQ};urR}qYRe!~Ao0f=h|n!vUCrFE -ly;+f9(34*4nH}>abrk5uekVUy#XJEXLzMd5#Q;$(MX5r)(wMF;L(Ayb;FVrh^~O#x?w^vx} -E(gg->Fh8JG!Q3%g02!hr80xd!7=u0p`8=%>bfW%aY-VYM8p7YH~`#oZq{hzGv{Bu~1-Qd4!zrW*|S0 -;w00T`(d+#MCQ_Wyp|IEW`Zw6>K)H<95ERWlUY$SVmvSCaK62B$aKa1`z&8FHhy~T!iH&c=(#pCOaa=EYd<d^XG4LA}GN(p@cUP2u8xls%}I0jq3i-B<- -^~_sf4=^}SP2hb5tn4m!rxeVai<23B>F=`xg*IcK4Kqei(Wd3IVao_ImqGh%7&JSv{nMw-)Fd4m)hJ8 -i(3^>36){`hLqPEihZX0|nSST90(OGi8a{4`PFby3&nb*}T*ZKzpN{a_QBOU!*1Yag8Zd=QbY*m`Qfqtd40 -^Uf%4P{U7C|2(xfXL0HJgI|K1mJ$QiWHweTC4+*x|}~+HsHb{1>!od$_Mb+0;YHGj}41bp`iEFj}2>5 -A#v^D9~+j0CbWO@)J;j!oYiQ!W1uX^VilNjlT^Bw=ewH-sBcNaG)#(jD&}r<`>z!+0{x0WX(oI$y#j5 -5y60W)ov7vE0QO4x7|x=n*HE)`7P3TLKR(z8Xj0s-NI@hAix?>8mDG -=bFg=*^;XpPji!u0?1I1MXl?Jh6&O=Vx~pp -yJuWNEkD`5JDnK?n;M2H}i(z9IYVgl&J-0<)PVXWPMn9gn%>R?Vj?wq}e?Hc*gi*ua1cBAJ+h+QM!MO -jS}FJwxK}&R!4RiqJreRmBjv;QyUG5t&LAF@JBKC55Pk|tb&Sp?0Y>G8e9JvFh-`ca0?&_6N9NzJ1s+ -2(IDm9M{$qP^j9AP!fbIi<0}t(AH*F@eB{YI-vq$?ywdI$MkD?N46(WlHwFERReb&vYQif&YUf+DM7) -OLKA(zWtRExrc&Ii|tLOtNZ|UxHXv>P{bitKNN+PMwl(2~svF0K5I7g2p%w&82XE9|6b0CKtI$O5o8628Q}5dkW1eefoi32YcES9^Qtw(HPFH4|M3VH>*My+Rn2} -B7z`W^@o^0P%hlzEof9*x9g~h-l|>4HcVjA`=jgFdI>bwR_r>qTL{_C`m~Ns;+b(38d))P6_;KQ?!gSgM(jYD;$xVeac;~~6b+eYntA8PjH&;-^x1Ie1hzOnV~oSC<0F4bAJjd$_ -T5Pn{%DH=|Co#9L>TWH^!v&4X{w?WMP# -p_@~1e|k-uw8Hj_yHmw1^=1NPnYXtMl+>S%i03~?$|+}!5<>TCPA{m#IejYmNS*IQm2z|v% -h_y}JFi@;na4@dW35r|xvmyAcy`}gZuLf=G;8624H&jeN+#rk%^bhVC9q)J5{+9rA~Bg$ON9j%)C&j> -sb+BEL7@7%0oaVzGJynYgoyeL*a`1U@5SiIy*z1ZmB{>wThaoBmCz#6h-B!aBUOW$87km6{dvcGNb)! -@7zhO|B*45-nuf<`q^jniGDkv43**szs&wEfw|rmaMD8Mj?*+zfO(`%?;^#5@Dc(C@4J0@K^5Ipk^<@ -_nFO0@3h{0(E&~?YY@n1oq%6o!)@Mr8*Oj#&uE+f+VbwOf85c|7jzC%SW7_v;A_n3|w+(D-o+y-12eC -sI9cz5^F2FG8{|wqG{A{)a8qVrJU1}%u>pj79{v?RDMW -H=xm<1Nppx+9`gw@`QgCTRVG5?}^_}m4TPbc4rQY0-sa0bRgdQW3j@p=5f#_W^;#-a*F8506lLIt<~e^5!9eln;1i;0D~mknaL7^%!^A+uBNmOz~l*K0(5PkEe`sM} -S`Fp0ywni&U;Z$T$&a>(&osen>H@2bhOr-OTa6$oB@ftm(V@fKmW3meeGx4s}dDTa~bx?|k=GSEpm?- -)0}41C9^?ilyIC@TBMP90<<0gP#^4hjxG;kGf~&8ysCYcRFBF)A)7%tS$7ouhrEq4JTDxP~YZ(MOXkk -3_*1xHNG9MnYV4&{v+`SDU@Dx)}Q49)cyN$$5f7#g<7V+nQAxv~<;gS4|R_(fA6|*Fzn0FBXoTHx9ax --0K01LRtT+k8a9Be|Te$8>S>s2O~?cX-~{Uj1vTixCh{Oi#vDKbVLUTZ7#w=@^iSYY``o*3oxEJ3%zk -(_V9cU1}+!(-eI}9UjyJ(<`F<6CaN&zvee@))WhOpKsb0rOk)ys9CvbQ`|-e>8IF46lioPeTq#(f#(X-Pq9hB*mmZp1UiYR2er7QR~ -%^vp;W1ZJp-uYf*a5rcbHgS%fMinQBFM*S+-!Mfq$t|*mrS-#x1MrUBZcr7ne*_8NN#bQRG#9dKKa5> -Lt^C)H+vQWl-_QMLjVGG1pM-Lmu2h?J7bK+~ZED4h`3!78_rAbl%AWkZ+toJ<&t>6^Y-=%jmGqGqAjY -pkr-37NAvxr$u2cWm8v75V{S!WkQViQNNlnd_ip21f_ncld@De8c(O7j2$;@rJX|S{7VFlBI3f!S1W= -w$Hz&H1o68{0%c~Fz~FPkfl7S#rVAaNkKU@`ae2vnJgB3$0j*R`JQg7E9PXQ4HyUY2^txe<3Ow2)yl$ -AI0@3AoUN`I!Mz^y+rSM5im@&C{xu_5M4__mIp1I3f0fr*gWumXxENvoyemJ!G%xoeM>_Zb*}lu0t3B3@6CbW@uRk&2 -a9!uX1XNEd*S86{D2A;$KSBtzlQMr494bECG{NF_PLbuJA0hEipQvR9>(l|6&pfB_WY^bdTk^yHJkM%*-?lpiU^1kmPOsk;7cc1 -#L>sM$W3@?d5RpGX)c*5Gx23#Dp_KbJGH!<9sjl)cyL!Lu+;bBldqS1S>1DIq%)XTk$jr3B#2fsHMkK -=Ki`6JcaL+!5(bwK7mE}F`?^wdX|-WY(6I_q8q-Qph0Jl}DYd67klMi0eUb`KDqI(c0EXf$H$Sb-#kc -TmDs7AP2L16Pdfpawc=Z&r+Kpa#Cf(ybWTKPf8v$4(t&Bmvm7I30^htrCXhn|UgyRk@wJk3o8YTzrL_ -(Q}z&+&*UFU7V!&gdwTe?GI!ngJ%e$qO&Wgk*~DHhM^~a%=zM89T*kelL#wF6XCk(&qiZu>INc3nR-? -Ki-Dlhxr2qw2ZPFmFtBZT)GLzr#1O?;2eq%op)^R}4xZjcOXeb0Hhlimm~1q7LEWPqw-&#Z$GlaiAluK@f>QE>$0TzW~Z@Erjqh|Bc=7Dr%j|3(HRhN&q9_OIN -w3t=*J0K$ZmcC$T-hDQ9D>1v5RyF>dE0gi6R1?kP2h-7YZpCGY%pBY$8R)C4KdoUrnIAw;_S3Hg0$3Q -hZ6=gJbGu-Ll@pTxNPNyrx&r%fTR3*ZQY=qjovghY>ecpm9>X4VFuY7S4?CcjONC{yk_ERkmXll-Dy; -&xp-6)TCc~p2I^#-pu#D-am=vfBLVJmGc`!^cNEJp91us6;L3O?RB|xq2uOIH8>Dgnxvw@!0GE<>-T -V5o2Oyx_v&xFB&|v>Rbkrd!I+h8%hzIv{s_|(F^ID&9Kp6Im`veXz6M4if?Yc@5m49xY`#lj5ss~IqI -%6LjQ;!@w9N%xKgrd#{R6!k%*4J5IIU{2g?`6_*SZI%s9mq~P#oDW-!e($C1~$W#H(+eWMlO)fX4MF9 -P3cgxvZrvGNj433i3C8HZk~4ith+3IG+HAKKR_SBH)PhLjx-l2eQVAH&k;j%y%Jzax0=mTO1|y!y;df -yv;ZlsD3Wpz-SY%2$V??LNROx8I?EIp|!6g{QW>hsrUD6aQYG69$w&UB=nQGJ-oozXk5?m_V5C4qqKk -Yw1Jc)Ix~oB?;18S7{6<{RR@SS8nZ?+_w(D&4YV*X5?4#>X|f>h)^5REV`?{g3{4}YtxNZyF&^vC-s) -{==3bzehJn7gu&j+rD06CsDs|v5CM3~s5O(hXL?sE?+`?_bMfQb@5(etP`@sSr7?bDl8UV(~Nm#!L?P -5}h8PN$M6jInlMbhW>uRVmK>!B{kdZ?{o7wm3I3C6Dx@`A4K -tldwRVJ>bSI}OBDR8-v=|=$Oy>Q~ypKiXw7+igP?SO6k3ZPkbX%gq&fTeJ>xc|fkr_I0M%r4a4Gdby)y=^zT9dX&Pp|TReWdbs0~RUP-H(5_q(4W5< -WDLn$;9V~(jnA>o=kaOr?+>2Om<#lK$4Pa&dVFcc_eSr$dVU(U-Xu-0i6n~?R{~|*dV8(T?M^e(gwXf -+*WY0%bVh2?TlUO1RIKhL-uo8p?4PCSoCGQ1-G#0tO{LFI>bEJpwI=qL%?%Q3SBU8h<47d&;_Ygs9kQ -li@VhDa_*!JF6}&?xCCsF(>~paOUwqn%c(qZ37g_#?~GOC1RIipUGQNr>S&=)x0>*1cX7)pa79Xcd0S -3lD_WOZx#bjeg@e5lb|ITANMgHy?J3VLK;lS&b`dU)lxP=W)0iUde}b7KW!lBL(J9pIeAJaQaQxuSzQ -SnX^ASWX+&zM-wWnvNZQpIsZN8d2Ou(j7&U`gDnt<-gZS&RKZ`O$HpE|Wuk -@#mOzD-dKtZG6ahXdCpM^H1ZBWF-Eabpgl)+`)3h1&9UdF6~fTNv59SObD$22g -|>7i(t2_^^ncQ;H1lZ9(28s-DZ$lj&Kp)fg-JnTaehkbZ#vFM1ANKE?a#iCRCl9GNHvFMb#By~N=Sae -F;WMJ=tUCIUrlGZL={3_FliO{34=j*rW4uX6~*n}#e;y>Xld@Ro79Z5yu|0llH$c4>K?2fe%Ac4>T_iEE?3?b3cLD|@%BG -r`7-RNu<3Ds&@k?p0WIAJ4TFw3 -xJLXz!-x}n>>Y9qcbyT*!Zp_7sjwHa*&y>F5AOYsZY=s-<74&Agdn6(6FydtiwH{BiN?q3$&sDj-hrB -Wk_<@F>U7BRE379s!u!W0kPfyV6C#xObWHo0kenoN1=NoTfth6X&sF7Q(2*2|crNF<2q>(!4=RdTAhl -R!X}T7xD^1vXRdtfLSWR8@?p3r-IT)a8IA3g2UE$Gs=e$s+Nok#MUMADDE^TyPB&%|;cfx8ZD=bK2tB -nkUelh*$ilqGNRq#E>)}}&1Z*lIqBwpn}?=$YX1Yc$0+Vb3U$=|}r-X-fW(40sfj?u7AZH$Jm&)ZlvT -&H)EIj+|`=^C%wx1n}izi&ZK5?bXwVxqW}%8Uli+~$6_oqdo5KC^bKt1?UpJF*IjVu9Ty8Ddzb9+asFU95gVr -f!7Tz*VmZ}c9Z=HWteA*}d?lLDe7PMRV-DOW|JX}8Ncb7r2ak6*HYEdk#NG4X3qL|G2;!u4YhiWf+0S -bCh&e{wLVnR#ctj(|>F0O$&Ycnu)UiOYP4o!_2$;K|~I26XKv|-BNzix+3o1@pxK`|N7hT*kyKuiuU1 -NGWD7$hHihwKAEF(O$w1|f%V*UoYS8hu*6>=t`&4|o-%4teKI5wC)QL(q9+$gAMskZ|sZc@?x)(TcKD -%7RJ!du?!OAv^cU208sRottEX-X(754w>R&?~GOE1RIipUGU|Du8vN*vir0JD(%ZZb&4HP(@yhKr_>R --%ZGpJ6ndSBy(@N^*SL`Mc9G{1Ts~X5(SSxrf^IlPEh*`6%MGWPC8;a&bHgcOlYzYpcJUecRZxOA@V3?7>3WCiq|>Z6AGbGiVcjaB^sfKbW+`q|9X~-4*N8HjPL-gU$zSKufz?&Sz{u>~j5_kJ -oA}?A@^HZ!NOx@4-&)*4GMQmiJgx&z!MAAhi)CN!Lb}B -w>4Dy2#r|(?Q=LPLncE&?T*#R`o5>=tb+MbLvY?_H;OD;(^du*=tEK@wX9oH4&UB;HrOAduQTtRm^!yr?1x+drjpk+*q -6<)C+)f+}Tq7{y^cApETLZ#K*}k!L$`o4~W3x_#Wakw=@ba}$$>sIy5PjK}82c>S7H&k-(tne&>I``a%c -G$Fq{@@g}$$&PFADr?wIk+tK2dCIcKK2gTrB5*;S=a}mNa2~pMM9ab4`yDO_`^3f^rCKU%Am^D+Itrc -Jy!4kUbzWaBclZ(=)O7WT;n3Bzqg^bfBdusIZ0^TPneHK9sTU^iDh(DdDa8`AUtE?IAEfN0;;e079mh -qn~e+2=T~=9TzTa&z{a9%-DQ8Q=HnI;);MHE-NzELbPXU$7VvNK|D;|X1)?5+?DsvuQQ5Yqy425z`I% -QFeN|#wdH}HU@=3u~ay$G^Kp`(MC=(w70V?zT3+VaoeE<%8RGVnyDF8y9P52Hpj^q$*Jc2ek?43M{CL -W)HDjDV*au@}A!qN;~Vx%WDi{iwW)v@j`qvm3~;1>Dr7l0)rKajrG^r7N<)!TS(7J>=s^yY|N5WzifF}131K1$8|`n -j@&3DD~UrkHX2$afSo1MOzi4p-=m<*S5=9B__NWBi=*uukhH5y)u51e(Cdk^>nPm%;64eYi{0Hf#BJb -{vM30?(F6{$7i8a^WcJUUy68v>5PuMgB8Eje5LNPN`2m8?Ot_z_Dp)T9APPFF-q)^9Z9acQd(D7*x<=pnTN6+)7Ds)m+{g)n>jdtr>t-u1Dt3%OF- -mz4>t0+GR#n9S4_+I9*?<)%ja`*tr1UVJkky0G@!YBQHsCuExA?$0#@A!mR;MsQsQ?v>GqM+`$$#pyLj -^Bwac;XUA{OqK#z1|jC@5d4Xyo2b3;oN#d=PWPn>k7DJ`Egw@bRjK|{70Aa;g(;Rn?zzZjZlv$u)Mhl -|5-RS)r1BT8-y%SC1nCyGdl;LJ(cHpT52L%{L$|QaBZ%yuI<-@g_-BAWV%!qwS4JCq01>YuxtzzoDiO -5u7=T=V{LHA+w*Moczc8X%J6BL$4Eg&Bdh(%eMvg>?r6E&CRJRq9%2 -Ep($Az5x5Be+;S@~Fp`#Nu&@u7b1q7$J>Wk8d;*DBr#N89~!gj6zu;qd2-GYF~0wRo?uQ2=4$v8Kbou -xQ*V|-!LYfK%=+PH;maP@Lc=$8^$z4*mmZp1UiWb6N{TJRPTGE+FmM>s!cn~t!q51brJ?T&x)c6eDA5 -MGw2E$aW{*{%I8?sg)d@Vqj5=3NO=XklLtlB1v#4mJWz&oO`GSK+mT+9OS8TQxM2^9xNJSb&3f;%_Oa -4TkFN0qR$u)^HTvDn8cOD0kuLb| -0|2s1WjItEyQjKz-_bjaVJz_v5u~=luL2McJo$AxoN>kxYUjY{< -hnH~+JG(Ez0jl5BdzD{HR(5aP(L_agoN8IZb#e0i2wW$3nJkwBZa5rArvBgDV6Zs_|XT?UN7o@718VI -kfgNTN2rvid<>wkaposM{_T;+G0674;Dn;nV=;nObEI5g;4JJ^pi`9^&?)210$ULukyiN~L4%sK -#@0t9}S3>x%R+K>sGF_M0xrH|CDBen`A=2+f0HqbcTGE{}LNZW_??+#^rqr#Eh3sM7u?&=^EnC&|?w? -!h8J!V&*fs$%-$M+`_sW#TLzuuR{G*?(_Eps#Qm$4qq^`p%?r##ASAUD-4an5?As&z?5Xl2m6v6^Rgy -v@!>eQG3tF!2|RiS*}nH+A;u=L_H!re27}PdmD;U$$puFM{YI$cJ&P7Ot?XZ>gdZq8L(o7fTHRj0GOt -%i2J2O)uBBDI_6VH6k5oVSgPLMb%2Wi3El^u0BFt=CD~PagH;$JCQ8YoqoVEC7$g;TX$?2()#7j$HFL -jTiTxpJmJ(Q6g*Isq9^6GuT@`SHgC^cO$ODPIIz`Q?*nFVf^(Me#aY{#2YMZX91Ks^c>TlDv3Z3OxtB -O@U%>sBL&i)EtDN({$eucm(n=OmHqRv0X@VHu`iv&6fbJ6(p%?4wKViLHnR2cgdlFa_OQx_dcaaN@uY -Hhht?De&rRK!n^Wwo4SR=LS)T5X7}pRNh7ZSAZY9WI$g0v(mIXnguT)B-86c6G6wWcJV1h16IDt3p8) -zjB~BYF@dh=q2!k5ZJ{kT(uy!f4VBX*7dBabd@qz7O}EXY)K@t-JCQ?v4ksO()-E@wG({4hSP#@4jy;yqbHuNNARh-pKv=*= -<0j-U~k$2Wc;@X{1k(LW&nRQc=G*wZ}L^ymfU9NyFDQDeSjx@6kgpTa84V13c58o$6GFXsHp^BFNYbiMZJGZX&>RW>nkXH4{+dW#HsL!V*#C*Fe(gArv^LS!bN -DqPI#TyF(>gNtH50o;eC81Gh4S;Y)$SMs7-yrYfq;FTCrHuO1`qUA1 -G-fv>BCz%UUVj=D<73{$y6U{?vT5w!LXS0z^`Cdq2TBXZ|BP#h_9TvU!sIW96+WcnHDR@X{ILWsv -@kUYN4Ty3@sJ{M`9KWh0EjkM2M?>8WtxZNm3IMnRLa8!v#G*f=P!2cGY(K$GNg0Jr!8si>=;Q4_wQX|%%|tX1fOxNaWz`F=<4oY#*KNQNrbVZ{m}k*EZl -n0?g4i0PM2DiXiEgu(uJtnNlHVSu>0uh+*70B}wW-O*n8C^HCb|BNySjDs{&H#s+v2yC|qp}nO|I)MAuT`Iz{zq!bTlZD!YU~tH3L%@M)+c^}U6_k?P -(;;hI>ljfr*9RQ66nk_7ZPj(bY8R*z6?=>5fmpxg300FAk#suy!eI%y_9ELH9)2Kk{k)KnX9FsP$0%x -|Ev5xl%U@0OK1b+{=ygFiR$ot!8Npfy84$KbnD8)$5-Tw1{PN`F;}Jnf+dcBba`hF -i!jlAA%0oZ1tiPWRNkE&-w5FFd?U2+Z~*w#2T=$^9O-q^XLh;W2DU -RF#VLC2Waa6^7EI^N05c2q}zvgfMJcQ({+z;Z~NCrYejE9Vv;b!mz4FiEao(ZB^v5va?76>7$s%sz?B -zJ_%W@N(Bg9$1jUj!PLm?pF4HYkrYTg@;1-v3N$H$eHEJKV`&qg>%>oDpUVe=Z;2w6=;gWILM5!rg0X -t!`W-dLNzd3geMPWIx40gH-q7uUDEYVW#&H<~@bb6$5kpus%R-$AZ92s;l)C9HYQH4f8O~?{DNHFmoc -`g6#Do?2{LwycFP=<#p-9E4_7f|wF@mu?A7Dt4oMj*qy~M(0;1fEYQ;XdEQ9EL><53 -*ZLzhNLZSozqiB*5BwY@j4wx9FOBZ2c>Sik~Nf>O&EXi(_H8^1%>JQJ(^G3_Mn0GoT}ni(hbCrh*kKF -aSj;!`yIy^zuPxBH}3-%(#`G09qagry@~@{26C;piWt=QwyYiu`qd*(K`U7J`*p}Rj;O>E09^H=JDZI -0O_na*P%_Mq41?OJBDVJ6>;ekv_-wSKs_Kocnen#0^xb*X%5 -9j4F2(9Z;l6ffkW6P;WTwIS)qvF!;b_B^4*7iLwsP!8W<>O08=;I6NqWAGn^hXZ!#k)^Oj8}CU77-f` -F-(;L<1K=g2RA?>Mg}Ds5Q!BaXTWA`-r_G1d}e)F8oxk(M7yeWCN6+_!$AqgU61q(x~n&)qq_cmDyWI -z$Amzs7r4RfaG=b0kzKQ$T4Ui{}ChEG&HH`sZ51($)O5P8BBXR^hJp8C8V7@7;KT|b0^qVc``cC -oda4-jE9x|~QX3B|>O`&yWW1=blG;Ce+C)oI!L$mlnDgZq1gXsyFuXIqNEWD*r$TZD{Y#%{Co^Ohf$u=`(6$91RmeJ=zyn*r+PPN4N= -05~6A2cBq7u7b6i$%8|1E~eRr3&5@R7J(%y(|Z##R2JBNN&Q`NpFc_aHc>fsw!c;J)ek)1{z`!)8(-aT-d?a{{n6Oj9K3b#t9EP2feW*K`;Hw=wIHT<9oRYF1&Vw< -3U?XdCNmZN(mSy<(y66a7g3|KovZLWBF5$`Mh(`AD4x62^~Yne<40MT`3&B~IK@mlc142PU=->9?};~jG!Tl{}%z<@ATHDIRDskk -eSeszAD-7^H4QTd1tp|Ov)1P9Bw-77cZOvSsFfNJdbn5Xjno$vgN0E3};*eWR=AwV({%3r=vX;oXNYP -2d_(~#)HSN6KHE_6mhJ1Lv**MHcDg;(T5pI`(ai@wQ2i-vW$v~2QmDXHe0I8_r-Z0Q%$K4$Rsc;Qc4S|(g2v#1m&v7%Z3>CO@Kjx|i%i$Vsj55w}hIqd -!cLYKv3BQX5MfC$%1F4k)pFz09G&iH%-o5`f}wLAmLzgMErAMUxi~kh71r262>pJUFE -Z;SdMX?MLM|=eRn+2jxSj|ZhjCsCaqkuZ4*g`n0EzD>4hXEIs0M(L4hf{vN5NS87^_erlpXys0`^1LQ -7nE#0NWCsM@gXLI=3p#o=9u0msOQC{U*XnRU~T>3}!xN$&jf+D2FEofSbcmFVTmfBx2tp2lOdwEOOk% -)%pl9H|BIsP1xGaiP&^65h?0)Xc)viiTpvVXWxDBGA6}|h{V_PA-jYkQES?}&-||ZeT%Q5ZH9@}QC}i -D_ys_b-|;gaWw>0vhkyf?Xa{)dE&z0zTlg^ZyZKnJ#Nwz(pQ#)pq~ucE?^mqt#%)08^(Nv%kP0=P03> -93@|=UB=ytC-8-jG_k1+^sU%C$9eAG$N+0fcD0iv(f7?1WaHht;Ec -(I4kU6*T&=XwN@{ZprQDiZ&!g0D`&)|P4jS{zG+IC5KU!6iZ%xx1j?5+RDbp^D%VA&A^sI?za0l?8B$ -u>1mAb<*9{H%24Au7e;Wab}=2?|kJwKoT>A4xfnQP+Pd^TWAh+!4UBCvv>q&Be$Rx+@Cen3fuk9&^YE -cPW{d!Xig%(rz|i`YY`iREyRQO7?7zm@ZE=~neQF+#iyIUB7j<>cJHA^bxe5|x1k$H2v?li=G-Yj7l? -L`2rzbXRuQ}?dCx|jkjam3z5}Q@?3qB0`DFwd52cY=xs6-PWY~O_kv3-OOAXrBci2~d24FrKYdr(Qn( -I8li!fHX=xyQ^u7+FzZvZgKw8so`o@3!`#Wqe&V9hx00X)9M=ghdMZKSpZFEG^L`v^hIbL0CMBFZ!pX ->-?JVx(%H)-Z4`w(-yl1(#I9w=#aY$X>5Di*vd`afntfXzywWD*bs>I65){sF$fc@<(}UPT+L_2IA5* -f*sxjnD|jDQn;yI>wJ8vFwk%)d;30uiNjS4dw^io)7;xz07T6%Z$op+mB0KR3K`x2_?(~gDiLsPMJpA -sE6t9I*HuhMh3hJ+D|5{rUKZL{u%;cidCwj2{3h?BoR9K3k -+fA`Cuz{~aF3YbCNLTE+^!fM(>Ovf<^< -^1k5Qkk~Wu08Shnj`4hqH3MO3dbhz)SrwSKoN>fCSRzAdPHJCO%!C(a7Rt5;*3-10#DANoN1tsf&)J& -;vK;ivu$`Qx>4LII&O>6I_FAyMi+imxb~aqdWkVL_O3$hkAZ<;`T+yPqk~dwFV#+>40)(Wg?GG-tX%} -fz=x(kgqa&kG=cdgw6T0YUh5}fZ!6wKLKRceDt+;j{womvvIQ&W7B@EakUhqyBuiaZYe=z|J13Sio{1 -6GCScvxO98|cJ(c^U$N=PdIg8j5xoqH*MyVGi=CaTCKRpEM*!);YyEZjA{BEil2R8jzM-Ftq(gCHK#L -le#;r^k%AvSY0@p9kcb}UecBLEoBfs)M;J%53ugF&+5RO`ZZNkv?O_R!Xfs@9huW(lSg5vc1OEnNzgz -k`aXW&;pyk#Pl0Y13VdH(XDi6Hf~jNT!Va`*?f%6#z&rj7CwG0;+Inm05m9VbLb> -=~&V!GO(9iY0JUBgDR`9T#% -2w5X;3FQ7oL}fXlNq;W;X`G?F@swlvbZo&_y53A<$}mzFF@VuygK0x#{b9kpIsGH{fAY01HLXYj-=_z -M<;xez0gg=-KhC1RQ}9o1r55IYLTw4ipyy0#jG-oBhnYbGSQOW1K#T%{j! -H4!9J~xa8!N}#%d~O&RJ3D*#Tw`SAM{;rwmntCw&g9aEI!+5ek|b;4LXvc?y+;>y3&+tx?)Dk0ggqbC -2U@>D0_jv@Mp{V|pU&iEWR)}tT?0={d1=-I+Eh7N?}u<3LeXt&OPw=C-;w?ZlT5|`QJZwTMM8 -je|SOeun|D+kb1$yVHs*n>|Jr#gDMx2-eCn;s2(akaw5^|;v=`11s%N-J#q_J5V}^wM{dy~a{K2ULIr -w~(jiWw&{`{GN6oc%^p5gt?HIT`zC*XTFIJT8B78^=Zt)lAPlorQSGSc{F6eV^hVLJ+Swq)R_+1ZkVYm9EW$LNBUJTJtEWZ@j -EdVSA!*p5=4Eg3kfezxS`8idu;Jr76S(DqDRL4k(&`NgY=UX&NfrNy -vRYK*ofaFiQu&%;r9v^^8our$t-i`FB(C@+$WYgo)GrOv2LbIsCbT1WZPW@6V<;l&2Yw=QbhmIq1gl6 -0;!9;{z)K%W}V}CjC4OrDeKA_QBSVkL -V3yWwYZEXn~7__i}ZVqnCN70(Kk6nY%SO&QTv7;h#3u@PaLBmC;3)D$&&4eU(=&_|dveSKSUo5#j4@c -eP_Do!E?LGHcT)dj{MR}22oFikM`X=H??*gYgc6KA%)^EDw9;uOazgJK?hG*2U$t&m`V=`*m*%b^N12 -M91=L%AX^hb=Z%Tds+F3g)#8m~nh+)COa2yUfq9RY9R&?5ZZz``B-t`j$OoJ4o@lESCERKcWcnLAXYb -y@TFJm?$94fiiDT3RPH2b#=9Gw0wMO!IkFC8_p-7qj0Et_~jRwNVWfE0NP<=NtLBO-lLyWBJcB`tj>y4)oDlGt@VyWAvngN -3~tPHAHvB(+1#xPp(QxYrJw7I?!x-C#if+=flM!NDc^h8;S^$KD}_{0T-R3)di&PohxL(xAk|!h -9N-ov_t;1VFVBZ*B;?%!vNIT*t_GBe~k~x!6p8(NQaE0+-O9kBUCpU#4c&+5YUYVkxOD%MCwL^xJ?%J -Zn#8k@F1xjVotjOi>OEJn=dVhZK5Dn+BRVjD{=ccNE3@Tfsh6s4Uv#4^)Sf?0|(=lj49H#XWWz(S8WVe09C}^lLejf5JBSALsKW9-h1DiE15MZ_Hv>)CTASVUZE`cX7`XMgMw;b4nIP -4W>IR9{MywXfHgdI)wim3MK^w_BIW&m2N}DI>a?wqPVwZ^Yf_JmQJV;u4rMcN)93-)8!MoXD8cY`UZn -zAC!GolBnT5Gv1aFpB?lxo6g1*}zXhlr_#JdfWR@5$0-)#_eg@?T(E?JvQNOFg;MVjd$@jB@RiPq*>l -7+I(dn5~K`-ezw25nv-IXSq^MV+=gjN&jJ5Wd^`H#=^d$hnKWP2k)`-#%{Mf=ipQc{7`asCkt^P!{fj -&H4H@#|%K&^hL#MZbM=)pf3wva~l(bgX`MjHMc>b_}Dw-Fe(Hil7+*tRPwVnB04g*Hq$zCxHc2J&c^r -L0&kz^wJi^l+9BpSKSbxHpBfSA*vBWgm=!G@X!zt7vm$oIJU+R_Tw!7FhC|FI50cs?X66^qII=wH=u~ -qb+Tzm|{X>J$O%}8v{m>wLlZVT)e`pv2l9Rntu92WvkxX0zB1@w6aCy`FE*#px-2Zzm_d|t@p45Wwo0 -HCERs{9;Hq`cypSB<;3C{?^aS|7~cDg>SuxfQZd}9*K$D@vZ!uP~FLPaI;POOiwf2~8iqVAt6RkOBRN -6-ZU*4FDlQEpciZd&^V!I}I~oji&r9#=N?g*TSY3Ib7{99?&`sV^(PjS88gB$V+yPV;*35gqC<-Z0n7O6VJ=tfTxWZ3OFf4D?}$dukj(F_Bd%(HEj%`nidT3-@NBG?%$t0`K1EHW#INIFv!Eo37-Kl^0=j&5jD>wq1KH_d}$ -&Nx1a0uj}!ijCcaEMy{+fpydPl&+t(?4=UGRkEI-sSi+S$DvEH1@@C`lfq#beQbJHPR?0`3KaAoGgrm;&Q3HjKW^yYnAbgg7sDtp9F9h-8ta| -Fj`Ub@{ywUQ|dH7=N~61Vw3-KiS|#!5@L~M=Qo$G4Ow|!RKXZtsm+LQ9kPMU8yRz_D=|d4kGBuDklw% -by~ry!p|yjGU%bFI0@B8CSijhj{MLIqF&JTvsh)W_Ck?KKkbm3WR3b~=fjB*kfK3R(9Gmb -gX=nPq6u9>JGq0s(pu6sE)5{L~ZUsTW=IZSy9! -p`(w_sqJ(QY%(Esl%-S0y&rwZzA&W)yGK(|N*Fp~8>Qg`}eg{|>aOG2GKWC)9 -{s)$^)wbvZz}yBmP4m@esQ8^vBnSMA-r52%qY45y8BCs5VDci%%>chU`YtbAAhhQEMPFCz}|(4QA?td -$dVzWo${NjB`MDx%A}>{ch_qa?k1?nEY{#ARH*Vn1KeKcpMxowCr@j^-!}$jjzKQ(=2?-r)?jLa)3u1 -Sp9PK&?dmgc9?)*|nm+mN$-NAz{h5EFu@m+6cP(Q%4q!Sb1%BN2ubQe4Ky2b=CZW^6YB0E3_F=H7APZ -M|#Nne5whVY0fK0Cg-yL*M&R(%IrM;!h=ulFPN<5_v&AGV}2n~`;YU-P$OB?7MJ7nv-we3KSOAXsAF8 -RMCELwXi{pQ7Bp2=l`Gmrm0sETMyiiAcW{c-$y4=3J{@!vqUY;6Q@U`8D9objq-sRI -kQX)NJqi~{bf13;U#?vN|6lJ8tp7Gx8vfHXcEKM|O!a2v+EW8g`9qQ5_4166YCDIl`$3`@!~&gw8&hd -}&pmyIQ#pLFddQHyJk_~W7H_h$Vl({J*A(!1Vn%vU}?%fq79O(C?u-PN`|W!{x$W$TvF`upE&^ur23Ht!nE|Dn-; -Hrr+ErqTNE8vBXaE?YN@*8kMl`)0dr-8MU@u^*c4vUSyH{jghfC=5_QHQA|>>D`) -ZylWXXv9*dD5h_bL{y)respI4x=`2}K}_f%;gvZAYShcZqqCxmv1Iy%z24kHtf#tBQI(DWAQPaO4yYo -z_b2N;s;9&hXMPXZMgTY`BK&ne5Po!4$s6yoK}O6kTSPg_svE9IKG2sFyY29?dBpuHaTo<9)Ry=J1IF -Z;6f3&0}%vI6pXm$8*fwfFm<0IVv!u4`9WBx;u)=dS*xd>g^SpqqKamW}4e ->+=O6<699?!&83hXVg38slNdhDL0K`>E;?HKFVhGU>s0L5~e3uFZQD1ty-pYW9|a?c%Vj1k8D2-z!tA -+TO7u66@RSvT^;U5bX2_c>M1a_cofKC7jIpVi^T$6x -9KhrJ}`-dtIE4(j6S8O1zrq$kt{7@k;##257^xR)w^1xy9VN-7*!KG-A1fxgtU#pMb#4dm**xKe@i1( -}5D(}*Gh>?vK4Zk8&6_VA;LmX6~Yj)g$50% -lIlv%D{pJyam+Uafuv}~Pl)47YuPDnMGF2pQTd!wq!VH79roD?wnmyG#$>rj(txH^{+<{Z -&ySMdraVAL1*Cudyv#p(xbWX35+i1U|iqVO9geMMgdb$VuP)R!B+o`cMw%hxtztYxj>+Lz;AP?eLJr~ -lxdpv{78t&-d#-&`m;#OPr?bk4cLvD&ur}f~Z8MpBsrx77{sGzrJgUeOi`v7hw&1ChCbn?4m`KEzRfzzezZ -1wO)gN8oS|LgO6as3PE$R$SE99mSa4NgXxn#Yuv3jtHDV9AcD@6bkn!45Zft61V-`OfM7#+9wJv805b -9&te_!MV|^yDM$DA#djco#V2sl`hA18@Y??eGMaa+U$a$J2erSC3D~)B~e5vEiv!v -q(mksTyk9x-+`XjY;FHFhd<~X9pJi>WJ;0SNC$CUaPXmza+gTwuM{Jm)%iG^IO80R~MGYFJ@FL0hyIK -ezaD8$d_1B`#z8v-L4fjUh|CFl%un59#4IJgHA%m$cEeLi&Jl&c$8;UP=r*&v<;1)bufPDGHUbH0I8* -S}7j&+?s#p?p((J_4ADKd)TE0;N7BI^v22H-_TSf}4uC%Ze+x32%{p%#@Y>*6ssB&N68#zro -NglBPb?8GKB@YLFVhEPtlP7#_^0{qKV}w|@r5urO5Ahf428p}V9qw1DFTN$GLs2JQ*MeSnMyd@rw!YS -^XXFSA6z;;C!g@`bj;`2z@!g2pa1|WugWToP>U6}gMJ(frh8a&u)7teX;%1lGhqvUc6lJox`I4S}#J4 -wNzDWwYs7H07cNw2OT-PoO%^mKBLDo4#gE5+|~M|>c*ouiSeKiYQ`P3*yI0*`91H?av@_0)u)Rl;FE^$=~LVVL -gJzrS^WpSB@nqUYd<6)pzK-?cQ4oRbk4RNfNVS= -JL~0-F=6o?v(2T9CF&j)&B)-6?cb|paZiOUtrmA?vxH%5Qdc8ZoDX -B1x)Z^TWDt)Hr26Wvx4n6!!;S@NT{HWBI73{RdqJu|-%nyjVAhACA6yO)LVf)EL4>T+3sNili3leml) -A|J#pO~l0N{7LB+a_Sa3;n;0x+D4ue3!NpniFkX`8n{S$WER6Fp{$PFK -;%Fm;P}ggYkg<^;;;+>w${^>Xo2qkz0;z5gYO}k(M3*RhNej}{&k<_f$`A61IV^Q6KlB4_JztUi7LlQ -z@KL%wacQSIdtZ@yHH@PkUxzSK{g-g0fj1MYsCt1_Hw42x;tRn9;nYVU=rW`c!bz-eF=9 -MlUz1uBBznK|np9AZ@FPN~+h@K)R(e~XP+Z7G0DWY9tS;yvO?(yNE>P-IA;|m??@#JOe3|cqVSRiOq|BE%2hfe^ -=~Gh<8uUYsn4^SNivwsSNdF^+@Y`C)pw-94*d)^%I`=|AY>gs(sp0q)Po2iOXF-9Bb@qd2!jp^i@hTi -nd3|_M%{XZ0rwFz7Nr>raG&ae4ufo%qhQ@TQW3-a3gcVMJQ*Og`g9-pU2XZ^lUl?vSPwZk@5!MZ@fBj -=r}`>efmRcgi_Z4fig*r&()>Hk$Sb6OsrQFX`2NeZceUgd|ZrA8m-DZs9dEw@n<5#zrBkl=|3 -_Y(4D)L7GIE)|2&qaKIbA(Y>fk;$`MHuykq8@$uKx%r#=De;S8WZ^gIW_`wp8051{Xj}VZ157F`)dfR -ZmXg|-wQ4<>f={n7=5Z+jVRDV@`0QMQJnPSaK?gu5l&eL*O7r5*d@+%J^G@UM@_aT} -K!GC@dnQ$A)1{v~w{84rW7$Q5ApzHWZULDuzyq#9#9CanwKP=cl$lXGUwH6S1833yD})^RSPb|M}XBf -R=l{?Uj1+9y)7iO>u^5k8a3ljLb?pwy=VSHP%;QR=FZ{C0#R44+A<yQd?5LoE)lpW)X0y>oeb3^ -#YG;0@%Y7QUIy2_FCL08Ll%ZwTTVzjmj+_d;0~U?qgEi+)j*u-`&<_=R$bLOR>q|v -&VeI@SYK(=kT0M!hA8UOTOgrB!xPfBrXsW(SJhqM*8SKt9-;HV6H>W~u!(VK&d~%Q-~PrS#b%W2#{PZOv}WWO7&o5IwwvV!Hl1fj!HVmdIlrH*3AXkeCb9>y -l3=KP7-4d)fcn$x2Lg}#iQSj_^%izzrBjO!5IJS`U;Zy!JHP~7t&y7{^h`M3cir?;P830gnS{zkqvz; -voGWi%>wQ(dt&#iw-#T>IgrJJ7$MZ{ViwOZa8gn#M?G+%7ylPS6?6;<*ClNv5q>Pqo~yh0em0;ZYDW^t9^)QK0+j3CtksfWS-vhNBSmdv3Q=px(__U5x(BOdr&%J8_6D=W(~u_$=(gk$$61?TfXB^2JhWpM5kk0~R4m3vd)UY;)>yq)jryZ6JmtAh2 -LdWNGXFVNUwvT0Qh5ATS$X1hqH*YHUc=>!;8(!@CD*WNLP2&SK -TD-%-`zjIeMz=ny#-bUC#!C-r7qYXnrXOOK1P~n7sNK>EH;#yL*RAy5?{Vd%kW1l{$cIoo>!ixpRifv -ZOB+g|K_cz$j-HvTj@+Q2M5npkn^rE;bXx6;4g9kFgm$QBFV8T2Z?r6$1(p(B=api#Hxg*zmFM-n-8= -4hxh%IWZPVfygJ{MhDSRL;)aKI(|<$IiDdvqN0&3oHW()s4SHmO{Y2~kBq3st^KT>ijJH8wl$pXi$R% -vRFO2k*g*8w0)8?udj9+CnFqon@IoOK>4yary03h4^V6w(2u6p} -kxou}l3@iNa*cO#`1YI6RwF~sD{tr=Uz2&+vNvSwU^AUq1U5XY!jIe(?{Q|z0KnNqQX2kW)?Ng4iMl> -oey)M#}06%e+*R3|Sc%4qD2D_6JWTXutu7XGj+k#@yhuFNFMoCO&nMt7scZ3o^#`KADmSj>HzyHYZst -uL~M4^w7b4|Vlk=M1-XxObN!A?(n*P<@3h*wPCxpYGu66w#4roaJN(!=AJy)ayLk!A2>gs$JA_*nFu@ -P}F;bvlM?j4&{-z?T^ZA8%?+Ck`g6ltE`jC+~CltH|>)2{MKi;D@b()d15Rz8?4!fEtU`o)Eh2+#1V+ -$E}F6bSrPTz^=O~b`OQHGkciKx9w*aQS~E{!cT%!Co@0|Gz#sVZ`T|*j4#gbu#uxfJhV`h`+5(B21q^ -MG@VNprk)fix2c9wYW}xw82q6g2UDJ($&d3AxV8r^N7%&&NTzmvyfP;>@~6Y#YvoT;1QF^wn5l6;F3$w+L}_JDP9-U=O$8dPny0Wb`?Uy?La5l<;zt>p*F -MHu6`8XEK}hireHZbr;c=RN%$Th@QHE}kMIEQ`0F{KDyvO=^{o<+jZ5TiH*$%m51!f!67*j``BqfSjo`04o!S6v$kui1ahhq-q*-K)B2f5?ee>mT*|oGV3r@%6g^la3}T<0VY2 -vGyD$3cN}loFvRnfu6B*9uLQ+Z=5cheA>U7fK_n3HJ^Ho$#alelM+3Ou0`7-J(D6sYD>Jx3w`KErX)m -i|96h|$;6G2Oc#rn;pc~O4Aimn=#RdyG+1{{j>>e=w&R3!Mbc;+p_&JujesJ$08?Et2CI0yigISGlmH -42|m=$Ble*M+`I}GqN{+ALTV)`#WkzuCBKPxf*niFqh?8Exu-Fpl%HC`$4!F|T@82_rozkDs6ACwq9W -an^shgaT;D*~RCJtEJV<Bc -&49QJos~Vt*KB4l0J(-g<1>r*MH4tA@rQC4?wazg~_^l2iCiRS%-Rhv%gVY72FVXX3L-#2dgZ{k}B{# -3kKebxyxZS`Mtq3xOVQPPX&jVMz|1I}NAeu*a`+p`8slkF!FoU@Px=Z$R9K(V^|yYvUSPUKZS2k3sTG -&o(BxzeL_7&04`0o~kjmR#W|vYpgu=ZxIjTtrD$6oF;pS0MGORddA?0;Ca+IPCF8XuCL*lxkRHU=1>j -$MI8KgvlCxx+ZXqS{M!~^Xstk({HKhdB{kvm;v~GYn09mLaX9g_=&5iwfMfOkJb363Rdo^G(2Fx>F!) -qG^^KG7wK{|DIgI8s(UkGPx^Bnw$NT%2xt!ClKp&?+Z1cOgz`mb<aSsfwL0Gqxa2 -SSZgVT2&gE)bB_aPr2FyewYcR+aP=^7c^;A(4;v*lXe768W1#TInboZK$ErtO&SR_X&um{SwNHa0F4>~Fwd7- -05k#pZvyk*1lYd`lz$WO{w8q!K0yCp|L{LhO9KQH000080LuVbTjP1}xz+&y04M|i03rYY0B~t=FJE? -LZe(wAFLG&PXfI!7ZEtF6Uvgz|Y+++%E^v8mQp;}JFc7@+D;Dg9Y!#tr0jUp#0yKpiAUWEBK*=jdi6j -aXmD*_g?_E+4+KvewC~{_YcDSX{4cKE>wR@O8tHAB3w&|^T2KYoGdptnn?{xl(h4qXfpUV?t-flMMV$TuGG7GwD-}M{rA -U~3H+z&a+QfH17*Jxk#?t@+^HOd81x`JjT=@Jway6LK##9ilXhb2sn8|-RTpkP4D;=v|PBl5olQ$uaD -cV#No(#;&>7{M+4(k+hANsUY=@X4?r10FouDFz|L~`kxkC=p;Z_Z5OD{#& -qN3Y=V;GT>!^QL3E^{74C(F#@d&lqPJomCc_Ec1w|nVtZDxa0B!_;$EUzot9_?#av19sjq;PDJ9(5-} -ZL3(8N^1x{RGs1%n_m6u$=x<|TWvoP%R@<@Kxl9Mq!kI1bs&e61QRYUvW-iyS0qlc~+av@rC3@UCJw1 -a6wq9Vw}wf>DpNC{s~qY-EPldz+!37w&X5j+&wS@?;^)EI59nhITxIZel53PngQ@)IK2e^5&U1QY-O0 -0;of09jieTGfm<0{{R}2mk;g0001RX>c!Jc4cm4Z*nhkX=7+FUuAf3Z*p{BW^ZzBVRSBVdBs&tZ`(K! -z3W#@xrfAnC3(}LTR5`O)Q!^uaT3JdVha>77>Tl&r9_vaBD;eAdr69tT_;)Wp+$8NHJq=m}N(F|AVgsFa7P$zAz<7w+vC(U_drMXY!vgjCon8ABLK(K$!YI4Xy^!%Dij_U_LZ1 -5g9&da`?{Ys@@ih7VYwyGx_$80y6aO%YXSPvKVHGKs$H>Y*J@wO>4-%u^>5t?Yj)#zG8+M>8O|ggYYA -|Rt>M!`|wRp!_^lJ;DGyH}sbZP%*3GS5rGX?l-uK;HCsv@Ib3$mn9?W)LvZk{bgSBXA!?d1@G`twxB) -BSX5|M2x@XTxYl+{9w3wDgpfiS&R4#cCxumoCoL_YmTOG1oMZ|2TxCl0Cjo4X{7oJj&BD^8{(_xiu2@ -V@5j{np@oU^G1qDoZG$=GP=IOs!`Esg!Vyd#jZt;ZQekTP=fB12)5>J4y#;S(`C~>5j3>zYr2N&+vI+ -my^#&w9pVOLqX-khQ(`qMkU~utea}Y7d@|K_u!XS0u&pY?8k@=G*?fLJU*6oFPe1;n&CKA^6UX@uP)h>@6a -WAK2ms3fSzAfex7(Wl000RA0015U003}la4%nWWo~3|axZdeV`wj5Wq5RDZgXjGZZ2?neNIabf4BH@x! -vzjV6U*IMj8h{fd#eT#FSis91v;g?fVGeBIN&PdT?jxJpt=oxG?G{TbS}XaXUJ_>cI72)c#`wo?iay- -YKYk*B6PUWAU~AYxS#9X6F2oy9O9KQH000080LuVbTkXNa7KZ@<03QPY02=@R0B~t=FJE?LZe(wAFLG -&PXfI!8X>4U~Z!U0ojZ({s8!-^P>nn=ciw9=?0fq!jSi)Y`kmP2}SRS`*ZDjQz^@I%M<0HxAN1PneAo -S|$Dpkt_KA~;RAVZ+Gg+28g1loHREtY*$%H$jwd+QIE``1;2Pue-XbEK5olfi+cHP(^mS)N*FztbWmu -A_uajpvWBUhoTV>}jEpqN=LTB$z-o3Xd-P2DW%GcXILVNC*@MCtx!vhBeSXv@IqyNFsA?KINW5F%KqP(|&$=uDU4lytw+ -eUT>OAwwc)QA$zGs!p+DA)frYbC8sWx>`h8UYS^8(IV6m?O^#5Wcs2ZbM?X921utbvvj0RqVS7ewGtH -j6VQPSmsV$P$Hoas-&^G3FX{^4!guvmU%F^lBoSQ7kP8;Ju=^(8GL8lQJC+)}o>UHx_T#i~UI%oA3o% -Tf#;eCkdKksR&G+he-0JtRp02}}S0B~t=FJE?LZ -e(wAFLG&PXfI!BZgX&DV{|TXd8Ha%kK4HM-M@lR9#*>Ql;ouk22>z-x4q`jd|rLs -*A4RDJ2NCDQj&cQPDpHvoEgqHXULVOx@uY`m9E9MUF@6c5}2vI73G9AQAbrj=-=IQm`vuA+osL71*l0}60QPQl(9 -H;R2lB|T4X0MXfvwN>wQT%Fm+b9<~5`>^>8BiVvb-ZTeoknyQDpIHpvlq|ug@C{K1Dp?N*3mAH4}m`<0 -T_f_f6+fsvG%*x8jl}+%#{2)TV0%E-MaFDa;a2vOTg6m($y-5y^F)3GmWF-U7NszjfqyRfk#{R=K)}# -^{?_0*AdNcv&C$Mzj*ueZHn2n7_y}Obs-zvh{_R-JFh1WEL3F$WaN -H$L^ccPOJ0(6eQb2NqwcQO@gsgycNJ&)OEXTMsn(bv9YJuvs_y^;sW;=6-jJb#H0Lk?hxO4oV9J3XvM2i -siGV2N+^(rW=&b;!zbb^9av04dM}s-fuMleNoujf7{c^;N4H-n5R^F}brVOYXF9U(t34l_>H#4E0SML9M42{6N>IS6qUPIX>om4Tl5H@o)(wD2{6!P5M<~c8OnaboE^{ryM8rvzA;ehzIPlunl%D -+vK8?SyGhwiag%_l!z30vvw)4JZWqFazA;I7_}CZ^h4AvB&B%RjcRl$kY^=lVXXadK+1HScC~n4uY{f -~(__;Jq{fZJSX=%(9v6xvxV?2s;%Hl$p -XBTkw{-0jKUjP6K4egjJwpFJv+xjy9I485!#wsS44Lk})U^gG}5lGXQu%Q@4l*W-pKE#t}J9;n70a9+ -xj%R|hU?b|Zg`4dq;>P(p$yYtfoqW@-`muLQkzw%gNP?Am~#!25+!66)fR_N?u5>%=Z1RZx -q!etw8#AQEAwj3s`9_S8f}>Pgzbb=CrLBkRwOg5`UL||p1$A!tn>4hudFMItHZ|EE_?^TpRt9Op4E^8tCVtQZ(SD6V -4o2Fumm8(Wv5>%HwZV?pSNqaPv#xQ_L6tqjX5VXSkSSVoU^#FmC`~Q_c -jbi$05}9_K$JEoHo}p~%I%4vHc3tqf5UJ8Z_$MgmNeu> -M1wsaM0`A1rl&c|c62PKD3Z%p(%hL#L_w{=BbXOXpP+K+M!#bDM()7&e8Ynr#l35-r|$6(`<2cJ(x~? -xWR)ZO>e-zmDjATfJPbg8rl`-nNt -*I6`5r{Mus7z+I{CJ-xTPOrl5%5O?;pWB9S&Gle^4~Ulr++C}s&=yx)p&^o5+Ure2fv>}f$~~Z4FvQ1 -;TB%!&@j}NRpfV&H(TivnU5|1L3p_&I8 -A{5;opd!yYn{*%eK+;Z64j_1avKpx%+87Id7i3nYWPUnj5{{%7q8sx;QmeOoz1ew*f2XrFcexjr!U@T -xjW!2(m0&_ZQFBN@R(5cOL>@2bZ*e(13xs_ZIP552(WMPtwxSq2M27xE7n?reqONm*-sgJ -!jm$E7UL>V*x7n5&{ -a3O1{H-CU$S1`3aoG3*3ima{Tt`!~0(ZI%< -BLfuHxpa(>gLz7`;;mMXHaY&5FlpeOOa*lzP4tS+5v$B71iFrn*Gx=X-cLz7riYkO -|r;xZ0rSj03Q9#oBaV3`UBbvjd+k>#Kib9iZZrvIyP}acLbPNjYH5X#Oc^;OtYrZ+p>VCUT1WQenD>| -NB1J$#&(Z=O-*y@xxbIK?{tc@_#aS90|XQR000O8%K%wh7OUnPw*vqGnG65`9{>OVaA|NaUv_0~WN&g -Wa%p2|FJEkLXJ2ww{bSpqZk#f+gq^Z*_vcG@8qriUe){{r -ZO7ynl1^@x$kL3Hn~ce1dMM5C*^5bZn_+589D!;XASK?WF+-uf{Cd2Eo(q+snF@!fU+ghKc+5#_G~38 -@?^=z@g2YH{Qv5E7ghWEx7BXqTH)43U?N}PUjB%*bCaW=#xBjb~=$Is}Lc)=9-Ym&X}iVqDqF*QQi|n -mZ4iI1=!iDGn=dncE^MVyrrU@A{wfXD-4F5RMw-ci$Kw8m7uwkJJ^ehrw{1qj~yNe{4&>Rdv?Q+ -2o&{PY;ThQA5iT`f47%Owg-1o?uh<(9pWm<*#`b#E1|D!Kd7s%^tiaf`NXE;{&hpH6gK(u>VbgC>#o=~(@#42{zJIU{x*OUfhmJqun&(L0;TJ8AG -~o=-WAeCi^HlhEZ|6^N0fo;ao;?}%FZwtqjo0Gs701(S!!|QOM$h0CbT6jQs_rox`|nIhO(E-uDJV?H -39c?eaE8zm)F^s5(G1lR`VutR;itIaO-k-Sf>f#E`{T-wcBg7@{*AkdK3Dc>qu}z5a3O1SE6!~9o(Un -Lm;XGsnpc+`D}Fd??BGxw0r?~c)C2$k82|tPaA|NaUv_0~WN& -gWa%p2|FJEkLZ*VSfdCgK?OT#b}eb2AB^r?mE4;aix!Dovvqm<0v)QHn##JrT--1 -EUXX}#S7jZT>k)Ye%~@IdJ0M&km?!SQEw_$0j$C9DoZo|Rc4glMn@ZLO#!Dvvs@{bY9smFUaA|NaUv_0~WN&gWa%p2|FJEqTY;0d4UKaCwzfO^d=X5WV+T44 -xD!tXE<8u%I9my||}RLadWD7!ygRtM0Ggq-sA@R-03ky!U4EW;()}9nfiy4&-icEWy(999KNwF8ygFF -85M@Ujeg43^PK@R|+p*WNZwHbQHs4Pts_vVkr4hlq>EW%nPNaQel6#NFCX1Fpj{lH44a;aoBgL!~ljL -FgF^{mtpNaIjW%2aiUX=T^XZltVJrUhACGLYmP;}E*UW}rpF*P!8?pfzR%>2-s`ow;hEhdDygxGs^m^ -m_sBuRRU#WtZ)*g@OEd43(yUf;h}%lU2;uN76Sjv9>KxmuA*UJV?4b-OZ7T(n(EClbU{D$yR6%#WaOP62ftrw6QK&Xd -2I#>KgG29znO<4KltWfNy4VHWa4|Z8h?Z64^T@31QY-O00;of09jiXp;s-11ONb+8vpc!J -c4cm4Z*nhkX=7+FUvOb;Wps39b1rasota&0BR3dE@BJ$Ty-C0%p0UUB6nc@8&<0AnYz(<5CAhIS88)$ -v?d|^kN~6)|dkjm%N|5tvq@z)u-K_lOu=`L~`_=pAblRxyW4}9`)F=1tXf*msEmvPRTlK{bdpn)>$Dg -mSH`}Z4-PdlvS$C_e-Qo4xes_KI@Ob~cIlip+o9CC^dZR9$y6t-R{iv#mdfB~y``BKNp6(xR|9pCU{& -e%l-Ew)Se(`HvjFkEMtMBdCMOBZLYnS7Mw%VX(tfp0t4V_n}o8@?FN6*JP7klk!;g_?|43&SQ>He1!c -vAUY|T|R$p=Xc#u_8xyT3|mdji!QK%(d(dHi#`5N@6r{KX-AOVe*29*=c-H=#O -*cPP1<2?hMrs_R-z_Z@({}f4yJ+&-1Dmd|o{lBPAYAre4sj1@`lnquV|^uHz+(C;B%8){gTF1#5cw5> -I~osLs7*hguj`O$lm%bwyAt-NIHa6N_6YSMRA}C7D$Qm6$_vQ`3e}6Fki6%N|>)$ASIcvfJuq+6$_+9`HF?&KT8$l%fl9j7BT^A#Dz$}p -ds=VD2PD5A_o=7SL7fg^A$1ZP`;7^2vNS00x5}nMNCSNucSaqkgucwN-|#&gA(N{DS#5?D=CnYylM!T -lptS8fs`O$NrA5}sJ$XRBJSG#$m*?a{`D!>{0fP?aD;7Wq^A!uEB=Z$8DN(* -+fs`m;u|U^YGG76M66PxwKne2|3+Hau&Rpe@66GrvNQv@w?oMwpR`yT!Z=JoC>{~X2A+Du#YKUtoof+ -a#>OoV}E-><)dG73?N`mlfQnGOo)CcB{V2f5 -EV9=c!Jc4cm4Z*nhkX=7+FUvOz-Yc6nkWl%wC!!QuM>lK4facBa$`cTNF$F}rZN>Q3s -B5G^hm0Za0Bdz5o&;=o}GdnX{Cp;isMjuY_!^o2GGA7T#_QseaIgI}6`}4ruuhze|pD1>JjQ7(1Xt8- -jd36T4nx=WpBS>e0^AZyRhfO}DB(OjzO%jyR+e}d1t80s#;=IM%xbia6y=c0}0rv9@$VNpNs -xXu0n5Wa;K#c@$^4Zme?-Q^|mpR)L8KW&8|2MEQ>?_z|Z7`jx*h5Q~~%<88=t&7A$7hq(N_Jsk#kinY -)JWT)VBg0#Hi>1QY-O00;of09jko7`$MQ1^@uf6aWAk0001RX>c!Jc4cm4Z*nhkX=7+FUvgn|X>TrYd -Cgc`Z`(E$e)q39NFF9{N|tPI#!{?U+o3?w4NZYP1cpjWRLoT(JyD92{`#E@MP2Nq>BE2-Xk(E)_wRg% -R@t~3%0}g_an^&WcdlvSO;}izv|X!q+wbJsRI=JW?2NsQ>&mw&yfgIwA3Ks&RSMVopHACcr|@UvT9>< -Ou~_7l@*Z9nw>s%pm#YQf=j7z%rF{Um`CI2Lv|H6ep-W@62W3H{1SG!8x1c<{QxC3dm**(6U}0~*=vB -0RV{E&ID|qj$=Iut|#tLv5S1);~HrCW#orz?>vsj=_Ii5AzyXsCSt!q`qM^bjfuVQF76v%oBXD>KWFqMz>vKbpaJx16Pt$`c^j{{ML0<0 -jemV+mRhHYb?~xbA_}|p|h37^qUrpXQfm?@6D!~_@=4~3!o=Xz#G?~I^QZTsVV71qe0=ncJCxJtaEP= -F{+P;10`2~x&W=Hd>HE|_U;pU0Y6`mBFpxs;q48&v5^AaHd?j1Ay`-m?Ht}(*fCVepLUD7_MC%S(;32 -#3P9;W#rv!;ugp$Ny1kdfBJA#TQ&w)zn?7$ey0+EkcH4$Bylig0l$bt}5%8?I8$74Y*fCs!#Sfg7kM4Rtu|=wk^r&Bp(^=0VC>RMPxbWB3Al?)tS`YKS`GoG6tGx1lC?QJY!lTrjuFFA#*7wTp|@+5+eDX4Ngz7Hn_Z8jQo9Djv6E|AuxFh1tKRkJ; -mY}(YuOJ!I@_y@o`Yh)j_eiEH`Y9M)85ffdGO`@1ZpJ$p0aZiU!!7s -!TzJ>zRS>Dwt{qdn~5L)7a9O02~k;^{#O_=IKEH21m%}frk-hbCAD6Mrh2YLt7E!(rv^kxvO{Ddh&!4 -*|^RY)GoSdTr<;jZE#3Rl7(z?teLab16483?2-R8w`BqnmI^&rNCSOKhAW&+WVJF@A3-5XPJnvxC|f- -T0ev0QiT1+W)vid6V&jM5WQx%L8Y&KeFce89l?RKn81Er1D)mF^D3@}x6ULz%XcFgQ+Vk7=I#+bLmew -6#>3A3KbR+3n-4jlyi-^NckCzrt-}|uQ!QJ65hGpETru8_DJQg0aC)S^Fx7hYKNz_rWh2c^mN?=F;3c -Q@(5u>SIs*iFi_by1qU5R_RRQSrZgOad&RE6w){StHFs)ySSsTm)nJCIi6m3Dfc36bNx7W^S+W$3v~5 -gB>mip|SmWQf#I=jD1pMJ9@_n9Jp$z!+6q*NzVj!2}daz&ERjy7^oVu=QQlj=)4Ah?v?T%@p`7XU^q$ -(YJ31C%7J9`B-2*YL|nnCiDYy+Bo7D^Qm7%IwWt9yOU^?$xp0d|IY*cajbWC*O6!v@%n6RV|s3`^XzJ -T_SjK&29NQVCwPAB!E($YlJ5y;3?`q)8#3Sg=$B~RjlV}9WsNv|{{PGo;(UZT!VkaW#PMIUfhmU)%tM -?~1YOUMP=v7uofPyviz1Zwx@qTKmBPt2y6no{mAdovV~6&_>x-wTVg&mv;O{BcI5<8gkWqsA3-^$yEN --FJi#0ZOrR^lXX)}$__g@caDqvxj)hIe8rf;Dmb)UrRFTI!gW;QxxhN*>_o_I3w$5G~bl8T-|X@Q9jY -KWvnu+6Nhb(X=^6`W)l>B+JaiGknB;Wb(D3pO2J=EWre+ONYiJ%+b&LmIyF9xe9M^j+GB6Xx&JBr1N9 -#xrg5X6^D4o8#tnLi+K#NH@k5o_btMs&@_QPq_PnYOL$OP)h>@6aWAK2ms3fSzE?2tyEtQ007TT0012 -T003}la4%nWWo~3|axZdeV`wj5b8u;HZe?4a#?PLy!G`G>^AI4!7z2hRq$(d^_wByi{PO`eA7Q&&&oezUlAHyyfJ -&Y@55IHzbkB6RGzx`6dlDmRI;~`~lAKYbfs8i%!JapzWJ=waEh#?OBhi1yqP-3XTKR7meo{yQXBpsX>YB9iQeF><4R^&&1O|CEvZ>)NXaODXQaW5 -oX#ozZ8RrEk+cCRIV5KWl{P7-Npkv(G(uCO4q9JMR?}*FP?Du2&4~l4b#92Fvf4oIG;5cf?w90jR!gc -fWlH^8a&%Zp=j0rVIiRMLth6^bq)4Ge$s_az57e*7DNa<=64RH}(rGm}q>bhz+L&BMO)^d!HA<#N`&l -k!bxIvpOHr6RFIbj>Ide2iTf=g0lZLf^b%6alu8NP2cM}=7?q)+Vhr{k+MoE#R -w|RUmM`i+AO&pYsLsGaIRD`7GF9&qz%`%h?Z+$G?S)b74Flo(ekx%CXM_%C7EMV<}#qAQrcJknZ@W?jsL^~p%JvrDS%kW8D=rm{8@EG@@mA*kVst -ThU%jqbnjMVn=&VhT{<_nvjmiV;@C3Bb`zxoS^5SHtA%vw<~s{SE3g2a8K{YQtYS{?)g~y-)K)~n-ux -~e~~eS45Ov7ZltZ<3UTfXe7aa$0G5LE_xb1677!me%v -M(SbE7si^p~52(nqBxvcZAb=DV8T3k0w}gj>cm&1S>D~ig6J2?THjgtn~FgL`~TFiA1EDPiG|D -O<9Rj-Rm6M_k`$uYfs{MltO7s4@-aMW}$v9mkoVTI^z09rjnSpM+{52U|_^3e0OTb)F;5bPM -J~n2Ulm-)TO=d6?t8*~L(_9Gfsr{=UN{m$-IZ5bfPvv}Xb47=JDB_7v-{^V2#0q6drAUJ&-Z=ED|XPI -^b);8i%!4~pfjwYkC_se3_?%c16Tnc%QXym_^)Eq3n)FTTkS#1-N06M|gcZ>cVgm!_e-#Hsyv(B)%2h -4l>9H?VhLUj(@YVt~xnT|RH`52T!jH3q8yLSGXgb0B6t;%z75?fni0E65mYGRecr^Cr)+M%7^sG+F9T -Dl#tT;C!gfAtgd7tG-)G+bJRlrRD`8;hc~ll$w`?giAt#P|64eA;B#rLMbCmx}}$zkg|c^g!BuybOM& -ZmJT(y)S3_x*r*Z^&I<`+LING(E22IoZYk^I+oC?U=0x=+ZYk>_Q`AG2u#d^>)yJl=k4PfqYK? -eIE7&*yC -*8ruRH+64n&E=&fm3mtoDonuR?9dlL2(>`SnxVPA$l1ACSZQCRb^7GN#HT7tC? -PRCwTat0Y=d4O^df_G1=dwqYp~W~U4uoJ-7BzHVPAoL74{nJb=cR~a@;au5m**18`cJ_8?bJ|x&_;WO -<-HFZP**IZ@|6@`xdO*u)ctG2i9F!_h5Yq>pm>HroRpQ3)pvH--Ue-_Ls2lv$ejp3F`r@Em#j>ZNqv5 ->oKe+b%}4negJz5_Cwg)uphyGO!o+|cFOCYuh|mF`WDu6wvS+Ip{KB)v2BI@4c}kTUB-8g^;JquE3Q? -|!Mat=C!u|G`y2Z7Q}%wAy}cFstk&PPy>Ty##`Yn35GrL(k+t3|!D-_eYAYEC8X?WI5M>0`kfRoBkFmCcwWnBn#^0+OMCKi|%lxlDzFDy!NU0;9EjuK>%bq%QYTs{v`>XG__t7KAFdCAF`$jW^i -mM5({#<#0l`X7nS5g_wXs16?GC4KL9mFZUeZb?+uNR=7gMJ?R1?c0@FY@;bTyoBlB#YFgWz;`u>0C>( -W}(kOzYKjE`X%U7&?ou(B`%@z8ArrEIhE41w5urZ)E7l+*J8hUQwllU+}G8CdOzy+{bo#SuQ;YPH5+_ -M>5I@8pwCz1X`|_++wXwa=EfYWasA_w#=LbM`U>=Q=!V-n_!hz{oQFO~d$!`9%cydCFy)H&Cwy-xoyDrb}qEyF$~cey)vq{b$UhcLEaJa -9z*KZ?nT5UvciVLXBH7{((9qDxMFuDAfM(#^qGgfS0ep|%lfb||oV%XhJSr`lKAprEG?IO9z{#?%vsQ -{cyE`d#SGNC$%lB!O-^yxs{OhbDKtLxVbe-wpP?5B)2L*LzN{gAT7B^6Nho@(!;_0>)P`zKHSFdP>tD0flgH{GU6sd`UscmBVYKDqX1@-n&?pu2!WNs#2XFI@T -aOHSf=vJtjt)oBp)+#UtoDZqd_PX+>0(){06}hloj?S@YS8uGEec7f`K#u=WXi|IE2*{p7G%dRQzyER -af6ERafKsU(()xE{7bt}utgUed-}?8wZsb5h@H -_!Hz6m=Tsc41geDQ8^MsdLNo)Ag0}{EDXs>opslVJ-c8xBEf2`|3o=y?QWNKA5Tj4k>g&;$DupfeTkl -?hC8sgSGO(mEz55eWQ4Ds;pO1TJrbacO{Q~Sba@@y1vW;yM%Kh;k-z=AQCQ$gi9jfvPh_lgoa4CA`-5 -OgjXc*QNudPy>NTq)k)KrOIqqqiH4dEJR1>;^CD*)Fl%MaHA%ahnLOMv_LvI;HZeo?FzabB!ea6YWYaX%+(;zhHX!~J4eP2gTMt5 -MubU^Rq$IjmsZOJfD#UM9H)7p{&Mn4Y3*}{ -zvRPUR72Av9-OztI84+U)%jqp&4RBtj38igVhMG305HSrLux?FO3y|Yhj)9s23+02w^1)L1V7Yvtmk*5c!HVmFlFTU -kIe5V1k1X_+$0f+@xti}8M>L=3ZyWoqls*S@taJu_+FgJ4^DJMzQ~`MEO<%nPlK0n_0ABiX20r=;Uws -MSr#F4|#U}Xa8Nf%+Gw|0>`05!TKyUi%N!~xcxtsiX5BTcGc2oc22=eKFQ}WmQj-T=s;H@9?(+7?J^6 ->Y({z~o+z4L*WAvxpo5QxX4X|8BzKmGJn%|CHUr|PXdjOwktjOwgBjcTpDjcTlXjH*_CMinbxqq3F1Q -OPR6sAv^r#H|92m{qXRTW14p-Z~o?jG6@5Fp~fqZsKoKH1V}5nfTe1O?+%BCf+tx6Hl8O6EB-u6Azm@ -lT*I8fc`&>^aVYB%WcXIrxiIjqFPjdqt~&?Sd5YBo142mQQ^Mcw-8`po&7A2umzxnJ4c0`~9I?A@7keoG*gzTbAskbmwLO4x -B-icC1=7M>N@6=p=jWtZ?DPXq~3%SE?PU>6{Abkm4NE9?FVz`OMu4NR;SU;H^!W47jm9`Ws3i}RYl2m``q~zq2^5?r84Sy^Ab*uW{S)jje)#}~aTD?eH)r+-dy_dGA_trAKcMr-+m -foy$YI3(bzD?*495460v+1_;GTLm=XtP11%?6D&8#LM!8u^r0`4>=20|XQR000O8%K%wh=1~*U -+yDRoUjYCB8UO$QaA|NaUv_0~WN&gWa%p2|FJE(XVPk79aCudbO-lnY5Qgvh6)$^Ape^)TQ1s+QL}>6 -*dWh|;8_XBUtg!#yBx{A{Hkl{SJCiX-4_=ZDU$8#K%y_khH_%I>50iz^c(@gE(1Q{}T%0KdM>dyl+31 -r#tAjv2!!SIfc;g((WIf^!Wqi)ou?=t)-Bm(#^eN?-DD%>*D~;foWFa>yY#*Ul;h*${xU~^Slos1Hw| -*v`b3l7~+V0QdUzp}Hb>jw>$seG*`Yi^!v%gmkCUgif)FkjkYZ9bMiNY0if%PIpPoaK64JCCqNQ>I7x -NI&ZegRNR0|XQR000O8%K%whVA>K~$Subb8T*&5p6c7HU0;e@agD--g45z;89SSIQ#EdI9}Ldr7si&E!h2zyNIqD-qr(b%PFwO -V1^Kly*bPwfJ}GGFje*LA{EqM{eKT?I`nt#LKzzUt@1>b8l>AE^v9>JpFUqxQ@U3ui&~f@5ytHYG>QGy-dAx -`D*KVY0`<)&F%Cu8bzXG-O7@WlH+>c$KQScNJ*qbDNfqm?rdh-L?j4+AOM0Oq<+7D7Tqm!wh{hn#Rc< -4x=uotFF3o7lJED~ahhal%-P|GDcghgC}YbsTtiE?NY`=5(j?xpTh5~8Dit}$1{@luo2=KnTF8<4!q1 -JtNH1gX9Hzl~$rBt1Ad5U-We3lnLu=+Pqd=rtI?vr8T|Vc@-a31}k(i$clC9$R-h7<|c?1u=e!t)ANo -H7(Z&y4kp0aJ$n~QYGpfyVF6vyl^+4g#fH;xbRflb(Xn($sU#}V!-_ValvmTWR%{SC^bf6!y_BZmOAq -~e}Mx55|Ov(yi{a8x5m;^gdX(BpB&R20LH@nN> -ja=vT<&+=`|E8gIt*PC6PTztF&(d~`Lz0>oh$vc^c?>9aCvcjI-5<;-vZ!^Uhn -wg;_?LkEN0_+)Y9wqg4oY8cE<1gV5{ULC42fg;44&%ft39U;7WJ}d7P#qj1oWRnNtZu4FF}*`dO3|mV -qM;+wcWv&UVN?4~8sBbLM;=vhBbnELW)bl4yiTau5%!YM!!{;2^79+6UN=i@FW3>Y4@bErkev~m>z%+`9M$qMHH7jU!);5 -gnUkHc`IM|ODBl4A-nhEH9uhermXt|w)Z0=ZltJ7aVPg&AXk6w_g9I$5FeZi%WAie#~nyh!{ -QtSf;3K}NUR!gz2oxFAEvB@T!UxfmtRLvfc?%;FEtk2y_F2L}v&UzDuYxUkg>fR4nSYlts7FrUA83G@y(mlOK*Dq=E5nRiZTpu9` -QKLJg~=4B`K#LZf4Q1yg!uW21BFkcrCJ2d`PnhP0|XqIm~kr-L6qVD!=7k*WofDMXm$Q-XETwd6%chS -2eot{m;>Iy_6#?{M|GJaQyLdcA;jj9C?=jw0W1@oJgL5f!m*=EV_-2-lL0)vjNie-tv5whws1!!I4bV -rsDnrddlPP_g%f)1a=FPEtkNgJ?!1o1_+*EzwScc9iE?@odO?dygIP6pkHL~FhDwNGTGYL(ocd8iwL| -KTCN<7x{kJ#t#}ZDxI)lV=$}>L*`i|JQye~P@Bp7G;GlW4W -K^2(1ixsvf=ClX~s=sxJ5HiLo2<|fKgIjp&d0^q#)2_Q3`UZYawvAlU{L&Gljj$XDW#MYo>Z`*Yi0Sg -hMS`F;c2Ub12IK!Fi%aT2-WCo7GnaROYazvjtwAp^}*nnwJ>eX;hm*38*J94q+y#GfM-QTgEnv2r%NV -gt3_(WCT@^(WaazarP5!MS?>}k)hfvG=<~}G$cB|7Iz!_&ZA@SJD#gQ5&1LP{yFTlnHftc7;jl>xaO^ -#0l{SJAmACQ@ok#Mj@PPsLj&DFj)sUGLobD-l8+3zE%9OhW8&Y+MTbeSm>?SA9g^-34D;4#yR?y_5x< -GzSdFqj{!g?c-V7~K*TBH|;bM9YF}hRV>2 -SO0V!PYNw2yRpru8QLxsQ>U?t7j4SNl*buFTt_OYN~qV5(9(+9w6i%ZtaT -XYq3I-fxZe9P~mBtdheBe1?5Y9ZXWaajc+oF{I9LWQlyyf-Kfm%U$SIi4+i6f#Z+rYNP%#uX;jqBdg0 -RTD^wQSwy~a$RMvL(l&Awy#^39f1~nH!q2t{y2};Aq?_%Qt+V*4^Ry$~b-(&sPF>A==UJ8E1xX=JzVDAL; -hWfjhcAzdSE#zV3^h-;C#(X`31milgoBt^yBu94jT;&G}I};*Su*6@;`2*NznT_9zR)!tZ426tgL}e( -;NW#B3P`c15t24-%ryaN@b^IdF+P=;ZlK$6z -z5o~rsNBu)IBZvCmPx5Hz}3qAzA~F;3{HY;$v&(S97sC1H+2xq4J{^|GA_4oRwkj7607nJ6j-IQ9R9S --n8TE_Gc}l9(HDiBQv;>9tw{Gx{Fd)jJDNqPoU2^P%k!h=FcgT!GvIOO@P@znF9Q?IU1Oc-B?M2>`DO -ENZNnpNH%aFL@%RGPV+Tb>5!!WBsO?-dgI3#H%(TlvplA?ExT?znt|3*9y5~75s7A+%+t>FM;*8e(c_ -Lv#RlJ=1sxPQwKjUY=27Ycci8Mzy1b1wOU5NiFq`d@y8E}Vc{DcQKUnIduCX1n1)}Y}azRy697*Or1Q --o*7)cMsP!|1)sRc?@vz5wJ51Wo=NY=LmV|2^Yg31X0xcms=J+ep|# -=h2Dbf_T+`K%95&{d%V?(;wayeFWf|7zWBBZT}_r*g>MysxmnOXyx;+kYCEX0Sya1eb#o!9x0%FOHwv -plST>4ECIe|4b!%!d4jY(#@K-htD~ZRs_}|VrX6?p7Cy(%d*BOQCe`Xs|0abA#N%%&W$M0qV)>VI;1z -bhB_0qitaoreWzi&PJaC|+xy1Z$E?|f)RXeY=2=`O>nfigQsKw6MtT=K~qI-91+VOpvmm(Rpy;P~D5c -ufR@fxWitii>3_5UuzL&hvO}M^!$)+wO=RWNW`TJv -Xu!l41F}nqmtwXODITM#$}XcTN&nuDV!Uc}*}HTD$1^cRxkWhw6x7O!C!L7&p2!Gh0fa0zQ|xY1@r(7 -qHpk0kXL|7o*7-&)G_`FOQLt#OVUr#QSrY0LDGSYpSld;WOJGr?td^)1iyd7~-@d!@j;2`I_e!=sW78 -yz0+mrbJfBUOtRIvq!=gk?B^*iIpvu`Qg;5|r8<)m-Ie9?qU0rzB)AN&y>zPrLG#Vlkx3M4m!mK}6#$ -CHi)*>@7v7!ZX2MTLJns_>810)VI3nSYECU`?-n*73h8 -@dNq_|4YKZ^zXnv{4V@QhsS^KkpFn3@Q;Z|iT``TA8h?L@muiQCCVrdzUr7A -pQ4RkqDSts_-qH>n(5CI61HVoFuL%EmRL4Kw*9m|fR6QfThcvkAd!hG9?JGO~(`0$gt6|^kU24&*-m% -_`OICHg)VoO9)ORD@6mG*F1PM{@c}~XT`H=kO0qZL#D$7-|F1%56HZcq5RHf+FqPUM{dNk|u~JEftm>@3=r3sbZP*red*hdB~2)yp= -P=KdTdo=?05}DS7D@LW~;Z5Js!RY8ez`D2{0pEPY(2VKk2dzc3NHSJf7yS!hqVwPOoCSEWS8!PqY7(p -fk-c3|-mPD293S6G*Ux(7axZ?kksM2;_+)#cIoT0!S*&3g-ayT*~6NMPBdG_pLp)a4$e|; -7eg3d~a;z#%9`VIM{EH?zVl#1Lb!lf!esvKC&m1(})babtk^leNuqe{&|;&qUxLZ&=bE;8YpGJdk%o7 -sJ3_tjaKpVaAf++?lQfPLe}T62v$dQkQn^cPbTdOPgA*lzMIH)_ku)NiBVr8t0Jm~s811yDr6|Fg>O{IEEkS5QrlS(W5HG)r?6Giivlg7OJPh}Q$6LpsxFrawANgKV?`Bc -lt=SHXe~woA!|8vU*w*CNjYf^O-)+3CNlmRRo3xP0lM8}`>a1yMSK&Jb5#^JX|QRr+g-}g8ni7v#*xb -eu=7|8C9M1oHHK@-#*@%X=a1Jp`5JPiY^83f7zyx76ZCUZOe1djQgy27DH-lw9j&2hzm;mV@%+bBj2g -EcY9UqVw@?gO5c+cKL`!DFi!cSq41*J<>XgAzkKyUm=Oob@sJQq<-o_Cv3NcHe-6gME1V1i7c -Y9SeuR|c?VG1lGJ=Zp0SGc)z4Di9XHyYii0S*m!FhS{aS&s{^bgp9iCWd3|a9}3jGLVuc$HQRaMhW>R -f1Y$rkD6VY$oz>Z{7->q%w&WT~{v1lXpsmLi>OR{6AMMU3JcEt_6d_~4Hvhy2J&AFJzAX)XvhRY?we& -Guiugm|~8$dDM9>gC`VU*3n{JV7Y!1>3$woO-2?b_-iy-co1nc -72o`6Rl2e`&aOL^R@mz=TeU*3LaQ9Aysm+H=~Zg#QU`0FBD|EA9|ohZk9pqGj}p&QA0V^eKPy?beY;D -Q4i;bUk{2d4qUf6x%>^n{I%sXwU^!JK(4Yi#D?P`rK51W&JN@i0S23qgm?5;hrPKeY`Bqa`zkHCoz+M!6qH0aaMsXn%LkM -OL_g=%joaO>KkK9TT%4Y47Ufq*NH+JPY{KkB28}w{^bTods -k8saU4quJ;Pi~q`-!)81Zm$@aVufO-QJJ7tHZrte#p-`tbABQ-KwayN@ -PUpeCzTI*Fi6?T`h^=fxqlsjm_hMZ#mQ#t#(@?#`IOaiGlP1*j_r}Ipo%bZHSo!eVBxetAuYN;V%&Jk2{IDON9FE%p(POt`26K1 -qN=U-Wv#>F~TR@p;e9r>(jq3!6_<552Fk{uP?sqtGbzqeOkEl?ldnoAuZz6PMZ?CoDDGf|4>T<1QY-O -00;of09jj17JZz&0ssJo2LJ#Z0001RX>c!Jc4cm4Z*nhkX=7+FUw3J4WN&wKE^v9RR9kP;Fcf~zudsr -|5@|$(2_7b;qKs}-$_nbrc(mN)WHlslWINp|#DC|+j?;wEib?pPbV7if@E_Gbx1({DsiQ82R=? -Z;!s_bwsy-3K>bBZX#-+QAbOcR?x?uEv6rm8NHh?S6`(k!$8gy6t8!XF -6eyg)j46zTS3EZ=E*ENNKZFLA29#g0h#Z+s_*Nq0N9;SFo?pV=SMQ90n(N?=yv-C$>g9ofNt4=g2kXx -d$^sfJqTVtO=PVPa|jwiwU5_!Gl(q -I|5w&y^2GOY-&lx9^5MC1p|q9Ka@zA^=;-r0fB`%TiGtwvG1V*+P^xW=Jg{=@7HY3SC_lc%tkaVw5cVR>LM1pkAP&mS*!wz$sFxXpH`TQ|JO$OLt7&)%Jn!i3Klm937q`gD&r?r%qi$U}xGtyl?^nr%y -I$q*a$M~mDo3ET0N)A#03`qb0B~t=FJE?LZe(wAFLG&PXfI!PX>Me1cXMBIWo~3;a%FNZaCxOxTa -ThJ6n^JdocM$ZOPtv^6E|@kc89DZ6U6nwH4RG9W`MG79mjvaZ2>Qfw^1Gt$~oWVoE|V{GzGEANwxrz= -9CF=1!mza!4?eh-xbd0@+w&6YGEF7VHmN>$w_Xbcbr9(*^JDWHd>SH>m!=ejMD^HV12v@=bJH+acdr6 -mReBW^eAN!$wGno2}F(;Db7R%)uWkUH0jYuDk*P*O|Et7K*KO197Cchl^jKbm_)?PoUP&*v-Z~Aq)eJ -xAXn0MU1<}(e1ieaaN7WRG#ZVU@C|ba^PJI~5vjcLqU31T-B(u@DM>iYLXIP#8Hgo@(g_Dr+!tbOLj! -l&f)^C(C?0`X)7ji<2O3wNIS2}}=I>ZmNM4Rr6LvD}Ev_;NX5C~PMoOqckqKJe5mR=Fr%}-|Lxw$A9_A|@qRlf@ak;pKo)#huu)`khD|-s -lY=25nb4=Hb)ctt6i$~nrSBNcSBz7J1U2h(Z8EyN5-p=ymjn9>FpGW%3?qK@i`QxttbS0S_rgg}oJNUkLd%r -$GI1A@*cQTfPhk9Or^m3xANKe2j&CXBYt;2^3aI@r4V|K&Ncb^!*gAa -VTe&_V71_1(dVVU`U^)yw++c1SG+QU_Ycxwf|C=bK|B|BxUWql9Gmh`Fu73iRPBP%c6yJ_GxB73>ML% -WZZ7GfI>QpRNUBz{MVTUSlSLMdBM;*UrP-C^+v%xEqMfW -_G43G2gp-Poj=RHBuKx9sU#*U!K^YzIM*hyg<@+j8Q-^_X>eo1MKdBXiRg8N8)-9Y!xOxP7}9FnGc+y -di~3EE7B)5e)NjjRZqD#x3KF5Xd1m+Ht8d=9YClcahYIIOs5;lll`;+FiEILi`G0QGk<{d=WC`6T*@@ -Rz)IcEgrjqD^GxH{F|`xNIGFN9)pP?t6UC8u3Iyw|CzB8Qv({EqXC47Qtluel#K-J(r?nIGxKPB^ATt -|?*v34-Rz3d!fS*Ng-SsJXk*+3cDmJ`nN`V}GZ^Nk$Xe$blL^W_hv(&5Ksb>-xYS8cH$ -wCZq)Os9vGGM*q$*8;CAyuP-Q4-9o9or1efK$(2!^Vz|@)n^Q%!RJR)nQf?H^e1QSXH+U@{F=oR!tOh -N-+oP_@aE1X4vAsEbF1#U;C015VBAX5@5%poAeDRSVOGZ=z&52S}1OQ;=9{zD>3WV!)OZs$YWB*|;AT -URcChy$L2I^9NIuMOJpeJ&ODho@V?)`niVodJpFb7Y}|^$X%1jgTLJqDqLCxCu@{tQnJ?o -dysn?{U5>kv2b=X(RBur*Qh)mBFlkjmN3)8?oHn~sySG^3;5&3!;xp71ReFv!lC|1RzvLqI`%UC`f__ -zvXXi;Sfi6HNpsGSLj@UjZ0`#3`fkz*ep9+kmwoDDlYO|Xt#z$Ekc#33svVV*!VSf>yQ&8tbWE}=!lOyIO4*2W -7rth;y7thu*n*jOd((Iiqtf=F!AbkSImJOHR#HO$GQY_O)7Yt8*5owC?_FV1Zr0?@$pSupSjIROww07 -YoFfj4gNX96Tag+7;-;5^jmY-?e)Z~$f%VzTj1kc)gQp?nN(P~^2tz(t=<$H2Cx}=wfL4F;)eCR8Ei9 -%Pa$2bsHC`3W-JmaCCFwrOYZtKK3&A83%E&_#U}J&$iad+9-150-_R~=Hwxdui=Da#(D&G0;jnp5Lzk -VdJFvPmQ75%Tnz>1PO{YIyj+3Pz^XeL~bfmyN1*C~0%~QNi_xmi2A`CBz!X7n!&W9=Q!G#Mhhx;P;$B -DlkgTH}pt?tJc{m-DjgwkBU9&Y*0+^X(b)qFRD`-W?UTj!sB?s>-dVz@L7NAm7Hm1>iH2xm>PKb{=i- -+t-h%dcFz{OZnYufOr;mA9_GeeL@0JMX^t{*4b9xrQFqqs?s!;ReWq$^EdyfK2;jy62bZ<^@cxjbIm2 -2-oBC@c8P)jf^f2e|hrLH%~tLwmka2oPJ$Se=LuFEvJ8%M}L&lhvoG5CtrQ~_`zT0^viPkGmcNcE2qa -~a9kcfEDs-qi`-6Fd(#p?X6uC)&prC{c!Jc4cm4Z*nhkX=7+FVPa!0aCwbV%TB{E5WMFrR^gIJB_9x}2&lx71LAHu?u -M8qUd!Gh`uD7zhVZI{4^iyd*_|1?4>W=94A{w2V7R;A6-8l29V0whGd!W&W6<4*l|Xe{fUL4C@0_qA2 -W=QS9T8?qPa)GFeDY#bWYQxs0uP$yg8{)ta0Cu<4@`jXEDV9ir=+U35~ -y`qSSN1qJDO#2}Ld>|%6V*?K&5a0FWM{L_!SX$D~M(ZMLXRw-1YVNxEk-^#jvJAOMBQ`#W&XrXcOG@+ -4Z8kI^546t*d+b4}cMpck5xbzS>SwmL>uD7#s{bYUW -Voi5n=DN=$XOq*PghM0KG`P5f{NTEsX8k -XB%9^kY5&#^IV^RBA%b`JT=EhX&N1zo89t1ws22|f_5DSJ7Y!Cn*ybV)$}-|1!fN96}lO9KQH000080 -LuVbTT1)I;(!PM08}Ud02u%P0B~t=FJE?LZe(wAFLG&PXfI)GX=iROaCz+-Ymd`L@Oys6*8OlcG6D5N -KSVVux>8D89s){LQRLf;H^~;iWN(g35&yk2`(E401@2HEY9vVPotd52&TDVgH5K84b)sp?p75$|Izg5 -j%Inz**Tk;n_1Y}F{O#F;{Pc$hr$0S<@-QWj0Fj=TEG17{!JC?vDftBH=gX$ziBt1J1N^ -aVcYtA*e^if;`K5)Z811(mGZ2r}W5sHr0eZH^Cj;<8J(@+aD=qE|e0aHQ*=)vFp)`vbfsbRlaagD7F* -QGwuzgCNv${|kK$_Wp=@#g}XJ>MkvopebQa8d_1$D?STecLeAQyW4?-W&Q7ha|O(CiV;)2dns5o$wlr$h9fS3LHtFLTlthL50M -~(|d{XMfV{&33f<)5$q|~{v`dC)`eTJ;_VC2HzJu|{`Dzid_02ui)bkXp&iej1a&~Qn|luSFOe}FsIc5r|ZO!_2*JU@SlZC8_(%7*d#f#{&6l)QJMqS -kbD_*Q1vN**t}8+w?t-m>gJHd6c-gMiX)IE)HK|GUh!Tt+GJdQM||^L$1p -KEB-Kx&s?ku%#l*Li`Bm~7I_(U;dlB -nV#k_AzXglOmY@|+-P^BrcRgXbiGm-Zj+Ok-pjyJLNL%=ko=UrN5eq)N9TDo3QkTZ2|$l$;jQs$ -ZNhN4=yRR$O4EpJG(3fncltz+Ll=H -RVsXI3ZAm=&Nc)YsIMSkAbvxHH*D(&ERC2^V`Lw+p%u&vva{3yhI6nkK%pu;rBY46xDOcfdRldt=a(Fp(SCAcYOBovf0*95<(?-WSo}>SfB21_r~IOwEhNa;k}uP -3gleZj*rolagSD$6KR#A~C0dCH?J^WPCQqZ0q^fm~)WFvK{Z^JDCpGS4!=Of20N`9m&^#r97x4&1Xu5 -R}u}>ItCvqx`$Q(g4~3*^#Jj(wy)_Vd$chnxmS83)3ZawC(uX7Np-^G9`33V_29;)Se6$~8aIF$MMy7 -x2Y{oFPPb1t*3o&_r%nr%dkNBbAgGIYY3@H8bMZOotDIQ6r}pUC%SkgiI8`&S>03$zRPI!c?t@-lUjh -9}yV~*oasxq7Z2>G;Dzi -4L)0wA(X>nH*rqIo02uu&N3w0h^ObR%;FhMZh5WPv(4wJY6`C -N@b~z53U%#p<4FHYIRhpP6Ld)0q`;>_NZR=^3-(Fw$N9!HHR9O0c##-fuW9NE+~cmiwTvZFg2|}2wAe -_c?DHfo=38#YdgBG=mMezhy>|N(vnJQc{|T9V3($wqb~OTq0>W1DD45f4V;$}&y4%St>+Mxe{io4Ah) -z;r9pd!56VV9(QHUhhX})PW_XRYffdnA(Q}rSf`su;ruwa*9j0Vd@uc3a9V;R;H -_i-Ee~g^W{sK@-0|XQR000O8%K%wh9WLn?vj+eG2^#O}mg8LPt~<@N)5QSWnPL(YXdOdPEZQSPmIR7U>fN@_vd^+l)^m<{wG~g%A4Rb -!5yaxr;o)~KBtP2uwyuO^g7@=~uhD7Q9jK<-X3#jehr9a*^H_e8AqMTb| -wSYJ7fM%e*SnqSp?IR#j06J>)Cws27dg7hJ!%!H#-?5VZUjo_jsPleBDdHV4pPa9igE4|)um==b}*;K -l6jr}%y`8opSDKkM$@GKfBT5z^zG6$gvp%i)iAcf(})^WB2Re6nD}&*1-;!(W$S*aMS;);N#8fA#L=+ -nZN|UfApPRz=!0EV%|Kx77nL1M7Dgm1Mk -$S|_f>rfH2}GqbV)XPvQYNGw^y$^uLHcx2;;oEMp6#lXYyC5D(uj4{SEAV32Q)BqzAY6nzIt0u<)L1a -eH1oW9$l6Vc3x=rQkJ$kQ-E?VbhmP71pOhb#ck*Saa#vwNIc4*YpX&H<{qXKggUZb!P)vj&=tC6o;JM -0VNfSm!`9wX!yo`JG31IWVCHww;sc3K&mL)`>@|J6bC(4kLRe58VT%*!3`FdM;`<|&u&gZ{Q39$PVk6hg^XKGP2@np(Im~-~xW -oD&yFlN}C!ZzOs@GFf#mfp`WRD9^&ytNe^L<&(#YT`}o2YnprQE)c0b1rNq7~dzog3MOkzE+srIB448KQ;_IfX+&CmSI{?H!<=#ID!?pNw6x!9F>2#R -mRl;);z2lXF*WoS0m=V&li;(iIz5Cev2zdvlJk*8;X>EkMA$T;*a_7#X2`^y^z!5DliPI~=7;>0sp4= -KM;7o7$XDjm+jeHesIHq{k-6Q=9eJgm`Mx9-9D9ZQf%O-lDT061WlDv=>7C -{h<#XK>-}p>G*m?*wq3I9vc< -@`{2;F;#-4L|7yTmjX)q5hJvNk2zb5De&yuH48K2dJer|ya1ate?xn`zv7*FK_?>8sz~G2=UuFkf?B6 -=WHdPUHpdOCr22e-@PLuUc(#l{LlQ>|`e3Hh3bIr^zIz^Q-j-xL(XP&5>%$fD5YXJuJC79mr1Kv$yO* -jR6o3s089bu55-OOAwPWz>O<4&X`3U?>sSVgYU!}W-vnTjVONmzRIkMfHm$HM$<6byti_Sp6A`Hgo=L -48S1_|nh7(RL;YR*R8F~(v{^vQsC0#&#tazg4+-ema(G(%=A%+1Z$ZOh|Qf?Xrg{?J8@I>!?tIE~E?> -MjKUra_FJFP4vi$u|5ry1IGAlz0bRmpWCCM(ABLPC+31$92+y;Zp+wO0405H^l-cUVaD|#N+Jh6q^72 -_Zuiqyy7~0*B3u603i{VO&lxp8ZKtd%P;am0*5QC?@JRA4vJLx3B?}(Fx2t$%=k*X&B87{M0q?>Iz11bi|f8>HDP` -@nXmD{Ttw^z%(+r{5U)K!VD#IUTK3$nhgHZp^23v!SudbyXpdHvSC$O@ij!8)5Er=mO6J?hvZ1bUegp -^cR00KXxa*t%}*M^6HRAE0)47%vp6@4@9BzlVM*1B_fMHjj&Vwj2-$g-QQ9Q#O_T8Ro0D+iK`{b1e=8 -#|zc`R_wUvbRyaxJ?0wZhD#u)J067U)*1A7ay`6KG_V#(EWNQBm2<^a(bg7xbsIPZGir@G7hCBL6N-s@&4acQfxr)wv?}cZni>dc)rzC@SLUaisMHVRfc9<2qt7p^Vn?m7hDNOx+}X)+Ybk)Qg}glN -#qNnYNi|3@9OPNlRfG;ZhJxRC~qh>Bv-ZEgI~DqhVT|bk(JrC1Xpg`Ug}*Z*p=Bt(Gk2{6GY7M&q$?D -az1sX5Lhh*2AUNCYxpQ8^!^!D5I-7m{^m1`_{`~Q -?#q4}`F}r+u1yh)h{`vdIQ3?mkpF_)E?Q$K>E}smcfus1tEx)_vH@E!imS5cRvs->btgl|3pF1$bup& -fJRIpqrILl@1yClgXz~5*za;{X61A3U2uw5Ph|VbG)I?QW4Ib)rlB!2I8fT -eb*n|%H_`{uCt0B3*aq~}>K9(PsGXVsi0yX@50a3?1SJws9^+SPF?tj(?50lX-ov(@OKy$^db -fzv*lyZSHAPmr-oK2h#IRBn?ngHl@%U*)@M@~nuX>A{MTW$pc -vkya@=#iFUx_CLrE+c%5gYD%?~&|8B|^Q#%RLy1Sd!y8@|LV5n=APs@s?6hTfedkX_@c10JplgB77)d -YcLPe<{7*czw_y!m8_0#$C-#z4{y0u6s*!KxNCc5B)GbEX#>kmj4A%O9KQH000080LuVbTkfoCU|b3S -0KO>z02crN0B~t=FJE?LZe(wAFLG&PXfI-KcrI{xU)YCr*oam-8Pi=Wj3Pe^}1n2bL%$zKj3ya>Bh)O8Xfx{%bk^l&EeTo|U<$qK3KP(}?3R -lywr2p3>&T{l#+rV?ZorI8}(Z;z^?H#2?tg>_o=xF550_oqBKJh|MBwcXDPg84AJd<1q9)EF|}NKJ+| -{XW-m1lP4k}iL(TJr?@o_hJk)HwT#~qZ|Wq`56p6w{9U4| -G~Pf2LA>tAPkVZaNsu%qK5lpc0B=?Yj+iOQRN81UyhblFsw=hfu}b$L-;UQ(Br)#YnSzLP}0kVr`Eb} -sRmJRow_662STA3MAB;wi3_e7B>KPI)ZF(DQX#MfPS{b{=2&v --3Ms8}KNNE80)OSbR4OBmh{!HH^a~S;atz3x>81{`q!GL!Zr~Tm}$IenCpfEl8?(cPYX!l- -Z@siFu-nwrS@>dS5P69(yPdYxIc*Aq|OyL4|j8sq<%9V)S{>`oJ2k!mB-p>C01NY$Gjf2C%;l0Db{@#Hz;SwD -Fh>CB!P7xJPmmGph!0l8BRe7$_wso*Q7?i`^2y{^Cg40FdV#enqoCd{gsRr?&yx>S8J`|p>FFD4kOOL -~NL`x6~394?G4B$Q}W5Ra+otZfGd>j{)E>AqkMj@yGGuIl~Ll80z6Uh=W6~3q9F&Bf=3J|9-t`qLD7? -Q36yi*C%fkL0gBgypq7=)a|AWM;q1wbU93}eh`Fi^@yn9UnWY9Jrg$5m?no!za$o_njmwbS3Ji3yd0X -jIIRgXK!vcqT^j1kzKsvOf}md|1tRp!+75WDxYlk@3)5knr -P6!5OAY5qDj~l(8FHP|md9eiBNHcVO2$6mCuvj2R5o9o+5u2kPORcuC5G>{vhAM7tQU_u>%(pwXvbb$ -pP?E=q1`x^@V^j?wSE-8;)kc7PFcG>bsL>jQNJ82_wysW{}1n;sV#!> -?fMxo7;+J1I*opW<^q3vBaK$dik2?&v-wM!`QZzk<%H}*kjH%FvvDKYy*UC;5uppMQ>Q`CzH9ssgzIZ -((e<>;+n^=oX5%A?X^%}K9Q?E)0Jzy>i+*$=C#e~ldANoBmcW%Z|;ig>%G27zA#c+u<@VDE;xAHqNRS -9mZv@ZtLndX?fUg=-wssLvmUd@kBf)v*B-Ca(#V#IRwbUcAU##`0U4p$tb=o{0nVB&7qT_Oad-Cb-Pr -BZGW$}yyufVr%=i*o!i8=1oE?wJ(#V?kN_m_*`Ef*g@O&e|x_6 -E1NcB{Y{TQ0t2c~^DTufhFWs&VnfrFhoqh^qg}Qh*sk^zUmA8oYcF*nceN?=0v4UM~Lr)q>5it_+FQV -3`Vo743v*6=9V(4+gjI?AFXR{G)YYW%+M6wO*dJouO#-&y#?1;X~mK6B6~WxvzdQKW -{@&r%?m-Qf)uvyvb!~RQb$9>X-cG+RwJrXbrnbeO{`cv5@#i$PEj~|yH;n3X{zDp)zbxlJ(j>kI65oS -=8tD_{TYP^x|NRyC*6sWuf%)fh{!<;c+uylaTM-rZ&#TzqHN}3EdF{7X;9Iw|!v19$ds8J`*Y6*b>%O -;~fA%yhp~~5HVLqpn{Ojf7LoM7sfVBDn{L@Gu5g$y4Pp-hXZs(s6-)GCk&kfv70h5CNv;rRdVaWYuR_ -T6q1-^AVEAYQ5z~A1w`|8gA{k;p4Z)s`%xN*_G7F#O1ZY`Q^CIZ}7-1Z^_L`U0nfUdojIuo#eiebA0_ -hZ;SC-n9v)DQ5}POVk1JHX^O2}Yh})uK{it6Tb{G{s=nFZHBaGLj}uPb^islAh7u%IsHfMAVljXdmpT -MV!7r@OhGG1}ClNBmCUXI%Q-@&04z76$XRhiwm7qvjI|DG%^>z>v7${n{wBsBP`dQ;wvH7>f#IFDE1~ -(uM4|e*#DpMm?SoR^h4(Z#YaE%xS~Vk1>Dcr@8CnL;C8YGJU;M}{G$@pSpR{M#vt*KB>w^mAM{54Nak -&DuheETdzm;p=EE=^(IiRFW^U|%vyX#%L{BuU$d}qVxFT#;DEbKKjcT;3lqUXkY->nb2-LMU>8IIKr8 -F=qB|2SOzv<+3W*eN9uKCogQp@TU{3@l=JafTgjYKf%rQiSH=@dja%|SexN+2xx -$fi)!j-^yY_f#;KzE3#`&s;oz<#DPa1wRr~tCPNqbyFIiS1Gy%Di_x_^*I35Rp!{%Ey{lEIeZE93TcD -;H9Bl+9X4w_;5#d8oBKX;()DUw>dp6rOHax;f1OUAEwf4?l{;bn+hYrlM&HbM-ztWG>P=v{Jxx<~3`& -ToDAItC)Vi2R`P0yNCT~r=Q|>0G6Pt9y4e0m}P)h>@6aWAK2ms3fSzF!mxc@2#000UU000>P003}la4 -%nWWo~3|axZdeV`wj9Wo&G7E^v9xSX*x!MHGJbuQ+mFc8QbPrlrA(BSM;#NKufQpgvgEcC#LPm#%k~o -!Q2&cqo*&Rze^um5K^+dFTs@NDa3trxNJac8wjA4WnSL(Ep%)C&hnu$V+lM|&lvgZ=`M!1J^`x5_*xcIPwXW= -3Ja}WjQaJ;c?O?zJ@IJ(W=kOi_9`(GyMvv09C4>?0#SQ}-8p~u2&>92BwIkb+EJ6zhu7}121Uy(~BiD -|F^sciS9L5qHew{@EhFs9^!yymg3KOUop_Dc@&02P{MJM%ob9*~Kt -k++@P}k?@R-4+xrHHX>a27Vlk>~rKlgT#E*R^c({Pw1neVvb>7uiP`2m*qCFrN)sAVm!Zj1QTNJO_kp -hfL&VR#P*p6pAtv+31Q@y2RK92quZJ96?Q$9T80eefh{x%2f&7eFP3u;|2@RhM`<)9O;vjnx6D|1eTC -?gt%KQ=pp!U3Rxw+%o-^T?a*-Cs0IjMg-VgV4sLL;HFwgz0fEIRT8gKf-hM8y-97AO`;D=VQ+2Yq&lH11Qnl~7#qgos1*!(^pWc` -d6+<6QmKSIkO}AvyGSPuuocU`RHSExLyZ%j=ZHEp_E;wR#EJ#NZ%^hlhw35sOL!ulP>vMTmjq(DOv-+ -hZUrD;1@w$__Kgp3oBK86_Y8;1y>&sE9{*^o!$71)tfu@R%1L=LnQ6jqk6ra@l? -DqoGd_LTVEhz6pSu}wjNL04l?ne8|Nj#XGFV{^&NrGfu$L>q{4c00)aTG1(ZhCLrwi -1L?CkWSRc)3O2s%6OKyyjTTVlf0M%JyT^yZVbH${?xbE1Ls8k#0>$B9D^2|FC?X;3g(q>M&<$k7~TCQ -kbNo*Gf>3K`uj=hBi{qOkIMUWK6B`MQVne -!`U6}8gXzw)LW%fdT|kvrd#Vd07Syw~B;0w(tH!tlRs8o77)lOFCpB}r&5&rPr9`?}h+1O}Q+}o`r4^ -OpN4EYVQkVX_ImLqTAN` -=?L8V$Li^waK^$WbBTRK7A?2nimIuu{OX=}rj{a77IDX^HNlU;-ss!WXCBv|eYq(KSh`IX5p7p%zMMB -%kBv6e9~X*XJo}77&0_dPd3{NhOPi%R^NmLhtN(p=wg7Y*ew-AENChm?3Z?|Guj9Li8VUwQU}|p*mk?Y(P)Q~$NT2(?f)@-ynNLjX29elHjU<;LE4pD80+{M -0=Avo2p%5TB-I!)>n)ItA4khf`B712nbY-6j^?DW1oW-RG6AtfH2BvU>bj(@{_6nqR!4o}YiI(J@HzQ -HK?RL6JPqMME2R$obN0#ENt(8{vc~L}i%eRFriBVOgqB@l*N3C?U*7f1c`}b>(ebO -bjw*DvDI-$c~B~)1yKo)e7t(eWiqrt7=(8+Y2nK+M-q;M)qK6w;v?_G_?O*avjT3WhAMfP&IrF09g4T*D=6My0 -&%{3~(e>mE**cdO=9DAf^-?!c`44Nyx11QY-O00;of09jjBE8c6z5&!@rM*sjB0001RX>c!Jc4cm4Z* -nhkX=7+FV{dG4axQRrLm8Dy;NaHkvC!q$i^-WNu!6sJY -GTCqS*vq82ghw7&?UO~3rm>kBXZv6gE;f?Ki^DET)_U;zUw^v{p8b6J?4Mu0deITD!mV6h?{~56h}XL -!N|P||2m-WtdN{ns}dplFehL6J -p3ztHraA@G`E3NMqpOb=hld>^Bt!U+nelEl}VwZ*MZFnw{6x9DgW-S4qQ`D~&^csN<_Q#^c`B=1CEu{c4M-z8BBh -p6eE1?VH7Z%;D@(ZVdH+y(h9uQI|B&D}Cxu4NV&!jn3DfbcWh)T|iU2+*3y$sKqm_a#*)20z(XV$PdP8pq9PE6bkZElPfaAZ_ -f4G|dIg5p}ENyAz+pxJ|z8SS@$J&waOixnh2?pXjNFm^P>J7?<144L|NL>ob~wcBTq7ZEmgK`88aVtH -clAxa^GB3LUW&BA0&lV_t+%){^uk!A)GRv=2ga_Nb&J)|*doM)n)Crb~IDYdaC+60SDm=&4avlC86w) -Dm$aOg}c21|{=aovfrMj>Ua_vGZP1|^2HG8bufVhVZ2S*OePp*X%42y(mPVENtWoT6XwYd=@pQ}45``VZ -RCSjn%eB$Zo=D$5H$l(DuutNTx7AU+F4~XBhc99!38PNoRHlgO^%Gxpf(HdbDA)1&LgB?llfv(C91-S2#UTKvW}ZwQEjBO^E8xIdo`g5~jog!w#?8x<3J0 -#JwjWtl^>2}O(nKDXQq39F{Fbq`Xb`3tFsXVE&p)ez8GGhB6+ -1-Vafm>X9+)&(Mn;B@Do#j1CxA00f}4Cz0gMT_RRGjxVT|y;lN`#R&PS?}OLS$jrKxx<_cjMZ=7C|R^ -a_(@88Qd`N+2p(?7KloD{x@)T?v_59?1|Bb;0onVzQncFnPpe4deU_H)j)eeg=*m=VF-4sPT0k)ZEhE -V31~)iXsM?N*IddJc?N^b40}5(C_sKR+dj;158q-Fc{huRTUx+X9O8qI8_L&$1IHO<*FdS8WYG)Ff|Z -h%?M1>R0xQNjPgOyHMXZi`HdYk`Sj6hcR(uU44%Im8EbhA4y03u%i&!1oy0 -3vCQxL0TTlX~(V3BFZw(e^nz#`L*ZQa*E@-QRE9NW4tL7&|o>2G0{%gZcFGyn4ALQ+TH>XmM!>XRiwt5=k`DRdy7cE^9lU^Q@=C( -+{9=TW|k!$VF=2p7fP7zk)z!F+-3vO@Wpp2*iHkPBmT$MOarDhT!a6+NyR>j>6~H{_e$~@4tb;Bj*IBwxmggh`3|s?IB3JsQ3Vk)lgvciU*he4Wn2{)vOf;-GX -Bk*7vUToyAKwJ~4v!F)74z*<4s#pJQ%vZ_)UwZ|R&G~3=Px-9P;7Stml0a{D73Q{icidkHeQ)3{v5_R -Biw|GTOM{7L<*L?S7g_U$c|dYMMZau6wZ{8dy}=w96FzLZ40#Q-L7uaa}u{nz4K!?_!f~PzGt?COBb^ -OzGz{^JY)d(niC~!&aK`^&|@+_zO>n-$Cu{FeBa8ocPrjnlixpx7abA5lohf^H-sZ%!0)C!o)vXd{-v -B$-X%H4M=t55j>jaFW?oKfq}*L{{Y4Drm@XS3b8)>za< -~Q8vkXr3P->|+9TfXiFGa<-xtKn=i7@u%ks3Q0+|+c;m2KSV>iI?%cTy&z_XJKGz|JDx=aAqZ6gP;By -i@!|2^O>q>BfeNhCz6@c*9C)zVbMGd}5|*6&f0Q?$>t!T1Lt7gXmA#G!2`!QpoX?xhkcq72R?3ser*2 -G}XCh$58KnYno(PR^x2c(5~Pim+fVKI$bGj-)5kFGKyz%*%8#M1iyjW1=et)`Xh|aYos^7%SCao$;rx -!pYBF}?6652Lc39{M!(l8a0q3KfN>Dp?Vyx)2ZNwba;ophc)NYd@V -f4W6^FoC_m<<8cqQqzc=U$}x)vD(KHe -^`*M*8sjtKv}=q6^{fph6%11iz|y)FTcDudA*3-v6FG@n+ -xapiq-9Rr&zN|zz#t4JMByc%nRk#f#ps=q+g)-J@Xbg_`y<$eRJgHdzhn>)vsy3H__XbNw8=Y#hhFoK -FP20l0+HNr>d6NXo`?kK9}zmLdh_@UZbX{NDUn_Emx-Nnn==SS=mpV!1cW=6XeDYpBZTEsuh1s8zglQ -yn*RM@YvJHBjR+_NLhS)uecS3g!JpO-k@d>2QB?{&PIs$-e><)pUw(_9Q6s`45depa6RxPuG|m4GfW? -Bb%n*Ss2;s(o~A#{~}wrb#|z|J;|5Yu<+&KaNYaXeV0Pr4m(2x;FhXh1H6%2Iu`BsO<~(@V(g?yJPBv -y_C7nU9lt0E=t;T74jC6X-9Os-S>?>SWJ3C=8JHLUITPbbzlQNDL3?&!Gyu%`m(+^oOw;hVQ)0^I@P$ -=K)kVEU_zaBN^d&E-_uJ#o;uVF15*C5*Ab&$+le$&X4an$I$}0xbORqW?;|Dgy;-j=E-j)F+7A)Dxo7 -kDAo*A8oPqOAk2v2H%4%Cvb$LqT{&FdQl_&QC-S|o04I6gwSffA16Gk(+Y8NaLe{I -HO;Oz)C)7A~cVMtn7KZeKneJw80wbIlP#U2GI7UTE_7yQtWRm4st~`psLn+9WPJc;+(1jy;;Ec4H92$ -p-!nu$_EB%K!)55VOzj<|yx;U?P3*RhqwY7enn4`{t&C_!Yb)*mOh;qT)Tg`6;0Pojtdkq}dh<>%W23 -ow?tLmmoq2`Uvs#MpskG5)|Ga|06Q7#~wT_jV}{Grik6OQ2`JZdfk3Q-H20=DRe?V7T-Z`?e^L=G@!w -lbJ434!B6F^)}Pd{C}?OZg9ct2$P9-420FQX=wt|AxMIB&vF;Upb4%N43T2(9_u} -Y{jHUP|F&m?H@%s<>odN;!joV*50ZN)v5>xOeYxI7{Nw<^x -gvNX0t%qneklnGR-`b<8tXz?`NH*Rj)VV}WJt9I+kfl9ls9B}(}{QvrFPsJZA=?1u7(w$MsQ~%uINwp -{kHE5RRd9Z@&SKTGyugtsS)^XiF#l-}^7*byKv06je7aLig)U!Sm3-IrzulrG4xEB(&j@35Pq)nV?@&)?|~c!=d({}F(r^FhfaGXGl -d81B#VV~I+?%Q}V$PF!N@-;1_e9NP&_#%W+&1xUWg= -dTb@n|kqTlXdCMV4v7%xBeSYO9KQH000080LuVbTWLfMPLcrt03HMY03iSX0B~t=FJE?LZe(wAFLG&P -XfI=LY;SU3baH8MY-MyVaCxOuyKcfj5bXUGBc}n$aU3EVq(lJww-m-|R3S(VpGc&op<4#l79eA*x@B6bA!B7$)Gej -yttP(7W4PmJ*w<5(Abc)!XEmLR_o`YJ<0J-8Xo+%{_#6BPbYxMekm{?`xfKuw(oFON2^DC_AI)jeRgk -C_p;>XhHk__+MDWMcA+IA``(k`waP_d~KITB~Y;1$)T!F7JZKYo6;#k-;rbqwPapbPIFLfv}jjZsMbt -G(nF^|pt79faFDU$VO>)?|yzVrM6zR^W6pLaNO==FmG8Opm!J8 -EW7@E9=03)-M0{m^+QnmwyV@1x)&WQ@lfO+lr+ZO=)R@6aWAK2ms3fSzDe=Qt^}r007 -$?000{R003}la4%nWWo~3|axZdeV`wj9Z)|mKZgVbhd8JufZ`(E$e)q2+JQScB#cf^uLcD#Sspj>sbv? -^M&21IOXaL~UqI%R8o;&XIt(I&lR%3|&37qCPq59bl&9SD>@|D@SVWl|L -Zs#p7QuFTvSLdd%TnCW6DY@f&ZshOx>3|0lX6BvHS|@`HE;F;}ZY$)p6q%f6~54T; -HiMc}huukVCxOZG&%!6e|>n$(;rj0fDgS!w3m?j+4N{Dl0}IdK;ocAsp?ZT&>*zIp&5n%WIqFa_Vcv;FApT14WO`r^aY#DIU2S}U<^Jhb_!sFIsUzSF*81ffZ%0OoRPY;qz0dKpG -)hCWFgiVG^if?>4<07iz$&aFwY7X2aA!M~ACZU@aQ63tlU_~MM&){cuW;H4KQ6a4pr?W~fo~l`L)#Q5 -5^w_@@>s&CJbM+2#M$<`A@QdiCa!`n_8C|}xvgd@Z@#J+%)*5)NgJ^w#ROa0}86xTkUu-?R6i!Mg$Yb -g@Nu~u&$8o=%RwQ6GA?%r7r*UcqCqj-+!A -;0-33?NV+9_!qzDvX~3YU1Im|0@dt`7ta8pc#$?U8`l`7h>7$n`tD+(u5>gf%TACA2acJ_0}Q_DQ_Ol -&{Q~GbXtTlQ|#7WLr)$Yvd78=X_>N9FwB-<UHw23k@i>)2AK4W7|j(3>Q&y2x+w1c9#HyPET9)$RI -)U;e$Zp0$$Q~*~#E}RoU^-&;fd`aSw{*fGCj;0B5er~oR_I -z|>kaJt;TxP+H!3QBBQaIq`JjWiZBrA%DalQ*N{NkZTlzMA-Lk`H8fKqQSDm4`Ps)ltQrvG*JtGium) -6!#o8egu2M)vmc&7&(TS0?i8wVecYcur#^9BlC@;a<8iH~ZSw)4tQqQD!>r5m8m_}NG|_4Wvh9!3!uW -@mzGimI=k@$R#Z^(Aa6IRkOg4--PNC$C2qaujbTOg(&)2`Y}H(Sf1Kx9A|>DRh<%ZC1ag7&i_@aNK(L -u`)P8Eju&fJWeZjjOF8VrUiWy?bHmR-((UPWu!U?cG7W*FJ`XE?*iH1yos!Yb`RGRTnqRj0?w& -4fO{3~PNr&3W!o9JJ-N!r%~X-D`8t?6Jn1S#>z~Mf>S;kV0V2p)>Lbjd!>eRll`1bruD+#x3e -Y^;GD-f3oV%hCEEe+9?gMH(g-g!-=18f?dF&v_-6R$g>b4}ErDSdSl){ESlsHcpEa1iirC;ALN082Qf --|D%VFcvd$WvO-z-b_&~f&8mii=kSxob9r(t8~&l|u1=$D!oBx|(&Cv0A%fr*Egxf#aIbQgh#tnX8V?sS^dYEjA0=aH_ARSSz{fz}bD*LmPrs&WjVz^wgik -xPC?sCgu@~8$|t7>V6Kg@$qgYN5 -DlsuOE9hr%J6MmaR$yrB}X*o0iP|BCGavaL3(QwjrUr_XoEW6>%y`8}(&(q1NQ^~aj8^y|DQj>jUp*H -$HK++@``wJ>QU91n-!tIEG&19BuUJwYgP?mRog1lHO$HU-rhvR3th&@}H2TxPKxWcCS>T=Iq1~$S(V{ -0%dX2H!dz44aIFTI?^7&Q1klWXx*dO0Tl+n{<;M*Qd^IDQC#p-6I`OTS4C_t4Rs)}(Rl-tMO%Hyz0Hx -ruyUS@WhJF`2uyETG4>xV7}#Dd{CJNDp#W0fVGb4k -ZL0G&6eLM6&A$denlTLI)k6#S_!<#s9DAH9EYi -Oo>vf|dA{D_oYHpeRIx4eP!RG9HQ(sE-FBQD*EBx -7sQweyyppD$%%TFPoZzXmDcNgs0@N0;MP2K%kX0JQ#jhY!LQgeqE&KK_Dr0s-$Bi2|ksq3}u!KveWHy -mG{M>AW;VK~_v9XAX2FHlPZ1QY-O00;of09jiH|6Dt1R{#Lc!Jc4cm4Z*nhkX=7+FV -{dMAZ){~QaCz;$Yj+zrvMBmpzoLg(_n35O+E$X8EbTCR6kAC&+m?JKd9vqtyjm@?DRoS-Io(Z58ZGU0 -f5iQT=a(GR13=-?O-gx8&OYHxBA9>Z7!-jn-+_6{otS -T;%sYHm6u6&Hc9y5?l&)n>UZ5#*`&G5n~QQvWu{q^P3Bo$=d~_t9u1ROk-D!1_1Qlh;@^|Mz1d4&e82bNKlTs4q -er`o8~XAho6obec}~Ax7L$g)?W;lA=dX&o(cj?3oPNB{me4BwI8a@ko~)L-&f&5t%7rfarc#5Pl>F-? -ziv$LY!w_GetWx6b?NIQ(`{F)1t2b+T;!7v)VaYH4S!iIN9o6WF)gc6RZK2M>0&jXr?X;i@O<$0)vNT -|{a1TjWn|{Ls>(_cS96hH<~*mPG`}o=EYd1p&K0h8QTM4SdJQkrSy`#s)X3fEmA^mwRmfT29OK -HHqfFH*r0{+mYs=R4#Ofmd5 -RLdPGAX`iUpeae@Zu+cfG6hQJ_f>wb2wcBC!e9KVo|Mbn9czlu%=;?OP1R%krC;UeFi`W3e(IOaja>} -K@MB1SEEAzJQ>Tz75;Yr_g<8%jE)!kk<=(fuZ(p6HCwoV)_YZbor6=F-z1~xR6@l;l=bOW$6aD0b>z_ -b%1M}K+7Ffa%{}=sn+QOPc`g5$dgb!O=#mr@HJIE(^$y^nR9b6e+`UCw`G+USR{M$Th6e-lA!o2us+I -pqFi&u4{O4k-I@M~-9FK^%cO$|KNNO*Dh>h0@;<7AvXfBr@A_SOEu9=?0_Z0m>pgO`Utz%R9PjubO#f -O#b2ane)0{%pIqwIx;oAVd|_%lzFtXEL+W+o+>)e -XZUxb$Rn@dTLn$Qmzi3U3d-Tfyqe9+s}RhroKLZCRW9LgKA#uMx~Sp1;w)vAtLj@@xIrcd<%^4Kah}t -};6A&xb#SOAZ1>>1Jq1=5=qpa%R5w&EJwAE4e}K&T;E<7&rgfvliZmS&W9$!-sAQTZByii|)M~I -hA!2-7FaR@-d48TvZZx0WNzRlItn2viX7lVEJ3}uLeqmrK1g9ZZUl)twa&>7jYjQmR+gRZ-{3x7ob5AoInfgu~BHqPVjLx5p20Y9%e{?nemXdqWH$b7 -hJGs-&;Bp{l7t^3h-h8qANx^>>We{m$73-Xp7BL4cI6)CamAn}U=t2FJqj|V;=Rc@+I#foc=sa5@Zgh&|pc;sppSF0tcCi -X{Od?V3!o#ni5&!J1wFzx --_Ej6zg8jxb(wL**5b?J8j4)A&DPfX)0KQ-R68M&@C0H8X@@YeJl -`8_1SpI1uXae%B#+-tjG-Yyn3XAgebe(wms}ZWhJ~k!9sa0_J+~MJbX_xcENc7=MW3&;e|7ND_VZ0@n -@J%)vpANSE99)#KuIQ+kT)fEYrmt_~tiEek0XX{+t~i^EtWn*&NzU@*5nNe3f;EvDEP3a5Lw2lXN@FJ --q%o3sHKiRHFrYZrJyYM+dHMYgs5_Tcc7B=dsBKuatppnbcQhJxjWTFp{c!8m5;Xg@qj}SJIMHDmWtu -l!U3CiRz@knE|uM|lk@`c20FNwSL?VQ0ckQK6g&yiofDzv1CXtT=%md|;Z72p=M|`2 -H5K|+qpu#&*6qY6D#m8K&)!9t#5dK1RFotz@&{E+Y{%7c8m|XL&cVjoa2sTq{c1)n$ch~CiwfhK8aW_ -8`(p{Wel*<)idwZ?-9^0c!#UN5eXJdUQKjJ5*1n4&jjf5+p?tHPjM6&D)1Ppn2-pPojn4$KshYV~GDT -H^X0Lx!;X2xH*O9@uUgjZSPaE+znJq@r8RYLG)+L@2@JlO%}v2*%gX!N1qYR<V&MItm>SF1 -XnuITwm)0l~}PWE@4O_kR2^o&2rU@^pHB1vOn>y1gM)E>q_2x!i!KH1DAGG$J9RF>T(BIyrWL{T3Q^P -<>yLZ;h}{}O(^U%aE$|O(P`$_la%f%O-F%czxyVubHi`dx^yahgi9Dx-Yx2ZK9hi@Db#x*ZN%_f>8-^ -bJlMF{p_^fz=i-2&Fzv|1-{p-*sz{y$#nHAEGq-g<^Si67%HS@^*_i^RgPz310s;+1_Ax8aE+7GUXzp -)Q3CvYJ>N((|6Rdk^RxMgu@KO~!o8JH}D!d*8`-rT%AycZ+`UR@$kv$15(N(Pgc5?X`7exR&+B~FKT; -x^J(1o6wYPid5-#sR8#Dx1wJ1x9;U9G3l>Cjg#f7z93}uZR`I{SE3nnx#S^R5djUZHlR9?K -p%nKR{e-xE}9J}4h_B5Q(Fr@Bjjh^A!LCM9Eax4wiudN)rLIrg44cLyXJ@J%ZKh6WY{*A#%%oo@Fcc) -lL|{KJ?n~?}$x9#V?gbzxwRIEXdS|QI3!VJy;s!ggFRASw`Afs>?5;zbkeI3D(bD2`5zJ5OlYUgv=1}&ZYHGy?HT42 -5sB9<)$m{j8c?D`wkO8D$Eut6*J13eq!cTY?k^22;VxBl58$>wUR7Fh#Ejbt*q0e7T=_hpaKUHPJz=I -LZr)n!HeY)Cw0#~Pg~wDx+P+-(hW6K&vqlKJH^F|wE)GcNnsk1_rqOa@*GO^xM=r&RR^{#ck=>?FR1g -2!vvgc-PFKN1r$%9{W{4HXBNj(gHnC$<|qJSE(Y%^GkTgKHPXR8nfGk8*W*$kCNSdGD{$$XE -?<#UQyj!^v7)_COJ_h@@JW9+NH|p+ -xm95^Udy6)y974@#H?I#^LFv#`zDGJ=^pko9S+5iI6g_0-&j0SFoyXmV&nu94@=iuqierR$}dRF~@j* -4C?g+%2K!4yXq%Y@3};rIU&(DqU8bax*-Is{f(I;CcAQMUB2;UQ)mc4FvKY{x -@E8>5ulP7GVnQsFu0H12v-A_QvQ#nD%IZyaOF<`vlMuGZG)mm;J{Xpc;KolgE!gT$!KPZLtkm!j -~+NU=tJzU=Uj>q5Ou7cdY&i>mx2~+&Q}nLkk-D;D{bh{}(iF_8Rp~UP24+I2?wyzAIo!sSiY$A80Co882M~l6M3t2FNOdh=eC+pL?w;&^bF}~BKaSI -*w+AQtulLfw?j0TPA0G6aD%>@7ru?6;|KEL(d9w+eevjcfGpzt8=;@(`?N0~*{6`}Xov3WEZD@DKprJ -V0p$%Qdkiz~OL;-`BT;H#H2<|3y+u#3pklfIdDn0(&@yXt6;ZOy)RMV`Qwz@%KsqHWRxW=2RQD}+zQ= -J^W-NUB29DDJ2rP#X>ovu=+r|H4&YjlFVt+A+DHHv8HsF#8Eo*n~@IQ2$~BVE@0!Ci+3)CmS2cvVgm( -)uO}!DtQ-?Bza{O7JP0N@Mfd-i_5NmbsyNI<3LqbVvH0!n-yt9EK_Av%h6JQ{+)rH)D`-2BMyl8>jiT -zT|&qbcKuCNepKRqom{9_nig@B4e&2e&}BuCz%CfUMTnUM_z0dQEnn#ap>3I^7~Dh*XAKxI&6LoJ -FrlC+EouZ0(@%d@;{JB}EbQ46#-TqV0&4};2($oO21hgw>^`ixjNW1nUr^!MUk&pj*nZ}C-UD>T~f+5 -ok0f#>t`Os%n2xO$^#Ef80e=uNv5Y}^C9Ym_u=?R^=q!O_-6eIsm7)y_6)QrAUNNELd)zr=FdWAX>jW -0D)wuH-=RfUWLsjy^WNBdido!GcOP_c5Of%Llc;%|MQg8%-##pEA?#qeHronorPR17b^X2{yAMQ?Z+J -RauWpXN}|q?suoBc0)To9l~$ex=v4Fp{pNumpVNiC9m@ABRILin{3sTP!6P3^_37V6a{59$&K%4()V3 ->#2Co6@9C6X4^5p3N5YsW;I&m+a)VE(MeFU^p}wTfuxxpszG#`nxEi*0gG;XOMPa3x`}9OsQw%r{I>I -F1)q%R4HRN&9)$noNU`g}=O?Zb*lqid;P;E)*G(!%$rX9M}@3Xal0KK~~zgSI6MTxHuiRQ>K93=+bIN -3VbaiHgZ9V{;B$_8X%2rvMQ@bna%{S|cPdk}G`P@+Wlcre6GfB(=O3q5A7+?8r9cT8K^E -A*-=;Ethm8=xN97XGm+a&ss?w2geD9^@HHA+WF&e^G}=B_6Y3R9C018F;`j94C6ae$47NXm?2*tJwyG -eR$A<>u0}t6WYSu&NL87-`wiBgh~{oAS{>j8~kEW6Wnniq^`xWC{tpj>lfvecHue6LJwMsO?H+-NPrN -12S)Hg-9`uF?_2%@6POfKrV&udJ2@|z;niaG@5#`><4<1;~QFSU -!f59^QTr{u)oK8HiVX(tV{&)FtJ*s1c_&Z*vwsp4syVhA@t45bu;W=C^+4O`DzAn8Yn!4&S!<}#3=ak --)1e0z8|1k}^R0=+JZv2J~>eFoaQ%$j0i3&ZiOqyG@{3InBWf?`$pT+;~DQ33`}sCskC8$-RWxg|tv+ -oU<*9I{EY{Zj{=%4>L&@#(-}eG%3N!g^7FFGcZMq3165eb_NH4B9oQb5k>iY`PCzMtz+jV$}Bu!y~;P -wNBSHP{QqU@7@3NpYM}iWU4_}@q!dN{bMMyf*89a`?gb!EBjiGW_IE(u4etX49m`Dd3sh}M;biJ0b7d -A!)xv1g9w_Ii$^Bvm=ZNSdE8OgnW%()8?8xAn||!h;0tFS&U5X5!q@N!CxeSNU43N0^UmH!{wsVv6g> -=FrYN@>17}fQx&4mx`CVTd^**|SV2;?XiFN4dz|(y3v8c+$jt(hEtANd$%X-A`H@DP#ER -FW;Aw?XGy;C_yx5AUdP-xPIq%RvvrP$DFzU*B;^pvZ&95h;@i1@|%D6nB^A$zZ_VPxz!^FubJwO8BlJ -7&WOVr&}_5IqHq{X;ZU522u3*#?B26`lA90feqlDxSqyq_uZn+E*m6w2e#0ju7`= -!n6Q3+OBtcHy8brffgeevY{@bQ1$-N@f!BTHMMyM(?Kn8?S|ngae=^g(u?%Zq4%D*ub{fh3k}&9e9jL -xwQd>kF3>V6$;Sn(I71JiYj)aJbv7@vCQ{GxKvOSZhHfo!o$-m<0m!%6GD|pdz^1vXrG#EI95$&2`O5 -$RMVzkJZJuO`jZF=!@+9MUb5#aqoQpHnjH~Z(43M2yvTGjLAZ7y{8l#2=@ZoB(wpW+dbys&EuA2Hi9j -}q2IQBwoP}PtRNuUZ35#D>%=V@_XH2p!?Dfnn_U@+tIs>aM+eZv4BOiMsYA6stliFR~pf4B~0VNlC%R -1uUmL{V(JHE1%P&k|2q^FCNIP@<7JJsLJ=;IZNLZUH{Qj+x)PFVnVwB#cTtagB+m$KGwg=bdsMjP58a -WGx+u-4~cb_S-boTG{Na%fLp>6Z4tkTwO0Le7Br=FYfBQ7kTHxjxXq9mb5M64HmClxO%}RPhzN*Bj{R -q9g8it$O`Qx>vw+bf^rrUFCv#VtSWYFMMjUrGy>enNKAfSpaSo8i##rFkl*`NJcYd%2*YeDrofsVECA -;|2wE@D5EuTqZ%8vbCP~)AA!&NYEGs*N?A)4kSjRc#yfLFWf?GfzKQ;hG4|#RVcl|Cqv?wU&p8lPD8O -$QMvYORTk4{^y%#G67-xl->pCOqFV)=!g;mhK-YL16dA$^KvWdP4*huvs5gXJD{Md03avdU25?9F7cm>Y}a -%^z(awhI39RtGi`92|yPz%+lZbr>n~|^ENwJZD5fqJy0onio%}(ND7Nics6MO~V -5WSf=s2YMh9@oKb;=faR5=!pII?I*>}lHF6Edr -_{ZSm{m*@pG!+mj@_NOT9*N`j^yQ9&xH>zd@>$f68a4w4n%fg~7$w3HlGz73IRz#bEdBj9a)1#>9?{t -nWpd2G2wCz$GIow#*(MTI@iX!fP)Sy)Yk=vWSaIWH8UbG4Sc)~_n-C@1t5d7o$x_T|c<93~*Kh4C=Yb -{?2;)#=5}@&bHBv4b$*@N!fwv#KdzCH8wyJS~HF){|L0up0zv7&SJJ(u)Kr*n|`K -!SIJ*>)KnO%9y=SsJOP+CL<1hCd`n+jm=@DpF%bk(GlT;-`)?h6pFRa=T;pz0Ns(vC}mwooi1ou#o?@Z(~xccjXfY*y>27oR96$;-vvL#rf;l(jN_G1b36^~iu!jNXFMlfmSH-k?5X-uHh2Z) -0mx4mRlV-PQ}=r&;qP2H%Sw6VE>yJm%=VxmsGwa&^h=c5GfT;=kztc^hz?OFmlgjVu^Q_8my^pSjJz? -wq#F1f&qcLzIs&pp3M=zJ3QKo>DQ!q6SQY(C1COGPn#coJPN;BoF(8o;?v5bZ?asj>tSC)pQJ2=UWG2 -U8b!Y7q+!hP6Fj8R3Nw3jIS=VtxJ5vyeU -cSNeCO-Se%hVYsmCnarnl_OK!3s)xxU}wc9#T -z@Wx%`mu^uv%bUwh_fe+(dslF(GvY)m#k*4 -*Qb!_C3cRj`;(v0=G!;8Hu7vwIEswx0IPoP%%KMZT=^k43qvecSWzFv{#)?NIeH&px0_z)iW*=HAjKc -5OU93|iYz+-JF_hituW8bvF19o@GPK3jU9=;XfA93KtWS(z4aUHIszd!hB-KEF|ujCKu-vM%$>aJh&h -tc=FjIDrMn{MTdOmPq2Wf0RS$R305oBu%=R@VY6HEqIf=rkgT=p6-vC#d{NN81a$2{}uy0P77E)XzNG -gC`qK4t`tqMXFR(ZIFXX?K>#Pt^mA;<>cM1S93EQ9p__9#2ua>e^LnBVTNn|hg@KVaTo27en2mX`6A=E4{$2qAdY -gDLn_+^{^a2*OLG=fZ`Xzm6ixT)fE6O5V5u_gjN$qEAr0Vbgmhsi!*yFy{nWWFhyki0jwE^ySNZ8im;B;T&4=fphsr*lV7sJck=Pqz=nLd|)aUSjI^H0_b -$jd?k;uVk`5r%Sy;H8C&F#C@vU=TCSJIoN>iRdRKq1O#=q%!*6omS^bP)7&UA86;6KiB}sxl*|@2Iiu -7uWQgk|iPJV6?I&*R1LzKWFAlj&kage!^e^8Y9=+HjW&)GBLF3kP(b1^c0&u}PDKD35H>1#S!G-aQ!` -E+iPxilgwMW`FF|E&pnu2MYDFx7G!Ae~C4My&YgXkU9+Vf~@_iD+%IuS -)D&MMTpY1t$>3vF-N_~MSfr|5bcmGOSS=`ce{|30=-=**XUI2@WM01whb%nrp-7P4Tj%~OFqeoscK*+ -;KXe4XAH(l_1p_$dM3?nD16IbVGv&Y78)pcR+MlnQ#31*w)&agz^1TwWeKbk&n}bm%Bw{$sul^1GZgHsNkiHnY(fs&eWh&zeiUz_;#W_mXTJxj1vno19GJ=fj{6^a8s;0 -o9xmS_q>?^e?xJn251IGt#K6Wn!S|U{-80?x^Tmi!9}n&GhOaUvKE9N1X)n>=G{E(&Lkt>fh1ct~v- -EzCC&K_C#D4(@r7F$SoHca;i_Hl&**&dnasve=?&PT4P`exMh%p2Jd~Z1^0%6H?01y+9nGn6w>}N6nn -`qIXuQM1H8a~QIaL|dzyeh-26Lc`fUo_!;1T17TlfI+n*%C9C -G9&jDMpj`WaqHf_wlfsAZk6rsb1hCLH5#IBb6TY%A!14x>=1Wzxd%#d39qA6;n5sU -RAyc*)cMH&v3_ZWU{Ryf4>u8l9V(X7WX6p=Y(#tMQv^y;=i1%} -#wqf3lcXD6U+ej+Jm>!6BhlZGNJ;Zlr&W)lz$oaW0C+{&uBElziuG?-n+9Ba -@;9|DG~Vn_X_o8@Py|qysBjiq$#Ry_7fhw%vV5;y -bFpd*MuGOoKuP0#Vyn;Z@&CRly&}}I#P{=qr&%v?+gIQ6w(hoUCZ&EUFIq%-MsW1IEbW7?)IiaINF(@AQl}PR^s*DyFTen7g*3fXN7FRbH7>Icpk -l9^bdm`lTwU?roP(A`ix;xXEZnBouY6}2BmwPrZlkhpU#!#dl4yukPJc3IVjA -JCOTq^s?ueab}J5id*w^6p@twk#aXA2x -@fZAlu%8wZ=)En?NNQY$Ji_kpvvREu&zfsZ#LeEs9INFR1f8;n43-n4@y3ffw})CnDNlx4=xurlrAe7HdI1%=V3U}pZ%e5Ufug -&a@8SZ -&efyv7B_Ymw3z)d|Bb#(IkB4jlImZ9}(2f2Kjl)MV&uo3r^am}Nl}LO<%pabF%s;f62Z;v~M;BA -cySvvp)QJdB>ahB;sEe~=UNkr61kLbbGcB~&am96%RXbx<-v?2na=hU6o_SuMUIr>Qw8fZw8(?lvry& -e-5TYfiz~DjIfKk*8%@j>+#X?8=6UK}AqNeoV8S3@;sHBAZd?s%d0k7gXz}b((%!k_IQD!_07NLNeX> -*5~`UmAz?>dQr$I}|zkn(V$H$Mkb=2Tw>aW-}1eGS*5HaIq7b#etEf#I0hI8u#suJm#tc95p)74A8&t -tVSSX@oP6Zf#E-(lzh`A7+9QoPY^T$3?S8@*CJnsT!KJuQC&X^Zv6FhqIxHlVf+}48vW~G$UzKG{$Kd -zoY!s?m9$~L7TaM)oAbb39uOT4_O5_h^RuW*bW9VV=ZcZkUs)j9tE$bn0(%Vwxh>-VgSYMDPmX4*FZI -ecu7j>=HvWahcBG;3Kq6qbJ$wSKEFoB&BgEqWKxA7X?p(q3_bf4r@JzF&#f3 -6f#l-xKiaoJ8w?2Z`!F&t)04HQ?{lMQBnR-9{-Us_bP#so|NTsb*C9Cd+3w!sP*rUWrI#dkAD??S0?A -S}^UR$++h$$T|6m5hyEW6W*|2kO+TPL6Qth3HgBSk71H1>|DOo5{#B9Ox6G*$ly@avl<;Cz;^h#JW^P -)mLhx0L-#65?}*k5*X;h5suB|&vLXqRh2?^95I(htbowZ(ah1GKrYVqMfre@0-e5ahF#dyqT*D6ri2F -(axUEKX(!Dz$097Ljt6`lKGQO%tI_jSR%L2u=IX&JL}$H8vqEJabNMiJt-}0-hQY4JMhIUR4gFIiS7> -u9L##iioJJfN|9`lMM*EbfzYWqZ>+qKM$wu6oAeNHHg)Uwc;d?PPg3O~(^n2vXei03d$miUmNb!a7b? -QucT8)Q4rxJLqjL2G~ZtY$||K;ZBHoASXSIs?P*oeX)oi5O@RbZ}$OBI$|J2!i$&3-WXWAUBmp7eugJ -1!FhbH2*<3{YWW>&GHHqzK|5jl4pl66q@)g5u{+KG~5*w0i2Q$GNv70qxZ0xj>U;vR@K`_ZAe`3wEnY -?Y@=SEo>eqd#&tb5umIh&4oz*1f~?Hs1Tq=c@SZ~o?@A78%1J3M75}EB6_BY`i+S-Or>rH~^B -brIlc17M-w+Bx!j~JaqnJ>>Fbiv5oK;!X?fJK6sc8K0)~q4>y$2ExsM}&!>e?oh_LRnxb+0{llH+-T> -b46#Nu&2)QX&2LIk0PbQ9Gj}bi&b7t}^o@)qNhfttKhb=+ZWsaPn31+_y+SC@Bq$&E99OZCN~q0~_{q -SwcLF0+L_muqmNxhJr9e#2>Mi(JfaGUN6rh;il`6(m0_^$64LMU6X<1Q(- -Z)5}C=>mCKEmxeVJ9NnCVKL5;BsQ-ee0%c8U2g>G&|MvMV@toRt-BXce0AWuE9xK@)${Nr2a(C|4}vhLwUGHar -!WSQ|!u*+4>xr%W>Bw80@-rn{4gYu99DVp5D8`QK!04iC3LRio_>nRSvGtp^|3Xh*;|V2{`rxW*HU|?JQ{#{ywN>H7WNY$j>D*>(iiw|CM6lD;7D+&>-cK=QyF9;2kNO9`&n5Jo4w!qENnO*?m5Cv$v0@Ys=nz|#jH_-hD*UJODOB;eRd0#No>2+h26{n -Fi@d15@MnUGZ&DTIoLA_0aS4xV<(Vu+%#gW8rtIL*8HcXoZ%2-!z{YAfdW6kM -MJd2WhWj1r$Ovf$Yx#`U1N<)4j}AFBt{ESCJ?EIFenR=&Nv$L->`tj%~*fA$J7h*lPQTH-HVsD2r(d@ -VampE5#Z4LDEAKNR$Gq<9=H=y93y0Vqsz>l8JsjldTkuoA -O8QzIAH3?CL%CX)=Ch)Yf%CUQVHVGH_UN>f9`K8Ik<4Srl4!NAGU#TNst`^N^3Zx&5x`2h?vV@&mKnUK-+#-}a@795cz=lx45YUrSiRcq~#nh^8XO-HH -y;C0sy)%c7%!Jz_FS7%c)iDI%^jsDGW#O3?I(H}W6~Bh%4|Uzm{~iq5cEg!JLhDh -eh-6l0Whzz`WK!{cpfyA(T`=Z=tsn8_>cfij=i@3_Cj9YN9w)P6IV7ix8d<_Y}R7OK`4x#tJYe&gdJ^ -HS&I{Kdm%Q6VGh|^%v#l0W197a^J*+9tmOhFJVUyBNr&|i5Yrkwx6NG_*nlfW+K{WEjbPWbo#xj|h^4 -LGx0_&X+SVH-*fuxLFxuu};P)nGCL(bdo?~rqCBt+2DLEBIOVS65mIT)Fh+-vkTQz2wMY5dQZ%{6udT -S6WMV7v(`aV1|+n3_PV-=j2?e0f@Wt-nBH7P-QsBkatsu9Bln{9BIha|D$e)tfgT9uFt&ZB`-`A0;n(ayM8@@x?m#1 -_P7flEFd*{@8AP9F8vW87T+!?#CZtnd+^vP`K|Hs8sfP^%GHiI9lL7KCAz%ip>Ntp{%AE9=YD)M#)ss -CGvY&6$9rN(ogd@>URGZUKLWsI&=L0hA??v){<))La>OZbe}U0Y7Ukjz0_98=)_r$PEEX)8lA| -x5f?mk#(gv!mb8AQJg(+!ud1_4n?FYVNdzF~1V{`TByVi7~-PqX}2AGn52rP-e3iOt%rj6q~Hs^jw0~ -MV1fe9qN@XQa-|Go(Mji=KSjOQmE(u27_BqKOK@|U@rR1B%SdNR*H=JVt#!&m|>(ZblEjjZ`i6P@{gk -lId@s}jO@FLAG=Q}R$Xk}c`R0riA<>!?qRkP-mgQ?1H{M-931u%tZfk+*DemeuMMNII_0kP55?@R-tC -eSpg_l3rXv&&QZf3hM13CfUK3w=@=l4T=j=V|Hzv~nZlwNyOy)?F;_J`wM&$hltvJ-RT%}yC%Y -gxEe7jzk?@;gb-#x72_t%~ZuDPdrP2`?T!8C{3@(IUMHwv?+2HZM85*>GRf;)bS`E0v@TKJJ*cQ;UL& -vLQm_gtIRSGnv{7OLs`WEH97b0!>LgGzH -m5XPGEHHJJs@2PZ=jBkmq3NsmaVlM6ccLYn_osKZpVebn}`YwchFL0+D -D`s=8sIt@VYylV+hICHq0QWQyj*ON@&TWy0H*bGNGuG#MA$@6o+Ui5MK^5pB35!s^dj^QKNc{2KlHTI -t?Zio&Ubb9*~@ZSxY*7=uFUG*V9Tx^@sIzDNCHtxt~Qc(qz8c_U&b@pG7hQ92{V*!tQ+9$LolXj>8{> -mie_!}P`Xy9eLxMUYHMn%X=v^l6Cy{t|N-@HOoS&)2Hj^UA$hg%2s1bQ%n<#cHR>gt^X8{t91X^3$nF -kGq7h{%gJO)h1UG4{tyE4n8set=7WuSwE6taDx$EUdT`UO>oiZoqd8zsz|)&;Mmjav5JC+i(4`V24a1 -r!lfjj0XcM$8&~9#6+T}l&x7YJ@x>k5EO286mo$xh48uJLMvON5AUcG*yLQy@$#-Xz!B~d_qDu -djPY8>JkfDgqo{gB^8eXPbxVDVwW3C0%5TmiM8dt4*SbwsnAth_a|It -EP1XZAuBNgszjn93LKx5>}C8K*X~gg2)w9dex<^fGHIR1i(L^(kp$x!fEdal&lN$oAlM7257E|pm^_u -_%!nZ+*lj%PRG=M9K$xWEF)^rngfp-T7pqyh!qs8skSR>Z?_%*+YTv?D&8EwcGjJKH1Wlv8>@E-zC*E -oW|aHQ%`#sbp-ob4tk;zFSda4er0$p)!3KG4rW7 -Gi`u+w6K52_m^l+IZTu5y>fa+H2l{JgA2cx^@NkFIv1o6hl2FV0XpgEwx>99-$pMha^++}Pl0D5}yr9 -Mhi%NV14Ay`IMATkIrdm&REB2#dO8K6ZbG#SPt#&9|@+)8IIWVQU7Vr+sX`wVk`dJ)5-v!ng?Lr_Csg -tjtXmoVW}!>*3_<6d?+dA~dx(fWIch3YHr`Ozrpj4X*4*hr%dUNzw=lYKQA0{@CYA7?zZj~vnMX_yu|^CoHG~#IXN#Nh;=!e)%j0}J3|Kd(uU -uy3*&-%@Ymtu^a$k<>J?~)v8y(?4CWkPG!P@YJ(Q|!?>fvFbgjqeRQ2+C+=TUR+Ul8#F-Ig5IBR#_P( -W3b?*RE))mqd#mAQ@3?gg^#Fpja2*gNBPndbYwUC|8lG3|5*v6w*)a%kMi3VzQBFc>KV=QX-tUMmoAq -Inh}Fwu$^m4SOVZOQ!_f{<|>;&C^OUt$>brE7<|9ws-#AQLK$_$Lm@0V*_8PX1pf6qw{*40qfXSqE!7 -v&=v^j1Ua0jps!pk0l(-D-xwhLDQMQo++^A`+62dW92Zf58_oElPS`xfo9CP4gl59Ae>}5T}>#XUo@M -C#Q5o8V`h_ZB4@R92M=XDoiKhmCM2fiH^GCL2jK#;d6xH5O!1yg2hU-!cWF^)ua6BI -la--jV7j`9CtoUF#?+pvu=#^U=hI=f^2#kZZ*Q!ut)rFo-)V~=_UCCsq@45{6U2^k$37M@~=riwtO5a -HiqjrL4M9)zyl*7+&)&yfnPASkz0fJ&`p9J3Lt}RehvlWFIe!NkczT_Z>uZpbito$ -&>`$7IrLJ5dETE9)aN(oJ)x<_A9iR#|9{BfvX4=Q%2Q+uuEV)H-!@Q0D)TYlOEJ$WHiUMD(L5lF+=D?Gp|1>?9%QF2Cq<^yQp<^<(RK@bM=NzmzT|M -uTGYcC)AI_Z{tLXvSa0kQhVYy#W|~;re9=Bh-qz -d&wcW$MD@lxEofeP-8g@HR`+Rj#pjQ$03&evpZOIN3j%R7GSBjxN8G9G6oKZMv#ybe2cbjei<%mg{TK -ds%fQA;O_>Ii5<1G3KY>sne^=u4V&&w;wWYzeiuslt0AuQG(yrbd0Z|ybK8+je3hfbfloZEfFUgzcce -+OIhf004B!rmGooubO5Vy6OeyGz)9Ri2ZajVUS@p@Y!xhT6X&L+$@V0)?y@-Y;*({Hr-P-M1Td2}e#v -->3(0o79-pXR7$QIeTF(+Wk&Xfs+nSPlIg=l6+E_(X~e?hgpP&)kAs?M^r6>sN)B(7Sz>{FzBVo?$R7 -R$TD{$3ss>}`Q1ddMUf33X=b=Ci22Xx-47otJD?#QNDNHVW8OMz$QIUFVVF7Hh2x654n;)(8 -ePA8qlnY3^}i*LNmb}sV(fXjzp#4Ucn5jdJZ}u1yNA`u8RalPCSjYo|MbHZwJM>e#gCc#d~_%^TY6X^ -b}c3lzQNt*J68kmUs{o0Y6>Y!e`wLI{Em6mT494(Z3qT^_u_ooVLPVlt^DQjsPKHZZzn($) -zsTyj(~dnr#XYQ#A87e}U-Q<7o3`H9tTke~5;JEtC#`ako>k?Ce334TWp1;M2L}N3|DUZEsFx5y$}uN -SmYjZrP~H@5Aaja2Z$SAl^)m;8eOEbAXLYKp+6B=CB9g4sxxi*(Qd|7$8qemJ+2rteB%s*Lk>-4#>3* -SlwjeJe#gQmF0H%e;*q)|>Rf39ihNyw9S7(%wr*V=GpXC?X#{!~Kn5xHlo?JA|vfg?6bXrc-#&M&-EJ -kH@{`9k_oQ3ddU7W*N=AyZrf1+4z{-j(yQO^M3NEy?-$%?iFzc}UpSQUBWJN!)Zj}UhZ?%8X#7INHp_ -~8(q9i~4V9Ui>;o26YSuv%{SO{w}u?N!ZGA)U_brL^#-3%cW*TJ0Wt3Z?z$f^gj1t#-O`vei@RAY~xm -wL*eN2)RTifc<=QKGHXki~%)KP*V^Nh0Vn@lh!q07 -S<^1iaw^=v&T@MV8n*Whl^{{TZ -?2uKZ~}ci58j&%(*2Oow(A1i@&rG9G{J9j1QC!C$>a9|7R;7fC=4=yn470T1AUdRRxmi~`7ZV;4JD~4dR@wPwwgcBk6 -Y$fDON@9`#MSgg5ns1$MQiX&r~gCuv?Zogjxa8MuA@`<#2qep>_6-uygdBjIQ||!>iC_~!cHk`du?+c -#{EJBbnKd2yS*j&5F1sXL3(7Kia_9dyMGZk$F)08UH2JZx0{sLXSMmI$0skXp9~}eIlCC5eQzHze3(y-*V46zDlEPzLE%tVA}%QYPHW -tU*OP;39rJ`~X|u{9wibcvQxX64<$t0h@9PyGAYD*-JDInyWc~Rh*()%=Vut<}N;m=aeV$#yp6>3?sA -vN-q{{5}qfXcy+;kIWwWg0+=sR1TJ3?Pk*Q@-qKYjV9(CaP^;9L!MbvBxmmrn`I(+U4YQ1#O8a9w*PAZGWCTeVTmUO$ENB6Z?R -!CgwH|5GXRVE5V=+EPS@OT|9)y)xpj$Ldm5A@jk8dmmG=jAssCo~`J=Qza1 -Y`&^5TBtO3l(EO%aHfVVs|izCtxUzEuwwDCl-5)UlpGwm2t)b6>fu)-sUpzswJn>5*%J81+1nGO=qWG -6%tKFgqNPP$)_rOUtHX^6*$Vn$VJ)s72RaT%Fm4x#jriLo`I;KARZ&n;4zDYsz4GJ)=kj>IVdl?q^1& -dZU3V7niKkPG&BR##Xs_ZTf9(E@SKBDeJ4{#KnL?Tp3Xa=~8e$<$Hr|Q8TV{0)cB`#M@OWYohC0xf{lkd=oxjFkl6nQq+G!k!+p7$cZvnC5$ -l31r0&Tae5h}#FG%!fEJ|vP!o(3^)_Zu+3{ZQ+K?=_mssfvnOzrn-DqgqjaF;{1ZWCwKqJEz$0sl5Pz -Ua2NO77*h(pR2P&rG0Xed|H?|yBCV>luF8<<(Znx995oS8|o0ZbE;4_h5Dy#at-H!WtD#{%2hqTN#t$Xa1DIU%6=9NvKd|i;gEwco~NhK3Eu#Yh8=;wKQlEz4y%hKr+6V0{9cEM;jH)5kqu3V++4AGeMS! -EuqB}328-o5h$!QuzQuW%ZIt`r${`Gz=YwI=J1H55;lzSHn4D7WB4B4T!L5wqF2t{H)?Kk^Ni3p{=L3 -?KDx!i;-;v2k&0~z7OybnIoF52{DU2}E?X|*GZ0&B+dl@0DwTmKZ*d~30kMZ0j@GR1;83qIaLsk)#vj --xlX{iclhU8{HturOC3uXisF_aMT-h3&zPV8VzsBzCek|i#C5rd@?3Oqj}vqIA`&WrH|z8f=R>Qe#6R -KHR7rqiVU(%en5C8_X`I~8JOjQ?v2pxFDuXo(Z&Es4nOYK=P&QXh6Q_5o(@0 -st|wg1nZ6sJb8UT+!H=Roxt-*K=zQ<#o2tS`h|X6>l}NF1qUSzS4qcN#4MHXPhP+J<@a!QTXitvB`p$ -f*y1+6%gw>1{MTGTVlF~!{BP-bZF~y&5CUi&CHo7|WTE9KynA%3z~tvuxmrx`9)+j$N4Ma_W8NdvdUO -lz!>{@Rjhyre@X;+z`9+>LcaPAxMp6Qtup+UN1B&V944Qpdb^ukUI=I(T@r@zJqQ1;NsDjAgf`QEy(V -%v#N%E{N$k1c5q66>}9K2X!A5r7n?V;a%etDKpr@6M}BG-!Syqg8_$>(Pgu7H`~)c&SqEDd%BW+B3Sv -qL3liuzhXN=>aw&|lTg%e`-R-@ZCYPxg*p?;q^GN>9Gud%Y+0A({$NvI&g#GfQOU|!eS-aAG1K-lPS+yJx47DzH$*yWbg1U -rwUq|{l!}amVX#<%ZHy!0(k!aH^ -f&EsQ`wfOlUbr^BHcNj?hJQB2AgEfHRM`?PMq9EamJw2S|pVj>?bm9pzI!vJ52hOK!qdEF3iD_M?9K2 -4tg>=cN`z2jbWEBkZTdoX3Lzrgosx1|Ff5#Ntw4Stc$rhv(I79Xi#xV{gh%}fA~nAm@sHD_{*g`Qn@J -!v)Hvmn}07<98g>y13@CAYV?7S`R}UBrFG-G;zhlT(7ZBaV8dx}46xQ?OBSfAw0b{cUd -v0WJ3uVFDO8jfkLjqDS;;Tm1!*I_*?GU|HvNn5?2|B&w>(mU;U9iQy+|JOWGk<)lY7P;@Ss$B)qajwE -!AJPecG6FU1zlsV1CW4Hre?Of4emK*g-_+qu9;c98UEPq&N(}MFq7X?9QX9#P#rI+RG>n~Ju^>agfGP -@x8mgk|kL}Sj1mC2L*c=-J;Km%L6cN9$B{$S^wIt-$*{F2{yCyr+XL{`tcbT2%_hiZ!5Cgz@uUG;zXA?(Bu -R{@d(ywwe-mGs(+mXyO9&p`&t0 -2Sg>_F6Yn}>>LIfyvRG7g#J+sQS3E`yPv_H0S`Iij*cC5dTNp+2dhTSrEbXOpur+a`nZS!;-GMcYP1} -lmI^Zo0u&n^ko#gLD-u+Ku;@o0Rg0E)n4)G!x|T-2Eh_MPQ&M)n5h#X!48ko!Z8YH4nW+u?ks=W_San -K9yOU&Epo!dq>H=~YrpDSJV|(IB&$Ell!RagB#1n5 -uFG)-pMTU+lj)KIRQ_m@Ggh4v1x14{%4NTe)V)i9G~<)p54GR8$FGgXr~tmeu3{3C+{u%%_@@xjIRq6 -pZ=AO!E1>)_ySXBULAJTZv9vlWcO4vrgogR2|HPZ|`hb*qt!f(#La|Z)u0(;e=e8d-#I#N~l2|>kv7B -Gd!gsPGRv&j*dt(u`3o%9|P?2u|$-4hu__E@&tV2R~(YSdxZv8F)yo@(^)vdGkp_Hu(V!NpRMUJrn#p -%4-d7d;*x~{&uS~m$>K4llkuF{&)1H_KP`5QLqiKj!Y?)*i!OfO(HNaj1@0v=#SB6kiedA&VG@qbK${ -j38yy2ct&3?|ouBnx*h9trJ|D!XX=F4RN}ikVB)8@m?rSbgAI{48G|~wlVh?!UldD{(a+*)u02WO)FD -4s<VJiIf3eSGgf-y9yD@NJ1WDNHfAj1Bi-rRj^TN=+>J?Pu-_T~U^L97+SG5}sHLG`H6C -(e}3Ok~-a;7mKvP)S%D(#}z{eUYwOpQ(ng2KxneWi3hoallJ1A+xYxh@ItrpW$?x{6Taehz6{&Z$Z|@ -4>b4eCGRTs>{nYD-KUUq69QvmY?))028G-jtU8D-veMRS~Wzp?wUn~fPmaRolHNEbVgB7o(JAZfr#R~ ->9&w&5%#MipLqyCEx=1@fr;&YWNd3uuTykmH3UYym-V4vqzChs&xT)7W1Ad}wz_~iK?pE(O1`UrZCw8 -_rh-EM(_d2zrjE47>}lgE#+;VdWyvXMFp+?b6%W{vgt -gpKHF|F=hT-hihwGekYNy-LGlI6U_^=K+MRJ&9oOz$qf=*Z>PG{gPf5k28`o_ZSnIrR>@PPh4h-{VXn -bz-#p)9EBmK@;Q#P!W5?{}CkawRqTnE0Y;nlWqnu-g4bROfoMU;K;3>e`ne9(%V6~n@O$xpE|iiO#`q -PXXm@36*u-50fqMN69OW0b+bk#Y? -wHDv({Mp)K{1DiZnAaTHoHOEPs1yfNlXF~${|V2gKLgTZSFG!i?wy}EL3`&>xdZXRW0Vhx8NhLXH9jr -I2CBy{j@vCyq;NXwa+HKmTs-qXJ&3#2!Jgp*2kWtIy1Vf8_*xx$>?x|d^v(E2m6pe+Pt}3xi|$I4N<$ -pf9!rdO*QHj*Zi~-G+iRMvFF+ixbda=hZ&DTYa&LN`*B!`fOLiWk0hHVSfI=Q%M3YMgG2j8@=OP1uxP -}b;Gf7J%PyI>BI%n3qHB!pmlUvL3@76T;(+NqO!@DcPXkO9Y?NF7Oh3@K;d9kFpqqE+sn)g4EBSu}x? -akfb0d(R&xrS}jV<)Z-f+F{#ij_m9#+H|fERs^RstRH=_HPF79bmO%+~bNa>S*;93>$48is`r)4w>QK -^}fccw}udz<){gUM0I28@-Vilo$P#(H9K(a1;2W#yY#hx#uWn|n&cc(!jssuh#z=b`c9OLBn~PxsD$w -+x3|gdq^|FhySpCcr>>Q($ghRG6`Zx~H^KLXOOwAP*z)SwSrylQQDhhcsqdfzgH?Cs#k{YPFJ|IXSJy -KLAs9VN{{~GB1FCcNK0@23j`2C_XJTwBwmac&w6XP-ItP^>@^P;^Kg;^tUwkpt|DKIL`@&z7RG0=i?m -hS3UXO*|$$fP*w)!P6KNG@oG3}OM>h@6SmZ`4CUk+nbiRYkHOLLn@sj$_c#e&lS&%>u}6M>V_vPU@k0 -)URN%jjl+g(CoQhQ-Lj5=hS3VscSd|b&6v+(Z00kYug)PM*rEpg -&$x$Q;byzP{m0?<_JH}JDGZsK!SLaF+WFMT$&=)HOBI -G4XLmuh7#Fa5=#80ztz_X|Q-3;t^GqB&yza40+JcZ(a-biVC(gg+TYclnSQ3_GtEcS0C$LyDv6#K5c5Y>7sEK$U -*F%SbNK?$}VeF;#Qp;7pr^$r9a;u>KYgHWc#x_uVMj@#Sr*?1V6p+^ZC3SLbT@j^bc?`dHX1Pw39r)> -p?=CW!2}>1Ftd#)@mC$OseXFiyeHUHqGWny?XcFd!b~rf5K41!+Z^lDf$Xut265r_?W2@?o_$ppSMEb -<`qRK&JOeO1MXRvISev_R22$M5RiO9sg#q0Z+8E^>!z4<2*Eg83t_>7{3`idS$#;pS5_08I)Ya}#KH;p51o-X$bs--r -;rRq7PG4<&q|PMCjE%x2Tce*SEjJXccBbG1mHf3^{Vo=5aN@QEORy0^V~_n&X81OAuy>HjcMJ^yT&e7 -;fNKY!?Sil2APn*KLXO9KQH000080LuVbTZ>H6N!|ef00;#D0384T0B~t=FJE?LZe(wAFLG&PXfI=LZ -gX^UVQFqIaCxniO>dkq5QgvkiV-J}sMTY%(pHMLQQAgHt7tD#R3R@6iz9y^+qBt#UxN)88nu`Bgz-D? -JTr`Gv=U%19g}BZqEp(!_s%jUDKFso^y8*He7QN?eSNr1l4j77m%h7L)RX#~-`=xVCWAsCHND13aqqV -RrA-pntX+2rxaH$2C5_^9+`(^)>A`cE0v=J;sOgF0@zUY6e?-%Bmm#S*HD(O5nlhQigt^a`rs;eRwng -~OYTJTpfP$Xg!78^uvtKAd_TuDey>rne>V85`=lV1{>BAIVg#2d%MzvSgCkvdj$frigHMkgjpl!l&EK1M%$8e!|`v= -kVbZJSd4v-Gd*As@q52YmqNv@7-fE&8oHxL*NKkQ3Zqp#~J1!o!VrUcVR}-F?OU(uq7P0=e1ftCsP)3{0-T@I?BqrWVUFQzqO*E{W$7i2KA|k>_tfj29$yAi!iBuwtD*|Esg%TE1U -rql^8o#!6-1XzqX7v{)?d5KBH67G)=Av{^O%F4Xj&*fc1Q?x^jxjQx^7P)h>@6aWAK2ms3fSz8R!V|T ->`004d!0015U003}la4%nWWo~3|axZdeV`wj9Z*Fv9X>Mh5b1ras-B{m`+cpe-&tIV#7+|kHx9MZh3@ -f@~>(*X};x=Fi0z;R`IkmE7Sn~bou>XA|CEIfBq+Pd%VSYJVB0rKJNr}0l(It~sG*)Y6m@I4ESoTrYx -zg1^7`rG?!amfpI&|Cdg28k3`!6r^7eBsy@$;+8A5vO+1wi~-iQvoX#(lk(#=5t+wUxT!%6(lu)Z#bp -g?_DI$Gz%X(J0OL=^|MyWHC0kA`x;e+v&<^CC1L-^TKVU#|zHB8P=Z -X9aKt#8q9}UK6wuF1d+lWIZshKUH6#hftzp2T1zf|dX7)xf&uRwB3qsk8g5|laq|NhzV}Djo5FjSmcW;N||$82VrAxxVIJ-Q!>&lbQA0WEi>_UvO`WG{h05;qtyts5w>lr(5rZ -vvgKeoFTv27mHa;LF)~<_Wm}R<{VIAimN@4K4lKusBILkVJc!*XlAPqkE4c%xSD=H2w#KK1s7 -JII+f3pwB8o8nBf$Tv`7?<=$uC@{oeSq!4c+>#yGH|T=sS=uds-y$c)x-_)-^R*7z|`t_P4Q&k~=|YA -#XF7XHjSz*bo-3*)1WuRR@`l@TaT9d&Gh>xfIY-6;XQ->Zi?o&Tq>FlYFI|r$I6}_7Q?aJwF6fG^t -~uJ6nrn>iyLoejA_-e$D$qFzM{dmqE#wlAQ3h4aATvaL!lYkDU8kcJe{=PO9Ym8jjqWarGyMZu(CGN9 -NQ$3fbXykPyGt%_GMp;Y~5AMI9KQsaypD&YPhMxhlsaLTkW_NF=Z_6LtA$*DXaN@ebuzVoF~+g8=mfe -;z$Ez%Q{ZJ(9IKi;Cqf7#ZHl3^uOI=x39RYV5oTfm_0oML8WsZ;xsv~?*7Bvd|_PFS@U7#bbzLPh@xs -s9E`+h^qEEr76x@Rg5HEG;N^0vh7Vn7lbt)=>1YNogcU%R1*OwYFHlPZ1QY-O00;of09jiiy)2={2LJ$982|tq0001RX>c!Jc4cm4Z*nhkX=7+FV{dMBa&K%daCx;^?~> -cL5&y2I*fJkXIxOeTX~wNGPO8(g=|fu} -uJr-!Aqq7LHaa7ZPN&Vx5KHE0EhXN|q3r`Pol5KIvci{kMbB8yEUCj_EvRE?h=|bYUQwFympA;rKR*c -;dqAD4}u3;CUHau2>=q#6C}?;BlXaxGNvbwkZy#hOpC#7InVe4&cwA1w6@k%qnPV=oTO2RH6i#3lo-v -OiWm|juCM>n+$G9e>|E^#zXw~(w}yXl<*4fH2g~!;48Rbajb`bG(#Hvnx!;gdhYyOPyWb+i~`&$>D0h%OljCf6eA+;>Vbj8L;lZ!{hDX8ul~j~ajksG^idiR{9NCkz%`R3;%xpy6 -1ZPn}LN;ZOU$n*-dJU9jH2r$u);;NC9lVm>)x!Z{o)SA+xg0!=X -owHW{Z=0B14KDrTJDPYH -Vsd*^&_hZEU3Fn$wtdvYPrc(?FURjj?VB%-Z$`&?*PC3n@9!Vqf6?o|Jq>Vq9N>EVqfcgI^yA^}b^AH -KKL7Nl$u%ausZRzYG8~Njb~)ajl%w3gabdgLzD3^Oeca&RO>R8(XXM>rgoYrq!ECq_LPj!X11W^)6LhOvi3pYPg*Hpo~ -T5YRTy?Y0K|Gm!WJ8k4DB{_g9rgzmXWtA#^wxV<$z)~4yXc|eAfSJkaK^s>g3i*opM(_7pUI`Y -~+Ng;I9~~i;PLHw!AHte=3*D^@%?59^P+7Mk~Sr`1}_ZRQHu}Ug%)Sr1243pSHN%)*`XarAURZszFzy -zyRIU_vqt9EDpgwn6F^&)kh5k-@{oE(KVu?2)d1kbe`E8=Wxy0mvsAINvf^^8s(^%w>QTH+%^ihpCM7 -x%CnzWOywN7z9Bu~>_l=zh}Wy6>h_=G^626!5iHT&6;lP}0#8jHi?=|-Fn3l(D){3mpDkkm^?`7bsTl -+8G)F_h4QFyrHu7=4I5sLr?Ch(x?gPpS2lld$Uu*K~73#Uc&du(1KrQ3X7#@&>@)9P#g2Nx^8D>bV{sjj{fJ8fzVVn)S@TH04MI&!l>F!rSCsoTUh# -@PdZB*qfPhH9btl;`AMjUO?aC|beCk5k#|t4PiZlv%JFDFyfLiwW6NWdglG}Hxh)7 -aLD2=x5FufYM^HW|aQ6Gb{1`>-nBE1M=BXM?Rbp`MZ1t-$A*q!-axdorcvq<)L)1tNk6+ZKfasu7HhAHY1^sJd-{#YwuYSl3N+7_I{9(DVDYs$$a; -UX48T`=5P6w1rjc -38vV0jiihiF%a)c>D&SC!HMABwfb26Pu7*~2gl0xi=r -7m8F#Dv3MnzweOLjUpvGX&>5}=lc$ME<8M>_EMBcA1lfWM9WG@jXuy^FBW?vNR>V^(h6?PkHzA{$4~G -7_^nJu|qd*9U@3s9k8Nb$sDM2d5~gs1BChV{S4W; -_P%>hf*fHJS$~@<^8oq^_Gpy8$*+fbqpJ^}oQbSSbNMrGVuxd6XI;%xKS&3TA17t{8i4(Qa&s45|uCxzNLNZR|TAyd -&GXIN>^vK>1#TB`9KrgFj-)DGz3tHNelhfr}$2fzQk()fD#W337sB^%qn*HSuo2h&3&PYGso*tQrPh5 -~ZLbPV1|%K1>Eml1hdt8+V|^QmAoU$AcLz(5Ziht1Ur=MYx^7YiA -gY<%*DiPDhmX>#5?_Ar?0NM>f@b-4|Xk7ipBqIE(6c(AM(d3vrQguX)xTr}JyLaAmOqFqnV-y8 -K43fvB#q7QFOlYR_Xwd~{Mx6$K7x5veq^`fq6(9kw&x*VFIntpBCjCd+}1$m;WwZJ`P+bhn+5S1B~aT`7U)}{ZpRlV>Ms9PVl{+VrqF}zaL(>o;!_kZzeG>rk2TW%ENsi&46Tr&t|k -ICKD@RVT8hZa6HIfwyz*eK2MSjti|y+2%@4`dyX4~H&D!JaDW~cXdA9-%fRl^wuam0_pHtP5LuqOfD0 -r|`rJ$rLxFLF5DX7^FX^waA^)Pxf4&bR4IxJEXq@8%=OyI!o2{>W6QlXbd7r_o*5t*AP7L%=9K#`=xG -ytV~tH+o_}%d;HsjWpL(W&%Z{$k#F`hD;F-xjyLgRO|70oejmx3Olemz$I!r@`MJbe$a;{W`8|mxl?8kx=^~MjOyb^<*7IH4(L6mhd0i)i72|0diaLyX@PECgj@ei1(ut -AsUtONoP`~SQ5A1`0wmRgNJORS7It@7n*EzY)Pa^AZ$-~wAErnl@?uzrcz6B!!qm#J4;e0|1;$<&$6 -IBH{TjKe_qq1pVzHoml-W0!EJK?OT4tH4@>vEfC_@$0!V=#gE}%9%5NT&~||!tp~6a?{*n( -q>F4D%LHO;|OURZ9o%=*_LJ5|9;V$ -yV~7eK4sp#d1T&9c+*OyL3W5=Zf#966df -Fvbfp4PvLWO=&G7pq%*5*J%jyvi0u^c07;nJ6**MNmYWNounW(&vzD!u%;+ha^sy%P3in4A;77{QKyQ -oSP5fnRyt<>0=knl%c -iW%aSCS}pl}xiD^WHRi4;iMSz)pN-XSUK;tf*2SmfN!4A0Dnrd8U4>CB?issY=zLLNYb>$|`I+|Zjho -11sH+uxV)p|wIuRxjZ%i62Y&Ls+g^y$H71=TqC^MYMh!j_<{TkiCZChlWcVH9m%86xy!lgUB6v0p*SQ -ENG!hZlZki{=?hbxm@$MX2sxPe!EyKL^V$+UjY7si$Y1G#^^WUxQrhmbF+jUmnGLbZcepDmm3%26Vrl -WQma^xJ@KK*5{TswmfWLeQ8Qy;tGq?Bc__GdnFbJqkgWo51^;$nkIaH(4bM<87zfj7ReNhnO|7h<6d~ --1vvZ++Io$^>j!e6ca{%h_II~ad-R*jN_Ytig1JE4EKd0B5tBunRImD~i -^n=2O$Sz^?2pc5IwB9+8NA{ZL&g}e%A=hC*)d`G1(cp9Iowvrg_uUE -xRF0v;19ysY?|-v88V;9%I7l|1n8>(YA!kt9PYmcvDMuioY(i9LT%rMdl%|{P~x84@Q!yIt(BfzYnSf -O2h~XXg4H!I;nDeep`S=Pm|@>Wc*1~G_Ts@WK>f4|r%3PYz-`hyKw6_?Z~2>Vs)pO6DhFcs!y+ka>~+ -@?p2ePI_&?*=uAepmeZ%yZQwv{(JwlNTu3b>)!@Ko?4EQOy1fCr-Aa -8hGE67bFYTo59$b?qhp0vDh?sE8+*Wff}Z-NBDo{+&8Eih%9_AxF7ZQ?RWr?UbD8fS%jl;@S~eN^Pvf -aGzYFp}V!O#PtU|G61yFascpBkd>rv!MmO00&(^>`828Gv4(JfQ&CCl{D8~IgZnIXj89gIBr8Ad1Znyf>NHcCXfBde -&x7J|@5oN1*ooJ4fFIRW^#M7t~0!h}KBHJ1riFk|{~PT?9}<`cKU(6BS7N88kc;{N4*y7 -&)JO9KQH000080LuVbTOvpy)9nBN0D=Ml02%-Q0B~t=FJE?LZe(wAFLG&PXfI`Qa&K~TE^v8`kwH?!F -bqZaIfW*xvf%*iI!t#^X2=N?+XYkCN+qQ|bBsB5n*vo;nL=uG_$NnlQht}4Au1k6#CNZWQeyCq0xjAwz=n8-s?N?}k6y&bml9q -~S`%hXAia4VMb3CAn(cKB$uaPfMmk{dbB8tQvx_z@s{P!`dDc_==8R@cBnAHJ%qK$L_T44Sn7F!{2Y< -L|Q?AJ(U}}O=DS-9T1sr}1xD47Db}s(GO79>g06Vhdx3Syzo@6aWAK2ms3fSzBt}ZxoFJ00 -5H)0015U003}la4%nWWo~3|axZdeV`wjCX>4U*aB^>Wc`k5ybyVAK+b|4$_g4^pvN2B|5TL`ltm}#mZ -65**Lsl57O()tM$&h5E!LWZHCHbbL^^2Y8@bK`ElsI%sTM#N~v^s!2bwchygn#cI3fTO7yeZdTZ`R-L -?r#gYmZt)~3uBWIz_>JG#9!0*XRK6>dN{TCV#m`BeJ6=%V#-wl^aOOJq){!70qgKHNs_AN#=sZR;-gk -Cr%Wj3*zO)aaf5lFNYnI(=A8i!4FMrTwqsyUh8bI{-sZq%4H|2qv9bn^T39q-6$K258F3_@ti}e)Qb= -LTGBemV#fVu!I_OFZpJ+ukZH?7WIlTV>_ex^W;~p2u%5tphY9u)ac)?em?)>PvZyMAq@3~Jyxb4$9%5 -|ZvgY^VhG$PAGMISq@I@ITGp}#p!06^|W)#xSj5Kgk(GVITjpee!%z+CJo>4wG7Y1Q->hMNurL!gt -DQI%^u0No(aOjPNXLP}Zr~OMC6Bu*S=@gdbzcinCL;I*DQ9GyOR!B5&Hut!9bKz@?gt{n7kH<26`7OK -hFV=Bwg%wI>=`XnuGFmYN#e|a-#=8z>ouw)Jt%S^i+#KHd>(CRr -&csv?#-r&GX;1z1hz@8{a%rabqW=`OBB?259V0Z0Z}*YHZ`N-mab&@@RIYn)O9Xd3j&30iP%T0Z>Z=1 -QY-O00;of09jk6yFJ%A1ONcE2><{b0001RX>c!Jc4cm4Z*nhkX=7+FW@&6?b7^{IE^v9ZRc&wEHW2=< -Uvbbr)HXEPNwE$v8Uyjr4gnT4STgJ-2xMA1*@PugBPq9T;QzimQg12xp~EmEp)K*@b8pWb!!Uf`RH>D -!GLS0`8f&Gktagyg0*(BPCWf`rT7AR}a;-|(TU(pue7=)*-)!SlmGisPdq$s(#d2Ps^8Vp62p;wbd2b -`bv=Ex8WCeCF4U{Tt3IwxPO_4#Wj;Lt^A(MHI8Y>$#1|by&4OEC39kmK}SfLgcL4dKE-6h908o~=U1S -+R3_5(Nu;$VzHVi(204NWC21$Zy36xpmMM*1&Bbs}O>WkP3$xOenT3_;K6Peyb`-=&*0SZ(pMG@Y%%|o!u!z$>d#b3mLVsvTm_zN~gX_7k>no(?#UpOkw{o(ULb=wTs58E9zB}2eKttVcSOxC?o&_n?wqR# -&xoBoQhTwfn5F-X#hNbvU2Q0hFXi&p5;3x~+a{*%zOQ!9|I`hXhM!bLQ_i;I&&9#BRt=+aQqwTkhG*o -(VzG$e{-zx{l$_X${-zzzDYI2S@>5r#NO2g;misK~-G`-!#78!dbjT^J%iPB -<+?szI{kxF}Fe}21PLtY58hfGs(jjK5yDC7ZQGQ0fY4cxA0P@<@K9`kvT+QwNG9qkLsQfam+DsZPc0v -E-sZ3q&K^KpBE+3Z%N@Zlc*m~rdyop8KZ#BbPte=!bMifMzt*1LvhyBYLB1q9asG!w3e6Y)zxUwT6E9 -Z5Rf(%q+e-9TN4lu|&($(5qXC;=oZ(G;xNi|N2ebtQwHP+GpW*mSzx@tHr0qCtO=^P{`Q#=In|P6NL> -Wsd^e5a)(tYPm@qR!A-c$>efPu+h2P)h ->@6aWAK2ms3fSzFV)i%mld0043&0018V003}la4%nWWo~3|axZdeV`wjEX=i9`X=iA3WpXZXdA%BIcb -i7`yMM)mbZf8#LJvQ}#;9(#J?Wk{=}EjNdm^+LiIGHtfHEkuYP-LE@15Z>l8|IK+m&quGxzCq&O=foS0Ng#%Ex>*bSk@%}ih?*g67xnzZ?j_hed0w;at$L# -M@K%L5O*BUM?pm0luqfbdcwabvj{NKh;HMzbwuE!D9S4VOFR-S<|CStctUan5=116VTOS)Ar-^M5yyN --vqc&)K?Xlalf0oz0Y#QUHX~jbmbBi!qdriZf|fKj3hw+n<+O-$uNa@=AX0M_4b&Nj9%JNnFr9_)m(f -%On`x!+SGn!T$YYcU;7Od4S%tt1WWy1-ZQL`P;iO5>j+QX(07wZl4!M!=u2K((lv1EYGB8xiVZ -cf(^P(}$ErXw{bSjw!Kipxe#OM#f$r3cc%%#i{op)Zq=mw?hRYHGA(-_SC(}YBEMo=R#E;l*n1Cjv`n -a-xSI3ce46`-(62|=Zl1*6bQk}#mYA={$74<;mF1+t3Vn>@_cS(Wr$F5e~R&7rjtj)rx#5bzK&{4&OI -vQlMgjIN`lXv@la>O`q?A_U`aiy)u1iyaklSrL_klUo1 --@=YAkwSc?=$E?V(-Cktmn?)GPSt;F1*(v8?K{%Uv8S! -JvqT_5U=6~^4WJJk=@mfLrfO#V*JZmb_C -sDhy!WDw(2SOwQ{tj;Tvq_MvkLs;|hxVdpbrX+RUn;RmkfD{##9O!X`Gn9>R!5FDZjiUB{eVKVvK=RV -eeyJ7DkvFIIKzEjgbw4~QhsSZ0K^if8;4GaF(gcls+t!q3Mfw=0yevzDQBKWSy2HD6^VEFME~UvYUbv -vP(&zIHrLXmmn;R02pymOA;o$B;)rK;gjgDo#Zgboq`O>oorv3tXcOM+%@V(4njnJ8`4zBbcJ$;hH_8 -u&&uffgv!{$Ui5xNM@dbZlPkek3j@Xmv--2sh_!TTCO>oI^m&_}!LKCF-J6sdQyL!{HkigXxIh}3zMN -Q*__+XDarKyJ{B2VH*bJ>{|8*}(<#3!~B)p7b?G-GXsOJoe%^bh8vf^cpUjGrE_?CmwVQ_K$e2W+JGc -HWpDBs<(I;1_=w8?HKq|FP&(rF}{;%ntN6Z*g4M9Ew6M&BV)Qw)4H(;#I|M3>(^s)8YkW5`?rWI52gh^=oLnd^T_#XF23Ju&U -Xk{_*j#ank2<@)f?_ix`kd3`{nvf4iIcBSQ;5 -tEMlh{4Z*wmj&#-GtsWAy6kQmgeu+U^eo{bG3Pl+ZG!f+y78fmwOpYbAPG4{x^fVq%hhkW`afg+{Wap -5#(9#Y`b1A#Hz1`?DSm!!Ql2U}P{15(3Mwi#4*Rq@zHtyF8~xdh%(G;Ir!?~XZXuwP~NdJfp%1<`_TO -^#CPeJ*R)Pw(GeoSE%E)!yc~#P>4jb%QxAN|ccw4`D+Oy){iK?5ZuoD=!q=<#K7nDGi_}hAsQN6$X&~ ->8kZ5pvzXi^U$T8hdz$u&yd`oTQANp+P&V3Zs)vr+HH+zW{;WY#w?qM>-)g<`V_Dj<=kJQo=chZ53eY -vn=`LG2pu0hFbDI_h)RvAK{}>+p_><)L9W3nvsE&q5eZm%j3BkZ|G{FjqF$Q!vCuZ@IK;`>Z5Zqf@6A -CQ=Cp65nj&%L)?T*aoDq|RFfY2Mw#g0Q?QOQZxC)e&-5r?iV(@;@?hHAmwjtHq)DN)WaG@k?bb8$_bH -s_LPLC8&@0X&S!Qk2O9z@blg&dVc(1+$yiU~8J9rkhH}aKnK_gs$gM(|_PNz*=X#9e -Fy3obtT0@O8IZF(xx5giA~G#5r(j@Lk+Nw}E=#izq{*dX6so1qcvf*SBmq0^(-&JgbwHGQL3j(sIqeY --jYxinr5{=Z7EB{Ai_19t*dz_}9FHsE%Q8*baKgM8xrK@M#RJNZOl-0sTjoizGAy=l*P0~duU1*7~CuX#s4Prqb82{ObdXgL0%=%QT?n)_Kh`3ELEcP!*Vp-q^_j -`OI7ZLdFnxzvktco?H2>UgU5YgjF>lzu+3~Gn)fM1AOxMQO_!AzYq<?qLssPgTgDWa|R45ko -tes)G4soc`NM`x>FF%v_VDGuNY4Dvv*%YBxxKYZ#?Fh>QOFY%&`fF(cD?-Z*HtVuZIXJ^xo2V2kK->iQEV5BxyxcWKUT=cjRbbJnW_v%ap;q23QOK_Mq%;5bC%G5HfTu`N-gT~qi~NE?Whi$v?g&} -PZ(!fd!|NST0fgl!r$*ZxV9vaKHOizeS=P?-@@OeKs$hSKrjvh#2yD_B&$aQqH+CtX!1000RB1*!)-J -;R%1U}aX3P6J$y}I#Obs!`tup}e45s~3Cc_F^z7W|_Ug#CNsDyqCUl$u-14`Ucv*y$cIVtMjkeJyAP{ -N4=osy@3wZ8f&1+kxKf16kPQ5NHM5ZY&7hZ>2X7}8*&b!?n!4T&3ggAK~;(1pMXBWp+ei-I2<_S~d(W -lEh1rr}KUBQ1xp?D-cju%k|Bkc#R33~ybM%g8QUvu@6aWAK2ms3fSzF*uyoM?T005^G000;O0 -03}la4%nWWo~3|axZdeV`wjGb8l`gaCz-mOK;;g5Wf3YOm&h0SwM4ip`u1M#kL1;&}2`+LZBte=0+w3 -l8)mp_P=+A)Qf&(i%r_o<`B!A;cz(f%{L=gT5X|Wc2kQLh;5^^g=dJEGMws)%9p$GMN --LGsY>-CxCxZ6fHn=S#nKIz( -WB(ZuPKK67PzJ6TqFH5>V~6(a3VQUEy_91;vJGR-n%Ytm-VR4k^6j72DO;Enf|1^wEHLrBoeQy^-!}y -$WU^v~#a_d@yAmMUy}BgE?dB4rSnEQd#no!HOACP-1y#^mkU9kyS3uPV=0ZLWpxdnu0YP5DREo&b-ID*3iGvt_l7*DDCK!PoWtuX2`Xw%b&(uZ&m7KH4`gtaoEE_x$Y&H%?=s#W0UHCLB=e(Q3>{q8hFcc9HDmfyKDY(;+yfRPDaR49MlpFzya*+9tP -((2$hq6Anp?nrw$suY}danSMm*w}pd6ml}xCEous~v*M_H%MK&zZ|R=SI#aa&?bs2Z;Ps -=&C{-8QW0$|7OZtgzFK}|en9wqYgnGS?*w8WChH1DKtQLR2-2MXizUd%zjQckphE7Z7EeE2Y*4}a``g -Z2uVleV$@%rH?poZ5BNWI0sxZP=aL;>zK1(tk&N+fo}VJ)7xJbKrlX{&pt{(|$Cu?*8XZ4uTF@@Zwmi=v -%OJ}0EfX@&+wR!Dq%*y4m^OU8W)8-m>?9Q4PAkR6xZ5JcZk;5?zFKJEDlLm{N3ZGb9?^7UC1quqvv&I -c=ZJo~h=v3s53nc1S9H6$e?MB^@Vd#IGSzD(935$xIodR~6tY$RH)qbo$Z|BB`La3Z#og`V`sQvVp@x -!)agBM;hNf*AhO|w@hAehsxr6@I*jsebV3b>@-3PgC5TZ`*IHnOznj^)ON~Z*d6e-sz3$aB0q%Qshrdbqnql0bxdV}yuT}D#PSve{*s=~|L=trx#KV|AR= -Y-|X_2+^&7H-@J0uGq**BML(w@V!m<0|^-ryhh2nS?L>89(W-zrYu_>1#9^{n3B^ENndcc^X8swMbkO -_ns#m_d2uwd5Tk=s;r}gHm3goP)h>@6aWAK2ms3fSz9>4SjdY6006WM000{R003}la4%nWWo~3|axZd -eV`wjGb#QrfWpXZXdEHiDPvbZcf6u3|a?&o7g$Wp;U+}uAfEr8`OSqO`rGC8uOT@R+mM{`M2F;3gJO?4gyf1>c)o+bY7 -hx{9?ynqu}AQi8RI1{Bl-j)l`2v^Sw)EymCT`aciJ3pQKGT#c~&oK1(u+x({Y>%mIVmi3fwhE$HW*5E -?7?W3X0Cl+3ft}>(zoTX1DXp>ys -71wZBl`Mf6$w%@DRF=!l;d9JtEst_8;BWaC8RW*&j8zW1=E+DdZ? -@0%TP&&0NBr=89S0`-=Ow66kuP<*#mx1WX%c}VC!|_&#E2vtg<+GGE^> -2n8bnA9k5aB9YtAU8kP%9`;F{8)f;+i|3zRJ -7m)Jk}!3fd&JmkZ_UYbqaBZ)_e5*5E6LaPd_+VZ^f9U^t#YX=N-_3e{L=u4iT|!q{jg8cespyF4^Yz -a2WScZ87}P%ne2`=yiOmoon;jY`Ln2h6O2(t7+~fT2>eezru`Eu$| -HZ=Q;#h**8}3z5|@VPx$VZ<*v%fdh{lxNKUgj+OS~#=g^u#|dQ(A{5U)v -H3id@3eV1GWU1gi=y6Lg!6%%FV8Dhp!3KR3k*>l*mySizVlvjwm?T`B7<2L}pzM0bf9PD2jT~qwrko( -?f(VdMmdH>7_jIk!5Mop`2Ulscfyu8`!l@yV?S^NBJkE|lRv=Y>jydHc*M({qS#wiXUC^f#~QUuPm)a -Mdc*-R<9Ga-H<7tHK8(|IIeh^@^ERC9D2j;t^ktW*Njd73O#4dITB$6B`+BD`_~u+l8gA&ok@64$w_X -Q?@6aWAK2ms3fSzCG%*B|Z;006W$ -000^Q003}la4%nWWo~3|axZdeV`wjIVR>(LbS`jt)f?@O+s5&~o`U6{k_kij(tiR}#Bq!yh~pY(r$8Y -H1dm)vj46^ODc>oFdxQ2(`c!?A&dl!Y7b)F2P2E;M?nLhF?9A+Z?`qq(hakTxg6hz<{TMv|`b7fI!`S -D=co5^hEf-sasYcO{ZCeivSq*hn2szkxqR+>+*JGnN?8FFD16l%-@`>6q_W -DzeUIcWx*R4vm(4_PmUw*H`)P4l|iHCkZ!pazPz0XV@!FXUswUnD_85?qU>lsvpUb%LL-#c-@gg=$~z -_JDa0ylR<=uK&)wR)AgJh?+s(-~$izumLAMLQiU_7ds%t;Ke6B)L9E6K}pb8#XiFU1v7}UideX&>+bv_KitNhd+M=7qz>~}1Yg<<#{xN6 -GkLscI0I;aUe=m|cB;RV4-_Lt1Ke1#xO(xx^2QCw{{u($;LG4FyE>ESw*goKI0u~|r(WSao-G3-KhmhB(Ir3^hH>!4 -A3+9p@HJ9(c@&YoxL&b#l5basxZ4I-vi`Oi3B2G4`$c(?(N)oy* -1HV#EZBuS-ty(q@K*&XY==i#wc0!wX9w1!-_w^~W~i_sK%1N`a`CL=as?c3YIRT$J2cpEN0N;SZ`j(r -n^YfA@GK(qPhKmKN+`Y-MJT@suU=`s`J7kh(HvVeY8I?(}#?Q|_tlbKZGe%kGJ5Qy%Xew9fIC{{U3pr -H~5g-9-jTk%WPMT|RfNr_#hlkQGoQ7c7v!RLw);?zRU{KgIvMfB$@8&s>)i?%)2yMa?>Pvg5!QzFd*l(81MWfOTEWhT1FVxKet7T5f5pU8kyVRdMK+h0=Y8CM~zIL|9_F+axPXUy;wz#j^}mw8=Vs@av)>h?I)OoER~RITF -{hD9}+GqnbxR9cUpqVC&8b3uf9iG2+eCL@HZda^tCL1}M4;Vbj#Kypgv#R@ujVh)3fGmVE)9o6mDY!W}5Kq97AA3e6`**IBpCQ?ve#N=4LDapSTD?S -Zm(%dyKWL8wNpy_yaIdvL7Zgka~r-k8&*MGMcZ+&dQxHRQ?GbBV^dgZ7^4cYcO1?rH1(I=q;=T0w@e4 -{ek@=$*GtW}s;Na(I{ohQktrjs$?fgj4sTCwhci6#JRA>)yyq^NX`GS>2)(-j)CJ=HsE(MaTiC@5XU? -dS2SS(b`l^h&yu<^YRrCzPL=wQ?XR$v^+>J{wEKK^4_;LGuR-2I%0g3@qxRa(}SXPJ7HJx}$dOfHEZ$JZ&ME_gd$Yt~f}XRs^tqqekY0Z&*@6r6Oh7v-kqH2+#w)srDhs>T;EI0n0GVplfe8-Bn=hW -yUuF6OZmTYXH|M@wm|`daoJC -m>kmwAIHsUl$aVK4qd|^oCriij3z*-EnZQer -Jkt#l7LrAS^VW6wF#4)P`m7$i9Dvy(Wlnub32$qyakpbrY-ioA%*7OWE0Q$u^d~ELBygNutRxsqW|$H -_Gk898Iy~6p+wuriaKBuQAub1AC{I}#1hz;7w^Gd>0FEIQIQw>poOXZ|F8g6;Im|SY>da4m&vujJ~%; -9pAy@N6_5&ajBZI#ihAm)tfjgVW~hy_+m7zm%v13bO}1MmmxO%2XO7i2cr?51ck{C-#(8m431*sFGZE -=rH_@=&i)}B4eKr)3^{}4mI~1Y1GNF^s%q=))wu8{&VYyC0E}!tKTUe1=Dc09COmc7wYAdvg(Kc(ZE6 -Q}nZDyDH7auK3r*2qVD{Q=N#iN`ATzb`rfopE7!=t9Uv!a>vt$2FCMRX{u( -K9H-L^I%<=)y5(d{(DyhJYsLvKLE&D2wOzo2ig$r&yrB2_5*^ec40)UStww#(g*?FlxCo<@BH$bm|J2 -qnj^!fkv7KWmTXX%}4Ck4ZOYb5`diQI1cirssSyJD@~>#4=XVlRb;92yk}HoyMm3)#&^eIPgk1&bDDP -@h|P?(ogRYkUiGR3F+56-Sxn;3ocX$^!1#PBZox+i#k#YBbJ!xFGm7IGqu~?a$$INAu3S6NJz3rD;5T -kV#I%j4C9JM^vm8h?hvt4ou1>P;1KXXs(C2?}vmzg@Q&O^O;&H=_5SDYsY*I5VBWy{w{V2-Fl~oeF7p -IlErW;*ZvVO4Wpz_s_V|)3A&x}6ehTMf_w|q;v2snC96#rdM@49kDm7c&a -)Y%`;+VRF04ISEf7jZ*7qJ$gwB+rg~X9A*zoy0q93npVmlja%sobYUTeo9zQO)i#9DO3S8)!Rx>kH_a -<9htT3m>x!b!uz%tQn0gDW}A;|cKG>tn{&m;cBlMj;j#E=E_FG5GP;5ipDKw|q1k)xXvFIcDQQLqnKw -JW<`lbuHIu*nfEnsXAod_4aCAaiHxre9$m=6$@l0gQWZue&fZ|&{ox7&xEZg47HTnM3Z17doN47+0VZ -1KotQPyunm=d4I;_u)4xHy5hcx1r76TrX&{KSj&Ur`diC7Lo?%ZTud3sL=M`qd?D5hXRKgYiU{B-9tWn8TuIE)#fS -E8{2L|tG+z%voMY=AEo)rAKK)+ -pfIXlt-j#$)bhj92^`h>XI>1T+mfV*;=YsD!(R@za&XsE475Ll7lAHH3M!1Mmizzs-ri3UBdP?u(!8m -&2qpx5a>&~0vo8BiWf!7fRT*gh98jxH#x~Pp<+wBf>Dt6@`%7bmwZB=|47c-hDtv0gZM&TWGblm17k` -6Hh$?NEnxxx6K0l9>{t&7=_EI-FH5>? -ko;9$eB|56OnUF?-ZN0(IxXoVDcTof7*UgMG_ct%)DuCBHnXw}sf*?B~yJ`u$>Cl7HD$IfxcBh`j4Mo7qj?<-YKt -7C(Skfd=+)6G&!e)Fd9;tgWt#|nL3xV(w9-3zRSmVer9UJa|&4K5h4lH#jB&Uv>IwhQSy9U;$lTNZRE -7Lf0Lt|+`tLV}~h<(k=4?GZ*5n%tg%uoQ-J4+lfCrUKNt6jXr1jHI57CmQIxm$M -I;i7Y|p9}&V{Tp2NDM^jmaj|dp(b$eGV`i`|-Dy1LT8!A#<8+ocW-9vKQgL9KQQmmn_l{++Cq*lGr`g -ph7Q&`)YdN@q)7j(GACl~v6$atGOS36Vh*lkHa3E`B<{W4VlkS%fU8GU5RQ*c&L*; -w#5KBigg{Z}m=_3n)a@2xzE8Hx(E|9wC~xU51+_llUcdyGj?;VhumH*fC{tZoPD+DpCM8~4BDE7HPF9AfA4l08aFdwEqF*S{-+9( -n~Az*)ylvToGtoj-(Tjf%%w3WJq;)G&b4wf>?iOgJ~JTJMeSOq(kc;HrtbOHNtx9O>v6E<)|M^OzjFL -eqKCT4h#`{5OOiF%(8FH~o2PeAeDjY_p!FRp}9_N{lgR_)Zju~>~!SIy9Vf^cq79mijX(IJx$WEwyjB4GJ{S4Vs(5tqMkiOw+3}2QG__&sVc*3VQh#(tG6r0NGfJJzAJn|VF -MaWp7bbLB$D>xjF5^n_drn}$Q&w`L$;NhfGS7MPd`>ABqWHR%)k8leLYvFy+0qJX(uzT(~k2Yi6NN(2 -=NCD%aao7@t$Y|QIcsL__qi&GYY7wwatY04Fql`TbJ1Sb`!LzFKgH+{nqAhHWL8Do`-$KEJ+n|pz%n< -gLAP}U9JRMG)nNoK_nS(QlMCW1CC%z<1Q6IM?*+jDFu_Z~k$1MqPCaxNVJbzq|P9*eyqe#QH*Fj)GK^ -cXxLp9MzC%il%#0Iv}LTm*0HH7^QHY6R}lNkm;OIO#5R8dw*?3l3TUgE`2FD -K6Y&OJ`>DTLc(UO%2Q3!VU#GGDm|sUBNgsb-2ML+1InfxzGO-aWJ~7IwtQsIOmXfs;81aNcg&)D6n+R -Bw&3OeKuiH;%-O!Y$i^OhaDqvf>X;G9R}xC6R<5b?6)}7gDt#a$dY$4QYTFhSSG2G30QG_tCd>%A=tVEeqAd`qU<3{*Z0C~2 -t7+FN`+?6^eny1znqt~zL~(68d|-4U>Jr@2XTR#vR>gG5U}~IOc#JP!1QKd)ZSy?JI}_8B -j3W9ny@}hT<;Nsk|bUYZ<7UQ0g&`2dTHryv^mQk3HW8QXkB{Xc&%Mr(q;mq=Y*R`ZGkBs@;gxi}+DHoDNVYk%b%GcaG5>Q5<0dsFf_? -HK}R5Fg5srUe2K83)>Q{_vQ1k#>2ATT<@vKAJ7AdFs|-F4ogzOWjZ1<>c{yE+k-+EX(9{}UAj<}Rn=5 -%ji{nigjX*a)-kxm=X!}EKrl*W#{>+YLDcyFEZaC`NC_8|F~=*iL22rQ0CprUy6P(uHgwKy?HB>@{QS -9`)TJzM9f{PH$cBN~W$l81%^MhIp7S5;U#Ni=?yT)AG{r5m5~_<^I!yb-VK!dcVx(OBtFJ_pw`7Z{1(&h_FVExyP#L_iU4WvsFIQNqsG?IzZu3gmP_IOn9%-Ry14D -olJ#gJ*;s3T9T~Q!*1vfHcr~osDOjJIU?HLLY6NG0Rla;%jYx(l?6O;n}NkNi33*iKTsWO~a(yfjmUH4Sf3aDSFYmZMOwB`P197?$J-4aw*H%GQ^aZ83`6(+vJWmi)8Hb2sE#&OLCq5}&q!D@<-?pQ-RogQO58w^G=cz+M;;1wY&$T3X-XW@U28Aq1Q -dQ$JVu5!Rt(WQ^+hF=QS9VyUi<@f$Fc5(4a3))kxH{Lh)oX!5!FV< -+YvrP#K%gec7lg(bt^W2?7yms);L38-d@C=n|zYu5EKHBdI_QAp5CG@`BfED15<3EPrE1yR4hs%~QQJ -NSGXM1U#Ni?N^~;6(_~x_s`1z9u{Q9%^_ -{;eh_wo4bo%{Iiy$5{!+Iu`b^CmiZ|K5Fk`_sAmc=7os_wno>{u2Iv=21TJC|~}~1HOFck^1dSkIHy1 -e46U)ApB|@dp9P32%HW%m{izP&+BN3mJPX$F1t&({rZ`I-ryx;$@aTdK)1){r0FV!KT^yT5 -?Zqf;PMt=AnSp -AN?d_azg9bHF4mZb^#aUeVLJhu2THl8-8H6%g-%K2BiV^CMk4(DTj{3ydK4yvBK(SpAK8Y~Hv_%`Y++ -PdE~Y4WMfNyaCcziY1cJ6i?$!G^{DSIQ5y -@n(daHr_}9cH*NLIpeg&+LGTr&X4j}t3Q%BMZpu_ph$HP(jDWHJ2T)4`1QY-O00;of09jiBjB{=30{{TQ4FCWi0001RX>c!Jc4cm4Z*nhkX=7+FY-x -67Uvgz`WMy(LaCzNVOK;;g5WeeI5IRYLRzQ0cpal%Cy9lsN3OVT|2naR=Z->xz7Yw$q(KlgnmfgmvCeVpcT7tmdt${}!6-QL) --DtGVwH;o70}j7>kEa4YX@#!fH+f6mE??is`Roa3OJmgl{2E%wYN0KDX*hHd%e(HJyF6xgN_IF9t3{_ -m!vA`{RMbK*^fU#9$AfYmxiaX;;ai36auBB6GZ6u3lT)c5XloA9g5g(# -Q@>oiADzE4T(7CG>L!AajV#5!@%Ie1xiKfYmGqh>1J_CFC4wXy?S(}1r2tGlqFS3yxRz0I+sO&m=)i3 -f0p5_#Nfu>X?5}Fo3{L$V%(w3miz>YR_1Ms5+#*SuX_z$TM$eS#b8yW-vioA2^~Tx0|m!T^N#c&>e0j}HsA6##uH}LchZeoQ7>+LUB9b-LTt99QTyO>V)f_dH$h!l=YNz`cp5DzayM>S`+6@jO5+is6MRgN -p%B(nK0CvyApb&Ycc(COa9p`-)kSZ{vTOkr7q|I;%GRtrFNEF=&1VMswqwHt^1gG?ig%`EEnRdbuo+= -R$=a8kvs}O`hXi_gO8wgp3t*4h -8=KP)h>@6aWAK2ms3fSzD%nIWB4r006Ww000{R003}la4%nWWo~3|axZdeV`wjIZ)ay|Zf7oVd9_+?b -K5o&{_bCavXd#*Q7FrCdamYpGd{;j+)Lt&t={!U^~kgcNsRfTfTR^qr}x`$7w;fQH#gIB5{Cp9y9;3V -*~Nlcv8z}^@@lim%FV%=RRu|thBjG2NhUhgyXcjqs`G79ra5I}@~+xk(N)D#LAFbpZJymJvXAg8I-18 -FtyS~;Ds6K*hbG-=@jlCwy8dSZ%r6?slDnLOn5=HbO -s)rP)WbdHn3WoZDYi|HBDo3MR&|b+7wW!@a3RXk8~8N`L|3E1xh*(yZR)$s7n{XbB6Bvr?N -=B<#N3p)N3AB>+rwO&+Tj=KesDbe(Nl1}dvHS%Yq1*eqcg5SLY>1-9HuU@R(MHmu4caM_Ez65?R?Gw^VL~ock%;)MoBtEko`W3m7mab; -EAPq>_F!nxV-GvWUE*>!i(_nlzP~X7u55etkZVG)J{);f!0*?Djw0w`6l94g1y+p}bTN&dRz;N}NX}R -szML#slEfkk}TkV(JCd0d`${n??p%1tU1LXVJNCS44V$ND> -I95g0021sdpkc~c*FC&1fr=nilVKaF)|`2r6n`v%>nwNfE0ylqz#g86#X~HS3-5=a_*}6J_8(RoePRU -w_Y`;TQgjJkbw4EJoRbg$twOK`T3i)QE@uOL5hI?GWCLXAK%*tgY8N)~m{Z5MftmqIl-vN(wmP`#YJh -}B1Bn)cGkh9VT+XkPytY`QCoMH<_nDQHfrP{&yyF_yX8JhaSxT-B5=5v@u4m5U~ -OUhR@(%|H5d^xA$G7!0Y+!VA#CY+G32AHDe(7*4HwJrB7dBVDpgdhpe&tQUig}`e}kPvpu%dOeU>1dC^1$g^K=NMIDdr -;q4kU?h0LS@`K1wBj-eB4DENp65d}NwN3#*2?Uof1LmOUzb;}u5WK6mD`JCQFYAp5o@7AUyHxFAu(BK -1%o%|^I&XVKVDut?{CjfUpf6hJKe9&-<{9T?Uu`r^NY(5*LKtV>h#Q^`FJ`%dn+4m`0uwDZ{EIxfAjO -J<=MN_YkLmp#|tZ*^BHuHPMk|h*JQbbyt`R0Lz67gd~JmD+-gQTIW%x|Q)DwhW&p;m7cwixL1@8-^n3 -HDGu0VECZaOzWX60hmTFwk{LuB1d#S)I0iYK%;m+ -%Ljob(Q{LjDZTTtNAvmykzVi>p>m -q>u6B06!19H3Ty##QwzEU2+UO$#3sXj-dp;LF*A5qbVM{%X$1uw0adqe$Zbcb%C39yyqx%Q>;m@&WO! -Vo}(kw@)jvb9TX~ffOE$by?Vjr`Xh&3|MhWg&nbe8v6udUgoif1id}hPcDJAOL1q{T*4ltihJ7JL0lu -@nz8FLodxmEnz4sh?f`iZizWEDo!ui>+yx<*+&d)29vTe}u+I+h0KyCo7HNF!;mxpmI+u@oL1%EFos` -D|AT&H!W!Z5bAPo+)a^tukg$9ROr^C1(>D{4zrX2fdHaJ8keX);i?hu{D#TuJT!hwJx%pD82AW})#H;^B2Fec`=QEuc~1&(U?Eet+ABzY?z?MXbTL0wbUXg+x8!kV`;?+zW8GDv6()^{Hf`U_z0it< -#!T}5Xey**-H{ezSUzgxt$-Vovw8|d;zn65q*bQtl*S$+OM9!R_)ytQX0DuDR2Q9b@3D|C3ntUiCBhs -<$5h(9?}0JJw!H1{Ygqm{3z?C^(}IOPp8UGA`v^JdUMItavw_o6LyDZ%kAz`&!Ir;TLs^*r%M^K%giP -s&|VP+^L;E*gI>X69W^YW@mm -yKpQsZC=2U>mXW|g)?%j1{mQcPxnW=AR}1Bijy{$*Z@e6KC$1_6T)C`vRcLQ&_1+8c(pRtSKghy_*P! -v*JStDY^!oV#>-Mh<%QY={-ccX+grRv7x%GyszRo|NOyVa8#@ -4vP5TyxSOK{Co#2s?3!p{K1KE-q{(y_>atVG+;4C_@qM+v;o4gx1-5$)Q2#AgNkT8`&m|7&8YZpyD@U -TV`T8F+(%SCAsFwr;_#OS3;pN4;6ihtUT2a -grNcqbWuwVni(h71ZK8w`M}f1>Q+iwO%5j5rMkX-P+D7bq1|OEr2>#X;(!mk$*;Cz>lDP_Rk?YX*|2h -EYCO@XLuVYUhDx$!Q(&nqLCR*=^`@cXuT`M+J?xoA9tXiu -@C;@`X+jz&rR^cUR)*=VyM(^8|*teJVNf>a4&`eaFY)iI?Ju`)iJ9#Sh)y6Myo<=k{xa=36gShTc?zHSFEiTc%0JWDOhDpvIicr7cBMAReWj$P=j^lF)W -?nC8IRi1FRWH -}pXr!X3w!_8>68ik==t4s&Q@pqi@MNtjyMo^X;x`p?wgQnnx%Qi!NR=>kY6 -g+@xuPJu~Qd``9;e>ZIgG7q=k)?;?{K^crHNa2G|D3R|Ogb2rOxY!8K(M{DL@P{HvQN3Pv3gyq%yNw5 -pP9n`SqgOr}SFi4Nh+Yj5p_VGReRQzC_!m$jQ;+A(hk?I)cYu>V3j+CUU?!dC0)|Blty*50A;w@zLN^QhL`~CJ_8|efi<);~5zG -=N3Te*YY_;z-R2&@~_XqoS3!Ezd4Q;z;f3Rn84tJX*3O7Hw+32=Y$kH`DdK0c+J|Z`Pi}*9AjyhdY8a -yy$0&VB1m>S496s>SW<5IQvnYZt?_YJ6*RAcn=t{QM=v^{6l7fF^vC0uN7LhzA1yh*TNCkgCjp~%n>W -Es+y|0?;xPbTMGem5`+LV!y=8q@C8^A=`A4k$3l8x$?j%W*I-S5j{O>dUyr+}TdnW!)>U1OiPWF_Ezm -s$gmQPtTfq&vJf8r0CD(Gm$XAlNImny2R!F0)YVgbLyGY0r`0=PiG@yrmvFCddobyc$6YQin>hnFv39 -1RdAX_1u^Fuf@`K011F+-In}q^`w>ObLEj^NFq?;PG6Ak0y22&?vA@Txs@@r8Mj*^vVx)Vlp+*1jKcP -ZHE?$RhR`Y0(b~+$oF`gYR6Z(X92DRmlfDG;LfpA%A6=k=E%xVl`b(~eqHPgPD{c!gMVoQ%TMlcC#uz -|WsIiZ=t;{7yN;{)t?-NuSQ!SFE&mv$phd+V{k@Y&NWWyb!2ByktYF5^ryl9zuFBJ!^ItCJ7a!gLWG* -JWAsCCgW}Id#I3n>glR)Uv!G8fzO9KQH000080LuVbTi#N$M^6d>00SWa02%-Q0B~t=FJE?LZe(wAFL -G&PXfJJHa%**PE^v9h8QXH(Huha#figQ2GNtLJkJ?oeC-EeiHkYy8?gOoIAQ6&TQzT1JmQ|1cea{6X0 -ZEk8ncd}sEdu8pocjgLR#jOyqE@3tU9N=GjmUT@&l*+BJW+Qt=gqQ$g{G`q2-!6CQZ)d^`^~n>iehr%rUTEP2==|+iKm49SOt*TtkReONDYCvh -Ng*eBT-(;Q1qG|W*x~z2z@JZIU>*^fW%)V~6d3VAkzL+2W6b?nFTd+?$Y3h^?fl}+m&@~T6cD`+XP@8wa(>~Gti9j;_< -9o|wtwwMFJ4^-!K;6GfHIorQfqM~mwvC)U`*8Em^0Bxuu383WuZ%uZbec@ur(r>0+7Z+H+73~b)yIeH -zJ{;Wh9=@$~=8eI0#h98uq}hdfeGA1`5;e&|)iOnr6IZ!3(*7f1-gShO8^9D!y#Jr0Vs@M4SGspo7COX`3bsck}&J+Y^b+@s_|T@v>cq~L<|pG+&nGvI -Zp?m-W@SI+)W^M)|ivo_iRqx~7jF0pASS~4AUR@{}hDiyOW`nq1vM#;IJ%L++;Qq4`7?h)+IVtPGQ*b -*y70$odCLMO%SHX5M;713r*voXo=x9g71jIDtqp?O(@kOhZTqC;>#V;~QXVrU#bC9HL8b0|~jL5 -u37j-iR5h0c-)z(g3qkw*ks-cvdfU?`6Ctd>|i;#oH1P3vjxUsc3LlAOW!sD-M+lax#k$@@Mj&*ksL( -(0{IFt;DRB^IO$uqpLJ(h3KP{caj+ZYREh{O8evyW6N0%#s?bCHnwOpI-x7=goQ;uWX(+&E?U;*y3&5 -I8$l4@7Sf~ha((@PG#-r@Mvar!=>EP;p#kD!vNrl7Riqjf=6l!IULa(VilS@|?c)Ak_A)1bFA#k++S@ -Unnm~wZfM0XK -m_(Qq#z+Fh8xTGwR$4|F_40A%5T(OtZTcMI8U&MUVtm;4i%{pabCy1X7st8!fUTMOlo-EWIz*d7iKoY --CO1B*$G!H?l^IBJE%_RDBBqlQ05_pE@zdPWs(uDo40rAIO28~FEnB&w*-f7G~$W~|Q*j9) -b?`pYyEaw9=?`@+CXT`e5Gd=`qG6#XJ+B@m|2@7jOM161K^@ecxL}o&AYM)8kWqe@{;Np#-J(U9t1AM -poFM)FJ`9GiC2lTEdw(8-$Qv+p)qQe!En&UC|=9q0l?vSUzP}|K?|2-D7vdl4L>$S4hlRNIkt)9_t(U -A@*hbt)d*Io9yAij2zNt?z96w6;Di@Ex*v;*VQ^w;@#GxpK -P%b5%*zy-?00G**Ey73AcR`mq10-Q1aJtpnl#i59QY{|AvnWB7xM(9OpgFO_b601=!;tfnQ1#iSi{j+ -e8lEeL4N>^rWAS$YP%OSOpy_8)D!In858v0C~aHHSl4oqwk4SC&UeP$z1X9Ao3#aVufpO{ZZqu=G7%D -DD}#DyRXcGM9Z)EeN+}Wzk~?~!-`v9uh^!iZqSkbz334FQ~GEiC3HK?uY~z?MU<#h1K;|8|RW*C4_{2_4L`C*6R_aq -Y6~bckPpX`DQ%;$l=TvpZFo80+T67GXi&m#X_v_z5>D=ohM3ukfY;vi -k!4MBc}JNg7Gns&_AK0tu7SfNBpd|6f&0RW}ZZD`!o!`m=#IaT_N92&u6iSNxp-|07U=C$aUF!CO$wj -ime=&i~4xgcRU*kUW`-TFjTN!@vwvnaL8FlPOO)x4I}6RVlTvabQ=!#7T=Az9)CBmtermPqSzUyi2{7KO?;=Gr;V_GCCf-B)=whZ$}TC++tZd*^h4?D)kYw5H -zfyIoI{S_rgo(ooE9m&#b@Eodu@y752#Z1;m79!2Hmbfd1!wsoM-g&RszbsHGoFtbjJec+5PUY|N@W$v>2yVZ!6{GQe@5s=TNGnkZ)NgA@NCyq=G-TfK# -Nu3;bB78hQ@c8ax=Wyp{ba(eC*#qgZeU3&JK0sdbImo{Kh2fh_VFK=wx=#utye(75Ll2+78GiUIJnn9 ->CSz@sS~ghRI&j+C++?|O=EdEsO=fv#K6GwCxAZv&^@NAI7oT`1a-i)$pmK}LkP&VE=~fl}RG0Ya -!l&1GJU)3j4aKe6PWq3`zDE?jfLSq`81yLSC5(hb;M3W -~&%b`f(YK#K<65QD4=T?~@$a(E)A$q@ywHMJ$t8Z$Gf?qtY5-?07=ACEzJWh+)`02eQ-Le=xquf0pdb -vt#is_$Ch`^*Zbh7D#qH##X(~NFJ>6_J(I(6D)qJ^L$^t)2O>wr1PZ4a2J>VUrDQK;?;!R#IZF=L4tk -IiK%q<-;35ra`jFI1+eSUv_@g4M=25o9aS4`?vRz(R+<0;g(I>SbYzlK*e%<{uvQNv&Q82AcbnCP<3Q -o1>S72o5cg#+b<((AnO@Ln`cf!m4y1CH$06y4wcu?>4qM4AzxC3U&+>Gs7;)BV>Rz8pcGeS{81{{c`- -0|XQR000O8%K%wh|Ew4UYy|)SxfB2Z8vpui*3~0jiO{dUdeC%er7_ImRPJUl(YN+ZNxqq(IlWBlXQ;FQ4D;1s!_a7U(_s?L -hj9BpPdBVF+Ub<7RWu1~F7a>m=YMBYh2GDP%^gU0Go+l7u#~0LuiyPoq>r&xe*+khNYG_Cd?U3-r!G- -#qR?u%v1+=jaW={J=;@dXafua?K%4)AwK^!jae!)X|=HhYsPsMjigBo*!r}b{u82N7W89T!H;~ju_}q -8L5#Nkk$#P&~X1Nsl)L^oQ;qsD=hrq;2^OqQX%yT9z2b$oEVbMfMFi1cE|5s#Q% -I#P_(*IBA<4y*QWf+AN00Ya&ohn}#syYIIM;10lULjWUzg#)&F_jEDp4k%V2?voFk2A5WYB{lGfh(m# -loA`Lv2PUPwQOYj+ni0kE=>xK2;tJR#8=zN9F5{LLs$;mwn`3smY(S?_<4mmGEsob;SQaIKtrS$kcY{ -OFqhqAN6Zsv|EAKNhu;5j0-0sqGA^E~%h*yU>P7&IfkY!u(_Aq$NttWR!x-^l -;?&%(Mir{3Hf%FPKfJD6>2ME0tdJ!!`e=D+23*uv_sy)hzUWswqsw<1%~k>j#;DF5ptttfMU+aA -YN3I#TEd{3p>u4$pSyZ0Axs1PCyX6le~jS$CQ(io_Zi&To1x?xRiTKbk(S0V7Qoonl<0kDKt$VX<>mO -nqkIQRYuK-c7(N^WM+=_GR

    62jP?CSQyYza=jLNAvi+fNESbb+%d+bFlr&%%2H7c3uxIY+K+WG(b -&38%B@a7-Y}J^OIjue>I5S)d)=?ppX0tM72jNdyt*Y{1!FXaaFZ3I!LS@KCG{2c -kM$&J<$%o9=+Fq44f{GYN06>Jwo;IvAlJ(}X}g&RKPi)tr7)dL}E^@me+YP?P@DtEh*%0iLq8L%UQqf -$A&hZQvYc)YQ}+#CsSHht#|?q&Mn}r{-ERcI2`dseMCMfdEwZWt2_ -&AOCyC8<|brUyNVmi?k8(`l?nta&XD)DgtytS`+%^z>*H=u{he!#%N -P|47T_b_xHoZ6vVy8vn8Unpb?Xr<5mEt-!4F7wF7m1`+YqUich9Tb0+|JBUs$|=My6d>u0B^chD!BOL -{b_mf_VncEAK(2z;1}L20>5^Ow}RFLE_Pkb2z(ScP7+?l`-=qdS83G6sug`(vjI0VOZs*zF1-$?oy`i`y`P} -%Op8qn_w6<#D}%jqp`P((30@A&u_I*B5(0XTij>bgw{WG@K3%n2AHmuV~Dy56xg&n?tzx#>NQ9D@T{9 -H_iN&qz(FRzj<`7!JRAF#>dS4cV-pt`NZ~2Nhzdwj&_x#~?d7KeqPZ1Dc37WY^X6_txE)Bqu!+T>zuBuEtPbov^MaDRTSy -jqw;~E8)IL{o%1P@!NzILDkw`XxGGDZbv>&qG84 -T5(l~r62}6X-d;Ct3;MsEm?l1QTo|4#y1?>3{&su+wi9ZU^{=j8zl1u9%QhyO}ZW)*mZ!&DL+bw0x^S -~Q3FvMBf=bgah`IMold5mL8u(XE6OD=#*r@3|eE0eAbQ`>eu#(|@bm};xKCy)Ca$$Z1=LK8EGqvO%*s -doVe_>6vHBR#n26&0BHNCS(FS>YtQ?B7q@FOs_YXY_?k>s!FPU2~0l+bYNbt=670ADG%nG114FTEOcu -?FZWyq4h~XUkx*sMz15$-dTEb8ZEJ;*+4D7r*+ReT>B%yvVIcT -LFGGnrt%m5q|{W2Ck*C|}GA5!4Z5ie3w-ZS`(VlOP|9gB$EBT;}XLgbcAo=eo*!SHvEQ(G! -JXEWSztst!s)Q7M96k%JFJoVRk>MfnUi7?AiB2(R(2GzUmDFYtb;$c0J^I)F&igZA*7o+v^wL7BCmwB -`qBqR`2KB#my4gR4|raIWghN;rbWxnJ_u@1^@9BI=E5y>gDXy|Qe=6qMdRG&Z7aFIm|`WIXT -wOW}|aWt1qTTrfL@@CiUK|S+b=FplY+}`S*HSv~!lSYlw?GOC0>(*2;dpNTO9|cYCC=eexItE;9naNw -;gJ>1J!+t&2LIU}d^G*#5v+ry>znm9(H1_e~(X#k}D{3xSA8^Wnn(0{bB!Fb1^lTpUuZXEd0gfP(dJaNV) -VE?c-VX3x~Afj@o_P}9A4^o#Y=bgN?-IMiw3C}*O*Hl3y;iLXjuI4rx?wPt-0f@UfXxh0O{!{EN&f@i -iPWHZ4O{pHk2v$$Z7@5W^*80AAW2-cJ_X8`qzs54{3wzZ5}=1{a}FVeBoHQES~lr_d@>VdQVWvgVr;a -)W@5~t&()y=M(hVT*jbVJ?c?(_Pgn&%XRZWx6BuL592dza0@mo;s^8=)%x$PYKrQLSEP-Xht>wR#L}a -sL%6|cq2Ihq*D13ZJDp>KBZ>L0RF;A^$O?VSOG`xHqAb$_;A^~LO1f!ifxQVya>Jzm>eTUWQQ|#o-O7 -b#*YK&veDScqozV`~y%+0|XQR000O8%K%wh4#pLNa{&MVJOcm#82|tPaA|NaUv_0~WN -&gWa%p2|FK}UJWpXZXd7YBMZi6rkhVOX_i(Nq4cmSkMYkQw2bvs8X7>Wi7irrPb&7NgXwvZ6$s_n8Ik -Ya!K@86`hrU!K_q3s7_8MeFqy}J`zRdSDfGL@(&hR31R&4TGs2vK&52wqf!Hg$zrA;6s^Np^}5Lbf&* -6buuSyNvM2(~ZIutFQ(-nM+F8)%Q>u&8WEm2WIf_2+u}i5ST>{4vrYKde#sNvb7DZ1=EReI`u4IKVR9 -v5us{n=+uZ+el4=NJdDKH2e1CSx_u)g;1h-k+$hAV>#zc&eVBad>V9%6J^Pnex84Ha4!%AItD0UFppB -zG&@QuC42Q)fKJtDYa`LC2Tg3NZ -G4aOPP3czgqKeEstP)ZoPU($w(|qWveUDiIM3ra{F)%mSb)c!)fN@mzGa4Iq!;i?2tCE_KexOg@v)L= -Je+bX_kp^P)h>@6aWAK2ms3fSz9;&K3}^8002Y{000{R003}la4%nWWo~3|axZdeV`wjMVQgh|bY(7Z -d6ib%ZsRr(efL)k)E6pIl-S+u76ue_H`t~@fp)Q*mm&xV*`_SEx_C)Cv0L=5-_Q^2mvlI!Znl!31`ZR(s0hSD6+_l%v%fZ$-sMAnFyMH4-cU`$^NWb% -$*t}UcV2&7DgITwkK5Eav{{*0)=pPHE>7=tAyFAYOu*p`e)rcoQJS~mi3cAh_C~4Q_#VVdzKIf0eVG2 -n|&9GFZSaV1tWHC2HfHMYvLoJZZzCp;+9#vOp%w>O6-DYxcK?z!{HLyc(903Dnt_aR65vPrW+&cD#?O -B3sJFtAi}?V&<|8{xy}pxGs&IWicvWJ+g2Ny?UO)R!5%=^ImkxFn&op7KR{sYg-VSW5}6O}vHpqE)rz -b7Y3L{W(?&7E4WpwyB{qr?=*XrG9}6IW6PUh@fKnl$9AJwzDJA6I*oKy?^@5!^sPF(RJWO`e1^OGf%O -&R3gjdQi{qzTCJcGVx_^hC~Qom^nKg2Sefp${6i$ui`kp7Ef8#d%e-Ica~LacDjj_3{ -D~do2Xq#sJ_eMvY1?Fs~cDD1c`HqHELPW{} -8UO;Jffh>lEIgfTQQ1BQl8n?(I8gSc^@3-D3;hcMij#BEJbY)Az3ix!h7tTO_Q`Q79QBG8a6*>|& -a1Q(oI{d8z9+IHIZQcS@(FElqb4a_`(B7$jr@r=IhVYYwGUeLESJ4H603^uFq1xP=+5$|Ajt0@*ZP8 -iF$d0Y6maj55jW-U!GvX+JY^ -?0Lj_^_r-7-IhiP)h>@6aWAK2ms3fSzF6kp*Keg005OM000>P003}la4%nWWo~3|axZdeV`wjMVQyt? -E^vA6T3?gfwh@2Vr$E_D}Bj?aWX -?=q+4E-0`)XbD!Hnvv?bcK82mP4an==6n22jz`5Bv&QTKXULqwSQp^(erP_>k+ri`$MH8p&_Q(HC%^t -6xR15Ibt$)q=rgi~};Xj{95vnA;n@5JS#Bjw&Kg`}~m)$678Hi%ne(18CRqI`xg1>A8|5dClxp&RR_F -E6KKfbv>|(t=k^jS=G14X3K;eaD)Ui49kV$a(+`Cx|(a2c)fb{%5q(Hy=Y{rB}l%5Rc>?kx3;ghze}= -AkAK}?4Q{X}MoP8IRxB!QyZ(i5Te+o$aua2^Pw&jvUl06q!dwJ -3vhV&7W*geV3GTDtT{h2Ji3CJ!G(mB8O)Bw?Ht;qy^=Z*3Nl{b;^YQM?$Fh_UZUP9vhTl8aj~4$Ys(V&}?FT7b|7k$e)NP -U1|%*$K=|gz}i<`b-(?*s)^R3!Uh@IrZI@%RF!@58wR*G0#S`(QDmAWmSmx*`BeKy0_JV#Ej -~Fq=tXkP(blk+jgj2#w)|;ez0T66%myXgts^6X^;A1KMeWQ(>1s7&}Iy!Y#Ru7~iN|^`OrW*nlAf_An -PcHudoT;PLR_@jntD5_QH%T!_HKO~a}h%+|op_Nz{xWPjYqw%57dz+nf|E>!4U3^y7zhth*94Ei&`nKb}ymg6)vP_B*NI`{oQUw@x0;Z1P|}lYxoNxGl#*pLvM#yWBs@$9{cy$+&A6|*CfHixAEZnE(Ih!bSTZ| -B)0WqL)tR(pRuHvfrL5LNI?D>ckUq}>UsFElh(hngm@>QW)+tV3`Y~k3eeHn&u;Wg`!06`p!C*W9--` -qycN`3q#Nr{+dLJa{kULNj2sI0gK!#at5W0DI=o3jVHv7^XqbDa5d%lK@!qnLkdsZ4;o2k7e-qzz+-M -k&BblH&El -RYWTJ0iRC18!380sRa;*;mLgLu>!=eEBwOR@Nb?B?nMw|5fweh4h#sa?jZ<1vDZ~)z)di5>@jfO^r&~D#O!*s&(lz5+VMmLQKR6%BJi+*iV=LO_&d@JXuxTh|xfixUSkEcWpKm;3q&bUi@amho -KxGyl{JNW6#F-Dya+PS?E`)OCuz@o~lp$#?z^jyP;XO2>JRVO*spLa8|&cRP&u5IQ`26a<#$20g)ddi -cR-lU*3GSH+86ZLc}v3lFbh_8(KkYFX07own&nAe{4%9MleUbxEatw>b&nQ^H*Q77hk|;#40$?{nmJZ -nNuS2Gc`f>-?H}*0U}=ifx#XIqq`0FrIohXGbTns3B$x2wL6uxUL|%u4QzBxD}(YI>iTn5%i-)9_*$IL1P>%}hq0j8qPM8YTNSYjC}*P~EhL<_SH-wL&WYBoleiY -i`E>|M75BH|K3dne&K3bPA=~J4c?^b4xM|svgHaX4P9w9#w{b3Fm)~4*YYhSE3<$W3Fm!`(Im5~In%$ -d;v?##jCqymk!;(@9l`x==jd#4GZdU_9qHrjKXbQ8dM}D -{CA~GUKc~L8b)LRMcKiC+}Ka&x;u}PeYBhle4F{Zt57}At^QKm?=c4ZyC$7sSW^Kb(Ss5LnWHBOjeBj -o^^e7b*NSlpTU;D@?LSi>qAGq!W$6BG^ECZCLflbefa6kU(a*BT6z|;qG;?YE}+&PttDRl- -ROLQYv+|*`hf!oLx#Bazb*nnr!WVzwXAX@!&ipS(VC}F0VL1184w^exU)D>rK=4V!kb_zO1`ZDSt1Ts -wx(J+1R9I-sg)d@4CVyL4gfG>7R|sXp2o-m2>^G+jXN=+pI<2w)sw|RQe&)S#vt~Nwcl{qMeApbw&GK -x#;EF4`l~$%VL#ptA0tz%i^EgLe|4@6h7J$u_}vd3ICQ|W>RH2)e&8$?06nv(Kk)i+1#pJs0RDp#y0h -1zS)%Z<%_m?zk3ZC1_Tl{Oga+qy#C)8PqQa~e){CE&z^rjp`^$4PQE?sX^N`T&~SCj -Y*aI|G;jOgvB4UTDE$zQLqXuT-(F9q~^yU`{8Rs;BCG#*LRH|1uUy({Wv(@xuRaW%zx)jfF$2`}OI%c -27WN29*oolzxvdi%Z=c_%V_$=1zsTNNV$$gc8^Qadpd@Q$x32b+kboG0T^@qSTkdhv`ZeA>26dnUvkv -1#+mb$%x51_IZ*#P@6tRG>2R`NAEC -dMRf>gGnL1z`|7?>L;G6Ac$AJDKfBsm$FYAf^6|}rA@^0G}h9>3Dby4e -n^~HNb=&#{@G%8mvLCR@V9gzd(k}r{|%$WEIlH@v@&iaLm!e;zs+x6vY_hZptHA|cP6U^wUYOeis*;7 -evG8M>l1e`!*rJVV!nZJQmoWeh6xSq!1?l%G_O?D_rk|$RM&gn|zwUEH#8c1}}7I|MR#dXY*0g@BPhvyT-@}6JWrgO9>aS%-}c2h0A3^)2L11gUNJGTYFg}s# -6@-`K$h33nzsnzD6sAnuwcab1shSrA~Rg|5^>qk{&59f+UL+I(ZEm3eqpKBeGZwTzbZ5XCqXflT;Xwh -*`1l$hL%LCnu7sj_2~Llv6s|YR+11CuiI^b2>Fcx_ZHi(SWYC5P$0NoU`oXTND`=CVx{^<|K0&+dwEH -GycT@}pjFdj4@(qMz+pRqwJ6vpdB3_E!-SSQ`LsYvT24vnr1XdSq3kQW?8_?q5+Bno|CT&Y}T$@e3+Po5(t-N69+;G@JQ~M-V|qks{0n@%l -T#DBdZwmC|hpV>s^NEEqjzdz_ghXFtr(J62K-Cpj5NMkcQO^GnkJcT87jh%^pLtww_N)sbB6TRA`^34 -94?$Qx_rCFaV|XOo2yXO3Eq8ndbDTS0sHW0y#Lnj%|&1gvth4UN0-qhlt3pX;~>Q3w?pUl{GL~pvEj) -!>DGN=oSq~1W`;cr{d+)7e9!*NIZsc^LKeky3mw0W>A2Bfx-nfLd`|pmG6pbH!*7rCQ@^F;Ofn`x7C9 -Fw!tZFo9*S5t$KqhA1x0>O*9e%V-*mkH*$7ysv2ftwT9VlcGYJo#@_{8-cnpsFyWBpcB^5#17Ea#kgop#bjkRRO|Qv!9f*Bg)+vE1*nVN?21M -2WYu-N+Fgi-=Hwle$%mpBXvimO&9qFs)0Pz+0_re#us1(!J>6K)xo-|mI81`>yPADaV@H{wtRn2CxfA -7JcV^*iuI-Ak)ON)y6RoI=KeB^C`jFdl&j5;gQBD-y?`O_cVTO$#Yr?vFU*;KiGw+dZ|3}p7qE+QW?z -Rv<%*oy*I`gvmNWAf2eeCbW?zTuY(a^d-G2~fyF)#80gzA4@Q<4e)RrN2WJG~8l%y^IOyn#{eC6eiz< -zH~ev}oeUQiEKGXde#gUsCm{yPu{Z7+3LvY1Hf%hJ!jToa_UvrsITW-ZJM&18TPgo2&tYoZAKF4OmRWf@h8%%SMSs5^x|R^YCyJfK&KWf7!H)5ZQd3n -MA2sgtfP3JqbjPL9mF+2N`tCN&1YKmh$*KGeL&jBctDpNQW|kosJg03e-(*`M4}ZEFhS4cu!(#NY9rV -a3#3xuVS>1*wrfBGRkjBzV^OkthlV2d0rVVt{q$*<|C -xl$MmFu`l-jjoxBUDah)gDlp6%EKwjlTVGu3XxIS}~Sc_twFWyeXy}j7qzjx2I3#0zrm}m4@4HCi7gFu@FYT0)r7Of( -R4BIhmtWD#wD23|F0$>;9^uSGnsVa&N~ -gu=1idyK?{Nq?pu~=csZA2_esT*{FVJjfv{=gx33D*%{3_6G;Zm$p!k$Bh+-Nb?`Rj1zENNL72+I~AR -`X0M5@1e)jtiU$l#U*Sk>WVtf_C{QojS5YnLNBm%O*2*K*wi6U)E)}a97=imYY$E&@31^0Hj`L3z*I< -OS__4O@zmp;NP=g0%7?f@Kn_wr3CGJkNb3D&w&M}Sjf!bzQ3+d@saD(C0YKc86^KP&CS>n&mBxg2_@h -!YRqUE8tRskhy<%iv$hN*XBKqkQTdoHU%lKcI9oyI2{!s$YM-new_-4p+plx5;#uDrrW8ba>g=pOUubbn|`OPH8_0Dk -98ektyd>+PYceq)kZtGIBWjxqrauM$FshvVnbWGIxC*19{N%<}QO_^pI@zXy2Q}=r2 -|_Vj7d$sU_}AlZl(n#p}!wwxN7v>IRkNGI7Rf#RWTxOcan?@ -giGQY?DLC?^Dx%ynjI#sy)Fa>mHTuZMWsOi_#9Gs~!EoNPEpA^lHT1_5WWE^^};*UYZ7X)0zbvE-NVX -sg5$qSWkgL;xZ6gEVDFP;zp06>=OZV&OA|#MGpSyL?34VvHhZKEZ|!DEq{l@eocKCf=lnaMB`ir`_9U -vZjbP_5R)y8$>c5W)y=rMcJ0+byvCrX@}cwPpz>TUzNDb^`NRGw~#RnofXN8#cf8l=w>DcGUcn%kdsd -EjMJ{)f%Tfmo&;NQ;c!GKiSe?O0lP*W2x*q*&JOy$aF|0Qe3jf#NF1lw@vVvFP{trAPQ`iy7LBBkBs4 -NrFPA083+Gk2@L8K|iD()x2AjW!BV{=a6tmjs2RU6`DDfhW!o}k?sLS6v>Qapa&Lj%b -gHw+`L44iQ(d+yu%no!F;qbflr&y*Hkp?r{ilZ9`^IFwf|&BhEt}Px4&?q{#3_3r${1WGBKn+U9$n7c -AKi~Nrz5lR4SVcTDkWkEh{Gto6XcBGhv81fw*%tC^~eO=!ii)L691)n+4lr=H2s-wSF`J!*y-52fcMBLmkHj&(p*{_poxiv+HyZg^=!8-~+&yVOG?l`EXVMbvt2g8r -jM7D7b3o*GE<6oIDPR8+pvISVLFD61ZG`n#d4n_yd9snI04LW&ZMdQBTRMSnfr}ewC#W>t5U(c7hwItJ|kXh*sVs%S1W0 -DXFIaY+{S$wD_nk{RB4B_~R9D-KocVY;sERK*)GU(9!#+oswUnCx3?Xp(M_X -aSSSJ75z{7pk(0mY39lf6gyBfAk|#lwYJ- -@FCZK{f~xsXonjZsm@eo$pnQv5-YbrK0FVoHRfDffP71B@s>B0oa#sIuTUQrgo?T3u1j|l`gLbdN@kM -hYUCESX-8$0w`m(D5@~aL#MK1ADD%DtJClpWBBjK_Y5Tp*{)qw5lkF>m*Vvz)%^^UAtc`50Km9T^r9oFQ9>23>re3 -cs3}bF?Xe2ksvS{2M;g0Tnrc#a^Nf7m$V--dz9IV%&9_WkL>GvE_LtSmw$z+bleh -wy1~tOe2(3qq>%jh#@+sPgaauaOZ7-?6sTS{ZFyNjLnd&74YzqGUk6OjdG-UIY{9Xg$>|0u)HaPg$S} -Te3S@PIn@A?@`U+2#>6||b?Plq`HKg_qDC=B-U-&Br)2CKtHlmTG7t3fo6dP;WKvrcOUTjEZ92Xo8Zs -N>z>ML>oLZPmZNEwZcgKkn85YZ!dc`CK*AGKbepjU?uDLHo0g$e~?YBS9q)0=foL>6+d*lNZg2MY{Iv -%6nZJaIHhHT{wGi!{OtqZaCvBwwqm6iW%>B?8s-1x2>_YPFu^uxo%;+-gazM70 -O!#>=m!rn2dEmPaOD9y9Xw~FcWo@rsaFP-LjS-rz4scn%10iOkwdukK2X#jW@ZFGX+BPDJk?TE9LtkMuKl8k+!*2)CR>_T1)vf($^k$W2&6?ZD-jnUa%) -Sn7KZRQL0Mdk`*ZL&h7CW1XXrBnkj|H1-X?n%THrI -(&xvi^u;I`LN5{ARZmqq@h25obz!%!N&QQFZr0Gzlds|dHjEJ+h9RQQZ{7ut!C7z0>Id82z+`jQSiI_`+bYdVan?4G}JPH3n*eCU)0_w&m1MBZU==y2>E@oj+}e^<0Sdj*ELi)dbxR -ga4sS|Ft9HZFJp!@HM~kt6SG2UhzVIL9lG@2Yt;S^9!Cb+}iOVBh9|VyE~+7~%L?&&{gXI_B_xtyi8P2gpe3e`{{sow2Gu!4b=8+X^bb4|TJHge0uO3+zS}>D!_d<_bik2rGK+3srSn~nBMc -Rc$3YoX#&Azb`P#fFN#BTv>=Vuo3;*0k7s~c0{UYYB#-AL_vHr(F%Oc -$?2Dcn$bv_c# -Xy1X=y4bcI&d0Lop{ZgS)2|T0)NVS>lIe7sj9oM^%mTNf^=){r0Ye*qwBJXL7B+J^s4vSb8v7eKC1mP -1JUrOcWVZfjV>buElA1Tm7>pe)@AaTrVxb@y8**{s4ifQ-mTq&f6~-KQ!GZ@!Ex6-uI`k95!UZ@ox*i -Vi0)Yvu3@~$3D~uEr94rBK8^DI3Ll{BCi!)&F-8X}YJOc@HRXeckFx9K>dkfars(kp{13mI@=do~t=IMYP5J2e4 -~F7LhgI4*O3deokBxu1{lJk0I)*8S=Uyy0lM^le%;W-Y_}K+>_m -CzeZV}aV^g9mQz2MtVmqg8f+?4Mln(d#K0iGRHHJyw3R2*7g?Z5+}N()f&lv_8NSlt1FGk>r%?SkQvG -n%r}k_wC%4{3zuO|>>*Nn}EX|T{4gsr)o#~hT;N)0Bjc7lZ@w3g6eb|2N+#NrF&cvQFe&f$Re(^!E9$ -clySpO|9>;bNd&j%?r`Z-#<<0IX$YJ#jPZ8RPXvX|Q5SK|I~T#Z{5xge2C;tk|S7SUS=Pz+vADdqUPX -JOzGepq&tpTOS_@b+= -2G~r`RUb=5SB?EW3ib%4%0Ro-yeNCRPPE!Do%ADZ*3=+8k&W{Wk- -gX-6hB@=N5SG_F*>@)fqR{zEy#P{S7M=vJ2k0eDL|d`64DQ>QL<6Vvz?P(9QD9IbqusCdNHH5~h6iyW(~eFpVrPz2%WAFi}mA@Q;LiHlWmZpQOfNVO%t -n}OcmIJ~Fk1TgW>p|B9FUtTeC2H_|@=9B0!A4oJGrMDcVvmB|f9MDzh2*Clmju3E=WVtHIewTg5`7pu -_a%fY_kgBU6AO!-Ek^^bgaa78@xac3?5Vh7qs;pDs+g`egMmiu*=geCpWf68cv74NOL -l`IKm0a;g4evcf>P{6Z!I8p`mFL4^Mr_djT+rr0dJ7lJ^f|rVMYT!acVBC)xsuvwgjAAn6n{T+Q~(Cx8z%nDN=O35vrcy@PDZKkZc8>e*b&#FHl59E -R{_SyuH~r}-D0tHDv0cD~^O`;P{*#zK(-Vnlp(!`X?g3a?!uu1BXgPc@^ZI!AwpJSivHC&4M750p{v8 -5|}Laq`^QaSiHx)H7lvnF1$;^I%k)2D**QhjkhkjCcwYRG4Sn;*f%z;BU^D-dFoxC5hup(Ur=GMytG$cI@ -en|w~I!<)iL5t`seM5sk84U4f6A&KLQxPQ`~+u^52ciJ0I|SjQs>YUE=rR#R^t6Q6ere?@NBC@$mSkI -!^k^pTl_Y6Msj1ZQ?(I2X~+G6Ask1BfjCVt?726LY+T2;zKAO`)v+fK)=Y>A#}`zzT8}fHQ0|J(MGxb -c-wPxPFsJfk1-wb)gyD!xBF#NeAwmOU9;-t#om?$;+ahpU}{? -db^JtstUaMmZ*AKAa-E;yGFxCf>U)8-G4O4R7aqNIB3jg4&N8k&eg+$#`{K9kavOY0( -y6nPKRqB?}A;0JtfceNcZR7EQiMUA?EvWn;++i3J{3aCM{GGH{TWD*Juk$T_2aE1N-Coj>gal@5+a?# -cv1O1fS<7u(<~81;0Tb%xUqt&T@9l&xrs` -lHEqNJL#lWngo}7hC$EC7Yo4kEXU(f!wE&H;&60bHSKL4K6@Vn4YU~)hRpj_%C*mVT#-t_lgxfs-Wf_ -sknI{z2c^99Z?{sy=fZzA|2xrnm6L~i+;xO9KQH00008 -0LuVbTS#u=wG=M^0QlY3pp#F7o!KnsS~=HJ|4bg0AyU^Iujur%tmro6NJO$!)6sG)QJeKA)P5d0Q-UoeQ6*` -Mk|Ixu|r)>+J6I$8WPHA0&S&o7ECO=F=aSc}-p9?2D|u(FG=XF;`y~*;0SSXZ~7I(W=%d^Sr!nZ^k!8 -*{Tet>8}9n_r)z#-N1lyYI{@X*%Zc1XZr4KUbche_4n1}c95J=vdX$%+F>|@fy}S6$?X|@HR%FQdXT{ -H>@Tmr8$bJ(@1Fhf^oJJ%`t-D1^TW4AnbqslANlQBHlJr#^IW}EWt+ciUuI=?o!9*BInIlEEG8{Ky~x -YFE++hP+QL{E7~dC7tDi7o&M!Y?Fg0hZrOx(4eusVFmmiln3p(o=G(opdl(5Cc&J-=9h-IymS9md -@OcpHTW2(8n^l%9A=@&SCVX$z1{Sd>I+MH~D;anG|L6Y7MAbCSMJYkAF)*MCLKHE-xF2x>yb?TGrY1B -0EXS3K0BlULWub4RyX*G0&UgUjY-Y^L9J|2-RH8;!wTmtk3v$ki4j?)pC&hVb!$7Z2gkf5a3B(PI=HR -A`EYe>zg_J52LQZ{EPjhGX1<-u0@7F@axOsU7=RNa9+GM&G;Uk)#qhhUDtWjjIT0N`jz_jZKlCDFhr# -V|At?6V!O6Ie1k8d^)2?sc7MKFl&}!+qs{;WZ`J`k59nWlaB!A2x2MmOkw&Yu$6tB_n9de(L$?$YA)C -UWv?ss!Am59;Jp&@5mjz_ikLTwXKzrs@)?OSO9E?vPi4kIGfQ?Uk1C70AodM7TqFl)sSMM#a?UeuY8A -p4)-cv;rlD0~KwiMO0m_W_zWVJ*pm+0XiLf!)^G|gu~w~Ml9$K$jC6dagLCyD78w5Wxl7#CCcXaQ&yi15~d{8k>G{_;vCFRAW17n=gA~lC$aCteoUJ*^IyXZu0rv4Et`f0v6!1k9FrRD_UI`M`MT -|Z0ZeA4j?Nn+8pap4wto6SXjH>(HwN&=P)gUT3L3l{I1T=tK?|6CO%uF3=K9I|q^+_MHb1 -DNs?!vUO2a;j>Z=FOxos68jZ0o5?My`QT37Afa&*yBts@S3uJOlOgvLz_5@MLi_xxOn(`wrZ=fn%A1QxQ2CcD=w)YSN>w>yTL#X;@L0`r(cpl@#;>`l?#9hB$GLMRh%h -mZJZ590)o6C09pYahJ6^?KLHLbxx6G~y}VSlJ*}@BxsbRiNgD4~|0MZyHeUfpj}_d(RN#isJ<3&ofSkLn9 -vbjsF(oN1rU3{d+(H2STqHt2kt)*ACna)D$XA#_gt>+6vurhQ8%)tFvS+JOOERU2cIr99l%!lOuAsG0 -QPbwj#=uL82Bo1?k3`>sv;;maN*Cr2E;Nx9=5XdW_sGJO%RfmGyBxSrAsOVlVxkx1KqD%`nN>ARRi@L -DY=*3Ho&icIB&$lNV~Dn5h%C_vYkoG*fl+R6a)Eqhf;)rT7g{HBj93mxjRYWJszfMuS3?Wo!8UIK8&- ->=BJOCA-{dYz*tTeS*lo)vPHLW=nLv`)ATye%JW#D3DW_jW2?M!sM~nJ)EGCMX|J2YJr(^ULr-J~agn -=~h8f2ex>S3xOkRe`;_o%}nShyG*x@xO&Y3ufCGE<_X4nvD=9fBTjuK+V683b5O1Uyr33)|&XBMbW_v -}dte(8S{sRrB8TL40oS^1Mt?B%ulv`_=c?<*}euYMq5kUgp`B)ZFOQy#z)1Zag!|3NJlRCeq>l8%U<0`yMRounD4H?FpIW{n^;d}wjOC<3Gs*^2WR03x)|DA-7 -V?7r>9yld1})+d{p%)Jq0Y*4P6n|SM&jAcS<%~fra*> -eFXegj%1qDJ2*NWI-Cx1@dg+|Av$ORxPKYWCyNU4cD-HvD -tO8w&1%{Zo{sg^6YNbo#)zVOa)hqlCjQ8}K5Ds{9Rg*QYRmrk~g0E!N>;ME#fOj~MRkm3$=rDBZxas1 -G_Gpn^3XtmJ@MFxb{4<-e_a+AIRK(J)nwP_$@g=ik&bF@zyfo$JU8DCbmD6q}(k_v59ele^-7~6KeOX -UiLD$wuh&7zv-795Co8=+zDE=Xmz*aH&g7AjL%JV^2P=$I)YMyJX1+daXUA_5G`IXi&yr~;|jStpJME -?Q-DcpO9Gt9doK)yh!OqZEsNBq|fk5SQt*+e~O#rJ4#0o=&sQv4L4#XX`<-$QM<;ep}>s13G?04MOro -!gQy&E~`2}FRQ~1ZuPkp&?sb$mA<|e(Nld8QN-K-WV;6)0JA8-x*c>5dM0p1}x&P9q@xHr -4%e%XG#**Ix-w?@vwr3q@xs@bf`dDKt~$9iO?*bV9Nvhv4`Krs+qzH9LOqXKzO0DyzRf^W9TZ9bpxzD -ZDSk9rwoN3n7p0h1LVzs1ZF(jJxX`2Z#E(X-INLlqa1N!?lr!RdJnu(Ah8UAutu@#)1tn*iX8oF -)!M6!cai^R?J<7VhtVPqUPeN0s$h^>@#XSvrjqUWH=|u__T{B+RCC@5RipmHZqyx(~cT8oe`;892WHN -p4}83ua-U~upQ>U=}ruE>jtWFeZF*8XUj42T09K+!Hv=tVgwlw(xMvhia-$4?wy*@@kZpJguA35{|bjX3j2ExrzH=kB@d@1d!Ckj!tq6L%+N{klB -b3(T5uvOQL&wtBU1dx!&$@M!db)f>aK*vKULk?;Yi`(;J_RlD2CrM;X}*s;XecC7@$kD-&e4RuXY3#4 -OWW;$W?x$%rsGjfD=b-X0^;q$zcjkq35s)AYHH+&;7vNUguW!&F8(k?Ey<+UwfJhzgo>^AS9L7Ry)86 -Ag)$Z`fhkctjysF^SZ9;W>bzbzeBNYQJn0a{1wzC_(nPp1()8@ysV)=xXw)-!{o;j?Qm -Ix-O!RuaP?xHxNI$0vq+652IXY)5&aR{{Ccfw8sND^axz<$latG$8e*-N7MNbxp>77DnpD+lJc>w+b0 -4&xWIO2AwpxfZbku=qtVnm7k5cV4SM@W1Eww1}60gRAYid(Pn(du&m0%~Kk?&@GSn_h!wCJb@nEJMuv -T2_lNr~9%ti`BuSbIPySu{yS8u*sPi9S4z@i6c~+4i7pdr-FhmsGahBZ*d9ixVITPqcBW42aw8=NWM{RS5!D%B3&bVi+7?zqS}&&r>~mqZVrH9@{5I_H*Uo5#c*h-s_fl`WaIi+d>VA=CBbP|0-2{TxTRow~scz6^MPDxNob<`eLPQxVHZAm(#JqU=jF6y;A>Fti&(a|>waT?5lJD38 -w&8X5DKrmbbcesR(dMG$*r~fA=Y7s(A9Dl7H*4H4qz%8U-(St3`9VJiVq85G(k!XM&P(!gZT7*gWjujEU8aE_jKtYx&jmz# -2o1y;j)l@lT4T3bo-IjqV_qeP;YKRhqQHP6Wl(+*w14XON?1W@K;pisH}u_AlA&!9O>@F4S9zd=lQG^ -8JP-24lVH~V&RE}E5hNq1eIF72mS4ki$%v#AhB<5w$3i91P`%kS&n$@+$0_XNaF`LQHBE?$pQHYk}3q -*vKph!7Q4jnyl4>j>|p5)#vimH{~-h)xuS)Y;%L@g>5QQT8@#SUfBFVs*UZQc+0NwLL)wuyP -94m;&g_10GfNyA$zl1NGPsBi>nmCtc|3+|TWP&{jSznmQLqYtE|qr@=}X`6S2{wF)%p>0 -UG*rK969gA$6&gZ< -dxd;wMdLuJOPw1?}DIUFj$H8}Y+{#GrtK#*z4C+3-2x|iGLTEaJ`ldC^mYJu(;~oRgvrV5=ImrQNPzp -6%20$S>F&VwXlSBNkrx^_^St%Gj%Meh+b8McDex9nhT^fghQ0urx5#MmC5V-V|%Jw(aXyz)6w`7G2J5 -fBYsLC;GJ)@PM*OMINi|ahfct|cf(31V0ztdi4z?h>kcgM2_H2d)1WavZNfHAPUq -Mokiz}$*lFukiX5{Rs~x7w~+wab-fmBRv)g``-`B=us1``R_mS4?6$MG~!qSsu$3p*6;LC24Uvy(`KZ -{-p4cM&&ODu1G*w*>X82>%3V1apj1)l2S_QZ`$rCP}`pw?fk0AH$nc00Ce$b7iLtSBt|3OvL8xEVYw) -6ActDezjrC!oy3@F@(yymM*v3MBk8eu7JEjb72W(+AaJTmM&dcTzlU3|mPFNGH<2e~#DCl*F&!iE>}K -lZ3!P0J|9nH6MfA9i2P2)#n}Jr6dT06Zotn3Pj=;!^ue9cn(jah;6K>kg^pW3|v+FaX!H*n^>jQ9ZT3 -l6$Y$7*&D+>_W`z^AG?SBRM16fzIS@Dj<0KQ~mxGW~OV@w=pC2(kz?qV781+ln@=U8{||NPJ6qyPNh` -sx4vkAJt1|6$Vq<3Imb@4_v8d~|dqaX~-f6)tu~(BCK2_$Wg%>%F%HPKL*`5AV4PA7H|tG}{K}qgcT} -S}^tj-ja^anmAKb=@Cv_>eCn`%{WwDyh$LhY_uichKmnjChXN1OG6ioNdFJJ^H^#`W6VM4AarH^gaceXu_N6eH#t5Rqj>W^r% -Gpo=xj09-|xD6K%wp>kqq}~$GXdry+dBeIgQm)>n4KY!^h{b>ByD}OupWM>~2VlciH3V-+ll{DH*n_3 -rJaDL$@h(qD>&sCHQrHk)RRk6n#K$+IHETJb7|mv^T3Og!82RWvc$Unpam(7Fh$!jz2W*bhxyqRg0or -wK=&Uk@^8fYg)WrR8w>WN=9%|z~g5nMk?_zwN3~&<)-pSFI(+?- -_RTJrL6N-31xEcho30$0^(UAOUP@Bi^fp2>QeA|xuhCQq>4BscKz>wjxfSq;BycpCXIm~H3fxT;PrdQ -}m<60&ICFI^TP^o3YMYQLz@WT5W8WL7SZ`DMH%I;kUj3J6)ds)qXc&1vsH!@ACRoOmK;WzT(^)Pva9Q -_ky>K`ed< -lisf8_r95z>BGo^{q?tdGdzqZ&Z3Zactyy9<&|hqMV{RbYgT&+lm4iX=x@A}$)L#74a|fz^HT#sU~9Yy#272o2e5p* -ZRtp}|4l5vmf;0I$$_$X6Haxun9||3Y!unM|j$k5@kJ(znp%@4QD&`TjW5$x> -gAjAaq;?f%3V}k;X63C<(p)ZmG8n!$b$^(B*d$A7y!C{*_)Ze$G0o)?UQ=y0(^OF^Y-TrxNGP~pS%y( -ja=ezaT{ZrAK#31e-kZ`N60Pv;g%G4F%mW*8ie_I5c0~uC<-5_zbe7j-l)K}8b+cMzrMWtQJyvI^jVD -$3{jlK?pgNycD{s}fI;3q!Mf^WUl+D;6hS=-Sa7RS~*80Y|6G!9yA*$kbYW2NK-(-kMsl8}h9K=(S^q -p6hHp(Rsq;fk)#*a)Hu4^D?#=i^=vOD;KjmBXjk(udSQeOOpGN%*h2@ZNH$1y3j2~pfo -?TZtxl|W`ZTXaN^RPnQVzp}VqrTYdglayy{5L(2m+X)0`|INk%IvmkMg*mi)8`1D%}jRwHjhJNzy)z+Z -QJ=?%xlGihehwLHfVWp$`>-79}TZlFCJlthM2W?=!1#&KfGSOuP3IyFUlAM-!4L5}@nTctIzpAXh53b -uudxila8s+v(q9=yvCjtQp>6CfwG5bEOM^K+SWdI2HasoWP_8nwck;Y|Ugq);V5I^< -=|_hC)R^9?}%sg9MO|ujD4#m?&_6}s>5i)!Iu0dnjO4W~7(wro(zvoMExGtM@ -5L8Dw&9L)5*$D!1w|QfgXngK@Rm#0xt86y9}RA7obgvzr|HUsw{GSY+< -3->xvppLR9H92nTf_?RQx6=0PfCZBp6xC3l|2JI*h3A>rmCB)C$=&{W9a|(m!e&4z1p=-H3?drQajrq -u)pUK0;roj7^qXg&j5}SK@>CEFzMi+y|=28!XJCD1Kg)OV-ShtAcqKwXd=+_+`W{csG0v&^`Da{6?57 -GU`-y&nJ@r+-$*Q-F;RGfveBtI8XiS8GpUQ+AT(M){p2x<) -Ey-kMVV6SfNa0%VZ-sr)&;~CC7(t(tgV4Do?;3j?56H(c@QJaM4C1jr+|o^#94wndBoI(_l$0;7k?~6 -tAkx#{QU6vVrXMz9NL5<@!fBL1T{*Ic5egSxG4IG?KVjI));1 -9#2sU8%#iYjsb!~YuPc4WiiDyWrit7?WQ4ffaw_TUMkt(MV^w(9dIE!>R!T6E>O;7IDFmgNKd5 -fq4=Z(WtW2AprE!Tj&FN=3Y%zv?@P-=%H?8vyb=th)+TphQY+7P)?V__JS(m4QuiaWzsQpigcVX$eZ) -vDb&3aGNE$6VUziMBWuehv7Vb+rO&ItQse=U`MAJu10WnbW=q?-~o9=`Auh)fU+rHCNcpO%4{a(!8$ssuc2X0&TT6sn*Z9@rMWlS#5*o_Sfjo7s_uQcBnOD7}`x3vOO -GVpl=B$$WqUDOFV;%3=k)Xux=E2w`gOI%OC5-p+iABQR-cF`o+MfRNDfs`-G*#zhaj#>Kj(1&id7n3} ->@3*n4z6bhOPw;ApN$@OV#^7!FKFtaTS1c60|pOCBBFYB*VtNjO?NpJf4}(s2VHyyg(0QLHO9{t2@M#o4n4MgVlVEr9giH=5~TS{ -T!Get~kje&00|lCO}%Ng{j~whdHun3e@Qd&NW5K7Y%9#s6^6dT=_eakxvesCABe{v;Mi!0sEna#ba|8 -6Q$WOp+kJ%x85+Xbox2L%QFz86gP&WshB=7q>5o02+uV;f+qN&y4wStIR~-H9jHs2VGGX -Z6&)t?jHLYy<%={kOcbl&_wQe6JI1 -1D(1>9u7c3kDT(8%uIQmJcJoKcs)&>S^;-=QV0+aPa4JhjP&=`m+r8#Fv`FunV@I@S%%OS=k0gT`E== -C3Fum~2k9+}N6GxxF#6F6Vl2*-Fi56wR1)G#N;{Z*62}KhMvsVTyFh;cL9h+5H>f-@t=!pxu4-```aP -?`r(*^2eK{I-_dhSu}E3$!W_hso~;8Rrre=umnP#Fh*fMR7#{wh-AN*28E(wTtG=Of#c9184piS06)OIsW=hU=M` -Z7>n%H;2+7pf72C7b*CDg&q{D^EvUUM~?rHo#EmSTxl!d^P$6Yj -ps|)O%|&Kvt^Dt^(t@gXuAUxp-wZbwC@ka@wV6q9jgJa5YX6)z2bcf`g04l(zRY{58yk$>}n9JND}=G -CJWjWnEIcx7t#2J>@GE?^tDVdim5xkFY4ITM;qW=#~q53=EpZbjnE(R5jnIMV#>C^_CRaM|=N35R -B8=ryL;u?KSW>SWom8=JbIYPPd?+z;}Y^BTK@+Pkkic2U0FRNpt%y=Mk@f8y`)eQl7zrJ=O>=~ivp?# -PSDyTVU79MO%AoAw_no9bHV*de%95^o=3JBH}(l4hv(zKW3K>4kqE;i{w%3YjuK;^6;=9?@n^2G!AR< -9>FHPuw&%Q*bk-FxJMDf9qYI7G_p7xd)P11isoN`3a*!l$CU-BrXSr803Cr+ELumJA&3RksjRMEhCW4 -Jb#g(WA$RGZoruk^Fc+}8(4!x2<{ph)4XrA@5G$+lX=zTW4^{zTQRd1!@s#5Ke`uQ>qhyT#I9Fho;a7_cW!zI_0Z|g%#emiR_w49#+F0tBKvHf8jgPKawNed -`bGYtI=JS?lWC{?jXy3Tz)PQ?DJlLZc+4eUxF^s{PSFdejgIZDXO4Wo|X6h_j8a40@0^ -q`X09n;>hzfY2I))_F3TD@l}S-8jd=ZS}HDv-w@NZj$vTUQgi3YCLCchWHT59X(o~f|%L6gOB?#jpWnABr#9{xJ -TApkfW(cGla;Q`|DMSD$<|0X4%;Lm&yWM$|xolveyE{>+KNQzFM{ -Md8bM-ACD{E<$?1!`C5BQf<-k)umF`#UIRiG<+-Hze$DNgDhVInY^t&hG7_rpd?hAG(I{wxMK2+}H=P -i?ZdhqPm7#OPLI6vz#doXStQAeWX?@a0On1ddW=GPms3EYj&j5dz~D3f#HgJ`Gkw*VCb6zFzGX3E)7^;F``D={o7lOT9KOFVzCSr0($UwdJ5N*X8YWjCt2sK;^rmQO-Ctj -8ATG!0W`9QnSLuF=9r%wn|+Cy|uMwBgL^JXairoHX<63Z-QR4;ih`C3msu~TnKVV1J -H&Qx3U?A>w_71>%c{rjH#Xpq@_AD3XshP~h4T;zoh%8ArdQMJqF -QLers1BDa;1I9Zg#Wcl!l`tk`(6QO;F5#v{WK%ELD+CX -rm(Dj^mjb|^#vEAUg>&)pM_LmzCFS{8U<>IUi$DIgyTZZJ5S*|UEU^yK=mtxJn<{)-5XRD~*lXVpIdsLX2ob^H9Gu4yKaT(Ut7xlO5lrY-mi=QGQ(6B1XJ(V@K+gdi840*e43)SmhzB;biw)ge -!U@nU0!c9+$|P!2Xd?8pjQ{2Qgn$OW_5i-Fnlij!+L4*=^2*bKIIy9MT8d7q+->-Cj -gX@-Eg0XEWSY)E%O(|Dx%NiT6_5J4Jmb>8(3hBfO)y%D~@+b)DKR7uZbTsxi(|UENuJ^~y0lieEj-F- -@MThMOQ&(M?X?Zh;iqB{Lz>QO?d~HO5mxbynzSg7pNfk*YVmQgFrTO+8r1GfI%7X~?lw-H{HYs4Xwo)U(w)M$p=_MU&49v7H(io29@oX>M!M#1AjL0%el+f -4=lF9faEGO9dwZxyXtT?9EBk{Ejf-#i_=$QJ4nDe{q9s~Gm(GM4cTXgks!%JBu&0l-9pqkg)tAi4^=F -#L=sD+3*pRCZ?#i90d=yg6(@3L(VCX~l59;DT^^3pa|uQ>ZSu9wXBUw-_7)tN&l4 -1Rl2v_gcGrcL3dM9!qaXESAClA55A`Wt58(H!xgFxkXsA30)nBqNT*$PUUT%R4D``y+CpL=1noW(pb~ -P|t8jg7f3_-ib6v{^w5}%{2KLefT06zoz83(rMIF-RRG1qj4J|1*t{#Hs9NN<#~?Nua(`%DYO{$vT66 -#;5=mrr_V*XoiTr!&*KV}g99|F9GBT5ACJkpbUelwJ>zljgp6J5?0S)%pl2a@{$706ISr%1nq)OOj{^-#ooW*%nEy -s9!M9Qx6QmVOd1vUqoOdNw0oVG>LqD0yEfaP^M9m%bWsIWWjG!R9<3bpRp@GdV^nPGw -Q!%Qb2KAvI6q{!;jCquoWhYLXeq_5LpH}lY#u!6jxE60%q$nG+bP`YyIWRTCRtM_I#xk0n>^93w6K;jGPRdf -F2!i}QRy0Hj*uV9|rwP`c#(5nedeBQXl-uE#64Xo~-3f6|>gJ!7s$%WDTAe%)C%7w+N&A#IrPJ1sfJW -xSufO`EK-M0%t6%g}USXhl-!ISkwuVB?Giq(l<)7X^t^(9IirgEf%f~Dx7Tu-GvF0IHg>|AHKxt+tzZ -44pj1=19=W+GcUu~A{Z!%+N4rH>OEW#3v^_s)U6y3p62;x*~oV4AL`aUguz0RcYbfeHissoqv1dFx90y+=Ksn&Uc=DH%{0o5k3eMx3`O!!G4p;?_q@7Ji7ac(?e`D<}%9{>O#0001RX>c -!Jc4cm4Z*nhkX=7+FaB^>Fa%FRKUt(c$E^v9RS?_M-ND==&Pto{vIy>t)lNI5^2(LJfJ#^^qIBl{*5F -)R~)1EPSY^&Se>~P`(?*Q+DcpRRDs{Y??Pm+^}lu7KauBxv3)xVzRIP`2J$8tQZYSte~J$2RC_bn%lT -x!`=RAqIba(!%$R4*O-t{?lZsnpVGXu7>#{AQ%A+_p3&-!^ASLBN#d;P+n#CSukaxyK;G>N -w%@um#mJhWrK3lCC*K*mj`uCK6yh<}|47k1CGUVeW$!4TTD|oz({KV!_153X$M2d?O{XZ1wB!>r%)gV -5N7`u?obK^S6Wq#2y*Qq%Vz>B^%0G9ffv#4ck?*JD7UcIesSYJ8E0E5~mexIp{9>3|PQM^!*MNuM420 -2vmLp&N@c!G|x5e!@e|k%D@_M~mt*W-WIRyG{x`HP!R{P(lk4?514E7Y|Mz`I@ -h=mKfnIoM1d=&*W&~IL5#0`YzCgGk_GPavyT@iX9!T7DKBGT~*T6^V4vUO@_<)Lj_&~rj5HFva<8+h^ -eD+9XB?t$ScncBbiFn&#Y^Y^g*R&daHeZ*pl7T91tEie9Y@>{Y=6@_X8NlaoJ|j+|Ej9r2WQw;!eRVG -Qt^T@a{nP~}iqK)W7|l&z`~(@+T{8p-5qjMbQ#W114|MlZF{RmYMuuLF0;jxGfJP(m;tT8rFNe1QZ6P -vctkd4T!0Ur2E_1v -vyGa-GftZ9@bJP&lgs_6`>aAmo8+roNRo1fdUW0?(-r)Q!!AKw&{QSED#bZxEE8LD0e;gNSqq5eplnk -JB?gUb3dOM6&Tbk;&N2r6c?$8!frDzfU3)RxTajFUcg4%cJ;966v`#`m)SI)A%oZ;vhEjz|;%+7)s~K -Ckx^w4~)C=sSCwNmRRnX9-EANReB$^3*5MX%E=%ukWE4^-+(cgbq%BA#5z((zl6d`e)Qe(Gxm2Z20vQ -=!J9WK=zax-a-eKH8By?5ffD7O7KVit8(_t%0Ac)9*;Z3qj@02{9Fk$2YhtJc4w@P+PHTBNM-pZT!wC -@<2bjBZI40ymCWOVE$EJ&6HN`3zW7e>%NfMgW0$G~Nw39LH=FQ(|e*KHYx@k|xz{lq@88%2>#M&iHB4 -?IlM=5jY1~Fily1o#rGG#f@&a0+b%45Qx9_~DE{ubJslC-ccS`u}fS*JtYKF92TC#=VPm-@2Emow&2E -TB;6eiT4td&4UE_i|N33me#K*GRl?>7?iDvzPci>ljqq)M4Vd-mPZ%7ErKL(j~D(p!Y-c -?oh*f=*2j9h90q^|W@)vQ$u~aWrJ{MDku`-JnKBF`W8T=;08Sjzd+Wjq?vNF~zPU-E{1fx3_R0VS_aN@SmH~Q_+n3shD1T5*{%`T||cI%ZcobAk9noFfJ_vOlU8^pnme! -^&AWw)meToNDbL(dyA=H+V3l>;>2{C37H$P^3-hUCi0gykY4+V5obCGlqu#65oVU -y>7iuZlJz+#%S2tlZ0zNW8WO~Cy--7K3Dq}6%VkvGSKvH(zFsalPu>Z|N}s956No2~`XMLAzBW!Zn -HWP_HKX`ynrqNFF^zdJ)~F8MiLGGFIjbJG!;^k-AwU{Syk7a|NT3`z{YPKdTvRCBaOs@_I{cf2eC}^w -lCJ-~SIjn=>nj;Dwgo&Pl0|J?NbN9zHPxX|UYgxdEF=(_Hu)yF#s`;w!^t0Y3aVKPC(V|`bX~a6O={YwI_R~LqD~XRh%`P8p2-jSU=@v?E`P_}geqyQ+GCU66pZ}`{07dEuCK1=_fXl9TCHh -!dhW(6C4&}YH#nISrCZXJujMC9yp>@#n8jiANOo0;?#Dvg84+ydvQKX-^^y4dt}M=)l%rOs38~^Kt@{=y)<^xvYQwWQ_%5tQs^#S02FtB!}c&6lk@gh&F -_L27_CML1$$*d6z_4l!WR*Cm47L$tPjd -OfwXL#Uvy}pJIXN1S^3e*JKxi=A#@4y;~jQLXCz{ewl&O46&%$)99e9OaIN(a1XRktroir%q_`m0i0v -icuTO9KQH000080LuVbTb0E6f!_-N0G}`b02%-Q0B~t=FJE?LZe(wAFLG&PXfJSbZ*6dNE^vA68eMbS -HuBxS0_8m%*YJ&m#^V_->~;oq~zZN)be2XJ>B!qLF`-g`pou!_z1^`-yF6Urk_~#{hJ8mRGcGiQ -0J~1Q+qe$DBzSlld8ePZUM=5n0og5w&N-1!?3yB3MN`RuF~=+8OMym@n&FT9w6H);1s=iycZK{2eD{U -6eVXBNXteUF43o@LXb=k+-yjN8NSVlDiVVYGfm4ZJQbRIPnqHiZe-$)&;`fmc_6@cCedn+3ZY$ycFA3 -@rED0Mq+8+W!nj5iB&Qu7yfQ7j>1x~4au^yF1sv?TUKpS130&qOvraH$SYp6fgD2449$hv;-Um+yw)h -?euHelAz0Uo+Cw!m^^ai3SbO$z@SD|Olb7uc=$MhdWkTAzted_gechE6m`2)SFc8>6u$+~*3=DLIrh$ -LBeQ=l*_+bQOC43`^;4(x&4%})Gro4)Mq`RhKZO|Or#G8&tu!oX_VCc+O`7tJXo5U-J^2o -Wz)h&bC5_ZUH9$pk02A=@76Q^6fS6tAebV-2R>2?i0Eay+a80++~GY -4v={V%fu|JXa*ymL}(QfMrf8#BD4W(oH^?-nD8jT00d2|WvN$4Cz(tc*$|@9QWMY_YW1dK==H4bAaqQ -q!2dmqB&M>e*p^mW0tZuUB^=|NDOiIVBAex%4kj5$^p=*D^eLr?R*_21hM39PezPe*BFA&0W!74n+hc -8M-D?Ga=rDn7<|V6i6q-YJLrT%G{B1f8L5Hl4Izu#_1ks$JoncwVMLLd)V-lnXT;8HYXGnwYdJM!;o` -4)m>n8IoYz~6l;*T;I7qPFzG90={frWCQ@MeZY;u~jL-=3}l&cMCp -O3jRNo!$^bLLR7f!C6ncOx|ns?bmi13%J~pPLpV%lW`wepz_-3jhw3-Ji{OgL>Zxicu*`rO3HZ^L#m|MR)%^_r;>8 -QHDJxdf9okR1#|&*AUFf37EMpUDw)utr;0%u`_8mw0`wM>8)xJp4@A&2yEc~ZV{5KZ=<6!tA4Zt{j9^ -C}^L8TlHXFCo;6KB%lr2oPyt7XRun+ntw;DmB6WLVz<*mIb_Gh5|B9*hI-pN5=yp}*6h^DWFDX~>lp` -a3=15exrkdeEK=_eTKC&~$$cU_)uClZ@6zY_J=|+A<6El?+XZV0FQSVL%UBLe+B$^S=9ft`x7Z7RBVB0Z=W8*5H~1#MZgcUmv~6jUy2kAV$nZl-Mn9*} -NF<1^`W8_|hQkDWqlr)ep7YsDe9EL(Nfjq^Z;}!*lGm}ysN8zIGm_ZZPxaFi=ttUBd=a06Hp8RW8w}kUq)rBt4SQ=i_RbNIHRCRG97hO9z)tH>No(w$|g=wK`! -#K2|eIh#qXFyiuEnTN3MpDSFy1>Isf6b7;c$hZ|VCHEs&p7w -^bM`$t!Cb8@*MA3jK|A3iA8Q?uq1ZsJHwuE13lUm(buX(IqjQdH>ph)RZGUr{?3zRm0^M0m&@GOc{?4 -s?2~#1d@4#^fEN3JiH|H^MD1700)g5^Z?!pG@n5lao&X#03rPWD3Qr4}5~F@Fjx7)r8=YK*P^J+&-U+ -a$91vqtK@kjm%3JJ+fCI;J+;#GE=4T9~&`nMSy1k1*;YqOV?0FRm@yE;nldr^Vtbvas;>=)a7?PW(|&8dm(R!B4~WP{wd{6KmO -7TIqVN*h;!Qk^;21Qv4l?C%<8ooVSW6{`HfrMw;a<6%C%--dBv43+3Y>ubl{t1IVjd%!+)Ees*@ft)j -UkZ(}JWH9EP9bVJOwq418cGb@7B|;9;2q!^8*aVLU#3+t-fX<(Tkc&7ORnCNxngKBw<{zc~w>87g6gTC>TCI>7ts -dQ$d`4paQ-19g-<;LO(t2Ym<5|x88eQ9xQ>_2W@5q -iqa?o%sDL6_3JwPohRq$3O%;eypK){H6{>T?5v^=ElcWa<(b@Vf2-GG)h`1k=9m2k(D)UjQ|1O<5H)u -HCvw_t6F$EpK-QZJJ7<1{nY1$5k-QtVtkU=!!5VGkAQ6R6)Xln_`|6J|(uSPzhTGwG3RUxg6qpZ+e7Axesexr^5){mYP;ZPQVdrcCv3XCAQlZDM0K -UM#cG~o3D}T>>=rO5l_Jj&PuF>FcoK+JVmo?N6R|PqB-_FjiB2dovZi&(IsNpyf$Ss%WlE=P~_q(m*! -nuSf`Q?F<@uCdb -WJBUdhq{e>}xs4nxm9oRjK9GxB0Vp6R<_`TJk+`{^_ftf7Lr8I@63{73T!NW1IMD?ZX}P;Me|;l7b#6poc84zUaSeRQ-0>|V+Q2!28JG)NZ{=S -Nqas>ODa*}^d-*k@S@7+)WJe4Y{Ra;9f9f}kA;ic}eEU6#&(Uy~uijm}dUtts`Tk;Tj)Bh~Whnz07YZ -FXP=U>JUR9-J_;njH3B7v(zoL&qmB(6zJp#&$F63ViYM%LQ$?y|eK(cgDgvMk9qg;p3FOYgm%_hv?gQ9E$87IF<@ -rtg<6`ST?xJV8jSF>@V5*sS8#27xHX=ve@CH54s$!VV~NWfCK@GRc& -*s=`iVMUT+%w8D>CV-T>H<@57mwS0etM+n1^i}Hv@Z0@ID@23yd#^zE_AGTKIh0sTZbdx@t%5FD#LHH -mDaxADcX3ok{R${hgD&0`=GjBA8FW(?ls%k4G-rPfgjBp7p5O%zV6rt6@9MT$A-hn^Gvkbbukq|Y1{3 -R>}P9aOgFB3 -K_2|KR8{zJ~nO8qI5~n$UK>~ot;e#*wLv1>H=uxh#^R<9zt3~tFvk%9!PaP{3v~6?7niK$Sh`iW5ofh ->*A8m*ZTJ3A^4K?Mxvgv7k~)H&Vw^&{c!Jc4cm4Z*nhkX=7+FaCt6td2nT90{~D<0|XQR000O8% -K%whUBa%{bpQYWrT_o{8UO$QaA|NaUv_0~WN&gWa%p2|FLGsPX>V>WaCt?GF%H8Z5CnT(v2v9(d4TKG -DWZG;;R8;B1HoB|e;?u$Y&EOh@vUU&QCR{w*U}UZn`wLPHIXsKCEGim6u8J`F4xsyWLcIwc^2xhXynv -N*pqF8e*%h!MnZRuNN(e8PVKXTMNd2?zIoF)Ya|X)|NRS4O9KQH000080LuVbThcic9To)u08bGB02l -xO0B~t=FJE?LZe(wAFLG&PXfJYQaB?nid9_zdZ{s!)zUx=8>LF606;c%F!L1gXrU?)?-E$*6M+>bcTrS+Y(kdcD1WR2kyeZGtQ&y@L*e*a-R?w;KI8CzL2aia`jV-#ItEoGl5hwG6VNnzVtNqu2ADx+yP=-jG~}o#S#DwD -?F$Jn={^8ioAG^?_k*2V>v@bG?Pg9P|>9MqQgYXl8GGXjytv#ugi7v4pvOVleBin9U~&*H_U$Io(uUJ -YXF-Nz7kTgIf}l40C2(bt~E=z7^@K1{c<6frrNiZhs~@(=m?~nMPOp$$fgEJ7fL3QKUjyN)Uj=zZ! -9+&n5{|jxx?fnd+V(v=NG5JLh+CQ~6c(?4kC^x9e}t0*-I)A$vaumC=+0Hlp$%5Sc={UF_#?-PmNdIx -AK%~2?j9em?$gJI`w!!fbjumE)#K6@8KcHByQ>dO4o0E-mQ2%MZa)0+IQ=;~Bh;`N+lP5a!qEwd%Uq} -%mVo2Lz%8DOhZP==3jWTQ5{rDmipI4>T?e`;X67f-j!}eGc6u7fJIX6|G$(|=0f+~H`3jqaX|8P^ga~ze>1=zQ2w(i{s8%%aRxg_(6@%meFA4nK}j*ZSX -Uu+tXnH$OlB;~|Dk9Wx2xQw3=w^DY;`0kKu@?xw5+rsBBrSymBPiAV~`PEvM_|5E3$lPbERRvUkO*i- -lbOyss!TiO35To8&ZssRTY5x3@m#Ny?=MFL9m98*gUYX(YH*|*TpleRGgY(KRK@?t{= -WfGO9KQH000080LuVbTasx5ZAAqD0QwOC02lxO0B~t=FJE?LZe(wAFLG&PXfJYgY-KKRdA(RsZ{#!(e -$THM!b6=S%~5nuD}q!VNC(n^R1SDJPLXfp>>6d0V0+u$tC08+{6c;SV>^!RWZP2pATQm-<8NlZ`DWs@ -5&aI-ap2ts_-^Qhf}c9xcS+Ldm4dc>)47#Pee;^BZC{VeYu2{XE$f@AWnB)Esmk0*{((8uzLR~+obgM -0o4X%B3<}-K9IjYbGjYY_zE$Ja-}b{%G2ym;XK!Eg1MfzQ9iwtDCh*txcE>sgA=T0Q{!!nI3yK|-d;S -K0CrPqysg!WFZy5<)Wvc}6O4IZbTr$E0OW)&9)k7`l4N%zjf`8~cMOzRE0;J^~%ZkKozZ4s3f%K)g(g -ud|w^E7R#PBYM-q1nYY-OHkbn6Rv2Wratl7Stx9L{j~0bffvqQo1Xaa5uBwW7l33g1ag|pvuQHSfrr%OZ~z&K?_WbH=-C1h -3&*TS0?rF}`wOhk;=h^$AIh-hk+UZjJpX;*ad?j`+2B3OgA}tDSf$XzntQN^E4?w6sczmvl>&S`zVN% -UNmj3B@hgAW+=e$Y=p|eb*j=BCoo_vz!YqjSe$NkzLgBi@spXtQ=QE0uY{mO)g0C?Ns&>== -t**!NzAs*4;NVV8E==^bf6na3%Hmh#T^QSb<1@h>J}<{_kzW~a7eS7TW#Mc+#jI1udjJr}I5Vfx(00F5hgQ!Nsg$HHOb3*msVmF{gUo -Ym|)%c@eQeva#Nq$P)CaIT#S1L>TDI(8-{n!+@0!Vn$eOmppjSJh0X0^?)J5BDjswxH|2GwM+Gxm+pU -{y(cB}fJI!yssAT4Cyeah(_l{;{|67JgAaVU=uLY(LM;#T7M5T+x2BOFeT9iD}Ya75OlMnD|}UAEGTM&IDX&DqqLcrPyc}rMRkpO>P4ZGy5m0-c83zj; -H!N+(j4*lCpOn_FW#9Yc387P!MAu^IhgN<#g6n9mY~n!mtM5BJySZTrSnI!@Od!^)D2DF>GDOE`~y%+0|XQR000O8%K%wh4~6wBbOZnZ8w&sc82|tPaA|NaUv_0~WN&gWa%p2|FLPs -WaAhuVd5u6=+FNiTq`6)xcMWNvlw~nmTi(@GmRw0rvXuV!jwIW%cN3@~8(A -}&muF@kxz?(IQVGFI%at@4-Iu_dR%r{LY1?vnSkz9(p4@s6J^p!rSKNNOyZ!Z-FF&VnBTp&();bGRr1 -03af-S -v{Z}1>h6Tm570gCU3**1Ctfg-uz-wu^P1BC`_qZmiXg-xdGCHwl!9JTy`Y&@^8aG%dJ4Q0jup0)eds_ -q9hAi27MQQv6I=rUMYdHN8R-@j-I704jr#plz*hc^a>4qDFl-Tdz|>;(#~BtEsPOEeM@Tz-~9;W0s#` ->B7}rJXixU2Bs@{go!nrAK;gijm<*_73xDE9A51tq;sr?$*Kz|l7?NLOJ4#Erbj!ggNZO!fDl6O)9W& -Wz$Ri4g$6NJ@EMrbVp21J7rk$_`tn&Fq10&dwI51~c?xgqbR9rZ!25wnn=H^^&qPbap-ZmN3sN061J2 -&bJ8m-2nLV{JAS^P|2c+)k|%Q46vL8cAtS*uSjdz(69dgl#cRqCQ6oJ#9WPwz%6;8mF%E-m9(^~ -FzP&cm%?h90((Kys&Zeei=!DT7f7vLBa^S4vAjOzL_0@h84_|#;=9kEO5E=zkY}V1KZRj0v=T~o5duu ->cSlsFWhKXMk5gE4ZLlo%h+{}ad!&12c}ayhWdj!O!NyOaz%-rraaVh&A(6^Djm%i8t$F75n2@3%n7Z -2)>e*q|XXBGVLoOp$C1LbKWfbOA+=+R5}~jMF_!v0u+V2 -QwnVT9I)yQ@+H=O1Mm&BnjRXuanx9*>R)4V(m-w7=yZNnJ&Wipv?%ZyEGr8ZC -tMR6qeRQGE?+>YU5OABsnq3XgOF>W_b+f@qd6t2>6EyHyhJ_hK1{aDMw`^6zgd -#hvMz&gC?amBw?{f9w`u%!!{lkZepMPUXJ-)LKFrkpwJj;YmAzQu+d+2jynVzQm%J{0R-(^?VYlQYHT -V1ceU(O2nA5cpJ1QY-O00;of09jic!Jc4cm4Z*nhkX=7+Fb7OL4Wo|BT -d2LkNj+-zLedjAiJduj>0goOVzEA0|u-*wq-lrreEK|HW-uTA%VG^Idf+0UduDst72j -Rc2-hb_@*qAg76ZKKfd2LhtKziuU{S?vaFwCD?^7f6?Tq(_MfxUUO8P{)5r0uXmB=MN;NyD_l6T=vn* -qMT56vGK0c|H!pMkOIM?(XyrbM2Jbe@GPgHbNKT&hxc55Xk&oJ5_w7e+)xkG9#QZ(?~YD$HtmFM{#=p -`HMNG$Y>bAzw=N%fhbdSF}*5Ya)ar7}SOw$zv^b0~LuF#DzG3a&_r?e=gO8Shph)sp7bE*70jkYRS1H -jfcbr2Rc$S4hMqRPd4Z1jRtY5+0=ph}f9t7EA3RPw9k+j`q+rOfcItaUzEDJ_`9R!kAl}GHoK#{ke4W -3u?zMh7zSXYsnPYoMneuCn-6GeIne1xE;Lz;L@VHZjlsIRr3c)NL^{xj@!j@bI+1qs?}I(A5g5j+NIH -c8ntgIb!~}b)Yh~XcH-9rem4R`aG<7PW#S4QzJaR*{@Cq5(~q;+sq?8l@0HR&q(&+j7Y)2qcCQydZ! -8;A)lB-EY5s%o7-p)FUq2399TBF+DT=?w+2(EUID|hmMo$MAD!fenjQK5vR5t5s`Ot)h -86zZ2Y^HllLc7ik8s7I9Wfh^)B1P4ipTek+<&cq1xqElN?vO;M@lxM?Z?X;gdjP+FbCodBMzAHF}1j= -noS`s>NrwHk~*8*9mluXPMw%r)e$UpT$d&T&+sga$YuRk;fVTx)gs~iF_{GEU>GGQvUTm2l&y>a -EB2a&qZdaX2hbjj%q+R6A`vX_i57H)!dzUmhk&bl%AERgob>L#bjzo^>!7oE73jZde4({(@AX0>RoMjI|Q^k3m#f=D>A~-y|JYl -#o46{SyeSLg-3aISUkE6qji<84|kE4s@tM?aY@cs?G|L*+VG5q)=1DzfpUcj$Eu`+yj{vkSge{p$^@B -Yl-CvT6+E55TQlnT5W*q!A5MZ69b*l|W*tt56_ -cC)L-w$(kIBeyM&Mo9$>A`${v~??B5_%wK`D&v(UlE`j$lymQ)k?BJ*ae{h^S-0ibC-0>1@>qi%#YLZ -b395NE*hw0!;9AAYUWdcc~WjSc#VhMVu$d=pD@wgVHszVG4>0?nUfaRM{{RZHGeLhdO-)a)&V0JkYO1 -ENr~PT4al1u=gcus-X-pL`uNP%|LNTb}a9J<3w`D%kaUOKrhV -TH0Icz+)<@b5D{xeS}+GRnFgw8mf&@e8rcbm;F{17hZ*i@21JoI#wao@;898YOBr$ur1vY-;hhBk`jAu*A+`H-8HL06}7C7zSFMa?(@PEZ9xA2FNt%ZuB~rXw+pq*Pag_*T3VVa1SMkzs$jcTgt{7D*DI>9W;idwAV(Y`wL80#fkhf!)W&U27?gGSc&ggFze1;f{ -*}X3d%*7Q9H3m|LL;;5LO7X85J@%<_ -*Ztt1j%NT_pb__$IDFOfhhamxCaVeU7ABAazHNc~b59dkvHIa9vFDbn -05OYsfJBHrD5KLQ~k($%LW-DqqFIGM2W*^+7@c -A}aO>daa~f3A&J*8-Bg38LxgD%5HA9NmlD1)buG62j`({DCD51M&c%}vQ%-&3TagjzK(05Dlu=xB8D` -ZMnR^!Ll{(j<6OEG4N8bhi%0!)taoc|YU?R2yg`w@f(QZ2I-O-)J*wlS&~QV)^W&vOq({&>a&5afRz3 -$aCiFlO7DaKAL{ZDZw1@8c8Z&2t&)f#?_s_fqq^lKgQuuVJnV;r5` -_*jHpSg>9`Ayi|xGpMiw3;4EmXde5d1S{4obo6=5~EZkv%it?s~5Q#m>4_L13(*>eU%`i30u%jYk4;( -|+UxLlxB>pgkEt9|*5-juD)H5Jb2vE5L9)o`wM9~Xbi!3dncB+k;p9Me#QCcZuZ9^B2e9DQBvyg~9bt -?66!G`t@{L5e-3Bo+-)rj|FxQ;W1^;JVPH<_pU20%C8KfK`=Dr^{8cn^EWQ^@oW$;E4g=>;Jza^&X6t -0hMpz`sD3KP%gp6-C-seDo3^PINrNLZ4|#^Le&PM5N?Xd*B}JbXsnP -#vO6V2$;dYgWF2PTtI+zfhu}4kuMk66z~WW`x(4_=*a$7i{x1y#kFw#n -exP$4&bbkd&6grnUe?S=};3`|ohB)Dn3%!nWqDsPPugyh$+{?{ybC=k8!|2>>9eLZ{w#7!C$XVT4LVq -bY8z(=lYL11b<_@MZ?l$N?N^ezS$35&dI6gnC181=4SL?`a{Ga2{(bh|mMAB<_h1kD>!G|ekNEu!ZBu -~Kv4&cAu?tY{t;$4ro77d#Usj*>O>H6p6SYeN`Hn@%F8Dz|%JKl?hERO^>GkI+#YmG>q5NA1tpC`0Y$ -ZE~!*0Bh6%7AmOyvPW|#psYkMsBv!9yd1U9)`~eQ9A3k<0K^NR0M8iv4%bN36n&I9;b5N?bsWDu5_&p -`YQ53b08i;yrFtM6`=qOlx(=MxTT_$Gs*XHA_E5%t9chdZtg!&z^dC*UxD^Hzl -s5oj#NwZnBgidlQUt7WJ?zqK3O3Mtsh7@>BKE^a$*D^-Ji5%~6tQ<6>B^wsDxjt0BIHO0JO5d0ws<&b -;!1VfddPq*+at>LdHrjy88h=1ut+!~>ktLp?nf%tNCR{*b_HV@mo6sQok)y92~NOs$!e4&qYRhv@*Lu -=PV_mR4L(3-N-!!Iv<9-VrqP6T2d&6VX;uM(}J(o|x%$scBr`+rfEMbQxs;+n=k -@x)Dw}SUh%|X7l+g=+32>sA8|DpCxBF8gK2}TrCgV8mV5NKH~1)g6De*Ysp`R}d$z2Rt^<8@aK=H2jM -J3M`=+d9e7<~O>*$URPH6=%D(St3)O>Tu7r^vbHfi%VgoR7)|n*PGS`xjtd*6&#Hhguv^pz}O0;%$pc -v?2tFKw=aHi7z+(Pda2~Za|z}b+jd*fhNR#F1$SJ1+==w|j{Kb&&o9cB&dc>v>#c@f#-Dk+8Q|dWux& -9E)JV)MP8MrDRKBw2ZGhKpWeN4bw`_9=vcqA6QxB=gS7BU4_Lx>@9w$!51oLe>3t&b -hus;;3g=5&atbV(RH;ZZg|-3mWv&g_9Z>l6qP>p6VjoK~rcW4pj!M0JJr~-8Lt^RQae?41pB}xJh>aE -dzhbnFy72LUy^z61Co=wY{5LlI*rhF5BgiGdc?3<*MUYC0hM+uWs}ks%QUp1HT|Q&}05E!86-)S! -j@Nr~dxnR+mNP!%g=N0-$mR+|8peKb|;J2#MLW1V8iLAgR_EbC0-=ea2HE{vii+R$70tjq9a%_okn9TjKF&VH1Qq&$XyV76 -EI#n$H~S`_QAw>p4rU_O_z>m1Vje@=X?l<)ah*u@_eT3}m@VkTHbr;?}LV^!#twlTEO<{d)tq70km -}QqA^x6Ymln)+Ycxxrt_8AKIuDr)(G5r@FbrXwq=Z8VC@^TLRfBVZwz6qjrpu^|v)G$4lAmC%E*!sgx -jE)@93&BENhX-iS^AT~9J(h3ihjcZj2gs$%@65Nd5R2Y5mSRD*TtP4BWnmXd&leykp280xV#!gn>z<9YK4g}^&*x+%xjim+J;9-#o+6WA8Kzvqs7gQ1+6t}Ub@l)7_}A{z -usSv~2?QiChJBaer~8>?mlk0LKtkk#=e8V|)PxR3cd54vd61b4`yYP^z!PSVJW-Q-6+Q50~3;=QHtiM -1xxeO=Y|FiG@v5j9ZL+zXLg-la&9v7878RO5EhIhiR#firh1WYR( -Kvl2$?&_oEnKN12dQJ^B9>Ghw_9T&yk$sPOTA%60)}@RrdB;lMY18DRt2yp8iE1elypz;ODwo6KkorQ -6HA%NLqaG56oZ>pTpfH!}!{U1T?Ua+8F5)q^887ey1OXWvACT=cP02skK+IkR7*T8g<0UhzqA-cQ}jv> -ZpDfQLmjrdw(KH%hPRP(Of4eb#qY)2}t)RL+*bS9sAV!Or=XV$eG3wGI>Y`arz_p~=(zINlfs0OG5Ut -1=vX2iaCJ3a)1b@M{_0f}rE+hoZ`o>sQm?4Tfs8F$G#)8aOxBK5bMRx8?}1Hecb -n>;jkL)3iE9q-7h8&Rq|%r5Z$?ba2@@@_-H5_Sf&}s4I-N2b>g!kOl+Z3_KZ& -`%InJg7NwBD2ptaaJgZrJdx>T}@q3*aXNp$hBu7t#VGgPi5$5#%^i>_OohfJ`9wP6qM%@Vhdg+Jy$V4 -8`$eMO!@6aWAK2ms3fSz8y+oGB&+0 -04Fn000{R003}la4%nWWo~3|axZdeV`wjPaA|ICWpXZXdCgZ*Z{s!$e$TJqJPfls*O%_mfCm&^aVzfD -9opW*5Cn#r$cb={Y)DFb!La{6O0s3iNrr8A56ep|Q>3Vm97vMn8hRLj3cjP}tbAsL*K(uWHuDp7#~@W&k^=6{b6b1ORPF_I2fb|tT<$>f)i&>CB=;69D -(UnVZ66{udkxYHLAS8xs%z*0s0O`dXm+#J=2)}vN25p3N(-~ih!t&`1IgiMR#OB<8AC}jdg0(nJus06 -_AS(C?--QG;(aSvV6$Uv4Qbu9RLAzLpr$L0CXLxM=C1O25BdY;iUqL0V~1NQTUo$e?P9={_#R;p%tNv -7Osw&sBd1k9gxOhNvl^<33$Cjw9o=V5okf8z5GI*Lkt-mHLzh5_hR018S&VVvt8bBbw38)j8St_gyaF -tEC`2fnjKJkjcIya+mWOfSblrtZU%LzA{?cz^pI5BYBZ`r>G+29CS|_6Q7EKxUwbvZvs2eNd1*EPt--VI4+C$8z`%SHtE?FibTQ;RjjT2X^ -L|e6Wt(YG3W}eUEoz=cuQ_ZTunAs2%fg!1?r{1GtfL?$h?pZh)>ROHOdS*_>bWK7|PR49IaMrAyTaMz -={(yYL9A+sYN8a@f=&N}bbg&|-Z3A`S+kA|yWO^?}sA)VNy!WSO)=Y4N-jZ|I<%A7Xh>|Cg(HSHvCyO -}V5+`i<3lr8>PnG5u_b7WGZu*ovtbix8;sURSF0WOf`CxB0QH~v(bR@K(DRL*y9G+mo=3^dSz&~SQPY -MGIeZpubst1E%{zgxSlsANu+8viu9Re3Y;j%|DC~q~$0iFrARR0a_ -zoH-(m+lO0hwYigZjtN29S~25qeP>p8#&+GCDcD`6nPBZyBiP=t7IEuwb!A;0n^hC{ -0C)LQYz;%?0N>w!>6)QOFrRRaHyRdKQ8{|*zsRxjnnPv?AU=t{n7fYa1_;&Kkv;WPG5AKxN58|;2O46!81dmvbo0Af6rC_i;3 -8y_?4$(5BQf%#<9r>y&l(gWrfN -!CbbhD!cQVF)77beH=Jl9X$SDtDZ>t^RaDc{O=-v#0ZTj?NskIzmO(`5W4E}ha?r5B6nPW@+O+AG)_x -GA8tO+vCR7LtMdla0jsNJvZ-L3UN{GY3n^e$}MIXrFbBhTZgTB=#w*ubdzm!6M}`GdOQyC}hlbr0 -Q2sb4fMse>7z6dXqUqNUeDix}bw&)iD8kVNpkYUY$CHoW@0xeNCJ6RM&DzX -3lj_;x@JLxt7Z7lKbxx43HNHzo8BXR5rM|9t_L7;~LS7d{$zaQZ1sxj-d+ggD?dO=x-+Y~7YEs|BN2U9+S+WSzcquC^P&6;*BM}1LUqENH8>VDKnru))Caek8R5~xV1nT`a&8!wt`D -xDdcTgcBNB+ojbfIh&4{Y5dfluARu%Hzu)cAK&zciQWBbqY9@GeVyx6zi(S3~!|LnA@(fb*X$=P8gw! -Z*T82dks1_Pr$;3)_PC^Pzyw9#Y>Oh)-@LkzUw+RrsXyR_%ZvYUSHuBJ0pnW3JDo -ff?QI@d}h_p>QE3tgmTz>s+IcA< -%!GIZ0~UbD^Qgk4uHsi;wv;ck0z)0 -wa_F)T?twPwWriw(QqEHHl7o>uvnR@Fu{OWJpCN#AdJ+77gFM;Z>=Vpbrw{Lc<*m%;X^I4!Cy+RSiu8U -IogP+HlnclK+~Lw`j~uE}UePAAU;6v_8FTXUp(5(zdL1e};dUY)iq4=nQ{d_CZ+&)GF~MnqS5(H(!5y -J2Q30pHvfN|Zm1UG8^aJV#yxrqm4KDTshXM(p@0E1#?rXdyq+j{9GZ#uNVb==#*-)%*YWAX|D>059_j -n8W!s{FML~CGN5H)Hyj0t(I9s4LjnA=_7$0$9DKlO_q?_M>!zKxPsYF}F&?bG!w) -FQ5rm-Nl{ZD)vjE275@YvD}Pj&NYOuW;8%kK{I?fTPs5OA!&M#NDeM{DqBhyLo?=UR}9=AkkB}%)re? -p@004kpp?X&kIKl8^A#7^;f_I=O@r|UxR|=yC^)sx|lZLWd%x~;nKV;qX%6P9ZB1==V*G^4ZZ%9Oe-t -v2il3)c()x0>0J4=P(U@}V^tB(n-Q+-K*2vyO9KQH000080LuVbTcN@$e7zO`0B&gj02u%P0B~t=FJE -?LZe(wAFLG&PXfJbgd2D4aaCz-LYi}DzlHc_!I$R(jeKb#^EytF_MZAu8*Ep>cNb(&HjzEvdrZh1*!< -`x0=IM05{p!*0>3PV(?FGRS1~lDWQ&rtv-Sz0|ak|;&MJ398d9W(-jaconMV064Ql#9nE_U-pvbd4lo -^NH5RC%GBld3ANWmU<7+qdcB!&>Ur#TUcY&9I=}eqtG~ -U19w_Jw1U@*J-z4SDoCj3QoY5T~2>1==)Q7jW&jC{EN+tQTH^5EcLu};cT% ->R^DC(Qp~@B)yRiz3~k5LLN=?)@w9D-uQ8@GY*=Y634D(3|slxz4L{J_lOw2GE4~u>n|xM&x!rAC_{x -8sm%eEI&bhN8;fh#hW~nOdo#;GXW4lbR@p}PC>iKfF*+~zxw4w_lQa$rSgOX@15=_5z#{V?h01Vr6rf -H@U2NOC*o_vhXuu7A*)@H@heeq@zqQm9t7anLm2Mwtt|4$6NApsYuV(no^YqbYO#Q6hlN(cw8OOUsu{ -kgzwbw5sJcoPAFhjh2lAp@l{8X2R3{?aZLUCu5ryf-Sd_cP4G~RBah0!^W06+LI$a1^EhYjMfdegWl8 -;h60;Uouod{S@dsyU|kSRPC;tJ$4BV@z(?=Om-e2>)g)e6ghPS&NWC`6XS4hJ#@^*B;ct-D@amv;PBU -s#mY{s0nv2{_c`j!t-gz;2mAtK{cuB7Tso1ondxWg3Z03|cib_eX4pxfMMOqZp#pdFWncxJJ3Pi^FBQ -X~uw!T?o`9HsVn2U1*@EvVfgRHe)g%-MzZPJ+BedlG{<^^g=Xk=4l0rr$4qrew3hYS{Tz1U6^2JDNuX_6sK78_UttFmfq5pa-l1;az-6>D9K -fqR;JJ)wa?^cd**G=E!g1!JPgDUA#^%sKZ$AfZy@t>!!<}WTT&R+lN?ZvBeuiFFAJcASdAFE_#ge~gHZHrf%T?(6R!G&mSA;my^gp$a5#WcC2HBdSL8}*~x+EA}Hd~DT^Zfiq5Z|> -Nrp4_&0wcX*petTWm?V8(Yv&LqAa9Z^4i;d>lY1ca{wwq_SJ<@O&aTPp&z`p;$K8D+7&3wBAamDS}zE -a<7`@975c@p>!H6Rf9P8gQkPB1egoscA!{h+3uU`kv&;go=O!m&{H?`=1SB#WI;S|~dKwGeg!2BVBpQ -iBN;Dh5W_+kTPl1DP5HzAh=ZIQ{@IRC*0BS)s%QwMjlmR^CG;{56>;2bs!IXII5#Rw6BRFoIF3Xv|1R -aOU`Vvc*^pH;)bi{uvQ-iTD)ak}C*FjondzLwjo$i>_(JqX0rofyX}s=XW#o`tO%$Eup&xgwSL41;C6 -sxML0gdE3)EnDV$n$j5a^>vV;{PwT+!Q-y#>PZ}VnHPA;9UbVjy!aj*`t9_mt3)2X<%B!gj*Kvej$7c -!QkB=fuvo}&><8T^b+V~tbP7WVO*fwrOZT1eI1)N)%3#HGKJYUdia)`fr4gN?Tsbd&NIjr-FKNLyoO$ -&^YN9sijm=#76z!pfBM3%lTu&ENg7C0?~EO#w{S^$;MwLmKAQxeyw(^D;OEx;^yTGmQ{NI2sFJN^J5%imfbGd9g_{4$^CkIw1 -Y_^mI<%T?=n`UZPZ}wf@%SxMq2Cz(N~25i8pRU?Zrjzao>LRJ{9BTHP>BJ_E^eCkj2WGI~aRD}5VA#d -ZT$MknHh7{UNa4&aW2GH3vnWOhYiX|8Px3*pZMkfTIwl=zE@xD!Z2a-z{`5eXyTqdd;`H*Z~) -?}NTLVL0SOdDMEH*B=_n0l%a3C*m-8i%c|YR`!=W_ONofV>SUY;xChxOQv9hdV`xM=c_P=vqI>$BOzTa)ovCYsJA{68If$Ceov9 -jvp%wFcKQboZlDiqu*^=drCWxt)(MPWD8*?;stSoY!u;xUve~YsFU+NqF(NV{cGJa&8L8$xn^Va#Tl@ -pW?$zL(5xmGuiB^bN8y>$dayuSkh~sf0{!Q*bFC$HOCM~zmDy6I=_s -mP=bLYM!*rk40J|=@g;V3G1^JK?WR6DdH1I6Okia!%DP -U7KfoTT;bsl4A=#&vdMgeqFdelM59^e>acJ*6&7Odt1dozfWcZ3gZGsv7h?%<3MkL*m+(sQPgn!5+!x -DT0WKxY~fn)aadq=S+^y)vZqE@^3LkZxk%b`aBtUWS<7d@V5z(oO7X2QhV8W@xD`-I7uO&9bA#=CGga -JnJjf(}-A4V{+YRfirCC1G?=ydHr=9JKoG8)iQGeA8EcI`*zbkPtrs?`!SKn+ncbEcp -pymgw2o&!J0V9Xj2I>!{9CP_Lf1x-&aE0rH0($o5z2J&?eDmn1`k{eCcorqHniGvgnl&)HVk -sy!U)m0rUo9xxwZzDJ!3=Lz%ZUGOeUlQstQcd(hm@+a8)@LZ6*(=AnJ&p!w8)tq1`k`yQx%Zr-X7|QX -2A!o0HsNjVYt(#hdfjWPg=pfK@lZyqoHYzq-rf5)(t0n9fS%wl3!Z*@8hUFbfX0(<*^sr^h9R(Hwp)9 -wSR+IYy1K^1A3}g#lp9{PL1&T)BXn2CL5i)-!NB3^>PfjO{ZXzIM~iGWqWCGP-Gb$fkXf9rZx=xG%D4 -4`fgKB74*WS*KlvKNNk;lm>^Sz-0B|<8h;oj1d{?yzrKH;$47udUc{h!ACpsY%GqN?AQP6Uff4(;WCt -;t59<-^H14zku25Oz1nU@6HK;uw!hJ@(QXduXqSD>){~>g!L!l*8D*PC0qfEb(|0OfIe4yg9CV}}S~H -Lh@wM;GRh{7``CO*|C1-d!wAkA>Ny6dqN=c~MwQAN4Q;MwIQqEELAfLqHgtJ!V`?nLl0#m4X(jFqPT2;dY;_z?$IWw9?4$WAfDuPfrJ@inM{MQih -1}|!M-#Fk&iByI&ve140U@er9HH%SG+F1LWHD?z)Wb_w^Ho}u70=`Y3XqjfK-FB$;kdskdNceJpb>-y -L?e4hF*;_&9R=KoOaM+13fHJ=Z(EzHs|0DKAW6Zn`gg3O(E(%r^A==BE@3YdYl(^V$d%9DVB{Kd=FW? -2&GF?d%f&vG>m?>X@{K9w1JYrc7w1-R_HGq>!PjlDlwN8BEaM;_0O`zAH_j2K>ZFZ&RcAJfG^8ahrRolc_c+cf!Nlek~-=c@l=_AQiyd -*YO&iHBM~$h)$kCtFqLW;+o_0eA?c?j@lCYGgP}Xcz=d;0zd;xxbh)3J0*kW-+n67HQ)Sew -0P#mr{m(s%_k1bwwH7SkL$vcT$oI|j%+Q5eild(G>^|Zc -$>#gVYY$K^4{sCjIqIo_!OcRSQq(pLB+lX6a>Z{_8J -^~S#h0nz;QE<}I+~0`PJUofiKdV6y|AD_vtlb%gDf;p$wjU^&)s`W~4jYEt4bR38I)3#y8?gz$q@u(K -y8X=<)(#G~o`Z&AZtivY&OBZkJ=QYTeEaqPYdyU^JFTrI=V>hHn@^6i^FH -sS)m}?R-OKZH)v;n1#9{e!QZw}z#v}eoDXa3!=Ik|dW(q+1%+$9Pv?=VT+}kbIb?9o2h`#FUO{70JHl -Kei;z;nuvxmTjG>jur7f2eNNewi98I(O`oJepFB|B(@pzWTMrp*hGa!t{^Tf@#@S4sN&VZ)FkB`Oog3 -0L)pljA1xZZG0tDMIRBsOavWIsse_l>K+>V42*=&v%0ccnqRETms^PT=0iI-uLg*_UVKUXhga(uARX( -FltA+CyyqYe-cADSb>}J@RLlhZ>p0pjndCjYnDQC+-vW -LEs_YOcH6;ooznx`P}_OeBHvsA2FFc>txbOw(Ekiu^mYKht`95miWdkIaDit;-TrDRhK3el$2%?O+*0 -M5x!#%e_L@zHBMCI;iUHR#ifFCfX6QhV*{X7CMy-Q;lcefa#sfXJY83m9D1Z4eFH?1yL0?8tmk+>`2Z -w$wYYu&Vz#sh1Mx3R0yT4!DFOs#@-9tY(!7;;h3A2zca7vsDjnTgULt}_eE<*qii6l1U@&zw?hRM9+D -xrU#5c2rDi8ZRAVeF8qcQDe1xbJB!PoG*_d9jvBK}cZIm&9}VWlMtP+T8uGodW3ic)ykN#_SCF)i2;A -{&Q{oR@TOSuj6)KTI?NhuZuT|FyB&}n?uHNOI)*$RFfPhVwtX1lFt0>oi{|>swJIo1v=LWQJ4cHOxj4Bvd;9 -9;kEcJLIUmoG*U4f(|CDCS{8P!_pwvB7-@}SuB}C@wh^Uq{-^WnJ?$wu|Imd$K#yGqcI2icXdHuxu=m -11M5v5}&Lmn5tX~iD|sX{pm`eLZoV?fNu<8p^*t210gvZ|IL -#4N-UMp&vRgE(-`CR4YmP$qE6l5*ig&s9zL6HW^7B7Y6+aUH9ez=*hVSdRbzt6#j%F`Q0*JSxb`WYu( -Y4!l6qsH-1Jgd_9hYbehXg(wp2czRfM>Y_)E}yIn#Y%LXoBs9}{>D)>`x%4h{~i>uS!@M&^I+pz?W -@EYv+u$^$I1Y;IAMP5(#7oZwpin+}zpsCOPW=IE;=ZT%w~e4@ocK_ufL@o0rt0z3;)jo)hU9e0Adbn6 -iMs<`{(pfMMBSD?#gakJYLdU|L{@_w3dp~|;5u@8teC!SLCyN5?kkG4DYT{KIJfa#on{I|j%wIG|E^Pozy!TcFyr3DKb{< -yli#J|p&9h8&`)qXv?n6$i$>nDc_+2_Yw~hZz{4F0 -0(pti&o?U?6ve!jQOywM_08+ruJ9t!*Z7m+@e2*_b4r-#%MT4$#T7{%D>uCBY8hV0`zc50e5%6Uj9lY<>Rs5bZ|hy -Mp7l{g;`6=B+qhUNZq%V6MYb#|rIL+KA~X#UN=GInrKAS<0ot&BWHm_T1b=X|ZLt=*q8~1lGL^41ZnH -MR{T#l8RhmgYMDUY0uxEJRgniBzPfo6U7!)Z=AWFjfaC7e*(LPJQaEk0dP)h>@6aWAK2ms3fSzF>n+0 -*_Y000w?000^Q003}la4%nWWo~3|axZdeV`wjPd2V!JcrI{x?LF;w+{ShP`4lMhqq`v%lANTc|9ai}<)RbLY;T`!#oF*KM-4^|gdTfQd>-L?x?Zt=wwk@_tYpl1pTB-_9zj%_sbJw@Uvfqg -QrdcrzR#}x-3hp(wKp*=3wydw!``J+h*J-^^=;mp;?D_6_FWTa&68!X{?Bx4b#YU{ocU$%4n=SUHsQB -qj{AnlZrBF|@kO62`m;5q00CmG~f9h3gA^I0W1&=xfpE^d`UB=Q -UuRMAtjfb@Ju9f&a=%;FpOyyMA8@Rpl8y6kYUMex?_PN?RDw8->wC$-a3fYI~5LG#mWJ#Cz7ZP0P1mi -CVPfk{|vc_P0&D8pcZA?KW4URU-x?+pE%NeGBX$#|9ha+wX45UUb_6hjKoF=dLd+h31p0=(<-;|7EkQ -SAkIlByZAXQ};z#8w2$Mp1Kj}ZLwMz-1?dy6Ls^n1S~q9O;kaeS5323;CK&%+;+YEx)fEV+rC4UDucE{Y#pUN%RJ45$zk%?J9-I_q*-BA6WbM{P? -o`P%3;jqUb<&D2SK*h*XyVouDz$%*l_5y6F^{$H58z!I>ywP`oIRbGSb`B33|%DC61t7DP-ErD=L`bo -AZxS5LqB?hHPGh)TCr(Zg)oL|GQaw`Kk7&*Gz}&%QkU_QiRA{`Iq$&j2poG7-q={zn#t9>7qUmJ7w`LeuyQ -MoWTlYRc^BLb|>tw>FZW(+h*xhwA>f9sz*+82jZ&Q2?rv@4op;}L_zwUU)`p-25briin1jAVpmpuSx? -kRU|;U)B~sXdyKaTB0r8&@)u_fK+TpaRKs=aaqJku9g4IuzdiG%{wgJSv4m=ghh6po46Oar|wesr!x1 -udkXF8=?v9fAMY0(!0AB=#V_8^+CcKy@<65x4p?Xp5!Uf=Y&$q_8)9k{)~S*R-`B{9KOQqur=D$xs%| -F}`38w&n;-HFyOVDW>3{aO?&15uaG3hZombz?&w93oZpHHf@zf2aXcum^h>g2K-MF8omnnWv|3{_tsK -j2`V``R-GQ8d={RB#zH!6M`pD?UekH;CB~8f%>PUkp#cHP%`VkuWYg!`xp^5*10iBXp!NL6Ni?%n2953R!=;sUJmyuzyElW4H9l8C>oiuR -rQfO7H2qgUU){NmZ0v;4_7FTQ>GDpz*o>(i%CpTGJF>iq2J=xB*EEIQi*F+PAXxe+sPgXbrNb{xlFfE -gJv448va0ahwPMm>V20B}T*|LVBS_--ZEhMR^yA~|BSPE_k8qWioj(WYhhJX(CtRAK}q5?_fPTPEY1v -A8b5v(c_2S4P;VQf*NJ;^06%-+;l5P+~=+5JeB7ZChM#ij$~rqGj_Q{NRIqxB_)>-J&nX-JwoQU<>4l -hB6Ti`A@-54wLqkjo~uxhhaN2MhX5cdgw_dt_x6lKrDlB_Ap|1^SrFfKF`%SA;bv>7}#&YfqH9s?H3n -BidRi-Py&mma`2bM@K}EnavhajNy2d{X4HBDL*LGan(^w}z5QZnQu!~cZK7P}M+~^~xmAlGm0^ -3|Ke?S4$XbBA8V_xWsuOm-lXViNM-FKXo@9SJo2Vppp~oxZd6p@`cg^xD+~)k~=bY*kpBtE!J7Ufvj| -*DS@7mhuy|)7n7sDX%V1t>(^7$}>*_53oxJ!!xgCNU=v7iAm$0MzGRW);J@R$iN#S=DiZ)18NUwH6xp -aUi3mp~btRlLs{jVyi;N5>K+nBjbbct`PrM|8zWu()DBu0Y`kRj6C4B+Q&6HKHc>x(F}hi*SLmFRF6s -72(%og$BLumeXHCMdtyTFv7;jeH1cP(2Doz<;Q6+Yx=mQv=%}W)B$_{kQbv`+@rUhdd`p)X^6N}9kP4h0=ZRLb4YcfFi0eb&S$;=0g8JUVGxo=8vN!# -Vk-hui~M70nrWQ*&D#r&q}dVH=#0XN-NEHSp|loRQ78qMN$(}3k1UllF?`ExwSqSl9bWW2Ixm8Q6o-< -nHO_+%>qS~&pyi&5KA32rL`i?=Me+ncfyT&FLtHA)&iZg1{Eg7A~TWd|l=?@WRaW&vc>b=8XEo%cm|9 -RI#OSvmTkH1L7}5iKqC}dXOR$-4uSyy256M;h -zgjlAd9-3I>h*Hl!q1VLNZkiO>|-evF@qTQ`QAggx<>+yctXtR3hmi=?|65PUb2ALfn+KQkA<20o@}Q -@dOCF4H2lTUt?5j=g4afoggsQtsJld)*DV;Roa@@XDemcyG>)RDJkB9#7 -eVAxFB&x@o%S1U{pgVN`sk#0OVIF_`g+A3F%M^qlGG2b_SMPIFJ*Us%mOU(5JTchl^uRdfaOk>w`#h% -K94CE-zb#k&%}Mkf{=srfIi2LJoS}j*>t(egIbha02cosn0#HqN*azwGq`mwAdF{ -c~|~9G{pE8sasq*!}*(`IIc$Im^w4gd3EsjAv58Kro -XE%$5~cAsC;sf_p|StTl}np+Jr4<*BHhr7P`t9_J4E4x_rHW~GE4OXl1CGb%?=oUDG(+==H9ikauU`s -LnYl^0X7I#pLF9E*jAva&x${LXT9E$=b01j7kxio{l>YNSYoQdHCkA^a+FDK;{!fVh=Xk~q4{Hrh~a(lvG{#P2x+yF#@=P&S(pV -%WKP$Qe%~l+l>JGh~dVb(=IVCyVn*0R4zBPNp^VSlu&kj!EWHxi>wip=1QKy9pgGlN5HAXnp -No(={b0gBg+#|&dGNx%;_kXDMB9(fyJPkn0B}8*NcCSJ&~;EwYC1%g}dhnM>myxLoDJ{ONR

    Tk!R7n@?)#Ws$g)ThTh4~OtSqg!v!;!o(>Ys6jmo@L3;=FopC=NqqI-T^0 -$|7lJ*UiWpxRln?7{t8YxUZ2=8$LnWx&UH&~$eDA*vI;RH_c@_e3?FTZJ+k|@#rQm}cyV-|_Bfad!j7 -$GB;sc%JLhRDidBx=3T8^6XX?hVz|a>4R%NRsAlNgscv55A4%MDedl~*?)Fs_kC@}LTOybv__!2#1(H -#(@8UIN}MPOz=1i(n(WW#MTBJ8r^HklG3ne&{C$cSvXO@b3TQ#VO?vS#`&nVg57wV#a7pUiqnLbD>1e -}|7#ma@qp6e1c-V;PqS*-~Wqz_T49!m5bhi>fR-CF`{&ugsMIy0OEyXrP`g5|YR8eREaTv`Z#tp2>tHJt)VM1+;tf -4oQG486>{xa9nOXw5Xph4neN6Jx0;W~N@}}vaGkTCU${^KPQm1z?eoyQD85y#;>uD=(V3)Ea>eUeMn; -qcL5#YNT^vZx5SUSEnFrfnNjkJFdMu@hS{h_xmav}PMula4xdNaNr%BMK1Dj!Y`Vj -<|MH^vfH$!eKS@(4I%83UWPLGX2$5x1lJ*5xrxT<7nz5k{7n93}w|R{WP|v)KF9(ldZXhuXjfk=D$Zwd%X*^G><`x95U!O%EcNgwV*Z4{35}=8Co00?<-N@i -6F0f$VKC+WW!fA?u@)Z4dZIJ=ah@*44cZy47kC17)c%U1(P?;&1U_NCv6y#IG|E1asvF@CHK!dud11>>7=+V7FI#CZFeJnooWCct9nmss!McA -B@^cl+H#$J(7}ZkjdPAd>ouKLgiW*7(+I70mQsyUF2IQ13OV`irZ?H0nT& -vGE0W?zem>zLAxa%ZQEc|Xw0O<5Eg94ZOM^jOYBB0)(P6TD&qptu*k>?!OI6zT<3%Y|QQAX4yKT{;e` -*AfIPJd@Yr|MHVh=srTi9c}8Hr6M|GJD0EoMXGBOdyMjHnxS(rHFo%YG*3tGisUZjIUuj!qhSsO&jj| -D*?6zNCd$L$dhB8A=R4+ZQR87Tc|;S2Np4_D!yGZO%VWmOhAHHZ3w1xQUb+wbFi8Q&;;)F501!-n6bm -9jNV`6yjEhRfp;6>s<@LEoNA?^4y4mT$5H%o@|8y17Mk&^dWAih8Moch#jEdPQv#XGqx6wy&|O5@jDF -#^k7EvLmkW`#Sb3c!z*`T1D?`D!Te{7sAH;tCn`8*Zpm!(LWSL`00K~@6@S{5E$i@2@lHf9_pkTS4Gi -HvT12;kT~$5eD#tr2dbTjO4G0*&lX#I8qKzMEP-1=zo7URu?3h!+{lsfBS=}2Af=i-W+;Y6lP9O)*37 -7@Pdlvt(j>BG#y7r~0R#qQ|7QpCMO@zmo`v$-{aZzxT$-^JkDtQCr_=rDg>i0N$!aU16nX~UiM&6b9iw)juB72f2aF6OwI49b$5N0v-EG<5Fh}2EiU8@3v*_{ -2$k2XrkhfC$8RG&C-EXEd`NlD_yXz`4({=i#$zas3<$5941F_)enSMXr5?(y!zPB(aOCIs?y{xBI>2o`m=Qm3jc>-EkDNgQdBs{$2QprgqGx7z!^5{uSfNCoEE-7Su -e{->wCoiaJRNNxh?KK7Bl_F+vMWZ8rn2wl^)R#11nTGJGptf0=x{D -!Xk}>~TjCa7X`-om{#WtHA|$VJ?WHrS*70vG_y^cXPXtJV5LbLrgkKG8aQrSBB>KwOVftBqMm~0=l3< -0AAcv8T+)Q@bay({+?J^UtGGmFRQUu@Y5%0!(;wZhiF7NhTFE%;Q0*DX}2vK*HgJw9%NYvml3~q#%f~x5`Nj3vjue^b(`|XaXdR4;z$IwT6T2odX(P -;CTM=5dv|7Jv5Jf5QFkdNR2Nw@G__P604*S^F&Hb}SH;Kd{DP$TP -n%_trc*W<Hh6nAAM^+T7$N7Z@{rSoLfWG3+A -FmQEvci#Xt#%uKd+vMtYoG#qpt7nX#(2z% -?wCRF6rnH5&LFXjoUBysfhFlJmXi|XM`)0Th6g}TaPbo*GI;9}XBjWS|~45Dj`Im6AC1|nBd4DYX+zH -c`CeoCedcpLk`bfM#+USlkb-FCJt@RjdOj!z2HTozhPiMU1dIGAna&1?^d1cTSwgf+1U^gS3#v*o#!9 -H$$000ZL5A{z?5M~U&yh=O6PHE-&?t`fC2CeX=R87I;4YkE1;O&jND@Zzeysd8plmbe=;{UmdLA1LRU -x+|yl>#>DLS?SQdkuW5P;pRQefZcjseh922J&4Xq%*TzEnB)b%8?F=^`(`M7Xch?XyreJFj)A(44_b) -wbm&mgDVxivd^9wIS2)CL@32r~pr%04Ez2^;<=ilxan%mVjbZ)Vs%mC0pDPW3|N7`t9k|xCu9{oS5Cv -C%8+-WP*{Q;d6$Pb^DKCkKHynCL6tvFERR4F|Rnd!?o6~>j~U~Uk>02MzX+JoA -LfTl@-AW6#?IYupCtNVW$T#)ZXE2E4?K`ON0Ov;50H@`3V1Z$5W+n -q!_Mlh0o5Bamjj#%*M7$Y$6IYhhp)k>cA3$yjk<#`i>U~|8C2Lnt5bolYmQbJ6Us!=h4^i -vOz+Q7`I{JYNGggq^#*Kuv#a{>Gn@;^Y7joRh=mMUib??A18qRzlmc-I=?@Z+!RiQNWX7NH9-Eyd&y+ -oDS3h^yX1jd$N0@BgkcHk{vL&{*YuIj8bc>mS@SZG)ePZ?Xt=9!6KIS``{=IU6Jw(PfC ->Eb(3h%T0kjEMqoBfywg06O_4J9S66IRhdO4&PN9>8go -yQ&PGy=-T(Ox_}W^o*t$p%NtbjGk|bw}W>YSusEIgDnok{}dbYEq@e3I~A$pP+}X&GR@ymku^eLOElv -UA*a1_Zew*-FL**70&OtI10|3@h}x$Ce_ImM-SB$T*0Pd_}{O>BKsIWCZ0UHqftkPmnIfg6r9t)7B{o -lWF$ifmq8JDJgbMgE7#)jFd>;)S2y#tMV^|C1Pwl$a|cH&F4vAy(G^dp7_o7xsc!0r#9IB}-=exp&ZA -u!cVsLiK9L2Hy)tRJ+kwQ*)S={35gAUAnaT;Qv;RGqUU5 -vH`(E8JXj3Y`)QI&Ao3+1vF_LhP{gzY-Dk(M`*sbev$4br^Mp>x6Y2@AcE1T?+r3>%+O2Ttq_8e^4(d -M86*bpYD!!U}5WNtXGF7ZrQL{ti?h0jKL3^@_J`%<9ChEY@RvAO6mF+$}N}L*BrC+3y -erQETl-^!;(PjsAt?5S=F|VtZFdNYqMwUT`P(>HT{}rR38vpS>W4pay`HGe{`LcyZIW>NIC2}x9A?vm -x#l2>fpv+c3$Nkn-aouI`_2suO&Y!<{{_2@MbDbY7Ku?mwi`e4u>B2TIZO$dn{|>ooY0gF -6284K9TtUO=gxr3fKIHZ26vtyc3;Qi-m+fC>7BY4zxEA3w%i!+(1S+4zF8cB+yF({SXDFJ7pO3Z9BUz -U%{aVvTBaK(en^TOE4u#wD29^s>7aKpy9)WdoF9;S1%>r~KO9R>qp@Rnr<@2(s?wHkl<6lOa9-f$mv4 -gOnHx5d7rD0PHxIF#qjTof}&yj`CP^P4ACM(2f_j=c)P -uaCyw7v-~mFhPXWTi<+@v25)=3P5VVEK5Stg9UbXa(YwecR~DVe$M> -MnhtWFhCpK+g$z#lK+gF?)14`|yYq6!X>gLZGpbaFa9j1CWTb%0McFj0)eHaewzN0gAqui++dv}26le -xTHG2heKh7I4EOHhM6!mQ}IYLKU4nFTHDId0ap%^a&zH>NI6!SX{v)NO^~p2~*G9VU=Rg@NiXvHS;n~ -8}k<>RPb8%`g5xZ3|)oVh<+E%%u-;S7`~jby&G+{O9S#UNG$?T*qQY%q?u(zGiV2sjm}%6JYaGG#U?; -5bgd8Yyq!yvI39W@(0?3rZxFe+4q-PktSnKr3C{zmPT!uw@7O8_A_;$vhIXUpIsIyRo^QYc&-3^s_)~ -^7to-ZfMSHz1+D@47x_zh4d4(?+YE@7vNQ@o&nXl=~AktUzw?lwwviOblV;1&tYpmVgl*=2L`eihff& --8jvW21U7|UX`^3zDt{nrBQRdSU??1vEYnP^Gw*T& -tm$PyWff>Q={v_q@(0(t03ioXU>t@)B~zco+QGjanuX-+wa|dx=`9(;3bO%FIjk}n3wSica3GGdDI@`>%vVj#H~6O69mr|WY0_ -j^PVK*xYzl85N{pC!5H(XsuvJG2}9`PFr%uG$9Ue0)SZoS*Q|RN?E}bRpRkzQr%AMhJYW{e4TpJdz^b -?p?mZ@rFBY^a8YdE!;fb5qKV_3BVP^eC?A;~i;>sq~5Lg}&Vea;lJ96eXpiNo~%(!Qh-iI_FoN105D- -X6WXT3g<%A3}F){F7!PPlA&m}J5kJm~qW@pz<0Zvf&JWp?I0Mrok~z9ifZh>dYF*uis;6zBGDnW^J+! -F2eCxpe}HMJYhIqijR!0{y{Gbn1+EK3KZkra9QiE3nTo*)~;bUYZgz@AF6+L;NjMuV<_|m2%6l`qa>- -GNqT_o6FFcFpKX<*`vX;FyeD68E=oj{|X~QXSbTaO2BqWrhcx=ivwgMN31d)1J6wypURNNo3Z0axohD -wN%;_0f(V@C>YZ3hfcH7JlnUJ8>{3Z;=KxdUFXt539|+Wb`=h&^%#Kw#qmc%FB>*oy+Bsre{{c`-0|XQR000O8%K%whJO8{DwI2Wgahw1E82|tPaA|NaUv_0~WN&gWa%p2|FL -YsIY-KKRdF?%GliS9T-}NhSbgm3a5NlVClXx#kMUj-)x!6ugI;T>5TUaa!?!p2I1^~6%OC`Vk`Z1552 -SD!HPUVA8r4@F(*7>FMt28Lqo_ldQ_2TsCFjSA9}%c5OG7xXmw}1WftK!9%uU`E3SKs_0qgOBM<-ni58mg|mYAXKrr@B|~zbQA>>g|4~OTOI=b-OJa{`RK&+rHW -^Rha<$pDjRTkLHhxTl`He~z)KWgY*y?m#k@pEz2wv8_AN?= -Zx(wAZdE$Qm3{ZM7-s~zRS=u1?D -t)zA-?9Xn`)~$-<7MC=#oG4*GzwSSp#ejLep~Db?wkD+eTNYd*FP#;$BiPeRYi#G{_tNstbp^#_;n7z -hwO1+q<1cFjOB*?YHnbqrU|3s9w7x=8QU)>n(xL)?(-Jh0hBv$ov;PY1 -TmSuQ({yCcfB(Ab_l+8`FytQn*36Da^cT2j_M0uYU^bh*tk(4w=2Dp~sUW$nhnso}-~ag*X)GQQgl{8{`Cb&w$U?bdV`j_aY^LUj-T!tU;peGd`^Y^$!kC -7^0`4Pz(mt13C`s`c4rx!Zx{>kpudK3vYD{2}oM$DGR{lKAq~^Y8vzz~7ga*S=g=1u<--gnG^u)62T2 -?ngM$S;GyCvQ@kN6#mvdtO<8Xiz9ryYH$0bsjkcAT>_9Gva+F(pFHkB!&cp6;1=5MviP>!3n6MPmIi9 -eVNI1%>z7s2fVg1Hg9fgzZ(zt>wF8V9(C*cGn7icClPo!%i@DG?O}Xo2V?&LF>C*BWQK^bOhH`z&IU81T@=%zM5I=z*Rw&fH$-T&)K2_M4d}PD=27 -HKVTc7CX)k|0E2hab~swgEoh-i@lp55b^RXN4$ROWRe?uw*}y2-!swmS<~%Hmql7@O^x)u%UgKFHTqK -0R1NH$#xMGwD7tQ3l)IF>vTyaWBWQsfhGu1o7LU{mcRr4^{-?M=T>nx%b=#=G{Y{*39F&*+dTY@F*$X -WAzAH-2^L)160tTni_MbyccrVQ5$v@nJW4U81#25>MRRViSnvKh+5@+_ITM(`}j0Y-zmWoN=*HnU!9d -h`w1SSyhOP+4 -f`nkefc}J$pob~BXv#ma$di<7W@a?PC$g?)WPi@q`=*&$S-)4VE$|TWa)}c#(4vg!s@R9*Evq36EMOI -&yQx>Jrc#92wiBdV<5rc&$w>W81d4?aJL*f{GY8B8941!gt`DbYRcH_(iy#GU(&8%$309B_X@KwA75^pQ87Cw}ZK3K -tU~<1{@?%fZJAmI0JWd0OThc@mKiP_M(6%r{s(Rc;~(n#~z|!p_XZ}2A{F*?sBx<=8OS~0PF4!{M%+d -PoDlJnQ`=FWI?h7u7(xG&A>jpcImY3-sXo-ehGmkvTe9=wrow2Y5{??P@JUyV35q -DIKxR)o)1-<7wO)fxwF3g?E=FM}40K&~?EI3?;xA}sVwyF~pSoO3~j3t6chHBp{3ecynt>-rYpao0mp*pvqHSOfd{ED7!B%$GSi27 -uqop=ZkRssQ{%sT{gO{*+0t(;n5S_PQZa(@=)wg4!bDcPu*;(C->U}czcT$nP?Vi>C|*6We_Vi}%9GQ -omfe8w=Sb!;rcmG{Qo<$Y;fx1dRXHoq;6?VvCYWKJbjyk;$j -j5Qtj;N0*z~d^7#0FKC62^;=uQNx_ETS3LW -AfUy0V3svC)X+t>Isv;O2u1l8(>eC}i21L&4#VvBoz}&S8aUW;qf;$FX?%H<|aK|*u9Ze7!rUp)Yiw; -=>=zNc1oS3Q(K%C4(Z4bl~sxet|JO6{3Yt*<-pLNw0x&x{3B0#(tDA!mxW>TnVa_;Dj1s4kK-Dx+pHc ->sRa_|9Ts_u2nE@cohXXY-PX#k8fMO;DW{NLOGuRzs>2$T!KyBajgx2fcR9@KDJn`%lNlBXc##>Q3Ae -!K4Km9^cLO@Ck0ZPIu-Il^vnje@{z@l1koyM^~&CDkdJsgcXadH22VJ|llB0KkR%$9zR7ciJ{$s(T`!o -ZX>@32&Reh=a@U#@!zow)XOrV0_g}c+!V<{6scMROW{nnOC&vo)Xvy`5EYjII|7VLvcSspppG>-Txh%Z#Js^cgf@ARHP;k$Ck5Ajk|)n%)kk21 -KUd*LGh)1*I`=3*2G-V$%qXM!xjRFw1MktD2&~rI$dp3o?#XNeVS#7&1oQpY%0|hTP8T_4h>}w5{|`o --!$nw5U?=0+}TQ%pX8d!1vbDkFY2AM!SbI@(^c7A*kfM8JeasdNQAU --VJ_4Or69Bw?IK=ldizct;F(#|L>zIlJe%Ys5P-t5FzwC4DT&z_$XN$0WbEu6;@+)Whp(C+__e<{boOz&UL*4@3n3E#!3av+H -;41S1Ub|=Xs{$tuDJ13|5&hTgO>3U^a50HlvNWJz0bB1?LOB(~Ts=d!!>9I8UA=v3{M)9=;D*l}h(Ov -iQKzP2CxQqd)YZV#9tO>qTH%j&w(kncKdx`m%{&(QkIC8}bR%jHocS{3sL4Enxl)vk6n-zKxb{dny`u*dg{4tw!r8`(Ut6L3uQ*sXC -7(1U5!7WByUXItUOq~)J#Z{|RbuhjI2-&e~aChUhaooha*=?|c$Q^PTU^B|gLo7+93E-60O37a-g%iHND1?1v2> -7Hps2RZ1!3lc*!$^n6{nmTjgXH5X4%YoKvN$M)Kd2r$a^xm>#O6#^+;r~YId1UPSh!Y!R^t`P48WSzO -a|mwQr0q5x*J^F5|AF2+Do^sqIg&BrtW44UQBGmFeh#II2Xg+sN-Ti4K2h5q`8xJ_B0@R5 -Uen(seiBusAYJG{NuT=vdY~c7KC4;KmOXu{3O^rv97ML?@f7(D@8ioaLYnPnU5ALotCTrv#4Uq^@l8~HFEtS3nzNco<@8o=H^otpE -9}ml!aqEr@vWj#O3OL87(ye3Ae>k?Vmbkch}yICi_^cK_*=UXHXtwzK))vIBc0=EDYmwmb>X~9s!u7H -255g2saeU<&8K+fdn=%RA@s$+D>d3pt{lLXmF>{?Hn32o<1v80w62t&_WUPvPdi#nl$2NU7(`5>wB -w?5T$lWoR5mJQu@bT5r~6lSfw~V+tRodHJ}Lo-1N_2sI24g)=EA1!a6@*!^r6w*^M-jo(Q -YeGyG*K;QIWUHle_GMyY9$yVd8a^Y!txYWyC}m;8Nz87Dbi6vejOR7LTR|1ul)sF-OFIHJ@+V#+LB -%>ts{!b)31suy3w0xHapOlsO(e4cv?uX=5RgCncmwz@TnITcztDc9)ziw&ar!qJmr|Ht3?F?%ylPcIg -7W#IgcVDm)9z<;#OlbO7@OavQ&Wi$|-TtJ6&KBKB`dP`^jyY&VyM;Aq;|6i~nV@;YPcT{+H#DE;Nd-4 -N!PV6TNbT}J2OwNzji-0Y5RiWq#IWh+>n{3Yt8slmg({<>zIeN>CI_tb5UT&dB=yD0=F+|_$R -o}sqogWl)zY3CnnDzGTHfC#5uLP9aG-P2fvy#F^n(=$nNs0e;pU`i3|$CFJ`|QaC9xUdu@j|0R!9LmxbCJoQ&nHv%Sw-HSvR14 -srxT*Qn-VyA*UbbuZdo~C-V<=qmD}rxl^)x`4z{Armb16bJifzhmY#DaY<6>@{wdqN^d7ND{wxN}z)+ -I%mb#$i-V39i5GqwF42BTr&^mQP1h9Len=aaoco0b6GZEOy<@PRRd=YzgU4N)ng%woP06kGO&~emcSR -K1^ja#oq%$ex4|EqVY6uB;oWCdS@Oal*vP6!xdT3P7I%)uVVML*~^4aVM6NL@E!e7wmgl+%&TsimDF4 -IH#tWG6-ZH%P<|m@bYlgi --#!6f0{m0IAxGGT&Y~`%ek1m3Lqky5xH{`}EHjAThpWxGJ7|?#OA#tm)2dJfLPzxAMTmAJ;_7zbxHu% -#CkbBA4U2T&b^|W@c)zw$MR?azbLT6dqeO^L^#!GDWzWv&1G_Ad3l>7+DJB?-NBRFC6 -w>(AolW;Mm;}wP!-a238&%skkV_j!=B^1>4%m<0Gnhmv!urI`|!bB@o`07$YsHCI%6YPbT8L=g6O$lR -WujbX6@IEHU>F-O)u5$Eg~ZZzz!L{uUQAF2KOz;*ZBgWp=_|{ -_7pTTXeT#N6rR)&n=^B0OLLjG_pO?sYETnI75rY6TXo?~tuD(VyG+mBO?j8}H(kAbH@eG=>C^s>GDQQ -a=|g+yn1+>H@|h-$n@Wf)`u<|r)Wf -)&E?n~S>Kc%HZ4I2kL3cxmFt{OgWmj8g3hlaQc$O?xkq4bO)lu6t+z3Qs+U37q|nF*tZ>lr;EunMH@8;cgwQ`T9e5054_kXzIZsCik+>8#Ba2X-7YrY?wPfp#0EaKTQmuWmrlUh2Kg13K(ekf -sx?B1#9G8Mbd)5x`Ld|`pjT9-upGsR=-NcM1C?I=pP(wr!|Pnnx&xE+xi@y7KSHNm2U<^e{2I}cid_{ -`Dg7{Q}k5Uy~T6;K|i+^IwO(b{%kMeAbArsy6Pky}^H*YUgjqBVGwo<~RMg&56loRCLk5e%}fymhVz= -D8!bilW&Y;$D%0H+*T0xplHdoZzUrqrpG27KAzFwLDH6yvr%Pb9rGJo(|(N(1`is)SGa*R5R05!e=fL -p&%%EBji+pOD3&?R!PG)M6=+IiSuD8WK?Y)?x4u)iyS20y&aw;!>q#NBq{41BS{ILxk%=2Y)!0ho;3b -kdM9kWc6L0kd5UC7s)&h%2o$&X3v;O%$&nBw??CbO!_f)PB%(!{U6GKHg{vDg>*o&X2f)i==o}O?oud -gh5SR>YG~0qV -}BG0-ap?O(0xZXHlnDkWy>3dGpkA6&yq}Ku;`?5S$mJbWL+%2L~p)DI49sbS_hoNsQKZj~+4>r^EG*4 -;tzlKe;~kTO9Ztof9LE7oXnuc=3^Yg#Koer0+-m<5@YMB)`YK4%P06`yJtROtFi&pk*gL+bY>sd{wmm8BKlK!}ghuqRL>QYn~Wl;~?rbAwZUh3(Nv5MuU|fh>cKmWP_rj3SCnLiuOBRcTO -`yjB=3}Qd!VOS2D@GcJm$~-)tA415y`QUyk5LqOD`RecQ8{%kZ8T%XyqVgv68G9T?`e#CxsF}~@IUI^ -T~or$B42BTFH=$mg}!OWW^<&cU5PjUr*JM)aHxvn3l9@A$N1U5TT@x31NM+*&9I7 -9Abxi9v^@AISRUg_mtou_p4F~B%+eURC+BmBG(AEV>Z6h~t;iLRjhd8R4q21Z!4~P5+3awBk{AqtKiA -{6of-F6GotH-j(j2=|EkCIf{?H4{++B}?@P}T1=JRL-4i5t699ZG>jfn1JJ}AJ$;FJzzpgk3c4}*;)@ -!Bi1a9`2Vi|);(xDp6n_2OXp8A%!s5;t5pw81z#9~eD#isW)_|EQsL;m1rBpMphmHD^4KEXPOmRx{m% ->8cYkJbc`%+Kkzw#eQn$lU!Rz(jh~;NPCTB#l*9b^uV)uNwK|^dmOyb4-tqRN7j!{-&WRkG;v!$`kb+ -wWfzAv$D|c-{Is-w#Lmy^fed;bmz(D(PQ~s_LjQQ^FBgGuM9fMROn5Wsj%k -H63$=5|FEZMmr1nG(8#*Uh8R(I(fn3b`Oy}GU&Xa0%tJ3i`yqALy{Y7a@^9<~6V!x$IYK>&rB1J9$E|pLHPJ{yu5UVo-kK_inOX5s>Y2ae{iW?XB2uJC%16@f(TQWY_TkMc@}=)w{ggC5**TS58VsSJeTOSNCk*@QOQ06ylkxPEgnDfbg!TEjG=j7D{f{>iN -aqwugo#V0-zz}gE30%hx#;qMFhj~p)JkmF0iPuVc|@}PBpoMUga8yGooN0i=J2v3>EM?~>_;*XH3x*~ --v{i4|5u0t_9Lt|mrmDii{46ou_qAyRlIPs(2EsFOrjoNNOzj>m6i@VsQ@*xXDbm^8+l#|%P3R&UM<8^_KcBrB$}ag!+4afS?fbgwJ-<_(aL;G2@ -u%CX59MapsOv)GkoqiqKBHwx00}RJPua(=J;rDDx`Nr$S2(ep8~A2dgtJ)QmD`!VR$`{U>1Ykz;1UkR -faL5g-GyM#Y|3jGTIFo2%I{#5e!E`RORVQNR3p0Xs%krnw${0Aj+*+iS|-mQXP*=8U*R2}Z`!8C<3M&uycyc=4vT*F*%z -N>r=Oo@pZ)xoQ3=IccFNd)Ioh)1)ozh|cY1otP5cwu@PDjUbe)TI~P)hEwP+m1)-+ -HC;J|g$mneNe!Ur;-K;j{I+z1~;&L+DYlMVs*Rpb6fcFFfK?zS{-wJsG>UjbnSet_FoTkCt_NnDj>55 -XyNRQe%1K70_j3^huDAzjNj(E(yi$n7vXugKt)!`xF#u^%M1GPN~ajf>~sK2gaG^nwI(%9W0xs6yVh< -JfEKgNI9nb`N=F4&pnyfM4ashuU&L0j#nst4lYm5XRCUHbCP}`^dex_0_W$S0_dk%@+pFTdhu`G00y< -2+owU(R9Sv1T-sPG|Lq&;Z1X!K*%v}M9Bo}zXFsYa{vS|F0|XQR000O8%K%whIwRA+{sRC2Dh&VtA^- -pYaA|NaUv_0~WN&gWa%p2|FLY&cZE0>{Y+rO}Wo>0HaCx;>QE%He5PsLMAo65yo*|MFB^hV|qwyM}aa -^?8R=@}ZH=z?@wq!^Q*8cn5k&-Qna<)Fyh?wYj_wl~_?&$2Q@<7!0V_rOw{BSHQMQ-z2P00GF^0L_OC -nSA2?xpAL&~y4w?#n7uRSq`QM!Z8~YHW?)aSMMd*?N35ub!6~UxXxY4D=>1;JkLZ_dM@mzpZPM%Ic6q -2I;dr$eBmrb9Hq!Cprb9&`Mrv`AGhGmIaaD@F1k|P!@Fw@bFv7LQOp#I{X3U7&vV;GYUM)9m%r1$W@k&PN22i?MSQt)7yNpc*M*f80 -PxHDt<<#5ICdy*h2=B)3c)DhT3HBk>p1H;CdnXKR&C+b#JISBV?}cmn0~Czi@pR3G!nuGJYR~LD<-Vu -`Pyjtw>$~-b`SOB(6iMr+6_3#fK$^vskO|n(k9B+y=wuTsDH;MYW3a_-?~ucZAq6|N{oz#VQ*Z>-xXxaemg#1|r$X5Wxh*@TiSRv2}6WblB3&=U-U3&!dD8-|NtvSc)VFw15Jw})S`$Pnm{{V*u2vz48# ->30{V%Pa@<6&~ZC5YTc!tn^HtqZD)da=mvYXeSwi+a$aK#Vg6dS-!CSKhnj`>rd&cU#$UPn&&jXI9t -p|l1c1J96^E+u!`+V#1GH}L*~2b32={k?pk9=TUhk@6aWAK2ms3fSzA+?Q -=FeD006+b000;O003}la4%nWWo~3|axZdeV`wjQWq5QhaCz;0+mhQxlIXj>0_N<7Nr~b}esXk}iqYEg -tY$4~!s?lcjW!R3CehunK!OIqZuM%lapHW#e#3s)e#yyORb^cOvgOOfdd-=z8bBedDl0Q9D{m+_ySnM -Jrnq;})SK*Lzg=~8UA0-s4|aLem3cLV7VG^=T5a;~$~3rKt@&oXE1JBk8}p#-ip^!wbw$I^yW3s4y); -k9_cHjM{_CHgEx-KdXJ7uy^Ot{|(ygc4TmIwAysGl^s^F`y%2mgIJqL*H{x@YS_h06lVtu;bnJ2GyU0 -H8o=ydhx97nR{3)Si`TWB=BH@U}yoGq)m-kCw--^(g*JM(Z?uHIVAz~9Srz?FG;)8zKeb$MN`i)B@A1 -!c3mDlfO@@l(8bQFK?eAaAxRDu;BuSXRY$nzco@q^q|4Hv?%^Z@U}-H@5vZaMh}(O|z^FI7vMR>S=)3m8Ka7A3Pm6aQVgQ~KZJxmjPO}l-9xd*TMx&?Mu7c&Kl?8qRh -Rd^)6;Nq=@&;SI$;xfp<=a&;gVy)%{jul>t%jcfu=(>0fiI?(Rn9E_7Yi6L0nJ|T^6l6$*W`qdP6OHG -&D&y~?RPMhOMsB!uWd2y41_3xM+xwW6aDZeG5x8Dmoq -fhElxLUab+Hxb4oUa}Zsgn97ix44a4?T)_Ob@a7r*^vZs&REW&4=Anj#fUdEqk#;u^qI9>YyqA!)rSw -_|#PJD$j^6Yr(Z>`Mq?M(tfUlW2vo)s)|K3$p=1)Uz(U03P~ZKx?7=9M&4w`}{UUy_@^f=Cbt>C5Rkj -A1B$%{pK8GbbXPnuJR^db+{|QlIH6*{^NC|ze@4w6>Ve1|>bH(+5zycouwrDEV7)F`xM -jSFI#S2p>I*^Sacj|!^ioFVzjtscdXURxO#ma&QdDixzhY(&Mphha4b1Jr_=q`lRac-4{GNLd9QgReZ -8;m$ep$;V0Z9Fd)tPL+r8n2Sdz;%NPDZ1dB1oM1INEzXnE#Zwg|xhYK@CIvo;9p^WTJXy8&ak~yQ+$d -4tBWF3e?D=!M`r{RfYeOa2WZ0;5OoVsYNr3yD>VO>j#AlK}XdM{)MG0cWv1YcCfAa2r$O&-N$g*lIBJ -D$2NEpnxg845C}t}Qx3sLGK$9jr4jpv2RcI!0lvS3-&P^L%g^ah5$`0#S#7pMnE5#k^UWI^%9}R`X%8 -#==8dS^GXv($o1vj2rvVe=VeJe#LleskII)5U*!2!hnZ*U1X4XPH^kN@mtD!8HZB=*eatRE1<`{a3xE -M{{4T7m}2<7QphD`K5Sy_DxP^~Ww8uYy-4c@y0XS&PoN^UK*idb&T4pc;kF_g0&Un -%$p!!`$` -Ll5lO!G4zz`}23cL@B~e{|F$NF)EvBiF(-^ItXsYHn7?e~i+DGzkB;6%=#3CILE -R2^{8e*2M83I?e*L2oI3*0PWyv?4xm_*#aKxgA=@yJ~s)}J0H66=2%1UdIx$qbS|WK=|Fz>6Xh2~NAD -d7I5^5B8pFVt8c$_6!IqOrsD!kr63A(7Rqt-clS -G3VY-Okw61=bh&63-NbAMTP&o^*pMRQQG)>A5e2xnBEe_yP+#1g`kwsn`77ELIrR0aL^u4Oi#bN6zIO -1@eb_n5Es1jIVm-enr`6kuV}Goa^WXK{pQniN$MRzEuQ6w@H3;nz8+sd|=|MThoXJIS^xS9T);E|oir -BrPHc06^Tn%)4=Pjtko3Z%HRnKCXKYL>7~V>^;Dro6#9}vY5tga0TaQ(S6(N_|Y_b94YSlhYDPLI4~F -$8cDqf>%|ryWga23PuMUqZg_F7H+lReSr9RFuZP@dm8OLGB!%c@qP9aP7ePVW)jRxefjbSoG%L4j_@^ -C{w6|+A@rMf|W>yhkJf7mtXcj;~|ZyMi0&!!1F^B)T~g9N8=J4Bpk1Js53KVriga|673J;{ppjS4Nr -`(4k5C}T1fd~O?#0vY_!!5*RFxt=ZKySjtrDf6nj%~nm3cX-nP&oA5b(O+B}+)SJaMZ3y($|70syWPG -MjR#HWs0=RLbJ;^u0+j)qZ6x!>4oVslwNs@+PJ9->KPGGCOi14}=UdpErIA>7l>q92Ljvox*?$xeVrX -7@3nVr0MRNYL7LYrBTini_uaJbT5)8)L2)KG=Z?e{bdM+)5Xd!M1>&6-pR}}!j!?u;ldITq-05Wpe_8 -ZFxQF)jqc=Bhqwd^vJSQx2^aHClE;aRS)yNl7;@846C4~r4mr+8u{$Uv7ovkxZ$uIVf|`@TjCAE7Il-Nhdm|{HcWK%zO8zx7^v4Nf^DJR;`ub+y)QSJ}g}01T?eG2~(Z?DG;R -{s9h!L#Urx#;IVT=4F>IXia>s5)1#)VpUwOzbodjP2!}V>Q)2a}0sWG?A_9jOy;vW=!SWmTlwe{A#?r -#qWQ!3=K4l7+(|@Nv+04DGNgAp#!Xzv$emlEx3A2jdGB%*tL$GRgNY28xRQ{Y&_E%s5{QM{Zn=^^>!P -~gFV~P3nA>QYB+DM{qZvL3YOfzSRZyz{`zUG<2%+$QcTH=ctlQF8h2&(WwHLn_2P-F4n-NjRBrJPGsIuLSUye*>{2 -x0uf8A+Z81$DOVbzYg7jmQK42ix_(G8si=jyh2ALs1q>OvQxm?QuwuU>!1~#4^#=egM((Oqct2vdZJV -#F)SNXp|Ut`za=k5{UG*^Hzn44$(bbFif(;g5+KGB`u0(WoDna_dviXZj`-pB2ih)s;ChBX-RN;^9R> -NMsVc$_zvdo=O3S$ky+!E(M7cvFtL?QnrzH`xO(^ag*iocc`*{7nn|DYQVtRoU2QCNHq&1vXgc@E>mn -BblmEwyhZ!Hnp9aK9Af8FtE|%%e07Bvi-24r%3znua3I?!9U1n=1c*JslU}nhw)JhzX~&i{zwft)en~P%n$(PlWr+{R_@8L+RGVhx>mif)u&c-Fd3lQx8+f>~nn4 -hUX&p=)9Flz!c?aFovlH^f7<0FJZ4pjDz4T~dW`YBswVbmQ&t{4=zO-FrA>xeNZa5$Wqfr -e9Zo}yrG$Al;uK$Bs#^hGP-4eV@qHz? -kJ+O9G3f{Xh0a5_jtbnDBF0ITv+tUgg`jONzze --k4r&SYV&##9~`0F>!_ih^VF`pVMA3KHqobZKmLDn`OlQKX1f1mL-lQXwzti0_kk*C2bvtU?((wtys467rtOgmWrzBUk`H#^!V -)XDNMI8>1*hO4h};7!KrE1e@a@IIU*~2f}APPh=*QcA`7uJe9U>m+Ab`wQ;ZVcp$KdCMH{s+w1=RoB3 -kTgx-gkA+u}0s%Ikvj@hG3Sbx|o3Ut@ig(6~K(dAj#AO5ixE8(({s%#Zk#)NeOUyW9PZ4LLa6&KNUaW -RDWt7d@IAAP*9OToUaCQ}M|{BggiFq1zE~1Nn@7Op{mgxkx2+cZB(G*1Wfq$ptRibo`Lf`17juo497R91Ik6vSc+T$Mzsk4ks%YB&3<^sk#WlwNnisM$IJ2Fjmp!?{msHz&HlB@M-l59Ey9IUEx^>#kGE{mI4S -6^OM#rRWt=hKM^AifjnzF>RnVcOOB{FmYu`dp(~y`?~A*xWA4ccv)A=no^L5|j34-aF3!^cs&wWh-)s -eNAkB&Nlfi?zXEdES1s|ibRu`XYW54OPG<+994;TytcFb4rD_yet$9g!~3_z?T68cTR~g7CB -x!PwSzJo(T=&`H{leO2C(rYW%Kz_Q(603wa)fi4CtQ{U1{sJu{1GCFn{@O9e_hrCF8ajw4RD~irWfc@ -L7J)PnSu*Jn5GpI?zey5WAX%FUxils3VftY`pcT1K~CK;U#=O@`nYQ2o&(g3ghJztq2Sq1?`=dOeT77 -D;j^-M|s^wn3dP -B;ysMB{8#b{)ND6CS4Gn6I}W42|$ZFH#gpr>lkp3XI%W8kep7VB6UxM~8uUN9?_56R}UmDg-@DU5hex -KrdOQ(sd#OeKJocr3@mh2i|aOA5-I;nvX_lcXU{9LCt6?ZX`$45sRy$3uNaDt1XBhO30E+GvJM@{93N -_jp2yAr8`*t$}z^UoM2#m-Y-|G!Hln8bnqLXrv_9L=$qD7N(=7UgU}`}VCLBf{&!!X2K|I9jW~~Kvro -en%X&-&lvyDaGpO5%>5ohN%ymb!*|cF|8~(g3j3|q2kr^xyD!Z}m5RBdloU#aBe}3SorolJ$=;oj#c=_0MLo&Jj#Lh0A@Fg -Ub63&*?1|kQ1J|Z0h|k+X5Gns$)TQ4Xj;7m8hsH562%2@RCN9Ipa6Q%v7->G|^ci9*YThtxdSXK(Nyg -CUvyEwRc!2^34@TQ+2B=IUcP9MnuFnSD@c!`D#_{Iuu{!IllEe4&(?(16>XYr=b? -p&)%?=;o~lTII21m%~WQ&0Sjk|v6xM>m;WU@| -9WYH{9|-wo}0JP&$%7Kyd~Oea+>XKn$F<4h_7Y!w6EAUO?-_hbODAaQFbQsTcv&U(s&8kGiHa?&Op1x -%T)Cw*Ifu*&0QQ2S@9K~I#rhCe14kgAWp+Z2_7n=vJ}oCC6L+l6ojC)ql;OTq1_**_J}6Rup7r+Yy!FZjngPx|5LP+1fevoxQObCbfhu_}?X -91Ei9qRpKK~LM7V)Z~-+|F*xgEWNIvqlTtPic22~FB9sO()sENfqRKDZ`6HG$9_C5h0JP_SeEI6zXJ0 -;j{mg<3Pr#aQo!lHh^4ZqRIl+5Iws?dPiV%D?14}ULWsBRgrrb5V|f}N`|NV9*`6@Q;mcHFG6|YHF?4_^-mt}K -%;IYs?0Qen0fVf@aJZ~WSX64er4HNTjb5^%1g`F0AMT;47K6KqF=Q(423fcf$$auL>q55p@|1w03p<= -V8C+%76ub40!tMT>JP8MSQq2@{ba5O6(U+1ebL*-Vh0AZBn*v -RQ&@gemB^!%WQbqFjK&}wJg>MBTAi0-R*5Q)rHyJ_L)qPvqRJz;XWBhZYCWrvtS~%0jCZv*Lw7?PW4% -Wfs7vD1;VihgGFW{psO_xkIk*&0FXTnT(f7i-&kmATr37PQ#(_xYs?@$IOnTVl@0KLKe!2kID4Z2X|9 -Lu+$JR3+doewBc=iUBoS3A-b#g2SA{byl#dq7$#ARKn{rk2os=4-Izg%HKGE9J+3GTK%3% -Y%}c$@7<)C})~rRb{%P4lKB!B+PkJ!ma(?*j7!!))OxejZs8mgJ2WgJJrRgORhuRSGo23@(IuDxlg&G -oq;-G_lGz9#Bz%^q3|~LhcOLc=&hfZx_`lRn#stQEh7;wqeY+{NVOOjTPzI^RG2ju{MdjW8_V(MyHMF -~m8?+!?NX-Gq!t}ZPH0Eji?zWPObrwpactc*W!Dv3V^1~GG^w2|1t#jfQiK06eM@<@UYj^`u)qt-B+C -0yHDdRk8Pb?oTR#tMyxIvMAEOvY&H!gvXLs6mUsYp$)GoIB4St@YSla+wjQ<_m`(rgvzXA=PX*X0i@j -ba;P6Fe;u9-s(rq*a1_WL0Qd-iD2=eN#@SwZ*|V6VC5@&Ga{$%@!(w?|rz?^dgqcGbSs(gCwj@ -tC+>+0wUzvjluXqSNjGef+3icgA;#V(ww|vvdVDZ<#}p3VTViG3{G$C$W*+q<$6JDZZG+x59o;;6U6W -oiql;C<~oNBhTle#!+UG+aAJIJ`ljVn?+38~4B)!@$fIj*G#At9=Q2A);H7FDl`{Myx_0<;0~gQo -w^__C+j68=9yWd0kT#aRxqlj-rvVv;yn5Wu7nkV$!LyE^9dpqQlRMM_z*1Gdi4gA$%nP^7H_0PK>WW$ -%a;?H{SV!>SZE=V!|Lld=`c=TrLPdj3?KXt%D0F=0Tad@w`><{OV_TfjDiA37RTDp-SaHb3l?JT{ot& -i8O*_tG)Q;)b(0Hu*Zrxzic%-RvMST?Db*m0JQZ -|_ZT-lIV4n~mol_I*-4GvbRq#@)9QP347G*Qh&CDOgG}xG4pOjTXe`7PT{U16YO6XZ5?^o8ct}=hRY> -+;@c|gB_`25@?kquWqPn*18qrWa1wAlk3a&?k}ez;_Y&R6zMFt3fx6M -*^tfX{m(Ie|1r-*qDsIou}!0^Pp6$UCNaQclF`pg`E7{jk6QH`vBv5dO`RocpxB*4W#!qq!&>A^K#2|g?upo7?;3LOxT0bKS})A5G6~e5B=8 -%-=-<~fnteppM#hQ;(zonONK&{(AW%JVzX!Ab01)kD)dBL>F*3E>wV}8iQ>2fT3F -K^4p7?_nh1O|qi5zL^A=gQlMd&PD#)_S*VZ-7=XZLP-T+Qzkh0TXNQ`DxZ!eAE$b4js#yXt2U}j_-Vl -VP{Ld)%(3;EveY?{pVa_&V}nrb_@xjAhyr}2e{fg4EX)iG -W6Kk5_KXe1}k(ZM9u>P0fys%d@|2Q<>j_+il7j_CF^AQZQatjpPnchjkwb5c0O0QPcwet-c{cxRU?@y -8hV&L#^B!3fB)bAip6!mfVpcw(d3Z`Fs#M;;0BOj<**)hA|J~R%!(k4da=X#@cz^6QyXgfVYKLZxmPw ->p`!~S^gHIoA&%h-^%}ig`Sw<9h0OApV0bM?@z5LYWkSHTqp;VIsFJZ02clsU-(*Myy6WKAFVtv~^dg -4LKH%p--@QomzY&es7sd%5#1Pp`2+8`uh=`RgjF1AzTc^$~u<4PdFZsJR;aeF=Zj%{^HI9PPjU!Q=4YnJ?0MKT~o<(71}~!OsOsOay;BkT33f -uq`5aX5(Kn)Ai_A)lHB{yaMF&I9G%&TQp{dD(EJ%-G(=IUg=k@#FXP -@;Cf;J%hPjBA`#!Slk!0Q7MbGH9p95~Kk+l6likkZZ^rEN;Tnp~z-_7c~KC4cQ`5!WS1rJb+WI{#=JR -5CrHE;9FQ3kohwkz{0;Q^S)qyJQG@$P1>vGY0s7Z0TTmL`bClg)+6H(!iSJgRlK@K*e_>UsUC~f!k8Q_>q*J#l)9ifjPR~>X0Kbmd5`s<-r -CIQP#9T^N1#d8JAL6)1T^I4E+*Lf{2*tt?{58N=34Z!Wwn=?;dWbU)am0`2gpc8XaCje0@%7tc8|DUP -d{+hfsG9p$?NZ6%i(x&BDNo1epte9C|K+q~&byNc+&ocdjZSJ&wY6c6pgHJJp#yV>qlhR}uhqdnI10S -PkUuy6pM!;zM6%qX{pCTskIaJ}&}wLiJXKASOc{X*fDw{bz~pI=a;M~NvcaJbHy#X!F;LSAksovtjUe -zQq+?Eee&PzL6x`UigCwF$cDE)llv6g6JIT>5MW8b@0WAx{ -{cO^an1KLpO?&NO@|B3aYBUSyF6WuQT@>idNL~dX1x`x?nbP3vfhnXj3i@*<3T5AFL!{M;IUDUb9h&q -J^+O5LAc1auCft;b?u%ZTxUa{t8gRvlw)%26#(!U`r@-0}jJ+t}Wc2Tz=v!C@iV#A= -HC4my+Q7KBq%ZL(#KBfK;G-Ksq3SoL;;)$C#%BbE_v?j!)4dzJ)}=3R#+NkYf8s^ajK=4hiHBuGp3#+R8RVL`WzuqsL9 -W~yXRi0nZ`MWf?j4AlJ3R)xUaK2`79&mor -(7|t>58N_Q=ey=uIszbLAst^yQ`|mOvhqoYF>VX4Jc>&(?D7Gpt?cdOj>rC@}qHag!ObqQbw=i|(e-q -f^s!`jHuzq%g^4yjG^+M~HxaITPECd}d`_s4<>Bnr8RSwRiAw!4bGP%O>Am7Is#c1mVM)2CykD!b<{@ -3=^9_Wgn%pgq?EPlm6>aepG-p2Oc`lnGjT2h{Or -*P)4ANaSlS}=~}W&h$u+2F5yex3FK&Bab<>v6CEYiaGD|2&x1vq3p1qk4l^Q}pxBhGp+q-Mo;rIFrOFkiGfp&R4q(c -VXgXlQR=<+Tn=`)#uD*cop(O(N@sst#75h5$;q&ao|6)6c9K3-}by1pF2Dc0L>?B{8>g%d(NLGpWT+rwsDjQqkoJm7Gsjt@O?HnZDVe?$;UkWz(Lw03lcWhnW~?dpx9(magSNH=! -#Xh#FSK+7)pMZl@!*=eY)@Jhx|gOUZ5?xUf;ExtYl^xTPp(;KpM49_f>r&a^L%CDNy}%)hZuPVirkz1 -xEj8oZX*{XZI)nFiMqxg~l&npLqVw^Ow&Y%n!<_+T4aSb;t<|w((M|_p4%jxq>swnM~Z2TiO!+`*SrR -lKl7|v(FZ^k)%b*ZZvk5+~rK%|?jJZ!|p_^h)-##h&u9PO)pKg8gk5P`Z7?|YDE0()qzNeqY3*hT~n`msCtn0as**Z?AKC?(I -lkYiYs;L>>ZQ+IrfH(>8~F~nt$YR=9l5u^2M(#Vt0$4QWfj98I4mhf<1$a`xvdrnXmO;!l7r!=#{4Ku -${_`?u?yJF`++p@|yUkwGjR+VYSjDj>_ZX54YHIyDt)qXx*?z#`_{3j%qUQcA9ae1<)951_r2$n|e)E ->bMmsqIyMrXozy#D^Plrx~}j{3u-eAr{hYiAUSc!faP*mwkuSY3t)QGY1Z&UR+Re|7yx>}ZZQ_5T$gL~xb3mR);Y!p7afDfV -93jDj%h(Z?DJXn7mU4u4L4t-p%4Cm>6H0qn*Fe+;Cie^xCYImXjY{Nf8daOv#J>~HY(c^dm@nK4jN|# -ek*+gW2$)q&=qc#2og3=z)!P>k?>qeZQC`)6_rdaNlm)0tNpIS>KNBJ)2wJ=OIVdvWiSL2+5@j%mU)+ -9?a&-6nZU)qnPtyNO)B;B~*50jbG3VPcng7Jeb)Y(#53g1vck?he5{ZHW@7Eug$AS0=s1P}kfL*4TeAYxZz&WX -_+TqS5P9>B=c!Jc4cm4Z*nhkX=7+FbZBL5WiD`e -tykY~+cpq>_g`_)UZhMB&{qQ6CPZ{HnB{j!uG8>$EE2g&2zci(q -Bi@H(TLZReVJT^=lt`}RaYOse!$VU)<{M}6om!hyKTnmH0f3k*#Q@ClYP?A*(zbY&FmhCH>8+)j@X_s -GK{eJQBdY#{`|GK`q3sc>ccE%mOr2e)(G<>mG6csZD);nJF&8O6uS(U8CCJ@H-G<}Eu6^<{b;;#x`NX&MY=Io2Y>EUP$A4@2^UcN;6ZJ}{*u+WOi8dR* -vR1017YafdnLJZmw}~l;H$7Zklp0Y0U>u>3k&gsQq^eJcVQ&pJG;$V?|IfD<>T|ymT(%8A4Ox3=Ryjb -=W&k$ueNEovQ6+NqD;HndI==*R9>0=1&SeoiAhHHE1(&hylQ?v=3|AJO^j{~Ov5AH;5xkPRtQCV04TTyu~+8tu(=$H{*+O=WY3Zjko^! -}5K;Ae>Nmi?`S9K*SdjuhV$E;zdRPtk19sBvD#!{h`2dIb6>&Fx;x;7UNisk!EMr^*flsv|iC)caIaC -JhVS2rB1>D2p04YOf1EU^IPx1$znJ$d#dJ1o5?EWv6IY((r|Uvrwy&JDfw2p=(@ddZ3+?5CAh@A=t^#T2Jvk7;~H+tpmI1`g0}(lF<;fd)8&T4W*2lK;0sG2eI&!VTeZq -9N>II@xWG5VNnB7se((}|!fy+XbSg_{_|1z~d@F}%v!xsrn6C5#-H@EbR+!&V~3WGi110K(}2aHgT(o -4eGpA!B@9f|GL!ZMZ6RXY;oGw^&)K(8dcd?n#sOF)W^3CBKttM(xQW=555v|n2?E0gtAaZU!_`h<|9c -I(cv;-4$6m5#5z72H9TbUoNxhUXt|zdMm|nrqOB3eVe;Fd_hb*oKU`BWAkCps#b&0`xWYL^J)FHyFG1 -t`i#rJ8peO&5iGAtWSzxPuw535AJlz@x;`EIv?tEH>Zsg>1X3l&2T)9mp|gUyBb^CoWHrJ@4Z{i-t2P -gE5yXHc|Fg#i;>=WW6O*BU`+T?+2ZvH#gVbvyyV7`G -qCv$F}-DV&`RGxRj+D2Jn=>zwIAMus!&ez`0|MNj@u(p59a`7`-MpkgLVD1|AQE|WH+=e06+rdz|v@X -U`#TPLtu!JCuvzy!v9D(>i`bmbWJir&EkYq5!*LjkA7|0Vc^AQ93Tn&;$Jl;<`6d&~1^3Gfp%nmyKRN -q>b3^#%WW_MB{wP`_KFg|xAc-&eI{@h?zI0|XQR000O8%K%whQFZoEM*si-W&i*H8UO$QaA|NaUv_0~ -WN&gWa%p2|FLY>SZDn&VaCu8B%Fk8MOG!;jEX^s2FD|LfNi9~$%q_?-Dp7E8b#n~$35gF5iS%&|=1K! -AEy+mDO@*ll0dcsvxS%Q(Y#|~Vs5&*d08mQ<1QY-O00;of09jirWX9x|9smFpi~s-~0001RX>c!Jc4c -m4Z*nhkX=7+FbaG*1Wny7tYc6nk?LA$O+{kh7`zz?tgE+d;oK}~I!{8`emUMSQbdn&+J_qXs#ToYO@^ -*$ClN{~N2?CrrZ#EDlh=at4AM)|wmq4(c0Qmy)UlLosJmoLAs`|Ux&6(Agk@L~M)v!ORtE;Q4tE;Pv# -d1~EZB&%aip|=+g1;_`lFhQ&F_Q^ZBY!oUW^Z0s%P7yZ3l=S_+p1cqVnwqk -W{hXHbv9$uY<40uPqW2I+D=98W~HHjQswIfdjv_N=#i|3uc~ZO9?8P-UIfqeaxiRq%MRx$^F1EK1sO#uM(of7b;C -_Sqd>t)$!8qurDESqYxE#EBc3@l8U~IpjIk35qQG%+ -bi9E7fGuT_2keNQeXP?*kg}?c1W|RBl}wTKKE8;IW>&3KUu!}!r8Jq}ZZs_(;)~4Klx=pVa_DU&{=Hu -{1=2~?CssTcjNkrUyo&dMVe3)s|d^iH%t9BOC+IjUo)+PFaj26mVCt)gs!{mXWMaK-M -K24MCuTo>BqoC<4g3(FB_&r4iR2#GXN3(tX&His}0cdJs!n$se0kBvgelQcK3NH94|1BKpypO=BzS!a -BPOpvJ|>ph-Bv(=~wlF4ZvR+sd&&Q_CX3JW^qeV9S-d1iB5oq~*>L5eS%=cG%Vovb`U40+?7$=)g7(6 -Z$z&!vD>{ukmd!^2Tzy^l2JmNs+<1wrr8qUtv1wn3T*n}|7OWQX -WgB2U05N5-CIiwehxhh45-WsSwNL>=X-EyBSkGEagBCn;P_Dvphv!F8C6a~{x$H) -56`j@~2`U4%);+1BH(I(CgtG!Qo0gy=#7m&$qCrk#0@AV5!tz&_un_x2t9`pLTf?4g_#l01aOFw4e31%h%eBwHRUZ}_l43XNYp0mX -9Cf^m+U^Iz2`nAg);zvlB}|W&&CU)29N4%3xG+X@Jy7Dg?%eNvRXXI0Nmzmb3z5kqC>>763ZEsNi2Bf -sGysBn{e-Sz+LL(Ew5#4i<_=PclFma{6I+1YUQfKFQ9C<$4*3d>Sm6P=Ge<4yFvq05qCTYxV?41U-Yw -HHjyp7A%`O)6^jFu{qR>6&I_!U?oZljM{o#&K!K86(AE5#bMfR(}slh(EMY_6(v#qpBHCR*5uV1M}sP -Y`dT|Jp;Y#_)ttYr20GW%mNU}UG7;t@UMv#DaNeUBKyk@&=8=>0%8y*gNr%z9|={32O34uqQ>eZVYiO))Sb<8Sdj7*lfVWnXz0zL^YIzj%+EFOF$w8cHFc;@; -brZ4^g>OTeXHW8fsfIaWAl@<6^6@Mo4<3KvV9tJ}#Q{v-N6&doDd17C_PZPzzv0qg#;jha4|d^*C*gt -Lj7}gf@+I68RQ*Rpv>ySOcBy^X%)>SOgN|w_T_s&67?_umTsKNhify!4Qv0Hz{0 -&Ghl+kC+?;X?TVJfN~|}BHbU+s@`s(Rb=gG!)J29}mQ|CE^vKx5kH+*VR-05$p>(#Wnsv?M0`?!E;cE -7dq6;98vZG}-!BufqJrPrpnx3X;mjpI{`V0#;EiPh3$<r|=|`wP$<{i!fxOpUCj|m+b56 -_EJ47+JgRFpYq3|=eBQ9u5!}mFfz(Q5x5v=6cmUdA3BoOv`6-I^!HAE}&}AVG*oC1o26hGLiLN6S<*x -0;1iR0b2!Sef{)*G#nk_+KAp3!sDfJ$bOEW_`;<}-q8aSU)RO-0dN);M*(W6d;vBw9r`s~EhFQU^JfR -|`vb=gd+8hUiEz-CUr?ZaC;?F@k>9cK%c)pU>(tye&eb(0-2-O>fA(?yOj#Emw~a@sApVh1Kwb*kt)X -BjA#K-gvt8${3&b2eQc9pMH8xA;|kVzkyC1D)Xiz&ntKdL_y@2`aBiF*9U)XL8dt^++NEt1K70O;mPay~$SCqt5K_h~UH{?A_9Qn$kRn@{4Ra&9lf6oRgh -pr6DFsr`9lxMOrB3oBnn50x~Yn06e|m127_Ez-$fc%Ch*}6r4t{pcg3eM3s0*!wTXFI&oCOdUSwiU6) -}20Fjd)8WjU;zvkbXb|`d=piA!Q$q*^sd1CaU0uHKA3qTx^m&+vpY1uMT5b&uMxq*T>uy9(!$4*AmdA -E$f)dsn@xAzToX1YgzqYqK)aD3%K-~>0|%1B>UWH+SuZZ5q63Sxzl_E@z-;L#`FH1fl>@m-M4hjOXlb -+QJA3(uw2Ga{a`zK|TTPewqfI0+h@r++Cwbh(8L&b!G4rS6o!+hHOsnp{x8vGfdid -PdT>x0xdb^NrDE_Bl%$HBk_{47LL>@<-XV(J&)Q?-I5Tz+JjD{KjLZ#Sa!?QAuiyV5FP7KFm);%X?UaYAatBdrCGB6D*q_CeCJW!5d`s=7MLfhnl~nTdu`A! -H+|Gu*#@8*7q+9d}CLGoS)A_b(FAi=Z!jZ(Y>5%FoHpCJcTcFt~~ZbcYc;8Fb(|RuiI9LNjnI6uT}C{ -RA6@4jwExLuUo*ro -P2{OGbm;z~vc7t55>bkQNCqvOlcbV9!IzrXRW6DJDZ{q6B&|%32tf>AGdz$&0!Mkf&4PAUZLx$gJx6CfQBd!8b))v~IW_# ->7H*|ec|{4=!Mw8nXihQ>1I4jBC_=&lPGG)4yW}MUMI={hM(6~1W}b*@! -w4N?qq~`%`Fakp@%c($`AT20^%XqG8z^=e=(!2%`;D5RGg;;y;b|uoR8F)&wQV~lYe%uiLxa5Ql=mUz -k2`g_Iul#FLt(^#;kgdqU?rqzPjOO2p&B$%)-H_!{-R36{OPI -de!G3HmYw9;T)?2Q7pqBYJplb8!U!it#+WS&@#&wfSM>S1heozZ^wdrY7{>$auxU)Opmtm}lDzE6)7~ -Vd0=3?KzDdhh9#&%F&?j1?`q-Y8dovC?-t;q9`wGP4UEgkJEj3aA&K=j_?18Q0&wd?5k%eIqv)4JwYO -5pFC61an6q?9795OIfyO|_0@SxJss=x0|6rzFtM9xFb6Y->td)^%M@Y!j(;#8K`q+IZ@QE};92;u?E; -`;+_k9z05Ke0o#Y@c7^yu@+5|v0?ikq$w3l)4_;eC_x@MFkEoSE60&wUBP8taG}1kHYRJIM+DkeXFLU -bS`aQ}O*UsZ58{+d3hjXoi5)NMn^k$S6|SB}<3!37!K5jUAzt{)jEfrc7`k#Y;OO -T-7<`~jKnQIZ%`hi)fdk3;L=jA;jA6V%bx2d^`h~1TY++ZUBu{b4tc -m>5Ekdl7&|f>#aWV%s!IQctFl+JR1ziWY1|0R?DB9om`zsqK0|q$^A4ta@jTowXt~dIHI1j?2f{^+AE9Aq~giaQ&K7i^d5?(5$MMy8CV(wGq( -M}G4#uRX;LH#)fZEz4k9G~8m$7!VLY2>Bb;MCK-w;7H?`5-?+(TGB8r>7Z0$ipP3y!4G@jFy -vGJ_Zq0l;&U&AGZtuRfmeTVqxQlO}g2AiUR?ab3_Y`=)!eK)$g6@-ip1yJxRSXc?*vG)3vgb?*rJgaz -%($MMLIg4Xpr*IYj(&kSMca-J?*;d0XfmGd6f06u>^q1WO`+5R%AZ)sy}~XfkSI38IHeh9H%7Kd -Xu|HY^n*Rj4~}ig{NO@e^ITi<{afA8EBLT6LY*kCa+gx-!4!c+%x2|4=Fg@H}dt47BD~$p~6N$YEvxG -~bOTof9k(ca%dbnC_k;C|S4(EK&t8IUFKaO#q@HPXVuC`>TqyZiSeQF5{Og~_$4~j6gFg51u -k|aXR4JODVI+G4&L{1p+vp%Ci$RF2y$eIl&cM-zRqy~#CbjG6`j>jN?v6lJs}N&nJ0gZSEq|BPt2Uxuw$ujQ**PlrZVWzz5VR)_mX!5DMb7i)%nPZi{xo -lEH!Y;w!-%8Wq(LqiCAh8$1%w$4}jfPh9Lu_WS85HKI}b+hR_^#kB;zwlrE$a1 -nxeyTeE|MyZ0X5I&^4)5P%+5-!aW?TaA?pfz-880(z81UvA*PY!4LB=1>h`n!56gJ_?=x_#C^)Ta4Uf -9M3dAHvEej{5N+=l3&fB?`1fMj>YCIY&Q0Ju5i1OwO|_!jtoo!hBZE9o+DkI`Fcb*(Jj8z?Zg(L#&wM -O;Ng0DS&!<4r*g!fZIQGgnN9Wk6$N}=NmBYOZvIb~m~76Yjr9x&O)$FmyUgknplT4s!FS7p;t<1sXH$ -%ssR@YTzA5}{K_NrAO#?q>aEyMr=oN@gva0;#yJdh#UAFRzq?bqk=9(OKDO$O>jn{U2gdf15P=Q7#tJ -!Q_*DUYDUM1ti-s@%vj#Xt_scde0deY+#-<4#jDehwZ)ECXS(q8Y!|j|R%AImd -`j)YZg@9cn9k9S)m9z&d@YFvA2cl}D%S>`EsH$jVrL*eo7fA%BlXC=C)sfYj-GnK(gNpNPy#bL9!#v# -WTXj|Mek3mMSiR`9Chs`0T4nS+0E^;)!_W)j$4fNjbH-?ZmFQ8rDeFdk87mP$~@QUz<@Pd1_QAy&nkB -B5;r?0WRY_@!Y@KGJCi{6HislCK4}GKv~Au`zS;>$#Ye8lM@A@lui1Wr)n-930c+F~2U}O4d`rz{Zp*gGh35>HWk- -ZvHS=o?yU0|%#tmU8YqH|ojf5~{fY6|b1fsk&Fr0&Q^+Db+{((9x -wiA$)CT&1HzN}NRk_$id36fdaa(#TqYX>rx@LIX5>uS#g&pyuWT)grpger&keOa$WDI-uiEo`@$9c(v -i4EY8XlcU)(Rg`Zegz|Eu5i>aEti7}UmFaGC^A$Tu0}>#Se{WbhrQ=THN(p7IP2?clVct{jm{czO5)L -M=o6QUg*f{S$K>p#*^Zq1txrUHTjg=twk?Hx?c{5qVKD1i_0F}Lc@7_Z7CZAae8T-8Y$wZ(iS4kn1XA -U3$O+Nb7|_TWaI`R>^xETik$}cy-yQ7+X9_seYX1d)kfv}o;ke)r)YR*8D{hOcz#__Hkmj-#q?dX&?* -!R_*>|OL$^EQ68pZHAAMi}q3s)(-wqhbq+^NB=+{AWebFrZ?Rm5);01ItAW;CnrF4j@ -#Z7g0%lkaG?H0*W*qmI%g_4?PP*b6y(UX7mC=m6sgm>hr4+3f4wB>1=AyAQ(PwRhufjlfHhOQ^p#lez -w%hJ$F&w`(0M8c!rO-)XSFu@HE=7va%Ly28fmeZbPS8MffkPFMSa4slT(ui{#k&yZEW9NuooYyEjiP%GdSHrrMo2O#b77%6VkND3t`TWw#-eREB8p__A@^JRFDkv88)aPh_FFZ;}b3ujBH{7BJW6%yopYh|*)+^_(Q -QcfRo(LQH>7#1ChOg$kb-)xW0%%{|Q`)hsZ6=P_FA{hNEaq4&&OJ|x`fmV(mQTtHnnP5E3P9~N{06!VA;Xw+$5G5);T -O+;{6F9SlmGqc-$R)fKl{5E|M2r){p&ye`p=$Y?hk(X;+H@EzyJKviy!{QZ~o(_zxmPMLU~C4;`z^C{ -NT@3;}r?geQjf<@4#Z}yHRF*`*mJbaTfaMsrL$tYgYQ4j1{)4^+3m?1^Rm2uk4=It(S42TkXz#aBO9KQH000080LuVbTbBnf7o-XR0P7?G02lxO0B~t=FJE?LZe(wAFLG&PXfJ -efWo0gKd9@nLZrsT6K3~zHb0F8$KqEia8h{2ITkExVZ7(Ek_AnVR)Uao!SA2mFO9SMP*D;3#@ghJFAO -Z3R`6PcqJ|R`r-DEe(k(8`M$U9_rRdv0qx_j4Up0avZh-|||TI6NTUe&ye>b&&X*P^O@_O7T!o<)hz& -YL3PlgS!!reTujg~q}6a4kxNr}rePY6ZB;vno%x27N9+ef=WIS3W!CSqv**B>a3=aQ%JCt0vK;zHf>h -aHmPVmY;9Lwa64i%A=|&+r{3<&y;7HM8!Awr2)?5r=F~L9iwOX;me=1I$B+-cjV55k1d3FGI`mEn4nl -7Rko;V(5CgD!q$1oB31)5;y0@rAT%7uX(jk0_V*dQOGPc@71n -KK86oQ@c=9p05Bx5OM0p5lZ-onIFbh!!*VBF__x>a(QESuCm|iFRzY6-f+Quu0n1*^13Jfq(S~u`k;> -LzS*rqb@z0=oe8^SyVy%GSA1Uf~Y7o&1ZSc_mTo{Rz^4dGGMMlvp#cvrGD=p9MJDC@zk2Z-4vv;n(n|zCP$L&mYD6UHkUWj~{*mf9mTm`fD8Thd;G%|M>XBAK_1Z{Z)S{yrmR7=4%#)A`^ -8Ox}B%;WbNA#Fz@HVcKS2l_V9U^jzBBYZ~Za99RZYn=nwpG&d|u1x4^ahp=-7CWLkgdC1U^{J9x_8<{ -6(`f=^|{2U2c(O2%6t1+}NlfDBAAZ7>}%bE;ws%(6>#o}~@e24K+Yzy|1|D(KYE2hkxCbc*QA9>wt|v -$(Fsi`Jqj6zX=5j$L$jIvJ5O{e3(nM!yP4WJm*cLq5w -O(sJjlNBukM(&6-q#NkY}54{8%r6LaRM2GP-uQ*~pig?EfRU^eeYKB%ydRw-HUGZwN(mi|t8dfCg=1# -C0|i93b&ZoH|s?Sz!M^BaB@Vtm28>hUSuT(5K84X@ieqU4;mV;!L;Nm?+9?kNtr-dMuLiliyHeMj<9; -}}SR<9L|EI?^qzvu~a~KY_i`ySL}B-hK;jU%Wg0%4a8UUjXSc$~N46XowIh**Yn^IW%t?UGk7va0M9T -dd@^v`&R!z_EN}p6hj!k0er~bdKI|`N%(t6(jzI(j9YxbmzA#=+X{OPe3Y3;U1oc@AeNVQSV1HVDM%& -)Hftk(dXYocp}296y3q!`64yJDbBErI+!&4L;?NW);hSi+3$K8~{Hiii!E%gg6?xiNE-s;rVAzt>TB+ -brpcDv(5j(#y{K{Ke5#b65EVtc-y9*R$>3OE-cgP_{k9KX^=eqk{*fTQD9kKg(up7_cD|sYbrSV-fS| -EE0)|pC;VQlKY61*}s)|7JaBYj%<(NZ8c#ffPGJajV9Tgj$U$1}{;65*8Is@+*p -+}aT4p+wqyh>SqFx4{^w#o##IG8O}g&WTz!^IH=@C{dziN&<-ZR>|^1|AUupWC4z&DJ7oY);8(_h}f_ -1LoML)S+CED*apZWVq5$#Y2HtO<7|0yKm%n+U)LCp0$0sjmijqQS!C8?(D5nn7!Gcq%-r_yq|1z2N%j -Sp02pbDLL#86e^FeBu}!rQYp$Gfa)~Z6meAZAt`*WL|ktJ4VgyQA&rc}9;$WtPXB!A{tT( -dSh)gMEgd2SJ^;3JGIf2^SApMN2P(*Pn$@fiFi#45&WnhINjh=Ub;SkPMNei+TgMP4QOmFDlD7!vU#g -YUX=uTPPzLiG^sKuq)#;5Vw2@aK#&NFp^!d@!bDIeX7NQxoi<4V?#s>cumn=*&2KGrP{=;oD*kzk}=? -8hC?ONNN3tQ%l-*SJraq$lqUnctN0zoJLi6>lA$|7gtr?_6oy9H6O%iq+R-&{ -l;OXuUnAX_ETw9LGs=Kn3uJxGTuVd`<8UUKyF(ouj#!lr%hza5AL5Nv(-G|PhD$$VJ)SMy2$IY`_d)8 -x-cMbbe7e1#ap`jF){nM(*yNHxs#X1QJHJ$9B5hLY4?CZ@9*(K2qQJZ*!VRF^ky|)>+e~CXq;se{cbN -*B`pXvme{L_?0ot}dwZGrmvmz)@JA{u~3p*{TH+Q8k@N#`~Cl9l*W$5*8kYH$UDmna&J?yWgi~ -&tmNmO6M^q6d|Nr(oQ2BL%+G?T=umHcRAfAZ_X%>o>WF)8JJ(++CqXE>OvBr-UhOaumF7Nrn`n3_e2G}uv>?7OI}~eP%C6~NJn4ca$I`U_;L;vyd;ea!u`X<`3qen#b2~o|H9F -J#P>X>259T9tQ{J@_*n)weNem+hNS)55?pIa=llVZw#KAX)XuU}*Nf6Ze-WP#v=BmdEZWBVjRLy1a$>is`ITHOn<*-Z@<)AUq9?-kRte6M_&G9+uW+ -p4eAi6Mguld_GX_?jl3LZLzDjIl5iFg**VRlxO*6gv!BwGt8Rwf|r$Y#)-^wDma5SF1rd%^4+_44@yL -of=JUeo;+n=!rm$O*>`!F#6Oulg%OR(!U(!7d0Vm-*`{jwRjaA{D>`R){{8nq{EICv7`f{56c6p`V#h -c3S#(L%-FKOstCJHPM)SJPd<|tYT;~bwS -71y4a7Nvm5rycg>KTt~p1QY-O00;of09jjISapJT0ssIo1ONaT0001RX>c!Jc4cm4Z*nhmZ*6R8FHA{ -8MNU&iT~f(z+b|Hl^D729C4i8&Irc_NbVO7eC_1i>vP4^KWzi6+8u|Oq$WGD}Fd%T2w~ceDA+xO+yg$%uS>aT9nm3=9!0>R!Bls9R=8`k4X^w}?DdF1c -p-Pl=jxW2ldohZ^)6y-sSCnfmkKykTxl6^GOnQ7sH&_5(kBR8$YQOKMdYH;k -I0HO5HI+EI^|hLRE%#>y&`=Om(@kN#;)pMn+hrs -AID#;Wm%AoxGc27XM`@5eqY_nT13E=tSOMSD)Wd@oVZagAh=hED+)yopDuzU^xo9s9S{=1GsLQCnJm5 -yKeE42O9KQH000080LuVbThBUA@azBp06GBx0384T0B~t=FJE?LZe(wAFLZBhY-ulFUukY>bYEXCaCu -#iL2H9R5QXpliiciWNF=qlKo1fPSy(mT2AY$YI7yf2u8eDG|9v&dsh637dEdM@0m^;yxs>`$W}u~L$T -)q`#A{?idBJ17)h%so|I7ONN`Q`ZPMrh_6i4O --7;TRaM=#RfP|{_^qv*!C&A$%v}|kOz3PIN|`PDuao?nc+l3|5kCsw!<7#NNDAbILYfE#wG$X4e6zVvG)1TgjZHKVxYaXV&y6Y`F{PSO64)&;n_Xki*0|XQR000O8%K%whUyVp -UWEubfvSa`N8~^|SaA|NaUv_0~WN&gWbZ>2JX)j-JVRCb2axQRr%{^;Z+s3xv{VS&8gYDKRK$6qe>(a -FxAgynJaDkk*hp-e`9%B_*awR$7a{2Ff?>!IA=wTa@c3rR40@lo){eI18vc)dW4u9J|4g(%%e0wI=*tx1~9nCDFv`6A_>$HLh>Npt4MagzIan8aDDCBMs-nHZig=R8v*PyG2jjBn1; -_@F)nCn^GddFeRnV)9-ZIt*q{GI@Vx7aX_K!7aAHL@3en)=I(`GSQJ$qGce$Z*A!2a -7ZV1YH1>Pm?IQ34ssM62Qbf!-3-@Sm2zD1^y!AY!0$zXG<6yqh6$b2w$EnI7ev;O3}HShQXAD8GHY(+ -XGmjD{Qu$F9S9Ln$jh^n{pmy>@JKV7NmYQ1?2){lF!7NCF_97;AEE#f*?u9h-{t!Bmo7G%2+-PC4>=Y -lkgsp4FRVq{6QMnhvCyapJ)9SFKC*Za(@iZakd*KFL?YS0&U3h7vpd;;VF*;o_WY>Z%zcd(|i_fWw{^ -RZX;@wD7ni3>Ki^ufC~h)o#EO@YZ&(mh64NeI71Era#1)7k(Z;TKspLXDR99BPzTc_AyIq+lW=c%t|8 -E*PWKb@Gf>C-aJHDi5`}1mWJydgA3#O?Gmm;WJOrbQ5LY%sJmnBr+mB88^++!83|N?f6@!+riJu~<

    p`qlnYc2-k+QtonCH=M(gtZ$+7 -qAMV)pyjQAx-c!SeJP`qDp`C_zuXsL@s%y~0_*zoS?_cA8MIC8lTxa={X+)TqiZlhV8%>PWYeDV42e) -*TfoWYxe!=tx9|1$jVcgH8E@6Z0{{NnP%zkmJh_y54do!z~cuU`M_`yYOUKejywq6e-wJRM#RkrVd1H -a;bUoi6aH0Fq;E712<(AImpN5XXA$>BM#9>_gf{CSbv@PwO64C=CyiMfD1yI{Bx6xXo0 -hKAtAIT$T*t41mg*SFUMV%)-?N|E%ms16+RFZ$Je&DJd4u|p48E-hKNJ@Fba*8DT6n;?XTT$!tHLY%m -8(seELe=SnIi*<{!X|Tfvg)Oi&*3qJ+{+BestM)Y}YgtHGZG%+IW)p5lpn^Kyk#IM{d))DY-&>;g-27 -4*h8#>Vg;zs#}H!Wfc_G^AvzOJRVD*N-F29M8udlamc$q2m&8_%7;D>x{n5&G)3{@j0gDTk#+nWM=4m -3-C}IBXap34O8nXtfZ_l4*QkCZA^U6O%D(pL`fDvx@r(UXF`s@yw8a1bHTSsVd6Rh`D)y_<65`-BC{I -sBEpLJ~dEcndk#_vqW!pcoaTw%RVD(h#U+dURqeUpBir-iD@&qCqOGbb2Ag6$j<3|7y*Qc_!BNR4Imx -Z*D&XGGNPbHu8`+(1LHc+A42-x@$)jTGCB7d+1?_lSM%@!GkXO_%)EGlk^;tW;+3Ar~eW;^T?4(PNeN -#f?oENX+VOGoYQr$WqKLE~u>i&Q`u8SypiWELOyjo-Tl|1%IkMnUqqoW}$&PX6@!?BK_p-DWw!3~;Zm -P64(kW60(_P^ET|#GiReg|_oC1*62&K^P}^-xxV1fXMk#h<%_abJ;P^A-5YREK6`shJBiFvQ&AETO-* -tAmL{q`*{SZcSkJQW1_?{MBTAv-7BQq-B9im=ja~j#QY7j5ut@Wa>MNh%qvg`T`<)@7k(7`Gwykh?Pf -6x``e-4lLZRr{UV<)a;c$HP-g5{zJrQU}}l3gX00pB -&S{?i)3|ATjmhjE0j}1BsCeWXU28xWMFYibimmE)1`@#dckGaU0IXGzj&SIl@rz307ewXwP7%7fb=S4 -hXU*BB-oa4d1rK_FPFpx`;R^ijV^#D4HBaHf~}ApKT^RIy#PqM~H0sQIO1#N#(C^`SQhQ^vWQKFfBiY -nh`;wWug8?A$W)mW;3(oGska2$YD@1h#B3=n&7{3E~albjd-1d(ndC)u^xyKod^Uviny!iSk{9dWIu` -`4X%Lw*F~HXRx`xVT5^u~83C@hW{eeZ_Q~$-p%l{hjhIf$CRi3Ocm)LLTD3(`fwzy6@JJFK>V%;`U`j -n{JeDKu)&NPB#{^oHE@<}!qcO?3wx+FNcnK;%(vC!ZfO2;d`H;s?u!|A?1% -823@z$xNxE81ysw9sTCo$xcc6zAPH7vIms3uE|@KD>=!h4>yz5%+nN_0sFUuq9H!&nsh+B-r7QEOP9= -%sXUW&dDeC?0ic=fBd<%Wyte(ABxAzt>p5sWaAk{#5HY9SLK8S0X|8F8oPO+hlc2Z79IBuP_rK)a$XI -`!4E=Zt=WDP1gw{{e>O+9HQc!nibSDK|hhb(9Sb66PFV|z0IK7gyML%psj?$%VOzCbcxLcY`+^@4qf3 -;-=4q!&(UeoG~BTP(|&`%Y3gT?D}qnv_vn*DY^w_c7*n|$02Yu~M@f9cQwgQWEyW%chB@bo*1vXGwv2 -QCUf^-`+u8B(V0cNp4h1^9HBB1K;wk+@?_uOonB05hJo-jvNWak*9Oj7FBYAi%&5<5P&4t4LDqqYaen -q=-P`}eOY|VBHbo}vHyCzZ4fSCGXBM$-fsh81&rnRICusWSvHM481Qt^Tf#A7Q%1EBP@hB8ZGX#mosr -|BK9&wQWuL%S=H4j?y5O^z(Ed+kB;RpLhtx)?f%rCMY*B}f5Gzv^%AUYD9c6L?CUqnMynU395YY%GUA -u5Pcd9I+khTwC}+F*!$OT)~<>!q=b;8PzCds#i4c`fQXWQ9*Q3l?--vUT(qy%nQKuIYP$Y0oj>I7lY3QBSl&xVq&@+_Em7RAzoFcBVPJD$^n>M@|v -ua?6XjdF!Rh~EhybZqz5$n>03{;RMb0;El4uLi-+WvA>IaaxWR!Vl!a?G;F34JLMKtx%lm^hYg;ICQq101T -*}uT)t#V&ws3tL#`WU+yIHgBrbGLpy1$KM40=KjwB0wRrHK0V23+$HJ|0V67QAb_mY$HAm0W~_K0<3Y -mWvwG*a5Hx~haeT+#7WAp;$+*8m)l~+B^M`B!7X7y3zVHd8a_9i -4__iN8-^(;-n1DZ)ewmmc+6VLq!iUUdl!EK7w{zvRHlw?%#!4nb^!1mpo=JhJ>Q40y!^i^X)?p$4ypE`()$x|>#?RapL1h3|EwC`#~b -@Ld|_ocWBJ^Cle)UJ<=IWE5$h-G%wI<6Jvp__{SM%`3ce`?`%gTbR7dexxsYc9P9>NTjY2SvF$f8-OF -WE?fQYFVf{bLzn+yWqzuLVtX{_wS7`j+KRevF+4LwQYePT$(`tihB=H3@jy?WKn)-gLOz@4OOX+R0?l -04Eh&D{BGYQR#b6ko-BxKM!*f7DO0>juUYzKjr^z%N(YbT+l0jJ;{w@Qn%lrRdSlAbvSk=a2*(iUyk? -omAR@&9GGthcF>zO;{jYl6!8KyFzo?xPBM{<4K-!;t70NbM)(ThGu`#XZ3C-oftQRWpBV3^zy`_2HPw -m%|ZrC!&Ywm~*cB;dS{10q|Uy?V`4I1&%mCfuM2qz|FSE!ITHL!eYk>9b3h@V&eh6;o<~Z -%0Y0nQ-Zr2p?t$azt>aiiqvxT!F97dTpEO4-f&3@eAO>W!r-|Nb5i{VgW?85Zs2ZA+dSKYW<$e->yR8 -s%>Dm@LGpZn90c%-gTivUdKJp@7LYK0-qbQ!i@m_ZFo4uLF^Zt -Wk?V)d0yt@CVj-VH9g@T;@~(5gwG%b0vduwK8)zUAJuWYLzwGfE~bprW*uyReHBgPoX?m_5(L;pV*)(TLCl%G1j2_8h5ij)vmU -u1PkT)SMalGNmj=6w$pBy4#ayHPrQ+zg+UFJHAkMX{~h{~(F?>>Q8*}+tbG#75Pki^8!LgO@)ku^LVA -#dL^+;v)Mb~63B}K^9&a$5UeURi8iZ-delZ638RV8vwWC&T{@Fs_paoKDOAxG8okDWj&vE{mIJMgXb6 -0f>8)+-7)HwCnS0q@QQvn&sb)&B=)2yeVi|uW8!tX;IL=~lmSe(t|nQATaM30icu -ddzphH;&Wi<1J>ysg{(s_c1{8MZ2H)o|%?tVob8KxgcA`zQYBben-}s^%jP;2L-WivrBQSq4s%-mg8qr@AtUI`_(G -HhVUFmXtJf23s0#v@l)GAT~%T#UOqYVHNB+H7CfE>{(Oa}{MJY#^f?-_lh>T42%sJ$N>w@TU7KEWPkA -=I2o?7Q08rlt=i^64Zzti+0ODutJeplM+7U2b<)H0L5dVG7;w0WSo)j6IK74J!!G0CK9m&MIxzr1`ri -Yn6dNC2VqtHoi*OY&OJe?N*So?Cvrlx?O+P5W|jlw?v>NY|7@B}D7)HGF0Tvy(#aiv$bs=AF^d7BBu+yGhOB0O!O -PpVkzixlN{>S`6?>P%EjzE&I*K8$!=9DuDw14m)GT=hqCzOFrS3qX$Uq6x_7>iz)xHWkQKvc;UIqE{# -;trd$ZB4I*9&r_O(Tv`(l%7dwTtsdL`msV48(#%`LA+F^CgCjqcRg9ed)XgiYJ%!JdFVOq!gII-$RVr -j?IMwDD-nyjFv7}ro>7QH^>o`COGTku(kMcB`Pem3|*TyjXP3g!>p_)#+?xaiQUZc`!7k#d{eviD`{L -uJ|tqc=;lTF%C8W=?(Tk_I)Ghh|#UN)5%bjHW7FYJ1cZYK~+=@DZdg_!9(#%sK|rQ6JE#qM%wu)lEH? -RM!Ji;7TITsu`iQ;VusQduFh46obKKBF(mxVE^luc`Y^a?(Cusrb=$+IF#}!Ee&RkzNreZ{dA9j9%Ij -7rW(m3B-mx^$nfLUeM)48Hz~lIcNOO1zuApp@1>JLbvu>9EzkDsTiiN+^7BNKk#M!m#*nwN(w@xHCLd -D#A=!RMc{zXG_tuPW{nkK*|%)3AqQ?UqEg=5Y({4KS|F}ejBRq!nN|!!O+?46sv4D8T%&I_>wej4Ez1 -^Uc2(z0H{L*a#->4IiUZzf7RJ83V4yj-QGlBI!-iVc2(zk-Zk) -R+4-{7hjAMiZGc!GwR^Ft6hmp^VGkY0iWW8(T%NkjHh>p%(sZM7vd^fx+hECDogj0(z_72F5RHQ$94@ -nao9tZJCEX;)6BSEq~@T-SK3^YQFeH8_Wt~GaC+H?Sc``V@D72H4zn=uv_KAuXU;jr+yIXcQT7M%Itp -(gbWM|Ftbn~i*DU>9rWB&qojt+fSS4B|y_A`VO!LZ3&CPe!Bn8I;P7@8*j$chiN;%+{(qK2!HrQ8G+= -naxFL|jTov%hZPpLRsvz_s(SG!-ue@?49K`Di(NJKHpF(}n(jZwM9nShCh5vT4 -^0bM8#qr{?+|ZeCcDA=dBlq9zHX1Mni(k0;?8&YK_pm^~HFZbjd -z=p+(d?u- -Bup)Q8n2~@=8*Cq$Wb@EN@GsjnWXFjYjPCV?GV@fypNU|>GKqWX{!7=)B5~g^{dWxsctBtXjx+P|gqk -eoCPp}Uk`&JvioV1$vYZ)NA#@dr)QP-1uRa%;c(aPI$>F?yvZNvfpUWLo6I$RK@WX>8}tM=$b{@2uO! -*A%Qe%(v?60yz>`!9TTRF1?yUccV^@io09(|P^<%U7=|19o?JU%`KB`L!3BnHi=y$*unsxZr)hsQgLl -rBOboeYDj{VhIPCo#+(Y;m7*MV|D(1y>xUp17#?Tb{*0>j>EO7ci<|~x4lANJjD8Pv4_M -U!eX4CGaF?Ho0rm&xayB)`*Y_*@6aWAK2ms3fSzCE`)!`2X005&5000;O0 -03}la4%nWWo~3|axZjmZER^TUvgzGaCxOxZExE)5dN-Tad5s!sU*m5mlt7fmZr`cEJ=gJE4F!tqR5oZ -RVEFJDq_d^?>mxub+Qh_48sCKdZtl*ZU^5~hhBMBjpbv%Q0#4_GkR=B4Eh`m#2~m_5vJfPSfG$ -^DNQfEZGM1FHywsV+l1REFm4;fowbX^?cSbT9o5u}jO1Pmc<;G1~utaj6<&~dI6j2f<3u2W5K`d+h`^ -_ji`!qWH<>LBh7{Ym+<*1NCXCt%#KiaKauPB>WK2O=&g>%Z{%xYBp2$e`3_a+uOInQ`3Erf04&wc#!A -PB;@gD?pEWBuvKtZ003HHvOW(cSnqGyyOcJ@(W4;JK6t8Vr0gG#p -4>>HKtXAd4IYZ-9#cv$;Bd}*==Jdv3F)|$NWrYEQwY$tuJs?G;(oR<2GBrZ@U$s8HFiiN~#ZhlNvOVs -MU4y%FE#vL~&?XWAwQ$#5haZ)~NccuM=jHWuJSKYMi$>R$7%A~P&EQHuK{K?{ZKExJ$8Qp&Pd^d7&6M -J;|`Rw#kQR3A$waKE@Kc0?9|EO)6nyRjrw0L^={OovkpfTL`Q8<`3>q&o5a_y~dpaxVQNrF^{{_)X~* -GH4Xnj(|2gsFadYNz&G!DLVV*i4AAquG$#n#$@g)ytQkpR8Ovl5Ngh=L#MY8fL2{*@ -Cb<+?mG{0GJX`%}iop;#(Y_His^Go_?!9rX#@~Hq=~oZ;fYYxlzW2?DxQ%I!!{JMu6=Wm?>fyAJ3sSn -9aoog_2niR0;Gb) -IF?EGH%#ZU)Q<>(=`vZ_x+ty($Op7Xlu3)&<^Qd%zjO$?+^cmo(Z=1QY-O -00;of09jkJ?-FOl0002~0000R0001RX>c!Jc4cm4Z*nhmZ*6R8FJE+daAk8YaCtqA!D_=m3`Fny6@pw -H)8-iHA-I@?HeiSkh2X=cR$P&_v}k3>{(U!X>2==Bn-!e8*N^o(eIh4ZA4X)OE1_=Tc<9~g7q=9!k(i -aHFYEDcTi85po}u@QfndMfZp6r?{wEs21iDiT3y-Map%FHv%VMIin-p;_N40ZTl^HBY;t6Puq6Rz86* -N8se+gJYzr=v9sQVldJhPZL5>Br3ny?Mx4PO2S=W{Y0s)uE8Zbf3WC#j^w=DqvP4^KWzi6+8u|Oq$WGD}Fd%T2w~ceDA+x -O+yg$%uS>aT9nm3=9!0>R!Bls9R=8`k4X^w}?DdF1cp-Pl=jxW2ldohZ^)6y-sSCnfmkKykTxl6^GOnQ7sH&_5(kBR8$YQOKM -dYH;kI0HO5HI+EI^|hLRE%#>y&`=Om(@kN#;)pM -n+hrsAID#;Wm%AoxGc27XM`@5eqY_nT13E=tSOMSD)Wd@oVZagAh=hED+)yopDuzU^xo9s9S{=1GsLQ -CnJm5yKeE42O9KQH000080LuVbTWdJ7;*J0S0I2`~03QGV0B~t=FJE?LZe(wAFLZBhY-wM2FJE72ZfS -I1UoLQYEsn7YgD@C`_dLasE-grC-8%>h4Wv5scX0Y?(Lx|GA=>)-Rq1jZ9N*nBKb>R76-U{Be4J#4-N -=(*{tjufVcc_hQMo!x;DEA0-4Lo$6X>QNFuJaAFY2~1rc*E#;k9<;e;zGqqp90KO+vg=@Q605bztm6Q -giUm@}hW2b{-O}k0H(2eq_5&;s;Pm0|XQR000O8%K%whn^YOTGYJ3yF&h8?9smFUaA|NaUv_0~WN&gW -bZ>2JXU?q?dgO$H&-m(8k!rmPrW -ZzjsEmY|A#;w)sGUW;B}XGb1HP;uUe6Amdp=9fz>hI!-uoqbTMsXK|EvIsx){aTrohtn}Q4rx?%O^_o -R@lE-&B|Ebmfl;HYYfWg^xV`P7fXbmq6}oZD4Xg8JbTx54 -Og^;;5Jnl`4xa{hOZNFbTt)HvAxU|*d3@SF8^jpB8NZ#pJAFU><-^(e#pTCeudZ)?`~8nUKm8ww{o|8 -wUVr=D_domyUydB3V|7MXSEEl3Mgfm9LzZ#~Yqk!l)j1#i>6}e3CeGR9;`HV{(2x85PG@|1{&6(EaVF -Q}k=#~~w^Vc|K84D+S-;=^!5F|_w+q82Ziq{N~o6I2%et639=aUoPWM;NJXOqZX(V<~fQYcvW9uvl?DR_G{@?xKwmR0dc`l3s~cCh=4HrMTy=3+qpy|uDNi -DC3U$1F1cfxXI{2CI-KRt|<61PzpCT9{Gz08LNy>_V9RLeLj|bzhif4O -CV_9TIGdNMf^QLuX99@}b!~B~NdMU!i*fifnVI(#2pfa}yP}~rFVf~7$e1-kat{t%rDhFcQrO7d$fM@ -QR$j7nxRHAGfCiOpnkdwkhFE18(@;o515z_S4I@o(U(bJ%w{_jwx7?TpVbaREs-VX0ER*xb? -g_XN#>51b{L --T^mXX!e0w~mO|5X}_M?CliuUU*%Qg}>E$JyO`M8o;rsB24wDYMBkCUljV({3!rcwn1A*83fdbroc)! -G*p<`*CA7{Oo4d_0F?w(ILX0z!d76CrWyg3rg2D)DVk{z#xA#OW4+as4;tY%>yR|DP0I*a?(t_91@V` -LmYd7-GvdcEJz$BY*;>r$aS*`7D+H&NspTK)1N5d5d2D3Ak$kl=X9yvx&qN2$s7%fW&&qKQCQo-;?O8 -?O4@z%haI0jAph|L%XkU>t_j#*sJTQB~UrD^7WEJ~aSX}1B_Pl}c|{^MzZ7_whEVV-k$aGHG%>Z|M{$8Xqh}!UsucibgK8_{0QKb+@b?Yf -8TEWAW+5GybpOD#zog!MKJy0*E}CjibQoeb^R+U}*Fi5Vq?T#E?y+{ZnUpgxk`H5!I#;&mmkF5hcYEn -ny|+|2ObvjH9Wk^8@Rm*TIa4FZ|BFIuR>I<#{jg-aJ6e`x*a-l$_fxP -gOzlgm4>y4_`r5Y8~5*%3#vw2=SV6c;4Vu!gM>aGUYGxHb>t|Ous#sD~X&LV+pHsvpDFLzcAFC%Hpss -jiH8@{(hyQ<}@dLKt(1G|UipRnUtkD+)uf?avkYr&MN1gOid-PHZO71N3=#1C;Z(c{6BZ5fC-3b)R}O -{m!BE!_mK8T$JYz!bDxn&9&i%9Q%iCy%aqiPn|i9)fXUv@wm~A_5Sg-T~lXz;iK*b?X2xj;kSjPp=l6t$QWdEg~^i53ols(T8Oz3=l2%-8(CURgRMgXQkzqPy!7aS$v33(wusYR>oboYEk7>l%|{n -ugCTAXF!3nss(Y%|q_7Awx`qK;y5-vlHkn>RVR(hsZQLf}WeufTt*BXu=}LdDVOphal$j*&dVGn9mEUm!Sx58cK@r^N^ -I>JwP`@jLXi&AqeBtE<_=vE?dDZDzi|U=ZmelG!$Wm5+ND&FzRe8q=#oB}apIdO!Lnv2gLyCcoTc1+4 -LR61bjYkG8KxazYC#v@hp-@vE`MPVy=xS0MO9&WNu40jIPUrQ4~YQWrXC^}bm?wj(M@qtv_FAK`P(Jy -(H?Q}y_Bxjd@Guv9CsiP%;Txm1Ibmez#~fpeCdKvFbQ0$^aK1ZUq{+rU_={HlY^x-goDp1Ul{H*H~j` -!zNA?KmNGBJtAZ~ahkoNQ@=3u*Dc^?p=DUUnLvIrQMnXV#$z2@Ffx}zI%%ehISY;cko|{US$x$R%R$k -ZjZoc2Wc)8izn)5>%nKhvn`DsY%|FtMkG?LF3wLP7`0Z>Z=1QY-O00;of09jirDFFB!0000Q0000S00 -01RX>c!Jc4cm4Z*nhmZ*6R8Uw1EXc`kH$aAjmuR`5+M%1$j(NXyJgRY=P(QV4JjP%t$!Eu{Mb#$CTw55<|QSF^bKizWvSw+Nzb1(2VE%n~CWy6w*NMS~~^{4_G`hmiqJCXYR)@ND -8~;!L9oNXkintvIqRppd4{Z?X`%7HqClN%?@wQoxWEpVX+%joV>*v3n-%}sHd2Si4}KW@m7=wN>>MQu -zZf(4)a?B~J7>Awc=Qpi-Yn1o&YxGkKZU*-2H4yC!cfk5AT`*&VHKpzQ=q*{k4Evt<08 -d0kVclt5=zc3>Lh`dXd0Znzh;(zzNMuvjepPUhd$xEVBf}pGu3m2Bn!Ss$2@}$uiA1nJiymiMUeX=Pd -C%MpC1I1n*Q*)I?SYovyLuOPNcvOV~=9k~mjN1BZ%hBhyXJHB=j2sal|XhP0(DR~l7BAxiUvs@MbZiU -rhbp64XR_y*l;@&~D^c3Qry4Xjn332ZI}I^)Y++@vs6nsZqsknw`Qh#94jMKc<^xr6OmP>b|9{!$}VN -ikAY8jWRwDYf~4Ze=YJ;9AzSkd;5T6_>fCU}lqc%Jgh`JUK+P)h>@ -6aWAK2ms3fSz8Y^bEsJY001Zj001HY003}la4%nWWo~3|axZjpb#rucbZ>HHFJE72ZfSI1UoLQYjZ?u -++%OQm=PO3$Kq76Uh2>DK+5@FjLTU>VRWArdPBKf396Q*Ph5UWT-pyuN5Na;5?3p)j-kT9Z{6(i$161 -fptyComJ@j6g2QZS<8o>{qumy6xC-7t)dLcxTsJ63?!ZSz@1^@YO6%Y9uombWr%B(DWfpjs0lX!O1 -h=<8uggb{&mWMA-mF8r3hSb3q%vs+=Qmf3`KM7n9KjkrK#7%X5xTzAssd&9$Sk1j37)xQxt2=HQez0X -*hWZGbMVaotYC$aeg2^;i!A2E_Z@2%-kRJ54#rZzPes+m3StJL8@fSk)g(7;4&fUu@p@y4*YOEoD(q- -q@+sv%-Jk~HnDLIGwRNwNoruWVr#j4_BR<~6Jf5|LCa%34Xt>En8C`~7JgfU$?l;>9b25iPNdOc?$p>Px -r85|PfwC0Cg+9h)Hs5?b{88nXWTG3Sy2(9Hh9W}jpRKb8!jfe%d`1ulElIcHj8WVM%J(wJ6R8y6ou5f -DA=4kF>we>b5Ju}9TF4woIfsUz!3AWzh@%(2T)4`1QY-O00;of09jioVaNzW3jhF>EC2u=0001RX>c! -Jc4cm4Z*nhma&>cbb98TVWiMZ0aA_`ZdCgi|Z`(K)e)q2+l83dit}-(Lx_}ohrcKkCVACc@G94^}K%p -ht=0+B^B-O;g{`ftIq$E-o$LR*e<{_4-bK`d|JR-$6g^-#gBG>FmXM8c@c0{P^s>rBb3c0E7l}H|#uJ -2SsW -vR5%LNf1JoR)bC)JaB_+DT}oI=TrMzN~?WV4|9B#uGB1lY>8bo{C3xFf-EJlpgJ1H5r`VlCMfo;gFt^ -_#2aaxs40PWWuCYF%V}XmJMq9NF~S(6>VrDzE6!}fgn+^931f-q7WME-st9^m#6W`@24mKIQ#OOC*D4 -eK?*40;Oa`vhjO`Oax^+5pBx=iYnC13v$skw~u&S1Wpm_qKN~if%OjE7p7 ->K6ipdmY$lICV&NDoOVGw{MOS&J-HMC1f4;b66R%t0?Wj^zm>Vo5@kj#~-x$VkpuihR~i1~Z}BW7a}g -1GU^*bnvgLJiLe?=pdD3|8ycges ->;B+cAg8s<$n*0(S{}TN~Xqv)YDJdk&1jTF6QW+QYm(5MEEu8R1V_CP7X{VH8ZE@o>9+`NuXTbe>iDzFK+R22PMD*j8T -eE&M%G#vYIbzA0tw|JO2p=zP=W -L-JWci6!Hal_!&8y5?hmy;&)mOYZeAGy$82Dd{k7{tG>1s#$>i#n7|v|`#umsb}z7bh3z@y+>ld~$wv -`sHR4GVL4GrdWYkNugSyV+N|#6*FNGmaA5oA!fL)5N>hCMA2-I;?uX#PhPn&c!Qqh5I23w_l3|-+jG( -L{lsKuQqC+ne3b*??S`x5u7caM?fvDMt(c_(_-!{paP8$wo={wrpen;^RCf#tu{MH!g<1nD^>&9c^nY -xPisP|j+0r=QP}dVZBp*47S<~bJ{R>$}s1(phL1I)D0-^>gM)IIMHn69)GSmeHvK{Q^0t?BJ0c^uO&c -`*&Vaiy(a*@9|x!meCOcdJXz#i$r8ITz(-!Op&Pbq -Jt<1hAXt}&Rd4EWfUj+X&}JHm5xc5CC;G72EMwDj?Q6VJZWOtPUfbbFUY-EU{5i2h}*kKSg#Jr4K6Oi -JcH=wSa%mJp%7l=QPLt_Z~?AN1ylK7bc>g7HZ0|KMM`Z?vZi^e*7Sk3R{K=6c)gex>OMZ5xp%Y3{&)|qwF;Vp8CeLCfx=I)nzM$)2PYOeE=M`oP?Ljw9yLG;s~1!aNm -naaz?fL7A<_`3nQNHQVJ@+JT7|96tSTGV_P&qYc)cPs)18gH3i4+Yizb&`!NQo0;usPYOp3iAI3$ZwV -+5xH0zc%%J_*Vd@OtD`Zvd?uT`_3H!USMX1-1G-9<_zim1v#-ek&@mb2g1VD!8{xCDxFpu9@z#c{h(v -pSZpT5l{+nl~Z^bM!n$(>qs{(3}0kHNUCM4^TIc-#Y|dhh^wr!6{-p8-Wc_xv)kBsVLA7%2MN6t4DP! -O1JOJm81K>86X5f{!9HV6(9Yd&Uu%YcL0^M^J@v=o+*gAJ%8%<2NV@sgJ)qb?#z-d)==ZSITB#kMy(r86l;JRfufP*_*Aw%Zz#2e$2{7(%UZ+iSq!@3al3 --ab(EO~f~UUR_KpbLZ}NTH7)8ttP@6a*Liaf4fAF4Jr9IPYt3>JB_SBW9Nj}<)%!rPWHd5zNXY&vzJ-@M6tmWkT%kGkC`_m0( -|ml^9=ur)kyFRJ}0c1mEY|9m0Liz;sh%V{HFCktt8KEimu~dbiVNpWhQ~62-m@?<*TS0KiC}RWI5sHM -<_;7SvUcRvinJ05$Y37NXnB>7@VaU=;8Wb0Nf>_UGaPUaRFc#=Va^(9StVT6Sxi2CT?Y6g63h4!vzWb -~iJRUtq!Wit!D|>1|Zn1?Dxhxt^$!?nPfC`~`9hOuf)Rq4rJRKTzt;rzRS4cc|`7nqWa2c0_l`=7fiX -7*_!ME=^PNjb^3mCU?eY$PC~Z#-T}7bg|w0f}U`*fmD3a40p(IYIjhy=l+~Q`vb8<(sQW5a+`@&j7x5 -a1eBZ|K(yGL^wwy5!hO2xN%mL$L5RM(DAsty{WZ)1E%ntrOgIsfTc62-B0y^$n*ad#85&7VmR6Xu}AIefY{ReLqHUTeM<g#;U3;fP4`I~vmwR%m)jl*&y}O94-TUM!T-$kT -BN%s>sW_GD?DW88eT}*H|8mTc5!DX>T?arcAeicqpMZ`e>G1&;e#Doe150HA8Na%qb`VPPvy(vI#BsT -g4=1_h?Mog_8G>XM-90{&~Rh`vszUTWT@`D5HtE-8L-hIaVLbw(a*QwN-8BLt5{8BOs(faZj43cWRrxBXh8eTlkxGC{Cqr_d1sq_0-TNZAZ -iwFdUoDfYOvcpbBurb<>cby*_Yp9a9g;Wj -nG@kdB^IS@NVQjg!})&YR{CYklY8Comf=gtaVQxj4s^!hop{MbGIjM@Ol_v(K)B^_vJ_3?@G+?S8I;_ -xCFI_-#3^HfTU=Y3h%VoNT4f -7lqosr^ZY^g;J}?5*k~EFp%WPH@aH`|;EPhT!7tg&){xy7pYIX!HnaLa|Md1_0?4BJv(ZM`v*%cLlV< -c^P)h>@6aWAK2ms3fSzF#Ww@#T8000+I001BW003}la4%nWWo~3|axZjpb#rucbZ>HHFJEn8V{daVaC -yBP{Zkve@pt|TJu_46dB&I*nlhf!mJsM;N+8ALy>>F0oAaHFubl5(Ne0|B|9f}!b<%zLplSU>V(YZ4) -$VG)v`4aBD?1lbZiQkKj&!7`8O&Q7olBOKvLo{Fn8+QoFyWM3N`lWB%bDUfe&)TKEZvZsDft^Y9FLAkn8hTcPn -406QNq_N1fubfWVs+;D@y_W-$pgVhhM(+YEQ}tA$jL*N -KV<~b5D*uz`U$cdM1zXe8xVQSC4YBzrZif!L~Kwu3%yh7WS;egq<1jga6y*`R|@yR1q0kqIeXv-X4+_O#_UiJO -wU=)=taKH;!r5#P+7;H(Nor|ZABlzS+{}-&eS_FsWilem3iGcySU6^`YZ>~TEf6oFwo5NEXrfl&=YG` -;o18TH21_>Q#;!qJ7p}~_(LDoBAlhP=7~}kZ|Dw8b9LnMy`n{vi9@$wAx@0j=&P{-5|E~O6pEaw2^P` -L&2%b4vF376ZOV{P#w7;_de7yDI5x1mFF#eI1H*~(V>vh|F{M1oeP%J8fxUSLc8mMmxmlrvFVgUyyW& -E9KBG(|b1+n*#4H0oz*!&jG>JAX6X5)DvHdw;r@{(PQ(>Mq5ov9#Jvu(wXN7leceIw8q@2GPmqPzbgUVSbX)xM5k(sniU<0Hx#n?bm -j23I6~Hic|QB)6;Bo`Ur6(`sWW+EbZx2{qKPkg_h$L9B6C&YAZdxgGcOel^n%PhFMmOhb#WYrp{v(7+!BE -3wXE7*?ojrOU@z&}25zq8<)I93_?3aL$rz$ZR0PLDv{a)dB(KS2Vj9 -ixL@L6uAY`T~QL*zTUfypj8#>w4ecXnQZc!S~!(5RQbI1*!c-n;cXi<#!pRJHO3{$<(vr+x9>#fMK9a -XcO%-Y?$Tb$;@Y-wx?0gJ{skOqS@TOs$^ra#be*ifgu?R*gQ9%)ycm58XjiQ_P?E+|A*VL1yX}jOCXuvQ68~mRg(d9S+3C74TR8C3D`LQM1)wR`A7)@qtK~-rk< -b*a|Da?CnPLe@{8>Y#>uokGURwm>t{pJ@~cS$*!z^F`P;8>%+naVQw(U0`y2)B*v--Gk&E@csBaY}=nbD8|~nM|W5q$nE&6MbDbf+S~1fLGidk#{V`PkYl -M|Gz~VISDgD6=idI{b5rr!jO!u6-CD{~s;lq^ggP!8@Bx1P%xTEuZqxx8fqPnJKy1H5z0$&5OH>QI2D)A>A -!5}Jv0GwRinH_hE+aZd%w*NmV&;3Q5MXsVqIgc@PMny8-sCEsj4pUdl8BUKp~2b$^XY5#Ei)6u{KQ~K -UTzP@m~2)F{EsHl!3S=G^2_3)?rd(s!7+Kg84de@aWK@TM$(KbRRzC|7t -j+|b`3TuQQA}$XHz+;Uyo>D%K}7H`Qwpox3~FAKFR2DxugDWPL^vb&G@~qu2+#!cL -xF%#Ap@1flcIqm47-hkC@Cf$F3X=2@FbHOdoov7xAYEuf$ZMMSCpyyS9^j&>y)g&0xzlr`>nU*L_eIT -qv`ZY$yBd0aPE$`)u4f`(Sk+~c=ucK$JOwRbNIG|gZ-t7zy8wXe~Vt;fqb_mN=5JQhSyfAUyE^PCraV -BF$^}$Dc#=Nwjc~)RQb2gf;hpv8Samt$H(LG_;5Vd2MS09@-N}}5}b{o$n~O-e5QA?|Mqup-~84A_&M -j|I$Y8Z315aHTIf0BpPV_40MDOS38S%^b$od2%(~8%FznN#HzsVHZ9w#a$iWB%u=Oxtbl#U{XhP&O)p -}#9Sd$8GVQv!dC7nSt~-?9IvZuqwbCr{mm}6QV8$A -gpE!W8GEptQrcGUIdG+}soWx8EgLHm)H#$9{OWsy6W4c2Srsi)+cnW}SFr|@iBDOAF30oEPKuGw{y`% -#BA-xPU@YJ`r5ow-q`%Y~MK@zb{&xzARf1hmS-ycRmy&<36R8?$E|lYw-g3UqB57AJ)LY2dA8o;Np$G -DdO)_2H_y| -tL{ANRZiq^EW}Krv-O1P;I$IQ$-jSflF8;R#6-Eb^9YB`G%Lsr*PAM@(bzTVwo+&73M}mh_a2>0I%kP7|xu{@0nwv12jS0hw*k -ads(-4u<{_M13(dmy1Fl=rU&;t+AgnBv_`PiZBZH2A}`X(}4z>DyanMrXU?9N)22)5vwLYuj!Q2Na+r -fHAU~4R}=AZ)W;5XeLvMHty$Wjf^hk$9e%uKBH!RQM-|`d$52I7k -hW5qj?_rgL0nJe6J_D?f4Q?1rR1)geg~#+uHe&@S!C9?ELmFxW2hQA5<%Wn}nuueNJAxcFiuj5#kKk0 -9KE!C}qmjClEKNgW=#Rb6>v~frM)KwU|Yv4fj=dO$K>RqLMx*l5gt~*>TiLzq%fJ3gFJv&+1X)|1 -1Qc|LFq#S3z7xE}cdV@g)Et0MIb<)<}*T?SqVVMdaU#&IM>?;9Q=isN!y45dReKp~Q8!DC79Zf6{d1~ -N@S)Zx7Y*+q`JUYLe-(Ft)Jvh6)3eJ8$Kl=;(dwO~8Q@Be`W5^4hSGbfg+R9`fn5xvPRxlVcxxKOl+( -YA#`l9%xf!VEbiusJIL$8fn%;%*p)ogp{2xtoq>(m+m3O;YwoTA=R@$$&pFN&PlKOoD5_^|ON5jUbkO5x7mz=x4E+h5 -O9Q&TAs@f<{=L%?W2d=aZ^WPtR^*X -QahA(bWl)8jnt9?kebttYRvdNqdWb}yl{b?7bqg80ROdTH?DE$HaMub@ilQ@G*eDa=6MHH7dPVQd8)=Vd$kK!|66e{TFmLoWx8FE6Wmg|ge3!L6zdz7D7E21M@16yD7sA$&aO$9%GIdAB -@l11?KDs;a+~i|E^|_8lAOZW%=@^od6|07nTM@o#L|AwGYK&>eJ61E@JCbsVjtDe4ZHb__m#s;mEGkH ->s<>V>`;;@iUa$9yTn{t_ZTwzYS@y{3$IHTZSCv!gfz$Pc))&5V^KFKURG*)nM5 -mm9_5IwFC&d$O6i0wgjn&(!NENUW{p*#<-zl2wmLpOK1gn+`_uP-;Rmxcp -B*eiuzlO_;bR10RSRbRAYz&`6FY4V&Og8_69EEvICN_SmmFoJjjmo|4q>BZA+ -yQ%-l`jkc%TpY=pYOt`K)*iPKwv`RmN$E6x8c%)X^j4*H{{C?WAL2dq(&5h -!Ps%C`V?5egeOZkhf&n0|v{r*4)s(}v!*mf;3(h%rTb@KAg!T$yM+IHLC+`6fest&hNYQa>(3cy_*>(T$!vHGsC -WIK+2967gKRR3{>#{O{a2Mglx^al2xc6(2Azt+H%J*FYWv|gxVS-)z6*NE61P5NW(@H#ZppJbMHsDn> -^aqcFZ`uN*+v4&lpfUzn9E*=Qm_hR!TRrzQFA2)y(Hwu2BS<^YzaIW+k@|tuz@4b`f(D(!U)+?VmBLQ -;|#)e7Kx7l5I7l9r?>x(gtfGe4`X$MEKs$Ck}Sh!3F$P`b(U*->B|75eEVu(FW#N01QS{z_qY{9z&M> -1g)Z3nD2Y!$*RT_;wQcgJ#Gg(4~_aC{?5Btt%N5 -n^9?t|5v9PX4tXsj(=$ChI)vtUpEPMwg4@8kkP+M%8Jc+Syr{seT#?GB)wi9ul&ryRsSGQ>_aU@gM6+ -Dtt|~l-Ag%RNhUszya#ea|Es~v^1mjUn}Z)K$ZR7-MqhgU~pxK0`sMmg=CWq>Nv+ybCBn9#*<&1dwo`BK3h6Qt8#dG($>b>4k1ceI|gM0)kvib(gd -N+u4gN1%IX-Un30y2gR&yNto@*>#{U6OO9KQH000080LuVbTb)jZkQf920Bi{W03rYY0B~t=FJE?LZe -(wAFLZKsb98fbZ*pZXUvF?_ZgX>NE^v9pR!wi?HW0n*R}7p2v9Xf8>E_S`O;Oi&5CdrzwG;G^rUE66E -Fu!Al2j6-?SJo3q-@9Q-D`c2X_9Z|&6^LI*Ns#LGARtcnTn?YcRNz|D~0_^>#ARxO~b`H2u>hlBIh|X -s3Bzlkep$^r#-`E{9i^?vD?DcN$1CFIt?3XnYY>g_&v4Dj33chX?CjNAo -nYT+@a`9?)J9T>_h<@^=7cUGdM)mV}Cj~xT_L6g@O$b?u*)zPBXi8K-4RNzu*D3q+bY9ad -`tr2Ue_%m9_23~np!J5Ayv>F&aSXv=nqV6bQI)P5#xW^LHHJZdbN?uU$Vv{t?lmxC6vU~%EJ$GrNs+U -F2IRe5Niir*iQqnO)zU58?ciIDr7kFXHg8`(wWF5g)#oPuj`tabDkyn?eJlrCwbHUNH&P~aTdvvDojX -F{vq1c^%mEvr%{9hO@1h72kHns4IrIXXeeN_B_P5eGbdAP~Eg<%pqezE)d -(I`rkhpc?ed?rhdU$m-awTCuRQ<93wtuWN?f@t6s;8GVm#D?5aa!%&wWRh_d3gVbI1YG==2TM;FdDre -owTy9N%dW4QGkZZeFrO~>e-RaBeQ)Mh8I@q*piF6wy2>4kKSMFWvV30dwsSfJmktFzW}Dxkk3K4z2R8 -pG@{%bI+`fAOXJ9>ZU&`NR*uYIFqj@IG*_HXXK%}iX%(>DfNu@jcZe0IW+o;>eN_o*_5{ee^8WYYS7= -%#;F?hl1ii$9)t$hEt;qXPAesN#IFE@&Rt~L|1EsyUd(}E;@4S2A`T|8%U7QL1Q>cdBb?T*lsvk~t*X -kCKWF?7o36}X|x}i7KlRMO?$aYdnZ7j7e3Gfftu}^(!zrvtPie2_I^}gh!w4K^QMzVnvGuCRfqk+ab3 -OmVj*zus6PFr4?hzoC_i-)^vf)~#G97<}L<5ef!a{k-HVmY6MU6ylc`$iy{rQxEG=8OSuToW`*>)=07 -O9KQH000080LuVbTWP>t&&2@%0BQsP04M+e0B~t=FJE?LZe(wAFLZKsb98fbZ*pZXUvqP8Ut@1>b97; -DbaO6nd6iSkZsRZvyz487_Rs`K`~d;7mtOX;FQAKFib9ZWI?=9784{h=e_zUWYqdyrQNstD8ge*8%C? -;k3^6)o=g%R&e)7;+Iw*AwU%jir6R-z9BKTkrmW+eSnHV8@ctWTQK^4$2f@P>YMH?B3-5>?dH^ -kJ6I70l5!^HIxHz=p#q=0W(tyt;Zr-uhvR0kW8trd#c5*5x1KMUi|Em3gaH+A0OX1Psa6_n8y(%Pp2M -zbAkU7=`^B+6P1$Hki{$B`&f`^5%znys=CqVtDORQ{$Eyxer~Y;%FdmT^dFv3p^NwlN?)eU{shU{9+j -%G0Uc4wN2jYf^@PDTgO+Q7tC1o*+QQK+x36B1(}ua%EX< -`rmK%MedVSf)oc-1saMwGdr)Hota%Z+3xZJdvTs&^o-Nwp_8yTWtoaceB343V`p9DTQ7z$sPy>A`y{) -bOg!%efkBsE@|fiXazOe;kr$`Hb~_$?=Ckp1=`Z^9OMr`gjQ0g%U*j;|L`Aq0-}Y&qJ&G?+d9td94hm -#cV*XU12%~VHC2_t&;lqBtMg<$%_)g%@r6C5!`C=v#>?}Xb{c!ZgAH5yJgx)Yab{N=H*Au -GkMtgvAJPgYXFz$@&;>K{cJun~}d*j@}H%Q-693|MRZ6P;4Yv&KAnM#fSNFzL>oWC;shl^r0T=bmBBB -OOIfMqZwHD=ViCDZ10i^7OoSNuF4S5m{!TBd$u}^P=WmntObHDLHTEWrsiH>0^2=QU}yyp!W$HzSOc< -sB>9L5Qe*RSAN|CCOg{d$IsGCR7U@|Y$CeHY#h!L?3z0;l0A*DzwvTC*#AA93e!K3oh>R^USbp};!#0{ -wT7N{m(t6Ow#(=y6%@qbx@AJWb+b8`|n{U!)6!iv&GI=`z`(e2=Z{tVbIZ|FS?&c^u&+&u$7-ZZcGs1 -PL%mn9@kQfh}uejYmQKYoq|sxkO-c-xtksal3^ms*2?pb# -`=d7UN#AuoZj&O+j@hR)knbW_(fMhr1mt9ZL69ltcNe|BO)vk)IlY^TW4YGk( -kJwL1kL)a(+B@8%PqfqVnAe2KPWKSf&z->wi}4{Mu^B95vyI{zdo8#l%Eo2dK)!cJ`sS*f})u!HpvgD_|HB{6MUq|^-{G?&;bAcF`;4$`+_o(BL|9wU~Tvyn1t|RF%0GSJ -)4Y2AHLPLzd8)}c;n1S){wH2jo2RHNgxVP0g3iBqUtgukeRN0WA<&^E$8kDf@Be~%*4KX>n4?$b8Mb9 -K$CP|wv{%@M+3ZdETwB<47M#$HP5?zeRU|k~LrlO%9q~9~-RWrwM*RI|eESWww1G4Za>^L -;(M6tPTQy?<()f20t~vUTFMs1HYS)Om2`Xb>FM)b^qg$>TbkppNHzpz5JWg+r#m -{vDY`}&#e~7Te6>#<(@7;&>*546m;t&+{SJk -Ao$n^{OKkQ6a4yW(#RxS0%Z16ip*y3xl -|hV$=XFdh+$!X5)q%YagmiTq?C0+Y|)hxgOB)7g7qeyAf|G%zk^o(_R=4exaW-uneaTV4ihXs=F~j=$ -9~e`?V%n5Ys{Q}2z{I!rT5AJjPm>nPLxWH%_w5$!0i2}{JOafnW5h1OoUiLzCSLLe9-lHYZq?AWLH(tF7 -BwC{zNq}|>3qT;cnk3H|@Kgqu=K}yzMmSd0ORGa7vZ<=ESSddIbX?paEC@E3t%>s645v?&QycmW^m;D -yuNmOD!l&!p*yx2zg4r-t15miJfjq2c|+oJ;of^;pT?14nI*+Neuq(p<}rhfLAWC*k&G&g92Z6!8P#H8+lSqx(_iDx^M(L_!cuNqLn6vp|<2|r@-jN^uCq)HJrVqkkhj2uf;` -IH9IE&^nv{QN-D!ZNo?)dbs0%rsifwQ<&@9*ao>(xZCfAS~L*Lai~<0C0IN3;FFH%itP`C_-1}wqYl?5j)Q(-akN!)Y#HXvM93dU^oX^=`M8F$dPqfrFn_0QKTNx02Rq3KTy_54Q(-;ln*te;U+2 -{Z<3v{Yx7nd*pV%I6>`)B^Y|}SS?B1T;kS&Bv8_Gcd_)s#8_W)HTN9u?MXnMUA0+f3P_RT`vGH?cuo< -3piUKCzrZM+cJAymYmfig_CyOePx-8;A>Uj}vkz-I#l68w|w(o2&Pe|$J%RDR<3GiVPr+~g(BqAdanx -PZiTnp5kLe}GAn0+L6T@SUBVv32R;4U(vP>?Q@gV?@1uNwPgi;8yve^oH|sIRM;Fhh>c8HH9T2d8=fV -eTM-LlEElDQlg@5M4&Q`bAb6G0=ddPsuF&BHSiWRS&TA|F13N2tV;50xuwvSh%}rU%Qw&Fgfn0X0~L) -UKKqfR)r6dU>c`ba-*3j`ICYCw3EweAXj1Z}uDxT^_s;vw0^)(Cz_>WdnKz^!ULs9HS*1LJHj?b+miB -z;TQDw(e*%{jn^ZZ}ApllV)yIp=PxUP1)0XCR6==8geFM&HczyOPT`m)`X5Ji2QP!=;dnQ0b56lE+3q -gLvEzfK~gcEnbs9IALttopI5P1_)BV5U9KL1ofXin6WYN@(DRev=PeX5@7Yp)`wN<~-oE-Ioe^(Tl^@ -rJt>5t1!)a){$0ekKfSUM}INpeWn?f{q~_A#xd4aC&a|fqw6Xwjtf%wi`Gkc8l!S`>S`+>pI$)8XuHm -$b`N!@7%bd2lrf5gTESus5a!Py)y39^sImb^`eZJK$_W(#*QxloL$E2%7Z=0{b+6lZ|jR78fXOzkyYA -uM4M_&J6V6BVU>5JMvIBuvsEJsBhIiC9oQSb6)sA@V*ZIgy!qPcMTc7AKwNyW6FaX}r9^ACu5OZ|#9o -9k+U{_@nehFRY>KewNBb04duMqBtHa((4m7+X&vB`d>L_G}Zb`eR1V -i#UHQLKz&NeE%&>Yip70Z558{tZN)&sYALeDYU;l?b$i%Jja_M@3g1c5$E($ag_gVw<=ko$~=JUVWoC -*(aKlzGt#qN0Wm7AHoo^E7VxG>@=W{5fIW+0DzccEk#eD#@N -lO%~bLnwRZ11?B<=<3IOMntz{<2_8Pe>BUPRFmT=r!*KMQ%_|j;7@nza?s6#G|w%3l-JhFgVpR0DlOS -460ToQeX(q#3Ly!zBuYw}99cEw9l(-@`aZYN*Dk}EB(D7rOewtUlgP!6)8coRcX7t7*M0?8c1_ps9L%f<>kXZNy*O#SWSRn9yJefhLVc= -0#PSWcSO}H%#R>~utO#?9aE&P7Cm*9tg5^iP{a&!Cdol~2gBkcWZfZhJ4srR;f}T(X(9~tTQA4;PNrr -g=Za?9ur0c1ov^-k2a{?C;HOimZs*Z#SD)lIA?V5+c -0?!+$Q-)d)+qlT1b{{j44&!w-y$g=EsG)EWXJsPmvG7UCT9OJ&|?o2 -d4C3wF^RRV*Ep)Vou3qt&q-xYiR@MM*CpT5Z8^Q#w!#o|Ms^&Ql=f5nMUI+fF`IV_Wa5^ew^Mnqo@ov -JNnw`;vPo4I*;X!Pu{>Dw4m2F?AD7UHnR^>|r&uUS%y|bAl8JvjS(c-$og13DGKgNRhW?{c_rd5=~!^ -P*Jl@Mx6tbA1RXMV*BM{_u~B3>s|vcS=PjKYP;SHM-6v`KI?<(#h_E`>{?^y*trW_Qh}weX3lFsTAHS -feNggw;ROg0!yd>69-{;{doZ7DFS56@VnM^q$!*sIN$~-BPult502wTVm>c+z;0;07@M`PR+a#NOhgi -9r;O|NZ@lg?PUJ|!aoul$T)yDsT4b3bg6iD!c^_jFbd4j@YI%DTI$+KTP|L%SKO_&p|D)nrgbf>K=fvLB -2^o~T+HA1Z*U|eY#p3<OVaA|NaUv_0~WN&gWbaHibbaQlXa%C@Yc`kH$ -aAjlz08mQ<1QY-O00;of09jj_7^gIp0ssJ81ONaZ0001RX>c!Jc4cm4Z*nhna%^mAVlyvHNkc_WQ$;R -xcywJ;TW{Jh6n^)wIOzf`x`&hVBj5ivC&<4NP!&tP;Yxt4pj5@HT!;lc@Cwfl32dZZ6>AMRp?I#hP%MGxTl -gXKMFQeysYF#lp_t6dR0^!ge4efsGJgSloRAmz-eo2=g6aZ@@D3$KMPQjw^A#TXOr}zA6SkB(C(LD`f -J4cZmh*MW6_jgL7L`Ey1;XYsUn*n~naK47S+NG<6%SCYc$yLw<7-r}NFRPLOVaA| -NaUv_0~WN&gWb#iQMX<{=kUtei%X>?y-E^v93R_kxuHW2@=zv7^xh+TL}*A2s9z`(pFZGyJ0k+|E3VG -wBPY;&PRm89%wfBlXxOU`3O{UOmE?|%2-J(#9xvh2)G)^N32EkW?LVyc4!H#;ypg{=HRvkgKa>lzEgN -wASp@OnFe4Hrn1HlW+4k;+V<)u>>jIM$`;Ccx?vDyc9@$<9W}3i5o@npR<+1Fx(Ou(g(=HJE!kNnXG$ -8YpqYT46xQ?Ut;;j<5uQ8wqSLc?s!49+?~2)@2F@rfQPaSrV{=awOxqcO@c+!$m8^mB&lGKo*bizQD% -D9yGIjk4oEI1c5v<@;pfN`LgMXayG|rk0(nynK<6Oz*7SClK>FlF+MR8` -3g3-Z={#cH;tGzArF;*)1hV5#JNccdCiJv6CMHxxS#qqqsYAkrExNbYNky=N7>!+b{-M5#FyGgM(g=* -AMmQ^PBi?Ga?<$)N-}7EiZm@*O3^7R_ungT?37C9FQf&E<0b_QU-B<;^Y37Z>oy<>KNqeUl7HE(J1*r -bK}Tsmi&AwUEW##D7y=@B-*V)}kYEN+^&`YK@;@jkaP+u8rbrD%ukEoWZhNUT3auaSA(Q8a+Kf-%<;; -Ybu25JZK;PbFC>+=Wl-b^>sqwPZOYD2r0 -;N-7W7Fs&058a;Vz4*M~3Ec;8(g2lhQ#lO$M+oS}P*eEQR4byh=feS$y0ELy@6EE#Scnu7EXVbCIuiP -Bd_9nZ}=&G5Y20+p-(=HFwANRM^3i%2MuO#2lG5j{HsU*uV0S)o8{k=?PvSwSZUrbiJ~?;~XH~a4J|cLO8b1efoG7gmBU)CPF+Kw -C<5LXy742?9j+_ofe7ER*__7nQKFhs&YFOd!Q%gP&}e0U~7SStgFa>b|6K8*tC%z&WXr7okPT>iwg5P -BmIxg$x)uytfKsnGYTNg?$Kb5kH(~y&?bRL0gikbM9hoJ-@m*c`$n<5%KpDX(YW2B;A_jNa5@WP`=7z -Bq19Mj^*hPz5Ik}QUo>u7VHF)tiR;!|NgPcV(W+WImAJsz#@N3eo}28%kr%MC-BKZyX$duAsjbO%-$; -4RYi{y9YdU*E8FJ_`KeSF1HH8Ui6gh3KzB>tkI3;yv=@!3j=2{Yfq8U=$QDJ$1e0)^F|8o;gFhF>=Nb5w;|lYe8Kj>iKZhxHK5oE($Mj~2(HogBH_8m05Hv -{Mg!(LD7E@$yr6Nkt2%@hf-3Q8=;NDa$f=OLMV?w#ld@#p!#4B<3_+LoeBv6@{+hRn=&yw2=}(+=J-Rr>X&tu+dYK4?VK4)>2 -jf~l~eIJmCdxu(}gUD&0GmEU-@Gq~x(s3l+~4y@C#l{Ei1u_`UHg@}6wKHm*ZWaQoA -k-`urf!MztAF+n)t`ySzo00b=*&3z4^T@31QY-O00;of09jkX!apXo4FCXaEC2u_0001RX>c!Jc4cm4 -Z*nhna%^mAVlyvaV{dG1Wn*+{Z*FrgaCxm-ZExE~68`RAv5*mnY9%60?@*ip2barB8#HNyBt>!X8T5) -=*;`X2LoRKphx_mM%)XGjBrQ9=(}zSBXJ_YaXP%kmrY^TENj9x&Ymp>OZmY6ZjIWzAZQ-$zHP-k% -;Zd^WToE%RKYN|r?v^L6T0z2{XWi>oDj-zvV&h5h?T>{3xF_EuB9uIsX%=JgPtPzZHh3!dS3x9sD)GQ -FAFjIm?(A?IpS*4tH6iBxW6ddhw-EsuurTP}0j>PYiXbKW$DpV^Q;nb^QmqEMnvlBf~+X3;;SPlZKCL --L(emU_t`TB7!qFyC%@-r}pOc`DXCy_w{8bf1lWsWVZF?3GNFC5nyaKGE9(o5<`(=qPGSRbeZBB|4f$ -?gmd{{N$cWXnP~}%~F5UB15Vq;dzc|Ib-Luk0OOl^8K%k$X=F(;<6C+Y{_One*XNy>`X$=E}%^EbDQT -_Rjz-BI?)0@SyN9ji9+{V`^;uD{r`u$yp^3vxRY -5y)@3oK9JVv)s+y~G!aPiw|I2^U@oWeGZ;=r5h`Ugiz_90X~9C}iOs3TvFRlc&o-y|bUoLU?C_sN=*^ -gSY(_KKB`b#{ctZz^M<9GxLitJ3=#{&XEZ~r!xYDZLbR~K?d*CK~Q)?A0x6?2sweHL@ppwsEo*^w57< -+^s^{33IhF=yoNK(a)=UljP9hm?nMbRg?9MaP_4KstB9ltXa_{+Ah^I!@d=E -{nQ6BkOjB$BPXx>oF(-wN_^r_6wz&MM4h6{fR__GWix*WyYR1@X4=(d!h(zy{2UXrRl2JrB=KXm0u>+ -AI&*53xjYMM&1_>U9mqm$snFiAY4uu7yKKD68? -T2?>2_5NMi~2@Bq{@=x#6>+M85!*05s!`FU*&ONM;k -%1uwsl4)5BpwnBmXFgj2U$A1do`REW6_11JkAt~I%*%nP0NYg$N2XCp1ZGkdw^hbfk -HKEp-+^idu#jBZsIsQ)N`_;`a*bX1rq_i*bWI8cD6t*ThGe0~x#UGEv~M-Hk8fVG@1K45G-mw?tQ>@; -aNj^7Wf*qsC|nR3uph}PcxreJch%OgQvceJI+vGTt%l~!9!l@@k-w|)Yh#Q_J9FjKVd6{6^T{2IUM0a -fvb2NJooVMVQ6*#c24Z@g4KJTN`iP=Yuro`MTND5*E5sf6gpmu8$<8QIrV#92lRnP_OrtuU~7ZCK@HqpX0>pFeje3| -VV%JQqU}(lnZX5^D9bEN`TkFWG!kmS<~T&+j3Imse>=yykzLeQ9^QG4=VJ0v}~zh5B-(uIGet=HJi9b -e~>-b{sGCQzycbb?nEhURv(&EO)cg^6&bC3H)z4Xt=**-+W^dfDioJ3cX-xx~0ySi@4)8!0Q3J_qq^| -4k_O`#v&7vV1TSQct!zBXF0Ou%Yhpg_tXK%Q0)^!a&Bc%MAkQ#A4%b-oJqEOonyUZ -fJ*=!W-fBuw1Li -(I@ulNbBCFIMFF2Q?&M%yS%?CQO_X;{cSi<8@^#zfD46cauLv?5MAO0&KA -%q|Vn)qLTUp%F+>sUmr;EHN>xS*_cNYW2{%*frDXH-M2gUM|6)EV8z)lXY>Dvwzf{g$KR>9Tq#n -#aN$=ui-%O-p&5O{Vpc?$gW<>fPT4&~Aj)?ybIv9ke+#$&Tp6xrv&F^@z(-G&FANz`ORA?L{{3dHuw2Q8dq0d1SO6YFhBg_H4L78vL(!$GWs&m4K^WHLZ2N&m)2Lv`qbe4xSUhV9jqa{dmrkR8u$FR7 -l>?z-FahNDuF@A$yqBQff3K36n&C%%4hz?1z8h8`FuG>0+_;n7fz3(B+cBk9l09c91I(glSf8QK#NLd#BdH+fzDXgMrc*WgYArG}5uL9-Ju;F8X>2R}HkF*}tZhFsQ5%7&@Ms<%;$e+?pvoQxVo^-|iyy&2#^U>4 -0M+2UZx?E8;pmlV#Xhz!Q{6Q&15*hx}^s1> -WjKHCs|*{o(az8ZrI&`o$~8(^LbxD|$df!oQu+P3R|c-Oo<`8vkV)WLk+5CsEQD`18OZGdG%fU8E)LG -+RzTe+S`PekJaGUZ#3(dEawN-1n~4W1->GSMmxL8vgo@uG-&2RKoV&!;9d2S_kNH>$D}NuWU5vGZdpH -mS{}C9J7D$lnyIRk>ViN}TWB^7gR12ROJWFzGdnK20O(BEA3Vkz2=RJoh`V-A43 -2#*mL`{t)w1w?%e}@}!BLG>aK~!lJ&z$JwORk|ayhpO0I2zrF0IT+=Q0;o@Q@Kc*14OgXx=4olfBO_7 -OkqGlYkS0yF4LYUz$LgQB7y|-2Fb+X~^defkKcJXp}BzD2RB^KU=7;G^#b()uj2q!3u@L@c#;V3LMRl -iG*3C4NLTrLQ6xko`^j|sg^KritWY2NB=lffb~)ueb5)B(?9w85XR367aXG%h8N-4@IVmZ43(z6lUQBsMKYX|4^>G)U%)=_!`R8367s -E$J5yct__loa5peig3zNY3gV3_Zn2_7RU}EIL(zR_~gZhx909@N6ia4MBmrgZ9QIMgUl`Udk``EOt;c -qqf?8Oc}uV6cP5bg_am%)9wQ^EdH-NXFV`t^zvl02rK7t?x3F9{00w!ceJqPsWJ%to>}aMv`7br=&!$ -&yFGA!& -w#IZfu2us{3TU9$`6m~lGK`HxETF9eZxZ^uemtuW+OVaA|NaUv_0~WN&gWb#iQMX<{=kUv_13b7^mGE^v8MRtU++ELK -R%%t=)M(gj8NC8^0JsVNF&nTZO{{w}Wm;eM_`!Cdk2WvNBQnfZC~@d~yIN``tyW_m^jN?ZU?O9KQH00 -0080LuVbTUT?1zE2kb09sD~03iSX0B~t=FJE?LZe(wAFLiQkY-wUMFJo_RZe?S1X>V>WaCz-K{de0ol -E3S(K$XjxYGvkg_fqZY?W(rY_|>(uw$i-4Y?h`-NMcQqEJ0e)_PT%j&I|w&04XQw>w9-URPBo?5SSSZ -<{N-kWxi!Wu&S%Nxu{R`(U8jYUu*_h -97_1Pne5aA}vdl{s3U*D+bNt%Z4++TYs_nUu4FlQPFT&_HTyvrOi{yjz#)l?aZl^ACGHCMr2FI-s^dh%|`utl}SFhE -=(rAdr)MTO=uu*)FLz?E3or&jSV@J!6p$==(RJ;P8@I$)?!u>#qY&L^1D-#!%as&8n(Qmay{leMGokr -(w0q%k8l!c$BP?Xu>XX#OV_b01BWl@{CKIQ5p(iVfOG#!-LU;y5XmQo&z~4Y6D#RY6EO~HUT!l3d6t0 -c`;@?epGT6#xckat2T*XRSthS%fc-eun-(%geAbaFk?%u_`qXHVY_mg5l)Nd#RWj428W!d#>AYV-bdz -TO-ez%`0b|lO_|^F>~&IbJvFDLJ2i7&6Kl!rk@w9^3NDCcno3>>Q`}$iFy`fH5>;B>nFkZA@-m6SbPU -D;wIyar#_pbuG_X5f3RDpVD9{Z-pn7mkPlL=emseQ^@<(W#r4@4Jtu#sM)m~J5i= -dk3k)fs<@%49SuY*_f+3EahOkZBlzB!+(mg_N#Tw6gpj+$(f4&3Br6|4XeFN-qCDoKVi0CkzB$?}<2f -B7yEK~jLclw1he0#A8*$AbuAmHa2C8PO_Ujx3j%x}1f#(KAq7h?Yte=S=qdYoh*onrVJ{w!89fX5^8ER;Uj~B*?CpkUB>g# -P1DSr*kGeolB-Q0EOlyu-It1xNC$xf@uj@1{fjrGLsr0{_FDgOkgP#q8|jnjWq$&x -4a!v#a0=hn3uqq%rr(;k$nvfAQcy9{%UimtXn+2yXuOUyqN!4v+siJN}>dLx3}oifVf!&+$%UmXmlZt -sM@BjVp*5=+}gROm2|VzVubT4dI4?2VmfR?vfao4t58GuMAPDMOD@jC^CUuf&&DH9*bj~L0BbAU}Uvt -X}ISlu&vs0SiC$1V*o&6mhl}tN!ufQX?kX0_z-89vRl62MO2%MBvYx;ooGNI@p7-M69QyKJ_4!Mpg?Wn8#132(X@-*V;%{uc$avDK1c=sybaey_p*-Z5Vn@?coz+z*nhImddu4N`s -BYA&QUk8A?7l0!g6VH#QynuP%q=z`V=!fPSPmXOW4d5O=#ztY -M`{vlV@%rH(76-G;f=Fk8RqYC7bq8Gz&gi)Cbq2euBJXjM*fFn52iTVv({;3A2V&wN~XvwR(%%l|w;! -IEq=a*>g!GQSM9RG!y1*vma`^Pf%I$#8@r4i7@UTc8tNY6d|jBvJT`_KMd2=^zh7x16j2-E426H+4h(K^DLugA=Nd(zbYZjmbbr)JfqjFha?Iy7q2pz{x9RRmJ|8cds~W~z(i(}!X-?0|TvCT+}D^&gEtX=Za>s^@)^>(<>4%# -96ShE&D2Y+#6m0%H*;FvlU#%ny;KtQn3lv=GAM8greXssU2!D&qt8ZJtAp!e&Lm<{8*~Ojoo?iF2?7= -x_lPykQCmNr;OV;#o9G$|erYu6FoEWECYZAW&@$gsc`Clr=K1M7g6)ovDrNRd~P_m#3E#_A>dvWDiwB -##vzu6Y~N+MV=YBfc-kXun1{zI4Hgk%iB)Q6iTL*yrwL$3oWdx92A+Onm%s1qCmK7r1V+R(O6X{3br{ -)ylep%b){UP;BckQ*9Z~2W(F1a-2u#B*EP-c@*lT0z%8SXtZJ{vcx6Bi2SeK^QHq -VYLnoYC`ONd)AWPMmNZ$*(4IN21OZUs&K#z-#;(d4fo-YUG&=p*XxCNiJ5)Ru^BXv0DclV6}Kx%K;mL -gmkp)_94)XdesB-W^!fsvMCszhl#>xmiHr&>d}jxWMt#Go^b;YmIY?wgK{Zo(bH@1p++T*@svL#*aos -Dj@FcR(!S$=vpZ47MPQxz)REx`cA=)9*U#h)gB&=+0oGd5u#Ra8x|fA3+|`F`kkPyEw4l4Xfzu4$^$O -oJ*&RV2yg!~w7mxalEKKDuShk{@H)$2X -m#(*%I& -ZUM-#dG&o$U`-=>k1qG(+ZoWsUxgGw5hXO5cOXrw5Ac7rC@C{a13``SF@vhy9HrWs2bohWkeYkR-TFY -Wd5#*VmyoTSeA)nY2r|{IitFFFBP4Ig&f+HCD|I>^1UF3UU~06eM#RNi*lP3K}fqhK||$~4Q11@ZW^| -u38y_?rENos&<^&bETHu%mQ-2m)ko8`uW{U>pEgeo+>a`OL1m96Xnln#>f^(L-r_@kYu6%eUCVnvsUT -@~BTNUQHQ1g9q)VAt0X^-~eC~7>8pK?K02`@9b?5|xAl!;@&o+M~R -Q~m8A-OlKtJRC)s52nt8R|1xf)5^ilNd&$cTXqpJG~(rAg4Md0>ZKk_nv~F5skVy^r)|3*zoV=-?MI< -VRRw1zw(r0i`DtI%vWj=q@9-D5ZA+bR`7k_Gq&jekhK*gYU1A>q7|TXasNuK%dr&*0nh%Wr1q!Rh?v?9KTiSe$)7zkIVWfV8I!UGTI8( -H!z|-eN*UUq%yStKr@)CRuetmW(y+^3S-T+NiiS`%sPMXwKo5mwL22G(1F`rJuVnG&>a>{nb&^mM`)E -WpftO?q)NIVjQc*v>aDjs&qA+jDA~=lGP~o#Q)lVtfDq5QtuMlXu#680gbFhQ}9R4Ugu1mkHvRB^YD1 -fnGoREA3XcRAG`IFyvn0IrI3iUvK4cP)JU@qiTP(raiiy-F5TJ+IL%6h4@hNK!TX`29Xa#{Oo{R8K=i -#ZYBge$*J&(sz$GqJn$n4@2>gP1hRjL_?Of0D4HW(Rr4kJ41{oSe+B7Qxm0uW!z-=BN4y%? -|ueDsls$^hp`o`V%G7@~jxf8dA`3A4|c;>Q(|0MTZRb0VI9g%33Yjx-nH#`!_V`U;t(bM^pE2NBRa4* -1&F(>VoQ=j`W-z@!!m-m{^}fjn6+6@D+0gPro1aLQwA5 -{B}Pa>~aW{bm8XLkJgDF;4wn;Fn~H!6WL`~x3GR+Y1y*mYOADmu1$ns2L`6tf2>Ha`}IT{U3or8xnc+ -nr*j{kCR}YRbmLx=_C;A6q)S$wfZY-bJeq$5h|b!>?XbM%1}L*L(%dhQN%oRMQvZLksnCMF1W_oHxIJ -Cc9#z1ShTK6{-jhxhp#0(CN1JR{9;^+aL$MMjL9V(sX3^<-Po1wDT6D)70w@9upMrd~a!)1uRUSivHD -dqGjI+xoX~tli3Vm_M$A_Lx)ZWB#D38<2W7{x6NQ@YgNE7gxf{Gp0N0ZmIU1`nRB%!P{HNcyy9P`fXFv2v?b7icU5)F?W6-83SZ-fbXSEZC*CV -$*h=D+81Mu`n+maZ2Ezb!kgWCg53Vg2R&Q3ZE`j3xohAxpjAT;!od%dU#{nh|Zn4xLwR3qBk& -bP@2oWDEBCNcsQC}UN^@G)dFKx1&j$M*z4quXXeJh+q}HRU<~9NsCFP2uEqKF-4om_pa=g6W-ZmG%y%qgr@Sb+w5!T48)NmsGCHDm -SaCi@mR&=o4eCLmZv+#>UX5`op)FY4pwkv{gn8D_Lz*fUvDKW1zZiOgb35fLo(V7{pXGyx7kr&pi9F& -;pdV?bwrB5gZn5GkWRYE$IYiRQ8$%AOt`X2V1brs&?y3J6{%bBm%A)NtMb2(%JeL(DVS3FpLE{8ulGwvRU2ZZMyPw(5YxQ+fAYDCF#EJKk6ijTl1>myrhE4GReXcOcKN=Q} -oLEu-QFHo1{@c(VoAJtL$+;h^JVKsSQ4p7O(1!c$I!8l8lf+}WSY0~+?p%Un -kR55C^XRNa#*nv+M5)?h^HCEm4sq(rXyn_y{FJbM21f8B=#sgFTVl2rJkp-wFH+I -JIe{PY+h{M>T^*&1!NIG1ug6qa$n{sEj!xN$J)w+I<3}l -F?s)H4jvyx*Bx@wZ|nd)X308eU}wfxmB)3VgcpUf`3=?FTPw%Ij{XRGFFP)lprULcbO -de@wqKbet!$eX+lEbD%Ztp28u-Njg-=6L$k9-yxU+!t%peEZ{VDP49H+8=&8Oq&|d2VHV+lG^!+3@0& -gB0D!?Hq*VncfH~J{UQ53QMwy6gP|oKu?vN9S*^Ka(sNynOsHhB0m>?T%Iis1*;5uPeq9dr|0kdMcp} -p)MNUxHzopQLk~)#mA*2UGrQhQox%2#c(DQrHL!aL;=s6VC^xca-r1AGgG};l$->w=AC{F1mMaDZkZX -+%ID?gWY$?8H>;Hb9FMF2chgUB6-BQ=t&e^@|rOt;@CegeLy$v;CK*qpuAKOX^Itjiq=hP)bp*9@;`7 -u1C=8OTeL@+vB@{|o9(*ZvA)4$a=7(dMhmf%Uuhs@5U53!2}e@j74(PF!OUOSDn(Z=Qm>R6poZ+7 -X}$ukFQ?Cjj9GWmvi+otDAB0Z>Z=1QY-O00;of09jk$+~6CSD*yo4od5tO0001RX>c!Jc4cm4Z*nhna -%^mAVlyveZ*Fd7V{~b6Zg6jJY%Xwl?S1=m+t!uv@A@mS)ObW@Daf+ZCe^Icacw2htRMO;xt>lo9|;kc -6ewdjhvYV|WeUR_=MsIpu&uZl${4HtD)aXZ~?yqCb|4_T`&FJIFDrhKKTZ&h_sY?O2HRliZQ2d -&g>n(^yRakO+i2Rx>(l#$C(qg2qx -C5Zni8+HQUr}fnE4*{Rw7ms{k`^2#aPLMkD_L*m}EW@WZ{^w&iS{b&H$X4d7dqtqtIp02Yn|pp23jau%}HTeO%P*LZ}slJrzOM86qlNVjr6xY2h(r0>nM$iMC -zlJY3gf@-3^~Ywe@9lkRkTAfo%z;Vy(%v301-LG(A%&%R4PR9=p3QJ~XS2!P-t3iHW_{T~yZZXifW~9 -ExCa9V$|P6Xy+sMAl~^i)H^+Q)azgVMjYfuEC)|9A9HcCr;{k-TF579+_KO?XPjUnJ{7|AKst08sPm+0b?%Oq; -J)-2W(p=+q!RbQ#f~bsv+pr0sQPYv{^pfTmy69r)Kr&3v(3Mi??;9=n4^tmt_mrEWMyRxmtp-P*g=Xn -~hslE~kkgK6HJG048QVaJ)EcY+x_qr^yP~v6x+qM{jFto(SfuJRME;%(K{GO7~C0vpVKs@LW!*A%CCv -PtS;o&IAblNz3{k&@*U9OL>~G>|DXWcI<$CFRqmYKmJVZHwApnDgSk=EBpn2^fdflK+49`1=uZx}$Hh>^l4A*|EA_JUx0{TpzK>3GMgGXTN#2H@hz30uZzr3BY(V{zl)@vllG}mSb<~+FfwPxt1@I)RbJc`d7rsVSbB#QGoWDIvSeA-S%-@Qi=aSAm{KS -EmLRGJTMvD_GuY0nBx2))AO+SzJG7iA}`BxpcBBP|gRAnyg_N3@{LK)Zq%Yy<%F_H7JxIr`2Mz4603& -7fW1992hhNr0N$PjO8Y4ooNCQVg#8ovcvybd7o_&OLmNpr5t}R*4d{Pn4R5=xBdDWI)}-+1QU?9&F+v -`!otFO6ACLlZfWucrinMI$$o@SsXqf2vGMUeb2F~aBY^&%*60M*5>*tQyWO&o6yL~KDMetU4I{MDnuduu@hu`2q8}XVs`uWwb4tVP#s*YYUJE) -p7}sAmTeL4m6@gUuMnDodVQWFg_%oM(VuJn0x|sf2U}DWOckLf(XfC7(BIU)E&}aII4Ay61~r4NbPGK -KEoIA$)Z}6JKK -FQr+MB31hdl%=Zn|^Xf@d7nY;NfPPW@(tT$M#C`y8t|{s6?z#D)&rz}RtORmhX7?eIywx93;uNcIv7B -BxA8?ZIS5a4HNTeRcNzix01_W>@EL&fa~vvJK_%o*vsvM1ewpf+Ioaa*TqjjDLd{XTxyhCLh1xoQik_n)phq7DZ0={Sm+%JKG6xI55TQAhjUKqC5yP}G7eE~P+uw -(tU~I7u@Ne7ke)2ssRAjGg!0b0kTeDn~*T9I->NcJvXh7{8=8IGj_*ZF%iL&Di$vs=B%>4cRFzc0W`i ->Nki94998--eGgDkWKipmlxpd*fiv{ZH7lgTUD^vwoT+BWs4UG?jk%VbZJr*Q+roua#PjNpboP%f(3h -4p8(pPW^sJipCWr8+>-b5%FMJk~S;d;+j$inKbrt3hKyt02fy%O+c`f$bA3O#_f0N%z6ib6eH-S~UWi -Poi&<}*~5=uabBrqR# -(T=+P>qq(4G@YDd(zN(3=pl+8OV9Nv42T?#oGb^t1fwrO^q&|LCw2byPC~?ToZy1ulJ{{u4yZqDlRx(@kAbigmU#(MAqm*EF{7YJR^PLlwaS`nVET!8&K#L@A0HtZ7`tJrd0pq*M4 -^3%w%Z3^INRouAN=kvlr{Tkz3!!RIEu=S^I9!gnpe)_u}wv^CZx8(bWnZJPll#!;T|(GQ_ZkH_mEC$1 -g=%CQuoy0A&Zun(6T`8iS+oPX9r2ezn><@rbSt=QX)2E4{1l29jXVg-!MehlO45V{3M^8Bv0CtCvZSw -WDhzVXyHpT<}#T{fsl;8ed@qPTAfz0t_&di -<^`T3kg#rd2xS?JfYO+sF``&I^L)N%amG8N(HQj9}D7Vfj6(ucmwlBTfuzCDHNSUzf(ykK~K&fmJHF -uBm5FJMj#@<20-d&y#3V}=8gPn8f#zpVb^n`>y@=;jtRY*Fqxxem?D@24EZU0{MO@j!@;R&#G5`$qVb -GY;;YfzVg@|l^8!E78gb3Qv#cNo?H?Phdi$J#~w$ah%Sur3-YepvaHZN9BC#@xYT& -IXvhF0)xWQ(w3LN1OYvpVs&FAsm%DKU@_NOuBJ(5Tv9q}veoXFhaBf-R2$IA;UuzEW_VQlfMblcuo`(y#3@iX9 -OFj-A1RKALcXl*Py0tzZfG9BUc_C-=LY%EgWf%NZYUgTUuB!oVhBv% -M$I`qbRW=oA`aocWp8G8M7BliAi{MnfM*MgaXaHUSaJUTwHVhYy2@} -tUrZU3TG`c*&Wl6@3zJMb9la}Inya@-C#5p9Fiepx?YuE$aE;Oj&FPI)jjd-5--@#WItNA5k;$vSo(tRDgDObgn=%V&8l;ChrMwi|EEWyH(nx9k|F67LwN -F%<&1ysixDYYxUKw%Eg_PeDxKsIgoq}bil9ViQ}rlI9P{G)@WHFCv(Eg!xaQY)^-BN>fZ6bKT>%2T$AfkD%EOabA7-K6kL^{Hzl^o(O5?p<1D*J4@nP_mnt(s(i=%O7u#n{hFYej -h}p4;KFito9-NL~a`91c(mwQONI5@$ccydU2F{a+h)8tT+2Y;h(d(k>KLyQbL -8;Bz<&Ph(UcLMHmbM~X|Jn+jQ*M6o{@vBPt3O?w9i6>=_xA1kv&*w9dbDEmO&^DHgX~ -3OjYvU#CTV1Onyi=-JcbuaAz8kB)zLl;s#t*2A(M7S*9~n;&L}Cc6BvZa*E~s-{xqA>2L+4}lVSl>#` -b%?5#qLk3n7-J(K;HSjDb60K -jv}MOK7>*Xg1=5$s2Hq*!Z-aGW{&Te*7bc_pT=?w$G+) -5U7aTa4GWcW*D@~N9lKO~@kxj5~g1{YSuSh9r?GRV{AGRXG~`5_-h9_qn1lx`X++62z -aQB#?1+pxc!F{X#lB#UUPAwm_|L&WOpDI2k8V5=1LI#7A=$fzE9qTuXB*#f&r8Z_L!W6>s%q8U3Xa^F}Ckvr&)Ae0KM3x7+CZtSDK9r8F9oE-*&?1RnH|J7{LgSU_WZU=lZk(GCc -BpBA&CTGm7(LvN8Ijge3juA5^4^ZIAjy2-kmIGytY-j9-3r(s}J7z{y;ih$ghN*b(wK)_psM%%Yv4VT==vjexXkVRd{61y+)&ag -Js#PMpesIOx1BV{p^O22c<6>_v{LaOd&^Xtp`Em=rP|9gT+FQRJ<44M*=@y*t4i4vEdpN|Qp4?9Pf-I -J)aCm1ZDum6Cuo92ZX+JDGCxTHz|WW2K1}oR$=0%A+#7C8^t)$cI*$Od%6?Z0133vZiR!8pbJI975-3 -8d)@MI*CN;g1lAyI9cJg77N9Fg~{fslmr7GECYB5!eEDkkt^eP?J0<7>=BU9q?zj~1)goxNiYdGna9O -24me2n_>x4EMsa#lKx2hBnFANxj+_MM4xMB-@Y3yylI`f62e*6|gEu#zX_0ae?D7+ptYm9l*hF@mBTO^Kl5m8~u?g0Zj;I8dpO0kQFZ# -XD#y{;?DaO*IWV1q%GO@V=Rl50m3hGSJwFt`{-q19cH`xbl9-%W`PKdonODs|`}C7IfA-m?pI?(XuO? -GQL#w(Kbqem`v}P7ltz_v8j1twF3sM}n#(A8YPLFX<#oG~GVtQCwk7i;QIvb_aCnYzu!W?g(b=>&(|6 -T+@#3aP!ChCthTqXt>yoX>#%I694@G|EO+}P9m^y&V86yFrN?aUy~RncM2>&#c+ei7D|XbSk-v50c+OEqEq-1~J}t|DmB}~PS!b9ugm!&xtddCX!0)Jc;lBrhxxLC@bH^s<=kHu#e -CK?Q%LGys6(vDqA*WZPdr*w^LyyTF;J6<2 -zUt*PlW|xDhc0mJ(;7DsnOdO!VJ3Z5O9r*+SqD8(XKYWbC60|}xfh^g6#uXHR)R6{TyaPw`xmth1#a4)!XWb__Is*J7-(O_etgGK%aWUi446SzYpw%1O)txd7_$AvXAB%q$M$L?+LS6FON> -MuChiO1ZnAA+C0wdyH_c8Da!`Ks(oVD5tSsaRDs)&cT27Tbz+0%%cJ{)C{B;pY@F`y~tSNKOR@ZSm3MsHNZlK~Gk1BX#>AShIsE(eufC;5-cmT^I%+h -f)Vp$tJak4iKhrlYjj?J -o)=OCl%-sC%cWf$IhuwgN0e*D0OAU(O6oA|n7L7kCSzflL}c!3bPs5B!m2>3Gy0di9KQ4qy!UWZZ~Bt -7P_z4N>7jUqc&b>-#K8uCw!3SZNOX4RlPytFZNQI+pKiSnj;cGi~7)FV38#(t4pn$tsR3xhBmd?e)c?Aehi6XO?=vcrKTO=ixv8$7zLF2E;+vP8>t -9%Zfp3vwi?#ZiN`sav9^A2vDD1X!uNtSLyCK#W%CbpSnn>SYOIbQv0n0wWTNtR0ztZkoDZp}mHqQu@j -iw#%@@7FxxY4m7t>!o#9E8&0UX;Z%c`t7~_rLVI$IZ#PU(CVR*A$2fYc=vqwzjh57If*J_>4rEMHpos -Nk|F^$6PJTN6nIt-67#w;A8~ull_mV)bMZG=}B53{+-1Z3IeivC{abN6#wX?UV;Z(EMC9ZRc9C)Dcwe -LW*r8=h0cPZO39U!RYY41?JQE(n?f~hg1aIMAM#55skVE)V%E#JL;d-n1QQwkhlWQ#LBxi>)zh&X -$EAV%1;zR#RM^B1&A0|tmjr+ewhQ@hr|jXy7z|V!BOmfU8^5vwJ(rYT^PwV0SYenI4kuKzX$Gj!f6`c -p6v27a)ChM1P!l&9%j>G#87`|oEMhx=m=d=5XkH76W+45B#ij~{9y=boN7Z{_SWfL+2KQdw7qB8EV(D -t^Qzm01++YY&-Q*@O*i>w{Ajp(OZ+8IPKxpltu*MAeNX$g~aV-*(j1W|XAG66;H+d-DU|$#S+Tj!7Y* -b$j`O+ACT1=-TgDuISeL}918dL}hg7Y5;zW1L4B8G|8hOF3v5L~R^x;l;Qpu<``IC>0z#HH6H3Z;wt# -x6d8p7En=VY|7W9Pl6P+Z(j^_?CVqS}e`R=`;Q=4upPj*Tbnn-cm$Yhr(*`6NT2`Cn&ptTNyVZ4DvQB5u;gFa|S!wlf*0&0| -^fLfa~JjpQRIvNtbmDfEGzLlc -I3#zI~?M}gr`DKGvS?G?ZEKydeA5{stHkiAxZkfV-@^V-C(r_rB*R{`md6R7Fm7shY*PR_X5D3M3=^P=FmO;cYo1$JY!=PqyN4Syl?9ih*u1i8&!ADIr} -3vZpH=sE*P=DYuINCI=tizkB<`?2l*fFVEk-oxT0=<_~A@56}rW08cMHn~i`bbt4tJ}bQqtEWLyqhyr+892|lQE3YVrW=ob`oG!XUw_6zLLiLT71 -FQ6n8Xp;I_@aCfuGTRQ^We{@LhIqQVO$-V(=+Ue+KqhAwAs)od)=SdtiVb;Aye19NU5Gf5MX|UPh1sGL8YEU?a3w6DnwVpU?p0`e8`q`bPyGiO<%8l~@}=qZI8lC%a -&1jIc)NEC*te*NYYt@KkkloId+)`Uz9-aR2hd9v3^k>DJ{AkoyGlkj8RITXcw5_`*Q6HGVWYz@~k!Od -CZ<`Uo}l(-uNbHQBLMPN$#k!R4?6Z=jn@0=)e!a*SSb;WC{EPv!Cv>wv0$>3BTJVApItQDE#=R04s&> -8L_pq5@JE6^Tx%U=~&*iP783@1>2$RFIh!I+Gd9U~!AK5_Q^c+f;|jk0(Dp`}qNlRd?Q&N^QmwIrc3u -wrY$`i^QWMX#zgEVv~*|9cc9_Bxg4NTwRwKar2Y -!v#ov!f_=#WG?)7QT^(lx=^%F>)YO!2Fm^KNl-;77AuRD41Mq-FY(z^W3x)Vtt!vvpzgtw=0HkL8b_Wk{`By1H -)XY=AanisB8$$0cAA|k?$8Z!KE|>`;P3maRcTl|pA8uthsNOqrMo}A(AK2ahQihLm;a$gzn}c*2eAF# -+|pNOrM$O~7>dNGajccNyAi{q)?+UbB)T2Z>LB?ICAjBLpAS(1cCWg%II;yQ2cpis|Ndq2yW``hQ!Gb -J<39dP`ZfHw0~=9W`r+&<7L4|j7cfghgUKe(#M(r5pu>eNA+4&=N`5zx+DQ+d#s*&eSYM0(#^SN|qFA -HtlEh{l@r*q{Ax|C5$9+f`@OZyoe`85uG7F8GGF -K!)ep=q+fqMcAEjP&^WBsJx+AH46vqC(0(k6%BG4+&X4G{kqq?!@zUy6qAHWS7X9iJttr~jg|`!-4N- -W0NB0avFlTKqIm-DE9e;EcKTDU7Es4La)h9gcaSlXyt)ryETI(ktg)pqXPIrW=PHhVR4q^^+H#$giKc -Jcgvu>!dXu`E`m~{~{B1^?nI55uup#KH1=<)Z_SlD8Wl)FEM`?v+5pr_w|(0N(j7uBi@b|Vb2qKVX9O -%oX=&mK=EFwW&?^&HC-@l0fqJ3tlxBl<4&|+|VQ?u${bP^ld7N>AHib|3n^glxlOP7Ft}X(N;IfJ;qq=40cDy97-Tjt42 -MPZ59z6f1#;q{JSR>icb;Py-FdBuzM+76o)w5U_+zF1$@LlKB4)=%n{<-)1nL9CENn`Y&QAa-P;dlQ8 -C07RC*)vGh1t)^EIeI6H-{}xn8{}lX`9s%cgFyT>!cX{iLX9zSw)9uOfx#WLu<7puO%RQA6Z{@`qcKX -l(mw-Of?o(qz794(gs^S!m*31mNDP6nzx-wac0$wO%3yN`UF{Rj#erh>i$_)1zu?wHLg#Z!M1hVk?Mp -Krq{jAWVPS2MKd;N8z2WR7ipfGAU}D4g&BpFfTtVIO7I0|TABnlCJRrk?T+SuT(8vhu63c;xNO=etOD -G-{im+%vDl$~gN8BAPiSGIMJJ3>-I}bT`wj(kXqUlaa)6m2rRNVQLODlIEWf%$Ojg&UWKOMn=3B^Af7 -5&)vGq^Z?rg4x;APNpT@`1#)`vJ=q4X>gH@EuzQCg6D1P=y64kZuZ|>G(kZ{XNCnBrTxZ0+S<7ga8}o -*+H9}Uz{M}-*Geyf=k4=j%6e|>=!!mqQ2umWbeh9v8#CCTyen~LB_3QQ4UL8u%F%IPD#G`ljz;o{MVxVhYBRmpzl&L -|g?@OiQeeLV{f^Y*n8GaeYOiCV^g$eIV{U5VJ!>GHvq|!h)jO<=#yE>$n)UAiykA0t`(|Ih_ojM&3N^ -`&9I~^a#n;yd%zl1C9e3N8I;r(Fj(D(OidRI`}1MwaY*dLhCPKIm9o5%N)@9Sn^^iyFTri@5WQ8XGt7 -`+_Qfrkne`^6{Mv<#H&g7M4AoGQG|eH0lW+D>P5yJ%;3B7*s+OuVcuxm#mGvLlG-L#3e!VR80{b9r!= -a7pO+9z1Ce(p@1yn6`?U#x`muGKIjwsH+uLPS5WOa*Y48V~Rk?In+yqm4j(fBSM6qfdHcOs>OUaC5a> --8P3kA`?!)d3jxj^YHTqUr2myobt3@KT*h`7zQ+Xi|DQem1^01}idfoQKMD+#jA(rE!0mlt-~53pEoL -UOmZYf48ZZM>r(0@+vL+D^jipe(@Up!D?zE8l9zsCGt<1$pb(V!+iceSiBWM!Mgvqi(iHXGWd*eSpK5 -d8{sLS<2ii?s&f_#r+tFdShrq^-xV&x1kO)ikTqpejcKKj5hG2YFG$E*uQ?{gnWiI;$7N|0@Od}p|xs -Z$uvfZ8?96ia1cX$mSi$(8DO8i-I-E!n^@Z(Y-t9b>woLRvhPvrie;!E#H(efQmWZje%lP&(3fRo5$! -*f5dxVB$U8=z`TcU+Lh7vn!0r)eSWcKC&35Q$>>AbGEA%kIy;IgRbW&XMt9vSb2_du;!t -&z6M24d_glnv(7beU_5I}1dFk3zJhsup=r*C`Df=c`++35oWyiBj=!ExrVs4(MNcv^*Jrl4 -XYqvS4{CfF!b~2@7GkZJ%jKWl(glJ6_~)nNlfz>UIMsESQIrt8+)> -@6aWAK2ms3fSzBJrNww<<000>v001EX003}la4%nWWo~3|axZmqY;0*_GcRR$V`Xr3X>V?GE^v9hSzT -`%yAgfYuOJi{kqbo$dS3$JIzXs@k7awQ8F?QieQko&1!Tkb;x#FCce%<#Y>gsqRv?WARy -PKs2_VAhDH=OC>m^&i|QskW2w*^zqiPhTPPm)5Dbw6b~7(v3Du;0()RPAU!(G$_Vd4q|EziP(Twq>?U -*CRU`IDNUHNOjrvZ2paGoj?Qq9UD2FspfYFHku<)rb9Q0J?1D|&QZJj@jYx`EgjEfNZ`fDZMlR#W6p_ -jclWPl+P)#rbf=PM9qKHt6qA8o)XeFi$|6W$;-HLw|!a5IN1?95LR9t!;X$$OXhqWs8UAvUX6vfVQW#kz1*F8qpkZI(=3< -{UTzWh_klJ3#T*gWDI+OMd;c!$5Gh#LMsOZ`R?wRi|)?5$T}TALnuv+R%Zg)1K3K5ryMYbksZ%-p@6e -w^3M3f_1*}Z1G%0u*qb>VK*cjy2d-%;XX^seKq_eqc5j9}qVn%DVAv9bYS^ -{ln?9QV!Jlr0Ga;CoWH2008Gn*lkug1PXgDn}G6t_yn3jN2rt%xRuxzs6Wrn;kmy}5wiZ5%aO_%mlXN!mHOs%b3?6}&B -04rK{i0t070qtaOe{VWXui3*UpRUC#<0BJcb0uiXVk5CHlZ-I*tot!+~9J7Za -sX;pHbBWST;~1OsfX0nItYwoV){`3IbO2V3Oa$C#T?>$dHJ$aqVMWBab3ERG0f>fzBW=;UQtgD1&g%ej1 -$mgX>e?~vxCf%vrvI01mduLI)`J`L`ftRclV&D8CQVsS1=%#{;Il940|)VSD(wz5D{*KRJ?!2VJWY>m -x%hc=bG!U{zgm77NZ89PcgMN~ctO7}^O-3iK@%REhFi@-7U*C~6sDumrZ*wz)y!D#%8WXai9?qL>2tyFp8Ua; -X8H#XRCYF57ikQpLkXE-cYW-{d_)WHxLIm#nIdMxNk6lqY}pB%n|Sf3q8ZkYzkMK5=2?bf2A|55zTM$k4hiZ*|_j}%M=KZ5Fk{G9NT)BV@c?AqMM9bUTpSK8s({Q}GI##m0 -HFiwXSHG5v?_Yc8|A`5?uo1Q+|2|z^HGu*M7OL{{lB03qQEw3+RHQ7Y^h(66-1{B@zkTdip(LtVmFpKk_O;JYHbMM%1@vLoYF%^Qu+J*tGUj!7`iBR$L@lBsi)!<+#e*~mz+-=4as%*=v&Tzfw1F#G^O -&u9e1J4D$vb?4lstf&r1fpB)3iT+GNvjo}vR(oXtLmx@Vj=eHkzKS&FhOf%9(dr8ooIOI65$V>9KnQd -{?c_3y7g4_(H(u_0%pMtX@a`T9vE+JXZ~uJ`rgve2bRi5Pt+4%VJf!C5n?0P1)N)`@1R!q*I1SrehBt -n5|iRKqOiZo<#TwpuZwH*rvn$too!XXrb2DOw^$adw+;dTTZ!t%q`aKDrdi$$AUAIURA_y0wDmp^jbX -6{{&H-OP9`z9-G>KA{7&-p!B5+3Xs==bf%p$94@i!hsMO-ovkr#vLIz>#Mq6K->V;3M{hZv8wQW5~pr --=bmzcRZy}}GPl;e(rE+ktY9rPZC6v0m_Wh=pSPf!V|a&>9J=}5V<9=dD<2X(4+(7fwj6Ch!=56m(^D -3zBd@AT5F2mNx=G>GT{ziZ`*16tdzv^tTdlgnDe%qswgWWePRIt@r#n3W6YV!(gxRE~bLSHa}hc>4tL -Eo6=$@4}GB>i`&Xx1ZnR?UOh-arQS5D|*W|yfWL-DBdscCEZwY-(l?NT#|t2F0|Vv^Ji7f{fJ8mh#pIWhjjHr7^4=N~_O{J`FQ3JR{O0Sr-m70Bd<`fK?_cQ^Ix40v8v<51C7>& -EkT1e66kURw671;cTqy{OS#kQZnxJ5F7+UpFoD8r07w6q(2k^Y`zu`NOXkzK<)hBDI4`+QCfjZ>v3Rm -Y}z(OhCM!)m{EO>4p4ZsGzD}e4N`G^~HH*>ZGqK`6|LiTv@JJ2$fjiJ2#_??JhjWmpS3;y|MX$x36?Q -v0VqbFWnffkN8dxLO=4yI?DaJDZQ9D(Df12><$YeY9dm-ot1BI>Gm?iMEx9j!z0YPQ?Q)kK8f(Yt-p7 -@XiTGPIcVe!5nPah9vufqjJqN`t~&+dW!K#+BMwp}SRP}&Hi%z%alxQm)dk-4(4Aj-l}gty<``NpFA~ -68gT%Je{tHk`0|XQR000O8%K%wh$0$oIAPN8ggCPI_9RL6TaA|NaUv_0~WN&gWb#iQMX<{=kW@%+?WO -FWXdEHriZ`;Tb|6iYC!5|Qslh(F;K?Aia>T~S70C5^5IkbgsP%CmJF{Vf!cNtqv)6d?SeUQtCe955?p -n?I3%bnSI{N}Y@m-&u_;kuGl$-|H&yCN?oVJnfRmE@s*9*xX8-?1d6l}u7$mv+gH%YDJ?g_4g(Yh+6c -7TvK87nZ9?9!8^4%-3X7aUsGe&m_-eh~)k{NqNS0d_rIYt5Uh}SW%=&#AK3ZM|mWm$7HzWEa -qjXv#KDgICZKd=m#8Sta^$g_0c!k|*D%~&lBa4yRHQxXG1CdhKR+DpLm65+ -Ch6PFg+fR~2dnR3b$oif#~MuDmp*bD7vO5)rwneoVsugm{z_HI -3;M|^Xe~aY?(Ptc@c0lqBfBJCs`g!>7;_d6R^NXvCHy_>u{b#kc@@}f*#Ku5*G51Sp_{g)A4P$Z{qy{I_SY{S)gh0|^==h*5&E%Y*@pWl!$eE)9XV^w6>Nj_-vZ+-eil&@ws798X2H@~vWRBzMxn!5WU -`==p!Cf-8X4b^L!-vsIB>YFYl0mv<=P%aW~i^V5{@lVduY(cdjYrKv&z*TgHnWuFp|5h`X>m8sqUnE( -vFD&Yu_)@Vm52?vhqAJgE`_6_%^5#2!SS1X$2x@~}VK^0X&@$rd{06M+P6PPlTUO^!0O -cK62|=s5ENgNdchsAiD(4Pt0JiebKS63MB`D%y<5-w-F@evQ2QlN!lXb-mL)0q_o4O`=x}MZe3dGHk;I! -t+I%4rA&zwz0*1^@rEQ^C;Y8oeYQ6d^DmoJn+oi{Zg7PUz29SZ*oC4KvabuaX4U)&z0hm!#cgo|$;gC -Vw!uCTkiq&@H_6~hpE+dQsaf=yf~D@Z)Tk5Z^~3S&+-nS6sW9W&Z|prAgf6MpHSPFx8_PV6r$X)#v?yIaZQ$p9Rp51xzm7bahK8k|xyY8niPLuiytvaln=*_)SY6+N!R -p`cX#fZ@>i+Pz*v1e2>*R~JN82&2mVmX};bTKx)H*l?vaE6m~if$e~3J|DI4F<$5ScuvMER*ol~H|&n -14Di~=ZR%D^KFINejLkimKsC*)anPYJIZJDL;P24q1<(9uLqOqY><3*As8Q^&_`6~OwX$Ym)@lwQTC! -sse)O^6o1R3_kAXc6!mUpcTdPQylEtak)tu^^}I_xR~z!o -82QjY5m4KbH5|3=-m}!~KyfQ0GRWcc7E$FuRT(IidHyC7v@FsQ|1S4oEnD=g2zwXgk$eztOyxj(XY7a -tnqJraK6bux@pISC@tR5V(??;m6_Q)U}dlCKXCwl{C_yfAHht>(8HfSthZn%iF^<(VDSu#lM;sm1Cd-MJe8T0IewT-RgZ_*Iz3bl* -tTgSR`(yY~i7{q7}b*ChR7Id)GUcdsLwTa5vR)}W6twQ?U|JZSz_*bLh27JIxt1hA-7UVX#(Z5+Q_xS -6q~(RD9GD$kqr7JG@f(}dcO4-tXk<4v4P0hm-^*><{e0001RX>c!Jc4cm4Z*nhna%^mAVlyvhX>4V1Z*z1m -aCwzg!H(ND5WVXw1}YYi3Y$S1AShhyC1?Wpkfca@DVjo}q|usPid0BycYVpXXGl@9EqgZ`0fr@V=FQA -|!(nT52Sw2iZZKFBAi7=~2fVUc4i1a)J4xbtB^0-{5Yu<98?5d0rESuBJfSC?H+zS6`!N~USu0RBcG~ -=eUk9}AonMRwJ+E*14bCQ3JS=t?Els;slH42N(7<)fgh{P$hdxUO*RFq0^M7Ccmn2DpEz~=$Ef$p?RK -v|a%Qp$olcs6-|2P=HuHwdUYlXUz`?aMg62V|Ed5s+^2dy^Hi4L9C(y|34I{}v)QP@qWmq{afhJq -h|JM~C1L0D(0i^eNMT9g7P8wF+~wc~BF6$%0gWKsRe1GAmaKy7rj|tJru;S-DMj@v%Fm8ZDJPvhdG7UKK -OkrY=SfvtNUMekR!?S}_I<_6BPbtl4bfC3~IxK*Yay2)t>86IyufJ5ldqkJ{%q2SkLQ>_ztTEftSX77=}a_&47k7dtKX(GzZSzXerM`6F9&g -LI@6aWAK2ms3fSz9}1+nApg000(F001HY003}la4%nWWo~3|axZmqY;0*_GcRy& -Z)|O0ZeeF-axQRrwLEKY+eWtE^(*G8C@keFY?JK0=o(Og+%#>2CTS3-yI5ocilRo=)D)@ZkhaxD|NA| -Ud6Tl8?yig45@+Vj%z3|#}Q_$~~ -x>%S%;XP9|G~o#nwA3;xFsAKv{ai;Zl*Q)Rah?`0|Lt~&mr<>p0GcZI5DYu1-dUGwq->=HtF;YQxU$b -z0-qz`d+zCPx<+{wK(Fy(D257CKc69GT3o3g;ie<`Y7zHHm3UC{G4#qE39wa4JmySBMKUJ%e<>Z`iBs -Yjmc{(#WshR1$+|2kPQHwQWtn;GTGZtTK($imyRJLrT@M3QxUNNPc<}Hg9EHQ!F*;-p4tYX_JD*H0E?(z9z5Lr>e}4bn1i*UXu+RJ_qQlUk+E08#SwGnG605{iaBR8UJU=ZDZ7QJR@FsQq>FG2hnDjX%7X_(H%sKY(`|d5rwPk5%3I$DP{CruQRw ->TT5&^x*>!`q{{MO6q7f^PfK#%qR`KHC)4wtgEKH8XM3}-M59BV$b;d4-Nwh5^M+% -N2?s+FKb~61tU0>YaVg_acST)F(QE|_W*|f_cpOp7s^%GsHLeZDJAUTUi@c|utMJwoc1SF+0+*7!W7QlK8;%Ckyd%Sz1P)N>G+0|}-5dW~j)z923DkM -p_c`a>lH-~?UAz~ckLS&$xH;w}FfplB=8$y<0R+XIY^Nzy=v*cqrw&OxrFz~>t}D`R6s$N`cTTcabqV -j$7a5Dx03vezBxb(kwy-nGy>QbnpRt9~PCwNVeMhHs!yaCU)OWUp$q@AugR_5DxqbQyXOnLp%gYG%M) -0)+$jiO3u1p;+IFDcmoI%iI+025^TT$#&`Kd_vac1p&D00Dcx8Vq{x5&156DB8Oi^Uv+fQx@q@C1>E1 -xNcs@j3tdMDi8r0n-xsYcIvHqL0Xh!3v9Jq;SxBq|M=e^=#I*K#SvGVxMgN}9p0X_n_4+T2(kf`m6fMk;Fd!R -#$E@oFoMO)qYg-VyWd3gBzrZY_7kwF1SvuqKUQ6;&tO8qrgNyHKn_*l^`)MO;yPS8la37&!;+Q4otV_ -8HgahKBnA0x*cU*~xl%4>WUZ3=M0qkv5(;9K0K4ki1&iGZ3OzD?~MXA(4tXpHA2v4^jjtOo5v#BwXV! -@;u;!$CbJQZN2^#nAo}o+@Y$WlXg)nY=QLPcSq5 -rZ+?c%a&o0sVZ?^s)VB(!tfkC8O8$#E&81eOc=VAlwz-6bDudU`sBwb+b_*jD?G=>tohtxIYB3S_2b9 -^aea`p!um9h3Kpkp!S@_1mCMAJw&SIEI-Rq-av#}l4Ph~fEXh;MBOqK|1AtP?SeydK(plT_8nU;)~JX -qP+Dd)4LVC}JPSo~fiiJNqC?}MoF>bRX^dEsry)5!V~+D4&BXKKQTU9kg8UA%M~|!#osf4E>O$5{`{I -Ba$yh#QFK0W9@*;byqp*hjB&8$8hD*`>(B*G{rR{XhX4D!Ei0AR18>t~i-Vm&pMY$6}R7-kHTSW$@ow -y`8xhA$c+Wpzl=m8u;yR4cu+~6qCL{l86>Wu;$+0)a8xzM`q*(9}%fYE=l-bEj1dhU}nm^J8BQO!;`F -$=4<=L-)&|4|6PPbP>$rli#I0qxnwQaNthk(>?yov>98@I8xj9ErmPo@{f&&Ti4J73d?NfUzVRyaH7j -B?Wh*L=+HR9>_mvi}DKSM_{~!C)STagQpX_Bz3>fY1niB+kU?W?f`}ufI$BUxJX)mYol~o7*`LTREX3V(kvWHt6^Wit1jvRj -c^EHZQCp5@oIUa9o>jh}tT3uwaE^S&&8Rf;b;v@1LT9lFxQWb50I7kny;U4Eo7f;a*{Gr%ABY!^7=ZM -V^^^nZ0S07iupAJ>m|%Pw&AtstB5%ta{e1p(T~y}LI)KAe0=;G3ZAQX-2LF9gX|mR>eB2fa-Rc@GZU? -6V^CEnXK0O$%3{F@?{_`nx -WD|trX@+rBI%#=IJ4FJ>qVT+etM-?>D0JTeQmJ1ra5{D~5(f3ofk(908)eZ*c3dWE9Uix#7)VKcki`pUpzyfAlC#1$xEBsxeW%A?Lo_A3DqpVw -xB|Lhu?{#E@Vch&?C+B=_XRA$L@bmVwqFX}(b~h{0H~-QogOKi*dAau?S7!1G)89S0(wf=8#_GE{$9}~reH7IOE`hcT3WyOOaqjmep;;@r0b$ -pB+YpJ#5Je<2bji0?^8q!R0$7+IiOJHottT-6uAtMorpO(1H)4MZ2Gq0oVGs5Qf(E0o9s3ZfX9m-|9<&QY9xVRj`>LY9y5|1$Inr&JdjALSOKrAz(?ZhA1$ZUrm{HWP)*iU{$Y1<%K4j4$oH4medsGJSEFc@1Brs2K&wadSLCx%kq0jzswsgRlFHeK~kD3wZ -iTk8uz^L|6z&ztsCr(|G+h&*+yMwT5vJOg8Tn7Tj?0HaiLj6yCOR46V)fcW>6-Ok_pm9QXoPWub1m+$ -4DPku20&UdL!vvlsjkswf*~@*=l{*OWz(YkJw>bpKB?c2YCpc~*vKZxAqI3b#%kY3G5Ej$(Xs%*kYRu -_#-rOWP(;Kc$QF1Cd2&Z_Z%?S&Pjic4~i+(hn?A*v?R_kM?=7ga*iJ%~Cen7xH5WeVc4&fgyvOgL1Fz -O|rmRyiCBed}8!7ks0+N-K)(MdVS2^jYoqF+xk|R? -oyOL_aQoP)90z5G5^q;fnlnEHb=`P3kdz+Hl5KqY8ze=MHX=od37cjXP0o}x=0L{EMona;f#YR+Mm})&SM(q$Q>y-DHjZkW7KSY2{ELoiS&KoW5STIlVu -DuXgfoE-k(TutnCpv66d4c=M%%t>^xi79;$x;W@e8id+blw9>)imKdQlPsG2^yg7ci>C&}H#r@kh}7zIC#wg%5{MaQ>?rIDpEqYHRhwEmz?Xut|YYW`WEoI -0h5|qDN|&GL3dJ0VKk@V|8>@cygFH@Z@$M)-dR0nD!*`;^((-U%vR@H7&I_k8>sFlE}Joqx5(JFwIH_ -PjUBq`l@41<#$^IpG^fOM01wZ0Yc>jG*%nHwILqQB+>y$K8iQ9X7~iz{$fCxHe?7+BxjR4kf1|n8qqe -5`0)Ay$%KCb-wo*DK)pM5yQY4M&LC-$-uK62A`V-|D~WmPg1hq`P7u?oYKtHLwR8pK5lT(QVP$_{vAW -IP{p~6Dy8uoG^vX_kC5}a2<8Kbk2Esk1dKI^?al2h0El8%8Qp!Apr@`i==@87>wJNME7#YxK*5UD#0h -%)5#M2l^z;J+jbkPk@aRT4CaNrD6qH_76?WqnJ)%_wb4m)oPZ7T-W=NYyKWM(_>+z>mNR@Jg?w0=zfv -)vnC8&D;YiPVFZ^+tyd%j3RPL4ezm-Z36MrRP?w$nKhYxE=GwY@aSLDiVaaFLx|Epnr-35aiG*5GV)k -GhaA_Jx}PKPYXJo-;FN!EfRWWWj=@ -(TvoDsyn~RnF#wKs*zs6tb^-C+86B=7q`KCD7MrOMS=tZyRtT8jCE`0!+#J6}uZYGwhSWq)@9x_<^|gV?H9C@^GIp~N(8Clk+fA0@P -yX7c?W+CaL4f}RSlTwT>3grS#g|$<{uX{u6IX9cFW-HuFqxB@xdBvX>TDNi@bc&^Xue8%K>rWlHvzi; -iXIxhLt!1SV9kmP;p`xWt&!HPCQzK{-KyHYv$il1CJtDS+SmpZ4tB-S>c#Dn>{q~r<5yNN;BCSfnUig -tCL69F2wc=5ZE6R2IxF8C$eB7dT(xuUKokc-q)u-PIe2guikqoDfZlzO#%Oiyc}r}5cDX-pr%fVuLcv -!SHwfN?BS7SsH`vWAUf6O7B`6VrbT9|2=wT_SWvknVC<)#Ti&4w?#9%Y%>9I*zIDN6M=v5_{8V9llML -tL%K$Ku$(nsM|e5U(@z?tdGr&KbzwR6nnQo|uc)(|-{y{x74T*-u^&6Be+)9iudSdhZEt*WhMtFlsvb -Cdg}Ze+~lQj!EOo}@+uMv6n;&Oi`f)`&7>RhNryQY{1u!a -NT4ng+Zv=WeK-(*2YdX_-^UgNtmzg7&&HOx7Yl;7uV;gmi<7*aWTj%J{%GE%^qHRcuBTVzgB(%x_#qJ -ZW+?!E8c?D?pwk*Cbk6$E{VNboeYp1^PP|R=pKh`ow$+ZO$nUt+{e@cm8Ec-uuQZE4BdtdN!byV403R -rLV9u7-4nD7f4BlPLZDHp_XaY{<%OyB~RwzxyYkjmX=(+8Zr1o#F$UmnxZE|3KYy!8;CP2l{oE;D0hR -qM_Bma9c8Vu>^k!<)*ulMn8({nL>^wp#3@18id?7&+YB&Rce9_dMXGMc5xKTdzNd1A9LTsfMBuuHT4f -NAg9;JqTlHg?{>qrhRS8fZ6X7xVE-#==z+()vO6pme@5Dfbl`@2M@i)O?yN{T^Sm2x@vh;;|6{Ff8l? --YIhr_XwN!77}32jMPy`k%XXIaU;)6sgbv{J0;*?8iWiKjjK+Y+Al)G;OPumMgIGqOHyC9(#~Kpvqn%`6nS|+&OP^%&9o|5l4P~5w -MY^si%MzB_}Zwvwj%M*lSvpi2Q!&uv?k7EE-GbgvqY2|m5QXOb1N&ZZIUTnq}=i_zv7!Gz7?jKZ{*9- -n;1iJp`P)p)_E@1c%zMI7GTWpF_|O@&vR&d$v#IviT~8X*gIi&DmBrU$z-z0xiM_`Vy4!=h>e{uCk&q -`irn{mRX&SSLatzie32!C*37Mkru&_u&2v$R5b|yuh`WTXAt5|JB(BV^srJ3Bo=ZA$K3Z96=@+1TwhOC%0PDdk3Cn`*F- -b66nM$k$NC&({M89Z-yg8lr{k8DQnBv=09`f8gY^A=e@^&5Z$asu-0Ds$JlubR0S`aZKUa_MZ{AS9RXIeJmj{L)X)inrCN0~`JZHI5Ppqy;KaaS -o(yG#uThT?2G~<>cPf=v56pcxIjr3x6uUdE@(%9Dv>lqzXy)C7cJeR);&0ssnDxu9X*x^RlPVm$Xvhc -a_10-6Oo4ig1bhmH|^l!qkV0D=bWU0cYx{*O~VOxomY=w!JDwi8+^F!QbuKUr$T+$eG$-YJ&D$#;Pzz -=;9g6RAH;qmtV{)ow(m8U0L({aZyVYZ>9y8McV@XD -ybla1UTny-A=^R|Ul{^-M#JTZxztSUOSWp+w4u|un>+8k8*mc>)etW^F{HFNbm%si8>A;s`(8jB_NUJ -szz=F4=K3+wJNUZ%8A~^nwbE8AuMMnOa|2>go>*2ZuQJAwk!i>-`aJHC{rKZx^Ik1-v#}IuOsYu@XOo -~@aA?U3gx6A17rd-_-r*xd?%SB%TDIXOiiX>c=&hhqZKSO^)gj~`>Od}-vBWA=T8Q-nsnADNG>>vPg~ -cqTU=dd+WT*)Rp(T<1PLvEK4hap+`GN=*Y=s4Xj#P`q8}A&Kceq8a2{^~QKd+m&hX5Z)ns<=M -jrh$Pk=y?`m5#Hy?}z_$V)qia>j_#}Gy`kmHd~QZS)IG8ecW&568oOCE(<2?W*nWsR8vwQnd}>etE)W -yLeyar9IQF~;HqDE8!ek~z2$vqH3e}H)3dWijkn -=21mW)}qkX{q;f3@F<$n?i$v3|phQ}RfKw&HKpTyg%)u=O|!wkXU7@RL??@p0rR+pRZ*z)GdZxwV*sd -9v~Ti=+9A9BUh@$!mtXsLCelS<6)sZHFYIT=^NpA_R3ncdEf>~`H0o1mj~W$@zVe58|~4)^Zdb=MvN+ -S>2>oTkjWGDv4)*My_If)1$(mv(9pAZ}F2*ZHA?HqS&jYEgralH$8zD+?<3O2zC3#qL>XcK|5(fvq9t -sR?!`NN!8)*%n@J_5D?CcRlq#q3)qOpd(=@?>z5Au5IwCv3L@Pi)WtK)FNoeHtg!-hgKB5B?E1Z4lia -`tbOpk%#5B(NLw%h!h+r1-d#I)&Y7PEU+Ezl+9@QW(Jy9yr`rF*sg+zVCj(XrhdJ=xNy(aKN==t+y5@ -Si7=m}7JMr%eyPNERPRzzV8WIIYr8kzwMU+NJt7Xo5!vgGtS6T4vVJ9B#!bf-@ZQ`H>l~q+y3MsF3vP6}@W`zVjy`)x(7B9$yfLQ%+7}1AetrXcebZh~k=k3`G|FLMelqu)o%HW(g -(s5e*WHT85C3`>3f=G`Tcvd_X69hx!u>R#lY|&HNxsBM;^rnvuMhP%TpM>)9?n;1TyNbO?P*WH=EL~+ -GoL=U88H<45iMOm!bld_1XSQPV5tN}6W4@ZkUhVPU%*I*T!g$t770pATJDZf5IESx)Z4A_Efd}!T$fV -{@M0(On@WB+{iCT@cFO#J0T7dstH~I&W8&ZCX;w!~Z`<5E&D(RoUu=9Y)m$A*>hI`o21Rg8TD8@{kn8 -nYtGdo_R0AUf+gV$g<;6v(4Nu{uRBcQw)rBZ8j3pB;p5L7p5_?dNoo{QIig;&>d?+@6^E&`S&I|xI`7 -cmQ0|XQR000O8%K%whP$0NEV;}$kv3CFf9{>OVaA|NaUv_0~WN&gWb#iQMX<{=ka%FRHZ*FsCE^vA6J -!^N{xRu}iE3oqEu{4o6wlhz6+)m;+9e0yB$M$U6ljca21j(!^Qsu*n`eA?j-Uk2zAV_<*v$OkQb!Hrk -1TKJ!i~GLtvdGtBGFjGDUC7BqBtWk)-;sl*sU$g+kq@A)M2h7)sgC-IaCJVH2!^z7}6zf3;7yu7_}G0;M!z -egU0@FHCp%Os1_GytcK6@9k;^v`6Y4%Q9OsBBIb0eO9SeRDf`d-3+=#Wi$2qK8+vKV4j_hd1!_n0dyB -7jNENeVDv@ar*)uo`i<@rDwMjmB-(RXQB(8+_%K#KyQDAk-et>jiU^Jkm2kq85c#o8;&!hcc?Zi -NKzo@2j!_QbOEbjwE=eWv!P6&^A0GH8p5aFh5PqQRLz35}*&J)Rt;YM}a+T$Ur1gn-S;<0FcT&(9Wc{ -Io&^x!@#$Ky0VBeo#0E9#Q?G^IGU!-Zi1)Ro73c^jS;$jBEQ^a$`Pz-8-NHUSbVhXW>0mM8_WLA|yXx -)2hTEvORh}21MHm=TU;B^y+b*nXpDd_Q$A3?Q)r(eP)gGU<_)um7hUsBSXk&P)X#eC)^y$*63-j$nrg -UU%MW`KHez~DB!0}*EnalqXVgw}Mdj%AX~CDPtbly@;uWNb7J0}| -s6czt9k_{_g6ZAIAbm_S_~=!G%VOA)v3GIlGp|8|{lTb1ax_gkB)ZE7KrH?;K${|;BNS)h%d{@NzmmNXg9&EAEcN`nTd -DH|I`50|oAckMjb1z!#7j^Segp!-KGucn)|u8#o|3t6_r`+|{+#PAWtuYNFGM`cwco -8f3g2V;$MAP`Xr16ISqcrg0Z3>G96XWR8iACEqn?SHbs#wjdt3LqT-hFlC?pdQUB=DX;YmX6N8N9b}^ -4?h)5pmkWQUVdwF4vA8?J>3nzp#U48fEl!U+Dci|sm)=DSs7Tfq7BJqy9$?*9$j9 -csqGyvY-CJSqR}`3I*)jWm%yjDaj_^5=f=7g+XOiKDUESDy#wPrV$mAe15U-Pu86>m**cxhz+j(FVd( -@s&Ec>tXz{!&3pmq<3%SN8+dIkUV>-nxgkHeFHnOO8h&dGUXjnu7#DW6Hi^i8}nG1iu2^_aaDKlyhv7Pi5aI -yGyVrAoy0u>Vp{+wMOIrHR=yDAUWuQ8V*{Ocdn}R)PzQ5;kQuhd^??SH)@p%ze$3e9tX_#tU2O6a_^Y -^tz70TBQMaI73RWQ9b;2Pp5?MB@yQR&ehx)j(e1Qd2W*z@o7b%eUbL}|-(<8dZpVNja@v{X|#zFONGi -m&zmSyG6LuRm$x!LVWO5U+%S{MA@NI8BlK<}F0qwk5Fb&6%da5K>IWC5IBtZSI|I<73)_YkMG!~>+|E -W6?q9gLL`fyW~fR~ma`R?*pJ@r331)w|n^cej&Q7w1>6*q;8#Jezbg;-7<0-3+=}W((3Y1O+UcdtYC? -y1c$Pzn$FNzPSDQ=7L6gesnw*&yP;<$8-GgE&llKQ$UbAvU_a4;b}P=44MawiSi65*E~f#p)SvknkOh -q;oRDffNhoV#P3-o^U7}Kagj4R(U$ILAARp->u7p|ZSp+z9yMNJ@5Q=Yc~2DWdr!mUc}DIdyDw4p1i0 ->9zGwmFmvC~tH*sC%>}R!Bs=YUgXcDkrP_WG9P;JiGbAX5J3e^;-ME%w*sHNb09bDjgP|rcb6oGaC{o -=k+r@x>+lP=bF*>Kq50Bsy0YrzJ2UwuN -vhzRt@j`A<%qtXp@$J%8W&%_-rhOW%8N!GFwfm-9`e@ -(lkavBUqsV|M5|XXBuANKi(TF;B#i3b7z9u)Z|GG5C^`IX`$R5e_y-+BWVlLv(y4?&S93rvf8C2M?l- -k$|dZu)+zqzPEvJf`J0=UKwitr-|niI06=YJs}qP(mf=8;XtWrVG=ou;=38m|4EDvVJUc)&Z(g(KZ6k -q9K+A<`&r#jPMa5)B(iT>z+$@rE4hk&3N}A);*sEt^2w${)M=A(T6g5YcL_S1_0{_(>%Wi5?+G}gG7!&3njQG -sDHulX~t~%hp9Zvyh9V08*c%`J8+XzUe%$f?+pzd7Unhh;=uiY@h;tUVm(seaaeCd5Vay?siQFJ&xex -jLQaiD_CXRH1fm%en=(9(I~=cz(inpA+e*mw)6og1Uhi+j&zE+XBL_O4H^Yeq22SUVWON1Eyb!k&gPs -$J*$C5I;t@rb+c(eA|U5zJ6n6@g!+ipsi=U@;zF|uu}b(^SItm12tUN)@J|ja=_xNj5IVjwNSrE_f(UKxYh< -FOFs1-+Pj%b5D6)71uQGYuNV%(Q@;EA41tlg$RY9B>QS=q9O4`zsJ*}P$?GQnj4Fp}2IMtGy&2Us69RD -DBrIsRoh?5lkU*WkLJ-yA7wS$M7*qj-_s(wM^+oo`(fwUtL`{WU9rc6$#CCrr%V4Qi0KV-aB=;s49Y! --@;g5Z+L1H1jyun`Z!7zOBQS{mM^9bULEc5I*l=rGrrTw(+Rofo9)Wx}L8MXkPRzKT?e*s%4Iw%}S%w -rCIl%cf|Y=T1SsGYCVu)#|H`Z~)RPwP^THekMXqcSs$xWU&C9v;z{Hnd{f*;@gwwM*w3ko=4B4lZd0v -2>0o8Qg|gLKNl*T)hrV>r#0)-E?9aS+phzWUp{q>=+&QKIDg++E(tnXvM4ulo-EPQU1LB2=+E<&l}Lj -Dcrz)qCpg9^=?Z!V9b@Czc3o7lmv_Fgjs8Y -*Zb@dt6--eMvO@-;eXf#Cy*pMkesI?A$WNFhw8)Va>cJBoC#kTv{WIU0!H3D4zWf7(kk3|TEM+T@#IF -uCMEaFmO1a*2V`O8W{&q#KU@gE$JK(gkU0QRnDfFO1DV4)%(6lr*X$vAt4Y4}xUO*Kn#pu`9>oje4?MUIb-#1GoZ2beyA?=*rf_$j#f)5VKd19zSV|3@toNfDT*T<8#LpFu0D!+Jx9*Lc^^!;{ -SFN|HGxk$cC*F04K512I(cRC7qxULPGYpNl(g|q$aJ{3QHc@WBWqTmolk_ooXr*>81>cgxDKwPDeS;lgk_K`oV -p=^vb@6etjzGrBe*>WeL(O7xaeaMtJvo2#;)cUK!^^zaO+q_4lyvBw~L)H7O9jAE)G)mHdjQ(V4x8i(}szyRK(Ae+9;r7sI{RB9xmwmK18w -|pUZ_w?T*Ud?kvfJv|^^abekHH+Vg~=Rs!+y-2V-wKMw_tHU$W&}N4W_B884C;xp)Fjo|Y2_CT;gJse3jnJ?#yeR7a(S6z8}AS -_upv;b(1k$`Rmyu&B-1vNm -{;Q=vbk+G6^q23!8I1rk*maz3M@l7k**rURnBgu)<0*&+A$5jtmzpa&tPR3MlM=4k=#fBE*sIpLL4hS -+;vNJ8F1;2L^tLOt)H^icLcf&Zc7w{|?-qt5+cFAh^=>3PGtNXDsl8=^sLPG*tc%M2t^fkFnR?ys97b -wIU-NgyRxcSXL132%)uA3x;Df=Qd$&&C5%XWC{+(w#d^=h!>tXTf3}nh*@UNXXpccqNh^!AW;Ct4uOU -&_qRyq?*@-2geSK*l&~`)XXcq4VD*h4e+rG3NdnrlXM3jqS)ywSAMS=>;FPbBF=V{3XQpsl<|-tuh?Q -AcJ46K_V{WN^7A|)9AmAtBk7?}<8C#MXv)LDi2;ilUNxy{9=^0-CofsOTqbj!jK`TCZE%=g3<3}pAjB -1iNhxbQ7I!%+6H6aSj87+Tu7OwOc@J8;LW3?oEIJAzLQ^cw&&e)q+{ksMQWBYbVC${Zl7R3jta^%`oM|v) -VhUp~k~y+2_Q9f8e3r^J9*;)nLK$X9Zjtd=leG8c3-LVq4qZv-Dp&dkc<_DnEzEEpBkKX2f%vZ7s0O; -)%F{2uTmk0mSp@RrOQk)Uf0MGTW%=dF*T>KI>33+fMw90F`0JCezyALFqoc@~hYpY?8K&C;LDRNwc2} -?#!1^H)D?G(tOD;Kb``ZgRK4J@ja>^N}i(IlTj(gl}Z&7BOz{OE?^qnFN= ->zcu8r5W_=YeWIzRaD9~y{l_XtqIFq}7Ij$i9dCqVa`PKTjf;me!J&Bgi8*B7_fFW%j}zq-C{wH@k)@ -Be!B{^A`T0+?O7D|JdC47$t#^~2|5aeQRu7>!zX)RG)8Ngln#qj-7c8O{6Rr7RQFQ>L(p3Fl(b4zdLo -1F3&b*7cerbD~WCDaXEm1mOtd%qtAofsEYIQM@Asa%mWY#X)8}QlmW>%8LreUC~zLHJBP?Plfykh$TG -QC*n`ZOQZ_oHLUtbCARe3eXYW*gpakM)s%F?05ci442OMj^7S|3Eews;FNc=%HE43~q#c}UDn9wnX~W -g5_+Gg*akKI3znz6#`WPDKw&4J-{@BBU{)Go&?9M0bBE%ktXHECE7b6&mXbowP)bCwdZdA%d_r!WjJn -=IF2=}cwO#jigNGcg9_NAGCsZ6sWRRa^L>NzF}#`l_->k`Xefo8oE;_h_gg1`~MxFJuvoiEcu8(2w{N -_KfgB#62)*|PxrtgGtZ#Teo$KoB&R*VK~yuGHsTY(L|iAG%2_Wbk(dB$9Ok;vG?_QaHLQJve!Ty=$yW -Y~jQ$776M|ag*KJRPUT(jgm9H>6G~7bm~lSI<>~rN+wAbcZ{%OiZa*)D*4hPeNrlfN(r6ySx_LII -C3sTjUqO;l9BFldJ}9H!tq}EE6u{$7(Vd7y9!7n@&|_zcr8H708i-6F5-cB;{!8FBNk?M!{ocCG0jHU -$dsAM;76jSgTqfP(u*Ra0`R;lP(l*UD{|CwEy>%KE&l2p?+1Ru($G02o%g)7^jH1S(%)!~tve6!Gde{X@;;5LWnQcg71x@dikJ0jwbMl ->6p#-oRQkgnfR;y?l=&=I386iq&ipGoRg(Wya|PUs_^9OxT#7bO4J}+}#~;+KBi#dK$ -*gGL`4O@vl(sR|mLPaxTeEJ^p$uPL7UG#1JO~b{5!QWyG+eyjZbub^OiA_uq|R@D6ZXREit~ETMS_qi -E^^V%1zvzwR6-Ckm-+G9 -%g?!9M1S@s10-c7AsC(P<{ZFtkT5p=9Ia*=vXzwV>vLF1yh&KI5xG9oGlEgg|09#;DKo2-SJFq~cp;ISJTKLhKS@xU_OSe*H$6>l -~vr*d$eW+aP86{PPZ`=wKU$1o(?s8}k(k)|DYQR&H0`~m5_UKN{(p6QA8^907c6s+LO`fdF<8Ok>e23 -otLf2}8&8Z8xmwBNSH&knsT^>PGBra9Rjui)>ec=U#8kAOUI2F6OCxo_8#m*dl2*VHrM;=gQgoYp -C_;l(UwKi%>|#JSvx|6Clpm!0>jZf?#a3pl95qZ#87U!ZWqdWVC6Z}Xw67V54tO1YPL%$RSN4nqv$q%Rj_EvaIwFT86=h|>;>s}8OSSqVZ1y=kTDZX0+U3f9(PrC-jblR -?^FQ(HlAvdMr1=b65`Eix*;S+-IosZ7#ML$UK -8HeO9=k8~)eePry1TE{H4lX<9Y;qIXqlWg#n%xz6lZQ -~;KJ8*7Y$rI;<&54*qyynvw$u9z$Y3&8F*|)_f#yiY7jPK+W7{LrY -=05vS1J1+`8cj7rlxu3~*B?RM-0Kp_d#H0C{JSjtSV@y3@ -hHfr0pwOa+M~USMkxc0GHa3M;}~Is`%0yTIbs*{8?u~z&N^37jYH%-czng&`mq(Ox%Y*$+$Jc?@^xI~HLyT?-Eeg0MX~`M&uwo-yn=Ttw!E+QHfZmG(;()NTNva^J6laDuk3N -3x0&xFJK*2=sPD&Xm&e<+v*djJM-4)9K}-`b;Zqmzd44RPS-Ql_Vf-pra&Go(4v03WV>zKCX69a2+ORFf~Y#`AzHkXbseE{;$`4 -C}Jg)FIjcWGTJO(CgGqjpn(Gogrwz5;QJ#l2;{rZA>1=PTkZ$MeC+Y8;ZPDwu2a9ZS5wq&{?NHUm2wt -=@o1D<}FgAX8rN-ur;JW@Fmj0oI%+EA*b?-_>ZY!iiOOhrY_k>dd|H+I!^`GCAG+$2!_~Ls&NxbvhJP -I81*fSA~_>pu+F?*kYcl75<|twq}Fh14lR>eLQN4t-{xEIM*d%?Ic2huHQVlaB^SQ)#_?{+*ZmyHUVg -`mSh}d6VazOw5`ps*hLfbNHQe6>N2$!8VUy-oJRsP_&bl1z}n&I`f5^36F`cDpppRCiu3ys*G=UEHbzd -F(12-8nmBuEnqIU{Bf=`G&tgAz$dpsX@KzKylWa4HQZS`v0__-^7#On~Y-0k -gg@A+p$ujbV9m+8xfSCeZm(>^JeQ(C3e%+s5GYeDw#XNxUsZDm(u8(AOEjF1=5D4Z9&KMN;p1(&X-(i^SgCar{kKK72EeXQHnl@>@R?|1Kac2OS(gLN&PxkVrV$dq)OP&0u1GeZV=U -#yeWUG;?qh4~}_AgQd-8@N}aKx|{=?=RtVg;RA&(6Ph9ZSwQSP=sS) -cwmxw3hgHk>dn=`lT!w?Ea^YXjRE+|wsYuitR8*H-CSWnSZr`K0kBrKM%*-j_@JK7Q -`p{OJOxleu9PwL?p_v7}v-QNn4!Tt4XC9^0QIq3`PU|8@yFrR)X~OrXk(U60GqYx(i;_*2xnGq=ZK=C -~hQ7hkJO^+9uM|4}DXbo8-tWzeNkbOd!8he}=wTy)=#k^=!}e6tRM2xQyZHoI}C{4EPyg2PN=HW);I& -y#G(a~+9i-1p?y-E^v7R08mQ<1QY-O00;of09jiqgXtBu0RRBK0{{Rq0001RX>c!Jc4cm4Z*nhna%^ -mAVlyveZ*FvQX<{#5VQ_F|Zf9w3WnX1(c4=~NZZ2?n#gjp6+%OP@@BS5oF5SRlLoR_p4hwNY4&4N2+a -5|WiZpAB${Hb!o$kM{WOf^88+xs89*@3xdT%U*D0aPxFpwTCf)6wqjp-ewi@*dL85INf2pjLAcAaqu= -q3}$4d}QmM1mA%@Dvy*7Db_P4<@$Kdz{->7uN-(Cm@f(A;*|FEsw=F4{X@VOR0;N}K|KX6a(@=Cnr@> -i1YqW*xCO?#VlHoEMPS2JK1{aiO+>!y8vyxV=-G__d6@fsIpx@L;})o{NOw>Y6CpSQ6Ri=8>&rvD)Ae -$Hv}>-Z=1QY-O00;of09jiK?vsgR3;+NeD*yl}0001RX>c!J -c4cm4Z*nhna%^mAVlyveZ*FvQX<{#7aByXAXK8L_E^v9ZTI+M$xDo%(zXDYzW9dew^O)%ePjww%;^aJ -+T-;chn@)Q}LnI{OOcFeTl&yB!zrDM75g;kWZd%PGrohEwAHRJ7&}1@+mRl}KA+k1Sq^iY^XG{`GDj{ --G(2{OgO`0ujNGsvRdm%PJcu`g4vfS{Joyh3+%jXa8Sd(ta4XbxNW#muWrm5ul*;$&4Hj6iEQk;t7j8 -+v>^UgAn%Clsl0~;#HjNSec1BhB-N3y(YYk~nb(~41!6Hx*-T;*Mq!b -Nk9wFHZ6d>tVAjMkLWG~v9fB-p9m7Fo@#Ba7WV?@1G#mO>f9EHMXs5ZR*d1(6j?`FgL=wH7%rL*2os%F?&DAZ)y -+MSgp0bCrt9a+mCA&B0VmcHxPi_ZOGRmz&Gv=F^*>E|)jSkIUuthpQjRIXP+SmYqb=>zG`vk-x&9LL0 -g5z(wBW?12O0=+P!zn;B{2Yf%=gY> -3l4jwVoHB9)7iWZgDk(j>77m#(D9+lD3jdlZ>*F6>h-HXDpI^YxzAC47~JV;r-Gl$o$JBvWN+wCH7Bi -+b+9{-WpVfsgdzGi&O-M{`|^Mx-M5hHh>?c;A-1U|d`9FTY-oma&Z2uu=|gE@i`LM# -S1A29w(fD_YQ#WvWO60Oz{&Nl0X?-(W80GRHQ^&5}URmqJqnlL4d13}x`-haA#kXzQ@Q#nPzAzOB(i4 -nIRaUZGkL-aOAqO*s~$a#$||FWy26)@9Z2{GzGKJl#>AV-AWvBAZj2h82~vhfaVn2ecZ=uXD7`x&^Ux -VerzpY%86t1yO%FW!8UZYZV~f|aM)rAC9)&2fJ!Wf?zPv8L%WWku*6|G*7_Yt76GMF_8`)T>a$?fD~0hSU@ -+_Y16#{td}CAwk -MYA<&DM}2pfx?~Z@aclk;jvH!1qvs@2EV(g$X%!cvCR5@R|kai^qTj#b<0doY;pKOzICs;f#*ZK0!fH -H93BJ`R?M&hov7Y01;Lj;XS;<)V3}a!+`q1(-bOg((EgypFkBevW*V-&3}CvOg=uT7m6PSw$IDM&mL1`D-`TvH+}+ -y@(O(n%rrel7Lt;HVHGk($1P_j`N3tKiMMx~EMuL7eyCB6Zc7uOdkQc$521;TcrYKtwI{3xv|6zU#EN -(v)-RFX43-rM7t#;a?@4LQ_%wNRwKmRo(qe^Yqz?2w9*+KzlNY*#cVm!{!9(~dwvkXLnohVy$K69H2g -Nv#8I~&d(EiVW2ML54pFs9>metBzTx`c6-u}swwthj8f?D)G18`vJY`pmCNB^u=DLu{ecO&TFcPU{Vu -U`>K;@s!QZ2kJkq?%}7aeYUA^E6x2H1qEd?&cL)K(lUSoS7BOu`Jz}US)t+SR|jW{;qYw4qQORPdd!$ -_#eIokamI21m@ror~ocopLkm{2Lo1^YI+?H&+h`+&ATj#MMEyVLSe&3&-P1+pYUiJDOn9oy7mzh?!uzKeQVKonFt$b3o8@TPO)Tq&>D;iI?|vhV$~IDP@gF0@ -6?>J34nY)zz|LIqXUBjQj{$7+OsW{XY*%agqzYne7+3K%`jL(%zJ57X%P!gV&)Ko{=|Omfoq|wOh#Y? -+5MSisEW7IyDiclj_Nlta5BGF@#Ya9?N^;g|?Bfr*5H1SdiB*UJvQv2nmKJjn2t6kWUJt2SIm%}?409cOfj!5CR$zFvo1CiO>MqY_*!Bkc4oz+)lsRKn0;p6 -l+fRdO!v%7&i8HopH`B>yBL{Bpo=bB2JthWc^AkT7c<@Q%$7=xkA$vazX$$WLH -LAViH$`tSt2@s`t$;+C$nCnG?8prPHnu7rImSq>0ZlnEI~Xm4}=&CfR6dep?fwafR3YM(eo(49s=tqK -Gkk?9H6}&&ye7&0DS~FzY5I5pqvHU)&uEE9T_~LUvOEnBkVN?{x(@(d{?!|x(kEptY565m(5yzF8==F -5nQ;!>VQ2_N#x2Kufu1^&{oB5Gf_2fPWm`ZNu-Hf(!`9);X2#;lXTi;&W0z)@R;_!PwM@v!4s9IX`45 -3nDz0F0Do;_Pt@jiz1iT{#8AWuUJ%%e615@(hI^Un+f)DjFrgl~C$XO+o9+z+eVbFaC^Qfaw%pJe9XO -$?*$xjGblj?wSfYYiS)@A#u}T~`tNQwlM(6qw=UDFHNpSeG2N`=-t&uW3u^wE8c@A~ds#iPMBkEfW{4 -GGjp+Ow1(KS2PUJL(9WEON0w)3+fP#b%DGLAP<{`axn-&cS!89%WyF7*!u9B4*Qvr@4s@fOy|sm}1AD --O>gu>Y@UO`!qB8a)Dv#{puP2i+x8^o(5O915^*a;5%F#_(4g2-|{!f8Zaf+Eva|^@~4_H%)7r!nP#x -f4T^M9Wsf1ifHVYj$fbVelQ7xxDm}Cv1X4h8nl`RXNF|MiAk|5 -PH&O-w0I(4N04D$d0B~t=FJE?LZe(wAFLiQkY-wUMFJo_RbaH88FK%>fZE$aIY%Xwl%~#)V+cpq>_g` -^S5!CKf#YtCSFklQp>LtO_v}j!PAqfOpI$LZcQ6s6i+0g%eM^Zm6Ck6H}EH9zRB7?h-sd0pb#SUAriTkd{t_UVf6?o`F(^jN~I;sN23d+N`x0jQy&!K3=*x2+2LU%VhlDyY*Gv&k -sTI?rA$-4dUMFl26Y&tW<33k$Md7<$Rh-6DKZGdb*W2i@qAjU!>MS&=1c??^grm#YgXwHwpbMAN!l{0ua4yLr6B2( -x$v*alkhC=eH1*o#HSh9z_XEA}}ZwN)T(%3E`^lT3E!CV_9a&%xM+DjEO{pz&H&&-@1EOu|i8<9J~H#G8$3rVcv#(9ELf|D -1&He6s1flFv;By?s!kYCo>1ybKoyB0%wpYiUADjL*(#4MPUBZomd-+sBvlwsPSnBDoV&F!SyVo)ohGw -%6y)49fo6t>Dsqs`A}?qdnoz#;T4k`MgLyiZ_TKJ+l6fYDsvNi22`OjCGqD^mmnOqda%sz|vWd$+fq4@Wy^5XX$Tn0L0_Oz -#3JH@d-&ZRfPg^60K_L2S*5vQBE`xD8EO+1%*F{RMnPX~=fu+8j+N88=fl-q)HaFv4xI -+oJ3iP%Z*Eh+FwNPzS0KgXhlPW044@_e|Ht)vQcL%`ib`+8hPrZ?8>3ZwB5Uh{fxZls};Ny)h?}Lbt_ -~_X#-`lK*9u!s{5RqGPq)ZeMpq1- -F)`JDk}LqR;lf$fz+L|uXUubB?77H1!p7bnX___$$szd283%pM*rD_Z_bD&>;O?oT@Rr --y+v&GW$`+Jxi5nw+1z?bI0z?UWZd?inPp2|~;xJ(i}c@A~M$bq+vjrb`RgAvPP`oEy6i`n-?+o -gI+pcf?FHd*)3L1b1F@R6IVKPA9#nf3j}`TkjIrC^3(0^Vdzm6N92^02(5e+cyGLeUu7i8$rjLb5o%Z -tFM%WFy11)zNlN1*L$jjj%zC#J&uEQK^&VP^;6S!A(e`)OivpgGwWqJUxT`ty`^HQtHtXZJawG9c{^*xp4<$Ls81SDeSea1qJQ -x%B3(8jFgVQnoa48`HFC-l+0V3DRn6lLNB%zECOyYPG1+g4nCd~vnBIDJyE29TQ`?RQ(dUw7u+JM97VV^yso|1Wj%zY)EG= ->A*%o7~Ju0y~*YRn@|SpXP8poes*kV?j0i%k*eizq>W(RhlL7tE;n${TXbfpa)Fz_$3voU%B4o*)sON -*qy3&>@Hi=Tovp-qC1t@m)JcVwr>X@wDfG}D#0!?iS&azw<*y+IVpvVl()rVf0Z;}8Pe+}Zohj^sFjE -x=QjOzYhq`XPL>?i{7rW%=JXp+&zG|7SQ4Kob*RkmI{N}@mpjj43T*vm+bg4;2i^BUxptp;wqQGtyj` -w-^wVU(UWYTwmF1|ZPQPZtZ|?68uG~?Ye*jQR0|XQR000O8%K%whPFm|b(G&mxxk3N{C;$KeaA|NaUv -_0~WN&gWb#iQMX<{=kV{dMBa%o~OaCvWVWo~nGY%XwlC2ZYoqpX6$5kYEHY_T8}MnR -GZj7+nMZaDn&seBymg;T!OT$+1mT<*WCa}kfLN~_CDN;r=}v2jYfB)yYX}b>h*fw-TPa%iHjAxy_|$c -s?0K(7i=MOcD*fDGG%W9b|I5QZdkDrO0YOpMPAN}Sf+}_3$~Rd+sHDByd=ICtdQ(M#RMZAd -TybwT~Iy^6BAqL2jMPk9rq+m%r%XuCHVI;j*sVrW>2bTkzuf+U5PM5!@NmC&roDPV16vZeIPe7FXo`m -2@9Kj&3Ap(rcl}HwW=iRMhMLZNqu^=6RIm`I`o-avWRq!Mc(a`f=v1zuw%0zm5`+gcSAD9&|Jz4JL`` -|b@?jy}S-)4ne<~&<%k!&5O@w!~Ili+Re-GD7}xn|i`0w1Lk8WBe+cPZZl$A-@3)%U!2awEX9gL*tz4 -nBbVU{RNNz&5LRzH(pVECZ$_zDWez2(N^t1m!^YT)-N{xtJHpmaRmNuNe$vWyV(g0Zo#NgctFH0D^_s -)E3Ic-%A)7fjNJM6`AwY8Vzl_*5+wicP`g!o}x*E0KP=xXqL$N{cy-;T&+Bh!Qa1uj^rmtV68Z-c-ES -`>XLG>p5O-Hl@H`79ZoX<@sh+FF3NO1oc?E-C*W{zgSkwLJe~y=?;uWp0b7J|TFB5)P4$YP_gEK3(Y= -{~KoR7PGqD8!UUQs^y}P@+CgXsYsDP2Hp>bwlX_;aSG0vt_6&GS&szR>gU&VCltz}du5ET%7ykI=tG6 -}}ZSrM-VtLK8ZS(c!7;PQH6#Y%+}Pp4Mm>A=Jz%-vyZZD4KgBHs?_1OCNpN&sN?AbbEXka@7zRE(>7O -)3rtE$R1(PCU*2)gOn&E&%vmb!p=nuh&=mMK(eJY=h)$kg!lYmo+<_&&}=6!Spl%h&kr^{vPu6@- -_3Ey76}8CRgu^BkQ;^L&a8%^!j+^D>W%Z73eIM8-v*7*)$ybJ=V*PMiQ6Gh-c#HLd$Jh$T^Cpw<*iND -k!;ylVaW{+H-AMR5;yBnq`KT$Dwbi!ihv$!AI?Wg$ZS-Sf5mbf}B*a#pCtW+t6mPbbtF -ypI#X@GkD=KCLxdjsvwAlXvp+tIm^6TJ76R#p=z%z6pHq-6AF-~ng~L&)opLq_a#uVXjr4OmY+7C8Wi -5KUY?|KfQsaN~fyzu{X&s8R*qyVKhcpN`mFUJ9i7z%%VWR(pV+*A^%u*vUZ38CW;tIb`cX`f(^7pfuxR}CQy9EMo2S^q<}XXBf@@bc<%+;obc#y69TpXtS*NqcpDH@UhzecwrcJs#h*=x@gV@ -o{o9KHp1E23 -ri>_41e}1F1eAV%A*fCpoCg%)r<=*$IQ;Sb)!AQeCx08m%qQQ!{oy;$lX<)hF|xxti%_7OuMPq5IfFC -}naR?iSV8?5RuT3acxiT>U@40C@KNM`7^a8{LlE18%9?dzaP3dM4>G+M+apL(P%YX7Mm1ZN!j1^qLIUsI-RMHB*->95U5m`q~^jKDgI#2Pj(xx`5K?2^(q>ZNx)fV@%o87 -~v-jLu>~XD+sa58(u2ge>kF@p1FLmVS%MGRC~ZNAU=1iTy7MbI$&{%jVq|+wp9M5vZ$%;?69>9K#mdi -mcaiZQ#6`8qtQulGW|V`}41LN*H -t=L4ESNZUswhaYNOpYs?nnqvUJdZp2!D_sQ$*MPuu_s7)C=*qgy2GwG~43?|7+~r$VX22w(>RGY}YI| -WjBNZEw2Q|G(5HDEORAa}%$#=9R@)>|ls`(pr4JK_%JB!nZ5H2jPU?J6Unjab~oLA^N58qzMuUy&^lQSCOO2d07HEDgSdMIkBvg(avg@of7)X?`+kxAk=_MyaCRqS`Ng{Mw2`$ -R}X3LcuNYcWyWqTZ&qPN0GpHQN7%!w`Y?Hhy`VBz&?qplT6f*NRSh;u56;Zk%CPNmg>d#l|T$iY&vqHXVMn5OlcXPBU%gmYAjwA>pk-k^p{$x)%oT!LRE)_PVy%||bfT49q7xvBd&%dlN${e -(Uc+<Kacc93?oU_3y4_Y9 -29%Jb7V#tQAD_$Ks{qnS*$aAKnuI3rplksI5`-jTHD8bI@6zHQE#HX#`J9N<60_l#PIBWBn#!2W?{PJ -cV<=d@WsW8F@_HSk_Y;}64H3*`E`rJVoB4QjY;h164zE?<85;N@I7wJ*jOcakjrwZ&l!*((^K -bxP1L<%4=szLIQdvlfe-Z*#!@>DY`M_;-L&pVN*X3Ws-5u<Mr6X;m>du6dW`k -|+1Y{*C18h3pKi8tg)K_K7Hgn+Fd5=<*&|Cz#-r*0>rJ`ow97eB<9Jj_82Q#zl+_bQ>A5NKhdYiuXth -`C8?UFS~NCSRBEVI~UT$4;*&io6{AK-H?)%C}Mi_E6Y|T!uZzUNqr|Rn$R?)*dVX(yRv|6dSd6_aG|Q -S?JiqZ@s^mydU5Fd_4|NZ!b><^^ETWIc@(9whBK}^!D_URj@8 -#r|Q0$1c5e@bfyI<%#Vwg$}n4Sg?5Gvl>&v#Fo{>E4gK9S9$hq$6>AV0>_L@;Z-VX}zaLOD%RSN|>I$ -u1%Hl$nc-&{_egUxp{s!Jr%{8#jjz$pctk?oPP({!Vvd>@KiE8{rQ=XUlNK9074=RGL{`a6c+#&h1NJ -Y*Q)bDPa36~A+%#^S6(>|gP;VO4(w+M;?Dnu4&fchx(?G9H6VuFAD=*rk%ZLIW#EJ_gLzr -b*;7F#X;ZpsXc9AcPsMgt{;njwtRisoCD3qDqUZ)mt&M%>mlt78-qPpFMyCorBL7JFbt+-1$veNS=?N -UY{(W=-;T`pfTK8l75Ge>_(V*S0P3`B)mS=8;BOQi38^lPkKAKwjqY`i?1{y#Fd?MEK!hZjpcp_wM9W -m-J*1wUg?8yZuyw;f%$sNqlC~gJHNcuDYufspf+Yv8`o(9!o`7{0}cKRJA-tM;kCaM;A@KcK4MUr_-13tk54nqbi$Wi>;$?!`WRS`+a?O1VVu?Fjy%vOv?SdeFWKiYhMrFNkG|{TnJE -(SNY>^tpS>r*5u5 -xt9uZ;3ML0$_l)%Gi24dR|p6Z93hNynfbbv4UZrq@Jp`!0xMS9^`U^^`HU~M-R3e$U;=mpDKOh++m13 -K1l$xCA)ZYqiviYAMy7!|6dcg;KAI=Y`En_uP{@TEVYki!4RqC~9*mn_^`%K)P0%<7M0Qc2qPcfh72a9{{4TyZ8d>_5`wBudscQ3SsO)yjM>^v#vZ(2z1fwotQ+rE -1TKaJEC0KeFQ{WbYz68nKgIMq1~N(e%Qqg2Lru{$M=d70;L8Z?9h6Z}5{N;7&aSy9meCX|8=h%-U?)l -}Y2MX{*ivAdl@4w`fjd_33mDj0X|m6l?_8{|Ino?_&xjCkO*8Grqd$brfu7eP_MqcG<+V6)`oGoO5>{ -vy7X#I&}J*$mI>q(ouC;jU1y<&9N$rNN!Sy>{0V*>1KR3p8PaE4{zUnygR@8bot+*(joCb!Ty0K%iY3 -bcBUTmhC}YaY?=eD?}W>s5j&uqxtHMgyHhL6`sNYmG+w+sU`F+if{5B`-ZtS`;IaePC=YCW0j6KzCK`G!? -XfcbNpWswa5S0V)oez{lAIWJ>K{~Dr~EORJE;iM4E4h@q{zV(mSXO+ixM1k#Q2@JyydQ6M@53Uol1`) ->L69?%N_bK8zFj#J9Oj-46B%&*D2gY_FHm;r(xRvt3u~@6g~it2iyiQ{q37uIj0}l7_R^P`xEd_j(kv -;7LelU@idF%jCTL1H%S!YcT$qp&GrlU0_FCWboNxJ9OvxmUY~bXttu8wsgmiND@bDo8(%M^#!|_{Pn{ -aeG89%-CJ0~UV2paGz0ItV_Gl;>Lvx%G%D9LG%NOtowGs#>nLmFnbJ6VEuh6Gc2$hqi{=k)d8#3Sbd-w|onPQWnr^g4jvkx?WKRCPW5AeTE9;Zn`BZnQw0VGR3?4R|@n8S!e^<;Yqa -#={I)a_j}J%hU!=xV>kvCB!mZ8(YU82rYNX_Se-#Z~cMo^uKK$*ahf*p``b>^0QT*hyQc_*XGHI(d!> -T2^<&=wdq1$whZGJdc!o&K2kC#`37^vT~zS6gadsiRme|Z(r@SmFV9I<7>vNhLRp?n?CjF({yUMOs7o -q3G`5a;29)i%+17)hUPp~iTw!%Hp&W-FZf*8mif*lg6XuDvzmeUQ8PD5@V8v<#oVVmPS*e13u~W)$#$ -<}+LvEi&M+??TbDAeQ5!t_tzpvaR6A`W`bksWaBg&u{si1$DT-~SnPGD%`0Z4Nh51HcP~@GA28^1PJA -PlwT%Qaz41r87m9Kg16I|cJ3^6SB2*3Bj=jVDpYOp$-`sRe-PtL=(rtO*cX>0Y-^#0zvq@QW{mjSbqp -i^RaHt~9BehO)+0JW@k)~z{n=ZF^9d7Q1lMQtxte`{vnSkZ5w(z)=>`vLy$AV4Rjo^%Tm8X|0{>QB{O -0rw1hS6odU;_;rH$Pt##QjI2b(&@O{oMU<#VfP)sprX|mgq-xHsgxe#_U&6raFhJ>hrKEgE1d8UD|$a -NANI1^!F;uOx|P?WS&f_Ej^v_l=U%y>uw~~-Z6wX -&XPVMr;18s?Wv2{EFf3dWYdDLG1T4_KjE}qj|Y4i0h)5cO;j{A*jAgB{Fc8m@hgE%|F2LCGlqw3>ccW -G*Z0YpQsiFvt8iHH>54j3Kg;Q7v0yo^w=p4J4XjPjNEtlI(h38r)}H9*j_Mhn%_+Jwr0{2Z+ -6I2654`Xvyfx{Ub{cQ0SkJBeddgZp1RLhO!fz;CLtlLBeAB;{#2;(wDx`iD$*}6ugn)7l3MSO$FxYq0 -v@?g@leKr!TlWr2+JOTM>h>!m=p(LX$>aE1Yo?t%{Tz_EM`|j^8c%jSvkBRoqxWaWRIh9Swg)+JfUYN -6931v`h*=QWSzC4BS9O_?o4eFcSIGzZmPY%1y-RdGTEVYxBlW-s@Lt5eE+wh*%*N8w9i?Wro^kv!lCe -+6F#!I}K*Rf;Vz+tv45qFe*j)H;LLN<_FHMU)4^yfCjX;Byy$Tef*BP6cX*4aivyZ+10#Hi>1QY-O00 -;of09ji%{d6_GB>(`9h5!IA0001RX>c!Jc4cm4Z*nhna%^mAVlyveZ*FvQX<{#PWn*=6Wpr|3ZgX&Na -&#_mdF6fkbK^GB;P3h?u$tN$c_MSi<0N0o7w@h{mSm-A -ZNxPb+hi$>wbrNxMW<@P0FfT`GUiA!Sg0?OUhu*Rn6gPc -=vV<+mw7?6%H;nlqf9!B5frcUj_$dMHkt-O9H(2k;7Ph2g#?;Gf5k7VOOzB67HPpt1I=OjQCj9et~gE -Zj)4YiXZp}n&EM-FZCB|=(zJ1TZbmoayKhfnf#O)m2bmNlV6_uOa-m3`my)?T2MQKqyNbe9X9KGSzXN -%I_?avaduku1Ns&c!;GsnsOHaHn5p7zBe1#hWaqMHyMAAe$dN7IH3ezLjLwHLD3S$CX#jEh&ryw+T0K -B@7U|rK?#~pf_dY(t?2BQU;L}Kz{Q#Wqb7dZ?IbLOo=JQ5{;mm!Q=oSk(#0)EsxSUk@MDlO8?Q(Mzhg -L*mfPoeoo1i&EnBZA~S^LQv9h?)8dWbdfc`50_AJKPjNj6WzI2-5|vTp;E_UIf`9%7K8e2XkC(*hxr2 -YP3vCnuYj+>-s)T7Oz2iVE4|6%$g%t(V7;09tZ!R1wjk|?I>2f;EBs -L<4vB)AWxINUjGHHbQv#@MS(ndw173xJR7DD^w&HEnK0jJ5eHi&jzXEh;@}Fv3ytsV92UXz2=MVdPQ! -b>D#LWM%Rm?vq8A>CcW+O=JBI&#U(Z_nh>ZQ$?Hqph5?TD2``rKs0$Z#7j|>9C-NBL+a?uydOp;s*S3 -!0M>;Y&8$xdv5&Pe8Y5kv`6Bp_HJ;J~yeiLflTLB>-E@;nWr0APZM3(GZZu7I}X683j4dbnEMsp`=^> -JvB^Y$1bK?->2M;&+gSz-=MI=|iJ1E@`{y*KxFtcoc>PAc#O(%Jm7~KC*ta#D67CaZ}FYC|~sjuoNU$7hqNM$8qDcWDT;j{EO&pAz5 -86a^(t6PRfc;p}q70VvRdIdBxsi+KU$B%rNiJxafNwbiS91wsyakqIbqtMC+!sbz>pj;b?)^}xM~*!k>fp&&XXD}LtHG_f`FwjbnT`OB3jlpRzP`MLkw%|J*E0`BhG$~*H~2$LKMt<0a5#cD_zW|- -MO?*ja`W@;_`}DU_&B+`7{SB$BfxO*{%XX-!32j_gYl=nxEOpId>B!$2|(g`VpGOceEB%S$2jf){u|E -5lWUyPaB@Ang+Kc+WjhccnT= -iWhzn=YE;oC+Rlv!@v)~uqsr85IY4~7VLZ+(BFW8sNMhxVJX8U%EIK18>x438Z7)UjzE|*1ke&uqFxp -f5d+og-^-nP0lK`%c4tDo5;xk~&o?rRmQi@7jhc7VpXSZ$tf%Osu1HB5cI4MkZVEA`HzS%r=!_vY{m0?s{jO|Xx~LBH|!i%!P}O2)wb=i--civa`Tt_xhB1@N3`tDpbc>AVut>oEvgu< -pSxNVEIA^Kmfs@!`3cWhFpQL5};#cz*qc_)z_6Gyp$i$XiUeL0usEo7O0=9n>63zl4 -Fgte9OFb$=dg+gGiwDpg#9qoT(77RY1<5}GXTJAW%yfQY0IJj^EpBz8X9G35Q16-g-TWv@F5tLuj3ei -Em9D&;FSSC$I-on-)o~uyMOt*yt|8~i;9B)0s~7qEYOHmb^{S0ku>L9xwKIrwWzUU+aFg#i~-VJDRb~ -rU%{_mDid9x)ddzbnU!qmkkmnc5`jhuZo!;K0b6^KCa(+B<`E4Fu#m|4pSG~cQOq!sn&vq>eQZS{eaU -3MOQ9tg(X9PJ;Uh}$a{^>5v_G^-L9r`hAvko9QX4qX6Hs9P!eRmj2S1y1Q^o<=emSgu5=s|Cc1{-T=* -_FwH8_ufDfhJ=n+dTl!6tJtKyz8{*wrN#q#YR*WJFcCPnr) -5V#Hmbf6Rpk79o`tbfWn5Y@}6j8|D9}pZ+-zET!{AZb=gb9!`+SEmB-0S#Z#brKdnH+YZQaIih`1fx? -T*`}T%HY^I9el#*yjEM@6+A!(=#{uaTUIER0pw4VjYkd`D>N;ef5GpRZS3D{_u>p&&2(4vcF)Zp34yXSCy> -%_d%|#39XPQ%qq`F6IWg&DAR{`r>KL$m`qS#lLZ$DG(plO>Qg)z;DDOIlLyqK)nUOUzyg;r1BV9&4{K2A>a&%Tx84umbbK(5i!@AQAKaRClz{W|6;$N^CSs5xHyz^UvMA!`J8`xwlkn`fmSy?vFjK6=x>_V9a#&mArLVi6R9KQEU{JiVtsz(c(&Rz98S%F -I~?*M+p=Wwk?&PdKPHo#`WOnHQ{saTEeJ$2bqFdZ_{)bS}(sG;359;__8I)QDVSIJMfvjb?5{!n>1&O -@O1SGsZ^j{xA<5r-qL|jIbjxWR?L-j1$T_h^zMVAeS{VHgr#g>pC@HA`vv;1%Mtc%kj(GU+XSKw3Zkj -XuyVPs&~D5f)o%pX~&}?Z=_{{P~IA#O}Y|oR~Q90n*Z3q*zTZcs&Q1r#G^F*=+7`>vybYBQ7&TgZ4iN -ENHb#NoPd3RDaJ&A9=yZK -)|RS+zL0vNhNPYqx)=4dXPIiw~eQ3zN5+6Kmhj#9DjTwHouhf{q87x1;H3hI`35qc)D+|ZL4tCYv>3%Ta@~;0trNR2S2C=l=`Kh^}kC067Ux*U@*Ek#S1i -XPJwAY4oGByXF&h|DdG99kU2USOxL|C*xU}#O38SAL|pW&RkeX0{ZsWft=dyGH#Pjx!R249v1yEdt;% -+LEB(V%F|{5%TF|mI`%vpzyGc`o=^u5Jv%*9^%z?!Nml)JGr5fpDNdH?f|rjd4L*+HrD@pip|OgxR-K+j -E?7xTg;e1S3CzuApK+#;>mp+JxDr!rekgpZ{ox7^@2qcoo%W_T$wa9Tu6BkL4o-+MRl|v2=x0ov03E= -0}#yC#N+a{T&I}Ris{y3s}d3?9FScOBbd*6XiBDL@nq3*17ROF*G -#w?67P`hHLMFW89t+wC#V6vhEamM^?sQ(CSbQ`NBs1~=nl%n=WYXdcA?S}k(Madpc4d0ickJDDPCp;X -f1bqZW7sPK`}QVz9J%c|%ifdMGXp-2r-|BzW)+l31h<2pNF%QX#v7UqK?lw}6E -Ehxi6CmF4z0Zfgkl6yFvB$k27*w^?V8p@KyqJ~T}9O7sIQ<-U$a@57N{dm{q-K{Xvn_G1c?vPkI;e7C -rMhg=)Yzg)*yboE8)(F9Lhz!n3)b^LbZtcdG*&weQyj9lRU+dgX&6w*Tfen$>98W672ZNW58iZ&tyx! -8Zt#zuXP(XNcSX?oeoUVph)Ucb0T&jFyk&5x3CKO1kAh~rcV!Sl7g|K~{`m>6V2}H&IoWozia&e+{3) -4&`n{7Z+zzV6lfEs;Dk2hs1AJXy{-h3m+tyc($fGgEq1$5@F2Iv^0kFc#64V;tk9?=~<~>M@ -PLk_5;Hd%#qBt;R1^@N-!mLF;pwr!WAqj`WmHv%dl0i`ZqUuyT=F|IUE$ko1{7n%#{IA{rv=<8eH96* -B!K`Qb@PcR=KxTr#Ab(dQ`Q-WN`|BwB?ZZYLxf;?)#Pir*LDDpeDLCyuCH|Qdl=wMhV7)Y$WVdeQ$Vq -D{=G4o-FG;N7jZpaP88ps3bak0?jC=pf| -MSfOQ<<|vaW<)fVV`oGZ*&n4CuakpBckb@p*G-x8?;(g6CwE+cSq$%&(LY21{@s(c7NIMQ-TA0sw4L&9oUba5gVLqJ-dc(dD=UrZ=C7KQ6lBkD_OS? -it;&ynkW*ac1P*m-& -lwm2O@~3)7`%Ghy5^)T2B|I`vB{QM14c2S9z>-MIR*JzwMz7~wob)k19X+5)iIo;IvzqZE^zHiUHMs_ -F0iQGp7q#7*ZVF^@E>fTxmtHrDx_unr?OqEXe}kSnh&IKXqRH^5(M&t5@W{v7u+eRwhHxOcmXQi;=9G -g#lA+hdJ&i>=I78mAU?YPm3}WYxIW9SKgDF{^O~8V6k3c0emaIq$8NxQ;xCm7?VC@q6isi-<9jffii* -j_IdwU?`4<)puV8d9A`l6f7yA*9)Husv@5|(7?(V4En)pZ?~>pSzFF%@AYU6v+LHY{Wsa@PTM$$6GCY+N#7HG31Oqx$mnQ8`NRh0&T#;tCj_CFJQM=7_(svr!x#TJMh?$&`}qz -uUFftj}=20F(@-J9ORriiRB9gmgv$UkOM5bORyBi$(Tf*z>S~x>##zQ-y?5C$^}Icp$eW2)G(4^RzJ; -4_(?)a(d7CNvqP(4gC*QIQf9M?w+&O0k67lr4Az^7e)LS+27i=tr2Z|cpkX-0WjL0t|s*j+nOLDMx4Nf4Vt*z|RmRD-3=P&~f)PXJf? -&RD5`1Xg>@4x*I+YF@a0@VQ~;AVR?1z`oer+)iJoE#Xy1^5g1`c?Uboky9gu6=e)3I_B%1{p7_ctoHC -tf&MmzEXADyfHBzn5LG2?Fi+DR-%iYaxhH`tg!95GMvwuThb?-hD06fi3c(PL|3m{7R%!wo?y5T@z5N -*NHIO|8elY&6m5z&u*JBa^U&7b&ow^K?JZOsPsfC9IcG(g_w=+QfNcPrAfx@K3zcAkx{TH(hlfE+((k -B`l(x*hM;TQI^!~;@K1D9W8XEaIT>I4Vg0Jp)wX%QqE>VduuuF85%)I7{&K;womZE%ihrWs#B5wvcFe -r}ka6uZBwjXHr6#=3%Aq|bsCg~)_0H@t99ho%zzks6oXV{D6?3NB;stmaWUHoJ!=aQYKDAA{Lx!et0O-Z8jmg|5ocHwby(!77wl$oC4?G0BAt(TrqpQQA1$tq -!H_teii)xuk84j?cwO6A&NnxAfLg*9ti-c2H5$dAZD^e;9aGH5$4^gOBOdAsxT%sa|y_%JZYLCv`BouBOdQT%=m2haOFB+sh|608}C;N`E -GIsv{+86Gp3SVjhQ+M5agXcuR_6uJRQbED862DHm!Jj7!vQ{C61ba-I+C+EaQVz1NZELK}jP5a_GIPu -yR_NeHQt3O}Iw*`TG^HkV`DR$W9?Jkl^UJ%Qn5s*@&_L;{gc%&cm8*TbbEAN!uM>N#H`nQP3YuwWG0@ -Azczu#d0To3t%0f~Ur5H2$!k7?4##ca69fi8*0Q;D3CEYZvZkxpbx{Cmv_(fp2@hy$|6!{IH@Wug08< -|!Fw_aae_^N102g!z0_jiu{LXR%(vaSORJM&($u?E^XjKZ-zBxOrJJur(_X?T3V;sb|m=7P8DMwA<^Y -SC1OFv~+{zO8)n#P*tdjI&;sC-{!9bz0zsQs_g&%B!SI=U{_Od$j;`C8%f|i)_Jd@F;_|+DZ%CdRG$8 -y0$q=955S@``|dn)!oIEcx;r&;1$I11TR-+!DC2mmruu-Sgdowke|vsCzrqVn6POJJjM9NbEd7jybhT -V?|2U8m`{i2-Wa|#si?5t0;zB-)#X5nny?fG;Y_wuJzJT%R$E4Xq&lB$6n(h)m+H3~#+?mgj|Jm5`ca -t&k4nWJb-(|UdWEvPA5$zWa8QF1rZ3bi&ydTSHJ&7h3?XFCw^ia$%0autc?my%;d-V1US^)*p(b8jhB -upWDb1GqJfB%PW1*F2inIt~zpAmSn1j43U_GP*`-Ql}T|3Z!3*o@XYQZ$!&=a=+S3Z#Fl}QD`o!v -cMz{cgsAPnXQunl1wem9gY-MFD(OfHo2IbX)!ZKaUI4Il(a^S1n^%l4Q9+S)FS=MV{XEa;-t+@s51B) -u?r1HVf~DgID4{gw?kD27XA&d{kbSS=J%Ua*i%c;ikJ>zH1RP{V}DYuE+jg`J-cfYkw>Pozaw91>}6{ -(ciq+)-q|;2qyFOYk_-H4@p&ItyL+8cQ$?Za66)M2a1|mvw87HH^*I%mSexT>cDF)_mZVrPSj(wbMw^ -cuK|`UotuTwg3FjW@Wb~e)Ns-sQFF<-Q~)RgH)Td?RT$P0SaMC4lwk -?G*t3R04hxt1K=4e;nztT9n0r0~$!GaH4VrjRoWy6-eM3Bnp*m15g*ZvMfP$;tR)%2(bG;sR;l+0Evj -_(TrW98D(@?S*rvF}aXCE4^rNo!j1&$x78Fv00ku9)=On^izM4A?swM%dJ?!jz??wDjo9uJqEnLD>ls -JMpF5XrY2`LxdncpAl~RB&_((NKh+`DW~e>g$Sk0vPOu%ywaTwFa_y^6m~0(S8gCso86D5nBwv3^NG{ -{x&iu5FVk08~fbK2EFJUCrIiCT~NF^pOlM{p+Uysop)Tw|WCmdsvj4NUsm44(Bi*5*36$`NDKnywoCV -+g9VHjVJz-r?Qe*RTPiy -SqrXc;c*NY&vP`RyXW4R>BsjLu{rKjanvA!)lw!0MEYyidrxuuj!`Z73Q}j)!olL!)jtJK?P7l^fpcJ -|01e5fAW{7&lx5En$KY#$6`lODc?!a(@(wPLBUh{hn3mWDrgFuO0LQcNCtssS!Jmb$_*h`a~V(Dx7$yZn -W@Oc)V*pSLmu6TD!jk^2s5{C%*^s0l?1%Yu*;x4wi2*eL2`Y1m+7|4|p?e9Pyb8Jj^5&U20M@Z?)A_@ -xQav+PtK1EH8;~5O~^&V7sgTdwey#Uex&J$)jIhsnFKIlxMe27gC?=$@Alrr;xx|Bm<2$pnZ<1oO4EY -hbH;h7hMivd*02u`tS<0!8vDhz6r9s9R7K8zs29&Vi8qsR(ET?wzQ6**B_$b&pH~=sGzw5qe~h{>O|@ -u5wN)cVW~bU`0C{3yHj!{(c3zz)p!lXjKL|YHoiyW(w4h6XAZn!U9eGq>E>M2-?mX#akSr|+I}a;SOj -aZ!t_NP3_xF#;Zjn4eXq|h8$69wQWAk8LM;RRp$*u;1z2NPN224pN{!<}%ea%UrpDQ6}@Fo%&KE$}EL -x3w@A?TN`C>X@)h5EG_5wk<(F*Cjt@4RwJY(lX8tA6uM8L`fB0yN>9QKt?`F+*?rWZlu~_+Q$?0Ya4+ ->KJdBE;UfR_{%vJ^KKITAfR}yms-4k_Hh{oTcahS~#6KW~cUoN|u3)18>y)K_Ol=lg1A4iE)-%rAiM! -H$RwP-R_@2v+tx5tghkccIMxX{`=?g?dYO5f0U4;y1S~`_2hb_Ml*oBvvtR9&e)NoxNULd% -F1lwMR=&IqxCqL46QEQP_TR~5N+Z6X{tW@if8%lan!Ah@Rlt*Z)K}uZfF~rp*nI;R(nI`{N-#JDDLeW -AGfM+OQJt1&r{(Zry*yB7?Aj`Ooe>q7nO-hC2oFTDX>f+Tx1g*e~CEYumpahRc|;QWTXvFx91qQno&K -jT|T1kIZ;bUWLI^VKw_>=!ps>A^m8N-P;p~@fwotM#c)pxe5#BA -KB=Gx$nrQ~NLxLFHJafb?vX<~mtpms15{VaoW~ER7pv+IjJ`Cp0w+HaG4!|P;yQq=x{z3BxHc^a` -hjHiC(YanGLSvS-zNkHl6luU_NUu3cx1A~tM2k^lsjqT++zP_hgEetw`@skuK5*zS`AAz3SI#b*r}PE -J6f!CZ^zwUjS}xJy(-)L!sQfd&P`GH^ty96 -#7o}j<))>##&_t7q>dF)XA6d+^@2V6P8`obE(ZQ2Gg+<)Ea=LE%4zXMd63y(L26tD-FoZ1augLgt(8; -@R>;ysQhKQ!$@Jk+j^}v$CuL<*!PH%HWz9Uxr)ztpAUPyE^Rp#~HGGHqMR8?c)XW03Sx_-4t{pwf%Q$ -*B}Foh%x05EI3u-?ytyF0n?WxCAWCU(UaZvuMTFc`M#<~ly@=JXJ%q9w~f>JU6-Xe!$)lDToufB_FBD -V6e4s!6_z)_p!2mi%>b0-~=C4fwiDohfL2enp`N7VgdHMtt|}Y30A8{G@NP*9{eheDO*}Tle)I+Pcl! -h}%r`dreJ>0|X -QR000O8%K%wh%TmZcSqK0Cxf=igBme*aaA|NaUv_0~WN&gWb#iQMX<{=kV{dMBa%o~Ob8lm7b1rastr -=@?+eY#`f5n19P|1+!w&xTFA}U(sfd+?5jg!+{Q3Ujg99cvtF3Vk75sLozo7o4seEFO<{9uv0^Pbsxt -e>#;#hRre%gei%HEO^9C#?iw7zQ_oT(YCcngS4_R6H*wt7>td%UqPaV6v$yQ7g6=HQWCD;$K_Z{gi`b -`&t~?DOZQQ44S$q^4&io=F_vN(%@aK?C+0k!Al086qLk+iM`!et{5+~Kz}?#Qi4IrcStBl8vCY9l@=* -i=M@9w3pGP&6)y^Qecp=nqhz96Lb+sMB_Szs0od)WB)HpK;$QP9Km&l)D?e5R906@Wd^+g~Z{HxrnwN -J#;#zIl9=K|1kP&;M7%!v<`beTE#0iaV894Ab2-ck~Zg0t;=LCsvzS7cN@%ZvGilW=wWeYe7Hv)TRG_ -Bz%6tI_Xw^L}(4s&>gNYUeSXg=0E33bVj7zOAtR&gm$qRym_64hE%lL&$+VkOEG#{pwxDU6UBQgK9EX -OJ-^SR-54Xz9>DNXV{ebKr_h4?r9-+>p*$Gz8H~i2#ZMMxupL1g^9mXl+q6+MmcZ6D63iB^iyCh!!ey -Z-WVVMjlR!90Or3^9+OM041G`5- -2yVZygGJHQJla)f7#728#{e;#by7su4JK}NYxunZsbcM5%r3CrVE} -!MV^*EjL6iZRmUUqicHVq$D1ZNh-d!JrRN+HPJ)2XCOoN)(`0KO`rnF>JS~_Bvj)-R4;8v;flw*LjZq -uYbQ(Ya-&mEKR@F-%yrF5=uyvdHGJqFeK(X2vmL6N67(bMge?Ep;GLPUVBy7qV55mc -wM4Vr4duJiXwx9_L2_lB~)}O%k=!-AQ*Uu9ns4|FVWMz(G!}ua1{f?)g?*_l^oBaHlt0@+(neAOkQVU -pIYWb!D16LSwWib(IZ8H>EGu2h_Tn0D2}2gvY331n(7KY3`24Ur7DTeaHXooqmJTaB9&1tE}^{qCi?c -f=+h;Y!8+eX2X!o-XxpxB$?GUCSMBS_VrFZp|JFR05E#!ADNxSPKpvtX-0EfRuqil|A3-##gS;^dmS_ -~Y14Aj!p!{@mbN!-SUFq*_w=hPmgeazwuW@HYfEI(<2(oZiOs&_7y@gYU@97#v#QVrgen$z|P$2sMHc -Ji)Q+N@f=Nw4IYYBABCQgGaX3qldu?>5RWgrNGw7{ci@=GV%qObqb;L+)%-yd03U>aU4%-q^4%(K2OI -`&(#`<%P%Ws*YFQqnIO?D4GUWv-HBA)(l>*affeB>r4{Jkguwt>wo3DABQ)2y5;QGhk@JY7zD~!)1rs -R8TJlTCOz3Wz-5VWbMgV!YLuGw2+=m&L;G{!V|qKa(O_N5#F~7*-0{ucJ3!VgGm0y~(XPAAJ*Z=Sjj;cc$tmQ+l+?BdQ -R_VG)J1mwMb|ijF#0u$hPyBe+bk#i3p!{4hs=%ri~E?CVlBi%qgUgiZVN*selOQ*gTvaJOZ3MgqG#k6 -&k{oW`Y{DLk8)4U+vBLtR|B_~~se1267%uIeTCe%RJ&YKf*-VusQZ+<92t4`V4lV-4BZW(-cCCB9fQ#M -pXji*uIW}G);UUGx?ZE%hn}x4AhJ|j-iXj#I`Ga51@%kV)4X_(U+% -&~9z;`Q3$_%Xj6yeP -vlLaS${?O^qN@vEBP+CrDJC_4&rkkaD_6AZmk-)l7c8pq?M0`8`2HiS$xY*9e2jL}JfP8k>tHUnIJV@ -T8(KgYC|y-!m6X=H3g*9f6m!>vyKbXq2EiMUOHrpIJE=74gV#O$Q{#o!kJxCLF4G{T_GR*e}LQ7Q0(~ -QBbEjU^*teouIsZJn7+t(xa~{hn)u&-fS@H&259@ZYkXG>Ej!x)tJk080}(6x;E9dy?Kd_o^Rf1Xi2n -)Ad}rWF;`R6xpu8B&6{C7XZuwjQ2LPDDbPPB#)jT}&{1sfAiAZty`tI2gLGX}f4OXq38^SI-+bMz9`N -v>g$J+~7o8b3l52Im%BQEb49(MyZHrdMWv?4|a#AZme0tZ5gGHhl(M=7lmqO{iema+!VU*)*fFr%O2; -;C{3y^sMIONdol5IBZ+i_)Z(ycm?Hin01E!9K@67)xsHlO0`iSYnW)H_--(#4H -ITiY5x^o*8vF_c28I1pb^n@A=d2wI=e&goPbFnk1Zuh3Z^x)Ql(boX1ozCL!gbWC|`73Tu&HzGBS+RC -Nx6@q%jC0Iy!z+&u&d1|H+`9wui^1p?q$g6F1XnKet>%**JtKF{6^W4`GoZo$08mQ<1QY-O00;of09j -iA00002000000000u0001RX>c!Jc4cm4Z*nhna%^mAVlyveZ*FvQX<{#5b7f<7a%FUKVQzD9Z*p`mUt -ei%X>?y-E^v7R08mQ<1QY-O00;of09jiu&n->v5C8z+L;wId0001RX>c!Jc4cm4Z*nhna%^mAVlyveZ -*FvQX<{#5b7f<7a%FUKVQzD9Z*p`mVrgzMn8E^v9p8*6XkNb)4-5Cv;+_) -lO3&A$e1ShVr-uudF-#hs&2R4?FTrsySmY626xxHtE#JC&1SQ)+}R=7dz;vyBo1jvtmw^$|00j9ZFfU -V^0M=wZymeVj(A(Lg&&daxgjy3+kzjF=|&;h2Hw75P%94TGBDlH_kyTPmODxwJ<}mO8cym*>sOtrLF -Am9ut|KJwy#P(%zO(m0CLn9a{M~jRfi{{ZFl~o5$6=R@ya0qZBN9+E^xO`IVFH9ZFLpL$O_{P_%79a4 -4bSXr?L|AXej{<~hYlg{hww|ILEw+(`EP;!xQj^3YLnAm|KcP3bD1eg{`mqw|NXL -t-(8nR|KfOa8tmw~b -zK+)-jP3_Z&>fhO5_R=fvCFmcSu)^@;KTR1pPBf)64JB%DK+feX57%J=+-E7#s7%dw{fz5oU13;D&Z% -`J|?bv&pbK%s01TzrxABdF>ledG>V{h?)%2f2@wPT074gn7W^6NN)#t<9X0C3SjrxO_Hkb-f6iVZp>E -+(X8?#sdj4ICvVwsGJBYwisj4@}57|A$%;wjlYf=RnM%n&3Mdo5fAy0T`?18XSn} -#eKo`A`qL4f`j5`qn&hxI|xX0J#IjXk-=4zR0~iH7M2jc1OuDJYfj~EBFC*xjM-@UFj(khY>@e4_H8` -Whom_$;Jeu&598(S?0!iAVlkL5zmwUB45r`7&*SN^L-b$f3*9itY{AHl@8*-S4(;QqHn|^;r(ekp;7( -^tG8x~Emq2(qV@wlrW8FZ4cltuRh0npwcrspo=i@dSFQ-U*G+U4XnGY7rv35ThEXe$RF`pSah%p4x)A -4k)08aXyK3#T!Gqe%?8+;Jsb}*SBJ4SDC51K4cR-(=3-xuSrw@Y$6n+$bmywO3z!OcYH>_91+3;toP)}H62zuzJ%eq)we`QpoVqq9@-;dt(&_#N86(^V -wCVg)Xi0iQj4gY??6$lW-CG@r=5g*kSWOc@Oh)$MkD(|R%=0W`E<1td%kAziUqwqb8mV|2r}FrxcboB -ZvXG>2vYvEBT__%jSLX6SklctNWo7T+0bP>ir_R&qV-DrrJ?m8n4fGcQjwSVpXCi=lDTMvMu* -mr1aZ~84zza!?!Zs4!(sXhq-hO1BMy)$@N}bf|vwM3F7xJ|AC&)=8xd*ODgE!$&LL;j5uI55w&)%ZLz|p--l8wA}_ -2O;}pmyrXqbOSQ4Ezji6+qD1~FLsY7wHXCziekR;U6aYb>0jq2z5eQiXi(jGi6)mjU -$DCDylO6sETIBIOpJi2bA?PT|P)I{2`?P^&HH0Y&=>>yMqQ>(I9ZW>tzrp0?=2B`jKFg@z0{Er347hf -r%R6TQI)1!)H@;lqUEW)2&0Mo=ee(Sh;FLVRJq -{dtRrxFojmz20>uHsZDorWUd`r(lg{G8xE1u&izPfT0+7twv2~dMkapJwt!@CE#`?9Z)>AuP%c?!

      RqLHYZCxvb8RPlT-Z*2=BT?qV1ehyY0W9=EyW+3|zVwq)vDi$Ql!avgHCNEcZ1U>5Q1>KxGKJuw>*#-Zl9(&D?2>b)=>=w6aE -Foq}Vv55frllAx-5rCNK1_ZxDL@%YPo2ONd4N+WS3zBy@ -)mjF$y>yFQ?w+-+aVB1TPYt^YMYbwJ?af%ZIV+@Wm?TGL9N4W{7HWWn+bLrKHAWLCwu5 -b}i=T$tFcB3!D5Do&{1zn(+A!M7u4BK7bUoKs8n45c!NK?u4{UxQGR_sxi*&(RhYTLv>K)GNvP|+)+f -$n@}Nme5e`toYEkd`*ZBmz)AOQO8F*)YO<}$I62;#*Oc;nU!KUL4B}|#L6=Prnm@~ZA{!i&MJ8NVtkk -8Bc!J`4<$`3S-r)m7Zx3l=3T4)xqPI%3?K?k3!sPY2i+S!GCdSYQynPq~s%E(*cO*5S|M_Ad~C)2 -$P*C`+~%`5NtDXw^`znr=<04$6~NGm2ovxm{UumaKnL4az6Ca!G22r|#m`# -9t%7x)om<%sKY&a)xrz|DY%Qm_UD}72o3pKWI++=tb=GmLt}Qs+_px7R}Rrw>)B3<6Clm@q>mHWDc3G -nrW@%2a6`0nAk2&E&FAAMJ2{l$<-;W86*VMz_p+}km`jnU(yS -CHl4Uo8OkCQhe&M66-Fvgv$hmpm6}fTVmGQ1sDfu11$m9mjsa?MA4ClHke@JsT7n2-@7Of~cW2AhBuu -K7h00cr9Q^XuXD62xq>az2*jJfLa?62RI(I62r$^Uuad1>W~)i-&ORoPKq#-Vi`0Hl4)gWTl* -4?jltVec&5;>Q=2JoH$Z|~0h(~FaNP{@Zx#l2BZW!f`H~&u3a*cR+oNn{e=+l3sAif!vZ4&?gxA>GT6 -6vda<&eRppCyx>bgH68T^>|Av5~%t6qOSkwV`r{FyXH>RBm~tz*0kbdQ2h3R~49a?o(_mAQevI3T-Ni -bRJp}zfY*TbTl#t5A=6|8Ri_KYrCuTd|Rf_A|}hws=-4ocSIlAd-^xll{;z -IMhlAf9djoL$^Oi52T}4z7HQm+AVA*Ft&@*4;W!x=yx&m4p!~??7>nblF9j4nLejo^NisS*eHS;*^Y< -0xX}h)z2Qpa&4Xk4LuXz;YOqU&B7Nk2Umr65}Nhn=Ypel=&oT`?YlhtE?$i9*&vJ&jRNjV;F;g_73=>Jt?KcJVdVYB&-w(B%e_CG1}$CP>unQU{(~>RaiG~5iPnqDj5v*+R#A*%Oc5&h${HF -Cdu8V_bK$J2`#wGvgdAM9a3T7`_$c(`9+RMzL-8vuGU=5urRw8*8lm1$Pv`KIZ9wL8OzrF8Q)_$xs+7 -R%=IKW#*l7JC~~>mN#0VZ`&1fO{S~RmRc4VVMo;CPx?22-6pnK|dQ-19=WV*LF8Ue<3|4lI0tiH5J;S{ -{4ilU-J4{|^WU@{;#I_?FDHq7biC%TMV{cBeV?O^dR}XR2z9RQjf_n{r@3p7OQc{c?0ASxYpxdau2!U -a!~h^)lhbGV3o;L6EK%JilRf1KIv*<{C73rU9OSgBHCER%`^ELUX^+dIqhhymdJ-L~9>f&x@_t{HfTC -l&_KnrJ=roBpVNMC*Gk?9zPI3B>nCgq0iLC`}Gk%yZeC86NJLd40crHgR=H}h4wr1HQO<|CZ9il&Xm_ -28!VmN`^>b5A*QCdi<{btT6}Y0uHPB-zcF#(}ts~QRU|<3Kq6~ZlMY@v%yetEp;dqPN?w`12X%MK}S06t^#7SKb_-6D5X0xZH^GxbwC*_;30 -R+%C7T6aBfT=h2B76JB0s5B(fWaEwlcQhCg8hO!^cwDCWn+#&pJWp3L2Ar(*p;V!|C!{I(9eCLl+J0oy~egyA@|vMHb -@>Q-_iDMd~&tJFMMGd2N2A>(6bT*{?5e2_pV9)htvvy79@&Y6#3zolamd;AoqMclJ>z?UCY~_C=ZNJ! -elNv>0ey*VE8|Vp|j#FR!pW9p3PCdEe{!WrE9}m9se~3D|{)r8tBIA^x|^$Rl8dmw}Z2*b4+-zTJKcL -AuW!gob?X#>7QN+B)}JZZVvG&M;TS~(QXIu_q4Z~&nDmdXT1vrRU3S+;&u%2Dn+bgj9%~as2ZU}D?re -XkCJwyQgNQVW^CJ4SyY$1DE@%R!zAe^w|N*Md0I<1?0t1vI(S=FDf*Gc<(DPB_;xx(*hk0V6JkME= -r+PnDfT}08mQ<1QY-O00;of09jkEDyO_z5dZ*UHUI!Q0001RX>c!Jc4cm4Z*nhna%^mAVlyveZ*FvQX -<{#5b7f<7a%FUKVQzD9Z*p`mY;Sj8Y-M(3Y%Xwl)f;VZ+qn6=e+AFQq4LcY(w7~I0NIeH>0Pk4DdKd+ -&?^L5rej@XQAbK~y}cM)H7ZxuO0&g@H>$zOYkb~lbDIj-b}Nc1FWUv|E;cJ&DlvzxS3)mE=_whVX`^20wpvKUNvr`TAT -H!w<37~_M%EfOS<6b^sI{sZf-Q1g<;nmNMFl$-R;=}+E!7B7ljU05M!;O42`FM+UUCi**20jJlCcGKG2ues%2XW3Aout!p0$UlyU6I^+KhiF$k(yvlOINKR@$td9$sR^>0O^Orv$# -4e0L4DF0+_**!L<%^UyLsCB(8fZWm2=(R|_yoLLI_Ua4-+)PEx-YF|e(QHSf2f#*Pn^8AyDUDE{YlFN -88NetIObT#g|w<)Q?7GkD3)&abX#+10r?0f1?)*EJxQj0F75$B*#m#rZE6Zw1 -b~cy)1hdVNl#zg0>4(_bgQRrGfL_T|ou@f0Dxyu3br{dzPRjYbQ#6j{E^5TgvK&*s}k*@XUY3~*s`Od -F#h@b@ppovHxoe!^7bMv-iRQ7?(I`R}0of|yy5h5-}eZ|l}JOpt;vR|@n7nKa*$BIL3xl}V$GjcQs`4 -MdPK)H6bWIfOrhNw;UOPRml~veD+HTFSNrS%w6ZstPG4I1zUWvT-m^=26quB`}G99bixi%8l}MeuC%r -ysA+DbzYEj04lm2Vyc6kZShY98iBaFjYl-)f%t&_!>Jg%j$oa*Ra>^mJ6X2MlHQS-D`DkY32hdh*EID -+B+bX8PwWk8rrQl?I2bhN@qszOFb`-)TGwE;SHRVuhLlZ`pKwtJq2K76VDK`>nS(0RE&C3|>E=2zqXF -K&9~vMy{lZ2_Rgc);)b3-lP&JE^zX9DW~ATBoew08w+$DMZ#iBl}=o!LY&hLd -JZ5L(Vf;27+Y|p5|G;k|wJ=%)AEhtR3Mo4lKA2MYB5Nf$x+d?h^2q9X+mLsOJ@gA^83N`qgvK&Ooym% -Z|Z>6Z09?WYQS9O-kF`&qM5G`oK-RX&X5&)hR;m#XuX~LkhIuN)}e3L%?{L&StOAjR7mhR7Msh_agdMG2zcq{OYu4zBOt&bP>)j7z1w&lYafAdn%TctuYnH(FO+ -jB;HV3Lo719`4V`T+#|{S3N?K%!^nb;#gBF@ei -VtCPKIU<6M}re8F-_!wkj}DO~~;N?3|S!OijG!S>9h#V$}OnmlLJpJ=$!SXQLu1#Mq&NsVI40djAI|6 -9p#qm4;Xgar}?AAU9p=hHDW@v~tXK5QIZt92LU0%GsKEFyj|IKSA -F6cTM4|?ujRO4LsJ1~GwPy-C6U-rd;D+Ud5TLN+DqR&hn4Ta<6nX*1W}jcWoH=D?@15(K%CWOZ#r$TFECsHh6A+`A66(9G#QkG-b_@9-T_{rhaoa -2(Js(FF!!x|aKnR2KEUR=z`kc`JQ)m9b-7GI*$_ef^G@+FO*G)mp+alT#X0y%8=nciO4h|9kQ=w)p_y -h5#(SZR)&|x}HIXMwW$AJ#uP9q=wXiC-*?`ysk9?pR6B!D+qyx%CN8IC@phoC^pUAGOp{XNcOunDjEl -8cLHds^Hs&I|+wb+_+{XvgE8zz&^ud^wvnvT3bzee7>K=EMTpHL&4?lgQJ$B%JhhFf(Bk`rH)tgPbrI -ShZRw@Gs-C_)rN-A2EZ%<|_dcS;nNDh>yKjqdr -JF|M{6SK%h2Xr8h>mb@F&Jp};|7$dRa_kS?*a(SX+rm`i{^d6`r5%I_Q4#GW|%RtuvsSAnU;vdAge*; -?iBq86TrUbNquC4NhW7-Wavn;A|*7W7zunFn7?zIM}P6b*8hnhhwPJ*@UjqbAYG-tJ5XVt?-F79YwUz -kz}wPDB6KOTI72upyJ|Y<-`Wj7L0L93yeqE4sA@Ir6RCn72i_V4^uc>m+&*m)+F4(;_DB8)%6#z3c6%%|F5Up+ra@^Lz_Kg -d0^bm$^pGXTlrn<`DSOnw>1X+PQCPAXuAk!QaHZJsUz=7D1>%O{Mc+5V&fd50FF`aNnWE3D3Ml{-IS-T^}p7V!9QZX$}v_hrx|1}&MXJJZ3;5=Ss+9zb{rmVzo8>4W5cD<(4T -yeiy~vtuN@S^1-tj%rRU=yVFtYdHJSRRe1+HtPpT&ie{nNz+y!u5vpKcVdSRltbNVh~s`g+3|@*<3O# -Ige_R8=#){e>syPM;VOIu{Z^GSMLv*HDlVXUQgShMur}%;i0ekXBdD%L*)Qu!WU(Wd3ftXoE2l!=^--cKmq`6lm -jR2E#(c*jv}C`%dT<6vLvC$b)bKb*q-#Ev`UFgSNv7=TPIc*+NSGv0;L0^4gak3|heK~BU^qKtjeCl8AtDA!y;)j{jWxT -({XMATQO4;b8pu1PmgOU7nCo?t(88P*^h;;btUmQX;-MY6Zl=x5#Q%oKm_lca|juqKBieHu6HsePpe> -6j%ba$e6TF#jol9w^d0QV$TglTV;bl#h@d^!D+9i|+pz^(h&Sm?lLBDi4#Bt -}89+Xwp~W>|=#oOYP6*5ujo3Xpu;Z$DaZph`5T{*(#|J!sj!*-YC3u0S4 -!8XHXEO+cjOHpguQ7y5Q@*igSbkAeTT@gSv`>J&Vj2UbjlsVx^(r0j0px;N#9|FsP2HvP4VWLN3WYsU -L>sk%LL>Z9Ef8u8!UF>w8s3)a5 -AFwLbg{pl3xD?1kC?FTIX)vL4P0?5sT|)kQ+wR);p(VfI-OKd!BK{D)99pyZ6muI8VYz -7qV;;MfAKZU>Y}Qxw7jW$4?lihWB4y!>TwM(!Ddeoh!ay;!eM+c0GqG3hV~)gDUPg2Bp?py{Q -x->!0gISG2db@rZS?2kTeGy -?O%5gn97pO%N#mTXo9upCjcw&emIT#0TCrf|ziR{?OP+zc`qG9t_z_Gyu8f52`JZE(@Xf!pEY2g -Vg0i*urTJyKt|5EjS6-<)|c#3AxikFT=LWQLo8Cf*??9u%&rX0xsblsBi^34pdDZa7`J -2QSRb*hY7bH0L%^EaaTQUVeQHs57e*a*XPssQCGyfg6|Qb%pU`*t>4b!|X2u -_c919`hG>0TMDUuqv#A&%poLEta68O_=r~vslR|U(gTu1Umz?~xq9FL2R%g=t%w76r8Zd@T=Cr><9)4 -6?qCmctUjxN6c)3Jc$3|enl1qf<`WW8%8jwgy9I^mHaThTY5vhF>u!!5q;KKg$1U%@W~(7?w>eFe -JjjE;~(er(!2 -7w&5lF$bR9vJ>}OgrSy>{m@BU=bzVKQI4In&_H}s!8K7(0#h^%rtGhmkbMvf{|oLNRAK;&h0pctS@!P -g^7``Z@^$w2^Q+m#<=Yvk+rK!;hxm2(WPB`=qp5h}Bus-(aq#I>9Qi>2bobZ|K8^-cKl~&fe&U9o!tl -R+4MxV2HKonk3TuTb%Vj1@td+rC;~$s|%C8r=0JQ;t3YrMq&9Qrr%)|7N;fS--!4*JdvbFqu$`52?dz -tak?fc2KM&T7aQnrKwvxGHApO2oKutW5t -W=fz!w7iuhnYxohDOWmqKzN_Irz9|?-#$)AIJ=rPIq`S_1jLwpYSoPt9kk8KTt~p1QY-O00;of09jiA -00002000000000e0001RX>c!Jc4cm4Z*nhna%^mAVlyvrVPk7yXJvCQUtei%X>?y-E^v7R08mQ<1QY- -O00;of09jjHlo_lhApihrhX4R00001RX>c!Jc4cm4Z*nhna%^mAVlyvrVPk7yXJvCQb7^=kaCyZ&`*Y -hilfUb)z$w$Ilq$T$PJ4Hmx>Lt-8r|pD$xhQ=z6?!?kj0uJRX%L1-+cf3?E?V8r(`*I=gCAQu)A0+b{ -C5WfoI|6aB+Wc -<+zIr}5Q?^N&{oASQ=rSHFw%_u}yEck#FJ*}IWA{_SFNe0eF(CqQg`dT}y7hSc%d(aFbmBOvvDe07F&-=9y!p}07lT#b)Do*YiZ#mC9T`Q{42bO%MXVqCs-IzJNyWBm>`$p==|dM$@rHKSK`C@$-84ne0vPs9lk -v|=E9(AM<<8l(~)?0czXEDF=3qp9jqY&GLPcZhht2{G7sV3(bf3;3@dYVes(p1*Adipa%GS|jW3T!;& -3v)M6tY|oS$MfQLun>PGkV{?3n4G(93NDAo%|A^4O3R?~V^ofZ8P@b44|{)9dy7{f}7?g2sVP%t0yc6 -L=;)^d)G52bP$lT7XwsgXu+LkMwRI5PQ9O0t)ZC^E}HF>AE6Vq8@?1rf~?yROm#v4 -eyWrZgDQJjoapy-ZL+fUN~XP@>v}oxkw9q&qF-});>W2<`gukCI&pIIQy2VxJh_uI4gG5d96{+qoo;-> ->K*dK}4LlD_q0O*^ZfIoCr(Tz(I0Neq*kA{>JfET}!Ns$E!Dr**kS1vN}LFHufA)^fBBTFc48bn^Q)~ -VBVhO)+t)5G5`$Nzn-SYCP$Y4Dfq$uO18yiL7Sa#|OQB8;P(c0ReWog>V;g2y`Fc@&_AHt>!$;)3VI@ -+$N)XxpU&zAnj9>_7(nKfF7yKja$V-!t)_tleIC@2ykpLsGH72c)yqF;ImppbmqJPjrK`h^FM93>df2P6XVF9-VlV(q5*=w`#BNRp2DZ4gs;VR4LqzMO;gg%Xr{0Dac -Hl8t|f1k;R!Vzp=Tvt(Tc1}cN!2E(CAln*nx%EUQn=^Xl|<0OfbGRiZM&T+UQKfK^H(->g3?zv~NmRa^_;b8^)e56I+!lSy1j-97%8zBxuj~U5IW5=*31D=O}T5Cp -`44fe*$}?C;QLptB^tAi;5w3b`Tp)x?=Q7Xp?%l!M4y%7XfsIp@5QYEtyk*imz;@aEwEC;=2E1In2x) -YZgYrrWdhf^3>7q@OYK3-H4#=bm_ft>4LXmXvjmbq<2UOex(+Slp5rA<@A!s3HFJEW8_8paEGh^p%AP -F<%~LXv0Xth1WQ|lY^l}8pN9sN28*gkwP4j$whbZEf}xsrdXEKuJIUKLGT3sJhXF>--HkA6-VYKC{1sSvfn@>YOtNsXG4WnkLx6hc -P-l+~$d%kD;mHVI)&5Xt)kRD8QlbLf}{7ER`S=ol*HI@rCJl?Iei-blResxVV&jGasp0v*LXdEeneEI -_D+rFB}5DlSsVtzvqb%CMrg*my=*_+n8~xE+dn=Y=`uD?$TT8#7U=>331eh*5`Ct|_%_tg5lO>*N>eJ -GZu5YGaGkhBoH98fT_nm1gS=)H^DyBzF9t`w9aF6Yz*d!Sk02+;;$oPwZbXWWzTc+;H(r)KJ&c -rjvfD%8ke{QxFH|Q(_m~MGB+J&=-K7BmqUf0b5v&B=Y*DY-Yw(A7^Q(1Y~XefGJoE*&e5x*Je{Rg%YdV5$;r=$}qQ4d>^T7^dkTGr&SsTmsYdU)YV8lhk -frs6v#-*M#PaXI@Hk)*e(0$u!p08rVFtlBM99oxO@}Ea4Y5vh*QHg9v9-XL3O1YsJfzp*fR<|C@rvdc -Q+}N@Gm0Rd|=Cx+;UNh^1uXLF%r&*{yfe(E{4ZlVWCjgUD%~eBRv -509V9Ha@ZJ4tXF9L+*-_mYwWH^|D7?>nyra4E4*c>TRK{I?lYm0WnKiK1Itjbco+Ziq20r9@DiR_c_; -q-vQgn?Lg)!4;Lj7A%UybZOGJneE^X^<3ToMh^YfnyoIi&92=`Ah|TolbITAP<4=AihjyVOUzhiebwK@Nm_xtZV^fVZEMfqJ@LzctZ~ZtgwVkYMjYsYL8T`kOULSxnNHZPJ+UBty- -ps*NsoXJc?)OCATOZ)o#dLQ{Yd?m6*RjyQKNf@i`7%N-sf-|C^NIs1bi2giu?8*Yn}{GR+|!913L=Fz -Tt9R-va1Huu1n)F_uot6(KrRG5wroYjJDTV=&UTQEMRGIGb5nguW}u6IZlM29!3K-#Z0y;3t?C9d -pxH^I*0Xdc|GaEXuH=@GqO@OuM7Q3p%>t~Z_aM~zu6$qr)ZB`JI1Pp|py50x#|X*E}4wN9i`i`Cl2=O -#FRx^!J*Oa=gZ~V%@z<$__hq%(#r^qlKGJY^B-C;Wg<=7FS2X}!A*P{$ -UnSog#Y!g_N}eA1lad3+gg-ahW(%xR^liYY2->ieo@pi4@BD;p+6{!n_gss#Q}J^I-rb -XpT!Y=bR=?x%7h2T?PxK0SV0Na0~j6h|#(Kxo}oObMw~DXmg)0lQ57s+2-h38;w>w!??3tBXsS+_6H`Q>dX+er=+0^vk%6P)pcE;gW -_kWt^t8O>|}#!R(f!4C`^lPfZAa<}xRk|DuN`kVLK$rSbUK2$ROMTbZQ=@AQ7z&Zu!aqF+4e#3qF7mV -#LXyHbxGV6Go{VN8zvoq1Auj>_vo#~pSBiTChUCu6~MkxFL?_C3B<(@XWPl~j`}!F3WOy(qY2J*o3og -6krf1r+IB@sCEw)K0AImcZ4I#)8)_EC>{6cU;O1S$fqpip0(qj(nz}9!%>-oeBgTgmE%k#If@K+3gXg -9N8(F4-2xJxqLg}@lps5$ds|k)?Dk1YIFBQxZ12sJA`rLqUK(`?!zJiUrm@nFQ+|INm$b! -#>aJE*d$Ze}^*2vK@JrmW1EQYSpviGs8*wTS3WIz5E|L&)xmGKr^gR)jT`SH+ -l}+C=MN878Zl72t4!a`^nx733)Vw9;+;c6u_oI4U|WMX2byzTnnS-xv2mGkaO?ix+Che>ZRC??8yBZ1l}Q|vb=&sXzagXRfqGf(H(% -T$cfXBEInepAY@hkLSk?b70-zB-Tde$}*Ij~5FM%5!}+t?Lha&3sX=W~@K7R_bY8j+f{l%e75zm&rt0 -_0fx)Wdp`(#$Ks_?Pn7y-k;EB9%USxoq|2l0Wp8JJ32G_XZjS}a2I5NfnieO-Ry|`M;t^uqT29}#Z6J -%JL2ZiQ%cUGoy#_pXvV8=kLIqzJbgTF9_?7DLfoy`VyvghWU)(|n>$F2M2XufV%O{@(2)_Yt~|b;)R- -{FN!B~!X2x_Li7<@Zyz>nl(=s(YhoIW0Np+k$`Mb-VG#hDad~c0o)~k4Y~*5je5T -UAE5R$ja~3>WVTA;hoB{3*?>J!*}K&IoP_o+!0kD0u44lUbUg3m+=HPAIgSrrju~l>!&b6bN5fI$OSlr`#AAWZ@vnRPeHa&?KP2JxZH{6;WW(>Z# -I=G^?MiKn)KJG;Mbu^|D(A$ilMONEd0i*9|28_JH4q6!8(Gau-c347LVr}m0*@Ly-S2l?vV|% -NZAQA=+#eS)+mKj?#t%HRvm93z#o4ytJjVi~gvkkx4yt`&xw!~ADM}((3u`Mx8U -#W1}=(k4J?uD_7w%V3}tsWb{1y`S$eD0g+aI<;A^xXKU9aq`Jzw~y|^LE|t<5xnZ1X54rpK0ImVsrTz -5HG@0jh6U1Q?HfGwU0fP -A<1BOC&yW6uuw2-nAAz*;*@Ebf#5YM_d?TpVqan(}{(#-ZFUVZ{TmTPpa_Px+`&J%m73bj-=5Z!LV)& -CMlJnJzL1zg3Vv4ijshsEY^hnD2`mGFe6UI|7R3PdFA&SrX%2f!K^w?Ab$im0A3fS*WH83x2RVXF2)W -LqoUfSq`=ilMh5f}IGAYo46xzx0q9ooY&K?I7GbjSEq|#V(J56#2lAIm%pa<0T{R`5bIhY@oJYQ^PXrpMggOT* -gVEKed5o*RoIFs6W1Z3~c7e4|`E+TFp$;sXg=xJs%wqI$(TW_jDSdrfH5{efH5MBfY;2HfT?<(HuQ+g -I_yj4m6vF3lAckrUYIiND`jK2%Hfe6dSzK001ASHp!s<@2*Q(3r+$n-+YK0MY3ctvVPm)xL0#C`LkEG -T4V{6?e1_V0`)L81hsq%&-YJob6j}C?a5c(j~IIf|MrQB6J23HX=bODwdYechMTNRVNTL&5dRB$Y(*N -&%wD!7F&Y)4fvjrNaD>sGGQ0Uqe^{Pw=d;{hSxAs~Vnma%D<#b$g3)+w9;0zKQp&qTFLhs{m*pJ(B-8LZ&VxfaUW9j;*~u-%NsE`9}nJKV3Yv=a~;$%KUtDP(7; -JiwuhVL_%u#Pb}s2>(EsUef9T-gS(O!PJbXozJ;1fK1|_dN?Mr>0s4pLAE1nGpFbHa-ssZrXW;2MHMg -0vhg2aw~qpY`xs&mwfVHfc6(Zy54lI#fj)9m=q+zT;MuBLH7-SB1O=T~3Ej|bwFwZ)ooCv?)oMBd1Uf -}oYS>enzTumpALWwV9es%Qb9#Z`EnQbs8sK(ZxoTe0>CmikqX-K@v@EVoZ_a_c1=-{9HSnHTw92oao^ -g1V&;x-B*+xH?jHkUEK6h2BgKlK$WS-X974frlaO1|=9*rZaCPwaplt;D@RWU{X;FrR0NY0}FBDL;B~ -JUu<`8^nEbR=OYocpyl(=fyT2fztI>M;}m -{tWq*XC8&5o?_!;XaLUO$!yjq?*r+oYfu9UB4f+{suN1i~e)wPsW0ObkFC^uQvfr7DKg@@1c-5!`_-=&f|Qq{p -l)CD>l0z7#5@MeZf%3bhW(dRALsFV)1UO -`uwbvhV===03d5T&G0j7As+IHTs#O7??SgGBQFMVhTc$r#1-4+cTz5X{$f2^2FbY<3;hX^zh#kBzqhu9+sk0qF73E-I+^HFrj -J)ZrBLns>D;2i|~Bz17dhl2*0Q)OyUE7W3f4t9XlhczHN(zIF^RaP_feC257v|q3J*#;5NLRf0)7V^Z -)W2sT7S{3FRuk(0aL%Yvc?)2Ll0Dttj(W=KtQ7AJ20G1XM>T&NyIE5LgX}tG%(3r_gWmP93rB*+5FLG -szucl;HqX1~pCUpr^__<2*a$}I^uIMb0OiO2g>pyl0IE$>7wvf|ho@RJbEwYL??|ACkjJ|c~W8_zAF6 -p#@TFoV?+*UOQ(63!{BDrPFk<%?|4ux>Fn-&T!azO-+Ht=!w2RAGWK*{-}V7!b8z29J1CX2k=m8^sf$ -WC%Ykw>lcx-bEzOPJdTRR2vA>D}6lwlfL-Jq#l2#*;z+?@_;OX}q(+f)M?#!$b>`NUD(%P-rl=y1KEs -5>|tX*5O&n6$3>-1Pu|4bX&*>S~c-9l6YG0K7s~CCO-qrl;y(7K&43LBw(W}O8HS8$fXu*xnn(0UGgK -kXuE#}x82#l&ci%iRJ&tL19)FAAM98%Vrtyov^xv%&;&>;t!B8Z0>u&UG+-;l9+j3<)nz8bX5q -aor1x6ML0z&zCj-j|Za6KF(tF&DgKuvzYqP3` -0};uYZ$EtH}D43k?i#~SCHNlDd+|_+&dR3YMLG2u;UorbIZaRSPD)Fqe_BbBxp7c3Q2#O99~=;Pu#=H -%keK~cuCr20k61h0RMRL?(phZfmRME*Bk0|X=Tl;jhXSUdPj!lymChd{%vtZ3jWPvXn?5Jt2JE~kd_K -fS}xzPYo_E{LQ~fY9%&1>OPz9${_eQCea(dQA9?GbLs`)LNkHNbUXyd>DpTJ+!NT=A$Wi -~Bhy4qqugA`8~R5NN*-ZbIEruD@|G0wlNLNB7Lr{0hon&P)kBJib${vQIQrFV;^(5Z`xmZ492jR9IH? -_@~VfvPPo>V8Od&n9i1q^a=~FLJu1X)vOMqM4QIjjX -AMCzOUDNc=+I{Dq_N}7#5Dw$LHKE1}FT?Jir^Yqdjp89ylPMiccjDgqdK%w-!g@>$hTSwG~%tJ#D%*9 -+561HP85d4`W!^^@IU08vddAKjHSs$AB)?l%Yjmv0CfISNmO -NAJE+i^4kw+;Suxu5CbODUh$F7xfWkkDH`QFPxIx^E;2mL&ozxemQo{>4 -wGdh-r#u6_Fc>X+n{#wUTi`F -eIYWz<0tSsFlR#bFkUk2K=THJ_ZE*9u#r@2@_is2XHzH1I?$ -(0ob0hozpH{07C>jd=U#*ZLhB=3E+Lnsyfi45$@SwGAj0X7P%Q^zhd5F^K~|XwOPfhJxHW@rd0VY$dMPMXE@qcf3`Z;#0B2 -FO0lfEBL47vO>h}Mp8R}ngihvuf7SZ&7?r0ovZxKc8k{Ay_Q|l`)ePI8a6kujvo=STzO36ksJQzBt8k -*fo41#|}%%g#dc0Mp^M63qM!aoq9x19PO-2?PSUtnaY01J9gA$yRL*CA4L-}+IE$F62lU0;lceVQwh) -D`P$ti~GyU?|43ggAQ~&($?jeB%_lpB3IMm9KP?`UaJVP_S=-!=Nv=uc)eh-={`@yoU_Og?RrUsun)K -{{9tV5HcZ|Z^!V0CBB?W?EmnC_<17f{8;~xCLge7IAN-G9Dty7P>#33L6}(&!F8TvE=zP%=_RuTZ8xC -XAq>-sH}Nx?Au5IctCs}y@xU{oJ{AK@1PcXM-EdhW2K7NiA4kfyU`1OcG2ky#N410(0j&iFp%D!nfDl -bY@MpgK61)T`XZ?~aDAt)}aI?rPf?%$5s1uLqXu~1O2ucA4!DZc(U4wm+x -fWf*5OsC7kh5E?iAB#8S8ErmS*-Tc6v1nTAzdhM5ra$#sQekCmK3>HKbX&f$tbfuq4IiV$3@+XMjjex -LZ=!xUDH)|S`;Pek1+XK%)(BJDYyD3g{3C^9F&KU#flki!Y$MQwCFu=@fhLgoMdHRLP|cSB3Xf_A2_{ -Ifexv*KQJGV4txItP)h>@6aWAK2ms3fSz7=A0006200000001%o003}la4%nWWo~3|axZmqY;0*_GcR -yqV{2h&WpgiLVPk7>Z*p{VFJE72ZfSI1UoLQY0{~D<0|XQR000O8%K%wh-ne74NCE%=i3I=vG5`PoaA -|NaUv_0~WN&gWb#iQMX<{=kaA9L>VP|D?FJfV1YjAIJbaO9lVQXb(X>4UKaCvorV1tSo`RSW7<^*iE9u$E#Lb3X(i8ltAR+&lmiHp5o8+02l -R2?vs}Q^*Vg16tX93IPUk`xAF2_x4z1;yTOVqJSsg9_ -`$pE&UU!xkoW>o&3h?7U!gxjk|`zRRJ!1z -K__OC}TwR5wLgYx{PUF$h1a#=A^fZVRXv{QwK-cKI5~_6K$k7AToyM45Bx_izJKuaal(b-lEeyOv05% -UDi)Q-cQC_VWl15U8>Sg1#k`Sx{vjap#o+j2^a8oT+C(udA|5`|K)CddB0f9Z`QZV@AFeYQ49%jG=i1 -K?eRfqMZjPQ+E_yy4Na?UHbwx=j=rI3m2kE^y8-)tmZvrca5%meI)wtL)NF~amoiTmapPsVP|D?FJfV1Y -jAIJbaO9vWnpV_WoBPyX>MU`X?kTYaCy~OTW{Mo6n@vQARGjf14qEHmmyt-w!;b(D28Du(1##UXo<2} -OQb>&8gh$^q -7;&8QN%A_9Gq)A&&(3YNVn4Bmhrpdh?uQ4~2-qz%Qd1e-usp9?*BX)MOg1Qm@wL@;>qPGVPm- -`MoiI-8+QtRjSuZ2#i2~p?N%EZ6H3+?BcZ*{y7WYY#6g9V&9dD(2{ZSNdd6O{s0m|QqtkQ-B67D3Jr> -n@y*R^1t%M#OA)QVNDDpu@@n~jCX)$V|AOTY+5E6X@504K?ae1@cWiR`eI#g-jpU9+_yT2aEHP)NX}W -gBkRd;?n=whAZ8jJ@BA0K{0woNNRIrk9UwL6US#Jo38bPFRAPBGM6162w*614cQ8ug`6Y-)j^`6zK^YSy|h2H5*bl^K$}p?-sp|t4HHiiGK#AXh#M{yviBN%M7(ocFc)k97I-)#*N{J -m!kJd|-ZQOYQh~Tf5CAbVB|wEh7DiA|J&@NOV6QkB7$pWdSxkw9UM|?1-rf*?R%-QwV_W_x2)K110W| -fxDi1JxudlZ3vSs?bf8PKTI*fAXZc%I>0Y>0!TB3$o#{xFv_Xte7K#Ytz#Wox;Buw3L% -a7JiULq@tc>fTD{8QX`10J=+F2@-)Wed40bNe^;t>Nh@YnZDBtD`-KtO -_z=>~=dHt;ds@KJ-U@B;k;yUh{zS7!bfv7 -qMwfD3lo5Wt{2h#B<>oN57#i -qJG}*Cmrf->)pCyBe%|dI=FN%#O{j@{U7=7yE*42%prR?)4JYO9kwj*5znOgd>3ty>D=Sz`BO>oC -$sHDYf<3;hPY|26N<=G@|!OB;uH2*u^a(+g@yzb@D%OI^1Qsg&ip;tgDXAQel$H0Tj95q-6EoQN+?_} -8nMu;TBO-WgHEG2bC}Q2PaLq?m5RqN2gh-vNh^fikco2mCa9q(wa*wl3IT-e&4YqZ`w0D_#W74Ts -e*JWYpvD0>VNc4t!)vxCUsJLY=e?Iznj{bMzSzA$ynpdxWlQeF%}j1+D51+6aV1s%M2rJKW2BZ8dVUg -^QMvn+;S<{9^Eo5W1-;aX9*{bJBTd+$Ek0v3KludVsd`zO$XpJoQ%VU3wM%&n4ZrPD=;R!mPo3=#SfB -&q3(x?)W;p$B!t=(N_+-xQ-hVev%VUqG%oZzHALXVh8)erE>JCVOGA7rlSI%djrba?}9)71yD-^1QY- -O00;of09jk+ikX_&0RRAl1ONae0001RX>c!Jc4cm4Z*nhna%^mAVlyvwbZKlaUtei%X>?y-E^v8mlfi -D=Fbsz8ehMMGY=F@>=poBIG-!|*c6!)}P;9ypVp|#|C+pi!jzvdl>XZ2W6eazn8`7NsXa+YB0tnR^O- -{&z)$QOArZ`EyiQk&UK~|@Wq}qx~cSbsOP_1$wsW7C^s>ZP03U`!F3>ItQv^bzRBH>fgjE6l{y6>@aO -80!4vT%b?lQstHkWKh^KvxT*z- -(>E7W#H^tIgBnOS^-;oTdK5+je-JTJuQS}bldpzFy><#d4PQnN-Bn?rjWZ{dZM!z2NaZR=<7Ias|2zA -mOmEMjLP_Q_jTZeB8p{bqLGPNvrp;2@a7p?^FtAKSx9>>sl)r#uqpp=8Dmb3FUZARcyR52Nu}h=zlus -B1I2pBTn>9ejY-KFhh -K2*A-pM^HM8k#4_C~SY30q7K#^K?S#=P9pqn1UOSezBCy+Sy{`T%DIv -aOLt}=JThIo%99hiIbu`(66t5g^>-x^eev0OyMSCG7Fu7-~1>^09#o`l_mM<3Iv==7 -1^S&feutEFfz&FFo7VT1NaE2+1KA?%R1w0Oo%?Dv3bifB#bH|)BKDtvF)9QE%UiHME!wCALXr*60pMh -ELQWvi1h`)6#>Ord@sgE7{PZ(%pA|SI(xdCQj*#bpeisd9f7EVd_YtFQNqApHAK$6)~V!Zbofnqn@aM -*H27e-pwKistJ`|3AYv8ePDg5VC4ihu>1y0t4jhEEIWOw;+^<$ -5ZPzc -CSj{9@`5G9+=e;WfSW`k(9*Y|8pLRE3x!G}JQA{Olsqt@Ak5>Qm@SN(@w$|}2RGM3L=@8*N9_}X -hnG~+gJxi3B4kW!E{+SOUnrFDR(}0SYrWmE-=g*Hn-hM8Ay8Cqd@_Am&?|;2}e3?&UK{Tf1MZEuk5IX -HDrH@10n9Nc&MFf+ZSMlI~E%-(X=%Y5k4IXHtRI`)|> -?p1b`Yy@^QSQF)?R+RHIL#@W4an{52Yc+yj&SN17)VjmhZ_!sXUbr!TtsFkMPWuwT3VlN5d`whn{hP>Tt^9p1wRr`C-2-wTCsq8e=fAWp!S64h6i%PNapbfkzL0f_OkC`23q}(uJc -|qaKx^4P7owTxMeEE&NoPuOaE`ds{r6@R!Vby9qhHtzzlWE9Yk`%2b@EV%W6!aC7@^S$6X$-+M9eaNK -_|M{b{@dN{uLtm9t8l=zHTKflXs4%8GKC>W_P{-cF+_dcg*avmOdCY<;84^YKC&~aqEG4Pf{GhAcA_F -Kp^}a@D)|m$6&X8KWx|D8=L0bXUMz+rAG`c;@kl$htVmPHCrT&fU+XeMmy_GMr#LSri=UHYHADLPiX7 -skY{A^GigaG=)&Fl()B|FYBQ8#p)% -!C!E2`Bvv&RhsKjt-z34y7D!m@P!TyYURyhwa?kFwUGvVzIBQ=s+l -<{OG~dFK^=fh}I$?c)Sm_$fb4$W7n146gu{df)X{y;Su}A7vch3*m-dg -Tw9HJ@si{= -tSvtK$n9L4b;!ug_y73r-4Fj`qaBvv&6%#@drKS^-b{uE4$7c=jMGnS9(U=VNkce#D7~Z6Ts -0DZ&y@6oESXN*~W=Z;{0R4Nyx11QY-O00;of09jiK&s+yy0ssI-1^@sd0001RX>c!Jc4cm4Z*nhna%^ -mAVlyvwbZKlaaB^>Wc`k5yeN;`4n=lZ)^D9R5utY*!dRr-{sNGhova8+nj2sw(C1TUDNjCq!W8;AHwO -oMle7yH&#trE`6vlKV$bGNLLPaPnmIHO2+4p>%5wj}&#W|Ip)AQ`Duk|Z@Yz)=b-% -LeB@%ck>fi5Z2T}$$G$6N>DI70xT;MZpJ2CLWg`p0U+ex!$8>-~N8BJug2dqH9kf8YPiAjgXCac14bK~c<`v#dOa3qo_a_kht -lkmm-CYeTPfw--w6kc$2kCpDJtNHM3}FqF}D1{Qf2kjVnU$XC8UmC`gQq_q4$H%8P8cA*`u_rhxXz5NIx!hu*7I52-#F-5}>$(V$#XSC5s2*;X%O~moqT;>37HYh -{f0BUb*>nlxEWHwqYU6?aFyu$CV#JxMdDYKN5_#JBMidE8*_~6IM0+mbIj7i?HJE+PwE8K{o -v!fEOCIk>pX^w?f81-afi>RE_LDpEq_T;ABCu5>iT%?bf!1BSpHuo^1xx90+WzqK{pQ0u`mN)lfHhl& -EMRGeiZrRR4UQ9%E;3R!~#^PQavQENU$|CkTq@5OZ88jZNrHku%iXw_z|17`{;}bb2{{c`-0|XQR000 -O8%K%whP6PIz`~Uy|@&Nzc!Jc4cm4Z*nhna%^mAVlyvwbZKlaa%FLKWpi{caCxOy>u=jO5d -W^fg3wUd#%#62iefO3qE70r!Ga`M(tQ~UnU>Bm8;R6NDvn$9f8QPTq7=K`!w6!Fy!ZX?p^a8Ki{njcO -AT?%L|!Ou8DATfl@{XaIS4jHx8??3y=-XKuU>*oR0{QN6v1Nxg%wJghUSCRuoE)n>kLj+x?gIobVFY7 -TiV-Gwlf$fCd3dvThE$|eyS*@5=%-o9ti6Ix(3VFwj%7MM -Vy}f2N8M9lCo@rn_O(A8xv?}LTBs|OZ5whNHfhn2opE8)TAnS+;pH{0yLXOUslGc+w;p`N#EUPQA|_VWZXg+1o6AY`Ni_>?egLRZ`Jw}BzBCP0g|ws- -$Pu-U>q>~+yfLb&%^_qTq+4ucUajIH}-fpWKQhN@eUC@B8f5RTko*Y>Tw?*yh?>3)Wiu-wvF1!++BrX -=)P~Zn3SZnhHaE)s1G!Ex!3Sl3C1>wi~>hKFSzEp=S(GyKIa=u*sU(Xr7ukl8zo9gj0&Q8?pS7yWmL= -TiyVZalY7P`TO;7uTB!`UoU%+9J7s9Du^LIEmbB5&tme{Co(#cyzb*TughX_%_bg -F)&O5Rjr==o9LohQbl@iANL0}Q|-RSZ*9}3V=XfSv~4|eM$)sSO*Rk9tI=$0vyEsAfM!(wl@O3q51Wn -%sF^B2*pu&1%hycaW-A&UW`ppqj^lIMUd#_W_%>(IHRkBC=k@tIedassipM59(7Q5?+L5PbjD1SeGJ8K#}Jsyhhl6u^vBYmKYpk -Mu!B~1sYL#P4wIN5O2^BmWQcq345p^P9W#%o-~rVJkk0~I?CXAKnsyqyzMma7L~B!z~OTZ{^XuB|4`6 -;1@b69(JJ4n{a?xV4ZMmS{CPiw)D0OHH#{I0-{h&jLz}GNf86%V_F%ai~_wn2h&42-xg?W|Pqv2+;Gx&fb-j>&c2j14ucp6grA3sIL!cVAEOoa5vSpw -Q}A=?C3rd(^0b1iV4c1s5(y;;jNqyk}qdmC-aXp|N(Skg#(7ut6ik69+}pJC{BieqSFRYZNqBGJ%l;8yz$V)bntF7`34eOxh?a*#-ip}kcZ#E`pwF@z -mN>*im45x?z;)ChL@bW3e?efvY=i;m=GIvS*IwK$<7P-%v{f1QY-O00;o -f09ji$DAlxJ1polH4FCWn0001RX>c!Jc4cm4Z*nhna%^mAVlyvwbZKlaa%FRHZ*FsCE^v9RR!wi?HW0 -o0R}kJqY`{^1=CXAW6zOiZ$syaIn=T3zg+ogti;YBjNy=WM=zs4Ul9FX7$f7`Xu_bao_~y+Umew?4u_ -#;LT3RfGYSzYjA(zh7t*1qNAB{@hM~zfN9+jI0kE{&>6F8@`EYh)3m8qSHrsvJNmREpq8d-0ys4^o1x4Iep@VBQGA~?#FE4>E|19>s( -9X&G+BE?Y~<>r8Vc2?PG-1SORh7y;V!pK4892Z7rR9CF6~f0>BAVmMT}W-k<+8oqsEL*-L4}B>smKR2 -ehq>m)^Pw3kYwo~1!ewF$L-UYbh!fx8uXhYV{1+=Vb2c~!D?OiwYPY?X1LanEG@o0iL(m_7x0vJH;YvouCYd>l^`)8UM)UsR!k7VFqTllbb6Xnz&<4tQVJBJC8F -+>lrjy1=!0PR^BWUGD@tWcsuD^;d6Y4tvr$+8X%*v7`{8UAZEb98?0-5Gj`~LTaI3G`wz&Mf(Ba$Zv6 -R*k9bsS#0f$(Z54F5P{CqJn4mFoHk@tx$Eo$qE587h8A25lqP#FG5d<#?vqW&krqwO(sk6pZ_w^2!d3 -yf5&|h4=%qc51VU~p)Rvo&f*`b5WgSYAa{@U?&;Y}d_r2FvtT{g0V{jM$gSg!pvb5YWU_isTz_)Ziop -UDlAQMl^Q-`Ula4DWMbQGXFzA$!yF2%*@+2^cQN)t -%E-xw;FUmZvD;|mOy`Fr}w{eoHFKK>zn`F1@Yr)l1m2CC%J}xroRQA?5(=ty}zO0fwE7eq4yydH^-lb -f0Nh)4cd7kPHRs50fHwms6M?eXj>pV-rQY@RnDv;gJ_X6d@&|Xwd8Rp -0SI#idk6|^%6IRt@8rtyDNCfvkJzsYQx!LlfCvccuU0jWU_U!JI?KxY -<87yX--*H%EQZj)GPg$I;2$y9VmnGXL)rPGQR{+OXEEhW+_M!{EjuEX1%NKuz83UX~42ozk-vE%Hdf4 -&7m{9ujuE=-5O%L>K#nsj<>I(v0>5&in=ayET^c{2-^X(iquX*demyOR6h&tAX2`gM9Cr?~;>Sy=%q_JsNt+X&>k(V -2HZN7+~B5zbB~IgS?Zi(NjM;=XOu!F>yEYUF>{#Bc6b<|L*tW%QoUM#{UVBzIypGdUN*gf0V;55Kl} -x7xL5ct_54t#(Lv(rAMXz1N;3Q5UYbqhLd`K23snOSF1=H4uP=&2GS~^7lAMqB&GWOKD`A?G+QUPbwN -uB1^uV-awF^|76b9OJ1}1n{bj*{JYY*@d@q1X3q%+z^=?JfqqV`#lsQ2I6YYVF9-v+suX*|auZ@|~gTrmA)@<}E;b6 -K63Rof6VS{LnC$Vt8@9i;Eb1uMF -I+yoysJR%ZB>W%YK!3*l&piL0E&X`1gtHUs~xWO%Ikif2nsTg`Hqy?|3zXH^**ir#RX0jT{;o&mA(+4 -c#5lfmhUaUCd&)-kRXyeI|>Cx}KwbV42puxs#CaTrozofKuIJfY?Yx%2o%(voHeuMN)Zd`^^#@+{JaN -)BiePo6|WzE}C4ZR6}gwZIO*$w*0FaKRHq0+K12wYb>^d_CG;*-h*h>>QjKvA61kNnZ*9iPzg4q*cj_ -JLT-+M=%EXVZd$mma`2AnkOY?kZFZxV(L5_*ptJ&K{wo9DowD;^{4lA8lMqEosgO{Jls#Zzbeeg}(09k9CPq^(Qh^TwIRb=e15Z8s2291hgMoXp@ -&zOvXIkp_fnFf$5-zY^zN^kYEKh1^%r2<0ZP(^Z=m`sqQLYD4UpXJo`mE|POf&4T=a>KiCQ$|Q$P)^O -nEb_ZGsb~V@nv;@p06SE}QLt7n#Vj>JVAPAfUBvFkT7Ii{a{&P4YCh2m?+ -Bw7*JY!2Cl!8e>cyKTQ1c%sy(B}{WqGHO#DF&8!2Mb5mO~ks_&cT+vM_qs%9_!BM8WABv71t)>i7CL(Swp^(U<@B}Fiq!_R3Y&o8Xw4S-OHb)e`AB7{(pri1V{(cL8!!Z1y2` -X0tGfH1P<7l4(_osPwYxx7fsad-uumT8OETSl5b4gGNF)7q*sF18UltudM+>WfhKMmmA(ii(< -)*Qzcsv8neeCYKtx59l&Rc?IseV6r4zf*1z9HV+9nu2_mF1fCZwhes`?moJxf0h5Iy>?8()^uW=uG!e -m4!2vfnH*ecgpe|82L=#I+I(E%nfr2#qT?0P!8{*UPF=tBdK)9Csg#r4GF$w3l3C>6|4?Qv=!*QdrB?l~%-nvaF8fh;qj;n&Go-m0-*f7&~^HAiC&(j(Hd;5<5 -T})q3XH$H-dON$k`sD^6XV+)v)BXne8S?&98PrzCH_VkA#`)>5&Lvw0tBi&i -MIl?2>!CXTt-Ols;W5SUvXEl(&NPZKkYNR2i_>bM!KGi@hC);|f^#poUp|d16N_}!V>gmXR6rn{FMP@ -xEPuE5YlFFT&jLaJq&qWJmXt7?rYP^zOu=`qUpZAh{KJT5!lM|I<_dyv>@Bp|5z{J9#>8a0(nzyve#k -#FhMmNyyl5sxpPEKk;Q57m&XoGvGB#gAIbywel_n-|diwks{-&;A4YLJn0a)7rThx=T2$-W&2cRZ)XqFjuqcx{>p`q3x{p@(YzM}86F-* -DC(_{n;ipW5@N}D`^n^zFg0R|gdFqW%4UPNAZSX}V^q} -!Q9|PUo2>M!>7@@6kvR+}hj$;XxGqTAS-W#{O=l@MSec&y-1--_*jv;z)yf<$5jV^A}Hhv*I03Xk8O+ -rNNYV?5|VCxn$1V8ULA=0x524wVXP|9?}`Jl%Rj6IR@eLxlj`0gL~Rs{eV*hnO>cD-qZjbToXD{_>*V -IQBN$FpF2=;PzJN!98wr`ty=d5b&Uc43^Dlh!ucGD2Fp+#TatCN1nbJbd^j?M&M-%0|$8>5fpAjm@w= -Jv)t;N7&AX(`hX|at4Mz+u=vobE29-E_h(K$sIUcz`@9f7zscHzu&|d7Z+_{<6DRH84gCPK{UdEjEhh -)m12P8tScH%Sub#WL-s4y3M`Wqw}31toQCfhMA^l}cYNF@XVh%WUM*2od?dNv{ve66w-+_~8v*Zo~BVs}rOrutQNl6o4$Z92g3kVJ3=X{UG@IYtdVt$UhUm3y4}XU -_b<%%}NxA$yk|H3q($mih7c7WoeIozyA}+M_`6kdA!WGJD}|%Nt5b9SLx)biNIk-I<7(tfhY)xsSW|2 -K=8idqso?4`45cHkHVk8QAB6E2F$Uyk`fss&XzpT7!o568nYn@WgokyN9$+|oCWIpfLIsj*Ihx=MrP5 -lprA=pJzS4U^V@L67xit>KhLu}kU89tjS@{`_Eo`-{}ulBt!=VqYcPVoD{g7sz#EkqWV4SppBF@;(+BT9xt>9v(sl#-uE9$(JzQozDE0Dwr2s-e!4W!f -WNgk8M+G>B}NP8gCb?nB9-ry*5hp9RK3-rgeIzV;+&4@#fejS>Sk7R$yxAPHUNzAX>m;u?zueU|Mf1zJmQnxmoVwCGZK> -z?I|Y&*{!inD7crxZ?KtrrM^@i}mvR(-)&>rL25<5uS$My7D5VQi-7G|JQ$(U%ww6|M1}(c=)G&*W~z --X(*r@3Ol=>?$#95q8m!uNF)U&;x)@Z9|0}iDFgIC~!z6Ga0@Zu|GbeCCyc#-CAA`7ChDczhSrU^mWY4j&WUyl&Jz2ss=6Zw?XH0;2 -~#e;U4*3d0~BwA?5HE^{;MU-|B=tz<8+XgC=VXfRTC7dOz&Y;;wZM4^~EEMDzpS+I{75L_c>KE+%zpl -o23-f>(WJL)%7Am6|%YrfRLFaRetuT+Q^5kAL2Ms?_HbMTMycD1k#E6v2`(dP;Y9C^4vC2!CSPrHE(t ->K04Tm}Ls$-lgLWW*_3ism#7iXB1!+7vlJcF7(fY@u$=zx`9Mz%@1 -+Ta=Oyc8QWQGDXwE@R8gD&o0lX>iP{u6^<~I8z%+Wqh(FC(s5-MH?Khb;?7~kdP0@wjjBSbHcTp2 -+8}Rm=mIg9))wlPa{`MTTK9)N69!sQM-&AZOUZ4L#i_v1Kh!?JmMhL4%`i0cMbsj)P!&Q;5heywsICG -gFDNA@q9lSK=Hv*Y!l&?p-vZ~$3Uc%~lB%6#9asj9d)Dsd?16`~bmwVru(&{CE?&FHpKti`V@b&p8HT -d!BF%maZ)?Ts9oDg|_>N~Q!enjsTUL53(|RSs8AyVz0D|j5iCEf8s*usel!ks83NZ?g(Zz~dSFKT<=>mA=BO-{YGIfKlbyK8nouudzsi2$bFc|nOYPWs~H(H@D2Qd1@)^4au& -8|8u!T(;#QC*}hMM&CpF|Wzu^4>*wt<`@=l^<0RFKq}eXI!h2+4lR!8WXPWY{<%0+dq=!mfEhT3VrGR -2w+s5EA~TL9=Gvy}(7u$0ABqO)s|Tu -LIxdyHze1&fhUvOaQ~N&XSRjd{Lc&Y$5&T9$uTG=iDoD{aASgiu9?sgHu?y?6E>)#Uui{ul3vtmj&W% -`!M!fT0aC)Ynxe2J}U?@fmZQinCD5_MZMRq1@Q>mh|FJ4fBnvW;!!~iq}Hnyy>4hR+4XlY|LxJ5tBnh -zy$VYju{>DGCuIvgW_%(CK|Zff`I>?0V}S>J;TO-^>FnG~3KK_noL;6SZkTAfnf;@>?}o|-tJKl(nhq -G5Z9&>!4#dpM|BXO{NuwYSsVWE$Usz2#oE5VtD;LYn*E&Kd**i@>;?|_t977(O@jn8ig1^!lI6!}h9t4rDcgq0@h*B3tr -Vsk$kYw7b)@9QVq;DPtvZza(Vf6|7}XRA!)lVkCGR!k}1i!YxC2HrrODp=galmQKiBdLDR4uM7>MDa? -NM0rRT$t=V%GL2WuCX<#po-Z~@6aWAK2ms3fSz9?X>2cZb8KHOaCyyK> -vQ9{k^ip00+Fk3q%2L3A3MqU&UdAiHEUHow$Jk1)+CjqAQF-gQ=~#tmX*!sZ@=!wg8)T&X6y3dT&^)6 -K%>#<{`CXk%!_=*qG-OUHid{HCRgjcs2HD>dAg}Yr0zQ%{d~#GWh!U(cEw|JotNgS+}X=5FEW`e%FZ0 -enaPZov6N<%rFc!{d>61yrY}3HVHk?nSgb3VXQgS}>oE8kXQhaN)rXW<^SoI7qy`A+P`q|B35&>&)`;n)m=!_@P!$<#4^u -K`OIU(Au*3CW}$*Y+dBZCI&e{XIbhV$nMR@&v(P<)8GLH`x&oCZv0tk*j~3s`rReBbQ{tatGqZptM;rMnLo>hA3O58Zyh^Gm1GNyMB*F=$8y|E|^?^jC2i&0q -yl$UPpgS^tuT?RLBLcc_}OYMbjxl_(SO*9}jNjDQ)U%QKLXT#2wlU7tXTrpsyr@|y|lTJ6??l~r*v+f -||rsp+Etdjj2kDmH@2IW2(g%BvXXo+WAxSq^WDtt>_8aaxFSlcJ)4G4_D79`CbbWU)8vyP}*|0W0KUS -s~wllj|Px!Ha_Lpa%%R-a$7d3y1pOuZRNsA4NhBjLh5duh@H#cB`a;IlJfp1+YZ^`)=@+73z07Rk6Ed -?1()fJ)9A0{#}hCc37rr)@X}xW1bI3k56E6|MhwNIJyqlw&3fi%;RTKbrg!QF*PJRR{~h`vQ#6 -tQZz;hT;1>zq2oV)6{!$tRo2oqQP=vvLe0v=pxkk}7O|Ym_>w*BswJ4`dHA7~bpr*_)9plQt$hLvyqu -lGl6h?oabjom<|vvb2x))dZx|e#9Ae<>-Qo7^ja77;M#J0j=n)AngLLAK#~xiP4Z0X5+cvxO`j-bOaLw8nQo}6suvkX`H-(;Hd -FfSG-6QnMy9b6PIqKGCk5wucXuk9ACF$D|_0fSD!t^GyBRe3T4xH(=I9K0PcbRx4vyoSqam3rNmyG%T -XD{;aLXqQ66B`*?z@x$>B=5o9hv*VJ+FOOlIW6k^tF^>*oAu>^bYmpENfIE>mFBo85h`A{ATv@r1;5x -udaEPPk_+l-JIq->sf*2}6Q5GWJvEKE~=wvj6fBS(;h#s*V_+WT}@|}sQ(v(odSuDIBJlnCsXmb4H@D -ur>;gZAe*?UFAydC~9JU|RP0X?Qx8P>t0ZYUEmOL`_^z9})(C=!^KIOcIIQjiaXY1jqC1DeQk -tN5_nl_*Y&E{U?Blfwpn9E*xx|c@&ERB43^J>@~=BriYQ{esAbHm#H|| -`&sPXd~XmyzxUQ5&3Qu1Me&k@U8&_TLk3LKZd(tBFLU`8Z~CfbcrD8`SHM;pCOEdgLZ0zv&8wxoNqEK -Srj@Z;+=@oWYXexLs3F)QVj&%$4E_`x5)zvp_qxLYz#h1u3KBMWhBIl-kvDB1Bk#k!U -ZUWGe&~7pMWp}o`K(e0gqI&?mn$ekwQLxu22B-vAJA;AB5B%Q!Ty*I9V4Gd~muWC$-``ru=jZEmSSA6 -zU43G7&4=t#UR8jzab&-$3t-ZPtr|13XZXAro6zuoiL}!7z{3Knd;3Jbe*VcL>?XmPiq$+sd2?+w_D( -z`Fz-Ym&!hDDxAMot%beAHvrYp1fcTlN@hK@a+`$wU9f)h|VmO<6q3ro5d*e)A#P4f!w0Br4w&B4H4Z --4?FgU?Har>EiGQPG}T~Wo9g=mRHH)t)wXp&zNIj -h0V88vXl5by^hrmdQsc_ux^IwGS!N-Nr!GPXS}-eM~4aMTvlYgU^7ecn#hHO^(32{2^En-^ik*_6qbV`T{A}&208Su+ -&i$Cbqe8Z4Y#$E5u&;E^V^8(Zv3%a{ALihFgdDv8fNU7shU5Iz38yebrPyz~D05l ->b%j$>fc~3V55##c=76E$>onW7v{i&ERvVV6_`VJ$3*` -aV*?H^td!kj{o`-&`__P{cMFK%NG12vDTrNR?{tqtms_Lp#7$k>*l1GZhtcxj9r%Y>Mb=BiW&-5rdyi -g~gU;87Fh_6f5f@c}g7q`K04t&#+C-j^6@T~2K%ZwjhY)pTs|kQ?-o6p8>Nz$&tWeBmIJIOMA2ZL8>D -|8TEx4mhB3rkWQ9x(W!!_-qC-j3gscgZ3sl+v#vZB~&{$wTxkoUx9R5`><9aOyX=+ia1ekwO_FbZVuLQ?P7lV;0+H&Nm{dB71pdW03D -^Qz=)?+3)*47~Xtqf8SRrFO<1f}joIMfr>gv>;zQ$-_pir5$5g=}2uD$OywR<%Pa)nHAXs_Cjsh<*qJ -Y)dD{Y$6bVQ0f2+Oai$uBSK43Z#tE!QC|OqS$DXfl*dB~xd -EE3!ReyLXI_!vpKxZ??C!lz+Km|Dc}k<^#3SPMw{G=~TrKfThzZ5uZ -*q%CnS1V35iTpmmnNh`t{dJK9-W3vI265)d`9)h%XnLtqY)QJv>*C8(vTZfOUIKh3Qoh11kJ6X5qTza -{RLa=@Te|W=qG2x(^T(N%-Q`~WS9f%x`Y?QTXU-?_|sC54}!VCi;XN4GfwirD -BaT4Z9UnfcF>m6w8ckHz2UVA)PqlSG14-vTi__dDHZ~Ain_BIHJt{Hp&f;Sm)=89L6ohzuhlM5gIl}Z -Q;Ui`tiLk?K~V~+h`TgcC=h9#KPjYi2o2F^7aKC?{W{7Pf&v{?U>c`IqDkSN{$;aLYMjrJzY9t01U3` -%s9yWKRlu?e?cY!sOMxox0*DHC`&qW1qrmImSWTy9$#`AES(HWs{;7_AlgGsdx?o3 -z`MVHE;#XUU7ze3R0fgkutrZ^}L5f4_3KY8&I6V>Sd@`9fQMfG~iltI~{>f&P9Y;d6(uc4QthgYN5s9 -~(Z7KaChf7#!;s2*}3DPs8`$Idpmw0$st5&fouQ|M2LJ2|Im3;a*O|Gg#Af7qZV8_|=kJq3_4rBjq>4YRn|t?i*MMYW= -oSCRvlM9W&*o@I(Y00Z47^e~PN@gSyJxY(I%~cb1xVqr2=$J+XL3#>gEZ -Q@M#}H5^RhZ#!qm$phw#_m?4)JYJCv&U`Qr^cI{k2d5weGy0r0JfWg>=^;l+-6z~G6*{-BI;Dp6a|oS -MW*NSb=sF)JldrwTdS)DgC)UH^Us?Loy5IVcuWyPvYGY=vIa(jZjUMc$BM-2#=#gdJ`m_d5vXC@Gw~{ -;?)zy`<6zwVR5m7lw}JP~#S@+>p!RiZ;GpLKR~vF6lJCk;&oRO( -HT#b%)`NBWUBE8-YTwu70L^d>5&~TK$S!<@Yts$tt*n0gmWk1;mfnDWc6j7d`Sf_j*pkgY)z -<7sGGOrr%j}=rLDULcV%->C799idX<17-r?$%v<6BlqpU(3EJP>t@LP+5Jd*I#SqJ}j`?a2X>4&(vUe4GM#r_)0c_hzDT=aZ0;BXq6 -9coLm?v+LLD}Kfw**LnfV#4!qB*#>CE$TdbliX~v>E@Kmbh;l{wB3wNFdudU>DQ?gOAH^Yx*r?`e2{t -i-oS>Bkq>ktCNFyk({Ee7}cW!J;LKZB72cVo>77Dg%=X3Xp-+hQhykIxxN4KskzkcH2eEQMTsn&r@(oHB} -edhNn2xcJU~6J(h~eq1b@^{bkJ!O-26(9KKY^!OnI{kTqboXrRI-v09%l5mWl9u6M{V8NCrTGPcN|^q -)N_i`KVDY@LBLO>I{$OqJr0F{b_7r%KjG4O+C61PBEI$ueU2hIkdN#C|Y*Q0QjD%3;#Xe&#Z#shm+W41H#;=7U^4+nJaWS;T`Itb!b9Yt|=KkBMQ?E%`)$@IN>xi(Qx?0& -7|U-QbqqdA%5CKc^+!hMl_dU^@o7S_Y)6I1(4RLC$}f|eRGRojO-BSYhhN|iXiE(MqA+SF^7KdZg7`q -LA=OQB5#0(NAfZOg*1LB>ug1ii6-xtE3?!F>fWHU%3q7&WG^PGS=l@ie8OjUOXPL1>0&BeNWQBSv`+w -T<4%Hm6u1DZ@>F3>r(w5!1vb@wN^_GKHZEu>zCK7GAPQS$`tx1g@Uufhwy5;Z7~RE}ZMX^`56af`v(# -gOKj6r&iu+dxxh3Yi>_*xYPrWcI{;~p7+N0>PshfKcS00s$Fx(!ry5XdP*Wyt~zQgIth6{R3N5p#h#9 -z^j-ac*`g9+GhoKnl&GcIvEB+$6NprPl39{(OV%@GQx(Brxqmep@ql -(k>bS`BIqv0w7n8Y(^L1YX7OIS3@q(sr@?l;+Q<9>9O?N~CUoQ&|LdCd_iW#C|Km^cs!&wp4?s>uSdZ$Q3Zt+hY -*f(Vu(u5uWQZNXti;*tjvgs?lhEuT=pQPYjZ1VKk~@wJ(M_+k`Q!d1cKX>6p*ZaM&i}wL-pekhD9<*v --?^!^GBPv9K@`nD;|m8?f1?GCII|>N6U4haCEcA-m%&(IV|TS^|oo_dmP?sfgO-W!Jg_6gJ7(GO#K%_ -SW4qGh&YSxo{mu@NwT_h3T)x0gsmr28#K*9)qmQl)PLfM)HhF~9{9RhDW}mXV+j8C -5(a)?8c`!z#xK89U!5p(t}B;yths*%y3I;zOYt%Nt4tgmx9_I@hK$brJ$$jiV)Nc>g2~dBwPa4jmp+xDu7Nshw{}J8v__}N)v{hk8x@0$IOx%Hgf;;dfz;F2TtYsjt!2%u0^WPtYVlxm<-2bDj0(1&r&Crc97p$DUj^khEdB<%{m%aaP)h>@6aWAK2ms -3fSzAAGx4P*F004d#001rk003}la4%nWWo~3|axZmqY;0*_GcR>?X>2cZb8KI2VRU0?UubW0bZ%j7Wi -D`etyud{8%YxW-G4=CP{f#+A;4`mXB{h{gzO?^feXoAblIHMp6S7zjJxNi+Zf+pe_wUaqYZ45dk~43e -pG!`_0^-DB+0vlq@{5AVzw~WX|a^)n97yToidur#w_V-;}-aQoZ;Vuq_t8zB`R7kq^6Zs`i8Er-(*RW -j7B?jVVs=OyM?k;s!9_6U+F?NilPB_T0rky0{L9zU-*bD$Qe4yo*aR -q*G|Q&4->L7aQJ!lr|fv-Z?z< -H##A;_(v_Y=%$$R!%=y|*duu=3YE`v>9ozr$Q*cR1+Bhn7z{5X{bgM?wYf)Jl^zG19Aed)%2>9{7-Do -H7rtOJJmTn1aC$!hdrBFUsP9Qk5b){U|>}I>;7C_?>3|mQ-gsn~pcKxZ3&jQL=V%RvGehS!npJJ=Kp5 -fESVCKXl{|EG~WPF87juoAV&uY4k-lg% -U&PV09`iXW{GNp1UMzTLQZX1ov1X+l9_dRLOK3y!EIyD)3c2G{0>h%Pr$Sfe2o -wP;}gHXFh!I4>U%y5IL40j1y`y@~vVtjmW>KB1Y>Q%dM5^wkZBKJKNAJI2^r0_3XZng~-U7tp>wFoWc -ne~*Ntr6E{?Y9Skx^OKW4pSrb9+aIJ0{#?x!svI3=?MZGFd8Ob^y8h+z{hL>Garusf!=c)H)F&?qs!= -F#ialL$rzjO731_SqDqoDca7U5O7oril!zq|tH}Ez}9aUm3t56sNJNOP0b>9=+Z)cq=2L7WM?Ikfc|H -fTvcW;-Ax#$_iTvX1=s+=$(Ja(^EBlp=NF2b(jqBw}^>DaGvO~^!DOI@VN$A4s>p0=4Is|ZyFz9RS5$ -}Q65PVet0sQAqglZV@}0pASKaTcBapTZQ$L33~9AH@l-l*_sFy^LWax(LweejBq`&iih$D81V6u`_O7&92vC>2B=69&bLISW -C#~Ma>>x{}nHCuAZcKwbT6#wCu;?$RUBVw?Y&)WTN-kH962GUxj3?1$!8R`1KJ~VQIeccj3BglY -fyeyO9oExZ%sd)t3@Z@V!pqUKXNz^y2V^r+gRcV8*KA+Jmew4NvW|n9v--iz(u_xrt+d#n>IN#R>w2? -oS?Eef>=0pCly$yh2XXrAAyK$fDpd(5;jSrdrnK9~4`}X-5gRHku;*)H6)&_vW3@A^}F5c*>dF3lexI -3vR2)_H|~d^}c2q{1?AlR&nk4@UD -(OUCb^c2CN-Q-x<$`aW!Xe@^JZizomcmZ$ud=JGI*hSF#%97nfe(bKOO`-`~sMWB%UfM;J1<}(SL_yT -@Jf;%^0)q?Q@c`;3|s>09FRYkt!qW)t{<=4EniXR=AtROU!#v?$lPibH3ye*G-**`QR_nkVRKn}_1Vf -wAu{)*NTX{^+|5e;}-hY6&Y!L4k-Wxej{7W*E-<+1S);xWil!s=VOb<1awUumDX(F@YT87Dq@xZb46c -2!3E>PjM4g)ee5poWE(a%JLIidzE<6PaeKiJYU!3CzJP2kcNTRMiWf4DJQD!} -gWIo0zKV|-IAFZri6fLlBmVC}_AmHwA|@0M(QP;wmj6x$9RT&J2;X`+KOK9-R9swhz%YNHup^~hFIr`?G${X~cNJKHzj^L+z*@}LcM<6C%0wXF(6*%O=46H13~uK1l0M)0MmQ- -1MJdeq}_{PgGF?Tayn4flQ<_ZHM`Y)_hdh&s_9pM84t5%pcJ0p;QH#^9yS-)juc!Jc4cm4Z*nhna%^mAVlyvwbZKla -b8~ETa$#y)#2fCM_u&^ux9OkhR5`hx7iTmDNrg$D|%f|IO;+vvBP%TaHF+qbrtW;<*qzC$+LHoZcJ#e0(gnq7t -8!t#|lPtQz_qFFx~)XsNAy9*rI!Z}ZFByN`c(y!p!&o3ppS{mt*bAB{#u$*pB@x-eW>r2k>W;4>bN>E -FdxB&Y0P_|7t}n0S`fNwo=*!=n`w*70R2?N$&8FtS1`B?=Tj8xi71t(@WJ$ed>PN4M4L4T)NUHp+3NW -O|Lfj|4np62Qw+A1piS8rh01mx{oux^xm)+2Zywy2=98M!2mo8M|?SzvITqqAs~9kF2(0U6)L*2~P>J -0&;?D#UQ+cbX%$LYY-ic;SCmyZds+RLvi7*#t~Nu_gEtsK`BLgPEtt1PFjZnMa>|-1?bw!E?wpCuFk|BeY#f;+FhSgB_k -VtuKF5ziGY!~i(xE1KPBv&1rlk4k<;WT-CR29U%3=$aKx>Zs!A=f=Q1>ao-gRlG$rDjp@&Bg^-*Z_)% -40Bqm(;fkqI)kQz&^cBaSyRAU7nvrwW3_lLMF{v;%h6+UYL9f8K-z9=09sCW|WDbg3S%zY{V)TdTr+m -QwxF|NR*F8K(d7rb*gRL=o+e6$m7~#bsI{UtQQ!?&!j4Jg}ifQ$o?oQZH|{Wcav#Kr)#&-{d6f|pN`1 -X^Z1O>Ycg-T%h{w$gdalf)F1BDg&#;wS)&aJ5kcnovJe1p2`HbsjO$Wksp2m_+z%Sqi0J01sBr0%61oQMU>g!O -{Hf#k=okY+0|@!ca61;N|2KGy?$uYr~ab7yAzOy5t)>zld<5F49Bxt2w(!>xr)=f8u2=u8h%UGRB1dr -!YFOU=^lDjIBgL$0$o|v~=}&`byEyM1}(3bdcNYG-GOcbi#)G{^e94Q6fFy9xoA%HWM!u_P8@w;U^r< -(e{(ZH_f3av4th<1KjAfUOSTYJXw$Ltxzk_B$yHA2>wgiE!-Sgt$-4W73ZRTf@2i4cr0#vDGn2qZ_2o -3_-jUgvw~YU8DEa+IjWiC<5Wmq1!%HyLz8w%Xj#yUybSLHN4IP4vF1^xg7_z>zY^aMf17p~Jm5ETwoi -@zKeBj|Y|y$CK+V5UEe}RI;YIDFY6Egou7aA1_3OMbS9&GpV>51o5XyhZQvuCk75zk|njBW*P)g-H_D -gNKT?e+#?TS+3Oj+YZ+lWJ%=x;`|Xz;gCR93N#{D0hv;`r{K{R82Gr7Z3kxa9aQnOajl3Wc#j -DXRWgWq>Y0@y)(3Y_ypZHLFLOcdX?0tcOv02&kC4fpM!MLPTvZ3rtEx(Zhw`)7zS?^@v|XznR{KWG0U -n~x9pXF2}&mtsjOS}4d}>}$g&epH_^$hTz^R?{*2n$Z=rY1El9Q2bB=*J0=(MHDgK)^KXtQwCcCiX2# -5V%U@fn7ZKss&k5^RJaPxQLFA<{6Oe%nph{`og2Xue?yYR`n{$9bUbI@4luqyFCH1<3wFM8g@8Vf; -vxJethvMZ=$jtH2a|b*_JkH`LSMkQo4Y3%tabq23{6e_ZAx!jb*BkL7BcEAsp&>MKed2b;p}!k=f2!z -JWn*4`02S1BjinIyxs`66T%c==VmM<8{ylKYJSbZCqrYmmmvm-SfeXoH?dNy7e{!H)lkM7$=w#r{$<` -kn2c#9lzmv#c)3$KER{|dx_kxH0AnAt`Y(~y5b%~mAm?5g`x?ZQA-0^ZXkC4s>0|0dQRieOVD}8svB% -;x5L^+lCriRC#(8;@i-_Yzb0rEW`MD>Ox?`QBKXiy=q;yMax45C@&<&J~my;8*6WFNI(JN+p3-S1POQ -jwManWfOB7^FpG?{3mn*syN^1MkDJTbzXy83fR5Tur;M`1y)OylBP~2ZJI3Lhh*e%$UY -k^_Cuj5LqvppG8FHtT9{AVw4a-pf{UQ<50g#r{+qFhF_+Ebvo+AKV8;4$q`Nx5Y(*BV!6fWWKe6XE^M -Sl5R%^ZB>$(=U0VI<*YBF-eV*S9y}}*q_GaGz1Sa~VXS;Nx{{T=+0|XQR000O8%K%wh{T7%r<_iD-xF --MrBLDyZaA|NaUv_0~WN&gWb#iQMX<{=kb#!TLFLY^bWp8zKE^v9(T3d75xD|f)ufW*ju{6V_-A&VNx -$DlRNjA+)c9WSn+eZl^Aqfpda0$?g`q1Cra}F*9L@IXDcBa)sB9XwsxqRmeuRFD4i^aM(eJ2(RCU>ps -4C5=U>fVTj{e5=ko*TInXJ;4eZYx+N*1WF`)2jR^>?_rqjP+VnY;|B;W7=A-m{CkNl`Odt%xt+~jq;1 -xp6lF?>H(YKcm;~on|F(*orI*s;C&q$%C2Ks)JvxQgvZ>^t~)ZQTZz;OGW; -y8Sl4$yg_Nb^Bd@GxzHeF0l9=9Pl%5_(oWXVdJg)KP>HpZo#6c9Nd+6;Rxo6tb`_0#uDRilzIL*p}&T+%&Y3wE9!SQ^Z5zONR1dF)M) -`|Z#InAwPllcd?$?PG@^I8{6#flxqKYojar|9283tL06A`;et{?$`cf;$#p8=6Zg7j*)0z -_K#IJOvXDq_l$1P^aOaNzKJS{J_y<_kAPN}8jpP_(#I7|td)?skKuO$#kJ{LN^c>C0h=KfIeB;su~t-huqboUhroblbFT_>XoXWUDlB8` -T5&@x;j+XQPNLWvg6;Sxx$Ad~?^F*OP;0BfnQTC-u%ps*Jaf;b4;qdP1e`#~X%!c8*w&%y;k)k3Qh*r -j?M_;BC&;$qdjE{Aajbi1TYb=Qj^B~7wo@x`paWat+f5C>0=7SR2Jdi_`Td5F{;N%j*b0duiT%ZaE1$xEl#GKwv4; -mC9_x09X39}cX!Q-St~7h` -%X?g`YINU9L4XDhq1q9L5Lr>xA%+EVs<8&Ns$*~AM;8R4m>Nq%lK7NOa>%Bv`)aW9R&KWJ`MVDc^ahc -Ve(F7MZO{Z4{LO!3djSzswBcZ8=#4C+cFx`k!k ->nJHa)S)BE9`7LZf9(?9L7XWci(f^=_Ze65B#{?z -p6qUzEJpB4TVX54XtBHe~`YvIB~%gF#c;9igD6qSoepX}9ru&kuTt&!`BlAh((Y)iP-s`n-ai$4$Rvq8lA^9_(5c -KIy(1FgXQ_6i2EbkSisU2FhUPEN!9FuL46eMKWf*YTs+ap+2eo^bfg$27mJw|^*Xb7c6F$)N_~6fDpU -V+UPRWY*ZIO%2Dk1v@Fgo~>wNj?%Uo!!+_&@j5ed@T$^G~iHsgQCJFOp{+)1xtEB!w@RU6)!i2WfGSP -h{CoQ5KyolBoz>QKL5c1}n5#j5hMo1e2wHVfMbN05XpxqPu`AR#XnS+XEe=+rAx#rV0EGdy>_%>~E+P -P7LoP8b954ra*E+NzW5ta|39~td -nJC=B{`?ryqShfKeukw9@F6qgDj7qSM+|}gJ>QD*qs@nW=nxDO>1;0_+h_(->MkGR*PHavCAS~u7q#rQC3mmjSF4P(tdp#6Lh1I$})>E~~6e2u+g@CHD?se>JBnD -%WC8mNINV8w@G0k)e46>5U#aYwuL?=t9ff|@Vl{T2=A>LueRtf+A5S^mhi{!`3%;zC2;EbT8kXC4UY_ -;$maAVJe{XRYw$x#k;6OcNaq(TKHoEaWW?3_HoM&p>K)UtsP)0AT7@VByaj$}6~HqvVvFi74oSmF)p2 -zybYB)GQ|du`aRhm3YU-i^zfsOAj{P1g>w(R7EDMcQUX7*^S_ADi8Q^@#B-65{J+v}(W9Gy=Q?&s||i -aMD>(CS;tF4udbcnREX6k+i494QTk`mfaMi!yXofSbAo=Nav?eV1P_fbLB-t2N(-P6wJu%ZluCvKRdP -g$o<&O0~)bf@!yy?PFLPTB6UrT?<8FMD2ZW=)CCD$EFIU&p+vVuCuPELJ`U)iZs -_Bl%S{HBOZQ~Cj0#cwv~Ylhy}`5BeN#r0aQyf=q$OZe)3n~g4m(rx8;)~CD4l(YY^536`#>jTL$6d?7 -`?}XiqmR}laGVcT~wv|LHbZc4f!s45j;;HWrod=5Br&!Mn`#@0wQF3vKcy$9Xn03K)R -0+tBhfLR5%@HPmj~5j772#Y96rHSNWWi2bF+j(ZP{9+u75a8w)aI+D3)Q1WL*`6`uPu)kqv+FTj -*8zP@qLbPnNwI`Ud82Y%7 -8ksNQ!i#Zc)-Xi9pR2~@qu_lf?gwRM+>Ro~f8HbfIY{*29OYd<&DiZ1-}(M?~0*%MmtK!rsgOoYUZt+ -YF9C`8tZ<_Gu=OVb*KKAbH0rg}G5KBBKrl^!ELXCV_=PLK(AZo*9dV9yv|CC=u@hMw3nm<}2t|8P;c` -@#tezVgwSW=x{yEm{t0%Tl;OO-&iTz*JT&lf#x%Zu!!YaC1ICm&r#iNzwJcf&&npd6C^9JB0*#cw&ez -M8o5eFfKEmQ+(MBv%!?rBka=D&7Xml?vgyi?vEiynq-gu8?+zW0LGTYsCFc!Jc4cm4Z*nhna%^mAVlyvwbZKlab#iPjaCxmedtci~lK=Zt^o -nII1tasGjBXf0f|GDZfHg3?8*qedwXrs`>zRW_Wov!S6|&X84U5^zDi9KXabU@nM{JcT*l^#-P< -&VgKMPj>0+n;zb$DViv{8)C+^JoB^YheF?G&^CI^C^77ogu$bUy10uW*gW)y{qA>?yuKFheI?h2=5{{ -Xtal|d8JLGJVWiz{Vde;ANcyj#l@byX0y?FD!_n~(IQ!x#EO%s6c6dni_MLZ0rFem1PA&bW>oA8t~m{ -rCTP~sP!aOU$QxaU42y#W9qwUhB^``4FyZ8nYlJme0sPn~Y}4us5y_9&iqXLFD@TA$2kJn2e=)_Jf0r -gwVj_In^!!Vz#5gxpMw)oX8ke7OF0WBu2`x2^An)e`bIfPOL$9Dx3Mxc+e{@_%)jzJG -wHSN^uGxTW1GeQ@2Zn?Lux{79{&aKGJy?NP+MDZ7G@?P?9KhfL)Pb{N+Bou8qa((+eRX>L=Ip4)9`VP -~+2!HMNv8$NaD4v9j(dK1i4r5`H7Q}Uv;FfSW1u&DxdSgxF1EMSTcGo8!XL -hizmMO7+5OIU%dp3|2{RK0_a&_k!*}~i3ajO&`4Er5Hv^yyhZS%}m~L;ufa3GL<=%Tq;E6K_XNG3F!k -+l-nQXGT#|+0{7DIA`g3_*{Ojhbx-}b(*n)u2}rS6tdG#nta+f31MDl~WNG_Mz^A(erkx$~Tq*1-b(o -hJG(Hg!50{q3dnb)pB$m~A&Q+gieG>xYU0dhEkUXJwC;7TW6?VG;!Kl&Z+!7#c8L? -sm@{dM>xRvG6uU1Q_<<~HPc^4Zcir%bikino%VYdz5XA)qrz4kzJ7Dmd;9MF@qhl;$%oUk^S}2mF0cO -h$3Oq|(Ho9@KE9g-|Nb0Kqj>gbl4kk+mxuY+jm@p?o!z}(UjAykH|zgxp!M0)jE2G1nDJ=X|Bi#mynb -M-Oqc_(8|(o*64Q?Sz(Qa!g&)sOd+za(MrL(u_i@!iV}=A5WCx_d=Nwl9p<_PT%grgzf2xNv`x!1b21zBN!}lx}O7W~1hB2BZo1{n^E(dw$lxl-dVQ -=4$##|Dg(lWARH996hMUcv0f2Or0fe*s2__H}4Po7&oO -n6B7NO4;r2`h-j`}yGOl~;|~=M#$W#Wx3+1$SnK@9J_au!#PR9dvjV0j*dtpc^a!rnF}}B8hVM(fDj| -e;J2OeA_>K_W{E)HK5T!v;)9swML1wxCsaFh~0Z(PDn({8RyZ+xt(Dw?*hkLo4t-9godD0#l&9Zv -k3*Xz^n+ji6}tLahOOA5pU$Bd_9Ojlr+d7#%11U6em841wnQH?Hh(Mt6j{89$YsK<1B^L1-7>(#$W@$ -c*MWBuGla58HH)5_#wpG$FRE`#H?ar2xVUw!dzo<$g3ez2C^`wk`ZlKt`2^JG0`EFMR7(y>@@R|EX7p -AgngnxE4FCeyJ!ROhfYf_K#&qK4FKHs)Gu&?3!AYY9*ZxX|Y3!>yW -NGBw9ykq*ExNARyQ5)hvmDRW>hH5J}+9Hd7vswW-6>zlCi(goN>nm-y*AnilYX!ypW@xya&RH`2DWA? -Iu3)(b|Qg~4YI+s?~ubsF@E0uviqE)yr=FoQIMjUF^ebSM*Mu_!?_3Uje>;VjV1@zB9Qg*A{On**E)4 -Q2(4*{Uofz7zzOE+vv#gMDy-`=CT{iOYwP`7_al9(2 -6vSACsM0sk!4zumlsNTG_}tpSayQ6cJim#fdyX{)0D%dBG4T+kRM6!Fl#AAkrkFxrANr6>!^VLn2|1Y -AB+es$9hPt=Hkh=1wr>0`CwxJ2uypW>O}#lS-H;El3yE^MTS7LqI>fd$blhfLoGYgf>z^Xaf*p0O -ZPVj|H$^8idUoGLaHq|sQK+`hBP6{H0@QQ~03(2G9Py1_^ovWW-Q)&ZfX@jdq~yVOkY!JE#oqTo2K%2 -NLR`|s=uc5fk=*~X#+)*RyaE+mKFoq9poZCiCj8`ZCmE=8{6Dpu_dYv(CX&_*~j@P)8ZeH_G{f)wES< -wmDMz_xAH?&4uXKsqM)po~)6S}ItL(<{ZsdAUIp4~X1KL91~q{-l+Z=0UR+9Le&DDV%%W~l3BHBE=?q -+_I}RXLHo=wqJPfAkLgWZ1$Hzz!`~deIhi%$hyMo<2h}9zI$?Zfg0m8^2!(S+ -rBc5&}vvu2OOdXtTbM6x@$QnPDhI)Dd&)Ww7sj4MPZXJSHU)$qqv5jF1GXnMU$Znq1|Y1p#I8G<%0;M -^{mM9LL?EmmCbeuV_9Fy8D`uqt`gBLFh*CIckumEK<{7w~J~qzP=cN-A^I3iMW@9GWv<~>5wOd;!_(i -VBS9WLW=l9FS+A&sLFCrgUAj12e3BCtr#LueGuvSScVWpw0M?~Q+pglzUfyE8NT|BZB#e;D8>vV=auk -BNdaqy9MK&MWci{>qLdl_*4`-5eZirbx*Dp8rw$DyYqD0~YU$(Gjzk;4_P$mm*4QRMT;WgVEV&3&> -+m1A|N=iA-wSCY*r!TmjtRi0u)w~mMW|1sDKa}DiQXvqhmvAhkd4uKV;>McUh~&QUl#)Y{xB?pL})d6!IA(n@1=nIlk3`Wb>eY{0Zv+YHwLA~x*$49QYpxFj$|lLojZ -QXp*!B@b8e$6|chYq8UO7?mkh=uL;dCoGyHzU(^)IA`Vt#3M6Iz^sm@S;Ub#gOJ`Ol>q&6*@S$i+%K? -eLFq|+2ZkGi7ek!IAayHJf^zqpdTCfUXplom5rl?2+%?#UUbLJjSfWAky# -M6H17nl4C+_zCGaR90(g*O5ybKfHv=Aoss`uO*J4}jZ*03Dgm6rg{5pW}Hozk}sKB@o$N?PuHSuS~N| -_BBLhmDJ!$eUHXfvwGa6$Grwz>7Pz_5Kzi(;WBDkr?mgsax9uG}@D-Db176wQXi#=!=oyr2OrW1C7y> -w3!>sQqHuU&o++bF0{OmQ^uyb<%QDYP(p<#EDG8#X!mRQIG|~AR#){*)gI40vK?k>B%TbB(!q3OPaXSioCwZ+jd(8AtN}L2o -xFYN-k{4)Zuq^&P7Cz-BaTV~IO0`5s1d$_9kd!b(U4ve*z%l&K`;2tMVYB%qvHvAIHx?O%}XB42@E3p -%C*wKOS?aV=m@SnbniAS40u!VGutTgo+yXCzGMY0>@)2*WFtBoa>WZy3eS{-wlN3{5gLnTGKLlEc0NdhiVo!qcJ> -J>O>--B3NFlyF7G>v$Xh#A8_k88?o!W!kMK0V^q{0CTXZWDU!1)y0*IUdXD-Jzd$o5@3aqDC)s3rg5g -M0#P<8Y0EUQQX~z)DZsR&5my)xR}6#5OXhAipYa7NN@-*z_`tjmN3Ck1O(we$Yr3>ir3sE9nBDII*-B -H#Cf6-#SK!AArAkMx=Q)rIQ%?k2g>Rk-1qXgddzWgyT%q3a$y;0pvle5gd3+jLOjk^~uAl=%Mk$(Pkm5 -oaNK?wlipi=3&ZzhR)^^IwhoV_05@vgzD59S%9c8aA9+$$X6OxhXvr)Y9`U%5bTXs>yEmxo$v;~GgO!4(pO+YrT -n1ICi5_@2{PT*BUUzSM=p{2Vj>(|QCU{{Kp8S)@X?jl`H<>8f#_LtZnL9uLU3nay%8Aq1IwKln>4$>3 -FS1Cx*8!Jn%Qr!YTDz@xo!2b)yPgUZ~u@PFhK*I2uL!Ef%*1k%@TkRgK&0d4v*ZFuXs?tKA5Xggj7ZJ -8NJwb#CRmd}O57lOG10MQfLbD{k576yXp9TVSoS|^B)U+ZB2~J=2$H}ND?@vX6Fz~RJBayhm3FEt9Bs -%mw#E?NXz*5>3Ll5UR;of}^Lg*V2I`sm5j3xt-Qr+IlEbaDI9(T#e!2n{4M-wOIkcrS{OgY$6#R(E{% -CJ>dZ}fuc5luMY=-jGa@9onC+@ESxVUtPT5m3D*x(d)PaoQd2DuB2G(3W&M*k`hqaW6?YFE^Tb3YyC} -zTD_&R}P-!)g8z2!O%;dWg}-%n~*4Fi`xuX(AMXs6$Wvu;xcR5S+B@6v;&9CW!#;{Ian%JcC=*IO_D0 -E8mmCDJdyYjI7XK)J``P#ir>)tY;h7~^W_aTaXVQWG-On@X80*H71^>cI%BGCT)joHERAT$IDYI_hK^ -TEsfsE(imq5sm*Rl^64p+z_YvUnHjP(?RYR4)IG37>D%+H3p!K_5T0BN98+4~hE$_Byx~#`=QkG^nAm -&)TNLAw`h`899dK3}FB?5m*=t1ki6~&HKy+-Y68-NYhNT>oMOX~63p>%wM*(yeZtY18%b?Tz -18>gJVs+SI2Jz}J&&2PB=4>?zZCLNpc3PEg;0k*yYgDQCnL4k9^JELRq#z7{V7;LnM0fne^K96P9OjcRx8Ltsah|9C1Gl%tQ`-lqQ7i`XgAj5=vpO;B$>wclSEt`Trv|rCR)YZ -*7;|U`mVame}E=|Z6;9!@D-GU--i%XI#&rVD$Xi<2H%a0AR#cGyJX)#Kxchg$+Ni -2m2`O13&iU#n46tuP%A5g-wLy8=9!hljK^`U=X@WJ~!2Wfp4?p*2+IAASYz|`MU>c4OslP=JdL*4>Ov -P_FtGj#wOu^KLF>Og8x!l`r&@&{pF>q)~=ab1G0($??GGF*7>`PDw_ta7W0OuMGxnJr@kyeugH8QIKM -cO~(e{=we=08mQ<1QY-O00;of09jjB)K=BM2LJ#Y6#xJr0001RX>c!Jc4cm4Z*nhna%^mAVlyvwbZKl -acVTICE^v9(SX*x!#}$6(ub5H~qFjnrYXkvYvw;gZsF74MYz2M@3}U!@mYi~SXEQUa6@mWvd}l7pB}F -;zTT9sDWzPBL+%K!@)>tpd+Qv*&{NhMc;fHT~H>s?tMb$NhH%2>=PFiWbl6o@ncK>=J@CtNZ)s#tU(- -?0WRZQrypw^4`;rDme8v6`O_$=Lsb+TGXtye3t6u(Yy&1bp4#~#8z#Wl5GtK -lE04ntqu~CWV&HiDzME?}DJpG*?l;7UW>N5dSp2a9h*qQuHJkS(rw+!0LXOT!7?$OBTZ8zweFk`<9$A -m3aH%y%T%Wi9$A_b0i>908q_lgz;MVEzV?9?15FAk}S@=^i*{^f@rlPL9NmH5^!+4TZn&jo~>&dy%+RAY$iy9yRyBdaIl)DB!`gr7np&?_@!ws{lIl3VbppK&S -0<;Q?U_b;X8y{H2>Hh3-QTBuSZ-~AR2XcCE!f9?Cfe0g5cKlbFoFRl!2w#8UyMz`JSjbaIvtF?FVt!u -#lzVu#;N+(x^vr^TBbT-XbHQI;ffrFd-zzPGOn{NCz%x1WVL%hlqoroMs^)SS6Ka%V=*l1g-5n1m9L| -w%1KpuMw}Ms1)p3k}Irgy|FIFNNW_7h1sH%LIs~!>JfN+TQDSUPSUcjt)+^NtpU|OtEMzPD11l{;4~( -K&L9i~*Cd@6ANGC=z|xRoC$`o+sFI_vx3X=?0*-}O{tCg?`5?um6L&uyV$Y1O8;q9$LoEBPa%|aWP5C -#`YNW$qwDBe(1>63Xqx`*{1heOOrj@^A9^d*hcM3gKeK0_<*3@e2D#hL>Y#vQK2^@SYh*<{Yt#6 -h*TW`{6>8EWM?6&90;s<=uoaHWI%E7ep+{Cz2Aujb;rRK+gDzhXU!HlD|LvE;2n#DJUa;KX{5IH+W&e -c-}pv0|eAD&TrMtC$z*W;&ms`30H|;8iorXOju*oiD#b(PXnE>7x*C@!wsds{A7x)wc0i|=y -WE@&W0r%UnI#rH=@sj1Eq-ZASHcTsvO-G#WMg|r*@>&=%-L>fvpDk0a96zI5dZTxnSk{ve&eu4L!U?8a+Jvn*mpTWw>?aPD5uE!ob%QOm?H{#|z@G0mCY|cb`7(~sb^YuD8R0zL -RI@iMFuQ)fN(ztf4-hSr_8d&mQcMgn|%r2`;+e~vX5>L1PiX{O&Jut%7iC!_RmmWh|fGll!gnpgoe%a -TH`d;@dk#RBRpIfUOm+)aB9Pi)pbR|4S&5)IJLBaE6f9NpkNu~TGL|Ow@kO$&J&6o!2|#*6B-&XQf^9 -lf|Pn1N_{KQyoBK-=8H;Ip-U1^L_i2SW*LG35?UU#hJ1u-aYR6ap2aB+xv);5gTBo(W5=x={Vu8`>x_ -oBhh7owS2b`?n%-wAv*X)5d1sz%pWl3I>&CQ{>rf3+XO@Napp>h&d>Fz^p<$!6r -GXe908`507c(LpGQ<67&{Q2!~+Tx4?!KCKx& -wp|nbk{1&G$p3xmw4rh45iK`+M&zT$74Ta<+#Ur#RI||cc9$`&WsZD2N6X$9Xrz|wc9qn7`3(oz77WQ -pa4J*Us7Z*yNd^P!(8;cbMAL3wkG~0YN@c4D&tRwHv#Z4s41It^F!4o4qrMV@I8c!$6klvB?XgWy8oP -Nn8-{$`duE!;5xsQt(7hcSE#g-pGXsRjCLdoh}8g$&tm38UqL{FDC*HxaIsSIVl2+szv^{KFyj-F@ab -p=krB2Z)bio05>`DXA=iL%2pJm(b04wsz`SMy%+V+XkZ#H&ZqBrD{6LJBe_86aQyb)(H1UGT(Gxd=TU -()>DM;s$c2OeaqP4GXJ;lV`cFOyh^?S0T?2K?nD~dVLy8w`&^M2SDZhJpVKh -o@}BDJndUQ{PPw|o?#-m3!yD|Whhh>RXKd61?y-E^v8JO928D0~7!N00;of09jja_pjHy0RRA20{{RI000000 -00000001_fh7R|0B~t=FJE76VQFq(UoLQYP)h*<6ay3h000O8%K%whB^=tD)dBzjss#W56#xJL00000 -00000q=60t003}la4%n9aA|NYa&>NQWpZC%E^v8JO928D0~7!N00;of09jkj$G|nf0000U0RR9D0000 -0000000001_fffb;0B~t=FK~G-ba`-PWKc^10u%!j000080LuVbTMcdTY&Zb`0RI6102u%P00000000 -000PTSQ2LJ$YX>c!JX>N37a&BR4FJE72ZfSI1UoLQYP)h*<6ay3h000O8%K%whgjc!iDG&evc{cz68U -O$Q0000000000q=9b;003}la4%nJZggdGZeeUMVs&Y3WM5@&b}n#vP)h*<6ay3h000O8%K%wh#Uk5%2 -MhoJX(#{y761SM0000000000q=C>F003}la4%nJZggdGZeeUMV_{=xWiD`eP)h*<6ay3h000O8%K%wh -@S@g#N)Z46=r{lX9smFU0000000000q=5@1003}la4%nJZggdGZeeUMV{dL|X=inEVRUJ4ZZ2?nP)h* -<6ay3h000O8%K%whP|GebiXi|1m~#LC8vpc!JX>N37a& -BR4FKuCIZZ2?nP)h*<6ay3h000O8%K%whkG}tUb_M_d%Mkzo8UO$Q0000000000q=9f*003}la4%nJZ -ggdGZeeUMaCvZYZ)#;@bS`jtP)h*<6ay3h000O8%K%whyO09K8w&sc7a;%uBme*a0000000000q=6n_ -003}la4%nJZggdGZeeUMb7gF1UvG7EWMOn=WM5-wWn*hDaCuNm0Rj{Q6aWAK2ms3fSzEm6vGX7b001& -00015U0000000000005+cduspyaA|NaUukZ1WpZv|Y%h0cWo2w%Vs&Y3WMy(LaCuNm0Rj{Q6aWAK2ms -3fSzD|mDhX%+004sk0012T0000000000005+c*L45@aA|NaUukZ1WpZv|Y%gPMX)j-2X>MtBUtcb8c~ -DCM0u%!j000080LuVbTgO#3@r?)o02v$r03!eZ00000000000HlF>b^riyX>c!JX>N37a&BR4FJo+JF -JX0bZ)0z5aBO9CX>V>WaCuNm0Rj{Q6aWAK2ms3fSzFom@aG2$000ak001EX0000000000005+cNPhqT -aA|NaUukZ1WpZv|Y%gPMX)j`7b7fy+Z*6U1Ze%WSc~DCM0u%!j000080LuVbTMa&|9YP@h05N$003QG -V00000000000HlGAiU0s`X>c!JX>N37a&BR4FJo+JFJoMd?cwb|0ZEaz0WG--dP)h*<6ay3h000O8%K%wh-O;>)6a@eP{|W#A82|tP0000000000q=9&~ -003}la4%nJZggdGZeeUMV{Bc!JX>N37a&BR4FJo+JFK}UUb7gWaaCuNm0Rj{Q6aWAK2m -s3fSzGJuA!B|8002A^001HY0000000000005+coz4INaA|NaUukZ1WpZv|Y%gPMX)kbcZ)b94b8}x}V -RCaWaCuNm0Rj{Q6aWAK2ms3fSzDNCi2n2q008GT001BW0000000000005+cThssmaA|NaUukZ1WpZv| -Y%gPMX)khRabII^ZEaz0WG--dP)h*<6ay3h000O8%K%whvI1y49ti*d!W;kq9RL6T0000000000q=Aj -%003}la4%nJZggdGZeeUMV{BbYEXCa -CuNm0Rj{Q6aWAK2ms3fSzCJ{NTMhS001T+0018V0000000000005+cbMXKGaA|NaUukZ1WpZv|Y%gPP -ZEaz0WOFZLVPj}zE^v8JO928D0~7!N00;of09jl7duma@0ssKS2mk;d00000000000001_f!O*00B~t -=FJEbHbY*gGVQepBZ*6U1Ze(*WV`yb#Yc6nkP)h*<6ay3h000O8%K%wh2-05in*{&>vJn6PBme*a000 -0000000q=D7_003}la4%nJZggdGZeeUMV{dJ3VQyq|FJo_QaBO9CX>V>WaCuNm0Rj{Q6aWAK2ms3fSz -Bi86XIeE004O>001Wd0000000000005+cvI7ACaA|NaUukZ1WpZv|Y%gPPZEaz0WOFZLZ*FF3XLWL6b -ZKvHE^v8JO928D0~7!N00;of09jikZx9Ik2mk<;8UO$v00000000000001_fmsg$0B~t=FJEbHbY*gG -VQepBZ*6U1Ze(*WWMyJ?XD)DgP)h*<6ay3h000O8%K%whWDMTSumu1B(-QyyA^-pY0000000000q=At -b0RV7ma4%nJZggdGZeeUMV{dJ3VQyq|FJy0bZftL1WG--dP)h*<6ay3h000O8%K%wh{#IW~@B;t<8w> -ydAOHXW0000000000q=A7R0RV7ma4%nJZggdGZeeUMV{dJ3VQyq|FJ^LOWqM^UaCuNm0Rj{Q6aWAK2m -s3fSz9b^5>VO#0055$0015U0000000000005+cuOa~eaA|NaUukZ1WpZv|Y%gPPZEaz0WOFZQVRL9Ma -CuNm0Rj{Q6aWAK2ms3fSzEL~pz+`V002}30015U0000000000005+c#U%j%aA|NaUukZ1WpZv|Y%gPP -ZEaz0WOFZQWo&RRaCuNm0Rj{Q6aWAK2ms3fSzD@O%Ud@F004Uw0018V0000000000005+c;U@tAaA|N -aUukZ1WpZv|Y%gPPZEaz0WOFZRZe(S6E^v8JO928D0~7!N00;of09jk3koovT1ONbO3;+Nj00000000 -000001_fmJR60B~t=FJEbHbY*gGVQepBZ*6U1Ze(*WX>N0HWn*+MaCuNm0Rj{Q6aWAK2ms3fSzA?UNT -#bE000Pg001EX0000000000005+c*fIeCaA|NaUukZ1WpZv|Y%gPPZEaz0WOFZRZgX^DY-}!Yc~DCM0 -u%!j000080LuVbTWK>c!JX>N37a&BR4FJo_QZDDR? -b1!UZb963nc~DCM0u%!j000080LuVbTbR62h2jPP0HqcH03HAU00000000000HlF>UjYDcX>c!JX>N3 -7a&BR4FJo_QZDDR?b1!UfV{0yOc~DCM0u%!j000080LuVbTXpFM(rO0)0G1X203ZMW00000000000Hl -GKW&r?jX>c!JX>N37a&BR4FJo_QZDDR?b1!pcVRB<=E^v8JO928D0~7!N00;of09jj52c?+22><|s9{ ->Oz00000000000001_fk18n0B~t=FJEbHbY*gGVQepBZ*6U1Ze(*Wb7*gOE^v8JO928D0~7!N00;of0 -9jiw&KKa_1ONaW4*&oo00000000000001_fjD>p0B~t=FJEbHbY*gGVQepBZ*6U1Ze(*Wb#7^Hb97;B -Y%XwlP)h*<6ay3h000O8%K%whIZn|;+6DjseHQ=#9{>OV0000000000q=8m_0RV7ma4%nJZggdGZeeU -MV{dJ3VQyq|FL!8VWo#~Rc~DCM0u%!j000080LuVbThnz7E;|7L09pe804D$d00000000000HlF!gaH -6>X>c!JX>N37a&BR4FJx(RbaH88b#!TOZgVeRUukY>bYEXCaCuNm0Rj{Q6aWAK2ms3fSzD&7p8M4T00 -1Tj001KZ0000000000005+c?S%mVaA|NaUukZ1WpZv|Y%gSKb98cPVs&(BZ*FrhVqtS-E^v8JO928D0 -~7!N00;of09jl1K{v9T0RRA@0{{Rh00000000000001_fdh#F0B~t=FJEbHbY*gGVQepCX>)XPX<~JB -X>V?GFKKRbbYX04Wn?aJc~DCM0u%!j000080LuVbTV1H44yOhH0OJ_|03-ka00000000000HlHCiU9y -{X>c!JX>N37a&BR4FJx(RbaH88b#!TOZgVelWNCABE^v8JO928D0~7!N00;of09jjltDV9+0ssJ11po -jf00000000000001_fzFWu0B~t=FJEbHbY*gGVQepCX>)XPX<~JBX>V?GFL!8VWo#~Rc~DCM0u%!j00 -0080LuVbTX2Jdbsqo#0384T03QGV00000000000HlFRlmP&6X>c!JX>N37a&BR4FKKRMWq2=NUukY>b -YEXCaCuNm0Rj{Q6aWAK2ms3fSz8-dbR&5a0022Z001BW0000000000005+crj!8yaA|NaUukZ1WpZv| -Y%ghUWMz0SV{dG1Wn*-2axQRrP)h*<6ay3h000O8%K%whJUcH~wIl!lvzGt>B>(^b0000000000q=8+ -i0RV7ma4%nJZggdGZeeUMX>Md?crS2aV{2h&WnX4#Ze(S0E^v8JO928D0~7!N00;of09jk#z}3a-2LJ -%SApig#00000000000001_fmO``0B~t=FJEbHbY*gGVQepHZe(S6FLQ5oa${w4E^v8JO928D0~7!N00 -;of09jj}_uI$r5C8yaHvj-500000000000001_fqT^f0B~t=FJEbHbY*gGVQepKZ)0I}X>V?GFJE72Z -fSI1UoLQYP)h*<6ay3h000O8%K%wh@MJ@%-3R~xR~G;PB>(^b0000000000q=BmD0RV7ma4%nJZggdG -ZeeUMY;R*>bZKvHb1z?HX>)XSbZKmJE^v8JO928D0~7!N00;of09jjtrH&%w2><{j9smF(000000000 -00001_fy(az0B~t=FJEbHbY*gGVQepKZ)0I}X>V?GFJE(cb7OCAW@%?GaCuNm0Rj{Q6aWAK2ms3fSz8 -W~Ai#J8008z00018V0000000000005+c@c97%aA|NaUukZ1WpZv|Y%gqYV_|e@Z*FrhVqtS-E^v8JO9 -28D0~7!N00;of09jk1nmd{v2LJ%T761Su00000000000001_fu;Td0B~t=FJEbHbY*gGVQepLWprU=V -RT_HUtei%X>?y-E^v8JO928D0~7!N00;of09jk|a&Ce80{{S*3IG5f00000000000001_fdd8t0B~t= -FJEbHbY*gGVQepLWprU=VRT_HUutu2ZZ2?nP)h*<6ay3h000O8%K%whwVq`iB^m$#Ok)559smFU0000 -000000q=7mL0swGna4%nJZggdGZeeUMZDn*}WMOn+FJfVHWiD`eP)h*<6ay3h000O8%K%whnQSPd&kF -zmFev~4CjbBd0000000000q=B0x0swGna4%nJZggdGZeeUMZDn*}WMOn+FK}yTUvg!0Z*_8GWpgfYc~ -DCM0u%!j000080LuVbTX{ifu3-QG0EYko04x9i00000000000HlGhFaiK@X>c!JX>N37a&BR4FKuOXV -Ps)+VJ~TIaBp&SY-wUIUtei%X>?y-E^v8JO928D0~7!N00;of09jjpU#A6o1ONc^3IG5r0000000000 -0001_fm$&F0B~t=FJEbHbY*gGVQepLWprU=VRT_HX>D+Ca&&BIVlQ7~Z*6d4bS`jtP)h*<6ay3h000O -8%K%who`0Lp5DEYQ<{{{d6)f4~#DF6Tf0000000000q=9xo0swGna4 -%nJZggdGZeeUMZDn*}WMOn+FKKOXZ*p{OX<{#5Wo~wJE^v8JO928D0~7!N00;of09jj~Qj4QH0000!0 -000W00000000000001_fxJfo0B~t=FJEbHbY*gGVQepLZ)9a`b1z?CX>MtBUtcb8c~DCM0u%!j00008 -0LuVbTg%^TJ4XQk0Pz9<03iSX00000000000HlF5NCE(GX>c!JX>N37a&BR4FKusRWo&aVV_|M&X=Gt -^WiD`eP)h*<6ay3h000O8%K%whn~;%qlm`F+n;8HAA^-pY0000000000q=C9g0swGna4%nJZggdGZee -UMZEs{{Y;!MUX>w&_bYFFHY%XwlP)h*<6ay3h000O8%K%wh|0GbUQUd@0rwIT6CIA2c0000000000q= -Al60swGna4%nJZggdGZeeUMZEs{{Y;!MWZ*py6bYEj{Zgg^QY%XwlP)h*<6ay3h000O8%K%wh-MySaCuNm0Rj{Q6aWAK2ms3fSzFdSe=5`k002W60015U0000000 -000005+c$8`b#aA|NaUukZ1WpZv|Y%gtZWMyn~FK~HmZ)0mNaCuNm0Rj{Q6aWAK2ms3fSzF)amiI~l0 -02J%0015U0000000000005+c*Lwm0aA|NaUukZ1WpZv|Y%gtZWMyn~FLPsPWo>0HaCuNm0Rj{Q6aWAK -2ms3fSz8L>y3AMx005g2001Na0000000000005+cU3~%oaA|NaUukZ1WpZv|Y%gtZWMyn~FLPyKa${& -;b7OCCWiD`eP)h*<6ay3h000O8%K%wh=H8j>3j+WE;0FKzCjbBd0000000000q=EN?0swGna4%nJZgg -dGZeeUMZEs{{Y;!MjWo%_*bZKvHUvP3|W^*oZc~DCM0u%!j000080LuVbTPB)h_Z|fR0Fw{^03`qb00 -000000000HlFOhXMd_X>c!JX>N37a&BR4FKusRWo&aVbYXI5WprO~d30!RZZ2?nP)h*<6ay3h000O8% -K%whly^$V+5-RpXbS)U9RL6T0000000000q=Bf60swGna4%nJZggdGZeeUMZEs{{Y;!MnXk}$=E^v8J -O928D0~7!N00;of09jjO5aR(c0000n0000X00000000000001_fxVCd0B~t=FJEbHbY*gGVQepMWpsC -Ma%(SNUukY>bYEXCaCuNm0Rj{Q6aWAK2ms3fSzDdJmTb!v007BQ0012T0000000000005+cEs+8MaA| -NaUukZ1WpZv|Y%gwQba!uZYcFASbZ9Pcc~DCM0u%!j000080LuVbTY+Npkb(vP0R9pH03HAU0000000 -0000HlF4rvdc!JX>N37a&BR4FK%UYcW-iQFJob2Xk{*Nc~DCM0u%!j000080LuVbTX&CK@Iem%0 -E#gH03iSX00000000000HlHOtpWgWX>c!JX>N37a&BR4FK%UYcW-iQFJy0bZftL1WG--dP)h*<6ay3h -000O8%K%whQy)_w-3b5y-W>n{BLDyZ0000000000q=9k00swGna4%nJZggdGZeeUMZe?_LZ*prdY+-t -NUw3F_Wo#~Rc~DCM0u%!j000080LuVbTl=Ch;~N+N0Q5=#03ZMW00000000000HlGB#{vLwX>c!JX>N -37a&BR4FK%UYcW-iQFLPycb7^mGE^v8JO928D0~7!N00;of09jjVsPjnB1pom14*&oj000000000000 -01_f#}}?0B~t=FJEbHbY*gGVQepMWpsCMa%(SjbZKmJE^v8JO928D0~7!N00;of09jkpymbr70ssIe2 -LJ#c00000000000001_f%N490B~t=FJEbHbY*gGVQepMWpsCMa%(SmZESLIV=i!cP)h*<6ay3h000O8 -%K%wh000000ssI200000Bme*a0000000000q=EYA0swGna4%nJZggdGZeeUMZ*XODVRUJ4ZgVeRUukY ->bYEXCaCuNm0Rj{Q6aWAK2ms3fSzC(}invJ!000IT001EX0000000000005+cKIsAgaA|NaUukZ1WpZ -v|Y%gzcWpZJ3X>V?GFJowBV{0yOc~DCM0u%!j000080LuVbTjw_%Eolk>0DdL_03rYY00000000000H -lG&@B#pEX>c!JX>N37a&BR4FK=*Va$$67Z*FrhW^!d^dSxzfc~DCM0u%!j000080LuVbTa4r$p_Lr~0 -P=7E03!eZ00000000000HlF$`vL%PX>c!JX>N37a&BR4FK=*Va$$67Z*FrhaB^jEVRB_IaCuNm0Rj{Q -6aWAK2ms3fSz7=A0006200000001fg0000000000005+cMHvGCaA|NaUukZ1WpZv|Y%gzcWpZJ3X>V? -GFJg6RY-BHAUukY>bYEXCaCuNm0Rj{Q6aWAK2ms3fSzD`a?XZpp005&B001ul0000000000005+cj~N -31aA|NaUukZ1WpZv|Y%gzcWpZJ3X>V?GFJg6RY-BHDb!lv5UvzR|V{2t{E^v8JO928D0~7!N00;of09 -jks1zWLQ0ssJw1pojl00000000000001_fodQF0B~t=FJEbHbY*gGVQepNaAk5~bZKvHb1!0bX>4RKZ -Dn*}WMOn+E^v8JO928D0~7!N00;of09jj`2F-DM0ssKz1poju00000000000001_fe#}C0B~t=FJEbH -bY*gGVQepNaAk5~bZKvHb1!0bX>4RKZDn*}WMOn+Uu9%zbYWs_WiD`eP)h*<6ay3h000O8%K%wh@IZ? -F@c{q;a0CDVCjbBd0000000000q=DNd0|0Poa4%nJZggdGZeeUMZ*XODVRUJ4ZgVeUb!lv5FL!8VWo# -~Rc~DCM0u%!j000080LuVbTb_oUv_t{`0LBFX05AXm00000000000HlEtC<6d+X>c!JX>N37a&BR4FK -=*Va$$67Z*FrhVs&Y3WG{DUWo2w%Wn^h|VPb4$E^v8JO928D0~7!N00;of09jkj5U=_)0000o0000i0 -0000000000001_fukw|0B~t=FJEbHbY*gGVQepNaAk5~bZKvHb1!Lbb97;BY%gD5X>MtBUtcb8c~DCM -0u%!j000080LuVbTb8bJ0=ppq05oj?04V?f00000000000HlE-D+2&c!JX>N37a&BR4FK=*Va$$6 -7Z*FrhX>N0LVQg$KcW7m0Y%XwlP)h*<6ay3h000O8%K%wh&&ODO=mP)%;R^r&9RL6T0000000000q=6 -$%0|0Poa4%nJZggdGZeeUMa%FKZUtei%X>?y-E^v8JO928D0~7!N00;of09jjo(zp0=6aWB~NB{sK00 -000000000001_fksgS0B~t=FJEbHbY*gGVQepQWpOWKZ*FsRa&=>LZ*p@kaCuNm0Rj{Q6aWAK2ms3fS -z9jkM;K`X0024)001cf0000000000005+c_GJSAaA|NaUukZ1WpZv|Y%g+UaW8UZabIL*aAj^}Wo~16 -UuSY}b#N|lc~DCM0u%!j000080LuVbTSjL|2AvlG0HRI+0384T00000000000HlGYX#)UoX>c!JX>N3 -7a&BR4FLGsZFLGsZUuJ1+WiD`eP)h*<6ay3h000O8%K%whCl=OM&maH*DS7|^AOHXW0000000000q=9 -~c0|0Poa4%nJZggdGZeeUMa%FKZa%FK}X>N0LVQg$JaCuNm0Rj{Q6aWAK2ms3fSzDA^6;L<>000aN00 -0~S0000000000005+ci=qPnaA|NaUukZ1WpZv|Y%g+UaW8UZabI&~bS`jtP)h*<6ay3h000O8%K%wh* -mO`p@EQOBBVGUiA^-pY0000000000q=EgV0|0Poa4%nJZggdGZeeUMa%FKZa%FK}b#7^Hb97;BY%Xwl -P)h*<6ay3h000O8%K%wh000000ssI200000Bme*a0000000000q=7BL0|0Poa4%nJZggdGZeeUMa%FR -GY;|;LZ*DJNUukY>bYEXCaCuNm0Rj{Q6aWAK2ms3fSzCD#z6ksP002P(001BW0000000000005+calr -!saA|NaUukZ1WpZv|Y%g+Ub8l>QbZKvHFJfVHWiD`eP)h*<6ay3h000O8%K%wh000000ssI200000D* -ylh0000000000q=Bo#0|0Poa4%nJZggdGZeeUMa%FRGY;|;LZ*DJaWoKbyc`sjIX>MtBUtcb8c~DCM0 -u%!j000080LuVbTaCNOGYT320Q_A504o3h00000000000HlHT!UF(sX>c!JX>N37a&BR4FLGsbZ)|mR -X>V>XY-ML*V|g!fWpi(Ac4cxdaCuNm0Rj{Q6aWAK2ms3fSz7=A0006200000001ul0000000000005+ -cN!QbZKvHFLGsbZ)|pDY-wUIUtei%X>?y-E^v8JO928D0~7!N00; -of09jifWzroa1polI6951!00000000000001_ftcL`0B~t=FJEbHbY*gGVQepQWpi(Ab#!TOZZC3Wb8 -l>RWo&6;FJfVHWiD`eP)h*<6ay3h000O8%K%wh$?Y`PQxgCH=1%|sF#rGn0000000000q=5zF0|0Poa -4%nJZggdGZeeUMa%FRGY;|;LZ*DJgWpi(Ac4cg7VlQK1Ze(d>VRU74E^v8JO928D0~7!N00;of09jku -m9w=KApiheg8%?700000000000001_fv5HZ0B~t=FJEbHbY*gGVQepQWpi(Ab#!TOZZC3Wb8l>RWo&6 -;FJ@t5bZ>HbE^v8JO928D0~7!N00;of09ji;=(m661^@tp7XSb^00000000000001_fe07`0B~t=FJE -bHbY*gGVQepQWpi(Ab#!TOZZC3Wb8l>RWo&6;FJ^CbZe(9$VQyq;WMOn=b1rasP)h*<6ay3h000O8%K -%wh1+IQqp$-55u`B=pF8}}l0000000000q=7;o1ORYpa4%nJZggdGZeeUMa%FRGY;|;LZ*DJgWpi(Ac -4cg7VlQxVZ+2;9WpXZXc~DCM0u%!j000080LuVbTlH+=S;PbY07VY~051Rl00000000000HlF1E(8E@ -X>c!JX>N37a&BR4FLGsbZ)|mRX>V>Xa%FRGY<6XAX<{#OWpHnDbY*fbaCuNm0Rj{Q6aWAK2ms3fSzC9 -QbZKvHFLGsbZ)|pDY-w -UIa%FLKX>w(4Wo~qHE^v8JO928D0~7!N00;of09jkY!d`&N5C8yuGyniE00000000000001_f&MuJ0B -~t=FJEbHbY*gGVQepQWpi(Ab#!TOZZC3Wb8l>RWo&6;FLGsbZ)|pDaxQRrP)h*<6ay3h000O8%K%wh0 -00000ssI2000009{>OV0000000000q=6Mn1ORYpa4%nJZggdGZeeUMb#!TLb1z?CX>MtBUtcb8c~DCM -0u%!j000080LuVbTZCzO(bok402U1Z03!eZ00000000000HlFbO9TLLX>c!JX>N37a&BR4FLiWjY;!M -PYGHC=V{cz{Wq5QhaCuNm0Rj{Q6aWAK2ms3fSzAhYIX~6`008#`000{R0000000000005+cY*7RNaA| -NaUukZ1WpZv|Y%g_mX>4;ZUu<{c00000000000001_fk##Z0B~t=FJEbHbY*gGVQepTbZKmJFJo_QaA9;VaCuNm0Rj{Q6aWAK2ms3f -Sz8DkHFfm{008D0001cf0000000000005+cQCkE6aA|NaUukZ1WpZv|Y%g_mX>4;ZV{dJ6VRUI?X>4h -9d0%v4XLBxac~DCM0u%!j000080LuVbTlvktub2P;0Pz3-03QGV00000000000HlG9VgvwiX>c!JX>N -37a&BR4FLiWjY;!MUVRU75X>DaLaCuNm0Rj{Q6aWAK2ms3fSzEup%C}Ml003|f001HY000000000000 -5+cVq*jVaA|NaUukZ1WpZv|Y%g_mX>4;ZWMy!2Wn*DV>WaCuNm0Rj{Q6aWAK2ms3fSz9J3(lhD<0 -04jt001cf0000000000005+c^k@VCaA|NaUukZ1WpZv|Y%g_mX>4;ZWNC6`V{~72a%^8{Wo&R|a&sZ3F;tX>c!JX>N37a&BR4FLiW -jY;!MVXJ=n*X>MySaCuNm0Rj{Q6aWAK2ms3fSzET&Zf!^f008_9001HY0000000000005+c>Tm=AaA| -NaUukZ1WpZv|Y%g_mX>4;ZWo~qGd2nxOZgg`laCuNm0Rj{Q6aWAK2ms3fSzFKDU`|yC004Ov001EX00 -00000000005+ca&`m&aA|NaUukZ1WpZv|Y%g_mX>4;ZW@&6?b9r-gWo<5Sc~DCM0u%!j000080LuVbT -O)$gb4~#O0I>o903ZMW00000000000HlEie*^$c!JX>N37a&BR4FLiWjY;!MWX>4V5d2nTOE^v8J -O928D0~7!N00;of09jl2Qmf$z1^@t#4gdfg00000000000001_fs%j(0B~t=FJEbHbY*gGVQepTbZKm -JFK29NVq-3Fc~DCM0u%!j000080LuVbTTK&QBIO1E0EQC)0384T00000000000HlG`h6Dg`X>c!JX>N -37a&BR4FLiWjY;!MYVRL9@b1rasP)h*<6ay3h000O8%K%whPt9!i1`YrKOfLWc9smFU0000000000q= -EE}1ORYpa4%nJZggdGZeeUMb#!TLb1!UfXJ=_{XD)DgP)h*<6ay3h000O8%K%whU<2TMR2={SPFnx~8 -vpc!JX>N37a&BR4FLiWjY;!MgVPk7yXK8L{E^v8JO928 -D0~7!N00;of09ji`nsm^00ssKD1pojX00000000000001_fyulC0B~t=FJEbHbY*gGVQepTbZKmJFLG -sca(OOrc~DCM0u%!j000080LuVbTaiEhCU6S?02dc!JX>N37a& -BR4FLiWjY;!Mjbz*RGZ)0V1b1rasP)h*<6ay3h000O8%K%whQoPCGtqTAETOOV0000000000q -=6^P1ORYpa4%nJZggdGZeeUMb#!TLb1!sdZE#;?X>u-bc~DCM0u%!j000080LuVbTUDbTByJ7>0Io6s -03ZMW00000000000HlEt*8~7?X>c!JX>N37a&BR4FLiWjY;!MlZg62^YiVw0E^v8JO928D0~7!N00;o -f09jjbkyNwv0ssI(1^@sX00000000000001_fxYDf0B~t=FJEbHbY*gGVQepTbZKmJFLiQkb1rasP)h -*<6ay3h000O8%K%whchf_09R&aYe+>WtApigX0000000000q=D+_1ORYpa4%nJZggdGZeeUMb#!TLb1 -!yja&&cJY-MhCE^v8JO928D0~7!N00;of09jjR$D#y01^@tb5dZ)j00000000000001_fky5G0B~t=F -JEbHbY*gGVQepTbZKmJFL!8VWo#~Rc~DCM0u%!j000080LuVbTQ9BN#V-K>08|110384T0000000000 -0HlGu^aKEKX>c!JX>N37a&BR4FLq;dFJE72ZfSI1UoLQYP)h*<6ay3h000O8%K%whcdOf9%me@cmJR> -_8vp(1ORYpa4%nJZggdGZeeUMc4KodVqtn=VR9~Tc~DCM0u%!j000080LuVbTbu -u)p28LY0LV%J02lxO00000000000HlE{`vd@RX>c!JX>N37a&BR4FLq;dFK20VE^v8JO928D0~7!N00 -;of09jirK51Vp1^@uZ6#xJp00000000000001_fg=(H0B~t=FJEbHbY*gGVQepUV{V?0Z*FvQZ)`4bc~DCM0u%!j000080LuVbTQ5`lfFch70J0c!Jc4cm4Z*nhEML|SOMJ{r4bWlqH0u%!j000080LuVbTLO|0dT9m#04 -oy!02crN00000000000HlHUP6YsPX>c!Jc4cm4Z*nhRUukY>bYEXCaCuNm0Rj{Q6aWAK2ms3fSzCVpf -yUeb00373000&M0000000000005+cl~n}*aA|NaUv_0~WN&gWc4cm4Z*nelcyv%p0Rj{Q6aWAK2ms3f -SzE^3KwV`4001rm001Na0000000000005+crB($1aA|NaUv_0~WN&gWV_{=xWn*t{baHQOFHA{8MNU& -iE_8TwP)h*<6ay3h000O8%K%wh3wIW|Kmh;%r2+r|Bme*a0000000000q=8IW1pshqa4%nWWo~3|axY -_HV`yb#Z*FvQZ)`7LUukY>bYEXCaCuNm0Rj{Q6aWAK2ms3fSzGyUPffW3007Ab001BW000000000000 -5+c&RGQjaA|NaUv_0~WN&gWV_{=xWn*t{baHQOFJEJAWG--dP)h*<6ay3h000O8%K%whbpx+`wg&(J# -~Ac!Jc4cm4Z*nhVVPj}zV{dMBa&K% -eV_{=xWiD`eP)h*<6ay3h000O8%K%wh!`*UJdK3Tvok{=zCIA2c0000000000q=D9F1pshqa4%nWWo~ -3|axY_HV`yb#Z*FvQZ)`7PZ*FvQZ)|L3axQRrP)h*<6ay3h000O8%K%whEP=Pl?gjt=!w>)fCjbBd00 -00000000q=A%t1pshqa4%nWWo~3|axY_HV`yb#Z*FvQZ)`7SX>4V8a$#_AWpXZXc~DCM0u%!j000080 -LuVbTPyCQ&a?&q01*=a044wc00000000000HlG(g9QL^X>c!Jc4cm4Z*nhVVPj}zV{dMBa&K%eXk~SB -X>)XGV{c!Jc4cm -4Z*nhVVPj}zV{dMBa&K%eaCt6td2nT9P)h*<6ay3h000O8%K%wht?6LD?*;$>D--|#B>(^b00000000 -00q=5j71pshqa4%nWWo~3|axY_HV`yb#Z*FvQZ)`7fWpZg@Y-xIBE^v8JO928D0~7!N00;of09jkuJb -z{70RRAr1poje00000000000001_fijW>0B~t=FJE?LZe(wAFJob2Xk}w>Zgg^QY%h0mVQ_F|axQRrP -)h*<6ay3h000O8%K%whCD2a%vj6}9F984mD*ylh0000000000q=8tJ1pshqa4%nWWo~3|axY_HV`yb# -Z*FvQZ)`7PVPj}zb1z?CX>MtBUtcb8c~DCM0u%!j000080LuVbTfq8&kf{X#02L4b04)Fj000000000 -00HlFbl?4EBX>c!Jc4cm4Z*nhVVPj}zV{dMBa&K%eV_{=xWpgiPX>4U*V_{=xWiD`eP)h*<6ay3h000 -O8%K%why!H3DMgjl;Y6SoQE&u=k0000000000q=87A1pshqa4%nWWo~3|axY_HV`yb#Z*FvQZ)`7PVP -j}zb1!mbWNC9>V_{=xWiD`eP)h*<6ay3h000O8%K%wh6ND==J^}y$-2(ss8vpc!Jc4cm4Z*nhVWpZ?BW@#^9UukY>bYEXCaCuNm0Rj{Q6aWAK2ms3fSz9k4t>K; -k009300018V0000000000005+c-=GBmaA|NaUv_0~WN&gWV`Xx5X=Z6JUteuuX>MO%E^v8JO928D0~7 -!N00;of09jk)QUn6dYXSf!as&V#00000000000001_fx4ju0B~t=FJE?LZe(wAFJonLbZKU3FJob2Wp -Z>baAj>!O928D0~7!N00;of09jj|`~%mK0{{SW4FCWe00000000000001_fy5060B~t=FJE?LZe(wAF -JonLbZKU3FJo_VWiD`eP)h*<6ay3h000O8%K%wh000000ssI2000008~^|S0000000000q=AkR2mo+t -a4%nWWo~3|axY_La&&2CX)kbjE_8WtWn@rG0Rj{Q6aWAK2ms3fSzC!J>(+t-002t_001cf000000000 -0005+c$`J?vaA|NaUv_0~WN&gWWMyz=Ze(R{V|ia^a&L8Tb1zIuLq$$gMJ{xBbWlqH0u%!j000080Lu -VbTY4cwo_YWP0Qvv`04V?f00000000000HlGF6bJxtX>c!Jc4cm4Z*nhWWpHI~WMyt+d0%I8Z*_2UFJ -E72ZfSI1UoLQYP)h*<6ay3h000O8%K%whO(cT{paK8@tp)%9DF6Tf0000000000q=8fw2mo+ta4%nWW -o~3|axY|MaAj^}Wo~16UuSY}b#QYpUteuuX>MO%E^v8JO928D0~7!N00;of09jks5C0D02LJ$R9{>O_ -00000000000001_fjbxo0B~t=FJE?LZe(wAFJxtKWo~3;Zew|0XL4_KaC0wTX>D+9Wo>0{bYXO9Z*DG -dc~DCM0u%!j000080LuVbTQ8A0QPlzf0ImiA05|{u00000000000HlFyAP4|(X>c!Jc4cm4Z*nhWWpH -I~WMyt+d0%I8Z*_2UFJEkFZggK{WpHI~WMyt+d0%I8Z*_2UE^v8JO928D0~7!N00;of09jkr0ZIQP0{ -{R?2LJ#q00000000000001_ftDi(0B~t=FJE?LZe(wAFJxtKWo~3;Zew|0XL4_KaC0wTaA|N~cXDBHa -Ak5XaCuNm0Rj{Q6aWAK2ms3fSz8B4&nTz>0012U001oj0000000000005+c1t$mqaA|NaUv_0~WN&gW -WMyz=Ze(R{V|ia^a&L8Tb1z?XZ*6Q}V{dJ6VRSBVc~DCM0u%!j000080LuVbTL1t600IC200000044w -c00000000000HlHVCkOy=X>c!Jc4cm4Z*nhWWpHI~WMyt+d0%I8Z*_2UFK~G-ba`-PWKc^10u%!j000 -080LuVbTdfb&ovaW50K+%{03QGV00000000000HlFFCX>c!Jc4cm4Z*nhWX>)XJX<{!-Nkc_WQ$ -;Rxcyv%p0Rj{Q6aWAK2ms3fSzB?y-E^v8JO928D0~7!N00;of09jjIN#!pyD*yodp#T6K00000000000001_fyp`u -0B~t=FJE?LZe(wAFJx(RbZlv2FJo_QaA9;VaCuNm0Rj{Q6aWAK2ms3fSzDdG)qt=I002ZP001BW0000 -000000005+cH)aR`aA|NaUv_0~WN&gWWNCABY-wUIa%FRGb#h~6b1rasP)h*<6ay3h000O8%K%whA>6 -p<1{VMTv`7E|9smFU0000000000q=6-J2mo+ta4%nWWo~3|axY|Qb98KJVlQ)Ja%pgMb1rasP)h*<6a -y3h000O8%K%whgi3i5>bw8|0Db`g8vpc!Jc4cm4Z*nhW -X>)XJX<{#QHZ(0^a&0bUcx6ya0Rj{Q6aWAK2ms3fSzBCX0~wRd0001{0RS5S0000000000005+c6@>} -_aA|NaUv_0~WN&gWWNCABY-wUIbT%|DWq4&!O928D0~7!N00;of09jkAOLI8vpy3jlCwa4%nWWo~3|axY|Qb98KJVlQ_yGA?C!Wl&220u%!j000080LuVbT -P;Kj+Vw2~005E#03QGV00000000000HlH1P7DBWX>c!Jc4cm4Z*nhWX>)XJX<{#THZ(0^a&0bUcx6ya -0Rj{Q6aWAK2ms3fSzG?_qApIw0001v0RS5S0000000000005+c41Em%aA|NaUv_0~WN&gWWNCABY-wU -IcQ!OGWq4&!O928D0~7!N00;of09jiO<6DGi4*&o?EC2u+00000000000001_fs;ZG0B~t=FJE?LZe( -wAFJx(RbaHPmOi4pUPE$otO928D0~7!N00;of09jl6U~t?c0RRBi0{{RX00000000000001_fiqGL0B -~t=FJE?LZe(wAFJx(RbaHPmUtei%X>?y-E^v8JO928D0~7!N00;of09jidRJ>_C0000$0000U000000 -00000001_fs#`W0B~t=FJE?LZe(wAFJx(RbaHPmUteuuX>MO%E^v8JO928D0~7!N00;of09ji~I&Et% -EdT%(!2kdp00000000000001_feKU(0B~t=FJE?LZe(wAFJx(RbaHPmWNCABa&Inhc~DCM0u%!j0000 -80LuVbTL1t600IC20000002=@R00000000000HlF(ferw0X>c!Jc4cm4Z*nhWX>)XPZ!d6pE_8WtWn@ -rG0Rj{Q6aWAK2ms3fSzE*6oA4?F000FB000{R0000000000005+ct$_{zaA|NaUv_0~WN&gWX=H9;FH -A{8MNU&iE^TB`O928D0~7!N00;of09jiD7xB9v0RRAG0{{RV00000000000001_fe?ib0B~t=FJE?LZ -e(wAFKJ|MVJ}}_X>MtBUtcb8c~DCM0u%!j000080LuVbTOgZYPdWnt0A39M02u%P00000000000HlFw -h7JI5X>c!Jc4cm4Z*nhbWNu+EV{c?-V=i!cP)h*<6ay3h000O8%K%why<-UJ#{d8TJOKaz8UO$Q0000 -000000q=DLr4ghdza4%nWWo~3|axZCQZecHDZ*6d4bS`jtP)h*<6ay3h000O8%K%whzBs-g_6q<2w=) -0$7ytkO0000000000q=DIr4ghdza4%nWWo~3|axZCQZecHDZ*pZWaCuNm0Rj{Q6aWAK2ms3fSz89D06 -Yvj0081K0RS8T0000000000005+c1(yy0aA|NaUv_0~WN&gWX=H9;FKJ|MVPs)+VJ>iaP)h*<6ay3h0 -00O8%K%whI$=a+B?ABeY6k!S9RL6T0000000000q=8G(4ghdza4%nWWo~3|axZCQZecHJZgg^CZf9k4 -E^v8JO928D0~7!N00;of09jlAIQrrj0000L0000W00000000000001_fveOG0B~t=FJE?LZe(wAFKJ| -MVJ~oDV{2h&WnW}rbYU)Vc~DCM0u%!j000080LuVbTL1t600IC20000002u%P00000000000HlEc)eZ -n~X>c!Jc4cm4Z*nhbWNu+EaCt6td2nT9P)h*<6ay3h000O8%K%whWnju4#Hj!PxuydE9RL6T0000000 -000q=7lr4ghdza4%nWWo~3|axZCQZecHVbaON|WMOn+E^v8JO928D0~7!N00;of09jjDwJ17>0RRAI0 -ssIT00000000000001_fjEB;0B~t=FJE?LZe(wAFKu&YaA9L>FGEjISxHVuP)h*<6ay3h000O8%K%wh -DhN0*;sF2vRRjP49{>OV0000000000q=ET>4*+m!a4%nWWo~3|axZOjXK-O-YcF44X>MtBUtcb8c~DC -M0u%!j000080LuVbTRS_Q)2#sj067Ez03iSX00000000000HlE!gAV|3X>c!Jc4cm4Z*nhfb7yd2V{0 -#Ecw=R7bZKvHb1rasP)h*<6ay3h000O8%K%whCE!Q*c?JLgUKRiV8UO$Q0000000000q=5v54*+m!a4 -%nWWo~3|axZOjXK-O-YcFMZbS`jtP)h*<6ay3h000O8%K%whBpktIHX8r{hJFA59{>OV0000000000q -=B`I4*+m!a4%nWWo~3|axZOjXK-O-YcFPDY;0m-V{0yOc~DCM0u%!j000080LuVbTi@=|yng@y0L1_R -0384T00000000000HlE^rw;&dX>c!Jc4cm4Z*nhiVPk7yXK8L{FHA{8MNU&iP)h*<6ay3h000O8%K%w -hu(zWg=?wq?z9#?xBme*a0000000000q=Dk64*+m!a4%nWWo~3|axZXUV{2h&X>MmPOi4pUPE$oLK~O -VP|P>XD>`iLq$$gMJ_^9L{Lis0u%!j000080LuVbTYvn?TrmLv0PX<*03iSX00000000000HlEjx -eowvX>c!Jc4cm4Z*nhiVPk7yXK8L{FJE72ZfSI1UoLQYP)h*<6ay3h000O8%K%wh4)5auMmPUuA4&W@&6?E^v8JO928D0~7!N00;o -f09jj@VL-t*4FCXqB>(^;00000000000001_fttS$0B~t=FJE?LZe(wAFK}UFYhh<;Zf7rFZDDSCY-w -(FcrI{xP)h*<6ay3h000O8%K%wh>DV)-!2|#Rh6(@xBLDyZ0000000000q=6F64*+m!a4%nWWo~3|ax -ZXUV{2h&X>MmPUu|`BY;0+6b$Bjtc~DCM0u%!j000080LuVbTg-yz;4ujR0PQCL03ZMW00000000000 -HlEw(hmS|X>c!Jc4cm4Z*nhiVPk7yXK8L{FJEwBa&u*JE^v8JO928D0~7!N00;of09jiwqw-46h8X>c!Jc4cm4Z*nhiVPk7yXK8L{FJE -+TYh`X}dS!AhaCuNm0Rj{Q6aWAK2ms3fSzC&M1n>e6000p$001BW0000000000005+co8%7waA|NaUv -_0~WN&gWaA9L>VP|P>XD@AGa%*LBb1rasP)h*<6ay3h000O8%K%whp_MniS0w-dvxfixApigX000000 -0000q=DP>4*+m!a4%nWWo~3|axZXUV{2h&X>MmPZDn*}WMOn+E^v8JO928D0~7!N00;of09jiA00002 -000000000U00000000000001_fpHlS0B~t=FJE?LZe(wAFK}UFYhh<;Zf7rWc`kH$aAjmrO928D0~7! -N00;of09jk=jVik>1ONbo3jhEm00000000000001_fv*`50B~t=FJE?LZe(wAFK}UFYhh<;Zf7rYWpQ -~Uox0057i001KZ0000000000005+cA07|@aA|NaUv -_0~WN&gWaA9L>VP|P>XD@SbWn*b(X=QSAE^v8JO928D0~7!N00;of09jkTktBWy8UO&VP|P>XD@YhX>4;YaCuNm0Rj{Q6aWA -K2ms3fSzCV>$`Wf4001yS001BW0000000000005+c)?p9;aA|NaUv_0~WN&gWaA9L>VP|P>XD@bTa&u -{KZZ2?nP)h*<6ay3h000O8%K%wh9o!vi{RaR5UlsrWDgXcg0000000000q=A2P5CCv#a4%nWWo~3|ax -ZXUV{2h&X>MmPY-wXMmPY-wXc!Jc4cm4Z*nhiYiD0_Wpi(Ja${w4FHA{ -8MNU&iP)h*<6ay3h000O8%K%whc~f!DR)PQkBjy1BB>(^b0000000000q=9Um5CCv#a4%nWWo~3|axZ -XeXJ2w?y-E^v8JO928D0~7!N00;of09jkC)*D@e0ssI(1ONaa000000000000 -01_fdwED0B~t=FJE?LZe(wAFK}#ObY^dIZDeV3b1zIuLq$$gMNmrt0u%!j000080LuVbTUKc!Jc4cm4Z*nhiY+-a}Z*py9X>xNfUtei%X>?y-E^v8JO928D -0~7!N00;of09jj;(H^6_0RRBu1pojf00000000000001_fvPDH0B~t=FJE?LZe(wAFK}#ObY^dIZDeV -3b1z?CZDDC{Utcb8c~DCM0u%!j000080LuVbTbgeR6F3I|05u~303!eZ00000000000HlGZD-i&2X>c -!Jc4cm4Z*nhiY+-a}Z*py9X>xNfVQyq{Z)s#MaCuNm0Rj{Q6aWAK2ms3fSzBUzjPzy)002QG0018V00 -00000000005+cATtpFaA|NaUv_0~WN&gWaBN|8W^ZzBWNC79FJW+LE^v8JO928D0~7!N00;of09jkHl -^i0`1ONcC7ytkv00000000000001_fx|ix0B~t=FJE?LZe(wAFK}#ObY^dIZDeV3b1!XSV{daVaCuNm -0Rj{Q6aWAK2ms3fSz7=A0006200000001EX0000000000005+c(?1aaaA|NaUv_0~WN&gWaBN|8W^Zz -BWNC79FK~G-ba`-PWKc^10u%!j000080LuVbTSm-MMR^GT0NN-303ZMW00000000000HlEwKoJ0NX>c -!Jc4cm4Z*nhiY+-a}Z*py9X>xNfb#7^RE^v8JO928D0~7!N00;of09jiL?Cz940RRBN0ssIb0000000 -0000001_fy+q|0B~t=FJE?LZe(wAFK}#ObY^dIZDeV3b1!yfa&u{KZZ2?nP)h*<6ay3h000O8%K%wht -(cJU%?SVijVAyABLDyZ0000000000q=8CH5dd&$a4%nWWo~3|axZXfVRUA1a&2U3a&s?tX>Me1cXKXq -c~DCM0u%!j000080LuVbTMGam%fSKw05b&u02}}S00000000000HlFfRS^JiX>c!Jc4cm4Z*nhid1q~ -9Zgg`mOi4pUPE$otO928D0~7!N00;of09jj&l9e3u1ONc13jhEh00000000000001_fm2u!0B~t=FJE -?LZe(wAFK~HhZDnqBb1z?CX>MtBUtcb8c~DCM0u%!j000080LuVbTW+1i5b6K`0AT?D03ZMW0000000 -0000HlG0T@e6qX>c!Jc4cm4Z*nhid1q~9Zgg`mUteuuX>MO%E^v8JO928D0~7!N00;of09jkdV|*r+0 -ssKE1^@sb00000000000001_fvjE;0B~t=FJE?LZe(wAFK~HhZDnqBb1!3WZgX#JWiD`eP)h*<6ay3h -000O8%K%whcJWzR`~m;~b_W0e9smFU0000000000q=9~65dd&$a4%nWWo~3|axZXsXKiI}baO9eX>4? -5axQRrP)h*<6ay3h000O8%K%whWi@i;;|2f#CJ_JtApigX0000000000q=B|%5dd&$a4%nWWo~3|axZ -XsXKiI}baO9eZ*py6baZ8ME^v8JO928D0~7!N00;of09jjC3Wi*XB>(^wiU0r|00000000000001_f! -J#i0B~t=FJE?LZe(wAFK~HhZDnqBb1!UVcx7@faCuNm0Rj{Q6aWAK2ms3fSzG0}oo(;|008X+001BW0 -000000000005+coRJX#aA|NaUv_0~WN&gWaCv8KWo~qHFKusRWo&6~WiD`eP)h*<6ay3h000O8%K%wh -e-4dud;$OfV+Q~L9smFU0000000000q=Cwl5dd&$a4%nWWo~3|axZXsXKiI}baO9oY;|X8ZZ2?nP)h* -<6ay3h000O8%K%whjivwtO#}b{01N;CAOHXW0000000000q=AB#5dd&$a4%nWWo~3|axZXsXKiI}baO -9qWoKo0Z*X)jaCuNm0Rj{Q6aWAK2ms3fSzC+35Vn8>000yW0018V0000000000005+c4Vw`FaA|NaUv -_0~WN&gWaCv8KWo~qHFLPsIZf<3AE^v8JO928D0~7!N00;of09jkkOAWKl2mk;r9{>O$00000000000 -001_fy$o|0B~t=FJE?LZe(wAFK~HhZDnqBb1!pnXlZVEWq5QhaCuNm0Rj{Q6aWAK2ms3fSzGOAcZ2!{ -000yj0012T0000000000005+c)~FExaA|NaUv_0~WN&gWaCv8KWo~qHFLQKxY-KKRc~DCM0u%!j0000 -80LuVbTU8^;g@FeE08$tL0384T00000000000HlEnun_=oX>c!Jc4cm4Z*nhid1q~9Zgg`mbZ={AZZ2 -?nP)h*<6ay3h000O8%K%wh4;A&2ZVc~DCM0u%!j000080LuVbTQuD)n71DQ0LYyH04M+e00000 -000000HlF$5)uG#X>c!Jc4cm4Z*nhid1q~9Zgg`mW@&76WpZ;bUtei%X>?y-E^v8JO928D0~7!N00;o -f09ji#+CPxa1^@sF6#xJ$00000000000001_fow7o0B~t=FJE?LZe(wAFK~HhZDnqBb1!CZa&2LBbY* -gLFJE72ZfSI1UoLQYP)h*<6ay3h000O8%K%whj*9)vAO!#bP!IqBD*ylh0000000000q=ADu5&&>%a4 -%nWWo~3|axZXsXKiI}baO9eZ*py6baZ8Mb1z?QVQ_G1Zf7oVc~DCM0u%!j000080LuVbTL6MwEk_Li0 -8=jj04D$d00000000000HlHJJ`wc!Jc4cm4Z*nhid1q~9Zgg`mY-M<5a&s?VUukY>bYEXCaCuNm -0Rj{Q6aWAK2ms3fSzCw9=Kmr;001H{0RSfe0000000000005+ccuW!iaA|NaUv_0~WN&gWaCv8KWo~q -HFKlIaWpZ;bUu|J?y-E^v8JO -928D0~7!N00;of09jjm{_Gta0{{T<3;+Np00000000000001_fjzqt0B~t=FJE?LZe(wAFK~HhZDnqB -b1!prd2D5KFJEn8aByjEXD)DgP)h*<6ay3h000O8%K%whM~lQneF6XgIRpRzBLDyZ0000000000q=B8 -j5&&>%a4%nWWo~3|axZXsaB^>IWn*+-Xm4+8b1zIuLq$$gMNmrt0u%!j000080LuVbTbLREok0Nr0J8 -!B04D$d00000000000HlFi!4d#)X>c!Jc4cm4Z*nhid2n)XYGq?|UubV{YjZDOUukY>bYEXCaCuNm0R -j{Q6aWAK2ms3fSz8}vF|INU003Az001Na0000000000005+c;ldICaA|NaUv_0~WN&gWaCvZYZ)#;@b -YEz1Z)5&&>% -a4%nWWo~3|axZXsaB^>IWn*+-Xm4+8b1!gtE_8WtWn@rG0Rj{Q6aWAK2ms3fSzEGZzdsiN001ol001) -p0000000000005+cna>gcaA|NaUv_0~WN&gWaCvZYZ)#;@bYEz1Z)BWpi^cUtei%X>? -y-E^v8JO928D0~7!N00;of09jiwC3Alr4FCYRF8}~G00000000000001_fdSDH0B~t=FJE?LZe(wAFK -~Hqa&Ky7V{~6=Z*OaJFJEbHUvP47V`X!5FJEbHUvP47V`X!5E^v8JO928D0~7!N00;of09jj-=AX;p4 -FCY1CjbB(00000000000001_fpXpw0B~t=FJE?LZe(wAFLGsZb!BsOb1zIuLq$$gMNmrt0u%!j00008 -0LuVbTM8r8<7Wo|0KpRg03ZMW00000000000HlG8>=FQQX>c!Jc4cm4Z*nhkWpQ<7b98erUtei%X>?y --E^v8JO928D0~7!N00;of09jjpxozqM0RRBA0RR9a00000000000001_fid(F0B~t=FJE?LZe(wAFLG -sZb!BsOb1z?Cc4cyNX>V>{UoLQYP)h*<6ay3h000O8%K%wh$SBO|hXMcq*98CoCjbBd0000000000q= -9zz5&&>%a4%nWWo~3|axZdaadl;LbaO9XX>N37a&BR4Uv+e8Y;!Jfc~DCM0u%!j000080LuVbTLRuU^ -j8@G06k{_03ZMW00000000000HlFJ_!0ncX>c!Jc4cm4Z*nhkWpQ<7b98erVPs)&bY*gLE^v8JO928D -0~7!N00;of09jje{x$+p1^@st82|tq00000000000001_f!7fe0B~t=FJE?LZe(wAFLGsZb!BsOb1z| -VX)bViP)h*<6ay3h000O8%K%whvfhoGdJ6ym$|nE-8~^|S0000000000q=8`=698~&a4%nWWo~3|axZ -daadl;LbaO9Zb#!PhaCuNm0Rj{Q6aWAK2ms3fSzCvi9~~P3006lG0012T0000000000005+c6(kb?aA -|NaUv_0~WN&gWa%FLKWpi|MFJonLbaO6nc~DCM0u%!j000080LuVbTO!V%e>?*K03HVb03HAU000000 -00000HlFyB@+N}X>c!Jc4cm4Z*nhkWpQ<7b98erV{dJ6VRSBVc~DCM0u%!j000080LuVbTh`v<*u)b6 -0G>zy03QGV00000000000HlHAC=&p1X>c!Jc4cm4Z*nhkWpQ<7b98erV{dP3X=QURaCuNm0Rj{Q6aWA -K2ms3fSzEw8U^;;X006KM001HY0000000000005+cE7}CjbEViU0r}00000 -000000001_fd@$w0B~t=FJE?LZe(wAFLGsZb!BsOb1!XgWMyn~E^v8JO928D0~7!N00;of09jj@A65m -q0RR9Y1ONaa00000000000001_fmCo40B~t=FJE?LZe(wAFLGsZb!BsOb1!gVV{2h&WpgfYc~DCM0u% -!j000080LuVbTf8{G|9Bw)04H|<03ZMW00000000000HlFSauWb>X>c!Jc4cm4Z*nhkWpQ<7b98erb7 -gaLX>V?GE^v8JO928D0~7!N00;of09jkCNtR`s1^@u!5C8xq00000000000001_fdP~g0B~t=FJE?LZ -e(wAFLGsZb!BsOb1!prVRUtKUt@1%WpgfYc~DCM0u%!j000080LuVbTTv2BxmyGP0ALFM03rYY00000 -000000HlH5niBwUX>c!Jc4cm4Z*nhkWpQ<7b98erb98cbV{~0384T00000000000HlF=pA!IZX>c!Jc4cm4Z*nhkWpQ<7b98erb#!TLb1rasP)h*<6ay -3h000O8%K%wh!Ktmw+W`Oo?*ae-9smFU0000000000q=C4|698~&a4%nWWo~3|axZdab8l>RWo&6;FH -A{8MNU&iP)h*<6ay3h000O8%K%whHq^9X)Bpeg9RdIVA^-pY0000000000q=C=M698~&a4%nWWo~3|a -xZdab8l>RWo&6;FJE72ZfSI1UoLQYP)h*<6ay3h000O8%K%whd5_9tRWo&6;FK}{ic4=f~a&sc!Jc4cm4Z*nhkWpi(Ac4cg7VlQxcE_8WtWn@rG0Rj{Q6aW -AK2ms3fSz8@my%N&`008v|001KZ0000000000005+cN7fSnaA|NaUv_0~WN&gWa%FRGY<6XAX<{#OWp -HnDbY*gLE^v8JO928D0~7!N00;of09jk`Mfc%Z2LJ#R82|tw00000000000001_fm_)V0B~t=FJE?LZ -e(wAFLGsbZ)|pDY-wUIb98cbV{~&aaCuNm0Rj{Q6aWAK2ms3fSzEDC;(PG`004jj001li0000000000 -005+c^WYNzaA|NaUv_0~WN&gWa%FRGY<6XAX<{#OWpi(Ac4cyNFJE72ZfSI1UoLQYP)h*<6ay3h000O -8%K%whXZ0Csr~&{02L=EDEC2ui0000000000q=7Et698~&a4%nWWo~3|axZdab8l>RWo&6;FLGsbZ)| -pDa&s?XVsmtIVPkYIaCuNm0Rj{Q6aWAK2ms3fSzEFR#o@yO008I)001oj0000000000005+cAmtMPaA -|NaUv_0~WN&gWa%FRGY<6XAX<{#OWpi(Ac4cyNFJp3PbY*gBZ*DGdc~DCM0u%!j000080LuVbTkdJj@ -J<2%0O$q)04@Lk00000000000HlE~=o0{NX>c!Jc4cm4Z*nhkWpi(Ac4cg7VlQ%Kb8l>RWpZ;bWq4y{ -aCB*JZgVbhc~DCM0u%!j000080LuVbTQ8?^u;Uj10F+(;04@Lk00000000000HlG)>JtEPX>c!Jc4cm -4Z*nhkWpi(Ac4cg7VlQ%Kb8l>RWpZ;ba%FRGY;|;LZ*DGdc~DCM0u%!j000080LuVbTmP2wN^1fD03Z -Ya02lxO00000000000HlHU0u%snX>c!Jc4cm4Z*nhkX=7+FOi4pUPE$otO928D0~7!N00;of09jjMgQ ->*;2mk=e7XSbp00000000000001_ftUpp0B~t=FJE?LZe(wAFLG&PXfI!1X>MtBUtcb8c~DCM0u%!j0 -00080LuVbTheV~Q;7=z0N5S?02}}S00000000000HlG^4HN)yX>c!Jc4cm4Z*nhkX=7+FUteuuX>MO% -E^v8JO928D0~7!N00;of09jjww)mB53IG7%CjbB-00000000000001_ft48)0B~t=FJE?LZe(wAFLG& -PXfI!5Wo&F;cWGpFXmc)bc~DCM0u%!j000080LuVbTYoO@fGlwU0Lvo+03ZMW00000000000HlFEBoq -K}X>c!Jc4cm4Z*nhkX=7+FUuA7?YH43%Z)9b2E^v8JO928D0~7!N00;of09jk(dGEQ_0RR9f1ONac00 -000000000001_fv1%e0B~t=FJE?LZe(wAFLG&PXfI!7ZEtF6Uvgz|Y+++%E^v8JO928D0~7!N00;of0 -9jieTGfm<0{{R}2mk;g00000000000001_fxVX$0B~t=FJE?LZe(wAFLG&PXfI!7cyMoWbYEs~a&2LB -E^v8JO928D0~7!N00;of09ji})VJH4000090RR9V00000000000001_fi#;G0B~t=FJE?LZe(wAFLG& -PXfI!7cywiMb7^mGE^v8JO928D0~7!N00;of09jk@!NV4Z0RR9W0{{RU00000000000001_feV}z0B~ -t=FJE?LZe(wAFLG&PXfI!8X>4U~Z!U0oP)h*<6ay3h000O8%K%wh-=?)RT?+sJxFrAp8~^|S0000000 -000q=Czw6aa8(a4%nWWo~3|axZdeV`wj5X>N0HWn*+MaCuNm0Rj{Q6aWAK2ms3fSz8vX<{P&I005Z`0 -018V0000000000005+cVyP4WaA|NaUv_0~WN&gWa%p2|FJEkLXJ2w)0B~t=FJE?LZe(wAFLG&PXfI!EZ*OoeaCuNm0Rj{ -Q6aWAK2ms3fSzEm*)3c)i0043Y0015U0000000000005+cvaS>WaA|NaUv_0~WN&gWa%p2|FJEqTY;0 -d4UKaCuNm0Rj{Q6aWAK2ms3fSz8yOS1pAE005U80012T0000000000005+ck+2j1aA|NaUv_0~WN -&gWa%p2|FJEwBY-MzGWpgfYc~DCM0u%!j000080LuVbTj}ubOyK|k0H*-}02u%P00000000000HlFYw -G;qwX>c!Jc4cm4Z*nhkX=7+FUvOz-Yc6nkP)h*<6ay3h000O8%K%wh(ips8j|Kn$%@hCt8UO$Q00000 -00000q=9L+6aa8(a4%nWWo~3|axZdeV`wj5a$$67Z!U0oP)h*<6ay3h000O8%K%wh#xbo_Uk(5O&rJX -T9RL6T0000000000q=7KK6aa8(a4%nWWo~3|axZdeV`wj5b8u;HZe?Db4E -^v8JO928D0~7!N00;of09jiOAN76G6953?SO5Sb00000000000001_f&0xA0B~t=FJE?LZe(wAFLG&P -XfI!PX>KzzUt@1>b8l>AE^v8JO928D0~7!N00;of09jj17JZz&0ssJo2LJ#Z00000000000001_feqp -m0B~t=FJE?LZe(wAFLG&PXfI!PX>Me1cXKXqc~DCM0u%!j000080LuVbTcif2J>3ET0N)A#03`qb000 -00000000HlEfc!Jc4cm4Z*nhkX=7+FUw3J4WN&wKUvgz`WMy(?axQRrP)h*<6ay3h000O8%K -%wh=jik%Lj?c;Obq}482|tP0000000000q=6&o6aa8(a4%nWWo~3|axZdeV`wj5cXDBHE^v8JO928D0 -~7!N00;of09jiOYei9w0RRAc0{{RQ00000000000001_ft~FX0B~t=FJE?LZe(wAFLG&PXfI)6V=i!c -P)h*<6ay3h000O8%K%whO8dp)fCvBpR44!d82|tP0000000000q=8@W6aa8(a4%nWWo~3|axZdeV`wj -7Y-wk1E^v8JO928D0~7!N00;of09jidF6kGu2LJ#G8vp -c!Jc4cm4Z*nhkX=7+FVqtPFaCuNm0Rj{Q6aWAK2ms3fSzGR`YG7On006!z000*N0000000000005+cj -|3F}aA|NaUv_0~WN&gWa%p2|FJf -P0000000000005+cfff}2aA|NaUv_0~WN&gWa%p2|FJo_PZ*nehc~DCM0u%!j000080LuVbTWLfMPLc -rt03HMY03iSX00000000000HlF`DHQ;4X>c!Jc4cm4Z*nhkX=7+FV{dG4a$j_EX>e?1bS`jtP)h*<6a -y3h000O8%K%who=j5llm`F++ZzA?8vpc!Jc4cm4Z*nhk -X=7+FV{dMAZ){~QaCuNm0Rj{Q6aWAK2ms3fSzC)t(n;O{000OD0012T0000000000005+czls$AaA|N -aUv_0~WN&gWa%p2|FJo_Rb98cHX>KlXc~DCM0u%!j000080LuVbTMW`;cf|z&0Dct!03HAU00000000 -000HlG~j1>TIX>c!Jc4cm4Z*nhkX=7+FV{dMBVQFqc!Jc4cm4Z*nhkX=7+FV{dMBa&K%daCuNm0Rj{Q6aWAK2ms3fS -zAOlSyLwm002QC001HY0000000000005+c*P0aoaA|NaUv_0~WN&gWa%p2|FJxtAVRdYDUvqSMY-MvU -aCuNm0Rj{Q6aWAK2ms3fSzAALI+oV~00035000~S0000000000005+cKB5%>aA|NaUv_0~WN&gWa%p2 -|FJx(9XKrtEWiD`eP)h*<6ay3h000O8%K%wh;RV|O%>w`cKM4Q;82|tP0000000000q=8SQ6##H)a4% -nWWo~3|axZdeV`wjBZEtF6E^v8JO928D0~7!N00;of09jiiNFmeh0001j0ssIS00000000000001_fm -5gz0B~t=FJE?LZe(wAFLG&PXfI`Qa&K~TE^v8JO928D0~7!N00;of09jjV-ft9*0ssJ$1^@sa000000 -00000001_fq1DE0B~t=FJE?LZe(wAFLG&PXfI}IY-L|?a&LHfE^v8JO928D0~7!N00;of09jk6yFJ%A -1ONcE2><{b00000000000001_fkCVl0B~t=FJE?LZe(wAFLG&PXfI}IY-Mw4dSxzfc~DCM0u%!j0000 -80LuVbThqIXO+yO+0CFV&03QGV00000000000HlGmuoVDsX>c!Jc4cm4Z*nhkX=7+FXlZ9?Y-wj`bY* -fbaCuNm0Rj{Q6aWAK2ms3fSzF*uyoM?T005^G000;O0000000000005+cG`tl6aA|NaUv_0~WN&gWa% -p2|FKTmdZZ2?nP)h*<6ay3h000O8%K%whIKf!Riv$1wvc!Jc4cm4Z*nhkX=7+FY+-qCb#yLpc~DCM0u%!j000080LuVbTiGDR&Z!Ro08Bdo02lxO0000 -0000000HlF_)D-}5X>c!Jc4cm4Z*nhkX=7+FY-x67E^v8JO928D0~7!N00;of09jiBjB{=30{{TQ4FC -Wi00000000000001_fnDSk0B~t=FJE?LZe(wAFLG&PXfJGOc4c33Wo~3;axQRrP)h*<6ay3h000O8%K -%whrhqvvY7PJZv@iew8vp00SWa02%-Q00000000000HlE@_Z0weX>c!Jc4cm4Z*nhkX=7+FZDDe2 -b#N|lc~DCM0u%!j000080LuVbTmP&W1Z)KW0J#(Z02=@R00000000000HlGg0TuvoX>c!Jc4cm4Z*nh -kX=7+FZDnC|b#i4caCuNm0Rj{Q6aWAK2ms3fSzG^SPo&lb001l#000{R0000000000005+cQwJ6RaA| -NaUv_0~WN&gWa%p2|FK}UGWNB_^E^v8JO928D0~7!N00;of09jiO#ub8d0RR9z0{{RS000000000000 -01_fnp670B~t=FJE?LZe(wAFLG&PXfJSKXJv9OaCuNm0Rj{Q6aWAK2ms3fSz9;&K3}^8002Y{000{R0 -000000000005+c3=b9naA|NaUv_0~WN&gWa%p2|FK}UOWps39E^v8JO928D0~7!N00;of09jkhSfMvZ -3IG6=D*yl)00000000000001_fdCX10B~t=FJE?LZe(wAFLG&PXfJSKZe?sPaCuNm0Rj{Q6aWAK2ms3 -fSz9pk4CHMe001YB000^Q0000000000005+cejXM8aA|NaUv_0~WN&gWa%p2|FK}{YbaZ(xaCuNm0Rj -{Q6aWAK2ms3fSzAbM;c!Jc4cm4Z*nhkX=7+ -Fa&>HFE^v8JO928D0~7!N00;of09jiPh4m|R1ONaV3jhEa00000000000001_f$oqN0B~t=FJE?LZe( -wAFLG&PXfJbPZ*XNUaCuNm0Rj{Q6aWAK2ms3fSzA9f{;pC2001`z000^Q0000000000005+cnUxj*aA -|NaUv_0~WN&gWa%p2|FLPsZWo2$IaCuNm0Rj{Q6aWAK2ms3fSzA^q8>EyJ005_8000{R00000000000 -05+cBbXKdaA|NaUv_0~WN&gWa%p2|FLPyQZDnqBE^v8JO928D0~7!N00;of09jiX&zvbH1pokc5C8xh -00000000000001_f$*#r0B~t=FJE?LZe(wAFLG&PXfJbcX>M+1axQRrP)h*<6ay3h000O8%K%wh0$S! -e_yhm|ND%-48UO$Q0000000000q=8Se765Q*a4%nWWo~3|axZdeV`wjPbYXOLb1rasP)h*<6ay3h000 -O8%K%whp~5SCy%qofZfO7j82|tP0000000000q=A38765Q*a4%nWWo~3|axZdeV`wjPba`xLE^v8JO9 -28D0~7!N00;of09jk{3Z_ui0RRB%1ONaU00000000000001_fpg9l0B~t=FJE?LZe(wAFLG&PXfJbgd -2D55E^v8JO928D0~7!N00;of09jk&MA_5+AOHXpjsO4}00000000000001_fq>8!0B~t=FJE?LZe(wA -FLG&PXfJblZggRIE^v8JO928D0~7!N00;of09ji*|GX8o9{>PxoB#kB00000000000001_fwu7$0B~t -=FJE?LZe(wAFLG&PXfJeOVr*qDaCuNm0Rj{Q6aWAK2ms3fSz9_I)4%=$001ft001HY0000000000005 -+cq7fGWaA|NaUv_0~WN&gWa%p2|FLY&cZE0>{Y+rO}Wo>0HaCuNm0Rj{Q6aWAK2ms3fSzA+?Q=FeD00 -6+b000;O0000000000005+c;1w4DaA|NaUv_0~WN&gWa%p2|FLY&ibS`jtP)h*<6ay3h000O8%K%whq -u(SR$^-xayAA*V82|tP0000000000q=B_Q7XWZ+a4%nWWo~3|axZdeV`wjQXk~3>E^v8JO928D0~7!N -00;of09jj6b@ot40001H0000Q00000000000001_fwn^z0B~t=FJE?LZe(wAFLG&PXfJeVWo>11E^v8 -JO928D0~7!N00;of09jirWX9x|9smFpi~s-~00000000000001_fi*-I0B~t=FJE?LZe(wAFLG&PXfJ -efVPj=tVPk79aCuNm0Rj{Q6aWAK2ms3fSzDI}FBhZ=008SG000;O0000000000005+c2V)liaA|NaUv -_0~WN&gWa%p2|FLZKcWiD`eP)h*<6ay3h000O8%K%whU08L3cLD$aFa!Vq82|tP0000000000q=DgW7 -XWZ+a4%nWWo~3|axZjmZER^TOi4pUPE$otO928D0~7!N00;of09jklI#2NI0000w0RR9U0000000000 -0001_fsb$(0B~t=FJE?LZe(wAFLZBhY-ulFUukY>bYEXCaCuNm0Rj{Q6aWAK2ms3fSzBL?NIqm5006R -N000~S0000000000005+cws98#aA|NaUv_0~WN&gWbZ>2JX)j-JVRCb2axQRrP)h*<6ay3h000O8%K% -whd3V*}4+Q`KqYMB57ytkO0000000000q=8h77XWZ+a4%nWWo~3|axZjmZER^TUvgzGaCuNm0Rj{Q6a -WAK2ms3fSzEL35@*E#008~~000{R0000000000005+cnUNO&aA|NaUv_0~WN&gWbZ>2JX)j-Nd2nTOE -^v8JO928D0~7!N00;of09jirDFFB!0000Q0000Q00000000000001_ftQjO0B~t=FJE?LZe(wAFLZBh -Y-ulWc`kH$aAjmrO928D0~7!N00;of09jjISapJT0ssIo1ONaV00000000000001_f$NeN0B~t=FJE? -LZe(wAFLZBhY-wM2FHA{8MNU&iP)h*<6ay3h000O8%K%whYdEsvjsO4vsQ>@~9{>OV0000000000q=B -227XWZ+a4%nWWo~3|axZjmZER^@cQ0RGX>MtBUtcb8c~DCM0u%!j000080LuVbTbooFzB35`05KZ?03 -HAU00000000000HlFtmKOkUX>c!Jc4cm4Z*nhmZ*6R8Uw1EGcXDZTWpXZXc~DCM0u%!j000080LuVbT -PrC5_#6NL02%-Q02}}S00000000000HlG{pBDgdX>c!Jc4cm4Z*nhmZ*6R8Uw1EXc`kH$aAjmrO928D -0~7!N00;of09jkb4+^k-0ssI$1ONaY00000000000001_fh(XF0B~t=FJE?LZe(wAFLZKsb98fbZ*pZ -XOi4pUPE$otO928D0~7!N00;of09jiPHFKz00ssIg1pojd00000000000001_f#jkW0B~t=FJE?LZe( -wAFLZKsb98fbZ*pZXUtei%X>?y-E^v8JO928D0~7!N00;of09jioVaNzW3jhF>EC2u=000000000000 -01_fqkVH0B~t=FJE?LZe(wAFLZKsb98fbZ*pZXUtw@*E^v8JO928D0~7!N00;of09jk!H@8li6951gP -yhfR00000000000001_f&8!+0B~t=FJE?LZe(wAFLZKsb98fbZ*pZXUu|JyZ*wkic~DCM0u%!j00008 -0LuVbTb)jZkQf920Bi{W03rYY00000000000HlG_#1{Z?X>c!Jc4cm4Z*nhma&>cbb98TVWiMZEaAj_ -Db8Iefc~DCM0u%!j000080LuVbTWP>t&&2@%0BQsP04M+e00000000000HlE{$rk`{X>c!Jc4cm4Z*n -hma&>cbb98TVWiMZIb8KH@Z*FsRVQzGDE^v8JO928D0~7!N00;of09jj6{tboz5C8y4MgRaJ0000000 -0000001_fi}w*0B~t=FJE?LZe(wAFLZKsb98fbZ*pZXUw3J4WN&wKE^v8JO928D0~7!N00;of09jiA0 -0002000000000V00000000000001_fp*&$0B~t=FJE?LZe(wAFLZKsb98fbZ*pZXaCt6td2nT9P)h*< -6ay3h000O8%K%whni!`vlmY+%TLb_A9{>OV0000000000q=B{D7XWZ+a4%nWWo~3|axZmqY;0*_GcQa -@Lq$$gMJ{xBbWlqH0u%!j000080LuVbTTg=^C~5@&00j*I03QGV00000000000HlG2-WLFHX>c!Jc4c -m4Z*nhna%^mAVlyvaUukY>bYEXCaCuNm0Rj{Q6aWAK2ms3fSzE!vKPI#d003+(001KZ000000000000 -5+cE#wyfaA|NaUv_0~WN&gWb#iQMX<{=kUt@1V?GE^v8JO928D0~7!N00;of09jkux;?u<0 -000$0000V00000000000001_fgV?DZ*OcaaCuNm0Rj{Q6aWAK2ms3fSzBJrNww<<000>v001EX0000000 -000005+c2sRi1aA|NaUv_0~WN&gWb#iQMX<{=kWq4y{aCB*JZgVbhc~DCM0u%!j000080LuVbTgNC%E -FcO10D~a_0384T00000000000HlF3Ko|gUX>c!Jc4cm4Z*nhna%^mAVlyvhX=Q9=b1rasP)h*<6ay3h -000O8%K%whhpWG%eggmihzS4y9{>OV0000000000q=Ah}7yxi-a4%nWWo~3|axZmqY;0*_GcRUoY-Mn -7b963nc~DCM0u%!j000080LuVbTRUdkn4cB^02WRF03rYY00000000000HlFPP8a}iX>c!Jc4cm4Z*n -hna%^mAVlyvrZ*OdEVQyh(WpXZXc~DCM0u%!j000080LuVbTL4*}jI#&;03#Xz03HAU00000000000H -lE@Wf%Z(X>c!Jc4cm4Z*nhna%^mAVlyvtWpQ<7b963nc~DCM0u%!j000080LuVbTTmdlI%6OJ0I_!f0 -3QGV00000000000HlEyZWsV?X>c!Jc4cm4Z*nhna%^mAVlyvtWpi+EZgXWWaCuNm0Rj{Q6aWAK2ms3f -Sz7=A0006200000001Wd0000000000005+cwvHG8aA|NaUv_0~WN&gWb#iQMX<{=kV{dMBa%o~OUtei -%X>?y-E^v8JO928D0~7!N00;of09jiqgXtBu0RRBK0{{Rq00000000000001_f&Gpc0B~t=FJE?LZe( -wAFLiQkY-wUMFJo_RbaH88FJEDBaAj_1X>Mg-Wo~w9a&K-faCuNm0Rj{Q6aWAK2ms3fSz8M3lZj*u00 -0~-001Ze0000000000005+c0g)I0aA|NaUv_0~WN&gWb#iQMX<{=kV{dMBa%o~OVQ_F|Zf9w3WiD`eP -)h*<6ay3h000O8%K%wh6=~@=QU(A3un_c!Jc4cm4Z*nhna%^mAVlyveZ*FvQX<{#Md2euKZgX>NE^v8JO928D0~7!N00;of09ji%{d6 -_GB>(`9h5!IA00000000000001_fmyj20B~t=FJE?LZe(wAFLiQkY-wUMFJo_RbaH88FLPyMb#i5Na$ -#c!Jc4cm4Z*nhna%^mAVlyveZ*FvQX<{#5b7f<7a%FUKVQzD9Z*p`mUtei%X>?y --E^v8JO928D0~7!N00;of09jiu&n->v5C8z+L;wId00000000000001_fmP@j0B~t=FJE?LZe(wAFLi -QkY-wUMFJo_RbaH88FJE(IV|8+6baG*Cb8v5RbT49QZe(e0XLBxac~DCM0u%!j000080LuVbTdpdnyj -c+d0Ae-(0672v00000000000HlGL`4|9jX>c!Jc4cm4Z*nhna%^mAVlyveZ*FvQX<{#5b7f<7a%FUKV -QzD9Z*p`mY;Sj8Y-M(3Y%XwlP)h*<6ay3h000O8%K%wh000000ssI200000C;$Ke0000000000q=88b -831r;a4%nWWo~3|axZmqY;0*_GcRyqV{2h&WpgiIUukY>bYEXCaCuNm0Rj{Q6aWAK2ms3fSzBC`8LTE -D003!+001KZ0000000000005+ckqa3BaA|NaUv_0~WN&gWb#iQMX<{=kaA9L>VP|D?FLP;lE^v8JO92 -8D0~7!N00;of09jiA00002000000000o00000000000001_f%q&L0B~t=FJE?LZe(wAFLiQkY-wUMFK -}UFYhh<)b1!0HV{344a&&VqUtei%X>?y-E^v8JO928D0~7!N00;of09jk!xMQD1^@ux6aWA^00000000000001_f#NS20B~t=FJE?LZe(wAFLiQkY-wU -MFK}UFYhh<)b1!0HV{344a&&VqcV%H~a%E;;W@&C=Y-xIBE^v8JO928D0~7!N00;of09jk+ikX_&0RR -Al1ONae00000000000001_fyFi%0B~t=FJE?LZe(wAFLiQkY-wUMFLiWjY%gD5X>MtBUtcb8c~DCM0u -%!j000080LuVbTkNXj51I!603{Ou044wc00000000000HlH6I2izNX>c!Jc4cm4Z*nhna%^mAVlyvwb -ZKlaV{dM5Wn*+{Z*DGdc~DCM0u%!j000080LuVbTMExy2VMdI07V7>03iSX00000000000HlGvKp6mV -X>c!Jc4cm4Z*nhna%^mAVlyvwbZKlaaB^>Wc`k5yP)h*<6ay3h000O8%K%whP6PIz`~Uy|@&Nz?X>2cXb!ByBE^v8JO928D0~7!N00;of09j -i%j&1R)1pok@4*&oo00000000000001_ft5rV0B~t=FJE?LZe(wAFLiQkY-wUMFLiWjY%g+Uadl;LbS -`jtP)h*<6ay3h000O8%K%whHYnA!U?X>2cYWpi+EZgXWWaCuNm0Rj{Q6aWAK2ms3fSzBKfFZT);001IZ001EX0000000000005+c -BTyLtaA|NaUv_0~WN&gWb#iQMX<{=kb#!TLFLGsca(OOrc~DCM0u%!j000080LuVbTR1R}=I#~%05n7 -Z03ZMW00000000000HlFzXBhx+X>c!Jc4cm4Z*nhna%^mAVlyvwbZKlab8~E8E^v8JO928D0~7!N00; -of09jic!Jc4cm4Z*nhna%^mAVlyvwbZKlabZKp6Z*_DoaCuNm0Rj{ -Q6aWAK2ms3fSz9aB=w25S007uG0018V0000000000005+cIh+{)aA|NaUv_0~WN&gWb#iQMX<{=kb#! -TLFLiQkE^v8JO928D0~7!N00;of09jjB)K=BM2LJ#Y6#xJr00000000000001_fsL;j0B~t=FJE?LZe -(wAFLiQkY-wUMFLiWjY%h0VX>=}dc~DCQ1^@s60J;IX0sn^p0E)L60000 -""" - - -if __name__ == "__main__": - main() diff --git a/parakeet-tdt-ctc-110m/output/Decoder.mlmodelc/analytics/coremldata.bin b/parakeet-tdt-ctc-110m/output/Decoder.mlmodelc/analytics/coremldata.bin deleted file mode 100644 index 18de02b6523b282aa52b4e816db388a961557778..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/output/Decoder.mlmodelc/analytics/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:fefed2be21166b9afcc5913192d8b86b4a621ee236b459a2cc754584020d4878 -size 243 diff --git a/parakeet-tdt-ctc-110m/output/Decoder.mlmodelc/coremldata.bin b/parakeet-tdt-ctc-110m/output/Decoder.mlmodelc/coremldata.bin deleted file mode 100644 index f6cebbf78c8aa9626ae27c1a9d8b681c3bf0f441..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/output/Decoder.mlmodelc/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:66953a7662cfdc8247e5f3da0ec38847d1e198a5b5dd94b51fa1c711f7b777ac -size 543 diff --git a/parakeet-tdt-ctc-110m/output/Decoder.mlmodelc/metadata.json b/parakeet-tdt-ctc-110m/output/Decoder.mlmodelc/metadata.json deleted file mode 100644 index 84ebb9b5a894782e6def44d0b3707784e36a1757..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/output/Decoder.mlmodelc/metadata.json +++ /dev/null @@ -1,123 +0,0 @@ -[ - { - "metadataOutputVersion" : "3.0", - "shortDescription" : "Hybrid TDT Decoder (110M)", - "outputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float16", - "formattedType" : "MultiArray (Float16 1 × 640 × 2)", - "shortDescription" : "", - "shape" : "[1, 640, 2]", - "name" : "decoder_output", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float16", - "formattedType" : "MultiArray (Float16 1 × 1 × 640)", - "shortDescription" : "", - "shape" : "[1, 1, 640]", - "name" : "h_out", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float16", - "formattedType" : "MultiArray (Float16 1 × 1 × 640)", - "shortDescription" : "", - "shape" : "[1, 1, 640]", - "name" : "c_out", - "type" : "MultiArray" - } - ], - "storagePrecision" : "Float16", - "modelParameters" : [ - - ], - "author" : "Fluid Inference", - "specificationVersion" : 8, - "mlProgramOperationTypeHistogram" : { - "Select" : 1, - "Ios17.squeeze" : 2, - "Ios17.gather" : 1, - "Ios17.cast" : 5, - "Ios17.lstm" : 1, - "Ios17.transpose" : 2, - "Ios17.add" : 1, - "Ios17.concat" : 1, - "Ios17.greaterEqual" : 1, - "Identity" : 1, - "Ios17.expandDims" : 2 - }, - "computePrecision" : "Mixed (Float16, Int16, Int32)", - "isUpdatable" : "0", - "stateSchema" : [ - - ], - "availability" : { - "macOS" : "14.0", - "tvOS" : "17.0", - "visionOS" : "1.0", - "watchOS" : "10.0", - "iOS" : "17.0", - "macCatalyst" : "17.0" - }, - "modelType" : { - "name" : "MLModelType_mlProgram" - }, - "inputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1 × 1)", - "shortDescription" : "", - "shape" : "[1, 1]", - "name" : "targets", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1)", - "shortDescription" : "", - "shape" : "[1]", - "name" : "target_lengths", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 1 × 640)", - "shortDescription" : "", - "shape" : "[1, 1, 640]", - "name" : "h_in", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 1 × 640)", - "shortDescription" : "", - "shape" : "[1, 1, 640]", - "name" : "c_in", - "type" : "MultiArray" - } - ], - "userDefinedMetadata" : { - "com.github.apple.coremltools.conversion_date" : "2026-01-01", - "com.github.apple.coremltools.source" : "torch==2.8.0", - "com.github.apple.coremltools.version" : "9.0", - "com.github.apple.coremltools.source_dialect" : "TorchScript" - }, - "generatedClassName" : "Decoder", - "method" : "predict" - } -] \ No newline at end of file diff --git a/parakeet-tdt-ctc-110m/output/Decoder.mlmodelc/model.mil b/parakeet-tdt-ctc-110m/output/Decoder.mlmodelc/model.mil deleted file mode 100644 index 3d8abb8f2d6b323d3d0fd6e3d34bc1f8145e5ca4..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/output/Decoder.mlmodelc/model.mil +++ /dev/null @@ -1,51 +0,0 @@ -program(1.0) -[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3510.2.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.8.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "9.0"}})] -{ - func main(tensor c_in, tensor h_in, tensor target_lengths, tensor targets) { - tensor y_batch_dims_0 = const()[name = tensor("y_batch_dims_0"), val = tensor(0)]; - tensor y_validate_indices_0 = const()[name = tensor("y_validate_indices_0"), val = tensor(false)]; - tensor decoder_prediction_embed_weight_to_fp16 = const()[name = tensor("decoder_prediction_embed_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; - tensor targets_to_int16_dtype_0 = const()[name = tensor("targets_to_int16_dtype_0"), val = tensor("int16")]; - tensor cast_4_dtype_0 = const()[name = tensor("cast_4_dtype_0"), val = tensor("int32")]; - tensor greater_equal_0_y_0 = const()[name = tensor("greater_equal_0_y_0"), val = tensor(0)]; - tensor targets_to_int16 = cast(dtype = targets_to_int16_dtype_0, x = targets)[name = tensor("cast_9")]; - tensor cast_4 = cast(dtype = cast_4_dtype_0, x = targets_to_int16)[name = tensor("cast_8")]; - tensor greater_equal_0 = greater_equal(x = cast_4, y = greater_equal_0_y_0)[name = tensor("greater_equal_0")]; - tensor slice_by_index_0 = const()[name = tensor("slice_by_index_0"), val = tensor(1025)]; - tensor add_1 = add(x = cast_4, y = slice_by_index_0)[name = tensor("add_1")]; - tensor select_0 = select(a = cast_4, b = add_1, cond = greater_equal_0)[name = tensor("select_0")]; - tensor y_cast_fp16_cast_uint16_axis_0 = const()[name = tensor("y_cast_fp16_cast_uint16_axis_0"), val = tensor(0)]; - tensor select_0_to_int16_dtype_0 = const()[name = tensor("select_0_to_int16_dtype_0"), val = tensor("int16")]; - tensor select_0_to_int16 = cast(dtype = select_0_to_int16_dtype_0, x = select_0)[name = tensor("cast_7")]; - tensor y_cast_fp16_cast_uint16_cast_uint16 = gather(axis = y_cast_fp16_cast_uint16_axis_0, batch_dims = y_batch_dims_0, indices = select_0_to_int16, validate_indices = y_validate_indices_0, x = decoder_prediction_embed_weight_to_fp16)[name = tensor("y_cast_fp16_cast_uint16_cast_uint16")]; - tensor input_1_perm_0 = const()[name = tensor("input_1_perm_0"), val = tensor([1, 0, 2])]; - tensor input_lstm_h0_squeeze_axes_0 = const()[name = tensor("input_lstm_h0_squeeze_axes_0"), val = tensor([0])]; - tensor h_in_to_fp16_dtype_0 = const()[name = tensor("h_in_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor h_in_to_fp16 = cast(dtype = h_in_to_fp16_dtype_0, x = h_in)[name = tensor("cast_6")]; - tensor input_lstm_h0_squeeze_cast_fp16 = squeeze(axes = input_lstm_h0_squeeze_axes_0, x = h_in_to_fp16)[name = tensor("input_lstm_h0_squeeze_cast_fp16")]; - tensor input_lstm_c0_squeeze_axes_0 = const()[name = tensor("input_lstm_c0_squeeze_axes_0"), val = tensor([0])]; - tensor c_in_to_fp16_dtype_0 = const()[name = tensor("c_in_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor c_in_to_fp16 = cast(dtype = c_in_to_fp16_dtype_0, x = c_in)[name = tensor("cast_5")]; - tensor input_lstm_c0_squeeze_cast_fp16 = squeeze(axes = input_lstm_c0_squeeze_axes_0, x = c_in_to_fp16)[name = tensor("input_lstm_c0_squeeze_cast_fp16")]; - tensor input_direction_0 = const()[name = tensor("input_direction_0"), val = tensor("forward")]; - tensor input_output_sequence_0 = const()[name = tensor("input_output_sequence_0"), val = tensor(true)]; - tensor input_recurrent_activation_0 = const()[name = tensor("input_recurrent_activation_0"), val = tensor("sigmoid")]; - tensor input_cell_activation_0 = const()[name = tensor("input_cell_activation_0"), val = tensor("tanh")]; - tensor input_activation_0 = const()[name = tensor("input_activation_0"), val = tensor("tanh")]; - tensor concat_1_to_fp16 = const()[name = tensor("concat_1_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1312128)))]; - tensor concat_2_to_fp16 = const()[name = tensor("concat_2_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4588992)))]; - tensor concat_0_to_fp16 = const()[name = tensor("concat_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7865856)))]; - tensor input_1_cast_fp16 = transpose(perm = input_1_perm_0, x = y_cast_fp16_cast_uint16_cast_uint16)[name = tensor("transpose_2")]; - tensor input_cast_fp16_0, tensor input_cast_fp16_1, tensor input_cast_fp16_2 = lstm(activation = input_activation_0, bias = concat_0_to_fp16, cell_activation = input_cell_activation_0, direction = input_direction_0, initial_c = input_lstm_c0_squeeze_cast_fp16, initial_h = input_lstm_h0_squeeze_cast_fp16, output_sequence = input_output_sequence_0, recurrent_activation = input_recurrent_activation_0, weight_hh = concat_2_to_fp16, weight_ih = concat_1_to_fp16, x = input_1_cast_fp16)[name = tensor("input_cast_fp16")]; - tensor var_30_axes_0 = const()[name = tensor("op_30_axes_0"), val = tensor([0])]; - tensor h_out = expand_dims(axes = var_30_axes_0, x = input_cast_fp16_1)[name = tensor("op_30_cast_fp16")]; - tensor var_31_axes_0 = const()[name = tensor("op_31_axes_0"), val = tensor([0])]; - tensor c_out = expand_dims(axes = var_31_axes_0, x = input_cast_fp16_2)[name = tensor("op_31_cast_fp16")]; - tensor transpose_0_perm_0 = const()[name = tensor("transpose_0_perm_0"), val = tensor([1, 2, 0])]; - tensor var_48 = const()[name = tensor("op_48"), val = tensor(2)]; - tensor var_49_interleave_0 = const()[name = tensor("op_49_interleave_0"), val = tensor(false)]; - tensor transpose_0_cast_fp16 = transpose(perm = transpose_0_perm_0, x = input_cast_fp16_0)[name = tensor("transpose_1")]; - tensor decoder_output = concat(axis = var_48, interleave = var_49_interleave_0, values = (transpose_0_cast_fp16, transpose_0_cast_fp16))[name = tensor("op_49_cast_fp16")]; - tensor target_lengths_tmp = identity(x = target_lengths)[name = tensor("target_lengths_tmp")]; - } -> (decoder_output, h_out, c_out); -} \ No newline at end of file diff --git a/parakeet-tdt-ctc-110m/output/Decoder.mlmodelc/weights/weight.bin b/parakeet-tdt-ctc-110m/output/Decoder.mlmodelc/weights/weight.bin deleted file mode 100644 index 4543ad4ea89b21ebed0de0d5ec3975d07eace7fb..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/output/Decoder.mlmodelc/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:dd90b58597ee2c172c672dffe13b1110898ba07394c1a15efc96cc8c6b18411b -size 7871040 diff --git a/parakeet-tdt-ctc-110m/output/Decoder.mlpackage/Data/com.apple.CoreML/model.mlmodel b/parakeet-tdt-ctc-110m/output/Decoder.mlpackage/Data/com.apple.CoreML/model.mlmodel deleted file mode 100644 index 70eaf8653921b3a9c2809f32857e034913cc2e86..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/output/Decoder.mlpackage/Data/com.apple.CoreML/model.mlmodel +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:75a807f2b8ec1cde294f77fe8a1faf50229f011a8e159d7d6c6beaab778ce8c1 -size 7501 diff --git a/parakeet-tdt-ctc-110m/output/Decoder.mlpackage/Data/com.apple.CoreML/weights/weight.bin b/parakeet-tdt-ctc-110m/output/Decoder.mlpackage/Data/com.apple.CoreML/weights/weight.bin deleted file mode 100644 index 4543ad4ea89b21ebed0de0d5ec3975d07eace7fb..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/output/Decoder.mlpackage/Data/com.apple.CoreML/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:dd90b58597ee2c172c672dffe13b1110898ba07394c1a15efc96cc8c6b18411b -size 7871040 diff --git a/parakeet-tdt-ctc-110m/output/Decoder.mlpackage/Manifest.json b/parakeet-tdt-ctc-110m/output/Decoder.mlpackage/Manifest.json deleted file mode 100644 index b24fa578e7986b6fc0ac0dba9f4481b1a819b0fe..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/output/Decoder.mlpackage/Manifest.json +++ /dev/null @@ -1,18 +0,0 @@ -{ - "fileFormatVersion": "1.0.0", - "itemInfoEntries": { - "6BDFA15B-7DF3-44B6-A308-B95282CBC63A": { - "author": "com.apple.CoreML", - "description": "CoreML Model Specification", - "name": "model.mlmodel", - "path": "com.apple.CoreML/model.mlmodel" - }, - "A04A1CE0-C74A-460D-B289-34CAF7B6F6C9": { - "author": "com.apple.CoreML", - "description": "CoreML Model Weights", - "name": "weights", - "path": "com.apple.CoreML/weights" - } - }, - "rootModelIdentifier": "6BDFA15B-7DF3-44B6-A308-B95282CBC63A" -} diff --git a/parakeet-tdt-ctc-110m/output/JointDecision.mlmodelc/analytics/coremldata.bin b/parakeet-tdt-ctc-110m/output/JointDecision.mlmodelc/analytics/coremldata.bin deleted file mode 100644 index f7d619494d2f2061ae52418edbb5c5fc2ec1f11e..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/output/JointDecision.mlmodelc/analytics/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:e51c4ee19e8d7e8392896e5d3b5464892f9e16ba271335ff7eb3d2eb4a015f27 -size 243 diff --git a/parakeet-tdt-ctc-110m/output/JointDecision.mlmodelc/coremldata.bin b/parakeet-tdt-ctc-110m/output/JointDecision.mlmodelc/coremldata.bin deleted file mode 100644 index cc76621d9ad55dc03a554246d198a4851607172f..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/output/JointDecision.mlmodelc/coremldata.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:5c45adaff9e0e6d8c49949666f07a6fcde0affc7a631a8f93d12a0e51c1393d3 -size 509 diff --git a/parakeet-tdt-ctc-110m/output/JointDecision.mlmodelc/metadata.json b/parakeet-tdt-ctc-110m/output/JointDecision.mlmodelc/metadata.json deleted file mode 100644 index 4d8cd2c6a2393dfafad11338ff96f388b2d59295..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/output/JointDecision.mlmodelc/metadata.json +++ /dev/null @@ -1,104 +0,0 @@ -[ - { - "metadataOutputVersion" : "3.0", - "shortDescription" : "Hybrid Joint Decision (110M)", - "outputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1 × 1 × 1)", - "shortDescription" : "", - "shape" : "[1, 1, 1]", - "name" : "token_id", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float16", - "formattedType" : "MultiArray (Float16 1 × 1 × 1)", - "shortDescription" : "", - "shape" : "[1, 1, 1]", - "name" : "token_prob", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Int32", - "formattedType" : "MultiArray (Int32 1 × 1 × 1)", - "shortDescription" : "", - "shape" : "[1, 1, 1]", - "name" : "duration", - "type" : "MultiArray" - } - ], - "storagePrecision" : "Float16", - "modelParameters" : [ - - ], - "author" : "Fluid Inference", - "specificationVersion" : 8, - "mlProgramOperationTypeHistogram" : { - "Ios17.reduceArgmax" : 2, - "Ios17.linear" : 3, - "Ios17.transpose" : 2, - "Ios17.log" : 1, - "Ios17.add" : 1, - "Ios17.sliceByIndex" : 2, - "Ios16.relu" : 1, - "Ios16.softmax" : 2, - "Ios17.expandDims" : 3, - "Ios17.squeeze" : 2, - "Ios17.cast" : 5, - "Ios17.gatherAlongAxis" : 1 - }, - "computePrecision" : "Mixed (Float16, Float32, Int16, Int32)", - "isUpdatable" : "0", - "stateSchema" : [ - - ], - "availability" : { - "macOS" : "14.0", - "tvOS" : "17.0", - "visionOS" : "1.0", - "watchOS" : "10.0", - "iOS" : "17.0", - "macCatalyst" : "17.0" - }, - "modelType" : { - "name" : "MLModelType_mlProgram" - }, - "inputSchema" : [ - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 512 × 1)", - "shortDescription" : "", - "shape" : "[1, 512, 1]", - "name" : "encoder_step", - "type" : "MultiArray" - }, - { - "hasShapeFlexibility" : "0", - "isOptional" : "0", - "dataType" : "Float32", - "formattedType" : "MultiArray (Float32 1 × 640 × 1)", - "shortDescription" : "", - "shape" : "[1, 640, 1]", - "name" : "decoder_step", - "type" : "MultiArray" - } - ], - "userDefinedMetadata" : { - "com.github.apple.coremltools.conversion_date" : "2026-01-01", - "com.github.apple.coremltools.source" : "torch==2.8.0", - "com.github.apple.coremltools.version" : "9.0", - "com.github.apple.coremltools.source_dialect" : "TorchScript" - }, - "generatedClassName" : "JointDecision", - "method" : "predict" - } -] \ No newline at end of file diff --git a/parakeet-tdt-ctc-110m/output/JointDecision.mlmodelc/model.mil b/parakeet-tdt-ctc-110m/output/JointDecision.mlmodelc/model.mil deleted file mode 100644 index 7057845c459a15cef5cb1a4baf0695b949c7f703..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/output/JointDecision.mlmodelc/model.mil +++ /dev/null @@ -1,66 +0,0 @@ -program(1.0) -[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3510.2.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.8.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "9.0"}})] -{ - func main(tensor decoder_step, tensor encoder_step) { - tensor input_1_perm_0 = const()[name = tensor("input_1_perm_0"), val = tensor([0, 2, 1])]; - tensor encoder_step_to_fp16_dtype_0 = const()[name = tensor("encoder_step_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor input_3_perm_0 = const()[name = tensor("input_3_perm_0"), val = tensor([0, 2, 1])]; - tensor decoder_step_to_fp16_dtype_0 = const()[name = tensor("decoder_step_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor joint_enc_weight_to_fp16 = const()[name = tensor("joint_enc_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; - tensor joint_enc_bias_to_fp16 = const()[name = tensor("joint_enc_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(655488)))]; - tensor encoder_step_to_fp16 = cast(dtype = encoder_step_to_fp16_dtype_0, x = encoder_step)[name = tensor("cast_5")]; - tensor input_1_cast_fp16 = transpose(perm = input_1_perm_0, x = encoder_step_to_fp16)[name = tensor("transpose_1")]; - tensor linear_0_cast_fp16 = linear(bias = joint_enc_bias_to_fp16, weight = joint_enc_weight_to_fp16, x = input_1_cast_fp16)[name = tensor("linear_0_cast_fp16")]; - tensor f_axes_0 = const()[name = tensor("f_axes_0"), val = tensor([2])]; - tensor f_cast_fp16 = expand_dims(axes = f_axes_0, x = linear_0_cast_fp16)[name = tensor("f_cast_fp16")]; - tensor joint_pred_weight_to_fp16 = const()[name = tensor("joint_pred_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(656832)))]; - tensor joint_pred_bias_to_fp16 = const()[name = tensor("joint_pred_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1476096)))]; - tensor decoder_step_to_fp16 = cast(dtype = decoder_step_to_fp16_dtype_0, x = decoder_step)[name = tensor("cast_4")]; - tensor input_3_cast_fp16 = transpose(perm = input_3_perm_0, x = decoder_step_to_fp16)[name = tensor("transpose_0")]; - tensor linear_1_cast_fp16 = linear(bias = joint_pred_bias_to_fp16, weight = joint_pred_weight_to_fp16, x = input_3_cast_fp16)[name = tensor("linear_1_cast_fp16")]; - tensor g_axes_0 = const()[name = tensor("g_axes_0"), val = tensor([1])]; - tensor g_cast_fp16 = expand_dims(axes = g_axes_0, x = linear_1_cast_fp16)[name = tensor("g_cast_fp16")]; - tensor input_5_cast_fp16 = add(x = f_cast_fp16, y = g_cast_fp16)[name = tensor("input_5_cast_fp16")]; - tensor input_7_cast_fp16 = relu(x = input_5_cast_fp16)[name = tensor("input_7_cast_fp16")]; - tensor joint_joint_net_2_weight_to_fp16 = const()[name = tensor("joint_joint_net_2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1477440)))]; - tensor joint_joint_net_2_bias_to_fp16 = const()[name = tensor("joint_joint_net_2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2795904)))]; - tensor linear_2_cast_fp16 = linear(bias = joint_joint_net_2_bias_to_fp16, weight = joint_joint_net_2_weight_to_fp16, x = input_7_cast_fp16)[name = tensor("linear_2_cast_fp16")]; - tensor var_35 = const()[name = tensor("op_35"), val = tensor(-1)]; - tensor joint_out_softmax_cast_fp16 = softmax(axis = var_35, x = linear_2_cast_fp16)[name = tensor("joint_out_softmax_cast_fp16")]; - tensor joint_out_epsilon_0 = const()[name = tensor("joint_out_epsilon_0"), val = tensor(0x1p-149)]; - tensor joint_out_cast_fp16 = log(epsilon = joint_out_epsilon_0, x = joint_out_softmax_cast_fp16)[name = tensor("joint_out_cast_fp16")]; - tensor input_begin_0 = const()[name = tensor("input_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor input_end_0 = const()[name = tensor("input_end_0"), val = tensor([1, 1, 1, 1025])]; - tensor input_end_mask_0 = const()[name = tensor("input_end_mask_0"), val = tensor([true, true, true, false])]; - tensor input_cast_fp16 = slice_by_index(begin = input_begin_0, end = input_end_0, end_mask = input_end_mask_0, x = joint_out_cast_fp16)[name = tensor("input_cast_fp16")]; - tensor duration_logits_begin_0 = const()[name = tensor("duration_logits_begin_0"), val = tensor([0, 0, 0, 1025])]; - tensor duration_logits_end_0 = const()[name = tensor("duration_logits_end_0"), val = tensor([1, 1, 1, 1030])]; - tensor duration_logits_end_mask_0 = const()[name = tensor("duration_logits_end_mask_0"), val = tensor([true, true, true, true])]; - tensor duration_logits_cast_fp16 = slice_by_index(begin = duration_logits_begin_0, end = duration_logits_end_0, end_mask = duration_logits_end_mask_0, x = joint_out_cast_fp16)[name = tensor("duration_logits_cast_fp16")]; - tensor var_48 = const()[name = tensor("op_48"), val = tensor(-1)]; - tensor probs_cast_fp16 = softmax(axis = var_48, x = input_cast_fp16)[name = tensor("probs_cast_fp16")]; - tensor token_id_1_axis_0 = const()[name = tensor("token_id_1_axis_0"), val = tensor(-1)]; - tensor token_id_1_keep_dims_0 = const()[name = tensor("token_id_1_keep_dims_0"), val = tensor(true)]; - tensor token_id_1_output_dtype_0 = const()[name = tensor("token_id_1_output_dtype_0"), val = tensor("int32")]; - tensor token_id_1_cast_fp16 = reduce_argmax(axis = token_id_1_axis_0, keep_dims = token_id_1_keep_dims_0, output_dtype = token_id_1_output_dtype_0, x = input_cast_fp16)[name = tensor("token_id_1_cast_fp16")]; - tensor token_id_axes_0 = const()[name = tensor("token_id_axes_0"), val = tensor([-1])]; - tensor token_id_1_cast_fp16_to_int16_dtype_0 = const()[name = tensor("token_id_1_cast_fp16_to_int16_dtype_0"), val = tensor("int16")]; - tensor token_id_1_cast_fp16_to_int16 = cast(dtype = token_id_1_cast_fp16_to_int16_dtype_0, x = token_id_1_cast_fp16)[name = tensor("cast_3")]; - tensor token_id_cast_int16 = squeeze(axes = token_id_axes_0, x = token_id_1_cast_fp16_to_int16)[name = tensor("token_id_cast_int16")]; - tensor token_id_cast_int16_to_int32_dtype_0 = const()[name = tensor("token_id_cast_int16_to_int32_dtype_0"), val = tensor("int32")]; - tensor var_57_axes_0 = const()[name = tensor("op_57_axes_0"), val = tensor([-1])]; - tensor token_id = cast(dtype = token_id_cast_int16_to_int32_dtype_0, x = token_id_cast_int16)[name = tensor("cast_2")]; - tensor var_57 = expand_dims(axes = var_57_axes_0, x = token_id)[name = tensor("op_57")]; - tensor var_58 = const()[name = tensor("op_58"), val = tensor(-1)]; - tensor token_prob_validate_indices_0 = const()[name = tensor("token_prob_validate_indices_0"), val = tensor(false)]; - tensor var_57_to_int16_dtype_0 = const()[name = tensor("op_57_to_int16_dtype_0"), val = tensor("int16")]; - tensor var_57_to_int16 = cast(dtype = var_57_to_int16_dtype_0, x = var_57)[name = tensor("cast_1")]; - tensor token_prob_cast_fp16_cast_int16 = gather_along_axis(axis = var_58, indices = var_57_to_int16, validate_indices = token_prob_validate_indices_0, x = probs_cast_fp16)[name = tensor("token_prob_cast_fp16_cast_int16")]; - tensor var_62_axes_0 = const()[name = tensor("op_62_axes_0"), val = tensor([-1])]; - tensor token_prob = squeeze(axes = var_62_axes_0, x = token_prob_cast_fp16_cast_int16)[name = tensor("op_62_cast_fp16")]; - tensor duration_axis_0 = const()[name = tensor("duration_axis_0"), val = tensor(-1)]; - tensor duration_keep_dims_0 = const()[name = tensor("duration_keep_dims_0"), val = tensor(false)]; - tensor duration_output_dtype_0 = const()[name = tensor("duration_output_dtype_0"), val = tensor("int32")]; - tensor duration = reduce_argmax(axis = duration_axis_0, keep_dims = duration_keep_dims_0, output_dtype = duration_output_dtype_0, x = duration_logits_cast_fp16)[name = tensor("duration_cast_fp16")]; - } -> (token_id, token_prob, duration); -} \ No newline at end of file diff --git a/parakeet-tdt-ctc-110m/output/JointDecision.mlmodelc/weights/weight.bin b/parakeet-tdt-ctc-110m/output/JointDecision.mlmodelc/weights/weight.bin deleted file mode 100644 index 747ec9d24e2d3e008bc869d2c404bb3d2653325c..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/output/JointDecision.mlmodelc/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:b3f771cb65b190f1873e39629676ed79b65a8361522f451b37bdba8b1106e6ff -size 2798028 diff --git a/parakeet-tdt-ctc-110m/output/JointDecision.mlpackage/Data/com.apple.CoreML/model.mlmodel b/parakeet-tdt-ctc-110m/output/JointDecision.mlpackage/Data/com.apple.CoreML/model.mlmodel deleted file mode 100644 index 00748a425eda1d051f13fc524fcfc160d1cb7ca1..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/output/JointDecision.mlpackage/Data/com.apple.CoreML/model.mlmodel +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:73d76247715622313f066e93874f5eff0706410e8572e1b3c51f6f7f2b14a78b -size 9790 diff --git a/parakeet-tdt-ctc-110m/output/JointDecision.mlpackage/Data/com.apple.CoreML/weights/weight.bin b/parakeet-tdt-ctc-110m/output/JointDecision.mlpackage/Data/com.apple.CoreML/weights/weight.bin deleted file mode 100644 index 747ec9d24e2d3e008bc869d2c404bb3d2653325c..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/output/JointDecision.mlpackage/Data/com.apple.CoreML/weights/weight.bin +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:b3f771cb65b190f1873e39629676ed79b65a8361522f451b37bdba8b1106e6ff -size 2798028 diff --git a/parakeet-tdt-ctc-110m/output/JointDecision.mlpackage/Manifest.json b/parakeet-tdt-ctc-110m/output/JointDecision.mlpackage/Manifest.json deleted file mode 100644 index 51edbc5622d98d269493af788ab212f926544b0c..0000000000000000000000000000000000000000 --- a/parakeet-tdt-ctc-110m/output/JointDecision.mlpackage/Manifest.json +++ /dev/null @@ -1,18 +0,0 @@ -{ - "fileFormatVersion": "1.0.0", - "itemInfoEntries": { - "601AC192-A217-4675-8F63-01F426D7BC52": { - "author": "com.apple.CoreML", - "description": "CoreML Model Weights", - "name": "weights", - "path": "com.apple.CoreML/weights" - }, - "76B2BF33-9547-4031-8668-01C0E473C306": { - "author": "com.apple.CoreML", - "description": "CoreML Model Specification", - "name": "model.mlmodel", - "path": "com.apple.CoreML/model.mlmodel" - } - }, - "rootModelIdentifier": "76B2BF33-9547-4031-8668-01C0E473C306" -} diff --git a/parakeet-tdt-v2-0.6b/.DS_Store b/parakeet-tdt-v2-0.6b/.DS_Store deleted file mode 100644 index 4a654c3e785ab1ee5f435f3eda3c58ea35b780d4..0000000000000000000000000000000000000000 Binary files a/parakeet-tdt-v2-0.6b/.DS_Store and /dev/null differ diff --git a/parakeet-tdt-v2-0.6b/coreml/.gitignore b/parakeet-tdt-v2-0.6b/coreml/.gitignore deleted file mode 100644 index e2e12b1c4fd2deeef4b38477dccac90f560a627c..0000000000000000000000000000000000000000 --- a/parakeet-tdt-v2-0.6b/coreml/.gitignore +++ /dev/null @@ -1,4 +0,0 @@ -mlpackages/ -parakeet_coreml/ -parakeet_coreml_quantized/ -compiled/ diff --git a/parakeet-tdt-v2-0.6b/coreml/README.md b/parakeet-tdt-v2-0.6b/coreml/README.md deleted file mode 100644 index 8ef855d29d09b61830ca71b880c99e953f44f667..0000000000000000000000000000000000000000 --- a/parakeet-tdt-v2-0.6b/coreml/README.md +++ /dev/null @@ -1,136 +0,0 @@ -# Parakeet‑TDT v2 (0.6B) — CoreML Export, Parity, and Quantization - -Tools to export NVIDIA Parakeet‑TDT v2 (0.6B) RNNT ASR to CoreML, validate numerical parity with the NeMo reference, measure latency, and explore quantization trade‑offs. All CoreML components use a fixed 15‑second audio window for export and validation. - -## Environment - -1. Create or reuse the local environment with `uv venv`. -2. Activate the repo `.venv` and install deps via `uv pip sync`. -3. Run everything through `uv run` to keep resolutions reproducible. - -## Test Environment - -All tests and measurements referenced here were run on an Apple M4 Pro with 48 GB of RAM. - -## Export CoreML packages - -Exports preprocessor, encoder, decoder, joint, and two fused variants (mel+encoder, joint+decision). Shapes and I/O match the fixed 15‑second window contract. - -``` -uv run python convert-parakeet.py convert \ - --nemo-path /path/to/parakeet-tdt-0.6b-v2.nemo \ - --output-dir parakeet_coreml -``` - -Notes -- Minimum deployment target: iOS 17. Export uses CPU_ONLY by default; runtime compute units can be set when loading the model (Python or Swift). -- Audio is 16 kHz, single‑channel. The 15 s window is enforced during export and validation. - -## Validate parity and speed (Torch vs CoreML) - -Runs Torch and CoreML side‑by‑side on the same 15 s input, records diffs and latency, and saves plots under `plots/compare-components/`. The tool updates `parakeet_coreml/metadata.json` with all measurements. - -``` -uv run python compare-components.py compare \ - --output-dir parakeet_coreml \ - --model-id nvidia/parakeet-tdt-0.6b-v2 \ - --runs 10 --warmup 3 -``` - -Output comparison: - -![./plots/compare-components/mel_encoder_time_l2.png](./plots/compare-components/mel_encoder_time_l2.png) -![./plots/compare-components/joint_decision_prob_u0.png](./plots/compare-components/joint_decision_prob_u0.png) -![./plots/compare-components/decoder_steps_l2.png](./plots/compare-components/decoder_steps_l2.png) - -Latency: - -![./plots/quantize/all/all_components_compile.png](./plots/quantize/all/all_components_compile.png) -![./plots/quantize/all/all_components_compression.png](./plots/quantize/all/all_components_compression.png) -![./plots/quantize/all/all_components_quality.png](./plots/quantize/all/all_components_quality.png) -![./plots/quantize/all/all_components_latency.png](./plots/quantize/all/all_components_latency.png) - - -Quants: - - - -### Key results (quality first) - -Numerical parity is strong across components on the fixed window: -- Preprocessor mel: match=true; max_abs≈0.484, max_rel≈2.00 (near‑zero bins inflate relative error). -- Encoder: match=true; max_abs≈0.0054, strong agreement over time (see plot). -- Decoder h/c state: match=true; value deltas within tolerance. -- Joint logits: match=true; max_abs≈0.099, distributions align (see top‑k plot). -- Joint+Decision: Fused CoreML head exactly matches decisions computed on CoreML logits (token_id/prob/duration). PyTorch logits produce slightly different argmax paths (expected from small logit differences). - -### Speed (latency and RTF) - -Component latency on a 15 s clip, Torch CPU vs CoreML (CPU+NE) from `parakeet_coreml/metadata.json`: -- Encoder: Torch 1030.48 ms → CoreML 25.44 ms (≈40.5× faster, RTF 0.00170) -- Preprocessor: 1.99 ms → 1.19 ms (≈1.68×) -- Joint: 28.34 ms → 22.66 ms (≈1.25×) -- Decoder (U=1): 7.51 ms → 4.32 ms (≈1.73×) - -Fused paths: -- Mel+Encoder (Torch separate vs CoreML fused): 1032.48 ms → 27.10 ms (≈38.1× faster) -- Joint+Decision (CoreML joint + CPU post vs fused CoreML head): 50.05 ms → 64.09 ms (fused is slower here; prefer CoreML joint + lightweight CPU decision on host). - -Plots -- Latency bars and speedups: `plots/compare-components/latency_summary.png`, `plots/compare-components/latency_speedup.png` -- Fused vs separate: `plots/compare-components/latency_fused_vs_separate.png`, `plots/compare-components/latency_fused_speedup.png` -- Quality visuals: mel composite (`mel_composite.png`), encoder L2 over time (`encoder_time_l2.png`), decoder step L2 (`decoder_steps_l2.png`), joint top‑k/time L2 (`joint_top50.png`, `joint_time_l2.png`), joint‑decision agreement (`joint_decision_token_agree.png`, `joint_decision_prob_u0.png`). - -## Quantization (size ‱ quality ‱ speed) - -`uv run python quantize_coreml.py` evaluates several variants and writes a roll‑up to `parakeet_coreml_quantized/quantization_summary.json`. Plots are mirrored to `plots/quantize//` (we include `plots/quantize/all/`). Quality here is reported as 1 − normalized L2 error (1.0 = identical). For JointDecision we report token‑id match rate, duration match, and token‑prob MAE. - -Quick highlights (ComputeUnits=ALL): -- int8 linear (per‑channel): ~2.0× smaller across components with minimal quality loss - - MelEncoder quality≈0.963; latency≈31.13 ms (baseline≈29.34 ms) - - JointDecision acc≈0.995; latency≈1.96 ms (baseline≈2.15 ms) -- int8 linear (per‑tensor symmetric): large encoder quality drop (≈0.50) — not recommended - -Quantization plots (ALL) -- Fused: `plots/quantize/all/fused_quality.png`, `fused_latency.png`, `fused_compression.png`, `fused_size.png` -- Component breakdown: `plots/quantize/all/all_components_quality.png`, `all_components_latency.png`, `all_components_compression.png`, `all_components_size.png`, `all_components_compile.png` - -## Reproduce the figures - -1) Export baseline CoreML packages -``` -uv run python convert-parakeet.py convert --model-id nvidia/parakeet-tdt-0.6b-v2 --output-dir parakeet_coreml -``` - -2) Compare Torch vs CoreML and generate parity/latency plots -``` -uv run python compare-components.py compare --output-dir parakeet_coreml --runs 10 --warmup 3 -``` - -3) Run quantization sweeps (mirrors plots into `plots/quantize//`) -``` -uv run python quantize_coreml.py \ - --input-dir parakeet_coreml \ - --output-root parakeet_coreml_quantized \ - --compute-units ALL --runs 10 -``` - -Examples -- Encoder 6‑bit palette only: - `uv run python quantize_coreml.py -c encoder-palettize` -- MelEncoder 6‑bit palette only: - `uv run python quantize_coreml.py -c mel-palettize` - (By default, the script derives the component whitelist from the selected - variants. Use `-m/--component` to explicitly restrict or `-m all` to force all.) - -## Notes & limits - -- Fixed 15‑second window shapes are required for all CoreML exports and validations. -- Latency measurements are host‑side CoreML predictions (CPU+NE or ALL); on‑device results can differ by chip/OS. -- For streaming decode, the exported decoder uses U=1 inputs with explicit LSTM state I/O. -- Minimum deployment target is iOS 17; models are saved as MLProgram and eligible for ANE when loaded with `ComputeUnits=ALL`. - -## Acknowledgements - -- Parakeet‑TDT v2 model from NVIDIA NeMo (`nvidia/parakeet-tdt-0.6b-v2`). -- This directory provides export/validation utilities and plots to help the community reproduce quality and performance on Apple devices. diff --git a/parakeet-tdt-v2-0.6b/coreml/agents.md b/parakeet-tdt-v2-0.6b/coreml/agents.md deleted file mode 100644 index d6765848d59f76846549712d8c31a4fce733a98a..0000000000000000000000000000000000000000 --- a/parakeet-tdt-v2-0.6b/coreml/agents.md +++ /dev/null @@ -1,7 +0,0 @@ -# Agent Notes - -- Preferred Python workflow uses `uv` (https://github.com/astral-sh/uv). - - Create and manage environments with `uv venv`. - - Install dependencies with `uv pip install` or `uv pip sync` as needed. -- When working in this repo, activate the local `.venv` and run tooling through `uv run` to keep resolutions reproducible. -- Keep CoreML conversions constrained to the fixed 15-second audio window when exporting or validating Parakeet components. diff --git a/parakeet-tdt-v2-0.6b/coreml/audio/yc_first_minute.wav b/parakeet-tdt-v2-0.6b/coreml/audio/yc_first_minute.wav deleted file mode 100644 index 8682a32b66200c635a60706e722c2b03b5701db0..0000000000000000000000000000000000000000 --- a/parakeet-tdt-v2-0.6b/coreml/audio/yc_first_minute.wav +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:004afd9a8b6265d321b1faeb3b22ca5efcc78b570566fca633327d73f545f3fa -size 10584142 diff --git a/parakeet-tdt-v2-0.6b/coreml/audio/yc_first_minute_16k.wav b/parakeet-tdt-v2-0.6b/coreml/audio/yc_first_minute_16k.wav deleted file mode 100644 index de71da9ab403797061b7051b820337a754336481..0000000000000000000000000000000000000000 --- a/parakeet-tdt-v2-0.6b/coreml/audio/yc_first_minute_16k.wav +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:555404115420c8b4a77fcfacc450ecfc767211ca7b449a9dcb292b1734256103 -size 1920090 diff --git a/parakeet-tdt-v2-0.6b/coreml/audio/yc_first_minute_16k_15s.wav b/parakeet-tdt-v2-0.6b/coreml/audio/yc_first_minute_16k_15s.wav deleted file mode 100644 index 0b8040f5f124d20a1d7e812c576c831a7573eddf..0000000000000000000000000000000000000000 --- a/parakeet-tdt-v2-0.6b/coreml/audio/yc_first_minute_16k_15s.wav +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:c79c8bc763b4efccb3e12f199ec0a59aa2edc5e9e4d21ca70fde8f36762d4147 -size 480078 diff --git a/parakeet-tdt-v2-0.6b/coreml/compare-components.py b/parakeet-tdt-v2-0.6b/coreml/compare-components.py deleted file mode 100644 index 114b74c388a2562cc2576f1ac331b8cd61dc518e..0000000000000000000000000000000000000000 --- a/parakeet-tdt-v2-0.6b/coreml/compare-components.py +++ /dev/null @@ -1,958 +0,0 @@ -#!/usr/bin/env python3 -"""Compare Parakeet TDT v2 Torch vs CoreML components on a fixed 15s window. - -Writes numeric diffs to the specified output directory (metadata.json) and -saves plots under a repo-tracked directory: plots//. -""" -from __future__ import annotations - -import json -import time -from dataclasses import dataclass -from pathlib import Path -from typing import Dict, Optional, Tuple - -import coremltools as ct -import numpy as np -import soundfile as sf -import torch -import typer - -import nemo.collections.asr as nemo_asr - -# Optional plotting -try: - import matplotlib - matplotlib.use("Agg") - import matplotlib.pyplot as plt - HAS_MPL = True -except Exception: - HAS_MPL = False - - -@dataclass -class ValidationSettings: - audio_path: Optional[Path] - seconds: float - seed: Optional[int] - rtol: float - atol: float - - -def _compute_length(seconds: float, sample_rate: int) -> int: - return int(round(seconds * sample_rate)) - - -def _prepare_audio( - validation_audio: Optional[Path], - sample_rate: int, - max_samples: int, - seed: Optional[int], -) -> torch.Tensor: - if validation_audio is None: - if seed is not None: - torch.manual_seed(seed) - return torch.randn(1, max_samples, dtype=torch.float32) - - data, sr = sf.read(str(validation_audio), dtype="float32") - if sr != sample_rate: - raise typer.BadParameter( - f"Validation audio sample rate {sr} does not match model rate {sample_rate}" - ) - if data.ndim > 1: - data = data[:, 0] - if data.size == 0: - raise typer.BadParameter("Validation audio is empty") - if data.size < max_samples: - data = np.pad(data, (0, max_samples - data.size)) - elif data.size > max_samples: - data = data[:max_samples] - return torch.from_numpy(data).unsqueeze(0).to(dtype=torch.float32) - - -def _np(x: torch.Tensor, dtype=None) -> np.ndarray: - arr = x.detach().cpu().numpy() - if dtype is not None: - return arr.astype(dtype, copy=False) - return arr - - -def _to_t(x) -> torch.Tensor: - if isinstance(x, torch.Tensor): - return x.detach().cpu() - elif isinstance(x, np.ndarray): - # Ensure a separate tensor (avoid shared memory weirdness) - return torch.from_numpy(np.array(x, copy=True)) - else: - return torch.tensor(x) - - -def _max_diffs(a, b, rtol: float, atol: float) -> Tuple[float, float, bool]: - # Use NumPy for comparisons to avoid invoking the PyTorch C-API in contexts - # where the GIL may not be held (which can trigger PyEval_SaveThread errors). - na = np.array(a, dtype=np.float32, copy=True) - nb = np.array(b, dtype=np.float32, copy=True) - if na.size == 0: - return 0.0, 0.0, True - diff = np.abs(na - nb) - max_abs = float(diff.max()) - denom = np.maximum(np.abs(na), np.abs(nb)) - with np.errstate(divide="ignore", invalid="ignore"): - rel = np.where(denom == 0.0, 0.0, diff / denom) - max_rel = float(rel.max()) - ok = bool(np.allclose(na, nb, rtol=rtol, atol=atol)) - return max_abs, max_rel, ok - - -def _plot_line(x_ref: np.ndarray, x_ml: np.ndarray, title: str, path: Path, also_delta: bool = False): - if not HAS_MPL: - return None - if also_delta: - fig, axes = plt.subplots(2, 1, figsize=(8, 5), sharex=True) - axes[0].plot(x_ref, label="torch", linewidth=1) - axes[0].plot(x_ml, label="coreml", linewidth=1, alpha=0.8) - axes[0].set_title(title) - axes[0].legend() - delta = np.asarray(x_ref) - np.asarray(x_ml) - axes[1].plot(delta, color="C3", linewidth=1) - axes[1].set_title("Delta (torch - coreml)") - axes[1].set_xlabel("time/step") - plt.tight_layout() - plt.savefig(path) - plt.close(fig) - else: - plt.figure(figsize=(8, 3)) - plt.plot(x_ref, label="torch", linewidth=1) - plt.plot(x_ml, label="coreml", linewidth=1, alpha=0.8) - plt.title(title) - plt.legend() - plt.tight_layout() - plt.savefig(path) - plt.close() - return str(path.name) - - -def _plot_image(img: np.ndarray, title: str, path: Path, vmin=None, vmax=None): - if not HAS_MPL: - return None - plt.figure(figsize=(6, 4)) - plt.imshow(img, aspect='auto', origin='lower', interpolation='nearest', vmin=vmin, vmax=vmax) - plt.title(title) - plt.colorbar(shrink=0.8) - plt.tight_layout() - plt.savefig(path) - plt.close() - return str(path.name) - - -def _plot_mel_composite( - mel_torch: np.ndarray, - mel_coreml: np.ndarray, - path: Path, - vmin=None, - vmax=None, -): - """Create a single PNG with mel torch, mel coreml, abs diff heatmap, and mean-over-time curves with delta.""" - if not HAS_MPL: - return None - mel_torch = np.asarray(mel_torch) - mel_coreml = np.asarray(mel_coreml) - absdiff = np.abs(mel_torch - mel_coreml) - mean_t = mel_torch.mean(axis=0) - mean_c = mel_coreml.mean(axis=0) - delta = mean_t - mean_c - - fig = plt.figure(figsize=(12, 8)) - gs = fig.add_gridspec(2, 2, height_ratios=[1, 1]) - ax1 = fig.add_subplot(gs[0, 0]) - im1 = ax1.imshow(mel_torch, aspect='auto', origin='lower', interpolation='nearest', vmin=vmin, vmax=vmax) - ax1.set_title("Mel (Torch)") - fig.colorbar(im1, ax=ax1, shrink=0.8) - - ax2 = fig.add_subplot(gs[0, 1]) - im2 = ax2.imshow(mel_coreml, aspect='auto', origin='lower', interpolation='nearest', vmin=vmin, vmax=vmax) - ax2.set_title("Mel (CoreML)") - fig.colorbar(im2, ax=ax2, shrink=0.8) - - ax3 = fig.add_subplot(gs[1, 0]) - im3 = ax3.imshow(absdiff, aspect='auto', origin='lower', interpolation='nearest') - ax3.set_title("Mel |diff|") - fig.colorbar(im3, ax=ax3, shrink=0.8) - - ax4 = fig.add_subplot(gs[1, 1]) - ax4.plot(mean_t, label="torch", linewidth=1) - ax4.plot(mean_c, label="coreml", linewidth=1, alpha=0.8) - ax4.plot(delta, label="delta", linewidth=1, color="C3") - ax4.set_title("Mel mean over time + delta") - ax4.legend() - - plt.tight_layout() - plt.savefig(path) - plt.close(fig) - return str(path.name) - - -def _plot_latency_bars( - labels, - torch_means, - torch_stds, - coreml_means, - coreml_stds, - path: Path, -): - if not HAS_MPL: - return None - x = np.arange(len(labels)) - width = 0.35 - fig, ax = plt.subplots(figsize=(8, 4)) - b1 = ax.bar(x - width/2, torch_means, width, yerr=torch_stds, label="torch", color="C0", alpha=0.9) - b2 = ax.bar(x + width/2, coreml_means, width, yerr=coreml_stds, label="coreml", color="C1", alpha=0.9) - ax.set_xticks(x, labels, rotation=15) - ax.set_ylabel("latency (ms)") - ax.set_title("Component latency (15s window inputs)") - ax.legend() - # Add value labels on bars - def _annotate(bars): - for bar in bars: - h = bar.get_height() - if np.isnan(h): - continue - ax.annotate(f"{h:.0f}", - xy=(bar.get_x() + bar.get_width()/2, h), - xytext=(0, 3), textcoords="offset points", - ha='center', va='bottom', fontsize=8) - _annotate(b1) - _annotate(b2) - plt.tight_layout() - plt.savefig(path) - plt.close(fig) - return str(path.name) - - -def _plot_speedup_bars(labels, torch_means, coreml_means, path: Path): - if not HAS_MPL: - return None - speedup = [] - for t, c in zip(torch_means, coreml_means): - if c and c > 0: - speedup.append(float(t) / float(c)) - else: - speedup.append(np.nan) - x = np.arange(len(labels)) - fig, ax = plt.subplots(figsize=(8, 4)) - bars = ax.bar(x, speedup, color="C2") - ax.set_xticks(x, labels, rotation=15) - ax.set_ylabel("torch/coreml speedup") - ax.set_title("CoreML speedup vs Torch (higher is better)") - ax.axhline(1.0, color="gray", linestyle="--", linewidth=1) - # Add value labels - for bar in bars: - h = bar.get_height() - if np.isnan(h): - continue - ax.annotate(f"{h:.2f}", - xy=(bar.get_x() + bar.get_width()/2, h), - xytext=(0, 3), textcoords="offset points", - ha='center', va='bottom', fontsize=8) - plt.tight_layout() - plt.savefig(path) - plt.close(fig) - return str(path.name) - - - -app = typer.Typer(add_completion=False, pretty_exceptions_show_locals=False) - - -@app.command() -def compare( - output_dir: Path = typer.Option(Path("parakeet_coreml"), help="Directory containing mlpackages + metadata.json"), - nemo_path: Optional[Path] = typer.Option(None, "--nemo-path", exists=True, resolve_path=True, help="Path to .nemo checkpoint"), - model_id: str = typer.Option("nvidia/parakeet-tdt-0.6b-v2", "--model-id", help="HF model id if --nemo-path omitted"), - validation_audio: Optional[Path] = typer.Option(None, exists=True, resolve_path=True, help="15s, 16kHz wav for validation (defaults to audio/yc_first_minute_16k_15s.wav if present)"), - seed: Optional[int] = typer.Option(None, help="Random seed for synthetic input when audio is not provided"), - rtol: float = typer.Option(1e-3, help="Relative tolerance for comparisons"), - atol: float = typer.Option(1e-4, help="Absolute tolerance for comparisons"), - runs: int = typer.Option(10, help="Timed runs per model for latency measurement"), - warmup: int = typer.Option(3, help="Warmup runs before timing (compilation, caches)"), - symbol_steps: int = typer.Option( - 32, - help="Number of sequential decoder steps to validate with streaming U=1 inputs", - ), -) -> None: - """Run Torch vs CoreML comparisons and update metadata.json with plots and diffs.""" - output_dir.mkdir(parents=True, exist_ok=True) - if symbol_steps < 1: - raise typer.BadParameter("symbol_steps must be >= 1") - - meta_path = output_dir / "metadata.json" - exported_meta: Dict[str, object] = {} - if meta_path.exists(): - try: - exported_meta = json.loads(meta_path.read_text()) - except Exception: - exported_meta = {} - exported_max_u = int(exported_meta.get("max_symbol_steps", 1)) - if exported_max_u != 1: - typer.echo( - f"Note: CoreML export reports max_symbol_steps={exported_max_u}; " - "comparison still drives decoder step-wise with U=1 inputs." - ) - if nemo_path is not None: - typer.echo(f"Loading NeMo model from {nemo_path}
") - asr_model = nemo_asr.models.EncDecRNNTBPEModel.restore_from(str(nemo_path), map_location="cpu") - else: - typer.echo(f"Downloading NeMo model via {model_id}
") - asr_model = nemo_asr.models.EncDecRNNTBPEModel.from_pretrained(model_id, map_location="cpu") - asr_model.eval() - - sample_rate = int(asr_model.cfg.preprocessor.sample_rate) - max_samples = _compute_length(15.0, sample_rate) - default_audio = (Path(__file__).parent / "audio" / "yc_first_minute_16k_15s.wav").resolve() - chosen_audio = validation_audio if validation_audio is not None else (default_audio if default_audio.exists() else None) - if chosen_audio is not None and validation_audio is None: - typer.echo(f"Using default validation audio: {chosen_audio}") - - audio_tensor = _prepare_audio(chosen_audio, sample_rate, max_samples, seed) - audio_length = torch.tensor([max_samples], dtype=torch.int32) - - asr_model.decoder._rnnt_export = True - # Disable fused loss/WER computation for simpler joint inference - asr_model.joint.set_fuse_loss_wer(False) - # Important: ensure the joint returns raw logits (not log-softmax) - # RNNTJoint applies log_softmax on CPU by default when `log_softmax is None`. - # Our exported CoreML joint emits pre-softmax logits, so make the Torch - # reference do the same to avoid systematic offsets in comparisons/plots. - try: - # Some versions expose this as a plain attribute - asr_model.joint.log_softmax = False - except Exception: - pass - - # Generate reference outputs directly from NeMo model components - with torch.inference_mode(): - # Preprocessor - direct NeMo call - mel_ref, mel_length_ref = asr_model.preprocessor( - input_signal=audio_tensor, - length=audio_length.to(dtype=torch.long) - ) - mel_length_ref = mel_length_ref.to(dtype=torch.int32) - - # Encoder - direct NeMo call - encoder_ref, encoder_length_ref = asr_model.encoder( - audio_signal=mel_ref, - length=mel_length_ref.to(dtype=torch.long) - ) - encoder_length_ref = encoder_length_ref.to(dtype=torch.int32) - - vocab_size = int(asr_model.tokenizer.vocab_size) - num_extra = int(asr_model.joint.num_extra_outputs) - decoder_hidden = int(asr_model.decoder.pred_hidden) - decoder_layers = int(asr_model.decoder.pred_rnn_layers) - blank_id = int(asr_model.decoder.blank_idx) - - blank_targets = torch.tensor([[blank_id]], dtype=torch.int32) - blank_target_lengths = torch.tensor([1], dtype=torch.int32) - blank_targets_long = blank_targets.to(dtype=torch.long) - blank_target_lengths_long = blank_target_lengths.to(dtype=torch.long) - - def _decoder_rollout_torch(num_steps: int) -> Tuple[torch.Tensor, torch.Tensor, torch.Tensor]: - outputs = [] - h_state = torch.zeros(decoder_layers, 1, decoder_hidden, dtype=torch.float32) - c_state = torch.zeros(decoder_layers, 1, decoder_hidden, dtype=torch.float32) - state = [h_state, c_state] - with torch.inference_mode(): - for _ in range(num_steps): - y, _, new_state = asr_model.decoder( - targets=blank_targets_long, - target_length=blank_target_lengths_long, - states=state, - ) - outputs.append(y.detach()) - state = [new_state[0].detach(), new_state[1].detach()] - if outputs: - decoder_seq = torch.cat(outputs, dim=-1) - else: - decoder_seq = torch.zeros(1, decoder_hidden, 0, dtype=torch.float32) - return decoder_seq, state[0], state[1] - - decoder_ref, h_ref, c_ref = _decoder_rollout_torch(symbol_steps) - - with torch.inference_mode(): - logits_ref = asr_model.joint( - encoder_outputs=encoder_ref, - decoder_outputs=decoder_ref, - ) - - # Convert tensors to numpy for CoreML - def _np32(x): - return np.array(x.detach().cpu().numpy(), dtype=np.float32, copy=True) - - # Prepare plot dir (write to repo-tracked plots//) - plots_root = Path(__file__).parent / "plots" - plots_dir = plots_root / Path(__file__).stem - plots_dir.mkdir(parents=True, exist_ok=True) - - encoder_np = _np32(encoder_ref) - decoder_ref_np = _np32(decoder_ref) - - summary: Dict[str, object] = { - "requested": True, - "status": "ok", - "atol": atol, - "rtol": rtol, - "symbol_steps": int(symbol_steps), - "audio_path": None if validation_audio is None else str(validation_audio), - "components": {}, - } - - # Preprocessor - pre = ct.models.MLModel(str(output_dir / "parakeet_preprocessor.mlpackage"), compute_units=ct.ComputeUnit.CPU_AND_NE) - t0 = time.perf_counter() - pre_out = pre.predict({"audio_signal": _np32(audio_tensor), "audio_length": _np32(audio_length).astype(np.int32)}) - t1 = time.perf_counter() - pre_first_ms = (t1 - t0) * 1000.0 - mel_ml = np.array(pre_out["mel"], dtype=np.float32, copy=True) - mel_len_ml = np.array(pre_out["mel_length"], dtype=np.int32, copy=True) - pre_atol, pre_rtol = max(atol, 1.0), max(rtol, 1e-2) - a_mel, r_mel, ok_mel = _max_diffs(_np32(mel_ref), mel_ml, pre_rtol, pre_atol) - ok_len = int(_np32(mel_length_ref).astype(np.int32)[0]) == int(np.array(mel_len_ml).astype(np.int32)[0]) - mel_t = _np32(mel_ref)[0] - mel_c = mel_ml[0] - vmin = float(min(mel_t.min(), mel_c.min())) - vmax = float(max(mel_t.max(), mel_c.max())) - pre_plots = { - "mel_composite.png": _plot_mel_composite(mel_t, mel_c, plots_dir / "mel_composite.png", vmin=vmin, vmax=vmax), - } - # Latency measurements: Torch and CoreML - def _time_coreml(model: ct.models.MLModel, inputs: Dict[str, np.ndarray]) -> Tuple[float, float]: - # Warmup - for _ in range(max(0, warmup)): - _ = model.predict(inputs) - times = [] - for _ in range(max(1, runs)): - t0 = time.perf_counter() - _ = model.predict(inputs) - t1 = time.perf_counter() - times.append((t1 - t0) * 1000.0) - arr = np.array(times, dtype=np.float64) - return float(arr.mean()), float(arr.std(ddof=1) if arr.size > 1 else 0.0) - - def _time_torch(fn, *args, **kwargs) -> Tuple[float, float]: - with torch.inference_mode(): - for _ in range(max(0, warmup)): - _ = fn(*args, **kwargs) - times = [] - for _ in range(max(1, runs)): - t0 = time.perf_counter() - _ = fn(*args, **kwargs) - t1 = time.perf_counter() - times.append((t1 - t0) * 1000.0) - arr = np.array(times, dtype=np.float64) - return float(arr.mean()), float(arr.std(ddof=1) if arr.size > 1 else 0.0) - - pre_torch_ms_mean, pre_torch_ms_std = _time_torch( - asr_model.preprocessor, input_signal=audio_tensor, length=audio_length.to(dtype=torch.long) - ) - pre_coreml_ms_mean, pre_coreml_ms_std = _time_coreml( - pre, - {"audio_signal": _np32(audio_tensor), "audio_length": _np32(audio_length).astype(np.int32)}, - ) - seconds = 15.0 - pre_coreml_rtf = float(pre_coreml_ms_mean / (seconds * 1000.0)) if pre_coreml_ms_mean > 0 else None - pre_torch_rtf = float(pre_torch_ms_mean / (seconds * 1000.0)) if pre_torch_ms_mean > 0 else None - - summary["components"]["preprocessor"] = { - "mel": {"max_abs": a_mel, "max_rel": r_mel, "match": bool(ok_mel)}, - "length_match": bool(ok_len), - "latency": { - "runs": int(runs), - "warmup": int(warmup), - "coreml_first_ms": pre_first_ms, - "torch_ms": {"mean": pre_torch_ms_mean, "std": pre_torch_ms_std}, - "coreml_ms": {"mean": pre_coreml_ms_mean, "std": pre_coreml_ms_std}, - "rtf": {"torch": pre_torch_rtf, "coreml": pre_coreml_rtf}, - }, - "plots": {k: v for k, v in pre_plots.items() if v}, - } - - # Encoder - enc = ct.models.MLModel(str(output_dir / "parakeet_encoder.mlpackage"), compute_units=ct.ComputeUnit.CPU_AND_NE) - t0 = time.perf_counter() - enc_out = enc.predict({"mel": _np32(mel_ref), "mel_length": _np32(mel_length_ref).astype(np.int32)}) - t1 = time.perf_counter() - enc_first_ms = (t1 - t0) * 1000.0 - enc_ml = np.array(enc_out["encoder"], dtype=np.float32, copy=True) - enc_len_ml = np.array(enc_out["encoder_length"], dtype=np.int32, copy=True) - a_enc, r_enc, ok_enc = _max_diffs(_np32(encoder_ref), enc_ml, max(rtol, 5e-3), max(atol, 5e-2)) - ok_enc_len = int(_np32(encoder_length_ref).astype(np.int32)[0]) == int(np.array(enc_len_ml).astype(np.int32)[0]) - enc_t = _np32(encoder_ref)[0] - enc_c = enc_ml[0] - enc_plots = { - "encoder_time_l2.png": _plot_line( - np.linalg.norm(enc_t, axis=0), # L2 norm over features (D) for each time step - np.linalg.norm(enc_c, axis=0), # enc_t shape is (D, T), so axis=0 is features - "Encoder L2 over time", - plots_dir / "encoder_time_l2.png", - also_delta=True, - ), - } - enc_torch_ms_mean, enc_torch_ms_std = _time_torch( - asr_model.encoder, audio_signal=mel_ref, length=mel_length_ref.to(dtype=torch.long) - ) - enc_coreml_ms_mean, enc_coreml_ms_std = _time_coreml( - enc, {"mel": _np32(mel_ref), "mel_length": _np32(mel_length_ref).astype(np.int32)} - ) - enc_coreml_rtf = float(enc_coreml_ms_mean / (seconds * 1000.0)) if enc_coreml_ms_mean > 0 else None - enc_torch_rtf = float(enc_torch_ms_mean / (seconds * 1000.0)) if enc_torch_ms_mean > 0 else None - - summary["components"]["encoder"] = { - "encoder": {"max_abs": a_enc, "max_rel": r_enc, "match": bool(ok_enc)}, - "length_match": bool(ok_enc_len), - "latency": { - "runs": int(runs), - "warmup": int(warmup), - "coreml_first_ms": enc_first_ms, - "torch_ms": {"mean": enc_torch_ms_mean, "std": enc_torch_ms_std}, - "coreml_ms": {"mean": enc_coreml_ms_mean, "std": enc_coreml_ms_std}, - "rtf": {"torch": enc_torch_rtf, "coreml": enc_coreml_rtf}, - }, - "plots": {k: v for k, v in enc_plots.items() if v}, - } - - # Decoder (sequential U=1 rollout) - dec = ct.models.MLModel(str(output_dir / "parakeet_decoder.mlpackage"), compute_units=ct.ComputeUnit.CPU_AND_NE) - - zero_state_np = np.zeros((decoder_layers, 1, decoder_hidden), dtype=np.float32) - blank_targets_np = np.array(blank_targets.detach().cpu().numpy(), dtype=np.int32, copy=True) - blank_target_lengths_np = np.array(blank_target_lengths.detach().cpu().numpy(), dtype=np.int32, copy=True) - - def _decoder_rollout_coreml(num_steps: int) -> Tuple[np.ndarray, np.ndarray, np.ndarray, float]: - outputs = [] - h_np = zero_state_np.copy() - c_np = zero_state_np.copy() - first_ms: Optional[float] = None - for i in range(num_steps): - t0_i = time.perf_counter() if i == 0 else None - res = dec.predict( - { - "targets": blank_targets_np, - "target_length": blank_target_lengths_np, - "h_in": h_np, - "c_in": c_np, - } - ) - if t0_i is not None: - t1_i = time.perf_counter() - first_ms = (t1_i - t0_i) * 1000.0 - outputs.append(np.array(res["decoder"], dtype=np.float32, copy=True)) - h_np = np.array(res["h_out"], dtype=np.float32, copy=True) - c_np = np.array(res["c_out"], dtype=np.float32, copy=True) - if outputs: - decoder_seq = np.concatenate(outputs, axis=-1) - else: - decoder_seq = np.zeros((1, decoder_hidden, 0), dtype=np.float32) - return decoder_seq, h_np, c_np, (0.0 if first_ms is None else float(first_ms)) - - dec_ml, h_ml, c_ml, dec_first_ms = _decoder_rollout_coreml(symbol_steps) - h_ref_np = _np32(h_ref) - c_ref_np = _np32(c_ref) - - a_dec, r_dec, ok_dec = _max_diffs(decoder_ref_np, dec_ml, max(rtol, 1e-2), max(atol, 1e-1)) - a_h, r_h, ok_h = _max_diffs(h_ref_np, h_ml, max(rtol, 1e-2), max(atol, 2.5e-1)) - a_c, r_c, ok_c = _max_diffs(c_ref_np, c_ml, max(rtol, 5e-2), max(atol, 1.5e0)) - - dec_t = decoder_ref_np[0] - dec_c = dec_ml[0] - dec_plots = { - "decoder_steps_l2.png": _plot_line( - np.linalg.norm(dec_t, axis=0), - np.linalg.norm(dec_c, axis=0), - "Decoder L2 over steps", - plots_dir / "decoder_steps_l2.png", - also_delta=True, - ), - } - - def _time_decoder_coreml() -> Tuple[float, float]: - for _ in range(max(0, warmup)): - _decoder_rollout_coreml(symbol_steps) - times = [] - for _ in range(max(1, runs)): - t0 = time.perf_counter() - _decoder_rollout_coreml(symbol_steps) - t1 = time.perf_counter() - times.append((t1 - t0) * 1000.0) - arr = np.array(times, dtype=np.float64) - return float(arr.mean()), float(arr.std(ddof=1) if arr.size > 1 else 0.0) - - dec_torch_ms_mean, dec_torch_ms_std = _time_torch(lambda: _decoder_rollout_torch(symbol_steps)) - dec_coreml_ms_mean, dec_coreml_ms_std = _time_decoder_coreml() - dec_coreml_rtf = float(dec_coreml_ms_mean / (seconds * 1000.0)) if dec_coreml_ms_mean > 0 else None - dec_torch_rtf = float(dec_torch_ms_mean / (seconds * 1000.0)) if dec_torch_ms_mean > 0 else None - - summary["components"]["decoder"] = { - "decoder": {"max_abs": a_dec, "max_rel": r_dec, "match": bool(ok_dec)}, - "h_out": {"max_abs": a_h, "max_rel": r_h, "match": bool(ok_h)}, - "c_out": {"max_abs": a_c, "max_rel": r_c, "match": bool(ok_c)}, - "latency": { - "runs": int(runs), - "warmup": int(warmup), - "coreml_first_ms": dec_first_ms, - "torch_ms": {"mean": dec_torch_ms_mean, "std": dec_torch_ms_std}, - "coreml_ms": {"mean": dec_coreml_ms_mean, "std": dec_coreml_ms_std}, - "rtf": {"torch": dec_torch_rtf, "coreml": dec_coreml_rtf}, - }, - "plots": {k: v for k, v in dec_plots.items() if v}, - } - - # Joint (sequential U=1 rollouts) - j = ct.models.MLModel(str(output_dir / "parakeet_joint.mlpackage"), compute_units=ct.ComputeUnit.CPU_AND_NE) - - def _joint_rollout_coreml(decoder_seq_np: np.ndarray) -> Tuple[np.ndarray, float]: - logits_steps = [] - first_ms: Optional[float] = None - for u in range(decoder_seq_np.shape[2]): - dec_slice = decoder_seq_np[:, :, u : u + 1] - t0_u = time.perf_counter() if u == 0 else None - res = j.predict({"encoder": encoder_np, "decoder": dec_slice}) - if t0_u is not None: - t1_u = time.perf_counter() - first_ms = (t1_u - t0_u) * 1000.0 - logits_steps.append(np.array(res["logits"], dtype=np.float32, copy=True)) - if not logits_steps: - raise RuntimeError("No decoder steps provided for joint rollout") - return np.concatenate(logits_steps, axis=2), (0.0 if first_ms is None else float(first_ms)) - - logits_ml, joint_first_ms = _joint_rollout_coreml(decoder_ref_np) - logits_ref_np = _np32(logits_ref) - a_j, r_j, ok_j = _max_diffs(logits_ref_np, logits_ml, max(rtol, 1e-2), max(atol, 1e-1)) - joint_plots = {} - if HAS_MPL: - lt = logits_ref_np[0, 0, 0, :] - lc = logits_ml[0, 0, 0, :] - top_idx = np.argsort(-np.abs(lt))[:50] - path = plots_dir / "joint_top50.png" - plt.figure(figsize=(8, 3)) - plt.plot(lt[top_idx], label="torch") - plt.plot(lc[top_idx], label="coreml", alpha=0.8) - plt.title("Joint logits (t=0,u=0) top-50 |torch|") - plt.legend(); plt.tight_layout(); plt.savefig(path); plt.close() - joint_plots["joint_top50.png"] = str(path.name) - - # Delta-over-time visualization (fix u=0; summarize over vocab) - jt = logits_ref_np[0, :, 0, :] - jc = logits_ml[0, :, 0, :] - l2_t = np.linalg.norm(jt, axis=1) - l2_c = np.linalg.norm(jc, axis=1) - path2 = plots_dir / "joint_time_l2.png" - _plot_line(l2_t, l2_c, "Joint L2 over time (u=0)", path2, also_delta=True) - joint_plots["joint_time_l2.png"] = str(path2.name) - - def _time_joint_coreml() -> Tuple[float, float]: - for _ in range(max(0, warmup)): - _joint_rollout_coreml(decoder_ref_np) - times = [] - for _ in range(max(1, runs)): - t0 = time.perf_counter() - _joint_rollout_coreml(decoder_ref_np) - t1 = time.perf_counter() - times.append((t1 - t0) * 1000.0) - arr = np.array(times, dtype=np.float64) - return float(arr.mean()), float(arr.std(ddof=1) if arr.size > 1 else 0.0) - - joint_torch_ms_mean, joint_torch_ms_std = _time_torch( - asr_model.joint, encoder_outputs=encoder_ref, decoder_outputs=decoder_ref - ) - joint_coreml_ms_mean, joint_coreml_ms_std = _time_joint_coreml() - joint_coreml_rtf = float(joint_coreml_ms_mean / (seconds * 1000.0)) if joint_coreml_ms_mean > 0 else None - joint_torch_rtf = float(joint_torch_ms_mean / (seconds * 1000.0)) if joint_torch_ms_mean > 0 else None - - summary["components"]["joint"] = { - "logits": {"max_abs": a_j, "max_rel": r_j, "match": bool(ok_j)}, - "latency": { - "runs": int(runs), - "warmup": int(warmup), - "coreml_first_ms": joint_first_ms, - "torch_ms": {"mean": joint_torch_ms_mean, "std": joint_torch_ms_std}, - "coreml_ms": {"mean": joint_coreml_ms_mean, "std": joint_coreml_ms_std}, - "rtf": {"torch": joint_torch_rtf, "coreml": joint_coreml_rtf}, - }, - "plots": joint_plots, - } - - # Fused components - # 1) Mel+Encoder fused vs separate - mel_enc_plots = {} - try: - mel_enc = ct.models.MLModel(str(output_dir / "parakeet_mel_encoder.mlpackage"), compute_units=ct.ComputeUnit.CPU_AND_NE) - t0 = time.perf_counter() - mel_enc_out = mel_enc.predict({ - "audio_signal": _np32(audio_tensor), - "audio_length": _np32(audio_length).astype(np.int32), - }) - t1 = time.perf_counter() - mel_enc_first_ms = (t1 - t0) * 1000.0 - mel_enc_ml = np.array(mel_enc_out["encoder"], dtype=np.float32, copy=True) - mel_enc_len_ml = np.array(mel_enc_out["encoder_length"], dtype=np.int32, copy=True) - # Compare fused output vs Torch reference encoder - a_melenc, r_melenc, ok_melenc = _max_diffs(_np32(encoder_ref), mel_enc_ml, max(rtol, 5e-3), max(atol, 5e-2)) - ok_melenc_len = int(_np32(encoder_length_ref).astype(np.int32)[0]) == int(mel_enc_len_ml.astype(np.int32)[0]) - # Also compare fused vs separate CoreML pipeline (pre -> enc) - a_melenc_vs_sep, r_melenc_vs_sep, ok_melenc_vs_sep = _max_diffs(enc_ml, mel_enc_ml, max(rtol, 5e-3), max(atol, 5e-2)) - - # Plots: L2 over time (fused vs torch) - enc_t_ref = _np32(encoder_ref)[0] - enc_c_fused = mel_enc_ml[0] - mel_enc_plots["mel_encoder_time_l2.png"] = _plot_line( - np.linalg.norm(enc_t_ref, axis=0), - np.linalg.norm(enc_c_fused, axis=0), - "Mel+Encoder (fused) L2 over time", - plots_dir / "mel_encoder_time_l2.png", - also_delta=True, - ) - - # Latency: fused CoreML vs separate (CoreML pre + CoreML enc) - mel_enc_coreml_ms_mean, mel_enc_coreml_ms_std = _time_coreml( - mel_enc, - {"audio_signal": _np32(audio_tensor), "audio_length": _np32(audio_length).astype(np.int32)}, - ) - sep_coreml_ms_mean = float(pre_coreml_ms_mean + enc_coreml_ms_mean) - sep_coreml_ms_std = float((pre_coreml_ms_std ** 2 + enc_coreml_ms_std ** 2) ** 0.5) - # Torch baseline (separate torch pre + enc) - sep_torch_ms_mean = float(pre_torch_ms_mean + enc_torch_ms_mean) - sep_torch_ms_std = float((pre_torch_ms_std ** 2 + enc_torch_ms_std ** 2) ** 0.5) - - mel_enc_coreml_rtf = float(mel_enc_coreml_ms_mean / (seconds * 1000.0)) if mel_enc_coreml_ms_mean > 0 else None - sep_coreml_rtf = float(sep_coreml_ms_mean / (seconds * 1000.0)) if sep_coreml_ms_mean > 0 else None - sep_torch_rtf = float(sep_torch_ms_mean / (seconds * 1000.0)) if sep_torch_ms_mean > 0 else None - - summary["components"]["mel_encoder"] = { - "encoder": {"max_abs": a_melenc, "max_rel": r_melenc, "match": bool(ok_melenc)}, - "length_match": bool(ok_melenc_len), - "vs_separate_coreml": {"max_abs": a_melenc_vs_sep, "max_rel": r_melenc_vs_sep, "match": bool(ok_melenc_vs_sep)}, - "latency": { - "runs": int(runs), - "warmup": int(warmup), - "fused_coreml_first_ms": mel_enc_first_ms, - "fused_coreml_ms": {"mean": mel_enc_coreml_ms_mean, "std": mel_enc_coreml_ms_std}, - "separate_coreml_ms": {"mean": sep_coreml_ms_mean, "std": sep_coreml_ms_std}, - "separate_torch_ms": {"mean": sep_torch_ms_mean, "std": sep_torch_ms_std}, - "rtf": {"fused_coreml": mel_enc_coreml_rtf, "separate_coreml": sep_coreml_rtf, "separate_torch": sep_torch_rtf}, - }, - "plots": {k: v for k, v in mel_enc_plots.items() if v}, - } - except Exception as e: - summary["components"]["mel_encoder_error"] = str(e) - - # 2) JointDecision fused vs CPU PyTorch post-processing - jd_plots = {} - try: - # Fused CoreML joint decision - jd = ct.models.MLModel(str(output_dir / "parakeet_joint_decision.mlpackage"), compute_units=ct.ComputeUnit.CPU_AND_NE) - def _joint_decision_rollout_coreml(decoder_seq_np: np.ndarray) -> Tuple[np.ndarray, np.ndarray, np.ndarray, float]: - token_ids = [] - token_probs = [] - durations = [] - first_ms: Optional[float] = None - for u in range(decoder_seq_np.shape[2]): - dec_slice = decoder_seq_np[:, :, u : u + 1] - t0_u = time.perf_counter() if u == 0 else None - res = jd.predict({"encoder": encoder_np, "decoder": dec_slice}) - if t0_u is not None: - t1_u = time.perf_counter() - first_ms = (t1_u - t0_u) * 1000.0 - token_ids.append(np.array(res["token_id"], dtype=np.int32, copy=True)) - token_probs.append(np.array(res["token_prob"], dtype=np.float32, copy=True)) - durations.append(np.array(res["duration"], dtype=np.int32, copy=True)) - if not token_ids: - raise RuntimeError("No decoder steps provided for joint decision rollout") - return ( - np.concatenate(token_ids, axis=2), - np.concatenate(token_probs, axis=2), - np.concatenate(durations, axis=2), - (0.0 if first_ms is None else float(first_ms)), - ) - - token_id_ml, token_prob_ml, duration_ml, jd_first_ms = _joint_decision_rollout_coreml(decoder_ref_np) - - # CPU PyTorch decision using Torch logits - vocab_with_blank = int(vocab_size) + 1 - with torch.inference_mode(): - logits_t = logits_ref - token_logits_t = logits_t[..., :vocab_with_blank] - duration_logits_t = logits_t[..., -num_extra:] if num_extra > 0 else None - token_ids_t = torch.argmax(token_logits_t, dim=-1).to(dtype=torch.int32) - token_probs_all_t = torch.softmax(token_logits_t, dim=-1) - token_prob_t = torch.gather( - token_probs_all_t, dim=-1, index=token_ids_t.long().unsqueeze(-1) - ).squeeze(-1) - if duration_logits_t is not None and duration_logits_t.numel() > 0: - duration_t = torch.argmax(duration_logits_t, dim=-1).to(dtype=torch.int32) - else: - duration_t = torch.zeros_like(token_ids_t, dtype=torch.int32) - - # Also derive CPU decision from CoreML joint logits for "separate" path - token_logits_c = _to_t(logits_ml)[..., :vocab_with_blank] - duration_logits_c = _to_t(logits_ml)[..., -num_extra:] if num_extra > 0 else None - token_ids_c = torch.argmax(token_logits_c, dim=-1).to(dtype=torch.int32) - token_probs_all_c = torch.softmax(token_logits_c, dim=-1) - token_prob_c = torch.gather( - token_probs_all_c, dim=-1, index=token_ids_c.long().unsqueeze(-1) - ).squeeze(-1) - if duration_logits_c is not None and duration_logits_c.numel() > 0: - duration_c = torch.argmax(duration_logits_c, dim=-1).to(dtype=torch.int32) - else: - duration_c = torch.zeros_like(token_ids_c, dtype=torch.int32) - - # Compare fused outputs to CPU PyTorch decisions - a_tid_t, r_tid_t, ok_tid_t = _max_diffs(_np(token_ids_t), token_id_ml, 0.0, 0.0) - a_tprob_t, r_tprob_t, ok_tprob_t = _max_diffs(_np(token_prob_t), token_prob_ml, max(rtol, 1e-2), max(atol, 1e-1)) - a_dur_t, r_dur_t, ok_dur_t = _max_diffs(_np(duration_t), duration_ml, 0.0, 0.0) - - a_tid_c, r_tid_c, ok_tid_c = _max_diffs(_np(token_ids_c), token_id_ml, 0.0, 0.0) - a_tprob_c, r_tprob_c, ok_tprob_c = _max_diffs(_np(token_prob_c), token_prob_ml, max(rtol, 1e-2), max(atol, 1e-1)) - a_dur_c, r_dur_c, ok_dur_c = _max_diffs(_np(duration_c), duration_ml, 0.0, 0.0) - - # Plots: token_prob over time for u=0 (fused vs torch CPU) - if HAS_MPL: - prob_t = _np(token_prob_t)[0, :, 0] - prob_ml = token_prob_ml[0, :, 0] - jd_plots["joint_decision_prob_u0.png"] = _plot_line( - prob_t, - prob_ml, - "JointDecision token_prob (u=0)", - plots_dir / "joint_decision_prob_u0.png", - also_delta=True, - ) - - # Agreement heatmap for token_id - agree = (_np(token_ids_t)[0] == token_id_ml[0]).astype(np.float32) - jd_plots["joint_decision_token_agree.png"] = _plot_image( - agree, - "token_id agreement (torch CPU vs fused)", - plots_dir / "joint_decision_token_agree.png", - vmin=0.0, - vmax=1.0, - ) - - # Latency: fused CoreML vs separate (CoreML joint + CPU PyTorch decision) - def _time_joint_decision_coreml() -> Tuple[float, float]: - for _ in range(max(0, warmup)): - _joint_decision_rollout_coreml(decoder_ref_np) - times = [] - for _ in range(max(1, runs)): - t0 = time.perf_counter() - _joint_decision_rollout_coreml(decoder_ref_np) - t1 = time.perf_counter() - times.append((t1 - t0) * 1000.0) - arr = np.array(times, dtype=np.float64) - return float(arr.mean()), float(arr.std(ddof=1) if arr.size > 1 else 0.0) - - jd_coreml_ms_mean, jd_coreml_ms_std = _time_joint_decision_coreml() - - # Time CPU post-processing only (Torch) on top of CoreML or Torch logits. Use Torch logits. - def _decision_torch_call(): - with torch.inference_mode(): - tl = logits_ref - tl_token = tl[..., :vocab_with_blank] - tl_ids = torch.argmax(tl_token, dim=-1) - tl_probs = torch.softmax(tl_token, dim=-1) - _ = torch.gather(tl_probs, -1, tl_ids.long().unsqueeze(-1)).squeeze(-1) - if num_extra > 0: - _ = torch.argmax(tl[..., -num_extra:], dim=-1) - return None - - jd_decision_torch_ms_mean, jd_decision_torch_ms_std = _time_torch(lambda: _decision_torch_call()) - sep_joint_plus_cpu_ms_mean = float(joint_coreml_ms_mean + jd_decision_torch_ms_mean) - sep_joint_plus_cpu_ms_std = float((joint_coreml_ms_std ** 2 + jd_decision_torch_ms_std ** 2) ** 0.5) - jd_coreml_rtf = float(jd_coreml_ms_mean / (seconds * 1000.0)) if jd_coreml_ms_mean > 0 else None - sep_joint_cpu_rtf = float(sep_joint_plus_cpu_ms_mean / (seconds * 1000.0)) if sep_joint_plus_cpu_ms_mean > 0 else None - - summary["components"]["joint_decision"] = { - "vs_torch_cpu": { - "token_id": {"max_abs": a_tid_t, "max_rel": r_tid_t, "match": bool(ok_tid_t)}, - "token_prob": {"max_abs": a_tprob_t, "max_rel": r_tprob_t, "match": bool(ok_tprob_t)}, - "duration": {"max_abs": a_dur_t, "max_rel": r_dur_t, "match": bool(ok_dur_t)}, - }, - "vs_coreml_joint_cpu": { - "token_id": {"max_abs": a_tid_c, "max_rel": r_tid_c, "match": bool(ok_tid_c)}, - "token_prob": {"max_abs": a_tprob_c, "max_rel": r_tprob_c, "match": bool(ok_tprob_c)}, - "duration": {"max_abs": a_dur_c, "max_rel": r_dur_c, "match": bool(ok_dur_c)}, - }, - "latency": { - "runs": int(runs), - "warmup": int(warmup), - "fused_coreml_first_ms": jd_first_ms, - "fused_coreml_ms": {"mean": jd_coreml_ms_mean, "std": jd_coreml_ms_std}, - "separate_joint_coreml_plus_cpu_ms": {"mean": sep_joint_plus_cpu_ms_mean, "std": sep_joint_plus_cpu_ms_std}, - "rtf": {"fused_coreml": jd_coreml_rtf, "separate_joint_coreml_plus_cpu": sep_joint_cpu_rtf}, - }, - "plots": {k: v for k, v in jd_plots.items() if v}, - } - except Exception as e: - summary["components"]["joint_decision_error"] = str(e) - - # Latency overview plots (saved alongside component plots) - latency_plots = {} - labels = ["preprocessor", "encoder", "decoder", "joint"] - torch_means = [pre_torch_ms_mean, enc_torch_ms_mean, dec_torch_ms_mean, joint_torch_ms_mean] - torch_stds = [pre_torch_ms_std, enc_torch_ms_std, dec_torch_ms_std, joint_torch_ms_std] - coreml_means = [pre_coreml_ms_mean, enc_coreml_ms_mean, dec_coreml_ms_mean, joint_coreml_ms_mean] - coreml_stds = [pre_coreml_ms_std, enc_coreml_ms_std, dec_coreml_ms_std, joint_coreml_ms_std] - lat_path = plots_dir / "latency_summary.png" - spd_path = plots_dir / "latency_speedup.png" - latency_plots["latency_summary.png"] = _plot_latency_bars( - labels, torch_means, torch_stds, coreml_means, coreml_stds, lat_path - ) - latency_plots["latency_speedup.png"] = _plot_speedup_bars( - labels, torch_means, coreml_means, spd_path - ) - - # Fused vs separate latency summary - fused_labels = ["mel+encoder", "joint_decision"] - fused_baseline_means = [ - float(pre_torch_ms_mean + enc_torch_ms_mean), - float(joint_coreml_ms_mean + jd_decision_torch_ms_mean if 'jd_coreml_ms_mean' in locals() else joint_coreml_ms_mean), - ] - fused_coreml_means = [ - float(mel_enc_coreml_ms_mean if 'mel_enc_coreml_ms_mean' in locals() else np.nan), - float(jd_coreml_ms_mean if 'jd_coreml_ms_mean' in locals() else np.nan), - ] - fused_latency_path = plots_dir / "latency_fused_vs_separate.png" - fused_speedup_path = plots_dir / "latency_fused_speedup.png" - latency_plots["latency_fused_vs_separate.png"] = _plot_latency_bars( - fused_labels, fused_baseline_means, [0, 0], fused_coreml_means, [0, 0], fused_latency_path - ) - latency_plots["latency_fused_speedup.png"] = _plot_speedup_bars( - fused_labels, fused_baseline_means, fused_coreml_means, fused_speedup_path - ) - - all_ok = ( - summary["components"]["preprocessor"]["mel"]["match"] - and summary["components"]["preprocessor"]["length_match"] - and summary["components"]["encoder"]["encoder"]["match"] - and summary["components"]["encoder"]["length_match"] - and summary["components"]["decoder"]["decoder"]["match"] - and summary["components"]["decoder"]["h_out"]["match"] - and summary["components"]["decoder"]["c_out"]["match"] - and summary["components"]["joint"]["logits"]["match"] - ) - summary["status"] = "ok" if all_ok else "mismatch" - - # Update metadata.json - meta_path = output_dir / "metadata.json" - try: - meta = json.loads(meta_path.read_text()) - except Exception: - meta = {} - meta["validation"] = summary - meta_path.write_text(json.dumps(meta, indent=2)) - - typer.echo(f"Validation {'passed' if all_ok else 'mismatched'}. Updated {meta_path}") - if HAS_MPL: - typer.echo(f"Saved plots to {plots_dir}") - - -if __name__ == "__main__": - app() diff --git a/parakeet-tdt-v2-0.6b/coreml/compile_modelc.py b/parakeet-tdt-v2-0.6b/coreml/compile_modelc.py deleted file mode 100644 index 9695e03b41fce9280f2ca2ee0e9585c26b2393cc..0000000000000000000000000000000000000000 --- a/parakeet-tdt-v2-0.6b/coreml/compile_modelc.py +++ /dev/null @@ -1,91 +0,0 @@ -#!/usr/bin/env python3 -"""Compile Core ML packages into ``.mlmodelc`` bundles via ``xcrun``. - -This script walks through the default Parakeet CoreML directories, finds -all ``*.mlpackage`` bundles, and compiles each of them with -``xcrun coremlcompiler`` into ``./compiled`` while preserving the -relative directory structure. -""" -from __future__ import annotations - -import shutil -import subprocess -import sys -from pathlib import Path - -BASE_DIR = Path(__file__).resolve().parent -OUTPUT_ROOT = BASE_DIR / "compiled" -SOURCE_DIRS = [BASE_DIR / "parakeet_coreml", BASE_DIR / "parakeet_coreml_quantized"] - - -def ensure_coremlcompiler() -> None: - """Ensure ``xcrun coremlcompiler`` is available for the active Xcode.""" - xcrun_path = shutil.which("xcrun") - if xcrun_path is None: - print("Error: 'xcrun' not found on PATH. Install Xcode command line tools.", file=sys.stderr) - sys.exit(1) - - try: - subprocess.run([ - xcrun_path, - "--find", - "coremlcompiler", - ], check=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE) - except subprocess.CalledProcessError: - print("Error: 'coremlcompiler' not found via xcrun. Check your Xcode installation.", file=sys.stderr) - sys.exit(1) - - -def gather_packages() -> list[Path]: - """Return a list of all ``*.mlpackage`` bundles under the source dirs.""" - packages: list[Path] = [] - for source in SOURCE_DIRS: - if not source.exists(): - print(f"Warning: {source.relative_to(BASE_DIR)} does not exist; skipping", file=sys.stderr) - continue - packages.extend(source.rglob("*.mlpackage")) - return packages - - -def compile_package(package: Path) -> None: - """Compile a single ``.mlpackage`` bundle using ``xcrun coremlcompiler``.""" - relative_pkg = package.relative_to(BASE_DIR) - output_dir = OUTPUT_ROOT / relative_pkg.parent - output_dir.mkdir(parents=True, exist_ok=True) - output_path = output_dir / f"{package.stem}.mlmodelc" - - if output_path.exists(): - shutil.rmtree(output_path) - - cmd = [ - "xcrun", - "coremlcompiler", - "compile", - str(package), - str(output_dir), - ] - - print(f"Compiling {relative_pkg} -> {output_path.relative_to(BASE_DIR)}") - subprocess.run(cmd, check=True) - - -def main() -> None: - ensure_coremlcompiler() - packages = gather_packages() - - if not packages: - print("No .mlpackage bundles found to compile.") - return - - for package in packages: - try: - compile_package(package) - except subprocess.CalledProcessError as exc: - print(f"Failed to compile {package}: {exc}", file=sys.stderr) - sys.exit(exc.returncode) - - print(f"Finished compiling {len(packages)} package(s) into {OUTPUT_ROOT.relative_to(BASE_DIR)}.") - - -if __name__ == "__main__": - main() diff --git a/parakeet-tdt-v2-0.6b/coreml/context/Efficient Sequence Transduction by Jointly Predicting Tokens and Durations.pdf b/parakeet-tdt-v2-0.6b/coreml/context/Efficient Sequence Transduction by Jointly Predicting Tokens and Durations.pdf deleted file mode 100644 index b3fdd0192810231ef4b0ac61b608a7e06d9d4f6c..0000000000000000000000000000000000000000 --- a/parakeet-tdt-v2-0.6b/coreml/context/Efficient Sequence Transduction by Jointly Predicting Tokens and Durations.pdf +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:a29e637f2f977a44bece33863aeb01b45b0fc73a9bb723aea2960986e41c5bba -size 3071390 diff --git a/parakeet-tdt-v2-0.6b/coreml/context/FAST CONFORMER WITH LINEARLY SCALABLE ATTENTION.pdf b/parakeet-tdt-v2-0.6b/coreml/context/FAST CONFORMER WITH LINEARLY SCALABLE ATTENTION.pdf deleted file mode 100644 index d0aef3ea5d69a76e18bbd94161a522246ccfd1fc..0000000000000000000000000000000000000000 --- a/parakeet-tdt-v2-0.6b/coreml/context/FAST CONFORMER WITH LINEARLY SCALABLE ATTENTION.pdf +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:eb8c7706a567ec5795333139f8cd41ff9c127ba99b6dc85d378eb3fa3213bf8d -size 586484 diff --git a/parakeet-tdt-v2-0.6b/coreml/context/coreml_component_io.md b/parakeet-tdt-v2-0.6b/coreml/context/coreml_component_io.md deleted file mode 100644 index 85a34b5927f4e7e98a4c672ebc7cc249992c3100..0000000000000000000000000000000000000000 --- a/parakeet-tdt-v2-0.6b/coreml/context/coreml_component_io.md +++ /dev/null @@ -1,96 +0,0 @@ -# CoreML Component I/O (Parakeet‑TDT‑0.6B‑v2) - -This file documents the CoreML I/O contracts for each exported sub‑module. All shapes below target the fixed 15‑second window at 16 kHz (240,000 samples). Batch is fixed to 1 for on-device use, and we do not plan to support other window sizes in CoreML (re-export if the requirement changes). - -## Conventions - -- Dtypes: `float32` for real‑valued tensors, `int32` for lengths and token IDs. -- Batch: `B=1` everywhere. -- Vocab: `V=8192` BPE tokens; RNNT adds blank (+1). TDT adds 5 duration logits appended to the class axis. -- Subsampling: overall encoder time reduction ×8. - -## PreprocessorWrapper - -- Inputs - - `audio_signal`: `[1, 240000]` float32, PCM range ~[-1, 1] - - `length`: `[1]` int32, number of valid samples (<= 240000) -- Outputs - - `mel`: `[1, 128, 1501]` float32 (center=True => frames = floor(240000/160)+1 = 1501) - - `mel_length`: `[1]` int32 (=1501 for full 15 s) - -## EncoderWrapper - -- Inputs - - `features`: `[1, 128, 1501]` float32 (mel) - - `length`: `[1]` int32 (=1501) -- Outputs - - `encoded`: `[1, 188, 1024]` float32 (wrapper has time‑major last: it transposes `[B,D,T] -> [B,T,D]`) - - `encoded_length`: `[1]` int32 (`ceil(1501/8)=188`) - -## DecoderWrapper (stateful LSTM, 2 layers) - -- Inputs - - `targets`: `[1, U]` int32 (token IDs in [0, V)) - - `target_lengths`: `[1]` int32 (=U) - - `h_in`: `[2, 1, 640]` float32 (LSTM hidden, L=2 layers) - - `c_in`: `[2, 1, 640]` float32 (LSTM cell, L=2 layers) -- Outputs - - `decoder_output`: `[1, U, 640]` float32 - - `h_out`: `[2, 1, 640]` float32 - - `c_out`: `[2, 1, 640]` float32 - -Notes - -- For step‑wise greedy decoding, set `U=1` and feed back `h_out`, `c_out`. -- For batched token scoring, `U` can be >1 to evaluate multiple symbols per call. - -## JointWrapper - -- Inputs - - `encoder_outputs`: `[1, 188, 1024]` float32 (time‑major) - - `decoder_outputs`: `[1, U, 640]` float32 -- Output - - `logits`: `[1, 188, U, 8192 + 1 + 5]` float32 - -Notes - -- Split the last dimension into `[token_logits: V+1]` and `[duration_logits: 5]` when post‑processing. -- `log_softmax` is disabled for export; apply on CPU if needed. - -## MelEncoderWrapper (fused) - -- Inputs - - `audio_signal`: `[1, 240000]` float32 - - `audio_length`: `[1]` int32 -- Outputs - - `encoder`: `[1, 188, 1024]` float32 - - `encoder_length`: `[1]` int32 - -Notes - -- Fuses preprocessor + encoder to avoid a mel round‑trip. Shapes match chaining `PreprocessorWrapper` then `EncoderWrapper` on the fixed 15 s window. - -## JointDecisionWrapper (fused post‑processing) - -- Inputs - - `encoder`: `[1, 188, 1024]` float32 - - `decoder`: `[1, U, 640]` float32 -- Outputs - - `token_id`: `[1, 188, U]` int32 (argmax over token logits, includes blank) - - `token_prob`: `[1, 188, U]` float32 (softmax probability of chosen token) - - `duration`: `[1, 188, U]` int32 (argmax over 5 duration logits) - -Notes - -- Embeds the common host‑side steps: split joint logits, softmax over token logits, argmax over both heads, and gather chosen token probability. Replaces separate `runJoint` → `splitLogits` → `softmax/argmax` in Swift. - -## Suggested CoreML tensor names - -- Preprocessor: `audio_signal`, `audio_length` → `mel`, `mel_length` -- Encoder: `mel`, `mel_length` → `encoder`, `encoder_length` -- Decoder: `targets`, `target_length`, `h_in`, `c_in` → `decoder`, `h_out`, `c_out` -- Joint: `encoder`, `decoder` → `logits` -- MelEncoder (fused): `audio_signal`, `audio_length` → `encoder`, `encoder_length` -- JointDecision (fused): `encoder`, `decoder` → `token_id`, `token_prob`, `duration` - -These names align with wrappers in `individual_components.py` and simplify downstream wiring. diff --git a/parakeet-tdt-v2-0.6b/coreml/context/coreml_conversion_plan.md b/parakeet-tdt-v2-0.6b/coreml/context/coreml_conversion_plan.md deleted file mode 100644 index cf3cd4c09dbd02b882cca3f37f0802f7bfa569b1..0000000000000000000000000000000000000000 --- a/parakeet-tdt-v2-0.6b/coreml/context/coreml_conversion_plan.md +++ /dev/null @@ -1,65 +0,0 @@ -# CoreML Conversion Plan (Parakeet‑TDT‑0.6B‑v2) - -This plan describes how we export Parakeet’s sub‑modules to CoreML, validate numerics, and prepare for on‑device decoding. The pipeline keeps a fixed 15‑second audio window. - -## Goals - -- Export preprocessor, encoder, decoder, and joint as separate mlpackages. -- Also export fused variants: `mel_encoder` (preprocessor+encoder) and `joint_decision` (joint + split/softmax/argmax). -- Preserve streaming decoder state I/O for on‑device greedy decoding. -- Validate component outputs against the NeMo reference on a 15‑second clip. - -## Environment - -- Use `uv venv` and run via `uv run` to ensure reproducible resolutions. -- Python 3.10.12, `torch 2.5.0`, `coremltools 8.3.0`, `nemo-toolkit 2.3.1`. - -## Export settings - -- `convert_to = "mlprogram"` -- `minimum_deployment_target` = `ct.target.iOS17` (we only target iOS 17+ for the CoreML deployment) -- `compute_units` = `.CPU_AND_GPU` (or `.ALL` on iOS) -- `compute_precision` = `ct.precision.FLOAT16` or `None` (start with fp32 for validation, then try fp16) -- Fixed window: 15 seconds at 16 kHz → shapes as in `context/coreml_component_io.md`. No variable-length support is required; CoreML graphs can assume the 15 s window. - -## Steps - -- Load the NeMo checkpoint with `ASRModel.from_pretrained("nvidia/parakeet-tdt-0.6b-v2")`. -- Extract modules: preprocessor, encoder, decoder, joint. -- Wrap modules with `PreprocessorWrapper`, `EncoderWrapper`, `DecoderWrapper`, `JointWrapper`; derive `MelEncoderWrapper` and `JointDecisionWrapper` for fused exports. -- Trace each wrapper with static 15‑second inputs (batch=1). Ensure outputs match I/O contracts in `coreml_component_io.md`. -- Define CoreML inputs with explicit names and fixed shapes (15 s audio → mel 1501 → encoder 188). Keeping the window fixed simplifies CoreML deployment; re‑export if window size changes. -- Convert via `ct.convert(...)` with the settings above. -- Save each mlpackage under the chosen `output_dir`, along with a `metadata.json` capturing shapes, sample rate, vocab size, tokenizer path, and export metadata (author, conversion commit, etc.). -- Emit fused graphs (`parakeet_mel_encoder.mlpackage` and `parakeet_joint_decision.mlpackage`) alongside standalone components to reduce host I/O and post‑processing. -- Provide a helper like `compile_modelc.py` to batch `xcrun coremlcompiler` invocations when packaging for release. - -## CLI layout - -- `convert_parakeet_components.py` → conversion-only entry point (no validation), parameterized by compute units, precision, and optional fused exports. Model inspection (encoder strides, etc.) should auto-populate config, inspired by Silero VAD’s converter. -- `compare_parakeet_models.py` → validation/plotting script that loads reference NeMo modules and CoreML packages, runs diffs per component, and reports max/mean errors plus RTF metrics. Use lightweight wrappers similar to Silero’s `PyTorchJITWrapper` / `CoreMLVADWrapper`. -- Future helpers: `compile_modelc.py` (reuse pattern from VAD project) and optional quantization experiments once parity is established. - -## Validation - -- Run inference on a known 16 kHz clip (trim/pad to 15 s): - - Compare preprocessor mel and lengths (max abs/rel diff thresholds: atol=1e-4, rtol=1e-3). - - Compare encoder outputs and lengths. - - Compare decoder outputs for a fixed target sequence and initial zero states. - - Compare joint logits on the same `(T_enc, U)` grid; split `[V+1]` token logits from the last 5 duration logits when computing metrics. -- Record per‑component diffs in `metadata.json` for auditability. - -## Decoding (device) - -- Implement greedy RNNT in app code calling CoreML: - - Either: use modular `joint` and perform split/softmax/argmax on host. - - Or: call fused `joint_decision` to receive `token_id`, `token_prob`, and `duration` directly. - - Preprocess → Encode once per window (or call fused `mel_encoder`). - - Maintain `(h, c)` across symbol steps. - - Blank handling: `token_id == V` indicates blank. - -## Known caveats - -- RNNT joint logits are large: `[188 × U × ~8200]` per window; consider fp16 and/or tiling over `U` to reduce memory. -- Length maths: mel frames use `center=True`, yielding 1501 frames for exactly 240,000 samples; encoder length computed via ceil divide by 8. -- For accurate timestamping, use the original NeMo decoder on server to validate any device‑side greedy implementation. diff --git a/parakeet-tdt-v2-0.6b/coreml/context/hf_model_card_notes.md b/parakeet-tdt-v2-0.6b/coreml/context/hf_model_card_notes.md deleted file mode 100644 index 082919ab08fdfb6067b8537d0af73b3c5001ff5a..0000000000000000000000000000000000000000 --- a/parakeet-tdt-v2-0.6b/coreml/context/hf_model_card_notes.md +++ /dev/null @@ -1,24 +0,0 @@ -# Hugging Face Model Card Notes: nvidia/parakeet-tdt-0.6b-v2 - -Key details extracted from the public model card for quick reference. - -## Summary -- Task: Multilingual automatic speech recognition (ASR) -- Architecture: FastConformer encoder + RNNT + TDT -- Params: ~0.6B -- License: CC‑BY‑4.0 - -## Languages (25 EU + Russian/Ukrainian) -bg, hr, cs, da, nl, et, fi, fr, de, el, hu, it, lv, lt, mt, pl, pt, ro, sk, sl, sv, es, en, ru, uk - -## Input / Output -- Input: 16 kHz mono audio (`.wav`, `.flac`) -- Output: text with punctuation and capitalization - -## Notes relevant to conversion -- The model card reports accurate word/segment timestamps; Parakeet TDT uses duration buckets which we expose from the joint head (last 5 logits along the class axis). -- Long‑audio inference is reported for GPU; our CoreML export intentionally uses a fixed 15‑second window to fit on‑device constraints. - -References -- HF repo: https://huggingface.co/nvidia/parakeet-tdt-0.6b-v2 -- NeMo FastConformer: https://docs.nvidia.com/deeplearning/nemo/user-guide/docs/en/main/asr/models.html#fast-conformer diff --git a/parakeet-tdt-v2-0.6b/coreml/context/mel_encoder_ane_behavior.md b/parakeet-tdt-v2-0.6b/coreml/context/mel_encoder_ane_behavior.md deleted file mode 100644 index e55cf015d7b17e89d1ce1dedeb54512a056cf5ae..0000000000000000000000000000000000000000 --- a/parakeet-tdt-v2-0.6b/coreml/context/mel_encoder_ane_behavior.md +++ /dev/null @@ -1,67 +0,0 @@ -# MelEncoder ANE behavior and iPhone 13 issue - -Summary of the failure seen on iPhone 13 (A15 / E5): - -> ANE model load has failed for on-device compiled macho. Must re-compile the E5 bundle. Invalid layer: Tensor dimensions N1D1C1H1W240000 are not within supported range, N[1-65536] D[1-16384] C[1-65536] H[1-16384] W[1-16384]. - -## Root cause - -- The fused MelEncoder model takes raw waveform input shaped `[1, 240000]` for the fixed 15 s window (16 kHz × 15 s). -- In Core ML’s internal layout this maps to `N1 D1 C1 H1 W240000`. -- Apple Neural Engine (ANE) enforces a per-dimension cap of ≀ 16384 for H and W. `W=240000` violates this, so any ANE partition that “sees” the waveform will fail validation on A15 (and other iPhone ANE generations). -- The failure is caused by the preprocessor stage inside the fused MelEncoder (waveform → STFT/mel). The standalone encoder (mel → encoder) uses inputs around `[1, 128, 1501]` and is ANE-safe. - -## Why earlier models “worked” - -- Prior exports either: - - Avoided fusing the waveform preprocessor with the encoder; or - - Exported a “mel→encoder” model only; or - - Relied on Core ML partitioning that kept the preprocessor on CPU/GPU while the encoder ran on ANE. -- Even with `computeUnits = .cpuAndNeuralEngine`, Core ML will fall back to CPU for non-ANE-eligible subgraphs. With split models, the preprocessor silently ran on CPU and the encoder on ANE. -- With the new fused MelEncoder, the compiler tries to start an ANE partition earlier (due to the encoder being ANE-friendly). That exposes the 240000-wide waveform to the ANE plan and triggers the dimension error. - -## Device differences (M‑series Macs vs iPhones) - -- Inputs are identical across devices; what differs is how Core ML partitions the graph. -- On M‑series Macs, the waveform preprocessor commonly stays on CPU/GPU; the ANE (if present) or GPU is used for the encoder. No ANE error is logged. -- On iPhones (A14/A15/A16/A17), the ANE compiler (E-series) may attempt to include earlier elementwise ops, making the NE subgraph input be the waveform itself, which fails validation with `W>16384`. -- Treat this limitation as universal across iPhones with ANE; the precise logging/behavior can vary, but no iPhone ANE can accept a subgraph that sees `W=240000`. - -## Current repo behavior and fix - -- The exporter now sets CPU+GPU compute units for waveform‑in components to avoid ANE attempts while preserving the 15 s window: - - `parakeet_preprocessor.mlpackage` - - `parakeet_mel_encoder.mlpackage` (fused waveform→mel→encoder) -- See the changes in `convert-parakeet.py`: - - `convert-parakeet.py:225` and `convert-parakeet.py:237` — preprocessor exported with `ct.ComputeUnit.CPU_AND_GPU`. - - `convert-parakeet.py:289` and `convert-parakeet.py:301` — fused mel+encoder exported with `ct.ComputeUnit.CPU_AND_GPU`. - -## Recommended runtime settings (iOS) - -- Preprocessor (waveform→mel): set `MLModelConfiguration.computeUnits = .cpuAndGPU` to skip ANE attempts and use CPU/GPU efficiently. -- Encoder / Decoder / Joint: set `computeUnits = .cpuAndNeuralEngine` (or `.all`) to leverage ANE. -- Using `.cpuAndNeuralEngine` for all models also works with the split setup (preprocessor falls back to CPU), but setting `.cpuAndGPU` on the preprocessor avoids ANE compile warnings and extra load-time overhead. - -## Using the split models (recommended) - -- Pipeline: - 1. `parakeet_preprocessor.mlmodelc` (CPU+GPU): input `audio_signal [1, 240000]`, `audio_length [1]` → output `mel [1, 128, 1501]`, `mel_length [1]`. - 2. `parakeet_encoder.mlmodelc` (ANE): input `mel [1, 128, 1501]`, `mel_length [1]` → output `encoder [1, 1024, 188]`, `encoder_length [1]`. - 3. Decoder / Joint / Decision: keep on ANE. -- Shapes are captured in `parakeet_coreml/metadata.json`. - -## If fused‑on‑ANE is required - -- The current fused API (`[1, 240000]` waveform input) cannot compile to ANE on iPhones due to the hard ≀16384 per‑dimension cap. -- Viable approach: introduce a chunked fused variant with input like `[1, 15, 16000]` and implement overlap‑aware preemphasis + STFT + mel inside the model to preserve exact parity with global STFT (center padding, hop=160, win=400). Ensure no op sees a dimension >16384. This changes the model API and requires careful seam handling. -- Alternative: stream ≀1.024 s windows with state and accumulate outputs, but this changes runtime control flow. - -## FAQ - -- Q: Which model causes the iPhone 13 error? - - A: The preprocessor stage inside the fused MelEncoder (not the standalone encoder). -- Q: Is this only iPhone 13? - - A: No. Treat it as affecting all iPhone ANE generations; some may just silently partition or fall back differently. -- Q: Why does M4 work? - - A: Likely because the preprocessor subgraph runs on CPU/GPU; the ANE never “sees” `W=240000` on macOS. - diff --git a/parakeet-tdt-v2-0.6b/coreml/context/nemo-joint-fuse-mode.md b/parakeet-tdt-v2-0.6b/coreml/context/nemo-joint-fuse-mode.md deleted file mode 100644 index 1c5eeae7a0f89608d7a6abddf40a1412c13c642a..0000000000000000000000000000000000000000 --- a/parakeet-tdt-v2-0.6b/coreml/context/nemo-joint-fuse-mode.md +++ /dev/null @@ -1,87 +0,0 @@ -# NeMo Joint Network Fuse Mode - -## Overview - -The NeMo RNNT Joint network has two operational modes controlled by the `fuse_loss_wer` flag. Understanding this is crucial for proper validation and inference. - -## Fuse Mode Details - -### Normal Mode (`fuse_loss_wer = False`) - -**Purpose**: Standard inference and model export -**Interface**: `joint(encoder_outputs, decoder_outputs)` -**Returns**: Raw logits `[B, T, U, V + 1]` -**Use case**: -- Model inference -- Model export (CoreML, ONNX, etc.) -- Validation and comparison - -```python -# Simple joint computation -logits = model.joint(encoder_outputs=encoder_out, decoder_outputs=decoder_out) -``` - -### Fused Mode (`fuse_loss_wer = True`) - -**Purpose**: Memory-optimized training with integrated loss/WER computation -**Interface**: Requires additional parameters -**Returns**: Loss and WER metrics -**Use case**: Training with memory constraints - -**Required parameters when fused**: -- `encoder_outputs` (mandatory) -- `decoder_outputs` (optional, required for loss) -- `encoder_lengths` (required) -- `transcripts` (optional, required for WER) -- `transcript_lengths` (required) - -**Memory optimization**: Processes batch in sub-batches to reduce memory usage - -```python -# Fused mode - more complex interface -result = model.joint( - encoder_outputs=encoder_out, - decoder_outputs=decoder_out, - encoder_lengths=enc_lengths, - transcripts=transcripts, # Ground truth tokens - transcript_lengths=trans_lengths -) -``` - -## Why We Disable Fused Mode for Validation - -### 1. Interface Simplicity -- Fused mode requires ground truth transcripts that we don't have during validation -- Normal mode only needs encoder/decoder outputs, matching CoreML model interface - -### 2. Export Compatibility -```python -def _prepare_for_export(self, **kwargs): - self._fuse_loss_wer = False # Automatically disabled for export - self.log_softmax = False -``` -NeMo automatically disables fused mode during model export, indicating this is the correct mode for inference. - -### 3. Validation Purpose -- We want raw logits for comparison, not loss/WER calculations -- Fused mode is training-oriented, not inference-oriented - -### 4. CoreML Conversion -CoreML models are converted from the non-fused mode, so validation should use the same mode to ensure fair comparison. - -## Implementation in compare-components.py - -```python -# Disable fused mode for proper inference comparison -asr_model.joint.set_fuse_loss_wer(False) - -# Now we can use the simple interface that matches CoreML -logits_ref = asr_model.joint( - encoder_outputs=encoder_ref, - decoder_outputs=decoder_ref -) -``` - -## Key Takeaway - -For validation and inference, always use `fuse_loss_wer = False` to get the standard joint network behavior that matches exported models and enables fair comparison with CoreML implementations. \ No newline at end of file diff --git a/parakeet-tdt-v2-0.6b/coreml/context/parakeet_tdt_v2_architecture.md b/parakeet-tdt-v2-0.6b/coreml/context/parakeet_tdt_v2_architecture.md deleted file mode 100644 index ca086b16c6a2b05b0bcf49a3d31cd9c98d7844d9..0000000000000000000000000000000000000000 --- a/parakeet-tdt-v2-0.6b/coreml/context/parakeet_tdt_v2_architecture.md +++ /dev/null @@ -1,61 +0,0 @@ -# Parakeet‑TDT‑0.6B‑v2 Architecture Overview - -## Model topology - -- Streaming‑capable RNN‑Transducer (RNNT) ASR model trained at 16 kHz. -- Time‑Depth Transducer (TDT) extends RNNT to jointly predict token and duration buckets (commonly [0, 1, 2, 3, 4]). -- Four major learnable blocks packaged in the `.nemo` checkpoint: - 1. `AudioToMelSpectrogramPreprocessor` – waveform → 128×T mel‑spectrogram. - 2. `ConformerEncoder` (FastConformer‑style) – 24 conformer blocks, `d_model=1024`, `n_heads=8`, depthwise CNN subsampling with total reduction ×8. - 3. `RNNTDecoder` – 2‑layer LSTM prediction network, hidden size 640, `blank_as_pad=True` for efficient batching and blank handling. - 4. `RNNTJoint` – linear projections from encoder (1024) and decoder (640) into joint space 640 with ReLU, then to class logits. - -## Audio front‑end - -- `sample_rate = 16000`, Hann window, 25 ms window, 10 ms hop (`n_fft=512`, `win_length=400`, `hop_length=160`, `center=True`). -- 128 mel filters (Slaney‑style), log‑magnitude power 2.0 and per‑feature normalization. -- Outputs `(mel, mel_length)` feed the encoder. For CoreML, both tensors must be surfaced to preserve valid‑lengths through subsampling. - -## Encoder (FastConformer) - -- Implemented as `nemo.collections.asr.modules.ConformerEncoder` configured per FastConformer. -- 24 blocks, `d_model=1024`, FFN expansion ×4, 8‑head relative‑position attention, conv kernel 9, batch‑norm. -- Dropouts: 0.1 general, 0.1 pre‑encoder, 0.0 embedding, 0.1 attention; stochastic depth disabled. -- Depthwise CNN subsampling yields overall time reduction ×8. Returns features `[B, D=1024, T_enc]` and lengths. - -## Decoder (`RNNTDecoder`) - -- Stateful LSTM prediction network (2 layers, hidden=640). See `nemo/collections/asr/modules/rnnt.py`. -- `blank_as_pad=True` adds a learnable pad/blank to the embedding; embedding returns zeros for blank, simplifying blank handling. -- Export toggles `_rnnt_export` and uses `.predict()` without prepending SOS so decoder outputs shape `[B, U, 640]`. -- Streaming state is a tuple `(h, c)`, each `[L=2, B, H=640]`. Utilities exist to batch and select states during beam search. - -## Joint network & TDT - -- `nemo.collections.asr.modules.RNNTJoint` with `joint_hidden=640`, activation=ReLU, dropout=0.2. -- Input shapes: encoder `[B, T_enc, 1024]`, decoder `[B, U, 640]` → output logits `[B, T_enc, U, V+1+E]`. -- Vocabulary `V=8192` (SentencePiece BPE). RNNT blank adds `+1`. TDT uses `num_extra_outputs=E=5`, appended in the last dimension of the joint logits. The TDT loss splits the tail 5 entries as duration logits for buckets `[0,1,2,3,4]`. -- Training often enables `fuse_loss_wer=true` with `fused_batch_size=4`; export/inference use just `joint.joint(...)`. - -## Auxiliary heads - -- No separate CTC decoder is attached in the public v2 checkpoint. -- Loss is TDT: `fastemit_lambda=0.0`, `durations=[0,1,2,3,4]`, `sigma=0.02`, `omega=0.1`. - -## Tokenizer - -- SentencePiece BPE (`tokenizer.model`) with 8192 subword units plus control symbols (e.g., punctuation/timestamp markers). RNNT adds blank internally. - -## Notes for CoreML export - -- Exportable subgraphs map 1:1 to NeMo modules: preprocessor, encoder, decoder, joint, and (optionally) a greedy decoder. In addition, we export two fused variants for on‑device efficiency: (1) `mel_encoder` which combines preprocessor+encoder, and (2) `joint_decision` which applies split/softmax/argmax to joint logits to output `token_id`, `token_prob`, and `duration`. -- Decoder state I/O: expose `h` and `c` separately as `[L=2, B, 640]` multi‑arrays to enable on‑device streaming. -- TDT duration logits are appended to the class dimension; downstream code must split `[V+1]` token logits from the last `5` duration logits. - -## Fixed 15‑second window (CoreML constraint) - -- Audio: `B=1`, `T_audio = 15s × 16kHz = 240,000` samples. -- Mel frames (center=True): `T_mel ≈ floor(T_audio / 160) + 1 = 1,500 + 1 = 1,501` → mel `[1, 128, 1501]`. -- Encoder length (×8 subsampling): `T_enc = ceil(1501 / 8) = 188` → encoder `[1, 1024, 188]` (wrapper transposes to `[1, 188, 1024]`). -- Decoder step budget: choose `U_max` to cap per‑window decoding (e.g., 256–384). Joint outputs then shape `[1, T_enc, U_max, 8192+1+5]`. - diff --git a/parakeet-tdt-v2-0.6b/coreml/context/silero_coreml_reference.md b/parakeet-tdt-v2-0.6b/coreml/context/silero_coreml_reference.md deleted file mode 100644 index 2efa8c5dad24884352d705092051bc1fedce9ca7..0000000000000000000000000000000000000000 --- a/parakeet-tdt-v2-0.6b/coreml/context/silero_coreml_reference.md +++ /dev/null @@ -1,53 +0,0 @@ -# Silero VAD CoreML Reference (Techniques to Reuse) - -This note distills patterns from `/models/vad/silero-vad/coreml` that we can reuse for Parakeet-TDT and highlights CoreML limits to plan around. - -## Tooling Patterns Worth Copying - -- **Dedicated CLI entry points**: `convert-coreml.py` performs all exports, while `compare-models.py` focuses solely on validation/plotting. Keeping convert vs. compare independent shortens iteration when only one side changes. -- **Module wrappers**: Lightweight `nn.Module` wrappers (`STFTModel`, `EncoderModel`, `DecoderModel`) map state-dict weights into export-friendly graphs. They surface state tensors explicitly (hidden/cell) and clamp activations where CoreML would otherwise overflow. -- **Shape contracts via CoreML types**: Inputs/outputs use `ct.TensorType` to document tensor layouts. Silero leans on `ct.RangeDim` for flexible windows; for Parakeet we can simplify by locking to the fixed 15 s window (240 k samples → 1501 mels → 188 encoder frames). -- **Metadata & publish scripts**: Conversion annotates author/version on each mlpackage and reports disk size. `compile_modelc.py` shows how to mass-compile mlpackages to `.mlmodelc` bundles via `xcrun`—handy when preparing release artifacts. -- **Multiple deployment variants**: The project exports both modular (STFT→Encoder→Decoder) and unified pipelines. The 256 ms unified variant demonstrates how to encode fixed iteration counts inside the traced graph (8 chunks + noisy-OR aggregation) for batch workflows. - -## Techniques Directly Applicable to Parakeet-TDT - -- **Separate convert & compare CLIs** - - `convert_parakeet_components.py` should mirror the Silero convert script: accept output directory, compute-unit/precision knobs, optional fused graph exports (e.g., preprocessor+encoder). - - A new `compare_parakeet_models.py` can follow the VAD wrapper pattern: run NeMo vs. CoreML components, collect per-layer diffs, and optionally chart mel/encoder/token probabilities over time. -- **Wrapper modules per component** - - We already have `PreprocessorWrapper`, `EncoderWrapper`, `DecoderWrapper`, `JointWrapper`; extend them with deterministic input padding/casts just like Silero’s wrappers to avoid CoreML runtime surprises. -- **Fixed-window CoreML shapes** - - Keep CoreML inputs fixed to the 15 s window. If we ever change the window length, re-export the models instead of relying on dynamic shape bounds. -- **Explicit state I/O** - - Follow Silero’s practice of exposing LSTM states as inputs/outputs every call. For Parakeet, that means separate `h`/`c` arrays of shape `[num_layers, batch, 640]`, enabling streaming decode loops outside CoreML. -- **Release automation** - - Borrow `compile_modelc.py` to generate `.mlmodelc`; consider a similar helper for quantization experiments once baseline parity is achieved. - -## Parakeet Components to Export - -1. **Audio front-end (Mel)** – same strategy as Silero STFT: wrap the NeMo preprocessor, trace with fixed 15 s audio (240 k samples), and export with constant shapes. -2. **Encoder (FastConformer)** – export with constant shapes for the 188 encoder frames corresponding to the 15 s window (transpose to time-major for CoreML compatibility). -3. **Decoder (RNNT Prediction Network)** – stateful LSTM; replicate Silero’s explicit state tensors but with `[2, 1, 640]` dims and integer targets. Provide zero-state helper for CoreML consumers. -4. **Joint network + TDT head** – trace head-only module; outputs `[B, T_enc, U, V+1+5]`. Document token vs. duration slices (see `tdt_decoding_notes.md`). -5. **Fused graphs** – export both: - - `mel_encoder` (preprocessor+encoder) to reduce I/O. - - `joint_decision` (joint + split/softmax/argmax) to avoid host-side post-processing on logits and return `token_id`, `token_prob`, and `duration` directly. - -## Likely CoreML Pain Points - -- **Joint tensor size**: At 15 s the joint produces ~188×U×(8192+6) floats. Even at `U=256`, that is >400 MB of fp32 activations. Plan for fp16 exports and/or tiled decoding (call joint per-step rather than full grid). -- **Dynamic loops**: RNNT greedy/beam search loops can’t be represented in a single CoreML graph (no `while` support). Expect to orchestrate decoding in Swift/Python by issuing repeated `decoder` and `joint` calls. -- **Large vocabulary softmax**: Full RNNT softmax over 8193 classes is CoreML-friendly but expensive. Avoid embedding beam-search logic in CoreML; keep it host-side. -- **Duration logits**: Splitting the final 5 logits inside CoreML is trivial, but combining them with token control flow should stay in host code (mirrors Silero’s noisy-OR aggregation done in Python before tracing). -- **Tokenizer & text post-processing**: SentencePiece decoding, punctuation fixes, and timestamp formatting remain in Python/Swift; CoreML returns logits only. -- **Feature normalization stats**: Stick to constant statistics baked into the wrapper as Silero does for STFT weights; do not rely on CoreML `LayerNormalization` with dynamic reductions over time. - -## Python / Host Responsibilities - -- Audio chunking, state caching, and greedy/beam decoding loops (leveraging exported decoder + joint). -- Token → text conversion (`sentencepiece`), duration bucket interpretation, punctuation/TDT alignment heuristics. -- Validation tooling (diff calculations, plotting) analogous to Silero’s `compare-models.py`. -- Optional post-processing (e.g., noise suppression, segmentation) that require dynamic control flow. - -Reuse these patterns to keep Parakeet’s conversion pipeline maintainable while respecting CoreML’s graph restrictions. diff --git a/parakeet-tdt-v2-0.6b/coreml/context/tdt_decoding_notes.md b/parakeet-tdt-v2-0.6b/coreml/context/tdt_decoding_notes.md deleted file mode 100644 index a1c5fb6ada701bc349921fbf288504b95fe34840..0000000000000000000000000000000000000000 --- a/parakeet-tdt-v2-0.6b/coreml/context/tdt_decoding_notes.md +++ /dev/null @@ -1,36 +0,0 @@ -# TDT Decoding Notes (RNNT + Duration Buckets) - -Parakeet‑TDT extends RNNT by appending 5 duration logits to the joint’s class axis. These describe coarse duration buckets associated with an emitted symbol. This note summarizes how to consume them alongside standard RNNT greedy decoding. - -## Joint output - -- Shape: `logits[B, T_enc, U, V+1+5]` where: - - `V=8192` BPE vocab, `+1` RNNT blank, `+5` TDT duration buckets. -- Split last dimension: - - `token_logits = logits[..., :V+1]` - - `duration_logits = logits[..., V+1:] # shape [..., 5]` - -## Greedy decoding sketch - -- Initialize decoder state `(h, c)` and `u=0`; iterate over encoder time `t=0..T_enc-1`. -- At each `(t,u)`: - - Either use raw joint logits: retrieve `token_logits[t, u]` and pick `k = argmax(token_logits)`. - - Or call the fused `JointDecisionWrapper` CoreML model to get `token_id[t,u]`, `token_prob[t,u]`, and `duration[t,u]` directly. - - If `k` is blank: advance `t += 1` (RNNT time advance), keep `u`. - - Else (emit token): append token `k` to hypothesis, update decoder state via `DecoderWrapper` with that token, increment `u += 1`. - - Duration (optional): `d = argmax(duration_logits[t, u_prev])`. Map `d ∈ {0..4}` to bucket durations per training config. Use to assign sub‑frame timestamp estimates. - -## Using duration buckets - -- Bucket IDs 0..4 correspond to the configured `tdt_durations = [0,1,2,3,4]` used during training. -- A simple heuristic for timestamps per emitted token: - - Keep track of current mel/encoder frame index. - - When emitting a token at `(t,u)`, associate bucket `d = argmax(duration_logits[t,u])`. - - Convert bucket to time by multiplying with the encoder frame stride (`8× hop_length / sr = 8×0.01 s = 80 ms` per encoder step) or use mel frame stride if preferred. -- For better quality, prefer server‑side timestamp decoding identical to NeMo’s implementation and use duration buckets as a hint when post‑processing on device. - -## Notes - -- NeMo trains the joint with `num_extra_outputs=5`. These logits are not part of the token softmax and should not affect argmax over `V+1`. -- The model card claims accurate word/segment timestamps; TDT buckets contribute to that, but exact use may vary per downstream aligner. -- If durations are not needed, you can ignore the last 5 logits entirely. diff --git a/parakeet-tdt-v2-0.6b/coreml/convert-parakeet.py b/parakeet-tdt-v2-0.6b/coreml/convert-parakeet.py deleted file mode 100644 index eaaa19ce95913996b3772611443ad6a7c0be7044..0000000000000000000000000000000000000000 --- a/parakeet-tdt-v2-0.6b/coreml/convert-parakeet.py +++ /dev/null @@ -1,619 +0,0 @@ -#!/usr/bin/env python3 -"""CLI for exporting Parakeet TDT v2 components to CoreML.""" -from __future__ import annotations - -import json -from dataclasses import asdict -from pathlib import Path -from typing import Dict, Optional, Tuple - -import coremltools as ct -import numpy as np -import soundfile as sf -import torch -import typer - -import nemo.collections.asr as nemo_asr - -from individual_components import ( - DecoderWrapper, - EncoderWrapper, - ExportSettings, - JointWrapper, - JointDecisionWrapper, - JointDecisionSingleStep, - PreprocessorWrapper, - MelEncoderWrapper, - _coreml_convert, -) - -DEFAULT_MODEL_ID = "nvidia/parakeet-tdt-0.6b-v2" -AUTHOR = "Fluid Inference" - - -def _compute_length(seconds: float, sample_rate: int) -> int: - return int(round(seconds * sample_rate)) - - -def _prepare_audio( - validation_audio: Optional[Path], - sample_rate: int, - max_samples: int, - seed: Optional[int], -) -> torch.Tensor: - if validation_audio is None: - if seed is not None: - torch.manual_seed(seed) - audio = torch.randn(1, max_samples, dtype=torch.float32) - return audio - - data, sr = sf.read(str(validation_audio), dtype="float32") - if sr != sample_rate: - raise typer.BadParameter( - f"Validation audio sample rate {sr} does not match model rate {sample_rate}" - ) - - if data.ndim > 1: - data = data[:, 0] - - if data.size == 0: - raise typer.BadParameter("Validation audio is empty") - - if data.size < max_samples: - pad_width = max_samples - data.size - data = np.pad(data, (0, pad_width)) - elif data.size > max_samples: - data = data[:max_samples] - - audio = torch.from_numpy(data).unsqueeze(0).to(dtype=torch.float32) - return audio - - -def _save_mlpackage(model: ct.models.MLModel, path: Path, description: str) -> None: - # Ensure iOS 17+ target for MLProgram ops and ANE readiness - try: - model.minimum_deployment_target = ct.target.iOS17 - except Exception: - pass - model.short_description = description - model.author = AUTHOR - path.parent.mkdir(parents=True, exist_ok=True) - model.save(str(path)) - - -def _tensor_shape(tensor: torch.Tensor) -> Tuple[int, ...]: - return tuple(int(dim) for dim in tensor.shape) - - -def _parse_compute_units(name: str) -> ct.ComputeUnit: - """Parse a human-friendly compute units string into ct.ComputeUnit. - - Accepted (case-insensitive): ALL, CPU_ONLY, CPU_AND_GPU, CPU_AND_NE. - """ - normalized = str(name).strip().upper() - mapping = { - "ALL": ct.ComputeUnit.ALL, - "CPU_ONLY": ct.ComputeUnit.CPU_ONLY, - "CPU_AND_GPU": ct.ComputeUnit.CPU_AND_GPU, - "CPU_AND_NE": ct.ComputeUnit.CPU_AND_NE, - "CPU_AND_NEURALENGINE": ct.ComputeUnit.CPU_AND_NE, - } - if normalized not in mapping: - raise typer.BadParameter( - f"Unknown compute units '{name}'. Choose from: " + ", ".join(mapping.keys()) - ) - return mapping[normalized] - - -def _parse_compute_precision(name: Optional[str]) -> Optional[ct.precision]: - """Parse compute precision string into ct.precision or None. - - Accepted (case-insensitive): FLOAT32, FLOAT16. If None/empty, returns None (tool default). - """ - if name is None: - return None - normalized = str(name).strip().upper() - if normalized == "": - return None - mapping = { - "FLOAT32": ct.precision.FLOAT32, - "FLOAT16": ct.precision.FLOAT16, - } - if normalized not in mapping: - raise typer.BadParameter( - f"Unknown compute precision '{name}'. Choose from: " + ", ".join(mapping.keys()) - ) - return mapping[normalized] - - -# Validation logic removed; use compare-compnents.py for comparisons. - - -# Fixed export choices: CPU_ONLY + FP32, min target iOS17 - - -app = typer.Typer(add_completion=False, pretty_exceptions_show_locals=False) - - -@app.command() -def convert( - nemo_path: Optional[Path] = typer.Option( - None, - "--nemo-path", - exists=True, - resolve_path=True, - help="Path to parakeet-tdt-0.6b-v2 .nemo checkpoint (skip to auto-download)", - ), - model_id: str = typer.Option( - DEFAULT_MODEL_ID, - "--model-id", - help="Model identifier to download when --nemo-path is omitted", - ), - output_dir: Path = typer.Option(Path("parakeet_coreml"), help="Directory where mlpackages and metadata will be written"), - preprocessor_cu: str = typer.Option( - "CPU_ONLY", - "--preprocessor-cu", - help="Compute units for preprocessor (default CPU_ONLY)", - ), - mel_encoder_cu: str = typer.Option( - "CPU_ONLY", - "--mel-encoder-cu", - help="Compute units for fused mel+encoder (default CPU_ONLY)", - ), - compute_precision: Optional[str] = typer.Option( - None, - "--compute-precision", - help="Export precision: FLOAT32 (default) or FLOAT16 to shrink non-quantized weights.", - ), -) -> None: - """Export all Parakeet sub-modules to CoreML with a fixed 15-second window.""" - # Runtime CoreML contract keeps U=1 so the prediction net matches the streaming decoder. - export_settings = ExportSettings( - output_dir=output_dir, - compute_units=ct.ComputeUnit.CPU_ONLY, # Default: CPU-only for all components - deployment_target=ct.target.iOS17, # iOS 17+ features and kernels - compute_precision=_parse_compute_precision(compute_precision), - max_audio_seconds=15.0, - max_symbol_steps=1, - ) - - typer.echo("Export configuration:") - typer.echo(asdict(export_settings)) - - output_dir.mkdir(parents=True, exist_ok=True) - pre_cu = _parse_compute_units(preprocessor_cu) - melenc_cu = _parse_compute_units(mel_encoder_cu) - - if nemo_path is not None: - typer.echo(f"Loading NeMo model from {nemo_path}
") - asr_model = nemo_asr.models.EncDecRNNTBPEModel.restore_from( - str(nemo_path), map_location="cpu" - ) - checkpoint_meta = { - "type": "file", - "path": str(nemo_path), - } - else: - typer.echo(f"Downloading NeMo model via {model_id}
") - asr_model = nemo_asr.models.EncDecRNNTBPEModel.from_pretrained( - model_id, map_location="cpu" - ) - checkpoint_meta = { - "type": "pretrained", - "model_id": model_id, - } - asr_model.eval() - - sample_rate = int(asr_model.cfg.preprocessor.sample_rate) - max_samples = _compute_length(export_settings.max_audio_seconds, sample_rate) - # Prefer a bundled 15s 16kHz audio if available - default_audio = (Path(__file__).parent / "audio" / "yc_first_minute_16k_15s.wav").resolve() - if not default_audio.exists(): - raise typer.BadParameter(f"Expected 15s trace audio at {default_audio}; add the file to proceed.") - typer.echo(f"Using trace audio: {default_audio}") - audio_tensor = _prepare_audio(default_audio, sample_rate, max_samples, seed=None) - audio_length = torch.tensor([max_samples], dtype=torch.int32) - - preprocessor = PreprocessorWrapper(asr_model.preprocessor.eval()) - encoder = EncoderWrapper(asr_model.encoder.eval()) - decoder = DecoderWrapper(asr_model.decoder.eval()) - joint = JointWrapper(asr_model.joint.eval()) - - decoder_export_flag = getattr(asr_model.decoder, "_rnnt_export", False) - asr_model.decoder._rnnt_export = True - - try: - with torch.inference_mode(): - mel_ref, mel_length_ref = preprocessor(audio_tensor, audio_length) - mel_length_ref = mel_length_ref.to(dtype=torch.int32) - encoder_ref, encoder_length_ref = encoder(mel_ref, mel_length_ref) - encoder_length_ref = encoder_length_ref.to(dtype=torch.int32) - - # Clone Tensors to drop the inference tensor flag before tracing - mel_ref = mel_ref.clone() - mel_length_ref = mel_length_ref.clone() - encoder_ref = encoder_ref.clone() - encoder_length_ref = encoder_length_ref.clone() - - vocab_size = int(asr_model.tokenizer.vocab_size) - num_extra = int(asr_model.joint.num_extra_outputs) - decoder_hidden = int(asr_model.decoder.pred_hidden) - decoder_layers = int(asr_model.decoder.pred_rnn_layers) - - targets = torch.full( - (1, export_settings.max_symbol_steps), - fill_value=asr_model.decoder.blank_idx, - dtype=torch.int32, - ) - target_lengths = torch.tensor( - [export_settings.max_symbol_steps], dtype=torch.int32 - ) - zero_state = torch.zeros( - decoder_layers, - 1, - decoder_hidden, - dtype=torch.float32, - ) - - with torch.inference_mode(): - decoder_ref, h_ref, c_ref = decoder(targets, target_lengths, zero_state, zero_state) - joint_ref = joint(encoder_ref, decoder_ref) - - decoder_ref = decoder_ref.clone() - h_ref = h_ref.clone() - c_ref = c_ref.clone() - joint_ref = joint_ref.clone() - - typer.echo("Tracing and converting preprocessor
") - # Ensure tracing happens on CPU explicitly - preprocessor = preprocessor.cpu() - audio_tensor = audio_tensor.cpu() - audio_length = audio_length.cpu() - traced_preprocessor = torch.jit.trace( - preprocessor, (audio_tensor, audio_length), strict=False - ) - traced_preprocessor.eval() - preprocessor_inputs = [ - # Allow variable-length audio up to the fixed 15s window using RangeDim - ct.TensorType( - name="audio_signal", - shape=(1, ct.RangeDim(1, max_samples)), - dtype=np.float32, - ), - ct.TensorType(name="audio_length", shape=(1,), dtype=np.int32), - ] - preprocessor_outputs = [ - ct.TensorType(name="mel", dtype=np.float32), - ct.TensorType(name="mel_length", dtype=np.int32), - ] - # Preprocessor compute units (parametrized; default CPU_ONLY) - preprocessor_model = _coreml_convert( - traced_preprocessor, - preprocessor_inputs, - preprocessor_outputs, - export_settings, - compute_units_override=pre_cu, - ) - preprocessor_path = output_dir / "parakeet_preprocessor.mlpackage" - _save_mlpackage( - preprocessor_model, - preprocessor_path, - "Parakeet preprocessor (15 s window)", - ) - - typer.echo("Tracing and converting encoder
") - traced_encoder = torch.jit.trace( - encoder, (mel_ref, mel_length_ref), strict=False - ) - traced_encoder.eval() - encoder_inputs = [ - ct.TensorType(name="mel", shape=_tensor_shape(mel_ref), dtype=np.float32), - ct.TensorType(name="mel_length", shape=(1,), dtype=np.int32), - ] - encoder_outputs = [ - ct.TensorType(name="encoder", dtype=np.float32), - ct.TensorType(name="encoder_length", dtype=np.int32), - ] - # Encoder: CPU only - encoder_model = _coreml_convert( - traced_encoder, - encoder_inputs, - encoder_outputs, - export_settings, - compute_units_override=ct.ComputeUnit.CPU_ONLY, - ) - encoder_path = output_dir / "parakeet_encoder.mlpackage" - _save_mlpackage( - encoder_model, - encoder_path, - "Parakeet encoder (15 s window)", - ) - - # Optional fused export: Preprocessor + Encoder - typer.echo("Tracing and converting fused mel+encoder
") - mel_encoder = MelEncoderWrapper(preprocessor, encoder) - traced_mel_encoder = torch.jit.trace( - mel_encoder, (audio_tensor, audio_length), strict=False - ) - traced_mel_encoder.eval() - mel_encoder_inputs = [ - # Keep fixed 15s window for fused Mel+Encoder - ct.TensorType(name="audio_signal", shape=(1, max_samples), dtype=np.float32), - ct.TensorType(name="audio_length", shape=(1,), dtype=np.int32), - ] - mel_encoder_outputs = [ - ct.TensorType(name="encoder", dtype=np.float32), - ct.TensorType(name="encoder_length", dtype=np.int32), - ] - # Fused mel+encoder compute units (parametrized; default CPU_ONLY) - mel_encoder_model = _coreml_convert( - traced_mel_encoder, - mel_encoder_inputs, - mel_encoder_outputs, - export_settings, - compute_units_override=melenc_cu, - ) - mel_encoder_path = output_dir / "parakeet_mel_encoder.mlpackage" - _save_mlpackage( - mel_encoder_model, - mel_encoder_path, - "Parakeet fused Mel+Encoder (15 s window)", - ) - - typer.echo("Tracing and converting decoder
") - traced_decoder = torch.jit.trace( - decoder, - (targets, target_lengths, zero_state, zero_state), - strict=False, - ) - traced_decoder.eval() - decoder_inputs = [ - ct.TensorType(name="targets", shape=_tensor_shape(targets), dtype=np.int32), - ct.TensorType(name="target_length", shape=(1,), dtype=np.int32), - ct.TensorType(name="h_in", shape=_tensor_shape(zero_state), dtype=np.float32), - ct.TensorType(name="c_in", shape=_tensor_shape(zero_state), dtype=np.float32), - ] - decoder_outputs = [ - ct.TensorType(name="decoder", dtype=np.float32), - ct.TensorType(name="h_out", dtype=np.float32), - ct.TensorType(name="c_out", dtype=np.float32), - ] - # Decoder: CPU only - decoder_model = _coreml_convert( - traced_decoder, - decoder_inputs, - decoder_outputs, - export_settings, - compute_units_override=ct.ComputeUnit.CPU_ONLY, - ) - decoder_path = output_dir / "parakeet_decoder.mlpackage" - _save_mlpackage( - decoder_model, - decoder_path, - "Parakeet decoder (RNNT prediction network)", - ) - - typer.echo("Tracing and converting joint
") - traced_joint = torch.jit.trace( - joint, - (encoder_ref, decoder_ref), - strict=False, - ) - traced_joint.eval() - joint_inputs = [ - ct.TensorType(name="encoder", shape=_tensor_shape(encoder_ref), dtype=np.float32), - ct.TensorType(name="decoder", shape=_tensor_shape(decoder_ref), dtype=np.float32), - ] - joint_outputs = [ - ct.TensorType(name="logits", dtype=np.float32), - ] - # Joint: CPU only - joint_model = _coreml_convert( - traced_joint, - joint_inputs, - joint_outputs, - export_settings, - compute_units_override=ct.ComputeUnit.CPU_ONLY, - ) - joint_path = output_dir / "parakeet_joint.mlpackage" - _save_mlpackage( - joint_model, - joint_path, - "Parakeet joint network (RNNT)", - ) - - # Joint + decision head (split logits, softmax, argmax) - typer.echo("Tracing and converting joint decision head
") - vocab_size = int(asr_model.tokenizer.vocab_size) - num_extra = int(asr_model.joint.num_extra_outputs) - joint_decision = JointDecisionWrapper(joint, vocab_size=vocab_size, num_extra=num_extra) - traced_joint_decision = torch.jit.trace( - joint_decision, - (encoder_ref, decoder_ref), - strict=False, - ) - traced_joint_decision.eval() - joint_decision_inputs = [ - ct.TensorType(name="encoder", shape=_tensor_shape(encoder_ref), dtype=np.float32), - ct.TensorType(name="decoder", shape=_tensor_shape(decoder_ref), dtype=np.float32), - ] - joint_decision_outputs = [ - ct.TensorType(name="token_id", dtype=np.int32), - ct.TensorType(name="token_prob", dtype=np.float32), - ct.TensorType(name="duration", dtype=np.int32), - ] - # JointDecision: CPU only - joint_decision_model = _coreml_convert( - traced_joint_decision, - joint_decision_inputs, - joint_decision_outputs, - export_settings, - compute_units_override=ct.ComputeUnit.CPU_ONLY, - ) - joint_decision_path = output_dir / "parakeet_joint_decision.mlpackage" - _save_mlpackage( - joint_decision_model, - joint_decision_path, - "Parakeet joint + decision head (split, softmax, argmax)", - ) - - # Single-step JointDecision for [1,1024,1] x [1,640,1] -> [1,1,1] - typer.echo("Tracing and converting single-step joint decision
") - jd_single = JointDecisionSingleStep(joint, vocab_size=vocab_size, num_extra=num_extra) - # Create single-step slices from refs - enc_step = encoder_ref[:, :, :1].contiguous() - dec_step = decoder_ref[:, :, :1].contiguous() - traced_jd_single = torch.jit.trace( - jd_single, - (enc_step, dec_step), - strict=False, - ) - traced_jd_single.eval() - jd_single_inputs = [ - ct.TensorType(name="encoder_step", shape=(1, enc_step.shape[1], 1), dtype=np.float32), - ct.TensorType(name="decoder_step", shape=(1, dec_step.shape[1], 1), dtype=np.float32), - ] - jd_single_outputs = [ - ct.TensorType(name="token_id", dtype=np.int32), - ct.TensorType(name="token_prob", dtype=np.float32), - ct.TensorType(name="duration", dtype=np.int32), - ] - # Single-step JointDecision: CPU only - jd_single_model = _coreml_convert( - traced_jd_single, - jd_single_inputs, - jd_single_outputs, - export_settings, - compute_units_override=ct.ComputeUnit.CPU_ONLY, - ) - jd_single_path = output_dir / "parakeet_joint_decision_single_step.mlpackage" - _save_mlpackage( - jd_single_model, - jd_single_path, - "Parakeet single-step joint decision (current frame)", - ) - - metadata: Dict[str, object] = { - "model_id": model_id, - "sample_rate": sample_rate, - "max_audio_seconds": export_settings.max_audio_seconds, - "max_audio_samples": max_samples, - "max_symbol_steps": export_settings.max_symbol_steps, - "vocab_size": vocab_size, - "joint_extra_outputs": num_extra, - "checkpoint": checkpoint_meta, - "coreml": { - "compute_units": export_settings.compute_units.name, - "compute_precision": ( - export_settings.compute_precision.name - if export_settings.compute_precision is not None - else "FLOAT32" - ), - }, - "components": { - "preprocessor": { - "inputs": { - "audio_signal": list(_tensor_shape(audio_tensor)), - "audio_length": [1], - }, - "outputs": { - "mel": list(_tensor_shape(mel_ref)), - "mel_length": [1], - }, - "path": preprocessor_path.name, - }, - "encoder": { - "inputs": { - "mel": list(_tensor_shape(mel_ref)), - "mel_length": [1], - }, - "outputs": { - "encoder": list(_tensor_shape(encoder_ref)), - "encoder_length": [1], - }, - "path": encoder_path.name, - }, - "mel_encoder": { - "inputs": { - "audio_signal": [1, max_samples], - "audio_length": [1], - }, - "outputs": { - "encoder": list(_tensor_shape(encoder_ref)), - "encoder_length": [1], - }, - "path": mel_encoder_path.name, - }, - "decoder": { - "inputs": { - "targets": list(_tensor_shape(targets)), - "target_length": [1], - "h_in": list(_tensor_shape(zero_state)), - "c_in": list(_tensor_shape(zero_state)), - }, - "outputs": { - "decoder": list(_tensor_shape(decoder_ref)), - "h_out": list(_tensor_shape(h_ref)), - "c_out": list(_tensor_shape(c_ref)), - }, - "path": decoder_path.name, - }, - "joint": { - "inputs": { - "encoder": list(_tensor_shape(encoder_ref)), - "decoder": list(_tensor_shape(decoder_ref)), - }, - "outputs": { - "logits": list(_tensor_shape(joint_ref)), - }, - "path": joint_path.name, - }, - "joint_decision": { - "inputs": { - "encoder": list(_tensor_shape(encoder_ref)), - "decoder": list(_tensor_shape(decoder_ref)), - }, - "outputs": { - "token_id": [ - _tensor_shape(encoder_ref)[0], - _tensor_shape(encoder_ref)[1], - _tensor_shape(decoder_ref)[1], - ], - "token_prob": [ - _tensor_shape(encoder_ref)[0], - _tensor_shape(encoder_ref)[1], - _tensor_shape(decoder_ref)[1], - ], - "duration": [ - _tensor_shape(encoder_ref)[0], - _tensor_shape(encoder_ref)[1], - _tensor_shape(decoder_ref)[1], - ], - }, - "path": joint_decision_path.name, - }, - "joint_decision_single_step": { - "inputs": { - "encoder_step": [1, _tensor_shape(encoder_ref)[2-1], 1], - "decoder_step": [1, _tensor_shape(decoder_ref)[2-1], 1], - }, - "outputs": { - "token_id": [1, 1, 1], - "token_prob": [1, 1, 1], - "duration": [1, 1, 1], - }, - "path": jd_single_path.name, - }, - }, - } - - metadata_path = output_dir / "metadata.json" - metadata_path.write_text(json.dumps(metadata, indent=2)) - typer.echo(f"Export complete. Metadata written to {metadata_path}") - - finally: - asr_model.decoder._rnnt_export = decoder_export_flag - - -if __name__ == "__main__": - app() diff --git a/parakeet-tdt-v2-0.6b/coreml/individual_components.py b/parakeet-tdt-v2-0.6b/coreml/individual_components.py deleted file mode 100644 index 5640330bbaa9e73ef18644458bcf023971ca70cf..0000000000000000000000000000000000000000 --- a/parakeet-tdt-v2-0.6b/coreml/individual_components.py +++ /dev/null @@ -1,229 +0,0 @@ -#!/usr/bin/env python3 -"""Export Parakeet TDT v2 RNNT components into CoreML and validate outputs.""" -from __future__ import annotations - -from dataclasses import dataclass -from pathlib import Path -from typing import Optional, Tuple - -import coremltools as ct -import torch - - -@dataclass -class ExportSettings: - output_dir: Path - compute_units: ct.ComputeUnit - deployment_target: Optional[ct.target.iOS17] - compute_precision: Optional[ct.precision] - max_audio_seconds: float - max_symbol_steps: int - - -@dataclass -class ValidationSettings: - audio_path: Optional[Path] - seconds: float - seed: Optional[int] - rtol: float - atol: float - skip: bool - - -@dataclass -class ValidationDiff: - name: str - max_abs_diff: float - max_rel_diff: float - - -@dataclass -class ValidationResult: - source: str - audio_num_samples: int - audio_seconds: float - token_length: int - atol: float - rtol: float - diffs: Tuple[ValidationDiff, ...] - - -class PreprocessorWrapper(torch.nn.Module): - def __init__(self, module: torch.nn.Module) -> None: - super().__init__() - self.module = module - - def forward(self, audio_signal: torch.Tensor, length: torch.Tensor) -> Tuple[torch.Tensor, torch.Tensor]: - mel, mel_length = self.module(input_signal=audio_signal, length=length.to(dtype=torch.long)) - return mel, mel_length - - -class EncoderWrapper(torch.nn.Module): - def __init__(self, module: torch.nn.Module) -> None: - super().__init__() - self.module = module - - def forward(self, features: torch.Tensor, length: torch.Tensor) -> Tuple[torch.Tensor, torch.Tensor]: - encoded, encoded_lengths = self.module(audio_signal=features, length=length.to(dtype=torch.long)) - return encoded, encoded_lengths - - -class DecoderWrapper(torch.nn.Module): - def __init__(self, module: torch.nn.Module) -> None: - super().__init__() - self.module = module - - def forward( - self, - targets: torch.Tensor, - target_lengths: torch.Tensor, - h_in: torch.Tensor, - c_in: torch.Tensor, - ) -> Tuple[torch.Tensor, torch.Tensor, torch.Tensor]: - state = [h_in, c_in] - decoder_output, _, new_state = self.module( - targets=targets.to(dtype=torch.long), - target_length=target_lengths.to(dtype=torch.long), - states=state, - ) - return decoder_output, new_state[0], new_state[1] - - -class JointWrapper(torch.nn.Module): - def __init__(self, module: torch.nn.Module) -> None: - super().__init__() - self.module = module - - def forward(self, encoder_outputs: torch.Tensor, decoder_outputs: torch.Tensor) -> torch.Tensor: - # Input: encoder_outputs [B, D, T], decoder_outputs [B, D, U] - # Transpose to match what projection layers expect - encoder_outputs = encoder_outputs.transpose(1, 2) # [B, T, D] - decoder_outputs = decoder_outputs.transpose(1, 2) # [B, U, D] - - # Apply projections - enc_proj = self.module.enc(encoder_outputs) # [B, T, 640] - dec_proj = self.module.pred(decoder_outputs) # [B, U, 640] - - # Explicit broadcasting along T and U to avoid converter ambiguity - x = enc_proj.unsqueeze(2) + dec_proj.unsqueeze(1) # [B, T, U, 640] - x = self.module.joint_net[0](x) # ReLU - x = self.module.joint_net[1](x) # Dropout (no-op in eval) - out = self.module.joint_net[2](x) # Linear -> logits [B, T, U, 8198] - return out - - -class MelEncoderWrapper(torch.nn.Module): - """Fused wrapper: waveform -> mel -> encoder. - - Inputs: - - audio_signal: [B, S] - - audio_length: [B] - - Outputs: - - encoder: [B, D, T_enc] - - encoder_length: [B] - """ - def __init__(self, preprocessor: PreprocessorWrapper, encoder: EncoderWrapper) -> None: - super().__init__() - self.preprocessor = preprocessor - self.encoder = encoder - - def forward(self, audio_signal: torch.Tensor, audio_length: torch.Tensor) -> Tuple[torch.Tensor, torch.Tensor]: - mel, mel_length = self.preprocessor(audio_signal, audio_length) - encoded, enc_len = self.encoder(mel, mel_length.to(dtype=torch.int32)) - return encoded, enc_len - - -class JointDecisionWrapper(torch.nn.Module): - """Joint + decision head: outputs label id, label prob, duration frames. - - Splits joint logits into token logits and duration logits, applies softmax - over tokens, argmax for both heads, and gathers probability of the chosen token. - - Inputs: - - encoder_outputs: [B, D, T] - - decoder_outputs: [B, D, U] - - Returns: - - token_id: [B, T, U] int32 - - token_prob: [B, T, U] float32 - - duration: [B, T, U] int32 (frames; duration buckets as defined by the model) - """ - def __init__(self, joint: JointWrapper, vocab_size: int, num_extra: int) -> None: - super().__init__() - self.joint = joint - self.vocab_with_blank = int(vocab_size) + 1 - self.num_extra = int(num_extra) - - def forward(self, encoder_outputs: torch.Tensor, decoder_outputs: torch.Tensor): - logits = self.joint(encoder_outputs, decoder_outputs) - token_logits = logits[..., : self.vocab_with_blank] - duration_logits = logits[..., -self.num_extra :] - - # Token selection - token_ids = torch.argmax(token_logits, dim=-1).to(dtype=torch.int32) - token_probs_all = torch.softmax(token_logits, dim=-1) - # gather expects int64 (long) indices; cast only for gather - token_prob = torch.gather( - token_probs_all, dim=-1, index=token_ids.long().unsqueeze(-1) - ).squeeze(-1) - - # Duration prediction (duration bucket argmax → frame increments) - duration = torch.argmax(duration_logits, dim=-1).to(dtype=torch.int32) - return token_ids, token_prob, duration - - -class JointDecisionSingleStep(torch.nn.Module): - """Single-step variant for streaming: encoder_step [1, 1024, 1] -> [1,1,1]. - - Inputs: - - encoder_step: [B=1, D=1024, T=1] - - decoder_step: [B=1, D=640, U=1] - - Returns: - - token_id: [1, 1, 1] int32 - - token_prob: [1, 1, 1] float32 - - duration: [1, 1, 1] int32 - """ - def __init__(self, joint: JointWrapper, vocab_size: int, num_extra: int) -> None: - super().__init__() - self.joint = joint - self.vocab_with_blank = int(vocab_size) + 1 - self.num_extra = int(num_extra) - - def forward(self, encoder_step: torch.Tensor, decoder_step: torch.Tensor): - # Reuse JointWrapper which expects [B, D, T] and [B, D, U] - logits = self.joint(encoder_step, decoder_step) # [1, 1, 1, V+extra] - token_logits = logits[..., : self.vocab_with_blank] - duration_logits = logits[..., -self.num_extra :] - - token_ids = torch.argmax(token_logits, dim=-1, keepdim=False).to(dtype=torch.int32) - token_probs_all = torch.softmax(token_logits, dim=-1) - token_prob = torch.gather( - token_probs_all, dim=-1, index=token_ids.long().unsqueeze(-1) - ).squeeze(-1) - duration = torch.argmax(duration_logits, dim=-1, keepdim=False).to(dtype=torch.int32) - return token_ids, token_prob, duration - - -def _coreml_convert( - traced: torch.jit.ScriptModule, - inputs, - outputs, - settings: ExportSettings, - compute_units_override: Optional[ct.ComputeUnit] = None, -) -> ct.models.MLModel: - cu = compute_units_override if compute_units_override is not None else settings.compute_units - kwargs = { - "convert_to": "mlprogram", - "inputs": inputs, - "outputs": outputs, - "compute_units": cu, - } - print("Converting:", traced.__class__.__name__) - print("Conversion kwargs:", kwargs) - if settings.deployment_target is not None: - kwargs["minimum_deployment_target"] = settings.deployment_target - if settings.compute_precision is not None: - kwargs["compute_precision"] = settings.compute_precision - return ct.convert(traced, **kwargs) diff --git a/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/decoder_steps_l2.png b/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/decoder_steps_l2.png deleted file mode 100644 index e9cc0ba8359231e77deff906e60dfa254568533b..0000000000000000000000000000000000000000 Binary files a/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/decoder_steps_l2.png and /dev/null differ diff --git a/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/encoder_time_l2.png b/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/encoder_time_l2.png deleted file mode 100644 index fa3053e9ba9fd5e0646cbc19b3dc7c3ca417ce5b..0000000000000000000000000000000000000000 Binary files a/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/encoder_time_l2.png and /dev/null differ diff --git a/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/joint_decision_prob_u0.png b/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/joint_decision_prob_u0.png deleted file mode 100644 index 1afb0785b0064069044b5e06f2270cf5248d112b..0000000000000000000000000000000000000000 Binary files a/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/joint_decision_prob_u0.png and /dev/null differ diff --git a/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/joint_decision_token_agree.png b/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/joint_decision_token_agree.png deleted file mode 100644 index 253ad19382995a4aced910a330b59e902b076bbd..0000000000000000000000000000000000000000 Binary files a/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/joint_decision_token_agree.png and /dev/null differ diff --git a/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/joint_time_l2.png b/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/joint_time_l2.png deleted file mode 100644 index dc8c869bd6444a83f95e68aec1aa51c4c0e04ef7..0000000000000000000000000000000000000000 Binary files a/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/joint_time_l2.png and /dev/null differ diff --git a/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/joint_top50.png b/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/joint_top50.png deleted file mode 100644 index 4292db8718f5779d18186ced8c43af327922563c..0000000000000000000000000000000000000000 Binary files a/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/joint_top50.png and /dev/null differ diff --git a/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/latency_fused_speedup.png b/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/latency_fused_speedup.png deleted file mode 100644 index a965cbcac59d25eba8a7dfe7837297df75b53dc2..0000000000000000000000000000000000000000 Binary files a/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/latency_fused_speedup.png and /dev/null differ diff --git a/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/latency_fused_vs_separate.png b/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/latency_fused_vs_separate.png deleted file mode 100644 index d637cdb676c2a30c57dcfb2fea7dc796fa5e9f0f..0000000000000000000000000000000000000000 Binary files a/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/latency_fused_vs_separate.png and /dev/null differ diff --git a/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/latency_speedup.png b/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/latency_speedup.png deleted file mode 100644 index 2688f5b6f2643641f97177665920216c340efa41..0000000000000000000000000000000000000000 Binary files a/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/latency_speedup.png and /dev/null differ diff --git a/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/latency_summary.png b/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/latency_summary.png deleted file mode 100644 index 3c51c8fca7ee437d5296a6e82c2a8f8523bbb444..0000000000000000000000000000000000000000 Binary files a/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/latency_summary.png and /dev/null differ diff --git a/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/mel_composite.png b/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/mel_composite.png deleted file mode 100644 index 50870080ad1be000c4b6fa4b3b8f333ebfaabf3d..0000000000000000000000000000000000000000 --- a/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/mel_composite.png +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:b46c8f85cf9be45d71847d38173d2a2912e285cda2f1d66c21980f769e39b055 -size 318316 diff --git a/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/mel_encoder_time_l2.png b/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/mel_encoder_time_l2.png deleted file mode 100644 index e713f7a17b7658678531e6a7eadafc6a4be26242..0000000000000000000000000000000000000000 Binary files a/parakeet-tdt-v2-0.6b/coreml/plots/compare-components/mel_encoder_time_l2.png and /dev/null differ diff --git a/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/all_components_compile.png b/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/all_components_compile.png deleted file mode 100644 index 11ce0c66d03ed9a5df57c487d06a3d9ebe6062a4..0000000000000000000000000000000000000000 --- a/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/all_components_compile.png +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:d3a9beaa0ed9d9893fa5a123240b7acdaca8e15a734deff20d94def1d4b4fd2d -size 103825 diff --git a/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/all_components_compression.png b/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/all_components_compression.png deleted file mode 100644 index d71982a212f6b7ec1998c6b6b874f05c9f8da353..0000000000000000000000000000000000000000 --- a/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/all_components_compression.png +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:dd7ad4cdd309d603728dae0726d779e274ad48798c69b3f64c1cddebec5a9561 -size 101215 diff --git a/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/all_components_latency.png b/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/all_components_latency.png deleted file mode 100644 index 51d3633deb37d6a46b3419a6d8f6831972597ee9..0000000000000000000000000000000000000000 Binary files a/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/all_components_latency.png and /dev/null differ diff --git a/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/all_components_quality.png b/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/all_components_quality.png deleted file mode 100644 index b702e1cae49f41220a043cbab815e943f37b8d9e..0000000000000000000000000000000000000000 --- a/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/all_components_quality.png +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:a25157bcfeaad91955b959303a8632bc6b3868dc61b5fdf4a1a1ccb91dc35cb1 -size 113723 diff --git a/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/all_components_size.png b/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/all_components_size.png deleted file mode 100644 index 8c0209963b2e28b05d8ea0dc7b2a075ff0e65b7b..0000000000000000000000000000000000000000 Binary files a/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/all_components_size.png and /dev/null differ diff --git a/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/fused_compression.png b/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/fused_compression.png deleted file mode 100644 index 25522432876d20012049ee5cf830ecd0bc8a44be..0000000000000000000000000000000000000000 Binary files a/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/fused_compression.png and /dev/null differ diff --git a/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/fused_latency.png b/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/fused_latency.png deleted file mode 100644 index b92312755973edc5c407f20fa28d42069bb7f57a..0000000000000000000000000000000000000000 Binary files a/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/fused_latency.png and /dev/null differ diff --git a/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/fused_quality.png b/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/fused_quality.png deleted file mode 100644 index 289f61fe3343cd8265674b9ffabd6e21ead217ae..0000000000000000000000000000000000000000 Binary files a/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/fused_quality.png and /dev/null differ diff --git a/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/fused_size.png b/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/fused_size.png deleted file mode 100644 index 8c794846f1c578688ed41de3c3a8df11c4ed161b..0000000000000000000000000000000000000000 Binary files a/parakeet-tdt-v2-0.6b/coreml/plots/quantize/cpu_and_ne/fused_size.png and /dev/null differ diff --git a/parakeet-tdt-v2-0.6b/coreml/pyproject.toml b/parakeet-tdt-v2-0.6b/coreml/pyproject.toml deleted file mode 100644 index d73aaacc5f26aa71ce35b13f3dd9fc69b9616bc7..0000000000000000000000000000000000000000 --- a/parakeet-tdt-v2-0.6b/coreml/pyproject.toml +++ /dev/null @@ -1,251 +0,0 @@ -[project] -name = "parakeet-coreml" -version = "0.1.0" -description = "Add your description here" -readme = "README.md" -requires-python = "==3.10.12" -dependencies = [ - "absl-py==2.3.0", - "accelerate==1.8.1", - "aiohappyeyeballs==2.6.1", - "aiohttp==3.12.13", - "aiosignal==1.3.2", - "alembic==1.16.2", - "annotated-types==0.7.0", - "antlr4-python3-runtime==4.9.3", - "anyio==4.9.0", - "appnope==0.1.4", - "argon2-cffi-bindings==21.2.0", - "argon2-cffi==25.1.0", - "arrow==1.3.0", - "asttokens==3.0.0", - "async-lru==2.0.5", - "async-timeout==5.0.1", - "attrs==25.3.0", - "audioread==3.0.1", - "babel==2.17.0", - "backports-datetime-fromisoformat==2.0.3", - "beautifulsoup4==4.13.4", - "bleach==6.2.0", - "braceexpand==0.1.7", - "cattrs==25.1.1", - "certifi==2025.6.15", - "cffi==1.17.1", - "charset-normalizer==3.4.2", - "click==8.2.1", - "cloudpickle==3.1.1", - "colorlog==6.9.0", - "comm==0.2.2", - "contourpy==1.3.2", - "coremltools==9.0b1", - "cycler==0.12.1", - "cytoolz==1.0.1", - "datasets==3.6.0", - "debugpy==1.8.14", - "decorator==5.2.1", - "defusedxml==0.7.1", - "dill==0.3.8", - "distance==0.1.3", - "docopt==0.6.2", - "editdistance==0.8.1", - "einops==0.8.1", - "exceptiongroup==1.3.0", - "executing==2.2.0", - "fastjsonschema==2.21.1", - "fiddle==0.3.0", - "filelock==3.18.0", - "fonttools==4.58.4", - "fqdn==1.5.1", - "frozenlist==1.7.0", - "fsspec==2024.12.0", - "future==1.0.0", - "g2p-en==2.1.0", - "gitdb==4.0.12", - "gitpython==3.1.44", - "graphviz==0.21", - "grpcio==1.73.1", - "h11==0.16.0", - "hf-xet==1.1.5", - "httpcore==1.0.9", - "httpx==0.28.1", - "huggingface-hub==0.33.1", - "hydra-core==1.3.2", - "idna==3.10", - "inflect==7.5.0", - "intervaltree==3.1.0", - "ipykernel==6.29.5", - "ipython==8.37.0", - "ipywidgets==8.1.7", - "isoduration==20.11.0", - "jedi==0.19.2", - "jinja2==3.1.6", - "jiwer==4.0.0", - "joblib==1.5.1", - "json5==0.12.0", - "jsonpointer==3.0.0", - "jsonschema==4.24.0", - "jsonschema-specifications==2025.4.1", - "jupyter==1.1.1", - "jupyter-console==6.6.3", - "jupyter-events==0.12.0", - "jupyter-lsp==2.2.5", - "jupyter-client==8.6.3", - "jupyter-core==5.8.1", - "jupyter-server==2.16.0", - "jupyter-server-terminals==0.5.3", - "jupyterlab==4.4.4", - "jupyterlab-pygments==0.3.0", - "jupyterlab-server==2.27.3", - "jupyterlab-widgets==3.0.15", - "kaldi-python-io==1.2.2", - "kaldiio==2.18.1", - "kiwisolver==1.4.8", - "lazy-loader==0.4", - "levenshtein==0.27.1", - "lhotse==1.30.3", - "libcst==1.8.2", - "librosa==0.11.0", - "lightning==2.4.0", - "lightning-utilities==0.14.3", - "lilcom==1.8.1", - "llvmlite==0.44.0", - "loguru==0.7.3", - "mako==1.3.10", - "markdown==3.8.2", - "markdown-it-py==3.0.0", - "markupsafe==3.0.2", - "marshmallow==4.0.0", - "matplotlib==3.10.3", - "matplotlib-inline==0.1.7", - "mdurl==0.1.2", - "mediapy==1.1.6", - "mistune==3.1.3", - "more-itertools==10.7.0", - "mpmath==1.3.0", - "msgpack==1.1.1", - "multidict==6.6.2", - "multiprocess==0.70.16", - "nbclient==0.10.2", - "nbconvert==7.16.6", - "nbformat==5.10.4", - "nemo-toolkit==2.3.1", - "nest-asyncio==1.6.0", - "networkx==3.4.2", - "nltk==3.9.1", - "notebook==7.4.3", - "notebook-shim==0.2.4", - "num2words==0.5.14", - "numba==0.61.0", - "numpy==1.26.4", - "omegaconf==2.3.0", - "onnx==1.17.0", - "optuna==4.4.0", - "overrides==7.7.0", - "packaging==24.2", - "pandas==2.3.0", - "pandocfilters==1.5.1", - "parso==0.8.4", - "peft==0.15.2", - "pexpect==4.9.0", - "pillow==11.2.1", - "plac==1.4.5", - "platformdirs==4.3.8", - "pooch==1.8.2", - "prometheus-client==0.22.1", - "prompt-toolkit==3.0.51", - "propcache==0.3.2", - "psutil==7.0.0", - "ptyprocess==0.7.0", - "pure-eval==0.2.3", - "pyaml==25.5.0", - "pyannote-core==5.0.0", - "pyannote-database==5.1.3", - "pyannote-metrics==3.2.1", - "pyarrow==20.0.0", - "pybind11==2.13.6", - "pycparser==2.22", - "pydantic==2.11.7", - "pydantic-core==2.33.2", - "pydub==0.25.1", - "pygments==2.19.2", - "pyloudnorm==0.1.1", - "pyparsing==3.2.3", - "python-dateutil==2.9.0.post0", - "python-json-logger==3.3.0", - "pytorch-lightning==2.5.2", - "pytz==2025.2", - "pyyaml==6.0.2", - "pyzmq==27.0.0", - "rapidfuzz==3.13.0", - "referencing==0.36.2", - "regex==2024.11.6", - "requests==2.32.4", - "resampy==0.4.3", - "rfc3339-validator==0.1.4", - "rfc3986-validator==0.1.1", - "rich==14.0.0", - "rpds-py==0.25.1", - "ruamel-yaml==0.18.14", - "ruamel-yaml-clib==0.2.12", - "sacremoses==0.1.1", - "safetensors==0.5.3", - "scikit-learn==1.5.1", - "scipy==1.15.3", - "send2trash==1.8.3", - "sentencepiece==0.2.0", - "sentry-sdk==2.32.0", - "setproctitle==1.3.6", - "shellingham==1.5.4", - "six==1.17.0", - "smmap==5.0.2", - "sniffio==1.3.1", - "sortedcontainers==2.4.0", - "soundfile==0.13.1", - "soupsieve==2.7", - "sox==1.5.0", - "soxr==0.5.0.post1", - "sqlalchemy==2.0.41", - "stack-data==0.6.3", - "tabulate==0.9.0", - "tensorboard==2.19.0", - "tensorboard-data-server==0.7.2", - "termcolor==3.1.0", - "terminado==0.18.1", - "text-unidecode==1.3", - "texterrors==0.5.1", - "threadpoolctl==3.6.0", - "tinycss2==1.4.0", - "tokenizers==0.21.2", - "tomli==2.2.1", - "toolz==1.0.0", - "torch==2.7.0", - "torchmetrics==1.7.3", - "tornado==6.5.1", - "tqdm==4.67.1", - "traitlets==5.14.3", - "transformers==4.51.3", - "typeguard==4.4.4", - "typer==0.16.0", - "types-python-dateutil==2.9.0.20250516", - "typing-inspection==0.4.1", - "typing-extensions==4.14.0", - "tzdata==2025.2", - "uri-template==1.3.0", - "urllib3==2.5.0", - "wandb==0.20.1", - "wcwidth==0.2.13", - "webcolors==24.11.1", - "webdataset==1.0.2", - "webencodings==0.5.1", - "websocket-client==1.8.0", - "werkzeug==3.1.3", - "wget==3.2", - "widgetsnbextension==4.0.14", - "wrapt==1.17.2", - "xxhash==3.5.0", - "yarl==1.20.1", - "pip>=25.1.1", - "seaborn>=0.13.2", - "pyannote-audio>=3.3.2", - "funasr>=1.2.6", -] diff --git a/parakeet-tdt-v2-0.6b/coreml/quantize_coreml.py b/parakeet-tdt-v2-0.6b/coreml/quantize_coreml.py deleted file mode 100644 index 2067bd179da56ba63fe260b8334c4adc56321970..0000000000000000000000000000000000000000 --- a/parakeet-tdt-v2-0.6b/coreml/quantize_coreml.py +++ /dev/null @@ -1,1207 +0,0 @@ -#!/usr/bin/env python3 -"""Quantize CoreML mlpackages and compare quality, compression, latency, and compile time. - -This script focuses on the fused models: - - parakeet_mel_encoder.mlpackage (waveform -> encoder) - - parakeet_joint_decision.mlpackage (joint + softmax/argmax) - -It also quantizes the rest of the components in the directory for completeness. - -Variants tried by default (examples): - - int8-linear (per-channel) - - palettize 6-bit (Mel-only) - - jd-only variants (int8 and palette) - - prune + int8 - -Outputs: - - A new directory per variant under // with quantized mlpackages - - quantization_summary.json with aggregate metrics (baseline vs each variant) - - Plots saved under /plots/ and mirrored to /plots/quantize// - - fused_quality.png / fused_latency.png / fused_compression.png / fused_size.png (latency chart also shows compile time) - - all_components_quality.png / all_components_latency.png / all_components_compression.png / all_components_size.png (latency chart also shows compile time) - -Notes: - - Uses CoreMLTools optimize.coreml linear/palettize/prune for MLProgram models. - - Keeps the fixed 15-second window shapes as per context/coreml_component_io.md. - - Run via `uv run` to ensure reproducible environment. - - Sets minimum deployment target to iOS 17 for all outputs and loads models with `compute_units=ALL` by default to enable ANE. - - Tracks offline compile time (mlpackage->mlmodelc) and includes that metric in plots. -""" -from __future__ import annotations - -import json -import platform -import shutil -import subprocess -import time -from dataclasses import dataclass -from pathlib import Path -from typing import Dict, List, Optional, Set, Tuple - -import numpy as np -import soundfile as sf -import typer - -import coremltools as ct -from coremltools.optimize.coreml import ( - OptimizationConfig, - OpLinearQuantizerConfig, - OpPalettizerConfig, - OpThresholdPrunerConfig, - linear_quantize_weights, - palettize_weights, - prune_weights, -) - - -# Optional plotting -try: - import matplotlib - matplotlib.use("Agg") - import matplotlib.pyplot as plt - HAS_MPL = True -except Exception: - HAS_MPL = False - - -BASE_DIR = Path(__file__).resolve().parent -BYTES_IN_MB = 1024 * 1024 - - -DISPLAY_LABEL_OVERRIDES = { - "mel6bit-palettize": "mel6bit", - "enc6bit-palettize": "enc6bit", - "int8-linear": "int8 linear", -} - - -# When no explicit components are provided, we derive the set of -# components to quantize from the selected variants' whitelists. -# If any selected variant has no whitelist (i.e., applies globally), -# we quantize all components. -DEFAULT_TARGET_COMPONENTS: Tuple[str, str] = () - - -# Formatting helpers ------------------------------------------------------- - - -def _format_labels(labels: List[str]) -> List[str]: - formatted: List[str] = [] - for label in labels: - pretty = DISPLAY_LABEL_OVERRIDES.get(label, label) - pretty = pretty.replace("_", " ") - pretty = pretty.replace("-", "\n") - formatted.append(pretty) - return formatted - - -def _friendly_component_name(name: str) -> str: - return name.replace("_", " ").title() - - -def _get_metric_series( - metrics: Dict[str, Dict[str, List[float]]], - component: str, - key: str, - length: int, -) -> List[float]: - values = metrics.get(component, {}).get(key) - if values is None: - return [float("nan")] * length - if len(values) >= length: - return list(values[:length]) - padded = list(values) - padded.extend([float("nan")] * (length - len(values))) - return padded - - -def _plot_bar_rows( - out_path: Path, - display_labels: List[str], - rows: List[Dict[str, object]], - title_suffix: str, -) -> None: - if not HAS_MPL: - return - if not rows or not display_labels: - return - - n_labels = len(display_labels) - fig_width = max(8.0, 1.3 * n_labels) - fig_height = max(2.6 * len(rows), 3.0) - - fig, axes = plt.subplots(len(rows), 1, figsize=(fig_width, fig_height), squeeze=False) - axes = axes.flatten() - x = np.arange(n_labels) - - for ax, row in zip(axes, rows): - values = np.asarray(row.get("values", [float("nan")] * n_labels), dtype=np.float64) - color = row.get("color") - bars = ax.bar(x, values, width=0.6, color=color) - ax.set_title(str(row.get("title", ""))) - ax.set_xticks(x) - ax.set_xticklabels(display_labels, rotation=25, ha="right") - ylim = row.get("ylim") - if isinstance(ylim, tuple) and len(ylim) == 2: - ax.set_ylim(ylim) - ax.grid(axis="y", linestyle="--", alpha=0.3) - # Add value labels for bars - for bar in bars: - h = bar.get_height() - if np.isnan(h): - continue - ax.annotate( - f"{h:.2f}", - xy=(bar.get_x() + bar.get_width() / 2, h), - xytext=(0, 3), - textcoords="offset points", - ha="center", - va="bottom", - fontsize=8, - ) - - if title_suffix: - fig.suptitle(title_suffix, fontsize=12) - fig.tight_layout(rect=(0, 0, 1, 0.95)) - else: - fig.tight_layout() - - out_path.parent.mkdir(parents=True, exist_ok=True) - plt.savefig(out_path) - plt.close(fig) - - -def _plot_fused_category_charts( - plot_dir: Path, - labels: List[str], - mel_quality: List[float], - mel_latency_ms: List[float], - mel_compression: List[float], - mel_size_mb: List[float], - jd_acc: List[float], - jd_latency_ms: List[float], - jd_compression: List[float], - jd_size_mb: List[float], - mel_compile_ms: Optional[List[float]] = None, - jd_compile_ms: Optional[List[float]] = None, - title_suffix: str = "", -) -> List[Path]: - if not HAS_MPL: - return [] - outputs: List[Path] = [] - display = _format_labels(labels) - prefix = f"Fused Components — {title_suffix}" if title_suffix else "Fused Components" - - quality_path = plot_dir / "fused_quality.png" - _plot_bar_rows( - quality_path, - display, - [ - { - "title": "MelEncoder quality (1 - norm err)", - "values": mel_quality, - "color": "C0", - "ylim": (0.0, 1.05), - }, - { - "title": "JointDecision token-id match rate", - "values": jd_acc, - "color": "C0", - "ylim": (0.0, 1.05), - }, - ], - f"{prefix} — Quality", - ) - outputs.append(quality_path) - - compression_path = plot_dir / "fused_compression.png" - _plot_bar_rows( - compression_path, - display, - [ - { - "title": "MelEncoder compression ratio", - "values": mel_compression, - "color": "C2", - }, - { - "title": "JointDecision compression ratio", - "values": jd_compression, - "color": "C2", - }, - ], - f"{prefix} — Compression", - ) - outputs.append(compression_path) - - size_path = plot_dir / "fused_size.png" - _plot_bar_rows( - size_path, - display, - [ - { - "title": "MelEncoder size (MB)", - "values": mel_size_mb, - "color": "C4", - }, - { - "title": "JointDecision size (MB)", - "values": jd_size_mb, - "color": "C4", - }, - ], - f"{prefix} — Size", - ) - outputs.append(size_path) - - latency_rows = [ - { - "title": "MelEncoder latency (ms)", - "values": mel_latency_ms, - "color": "C1", - }, - { - "title": "JointDecision latency (ms)", - "values": jd_latency_ms, - "color": "C1", - }, - ] - if mel_compile_ms is not None and jd_compile_ms is not None: - latency_rows.extend( - [ - { - "title": "MelEncoder compile (ms)", - "values": mel_compile_ms, - "color": "C3", - }, - { - "title": "JointDecision compile (ms)", - "values": jd_compile_ms, - "color": "C3", - }, - ] - ) - - latency_path = plot_dir / "fused_latency.png" - _plot_bar_rows( - latency_path, - display, - latency_rows, - f"{prefix} — Latency", - ) - outputs.append(latency_path) - - return outputs - - -def _plot_all_component_category_charts( - plot_dir: Path, - labels: List[str], - metrics: Dict[str, Dict[str, List[float]]], - title_suffix: str = "", -) -> List[Path]: - if not HAS_MPL: - return [] - outputs: List[Path] = [] - display = _format_labels(labels) - prefix = f"Component Breakdown — {title_suffix}" if title_suffix else "Component Breakdown" - comp_order = [ - "preprocessor", - "encoder", - "mel_encoder", - "decoder", - "joint", - "joint_decision", - ] - n = len(labels) - - quality_rows: List[Dict[str, object]] = [] - for comp in comp_order: - friendly = _friendly_component_name(comp) - if comp == "joint_decision": - values = _get_metric_series(metrics, comp, "acc", n) - title = f"{friendly} token-id match rate" - else: - values = _get_metric_series(metrics, comp, "quality", n) - title = f"{friendly} quality (1 - norm err)" - quality_rows.append({"title": title, "values": values, "color": "C0", "ylim": (0.0, 1.05)}) - - compression_rows: List[Dict[str, object]] = [] - for comp in comp_order: - friendly = _friendly_component_name(comp) - compression_rows.append( - { - "title": f"{friendly} compression ratio", - "values": _get_metric_series(metrics, comp, "compression", n), - "color": "C2", - } - ) - - size_rows: List[Dict[str, object]] = [] - for comp in comp_order: - friendly = _friendly_component_name(comp) - size_rows.append( - { - "title": f"{friendly} size (MB)", - "values": _get_metric_series(metrics, comp, "size_mb", n), - "color": "C4", - } - ) - - latency_rows: List[Dict[str, object]] = [] - for comp in comp_order: - friendly = _friendly_component_name(comp) - latency_rows.append( - { - "title": f"{friendly} latency (ms)", - "values": _get_metric_series(metrics, comp, "latency_ms", n), - "color": "C1", - } - ) - - compile_rows: List[Dict[str, object]] = [] - for comp in comp_order: - friendly = _friendly_component_name(comp) - compile_rows.append( - { - "title": f"{friendly} compile (ms)", - "values": _get_metric_series(metrics, comp, "compile_ms", n), - "color": "C3", - } - ) - - quality_path = plot_dir / "all_components_quality.png" - _plot_bar_rows(quality_path, display, quality_rows, f"{prefix} — Quality") - outputs.append(quality_path) - - compression_path = plot_dir / "all_components_compression.png" - _plot_bar_rows(compression_path, display, compression_rows, f"{prefix} — Compression") - outputs.append(compression_path) - - size_path = plot_dir / "all_components_size.png" - _plot_bar_rows(size_path, display, size_rows, f"{prefix} — Size") - outputs.append(size_path) - - latency_path = plot_dir / "all_components_latency.png" - _plot_bar_rows(latency_path, display, latency_rows, f"{prefix} — Latency") - outputs.append(latency_path) - - compile_path = plot_dir / "all_components_compile.png" - _plot_bar_rows(compile_path, display, compile_rows, f"{prefix} — Compile") - outputs.append(compile_path) - - return outputs - - -app = typer.Typer(add_completion=False, pretty_exceptions_show_locals=False) - - -@dataclass -class VariantConfig: - name: str - # List of (op_kind, opt_config) steps applied in order; op_kind in {'linear','palettize','prune'} - steps: List[Tuple[str, OptimizationConfig]] - category: str - whitelist: Optional[List[str]] = None # component names to apply; others copied - - -def _dir_size_bytes(path: Path) -> int: - total = 0 - for p in path.rglob("*"): - if p.is_file(): - try: - total += p.stat().st_size - except OSError: - pass - return total - - -def _load_metadata(input_dir: Path) -> Dict[str, object]: - meta_path = input_dir / "metadata.json" - if not meta_path.exists(): - raise typer.BadParameter(f"Expected metadata.json in {input_dir}") - return json.loads(meta_path.read_text()) - - -def _prepare_audio( - seconds: float, - sample_rate: int, - audio_path: Optional[Path], -) -> Tuple[np.ndarray, np.ndarray]: - max_samples = int(round(seconds * sample_rate)) - if audio_path is None: - # random but deterministic sample - rng = np.random.default_rng(1234) - audio = rng.standard_normal(size=(1, max_samples), dtype=np.float32).astype(np.float32) - else: - data, sr = sf.read(str(audio_path), dtype="float32") - if sr != sample_rate: - raise typer.BadParameter( - f"Validation audio sample rate {sr} != expected {sample_rate}" - ) - if data.ndim > 1: - data = data[:, 0] - if data.size < max_samples: - data = np.pad(data, (0, max_samples - data.size)) - elif data.size > max_samples: - data = data[:max_samples] - audio = data.reshape(1, -1).astype(np.float32, copy=False) - length = np.array([max_samples], dtype=np.int32) - return audio, length - - -def _predict_latency(model: ct.models.MLModel, inputs: Dict[str, np.ndarray], runs: int = 10, warmup: int = 3) -> Tuple[float, float]: - # Warmup - for _ in range(max(0, warmup)): - _ = model.predict(inputs) - # Timed runs - times: List[float] = [] - for _ in range(max(1, runs)): - t0 = time.perf_counter() - _ = model.predict(inputs) - t1 = time.perf_counter() - times.append((t1 - t0) * 1000.0) # ms - arr = np.array(times, dtype=np.float64) - return float(arr.mean()), float(arr.std(ddof=1) if arr.size > 1 else 0.0) - - -def _max_abs_rel(a: np.ndarray, b: np.ndarray) -> Tuple[float, float]: - a = np.asarray(a, dtype=np.float32) - b = np.asarray(b, dtype=np.float32) - if a.size == 0: - return 0.0, 0.0 - diff = np.abs(a - b) - max_abs = float(diff.max()) - denom = np.maximum(np.abs(a), np.abs(b)) - with np.errstate(divide="ignore", invalid="ignore"): - rel = np.where(denom == 0.0, 0.0, diff / denom) - max_rel = float(rel.max()) - return max_abs, max_rel - - -def _save_mlpackage(model: ct.models.MLModel, path: Path, description: str) -> None: - path.parent.mkdir(parents=True, exist_ok=True) - # Ensure iOS 17 target for proper MLProgram ops (e.g., blockwise shift/scale) - try: - model.minimum_deployment_target = ct.target.iOS17 - except Exception: - pass - model.short_description = description - model.author = "Fluid Inference" - model.save(str(path)) - - -def _offline_compile_time_ms(model_path: Path) -> float: - """Compile mlpackage -> mlmodelc and return wall time in ms (host offline compile). - - Returns NaN on failure. - """ - compiled_dir: Optional[Path] = None - expected_dir = model_path.with_suffix(".mlmodelc") - try: - # Delete any prior compile artifact in-place so we can measure a fresh build - if expected_dir.exists(): - shutil.rmtree(expected_dir, ignore_errors=True) - - t0 = time.perf_counter() - compiled_path = ct.utils.compile_model(str(model_path)) - t1 = time.perf_counter() - compiled_dir = Path(compiled_path) - return (t1 - t0) * 1000.0 - except Exception: - return float("nan") - finally: - if ( - compiled_dir is not None - and compiled_dir.exists() - and compiled_dir == expected_dir - ): - shutil.rmtree(compiled_dir, ignore_errors=True) - -def _chip_spec_string(compute_units: str) -> str: - try: - chip = subprocess.check_output(["sysctl", "-n", "machdep.cpu.brand_string"]).decode().strip() - except Exception: - chip = platform.processor() or platform.machine() - mac_ver = platform.mac_ver()[0] or platform.platform() - return f"Host: {chip} ‱ macOS {mac_ver} ‱ CoreMLTools {ct.__version__} ‱ ComputeUnits={compute_units} ‱ Min Target: iOS17" - - -def _quantize_dir( - input_dir: Path, - output_dir: Path, - variant: VariantConfig, - global_whitelist: Optional[Set[str]] = None, -) -> Dict[str, str]: - """Quantize all mlpackages in input_dir into output_dir using given variant config. - - Returns a map of component name -> saved relative path. - """ - meta = _load_metadata(input_dir) - comps = meta.get("components", {}) - saved: Dict[str, str] = {} - for name, cfg in comps.items(): - src_name = cfg.get("path") - if not src_name: - continue - src_path = input_dir / src_name - if not src_path.exists(): - continue - dst_path = output_dir / src_name - # Use CPU+GPU for preprocessor to avoid NE preprocessor input size issues; others use CPU+NE - cu = ct.ComputeUnit.CPU_AND_GPU if name == "preprocessor" else ct.ComputeUnit.CPU_AND_NE - base_model = ct.models.MLModel(str(src_path), compute_units=cu) - # Target iOS17 when running optimizations so the right ops are chosen - try: - base_model.minimum_deployment_target = ct.target.iOS17 - except Exception: - pass - - if name == "decoder": - typer.echo(f"[{variant.name}] Skipping decoder quantization; copying baseline: {src_name}") - _save_mlpackage(base_model, dst_path, "Baseline copy (decoder quantization disabled) - decoder") - saved[name] = dst_path.name - continue - - skip_reasons: List[str] = [] - if variant.whitelist is not None and name not in variant.whitelist: - skip_reasons.append("not targeted by variant") - if global_whitelist is not None and name not in global_whitelist: - skip_reasons.append("not in requested components") - - if skip_reasons: - reason = "; ".join(skip_reasons) - typer.echo(f"[{variant.name}] Skipping {name} ({reason}); copying baseline: {src_name}") - _save_mlpackage(base_model, dst_path, f"Baseline copy ({reason}) - {name}") - saved[name] = dst_path.name - continue - - typer.echo(f"[{variant.name}] Quantizing {name}: {src_name}") - - try: - q_model = base_model - for step_kind, step_cfg in variant.steps: - if step_kind == 'linear': - q_model = linear_quantize_weights(q_model, step_cfg) - elif step_kind == 'palettize': - q_model = palettize_weights(q_model, step_cfg) - elif step_kind == 'prune': - q_model = prune_weights(q_model, step_cfg) - else: - raise ValueError(f"Unknown variant step: {step_kind}") - except Exception as e: - # If quantization fails (e.g., unsupported op), fall back to copying baseline. - typer.echo(f" ! Failed to quantize {name} with {variant.name}: {e}. Copying baseline.") - _save_mlpackage(base_model, dst_path, f"Baseline copy (failed to quantize) - {name}") - else: - _save_mlpackage(q_model, dst_path, f"{variant.name} quantized - {name}") - saved[name] = dst_path.name - # Persist a variant metadata shim - out_meta = { - "variant": variant.name, - "base_dir": str(input_dir.resolve()), - "components": saved, - } - (output_dir / "quantization_metadata.json").write_text(json.dumps(out_meta, indent=2)) - return saved - - -@app.command() -def quantize( - input_dir: Path = typer.Option(Path("parakeet_coreml"), help="Directory containing baseline mlpackages + metadata.json"), - output_root: Path = typer.Option(Path("parakeet_coreml_quantized"), help="Root output dir for quantized variants"), - validation_audio: Optional[Path] = typer.Option(None, exists=True, resolve_path=True, help="Optional 15s, 16kHz wav for evaluation (defaults to bundled audio if present)"), - compute_units: str = typer.Option("CPU_AND_NE", help="Compute units for evaluation of non-preprocessor models. Preprocessor is forced to CPU_AND_GPU."), - runs: int = typer.Option(10, help="Timed runs per model for latency measurement"), - categories: Optional[List[str]] = typer.Option( - None, - "--category", - "-c", - help="Only run quantization variants in these categories (e.g., linear, mel-palettize). Can be repeated.", - ), - components: Optional[List[str]] = typer.Option( - None, - "--component", - "-m", - help="Component names to quantize. Defaults to mel_encoder and joint_decision. Use 'all' to keep every component enabled.", - ), -) -> None: - """Quantize models, then compare quality, compression, latency, and compile time. - - Variants include int8-linear (per-channel/per-tensor/block), palettization (6-bit), - jd-only probes, and prune+int8. Baseline is the pre-converted models. - """ - meta = _load_metadata(input_dir) - sr = int(meta.get("sample_rate", 16000)) - seconds = float(meta.get("max_audio_seconds", 15.0)) - - components_meta = meta.get("components", {}) - component_lookup = {name.lower(): name for name in components_meta.keys()} - # Defer computing component_filter until after variant/category selection so - # that, by default, we quantize exactly the components targeted by the - # selected variants (instead of a hard-coded subset). - component_filter: Optional[Set[str]] = None - - # Default audio if present - default_audio = (BASE_DIR / "audio" / "yc_first_minute_16k_15s.wav").resolve() - audio_path = validation_audio if validation_audio is not None else (default_audio if default_audio.exists() else None) - if audio_path is not None and validation_audio is None: - typer.echo(f"Using default validation audio: {audio_path}") - audio, audio_len = _prepare_audio(seconds, sr, audio_path) - - # Load baseline models and helpers for inputs - # Baseline models for input preparation - # Force preprocessor to CPU+GPU and all other components to CPU+NE for evaluation - pre_base = ct.models.MLModel(str(input_dir / "parakeet_preprocessor.mlpackage"), compute_units=ct.ComputeUnit.CPU_AND_GPU) - enc_base = ct.models.MLModel(str(input_dir / "parakeet_encoder.mlpackage"), compute_units=ct.ComputeUnit.CPU_AND_NE) - mel_encoder_base = ct.models.MLModel(str(input_dir / "parakeet_mel_encoder.mlpackage"), compute_units=ct.ComputeUnit.CPU_AND_NE) - decoder_base = ct.models.MLModel(str(input_dir / "parakeet_decoder.mlpackage"), compute_units=ct.ComputeUnit.CPU_AND_NE) - joint_decision_base = ct.models.MLModel(str(input_dir / "parakeet_joint_decision.mlpackage"), compute_units=ct.ComputeUnit.CPU_AND_NE) - joint_base = ct.models.MLModel(str(input_dir / "parakeet_joint.mlpackage"), compute_units=ct.ComputeUnit.CPU_AND_NE) - - # Prepare typical inputs once using baseline models - pre_out = pre_base.predict({"audio_signal": audio, "audio_length": audio_len}) - mel_ref = np.array(pre_out["mel"], dtype=np.float32, copy=True) - mel_len = np.array(pre_out["mel_length"], dtype=np.int32, copy=True) - - enc_out = enc_base.predict({"mel": mel_ref, "mel_length": mel_len}) - encoder_ref = np.array(enc_out["encoder"], dtype=np.float32, copy=True) - encoder_len = np.array(enc_out["encoder_length"], dtype=np.int32, copy=True) - - # Decoder inputs from metadata - dec_in = meta["components"]["decoder"]["inputs"] - targets_shape = tuple(int(x) for x in dec_in["targets"]) # e.g., (1, 256) - h_shape = tuple(int(x) for x in dec_in["h_in"]) # e.g., (2, 1, 640) - # Use zeros as targets for reproducibility without needing blank idx - targets = np.zeros(targets_shape, dtype=np.int32) - target_length = np.array([targets_shape[1]], dtype=np.int32) - h0 = np.zeros(h_shape, dtype=np.float32) - c0 = np.zeros(h_shape, dtype=np.float32) - dec_out = decoder_base.predict({ - "targets": targets, - "target_length": target_length, - "h_in": h0, - "c_in": c0, - }) - decoder_ref = np.array(dec_out["decoder"], dtype=np.float32, copy=True) - - # Baseline sizes per component - pre_base_size = _dir_size_bytes(input_dir / "parakeet_preprocessor.mlpackage") - enc_base_size = _dir_size_bytes(input_dir / "parakeet_encoder.mlpackage") - mel_base_size = _dir_size_bytes(input_dir / "parakeet_mel_encoder.mlpackage") - dec_base_size = _dir_size_bytes(input_dir / "parakeet_decoder.mlpackage") - joint_base_size = _dir_size_bytes(input_dir / "parakeet_joint.mlpackage") - jd_base_size = _dir_size_bytes(input_dir / "parakeet_joint_decision.mlpackage") - - # Baseline latencies - pre_base_inputs = {"audio_signal": audio, "audio_length": audio_len} - enc_base_inputs = {"mel": mel_ref, "mel_length": mel_len} - mel_base_inputs = {"audio_signal": audio, "audio_length": audio_len} - dec_base_inputs = {"targets": targets, "target_length": target_length, "h_in": h0, "c_in": c0} - joint_base_inputs = {"encoder": encoder_ref, "decoder": decoder_ref} - jd_base_inputs = {"encoder": encoder_ref, "decoder": decoder_ref} - pre_base_ms, _ = _predict_latency(pre_base, pre_base_inputs, runs=runs) - enc_base_ms, _ = _predict_latency(enc_base, enc_base_inputs, runs=runs) - mel_base_ms, _ = _predict_latency(mel_encoder_base, mel_base_inputs, runs=runs) - dec_base_ms, _ = _predict_latency(decoder_base, dec_base_inputs, runs=runs) - joint_base_ms, _ = _predict_latency(joint_base, joint_base_inputs, runs=runs) - jd_base_ms, _ = _predict_latency(joint_decision_base, jd_base_inputs, runs=runs) - # Cache baseline joint logits for comparisons - logits_base = np.array(joint_base.predict(joint_base_inputs)["logits"], dtype=np.float32, copy=True) - - # Variants - variants: List[VariantConfig] = [ - VariantConfig( - name="int8-linear", - steps=[( - "linear", - OptimizationConfig(global_config=OpLinearQuantizerConfig(mode="linear", granularity="per_channel")), - )], - category="linear", - ), - # 6-bit palettization for MelEncoder only - VariantConfig( - name="mel6bit-palettize", - steps=[( - "palettize", - OptimizationConfig( - global_config=OpPalettizerConfig(mode="kmeans", nbits=6) - ), - )], - category="mel-palettize", - whitelist=["mel_encoder"], - ), - # 6-bit palettization for Encoder only - VariantConfig( - name="enc6bit-palettize", - steps=[( - "palettize", - OptimizationConfig( - global_config=OpPalettizerConfig(mode="kmeans", nbits=6) - ), - )], - category="encoder-palettize", - whitelist=["encoder"], - ), - # (removed) Global palettization variants - ] - - available_categories = {variant.category for variant in variants} - category_lookup = {cat.lower(): cat for cat in available_categories} - selected_categories: Set[str] - if categories: - normalized_categories = [cat.strip().lower() for cat in categories if cat.strip()] - if normalized_categories: - invalid_categories = [cat for cat in normalized_categories if cat not in category_lookup] - if invalid_categories: - available_str = ", ".join(sorted(available_categories)) or "none" - bad = ", ".join(sorted(set(invalid_categories))) - raise typer.BadParameter( - f"Unknown category (--category): {bad}. Available categories: {available_str}." - ) - selected_categories = {category_lookup[cat] for cat in normalized_categories} - variants = [variant for variant in variants if variant.category in selected_categories] - else: - selected_categories = available_categories - else: - selected_categories = available_categories - - if not variants: - typer.echo("No quantization variants matched the requested categories; nothing to do.") - raise typer.Exit(code=0) - - typer.echo("Running variant categories: " + ", ".join(sorted(selected_categories))) - - # Resolve the component whitelist now that variants are known. - if components: - normalized_components = [comp.strip().lower() for comp in components if comp.strip()] - if any(comp == "all" for comp in normalized_components): - component_filter = None - else: - resolved: List[str] = [] - invalid: List[str] = [] - for comp in normalized_components: - match = component_lookup.get(comp) - if match is None: - invalid.append(comp) - else: - resolved.append(match) - if invalid: - available = ", ".join(sorted(component_lookup.values())) or "none" - bad = ", ".join(sorted(set(invalid))) - raise typer.BadParameter( - f"Unknown component(s) for --component: {bad}. Available components: {available}." - ) - component_filter = set(resolved) - else: - # Derive from selected variants' whitelists - derived: Set[str] = set() - has_global = False - for v in variants: - if v.whitelist is None: - has_global = True - break - derived.update(v.whitelist) - component_filter = None if has_global or not derived else derived - - if component_filter is None: - typer.echo("Quantizing components: all components (derived)") - else: - typer.echo("Quantizing components: " + ", ".join(sorted(component_filter)) + " (derived)") - - # Aggregate results (baseline + variants) - summary: Dict[str, Dict[str, object]] = {} - variants_names: List[str] = [] - # Build arrays including baseline as first label - fused_labels: List[str] = ["baseline"] - mel_quality_scores: List[float] = [1.0] - mel_latency_ms: List[float] = [mel_base_ms] - mel_compression: List[float] = [1.0] - mel_size_mb: List[float] = [float(mel_base_size) / BYTES_IN_MB] - # Offline compile time (host) for fused models - mel_compile_ms: List[float] = [_offline_compile_time_ms(input_dir / "parakeet_mel_encoder.mlpackage")] - jd_accuracy: List[float] = [1.0] - jd_latency_ms: List[float] = [jd_base_ms] - jd_compression: List[float] = [1.0] - jd_size_mb: List[float] = [float(jd_base_size) / BYTES_IN_MB] - jd_compile_ms: List[float] = [_offline_compile_time_ms(input_dir / "parakeet_joint_decision.mlpackage")] - - # For the all-components chart, collect per-component metrics similarly - all_metrics: Dict[str, Dict[str, List[float]]] = { - "preprocessor": { - "quality": [1.0], - "compression": [1.0], - "latency_ms": [pre_base_ms], - "compile_ms": [_offline_compile_time_ms(input_dir / "parakeet_preprocessor.mlpackage")], - "size_mb": [float(pre_base_size) / BYTES_IN_MB], - }, - "encoder": { - "quality": [1.0], - "compression": [1.0], - "latency_ms": [enc_base_ms], - "compile_ms": [_offline_compile_time_ms(input_dir / "parakeet_encoder.mlpackage")], - "size_mb": [float(enc_base_size) / BYTES_IN_MB], - }, - "mel_encoder": { - "quality": [1.0], - "compression": [1.0], - "latency_ms": [mel_base_ms], - "compile_ms": [mel_compile_ms[0]], - "size_mb": [float(mel_base_size) / BYTES_IN_MB], - }, - "decoder": { - "quality": [1.0], - "compression": [1.0], - "latency_ms": [dec_base_ms], - "compile_ms": [_offline_compile_time_ms(input_dir / "parakeet_decoder.mlpackage")], - "size_mb": [float(dec_base_size) / BYTES_IN_MB], - }, - "joint": { - "quality": [1.0], - "compression": [1.0], - "latency_ms": [joint_base_ms], - "compile_ms": [_offline_compile_time_ms(input_dir / "parakeet_joint.mlpackage")], - "size_mb": [float(joint_base_size) / BYTES_IN_MB], - }, - "joint_decision": { - "acc": [1.0], - "compression": [1.0], - "latency_ms": [jd_base_ms], - "compile_ms": [jd_compile_ms[0]], - "size_mb": [float(jd_base_size) / BYTES_IN_MB], - }, - } - - # Populate baseline entry - summary["baseline"] = { - "components": { - "preprocessor": { - "quality": 1.0, - "latency_ms": pre_base_ms, - "size_bytes": float(pre_base_size), - "size_mb": float(pre_base_size) / BYTES_IN_MB, - "compression_ratio": 1.0, - "compile_ms": all_metrics["preprocessor"]["compile_ms"][0], - }, - "encoder": { - "quality": 1.0, - "latency_ms": enc_base_ms, - "size_bytes": float(enc_base_size), - "size_mb": float(enc_base_size) / BYTES_IN_MB, - "compression_ratio": 1.0, - "compile_ms": all_metrics["encoder"]["compile_ms"][0], - }, - "mel_encoder": { - "quality": 1.0, - "latency_ms": mel_base_ms, - "size_bytes": float(mel_base_size), - "size_mb": float(mel_base_size) / BYTES_IN_MB, - "compression_ratio": 1.0, - "compile_ms": all_metrics["mel_encoder"]["compile_ms"][0], - }, - "decoder": { - "quality": 1.0, - "latency_ms": dec_base_ms, - "size_bytes": float(dec_base_size), - "size_mb": float(dec_base_size) / BYTES_IN_MB, - "compression_ratio": 1.0, - "compile_ms": all_metrics["decoder"]["compile_ms"][0], - }, - "joint": { - "quality": 1.0, - "latency_ms": joint_base_ms, - "size_bytes": float(joint_base_size), - "size_mb": float(joint_base_size) / BYTES_IN_MB, - "compression_ratio": 1.0, - "compile_ms": all_metrics["joint"]["compile_ms"][0], - }, - "joint_decision": { - "acc": 1.0, - "latency_ms": jd_base_ms, - "size_bytes": float(jd_base_size), - "size_mb": float(jd_base_size) / BYTES_IN_MB, - "compression_ratio": 1.0, - "compile_ms": all_metrics["joint_decision"]["compile_ms"][0], - }, - } - } - - for var in variants: - variants_names.append(var.name) - out_dir = output_root / var.name - out_dir_exists = out_dir.exists() - out_dir.mkdir(parents=True, exist_ok=True) - - expected_components = [] - for comp_cfg in meta.get("components", {}).values(): - rel = comp_cfg.get("path") - if rel: - expected_components.append(out_dir / rel) - - missing = [p for p in expected_components if not p.exists()] - if out_dir_exists and not missing: - typer.echo(f"[{var.name}] Output already present at {out_dir}; skipping quantization step.") - else: - if out_dir_exists and missing: - missing_names = ", ".join(sorted(p.name for p in missing)) or "unknown" - typer.echo(f"[{var.name}] Output directory exists but is incomplete (missing: {missing_names}). Re-quantizing.") - shutil.rmtree(out_dir, ignore_errors=True) - out_dir.mkdir(parents=True, exist_ok=True) - _quantize_dir(input_dir, out_dir, var, component_filter) - - # Load quantized models and pre-compute offline compile time - pre_q_path = out_dir / "parakeet_preprocessor.mlpackage" - enc_q_path = out_dir / "parakeet_encoder.mlpackage" - mel_q_path = out_dir / "parakeet_mel_encoder.mlpackage" - dec_q_path = out_dir / "parakeet_decoder.mlpackage" - joint_q_path = out_dir / "parakeet_joint.mlpackage" - jd_q_path = out_dir / "parakeet_joint_decision.mlpackage" - - pre_compile_ms = _offline_compile_time_ms(pre_q_path) - enc_compile_ms = _offline_compile_time_ms(enc_q_path) - mel_compile_q_ms = _offline_compile_time_ms(mel_q_path) - dec_compile_ms = _offline_compile_time_ms(dec_q_path) - joint_compile_ms = _offline_compile_time_ms(joint_q_path) - jd_compile_q_ms = _offline_compile_time_ms(jd_q_path) - - # Match compute units for quantized artifacts: preprocessor on CPU+GPU; others on CPU+NE - pre_q = ct.models.MLModel(str(pre_q_path), compute_units=ct.ComputeUnit.CPU_AND_GPU) - enc_q = ct.models.MLModel(str(enc_q_path), compute_units=ct.ComputeUnit.CPU_AND_NE) - mel_q = ct.models.MLModel(str(mel_q_path), compute_units=ct.ComputeUnit.CPU_AND_NE) - dec_q = ct.models.MLModel(str(dec_q_path), compute_units=ct.ComputeUnit.CPU_AND_NE) - joint_q = ct.models.MLModel(str(joint_q_path), compute_units=ct.ComputeUnit.CPU_AND_NE) - jd_q = ct.models.MLModel(str(jd_q_path), compute_units=ct.ComputeUnit.CPU_AND_NE) - - # Preprocessor quality vs baseline - pre_q_out = pre_q.predict(pre_base_inputs) - mel_q_in = np.array(pre_q_out["mel"], dtype=np.float32, copy=True) - a_pre, r_pre = _max_abs_rel(mel_ref, mel_q_in) - l2_pre_ref = float(np.linalg.norm(mel_ref)) - l2_pre_err = float(np.linalg.norm(mel_ref - mel_q_in)) - pre_norm_err = (l2_pre_err / (l2_pre_ref + 1e-8)) if l2_pre_ref > 0 else 0.0 - pre_quality = float(max(0.0, 1.0 - pre_norm_err)) - pre_q_ms, _ = _predict_latency(pre_q, pre_base_inputs, runs=runs) - pre_q_size = _dir_size_bytes(out_dir / "parakeet_preprocessor.mlpackage") - all_metrics["preprocessor"]["quality"].append(pre_quality) - all_metrics["preprocessor"]["latency_ms"].append(pre_q_ms) - all_metrics["preprocessor"]["compression"].append(float(pre_base_size) / float(pre_q_size if pre_q_size > 0 else 1)) - all_metrics["preprocessor"].setdefault("compile_ms", []).append(pre_compile_ms) - all_metrics["preprocessor"]["size_mb"].append(float(pre_q_size) / BYTES_IN_MB) - - # Encoder quality vs baseline (feed baseline mel to both) - enc_q_out = enc_q.predict({"mel": mel_ref, "mel_length": mel_len}) - encoder_q = np.array(enc_q_out["encoder"], dtype=np.float32, copy=True) - l2_enc_ref = float(np.linalg.norm(encoder_ref)) - l2_enc_err = float(np.linalg.norm(encoder_ref - encoder_q)) - enc_norm_err = (l2_enc_err / (l2_enc_ref + 1e-8)) if l2_enc_ref > 0 else 0.0 - enc_quality = float(max(0.0, 1.0 - enc_norm_err)) - enc_q_ms, _ = _predict_latency(enc_q, enc_base_inputs, runs=runs) - enc_q_size = _dir_size_bytes(out_dir / "parakeet_encoder.mlpackage") - all_metrics["encoder"]["quality"].append(enc_quality) - all_metrics["encoder"]["latency_ms"].append(enc_q_ms) - all_metrics["encoder"]["compression"].append(float(enc_base_size) / float(enc_q_size if enc_q_size > 0 else 1)) - all_metrics["encoder"].setdefault("compile_ms", []).append(enc_compile_ms) - all_metrics["encoder"]["size_mb"].append(float(enc_q_size) / BYTES_IN_MB) - - # MelEncoder quality - mel_q_out = mel_q.predict(mel_base_inputs) - enc_q_fused = np.array(mel_q_out["encoder"], dtype=np.float32, copy=True) - a_mel, r_mel = _max_abs_rel(encoder_ref, enc_q_fused) - # Normalize error into a [0,1] quality score: 1 / (1 + normalized L2) - # Use relative measure derived from L2 norms as more stable than max. - l2_ref = float(np.linalg.norm(encoder_ref)) - l2_err = float(np.linalg.norm(encoder_ref - enc_q_fused)) - norm_err = (l2_err / (l2_ref + 1e-8)) if l2_ref > 0 else 0.0 - mel_quality = float(max(0.0, 1.0 - norm_err)) - mel_q_ms, _ = _predict_latency(mel_q, mel_base_inputs, runs=runs) - mel_q_size = _dir_size_bytes(out_dir / "parakeet_mel_encoder.mlpackage") - mel_ratio = float(mel_base_size) / float(mel_q_size if mel_q_size > 0 else 1) - mel_size_mb.append(float(mel_q_size) / BYTES_IN_MB) - all_metrics["mel_encoder"]["quality"].append(mel_quality) - all_metrics["mel_encoder"]["latency_ms"].append(mel_q_ms) - all_metrics["mel_encoder"]["compression"].append(float(mel_base_size) / float(mel_q_size if mel_q_size > 0 else 1)) - all_metrics["mel_encoder"].setdefault("compile_ms", []).append(mel_compile_q_ms) - all_metrics["mel_encoder"]["size_mb"].append(float(mel_q_size) / BYTES_IN_MB) - - # JointDecision quality: token-id and duration match rates - jd_base_out = joint_decision_base.predict(jd_base_inputs) - token_id_base = np.array(jd_base_out["token_id"], dtype=np.int32, copy=True) - duration_base = np.array(jd_base_out["duration"], dtype=np.int32, copy=True) - token_prob_base = np.array(jd_base_out["token_prob"], dtype=np.float32, copy=True) - - jd_q_out = jd_q.predict(jd_base_inputs) - token_id_q = np.array(jd_q_out["token_id"], dtype=np.int32, copy=True) - duration_q = np.array(jd_q_out["duration"], dtype=np.int32, copy=True) - token_prob_q = np.array(jd_q_out["token_prob"], dtype=np.float32, copy=True) - - # Accuracy metrics - id_match = float((token_id_q == token_id_base).mean()) - dur_match = float((duration_q == duration_base).mean()) - # Aggregate a single "accuracy" number as token-id match rate (primary) - jd_acc = id_match - jd_q_ms, _ = _predict_latency(jd_q, jd_base_inputs, runs=runs) - jd_q_size = _dir_size_bytes(out_dir / "parakeet_joint_decision.mlpackage") - jd_ratio = float(jd_base_size) / float(jd_q_size if jd_q_size > 0 else 1) - jd_size_mb.append(float(jd_q_size) / BYTES_IN_MB) - all_metrics["joint_decision"].setdefault("acc", []).append(jd_acc) - all_metrics["joint_decision"]["latency_ms"].append(jd_q_ms) - all_metrics["joint_decision"]["compression"].append(float(jd_base_size) / float(jd_q_size if jd_q_size > 0 else 1)) - all_metrics["joint_decision"].setdefault("compile_ms", []).append(jd_compile_q_ms) - all_metrics["joint_decision"]["size_mb"].append(float(jd_q_size) / BYTES_IN_MB) - - # Decoder quality vs baseline - dec_q_out = dec_q.predict(dec_base_inputs) - decoder_q = np.array(dec_q_out["decoder"], dtype=np.float32, copy=True) - l2_dec_ref = float(np.linalg.norm(decoder_ref)) - l2_dec_err = float(np.linalg.norm(decoder_ref - decoder_q)) - dec_norm_err = (l2_dec_err / (l2_dec_ref + 1e-8)) if l2_dec_ref > 0 else 0.0 - dec_quality = float(max(0.0, 1.0 - dec_norm_err)) - dec_q_ms, _ = _predict_latency(dec_q, dec_base_inputs, runs=runs) - dec_q_size = _dir_size_bytes(out_dir / "parakeet_decoder.mlpackage") - all_metrics["decoder"]["quality"].append(dec_quality) - all_metrics["decoder"]["latency_ms"].append(dec_q_ms) - all_metrics["decoder"]["compression"].append(float(dec_base_size) / float(dec_q_size if dec_q_size > 0 else 1)) - all_metrics["decoder"].setdefault("compile_ms", []).append(dec_compile_ms) - all_metrics["decoder"]["size_mb"].append(float(dec_q_size) / BYTES_IN_MB) - - # Joint quality vs baseline (compare logits) - joint_q_out = joint_q.predict(joint_base_inputs) - logits_q = np.array(joint_q_out["logits"], dtype=np.float32, copy=True) - l2_joint_ref = float(np.linalg.norm(logits_base)) - l2_joint_err = float(np.linalg.norm(logits_base - logits_q)) - joint_norm_err = (l2_joint_err / (l2_joint_ref + 1e-8)) if l2_joint_ref > 0 else 0.0 - joint_quality = float(max(0.0, 1.0 - joint_norm_err)) - joint_q_ms, _ = _predict_latency(joint_q, joint_base_inputs, runs=runs) - joint_q_size = _dir_size_bytes(out_dir / "parakeet_joint.mlpackage") - all_metrics["joint"]["quality"].append(joint_quality) - all_metrics["joint"]["latency_ms"].append(joint_q_ms) - all_metrics["joint"]["compression"].append(float(joint_base_size) / float(joint_q_size if joint_q_size > 0 else 1)) - all_metrics["joint"].setdefault("compile_ms", []).append(joint_compile_ms) - all_metrics["joint"]["size_mb"].append(float(joint_q_size) / BYTES_IN_MB) - - # Decoder deltas for JSON - a_dec, r_dec = _max_abs_rel(decoder_ref, decoder_q) - # Joint deltas for JSON - a_joint, r_joint = _max_abs_rel(logits_base, logits_q) - - # Store metrics - summary[var.name] = { - "components": { - "preprocessor": { - "quality": pre_quality, - "latency_ms": pre_q_ms, - "size_bytes": float(pre_q_size), - "size_mb": float(pre_q_size) / BYTES_IN_MB, - "compression_ratio": float(pre_base_size) / float(pre_q_size if pre_q_size > 0 else 1), - "max_abs": a_pre, - "max_rel": r_pre, - "compile_ms": pre_compile_ms, - }, - "encoder": { - "quality": enc_quality, - "latency_ms": enc_q_ms, - "size_bytes": float(enc_q_size), - "size_mb": float(enc_q_size) / BYTES_IN_MB, - "compression_ratio": float(enc_base_size) / float(enc_q_size if enc_q_size > 0 else 1), - "compile_ms": enc_compile_ms, - }, - "mel_encoder": { - "quality": mel_quality, - "latency_ms": mel_q_ms, - "size_bytes": float(mel_q_size), - "size_mb": float(mel_q_size) / BYTES_IN_MB, - "compression_ratio": mel_ratio, - "max_abs": a_mel, - "max_rel": r_mel, - "compile_ms": mel_compile_q_ms, - }, - "decoder": { - "quality": dec_quality, - "latency_ms": dec_q_ms, - "size_bytes": float(dec_q_size), - "size_mb": float(dec_q_size) / BYTES_IN_MB, - "compression_ratio": float(dec_base_size) / float(dec_q_size if dec_q_size > 0 else 1), - "max_abs": a_dec, - "max_rel": r_dec, - "compile_ms": dec_compile_ms, - }, - "joint": { - "quality": joint_quality, - "latency_ms": joint_q_ms, - "size_bytes": float(joint_q_size), - "size_mb": float(joint_q_size) / BYTES_IN_MB, - "compression_ratio": float(joint_base_size) / float(joint_q_size if joint_q_size > 0 else 1), - "max_abs": a_joint, - "max_rel": r_joint, - "compile_ms": joint_compile_ms, - }, - "joint_decision": { - "acc": jd_acc, - "duration_match": dur_match, - "prob_mae": float(np.mean(np.abs(token_prob_q - token_prob_base))), - "latency_ms": jd_q_ms, - "size_bytes": float(jd_q_size), - "size_mb": float(jd_q_size) / BYTES_IN_MB, - "compression_ratio": jd_ratio, - "compile_ms": jd_compile_q_ms, - }, - } - } - - fused_labels.append(var.name) - mel_quality_scores.append(mel_quality) - mel_latency_ms.append(mel_q_ms) - mel_compression.append(mel_ratio) - jd_accuracy.append(jd_acc) - jd_latency_ms.append(jd_q_ms) - jd_compression.append(jd_ratio) - mel_compile_ms.append(mel_compile_q_ms) - jd_compile_ms.append(jd_compile_q_ms) - - # Write summary JSON - out_root = output_root - out_root.mkdir(parents=True, exist_ok=True) - (out_root / "quantization_summary.json").write_text(json.dumps(summary, indent=2)) - - # Plot - plot_dir = out_root / "plots" - title_suffix = _chip_spec_string(compute_units) - fused_paths = _plot_fused_category_charts( - plot_dir, - fused_labels, - mel_quality_scores, - mel_latency_ms, - mel_compression, - mel_size_mb, - jd_accuracy, - jd_latency_ms, - jd_compression, - jd_size_mb, - mel_compile_ms, - jd_compile_ms, - title_suffix, - ) - component_paths = _plot_all_component_category_charts( - plot_dir, - fused_labels, - all_metrics, - title_suffix, - ) - - typer.echo(f"Wrote summary JSON: {out_root / 'quantization_summary.json'}") - if HAS_MPL: - all_plot_paths = fused_paths + component_paths - repo_plot_dir = BASE_DIR / "plots" / "quantize" / compute_units.lower() - repo_plot_dir.mkdir(parents=True, exist_ok=True) - for path in all_plot_paths: - typer.echo(f"Wrote plot: {path}") - if path.exists(): - dest = repo_plot_dir / path.name - shutil.copy2(path, dest) - typer.echo(f"Mirrored plot: {dest}") - else: - typer.echo("matplotlib unavailable; skipped plotting.") - - -if __name__ == "__main__": - app() diff --git a/parakeet-tdt-v2-0.6b/coreml/speech_to_text_streaming_infer_rnnt.py b/parakeet-tdt-v2-0.6b/coreml/speech_to_text_streaming_infer_rnnt.py deleted file mode 100644 index 07d8fcc04df44b41c758ad52f518d36a1ad797ae..0000000000000000000000000000000000000000 --- a/parakeet-tdt-v2-0.6b/coreml/speech_to_text_streaming_infer_rnnt.py +++ /dev/null @@ -1,341 +0,0 @@ -"""Streaming inference helper that stitches Parakeet CoreML components using the RNNT greedy loop from Nemo.""" -from __future__ import annotations - -import argparse -import json -import logging -from pathlib import Path -from typing import Iterable, List, Optional, Sequence - -import coremltools as ct -import librosa -import numpy as np -import torch - -from parakeet_components import CoreMLModelBundle - -LOGGER = logging.getLogger("parakeet_streaming") - - -class BatchedHyps: - """Minimal port of Nemo's batched hypothesis buffer.""" - - def __init__( - self, - batch_size: int, - init_length: int, - device: torch.device, - float_dtype: torch.dtype, - ) -> None: - if init_length <= 0: - raise ValueError("init_length must be > 0") - self._max_length = init_length - self.current_lengths = torch.zeros(batch_size, device=device, dtype=torch.long) - self.transcript = torch.zeros((batch_size, self._max_length), device=device, dtype=torch.long) - self.timestamps = torch.zeros((batch_size, self._max_length), device=device, dtype=torch.long) - self.scores = torch.zeros(batch_size, device=device, dtype=float_dtype) - self.last_timestamp = torch.full((batch_size,), -1, device=device, dtype=torch.long) - self.last_timestamp_lasts = torch.zeros(batch_size, device=device, dtype=torch.long) - self._batch_indices = torch.arange(batch_size, device=device) - self._ones = torch.ones_like(self._batch_indices) - - def add_results( - self, - active_mask: torch.Tensor, - labels: torch.Tensor, - time_indices: torch.Tensor, - scores: torch.Tensor, - ) -> None: - self.scores = torch.where(active_mask, self.scores + scores, self.scores) - self.transcript[self._batch_indices, self.current_lengths] = labels - self.timestamps[self._batch_indices, self.current_lengths] = time_indices - torch.where( - torch.logical_and(active_mask, self.last_timestamp == time_indices), - self.last_timestamp_lasts + 1, - self.last_timestamp_lasts, - out=self.last_timestamp_lasts, - ) - torch.where( - torch.logical_and(active_mask, self.last_timestamp != time_indices), - self._ones, - self.last_timestamp_lasts, - out=self.last_timestamp_lasts, - ) - torch.where(active_mask, time_indices, self.last_timestamp, out=self.last_timestamp) - self.current_lengths += active_mask - - -class CoreMLStreamingDecoder: - """Use exported decoder and joint CoreML models with Nemo's greedy RNNT loop.""" - - def __init__( - self, - decoder_model: ct.models.MLModel, - joint_model: ct.models.MLModel, - *, - vocab_size: int, - blank_id: int, - num_layers: int, - hidden_size: int, - durations: Sequence[int] = (0, 1, 2, 3, 4), - max_symbols: int = 10, - device: torch.device = torch.device("cpu"), - ) -> None: - self.decoder_model = decoder_model - self.joint_model = joint_model - self.vocab_size = vocab_size - self.blank_id = blank_id - self.num_layers = num_layers - self.hidden_size = hidden_size - self.durations = torch.tensor(durations, dtype=torch.long, device=device) - self.max_symbols = max_symbols - self.device = device - - def _predict_decoder(self, labels: torch.Tensor, h_in: np.ndarray, c_in: np.ndarray) -> tuple[torch.Tensor, np.ndarray, np.ndarray]: - outputs = self.decoder_model.predict( - { - "targets": np.array([labels.cpu().numpy()], dtype=np.int32), - "target_lengths": np.array([labels.numel()], dtype=np.int32), - "h_in": h_in, - "c_in": c_in, - } - ) - decoder_output = torch.from_numpy(outputs["decoder_output"]).to(self.device) - return decoder_output, outputs["h_out"], outputs["c_out"] - - def _predict_joint(self, encoder_frame: torch.Tensor, decoder_output: torch.Tensor) -> torch.Tensor: - outputs = self.joint_model.predict( - { - "encoder_outputs": encoder_frame.unsqueeze(1).cpu().numpy().astype(np.float32), - "decoder_outputs": decoder_output.cpu().numpy().astype(np.float32), - } - ) - logits = torch.from_numpy(outputs["logits"]).to(self.device) - return logits.squeeze(1).squeeze(1) - - def decode(self, encoder_output: torch.Tensor, encoder_lengths: torch.Tensor) -> List[List[int]]: - batch_size, max_time, _ = encoder_output.shape - encoder_output = encoder_output.to(self.device) - encoder_lengths = encoder_lengths.to(self.device) - - float_dtype = encoder_output.dtype - batch_indices = torch.arange(batch_size, device=self.device) - labels = torch.full((batch_size,), fill_value=self.blank_id, device=self.device, dtype=torch.long) - time_indices = torch.zeros_like(labels) - safe_time_indices = torch.zeros_like(labels) - time_indices_current = torch.zeros_like(labels) - last_timesteps = encoder_lengths - 1 - active_mask = encoder_lengths > 0 - advance_mask = torch.empty_like(active_mask) - active_mask_prev = torch.empty_like(active_mask) - became_inactive = torch.empty_like(active_mask) - - hyps = BatchedHyps( - batch_size=batch_size, - init_length=max_time * self.max_symbols if self.max_symbols else max_time, - device=self.device, - float_dtype=float_dtype, - ) - - h_in = np.zeros((self.num_layers, batch_size, self.hidden_size), dtype=np.float32) - c_in = np.zeros((self.num_layers, batch_size, self.hidden_size), dtype=np.float32) - - while active_mask.any(): - active_mask_prev.copy_(active_mask) - decoder_output, h_in, c_in = self._predict_decoder(labels, h_in, c_in) - logits = self._predict_joint(encoder_output[batch_indices, safe_time_indices], decoder_output) - - scores, labels = logits[:, : self.vocab_size].max(dim=-1) - duration_indices = logits[:, self.vocab_size : self.vocab_size + len(self.durations)].argmax(dim=-1) - durations = self.durations[duration_indices] - - blank_mask = labels == self.blank_id - durations.masked_fill_(torch.logical_and(durations == 0, blank_mask), 1) - time_indices_current.copy_(time_indices) - time_indices += durations - torch.minimum(time_indices, last_timesteps, out=safe_time_indices) - torch.less(time_indices, encoder_lengths, out=active_mask) - torch.logical_and(active_mask, blank_mask, out=advance_mask) - - while advance_mask.any(): - torch.where(advance_mask, time_indices, time_indices_current, out=time_indices_current) - logits = self._predict_joint(encoder_output[batch_indices, safe_time_indices], decoder_output) - more_scores, more_labels = logits[:, : self.vocab_size].max(dim=-1) - labels = torch.where(advance_mask, more_labels, labels) - scores = torch.where(advance_mask, more_scores, scores) - duration_indices = logits[:, self.vocab_size : self.vocab_size + len(self.durations)].argmax(dim=-1) - durations = self.durations[duration_indices] - blank_mask = labels == self.blank_id - durations.masked_fill_(torch.logical_and(durations == 0, blank_mask), 1) - torch.where(advance_mask, time_indices + durations, time_indices, out=time_indices) - torch.minimum(time_indices, last_timesteps, out=safe_time_indices) - torch.less(time_indices, encoder_lengths, out=active_mask) - torch.logical_and(active_mask, blank_mask, out=advance_mask) - - torch.ne(active_mask, active_mask_prev, out=became_inactive) - hyps.add_results(active_mask, labels, time_indices_current, scores) - - if self.max_symbols is not None: - force_blank = torch.logical_and( - active_mask, - torch.logical_and( - torch.logical_and(labels != self.blank_id, hyps.last_timestamp_lasts >= self.max_symbols), - hyps.last_timestamp == time_indices, - ), - ) - time_indices += force_blank - torch.minimum(time_indices, last_timesteps, out=safe_time_indices) - torch.less(time_indices, encoder_lengths, out=active_mask) - - results: List[List[int]] = [] - for hyp in hyps.transcript: - tokens = [int(token) for token in hyp.tolist() if 0 < token < self.vocab_size] - results.append(tokens) - return results - - -class StreamingTranscriber: - def __init__( - self, - bundle: CoreMLModelBundle, - *, - blank_id: Optional[int] = None, - num_layers: int = 2, - hidden_size: int = 640, - durations: Sequence[int] = (0, 1, 2, 3, 4), - ) -> None: - self.preprocessor = ct.models.MLModel(str(bundle.preprocessor), compute_units=ct.ComputeUnit.CPU_ONLY) - self.encoder = ct.models.MLModel(str(bundle.encoder), compute_units=ct.ComputeUnit.CPU_ONLY) - self.decoder = ct.models.MLModel(str(bundle.decoder), compute_units=ct.ComputeUnit.CPU_ONLY) - self.joint = ct.models.MLModel(str(bundle.joint), compute_units=ct.ComputeUnit.CPU_ONLY) - self.tokenizer = self._load_tokenizer(bundle.tokenizer) - - vocab_size = max(self.tokenizer.keys()) + 1 - if blank_id is None: - blank_id = vocab_size - 1 - self.decoder_helper = CoreMLStreamingDecoder( - self.decoder, - self.joint, - vocab_size=vocab_size, - blank_id=blank_id, - num_layers=num_layers, - hidden_size=hidden_size, - durations=durations, - ) - self.blank_id = blank_id - - @staticmethod - def _load_tokenizer(tokenizer_path: Optional[Path]) -> dict[int, str]: - if tokenizer_path is None: - raise ValueError("Tokenizer JSON is required") - with Path(tokenizer_path).open() as f: - data = json.load(f) - return {int(k): v for k, v in data.items()} - - def _tokens_to_text(self, tokens: Iterable[int]) -> str: - pieces: List[str] = [] - for token in tokens: - piece = self.tokenizer.get(token) - if piece is None: - continue - if piece.startswith("▁"): - if pieces: - pieces.append(" ") - pieces.append(piece[1:]) - else: - pieces.append(piece) - return "".join(pieces).strip() - - def _preprocess(self, audio: np.ndarray) -> tuple[np.ndarray, int]: - audio_2d = audio.reshape(1, -1).astype(np.float32) - length = np.array([audio_2d.shape[-1]], dtype=np.int32) - outputs = self.preprocessor.predict({ - "audio_signal": audio_2d, - "audio_length": length, - }) - return outputs["melspectrogram"], int(outputs["melspectrogram_length"][0]) - - def _encode(self, mel: np.ndarray, mel_length: int) -> tuple[torch.Tensor, torch.Tensor]: - outputs = self.encoder.predict({ - "melspectrogram": mel.astype(np.float32), - "melspectrogram_length": np.array([mel_length], dtype=np.int32), - }) - encoder_output = outputs["encoder_output"] - if encoder_output.ndim == 3: - encoder_output = np.transpose(encoder_output, (0, 2, 1)) - length = torch.tensor(outputs["encoder_output_length"], dtype=torch.long) - return torch.from_numpy(encoder_output.astype(np.float32)), length - - def transcribe(self, audio_path: Path) -> str: - audio, _ = librosa.load(str(audio_path), sr=16000) - mel, mel_length = self._preprocess(audio) - encoder_output, encoder_length = self._encode(mel, mel_length) - token_ids = self.decoder_helper.decode(encoder_output, encoder_length)[0] - return self._tokens_to_text(token_ids) - - def transcribe_many(self, audio_paths: Sequence[Path]) -> List[str]: - results: List[str] = [] - for path in audio_paths: - LOGGER.info("Transcribing %s", path) - results.append(self.transcribe(path)) - return results - - -def _resolve_bundle(args: argparse.Namespace) -> CoreMLModelBundle: - base = Path(args.model_dir) if args.model_dir else None - if base is None and not all([args.preprocessor, args.encoder, args.decoder, args.joint, args.tokenizer]): - raise ValueError("Either --model-dir or explicit model paths are required") - return CoreMLModelBundle( - preprocessor=Path(args.preprocessor) if args.preprocessor else base / "Melspectrogram.mlpackage", - encoder=Path(args.encoder) if args.encoder else base / "ParakeetEncoder.mlpackage", - decoder=Path(args.decoder) if args.decoder else base / "ParakeetDecoder.mlpackage", - joint=Path(args.joint) if args.joint else base / "RNNTJoint.mlpackage", - tokenizer=Path(args.tokenizer) if args.tokenizer else base / "tokenizer.json", - ) - - -def _build_parser() -> argparse.ArgumentParser: - parser = argparse.ArgumentParser(description="Streaming RNNT inference with CoreML components") - parser.add_argument("--model-dir", type=Path, help="Directory containing exported CoreML models") - parser.add_argument("--preprocessor", type=Path, help="Path to the preprocessor .mlpackage") - parser.add_argument("--encoder", type=Path, help="Path to the encoder .mlpackage") - parser.add_argument("--decoder", type=Path, help="Path to the decoder .mlpackage") - parser.add_argument("--joint", type=Path, help="Path to the joint .mlpackage") - parser.add_argument("--tokenizer", type=Path, help="Path to tokenizer JSON") - parser.add_argument("audio", nargs="+", help="Audio files to transcribe") - parser.add_argument("--blank-id", type=int, help="Blank token id") - parser.add_argument("--num-layers", type=int, default=2, help="Prediction network layer count") - parser.add_argument("--hidden-size", type=int, default=640, help="Prediction network hidden size") - parser.add_argument("--durations", type=int, nargs="+", default=[0, 1, 2, 3, 4], help="RNNT duration bucket values") - parser.add_argument("--verbose", "-v", action="count", default=0, help="Increase log verbosity") - return parser - - -def _configure_logging(verbosity: int) -> None: - level = logging.WARNING - (10 * verbosity) - logging.basicConfig(level=max(logging.DEBUG, level), format="[%(levelname)s] %(message)s") - - -def main(argv: Optional[Sequence[str]] = None) -> None: - parser = _build_parser() - args = parser.parse_args(argv) - _configure_logging(args.verbose) - - try: - bundle = _resolve_bundle(args) - transcriber = StreamingTranscriber( - bundle, - blank_id=args.blank_id, - num_layers=args.num_layers, - hidden_size=args.hidden_size, - durations=tuple(args.durations), - ) - transcripts = transcriber.transcribe_many([Path(p) for p in args.audio]) - for path, text in zip(args.audio, transcripts): - print(f"{path}: {text}") - except ValueError as exc: - parser.error(str(exc)) - - -if __name__ == "__main__": # pragma: no cover - main() diff --git a/parakeet-tdt-v2-0.6b/coreml/uv.lock b/parakeet-tdt-v2-0.6b/coreml/uv.lock deleted file mode 100644 index 627100c36a840640c5f07947fd14ac1ef80534a0..0000000000000000000000000000000000000000 --- a/parakeet-tdt-v2-0.6b/coreml/uv.lock +++ /dev/null @@ -1,4725 +0,0 @@ -version = 1 -requires-python = "==3.10.12" -resolution-markers = [ - "sys_platform != 'linux'", - "sys_platform == 'linux'", -] - -[[package]] -name = "absl-py" -version = "2.3.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/03/15/18693af986560a5c3cc0b84a8046b536ffb2cdb536e03cce897f2759e284/absl_py-2.3.0.tar.gz", hash = "sha256:d96fda5c884f1b22178852f30ffa85766d50b99e00775ea626c23304f582fc4f", size = 116400 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/87/04/9d75e1d3bb4ab8ec67ff10919476ccdee06c098bcfcf3a352da5f985171d/absl_py-2.3.0-py3-none-any.whl", hash = "sha256:9824a48b654a306168f63e0d97714665f8490b8d89ec7bf2efc24bf67cf579b3", size = 135657 }, -] - -[[package]] -name = "accelerate" -version = "1.8.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "huggingface-hub" }, - { name = "numpy" }, - { name = "packaging" }, - { name = "psutil" }, - { name = "pyyaml" }, - { name = "safetensors" }, - { name = "torch" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/bd/c2/b9e33ad13232606dded4c546e654fb06a15f1dbcbd95d81c9f9dd3ccc771/accelerate-1.8.1.tar.gz", hash = "sha256:f60df931671bc4e75077b852990469d4991ce8bd3a58e72375c3c95132034db9", size = 380872 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/91/d9/e044c9d42d8ad9afa96533b46ecc9b7aea893d362b3c52bd78fb9fe4d7b3/accelerate-1.8.1-py3-none-any.whl", hash = "sha256:c47b8994498875a2b1286e945bd4d20e476956056c7941d512334f4eb44ff991", size = 365338 }, -] - -[[package]] -name = "aiohappyeyeballs" -version = "2.6.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/26/30/f84a107a9c4331c14b2b586036f40965c128aa4fee4dda5d3d51cb14ad54/aiohappyeyeballs-2.6.1.tar.gz", hash = "sha256:c3f9d0113123803ccadfdf3f0faa505bc78e6a72d1cc4806cbd719826e943558", size = 22760 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/0f/15/5bf3b99495fb160b63f95972b81750f18f7f4e02ad051373b669d17d44f2/aiohappyeyeballs-2.6.1-py3-none-any.whl", hash = "sha256:f349ba8f4b75cb25c99c5c2d84e997e485204d2902a9597802b0371f09331fb8", size = 15265 }, -] - -[[package]] -name = "aiohttp" -version = "3.12.13" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "aiohappyeyeballs" }, - { name = "aiosignal" }, - { name = "async-timeout", marker = "python_full_version == '3.10.12'" }, - { name = "attrs" }, - { name = "frozenlist" }, - { name = "multidict" }, - { name = "propcache" }, - { name = "yarl" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/42/6e/ab88e7cb2a4058bed2f7870276454f85a7c56cd6da79349eb314fc7bbcaa/aiohttp-3.12.13.tar.gz", hash = "sha256:47e2da578528264a12e4e3dd8dd72a7289e5f812758fe086473fab037a10fcce", size = 7819160 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/8b/2d/27e4347660723738b01daa3f5769d56170f232bf4695dd4613340da135bb/aiohttp-3.12.13-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:5421af8f22a98f640261ee48aae3a37f0c41371e99412d55eaf2f8a46d5dad29", size = 702090 }, - { url = "https://files.pythonhosted.org/packages/10/0b/4a8e0468ee8f2b9aff3c05f2c3a6be1dfc40b03f68a91b31041d798a9510/aiohttp-3.12.13-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:0fcda86f6cb318ba36ed8f1396a6a4a3fd8f856f84d426584392083d10da4de0", size = 478440 }, - { url = "https://files.pythonhosted.org/packages/b9/c8/2086df2f9a842b13feb92d071edf756be89250f404f10966b7bc28317f17/aiohttp-3.12.13-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:4cd71c9fb92aceb5a23c4c39d8ecc80389c178eba9feab77f19274843eb9412d", size = 466215 }, - { url = "https://files.pythonhosted.org/packages/a7/3d/d23e5bd978bc8012a65853959b13bd3b55c6e5afc172d89c26ad6624c52b/aiohttp-3.12.13-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:34ebf1aca12845066c963016655dac897651e1544f22a34c9b461ac3b4b1d3aa", size = 1648271 }, - { url = "https://files.pythonhosted.org/packages/31/31/e00122447bb137591c202786062f26dd383574c9f5157144127077d5733e/aiohttp-3.12.13-cp310-cp310-manylinux_2_17_armv7l.manylinux2014_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:893a4639694c5b7edd4bdd8141be296042b6806e27cc1d794e585c43010cc294", size = 1622329 }, - { url = "https://files.pythonhosted.org/packages/04/01/caef70be3ac38986969045f21f5fb802ce517b3f371f0615206bf8aa6423/aiohttp-3.12.13-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:663d8ee3ffb3494502ebcccb49078faddbb84c1d870f9c1dd5a29e85d1f747ce", size = 1694734 }, - { url = "https://files.pythonhosted.org/packages/3f/15/328b71fedecf69a9fd2306549b11c8966e420648a3938d75d3ed5bcb47f6/aiohttp-3.12.13-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:f0f8f6a85a0006ae2709aa4ce05749ba2cdcb4b43d6c21a16c8517c16593aabe", size = 1737049 }, - { url = "https://files.pythonhosted.org/packages/e6/7a/d85866a642158e1147c7da5f93ad66b07e5452a84ec4258e5f06b9071e92/aiohttp-3.12.13-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1582745eb63df267c92d8b61ca655a0ce62105ef62542c00a74590f306be8cb5", size = 1641715 }, - { url = "https://files.pythonhosted.org/packages/14/57/3588800d5d2f5f3e1cb6e7a72747d1abc1e67ba5048e8b845183259c2e9b/aiohttp-3.12.13-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:d59227776ee2aa64226f7e086638baa645f4b044f2947dbf85c76ab11dcba073", size = 1581836 }, - { url = "https://files.pythonhosted.org/packages/2f/55/c913332899a916d85781aa74572f60fd98127449b156ad9c19e23135b0e4/aiohttp-3.12.13-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:06b07c418bde1c8e737d8fa67741072bd3f5b0fb66cf8c0655172188c17e5fa6", size = 1625685 }, - { url = "https://files.pythonhosted.org/packages/4c/34/26cded195f3bff128d6a6d58d7a0be2ae7d001ea029e0fe9008dcdc6a009/aiohttp-3.12.13-cp310-cp310-musllinux_1_2_armv7l.whl", hash = "sha256:9445c1842680efac0f81d272fd8db7163acfcc2b1436e3f420f4c9a9c5a50795", size = 1636471 }, - { url = "https://files.pythonhosted.org/packages/19/21/70629ca006820fccbcec07f3cd5966cbd966e2d853d6da55339af85555b9/aiohttp-3.12.13-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:09c4767af0b0b98c724f5d47f2bf33395c8986995b0a9dab0575ca81a554a8c0", size = 1611923 }, - { url = "https://files.pythonhosted.org/packages/31/80/7fa3f3bebf533aa6ae6508b51ac0de9965e88f9654fa679cc1a29d335a79/aiohttp-3.12.13-cp310-cp310-musllinux_1_2_ppc64le.whl", hash = "sha256:f3854fbde7a465318ad8d3fc5bef8f059e6d0a87e71a0d3360bb56c0bf87b18a", size = 1691511 }, - { url = "https://files.pythonhosted.org/packages/0f/7a/359974653a3cdd3e9cee8ca10072a662c3c0eb46a359c6a1f667b0296e2f/aiohttp-3.12.13-cp310-cp310-musllinux_1_2_s390x.whl", hash = "sha256:2332b4c361c05ecd381edb99e2a33733f3db906739a83a483974b3df70a51b40", size = 1714751 }, - { url = "https://files.pythonhosted.org/packages/2d/24/0aa03d522171ce19064347afeefadb008be31ace0bbb7d44ceb055700a14/aiohttp-3.12.13-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:1561db63fa1b658cd94325d303933553ea7d89ae09ff21cc3bcd41b8521fbbb6", size = 1643090 }, - { url = "https://files.pythonhosted.org/packages/86/2e/7d4b0026a41e4b467e143221c51b279083b7044a4b104054f5c6464082ff/aiohttp-3.12.13-cp310-cp310-win32.whl", hash = "sha256:a0be857f0b35177ba09d7c472825d1b711d11c6d0e8a2052804e3b93166de1ad", size = 427526 }, - { url = "https://files.pythonhosted.org/packages/17/de/34d998da1e7f0de86382160d039131e9b0af1962eebfe53dda2b61d250e7/aiohttp-3.12.13-cp310-cp310-win_amd64.whl", hash = "sha256:fcc30ad4fb5cb41a33953292d45f54ef4066746d625992aeac33b8c681173178", size = 450734 }, -] - -[[package]] -name = "aiosignal" -version = "1.3.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "frozenlist" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/ba/b5/6d55e80f6d8a08ce22b982eafa278d823b541c925f11ee774b0b9c43473d/aiosignal-1.3.2.tar.gz", hash = "sha256:a8c255c66fafb1e499c9351d0bf32ff2d8a0321595ebac3b93713656d2436f54", size = 19424 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/ec/6a/bc7e17a3e87a2985d3e8f4da4cd0f481060eb78fb08596c42be62c90a4d9/aiosignal-1.3.2-py2.py3-none-any.whl", hash = "sha256:45cde58e409a301715980c2b01d0c28bdde3770d8290b5eb2173759d9acb31a5", size = 7597 }, -] - -[[package]] -name = "alembic" -version = "1.16.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "mako" }, - { name = "sqlalchemy" }, - { name = "tomli", marker = "python_full_version == '3.10.12'" }, - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/9c/35/116797ff14635e496bbda0c168987f5326a6555b09312e9b817e360d1f56/alembic-1.16.2.tar.gz", hash = "sha256:e53c38ff88dadb92eb22f8b150708367db731d58ad7e9d417c9168ab516cbed8", size = 1963563 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/dd/e2/88e425adac5ad887a087c38d04fe2030010572a3e0e627f8a6e8c33eeda8/alembic-1.16.2-py3-none-any.whl", hash = "sha256:5f42e9bd0afdbd1d5e3ad856c01754530367debdebf21ed6894e34af52b3bb03", size = 242717 }, -] - -[[package]] -name = "aliyun-python-sdk-core" -version = "2.16.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "cryptography" }, - { name = "jmespath" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/3e/09/da9f58eb38b4fdb97ba6523274fbf445ef6a06be64b433693da8307b4bec/aliyun-python-sdk-core-2.16.0.tar.gz", hash = "sha256:651caad597eb39d4fad6cf85133dffe92837d53bdf62db9d8f37dab6508bb8f9", size = 449555 } - -[[package]] -name = "aliyun-python-sdk-kms" -version = "2.16.5" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "aliyun-python-sdk-core" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/a8/2c/9877d0e6b18ecf246df671ac65a5d1d9fecbf85bdcb5d43efbde0d4662eb/aliyun-python-sdk-kms-2.16.5.tar.gz", hash = "sha256:f328a8a19d83ecbb965ffce0ec1e9930755216d104638cd95ecd362753b813b3", size = 12018 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/11/5c/0132193d7da2c735669a1ed103b142fd63c9455984d48c5a88a1a516efaa/aliyun_python_sdk_kms-2.16.5-py2.py3-none-any.whl", hash = "sha256:24b6cdc4fd161d2942619479c8d050c63ea9cd22b044fe33b60bbb60153786f0", size = 99495 }, -] - -[[package]] -name = "annotated-types" -version = "0.7.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/ee/67/531ea369ba64dcff5ec9c3402f9f51bf748cec26dde048a2f973a4eea7f5/annotated_types-0.7.0.tar.gz", hash = "sha256:aff07c09a53a08bc8cfccb9c85b05f1aa9a2a6f23728d790723543408344ce89", size = 16081 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/78/b6/6307fbef88d9b5ee7421e68d78a9f162e0da4900bc5f5793f6d3d0e34fb8/annotated_types-0.7.0-py3-none-any.whl", hash = "sha256:1f02e8b43a8fbbc3f3e0d4f0f4bfc8131bcb4eebe8849b8e5c773f3a1c582a53", size = 13643 }, -] - -[[package]] -name = "antlr4-python3-runtime" -version = "4.9.3" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/3e/38/7859ff46355f76f8d19459005ca000b6e7012f2f1ca597746cbcd1fbfe5e/antlr4-python3-runtime-4.9.3.tar.gz", hash = "sha256:f224469b4168294902bb1efa80a8bf7855f24c99aef99cbefc1bcd3cce77881b", size = 117034 } - -[[package]] -name = "anyio" -version = "4.9.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "exceptiongroup", marker = "python_full_version == '3.10.12'" }, - { name = "idna" }, - { name = "sniffio" }, - { name = "typing-extensions", marker = "python_full_version == '3.10.12'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/95/7d/4c1bd541d4dffa1b52bd83fb8527089e097a106fc90b467a7313b105f840/anyio-4.9.0.tar.gz", hash = "sha256:673c0c244e15788651a4ff38710fea9675823028a6f08a5eda409e0c9840a028", size = 190949 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/a1/ee/48ca1a7c89ffec8b6a0c5d02b89c305671d5ffd8d3c94acf8b8c408575bb/anyio-4.9.0-py3-none-any.whl", hash = "sha256:9f76d541cad6e36af7beb62e978876f3b41e3e04f2c1fbf0884604c0a9c4d93c", size = 100916 }, -] - -[[package]] -name = "appnope" -version = "0.1.4" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/35/5d/752690df9ef5b76e169e68d6a129fa6d08a7100ca7f754c89495db3c6019/appnope-0.1.4.tar.gz", hash = "sha256:1de3860566df9caf38f01f86f65e0e13e379af54f9e4bee1e66b48f2efffd1ee", size = 4170 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/81/29/5ecc3a15d5a33e31b26c11426c45c501e439cb865d0bff96315d86443b78/appnope-0.1.4-py2.py3-none-any.whl", hash = "sha256:502575ee11cd7a28c0205f379b525beefebab9d161b7c964670864014ed7213c", size = 4321 }, -] - -[[package]] -name = "argon2-cffi" -version = "25.1.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "argon2-cffi-bindings" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/0e/89/ce5af8a7d472a67cc819d5d998aa8c82c5d860608c4db9f46f1162d7dab9/argon2_cffi-25.1.0.tar.gz", hash = "sha256:694ae5cc8a42f4c4e2bf2ca0e64e51e23a040c6a517a85074683d3959e1346c1", size = 45706 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/4f/d3/a8b22fa575b297cd6e3e3b0155c7e25db170edf1c74783d6a31a2490b8d9/argon2_cffi-25.1.0-py3-none-any.whl", hash = "sha256:fdc8b074db390fccb6eb4a3604ae7231f219aa669a2652e0f20e16ba513d5741", size = 14657 }, -] - -[[package]] -name = "argon2-cffi-bindings" -version = "21.2.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "cffi" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/b9/e9/184b8ccce6683b0aa2fbb7ba5683ea4b9c5763f1356347f1312c32e3c66e/argon2-cffi-bindings-21.2.0.tar.gz", hash = "sha256:bb89ceffa6c791807d1305ceb77dbfacc5aa499891d2c55661c6459651fc39e3", size = 1779911 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/d4/13/838ce2620025e9666aa8f686431f67a29052241692a3dd1ae9d3692a89d3/argon2_cffi_bindings-21.2.0-cp36-abi3-macosx_10_9_x86_64.whl", hash = "sha256:ccb949252cb2ab3a08c02024acb77cfb179492d5701c7cbdbfd776124d4d2367", size = 29658 }, - { url = "https://files.pythonhosted.org/packages/b3/02/f7f7bb6b6af6031edb11037639c697b912e1dea2db94d436e681aea2f495/argon2_cffi_bindings-21.2.0-cp36-abi3-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:9524464572e12979364b7d600abf96181d3541da11e23ddf565a32e70bd4dc0d", size = 80583 }, - { url = "https://files.pythonhosted.org/packages/ec/f7/378254e6dd7ae6f31fe40c8649eea7d4832a42243acaf0f1fff9083b2bed/argon2_cffi_bindings-21.2.0-cp36-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b746dba803a79238e925d9046a63aa26bf86ab2a2fe74ce6b009a1c3f5c8f2ae", size = 86168 }, - { url = "https://files.pythonhosted.org/packages/74/f6/4a34a37a98311ed73bb80efe422fed95f2ac25a4cacc5ae1d7ae6a144505/argon2_cffi_bindings-21.2.0-cp36-abi3-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:58ed19212051f49a523abb1dbe954337dc82d947fb6e5a0da60f7c8471a8476c", size = 82709 }, - { url = "https://files.pythonhosted.org/packages/74/2b/73d767bfdaab25484f7e7901379d5f8793cccbb86c6e0cbc4c1b96f63896/argon2_cffi_bindings-21.2.0-cp36-abi3-musllinux_1_1_aarch64.whl", hash = "sha256:bd46088725ef7f58b5a1ef7ca06647ebaf0eb4baff7d1d0d177c6cc8744abd86", size = 83613 }, - { url = "https://files.pythonhosted.org/packages/4f/fd/37f86deef67ff57c76f137a67181949c2d408077e2e3dd70c6c42912c9bf/argon2_cffi_bindings-21.2.0-cp36-abi3-musllinux_1_1_i686.whl", hash = "sha256:8cd69c07dd875537a824deec19f978e0f2078fdda07fd5c42ac29668dda5f40f", size = 84583 }, - { url = "https://files.pythonhosted.org/packages/6f/52/5a60085a3dae8fded8327a4f564223029f5f54b0cb0455a31131b5363a01/argon2_cffi_bindings-21.2.0-cp36-abi3-musllinux_1_1_x86_64.whl", hash = "sha256:f1152ac548bd5b8bcecfb0b0371f082037e47128653df2e8ba6e914d384f3c3e", size = 88475 }, - { url = "https://files.pythonhosted.org/packages/8b/95/143cd64feb24a15fa4b189a3e1e7efbaeeb00f39a51e99b26fc62fbacabd/argon2_cffi_bindings-21.2.0-cp36-abi3-win32.whl", hash = "sha256:603ca0aba86b1349b147cab91ae970c63118a0f30444d4bc80355937c950c082", size = 27698 }, - { url = "https://files.pythonhosted.org/packages/37/2c/e34e47c7dee97ba6f01a6203e0383e15b60fb85d78ac9a15cd066f6fe28b/argon2_cffi_bindings-21.2.0-cp36-abi3-win_amd64.whl", hash = "sha256:b2ef1c30440dbbcba7a5dc3e319408b59676e2e039e2ae11a8775ecf482b192f", size = 30817 }, - { url = "https://files.pythonhosted.org/packages/5a/e4/bf8034d25edaa495da3c8a3405627d2e35758e44ff6eaa7948092646fdcc/argon2_cffi_bindings-21.2.0-cp38-abi3-macosx_10_9_universal2.whl", hash = "sha256:e415e3f62c8d124ee16018e491a009937f8cf7ebf5eb430ffc5de21b900dad93", size = 53104 }, -] - -[[package]] -name = "arrow" -version = "1.3.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "python-dateutil" }, - { name = "types-python-dateutil" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/2e/00/0f6e8fcdb23ea632c866620cc872729ff43ed91d284c866b515c6342b173/arrow-1.3.0.tar.gz", hash = "sha256:d4540617648cb5f895730f1ad8c82a65f2dad0166f57b75f3ca54759c4d67a85", size = 131960 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/f8/ed/e97229a566617f2ae958a6b13e7cc0f585470eac730a73e9e82c32a3cdd2/arrow-1.3.0-py3-none-any.whl", hash = "sha256:c728b120ebc00eb84e01882a6f5e7927a53960aa990ce7dd2b10f39005a67f80", size = 66419 }, -] - -[[package]] -name = "asteroid-filterbanks" -version = "0.4.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "numpy" }, - { name = "torch" }, - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/90/fa/5c2be1f96dc179f83cdd3bb267edbd1f47d08f756785c016d5c2163901a7/asteroid-filterbanks-0.4.0.tar.gz", hash = "sha256:415f89d1dcf2b13b35f03f7a9370968ac4e6fa6800633c522dac992b283409b9", size = 24599 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/c5/7c/83ff6046176a675e6a1e8aeefed8892cd97fe7c46af93cc540d1b24b8323/asteroid_filterbanks-0.4.0-py3-none-any.whl", hash = "sha256:4932ac8b6acc6e08fb87cbe8ece84215b5a74eee284fe83acf3540a72a02eaf5", size = 29912 }, -] - -[[package]] -name = "asttokens" -version = "3.0.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/4a/e7/82da0a03e7ba5141f05cce0d302e6eed121ae055e0456ca228bf693984bc/asttokens-3.0.0.tar.gz", hash = "sha256:0dcd8baa8d62b0c1d118b399b2ddba3c4aff271d0d7a9e0d4c1681c79035bbc7", size = 61978 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/25/8a/c46dcc25341b5bce5472c718902eb3d38600a903b14fa6aeecef3f21a46f/asttokens-3.0.0-py3-none-any.whl", hash = "sha256:e3078351a059199dd5138cb1c706e6430c05eff2ff136af5eb4790f9d28932e2", size = 26918 }, -] - -[[package]] -name = "async-lru" -version = "2.0.5" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "typing-extensions", marker = "python_full_version == '3.10.12'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/b2/4d/71ec4d3939dc755264f680f6c2b4906423a304c3d18e96853f0a595dfe97/async_lru-2.0.5.tar.gz", hash = "sha256:481d52ccdd27275f42c43a928b4a50c3bfb2d67af4e78b170e3e0bb39c66e5bb", size = 10380 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/03/49/d10027df9fce941cb8184e78a02857af36360d33e1721df81c5ed2179a1a/async_lru-2.0.5-py3-none-any.whl", hash = "sha256:ab95404d8d2605310d345932697371a5f40def0487c03d6d0ad9138de52c9943", size = 6069 }, -] - -[[package]] -name = "async-timeout" -version = "5.0.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/a5/ae/136395dfbfe00dfc94da3f3e136d0b13f394cba8f4841120e34226265780/async_timeout-5.0.1.tar.gz", hash = "sha256:d9321a7a3d5a6a5e187e824d2fa0793ce379a202935782d555d6e9d2735677d3", size = 9274 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/fe/ba/e2081de779ca30d473f21f5b30e0e737c438205440784c7dfc81efc2b029/async_timeout-5.0.1-py3-none-any.whl", hash = "sha256:39e3809566ff85354557ec2398b55e096c8364bacac9405a7a1fa429e77fe76c", size = 6233 }, -] - -[[package]] -name = "attrs" -version = "25.3.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/5a/b0/1367933a8532ee6ff8d63537de4f1177af4bff9f3e829baf7331f595bb24/attrs-25.3.0.tar.gz", hash = "sha256:75d7cefc7fb576747b2c81b4442d4d4a1ce0900973527c011d1030fd3bf4af1b", size = 812032 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/77/06/bb80f5f86020c4551da315d78b3ab75e8228f89f0162f2c3a819e407941a/attrs-25.3.0-py3-none-any.whl", hash = "sha256:427318ce031701fea540783410126f03899a97ffc6f61596ad581ac2e40e3bc3", size = 63815 }, -] - -[[package]] -name = "audioread" -version = "3.0.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/db/d2/87016ca9f083acadffb2d8da59bfa3253e4da7eeb9f71fb8e7708dc97ecd/audioread-3.0.1.tar.gz", hash = "sha256:ac5460a5498c48bdf2e8e767402583a4dcd13f4414d286f42ce4379e8b35066d", size = 116513 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/57/8d/30aa32745af16af0a9a650115fbe81bde7c610ed5c21b381fca0196f3a7f/audioread-3.0.1-py3-none-any.whl", hash = "sha256:4cdce70b8adc0da0a3c9e0d85fb10b3ace30fbdf8d1670fd443929b61d117c33", size = 23492 }, -] - -[[package]] -name = "babel" -version = "2.17.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/7d/6b/d52e42361e1aa00709585ecc30b3f9684b3ab62530771402248b1b1d6240/babel-2.17.0.tar.gz", hash = "sha256:0c54cffb19f690cdcc52a3b50bcbf71e07a808d1c80d549f2459b9d2cf0afb9d", size = 9951852 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/b7/b8/3fe70c75fe32afc4bb507f75563d39bc5642255d1d94f1f23604725780bf/babel-2.17.0-py3-none-any.whl", hash = "sha256:4d0b53093fdfb4b21c92b5213dba5a1b23885afa8383709427046b21c366e5f2", size = 10182537 }, -] - -[[package]] -name = "backports-datetime-fromisoformat" -version = "2.0.3" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/71/81/eff3184acb1d9dc3ce95a98b6f3c81a49b4be296e664db8e1c2eeabef3d9/backports_datetime_fromisoformat-2.0.3.tar.gz", hash = "sha256:b58edc8f517b66b397abc250ecc737969486703a66eb97e01e6d51291b1a139d", size = 23588 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/42/4b/d6b051ca4b3d76f23c2c436a9669f3be616b8cf6461a7e8061c7c4269642/backports_datetime_fromisoformat-2.0.3-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:5f681f638f10588fa3c101ee9ae2b63d3734713202ddfcfb6ec6cea0778a29d4", size = 27561 }, - { url = "https://files.pythonhosted.org/packages/6d/40/e39b0d471e55eb1b5c7c81edab605c02f71c786d59fb875f0a6f23318747/backports_datetime_fromisoformat-2.0.3-cp310-cp310-macosx_11_0_universal2.whl", hash = "sha256:cd681460e9142f1249408e5aee6d178c6d89b49e06d44913c8fdfb6defda8d1c", size = 34448 }, - { url = "https://files.pythonhosted.org/packages/f2/28/7a5c87c5561d14f1c9af979231fdf85d8f9fad7a95ff94e56d2205e2520a/backports_datetime_fromisoformat-2.0.3-cp310-cp310-macosx_11_0_x86_64.whl", hash = "sha256:ee68bc8735ae5058695b76d3bb2aee1d137c052a11c8303f1e966aa23b72b65b", size = 27093 }, - { url = "https://files.pythonhosted.org/packages/80/ba/f00296c5c4536967c7d1136107fdb91c48404fe769a4a6fd5ab045629af8/backports_datetime_fromisoformat-2.0.3-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8273fe7932db65d952a43e238318966eab9e49e8dd546550a41df12175cc2be4", size = 52836 }, - { url = "https://files.pythonhosted.org/packages/e3/92/bb1da57a069ddd601aee352a87262c7ae93467e66721d5762f59df5021a6/backports_datetime_fromisoformat-2.0.3-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:39d57ea50aa5a524bb239688adc1d1d824c31b6094ebd39aa164d6cadb85de22", size = 52798 }, - { url = "https://files.pythonhosted.org/packages/df/ef/b6cfd355982e817ccdb8d8d109f720cab6e06f900784b034b30efa8fa832/backports_datetime_fromisoformat-2.0.3-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:ac6272f87693e78209dc72e84cf9ab58052027733cd0721c55356d3c881791cf", size = 52891 }, - { url = "https://files.pythonhosted.org/packages/37/39/b13e3ae8a7c5d88b68a6e9248ffe7066534b0cfe504bf521963e61b6282d/backports_datetime_fromisoformat-2.0.3-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:44c497a71f80cd2bcfc26faae8857cf8e79388e3d5fbf79d2354b8c360547d58", size = 52955 }, - { url = "https://files.pythonhosted.org/packages/1e/e4/70cffa3ce1eb4f2ff0c0d6f5d56285aacead6bd3879b27a2ba57ab261172/backports_datetime_fromisoformat-2.0.3-cp310-cp310-win_amd64.whl", hash = "sha256:6335a4c9e8af329cb1ded5ab41a666e1448116161905a94e054f205aa6d263bc", size = 29323 }, - { url = "https://files.pythonhosted.org/packages/be/03/7eaa9f9bf290395d57fd30d7f1f2f9dff60c06a31c237dc2beb477e8f899/backports_datetime_fromisoformat-2.0.3-pp310-pypy310_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:90e202e72a3d5aae673fcc8c9a4267d56b2f532beeb9173361293625fe4d2039", size = 28980 }, - { url = "https://files.pythonhosted.org/packages/47/80/a0ecf33446c7349e79f54cc532933780341d20cff0ee12b5bfdcaa47067e/backports_datetime_fromisoformat-2.0.3-pp310-pypy310_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2df98ef1b76f5a58bb493dda552259ba60c3a37557d848e039524203951c9f06", size = 28449 }, -] - -[[package]] -name = "beautifulsoup4" -version = "4.13.4" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "soupsieve" }, - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/d8/e4/0c4c39e18fd76d6a628d4dd8da40543d136ce2d1752bd6eeeab0791f4d6b/beautifulsoup4-4.13.4.tar.gz", hash = "sha256:dbb3c4e1ceae6aefebdaf2423247260cd062430a410e38c66f2baa50a8437195", size = 621067 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/50/cd/30110dc0ffcf3b131156077b90e9f60ed75711223f306da4db08eff8403b/beautifulsoup4-4.13.4-py3-none-any.whl", hash = "sha256:9bbbb14bfde9d79f38b8cd5f8c7c85f4b8f2523190ebed90e950a8dea4cb1c4b", size = 187285 }, -] - -[[package]] -name = "bleach" -version = "6.2.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "webencodings" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/76/9a/0e33f5054c54d349ea62c277191c020c2d6ef1d65ab2cb1993f91ec846d1/bleach-6.2.0.tar.gz", hash = "sha256:123e894118b8a599fd80d3ec1a6d4cc7ce4e5882b1317a7e1ba69b56e95f991f", size = 203083 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/fc/55/96142937f66150805c25c4d0f31ee4132fd33497753400734f9dfdcbdc66/bleach-6.2.0-py3-none-any.whl", hash = "sha256:117d9c6097a7c3d22fd578fcd8d35ff1e125df6736f554da4e432fdd63f31e5e", size = 163406 }, -] - -[package.optional-dependencies] -css = [ - { name = "tinycss2" }, -] - -[[package]] -name = "braceexpand" -version = "0.1.7" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/54/93/badd4f5ccf25209f3fef2573073da9fe4a45a3da99fca2f800f942130c0f/braceexpand-0.1.7.tar.gz", hash = "sha256:e6e539bd20eaea53547472ff94f4fb5c3d3bf9d0a89388c4b56663aba765f705", size = 7777 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/fa/93/e8c04e80e82391a6e51f218ca49720f64236bc824e92152a2633b74cf7ab/braceexpand-0.1.7-py2.py3-none-any.whl", hash = "sha256:91332d53de7828103dcae5773fb43bc34950b0c8160e35e0f44c4427a3b85014", size = 5923 }, -] - -[[package]] -name = "cattrs" -version = "25.1.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "attrs" }, - { name = "exceptiongroup", marker = "python_full_version == '3.10.12'" }, - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/57/2b/561d78f488dcc303da4639e02021311728fb7fda8006dd2835550cddd9ed/cattrs-25.1.1.tar.gz", hash = "sha256:c914b734e0f2d59e5b720d145ee010f1fd9a13ee93900922a2f3f9d593b8382c", size = 435016 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/18/b0/215274ef0d835bbc1056392a367646648b6084e39d489099959aefcca2af/cattrs-25.1.1-py3-none-any.whl", hash = "sha256:1b40b2d3402af7be79a7e7e097a9b4cd16d4c06e6d526644b0b26a063a1cc064", size = 69386 }, -] - -[[package]] -name = "certifi" -version = "2025.6.15" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/73/f7/f14b46d4bcd21092d7d3ccef689615220d8a08fb25e564b65d20738e672e/certifi-2025.6.15.tar.gz", hash = "sha256:d747aa5a8b9bbbb1bb8c22bb13e22bd1f18e9796defa16bab421f7f7a317323b", size = 158753 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/84/ae/320161bd181fc06471eed047ecce67b693fd7515b16d495d8932db763426/certifi-2025.6.15-py3-none-any.whl", hash = "sha256:2e0c7ce7cb5d8f8634ca55d2ba7e6ec2689a2fd6537d8dec1296a477a4910057", size = 157650 }, -] - -[[package]] -name = "cffi" -version = "1.17.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "pycparser" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/fc/97/c783634659c2920c3fc70419e3af40972dbaf758daa229a7d6ea6135c90d/cffi-1.17.1.tar.gz", hash = "sha256:1c39c6016c32bc48dd54561950ebd6836e1670f2ae46128f67cf49e789c52824", size = 516621 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/90/07/f44ca684db4e4f08a3fdc6eeb9a0d15dc6883efc7b8c90357fdbf74e186c/cffi-1.17.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:df8b1c11f177bc2313ec4b2d46baec87a5f3e71fc8b45dab2ee7cae86d9aba14", size = 182191 }, - { url = "https://files.pythonhosted.org/packages/08/fd/cc2fedbd887223f9f5d170c96e57cbf655df9831a6546c1727ae13fa977a/cffi-1.17.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:8f2cdc858323644ab277e9bb925ad72ae0e67f69e804f4898c070998d50b1a67", size = 178592 }, - { url = "https://files.pythonhosted.org/packages/de/cc/4635c320081c78d6ffc2cab0a76025b691a91204f4aa317d568ff9280a2d/cffi-1.17.1-cp310-cp310-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:edae79245293e15384b51f88b00613ba9f7198016a5948b5dddf4917d4d26382", size = 426024 }, - { url = "https://files.pythonhosted.org/packages/b6/7b/3b2b250f3aab91abe5f8a51ada1b717935fdaec53f790ad4100fe2ec64d1/cffi-1.17.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:45398b671ac6d70e67da8e4224a065cec6a93541bb7aebe1b198a61b58c7b702", size = 448188 }, - { url = "https://files.pythonhosted.org/packages/d3/48/1b9283ebbf0ec065148d8de05d647a986c5f22586b18120020452fff8f5d/cffi-1.17.1-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ad9413ccdeda48c5afdae7e4fa2192157e991ff761e7ab8fdd8926f40b160cc3", size = 455571 }, - { url = "https://files.pythonhosted.org/packages/40/87/3b8452525437b40f39ca7ff70276679772ee7e8b394934ff60e63b7b090c/cffi-1.17.1-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:5da5719280082ac6bd9aa7becb3938dc9f9cbd57fac7d2871717b1feb0902ab6", size = 436687 }, - { url = "https://files.pythonhosted.org/packages/8d/fb/4da72871d177d63649ac449aec2e8a29efe0274035880c7af59101ca2232/cffi-1.17.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2bb1a08b8008b281856e5971307cc386a8e9c5b625ac297e853d36da6efe9c17", size = 446211 }, - { url = "https://files.pythonhosted.org/packages/ab/a0/62f00bcb411332106c02b663b26f3545a9ef136f80d5df746c05878f8c4b/cffi-1.17.1-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:045d61c734659cc045141be4bae381a41d89b741f795af1dd018bfb532fd0df8", size = 461325 }, - { url = "https://files.pythonhosted.org/packages/36/83/76127035ed2e7e27b0787604d99da630ac3123bfb02d8e80c633f218a11d/cffi-1.17.1-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:6883e737d7d9e4899a8a695e00ec36bd4e5e4f18fabe0aca0efe0a4b44cdb13e", size = 438784 }, - { url = "https://files.pythonhosted.org/packages/21/81/a6cd025db2f08ac88b901b745c163d884641909641f9b826e8cb87645942/cffi-1.17.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:6b8b4a92e1c65048ff98cfe1f735ef8f1ceb72e3d5f0c25fdb12087a23da22be", size = 461564 }, - { url = "https://files.pythonhosted.org/packages/f8/fe/4d41c2f200c4a457933dbd98d3cf4e911870877bd94d9656cc0fcb390681/cffi-1.17.1-cp310-cp310-win32.whl", hash = "sha256:c9c3d058ebabb74db66e431095118094d06abf53284d9c81f27300d0e0d8bc7c", size = 171804 }, - { url = "https://files.pythonhosted.org/packages/d1/b6/0b0f5ab93b0df4acc49cae758c81fe4e5ef26c3ae2e10cc69249dfd8b3ab/cffi-1.17.1-cp310-cp310-win_amd64.whl", hash = "sha256:0f048dcf80db46f0098ccac01132761580d28e28bc0f78ae0d58048063317e15", size = 181299 }, -] - -[[package]] -name = "charset-normalizer" -version = "3.4.2" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/e4/33/89c2ced2b67d1c2a61c19c6751aa8902d46ce3dacb23600a283619f5a12d/charset_normalizer-3.4.2.tar.gz", hash = "sha256:5baececa9ecba31eff645232d59845c07aa030f0c81ee70184a90d35099a0e63", size = 126367 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/95/28/9901804da60055b406e1a1c5ba7aac1276fb77f1dde635aabfc7fd84b8ab/charset_normalizer-3.4.2-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:7c48ed483eb946e6c04ccbe02c6b4d1d48e51944b6db70f697e089c193404941", size = 201818 }, - { url = "https://files.pythonhosted.org/packages/d9/9b/892a8c8af9110935e5adcbb06d9c6fe741b6bb02608c6513983048ba1a18/charset_normalizer-3.4.2-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:b2d318c11350e10662026ad0eb71bb51c7812fc8590825304ae0bdd4ac283acd", size = 144649 }, - { url = "https://files.pythonhosted.org/packages/7b/a5/4179abd063ff6414223575e008593861d62abfc22455b5d1a44995b7c101/charset_normalizer-3.4.2-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:9cbfacf36cb0ec2897ce0ebc5d08ca44213af24265bd56eca54bee7923c48fd6", size = 155045 }, - { url = "https://files.pythonhosted.org/packages/3b/95/bc08c7dfeddd26b4be8c8287b9bb055716f31077c8b0ea1cd09553794665/charset_normalizer-3.4.2-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:18dd2e350387c87dabe711b86f83c9c78af772c748904d372ade190b5c7c9d4d", size = 147356 }, - { url = "https://files.pythonhosted.org/packages/a8/2d/7a5b635aa65284bf3eab7653e8b4151ab420ecbae918d3e359d1947b4d61/charset_normalizer-3.4.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8075c35cd58273fee266c58c0c9b670947c19df5fb98e7b66710e04ad4e9ff86", size = 149471 }, - { url = "https://files.pythonhosted.org/packages/ae/38/51fc6ac74251fd331a8cfdb7ec57beba8c23fd5493f1050f71c87ef77ed0/charset_normalizer-3.4.2-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:5bf4545e3b962767e5c06fe1738f951f77d27967cb2caa64c28be7c4563e162c", size = 151317 }, - { url = "https://files.pythonhosted.org/packages/b7/17/edee1e32215ee6e9e46c3e482645b46575a44a2d72c7dfd49e49f60ce6bf/charset_normalizer-3.4.2-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:7a6ab32f7210554a96cd9e33abe3ddd86732beeafc7a28e9955cdf22ffadbab0", size = 146368 }, - { url = "https://files.pythonhosted.org/packages/26/2c/ea3e66f2b5f21fd00b2825c94cafb8c326ea6240cd80a91eb09e4a285830/charset_normalizer-3.4.2-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:b33de11b92e9f75a2b545d6e9b6f37e398d86c3e9e9653c4864eb7e89c5773ef", size = 154491 }, - { url = "https://files.pythonhosted.org/packages/52/47/7be7fa972422ad062e909fd62460d45c3ef4c141805b7078dbab15904ff7/charset_normalizer-3.4.2-cp310-cp310-musllinux_1_2_ppc64le.whl", hash = "sha256:8755483f3c00d6c9a77f490c17e6ab0c8729e39e6390328e42521ef175380ae6", size = 157695 }, - { url = "https://files.pythonhosted.org/packages/2f/42/9f02c194da282b2b340f28e5fb60762de1151387a36842a92b533685c61e/charset_normalizer-3.4.2-cp310-cp310-musllinux_1_2_s390x.whl", hash = "sha256:68a328e5f55ec37c57f19ebb1fdc56a248db2e3e9ad769919a58672958e8f366", size = 154849 }, - { url = "https://files.pythonhosted.org/packages/67/44/89cacd6628f31fb0b63201a618049be4be2a7435a31b55b5eb1c3674547a/charset_normalizer-3.4.2-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:21b2899062867b0e1fde9b724f8aecb1af14f2778d69aacd1a5a1853a597a5db", size = 150091 }, - { url = "https://files.pythonhosted.org/packages/1f/79/4b8da9f712bc079c0f16b6d67b099b0b8d808c2292c937f267d816ec5ecc/charset_normalizer-3.4.2-cp310-cp310-win32.whl", hash = "sha256:e8082b26888e2f8b36a042a58307d5b917ef2b1cacab921ad3323ef91901c71a", size = 98445 }, - { url = "https://files.pythonhosted.org/packages/7d/d7/96970afb4fb66497a40761cdf7bd4f6fca0fc7bafde3a84f836c1f57a926/charset_normalizer-3.4.2-cp310-cp310-win_amd64.whl", hash = "sha256:f69a27e45c43520f5487f27627059b64aaf160415589230992cec34c5e18a509", size = 105782 }, - { url = "https://files.pythonhosted.org/packages/20/94/c5790835a017658cbfabd07f3bfb549140c3ac458cfc196323996b10095a/charset_normalizer-3.4.2-py3-none-any.whl", hash = "sha256:7f56930ab0abd1c45cd15be65cc741c28b1c9a34876ce8c17a2fa107810c0af0", size = 52626 }, -] - -[[package]] -name = "click" -version = "8.2.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "colorama", marker = "platform_system == 'Windows'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/60/6c/8ca2efa64cf75a977a0d7fac081354553ebe483345c734fb6b6515d96bbc/click-8.2.1.tar.gz", hash = "sha256:27c491cc05d968d271d5a1db13e3b5a184636d9d930f148c50b038f0d0646202", size = 286342 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/85/32/10bb5764d90a8eee674e9dc6f4db6a0ab47c8c4d0d83c27f7c39ac415a4d/click-8.2.1-py3-none-any.whl", hash = "sha256:61a3265b914e850b85317d0b3109c7f8cd35a670f963866005d6ef1d5175a12b", size = 102215 }, -] - -[[package]] -name = "cloudpickle" -version = "3.1.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/52/39/069100b84d7418bc358d81669d5748efb14b9cceacd2f9c75f550424132f/cloudpickle-3.1.1.tar.gz", hash = "sha256:b216fa8ae4019d5482a8ac3c95d8f6346115d8835911fd4aefd1a445e4242c64", size = 22113 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/7e/e8/64c37fadfc2816a7701fa8a6ed8d87327c7d54eacfbfb6edab14a2f2be75/cloudpickle-3.1.1-py3-none-any.whl", hash = "sha256:c8c5a44295039331ee9dad40ba100a9c7297b6f988e50e87ccdf3765a668350e", size = 20992 }, -] - -[[package]] -name = "colorama" -version = "0.4.6" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/d8/53/6f443c9a4a8358a93a6792e2acffb9d9d5cb0a5cfd8802644b7b1c9a02e4/colorama-0.4.6.tar.gz", hash = "sha256:08695f5cb7ed6e0531a20572697297273c47b8cae5a63ffc6d6ed5c201be6e44", size = 27697 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/d1/d6/3965ed04c63042e047cb6a3e6ed1a63a35087b6a609aa3a15ed8ac56c221/colorama-0.4.6-py2.py3-none-any.whl", hash = "sha256:4f1d9991f5acc0ca119f9d443620b77f9d6b33703e51011c16baf57afb285fc6", size = 25335 }, -] - -[[package]] -name = "colorlog" -version = "6.9.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "colorama", marker = "sys_platform == 'win32'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/d3/7a/359f4d5df2353f26172b3cc39ea32daa39af8de522205f512f458923e677/colorlog-6.9.0.tar.gz", hash = "sha256:bfba54a1b93b94f54e1f4fe48395725a3d92fd2a4af702f6bd70946bdc0c6ac2", size = 16624 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/e3/51/9b208e85196941db2f0654ad0357ca6388ab3ed67efdbfc799f35d1f83aa/colorlog-6.9.0-py3-none-any.whl", hash = "sha256:5906e71acd67cb07a71e779c47c4bcb45fb8c2993eebe9e5adcd6a6f1b283eff", size = 11424 }, -] - -[[package]] -name = "comm" -version = "0.2.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "traitlets" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/e9/a8/fb783cb0abe2b5fded9f55e5703015cdf1c9c85b3669087c538dd15a6a86/comm-0.2.2.tar.gz", hash = "sha256:3fd7a84065306e07bea1773df6eb8282de51ba82f77c72f9c85716ab11fe980e", size = 6210 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/e6/75/49e5bfe642f71f272236b5b2d2691cf915a7283cc0ceda56357b61daa538/comm-0.2.2-py3-none-any.whl", hash = "sha256:e6fb86cb70ff661ee8c9c14e7d36d6de3b4066f1441be4063df9c5009f0a64d3", size = 7180 }, -] - -[[package]] -name = "contourpy" -version = "1.3.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "numpy" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/66/54/eb9bfc647b19f2009dd5c7f5ec51c4e6ca831725f1aea7a993034f483147/contourpy-1.3.2.tar.gz", hash = "sha256:b6945942715a034c671b7fc54f9588126b0b8bf23db2696e3ca8328f3ff0ab54", size = 13466130 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/12/a3/da4153ec8fe25d263aa48c1a4cbde7f49b59af86f0b6f7862788c60da737/contourpy-1.3.2-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:ba38e3f9f330af820c4b27ceb4b9c7feee5fe0493ea53a8720f4792667465934", size = 268551 }, - { url = "https://files.pythonhosted.org/packages/2f/6c/330de89ae1087eb622bfca0177d32a7ece50c3ef07b28002de4757d9d875/contourpy-1.3.2-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:dc41ba0714aa2968d1f8674ec97504a8f7e334f48eeacebcaa6256213acb0989", size = 253399 }, - { url = "https://files.pythonhosted.org/packages/c1/bd/20c6726b1b7f81a8bee5271bed5c165f0a8e1f572578a9d27e2ccb763cb2/contourpy-1.3.2-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:9be002b31c558d1ddf1b9b415b162c603405414bacd6932d031c5b5a8b757f0d", size = 312061 }, - { url = "https://files.pythonhosted.org/packages/22/fc/a9665c88f8a2473f823cf1ec601de9e5375050f1958cbb356cdf06ef1ab6/contourpy-1.3.2-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:8d2e74acbcba3bfdb6d9d8384cdc4f9260cae86ed9beee8bd5f54fee49a430b9", size = 351956 }, - { url = "https://files.pythonhosted.org/packages/25/eb/9f0a0238f305ad8fb7ef42481020d6e20cf15e46be99a1fcf939546a177e/contourpy-1.3.2-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:e259bced5549ac64410162adc973c5e2fb77f04df4a439d00b478e57a0e65512", size = 320872 }, - { url = "https://files.pythonhosted.org/packages/32/5c/1ee32d1c7956923202f00cf8d2a14a62ed7517bdc0ee1e55301227fc273c/contourpy-1.3.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ad687a04bc802cbe8b9c399c07162a3c35e227e2daccf1668eb1f278cb698631", size = 325027 }, - { url = "https://files.pythonhosted.org/packages/83/bf/9baed89785ba743ef329c2b07fd0611d12bfecbedbdd3eeecf929d8d3b52/contourpy-1.3.2-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:cdd22595308f53ef2f891040ab2b93d79192513ffccbd7fe19be7aa773a5e09f", size = 1306641 }, - { url = "https://files.pythonhosted.org/packages/d4/cc/74e5e83d1e35de2d28bd97033426b450bc4fd96e092a1f7a63dc7369b55d/contourpy-1.3.2-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:b4f54d6a2defe9f257327b0f243612dd051cc43825587520b1bf74a31e2f6ef2", size = 1374075 }, - { url = "https://files.pythonhosted.org/packages/0c/42/17f3b798fd5e033b46a16f8d9fcb39f1aba051307f5ebf441bad1ecf78f8/contourpy-1.3.2-cp310-cp310-win32.whl", hash = "sha256:f939a054192ddc596e031e50bb13b657ce318cf13d264f095ce9db7dc6ae81c0", size = 177534 }, - { url = "https://files.pythonhosted.org/packages/54/ec/5162b8582f2c994721018d0c9ece9dc6ff769d298a8ac6b6a652c307e7df/contourpy-1.3.2-cp310-cp310-win_amd64.whl", hash = "sha256:c440093bbc8fc21c637c03bafcbef95ccd963bc6e0514ad887932c18ca2a759a", size = 221188 }, - { url = "https://files.pythonhosted.org/packages/33/05/b26e3c6ecc05f349ee0013f0bb850a761016d89cec528a98193a48c34033/contourpy-1.3.2-pp310-pypy310_pp73-macosx_10_15_x86_64.whl", hash = "sha256:fd93cc7f3139b6dd7aab2f26a90dde0aa9fc264dbf70f6740d498a70b860b82c", size = 265681 }, - { url = "https://files.pythonhosted.org/packages/2b/25/ac07d6ad12affa7d1ffed11b77417d0a6308170f44ff20fa1d5aa6333f03/contourpy-1.3.2-pp310-pypy310_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:107ba8a6a7eec58bb475329e6d3b95deba9440667c4d62b9b6063942b61d7f16", size = 315101 }, - { url = "https://files.pythonhosted.org/packages/8f/4d/5bb3192bbe9d3f27e3061a6a8e7733c9120e203cb8515767d30973f71030/contourpy-1.3.2-pp310-pypy310_pp73-win_amd64.whl", hash = "sha256:ded1706ed0c1049224531b81128efbd5084598f18d8a2d9efae833edbd2b40ad", size = 220599 }, -] - -[[package]] -name = "coremltools" -version = "9.0b1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "attrs" }, - { name = "cattrs" }, - { name = "numpy" }, - { name = "packaging" }, - { name = "protobuf" }, - { name = "pyaml" }, - { name = "sympy" }, - { name = "tqdm" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/7b/4b/10f7409775150955bb7649124c592480d764491ebfacd10835130f964485/coremltools-9.0b1.tar.gz", hash = "sha256:55c0e91b0362865041cb3d8625065b9c2b9ac741b5bc8506ae3ad72012178c53", size = 1610618 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/28/47/8e4dd7b739043be8c19a89e505b751797e0d95a4827ae5a136b414890bf5/coremltools-9.0b1-cp310-none-macosx_10_15_x86_64.whl", hash = "sha256:857213355eadd3aeaa61796adff095a4c1b2d925936d866fcf9d3ac0d8c11fb2", size = 2784755 }, - { url = "https://files.pythonhosted.org/packages/81/bc/12179b5d17f4fa75d321b3ace57689a89bfb3f9728e2aed6cf49698a1c8a/coremltools-9.0b1-cp310-none-macosx_11_0_arm64.whl", hash = "sha256:beb692994682bb7e9d5739580a15a50d500bccdfabeb36d541b1f0db483fd438", size = 2759581 }, - { url = "https://files.pythonhosted.org/packages/12/a9/98c8177de8771fe79d58cc8c06be0a0c5900aa9cd3292db61350e17f35dc/coremltools-9.0b1-cp310-none-manylinux1_x86_64.whl", hash = "sha256:4cb3f710a0edcf82c110ec041b6c57cd15c76f2c5c9c015ea49afcee2975c1be", size = 2303866 }, -] - -[[package]] -name = "crcmod" -version = "1.7" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/6b/b0/e595ce2a2527e169c3bcd6c33d2473c1918e0b7f6826a043ca1245dd4e5b/crcmod-1.7.tar.gz", hash = "sha256:dc7051a0db5f2bd48665a990d3ec1cc305a466a77358ca4492826f41f283601e", size = 89670 } - -[[package]] -name = "cryptography" -version = "45.0.7" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "cffi", marker = "platform_python_implementation != 'PyPy'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/a7/35/c495bffc2056f2dadb32434f1feedd79abde2a7f8363e1974afa9c33c7e2/cryptography-45.0.7.tar.gz", hash = "sha256:4b1654dfc64ea479c242508eb8c724044f1e964a47d1d1cacc5132292d851971", size = 744980 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/0c/91/925c0ac74362172ae4516000fe877912e33b5983df735ff290c653de4913/cryptography-45.0.7-cp311-abi3-macosx_10_9_universal2.whl", hash = "sha256:3be4f21c6245930688bd9e162829480de027f8bf962ede33d4f8ba7d67a00cee", size = 7041105 }, - { url = "https://files.pythonhosted.org/packages/fc/63/43641c5acce3a6105cf8bd5baeceeb1846bb63067d26dae3e5db59f1513a/cryptography-45.0.7-cp311-abi3-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:67285f8a611b0ebc0857ced2081e30302909f571a46bfa7a3cc0ad303fe015c6", size = 4205799 }, - { url = "https://files.pythonhosted.org/packages/bc/29/c238dd9107f10bfde09a4d1c52fd38828b1aa353ced11f358b5dd2507d24/cryptography-45.0.7-cp311-abi3-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:577470e39e60a6cd7780793202e63536026d9b8641de011ed9d8174da9ca5339", size = 4430504 }, - { url = "https://files.pythonhosted.org/packages/62/62/24203e7cbcc9bd7c94739428cd30680b18ae6b18377ae66075c8e4771b1b/cryptography-45.0.7-cp311-abi3-manylinux_2_28_aarch64.whl", hash = "sha256:4bd3e5c4b9682bc112d634f2c6ccc6736ed3635fc3319ac2bb11d768cc5a00d8", size = 4209542 }, - { url = "https://files.pythonhosted.org/packages/cd/e3/e7de4771a08620eef2389b86cd87a2c50326827dea5528feb70595439ce4/cryptography-45.0.7-cp311-abi3-manylinux_2_28_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:465ccac9d70115cd4de7186e60cfe989de73f7bb23e8a7aa45af18f7412e75bf", size = 3889244 }, - { url = "https://files.pythonhosted.org/packages/96/b8/bca71059e79a0bb2f8e4ec61d9c205fbe97876318566cde3b5092529faa9/cryptography-45.0.7-cp311-abi3-manylinux_2_28_x86_64.whl", hash = "sha256:16ede8a4f7929b4b7ff3642eba2bf79aa1d71f24ab6ee443935c0d269b6bc513", size = 4461975 }, - { url = "https://files.pythonhosted.org/packages/58/67/3f5b26937fe1218c40e95ef4ff8d23c8dc05aa950d54200cc7ea5fb58d28/cryptography-45.0.7-cp311-abi3-manylinux_2_34_aarch64.whl", hash = "sha256:8978132287a9d3ad6b54fcd1e08548033cc09dc6aacacb6c004c73c3eb5d3ac3", size = 4209082 }, - { url = "https://files.pythonhosted.org/packages/0e/e4/b3e68a4ac363406a56cf7b741eeb80d05284d8c60ee1a55cdc7587e2a553/cryptography-45.0.7-cp311-abi3-manylinux_2_34_x86_64.whl", hash = "sha256:b6a0e535baec27b528cb07a119f321ac024592388c5681a5ced167ae98e9fff3", size = 4460397 }, - { url = "https://files.pythonhosted.org/packages/22/49/2c93f3cd4e3efc8cb22b02678c1fad691cff9dd71bb889e030d100acbfe0/cryptography-45.0.7-cp311-abi3-musllinux_1_2_aarch64.whl", hash = "sha256:a24ee598d10befaec178efdff6054bc4d7e883f615bfbcd08126a0f4931c83a6", size = 4337244 }, - { url = "https://files.pythonhosted.org/packages/04/19/030f400de0bccccc09aa262706d90f2ec23d56bc4eb4f4e8268d0ddf3fb8/cryptography-45.0.7-cp311-abi3-musllinux_1_2_x86_64.whl", hash = "sha256:fa26fa54c0a9384c27fcdc905a2fb7d60ac6e47d14bc2692145f2b3b1e2cfdbd", size = 4568862 }, - { url = "https://files.pythonhosted.org/packages/29/56/3034a3a353efa65116fa20eb3c990a8c9f0d3db4085429040a7eef9ada5f/cryptography-45.0.7-cp311-abi3-win32.whl", hash = "sha256:bef32a5e327bd8e5af915d3416ffefdbe65ed975b646b3805be81b23580b57b8", size = 2936578 }, - { url = "https://files.pythonhosted.org/packages/b3/61/0ab90f421c6194705a99d0fa9f6ee2045d916e4455fdbb095a9c2c9a520f/cryptography-45.0.7-cp311-abi3-win_amd64.whl", hash = "sha256:3808e6b2e5f0b46d981c24d79648e5c25c35e59902ea4391a0dcb3e667bf7443", size = 3405400 }, - { url = "https://files.pythonhosted.org/packages/63/e8/c436233ddf19c5f15b25ace33979a9dd2e7aa1a59209a0ee8554179f1cc0/cryptography-45.0.7-cp37-abi3-macosx_10_9_universal2.whl", hash = "sha256:bfb4c801f65dd61cedfc61a83732327fafbac55a47282e6f26f073ca7a41c3b2", size = 7021824 }, - { url = "https://files.pythonhosted.org/packages/bc/4c/8f57f2500d0ccd2675c5d0cc462095adf3faa8c52294ba085c036befb901/cryptography-45.0.7-cp37-abi3-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:81823935e2f8d476707e85a78a405953a03ef7b7b4f55f93f7c2d9680e5e0691", size = 4202233 }, - { url = "https://files.pythonhosted.org/packages/eb/ac/59b7790b4ccaed739fc44775ce4645c9b8ce54cbec53edf16c74fd80cb2b/cryptography-45.0.7-cp37-abi3-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:3994c809c17fc570c2af12c9b840d7cea85a9fd3e5c0e0491f4fa3c029216d59", size = 4423075 }, - { url = "https://files.pythonhosted.org/packages/b8/56/d4f07ea21434bf891faa088a6ac15d6d98093a66e75e30ad08e88aa2b9ba/cryptography-45.0.7-cp37-abi3-manylinux_2_28_aarch64.whl", hash = "sha256:dad43797959a74103cb59c5dac71409f9c27d34c8a05921341fb64ea8ccb1dd4", size = 4204517 }, - { url = "https://files.pythonhosted.org/packages/e8/ac/924a723299848b4c741c1059752c7cfe09473b6fd77d2920398fc26bfb53/cryptography-45.0.7-cp37-abi3-manylinux_2_28_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:ce7a453385e4c4693985b4a4a3533e041558851eae061a58a5405363b098fcd3", size = 3882893 }, - { url = "https://files.pythonhosted.org/packages/83/dc/4dab2ff0a871cc2d81d3ae6d780991c0192b259c35e4d83fe1de18b20c70/cryptography-45.0.7-cp37-abi3-manylinux_2_28_x86_64.whl", hash = "sha256:b04f85ac3a90c227b6e5890acb0edbaf3140938dbecf07bff618bf3638578cf1", size = 4450132 }, - { url = "https://files.pythonhosted.org/packages/12/dd/b2882b65db8fc944585d7fb00d67cf84a9cef4e77d9ba8f69082e911d0de/cryptography-45.0.7-cp37-abi3-manylinux_2_34_aarch64.whl", hash = "sha256:48c41a44ef8b8c2e80ca4527ee81daa4c527df3ecbc9423c41a420a9559d0e27", size = 4204086 }, - { url = "https://files.pythonhosted.org/packages/5d/fa/1d5745d878048699b8eb87c984d4ccc5da4f5008dfd3ad7a94040caca23a/cryptography-45.0.7-cp37-abi3-manylinux_2_34_x86_64.whl", hash = "sha256:f3df7b3d0f91b88b2106031fd995802a2e9ae13e02c36c1fc075b43f420f3a17", size = 4449383 }, - { url = "https://files.pythonhosted.org/packages/36/8b/fc61f87931bc030598e1876c45b936867bb72777eac693e905ab89832670/cryptography-45.0.7-cp37-abi3-musllinux_1_2_aarch64.whl", hash = "sha256:dd342f085542f6eb894ca00ef70236ea46070c8a13824c6bde0dfdcd36065b9b", size = 4332186 }, - { url = "https://files.pythonhosted.org/packages/0b/11/09700ddad7443ccb11d674efdbe9a832b4455dc1f16566d9bd3834922ce5/cryptography-45.0.7-cp37-abi3-musllinux_1_2_x86_64.whl", hash = "sha256:1993a1bb7e4eccfb922b6cd414f072e08ff5816702a0bdb8941c247a6b1b287c", size = 4561639 }, - { url = "https://files.pythonhosted.org/packages/71/ed/8f4c1337e9d3b94d8e50ae0b08ad0304a5709d483bfcadfcc77a23dbcb52/cryptography-45.0.7-cp37-abi3-win32.whl", hash = "sha256:18fcf70f243fe07252dcb1b268a687f2358025ce32f9f88028ca5c364b123ef5", size = 2926552 }, - { url = "https://files.pythonhosted.org/packages/bc/ff/026513ecad58dacd45d1d24ebe52b852165a26e287177de1d545325c0c25/cryptography-45.0.7-cp37-abi3-win_amd64.whl", hash = "sha256:7285a89df4900ed3bfaad5679b1e668cb4b38a8de1ccbfc84b05f34512da0a90", size = 3392742 }, - { url = "https://files.pythonhosted.org/packages/13/3e/e42f1528ca1ea82256b835191eab1be014e0f9f934b60d98b0be8a38ed70/cryptography-45.0.7-pp310-pypy310_pp73-macosx_10_9_x86_64.whl", hash = "sha256:de58755d723e86175756f463f2f0bddd45cc36fbd62601228a3f8761c9f58252", size = 3572442 }, - { url = "https://files.pythonhosted.org/packages/59/aa/e947693ab08674a2663ed2534cd8d345cf17bf6a1facf99273e8ec8986dc/cryptography-45.0.7-pp310-pypy310_pp73-manylinux_2_28_aarch64.whl", hash = "sha256:a20e442e917889d1a6b3c570c9e3fa2fdc398c20868abcea268ea33c024c4083", size = 4142233 }, - { url = "https://files.pythonhosted.org/packages/24/06/09b6f6a2fc43474a32b8fe259038eef1500ee3d3c141599b57ac6c57612c/cryptography-45.0.7-pp310-pypy310_pp73-manylinux_2_28_x86_64.whl", hash = "sha256:258e0dff86d1d891169b5af222d362468a9570e2532923088658aa866eb11130", size = 4376202 }, - { url = "https://files.pythonhosted.org/packages/00/f2/c166af87e95ce6ae6d38471a7e039d3a0549c2d55d74e059680162052824/cryptography-45.0.7-pp310-pypy310_pp73-manylinux_2_34_aarch64.whl", hash = "sha256:d97cf502abe2ab9eff8bd5e4aca274da8d06dd3ef08b759a8d6143f4ad65d4b4", size = 4141900 }, - { url = "https://files.pythonhosted.org/packages/16/b9/e96e0b6cb86eae27ea51fa8a3151535a18e66fe7c451fa90f7f89c85f541/cryptography-45.0.7-pp310-pypy310_pp73-manylinux_2_34_x86_64.whl", hash = "sha256:c987dad82e8c65ebc985f5dae5e74a3beda9d0a2a4daf8a1115f3772b59e5141", size = 4375562 }, - { url = "https://files.pythonhosted.org/packages/36/d0/36e8ee39274e9d77baf7d0dafda680cba6e52f3936b846f0d56d64fec915/cryptography-45.0.7-pp310-pypy310_pp73-win_amd64.whl", hash = "sha256:c13b1e3afd29a5b3b2656257f14669ca8fa8d7956d509926f0b130b600b50ab7", size = 3322781 }, -] - -[[package]] -name = "cycler" -version = "0.12.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/a9/95/a3dbbb5028f35eafb79008e7522a75244477d2838f38cbb722248dabc2a8/cycler-0.12.1.tar.gz", hash = "sha256:88bb128f02ba341da8ef447245a9e138fae777f6a23943da4540077d3601eb1c", size = 7615 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/e7/05/c19819d5e3d95294a6f5947fb9b9629efb316b96de511b418c53d245aae6/cycler-0.12.1-py3-none-any.whl", hash = "sha256:85cef7cff222d8644161529808465972e51340599459b8ac3ccbac5a854e0d30", size = 8321 }, -] - -[[package]] -name = "cytoolz" -version = "1.0.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "toolz" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/a7/f9/3243eed3a6545c2a33a21f74f655e3fcb5d2192613cd3db81a93369eb339/cytoolz-1.0.1.tar.gz", hash = "sha256:89cc3161b89e1bb3ed7636f74ed2e55984fd35516904fc878cae216e42b2c7d6", size = 626652 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/1e/d9/f13d66c16cff1fa1cb6c234698029877c456f35f577ef274aba3b86e7c51/cytoolz-1.0.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:cec9af61f71fc3853eb5dca3d42eb07d1f48a4599fa502cbe92adde85f74b042", size = 403515 }, - { url = "https://files.pythonhosted.org/packages/4b/2d/4cdf848a69300c7d44984f2ebbebb3b8576e5449c8dea157298f3bdc4da3/cytoolz-1.0.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:140bbd649dbda01e91add7642149a5987a7c3ccc251f2263de894b89f50b6608", size = 383936 }, - { url = "https://files.pythonhosted.org/packages/72/a4/ccfdd3f0ed9cc818f734b424261f6018fc61e3ec833bf85225a9aca0d994/cytoolz-1.0.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e90124bdc42ff58b88cdea1d24a6bc5f776414a314cc4d94f25c88badb3a16d1", size = 1934569 }, - { url = "https://files.pythonhosted.org/packages/50/fc/38d5344fa595683ad10dc819cfc1d8b9d2b3391ccf3e8cb7bab4899a01f5/cytoolz-1.0.1-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:e74801b751e28f7c5cc3ad264c123954a051f546f2fdfe089f5aa7a12ccfa6da", size = 2015129 }, - { url = "https://files.pythonhosted.org/packages/28/29/75261748dc54a20a927f33641f4e9aac674cfc6d3fbd4f332e10d0b37639/cytoolz-1.0.1-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:582dad4545ddfb5127494ef23f3fa4855f1673a35d50c66f7638e9fb49805089", size = 2000506 }, - { url = "https://files.pythonhosted.org/packages/00/ae/e4ead004cc2698281d153c4a5388638d67cdb5544d6d6cc1e5b3db2bd2a3/cytoolz-1.0.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:dd7bd0618e16efe03bd12f19c2a26a27e6e6b75d7105adb7be1cd2a53fa755d8", size = 1957537 }, - { url = "https://files.pythonhosted.org/packages/4a/ff/4f3aa07f4f47701f7f63df60ce0a5669fa09c256c3d4a33503a9414ea5cc/cytoolz-1.0.1-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:d74cca6acf1c4af58b2e4a89cc565ed61c5e201de2e434748c93e5a0f5c541a5", size = 1863331 }, - { url = "https://files.pythonhosted.org/packages/a2/29/654f57f2a9b8e9765a4ab876765f64f94530b61fc6471a07feea42ece6d4/cytoolz-1.0.1-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:823a3763828d8d457f542b2a45d75d6b4ced5e470b5c7cf2ed66a02f508ed442", size = 1849938 }, - { url = "https://files.pythonhosted.org/packages/bc/7b/11f457db6b291060a98315ab2c7198077d8bddeeebe5f7126d9dad98cc54/cytoolz-1.0.1-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:51633a14e6844c61db1d68c1ffd077cf949f5c99c60ed5f1e265b9e2966f1b52", size = 1852345 }, - { url = "https://files.pythonhosted.org/packages/6b/92/0dccc96ce0323be236d404f5084479b79b747fa0e74e43a270e95868b5f9/cytoolz-1.0.1-cp310-cp310-musllinux_1_2_ppc64le.whl", hash = "sha256:f3ec9b01c45348f1d0d712507d54c2bfd69c62fbd7c9ef555c9d8298693c2432", size = 1989877 }, - { url = "https://files.pythonhosted.org/packages/a3/c8/1c5203a81200bae51aa8f7b5fad613f695bf1afa03f16251ca23ecb2ef9f/cytoolz-1.0.1-cp310-cp310-musllinux_1_2_s390x.whl", hash = "sha256:1855022b712a9c7a5bce354517ab4727a38095f81e2d23d3eabaf1daeb6a3b3c", size = 1994492 }, - { url = "https://files.pythonhosted.org/packages/e2/8a/04bc193c4d7ced8ef6bb62cdcd0bf40b5e5eb26586ed2cfb4433ec7dfd0a/cytoolz-1.0.1-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:9930f7288c4866a1dc1cc87174f0c6ff4cad1671eb1f6306808aa6c445857d78", size = 1896077 }, - { url = "https://files.pythonhosted.org/packages/21/a5/bee63a58f51d2c74856db66e6119a014464ff8cb1c9387fa4bd2d94e49b0/cytoolz-1.0.1-cp310-cp310-win32.whl", hash = "sha256:a9baad795d72fadc3445ccd0f122abfdbdf94269157e6d6d4835636dad318804", size = 322135 }, - { url = "https://files.pythonhosted.org/packages/e8/16/7abfb1685e8b7f2838264551ee33651748994813f566ac4c3d737dfe90e5/cytoolz-1.0.1-cp310-cp310-win_amd64.whl", hash = "sha256:ad95b386a84e18e1f6136f6d343d2509d4c3aae9f5a536f3dc96808fcc56a8cf", size = 363599 }, - { url = "https://files.pythonhosted.org/packages/d9/f7/ef2a10daaec5c0f7d781d50758c6187eee484256e356ae8ef178d6c48497/cytoolz-1.0.1-pp310-pypy310_pp73-macosx_10_15_x86_64.whl", hash = "sha256:83d19d55738ad9c60763b94f3f6d3c6e4de979aeb8d76841c1401081e0e58d96", size = 345702 }, - { url = "https://files.pythonhosted.org/packages/c8/14/53c84adddedb67ff1546abb86fea04d26e24298c3ceab8436d20122ed0b9/cytoolz-1.0.1-pp310-pypy310_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:f112a71fad6ea824578e6393765ce5c054603afe1471a5c753ff6c67fd872d10", size = 385695 }, - { url = "https://files.pythonhosted.org/packages/bd/80/3ae356c5e7b8d7dc7d1adb52f6932fee85cd748ed4e1217c269d2dfd610f/cytoolz-1.0.1-pp310-pypy310_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:5a515df8f8aa6e1eaaf397761a6e4aff2eef73b5f920aedf271416d5471ae5ee", size = 406261 }, - { url = "https://files.pythonhosted.org/packages/0c/31/8e43761ffc82d90bf9cab7e0959712eedcd1e33c211397e143dd42d7af57/cytoolz-1.0.1-pp310-pypy310_pp73-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:92c398e7b7023460bea2edffe5fcd0a76029580f06c3f6938ac3d198b47156f3", size = 397207 }, - { url = "https://files.pythonhosted.org/packages/d1/b9/fe9da37090b6444c65f848a83e390f87d8cb43d6a4df46de1556ad7e5ceb/cytoolz-1.0.1-pp310-pypy310_pp73-win_amd64.whl", hash = "sha256:3237e56211e03b13df47435b2369f5df281e02b04ad80a948ebd199b7bc10a47", size = 343358 }, -] - -[[package]] -name = "datasets" -version = "3.6.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "dill" }, - { name = "filelock" }, - { name = "fsspec", extra = ["http"] }, - { name = "huggingface-hub" }, - { name = "multiprocess" }, - { name = "numpy" }, - { name = "packaging" }, - { name = "pandas" }, - { name = "pyarrow" }, - { name = "pyyaml" }, - { name = "requests" }, - { name = "tqdm" }, - { name = "xxhash" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/1a/89/d3d6fef58a488f8569c82fd293ab7cbd4250244d67f425dcae64c63800ea/datasets-3.6.0.tar.gz", hash = "sha256:1b2bf43b19776e2787e181cfd329cb0ca1a358ea014780c3581e0f276375e041", size = 569336 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/20/34/a08b0ee99715eaba118cbe19a71f7b5e2425c2718ef96007c325944a1152/datasets-3.6.0-py3-none-any.whl", hash = "sha256:25000c4a2c0873a710df127d08a202a06eab7bf42441a6bc278b499c2f72cd1b", size = 491546 }, -] - -[[package]] -name = "debugpy" -version = "1.8.14" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/bd/75/087fe07d40f490a78782ff3b0a30e3968936854105487decdb33446d4b0e/debugpy-1.8.14.tar.gz", hash = "sha256:7cd287184318416850aa8b60ac90105837bb1e59531898c07569d197d2ed5322", size = 1641444 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/fc/df/156df75a41aaebd97cee9d3870fe68f8001b6c1c4ca023e221cfce69bece/debugpy-1.8.14-cp310-cp310-macosx_14_0_x86_64.whl", hash = "sha256:93fee753097e85623cab1c0e6a68c76308cd9f13ffdf44127e6fab4fbf024339", size = 2076510 }, - { url = "https://files.pythonhosted.org/packages/69/cd/4fc391607bca0996db5f3658762106e3d2427beaef9bfd363fd370a3c054/debugpy-1.8.14-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3d937d93ae4fa51cdc94d3e865f535f185d5f9748efb41d0d49e33bf3365bd79", size = 3559614 }, - { url = "https://files.pythonhosted.org/packages/1a/42/4e6d2b9d63e002db79edfd0cb5656f1c403958915e0e73ab3e9220012eec/debugpy-1.8.14-cp310-cp310-win32.whl", hash = "sha256:c442f20577b38cc7a9aafecffe1094f78f07fb8423c3dddb384e6b8f49fd2987", size = 5208588 }, - { url = "https://files.pythonhosted.org/packages/97/b1/cc9e4e5faadc9d00df1a64a3c2d5c5f4b9df28196c39ada06361c5141f89/debugpy-1.8.14-cp310-cp310-win_amd64.whl", hash = "sha256:f117dedda6d969c5c9483e23f573b38f4e39412845c7bc487b6f2648df30fe84", size = 5241043 }, - { url = "https://files.pythonhosted.org/packages/97/1a/481f33c37ee3ac8040d3d51fc4c4e4e7e61cb08b8bc8971d6032acc2279f/debugpy-1.8.14-py2.py3-none-any.whl", hash = "sha256:5cd9a579d553b6cb9759a7908a41988ee6280b961f24f63336835d9418216a20", size = 5256230 }, -] - -[[package]] -name = "decorator" -version = "5.2.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/43/fa/6d96a0978d19e17b68d634497769987b16c8f4cd0a7a05048bec693caa6b/decorator-5.2.1.tar.gz", hash = "sha256:65f266143752f734b0a7cc83c46f4618af75b8c5911b00ccb61d0ac9b6da0360", size = 56711 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/4e/8c/f3147f5c4b73e7550fe5f9352eaa956ae838d5c51eb58e7a25b9f3e2643b/decorator-5.2.1-py3-none-any.whl", hash = "sha256:d316bb415a2d9e2d2b3abcc4084c6502fc09240e292cd76a76afc106a1c8e04a", size = 9190 }, -] - -[[package]] -name = "defusedxml" -version = "0.7.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/0f/d5/c66da9b79e5bdb124974bfe172b4daf3c984ebd9c2a06e2b8a4dc7331c72/defusedxml-0.7.1.tar.gz", hash = "sha256:1bb3032db185915b62d7c6209c5a8792be6a32ab2fedacc84e01b52c51aa3e69", size = 75520 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/07/6c/aa3f2f849e01cb6a001cd8554a88d4c77c5c1a31c95bdf1cf9301e6d9ef4/defusedxml-0.7.1-py2.py3-none-any.whl", hash = "sha256:a352e7e428770286cc899e2542b6cdaedb2b4953ff269a210103ec58f6198a61", size = 25604 }, -] - -[[package]] -name = "dill" -version = "0.3.8" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/17/4d/ac7ffa80c69ea1df30a8aa11b3578692a5118e7cd1aa157e3ef73b092d15/dill-0.3.8.tar.gz", hash = "sha256:3ebe3c479ad625c4553aca177444d89b486b1d84982eeacded644afc0cf797ca", size = 184847 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/c9/7a/cef76fd8438a42f96db64ddaa85280485a9c395e7df3db8158cfec1eee34/dill-0.3.8-py3-none-any.whl", hash = "sha256:c36ca9ffb54365bdd2f8eb3eff7d2a21237f8452b57ace88b1ac615b7e815bd7", size = 116252 }, -] - -[[package]] -name = "distance" -version = "0.1.3" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/5c/1a/883e47df323437aefa0d0a92ccfb38895d9416bd0b56262c2e46a47767b8/Distance-0.1.3.tar.gz", hash = "sha256:60807584f5b6003f5c521aa73f39f51f631de3be5cccc5a1d67166fcbf0d4551", size = 180271 } - -[[package]] -name = "docopt" -version = "0.6.2" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/a2/55/8f8cab2afd404cf578136ef2cc5dfb50baa1761b68c9da1fb1e4eed343c9/docopt-0.6.2.tar.gz", hash = "sha256:49b3a825280bd66b3aa83585ef59c4a8c82f2c8a522dbe754a8bc8d08c85c491", size = 25901 } - -[[package]] -name = "editdistance" -version = "0.8.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/d5/18/9f4f975ca87a390832b1c22478f3702fcdf739f83211e24d054b7551270d/editdistance-0.8.1.tar.gz", hash = "sha256:d1cdf80a5d5014b0c9126a69a42ce55a457b457f6986ff69ca98e4fe4d2d8fed", size = 50006 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/09/c9/302658ce7f4c537a4e85cf578d11bbf7af120a712e1d78fedc6cb8823c65/editdistance-0.8.1-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:adeb705f32b93accc74960d227875abff150ee42d676e428536361fe5f8f5388", size = 106150 }, - { url = "https://files.pythonhosted.org/packages/45/80/0b3c7d2c0e183725986fea5dd2df11f0b4b46320e9a64f6077a121ab1f64/editdistance-0.8.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:3de77951b105d0972deec7684a0b3d1a9dee69c9b5d34f6e2acc0d76cd4a1c52", size = 80551 }, - { url = "https://files.pythonhosted.org/packages/b5/14/681460965c6a4a48321b07f88de2273d097fdca0491ff55db891aacbd291/editdistance-0.8.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:5e88efb052d45e924606c305cb833a80579dca3e8e4ff01309d50ba2c1c0bbd5", size = 79142 }, - { url = "https://files.pythonhosted.org/packages/ed/0d/abdbc8e394a9461cf2ae27c16564fadaa65f52bd242dd1582ae5e7736dc3/editdistance-0.8.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0247e7a1e9c66ea75211a97e725366bff19a52aac2c838ed5f90025630e976dd", size = 396768 }, - { url = "https://files.pythonhosted.org/packages/c2/fb/2940d26ebda12efd280ae939436f17ac482930d862df9e774cb8b771ab03/editdistance-0.8.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:67d143429a49ab552411505f550a0fb4285a1d4336e096804d233ec495ac20fc", size = 401846 }, - { url = "https://files.pythonhosted.org/packages/53/cc/c63d75c7f387d4df0645682c1ab8706c2dfe5c9c0c4999723ce9a3ba0853/editdistance-0.8.1-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ca9d3be2b10e5d44a950a4bd1e84bca9ebbecd364bce0cf5693bf8224c78eaef", size = 397543 }, - { url = "https://files.pythonhosted.org/packages/8e/38/bb0f734a7571e093184606b930734b12da5b6bff2635eba9312fe4536dd9/editdistance-0.8.1-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:5c72aa1df8535f2e2b3d8773a1a7da091bc1a7e52bb396e7e48d375ba687e7b2", size = 898934 }, - { url = "https://files.pythonhosted.org/packages/1c/9f/624fc7a09918f850a057465f02e86f269e139a457f48ff8cabfb12701756/editdistance-0.8.1-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:9a606c34a2a6cc190e4fffc856b36333cdcf1f1fab5b22bd3088e585c22d6ca0", size = 959637 }, - { url = "https://files.pythonhosted.org/packages/5e/5c/7fa6cc277f91c477ee370807d51c1826891dc6dfc307544223ce7f2687de/editdistance-0.8.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:5af173d442ffac33b7c7990132f97f88818a3abf4b21c0c702a7022df37c0c5c", size = 911024 }, - { url = "https://files.pythonhosted.org/packages/ad/97/556215f71184291155aee340a6d34f0676e7238fdfd10615b6b775ce25fe/editdistance-0.8.1-cp310-cp310-win32.whl", hash = "sha256:fd64b58f5a7b59afd9d75982aaeeacd2a98498bf472fa0360c122ffe6ea4c871", size = 80834 }, - { url = "https://files.pythonhosted.org/packages/c8/d1/7ec5f5cbb95838d0eff7f980a660c81acd1363d658f2f5d4ceba38877c5a/editdistance-0.8.1-cp310-cp310-win_amd64.whl", hash = "sha256:6c7c62c3cae45ca1fa01bb2722b297b9de1e3a244ac44cfba88bdcb488fe6aee", size = 79614 }, - { url = "https://files.pythonhosted.org/packages/d4/4c/c9d02eeb47815d35f8d324b52f6704ea7beb032bcb209358cac44047d413/editdistance-0.8.1-pp310-pypy310_pp73-macosx_10_9_x86_64.whl", hash = "sha256:a4a90c6b03094c07358572027a8d0a13cca7450b1aa6caca98a5f1fa4f0b8961", size = 76455 }, - { url = "https://files.pythonhosted.org/packages/af/b0/2818fa6a24595dac069b0bfb9d05658406779a1ded8fd2b0c9066396cf99/editdistance-0.8.1-pp310-pypy310_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:510a4f9ced348a4fd89ae2e102357d4d801a771e29bb2bc2f130a1692193407f", size = 84104 }, - { url = "https://files.pythonhosted.org/packages/1f/d1/3d5e09bcf7fdb7aed705bf74047a8634bd2b8fd92177c25a2547e6dbadfb/editdistance-0.8.1-pp310-pypy310_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:4787fa7228ba6a34b430066d174320f011d605015baa7299c2c4911e6ea6bd46", size = 89058 }, - { url = "https://files.pythonhosted.org/packages/cd/88/fca5d7b1a1edf66ce1e5b6b60bff75842e6814b4f5facbdf4585d88c912d/editdistance-0.8.1-pp310-pypy310_pp73-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ee02601375073afccd6b4d811129ce1cb696d47db734784d8dbd1fddcea75447", size = 84635 }, - { url = "https://files.pythonhosted.org/packages/a9/91/0e6285bbe2358d81fd16313d30306b2d0036387348f7bc11d8c076ca3c72/editdistance-0.8.1-pp310-pypy310_pp73-win_amd64.whl", hash = "sha256:bc7ad9f9a20e6f351523de77c59249f005242e3f317b5de45d02c378d24f6531", size = 77389 }, -] - -[[package]] -name = "einops" -version = "0.8.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/e5/81/df4fbe24dff8ba3934af99044188e20a98ed441ad17a274539b74e82e126/einops-0.8.1.tar.gz", hash = "sha256:de5d960a7a761225532e0f1959e5315ebeafc0cd43394732f103ca44b9837e84", size = 54805 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/87/62/9773de14fe6c45c23649e98b83231fffd7b9892b6cf863251dc2afa73643/einops-0.8.1-py3-none-any.whl", hash = "sha256:919387eb55330f5757c6bea9165c5ff5cfe63a642682ea788a6d472576d81737", size = 64359 }, -] - -[[package]] -name = "exceptiongroup" -version = "1.3.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "typing-extensions", marker = "python_full_version == '3.10.12'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/0b/9f/a65090624ecf468cdca03533906e7c69ed7588582240cfe7cc9e770b50eb/exceptiongroup-1.3.0.tar.gz", hash = "sha256:b241f5885f560bc56a59ee63ca4c6a8bfa46ae4ad651af316d4e81817bb9fd88", size = 29749 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/36/f4/c6e662dade71f56cd2f3735141b265c3c79293c109549c1e6933b0651ffc/exceptiongroup-1.3.0-py3-none-any.whl", hash = "sha256:4d111e6e0c13d0644cad6ddaa7ed0261a0b36971f6d23e7ec9b4b9097da78a10", size = 16674 }, -] - -[[package]] -name = "executing" -version = "2.2.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/91/50/a9d80c47ff289c611ff12e63f7c5d13942c65d68125160cefd768c73e6e4/executing-2.2.0.tar.gz", hash = "sha256:5d108c028108fe2551d1a7b2e8b713341e2cb4fc0aa7dcf966fa4327a5226755", size = 978693 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/7b/8f/c4d9bafc34ad7ad5d8dc16dd1347ee0e507a52c3adb6bfa8887e1c6a26ba/executing-2.2.0-py2.py3-none-any.whl", hash = "sha256:11387150cad388d62750327a53d3339fad4888b39a6fe233c3afbb54ecffd3aa", size = 26702 }, -] - -[[package]] -name = "fastjsonschema" -version = "2.21.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/8b/50/4b769ce1ac4071a1ef6d86b1a3fb56cdc3a37615e8c5519e1af96cdac366/fastjsonschema-2.21.1.tar.gz", hash = "sha256:794d4f0a58f848961ba16af7b9c85a3e88cd360df008c59aac6fc5ae9323b5d4", size = 373939 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/90/2b/0817a2b257fe88725c25589d89aec060581aabf668707a8d03b2e9e0cb2a/fastjsonschema-2.21.1-py3-none-any.whl", hash = "sha256:c9e5b7e908310918cf494a434eeb31384dd84a98b57a30bcb1f535015b554667", size = 23924 }, -] - -[[package]] -name = "fiddle" -version = "0.3.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "absl-py" }, - { name = "graphviz" }, - { name = "libcst" }, - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/73/36/7a4fac76351619b36bbc7937abf59f7b601326dc4efc253b3c16819f782a/fiddle-0.3.0.tar.gz", hash = "sha256:5d083d3299a479868345513385a6c5546141bd92086c15d3dcbf8008a90075d3", size = 277884 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/3b/98/a38e949a91ff9e15874487fd8329ff53c25f3413c0cfc809eb6ff7eb7fa1/fiddle-0.3.0-py3-none-any.whl", hash = "sha256:f4824541c103a94a2f33f6c93eeddf6007c3a7300440087a95907f3e74362e61", size = 419830 }, -] - -[[package]] -name = "filelock" -version = "3.18.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/0a/10/c23352565a6544bdc5353e0b15fc1c563352101f30e24bf500207a54df9a/filelock-3.18.0.tar.gz", hash = "sha256:adbc88eabb99d2fec8c9c1b229b171f18afa655400173ddc653d5d01501fb9f2", size = 18075 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/4d/36/2a115987e2d8c300a974597416d9de88f2444426de9571f4b59b2cca3acc/filelock-3.18.0-py3-none-any.whl", hash = "sha256:c401f4f8377c4464e6db25fff06205fd89bdd83b65eb0488ed1b160f780e21de", size = 16215 }, -] - -[[package]] -name = "fonttools" -version = "4.58.4" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/2e/5a/1124b2c8cb3a8015faf552e92714040bcdbc145dfa29928891b02d147a18/fonttools-4.58.4.tar.gz", hash = "sha256:928a8009b9884ed3aae17724b960987575155ca23c6f0b8146e400cc9e0d44ba", size = 3525026 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/ed/86/d22c24caa574449b56e994ed1a96d23b23af85557fb62a92df96439d3f6c/fonttools-4.58.4-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:834542f13fee7625ad753b2db035edb674b07522fcbdd0ed9e9a9e2a1034467f", size = 2748349 }, - { url = "https://files.pythonhosted.org/packages/f9/b8/384aca93856def00e7de30341f1e27f439694857d82c35d74a809c705ed0/fonttools-4.58.4-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:2e6c61ce330142525296170cd65666e46121fc0d44383cbbcfa39cf8f58383df", size = 2318565 }, - { url = "https://files.pythonhosted.org/packages/1a/f2/273edfdc8d9db89ecfbbf659bd894f7e07b6d53448b19837a4bdba148d17/fonttools-4.58.4-cp310-cp310-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:e9c75f8faa29579c0fbf29b56ae6a3660c6c025f3b671803cb6a9caa7e4e3a98", size = 4838855 }, - { url = "https://files.pythonhosted.org/packages/13/fa/403703548c093c30b52ab37e109b369558afa221130e67f06bef7513f28a/fonttools-4.58.4-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:88dedcedbd5549e35b2ea3db3de02579c27e62e51af56779c021e7b33caadd0e", size = 4767637 }, - { url = "https://files.pythonhosted.org/packages/6e/a8/3380e1e0bff6defb0f81c9abf274a5b4a0f30bc8cab4fd4e346c6f923b4c/fonttools-4.58.4-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:ae80a895adab43586f4da1521d58fd4f4377cef322ee0cc205abcefa3a5effc3", size = 4819397 }, - { url = "https://files.pythonhosted.org/packages/cd/1b/99e47eb17a8ca51d808622a4658584fa8f340857438a4e9d7ac326d4a041/fonttools-4.58.4-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:0d3acc7f0d151da116e87a182aefb569cf0a3c8e0fd4c9cd0a7c1e7d3e7adb26", size = 4926641 }, - { url = "https://files.pythonhosted.org/packages/31/75/415254408f038e35b36c8525fc31feb8561f98445688dd2267c23eafd7a2/fonttools-4.58.4-cp310-cp310-win32.whl", hash = "sha256:1244f69686008e7e8d2581d9f37eef330a73fee3843f1107993eb82c9d306577", size = 2201917 }, - { url = "https://files.pythonhosted.org/packages/c5/69/f019a15ed2946317c5318e1bcc8876f8a54a313848604ad1d4cfc4c07916/fonttools-4.58.4-cp310-cp310-win_amd64.whl", hash = "sha256:2a66c0af8a01eb2b78645af60f3b787de5fe5eb1fd8348163715b80bdbfbde1f", size = 2246327 }, - { url = "https://files.pythonhosted.org/packages/0b/2f/c536b5b9bb3c071e91d536a4d11f969e911dbb6b227939f4c5b0bca090df/fonttools-4.58.4-py3-none-any.whl", hash = "sha256:a10ce13a13f26cbb9f37512a4346bb437ad7e002ff6fa966a7ce7ff5ac3528bd", size = 1114660 }, -] - -[[package]] -name = "fqdn" -version = "1.5.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/30/3e/a80a8c077fd798951169626cde3e239adeba7dab75deb3555716415bd9b0/fqdn-1.5.1.tar.gz", hash = "sha256:105ed3677e767fb5ca086a0c1f4bb66ebc3c100be518f0e0d755d9eae164d89f", size = 6015 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/cf/58/8acf1b3e91c58313ce5cb67df61001fc9dcd21be4fadb76c1a2d540e09ed/fqdn-1.5.1-py3-none-any.whl", hash = "sha256:3a179af3761e4df6eb2e026ff9e1a3033d3587bf980a0b1b2e1e5d08d7358014", size = 9121 }, -] - -[[package]] -name = "frozenlist" -version = "1.7.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/79/b1/b64018016eeb087db503b038296fd782586432b9c077fc5c7839e9cb6ef6/frozenlist-1.7.0.tar.gz", hash = "sha256:2e310d81923c2437ea8670467121cc3e9b0f76d3043cc1d2331d56c7fb7a3a8f", size = 45078 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/af/36/0da0a49409f6b47cc2d060dc8c9040b897b5902a8a4e37d9bc1deb11f680/frozenlist-1.7.0-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:cc4df77d638aa2ed703b878dd093725b72a824c3c546c076e8fdf276f78ee84a", size = 81304 }, - { url = "https://files.pythonhosted.org/packages/77/f0/77c11d13d39513b298e267b22eb6cb559c103d56f155aa9a49097221f0b6/frozenlist-1.7.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:716a9973a2cc963160394f701964fe25012600f3d311f60c790400b00e568b61", size = 47735 }, - { url = "https://files.pythonhosted.org/packages/37/12/9d07fa18971a44150593de56b2f2947c46604819976784bcf6ea0d5db43b/frozenlist-1.7.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:a0fd1bad056a3600047fb9462cff4c5322cebc59ebf5d0a3725e0ee78955001d", size = 46775 }, - { url = "https://files.pythonhosted.org/packages/70/34/f73539227e06288fcd1f8a76853e755b2b48bca6747e99e283111c18bcd4/frozenlist-1.7.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3789ebc19cb811163e70fe2bd354cea097254ce6e707ae42e56f45e31e96cb8e", size = 224644 }, - { url = "https://files.pythonhosted.org/packages/fb/68/c1d9c2f4a6e438e14613bad0f2973567586610cc22dcb1e1241da71de9d3/frozenlist-1.7.0-cp310-cp310-manylinux_2_17_armv7l.manylinux2014_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:af369aa35ee34f132fcfad5be45fbfcde0e3a5f6a1ec0712857f286b7d20cca9", size = 222125 }, - { url = "https://files.pythonhosted.org/packages/b9/d0/98e8f9a515228d708344d7c6986752be3e3192d1795f748c24bcf154ad99/frozenlist-1.7.0-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ac64b6478722eeb7a3313d494f8342ef3478dff539d17002f849101b212ef97c", size = 233455 }, - { url = "https://files.pythonhosted.org/packages/79/df/8a11bcec5600557f40338407d3e5bea80376ed1c01a6c0910fcfdc4b8993/frozenlist-1.7.0-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:f89f65d85774f1797239693cef07ad4c97fdd0639544bad9ac4b869782eb1981", size = 227339 }, - { url = "https://files.pythonhosted.org/packages/50/82/41cb97d9c9a5ff94438c63cc343eb7980dac4187eb625a51bdfdb7707314/frozenlist-1.7.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:1073557c941395fdfcfac13eb2456cb8aad89f9de27bae29fabca8e563b12615", size = 212969 }, - { url = "https://files.pythonhosted.org/packages/13/47/f9179ee5ee4f55629e4f28c660b3fdf2775c8bfde8f9c53f2de2d93f52a9/frozenlist-1.7.0-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1ed8d2fa095aae4bdc7fdd80351009a48d286635edffee66bf865e37a9125c50", size = 222862 }, - { url = "https://files.pythonhosted.org/packages/1a/52/df81e41ec6b953902c8b7e3a83bee48b195cb0e5ec2eabae5d8330c78038/frozenlist-1.7.0-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:24c34bea555fe42d9f928ba0a740c553088500377448febecaa82cc3e88aa1fa", size = 222492 }, - { url = "https://files.pythonhosted.org/packages/84/17/30d6ea87fa95a9408245a948604b82c1a4b8b3e153cea596421a2aef2754/frozenlist-1.7.0-cp310-cp310-musllinux_1_2_armv7l.whl", hash = "sha256:69cac419ac6a6baad202c85aaf467b65ac860ac2e7f2ac1686dc40dbb52f6577", size = 238250 }, - { url = "https://files.pythonhosted.org/packages/8f/00/ecbeb51669e3c3df76cf2ddd66ae3e48345ec213a55e3887d216eb4fbab3/frozenlist-1.7.0-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:960d67d0611f4c87da7e2ae2eacf7ea81a5be967861e0c63cf205215afbfac59", size = 218720 }, - { url = "https://files.pythonhosted.org/packages/1a/c0/c224ce0e0eb31cc57f67742071bb470ba8246623c1823a7530be0e76164c/frozenlist-1.7.0-cp310-cp310-musllinux_1_2_ppc64le.whl", hash = "sha256:41be2964bd4b15bf575e5daee5a5ce7ed3115320fb3c2b71fca05582ffa4dc9e", size = 232585 }, - { url = "https://files.pythonhosted.org/packages/55/3c/34cb694abf532f31f365106deebdeac9e45c19304d83cf7d51ebbb4ca4d1/frozenlist-1.7.0-cp310-cp310-musllinux_1_2_s390x.whl", hash = "sha256:46d84d49e00c9429238a7ce02dc0be8f6d7cd0cd405abd1bebdc991bf27c15bd", size = 234248 }, - { url = "https://files.pythonhosted.org/packages/98/c0/2052d8b6cecda2e70bd81299e3512fa332abb6dcd2969b9c80dfcdddbf75/frozenlist-1.7.0-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:15900082e886edb37480335d9d518cec978afc69ccbc30bd18610b7c1b22a718", size = 221621 }, - { url = "https://files.pythonhosted.org/packages/c5/bf/7dcebae315436903b1d98ffb791a09d674c88480c158aa171958a3ac07f0/frozenlist-1.7.0-cp310-cp310-win32.whl", hash = "sha256:400ddd24ab4e55014bba442d917203c73b2846391dd42ca5e38ff52bb18c3c5e", size = 39578 }, - { url = "https://files.pythonhosted.org/packages/8f/5f/f69818f017fa9a3d24d1ae39763e29b7f60a59e46d5f91b9c6b21622f4cd/frozenlist-1.7.0-cp310-cp310-win_amd64.whl", hash = "sha256:6eb93efb8101ef39d32d50bce242c84bcbddb4f7e9febfa7b524532a239b4464", size = 43830 }, - { url = "https://files.pythonhosted.org/packages/ee/45/b82e3c16be2182bff01179db177fe144d58b5dc787a7d4492c6ed8b9317f/frozenlist-1.7.0-py3-none-any.whl", hash = "sha256:9a5af342e34f7e97caf8c995864c7a396418ae2859cc6fdf1b1073020d516a7e", size = 13106 }, -] - -[[package]] -name = "fsspec" -version = "2024.12.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/ee/11/de70dee31455c546fbc88301971ec03c328f3d1138cfba14263f651e9551/fsspec-2024.12.0.tar.gz", hash = "sha256:670700c977ed2fb51e0d9f9253177ed20cbde4a3e5c0283cc5385b5870c8533f", size = 291600 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/de/86/5486b0188d08aa643e127774a99bac51ffa6cf343e3deb0583956dca5b22/fsspec-2024.12.0-py3-none-any.whl", hash = "sha256:b520aed47ad9804237ff878b504267a3b0b441e97508bd6d2d8774e3db85cee2", size = 183862 }, -] - -[package.optional-dependencies] -http = [ - { name = "aiohttp" }, -] - -[[package]] -name = "funasr" -version = "1.2.7" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "editdistance" }, - { name = "hydra-core" }, - { name = "jaconv" }, - { name = "jamo" }, - { name = "jieba" }, - { name = "kaldiio" }, - { name = "librosa" }, - { name = "modelscope" }, - { name = "oss2" }, - { name = "pytorch-wpe" }, - { name = "pyyaml" }, - { name = "requests" }, - { name = "scipy" }, - { name = "sentencepiece" }, - { name = "soundfile" }, - { name = "tensorboardx" }, - { name = "torch-complex" }, - { name = "tqdm" }, - { name = "umap-learn" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/4d/ca/0f3e0df85bfea75acd5950a3c8da67c857af700b7e781eaa47322e18e593/funasr-1.2.7.tar.gz", hash = "sha256:fe943487bab0137813b213ffab14e819d8762fe1df6f776bb2db0bf89958d5e2", size = 555042 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/f7/6f/491bc744f9d23be35d848479dd21ba6576788e4a2cffbdd410725669fe5c/funasr-1.2.7-py3-none-any.whl", hash = "sha256:b53f748e479e5bf6af172407c50eccaa6818ed91bdf8656abcd7ea6c5e3d2b0d", size = 703175 }, -] - -[[package]] -name = "future" -version = "1.0.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/a7/b2/4140c69c6a66432916b26158687e821ba631a4c9273c474343badf84d3ba/future-1.0.0.tar.gz", hash = "sha256:bd2968309307861edae1458a4f8a4f3598c03be43b97521076aebf5d94c07b05", size = 1228490 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/da/71/ae30dadffc90b9006d77af76b393cb9dfbfc9629f339fc1574a1c52e6806/future-1.0.0-py3-none-any.whl", hash = "sha256:929292d34f5872e70396626ef385ec22355a1fae8ad29e1a734c3e43f9fbc216", size = 491326 }, -] - -[[package]] -name = "g2p-en" -version = "2.1.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "distance" }, - { name = "inflect" }, - { name = "nltk" }, - { name = "numpy" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/5f/22/2c7acbe6164ed6cfd4301e9ad2dbde69c68d22268a0f9b5b0ee6052ed3ab/g2p_en-2.1.0.tar.gz", hash = "sha256:32ecb119827a3b10ea8c1197276f4ea4f44070ae56cbbd01f0f261875f556a58", size = 3116166 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/d7/d9/b77dc634a7a0c0c97716ba97dd0a28cbfa6267c96f359c4f27ed71cbd284/g2p_en-2.1.0-py3-none-any.whl", hash = "sha256:2a7aabf1fc7f270fcc3349881407988c9245173c2413debbe5432f4a4f31319f", size = 3117464 }, -] - -[[package]] -name = "gitdb" -version = "4.0.12" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "smmap" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/72/94/63b0fc47eb32792c7ba1fe1b694daec9a63620db1e313033d18140c2320a/gitdb-4.0.12.tar.gz", hash = "sha256:5ef71f855d191a3326fcfbc0d5da835f26b13fbcba60c32c21091c349ffdb571", size = 394684 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/a0/61/5c78b91c3143ed5c14207f463aecfc8f9dbb5092fb2869baf37c273b2705/gitdb-4.0.12-py3-none-any.whl", hash = "sha256:67073e15955400952c6565cc3e707c554a4eea2e428946f7a4c162fab9bd9bcf", size = 62794 }, -] - -[[package]] -name = "gitpython" -version = "3.1.44" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "gitdb" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/c0/89/37df0b71473153574a5cdef8f242de422a0f5d26d7a9e231e6f169b4ad14/gitpython-3.1.44.tar.gz", hash = "sha256:c87e30b26253bf5418b01b0660f818967f3c503193838337fe5e573331249269", size = 214196 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/1d/9a/4114a9057db2f1462d5c8f8390ab7383925fe1ac012eaa42402ad65c2963/GitPython-3.1.44-py3-none-any.whl", hash = "sha256:9e0e10cda9bed1ee64bc9a6de50e7e38a9c9943241cd7f585f6df3ed28011110", size = 207599 }, -] - -[[package]] -name = "graphviz" -version = "0.21" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/f8/b3/3ac91e9be6b761a4b30d66ff165e54439dcd48b83f4e20d644867215f6ca/graphviz-0.21.tar.gz", hash = "sha256:20743e7183be82aaaa8ad6c93f8893c923bd6658a04c32ee115edb3c8a835f78", size = 200434 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/91/4c/e0ce1ef95d4000ebc1c11801f9b944fa5910ecc15b5e351865763d8657f8/graphviz-0.21-py3-none-any.whl", hash = "sha256:54f33de9f4f911d7e84e4191749cac8cc5653f815b06738c54db9a15ab8b1e42", size = 47300 }, -] - -[[package]] -name = "greenlet" -version = "3.2.4" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/03/b8/704d753a5a45507a7aab61f18db9509302ed3d0a27ac7e0359ec2905b1a6/greenlet-3.2.4.tar.gz", hash = "sha256:0dca0d95ff849f9a364385f36ab49f50065d76964944638be9691e1832e9f86d", size = 188260 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/7d/ed/6bfa4109fcb23a58819600392564fea69cdc6551ffd5e69ccf1d52a40cbc/greenlet-3.2.4-cp310-cp310-macosx_11_0_universal2.whl", hash = "sha256:8c68325b0d0acf8d91dde4e6f930967dd52a5302cd4062932a6b2e7c2969f47c", size = 271061 }, - { url = "https://files.pythonhosted.org/packages/2a/fc/102ec1a2fc015b3a7652abab7acf3541d58c04d3d17a8d3d6a44adae1eb1/greenlet-3.2.4-cp310-cp310-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:94385f101946790ae13da500603491f04a76b6e4c059dab271b3ce2e283b2590", size = 629475 }, - { url = "https://files.pythonhosted.org/packages/c5/26/80383131d55a4ac0fb08d71660fd77e7660b9db6bdb4e8884f46d9f2cc04/greenlet-3.2.4-cp310-cp310-manylinux2014_ppc64le.manylinux_2_17_ppc64le.whl", hash = "sha256:f10fd42b5ee276335863712fa3da6608e93f70629c631bf77145021600abc23c", size = 640802 }, - { url = "https://files.pythonhosted.org/packages/9f/7c/e7833dbcd8f376f3326bd728c845d31dcde4c84268d3921afcae77d90d08/greenlet-3.2.4-cp310-cp310-manylinux2014_s390x.manylinux_2_17_s390x.whl", hash = "sha256:c8c9e331e58180d0d83c5b7999255721b725913ff6bc6cf39fa2a45841a4fd4b", size = 636703 }, - { url = "https://files.pythonhosted.org/packages/e9/49/547b93b7c0428ede7b3f309bc965986874759f7d89e4e04aeddbc9699acb/greenlet-3.2.4-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:58b97143c9cc7b86fc458f215bd0932f1757ce649e05b640fea2e79b54cedb31", size = 635417 }, - { url = "https://files.pythonhosted.org/packages/7f/91/ae2eb6b7979e2f9b035a9f612cf70f1bf54aad4e1d125129bef1eae96f19/greenlet-3.2.4-cp310-cp310-manylinux_2_24_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:c2ca18a03a8cfb5b25bc1cbe20f3d9a4c80d8c3b13ba3df49ac3961af0b1018d", size = 584358 }, - { url = "https://files.pythonhosted.org/packages/f7/85/433de0c9c0252b22b16d413c9407e6cb3b41df7389afc366ca204dbc1393/greenlet-3.2.4-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:9fe0a28a7b952a21e2c062cd5756d34354117796c6d9215a87f55e38d15402c5", size = 1113550 }, - { url = "https://files.pythonhosted.org/packages/a1/8d/88f3ebd2bc96bf7747093696f4335a0a8a4c5acfcf1b757717c0d2474ba3/greenlet-3.2.4-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:8854167e06950ca75b898b104b63cc646573aa5fef1353d4508ecdd1ee76254f", size = 1137126 }, - { url = "https://files.pythonhosted.org/packages/d6/6f/b60b0291d9623c496638c582297ead61f43c4b72eef5e9c926ef4565ec13/greenlet-3.2.4-cp310-cp310-win_amd64.whl", hash = "sha256:73f49b5368b5359d04e18d15828eecc1806033db5233397748f4ca813ff1056c", size = 298654 }, -] - -[[package]] -name = "grpcio" -version = "1.73.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/79/e8/b43b851537da2e2f03fa8be1aef207e5cbfb1a2e014fbb6b40d24c177cd3/grpcio-1.73.1.tar.gz", hash = "sha256:7fce2cd1c0c1116cf3850564ebfc3264fba75d3c74a7414373f1238ea365ef87", size = 12730355 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/8f/51/a5748ab2773d893d099b92653039672f7e26dd35741020972b84d604066f/grpcio-1.73.1-cp310-cp310-linux_armv7l.whl", hash = "sha256:2d70f4ddd0a823436c2624640570ed6097e40935c9194482475fe8e3d9754d55", size = 5365087 }, - { url = "https://files.pythonhosted.org/packages/ae/12/c5ee1a5dfe93dbc2eaa42a219e2bf887250b52e2e2ee5c036c4695f2769c/grpcio-1.73.1-cp310-cp310-macosx_11_0_universal2.whl", hash = "sha256:3841a8a5a66830261ab6a3c2a3dc539ed84e4ab019165f77b3eeb9f0ba621f26", size = 10608921 }, - { url = "https://files.pythonhosted.org/packages/c4/6d/b0c6a8120f02b7d15c5accda6bfc43bc92be70ada3af3ba6d8e077c00374/grpcio-1.73.1-cp310-cp310-manylinux_2_17_aarch64.whl", hash = "sha256:628c30f8e77e0258ab788750ec92059fc3d6628590fb4b7cea8c102503623ed7", size = 5803221 }, - { url = "https://files.pythonhosted.org/packages/a6/7a/3c886d9f1c1e416ae81f7f9c7d1995ae72cd64712d29dab74a6bafacb2d2/grpcio-1.73.1-cp310-cp310-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:67a0468256c9db6d5ecb1fde4bf409d016f42cef649323f0a08a72f352d1358b", size = 6444603 }, - { url = "https://files.pythonhosted.org/packages/42/07/f143a2ff534982c9caa1febcad1c1073cdec732f6ac7545d85555a900a7e/grpcio-1.73.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:68b84d65bbdebd5926eb5c53b0b9ec3b3f83408a30e4c20c373c5337b4219ec5", size = 6040969 }, - { url = "https://files.pythonhosted.org/packages/fb/0f/523131b7c9196d0718e7b2dac0310eb307b4117bdbfef62382e760f7e8bb/grpcio-1.73.1-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:c54796ca22b8349cc594d18b01099e39f2b7ffb586ad83217655781a350ce4da", size = 6132201 }, - { url = "https://files.pythonhosted.org/packages/ad/18/010a055410eef1d3a7a1e477ec9d93b091ac664ad93e9c5f56d6cc04bdee/grpcio-1.73.1-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:75fc8e543962ece2f7ecd32ada2d44c0c8570ae73ec92869f9af8b944863116d", size = 6774718 }, - { url = "https://files.pythonhosted.org/packages/16/11/452bfc1ab39d8ee748837ab8ee56beeae0290861052948785c2c445fb44b/grpcio-1.73.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:6a6037891cd2b1dd1406b388660522e1565ed340b1fea2955b0234bdd941a862", size = 6304362 }, - { url = "https://files.pythonhosted.org/packages/1e/1c/c75ceee626465721e5cb040cf4b271eff817aa97388948660884cb7adffa/grpcio-1.73.1-cp310-cp310-win32.whl", hash = "sha256:cce7265b9617168c2d08ae570fcc2af4eaf72e84f8c710ca657cc546115263af", size = 3679036 }, - { url = "https://files.pythonhosted.org/packages/62/2e/42cb31b6cbd671a7b3dbd97ef33f59088cf60e3cf2141368282e26fafe79/grpcio-1.73.1-cp310-cp310-win_amd64.whl", hash = "sha256:6a2b372e65fad38842050943f42ce8fee00c6f2e8ea4f7754ba7478d26a356ee", size = 4340208 }, -] - -[[package]] -name = "h11" -version = "0.16.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/01/ee/02a2c011bdab74c6fb3c75474d40b3052059d95df7e73351460c8588d963/h11-0.16.0.tar.gz", hash = "sha256:4e35b956cf45792e4caa5885e69fba00bdbc6ffafbfa020300e549b208ee5ff1", size = 101250 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/04/4b/29cac41a4d98d144bf5f6d33995617b185d14b22401f75ca86f384e87ff1/h11-0.16.0-py3-none-any.whl", hash = "sha256:63cf8bbe7522de3bf65932fda1d9c2772064ffb3dae62d55932da54b31cb6c86", size = 37515 }, -] - -[[package]] -name = "hf-xet" -version = "1.1.5" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/ed/d4/7685999e85945ed0d7f0762b686ae7015035390de1161dcea9d5276c134c/hf_xet-1.1.5.tar.gz", hash = "sha256:69ebbcfd9ec44fdc2af73441619eeb06b94ee34511bbcf57cd423820090f5694", size = 495969 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/00/89/a1119eebe2836cb25758e7661d6410d3eae982e2b5e974bcc4d250be9012/hf_xet-1.1.5-cp37-abi3-macosx_10_12_x86_64.whl", hash = "sha256:f52c2fa3635b8c37c7764d8796dfa72706cc4eded19d638331161e82b0792e23", size = 2687929 }, - { url = "https://files.pythonhosted.org/packages/de/5f/2c78e28f309396e71ec8e4e9304a6483dcbc36172b5cea8f291994163425/hf_xet-1.1.5-cp37-abi3-macosx_11_0_arm64.whl", hash = "sha256:9fa6e3ee5d61912c4a113e0708eaaef987047616465ac7aa30f7121a48fc1af8", size = 2556338 }, - { url = "https://files.pythonhosted.org/packages/6d/2f/6cad7b5fe86b7652579346cb7f85156c11761df26435651cbba89376cd2c/hf_xet-1.1.5-cp37-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:fc874b5c843e642f45fd85cda1ce599e123308ad2901ead23d3510a47ff506d1", size = 3102894 }, - { url = "https://files.pythonhosted.org/packages/d0/54/0fcf2b619720a26fbb6cc941e89f2472a522cd963a776c089b189559447f/hf_xet-1.1.5-cp37-abi3-manylinux_2_28_aarch64.whl", hash = "sha256:dbba1660e5d810bd0ea77c511a99e9242d920790d0e63c0e4673ed36c4022d18", size = 3002134 }, - { url = "https://files.pythonhosted.org/packages/f3/92/1d351ac6cef7c4ba8c85744d37ffbfac2d53d0a6c04d2cabeba614640a78/hf_xet-1.1.5-cp37-abi3-musllinux_1_2_aarch64.whl", hash = "sha256:ab34c4c3104133c495785d5d8bba3b1efc99de52c02e759cf711a91fd39d3a14", size = 3171009 }, - { url = "https://files.pythonhosted.org/packages/c9/65/4b2ddb0e3e983f2508528eb4501288ae2f84963586fbdfae596836d5e57a/hf_xet-1.1.5-cp37-abi3-musllinux_1_2_x86_64.whl", hash = "sha256:83088ecea236d5113de478acb2339f92c95b4fb0462acaa30621fac02f5a534a", size = 3279245 }, - { url = "https://files.pythonhosted.org/packages/f0/55/ef77a85ee443ae05a9e9cba1c9f0dd9241eb42da2aeba1dc50f51154c81a/hf_xet-1.1.5-cp37-abi3-win_amd64.whl", hash = "sha256:73e167d9807d166596b4b2f0b585c6d5bd84a26dea32843665a8b58f6edba245", size = 2738931 }, -] - -[[package]] -name = "httpcore" -version = "1.0.9" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "certifi" }, - { name = "h11" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/06/94/82699a10bca87a5556c9c59b5963f2d039dbd239f25bc2a63907a05a14cb/httpcore-1.0.9.tar.gz", hash = "sha256:6e34463af53fd2ab5d807f399a9b45ea31c3dfa2276f15a2c3f00afff6e176e8", size = 85484 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/7e/f5/f66802a942d491edb555dd61e3a9961140fd64c90bce1eafd741609d334d/httpcore-1.0.9-py3-none-any.whl", hash = "sha256:2d400746a40668fc9dec9810239072b40b4484b640a8c38fd654a024c7a1bf55", size = 78784 }, -] - -[[package]] -name = "httpx" -version = "0.28.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "anyio" }, - { name = "certifi" }, - { name = "httpcore" }, - { name = "idna" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/b1/df/48c586a5fe32a0f01324ee087459e112ebb7224f646c0b5023f5e79e9956/httpx-0.28.1.tar.gz", hash = "sha256:75e98c5f16b0f35b567856f597f06ff2270a374470a5c2392242528e3e3e42fc", size = 141406 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/2a/39/e50c7c3a983047577ee07d2a9e53faf5a69493943ec3f6a384bdc792deb2/httpx-0.28.1-py3-none-any.whl", hash = "sha256:d909fcccc110f8c7faf814ca82a9a4d816bc5a6dbfea25d6591d6985b8ba59ad", size = 73517 }, -] - -[[package]] -name = "huggingface-hub" -version = "0.33.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "filelock" }, - { name = "fsspec" }, - { name = "hf-xet", marker = "platform_machine == 'aarch64' or platform_machine == 'amd64' or platform_machine == 'arm64' or platform_machine == 'x86_64'" }, - { name = "packaging" }, - { name = "pyyaml" }, - { name = "requests" }, - { name = "tqdm" }, - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/a4/01/bfe0534a63ce7a2285e90dbb33e8a5b815ff096d8f7743b135c256916589/huggingface_hub-0.33.1.tar.gz", hash = "sha256:589b634f979da3ea4b8bdb3d79f97f547840dc83715918daf0b64209c0844c7b", size = 426728 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/d0/fb/5307bd3612eb0f0e62c3a916ae531d3a31e58fb5c82b58e3ebf7fd6f47a1/huggingface_hub-0.33.1-py3-none-any.whl", hash = "sha256:ec8d7444628210c0ba27e968e3c4c973032d44dcea59ca0d78ef3f612196f095", size = 515377 }, -] - -[[package]] -name = "hydra-core" -version = "1.3.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "antlr4-python3-runtime" }, - { name = "omegaconf" }, - { name = "packaging" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/6d/8e/07e42bc434a847154083b315779b0a81d567154504624e181caf2c71cd98/hydra-core-1.3.2.tar.gz", hash = "sha256:8a878ed67216997c3e9d88a8e72e7b4767e81af37afb4ea3334b269a4390a824", size = 3263494 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/c6/50/e0edd38dcd63fb26a8547f13d28f7a008bc4a3fd4eb4ff030673f22ad41a/hydra_core-1.3.2-py3-none-any.whl", hash = "sha256:fa0238a9e31df3373b35b0bfb672c34cc92718d21f81311d8996a16de1141d8b", size = 154547 }, -] - -[[package]] -name = "hyperpyyaml" -version = "1.2.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "pyyaml" }, - { name = "ruamel-yaml" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/52/e3/3ac46d9a662b037f699a6948b39c8d03bfcff0b592335d5953ba0c55d453/HyperPyYAML-1.2.2.tar.gz", hash = "sha256:bdb734210d18770a262f500fe5755c7a44a5d3b91521b06e24f7a00a36ee0f87", size = 17085 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/33/c9/751b6401887f4b50f9307cc1e53d287b3dc77c375c126aeb6335aff73ccb/HyperPyYAML-1.2.2-py3-none-any.whl", hash = "sha256:3c5864bdc8864b2f0fbd7bc495e7e8fdf2dfd5dd80116f72da27ca96a128bdeb", size = 16118 }, -] - -[[package]] -name = "idna" -version = "3.10" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/f1/70/7703c29685631f5a7590aa73f1f1d3fa9a380e654b86af429e0934a32f7d/idna-3.10.tar.gz", hash = "sha256:12f65c9b470abda6dc35cf8e63cc574b1c52b11df2c86030af0ac09b01b13ea9", size = 190490 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/76/c6/c88e154df9c4e1a2a66ccf0005a88dfb2650c1dffb6f5ce603dfbd452ce3/idna-3.10-py3-none-any.whl", hash = "sha256:946d195a0d259cbba61165e88e65941f16e9b36ea6ddb97f00452bae8b1287d3", size = 70442 }, -] - -[[package]] -name = "inflect" -version = "7.5.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "more-itertools" }, - { name = "typeguard" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/78/c6/943357d44a21fd995723d07ccaddd78023eace03c1846049a2645d4324a3/inflect-7.5.0.tar.gz", hash = "sha256:faf19801c3742ed5a05a8ce388e0d8fe1a07f8d095c82201eb904f5d27ad571f", size = 73751 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/8a/eb/427ed2b20a38a4ee29f24dbe4ae2dafab198674fe9a85e3d6adf9e5f5f41/inflect-7.5.0-py3-none-any.whl", hash = "sha256:2aea70e5e70c35d8350b8097396ec155ffd68def678c7ff97f51aa69c1d92344", size = 35197 }, -] - -[[package]] -name = "intervaltree" -version = "3.1.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "sortedcontainers" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/50/fb/396d568039d21344639db96d940d40eb62befe704ef849b27949ded5c3bb/intervaltree-3.1.0.tar.gz", hash = "sha256:902b1b88936918f9b2a19e0e5eb7ccb430ae45cde4f39ea4b36932920d33952d", size = 32861 } - -[[package]] -name = "ipykernel" -version = "6.29.5" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "appnope", marker = "platform_system == 'Darwin'" }, - { name = "comm" }, - { name = "debugpy" }, - { name = "ipython" }, - { name = "jupyter-client" }, - { name = "jupyter-core" }, - { name = "matplotlib-inline" }, - { name = "nest-asyncio" }, - { name = "packaging" }, - { name = "psutil" }, - { name = "pyzmq" }, - { name = "tornado" }, - { name = "traitlets" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/e9/5c/67594cb0c7055dc50814b21731c22a601101ea3b1b50a9a1b090e11f5d0f/ipykernel-6.29.5.tar.gz", hash = "sha256:f093a22c4a40f8828f8e330a9c297cb93dcab13bd9678ded6de8e5cf81c56215", size = 163367 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/94/5c/368ae6c01c7628438358e6d337c19b05425727fbb221d2a3c4303c372f42/ipykernel-6.29.5-py3-none-any.whl", hash = "sha256:afdb66ba5aa354b09b91379bac28ae4afebbb30e8b39510c9690afb7a10421b5", size = 117173 }, -] - -[[package]] -name = "ipython" -version = "8.37.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "colorama", marker = "sys_platform == 'win32'" }, - { name = "decorator" }, - { name = "exceptiongroup", marker = "python_full_version == '3.10.12'" }, - { name = "jedi" }, - { name = "matplotlib-inline" }, - { name = "pexpect", marker = "sys_platform != 'emscripten' and sys_platform != 'win32'" }, - { name = "prompt-toolkit" }, - { name = "pygments" }, - { name = "stack-data" }, - { name = "traitlets" }, - { name = "typing-extensions", marker = "python_full_version == '3.10.12'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/85/31/10ac88f3357fc276dc8a64e8880c82e80e7459326ae1d0a211b40abf6665/ipython-8.37.0.tar.gz", hash = "sha256:ca815841e1a41a1e6b73a0b08f3038af9b2252564d01fc405356d34033012216", size = 5606088 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/91/d0/274fbf7b0b12643cbbc001ce13e6a5b1607ac4929d1b11c72460152c9fc3/ipython-8.37.0-py3-none-any.whl", hash = "sha256:ed87326596b878932dbcb171e3e698845434d8c61b8d8cd474bf663041a9dcf2", size = 831864 }, -] - -[[package]] -name = "ipywidgets" -version = "8.1.7" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "comm" }, - { name = "ipython" }, - { name = "jupyterlab-widgets" }, - { name = "traitlets" }, - { name = "widgetsnbextension" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/3e/48/d3dbac45c2814cb73812f98dd6b38bbcc957a4e7bb31d6ea9c03bf94ed87/ipywidgets-8.1.7.tar.gz", hash = "sha256:15f1ac050b9ccbefd45dccfbb2ef6bed0029d8278682d569d71b8dd96bee0376", size = 116721 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/58/6a/9166369a2f092bd286d24e6307de555d63616e8ddb373ebad2b5635ca4cd/ipywidgets-8.1.7-py3-none-any.whl", hash = "sha256:764f2602d25471c213919b8a1997df04bef869251db4ca8efba1b76b1bd9f7bb", size = 139806 }, -] - -[[package]] -name = "isoduration" -version = "20.11.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "arrow" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/7c/1a/3c8edc664e06e6bd06cce40c6b22da5f1429aa4224d0c590f3be21c91ead/isoduration-20.11.0.tar.gz", hash = "sha256:ac2f9015137935279eac671f94f89eb00584f940f5dc49462a0c4ee692ba1bd9", size = 11649 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/7b/55/e5326141505c5d5e34c5e0935d2908a74e4561eca44108fbfb9c13d2911a/isoduration-20.11.0-py3-none-any.whl", hash = "sha256:b2904c2a4228c3d44f409c8ae8e2370eb21a26f7ac2ec5446df141dde3452042", size = 11321 }, -] - -[[package]] -name = "jaconv" -version = "0.4.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/d2/e1/670cefc7f00b0e1890e114a37a98ea425f7e06131342aeb9636856ac663c/jaconv-0.4.0.tar.gz", hash = "sha256:32da74b247f276e09a52d6b35c153df2387965cb85a6f034cc8af21d446f8161", size = 17402 } - -[[package]] -name = "jamo" -version = "0.4.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/b1/a2/bda770579809726e929ca6356743f9f50f64a2cbaee578fa9d4824afb00e/jamo-0.4.1.tar.gz", hash = "sha256:ea65cf9d35338d0e0af48d75ff426d8a369b0ebde6f07051c3ac37256f56d025", size = 7386 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/ac/cc/49812faae67f9a24be6ddaf58a2cf7e8c3cbfcf5b762d9414f7103d2ea2c/jamo-0.4.1-py3-none-any.whl", hash = "sha256:d4b94fd23324c606ed2fbc4037c603e2c3a7ae9390c05d3473aea1ccb6b1c3fb", size = 9543 }, -] - -[[package]] -name = "jedi" -version = "0.19.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "parso" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/72/3a/79a912fbd4d8dd6fbb02bf69afd3bb72cf0c729bb3063c6f4498603db17a/jedi-0.19.2.tar.gz", hash = "sha256:4770dc3de41bde3966b02eb84fbcf557fb33cce26ad23da12c742fb50ecb11f0", size = 1231287 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/c0/5a/9cac0c82afec3d09ccd97c8b6502d48f165f9124db81b4bcb90b4af974ee/jedi-0.19.2-py2.py3-none-any.whl", hash = "sha256:a8ef22bde8490f57fe5c7681a3c83cb58874daf72b4784de3cce5b6ef6edb5b9", size = 1572278 }, -] - -[[package]] -name = "jieba" -version = "0.42.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/c6/cb/18eeb235f833b726522d7ebed54f2278ce28ba9438e3135ab0278d9792a2/jieba-0.42.1.tar.gz", hash = "sha256:055ca12f62674fafed09427f176506079bc135638a14e23e25be909131928db2", size = 19214172 } - -[[package]] -name = "jinja2" -version = "3.1.6" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "markupsafe" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/df/bf/f7da0350254c0ed7c72f3e33cef02e048281fec7ecec5f032d4aac52226b/jinja2-3.1.6.tar.gz", hash = "sha256:0137fb05990d35f1275a587e9aee6d56da821fc83491a0fb838183be43f66d6d", size = 245115 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/62/a1/3d680cbfd5f4b8f15abc1d571870c5fc3e594bb582bc3b64ea099db13e56/jinja2-3.1.6-py3-none-any.whl", hash = "sha256:85ece4451f492d0c13c5dd7c13a64681a86afae63a5f347908daf103ce6d2f67", size = 134899 }, -] - -[[package]] -name = "jiwer" -version = "4.0.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "click" }, - { name = "rapidfuzz" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/12/1e/963dfc249d5bbe88079b57a22556e981ddd9208e4b6116e48bdbfc01f26b/jiwer-4.0.0.tar.gz", hash = "sha256:ae9c051469102a61ef0927100baeeb4546f78d180c9b0948281d08eaf44c191e", size = 28074 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/69/c9/172c525330c739a068c01050759a6f855ce16212db10a0359e690a03ac48/jiwer-4.0.0-py3-none-any.whl", hash = "sha256:7efaf0bd336b095d99ddef9dd67e1ee829d75d58aa2a81d9639870b01d6d95ea", size = 23034 }, -] - -[[package]] -name = "jmespath" -version = "0.10.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/3c/56/3f325b1eef9791759784aa5046a8f6a1aff8f7c898a2e34506771d3b99d8/jmespath-0.10.0.tar.gz", hash = "sha256:b85d0567b8666149a93172712e68920734333c0ce7e89b78b3e987f71e5ed4f9", size = 21607 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/07/cb/5f001272b6faeb23c1c9e0acc04d48eaaf5c862c17709d20e3469c6e0139/jmespath-0.10.0-py2.py3-none-any.whl", hash = "sha256:cdf6525904cc597730141d61b36f2e4b8ecc257c420fa2f4549bac2c2d0cb72f", size = 24489 }, -] - -[[package]] -name = "joblib" -version = "1.5.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/dc/fe/0f5a938c54105553436dbff7a61dc4fed4b1b2c98852f8833beaf4d5968f/joblib-1.5.1.tar.gz", hash = "sha256:f4f86e351f39fe3d0d32a9f2c3d8af1ee4cec285aafcb27003dda5205576b444", size = 330475 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/7d/4f/1195bbac8e0c2acc5f740661631d8d750dc38d4a32b23ee5df3cde6f4e0d/joblib-1.5.1-py3-none-any.whl", hash = "sha256:4719a31f054c7d766948dcd83e9613686b27114f190f717cec7eaa2084f8a74a", size = 307746 }, -] - -[[package]] -name = "json5" -version = "0.12.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/12/be/c6c745ec4c4539b25a278b70e29793f10382947df0d9efba2fa09120895d/json5-0.12.0.tar.gz", hash = "sha256:0b4b6ff56801a1c7dc817b0241bca4ce474a0e6a163bfef3fc594d3fd263ff3a", size = 51907 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/41/9f/3500910d5a98549e3098807493851eeef2b89cdd3032227558a104dfe926/json5-0.12.0-py3-none-any.whl", hash = "sha256:6d37aa6c08b0609f16e1ec5ff94697e2cbbfbad5ac112afa05794da9ab7810db", size = 36079 }, -] - -[[package]] -name = "jsonpointer" -version = "3.0.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/6a/0a/eebeb1fa92507ea94016a2a790b93c2ae41a7e18778f85471dc54475ed25/jsonpointer-3.0.0.tar.gz", hash = "sha256:2b2d729f2091522d61c3b31f82e11870f60b68f43fbc705cb76bf4b832af59ef", size = 9114 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/71/92/5e77f98553e9e75130c78900d000368476aed74276eb8ae8796f65f00918/jsonpointer-3.0.0-py2.py3-none-any.whl", hash = "sha256:13e088adc14fca8b6aa8177c044e12701e6ad4b28ff10e65f2267a90109c9942", size = 7595 }, -] - -[[package]] -name = "jsonschema" -version = "4.24.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "attrs" }, - { name = "jsonschema-specifications" }, - { name = "referencing" }, - { name = "rpds-py" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/bf/d3/1cf5326b923a53515d8f3a2cd442e6d7e94fcc444716e879ea70a0ce3177/jsonschema-4.24.0.tar.gz", hash = "sha256:0b4e8069eb12aedfa881333004bccaec24ecef5a8a6a4b6df142b2cc9599d196", size = 353480 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/a2/3d/023389198f69c722d039351050738d6755376c8fd343e91dc493ea485905/jsonschema-4.24.0-py3-none-any.whl", hash = "sha256:a462455f19f5faf404a7902952b6f0e3ce868f3ee09a359b05eca6673bd8412d", size = 88709 }, -] - -[package.optional-dependencies] -format-nongpl = [ - { name = "fqdn" }, - { name = "idna" }, - { name = "isoduration" }, - { name = "jsonpointer" }, - { name = "rfc3339-validator" }, - { name = "rfc3986-validator" }, - { name = "uri-template" }, - { name = "webcolors" }, -] - -[[package]] -name = "jsonschema-specifications" -version = "2025.4.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "referencing" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/bf/ce/46fbd9c8119cfc3581ee5643ea49464d168028cfb5caff5fc0596d0cf914/jsonschema_specifications-2025.4.1.tar.gz", hash = "sha256:630159c9f4dbea161a6a2205c3011cc4f18ff381b189fff48bb39b9bf26ae608", size = 15513 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/01/0e/b27cdbaccf30b890c40ed1da9fd4a3593a5cf94dae54fb34f8a4b74fcd3f/jsonschema_specifications-2025.4.1-py3-none-any.whl", hash = "sha256:4653bffbd6584f7de83a67e0d620ef16900b390ddc7939d56684d6c81e33f1af", size = 18437 }, -] - -[[package]] -name = "julius" -version = "0.2.7" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "torch" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/a1/19/c9e1596b5572c786b93428d0904280e964c930fae7e6c9368ed9e1b63922/julius-0.2.7.tar.gz", hash = "sha256:3c0f5f5306d7d6016fcc95196b274cae6f07e2c9596eed314e4e7641554fbb08", size = 59640 } - -[[package]] -name = "jupyter" -version = "1.1.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "ipykernel" }, - { name = "ipywidgets" }, - { name = "jupyter-console" }, - { name = "jupyterlab" }, - { name = "nbconvert" }, - { name = "notebook" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/58/f3/af28ea964ab8bc1e472dba2e82627d36d470c51f5cd38c37502eeffaa25e/jupyter-1.1.1.tar.gz", hash = "sha256:d55467bceabdea49d7e3624af7e33d59c37fff53ed3a350e1ac957bed731de7a", size = 5714959 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/38/64/285f20a31679bf547b75602702f7800e74dbabae36ef324f716c02804753/jupyter-1.1.1-py2.py3-none-any.whl", hash = "sha256:7a59533c22af65439b24bbe60373a4e95af8f16ac65a6c00820ad378e3f7cc83", size = 2657 }, -] - -[[package]] -name = "jupyter-client" -version = "8.6.3" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "jupyter-core" }, - { name = "python-dateutil" }, - { name = "pyzmq" }, - { name = "tornado" }, - { name = "traitlets" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/71/22/bf9f12fdaeae18019a468b68952a60fe6dbab5d67cd2a103cac7659b41ca/jupyter_client-8.6.3.tar.gz", hash = "sha256:35b3a0947c4a6e9d589eb97d7d4cd5e90f910ee73101611f01283732bd6d9419", size = 342019 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/11/85/b0394e0b6fcccd2c1eeefc230978a6f8cb0c5df1e4cd3e7625735a0d7d1e/jupyter_client-8.6.3-py3-none-any.whl", hash = "sha256:e8a19cc986cc45905ac3362915f410f3af85424b4c0905e94fa5f2cb08e8f23f", size = 106105 }, -] - -[[package]] -name = "jupyter-console" -version = "6.6.3" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "ipykernel" }, - { name = "ipython" }, - { name = "jupyter-client" }, - { name = "jupyter-core" }, - { name = "prompt-toolkit" }, - { name = "pygments" }, - { name = "pyzmq" }, - { name = "traitlets" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/bd/2d/e2fd31e2fc41c14e2bcb6c976ab732597e907523f6b2420305f9fc7fdbdb/jupyter_console-6.6.3.tar.gz", hash = "sha256:566a4bf31c87adbfadf22cdf846e3069b59a71ed5da71d6ba4d8aaad14a53539", size = 34363 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/ca/77/71d78d58f15c22db16328a476426f7ac4a60d3a5a7ba3b9627ee2f7903d4/jupyter_console-6.6.3-py3-none-any.whl", hash = "sha256:309d33409fcc92ffdad25f0bcdf9a4a9daa61b6f341177570fdac03de5352485", size = 24510 }, -] - -[[package]] -name = "jupyter-core" -version = "5.8.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "platformdirs" }, - { name = "pywin32", marker = "platform_python_implementation != 'PyPy' and sys_platform == 'win32'" }, - { name = "traitlets" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/99/1b/72906d554acfeb588332eaaa6f61577705e9ec752ddb486f302dafa292d9/jupyter_core-5.8.1.tar.gz", hash = "sha256:0a5f9706f70e64786b75acba995988915ebd4601c8a52e534a40b51c95f59941", size = 88923 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/2f/57/6bffd4b20b88da3800c5d691e0337761576ee688eb01299eae865689d2df/jupyter_core-5.8.1-py3-none-any.whl", hash = "sha256:c28d268fc90fb53f1338ded2eb410704c5449a358406e8a948b75706e24863d0", size = 28880 }, -] - -[[package]] -name = "jupyter-events" -version = "0.12.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "jsonschema", extra = ["format-nongpl"] }, - { name = "packaging" }, - { name = "python-json-logger" }, - { name = "pyyaml" }, - { name = "referencing" }, - { name = "rfc3339-validator" }, - { name = "rfc3986-validator" }, - { name = "traitlets" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/9d/c3/306d090461e4cf3cd91eceaff84bede12a8e52cd821c2d20c9a4fd728385/jupyter_events-0.12.0.tar.gz", hash = "sha256:fc3fce98865f6784c9cd0a56a20644fc6098f21c8c33834a8d9fe383c17e554b", size = 62196 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/e2/48/577993f1f99c552f18a0428731a755e06171f9902fa118c379eb7c04ea22/jupyter_events-0.12.0-py3-none-any.whl", hash = "sha256:6464b2fa5ad10451c3d35fabc75eab39556ae1e2853ad0c0cc31b656731a97fb", size = 19430 }, -] - -[[package]] -name = "jupyter-lsp" -version = "2.2.5" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "jupyter-server" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/85/b4/3200b0b09c12bc3b72d943d923323c398eff382d1dcc7c0dbc8b74630e40/jupyter-lsp-2.2.5.tar.gz", hash = "sha256:793147a05ad446f809fd53ef1cd19a9f5256fd0a2d6b7ce943a982cb4f545001", size = 48741 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/07/e0/7bd7cff65594fd9936e2f9385701e44574fc7d721331ff676ce440b14100/jupyter_lsp-2.2.5-py3-none-any.whl", hash = "sha256:45fbddbd505f3fbfb0b6cb2f1bc5e15e83ab7c79cd6e89416b248cb3c00c11da", size = 69146 }, -] - -[[package]] -name = "jupyter-server" -version = "2.16.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "anyio" }, - { name = "argon2-cffi" }, - { name = "jinja2" }, - { name = "jupyter-client" }, - { name = "jupyter-core" }, - { name = "jupyter-events" }, - { name = "jupyter-server-terminals" }, - { name = "nbconvert" }, - { name = "nbformat" }, - { name = "overrides" }, - { name = "packaging" }, - { name = "prometheus-client" }, - { name = "pywinpty", marker = "os_name == 'nt'" }, - { name = "pyzmq" }, - { name = "send2trash" }, - { name = "terminado" }, - { name = "tornado" }, - { name = "traitlets" }, - { name = "websocket-client" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/41/c8/ba2bbcd758c47f1124c4ca14061e8ce60d9c6fd537faee9534a95f83521a/jupyter_server-2.16.0.tar.gz", hash = "sha256:65d4b44fdf2dcbbdfe0aa1ace4a842d4aaf746a2b7b168134d5aaed35621b7f6", size = 728177 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/46/1f/5ebbced977171d09a7b0c08a285ff9a20aafb9c51bde07e52349ff1ddd71/jupyter_server-2.16.0-py3-none-any.whl", hash = "sha256:3d8db5be3bc64403b1c65b400a1d7f4647a5ce743f3b20dbdefe8ddb7b55af9e", size = 386904 }, -] - -[[package]] -name = "jupyter-server-terminals" -version = "0.5.3" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "pywinpty", marker = "os_name == 'nt'" }, - { name = "terminado" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/fc/d5/562469734f476159e99a55426d697cbf8e7eb5efe89fb0e0b4f83a3d3459/jupyter_server_terminals-0.5.3.tar.gz", hash = "sha256:5ae0295167220e9ace0edcfdb212afd2b01ee8d179fe6f23c899590e9b8a5269", size = 31430 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/07/2d/2b32cdbe8d2a602f697a649798554e4f072115438e92249624e532e8aca6/jupyter_server_terminals-0.5.3-py3-none-any.whl", hash = "sha256:41ee0d7dc0ebf2809c668e0fc726dfaf258fcd3e769568996ca731b6194ae9aa", size = 13656 }, -] - -[[package]] -name = "jupyterlab" -version = "4.4.4" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "async-lru" }, - { name = "httpx" }, - { name = "ipykernel" }, - { name = "jinja2" }, - { name = "jupyter-core" }, - { name = "jupyter-lsp" }, - { name = "jupyter-server" }, - { name = "jupyterlab-server" }, - { name = "notebook-shim" }, - { name = "packaging" }, - { name = "setuptools" }, - { name = "tomli", marker = "python_full_version == '3.10.12'" }, - { name = "tornado" }, - { name = "traitlets" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/e2/4d/7ca5b46ea56742880d71a768a9e6fb8f8482228427eb89492d55c5d0bb7d/jupyterlab-4.4.4.tar.gz", hash = "sha256:163fee1ef702e0a057f75d2eed3ed1da8a986d59eb002cbeb6f0c2779e6cd153", size = 23044296 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/f8/82/66910ce0995dbfdb33609f41c99fe32ce483b9624a3e7d672af14ff63b9f/jupyterlab-4.4.4-py3-none-any.whl", hash = "sha256:711611e4f59851152eb93316c3547c3ec6291f16bb455f1f4fa380d25637e0dd", size = 12296310 }, -] - -[[package]] -name = "jupyterlab-pygments" -version = "0.3.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/90/51/9187be60d989df97f5f0aba133fa54e7300f17616e065d1ada7d7646b6d6/jupyterlab_pygments-0.3.0.tar.gz", hash = "sha256:721aca4d9029252b11cfa9d185e5b5af4d54772bb8072f9b7036f4170054d35d", size = 512900 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/b1/dd/ead9d8ea85bf202d90cc513b533f9c363121c7792674f78e0d8a854b63b4/jupyterlab_pygments-0.3.0-py3-none-any.whl", hash = "sha256:841a89020971da1d8693f1a99997aefc5dc424bb1b251fd6322462a1b8842780", size = 15884 }, -] - -[[package]] -name = "jupyterlab-server" -version = "2.27.3" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "babel" }, - { name = "jinja2" }, - { name = "json5" }, - { name = "jsonschema" }, - { name = "jupyter-server" }, - { name = "packaging" }, - { name = "requests" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/0a/c9/a883ce65eb27905ce77ace410d83587c82ea64dc85a48d1f7ed52bcfa68d/jupyterlab_server-2.27.3.tar.gz", hash = "sha256:eb36caca59e74471988f0ae25c77945610b887f777255aa21f8065def9e51ed4", size = 76173 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/54/09/2032e7d15c544a0e3cd831c51d77a8ca57f7555b2e1b2922142eddb02a84/jupyterlab_server-2.27.3-py3-none-any.whl", hash = "sha256:e697488f66c3db49df675158a77b3b017520d772c6e1548c7d9bcc5df7944ee4", size = 59700 }, -] - -[[package]] -name = "jupyterlab-widgets" -version = "3.0.15" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/b9/7d/160595ca88ee87ac6ba95d82177d29ec60aaa63821d3077babb22ce031a5/jupyterlab_widgets-3.0.15.tar.gz", hash = "sha256:2920888a0c2922351a9202817957a68c07d99673504d6cd37345299e971bb08b", size = 213149 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/43/6a/ca128561b22b60bd5a0c4ea26649e68c8556b82bc70a0c396eebc977fe86/jupyterlab_widgets-3.0.15-py3-none-any.whl", hash = "sha256:d59023d7d7ef71400d51e6fee9a88867f6e65e10a4201605d2d7f3e8f012a31c", size = 216571 }, -] - -[[package]] -name = "kaldi-python-io" -version = "1.2.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "numpy" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/80/45/e3e542ffa8970ebd782fcece35e2295de9c60e8c396c2c1a403410d1b24e/kaldi-python-io-1.2.2.tar.gz", hash = "sha256:4ebb4029c6c58296cc0abf96edff02832ba341d290ed37624a8d00105f0f7c00", size = 8814 } - -[[package]] -name = "kaldiio" -version = "2.18.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "numpy" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/8d/85/92435e8e62eb3d43eded9f24643fc2a6dbce031cebceed11528147c7873f/kaldiio-2.18.1.tar.gz", hash = "sha256:0283d197fac6ac683f7a9e6af8d18aad9dbd2c4a997f22e45294f2ac1ee3c432", size = 35570 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/ae/e3/6c3b42233225f398f7a72988b524f654ae818cca0d441db847a2761203e9/kaldiio-2.18.1-py3-none-any.whl", hash = "sha256:397a4cd18977acaae7acabfba6807ee0a6978c620064381a266eac15b3c1a0a0", size = 29330 }, -] - -[[package]] -name = "kiwisolver" -version = "1.4.8" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/82/59/7c91426a8ac292e1cdd53a63b6d9439abd573c875c3f92c146767dd33faf/kiwisolver-1.4.8.tar.gz", hash = "sha256:23d5f023bdc8c7e54eb65f03ca5d5bb25b601eac4d7f1a042888a1f45237987e", size = 97538 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/47/5f/4d8e9e852d98ecd26cdf8eaf7ed8bc33174033bba5e07001b289f07308fd/kiwisolver-1.4.8-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:88c6f252f6816a73b1f8c904f7bbe02fd67c09a69f7cb8a0eecdbf5ce78e63db", size = 124623 }, - { url = "https://files.pythonhosted.org/packages/1d/70/7f5af2a18a76fe92ea14675f8bd88ce53ee79e37900fa5f1a1d8e0b42998/kiwisolver-1.4.8-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:c72941acb7b67138f35b879bbe85be0f6c6a70cab78fe3ef6db9c024d9223e5b", size = 66720 }, - { url = "https://files.pythonhosted.org/packages/c6/13/e15f804a142353aefd089fadc8f1d985561a15358c97aca27b0979cb0785/kiwisolver-1.4.8-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:ce2cf1e5688edcb727fdf7cd1bbd0b6416758996826a8be1d958f91880d0809d", size = 65413 }, - { url = "https://files.pythonhosted.org/packages/ce/6d/67d36c4d2054e83fb875c6b59d0809d5c530de8148846b1370475eeeece9/kiwisolver-1.4.8-cp310-cp310-manylinux_2_12_i686.manylinux2010_i686.whl", hash = "sha256:c8bf637892dc6e6aad2bc6d4d69d08764166e5e3f69d469e55427b6ac001b19d", size = 1650826 }, - { url = "https://files.pythonhosted.org/packages/de/c6/7b9bb8044e150d4d1558423a1568e4f227193662a02231064e3824f37e0a/kiwisolver-1.4.8-cp310-cp310-manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:034d2c891f76bd3edbdb3ea11140d8510dca675443da7304205a2eaa45d8334c", size = 1628231 }, - { url = "https://files.pythonhosted.org/packages/b6/38/ad10d437563063eaaedbe2c3540a71101fc7fb07a7e71f855e93ea4de605/kiwisolver-1.4.8-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d47b28d1dfe0793d5e96bce90835e17edf9a499b53969b03c6c47ea5985844c3", size = 1408938 }, - { url = "https://files.pythonhosted.org/packages/52/ce/c0106b3bd7f9e665c5f5bc1e07cc95b5dabd4e08e3dad42dbe2faad467e7/kiwisolver-1.4.8-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:eb158fe28ca0c29f2260cca8c43005329ad58452c36f0edf298204de32a9a3ed", size = 1422799 }, - { url = "https://files.pythonhosted.org/packages/d0/87/efb704b1d75dc9758087ba374c0f23d3254505edaedd09cf9d247f7878b9/kiwisolver-1.4.8-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:d5536185fce131780ebd809f8e623bf4030ce1b161353166c49a3c74c287897f", size = 1354362 }, - { url = "https://files.pythonhosted.org/packages/eb/b3/fd760dc214ec9a8f208b99e42e8f0130ff4b384eca8b29dd0efc62052176/kiwisolver-1.4.8-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:369b75d40abedc1da2c1f4de13f3482cb99e3237b38726710f4a793432b1c5ff", size = 2222695 }, - { url = "https://files.pythonhosted.org/packages/a2/09/a27fb36cca3fc01700687cc45dae7a6a5f8eeb5f657b9f710f788748e10d/kiwisolver-1.4.8-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:641f2ddf9358c80faa22e22eb4c9f54bd3f0e442e038728f500e3b978d00aa7d", size = 2370802 }, - { url = "https://files.pythonhosted.org/packages/3d/c3/ba0a0346db35fe4dc1f2f2cf8b99362fbb922d7562e5f911f7ce7a7b60fa/kiwisolver-1.4.8-cp310-cp310-musllinux_1_2_ppc64le.whl", hash = "sha256:d561d2d8883e0819445cfe58d7ddd673e4015c3c57261d7bdcd3710d0d14005c", size = 2334646 }, - { url = "https://files.pythonhosted.org/packages/41/52/942cf69e562f5ed253ac67d5c92a693745f0bed3c81f49fc0cbebe4d6b00/kiwisolver-1.4.8-cp310-cp310-musllinux_1_2_s390x.whl", hash = "sha256:1732e065704b47c9afca7ffa272f845300a4eb959276bf6970dc07265e73b605", size = 2467260 }, - { url = "https://files.pythonhosted.org/packages/32/26/2d9668f30d8a494b0411d4d7d4ea1345ba12deb6a75274d58dd6ea01e951/kiwisolver-1.4.8-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:bcb1ebc3547619c3b58a39e2448af089ea2ef44b37988caf432447374941574e", size = 2288633 }, - { url = "https://files.pythonhosted.org/packages/98/99/0dd05071654aa44fe5d5e350729961e7bb535372935a45ac89a8924316e6/kiwisolver-1.4.8-cp310-cp310-win_amd64.whl", hash = "sha256:89c107041f7b27844179ea9c85d6da275aa55ecf28413e87624d033cf1f6b751", size = 71885 }, - { url = "https://files.pythonhosted.org/packages/6c/fc/822e532262a97442989335394d441cd1d0448c2e46d26d3e04efca84df22/kiwisolver-1.4.8-cp310-cp310-win_arm64.whl", hash = "sha256:b5773efa2be9eb9fcf5415ea3ab70fc785d598729fd6057bea38d539ead28271", size = 65175 }, - { url = "https://files.pythonhosted.org/packages/1f/f9/ae81c47a43e33b93b0a9819cac6723257f5da2a5a60daf46aa5c7226ea85/kiwisolver-1.4.8-pp310-pypy310_pp73-macosx_10_15_x86_64.whl", hash = "sha256:e7a019419b7b510f0f7c9dceff8c5eae2392037eae483a7f9162625233802b0a", size = 60403 }, - { url = "https://files.pythonhosted.org/packages/58/ca/f92b5cb6f4ce0c1ebfcfe3e2e42b96917e16f7090e45b21102941924f18f/kiwisolver-1.4.8-pp310-pypy310_pp73-macosx_11_0_arm64.whl", hash = "sha256:286b18e86682fd2217a48fc6be6b0f20c1d0ed10958d8dc53453ad58d7be0bf8", size = 58657 }, - { url = "https://files.pythonhosted.org/packages/80/28/ae0240f732f0484d3a4dc885d055653c47144bdf59b670aae0ec3c65a7c8/kiwisolver-1.4.8-pp310-pypy310_pp73-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:4191ee8dfd0be1c3666ccbac178c5a05d5f8d689bbe3fc92f3c4abec817f8fe0", size = 84948 }, - { url = "https://files.pythonhosted.org/packages/5d/eb/78d50346c51db22c7203c1611f9b513075f35c4e0e4877c5dde378d66043/kiwisolver-1.4.8-pp310-pypy310_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7cd2785b9391f2873ad46088ed7599a6a71e762e1ea33e87514b1a441ed1da1c", size = 81186 }, - { url = "https://files.pythonhosted.org/packages/43/f8/7259f18c77adca88d5f64f9a522792e178b2691f3748817a8750c2d216ef/kiwisolver-1.4.8-pp310-pypy310_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c07b29089b7ba090b6f1a669f1411f27221c3662b3a1b7010e67b59bb5a6f10b", size = 80279 }, - { url = "https://files.pythonhosted.org/packages/3a/1d/50ad811d1c5dae091e4cf046beba925bcae0a610e79ae4c538f996f63ed5/kiwisolver-1.4.8-pp310-pypy310_pp73-win_amd64.whl", hash = "sha256:65ea09a5a3faadd59c2ce96dc7bf0f364986a315949dc6374f04396b0d60e09b", size = 71762 }, -] - -[[package]] -name = "lazy-loader" -version = "0.4" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "packaging" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/6f/6b/c875b30a1ba490860c93da4cabf479e03f584eba06fe5963f6f6644653d8/lazy_loader-0.4.tar.gz", hash = "sha256:47c75182589b91a4e1a85a136c074285a5ad4d9f39c63e0d7fb76391c4574cd1", size = 15431 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/83/60/d497a310bde3f01cb805196ac61b7ad6dc5dcf8dce66634dc34364b20b4f/lazy_loader-0.4-py3-none-any.whl", hash = "sha256:342aa8e14d543a154047afb4ba8ef17f5563baad3fc610d7b15b213b0f119efc", size = 12097 }, -] - -[[package]] -name = "levenshtein" -version = "0.27.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "rapidfuzz" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/7e/b3/b5f8011483ba9083a0bc74c4d58705e9cf465fbe55c948a1b1357d0a2aa8/levenshtein-0.27.1.tar.gz", hash = "sha256:3e18b73564cfc846eec94dd13fab6cb006b5d2e0cc56bad1fd7d5585881302e3", size = 382571 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/b3/b1/9906a75b98dd9c008015a72d7658be53851e361a35492631edf1b1f334ab/levenshtein-0.27.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:13d6f617cb6fe63714c4794861cfaacd398db58a292f930edb7f12aad931dace", size = 174542 }, - { url = "https://files.pythonhosted.org/packages/3b/57/e26e0164a93fb045316856603111d95538cac8224a3709e4ac96a6bb74f3/levenshtein-0.27.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:ca9d54d41075e130c390e61360bec80f116b62d6ae973aec502e77e921e95334", size = 156367 }, - { url = "https://files.pythonhosted.org/packages/6d/dd/92fcb71d48c1fe69c46c211156adafb8175037dc63e80e970106aef3f9d5/levenshtein-0.27.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:2de1f822b5c9a20d10411f779dfd7181ce3407261436f8470008a98276a9d07f", size = 152189 }, - { url = "https://files.pythonhosted.org/packages/5e/23/3f331f5fbfa93634126439cfc8c01b31f7ef1fbedb81663581e27a69da4d/levenshtein-0.27.1-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:81270392c2e45d1a7e1b3047c3a272d5e28bb4f1eff0137637980064948929b7", size = 184271 }, - { url = "https://files.pythonhosted.org/packages/5a/76/d6ac541a1a80bdc5c98584a6a2d2301e677af4cb2e4092247207791b56a6/levenshtein-0.27.1-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:2d30c3ea23a94dddd56dbe323e1fa8a29ceb24da18e2daa8d0abf78b269a5ad1", size = 185078 }, - { url = "https://files.pythonhosted.org/packages/2d/ed/d0c5abe8cfcf6a7f2a4197e889e12b7a0c2145a0ef3354b1c000bf367305/levenshtein-0.27.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0f3e0bea76695b9045bbf9ad5f67ad4cc01c11f783368f34760e068f19b6a6bc", size = 161505 }, - { url = "https://files.pythonhosted.org/packages/f3/28/a5b78e1818211bc6407590876bbdcc6d79671e529a0c186780492c1f2136/levenshtein-0.27.1-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:cdd190e468a68c31a5943368a5eaf4e130256a8707886d23ab5906a0cb98a43c", size = 246968 }, - { url = "https://files.pythonhosted.org/packages/77/7f/981b903583956cb67b33bed39d9840ab5e4c7062bceec564b7bf2c3f6f49/levenshtein-0.27.1-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:7c3121314bb4b676c011c33f6a0ebb462cfdcf378ff383e6f9e4cca5618d0ba7", size = 1116000 }, - { url = "https://files.pythonhosted.org/packages/75/1d/c4be47d5f436fd310373c5ebdf05828c1d95be9a44c3e94f29c40937b30c/levenshtein-0.27.1-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:f8ef378c873efcc5e978026b69b45342d841cd7a2f273447324f1c687cc4dc37", size = 1401162 }, - { url = "https://files.pythonhosted.org/packages/91/e4/0b107676efe3ecd5fada1ed3a3bbddd4c829e2ef34e980b76374c116235b/levenshtein-0.27.1-cp310-cp310-musllinux_1_2_ppc64le.whl", hash = "sha256:ff18d78c5c16bea20876425e1bf5af56c25918fb01bc0f2532db1317d4c0e157", size = 1225141 }, - { url = "https://files.pythonhosted.org/packages/29/f0/f3f88d766fdbb1d39fe98dc5527223bae099444e501550ae088c47ddd97b/levenshtein-0.27.1-cp310-cp310-musllinux_1_2_s390x.whl", hash = "sha256:13412ff805afbfe619d070280d1a76eb4198c60c5445cd5478bd4c7055bb3d51", size = 1419707 }, - { url = "https://files.pythonhosted.org/packages/b8/1c/f51ac1db4064a85effa50df240250e413f428164301d836c312baf09381e/levenshtein-0.27.1-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:a2adb9f263557f7fb13e19eb2f34595d86929a44c250b2fca6e9b65971e51e20", size = 1189284 }, - { url = "https://files.pythonhosted.org/packages/e0/67/5ace76bc964b93ed6203a9f8c4dcde1a50e336468f7da3a21dd29febaf46/levenshtein-0.27.1-cp310-cp310-win32.whl", hash = "sha256:6278a33d2e0e909d8829b5a72191419c86dd3bb45b82399c7efc53dabe870c35", size = 88036 }, - { url = "https://files.pythonhosted.org/packages/06/e0/d9737dbbe85842ddb300cb7974fc065edc56ec647652863f95ac1977d378/levenshtein-0.27.1-cp310-cp310-win_amd64.whl", hash = "sha256:5b602b8428ee5dc88432a55c5303a739ee2be7c15175bd67c29476a9d942f48e", size = 99922 }, - { url = "https://files.pythonhosted.org/packages/27/b8/13e22789ab700db0da98f973a508643dbe2d25bd0fb5dc53239e0e2852c1/levenshtein-0.27.1-cp310-cp310-win_arm64.whl", hash = "sha256:48334081fddaa0c259ba01ee898640a2cf8ede62e5f7e25fefece1c64d34837f", size = 87846 }, - { url = "https://files.pythonhosted.org/packages/25/ed/37e2d1f5e690d7376cd7e8bdd19411479ff352a3df9ab5f845dd680ef779/levenshtein-0.27.1-pp310-pypy310_pp73-macosx_10_15_x86_64.whl", hash = "sha256:c92a222ab95b8d903eae6d5e7d51fe6c999be021b647715c18d04d0b0880f463", size = 170482 }, - { url = "https://files.pythonhosted.org/packages/6d/9f/30b1144b9d1da74743e7d7cdf47575b7013c9767e608c7454dbd318aacd2/levenshtein-0.27.1-pp310-pypy310_pp73-macosx_11_0_arm64.whl", hash = "sha256:71afc36b4ee950fa1140aff22ffda9e5e23280285858e1303260dbb2eabf342d", size = 153106 }, - { url = "https://files.pythonhosted.org/packages/b1/c5/18d0bec94a166cebaefa3db4beab9a7e0d75412b52e9626f5dce1ca8d149/levenshtein-0.27.1-pp310-pypy310_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:58b1daeebfc148a571f09cfe18c16911ea1eaaa9e51065c5f7e7acbc4b866afa", size = 150984 }, - { url = "https://files.pythonhosted.org/packages/55/b4/4b80eb0c96caabdb683256cac9cc2cc9a73dee8ea80ab7cc3ee8aebd603f/levenshtein-0.27.1-pp310-pypy310_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:105edcb14797d95c77f69bad23104314715a64cafbf4b0e79d354a33d7b54d8d", size = 158673 }, - { url = "https://files.pythonhosted.org/packages/81/14/a43daefbc6d5e5561176150363cbac73003795b85ae136ffd4d0691af3fb/levenshtein-0.27.1-pp310-pypy310_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:d9c58fb1ef8bdc8773d705fbacf628e12c3bb63ee4d065dda18a76e86042444a", size = 244419 }, - { url = "https://files.pythonhosted.org/packages/d0/55/34f133f4f0998d7335bd96b9d315dc888b118e48e999c3d2c621b84965b9/levenshtein-0.27.1-pp310-pypy310_pp73-win_amd64.whl", hash = "sha256:e52270591854af67217103955a36bd7436b57c801e3354e73ba44d689ed93697", size = 97932 }, -] - -[[package]] -name = "lhotse" -version = "1.30.3" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "audioread" }, - { name = "click" }, - { name = "cytoolz" }, - { name = "intervaltree" }, - { name = "lilcom" }, - { name = "numpy" }, - { name = "packaging" }, - { name = "pyyaml" }, - { name = "soundfile" }, - { name = "tabulate" }, - { name = "torch" }, - { name = "tqdm" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/a4/28/4ad81e95f2abaf101532cc7f599223abbfa0faa19caa1232496e7cb60f19/lhotse-1.30.3.tar.gz", hash = "sha256:f39eaeab795b85d35cf7e6ed315246252c6565ee4d0f767781ba23a8992f2df9", size = 643240 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/f7/35/a0ddacd91f13267653262eba0933d2750fe63cff2fddcb8d73bb584090cb/lhotse-1.30.3-py3-none-any.whl", hash = "sha256:80df401000ac4b016721bcf27a575d72a4f31ace914bb95db2e118e61556a3ac", size = 851420 }, -] - -[[package]] -name = "libcst" -version = "1.8.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "pyyaml", marker = "python_full_version == '3.10.12'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/89/aa/b52d195b167958fe1bd106a260f64cc80ec384f6ac2a9cda874d8803df06/libcst-1.8.2.tar.gz", hash = "sha256:66e82cedba95a6176194a817be4232c720312f8be6d2c8f3847f3317d95a0c7f", size = 881534 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/3c/2e/1d7f67d2ef6f875e9e8798c024f7cb3af3fe861e417bff485c69b655ac96/libcst-1.8.2-cp310-cp310-macosx_10_12_x86_64.whl", hash = "sha256:67d9720d91f507c87b3e5f070627ad640a00bc6cfdf5635f8c6ee9f2964cf71c", size = 2195106 }, - { url = "https://files.pythonhosted.org/packages/82/d0/3d94fee2685f263fd8d85a83e2537fcc78b644eae450738bf2c72604f0df/libcst-1.8.2-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:94b7c032b72566077614a02baab1929739fd0af0cc1d46deaba4408b870faef2", size = 2080577 }, - { url = "https://files.pythonhosted.org/packages/14/87/c9b49bebb9a930fdcb59bf841f1c45719d2a4a39c3eb7efacfd30a2bfb0a/libcst-1.8.2-cp310-cp310-manylinux1_i686.manylinux2014_i686.manylinux_2_17_i686.manylinux_2_5_i686.whl", hash = "sha256:11ea148902e3e1688afa392087c728ac3a843e54a87d334d1464d2097d3debb7", size = 2404076 }, - { url = "https://files.pythonhosted.org/packages/49/fa/9ca145aa9033f9a8362a5663ceb28dfb67082574de8118424b6b8e445e7a/libcst-1.8.2-cp310-cp310-manylinux_2_28_aarch64.whl", hash = "sha256:22c9473a2cc53faabcc95a0ac6ca4e52d127017bf34ba9bc0f8e472e44f7b38e", size = 2219813 }, - { url = "https://files.pythonhosted.org/packages/0c/25/496a025c09e96116437a57fd34abefe84c041d930f832c6e42d84d9e028c/libcst-1.8.2-cp310-cp310-manylinux_2_28_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:b5269b96367e65793a7714608f6d906418eb056d59eaac9bba980486aabddbed", size = 2189782 }, - { url = "https://files.pythonhosted.org/packages/b3/75/826b5772192826d70480efe93bab3e4f0b4a24d31031f45547257ad5f9a8/libcst-1.8.2-cp310-cp310-manylinux_2_28_x86_64.whl", hash = "sha256:d20e932ddd9a389da57b060c26e84a24118c96ff6fc5dcc7b784da24e823b694", size = 2312403 }, - { url = "https://files.pythonhosted.org/packages/93/f4/316fa14ea6c61ea8755672d60e012558f0216300b3819e72bebc7864a507/libcst-1.8.2-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:a553d452004e44b841788f6faa7231a02157527ddecc89dbbe5b689b74822226", size = 2280566 }, - { url = "https://files.pythonhosted.org/packages/fc/52/74b69350db379b1646739288b88ffab2981b2ad48407faf03df3768d7d2f/libcst-1.8.2-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:7fe762c4c390039b79b818cbc725d8663586b25351dc18a2704b0e357d69b924", size = 2388508 }, - { url = "https://files.pythonhosted.org/packages/bc/c6/fa92699b537ed65e93c2869144e23bdf156ec81ae7b84b4f34cbc20d6048/libcst-1.8.2-cp310-cp310-win_amd64.whl", hash = "sha256:5c513e64eff0f7bf2a908e2d987a98653eb33e1062ce2afd3a84af58159a24f9", size = 2093260 }, - { url = "https://files.pythonhosted.org/packages/b0/ac/4ec4ae9da311f72cd97e930c325bb605e9ad0baaafcafadb0588e1dc5c4e/libcst-1.8.2-cp310-cp310-win_arm64.whl", hash = "sha256:41613fe08e647213546c7c59a5a1fc5484666e7d4cab6e80260c612acbb20e8c", size = 1985236 }, -] - -[[package]] -name = "librosa" -version = "0.11.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "audioread" }, - { name = "decorator" }, - { name = "joblib" }, - { name = "lazy-loader" }, - { name = "msgpack" }, - { name = "numba" }, - { name = "numpy" }, - { name = "pooch" }, - { name = "scikit-learn" }, - { name = "scipy" }, - { name = "soundfile" }, - { name = "soxr" }, - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/64/36/360b5aafa0238e29758729e9486c6ed92a6f37fa403b7875e06c115cdf4a/librosa-0.11.0.tar.gz", hash = "sha256:f5ed951ca189b375bbe2e33b2abd7e040ceeee302b9bbaeeffdfddb8d0ace908", size = 327001 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/b5/ba/c63c5786dfee4c3417094c4b00966e61e4a63efecee22cb7b4c0387dda83/librosa-0.11.0-py3-none-any.whl", hash = "sha256:0b6415c4fd68bff4c29288abe67c6d80b587e0e1e2cfb0aad23e4559504a7fa1", size = 260749 }, -] - -[[package]] -name = "lightning" -version = "2.4.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "fsspec", extra = ["http"] }, - { name = "lightning-utilities" }, - { name = "packaging" }, - { name = "pytorch-lightning" }, - { name = "pyyaml" }, - { name = "torch" }, - { name = "torchmetrics" }, - { name = "tqdm" }, - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/56/d0/78ea244ac044cd4df15aa8294a50ff3561fb177e7e5ba788aaa542046cae/lightning-2.4.0.tar.gz", hash = "sha256:9156604cc56e4b2b603f34fa7f0fe5107375c8e6d85e74544b319a15faa9ed0e", size = 620632 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/a3/2c/85eaf42c983b0cd81bcda5876da2c8e2a9fd347908666ea9855724369171/lightning-2.4.0-py3-none-any.whl", hash = "sha256:560163af9711cf59055c448232c473150a299089efce0d2be3cc3288082d8768", size = 810971 }, -] - -[[package]] -name = "lightning-utilities" -version = "0.14.3" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "packaging" }, - { name = "setuptools" }, - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/0f/bb/63a6a8c9e7a96b6ba92647fa5b1595c2dbee29f8178705adb4704d82ecba/lightning_utilities-0.14.3.tar.gz", hash = "sha256:37e2f83f273890052955a44054382c211a303012ee577619efbaa5df9e65e9f5", size = 30346 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/1a/c1/31b3184cba7b257a4a3b5ca5b88b9204ccb7aa02fe3c992280899293ed54/lightning_utilities-0.14.3-py3-none-any.whl", hash = "sha256:4ab9066aa36cd7b93a05713808901909e96cc3f187ea6fd3052b2fd91313b468", size = 28894 }, -] - -[[package]] -name = "lilcom" -version = "1.8.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "numpy" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/8b/b5/97422825e61a2683dd39d78165fc470c1153b4a908dd5cd0711c0e9d262c/lilcom-1.8.1.tar.gz", hash = "sha256:69c62037c92e71e601ac3bb3ae19811f22ceffbdf58b0fdbf81cc6a0ec6fc3c5", size = 45813 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/e1/5c/51ece6a8f6ed39c27eb18bb1c09fe6508613bc67bcc5872caa788cdc9eef/lilcom-1.8.1-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:7e896dd6172011d320c26a4ca648edddf1e6b47bb1de749cf27ac096d73f63c9", size = 122436 }, - { url = "https://files.pythonhosted.org/packages/58/fd/8008d8b82965eeb2e35a46edb118e195d10bc424e3d80f2e3dafe3044e1f/lilcom-1.8.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:6b7c5488267c9c3c6ab49c3e3ec287031029d7de57f4c2307264dd75f5e77211", size = 86593 }, - { url = "https://files.pythonhosted.org/packages/7b/02/30e76ca2bfc3fe19adbcca254dde03ba48a6e623eb55df2752f4c06cfeb0/lilcom-1.8.1-cp310-cp310-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:7460e2dd481b38fc8fa1f0cc48c5aaa1b4206d0c27130c001bf23f1d4c99143b", size = 98968 }, - { url = "https://files.pythonhosted.org/packages/d7/cd/e3e45ef07c2e19f5718d7c9d33b40438b83a3397bdd4c70bd3d1bbf059ad/lilcom-1.8.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9bdeea4aa3d7cad9e38cac99ef012a61f843a87a661b9475f565787f73828ec5", size = 92337 }, - { url = "https://files.pythonhosted.org/packages/84/da/3b8a45be54f4d12a2d118c84d9d2aab81e5b75ae4a7b8469213a28b121dd/lilcom-1.8.1-cp310-cp310-win_amd64.whl", hash = "sha256:c254a74156ff8da2e43c8b891ed47eefcf0c1fee82a0c5984388052d97573a1b", size = 68524 }, -] - -[[package]] -name = "llvmlite" -version = "0.44.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/89/6a/95a3d3610d5c75293d5dbbb2a76480d5d4eeba641557b69fe90af6c5b84e/llvmlite-0.44.0.tar.gz", hash = "sha256:07667d66a5d150abed9157ab6c0b9393c9356f229784a4385c02f99e94fc94d4", size = 171880 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/41/75/d4863ddfd8ab5f6e70f4504cf8cc37f4e986ec6910f4ef8502bb7d3c1c71/llvmlite-0.44.0-cp310-cp310-macosx_10_14_x86_64.whl", hash = "sha256:9fbadbfba8422123bab5535b293da1cf72f9f478a65645ecd73e781f962ca614", size = 28132306 }, - { url = "https://files.pythonhosted.org/packages/37/d9/6e8943e1515d2f1003e8278819ec03e4e653e2eeb71e4d00de6cfe59424e/llvmlite-0.44.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:cccf8eb28f24840f2689fb1a45f9c0f7e582dd24e088dcf96e424834af11f791", size = 26201096 }, - { url = "https://files.pythonhosted.org/packages/aa/46/8ffbc114def88cc698906bf5acab54ca9fdf9214fe04aed0e71731fb3688/llvmlite-0.44.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:7202b678cdf904823c764ee0fe2dfe38a76981f4c1e51715b4cb5abb6cf1d9e8", size = 42361859 }, - { url = "https://files.pythonhosted.org/packages/30/1c/9366b29ab050a726af13ebaae8d0dff00c3c58562261c79c635ad4f5eb71/llvmlite-0.44.0-cp310-cp310-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:40526fb5e313d7b96bda4cbb2c85cd5374e04d80732dd36a282d72a560bb6408", size = 41184199 }, - { url = "https://files.pythonhosted.org/packages/69/07/35e7c594b021ecb1938540f5bce543ddd8713cff97f71d81f021221edc1b/llvmlite-0.44.0-cp310-cp310-win_amd64.whl", hash = "sha256:41e3839150db4330e1b2716c0be3b5c4672525b4c9005e17c7597f835f351ce2", size = 30332381 }, -] - -[[package]] -name = "loguru" -version = "0.7.3" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "colorama", marker = "sys_platform == 'win32'" }, - { name = "win32-setctime", marker = "sys_platform == 'win32'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/3a/05/a1dae3dffd1116099471c643b8924f5aa6524411dc6c63fdae648c4f1aca/loguru-0.7.3.tar.gz", hash = "sha256:19480589e77d47b8d85b2c827ad95d49bf31b0dcde16593892eb51dd18706eb6", size = 63559 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/0c/29/0348de65b8cc732daa3e33e67806420b2ae89bdce2b04af740289c5c6c8c/loguru-0.7.3-py3-none-any.whl", hash = "sha256:31a33c10c8e1e10422bfd431aeb5d351c7cf7fa671e3c4df004162264b28220c", size = 61595 }, -] - -[[package]] -name = "mako" -version = "1.3.10" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "markupsafe" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/9e/38/bd5b78a920a64d708fe6bc8e0a2c075e1389d53bef8413725c63ba041535/mako-1.3.10.tar.gz", hash = "sha256:99579a6f39583fa7e5630a28c3c1f440e4e97a414b80372649c0ce338da2ea28", size = 392474 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/87/fb/99f81ac72ae23375f22b7afdb7642aba97c00a713c217124420147681a2f/mako-1.3.10-py3-none-any.whl", hash = "sha256:baef24a52fc4fc514a0887ac600f9f1cff3d82c61d4d700a1fa84d597b88db59", size = 78509 }, -] - -[[package]] -name = "markdown" -version = "3.8.2" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/d7/c2/4ab49206c17f75cb08d6311171f2d65798988db4360c4d1485bd0eedd67c/markdown-3.8.2.tar.gz", hash = "sha256:247b9a70dd12e27f67431ce62523e675b866d254f900c4fe75ce3dda62237c45", size = 362071 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/96/2b/34cc11786bc00d0f04d0f5fdc3a2b1ae0b6239eef72d3d345805f9ad92a1/markdown-3.8.2-py3-none-any.whl", hash = "sha256:5c83764dbd4e00bdd94d85a19b8d55ccca20fe35b2e678a1422b380324dd5f24", size = 106827 }, -] - -[[package]] -name = "markdown-it-py" -version = "3.0.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "mdurl" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/38/71/3b932df36c1a044d397a1f92d1cf91ee0a503d91e470cbd670aa66b07ed0/markdown-it-py-3.0.0.tar.gz", hash = "sha256:e3f60a94fa066dc52ec76661e37c851cb232d92f9886b15cb560aaada2df8feb", size = 74596 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/42/d7/1ec15b46af6af88f19b8e5ffea08fa375d433c998b8a7639e76935c14f1f/markdown_it_py-3.0.0-py3-none-any.whl", hash = "sha256:355216845c60bd96232cd8d8c40e8f9765cc86f46880e43a8fd22dc1a1a8cab1", size = 87528 }, -] - -[[package]] -name = "markupsafe" -version = "3.0.2" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/b2/97/5d42485e71dfc078108a86d6de8fa46db44a1a9295e89c5d6d4a06e23a62/markupsafe-3.0.2.tar.gz", hash = "sha256:ee55d3edf80167e48ea11a923c7386f4669df67d7994554387f84e7d8b0a2bf0", size = 20537 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/04/90/d08277ce111dd22f77149fd1a5d4653eeb3b3eaacbdfcbae5afb2600eebd/MarkupSafe-3.0.2-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:7e94c425039cde14257288fd61dcfb01963e658efbc0ff54f5306b06054700f8", size = 14357 }, - { url = "https://files.pythonhosted.org/packages/04/e1/6e2194baeae0bca1fae6629dc0cbbb968d4d941469cbab11a3872edff374/MarkupSafe-3.0.2-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:9e2d922824181480953426608b81967de705c3cef4d1af983af849d7bd619158", size = 12393 }, - { url = "https://files.pythonhosted.org/packages/1d/69/35fa85a8ece0a437493dc61ce0bb6d459dcba482c34197e3efc829aa357f/MarkupSafe-3.0.2-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:38a9ef736c01fccdd6600705b09dc574584b89bea478200c5fbf112a6b0d5579", size = 21732 }, - { url = "https://files.pythonhosted.org/packages/22/35/137da042dfb4720b638d2937c38a9c2df83fe32d20e8c8f3185dbfef05f7/MarkupSafe-3.0.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:bbcb445fa71794da8f178f0f6d66789a28d7319071af7a496d4d507ed566270d", size = 20866 }, - { url = "https://files.pythonhosted.org/packages/29/28/6d029a903727a1b62edb51863232152fd335d602def598dade38996887f0/MarkupSafe-3.0.2-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:57cb5a3cf367aeb1d316576250f65edec5bb3be939e9247ae594b4bcbc317dfb", size = 20964 }, - { url = "https://files.pythonhosted.org/packages/cc/cd/07438f95f83e8bc028279909d9c9bd39e24149b0d60053a97b2bc4f8aa51/MarkupSafe-3.0.2-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:3809ede931876f5b2ec92eef964286840ed3540dadf803dd570c3b7e13141a3b", size = 21977 }, - { url = "https://files.pythonhosted.org/packages/29/01/84b57395b4cc062f9c4c55ce0df7d3108ca32397299d9df00fedd9117d3d/MarkupSafe-3.0.2-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:e07c3764494e3776c602c1e78e298937c3315ccc9043ead7e685b7f2b8d47b3c", size = 21366 }, - { url = "https://files.pythonhosted.org/packages/bd/6e/61ebf08d8940553afff20d1fb1ba7294b6f8d279df9fd0c0db911b4bbcfd/MarkupSafe-3.0.2-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:b424c77b206d63d500bcb69fa55ed8d0e6a3774056bdc4839fc9298a7edca171", size = 21091 }, - { url = "https://files.pythonhosted.org/packages/11/23/ffbf53694e8c94ebd1e7e491de185124277964344733c45481f32ede2499/MarkupSafe-3.0.2-cp310-cp310-win32.whl", hash = "sha256:fcabf5ff6eea076f859677f5f0b6b5c1a51e70a376b0579e0eadef8db48c6b50", size = 15065 }, - { url = "https://files.pythonhosted.org/packages/44/06/e7175d06dd6e9172d4a69a72592cb3f7a996a9c396eee29082826449bbc3/MarkupSafe-3.0.2-cp310-cp310-win_amd64.whl", hash = "sha256:6af100e168aa82a50e186c82875a5893c5597a0c1ccdb0d8b40240b1f28b969a", size = 15514 }, -] - -[[package]] -name = "marshmallow" -version = "4.0.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "backports-datetime-fromisoformat", marker = "python_full_version == '3.10.12'" }, - { name = "typing-extensions", marker = "python_full_version == '3.10.12'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/1e/ff/26df5a9f5ac57ccf693a5854916ab47243039d2aa9e0fe5f5a0331e7b74b/marshmallow-4.0.0.tar.gz", hash = "sha256:3b6e80aac299a7935cfb97ed01d1854fb90b5079430969af92118ea1b12a8d55", size = 220507 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/d6/26/6cc45d156f44dbe1d5696d9e54042e4dcaf7b946c0b86df6a97d29706f32/marshmallow-4.0.0-py3-none-any.whl", hash = "sha256:e7b0528337e9990fd64950f8a6b3a1baabed09ad17a0dfb844d701151f92d203", size = 48420 }, -] - -[[package]] -name = "matplotlib" -version = "3.10.3" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "contourpy" }, - { name = "cycler" }, - { name = "fonttools" }, - { name = "kiwisolver" }, - { name = "numpy" }, - { name = "packaging" }, - { name = "pillow" }, - { name = "pyparsing" }, - { name = "python-dateutil" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/26/91/d49359a21893183ed2a5b6c76bec40e0b1dcbf8ca148f864d134897cfc75/matplotlib-3.10.3.tar.gz", hash = "sha256:2f82d2c5bb7ae93aaaa4cd42aca65d76ce6376f83304fa3a630b569aca274df0", size = 34799811 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/d0/ea/2bba25d289d389c7451f331ecd593944b3705f06ddf593fa7be75037d308/matplotlib-3.10.3-cp310-cp310-macosx_10_12_x86_64.whl", hash = "sha256:213fadd6348d106ca7db99e113f1bea1e65e383c3ba76e8556ba4a3054b65ae7", size = 8167862 }, - { url = "https://files.pythonhosted.org/packages/41/81/cc70b5138c926604e8c9ed810ed4c79e8116ba72e02230852f5c12c87ba2/matplotlib-3.10.3-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:d3bec61cb8221f0ca6313889308326e7bb303d0d302c5cc9e523b2f2e6c73deb", size = 8042149 }, - { url = "https://files.pythonhosted.org/packages/4a/9a/0ff45b6bfa42bb16de597e6058edf2361c298ad5ef93b327728145161bbf/matplotlib-3.10.3-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8c21ae75651c0231b3ba014b6d5e08fb969c40cdb5a011e33e99ed0c9ea86ecb", size = 8453719 }, - { url = "https://files.pythonhosted.org/packages/85/c7/1866e972fed6d71ef136efbc980d4d1854ab7ef1ea8152bbd995ca231c81/matplotlib-3.10.3-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a49e39755580b08e30e3620efc659330eac5d6534ab7eae50fa5e31f53ee4e30", size = 8590801 }, - { url = "https://files.pythonhosted.org/packages/5d/b9/748f6626d534ab7e255bdc39dc22634d337cf3ce200f261b5d65742044a1/matplotlib-3.10.3-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:cf4636203e1190871d3a73664dea03d26fb019b66692cbfd642faafdad6208e8", size = 9402111 }, - { url = "https://files.pythonhosted.org/packages/1f/78/8bf07bd8fb67ea5665a6af188e70b57fcb2ab67057daa06b85a08e59160a/matplotlib-3.10.3-cp310-cp310-win_amd64.whl", hash = "sha256:fd5641a9bb9d55f4dd2afe897a53b537c834b9012684c8444cc105895c8c16fd", size = 8057213 }, - { url = "https://files.pythonhosted.org/packages/3d/d1/f54d43e95384b312ffa4a74a4326c722f3b8187aaaa12e9a84cdf3037131/matplotlib-3.10.3-pp310-pypy310_pp73-macosx_10_15_x86_64.whl", hash = "sha256:86ab63d66bbc83fdb6733471d3bff40897c1e9921cba112accd748eee4bce5e4", size = 8162896 }, - { url = "https://files.pythonhosted.org/packages/24/a4/fbfc00c2346177c95b353dcf9b5a004106abe8730a62cb6f27e79df0a698/matplotlib-3.10.3-pp310-pypy310_pp73-macosx_11_0_arm64.whl", hash = "sha256:a48f9c08bf7444b5d2391a83e75edb464ccda3c380384b36532a0962593a1751", size = 8039702 }, - { url = "https://files.pythonhosted.org/packages/6a/b9/59e120d24a2ec5fc2d30646adb2efb4621aab3c6d83d66fb2a7a182db032/matplotlib-3.10.3-pp310-pypy310_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:cb73d8aa75a237457988f9765e4dfe1c0d2453c5ca4eabc897d4309672c8e014", size = 8594298 }, -] - -[[package]] -name = "matplotlib-inline" -version = "0.1.7" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "traitlets" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/99/5b/a36a337438a14116b16480db471ad061c36c3694df7c2084a0da7ba538b7/matplotlib_inline-0.1.7.tar.gz", hash = "sha256:8423b23ec666be3d16e16b60bdd8ac4e86e840ebd1dd11a30b9f117f2fa0ab90", size = 8159 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/8f/8e/9ad090d3553c280a8060fbf6e24dc1c0c29704ee7d1c372f0c174aa59285/matplotlib_inline-0.1.7-py3-none-any.whl", hash = "sha256:df192d39a4ff8f21b1895d72e6a13f5fcc5099f00fa84384e0ea28c2cc0653ca", size = 9899 }, -] - -[[package]] -name = "mdurl" -version = "0.1.2" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/d6/54/cfe61301667036ec958cb99bd3efefba235e65cdeb9c84d24a8293ba1d90/mdurl-0.1.2.tar.gz", hash = "sha256:bb413d29f5eea38f31dd4754dd7377d4465116fb207585f97bf925588687c1ba", size = 8729 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/b3/38/89ba8ad64ae25be8de66a6d463314cf1eb366222074cfda9ee839c56a4b4/mdurl-0.1.2-py3-none-any.whl", hash = "sha256:84008a41e51615a49fc9966191ff91509e3c40b939176e643fd50a5c2196b8f8", size = 9979 }, -] - -[[package]] -name = "mediapy" -version = "1.1.6" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "ipython" }, - { name = "matplotlib" }, - { name = "numpy" }, - { name = "pillow" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/9e/b2/451be65c13d2d69b7601eded7ddd3f150884486715a9b3a705ffb08d0177/mediapy-1.1.6.tar.gz", hash = "sha256:9f44b760400964d8bea5121a213f94dc9a225d026d6a819901283a695e585634", size = 25459 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/e3/a0/0d55c59ea8a5f1b13b3eb931e1c0eab9c2a07322cad79cb51a596d2d2a5c/mediapy-1.1.6-py3-none-any.whl", hash = "sha256:c74370808b445666f95272bfdf0eb5707a43b7e05e5527f2dd0830e6892f976f", size = 24955 }, -] - -[[package]] -name = "mistune" -version = "3.1.3" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "typing-extensions", marker = "python_full_version == '3.10.12'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/c4/79/bda47f7dd7c3c55770478d6d02c9960c430b0cf1773b72366ff89126ea31/mistune-3.1.3.tar.gz", hash = "sha256:a7035c21782b2becb6be62f8f25d3df81ccb4d6fa477a6525b15af06539f02a0", size = 94347 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/01/4d/23c4e4f09da849e127e9f123241946c23c1e30f45a88366879e064211815/mistune-3.1.3-py3-none-any.whl", hash = "sha256:1a32314113cff28aa6432e99e522677c8587fd83e3d51c29b82a52409c842bd9", size = 53410 }, -] - -[[package]] -name = "modelscope" -version = "1.30.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "filelock" }, - { name = "requests" }, - { name = "setuptools" }, - { name = "tqdm" }, - { name = "urllib3" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/cc/cb/2c8a53914c17c2de204ea45b6ece2e54dd5afb6973cc3b3a295a05bdbaf3/modelscope-1.30.0.tar.gz", hash = "sha256:1ebedc419881852a784890330e3b6638ab934fa57f82ffa7e2b81b191fd5f76e", size = 4439412 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/8b/2b/3f7efc0538766ecfc082e47d8358abdfad764397665e8576e0c00edaf0c1/modelscope-1.30.0-py3-none-any.whl", hash = "sha256:a391c63a20f286e50010498ec3cc2aa1a86210a9dd74c55cfa4609fd2f8a999f", size = 5922191 }, -] - -[[package]] -name = "more-itertools" -version = "10.7.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/ce/a0/834b0cebabbfc7e311f30b46c8188790a37f89fc8d756660346fe5abfd09/more_itertools-10.7.0.tar.gz", hash = "sha256:9fddd5403be01a94b204faadcff459ec3568cf110265d3c54323e1e866ad29d3", size = 127671 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/2b/9f/7ba6f94fc1e9ac3d2b853fdff3035fb2fa5afbed898c4a72b8a020610594/more_itertools-10.7.0-py3-none-any.whl", hash = "sha256:d43980384673cb07d2f7d2d918c616b30c659c089ee23953f601d6609c67510e", size = 65278 }, -] - -[[package]] -name = "mpmath" -version = "1.3.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/e0/47/dd32fa426cc72114383ac549964eecb20ecfd886d1e5ccf5340b55b02f57/mpmath-1.3.0.tar.gz", hash = "sha256:7a28eb2a9774d00c7bc92411c19a89209d5da7c4c9a9e227be8330a23a25b91f", size = 508106 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/43/e3/7d92a15f894aa0c9c4b49b8ee9ac9850d6e63b03c9c32c0367a13ae62209/mpmath-1.3.0-py3-none-any.whl", hash = "sha256:a0b2b9fe80bbcd81a6647ff13108738cfb482d481d826cc0e02f5b35e5c88d2c", size = 536198 }, -] - -[[package]] -name = "msgpack" -version = "1.1.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/45/b1/ea4f68038a18c77c9467400d166d74c4ffa536f34761f7983a104357e614/msgpack-1.1.1.tar.gz", hash = "sha256:77b79ce34a2bdab2594f490c8e80dd62a02d650b91a75159a63ec413b8d104cd", size = 173555 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/33/52/f30da112c1dc92cf64f57d08a273ac771e7b29dea10b4b30369b2d7e8546/msgpack-1.1.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:353b6fc0c36fde68b661a12949d7d49f8f51ff5fa019c1e47c87c4ff34b080ed", size = 81799 }, - { url = "https://files.pythonhosted.org/packages/e4/35/7bfc0def2f04ab4145f7f108e3563f9b4abae4ab0ed78a61f350518cc4d2/msgpack-1.1.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:79c408fcf76a958491b4e3b103d1c417044544b68e96d06432a189b43d1215c8", size = 78278 }, - { url = "https://files.pythonhosted.org/packages/e8/c5/df5d6c1c39856bc55f800bf82778fd4c11370667f9b9e9d51b2f5da88f20/msgpack-1.1.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:78426096939c2c7482bf31ef15ca219a9e24460289c00dd0b94411040bb73ad2", size = 402805 }, - { url = "https://files.pythonhosted.org/packages/20/8e/0bb8c977efecfe6ea7116e2ed73a78a8d32a947f94d272586cf02a9757db/msgpack-1.1.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8b17ba27727a36cb73aabacaa44b13090feb88a01d012c0f4be70c00f75048b4", size = 408642 }, - { url = "https://files.pythonhosted.org/packages/59/a1/731d52c1aeec52006be6d1f8027c49fdc2cfc3ab7cbe7c28335b2910d7b6/msgpack-1.1.1-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:7a17ac1ea6ec3c7687d70201cfda3b1e8061466f28f686c24f627cae4ea8efd0", size = 395143 }, - { url = "https://files.pythonhosted.org/packages/2b/92/b42911c52cda2ba67a6418ffa7d08969edf2e760b09015593c8a8a27a97d/msgpack-1.1.1-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:88d1e966c9235c1d4e2afac21ca83933ba59537e2e2727a999bf3f515ca2af26", size = 395986 }, - { url = "https://files.pythonhosted.org/packages/61/dc/8ae165337e70118d4dab651b8b562dd5066dd1e6dd57b038f32ebc3e2f07/msgpack-1.1.1-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:f6d58656842e1b2ddbe07f43f56b10a60f2ba5826164910968f5933e5178af75", size = 402682 }, - { url = "https://files.pythonhosted.org/packages/58/27/555851cb98dcbd6ce041df1eacb25ac30646575e9cd125681aa2f4b1b6f1/msgpack-1.1.1-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:96decdfc4adcbc087f5ea7ebdcfd3dee9a13358cae6e81d54be962efc38f6338", size = 406368 }, - { url = "https://files.pythonhosted.org/packages/d4/64/39a26add4ce16f24e99eabb9005e44c663db00e3fce17d4ae1ae9d61df99/msgpack-1.1.1-cp310-cp310-win32.whl", hash = "sha256:6640fd979ca9a212e4bcdf6eb74051ade2c690b862b679bfcb60ae46e6dc4bfd", size = 65004 }, - { url = "https://files.pythonhosted.org/packages/7d/18/73dfa3e9d5d7450d39debde5b0d848139f7de23bd637a4506e36c9800fd6/msgpack-1.1.1-cp310-cp310-win_amd64.whl", hash = "sha256:8b65b53204fe1bd037c40c4148d00ef918eb2108d24c9aaa20bc31f9810ce0a8", size = 71548 }, -] - -[[package]] -name = "multidict" -version = "6.6.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "typing-extensions", marker = "python_full_version == '3.10.12'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/03/5d/d72502cd6dd64b0c5a5117b1701f05c38e94ffb4a1b4ab65ff0cd9b974e8/multidict-6.6.2.tar.gz", hash = "sha256:c1e8b8b0523c0361a78ce9b99d9850c51cf25e1fa3c5686030ce75df6fdf2918", size = 100939 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/1d/9d/9bcb4da29ff4e5a5dd7dccefaf49de8acae5b027e1a8b53296ac61ba14ab/multidict-6.6.2-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:cfd9c74d337e710d7ee26e72a7dbedbd60e0c58d3df7c5ccbb748857e977783c", size = 76829 }, - { url = "https://files.pythonhosted.org/packages/8d/40/4ca4b3eb34d4b57bb0a7385ca206fc28bc62aeb99daee47e72463efcdfc6/multidict-6.6.2-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:9d2c5867a1bd182041a950e9ec3dd3622926260434655bd5d94a62d889100787", size = 44799 }, - { url = "https://files.pythonhosted.org/packages/f5/d7/30ef84053dcb9f4a3ae9b0057b89da3236683ece29ded9b489793addf660/multidict-6.6.2-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:bc8551dd0000ce3f1d909906415ec18970fedb78e685dcac3a0b331a3422d810", size = 44476 }, - { url = "https://files.pythonhosted.org/packages/dd/3a/4cc34184902534abd2f82d9cfd159a333fd56602aa4de4644aaa441f3e6b/multidict-6.6.2-cp310-cp310-manylinux1_i686.manylinux2014_i686.manylinux_2_17_i686.manylinux_2_5_i686.whl", hash = "sha256:9a23d9360f656c316518c8534685ca7c9f18877f782c11bcfce97ff1012ba256", size = 225204 }, - { url = "https://files.pythonhosted.org/packages/f0/20/fb362a4b56a050c10480a81d4d04ce461e01b2f8d02f1e41d2367849670d/multidict-6.6.2-cp310-cp310-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:37fe64cfc6f73fce34f2ef9e099efb8333650b85b50929ba37789311283f476f", size = 244463 }, - { url = "https://files.pythonhosted.org/packages/4e/a0/84aec746dc7e5db95da3c5777aafd8b78ff91a66b3e7d55bcea783d5b3e3/multidict-6.6.2-cp310-cp310-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:2429b25566ff8c12cdf472ee82084ea96ea085675822d6d85aee85efd1d36cc0", size = 221250 }, - { url = "https://files.pythonhosted.org/packages/d9/9b/549656e890c5134b666928fd56d99b7d7eb1579ab62e821a0a3a07d20738/multidict-6.6.2-cp310-cp310-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:66c596bd9bc833bad98445539ad53165b214c2c87bf386dbb819fabd1acdb462", size = 255154 }, - { url = "https://files.pythonhosted.org/packages/c6/de/8ca2242eda642d264a6e6f43a8c1ad9fee5e5aff15b39db8b00afaba5971/multidict-6.6.2-cp310-cp310-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:eb8c006b565a0e53b298e9d48ef5aafe343f77de65c4fa7adb3d3b752a22d10b", size = 251359 }, - { url = "https://files.pythonhosted.org/packages/e5/d7/34c3743e2dce6777c45dff9f451297b0e9a64e145ba3b59c6d5a8834a245/multidict-6.6.2-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:7d09a7ade505d4556aa00c18f5635c9e7fe5973b98fee4e034162b02e48da7bc", size = 242695 }, - { url = "https://files.pythonhosted.org/packages/33/ab/20d63595785766d1d0aac9850b972b9ff20d533371a9140d499904dc7ace/multidict-6.6.2-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:6c95573274630213748ecee465410d4e5e44532d97ba9b09481968efd3c1fd2c", size = 240935 }, - { url = "https://files.pythonhosted.org/packages/6c/1e/a7c9b9a756ad45f2c9750471750345eb8ed8b7a921f742cec30fa70a4070/multidict-6.6.2-cp310-cp310-musllinux_1_2_armv7l.whl", hash = "sha256:e45ebeeee1ce0f9a68151cee1afe02eef56f3b6977a580873c179175e5108275", size = 233923 }, - { url = "https://files.pythonhosted.org/packages/44/c6/bb6e4cca146748e2b787d9338009e8c845af48808600b0769215b6a1fea7/multidict-6.6.2-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:329ecbdd77402648ebcb077b342ad6e67396dcf377c67418a733e88476ff3a11", size = 241715 }, - { url = "https://files.pythonhosted.org/packages/bf/24/d3c01293132168f6a29b2a5490ce4a07d34f0bdb5d73a4b2a177623b88bb/multidict-6.6.2-cp310-cp310-musllinux_1_2_ppc64le.whl", hash = "sha256:f449699e273579a7eda79e36a8b7a6aae06a601d115c54e1aeebf08e07ea3ea1", size = 251433 }, - { url = "https://files.pythonhosted.org/packages/0f/c1/dd47ff9571905e722ce9d71161d21bb970d9632224fa7bfdfe4ae59073c3/multidict-6.6.2-cp310-cp310-musllinux_1_2_s390x.whl", hash = "sha256:ed4bb96a59976e4da7e1fbe3a7c37bcb4a16f3b20c5bba8af9a0ce459e14039a", size = 243316 }, - { url = "https://files.pythonhosted.org/packages/1b/51/73906c1101792b8c6232ecbbbb2fe943a01d820b502a3e882b3ed986bad6/multidict-6.6.2-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:4d05496c2779af4698ba8a841b226247a9a515210eff3a029f48d5345255b1d3", size = 238648 }, - { url = "https://files.pythonhosted.org/packages/a8/8d/0174c5602a531da3c1c0e6e8497bd98702ad1793ff3a9988628de8d75a41/multidict-6.6.2-cp310-cp310-win32.whl", hash = "sha256:f96af5fbf6bab448d6dab34e8126f32f86de65034539d4a7077193f7b64a08f6", size = 41362 }, - { url = "https://files.pythonhosted.org/packages/b1/58/40b720fd0a9ba2f492497c27c7d047606b20540be64a4315693054bd1d09/multidict-6.6.2-cp310-cp310-win_amd64.whl", hash = "sha256:44468089034383be86735f64f5d7daa6a1297e338b739403871a63750b95866d", size = 45892 }, - { url = "https://files.pythonhosted.org/packages/72/53/1ab0ca0093516836b3e89aa9a9e7247f06109300a24b7d9fa3c983122394/multidict-6.6.2-cp310-cp310-win_arm64.whl", hash = "sha256:4e36b00dfb630a81f8efd4eb8a67b5b45f0918da3f2c8c4c14d16fc12b682d33", size = 42983 }, - { url = "https://files.pythonhosted.org/packages/0c/30/7b7d121f76ea3ea7561814531e5cc19e75e9b6646818491179c2c875b591/multidict-6.6.2-py3-none-any.whl", hash = "sha256:a7d14275ff2f85a8ff3c2a32e30f94b9fc8a2125b59a4ecc32271a347fad6e78", size = 12312 }, -] - -[[package]] -name = "multiprocess" -version = "0.70.16" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "dill" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/b5/ae/04f39c5d0d0def03247c2893d6f2b83c136bf3320a2154d7b8858f2ba72d/multiprocess-0.70.16.tar.gz", hash = "sha256:161af703d4652a0e1410be6abccecde4a7ddffd19341be0a7011b94aeb171ac1", size = 1772603 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/ef/76/6e712a2623d146d314f17598df5de7224c85c0060ef63fd95cc15a25b3fa/multiprocess-0.70.16-pp310-pypy310_pp73-macosx_10_13_x86_64.whl", hash = "sha256:476887be10e2f59ff183c006af746cb6f1fd0eadcfd4ef49e605cbe2659920ee", size = 134980 }, - { url = "https://files.pythonhosted.org/packages/0f/ab/1e6e8009e380e22254ff539ebe117861e5bdb3bff1fc977920972237c6c7/multiprocess-0.70.16-pp310-pypy310_pp73-manylinux_2_28_x86_64.whl", hash = "sha256:d951bed82c8f73929ac82c61f01a7b5ce8f3e5ef40f5b52553b4f547ce2b08ec", size = 134982 }, - { url = "https://files.pythonhosted.org/packages/bc/f7/7ec7fddc92e50714ea3745631f79bd9c96424cb2702632521028e57d3a36/multiprocess-0.70.16-py310-none-any.whl", hash = "sha256:c4a9944c67bd49f823687463660a2d6daae94c289adff97e0f9d696ba6371d02", size = 134824 }, - { url = "https://files.pythonhosted.org/packages/ea/89/38df130f2c799090c978b366cfdf5b96d08de5b29a4a293df7f7429fa50b/multiprocess-0.70.16-py38-none-any.whl", hash = "sha256:a71d82033454891091a226dfc319d0cfa8019a4e888ef9ca910372a446de4435", size = 132628 }, - { url = "https://files.pythonhosted.org/packages/da/d9/f7f9379981e39b8c2511c9e0326d212accacb82f12fbfdc1aa2ce2a7b2b6/multiprocess-0.70.16-py39-none-any.whl", hash = "sha256:a0bafd3ae1b732eac64be2e72038231c1ba97724b60b09400d68f229fcc2fbf3", size = 133351 }, -] - -[[package]] -name = "nbclient" -version = "0.10.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "jupyter-client" }, - { name = "jupyter-core" }, - { name = "nbformat" }, - { name = "traitlets" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/87/66/7ffd18d58eae90d5721f9f39212327695b749e23ad44b3881744eaf4d9e8/nbclient-0.10.2.tar.gz", hash = "sha256:90b7fc6b810630db87a6d0c2250b1f0ab4cf4d3c27a299b0cde78a4ed3fd9193", size = 62424 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/34/6d/e7fa07f03a4a7b221d94b4d586edb754a9b0dc3c9e2c93353e9fa4e0d117/nbclient-0.10.2-py3-none-any.whl", hash = "sha256:4ffee11e788b4a27fabeb7955547e4318a5298f34342a4bfd01f2e1faaeadc3d", size = 25434 }, -] - -[[package]] -name = "nbconvert" -version = "7.16.6" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "beautifulsoup4" }, - { name = "bleach", extra = ["css"] }, - { name = "defusedxml" }, - { name = "jinja2" }, - { name = "jupyter-core" }, - { name = "jupyterlab-pygments" }, - { name = "markupsafe" }, - { name = "mistune" }, - { name = "nbclient" }, - { name = "nbformat" }, - { name = "packaging" }, - { name = "pandocfilters" }, - { name = "pygments" }, - { name = "traitlets" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/a3/59/f28e15fc47ffb73af68a8d9b47367a8630d76e97ae85ad18271b9db96fdf/nbconvert-7.16.6.tar.gz", hash = "sha256:576a7e37c6480da7b8465eefa66c17844243816ce1ccc372633c6b71c3c0f582", size = 857715 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/cc/9a/cd673b2f773a12c992f41309ef81b99da1690426bd2f96957a7ade0d3ed7/nbconvert-7.16.6-py3-none-any.whl", hash = "sha256:1375a7b67e0c2883678c48e506dc320febb57685e5ee67faa51b18a90f3a712b", size = 258525 }, -] - -[[package]] -name = "nbformat" -version = "5.10.4" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "fastjsonschema" }, - { name = "jsonschema" }, - { name = "jupyter-core" }, - { name = "traitlets" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/6d/fd/91545e604bc3dad7dca9ed03284086039b294c6b3d75c0d2fa45f9e9caf3/nbformat-5.10.4.tar.gz", hash = "sha256:322168b14f937a5d11362988ecac2a4952d3d8e3a2cbeb2319584631226d5b3a", size = 142749 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/a9/82/0340caa499416c78e5d8f5f05947ae4bc3cba53c9f038ab6e9ed964e22f1/nbformat-5.10.4-py3-none-any.whl", hash = "sha256:3b48d6c8fbca4b299bf3982ea7db1af21580e4fec269ad087b9e81588891200b", size = 78454 }, -] - -[[package]] -name = "nemo-toolkit" -version = "2.3.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "fsspec" }, - { name = "huggingface-hub" }, - { name = "numba" }, - { name = "numpy" }, - { name = "onnx" }, - { name = "protobuf" }, - { name = "python-dateutil" }, - { name = "ruamel-yaml" }, - { name = "scikit-learn" }, - { name = "setuptools" }, - { name = "tensorboard" }, - { name = "text-unidecode" }, - { name = "torch" }, - { name = "tqdm" }, - { name = "wget" }, - { name = "wrapt" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/b1/9e/534533c7b80578b47e4d59f095f8e2b782afcbf8bb3f090ea882e42201ea/nemo_toolkit-2.3.1.tar.gz", hash = "sha256:8ac1444651fd36a315efad10b2837a91cba9b79f63c00da211454b451e627853", size = 4364579 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/b2/b4/821c0253875f9cd206ff9c5c2f0245b42947e84a201c7c3d9118511b737e/nemo_toolkit-2.3.1-py3-none-any.whl", hash = "sha256:2ecbc805ab368123fe047dd712ee5a6494c2a14eb00bb54481f5c12fcafd077a", size = 5965438 }, -] - -[[package]] -name = "nest-asyncio" -version = "1.6.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/83/f8/51569ac65d696c8ecbee95938f89d4abf00f47d58d48f6fbabfe8f0baefe/nest_asyncio-1.6.0.tar.gz", hash = "sha256:6f172d5449aca15afd6c646851f4e31e02c598d553a667e38cafa997cfec55fe", size = 7418 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/a0/c4/c2971a3ba4c6103a3d10c4b0f24f461ddc027f0f09763220cf35ca1401b3/nest_asyncio-1.6.0-py3-none-any.whl", hash = "sha256:87af6efd6b5e897c81050477ef65c62e2b2f35d51703cae01aff2905b1852e1c", size = 5195 }, -] - -[[package]] -name = "networkx" -version = "3.4.2" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/fd/1d/06475e1cd5264c0b870ea2cc6fdb3e37177c1e565c43f56ff17a10e3937f/networkx-3.4.2.tar.gz", hash = "sha256:307c3669428c5362aab27c8a1260aa8f47c4e91d3891f48be0141738d8d053e1", size = 2151368 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/b9/54/dd730b32ea14ea797530a4479b2ed46a6fb250f682a9cfb997e968bf0261/networkx-3.4.2-py3-none-any.whl", hash = "sha256:df5d4365b724cf81b8c6a7312509d0c22386097011ad1abe274afd5e9d3bbc5f", size = 1723263 }, -] - -[[package]] -name = "nltk" -version = "3.9.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "click" }, - { name = "joblib" }, - { name = "regex" }, - { name = "tqdm" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/3c/87/db8be88ad32c2d042420b6fd9ffd4a149f9a0d7f0e86b3f543be2eeeedd2/nltk-3.9.1.tar.gz", hash = "sha256:87d127bd3de4bd89a4f81265e5fa59cb1b199b27440175370f7417d2bc7ae868", size = 2904691 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/4d/66/7d9e26593edda06e8cb531874633f7c2372279c3b0f46235539fe546df8b/nltk-3.9.1-py3-none-any.whl", hash = "sha256:4fa26829c5b00715afe3061398a8989dc643b92ce7dd93fb4585a70930d168a1", size = 1505442 }, -] - -[[package]] -name = "notebook" -version = "7.4.3" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "jupyter-server" }, - { name = "jupyterlab" }, - { name = "jupyterlab-server" }, - { name = "notebook-shim" }, - { name = "tornado" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/dc/21/4f83b15e483da4f4f63928edd0cb08b6e7d33f8a15c23b116a90c44c6235/notebook-7.4.3.tar.gz", hash = "sha256:a1567481cd3853f2610ee0ecf5dfa12bb508e878ee8f92152c134ef7f0568a76", size = 13881668 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/76/1b/16c809d799e3ddd7a97c8b43734f79624b74ddef9707e7d92275a13777bc/notebook-7.4.3-py3-none-any.whl", hash = "sha256:9cdeee954e04101cadb195d90e2ab62b7c9286c1d4f858bf3bb54e40df16c0c3", size = 14286402 }, -] - -[[package]] -name = "notebook-shim" -version = "0.2.4" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "jupyter-server" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/54/d2/92fa3243712b9a3e8bafaf60aac366da1cada3639ca767ff4b5b3654ec28/notebook_shim-0.2.4.tar.gz", hash = "sha256:b4b2cfa1b65d98307ca24361f5b30fe785b53c3fd07b7a47e89acb5e6ac638cb", size = 13167 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/f9/33/bd5b9137445ea4b680023eb0469b2bb969d61303dedb2aac6560ff3d14a1/notebook_shim-0.2.4-py3-none-any.whl", hash = "sha256:411a5be4e9dc882a074ccbcae671eda64cceb068767e9a3419096986560e1cef", size = 13307 }, -] - -[[package]] -name = "num2words" -version = "0.5.14" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "docopt" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/f6/58/ad645bd38b4b648eb2fc2ba1b909398e54eb0cbb6a7dbd2b4953e38c9621/num2words-0.5.14.tar.gz", hash = "sha256:b066ec18e56b6616a3b38086b5747daafbaa8868b226a36127e0451c0cf379c6", size = 218213 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/d6/5b/545e9267a1cc080c8a1be2746113a063e34bcdd0f5173fd665a5c13cb234/num2words-0.5.14-py3-none-any.whl", hash = "sha256:1c8e5b00142fc2966fd8d685001e36c4a9911e070d1b120e1beb721fa1edb33d", size = 163525 }, -] - -[[package]] -name = "numba" -version = "0.61.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "llvmlite" }, - { name = "numpy" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/3c/88/c13a935f200fda51384411e49840a8e7f70c9cb1ee8d809dd0f2477cf7ef/numba-0.61.0.tar.gz", hash = "sha256:888d2e89b8160899e19591467e8fdd4970e07606e1fbc248f239c89818d5f925", size = 2816484 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/77/97/8568a025b9ab8b4d53491e70d4206d5f3fc71fbe94f3097058e01ad8e7ff/numba-0.61.0-cp310-cp310-macosx_10_14_x86_64.whl", hash = "sha256:9cab9783a700fa428b1a54d65295122bc03b3de1d01fb819a6b9dbbddfdb8c43", size = 2769008 }, - { url = "https://files.pythonhosted.org/packages/8c/ab/a88c20755f66543ee01c85c98b866595b92e1bd0ed80565a4889e22929a8/numba-0.61.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:46c5ae094fb3706f5adf9021bfb7fc11e44818d61afee695cdee4eadfed45e98", size = 2771815 }, - { url = "https://files.pythonhosted.org/packages/ae/f4/b357913089ecec1a9ddc6adc04090396928f36a484a5ab9e71b24ddba4cd/numba-0.61.0-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:6fb74e81aa78a2303e30593d8331327dfc0d2522b5db05ac967556a26db3ef87", size = 3820233 }, - { url = "https://files.pythonhosted.org/packages/ea/60/0e21bcf3baaf10e39d48cd224618e46a6b75d3394f465c37ce57bf98cbfa/numba-0.61.0-cp310-cp310-manylinux_2_28_aarch64.whl", hash = "sha256:0ebbd4827091384ab8c4615ba1b3ca8bc639a3a000157d9c37ba85d34cd0da1b", size = 3514707 }, - { url = "https://files.pythonhosted.org/packages/a0/08/45c136ab59e6b11e61ce15a0d17ef03fd89eaccb0db05ad67912aaf5218a/numba-0.61.0-cp310-cp310-win_amd64.whl", hash = "sha256:43aa4d7d10c542d3c78106b8481e0cbaaec788c39ee8e3d7901682748ffdf0b4", size = 2827753 }, -] - -[[package]] -name = "numpy" -version = "1.26.4" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/65/6e/09db70a523a96d25e115e71cc56a6f9031e7b8cd166c1ac8438307c14058/numpy-1.26.4.tar.gz", hash = "sha256:2a02aba9ed12e4ac4eb3ea9421c420301a0c6460d9830d74a9df87efa4912010", size = 15786129 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/a7/94/ace0fdea5241a27d13543ee117cbc65868e82213fb31a8eb7fe9ff23f313/numpy-1.26.4-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:9ff0f4f29c51e2803569d7a51c2304de5554655a60c5d776e35b4a41413830d0", size = 20631468 }, - { url = "https://files.pythonhosted.org/packages/20/f7/b24208eba89f9d1b58c1668bc6c8c4fd472b20c45573cb767f59d49fb0f6/numpy-1.26.4-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:2e4ee3380d6de9c9ec04745830fd9e2eccb3e6cf790d39d7b98ffd19b0dd754a", size = 13966411 }, - { url = "https://files.pythonhosted.org/packages/fc/a5/4beee6488160798683eed5bdb7eead455892c3b4e1f78d79d8d3f3b084ac/numpy-1.26.4-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d209d8969599b27ad20994c8e41936ee0964e6da07478d6c35016bc386b66ad4", size = 14219016 }, - { url = "https://files.pythonhosted.org/packages/4b/d7/ecf66c1cd12dc28b4040b15ab4d17b773b87fa9d29ca16125de01adb36cd/numpy-1.26.4-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ffa75af20b44f8dba823498024771d5ac50620e6915abac414251bd971b4529f", size = 18240889 }, - { url = "https://files.pythonhosted.org/packages/24/03/6f229fe3187546435c4f6f89f6d26c129d4f5bed40552899fcf1f0bf9e50/numpy-1.26.4-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:62b8e4b1e28009ef2846b4c7852046736bab361f7aeadeb6a5b89ebec3c7055a", size = 13876746 }, - { url = "https://files.pythonhosted.org/packages/39/fe/39ada9b094f01f5a35486577c848fe274e374bbf8d8f472e1423a0bbd26d/numpy-1.26.4-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:a4abb4f9001ad2858e7ac189089c42178fcce737e4169dc61321660f1a96c7d2", size = 18078620 }, - { url = "https://files.pythonhosted.org/packages/d5/ef/6ad11d51197aad206a9ad2286dc1aac6a378059e06e8cf22cd08ed4f20dc/numpy-1.26.4-cp310-cp310-win32.whl", hash = "sha256:bfe25acf8b437eb2a8b2d49d443800a5f18508cd811fea3181723922a8a82b07", size = 5972659 }, - { url = "https://files.pythonhosted.org/packages/19/77/538f202862b9183f54108557bfda67e17603fc560c384559e769321c9d92/numpy-1.26.4-cp310-cp310-win_amd64.whl", hash = "sha256:b97fe8060236edf3662adfc2c633f56a08ae30560c56310562cb4f95500022d5", size = 15808905 }, -] - -[[package]] -name = "nvidia-cublas-cu12" -version = "12.6.4.1" -source = { registry = "https://pypi.org/simple" } -wheels = [ - { url = "https://files.pythonhosted.org/packages/af/eb/ff4b8c503fa1f1796679dce648854d58751982426e4e4b37d6fce49d259c/nvidia_cublas_cu12-12.6.4.1-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:08ed2686e9875d01b58e3cb379c6896df8e76c75e0d4a7f7dace3d7b6d9ef8eb", size = 393138322 }, - { url = "https://files.pythonhosted.org/packages/97/0d/f1f0cadbf69d5b9ef2e4f744c9466cb0a850741d08350736dfdb4aa89569/nvidia_cublas_cu12-12.6.4.1-py3-none-manylinux_2_27_aarch64.whl", hash = "sha256:235f728d6e2a409eddf1df58d5b0921cf80cfa9e72b9f2775ccb7b4a87984668", size = 390794615 }, -] - -[[package]] -name = "nvidia-cuda-cupti-cu12" -version = "12.6.80" -source = { registry = "https://pypi.org/simple" } -wheels = [ - { url = "https://files.pythonhosted.org/packages/e6/8b/2f6230cb715646c3a9425636e513227ce5c93c4d65823a734f4bb86d43c3/nvidia_cuda_cupti_cu12-12.6.80-py3-none-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:166ee35a3ff1587f2490364f90eeeb8da06cd867bd5b701bf7f9a02b78bc63fc", size = 8236764 }, - { url = "https://files.pythonhosted.org/packages/25/0f/acb326ac8fd26e13c799e0b4f3b2751543e1834f04d62e729485872198d4/nvidia_cuda_cupti_cu12-12.6.80-py3-none-manylinux2014_aarch64.whl", hash = "sha256:358b4a1d35370353d52e12f0a7d1769fc01ff74a191689d3870b2123156184c4", size = 8236756 }, - { url = "https://files.pythonhosted.org/packages/49/60/7b6497946d74bcf1de852a21824d63baad12cd417db4195fc1bfe59db953/nvidia_cuda_cupti_cu12-12.6.80-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:6768bad6cab4f19e8292125e5f1ac8aa7d1718704012a0e3272a6f61c4bce132", size = 8917980 }, - { url = "https://files.pythonhosted.org/packages/a5/24/120ee57b218d9952c379d1e026c4479c9ece9997a4fb46303611ee48f038/nvidia_cuda_cupti_cu12-12.6.80-py3-none-manylinux2014_x86_64.whl", hash = "sha256:a3eff6cdfcc6a4c35db968a06fcadb061cbc7d6dde548609a941ff8701b98b73", size = 8917972 }, -] - -[[package]] -name = "nvidia-cuda-nvrtc-cu12" -version = "12.6.77" -source = { registry = "https://pypi.org/simple" } -wheels = [ - { url = "https://files.pythonhosted.org/packages/f4/2f/72df534873235983cc0a5371c3661bebef7c4682760c275590b972c7b0f9/nvidia_cuda_nvrtc_cu12-12.6.77-py3-none-manylinux2014_aarch64.whl", hash = "sha256:5847f1d6e5b757f1d2b3991a01082a44aad6f10ab3c5c0213fa3e25bddc25a13", size = 23162955 }, - { url = "https://files.pythonhosted.org/packages/75/2e/46030320b5a80661e88039f59060d1790298b4718944a65a7f2aeda3d9e9/nvidia_cuda_nvrtc_cu12-12.6.77-py3-none-manylinux2014_x86_64.whl", hash = "sha256:35b0cc6ee3a9636d5409133e79273ce1f3fd087abb0532d2d2e8fff1fe9efc53", size = 23650380 }, -] - -[[package]] -name = "nvidia-cuda-runtime-cu12" -version = "12.6.77" -source = { registry = "https://pypi.org/simple" } -wheels = [ - { url = "https://files.pythonhosted.org/packages/8f/ea/590b2ac00d772a8abd1c387a92b46486d2679ca6622fd25c18ff76265663/nvidia_cuda_runtime_cu12-12.6.77-py3-none-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:6116fad3e049e04791c0256a9778c16237837c08b27ed8c8401e2e45de8d60cd", size = 908052 }, - { url = "https://files.pythonhosted.org/packages/b7/3d/159023799677126e20c8fd580cca09eeb28d5c5a624adc7f793b9aa8bbfa/nvidia_cuda_runtime_cu12-12.6.77-py3-none-manylinux2014_aarch64.whl", hash = "sha256:d461264ecb429c84c8879a7153499ddc7b19b5f8d84c204307491989a365588e", size = 908040 }, - { url = "https://files.pythonhosted.org/packages/e1/23/e717c5ac26d26cf39a27fbc076240fad2e3b817e5889d671b67f4f9f49c5/nvidia_cuda_runtime_cu12-12.6.77-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:ba3b56a4f896141e25e19ab287cd71e52a6a0f4b29d0d31609f60e3b4d5219b7", size = 897690 }, - { url = "https://files.pythonhosted.org/packages/f0/62/65c05e161eeddbafeca24dc461f47de550d9fa8a7e04eb213e32b55cfd99/nvidia_cuda_runtime_cu12-12.6.77-py3-none-manylinux2014_x86_64.whl", hash = "sha256:a84d15d5e1da416dd4774cb42edf5e954a3e60cc945698dc1d5be02321c44dc8", size = 897678 }, -] - -[[package]] -name = "nvidia-cudnn-cu12" -version = "9.5.1.17" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "nvidia-cublas-cu12" }, -] -wheels = [ - { url = "https://files.pythonhosted.org/packages/99/93/a201a12d3ec1caa8c6ac34c1c2f9eeb696b886f0c36ff23c638b46603bd0/nvidia_cudnn_cu12-9.5.1.17-py3-none-manylinux_2_28_aarch64.whl", hash = "sha256:9fd4584468533c61873e5fda8ca41bac3a38bcb2d12350830c69b0a96a7e4def", size = 570523509 }, - { url = "https://files.pythonhosted.org/packages/2a/78/4535c9c7f859a64781e43c969a3a7e84c54634e319a996d43ef32ce46f83/nvidia_cudnn_cu12-9.5.1.17-py3-none-manylinux_2_28_x86_64.whl", hash = "sha256:30ac3869f6db17d170e0e556dd6cc5eee02647abc31ca856634d5a40f82c15b2", size = 570988386 }, -] - -[[package]] -name = "nvidia-cufft-cu12" -version = "11.3.0.4" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "nvidia-nvjitlink-cu12" }, -] -wheels = [ - { url = "https://files.pythonhosted.org/packages/1f/37/c50d2b2f2c07e146776389e3080f4faf70bcc4fa6e19d65bb54ca174ebc3/nvidia_cufft_cu12-11.3.0.4-py3-none-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:d16079550df460376455cba121db6564089176d9bac9e4f360493ca4741b22a6", size = 200164144 }, - { url = "https://files.pythonhosted.org/packages/ce/f5/188566814b7339e893f8d210d3a5332352b1409815908dad6a363dcceac1/nvidia_cufft_cu12-11.3.0.4-py3-none-manylinux2014_aarch64.whl", hash = "sha256:8510990de9f96c803a051822618d42bf6cb8f069ff3f48d93a8486efdacb48fb", size = 200164135 }, - { url = "https://files.pythonhosted.org/packages/8f/16/73727675941ab8e6ffd86ca3a4b7b47065edcca7a997920b831f8147c99d/nvidia_cufft_cu12-11.3.0.4-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:ccba62eb9cef5559abd5e0d54ceed2d9934030f51163df018532142a8ec533e5", size = 200221632 }, - { url = "https://files.pythonhosted.org/packages/60/de/99ec247a07ea40c969d904fc14f3a356b3e2a704121675b75c366b694ee1/nvidia_cufft_cu12-11.3.0.4-py3-none-manylinux2014_x86_64.whl", hash = "sha256:768160ac89f6f7b459bee747e8d175dbf53619cfe74b2a5636264163138013ca", size = 200221622 }, -] - -[[package]] -name = "nvidia-cufile-cu12" -version = "1.11.1.6" -source = { registry = "https://pypi.org/simple" } -wheels = [ - { url = "https://files.pythonhosted.org/packages/b2/66/cc9876340ac68ae71b15c743ddb13f8b30d5244af344ec8322b449e35426/nvidia_cufile_cu12-1.11.1.6-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:cc23469d1c7e52ce6c1d55253273d32c565dd22068647f3aa59b3c6b005bf159", size = 1142103 }, - { url = "https://files.pythonhosted.org/packages/17/bf/cc834147263b929229ce4aadd62869f0b195e98569d4c28b23edc72b85d9/nvidia_cufile_cu12-1.11.1.6-py3-none-manylinux_2_27_aarch64.whl", hash = "sha256:8f57a0051dcf2543f6dc2b98a98cb2719c37d3cee1baba8965d57f3bbc90d4db", size = 1066155 }, -] - -[[package]] -name = "nvidia-curand-cu12" -version = "10.3.7.77" -source = { registry = "https://pypi.org/simple" } -wheels = [ - { url = "https://files.pythonhosted.org/packages/42/ac/36543605358a355632f1a6faa3e2d5dfb91eab1e4bc7d552040e0383c335/nvidia_curand_cu12-10.3.7.77-py3-none-manylinux2014_aarch64.whl", hash = "sha256:6e82df077060ea28e37f48a3ec442a8f47690c7499bff392a5938614b56c98d8", size = 56289881 }, - { url = "https://files.pythonhosted.org/packages/73/1b/44a01c4e70933637c93e6e1a8063d1e998b50213a6b65ac5a9169c47e98e/nvidia_curand_cu12-10.3.7.77-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:a42cd1344297f70b9e39a1e4f467a4e1c10f1da54ff7a85c12197f6c652c8bdf", size = 56279010 }, - { url = "https://files.pythonhosted.org/packages/4a/aa/2c7ff0b5ee02eaef890c0ce7d4f74bc30901871c5e45dee1ae6d0083cd80/nvidia_curand_cu12-10.3.7.77-py3-none-manylinux2014_x86_64.whl", hash = "sha256:99f1a32f1ac2bd134897fc7a203f779303261268a65762a623bf30cc9fe79117", size = 56279000 }, - { url = "https://files.pythonhosted.org/packages/a6/02/5362a9396f23f7de1dd8a64369e87c85ffff8216fc8194ace0fa45ba27a5/nvidia_curand_cu12-10.3.7.77-py3-none-manylinux_2_27_aarch64.whl", hash = "sha256:7b2ed8e95595c3591d984ea3603dd66fe6ce6812b886d59049988a712ed06b6e", size = 56289882 }, -] - -[[package]] -name = "nvidia-cusolver-cu12" -version = "11.7.1.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "nvidia-cublas-cu12" }, - { name = "nvidia-cusparse-cu12" }, - { name = "nvidia-nvjitlink-cu12" }, -] -wheels = [ - { url = "https://files.pythonhosted.org/packages/93/17/dbe1aa865e4fdc7b6d4d0dd308fdd5aaab60f939abfc0ea1954eac4fb113/nvidia_cusolver_cu12-11.7.1.2-py3-none-manylinux2014_aarch64.whl", hash = "sha256:0ce237ef60acde1efc457335a2ddadfd7610b892d94efee7b776c64bb1cac9e0", size = 157833628 }, - { url = "https://files.pythonhosted.org/packages/f0/6e/c2cf12c9ff8b872e92b4a5740701e51ff17689c4d726fca91875b07f655d/nvidia_cusolver_cu12-11.7.1.2-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:e9e49843a7707e42022babb9bcfa33c29857a93b88020c4e4434656a655b698c", size = 158229790 }, - { url = "https://files.pythonhosted.org/packages/9f/81/baba53585da791d043c10084cf9553e074548408e04ae884cfe9193bd484/nvidia_cusolver_cu12-11.7.1.2-py3-none-manylinux2014_x86_64.whl", hash = "sha256:6cf28f17f64107a0c4d7802be5ff5537b2130bfc112f25d5a30df227058ca0e6", size = 158229780 }, - { url = "https://files.pythonhosted.org/packages/7c/5f/07d0ba3b7f19be5a5ec32a8679fc9384cfd9fc6c869825e93be9f28d6690/nvidia_cusolver_cu12-11.7.1.2-py3-none-manylinux_2_27_aarch64.whl", hash = "sha256:dbbe4fc38ec1289c7e5230e16248365e375c3673c9c8bac5796e2e20db07f56e", size = 157833630 }, -] - -[[package]] -name = "nvidia-cusparse-cu12" -version = "12.5.4.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "nvidia-nvjitlink-cu12" }, -] -wheels = [ - { url = "https://files.pythonhosted.org/packages/eb/eb/6681efd0aa7df96b4f8067b3ce7246833dd36830bb4cec8896182773db7d/nvidia_cusparse_cu12-12.5.4.2-py3-none-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:d25b62fb18751758fe3c93a4a08eff08effedfe4edf1c6bb5afd0890fe88f887", size = 216451147 }, - { url = "https://files.pythonhosted.org/packages/d3/56/3af21e43014eb40134dea004e8d0f1ef19d9596a39e4d497d5a7de01669f/nvidia_cusparse_cu12-12.5.4.2-py3-none-manylinux2014_aarch64.whl", hash = "sha256:7aa32fa5470cf754f72d1116c7cbc300b4e638d3ae5304cfa4a638a5b87161b1", size = 216451135 }, - { url = "https://files.pythonhosted.org/packages/06/1e/b8b7c2f4099a37b96af5c9bb158632ea9e5d9d27d7391d7eb8fc45236674/nvidia_cusparse_cu12-12.5.4.2-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:7556d9eca156e18184b94947ade0fba5bb47d69cec46bf8660fd2c71a4b48b73", size = 216561367 }, - { url = "https://files.pythonhosted.org/packages/43/ac/64c4316ba163e8217a99680c7605f779accffc6a4bcd0c778c12948d3707/nvidia_cusparse_cu12-12.5.4.2-py3-none-manylinux2014_x86_64.whl", hash = "sha256:23749a6571191a215cb74d1cdbff4a86e7b19f1200c071b3fcf844a5bea23a2f", size = 216561357 }, -] - -[[package]] -name = "nvidia-cusparselt-cu12" -version = "0.6.3" -source = { registry = "https://pypi.org/simple" } -wheels = [ - { url = "https://files.pythonhosted.org/packages/62/da/4de092c61c6dea1fc9c936e69308a02531d122e12f1f649825934ad651b5/nvidia_cusparselt_cu12-0.6.3-py3-none-manylinux2014_aarch64.whl", hash = "sha256:8371549623ba601a06322af2133c4a44350575f5a3108fb75f3ef20b822ad5f1", size = 156402859 }, - { url = "https://files.pythonhosted.org/packages/3b/9a/72ef35b399b0e183bc2e8f6f558036922d453c4d8237dab26c666a04244b/nvidia_cusparselt_cu12-0.6.3-py3-none-manylinux2014_x86_64.whl", hash = "sha256:e5c8a26c36445dd2e6812f1177978a24e2d37cacce7e090f297a688d1ec44f46", size = 156785796 }, -] - -[[package]] -name = "nvidia-nccl-cu12" -version = "2.26.2" -source = { registry = "https://pypi.org/simple" } -wheels = [ - { url = "https://files.pythonhosted.org/packages/69/5b/ca2f213f637305633814ae8c36b153220e40a07ea001966dcd87391f3acb/nvidia_nccl_cu12-2.26.2-py3-none-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:5c196e95e832ad30fbbb50381eb3cbd1fadd5675e587a548563993609af19522", size = 291671495 }, - { url = "https://files.pythonhosted.org/packages/67/ca/f42388aed0fddd64ade7493dbba36e1f534d4e6fdbdd355c6a90030ae028/nvidia_nccl_cu12-2.26.2-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:694cf3879a206553cc9d7dbda76b13efaf610fdb70a50cba303de1b0d1530ac6", size = 201319755 }, -] - -[[package]] -name = "nvidia-nvjitlink-cu12" -version = "12.6.85" -source = { registry = "https://pypi.org/simple" } -wheels = [ - { url = "https://files.pythonhosted.org/packages/9d/d7/c5383e47c7e9bf1c99d5bd2a8c935af2b6d705ad831a7ec5c97db4d82f4f/nvidia_nvjitlink_cu12-12.6.85-py3-none-manylinux2010_x86_64.manylinux_2_12_x86_64.whl", hash = "sha256:eedc36df9e88b682efe4309aa16b5b4e78c2407eac59e8c10a6a47535164369a", size = 19744971 }, - { url = "https://files.pythonhosted.org/packages/31/db/dc71113d441f208cdfe7ae10d4983884e13f464a6252450693365e166dcf/nvidia_nvjitlink_cu12-12.6.85-py3-none-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:cf4eaa7d4b6b543ffd69d6abfb11efdeb2db48270d94dfd3a452c24150829e41", size = 19270338 }, -] - -[[package]] -name = "nvidia-nvtx-cu12" -version = "12.6.77" -source = { registry = "https://pypi.org/simple" } -wheels = [ - { url = "https://files.pythonhosted.org/packages/b9/93/80f8a520375af9d7ee44571a6544653a176e53c2b8ccce85b97b83c2491b/nvidia_nvtx_cu12-12.6.77-py3-none-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:f44f8d86bb7d5629988d61c8d3ae61dddb2015dee142740536bc7481b022fe4b", size = 90549 }, - { url = "https://files.pythonhosted.org/packages/2b/53/36e2fd6c7068997169b49ffc8c12d5af5e5ff209df6e1a2c4d373b3a638f/nvidia_nvtx_cu12-12.6.77-py3-none-manylinux2014_aarch64.whl", hash = "sha256:adcaabb9d436c9761fca2b13959a2d237c5f9fd406c8e4b723c695409ff88059", size = 90539 }, - { url = "https://files.pythonhosted.org/packages/56/9a/fff8376f8e3d084cd1530e1ef7b879bb7d6d265620c95c1b322725c694f4/nvidia_nvtx_cu12-12.6.77-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:b90bed3df379fa79afbd21be8e04a0314336b8ae16768b58f2d34cb1d04cd7d2", size = 89276 }, - { url = "https://files.pythonhosted.org/packages/9e/4e/0d0c945463719429b7bd21dece907ad0bde437a2ff12b9b12fee94722ab0/nvidia_nvtx_cu12-12.6.77-py3-none-manylinux2014_x86_64.whl", hash = "sha256:6574241a3ec5fdc9334353ab8c479fe75841dbe8f4532a8fc97ce63503330ba1", size = 89265 }, -] - -[[package]] -name = "omegaconf" -version = "2.3.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "antlr4-python3-runtime" }, - { name = "pyyaml" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/09/48/6388f1bb9da707110532cb70ec4d2822858ddfb44f1cdf1233c20a80ea4b/omegaconf-2.3.0.tar.gz", hash = "sha256:d5d4b6d29955cc50ad50c46dc269bcd92c6e00f5f90d23ab5fee7bfca4ba4cc7", size = 3298120 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/e3/94/1843518e420fa3ed6919835845df698c7e27e183cb997394e4a670973a65/omegaconf-2.3.0-py3-none-any.whl", hash = "sha256:7b4df175cdb08ba400f45cae3bdcae7ba8365db4d165fc65fd04b050ab63b46b", size = 79500 }, -] - -[[package]] -name = "onnx" -version = "1.17.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "numpy" }, - { name = "protobuf" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/9a/54/0e385c26bf230d223810a9c7d06628d954008a5e5e4b73ee26ef02327282/onnx-1.17.0.tar.gz", hash = "sha256:48ca1a91ff73c1d5e3ea2eef20ae5d0e709bb8a2355ed798ffc2169753013fd3", size = 12165120 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/2e/29/57053ba7787788ac75efb095cfc1ae290436b6d3a26754693cd7ed1b4fac/onnx-1.17.0-cp310-cp310-macosx_12_0_universal2.whl", hash = "sha256:38b5df0eb22012198cdcee527cc5f917f09cce1f88a69248aaca22bd78a7f023", size = 16645616 }, - { url = "https://files.pythonhosted.org/packages/75/0d/831807a18db2a5e8f7813848c59272b904a4ef3939fe4d1288cbce9ea735/onnx-1.17.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d545335cb49d4d8c47cc803d3a805deb7ad5d9094dc67657d66e568610a36d7d", size = 15908420 }, - { url = "https://files.pythonhosted.org/packages/dd/5b/c4f95dbe652d14aeba9afaceb177e9ffc48ac3c03048dd3f872f26f07e34/onnx-1.17.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3193a3672fc60f1a18c0f4c93ac81b761bc72fd8a6c2035fa79ff5969f07713e", size = 16046244 }, - { url = "https://files.pythonhosted.org/packages/08/a9/c1f218085043dccc6311460239e253fa6957cf12ee4b0a56b82014938d0b/onnx-1.17.0-cp310-cp310-win32.whl", hash = "sha256:0141c2ce806c474b667b7e4499164227ef594584da432fd5613ec17c1855e311", size = 14423516 }, - { url = "https://files.pythonhosted.org/packages/0e/d3/d26ebf590a65686dde6b27fef32493026c5be9e42083340d947395f93405/onnx-1.17.0-cp310-cp310-win_amd64.whl", hash = "sha256:dfd777d95c158437fda6b34758f0877d15b89cbe9ff45affbedc519b35345cf9", size = 14528496 }, -] - -[[package]] -name = "optuna" -version = "4.4.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "alembic" }, - { name = "colorlog" }, - { name = "numpy" }, - { name = "packaging" }, - { name = "pyyaml" }, - { name = "sqlalchemy" }, - { name = "tqdm" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/a5/e0/b303190ae8032d12f320a24c42af04038bacb1f3b17ede354dd1044a5642/optuna-4.4.0.tar.gz", hash = "sha256:a9029f6a92a1d6c8494a94e45abd8057823b535c2570819072dbcdc06f1c1da4", size = 467708 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/5c/5e/068798a8c7087863e7772e9363a880ab13fe55a5a7ede8ec42fab8a1acbb/optuna-4.4.0-py3-none-any.whl", hash = "sha256:fad8d9c5d5af993ae1280d6ce140aecc031c514a44c3b639d8c8658a8b7920ea", size = 395949 }, -] - -[[package]] -name = "oss2" -version = "2.19.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "aliyun-python-sdk-core" }, - { name = "aliyun-python-sdk-kms" }, - { name = "crcmod" }, - { name = "pycryptodome" }, - { name = "requests" }, - { name = "six" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/df/b5/f2cb1950dda46ac2284d6c950489fdacd0e743c2d79a347924d3cc44b86f/oss2-2.19.1.tar.gz", hash = "sha256:a8ab9ee7eb99e88a7e1382edc6ea641d219d585a7e074e3776e9dec9473e59c1", size = 298845 } - -[[package]] -name = "overrides" -version = "7.7.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/36/86/b585f53236dec60aba864e050778b25045f857e17f6e5ea0ae95fe80edd2/overrides-7.7.0.tar.gz", hash = "sha256:55158fa3d93b98cc75299b1e67078ad9003ca27945c76162c1c0766d6f91820a", size = 22812 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/2c/ab/fc8290c6a4c722e5514d80f62b2dc4c4df1a68a41d1364e625c35990fcf3/overrides-7.7.0-py3-none-any.whl", hash = "sha256:c7ed9d062f78b8e4c1a7b70bd8796b35ead4d9f510227ef9c5dc7626c60d7e49", size = 17832 }, -] - -[[package]] -name = "packaging" -version = "24.2" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/d0/63/68dbb6eb2de9cb10ee4c9c14a0148804425e13c4fb20d61cce69f53106da/packaging-24.2.tar.gz", hash = "sha256:c228a6dc5e932d346bc5739379109d49e8853dd8223571c7c5b55260edc0b97f", size = 163950 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/88/ef/eb23f262cca3c0c4eb7ab1933c3b1f03d021f2c48f54763065b6f0e321be/packaging-24.2-py3-none-any.whl", hash = "sha256:09abb1bccd265c01f4a3aa3f7a7db064b36514d2cba19a2f694fe6150451a759", size = 65451 }, -] - -[[package]] -name = "pandas" -version = "2.3.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "numpy", marker = "python_full_version == '3.10.12'" }, - { name = "python-dateutil" }, - { name = "pytz" }, - { name = "tzdata" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/72/51/48f713c4c728d7c55ef7444ba5ea027c26998d96d1a40953b346438602fc/pandas-2.3.0.tar.gz", hash = "sha256:34600ab34ebf1131a7613a260a61dbe8b62c188ec0ea4c296da7c9a06b004133", size = 4484490 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/e2/2d/df6b98c736ba51b8eaa71229e8fcd91233a831ec00ab520e1e23090cc072/pandas-2.3.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:625466edd01d43b75b1883a64d859168e4556261a5035b32f9d743b67ef44634", size = 11527531 }, - { url = "https://files.pythonhosted.org/packages/77/1c/3f8c331d223f86ba1d0ed7d3ed7fcf1501c6f250882489cc820d2567ddbf/pandas-2.3.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:a6872d695c896f00df46b71648eea332279ef4077a409e2fe94220208b6bb675", size = 10774764 }, - { url = "https://files.pythonhosted.org/packages/1b/45/d2599400fad7fe06b849bd40b52c65684bc88fbe5f0a474d0513d057a377/pandas-2.3.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:f4dd97c19bd06bc557ad787a15b6489d2614ddaab5d104a0310eb314c724b2d2", size = 11711963 }, - { url = "https://files.pythonhosted.org/packages/66/f8/5508bc45e994e698dbc93607ee6b9b6eb67df978dc10ee2b09df80103d9e/pandas-2.3.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:034abd6f3db8b9880aaee98f4f5d4dbec7c4829938463ec046517220b2f8574e", size = 12349446 }, - { url = "https://files.pythonhosted.org/packages/f7/fc/17851e1b1ea0c8456ba90a2f514c35134dd56d981cf30ccdc501a0adeac4/pandas-2.3.0-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:23c2b2dc5213810208ca0b80b8666670eb4660bbfd9d45f58592cc4ddcfd62e1", size = 12920002 }, - { url = "https://files.pythonhosted.org/packages/a1/9b/8743be105989c81fa33f8e2a4e9822ac0ad4aaf812c00fee6bb09fc814f9/pandas-2.3.0-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:39ff73ec07be5e90330cc6ff5705c651ace83374189dcdcb46e6ff54b4a72cd6", size = 13651218 }, - { url = "https://files.pythonhosted.org/packages/26/fa/8eeb2353f6d40974a6a9fd4081ad1700e2386cf4264a8f28542fd10b3e38/pandas-2.3.0-cp310-cp310-win_amd64.whl", hash = "sha256:40cecc4ea5abd2921682b57532baea5588cc5f80f0231c624056b146887274d2", size = 11082485 }, -] - -[[package]] -name = "pandocfilters" -version = "1.5.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/70/6f/3dd4940bbe001c06a65f88e36bad298bc7a0de5036115639926b0c5c0458/pandocfilters-1.5.1.tar.gz", hash = "sha256:002b4a555ee4ebc03f8b66307e287fa492e4a77b4ea14d3f934328297bb4939e", size = 8454 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/ef/af/4fbc8cab944db5d21b7e2a5b8e9211a03a79852b1157e2c102fcc61ac440/pandocfilters-1.5.1-py2.py3-none-any.whl", hash = "sha256:93be382804a9cdb0a7267585f157e5d1731bbe5545a85b268d6f5fe6232de2bc", size = 8663 }, -] - -[[package]] -name = "parakeet-coreml" -version = "0.1.0" -source = { virtual = "." } -dependencies = [ - { name = "absl-py" }, - { name = "accelerate" }, - { name = "aiohappyeyeballs" }, - { name = "aiohttp" }, - { name = "aiosignal" }, - { name = "alembic" }, - { name = "annotated-types" }, - { name = "antlr4-python3-runtime" }, - { name = "anyio" }, - { name = "appnope" }, - { name = "argon2-cffi" }, - { name = "argon2-cffi-bindings" }, - { name = "arrow" }, - { name = "asttokens" }, - { name = "async-lru" }, - { name = "async-timeout" }, - { name = "attrs" }, - { name = "audioread" }, - { name = "babel" }, - { name = "backports-datetime-fromisoformat" }, - { name = "beautifulsoup4" }, - { name = "bleach" }, - { name = "braceexpand" }, - { name = "cattrs" }, - { name = "certifi" }, - { name = "cffi" }, - { name = "charset-normalizer" }, - { name = "click" }, - { name = "cloudpickle" }, - { name = "colorlog" }, - { name = "comm" }, - { name = "contourpy" }, - { name = "coremltools" }, - { name = "cycler" }, - { name = "cytoolz" }, - { name = "datasets" }, - { name = "debugpy" }, - { name = "decorator" }, - { name = "defusedxml" }, - { name = "dill" }, - { name = "distance" }, - { name = "docopt" }, - { name = "editdistance" }, - { name = "einops" }, - { name = "exceptiongroup" }, - { name = "executing" }, - { name = "fastjsonschema" }, - { name = "fiddle" }, - { name = "filelock" }, - { name = "fonttools" }, - { name = "fqdn" }, - { name = "frozenlist" }, - { name = "fsspec" }, - { name = "funasr" }, - { name = "future" }, - { name = "g2p-en" }, - { name = "gitdb" }, - { name = "gitpython" }, - { name = "graphviz" }, - { name = "grpcio" }, - { name = "h11" }, - { name = "hf-xet" }, - { name = "httpcore" }, - { name = "httpx" }, - { name = "huggingface-hub" }, - { name = "hydra-core" }, - { name = "idna" }, - { name = "inflect" }, - { name = "intervaltree" }, - { name = "ipykernel" }, - { name = "ipython" }, - { name = "ipywidgets" }, - { name = "isoduration" }, - { name = "jedi" }, - { name = "jinja2" }, - { name = "jiwer" }, - { name = "joblib" }, - { name = "json5" }, - { name = "jsonpointer" }, - { name = "jsonschema" }, - { name = "jsonschema-specifications" }, - { name = "jupyter" }, - { name = "jupyter-client" }, - { name = "jupyter-console" }, - { name = "jupyter-core" }, - { name = "jupyter-events" }, - { name = "jupyter-lsp" }, - { name = "jupyter-server" }, - { name = "jupyter-server-terminals" }, - { name = "jupyterlab" }, - { name = "jupyterlab-pygments" }, - { name = "jupyterlab-server" }, - { name = "jupyterlab-widgets" }, - { name = "kaldi-python-io" }, - { name = "kaldiio" }, - { name = "kiwisolver" }, - { name = "lazy-loader" }, - { name = "levenshtein" }, - { name = "lhotse" }, - { name = "libcst" }, - { name = "librosa" }, - { name = "lightning" }, - { name = "lightning-utilities" }, - { name = "lilcom" }, - { name = "llvmlite" }, - { name = "loguru" }, - { name = "mako" }, - { name = "markdown" }, - { name = "markdown-it-py" }, - { name = "markupsafe" }, - { name = "marshmallow" }, - { name = "matplotlib" }, - { name = "matplotlib-inline" }, - { name = "mdurl" }, - { name = "mediapy" }, - { name = "mistune" }, - { name = "more-itertools" }, - { name = "mpmath" }, - { name = "msgpack" }, - { name = "multidict" }, - { name = "multiprocess" }, - { name = "nbclient" }, - { name = "nbconvert" }, - { name = "nbformat" }, - { name = "nemo-toolkit" }, - { name = "nest-asyncio" }, - { name = "networkx" }, - { name = "nltk" }, - { name = "notebook" }, - { name = "notebook-shim" }, - { name = "num2words" }, - { name = "numba" }, - { name = "numpy" }, - { name = "omegaconf" }, - { name = "onnx" }, - { name = "optuna" }, - { name = "overrides" }, - { name = "packaging" }, - { name = "pandas" }, - { name = "pandocfilters" }, - { name = "parso" }, - { name = "peft" }, - { name = "pexpect" }, - { name = "pillow" }, - { name = "pip" }, - { name = "plac" }, - { name = "platformdirs" }, - { name = "pooch" }, - { name = "prometheus-client" }, - { name = "prompt-toolkit" }, - { name = "propcache" }, - { name = "psutil" }, - { name = "ptyprocess" }, - { name = "pure-eval" }, - { name = "pyaml" }, - { name = "pyannote-audio" }, - { name = "pyannote-core" }, - { name = "pyannote-database" }, - { name = "pyannote-metrics" }, - { name = "pyarrow" }, - { name = "pybind11" }, - { name = "pycparser" }, - { name = "pydantic" }, - { name = "pydantic-core" }, - { name = "pydub" }, - { name = "pygments" }, - { name = "pyloudnorm" }, - { name = "pyparsing" }, - { name = "python-dateutil" }, - { name = "python-json-logger" }, - { name = "pytorch-lightning" }, - { name = "pytz" }, - { name = "pyyaml" }, - { name = "pyzmq" }, - { name = "rapidfuzz" }, - { name = "referencing" }, - { name = "regex" }, - { name = "requests" }, - { name = "resampy" }, - { name = "rfc3339-validator" }, - { name = "rfc3986-validator" }, - { name = "rich" }, - { name = "rpds-py" }, - { name = "ruamel-yaml" }, - { name = "ruamel-yaml-clib" }, - { name = "sacremoses" }, - { name = "safetensors" }, - { name = "scikit-learn" }, - { name = "scipy" }, - { name = "seaborn" }, - { name = "send2trash" }, - { name = "sentencepiece" }, - { name = "sentry-sdk" }, - { name = "setproctitle" }, - { name = "shellingham" }, - { name = "six" }, - { name = "smmap" }, - { name = "sniffio" }, - { name = "sortedcontainers" }, - { name = "soundfile" }, - { name = "soupsieve" }, - { name = "sox" }, - { name = "soxr" }, - { name = "sqlalchemy" }, - { name = "stack-data" }, - { name = "tabulate" }, - { name = "tensorboard" }, - { name = "tensorboard-data-server" }, - { name = "termcolor" }, - { name = "terminado" }, - { name = "text-unidecode" }, - { name = "texterrors" }, - { name = "threadpoolctl" }, - { name = "tinycss2" }, - { name = "tokenizers" }, - { name = "tomli" }, - { name = "toolz" }, - { name = "torch" }, - { name = "torchmetrics" }, - { name = "tornado" }, - { name = "tqdm" }, - { name = "traitlets" }, - { name = "transformers" }, - { name = "typeguard" }, - { name = "typer" }, - { name = "types-python-dateutil" }, - { name = "typing-extensions" }, - { name = "typing-inspection" }, - { name = "tzdata" }, - { name = "uri-template" }, - { name = "urllib3" }, - { name = "wandb" }, - { name = "wcwidth" }, - { name = "webcolors" }, - { name = "webdataset" }, - { name = "webencodings" }, - { name = "websocket-client" }, - { name = "werkzeug" }, - { name = "wget" }, - { name = "widgetsnbextension" }, - { name = "wrapt" }, - { name = "xxhash" }, - { name = "yarl" }, -] - -[package.metadata] -requires-dist = [ - { name = "absl-py", specifier = "==2.3.0" }, - { name = "accelerate", specifier = "==1.8.1" }, - { name = "aiohappyeyeballs", specifier = "==2.6.1" }, - { name = "aiohttp", specifier = "==3.12.13" }, - { name = "aiosignal", specifier = "==1.3.2" }, - { name = "alembic", specifier = "==1.16.2" }, - { name = "annotated-types", specifier = "==0.7.0" }, - { name = "antlr4-python3-runtime", specifier = "==4.9.3" }, - { name = "anyio", specifier = "==4.9.0" }, - { name = "appnope", specifier = "==0.1.4" }, - { name = "argon2-cffi", specifier = "==25.1.0" }, - { name = "argon2-cffi-bindings", specifier = "==21.2.0" }, - { name = "arrow", specifier = "==1.3.0" }, - { name = "asttokens", specifier = "==3.0.0" }, - { name = "async-lru", specifier = "==2.0.5" }, - { name = "async-timeout", specifier = "==5.0.1" }, - { name = "attrs", specifier = "==25.3.0" }, - { name = "audioread", specifier = "==3.0.1" }, - { name = "babel", specifier = "==2.17.0" }, - { name = "backports-datetime-fromisoformat", specifier = "==2.0.3" }, - { name = "beautifulsoup4", specifier = "==4.13.4" }, - { name = "bleach", specifier = "==6.2.0" }, - { name = "braceexpand", specifier = "==0.1.7" }, - { name = "cattrs", specifier = "==25.1.1" }, - { name = "certifi", specifier = "==2025.6.15" }, - { name = "cffi", specifier = "==1.17.1" }, - { name = "charset-normalizer", specifier = "==3.4.2" }, - { name = "click", specifier = "==8.2.1" }, - { name = "cloudpickle", specifier = "==3.1.1" }, - { name = "colorlog", specifier = "==6.9.0" }, - { name = "comm", specifier = "==0.2.2" }, - { name = "contourpy", specifier = "==1.3.2" }, - { name = "coremltools", specifier = "==9.0b1" }, - { name = "cycler", specifier = "==0.12.1" }, - { name = "cytoolz", specifier = "==1.0.1" }, - { name = "datasets", specifier = "==3.6.0" }, - { name = "debugpy", specifier = "==1.8.14" }, - { name = "decorator", specifier = "==5.2.1" }, - { name = "defusedxml", specifier = "==0.7.1" }, - { name = "dill", specifier = "==0.3.8" }, - { name = "distance", specifier = "==0.1.3" }, - { name = "docopt", specifier = "==0.6.2" }, - { name = "editdistance", specifier = "==0.8.1" }, - { name = "einops", specifier = "==0.8.1" }, - { name = "exceptiongroup", specifier = "==1.3.0" }, - { name = "executing", specifier = "==2.2.0" }, - { name = "fastjsonschema", specifier = "==2.21.1" }, - { name = "fiddle", specifier = "==0.3.0" }, - { name = "filelock", specifier = "==3.18.0" }, - { name = "fonttools", specifier = "==4.58.4" }, - { name = "fqdn", specifier = "==1.5.1" }, - { name = "frozenlist", specifier = "==1.7.0" }, - { name = "fsspec", specifier = "==2024.12.0" }, - { name = "funasr", specifier = ">=1.2.6" }, - { name = "future", specifier = "==1.0.0" }, - { name = "g2p-en", specifier = "==2.1.0" }, - { name = "gitdb", specifier = "==4.0.12" }, - { name = "gitpython", specifier = "==3.1.44" }, - { name = "graphviz", specifier = "==0.21" }, - { name = "grpcio", specifier = "==1.73.1" }, - { name = "h11", specifier = "==0.16.0" }, - { name = "hf-xet", specifier = "==1.1.5" }, - { name = "httpcore", specifier = "==1.0.9" }, - { name = "httpx", specifier = "==0.28.1" }, - { name = "huggingface-hub", specifier = "==0.33.1" }, - { name = "hydra-core", specifier = "==1.3.2" }, - { name = "idna", specifier = "==3.10" }, - { name = "inflect", specifier = "==7.5.0" }, - { name = "intervaltree", specifier = "==3.1.0" }, - { name = "ipykernel", specifier = "==6.29.5" }, - { name = "ipython", specifier = "==8.37.0" }, - { name = "ipywidgets", specifier = "==8.1.7" }, - { name = "isoduration", specifier = "==20.11.0" }, - { name = "jedi", specifier = "==0.19.2" }, - { name = "jinja2", specifier = "==3.1.6" }, - { name = "jiwer", specifier = "==4.0.0" }, - { name = "joblib", specifier = "==1.5.1" }, - { name = "json5", specifier = "==0.12.0" }, - { name = "jsonpointer", specifier = "==3.0.0" }, - { name = "jsonschema", specifier = "==4.24.0" }, - { name = "jsonschema-specifications", specifier = "==2025.4.1" }, - { name = "jupyter", specifier = "==1.1.1" }, - { name = "jupyter-client", specifier = "==8.6.3" }, - { name = "jupyter-console", specifier = "==6.6.3" }, - { name = "jupyter-core", specifier = "==5.8.1" }, - { name = "jupyter-events", specifier = "==0.12.0" }, - { name = "jupyter-lsp", specifier = "==2.2.5" }, - { name = "jupyter-server", specifier = "==2.16.0" }, - { name = "jupyter-server-terminals", specifier = "==0.5.3" }, - { name = "jupyterlab", specifier = "==4.4.4" }, - { name = "jupyterlab-pygments", specifier = "==0.3.0" }, - { name = "jupyterlab-server", specifier = "==2.27.3" }, - { name = "jupyterlab-widgets", specifier = "==3.0.15" }, - { name = "kaldi-python-io", specifier = "==1.2.2" }, - { name = "kaldiio", specifier = "==2.18.1" }, - { name = "kiwisolver", specifier = "==1.4.8" }, - { name = "lazy-loader", specifier = "==0.4" }, - { name = "levenshtein", specifier = "==0.27.1" }, - { name = "lhotse", specifier = "==1.30.3" }, - { name = "libcst", specifier = "==1.8.2" }, - { name = "librosa", specifier = "==0.11.0" }, - { name = "lightning", specifier = "==2.4.0" }, - { name = "lightning-utilities", specifier = "==0.14.3" }, - { name = "lilcom", specifier = "==1.8.1" }, - { name = "llvmlite", specifier = "==0.44.0" }, - { name = "loguru", specifier = "==0.7.3" }, - { name = "mako", specifier = "==1.3.10" }, - { name = "markdown", specifier = "==3.8.2" }, - { name = "markdown-it-py", specifier = "==3.0.0" }, - { name = "markupsafe", specifier = "==3.0.2" }, - { name = "marshmallow", specifier = "==4.0.0" }, - { name = "matplotlib", specifier = "==3.10.3" }, - { name = "matplotlib-inline", specifier = "==0.1.7" }, - { name = "mdurl", specifier = "==0.1.2" }, - { name = "mediapy", specifier = "==1.1.6" }, - { name = "mistune", specifier = "==3.1.3" }, - { name = "more-itertools", specifier = "==10.7.0" }, - { name = "mpmath", specifier = "==1.3.0" }, - { name = "msgpack", specifier = "==1.1.1" }, - { name = "multidict", specifier = "==6.6.2" }, - { name = "multiprocess", specifier = "==0.70.16" }, - { name = "nbclient", specifier = "==0.10.2" }, - { name = "nbconvert", specifier = "==7.16.6" }, - { name = "nbformat", specifier = "==5.10.4" }, - { name = "nemo-toolkit", specifier = "==2.3.1" }, - { name = "nest-asyncio", specifier = "==1.6.0" }, - { name = "networkx", specifier = "==3.4.2" }, - { name = "nltk", specifier = "==3.9.1" }, - { name = "notebook", specifier = "==7.4.3" }, - { name = "notebook-shim", specifier = "==0.2.4" }, - { name = "num2words", specifier = "==0.5.14" }, - { name = "numba", specifier = "==0.61.0" }, - { name = "numpy", specifier = "==1.26.4" }, - { name = "omegaconf", specifier = "==2.3.0" }, - { name = "onnx", specifier = "==1.17.0" }, - { name = "optuna", specifier = "==4.4.0" }, - { name = "overrides", specifier = "==7.7.0" }, - { name = "packaging", specifier = "==24.2" }, - { name = "pandas", specifier = "==2.3.0" }, - { name = "pandocfilters", specifier = "==1.5.1" }, - { name = "parso", specifier = "==0.8.4" }, - { name = "peft", specifier = "==0.15.2" }, - { name = "pexpect", specifier = "==4.9.0" }, - { name = "pillow", specifier = "==11.2.1" }, - { name = "pip", specifier = ">=25.1.1" }, - { name = "plac", specifier = "==1.4.5" }, - { name = "platformdirs", specifier = "==4.3.8" }, - { name = "pooch", specifier = "==1.8.2" }, - { name = "prometheus-client", specifier = "==0.22.1" }, - { name = "prompt-toolkit", specifier = "==3.0.51" }, - { name = "propcache", specifier = "==0.3.2" }, - { name = "psutil", specifier = "==7.0.0" }, - { name = "ptyprocess", specifier = "==0.7.0" }, - { name = "pure-eval", specifier = "==0.2.3" }, - { name = "pyaml", specifier = "==25.5.0" }, - { name = "pyannote-audio", specifier = ">=3.3.2" }, - { name = "pyannote-core", specifier = "==5.0.0" }, - { name = "pyannote-database", specifier = "==5.1.3" }, - { name = "pyannote-metrics", specifier = "==3.2.1" }, - { name = "pyarrow", specifier = "==20.0.0" }, - { name = "pybind11", specifier = "==2.13.6" }, - { name = "pycparser", specifier = "==2.22" }, - { name = "pydantic", specifier = "==2.11.7" }, - { name = "pydantic-core", specifier = "==2.33.2" }, - { name = "pydub", specifier = "==0.25.1" }, - { name = "pygments", specifier = "==2.19.2" }, - { name = "pyloudnorm", specifier = "==0.1.1" }, - { name = "pyparsing", specifier = "==3.2.3" }, - { name = "python-dateutil", specifier = "==2.9.0.post0" }, - { name = "python-json-logger", specifier = "==3.3.0" }, - { name = "pytorch-lightning", specifier = "==2.5.2" }, - { name = "pytz", specifier = "==2025.2" }, - { name = "pyyaml", specifier = "==6.0.2" }, - { name = "pyzmq", specifier = "==27.0.0" }, - { name = "rapidfuzz", specifier = "==3.13.0" }, - { name = "referencing", specifier = "==0.36.2" }, - { name = "regex", specifier = "==2024.11.6" }, - { name = "requests", specifier = "==2.32.4" }, - { name = "resampy", specifier = "==0.4.3" }, - { name = "rfc3339-validator", specifier = "==0.1.4" }, - { name = "rfc3986-validator", specifier = "==0.1.1" }, - { name = "rich", specifier = "==14.0.0" }, - { name = "rpds-py", specifier = "==0.25.1" }, - { name = "ruamel-yaml", specifier = "==0.18.14" }, - { name = "ruamel-yaml-clib", specifier = "==0.2.12" }, - { name = "sacremoses", specifier = "==0.1.1" }, - { name = "safetensors", specifier = "==0.5.3" }, - { name = "scikit-learn", specifier = "==1.5.1" }, - { name = "scipy", specifier = "==1.15.3" }, - { name = "seaborn", specifier = ">=0.13.2" }, - { name = "send2trash", specifier = "==1.8.3" }, - { name = "sentencepiece", specifier = "==0.2.0" }, - { name = "sentry-sdk", specifier = "==2.32.0" }, - { name = "setproctitle", specifier = "==1.3.6" }, - { name = "shellingham", specifier = "==1.5.4" }, - { name = "six", specifier = "==1.17.0" }, - { name = "smmap", specifier = "==5.0.2" }, - { name = "sniffio", specifier = "==1.3.1" }, - { name = "sortedcontainers", specifier = "==2.4.0" }, - { name = "soundfile", specifier = "==0.13.1" }, - { name = "soupsieve", specifier = "==2.7" }, - { name = "sox", specifier = "==1.5.0" }, - { name = "soxr", specifier = "==0.5.0.post1" }, - { name = "sqlalchemy", specifier = "==2.0.41" }, - { name = "stack-data", specifier = "==0.6.3" }, - { name = "tabulate", specifier = "==0.9.0" }, - { name = "tensorboard", specifier = "==2.19.0" }, - { name = "tensorboard-data-server", specifier = "==0.7.2" }, - { name = "termcolor", specifier = "==3.1.0" }, - { name = "terminado", specifier = "==0.18.1" }, - { name = "text-unidecode", specifier = "==1.3" }, - { name = "texterrors", specifier = "==0.5.1" }, - { name = "threadpoolctl", specifier = "==3.6.0" }, - { name = "tinycss2", specifier = "==1.4.0" }, - { name = "tokenizers", specifier = "==0.21.2" }, - { name = "tomli", specifier = "==2.2.1" }, - { name = "toolz", specifier = "==1.0.0" }, - { name = "torch", specifier = "==2.7.0" }, - { name = "torchmetrics", specifier = "==1.7.3" }, - { name = "tornado", specifier = "==6.5.1" }, - { name = "tqdm", specifier = "==4.67.1" }, - { name = "traitlets", specifier = "==5.14.3" }, - { name = "transformers", specifier = "==4.51.3" }, - { name = "typeguard", specifier = "==4.4.4" }, - { name = "typer", specifier = "==0.16.0" }, - { name = "types-python-dateutil", specifier = "==2.9.0.20250516" }, - { name = "typing-extensions", specifier = "==4.14.0" }, - { name = "typing-inspection", specifier = "==0.4.1" }, - { name = "tzdata", specifier = "==2025.2" }, - { name = "uri-template", specifier = "==1.3.0" }, - { name = "urllib3", specifier = "==2.5.0" }, - { name = "wandb", specifier = "==0.20.1" }, - { name = "wcwidth", specifier = "==0.2.13" }, - { name = "webcolors", specifier = "==24.11.1" }, - { name = "webdataset", specifier = "==1.0.2" }, - { name = "webencodings", specifier = "==0.5.1" }, - { name = "websocket-client", specifier = "==1.8.0" }, - { name = "werkzeug", specifier = "==3.1.3" }, - { name = "wget", specifier = "==3.2" }, - { name = "widgetsnbextension", specifier = "==4.0.14" }, - { name = "wrapt", specifier = "==1.17.2" }, - { name = "xxhash", specifier = "==3.5.0" }, - { name = "yarl", specifier = "==1.20.1" }, -] - -[[package]] -name = "parso" -version = "0.8.4" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/66/94/68e2e17afaa9169cf6412ab0f28623903be73d1b32e208d9e8e541bb086d/parso-0.8.4.tar.gz", hash = "sha256:eb3a7b58240fb99099a345571deecc0f9540ea5f4dd2fe14c2a99d6b281ab92d", size = 400609 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/c6/ac/dac4a63f978e4dcb3c6d3a78c4d8e0192a113d288502a1216950c41b1027/parso-0.8.4-py2.py3-none-any.whl", hash = "sha256:a418670a20291dacd2dddc80c377c5c3791378ee1e8d12bffc35420643d43f18", size = 103650 }, -] - -[[package]] -name = "peft" -version = "0.15.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "accelerate" }, - { name = "huggingface-hub" }, - { name = "numpy" }, - { name = "packaging" }, - { name = "psutil" }, - { name = "pyyaml" }, - { name = "safetensors" }, - { name = "torch" }, - { name = "tqdm" }, - { name = "transformers" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/33/65/faa18cd8ffbe0f742c3f2559770646cce2574b9cd28a2a05e8d36f64e968/peft-0.15.2.tar.gz", hash = "sha256:7059029f4d42a092ded1aa117dd366a46084aef638bdd593f6ab0195d5427fcd", size = 472952 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/68/85/8e6ea3d1089f2b6de3c1cd34bbbd7560912af9d34b057be3b8b8fefe1da3/peft-0.15.2-py3-none-any.whl", hash = "sha256:0dfc942b03b7af4b7267cd4e30b15e3a4a1d277adc581ce6245fc13f1f93d0a0", size = 411051 }, -] - -[[package]] -name = "pexpect" -version = "4.9.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "ptyprocess" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/42/92/cc564bf6381ff43ce1f4d06852fc19a2f11d180f23dc32d9588bee2f149d/pexpect-4.9.0.tar.gz", hash = "sha256:ee7d41123f3c9911050ea2c2dac107568dc43b2d3b0c7557a33212c398ead30f", size = 166450 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/9e/c3/059298687310d527a58bb01f3b1965787ee3b40dce76752eda8b44e9a2c5/pexpect-4.9.0-py2.py3-none-any.whl", hash = "sha256:7236d1e080e4936be2dc3e326cec0af72acf9212a7e1d060210e70a47e253523", size = 63772 }, -] - -[[package]] -name = "pillow" -version = "11.2.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/af/cb/bb5c01fcd2a69335b86c22142b2bccfc3464087efb7fd382eee5ffc7fdf7/pillow-11.2.1.tar.gz", hash = "sha256:a64dd61998416367b7ef979b73d3a85853ba9bec4c2925f74e588879a58716b6", size = 47026707 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/0d/8b/b158ad57ed44d3cc54db8d68ad7c0a58b8fc0e4c7a3f995f9d62d5b464a1/pillow-11.2.1-cp310-cp310-macosx_10_10_x86_64.whl", hash = "sha256:d57a75d53922fc20c165016a20d9c44f73305e67c351bbc60d1adaf662e74047", size = 3198442 }, - { url = "https://files.pythonhosted.org/packages/b1/f8/bb5d956142f86c2d6cc36704943fa761f2d2e4c48b7436fd0a85c20f1713/pillow-11.2.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:127bf6ac4a5b58b3d32fc8289656f77f80567d65660bc46f72c0d77e6600cc95", size = 3030553 }, - { url = "https://files.pythonhosted.org/packages/22/7f/0e413bb3e2aa797b9ca2c5c38cb2e2e45d88654e5b12da91ad446964cfae/pillow-11.2.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:b4ba4be812c7a40280629e55ae0b14a0aafa150dd6451297562e1764808bbe61", size = 4405503 }, - { url = "https://files.pythonhosted.org/packages/f3/b4/cc647f4d13f3eb837d3065824aa58b9bcf10821f029dc79955ee43f793bd/pillow-11.2.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c8bd62331e5032bc396a93609982a9ab6b411c05078a52f5fe3cc59234a3abd1", size = 4490648 }, - { url = "https://files.pythonhosted.org/packages/c2/6f/240b772a3b35cdd7384166461567aa6713799b4e78d180c555bd284844ea/pillow-11.2.1-cp310-cp310-manylinux_2_28_aarch64.whl", hash = "sha256:562d11134c97a62fe3af29581f083033179f7ff435f78392565a1ad2d1c2c45c", size = 4508937 }, - { url = "https://files.pythonhosted.org/packages/f3/5e/7ca9c815ade5fdca18853db86d812f2f188212792780208bdb37a0a6aef4/pillow-11.2.1-cp310-cp310-manylinux_2_28_x86_64.whl", hash = "sha256:c97209e85b5be259994eb5b69ff50c5d20cca0f458ef9abd835e262d9d88b39d", size = 4599802 }, - { url = "https://files.pythonhosted.org/packages/02/81/c3d9d38ce0c4878a77245d4cf2c46d45a4ad0f93000227910a46caff52f3/pillow-11.2.1-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:0c3e6d0f59171dfa2e25d7116217543310908dfa2770aa64b8f87605f8cacc97", size = 4576717 }, - { url = "https://files.pythonhosted.org/packages/42/49/52b719b89ac7da3185b8d29c94d0e6aec8140059e3d8adcaa46da3751180/pillow-11.2.1-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:cc1c3bc53befb6096b84165956e886b1729634a799e9d6329a0c512ab651e579", size = 4654874 }, - { url = "https://files.pythonhosted.org/packages/5b/0b/ede75063ba6023798267023dc0d0401f13695d228194d2242d5a7ba2f964/pillow-11.2.1-cp310-cp310-win32.whl", hash = "sha256:312c77b7f07ab2139924d2639860e084ec2a13e72af54d4f08ac843a5fc9c79d", size = 2331717 }, - { url = "https://files.pythonhosted.org/packages/ed/3c/9831da3edea527c2ed9a09f31a2c04e77cd705847f13b69ca60269eec370/pillow-11.2.1-cp310-cp310-win_amd64.whl", hash = "sha256:9bc7ae48b8057a611e5fe9f853baa88093b9a76303937449397899385da06fad", size = 2676204 }, - { url = "https://files.pythonhosted.org/packages/01/97/1f66ff8a1503d8cbfc5bae4dc99d54c6ec1e22ad2b946241365320caabc2/pillow-11.2.1-cp310-cp310-win_arm64.whl", hash = "sha256:2728567e249cdd939f6cc3d1f049595c66e4187f3c34078cbc0a7d21c47482d2", size = 2414767 }, - { url = "https://files.pythonhosted.org/packages/33/49/c8c21e4255b4f4a2c0c68ac18125d7f5460b109acc6dfdef1a24f9b960ef/pillow-11.2.1-pp310-pypy310_pp73-macosx_10_15_x86_64.whl", hash = "sha256:9b7b0d4fd2635f54ad82785d56bc0d94f147096493a79985d0ab57aedd563156", size = 3181727 }, - { url = "https://files.pythonhosted.org/packages/6d/f1/f7255c0838f8c1ef6d55b625cfb286835c17e8136ce4351c5577d02c443b/pillow-11.2.1-pp310-pypy310_pp73-macosx_11_0_arm64.whl", hash = "sha256:aa442755e31c64037aa7c1cb186e0b369f8416c567381852c63444dd666fb772", size = 2999833 }, - { url = "https://files.pythonhosted.org/packages/e2/57/9968114457bd131063da98d87790d080366218f64fa2943b65ac6739abb3/pillow-11.2.1-pp310-pypy310_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:f0d3348c95b766f54b76116d53d4cb171b52992a1027e7ca50c81b43b9d9e363", size = 3437472 }, - { url = "https://files.pythonhosted.org/packages/b2/1b/e35d8a158e21372ecc48aac9c453518cfe23907bb82f950d6e1c72811eb0/pillow-11.2.1-pp310-pypy310_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:85d27ea4c889342f7e35f6d56e7e1cb345632ad592e8c51b693d7b7556043ce0", size = 3459976 }, - { url = "https://files.pythonhosted.org/packages/26/da/2c11d03b765efff0ccc473f1c4186dc2770110464f2177efaed9cf6fae01/pillow-11.2.1-pp310-pypy310_pp73-manylinux_2_28_aarch64.whl", hash = "sha256:bf2c33d6791c598142f00c9c4c7d47f6476731c31081331664eb26d6ab583e01", size = 3527133 }, - { url = "https://files.pythonhosted.org/packages/79/1a/4e85bd7cadf78412c2a3069249a09c32ef3323650fd3005c97cca7aa21df/pillow-11.2.1-pp310-pypy310_pp73-manylinux_2_28_x86_64.whl", hash = "sha256:e616e7154c37669fc1dfc14584f11e284e05d1c650e1c0f972f281c4ccc53193", size = 3571555 }, - { url = "https://files.pythonhosted.org/packages/69/03/239939915216de1e95e0ce2334bf17a7870ae185eb390fab6d706aadbfc0/pillow-11.2.1-pp310-pypy310_pp73-win_amd64.whl", hash = "sha256:39ad2e0f424394e3aebc40168845fee52df1394a4673a6ee512d840d14ab3013", size = 2674713 }, -] - -[[package]] -name = "pip" -version = "25.2" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/20/16/650289cd3f43d5a2fadfd98c68bd1e1e7f2550a1a5326768cddfbcedb2c5/pip-25.2.tar.gz", hash = "sha256:578283f006390f85bb6282dffb876454593d637f5d1be494b5202ce4877e71f2", size = 1840021 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/b7/3f/945ef7ab14dc4f9d7f40288d2df998d1837ee0888ec3659c813487572faa/pip-25.2-py3-none-any.whl", hash = "sha256:6d67a2b4e7f14d8b31b8b52648866fa717f45a1eb70e83002f4331d07e953717", size = 1752557 }, -] - -[[package]] -name = "plac" -version = "1.4.5" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/23/09/26ef2d614cabdcc52a7f383d0dc7967bf46be3c9700898c594e37b710c3d/plac-1.4.5.tar.gz", hash = "sha256:5f05bf85235c017fcd76c73c8101d4ff8e96beb3dc58b9a37de49cac7de82d14", size = 38988 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/15/36/38676114a0dbee137ec366daa86603d667a07e9a52667d5ebf5c580100ba/plac-1.4.5-py2.py3-none-any.whl", hash = "sha256:87187786b4e446688b1cf5112e18fed8a23ab3b316c25fe91266a10bd1736b16", size = 22468 }, -] - -[[package]] -name = "platformdirs" -version = "4.3.8" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/fe/8b/3c73abc9c759ecd3f1f7ceff6685840859e8070c4d947c93fae71f6a0bf2/platformdirs-4.3.8.tar.gz", hash = "sha256:3d512d96e16bcb959a814c9f348431070822a6496326a4be0911c40b5a74c2bc", size = 21362 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/fe/39/979e8e21520d4e47a0bbe349e2713c0aac6f3d853d0e5b34d76206c439aa/platformdirs-4.3.8-py3-none-any.whl", hash = "sha256:ff7059bb7eb1179e2685604f4aaf157cfd9535242bd23742eadc3c13542139b4", size = 18567 }, -] - -[[package]] -name = "pooch" -version = "1.8.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "packaging" }, - { name = "platformdirs" }, - { name = "requests" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/c6/77/b3d3e00c696c16cf99af81ef7b1f5fe73bd2a307abca41bd7605429fe6e5/pooch-1.8.2.tar.gz", hash = "sha256:76561f0de68a01da4df6af38e9955c4c9d1a5c90da73f7e40276a5728ec83d10", size = 59353 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/a8/87/77cc11c7a9ea9fd05503def69e3d18605852cd0d4b0d3b8f15bbeb3ef1d1/pooch-1.8.2-py3-none-any.whl", hash = "sha256:3529a57096f7198778a5ceefd5ac3ef0e4d06a6ddaf9fc2d609b806f25302c47", size = 64574 }, -] - -[[package]] -name = "primepy" -version = "1.3" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/35/77/0cfa1b4697cfb5336f3a96e8bc73327f64610be3a64c97275f1801afb395/primePy-1.3.tar.gz", hash = "sha256:25fd7e25344b0789a5984c75d89f054fcf1f180bef20c998e4befbac92de4669", size = 3914 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/74/c1/bb7e334135859c3a92ec399bc89293ea73f28e815e35b43929c8db6af030/primePy-1.3-py3-none-any.whl", hash = "sha256:5ed443718765be9bf7e2ff4c56cdff71b42140a15b39d054f9d99f0009e2317a", size = 4040 }, -] - -[[package]] -name = "prometheus-client" -version = "0.22.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/5e/cf/40dde0a2be27cc1eb41e333d1a674a74ce8b8b0457269cc640fd42b07cf7/prometheus_client-0.22.1.tar.gz", hash = "sha256:190f1331e783cf21eb60bca559354e0a4d4378facecf78f5428c39b675d20d28", size = 69746 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/32/ae/ec06af4fe3ee72d16973474f122541746196aaa16cea6f66d18b963c6177/prometheus_client-0.22.1-py3-none-any.whl", hash = "sha256:cca895342e308174341b2cbf99a56bef291fbc0ef7b9e5412a0f26d653ba7094", size = 58694 }, -] - -[[package]] -name = "prompt-toolkit" -version = "3.0.51" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "wcwidth" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/bb/6e/9d084c929dfe9e3bfe0c6a47e31f78a25c54627d64a66e884a8bf5474f1c/prompt_toolkit-3.0.51.tar.gz", hash = "sha256:931a162e3b27fc90c86f1b48bb1fb2c528c2761475e57c9c06de13311c7b54ed", size = 428940 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/ce/4f/5249960887b1fbe561d9ff265496d170b55a735b76724f10ef19f9e40716/prompt_toolkit-3.0.51-py3-none-any.whl", hash = "sha256:52742911fde84e2d423e2f9a4cf1de7d7ac4e51958f648d9540e0fb8db077b07", size = 387810 }, -] - -[[package]] -name = "propcache" -version = "0.3.2" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/a6/16/43264e4a779dd8588c21a70f0709665ee8f611211bdd2c87d952cfa7c776/propcache-0.3.2.tar.gz", hash = "sha256:20d7d62e4e7ef05f221e0db2856b979540686342e7dd9973b815599c7057e168", size = 44139 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/ab/14/510deed325e262afeb8b360043c5d7c960da7d3ecd6d6f9496c9c56dc7f4/propcache-0.3.2-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:22d9962a358aedbb7a2e36187ff273adeaab9743373a272976d2e348d08c7770", size = 73178 }, - { url = "https://files.pythonhosted.org/packages/cd/4e/ad52a7925ff01c1325653a730c7ec3175a23f948f08626a534133427dcff/propcache-0.3.2-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:0d0fda578d1dc3f77b6b5a5dce3b9ad69a8250a891760a548df850a5e8da87f3", size = 43133 }, - { url = "https://files.pythonhosted.org/packages/63/7c/e9399ba5da7780871db4eac178e9c2e204c23dd3e7d32df202092a1ed400/propcache-0.3.2-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:3def3da3ac3ce41562d85db655d18ebac740cb3fa4367f11a52b3da9d03a5cc3", size = 43039 }, - { url = "https://files.pythonhosted.org/packages/22/e1/58da211eb8fdc6fc854002387d38f415a6ca5f5c67c1315b204a5d3e9d7a/propcache-0.3.2-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:9bec58347a5a6cebf239daba9bda37dffec5b8d2ce004d9fe4edef3d2815137e", size = 201903 }, - { url = "https://files.pythonhosted.org/packages/c4/0a/550ea0f52aac455cb90111c8bab995208443e46d925e51e2f6ebdf869525/propcache-0.3.2-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:55ffda449a507e9fbd4aca1a7d9aa6753b07d6166140e5a18d2ac9bc49eac220", size = 213362 }, - { url = "https://files.pythonhosted.org/packages/5a/af/9893b7d878deda9bb69fcf54600b247fba7317761b7db11fede6e0f28bd0/propcache-0.3.2-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:64a67fb39229a8a8491dd42f864e5e263155e729c2e7ff723d6e25f596b1e8cb", size = 210525 }, - { url = "https://files.pythonhosted.org/packages/7c/bb/38fd08b278ca85cde36d848091ad2b45954bc5f15cce494bb300b9285831/propcache-0.3.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9da1cf97b92b51253d5b68cf5a2b9e0dafca095e36b7f2da335e27dc6172a614", size = 198283 }, - { url = "https://files.pythonhosted.org/packages/78/8c/9fe55bd01d362bafb413dfe508c48753111a1e269737fa143ba85693592c/propcache-0.3.2-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:5f559e127134b07425134b4065be45b166183fdcb433cb6c24c8e4149056ad50", size = 191872 }, - { url = "https://files.pythonhosted.org/packages/54/14/4701c33852937a22584e08abb531d654c8bcf7948a8f87ad0a4822394147/propcache-0.3.2-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:aff2e4e06435d61f11a428360a932138d0ec288b0a31dd9bd78d200bd4a2b339", size = 199452 }, - { url = "https://files.pythonhosted.org/packages/16/44/447f2253d859602095356007657ee535e0093215ea0b3d1d6a41d16e5201/propcache-0.3.2-cp310-cp310-musllinux_1_2_armv7l.whl", hash = "sha256:4927842833830942a5d0a56e6f4839bc484785b8e1ce8d287359794818633ba0", size = 191567 }, - { url = "https://files.pythonhosted.org/packages/f2/b3/e4756258749bb2d3b46defcff606a2f47410bab82be5824a67e84015b267/propcache-0.3.2-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:6107ddd08b02654a30fb8ad7a132021759d750a82578b94cd55ee2772b6ebea2", size = 193015 }, - { url = "https://files.pythonhosted.org/packages/1e/df/e6d3c7574233164b6330b9fd697beeac402afd367280e6dc377bb99b43d9/propcache-0.3.2-cp310-cp310-musllinux_1_2_ppc64le.whl", hash = "sha256:70bd8b9cd6b519e12859c99f3fc9a93f375ebd22a50296c3a295028bea73b9e7", size = 204660 }, - { url = "https://files.pythonhosted.org/packages/b2/53/e4d31dd5170b4a0e2e6b730f2385a96410633b4833dc25fe5dffd1f73294/propcache-0.3.2-cp310-cp310-musllinux_1_2_s390x.whl", hash = "sha256:2183111651d710d3097338dd1893fcf09c9f54e27ff1a8795495a16a469cc90b", size = 206105 }, - { url = "https://files.pythonhosted.org/packages/7f/fe/74d54cf9fbe2a20ff786e5f7afcfde446588f0cf15fb2daacfbc267b866c/propcache-0.3.2-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:fb075ad271405dcad8e2a7ffc9a750a3bf70e533bd86e89f0603e607b93aa64c", size = 196980 }, - { url = "https://files.pythonhosted.org/packages/22/ec/c469c9d59dada8a7679625e0440b544fe72e99311a4679c279562051f6fc/propcache-0.3.2-cp310-cp310-win32.whl", hash = "sha256:404d70768080d3d3bdb41d0771037da19d8340d50b08e104ca0e7f9ce55fce70", size = 37679 }, - { url = "https://files.pythonhosted.org/packages/38/35/07a471371ac89d418f8d0b699c75ea6dca2041fbda360823de21f6a9ce0a/propcache-0.3.2-cp310-cp310-win_amd64.whl", hash = "sha256:7435d766f978b4ede777002e6b3b6641dd229cd1da8d3d3106a45770365f9ad9", size = 41459 }, - { url = "https://files.pythonhosted.org/packages/cc/35/cc0aaecf278bb4575b8555f2b137de5ab821595ddae9da9d3cd1da4072c7/propcache-0.3.2-py3-none-any.whl", hash = "sha256:98f1ec44fb675f5052cccc8e609c46ed23a35a1cfd18545ad4e29002d858a43f", size = 12663 }, -] - -[[package]] -name = "protobuf" -version = "4.24.4" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/52/5c/f2c0778278259089952f94b0884ca27a001a17ffbd992ebe30c841085f4c/protobuf-4.24.4.tar.gz", hash = "sha256:5a70731910cd9104762161719c3d883c960151eea077134458503723b60e3667", size = 383850 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/00/0f/3dc2f86e9c3d6e73c56d915a3563ecc96ebee8144fb39614f0d6c1fb023d/protobuf-4.24.4-cp310-abi3-win32.whl", hash = "sha256:ec9912d5cb6714a5710e28e592ee1093d68c5ebfeda61983b3f40331da0b1ebb", size = 409968 }, - { url = "https://files.pythonhosted.org/packages/c2/59/f89c04923d68595d359f4cd7adbbdf5e5d791257945f8873d88b2fd1f979/protobuf-4.24.4-cp310-abi3-win_amd64.whl", hash = "sha256:1badab72aa8a3a2b812eacfede5020472e16c6b2212d737cefd685884c191085", size = 430493 }, - { url = "https://files.pythonhosted.org/packages/88/12/efb5896c901382548ecb58d0449885a8f9aa62bb559d65e5a8a47f122629/protobuf-4.24.4-cp37-abi3-macosx_10_9_universal2.whl", hash = "sha256:8e61a27f362369c2f33248a0ff6896c20dcd47b5d48239cb9720134bef6082e4", size = 409417 }, - { url = "https://files.pythonhosted.org/packages/db/61/9c7b481771fe4702fb3be1152812fecec9b06f9c36d523ad52b98cb46800/protobuf-4.24.4-cp37-abi3-manylinux2014_aarch64.whl", hash = "sha256:bffa46ad9612e6779d0e51ae586fde768339b791a50610d85eb162daeb23661e", size = 310587 }, - { url = "https://files.pythonhosted.org/packages/c8/2c/03046cac73f46bfe98fc846ef629cf4f84c2f59258216aa2cc0d22bfca8f/protobuf-4.24.4-cp37-abi3-manylinux2014_x86_64.whl", hash = "sha256:b493cb590960ff863743b9ff1452c413c2ee12b782f48beca77c8da3e2ffe9d9", size = 311559 }, - { url = "https://files.pythonhosted.org/packages/e5/a7/bb962b8b981dd890a44a34d0e922b76c32e5db443ff9f9b9ce6149069070/protobuf-4.24.4-py3-none-any.whl", hash = "sha256:80797ce7424f8c8d2f2547e2d42bfbb6c08230ce5832d6c099a37335c9c90a92", size = 175677 }, -] - -[[package]] -name = "psutil" -version = "7.0.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/2a/80/336820c1ad9286a4ded7e845b2eccfcb27851ab8ac6abece774a6ff4d3de/psutil-7.0.0.tar.gz", hash = "sha256:7be9c3eba38beccb6495ea33afd982a44074b78f28c434a1f51cc07fd315c456", size = 497003 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/ed/e6/2d26234410f8b8abdbf891c9da62bee396583f713fb9f3325a4760875d22/psutil-7.0.0-cp36-abi3-macosx_10_9_x86_64.whl", hash = "sha256:101d71dc322e3cffd7cea0650b09b3d08b8e7c4109dd6809fe452dfd00e58b25", size = 238051 }, - { url = "https://files.pythonhosted.org/packages/04/8b/30f930733afe425e3cbfc0e1468a30a18942350c1a8816acfade80c005c4/psutil-7.0.0-cp36-abi3-macosx_11_0_arm64.whl", hash = "sha256:39db632f6bb862eeccf56660871433e111b6ea58f2caea825571951d4b6aa3da", size = 239535 }, - { url = "https://files.pythonhosted.org/packages/2a/ed/d362e84620dd22876b55389248e522338ed1bf134a5edd3b8231d7207f6d/psutil-7.0.0-cp36-abi3-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:1fcee592b4c6f146991ca55919ea3d1f8926497a713ed7faaf8225e174581e91", size = 275004 }, - { url = "https://files.pythonhosted.org/packages/bf/b9/b0eb3f3cbcb734d930fdf839431606844a825b23eaf9a6ab371edac8162c/psutil-7.0.0-cp36-abi3-manylinux_2_12_x86_64.manylinux2010_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4b1388a4f6875d7e2aff5c4ca1cc16c545ed41dd8bb596cefea80111db353a34", size = 277986 }, - { url = "https://files.pythonhosted.org/packages/eb/a2/709e0fe2f093556c17fbafda93ac032257242cabcc7ff3369e2cb76a97aa/psutil-7.0.0-cp36-abi3-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a5f098451abc2828f7dc6b58d44b532b22f2088f4999a937557b603ce72b1993", size = 279544 }, - { url = "https://files.pythonhosted.org/packages/50/e6/eecf58810b9d12e6427369784efe814a1eec0f492084ce8eb8f4d89d6d61/psutil-7.0.0-cp37-abi3-win32.whl", hash = "sha256:ba3fcef7523064a6c9da440fc4d6bd07da93ac726b5733c29027d7dc95b39d99", size = 241053 }, - { url = "https://files.pythonhosted.org/packages/50/1b/6921afe68c74868b4c9fa424dad3be35b095e16687989ebbb50ce4fceb7c/psutil-7.0.0-cp37-abi3-win_amd64.whl", hash = "sha256:4cf3d4eb1aa9b348dec30105c55cd9b7d4629285735a102beb4441e38db90553", size = 244885 }, -] - -[[package]] -name = "ptyprocess" -version = "0.7.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/20/e5/16ff212c1e452235a90aeb09066144d0c5a6a8c0834397e03f5224495c4e/ptyprocess-0.7.0.tar.gz", hash = "sha256:5c5d0a3b48ceee0b48485e0c26037c0acd7d29765ca3fbb5cb3831d347423220", size = 70762 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/22/a6/858897256d0deac81a172289110f31629fc4cee19b6f01283303e18c8db3/ptyprocess-0.7.0-py2.py3-none-any.whl", hash = "sha256:4b41f3967fce3af57cc7e94b888626c18bf37a083e3651ca8feeb66d492fef35", size = 13993 }, -] - -[[package]] -name = "pure-eval" -version = "0.2.3" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/cd/05/0a34433a064256a578f1783a10da6df098ceaa4a57bbeaa96a6c0352786b/pure_eval-0.2.3.tar.gz", hash = "sha256:5f4e983f40564c576c7c8635ae88db5956bb2229d7e9237d03b3c0b0190eaf42", size = 19752 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/8e/37/efad0257dc6e593a18957422533ff0f87ede7c9c6ea010a2177d738fb82f/pure_eval-0.2.3-py3-none-any.whl", hash = "sha256:1db8e35b67b3d218d818ae653e27f06c3aa420901fa7b081ca98cbedc874e0d0", size = 11842 }, -] - -[[package]] -name = "pyaml" -version = "25.5.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "pyyaml" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/c1/40/94f10f32ab952c5cca713d9ac9d8b2fdc37392d90eea403823eeac674c24/pyaml-25.5.0.tar.gz", hash = "sha256:5799560c7b1c9daf35a7a4535f53e2c30323f74cbd7cb4f2e715b16dd681a58a", size = 29812 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/aa/7d/1b5061beff826f902285827261485a058b943332eba8a5532a0164735205/pyaml-25.5.0-py3-none-any.whl", hash = "sha256:b9e0c4e58a5e8003f8f18e802db49fd0563ada587209b13e429bdcbefa87d035", size = 26422 }, -] - -[[package]] -name = "pyannote-audio" -version = "3.4.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "asteroid-filterbanks" }, - { name = "einops" }, - { name = "huggingface-hub" }, - { name = "lightning" }, - { name = "omegaconf" }, - { name = "pyannote-core" }, - { name = "pyannote-database" }, - { name = "pyannote-metrics" }, - { name = "pyannote-pipeline" }, - { name = "pytorch-metric-learning" }, - { name = "rich" }, - { name = "semver" }, - { name = "soundfile" }, - { name = "speechbrain" }, - { name = "tensorboardx" }, - { name = "torch" }, - { name = "torch-audiomentations" }, - { name = "torchaudio" }, - { name = "torchmetrics" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/ec/1e/efe9619c38f1281ddf21640654d8ea9e3f67c459b76f78657b26d8557bbe/pyannote_audio-3.4.0.tar.gz", hash = "sha256:d523d883cb8d37cb6daf99f3ba83f9138bb193646ad71e6eae7deb89d8ddd642", size = 804850 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/79/13/620c6f711b723653092fd063bfee82a6af5ea3a4d3c42efc53ce623a7f4d/pyannote_audio-3.4.0-py2.py3-none-any.whl", hash = "sha256:36e38f058059f46da3478dda581cda53d9d85a21173a3e70bbdbc3ba93b5e1b7", size = 897789 }, -] - -[[package]] -name = "pyannote-core" -version = "5.0.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "numpy" }, - { name = "scipy" }, - { name = "sortedcontainers" }, - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/65/03/feaf7534206f02c75baf151ce4b8c322b402a6f477c2be82f69d9269cbe6/pyannote.core-5.0.0.tar.gz", hash = "sha256:1a55bcc8bd680ba6be5fa53efa3b6f3d2cdd67144c07b6b4d8d66d5cb0d2096f", size = 59247 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/84/c4/370bc8ba66815a5832ece753a1009388bb07ea353d21c83f2d5a1a436f2c/pyannote.core-5.0.0-py3-none-any.whl", hash = "sha256:04920a6754492242ce0dc6017545595ab643870fe69a994f20c1a5f2da0544d0", size = 58475 }, -] - -[[package]] -name = "pyannote-database" -version = "5.1.3" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "pandas" }, - { name = "pyannote-core" }, - { name = "pyyaml" }, - { name = "typer" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/a9/ae/de36413d69a46be87cb612ebbcdc4eacbeebce3bc809124603e44a88fe26/pyannote.database-5.1.3.tar.gz", hash = "sha256:0eaf64c1cc506718de60d2d702f1359b1ae7ff252ee3e4799f1c5e378cd52c31", size = 49957 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/a1/64/92d51a3a05615ba58be8ba62a43f9f9f952d9f3646f7e4fb7826e5a3a24e/pyannote.database-5.1.3-py3-none-any.whl", hash = "sha256:37887844c7dfbcc075cb591eddc00aff45fae1ed905344e1f43e0090e63bd40a", size = 48127 }, -] - -[[package]] -name = "pyannote-metrics" -version = "3.2.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "docopt" }, - { name = "matplotlib" }, - { name = "numpy" }, - { name = "pandas" }, - { name = "pyannote-core" }, - { name = "pyannote-database" }, - { name = "scikit-learn" }, - { name = "scipy" }, - { name = "sympy" }, - { name = "tabulate" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/39/2b/6c5f01d3c49aa1c160765946e23782ca6436ae8b9bc514b56319ff5f16e7/pyannote.metrics-3.2.1.tar.gz", hash = "sha256:08024255a3550e96a8e9da4f5f4af326886548480de891414567c8900920ee5c", size = 49086 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/6c/7d/035b370ab834b30e849fe9cd092b7bd7f321fcc4a2c56b84e96476b7ede5/pyannote.metrics-3.2.1-py3-none-any.whl", hash = "sha256:46be797cdade26c82773e5018659ae610145260069c7c5bf3d3c8a029ade8e22", size = 51386 }, -] - -[[package]] -name = "pyannote-pipeline" -version = "3.0.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "docopt" }, - { name = "filelock" }, - { name = "optuna" }, - { name = "pyannote-core" }, - { name = "pyannote-database" }, - { name = "pyyaml" }, - { name = "scikit-learn" }, - { name = "tqdm" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/35/04/4bcfe0dd588577a188328b806f3a7213d8cead0ce5fe5784d01fd57df93f/pyannote.pipeline-3.0.1.tar.gz", hash = "sha256:021794e26a2cf5d8fb5bb1835951e71f5fac33eb14e23dfb7468e16b1b805151", size = 34486 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/83/42/1bf7cbf061ed05c580bfb63bffdd3f3474cbd5c02bee4fac518eea9e9d9e/pyannote.pipeline-3.0.1-py3-none-any.whl", hash = "sha256:819bde4c4dd514f740f2373dfec794832b9fc8e346a35e43a7681625ee187393", size = 31517 }, -] - -[[package]] -name = "pyarrow" -version = "20.0.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/a2/ee/a7810cb9f3d6e9238e61d312076a9859bf3668fd21c69744de9532383912/pyarrow-20.0.0.tar.gz", hash = "sha256:febc4a913592573c8d5805091a6c2b5064c8bd6e002131f01061797d91c783c1", size = 1125187 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/5b/23/77094eb8ee0dbe88441689cb6afc40ac312a1e15d3a7acc0586999518222/pyarrow-20.0.0-cp310-cp310-macosx_12_0_arm64.whl", hash = "sha256:c7dd06fd7d7b410ca5dc839cc9d485d2bc4ae5240851bcd45d85105cc90a47d7", size = 30832591 }, - { url = "https://files.pythonhosted.org/packages/c3/d5/48cc573aff00d62913701d9fac478518f693b30c25f2c157550b0b2565cb/pyarrow-20.0.0-cp310-cp310-macosx_12_0_x86_64.whl", hash = "sha256:d5382de8dc34c943249b01c19110783d0d64b207167c728461add1ecc2db88e4", size = 32273686 }, - { url = "https://files.pythonhosted.org/packages/37/df/4099b69a432b5cb412dd18adc2629975544d656df3d7fda6d73c5dba935d/pyarrow-20.0.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:6415a0d0174487456ddc9beaead703d0ded5966129fa4fd3114d76b5d1c5ceae", size = 41337051 }, - { url = "https://files.pythonhosted.org/packages/4c/27/99922a9ac1c9226f346e3a1e15e63dee6f623ed757ff2893f9d6994a69d3/pyarrow-20.0.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:15aa1b3b2587e74328a730457068dc6c89e6dcbf438d4369f572af9d320a25ee", size = 42404659 }, - { url = "https://files.pythonhosted.org/packages/21/d1/71d91b2791b829c9e98f1e0d85be66ed93aff399f80abb99678511847eaa/pyarrow-20.0.0-cp310-cp310-manylinux_2_28_aarch64.whl", hash = "sha256:5605919fbe67a7948c1f03b9f3727d82846c053cd2ce9303ace791855923fd20", size = 40695446 }, - { url = "https://files.pythonhosted.org/packages/f1/ca/ae10fba419a6e94329707487835ec721f5a95f3ac9168500bcf7aa3813c7/pyarrow-20.0.0-cp310-cp310-manylinux_2_28_x86_64.whl", hash = "sha256:a5704f29a74b81673d266e5ec1fe376f060627c2e42c5c7651288ed4b0db29e9", size = 42278528 }, - { url = "https://files.pythonhosted.org/packages/7a/a6/aba40a2bf01b5d00cf9cd16d427a5da1fad0fb69b514ce8c8292ab80e968/pyarrow-20.0.0-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:00138f79ee1b5aca81e2bdedb91e3739b987245e11fa3c826f9e57c5d102fb75", size = 42918162 }, - { url = "https://files.pythonhosted.org/packages/93/6b/98b39650cd64f32bf2ec6d627a9bd24fcb3e4e6ea1873c5e1ea8a83b1a18/pyarrow-20.0.0-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:f2d67ac28f57a362f1a2c1e6fa98bfe2f03230f7e15927aecd067433b1e70ce8", size = 44550319 }, - { url = "https://files.pythonhosted.org/packages/ab/32/340238be1eb5037e7b5de7e640ee22334417239bc347eadefaf8c373936d/pyarrow-20.0.0-cp310-cp310-win_amd64.whl", hash = "sha256:4a8b029a07956b8d7bd742ffca25374dd3f634b35e46cc7a7c3fa4c75b297191", size = 25770759 }, -] - -[[package]] -name = "pybind11" -version = "2.13.6" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/d2/c1/72b9622fcb32ff98b054f724e213c7f70d6898baa714f4516288456ceaba/pybind11-2.13.6.tar.gz", hash = "sha256:ba6af10348c12b24e92fa086b39cfba0eff619b61ac77c406167d813b096d39a", size = 218403 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/13/2f/0f24b288e2ce56f51c920137620b4434a38fd80583dbbe24fc2a1656c388/pybind11-2.13.6-py3-none-any.whl", hash = "sha256:237c41e29157b962835d356b370ededd57594a26d5894a795960f0047cb5caf5", size = 243282 }, -] - -[[package]] -name = "pycparser" -version = "2.22" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/1d/b2/31537cf4b1ca988837256c910a668b553fceb8f069bedc4b1c826024b52c/pycparser-2.22.tar.gz", hash = "sha256:491c8be9c040f5390f5bf44a5b07752bd07f56edf992381b05c701439eec10f6", size = 172736 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/13/a3/a812df4e2dd5696d1f351d58b8fe16a405b234ad2886a0dab9183fb78109/pycparser-2.22-py3-none-any.whl", hash = "sha256:c3702b6d3dd8c7abc1afa565d7e63d53a1d0bd86cdc24edd75470f4de499cfcc", size = 117552 }, -] - -[[package]] -name = "pycryptodome" -version = "3.23.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/8e/a6/8452177684d5e906854776276ddd34eca30d1b1e15aa1ee9cefc289a33f5/pycryptodome-3.23.0.tar.gz", hash = "sha256:447700a657182d60338bab09fdb27518f8856aecd80ae4c6bdddb67ff5da44ef", size = 4921276 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/db/6c/a1f71542c969912bb0e106f64f60a56cc1f0fabecf9396f45accbe63fa68/pycryptodome-3.23.0-cp37-abi3-macosx_10_9_universal2.whl", hash = "sha256:187058ab80b3281b1de11c2e6842a357a1f71b42cb1e15bce373f3d238135c27", size = 2495627 }, - { url = "https://files.pythonhosted.org/packages/6e/4e/a066527e079fc5002390c8acdd3aca431e6ea0a50ffd7201551175b47323/pycryptodome-3.23.0-cp37-abi3-macosx_10_9_x86_64.whl", hash = "sha256:cfb5cd445280c5b0a4e6187a7ce8de5a07b5f3f897f235caa11f1f435f182843", size = 1640362 }, - { url = "https://files.pythonhosted.org/packages/50/52/adaf4c8c100a8c49d2bd058e5b551f73dfd8cb89eb4911e25a0c469b6b4e/pycryptodome-3.23.0-cp37-abi3-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:67bd81fcbe34f43ad9422ee8fd4843c8e7198dd88dd3d40e6de42ee65fbe1490", size = 2182625 }, - { url = "https://files.pythonhosted.org/packages/5f/e9/a09476d436d0ff1402ac3867d933c61805ec2326c6ea557aeeac3825604e/pycryptodome-3.23.0-cp37-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c8987bd3307a39bc03df5c8e0e3d8be0c4c3518b7f044b0f4c15d1aa78f52575", size = 2268954 }, - { url = "https://files.pythonhosted.org/packages/f9/c5/ffe6474e0c551d54cab931918127c46d70cab8f114e0c2b5a3c071c2f484/pycryptodome-3.23.0-cp37-abi3-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:aa0698f65e5b570426fc31b8162ed4603b0c2841cbb9088e2b01641e3065915b", size = 2308534 }, - { url = "https://files.pythonhosted.org/packages/18/28/e199677fc15ecf43010f2463fde4c1a53015d1fe95fb03bca2890836603a/pycryptodome-3.23.0-cp37-abi3-musllinux_1_2_aarch64.whl", hash = "sha256:53ecbafc2b55353edcebd64bf5da94a2a2cdf5090a6915bcca6eca6cc452585a", size = 2181853 }, - { url = "https://files.pythonhosted.org/packages/ce/ea/4fdb09f2165ce1365c9eaefef36625583371ee514db58dc9b65d3a255c4c/pycryptodome-3.23.0-cp37-abi3-musllinux_1_2_i686.whl", hash = "sha256:156df9667ad9f2ad26255926524e1c136d6664b741547deb0a86a9acf5ea631f", size = 2342465 }, - { url = "https://files.pythonhosted.org/packages/22/82/6edc3fc42fe9284aead511394bac167693fb2b0e0395b28b8bedaa07ef04/pycryptodome-3.23.0-cp37-abi3-musllinux_1_2_x86_64.whl", hash = "sha256:dea827b4d55ee390dc89b2afe5927d4308a8b538ae91d9c6f7a5090f397af1aa", size = 2267414 }, - { url = "https://files.pythonhosted.org/packages/59/fe/aae679b64363eb78326c7fdc9d06ec3de18bac68be4b612fc1fe8902693c/pycryptodome-3.23.0-cp37-abi3-win32.whl", hash = "sha256:507dbead45474b62b2bbe318eb1c4c8ee641077532067fec9c1aa82c31f84886", size = 1768484 }, - { url = "https://files.pythonhosted.org/packages/54/2f/e97a1b8294db0daaa87012c24a7bb714147c7ade7656973fd6c736b484ff/pycryptodome-3.23.0-cp37-abi3-win_amd64.whl", hash = "sha256:c75b52aacc6c0c260f204cbdd834f76edc9fb0d8e0da9fbf8352ef58202564e2", size = 1799636 }, - { url = "https://files.pythonhosted.org/packages/18/3d/f9441a0d798bf2b1e645adc3265e55706aead1255ccdad3856dbdcffec14/pycryptodome-3.23.0-cp37-abi3-win_arm64.whl", hash = "sha256:11eeeb6917903876f134b56ba11abe95c0b0fd5e3330def218083c7d98bbcb3c", size = 1703675 }, - { url = "https://files.pythonhosted.org/packages/9f/7c/f5b0556590e7b4e710509105e668adb55aa9470a9f0e4dea9c40a4a11ce1/pycryptodome-3.23.0-pp27-pypy_73-manylinux2010_x86_64.whl", hash = "sha256:350ebc1eba1da729b35ab7627a833a1a355ee4e852d8ba0447fafe7b14504d56", size = 1705791 }, - { url = "https://files.pythonhosted.org/packages/33/38/dcc795578d610ea1aaffef4b148b8cafcfcf4d126b1e58231ddc4e475c70/pycryptodome-3.23.0-pp27-pypy_73-win32.whl", hash = "sha256:93837e379a3e5fd2bb00302a47aee9fdf7940d83595be3915752c74033d17ca7", size = 1780265 }, - { url = "https://files.pythonhosted.org/packages/d9/12/e33935a0709c07de084d7d58d330ec3f4daf7910a18e77937affdb728452/pycryptodome-3.23.0-pp310-pypy310_pp73-macosx_10_15_x86_64.whl", hash = "sha256:ddb95b49df036ddd264a0ad246d1be5b672000f12d6961ea2c267083a5e19379", size = 1623886 }, - { url = "https://files.pythonhosted.org/packages/22/0b/aa8f9419f25870889bebf0b26b223c6986652bdf071f000623df11212c90/pycryptodome-3.23.0-pp310-pypy310_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d8e95564beb8782abfd9e431c974e14563a794a4944c29d6d3b7b5ea042110b4", size = 1672151 }, - { url = "https://files.pythonhosted.org/packages/d4/5e/63f5cbde2342b7f70a39e591dbe75d9809d6338ce0b07c10406f1a140cdc/pycryptodome-3.23.0-pp310-pypy310_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:14e15c081e912c4b0d75632acd8382dfce45b258667aa3c67caf7a4d4c13f630", size = 1664461 }, - { url = "https://files.pythonhosted.org/packages/d6/92/608fbdad566ebe499297a86aae5f2a5263818ceeecd16733006f1600403c/pycryptodome-3.23.0-pp310-pypy310_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:a7fc76bf273353dc7e5207d172b83f569540fc9a28d63171061c42e361d22353", size = 1702440 }, - { url = "https://files.pythonhosted.org/packages/d1/92/2eadd1341abd2989cce2e2740b4423608ee2014acb8110438244ee97d7ff/pycryptodome-3.23.0-pp310-pypy310_pp73-win_amd64.whl", hash = "sha256:45c69ad715ca1a94f778215a11e66b7ff989d792a4d63b68dc586a1da1392ff5", size = 1803005 }, -] - -[[package]] -name = "pydantic" -version = "2.11.7" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "annotated-types" }, - { name = "pydantic-core" }, - { name = "typing-extensions" }, - { name = "typing-inspection" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/00/dd/4325abf92c39ba8623b5af936ddb36ffcfe0beae70405d456ab1fb2f5b8c/pydantic-2.11.7.tar.gz", hash = "sha256:d989c3c6cb79469287b1569f7447a17848c998458d49ebe294e975b9baf0f0db", size = 788350 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/6a/c0/ec2b1c8712ca690e5d61979dee872603e92b8a32f94cc1b72d53beab008a/pydantic-2.11.7-py3-none-any.whl", hash = "sha256:dde5df002701f6de26248661f6835bbe296a47bf73990135c7d07ce741b9623b", size = 444782 }, -] - -[[package]] -name = "pydantic-core" -version = "2.33.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/ad/88/5f2260bdfae97aabf98f1778d43f69574390ad787afb646292a638c923d4/pydantic_core-2.33.2.tar.gz", hash = "sha256:7cb8bc3605c29176e1b105350d2e6474142d7c1bd1d9327c4a9bdb46bf827acc", size = 435195 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/e5/92/b31726561b5dae176c2d2c2dc43a9c5bfba5d32f96f8b4c0a600dd492447/pydantic_core-2.33.2-cp310-cp310-macosx_10_12_x86_64.whl", hash = "sha256:2b3d326aaef0c0399d9afffeb6367d5e26ddc24d351dbc9c636840ac355dc5d8", size = 2028817 }, - { url = "https://files.pythonhosted.org/packages/a3/44/3f0b95fafdaca04a483c4e685fe437c6891001bf3ce8b2fded82b9ea3aa1/pydantic_core-2.33.2-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:0e5b2671f05ba48b94cb90ce55d8bdcaaedb8ba00cc5359f6810fc918713983d", size = 1861357 }, - { url = "https://files.pythonhosted.org/packages/30/97/e8f13b55766234caae05372826e8e4b3b96e7b248be3157f53237682e43c/pydantic_core-2.33.2-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0069c9acc3f3981b9ff4cdfaf088e98d83440a4c7ea1bc07460af3d4dc22e72d", size = 1898011 }, - { url = "https://files.pythonhosted.org/packages/9b/a3/99c48cf7bafc991cc3ee66fd544c0aae8dc907b752f1dad2d79b1b5a471f/pydantic_core-2.33.2-cp310-cp310-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:d53b22f2032c42eaaf025f7c40c2e3b94568ae077a606f006d206a463bc69572", size = 1982730 }, - { url = "https://files.pythonhosted.org/packages/de/8e/a5b882ec4307010a840fb8b58bd9bf65d1840c92eae7534c7441709bf54b/pydantic_core-2.33.2-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:0405262705a123b7ce9f0b92f123334d67b70fd1f20a9372b907ce1080c7ba02", size = 2136178 }, - { url = "https://files.pythonhosted.org/packages/e4/bb/71e35fc3ed05af6834e890edb75968e2802fe98778971ab5cba20a162315/pydantic_core-2.33.2-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:4b25d91e288e2c4e0662b8038a28c6a07eaac3e196cfc4ff69de4ea3db992a1b", size = 2736462 }, - { url = "https://files.pythonhosted.org/packages/31/0d/c8f7593e6bc7066289bbc366f2235701dcbebcd1ff0ef8e64f6f239fb47d/pydantic_core-2.33.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6bdfe4b3789761f3bcb4b1ddf33355a71079858958e3a552f16d5af19768fef2", size = 2005652 }, - { url = "https://files.pythonhosted.org/packages/d2/7a/996d8bd75f3eda405e3dd219ff5ff0a283cd8e34add39d8ef9157e722867/pydantic_core-2.33.2-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:efec8db3266b76ef9607c2c4c419bdb06bf335ae433b80816089ea7585816f6a", size = 2113306 }, - { url = "https://files.pythonhosted.org/packages/ff/84/daf2a6fb2db40ffda6578a7e8c5a6e9c8affb251a05c233ae37098118788/pydantic_core-2.33.2-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:031c57d67ca86902726e0fae2214ce6770bbe2f710dc33063187a68744a5ecac", size = 2073720 }, - { url = "https://files.pythonhosted.org/packages/77/fb/2258da019f4825128445ae79456a5499c032b55849dbd5bed78c95ccf163/pydantic_core-2.33.2-cp310-cp310-musllinux_1_1_armv7l.whl", hash = "sha256:f8de619080e944347f5f20de29a975c2d815d9ddd8be9b9b7268e2e3ef68605a", size = 2244915 }, - { url = "https://files.pythonhosted.org/packages/d8/7a/925ff73756031289468326e355b6fa8316960d0d65f8b5d6b3a3e7866de7/pydantic_core-2.33.2-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:73662edf539e72a9440129f231ed3757faab89630d291b784ca99237fb94db2b", size = 2241884 }, - { url = "https://files.pythonhosted.org/packages/0b/b0/249ee6d2646f1cdadcb813805fe76265745c4010cf20a8eba7b0e639d9b2/pydantic_core-2.33.2-cp310-cp310-win32.whl", hash = "sha256:0a39979dcbb70998b0e505fb1556a1d550a0781463ce84ebf915ba293ccb7e22", size = 1910496 }, - { url = "https://files.pythonhosted.org/packages/66/ff/172ba8f12a42d4b552917aa65d1f2328990d3ccfc01d5b7c943ec084299f/pydantic_core-2.33.2-cp310-cp310-win_amd64.whl", hash = "sha256:b0379a2b24882fef529ec3b4987cb5d003b9cda32256024e6fe1586ac45fc640", size = 1955019 }, - { url = "https://files.pythonhosted.org/packages/30/68/373d55e58b7e83ce371691f6eaa7175e3a24b956c44628eb25d7da007917/pydantic_core-2.33.2-pp310-pypy310_pp73-macosx_10_12_x86_64.whl", hash = "sha256:5c4aa4e82353f65e548c476b37e64189783aa5384903bfea4f41580f255fddfa", size = 2023982 }, - { url = "https://files.pythonhosted.org/packages/a4/16/145f54ac08c96a63d8ed6442f9dec17b2773d19920b627b18d4f10a061ea/pydantic_core-2.33.2-pp310-pypy310_pp73-macosx_11_0_arm64.whl", hash = "sha256:d946c8bf0d5c24bf4fe333af284c59a19358aa3ec18cb3dc4370080da1e8ad29", size = 1858412 }, - { url = "https://files.pythonhosted.org/packages/41/b1/c6dc6c3e2de4516c0bb2c46f6a373b91b5660312342a0cf5826e38ad82fa/pydantic_core-2.33.2-pp310-pypy310_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:87b31b6846e361ef83fedb187bb5b4372d0da3f7e28d85415efa92d6125d6e6d", size = 1892749 }, - { url = "https://files.pythonhosted.org/packages/12/73/8cd57e20afba760b21b742106f9dbdfa6697f1570b189c7457a1af4cd8a0/pydantic_core-2.33.2-pp310-pypy310_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:aa9d91b338f2df0508606f7009fde642391425189bba6d8c653afd80fd6bb64e", size = 2067527 }, - { url = "https://files.pythonhosted.org/packages/e3/d5/0bb5d988cc019b3cba4a78f2d4b3854427fc47ee8ec8e9eaabf787da239c/pydantic_core-2.33.2-pp310-pypy310_pp73-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:2058a32994f1fde4ca0480ab9d1e75a0e8c87c22b53a3ae66554f9af78f2fe8c", size = 2108225 }, - { url = "https://files.pythonhosted.org/packages/f1/c5/00c02d1571913d496aabf146106ad8239dc132485ee22efe08085084ff7c/pydantic_core-2.33.2-pp310-pypy310_pp73-musllinux_1_1_aarch64.whl", hash = "sha256:0e03262ab796d986f978f79c943fc5f620381be7287148b8010b4097f79a39ec", size = 2069490 }, - { url = "https://files.pythonhosted.org/packages/22/a8/dccc38768274d3ed3a59b5d06f59ccb845778687652daa71df0cab4040d7/pydantic_core-2.33.2-pp310-pypy310_pp73-musllinux_1_1_armv7l.whl", hash = "sha256:1a8695a8d00c73e50bff9dfda4d540b7dee29ff9b8053e38380426a85ef10052", size = 2237525 }, - { url = "https://files.pythonhosted.org/packages/d4/e7/4f98c0b125dda7cf7ccd14ba936218397b44f50a56dd8c16a3091df116c3/pydantic_core-2.33.2-pp310-pypy310_pp73-musllinux_1_1_x86_64.whl", hash = "sha256:fa754d1850735a0b0e03bcffd9d4b4343eb417e47196e4485d9cca326073a42c", size = 2238446 }, - { url = "https://files.pythonhosted.org/packages/ce/91/2ec36480fdb0b783cd9ef6795753c1dea13882f2e68e73bce76ae8c21e6a/pydantic_core-2.33.2-pp310-pypy310_pp73-win_amd64.whl", hash = "sha256:a11c8d26a50bfab49002947d3d237abe4d9e4b5bdc8846a63537b6488e197808", size = 2066678 }, -] - -[[package]] -name = "pydub" -version = "0.25.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/fe/9a/e6bca0eed82db26562c73b5076539a4a08d3cffd19c3cc5913a3e61145fd/pydub-0.25.1.tar.gz", hash = "sha256:980a33ce9949cab2a569606b65674d748ecbca4f0796887fd6f46173a7b0d30f", size = 38326 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/a6/53/d78dc063216e62fc55f6b2eebb447f6a4b0a59f55c8406376f76bf959b08/pydub-0.25.1-py2.py3-none-any.whl", hash = "sha256:65617e33033874b59d87db603aa1ed450633288aefead953b30bded59cb599a6", size = 32327 }, -] - -[[package]] -name = "pygments" -version = "2.19.2" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/b0/77/a5b8c569bf593b0140bde72ea885a803b82086995367bf2037de0159d924/pygments-2.19.2.tar.gz", hash = "sha256:636cb2477cec7f8952536970bc533bc43743542f70392ae026374600add5b887", size = 4968631 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/c7/21/705964c7812476f378728bdf590ca4b771ec72385c533964653c68e86bdc/pygments-2.19.2-py3-none-any.whl", hash = "sha256:86540386c03d588bb81d44bc3928634ff26449851e99741617ecb9037ee5ec0b", size = 1225217 }, -] - -[[package]] -name = "pyloudnorm" -version = "0.1.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "future" }, - { name = "numpy" }, - { name = "scipy" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/75/b5/39d59c44ecd828fabfdbd796b50a561e6543ca90ef440ab307374f107856/pyloudnorm-0.1.1.tar.gz", hash = "sha256:63cd4e197dea4e7795160ea08ed02d318091bce883e436a6dbc5963326b71e1e", size = 8588 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/58/f5/6724805521ab4e723a12182f92374031032aff28a8a89dc8505c52b79032/pyloudnorm-0.1.1-py3-none-any.whl", hash = "sha256:d7f12ebdd097a464d87ce2878fc4d942f15f8233e26cc03f33fefa226f869a14", size = 9636 }, -] - -[[package]] -name = "pynndescent" -version = "0.5.13" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "joblib" }, - { name = "llvmlite" }, - { name = "numba" }, - { name = "scikit-learn" }, - { name = "scipy" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/7e/58/560a4db5eb3794d922fe55804b10326534ded3d971e1933c1eef91193f5e/pynndescent-0.5.13.tar.gz", hash = "sha256:d74254c0ee0a1eeec84597d5fe89fedcf778593eeabe32c2f97412934a9800fb", size = 2975955 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/d2/53/d23a97e0a2c690d40b165d1062e2c4ccc796be458a1ce59f6ba030434663/pynndescent-0.5.13-py3-none-any.whl", hash = "sha256:69aabb8f394bc631b6ac475a1c7f3994c54adf3f51cd63b2730fefba5771b949", size = 56850 }, -] - -[[package]] -name = "pyparsing" -version = "3.2.3" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/bb/22/f1129e69d94ffff626bdb5c835506b3a5b4f3d070f17ea295e12c2c6f60f/pyparsing-3.2.3.tar.gz", hash = "sha256:b9c13f1ab8b3b542f72e28f634bad4de758ab3ce4546e4301970ad6fa77c38be", size = 1088608 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/05/e7/df2285f3d08fee213f2d041540fa4fc9ca6c2d44cf36d3a035bf2a8d2bcc/pyparsing-3.2.3-py3-none-any.whl", hash = "sha256:a749938e02d6fd0b59b356ca504a24982314bb090c383e3cf201c95ef7e2bfcf", size = 111120 }, -] - -[[package]] -name = "python-dateutil" -version = "2.9.0.post0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "six" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/66/c0/0c8b6ad9f17a802ee498c46e004a0eb49bc148f2fd230864601a86dcf6db/python-dateutil-2.9.0.post0.tar.gz", hash = "sha256:37dd54208da7e1cd875388217d5e00ebd4179249f90fb72437e91a35459a0ad3", size = 342432 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/ec/57/56b9bcc3c9c6a792fcbaf139543cee77261f3651ca9da0c93f5c1221264b/python_dateutil-2.9.0.post0-py2.py3-none-any.whl", hash = "sha256:a8b2bc7bffae282281c8140a97d3aa9c14da0b136dfe83f850eea9a5f7470427", size = 229892 }, -] - -[[package]] -name = "python-json-logger" -version = "3.3.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/9e/de/d3144a0bceede957f961e975f3752760fbe390d57fbe194baf709d8f1f7b/python_json_logger-3.3.0.tar.gz", hash = "sha256:12b7e74b17775e7d565129296105bbe3910842d9d0eb083fc83a6a617aa8df84", size = 16642 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/08/20/0f2523b9e50a8052bc6a8b732dfc8568abbdc42010aef03a2d750bdab3b2/python_json_logger-3.3.0-py3-none-any.whl", hash = "sha256:dd980fae8cffb24c13caf6e158d3d61c0d6d22342f932cb6e9deedab3d35eec7", size = 15163 }, -] - -[[package]] -name = "pytorch-lightning" -version = "2.5.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "fsspec", extra = ["http"] }, - { name = "lightning-utilities" }, - { name = "packaging" }, - { name = "pyyaml" }, - { name = "torch" }, - { name = "torchmetrics" }, - { name = "tqdm" }, - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/01/3e/728fbdc671d07727ad447f9401d98a43570573965beb3cb2060f9a330b4f/pytorch_lightning-2.5.2.tar.gz", hash = "sha256:f817087d611be8d43b777dd4e543d72703e235510936677a13e6c29f7fd790e3", size = 636859 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/e2/42/47c186c8f9e956e559c89e6c764d5d5d0d0af517c04ca0ad39bd0a357d3a/pytorch_lightning-2.5.2-py3-none-any.whl", hash = "sha256:17cfdf89bd98074e389101f097cdf34c486a1f5c6d3fdcefbaf4dea7f97ff0bf", size = 825366 }, -] - -[[package]] -name = "pytorch-metric-learning" -version = "2.9.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "numpy" }, - { name = "scikit-learn" }, - { name = "torch" }, - { name = "tqdm" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/9b/80/6e61b1a91debf4c1b47d441f9a9d7fe2aabcdd9575ed70b2811474eb95c3/pytorch-metric-learning-2.9.0.tar.gz", hash = "sha256:27a626caf5e2876a0fd666605a78cb67ef7597e25d7a68c18053dd503830701f", size = 84530 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/46/7d/73ef5052f57b7720cad00e16598db3592a5ef4826745ffca67a2f085d4dc/pytorch_metric_learning-2.9.0-py3-none-any.whl", hash = "sha256:d51646006dc87168f00cf954785db133a4c5aac81253877248737aa42ef6432a", size = 127801 }, -] - -[[package]] -name = "pytorch-wpe" -version = "0.0.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "numpy" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/db/39/8d94737fd6fab4028687575099566a125100f3ba8c638f861506747d7b7c/pytorch_wpe-0.0.1.tar.gz", hash = "sha256:fc7e706b5411800c4483fe94db7dcd82ecf6c57bc013af529ab4fb675c9cc29c", size = 4457 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/23/de/c47967a11bfe68cb28d2f19e55c7027993c3721eba79813db65d245e4ced/pytorch_wpe-0.0.1-py3-none-any.whl", hash = "sha256:fa0dc9f818fba81b36c1a51a53331cf6ed975f29b33f23e07b0deb4bee82eaad", size = 8080 }, -] - -[[package]] -name = "pytz" -version = "2025.2" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/f8/bf/abbd3cdfb8fbc7fb3d4d38d320f2441b1e7cbe29be4f23797b4a2b5d8aac/pytz-2025.2.tar.gz", hash = "sha256:360b9e3dbb49a209c21ad61809c7fb453643e048b38924c765813546746e81c3", size = 320884 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/81/c4/34e93fe5f5429d7570ec1fa436f1986fb1f00c3e0f43a589fe2bbcd22c3f/pytz-2025.2-py2.py3-none-any.whl", hash = "sha256:5ddf76296dd8c44c26eb8f4b6f35488f3ccbf6fbbd7adee0b7262d43f0ec2f00", size = 509225 }, -] - -[[package]] -name = "pywin32" -version = "311" -source = { registry = "https://pypi.org/simple" } -wheels = [ - { url = "https://files.pythonhosted.org/packages/7b/40/44efbb0dfbd33aca6a6483191dae0716070ed99e2ecb0c53683f400a0b4f/pywin32-311-cp310-cp310-win32.whl", hash = "sha256:d03ff496d2a0cd4a5893504789d4a15399133fe82517455e78bad62efbb7f0a3", size = 8760432 }, - { url = "https://files.pythonhosted.org/packages/5e/bf/360243b1e953bd254a82f12653974be395ba880e7ec23e3731d9f73921cc/pywin32-311-cp310-cp310-win_amd64.whl", hash = "sha256:797c2772017851984b97180b0bebe4b620bb86328e8a884bb626156295a63b3b", size = 9590103 }, - { url = "https://files.pythonhosted.org/packages/57/38/d290720e6f138086fb3d5ffe0b6caa019a791dd57866940c82e4eeaf2012/pywin32-311-cp310-cp310-win_arm64.whl", hash = "sha256:0502d1facf1fed4839a9a51ccbcc63d952cf318f78ffc00a7e78528ac27d7a2b", size = 8778557 }, -] - -[[package]] -name = "pywinpty" -version = "3.0.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/06/df/429cc505dc5f77ab0612c4b60bca2e3dcc81f6c321844ee017d6dc0f4a95/pywinpty-3.0.0.tar.gz", hash = "sha256:68f70e68a9f0766ffdea3fc500351cb7b9b012bcb8239a411f7ff0fc8f86dcb1", size = 28551 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/15/f9/13d62974debb0c74ce3fa3d96b32cee6fce4f2d634789217e67aebf339f6/pywinpty-3.0.0-cp310-cp310-win_amd64.whl", hash = "sha256:327b6034e0dc38352c1c99a7c0b3e54941b4e506a5f21acce63609cd2ab6cce2", size = 2050843 }, -] - -[[package]] -name = "pyyaml" -version = "6.0.2" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/54/ed/79a089b6be93607fa5cdaedf301d7dfb23af5f25c398d5ead2525b063e17/pyyaml-6.0.2.tar.gz", hash = "sha256:d584d9ec91ad65861cc08d42e834324ef890a082e591037abe114850ff7bbc3e", size = 130631 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/9b/95/a3fac87cb7158e231b5a6012e438c647e1a87f09f8e0d123acec8ab8bf71/PyYAML-6.0.2-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:0a9a2848a5b7feac301353437eb7d5957887edbf81d56e903999a75a3d743086", size = 184199 }, - { url = "https://files.pythonhosted.org/packages/c7/7a/68bd47624dab8fd4afbfd3c48e3b79efe09098ae941de5b58abcbadff5cb/PyYAML-6.0.2-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:29717114e51c84ddfba879543fb232a6ed60086602313ca38cce623c1d62cfbf", size = 171758 }, - { url = "https://files.pythonhosted.org/packages/49/ee/14c54df452143b9ee9f0f29074d7ca5516a36edb0b4cc40c3f280131656f/PyYAML-6.0.2-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8824b5a04a04a047e72eea5cec3bc266db09e35de6bdfe34c9436ac5ee27d237", size = 718463 }, - { url = "https://files.pythonhosted.org/packages/4d/61/de363a97476e766574650d742205be468921a7b532aa2499fcd886b62530/PyYAML-6.0.2-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:7c36280e6fb8385e520936c3cb3b8042851904eba0e58d277dca80a5cfed590b", size = 719280 }, - { url = "https://files.pythonhosted.org/packages/6b/4e/1523cb902fd98355e2e9ea5e5eb237cbc5f3ad5f3075fa65087aa0ecb669/PyYAML-6.0.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ec031d5d2feb36d1d1a24380e4db6d43695f3748343d99434e6f5f9156aaa2ed", size = 751239 }, - { url = "https://files.pythonhosted.org/packages/b7/33/5504b3a9a4464893c32f118a9cc045190a91637b119a9c881da1cf6b7a72/PyYAML-6.0.2-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:936d68689298c36b53b29f23c6dbb74de12b4ac12ca6cfe0e047bedceea56180", size = 695802 }, - { url = "https://files.pythonhosted.org/packages/5c/20/8347dcabd41ef3a3cdc4f7b7a2aff3d06598c8779faa189cdbf878b626a4/PyYAML-6.0.2-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:23502f431948090f597378482b4812b0caae32c22213aecf3b55325e049a6c68", size = 720527 }, - { url = "https://files.pythonhosted.org/packages/be/aa/5afe99233fb360d0ff37377145a949ae258aaab831bde4792b32650a4378/PyYAML-6.0.2-cp310-cp310-win32.whl", hash = "sha256:2e99c6826ffa974fe6e27cdb5ed0021786b03fc98e5ee3c5bfe1fd5015f42b99", size = 144052 }, - { url = "https://files.pythonhosted.org/packages/b5/84/0fa4b06f6d6c958d207620fc60005e241ecedceee58931bb20138e1e5776/PyYAML-6.0.2-cp310-cp310-win_amd64.whl", hash = "sha256:a4d3091415f010369ae4ed1fc6b79def9416358877534caf6a0fdd2146c87a3e", size = 161774 }, -] - -[[package]] -name = "pyzmq" -version = "27.0.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "cffi", marker = "implementation_name == 'pypy'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/f1/06/50a4e9648b3e8b992bef8eb632e457307553a89d294103213cfd47b3da69/pyzmq-27.0.0.tar.gz", hash = "sha256:b1f08eeb9ce1510e6939b6e5dcd46a17765e2333daae78ecf4606808442e52cf", size = 280478 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/9c/09/1681d4b047626d352c083770618ac29655ab1f5c20eee31dc94c000b9b7b/pyzmq-27.0.0-cp310-cp310-macosx_10_15_universal2.whl", hash = "sha256:b973ee650e8f442ce482c1d99ca7ab537c69098d53a3d046676a484fd710c87a", size = 1329291 }, - { url = "https://files.pythonhosted.org/packages/9d/b2/9c9385225fdd54db9506ed8accbb9ea63ca813ba59d43d7f282a6a16a30b/pyzmq-27.0.0-cp310-cp310-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:661942bc7cd0223d569d808f2e5696d9cc120acc73bf3e88a1f1be7ab648a7e4", size = 905952 }, - { url = "https://files.pythonhosted.org/packages/41/73/333c72c7ec182cdffe25649e3da1c3b9f3cf1cede63cfdc23d1384d4a601/pyzmq-27.0.0-cp310-cp310-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:50360fb2a056ffd16e5f4177eee67f1dd1017332ea53fb095fe7b5bf29c70246", size = 666165 }, - { url = "https://files.pythonhosted.org/packages/a5/fe/fc7b9c1a50981928e25635a926653cb755364316db59ccd6e79cfb9a0b4f/pyzmq-27.0.0-cp310-cp310-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:cf209a6dc4b420ed32a7093642843cbf8703ed0a7d86c16c0b98af46762ebefb", size = 853755 }, - { url = "https://files.pythonhosted.org/packages/8c/4c/740ed4b6e8fa160cd19dc5abec8db68f440564b2d5b79c1d697d9862a2f7/pyzmq-27.0.0-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:c2dace4a7041cca2fba5357a2d7c97c5effdf52f63a1ef252cfa496875a3762d", size = 1654868 }, - { url = "https://files.pythonhosted.org/packages/97/00/875b2ecfcfc78ab962a59bd384995186818524ea957dc8ad3144611fae12/pyzmq-27.0.0-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:63af72b2955fc77caf0a77444baa2431fcabb4370219da38e1a9f8d12aaebe28", size = 2033443 }, - { url = "https://files.pythonhosted.org/packages/60/55/6dd9c470c42d713297c5f2a56f7903dc1ebdb4ab2edda996445c21651900/pyzmq-27.0.0-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:e8c4adce8e37e75c4215297d7745551b8dcfa5f728f23ce09bf4e678a9399413", size = 1891288 }, - { url = "https://files.pythonhosted.org/packages/28/5d/54b0ef50d40d7c65a627f4a4b4127024ba9820f2af8acd933a4d30ae192e/pyzmq-27.0.0-cp310-cp310-win32.whl", hash = "sha256:5d5ef4718ecab24f785794e0e7536436698b459bfbc19a1650ef55280119d93b", size = 567936 }, - { url = "https://files.pythonhosted.org/packages/18/ea/dedca4321de748ca48d3bcdb72274d4d54e8d84ea49088d3de174bd45d88/pyzmq-27.0.0-cp310-cp310-win_amd64.whl", hash = "sha256:e40609380480b3d12c30f841323f42451c755b8fece84235236f5fe5ffca8c1c", size = 628686 }, - { url = "https://files.pythonhosted.org/packages/d4/a7/fcdeedc306e71e94ac262cba2d02337d885f5cdb7e8efced8e5ffe327808/pyzmq-27.0.0-cp310-cp310-win_arm64.whl", hash = "sha256:6b0397b0be277b46762956f576e04dc06ced265759e8c2ff41a0ee1aa0064198", size = 559039 }, - { url = "https://files.pythonhosted.org/packages/93/a7/9ad68f55b8834ede477842214feba6a4c786d936c022a67625497aacf61d/pyzmq-27.0.0-cp312-abi3-macosx_10_15_universal2.whl", hash = "sha256:cbabc59dcfaac66655c040dfcb8118f133fb5dde185e5fc152628354c1598e52", size = 1305438 }, - { url = "https://files.pythonhosted.org/packages/ba/ee/26aa0f98665a22bc90ebe12dced1de5f3eaca05363b717f6fb229b3421b3/pyzmq-27.0.0-cp312-abi3-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:cb0ac5179cba4b2f94f1aa208fbb77b62c4c9bf24dd446278b8b602cf85fcda3", size = 895095 }, - { url = "https://files.pythonhosted.org/packages/cf/85/c57e7ab216ecd8aa4cc7e3b83b06cc4e9cf45c87b0afc095f10cd5ce87c1/pyzmq-27.0.0-cp312-abi3-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:53a48f0228eab6cbf69fde3aa3c03cbe04e50e623ef92ae395fce47ef8a76152", size = 651826 }, - { url = "https://files.pythonhosted.org/packages/69/9a/9ea7e230feda9400fb0ae0d61d7d6ddda635e718d941c44eeab22a179d34/pyzmq-27.0.0-cp312-abi3-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:111db5f395e09f7e775f759d598f43cb815fc58e0147623c4816486e1a39dc22", size = 839750 }, - { url = "https://files.pythonhosted.org/packages/08/66/4cebfbe71f3dfbd417011daca267539f62ed0fbc68105357b68bbb1a25b7/pyzmq-27.0.0-cp312-abi3-musllinux_1_2_aarch64.whl", hash = "sha256:c8878011653dcdc27cc2c57e04ff96f0471e797f5c19ac3d7813a245bcb24371", size = 1641357 }, - { url = "https://files.pythonhosted.org/packages/ac/f6/b0f62578c08d2471c791287149cb8c2aaea414ae98c6e995c7dbe008adfb/pyzmq-27.0.0-cp312-abi3-musllinux_1_2_i686.whl", hash = "sha256:c0ed2c1f335ba55b5fdc964622254917d6b782311c50e138863eda409fbb3b6d", size = 2020281 }, - { url = "https://files.pythonhosted.org/packages/37/b9/4f670b15c7498495da9159edc374ec09c88a86d9cd5a47d892f69df23450/pyzmq-27.0.0-cp312-abi3-musllinux_1_2_x86_64.whl", hash = "sha256:e918d70862d4cfd4b1c187310015646a14e1f5917922ab45b29f28f345eeb6be", size = 1877110 }, - { url = "https://files.pythonhosted.org/packages/66/31/9dee25c226295b740609f0d46db2fe972b23b6f5cf786360980524a3ba92/pyzmq-27.0.0-cp312-abi3-win32.whl", hash = "sha256:88b4e43cab04c3c0f0d55df3b1eef62df2b629a1a369b5289a58f6fa8b07c4f4", size = 559297 }, - { url = "https://files.pythonhosted.org/packages/9b/12/52da5509800f7ff2d287b2f2b4e636e7ea0f001181cba6964ff6c1537778/pyzmq-27.0.0-cp312-abi3-win_amd64.whl", hash = "sha256:dce4199bf5f648a902ce37e7b3afa286f305cd2ef7a8b6ec907470ccb6c8b371", size = 619203 }, - { url = "https://files.pythonhosted.org/packages/93/6d/7f2e53b19d1edb1eb4f09ec7c3a1f945ca0aac272099eab757d15699202b/pyzmq-27.0.0-cp312-abi3-win_arm64.whl", hash = "sha256:56e46bbb85d52c1072b3f809cc1ce77251d560bc036d3a312b96db1afe76db2e", size = 551927 }, - { url = "https://files.pythonhosted.org/packages/09/6f/be6523a7f3821c0b5370912ef02822c028611360e0d206dd945bdbf9eaef/pyzmq-27.0.0-pp310-pypy310_pp73-macosx_10_15_x86_64.whl", hash = "sha256:656c1866505a5735d0660b7da6d7147174bbf59d4975fc2b7f09f43c9bc25745", size = 835950 }, - { url = "https://files.pythonhosted.org/packages/c6/1e/a50fdd5c15018de07ab82a61bc460841be967ee7bbe7abee3b714d66f7ac/pyzmq-27.0.0-pp310-pypy310_pp73-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:74175b9e12779382432dd1d1f5960ebe7465d36649b98a06c6b26be24d173fab", size = 799876 }, - { url = "https://files.pythonhosted.org/packages/88/a1/89eb5b71f5a504f8f887aceb8e1eb3626e00c00aa8085381cdff475440dc/pyzmq-27.0.0-pp310-pypy310_pp73-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:d8c6de908465697a8708e4d6843a1e884f567962fc61eb1706856545141d0cbb", size = 567400 }, - { url = "https://files.pythonhosted.org/packages/56/aa/4571dbcff56cfb034bac73fde8294e123c975ce3eea89aff31bf6dc6382b/pyzmq-27.0.0-pp310-pypy310_pp73-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:c644aaacc01d0df5c7072826df45e67301f191c55f68d7b2916d83a9ddc1b551", size = 747031 }, - { url = "https://files.pythonhosted.org/packages/46/e0/d25f30fe0991293c5b2f5ef3b070d35fa6d57c0c7428898c3ab4913d0297/pyzmq-27.0.0-pp310-pypy310_pp73-win_amd64.whl", hash = "sha256:10f70c1d9a446a85013a36871a296007f6fe4232b530aa254baf9da3f8328bc0", size = 544726 }, -] - -[[package]] -name = "rapidfuzz" -version = "3.13.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/ed/f6/6895abc3a3d056b9698da3199b04c0e56226d530ae44a470edabf8b664f0/rapidfuzz-3.13.0.tar.gz", hash = "sha256:d2eaf3839e52cbcc0accbe9817a67b4b0fcf70aaeb229cfddc1c28061f9ce5d8", size = 57904226 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/de/27/ca10b3166024ae19a7e7c21f73c58dfd4b7fef7420e5497ee64ce6b73453/rapidfuzz-3.13.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:aafc42a1dc5e1beeba52cd83baa41372228d6d8266f6d803c16dbabbcc156255", size = 1998899 }, - { url = "https://files.pythonhosted.org/packages/f0/38/c4c404b13af0315483a6909b3a29636e18e1359307fb74a333fdccb3730d/rapidfuzz-3.13.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:85c9a131a44a95f9cac2eb6e65531db014e09d89c4f18c7b1fa54979cb9ff1f3", size = 1449949 }, - { url = "https://files.pythonhosted.org/packages/12/ae/15c71d68a6df6b8e24595421fdf5bcb305888318e870b7be8d935a9187ee/rapidfuzz-3.13.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7d7cec4242d30dd521ef91c0df872e14449d1dffc2a6990ede33943b0dae56c3", size = 1424199 }, - { url = "https://files.pythonhosted.org/packages/dc/9a/765beb9e14d7b30d12e2d6019e8b93747a0bedbc1d0cce13184fa3825426/rapidfuzz-3.13.0-cp310-cp310-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e297c09972698c95649e89121e3550cee761ca3640cd005e24aaa2619175464e", size = 5352400 }, - { url = "https://files.pythonhosted.org/packages/e2/b8/49479fe6f06b06cd54d6345ed16de3d1ac659b57730bdbe897df1e059471/rapidfuzz-3.13.0-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ef0f5f03f61b0e5a57b1df7beafd83df993fd5811a09871bad6038d08e526d0d", size = 1652465 }, - { url = "https://files.pythonhosted.org/packages/6f/d8/08823d496b7dd142a7b5d2da04337df6673a14677cfdb72f2604c64ead69/rapidfuzz-3.13.0-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:d8cf5f7cd6e4d5eb272baf6a54e182b2c237548d048e2882258336533f3f02b7", size = 1616590 }, - { url = "https://files.pythonhosted.org/packages/38/d4/5cfbc9a997e544f07f301c54d42aac9e0d28d457d543169e4ec859b8ce0d/rapidfuzz-3.13.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9256218ac8f1a957806ec2fb9a6ddfc6c32ea937c0429e88cf16362a20ed8602", size = 3086956 }, - { url = "https://files.pythonhosted.org/packages/25/1e/06d8932a72fa9576095234a15785136407acf8f9a7dbc8136389a3429da1/rapidfuzz-3.13.0-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:e1bdd2e6d0c5f9706ef7595773a81ca2b40f3b33fd7f9840b726fb00c6c4eb2e", size = 2494220 }, - { url = "https://files.pythonhosted.org/packages/03/16/5acf15df63119d5ca3d9a54b82807866ff403461811d077201ca351a40c3/rapidfuzz-3.13.0-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:5280be8fd7e2bee5822e254fe0a5763aa0ad57054b85a32a3d9970e9b09bbcbf", size = 7585481 }, - { url = "https://files.pythonhosted.org/packages/e1/cf/ebade4009431ea8e715e59e882477a970834ddaacd1a670095705b86bd0d/rapidfuzz-3.13.0-cp310-cp310-musllinux_1_2_ppc64le.whl", hash = "sha256:fd742c03885db1fce798a1cd87a20f47f144ccf26d75d52feb6f2bae3d57af05", size = 2894842 }, - { url = "https://files.pythonhosted.org/packages/a7/bd/0732632bd3f906bf613229ee1b7cbfba77515db714a0e307becfa8a970ae/rapidfuzz-3.13.0-cp310-cp310-musllinux_1_2_s390x.whl", hash = "sha256:5435fcac94c9ecf0504bf88a8a60c55482c32e18e108d6079a0089c47f3f8cf6", size = 3438517 }, - { url = "https://files.pythonhosted.org/packages/83/89/d3bd47ec9f4b0890f62aea143a1e35f78f3d8329b93d9495b4fa8a3cbfc3/rapidfuzz-3.13.0-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:93a755266856599be4ab6346273f192acde3102d7aa0735e2f48b456397a041f", size = 4412773 }, - { url = "https://files.pythonhosted.org/packages/b3/57/1a152a07883e672fc117c7f553f5b933f6e43c431ac3fd0e8dae5008f481/rapidfuzz-3.13.0-cp310-cp310-win32.whl", hash = "sha256:3abe6a4e8eb4cfc4cda04dd650a2dc6d2934cbdeda5def7e6fd1c20f6e7d2a0b", size = 1842334 }, - { url = "https://files.pythonhosted.org/packages/a7/68/7248addf95b6ca51fc9d955161072285da3059dd1472b0de773cff910963/rapidfuzz-3.13.0-cp310-cp310-win_amd64.whl", hash = "sha256:e8ddb58961401da7d6f55f185512c0d6bd24f529a637078d41dd8ffa5a49c107", size = 1624392 }, - { url = "https://files.pythonhosted.org/packages/68/23/f41c749f2c61ed1ed5575eaf9e73ef9406bfedbf20a3ffa438d15b5bf87e/rapidfuzz-3.13.0-cp310-cp310-win_arm64.whl", hash = "sha256:c523620d14ebd03a8d473c89e05fa1ae152821920c3ff78b839218ff69e19ca3", size = 865584 }, - { url = "https://files.pythonhosted.org/packages/d5/e1/f5d85ae3c53df6f817ca70dbdd37c83f31e64caced5bb867bec6b43d1fdf/rapidfuzz-3.13.0-pp310-pypy310_pp73-macosx_10_15_x86_64.whl", hash = "sha256:fe5790a36d33a5d0a6a1f802aa42ecae282bf29ac6f7506d8e12510847b82a45", size = 1904437 }, - { url = "https://files.pythonhosted.org/packages/db/d7/ded50603dddc5eb182b7ce547a523ab67b3bf42b89736f93a230a398a445/rapidfuzz-3.13.0-pp310-pypy310_pp73-macosx_11_0_arm64.whl", hash = "sha256:cdb33ee9f8a8e4742c6b268fa6bd739024f34651a06b26913381b1413ebe7590", size = 1383126 }, - { url = "https://files.pythonhosted.org/packages/c4/48/6f795e793babb0120b63a165496d64f989b9438efbeed3357d9a226ce575/rapidfuzz-3.13.0-pp310-pypy310_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8c99b76b93f7b495eee7dcb0d6a38fb3ce91e72e99d9f78faa5664a881cb2b7d", size = 1365565 }, - { url = "https://files.pythonhosted.org/packages/f0/50/0062a959a2d72ed17815824e40e2eefdb26f6c51d627389514510a7875f3/rapidfuzz-3.13.0-pp310-pypy310_pp73-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:6af42f2ede8b596a6aaf6d49fdee3066ca578f4856b85ab5c1e2145de367a12d", size = 5251719 }, - { url = "https://files.pythonhosted.org/packages/e7/02/bd8b70cd98b7a88e1621264778ac830c9daa7745cd63e838bd773b1aeebd/rapidfuzz-3.13.0-pp310-pypy310_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6c0efa73afbc5b265aca0d8a467ae2a3f40d6854cbe1481cb442a62b7bf23c99", size = 2991095 }, - { url = "https://files.pythonhosted.org/packages/9f/8d/632d895cdae8356826184864d74a5f487d40cb79f50a9137510524a1ba86/rapidfuzz-3.13.0-pp310-pypy310_pp73-win_amd64.whl", hash = "sha256:7ac21489de962a4e2fc1e8f0b0da4aa1adc6ab9512fd845563fecb4b4c52093a", size = 1553888 }, -] - -[[package]] -name = "referencing" -version = "0.36.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "attrs" }, - { name = "rpds-py" }, - { name = "typing-extensions", marker = "python_full_version == '3.10.12'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/2f/db/98b5c277be99dd18bfd91dd04e1b759cad18d1a338188c936e92f921c7e2/referencing-0.36.2.tar.gz", hash = "sha256:df2e89862cd09deabbdba16944cc3f10feb6b3e6f18e902f7cc25609a34775aa", size = 74744 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/c1/b1/3baf80dc6d2b7bc27a95a67752d0208e410351e3feb4eb78de5f77454d8d/referencing-0.36.2-py3-none-any.whl", hash = "sha256:e8699adbbf8b5c7de96d8ffa0eb5c158b3beafce084968e2ea8bb08c6794dcd0", size = 26775 }, -] - -[[package]] -name = "regex" -version = "2024.11.6" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/8e/5f/bd69653fbfb76cf8604468d3b4ec4c403197144c7bfe0e6a5fc9e02a07cb/regex-2024.11.6.tar.gz", hash = "sha256:7ab159b063c52a0333c884e4679f8d7a85112ee3078fe3d9004b2dd875585519", size = 399494 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/95/3c/4651f6b130c6842a8f3df82461a8950f923925db8b6961063e82744bddcc/regex-2024.11.6-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:ff590880083d60acc0433f9c3f713c51f7ac6ebb9adf889c79a261ecf541aa91", size = 482674 }, - { url = "https://files.pythonhosted.org/packages/15/51/9f35d12da8434b489c7b7bffc205c474a0a9432a889457026e9bc06a297a/regex-2024.11.6-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:658f90550f38270639e83ce492f27d2c8d2cd63805c65a13a14d36ca126753f0", size = 287684 }, - { url = "https://files.pythonhosted.org/packages/bd/18/b731f5510d1b8fb63c6b6d3484bfa9a59b84cc578ac8b5172970e05ae07c/regex-2024.11.6-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:164d8b7b3b4bcb2068b97428060b2a53be050085ef94eca7f240e7947f1b080e", size = 284589 }, - { url = "https://files.pythonhosted.org/packages/78/a2/6dd36e16341ab95e4c6073426561b9bfdeb1a9c9b63ab1b579c2e96cb105/regex-2024.11.6-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d3660c82f209655a06b587d55e723f0b813d3a7db2e32e5e7dc64ac2a9e86fde", size = 782511 }, - { url = "https://files.pythonhosted.org/packages/1b/2b/323e72d5d2fd8de0d9baa443e1ed70363ed7e7b2fb526f5950c5cb99c364/regex-2024.11.6-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:d22326fcdef5e08c154280b71163ced384b428343ae16a5ab2b3354aed12436e", size = 821149 }, - { url = "https://files.pythonhosted.org/packages/90/30/63373b9ea468fbef8a907fd273e5c329b8c9535fee36fc8dba5fecac475d/regex-2024.11.6-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:f1ac758ef6aebfc8943560194e9fd0fa18bcb34d89fd8bd2af18183afd8da3a2", size = 809707 }, - { url = "https://files.pythonhosted.org/packages/f2/98/26d3830875b53071f1f0ae6d547f1d98e964dd29ad35cbf94439120bb67a/regex-2024.11.6-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:997d6a487ff00807ba810e0f8332c18b4eb8d29463cfb7c820dc4b6e7562d0cf", size = 781702 }, - { url = "https://files.pythonhosted.org/packages/87/55/eb2a068334274db86208ab9d5599ffa63631b9f0f67ed70ea7c82a69bbc8/regex-2024.11.6-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:02a02d2bb04fec86ad61f3ea7f49c015a0681bf76abb9857f945d26159d2968c", size = 771976 }, - { url = "https://files.pythonhosted.org/packages/74/c0/be707bcfe98254d8f9d2cff55d216e946f4ea48ad2fd8cf1428f8c5332ba/regex-2024.11.6-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:f02f93b92358ee3f78660e43b4b0091229260c5d5c408d17d60bf26b6c900e86", size = 697397 }, - { url = "https://files.pythonhosted.org/packages/49/dc/bb45572ceb49e0f6509f7596e4ba7031f6819ecb26bc7610979af5a77f45/regex-2024.11.6-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:06eb1be98df10e81ebaded73fcd51989dcf534e3c753466e4b60c4697a003b67", size = 768726 }, - { url = "https://files.pythonhosted.org/packages/5a/db/f43fd75dc4c0c2d96d0881967897926942e935d700863666f3c844a72ce6/regex-2024.11.6-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:040df6fe1a5504eb0f04f048e6d09cd7c7110fef851d7c567a6b6e09942feb7d", size = 775098 }, - { url = "https://files.pythonhosted.org/packages/99/d7/f94154db29ab5a89d69ff893159b19ada89e76b915c1293e98603d39838c/regex-2024.11.6-cp310-cp310-musllinux_1_2_ppc64le.whl", hash = "sha256:fdabbfc59f2c6edba2a6622c647b716e34e8e3867e0ab975412c5c2f79b82da2", size = 839325 }, - { url = "https://files.pythonhosted.org/packages/f7/17/3cbfab1f23356fbbf07708220ab438a7efa1e0f34195bf857433f79f1788/regex-2024.11.6-cp310-cp310-musllinux_1_2_s390x.whl", hash = "sha256:8447d2d39b5abe381419319f942de20b7ecd60ce86f16a23b0698f22e1b70008", size = 843277 }, - { url = "https://files.pythonhosted.org/packages/7e/f2/48b393b51900456155de3ad001900f94298965e1cad1c772b87f9cfea011/regex-2024.11.6-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:da8f5fc57d1933de22a9e23eec290a0d8a5927a5370d24bda9a6abe50683fe62", size = 773197 }, - { url = "https://files.pythonhosted.org/packages/45/3f/ef9589aba93e084cd3f8471fded352826dcae8489b650d0b9b27bc5bba8a/regex-2024.11.6-cp310-cp310-win32.whl", hash = "sha256:b489578720afb782f6ccf2840920f3a32e31ba28a4b162e13900c3e6bd3f930e", size = 261714 }, - { url = "https://files.pythonhosted.org/packages/42/7e/5f1b92c8468290c465fd50c5318da64319133231415a8aa6ea5ab995a815/regex-2024.11.6-cp310-cp310-win_amd64.whl", hash = "sha256:5071b2093e793357c9d8b2929dfc13ac5f0a6c650559503bb81189d0a3814519", size = 274042 }, -] - -[[package]] -name = "requests" -version = "2.32.4" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "certifi" }, - { name = "charset-normalizer" }, - { name = "idna" }, - { name = "urllib3" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/e1/0a/929373653770d8a0d7ea76c37de6e41f11eb07559b103b1c02cafb3f7cf8/requests-2.32.4.tar.gz", hash = "sha256:27d0316682c8a29834d3264820024b62a36942083d52caf2f14c0591336d3422", size = 135258 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/7c/e4/56027c4a6b4ae70ca9de302488c5ca95ad4a39e190093d6c1a8ace08341b/requests-2.32.4-py3-none-any.whl", hash = "sha256:27babd3cda2a6d50b30443204ee89830707d396671944c998b5975b031ac2b2c", size = 64847 }, -] - -[[package]] -name = "resampy" -version = "0.4.3" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "numba" }, - { name = "numpy" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/29/f1/34be702a69a5d272e844c98cee82351f880985cfbca0cc86378011078497/resampy-0.4.3.tar.gz", hash = "sha256:a0d1c28398f0e55994b739650afef4e3974115edbe96cd4bb81968425e916e47", size = 3080604 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/4d/b9/3b00ac340a1aab3389ebcc52c779914a44aadf7b0cb7a3bf053195735607/resampy-0.4.3-py3-none-any.whl", hash = "sha256:ad2ed64516b140a122d96704e32bc0f92b23f45419e8b8f478e5a05f83edcebd", size = 3076529 }, -] - -[[package]] -name = "rfc3339-validator" -version = "0.1.4" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "six" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/28/ea/a9387748e2d111c3c2b275ba970b735e04e15cdb1eb30693b6b5708c4dbd/rfc3339_validator-0.1.4.tar.gz", hash = "sha256:138a2abdf93304ad60530167e51d2dfb9549521a836871b88d7f4695d0022f6b", size = 5513 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/7b/44/4e421b96b67b2daff264473f7465db72fbdf36a07e05494f50300cc7b0c6/rfc3339_validator-0.1.4-py2.py3-none-any.whl", hash = "sha256:24f6ec1eda14ef823da9e36ec7113124b39c04d50a4d3d3a3c2859577e7791fa", size = 3490 }, -] - -[[package]] -name = "rfc3986-validator" -version = "0.1.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/da/88/f270de456dd7d11dcc808abfa291ecdd3f45ff44e3b549ffa01b126464d0/rfc3986_validator-0.1.1.tar.gz", hash = "sha256:3d44bde7921b3b9ec3ae4e3adca370438eccebc676456449b145d533b240d055", size = 6760 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/9e/51/17023c0f8f1869d8806b979a2bffa3f861f26a3f1a66b094288323fba52f/rfc3986_validator-0.1.1-py2.py3-none-any.whl", hash = "sha256:2f235c432ef459970b4306369336b9d5dbdda31b510ca1e327636e01f528bfa9", size = 4242 }, -] - -[[package]] -name = "rich" -version = "14.0.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "markdown-it-py" }, - { name = "pygments" }, - { name = "typing-extensions", marker = "python_full_version == '3.10.12'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/a1/53/830aa4c3066a8ab0ae9a9955976fb770fe9c6102117c8ec4ab3ea62d89e8/rich-14.0.0.tar.gz", hash = "sha256:82f1bc23a6a21ebca4ae0c45af9bdbc492ed20231dcb63f297d6d1021a9d5725", size = 224078 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/0d/9b/63f4c7ebc259242c89b3acafdb37b41d1185c07ff0011164674e9076b491/rich-14.0.0-py3-none-any.whl", hash = "sha256:1c9491e1951aac09caffd42f448ee3d04e58923ffe14993f6e83068dc395d7e0", size = 243229 }, -] - -[[package]] -name = "rpds-py" -version = "0.25.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/8c/a6/60184b7fc00dd3ca80ac635dd5b8577d444c57e8e8742cecabfacb829921/rpds_py-0.25.1.tar.gz", hash = "sha256:8960b6dac09b62dac26e75d7e2c4a22efb835d827a7278c34f72b2b84fa160e3", size = 27304 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/cb/09/e1158988e50905b7f8306487a576b52d32aa9a87f79f7ab24ee8db8b6c05/rpds_py-0.25.1-cp310-cp310-macosx_10_12_x86_64.whl", hash = "sha256:f4ad628b5174d5315761b67f212774a32f5bad5e61396d38108bd801c0a8f5d9", size = 373140 }, - { url = "https://files.pythonhosted.org/packages/e0/4b/a284321fb3c45c02fc74187171504702b2934bfe16abab89713eedfe672e/rpds_py-0.25.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:8c742af695f7525e559c16f1562cf2323db0e3f0fbdcabdf6865b095256b2d40", size = 358860 }, - { url = "https://files.pythonhosted.org/packages/4e/46/8ac9811150c75edeae9fc6fa0e70376c19bc80f8e1f7716981433905912b/rpds_py-0.25.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:605ffe7769e24b1800b4d024d24034405d9404f0bc2f55b6db3362cd34145a6f", size = 386179 }, - { url = "https://files.pythonhosted.org/packages/f3/ec/87eb42d83e859bce91dcf763eb9f2ab117142a49c9c3d17285440edb5b69/rpds_py-0.25.1-cp310-cp310-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:ccc6f3ddef93243538be76f8e47045b4aad7a66a212cd3a0f23e34469473d36b", size = 400282 }, - { url = "https://files.pythonhosted.org/packages/68/c8/2a38e0707d7919c8c78e1d582ab15cf1255b380bcb086ca265b73ed6db23/rpds_py-0.25.1-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:f70316f760174ca04492b5ab01be631a8ae30cadab1d1081035136ba12738cfa", size = 521824 }, - { url = "https://files.pythonhosted.org/packages/5e/2c/6a92790243569784dde84d144bfd12bd45102f4a1c897d76375076d730ab/rpds_py-0.25.1-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:e1dafef8df605fdb46edcc0bf1573dea0d6d7b01ba87f85cd04dc855b2b4479e", size = 411644 }, - { url = "https://files.pythonhosted.org/packages/eb/76/66b523ffc84cf47db56efe13ae7cf368dee2bacdec9d89b9baca5e2e6301/rpds_py-0.25.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0701942049095741a8aeb298a31b203e735d1c61f4423511d2b1a41dcd8a16da", size = 386955 }, - { url = "https://files.pythonhosted.org/packages/b6/b9/a362d7522feaa24dc2b79847c6175daa1c642817f4a19dcd5c91d3e2c316/rpds_py-0.25.1-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:e87798852ae0b37c88babb7f7bbbb3e3fecc562a1c340195b44c7e24d403e380", size = 421039 }, - { url = "https://files.pythonhosted.org/packages/0f/c4/b5b6f70b4d719b6584716889fd3413102acf9729540ee76708d56a76fa97/rpds_py-0.25.1-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:3bcce0edc1488906c2d4c75c94c70a0417e83920dd4c88fec1078c94843a6ce9", size = 563290 }, - { url = "https://files.pythonhosted.org/packages/87/a3/2e6e816615c12a8f8662c9d8583a12eb54c52557521ef218cbe3095a8afa/rpds_py-0.25.1-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:e2f6a2347d3440ae789505693a02836383426249d5293541cd712e07e7aecf54", size = 592089 }, - { url = "https://files.pythonhosted.org/packages/c0/08/9b8e1050e36ce266135994e2c7ec06e1841f1c64da739daeb8afe9cb77a4/rpds_py-0.25.1-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:4fd52d3455a0aa997734f3835cbc4c9f32571345143960e7d7ebfe7b5fbfa3b2", size = 558400 }, - { url = "https://files.pythonhosted.org/packages/f2/df/b40b8215560b8584baccd839ff5c1056f3c57120d79ac41bd26df196da7e/rpds_py-0.25.1-cp310-cp310-win32.whl", hash = "sha256:3f0b1798cae2bbbc9b9db44ee068c556d4737911ad53a4e5093d09d04b3bbc24", size = 219741 }, - { url = "https://files.pythonhosted.org/packages/10/99/e4c58be18cf5d8b40b8acb4122bc895486230b08f978831b16a3916bd24d/rpds_py-0.25.1-cp310-cp310-win_amd64.whl", hash = "sha256:3ebd879ab996537fc510a2be58c59915b5dd63bccb06d1ef514fee787e05984a", size = 231553 }, - { url = "https://files.pythonhosted.org/packages/78/ff/566ce53529b12b4f10c0a348d316bd766970b7060b4fd50f888be3b3b281/rpds_py-0.25.1-pp310-pypy310_pp73-macosx_10_12_x86_64.whl", hash = "sha256:b24bf3cd93d5b6ecfbedec73b15f143596c88ee249fa98cefa9a9dc9d92c6f28", size = 373931 }, - { url = "https://files.pythonhosted.org/packages/83/5d/deba18503f7c7878e26aa696e97f051175788e19d5336b3b0e76d3ef9256/rpds_py-0.25.1-pp310-pypy310_pp73-macosx_11_0_arm64.whl", hash = "sha256:0eb90e94f43e5085623932b68840b6f379f26db7b5c2e6bcef3179bd83c9330f", size = 359074 }, - { url = "https://files.pythonhosted.org/packages/0d/74/313415c5627644eb114df49c56a27edba4d40cfd7c92bd90212b3604ca84/rpds_py-0.25.1-pp310-pypy310_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d50e4864498a9ab639d6d8854b25e80642bd362ff104312d9770b05d66e5fb13", size = 387255 }, - { url = "https://files.pythonhosted.org/packages/8c/c8/c723298ed6338963d94e05c0f12793acc9b91d04ed7c4ba7508e534b7385/rpds_py-0.25.1-pp310-pypy310_pp73-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:7c9409b47ba0650544b0bb3c188243b83654dfe55dcc173a86832314e1a6a35d", size = 400714 }, - { url = "https://files.pythonhosted.org/packages/33/8a/51f1f6aa653c2e110ed482ef2ae94140d56c910378752a1b483af11019ee/rpds_py-0.25.1-pp310-pypy310_pp73-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:796ad874c89127c91970652a4ee8b00d56368b7e00d3477f4415fe78164c8000", size = 523105 }, - { url = "https://files.pythonhosted.org/packages/c7/a4/7873d15c088ad3bff36910b29ceb0f178e4b3232c2adbe9198de68a41e63/rpds_py-0.25.1-pp310-pypy310_pp73-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:85608eb70a659bf4c1142b2781083d4b7c0c4e2c90eff11856a9754e965b2540", size = 411499 }, - { url = "https://files.pythonhosted.org/packages/90/f3/0ce1437befe1410766d11d08239333ac1b2d940f8a64234ce48a7714669c/rpds_py-0.25.1-pp310-pypy310_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c4feb9211d15d9160bc85fa72fed46432cdc143eb9cf6d5ca377335a921ac37b", size = 387918 }, - { url = "https://files.pythonhosted.org/packages/94/d4/5551247988b2a3566afb8a9dba3f1d4a3eea47793fd83000276c1a6c726e/rpds_py-0.25.1-pp310-pypy310_pp73-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:ccfa689b9246c48947d31dd9d8b16d89a0ecc8e0e26ea5253068efb6c542b76e", size = 421705 }, - { url = "https://files.pythonhosted.org/packages/b0/25/5960f28f847bf736cc7ee3c545a7e1d2f3b5edaf82c96fb616c2f5ed52d0/rpds_py-0.25.1-pp310-pypy310_pp73-musllinux_1_2_aarch64.whl", hash = "sha256:3c5b317ecbd8226887994852e85de562f7177add602514d4ac40f87de3ae45a8", size = 564489 }, - { url = "https://files.pythonhosted.org/packages/02/66/1c99884a0d44e8c2904d3c4ec302f995292d5dde892c3bf7685ac1930146/rpds_py-0.25.1-pp310-pypy310_pp73-musllinux_1_2_i686.whl", hash = "sha256:454601988aab2c6e8fd49e7634c65476b2b919647626208e376afcd22019eeb8", size = 592557 }, - { url = "https://files.pythonhosted.org/packages/55/ae/4aeac84ebeffeac14abb05b3bb1d2f728d00adb55d3fb7b51c9fa772e760/rpds_py-0.25.1-pp310-pypy310_pp73-musllinux_1_2_x86_64.whl", hash = "sha256:1c0c434a53714358532d13539272db75a5ed9df75a4a090a753ac7173ec14e11", size = 558691 }, - { url = "https://files.pythonhosted.org/packages/41/b3/728a08ff6f5e06fe3bb9af2e770e9d5fd20141af45cff8dfc62da4b2d0b3/rpds_py-0.25.1-pp310-pypy310_pp73-win_amd64.whl", hash = "sha256:f73ce1512e04fbe2bc97836e89830d6b4314c171587a99688082d090f934d20a", size = 231651 }, -] - -[[package]] -name = "ruamel-yaml" -version = "0.18.14" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "ruamel-yaml-clib", marker = "platform_python_implementation == 'CPython' and python_full_version == '3.10.12'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/39/87/6da0df742a4684263261c253f00edd5829e6aca970fff69e75028cccc547/ruamel.yaml-0.18.14.tar.gz", hash = "sha256:7227b76aaec364df15936730efbf7d72b30c0b79b1d578bbb8e3dcb2d81f52b7", size = 145511 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/af/6d/6fe4805235e193aad4aaf979160dd1f3c487c57d48b810c816e6e842171b/ruamel.yaml-0.18.14-py3-none-any.whl", hash = "sha256:710ff198bb53da66718c7db27eec4fbcc9aa6ca7204e4c1df2f282b6fe5eb6b2", size = 118570 }, -] - -[[package]] -name = "ruamel-yaml-clib" -version = "0.2.12" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/20/84/80203abff8ea4993a87d823a5f632e4d92831ef75d404c9fc78d0176d2b5/ruamel.yaml.clib-0.2.12.tar.gz", hash = "sha256:6c8fbb13ec503f99a91901ab46e0b07ae7941cd527393187039aec586fdfd36f", size = 225315 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/70/57/40a958e863e299f0c74ef32a3bde9f2d1ea8d69669368c0c502a0997f57f/ruamel.yaml.clib-0.2.12-cp310-cp310-macosx_13_0_arm64.whl", hash = "sha256:11f891336688faf5156a36293a9c362bdc7c88f03a8a027c2c1d8e0bcde998e5", size = 131301 }, - { url = "https://files.pythonhosted.org/packages/98/a8/29a3eb437b12b95f50a6bcc3d7d7214301c6c529d8fdc227247fa84162b5/ruamel.yaml.clib-0.2.12-cp310-cp310-manylinux2014_aarch64.whl", hash = "sha256:a606ef75a60ecf3d924613892cc603b154178ee25abb3055db5062da811fd969", size = 633728 }, - { url = "https://files.pythonhosted.org/packages/35/6d/ae05a87a3ad540259c3ad88d71275cbd1c0f2d30ae04c65dcbfb6dcd4b9f/ruamel.yaml.clib-0.2.12-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:fd5415dded15c3822597455bc02bcd66e81ef8b7a48cb71a33628fc9fdde39df", size = 722230 }, - { url = "https://files.pythonhosted.org/packages/7f/b7/20c6f3c0b656fe609675d69bc135c03aac9e3865912444be6339207b6648/ruamel.yaml.clib-0.2.12-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:f66efbc1caa63c088dead1c4170d148eabc9b80d95fb75b6c92ac0aad2437d76", size = 686712 }, - { url = "https://files.pythonhosted.org/packages/cd/11/d12dbf683471f888d354dac59593873c2b45feb193c5e3e0f2ebf85e68b9/ruamel.yaml.clib-0.2.12-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:22353049ba4181685023b25b5b51a574bce33e7f51c759371a7422dcae5402a6", size = 663936 }, - { url = "https://files.pythonhosted.org/packages/72/14/4c268f5077db5c83f743ee1daeb236269fa8577133a5cfa49f8b382baf13/ruamel.yaml.clib-0.2.12-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:932205970b9f9991b34f55136be327501903f7c66830e9760a8ffb15b07f05cd", size = 696580 }, - { url = "https://files.pythonhosted.org/packages/30/fc/8cd12f189c6405a4c1cf37bd633aa740a9538c8e40497c231072d0fef5cf/ruamel.yaml.clib-0.2.12-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:a52d48f4e7bf9005e8f0a89209bf9a73f7190ddf0489eee5eb51377385f59f2a", size = 663393 }, - { url = "https://files.pythonhosted.org/packages/80/29/c0a017b704aaf3cbf704989785cd9c5d5b8ccec2dae6ac0c53833c84e677/ruamel.yaml.clib-0.2.12-cp310-cp310-win32.whl", hash = "sha256:3eac5a91891ceb88138c113f9db04f3cebdae277f5d44eaa3651a4f573e6a5da", size = 100326 }, - { url = "https://files.pythonhosted.org/packages/3a/65/fa39d74db4e2d0cd252355732d966a460a41cd01c6353b820a0952432839/ruamel.yaml.clib-0.2.12-cp310-cp310-win_amd64.whl", hash = "sha256:ab007f2f5a87bd08ab1499bdf96f3d5c6ad4dcfa364884cb4549aa0154b13a28", size = 118079 }, -] - -[[package]] -name = "sacremoses" -version = "0.1.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "click" }, - { name = "joblib" }, - { name = "regex" }, - { name = "tqdm" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/1d/51/fbdc4af4f6e85d26169e28be3763fe50ddfd0d4bf8b871422b0788dcc4d2/sacremoses-0.1.1.tar.gz", hash = "sha256:b6fd5d3a766b02154ed80b962ddca91e1fd25629c0978c7efba21ebccf663934", size = 883188 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/0b/f0/89ee2bc9da434bd78464f288fdb346bc2932f2ee80a90b2a4bbbac262c74/sacremoses-0.1.1-py3-none-any.whl", hash = "sha256:31e04c98b169bfd902144824d191825cd69220cdb4ae4bcf1ec58a7db5587b1a", size = 897476 }, -] - -[[package]] -name = "safetensors" -version = "0.5.3" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/71/7e/2d5d6ee7b40c0682315367ec7475693d110f512922d582fef1bd4a63adc3/safetensors-0.5.3.tar.gz", hash = "sha256:b6b0d6ecacec39a4fdd99cc19f4576f5219ce858e6fd8dbe7609df0b8dc56965", size = 67210 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/18/ae/88f6c49dbd0cc4da0e08610019a3c78a7d390879a919411a410a1876d03a/safetensors-0.5.3-cp38-abi3-macosx_10_12_x86_64.whl", hash = "sha256:bd20eb133db8ed15b40110b7c00c6df51655a2998132193de2f75f72d99c7073", size = 436917 }, - { url = "https://files.pythonhosted.org/packages/b8/3b/11f1b4a2f5d2ab7da34ecc062b0bc301f2be024d110a6466726bec8c055c/safetensors-0.5.3-cp38-abi3-macosx_11_0_arm64.whl", hash = "sha256:21d01c14ff6c415c485616b8b0bf961c46b3b343ca59110d38d744e577f9cce7", size = 418419 }, - { url = "https://files.pythonhosted.org/packages/5d/9a/add3e6fef267658075c5a41573c26d42d80c935cdc992384dfae435feaef/safetensors-0.5.3-cp38-abi3-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:11bce6164887cd491ca75c2326a113ba934be596e22b28b1742ce27b1d076467", size = 459493 }, - { url = "https://files.pythonhosted.org/packages/df/5c/bf2cae92222513cc23b3ff85c4a1bb2811a2c3583ac0f8e8d502751de934/safetensors-0.5.3-cp38-abi3-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:4a243be3590bc3301c821da7a18d87224ef35cbd3e5f5727e4e0728b8172411e", size = 472400 }, - { url = "https://files.pythonhosted.org/packages/58/11/7456afb740bd45782d0f4c8e8e1bb9e572f1bf82899fb6ace58af47b4282/safetensors-0.5.3-cp38-abi3-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:8bd84b12b1670a6f8e50f01e28156422a2bc07fb16fc4e98bded13039d688a0d", size = 522891 }, - { url = "https://files.pythonhosted.org/packages/57/3d/fe73a9d2ace487e7285f6e157afee2383bd1ddb911b7cb44a55cf812eae3/safetensors-0.5.3-cp38-abi3-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:391ac8cab7c829452175f871fcaf414aa1e292b5448bd02620f675a7f3e7abb9", size = 537694 }, - { url = "https://files.pythonhosted.org/packages/a6/f8/dae3421624fcc87a89d42e1898a798bc7ff72c61f38973a65d60df8f124c/safetensors-0.5.3-cp38-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:cead1fa41fc54b1e61089fa57452e8834f798cb1dc7a09ba3524f1eb08e0317a", size = 471642 }, - { url = "https://files.pythonhosted.org/packages/ce/20/1fbe16f9b815f6c5a672f5b760951e20e17e43f67f231428f871909a37f6/safetensors-0.5.3-cp38-abi3-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:1077f3e94182d72618357b04b5ced540ceb71c8a813d3319f1aba448e68a770d", size = 502241 }, - { url = "https://files.pythonhosted.org/packages/5f/18/8e108846b506487aa4629fe4116b27db65c3dde922de2c8e0cc1133f3f29/safetensors-0.5.3-cp38-abi3-musllinux_1_2_aarch64.whl", hash = "sha256:799021e78287bac619c7b3f3606730a22da4cda27759ddf55d37c8db7511c74b", size = 638001 }, - { url = "https://files.pythonhosted.org/packages/82/5a/c116111d8291af6c8c8a8b40628fe833b9db97d8141c2a82359d14d9e078/safetensors-0.5.3-cp38-abi3-musllinux_1_2_armv7l.whl", hash = "sha256:df26da01aaac504334644e1b7642fa000bfec820e7cef83aeac4e355e03195ff", size = 734013 }, - { url = "https://files.pythonhosted.org/packages/7d/ff/41fcc4d3b7de837963622e8610d998710705bbde9a8a17221d85e5d0baad/safetensors-0.5.3-cp38-abi3-musllinux_1_2_i686.whl", hash = "sha256:32c3ef2d7af8b9f52ff685ed0bc43913cdcde135089ae322ee576de93eae5135", size = 670687 }, - { url = "https://files.pythonhosted.org/packages/40/ad/2b113098e69c985a3d8fbda4b902778eae4a35b7d5188859b4a63d30c161/safetensors-0.5.3-cp38-abi3-musllinux_1_2_x86_64.whl", hash = "sha256:37f1521be045e56fc2b54c606d4455573e717b2d887c579ee1dbba5f868ece04", size = 643147 }, - { url = "https://files.pythonhosted.org/packages/0a/0c/95aeb51d4246bd9a3242d3d8349c1112b4ee7611a4b40f0c5c93b05f001d/safetensors-0.5.3-cp38-abi3-win32.whl", hash = "sha256:cfc0ec0846dcf6763b0ed3d1846ff36008c6e7290683b61616c4b040f6a54ace", size = 296677 }, - { url = "https://files.pythonhosted.org/packages/69/e2/b011c38e5394c4c18fb5500778a55ec43ad6106126e74723ffaee246f56e/safetensors-0.5.3-cp38-abi3-win_amd64.whl", hash = "sha256:836cbbc320b47e80acd40e44c8682db0e8ad7123209f69b093def21ec7cafd11", size = 308878 }, -] - -[[package]] -name = "scikit-learn" -version = "1.5.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "joblib" }, - { name = "numpy" }, - { name = "scipy" }, - { name = "threadpoolctl" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/92/72/2961b9874a9ddf2b0f95f329d4e67f67c3301c1d88ba5e239ff25661bb85/scikit_learn-1.5.1.tar.gz", hash = "sha256:0ea5d40c0e3951df445721927448755d3fe1d80833b0b7308ebff5d2a45e6414", size = 6958368 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/f8/59/d8ea8c05e61d2afa988dfcfe47526595b531e94d23babf58d2e00a35f646/scikit_learn-1.5.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:781586c414f8cc58e71da4f3d7af311e0505a683e112f2f62919e3019abd3745", size = 12102257 }, - { url = "https://files.pythonhosted.org/packages/1f/c6/ba8e5691acca616adc8f0d6f8f5e79d55b927530aa404ee712b077acf0cf/scikit_learn-1.5.1-cp310-cp310-macosx_12_0_arm64.whl", hash = "sha256:f5b213bc29cc30a89a3130393b0e39c847a15d769d6e59539cd86b75d276b1a7", size = 10975310 }, - { url = "https://files.pythonhosted.org/packages/5c/c6/e362563cc7dfe37e4699cbf2b2d22c2854be227c254976de1c4854fc6e84/scikit_learn-1.5.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:1ff4ba34c2abff5ec59c803ed1d97d61b036f659a17f55be102679e88f926fac", size = 12496508 }, - { url = "https://files.pythonhosted.org/packages/f2/60/6c589c91e474721efdcec82ea9cc5c743359e52637e46c364ee5236666ef/scikit_learn-1.5.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:161808750c267b77b4a9603cf9c93579c7a74ba8486b1336034c2f1579546d21", size = 13352348 }, - { url = "https://files.pythonhosted.org/packages/f1/13/de29b945fb28fc0c24159d3a83f1250c5232c1c9abac12434c7c3447e9cc/scikit_learn-1.5.1-cp310-cp310-win_amd64.whl", hash = "sha256:10e49170691514a94bb2e03787aa921b82dbc507a4ea1f20fd95557862c98dc1", size = 10966250 }, -] - -[[package]] -name = "scipy" -version = "1.15.3" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "numpy" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/0f/37/6964b830433e654ec7485e45a00fc9a27cf868d622838f6b6d9c5ec0d532/scipy-1.15.3.tar.gz", hash = "sha256:eae3cf522bc7df64b42cad3925c876e1b0b6c35c1337c93e12c0f366f55b0eaf", size = 59419214 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/78/2f/4966032c5f8cc7e6a60f1b2e0ad686293b9474b65246b0c642e3ef3badd0/scipy-1.15.3-cp310-cp310-macosx_10_13_x86_64.whl", hash = "sha256:a345928c86d535060c9c2b25e71e87c39ab2f22fc96e9636bd74d1dbf9de448c", size = 38702770 }, - { url = "https://files.pythonhosted.org/packages/a0/6e/0c3bf90fae0e910c274db43304ebe25a6b391327f3f10b5dcc638c090795/scipy-1.15.3-cp310-cp310-macosx_12_0_arm64.whl", hash = "sha256:ad3432cb0f9ed87477a8d97f03b763fd1d57709f1bbde3c9369b1dff5503b253", size = 30094511 }, - { url = "https://files.pythonhosted.org/packages/ea/b1/4deb37252311c1acff7f101f6453f0440794f51b6eacb1aad4459a134081/scipy-1.15.3-cp310-cp310-macosx_14_0_arm64.whl", hash = "sha256:aef683a9ae6eb00728a542b796f52a5477b78252edede72b8327a886ab63293f", size = 22368151 }, - { url = "https://files.pythonhosted.org/packages/38/7d/f457626e3cd3c29b3a49ca115a304cebb8cc6f31b04678f03b216899d3c6/scipy-1.15.3-cp310-cp310-macosx_14_0_x86_64.whl", hash = "sha256:1c832e1bd78dea67d5c16f786681b28dd695a8cb1fb90af2e27580d3d0967e92", size = 25121732 }, - { url = "https://files.pythonhosted.org/packages/db/0a/92b1de4a7adc7a15dcf5bddc6e191f6f29ee663b30511ce20467ef9b82e4/scipy-1.15.3-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:263961f658ce2165bbd7b99fa5135195c3a12d9bef045345016b8b50c315cb82", size = 35547617 }, - { url = "https://files.pythonhosted.org/packages/8e/6d/41991e503e51fc1134502694c5fa7a1671501a17ffa12716a4a9151af3df/scipy-1.15.3-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9e2abc762b0811e09a0d3258abee2d98e0c703eee49464ce0069590846f31d40", size = 37662964 }, - { url = "https://files.pythonhosted.org/packages/25/e1/3df8f83cb15f3500478c889be8fb18700813b95e9e087328230b98d547ff/scipy-1.15.3-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:ed7284b21a7a0c8f1b6e5977ac05396c0d008b89e05498c8b7e8f4a1423bba0e", size = 37238749 }, - { url = "https://files.pythonhosted.org/packages/93/3e/b3257cf446f2a3533ed7809757039016b74cd6f38271de91682aa844cfc5/scipy-1.15.3-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:5380741e53df2c566f4d234b100a484b420af85deb39ea35a1cc1be84ff53a5c", size = 40022383 }, - { url = "https://files.pythonhosted.org/packages/d1/84/55bc4881973d3f79b479a5a2e2df61c8c9a04fcb986a213ac9c02cfb659b/scipy-1.15.3-cp310-cp310-win_amd64.whl", hash = "sha256:9d61e97b186a57350f6d6fd72640f9e99d5a4a2b8fbf4b9ee9a841eab327dc13", size = 41259201 }, -] - -[[package]] -name = "seaborn" -version = "0.13.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "matplotlib" }, - { name = "numpy" }, - { name = "pandas" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/86/59/a451d7420a77ab0b98f7affa3a1d78a313d2f7281a57afb1a34bae8ab412/seaborn-0.13.2.tar.gz", hash = "sha256:93e60a40988f4d65e9f4885df477e2fdaff6b73a9ded434c1ab356dd57eefff7", size = 1457696 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/83/11/00d3c3dfc25ad54e731d91449895a79e4bf2384dc3ac01809010ba88f6d5/seaborn-0.13.2-py3-none-any.whl", hash = "sha256:636f8336facf092165e27924f223d3c62ca560b1f2bb5dff7ab7fad265361987", size = 294914 }, -] - -[[package]] -name = "semver" -version = "3.0.4" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/72/d1/d3159231aec234a59dd7d601e9dd9fe96f3afff15efd33c1070019b26132/semver-3.0.4.tar.gz", hash = "sha256:afc7d8c584a5ed0a11033af086e8af226a9c0b206f313e0301f8dd7b6b589602", size = 269730 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/a6/24/4d91e05817e92e3a61c8a21e08fd0f390f5301f1c448b137c57c4bc6e543/semver-3.0.4-py3-none-any.whl", hash = "sha256:9c824d87ba7f7ab4a1890799cec8596f15c1241cb473404ea1cb0c55e4b04746", size = 17912 }, -] - -[[package]] -name = "send2trash" -version = "1.8.3" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/fd/3a/aec9b02217bb79b87bbc1a21bc6abc51e3d5dcf65c30487ac96c0908c722/Send2Trash-1.8.3.tar.gz", hash = "sha256:b18e7a3966d99871aefeb00cfbcfdced55ce4871194810fc71f4aa484b953abf", size = 17394 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/40/b0/4562db6223154aa4e22f939003cb92514c79f3d4dccca3444253fd17f902/Send2Trash-1.8.3-py3-none-any.whl", hash = "sha256:0c31227e0bd08961c7665474a3d1ef7193929fedda4233843689baa056be46c9", size = 18072 }, -] - -[[package]] -name = "sentencepiece" -version = "0.2.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/c9/d2/b9c7ca067c26d8ff085d252c89b5f69609ca93fb85a00ede95f4857865d4/sentencepiece-0.2.0.tar.gz", hash = "sha256:a52c19171daaf2e697dc6cbe67684e0fa341b1248966f6aebb541de654d15843", size = 2632106 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/f6/71/98648c3b64b23edb5403f74bcc906ad21766872a6e1ada26ea3f1eb941ab/sentencepiece-0.2.0-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:188779e1298a1c8b8253c7d3ad729cb0a9891e5cef5e5d07ce4592c54869e227", size = 2408979 }, - { url = "https://files.pythonhosted.org/packages/77/9f/7efbaa6d4c0c718a9affbecc536b03ca62f99f421bdffb531c16030e2d2b/sentencepiece-0.2.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:bed9cf85b296fa2b76fc2547b9cbb691a523864cebaee86304c43a7b4cb1b452", size = 1238845 }, - { url = "https://files.pythonhosted.org/packages/1c/e4/c2541027a43ec6962ba9b601805d17ba3f86b38bdeae0e8ac65a2981e248/sentencepiece-0.2.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:d7b67e724bead13f18db6e1d10b6bbdc454af574d70efbb36f27d90387be1ca3", size = 1181472 }, - { url = "https://files.pythonhosted.org/packages/fd/46/316c1ba6c52b97de76aff7b9da678f7afbb52136afb2987c474d95630e65/sentencepiece-0.2.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:2fde4b08cfe237be4484c6c7c2e2c75fb862cfeab6bd5449ce4caeafd97b767a", size = 1259151 }, - { url = "https://files.pythonhosted.org/packages/aa/5a/3c48738a0835d76dd06c62b6ac48d39c923cde78dd0f587353bdcbb99851/sentencepiece-0.2.0-cp310-cp310-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:4c378492056202d1c48a4979650981635fd97875a00eabb1f00c6a236b013b5e", size = 1355931 }, - { url = "https://files.pythonhosted.org/packages/a6/27/33019685023221ca8ed98e8ceb7ae5e166032686fa3662c68f1f1edf334e/sentencepiece-0.2.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1380ce6540a368de2ef6d7e6ba14ba8f3258df650d39ba7d833b79ee68a52040", size = 1301537 }, - { url = "https://files.pythonhosted.org/packages/ca/e4/55f97cef14293171fef5f96e96999919ab5b4d1ce95b53547ad653d7e3bf/sentencepiece-0.2.0-cp310-cp310-win32.whl", hash = "sha256:a1151d6a6dd4b43e552394aed0edfe9292820272f0194bd56c7c1660a0c06c3d", size = 936747 }, - { url = "https://files.pythonhosted.org/packages/85/f4/4ef1a6e0e9dbd8a60780a91df8b7452ada14cfaa0e17b3b8dfa42cecae18/sentencepiece-0.2.0-cp310-cp310-win_amd64.whl", hash = "sha256:d490142b0521ef22bc1085f061d922a2a6666175bb6b42e588ff95c0db6819b2", size = 991525 }, -] - -[[package]] -name = "sentry-sdk" -version = "2.32.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "certifi" }, - { name = "urllib3" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/10/59/eb90c45cb836cf8bec973bba10230ddad1c55e2b2e9ffa9d7d7368948358/sentry_sdk-2.32.0.tar.gz", hash = "sha256:9016c75d9316b0f6921ac14c8cd4fb938f26002430ac5be9945ab280f78bec6b", size = 334932 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/01/a1/fc4856bd02d2097324fb7ce05b3021fb850f864b83ca765f6e37e92ff8ca/sentry_sdk-2.32.0-py2.py3-none-any.whl", hash = "sha256:6cf51521b099562d7ce3606da928c473643abe99b00ce4cb5626ea735f4ec345", size = 356122 }, -] - -[[package]] -name = "setproctitle" -version = "1.3.6" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/9e/af/56efe21c53ac81ac87e000b15e60b3d8104224b4313b6eacac3597bd183d/setproctitle-1.3.6.tar.gz", hash = "sha256:c9f32b96c700bb384f33f7cf07954bb609d35dd82752cef57fb2ee0968409169", size = 26889 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/7d/db/8214810cae49e2e474ea741aaa7d6111486f27377e864f0eb6d297c9be56/setproctitle-1.3.6-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:ebcf34b69df4ca0eabaaaf4a3d890f637f355fed00ba806f7ebdd2d040658c26", size = 17412 }, - { url = "https://files.pythonhosted.org/packages/a4/45/909b0dcd68b16d2e58de0e861c0c0b67748ccc87ff9b59136e9710b528b1/setproctitle-1.3.6-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:1aa1935aa2195b76f377e5cb018290376b7bf085f0b53f5a95c0c21011b74367", size = 11966 }, - { url = "https://files.pythonhosted.org/packages/8a/f4/f1cd54fedae1cdacf1d1db833dc096bfb1f029451f60e68563e4c26ed2f7/setproctitle-1.3.6-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:13624d9925bb481bc0ccfbc7f533da38bfbfe6e80652314f789abc78c2e513bd", size = 31350 }, - { url = "https://files.pythonhosted.org/packages/5a/5f/f159b22d286a349633d4090090b8e6632fb84575a64f189b68e70a613c65/setproctitle-1.3.6-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:97a138fa875c6f281df7720dac742259e85518135cd0e3551aba1c628103d853", size = 32704 }, - { url = "https://files.pythonhosted.org/packages/9c/25/e5ea2673d951dafc04b6544d7b33dd9283733d715c8f40e93d39ae35d6a0/setproctitle-1.3.6-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:c86e9e82bfab579327dbe9b82c71475165fbc8b2134d24f9a3b2edaf200a5c3d", size = 29833 }, - { url = "https://files.pythonhosted.org/packages/67/2b/c3cbd4a4462c1143465d8c151f1d51bbfb418e60a96a754329d28d416575/setproctitle-1.3.6-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6af330ddc2ec05a99c3933ab3cba9365357c0b8470a7f2fa054ee4b0984f57d1", size = 30884 }, - { url = "https://files.pythonhosted.org/packages/27/04/b43a622a9fbf0f216a50b523067d3b07739ede2106a7226223e33abf6659/setproctitle-1.3.6-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:109fc07b1cd6cef9c245b2028e3e98e038283342b220def311d0239179810dbe", size = 30798 }, - { url = "https://files.pythonhosted.org/packages/41/60/8eb197b56b0a3110eef2a1d2ddb61b3f5809dbab9d975aa40c86e5d4b312/setproctitle-1.3.6-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:7df5fcc48588f82b6cc8073db069609ddd48a49b1e9734a20d0efb32464753c4", size = 29758 }, - { url = "https://files.pythonhosted.org/packages/db/1d/c394322a5425c12f4ada0696eb6d194768752d4e3acaca0c9d593025feb4/setproctitle-1.3.6-cp310-cp310-musllinux_1_2_ppc64le.whl", hash = "sha256:2407955dc359d735a20ac6e797ad160feb33d529a2ac50695c11a1ec680eafab", size = 32157 }, - { url = "https://files.pythonhosted.org/packages/e7/24/ce080682b144f057814efbe95daac09149e90f7689e2515897817a941686/setproctitle-1.3.6-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:38ca045626af693da042ac35d7332e7b9dbd52e6351d6973b310612e3acee6d6", size = 30291 }, - { url = "https://files.pythonhosted.org/packages/a7/4f/4db265493567865428dcec376f8142a9c65d27c10c3ac915d173b4053afb/setproctitle-1.3.6-cp310-cp310-win32.whl", hash = "sha256:9483aa336687463f5497dd37a070094f3dff55e2c888994f8440fcf426a1a844", size = 11492 }, - { url = "https://files.pythonhosted.org/packages/38/b0/64c3948f7957db44b4c5edfe9c197de493144dc915ddf95cf36aeca0dc52/setproctitle-1.3.6-cp310-cp310-win_amd64.whl", hash = "sha256:4efc91b437f6ff2578e89e3f17d010c0a0ff01736606473d082913ecaf7859ba", size = 12204 }, - { url = "https://files.pythonhosted.org/packages/d0/2b/f19977b646b64c1440dade2c385c89c39f74ce5254defa102dfd9c163e0b/setproctitle-1.3.6-pp310-pypy310_pp73-macosx_11_0_arm64.whl", hash = "sha256:3cde5b83ec4915cd5e6ae271937fd60d14113c8f7769b4a20d51769fe70d8717", size = 11471 }, - { url = "https://files.pythonhosted.org/packages/78/46/db58cf700f1408cf0f63d3f939f7b077bd450da8e037310f70e74c41262f/setproctitle-1.3.6-pp310-pypy310_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:797f2846b546a8741413c57d9fb930ad5aa939d925c9c0fa6186d77580035af7", size = 13520 }, - { url = "https://files.pythonhosted.org/packages/5c/46/0b89e7ebe77543e721c67077ad93fc8c7c3c31a8db3b12e00d02950f6adc/setproctitle-1.3.6-pp310-pypy310_pp73-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4ac3eb04bcf0119aadc6235a2c162bae5ed5f740e3d42273a7228b915722de20", size = 13094 }, - { url = "https://files.pythonhosted.org/packages/f7/78/03f2e42eb83bce6f853d7751ae95f8a3ae7408145a0b6cdd717be01497d7/setproctitle-1.3.6-pp310-pypy310_pp73-win_amd64.whl", hash = "sha256:0e6b5633c94c5111f7137f875e8f1ff48f53b991d5d5b90932f27dc8c1fa9ae4", size = 12241 }, -] - -[[package]] -name = "setuptools" -version = "80.9.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/18/5d/3bf57dcd21979b887f014ea83c24ae194cfcd12b9e0fda66b957c69d1fca/setuptools-80.9.0.tar.gz", hash = "sha256:f36b47402ecde768dbfafc46e8e4207b4360c654f1f3bb84475f0a28628fb19c", size = 1319958 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/a3/dc/17031897dae0efacfea57dfd3a82fdd2a2aeb58e0ff71b77b87e44edc772/setuptools-80.9.0-py3-none-any.whl", hash = "sha256:062d34222ad13e0cc312a4c02d73f059e86a4acbfbdea8f8f76b28c99f306922", size = 1201486 }, -] - -[[package]] -name = "shellingham" -version = "1.5.4" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/58/15/8b3609fd3830ef7b27b655beb4b4e9c62313a4e8da8c676e142cc210d58e/shellingham-1.5.4.tar.gz", hash = "sha256:8dbca0739d487e5bd35ab3ca4b36e11c4078f3a234bfce294b0a0291363404de", size = 10310 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/e0/f9/0595336914c5619e5f28a1fb793285925a8cd4b432c9da0a987836c7f822/shellingham-1.5.4-py2.py3-none-any.whl", hash = "sha256:7ecfff8f2fd72616f7481040475a65b2bf8af90a56c89140852d1120324e8686", size = 9755 }, -] - -[[package]] -name = "six" -version = "1.17.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/94/e7/b2c673351809dca68a0e064b6af791aa332cf192da575fd474ed7d6f16a2/six-1.17.0.tar.gz", hash = "sha256:ff70335d468e7eb6ec65b95b99d3a2836546063f63acc5171de367e834932a81", size = 34031 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/b7/ce/149a00dd41f10bc29e5921b496af8b574d8413afcd5e30dfa0ed46c2cc5e/six-1.17.0-py2.py3-none-any.whl", hash = "sha256:4721f391ed90541fddacab5acf947aa0d3dc7d27b2e1e8eda2be8970586c3274", size = 11050 }, -] - -[[package]] -name = "smmap" -version = "5.0.2" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/44/cd/a040c4b3119bbe532e5b0732286f805445375489fceaec1f48306068ee3b/smmap-5.0.2.tar.gz", hash = "sha256:26ea65a03958fa0c8a1c7e8c7a58fdc77221b8910f6be2131affade476898ad5", size = 22329 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/04/be/d09147ad1ec7934636ad912901c5fd7667e1c858e19d355237db0d0cd5e4/smmap-5.0.2-py3-none-any.whl", hash = "sha256:b30115f0def7d7531d22a0fb6502488d879e75b260a9db4d0819cfb25403af5e", size = 24303 }, -] - -[[package]] -name = "sniffio" -version = "1.3.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/a2/87/a6771e1546d97e7e041b6ae58d80074f81b7d5121207425c964ddf5cfdbd/sniffio-1.3.1.tar.gz", hash = "sha256:f4324edc670a0f49750a81b895f35c3adb843cca46f0530f79fc1babb23789dc", size = 20372 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/e9/44/75a9c9421471a6c4805dbf2356f7c181a29c1879239abab1ea2cc8f38b40/sniffio-1.3.1-py3-none-any.whl", hash = "sha256:2f6da418d1f1e0fddd844478f41680e794e6051915791a034ff65e5f100525a2", size = 10235 }, -] - -[[package]] -name = "sortedcontainers" -version = "2.4.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/e8/c4/ba2f8066cceb6f23394729afe52f3bf7adec04bf9ed2c820b39e19299111/sortedcontainers-2.4.0.tar.gz", hash = "sha256:25caa5a06cc30b6b83d11423433f65d1f9d76c4c6a0c90e3379eaa43b9bfdb88", size = 30594 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/32/46/9cb0e58b2deb7f82b84065f37f3bffeb12413f947f9388e4cac22c4621ce/sortedcontainers-2.4.0-py2.py3-none-any.whl", hash = "sha256:a163dcaede0f1c021485e957a39245190e74249897e2ae4b2aa38595db237ee0", size = 29575 }, -] - -[[package]] -name = "soundfile" -version = "0.13.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "cffi" }, - { name = "numpy" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/e1/41/9b873a8c055582859b239be17902a85339bec6a30ad162f98c9b0288a2cc/soundfile-0.13.1.tar.gz", hash = "sha256:b2c68dab1e30297317080a5b43df57e302584c49e2942defdde0acccc53f0e5b", size = 46156 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/64/28/e2a36573ccbcf3d57c00626a21fe51989380636e821b341d36ccca0c1c3a/soundfile-0.13.1-py2.py3-none-any.whl", hash = "sha256:a23c717560da2cf4c7b5ae1142514e0fd82d6bbd9dfc93a50423447142f2c445", size = 25751 }, - { url = "https://files.pythonhosted.org/packages/ea/ab/73e97a5b3cc46bba7ff8650a1504348fa1863a6f9d57d7001c6b67c5f20e/soundfile-0.13.1-py2.py3-none-macosx_10_9_x86_64.whl", hash = "sha256:82dc664d19831933fe59adad199bf3945ad06d84bc111a5b4c0d3089a5b9ec33", size = 1142250 }, - { url = "https://files.pythonhosted.org/packages/a0/e5/58fd1a8d7b26fc113af244f966ee3aecf03cb9293cb935daaddc1e455e18/soundfile-0.13.1-py2.py3-none-macosx_11_0_arm64.whl", hash = "sha256:743f12c12c4054921e15736c6be09ac26b3b3d603aef6fd69f9dde68748f2593", size = 1101406 }, - { url = "https://files.pythonhosted.org/packages/58/ae/c0e4a53d77cf6e9a04179535766b3321b0b9ced5f70522e4caf9329f0046/soundfile-0.13.1-py2.py3-none-manylinux_2_28_aarch64.whl", hash = "sha256:9c9e855f5a4d06ce4213f31918653ab7de0c5a8d8107cd2427e44b42df547deb", size = 1235729 }, - { url = "https://files.pythonhosted.org/packages/57/5e/70bdd9579b35003a489fc850b5047beeda26328053ebadc1fb60f320f7db/soundfile-0.13.1-py2.py3-none-manylinux_2_28_x86_64.whl", hash = "sha256:03267c4e493315294834a0870f31dbb3b28a95561b80b134f0bd3cf2d5f0e618", size = 1313646 }, - { url = "https://files.pythonhosted.org/packages/fe/df/8c11dc4dfceda14e3003bb81a0d0edcaaf0796dd7b4f826ea3e532146bba/soundfile-0.13.1-py2.py3-none-win32.whl", hash = "sha256:c734564fab7c5ddf8e9be5bf70bab68042cd17e9c214c06e365e20d64f9a69d5", size = 899881 }, - { url = "https://files.pythonhosted.org/packages/14/e9/6b761de83277f2f02ded7e7ea6f07828ec78e4b229b80e4ca55dd205b9dc/soundfile-0.13.1-py2.py3-none-win_amd64.whl", hash = "sha256:1e70a05a0626524a69e9f0f4dd2ec174b4e9567f4d8b6c11d38b5c289be36ee9", size = 1019162 }, -] - -[[package]] -name = "soupsieve" -version = "2.7" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/3f/f4/4a80cd6ef364b2e8b65b15816a843c0980f7a5a2b4dc701fc574952aa19f/soupsieve-2.7.tar.gz", hash = "sha256:ad282f9b6926286d2ead4750552c8a6142bc4c783fd66b0293547c8fe6ae126a", size = 103418 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/e7/9c/0e6afc12c269578be5c0c1c9f4b49a8d32770a080260c333ac04cc1c832d/soupsieve-2.7-py3-none-any.whl", hash = "sha256:6e60cc5c1ffaf1cebcc12e8188320b72071e922c2e897f737cadce79ad5d30c4", size = 36677 }, -] - -[[package]] -name = "sox" -version = "1.5.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "numpy" }, - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/3d/a2/d8e0d8fd7abf509ead4a2cb0fb24e5758b5330166bf9223d5cb9f98a7e8d/sox-1.5.0.tar.gz", hash = "sha256:12c7be5bb1f548d891fe11e82c08cf5f1a1d74e225298f60082e5aeb2469ada0", size = 63905 } - -[[package]] -name = "soxr" -version = "0.5.0.post1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "numpy" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/02/c0/4429bf9b3be10e749149e286aa5c53775399ec62891c6b970456c6dca325/soxr-0.5.0.post1.tar.gz", hash = "sha256:7092b9f3e8a416044e1fa138c8172520757179763b85dc53aa9504f4813cff73", size = 170853 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/7d/96/bee1eb69d66fc28c3b219ba9b8674b49d3dcc6cd2f9b3e5114ff28cf88b5/soxr-0.5.0.post1-cp310-cp310-macosx_10_14_x86_64.whl", hash = "sha256:7406d782d85f8cf64e66b65e6b7721973de8a1dc50b9e88bc2288c343a987484", size = 203841 }, - { url = "https://files.pythonhosted.org/packages/1f/5d/56ad3d181d30d103128f65cc44f4c4e24c199e6d5723e562704e47c89f78/soxr-0.5.0.post1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:fa0a382fb8d8e2afed2c1642723b2d2d1b9a6728ff89f77f3524034c8885b8c9", size = 160192 }, - { url = "https://files.pythonhosted.org/packages/7f/09/e43c39390e26b4c1b8d46f8a1c252a5077fa9f81cc2326b03c3d2b85744e/soxr-0.5.0.post1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8b01d3efb95a2851f78414bcd00738b0253eec3f5a1e5482838e965ffef84969", size = 221176 }, - { url = "https://files.pythonhosted.org/packages/ba/e6/059070b4cdb7fdd8ffbb67c5087c1da9716577127fb0540cd11dbf77923b/soxr-0.5.0.post1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:fcc049b0a151a65aa75b92f0ac64bb2dba785d16b78c31c2b94e68c141751d6d", size = 252779 }, - { url = "https://files.pythonhosted.org/packages/ad/64/86082b6372e5ff807dfa79b857da9f50e94e155706000daa43fdc3b59851/soxr-0.5.0.post1-cp310-cp310-win_amd64.whl", hash = "sha256:97f269bc26937c267a2ace43a77167d0c5c8bba5a2b45863bb6042b5b50c474e", size = 166881 }, - { url = "https://files.pythonhosted.org/packages/5d/e3/d422d279e51e6932e7b64f1170a4f61a7ee768e0f84c9233a5b62cd2c832/soxr-0.5.0.post1-cp312-abi3-macosx_10_14_x86_64.whl", hash = "sha256:fef509466c9c25f65eae0ce1e4b9ac9705d22c6038c914160ddaf459589c6e31", size = 199993 }, - { url = "https://files.pythonhosted.org/packages/20/f1/88adaca3c52e03bcb66b63d295df2e2d35bf355d19598c6ce84b20be7fca/soxr-0.5.0.post1-cp312-abi3-macosx_11_0_arm64.whl", hash = "sha256:4704ba6b13a3f1e41d12acf192878384c1c31f71ce606829c64abdf64a8d7d32", size = 156373 }, - { url = "https://files.pythonhosted.org/packages/b8/38/bad15a9e615215c8219652ca554b601663ac3b7ac82a284aca53ec2ff48c/soxr-0.5.0.post1-cp312-abi3-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:bd052a66471a7335b22a6208601a9d0df7b46b8d087dce4ff6e13eed6a33a2a1", size = 216564 }, - { url = "https://files.pythonhosted.org/packages/e1/1a/569ea0420a0c4801c2c8dd40d8d544989522f6014d51def689125f3f2935/soxr-0.5.0.post1-cp312-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a3f16810dd649ab1f433991d2a9661e9e6a116c2b4101039b53b3c3e90a094fc", size = 248455 }, - { url = "https://files.pythonhosted.org/packages/bc/10/440f1ba3d4955e0dc740bbe4ce8968c254a3d644d013eb75eea729becdb8/soxr-0.5.0.post1-cp312-abi3-win_amd64.whl", hash = "sha256:b1be9fee90afb38546bdbd7bde714d1d9a8c5a45137f97478a83b65e7f3146f6", size = 164937 }, -] - -[[package]] -name = "speechbrain" -version = "1.0.3" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "huggingface-hub" }, - { name = "hyperpyyaml" }, - { name = "joblib" }, - { name = "numpy" }, - { name = "packaging" }, - { name = "scipy" }, - { name = "sentencepiece" }, - { name = "torch" }, - { name = "torchaudio" }, - { name = "tqdm" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/ab/10/87e666544a4e0cec7cbdc09f26948994831ae0f8bbc58de3bf53b68285ff/speechbrain-1.0.3.tar.gz", hash = "sha256:fcab3c6e90012cecb1eed40ea235733b550137e73da6bfa2340ba191ec714052", size = 747735 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/58/13/e61f1085aebee17d5fc2df19fcc5177c10379be52578afbecdd615a831c9/speechbrain-1.0.3-py3-none-any.whl", hash = "sha256:9859d4c1b1fb3af3b85523c0c89f52e45a04f305622ed55f31aa32dd2fba19e9", size = 864091 }, -] - -[[package]] -name = "sqlalchemy" -version = "2.0.41" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "greenlet", marker = "(platform_machine == 'AMD64' and python_full_version == '3.10.12') or (platform_machine == 'WIN32' and python_full_version == '3.10.12') or (platform_machine == 'aarch64' and python_full_version == '3.10.12') or (platform_machine == 'amd64' and python_full_version == '3.10.12') or (platform_machine == 'ppc64le' and python_full_version == '3.10.12') or (platform_machine == 'win32' and python_full_version == '3.10.12') or (platform_machine == 'x86_64' and python_full_version == '3.10.12')" }, - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/63/66/45b165c595ec89aa7dcc2c1cd222ab269bc753f1fc7a1e68f8481bd957bf/sqlalchemy-2.0.41.tar.gz", hash = "sha256:edba70118c4be3c2b1f90754d308d0b79c6fe2c0fdc52d8ddf603916f83f4db9", size = 9689424 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/e9/12/d7c445b1940276a828efce7331cb0cb09d6e5f049651db22f4ebb0922b77/sqlalchemy-2.0.41-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:b1f09b6821406ea1f94053f346f28f8215e293344209129a9c0fcc3578598d7b", size = 2117967 }, - { url = "https://files.pythonhosted.org/packages/6f/b8/cb90f23157e28946b27eb01ef401af80a1fab7553762e87df51507eaed61/sqlalchemy-2.0.41-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:1936af879e3db023601196a1684d28e12f19ccf93af01bf3280a3262c4b6b4e5", size = 2107583 }, - { url = "https://files.pythonhosted.org/packages/9e/c2/eef84283a1c8164a207d898e063edf193d36a24fb6a5bb3ce0634b92a1e8/sqlalchemy-2.0.41-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:b2ac41acfc8d965fb0c464eb8f44995770239668956dc4cdf502d1b1ffe0d747", size = 3186025 }, - { url = "https://files.pythonhosted.org/packages/bd/72/49d52bd3c5e63a1d458fd6d289a1523a8015adedbddf2c07408ff556e772/sqlalchemy-2.0.41-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:81c24e0c0fde47a9723c81d5806569cddef103aebbf79dbc9fcbb617153dea30", size = 3186259 }, - { url = "https://files.pythonhosted.org/packages/4f/9e/e3ffc37d29a3679a50b6bbbba94b115f90e565a2b4545abb17924b94c52d/sqlalchemy-2.0.41-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:23a8825495d8b195c4aa9ff1c430c28f2c821e8c5e2d98089228af887e5d7e29", size = 3126803 }, - { url = "https://files.pythonhosted.org/packages/8a/76/56b21e363f6039978ae0b72690237b38383e4657281285a09456f313dd77/sqlalchemy-2.0.41-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:60c578c45c949f909a4026b7807044e7e564adf793537fc762b2489d522f3d11", size = 3148566 }, - { url = "https://files.pythonhosted.org/packages/3b/92/11b8e1b69bf191bc69e300a99badbbb5f2f1102f2b08b39d9eee2e21f565/sqlalchemy-2.0.41-cp310-cp310-win32.whl", hash = "sha256:118c16cd3f1b00c76d69343e38602006c9cfb9998fa4f798606d28d63f23beda", size = 2086696 }, - { url = "https://files.pythonhosted.org/packages/5c/88/2d706c9cc4502654860f4576cd54f7db70487b66c3b619ba98e0be1a4642/sqlalchemy-2.0.41-cp310-cp310-win_amd64.whl", hash = "sha256:7492967c3386df69f80cf67efd665c0f667cee67032090fe01d7d74b0e19bb08", size = 2110200 }, - { url = "https://files.pythonhosted.org/packages/1c/fc/9ba22f01b5cdacc8f5ed0d22304718d2c758fce3fd49a5372b886a86f37c/sqlalchemy-2.0.41-py3-none-any.whl", hash = "sha256:57df5dc6fdb5ed1a88a1ed2195fd31927e705cad62dedd86b46972752a80f576", size = 1911224 }, -] - -[[package]] -name = "stack-data" -version = "0.6.3" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "asttokens" }, - { name = "executing" }, - { name = "pure-eval" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/28/e3/55dcc2cfbc3ca9c29519eb6884dd1415ecb53b0e934862d3559ddcb7e20b/stack_data-0.6.3.tar.gz", hash = "sha256:836a778de4fec4dcd1dcd89ed8abff8a221f58308462e1c4aa2a3cf30148f0b9", size = 44707 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/f1/7b/ce1eafaf1a76852e2ec9b22edecf1daa58175c090266e9f6c64afcd81d91/stack_data-0.6.3-py3-none-any.whl", hash = "sha256:d5558e0c25a4cb0853cddad3d77da9891a08cb85dd9f9f91b9f8cd66e511e695", size = 24521 }, -] - -[[package]] -name = "sympy" -version = "1.14.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "mpmath" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/83/d3/803453b36afefb7c2bb238361cd4ae6125a569b4db67cd9e79846ba2d68c/sympy-1.14.0.tar.gz", hash = "sha256:d3d3fe8df1e5a0b42f0e7bdf50541697dbe7d23746e894990c030e2b05e72517", size = 7793921 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/a2/09/77d55d46fd61b4a135c444fc97158ef34a095e5681d0a6c10b75bf356191/sympy-1.14.0-py3-none-any.whl", hash = "sha256:e091cc3e99d2141a0ba2847328f5479b05d94a6635cb96148ccb3f34671bd8f5", size = 6299353 }, -] - -[[package]] -name = "tabulate" -version = "0.9.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/ec/fe/802052aecb21e3797b8f7902564ab6ea0d60ff8ca23952079064155d1ae1/tabulate-0.9.0.tar.gz", hash = "sha256:0095b12bf5966de529c0feb1fa08671671b3368eec77d7ef7ab114be2c068b3c", size = 81090 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/40/44/4a5f08c96eb108af5cb50b41f76142f0afa346dfa99d5296fe7202a11854/tabulate-0.9.0-py3-none-any.whl", hash = "sha256:024ca478df22e9340661486f85298cff5f6dcdba14f3813e8830015b9ed1948f", size = 35252 }, -] - -[[package]] -name = "tensorboard" -version = "2.19.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "absl-py" }, - { name = "grpcio" }, - { name = "markdown" }, - { name = "numpy" }, - { name = "packaging" }, - { name = "protobuf" }, - { name = "setuptools" }, - { name = "six" }, - { name = "tensorboard-data-server" }, - { name = "werkzeug" }, -] -wheels = [ - { url = "https://files.pythonhosted.org/packages/5d/12/4f70e8e2ba0dbe72ea978429d8530b0333f0ed2140cc571a48802878ef99/tensorboard-2.19.0-py3-none-any.whl", hash = "sha256:5e71b98663a641a7ce8a6e70b0be8e1a4c0c45d48760b076383ac4755c35b9a0", size = 5503412 }, -] - -[[package]] -name = "tensorboard-data-server" -version = "0.7.2" -source = { registry = "https://pypi.org/simple" } -wheels = [ - { url = "https://files.pythonhosted.org/packages/7a/13/e503968fefabd4c6b2650af21e110aa8466fe21432cd7c43a84577a89438/tensorboard_data_server-0.7.2-py3-none-any.whl", hash = "sha256:7e0610d205889588983836ec05dc098e80f97b7e7bbff7e994ebb78f578d0ddb", size = 2356 }, - { url = "https://files.pythonhosted.org/packages/b7/85/dabeaf902892922777492e1d253bb7e1264cadce3cea932f7ff599e53fea/tensorboard_data_server-0.7.2-py3-none-macosx_10_9_x86_64.whl", hash = "sha256:9fe5d24221b29625dbc7328b0436ca7fc1c23de4acf4d272f1180856e32f9f60", size = 4823598 }, - { url = "https://files.pythonhosted.org/packages/73/c6/825dab04195756cf8ff2e12698f22513b3db2f64925bdd41671bfb33aaa5/tensorboard_data_server-0.7.2-py3-none-manylinux_2_31_x86_64.whl", hash = "sha256:ef687163c24185ae9754ed5650eb5bc4d84ff257aabdc33f0cc6f74d8ba54530", size = 6590363 }, -] - -[[package]] -name = "tensorboardx" -version = "2.6.4" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "numpy" }, - { name = "packaging" }, - { name = "protobuf" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/2b/c5/d4cc6e293fb837aaf9f76dd7745476aeba8ef7ef5146c3b3f9ee375fe7a5/tensorboardx-2.6.4.tar.gz", hash = "sha256:b163ccb7798b31100b9f5fa4d6bc22dad362d7065c2f24b51e50731adde86828", size = 4769801 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/e0/1d/b5d63f1a6b824282b57f7b581810d20b7a28ca951f2d5b59f1eb0782c12b/tensorboardx-2.6.4-py3-none-any.whl", hash = "sha256:5970cf3a1f0a6a6e8b180ccf46f3fe832b8a25a70b86e5a237048a7c0beb18e2", size = 87201 }, -] - -[[package]] -name = "termcolor" -version = "3.1.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/ca/6c/3d75c196ac07ac8749600b60b03f4f6094d54e132c4d94ebac6ee0e0add0/termcolor-3.1.0.tar.gz", hash = "sha256:6a6dd7fbee581909eeec6a756cff1d7f7c376063b14e4a298dc4980309e55970", size = 14324 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/4f/bd/de8d508070629b6d84a30d01d57e4a65c69aa7f5abe7560b8fad3b50ea59/termcolor-3.1.0-py3-none-any.whl", hash = "sha256:591dd26b5c2ce03b9e43f391264626557873ce1d379019786f99b0c2bee140aa", size = 7684 }, -] - -[[package]] -name = "terminado" -version = "0.18.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "ptyprocess", marker = "os_name != 'nt'" }, - { name = "pywinpty", marker = "os_name == 'nt'" }, - { name = "tornado" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/8a/11/965c6fd8e5cc254f1fe142d547387da17a8ebfd75a3455f637c663fb38a0/terminado-0.18.1.tar.gz", hash = "sha256:de09f2c4b85de4765f7714688fff57d3e75bad1f909b589fde880460c753fd2e", size = 32701 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/6a/9e/2064975477fdc887e47ad42157e214526dcad8f317a948dee17e1659a62f/terminado-0.18.1-py3-none-any.whl", hash = "sha256:a4468e1b37bb318f8a86514f65814e1afc977cf29b3992a4500d9dd305dcceb0", size = 14154 }, -] - -[[package]] -name = "text-unidecode" -version = "1.3" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/ab/e2/e9a00f0ccb71718418230718b3d900e71a5d16e701a3dae079a21e9cd8f8/text-unidecode-1.3.tar.gz", hash = "sha256:bad6603bb14d279193107714b288be206cac565dfa49aa5b105294dd5c4aab93", size = 76885 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/a6/a5/c0b6468d3824fe3fde30dbb5e1f687b291608f9473681bbf7dabbf5a87d7/text_unidecode-1.3-py2.py3-none-any.whl", hash = "sha256:1311f10e8b895935241623731c2ba64f4c455287888b18189350b67134a822e8", size = 78154 }, -] - -[[package]] -name = "texterrors" -version = "0.5.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "levenshtein" }, - { name = "loguru" }, - { name = "numpy" }, - { name = "plac" }, - { name = "pybind11" }, - { name = "regex" }, - { name = "termcolor" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/9b/47/9a391643961698df3c804172f005e8b56c9693c14c4170abd9d3c961e971/texterrors-0.5.1.tar.gz", hash = "sha256:7fa24b2ca6ed5e05681b5cfdbb6c1fd0e4ae6518f8939e9782294f620d4eb3b1", size = 23813 } - -[[package]] -name = "threadpoolctl" -version = "3.6.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/b7/4d/08c89e34946fce2aec4fbb45c9016efd5f4d7f24af8e5d93296e935631d8/threadpoolctl-3.6.0.tar.gz", hash = "sha256:8ab8b4aa3491d812b623328249fab5302a68d2d71745c8a4c719a2fcaba9f44e", size = 21274 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/32/d5/f9a850d79b0851d1d4ef6456097579a9005b31fea68726a4ae5f2d82ddd9/threadpoolctl-3.6.0-py3-none-any.whl", hash = "sha256:43a0b8fd5a2928500110039e43a5eed8480b918967083ea48dc3ab9f13c4a7fb", size = 18638 }, -] - -[[package]] -name = "tinycss2" -version = "1.4.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "webencodings" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/7a/fd/7a5ee21fd08ff70d3d33a5781c255cbe779659bd03278feb98b19ee550f4/tinycss2-1.4.0.tar.gz", hash = "sha256:10c0972f6fc0fbee87c3edb76549357415e94548c1ae10ebccdea16fb404a9b7", size = 87085 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/e6/34/ebdc18bae6aa14fbee1a08b63c015c72b64868ff7dae68808ab500c492e2/tinycss2-1.4.0-py3-none-any.whl", hash = "sha256:3a49cf47b7675da0b15d0c6e1df8df4ebd96e9394bb905a5775adb0d884c5289", size = 26610 }, -] - -[[package]] -name = "tokenizers" -version = "0.21.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "huggingface-hub" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/ab/2d/b0fce2b8201635f60e8c95990080f58461cc9ca3d5026de2e900f38a7f21/tokenizers-0.21.2.tar.gz", hash = "sha256:fdc7cffde3e2113ba0e6cc7318c40e3438a4d74bbc62bf04bcc63bdfb082ac77", size = 351545 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/1d/cc/2936e2d45ceb130a21d929743f1e9897514691bec123203e10837972296f/tokenizers-0.21.2-cp39-abi3-macosx_10_12_x86_64.whl", hash = "sha256:342b5dfb75009f2255ab8dec0041287260fed5ce00c323eb6bab639066fef8ec", size = 2875206 }, - { url = "https://files.pythonhosted.org/packages/6c/e6/33f41f2cc7861faeba8988e7a77601407bf1d9d28fc79c5903f8f77df587/tokenizers-0.21.2-cp39-abi3-macosx_11_0_arm64.whl", hash = "sha256:126df3205d6f3a93fea80c7a8a266a78c1bd8dd2fe043386bafdd7736a23e45f", size = 2732655 }, - { url = "https://files.pythonhosted.org/packages/33/2b/1791eb329c07122a75b01035b1a3aa22ad139f3ce0ece1b059b506d9d9de/tokenizers-0.21.2-cp39-abi3-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:4a32cd81be21168bd0d6a0f0962d60177c447a1aa1b1e48fa6ec9fc728ee0b12", size = 3019202 }, - { url = "https://files.pythonhosted.org/packages/05/15/fd2d8104faa9f86ac68748e6f7ece0b5eb7983c7efc3a2c197cb98c99030/tokenizers-0.21.2-cp39-abi3-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:8bd8999538c405133c2ab999b83b17c08b7fc1b48c1ada2469964605a709ef91", size = 2934539 }, - { url = "https://files.pythonhosted.org/packages/a5/2e/53e8fd053e1f3ffbe579ca5f9546f35ac67cf0039ed357ad7ec57f5f5af0/tokenizers-0.21.2-cp39-abi3-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:5e9944e61239b083a41cf8fc42802f855e1dca0f499196df37a8ce219abac6eb", size = 3248665 }, - { url = "https://files.pythonhosted.org/packages/00/15/79713359f4037aa8f4d1f06ffca35312ac83629da062670e8830917e2153/tokenizers-0.21.2-cp39-abi3-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:514cd43045c5d546f01142ff9c79a96ea69e4b5cda09e3027708cb2e6d5762ab", size = 3451305 }, - { url = "https://files.pythonhosted.org/packages/38/5f/959f3a8756fc9396aeb704292777b84f02a5c6f25c3fc3ba7530db5feb2c/tokenizers-0.21.2-cp39-abi3-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:b1b9405822527ec1e0f7d8d2fdb287a5730c3a6518189c968254a8441b21faae", size = 3214757 }, - { url = "https://files.pythonhosted.org/packages/c5/74/f41a432a0733f61f3d21b288de6dfa78f7acff309c6f0f323b2833e9189f/tokenizers-0.21.2-cp39-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:fed9a4d51c395103ad24f8e7eb976811c57fbec2af9f133df471afcd922e5020", size = 3121887 }, - { url = "https://files.pythonhosted.org/packages/3c/6a/bc220a11a17e5d07b0dfb3b5c628621d4dcc084bccd27cfaead659963016/tokenizers-0.21.2-cp39-abi3-musllinux_1_2_aarch64.whl", hash = "sha256:2c41862df3d873665ec78b6be36fcc30a26e3d4902e9dd8608ed61d49a48bc19", size = 9091965 }, - { url = "https://files.pythonhosted.org/packages/6c/bd/ac386d79c4ef20dc6f39c4706640c24823dca7ebb6f703bfe6b5f0292d88/tokenizers-0.21.2-cp39-abi3-musllinux_1_2_armv7l.whl", hash = "sha256:ed21dc7e624e4220e21758b2e62893be7101453525e3d23264081c9ef9a6d00d", size = 9053372 }, - { url = "https://files.pythonhosted.org/packages/63/7b/5440bf203b2a5358f074408f7f9c42884849cd9972879e10ee6b7a8c3b3d/tokenizers-0.21.2-cp39-abi3-musllinux_1_2_i686.whl", hash = "sha256:0e73770507e65a0e0e2a1affd6b03c36e3bc4377bd10c9ccf51a82c77c0fe365", size = 9298632 }, - { url = "https://files.pythonhosted.org/packages/a4/d2/faa1acac3f96a7427866e94ed4289949b2524f0c1878512516567d80563c/tokenizers-0.21.2-cp39-abi3-musllinux_1_2_x86_64.whl", hash = "sha256:106746e8aa9014a12109e58d540ad5465b4c183768ea96c03cbc24c44d329958", size = 9470074 }, - { url = "https://files.pythonhosted.org/packages/d8/a5/896e1ef0707212745ae9f37e84c7d50269411aef2e9ccd0de63623feecdf/tokenizers-0.21.2-cp39-abi3-win32.whl", hash = "sha256:cabda5a6d15d620b6dfe711e1af52205266d05b379ea85a8a301b3593c60e962", size = 2330115 }, - { url = "https://files.pythonhosted.org/packages/13/c3/cc2755ee10be859c4338c962a35b9a663788c0c0b50c0bdd8078fb6870cf/tokenizers-0.21.2-cp39-abi3-win_amd64.whl", hash = "sha256:58747bb898acdb1007f37a7bbe614346e98dc28708ffb66a3fd50ce169ac6c98", size = 2509918 }, -] - -[[package]] -name = "tomli" -version = "2.2.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/18/87/302344fed471e44a87289cf4967697d07e532f2421fdaf868a303cbae4ff/tomli-2.2.1.tar.gz", hash = "sha256:cd45e1dc79c835ce60f7404ec8119f2eb06d38b1deba146f07ced3bbc44505ff", size = 17175 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/6e/c2/61d3e0f47e2b74ef40a68b9e6ad5984f6241a942f7cd3bbfbdbd03861ea9/tomli-2.2.1-py3-none-any.whl", hash = "sha256:cb55c73c5f4408779d0cf3eef9f762b9c9f147a77de7b258bef0a5628adc85cc", size = 14257 }, -] - -[[package]] -name = "toolz" -version = "1.0.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/8a/0b/d80dfa675bf592f636d1ea0b835eab4ec8df6e9415d8cfd766df54456123/toolz-1.0.0.tar.gz", hash = "sha256:2c86e3d9a04798ac556793bced838816296a2f085017664e4995cb40a1047a02", size = 66790 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/03/98/eb27cc78ad3af8e302c9d8ff4977f5026676e130d28dd7578132a457170c/toolz-1.0.0-py3-none-any.whl", hash = "sha256:292c8f1c4e7516bf9086f8850935c799a874039c8bcf959d47b600e4c44a6236", size = 56383 }, -] - -[[package]] -name = "torch" -version = "2.7.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "filelock" }, - { name = "fsspec" }, - { name = "jinja2" }, - { name = "networkx" }, - { name = "nvidia-cublas-cu12", marker = "platform_machine == 'x86_64' and platform_system == 'Linux'" }, - { name = "nvidia-cuda-cupti-cu12", marker = "platform_machine == 'x86_64' and platform_system == 'Linux'" }, - { name = "nvidia-cuda-nvrtc-cu12", marker = "platform_machine == 'x86_64' and platform_system == 'Linux'" }, - { name = "nvidia-cuda-runtime-cu12", marker = "platform_machine == 'x86_64' and platform_system == 'Linux'" }, - { name = "nvidia-cudnn-cu12", marker = "platform_machine == 'x86_64' and platform_system == 'Linux'" }, - { name = "nvidia-cufft-cu12", marker = "platform_machine == 'x86_64' and platform_system == 'Linux'" }, - { name = "nvidia-cufile-cu12", marker = "platform_machine == 'x86_64' and platform_system == 'Linux'" }, - { name = "nvidia-curand-cu12", marker = "platform_machine == 'x86_64' and platform_system == 'Linux'" }, - { name = "nvidia-cusolver-cu12", marker = "platform_machine == 'x86_64' and platform_system == 'Linux'" }, - { name = "nvidia-cusparse-cu12", marker = "platform_machine == 'x86_64' and platform_system == 'Linux'" }, - { name = "nvidia-cusparselt-cu12", marker = "platform_machine == 'x86_64' and platform_system == 'Linux'" }, - { name = "nvidia-nccl-cu12", marker = "platform_machine == 'x86_64' and platform_system == 'Linux'" }, - { name = "nvidia-nvjitlink-cu12", marker = "platform_machine == 'x86_64' and platform_system == 'Linux'" }, - { name = "nvidia-nvtx-cu12", marker = "platform_machine == 'x86_64' and platform_system == 'Linux'" }, - { name = "sympy" }, - { name = "triton", marker = "platform_machine == 'x86_64' and platform_system == 'Linux'" }, - { name = "typing-extensions" }, -] -wheels = [ - { url = "https://files.pythonhosted.org/packages/46/c2/3fb87940fa160d956ee94d644d37b99a24b9c05a4222bf34f94c71880e28/torch-2.7.0-cp310-cp310-manylinux_2_28_aarch64.whl", hash = "sha256:c9afea41b11e1a1ab1b258a5c31afbd646d6319042bfe4f231b408034b51128b", size = 99158447 }, - { url = "https://files.pythonhosted.org/packages/cc/2c/91d1de65573fce563f5284e69d9c56b57289625cffbbb6d533d5d56c36a5/torch-2.7.0-cp310-cp310-manylinux_2_28_x86_64.whl", hash = "sha256:0b9960183b6e5b71239a3e6c883d8852c304e691c0b2955f7045e8a6d05b9183", size = 865164221 }, - { url = "https://files.pythonhosted.org/packages/7f/7e/1b1cc4e0e7cc2666cceb3d250eef47a205f0821c330392cf45eb08156ce5/torch-2.7.0-cp310-cp310-win_amd64.whl", hash = "sha256:2ad79d0d8c2a20a37c5df6052ec67c2078a2c4e9a96dd3a8b55daaff6d28ea29", size = 212521189 }, - { url = "https://files.pythonhosted.org/packages/dc/0b/b2b83f30b8e84a51bf4f96aa3f5f65fdf7c31c591cc519310942339977e2/torch-2.7.0-cp310-none-macosx_11_0_arm64.whl", hash = "sha256:34e0168ed6de99121612d72224e59b2a58a83dae64999990eada7260c5dd582d", size = 68559462 }, -] - -[[package]] -name = "torch-audiomentations" -version = "0.12.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "julius" }, - { name = "torch" }, - { name = "torch-pitch-shift" }, - { name = "torchaudio" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/31/8d/2f8fd7e34c75f5ee8de4310c3bd3f22270acd44d1f809e2fe7c12fbf35f8/torch_audiomentations-0.12.0.tar.gz", hash = "sha256:b02d4c5eb86376986a53eb405cca5e34f370ea9284411237508e720c529f7888", size = 52094 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/21/9d/1ee04f49c15d2d632f6f7102061d7c07652858e6d91b58a091531034e84f/torch_audiomentations-0.12.0-py3-none-any.whl", hash = "sha256:1b80b91d2016ccf83979622cac8f702072a79b7dcc4c2bee40f00b26433a786b", size = 48506 }, -] - -[[package]] -name = "torch-complex" -version = "0.4.4" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "numpy" }, - { name = "packaging" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/bf/2b/17cb15a383cf2135330371e034d13b9043dc6d8bd07c871b5aa3064fbed1/torch_complex-0.4.4.tar.gz", hash = "sha256:4153fd6b24a0bad689e6f193bfbd00f38283b1890d808bef684ddc6d1f63fd3f", size = 10025 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/f4/c5/9b4d756a7ada951e9b17dcc636f98ed1073c737ae809b150ef408afb6298/torch_complex-0.4.4-py3-none-any.whl", hash = "sha256:6ab4ecd4f3a16e3adb70a7f7cd2e769a9dfd07d7a8e27d04ff9c621ebbe34b13", size = 9125 }, -] - -[[package]] -name = "torch-pitch-shift" -version = "1.2.5" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "packaging" }, - { name = "primepy" }, - { name = "torch" }, - { name = "torchaudio" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/79/a6/722a832bca75d5079f6731e005b3d0c2eec7c6c6863d030620952d143d57/torch_pitch_shift-1.2.5.tar.gz", hash = "sha256:6e1c7531f08d0f407a4c55e5ff8385a41355c5c5d27ab7fa08632e51defbd0ed", size = 4725 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/27/4c/96ac2a09efb56cc3c41fb3ce9b6f4d8c0604499f7481d4a13a7b03e21382/torch_pitch_shift-1.2.5-py3-none-any.whl", hash = "sha256:6f8500cbc13f1c98b11cde1805ce5084f82cdd195c285f34287541f168a7c6a7", size = 5005 }, -] - -[[package]] -name = "torchaudio" -version = "2.7.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "torch" }, -] -wheels = [ - { url = "https://files.pythonhosted.org/packages/34/26/abc66c79092ad2eaaade546dc93e23d99ddf2513988261b943d274f5c01a/torchaudio-2.7.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:1c4a646c9e9347836c09e965eebc58dd028ec6ef34c46d3e7891bffd8dc645ea", size = 1842304 }, - { url = "https://files.pythonhosted.org/packages/ee/f7/17b8fbce19280424e612f254e1b89faf3c7640c022667a480307f2f3ca76/torchaudio-2.7.0-cp310-cp310-manylinux_2_28_aarch64.whl", hash = "sha256:9e4073992f4f8e7113e4b505d95095361ceb2f21dd7b9310776160a24266f8f6", size = 1680682 }, - { url = "https://files.pythonhosted.org/packages/f2/df/ee0097fc41f718152026541c4c6cdeea830bc09903cc36a53037942a6d3d/torchaudio-2.7.0-cp310-cp310-manylinux_2_28_x86_64.whl", hash = "sha256:f7c99f7c062d6a56a3e281e3c2b779099e64cad1ce78891df61c4d19ce40742e", size = 3444849 }, - { url = "https://files.pythonhosted.org/packages/65/a6/e1903c1b3787f0408d30624536d2ae30da9f749720f3cf272a4fb7abc490/torchaudio-2.7.0-cp310-cp310-win_amd64.whl", hash = "sha256:a5443422640cbe532aaacd83ad2ee6911b0451f7f50e6b3755015e92df579d37", size = 2492239 }, -] - -[[package]] -name = "torchmetrics" -version = "1.7.3" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "lightning-utilities" }, - { name = "numpy" }, - { name = "packaging" }, - { name = "torch" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/48/22/8b16c4ec34d93ee15024924cbbe84fbd235bb3e1df2cc8f48c865c1528e7/torchmetrics-1.7.3.tar.gz", hash = "sha256:08450a19cdb67ba1608aac0b213e5dc73033e11b60ad4719696ebcede591621e", size = 566545 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/6f/f2/bed7da46003c26ed44fc7fa3ecc98a84216f0d4758e5e6a3693754d490d9/torchmetrics-1.7.3-py3-none-any.whl", hash = "sha256:7b6fd43e92f0a1071c8bcb029637f252b0630699140a93ed8817ce7afe9db34e", size = 962639 }, -] - -[[package]] -name = "tornado" -version = "6.5.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/51/89/c72771c81d25d53fe33e3dca61c233b665b2780f21820ba6fd2c6793c12b/tornado-6.5.1.tar.gz", hash = "sha256:84ceece391e8eb9b2b95578db65e920d2a61070260594819589609ba9bc6308c", size = 509934 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/77/89/f4532dee6843c9e0ebc4e28d4be04c67f54f60813e4bf73d595fe7567452/tornado-6.5.1-cp39-abi3-macosx_10_9_universal2.whl", hash = "sha256:d50065ba7fd11d3bd41bcad0825227cc9a95154bad83239357094c36708001f7", size = 441948 }, - { url = "https://files.pythonhosted.org/packages/15/9a/557406b62cffa395d18772e0cdcf03bed2fff03b374677348eef9f6a3792/tornado-6.5.1-cp39-abi3-macosx_10_9_x86_64.whl", hash = "sha256:9e9ca370f717997cb85606d074b0e5b247282cf5e2e1611568b8821afe0342d6", size = 440112 }, - { url = "https://files.pythonhosted.org/packages/55/82/7721b7319013a3cf881f4dffa4f60ceff07b31b394e459984e7a36dc99ec/tornado-6.5.1-cp39-abi3-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:b77e9dfa7ed69754a54c89d82ef746398be82f749df69c4d3abe75c4d1ff4888", size = 443672 }, - { url = "https://files.pythonhosted.org/packages/7d/42/d11c4376e7d101171b94e03cef0cbce43e823ed6567ceda571f54cf6e3ce/tornado-6.5.1-cp39-abi3-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:253b76040ee3bab8bcf7ba9feb136436a3787208717a1fb9f2c16b744fba7331", size = 443019 }, - { url = "https://files.pythonhosted.org/packages/7d/f7/0c48ba992d875521ac761e6e04b0a1750f8150ae42ea26df1852d6a98942/tornado-6.5.1-cp39-abi3-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:308473f4cc5a76227157cdf904de33ac268af770b2c5f05ca6c1161d82fdd95e", size = 443252 }, - { url = "https://files.pythonhosted.org/packages/89/46/d8d7413d11987e316df4ad42e16023cd62666a3c0dfa1518ffa30b8df06c/tornado-6.5.1-cp39-abi3-musllinux_1_2_aarch64.whl", hash = "sha256:caec6314ce8a81cf69bd89909f4b633b9f523834dc1a352021775d45e51d9401", size = 443930 }, - { url = "https://files.pythonhosted.org/packages/78/b2/f8049221c96a06df89bed68260e8ca94beca5ea532ffc63b1175ad31f9cc/tornado-6.5.1-cp39-abi3-musllinux_1_2_i686.whl", hash = "sha256:13ce6e3396c24e2808774741331638ee6c2f50b114b97a55c5b442df65fd9692", size = 443351 }, - { url = "https://files.pythonhosted.org/packages/76/ff/6a0079e65b326cc222a54720a748e04a4db246870c4da54ece4577bfa702/tornado-6.5.1-cp39-abi3-musllinux_1_2_x86_64.whl", hash = "sha256:5cae6145f4cdf5ab24744526cc0f55a17d76f02c98f4cff9daa08ae9a217448a", size = 443328 }, - { url = "https://files.pythonhosted.org/packages/49/18/e3f902a1d21f14035b5bc6246a8c0f51e0eef562ace3a2cea403c1fb7021/tornado-6.5.1-cp39-abi3-win32.whl", hash = "sha256:e0a36e1bc684dca10b1aa75a31df8bdfed656831489bc1e6a6ebed05dc1ec365", size = 444396 }, - { url = "https://files.pythonhosted.org/packages/7b/09/6526e32bf1049ee7de3bebba81572673b19a2a8541f795d887e92af1a8bc/tornado-6.5.1-cp39-abi3-win_amd64.whl", hash = "sha256:908e7d64567cecd4c2b458075589a775063453aeb1d2a1853eedb806922f568b", size = 444840 }, - { url = "https://files.pythonhosted.org/packages/55/a7/535c44c7bea4578e48281d83c615219f3ab19e6abc67625ef637c73987be/tornado-6.5.1-cp39-abi3-win_arm64.whl", hash = "sha256:02420a0eb7bf617257b9935e2b754d1b63897525d8a289c9d65690d580b4dcf7", size = 443596 }, -] - -[[package]] -name = "tqdm" -version = "4.67.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "colorama", marker = "platform_system == 'Windows'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/a8/4b/29b4ef32e036bb34e4ab51796dd745cdba7ed47ad142a9f4a1eb8e0c744d/tqdm-4.67.1.tar.gz", hash = "sha256:f8aef9c52c08c13a65f30ea34f4e5aac3fd1a34959879d7e59e63027286627f2", size = 169737 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/d0/30/dc54f88dd4a2b5dc8a0279bdd7270e735851848b762aeb1c1184ed1f6b14/tqdm-4.67.1-py3-none-any.whl", hash = "sha256:26445eca388f82e72884e0d580d5464cd801a3ea01e63e5601bdff9ba6a48de2", size = 78540 }, -] - -[[package]] -name = "traitlets" -version = "5.14.3" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/eb/79/72064e6a701c2183016abbbfedaba506d81e30e232a68c9f0d6f6fcd1574/traitlets-5.14.3.tar.gz", hash = "sha256:9ed0579d3502c94b4b3732ac120375cda96f923114522847de4b3bb98b96b6b7", size = 161621 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/00/c0/8f5d070730d7836adc9c9b6408dec68c6ced86b304a9b26a14df072a6e8c/traitlets-5.14.3-py3-none-any.whl", hash = "sha256:b74e89e397b1ed28cc831db7aea759ba6640cb3de13090ca145426688ff1ac4f", size = 85359 }, -] - -[[package]] -name = "transformers" -version = "4.51.3" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "filelock" }, - { name = "huggingface-hub" }, - { name = "numpy" }, - { name = "packaging" }, - { name = "pyyaml" }, - { name = "regex" }, - { name = "requests" }, - { name = "safetensors" }, - { name = "tokenizers" }, - { name = "tqdm" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/f1/11/7414d5bc07690002ce4d7553602107bf969af85144bbd02830f9fb471236/transformers-4.51.3.tar.gz", hash = "sha256:e292fcab3990c6defe6328f0f7d2004283ca81a7a07b2de9a46d67fd81ea1409", size = 8941266 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/a9/b6/5257d04ae327b44db31f15cce39e6020cc986333c715660b1315a9724d82/transformers-4.51.3-py3-none-any.whl", hash = "sha256:fd3279633ceb2b777013234bbf0b4f5c2d23c4626b05497691f00cfda55e8a83", size = 10383940 }, -] - -[[package]] -name = "triton" -version = "3.3.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "setuptools" }, -] -wheels = [ - { url = "https://files.pythonhosted.org/packages/76/04/d54d3a6d077c646624dc9461b0059e23fd5d30e0dbe67471e3654aec81f9/triton-3.3.0-cp310-cp310-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:fad99beafc860501d7fcc1fb7045d9496cbe2c882b1674640304949165a916e7", size = 156441993 }, -] - -[[package]] -name = "typeguard" -version = "4.4.4" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/c7/68/71c1a15b5f65f40e91b65da23b8224dad41349894535a97f63a52e462196/typeguard-4.4.4.tar.gz", hash = "sha256:3a7fd2dffb705d4d0efaed4306a704c89b9dee850b688f060a8b1615a79e5f74", size = 75203 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/1b/a9/e3aee762739c1d7528da1c3e06d518503f8b6c439c35549b53735ba52ead/typeguard-4.4.4-py3-none-any.whl", hash = "sha256:b5f562281b6bfa1f5492470464730ef001646128b180769880468bd84b68b09e", size = 34874 }, -] - -[[package]] -name = "typer" -version = "0.16.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "click" }, - { name = "rich" }, - { name = "shellingham" }, - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/c5/8c/7d682431efca5fd290017663ea4588bf6f2c6aad085c7f108c5dbc316e70/typer-0.16.0.tar.gz", hash = "sha256:af377ffaee1dbe37ae9440cb4e8f11686ea5ce4e9bae01b84ae7c63b87f1dd3b", size = 102625 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/76/42/3efaf858001d2c2913de7f354563e3a3a2f0decae3efe98427125a8f441e/typer-0.16.0-py3-none-any.whl", hash = "sha256:1f79bed11d4d02d4310e3c1b7ba594183bcedb0ac73b27a9e5f28f6fb5b98855", size = 46317 }, -] - -[[package]] -name = "types-python-dateutil" -version = "2.9.0.20250516" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/ef/88/d65ed807393285204ab6e2801e5d11fbbea811adcaa979a2ed3b67a5ef41/types_python_dateutil-2.9.0.20250516.tar.gz", hash = "sha256:13e80d6c9c47df23ad773d54b2826bd52dbbb41be87c3f339381c1700ad21ee5", size = 13943 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/c5/3f/b0e8db149896005adc938a1e7f371d6d7e9eca4053a29b108978ed15e0c2/types_python_dateutil-2.9.0.20250516-py3-none-any.whl", hash = "sha256:2b2b3f57f9c6a61fba26a9c0ffb9ea5681c9b83e69cd897c6b5f668d9c0cab93", size = 14356 }, -] - -[[package]] -name = "typing-extensions" -version = "4.14.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/d1/bc/51647cd02527e87d05cb083ccc402f93e441606ff1f01739a62c8ad09ba5/typing_extensions-4.14.0.tar.gz", hash = "sha256:8676b788e32f02ab42d9e7c61324048ae4c6d844a399eebace3d4979d75ceef4", size = 107423 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/69/e0/552843e0d356fbb5256d21449fa957fa4eff3bbc135a74a691ee70c7c5da/typing_extensions-4.14.0-py3-none-any.whl", hash = "sha256:a1514509136dd0b477638fc68d6a91497af5076466ad0fa6c338e44e359944af", size = 43839 }, -] - -[[package]] -name = "typing-inspection" -version = "0.4.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/f8/b1/0c11f5058406b3af7609f121aaa6b609744687f1d158b3c3a5bf4cc94238/typing_inspection-0.4.1.tar.gz", hash = "sha256:6ae134cc0203c33377d43188d4064e9b357dba58cff3185f22924610e70a9d28", size = 75726 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/17/69/cd203477f944c353c31bade965f880aa1061fd6bf05ded0726ca845b6ff7/typing_inspection-0.4.1-py3-none-any.whl", hash = "sha256:389055682238f53b04f7badcb49b989835495a96700ced5dab2d8feae4b26f51", size = 14552 }, -] - -[[package]] -name = "tzdata" -version = "2025.2" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/95/32/1a225d6164441be760d75c2c42e2780dc0873fe382da3e98a2e1e48361e5/tzdata-2025.2.tar.gz", hash = "sha256:b60a638fcc0daffadf82fe0f57e53d06bdec2f36c4df66280ae79bce6bd6f2b9", size = 196380 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/5c/23/c7abc0ca0a1526a0774eca151daeb8de62ec457e77262b66b359c3c7679e/tzdata-2025.2-py2.py3-none-any.whl", hash = "sha256:1a403fada01ff9221ca8044d701868fa132215d84beb92242d9acd2147f667a8", size = 347839 }, -] - -[[package]] -name = "umap-learn" -version = "0.5.7" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "numba" }, - { name = "numpy" }, - { name = "pynndescent" }, - { name = "scikit-learn" }, - { name = "scipy" }, - { name = "tqdm" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/6f/d4/9ed627905f7993349671283b3c5bf2d9f543ef79229fa1c7e01324eb900c/umap-learn-0.5.7.tar.gz", hash = "sha256:b2a97973e4c6ffcebf241100a8de589a4c84126a832ab40f296c6d9fcc5eb19e", size = 92680 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/3c/8f/671c0e1f2572ba625cbcc1faeba9435e00330c3d6962858711445cf1e817/umap_learn-0.5.7-py3-none-any.whl", hash = "sha256:6a7e0be2facfa365a5ed6588447102bdbef32a0ef449535c25c97ea7e680073c", size = 88815 }, -] - -[[package]] -name = "uri-template" -version = "1.3.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/31/c7/0336f2bd0bcbada6ccef7aaa25e443c118a704f828a0620c6fa0207c1b64/uri-template-1.3.0.tar.gz", hash = "sha256:0e00f8eb65e18c7de20d595a14336e9f337ead580c70934141624b6d1ffdacc7", size = 21678 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/e7/00/3fca040d7cf8a32776d3d81a00c8ee7457e00f80c649f1e4a863c8321ae9/uri_template-1.3.0-py3-none-any.whl", hash = "sha256:a44a133ea12d44a0c0f06d7d42a52d71282e77e2f937d8abd5655b8d56fc1363", size = 11140 }, -] - -[[package]] -name = "urllib3" -version = "2.5.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/15/22/9ee70a2574a4f4599c47dd506532914ce044817c7752a79b6a51286319bc/urllib3-2.5.0.tar.gz", hash = "sha256:3fc47733c7e419d4bc3f6b3dc2b4f890bb743906a30d56ba4a5bfa4bbff92760", size = 393185 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/a7/c2/fe1e52489ae3122415c51f387e221dd0773709bad6c6cdaa599e8a2c5185/urllib3-2.5.0-py3-none-any.whl", hash = "sha256:e6b01673c0fa6a13e374b50871808eb3bf7046c4b125b216f6bf1cc604cff0dc", size = 129795 }, -] - -[[package]] -name = "wandb" -version = "0.20.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "click" }, - { name = "gitpython" }, - { name = "packaging" }, - { name = "platformdirs" }, - { name = "protobuf", marker = "python_full_version == '3.10.12'" }, - { name = "psutil" }, - { name = "pydantic" }, - { name = "pyyaml" }, - { name = "requests" }, - { name = "sentry-sdk" }, - { name = "setproctitle" }, - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/62/1f/92be0ca87fb49eb48c16dcf0845a3579a57c4734fec2b95862cf5a0494a0/wandb-0.20.1.tar.gz", hash = "sha256:dbd3fc60dfe7bf83c4de24b206b99b44949fef323f817a783883db72fc5f3bfe", size = 40320062 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/c9/18/afcc37d0b93dd6f6d0f0c5683b9cfff9416ae1539931f58932a2938c0070/wandb-0.20.1-py3-none-any.whl", hash = "sha256:e6395cabf074247042be1cf0dc6ab0b06aa4c9538c2e1fdc5b507a690ce0cf17", size = 6458872 }, - { url = "https://files.pythonhosted.org/packages/e6/b5/70f9e2a3d1380b729ae5853763d938edc50072df357f79bbd19b9aae8e3f/wandb-0.20.1-py3-none-macosx_10_14_x86_64.whl", hash = "sha256:2475a48c693adf677d40da9e1c8ceeaf86d745ffc3b7e3535731279d02f9e845", size = 22517483 }, - { url = "https://files.pythonhosted.org/packages/cc/7e/4eb9aeb2fd974d410a8f2eb11b0219536503913a050d46a03206151705c8/wandb-0.20.1-py3-none-macosx_11_0_arm64.whl", hash = "sha256:99cce804c31ec1e0d1e691650a7d51773ed7329c41745d56384fa3655a0e9b2c", size = 22034511 }, - { url = "https://files.pythonhosted.org/packages/34/38/1df22c2273e6f7ab0aae4fd032085d6d92ab112f5b261646e7dc5e675cfe/wandb-0.20.1-py3-none-macosx_11_0_x86_64.whl", hash = "sha256:ce3ee412677a1679e04b21e03a91e1e02eb90faf658d682bee86c33cf5f32e09", size = 22720771 }, - { url = "https://files.pythonhosted.org/packages/38/96/78fc7a7ea7158d136c84f481423f8736c9346a2387287ec8a6d92019975c/wandb-0.20.1-py3-none-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5e58ca32c7147161158f09b0fb5f5896876f8569d0d10ae7b64d0510c868ce33", size = 21537453 }, - { url = "https://files.pythonhosted.org/packages/88/c9/41b8bdb493e5eda32b502bc1cc49d539335a92cacaf0ef304d7dae0240aa/wandb-0.20.1-py3-none-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:591506ecbdd396648cc323ba270f3ab4aed3158e1dbfa7636c09f9f7f0253e1c", size = 23161349 }, - { url = "https://files.pythonhosted.org/packages/7d/f2/79e783cc50a47d373dfbda862eb5396de8139167e8c6443a16ef0166106f/wandb-0.20.1-py3-none-musllinux_1_2_aarch64.whl", hash = "sha256:382508532db09893f81cc926b1d333caa4c8a7db057878899fadf929bbdb3b56", size = 21550624 }, - { url = "https://files.pythonhosted.org/packages/26/32/23890a726302e7be28bda9fff47ce9b491af64e339aba4d32b3b8d1a7aaf/wandb-0.20.1-py3-none-musllinux_1_2_x86_64.whl", hash = "sha256:29ea495e49393db860f17437fe37e48018da90436ce10949b471780f09293bd7", size = 23237996 }, - { url = "https://files.pythonhosted.org/packages/af/94/296e520b086b2a4f10e99bcea3cd5856421b9c004824663501e3789a713b/wandb-0.20.1-py3-none-win32.whl", hash = "sha256:455ee0a652e59ab1e4b546fa1dc833dd3063aa7e64eb8abf95d22f0e9f08c574", size = 22518456 }, - { url = "https://files.pythonhosted.org/packages/52/5f/c44ad7b2a062ca5f4da99ae475cea274c38f6ec37bdaca1b1c653ee87274/wandb-0.20.1-py3-none-win_amd64.whl", hash = "sha256:6d2431652f096b7e394c29a99135a6441c02ed3198b963f0b351a5b5e56aeca0", size = 22518459 }, -] - -[[package]] -name = "wcwidth" -version = "0.2.13" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/6c/63/53559446a878410fc5a5974feb13d31d78d752eb18aeba59c7fef1af7598/wcwidth-0.2.13.tar.gz", hash = "sha256:72ea0c06399eb286d978fdedb6923a9eb47e1c486ce63e9b4e64fc18303972b5", size = 101301 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/fd/84/fd2ba7aafacbad3c4201d395674fc6348826569da3c0937e75505ead3528/wcwidth-0.2.13-py2.py3-none-any.whl", hash = "sha256:3da69048e4540d84af32131829ff948f1e022c1c6bdb8d6102117aac784f6859", size = 34166 }, -] - -[[package]] -name = "webcolors" -version = "24.11.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/7b/29/061ec845fb58521848f3739e466efd8250b4b7b98c1b6c5bf4d40b419b7e/webcolors-24.11.1.tar.gz", hash = "sha256:ecb3d768f32202af770477b8b65f318fa4f566c22948673a977b00d589dd80f6", size = 45064 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/60/e8/c0e05e4684d13459f93d312077a9a2efbe04d59c393bc2b8802248c908d4/webcolors-24.11.1-py3-none-any.whl", hash = "sha256:515291393b4cdf0eb19c155749a096f779f7d909f7cceea072791cb9095b92e9", size = 14934 }, -] - -[[package]] -name = "webdataset" -version = "1.0.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "braceexpand" }, - { name = "numpy" }, - { name = "pyyaml" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/5a/3a/68800d92e065cf4750ebecf973b13979c0c929b439e1293012938862038d/webdataset-1.0.2.tar.gz", hash = "sha256:7f0498be827cfa46cc5430a58768a24e2c6a410676a61be1838f53d61afdaab4", size = 80090 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/d9/00/aca6beb3658dab4ed3dbb41a78e6e7f31342e0b41d28088f205525751601/webdataset-1.0.2-py3-none-any.whl", hash = "sha256:3dbfced32b25c0d199c6b9787937b6f85742bc3c84f652c846893075c1c082d9", size = 74956 }, -] - -[[package]] -name = "webencodings" -version = "0.5.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/0b/02/ae6ceac1baeda530866a85075641cec12989bd8d31af6d5ab4a3e8c92f47/webencodings-0.5.1.tar.gz", hash = "sha256:b36a1c245f2d304965eb4e0a82848379241dc04b865afcc4aab16748587e1923", size = 9721 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/f4/24/2a3e3df732393fed8b3ebf2ec078f05546de641fe1b667ee316ec1dcf3b7/webencodings-0.5.1-py2.py3-none-any.whl", hash = "sha256:a0af1213f3c2226497a97e2b3aa01a7e4bee4f403f95be16fc9acd2947514a78", size = 11774 }, -] - -[[package]] -name = "websocket-client" -version = "1.8.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/e6/30/fba0d96b4b5fbf5948ed3f4681f7da2f9f64512e1d303f94b4cc174c24a5/websocket_client-1.8.0.tar.gz", hash = "sha256:3239df9f44da632f96012472805d40a23281a991027ce11d2f45a6f24ac4c3da", size = 54648 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/5a/84/44687a29792a70e111c5c477230a72c4b957d88d16141199bf9acb7537a3/websocket_client-1.8.0-py3-none-any.whl", hash = "sha256:17b44cc997f5c498e809b22cdf2d9c7a9e71c02c8cc2b6c56e7c2d1239bfa526", size = 58826 }, -] - -[[package]] -name = "werkzeug" -version = "3.1.3" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "markupsafe" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/9f/69/83029f1f6300c5fb2471d621ab06f6ec6b3324685a2ce0f9777fd4a8b71e/werkzeug-3.1.3.tar.gz", hash = "sha256:60723ce945c19328679790e3282cc758aa4a6040e4bb330f53d30fa546d44746", size = 806925 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/52/24/ab44c871b0f07f491e5d2ad12c9bd7358e527510618cb1b803a88e986db1/werkzeug-3.1.3-py3-none-any.whl", hash = "sha256:54b78bf3716d19a65be4fceccc0d1d7b89e608834989dfae50ea87564639213e", size = 224498 }, -] - -[[package]] -name = "wget" -version = "3.2" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/47/6a/62e288da7bcda82b935ff0c6cfe542970f04e29c756b0e147251b2fb251f/wget-3.2.zip", hash = "sha256:35e630eca2aa50ce998b9b1a127bb26b30dfee573702782aa982f875e3f16061", size = 10857 } - -[[package]] -name = "widgetsnbextension" -version = "4.0.14" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/41/53/2e0253c5efd69c9656b1843892052a31c36d37ad42812b5da45c62191f7e/widgetsnbextension-4.0.14.tar.gz", hash = "sha256:a3629b04e3edb893212df862038c7232f62973373869db5084aed739b437b5af", size = 1097428 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/ca/51/5447876806d1088a0f8f71e16542bf350918128d0a69437df26047c8e46f/widgetsnbextension-4.0.14-py3-none-any.whl", hash = "sha256:4875a9eaf72fbf5079dc372a51a9f268fc38d46f767cbf85c43a36da5cb9b575", size = 2196503 }, -] - -[[package]] -name = "win32-setctime" -version = "1.2.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/b3/8f/705086c9d734d3b663af0e9bb3d4de6578d08f46b1b101c2442fd9aecaa2/win32_setctime-1.2.0.tar.gz", hash = "sha256:ae1fdf948f5640aae05c511ade119313fb6a30d7eabe25fef9764dca5873c4c0", size = 4867 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/e1/07/c6fe3ad3e685340704d314d765b7912993bcb8dc198f0e7a89382d37974b/win32_setctime-1.2.0-py3-none-any.whl", hash = "sha256:95d644c4e708aba81dc3704a116d8cbc974d70b3bdb8be1d150e36be6e9d1390", size = 4083 }, -] - -[[package]] -name = "wrapt" -version = "1.17.2" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/c3/fc/e91cc220803d7bc4db93fb02facd8461c37364151b8494762cc88b0fbcef/wrapt-1.17.2.tar.gz", hash = "sha256:41388e9d4d1522446fe79d3213196bd9e3b301a336965b9e27ca2788ebd122f3", size = 55531 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/5a/d1/1daec934997e8b160040c78d7b31789f19b122110a75eca3d4e8da0049e1/wrapt-1.17.2-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:3d57c572081fed831ad2d26fd430d565b76aa277ed1d30ff4d40670b1c0dd984", size = 53307 }, - { url = "https://files.pythonhosted.org/packages/1b/7b/13369d42651b809389c1a7153baa01d9700430576c81a2f5c5e460df0ed9/wrapt-1.17.2-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:b5e251054542ae57ac7f3fba5d10bfff615b6c2fb09abeb37d2f1463f841ae22", size = 38486 }, - { url = "https://files.pythonhosted.org/packages/62/bf/e0105016f907c30b4bd9e377867c48c34dc9c6c0c104556c9c9126bd89ed/wrapt-1.17.2-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:80dd7db6a7cb57ffbc279c4394246414ec99537ae81ffd702443335a61dbf3a7", size = 38777 }, - { url = "https://files.pythonhosted.org/packages/27/70/0f6e0679845cbf8b165e027d43402a55494779295c4b08414097b258ac87/wrapt-1.17.2-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0a6e821770cf99cc586d33833b2ff32faebdbe886bd6322395606cf55153246c", size = 83314 }, - { url = "https://files.pythonhosted.org/packages/0f/77/0576d841bf84af8579124a93d216f55d6f74374e4445264cb378a6ed33eb/wrapt-1.17.2-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:b60fb58b90c6d63779cb0c0c54eeb38941bae3ecf7a73c764c52c88c2dcb9d72", size = 74947 }, - { url = "https://files.pythonhosted.org/packages/90/ec/00759565518f268ed707dcc40f7eeec38637d46b098a1f5143bff488fe97/wrapt-1.17.2-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b870b5df5b71d8c3359d21be8f0d6c485fa0ebdb6477dda51a1ea54a9b558061", size = 82778 }, - { url = "https://files.pythonhosted.org/packages/f8/5a/7cffd26b1c607b0b0c8a9ca9d75757ad7620c9c0a9b4a25d3f8a1480fafc/wrapt-1.17.2-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:4011d137b9955791f9084749cba9a367c68d50ab8d11d64c50ba1688c9b457f2", size = 81716 }, - { url = "https://files.pythonhosted.org/packages/7e/09/dccf68fa98e862df7e6a60a61d43d644b7d095a5fc36dbb591bbd4a1c7b2/wrapt-1.17.2-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:1473400e5b2733e58b396a04eb7f35f541e1fb976d0c0724d0223dd607e0f74c", size = 74548 }, - { url = "https://files.pythonhosted.org/packages/b7/8e/067021fa3c8814952c5e228d916963c1115b983e21393289de15128e867e/wrapt-1.17.2-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:3cedbfa9c940fdad3e6e941db7138e26ce8aad38ab5fe9dcfadfed9db7a54e62", size = 81334 }, - { url = "https://files.pythonhosted.org/packages/4b/0d/9d4b5219ae4393f718699ca1c05f5ebc0c40d076f7e65fd48f5f693294fb/wrapt-1.17.2-cp310-cp310-win32.whl", hash = "sha256:582530701bff1dec6779efa00c516496968edd851fba224fbd86e46cc6b73563", size = 36427 }, - { url = "https://files.pythonhosted.org/packages/72/6a/c5a83e8f61aec1e1aeef939807602fb880e5872371e95df2137142f5c58e/wrapt-1.17.2-cp310-cp310-win_amd64.whl", hash = "sha256:58705da316756681ad3c9c73fd15499aa4d8c69f9fd38dc8a35e06c12468582f", size = 38774 }, - { url = "https://files.pythonhosted.org/packages/2d/82/f56956041adef78f849db6b289b282e72b55ab8045a75abad81898c28d19/wrapt-1.17.2-py3-none-any.whl", hash = "sha256:b18f2d1533a71f069c7f82d524a52599053d4c7166e9dd374ae2136b7f40f7c8", size = 23594 }, -] - -[[package]] -name = "xxhash" -version = "3.5.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/00/5e/d6e5258d69df8b4ed8c83b6664f2b47d30d2dec551a29ad72a6c69eafd31/xxhash-3.5.0.tar.gz", hash = "sha256:84f2caddf951c9cbf8dc2e22a89d4ccf5d86391ac6418fe81e3c67d0cf60b45f", size = 84241 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/bb/8a/0e9feca390d512d293afd844d31670e25608c4a901e10202aa98785eab09/xxhash-3.5.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:ece616532c499ee9afbb83078b1b952beffef121d989841f7f4b3dc5ac0fd212", size = 31970 }, - { url = "https://files.pythonhosted.org/packages/16/e6/be5aa49580cd064a18200ab78e29b88b1127e1a8c7955eb8ecf81f2626eb/xxhash-3.5.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:3171f693dbc2cef6477054a665dc255d996646b4023fe56cb4db80e26f4cc520", size = 30801 }, - { url = "https://files.pythonhosted.org/packages/20/ee/b8a99ebbc6d1113b3a3f09e747fa318c3cde5b04bd9c197688fadf0eeae8/xxhash-3.5.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7c5d3e570ef46adaf93fc81b44aca6002b5a4d8ca11bd0580c07eac537f36680", size = 220927 }, - { url = "https://files.pythonhosted.org/packages/58/62/15d10582ef159283a5c2b47f6d799fc3303fe3911d5bb0bcc820e1ef7ff4/xxhash-3.5.0-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:7cb29a034301e2982df8b1fe6328a84f4b676106a13e9135a0d7e0c3e9f806da", size = 200360 }, - { url = "https://files.pythonhosted.org/packages/23/41/61202663ea9b1bd8e53673b8ec9e2619989353dba8cfb68e59a9cbd9ffe3/xxhash-3.5.0-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:5d0d307d27099bb0cbeea7260eb39ed4fdb99c5542e21e94bb6fd29e49c57a23", size = 428528 }, - { url = "https://files.pythonhosted.org/packages/f2/07/d9a3059f702dec5b3b703737afb6dda32f304f6e9da181a229dafd052c29/xxhash-3.5.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c0342aafd421795d740e514bc9858ebddfc705a75a8c5046ac56d85fe97bf196", size = 194149 }, - { url = "https://files.pythonhosted.org/packages/eb/58/27caadf78226ecf1d62dbd0c01d152ed381c14c1ee4ad01f0d460fc40eac/xxhash-3.5.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3dbbd9892c5ebffeca1ed620cf0ade13eb55a0d8c84e0751a6653adc6ac40d0c", size = 207703 }, - { url = "https://files.pythonhosted.org/packages/b1/08/32d558ce23e1e068453c39aed7b3c1cdc690c177873ec0ca3a90d5808765/xxhash-3.5.0-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:4cc2d67fdb4d057730c75a64c5923abfa17775ae234a71b0200346bfb0a7f482", size = 216255 }, - { url = "https://files.pythonhosted.org/packages/3f/d4/2b971e2d2b0a61045f842b622ef11e94096cf1f12cd448b6fd426e80e0e2/xxhash-3.5.0-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:ec28adb204b759306a3d64358a5e5c07d7b1dd0ccbce04aa76cb9377b7b70296", size = 202744 }, - { url = "https://files.pythonhosted.org/packages/19/ae/6a6438864a8c4c39915d7b65effd85392ebe22710412902487e51769146d/xxhash-3.5.0-cp310-cp310-musllinux_1_2_ppc64le.whl", hash = "sha256:1328f6d8cca2b86acb14104e381225a3d7b42c92c4b86ceae814e5c400dbb415", size = 210115 }, - { url = "https://files.pythonhosted.org/packages/48/7d/b3c27c27d1fc868094d02fe4498ccce8cec9fcc591825c01d6bcb0b4fc49/xxhash-3.5.0-cp310-cp310-musllinux_1_2_s390x.whl", hash = "sha256:8d47ebd9f5d9607fd039c1fbf4994e3b071ea23eff42f4ecef246ab2b7334198", size = 414247 }, - { url = "https://files.pythonhosted.org/packages/a1/05/918f9e7d2fbbd334b829997045d341d6239b563c44e683b9a7ef8fe50f5d/xxhash-3.5.0-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:b96d559e0fcddd3343c510a0fe2b127fbff16bf346dd76280b82292567523442", size = 191419 }, - { url = "https://files.pythonhosted.org/packages/08/29/dfe393805b2f86bfc47c290b275f0b7c189dc2f4e136fd4754f32eb18a8d/xxhash-3.5.0-cp310-cp310-win32.whl", hash = "sha256:61c722ed8d49ac9bc26c7071eeaa1f6ff24053d553146d5df031802deffd03da", size = 30114 }, - { url = "https://files.pythonhosted.org/packages/7b/d7/aa0b22c4ebb7c3ccb993d4c565132abc641cd11164f8952d89eb6a501909/xxhash-3.5.0-cp310-cp310-win_amd64.whl", hash = "sha256:9bed5144c6923cc902cd14bb8963f2d5e034def4486ab0bbe1f58f03f042f9a9", size = 30003 }, - { url = "https://files.pythonhosted.org/packages/69/12/f969b81541ee91b55f1ce469d7ab55079593c80d04fd01691b550e535000/xxhash-3.5.0-cp310-cp310-win_arm64.whl", hash = "sha256:893074d651cf25c1cc14e3bea4fceefd67f2921b1bb8e40fcfeba56820de80c6", size = 26773 }, - { url = "https://files.pythonhosted.org/packages/ab/9a/233606bada5bd6f50b2b72c45de3d9868ad551e83893d2ac86dc7bb8553a/xxhash-3.5.0-pp310-pypy310_pp73-macosx_10_15_x86_64.whl", hash = "sha256:2014c5b3ff15e64feecb6b713af12093f75b7926049e26a580e94dcad3c73d8c", size = 29732 }, - { url = "https://files.pythonhosted.org/packages/0c/67/f75276ca39e2c6604e3bee6c84e9db8a56a4973fde9bf35989787cf6e8aa/xxhash-3.5.0-pp310-pypy310_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:fab81ef75003eda96239a23eda4e4543cedc22e34c373edcaf744e721a163986", size = 36214 }, - { url = "https://files.pythonhosted.org/packages/0f/f8/f6c61fd794229cc3848d144f73754a0c107854372d7261419dcbbd286299/xxhash-3.5.0-pp310-pypy310_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4e2febf914ace002132aa09169cc572e0d8959d0f305f93d5828c4836f9bc5a6", size = 32020 }, - { url = "https://files.pythonhosted.org/packages/79/d3/c029c99801526f859e6b38d34ab87c08993bf3dcea34b11275775001638a/xxhash-3.5.0-pp310-pypy310_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:5d3a10609c51da2a1c0ea0293fc3968ca0a18bd73838455b5bca3069d7f8e32b", size = 40515 }, - { url = "https://files.pythonhosted.org/packages/62/e3/bef7b82c1997579c94de9ac5ea7626d01ae5858aa22bf4fcb38bf220cb3e/xxhash-3.5.0-pp310-pypy310_pp73-win_amd64.whl", hash = "sha256:5a74f23335b9689b66eb6dbe2a931a88fcd7a4c2cc4b1cb0edba8ce381c7a1da", size = 30064 }, -] - -[[package]] -name = "yarl" -version = "1.20.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "idna" }, - { name = "multidict" }, - { name = "propcache" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/3c/fb/efaa23fa4e45537b827620f04cf8f3cd658b76642205162e072703a5b963/yarl-1.20.1.tar.gz", hash = "sha256:d017a4997ee50c91fd5466cef416231bb82177b93b029906cefc542ce14c35ac", size = 186428 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/cb/65/7fed0d774abf47487c64be14e9223749468922817b5e8792b8a64792a1bb/yarl-1.20.1-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:6032e6da6abd41e4acda34d75a816012717000fa6839f37124a47fcefc49bec4", size = 132910 }, - { url = "https://files.pythonhosted.org/packages/8a/7b/988f55a52da99df9e56dc733b8e4e5a6ae2090081dc2754fc8fd34e60aa0/yarl-1.20.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:2c7b34d804b8cf9b214f05015c4fee2ebe7ed05cf581e7192c06555c71f4446a", size = 90644 }, - { url = "https://files.pythonhosted.org/packages/f7/de/30d98f03e95d30c7e3cc093759982d038c8833ec2451001d45ef4854edc1/yarl-1.20.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:0c869f2651cc77465f6cd01d938d91a11d9ea5d798738c1dc077f3de0b5e5fed", size = 89322 }, - { url = "https://files.pythonhosted.org/packages/e0/7a/f2f314f5ebfe9200724b0b748de2186b927acb334cf964fd312eb86fc286/yarl-1.20.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:62915e6688eb4d180d93840cda4110995ad50c459bf931b8b3775b37c264af1e", size = 323786 }, - { url = "https://files.pythonhosted.org/packages/15/3f/718d26f189db96d993d14b984ce91de52e76309d0fd1d4296f34039856aa/yarl-1.20.1-cp310-cp310-manylinux_2_17_armv7l.manylinux2014_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:41ebd28167bc6af8abb97fec1a399f412eec5fd61a3ccbe2305a18b84fb4ca73", size = 319627 }, - { url = "https://files.pythonhosted.org/packages/a5/76/8fcfbf5fa2369157b9898962a4a7d96764b287b085b5b3d9ffae69cdefd1/yarl-1.20.1-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:21242b4288a6d56f04ea193adde174b7e347ac46ce6bc84989ff7c1b1ecea84e", size = 339149 }, - { url = "https://files.pythonhosted.org/packages/3c/95/d7fc301cc4661785967acc04f54a4a42d5124905e27db27bb578aac49b5c/yarl-1.20.1-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:bea21cdae6c7eb02ba02a475f37463abfe0a01f5d7200121b03e605d6a0439f8", size = 333327 }, - { url = "https://files.pythonhosted.org/packages/65/94/e21269718349582eee81efc5c1c08ee71c816bfc1585b77d0ec3f58089eb/yarl-1.20.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1f8a891e4a22a89f5dde7862994485e19db246b70bb288d3ce73a34422e55b23", size = 326054 }, - { url = "https://files.pythonhosted.org/packages/32/ae/8616d1f07853704523519f6131d21f092e567c5af93de7e3e94b38d7f065/yarl-1.20.1-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:dd803820d44c8853a109a34e3660e5a61beae12970da479cf44aa2954019bf70", size = 315035 }, - { url = "https://files.pythonhosted.org/packages/48/aa/0ace06280861ef055855333707db5e49c6e3a08840a7ce62682259d0a6c0/yarl-1.20.1-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:b982fa7f74c80d5c0c7b5b38f908971e513380a10fecea528091405f519b9ebb", size = 338962 }, - { url = "https://files.pythonhosted.org/packages/20/52/1e9d0e6916f45a8fb50e6844f01cb34692455f1acd548606cbda8134cd1e/yarl-1.20.1-cp310-cp310-musllinux_1_2_armv7l.whl", hash = "sha256:33f29ecfe0330c570d997bcf1afd304377f2e48f61447f37e846a6058a4d33b2", size = 335399 }, - { url = "https://files.pythonhosted.org/packages/f2/65/60452df742952c630e82f394cd409de10610481d9043aa14c61bf846b7b1/yarl-1.20.1-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:835ab2cfc74d5eb4a6a528c57f05688099da41cf4957cf08cad38647e4a83b30", size = 338649 }, - { url = "https://files.pythonhosted.org/packages/7b/f5/6cd4ff38dcde57a70f23719a838665ee17079640c77087404c3d34da6727/yarl-1.20.1-cp310-cp310-musllinux_1_2_ppc64le.whl", hash = "sha256:46b5e0ccf1943a9a6e766b2c2b8c732c55b34e28be57d8daa2b3c1d1d4009309", size = 358563 }, - { url = "https://files.pythonhosted.org/packages/d1/90/c42eefd79d0d8222cb3227bdd51b640c0c1d0aa33fe4cc86c36eccba77d3/yarl-1.20.1-cp310-cp310-musllinux_1_2_s390x.whl", hash = "sha256:df47c55f7d74127d1b11251fe6397d84afdde0d53b90bedb46a23c0e534f9d24", size = 357609 }, - { url = "https://files.pythonhosted.org/packages/03/c8/cea6b232cb4617514232e0f8a718153a95b5d82b5290711b201545825532/yarl-1.20.1-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:76d12524d05841276b0e22573f28d5fbcb67589836772ae9244d90dd7d66aa13", size = 350224 }, - { url = "https://files.pythonhosted.org/packages/ce/a3/eaa0ab9712f1f3d01faf43cf6f1f7210ce4ea4a7e9b28b489a2261ca8db9/yarl-1.20.1-cp310-cp310-win32.whl", hash = "sha256:6c4fbf6b02d70e512d7ade4b1f998f237137f1417ab07ec06358ea04f69134f8", size = 81753 }, - { url = "https://files.pythonhosted.org/packages/8f/34/e4abde70a9256465fe31c88ed02c3f8502b7b5dead693a4f350a06413f28/yarl-1.20.1-cp310-cp310-win_amd64.whl", hash = "sha256:aef6c4d69554d44b7f9d923245f8ad9a707d971e6209d51279196d8e8fe1ae16", size = 86817 }, - { url = "https://files.pythonhosted.org/packages/b4/2d/2345fce04cfd4bee161bf1e7d9cdc702e3e16109021035dbb24db654a622/yarl-1.20.1-py3-none-any.whl", hash = "sha256:83b8eb083fe4683c6115795d9fc1cfaf2cbbefb19b3a1cb68f6527460f483a77", size = 46542 }, -] diff --git a/preprocessor_config.json b/preprocessor_config.json deleted file mode 100644 index 54683c2afb1e841dfe59d03ad92c060a6cf50c0d..0000000000000000000000000000000000000000 --- a/preprocessor_config.json +++ /dev/null @@ -1,13 +0,0 @@ -{ - "feature_extractor_type": "ParakeetFeatureExtractor", - "feature_size": 80, - "hop_length": 160, - "n_fft": 512, - "padding_side": "right", - "padding_value": 0.0, - "preemphasis": 0.97, - "processor_class": "ParakeetProcessor", - "return_attention_mask": true, - "sampling_rate": 16000, - "win_length": 400 -} diff --git a/pyproject.toml b/pyproject.toml deleted file mode 100644 index 70fc1fd40323347db84ccf41d86eaea2b4561b85..0000000000000000000000000000000000000000 --- a/pyproject.toml +++ /dev/null @@ -1,65 +0,0 @@ -[project] -name = "parakeet-tdt-ctc-110m-coreml" -version = "1.0.0" -description = "NVIDIA Parakeet-TDT-CTC-110M converted to CoreML format for Apple Silicon" -readme = "README.md" -requires-python = ">=3.10" -license = { text = "Apache-2.0" } -authors = [ - { name = "FluidInference" } -] -keywords = ["asr", "speech-recognition", "coreml", "apple-silicon", "nvidia", "parakeet"] -classifiers = [ - "Development Status :: 4 - Beta", - "Intended Audience :: Developers", - "License :: OSI Approved :: Apache Software License", - "Operating System :: MacOS", - "Programming Language :: Python :: 3", - "Programming Language :: Python :: 3.10", - "Programming Language :: Python :: 3.11", - "Programming Language :: Python :: 3.12", - "Topic :: Scientific/Engineering :: Artificial Intelligence", - "Topic :: Multimedia :: Sound/Audio :: Speech", -] - -dependencies = [ - "coremltools>=7.0", - "numpy>=1.24.0", - "scipy>=1.10.0", -] - -[project.optional-dependencies] -convert = [ - "torch>=2.0.0", - "nemo-toolkit[asr]>=1.20.0", -] -audio = [ - "librosa>=0.10.0", -] -dev = [ - "pytest>=7.0.0", - "ruff>=0.1.0", -] - -[project.urls] -Homepage = "https://huggingface.co/FluidInference/parakeet-tdt-ctc-110m-coreml" -Repository = "https://huggingface.co/FluidInference/parakeet-tdt-ctc-110m-coreml" -Issues = "https://github.com/FluidInference/fluidaudio/issues" - -[project.scripts] -parakeet-inference = "scripts.inference:main" - -[build-system] -requires = ["hatchling"] -build-backend = "hatchling.build" - -[tool.hatch.build.targets.wheel] -packages = ["scripts"] - -[tool.ruff] -line-length = 100 -target-version = "py310" - -[tool.ruff.lint] -select = ["E", "F", "I", "N", "W"] -ignore = ["E501"] diff --git a/scripts/run_benchmarks.py b/scripts/run_benchmarks.py deleted file mode 100644 index d8e7f9c40658f6325c539d2ee11f8a25d3e9d220..0000000000000000000000000000000000000000 --- a/scripts/run_benchmarks.py +++ /dev/null @@ -1,273 +0,0 @@ -#!/usr/bin/env python3 -""" -FluidAudio Benchmark Suite - -Runs ASR, VAD, and Diarization benchmarks and saves results to JSON. -Compare results against Documentation/Benchmarks.md baselines. - -Usage: - python run_benchmarks.py # Run all benchmarks - python run_benchmarks.py --quick # Quick smoke test - python run_benchmarks.py --asr-only # ASR benchmark only - python run_benchmarks.py --vad-only # VAD benchmark only - python run_benchmarks.py --diar-only # Diarization only -""" - -import argparse -import json -import subprocess -import sys -from datetime import datetime -from pathlib import Path - - -# Baseline values from Documentation/Benchmarks.md -BASELINES = { - "asr": { - "wer_percent": 5.8, - "rtfx_min": 200, # M4 Pro: ~210x - "description": "LibriSpeech test-clean, Parakeet TDT 0.6B" - }, - "vad": { - "f1_percent": 85.0, - "rtfx_min": 500, - "description": "VOiCES dataset, Silero VAD" - }, - "diarization": { - "der_percent": 17.7, - "rtfx_min": 1.0, - "description": "AMI SDM, pyannote-based" - } -} - - -def run_command(cmd: list[str], output_file: Path | None = None) -> tuple[int, str]: - """Run a command and optionally save output.""" - print(f"Running: {' '.join(cmd)}") - - result = subprocess.run( - cmd, - capture_output=True, - text=True - ) - - output = result.stdout + result.stderr - - if output_file: - output_file.write_text(output) - - return result.returncode, output - - -def build_release() -> bool: - """Build the project in release mode.""" - print("\n" + "=" * 60) - print("Building release...") - print("=" * 60) - - returncode, _ = run_command(["swift", "build", "-c", "release"]) - - if returncode != 0: - print("ERROR: Build failed!") - return False - - print("Build successful.") - return True - - -def run_asr_benchmark(output_dir: Path, quick: bool = False) -> dict | None: - """Run ASR benchmark on LibriSpeech test-clean.""" - print("\n" + "=" * 60) - print("ASR Benchmark (LibriSpeech test-clean)") - print("=" * 60) - - max_files = "100" if quick else "all" - output_json = output_dir / f"asr_results.json" - - cmd = [ - "swift", "run", "-c", "release", "fluidaudio", "asr-benchmark", - "--subset", "test-clean", - "--max-files", max_files, - "--output", str(output_json) - ] - - returncode, output = run_command(cmd, output_dir / "asr_log.txt") - - if returncode != 0: - print(f"ERROR: ASR benchmark failed!") - return None - - if output_json.exists(): - return json.loads(output_json.read_text()) - - return None - - -def run_vad_benchmark(output_dir: Path, quick: bool = False) -> dict | None: - """Run VAD benchmark.""" - print("\n" + "=" * 60) - print("VAD Benchmark") - print("=" * 60) - - dataset = "mini50" if quick else "voices-subset" - output_json = output_dir / f"vad_results.json" - - cmd = [ - "swift", "run", "-c", "release", "fluidaudio", "vad-benchmark", - "--dataset", dataset, - "--all-files", - "--threshold", "0.5", - "--output", str(output_json) - ] - - returncode, output = run_command(cmd, output_dir / "vad_log.txt") - - if returncode != 0: - print(f"ERROR: VAD benchmark failed!") - return None - - if output_json.exists(): - return json.loads(output_json.read_text()) - - return None - - -def run_diarization_benchmark(output_dir: Path, quick: bool = False) -> dict | None: - """Run diarization benchmark on AMI SDM.""" - print("\n" + "=" * 60) - print("Diarization Benchmark (AMI SDM)") - print("=" * 60) - - output_json = output_dir / f"diarization_results.json" - - cmd = [ - "swift", "run", "-c", "release", "fluidaudio", "diarization-benchmark", - "--auto-download", - "--output", str(output_json) - ] - - if quick: - cmd.extend(["--single-file", "ES2004a"]) - - returncode, output = run_command(cmd, output_dir / "diarization_log.txt") - - if returncode != 0: - print(f"ERROR: Diarization benchmark failed!") - return None - - if output_json.exists(): - return json.loads(output_json.read_text()) - - return None - - -def compare_results(results: dict) -> None: - """Compare results against baselines.""" - print("\n" + "=" * 60) - print("Results vs Baselines (Documentation/Benchmarks.md)") - print("=" * 60) - - if "asr" in results and results["asr"]: - asr = results["asr"] - baseline = BASELINES["asr"] - wer = asr.get("wer", asr.get("average_wer", 0)) * 100 - rtfx = asr.get("rtfx", asr.get("median_rtfx", 0)) - - wer_status = "✓" if wer <= baseline["wer_percent"] * 1.1 else "✗" - rtfx_status = "✓" if rtfx >= baseline["rtfx_min"] * 0.8 else "✗" - - print(f"\nASR ({baseline['description']}):") - print(f" WER: {wer:.1f}% (baseline: {baseline['wer_percent']}%) {wer_status}") - print(f" RTFx: {rtfx:.1f}x (baseline: {baseline['rtfx_min']}x+) {rtfx_status}") - - if "vad" in results and results["vad"]: - vad = results["vad"] - baseline = BASELINES["vad"] - f1 = vad.get("f1_score", 0) - rtfx = vad.get("rtfx", 0) - - f1_status = "✓" if f1 >= baseline["f1_percent"] * 0.9 else "✗" - rtfx_status = "✓" if rtfx >= baseline["rtfx_min"] * 0.5 else "✗" - - print(f"\nVAD ({baseline['description']}):") - print(f" F1: {f1:.1f}% (baseline: {baseline['f1_percent']}%+) {f1_status}") - print(f" RTFx: {rtfx:.1f}x (baseline: {baseline['rtfx_min']}x+) {rtfx_status}") - - if "diarization" in results and results["diarization"]: - diar = results["diarization"] - baseline = BASELINES["diarization"] - der = diar.get("der", diar.get("average_der", 0)) * 100 - rtfx = diar.get("rtfx", diar.get("average_rtfx", 0)) - - der_status = "✓" if der <= baseline["der_percent"] * 1.2 else "✗" - rtfx_status = "✓" if rtfx >= baseline["rtfx_min"] else "✗" - - print(f"\nDiarization ({baseline['description']}):") - print(f" DER: {der:.1f}% (baseline: {baseline['der_percent']}%) {der_status}") - print(f" RTFx: {rtfx:.1f}x (baseline: {baseline['rtfx_min']}x+) {rtfx_status}") - - -def main(): - parser = argparse.ArgumentParser(description="FluidAudio Benchmark Suite") - parser.add_argument("--quick", action="store_true", help="Quick smoke test with smaller datasets") - parser.add_argument("--asr-only", action="store_true", help="Run ASR benchmark only") - parser.add_argument("--vad-only", action="store_true", help="Run VAD benchmark only") - parser.add_argument("--diar-only", action="store_true", help="Run diarization benchmark only") - parser.add_argument("--output-dir", type=str, help="Output directory for results") - args = parser.parse_args() - - # Determine which benchmarks to run - run_all = not (args.asr_only or args.vad_only or args.diar_only) - - # Setup output directory - timestamp = datetime.now().strftime("%Y%m%d_%H%M%S") - if args.output_dir: - output_dir = Path(args.output_dir) - else: - output_dir = Path("benchmark-results") / timestamp - - output_dir.mkdir(parents=True, exist_ok=True) - - print("=" * 60) - print("FluidAudio Benchmark Suite") - print("=" * 60) - print(f"Mode: {'Quick' if args.quick else 'Full'}") - print(f"Output: {output_dir}") - print(f"Time: {timestamp}") - - # Build first - if not build_release(): - sys.exit(1) - - results = {} - - # Run benchmarks - if run_all or args.asr_only: - results["asr"] = run_asr_benchmark(output_dir, args.quick) - - if run_all or args.vad_only: - results["vad"] = run_vad_benchmark(output_dir, args.quick) - - if run_all or args.diar_only: - results["diarization"] = run_diarization_benchmark(output_dir, args.quick) - - # Save combined results - combined_output = output_dir / "benchmark_results.json" - combined_output.write_text(json.dumps({ - "timestamp": timestamp, - "mode": "quick" if args.quick else "full", - "baselines": BASELINES, - "results": results - }, indent=2)) - - # Compare against baselines - compare_results(results) - - print("\n" + "=" * 60) - print("Benchmark complete!") - print("=" * 60) - print(f"Results saved to: {combined_output}") - - -if __name__ == "__main__": - main() diff --git a/special_tokens_map.json b/special_tokens_map.json deleted file mode 100644 index 815fbdc56abc60a0f29e51ff2373c998d7779805..0000000000000000000000000000000000000000 --- a/special_tokens_map.json +++ /dev/null @@ -1,16 +0,0 @@ -{ - "pad_token": { - "content": "", - "lstrip": false, - "normalized": false, - "rstrip": false, - "single_word": false - }, - "unk_token": { - "content": "", - "lstrip": false, - "normalized": false, - "rstrip": false, - "single_word": false - } -} diff --git a/tokenizer.json b/tokenizer.json deleted file mode 100644 index 21e87f185d71820f5089456aa9c79c97c0a0ce8b..0000000000000000000000000000000000000000 --- a/tokenizer.json +++ /dev/null @@ -1,2456 +0,0 @@ -{ - "version": "1.0", - "truncation": null, - "padding": null, - "added_tokens": [ - { - "id": 0, - "content": "", - "single_word": false, - "lstrip": false, - "rstrip": false, - "normalized": false, - "special": true - }, - { - "id": 1024, - "content": "", - "single_word": false, - "lstrip": false, - "rstrip": false, - "normalized": false, - "special": true - } - ], - "normalizer": { - "type": "Sequence", - "normalizers": [ - { - "type": "Precompiled", - "precompiled_charsmap": "ALQCAACEAAAAAACAAQAAgMz8AgC4BQAAjSIAgMzkAgC4PQAAgSIAgMzsAgC4BQAAkSIAgMw8AADNvAAAngkAgKEJAICkCQCAgx0AAIAZAACBGQAAQx0AgDsdAIBTHQCASx0AgIAxAACBMQAApwkAgIkxAAA9WAMAPEgDAEMKAIA+aAMAAYUAAIQBAQADjQAAAokAAAWVAAAEkQAAB50AAAaZAAAJqQAACKEAAAutAAAKpQAADbkAAAy9AAAPvQAADrkAABHFAAAQwQAAE80AABLJAAAV1QAAFNEAABfdAAAW2QAAGeUAABjhAAAb7QAAGukAAB31AAAc8QAAH/0AAB75AABhOAkAax0AgGNADgBi8AgAZSgPAGSADgBn2A8AZvAPAGlwDABoMAwAa/AMAGrYDABtSA0AbBwNAG8QEgBubA0ASQoAgHAMEwBzqBMAcuwTAHUoEAB0TBAAd9ARAHYUEAB50BYAePQQAGMdAIB69BYAex0AgHMdAIB/fQEAiQwAgEGAAgDhCwCAQxgAAELAAABFSAAARGAAAEeQBgBGhAEASSgGAEhsAQBLOAcASvAHAE1wBwBMRAcAT/AEAE7MBACqCQCAUCwFAFOgCgBSEAUAVQAKAFRQCgBX0AgAVhALAFlICABYuAgAhBEAAFo8CACA9QAAgZ0AANsLAIAzHQCAg2kCAIJFAgCBNQIAgDUCAIdtAwCGVQMAgTkAAIRlAgAaDACAigEEAInVAwCI7QMAjwkAAKsLAIAsDACAjAkAADIMAICJMQMAkQkAAMzYAABbHQCAgx0AgMMaAIBPCgCAgGUDAIENAwCGPQAAgx0DAMwQAgDNhAEAgikAAMx0AwCjgQYAyxoAgICxAgCBsQIA0xoAgIEpAAClwQAA2xoAgMzoAwDNYAIAVQoAgKjxAABbCgCAYQoAgGcKAIDjGgCAgWkAAMzcBACCEQEA6xoAgG0KAIDzGgCAAxsAgAsbAID7GgCAtgkAgMygBADN3AQAzAgBALkJAICrHQCAhhEBAOEAKwDgfCcA44hIAuIMOAKjHQCAh5EBALsdAICzHQCAgNkBAIE1AADMxAIA6kRkApsdAIATGwCA7/hnAoERBwCC8QEA8BiGAolVAACB5QEAGxsAgIfhAQCAbQAAgQ0AAIN5AAB5CgCAgXkAAICVAQDMOAEAzRQBAIzBAQB/CgCAvwkAgKMVAQDDlBcAwpwUAMWEFwDEUBcAx+wXAMaAEgCTHQCAiwoAgMvQFgDK4BYAzRQWADgMAIDPvCAAzpwZANHMJADQ2CUA0+gkALFRAQA+DACAp90HAMMdAIDWvCQA2cgnANjUIgDb+CcAMxsAgIftBwCFCgCAzPgEACMbAIArHQCAh8kGALMJAICR3QcAvAkAgCsbAIBzCgCAOxsAgIsdAICPDACAjPkGAA4MAICA1QYAgcEGAMzEAgDNBAUAglEAAIN1BwCArQYAgbkGAIY1BwCHKQcAhEEAAJEKAICn7QAAQxsAgIjpBwCJzQcAlwoAgI/BBwCM3QcAnQoAgO0LAICnXQYAsJ0AAKMKAICpCgCAo0EGAEsbAIBbGwCAgAwAgFMbAIBjGwCArXEGAGsbAIDCCQCAzPgDAM0sAwDFCQCAo+UAAMgJAICMTQAAtQoAgKfxAAC7CgCAsT0GAIedAACGlQAAqB0HAISJAADBCgCAgqkAAIHVAACtAQcAzQoAgJE9AACCmQEAywkAgM0MBQDMCAUAgT0AAIeFAQCIvQEAexsAgMsdAICxCwCAjJEBAEQMAIBKDACA0x0AgID1AQCBhQEAgoEBAIOdAQCEiQEAxwoAgIapAQCHXQAAiG0AAIlNAABzGwCAzBACAIxdAACCDQAA0woAgI9JAACw6QAAgxsAgPMLAICjKQEAgCUBAIFVAQCLGwCApzUBAMykAQDNEAIA2QoAgJMbAICBNQAA3woAgK4JAQDrCgCAzOgBAM0oAgCbGwCAo/EAAIQFAACjGwCA5QoAgLMbAICotQAAqxsAgIFdAAC7GwCAzPwBAM3AAQDDGwCAyxsAgIGFAwAUDACAgeUDAPEKAICH6QMAzgkAgIylAwDTGwCA/QoAgK0JAIDbGwCAgZkDAIHdAwCMvQMAzSQBAMwgAQDMEAIAzTACAIH5AACHUQAAgFUAAIFZAAD3CgCAg0kAAIxBAADrGwCA4xsAgNEJAICBfQAAgHEAAMwgAwDNsAMAo30DANQJAICjEQMA8x0AgIEtAQCx/QAApzEDAK1BAwDrHQCAo20DAAMeAID7HQCA8xsAgKdtAwCANQAAgR0AALFtAwCILQAAmwwAgKeVAACBcQAAgFkAAINxAACj9QAAgVEAAK2BAAD7GwCAsQkDAIldAACEPQAAzDgBAISdAQCBGQAAgAkAAIRlAAADHACAzNAHAMzwBwALHACAkYkAAMxMBgDNBAYAzHAGAM10BgDMQAcAmy0PAMyoBwDNrAcAhg0AAIdVDwCEQQ8ADAsAgIIBDACDVQ8AgDUBAIHZAQCnDACAj+kAAIztAACVDACA4x0AgIv1AACIbQ8AiQ0AABILAIC3CwCAgiUAAFAMAICBQQAAVgwAgBseAIATHgCAKx4AgCMeAIAzHgCACx4AgIApAACBKQAA/wsAgBMcAICEeQAAGxwAgIFNAQCAoQEAGwsAgKP9DwDMOAIAzUgDACMcAICBWQAAzXwCAMykDQAnCwCAXAwAgKjJDwCHOQAA2gkAgImhDwAGCwCAkREAAKEMAIDdCQCAnAsAgGIMAICAuQ8AgbkPANsdAICDjQ8A+QsAgCscAICEBQAAMxwAgCELAIA7HACALQsAgIGdDwCHIQAAh7UPAMyoAgDN6AIAzLQMAM3cDACmzQAAp8UAAFMcAICPgQ8AjIkPAKPlAAAzCwCAQxwAgD8LAICxyQAAhwUAAFscAIBLHACAhz0AAGMcAIB0DACAOQsAgKMFDwCB+QAAzKgDAGscAIBLCwCAjEkAAKPxAABzHACAegwAgEULAICnlQAAgxwAgHscAIDMrAMAzcgAAOAJAICHaQAA4wkAgIG9AACCeQAA5gkAgIe5AQBRCwCAkaUAAIEdAACjHACAVwsAgIgFAACrHACAm5EAAF0LAIDpCQCAjJEBANULAIDJCwCAwwsAgM8LAICDRQAAgrkBAIG5AQCApQEAQx4AgIZxAABjCwCAhEkAAIsVAACKPQAAiTkAAIhFAACP+QAAaQsAgL0LAICMBQAAp1EBAKZJAQBoDACAsHkAAKNZAQCMqQAAgKkAAIGpAACBlQAAgJUAAK1xAQBuDACApQsAgISNAABTHgCASx4AgKMhAABjHgCAWx4AgGseAICBbQAAgG0AALEFAQCkOQAAOx4AgIscAIBvCwCAqAUAAJscAICTHACArQkAAMywAQCBvQMAgL0DAIPNAwCzHACAuxwAgMMcAIDMvAEAzYQBAInpAwDMHAEAgdkCAIDFAgDNOAEAzDwBAMxoAgDNRAIAg00AAMscAICH2QAAhy0AAIBFAACBEQAAggUAAHULAIDbHACA0xwAgOMcAIDMOAIAiBUAAIjhAACAbQAAgTkAAMyEAgDNUAEAo0UDAIQ5AQDrHACA8xwAgMzcAwDNSAIAcx4AgOwJAIB7CwCAix4AgK0MAICBbQAA+xwAgIELAICj0QAAgx4AgHseAIDMiAQAgXUAAIB1AACECwCAo7UAAMwABADNVAIAAx0AgIoLAICETQEAkAsAgAsdAIATHQCAzNAOAMwsAQDMAAUAzVwFAO8JAIDyCQCAzJgOAIHBAADMzA8AzDwOAMwIAQDNnA4AzNQPAM14DwDMPA4AzTgOAIHlAQCA5QEAg+UBAILlAQDXCQCAhOUBAIfhAQBHHQCAiaUBAIjZAQCByQcAPx0AgFcdAIBPHQCAzDQBAPgJAICA3QAAgekAAEYKAICD/QAAgM0AAIH5AACBEQcAbx0AgGcdAICJ0QAAzCgBAH8dAIB3HQCA5AsAgMw0AQDeCwCAgF0AAIFlAACjAQEAg2EAAIFxAACASQAANx0AgB0MAICuCwCAiVUAAC8MAIA1DACAXx0AgIcdAIDHGgCAUgoAgIIdAACDeQcAgBkHAIEZBwCGIQAAhykAAISRBwD1CQCAimkAALHZBgCIaQAAifUHAEwKAICP3QcAjNkHAIwMAID7CQCALx0AgP4JAICRoQcAgEEHAIFBBwCHBQAAzxoAgIKRBwDXGgCA3xoAgKOVBgCGhQcAp+0AAMyQAgDN4AUAsekAAKPBAABYCgCAXgoAgGQKAIBqCgCAAQoAgKVlBwDnGgCAzLgDAKhVBwDvGgCAcAoAgPcaAIAHGwCADxsAgP8aAIAECgCAo60AAAcKAICMJQYACgoAgIxNAACvHQCAgm0AAIE9BgCCAQYAgWUAAKcdAICHZQAAvx0AgIcRBgCHrQEAtx0AgMxQAgDNxAIAgeEBAIDJAQCD4QEAkYkAAID9AQCB1QEAnx0AgIydAQCJNQAAdgoAgIB1AACBXQAAhi0AAIc1AACEfQAAFxsAgIKFAQCDfQAAgJ0BAIGRAQAfGwCAj+kAAIzhAAB8CgCAggoAgA0KAICIDQAAifkAAKc5AQCXHQCAjgoAgDsMAICjJQEAQQwAgLBZAACPHQCAggUAAMcdAICtFQEAkgwAgDcbAICGBQAAiAoAgCcbAIAvGwCAp2kAAIANAQCBAQEAhzEAAKNJAACxGQEAzBACAD8bAIARDACAlAoAgK1RAADM1AEAzfgBAKhBAABHGwCAzTgBAMw8AQCB7QMAmgoAgKAKAICMDQAA8AsAgKYKAICBxQMAzGgCAKwKAICCxQMATxsAgITJAwCHKQAAhjEAAF8bAICCbQAAgwwAgFcbAICHYQAAZxsAgG8bAIAbHQCAzKgDAM2sAgCB+QAAiC0AABAKAIATCgCAFgoAgIw1AAC4CgCAvgoAgLHVAADECgCAfxsAgM8dAIC0CwCAzDABAEcMAIBNDACA1x0AgMwEAQDKCgCAdxsAgKelAADWCgCAo40AAMwUAgCAuQAAgbkAAKeFAAALDACAgmUAAIcbAICMNQAA9gsAgMzsHADN/AMAjxsAgK6tAADcCgCAlxsAgMzABgDN0AYAsL0BAMyQBwDiCgCAgckBAMwYHQDNIAIAhBEAAO4KAIDNuAYAzKwGAKcbAIDoCgCAgSkAALcbAICvGwCAo+0BAMxAHQDNEAIAvxsAgMcbAICBCQAAzxsAgMxAHQDN0AIAqNkBABcMAIDMkAcAzBwBAMxgBgDNZAYA9AoAgB8KAIDXGwCAkSkBAAALAICBzR8A3xsAgPoKAIDvGwCA5xsAgMzEBgDNwAYAgTEAAIDZAAAiCgCAJQoAgIK5AQCDRQEAgLkBAIG5AQCGXQEA9x0AgIRdAQDvHQCAzcAAAMzwAACIARwAiXkBAAceAICPVQEAjGEBAP8dAICB3R4AgRUfAJ8bAICBXR8AjIEfAIdBHwDMGAMAzWgDAIBNHwCBpR8AKAoAgIOpHwCMFR8AjNEeACsKAICHtR8AgJUfAIGZHwCBEQAAg70fAICFHwCBiR8A9xsAgIQ9AACeDACAiZkfAP8bAICIBQAACQsAgAccAICADQAAgf0AAA8cAICj2R8Ao3keAKOFAAAPCwCArTUfAKdhHgCnqR8ApAwAgIQNAACqDACAozUfAC4KAICtiR8AhHEAAKchHwCxPR4AsYUfAJgMAIDnHQCAFQsAgLoLAIDMtBwAzbAcAFMMAICxQR8AWQwAgJ8LAIAfHgCAFx4AgC8eAIAnHgCAgLkeAIG5HgCCIQEAgzUBAIRhAQA3HgCAhokBAIe9AQCIkQEAiekBAN8dAICL/QEAjOUBAIINAAAPHgCAj90BAIO5AQCRrQEAgb0BAIC9AQCAoQEAgaEBAPwLAIACDACAhD0AABccAICJlQEAm4EBAIHNHgCAzR4AzPwCAM3wAgCB5QAAHxwAgIHtAACjpQAAzJABAM1cAgCHHQAAHgsAgKj5AAAnHACAKgsAgF8MAIBlDACALxwAgIQFAAA3HACAo9UAACQLAIA/HACAgVEAAMz0AQDN0AEAMAsAgIc9AABXHACANgsAgEccAIBCCwCAhwUAAF8cAIBPHACAh/EDAIHZAwCBmQMAgZEAAGccAIB3DACAjPkDAMwkAQCHuQMAgfkDADwLAIDMZAIAgskDAIyZAwBvHACAh9EDAI+RAwCB3QYAkfUDAMwABADN7AMAh2UAAB8dAIBOCwCAdxwAgH0MAIBICwCAzBgBAIg5AACHHACAfxwAgMxcAwCMJQAAMQoAgMwsAQCx/QAAozkDADQKAIA3CgCApxwAgKdZAwDMdAMAiAkAAKNRAwCvHACAYAsAgINtDQCnnQAApq0AAKOdAACxDQMAzCgBANgLAICntQAAprUAAMwLAIDMMAEAgdUHAMYLAIDMKAEA0gsAgEceAIBmCwCArYkAAGwLAICAzQEAgd0BAMxEAQDNnB4AhPUBAMALAIDMWAEAzUwBAIDtAQCB/QEAg7UAAGsMAICM3QEAcQwAgMwIHgCM8QYAzDgBAM08AQBXHgCAiREAAIEFBgBPHgCAZx4AgF8eAIBvHgCAgz0AAIAhAACBOQAAgDkAAIEhAAA/HgCAjxwAgMwoAQCB2QYAcgsAgIH9BgDMJAEAnxwAgJccAIC3HACAgCEBAIE1AQCjBQAAvxwAgMccAIDPHACAzIwFAM1AAgC3HAMAeAsAgIfNBwDfHACA1xwAgCMdAIDNiAAAzJAAAIzdBQCjhQAAGQoAgMzgAgDnHACAiNUHAIFNAACATQAAVAsAgO8cAIBaCwCAkTkHADoKAICIxQcAqAsAgIrJBwD3HACAmz0AAIflBwB3HgCAgYUHAICFBwA9CgCAgvkHAILVBgCDRQAAgMkGAIHdBgCG4QYAfgsAgIRRAACPHgCAipUGAIuZBgCIeQAAiZ0GALAMAICPWQcAjG0HAP8cAIDMgAMAzSQCALARBwBACgCAhx4AgCcdAIB/HgCAhwsAgICNAACBnQAAzOwDAM3oBAAHHQCAjQsAgKNJBwCTCwCADx0AgKO9BwAXHQCAGwAAgOoHAIALAACApKUHAOsEAICKBQCAAwAAgKhhBwDfDQCAZQAAgMgDAIAeCQCArWkHAIAtAQCBPQEAgl0BAINRAQCEYQEAuAQAgKwEAICHYQEAiK0BAIm1AQCKvQEAjykVALwFAIAgDACAzHgCAM3YBQCB3QEAgXEAAOcLAICC/QEAhBkAACYMAICH7QEAIwwAgMw0BADNMAQA6gsAgJ9pFQApDACAjMkBAM34BADM8AIAsUkBACEHAICB1QAAoxUBAKCZFQB2CACARgcAgIT1AADMKAQAzSwEAMYIAICveQEAqH0BADcNAICqaQEAVQkAgLQlAQC1KQEAowkBAAUMAIDqBgCA7gYAgLIFAQCzPQEAvPUAAL39AAC+2QAAOwgAgLgBAQC5AQEAugEBADwHAIBDBwCAhgwAALOdAwCyiQMAtggAgIC9AwBsBwCAbwcAgBUJAIDkBgCA5wYAgDgIAICJhQMAzOQHAL+hAwAIDACA2gwAgIxlAADN5AwAzCQMAIlBAACIVQAAi0UAAIpFAACFtQMAhLUDAIeVAwCGgQMABA0AgAcNAIAKDQCAmCwAABMAAICmyAAAzYwGAMyoBgCFaQAAFwAAgDEAAIBpAACAzPADAAcAAIA1AACA1AwAgLGVAAArDQCAs5UAALKVAAA7DQCAPg0AgEYNAIBBDQCANA0AgHUAAICmBgCAJQAAgJsJAIAjIQCAv1UDAEkNAIAfIQCAGyEAgGcgAIC4bAAAlGUNAJIAAgCcrQEAnaUBAJqJAQCbiQEAmJkBAJmJAQDMIAYAzQQGAMxABgDNXAYAzDwHAM04BwDMvAcAhXUAAIABDwCBDQ8AbyAAgLqZAQCFBQAAdyAAgF8gAIC+hQEAgSkPAIAlDwBrIACAgiEPAIUpAAC0pQEAhREAAHMgAICziQ8AsoUPALHJAQCwAQwAt4EPALbtAQC17QEAtO0BAIFlAQCAZQEAg2EBALi1DwDMPAsAhHkBAIDhDwCB3Q8AeyAAgGMgAIDMyAQAzbgEAIWtAACFFQAAJyEAgD8hAIDM6BkAzbQZAKRdAQBMDQCAok0CAKPxDwCgVQEAod0PAIIIAIBxCQCAPgkAgPMeAIBvCQCA+x4AgHoJAID3HgCAtAgAgJMNAACzHgCA/x4AgITVDACF6Q4AlGkAAIfdDgC7HgCAmbQCAMMeAIDLHgCAtx4AgEMhAIC/HgCAn3QBAMceAICRGA0AgI0OAIGBDgCGhQ4AlYwDAISJDgCXRAIAghEAAKm4AACA0QAAge0AAM8eAIBPDQCA6x4AgIVZDwCDiQAAoTQNAIFFDgCASQ4A7x4AgKU0AQCFYQ8AzPAUACMfAIC5xAUAzMgDAM3cAwCA3QAAgcEAACsfAIC/kAUAhREAALHsBwCA9QAAgcEAAKcgAIC1jAYAMx8AgLdABgCA3Q4AgekOAMwoAgDNtAIAgM0OAIH5DgCFKQAAg4UBAIB1AQCBsQEAgPEBAIHVAQCvIACAOx8AgIUFAAC3IACAgJkBAIG9AQCCfQAAk9UBAJThAQCFDQAAnyAAgCcfAICACQAAgRkAAC8fAICTrQEAlC0AAKsgAICFDQAANx8AgIUFAACzIACAPx8AgIUpAACCGQAAhTUAAIDxAACB4QAAuyAAgKMgAIBHIQCAhQUAAGchAICDdQEAgO0BAIEpAQDM8AEAzbABAFINAIBjIQCAXyEAgKkNAIBjHwCAax8AgIA9AACBDQAAcx8AgHsfAICALQAAgR0AAIIVAABnHwCAzSwBAG8fAIB3HwCAfx8AgIjFAwCrIQCAzJACAM28AgCE7QMAVQ0AgIb5AwCjHwCAgIEDAIH9AwCAPQAAgTUAAIFJAACAQQAAzdwBAIJBAACrHwCApx8AgK8fAIDNMAEAlJ0DAJMhAIDN8AEAzAwBAIG5AwCAxQMAg6EDAJOlAwCArQAAgdUAAICdAACBqQAAjyEAgFgNAICBwQAAgMkAAIC1AACBgQAAiyEAgINpBADMcAMAzbQDAIchAIDNPAEArA0AgJMBBADNjAIAzPQCAIANAACBNQAAlNkGANcfAIDbHwCA3x8AgMwIAQDNHAEAgREAAIApAACvIQCAghkAAICRAQCBkQEAzWgFAMyUAgDMEAkAzSgWAMxYDgDNeA4AzBQNAM3YCgDMKAwAzYwNAMzgFwDM4AoAzDgLAM30CACFEQAAWw0AgIBRBwCBUQcA5yAAgM2QDgCFBQAA7yAAgMzYDgDN7AEA9yAAgM0ADgCFGQAAzfAPAM08DgDNVA4AzGgBAM1sAQDfIACAZAgAgJSZBwDMwDsAgGEBAIHZAACFKQAAzWQOAMx4AQDNfAEAga0HAICtBwCFZQAAgp0HAIBRAQCBUQEAlOEHAM3AAACEeQEAk8UHAIZhAQDrIACAiCEBAIUNAADzIACAzRgBAMzYAADNtAAAgN0HAIHNBwCfHwCAhQkAANMfAID7IACAAyAAgOMgAIALIACAEyAAgBsgAIAPIACAByAAgLMhAIAXIACAHyAAgMy4AgDNHAMAgGUAAIF1AACCfQAAIyAAgIUJAACFQQAAByEAgK8NAICAmQYAgSEHAIUZAACDfQAADyEAgIVZAAADIQCA/yAAgIDNAACB2QAAkx4AgIURAACE6QAAmx4AgIblAABHIACAgDUAAIENAACjHgCAhR0AAE8gAICrHgCAhQUAAFcgAICAVQAAgW0AAIJ9AACTRQAAlA0AAIUNAAA/IACAlx4AgIAJAACBEQAAnx4AgIUdAABLIACApx4AgIUFAABTIACAgOkBAIHxAQCCBQAArx4AgIUJAACFCQAAWyAAgEMgAICAbQEAgXkBAIIZAACDpQEAEyEAgIV1AACFBQAAFyEAgAshAIAnIACAzMgCAM3cAgCyDQCA0x4AgIA5AACBOQAA2x4AgOMeAIDXHgCA3x4AgIAdAACBDQAA5x4AgCsgAICAxQAAgdUAAM3AAADMJAIAgNUAAIHFAACFOQAAg8kAACshAIC1DQCAgNUAAIEJAACFBQAAMyEAgAMfAICHIACAgAkAAIERAAALHwCAk5kAAJS5AAATHwCAhWUAAIU9AACPIACAk10AABsfAICFEQAAzXAFAMx0BQCUATwAlyAAgH8gAIDNKAEAiyAAgJMgAICFGQAAmyAAgIMgAIA7IQCALyEAgC8gAICFJQAAhTkAAMz4AgDNxAMAzTwBALgNAICBlQMAgI0DAM3EAQCCpQMAhVEAAIVJAADMKAEAzSwBAM04AQDMPAEAgGk+AIFpPgBPIQCASyEAgM04PADMVDwAgdE8AJOdPgDMSAEAzcgCAM00AQBTIQCAlLk+AF4NAICAoT4AgaE+AIKhPgCIjTwAWyEAgIWtAACALQAAgSEAAIXVPwCbHwCAgO0AAIHxAACGpQAASx8AgISpAADNJAEAzSgBAFMfAICI+T4AhfE/AFsfAIBPHwCAhcU/AM0wAQDNEAEAzfQGAIDdAQCB6QEAzbwGAM1wBgDM4AYAzVwBAMxoBgDNkAYAzWQGAM14BgDMrAcAzagHAMzoBwDNyAcAgk0/AIP9AgCANQIAgekCAFcfAIBfHwCAgAU9AIV9AQBXIQCAMyAAgM0UAQAvDgCAge0BAIDhAQDNPAEAgs0BAM0sAQCCdQEAgW0BAIBZAQCAZQEAgcUAAIsfAIDNJAEAzTgBAILxAACB+QAAgFkBAIApAACBcQAAzBgBAM18AQDNLAEAkx8AgIEdAACAHQAAjx8AgJcfAIB3IQCAzSQBAMzkPQDNXA8AzegAAMwMAQCA1QEAgckBAIKZAACD5T8ADx8AgBcfAIAfHwCANyEAgCkOAIB7IQCAQx8AgDcgAIBHHwCAMg4AgIBNPwCBQT8Agx8AgG8hAICHHwCAayEAgIAlPwCBKT8Ak5E/AIN9AAAsDgCAlEEAAMzYAgDNrAIAcyEAgJNVAACACQAAgR0AALsNAICDIQCAlEEAALMfAICAnQAAgaEAAIAdAACBEQAAhKUAALsfAICGpQAAwx8AgIjxAACC0QAAgdkAAIDNAACAJQAAgSkAAIIFAADLHwCAtx8AgL8fAIDHHwCAk7EAAJQRAADPHwCAgB0AAIEVAACAJQAAgS0AAII9AAB/IQCAgO0AAIHRAACCFQAAg4EAAIHQPQA7IACAzCACAM3cAQCFeAIAlyEAgDUOAICfIQCAiRgDAOMfAICALQAAgTUAAIAJAACBbQAA6x8AgMcgAICRsQAAkKkAAJPdOwCSAQQAlaUAAJSVOwDzHwCAlqEAAIUJAACTQQAAzyAAgPsfAICFBQAA1yAAgJT1AAC/IACAgLkAAIHdAACC5QAA5x8AgO8fAICF6QAAgAkAAIE1AACFBQAAyyAAgPcfAICFHQAA0yAAgP8fAICFBQAA2yAAgLHBBQCwxQMAwyAAgLLFAwC12QUAtM0DAKMhAICFOQAAuf0DAKchAICbIQCAwQ0AgNMNAIAdDgCABx8AgAsOAIDZDQCAzIgCABEOAIDN4D4AzZABAMwkAQB2DQCAlA0AgEcOAICDDgCAgLEAAM3UPgDN5D4AiQ4AgMy8PgDNuD4AgNEDAIHtAwCC/QMAhmkAAEQOAICFnQMAzTwBAD4OAIDM6AIAzTw/AIjlAADNGAEAjw4AgIhBAABBDgCAfQ4AgM0sAQCbDgCAgNUAAKEOAICG4QAAhukAAE0OAIDNJAEApw4AgM0QAQCI0QAAiCkAAMz4AgBTDgCAzfgCAMwkAQCtDgCAhS0DAMygPgDNbD4AgNUDAIHNAwCCAQMAg/kDAMxkAwDNzAIASg4AgM0kAQDMDAIAzQgCAIERAADMnAMAzLA+AM20PgDMxD4AzcA+AMyAPgDNuD4Asw4AgMyEAgDMmD8AzVA+AMwgPgDNoD4AzQw/AM0wPwDNeD8AzQQ/AIhZAADFDgCAzfgBAMzEAQBQDgCAyw4AgNEOAIDMFAIAzAgBAM3IAQCIBQAA1w4AgN0OAIDMKAIAvw4AgIgNAACG0QAAgB0BAITNAACI9QAAzDwCAIQ1AQDMRAIAhikBAIYOAICIZQEAjA4AgKdEBQBoDgCAi+0AAIjtAACBDQAAiCUAAIZlAADMcAIAzXQCAMwwAgDN2AUAYg4AgJIOAICAOQAAZQ4AgMzgBQCADgCAzCgBAM0UAQCGJQAAiFUAAA4OAICGhDAAyg0AgIDVBwCG/QcAng4AgMwkAgCIPQAApA4AgHEOAICIPQAAqg4AgMxIAgDNeAIAVg4AgLAOAICXwAUAlnAFAJUYBQCAaQAAk1gFAIE5AACIZQAAkPg8AIZZAACeqAUAhEUAAG4OAIDM1AIAmrQFAIBdAACYrAUAp+wEAIgRAADM2AIAzdwCAKO8BAC2DgCAzGACAMgOAIB0DgCAzg4AgK0IBADUDgCAq/QEAMwsAgCIBQAA2g4AgLfoAwC2HAQAtSgEAMwAAgCzKAQAi3kAAIh9AACwdAQAhkEAAL6kAwCEdQAAiB0AAOAOAIC6TAMAzNwDALj8AwCDqAIAiA0AAMIOAICIFQAAh5QCAMw4AgBrDgCAzAQCAIvcAgCPDQAAdw4AgI8ZAADMIAIAeg4AgI3wAgCIdQAAmCADAJksAwCVDgCAmg0AgMxMAgCWcAMAzCQCAIg9AACYDgCAzCwCAIgFAAC5DgCAzCQCAIgNAAC8DgCAh/UAAKjUAwCpxAMA4w4AgNlgAgDYDwCA2w8AgOEPAICUNQAAkzEAANloAgDeDwCA2UwCAJQFAADkDwCAlSEAAJQpAABWEACAehYAgEkXAIDYFgCA2WACAD0XAIC12AMAtPADAJQ1AADZWAIAYBcAgJQFAADZVAIAlA0AADcXAIDgdAEAisgAALwVAACIyAAA4IACAI0XAICBoAAApOwCAKTIAgCoXAAAvA0AAJ8XAIDghAIAvAUAAKMXAICk+AIA4PQCALDMAwCV0AAAYxcAgLPgAwCmyAIAp2ACAJLYAABqFwCAvsEAAHEXAICXwQAAeBcAgH8XAICGFwCAzXg/AMy8PwC+gA0AkRcAgLx4DAC9gA0AuvQMALtUDAC49AwAmBcAgLwXAIC3uAwAwBcAgLWMDACyoAMAs6AMAKcXAICxQAMArnACAK9kAwC4BQMArUgDAK4XAIC1FwCAqEQDAKnYAwDgFwCAp9gDAKRoAgCliAMAtjUDALc9AwCSyAIAtT0DAJldAQCYTQEAm2UBAJppAQCdZQEAnGUBAJ+FAQCemQEAh5wCAL6tAACWpQAAl70AAMw0BQDNjDcAzLg4AM2sOACflQEAth0AAJ2ZAQCc9QEAs7EBAK54AgDnFwCAxBcAgJk9AADLFwCAmxkAAJoJAADSFwCA2RcAgOBIAgCeCQAArFwCAK30AgAAGACA/BcAgAQYAIDuFwCAh2ADAPUXAICvVAIAvhEAAJcFAAAIGACA4KwCAAwYAICG+AMAh+wDAOC0AgAUGACAr0gCAK6QAgDgPAIAvg0AABAYAICXGQAA4NgCAIaEAwCWEQAAvwAMAJ1tAACcYQAAGBgAgLFMAgCzUAIAlQ0AABwYAICGnAMA4MgCALMEAgCCBQAAKBgAgLNQAgCVDQAALBgAgCAYAIAkGACA4LQCAIaMAwCH3AMAvg0AAJVpAACWeQAAMBgAgLToAgC1UAIAlwUAADgYAIDg1AIAtPQCAL4ZAADgoAIANBgAgODUAgCZjAMAt9QCAIoFAAA8GACAQBgAgIoVAAC3NAIAjx0AAEQYAIBIGACAswUAAEwYAICzBQAAYRgAgJwJAACdCQAAUxgAgFoYAICMBQAAaBgAgHMYAIB6GACAgRgAgJ9JAACIGACAjxgAgGwYAICWGACAnRgAgN8YAIDVGACA8BgAgOYYAICkGACAg8kBAIH5AQCyGACAuRgAgMAYAIDHGACAzhgAgKsYAICAtAIApYgDAOEIAgCuHQAA9xgAgLwJAACN9QEA+xgAgOEAAgCSlQEA45QQAJNFAACXiQEAhRQAAId4AQCGAAQAVzoAgFs6AIBfOgCAYzoAgGc6AICdeQAA74xoAJyhAQBrOgCAbzoAgKKZAABzOgCAdzoAgHs6AIB/OgCAp4kAAIM6AICHOgCAqUkBAIs6AICsqQAAjzoAgJM6AICXOgCAsyUBAJs6AICfOgCAozoAgLchAQC2OQEAtTEBAKc6AICrOgCAufkAALkRAQC4GQEArzoAgLM6AIC3OgCAuzoAgICwAQCEiAIAvzoAgIPIAQCEVAMAhFwEAMM6AICEXAUAgN0DAIEtAACCMQAAvjwCAMs6AIDPOgCAh4gDAIacBACzLQMA0zoAgNc6AIC+AAQAvhwFALbRAwC12QMA2zoAgLv5AwC68QMAmljTAYTgBwC/xQMAvtkDAL3dAwC83QMAvgAYAKUFAwCmDQMA3zoAgIQcGADjOgCA5zoAgKPxAwCsAQMArQEDAK4FAwCvGQMArKQbAq3cGgKqLQMAqyUDAL5MGQC+SBoA6zoAgL6AGwC04BoCtdQdArYwHgLvCAIA7zoAgOGgAQC6OBoC4/gCALoAAAC9ZBwCvvQcAr8AEAKRBNMBkOT2AeBEAQCSCD4C8zoAgPc6AID7OgCA/zoAgL6sHAADOwCABzsAgAs7AIAPOwCAEzsAgBc7AIAbOwCAgbBtAICAAQCDHFIAgth3AIUgmgCEkL4AhwjPAIaM5gCJbDcBiOAsAYsYfgGK2BMBjeClAYzwWgGP/OsBjliPAbDVFwCxAWgAso1rALOdawC0SWsAtZVvAB87AIDgcAEAIzsAgCc7AIArOwCALzsAgIAZAACBGQAAggUAADM7AIA7OwCAoaUCAKJJBwCjQQcApEEGAKXVGwCm3RsAp8EaAKgBHACp4R8AqkkfAKsBEACs9RMAra0TAK4BFACv+RcAqDEGAKkxBgCqTQYAq0UGAKxNBgCtmQYAro0GAK+FBgCGgAMAhxgDAD87AIBDOwCARzsAgEs7AIBPOwCAUzsAgLhtBwC5dQcAun0HALt1BwC8bQcAvc0HAL75BwC/+QcAsKkGALGFBgCyeQcAs3kHALRpBwC1aQcAtl0HALdVBwDHOgCAs8EGAFc7AIA3OwCAth0GAFs7AIBfOwCAtcEGALppBgC7RQYAYzsAgGc7AIC+qQcAv6kHALypBwC9qQcAo4UGAGs7AIBvOwCAczsAgHc7AICmWQYApYUGAHs7AICrAQYAqi0GAH87AICDOwCAr+0HAK7tBwCt7QcArO0HAKjBBgCpLQEAqiUBAKs9AQCsJQEArS0BAK4lAQCvlQEAhzsAgIs7AICPOwCAkzsAgJc7AICCvQAAgb0AAIC9AAC4nQEAua0BALqlAQC7bQAAvHUAAL19AAC+dQAAv20AALD1AQCx/QEAssEBALPBAQC0tQEAtb0BALa1AQC3rQEAmzsAgJ87AICjOwCAs6EBAKc7AIC1oQEAtqEBAKs7AICGgAEAh8QBALo9AQC7NQEAvBkBAL0ZAQC+fQEAv3UBAKPtAQCvOwCAszsAgLc7AIC7OwCApu0BAKXtAQC/OwCAq3kBAKpxAQDDOwCAxzsAgK85AQCuMQEArVUBAKxVAQDLOwCAzzsAgNM7AIDXOwCA2zsAgOGsAQDfOwCA42AGAOM7AIDnOwCA6zsAgO9UBgDvOwCA8zsAgL60GgD3OwCA+zsAgP87AICGaBwAh4wDAAM8AIAHPACACzwAgA88AICAOQAAgTkAAIIFAAATPACAGzwAgB88AIAjPACAJzwAgKgdAwCpQQMAqkEDAKtBAwCsQQMArUkDAK5xAwCvcQMAhCAdACs8AIAvPACAMzwAgDc8AIA7PACAPzwAgEM8AIC46QAAufUAALr9AAC78QAAvJEAAL2RAAC+iQAAv4kAALDhAACx4QAAsuEAALPhAAC04QAAte0AALbZAAC32QAA4wwHAOEgBwDhMAEA4wgHAEc8AIBLPACATzwAgFM8AIBXPACAWzwAgF88AIBjPACA75gHAGc8AIBrPACA74gHALOJAgBvPACAczwAgL6AGgB3PACAtokCALWJAgB7PACAu2UBALplAQB/PACAgzwAgL9pAQC+ZQEAvXUBALx1AQC3PQYAtj0GALU9BgC0IQYAszUGALI1BgCxAQYAsAkGAL9ZBgC+UQYAvVkGALxNBgC7bQYAunkGALlxBgC4eQYAgJ0AAIGtAACCpQAAizwAgI88AICTPACAlzwAgJs8AICvcQYArmkGAK1tBgCsbQYAq4EGAKqZBgCpkQYAqJkGABc8AICHPACAnzwAgKPFHQCjPACApcUdAKbFHQCnPACAhgADAIdkAwCqKR4AqykeAKw5HgCtOR4ArikeAK8lHgCzOR4AqzwAgK88AICzPACAtzwAgLb9HgC1/R4AuzwAgLvZHgC60R4AvzwAgMM8AIC/aR8AvmEfAL1pHwC8wR4AqPEeAKnxHgCq8R4Aq/EeAKw1HgCtPR4ArjUeAK8tHgDHPACAyzwAgM88AIDTPACA1zwAgNs8AIDfPACA4zwAgLjlHwC57R8AuuUfALv5HwC86R8AvZEfAL6RHwC/jR8AsFUeALFdHgCyVR4As/0fALTlHwC17R8AtuUfALfdHwCjeR8A5zwAgOs8AIDvPACA8zwAgKa9HwClvR8A9zwAgKuZHwCqkR8AhogAAIdMAQCvKR4AriEeAK0pHgCsgR8AgEkAAIFJAACCWQAAs5keAPs8AIC1iR4AtlEBAP88AIADPQCABz0AgLotAQC7JQEAvD0BAL0lAQC+JQEAvxUBAKhNHgCpVR4Aql0eAKtVHgCsTR4ArZ0BAK6JAQCvgQEAhKwBAAs9AIAPPQCAEz0AgBc9AIAbPQCAHz0AgCM9AIC4ZQEAuW0BALplAQC7fQEAvGUBAL1tAQC+ZQEAv9kAALClAQCxrQEAsqUBALO9AQC0rQEAtZ0BALaVAQC3XQEAo9UdACc9AIArPQCALz0AgDM9AICmHQIApcUdADc9AICraQIAqmECADs9AIA/PQCAr1kCAK5pAgCtaQIArHECAEM9AIBHPQCASz0AgE89AIBTPQCAVz0AgFs9AIBfPQCAgDkAAIE5AACCBQAAYz0AgGs9AIBvPQCAh0ADAIZcBACETAQAcz0AgHc9AICEBAUA4yABAHs9AIDhqAEAfz0AgO+UGgCDPQCAhz0AgIs9AICPPQCAkz0AgJc9AICbPQCAs6EDAJ89AICjPQCApz0AgKs9AIC2fQMAtX0DAK89AIC7WQMAulEDALM9AIC3PQCAv/0AAL79AAC9/QAAvEEDAKhRAgCpWQIAqmkCAKtpAgCstQIArb0CAK61AgCvrQIAhKgHALs9AIC/PQCAwz0AgIKpAADHPQCAgKkAAIGpAAC4aQEAuWkBALoJAQC7CQEAvBkBAL0ZAQC+CQEAvwkBALDVAgCx3QIAstUCALNpAQC0eQEAtXkBALZpAQC3YQEA4bgBAOHUHwDjOB8A4wwbAMs9AIDPPQCA0z0AgNs9AIDfPQCA4z0AgOc9AIDrPQCAvjwJAO89AIDvhBsA74QbAKOhAgDzPQCAhugEAIe8BQD3PQCApn0CAKV9AgD7PQCAq1kCAKpRAgD/PQCAAz4AgK/9AQCu/QEArf0BAKxBAgCzhQYA1z0AgAc+AIALPgCADz4AgLaJBgC1jQYAEz4AgLuRBgC6iQYAFz4AgBs+AIC/9QYAvokGAL2BBgC8iQYAHz4AgCM+AIAnPgCAKz4AgC8+AIAzPgCANz4AgO+EHQA7PgCA4QAEAD8+AIDj/AQAgBEAAIEdAACCBQAAQz4AgKjxBgCp8QYAqg0GAKsFBgCsBQYArQkGAK49BgCvNQYARz4AgEs+AICGiAAAhxADAE8+AIBTPgCAVz4AgFs+AIC4EQYAuRkGALohBgC7IQYAvPUHAL39BwC+9QcAv+kHALBNBgCxVQYAsl0GALNVBgC0TQYAtTEGALYxBgC3MQYAo4UHAF8+AIBjPgCAZz4AgGs+AICmiQcApY0HAG8+AICrkQcAqokHAHM+AIB3PgCAr/UHAK6JBwCtgQcArIkHAHs+AICz4QYAfz4AgIM+AIC25QYAhz4AgIs+AIC18QYAur0GALuNBgCPPgCAkz4AgL59AQC/ZQEAvJUGAL11AQCoHQYAqSUGAKotBgCrJQYArD0GAK0hBgCuXQYAr00GAJc+AICbPgCAnz4AgKM+AICnPgCAgrkDAIGxAwCAuQMAuO0BALmFAQC6jQEAu4UBALydAQC9hQEAvo0BAL+FAQCwPQYAsQ0GALIFBgCz5QEAtP0BALXlAQC25QEAt9UBAKOlBQCrPgCArz4AgLM+AIC7PgCApqEFAKW1BQC/PgCAq8kFAKr5BQCGCAwAhxwDAK8hAgCuOQIArTECAKzRBQDDPgCAs/ECAMc+AIDLPgCAtlUDAM8+AIDTPgCAteECALpxAwC7eQMA1z4AgNs+AIC+MQMAvz0DALxRAwC9UQMAqCUCAKk1AgCqPQIAqzUCAKwtAgCtkQMArpEDAK+RAwDfPgCA4z4AgOc+AIDrPgCArAAAAO8+AIDzPgCA9z4AgLiZAwC5rQMAuqUDALttAwC8dQMAvX0DAL51AwC/bQMAsPEDALH5AwCywQMAs8EDALSxAwC1vQMAtrUDALepAwD7PgCA/z4AgAM/AIAHPwCACz8AgA8/AIATPwCA76gaAL5oDADhlAEAFz8AgOMcBgCADQAAgXEAAIJxAAAbPwCAo/UDAB8/AIAjPwCAhEwCACs/AICmUQIApeUDAC8/AICrfQIAqnUCAIbIDACHLA0ArzkCAK41AgCtVQIArFUCAOFQBgAzPwCA4xQHAITADAA3PwCAOz8AgD8/AIBDPwCARz8AgEs/AIBPPwCAUz8AgFc/AIBbPwCA73gbAL74DwBfPwCAYz8AgGc/AICzjQEAaz8AgLWZAQC2jQEAbz8AgGc9AIBzPwCAuoUBALtNAQC8VQEAvV0BAL5VAQC/SQEAo0EOACc/AIB3PwCAez8AgH8/AICmQQ4ApVUOAIM/AICrgQ4AqkkOAIc/AICLPwCAr4UOAK6ZDgCtkQ4ArJkOAIBtAACBCQAAgh0AAI8/AIDvGAkAkz8AgJc/AICbPwCA4zwNAJ8/AIDhWAwAoz8AgIbQAACHvAMApz8AgKs/AICokQ4AqZkOAKrJDgCrxQ4ArN0OAK3BDgCuwQ4Ar/UOAIToAACvPwCAsz8AgLc/AIC7PwCAvz8AgMM/AIDHPwCAuMEPALnBDwC6wQ8Au8EPALzBDwC9wQ8AvsEPAL/1DwCwjQ4AsUUOALJNDgCzRQ4AtF0OALVBDgC2QQ4At0EOAKhRDgCpWQ4Aqo0OAKudDgCshQ4ArY0OAK6FDgCvvQ4Ayz8AgM8/AIDTPwCA1z8AgNs/AIDfPwCA4z8AgOc/AIC4kQ4AuZkOALqtDgC7RQEAvF0BAL1FAQC+RQEAv3UBALDFDgCxzQ4AssUOALPdDgC0xQ4AtbUOALa9DgC3tQ4AswUOAOs/AIDvPwCA8z8AgPc/AIC2DQ4AtQ0OAPs/AIC7CQ4AugEOAP8/AIADQACAv3EOAL4BDgC9CQ4AvBEOAIJtAACjQQ4AgFUAAIFlAACmSQ4AC0AAgA9AAIClSQ4AqkUOAKtNDgCGSAAAh3gAAK5FDgCvNQ4ArFUOAK1NDgCoXQIAqWECAKplAgCrdQIArG0CAK2xAgCusQIAr7ECAITsBAATQACAF0AAgBtAAIAfQACAI0AAgCdAAIArQACAuHEDALlxAwC6cQMAu3EDALzVAwC93QMAvtUDAL/NAwCw0QIAsdECALLRAgCz0QIAtFEDALVRAwC2UQMAt1EDAC9AAICz6QIAM0AAgL6ABAC2NQIAN0AAgDtAAIC14QIAuhECALsRAgA/QACAQ0AAgL6RAwC/kQMAvAECAL0BAgBHQACAS0AAgKOlAgBPQACApa0CAFNAAIBXQACApnkCAFtAAIBfQACAq10CAKpdAgCtTQIArE0CAK/dAwCu3QMAqNUCAKndAgCqLQEAqyUBAKw9AQCtJQEAri0BAK8lAQBjQACAZ0AAgGtAAIBvQACAc0AAgHtAAIB/QACAg0AAgLiFAQC5iQEAup0BALuVAQC8sQEAvbEBAL55AAC/eQAAsF0BALHlAQCy4QEAs/kBALTpAQC13QEAttUBALe9AQDh8A4Ah0AAgOMUDgCLQACAgb0AAIC9AACPQACAgq0AAIYABACH7AUAk0AAgJdAAICbQACAn0AAgO9gDgCjQACAp0AAgKtAAICFXH0Ar0AAgLNAAIDjZAEAt0AAgOG0AQC7QACA76AOAL9AAIC3PgCAhPgFAMNAAIDHQACAy0AAgLMlBgB3QACAz0AAgNNAAIDXQACAtiUGALU1BgDbQACAu6EGALoZBgDfQACA40AAgL+ZBgC+rQYAva0GALy1BgCCbQAA7zAEAIBVAACBZQAAvlwDAOdAAICG+AAAh2wDAOtAAIDvQACA80AAgPdAAID7QACA40QEAP9AAIDhjAcAo6UGAANBAIAHQQCAC0EAgA9BAICmpQYApbUGABNBAICrIQYAqpkGABdBAIAbQQCArxkGAK4tBgCtLQYArDUGAB9BAICz+QcAI0EAgCdBAIC2SQcAK0EAgC9BAIC1UQcAulEHALtRBwAzQQCAN0EAgL41BwC/OQcAvEUHAL09BwCoNQYAqT0GAKo1BgCriQYArJ0GAK2NBgCusQYAr7EGADtBAIA/QQCAQ0EAgEdBAICADQAAgbEAAIKxAABLQQCAuKEGALmtBgC6vQYAu7UGALytBgC9XQEAvlUBAL9NAQCw0QYAsdEGALLVBgCzrQYAtLUGALW5BgC2qQYAt6UGAKO9BgBPQQCAU0EAgISEAgC+kAEApg0GAKUVBgBbQQCAqxUGAKoVBgCGCAAAh3wBAK99BgCucQYArXkGAKwBBgBfQQCAs60BAGNBAIBnQQCAtqkBAGtBAIBvQQCAta0BALptAQC7dQEAc0EAgHdBAIC+XQEAvzUBALxlAQC9VQEAqGECAKlhAgCqYQIAq2ECAKxhAgCtbQIArp0CAK+VAgB7QQCAf0EAgINBAICHQQCAi0EAgI9BAICTQQCAl0EAgLiVAgC5nQIAuqECALuhAgC8cQMAvXEDAL5xAwC/cQMAsO0CALH1AgCy9QIAs8UCALTdAgC1tQIAtrECALexAgCbQQCAn0EAgKNBAICj5QIAp0EAgKXlAgCm4QIAq0EAgK9BAICzQQCAqiUCAKs9AgCsLQIArR0CAK4VAgCvfQIAt0EAgLtBAIC/QQCAhEB8AIAVAACBHQAAggUAAMNBAIC+7HwAy0EAgIZIfQCHCAMAz0EAgNNBAIDXQQCA20EAgKidAgCpxQIAqsECAKvBAgCsxQIArc0CAK7xAgCv8QIA30EAgONBAIDnQQCA60EAgMkAAADvQQCA80EAgPdBAIC4wQEAucEBALrBAQC73QEAvM0BAL31AQC+/QEAv50BALBBAQCxQQEAskEBALNBAQC0QQEAtUEBALZBAQC3QQEA4TgGAPtBAIDjaAYA/0EAgANCAIAHQgCAC0IAgISUfQC+rHwAD0IAgBNCAIAXQgCAvrh/ABtCAIDvEAEAH0IAgCNCAIAnQgCAK0IAgC9CAIDhkAEAM0IAgONEAAA7QgCAgS0AAIAtAADvgAAAgjkAAD9CAIBDQgCAB0AAgEdCAIDhsH8Ax0EAgOPUfABLQgCAN0IAgE9CAICGuAAAh9QCAFNCAIBXQgCAW0IAgF9CAIBjQgCAZ0IAgO8gfABrQgCAs4l9AG9CAIBzQgCAd0IAgHtCAIC2jX0AtY19AH9CAIC7RX4AukV+AINCAICHQgCAv0V+AL5FfgC9VX4AvFV+AKNJfQCLQgCAj0IAgJNCAICXQgCApk19AKVNfQCbQgCAq4V+AKqFfgCfQgCAo0IAgK+FfgCuhX4ArZV+AKyVfgCCbQAAszF+AIBVAACBZQAAtvF/AITcAwCnQgCAtSF+ALrNfwC70X8AhgAEAIfUAAC+dX8Av3l/ALzBfwC9wX8AqOV/AKn1fwCq/X8Aq/V/AKztfwCtNX4Arj1+AK81fgCrQgCAr0IAgLNCAIC3QgCAu0IAgL9CAIDDQgCAx0IAgLjZfgC54X4AuuF+ALvhfgC85X4Avel+AL6ZfgC/mX4AsE1+ALFRfgCyUX4As1F+ALT1fgC1+X4Atul+ALfpfgCjdX8Ay0IAgM9CAIDTQgCA10IAgKa1fgClZX8A20IAgKuVfgCqiX4A30IAgONCAICvPX4ArjF+AK2FfgCshX4A50IAgLMxfgDrQgCA70IAgLbFAQDzQgCA90IAgLXRAQC6yQEAu8kBAPtCAID/QgCAvs0BAL+xAQC8yQEAvckBAKjdfQCp9X0Aqv19AKvxfQCsHQIArQECAK45AgCvOQIAA0MAgAdDAIALQwCAD0MAgIIFAAATQwCAgBEAAIERAAC4EQIAuRkCALohAgC7IQIAvNUCAL3dAgC+1QIAv80CALBJAgCxSQIAslkCALNZAgC0TQIAtTECALYxAgC3MQIAvgADAKNxfQCEiAIAvoAEAKaFAgAbQwCAH0MAgKWRAgCqiQIAq4kCAIYoBACHDAMAro0CAK/xAgCsiQIArYkCACNDAICEyAMAhcwFALPlAwAnQwCAteUDALbtAwArQwCAL0MAgDNDAIC6bQMAu2UDALx9AwC9ZQMAvmUDAL9VAwA3QwCAO0MAgL8ABACjJQIAP0MAgKUlAgCmLQIAQ0MAgEdDAIBLQwCAqq0CAKulAgCsvQIAraUCAK6lAgCvlQIAT0MAgFNDAIBXQwCAW0MAgF9DAIDjzAMAY0MAgOGsAQBnQwCA7xwDAGtDAIBvQwCAc0MAgHdDAIB7QwCAf0MAgOFwfwBXQQCA4wR+AINDAICLQwCA4ZQBAI9DAIDjWAEAgNkAAIHZAACCJQAA7+R+AJNDAICXQwCA7+B+AJtDAICzAQEAn0MAgIboBwCHLAQAo0MAgLY1AQC1BQEAp0MAgLvxAAC64QAAq0MAgK9DAIC/sQAAvtEAAL3ZAAC84QAAF0MAgIdDAICzQwCAt0MAgKEBBACgEQQAoxkAAKLFBACotQYAqb0GAKrpBgCr/QYArO0GAK3VBgCu3QYArz0HALBFBwCxVQcAslUHALNtBwC0dQcAtRUHALYdBwC3FQcAuC0HALk1BwC6MQcAuw0HALwZBwC9GQcAvgkHAL8JBwCjQQYAu0MAgL9DAIDDQwCAx0MAgKZ1BgClRQYAy0MAgKuxBwCqoQcAj8ltAM9DAICv8QcArpEHAK2ZBwCsoQcAld11AJTBdACXzXAAli1zAJFdaACQVWgAk9l0AJJNaQCd5XgAnB17AJ9tBwCeuXgAmR1/AJhVcACboXwAmvl8AIJhbACDhWkA00MAgNdDAICGEXUAhxF1AISVaQCFjWgAij10AIvFcgDbQwCA30MAgI7dfgCPMX0AjD1xAI2dcQCSGX0Ak716AONDAIDvkAkAltUGAJdRBQCUXXkAlQl5AJpxBQCbvQUA50MAgOtDAIDvQwCA4agFAJx5AQDjuAgAoYUBAPNDAICjqQ0AogEMAKUBCACkOQ0Ap6kJAKa9CQCppRUAqAEUAKsBFACq/RUArbkRAKyxEQCvARwArqEQALH9HACw5R0As+kZALIBGAC1ASQAtH0ZAIQUAAC+FAAAgI0AAIGVAACCbQAA+0MAgIZQDwCHZAAA/0MAgANEAIC61QcAu90HALjBBwC5wQcAvjEEAL8xBAC88QcAvfEHALKtBwCztQcAsK0HALGlBwC2nQcAt/UHALSlBwC1lQcAqmkHAKtpBwCoaQcAqWkHAK5pBwCvaQcArGkHAK1pBwAHRACAC0QAgA9EAIATRACAF0QAgBtEAIAfRACAI0QAgKgRBQCpHQUAqjkFAKs5BQCsLQUArVEFAK5JBQCvQQUAJ0QAgCtEAIAvRACAM0QAgDdEAIA7RACAP0QAgENEAIC4XQIAuWkCALrBAwC7wQMAvPkDAL35AwC+kQMAv7UDALAJBQCxCQUAsuECALPhAgC0dQIAtX0CALZ1AgC3bQIAs7EEAIQAAgC+BA0AR0QAgEtEAIC20QQAtaUEAE9EAIC7zQQAus0EAFNEAIBXRACAv7kDAL6xAwC9NQMAvDUDAFtEAICj9QQAX0QAgGNEAICmlQQAa0QAgG9EAICl4QQAqokEAKuJBACHqA0AhswMAK71AwCv/QMArHEDAK1xAwDhUAYA4TQHAONAAADjWAcAgNEAAIHdAACC1QAAc0QAgHdEAIB7RACAf0QAgINEAICHRACAi0QAgO+cAADvyAcAj0QAgJNEAICzNQIAl0QAgLW1AQCbRACAn0QAgLa1AQC+7AwAo0QAgLuRAQC6mQEAvVEBALyJAQC/UQEAvlkBAKjtDQCp/Q0AqvUNAKttDgCsdQ4ArX0OAK51DgCvbQ4AZ0QAgKdEAICrRACAr0QAgLNEAIC3RACAu0QAgL9EAIC49Q4Auf0OALr1DgC7QQ8AvEEPAL1JDwC+cQ8Av3EPALAVDgCxHQ4AshUOALPNDgC01Q4Atd0OALbVDgC3zQ4Ao30NAMNEAIDHRACAy0QAgM9EAICm/Q4Apf0OANNEAICr2Q4AqtEOAISoAgDXRACArxkOAK4RDgCtGQ4ArMEOAIBNAACBVQAAglUAALNRDwDbRACAtXEPALZxDwDfRACAhuAAAIcEAwC6XQ8Auy0PALw1DwC9OQ8Avi0PAL8lDwCoVQ4AqV0OAKqVDgCrrQ4ArLUOAK29DgCutQ4Ar60OAONEAIDnRACA60QAgO9EAIDzRACA90QAgPtEAID/RACAuGkBALlpAQC6eQEAu3kBALxpAQC9aQEAvt0BAL/VAQCw1Q4AsaUOALKtDgCzoQ4AtKUOALWtDgC2nQ4At1kBAKMdDgADRQCAB0UAgPdDAIALRQCApj0OAKU9DgAPRQCAq2EOAKoRDgATRQCAF0UAgK9pDgCuYQ4ArXUOAKx5DgAbRQCAH0UAgCNFAIAnRQCAK0UAgC9FAIAzRQCAN0UAgIANAACBFQAAgh0AADtFAIA/RQCAQ0UAgIR4AQC+FAAA4xQPAEtFAIDh4A0AhAADAIawBACHFAMAT0UAgFNFAIBXRQCAW0UAgF9FAIBjRQCA78APAGdFAIBrRQCAb0UAgHNFAIB3RQCAe0UAgLNtAwB/RQCAtX0DALZ1AwCDRQCAh0UAgItFAIC6UQMAu1EDALz1AwC9/QMAvukDAL/hAwCPRQCAk0UAgJdFAICbRQCAn0UAgKNFAICnRQCAq0UAgKhxAgCpeQIAqokDAKuJAwCsmQMArZkDAK6JAwCviQMAsPkDALH5AwCyTQMAs0UDALRBAwC1SQMAtnEDALdxAwC4IQMAuSEDALohAwC7IQMAvCEDAL0hAwC+IQMAvyEDAICdAQCBEQAAghEAAIQEBQDvFAAAr0UAgLNFAIC+EAUA48gAALtFAIDh0AEAv0UAgMNFAIDHRQCAy0UAgM9FAICqeQIAq3kCAIboBACHYAUArsECAK/JAgCs3QIArdUCANNFAICjRQIA10UAgNtFAICmXQIA30UAgONFAIClVQIA50UAgOtFAIDvRQCA80UAgPdFAID7RQCA/0UAgO+EDgC+rAQA4dAOAANGAIDjFAEAB0YAgAtGAIAPRgCAE0YAgLPdAQAXRgCAG0YAgB9GAIAjRgCAtv0BALX9AQArRgCAu90BALrdAQCE4AQAL0YAgL+hAQC+vQEAvb0BALy9AQCoBQYAqR0GAKoVBgCrLQYArDUGAK09BgCuNQYArykGALdFAICC9QcAgeUHAIDlBwAnRgCAM0YAgIYcAACHsAMAuCUGALnFBgC6zQYAu8UGALzdBgC9xQYAvs0GAL/FBgCwWQYAsVkGALIpBgCzKQYAtDkGALUlBgC2JQYAtx0GAKOdBgA3RgCAO0YAgD9GAIBDRgCApr0GAKW9BgBHRgCAq50GAKqdBgBLRgCAT0YAgK/hBgCu/QYArf0GAKz9BgBTRgCAs/UHAFdGAIBbRgCAtu0HAF9GAIBjRgCAteUHALqNBwC7kQcAZ0YAgGtGAIC+dQcAv30HALyBBwC9fQcAqCUGAKkpBgCqOQYAqzkGAKwpBgCtKQYArnkGAK91BgBvRgCAc0YAgHdGAIB7RgCAf0YAgINGAICHRgCAi0YAgLjVBgC53QYAuuEGALv9BgC85QYAve0GAL7lBgC/mQYAsA0GALERBgCyEQYAs+0GALT1BgC1/QYAtvUGALftBgCjsQYAgi0AAIEVAACAsQAAR0UAgKapBgCloQYAj0YAgKvVBgCqyQYAk0YAgL5oAQCvOQYArjEGAK05BgCsxQYAm0YAgLPxAQCGaAAAh3wBALZdAQCfRgCAo0YAgLVVAQC6SQEAu0kBAKdGAICrRgCAvj0BAL8hAQC8OQEAvTUBAK9GAICzRgCAhAQDAL6AHAC3RgCA4RwGALtGAIDjAAYAvwguAL9GAIDDRgCA78gHAMdGAIDLRgCAz0YAgNNGAIDXRgCA20YAgKN9AgDfRgCApdkCAONGAIDnRgCAptECAOtGAIDvRgCAq8UCAKrFAgCtuQIArLUCAK+tAgCusQIAqW0FAKhZBQCrDQIAqrkCAK0dAgCsHQIArwUCAK4NAgC+aB0A80YAgPdGAID7RgCAgB0AAIEJAACCmQEA/0YAgLnhAwC4KQIAu+EDALrpAwC94QMAvPkDAL/hAwC+6QMAsU0CALBNAgCzIQIAsi0CALUlAgC0OQIAtxECALYlAgCowQIAqdECAKrRAgCr5QIArP0CAK0VAQCuHQEArw0BAANHAIALRwCAD0cAgBNHAIAXRwCAG0cAgB9HAIAjRwCAuAUBALkJAQC6HQEAuxUBALwxAQC9MQEAvv0BAL/1AQCweQEAsUEBALJBAQCzXQEAtEUBALVNAQC2RQEAtz0BAIagHQCHxB0AJ0cAgO/YAAArRwCAL0cAgDNHAIDvxAYAhGwcAOH0BgA3RwCA47AGADtHAIDhlAEAP0cAgONEBgCzGQIAQ0cAgEdHAIBLRwCAhewsALbVAQC1NQIAT0cAgLvFAQC6/QEAU0cAgFdHAIC/yQEAvsEBAL3JAQC81QEAo9kdAAdHAIBbRwCAX0cAgGNHAICmFR4ApfUdAGdHAICrBR4Aqj0eAGtHAIBvRwCArwkeAK4BHgCtCR4ArBUeAIBpAACBaQAAggUAAHNHAIB3RwCAe0cAgIcQAwCGfAMAf0cAgINHAICHRwCAi0cAgI9HAICTRwCAl0cAgJtHAICopR8Aqa0fAKqlHwCrvR8ArKUfAK2tHwCupR8ArxUfAJ9HAICjRwCAp0cAgKtHAICvRwCAs0cAgLdHAIC7RwCAuA0fALkZHwC6IR8AuyEfALzZAAC92QAAvskAAL/BAACwcR8AsXEfALJxHwCzRR8AtEEfALVNHwC2PR8AtzUfALMtHgC/RwCAw0cAgMdHAIDLRwCAti0eALUtHgDPRwCAu7UeALq1HgDTRwCA10cAgL+JHgC+hR4AvZEeALylHgCCKQAAo2keAIAdAACBFQAApmkeANtHAIDfRwCApWkeAKrxHgCr8R4A40cAgITgAQCuwR4Ar80eAKzhHgCt1R4AqNUBAKnlAQCq7QEAq+UBAKz9AQCt5QEAru0BAK/lAQC+oAEAl0YAgOdHAIDrRwCAhhAAAId0AQDvRwCA80cAgLh9AQC5wQAAusEAALvBAAC8wQAAvckAAL7xAAC/8QAAsJ0BALFFAQCyTQEAs0UBALRdAQC1RQEAtk0BALdFAQD3RwCA+0cAgP9HAIADSACAB0gAgO80AgDv7B4AC0gAgOHwHQDj4AIA4zAeAOGEAQAPSACAE0gAgBdIAIAbSACAsyUCAJQAAAAfSACAI0gAgCdIAIC2JQIAtTUCACtIAIC7wQIAuhkCAC9IAIAzSACAv8ECAL7ZAgC90QIAvNkCADdIAIA7SACAP0gAgKPpAgBDSACApfkCAKbpAgBHSACAS0gAgE9IAICq1QIAqw0CAKwVAgCtHQIArhUCAK8NAgCAYQAAgWEAAIIFAABTSACAW0gAgIQABAC+FAQAX0gAgIbABACHUAMAY0gAgGdIAIBrSACAb0gAgHNIAIB3SACAqK0CAKm9AgCqtQIAqw0BAKwVAQCtHQEArhUBAK8NAQCE7AQAe0gAgH9IAICDSACAh0gAgItIAICPSACAk0gAgLgdAQC5LQEAuiUBALvNAQC81QEAvd0BAL7JAQC/wQEAsH0BALFVAQCyXQEAs1UBALRNAQC1PQEAtjUBALctAQDhGB4Al0gAgOM4HgCbSACAn0gAgKNIAICnSACAq0gAgK9IAICzSACAvmAEALdIAICBdQAAgHUAAO/gHwCCbQAAu0gAgL9IAICG6AQAh3wFAMNIAIDhkAEAy0gAgOOgAADPSACA00gAgNdIAIDvtAAA20gAgN9IAIDjSACA50gAgLUFBgBXSACAx0gAgLYFBgDrSACA70gAgLOlBQDzSACAvRkGALwRBgC/YQYAvhEGAPdIAID7SACAuwkGALohBgCj/QUA/0gAgANJAIAHSQCAC0kAgKZdBgClXQYAD0kAgKtRBgCqeQYAE0kAgBdJAICvOQYArkkGAK1BBgCsSQYAqFEGAKlZBgCqYQYAq2EGAKxhBgCtYQYArmEGAK9hBgAbSQCAH0kAgCNJAIAnSQCAgA0AAIGxAQCCsQEAK0kAgLhNBwC5VQcAul0HALtVBwC8TQcAvXUHAL59BwC/cQcAsMUHALHNBwCyxQcAs90HALTFBwC1zQcAtsUHALd5BwCz6QcAL0kAgDNJAICEwAEAvtgBALbhBwC16QcAN0kAgLsJBgC6AQYAhogAAIesAQC/CQYAvgEGAL0JBgC8EQYAO0kAgKOtBwA/SQCAQ0kAgKalBwBHSQCAS0kAgKWtBwCqRQYAq00GAE9JAIBTSQCArkUGAK9NBgCsVQYArU0GAKhZBgCpZQYAqm0GAKtlBgCsYQYArWEGAK5hBgCvYQYAhKwBAFdJAIBbSQCAX0kAgGNJAIBnSQCAa0kAgG9JAIC4kQEAuZkBALqhAQC7oQEAvHEBAL1xAQC+cQEAv3EBALDxAQCx8QEAsvUBALPdAQC0xQEAtbEBALaxAQC3sQEAs+UFAHNJAIB3SQCAe0kAgH9JAIC24QUAtekFAINJAIC7NQIAujUCAIdJAICLSQCAv3UCAL4BAgC9CQIAvCECAI9JAICjoQUAk0kAgJdJAICmpQUAm0kAgJ9JAIClrQUAqnECAKtxAgCjSQCAvigDAK5FAgCvMQIArGUCAK1NAgCA1QAAgd0AAILhAACrSQCA4yABAK9JAIDhqAEAs0kAgO80AgC3SQCAhggMAIdoAwCsAAAAu0kAgL9JAIDDSQCAs40DAMdJAIDLSQCAhIAMAM9JAIC2vQMAtYEDANNJAIC7TQMAuk0DANdJAIDbSQCAv00DAL5NAwC9TQMAvE0DAKhBAgCpTQIAqkUCAKtZAgCsSQIArX0CAK51AgCvuQIAvmgNAN9JAIDjSQCA50kAgIRsDADrSQCA70kAgPNJAIC4TQEAuVUBALpVAQC7ZQEAvH0BAL0VAQC+EQEAvxEBALDJAgCxyQIAstkCALPZAgC0yQIAtckCALZ9AQC3dQEA4XgHAOOYAADjuAYA4VwGAPdJAID7SQCA/0kAgANKAIAHSgCAC0oAgA9KAIATSgCA7AAAAO9cAADv6AYAG0oAgIFpAACAYQAAo4UCAIJhAACliQIAH0oAgCNKAICmtQIAhkAMAIfEDACrRQIAqkUCAK1FAgCsRQIAr0UCAK5FAgCojQ4AqZEOAKqVDgCrqQ4ArKUOAK2tDgCupQ4Ar9kOABdKAIAnSgCAK0oAgC9KAIAzSgCAN0oAgDtKAIA/SgCAuHUPALl9DwC6dQ8Au90PALzFDwC9zQ8AvsUPAL/9DwCwqQ4AsbUOALK1DgCzhQ4AtJ0OALVRDwC2UQ8At1EPALMdDgBDSgCAR0oAgEtKAIBPSgCAti0OALUtDgBTSgCAu3EOALptDgBXSgCAW0oAgL+VDwC+WQ4AvVEOALxhDgBfSgCAo1kOAGNKAIBnSgCApmkOAGtKAIBvSgCApWkOAKopDgCrNQ4Ac0oAgHdKAICuHQ4Ar9EPAKwlDgCtFQ4AqL0OAKnRDgCq0Q4AqykBAKw5AQCtOQEArikBAK8pAQCADQAAgRUAAIIdAAB7SgCAf0oAgINKAIC+dAIAh0oAgLjtAQC5hQEAuoEBALuBAQC8hQEAvY0BAL6xAQC/sQEAsFkBALFZAQCy7QEAs+UBALT9AQC15QEAtuUBALfVAQCLSgCAtqkBALWhAQCPSgCAs0kOAJNKAICGOAAAh9wBAL8xAQC+KQEAvSEBALwpAQC7jQEAuo0BAKdJAICXSgCAoxkOAJtKAICfSgCAo0oAgKdKAICm+QEApfEBAKtKAICr3QEAqt0BAK9KAICzSgCAr2EBAK55AQCtcQEArHkBALdKAIDv3A8Au0oAgL9KAIDDSgCAx0oAgMtKAIDPSgCA00oAgNdKAIDbSgCA30oAgONKAIDj6A4A50oAgOGMDgCAEQAAgREAAIIRAACEQAIA60oAgO9KAIDzSgCAvhADAIbABACHRAMA+0oAgP9KAIADSwCAB0sAgAtLAIAPSwCA7yQCABNLAIAXSwCAG0sAgB9LAIAjSwCAJ0sAgCtLAICE7AQAL0sAgDNLAIA3SwCA4+wCADtLAIDhOAEAP0sAgLNVAwBDSwCAR0sAgEtLAIBPSwCAth0DALUdAwBTSwCAuwkDALo5AwBXSwCAW0sAgL/9AAC+/QAAvfkAALwRAwCogQIAqYkCAKqdAgCrsQIArNUCAK3dAgCu1QIAr80CAIDNAQCBCQAAghkAAF9LAIBjSwCAa0sAgL5wBQBvSwCAuFkBALlZAQC6aQEAu2kBALx5AQC9eQEAvmkBAL9lAQCwvQIAsY0CALKFAgCzbQEAtHkBALV5AQC2aQEAt2kBAIYgBACHCAUAc0sAgHdLAIB7SwCAf0sAgINLAIDvXAAAhOwEAOFcDgCHSwCA44wOAItLAICPSwCAk0sAgJdLAICjVQIAm0sAgJ9LAICjSwCAp0sAgKYdAgClHQIAq0sAgKsJAgCqOQIAr0sAgLNLAICv/QEArv0BAK35AQCsEQIAqGkGAKlpBgCqeQYAq3kGAKxpBgCtaQYArp0GAK+VBgBnSwCAt0sAgLtLAIC/SwCAw0sAgMdLAIDLSwCAz0sAgLj1BgC5+QYAuo0GALuFBgC8nQYAvYUGAL6FBgC/tQYAsO0GALH1BgCy/QYAs/UGALTtBgC10QYAttEGALfRBgCz8QYAghUAAIG1AACAtQAA00sAgLbpBgC14QYAvtQDALsxBgC6KQYA10sAgNtLAIC/FQYAvikGAL0hBgC8KQYA30sAgKO1BgCGyAAAh8gAAKatBgDjSwCA50sAgKWlBgCqbQYAq3UGAOtLAIDvSwCArm0GAK9RBgCsbQYArWUGAKg1BgCpOQYAqoEGAKuBBgCsgQYArYEGAK6BBgCvtQYA80sAgPdLAID7SwCA/0sAgANMAIAHTACAC0wAgA9MAIC4nQYAua0GALqlBgC7aQEAvHkBAL15AQC+aQEAv2kBALDRBgCx0QYAstEGALPRBgC0tQYAtb0GALa1BgC3rQYAswkGABNMAIAXTACAG0wAgB9MAIC2AQYAtQkGACNMAIC7FQYAuhUGACdMAIArTACAv3kGAL5xBgC9BQYAvAUGAC9MAICjTQYAM0wAgPdKAICmRQYAN0wAgDtMAIClTQYAqlEGAKtRBgA/TACAQ0wAgK41BgCvPQYArEEGAK1BBgCB6QMAgN0DAISIAwCC4QMAhrA8AIeIAgC+VAMAS0wAgE9MAIBTTACAV0wAgFtMAIBfTACAY0wAgGdMAIBrTACA4/AGAG9MAIDhMAYAhAA8AHNMAIB3TACAe0wAgH9MAICDTACAhTQ9AIdMAICLTACA77AHAI9MAICTTACAl0wAgJtMAICfTACAo0wAgL7EPACnTACAgp0BAIGdAQCAnQEAqA0CAKllAgCqfQIAq3UCAKxZAgCtWQIArpkDAK+ZAwCw6QMAsekDALL5AwCz+QMAtOkDALXpAwC2XQMAt1UDALhtAwC5dQMAunUDALtFAwC8XQMAvTUDAL4xAwC/KQMAq0wAgK9MAICzTACAu0wAgOFgAwDv9AMA40QCAL9MAIDDTACA4zwDAO/0NwDh/AEAx0wAgMtMAIDPTACA00wAgIZkPwCHaD0AhTQhALOZAwDXTACAtb0DALa1AwDbTACA30wAgONMAIC6QQIAu0ECALxBAgC9QQIAvkECAL9BAgDnTACA60wAgO9MAIDzTACA90wAgPtMAID/TACA7/gBAIRoPADhPAYAA00AgOMcBgAHTQCAC00AgA9NAIATTQCAoxUDABdNAIAbTQCAH00AgCNNAICmOQMApTEDACtNAICrzQIAqs0CAL5kPgAvTQCAr80CAK7NAgCtzQIArM0CAKgdPgCpJT4Aqi0+AKslPgCsPT4ArSU+AK4tPgCvJT4At0wAgIL1PwCB5T8AgOU/ACdNAIAzTQCAhgAEAIecAwC4LT4AuTE+ALoxPgC7MT4AvNE+AL3RPgC+0T4Av80+ALBdPgCxIT4Asjk+ALM5PgC0KT4AtSk+ALYZPgC3FT4As6U+ADdNAIA7TQCAP00AgENNAIC2pT4AtbU+AEdNAIC75T4Aupk+AEtNAIBPTQCAv+0+AL7tPgC97T4AvO0+AFNNAICj4T4AV00AgFtNAICm4T4AX00AgGNNAICl8T4Aqt0+AKuhPgBnTQCAa00AgK6pPgCvqT4ArKk+AK2pPgCPBSUAsyU+AG9NAIBzTQCAtik+AHdNAIB7TQCAtSk+ALp9PgC7RT4Af00AgINNAIC+tT4Av70+ALxdPgC9vT4An304AJ5lOQCd8TgAnFE0AJtZNQCaUTUAmfEwAJgNMQCXZTEAlsEwAJVZLQCUTS0Ak+EsAJLZKQCRWSkAkPEoALSlGQC13RgAh00AgIQIAACwkRUAsQEVALIBGACzvRkAgA0AAIGtAwCCpQMAi00AgKNhAACiHT0AoZk9AKBxPACkxQUApUEEAKYBCACn4QkAR0wAgKH1AQCi6QEAo90FAKwBEACtxREArtkRAK85EACoZQgAqQEMAKrZDQCrCQ0AijEuAIuhMwCPTQCAk00AgI65MwCPETYAjB0yAI1NMgCCJSYAg6krAL5kAwCEYAQAhqEvAIcVLgCEGSoAhZEqAJphPgCb7T4AhsgEAIfcAwCbTQCA4Vw+AJyJAwDjAD4Akmk2AJN5NwCfTQCA7xg+AJZNOwCXuT8AlME7AJVdOgCpnT0AqIk9AKu5PQCqrT0Arak9AKyhPQCvyT0ArqE9AL7oBACjTQCAp00AgKtNAICvTQCAs00AgLdNAIC7TQCAuVk9ALhRPQC7eT0AumU9AL1pPQC8YT0Avx09AL5hPQCxgT0AsLk9ALNpPQCyiT0AtXk9ALRxPQC3aT0AtnE9AKMhPAC/TQCAw00AgMdNAIDLTQCApi08AKUtPADPTQCAq0E8AKp5PADTTQCA100AgK+5PACusTwArbk8AKxZPADbTQCA300AgLN9AwDjTQCAtdkDAOdNAIDrTQCAttEDAO9NAIDzTQCAu8UDALrFAwC9uQMAvLUDAL+tAwC+sQMA900AgPtNAID/TQCA71wDAIAVAACBHQAAgjEAAO+MPgCE7AQA4fw+AANOAIDjHD4AC04AgOGUAQAPTgCA4yAAAKP1AwATTgCAh+gEAIZsBAAXTgCAplkDAKVRAwAbTgCAq00DAKpNAwAfTgCAI04AgK8lAwCuOQMArTEDAKw9AwCXTQCAB04AgCdOAIArTgCAL04AgDNOAIA3TgCAO04AgKhxBgCpTQYAqo0GAKuFBgCsnQYArYUGAK6NBgCvhQYAsP0GALFBBwCyQQcAs0EHALRBBwC1SQcAtnEHALdxBwC4IQcAuSEHALolBwC7OQcAvCkHAL0VBwC+HQcAv/0HALMlBgA/TgCAQ04AgEdOAIBLTgCAtiUGALU1BgBPTgCAu6UHALoZBgBTTgCAV04AgL+tBwC+pQcAvbUHALy1BwBbTgCAo2EGAF9OAIBjTgCApmEGAGdOAIBrTgCApXEGAKpdBgCr4QcAb04AgHNOAICu4QcAr+kHAKzxBwCt8QcAqLEGAKm9BgCqzQYAq90GAKzNBgCt/QYArvUGAK8VAQCA+QEAgc0BAILFAQC+ZAIAhpAAAIcAAQB7TgCAf04AgLjRAQC52QEAuuEBALvhAQC8kQEAvZ0BAL6VAQC/iQEAsG0BALF1AQCyfQEAs3UBALRtAQC18QEAtvEBALfxAQCzRQYAd04AgINOAICHTgCAi04AgLZ9BgC1RQYAj04AgLuxAQC6qQEAk04AgJdOAIC/NQEAvqkBAL2hAQC8qQEAm04AgKMBBgCfTgCAo04AgKY5BgCnTgCAq04AgKUBBgCq7QEAq/UBAK9OAICzTgCAru0BAK9xAQCs7QEAreUBAOEoAQC3TgCA41ACALtOAIC/TgCAw04AgMdOAIDLTgCAz04AgNNOAIDXTgCA204AgIFxAACAGQAA75wCAIJ5AADfTgCA404AgITIAgCzxQMA604AgLXFAwC2xQMAvhADAIbADACHRAwAuqkDALulAwC8vQMAvaEDAL6hAwC/lQMArhEGAK8ZBgCsAQYArQEGAKqlBgCrEQYAqEU5AKlxOQDvTgCA804AgPdOAID7TgCA/04AgANPAIAHTwCAC08AgL7tBwC/TQcAvNEHAL3lBwC63QcAu8EHALg1BgC51QcAtjkGALcNBgC0JQYAtTkGALIxBgCzPQYAsFEGALFRBgCoOQIAqTkCAKqBAgCrgQIArIECAK2JAgCusQIAr7ECAIRsDQAPTwCAvmANABNPAIAXTwCAG08AgB9PAIAjTwCAuE0BALlVAQC6XQEAu1UBALxNAQC9dQEAvn0BAL91AQCwoQIAsa0CALKlAgCzuQIAtKkCALWdAgC2lQIAt3kBAOFUBgDh1AcA4zgGAOOwBwAnTwCAK08AgC9PAIAzTwCAhOQMADdPAIA7TwCAP08AgENPAIBHTwCA72wAAO/kBwCjSQIAS08AgE9PAIBTTwCAW08AgKZJAgClSQIAX08AgKspAgCqJQIAhkgMAIfcDACvGQIAri0CAK0tAgCsMQIAqFEOAKmlDgCqrQ4Aq6UOAKy9DgCtpQ4Arq0OAK+lDgCA5Q8Age0PAILlDwBXTwCAY08AgGdPAIBrTwCAb08AgLjVDwC53Q8AutUPALvpDwC8+Q8AvfkPAL7pDwC/6Q8AsN0OALFBDwCyRQ8As10PALRFDwC1TQ8AtkUPALftDwCzJQ4Ac08AgHdPAIB7TwCAf08AgLYlDgC1NQ4Ag08AgLuFDwC6GQ4Ah08AgItPAIC/iQ8AvoEPAL2JDwC8kQ8Aj08AgKNhDgCTTwCAl08AgKZhDgCbTwCAn08AgKVxDgCqXQ4Aq8EPAKNPAICnTwCArsUPAK/NDwCs1Q8Arc0PAKjRDgCp2Q4AqjkBAKs5AQCsKQEArSkBAK6dAQCvlQEAq08AgK9PAICzTwCAt08AgIANAACBtQAAgr0AALtPAIC4lQEAuZ0BALqhAQC7oQEAvHEAAL1xAAC+cQAAv3EAALDtAQCx9QEAsvUBALPFAQC03QEAtbUBALaxAQC3sQEAv08AgMNPAICzuQEAvsACALWpAQDHTwCAy08AgLahAQCGgAEAh8QBALs5AQC6IQEAvRkBALwpAQC/eQEAvhEBAKPxAQDPTwCA504AgNNPAIDXTwCApukBAKXhAQDbTwCAq3EBAKppAQDfTwCA408AgK8xAQCuWQEArVEBAKxhAQDnTwCA608AgO9PAIDzTwCA4agBAPdPAIDjQAIA+08AgL8oFQD/TwCA73QCAANQAIAHUACAC1AAgA9QAIATUACAF1AAgON0DwCEiAMA4TQOABtQAIAfUACAI1AAgCdQAICADQAAgRUAAIIRAAArUACAL1AAgO+kDwAzUACAO1AAgKgZAwCpQQMAqkUDAKtdAwCsTQMArX0DAK51AwCvnQAAhaQVAL58AwCGCAQAhxwDAD9QAIBDUACAR1AAgEtQAIC49QAAuf0AALr1AAC7jQAAvIEAAL2BAAC+gQAAv4EAALDlAACx7QAAsuUAALP5AAC07QAAtdEAALbVAAC3zQAAT1AAgFNQAIBXUACAs8ECAFtQAIC1yQIAtvECAF9QAIBjUACAZ1AAgLotAQC7JQEAvD0BAL0hAQC+JQEAvxkBAKapAgCESAIAa1AAgKWRAgBvUACAo5kCAHNQAIB3UACArn0BAK9BAQCsZQEArXkBAKp1AQCrfQEAe1AAgH9QAICDUACAh1AAgItQAICPUACA7+QAAJNQAICXUACAm1AAgOMQDgCfUACA4VgOAKNQAICALQAAgREAAIIVAAC+sAUAs3UBAKtQAICHFAUAhmwEAK9QAIC21QAAtWUBALNQAIC7/QAAuvUAALdQAIC7UACAv6EAAL69AAC93QAAvN0AAKh9BgCptQYAqr0GAKu1BgCsrQYArRUHAK4dBwCvFQcAp1AAgL9QAIDDUACAx1AAgMtQAIDPUACA01AAgNdQAIC4OQcAuTkHALrJBwC7yQcAvNkHAL3ZBwC+zQcAv8UHALBxBwCxeQcAskkHALNJBwC0OQcAtSUHALYhBwC3IQcAozUGANtQAIDfUACA41AAgOdQAICmlQcApSUGAOtQAICrvQcAqrUHAO9QAIDzUACAr+EHAK79BwCtnQcArJ0HAPdQAID7UACA/1AAgANRAIAHUQCAgj0AAIE9AACAPQAAC1EAgA9RAIATUQCAhKADAL6kAwAXUQCAhvgAAIfgAACoxQYAqdUGAKrVBgCr5QYArP0GAK0xAQCuMQEArzEBABtRAIAfUQCAI1EAgCdRAIArUQCAL1EAgDNRAIA3UQCAuN0BALntAQC65QEAu40BALyVAQC9nQEAvpUBAL+NAQCwUQEAsVEBALJRAQCzUQEAtPUBALX9AQC29QEAt+0BALNdBgA7UQCAP1EAgENRAIBHUQCAtrEBALV1BgBLUQCAu5UBALqVAQBPUQCAU1EAgL85AQC+MQEAvYUBALyFAQClLQYAV1EAgFtRAICm6QEAX1EAgGNRAICjBQYAZ1EAgK3dAQCs3QEAr2EBAK5pAQBrUQCAN1AAgKvNAQCqzQEAb1EAgHNRAICExAMAvwD0AHdRAICCPQAAgT0AAIA9AAB7UQCAf1EAgINRAIC+YAMAi1EAgI9RAICTUQCAl1EAgIbgHACHAAMA7wwHAJtRAICfUQCAo1EAgKdRAICrUQCAr1EAgLNRAIC3UQCAu1EAgOHABgC/UQCA4ywHAMNRAIDHUQCAy1EAgM9RAIDTUQCA11EAgNtRAIDfUQCA41EAgKiBAwCpgQMAqoEDAKuBAwCsgQMArYEDAK6BAwCvgQMAsEUDALFNAwCyRQMAs10DALRNAwC1fQMAtnUDALcZAwC4KQMAuTUDALo9AwC7MQMAvAEDAL31AAC+/QAAv+0AALMpAgDnUQCA61EAgO9RAIDzUQCAtiECALUpAgCEUB0Au6kCALqhAgD7UQCA/1EAgL+ZAgC+qQIAvakCALyxAgCBTQAAgE0AAO+cAwCCXQAAhvAcAId4HQC+EB0AA1IAgAdSAIALUgCAD1IAgBNSAIDhkAEAF1IAgONgAwAbUgCAH1IAgCNSAIAnUgCAK1IAgC9SAIAzUgCAN1IAgO+UAQCE7BwA4XAGADtSAIDjUAEAP1IAgENSAIBHUgCAS1IAgKPpAgBPUgCAU1IAgFdSAIBbUgCApuECAKXpAgBfUgCAq2kCAKphAgBjUgCAvqgcAK9ZAgCuaQIArWkCAKxxAgCoMR4AqTEeAKoxHgCrMR4ArF0eAK1FHgCuTR4Ar0UeAPdRAICCzR8AgfUfAID9HwBnUgCAa1IAgIYcAACH+AMAuMUeALnNHgC6xR4Au90eALzFHgC9zR4AvsUeAL9ZHwCwPR4AsQUeALINHgCzBR4AtB0eALUBHgC2BR4At/0eALO5HgBvUgCAc1IAgHdSAIB7UgCAtsUeALXVHgB/UgCAu8EeALr5HgCDUgCAh1IAgL/FHgC+2R4AvdEeALzZHgCLUgCAo/0eAI9SAICTUgCApoEeAJdSAICbUgCApZEeAKq9HgCrhR4An1IAgKNSAICunR4Ar4EeAKydHgCtlR4AqCkeAKkpHgCqVR4Aq20eAKx1HgCtfR4ArnUeAK9pHgCnUgCAq1IAgK9SAICzUgCAt1IAgLtSAIC/UgCAw1IAgLjpHgC59R4Auv0eALv1HgC87R4AvZEeAL6RHgC/kR4AsB0eALHlHgCy7R4As+UeALT9HgC15R4Atu0eALflHgCz3R4Ax1IAgMtSAIDPUgCA01IAgLb9HgC1/R4AhFgBALshHgC62R4AvigAANtSAIC/IR4AvjkeAL0xHgC8OR4AgU0AAIBNAACjlR4Agl0AAKW1HgDXUgCA31IAgKa1HgCHUQCA41IAgKtpHgCqkR4ArXkeAKxxHgCvaR4ArnEeAIYABACHRAMAs4ECAOdSAIC1gQIA61IAgO9SAIC2gQIAiAAAAPNSAIC74QIAuu0CAL3lAgC8+QIAv9ECAL7lAgD3UgCA+1IAgIREAwC+jAMA4UgCAP9SAIDjAAIA7/wfAANTAIDhPB4A79wCAONgHwAHUwCAC1MAgA9TAIATUwCAqQUCAKixAgCrBQIAqgUCAK0NAgCsBQIArzUCAK41AgCEbAUAF1MAgBtTAIAfUwCAI1MAgCdTAIArUwCAL1MAgLnpAwC44QMAu/kDALrhAwC96QMAvOEDAL9dAwC+4QMAsSkCALAlAgCzPQIAsiECALUZAgC0LQIAt9kDALYRAgAzUwCAN1MAgDtTAICjhQMAP1MAgKWFAwCmhQMAQ1MAgEtTAIBPUwCAqukDAKvlAwCs/QMAreEDAK7hAwCv1QMAgEkAAIFVAACCVQAAo6kCAL6YBAClQQEApkEBAFNTAICG4AUAh+AFAKotAQCrOQEArBEBAK0FAQCuDQEArwUBAFdTAIBbUwCAX1MAgO/cAABjUwCAZ1MAgGtTAIDviB4AhCwHAOHsHgBvUwCA4xweAHNTAIDhlAEAd1MAgOMwAACzJQIAhWDmAHtTAIB/UwCAg1MAgLbNAQC1zQEAh1MAgLu1AQC6oQEAi1MAgI9TAIC/iQEAvoEBAL2JAQC8nQEAR1MAgJNTAICXUwCAm1MAgJ9TAICjUwCAp1MAgKtTAICoAQcAqQEHAKp1BwCrrQcArLUHAK29BwCuqQcAr6kHALDZBwCx7QcAsvkHALP1BwC0mQcAtZkHALaJBwC3gQcAuIkHALmJBwC6bQAAu2UAALx9AAC9ZQAAvm0AAL9lAACBCQAAgJkAAK9TAICCHQAAs1MAgLdTAIC7UwCAv1MAgKgNBQCpfQUAqk0FAKuhBgCspQYAra0GAK6dBgCv/QYAsIUGALGRBgCyqQYAs70GALSlBgC1rQYAtqUGALd5BgC4SQYAuUkGALpZBgC7WQYAvEkGAL1JBgC++QcAv/kHALNdBgDDUwCAhigCAIcsAQDHUwCAtp0GALWdBgDLUwCAu4kGALq9BgDPUwCA01MAgL/9BgC+/QYAvYEGALyNBgDXUwCAoxkGANtTAIDfUwCAptkGAONTAIDnUwCApdkGAKr5BgCrzQYA61MAgO9TAICuuQYAr7kGAKzJBgCtxQYAqBkBAKkZAQCqjQAAq50AAKyNAACtvQAArrUAAK/dAADzUwCA91MAgPtTAID/UwCAA1QAgAdUAIALVACAD1QAgLhpAAC5aQAAunkAALt5AAC8aQAAvWkAAL7dAwC/1QMAsKkAALGpAACyvQAAs7UAALSZAAC1mQAAtlkAALdZAAC+LAIAE1QAgBdUAIAbVACAH1QAgCNUAIArVACAL1QAgIAtAACBNQAAgj0AADNUAICGkAwAh+gCADdUAIA7VACAs0UDAD9UAIBDVACAR1QAgEtUAIC2fQMAtUUDAE9UAIC7LQMAui0DAFNUAIBXVACAvx0DAL4dAwC9IQMAvCkDAKvNAwCqzQMAW1QAgF9UAICv/QMArv0DAK3BAwCsyQMAo6UDAGNUAIBnVACAa1QAgG9UAICmnQMApaUDAHNUAIB3VACAe1QAgH9UAICDVACAh1QAgII9AACBPQAAgD0AAItUAICPVACAk1QAgIRgAwCG0AwAhzADAJtUAICfVACAvkQCAKNUAICnVACAq1QAgOEAAACvVACA46gGALNUAICE7AwAt1QAgO/QAwC7VACAv1QAgMNUAIDHVACAy1QAgLNtAQDPVACA01QAgNdUAIDbVACAthEBALVlAQDfVACAuz0BALo1AQDjVACA51QAgL/9AQC+/QEAvRUBALwVAQDrVACA4fwGAO9UAIDjPAcA81QAgPdUAID7VACA/1QAgANVAIC+bAwAC1UAgA9VAIATVQCAF1UAgBtVAIDvFAYAgV0AAIBdAACj5QEAgm0AAKXtAQAfVQCAI1UAgKaZAQCHqAwAhuQMAKu1AQCqvQEArZ0BAKydAQCvdQEArnUBAKgZDgCpGQ4AqiUOAKs1DgCsLQ4ArVEOAK5RDgCvUQ4Al1QAgAdVAIAnVQCAK1UAgC9VAIAzVQCAN1UAgDtVAIC47Q4AufUOALr1DgC7jQ4AvJUOAL2dDgC+lQ4Av40OALAxDgCxOQ4AsgEOALMBDgC0+Q4AtfkOALbdDgC31Q4AqHkOAKl5DgCqjQ8Aq4UPAKydDwCtgQ8AroUPAK+5DwA/VQCAQ1UAgEdVAIBLVQCAT1UAgFNVAIBXVQCAW1UAgLiRDwC5mQ8AuqEPALuhDwC8UQ8AvV0PAL5JDwC/SQ8AsM0PALHVDwCy3Q8As9UPALTNDwC1sQ8AtrEPALexDwCzBQ4AX1UAgGNVAIBnVQCAa1UAgLYBDgC1FQ4Ab1UAgLsRDgC6CQ4Ac1UAgISgAQC/dQ4AvgkOAL0BDgC8CQ4AgmkAAKNBDgCAWQAAgVEAAKZFDgC+WAEAd1UAgKVRDgCqTQ4Aq1UOAIbIAACHrAEArk0OAK8xDgCsTQ4ArUUOAHtVAIB/VQCAg1UAgIdVAICLVQCAj1UAgCdUAICTVQCAqAkOAKkJDgCqGQ4AqxkOAKwJDgCtYQ4ArmEOAK+VAQCw7QEAsfUBALL9AQCz9QEAtO0BALV1AQC2fQEAt3UBALhNAQC5VQEAul0BALtVAQC8TQEAvfEAAL7xAAC/8QAAl1UAgJtVAICfVQCAo1UAgKdVAIDj6A4Aq1UAgOE0DgC+AAQA79wPAK9VAICzVQCAt1UAgLtVAIC/VQCAw1UAgLPxDQDHVQCAy1UAgM9VAIDTVQCAtoENALXhDQDXVQCAu1ECALpJAgDbVQCA31UAgL/RAgC+SQIAvUECALxJAgCjMQ0A41UAgISIAwDrVQCA71UAgKZBDQClIQ0A81UAgKuRAgCqiQIA91UAgPtVAICvEQIArokCAK2BAgCsiQIAgKkAAIGpAACCTQAA/1UAgOFkEgDjTAIA4wgLAOGsAQADVgCA7zwCAO8YFgAHVgCAhlAGAIdIAwALVgCAD1YAgKiBAgCpgQIAqoECAKuBAgCsgQIArYECAK6FAgCvHQEAE1YAgBdWAIAbVgCAH1YAgCNWAIAnVgCAK1YAgIS4BQC4dQEAuX0BALp1AQC7CQEAvBkBAL0ZAQC+CQEAvwEBALBlAQCxbQEAsmUBALN9AQC0aQEAtV0BALZVAQC3TQEAL1YAgDNWAIA3VgCAO1YAgD9WAIBDVgCA7zQAAO/ADgDhXA4A4UwPAOOUAADjnA4AR1YAgIJlAACBfQAAgH0AAEtWAIBPVgCAvsQHALNFAgBTVgCAtUUCALZNAgBbVgCAhkAGAIeQBAC67QEAu+UBALz9AQC95QEAvuEBAL/VAQCflQgAngUIAJ3dDQCcPQwAmzEMAJr1DQCZ7RAAmD0QAJfVEQCWsRUAlQUUAJTlFQCTtRkAkjEYAJE5GACQDRwAj2EcAOdVAICz1QYAX1YAgLX9BgBXVgCAY1YAgLaRBgBnVgCAa1YAgLuVBgC6lQYAvVUHALxVBwC/VQcAvlUHAG9WAIBzVgCAqo0GAKuFBgCsnQYArYUGAK6BBgCvtQYAhKgAAHdWAIB7VgCAoyUFAH9WAIClJQUApi0FAINWAICHVgCAi1YAgI9WAICTVgCAl1YAgJtWAICfVgCAo1YAgKdWAICrVgCAr1YAgLNWAICjqQUAotEEAKHZBACgZQUAgiEdAIM1HQC3VgCAu1YAgIaVGACH3RQAhBkZAIUZGQCKDRUAi7EUAL9WAIDDVgCAjsURAI/VDACMzRAAjR0RAJJhDQCTdQ0AvkwAAMtWAICWxQkAl80EAJSNDACVXQkAmkEFAJtBBQCGyP8Ah0wAAIFZAACAeQAAnCEEAIJRAAChxQEAz1YAgKMB/ACi2QEApRX9AKS1/QCnufkApgH4AKkJ+AColfkAqwX1AKqt9QCtsfEArAHwAK8d8ACurfEAseHtALAB7ACzAegAsv3sALVd6QC09ekA01YAgNdWAIDbVgCA31YAgONWAIDnVgCA61YAgO9WAIDzVgCA91YAgKiNBACplQQAqpUEAKulBACsvQQArdkEAK75BACv8QQAhGz8APtWAID/VgCAA1cAgAdXAIALVwCAD1cAgBNXAIC4eQUAucUFALrNBQC7xQUAvN0FAL3FBQC+zQUAv+0FALCZBACxmQQAskkFALNJBQC0WQUAtVkFALZJBQC3SQUAox0EAL7M/AAXVwCAG1cAgB9XAICmWQQApTUEACNXAICrXQQAql0EACdXAIArVwCAr50FAK6dBQCtnQUArJ0FAC9XAICznQIAM1cAgDtXAIC2UQIAP1cAgENXAIC1uQIAukkCALtVAgCGSP0Ah8D8AL41AgC/PQIAvEUCAL09AgCo3QQAqUkDAKpRAwCrbQMArHUDAK2VAwCunQMAr7kDAICNAQCB5QEAguEBAEdXAIBLVwCAT1cAgFNXAIBXVwCAuJUDALmdAwC6lQMAu60DALy1AwC9vQMAvrUDAL9VAgCwyQMAsdUDALLVAwCzrQMAtLUDALW9AwC2tQMAt60DAFtXAIBfVwCAo9EDAGNXAICl9QMAZ1cAgGtXAICmHQMAb1cAgHNXAICrGQMAqgUDAK1xAwCsCQMAr3EDAK55AwDhKAcAd1cAgOPkBgB7VwCA4SgGAH9XAIDjaAEAg1cAgIdXAICLVwCA71gAAI9XAICTVwCAl1cAgO/IBgCbVwCAqE39AKmB/QCq0f0Aq9H9AKzx/QCt8f0ArvH9AK/x/QA3VwCAghEAAIEZAACA0f8An1cAgKNXAICEdAMAvnQDALh1/gC5ff4AunX+ALvF/gC83f4AvcX+AL7F/gC/9f4AsJH9ALGR/QCykf0As5H9ALRV/gC1Xf4AtlX+ALdN/gCzWf0Ap1cAgIasAACHRAMAq1cAgLZx/QC1ef0Ar1cAgLtV/QC6Vf0As1cAgLdXAIC/mf4AvpH+AL1F/QC8Rf0Au1cAgKMd/QC/VwCAw1cAgKY1/QDHVwCAy1cAgKU9/QCqEf0AqxH9AM9XAIDTVwCArtX+AK/d/gCsAf0ArQH9AKjN/wCp0f8AqtH/AKsh/gCsIf4ArSH+AK4h/gCvIf4A11cAgNtXAIDfVwCA41cAgOdXAIDrVwCA71cAgPNXAIC4jf4AuZH+ALqV/gC7rf4AvLX+AL25/gC+qf4Av6n+ALDh/gCx4f4AsuX+ALP5/gC06f4AtdX+ALbd/gC3uf4As1n/APdXAIDHVgCA+1cAgP9XAIC2of4Atan+AANYAIC7Jf4AuiX+AAdYAIALWACAvxH+AL4t/gC9Lf4AvDH+AIIZAACjHf8AgGUAAIEZAACm5f4AD1gAgBNYAICl7f4AqmH+AKth/gCEZAEAviAAAK5p/gCvVf4ArHX+AK1p/gAbWACA4zT+AB9YAIDhfP0AhrAEAIcIAwAjWACAJ1gAgCtYAIAvWACAhCQDAIQkBAAzWACA70j+ADdYAIA7WACAs+kCAD9YAIC+RAQAvkAFAENYAIC2nQIAtZkCAEdYAIC7iQIAur0CAEtYAIBPWACAv1kDAL5RAwC9WQMAvJECAKkdAgCoFQIAqyUCAKolAgCtWQIArFUCAK9NAgCuUQIAvmQGAFNYAIBXWACAW1gAgF9YAIBjWACAZ1gAgGtYAIC5+QMAuPEDALtNAwC68QMAvUEDALxZAwC/cQMAvkEDALEJAgCwPQIAs8kDALIBAgC12QMAtNEDALfJAwC20QMA4ZABAG9YAIDj8AAAc1gAgHdYAICCPQAAgT0AAIA9AAB7WACAf1gAgINYAICLWACAj1gAgJNYAIDvLAAAl1gAgKPpAwCbWACAhugEAIdgBQCfWACApp0DAKWZAwCjWACAq4kDAKq9AwCnWACAq1gAgK9ZAgCuUQIArVkCAKyRAwCvWACAs1gAgLdYAIC7WACAv1gAgMNYAIDHWACA71gBAISgBADhVP8Ay1gAgOOEAQDPWACA01gAgNdYAIDbWACAs9kBAN9YAICFzBkA41gAgOdYAIC28QEAtfkBAOtYAIC7pQEAutkBAO9YAIDzWACAv50BAL6dAQC9pQEAvK0BAKgBBgCpDQYAqhEGAKsRBgCsMQYArTEGAK4pBgCvJQYAh1gAgILJBwCBwQcAgPEHAPdYAID7WACAhhwAAIf8AwC47QYAufUGALr9BgC79QYAvO0GAL1RBwC+VQcAv00HALBdBgCxIQYAsjkGALMxBgC0GQYAtRkGALbdBgC31QYAo5kGAP9YAIADWQCAB1kAgAtZAICmsQYApbkGAA9ZAICr5QYAqpkGABNZAIAXWQCAr90GAK7dBgCt5QYArO0GABtZAICz8QcAH1kAgCNZAIC2gQcAJ1kAgCtZAIC1mQcAuo0HALtlBwAvWQCAM1kAgL59BwC/ZQcAvH0HAL11BwCoLQYAqTUGAKo9BgCrMQYArFUGAK1FBgCuRQYAr3UGADdZAIA7WQCAP1kAgENZAIBHWQCAS1kAgE9ZAIBTWQCAuOkGALn1BgC6/QYAu/UGALztBgC9kQYAvpUGAL+NBgCwDQYAseUGALLtBgCz5QYAtP0GALXlBgC27QYAt+UGAKO1BgBXWQCAW1kAgF9ZAIBjWQCApsUGAKXdBgAXWACAqyEGAKrJBgBnWQCAa1kAgK8hBgCuOQYArTEGAKw5BgCASQAAgUkAAIJZAACzRQEAb1kAgLVFAQC2RQEAc1kAgIZAAACHZAAAuikBALslAQC8PQEAvSEBAL4hAQC/FQEAd1kAgHtZAICEBAMAvgAMAOMoBgDv4AIA4RAGAH9ZAIDvkAYA4zwCAINZAIDh1AEAh1kAgItZAICPWQCAk1kAgJdZAICbWQCAo8ECAJ9ZAIClwQIAo1kAgKdZAICmwQIAq1kAgK9ZAICroQIAqq0CAK2lAgCsuQIAr5ECAK6lAgCpBQIAqLECAKsFAgCqBQIArQ0CAKwFAgCvNQIArjUCAISoDACzWQCAt1kAgLtZAIC/WQCAw1kAgMdZAIDLWQCAuekDALjhAwC7+QMAuuEDAL3pAwC84QMAv10DAL7hAwCxKQIAsCUCALM9AgCyIQIAtRkCALQtAgC32QMAthECAKitAgCp1QIAqtUCAKsNAQCsFQEArQkBAK4xAQCvLQEAz1kAgNNZAIDbWQCA31kAgONZAIDnWQCA61kAgO9ZAIC4IQEAuSEBALrtAQC75QEAvP0BAL3lAQC+7QEAv+UBALBVAQCxXQEAslUBALMtAQC0NQEAtTkBALYtAQC3JQEAgD0BAIGlAACCrQAA79QHAPNZAID3WQCA+1kAgO8oBwC+LAwA4fQGAP9ZAIDjkAcAA1oAgOGUAQAHWgCA4wwGALMdAgALWgCAh0QNAIZMDQAPWgCAtskBALXdAQATWgCAu9kBALrRAQAXWgCAG1oAgL+9AQC+sQEAvbkBALzBAQDXWQCAH1oAgCNaAIAnWgCAK1oAgC9aAIAzWgCAN1oAgKgJDwCpCQ8AqhkPAKsZDwCsCQ8ArQkPAK6pDwCvqQ8AsNkPALHtDwCy+Q8As/UPALSVDwC1hQ8AtoUPALe1DwC4jQ8AuWEAALphAAC7YQAAvGEAAL1hAAC+YQAAv2EAAKNdDQCCLQAAgRUAAIAdAAA7WgCApokOAKWdDgA/WgCAq5kOAKqRDgBDWgCAR1oAgK/9DgCu8Q4ArfkOAKyBDgBLWgCAs/UPAIboAwCHvAMAtu0PAE9aAIBTWgCAteUPALp5DwC7TQ8AV1oAgFtaAIC+NQ8AvyUPALxJDwC9RQ8AozEOAF9aAIBjWgCAZ1oAgGtaAICmKQ4ApSEOAG9aAICriQ4Aqr0OAHNaAIB3WgCAr+EOAK7xDgCtgQ4ArI0OAHtaAIB/WgCAg1oAgIdaAICLWgCAj1oAgJNaAICXWgCAm1oAgJ9aAICjWgCAp1oAgIANAACB1QAAgt0AAKtaAICoQQEAqVEBAKpRAQCrZQEArH0BAK2RAACukQAAr5EAAK9aAICzWgCAhGQBAL5kAQCGkAEAh4QAALtaAIC/WgCAuJEAALmRAAC6kQAAu5EAALyxAAC9sQAAvrEAAL+xAACw8QAAsfkAALLBAACzwQAAtLEAALWxAAC2sQAAt7EAALPZAgDDWgCAvnADAL5EBADHWgCAthEDALX1AgDLWgCAuz0DALo1AwDPWgCA01oAgL91AwC+dQMAvRUDALwVAwDXWgCAo50CANtaAIDfWgCAplUDAONaAIDnWgCApbECAKpxAwCreQMA61oAgO9aAICuMQMArzEDAKxRAwCtUQMAqDkDAKk5AwCqjQAAq50AAKyNAACtvQAArrUAAK/dAADzWgCA91oAgPtaAID/WgCAA1sAgAdbAIALWwCAD1sAgLhpAAC5aQAAunkAALt5AAC8aQAAvWkAAL7ZAQC/2QEAsKkAALGpAACyvQAAs7UAALSZAAC1mQAAtlkAALdZAAATWwCAF1sAgBtbAIAfWwCA70QAACNbAICGmAUAh+QCAOOYAACEqAIA4fgBACtbAICAOQAAgTkAAIItAAAvWwCAs0UBADNbAIA3WwCAO1sAgD9bAIC2fQEAtUUBAENbAIC7LQEAui0BAEdbAIBLWwCAvx0BAL4dAQC9IQEAvCkBAE9bAIDhUA4AU1sAgOM8DwBXWwCAW1sAgF9bAIBjWwCAZ1sAgGtbAIDjAAAAb1sAgHNbAIB3WwCAhPQFAO/kDgCuqQEAr6kBAKydAQCtlQEAqpkBAKuZAQB7WwCAf1sAgKbJAQCDWwCAh1sAgKXxAQCC/QcAo/EBAID9BwCB9QcAJ1sAgItbAICPWwCAk1sAgJdbAICbWwCAhrgDAIeQAwCoDQcAqRkHAKptBwCrZQcArH0HAK1lBwCuZQcAr1UHALAtBwCxxQcAssEHALPdBwC0xQcAtc0HALbFBwC3/QcAuMUHALnJBwC62QcAu9kHALypBwC9qQcAvp0HAL+VBwCzxQcAn1sAgKNbAICnWwCAq1sAgLbFBwC11QcAr1sAgLshBwC6yQcAs1sAgLdbAIC/KQcAviEHAL0pBwC8NQcAu1sAgKOBBwC/WwCAw1sAgKaBBwDHWwCAy1sAgKWRBwCqjQcAq2UHAM9bAIDTWwCArmUHAK9tBwCscQcArW0HAKgVAQCpgQEAqoEBAKuBAQCsgQEArYkBAK6xAQCvsQEA11sAgNtbAIDfWwCA41sAgOdbAIDrWwCA71sAgPNbAIC4ZQAAuW0AALplAAC7fQAAvGUAAL1tAAC+ZQAAv90AALChAQCxrQEAsqUBALO5AQC0qQEAtZ0BALaVAQC3XQAA91sAgIIdAACBHQAAgB0AAPtbAID/WwCAA1wAgL5YAQCErAIAB1wAgIcIAQCGjAEAC1wAgLdaAIAPXACAE1wAgLNJAQAXXACAG1wAgB9cAIAjXACAtkkBALVJAQAnXACAuykBALolAQArXACAL1wAgL8ZAQC+LQEAvS0BALwxAQC+2AMAM1wAgO/4BgA3XACAO1wAgD9cAIDv4AIAQ1wAgOGUAQBHXACA43QCAEtcAIDhmAUAT1wAgOMMBwBTXACAV1wAgFtcAICjwQIAhIwDAKXBAgBfXACAY1wAgKbBAgBnXACAa1wAgKuhAgCqrQIAraUCAKy5AgCvkQIArqUCAKgxAwCpPQMAqjUDAKtJAwCsWQMArVkDAK5JAwCvQQMAgMUAAIEJAACCGQAAb1wAgHNcAIB7XACAh2wDAIYcHAC47QAAufEAALr1AAC7jQAAvJUAAL2BAAC+gQAAv70AALAJAwCxCQMAsu0AALPhAAC04QAAteEAALblAAC32QAAf1wAgINcAICHXACAs7ECAItcAIC13QIAttUCAI9cAICTXACAl1wAgLrBAgC7wQIAvDUBAL05AQC+KQEAvykBAKaNAgCbXACAn1wAgKWFAgCjXACAo+kCAKdcAICrXACArnEBAK9xAQCsbQEArWEBAKqZAgCrmQIAr1wAgLNcAIC3XACA4YQGALtcAIDjJAYAv1wAgOGUAQDDXACA4ywAAL7oHQDHXACAy1wAgO/IAACE/B0AvvAcAM9cAIDvSAcA01wAgNdcAIDbXACA31wAgIEdAACAHQAA41wAgIIFAACGQBwAh8QcAOtcAIDvXACA81wAgPdcAID7XACA/1wAgKi1HgCpBR8Aqg0fAKsFHwCsAR8ArQkfAK45HwCvOR8A51wAgANdAIAHXQCAC10AgA9dAIATXQCAF10AgBtdAIC4yR8AudUfALrRHwC76R8AvPkfAL3tHwC+mR8Av5kfALAlHwCxLR8AsjkfALM1HwC0LR8AtQ0fALYFHwC3/R8As4UfAB9dAIAjXQCAJ10AgCtdAIC2iR8AtYkfAC9dAIC76R8AuuEfADNdAIA3XQCAv8kfAL7pHwC94R8AvO0fADtdAICjwR8AP10AgENdAICmzR8AR10AgEtdAIClzR8AqqUfAKutHwBPXQCAU10AgK6tHwCvjR8ArKkfAK2lHwCo6R4AqekeAKr5HgCr+R4ArOkeAK3pHgCuPQEArzUBAID5AQCBzQEAgsUBAIRgAgBXXQCAW10AgIdoAQCGnAAAuNEBALnZAQC64QEAu+EBALyRAQC9nQEAvpUBAL+JAQCwTQEAsVUBALJdAQCzVQEAtE0BALXxAQC28QEAt/EBALNxHgBfXQCAY10AgGddAIBrXQCAtmkeALVhHgBvXQCAu5EBALqJAQBzXQCAd10AgL81AQC+iQEAvYEBALyJAQB7XQCAd1wAgKM5HgB/XQCApSkeAINdAICHXQCApiEeAItdAICPXQCAq9kBAKrBAQCtyQEArMEBAK99AQCuwQEAk10AgJddAICbXQCAn10AgKNdAICnXQCAq10AgK9dAICzXQCAt10AgLtdAIC/XQCAw10AgMtdAIDPXQCAvnADAOHkHgCESAIA4+gfAIQABACAeQAAgXkAAIJpAADTXQCAhsAEAIdEAwDXXQCA210AgN9dAIDjXQCA7yAfAOddAIDrXQCA710AgPNdAIDvSAIA910AgPtdAID/XQCAA14AgL7oBAAHXgCAC14AgA9eAIATXgCA4ZABABdeAIDj6AIAs0kDABteAIAfXgCAI14AgCdeAIC2SQMAtUkDACteAIC7LQMAuiUDAC9eAIAzXgCAvxUDAL4VAwC9IQMAvCkDAKg1AgCpgQIAqoECAKuBAgCsgQIArYkCAK6xAgCvsQIAgP0BAIHNAQCCxQEAO14AgIaQBACHBAUAP14AgIRwBAC4SQEAuUkBALpZAQC7WQEAvEkBAL1JAQC+eQEAv3kBALChAgCxqQIAsr0CALO1AgC0kQIAtZECALZ5AQC3eQEAQ14AgEdeAIBLXgCAT14AgFNeAIBXXgCAW14AgO/QHgC+6AQA4VweAF9eAIDjkAAAY14AgGdeAIBrXgCAb14AgKNJAgBzXgCAd14AgHteAIB/XgCApkkCAKVJAgCDXgCAqy0CAKolAgCHXgCAi14AgK8VAgCuFQIArSECAKwpAgCoNQYAqT0GAKpVBgCrZQYArH0GAK1lBgCubQYAr2EGADdeAICPXgCAk14AgJdeAICADQAAgbEAAIKxAACbXgCAuOkGALnpBgC6+QYAu/UGALyVBgC9nQYAvpUGAL+NBgCw4QYAseEGALLhBgCz/QYAtOUGALXtBgC25QYAt9kGALPdBgCfXgCAo14AgKdeAICrXgCAtuUGALX1BgCvXgCAuyUGALolBgCGmAAAh6wAAL8pBgC+IQYAvSkGALw1BgCzXgCAo5kGALdeAIC7XgCApqEGAL9eAIDDXgCApbEGAKphBgCrYQYAx14AgMteAICuZQYAr20GAKxxBgCtbQYAqC0GAKk9BgCqiQYAq4kGAKyZBgCtmQYArokGAK+JBgDPXgCA014AgNdeAIDbXgCA314AgONeAIDnXgCA614AgLiNBgC5lQYAupUGALulBgC8vQYAvXEBAL5xAQC/cQEAsPkGALHNBgCy2QYAs9kGALTJBgC1yQYAtr0GALe1BgCzAQYA714AgPNeAID3XgCA+14AgLYZBgC1EQYA/14AgLsJBgC6PQYAA18AgAdfAIC/DQYAvg0GAL0NBgC8DQYAC18AgKNFBgDHXQCAD18AgKZdBgATXwCAhFgAAKVVBgCqeQYAq00GAL5oAQAXXwCArkkGAK9JBgCsSQYArUkGAIDBAwCByQMAgt0DAKPNAgAbXwCApdkCAKbNAgAfXwCAhoANAIeUAwCqxQIAqw0DAKwVAwCtHQMArhUDAK8NAwDhnBcA4xgGAOMUAwDhNAYA7xgCACNfAIAnXwCAK18AgOPQAgAvXwCA4VACADNfAIA3XwCA7ywGAO/kJQA7XwCArE0CAK1RAgCuUQIAr2UCAKgBAgCpCQIAqlkCAKtVAgCE7A0AP18AgENfAIBHXwCAvvgNAEtfAIBPXwCAU18AgLxRAwC9WQMAvmEDAL9hAwC47QMAuVEDALpRAwC7UQMAtM0DALXVAwC23QMAt9UDALAdAgCx1QMAst0DALPVAwDjyAAAV18AgOG4AQBbXwCAhFQPAF9fAIBjXwCAZ18AgKHpAgCgFQYAo6UDAKINAwDvIAAAa18AgG9fAIBzXwCAd18AgHtfAICFNCYAs40DAH9fAIC1mQMAto0DAINfAICGwA8Ah5QNALqFAwC7TQIAvFUCAL1dAgC+VQIAv00CAItfAICPXwCAk18AgJdfAICbXwCAn18AgI/d6wDvxAYAvuAPAOGMBgCjXwCA44AGAID1AACB5QAAguUAAKdfAICZbR8AmMUfAJvJGwCaeRoAnXUaAJzFGwCf+QcAnhkGAJFpFgCQsesAk20XAJLNFwCV0RMAlGkSAJdREgCWzRMAg1XkAIJB5ACHXwCAq18AgIeNHQCGkRgAhTkYAISVGQCLERwAigUcAK9fAICzXwCAj4UVAI6ZEACNORAAjJUdAJNRFACSRRQAt18AgLtfAICXYQkAlnUIAJWdCQCU+RUAm0EMAJqtDQC/XwCAw18AgMdfAIDLXwCAz18AgJzxDAChbQ0A018AgKMBBACihQAApZkEAKSRBACnGTgApsUFAKkJOACoKTgAq4k8AKoBPACtATAArB08AK8pMACunTAAseE0ALABNACzASgAsv00ALXZKAC00SgA118AgNtfAIDfXwCA418AgOdfAIDrXwCAgB0AAIEJAACC2QEA718AgKgRDwCpGQ8Aql0PAKtVDwCsTQ8ArXEPAK51DwCvbQ8A818AgPtfAICGiAAAhxABAP9fAIADYACAB2AAgAtgAIC4TQ4AuVEOALpRDgC7UQ4AvGUOAL1tDgC+ZQ4Avx0OALAdDwCxwQ8AssEPALPBDwC0xQ8Atc0PALbFDwC3eQ4As9UPAA9gAIATYACAF2AAgBtgAIC28Q8AtcUPAB9gAIC7BQ8AutkPACNgAIAnYACAvwkPAL4BDwC9FQ8AvBUPACtgAICjkQ8AL2AAgDNgAICmtQ8AN2AAgDtgAIClgQ8Aqp0PAKtBDwA/YACAQ2AAgK5FDwCvTQ8ArFEPAK1RDwCogQ0AqYENAKqBDQCrgQ0ArIENAK2BDQCusQ0Ar6ENAEdgAIBLYACAT2AAgFNgAIBXYACAgrkAAIG9AACAvQAAuDUCALk9AgC6zQIAu5UCALyNAgC9tQIAvr0CAL+1AgCwbQIAsU0CALJFAgCzJQIAtD0CALUdAgC2FQIAtw0CAFtgAIBfYACAswENAGNgAIC1AQ0Aa2AAgISUAwC2CQ0AviwEAG9gAIC7gQIAuqECAL35AgC8mQIAv9ECAL7xAgBzYACAd2AAgHtgAICjRQ0Af2AAgKVFDQCmTQ0Ag2AAgIbgBACHpAQAquUCAKvFAgCs3QIArb0CAK61AgCvlQIAqCUCAKk1AgCqPQIAqzUCAKwtAgCtkQIArpECAK+RAgCHYACAi2AAgI9gAICTYACAzAAAAJdgAICbYACAn2AAgLiZAgC5rQIAuqUCALttAQC8dQEAvX0BAL51AQC/bQEAsPECALH5AgCywQIAs8ECALSxAgC1vQIAtrUCALepAgCjYACA44QOAKdgAIDh9A4Aq2AAgK9gAICzYACAt2AAgIQgBQC7YACAv2AAgMNgAIDHYACA7+wOAMtgAIDPYACAs/UCANNgAICG6AQAh4wEAL5cBAC2UQIAteUCANtgAIC7fQIAunUCAN9gAIDjYACAvzkCAL41AgC9VQIAvFUCAKM1BQBnYACA12AAgOdgAIDrYACAppEFAKUlBQDvYACAq70FAKq1BQDzYACA92AAgK/5BQCu9QUArZUFAKyVBQCA+QcAgfkHAIKNBwCzjQYA+2AAgLWdBgC2iQYA/2AAgANhAIAHYQCAuk0HALtFBwC8XQcAvUEHAL5BBwC/QQcAC2EAgA9hAID3XwCAE2EAgBdhAIAbYQCAH2EAgCNhAICoNQYAqQEGAKppBgCraQYArHkGAK1lBgCuZQYAr50HALDlBwCx7QcAsuUHALP5BwC06QcAtekHALZZBwC3VQcAuHEHALlxBwC6cQcAu3EHALxVBwC9XQcAvlUHAL9NBwCjwQcAJ2EAgCthAIAvYQCAM2EAgKbFBwCl0QcAN2EAgKsJBgCqAQYAO2EAgD9hAICvDQYArg0GAK0NBgCsEQYAgGkAAIFpAACCBQAAQ2EAgL6YAQCEmAEAR2EAgEthAICGADwAh8QBAE9hAIBTYQCAV2EAgFthAIBfYQCAY2EAgKhdBgCpbQYAqmUGAKuBAQCsgQEArYkBAK6xAQCvsQEAZ2EAgGthAIBvYQCAc2EAgHdhAIB7YQCAf2EAgINhAIC4VQEAuV0BALpVAQC7yQAAvNkAAL3ZAAC+yQAAv8EAALCxAQCxuQEAsokBALOJAQC0cQEAtXEBALZ1AQC3bQEAs+0FAIdhAICLYQCAj2EAgJNhAIC2CQIAtQkCAJdhAIC7fQIAunUCAJthAICfYQCAv7UCAL61AgC9XQIAvF0CAL5gAgCjqQUAo2EAgKdhAICmTQIAq2EAgK9hAIClTQIAqjECAKs5AgCzYQCAhOADAK7xAgCv8QIArBkCAK0ZAgC+iDwAu2EAgKotAwCrJQMArD0DAK0lAwCuLQMAryUDAID1AACB/QAAgsEAAKPBAwC/YQCApcEDAKbBAwDDYQCAhmA8AIdUAwDHYQCAy2EAgM9hAIDjqAIA02EAgOGkAQDXYQCA71wCANthAIDfYQCA42EAgOdhAIDrYQCA72EAgPNhAIDjjAcA92EAgOE8BAD7YQCA/2EAgANiAIAHYgCAhCACAAtiAIAPYgCAE2IAgBdiAIDvbAcAG2IAgB9iAICzLQIAhEQ9ACNiAIArYgCAL2IAgLYtAgC1LQIAM2IAgLvJAgC6wQIAN2IAgDtiAIC/yQIAvsECAL3JAgC80QIA4XgHAOPAAADjOAYA4VwGAICpAACBqQAAgtEAAD9iAIBDYgCAR2IAgL6kPABLYgCAT2IAgO8cAADvkAYAU2IAgIZgPACHBD0AV2IAgLNxAQBbYgCAtRkBALYJAQBfYgCAY2IAgGdiAIC6AQEAuwEBALwBAQC9AQEAvgEBAL8BAQCohT4AqbU+AKq1PgCrxT4ArN0+AK3FPgCuwT4Ar/0+AGtiAIBvYgCAc2IAgHdiAIB7YgCAf2IAgINiAICHYgCAuFE/ALlRPwC6UT8Au1E/ALx1PwC9fT8AvnU/AL9tPwCwiT4AsYk+ALKZPgCzmT4AtIk+ALWJPgC2eT8At3U/ALdhAICjOT4Ai2IAgCdiAICmQT4Aj2IAgJNiAIClUT4Aqkk+AKtJPgCXYgCAm2IAgK5JPgCvST4ArEk+AK1JPgCASQAAgVEAAIJRAACzkT8An2IAgLW5PwC2RT8Ao2IAgIZAAACHBAMAukU/ALtdPwC8TT8AvT0/AL4pPwC/IT8AqE0+AKlVPgCqVT4Aq2U+AKx9PgCtiT4Arrk+AK+5PgCnYgCAq2IAgK9iAICzYgCAt2IAgLtiAIC/YgCAw2IAgLhhAQC5YQEAumEBALthAQC8YQEAvWEBAL5hAQC/YQEAsM0+ALHVPgCy1T4As6U+ALShPgC1qT4Atpk+ALeZPgCj3T4Ax2IAgMtiAIDPYgCA02IAgKYJPgCl9T4A12IAgKsRPgCqCT4A22IAgN9iAICvbT4ArmU+AK1xPgCsAT4A42IAgOdiAIDrYgCA72IAgPNiAID3YgCA+2IAgP9iAICAOQAAgTkAAIIFAAADYwCAvrgBAIS4AQALYwCAD2MAgKitAgCp1QIAqtUCAKstAwCsNQMArT0DAK41AwCvLQMAE2MAgBdjAIAbYwCAH2MAgCNjAIAnYwCAK2MAgC9jAIC46QMAuekDALqJAwC7iQMAvJkDAL2ZAwC+iQMAv4kDALBVAwCxXQMAslUDALPpAwC0+QMAtfkDALbpAwC34QMAs10CADNjAICGKAQAh8wDADdjAIC2vQMAtb0DADtjAIC7mQMAupEDAD9jAIBDYwCAvz0DAL49AwC9PQMAvIEDAIUAFACjGQIAR2MAgEtjAICm+QMAT2MAgFNjAICl+QMAqtUDAKvdAwBXYwCAW2MAgK55AwCveQMArMUDAK15AwDjVD4A4dw/AOHQPgDjPD4AX2MAgO8cAABjYwCAZ2MAgGtjAIDjwAAAb2MAgOHUAQDvYD4Ac2MAgHtjAIDvRD8AgGEAAIFtAACCfQAAhAAFAIbwBACHnAUAvhAFAH9jAICDYwCAh2MAgItjAICPYwCAk2MAgJdjAICbYwCAn2MAgLiJPQC5iT0Aupk9ALuRPQC8uT0Avbk9AL7RPQC/0T0AsAU+ALENPgCyBT4Asx0+ALQFPgC1DT4AtgU+ALe5PQConT4Aqa0+AKqlPgCrvT4ArKU+AK2tPgCupT4Ar30+AISsBAC+rAQAo2MAgKdjAICrYwCAr2MAgLNjAIC3YwCAqPkFAKn5BQCqKQYAqykGAKw5BgCtOQYArikGAK8pBgB3YwCAu2MAgL9jAIDDYwCAx2MAgMtjAIDPYwCA02MAgLiNBgC5kQYAupEGALulBgC8vQYAvUUHAL5BBwC/QQcAsFkGALFZBgCy7QYAs/0GALTtBgC13QYAttUGALe1BgCzoQYA12MAgNtjAIDfYwCA42MAgLa5BgC1sQYA62MAgLudBgC6nQYA52MAgAdjAIC/GQYAvikGAL0pBgC8OQYAglEAAKPlBgCAQQAAgUEAAKb9BgDvYwCA82MAgKX1BgCq2QYAq9kGAIZIAACHbAAArm0GAK9dBgCsfQYArW0GAKg5BgCpWQYAqmkGAKtpBgCseQYArXkGAK5pBgCvaQYA92MAgPtjAID/YwCAA2QAgAdkAIALZACAD2QAgBNkAIC4ZQEAuW0BALplAQC7fQEAvGUBAL1tAQC+ZQEAv9kBALAZBgCxGQYAsoEGALOBBgC0gQYAtYEGALaBBgC3gQYAs+EGABdkAIAbZACAH2QAgCNkAIC2+QYAtfEGACdkAIC73QYAut0GACtkAIAvZACAv0UGAL5FBgC9VQYAvFUGADNkAICjpQYAN2QAgDtkAICmvQYAP2QAgENkAICltQYAqpkGAKuZBgBHZACAS2QAgK4BBgCvAQYArBEGAK0RBgConQIAqdECAKrRAgCrLQMArDUDAK09AwCuNQMAry0DAE9kAIBTZACAvmQCAFtkAIBfZACAY2QAgGdkAIBrZACAuOkDALnpAwC6iQMAu4UDALydAwC9gQMAvoEDAL+1AwCwVQMAsV0DALJVAwCz6QMAtPkDALX5AwC26QMAt+EDAIBtAwCBpQAAgq0AALNVAgBvZACAtbEDALaxAwBzZACAhOACAHdkAIC6nQMAu5UDALyNAwC9MQMAvjEDAL8xAwCjGQIAe2QAgIVwaQB/ZACAg2QAgKb9AwCl/QMAh2QAgKvZAwCq0QMAhkgMAIe8AwCvfQMArn0DAK19AwCswQMAi2QAgI9kAICTZACAl2QAgO+wBgDvxAMAm2QAgJ9kAIDjfAYA45QDAOG4BwDh3AEAo2QAgKdkAICrZACAr2QAgLNkAIC3ZACAhEQCAL5YDQCADQAAgTUAAII9AAC7ZACAv2QAgMNkAICGyAwAh1wNAMtkAIDPZACA02QAgNdkAIDbZACA32QAgONkAIDnZACA62QAgO9kAIDzZACA74AGAISsDQDh7AYA92QAgONcBgD7ZACA/2QAgANlAIAHZQCAs/UBAAtlAIAPZQCAE2UAgBdlAIC2RQEAteUBABtlAIC7LQEAuiEBAB9lAIAjZQCAv/UAAL71AAC9JQEAvC0BAKgtDgCpNQ4Aqj0OAKs1DgCsLQ4ArYUOAK6FDgCvuQ4Ax2QAgCdlAIArZQCAL2UAgIAZAACBGQAAggUAADNlAIC4WQ8AuVkPALp5DwC7eQ8AvGkPAL1pDwC+GQ8AvxkPALClDgCxqQ4AsrkOALOxDgC0cQ8AtXEPALZxDwC3cQ8Apb0OAL6IAwA7ZQCAph0OADdlAIA/ZQCAo60OAENlAICtfQ4ArHUOAK+tDwCurQ8AV2QAgEdlAICrdQ4AqnkOALO5DwBLZQCAhmgAAIcMAwBPZQCAtlEPALVZDwBTZQCAu3UPALp1DwBXZQCAW2UAgL9FDwC+RQ8AvVEPALxlDwCocQ4AqXEOAKpxDgCrcQ4ArJEOAK2RDgCukQ4Ar5EOAF9lAIBjZQCAZ2UAgGtlAIBvZQCAc2UAgHdlAIB7ZQCAuIUOALmNDgC6hQ4Au50OALyNDgC9vQ4AvrUOAL95AQCw8Q4AsfEOALLxDgCzxQ4AtMEOALXBDgC2wQ4At8EOAKP5DgB/ZQCAg2UAgIdlAICLZQCAphEOAKUZDgCPZQCAqzUOAKo1DgCTZQCAl2UAgK8FDgCuBQ4ArREOAKwlDgCADQAAgRUAAIIdAACbZQCAn2UAgKNlAICElAEAvpQBAIZABwCH5AAAq2UAgK9lAICzZQCAt2UAgLtlAIC/ZQCAqIkCAKmRAgCqlQIAq7kCAKzVAgCtxQIArsUCAK/1AgDDZQCAx2UAgMtlAIDPZQCAvnwDANNlAIDXZQCA22UAgLh9AwC5wQMAusEDALvBAwC8wQMAvckDAL7xAwC/8QMAsI0CALFFAwCyTQMAs0UDALRdAwC1RQMAtk0DALdFAwCzHQIA32UAgONlAIDnZQCA62UAgLZFAgC1XQIA72UAgLuBAwC6SQIA82UAgPdlAIC/gQMAvpkDAL2RAwC8mQMA+2UAgKNZAgD/ZQCAA2YAgKYBAgAHZgCAC2YAgKUZAgCqDQIAq8UDAA9mAIATZgCArt0DAK/FAwCs3QMArdUDAIDZAQCB7QEAguUBAO+4DgAbZgCA4cQBAISYAgDj1AAAH2YAgL7sBAAjZgCA7wgAACdmAIDhxA8AK2YAgONkDgCGAAUAh2gFAC9mAICzvQIAM2YAgLWtAgC2pQIAN2YAgDtmAIA/ZgCAukEBALtBAQC8RQEAvU0BAL5FAQC/+QEAQ2YAgEdmAIBLZgCAT2YAgFNmAIBXZgCAW2YAgO/gAQCEbAQA4dQOAF9mAIDjHA4AY2YAgGdmAIBrZgCAb2YAgKMxAgBzZgCAhCQHAHdmAIB7ZgCApikCAKUhAgB/ZgCAq80BAKrNAQCDZgCAi2YAgK91AQCuyQEArcEBAKzJAQCo6QUAqekFAKr5BQCr+QUArOkFAK3pBQCuOQYArzkGABdmAICCzQcAgfUHAID9BwCHZgCAj2YAgIYYAwCHkAMAuNEGALnZBgC64QYAu+EGALyRBgC9nQYAvpUGAL+JBgCwSQYAsUkGALJdBgCzVQYAtE0GALXxBgC28QYAt/EGALDhBwCx4QcAsgkHALMJBwC0GQcAtRkHALYJBwC3CQcAuDkHALkNBwC6GQcAuxkHALwJBwC9CQcAvn0HAL9xBwCTZgCAp2UAgJdmAICbZgCAn2YAgKNmAICnZgCAq2YAgKjxBwCpxQcAqsEHAKvdBwCsyQcArb0HAK6pBwCvoQcAsykGAK9mAICzZgCAt2YAgLtmAIC2XQYAtSEGAL9mAIC7RQYAukUGAMNmAIDHZgCAv70GAL69BgC9vQYAvL0GAMtmAICjbQYAz2YAgNNmAICmGQYA12YAgNtmAIClZQYAqgEGAKsBBgDfZgCA42YAgK75BgCv+QYArPkGAK35BgCobQYAqbEBAKpJAQCrRQEArF0BAK1FAQCuTQEAr0UBAOdmAICCHQAAgR0AAIAdAADrZgCA72YAgPNmAIC+VAEAuIEAALmNAAC6hQAAu5kAALyJAAC9vQAAvrUAAL99AACwPQEAseEAALLhAACz4QAAtOEAALXpAAC20QAAt9EAALsFAwC62QIAhiwCAIcsAwC/DQMAvgUDAL0VAwC8FQMAs+ECAPtmAID/ZgCAhCwDAANnAIC25QIAtfUCAAdnAICqnQIAq0EDAAtnAIAPZwCArkEDAK9JAwCsUQMArVEDABNnAICjpQIAF2cAgBtnAICmoQIAH2cAgCNnAIClsQIAqakAAKihAACrtQAAqr0AAK3dAACs3QAAr/EAAK79AAC+LBwAJ2cAgCtnAIAvZwCAM2cAgDdnAIA7ZwCAP2cAgLl9AAC4fQAAu80BALrNAQC93QEAvN0BAL/NAQC+zQEAsZUAALCJAACzTQAAspUAALVdAAC0XQAAt00AALZNAABDZwCAR2cAgEtnAIBPZwCAU2cAgFdnAIBbZwCAX2cAgIA5AACBOQAAggUAAGNnAIBrZwCAb2cAgIf4AgCGfB0A4bgEAL7IHADjQAYAc2cAgHdnAIB7ZwCAf2cAgINnAICHZwCAi2cAgI9nAICTZwCAl2cAgJtnAIDvsAcAn2cAgKNnAICnZwCAq2cAgO/IAACvZwCAs2cAgLdnAIDvQAYAu2cAgOH8BgC/ZwCA4xwGAMNnAIDhlAEAx2cAgONkBgCAEQAAgRkAAIIpAACz/QEAy2cAgLWdAQC2lQEAz2cAgNNnAICEbB0AuoUBALuZAQC8iQEAvVEBAL5RAQC/UQEAozEeAGdnAIDXZwCA22cAgN9nAICmWR4ApVEeAONnAICrVR4AqkkeAIYIAwCHbAMAr50eAK6dHgCtnR4ArEUeAOdnAICzCR8A62cAgO9nAIC2CR8A82cAgPdnAIC1CR8AugUfALsNHwD7ZwCA/2cAgL4FHwC/CR8AvBUfAL0NHwCw5R8Ase0fALLlHwCz/R8AtOUfALXpHwC2GR8AtxkfALgpHwC5NR8Auj0fALs1HwC8ER8AvR0fAL4JHwC/BR8AA2gAgAdoAID3ZgCAC2gAgA9oAIATaACAF2gAgBtoAICo0R8AqdEfAKqlHwCrvR8ArKUfAK2tHwCupR8Ar50fAKNNHgAfaACAI2gAgCdoAIAraACApk0eAKVNHgAvaACAq0keAKpBHgAzaACAN2gAgK9NHgCuQR4ArUkeAKxRHgCADQAAgRUAAIIdAAA7aACAP2gAgENoAICEtAEAvrQBAL/oAQBLaACAhkgHAIc0AACEvAYAT2gAgFNoAIC+tAYAqI0BAKmVAQCqlQEAq80BAKzZAQCt2QEArs0BAK/FAQBXaACAW2gAgF9oAIBjaACAZ2gAgGtoAIBvaACAc2gAgLgdAQC5wQAAusEAALvBAAC8wQAAvckAAL7xAAC/8QAAsIkBALGJAQCyKQEAsykBALQ9AQC1JQEAti0BALclAQC7bQIAum0CAHdoAIB7aACAv8ECAL7ZAgC93QIAvN0CALM9AgB/aACAg2gAgIdoAICE/AYAtnkCALVxAgCLaACAqikCAKspAgCPaACAk2gAgK6dAgCvhQIArJkCAK2ZAgCXaACAo3kCAJtoAICfaACApj0CAKNoAICnaACApTUCAIJtJwCDjSoAhqgFAIdsAwCGmS4Ah80vAIQRLgCFmS4AiiESAIspEgCraACAr2gAgI6RFgCPHRYAjBESAI0RFgCScRoAk+UaALNoAIDvlHYAlvEeAJflHgCUSRoAlRkeAJopAgCb4QIAu2gAgL9oAIDDaACA4SASAJzxAgDjIBYAnyEfAJ7BHwCdmRsAnC0bAJuhGwCavRcAmTkXAJixFwCXiRMAlqkTAJWpEwCUdS4AkzkvAJIxLwCRsS8AkDUrAI+tJgDjeB8A0gAAAOFcHwCCmQEAx2gAgIDxAQCB8QEAvqgHAMtoAIDPaACA02gAgIS8BgDvLB8A12gAgNtoAIDhpB4A48wAAON8HgDhvAEA32gAgONoAIDnaACAhJwGAOtoAIC+bAYA72gAgPNoAID3aACA7xAAAO8EHgD7aACA/2gAgANpAIAHaQCAC2kAgA9pAIATaQCAF2kAgBtpAICAPQAAgQkAAILJBwAfaQCAo/kDAKLxAwChMQMAoM0fALBJcQCxAXwAsgl8ALMhfQC0AXgAtRV4AEdoAIC3aACAI2kAgL4oDgCGDAAAh4wDACdpAIAraQCAL2kAgDNpAIA3aQCAoV0AAKJVAACjfQAApAEMAKUVDACm9QwApwEIAKghCACpxQgAqgF0AKsJdACsAXQArR11AK55cACveXAAqOUFAKnxBQCq8QUAqy0FAKw1BQCtPQUArjUFAK8tBQA7aQCAP2kAgENpAIBHaQCAS2kAgE9pAIBTaQCAV2kAgLj9BgC5jQYAuoUGALutBgC8uQYAvbkGAL6tBgC/pQYAsFUFALFdBQCyVQUAs+UGALT9BgC10QYAttEGALfRBgCzeQQAW2kAgF9pAIBjaQCAZ2kAgLa9BAC1vQQAa2kAgLuZBAC6kQQAb2kAgHNpAIC/FQcAvjkHAL0xBwC8gQQAd2kAgKM9BAB7aQCAf2kAgKb5BACDaQCAh2kAgKX5BACq1QQAq90EAItpAICPaQCArn0HAK9RBwCsxQQArXUHAKhpBwCpaQcAqnkHAKvZBgCs9QYArf0GAK71BgCv5QYAgMkAAIHJAACCBQAAk2kAgIZwDwCHNAAAm2kAgJ9pAIC4fQYAuQUGALoNBgC7BQYAvB0GAL0FBgC+DQYAvwUGALCdBgCxdQYAsn0GALN1BgC0UQYAtV0GALZVBgC3TQYAs/EEAKNpAICnaQCAq2kAgK9pAIC2fQUAtX0FALNpAIC7sQUAulkFALdpAIC7aQCAv5kFAL6VBQC9oQUAvKkFAL9pAICjtQQAw2kAgMdpAICmOQUAy2kAgM9pAIClOQUAqh0FAKv1BQDTaQCA12kAgK7RBQCv3QUArO0FAK3lBQCpuQIAqLECAKvJAgCqsQIArTUCAKw1AgCvNQIArjUCANtpAIDfaQCA42kAgOdpAIDraQCA72kAgPNpAID3aQCAuekDALjZAwC7iQMAuuEDAL2dAwC8nQMAv4EDAL6JAwCxVQIAsFUCALNVAgCyVQIAtfkDALTxAwC36QMAtvEDALM9AwD7aQCA/2kAgANqAIALagCAtrEDALW5AwAPagCAu5UDALqVAwCGiAwAh6ANAL85AgC+MQIAvYUDALyFAwATagCAo3kDABdqAIAbagCApvUDAB9qAIAjagCApf0DAKrRAwCr0QMAJ2oAgCtqAICudQIAr30CAKzBAwCtwQMAgIUAAIGNAACChQAA79AGAOOwBwDj9AQA4QgHAOHsBADvOAYA7yAEAL6kDAAvagCAM2oAgOGEAQA3agCA49wGADtqAIA/agCAhMANALPJAQBDagCAtdkBALbJAQBHagCAS2oAgE9qAIC6xQEAu60BALy5AQC9uQEAvq0BAL+lAQCwLQ4AsUUOALJBDgCzQQ4AtEUOALVNDgC2cQ4At3EOALiBDgC5gQ4AuoEOALuBDgC8gQ4AvYEOAL6BDgC/gQ4AB2oAgFNqAIBXagCAW2oAgJdpAIBfagCAY2oAgGdqAICo2Q0AqdkNAKptDgCrZQ4ArH0OAK1lDgCuZQ4Ar1UOAKOFDgCCLQAAgRUAAIAdAABragCApoUOAKWVDgBvagCAq+EOAKqJDgBzagCAd2oAgK/pDgCu4Q4ArfUOAKz1DgB7agCAs4UPAIZoAACHHAMAtoUPAH9qAICDagCAtZEPALqNDwC7SQ8Ah2oAgItqAIC+MQ8AvzEPALxJDwC9RQ8AqBEOAKkZDgCqSQ4Aq0UOAKxdDgCtQQ4ArkEOAK91DgCPagCAk2oAgJdqAICbagCAn2oAgKNqAICnagCAq2oAgLihDgC5oQ4Aug0BALsFAQC8HQEAvQEBAL4BAQC/AQEAsA0OALHJDgCy2Q4As9UOALSxDgC1sQ4AtqkOALehDgCjwQ4Ar2oAgLNqAIC3agCAu2oAgKbBDgCl1Q4Av2oAgKsNDgCqyQ4Aw2oAgMdqAICvdQ4ArnUOAK0BDgCsDQ4Ay2oAgM9qAIDTagCA12oAgIANAACBNQAAgj0AANtqAIDfagCA42oAgISEAQC+hAEAhjAHAIf4AADragCA72oAgKjBAgCp0QIAqtECAKvlAgCs/QIArTUDAK49AwCvNQMA82oAgPdqAID7agCA/2oAgANrAIAHawCAC2sAgA9rAIC40QMAudkDALrhAwC74QMAvJEDAL2RAwC+kQMAv5EDALBNAwCxVQMAsl0DALNVAwC0TQMAtfEDALbxAwC38QMAu7EDALqpAwATawCAvoQDAL8VAwC+qQMAvaEDALypAwCzeQIAF2sAgBtrAIAfawCAI2sAgLaVAwC1VQIAJ2sAgKrtAwCr9QMAK2sAgC9rAICu7QMAr1EDAKztAwCt5QMAM2sAgKM9AgA3awCAO2sAgKbRAwA/awCAQ2sAgKURAgBHawCAgiEAAIEVAACAFQAA7wQAAISUAgBLawCAT2sAgOPYAABTawCA4fgBAFtrAIBfawCAY2sAgGdrAIBrawCAhmAFAIcIBQBvawCAs20BAHNrAIC1fQEAtnUBAHdrAIB7awCAf2sAgLpRAQC7UQEAvPkBAL3RAQC+0QEAv9EBAINrAICjpQEAh2sAgItrAICmvQEAj2sAgJNrAICltQEAqpkBAKuZAQCXawCAm2sAgK4ZAQCvGQEArDEBAK0ZAQCfawCA4fQOAKNrAIDjFA4A9AAAAOF8DACnawCA41AKAKtrAICvawCAviAEAO8wDQCzawCAt2sAgIQ0BADvrA4AsDkGALE5BgCygQYAs6kGALS5BgC1uQYAtqkGALehBgC46QYAuekGALrJBgC7xQYAvN0GAL3BBgC+wQYAvz0HAFdrAICCHQAAgR0AAIAdAAC7awCAv2sAgMNrAIDnagCAqJkFAKmZBQCqSQYAq0kGAKxZBgCtWQYArkkGAK9JBgCorQcAqbUHAKq9BwCrtQcArK0HAK3dBwCuyQcAr8EHAMdrAIDLawCAhogDAIcQAwDPawCA02sAgNdrAIDbawCAuG0HALkFBwC6AQcAuxUHALwxBwC9MQcAvikHAL8pBwCwgQcAsYEHALJpBwCzZQcAtH0HALVhBwC2YQcAt1UHALM1BgDfawCA42sAgOdrAIDrawCAtl0GALUlBgDvawCAu0UGALpFBgDzawCA92sAgL+lBgC+uQYAvbEGALy9BgD7awCAo3EGAP9rAIADbACAphkGAAdsAIALbACApWEGAKoBBgCrAQYAD2wAgBNsAICu/QYAr+EGAKz5BgCt9QYAqCUBAKk1AQCqPQEAqzUBAKwtAQCtkQAArpEAAK+RAAAXbACAG2wAgB9sAIAjbACAJ2wAgIK9AwCBvQMAgL0DALiZAAC5rQAAuqUAALttAAC8dQAAvX0AAL51AAC/bQAAsPEAALH5AACywQAAs8EAALSxAAC1vQAAtrUAALepAAArbACAL2wAgDNsAICEgAIAvhwCADtsAICG+HwAh8wCAISsAwA/bACAQ2wAgEdsAIBLbACAT2wAgFNsAIBXbACAs/UCAFtsAIBfbACAkgAAAGNsAIC2UQMAteUCAGdsAIC7fQMAunUDAGtsAIBvbACAvzkDAL41AwC9VQMAvFUDAKM1AgBzbACAd2wAgHtsAIB/bACAppEDAKUlAgCDbACAq70DAKq1AwCHbACAi2wAgK/5AwCu9QMArZUDAKyVAwC+wAMAj2wAgJNsAICXbACAgA0AAIE1AACCPQAAm2wAgJ9sAICjbACAhsh8AIcAAwCrbACAr2wAgLNsAIC3bACAu2wAgL9sAIDDbACAx2wAgMtsAIDPbACA02wAgO/0AwCE7HwA4ZQBANdsAIDjMAMA22wAgN9sAIDjbACA52wAgLNpAQDrbACA72wAgPNsAID3bACAtmEBALVpAQD7bACAuykBALohAQD/bACAA20AgL8dAQC+HQEAvSUBALwtAQAHbQCAC20AgA9tAICjpQEAE20AgKWlAQCmrQEAvlR8AIaAfACH7HwAqu0BAKvlAQCs4QEArekBAK7RAQCv0QEAG20AgOGcBgCEBH8A4yQGAOPUBgAfbQCA4TAEACNtAIDvlAcAgnUAAIFhAACAaQAAJ20AgCttAIAvbQCA7+wGALiNfgC5lX4AupV+ALulfgC8vX4AvdF+AL7RfgC/0X4AsGV+ALFtfgCyeX4As3F+ALRZfgC1WX4Atr1+ALe1fgCoVX4AqWF+AKphfgCrYX4ArGF+AK1hfgCuYX4Ar2F+ADNtAICnbACAN2wAgDdtAIAXbQCAO20AgD9tAIBDbQCAqHF+AKlxfgCqcX4Aq3F+AKyRfwCtkX8ArpF/AK+RfwBHbQCAS20AgE9tAIBTbQCAV20AgFttAIBfbQCAY20AgLiFfwC5jX8AuoV/ALudfwC8jX8Avb1/AL61fwC/XX8AsPF/ALHxfwCy8X8As8V/ALTBfwC1wX8AtsF/ALfBfwCz+X8AZ20AgGttAIBvbQCAc20AgLYRfgC1GX4Ad20AgLs1fgC6NX4Ae20AgH9tAIC/BX4AvgV+AL0RfgC8JX4AghUAAKO9fwCAYQAAgWEAAKZVfgCDbQCAvpABAKVdfgCqcX4Aq3F+AIdtAICLbQCArkF+AK9BfgCsYX4ArVV+AKhBfgCpUX4AqlV+AKt9fgCsZX4ArW1+AK75AQCv8QEAhgAAAIc0AQCPbQCAk20AgJdtAICbbQCAn20AgKNtAIC4dQEAuX0BALp1AQC7yQAAvNkAAL3ZAAC+yQAAv8EAALCVAQCxnQEAspUBALNNAQC0VQEAtV0BALZVAQC3TQEAs919AKdtAICrbQCAr20AgLNtAIC27X0Ate19ALdtAIC7WQIAulECALttAIC/bQCAv5kCAL6RAgC9mQIAvEECAMNtAICjmX0Ax20AgMttAICmqX0Az20AgNNtAIClqX0AqhUCAKsdAgDXbQCA220AgK7VAgCv3QIArAUCAK3dAgDfbQCA420AgOdtAIDrbQCAgB0AAIEJAACCOQAA720AgPNtAIC+AAQA+20AgP9tAIADbgCAB24AgAtuAIAPbgCAhIwDABNuAICHCAMAhuwEABduAIDviAIAG24AgB9uAICEbAQA4zQCACNuAIDhVAEAJ24AgCtuAIAvbgCAM24AgKhtAgCprQIAqqUCAKu9AgCspQIAra0CAK6lAgCvGQEAvqwEADduAIA7bgCAP24AgENuAIBHbgCAS24AgE9uAIC4DQEAuREBALoRAQC7JQEAvD0BAL3VAQC+3QEAv9UBALBpAQCxaQEAsnkBALNxAQC0WQEAtVkBALY5AQC3NQEAsy0CAFNuAIBXbgCAW24AgF9uAIC2LQIAtS0CAGNuAIC7rQEAuq0BAGtuAIBvbgCAv50BAL6dAQC9pQEAvK0BAIBNAACBVQAAglUAAO9sAABzbgCA7+x/AO+8fgB3bgCA4RB/AOPUfwDj2H4A4ex/AHtuAIDhTH4Af24AgOMkfgD3bQCAZ24AgKsFBgCqBQYArQ0GAKwFBgCvNQYArjUGAIYAAwCHKAMAo4UFAINuAIClhQUAh24AgItuAICmhQUAs/EGAI9uAICTbgCAl24AgJtuAIC26QYAteEGAJ9uAIC7vQYAur0GAKNuAICnbgCAv4kGAL6BBgC9iQYAvJUGAKgpBgCpKQYAqjkGAKs5BgCsKQYArSkGAK5dBgCvTQYAq24AgK9uAICzbgCAt24AgLtuAIC/bgCAw24AgMduAIC46QcAuekHALr5BwC7+QcAvOkHAL3pBwC+XQcAv1UHALA5BgCxOQYAsgEGALMdBgC0BQYAtQ0GALYFBgC32QcAo7EHAIItAACBFQAAgB0AAMtuAICmqQcApaEHAM9uAICr/QcAqv0HANNuAICEpAIAr8kHAK7BBwCtyQcArNUHAL7MAQCzlQYA124AgNtuAIC2qQYA324AgONuAIC1rQYAulkBALshAQCGyAAAhwwBAL4hAQC/KQEAvDEBAL0xAQCoKQYAqSkGAKpZBgCrUQYArGEGAK1tBgCutQEAr6kBAITgAQDnbgCA624AgO9uAIDzbgCA924AgPtuAID/bgCAuGEBALlhAQC6YQEAu2EBALxhAQC9YQEAvmEBAL9hAQCw2QEAsaEBALKhAQCzoQEAtKEBALWpAQC2kQEAt5EBAKPRBQADbwCAB28AgAtvAIAPbwCApu0FAKXpBQATbwCAq2UCAKodAgAXbwCAG28AgK9tAgCuZQIArXUCAKx1AgAfbwCAI28AgCdvAIArbwCAL28AgDNvAIA3bwCAO28AgIA9AACBCQAAghkAAD9vAIBDbwCAS28AgL48AwBPbwCAhgAMAIcUAwBTbwCAs9UDAFdvAIC1PQMAtjUDAFtvAIBfbwCAv4wKALoRAwC7EQMAvLUAAL29AAC+tQAAv60AAGNvAIDjdAEAZ28AgOG8AQBrbwCAb28AgHNvAIB3bwCAe28AgH9vAICDbwCAh28AgItvAIDvdAIAj28AgJNvAICoTQIAqVECAKpRAgCrqQIArLkCAK25AgCuqQIAr6kCAIRsDQCXbwCAm28AgJ9vAICjbwCAp28AgKtvAIC+dA0AuG0BALkFAQC6DQEAuwUBALwdAQC9BQEAvg0BAL8FAQCw2QIAsdkCALJtAQCzZQEAtH0BALVlAQC2ZQEAt1UBAOG4AQDhUAcA47QAAON8BwCAqQAAgQkAAII5AACvbwCAs28AgLtvAIC/bwCAw28AgO4AAADHbwCA7wAAAO9kBgCGYAwAh+QMAKORAgDLbwCApXkCAM9vAIDTbwCApnECANdvAIDbbwCAq1UCAKpVAgCt+QEArPEBAK/pAQCu8QEAt28AgEdvAIDfbwCA428AgOdvAIDrbwCA728AgPNvAICoVQ4AqVkOAKqhDgCrvQ4ArK0OAK2VDgCu+Q4Ar/UOALCRDgCxkQ4AspEOALORDgC0sQ4AtbEOALaxDgC3sQ4AuJEOALmdDgC6lQ4Au0kPALxZDwC9WQ8AvkkPAL9JDwCzCQ4A928AgPtvAID/bwCAA3AAgLY1DgC1BQ4AB3AAgLt1DgC6dQ4AC3AAgA9wAIC/VQ4AvlUOAL1lDgC8ZQ4AE3AAgKNNDgAXcACAG3AAgKZxDgAfcACAI3AAgKVBDgCqMQ4AqzEOAISkAwC+pAMArhEOAK8RDgCsIQ4ArSEOAKilDgCprQ4AqqUOAKu5DgCs3Q4ArcEOAK7BDgCv/Q4AgO0BAIHxAQCC8QEAJ3AAgIaQAQCHtAEAK3AAgC9wAIC4yQEAuckBALrZAQC70QEAvPkBAL35AQC+mQEAv5UBALCFDgCxbQEAsmUBALN9AQC0ZQEAtW0BALZlAQC3+QEAsy0OADNwAIA3cACAO3AAgD9wAIC2QQ4AtVUOAENwAIC7qQEAukEOAEdwAIBLcACAv6kBAL6hAQC9qQEAvLEBAE9wAICjaQ4AU3AAgFdwAICmBQ4AW3AAgF9wAIClEQ4AqgUOAKvtAQBjcACAZ3AAgK7lAQCv7QEArPUBAK3tAQCoOQMAqTkDAKqNAwCrhQMArJ0DAK2FAwCuhQMAr7UDAGtwAIBvcACAc3AAgHdwAIB7cACAf3AAgINwAICHcACAuGEAALlhAAC6YQAAu2EAALxhAAC9YQAAvmEAAL9hAACwzQMAsaUDALKhAwCzoQMAtKUDALWtAwC2kQMAt5EDAIANAACBEQAAghEAAItwAIDv9AIAj3AAgJNwAIC+HAMA4xQCAISIAgDhgAEAm3AAgJ9wAICjcACAh8gDAIY8BAC7AQMAumkDAKdwAICrcACAvwkDAL4BAwC9FQMAvBUDALNlAwCvcACAs3AAgLdwAIC7cACAtmUDALV1AwC/cACAw3AAgMdwAIDLcACAo4kCAM9wAIClmQIApokCANNwAICELAIA13AAgKqFAgCr7QIArPkCAK35AgCu7QIAr+UCANtwAIDfcACAvkQFAIRMBQDjcACA53AAgOtwAIDvcACA83AAgPdwAID7cACA/3AAgIAZAACBGQAAggUAAANxAIDhGA8A4VwOAOO4DgDjdAEAC3EAgA9xAIATcQCAF3EAgIYABACHZAUAG3EAgB9xAIAjcQCAJ3EAgO98DgDvqAEAs3UBACtxAIAvcQCAM3EAgDdxAIC2MQEAtRUBADtxAIC7HQEAuhUBAD9xAIBDcQCAv+EAAL79AAC9/QAAvP0AAAdxAIBHcQCAS3EAgE9xAICXcACAU3EAgFdxAIBbcQCAqI0GAKmVBgCqnQYAq+UGAKz9BgCt0QYArtEGAK/RBgCwsQYAsbkGALJJBwCzSQcAtFkHALVFBwC2RQcAt3kHALghBwC5IQcAujkHALs5BwC8KQcAvSkHAL4ZBwC/GQcAozUGAF9xAIBjcQCAZ3EAgGtxAICmcQYApVUGAG9xAICrXQYAqlUGAHNxAIC+oAMAr6EHAK69BwCtvQcArL0HAIBRAACBWQAAgmEAALNVBwCF9AAAtX0HALZ1BwB3cQCAhgAcAIfkAQC6LQcAuyUHALw9BwC9JQcAviUHAL8VBwCokQYAqZEGAKqRBgCrkQYArLkGAK25BgCuqQYAr6kGAHtxAIB/cQCAg3EAgIdxAICiIQEAozUBAKA5BQChEQQAuEkBALlJAQC6XQEAu1UBALxNAQC90QEAvtEBAL/RAQCwpQYAsa0GALKlBgCzvQYAtK0GALWdBgC2lQYAt3kBAKMZBgCPnXkAi3EAgI9xAICTcQCApjkGAKUxBgCXcQCAq2kGAKphBgCbcQCAn3EAgK9ZBgCuaQYArWkGAKxxBgCeiQgAn8EFAJzJCQCdyQkAmqENAJu9DACYsQ0AmbkNAJahcQCXRXEAlEV1AJWxcQCSoXUAk7V1AJDleQCRzXkAil1yAItFcgCjcQCAvoAcAI51DgCPZQ4AjLlyAI11DgCCOXoAgzl6AKdxAICrcQCAhnF2AIeZdgCECXoAhW12AJptBwCbVQIAr3EAgLNxAIC3cQCA4ZAAAJxZAgDjCBoAkgkPAJNlCgC7cQCA7zgWAJZ1BgCXdQYAlH0KAJU1CwCpjRYAqIUWAKsBEACqMRYArXESAKy1EgCvuS4ArgEsAKF9AgC/cQCAo6EeAKKpHgClsRoApPUfAKflGwCmsRoAhMwDAIRMHADDcQCAx3EAgMtxAIDPcQCA03EAgNdxAICxASgAsNkuALONKgCy6SoAtfUmALQBJACEcB0A23EAgID9AQCBFQAAgh0AAL6AHADfcQCA43EAgIe4AgCGPB0A63EAgO9xAIDzcQCA93EAgPtxAID/cQCAA3IAgAdyAIALcgCAD3IAgBNyAIAXcgCA44ADABtyAIDhoAEAH3IAgO+UAwAjcgCAJ3IAgCtyAIAvcgCAM3IAgDdyAIA7cgCAP3IAgOE8BgBDcgCA49AGAEdyAIDhMAcAS3IAgOOsBgCAOQAAgRUAAIIdAADvHAYAT3IAgFNyAIC+uB8A7+gBALPpAgBbcgCAh8QcAIbsHABfcgCAtlkCALVRAgBjcgCAu00CALpNAgBncgCAa3IAgL+5AQC+2QEAvdEBALz1AQCjKR0A53EAgFdyAIBvcgCAc3IAgKaZHQClkR0Ad3IAgKuNHQCqjR0Ae3IAgH9yAICveR4ArhkeAK0RHgCsNR4Ag3IAgLNtHwCHcgCAi3IAgLZlHwCPcgCAk3IAgLVtHwC6IR8AuyEfAJdyAICbcgCAviUfAL8pHwC8MR8AvTEfAKihHwCpoR8AqqEfAKuhHwCsoR8AraEfAK6hHwCvoR8An3IAgKNyAICncgCAq3IAgK9yAICzcgCAt3IAgLtyAIC4rR8AubUfALq9HwC7tR8AvK0fAL1VHwC+UR8Av00fALChHwCxoR8AsqEfALOhHwC0pR8AtakfALadHwC3lR8AoykeAIIZAACBGQAAgLEBAL9yAICmIR4ApSkeAMNyAICrZR4AqmUeAIaIAACH/AEAr20eAK5hHgCtdR4ArHUeAMdyAICzmR4Ay3IAgM9yAIC2XQEA03IAgNdyAIC1sR4AukkBALtJAQDbcgCA33IAgL49AQC/IQEAvDkBAL01AQCoRR4AqVUeAKpVHgCrZR4ArH0eAK2ZAQCuiQEAr4EBAISsAADjcgCA53IAgOtyAIDvcgCA83IAgPdyAID7cgCAuK0BALllAQC6bQEAu2UBALx9AQC9ZQEAvm0BAL9lAQCwyQEAsckBALKpAQCzpQEAtL0BALWhAQC2oQEAt5UBALhpHAC5oRwAusEcALvBHAC8wRwAvcEcAL7BHAC/wRwAsIkfALGJHwCyIRwAswUcALQdHAC1fRwAtnUcALdtHACoYR8AqWEfAKphHwCrYR8ArNkfAK3ZHwCuyR8Ar8EfAP9yAIADcwCAB3MAgAtzAIAPcwCAE3MAgBdzAIAbcwCAH3MAgCNzAIC+AAQAo1EdACdzAICleR0AppUCACtzAIAvcwCAM3MAgKqBAgCrgQIArPECAK39AgCu9QIAr+kCADtzAIDh9AEAP3MAgON8AQCATQAAgXUAAIJ9AABDcwCAhsAEAIekBABHcwCAS3MAgE9zAIBTcwCAV3MAgO+MAgCoSQIAqUkCAKpdAgCrVQIArHkCAK15AgCuvQIAr7UCAISgBQBbcwCAX3MAgGNzAIC+vAQAZ3MAgGtzAIBvcwCAuC0BALk1AQC6PQEAuzUBALwtAQC91QEAvt0BAL/NAQCwzQIAsdUCALLdAgCz1QIAtM0CALUVAQC2HQEAtxUBAOGEHgDjbB8A41wfAOFYHgBzcwCAd3MAgHtzAIB/cwCAg3MAgIdzAICLcwCAj3MAgOkAAADv9B4A70weAJNzAICzlQIAl3MAgJtzAICfcwCAo3MAgLa5AgC1sQIAq3MAgLtRAgC6SQIAhsgEAIesBAC/kQEAvkkCAL1BAgC8SQIAN3MAgKNRBQCvcwCAp3MAgKZ9BQCzcwCAt3MAgKV1BQCqjQUAq5UFALtzAIC/cwCAro0FAK9VBgCsjQUArYUFAICJBwCBiQcAgpkHALORBgDDcwCAtbkGALapBgDHcwCAy3MAgM9zAIC6TQcAu0UHALxdBwC9QQcAvkEHAL9BBwCoQQYAqU0GAKpVBgCrZQYArH0GAK1lBgCubQYAr2UGANNzAIDXcwCA23MAgN9zAIDjcwCA53MAgOtzAIDvcwCAuFkHALlZBwC6aQcAu2kHALx5BwC9eQcAvmUHAL8ZBwCwxQcAsc0HALLFBwCz2QcAtMkHALXJBwC2aQcAt2kHAKPdBwDzcwCA93MAgPtzAID/cwCApuUHAKX1BwADdACAqwkGAKoBBgAHdACAC3QAgK8NBgCuDQYArQ0GAKwRBgCAbQAAgQkAAIIZAAAPdACAE3QAgISYAQC+kAEAF3QAgIbAAACH5AEAG3QAgB90AIAjdACAJ3QAgCt0AIAvdACAqF0GAKmNAQCqnQEAq5UBAKy5AQCtuQEArskBAK/BAQCEoAAAM3QAgDd0AIA7dACAP3QAgEN0AIBHdACAS3QAgLh5AQC5eQEAus0AALvFAAC83QAAvcUAAL7FAAC/9QAAsIEBALGBAQCySQEAs0kBALRZAQC1WQEAtkkBALdJAQCzFQIAT3QAgFN0AIBXdACAW3QAgLY5AgC1MQIAX3QAgLtFAgC6RQIAY3QAgGd0AIC/nQIAvp0CAL2dAgC8nQIAhXw+AKNRAgBrdACAb3QAgKZ9AgBzdACAd3QAgKV1AgCqAQIAqwECAHt0AIB/dACArtkCAK/ZAgCs2QIArdkCAIDpAACB6QAAggUAAIN0AIC+AAwAi3QAgIeoAwCGvAwAj3QAgJN0AICXdACAm3QAgJ90AICjdACAp3QAgKt0AICvdACAs3QAgLd0AIC7dACA42ABAL90AIDhoAEAw3QAgO+IAgDHdACAy3QAgM90AIDTdACA13QAgNt0AIDfdACAqGkCAKlpAgCqeQIAq3kCAKxpAgCtaQIArr0CAK+1AgC+rAwA43QAgOd0AIDrdACAgB0AAIEJAACCqQAA73QAgLhRAQC5WQEAumEBALthAQC8GQEAvRkBAL4NAQC/BQEAsM0CALHVAgCy3QIAs9UCALTNAgC1cQEAtnEBALdxAQDjxAAA4XwHAOF4BgDjvAYA83QAgIQYDQCGuAwAhzwNAL4sDwD7dACA/3QAgAN1AIDvEAAAB3UAgAt1AIDvdAYAD3UAgBN1AIAXdQCAs70CABt1AIC1rQIAtqUCAB91AIAjdQCAJ3UAgLpFAgC7XQIAvEUCAL1NAgC+RQIAv/kBAId0AIClfQ0ApnUNAPd0AIArdQCAL3UAgDN1AICjbQ0ArJUNAK2dDQCulQ0ArykOADd1AIA7dQCAqpUNAKuNDQCz5Q4AP3UAgEN1AIBHdQCAS3UAgLblDgC19Q4AT3UAgLuhDgC62Q4AU3UAgFd1AIC/pQ4AvrkOAL2xDgC8uQ4AqBUOAKklDgCqLQ4AqyUOAKw9DgCtJQ4Ari0OAK8lDgCADQAAgRUAAIIdAABbdQCAX3UAgGN1AICEMAMAZ3UAgLgpDgC5KQ4AujkOALs5DgC8KQ4AvSkOAL79DwC/9Q8AsF0OALElDgCyLQ4AsyUOALQ9DgC1IQ4AtiUOALcZDgCjpQ8Aa3UAgIYoAQCHTAEAb3UAgKalDwCltQ8Ac3UAgKvhDwCqmQ8Ad3UAgHt1AICv5Q8ArvkPAK3xDwCs+Q8Af3UAgLPpDgCDdQCAh3UAgLaRDgCLdQCAj3UAgLXlDgC6sQ4Au7kOAJN1AICXdQCAvmEBAL9hAQC8mQ4AvZkOAKglDgCpLQ4AqiUOAKs5DgCsKQ4ArVUOAK5dDgCvVQ4Am3UAgJ91AICjdQCAp3UAgKt1AICvdQCAs3UAgLd1AIC49QEAuYEBALqBAQC7gQEAvIEBAL2JAQC+sQEAv7EBALAxDgCxOQ4AsgkOALMJDgC04QEAteEBALbhAQC3zQEAo60NALt1AIC/dQCAw3UAgMd1AICm1Q0ApaENAMt1AICr/Q0AqvUNAM91AIDTdQCAryUCAK4lAgCt3Q0ArN0NAIBdAACBbQAAgmUAALNRAwC+nAMAtXkDALYZAwDbdQCAhOACAN91AIC6PQMAuzUDALwZAwC9GQMAvtkDAL/ZAwCohQMAqZUDAKqVAwCrpQMArL0DAK3VAwCu0QMAr9EDAIYABACHNAMAv6AzAON1AIDndQCA63UAgO91AIDzdQCAuHEDALlxAwC6cQMAu3EDALzVAAC93QAAvtUAAL/NAACwtQMAsb0DALKBAwCzgQMAtFEDALVRAwC2UQMAt1EDAO+oAwD3dQCA+3UAgP91AICEHAIAA3YAgAd2AIALdgCAviwFAA92AIATdgCAF3YAgONAAwAbdgCA4SgAAB92AICjXQIAI3YAgCd2AIArdgCAL3YAgKYVAgCldQIAM3YAgKs5AgCqMQIAN3YAgDt2AICv1QIArtUCAK0VAgCsFQIA4ygBAOEADwDhCA4A4wgOAID9AACBCQAAgjkAAD92AIBDdgCAS3YAgE92AIBTdgCA7+gOAFd2AIBbdgCA72QOALNtAQBfdgCAhugEAIcMBQBjdgCAtm0BALVtAQBndgCAu+0AALrtAABrdgCAb3YAgL/VAAC+6QAAveEAALzpAACoXQYAqWEGAKqlBgCrvQYArKUGAK2tBgCupQYArxkHAEd2AIBzdgCAd3YAgHt2AIB/dgCAg3YAgId2AICLdgCAuHUHALl5BwC6DQcAuwUHALwdBwC9BQcAvgUHAL81BwCwaQcAsWkHALJ9BwCzdQcAtG0HALVRBwC2UQcAt1EHAKMtBgCPdgCAk3YAgJd2AICbdgCApi0GAKUtBgCfdgCAq60HAKqtBwCjdgCAp3YAgK+VBwCuqQcAraEHAKypBwCADQAAgRUAAIIdAACrdgCAr3YAgLN2AICEVAMAvlwAALd2AIC7dgCAhugAAIdMAwC/dgCAw3YAgMd2AIDLdgCAz3YAgOMEBADTdgCA4bQFANd2AIDbdgCA33YAgON2AIDndgCA63YAgO92AIDzdgCA93YAgO/sBAD7dgCA/3YAgLPtBgADdwCAB3cAgAt3AIAPdwCAtpEGALXhBgATdwCAu40GALqNBgAXdwCAG3cAgL9BAQC+WQEAvVEBALxZAQCoJQYAqS0GAKolBgCrOQYArCkGAK1RBgCuSQYAr0EGAIDNAACBCQAAghkAAB93AIAjdwCAhCwBAL40AAArdwCAuP0BALlBAQC6QQEAu0EBALxBAQC9SQEAvnEBAL9xAQCwCQYAsQkGALLNAQCzxQEAtN0BALXFAQC2zQEAt8UBAIagPACHRAMAL3cAgKOhBQAzdwCApa0FAKbdBQA3dwCAO3cAgL4oPACqwQUAq8EFAKwVAgCtHQIArhUCAK8NAgC2QQMAP3cAgEN3AIC1sQIAR3cAgLOhAgBLdwCAT3cAgL5FAwC/TQMAvHUDAL1NAwC6ZQMAu20DAFN3AIBXdwCAW3cAgF93AIDXdQCAY3cAgGd3AIBrdwCAb3cAgHN3AICoRQIAqVUCAKpdAgCrVQIArE0CAK21AwCusQMAr60DALDVAwCx3QMAstUDALPtAwC09QMAtf0DALb1AwC37QMAuNkDALnZAwC6rQMAu6UDALy9AwC9pQMAvqUDAL+VAwCj9QMAd3cAgHt3AIB/dwCAg3cAgKYVAgCl5QMAh3cAgKs5AgCqMQIAi3cAgI93AICvGQIArhECAK0ZAgCsIQIAgGkAAIFpAACCBQAAk3cAgJt3AICfdwCAo3cAgO8cAACEbAIA4ZQBAKd3AIDjyAAAq3cAgK93AICGWDwAh1A9ALN3AIC3dwCAu3cAgISEPQC/dwCAw3cAgMd3AIDvuAEAvmw8AOF0BgDLdwCA42QBAM93AIDTdwCA13cAgNt3AICz0QEA33cAgON3AIDndwCA63cAgLaRAQC1+QEA73cAgLu9AQC6vQEA83cAgPd3AIC/dQEAvnUBAL2FAQC8hQEAqL09AKkNPgCqGT4AqxE+AKwxPgCtUT4ArlE+AK9NPgCXdwCAgh0AAIEdAACAHQAA+3cAgP93AIADeACAB3gAgLjVPgC53T4AutU+ALtJPwC8WT8AvVk/AL5JPwC/QT8AsDk+ALE5PgCyET4AsxE+ALTxPgC18T4AtvU+ALftPgCjkT4AC3gAgIYoAACHwAMAD3gAgKbRPgCluT4AE3gAgKv9PgCq/T4AF3gAgBt4AICvNT4ArjU+AK3FPgCsxT4AH3gAgLOdPwAjeACAJ3gAgLalPwAreACAL3gAgLWtPwC6aT8Au3U/ADN4AIA3eACAvlk/AL9FPwC8bT8AvWU/ADt4AIA/eACAQ3gAgEd4AIDjYDwAS3gAgOEAPQBPeACA7/w9AFN4AIBXeACAW3gAgF94AIBjeACAZ3gAgGt4AICjGT4AghkAAIEZAACAcQAAb3gAgKYhPgClKT4Ac3gAgKvxPgCq7T4AhCQBAL4kAQCvwT4Art0+AK3hPgCs6T4AqNE+AKnRPgCq0T4Aq+U+AKzhPgCt4T4Arhk+AK8ZPgCGAAAAh4QAAHt4AIB/eACAg3gAgId4AICLeACAj3gAgLh9PgC5AT4AugE+ALsBPgC8AT4AvQk+AL4xPgC/MT4AsGk+ALF1PgCyfT4As3U+ALRZPgC1RT4Atk0+ALdFPgCohQIAqZUCAKqVAgCrpQIArL0CAK3VAgCu0QIAr9ECAJN4AICXeACAm3gAgL8k5gGfeACAo3gAgKd4AICreACAuFUDALlZAwC6bQMAu2UDALx9AwC9ZQMAvm0DAL9lAwCwtQIAsb0CALKBAgCzgQIAtHEDALVxAwC2cQMAt3EDALMdAgCveACAs3gAgLd4AICEiAMAtlUCALU1AgAndwCAu3kCALpxAgC7eACAv3gAgL+1AwC+tQMAvVUCALxVAgDDeACAo1kCAMd4AIDLeACAphECAM94AIDTeACApXECAKo1AgCrPQIA13gAgNt4AICu8QMAr/EDAKwRAgCtEQIAqKkCAKmpAgCquQIAq7kCAKypAgCtqQIArjkBAK85AQCAzQEAgQkAAIIZAADfeACA43gAgL64BQDreACA73gAgLjpAQC56QEAuokBALuFAQC8nQEAvYEBAL6BAQC/tQEAsEkBALFVAQCyXQEAs1UBALRNAQC18QEAtvEBALfxAQDvFAAA83gAgIaoBQCH3AUA93gAgIRYBAD7eACA78Q+AP94AIDhxD4AA3kAgOMwPgDjyAAAB3kAgOEoAQALeQCAtn0CAA95AIATeQCAtXUCABd5AICzZQIAG3kAgB95AIC+3QEAv2EBALzdAQC91QEAutkBALvFAQAjeQCAJ3kAgKOxBQDneACAK3kAgC95AIAzeQCApqkFAKWhBQA3eQCAqxEGAKoNBgA7eQCAP3kAgK+1BgCuCQYArQEGAKwJBgBDeQCAR3kAgEt5AIBPeQCAgBkAAIEZAACCBQAAU3kAgL5sAwBXeQCAhsgAAIccAwBbeQCAX3kAgGN5AIBneQCAqLkHAKm5BwCqDQcAqx0HAKwJBwCtNQcArjEHAK8pBwCEqAMAa3kAgG95AIBzeQCAd3kAgHt5AIB/eQCAg3kAgLjJAAC5yQAAutkAALvRAAC8+QAAvfkAAL6ZAAC/mQAAsF0HALEhBwCyIQcAsz0HALQpBwC1KQcAtgEHALcBBwCzhQYAh3kAgIt5AICPeQCAk3kAgLa1BgC1gQYAl3kAgLvlBgC6mQYAm3kAgJ95AIC/7QYAvu0GAL3pBgC89QYAo3kAgKd5AICreQCAr3kAgLN5AIC3eQCAu3kAgO+QBAC/eQCA4dwGAMN5AIDj7AUAgCkAAIEVAACCEQAAvnwBAKMFBgDLeQCAhigAAIdMAQDPeQCApjUGAKUBBgDTeQCAq2UGAKoZBgDXeQCA23kAgK9tBgCubQYArWkGAKx1BgDfeQCAs70BAON5AIDneQCAtnkBAOt5AIDveQCAtXkBALpVAQC7XQEA83kAgPd5AIC++QAAv/kAALxFAQC9+QAAqHECAKlxAgCqcQIAq3ECAKy1AgCtvQIArrUCAK+tAgCE7AwA+3kAgP95AIADegCAB3oAgAt6AIAPegCAE3oAgLhpAwC5aQMAugkDALsJAwC8GQMAvRkDAL4JAwC/CQMAsNUCALHdAgCy1QIAs2kDALR5AwC1eQMAtmkDALdhAwAXegCAG3oAgB96AICj9QIAI3oAgKUxAgCmMQIAJ3oAgCt6AIAvegCAqh0CAKsVAgCsDQIArbEDAK6xAwCvsQMAgGEAAIFhAACCBQAAM3oAgIbwDACHYAMAvhAMADt6AIB3eACAP3oAgEN6AIBHegCAS3oAgE96AIBTegCAV3oAgKiFAgCplQIAqpUCAKulAgCsvQIArdUCAK7RAgCv0QIAW3oAgF96AIBjegCAZ3oAgGt6AIBvegCAc3oAgHd6AIC4dQEAuX0BALp1AQC7zQEAvNUBAL3dAQC+yQEAv8EBALC1AgCxvQIAsoECALOBAgC0VQEAtV0BALZVAQC3TQEA4RAGAIRIDADjDAYAe3oAgISYDAB/egCAg3oAgId6AICLegCAj3oAgJN6AICXegCAgXUAAIB1AADvIAEAgnUAAJt6AICfegCAo3oAgL7ADACFtA4A4RACAO9cAADjABYA4ZABAKt6AIDjWAEA7zwHAK96AICzegCAhgAIAIe4DACznQ0AN3oAgLd6AIC7egCAv3oAgLbVDQC1tQ0Aw3oAgLv5DQC68Q0Ax3oAgMt6AIC/GQ4AvhEOAL3VDQC81Q0Az3oAgKPZDQDTegCA13oAgKaRDQDbegCA33oAgKXxDQCqtQ0Aq70NAON6AIDnegCArlUOAK9dDgCskQ0ArZENAKhdDgCpYQ4AqmEOAKthDgCsYQ4ArWEOAK5hDgCvYQ4A63oAgO96AIDzegCA93oAgPt6AID/egCAA3sAgAd7AIC4TQ8AuVEPALpRDwC7UQ8AvHEPAL1xDwC+cQ8Av3EPALDBDwCxwQ8AssEPALPBDwC0wQ8AtcEPALbBDwC3wQ8As+kPAAt7AIC+gAEAD3sAgKd6AIC24Q8AtekPABN7AIC7BQ4AugUOABt7AIAXewCAvwUOAL4FDgC9FQ4AvBUOAIFNAACAQQAA72gNAIJRAACG8AcAh9QBAB97AIAjewCAJ3sAgIRwAQArewCAL3sAgOHgDgAzewCA40gNADd7AICjaQ8AO3sAgD97AIBDewCAR3sAgKZhDwClaQ8AS3sAgKuFDgCqhQ4AT3sAgFN7AICvhQ4AroUOAK2VDgCslQ4AV3sAgLMxDgBbewCAX3sAgLbBAQBjewCAZ3sAgLXRAQC6zQEAu6UBAGt7AIBvewCAvqUBAL+tAQC8sQEAvbEBAI/dJgCj8Q0Ac3sAgHd7AICmAQIAe3sAgH97AIClEQIAqg0CAKtlAgCDewCAviAEAK5lAgCvbQIArHECAK1xAgCfoQwAnnkKAJ1pCgCc0QgAm7E2AJp1NgCZ0TQAmOEyAJdtMgCWZTIAlTU/AJRhPgCTcT4AkjU7AJFxOgCQeToAgJUAAIGdAACCoQAAi3sAgO9EAgDhdA8Aj3sAgOMcDwDj1AEAk3sAgOHgAQDvXAEAo7UCAKJBAACh3Q4AoLkOALWpAwCXewCAhMAEALahAwCG8AUAh+QEALOFAwCbewCAvXEDALxpAwC/QQMAvnEDAJ97AIDHeQCAu3EDALp5AwCC3ScAgwE7AL6EBwC+wAYAhhE/AIcZPwCEETsAhV06AIp9PgCLJTMAo3sAgKd7AICOuTUAjxU3AIw1MwCNgTMAkqE3AJPZCQC+xBkAq3sAgJaxDQCXUQ8AlHkLAJVhCwCaBQ8Am5EBAK97AICzewCAt3sAgN0AAACcfQMAu3sAgOFIDwC/ewCA4xwOAMN7AIDHewCAy3sAgM97AIDTewCAsUEXALChFwCzqesBsgHoAbUB7AG0EesB74wOANd7AICpxR8AqAEcAKsBEACqkR8ArdkTAKzREwCv2RcArgUTAKHxAgDbewCAo8kHAKLBAgClARgApGUHAKehGwCm+RsAqCkFAKldBQCqVQUAq20FAKx5BQCteQUArm0FAK9hBQCHewCA33sAgON7AIDnewCAgA0AAIGxAACCsQAA63sAgLiJBQC5iQUAup0FALuVBQC8uQUAvbkFAL5RBgC/UQYAsOUFALHtBQCy5QUAs/0FALTtBQC13QUAttUFALe9BQCj3QUA73sAgPN7AICEDAAA93sAgKb5BQCl8QUA+3sAgKspBQCqIQUAhpgAAIegAACvGQUArikFAK0pBQCsMQUA/3sAgLNhBgADfACAB3wAgLYhBgALfACAD3wAgLUBBgC6rQcAu40HABN8AIAXfACAvo0HAL9xBwC8lQcAvY0HAL65BQC/uQUAvLkFAL25BQC6uQUAu7kFALi5BQC5uQUAtkkFALdJBQC0fQUAtXUFALJ5BQCzeQUAsBUFALF9BQCuXQUAr20FAKxFBQCtXQUAqqUKAKtdBQCovQoAqa0KABt8AIAffACAI3wAgCd8AIArfACAL3wAgDN8AIA3fACAqA0HAKkdBwCqLQcAq0kHAKxNBwCtZQcArrEGAK+xBgA7fACAP3wAgEN8AIBHfACAS3wAgE98AIBTfACAV3wAgLhVBgC5XQYAulUGALtxBgC8NQYAvfEBAL7xAQC/8QEAsK0GALGNBgCyhQYAs50GALSNBgC1cQYAtnUGALdtBgCjpQQAgi0AAIEVAACAHQAAW3wAgKblBAClxQQAX3wAgKtJBQCqaQUAY3wAgGt8AICvtQUArkkFAK1JBQCsUQUAhmAcAIcIAwBvfACAs4UCAHN8AIC1gQIAtoECAHd8AIB7fACAf3wAgLoJAwC7CQMAvBkDAL0ZAwC+CQMAvwkDAKxVAgCtXQIArmECAK9hAgCoDQIAqVUCAKpRAgCrUQIAhKwDAIN8AICHfACAi3wAgIT8HQCPfACAk3wAgJd8AIC8cQMAvXEDAL5xAwC/cQMAuHEDALlxAwC6cQMAu3EDALSRAwC1kQMAtpEDALeRAwCwkQMAsZEDALKRAwCzkQMAm3wAgJ98AICjfACAp3wAgKt8AIDhpAEAr3wAgOOAAQC+aBwAs3wAgLd8AIDv2AYAu3wAgL98AIDDfACAx3wAgKOJAwCCLQAAgRUAAIAdAADLfACApo0DAKWNAwDPfACAqwUCAKoFAgDTfACA23wAgK8FAgCuBQIArRUCAKwVAgCGIBwAh8QdAN98AIDjfACA53wAgOt8AIDvfACA72wGAPN8AIDhbAcA93wAgON0BwD7fACA/3wAgAN9AIAHfQCAs5EBAAt9AIAPfQCAE30AgBd9AIC2sQEAtbkBABt9AIC7VQEAukkBAB99AIAjfQCAv/UAAL71AAC9RQEAvEUBAKNRHgDXfACAJ30AgCt9AIAvfQCApnEeAKV5HgAzfQCAq5UeAKqJHgA3fQCAO30AgK81HwCuNR8ArYUeAKyFHgCAbQAAgRUAAIIdAADv/BkAP30AgEN9AIBHfQCAS30AgIbAAACHrAMAT30AgFN9AIBXfQCA4SwcAFt9AIDjzBwAqK0eAKnNHgCq2R4Aq9EeAKzxHgCt8R4Arj0eAK81HgCE7AAAX30AgGN9AIBnfQCAa30AgG99AIBzfQCAd30AgLjRHwC53R8Auu0fALvlHwC84R8AveEfAL7hHwC/4R8AsE0eALFRHgCyUR4As1EeALTxHwC18R8AtvEfALfxHwCobR4AqY0eAKqFHgCrnR4ArIUeAK2NHgCuuR4Ar7UeAHt9AIB/fQCAg30AgId9AICLfQCAj30AgJN9AICXfQCAuJ0eALmtHgC6pR4Au0UBALxdAQC9RQEAvkUBAL91AQCw0R4AsdEeALLRHgCz0R4AtLUeALW9HgC2tR4At60eALMNHgCbfQCAn30AgKN9AICnfQCAtg0eALUNHgCrfQCAuxUeALoVHgCvfQCAs30AgL95HgC+cR4AvQUeALwFHgCCbQAAo0keAIBVAACBZQAApkkeAL6cAQC7fQCApUkeAKpRHgCrUR4Ah3wAAIZMAACuNR4Arz0eAKxBHgCtQR4AqF0CAKltAgCqZQIAq30CAKxpAgCtsQIArrECAK+xAgCE7AQAv30AgMN9AIDHfQCAy30AgM99AIDTfQCA130AgLhxAwC5cQMAunEDALtxAwC81QMAvd0DAL7VAwC/zQMAsNECALHRAgCy0QIAs9ECALRRAwC1UQMAtlEDALdRAwCz7QIA230AgN99AIC+gAQA430AgLYxAgC14QIA530AgLsVAgC6FQIA630AgO99AIC/lQMAvpUDAL0FAgC8BQIA830AgKOpAgD3fQCA+30AgKZ1AgD/fQCAA34AgKWlAgCqUQIAq1ECAAd+AIALfgCArtEDAK/RAwCsQQIArUECAKjZAgCpIQEAqiEBAKshAQCsIQEArSEBAK4hAQCvIQEAD34AgBN+AIAXfgCAviAEABt+AIAffgCAI34AgCt+AIC4jQEAuZEBALqRAQC7pQEAvL0BAL11AAC+fQAAv3UAALDlAQCx7QEAsvkBALPxAQC02QEAtdkBALa5AQC3tQEA4RgeAC9+AIDjKB8AM34AgIGlAACApQAAN34AgIKlAACGAAQAh/QFADt+AIA/fgCAQ34AgEd+AIDvYB4AS34AgE9+AIBTfgCAhfD0AVd+AIBbfgCA42QBAF9+AIDhpAEAY34AgO/IAABnfgCAa34AgGd8AICE/AUAb34AgHN+AICzKQYAJ34AgHd+AIB7fgCAf34AgLYhBgC1KQYAg34AgLupBgC6oQYAh34AgIt+AIC/nQYAvp0GAL2lBgC8rQYA4bQHAI9+AIDjeAQAk34AgIB9AACBEQAAghUAAJd+AICGwAAAh1gDAJt+AICffgCAo34AgKd+AIDvDAQAq34AgKOpBgCvfgCAs34AgLd+AIC7fgCApqEGAKWpBgC/fgCAqykGAKohBgDDfgCAx34AgK8dBgCuHQYArSUGAKwtBgDLfgCAs0kHAM9+AIDTfgCAtn0HANd+AIDbfgCAtXUHALpdBwC7JQcA334AgON+AIC+IQcAvy0HALw9BwC9MQcAqD0GAKmBBgCqhQYAq5UGAKy5BgCtuQYArqkGAK+pBgDnfgCA634AgO9+AIDzfgCA934AgIK5AACBsQAAgLkAALitBgC5vQYAurUGALtFAQC8XQEAvUUBAL5FAQC/dQEAsN0GALGlBgCyrQYAs6EGALShBgC1rQYAtpkGALeVBgCjDQYA+34AgP9+AIADfwCAhJgCAKY5BgClMQYAvpwBAKthBgCqGQYAhggAAId8AQCvaQYArmUGAK11BgCseQYAC38AgLO1AQAPfwCAE38AgLZVAQAXfwCAG38AgLWhAQC6cQEAu3kBAB9/AIAjfwCAvjEBAL89AQC8UQEAvVEBAKhpAgCpaQIAqnkCAKt5AgCsbQIArZECAK6RAgCvkQIAJ38AgCt/AIAvfwCAM38AgDd/AIA7fwCAP38AgEN/AIC4mQIAua0CALqlAgC7bQMAvHUDAL19AwC+dQMAv20DALDxAgCx+QIAssECALPBAgC0sQIAtb0CALa1AgC3qQIAR38AgEt/AIBPfwCAo/0CAFN/AICl6QIAph0CAFd/AIBbfwCAX38AgKo5AgCrMQIArBkCAK0ZAgCueQIAr3UCAGN/AIBnfwCAa38AgIQADACAGQAAgQkAAII5AABvfwCAc38AgHt/AIB/fwCAvuAMAIN/AICHfwCAhlgNAIcMAwCowQIAqc0CAKrFAgCr2QIArMkCAK39AgCu9QIArz0BAIt/AICPfwCAk38AgJd/AICbfwCAn38AgKN/AIC+MAwAuMUBALnNAQC62QEAu9EBALzxAQC98QEAvpkBAL+ZAQCwRQEAsU0BALJFAQCzXQEAtEUBALVNAQC2RQEAt/0BAOE4BgCnfwCA42wGAKt/AICvfwCAs38AgLd/AIC7fwCAhKgNAL9/AIDDfwCAx38AgL6wDwDLfwCA72wGAM9/AIDTfwCAt30AgNd/AIDbfwCA41AAAN9/AIDhoAEA438AgO+EAADrfwCAhyANAIZMDwCAPQAAgSEAAIIlAADvfwCAs80NAHd/AIDnfwCA838AgPd/AIC2/Q0AtcENAPt/AIC7CQ4AugEOAP9/AIADgACAvwkOAL4BDgC9CQ4AvBEOAAeAAIDjmAwAC4AAgOH8DwAPgACAE4AAgBeAAIAbgACAH4AAgCOAAIAngACAK4AAgC+AAIDvYAwAM4AAgDeAAICjTQ0AO4AAgD+AAIBDgACAR4AAgKZ9DQClQQ0AS4AAgKuJDgCqgQ4AT4AAgFOAAICviQ4AroEOAK2JDgCskQ4Agm0AALM1DgCAVQAAgWUAALb1DwCE3AMAV4AAgLX9DwC60Q8Au9EPAIYABACH3AAAvn0PAL9lDwC8wQ8AvXkPAKjlDwCp7Q8AqvkPAKv5DwCsMQ4ArTEOAK4xDgCvMQ4AW4AAgF+AAIBjgACAZ4AAgGuAAIBvgACAc4AAgHeAAIC43Q4AueEOALrhDgC74Q4AvOUOAL3pDgC+mQ4Av5UOALBRDgCxUQ4AslEOALPpDgC0/Q4AteUOALbtDgC35Q4Ao3EPAHuAAIB/gACAg4AAgIeAAICmsQ4ApbkOAIuAAICrlQ4AqpUOAI+AAICTgACAryEOAK45DgCtPQ4ArIUOAJeAAICzyQEAm4AAgJ+AAIC2+QEAo4AAgKeAAIC1wQEAuqkBALu1AQCrgACAr4AAgL6tAQC/lQEAvK0BAL2lAQCo5Q0AqfkNAKoFAgCrHQIArA0CAK09AgCuNQIAr10CALOAAIC3gACAu4AAgL+AAICAGQAAgRkAAIIFAADDgACAuC0CALk1AgC6MQIAuzECALzVAgC93QIAvtUCAL/NAgCwKQIAsTUCALI9AgCzNQIAtC0CALUVAgC2HQIAtxUCAMuAAICEnAIAz4AAgKOBAgDTgACApYkCAKaxAgDXgACAhiAEAIfUAwCq4QIAq/0CAKzlAgCt7QIAruUCAK/dAgC29QMAvkQDAIWM/QG1/QMA24AAgLP9AwDfgACA44AAgL59AwC/TQMAvGUDAL19AwC6dQMAu30DAOeAAIDrgACA74AAgPOAAICEBAIAoyUCAPeAAIClJQIApi0CAPuAAID/gACAA4EAgKqtAgCrpQIArL0CAK2lAgCupQIAr5UCAAeBAIALgQCAD4EAgBOBAIAXgQCA48ADABuBAIDhrAEAH4EAgO9YAwAjgQCAJ4EAgIANAACB5QAAgu0AACuBAIDhYA8A40ABAOM4DgDheA4AL4EAgDOBAIC+lAUAO4EAgIYABACHZAUAP4EAgEOBAIBHgQCA7/wOAO98DgBLgQCAs1EBAE+BAIAHfwCAU4EAgFeBAIC2DQEAtQkBAFuBAIC74QAAuhkBAF+BAIBjgQCAv9EAAL7pAAC96QAAvPkAAMeAAIA3gQCAZ4EAgGuBAIBvgQCAc4EAgHeBAIB7gQCAqKEGAKmtBgCquQYAq7EGAKzhBgCt7QYAruUGAK/FBgCwvQYAsUUHALJNBwCzXQcAtE0HALV1BwC2fQcAtx0HALglBwC5LQcAuiUHALs9BwC8KQcAvRUHAL4RBwC/EQcAoxEGAH+BAICDgQCAh4EAgIuBAICmTQYApUkGAI+BAICroQcAqlkGAJOBAICXgQCAr5EHAK6pBwCtqQcArLkHAIANAACBFQAAgh0AAJuBAICfgQCAo4EAgISUAwC+lAMAp4EAgKuBAICGyAAAh4wAAK+BAICzgQCAt4EAgLuBAIConQYAqa0GAKqlBgCrvQYArK0GAK3RBgCu1QYAr80GAL+BAIDDgQCAx4EAgMuBAIDPgQCA04EAgNeBAIDbgQCAuF0BALnBAQC6wQEAu8EBALzBAQC9yQEAvvEBAL/xAQCwvQYAsY0GALKFBgCzZQEAtH0BALVlAQC2bQEAt2UBALMtBgDfgQCA44EAgOeBAIDrgQCAtlEGALUlBgDvgQCAu0kGALp5BgDzgQCA94EAgL+hAQC+uQEAvbEBALxRBgD7gQCAo2kGAP+BAIADggCAphUGAAeCAIALggCApWEGAKo9BgCrDQYAD4IAgBOCAICu/QEAr+UBAKwVBgCt9QEAutUHALvdBwC4wQcAucEHAL4xBAC/MQQAvPEHAL3xBwCyrQcAs7UHALCtBwCxpQcAtp0HALf1BwC0pQcAtZUHAKppBwCraQcAqGkHAKlpBwCuaQcAr2kHAKxpBwCtaQcAgLkDAIGNAwCChQMAhKgDAIZQ/AGHCAMAvjQDABuCAICoZQIAqXUCAKp9AgCrdQIArG0CAK21AwCuvQMAr7UDAB+CAIAjggCAJ4IAgCuCAIAvggCAM4IAgDeCAIA7ggCAuFEDALlZAwC6YQMAu2EDALwRAwC9HQMAvhUDAL8JAwCwzQMAsdUDALLdAwCz1QMAtM0DALVxAwC2cQMAt3EDAD+CAIBDggCAs/0DAEeCAIC17QMAS4IAgE+CAIC2PQIAU4IAgFeCAIC7GQIAugECAL0JAgC8AQIAv70CAL4BAgBbggCAX4IAgITE/QG+wPwBY4IAgGeCAIBrggCA79wDAG+CAIDhlAEAc4IAgOMQAwB3ggCAgu0AAIHtAACA7QAA4TgGAOE8BwDjQAEA45QGAHuCAIB/ggCAg4IAgIuCAICGgPwBh+j9AY+CAICTggCAl4IAgJuCAIDvnAEA79wGAKM1AwCfggCAo4IAgKeCAICrggCApvUCAKUlAwCvggCAq9ECAKrJAgCzggCAt4IAgK91AgCuyQIArcECAKzJAgCHggCAu4IAgL+CAIDDggCA76T9AceCAIDLggCAz4IAgON4/QHTggCA4UD8AdeCAIDbggCA34IAgOOCAIDnggCAs+X+AYItAACBFQAAgB0AAOuCAIC25f4BtfX+Ae+CAIC7Yf8Butn+AfOCAICE5AMAv2n/Ab5h/wG9df8BvHn/Aaj9/gGpJf4Bqi3+Aasl/gGsPf4BrSX+Aa4t/gGvJf4BviwAAPeCAICGiAAAh+wAAPuCAID/ggCAA4MAgAeDAIC4gf8BuYH/AbqZ/wG7mf8BvIn/Ab21/wG+sf8Bv63/AbBd/gGx5f8Bsu3/AbPh/wG05f8Bte3/AbbZ/wG32f8Bo6X/AQuDAIAPgwCAE4MAgBeDAICmpf8BpbX/ARuDAICrIf4Bqpn/AR+DAIAjgwCAryn+Aa4h/gGtNf4BrDn+ASeDAICz6f4BK4MAgC+DAIC2lf4BM4MAgDeDAIC16f4BurH+Abu5/gE7gwCAP4MAgL51AQC/fQEAvJH+Ab2R/gGoHf4BqS3+Aaol/gGrPf4BrCX+Aa1R/gGuUf4Br1H+AUODAIBHgwCAS4MAgE+DAIBTgwCAV4MAgFuDAIBfgwCAuNkBALnZAQC67QEAu+EBALzhAQC94QEAvuEBAL/hAQCwMf4BsTn+AbIB/gGzAf4BtPUBALX9AQC29QEAt+kBAKOt/QFjgwCAvkwDAGuDAIBvgwCAptH9AaWt/QFzgwCAq/39Aar1/QF3gwCAe4MAgK85AgCuMQIArdX9AazV/QGA+QMAgfkDAIJNAACFdCAAf4MAgITYAwCE1AQAg4MAgIZABACHVAMAh4MAgIuDAICPgwCAk4MAgJeDAIC+8AUAqDECAKkxAgCqMQIAqzECAKyVAwCtnQMArpUDAK+NAwCbgwCAn4MAgKODAICngwCAhHwHAKuDAICvgwCAs4MAgLipAwC5qQMAumkDALtpAwC8eQMAvXkDAL5pAwC/aQMAsP0DALHNAwCyxQMAs60DALS5AwC1uQMAtq0DALelAwC3gwCAu4MAgL+DAIDDgwCAx4MAgMuDAIDv6AMAz4MAgOGQAQDTgwCA42wDANuDAICAJQAAgSkAAIIdAADfgwCAs/kDAOODAICGaAcAh1wFAOeDAIC2XQIAtV0CAOuDAIC7SQIAunkCAO+DAIDzgwCAvz0CAL49AgC9OQIAvFECAPeDAIDhPP4BvkAGAOPwAQD7gwCA/4MAgAOEAIAHhACAC4QAgA+EAIAThACAF4IAgBeEAIAbhACAH4QAgO/kAQAjhACAJ4QAgKNxAwArhACApdUCAC+EAIAzhACAptUCADeEAIA7hACAq8ECAKrxAgCtsQIArNkCAK+1AgCutQIA4dz8AdeDAIDjUAQA74gEAID1BwCBCQAAgj0AAD+EAICEJAEAQ4QAgEeEAIBLhACAT4QAgOFMBADv5BwA43QEALNdBgBThACAhgAMAIfgAwBXhACAtgUGALV1BgBbhACAuxEGALoJBgBfhACAY4QAgL/VBgC+1QYAvQEGALwJBgCojQYAqZUGAKqVBgCrpQYArL0GAK3FBgCuxQYAr/UGAGeEAIBrhACAb4QAgHOEAIB3hACAe4QAgH+EAICDhACAuHUGALl9BgC6dQYAu80HALzVBwC93QcAvtUHAL/NBwCwjQYAsZUGALKdBgCzlQYAtFEGALVRBgC2UQYAt1EGAKMdBwCPFewBh4QAgIuEAICPhACApkUHAKU1BwCThACAq1EHAKpJBwCXhACAm4QAgK+VBwCulQcArUEHAKxJBwCeRfkBn6X5AZyR/QGdTfkBmlX9AZtd/QGYBfEBmZX+AZal8gGXYfEBlG31AZU19QGS4ekBk4X2AZBV7AGRXekBsbEdALClHQCziRkAskEcALUBJAC09RkAn4QAgKOEAICnhACAgqkDAIGhAwCAaQAAohUFAKMFAgCgFQYAob0FAKHFAQCrhACAo80NAKLlAQClAQgApN0NAKfRCQCm2QkAqQEUAKilCACrxRQAqs0VAK3REQCsARAArwEcAK51EQCCEe8BgynvAa+EAICzhACAhuH1AYcR9gGEOeoBhY3qAYp59gGL4fEBvqQMALuEAICO+f0BjzH+AYw98gGNYfIBkkn+AZOd/gGHCAwAhmwMAJax+gGX+QUAlFn6AZVZ+gGaYQYAm8EGAL+EAIDDhACAx4QAgMuEAICcyQEAz4QAgKitBQCpuQUAqs0FAKvdBQCszQUArf0FAK71BQCvHQUA04QAgNeEAIDbhACA34QAgOOEAIDnhACA64QAgO+EAIC4dQUAuX0FALoJBQC7CQUAvB0FAL0BBQC+AQUAvz0FALBxBQCxcQUAsnEFALNxBQC0UQUAtVEFALZRBQC3TQUAs0UEAPOEAID3hACA+4QAgP+EAIC2fQQAtUUEAAOFAIC7tQQAurUEAAeFAIALhQCAv5UEAL6VBAC9pQQAvKUEAA+FAICjAQQAE4UAgBeFAICmOQQAG4UAgB+FAIClAQQAqvEEAKvxBAAjhQCAhOwNAK7RBACv0QQArOEEAK3hBADh0AYAhAwMAOMoBwC+AAwAK4UAgO9EAwCGuAwAhywNAC+FAIDjlAEAM4UAgOH8AQBngwCAN4UAgO/IBgA7hQCAP4UAgEOFAICzjQMAR4UAgLWNAwBLhQCAT4UAgLa1AwBThQCAV4UAgLtBAwC6SQMAvUEDALxZAwC/QQMAvkkDAKNFDAC3hACAJ4UAgFuFAIBfhQCApn0MAKVFDABjhQCAq4kMAKqBDABnhQCAa4UAgK+JDACugQwArYkMAKyRDACAFQ8AgR0PAIIhDwCzIQ4Ab4UAgLUhDgC2JQ4Ac4UAgHeFAIB7hQCAusEOALvBDgC8wQ4AvcEOAL7BDgC/wQ4AqK0OAKntDgCq5Q4Aq/0OAKzlDgCt6Q4ArjkOAK85DgB/hQCAg4UAgIeFAICLhQCAgB0AAIEJAACCvQEAj4UAgLjNDwC51Q8AutUPALvlDwC8/Q8AvZUPAL6RDwC/kQ8AsEkOALFJDgCyWQ4As1kOALRJDgC1SQ4Atv0PALf1DwCjbQ8Ak4UAgL6EAQCbhQCAn4UAgKZpDwClbQ8Ao4UAgKuNDwCqjQ8AhogAAIdsAQCvjQ8Aro0PAK2NDwCsjQ8Ap4UAgLPtDgCrhQCAr4UAgLaRDgCzhQCAt4UAgLXhDgC6tQ4Au70OALuFAIC/hQCAvn0BAL9lAQC8mQ4AvZkOAKgRDgCpJQ4AqiEOAKs5DgCsLQ4ArVUOAK5dDgCvUQ4AhKgAAMOFAIDHhQCAy4UAgM+FAIDThQCA14UAgNuFAIC47QEAuZUBALqVAQC7rQEAvLUBAL11AQC+fQEAv3UBALA1DgCxPQ4AsgkOALMJDgC0/QEAteUBALblAQC31QEAo6kNAN+FAIDjhQCA54UAgOuFAICm1Q0ApaUNAO+FAICr+Q0AqvENAPOFAID3hQCAryECAK45AgCt3Q0ArN0NAIANAACBFQAAgh0AAPuFAID/hQCAA4YAgIeQAwCGfAQAvuwEAAuGAIAPhgCAE4YAgBeGAIAbhgCAH4YAgCOGAICyLQ4AszUOALAtDgCxJQ4Ati0OALedDwC0LQ4AtSUOALq9DwC7jQ8AuKUPALm9DwC+LQ8AvxUPALyVDwC9JQ8AJ4YAgCuGAIAvhgCAM4YAgDeGAIA7hgCAP4YAgEOGAICqpQ4Aq7UOAKjFDgCp3Q4Arp0OAK9VDgCspQ4ArZUOAKgNAgCpFQIAqhUCAKtNAgCsWQIArVkCAK5NAgCvRQIAhKgFAEeGAIBLhgCAT4YAgIS4BABThgCAV4YAgFuGAIC4/QIAuUEBALpBAQC7QQEAvEEBAL1JAQC+cQEAv3EBALAJAgCxCQIAss0CALPFAgC03QIAtcUCALbNAgC3xQIA4dQPAOMQDgDj9A4A4QwOAF+GAIBjhgCAZ4YAgGuGAIBvhgCAc4YAgL4kBAB7hgCA7AAAAO9EAADvzA4Af4YAgIJlAACz2QIAgFUAAIFtAAC2nQIAg4YAgIeGAIC1lQIAuokCALuJAgCGqAQAh+AEAL5dAgC/RQIAvF0CAL1VAgCjHQUAB4YAgHeGAICLhgCAj4YAgKZZBQClUQUAk4YAgKtNBQCqTQUAl4YAgJuGAICvgQUArpkFAK2RBQCsmQUAn4YAgLMpBgCjhgCAp4YAgLYpBgCrhgCAr4YAgLUpBgC6pQYAu60GALOGAIC3hgCAvqUGAL+tBgC8tQYAva0GAKjlBgCp7QYAquUGAKv9BgCs5QYAre0GAK7lBgCvXQYAu4YAgL+GAIDDhgCAx4YAgMuGAIDPhgCA04YAgNeGAIC46QcAuekHALr9BwC79QcAvO0HAL1FBwC+TQcAv0UHALAlBgCxLQYAsiUGALM9BgC0JQYAtS0GALYlBgC32QcAo20HAIItAACBFQAAgB0AANuGAICmbQcApW0HAN+GAICr6QcAquEHAOOGAIC+oAEAr+kHAK7hBwCt6QcArPEHAOeGAICzkQYAhugAAIcsAQC2QQEA64YAgO+GAIC1UQEAuk0BALslAQDzhgCA94YAgL4lAQC/LQEAvDEBAL0xAQCwrQEAscUBALLBAQCzwQEAtMUBALXNAQC28QEAt/EBALgBAQC5AQEAugEBALsBAQC8AQEAvQEBAL4BAQC/AQEA+4YAgP+GAIADhwCAB4cAgJeFAIALhwCAD4cAgBOHAICoTQYAqVkGAKo9BgCrNQYArP0BAK3lAQCu5QEAr9UBAKPVBQAXhwCAG4cAgB+HAIAjhwCApgUCAKUVAgAnhwCAq2ECAKoJAgArhwCAL4cAgK9pAgCuYQIArXUCAKx1AgAzhwCAN4cAgDuHAIA/hwCAQ4cAgOFkBQBHhwCA4+wFAIARAACBEQAAghEAAO/0BgBLhwCAT4cAgFOHAIC+MAMAhMQCAFuHAICz4QMAhMAcALVRAwBfhwCAY4cAgLZZAwBnhwCAa4cAgLtxAwC6eQMAvbUAALxpAwC/tQAAvrUAAG+HAIDhlAEAc4cAgONcAgCGcBwAh0QDAHeHAIB7hwCAf4cAgIOHAICHhwCAi4cAgI+HAICThwCAl4cAgO94AgCoVQIAqV0CAKphAgCrYQIArNECAK3RAgCu0QIAr9ECAJuHAICfhwCAo4cAgKeHAICrhwCAr4cAgLOHAIC3hwCAuGkBALlpAQC6CQEAuwkBALwZAQC9GQEAvgkBAL8FAQCwtQIAsb0CALK1AgCzaQEAtHkBALV5AQC2aQEAt2EBAOHEBwDjpAYA47gGAOF8BgCADQAAgTUAAII9AAC7hwCAv4cAgMOHAIC+4B0Ay4cAgM+HAIDvYAAA7+gGANOHAICjqQIA14cAgNuHAIDfhwCA44cAgKYRAgClGQIA54cAgKs5AgCqMQIAhkgcAIfMHACv/QEArv0BAK39AQCsIQIAqIUeAKmRHgCqkR4Aq60eAKy1HgCt1R4ArtEeAK/FHgDHhwCA64cAgO+HAIDzhwCA94cAgPuHAID/hwCAA4gAgLhhHwC5YR8AumEfALthHwC8YR8AvWEfAL5hHwC/YR8AsL0eALGFHgCyjR4As4UeALSdHgC1hR4Ato0eALeFHgCzGR4AB4gAgAuIAIAPiACAE4gAgLZVHgC1PR4AF4gAgLtBHgC6eR4AG4gAgB+IAIC/QR4AvlkeAL1RHgC8WR4AI4gAgKNdHgAniACAK4gAgKYRHgAviACAM4gAgKV5HgCqPR4AqwUeAISkAwC+qAMArh0eAK8FHgCsHR4ArRUeAKitHgCptR4AqrUeAKvJHgCs2R4ArdkeAK7JHgCvwR4AgO0BAIHxAQCC8QEAN4gAgIaQAACHdAEAO4gAgD+IAIC4yQEAuckBALrZAQC70QEAvPkBAL35AQC+mQEAv5UBALBFAQCxTQEAskUBALNdAQC0RQEAtU0BALZFAQC3+QEAsz0eAEOIAIBHiACAS4gAgE+IAIC2WR4AtVEeAFOIAIC7iQEAuoEBAFeIAIBbiACAv4kBAL6BAQC9iQEAvJEBAF+IAIBjiACAo3UeAGeIAIClGR4Aa4gAgG+IAICmER4AV4cAgHOIAICrwQEAqskBAK3BAQCs2QEAr8EBAK7JAQB3iACAe4gAgH+IAICDiACAh4gAgIQYAgCLiACAj4gAgJOIAICXiACAm4gAgJ+IAICjiACAq4gAgK+IAIC+cAMAgGkAAIFpAACCeQAAhAAEAIbwBACHdAMAs4gAgO8MHwC3iACA4aweALuIAIDj8B4Av4gAgMOIAIDHiACAy4gAgM+IAIDTiACA14gAgNuIAIDvVAIA34gAgOOIAIDniACA46QCAOuIAIDhgAEA74gAgPOIAID3iACA+4gAgP+IAICzRQMAA4kAgAeJAIALiQCAD4kAgLZFAwC1VQMAE4kAgLshAwC6SQMAvqAEABuJAIC/KQMAviEDAL01AwC8OQMAqDkCAKk5AgCqjQIAq4UCAKydAgCthQIAroUCAK+1AgCA7QEAgfUBAIL1AQAfiQCAhpAEAIcEBQAjiQCAJ4kAgLhFAQC5TQEAukUBALtdAQC8SQEAvUkBAL55AQC/eQEAsM0CALGlAgCyrQIAs6ECALSlAgC1rQIAtp0CALd9AQAriQCAL4kAgDOJAIA3iQCAO4kAgD+JAIBDiQCA74gBAITsBADhVB4AR4kAgONUAQBLiQCAT4kAgFOJAIBXiQCAo0UCAFuJAIBfiQCAY4kAgGeJAICmRQIApVUCAGuJAICrIQIAqkkCAG+JAIBziQCArykCAK4hAgCtNQIArDkCAKg1BgCpPQYAqlEGAKttBgCseQYArWUGAK5tBgCvZQYAF4kAgHeJAIB7iQCAf4kAgIAZAACBGQAAggUAAIOJAIC45QYAuekGALr5BgC7+QYAvOkGAL3pBgC+nQYAv5UGALAdBgCx5QYAsu0GALPlBgC0/QYAteEGALbhBgC34QYAs9kGAL7QAwCHiQCAi4kAgI+JAIC25QYAtfEGAJOJAIC7IQYAutkGAIaYAACHeAMAvyUGAL45BgC9MQYAvDkGAJeJAICjnQYAm4kAgJ+JAICmoQYAo4kAgKeJAICltQYAqp0GAKtlBgCriQCAr4kAgK59BgCvYQYArH0GAK11BgCo7QcAqSkGAKoxBgCrMQYArJEGAK2RBgCukQYAr5EGALOJAIC3iQCAu4kAgL+JAIDDiQCAx4kAgMuJAIDPiQCAuIUGALmNBgC6hQYAu50GALyNBgC9vQYAvrUGAL95AQCw8QYAsfEGALLxBgCzxQYAtMEGALXBBgC2wQYAt8EGALO5BgDTiQCA14kAgNuJAIDfiQCAthEGALUZBgDjiQCAuzUGALo1BgDniQCA64kAgL8FBgC+BQYAvREGALwlBgClQQYA74kAgPOJAICmSQYAgRUAAIB5AACj4QYAghUAAK1JBgCsfQYAr10GAK5dBgCENAEAp4gAgKttBgCqbQYAvswDAPuJAICzlQIA/4kAgLXZAgADigCAB4oAgLbRAgCGgAwAhzgDALvFAgC6xQIAvRUDALwVAwC/FQMAvhUDAAuKAIAPigCA71gGAIRAAwATigCAF4oAgBuKAIAfigCAI4oAgCeKAIArigCAL4oAgOE4BgAzigCA4yQGAL5wDACsSQIArUkCAK5dAgCvVQIAqB0CAKkFAgCqBQIAq10CAISoDAA3igCAO4oAgD+KAIC+vA0AQ4oAgEeKAIBLigCAvE0DAL1VAwC+VQMAv2UDALjpAwC56QMAul0DALtVAwC0yQMAtckDALbZAwC32QMAsBkCALEZAgCy2QMAs9kDAE+KAIDj5AAAU4oAgOG8AQBXigCAgj0AAIE9AACAPQAAW4oAgF+KAIBjigCAa4oAgG+KAIDvzAMAc4oAgHeKAICj3QMAe4oAgIboDACHYA0Af4oAgKaZAwClkQMAg4oAgKuNAwCqjQMAh4oAgIuKAICvXQIArl0CAK1dAgCsXQIAj4oAgJOKAICXigCAm4oAgJ+KAICjigCAp4oAgO/gAQCEvAwA4YwGAKuKAIDjHAYAr4oAgLOKAIC3igCAu4oAgLPVAQC/igCAw4oAgMeKAIDLigCAtpEBALWZAQDPigCAu70BALq9AQDTigCA24oAgL+dAQC+nQEAvZ0BALydAQCoBQ4AqQkOAKodDgCrFQ4ArFEOAK1RDgCuSQ4Ar0kOAGeKAICCzQ8AgfUPAID9DwDXigCA34oAgIYcAACHsAMAuOkOALnpDgC6/Q4Au/UOALztDgC9VQ8AvlEPAL9NDwCwOQ4AsTkOALIJDgCzCQ4AtBkOALUZDgC2DQ4At9kOAKOVDgDjigCA54oAgOuKAIDvigCAptEOAKXZDgDzigCAq/0OAKr9DgD3igCA+4oAgK/dDgCu3Q4Ard0OAKzdDgD/igCAs/0PAAOLAIAHiwCAtoEPAAuLAIAPiwCAtZkPALqNDwC7ZQ8AE4sAgBeLAIC+fQ8Av2UPALx9DwC9dQ8AqC0OAKk1DgCqMQ4AqzEOAKxVDgCtRQ4ArkUOAK91DgAbiwCAH4sAgCOLAIAniwCAK4sAgC+LAIAziwCAN4sAgLjpDgC59Q4Auv0OALv1DgC87Q4AvZEOAL6RDgC/kQ4AsA0OALHlDgCy7Q4As+UOALT9DgC15Q4Atu0OALflDgCjuQ4Agi0AAIEVAACAHQAAO4sAgKbFDgCl3Q4AP4sAgKshDgCqyQ4AQ4sAgL4sAQCvIQ4ArjkOAK0xDgCsOQ4AS4sAgLZVAQC1RQEAR4sAgLNVAQBPiwCAhngAAIdcAAC/OQEAvjEBAL0lAQC8JQEAuzEBALpZAQD3iQCAU4sAgFeLAIBbiwCAhAQDAKOJAgBfiwCApZkCAKaJAgBjiwCAvyg5AGeLAICqhQIAq+0CAKz5AgCt+QIAru0CAK/lAgDjWAIA78AOAOGIAQBriwCAb4sAgHOLAIB3iwCAe4sAgH+LAICDiwCAh4sAgIuLAIDvKAIA4ygOAI+LAIDhRA4AqbUCAKhpDQCrAQIAqgkCAK0BAgCsGQIArzECAK4BAgC+AAQAk4sAgJeLAICbiwCAn4sAgKOLAICniwCAq4sAgLnlAwC45QMAu+UDALrlAwC95QMAvOUDAL/lAwC+5QMAsSECALBJAgCzJQIAsiUCALUpAgC0IQIAtxUCALYVAgCowQIAqdECAKr1AgCrDQEArBUBAK0FAQCuBQEArzkBAK+LAICziwCAu4sAgL+LAIDDiwCAx4sAgMuLAIDPiwCAuC0BALk9AQC67QEAu+UBALz9AQC95QEAvu0BAL/lAQCwLQEAsTUBALI9AQCzNQEAtC0BALUVAQC2HQEAtxUBAIA9AQCBpQAAgq0AAO/YAACGsAUAh9gFANOLAIDv1A8AhGwEAOH0DgDXiwCA4xwPANuLAIDhlAEA34sAgOMMDgCzPQIA44sAgOeLAIDriwCA74sAgLbFAQC13QEA84sAgLuxAQC6qQEA94sAgPuLAIC/kQEAvqkBAL2hAQC8qQEAt4sAgP+LAICqRQYAq10GAKxFBgCtTQYArkUGAK99BgADjACAB4wAgAuMAICj0QUAD4wAgKUxBgCmKQYAE4wAgBeMAICCHQAAgR0AAIAdAAAbjACAH4wAgCOMAIC+lAMAJ4wAgCuMAICGSAMAh8wDAC+MAIAzjACAN4wAgDuMAICoqQcAqakHAKq5BwCruQcArKkHAK2pBwCuAQcArzUHAD+MAIBDjACAR4wAgEuMAIBPjACAU4wAgFeMAIBbjACAuC0HALnBAAC66QAAu+kAALz5AAC95QAAvuUAAL+dAACwUQcAsV0HALItBwCzJQcAtD0HALUlBwC2JQcAtxUHALMxBgBfjACAY4wAgGeMAIBrjACAtikGALUhBgBvjACAu5kGALqVBgBzjACAd4wAgL/hBgC++QYAvfEGALz5BgB7jACAo3UGAH+MAICDjACApm0GAIeMAICLjACApWUGAKrRBgCr3QYAj4wAgJOMAICuvQYAr6UGAKy9BgCttQYAqOUBAKn1AQCq/QEAq/UBAKztAQCtNQEArj0BAK81AQCA+QAAgc0AAILFAACEYAEAvngBAJuMAICHrAAAhpABALjRAAC52QAAuuEAALvhAAC8kQAAvZ0AAL6VAAC/iQAAsE0BALFVAQCyXQEAs1UBALRNAQC18QAAtvEAALfxAACzdQIAn4wAgKOMAICnjACAq4wAgLa1AgC1ZQIAr4wAgLuRAgC6iQIAs4wAgLeMAIC/NQMAvokCAL2BAgC8iQIAu4wAgKMxAgC/jACAhMADAKbxAgDDjACAx4wAgKUhAgCqzQIAq9UCAMuMAIDPjACArs0CAK9xAwCszQIArcUCAKuNAACqjQAAqY0AAKg5AwCvvQAArr0AAK2FAACsjQAAqgAAAKsAAADTjACA14wAgNuMAIDfjACA44wAgOeMAIC7fQAAun0AALl9AAC4fQAAv90BAL7dAQC93QEAvN0BALO5AACysQAAsaEAALCtAAC3XQAAtl0AALWVAAC0lQAA64wAgO+MAIDzjACA94wAgIE1AACADQAA+4wAgII1AAC+rD0A/4wAgAONAICFaD0AC40AgA+NAICGODwAh8ACALNJAQATjQCA0AAAABeNAIAbjQCAtkkBALVJAQAfjQCAuykBALolAQAjjQCAJ40AgL8dAQC+HQEAvSEBALwpAQDjNDYA4QwGAOGwAgDjPAYAK40AgC+NAIAzjQCAN40AgIQsPwC+oD8AO40AgD+NAIDvfDcAQ40AgEeNAIDvGAEAS40AgE+NAICGaD4Ah8w/AFONAIBXjQCAW40AgO+UAABfjQCA4ZQBAGONAIDjUAAAZ40AgILpPwCB6T8AgPE/AKMJPgCPASQAB40AgGuNAIBvjQCApgk+AKUJPgBzjQCAq2k+AKplPgB3jQCAe40AgK9dPgCuXT4ArWE+AKxpPgCeYTgAn3U4AJzBNACdtTkAmqU1AJt1NACYeTAAmXExAJYhLQCXhTEAlG0sAJVlLACSeSgAk6UtAJBRJACReSgAsQ0UALAFFACzARgAslUUALV5GAC0tRgAf40AgIONAICHjQCAi40AgI+NAICTjQCAotE8AKMlAQCgdTkAob08AKHJAACXjQCAowEEAKLlAAClHQQApPUEAKf5CACmAQgAqQEMAKhtCACrzQwAqs0MAK3REACsARAAr9URAK7ZEACCBSUAgy0lAJuNAICfjQCAhsEsAIcRLQCEHSkAhRUpAIopLQCLZSwAo40AgKeNAICOHTAAj8E0AIzZMACNHTEAkmE1AJPNNQCrjQCAr40AgJZhOQCXmTgAlKE4AJV9OQCaYT0AmwU9ALONAIC3jQCAu40AgL+NAICc6QAAw40AgMeNAIDLjQCAz40AgNONAICXjACA140AgNuNAIDfjQCAqJE+AKmRPgCq7T4Aq+E+AKzhPgCt6T4ArtE+AK/RPgCwUT4AsVE+ALJRPgCzUT4AtHk+ALV5PgC2bT4At2U+ALghPgC5IT4Aujk+ALs5PgC8KT4AvRU+AL4RPgC/DT4AgJkDAIGZAwCCBQAA440AgL5UAwDhsD0A640AgONAPgCEOAIA740AgPONAIDv9D8A940AgPuNAICGmAQAhxwDALMFPQCECAQA/40AgAOOAIAHjgCAtgk9ALUJPQALjgCAu/U9ALr1PQAPjgCAE44AgL/dPQC+3T0AveU9ALzlPQAXjgCAG44AgKPNPQC+xAQApcE9AB+OAIAjjgCApsE9ACeOAIArjgCAqz09AKo9PQCtLT0ArC09AK8VPQCuFT0AtmkCAC+OAIAzjgCAtWkCADeOAICzSQIAO44AgD+OAIC+qQMAv6kDALzBAwC9wQMAuvkDALv5AwBDjgCAR44AgKgtAwCpnQMAqpUDAKutAwCstQMArb0DAK61AwCv2QMAgA0AAIEVAACCHQAAS44AgE+OAIBTjgCAh7QFAIacBAC4MQIAuTECALo1AgC7zQIAvNUCAL3dAgC+1QIAv8kCALBpAgCxaQIAskECALNBAgC0OQIAtTkCALYRAgC3EQIAW44AgOM0PgBfjgCA4aw+AGOOAIDvfAMAZ44AgGuOAIBvjgCA45QDAHOOAIDhfD4Ad44AgO/oPgB7jgCAf44AgIOOAICHjgCAo1UDAIuOAICldQMAj44AgJOOAICmdQMAl44AgJuOAICr5QIAquUCAK3dAgCs3QIAr7UCAK61AgCoGQYAqSEGAKohBgCrPQYArCUGAK1dBgCuVQYAr00GAFeOAICfjgCAo44AgKeOAICrjgCAr44AgLOOAIC3jgCAuOUGALmBBgC6gQYAu50GALyJBgC9iQYAvqEGAL+hBgCwPQYAsQ0GALIFBgCz7QYAtPUGALXhBgC24QYAt90GALOpBgCCLQAAgRUAAIAdAAC7jgCAtt0GALWtBgC/jgCAu8kGALr5BgDDjgCAhOADAL8lBgC+MQYAvTkGALzRBgC+iAMAo+0GAOeNAIDHjgCAppkGAMuOAIDPjgCApekGAKq9BgCrjQYAhkgAAIdsAACudQYAr2EGAKyVBgCtfQYAqIEGAKmNBgCqmQYAq5UGAKyNBgCttQYArrEGAK+tBgDTjgCA144AgNuOAIDfjgCA444AgOeOAIDrjgCA744AgLilBgC5YQEAumEBALthAQC8YQEAvWEBAL5hAQC/YQEAsNkGALHZBgCyqQYAs6kGALS9BgC1oQYAtqEGALedBgCzEQYA844AgPeOAID7jgCA/44AgLY1BgC1BQYAA48AgLsdBgC6HQYAB48AgAuPAIC/ZQYAvnkGAL19BgC8fQYAD48AgKNVBgATjwCAF48AgKZxBgAbjwCAH48AgKVBBgCqWQYAq1kGACOPAIAnjwCArj0GAK8hBgCsOQYArTkGAKjVAgCp3QIAqikDAKspAwCsOQMArTkDAK4pAwCvKQMAK48AgC+PAIAzjwCAO48AgD+PAIBDjwCAvrgDAEePAIC47QMAuYUDALqBAwC7gQMAvIUDAL2NAwC+sQMAv7EDALBZAwCxWQMAsu0DALPlAwC0/QMAteUDALblAwC31QMAgKEAAIGhAACCoQAAvoAMAEuPAICEmAIAT48AgFOPAICGAAwAh/QDAFePAIBbjwCAX48AgGOPAIBnjwCAhLADALPhAwBrjwCAb48AgHOPAIB3jwCAtvkDALXxAwB7jwCAu90DALrdAwB/jwCAg48AgL9hAwC+eQMAvXEDALx5AwCHjwCAi48AgI+PAICjLQIAk48AgKU9AgCmNQIAl48AgJuPAICfjwCAqhECAKsRAgCstQIArb0CAK61AgCvrQIA48QDAOMQBwDhuAEA4WwHAIBxAACBcQAAggUAAKOPAICGwAwAh1QNAKuPAICvjwCA77ADAO8ABwCzjwCAt48AgLuPAIC/jwCAw48AgMePAIDLjwCAz48AgNOPAIDvpAEAhKANAOGABgDXjwCA4xABANuPAIDfjwCA448AgOePAICz9QEA648AgO+PAIDzjwCA948AgLZNAQC1SQEA+48AgLtRAQC6SQEA/48AgAOQAIC/OQEAvjEBAL1BAQC8SQEAqC0OAKk1DgCqPQ4AqzEOAKyBDgCtjQ4AroUOAK+1DgCnjwCAB5AAgAuQAIAPkACAgBkAAIEZAACCBQAAE5AAgLidDgC5rQ4AuqUOALtNDwC8VQ8AvV0PAL5JDwC/QQ8AsM0OALHVDgCy3Q4As9UOALS1DgC1vQ4AtrUOALetDgCjtQ4AvogDABeQAIAbkACAH5AAgKYNDgClCQ4AI5AAgKsRDgCqCQ4AhggAAIdsAwCveQ4ArnEOAK0BDgCsCQ4AJ5AAgCuQAIAvkACAs7UPADOQAIC1VQ8Atl0PADePAIA3kACAO5AAgLp5DwC7eQ8AvGkPAL1dDwC+SQ8Av0kPAKhpDgCpaQ4AqnEOAKtxDgCskQ4ArZEOAK6RDgCvkQ4AP5AAgEOQAIBHkACAS5AAgE+QAIBTkACAV5AAgFuQAIC4hQ4AuY0OALqFDgC7nQ4AvI0OAL29DgC+tQ4Av3kBALDxDgCx8Q4AsvEOALPFDgC0wQ4AtcEOALbBDgC3wQ4Ao/kOAF+QAIBjkACAZ5AAgGuQAICmEQ4ApRkOAG+QAICrNQ4AqjUOAHOQAIB3kACArwUOAK4FDgCtEQ4ArCUOAIANAACBFQAAgh0AAHuQAIB/kACAg5AAgISUAQC+lAEAhkAHAIf0AACLkACAj5AAgJOQAICXkACAm5AAgJ+QAICojQIAqZUCAKqVAgCrzQIArNUCAK3dAgCuyQIAr/0CAKOQAICnkACAq5AAgK+QAIC/ABQAs5AAgLeQAIC7kACAuH0DALnBAwC6wQMAu8EDALzBAwC9yQMAvvEDAL/xAwCwhQIAsUUDALJNAwCzRQMAtF0DALVFAwC2TQMAt0UDALMdAgC/kACAw5AAgMeQAIDLkACAtl0CALVdAgDPkACAu4EDALpBAgDTkACA15AAgL+BAwC+mQMAvZEDALyZAwDbkACAo1kCAN+QAIDjkACAphkCAOeQAIDrkACApRkCAKoFAgCrxQMA75AAgPOQAICu3QMAr8UDAKzdAwCt1QMA+5AAgOPMAACEBAIA4bwBAIDJAQCB/QEAgvUBAL4QBQD/kACAvigEAAORAIAHkQCAC5EAgO8QAAAPkQCAE5EAgIbgBACH9AIAF5EAgBuRAIDj/A8AH5EAgOHgDwAjkQCA7xQPACeRAIArkQCAL5EAgDORAIA3kQCAO5EAgD+RAIBDkQCAR5EAgEuRAIBPkQCAU5EAgFeRAIBbkQCA7+ABAIUEEgDh3A4AX5EAgOMcDgCAKQAAgR0AAIIFAABjkQCAszECAGuRAICEzAUAb5EAgHORAIC2KQIAtSECAHeRAIC7zQEAus0BAHuRAIB/kQCAv3UBAL7JAQC9wQEAvMkBAKjpBQCp6QUAqvkFAKv5BQCs6QUArekFAK45BgCvOQYA95AAgGeRAICGiAAAhwADAIORAICHkQCAi5EAgI+RAIC40QYAudkGALrhBgC74QYAvJEGAL2dBgC+lQYAv4kGALBJBgCxSQYAsl0GALNVBgC0TQYAtfEGALbxBgC38QYAo3EFAJORAICXkQCAm5EAgJ+RAICmaQUApWEFAKORAICrjQYAqo0GAKeRAICrkQCArzUGAK6JBgCtgQYArIkGAK+RAICzkQCAs+EHALeRAIC14QcAu5EAgL+RAIC25QcAh5AAgMORAIC7vQcAuqEHAL2VBwC8qQcAv5UHAL6VBwCoAQYAqSUGAKohBgCrIQYArCEGAK0tBgCuJQYAr1UGAMeRAICCHQAAgR0AAIAdAADLkQCAz5EAgNORAIC+MAEAuDkGALk5BgC6yQYAu8kGALzZBgC92QYAvskGAL/JBgCwLQYAsTEGALI1BgCzCQYAtBkGALUZBgC2CQYAtwkGAKOpBgCEjAIAhigfAIdEAQDbkQCApq0GAKWpBgDfkQCAq/UGAKrpBgDjkQCA55EAgK/dBgCu3QYArd0GAKzhBgDrkQCAsxUGAO+RAIDzkQCAtj0GAPeRAID7kQCAtTUGALrZAQC72QEA/5EAgAOSAIC+fQEAv2UBALx9AQC9dQEAqMUFAKnJBQCq2QUAq9EFAKz5BQCt+QUArikCAK8pAgAHkgCAC5IAgA+SAIATkgCAjAAAABeSAIAbkgCAH5IAgLjtAgC5hQIAuo0CALuBAgC8hQIAvY0CAL69AgC/fQMAsFkCALFZAgCy7QIAs+UCALT9AgC15QIAtuUCALfVAgCjUQUAI5IAgCeSAIArkgCAL5IAgKZ5BQClcQUAM5IAgKudAgCqnQIAN5IAgDuSAICvIQIArjkCAK0xAgCsOQIAghEAAD+SAICAZQAAgQkAAEOSAIC+mAMAS5IAgE+SAICEJAMAU5IAgIdoAwCGjBwAV5IAgFuSAIBfkgCAY5IAgGeSAIBrkgCAs6ECAITAHAC10QIAb5IAgHOSAIC21QIAd5IAgHuSAIC7wQIAuvUCAL0RAQC82QIAvxEBAL4ZAQB/kgCAg5IAgIeSAICLkgCAj5IAgJOSAICXkgCA77gGAJuSAIDhnAQAn5IAgON0BgCjkgCAp5IAgKuSAICvkgCAgPkAAIH5AACCBQAAs5IAgL5YHACEWB8A71wAAO9ABgDhkAEA4fwGAOM8AADjdAYAu5IAgL+SAICGmBwAh/QcAKNpAgC+DB8Aw5IAgMeSAIDLkgCAph0CAKUZAgDPkgCAqwkCAKo9AgDTkgCA15IAgK/ZAQCu0QEArdkBAKwRAgCokR0AqZkdAKqhHQCroR0ArNEdAK3dHQCu1R0Ar8kdAEeSAIC3kgCA25IAgN+SAIDjkgCA55IAgOuSAIDvkgCAuHkeALl5HgC6zR4Au8UeALzdHgC9xR4AvsUeAL/1HgCwuR0AsY0dALKFHQCzTR4AtFUeALVdHgC2VR4At0keALjNHwC51R8Aut0fALvVHwC88R8Avf0fAL7pHwC/6R8AsKUfALGxHwCysR8As40fALSVHwC19R8Atv0fALf1HwCoGR4AqRkeAKotHgCrPR4ArCUeAK0tHgCuJR4Ar90fAPOSAID3kgCA+5IAgP+SAIADkwCA15EAgAeTAIALkwCAs+UfAA+TAIATkwCAF5MAgBuTAIC27R8Ate0fAB+TAIC7NR4AuiEeACOTAIAnkwCAv3EeAL4RHgC9GR4AvCUeAIJpAACjoR8AgFkAAIFRAACmqR8AK5MAgC+TAIClqR8AqmUeAKtxHgCGAAQAh+wBAK5VHgCvNR4ArGEeAK1dHgCoMR4AqTEeAKpBHgCrQR4ArEEeAK1JHgCucR4Ar3EeADOTAIA3kwCAO5MAgD+TAIBDkwCAR5MAgEuTAIBPkwCAuCkBALkpAQC6OQEAuzUBALwtAQC90QAAvtEAAL/RAACwyQEAsckBALLZAQCz2QEAtMkBALXJAQC2GQEAtxkBALPJHQBTkwCAV5MAgFuTAIBfkwCAtskdALXJHQBjkwCAuw0CALoNAgBnkwCAa5MAgL8NAgC+DQIAvQ0CALwNAgBvkwCAo40dAHOTAIB3kwCApo0dAHuTAIB/kwCApY0dAKpJAgCrSQIAg5MAgIeTAICuSQIAr0kCAKxJAgCtSQIAgA0AAIERAACCEQAAi5MAgO/MAgCPkwCAk5MAgISQAgDjLAIAvigDAOHYAQCbkwCAhhAEAIfUAwCfkwCAo5MAgLNhAwCnkwCAq5MAgK+TAICzkwCAtnkDALVxAwC3kwCAu10DALpdAwC7kwCAv5MAgL/hAAC++QAAvfEAALz5AACjoQIAw5MAgMeTAIDLkwCAz5MAgKa5AgClsQIA05MAgKudAgCqnQIA15MAgNuTAICvIQEArjkBAK0xAQCsOQEA35MAgOOTAIDvZB8A55MAgOuTAIDvkwCA85MAgPeTAICADQAAgREAAIIVAAD7kwCA4eAcAP+TAIDjiB8AA5QAgISAAgC+jAUAh0gFAIYsBAALlACAD5QAgO+kHgDv9B4A4QAeAOFQHwDjLB4A47AeABOUAIAXlACAG5QAgB+UAIAjlACAJ5QAgISEBACzcQEAK5QAgLUdAQC2FQEAL5QAgDOUAIA3lACAugEBALsBAQC89QAAvf0AAL71AAC/7QAAqK0GAKm9BgCqtQYAq8kGAKzZBgCt2QYArskGAK/BBgA7lACAP5QAgEOUAIBHlACAS5QAgE+UAIBTlACAV5QAgLhtBwC5BQcAug0HALsBBwC8AQcAvQEHAL4BBwC/AQcAsIkGALGJBgCybQcAs2UHALR9BwC1ZQcAtmUHALdVBwCXkwCAozkGAFuUAIAHlACApl0GAF+UAIBjlACApVUGAKpJBgCrSQYAZ5QAgGuUAICuvQcAr6UHAKy9BwCttQcAgG0AAIEJAACCGQAAb5QAgHOUAIC+nAMAd5QAgHuUAICGQAAAh2AAAH+UAICDlACAh5QAgIuUAICPlACAk5QAgKiRBgCpkQYAqrkGAKu5BgCsqQYArakGAK7ZBgCv2QYAl5QAgJuUAICflACAo5QAgKeUAICrlACAr5QAgLOUAIC4cQEAuXEBALpxAQC7cQEAvNkBAL3BAQC+wQEAv/UBALCxBgCxuQYAsokGALOJBgC0UQEAtVEBALZRAQC3UQEAszEGALeUAIC7lACAv5QAgMOUAIC2KQYAtSEGAMeUAIC7fQYAunUGAMuUAIDPlACAv5UBAL6VAQC9XQYAvF0GANOUAICjdQYA15QAgNuUAICmbQYA35QAgOOUAIClZQYAqjEGAKs5BgCErAEAvqABAK7RAQCv0QEArBkGAK0ZBgCo3QIAqe0CAKrlAgCr/QIArOUCAK3tAgCu5QIArz0DAOuUAIDvlACA85QAgL5kDAD3lACA+5QAgP+UAIADlQCAuMkDALnJAwC62QMAu9EDALz5AwC9+QMAvpkDAL+VAwCwRQMAsU0DALJFAwCzXQMAtEUDALVNAwC2RQMAt/kDAIFVAwCASQMAs2UCAIJVAwC1ZQIAB5UAgAuVAIC2ZQIAhgAMAIfkAwC7gQMAuokDAL2BAwC8mQMAv4EDAL6JAwCjLQIAD5UAgBOVAIAXlQCAG5UAgKYtAgClLQIAH5UAgKvJAwCqwQMAI5UAgCeVAICvyQMArsEDAK3JAwCs0QMA49gGAOGsBwDhnAYA45wGACuVAICEWA0AL5UAgDOVAIA3lQCAO5UAgD+VAIBDlQCA7xwBAEeVAIBLlQCA70AGAIB5AACBFQAAghEAAIQADABPlQCA46wAAFOVAIDhpAEAW5UAgO9wAACGyAwAh6QNAF+VAIBjlQCAZ5UAgGuVAIC6yQUAu8kFALilBQC5zQUAvvkFAL/5BQC8zQUAvcUFALKlBQCzrQUAsBEGALERBgC2rQUAt50FALS1BQC1rQUAqmEGAKthBgConQYAqZUGAK5hBgCvYQYArHEGAK1xBgBvlQCAc5UAgHeVAIB7lQCAf5UAgIOVAIC+sAwAh5UAgKghDgCpIQ4AqiEOAKs9DgCsJQ4ArS0OAK4lDgCviQ4AV5UAgIuVAICPlQCAk5UAgJeVAICblQCAn5UAgKOVAIC4UQ8AuV0PALpVDwC7bQ8AvHUPAL19DwC+dQ8Av2kPALD5DgCxoQ4AsqEOALOhDgC0oQ4AtakOALaRDgC3kQ4As6kOAKeVAIDnlACAq5UAgK+VAIC2rQ4Ata0OALOVAIC7ZQ4Auj0OALeVAIC7lQCAv20OAL5lDgC9dQ4AvHUOAIIZAACj7Q4AgGUAAIEZAACm6Q4Av5UAgMOVAICl6Q4AqnkOAKshDgDHlQCAy5UAgK4hDgCvKQ4ArDEOAK0xDgCoYQ4AqXUOAKp9DgCrdQ4ArG0OAK31DgCu/Q4Ar/UOAIaAAQCHpAEAz5UAgNOVAIDXlQCA25UAgN+VAIDjlQCAuHUBALl9AQC6dQEAu8kBALzdAQC9xQEAvsUBAL/1AQCwjQ4AsZUOALKdDgCzkQ4AtFUBALVdAQC2VQEAt00BALP1DgDnlQCA65UAgO+VAIDzlQCAtnUOALXlDgD3lQCAu1EOALpJDgD7lQCA/5UAgL+ZAQC+kQEAvUUOALxJDgADlgCAo7EOAAeWAIALlgCApjEOAA+WAIATlgCApaEOAKoNDgCrFQ4AF5YAgBuWAICu1QEAr90BAKwNDgCtAQ4AqO0CAKktAwCqJQMAqz0DAKwlAwCtLQMAriUDAK+ZAwAflgCAI5YAgCeWAIArlgCAL5YAgDOWAIC+dAIAO5YAgLiNAwC5kQMAupEDALulAwC8vQMAvXUAAL59AAC/dQAAsOkDALHpAwCy+QMAs/EDALTZAwC12QMAtrkDALe1AwCArQAAgbUAAIK9AACzoQMAP5YAgLWhAwC2oQMAQ5YAgITgAgBHlgCAuiEDALshAwC8IQMAvSkDAL4RAwC/EQMAo+0DAIXABACFtG8AS5YAgE+WAICm7QMApe0DAFOWAICrbQMAqm0DAIZIBQCHbAMAr10DAK5dAwCtZQMArG0DAFeWAIDjAA4A71hsAOG0DwBblgCAX5YAgGOWAIBnlgCAoakDAKD9DwCjwQMAog0DAOHgAwDv4A8A4+QDAGuWAIBvlgCAc5YAgIQEBAC+BAQAd5YAgO+UAwB7lgCAf5YAgIOWAIDj1AMAh5YAgOFUAACLlgCAj5YAgJOWAICXlgCAgA0AAIEVAACCHQAAm5YAgJ+WAICjlgCAj5EbAO+cDgCE4AcA4dQOAKuWAIDj8A4Ar5YAgLOWAICGGAcAh5AEAJnlFwCY5RcAm+kLAJo5CwCd/QoAnPELAJ9VDwCeXQ8AkSkfAJDNGwCTJR8Aks0fAJXREwCUKRMAlxkXAJZ1EwCM4RAAjSUQAI4tEACP+QwAN5YAgKeWAICKORQAi5UUAITpGACFBRgAhuUYAIfxFAC3lgCAu5YAgIIxHACDFRwAnKkEAL+WAIDDlgCAx5YAgMuWAIDPlgCAmtEEAJt9BACUTQ0AleUIAJblCACXtQgA05YAgNeWAICSWQwAk1kMAKGRAADblgCAowF8AKKZAACluXwApJF8AKeZeACm4X0AqYF5AKiheACriXQAqgF0AK0BcACsWXQAr4VwAK6dcACx4WwAsAFsALMBaACyHWwAtfVoALT1aADflgCA45YAgOeWAIDrlgCA75YAgPOWAID3lgCA+5YAgP+WAIADlwCAqD0HAKmVBwCqlQcAq6kHAKzdBwCtxQcArsUHAK8dBgAHlwCAgh0AAIEdAACAHQAAC5cAgA+XAIATlwCAvmABALgZBgC5GQYAuikGALslBgC8IQYAvSEGAL4hBgC/IQYAsHEGALFxBgCycQYAs3EGALRNBgC1NQYAtj0GALctBgCzHQcAG5cAgIYoAACHqAAAH5cAgLZFBwC1VQcAI5cAgLu1BgC6tQYAJ5cAgCuXAIC/8QYAvokGAL2lBgC8pQYAL5cAgKNZBwAzlwCAN5cAgKYBBwA7lwCAP5cAgKURBwCq8QYAq/EGAEOXAIBHlwCArs0GAK+1BgCs4QYAreEGAKipBQCptQUAqr0FAKs9AgCsJQIArVECAK5RAgCvUQIAS5cAgE+XAIBTlwCAV5cAgIQ8AwBblwCAX5cAgGOXAIC4pQIAua0CALqlAgC7vQIAvKUCAL2tAgC+pQIAv30DALAxAgCxMQIAshkCALMZAgC09QIAta0CALalAgC3nQIAZ5cAgGuXAIBvlwCAszkFAHOXAIC1oQIAtt0CAHeXAIB7lwCAf5cAgLr5AgC7+QIAvMECAL3BAgC+PQIAv2UCAIOXAICmgQIApf0CAIuXAICjZQUAvlh8AIbYfACHnHwArzkCAK5hAgCtnQIArJ0CAKulAgCqpQIAj5cAgJOXAICohQIAqZUCAKqVAgCrpQIArL0CAK3VAgCu0QIAr9ECAIGFAQCAhQEAl5cAgILtAQCblwCAn5cAgKOXAICnlwCAuHUBALl9AQC6dQEAu80BALzVAQC93QEAvskBAL/BAQCwtQIAsb0CALKBAgCzgQIAtFEBALVRAQC2UQEAt1EBAKuXAICvlwCAs5cAgLeXAIDhMAYA4WQHAOMoBgDjxAYAhCB9ALuXAIDvbAAA7xgGAL+XAIDDlwCAx5cAgMuXAICzXQIAvkh8AM+XAIDTlwCA15cAgLYVAgC1dQIA25cAgLs5AgC6MQIA35cAgOOXAIC/1QEAvtUBAL0VAgC8FQIAo519AIeXAIDnlwCA65cAgO+XAICm1X0ApbV9APOXAICr+X0AqvF9APeXAID7lwCArxV+AK4VfgCt1X0ArNV9AIBNAACBVQAAglUAALOxfgD/lwCAtWV/ALZtfwADmACAhkADAIcEAwC66X8Au+l/ALz5fwC9+X8Avt1/AL/NfwAHmACAC5gAgBeXAIAPmACAE5gAgBeYAIAbmACAH5gAgKhtfgCpXX4AqlV+AKuFfwCsgX8ArYF/AK6BfwCvgX8AsEF/ALFBfwCyQX8As0F/ALR1fwC1ZX8Atm1/ALdlfwC4XX8AuS1/ALolfwC7PX8AvC1/AL0dfwC+FX8Av/UAAKP9fwAjmACAJ5gAgCuYAIAvmACApiF+AKUpfgAzmACAq6V+AKqlfgA3mACAO5gAgK+BfgCukX4ArbV+AKy1fgA/mACAQ5gAgEeYAIBLmACAT5gAgFOYAIBXmACAW5gAgIA9AACBCQAAghkAAF+YAIBjmACAhLgBAL6wAQBnmACAqK0BAKnVAQCq1QEAqw0BAKwVAQCtGQEArgkBAK8JAQCGAAQAhwQBAGuYAIBvmACAc5gAgHeYAIB7mACAf5gAgLjtAAC5hQAAuo0AALuFAAC8nQAAvYUAAL6NAAC/hQAAsHkBALF5AQCy7QAAs+UAALT9AAC15QAAtuUAALfVAACzXQIAg5gAgIeYAICLmACAj5gAgLaZAgC1nQIAk5gAgLu9AgC6vQIAl5gAgJuYAIC/IQMAvjkDAL0xAwC8OQMAvigDAKMZAgCfmACAo5gAgKbdAgCnmACAq5gAgKXZAgCq+QIAq/kCAK+YAICzmACArn0DAK9lAwCsfQMArXUDAL7IBAC3mACAu5gAgL7EBQC/mACAw5gAgMeYAIDLmACAgD0AAIEJAACCGQAAz5gAgNOYAICEOAMA25gAgN+YAIDveAIA45gAgIZIBACHVAMA55gAgOuYAIDvmACA85gAgPeYAID7mACA/5gAgAOZAIDjVAIAB5kAgOFAAQALmQCAD5kAgOMkfwATmQCA4Zx8ABeZAIAbmQCAH5kAgCOZAICEbAUAJ5kAgCuZAIAvmQCAM5kAgO8YfwA3mQCAO5kAgLPxAgA/mQCAQ5kAgEuZAIBPmQCAtukCALXhAgBTmQCAu3EBALppAQCHoAUAhswEAL85AQC+WQEAvVEBALxhAQDhQH8AV5kAgOM4fgCEwAQAgtkAAO8UAACApQAAgdkAAFuZAIDjwAAAX5kAgOHUAQBjmQCAZ5kAgO+EfgBrmQCAqs0BAKvVAQBvmQCAc5kAgK79AQCvnQEArMUBAK31AQB3mQCAo1UCAHuZAIB/mQCApk0CAIOZAICHmQCApUUCANeYAIBHmQCAi5kAgI+ZAICTmQCAl5kAgJuZAICfmQCAqJkGAKmZBgCq7QYAq/0GAKzlBgCt7QYAruUGAK/dBgCwpQYAsa0GALKlBgCzuQYAtK0GALVVBwC2UQcAt00HALh1BwC5fQcAunUHALtJBwC8WQcAvVkHAL5JBwC/RQcAs0UGAKOZAICnmQCAq5kAgK+ZAIC2TQYAtU0GALOZAIC7SQYAukEGAIYIAACHjAAAv7EHAL5JBgC9TQYAvFEGAIJdAACjAQYAgEUAAIFdAACmCQYAu5kAgL+ZAIClCQYAqgUGAKsNBgDDmQCAx5kAgK4NBgCv9QcArBUGAK0JBgCoTQYAqVUGAKpVBgCriQYArLEGAK29BgCuqQYAr6kGALeZAIDLmQCAz5kAgNOZAIDXmQCA25kAgN+ZAIDjmQCAuEkBALlJAQC6WQEAu1kBALxJAQC9SQEAvt0BAL/VAQCw3QYAsa0GALKlBgCzjQYAtJkGALWZBgC2jQYAt4UGALPdBgDnmQCA65kAgO+ZAIDzmQCAtj0GALU5BgD3mQCAu2kGALoZBgD7mQCA/5kAgL9dBgC+XQYAvVkGALxxBgADmgCAo5kGAAeaAIALmgCApnkGAA+aAIATmgCApX0GAKpdBgCrLQYAF5oAgBuaAICuGQYArxkGAKw1BgCtHQYAqNUCAKndAgCq4QIAq+ECAKw1AwCtPQMArjUDAK8tAwCAzQMAgQkAAIIZAAAfmgCAI5oAgIQYAgC+dAMAK5oAgLjpAwC56QMAuokDALuFAwC8nQMAvYEDAL6BAwC/tQMAsFUDALFdAwCyVQMAs+kDALT5AwC1+QMAtukDALfhAwCGIAwAhxADAC+aAIAzmgCAN5oAgDuaAIA/mgCA71wCAEOaAIDhFAAAR5oAgOOIAgC++AwAS5oAgE+aAIBTmgCAu/kDALrxAwC+gA0AV5oAgL9dAwC+XQMAvV0DALzhAwCzCQIAW5oAgF+aAIBjmgCAZ5oAgLbdAwC13QMAa5oAgKipBgCpqQYAqrkGAKu5BgCsqQYArakGAK4dBQCvFQUAb5oAgHOaAIB3mgCAe5oAgH+aAICDmgCAh5oAgIuaAIC4GQUAuS0FALolBQC7yQUAvNkFAL3FBQC+zQUAv8UFALBtBQCxdQUAsnUFALNFBQC0XQUAtT0FALY1BQC3KQUA4fQGAOFUBwDjFAYA47wGAIEJAACAqQAAj5oAgII5AACE7A0Ak5oAgIeIDACGDAwAm5oAgJ+aAIDvzAcA78QHAKMpAwCjmgCAp5oAgKuaAICvmgCApv0CAKX9AgCzmgCAq9kCAKrRAgC3mgCAu5oAgK99AgCufQIArX0CAKzBAgCoPQ4AqY0OAKqFDgCrnQ4ArIUOAK2NDgCuuQ4Ar7UOAJeaAIC/mgCAw5oAgMeaAIDLmgCAz5oAgNOaAIDXmgCAuL0OALllDwC6bQ8Au2UPALx9DwC9ZQ8Avm0PAL9lDwCw1Q4Asd0OALLVDgCzoQ4AtJUOALWdDgC2lQ4At40OALMNDgDbmgCA35oAgOOaAIDnmgCAtg0OALUNDgDrmgCAuxkOALoRDgDvmgCAJ5oAgL9ZDgC+UQ4AvXUOALwBDgDzmgCAo0kOAPeaAID7mgCApkkOAP+aAIADmwCApUkOAKpVDgCrXQ4AhKQDAAebAICuFQ4Arx0OAKxFDgCtMQ4AqLEOAKmxDgCqzQ4Aq8UOAKzdDgCtxQ4ArsUOAK/1DgCA7QEAgfEBAILxAQALmwCAhpABAIe0AQAPmwCAE5sAgLjFAQC5zQEAusUBALvdAQC8zQEAvf0BAL6ZAQC/lQEAsI0OALFBAQCyQQEAs0EBALRBAQC1QQEAtkEBALdBAQCzRQ4AF5sAgBubAIAfmwCAI5sAgLZFDgC1VQ4AJ5sAgLuFAQC6SQ4AK5sAgC+bAIC/hQEAvoUBAL2VAQC8lQEAM5sAgKMBDgA3mwCAO5sAgKYBDgA/mwCAQ5sAgKURDgCqDQ4Aq8EBAEebAIBLmwCArsEBAK/BAQCs0QEArdEBAKgtAwCpPQMAqjUDAKuJAwCsmQMArZkDAK6JAwCvgQMAT5sAgFObAIBXmwCAW5sAgF+bAIBjmwCAZ5sAgGubAIC4rQMAuWUAALptAAC7ZQAAvH0AAL1lAAC+bQAAv2UAALDJAwCxyQMAsqkDALOlAwC0vQMAtaEDALahAwC3lQMAgL0AAIEJAACCGQAAb5sAgHObAIC+2AMAe5sAgH+bAICErAIAg5sAgIfoAwCGDAQAh5sAgIubAICPmwCAk5sAgLP9AwCXmwCAm5sAgJ+bAICjmwCAtlkDALVRAwCnmwCAu00DALpNAwCrmwCAr5sAgL8lAwC+OQMAvTEDALw9AwCzmwCAt5sAgLubAIC/mwCA71gPAMObAIDHmwCAy5sAgOOQDgDPmwCA4bAPANObAIDXmwCA25sAgN+bAIDjmwCAgHUAAIF9AACCdQAAhBgFAO88AwDrmwCAvhQFAO+bAIDj0AMA85sAgOFAAAD3mwCAhtAEAIdYBQD7mwCA/5sAgAOcAIAHnACAC5wAgA+cAIATnACAF5wAgBucAIDvrA8AhOwEAOEQDgAfnACA41QBACOcAIAnnACAK5wAgC+cAICj/QIAM5wAgDecAIA7nACAP5wAgKZZAgClUQIAQ5wAgKtNAgCqTQIAR5wAgEucAICvJQIArjkCAK0xAgCsPQIAqJkGAKmZBgCqrQYAq70GAKylBgCtrQYArqUGAK/ZBgDnmwCAghEAAIEZAACAwQcAT5wAgFOcAIC+cAMAV5wAgLhJBwC5SQcAul0HALtVBwC8TQcAvXEHAL51BwC/bQcAsKkGALGpBgCyuQYAs7EGALSZBgC1mQYAtnkHALd5BwC1NQYAW5wAgF+cAIC2NQYAhjAAAIdcAwCzPQYAY5wAgL19BgC8dQYAv0UGAL5FBgB3mwCAZ5wAgLt1BgC6dQYAo2UGAGucAIBvnACAc5wAgHecAICmbQYApW0GAHucAICrLQYAqi0GAH+cAICDnACArx0GAK4dBgCtJQYArC0GAKhVBgCpWQYAqm0GAKthBgCsaQYArWkGAK6ZBgCvmQYAh5wAgIucAICPnACAk5wAgJecAICbnACAn5wAgKOcAIC4+QYAufkGALqNBgC7hQYAvJ0GAL2FBgC+hQYAv7UGALDpBgCx6QYAsvkGALP5BgC06QYAtd0GALbJBgC3yQYAs+UGAKecAICrnACAr5wAgLOcAIC26QYAteEGALecAIC7LQYAui0GALucAIC/nACAvxkGAL4tBgC9LQYAvC0GAIIVAACjoQYAgGEAAIFhAACmrQYAw5wAgL6QAQClpQYAqmkGAKtpBgCEpAEAy5wAgK5pBgCvXQYArGkGAK1pBgCohQIAqY0CAKqVAgCruQIArNUCAK3dAgCu1QIAr80CAIaAHACHZAMAz5wAgL5gAwDTnACA15wAgNucAIDfnACAuHUDALl9AwC6dQMAu8kDALzZAwC92QMAvskDAL/BAwCwvQIAsY0CALKFAgCzTQMAtFUDALVdAwC2VQMAt00DALMdAgDjnACAhAgDAOecAIDrnACAtl0CALVdAgDvnACAu0kCALp5AgDznACA95wAgL+ZAwC+kQMAvZkDALxRAgCwAAAAo1kCAPucAID/nACAphkCAAOdAIAHnQCApRkCAKo9AgCrDQIAC50AgA+dAICu1QMAr90DAKwVAgCt3QMAE50AgBedAIAbnQCA76wGAB+dAIAjnQCAJ50AgCudAIC+6BwAL50AgDOdAIA7nQCAP50AgOGABwBDnQCA42AGAIBdAACBYQAAgmEAALN9AQBHnQCAtW0BALZlAQBLnQCAhiAdAIdYHQC6+QEAu/EBALzZAQC92QEAvrEBAL+xAQDvoAAAT50AgFOdAIBXnQCAW50AgF+dAIBjnQCA71wBAIRsHADhzAYAZ50AgOMcBgDjSAAAa50AgOEwAQBvnQCAo/EBAHOdAICFABQAd50AgHudAICm6QEApeEBAH+dAICrfQEAqnUBAIOdAICHnQCArz0BAK49AQCtVQEArFUBAKjtHQCpLR4AqjkeAKs5HgCsKR4ArSkeAK6dHgCvkR4AN50AgIudAICPnQCAk50AgJedAICC+QAAgfEAAID9AAC4qR4AuakeALpJHwC7SR8AvFkfAL1FHwC+TR8Av0UfALDxHgCx+R4AssEeALPBHgC0uR4AtbkeALatHgC3pR4AsBEfALERHwCyER8AsyUfALQlHwC1KR8Atl0fALdRHwC4cR8AuXkfALpBHwC7QR8AvJUAAL2dAAC+lQAAv40AAJudAIDHnACAn50AgKOdAICnnQCAq50AgIb4AwCH0AAAqM0fAKnVHwCq0R8Aq70fAKytHwCtcR8ArnEfAK9xHwCzOR4Ar50AgLOdAIC3nQCAu50AgLaRHgC1RR4Av50AgLu1HgC6tR4Aw50AgMedAIC/jR4AvoEeAL2RHgC8pR4Ay50AgKN9HgDPnQCA050AgKbVHgDXnQCA250AgKUBHgCq8R4Aq/EeAN+dAIDjnQCArsUeAK/JHgCs4R4ArdUeAKhVAQCpgQAAqoEAAKuBAACsgQAArYkAAK6xAACvsQAA550AgOudAIDvnQCA850AgPedAID7nQCA/50AgAOeAIC4ZQAAuW0AALplAAC7fQAAvGUAAL1tAAC+ZQAAv90DALChAACxrQAAsqUAALO5AAC0qQAAtZ0AALaVAAC3XQAAB54AgIIdAACBHQAAgB0AAAueAIAPngCAE54AgL4UAgAbngCAhKgCAB+eAIAjngCAJ54AgCueAIAvngCAjwAAALNJAwAzngCAhugEAIesAgA3ngCAtkkDALVJAwA7ngCAuykDALolAwA/ngCAQ54AgL8ZAwC+LQMAvS0DALwxAwBHngCAo40DAEueAIBPngCApo0DAFOeAIBXngCApY0DAKrhAwCr7QMAW54AgF+eAICu6QMAr90DAKz1AwCt6QMAvoQDAGOeAIBnngCAa54AgG+eAIBzngCAd54AgHueAICAPQAAgQkAAIIZAAB/ngCAg54AgIueAICENAMAj54AgLMtAQCTngCAh8wCAIZMBQCXngCAti0BALUtAQCbngCAu0kBALp5AQCfngCAo54AgL+9AQC+vQEAvbkBALxRAQDheB8Ap54AgOPQHwCrngCAr54AgOGUAQCzngCA42gDALeeAIC7ngCAv54AgO+IAwDDngCAx54AgO+sHwDLngCAz54AgNOeAIDXngCA254AgN+eAIDjngCA554AgO9EHgDrngCA4dweAO+eAIDjHB4A854AgPueAID/ngCAA58AgIFpAACAZQAAo+UBAIJ9AACl5QEAB58AgIQUBACm5QEAvigEAAufAICrgQEAqrEBAK1xAQCsmQEAr3UBAK51AQCoIQYAqS0GAKolBgCrPQYArCUGAK0tBgCuXQYAr00GAIeeAID3ngCAhggDAIeMAwAPnwCAE58AgBefAIAbnwCAuOkGALnpBgC6jQYAu4UGALydBgC9hQYAvo0GAL+FBgCwPQYAsQ0GALIFBgCz7QYAtPkGALX5BgC27QYAt+UGALDNBwCx1QcAstEHALPtBwC09QcAtf0HALbpBwC36QcAuN0HALklBwC6LQcAuyUHALw9BwC9JQcAvi0HAL8lBwAfnwCAI58AgBeeAIAnnwCAK58AgC+fAIAznwCAN58AgKgVBgCpGQYAqu0HAKv9BwCs7QcArd0HAK7VBwCvuQcAswUGADufAIA/nwCAQ58AgEefAIC2PQYAtQUGAEufAIC7cQYAumkGAE+fAIBTnwCAv1kGAL5RBgC9WQYAvGUGAFefAICjQQYAW58AgF+fAICmeQYAY58AgIS0AQClQQYAqi0GAKs1BgC+gAEAa58AgK4VBgCvHQYArCEGAK0dBgCoNQYAqT0GAKo1BgCrWQYArHUGAK2lAQCurQEAr6UBAIDpAACB6QAAgv0AAL8kAQCGMA8Ah+QAAG+fAIBznwCAuMUAALnNAAC6xQAAu90AALzNAAC9/QAAvvUAAL+dAACw3QEAsSUBALItAQCzIQEAtCEBALUhAQC2IQEAtyEBALvBAgC6OQIAd58AgHufAIC/xQIAvsUCAL3VAgC82QIAs50FAH+fAICDnwCAh58AgIwAAAC2BQIAtd0FAIufAICqfQIAq4UCAI+fAICTnwCAroECAK+BAgCsnQIArZECAJefAICj2QUAm58AgJ+fAICmQQIAo58AgKefAIClmQUAgpFqAIORagCrnwCAr58AgIa5FgCH6RcAhBEWAIWZFgCKoRIAi6ESALOfAIC3nwCAjpEeAI9ZHgCMmRMAjREeAJJxGgCT5RoAu58AgO/oJACW8QYAlwUGAJTlGgCVGQYAmikCAJvFAgC/nwCAw58AgMefAIDhKBsAnN0CAOMgDwCfIQcAnsEHAJ01GwCcLRsAm6EbAJr5HwCZOR8AmLEfAJcBEgCWIRMAlSkTAJRRFgCTGRcAkjEXAJGxFwCQKWsAj1FrAOOsBwCEBA0A4RwHAIANAACBNQAAgj0AAMufAIDPnwCA058AgL4gDQDbnwCA358AgO9MBwCGWAwAh2ANAOOfAIDnnwCA658AgO+fAICEXA8A858AgO8IAADvhAYA4ZABAOGwBgDj4AAA42QGAPefAID7nwCA/58AgAOgAIAHoACAC6AAgL4ADwCEQA4AD6AAgBOgAIAXoACAG6AAgB+gAIAjoACAJ6AAgCugAICj1QMAotUDAKExAwCgLQcAZ58AgNefAIAvoACAM6AAgDegAICCmQAAgZEAAICZAACoTQ0AqZ0NAKqVDQCrJQ4ArD0OAK0RDgCuEQ4ArxEOALB9DgCxDQ4AsgUOALMtDgC0OQ4AtTkOALYtDgC3JQ4AuOkOALnpDgC6wQ4Au8EOALy5DgC9nQ4AvpUOAL+NDgCzPQ0AO6AAgD+gAIBDoACAR6AAgLaxDgC1lQ4AS6AAgLvpDgC6mQ4AhogAAIfkAAC/3Q4Avt0OAL3ZDgC88Q4AT6AAgKN5DQC+hAEAhIAGAKb1DgBToACAV6AAgKXRDgCq3Q4Aq60OAFugAIBfoACArpkOAK+ZDgCstQ4ArZ0OALIFNQCzGTQAsG0wALENNQBjoACAZ6AAgLQBKAC1PSkAa6AAgG+gAIBzoACAd6AAgHugAIB/oACAg6AAgIegAICiRQEAo9UBAIugAIChTQEAps0FAKcBOACkAQQApX0FAKoBPACrRT0AqEk5AKnlOQCudTEAr30xAKxdPQCtATAAqO0OAKn1DgCqCQ4AqwkOAKwZDgCtGQ4Arg0OAK8tDgCPoACAk6AAgJegAICboACAn6AAgKOgAICnoACAq6AAgLgdDgC5JQ4Aui0OALslDgC8PQ4Avd0BAL7VAQC/zQEAsFUOALFdDgCyVQ4Asy0OALQ1DgC1JQ4Ati0OALclDgCzgQ0Ar6AAgLOgAIC7oACAv6AAgLaZDQC1kQ0AvlQEALuZDQC6kQ0AhogEAIe8AwC/4Q0AvvENAL35DQC8gQ0AgkkAAKPFDQCA9QMAgUkAAKbdDQDDoACAx6AAgKXVDQCq1Q0Aq90NAMugAIDPoACArrUNAK+lDQCsxQ0Arb0NAKgdAgCpRQIAql0CAKtVAgCseQIArXkCAK6JAwCviQMA06AAgNegAIDboACA36AAgIT8BQDjoACA56AAgOugAIC4iQMAuWUDALptAwC7ZQMAvH0DAL1lAwC+bQMAv2UDALDBAwCxwQMAssEDALPBAwC0wQMAtcEDALbBAwC3wQMA76AAgPOgAID3oACA+6AAgP+gAIDhpAEAA6EAgOPADgC+aAQAB6EAgAuhAIDvHAEAD6EAgBOhAIAXoQCAG6EAgLOVAwAfoQCAI6EAgCuhAIAvoQCAtrkDALWxAwAzoQCAu0UCALpFAgCGqAQAh6QFAL9FAgC+RQIAvVUCALxVAgDh4A4A4SwMAOMIDgDj1A4AgK0AAIHRAACC0QAAN6EAgDuhAIA/oQCAQ6EAgEehAIBLoQCAT6EAgO+IDgDvLA4AoxUDAFOhAICFxCsAV6EAgFuhAICmOQMApTEDAF+hAICrxQIAqsUCAGOhAIBnoQCAr8UCAK7FAgCt1QIArNUCAKgNBgCpFQYAql0GAKtVBgCseQYArXkGAK65BgCvuQYAJ6EAgGuhAIBvoQCAc6EAgHehAIB7oQCAf6EAgIOhAIC4TQcAuVUHALpRBwC7aQcAvHkHAL1lBwC+bQcAv2UHALDJBgCxyQYAst0GALPVBgC0zQYAtXUHALZ9BwC3dQcAs9UGAIehAICLoQCAj6EAgJOhAIC2+QYAtfEGAJehAIC7DQYAug0GAIYIAACHLAAAv7EHAL4JBgC9AQYAvAkGAIJRAACjkQYAgEEAAIFBAACmvQYAm6EAgJ+hAICltQYAqkkGAKtJBgCjoQCAp6EAgK5NBgCv9QcArE0GAK1FBgCwsQYAsbEGALLNBgCzwQYAtMEGALXJBgC28QYAt/EGALgFAQC5DQEAugUBALsdAQC8BQEAvQ0BAL4FAQC/uQEAq6EAgK+hAICzoQCAt6EAgLuhAIC/oQCAt6AAgMOhAICoLQYAqTUGAKo1BgCr8QYArNEGAK3RBgCu0QYAr9EGALPdBgDHoQCAy6EAgM+hAIDToQCAtjEGALU5BgDXoQCAuxUGALoVBgDboQCA36EAgL9tBgC+ZQYAvXUGALx5BgDjoQCAo5kGAOehAIDroQCApnUGAO+hAIDzoQCApX0GAKpRBgCrUQYA96EAgPuhAICuIQYArykGAKw9BgCtMQYAqNUCAKndAgCq4QIAq+ECAKxRAwCtUQMArlEDAK9RAwD/oQCAA6IAgL7sAwALogCAD6IAgBOiAIAXogCAG6IAgLjpAwC56QMAuokDALuFAwC8nQMAvYEDAL6BAwC/tQMAsDEDALExAwCyNQMAs+kDALT5AwC1+QMAtukDALfhAwCAbQMAgaUAAIKtAACzZQIAH6IAgLXVAwC23QMAI6IAgITgAgAnogCAuvkDALv5AwC87QMAvTEDAL4xAwC/MQMAh+wDAIZkPACyAAAAK6IAgC+iAIDjCAQAM6IAgOHsBgA3ogCA7wAGADuiAIA/ogCAQ6IAgEeiAIBLogCAT6IAgFOiAIBXogCAW6IAgF+iAIDjoAMAY6IAgOGoAQBnogCA7/ADAIIdAACBHQAAgB0AAGuiAIBvogCAc6IAgHuiAIC+TD0Af6IAgKOhAwC+QDwApRECAIOiAICHogCAphkCAIRsAgCLogCAqz0CAKo9AgCt9QIArCkCAK/1AgCu9QIAhkA8AIe0PQCPogCAk6IAgJeiAICbogCAn6IAgO9EBgCjogCA4dQGAKeiAIDjDAcAq6IAgK+iAICzogCAt6IAgLP1AQC7ogCAv6IAgMOiAIDHogCAtkUBALXlAQDLogCAuzEBALopAQDPogCA06IAgL8dAQC+HQEAvRkBALwlAQCoLT4AqTU+AKo9PgCrNT4ArC0+AK2FPgCuhT4Ar7k+AHeiAIDXogCA26IAgN+iAICAGQAAgRkAAIIFAADjogCAuLk+ALm5PgC6ST8Au0k/ALxZPwC9WT8Avk0/AL9BPwCwrT4AsbU+ALKxPgCzjT4AtJk+ALWZPgC2iT4At4k+AKO1PgCEjAIA56IAgOuiAIDvogCApgU+AKWlPgDzogCAq3E+AKppPgCGCAAAh2gDAK9dPgCuXT4ArVk+AKxlPgD3ogCAs5E/APuiAID/ogCAtlk/AAOjAIAHowCAtbk/ALp1PwC7fT8AC6MAgA+jAIC+QT8Av0E/ALxZPwC9VT8AsJU+ALGdPgCyqT4As6U+ALShPgC1oT4AtqE+ALehPgC45T4Aue0+ALrlPgC7/T4AvO0+AL3dPgC+1T4AvxkBABOjAIAXowCAG6MAgB+jAIAjowCAB6IAgCejAIArowCAqF0+AKkhPgCqPT4AqzU+AKwVPgCt/T4ArvU+AK/tPgCj1T4AL6MAgDOjAIA3owCAO6MAgKYdPgCl/T4AP6MAgKs5PgCqMT4AQ6MAgEejAICvBT4ArgU+AK0RPgCsHT4AgREAAIANAABLowCAghkAAE+jAIBTowCAhJQBAL4QAACGQAcAhwABAFujAIBfowCAY6MAgGejAIBrowCAb6MAgKiNAgCplQIAqpUCAKvNAgCs2QIArdkCAK7NAgCvxQIAc6MAgHejAIB7owCAf6MAgIwAAACDowCAh6MAgIujAIC4HQMAucEDALrBAwC7wQMAvMEDAL3JAwC+8QMAv/EDALCJAgCxiQIAsikDALMpAwC0OQMAtTkDALYpAwC3JQMAsx0CAI+jAICTowCAl6MAgJujAIC2WQIAtVECAJ+jAIC7TQIAuk0CAKOjAICnowCAv/0DAL79AwC9/QMAvP0DAKujAICvowCAs6MAgLejAIDhDD4Au6MAgOOoPwC/owCAgT0AAIAxAADvUD8Agh0AAMOjAIC++AQAhhgFAIdMAwCEDAIA48wAAMujAIDhvAEAz6MAgNOjAIDXowCA26MAgN+jAICELAUA46MAgOejAIDrowCA7xAAAO+jAIDzowCAo90DAPejAID7owCA/6MAgAOkAICmmQMApZEDAAekAICrjQMAqo0DAAukAIAPpACArz0CAK49AgCtPQIArD0CABOkAIAXpACAG6QAgB+kAIAjpACAJ6QAgCukAIDvKD4AL6QAgOE8PgAzpACA4zgBAIApAACBFQAAghEAADukAICzMQIAvsgEAITABAA/pACAQ6QAgLYpAgC1IQIAR6QAgLvNAQC6zQEAS6QAgE+kAIC/dQEAvskBAL3BAQC8yQEAqOkFAKnpBQCq+QUAq/kFAKzpBQCt6QUArjkGAK85BgDHowCAN6QAgIaIAACHQAMAU6QAgFekAIBbpACAX6QAgLjRBgC52QYAuuEGALvhBgC8kQYAvZEGAL6RBgC/kQYAsEkGALFJBgCyXQYAs1UGALRNBgC18QYAtvEGALfxBgCjcQUAY6QAgGekAIBrpACAb6QAgKZpBQClYQUAc6QAgKuNBgCqjQYAd6QAgHukAICvNQYArokGAK2BBgCsiQYAf6QAgLPRBwCDpACAh6QAgLbxBwCLpACAj6QAgLXBBwC60QcAu90HAJOkAICXpACAvrkHAL+5BwC8xQcAvbkHALhpBgC5aQYAuokGALuJBgC8mQYAvZkGAL6JBgC/iQYAsBEGALEdBgCyFQYAs2kGALR5BgC1eQYAtmkGALdhBgCoSQYAqVUGAKpdBgCrVQYArE0GAK11BgCucQYAr3EGAFejAICCHQAAgR0AAIAdAACbpACAn6QAgKOkAIC+cAEAo5UGAKukAICGKAAAh0gBAK+kAICmtQYApYUGALOkAICrmQYAqpUGALekAIC7pACAr/0GAK79BgCt/QYArIEGAL+kAICzFQYAw6QAgMekAIC2PQYAy6QAgM+kAIC1NQYAutkBALvZAQDTpACA16QAgL59AQC/ZQEAvH0BAL11AQCovQUAqckFAKrZBQCr0QUArPkFAK35BQCuKQIArykCANukAIDfpACA46QAgOekAICMAAAA66QAgO+kAIDzpACAuO0CALmFAgC6gQIAu4ECALyFAgC9jQIAvrECAL+xAgCwWQIAsVkCALLtAgCz5QIAtP0CALXlAgC25QIAt9UCAKNRBQD3pACA+6QAgP+kAIADpQCApnkFAKVxBQAHpQCAq50CAKqdAgALpQCAD6UAgK8hAgCuOQIArTECAKw5AgCBbQAAgG0AABOlAICCBQAAvlwMABulAIAfpQCA79AGAITsAwDhHAUAI6UAgOP8BwAnpQCAK6UAgIbYDACHvAwAqIUCAKmVAgCqlQIAq6UCAKy9AgCt1QIArtECAK/RAgAvpQCAM6UAgDelAIA7pQCAP6UAgEOlAIBHpQCAS6UAgLh1AQC5fQEAunUBALvJAQC82QEAvdkBAL7JAQC/wQEAsLUCALG9AgCygQIAs4ECALRRAQC1UQEAtlEBALdRAQBPpQCAhAQNAFOlAIBXpQCAvhwMAFulAIDvHAAA76AGAOGQAQDhRAcA43AGAOOYBgBfpQCAY6UAgGelAIBrpQCAs10CAG+lAIBzpQCAd6UAgHulAIC2FQIAtXUCAH+lAIC7OQIAujECAIOlAICLpQCAv9UBAL7VAQC9FQIAvBUCAKOdDQAXpQCAh6UAgI+lAICTpQCAptUNAKW1DQCXpQCAq/kNAKrxDQCGCAMAh2ADAK8VDgCuFQ4ArdUNAKzVDQCAkQ8AgZkPAIKhDwCzpQ4Am6UAgLWhDgC2eQ8An6UAgKOlAICnpQCAukUPALtdDwC8RQ8AvU0PAL5FDwC//Q8AqFUOAKldDgCqYQ4Aq30OAKxlDgCttQ8Arr0PAK+1DwCrpQCAr6UAgLOlAIC3pQCAu6UAgL+lAIDDpQCAx6UAgLhVDwC5dQ8Aun0PALt1DwC8bQ8AvREPAL4RDwC/EQ8AsM0PALHVDwCy3Q8As9UPALTNDwC1dQ8AtnEPALdxDwCj6Q8Ay6UAgM+lAIDTpQCA16UAgKY1DgCl7Q8A26UAgKsRDgCqCQ4A36UAgOOlAICvsQ4ArgkOAK0BDgCsCQ4A56UAgIIdAACBHQAAgB0AAOulAIDvpQCA86UAgL6UAQCErAEA96UAgIfgAQCGzAAA+6UAgP+lAIADpgCAp6QAgKhtDgCpiQEAqpkBAKuRAQCswQEArckBAK75AQCv+QEAhKAAAAemAIALpgCAD6YAgBOmAIAXpgCAG6YAgB+mAIC4xQAAuc0AALrFAAC73QAAvM0AAL39AAC+9QAAv50AALBBAQCxQQEAskEBALNBAQC0QQEAtUEBALZBAQC3QQEAsxECACOmAIAnpgCAK6YAgC+mAIC2SQIAtUkCADOmAIC7hQIAuoUCADemAIA7pgCAv4UCAL6FAgC9lQIAvJUCAIU8GgCjVQIAP6YAgEOmAICmDQIAR6YAgEumAIClDQIAqsECAKvBAgBPpgCAU6YAgK7BAgCvwQIArNECAK3RAgCCGQAAV6YAgIAZAACBGQAAW6YAgF+mAIBjpgCAa6YAgL4ABABvpgCAc6YAgHemAIB7pgCAf6YAgIOmAICHpgCA7+gOAIumAICG6AQAh1ADAI+mAICTpgCA74ACAJemAIDhlAEAm6YAgONYAQCfpgCA4wAOAKOmAIDhaA0Ap6YAgKhxAgCpcQIAqnECAKupAgCsuQIArbkCAK6pAgCvqQIAhKwFAKumAICvpgCAs6YAgLemAIC7pgCAv6YAgMOmAIC4bQEAuQ0BALoFAQC7GQEAvAkBAL09AQC+NQEAv9kBALDZAgCx2QIAsm0BALNlAQC0fQEAtWUBALZlAQC3VQEA4WAPAOP0AADjHA4A4bwBAMemAICCOQAAgTEAAIA9AADLpgCAvigEAM+mAIDTpgCAvjwHAO8QAADv0A4A26YAgIbgBACHyAQA36YAgLO1AgDjpgCAtX0CALZ1AgDnpgCA66YAgO+mAIC6UQIAu1ECALz1AQC9/QEAvvUBAL/tAQBnpgCA16YAgKqxBQCrsQUArBUGAK0dBgCuFQYArw0GAPOmAID3pgCA+6YAgKNVBQD/pgCApZ0FAKaVBQADpwCAs+kGAAenAIALpwCAD6cAgBOnAIC24QYAtekGABenAIC7sQYAuqEGABunAIAfpwCAv50GAL6RBgC9pQYAvKkGAKgdBgCpIQYAqiEGAKshBgCsIQYArSEGAK4hBgCvIQYAI6cAgCenAIArpwCAL6cAgDOnAIA3pwCAO6cAgD+nAIC45QcAue0HALrlBwC7/QcAvOUHAL3tBwC+5QcAv00HALAlBgCxNQYAsj0GALMxBgC0FQYAtRkGALYNBgC3AQYAo6kHAIIVAACBtQEAgLUBAEOnAICmoQcApakHAEenAICr8QcAquEHAISgAgBLpwCAr90HAK7RBwCt5QcArOkHAE+nAICzlQYAhugAAIcYAQC2tQYAU6cAgFenAIC1vQYAukkBALtVAQBbpwCAX6cAgL45AQC/OQEAvEUBAL05AQCoPQYAqU0GAKpZBgCrUQYArHEGAK1xBgCuuQEAr7kBAISsAQBjpwCAZ6cAgGunAIBvpwCAc6cAgHenAIB7pwCAuKkBALmpAQC6aQEAu2kBALx5AQC9eQEAvmkBAL9pAQCwyQEAsdUBALLVAQCzqQEAtLkBALW5AQC2qQEAt6EBAKPRBQB/pwCAg6cAgIenAICLpwCApvEFAKX5BQCPpwCAqxECAKoNAgCTpwCAl6cAgK99AgCufQIArX0CAKwBAgCbpwCAn6cAgKOnAICnpwCAgTEAAIANAACrpwCAgjkAAK+nAICzpwCAviQDALunAIC/pwCAw6cAgIbYHACHTAMAx6cAgMunAIDPpwCAhMAcAOMgAQDTpwCA4cgBANenAIDvMAIA26cAgN+nAIDjpwCA56cAgOunAIDvpwCA86cAgLOVAwD3pwCA+6cAgP+nAIADqACAtrkDALWxAwAHqACAu1EDALpJAwALqACAD6gAgL/1AAC+SQMAvUEDALxJAwCoLQIAqUUCAKpdAgCrVQIArHkCAK15AgCuvQIAr7UCAL5oHQATqACAF6gAgBuoAICAHQAAgQkAAIKpAAAfqACAuFEBALlZAQC6YQEAu2EBALwRAQC9EQEAvhEBAL8RAQCwzQIAsdUCALLdAgCz1QIAtM0CALVxAQC2cQEAt3EBAOFYBgDhVAcA47AAAOO8BgAjqACAK6gAgIYYHACHVB0AL6gAgDOoAIA3qACAO6gAgL74HAA/qACA7/AGAO/gBgCjlQIAQ6gAgEeoAIBLqACAT6gAgKa5AgClsQIAU6gAgKtRAgCqSQIAV6gAgFuoAICv9QEArkkCAK1BAgCsSQIAqG0eAKl1HgCqfR4Aq40eAKyVHgCtnR4Aro0eAK+BHgAnqACAX6gAgGOoAIBnqACAa6gAgG+oAIBzqACAd6gAgLiJHgC5iR4AupkeALuRHgC8uR4AvbkeAL59HwC/dR8AsMUeALHNHgCyxR4As90eALTFHgC1zR4AtsUeALe5HgCz9R4Ae6gAgH+oAICDqACAh6gAgLYdHgC1HR4Ai6gAgLsJHgC6AR4Aj6gAgJOoAIC/CR4AvgEeAL0JHgC8ER4Agm0AAKOxHgCAVQAAgWUAAKZZHgCEmAMAv9ABAKVZHgCqRR4Aq00eAIYABACHmAEArkUeAK9NHgCsVR4ArU0eAJuoAICfqACAhCQAAKOoAICnqACAq6gAgLenAICXqACAqLUeAKmFHgCqjR4Aq4UeAKydHgCtgR4Arv0eAK/1HgCwjR4AsZUeALKVHgCzpR4AtL0eALVxAQC2cQEAt3EBALhRAQC5UQEAulEBALtRAQC89QEAvf0BAL71AQC/7QEAsyUeAL4IBwCvqACAs6gAgLeoAIC2IR4AtTUeALuoAIC7cR4AumkeAL+oAIDDqACAv5UBAL5ZHgC9UR4AvGEeAMeoAICjYR4Ay6gAgM+oAICmZR4A06gAgNeoAIClcR4Aqi0eAKs1HgDbqACA36gAgK4dHgCv0QEArCUeAK0VHgDhVBoA46gAgONcCgDnqACA66gAgO+oAIDzqACA96gAgPuoAIC+qAUA/6gAgAOpAICPMSoAC6kAgO/E+wAPqQCAk2EuAJIdLwCR2SoAkEkqAJfZEgCWdRIAlQ0TAJTBLgCbHRsAmkEWAJlJFgCYDRcAn3EeAJ4RGwCdcRoAnHkaAKOhAgCinQMAoZUfAKCJHgDjiAEA4wgeAOFoAADh/B4A79wBAO98HwC1if4AtAH8ALMB+gCylfoAsQH4ALAR9gCv4fYArgH0AK0l8gCs7fIAqwHwAKrpDwCp1Q4AqN0OAKcBDACmyQoApe0KAKQBCACj4QYAovEGAKHlAwATqQCAggErAIMBKwAXqQCAG6kAgIYxLwCHiS8AhIkrAIVFLgCKdRIAiwUTAIYIBQCHbAUAjhEXAI8RFwCMsRMAjV0WAJI9GgCTQRsAhMgFAIQABwCWUR8Al1EfAJRRGwCVORoAmn0eAJt9AgAfqQCAI6kAgIFZAQCAVQEAnFkDAIJRAQC+yAcAJ6kAgCupAIAvqQCAM6kAgDepAIA7qQCA79QeAD+pAIDhJB4AQ6kAgONoAQBHqQCAS6kAgE+pAIBTqQCAu2kCALpZAgBXqQCAW6kAgL8dAgC+HQIAvRkCALxxAgCz7QIAX6kAgGOpAIBnqQCAa6kAgLZ9AgC17QIAb6kAgKMNBQAHqQCAc6kAgHupAIB3qQCApp0FAKUNBQB/qQCAq4kFAKq5BQCGCAMAh3wDAK/9BQCu/QUArfkFAKyRBQCAsQcAgbkHAIJBAACzsQYAg6kAgLVZBwC2MQcAh6kAgIupAICPqQCAuuEHALvhBwC84QcAveEHAL7hBwC/3QcAqLUGAKm5BgCqdQYAq4UHAKydBwCt/QcArvUHAK8ZBwCTqQCAl6kAgJupAICfqQCAo6kAgKepAICrqQCAr6kAgLh1BwC5fQcAunUHALsFBwC8HQcAvTEHAL4xBwC/MQcAsGkHALFpBwCyeQcAs3kHALRpBwC1VQcAtlEHALdNBwCj/QcAs6kAgLepAIC7qQCAv6kAgKZ9BgClFQYAw6kAgKutBgCqrQYAx6kAgMupAICvkQYArq0GAK2tBgCsrQYAz6kAgNOpAIDXqQCA26kAgIAdAACBCQAAgjkAAN+pAIDjqQCA66kAgIbIAACHpAEA76kAgPOpAID3qQCA+6kAgKiNAQCpmQEAqtkBAKvRAQCs8QEArfEBAK45AQCvOQEAhKAAAP+pAIADqgCAB6oAgAuqAIAPqgCAE6oAgBeqAIC4zQAAudUAALrVAAC75QAAvP0AAL2VAAC+nQAAv5UAALBJAQCxSQEAslkBALNZAQC0SQEAtUkBALb9AAC39QAAugUEALsJBAC44QcAueEHAL4JBAC/CQQAvAkEAL0JBACyjQcAs+UHALC1BwCxhQcAtuUHALftBwC08QcAtfEHAKpNBwCrVQcAqEkHAKlJBwCu3QcAr8UHAKxNBwCt1QcAG6oAgB+qAIAjqgCAJ6oAgCuqAIAvqgCAM6oAgDeqAICz0QIAO6oAgD+qAIC+AAwAQ6oAgLbxAgC1+QIAR6oAgLsNAgC6DQIAS6oAgE+qAIC/DQIAvg0CAL0NAgC8DQIAghUAAKOVAgCAYQAAgWEAAKa1AgBTqgCAW6oAgKW9AgCqSQIAq0kCAIbIDACHrAwArkkCAK9JAgCsSQIArUkCAKhlAgCpdQIAqn0CAKt1AgCsbQIArbECAK6xAgCvsQIAhKANAF+qAIBjqgCAZ6oAgGuqAIBvqgCAc6oAgHeqAIC4MQEAuTEBALoxAQC7MQEAvNUBAL3dAQC+yQEAv8EBALDRAgCx0QIAstECALPRAgC0EQEAtREBALYRAQC3EQEA4bAGAHuqAIDj0AYAhEAPAH+qAIDhpAEAg6oAgOPABgCHqgCAi6oAgI+qAIDv1AYA7AAAAJOqAIDvZAcAl6oAgJuqAICfqgCAo6oAgLO5AgCnqgCAtakCALZ9AgCrqgCAr6oAgLOqAIC6WQIAu1kCALxJAgC9SQIAvpkBAL+ZAQCjdQ0AV6oAgLeqAIC7qgCAv6oAgKaxDQClZQ0Aw6oAgKuVDQCqlQ0AvqQDAMeqAICvVQ4ArlUOAK2FDQCshQ0AgE0AAIFVAACCVQAAs2UPAMuqAIC1ZQ8Atm0PAM+qAICGQAMAhxQDALrtDwC7/Q8AvOkPAL3VDwC+3Q8Av9UPAKhZDgCpoQ8AqqEPAKuhDwCsoQ8AraEPAK6hDwCvoQ8A06oAgNeqAIDbqgCA36oAgOOqAIDnqgCA66oAgO+qAIC4AQ8AuQEPALoBDwC7HQ8AvA0PAL01DwC+PQ8Av9UAALBlDwCxdQ8AsnEPALNNDwC0VQ8AtV0PALZNDwC3QQ8AoykOAPOqAID3qgCA+6oAgP+qAICmIQ4ApSkOAAOrAICrsQ4AqqEOAAerAIALqwCAr5kOAK6RDgCtmQ4ArKUOAA+rAIATqwCAF6sAgBurAIDvJA0AH6sAgCOrAIAnqwCA49AOACurAIDhGA4AL6sAgIAVAACBGQAAggUAADOrAICo0QEAqdkBAKopAQCrKQEArDkBAK05AQCuKQEArykBAL5oAQA7qwCAhsgBAIesAAA/qwCAQ6sAgEerAIBLqwCAuO0AALmFAAC6jQAAu4UAALydAAC9gQAAvoEAAL+BAACwWQEAsVkBALLtAACz5QAAtP0AALXlAAC25QAAt9UAALOhAgBPqwCAU6sAgFerAIBbqwCAtrkCALWxAgBfqwCAu50CALqdAgBjqwCAZ6sAgL8hAwC+OQMAvTEDALw5AwCF+PUAo+UCAGurAIBvqwCApv0CAHOrAIB3qwCApfUCAKrZAgCr2QIAe6sAgH+rAICufQMAr2UDAKx9AwCtdQMAuOkAALnpAAC6aQAAu2kAALx5AAC9ZQAAvm0AAL9lAACwsQAAsbkAALKBAACzgQAAtPkAALX5AAC27QAAt+UAAKhlAwCpdQMAqn0DAKt1AwCsbQMArdEAAK7RAACv0QAAg6sAgIerAICLqwCA56kAgI+rAICTqwCAl6sAgJurAICA/QEAgQkAAIIZAACfqwCAo6sAgL5EAgCrqwCAr6sAgISsAgCzqwCAh/gCAIasBQC3qwCAu6sAgL+rAIDDqwCAs/UCAMerAIDLqwCAz6sAgNOrAIC2UQEAteUCANerAIC7fQEAunUBANurAIDfqwCAvz0BAL49AQC9VQEAvFUBAOFwDwDjqwCA47gOAITABQDvyAAA56sAgOurAIDvqwCA4zwOAPOrAIDh0AEA96sAgIR0BwD7qwCA72gBAP+rAIADrACApXkCAKbNAQAHrACAgCEAAIEhAACC3QcAo2kCAKzJAQCtyQEArqEBAK+hAQALrACAD6wAgKrpAQCr4QEAp6sAgBOsAIC+QAIAF6wAgIYwAwCHMAMAG6wAgB+sAICoOQcAqTkHAKoNBwCrHQcArAUHAK0NBwCuBQcAr3kHALAJBwCxCQcAshkHALMRBwC0OQcAtTkHALbdBwC3yQcAuPkHALn5BwC6zQcAu8EHALzFBwC9yQcAvrkHAL+xBwCzpQcAI6wAgCesAIArrACAL6wAgLatBwC1rQcAM6wAgLvtBwC67QcAN6wAgDusAIC/3QcAvt0HAL3lBwC87QcAP6wAgKPhBwBDrACAR6wAgKbpBwBLrACAT6wAgKXpBwCqqQcAq6kHAFOsAIBXrACArpkHAK+ZBwCsqQcAraEHAFusAIBfrACAY6wAgGesAIBrrACAb6wAgHOsAIB3rACAgREAAIANAAB7rACAghkAAH+sAICDrACAvuQBAIesAICG4AAAhxgBAIusAICPrACAk6wAgJesAICbrACA77AEAJ+sAIDh1AYAo6wAgONcBACnrACAq6wAgK+sAICzrACAqJkBAKmZAQCqDQEAqwUBAKwdAQCtBQEArgUBAK81AQCEiAEAt6wAgLusAIC/rACAw6wAgMesAIDLrACAz6wAgLjBAAC5wQAAusEAALvBAAC8wQAAvcEAAL7BAAC/wQAAsE0BALElAQCyIQEAsyEBALQlAQC1LQEAthEBALcRAQDTrACA16wAgLONAgDbrACAtZ0CAN+sAIDjrACAto0CAOesAIDrrACAu+kCALqBAgC9/QIAvP0CAL/hAgC+6QIA76wAgKbVAgClxQIAvggDAKPVAgCCLQAAgRkAAIB5AACvuQIArrECAK2lAgCspQIAq7ECAKrZAgDzrACA+6wAgO80AgD/rACAhxgDAIYs/AADrQCAB60AgAutAIAPrQCAE60AgBetAIAbrQCAH60AgOMAAQAjrQCA4eABACetAIC6tQMAu70DACutAIAvrQCAvnkDAL95AwC8pQMAvXkDADerAICztQMAM60AgDetAIC2kQMAO60AgD+tAIC1pQMAqEkCAKlJAgCqWQIAq1kCAKxJAgCtdQIArnECAK9tAgC+aP0AvqT/AEOtAIBHrQCAS60AgE+tAIBTrQCAV60AgLj5AgC5+QIAukkBALtJAQC8XQEAvUEBAL5BAQC/fQEAsBUCALEdAgCyFQIAs8kCALTZAgC12QIAtskCALfJAgDjIAYA4bAGAOGAAQDjEAYAgA0AAIE1AACCPQAAW60AgF+tAIBjrQCAa60AgG+tAIDvcAAAc60AgHetAIDvTAEAhIz9AHutAICjmQIAf60AgKWJAgCDrQCAh60AgKa9AgCGwPwAh+T8AKuRAgCqmQIArVUCAKyJAgCvVQIArlUCAKh9/gCpgf4Aqpn+AKuZ/gCsif4ArYn+AK65/gCvuf4AZ60AgIutAICPrQCAk60AgJetAICbrQCAn60AgKOtAIC4tf4Aub3+ALph/wC7Yf8AvGH/AL1h/wC+Yf8Av2H/ALDJ/gCxyf4Ast3+ALPR/gC0uf4Atbn+ALaR/gC3kf4AsxH+AKetAICrrQCAr60AgLOtAIC2Cf4AtQH+ALetAIC7Df4Aug3+ALutAIC/rQCAv33+AL59/gC9Bf4AvAn+AMOtAICjVf4Ax60AgMutAICmTf4Az60AgNOtAIClRf4Aqkn+AKtJ/gCEKAMA160AgK45/gCvOf4ArE3+AK1B/gCAzQEAgdEBAILRAQCzuf4A260AgLXR/gC21f4A360AgIZgAQCHYAEAug0BALsFAQC8HQEAvQUBAL4NAQC/BQEA460AgOetAIDrrQCA760AgPOtAIDhwP0A960AgOOM/AD7rQCA/60AgAOuAIDvtPwAB64AgAuuAIAPrgCAE64AgKgp/gCpKf4Aqj3+AKs1/gCsVf4ArVn+AK5N/gCvRf4AF64AgBuuAIAfrgCAI64AgCeuAIArrgCAL64AgDOuAIC4SQEAuUkBALpZAQC7UQEAvHkBAL15AQC+GQEAvxUBALDFAQCxzQEAssUBALPdAQC0xQEAtc0BALbFAQC3eQEAN64AgDuuAIA/rgCAo7n9AEOuAICl0f0AptX9AITQAwBSrgCAvuACAKoNAgCrBQIArB0CAK0FAgCuDQIArwUCAIFJAACAQQAAowkDAIJdAAClGQMAVq4AgFquAICmEQMAhsAEAIfkAwCrDQMAqg0DAK0BAwCsHQMArwEDAK4JAwCw4QMAseEDALLhAwCz/QMAtOUDALXtAwC25QMAtz0DALgFAwC5DQMAugUDALsdAwC8BQMAvQ0DAL4FAwC/vQAAXq4AgGKuAIBmrgCAaq4AgPesAIBurgCAcq4AgHauAICo8QMAqfkDAKqpAwCrqQMArLkDAK25AwCuqQMAr6UDALNBAgB6rgCAfq4AgIKuAICGrgCAtlkCALVRAgCKrgCAu0UCALpFAgCOrgCAkq4AgL9JAgC+QQIAvUkCALxVAgCWrgCAmq4AgJ6uAICirgCA74wDAKauAICqrgCArq4AgONsAwCyrgCA4VAAALauAIC6rgCAvngFAMKuAICEcAIAgOUAAIHpAACC+QAAxq4AgIawBACHVAUAyq4AgO9A/gDOrgCA4Vz+ANKuAIDjVAEA1q4AgNquAIDergCA4q4AgLOZAQDmrgCA6q4AgO6uAIDyrgCAth0BALUdAQD2rgCAuz0BALo9AQD6rgCA/q4AgL/hAAC++QAAvfEAALz5AACoIQYAqVEGAKpRBgCrzQYArNUGAK3dBgCu1QYAr8kGAL6uAIACrwCABq8AgAqvAIAOrwCAEq8AgBavAIAarwCAuG0HALkFBwC6DQcAuwUHALwdBwC9AQcAvgEHAL8BBwCwuQYAsbkGALJtBwCzZQcAtH0HALVlBwC2ZQcAt1UHAKPZBgAerwCAIq8AgCavAIAqrwCApl0GAKVdBgCEnAIAq30GAKp9BgC+JAMALq8AgK+hBwCuuQcArbEHAKy5BwCASQAAgUkAAIJZAACzVQcAMq8AgLV9BwC2aQcANq8AgIZAAACHVAMAulUHALspBwC8OQcAvTkHAL4pBwC/IQcAo5kGADqvAIA+rwCAQq8AgEavAICmpQYApbEGAEqvAICr5QYAqpkGAE6vAIBSrwCAr+0GAK7lBgCt9QYArPUGAOE4BQBWrwCA4yQEAFqvAIBerwCAYq8AgGavAIBqrwCAbq8AgHKvAIB2rwCAeq8AgH6vAICCrwCA7/QEAIavAICo+QYAqQkGAKoRBgCrLQYArDkGAK0lBgCuLQYAryUGAIqvAICOrwCAkq8AgJavAICAGQAAgRkAAIIFAACarwCAuOUBALntAQC65QEAu/0BALzlAQC97QEAvuUBAL9ZAQCwXQYAsSEGALIhBgCzIQYAtCEGALUpBgC2EQYAtxEGAKjRAgCp2QIAqg0DAKsFAwCsHQMArQUDAK4FAwCvNQMAvmQCAKKvAICmrwCAqq8AgK6vAICyrwCAtq8AgLqvAIC4JQMAuS0DALolAwC7PQMAvCUDAL0pAwC++QMAv/kDALBNAwCxIQMAsiUDALM9AwC0JQMAtS0DALYlAwC3HQMAs4UDAITIAgC+rwCAhAgDAMKvAIC2hQMAtZUDAMavAIC75QMAuokDAIYIDACHnAMAv+kDAL7hAwC96QMAvPEDAIXsCgBHrgCAo80DAMqvAICl3QMAzq8AgNKvAICmzQMA1q8AgNqvAICrrQMAqsEDAK2hAwCsuQMAr6EDAK6pAwDerwCA4q8AgOavAIDqrwCA78gDAO6vAIDyrwCA9q8AgOO0AwD6rwCA4dABAP6vAICADQAAgXUAAIJ9AAACsACABrAAgAqwAICzZQEAvgQCALVlAQASsACAFrAAgLZlAQCGQA0Ah1gNALv1AQC6/QEAvaUBALy5AQC/mQEAvqUBABqwAIAesACAIrAAgIQADAAmsACAKrAAgC6wAIDvzAEAMrAAgOEsBgA2sACA4yABAOwAAAA6sACAPrAAgEKwAIBGsACAo+kBAEqwAIBOsACApukBAFKwAIBWsACApekBAKpxAQCreQEAWrAAgF6wAICuKQEArxUBAKw1AQCtKQEAqCUOAKktDgCqJQ4Aqz0OAKwlDgCtLQ4AriUOAK+VDgAOsACAYrAAgGawAIBqsACAbrAAgIKdAACBnQAAgJ0AALhFDwC5TQ8AukUPALtZDwC8SQ8AvUkPAL59DwC/cQ8AsPEOALH5DgCypQ4As7kOALSpDgC1lQ4Atp0OALd9DwCo1Q8Aqd0PAKoJDwCrCQ8ArBkPAK0FDwCuDQ8ArwUPAHKwAIB2sACAerAAgL6gAwB+sACAgrAAgId4AwCGEAAAuBUPALkdDwC6IQ8AuyEPALz1AAC9/QAAvvUAAL/tAACwQQ8AsU0PALJdDwCzVQ8AtE0PALU1DwC2MQ8AtzEPAIawAIDvsAwAirAAgI6wAICSsACAlrAAgJqwAICesACAorAAgKawAICqsACArrAAgLKwAIDjqA0AtrAAgOGMDQCzwQ4AurAAgL6wAIDCsACAxrAAgLbFDgC10Q4AyrAAgLvJDgC6xQ4AzrAAgNKwAIC/sQ4AvskOAL3BDgC8yQ4AowEOANawAIDasACA3rAAgOKwAICmBQ4ApREOAOawAICrCQ4AqgUOAOqwAICErAIAr3EOAK4JDgCtAQ4ArAkOAIBRAACBWQAAgmEAALPFAAC+zAEAtcUAALbNAADysACAhkAHAIcUAQC6yQAAu8kAALzZAAC92QAAvskAAL/FAACrDQMAqg0DAKkJAwCouQIArw0DAK4NAwCtDQMArA0DAL5gAwD2sACA+rAAgP6wAIACsQCABrEAgAqxAIC+MAUAuykDALoZAwC5GQMAuAEDAL/dAwC+3QMAvd0DALwxAwCzTQMAsk0DALFNAwCwTQMAtzkDALYxAwC1QQMAtE0DAA6xAICmkQMApZkDABKxAICjmQMAFrEAgBqxAIAesQCAr5kDAK6VAwCthQMArIUDAKuVAwCqlQMAnq8AgCKxAIAmsQCAKrEAgC6xAIAysQCANrEAgDqxAIA+sQCAQrEAgEaxAIBKsQCATrEAgFKxAICAHQAAgQkAAIL9AQBWsQCAvwgHAFqxAIBisQCA7yQAAGaxAICElAIAarEAgG6xAICH4AIAhgQFAL4AGABysQCAdrEAgOGQAQB6sQCA44AAAH6xAICCsQCAhrEAgLNlAQCKsQCAtWUBALZtAQCOsQCAkrEAgJaxAIC65QEAu/kBALzpAQC96QEAvsUBAL+9AQCasQCAnrEAgKKxAIC+xBkAprEAgKqxAICusQCA78gBALKxAIDh3A4AtrEAgOMwDgC6sQCAvrEAgMKxAICEMAQAgHkAAIEVAACCFQAAo+UBAMaxAICl5QEApu0BAMqxAICGQAYAh5AHAKplAQCreQEArGkBAK1pAQCuRQEArz0BAKjdBQCpIQYAqiEGAKshBgCsIQYArSEGAK4hBgCvnQYAXrEAgM6xAIDSsQCAhDABANaxAIDasQCA3rEAgOKxAIC4jQYAuZUGALqdBgC7lQYAvI0GAL21BgC+vQYAv7UGALDtBgCx8QYAsvEGALPxBgC0zQYAtbUGALa9BgC3tQYAqIkHAKmVBwCqkQcAq5EHAKy9BwCtpQcArqEHAK/dBwDmsQCA6rEAgO6xAIDysQCA9rEAgPqxAID+sQCAArIAgLhJBwC5VQcAul0HALtVBwC8cQcAvX0HAL5pBwC/aQcAsKUHALGtBwCyuQcAs7EHALSRBwC1kQcAtnkHALd5BwAGsgCACrIAgA6yAIASsgCA78gFAOHACQAWsgCA48AZAOMkBAAasgCA4dAGAO/cKACinQMAoxUBAKAZBQChjQUAs1kGAB6yAIAisgCAJrIAgCqyAIC2ZQYAtXUGAC6yAIC7KQYAuiEGADKyAIA2sgCAvxUGAL4VBgC9JQYAvC0GAKOZBgCPmfwAOrIAgEKyAIBGsgCApqUGAKW1BgBKsgCAq+kGAKrhBgCGKB8Ah5wAAK/VBgCu1QYAreUGAKztBgCebQkAn30HAJwNCwCd7QkAmvENAJs5DQCY5fAAmQ0PAJbh8QCX6fEAlMX1AJUN8wCSHfcAk/H1AJD9+QCR7fkAgh3/AIMB+gBOsgCAUrIAgIYV9gCHOfYAhAn6AIXx9ACKwfAAiyXyAFayAIBasgCAjuEMAI8VDgCMNfIAjQHzAJKtDgCTgQgAXrIAgGKyAICW6QQAl3UGAJR5CgCV8QoAmtEGAJvJAABmsgCAarIAgIEdAwCAHQMAnFkCAIL1AwCrARAAqpUWAKmNFgCojRYAr5UuAK4BLACt/RIArJkSAKOlHgCipR4AoY0CAO6wAICnGRoAppUaAKUBGACknR8AbrIAgHKyAIB2sgCAerIAgH6yAICCsgCAhrIAgIqyAICz5SoAsuUqALGtLwCw5S4AjrIAgJKyAIC1ASQAtBEqAKgpAwCpNQMAqj0DAKs1AwCsLQMArbUDAK69AwCvtQMAlrIAgJqyAICesgCAorIAgIAdAACBCQAAgrkAAKayAIC4TQIAuV0CALptAgC7CQIAvBkCAL0ZAgC+CQIAvwECALDNAwCx1QMAst0DALPVAwC0zQMAtXUCALZ9AgC3dQIAqrIAgITIHQCysgCAvgwfALayAIC6sgCA70gGAO9YBwDhWAYA4ZgGAOOUAQDjAAYAhhAcAId8HQC+9B4AvrIAgMKyAIC2ZQMAtfUDAMayAICz5QMAyrIAgM6yAIDSsgCAv+ECAL5ZAwC9UQMAvFkDALtBAwC6WQMA1rIAgNqyAIA+sgCArrIAgN6yAIDisgCA5rIAgOqyAIDusgCA8rIAgKitHQCptR0AqrUdAKslHgCsPR4ArR0eAK4VHgCvdR4AsA0eALEtHgCyJR4As40eALSVHgC1nR4AtpUeALeNHgC4tR4Aub0eALq1HgC7nR4AvIUeAL1VHwC+XR8Av1UfALMdHQD2sgCA+rIAgP6yAIACswCAtr0eALWVHgAGswCAu8keALrpHgAKswCADrMAgL95HgC+cR4AvXkeALzRHgCCKQAAo1kdAIAdAACBFQAApvkeABKzAIAWswCApdEeAKqtHgCrjR4AGrMAgITgAwCuNR4Arz0eAKyVHgCtPR4AqIkeAKmVHgCqnR4Aq7EeAKzRHgCt2R4Ars0eAK/FHgAeswCAIrMAgIaIAACHbAEAJrMAgCqzAIAuswCAMrMAgLhdAQC5wQEAusEBALvBAQC8wQEAvckBAL7xAQC/8QEAsL0eALGdHgCylR4As2UBALR9AQC1ZQEAtm0BALdlAQCqLR0AqzUdADazAIA6swCAri0dAK+VHACsLR0ArSUdAISMAQCjkR0APrMAgEKzAICmER0ARrMAgEqzAIClgR0As1UeAE6zAIBSswCAVrMAgFqzAIC2GR4AtRkeAF6zAIC7GR4AujkeAGKzAIBmswCAv+EBAL75AQC98QEAvAEeAGqzAIBuswCAcrMAgKOZHQB2swCApdUdAKbVHQB6swCAfrMAgIKzAICq9R0Aq9UdAKzNHQCtPQIArjUCAK8tAgCAZQAAgRUAAIIdAACEAAQAhrMAgIqzAICHcAMAhvwEAJKzAICWswCAmrMAgJ6zAICiswCAprMAgKqzAICuswCAvsgEALKzAIC2swCAurMAgL6zAIDCswCAxrMAgO/cHwDKswCA4ZQBAM6zAIDjHAEA0rMAgNazAIDaswCA3rMAgLt1AwC6aQMAvkgGAOKzAIC/HQMAvh0DAL0dAwC8ZQMAs9UDAOazAIDqswCA7rMAgPKzAIC2fQMAtcUDAIRwBQCoJQIAqTUCAKo9AgCrNQIArC0CAK2dAgCulQIAr7UCAIIVAAD2swCAgNkBAIEJAADEAAAA+rMAgAK0AIAGtACAuKkCALmpAgC6SQEAu0kBALxZAQC9RQEAvkUBAL99AQCwzQIAsdECALLRAgCzqQIAtLkCALW5AgC2qQIAt6ECAOEoHgDhNBwA43QBAOMYHgAKtACADrQAgIa4BACHVAUAhDgHABK0AIAWtACAGrQAgL6sBwAetACA78weAO/IGgCj9QIAIrQAgCa0AIAqtACALrQAgKZdAgCl5QIAMrQAgKtVAgCqSQIANrQAgDq0AICvPQIArj0CAK09AgCsRQIAqGEGAKlhBgCqYQYAq2EGAKxhBgCtYQYArmEGAK9hBgD+swCAPrQAgEK0AIBGtACASrQAgE60AIBStACAVrQAgLjxBgC58QYAuvEGALvxBgC8nQYAvbEGAL6xBgC/sQYAsOUGALHtBgCy5QYAs/0GALTlBgC17QYAttkGALfVBgCz6QYAWrQAgF60AIBitACAZrQAgLbhBgC16QYAarQAgLspBgC6IQYAbrQAgHK0AIC/KQYAviEGAL0pBgC8MQYAgl0AAKOtBgCARQAAgV0AAKalBgB2tACAerQAgKWtBgCqZQYAq20GAIYADACHQAMArmUGAK9tBgCsdQYArW0GAH60AIDvfAUAgrQAgIa0AICKtACAjrQAgJK0AICWtACAmrQAgJ60AICitACAprQAgKq0AIDjaAUArrQAgOF4BQCz0QYAsrQAgLa0AIC6tACAvrQAgLb9BgC1/QYAwrQAgLupBgC6oQYAxrQAgMq0AIC/mQYAvqkGAL2pBgC8sQYAqLkGAKm5BgCqGQYAqxkGAKw1BgCtPQYArjUGAK8pBgDOtACAgh0AAIEdAACAHQAA0rQAgNa0AIDatACA4rQAgLjpAQC56QEAuvkBALv5AQC86QEAvekBAL5dAQC/VQEAsCUGALEtBgCyJQYAsz0GALQtBgC1HQYAthUGALfZAQCGgAwAh+QCAOa0AICjnQUA6rQAgKWxBQCmsQUA7rQAgPK0AID2tACAqu0FAKvlBQCs/QUAreUFAK7lBQCv1QUAtk0DAPq0AICExAMAtUUDAP60AICzjQIAArUAgAa1AIC+SQMAv0kDALxJAwC9SQMAumkDALtpAwAKtQCADrUAgBK1AICmiQMApYEDABa1AICjSQIAGrUAgB61AIAitQCAr40DAK6NAwCtjQMArI0DAKutAwCqrQMAjrMAgCa1AIAqtQCALrUAgIW0PQAytQCANrUAgDq1AIA+tQCAQrUAgIA9AACBCQAAgh0AAEa1AIC+sAMASrUAgIc4AwCG3AwAUrUAgFa1AIBatQCAXrUAgGK1AIDvXAYAZrUAgGq1AIC+6AwA45QGAG61AIDh3AEAcrUAgHa1AIB6tQCAfrUAgLNRAQCCtQCAhrUAgIq1AICOtQCAtnEBALV5AQCStQCAuz0BALo9AQCWtQCAmrUAgL/9AQC+9QEAvQUBALwFAQCetQCAorUAgKa1AICEQAwAqrUAgK61AICytQCA76wHALa1AIDhJAYAurUAgONABwCGkAwAh/wMAMK1AIDGtQCAgFkAAIFlAACCYQAAo90BAMq1AICl9QEApv0BAM61AIDStQCA1rUAgKqxAQCrsQEArIkBAK2JAQCueQEAr3EBAN60AIBOtQCA2rUAgN61AIC+tQCA4rUAgOa1AIDqtQCAqJ0NAKktDgCqOQ4AqzEOAKwRDgCtEQ4Arn0OAK9tDgCwGQ4AsRkOALIxDgCzMQ4AtNEOALXZDgC2zQ4At8UOALj9DgC52Q4AuqkOALupDgC8vQ4AvaUOAL6tDgC/pQ4AqIEPAKmBDwCqgQ8Aq4EPAKyBDwCtjQ8AroUPAK+1DwDutQCA8rUAgPa1AID6tQCA/rUAgAK2AIAGtgCACrYAgLidDwC5rQ8AuqUPALtNDwC8VQ8AvV0PAL5JDwC/SQ8AsNEPALHRDwCy0Q8As9EPALS1DwC1vQ8AtrUPALetDwCzCQ4ADrYAgBK2AIAWtgCAGrYAgLYNDgC1CQ4AHrYAgLsVDgC6FQ4AIrYAgCa2AIC/eQ4AvnEOAL0FDgC8BQ4AghUAAKNNDgCAYQAAgWEAAKZJDgAqtgCAvhABAKVNDgCqUQ4Aq1EOAIQkAQAytgCArjUOAK89DgCsQQ4ArUEOAKg5DgCpOQ4AqlkOAKtRDgCscQ4ArXEOAK6RAQCvkQEAhgAAAIeEAAA2tgCAOrYAgD62AIBCtgCARrYAgEq2AIC4dQEAuX0BALp1AQC7yQAAvNkAAL3ZAAC+yQAAv8EAALD1AQCx/QEAsvUBALNNAQC0VQEAtV0BALZVAQC3TQEAuk0PALtVDwC4TQ8AuUUPAL59DwC/tQ8AvEUPAL11DwCyAQ8AswEPALAxDwCxMQ8AtgEPALcNDwC0EQ8AtREPAKqZDgCrRQ8AqOUOAKmZDgCuQQ8Ar0EPAKxRDwCtUQ8ATrYAgFK2AIBWtgCAWrYAgF62AIBitgCAZrYAgGq2AICzUQ0AbrYAgHK2AIB2tgCAerYAgLZxDQC1eQ0AfrYAgLu5AgC6sQIAgrYAgIa2AIC/GQIAvhECAL0ZAgC8oQIAirYAgKMVDQCOtgCAkrYAgKY1DQCWtgCAmrYAgKU9DQCq9QIAq/0CAIToAwCitgCArlUCAK9dAgCs5QIArV0CAKhtAgCprQIAqqUCAKu9AgCspQIAra0CAK6lAgCvfQEAgO0BAIHxAQCC8QEAvqAFAKa2AICqtgCAh2gFAIYcBQC4yQEAuckBALrZAQC70QEAvPkBAL35AQC+mQEAv5UBALAFAQCxDQEAsgUBALMdAQC0BQEAtQ0BALYFAQC3+QEA4WQPAOGcDwDjFA4A49QPAK62AIDhPA4AsrYAgOPkAAC+rAQAtrYAgLq2AIDvDAAAvrYAgMK2AIDvYA4A77QPAMa2AIDKtgCAhEQEALNhAgDOtgCAtWECALZhAgDStgCA1rYAgNq2AIC6jQEAu4UBALydAQC9hQEAvo0BAL+FAQCjrQUAnrYAgN62AIDitgCA5rYAgKatBQClrQUA6rYAgKtJBgCqQQYA7rYAgPK2AICvSQYArkEGAK1JBgCsUQYA9rYAgPq2AID+tgCAArcAgIAdAACBCQAAgjkAAAa3AIAKtwCADrcAgIbIAACHIAMAErcAgBa3AIAatwCAHrcAgKhtBgCptQcAqr0HAKsdBwCsCQcArTEHAK4xBwCvLQcAhKgDACK3AIAmtwCAKrcAgC63AIAytwCANrcAgDq3AIC4zQAAudUAALrVAAC75QAAvP0AAL2VAAC+nQAAv5UAALBVBwCxJQcAsi0HALM9BwC0LQcAtRUHALYdBwC39QAAPrcAgOG8BgBCtwCA4/QFAEa3AIBKtwCATrcAgFK3AIBWtwCAWrcAgF63AIBitwCAZrcAgGq3AIButwCA7+gEALN1BgCCLQAAgRUAAIAdAABytwCAtvEGALXBBgB2twCAu6EGALrRBgB6twCAvmwBAL+RBgC+qQYAvakGALy5BgCjtQYAgrcAgIYoAACHTAEAhrcAgKYxBgClAQYAircAgKthBgCqEQYAjrcAgJK3AICvUQYArmkGAK1pBgCseQYAlrcAgLO9AQCatwCAnrcAgLZ5AQCitwCAprcAgLV5AQC6VQEAu10BAKq3AICutwCAvvkAAL/lAAC8RQEAvf0AAKhxAgCpcQIAqnECAKtxAgCstQIArb0CAK61AgCvrQIAhOw8ALK3AIC2twCAurcAgL63AIDCtwCAxrcAgMq3AIC4XQMAuWUDALptAwC7ZQMAvH0DAL1lAwC+bQMAv2UDALDVAgCx3QIAstUCALNtAwC0eQMAtWUDALZtAwC3ZQMALrYAgM63AIDStwCAo/UCANa3AIClMQIApjECANq3AIDetwCA4rcAgKodAgCrFQIArA0CAK21AwCusQMAr60DAIBlAACBCQAAghkAAOa3AIDqtwCA8rcAgL4QPAD2twCAhsA8AIcgAwD6twCA/rcAgAK4AIAGuACACrgAgA64AICohQIAqZUCAKqVAgCrpQIArL0CAK3VAgCu0QIAr9ECABK4AIAWuACAGrgAgB64AIAiuACAJrgAgCq4AIAuuACAuHUBALl9AQC6dQEAu8kBALzZAQC9xQEAvsUBAL/9AQCwtQIAsb0CALKBAgCzgQIAtFUBALVdAQC2VQEAt00BAOGkBgAyuACA41AGAL6APACEHDwAvoA/ADa4AIA6uACAPrgAgEK4AIBGuACASrgAgE64AIBSuACA7+AGAFa4AICBfQAAgHEAAFq4AICCBQAAYrgAgGa4AIDvTAAAargAgOGQAQBuuACA41gBAHK4AIB2uACAergAgIZYPwCH/DwAs509AO63AIBeuACAfrgAgIK4AIC21T0AtbU9AIa4AIC7+T0AuvE9AIq4AICOuACAvxk+AL4RPgC91T0AvNU9AJK4AICj2T0AlrgAgJq4AICmkT0AnrgAgKK4AICl8T0AqrU9AKu9PQCmuACAqrgAgK5VPgCvXT4ArJE9AK2RPQCoVT4AqVk+AKphPgCrYT4ArGE+AK1hPgCuYT4Ar2E+AISoAwCuuACAsrgAgLa4AIC6uACAvrgAgMK4AIDGuACAuEU/ALldPwC6VT8Au20/ALx1PwC9fT8AvnU/AL9tPwCwwT8AscE/ALLBPwCzwT8AtME/ALXBPwC2wT8At8E/AIC5AQCBuQEAggUAAMq4AIDhgD4A0rgAgOMoPQDWuACAhoAAAIcEAQDvCD0A2rgAgN64AIDiuACA5rgAgOq4AICzqT8AzrgAgO64AIDyuACA9rgAgLahPwC1qT8A+rgAgLtFPgC6RT4A/rgAgAK5AIC/RT4AvkU+AL1VPgC8VT4Ao2k/AAa5AIAKuQCADrkAgBK5AICmYT8ApWk/ABa5AICrhT4AqoU+ABq5AIAeuQCAr4U+AK6FPgCtlT4ArJU+ACK5AICzGT4AJrkAgCq5AIC2IT4ALrkAgDK5AIC1MT4AuvEBALv5AQA2uQCAOrkAgL6xAQC/vQEAvNEBAL3RAQCo0T0AqdE9AKrVPQCr6T0ArP09AK3lPQCu7T0ArxECAID5AwCBzQMAgsUDAIQkAwC+AAQAQrkAgIesAwCGvAQAuBkCALktAgC6JQIAu+kCALz5AgC9+QIAvukCAL/pAgCwcQIAsXkCALJBAgCzQQIAtDECALU9AgC2NQIAtykCAKVtPQBGuQCASrkAgKZ9PQBOuQCAfrcAgKNFPQBSuQCArY0CAKyNAgCv4QIAru0CAKwAAABWuQCAq6UCAKqtAgDh+AEAWrkAgOP0AgCEwAQAXrkAgGK5AIBmuQCAarkAgG65AIByuQCAdrkAgHq5AIB+uQCAgrkAgO8wAgCGuQCAqBUCAKkZAgCqJQIAqz0CAKwlAgCtLQIAriUCAK9VAgCKuQCAjrkAgJK5AICWuQCAmrkAgJ65AICEsAQAorkAgLjRAgC52QIAuuECALvhAgC8kQIAvZ0CAL6VAgC/iQIAsC0CALE1AgCyNQIAswUCALQdAgC18QIAtvECALfxAgDheD8A4zQBAOMIPgDhbD4AgQkAAICpAACmuQCAgj0AAKq5AICyuQCAtrkAgL4gBAC6uQCA79g+AO/MPgC+uQCAwrkAgLPpAgCG6AQAh8AEALbpAgDGuQCAyrkAgLXpAgC6rQIAu7UCAM65AIDSuQCAvp0CAL9xAgC8pQIAvZUCAD65AICuuQCA1rkAgNq5AIDeuQCA4rkAgOa5AIDquQCAqBUGAKmhBgCqoQYAq70GAKytBgCtgQYArv0GAK/tBgCwlQYAsZ0GALKVBgCzrQYAtLUGALW9BgC2tQYAt60GALiVBgC5mQYAukkHALtJBwC8WQcAvVkHAL5JBwC/SQcArN0FAK3tBQCu5QUArwkFAO65AIDyuQCAqtUFAKvNBQD2uQCApZEFAKaRBQD6uQCA/rkAgAK6AIAGugCAo5EFALNJBgAKugCADroAgBK6AIAWugCAtmEGALVFBgAaugCAuzkGALoxBgC+ZAAAHroAgL8ZBgC+EQYAvRkGALwhBgCjiQcAgtkBAIHZAQCAwQEAIroAgKahBwClhQcAJroAgKv5BwCq8QcAhggBAId8AQCv2QcArtEHAK3ZBwCs4QcAKroAgLP1BgAuugCAMroAgLaFBgA2ugCAOroAgLWdBgC6jQYAu20BAD66AIBCugCAvmUBAL9tAQC8dQEAvW0BAKglBgCpLQYAqjkGAKsxBgCsUQYArUEGAK5BBgCvdQYARroAgEq6AIBOugCAUroAgFa6AIBaugCAXroAgGK6AIC4VQEAuWUBALplAQC7fQEAvGUBAL1tAQC+HQEAvxUBALANBgCx7QEAsuUBALP9AQC05QEAte0BALblAQC3bQEAo7EFAGa6AIBqugCAvkgDAL5YDACmwQUApdkFAG66AICrKQIAqskFAHK6AIB2ugCArykCAK4hAgCtKQIArDECAHq6AIB+ugCAgroAgIa6AICAGQAAgRkAAIIFAACKugCAhKwDAJK6AICHGAMAhswMAJa6AICaugCAnroAgKK6AICokQMAqZkDAKrJAwCrxQMArN0DAK3BAwCuwQMAr/UDAKa6AICqugCArroAgLK6AIC2ugCAuroAgL66AIDCugCAuH0DALnBAAC6wQAAu9EAALz5AAC9+QAAvpkAAL+ZAACwjQMAsUUDALJNAwCzRQMAtF0DALVFAwC2TQMAt0UDALNBAgDGugCAyroAgL8EDwDOugCAtkECALVVAgDSugCAu4ECALpJAgDWugCA2roAgL+BAgC+mQIAvZECALyZAgDeugCA4roAgOa6AIDqugCA76QDAO66AIDyugCA9roAgOMQAwD6ugCA4VgAAIQgDQCAKQAAgSkAAIIdAAACuwCA4VAGAOGgBwDjoAYA41AHAIWUDAAGuwCA70gbAAq7AIDhJAIADrsAgONwGgASuwCAFrsAgBq7AIDvqAEA7+gGAIagDwCHDA0Ao4kCAB67AIClnQIAIrsAgCa7AICmiQIAKrsAgC67AICrSQIAqoECAK1ZAgCsUQIAr0kCAK5RAgCoZQ4AqXUOAKp9DgCrdQ4ArG0OAK21DgCuvQ4Ar7UOAP66AIAyuwCANrsAgDq7AIA+uwCASbsAgE27AIBRuwCAuF0PALltDwC6ZQ8Auw0PALwVDwC9HQ8AvhUPAL8JDwCwzQ4AsdUOALLdDgCz1Q4AtM0OALVxDwC2cQ8At20PALP1DgBVuwCAWbsAgF27AIBhuwCAtjUOALXlDgBluwCAuxEOALoJDgBpuwCAbbsAgL+1DwC+CQ4AvQEOALwJDgCCFQAAo7EOAIBhAACBYQAApnEOAHG7AIC+EAEApaEOAKpNDgCrVQ4AebsAgIQgAQCuTQ4Ar/EPAKxNDgCtRQ4An0UIAJ4NCQCdDQkAnJkLAJt1NQCaETUAmZk3AJgNMQCXJTEAliUxAJWBPQCUDT0Ak4k/AJIVOACRPTkAkD05AI9lJQDvrA0AhgAEAIegAQB9uwCAgbsAgIW7AIDv6AEAibsAgOE0AgCNuwCA4zQBAJG7AIDjCAwAlbsAgOEIDQChoQEAmbsAgKMJBQCibQMApc0EAKQRBQCnHRkAph0ZAKmhHQCoORkAq+kcAKqpHQCtkREArAEQAK8BFACuUREAsfkVALDlFQCz6WkAsgFoALUBbAC0eWkAnbsAgKG7AICluwCAqbsAgK27AICxuwCAowkDAKIZDQCh/Q0AoP0NAIIlJgCDBToAtbsAgLm7AICGqTwAhzU+AIQdOgCFPTsAiok+AIslMgC9uwCAwbsAgI6xNACPMTYAjD0yAI0tMgCSJTYAk9EIAIREAwC+wAQAlhULAJdVDgCUXQoAlVUKAJplDgCbiQ4AxbsAgMm7AIDNuwCA0bsAgJyBAADVuwCAuLUCALm9AgC6tQIAuwkCALwZAgC9GQIAvgkCAL8BAgCwdQ0AsX0NALJJDQCzSQ0AtJUCALWdAgC2lQIAt40CAKi9DQCpUQ0AqlUNAKtpDQCsfQ0ArWUNAK5tDQCvEQ0AdbsAgILtAQCBHQAAgB0AANm7AIDduwCAjroAgL5wBQCznQwAhIwFAOG7AIDpuwCA7bsAgLalDAC1tQwA8bsAgLv5DAC68QwAhigFAIcgBQC/GQMAvhEDAL3dDAC83QwA9bsAgKPZDAD5uwCA/bsAgKbhDAABvACABbwAgKXxDACqtQwAq70MAAm8AIANvACArlUDAK9dAwCsmQwArZkMABG8AIAVvACAGbwAgB28AIAhvACAJbwAgCm8AIDvvAEALbwAgOF8DgAxvACA41ABADW8AIA5vACAPbwAgEG8AICzlQIARbwAgEm8AIBNvACAUbwAgLa9AgC1uQIAWbwAgLs5AgC6YQIAhsgEAIesBAC/GQIAvhECAL0ZAgC8IQIAo1UFAILVBwCBxQcAgMUHAF28AICmfQUApXkFAGG8AICr+QUAqqEFAGW8AIBpvACAr9kFAK7RBQCt2QUArOEFAG28AICzWQcAcbwAgHW8AIC2HQcAebwAgH28AIC1FQcAugkHALsJBwCBvACAhbwAgL75BwC/+QcAvPkHAL35BwDluwCAVbwAgIm8AICNvACAkbwAgJW8AICZvACAnbwAgKitBwCptQcAqrUHAKvtBwCs+QcArfkHAK7tBwCv5QcAsKkHALGpBwCySQcAs0kHALRZBwC1WQcAtkkHALdJBwC4eQcAuUUHALpBBwC7XQcAvEUHAL1NBwC+RQcAvzkHAKMdBgChvACApbwAgKm8AICtvACAplkGAKVRBgCxvACAq00GAKpNBgC1vACAubwAgK+9BgCuvQYArb0GAKy9BgCAbQAAgQkAAIIZAAC9vACAwbwAgISYAQC+kAEAxbwAgIYAHACHxAEAybwAgM28AIDRvACA1bwAgNm8AIDdvACAqF0GAKmVAQCqlQEAq6UBAKy9AQCt1QEArtEBAK/RAQDhvACA5bwAgOm8AIDtvACA8bwAgPW8AID5vACA/bwAgLhZAQC5WQEAus0AALvFAAC83QAAvcUAAL7FAAC/9QAAsLUBALG9AQCygQEAs4EBALR5AQC1eQEAtmkBALdpAQCzHQIAAb0AgAW9AIC+gBwACb0AgLZVAgC1NQIADb0AgLt5AgC6cQIAEb0AgBW9AIC/vQIAvr0CAL1VAgC8VQIAGb0AgKNZAgAdvQCAIb0AgKYRAgAlvQCAKb0AgKVxAgCqNQIAqz0CAC29AIAxvQCArvkCAK/5AgCsEQIArRECADm9AIA9vQCAvgQdAL4AHgBBvQCARb0AgEm9AIBNvQCAgPkAAIHNAACCxQAAhCADAIawHACHlAMAUb0AgFW9AIBZvQCAXb0AgGG9AIBlvQCA42wCAGm9AIDhoAEAbb0AgO8UAgBxvQCAdb0AgHm9AIB9vQCAgb0AgIW9AICJvQCA4fAGAOE0BgDjTAAA4xgGAI29AICRvQCAlb0AgJm9AICAPQAAgQkAAIIZAACdvQCAob0AgIS8HQDvmAAA7zgHALMxAgDRAAAAh9gdAIZsHACpvQCAtikCALUhAgCtvQCAu80CALrNAgCxvQCAtb0AgL/NAgC+zQIAvc0CALzNAgCyXQYAs2UGALANBgCxVQYAtn0GALedBQC0fQYAtXUGALqNBQC7zQUAuKUFALmFBQC+xQUAv8kFALzVBQC9zQUAub0AgL29AIDBvQCAxb0AgMm9AIDNvQCA0b0AgNW9AICqtQYAq70GAKgBBwCpvQYAroEGAK+NBgCsmQYArZUGAKNxHQDZvQCA3b0AgOG9AIDlvQCApmkdAKVhHQDpvQCAq40dAKqNHQDtvQCA8b0AgK+NHQCujR0ArY0dAKyNHQD1vQCAs9UeAPm9AID9vQCAts0eAAG+AIAFvgCAtcUeALqhHgC7oR4ACb4AgA2+AIC+pR4Av6keALyxHgC9sR4ANb0AgKW9AIARvgCAhAQDAID5AACB+QAAghEAABW+AICoIR4AqSEeAKo5HgCrOR4ArCkeAK0pHgCuAR4ArwEeALABHgCxAR4AsgEeALMBHgC0BR4AtQkeALY9HgC3NR4AuA0eALkVHgC6HR4AuxUeALwNHgC95R8Avu0fAL/lHwCjkR8AGb4AgIYoAQCHSAEAHb4AgKaJHwClgR8AIb4AgKvlHwCq5R8AJb4AgCm+AICv7R8AruEfAK31HwCs9R8ALb4AgLMtHgAxvgCANb4AgLaVHgA5vgCAPb4AgLWdHgC6sR4Au7EeAEG+AIBFvgCAvnUBAL99AQC8oR4AvaEeAKjRHgCp2R4AquEeAKvhHgCsUR4ArVEeAK5RHgCvUR4ASb4AgE2+AIBRvgCAVb4AgFm+AIBdvgCAYb4AgGW+AIC43QEAue0BALrlAQC7jQEAvJkBAL2ZAQC+jQEAv4UBALAxHgCxMR4AsjEeALMxHgC09QEAtf0BALb1AQC37QEAo2kdAGm+AIBtvgCAcb4AgHW+AICm0R0ApdkdAHm+AICr9R0AqvUdAH2+AICBvgCArzkCAK4xAgCt5R0ArOUdAIFpAACAWQAAvgAEAIJhAACJvgCAjb4AgJG+AICVvgCAhOwDAJm+AICHiAMAhuwEAJ2+AIChvgCApb4AgKm+AICohQMAqZUDAKqVAwCrpQMArL0DAK3VAwCu0QMAr9EDAK2+AICxvgCAtb4AgLm+AIC9vgCAwb4AgMW+AIDJvgCAuHEDALlxAwC6cQMAu3EDALzVAAC93QAAvtUAAL/NAACwtQMAsb0DALKBAwCzgQMAtFEDALVRAwC2UQMAt1EDAOFUHgDhrB8A45QBAOMoHgDjYAMAzb4AgOEIAADRvgCA75ADANW+AIDZvgCA3b4AgOG+AIDlvgCA70wfAO9MHwCzXQIA6b4AgO2+AIDxvgCA+b4AgLYVAgC1dQIA/b4AgLs5AgC6MQIAhCQFAL7gBAC/1QIAvtUCAL0VAgC8FQIAuJEdALmZHQC6oR0Au6EdALzRHQC93R0AvtUdAL/JHQCwCR4AsQkeALIZHgCzGR4AtAkeALUJHgC2vR0At7UdAKipHgCpqR4AqrkeAKu5HgCsqR4ArakeAK55HgCveR4AgKUAAIGtAACCpQAAAb8AgIbQBACH+AQABb8AgAm/AICFvgCA9b4AgA2/AIARvwCAFb8AgBm/AIAdvwCAIb8AgKhxBgCpcQYAqnEGAKtxBgCsVQYArUUGAK5NBgCvRQYAsD0GALHlBgCy7QYAs+UGALT9BgC15QYAtu0GALflBgC43QYAuXEHALp1BwC7SQcAvFkHAL1ZBwC+SQcAv0kHALPZBgAlvwCAKb8AgC2/AIAxvwCAtuUGALX9BgA1vwCAuwEGALrZBgA5vwCAPb8AgL8BBgC+GQYAvREGALwZBgBBvwCAo9kFAEW/AIBJvwCAppEFAE2/AIBRvwCApfEFAKq1BQCrvQUAVb8AgFm/AICuUQUAr1EFAKyRBQCtkQUAo1kHAIIZAACBGQAAgOEBAF2/AICmZQcApX0HAGG/AICrgQcAqlkHAISgAgC+rAEAr4EHAK6ZBwCtkQcArJkHAGW/AICzqQYAhugAAIcsAQC2WQEAab8AgG2/AIC1oQYAunUBALt9AQBxvwCAdb8AgL75AQC/+QEAvGUBAL35AQCo0QYAqdkGAKplBgCrdQYArG0GAK2dAQCulQEAr40BAITsAQB5vwCAfb8AgIG/AICFvwCAib8AgI2/AICRvwCAuGkBALlpAQC6CQEAuwUBALwdAQC9AQEAvgEBAL81AQCw9QEAsf0BALL1AQCzaQEAtHkBALV5AQC2aQEAt2EBAJW/AICZvwCAnb8AgKPhBQChvwCApekFAKYRAgClvwCAqb8AgK2/AICqPQIAqzUCAKwtAgCtsQIArrECAK+xAgCxvwCAtb8AgL4EAwCEAAwAub8AgL2/AIDBvwCAxb8AgIANAACBFQAAgh0AAMm/AIDNvwCA0b8AgIdEAwCG3AwAs+kDANm/AIDdvwCA4b8AgOW/AIC2PQMAtT0DAOm/AIC7GQMAuhEDAO2/AIDxvwCAv7kAAL6xAAC9uQAAvAEDAPW/AIDhlAEA+b8AgON8AQD9vwCAAcAAgAXAAIAJwACADcAAgBHAAIAVwACAGcAAgB3AAIAhwACAJcAAgO9MAgCoVQIAqV0CAKphAgCrYQIArLUCAK29AgCutQIAr60CAL5oDQApwACALcAAgDHAAIA1wACAgq0AAIGtAACArQAAuGEBALlhAQC6CQEAuwkBALwBAQC9AQEAvgEBAL8BAQCw1QIAsd0CALLVAgCzbQEAtHUBALV9AQC2aQEAt2EBAOFoBgDh8AcA47AAAOP0BgA5wACAPcAAgEHAAIBJwACATcAAgFHAAIBVwACAWcAAgL78DABdwACA72wAAO8oBgCjqQIAYcAAgIZoDACHBA0AZcAAgKZ9AgClfQIAacAAgKtZAgCqUQIAbcAAgHHAAICv+QEArvEBAK35AQCsQQIAqIUOAKmNDgCqhQ4Aq50OAKyNDgCtvQ4ArrUOAK/dDgBFwACAdcAAgHnAAIB9wACAgcAAgIXAAICJwACAjcAAgLitDgC5tQ4Aur0OALu1DgC8dQ8AvX0PAL51DwC/bQ8AsKkOALG1DgCyvQ4As7UOALStDgC1lQ4Atp0OALeVDgCzDQ4AkcAAgJXAAICZwACAncAAgLY9DgC1BQ4AocAAgLtxDgC6bQ4ApcAAgKnAAIC/UQ4AvmkOAL1hDgC8aQ4AghkAAKNJDgCAZQAAgRkAAKZ5DgCtwACAscAAgKVBDgCqKQ4AqzUOAIS8AwC1wACAri0OAK8VDgCsLQ4ArSUOAKidDgCppQ4Aqq0OAKulDgCsvQ4AraEOAK7dDgCvzQ4AhiABAIdkAQC5wACAvcAAgMHAAIDFwACAycAAgM3AAIC4eQEAuXkBALrNAQC7xQEAvN0BAL3FAQC+xQEAv/UBALC9DgCxjQ4AsoUOALNJAQC0WQEAtVkBALZJAQC3SQEAtS0OANHAAIDVwACAtjkOANnAAIDdwACAsz0OAOHAAIC9hQEAvEkOAL+FAQC+hQEA5cAAgNW/AIC7UQ4AumEOAKNlDgDpwACA7cAAgPHAAID1wACApmEOAKV1DgD5wACAqwkOAKo5DgD9wACAAcEAgK/dAQCu3QEArd0BAKwRDgAFwQCACcEAgO/QDwANwQCAEcEAgBXBAIAZwQCAHcEAgCHBAIC+aAMAKcEAgC3BAIDhVA4AMcEAgONkDgA1wQCAgFkAAIFZAACCaQAAhIwDAIbwBACHFAMAOcEAgD3BAIBBwQCARcEAgEnBAIBNwQCAUcEAgFXBAIBZwQCAXcEAgGHBAIBlwQCAacEAgG3BAIBxwQCAdcEAgHnBAIB9wQCAqIkDAKmJAwCqmQMAq5kDAKyJAwCtiQMArj0DAK81AwCwUQMAsVEDALJVAwCzfQMAtBUDALUdAwC2FQMAtw0DALg9AwC5DQMAugUDALvtAAC89QAAvfkAAL7pAAC/6QAAgcEAgIXBAICJwQCAsz0CAI3BAIC1LQIAtiUCAJHBAIC+aAUAmcEAgLq5AgC7uQIAvK0CAL2FAgC+/QIAv/UCAIBJAACBVQAAglUAAIQABQDvjAMAvhgEAId0BQCG/AQA4zwDAJ3BAIDhUAAAocEAgKXBAICpwQCArcEAgLHBAIC1wQCAucEAgL3BAIDBwQCAxcEAgMnBAIDNwQCA79QOAL4oBgDhdA4A0cEAgONUAQDVwQCA2cEAgN3BAIDhwQCAo/ECAOXBAIDpwQCA7cEAgPHBAICm6QIApeECAPXBAICrdQIAqnUCAPnBAID9wQCArzkCAK4xAgCtSQIArGECAKgpBgCpKQYAqj0GAKsxBgCsSQYArUkGAK55BgCveQYAlcEAgIIVAACBxQcAgMUHAAHCAICEaAMABcIAgAnCAIC4yQYAuckGALrZBgC72QYAvMkGAL3JBgC+WQcAv1kHALAJBgCxCQYAshkGALMZBgC0CQYAtQkGALb5BgC3+QYAs7UGAA3CAICGrAAAh0ADABHCAIC2yQYAtcEGABXCAIC7zQYAus0GABnCAIAdwgCAv80GAL7NBgC9zQYAvM0GACHCAICj8QYAJcIAgCnCAICmjQYALcIAgDHCAIClhQYAqokGAKuJBgA1wgCAOcIAgK6JBgCviQYArIkGAK2JBgCoJQYAqWEGAKplBgCrfQYArGUGAK1tBgCuZQYAr50GAD3CAIBBwgCARcIAgEnCAIBNwgCAUcIAgFXCAIBZwgCAuPUGALn9BgC69QYAu4kGALyZBgC9mQYAvokGAL+BBgCw5QYAse0GALLlBgCz/QYAtOUGALXtBgC20QYAt80GAF3CAIC2/QYAtf0GAGHCAICz/QYAZcIAgGnCAIBtwgCAvzkGAL4xBgC9OQYAvCEGALs5BgC6MQYAJcEAgHHCAICjrQYAgnkAAIFVAACAVQAAhFwBAKatBgClrQYAecIAgKtpBgCqYQYAhkh/AIfkAACvaQYArmEGAK1pBgCscQYAfcIAgO/cBwCBwgCAhcIAgInCAICNwgCAkcIAgJXCAICZwgCAhKADAJ3CAIC/JHkAocIAgONoBwClwgCA4XQGALPRAgCpwgCAvgQDAISAfQCtwgCAtvkCALXxAgCxwgCAu7UCALqpAgC1wgCAucIAgL9RAwC+mQIAvZECALylAgCpBQIAqLkCAKsVAgCqHQIArT0CAKw9AgCvUQIArl0CAL5ofQC9wgCAwcIAgMXCAIDJwgCAzcIAgNHCAIDVwgCAufEDALjpAwC78QMAuvkDAL1RAwC86QMAv00DAL5RAwCxNQIAsCkCALMBAgCyNQIAtdEDALQZAgC30QMAttkDAIIpAACjlQMAgB0AAIEVAACmvQMA2cIAgN3CAICltQMAqu0DAKvxAwDhwgCA6cIAgK7dAwCvFQIArOEDAK3VAwCGYH0Ah3h9ALNBAQCEAH8AtUEBAO3CAIDxwgCAtkkBAPXCAID5wgCAu0EBALpNAQC9SQEAvEUBAL8pAQC+OQEA/cIAgO/cBgABwwCABcMAgAnDAIANwwCAEcMAgO8wBgCELH4A4eAGABXDAIDjiAEAGcMAgON0AAAdwwCA4SwBAKPJAQAhwwCAJcMAgIVweQApwwCApsEBAKXJAQAtwwCAq8kBAKrFAQAxwwCANcMAgK+hAQCusQEArcEBAKzNAQCo3X0AqQV+AKoBfgCrAX4ArAF+AK0BfgCuAX4ArwF+AOXCAIA5wwCAPcMAgEHDAIBFwwCAgp0AAIGdAACAnQAAuC1+ALnhfgC64X4Au+F+ALzhfgC94X4AvuF+AL/hfgCwQX4AsU1+ALJZfgCzVX4AtDV+ALUlfgC2JX4AtxV+AKitfwCp0X8AqtF/AKvtfwCs9X8ArRV/AK4RfwCvEX8AScMAgE3DAIBRwwCAVcMAgIbwAwCHuAAAWcMAgF3DAIC4EX8AuRl/ALohfwC7IX8AvPUAAL39AAC+9QAAv+0AALBxfwCxcX8AsnF/ALNFfwC0QX8AtU1/ALY9fwC3NX8As1l+AGHDAIBlwwCAacMAgG3DAIC2lX4AtX1+AHHDAIC7tX4AurV+AHXDAIB5wwCAv4l+AL6FfgC9kX4AvKV+AH3DAICjHX4AgcMAgIXDAICm0X4AicMAgI3DAIClOX4AqvF+AKvxfgCRwwCAlcMAgK7BfgCvzX4ArOF+AK3VfgCwrQAAscUAALLBAACzwQAAtMUAALXNAAC28QAAt/EAALhhAAC5YQAAumEAALt9AAC8ZQAAvW0AAL5lAAC/vQMAmcMAgJ3DAIChwwCAdcIAgKXDAICpwwCArcMAgLHDAICoWQEAqVkBAKrtAACr5QAArP0AAK3lAACu5QAAr9UAALXDAICCHQAAgR0AAIAdAAC5wwCAvcMAgMHDAIC+VAIAhoAEAIfsAgDJwwCAzcMAgNHDAIDVwwCA2cMAgL54AwDjdH4A3cMAgOG4fQDhwwCA5cMAgOnDAIDtwwCA8cMAgPXDAID5wwCA/cMAgAHEAIDvwH4ABcQAgAnEAIANxACAs4UDABHEAIAVxACAGcQAgB3EAIC2hQMAtZUDACHEAIC74QMAuokDAL4kBgAlxACAv+kDAL7hAwC99QMAvPUDAIIpAACjwQMAgB0AAIEVAACmwQMAKcQAgC3EAICl0QMAqs0DAKulAwAxxACAheAFAK6lAwCvrQMArLEDAK2xAwDh+AMAOcQAgONcHwA9xACA7/QDAEHEAICGPAcAh6wCAON8fgBFxACA4YABAEnEAIBNxACAUcQAgO/kEwBVxACAs3EBAFnEAIBdxACAYcQAgGXEAIC2EQEAtWEBAGnEAIC7OQEAujEBAG3EAIBxxACAvxkBAL4RAQC9GQEAvCEBAHXEAIB5xACAfcQAgIHEAICFxACAicQAgI3EAIDvxH8AkcQAgOH8fgCVxACA4/B/AIANAACBdQAAgn0AAJnEAICdxACAocQAgKP5AQC+AAgApekBAKnEAICtxACAppkBAISoBQCxxACAq7EBAKq5AQCtkQEArKkBAK+RAQCumQEAqCkGAKkpBgCqOQYAqzkGAKwpBgCtUQYArlUGAK9NBgA1xACAhCABALXEAIClxACAo+EBAKKZBAChGQQAoPEFALg5BgC5OQYAus0GALvFBgC83QYAvcUGAL7FBgC/8QYAsDUGALE9BgCyNQYAsw0GALQVBgC1HQYAthUGALcJBgCPoWwAs5EHAIYoAQCHfAMAtqEHALnEAIC9xACAtbEHALrlBwC77QcAwcQAgMXEAIC+7QcAv90HALz1BwC97QcAn/l4AJ7leACdcXkAnCF8AJvxfACaYX0AmZlxAJjZcACX4XAAlnl0AJVtdACUbXQAk61pAJJxaACReWgAkB1uAIIhbQCD5W8AycQAgM3EAICGTWgAh5V1AISZaQCFmWkAiqV1AIu5dQDRxACA1cQAgI5xcACPgXwAjDlxAI05cQCSYX0Ak6l9ANnEAIDdxACAlml5AJeZBACU4XgAlX15AJpBBQCbyQUA4cQAgOXEAIDpxACA7cQAgJypAADxxACAo4ENAKKpAQChqQEA9cQAgKexCQCmAQgApU0NAKSZDQCrkRUAqoUVAKkBFACocQkArx0QAK7pEQCtvREArAEQALMBGACy8RwAscEdALDJHQDFwwCA+cQAgLXhGAC0/RkA/cQAgAHFAIAFxQCACcUAgIAdAACBCQAAgv0DAA3FAICjFQUAEcUAgIaIDACHPAMAGcUAgKYlBQClNQUAHcUAgKtpBQCqYQUAIcUAgCXFAICvWQUArmkFAK1pBQCscQUAKcUAgC3FAICEBAwAMcUAgDXFAIDhbAYAOcUAgOPsewA9xQCAQcUAgEXFAIDvqAYAScUAgE3FAIBRxQCAVcUAgKmNBQCogQUAq60FAKqZBQCtoQUArLkFAK+lBQCuqQUAhGgNAFnFAIBdxQCAYcUAgGXFAIBpxQCAbcUAgL70DAC5SQUAuEEFALtZBQC6QQUAvUkFALxBBQC/cQUAvn0FALGpBQCwoQUAs7kFALKhBQC1mQUAtKkFALd5BQC2kQUAqNUEAKndBACq7QQAqyUDAKyFAwCtjQMArrEDAK+xAwBxxQCAdcUAgHnFAIB9xQCAgBkAAIEZAACCBQAAgcUAgLgxAgC5MQIAujUCALvBAgC8hQIAvbUCAL69AgC/tQIAsGkCALFpAgCyQQIAs0ECALQ5AgC1OQIAthECALcRAgCGoAwAh0wNAInFAICNxQCA76QGAJHFAICVxQCA78wHAOOUAQDhpAYA4TgBAONcBgCZxQCAncUAgKHFAIClxQCAqcUAgK3FAICzLQQAscUAgLVFAwC1xQCAucUAgLZFAwC9xQCAwcUAgLvlAgC65QIAvd0CALzdAgC/tQIAvrUCABXFAICFxQCAxcUAgMnFAIDNxQCA0cUAgNXFAIDZxQCAqDEOAKk5DgCqAQ4AqwEOAKxxDgCtcQ4ArnUOAK9tDgCwGQ4AsSUOALItDgCzJQ4AtCEOALUhDgC2IQ4AtyEOALjFDgC5zQ4AusUOALvdDgC8xQ4Avc0OAL5ZDwC/WQ8As6kOAN3FAIDhxQCA5cUAgOnFAIC20Q4AtdkOAO3FAIC7wQ4Auv0OAPHFAIC+LAAAv8UOAL7FDgC90Q4AvNkOAIJpAACj7Q4AgFkAAIFRAACmlQ4A9cUAgPnFAIClnQ4AqrkOAKuFDgCGyAAAh6wAAK6BDgCvgQ4ArJ0OAK2VDgD9xQCAs5EOAAHGAIAFxgCAtqUOAAnGAIANxgCAta0OALrhDgC74Q4AEcYAgBXGAIC+6Q4Av9UOALz1DgC96Q4Ao6UKABnGAIAdxgCAIcYAgCXGAICmzQ0Apc0NACnGAICrbQwAqm0MAC3GAIAxxgCArz0MAK49DACtVQwArFUMAKgJDgCpCQ4Aqh0OAKsVDgCsIQ4ArSEOAK4hDgCvIQ4ANcYAgDnGAIA9xgCAQcYAgEXGAIBJxgCATcYAgFHGAIC4zQEAudUBALrdAQC71QEAvM0BAL1RAQC+UQEAv1EBALAhDgCxIQ4AsiUOALM5DgC0KQ4AtRUOALYdDgC39QEAVcYAgFnGAIBdxgCAo5kNAGHGAIClpQ0Apq0NAL7cAgCE7AMAacYAgKrpDQCr6Q0ArP0NAK3hDQCu4Q0Ar90NAIBFAACBTQAAglkAAKNFAwBtxgCApUEDAKZBAwBxxgCAhsAEAIcAAwCqLQMAqyUDAKw9AwCtJQMAriUDAK8VAwCoWQIAqYUDAKqBAwCrgQMArIUDAK2NAwCusQMAr7EDAHXGAIB5xgCAfcYAgIHGAICFxgCAicYAgI3GAICRxgCAuGUDALltAwC6ZQMAu30DALxlAwC9bQMAvmUDAL/dAACwpQMAsa0DALKlAwCzvQMAtK0DALWdAwC2lQMAt10DALMJAgCVxgCAmcYAgJ3GAIChxgCAtg0CALUNAgClxgCAu2kCALphAgCpxgCArcYAgL9ZAgC+aQIAvWkCALxxAgCxxgCAtcYAgLnGAIC9xgCA4aABAMHGAIDjaAMAxcYAgIEVAACAFQAA74wDAIIVAADJxgCAzcYAgNHGAIC+cAUA4RgOAOGUDwDjOA8A49QPAISUAgDZxgCA3cYAgOHGAIDlxgCA6cYAgO3GAIDxxgCA9cYAgPnGAIDv7AEA7/gPAIZgBACHBAUAs5UBAITMBQC1dQEA/cYAgAHHAIC2dQEABccAgAnHAIC7UQEAulkBAL31AAC8SQEAv/UAAL71AACoJQYAqVUGAKpVBgCrrQYArLUGAK29BgCutQYAr60GANXGAIANxwCAEccAgBXHAIAZxwCAHccAgCHHAIAlxwCAuGkHALlpBwC6CQcAuwkHALwZBwC9GQcAvg0HAL8BBwCw1QYAsd0GALLVBgCzaQcAtHkHALV5BwC2aQcAt2EHAKPdBgApxwCALccAgDHHAIA1xwCApj0GAKU9BgA5xwCAqxkGAKoRBgA9xwCAQccAgK+9BwCuvQcArb0HAKwBBgCAXQAAgW0AAIJlAACzUQcAvtgDALVxBwC2cQcARccAgIbgAACHFAMAul0HALs5BwC8KQcAvRUHAL4dBwC/2QAAqJUGAKmdBgCqlQYAq60GAKy1BgCtvQYArrUGAK+tBgBJxwCATccAgFHHAIBVxwCAWccAgF3HAIBhxwCAZccAgLhxAQC5cQEAunEBALtxAQC81QEAvd0BAL7VAQC/zQEAsNUGALGxBgCysQYAs40GALSVBgC1UQEAtlEBALdRAQBpxwCAoxkGAG3HAIBxxwCApjkGAGXGAIB1xwCApTkGAKoVBgCrcQYAeccAgH3HAICuVQYAr5EBAKxhBgCtXQYAgccAgIXHAICJxwCAjccAgJHHAICVxwCAmccAgJ3HAIChxwCApccAgKnHAICtxwCAgBkAAIEZAACCBQAAsccAgISAAgC+gAMAhwwDAIasHADhaAYAuccAgOOYBwC9xwCAwccAgMXHAIDvrAcAyccAgM3HAIDRxwCA1ccAgNnHAIDdxwCA4ccAgOXHAICzZQMA6ccAgLVlAwC2bQMA7ccAgPHHAID1xwCAuukDALvlAwC8/QMAve0DAL7RAwC/0QMA+ccAgP3HAIAByACABcgAgAnIAIANyACAEcgAgBXIAICogQMAqYEDAKqBAwCrgQMArIEDAK2BAwCugQMAr4EDALBBAwCxTQMAskUDALNVAwC0eQMAtXkDALYZAwC3GQMAuCkDALkpAwC6OQMAuzkDALwpAwC9KQMAvhkDAL8ZAwCBGQAAgBEAAKMhAgCCLQAApSECABnIAIAdyACApikCACHIAIApyACAq6ECAKqtAgCtqQIArLkCAK+VAgCulQIAhEwCAL5IHQCHZB0AhuwcAONAAwAtyACA4aABADHIAIDvnAMANcgAgDnIAIA9yACAQcgAgEXIAIBJyACATcgAgFHIAIBVyACAWcgAgF3IAIBhyACAZcgAgGnIAIDvtAEAhKgdAOF8BgBtyACA43AGAHHIAIB1yACAecgAgH3IAICz4QEAgcgAgIXIAICJyACAjcgAgLblAQC19QEAkcgAgLuhAQC62QEAvuQcAJnIAIC/rQEAvqUBAL2xAQC8uQEAqBUeAKkZHgCqKR4AqykeAKw9HgCtJR4Ari0eAK8lHgAlyACAgvkfAIH5HwCA4R8AlcgAgJ3IAICGHAAAh7ADALjBHgC5wR4AusEeALvBHgC8wR4AvcEeAL7BHgC/wR4AsF0eALElHgCyLR4AsyUeALQhHgC1KR4AthkeALcZHgCjoR4AocgAgKXIAICpyACArcgAgKalHgCltR4AscgAgKvhHgCqmR4AtcgAgLnIAICv7R4AruUeAK3xHgCs+R4AvcgAgLOZHwDByACAxcgAgLa9HwDJyACAzcgAgLW1HwC6mR8Au5kfANHIAIDVyACAvnkfAL95HwC8eR8AvXkfAKglHgCpUR4AqlUeAKtpHgCseR4ArXkeAK5pHgCvaR4A2cgAgN3IAIDhyACA5cgAgOnIAIDtyACA8cgAgPXIAIC42R4Aue0eALr5HgC7+R4AvOkeAL3pHgC+nR4Av5UeALAZHgCxGR4AsukeALPpHgC0+R4AtfkeALbpHgC36R4Ao90eAIIpAACBFQAAgB0AAPnIAICm+R4ApfEeAP3IAICr3R4Aqt0eALXHAIAByQCArz0eAK49HgCtPR4ArD0eAITIAgCzQQEAvgwBAAnJAIC2QQEADckAgBHJAIC1UQEAuk0BALslAQCGSAAAh1ABAL4lAQC/LQEAvDEBAL0xAQAVyQCAGckAgIQEAwC+gAQAHckAgO+oHwAhyQCAJckAgL8oMQDjdB8AKckAgOE4HgAtyQCAMckAgDXJAIA5yQCAPckAgEHJAICjzQIARckAgKXdAgBJyQCATckAgKbNAgBRyQCAVckAgKupAgCqwQIArb0CAKy9AgCvoQIArqkCAKm1AgCoaR0AqwECAKoJAgCtAQIArBkCAK8xAgCuAQIAhGwFAFnJAIBdyQCAYckAgGXJAICCnQEAgZ0BAICdAQC55QMAuOUDALvlAwC65QMAveUDALzlAwC/5QMAvuUDALEhAgCwSQIAsyUCALIlAgC1KQIAtCECALcVAgC2FQIAqM0CAKnRAgCq0QIAqw0BAKwVAQCtBQEArgEBAK8BAQBpyQCAbckAgHHJAIB5yQCAvvgEAH3JAICByQCAhckAgLgVAQC5HQEAuikBALspAQC89QEAvf0BAL71AQC/7QEAsEkBALFVAQCyXQEAs1UBALRNAQC1NQEAtj0BALcxAQCGoAUAh8gFAInJAIDvvAAAjckAgJHJAICVyQCA74weAIQsBwDh8B4AmckAgOMcHgCdyQCA4ZQBAKHJAIDjbAAAsxkCAKXJAICpyQCArckAgIQACAC2xQEAtd0BALHJAIC70QEAus0BALXJAIC5yQCAv7EBAL7JAQC9wQEAvMkBAKPZBQB1yQCAvckAgMHJAIDFyQCApgUGAKUdBgDJyQCAqxEGAKoNBgDNyQCA0ckAgK9xBgCuCQYArQEGAKwJBgDVyQCAgh0AAIEdAACAHQAA2ckAgN3JAIDhyQCA5ckAgIZAAwCHxAMA6ckAgO3JAIDxyQCA9ckAgPnJAID9yQCAqK0HAKmxBwCqsQcAq7EHAKwZBwCtBQcArg0HAK8FBwABygCABcoAgAnKAIANygCAEcoAgBXKAIAZygCAHcoAgLgtBwC5zQAAusUAALvdAAC8zQAAvf0AAL71AAC/nQAAsEkHALFVBwCyUQcAsykHALQ5BwC1OQcAtiUHALcVBwCzOQYAIcoAgCXKAIApygCALcoAgLaFBgC1kQYAMcoAgLuRBgC6jQYANcoAgDnKAIC//QYAvv0GAL39BgC8hQYAPcoAgKN9BgBBygCARcoAgKbBBgBJygCATcoAgKXVBgCqyQYAq9UGAFHKAIC+bAEArrkGAK+5BgCswQYArbkGAKjpAQCp6QEAqvkBAKv5AQCs6QEArekBAK45AQCvOQEAgPUAAIH9AACCwQAAVcoAgIYQAACHdAEAWcoAgAXJAIC4zQAAudUAALrVAAC75QAAvP0AAL2VAAC+kQAAv5EAALBJAQCxSQEAslkBALNZAQC0SQEAtUkBALb9AAC39QAA7/QGAF3KAIBhygCAZcoAgO8wAgBpygCAbcoAgHHKAIDj4AcAdcoAgOGAAQB5ygCA4ygGAH3KAIDhyAUAgcoAgLMxAgCFygCAicoAgJYAAACNygCAtikCALUhAgCRygCAu80CALrNAgCVygCAmcoAgL/NAgC+zQIAvc0CALzNAgCdygCAocoAgKXKAICj/QIAqcoAgKXtAgCm5QIArcoAgLHKAIC1ygCAqgECAKsBAgCsAQIArQECAK4BAgCvAQIAgA0AAIEVAACCHQAAucoAgL3KAIDBygCAvlQMAMnKAICGwAwAhyQDAM3KAIDRygCA1coAgNnKAIDdygCA4coAgKi5AgCpAQEAqgEBAKsBAQCsBQEArQ0BAK4FAQCvOQEAhKgNAOXKAIDpygCA7coAgPHKAID1ygCA+coAgP3KAIC4LQEAucUBALrNAQC7xQEAvMEBAL3JAQC++QEAv/kBALBNAQCxUQEAslUBALMpAQC0OQEAtSUBALYlAQC3FQEA4RgGAAHLAIDjOAcABcsAgAnLAIC+WAwADcsAgBHLAICEbA8AFcsAgL5gDwAZywCAHcsAgCHLAIDvcAYAJcsAgIAVAACBGQAAgi0AAITMDwDjYAYAKcsAgOGgAQAtywCA73QAADHLAICGyAwAh/wMADnLAIA9ywCAQcsAgEXLAICjCQ4AxcoAgDXLAIBJywCATcsAgKYNDgClDQ4AUcsAgKsVDgCqCQ4AVcsAgFnLAICvYQ4Arn0OAK19DgCsAQ4AXcsAgLOpDgBhywCAZcsAgLapDgBpywCAbcsAgLWpDgC6SQ8Au0kPAHHLAIB1ywCAvkkPAL9JDwC8SQ8AvUkPAKhdDgCpbQ4AqmUOAKt9DgCsZQ4ArW0OAK5lDgCvuQ8AecsAgH3LAICBywCAhcsAgInLAICNywCAkcsAgJXLAIC4UQ8AuV0PALpVDwC7aQ8AvH0PAL1lDwC+bQ8Av2EPALDJDwCxyQ8AstkPALPZDwC0yQ8AtckPALZ9DwC3cQ8AmcsAgLURDwC2EQ8AncsAgIARAACBGQAAgikAALMVDwC8HQ8AvWEPAL5hDwC/fQ8AocsAgKXLAIC6FQ8AuwkPAKOtDwCpywCAhugAAIfIAQCtywCApq0PAKWtDwCxywCAq00OAKpNDgC1ywCAucsAgK9NDgCuTQ4ArU0OAKxNDgCocQ4AqXEOAKpxDgCrcQ4ArJ0BAK2FAQCuhQEAr7UBAL7sAAC9ywCAwcsAgMXLAIDJywCAzcsAgNHLAIDVywCAuGEBALlhAQC6YQEAu2EBALxhAQC9YQEAvmEBAL9hAQCwzQEAsaUBALKhAQCzoQEAtKUBALWtAQC2kQEAt5EBALP5DQDZywCA3csAgOHLAIDlywCAtgUCALUVAgDpywCAu2ECALoJAgDtywCA8csAgL9pAgC+YQIAvXUCALx1AgD1ywCAo70NAPnLAID9ywCApkECAAHMAIAFzACApVECAKpNAgCrJQIACcwAgA3MAICuJQIAry0CAKwxAgCtMQIAge0AAIDtAADv0AEAgh0AABHMAIAZzACAhjgEAIdQAwAdzACAIcwAgCXMAIApzACA4eABAC3MAIDjZA8AMcwAgDXMAIA5zACAPcwAgLORAwBBzACAtbkDALZ9AwBFzACAScwAgE3MAIC6WQMAu1kDALxJAwC9SQMAvv0AAL/1AACoRQIAqVUCAKpVAgCrZQIArH0CAK2xAgCusQIAr7ECAL5oBQBRzACAVcwAgFnMAIBdzACAYcwAgGXMAIBpzACAuF0BALltAQC6ZQEAuw0BALwZAQC9GQEAvg0BAL8FAQCw0QIAsdECALLRAgCz0QIAtHUBALV9AQC2dQEAt20BAOF4DwDjNA4A47gOAOF8DgBtzACAccwAgHXMAIB5zACAfcwAgIHMAICJzACAjcwAgJHMAIDv5A4A79QOAJXMAICjnQIAgmEAAIFpAACAUQAAhJwFAKZxAgCltQIAmcwAgKtVAgCqVQIAhkgEAIfMBACv+QEArvEBAK1FAgCsRQIAqJUGAKmlBgCqrQYAq6UGAKy9BgCtoQYArqUGAK/dBgCFzACAncwAgKHMAIClzACAqcwAgK3MAICxzACAtcwAgLhtBwC5dQcAun0HALt1BwC8bQcAvcUHAL7NBwC/xQcAsKUGALGtBgCyuQYAs7EGALSRBgC1kQYAtl0HALdVBwCzJQYAucwAgL3MAIDBzACAxcwAgLYhBgC1NQYAycwAgLtpBgC6YQYAzcwAgNHMAIC/VQYAvlUGAL1lBgC8bQYA1cwAgKNhBgDZzACA3cwAgKZlBgDhzACA5cwAgKVxBgCqJQYAqy0GAOnMAIDtzACArhEGAK8RBgCsKQYArSEGAKipBgCpqQYAqrkGAKuxBgCszQYArTEBAK4xAQCvMQEAgMkBAIHJAQCCBQAA8cwAgL54AgCEeAIA9cwAgPnMAIC43QEAue0BALrlAQC7jQEAvJkBAL2ZAQC+jQEAv4UBALBRAQCxUQEAslEBALNRAQC09QEAtf0BALb1AQC37QEAszEGAP3MAICGKAAAh9wBAAHNAIC2sQEAtUUGAAXNAIC7lQEAupUBAAnNAIANzQCAvzkBAL4xAQC9hQEAvIUBABXMAICjdQYAEc0AgBXNAICm9QEAGc0AgB3NAIClAQYAqtEBAKvRAQAhzQCAJc0AgK51AQCvfQEArMEBAK3BAQApzQCALc0AgDHNAIA1zQCAOc0AgD3NAIBBzQCARc0AgEnNAIBNzQCAUc0AgFXNAIBZzQCAXc0AgGHNAIC+cAMAhQA8AOHEBgCERAIA44wHAIBhAACBYQAAgmEAAO9oAwCFRDwA4RACAGnNAIDj2CsAhlA9AIf0AwBtzQCA76QHAHHNAIDvQAIAdc0AgHnNAIB9zQCAgc0AgIXNAICJzQCAhDw8AI3NAICRzQCAlc0AgJnNAIDj7AIAnc0AgOEsAQCzUQMAoc0AgKXNAICpzQCArc0AgLZ5AwC1cQMAsc0AgLs5AwC6MQMAtc0AgLnNAIC/9QAAvvUAAL0VAwC8FQMAqD0CAKmBAgCqmQIAq5ECAKy5AgCtuQIArtECAK/RAgCEqD8Avqg/AL3NAIDBzQCAxc0AgMnNAIDNzQCA0c0AgLhRAQC5UQEAulEBALtRAQC8cQEAvXEBAL5xAQC/cQEAsLUCALG9AgCygQIAs4ECALRxAQC1cQEAtnEBALdxAQCAtQAAgb0AAIK1AADZzQCAhrA/AIfgPADdzQCA71QAAL4sPgDhVAYA4c0AgOOIAADlzQCA6c0AgO3NAIDxzQCAo1ECAPXNAIC/2CYA+c0AgP3NAICmeQIApXECAAHOAICrOQIAqjECAAXOAIAJzgCAr/UBAK71AQCtFQIArBUCAJAtJACRBSgAkg0oAJPZKACUhS0AlTUsAJbFLACXtTEAmAEwAJkVMACalTUAmyk0AJxtNACdmTUAnj04AJ81OABlzQCAttU+ALXFPgDVzQCAs9E+AA3OAIARzgCAFc4AgL/ZPgC+1T4AvcU+ALzFPgC71T4Auuk+ABnOAICPXSQAqeUJAKgVCACrBQwAqg0MAK0BEACsAQwAr0EQAK69EACh4QAAHc4AgKMBBACi4QAApZ0EAKSVBACnuQgApgEIAKD1OQChBT0Aouk8AKP1PQAhzgCAJc4AgCnOAIAtzgCAscEUALABFACzARgAsn0UALXVGAC01RgAMc4AgDXOAICCISUAgyklADnOAIA9zgCAhsUpAIeBLACEGSkAhRkpAIoBLQCL+S0AQc4AgEnOAICOATEAj4k0AIyRMACNHTEAkkU1AJMZNQCG6AcAh+wBAJZZOQCXYTgAlPU0AJVZOQCaoTwAm0U9AE3OAIBRzgCAgX0AAIB9AACcQTwAglUAAKjpPwCp/T8Aqgk/AKsFPwCsHT8ArQU/AK4NPwCvBT8AVc4AgFnOAIBdzgCAYc4AgGXOAIBpzgCAbc4AgHHOAIC4DT8AuRU/ALoVPwC7JT8AvD0/AL39PgC+9T4Av+0+ALB9PwCxQT8AskE/ALNBPwC0QT8AtU0/ALY9PwC3NT8Ao4E8AHXOAIB5zgCAfc4AgIHOAICmhTwApZU8AIXOAICrhTwAqrk8AInOAICNzgCAr4k8AK6FPACtlTwArJU8AITIAwCz7T0Akc4AgJXOAIC26T0Amc4AgJ3OAIC16T0Auq09ALu1PQChzgCApc4AgL6dPQC/IQIAvKU9AL2VPQCoDT0AqR09AKohPQCrPT0ArCU9AK0tPQCuJT0Ar1k9AIANAACBFQAAgh0AAKnOAICtzgCAsc4AgLnOAIC+uAMAuLkCALlhAgC6GQIAuxkCALwJAgC9CQIAviECAL8hAgCwLT0AsTU9ALI1PQCzBT0AtB09ALWhAgC2oQIAt6ECAKOpPAC9zgCAhigFAIfsAgDBzgCApq08AKWtPADFzgCAq/E8AKrpPADJzgCAzc4AgK9lAwCu2TwArdE8AKzhPADRzgCAsykCANXOAIDZzgCAtvkCAN3OAIDhzgCAtfkCALrVAgC73QIA5c4AgOnOAIC+eQEAv3kBALzFAgC9eQEA7c4AgPHOAICj5QIA9c4AgKU1AgD5zgCA/c4AgKY1AgABzwCABc8AgKsRAgCqGQIArbUBAKwJAgCvtQEArrUBAOPwPgDhrD8A4UA+AON8PwAJzwCADc8AgBHPAIAVzwCAgA0AAIERAACCEQAAGc8AgO+oPgAdzwCAIc8AgO8gPgCoLQUAqW0FAKplBQCrrQUArLUFAK29BQCutQUAr60FALXOAICE6AMAvuADACXPAICGEAMAh5gDACnPAIAtzwCAuGkGALlpBgC6AQYAuwEGALwFBgC9DQYAvjEGAL8xBgCw1QUAsd0FALLVBQCzaQYAtHkGALV5BgC2aQYAt2EGAKg5BgCpgQcAqpkHAKuRBwCsuQcArbkHAK7ZBwCv1QcAMc8AgDXPAIBFzgCAOc8AgD3PAIBBzwCARc8AgEnPAIC4VQcAuV0HALppBwC7aQcAvAEHAL0BBwC+AQcAvwEHALCtBwCxsQcAsrEHALOFBwC0nQcAtXUHALZ9BwC3cQcAsxEGAE3PAIBRzwCAVc8AgFnPAIC2OQYAtTEGAF3PAIC7dQYAumkGAGHPAIBlzwCAv7EGAL5ZBgC9UQYAvGUGAGnPAICjVQYAbc8AgHHPAICmfQYAdc8AgHnPAICldQYAqi0GAKsxBgB9zwCAgc8AgK4dBgCv9QYArCEGAK0VBgCouQEAqbkBAKopAQCrKQEArD0BAK0lAQCuLQEAryUBAIXPAICCHQAAgR0AAIAdAACJzwCAjc8AgJHPAIC+cAEAuIEAALmNAAC6hQAAu5kAALyJAAC9vQAAvrUAAL99AACwXQEAseEAALLhAACz4QAAtOEAALXpAAC20QAAt9EAAITIAgCzpQIAhzgDAIYoAgC2oQIAmc8AgJ3PAIC1sQIAup0CALshAwC+bAMAoc8AgL4hAwC/KQMAvDEDAL0xAwCj4QIApc8AgKnPAICtzwCAsc8AgKblAgCl9QIAtc8AgKtlAwCq2QIAuc8AgL3PAICvbQMArmUDAK11AwCsdQMAqZkAAKiRAACrzQAAqqEAAK3dAACs3QAAr8UAAK7NAAC+LA0Awc8AgMXPAIDJzwCAzc8AgNHPAIDVzwCA2c8AgLnBAQC4eQAAu8EBALrJAQC9wQEAvNkBAL/FAQC+xQEAsY0AALCNAACzQQAAskkAALVBAAC0WQAAt0EAALZJAADdzwCA4c8AgOXPAIDpzwCA7c8AgO9QBwDxzwCA9c8AgL74DwDjdAcA+c8AgOF8BACAGQAAgQkAAIJ5AAD9zwCAAdAAgLNpAQAJ0ACAhMQCALYdAQAN0ACAEdAAgLUVAQC6CQEAuwkBAIboDQCH6A0Avt0BAL/FAQC83QEAvdUBABXQAIAZ0ACAHdAAgCHQAIDv1AAAJdAAgCnQAIDvTAEA47ADAOG0BgDhgAEA45gBAC3QAIAx0ACANdAAgDnQAIA90ACAQdAAgKPlAQCEwA0ApZkBAEXQAIBJ0ACAppEBAE3QAIBR0ACAq4UBAKqFAQCtWQEArFEBAK9JAQCuUQEABdAAgFXQAIBZ0ACAXdAAgGHQAIBl0ACAadAAgG3QAICoaQ8AqXEPAKpxDwCrrQ8ArLUPAK29DwCutQ8Ar6kPALDZDwCx9Q8Asv0PALP1DwC07Q8AtZUPALadDwC3iQ8AuLkPALmFDwC6jQ8Au2kAALx5AAC9eQAAvmkAAL9pAACBnQAAgJ0AAHHQAICCBQAAddAAgHnQAIB90ACAgdAAgIaAAwCH9AMAhdAAgInQAICN0ACAkdAAgJXQAICVzwCAs5kPAJnQAICd0ACAodAAgKXQAIC2XQ8AtV0PAKnQAIC7UQ8Aun0PAK3QAICx0ACAvzEPAL5JDwC9QQ8AvEkPAKNZDgC10ACAudAAgL3QAIDB0ACApp0OAKWdDgDF0ACAq5EOAKq9DgDJ0ACAzdAAgK/xDgCuiQ4ArYEOAKyJDgDR0ACA1dAAgNnQAIDd0ACAgBkAAIEZAACCBQAA4dAAgISgAQDl0ACAh+gBAIYABADp0ACA7dAAgPHQAID10ACAqBUBAKkdAQCqFQEAqyUBAKw9AQCtJQEAri0BAK8lAQD50ACA/dAAgAHRAIAF0QCACdEAgA3RAIAR0QCAFdEAgLjJAAC5yQAAutkAALvRAAC8+QAAvfkAAL6ZAAC/mQAAsCUBALEtAQCyJQEAsz0BALQtAQC1HQEAthUBALf5AAAZ0QCAHdEAgCHRAICzkQIAJdEAgLW5AgC2qQIAKdEAgC3RAIAx0QCAuu0CALvlAgC8/QIAveUCAL7lAgC/1QIApvECADXRAIA50QCApeECAD3RAICjyQIAQdEAgEXRAICuvQIAr40CAKylAgCtvQIAqrUCAKu9AgBJ0QCATdEAgID5AACB+QAAggUAAFHRAIC+yAMAhBgDAFnRAIBd0QCAYdEAgGXRAIBp0QCAbdEAgHHRAIB10QCAhhgEAIecAwB50QCAfdEAgIHRAICF0QCAidEAgI3RAIDvsAIAkdEAgOGUAQCV0QCA42wCAJnRAICd0QCAodEAgKXRAICp0QCA79APAK3RAICx0QCAtdEAgLnRAIDhrAEAvdEAgONsAACAMQAAgT0AAIIdAADv9A4A42wOAMHRAIDhLA8AvnAFALM5AgCEDAUAhugEAIdgBQDcAAAAtvECALX5AgDJ0QCAu9UCALrVAgDN0QCA0dEAgL91AQC+dQEAvcUCALzFAgDV0QCA4fQOANnRAIDjUA4A3dEAgOHRAIDl0QCA6dEAgO3RAIDx0QCA9dEAgPnRAID90QCAAdIAgAXSAIDv5A8ApmUCAAnSAIAN0gCApW0CABHSAICjrQIAFdIAgBnSAICu4QEAr+EBAKxRAgCtUQIAqkECAKtBAgAd0gCAIdIAgKiZBgCpmQYAqqkGAKupBgCsuQYArbkGAK6pBgCvqQYAJdIAgIIdAACBHQAAgB0AACnSAIAt0gCAMdIAgL50AwC4rQYAubUGALq9BgC7tQYAvK0GAL1RBwC+UQcAv1EHALChBgCxoQYAsqEGALOhBgC0oQYAtaEGALalBgC3mQYAVdEAgLMlBgCExAMAxdEAgLY9BgA10gCAOdIAgLU1BgC6YQYAu2EGAIYIAACHiAAAvmEGAL9hBgC8cQYAvXEGAKNhBgA90gCAQdIAgEXSAIBJ0gCApnkGAKVxBgBN0gCAqyUGAKolBgBR0gCAVdIAgK8lBgCuJQYArTUGAKw1BgCoXQYAqW0GAKplBgCrjQYArJkGAK2FBgCujQYAr4UGAFnSAIBd0gCAYdIAgGXSAIBp0gCAbdIAgHHSAIB10gCAuIUGALmNBgC6mQYAu5UGALyNBgC9rQYAvqUGAL99AQCw/QYAscUGALLNBgCzxQYAtN0GALXFBgC2zQYAt8UGALPtBgB50gCAfdIAgIHSAICF0gCAtgUGALURBgCJ0gCAuwEGALo5BgCN0gCAkdIAgL8BBgC+GQYAvREGALwZBgCV0gCAo6kGAJnSAICd0gCApkEGAKHSAICElAEApVUGAKp9BgCrRQYAvqABAKnSAICuXQYAr0UGAKxdBgCtVQYAqJkCAKnBAgCqwQIAq8ECAKzBAgCtyQIArvECAK/xAgCB7QMAgO0DAK3SAICC+QMAhpAcAId0AwCx0gCAtdIAgLjFAwC5zQMAusUDALvdAwC8zQMAvf0DAL71AwC/nQMAsEEDALFBAwCyQQMAs0EDALRBAwC1QQMAtkEDALdBAwCzSQIAudIAgL3SAIDB0gCAxdIAgLZJAgC1SQIAydIAgLuFAwC6hQMAzdIAgNHSAIC/hQMAvoUDAL2VAwC8lQMA1dIAgKMNAgDZ0gCA3dIAgKYNAgDh0gCA5dIAgKUNAgCqwQMAq8EDAOnSAIDt0gCArsEDAK/BAwCs0QMArdEDAOOYAQDhpAcA4VgGAONYBgDhoAEA8dIAgOPQAAD10gCA+dIAgP3SAIDvOAAAAdMAgO/0AQAF0wCACdMAgO/4BgCAeQAAgRUAAIIdAACEAB0ADdMAgBHTAIC+EB0AGdMAgIbAHACHrB0AHdMAgCHTAIAl0wCAKdMAgC3TAIAx0wCAu8UFALqhBQC5qQUAuJEFAL/NBQC+zQUAvckFALzVBQCzHQYAsh0GALEdBgCwHQYAt6EFALa9BQC1vQUAtL0FAKu9BgCqvQYAqb0GAKi9BgCvfQYArn0GAK19BgCsfQYANdMAgDnTAIA90wCAQdMAgEXTAIBJ0wCATdMAgFHTAICo7R0AqS0eAKoxHgCrMR4ArJUeAK2dHgCulR4Ar40eABXTAIBV0wCAWdMAgF3TAIBh0wCAZdMAgGnTAIBt0wCAuKkeALmpHgC6XR8Au1EfALxxHwC9cR8AvnUfAL9pHwCw/R4Asc0eALLFHgCzrR4AtLkeALW5HgC2rR4At6UeALO5HgBx0wCAddMAgHnTAICl0gCAth0eALUdHgB90wCAuwkeALo5HgCB0wCAhOADAL99HgC+fR4AvXkeALwRHgCCaQAAo/0eAIBFAACBUQAAplkeAL6cAwCF0wCApVkeAKp9HgCrTR4AhkgAAIdsAACuOR4ArzkeAKxVHgCtPR4AqF0eAKltHgCqZR4Aq30eAKxlHgCtbR4ArmUeAK/9HgCJ0wCAjdMAgJHTAICV0wCAmdMAgJ3TAICh0wCApdMAgLhpAQC5aQEAunkBALt5AQC8aQEAvWkBAL7dAQC/1QEAsIUeALGNHgCyhR4As50eALSFHgC1jR4AtoUeALdZAQCz7R4AqdMAgK3TAICx0wCAtdMAgLbtHgC17R4AudMAgLtJHgC6QR4AvdMAgMHTAIC/SR4AvkEeAL1JHgC8UR4AxdMAgKOpHgDJ0wCAzdMAgKapHgDR0wCA1dMAgKWpHgCqBR4Aqw0eANnTAIDd0wCArgUeAK8NHgCsFR4ArQ0eAKghAwCpIQMAqiEDAKshAwCsIQMArSEDAK4hAwCvIQMA4dMAgOXTAIDp0wCAvmACAO3TAIDx0wCA+dMAgP3TAIC4iQMAuYkDALqdAwC7lQMAvLkDAL25AwC+eQAAv3kAALDlAwCx7QMAsuUDALP9AwC07QMAtd0DALbVAwC3vQMAgKkAAIG1AACCvQAAs6UDAAHUAIC1pQMAtq0DAAXUAICE4AIACdQAgLotAwC7JQMAvD0DAL0lAwC+JQMAvxUDAKPpAwAN1ACAhmgEAIeAAwAR1ACApuEDAKXpAwAV1ACAq2kDAKphAwAZ1ACAHdQAgK9ZAwCuaQMArWkDAKxxAwAh1ACAJdQAgCnUAIAt1ACAMdQAgOE8HwA11ACA40AeADnUAIA91ACAQdQAgO+MHgBF1ACASdQAgE3UAIBR1ACAVdQAgIIlAACBEQAAgB0AAFnUAIDj5AMAXdQAgOGsAQBh1ACA77ADAIRkAgC+YAUAhtAEAIdEBQBp1ACAbdQAgHHUAIB11ACAedQAgH3UAICB1ACAhdQAgInUAIDvsAEAhKQFAOHcHgCN1ACA4xABAJHUAICV1ACAmdQAgJ3UAICzUQEAodQAgKXUAICp1ACArdQAgLYRAQC1fQEAsdQAgLsNAQC6DQEAtdQAgLnUAIC//QAAvv0AAL39AAC8/QAAqDkGAKk5BgCqmQYAq5EGAKy1BgCt0QYArskGAK/BBgBl1ACAvdQAgMHUAIDF1ACAgA0AAIGxAACCsQAAydQAgLhhBwC5YQcAumEHALt9BwC8ZQcAvW0HAL5lBwC/HQcAsIkGALGJBgCyaQcAs2kHALR5BwC1eQcAtmkHALdlBwCjEQYAzdQAgNHUAIC+gAMA1dQAgKZRBgClPQYA2dQAgKtNBgCqTQYAhggAAId8AwCvvQcArr0HAK29BwCsvQcA3dQAgOHUAICzSQcA5dQAgLVZBwDp1ACA7dQAgLZRBwDx1ACA9dMAgLtBBwC6dQcAvUUHALxFBwC/RQcAvkUHAKh5BgCpeQYAqokGAKuJBgCsmQYArZkGAK6JBgCviQYA9dQAgPnUAID91ACAAdUAgAXVAIAJ1QCADdUAgBHVAIC4jQYAuZUGALqVBgC7pQYAvL0GAL1xAQC+cQEAv3EBALD5BgCxzQYAstkGALPZBgC0yQYAtckGALa9BgC3tQYAowEGABXVAIAZ1QCAHdUAgCHVAICmGQYApREGACXVAICrCQYAqj0GACnVAIAt1QCArw0GAK4NBgCtDQYArA0GADHVAIA11QCAOdUAgD3VAICAGQAAgRkAAIIFAABB1QCAhKwBAL6sAQCH6AAAhkwPAEnVAIBN1QCAUdUAgFXVAIConQIAqcUCAKrNAgCrwQIArMUCAK3NAgCu+QIArz0DAFnVAIBd1QCAYdUAgGXVAIC+PAwAadUAgG3VAIBx1QCAuMkDALnJAwC62QMAu9EDALz5AwC9+QMAvpkDAL+ZAwCwRQMAsU0DALJFAwCzXQMAtEUDALVNAwC2RQMAt/kDALNFAgB11QCAedUAgH3VAICB1QCAtk0CALVNAgCF1QCAu4kDALqBAwCJ1QCAjdUAgL+JAwC+gQMAvYkDALyRAwCR1QCAowECAJXVAICZ1QCApgkCAJ3VAICh1QCApQkCAKrFAwCrzQMApdUAgKnVAICuxQMAr80DAKzVAwCtzQMAgO0BAIEVAACCEQAAhAACAK3VAIDhpAEAsdUAgOPsAAC51QCAvdUAgMHVAIDvMAAAxdUAgMnVAIDN1QCA0dUAgIbgDACH9AIA1dUAgNnVAIDd1QCA4dUAgO/MBgDl1QCA4bAHAOnVAIDjEAYA7dUAgPHVAID11QCA+dUAgP3VAIAB1gCABdYAgAnWAIAN1gCAEdYAgBXWAIAZ1gCA7+gBAIUYDwDhzAYAHdYAgOMcBgCAKQAAgR0AAIIFAAAh1gCAszkCAITMDQCGaA8Ah/wMAOHQ0gO28QEAtfkBACnWAIC72QEAutEBAL7kDAAt1gCAv30BAL59AQC9fQEAvMEBAKjxDQCp8Q0AqvENAKvxDQCsMQ4ArTEOAK4xDgCvMQ4AtdUAgCXWAIAx1gCANdYAgDnWAIA91gCAQdYAgEXWAIC46Q4AuekOALqJDgC7hQ4AvJ0OAL2BDgC+gQ4Av7UOALBVDgCxXQ4AslUOALPpDgC0+Q4AtfkOALbpDgC34Q4Ao3kNAEnWAIBN1gCAUdYAgFXWAICmsQ4ApbkOAFnWAICrmQ4AqpEOAF3WAIBh1gCArz0OAK49DgCtPQ4ArIEOAGXWAICz7Q8AadYAgG3WAIC26Q8AcdYAgHXWAIC16Q8Auq0PALu1DwBF1QCAedYAgL6VDwC/mQ8AvK0PAL2hDwCoIQ4AqSEOAKohDgCrPQ4ArCUOAK0tDgCuJQ4Ar1UOAH3WAICB1gCAhdYAgInWAICAHQAAgQkAAIK9AACN1gCAuDkOALk5DgC6yQ4Au8kOALzZDgC92Q4AvskOAL/JDgCwLQ4AsTUOALI9DgCzMQ4AtBUOALUZDgC2CQ4AtwkOAKOpDgCR1gCAhIACAL6AAQCFAAQApq0OAKWtDgCZ1gCAq/EOAKrpDgCGKAcAhxgAAK/dDgCu0Q4AreUOAKzpDgCd1gCAs+0BAKHWAICl1gCAtuUBAKnWAICt1gCAte0BALplAQC7bQEAsdYAgLXWAIC+bQEAv10BALx1AQC9bQEAqN0NAKnpDQCqIQIAqyECAKwhAgCtIQIAriECAK8hAgC51gCAvdYAgMHWAIDF1gCAohECAKMRAgCgqQ4AodUCALiJAgC5iQIAup0CALuVAgC8vQIAvXUDAL59AwC/dQMAsOUCALHtAgCy5QIAs/0CALTtAgC13QIAttUCALe9AgCjqQIAj8UaAMnWAIDN1gCA0dYAgKahAgClqQIA1dYAgKspAgCqIQIA2dYAgN3WAICvGQIArikCAK0pAgCsMQIAniUOAJ/lDgCc6QoAnRUKAJpFFgCbRQoAmFkWAJlRFgCWcRIAl4ETAJRVEgCV7RIAktEeAJPZHgCQtRoAkVUeAISpHwCFJR8AhiUfAIexEwDh1gCA5dYAgIJZGwCDURsAjEUSAI2lFwCOpRcAj7kXAIA5+wHp1gCAijkTAIutEwCUmQsAlaEPAJZpDwCX3Q8A7dYAgO+cDwCSyQsAk30LAJxFAwDjeA4A8dYAgOGYDAD11gCAhHgCAJqRAwCbXQMA4QQAAL6IBQDj3OoD+dYAgP3WAIAB1wCA7+wAAO+MDgDhcA4A4fwOAOMwAADjeA4AgSEAAIA5AADvtO0DgikAALMJAgAJ1wCAhmgEAIcsBQAN1wCAtg0CALUNAgAR1wCAu8UBALrFAQAV1wCAGdcAgL99AQC+fQEAvdUBALzVAQCV1gCABdcAgB3XAIAh1wCAJdcAgCnXAIAt1wCAMdcAgKi9BQCp5QUAquEFAKvhBQCs5QUAre0FAK7RBQCv0QUAsGEGALFhBgCyYQYAs2EGALTZBgC12QYAtskGALfBBgC4yQYAuckGALp5BwC7eQcAvEUHAL0lBwC+EQcAvw0HAKNJBQA11wCAOdcAgD3XAIBB1wCApk0FAKVNBQBF1wCAq4UGAKqFBgBJ1wCATdcAgK89BgCuPQYArZUGAKyVBgBR1wCAVdcAgFnXAIBd1wCAYdcAgGXXAIBp1wCAbdcAgIA5AACBOQAAggUAAHHXAIC+uAMAhLgDAHnXAIB91wCAqMUGAKnVBgCq1QYAq+UGAKz9BgCtHQEArhUBAK8NAQB11wCAgdcAgIaIAQCHHAEAhdcAgInXAICN1wCAkdcAgLjpAQC56QEAuokBALuJAQC8mQEAvZkBAL6JAQC/iQEAsHUBALF9AQCydQEAs+kBALT5AQC1+QEAtukBALfhAQCzXQYAldcAgJnXAICd1wCAhLwBALadAQC1dQYAodcAgLu5AQC6sQEApdcAgKnXAIC/PQEAvj0BAL09AQC8oQEArdcAgKMZBgCx1wCAtdcAgKbZAQC51wCAvdcAgKUxBgCq9QEAq/0BAMHXAIDF1wCArnkBAK95AQCs5QEArXkBAKj5AgCp+QIAqi0DAKs9AwCsJQMArS0DAK4lAwCvmQMAydcAgM3XAIDR1wCA1dcAgIANAACBsQAAgrEAANnXAIC4lQMAuZ0DALqhAwC7oQMAvHEAAL1xAAC+cQAAv3EAALDpAwCx6QMAsvUDALPFAwC03QMAtbUDALaxAwC3sQMAvswDAN3XAIDh1wCA6dcAgO3XAIDx1wCA9dcAgO/kAgD51wCA4ZQBAP3XAIDjLAEAAdgAgAXYAICHGAMAhhz8A7tNAwC6TQMACdgAgA3YAIC/EQMAvnkDAL1xAwC8QQMAs8UDAITo/AMR2ACAFdgAgBnYAIC2zQMAtc0DAB3YAICkAfwDpSX/A6bZ/wOnAfgDIdgAgKEVAwCiHQMAoz0CAKwR9wOtAfADri3zA68B8wOoEfsDqZn7A6oB9AOrHfcDtAHoA7Vl6wO+xPwDhMT8A7AB7AOxVe8Dsk3vA7Nx7gMl2ACAKdgAgC3YAIAx2ACANdgAgDnYAIA92ACAQdgAgOFQBgDhNAQA42wBAOPoBgBF2ACASdgAgE3YAIBR2ACAgDUAAIE9AACCNQAAWdgAgF3YAIBh2ACA77ABAO/ABgCj5QIAZdgAgIbo/AOHfP0DadgAgKbtAgCl7QIAbdgAgKttAgCqbQIAcdgAgHXYAICvMQIArlkCAK1RAgCsYQIAqI3+A6mV/gOqnf4Dq5X+A6yx/gOtvf4Drqn+A6+p/gNV2ACAedgAgH3YAICB2ACAhdgAgInYAICN2ACAkdgAgLgl/wO5Lf8DuiX/A7s9/wO8Jf8DvS3/A74l/wO/zf8DsKn+A7Gp/gOygf4Ds4H+A7SB/gO1if4Dtmn/A7cd/wOV2ACA4SD8A5nYAIDjePwDndgAgKHYAICl2ACAqdgAgK3YAICx2ACAtdgAgLnYAICAHQAAgXEAAIJxAADvDP0Ds1X+A73YAIDB2ACAvkAAAMXYAIC2ff4DtXn+A8nYAIC7Lf4Dui3+A4boAACHrAAAvw3+A74F/gO9Ff4DvBX+A6OV/wPN2ACA0dgAgNXYAIDZ2ACApr3/A6W5/wPd2ACAq+3/A6rt/wPh2ACA5dgAgK/N/wOuxf8DrdX/A6zV/wPp2ACAs/H+A+3YAIDx2ACAto3+A/XYAID52ACAtY3+A7pFAQC7TQEA/dgAgAHZAIC+RQEAv00BALxVAQC9TQEAqC3+A6k1/gOqPf4Dq0n+A6xB/gOtSf4DrnH+A69x/gMF2QCACdkAgA3ZAIAR2QCAFdkAgBnZAIAd2QCAIdkAgLhJAQC5VQEAul0BALtVAQC8TQEAvXUBAL59AQC/dQEAsMUBALHNAQCyxQEAs90BALTFAQC1zQEAtsUBALd9AQCjtf0DJdkAgCnZAICExAMALdkAgKbJ/QOlyf0DMdkAgKsJAgCqAQIAOdkAgL7sAgCvCQIArgECAK0JAgCsEQIAgEkAAIFVAACCVQAAo0UDAD3ZAIClRQMApkUDAEHZAICGwAQAhxQDAKopAwCrJQMArD0DAK0hAwCuIQMArxUDAEXZAIBJ2QCATdkAgFHZAIBV2QCAWdkAgF3ZAIBh2QCAqH0CAKmhAwCqoQMAq6EDAKyhAwCtqQMArpEDAK+RAwCwgQMAsY0DALKFAwCzmQMAtIkDALW9AwC2tQMAt30DALhFAwC5TQMAukUDALtdAwC8RQMAvU0DAL5FAwC/+QAA5dcAgLMNAgBl2QCAadkAgLYNAgBt2QCAcdkAgLUNAgC6YQIAu20CAHXZAIB52QCAvmkCAL9dAgC8dQIAvWkCAH3ZAICB2QCAhdkAgInZAICN2QCA4aQBAJHZAIDjQAMAldkAgJnZAICd2QCA77gDAIAVAACBHQAAggUAAKHZAICEgAIAvsgFAIcYBQCGLAQAqdkAgK3ZAICx2QCA76gBALXZAIDhdP4DudkAgOPw/gO92QCAwdkAgMXZAIDJ2QCAzdkAgNHZAIDV2QCAs5EBANnZAIC1UQEAtlEBAN3ZAIDh2QCA5dkAgLp9AQC7dQEAvG0BAL39AAC+9QAAv+kAAKgpBgCpVQYAqlUGAKuNBgCslQYArZ0GAK6VBgCvjQYApdkAgOnZAIDt2QCA8dkAgPXZAID52QCA/dkAgAHaAIC4bQcAuQUHALoNBwC7BQcAvB0HAL0FBwC+AQcAvz0HALD1BgCx/QYAsvUGALNlBwC0fQcAtWEHALZhBwC3VQcA4xAFAAXaAIDh8AQACdoAgIAdAACBCQAAgjkAAA3aAIAR2gCAhOgDAL7gAwAV2gCA78wFABnaAICHOAAAhhgAAKOdBgAd2gCAIdoAgCXaAIAp2gCApl0GAKVdBgAt2gCAq3kGAKpxBgAx2gCANdoAgK/lBwCu+QcArfEHAKxhBgCokQYAqZEGAKqRBgCrrQYArLkGAK2lBgCurQYAr6UGADnaAIA92gCAQdoAgEXaAIBJ2gCATdoAgFHaAIBV2gCAuGUBALltAQC6ZQEAu30BALxlAQC9bQEAvmUBAL/ZAQCw3QYAsaUGALKtBgCzpQYAtKEGALWpBgC2mQYAt5kGALMZBgBZ2gCAXdoAgGHaAIBl2gCAtiUGALUxBgBp2gCAu2EGALoZBgBt2gCAcdoAgL9tBgC+ZQYAvXEGALx5BgB12gCAo10GAHnaAIB92gCApmEGAIHaAICEmAEApXUGAKpdBgCrJQYAvqQBAInaAICuIQYArykGAKw9BgCtNQYAqcUCAKixAgCrxQIAqsUCAK3NAgCsxQIAr/UCAK71AgCN2gCAkdoAgJXaAICZ2gCAndoAgKHaAICl2gCAqdoAgLnJAwC4wQMAu9kDALrBAwC9+QMAvMkDAL+ZAwC+8QMAsUUDALBFAwCzRQMAskUDALVFAwC0RQMAt0UDALZFAwCASQMAgUkDAIJdAwCzRQIAvtwMALVFAgC2RQIArdoAgIYADACH5AMAuokDALuJAwC8mQMAvZkDAL6JAwC/iQMAowkCALHaAIC12gCAudoAgL3aAICmCQIApQkCAMHaAICrxQMAqsUDAMXaAIDJ2gCAr8UDAK7FAwCt1QMArNUDAM3aAIDR2gCA1doAgDXZAIDvAAAA2doAgN3aAIDh2gCA4+gAAOXaAIDhjAEA6doAgO3aAIDx2gCA+doAgP3aAICAbQAAgXUAAIJ9AACEQAIAhvAMAId4DQAB2wCABdsAgAnbAIAN2wCAEdsAgBXbAIAZ2wCAHdsAgCHbAIAl2wCAKdsAgC3bAIAx2wCANdsAgDnbAIA92wCAQdsAgO/MAQCE7AwA4TAGAEXbAIDjGAEASdsAgE3bAIBR2wCAVdsAgLPlAQBZ2wCAhIQPAF3bAIBh2wCAtuUBALX1AQBp2wCAu30BALrZAQC+oAwAbdsAgL8hAQC+OQEAvTEBALw5AQCo7Q0AqSUOAKotDgCrJQ4ArD0OAK0lDgCuLQ4AryUOAPXaAICC9Q8AgeUPAIDpDwBl2wCAcdsAgIaYAACHDAMAuK0OALlFDwC6TQ8Au0UPALxFDwC9TQ8AvkUPAL95DwCwXQ4AsfkOALKtDgCzpQ4AtL0OALWlDgC2pQ4At5UOAHXbAIDv7AwAedsAgH3bAICB2wCAhdsAgInbAICN2wCAvugAAJHbAICV2wCAmdsAgJ3bAIDj6A0AodsAgOEEDACj5Q4ApdsAgKnbAICt2wCAsdsAgKblDgCl9Q4AtdsAgKt9DgCq2Q4AudsAgL3bAICvIQ4ArjkOAK0xDgCsOQ4AqDkOAKk5DgCqUQ4Aq1EOAKxxDgCtcQ4ArnEOAK9xDgDB2wCAxdsAgMnbAIDN2wCAgBkAAIEZAACCBQAA0dsAgLjRDgC50Q4AutEOALvlDgC84Q4AveEOAL7hDgC/4Q4AsBEOALERDgCyEQ4AsxEOALTxDgC18Q4AtvEOALfxDgCz2Q4A2dsAgIYoAACHuAAA3dsAgLbxDgC1+Q4A4dsAgLvVDgC61Q4A5dsAgOnbAIC/NQ4AvjUOAL3FDgC8xQ4A7dsAgKOdDgDx2wCA9dsAgKa1DgD52wCA/dsAgKW9DgCqkQ4Aq5EOAAHcAIAF3ACArnEOAK9xDgCsgQ4ArYEOAKjdDQCp6Q0Aqj0CAKuNAgCsmQIArZkCAK6JAgCviQIAvqwEAAncAIAN3ACAhCADABHcAIAV3ACAGdwAgB3cAIC4iQIAuYkCALqZAgC7kQIAvLkCAL25AgC+eQMAv3kDALD5AgCx+QIAss0CALPFAgC03QIAtcUCALbBAgC3uQIAs7UCACHcAIAl3ACAKdwAgC3cAIC2GQIAtRECADHcAIC7PQIAuj0CADXcAIA53ACAvwECAL4ZAgC9EQIAvBkCAD3cAICj8QIAQdwAgEncAICmXQIATdwAgFHcAIClVQIAqnkCAKt5AgCGSAUAh6wEAK5dAgCvRQIArF0CAK1VAgCohQIAqZUCAKqVAgCrpQIArL0CAK3VAgCu0QIAr9ECAFXcAIBZ3ACAXdwAgGHcAICB8QEAgJkBAIXaAICC9QEAuHkBALl5AQC6zQEAu8UBALzdAQC9xQEAvsUBAL/1AQCwtQIAsb0CALKBAgCzgQIAtFUBALVdAQC2SQEAt0kBAGXcAIBp3ACAbdwAgO/UAQCEEAUAcdwAgHXcAIDvjA4AvuwFAOHsDgB53ACA4xwOAH3cAIDhlAEAgdwAgONkDgCzXQIAhdwAgIncAICN3ACAkdwAgLYVAgC1dQIAldwAgLs5AgC6MQIAmdwAgJ3cAIC/2QEAvtEBAL0VAgC8FQIAo50FAEXcAICh3ACApdwAgKncAICm1QUApbUFAK3cAICr+QUAqvEFALHcAIC13ACArxkGAK4RBgCt1QUArNUFAIBRAACBWQAAgmEAALOVBgC53ACAtXEHALZxBwC93ACAhkADAIdUAwC67QcAu+UHALzlBwC97QcAvtEHAL/NBwDB3ACAxdwAgMncAIDN3ACA0dwAgNXcAIDvQAQA2dwAgOEwBwDd3ACA45QEAOHcAIDl3ACA6dwAgO3cAIDx3ACAoxkGAPXcAID53ACA/dwAgAHdAICm/QcApf0HAAXdAICraQcAqmEHAAndAIAN3QCAr0EHAK5dBwCtYQcArGkHAKjNBwCp0QcAqtEHAKstBgCsNQYArT0GAK41BgCvnQYAEd0AgBXdAIAZ3QCAHd0AgIAZAACBGQAAggUAACHdAIC4iQYAuYkGALqZBgC7kQYAvLkGAL25BgC+UQEAv1EBALDlBgCx7QYAsv0GALP1BgC02QYAtcUGALbBBgC3uQYAqNEBAKnZAQCqCQEAqwkBAKwZAQCtGQEArgkBAK8JAQCEYAEAvnwBAIeoAACGjAEAKd0AgC3dAIAx3QCANd0AgLgJAQC5CQEAuhkBALsRAQC8OQEAvTkBAL75AAC/+QAAsH0BALFBAQCyRQEAs10BALRFAQC1TQEAtkUBALc5AQA53QCAPd0AgEHdAICzjQIARd0AgLWdAgC2lQIASd0AgE3dAIBR3QCAurUCALuJAgC8nQIAvYUCAL6NAgC/hQIAps0CAFXdAIBZ3QCApcUCAF3dAICj1QIAYd0AgGXdAICu1QIAr90CAKzFAgCt3QIAqu0CAKvRAgCE9AMAad0AgKgxAwCpMQMAqjEDAKsxAwCskQAArZEAAK6RAACvjQAAbd0AgHHdAIB13QCAed0AgH3dAICB3QCAhd0AgIndAIC4vQAAuWUAALptAAC7ZQAAvH0AAL1lAAC+bQAAv2UAALD9AACxxQAAss0AALOpAAC0uQAAtaUAALahAAC3oQAAgL0BAIEJAACCGQAAjd0AgJHdAIC+WAIAhxQdAIacHQCEbB0A1dsAgJndAICd3QCAvrwcAKHdAICl3QCAqd0AgLP5AgCt3QCAsd0AgLXdAIC53QCAtlEBALVZAQC+3B8Au0EBALp5AQC93QCAwd0AgL8hAQC+PQEAvT0BALxZAQDhcAcAxd0AgOMIBgDJ3QCA78wAAM3dAIDR3QCA1d0AgOMQAADZ3QCA4dABAN3dAICGkBwAh/QcAO/gBgDh3QCAo3kCAOXdAIDp3QCA7d0AgPHdAICm0QEApdkBAPXdAICrwQEAqvkBAPndAID93QCAr6EBAK69AQCtvQEArNkBAJXdAICCFQAAgeUfAIDlHwAB3gCABd4AgAneAIAN3gCAqAkfAKkJHwCqHR8AqxUfAKwNHwCtcR8ArnEfAK9xHwCwER8AsS0fALIlHwCzyR8AtN0fALXBHwC2wR8At8EfALjFHwC5yR8AutUfALupHwC8uR8AvbkfAL6pHwC/oR8As7UfABHeAIAV3gCAGd4AgB3eAIC20R8AtaUfACHeAIC7yR8AuvUfACXeAIAp3gCAvyUfAL45HwC9PR8AvNEfAC3eAIAx3gCANd4AgDneAIA93gCA4WAfAEHeAIDjtBwARd4AgEneAIBN3gCA7wAdAFHeAIBV3gCAWd4AgF3eAICjNR4AYd4AgGXeAIBp3gCAbd4AgKZRHgClJR4Acd4AgKtJHgCqdR4AhKgCAHXeAICvpR4ArrkeAK29HgCsUR4AgE0AAIFVAACCVQAAs8kBAHneAIC12QEAtskBAH3eAICGoAAAhwQBALrFAQC7rQEAvLUBAL29AQC+tQEAv60BAKiZAQCpmQEAqg0BAKsFAQCsHQEArQUBAK4FAQCvNQEAgd4AgIXeAICJ3gCAjd4AgJHeAICV3gCAmd4AgJ3eAIC4JQEAuS0BALo5AQC7OQEAvCkBAL0pAQC+3QAAv9UAALBNAQCxJQEAsi0BALMlAQC0PQEAtSUBALYhAQC3HQEAod4AgKXeAICp3gCAo4kCAK3eAIClmQIApokCALHeAIC13gCAud4AgKqFAgCr7QIArPUCAK39AgCu9QIAr+0CAL3eAIDB3gCAxd4AgIRAAgDJ3gCAzd4AgNHeAIDV3gCAgA0AAIEVAACCHQAA2d4AgN3eAIDh3gCAh7QDAIbcBAC+zAMA6d4AgO3eAIDx3gCA7+gCAPXeAID53gCA/d4AgOP8AgAB3wCA4dABAAXfAIAJ3wCADd8AgBHfAIAV3wCAs2EDABnfAIAd3wCAId8AgCXfAIC2eQMAtXEDACnfAIC7XQMAul0DAC3fAIAx3wCAv+EAAL79AAC9/QAAvP0AALC5AgCxuQIAsgkBALMJAQC0GQEAtQUBALYFAQC3PQEAuAUBALllAQC6bQEAu2UBALxhAQC9YQEAvmEBAL9hAQCFXAcANd8AgDnfAIA93wCAJd0AgEHfAIBF3wCASd8AgKgxAgCpOQIAqskCAKvJAgCs2QIArdkCAK7JAgCvyQIAhMwFAOGAHgBN3wCA47weAOE4HgBR3wCA46AAAL4QBABZ3wCAXd8AgO8MHgBh3wCAZd8AgGnfAIBt3wCA73QeAKNhAgCCUQAAgUEAAICRAABx3wCApnkCAKVxAgB13wCAq10CAKpdAgCGyAQAhzwFAK/hAQCu/QEArf0BAKz9AQCohQYAqY0GAKqFBgCrmQYArIkGAK2JBgCuvQYAr7EGAFXfAIB53wCAfd8AgIHfAICF3wCAid8AgI3fAICR3wCAuJ0GALmtBgC6pQYAuwkHALwZBwC9GQcAvg0HAL8FBwCw0QYAsdEGALLRBgCz0QYAtLUGALW9BgC2tQYAt60GALMNBgCV3wCAmd8AgJ3fAICh3wCAtgkGALUBBgCl3wCAuxUGALoVBgCp3wCArd8AgL95BgC+cQYAvQUGALwFBgCx3wCA4aAEALXfAIDjXAUAgA0AAIE1AACCPQAAud8AgL3fAIDB3wCAhGADAL5sAAC/8AEAhZAAAMXfAIDvmAUAo40HAIQIAACGAAwAh4wAAMnfAICmiQcApYEHAM3fAICrlQcAqpUHANHfAIDV3wCAr/kHAK7xBwCthQcArIUHANnfAICz6QYA3d8AgOHfAIC26QYA5d8AgOnfAIC16QYAukUBALtNAQDt3wCA8d8AgL5FAQC/TQEAvFUBAL1NAQCoIQYAqSEGAKolBgCrPQYArCUGAK0tBgCuSQYAr0EGAPXfAID53wCA/d8AgAHgAIAF4ACACeAAgA3gAIAR4ACAuEkBALlJAQC6WQEAu1EBALx5AQC9eQEAvhkBAL8VAQCwxQEAsc0BALLFAQCz3QEAtMUBALXNAQC2xQEAt3kBABXgAIAZ4ACAHeAAgKOhBQAh4ACApaEFAKahBQAl4ACAjyHqAyngAICqDQIAqwUCAKwdAgCtBQIArg0CAK8FAgCX7RIAlmUSAJVFEQCUnRYAk3EWAJJVFQCReesDkFnqA59hBgCeNQUAnUUaAJxpGgCbVRkAmkUeAJlZHgCYRR0A4WAAAC3gAIDjTD4AMeAAgKOxAgCi1QEAobUHAKCJBgCxATgAsAk+ALOVOgCyjToAtbUmALQBJADvaDoAvjAMAKnJNgCowTYAqwEwAKrhNwCtzTMArPUyAK/5PgCuATwAoRkCADngAICjbQ4Aom0OAKX1CgCkAQgAp4ULAKaZCgCGAA0Ah0QNAIIJ6wODCesDhDHqA4UVFACGORcAh80XAISgDQA94ACAiiUQAIsNEwCMnRMAjQ0cAI4ZHwCPDR8A5d4AgO8AAwCSbRgAk0kbAJR9GwCVBQQAllkHAJdJBwBB4ACAReAAgJpFBgCbLQAAnFEDAONgAABJ4ACA4WwAAIClAQCBAQEAggUBAL4ADABN4ACAUeAAgFXgAIDviAEAWeAAgOFUBgBd4ACA41QBAGHgAIBl4ACAaeAAgG3gAICz6QIAceAAgHXgAIB54ACAfeAAgLadAgC1mQIAgeAAgLuJAgC6vQIAheAAgIngAIC/WQIAvlECAL1ZAgC8kQIAoykNAI3gAICR4ACAleAAgJngAICmXQ0ApVkNAJ3gAICrSQ0Aqn0NAKHgAICp4ACAr5kNAK6RDQCtmQ0ArFENAIBRAACBWQAAgmEAALMtDwCt4ACAtS0PALbJDwCx4ACAhkADAIcIAwC6yQ8Au8UPALzBDwC9wQ8AvsEPAL/BDwA14ACApeAAgLXgAIC54ACAveAAgMHgAIDF4ACAyeAAgKhFDgCpgQ8AqskPAKvJDwCsyQ8ArSUPAK4tDwCvJQ8AsGEPALFtDwCyeQ8As3kPALRpDwC1aQ8Ath0PALcVDwC4LQ8AuTUPALo1DwC7BQ8AvB0PAL3xAAC+8QAAv/EAAKNhDgDN4ACAhMQBANHgAIDV4ACApoUOAKVhDgDZ4ACAq4kOAKqFDgDd4ACA4eAAgK+NDgCujQ4ArY0OAKyNDgDl4ACA6eAAgO3gAIDx4ACA9eAAgPngAID94ACAAeEAgAXhAICCHQAAgR0AAIAdAAAJ4QCADeEAgBHhAIC+tAEAqK0BAKnVAQCq1QEAqwUBAKwdAQCtBQEArg0BAK8FAQCGgAEAhxgBABnhAIAd4QCAIeEAgCXhAIAp4QCALeEAgLiFAAC5jQAAuoUAALudAAC8hQAAvY0AAL6FAAC/vQAAsH0BALHhAACy5QAAs/0AALTtAAC13QAAttUAALe9AACzXQIAMeEAgDXhAIA54QCAPeEAgLaFAgC1lQIAQeEAgLslAwC6uQIAReEAgEnhAIC/GQMAvikDAL0pAwC8MQMAvswEAKMZAgBN4QCAUeEAgKbBAgBV4QCAWeEAgKXRAgCq/QIAq2EDAF3hAIBh4QCArm0DAK9dAwCsdQMArW0DAKgpAwCpKQMAqjkDAKs5AwCsKQMArSkDAK6dAACvlQAAZeEAgGnhAIBt4QCAceEAgHXhAICCqQEAga0BAICtAQC4mQAAua0AALqlAAC7bQAAvHUAAL19AAC+dQAAv20AALDtAACx9QAAsvUAALPFAAC03QAAtb0AALa1AAC3qQAA4XgBAOEcDgDjEAAA4zwOAHnhAIB94QCAvhQEAIHhAICErAIAieEAgId4BQCGDAUAjeEAgJHhAIDvvAAA70gOALPxAgCV4QCAmeEAgJ3hAICh4QCAtukCALXhAgCl4QCAu3EBALppAQCp4QCAhKAEAL85AQC+WQEAvVEBALxhAQCt4QCAhIwEALHhAICEADgAteEAgLnhAIC94QCAweEAgKqJDgCriQ4AqLkOAKmxDgCu/Q4Ar+EOAKz5DgCt9Q4Asq0OALNlDgCwkQ4AsaUOALZ9DgC3ZQ4AtH0OALV1DgC6XQ4Au+UNALhdDgC5VQ4AvuENAL/pDQC8/Q0AvfUNAKOxBQCF4QCAxeEAgMnhAIDN4QCApqkFAKWhBQDR4QCAqzEGAKopBgDV4QCA2eEAgK95BgCuGQYArREGAKwhBgDd4QCA4eEAgOXhAIDp4QCAgB0AAIEJAACCOQAA7eEAgPHhAID14QCAhsgAAIcMAwD54QCA/eEAgAHiAIAF4gCAqKUHAKm1BwCqvQcAq8kHAKzZBwCt2QcArskHAK/BBwC+oAAACeIAgA3iAIAR4gCAFeIAgBniAIAd4gCAIeIAgLjNAAC51QAAutUAALvlAAC8/QAAvZUAAL6dAAC/lQAAsIkHALFlBwCyYQcAs30HALRlBwC1bQcAtmUHALf1AACzNQYAJeIAgCniAIAt4gCAMeIAgLZZBgC1UQYANeIAgLuhBgC6TQYAOeIAgD3iAIC/qQYAvqEGAL2pBgC8tQYAQeIAgEXiAIDv8AUASeIAgE3iAIBR4gCAVeIAgFniAICAPQAAgQkAAIIdAABd4gCA4cgGAGHiAIDjSAQAZeIAgKO1BgBp4gCAhigAAIdAAQBt4gCAptkGAKXRBgBx4gCAqyEGAKrNBgB14gCAeeIAgK8pBgCuIQYArSkGAKw1BgB94gCAs70BAIHiAICF4gCAtnkBAIniAICN4gCAtXkBALpVAQC7XQEAkeIAgJXiAIC++QAAv/kAALxFAQC9+QAAqHECAKlxAgCqcQIAq3ECAKy1AgCtvQIArrUCAK+tAgC+rDwAmeIAgJ3iAICh4gCApeIAgKniAICt4gCAseIAgLhpAwC5aQMAugkDALsJAwC8HQMAvQUDAL4NAwC/BQMAsNUCALHdAgCy1QIAs2kDALR5AwC1eQMAtmkDALdhAwC14gCAueIAgL3iAICj9QIAweIAgKUxAgCmMQIAxeIAgMniAIDN4gCAqh0CAKsVAgCsDQIArbEDAK6xAwCvsQMA7xgCAIIVAACBbQAAgG0AANHiAIDZ4gCAhvg8AIcYAwDd4gCA4eIAgOXiAIDp4gCA42wHABXhAIDhaAEA7eIAgKiFAgCplQIAqpUCAKulAgCsvQIArdUCAK7RAgCv0QIA8eIAgPXiAID54gCA/eIAgAHjAIAF4wCACeMAgA3jAIC4dQEAuX0BALp1AQC7zQEAvNUBAL3dAQC+yQEAv8EBALC1AgCxvQIAsoECALOBAgC0VQEAtV0BALZVAQC3TQEA4bQGABHjAIDj9AYAFeMAgIQYPQAZ4wCAHeMAgCHjAIAl4wCAKeMAgC3jAIAx4wCANeMAgDnjAIDvWAYAPeMAgIF9AACAcQAAQeMAgIIFAABJ4wCATeMAgO+AAQC+VDwA4ZABAFHjAIDjfAYAVeMAgFnjAIBd4wCAhtg8AIf0PACjnT0A1eIAgEXjAIBh4wCAZeMAgKbVPQCltT0AaeMAgKv5PQCq8T0AbeMAgHHjAICvGT4ArhE+AK3VPQCs1T0AdeMAgLOhPgB54wCAfeMAgLatPgCB4wCAheMAgLWxPgC6ST8Au0k/AInjAICN4wCAvkk/AL9JPwC8ST8AvUk/AKhVPgCpZT4Aqm0+AKtlPgCsfT4ArWk+AK65PwCvuT8AkeMAgJXjAICZ4wCAneMAgKHjAICl4wCAqeMAgK3jAIC4VT8AuV0/ALpVPwC7bT8AvHU/AL19PwC+dT8Av20/ALDJPwCxyT8Astk/ALPZPwC0yT8Atck/ALZ9PwC3cT8AghUAAKPhPwCAsQEAgbEBAKbtPwCx4wCAvtABAKXxPwCqCT4Aqwk+AITkAQC14wCArgk+AK8JPgCsCT4ArQk+ALPdPAC54wCAhugAAIfMAQC94wCAtpU8ALX1PADB4wCAu7k8ALqxPADF4wCAyeMAgL9ZPwC+UT8AvZU8ALyVPACoUT4AqVE+AKptPgCrYT4ArGE+AK1hPgCulQEAr40BAISgAQDN4wCA0eMAgNXjAIDZ4wCA3eMAgOHjAIDl4wCAuKkBALmpAQC6aQEAu2kBALx5AQC9eQEAvmkBAL9pAQCw/QEAsc0BALLFAQCzrQEAtLkBALW5AQC2rQEAt6UBALPlPQDp4wCA7eMAgPHjAID14wCAtuE9ALXpPQD54wCAuwkCALo5AgD94wCAAeQAgL99AgC+fQIAvXkCALwRAgAF5ACAo6E9AAnkAIAN5ACApqU9ABHkAIAV5ACApa09AKp9AgCrTQIAGeQAgB3kAICuOQIArzkCAKxVAgCtPQIAgOkAAIHpAACCHQAAvsADAO/kAgAh5ACAh1QDAIY8BADjEAEAKeQAgOH4AQAt5ACAMeQAgDXkAIA55ACAPeQAgEHkAIBF5ACASeQAgLORAwBN5ACAtbkDALZ9AwBR5ACAVeQAgFnkAIC6WQMAu1kDALxJAwC9SQMAvv0AAL/1AACoRQIAqVUCAKpVAgCrZQIArH0CAK2xAgCusQIAr7ECAIRsBQBd5ACAYeQAgGXkAIBp5ACAbeQAgL5wBQBx5ACAuF0BALltAQC6ZQEAuw0BALwZAQC9GQEAvg0BAL8FAQCw0QIAsdECALLRAgCz0QIAtHUBALV9AQC2dQEAt20BAOFAPwDjvAAA4wg+AOFsPgB15ACAeeQAgH3kAICB5ACAheQAgInkAICN5ACAkeQAgL5sBwDvVAAA75w+AJnkAICjnQIAgmkAAIFhAACAaQAAneQAgKZxAgCltQIAoeQAgKtVAgCqVQIAhsgEAIfsBACv+QEArvEBAK1FAgCsRQIAqKUGAKmpBgCquQYAq7kGAKypBgCtqQYArtkGAK/ZBgCV5ACApeQAgKnkAICt5ACAseQAgLXkAIC55ACAveQAgLhxBwC5cQcAunUHALvdBwC8xQcAvc0HAL7FBwC//QcAsKkGALG1BgCytQYAs40GALSVBgC1UQcAtlEHALdRBwCzMQYAweQAgMXkAIDJ5ACAzeQAgLYpBgC1IQYA0eQAgLtxBgC6bQYA1eQAgNnkAIC/lQcAvlEGAL1ZBgC8YQYA3eQAgKN1BgDh5ACA5eQAgKZtBgDp5ACA7eQAgKVlBgCqKQYAqzUGAPHkAID15ACArhUGAK/RBwCsJQYArR0GAIANAACBFQAAgh0AAPnkAID95ACAAeUAgITcAQAF5QCAhoAAAIcgAQAJ5QCADeUAgBHlAIAV5QCAGeUAgB3lAIAh5QCA43QEACXlAIDhyAUAKeUAgC3lAIAx5QCANeUAgDnlAIA95QCAQeUAgEXlAIBJ5QCA77QEAE3lAIBR5QCAqD0GAKlVBgCqVQYAq6kBAKy5AQCtuQEArqkBAK+pAQCErAEAVeUAgFnlAIBd5QCAYeUAgGXlAIBp5QCAbeUAgLhtAQC5BQEAugEBALsBAQC8BQEAvQ0BAL4xAQC/MQEAsNkBALHZAQCybQEAs2UBALR9AQC1ZQEAtmUBALdVAQCBvQMAgL0DALPVBQCCGQAAtTkCAHHlAIC+VAMAtjECAHnlAIB95QCAuxUCALoVAgC9uQIAvLECAL+pAgC+sQIAgeUAgKZpAgClYQIAhAAMAKONBQCF5QCAhvgMAId8AwCv8QIArukCAK3hAgCs6QIAq00CAKpNAgCJ5QCAjeUAgJHlAICV5QCAmeUAgJ3lAIDjIAEAoeUAgOGgAQCl5QCA70ACAKnlAICt5QCAseUAgLXlAIC55QCAveUAgMHlAICz8QMAxeUAgCXkAIDJ5QCAzeUAgLbpAwC14QMA0eUAgLu1AwC6tQMA1eUAgNnlAIC/lQMAvpUDAL2lAwC8pQMAqCkCAKkpAgCqOQIAqzkCAKwpAgCtKQIArlkCAK9VAgCAzQEAgQkAAIIZAADd5QCA4eUAgL58DQCHtA0AhhwMALgxAgC5PQIAujUCALvpAgC8+QIAvfkCAL7pAgC/6QIAsDECALExAgCyMQIAszECALQRAgC1EQIAthECALcRAgDp5QCA7eUAgPHlAID15QCA+eUAgP3lAIAB5gCA79QGAAXmAIDhVAYACeYAgOOkAACsDBUADeYAgBHmAIAV5gCAo/ECABnmAIAd5gCAIeYAgCXmAICm6QIApeECACnmAICrtQIAqrUCAC3mAIAx5gCAr5UCAK6VAgCtpQIArKUCAKghDgCpIQ4AqkkOAKtZDgCsaQ4ArWkOAK6ZDgCvmQ4A5eUAgDXmAIA55gCAPeYAgEHmAIBF5gCASeYAgE3mAIC49Q4Auf0OALr1DgC7iQ4AvJ0OAL2FDgC+hQ4Av7UOALDpDgCx6Q4Asv0OALPxDgC01Q4Atd0OALbVDgC3zQ4As8EOAIIVAACBtQAAgLUAAFHmAIC26Q4AteEOAL4QAAC7LQ4Aui0OAIRkAwBV5gCAvxkOAL4RDgC9JQ4AvCkOAFnmAICjhQ4AhogAAIdsAwCmrQ4AXeYAgGHmAIClpQ4AqmkOAKtpDgBl5gCAaeYAgK5VDgCvXQ4ArG0OAK1hDgCziQ4AbeYAgHHmAIB15gCAeeYAgLaBDgC1iQ4AfeYAgLuVDgC6jQ4AgeYAgIXmAIC/+Q4AvvEOAL2FDgC8hQ4AieYAgI3mAICR5gCAleYAgOMMDQCZ5gCA4RgNAJ3mAIDvrAwAoeYAgKXmAICp5gCAreYAgLHmAIC15gCAueYAgKgBDgCpAQ4AqgEOAKsBDgCsAQ4ArQEOAK4BDgCvPQ4AgN0AAIEJAACCGQAAveYAgMHmAICEPAEAvnQAAMnmAIC4HQ4AuS0OALolDgC76QEAvPkBAL35AQC+6QEAv+kBALBJDgCxUQ4AslEOALNRDgC0NQ4AtT0OALY1DgC3LQ4Ao4kNAM3mAICGrAQAhzwDANHmAICmgQ0ApYkNANXmAICrlQ0Aqo0NANnmAIDd5gCAr/kNAK7xDQCthQ0ArIUNAOHmAICznQIAhEgDAL5ABAC2VQMA5eYAgOnmAIC1sQIAunEDALt5AwDt5gCA8eYAgL4xAwC/MQMAvFEDAL1RAwCwkQMAsZkDALKhAwCzoQMAtNEDALXRAwC20QMAt9EDALj1AwC5+QMAus0DALvFAwC83QMAvcUDAL7NAwC/xQMA9eYAgPnmAID95gCAAecAgIV8GQAF5wCACecAgHXlAICoIQIAqTECAKoxAgCrBQIArB0CAK3xAwCu8QMAr/EDAA3nAIAR5wCAFecAgBnnAIDvUAAAHecAgCHnAIAl5wCA44QAACnnAIDh+AEALecAgIAVAACBGQAAggUAADHnAICjmQMAOecAgIZoBACHYAUAPecAgKZRAgCltQMAQecAgKt9AgCqdQIARecAgEnnAICvNQIArjUCAK1VAgCsVQIATecAgFHnAIBV5wCAWecAgF3nAIBh5wCAZecAgO/4AQC+bAQA4YAOAGnnAIDjFAEAbecAgHHnAIB15wCAeecAgH3nAICB5wCAhecAgLPdAQCJ5wCAtf0BALb1AQCN5wCAkecAgJXnAIC6sQEAu4UBALydAQC9NQEAvj0BAL81AQCpBQYAqLkFAKsVBgCqHQYArT0GAKw9BgCvTQYArl0GADXnAICCHQAAgR0AAIAdAACZ5wCAnecAgKHnAICl5wCAuUEHALidBgC7QQcAukkHAL1FBwC8WQcAv0UHAL5FBwCxCQYAsD0GALOpBgCyAQYAtbkGALSxBgC3rQYAtrEGAKORBgCEjAIAhigAAIfAAwCp5wCAprkGAKWxBgCt5wCAq8kGAKr9BgCx5wCAtecAgK95BgCucQYArXkGAKzRBgC55wCAs5kHAL3nAIDB5wCAtlEHAMXnAIDJ5wCAtbEHALptBwC7dQcAzecAgNHnAIC+WQcAv0UHALxtBwC9ZQcA1ecAgNnnAIDd5wCA4ecAgOXnAIDp5wCA7ecAgO+oBQDx5wCA4TQFAPXnAIDjdAUA+ecAgP3nAIAB6ACABegAgKMdBgCCLQAAgRUAAIAdAAAJ6ACAptUGAKU1BgAN6ACAq/EGAKrpBgAR6ACAhCgBAK/BBgCu3QYAreEGAKzpBgCoxQYAqdUGAKrVBgCr5QYArP0GAK0VBgCuHQYArxUGAL7sAQAZ6ACAhggAAIcgAAAd6ACAIegAgCXoAIAp6ACAuH0GALkFBgC6DQYAuwUGALwBBgC9CQYAvjkGAL85BgCwbQYAsXUGALJ9BgCzdQYAtFkGALVFBgC2TQYAt0UGAKiRAgCpmQIAqqECAKuhAgCs0QIArd0CAK7VAgCvyQIALegAgDHoAIA16ACAvyweADnoAIA96ACAQegAgEXoAIC4VQMAuV0DALppAwC7ZQMAvGEDAL1hAwC+YQMAv2EDALC5AgCxjQIAsoUCALNtAwC0dQMAtX0DALZ1AwC3bQMASegAgE3oAICzIQIAUegAgLVRAgCEiAMAVegAgLZVAgDF5gCAvigcALtBAgC6dQIAvbEDALxZAgC/sQMAvrkDAKNpAgBZ6ACAXegAgGHoAIBl6ACAph0CAKUZAgBp6ACAqwkCAKo9AgBt6ACAcegAgK/5AwCu8QMArfkDAKwRAgCopQIAqbUCAKq9AgCrtQIArK0CAK01AQCuPQEArzUBAL4sHAB16ACAeegAgH3oAICB6ACAiegAgIdoHQCGHB0AuIUBALmNAQC6hQEAu50BALyNAQC9vQEAvrUBAL95AACwUQEAsVEBALJRAQCzUQEAtPEBALXxAQC29QEAt+UBAO/YAACCtQAAgaUAAIClAACN6ACAkegAgJXoAIDvxAYAmegAgOH0BgCd6ACA4zgBAOPMAACh6ACA4SgBAKXoAICp6ACAtuUBALV1AgCEQBwAs2UCAK3oAICx6ACAtegAgL9lAQC+ZQEAvdUBALzVAQC7xQEAusUBALnoAIC96ACAo7UdAIXoAIDB6ACAxegAgMnoAICmNR4ApaUdAM3oAICrFR4AqhUeANHoAIDV6ACAr7UeAK61HgCtBR4ArAUeANnoAIDd6ACA4egAgOXoAICADQAAgTUAAII9AADp6ACA7egAgPHoAIC1BQAAdxoAgOG0AgCs2AIAtQUAAHsaAICotR8AqRUfAKodHwCrFR8ArDEfAK09HwCuLR8AryEfAOG0AgCs2AIAtQUAAH8aAIDhtAIArNgCALUFAACDGgCAuNEAALnZAAC64QAAu+EAALyRAAC9kQAAvpEAAL+RAACwIR8AsTEfALIxHwCzMR8AtAkfALUJHwC28QAAt/EAAOG0AgCs3AIA71QdALUdAACHGgCA4bwCAKzQAgC1KQAAoyUBAKKRAwChFR0AoA0dAOGAHgCLGgCA47wdAOHEAgCz1R4AtQkAAKzYAgCPGgCA4bwCALb9HgC1+R4ArOACALu1HgC6pR4AtQUAAJMaAIC/jR4Avo0eAL2lHgC8pR4AoxUeAOG8AgCs0AIAtREAAI9pJQCmPR4ApTkeAJcaAICrdR4AqmUeAOG0AgCseAEAr00eAK5NHgCtZR4ArGUeAJvdFACa5RUAmQEXAJjhEACfcR8AnnkZAJ35GQCcARsAk+UtAJIRLwCRbSkAkG0pAJf5EQCW8REAlYUsAJSZLQC1JQAA4ZQCAILxJgCDjSoAhJUqAIXhLACGHS4Ah3kuAKy0AgCbGgCAilUvAIspEgCMORIAjRkTAI7xFACPHRYAtQUAAJ8aAICSVRcAk5EYAJRxGgCV+RoAlvkcAJd9HgCC4AMAlgsAgJpVHgCb2QAAnHUCAIYMAIC2DACAuIkKAKwBBACthQYAroEGAMwQAgDMfAMAuQwAgKMaAIDFDACAyAwAgMsMAIADCwCAgaUyAr8MAIAV6ACAmpUGAJtVIwK8kQYAvbEAAL6RBgC/rQYAuOkGALmVBgC6kQYApxoAgLTBBgC1zQYAts0GALfdBgCw/QYAseUGALKdAACz5QYAhVTHA6saAICH/AAAuAEKALMaAIDsDACAtxoAgIzlDACNpAEAzPACAMQNAIDHDQCAiRQAALgZCgCLDAAAIA4AgFkOAIC8DACAwgwAgBwKAICRwAEAzgwAgLhtCgDRDACA1wwAgN0MAIDgDACA4wwAgLsaAIAuDQCA6QwAgL8aAIDhpB4AMQ0AgONUHgCvJXMAzCgCAPIMAIDvDACA9QwAgPgMAID7DACAzIACAJS4AwD+DACAkhQCAO9gHgCQAAIAAQ0AgA0NAIC48QoAEA0AgKILAIATDQCAiSkLABYNAICvGgCAvDABAL/EAQC+7AEAGQ0AgMzsAgC4xQoAukQBALAJAIAcDQCAygYAgN8GAIDyBgCAIg0AgPoGAIAlDQCACgcAgC0HAIAYBwCA+QcAgC8HAICvDQCAOgcAgLUNAIBKBwCAtXkAAGoHAIC3cSoCdQcAgLFhAAB3BwCAsw0pApAHAIC96QAAowcAgP0HAICwBwCAuRkrAsYHAIC7WRQCIggAgF0JAIA/CACANQ4AgF4IAIA5AACAhAgAgHEAAIDKCACAKwAAgCMJAIA9AACAXwkAgEMAAIBhCQCASAgAgG0IAIBJAACAAwgAgFMAAIB8CQCAWQAAgCgNAIBfAACAuw0iAtYNAIDMFDYCHwAAgL9lAAC+EQAAvW0AAOgHAICAaQEAgXUBAIJxAQCD3SEChGkHAIWBBwCGgQcAh3EBAIihAQCJrQEAirUHAIuNBwCMlQcAjaUBAE8AAICPpQEAkOEBAJHtBwCSsSECk/0HAJSNBwCVUQYAlvEBAJfZAQCY0QEAmXUGAJp9BgCb1QEAnGkGAJ2ZFAKeUQYAn1EGAKB1FAKhuQYAokkBAKOFLQKkIQEApS0BAKZ1FAKntQYAqKERAqlRFAKqlQYAtyEAgMy8NQLNPDUCbQAAgKoDAICsAwCArwMAgMMhAIDKIQCA4SEAgOghAIDJAACADwAAgLihBgC6BgCAtwYAgMwAAIDUIQCAtQMAgN0FAIAYBgCAugUCALvVAgC46QUAuf0FAL7JAgC/5RcCvA0CAL0BAgCy4QUAs+EFALCNBQCxnQUAtuUFALfpBQC09QUAte0FAKo9BQCrwQUAqD0FAKk1BQCuzQUAr/UFAKzNBQCtxQUAoj0FAKMFBQCg1QIAoTkFAKYdBQCnBQUApB0FAKUVBQC/BgCAm8EFAD4GAIBVBgCAnt0FAJ8xBACcUQIAndUFAHIGAICJBgCApAMAgDYiAIDbAACAoAMAgJIHAIDxBwCA9QcAgJMJAIAFCACACQgAgJkLAICXCQCAsgoAgHIHAICOBwCAmgcAgKUHAICtBwCArQkAgAEPAIAYDwCAJQ8AgMwEMwLNsDACzCAzAs3gMALMEDACzGgwAsxYMALNjDACzGgxAs0UMQLM1DECzRQ2AsxwIALN0CcCzDA2AswkMQLMDDwCzWg/AswYPwLNND8CzBg9As3AMgLMRDwCzBg5Asw4MgLNqDICzIgyAs34MwLMfDMCzUAzAswoMwLNCDMCzMghAs0kJgLMrCYCzEA4AsyYJQLNyDoCzBwkAs0QJALMhDsCzag7AsysJQLNvDoCzKw4Asz4JwLM4DgCzXQ4Ai0PAID2BgCAZw0AgI4NAIDNICoCzBwrAqoGAIAyIgCAzKQgAs2gJwLMOCYCygQAgMw4OgLNPDsCzBA5As1gPgLMoAMAvj0NAL3tLALWBACAu1UjAgcJAIC5PSICzwYAgNwHAIClBACApg0AgLIEAIBvBQCA9AYAgL4EAIB1BQCAr70MAK6ZLgKtpQwAwgUAgKvFIgIDBgCAxAQAgCMGAIDQBACAyAUAgCkGAIBdBgCAowEYAqAEAIAaBwCAHQcAgJ9dDACeUQwAnUUMACcHAICbWSECsgcAgLQHAIC3BwCAuwcAgCoHAIDRBwCA0wcAgJMtJgLWBwCAbwgAgHIIAICPBQwAjnEMAI1lDAB8CACAi0UgAmMJAICJNS8CZgkAgGoJAIB/CACAcwkAgHYJAIC9AwCABiIAgIFdDACAYQwAgAABAIEYAACCAAQACiIAgIQQBwCFFAYAhuQIAIc8AgCILAUAiaQFAIoAeAAOIgCAjJAGABIiAIAaIgCAFiIAgLgRAACRxHsAkkh6AJNMeQAiIgCAzOgCAJbwCQC4OQAAkMAJACoiAICS8AkAzPgCAJS0CQC4DQAALiIAgMwcAgC4BQAAOiIAgMzkAgC4HQAAPiIAgEIiAIBJIgCAYCIAgKiMCACp5HsAZyIAgKt8BADM5AIAuA0AAHEiAIDMlAIAdSIAgLGUewC4CQAAuBUAAMz8AgC15AgAeSIAgMzYAgB9IgCAuAUAALroBQC7hAQAvAB8AL30fwC++H0Av/xyAIAJOgKBDToCggE6AoMFOgKEGToChR06AoYROgKHFToCiCk6AoktOgKKIToCiyU6Aow5OgLMhAIAjjE6Ao81OgKJIgCAkekPAIUiAIDMzAIAuBEAAJUiAIDM3AIAl+UPALgpAAC4MQAAzPgCAJkiAIC4HQAAzDwCAJ0iAIDMLAIAzIQCAKEiAIClIgCAzIQCAKQtDwClVQ8Apl0PAKkiAICoqToCqa06ArjRAACtIgCAuDUAALEiAIDMUAMAr7U6AswsAwDMFAMA1SIAgLMFDwC0HQ8AzAADALYJDwC3CQ8Avmh9ALhtAAC4dQAAzEgDALwpDwDZIgCAviUPAMxQAwCH5Q4AzOg6ArilAQC4uQEAzPA1As2kMwLMgCICzXwlAs2UNgLMBCkCzew7AsxkOgK4+QEAuMEBAInVDgCI1Q4Al7EOALgNAAC1IgCAuB0AALkiAIC4DQAAvSIAgMEiAICfaTsC3SIAgOEiAIC4MQAAzMQCAMz4AgDFIgCAySIAgLjlAADNIgCA0SIAgLjlAAC46QAAuOkAAMzMMwLlIgCAuCUAAMzoMwK4IQAA6SIAgO0iAIC4KQAAzNgCALgRAAC3TQ0Atk0NALU1DgC0NQ4AuGEAAPEiAICxGQ8AsCkOAL/1AwC+UQ0AvVkNALw1DAC7XQ0Aul0NALldDQC4XQ0AgL0KAIHFCgCCFQQAg8kKAMx8BQCF3QoAhtUKAIfNCgDMUAUAifEKAIq5CACLDQgAjBEIAI0VCACOtScCj+UKAJBpCACRbQgAknEIAJNtJALMJAUAlR0IAJaFCgDMHAUAzCgFAJk9CACaiQoAmw0IAJwRCACdFQgAzCwFAMwgBQCgZQoAoW0KAKJlCgDMvAUApLEEAMzoAgCmsQQAuEkHAKiBBAAbIwCAqpkIAKtdCgCsuQgArakEAB8jAICvNQgAsNEIALHxBADQAwCAs40IALQpKAK1IQoAtiEKALchCgC4IQsAuSUIAOkDAIC7KQsAvA0dAr3dDwC+MQsAvzELAIDdCgCpoQEAqrEBAAIEAIAbBACAhRkJAIaZCQCHlQkAiOEJAIklJQIuBACAQQQAgFQEAIBnBACAegQAgI0EAICQrQoAkUkFAJJtBQCTYQUAlGEFAJVtBQCWZQUAlxEFAJg1BQCZPQUAmjUFAJsNBQCcFQUAnR0FAJ4VBQCfCQUAoKkJAKH9BQCi9QUAowEFAKQFBQClDQUApgUFAKc9BQCoBQUAqQ0FAKoFBQCrGQUArIkJAK2pBQCutQkAr/0JALABCQCxfQUAsnUFALMBBQC0aQkAtQEFALYFBQC3PQUAuAUFALnhJQK6AQUAuwEFALzRJQK9PQkAvnkJAL9dCQCDMAUAoXgHAPMEAIDdAACApHgHAKVMBwATAQCAHAEAgIt8BAAgAQCAJAEAgIhIBAAoAQCALAEAgDABAIA0AQCA4QAAgOYAAICyDAcAswAHAOsAAIDwAACAtugHALfgBwD1AACA+gAAgLrgBwC7nAcAvIQHAL2QBwD/AACAn9F+AKPMBAAEAQCACQEAgIMABAAOAQCAhXQEAKUgBAAXAQCAiEwEAM0DAIDwBACAjwUAgK8tBwCNsAcArSEHAKwpBwDiBQCAHQYAgEMGAICwZQcAWgYAgHcGAICOBgCA0wMAgOwDAIAFBACAHgQAgDEEAIBEBACAVwQAgGoEAIC8fAQAgt0rAoPlKwKA/QoAgfkrAoaZCQCHmQkAhOEKAIXhCgCKiQkAi4kJAIiJCQCJiQkAjoUJAH0EAICM4QgAjY0JAJK5KwKTQScCkJkrApHFCwCWyQsAl3UnApTFDQCV0SQCmskLAJvZKgKYyQsAmXkHAJAEAID2BACAnP0LAKABAICkAQCAqAEAgKwBAICwAQCAtAEAgLgBAIC8AQCAwAEAgJw9fwCzFX8AqBEJAMQBAIDIAQCAzAEAgNABAICC2H4A1AEAgNgBAIDcAQCA4AEAgOQBAIDoAQCA7AEAgPABAID0AQCA+AEAgPwBAIAAAgCABAIAgDwJAIBTIgCAogYAgKD1VAKh2VQCoulUAqP1dQCk7XUApZ12AKaVdgCnvXYAqIV2AKntfACqwXwAqyF9AKwhfQCtHX0ArhV9AK8NfQCwdX0AsX19ALJ1fQCzRX4AtF1+ALVNfgC2RX4At3l+ALhJfgC5VX4Aul1+ALtVfgC8TX4AvTV+AL49fgC/LX4AlQcAgKwGAIDaBwCArwYAgLRtfwC1EQAAthUAAAkjAIC8oVgCvTF4AA8jAIDTMQCAPzoAgJ8qAIDDKgCAzyoAgN8qAIDnKgCA8yoAgPsqAIADKwCADysAgGorAICCKwCAkisAgKIrAICyKwCAwisAgOIrAIDmKwCA6isAgB4sAICAVX8AgWV/AIJtfwCDdX8AhJV/AIWdfwCGiX8Ah4F/AIiFfwCJjX8AioV/AIvtfwCM9X8Ajf1/AI7pfwCP6X8AkJl/AJGZfwCSqX8Ak6l/AJS5fwCVuX8Alql/AJepfwCYmX8AmVF+AJoZfgCbGX4AnA1+AJ31fgCe/X4An/V+AKANfgChFX4Aoh1+AKMVfgCkDX4ApTl+AKYpfgCnKX4AqBl+AKllfgCqbX4Aq2V+AKx9fgCtZX4Arm1+AK9lfgCwHX4AsSV+ALItfgCzJX4AtD1+ALUlfgC2WXcAt9V1ALj9eQC56XUAuvl1ALvZeQC86XUAvdV1AL7RdQC/2XUAgDF2AIE9dgCCSXYAg0V2AIRBdgCFTXYAhvl0AId9dgCIoQIAiU12AIpZdgCLuXoAjEl2AI2degCOsQIAjx16AJCRVgKRKXYAkoF2AJPNdgCU2XYAlel2AJbJdgCX0VkCmKF2AJllWgKa8XYAm01aApzRdgCdYXoAnoFWAp/VdgCgdX0AoY1aAqI1VwKjCXYApCF2AKUtdgCmiVoCp5laAqi5WgKpdXYAql13AEYsAIBWLACAXiwAgGIsAIBuLACAiiwAgI4sAICmLACAqiwAgLIsAIDCLACAXi0AgHItAICyLQCAxi0AgM4tAIDSLQCA4i0AgAUuAIAxLgCAPS4AgJlpCgBdLgCAaS4AgG0uAIBxLgCAiS4AgI0uAIC5LgCAgvx6AIPoegDFLgCAzS4AgIZwewCHcHsA1S4AgOUuAID0LgCA/C4AgCgvAIAsLwCANC8AgDgvAIBALwCASC8AgJJAfABYLwCAdC8AgJHkewDsLwCAADAAgAQwAICEMACAiDAAgKvUfACo6HwAqeB8AJwwAICgMACAqDAAgLAwAICisHwAuDAAgMQwAID6MACAzEBJAs0ASQLM/EoCzWhLAgoxAIAeMQCAmzEAgKcxAIC3MQCAwzEAgM8xAIDXMQCAsqh8ALOofADbMQCA3zEAgOMxAIDnMQCAtER8ALVYfACAfQcAgYUHAIKZBwCDmQcAhIkHAIWJBwCGvQcAh7UHAIiNBwCJ5QcAiu0HAIvlBwCM/QcAjeUHAI7tBwCP5QcAkJ0HAJGtBwCSpQcAk70HAJSlBwCVVQEAll0BAJdVAQCYbQEAmXUBAJp9AQCbdQEAnG0BAJ1VAQCeXQEAn1UBAKCtAQChtQEAor0BAKO1AQCkrQEApdUBAKbdAQCn1QEAqO0BAKn1AQCq/QEAq/UBAKztAQCt1QEArt0BAK/VAQCwrQEAsbUBALK9AQCztQEAtK0BALVVAQC2XQEAt1UBALhtAQC5dQEAun0BALt1AQC8bQEAvVUBAL5dAQC/VQEAnzIAgOcyAIDzMgCA9zIAgPsyAID/MgCABzMAgAszAIAfMwCAOzMAgEMzAICDMwCAhzMAgI8zAICTMwCAmzMAgJ8zAIDDMwCAxzMAgOMzAIDnMwCA6zMAgO8zAIADNACAJzQAgCs0AIAvNACAUzQAgJM0AICXNACAtzQAgMc0AIDPNACA7zQAgBM1AIBXNQCAXzUAgHM1AIB/NQCAhzUAgI81AICTNQCAlzUAgK81AICzNQCAzzUAgNc1AIDfNQCA4zUAgO81AID3NQCA+zUAgP81AIAHNgCACzYAgKs2AIC/NgCA8zYAgPc2AID/NgCAnpkMACs3AIAzNwCAOzcAgICtAwCBtQMAgr0DAIO1AwCErQMAhdUDAIbdAwCH1QMAiO0DAIn1AwCK/QMAi/UDAIztAwCN1QMAjt0DAI/VAwCQrQMAkbEDAJKxAwCTsQMAlAEMAJVVDgCWXQ4Al1UOAJhtDgCZdQ4Amn0OAJt1DgCcbQ4AnVUOAJ5dDgCfVQ4AoK0OAKG1DgCivQ4Ao7UOAKStDgCl1Q4Apt0OAKfVDgCo7Q4AqfUOAKr9DgCr9Q4ArO0OAK3VDgCu3Q4Ar9UOALCtDgCxtQ4Asr0OALO1DgC0rQ4AtVUOALZdDgC3VQ4AuG0OALl1DgC6fQ4Au3UOALxtDgC9VQ4Avl0OAL9VDgC8aQQAvXUEAL59BAC/dQQAuEUEALlNBAC6RQQAu3kEALRlBAC1bQQAtmUEALd9BACwHQQAsQ0EALIFBACzfQQArFUEAK1dBACuVQQAr2UEAKhVBACpXQQAqlUEAKtNBACkkQUApZEFAKaRBQCnkQUAoJEFAKGRBQCikQUAo5EFAJxRBQCdUQUAnlEFAJ9RBQCYUQUAmVEFAJpRBQCbUQUAlBEFAJURBQCWEQUAlxEFAJBFBwCRTQcAkkUHAJMRBQCMJQcAjS0HAI4lBwCPPQcAiAUHAIkNBwCKBQcAiz0HAIQlBwCFLQcAhiUHAIc9BwCARQcAgU0HAIJFBwCDPQcAfzcAgIM3AICLNwCAjzcAgJM3AIC/NwCAwzcAgMs3AIDfNwCA4zcAgP83AIAHOACACzgAgC84AIBPOACAYzgAgGc4AIBvOACAmzgAgJ84AICvOACA0zgAgN84AIDvOACABzkAgA85AIATOQCAFzkAgBs5AIAnOQCAKzkAgDM5AIBPOQCAUzkAgFc5AIBvOQCAczkAgHs5AICPOQCAkzkAgJc5AICfOQCAozkAgKc5AICrOQCArzkAgL85AIDXOQCA2zkAgOc5AIDrOQCA7zkAgPM5AID7OQCA/zkAgAM6AIAPOgCAFzoAgB86AIAjOgCAKzoAgC86AIAzOgCAOzoAgICtAQCBtQEAgr0BAIO1AQCErQEAhdUBAIbdAQCH1QEAiO0BAIn1AQCK/QEAi/UBAIztAQCN1QEAjt0BAI/VAQCQrQEAkbUBAJK9AQCTtQEAlK0BAJUNAABDOgCAQyMAgHIsAIB2LACAKyQAgIJ4AgCZBQAAuSMAgILQAgC+mAIAgIAAAIGYAACC5AYAg4gEAITUGwCFlBoAhhgfAJ7pAACIxB4AiQAQAIqoEwCLrBEAjAAoAI20KwCOuCoAj7wpAON0dADjoAIA47gCAJkdAAC9IwCAvnQCAJ4JAADv2AIAgmwCAMEjAICZDQAA4wQCAO/ocQDvcAIA79QCAL6MAADjBAMAxSMAgOMcAwDJIwCA4yADAM0jAIDjTAMA0SMAgO/kAwDVIwCA74QDANkjAIDv2AMA3SMAgO/sAwDhIwCA48gDAOPcAwDvoAQA4+gDAEM3AIDjqAMA5SMAgO+kBADXAAAA70wDAOkjAIDjAAQA7yADAO88AwDjeAQA70QDAO0jAIDxIwCA9SMAgPkjAIDvWAQA/SMAgO9cBADjgAQA44AEAAEkAIDjmAQA73QEAAUkAIAJJACADSQAgBEkAIAcJACA72AEAOOIBAAgJACAvQAAgMIAAIA3JACAJCQAgHMpAIAOJQCAcSUAgLQlAIDgJQCA46QEAO9EBAAKJgCAgAlLAoZIAQCe7QIAgnACAL58AgCeEQEAmR0BAIJcAgCPQAEAmSkBAI1sAQC+YAIAi3gBAJ45AQCCGAIAvggCAJfUAgCZUQEAldQCAJ5ZAQCT1AIAvlgCAJHUAgCCVAIAn7gCAJYAAACdqAIAmXEBAJu4AgCeaQEAmbQCAJlZAQCCmAIAppQCAJ6tAQCkeAIAgmgCAJ65AQCheAIAmbEBAK+EAgAvJgCAvlgCAIJwAgC+XAIARCYAgJmNAQCokAIAvnwCAJ7xAQC1sAIAmfEBAL5QAgCykAIAtoUCAJmFAQC4XQkAuYUCALqNAgBMJgCAu/QEAIJcAgCexQEAuPQEAL58AgCeXQYAgmACAJllBgC+cAIAgmwCAJ5xBgCZnQYAvnwCAJ6lBgCCYAIAmakGAL58AgCesQYAghwCAL4UAgCZyQYAvkwCAIJMAgCawQYAntkGAJ/ZBgCCvAIAvrACAJn1BgCC3AUAvogCAJrJBgCe5QYAn9EGAIK4AgCCqAIAmTEGAOP0AgDj/AIAmjkGAJ8lBgCeJQYAniEGAJ8hBgC+GAIAmR0GAJoVBgC+DAIAmXEGAO8kAgDv4AIAmnEGAJ4BBgCfCQYA4xQCAJkVBgDjLAIAgnwCAJ4BBgBoJgCA7/ACAJkFBgC+fAIAng0GAHAmAICCdAIA76wCAIKUAwCb0QcA4/ADAL5QAgCZ6QcAn80HAOM0AgCdXAIAnMkHAJ7FBwDv7AIAgmACAJnFBwC+cAIA78QCAJ7RBwCCEAIA43QCAL5sAgCZpQcAvlQCAJ69BwCZpQcAgmgCAOMMAgCekQcAmZkHAIJoAgDv2AIA78wCAL6IAgCCkAIAvpgCALsEAACeeQcAuQAAAJkpBgC/XAAA4/QCAL0EAACeOQYAs7gDAO8oAgCxeAMAgnwCALc8AACZAQYAtfwDAJ4JBgCrSAMAvgACAOO4AgCZIQYArxwDAIJ8AgCtQAMAvkQCAJ4NBgCCbAIAvpwBAJkxAQCCgAEApqgCAO98AgCCvAEA46QBAOPYAQDj5AEAntECAOPoAQCZ5QIAvnwCAJ7tAgDvwAIAmQkAAL5sAgB4JgCA7wwBAO8QAQDvKAEAnhEAAOO0AQDjHAIAmS0AAL5sAgCCfAIAglACAJ49AADjAAIAmeUAAIomAIC+aAIA7/wCAO+gAgDvwAIAnv0AAJnxAADjAAIAliYAgIKQAgCeJgCAvnwCAJ4ZAADjDAIAglgCAJkNAACCBAIA7+QCALQmAIDv2AIAvnQCAJ4VAACZZQEA42wCAJnJAQCanQEAvmQCAJ7dAQCfgQEA4wwCAIJYAgCZlQEAvrQDAO/0AgCalQEA78QCAIKsAwCwRSoAgvQDAIHIAgDj+AMAvjACAPImAIDj5AMAhMQCAL44AgCGEAIA78ACAIgwAgCeXQAAn1UAAJqxAADvxAIAj3QCAJlBAACePQAAn8UAAIKMAgCSGAIAldgDAAMnAICeyQAAn8kAAJr1AACY+AMAm2ADAJn9AACCqAMAJScAgDAnAICAJwCAlCcAgOPEAgC+IAIAvlACALYnAIDEJwCA4+ACAAwoAICZPQAAnn0AAO/cAgCaHQAAGigAgO/EAgCvrAMAghACALEAHACwlAMAmSkAALJMHACeTQAAn2UAAHcpAIDjaAIAeykAgL50AgCeCQAA7/ACAL08HACCbAIAv80fAJn9HwB/KQCAvnQCAJ4JAADjFAIAgmwCAIMpAICZDQAAvkQCAJ41AACCaAIAmQUAAI8pAIC+cAIA7yUAgJ4VAADv2AIA42wCAP8YAIADGQCA47AdACcaAIAHGQCAKxoAgC8aAIALGQCAMxoAgDcaAIA7GgCA77gCAD8aAIBDGgCA74ACALHFAAAPGQCAs9kAALLFAAC1yQAAtMEAALf5AAC2wQAAuWEAALghAAC7wQAAuskAAL3FAAC82QAAv30AAL7FAAATGQCARxoAgFMZAIAXGQCAGxkAgB8ZAIBnGQCA7xB4A7MZAIDh3E0DtxkAgONYeQO7GQCAvxkAgMMZAIDHGQCAgMkBAIHJAQCC2QEAg9kBAITJAQCFdQIAhgEEAIcdBQCIJQUAiTUFAIo9BQCLbQUAjHUFAI1lBQCObQUAj90BAJCpAQCRtQEAkr0BAJO1AQCUrQEAlVUDAJZdAwCXVQMAmG0DAJl1AwCafQMAm3EDAJxRAwCdUQMAnlEDAJ9NAwCgtQMAob0DAKK1AwCjxQMApMEDAKXBAwCmxQMAp/0DAKjFAwCpzQMAqsUDAKvZAwCsyQMArTUDAK49AwCvNQMAsE0DALFVAwCyXQMAs1UDALRNAwC1dQMAtn0DALd1AwC4TQMAuVUDALpdAwC7VQMAvE0DAL01AwC+PQMAvzEDAMsZAIDPGQCA0xkAgNcZAIDbGQCA3xkAgPAQAgDjGQCA5xkAgOsZAIDvGQCAgpwCAPMZAID3GQCA+xkAgP8ZAICc9TYAnf02AAMaAICRaAIArxkAgEsZAIC6ZdUATxkAgEsaAIBPGgCAUxoAgFcaAIDwSAIAWxoAgF8aAICRkAIAYxoAgFcZAIBnGgCAaxoAgFsZAIBfGQCAYxkAgGsZAIBvGQCAcxkAgHcZAIB7GQCAfxkAgIMZAICHGQCAixkAgI8ZAICTGQCAuo3VAJcZAICbGQCAnxkAgG8aAIBzGgCAoxkAgILEAgCnGQCAqxkAgAcaAIALGgCADxoAgPA4AwDh3NICExoAgONUxgIXGgCAGxoAgB8aAIAjGgCAqyoAgGItAICvKgCAxyoAgLMqAICjMwCAuyoAgO+k4AKjKgCA4xSHAuGQ8wLhjJ0C45D3AuGMqwLhkLYC4wBTA+OAogIjGQCA7S0AgO+0SwPvvLMC7ySLAnYtAIAIAgCA7ySXApEIBwAOAgCAFAIAgBoCAIAgAgCAJgIAgCwCAIAyAgCAOAIAgD4CAIBEAgCASgIAgFACAIBWAgCANAMAgDoDAIDhgHgC4wAeAuMUagLhEBMC4aAPAkADAIDjhA4C7yw/AkYDAIDhUDsC7zQ7AuM8IgJMAwCA7ygfAu8MEgJSAwCAKxkAgC8ZAIBYAwCAXgMAgDMZAIA3GQCAdgMAgIIDAICIAwCAjgMAgJQDAICaAwCAfAMAgGQDAIBtAwCAXAIAgDsZAIA/GQCAdAIAgGgCAIBDGQCARxkAgLwCAIB6AgCAmAIAgGICAICSAgCAbgIAgKQCAIDUAgCA8gIAgOwCAICAUQYAgVEGAIJRBgCDUQYAhHEGAIV9BgCGdQYAh20GAIhVBgCJXQYAiq0HAIuhBwCMoQcAjaEHAI6hBwDgAgCALgMAgMICAICSfRQAkwEUAJTNBwCV9QcAlv0HAJf1BwCYzQcAmdUHAJotFACb2QcAnM0HAJ2RBwCejQcAnykUAJklAQCYJQEAmyUBAJolAQCdJQEAnCUBACcZAICeJQEAkcEGAJDxBgCT1QYAkt0GAJU9AQCUPQEAlyUBAJYlAQCJ5QYAiOUGAIvlBgCK5QYAjeUGAIzlBgCP5QYAjuUGAIHlBgCAHQYAg+UGAILlBgCF5QYAhOUGAIflBgCG5QYAuaUDALilAwC7pQMAuqUDAL2lAwC8pQMAv6UDAL6lAwCxpQMAsKUDALOlAwCypQMAtaUDALSlAwC3pQMAtqUDAKmxAQCoqQEAq7EBAKq5AQCtbQEArKkBAK8dAQCuHQEAoakBAKDZAQCjsQEAoqEBAKWRAQCkqQEAp5EBAKaZAQDOAgCA5gIAgNoCAIAEAwCAsAIAgPgCAIAiAwCACgMAgJ4CAICAAgCAtgIAgMgCAID+AgCAhgIAgCgDAICqAgCAEAMAgIwCAIAWAwCAHAMAgBYtAID4LgCA1zQAgIcHAIAGBQCAFQUAgCQFAIAzBQCAQgUAgEsFAIBUBQCAXQUAgIIMAwBmBQCAkgUAgJsFAICkBQCA40huA6cFAIDhTG4DqgUAgO/0AQOtBQCAVzoAgLdMAIDnVQCAR2gAgHdxAICnegCAB40AgGefAICXqACA/roAgDXEAIBlzQCAldYAgMXfAIBCuwCAS64AgBelAID/KgCAlisAgKcqAIDrKgCATjEAgA4xAIBbNACA4iwAgBMzAICXNwCAbzQAgCosAICfNACAqzMAgB84AIBmKwCAkiwAgAcyAIA3OQCAKisAgLorAICrMQCAyS4AgNYsAIBmLACARS4AgDkuAID7MwCAJisAgLop0ACrNwCAgiwAgNotAICwBQCAswUAgLYFAIDh1D8D4VgaA+PcLwPjUA4D4RTyA+FA0wPjQOoD40DDA7kFAIDlBQCA73jrA+9c8gPoBQCA6wUAgO9E3gPvmCUD4bSLA+E8lwPjfKID45iLA+EwQQDhUKwD4xx/AOOIRgDuBQCA8QUAgO84ewDv4EEA9AUAgPcFAIDvzIoD7yCHA4C1GACByRgAghULAIMtCwCE4Q4AheEOAIbhDgCH/RgAiN0OAImZGgCK6RsAiy0dAIzFGwCN8RsAjn0QAI/hGgCQ6RsAkUUPAJJNDwCTRQ8AlF0PAJVFDwCWTQ8Al0UPAJh9DwCZRQ8Amk0PAJuZGgCcWQ8AnVkPAJ5JDwCfSQ8AoLkPAKG5DwCiyQ8Ao8kPAKS1CwClvQsAprULAKfVDwCo7Q8AqfUPAKr9DwCr9Q8ArO0PAK3VDwCu0Q8Ar9EPALCxDwCxsQ8AsrEPALOxDwC0cQ8AtXEPALZpDwC3aQ8AuAEPALkBDwC6GQ8AuxkPALzxAQC98QEAvvEBAL/xAQD6BQCA/QUAgAAGAIAgBgCA4QQAgIAFAIDTBQCADgYAgDQGAIBLBgCAaAYAgH8GAICWBgCA3QMAgPYDAIAPBACAEgcAgEQIAIBBCACAPwcAgD8kAIB4JACAqSQAgM4kAIC/JgCAyiYAgM4mAIDSJgCA1iYAgDUoAIB0KACAnCgAgKAoAIDFKACAzSgAgOkoAID7KACA/ygAgAMpAIAbKQCANikAgPAUNgBRKQCAHysAgEMkAIBQJACAXSQAgGokAIB8JACAiSQAgJskAICtJACAvSQAgNIkAIDcJACA6iQAgPQkAIABJQCAEiUAgBwlAIB1JQCAfCUAgColAICGJQCAgBEDAIERAwCCEQMAgxEDAIQxAwCFMQMAhjEDAIcxAwCIEQMAiREDAIoRAwCLEQMAjHEDAI1xAwCOcQMAj3EDAJARAwCREQMAkgEEAJMVAwCUDQMAlVUGAJZdBgCXVQYAmG0GAJl1BgCafQYAm3UGAJxtBgCdNQYAnj0GAJ81BgCgzQYAodUGAKLdBgCj1QYApPEDAKXxAwCm8QMAp/EDAKjRAwCp+QYAqikGAKspBgCsOQYArTkGAK7NAwCvxQMAsL0DALFFAwCyTQMAs0UDALRdAwC1RQMAtk0DALdFAwC4fQMAuUUDALpBAwC7fQYAvGUGAL1tBgC+ZQYAv1EDAKkVDwCoAQ8Aq00PAKpNDwCtRQ8ArEUPAK+pDQCusQ0AoXULAKBhCwCj6QsAok0LAKXhCwCk+QsApzkPAKZdCAC5mQ0AuJENALupDQC6kQ0AvbkNALyxDQA3JQCAvrENALHZDQCw0Q0As6kNALLRDQC1uQ0AtLENALepDQC2sQ0APiUAgE4lAIBhJQCAuCUAgMIlAICXJQCApyUAgNYlAICB5Q0AgOUNAIPlDQCC5Q0AheUNAITlDQCH4Q0Ahi0YAJlFDQCYuQ0Am0UNAJpFDQCdSQ0AnEUNAJ9xDQCefQ0AkYENAJC5DQCTgQ0AkokNAJWBDQCUmQ0Al4ENAJaJDQDmJACAJiUAgJMlAIDSJQCA5CUAgA4mAIAzJgCASCYAgPYlAIAAJgCAEiYAgB8mAIA3JgCAVCYAgF4mAIB8JgCAUCYAgGwmAIB0JgCAhiYAgJImAICaJgCAqSYAgOQmAICiJgCAuCYAgK0mAIDDJgCA2iYAgOgmAIAHJwCAFycAgCEnAIBVJwCAmCcAgO0nAIBVKQCAYykAgGcpAIBrKQCA9iYAgDQnAIBEJwCATicAgCknAIBZJwCAaScAgIQnAIB2JwCAnCcAgMgnAIDPJwCArCcAgNknAIDjJwCAuicAgB4oAIAQKACA8ScAgCsoAID4JwCAAigAgDkoAIBGKACAUCgAgFooAIBkKACAeCgAgIUoAICMKACApCgAgKsoAIC4KACA0SgAgNsoAIDtKACABykAgBQpAIAfKQCAKSkAgDopAIBBKQCAWSkAgMMDAIDmBACAhQUAgNgFAIATBgCAOQYAgFAGAIBtBgCAhAYAgJsGAIDjAwCA/AMAgBUEAIAoBACAOwQAgE4EAIBhBACAdAQAgIcEAICaBACAAAUAgA8FAIAeBQCALQUAgDwFAIBmCACAJwgAgMEGAID/BwCAIAkAgOM4EwA2CQCALQgAgDAIAIA0CACAJAcAgOkuAIDXMACA5i0AgMgwAIBSMQCAKgkAgO/kEwAJCQCA4g0AgNIIAICGCACAMQcAgEwHAID8BgCADQgAgJcIAIAtCQCADAkAgOYNAIDyDQCA3ggAgJwIAIAVBwCAiQgAgFUHAID/BgCAqQcAgJckAID2DQCA5QgAgCoIAICfCACAWwgAgBgJAID6DQCA6AgAgBcIAICiCACA6wgAgBoIAIDMCACApQgAgO8IAIAeCACAzwgAgKkIAID6CACAAAkAgIsHAICNCACAWQcAgAMHAIBACQCARAkAgEwJAIA5CQCAGwkAgP4NAID3CACAMAkAgA8JAIDqDQCA1QgAgJEIAIBgBwCAMwkAgBIJAIDuDQCA2AgAgJQIAIBjBwCAsAgAgGYHAIDjQBIA4/ATAOP0EwDj2BMA4wgNAOPoEgDjrBIA42QSAO/EDQDv2A0A7zQSAO9YEgDvQBIA79QSAO/IEgDvIBMA7AcAgMwGAIARCACAFAgAgNgGAIDUBgCAJAgAgAcHAIBqCACADAcAgHkIAIA0BwCANwcAgK0IAIC5CACAvAgAgOPQEADjsBAA46gQAON4EQDjTBAA4zAQAOPoEADj/BAA74QQAO+YEADvLBAA7yAQAO8EEADvCBAA73AQAO9MEADjmBMA4zAQAOMwEADjIBAA43wTAONAEwDjWBMA44ATAO/UEwDvqBMA74ATAO98EwDvRBMA7yQTAO8YEwDv7BAAgO08AIH1PACC/TwAg/U8AITtPACFFT0Ahh09AIcVPQCILT0AiTU9AIo9PQCLNT0AjC09AI0VPQCOHT0AjxU9AJBtPQCRdT0Akn09AJN1PQCUbT0AlRU9AJYdPQCXFT0AmC09AJk1PQCaPT0AmzU9AJwtPQCdFT0Anh09AJ8VPQCg7T0AofU9AKL9PQCj9T0ApO09AKUVPQCmHT0ApxU9AKgtPQCpNT0Aqj09AKs1PQCsLT0ArRU9AK4dPQCvFT0AsG09ALF1PQCyfT0As3U9ALRtPQC1FT0AthE9ALcRPQC4MT0AuTE9ALoxPQC7MT0AvBE9AL0RPQC+ET0AvxE9AIDxPACB/TwAgvU8AIMNPwCEFT8AhR0/AIYVPwCHDT8AiDU/AIk9PwCKNT8Aiw0/AIwVPwCNHT8AjhU/AI8NPwCQdT8AkX0/AJJ1PwCTDT8AlBU/AJUZPwCWCT8Alwk/AJg5PwCZOT8Amgk/AJsJPwCcGT8AnRk/AJ4JPwCfCT8AoPk/AKH5PwCiCT8Aowk/AKQZPwClGT8Apgk/AKcJPwCoOT8AqTk/AKoJPwCrCT8ArBk/AK0ZPwCuCT8Arwk/ALB5PwCxeT8Asgk/ALMJPwC0GT8AtRk/ALYJPwC3CT8AuDk/ALk5PwC6CT8Auwk/ALwZPwC9GT8Avgk/AL8JPwCA+TwAgfk8AIJJPQCDST0AhFk9AIVZPQCGST0Ah0k9AIh5PQCJeT0Aikk9AItJPQCMWT0AjVk9AI5JPQCPST0AkDk9AJE5PQCSAQQAk00GAJRVBgCVXQYAllUGAJdNBgCYdQYAmX0GAJp1BgCbTQYAnFUGAJ1dBgCeVQYAn00GAKC1BgChvQYAorUGAKPNBgCk1QYApd0GAKbVBgCnzQYAqPUGAKn9BgCq9QYAq80GAKzVBgCt3QYArtUGAK/NBgCwtQYAsb0GALK1BgCzTQYAtFUGALVdBgC2VQYAt00GALh1BgC5fQYAunUGALtNBgC8VQYAvV0GAL5VBgC/TQYArH0/AK2lPwCurT8Ar6U/AKh9PwCpZT8Aqm0/AKtlPwCkHT8ApUU/AKZNPwCnRT8AoB0/AKEFPwCiDT8AowU/ALydPwC9pT8Avq0/AL+lPwC4nT8AuYU/ALqNPwC7hT8AtN0/ALWlPwC2rT8At6U/ALDdPwCxxT8Ass0/ALPFPwCMYToAjWE6AI5hOgCPYToAiEE6AIlBOgCKQToAi0E6AIRhOgCFYToAhmE6AIdhOgCAAToAgQE6AIIBOgCDAToAnF04AJ3lPwCe7T8An+U/AJhdOACZRTgAmk04AJtFOACUuTgAlWU4AJZtOACXZTgAkAE6AJEBOgCSAToAkwE6AMMIAIDbCACA4QgAgPMIAIB5BwCAJQkAgHwHAICEBwCAVwkAgKAHAIDOBwCAwAcAgMQGAIDcBACAewUAgM4FAIAJBgCALwYAgEYGAIBjBgCAegYAgJEGAIDXAwCA8AMAgAkEAIAiBACANQQAgEgEAIBbBACAbgQAgIEEAICUBACA+gQAgAkFAIAYBQCAJwUAgDYFAIBFBQCATgUAgFcFAIBgBQCAaQUAgJUFAICeBQCAYAgAgFwOAIBfDgCASzoAgLgJAAC5CQAArwoAgBgLAIBHOgCATzoAgOYMAIBTOgCAHw0AgIc3AID+MACArzcAgGcyAIDLKgCAxiwAgPktAICaMDUAKi0AgPUtAIDkLwCA3zMAgJ80AwBvNQCAnSQpAJzxAQCd8QEAnvEBAJ/xAQCnNgCA4zYAgBc3AIArOACAgzEAgA8yAIC7MgCAUzMAgG82AIBXOACAgzkAgO8qAICaLACAlzEAgN8yAICjNgCA0zkAgKEuAICHMgCAkzYAgCc3AIAYMACAyzUAgO82AIAULwCAEjEAgCcyAIArMwCANzgAgDYrAIDOKwCAOiwAgIAwAICPMQCA2zIAgP8zAICbNgCAszYAgNc3AID/OACAszkAgM85AIA7NACArYwCAHs0AIAzNQCAUzYAgIs4AIBbNwCAqRUBAK4tAIAwLwCAsAAdALEEHgCyABgAswwZALToGgC1AOQAthjlALcc5wC4AOAAuSThALog4gC7AOwAvDTtAL007gC+OO8AvwDoAOs0AICrNQCAvwgAgA8zAICkUAMApQAMAKZYDgCnAAgAqGAKAKmQCwCqaBUAq2wXAKwAEACtdBEArnwSAK8AHABDNACApzcAgPc4AICqLQCAfS4AgIcxAIA7MgCAbzIAgCM1AIBLNQCAtzgAgDYsAIC3NQCA2isAgNYrAICnNACANzUAgGs2AIC/OACAdzcAgBwwAIBnNwCA1yoAgFEuAICILwCAPzMAgL8zAIBaLACASzQAgEYrAIBsLwCAtyoAgIDlAwCB7QMAgi0vAIPhAwCE4QMAheEDAIbhAwCH4S8AiN0vAInZAwCKbS8AiykCAIw5AgCNOQIAjikCAI8lAgCQcQIAkXECAJJxAgCTcQIAlBECAJURAgCWEQIAlxECAJgxAgCZMQIAmjECAJsxAgCcEQIAnRECAJ4RAgCfEQIAoPECAKHxAgCi8QIAo/ECAKQRAgClEQIAphECAKcRAgCoMQIAqTECAKoxAgCrMQIArBECAK0RAgCuEQIArxECALBxAgCxcQIAsnECALNxAgC0TSIAtRUCALYdAgC3FQIAuC0CALk1AgC6PQIAuzUCALwRAgC9EQIAvhECAL8RAgCVuQ4AlLEOAJfJDgCWsQ4AkbkOAJCxDgCTqQ4AkrEOAJ31DgCcZQ0An/UOAJ71DgCZ+Q4AmPEOAJvpDgCa8Q4AhQUOAIQFDgCHyQ4AhgEOAIGhDQCAlSAAg6UNAIKlDQCN+Q4AjPEOAI/JDgCO8Q4AifkOAIjxDgCL6Q4AivEOALWxAQC0qQEAt7EBALa5AQCxvSAAsLUBALOxAQCyuQEAvfEBALzpAQC/8QEAvvkBALnxAQC4iQEAu/EBALr5AQClNQ4ApDUOAKc1DgCmNQ4AoTUOAKA1DgCjNQ4AojUOAK31AQCs9QEAr/UBAK71AQCp9QEAqPUBAKv1AQCq9QEA+zEAgJgwAIAfNQCAriwAgJotAIALNACAczYAgEs3AIDHMQCA8zEAgCwwAIArNgCATDAAgLszAIALKwCAjisAgNIrAIBjMQCACzUAgAM2AIBXNwCAazgAgEIsAID2LACAJC8AgLQwAICLMgCATzQAgKc4AICLOQCA3zkAgPc5AID2MACAszEAgPs3AIDwLgCAzC8AgOgvAIB4MACAezIAgMcyAIB3MwCAmzQAgD81AICjNQCA6zcAgHs2AIATOACAjzgAgPYrAIAiLACACi0AgLcyAIADNwCAEC8AgIAvAIBEMACAvzEAgOc0AIAzMwCAGysAgGYtAIC1LgCAjC8AgIBdAwCB8SgAgmkDAINpAwCEeQMAhXkDAIZpAwCHaQMAiFkDAIlZAwCK1SoAi60DAIy1AwCNvQMAjrUDAI+tAwCQ1QMAkd0DAJLVAwCT7QMAlPUDAJX9AwCW9QMAl+0DAJjVAwCZ3QMAmtUDAJutAwCctQMAnb0DAJ61AwCfrQMAoFUDAKGZAwCiUQMAo1EDAKRxAwClcQMApnEDAKdxAwCoUQMAqVEDAKp1DACrVQMArE0DAK21AQCuvQEAr7UBALDNAQCx1QEAst0BALPVAQC0zQEAtfUBALb9AQC39QEAuM0BALnVAQC63QEAu9UBALzNAQC9tQEAvr0BAL+9DwBPMwCAazMAgHs1AICbNQCAczgAgPM4AIADOQCAPzkAgDorAICPNACAXzgAgNs4AICkLwCA9yoAgF4rAIBVLgCAdS4AgKQwAIDTMgCA2zMAgIc2AIAnOACA5jAAgLM4AIAaLACAMjEAgD4xAIAfMgCAVzIAgFszAIC3MwCANzQAgBs1AIBLOQCA+C8AgMM4AIBOKwCAmS4AgD8yAIDvNwCAXC8AgKwvAIBGMQCAyzgAgP4rAIDmLACAhS4AgM8wAIAiMQCAbzEAgAMyAIBXMwCAyzMAgGc1AIAHNwCAEzcAgOc4AIBqLACAWzIAgOosAIDXMgCAezMAgJc2AIDPOACAl50HAJadBwCVnQcAlJ0HAJOdBwCSnQcAke06AJCZBwCfmQcAnpEHAJ2ZBwCcgQcAm9UKAJqdBwCZnQcAmJ0HAIchBwCGGQcAhREHAIS1JACDBQcAggUHAIEVBwCAFQcAj+EHAI4ZBwCNEQcAjBkHAIsBBwCKGQcAiREHAIgZBwC3/SkAtpUBALWFAQC0hQEAs5UBALKVAQCxZQcAsGUHAL+RAQC+iQEAvYEBALyJAQC7kQEAuqkBALmhAQC4qQEApxkHAKYRBwClGQcApAEHAKMZBwCiEQcAoRkHAKBhBwCvFQcArhUHAK0FBwCsBQcAqxUHAKoVBwCpuSQAqCEHALs5AIDjOQCAOjEAgDcyAIDTNQCA0zQAgPc0AIAnMwCArzIAgHM3AIATKwCAOzYAgAIsAIDyKwCAAC8AgCAwAIADNQCAQS4AgBMyAIDyMACA9zcAgLs4AIAcLwCAbisAgEItAICWLQCA4jAAgN4rAIAvMwCA8zMAgFc0AIBzNACAdzQAgIs0AIALOQCA+zQAgJ82AIBjNwCAFzgAgEM4AIBfOQCAYzkAgGc5AIDLOQCAOzgAgNc4AIA+KwCAYisAgHYrAIAyLACAPiwAgH4sAIAyLQCATi0AgFYtAICSLQCAni0AgIEuAICYLwCAwC8AgMgvAICR5BgA4C8AgIwwAICANQMAgT0DAII1AwCDTQMAhFUDAIVdAwCGVQMAh00DAIjJKwCJcQMAitErAIt1AwCMbQMAjVUDALwwAIDqMACAkCUDAGcxAICSIQMAKzIAgEcyAICVOQMAlikDAJcpAwCYGQMAmRkDAJrpAwCb6QMAnPkDAJ35AwCe4SsAdzIAgKARAwDLMgCAoh0DAOsyAIBfMwCApQ0DAKYFAwA/NACAYzQAgF80AICqCQMAqwkDAKwZAwCtGQMArgkDAK8JAwCweQMAsXkDALIJAwCzCQMAtBkDALUZAwC2CQMAtwkDALg5AwC5OQMAugkDALsJAwC85S0AvR0DAL4VAwC/DQMAvY0fALxhAgC/nR8Avp0fALmRHwC4xQIAu5EfALqZHwC1VR8AtFUfALdVHwC2VR8AsVUfALBVHwCzVR8AslUfAK0dHwCsHR8AZzQAgGs0AICpHR8AqB0fAKsNHwCqDR8ApSEfAKRZHwCn8QIApikfAKFBHwCgWR8Ao0EfAKJJHwCdpR8AnKUfAJ+hHwCeqR8AmYUfAJiFHwCbhR8AmoUfAJWpHwCUoR8AlwUGAJahHwCRqTgAkPkAAJO5HwCSARwAjWEBAIzNOACPgQAAjmkBAIldAQCIUQEAi0UBAIpNAQCFrQEAhKEBAIeVAQCGvQEAgQ0CAIANAgCDxQEAgsUBAIc0AICrNACAvzQAgNs0AIBHNQCATzUAgGM1AICLNQCA2zUAgA82AIB3NgCAHzcAgDc3AIBrNwCAbzcAgLM3AIC3NwCADzgAgOs4AIAvOQCARzkAgJAvAICm6gCA8zUAgL8qAIDKKwCAiisAgDIrAIByKwCAnisAgC4sAIBKLACAHi0AgC4tAIBKLQCApi0AgPEtAID9LQCAGS4AgCkuAIAYLwCAIC8AgFAvAIBwLwCAoC8AgLgvAICoLwCAvC8AgPwvAIBUMACAYDAAgGgwAICQMACAFjEAgCoxAIBrMgCAYzIAgJMyAIAjNACA7zIAgCMzAIBvMwCAizMAgK8zAICAmQEAgZkBAIKpAQCDqQEAhLkBAIW5AQCGqQEAh6kBAIiZAQCJ1RwAipUBAIvVHACM8QEAjfEBAI7xAQCP8QEAkJEBAJEtHACS3RQAk5kBAJSJAQCVhTIAlnkYAJcxGgCYvQEAmYUBAJoVHwCbiQEAnPUfAJ2dAQCelQEAn40BAKDxHAChcQEAonEBAKNxAQCkkQMApZEDAKbtHACnlQMAqK0DAKm1AwCqvQMAq7UDAKytAwCtuQEArpkDAK+ZAwCwbRgAse0DALLVAQCz4QMAtOEDALXhAwC24QMAt+EDALjRAQC5pQMAun0cALupAwC8xQEAvSEXAL6xAwC/xQEA0zMAgNczAID3MwCABzQAgBs0AIAXNACARzQAgMM0AIDzNACAKzUAgFs1AIA/NgCAZzYAgNs2AIAjNwCALzcAgE83AIBTNwCAXzcAgHs3AIDzNwCAIzgAgFs4AIB7OACAxzgAgB85AIA7OQCAmzkAgD3qAIA46gCAauoAgOcpAIAPKgCAEyoAgOzqAIAZ6wCAkesAgCc6AIA3OgCASggAgFUIAIBYCACATQgAgFEIAIBaCQCA9w4AgOgOAIDtDgCA/A4AgPIOAIBRDwCA0A8AgIcPAIA1DwCAYA8AgG0PAIB1DwCAow8AgMgPAIC+DwCAww8AgLAPAIC3DwCABA8AgIBNAQCBRQMAglkBAINZAQCESQEAhUkBAIZ5AQCHVQMAiKkeAIlBAQCKZQMAi0UBAIxhAwCNWQEAjsU7AI9NAQCQNQEAkT0BAJI1AQCTzQEAlNUBAJXdAQCW1QEAl80BAJj1AQCZ/QEACQ8AgA4PAIAbDwCAKA8AgDAPAIA4DwCAQg8AgEcPAIBMDwCAVg8AgFsPAIBjDwCAcA8AgHgPAIB9DwCAgg8AgIoPAICPDwCAmQ8AgJ4PAICmDwCAgzQAgKsPAIDLDwCAPQ8AgCAPAIBoDwCAlA8AgBMPAIDjFgCA7BYAgO8WAID1FgCA6RYAgPIWAIDmFgCAGRcAgBwXAID7FgCAnc0GAPgWAICfwQYA/hYAgAEXAIAKFwCABxcAgJSZBgCVmQYAlukGAJfpBgANFwCABBcAgBMXAICTiQYAEBcAgB8XAIAlFwCAKxcAgCgXAIAuFwCAMRcAgDoXAICEzQYAhdUGAIbZBgA0FwCAgO0GAIHVBgCC3QYAg9UGALwZBwBdFwCAvhUHACIXAIC4GQcAuRkHALoJBwC7CQcAtN0HALUlBwC2LQcARhcAgLDdBwCxxQcAss0HALPFBwCsNQYArT0GAK41BgCvpQcAqDkGAKl9QgCqNQYAqy0GAKQ5BgClOQYApgkGAKcJBgCgIQYAoQFCAKI5QwCjKQYAgKEGAIGhBgBDFwCAg6UGAIS9BgBOFwCAhqkGAIepBgCImQYAieUGAIrtBgCL5QYAjP0GAI3lBgCO7QYAj+UGAJCdBgCRmQYAkqkGAJOtBgCUsQYAlbUGAJa9BgCXuQYAmIUGAJmBBgCagQYAm4UGAJyZBgCdnQYAnpUGAJ+RBgCgbQYAoWkGAKJ5BgCjfQYApGEGAKVlBgCmbQYAp2kGAKhVBgCpUQYAqlEGAKtVBgCsSQYArU0GAK5FBgCvQQYAsD0GALE5BgCyyQEAs80BALTRAQC11QEAttEBALfVAQC46QEAue0BALr5AQC7/QEAvOEBAL3lAQC+7QEAv+kBAIEVAgCAEQIAgxECAIIVAgCFDQIAhAkCAIcpAgCGLQIAiRUCAIgRAgCLEQIAihUCAI1xAgCMdQIAj30CAI55AgCRBQIAkAECAJMBAgCSBQIAlRkCAJQdAgCXFQIAlhECAJktAgCYKQIAmzkCAJo9AgCdIQIAnCUCAJ8tAgCeKQIAodkCAKDdAgCj0QIAotUCAKVpUwKkbVMCp8UCAKbBAgCp/QIAqPkCAKvFAgCqwQIArd0CAKzZAgCvPQIArjkCALEpUwKwLVMCVBcAgEAXAIBRFwCAWhcAgH8WAIDnDwCANxAAgBQQAIAoEACAIxAAgC0QAIAyEACAGRAAgFcXAICEnA4A7A8AgPEPAIAFEACAHhAAgF4QAIBjEACAbxAAgIUQAICUEACAmRAAgKQQAIC+EACA0RAAgNmIUAL1EACAJxEAgCwRAIA0EQCAQxEAgFIRAIBXEQCAXxEAgIIRAICpEQCAtREAgNURAIDaEQCA3xEAgBkSAIAsEgCAOBIAgFASAIDKEgCAIBMAgDkTAIA+EwCAURMAgGITAIB0EwCAeRMAgKATAICoEwCAvRMAgOQTAIDpEwCAQxQAgEgUAIBNFACAWRQAgGUUAIBqFACAchQAgH4UAICYFACAnRQAgNlcUAKlFACAqhQAgK8UAIC0FACAuRQAgL4UAIDRFACAn8kOAJ7NDgCdwV0CnBkNAJsFDQCaHQ0AmRENAJixDACXjQwAlqkMAJWlDACUoQwAk70MANYUAIDyFACADBUAgCYVAIAyFQCAShUAgE8VAIBcFQCAfRUAgKAVAIC6FQCAxhUAgMsVAIDTFQCA9BUAgA4WAIAdFgCAOhYAgD8WAIC/fQ4AvnkOAL11DgC8cQ4Au2kOALptDgC5YQ4AuGkOALdVDgC2UQ4AtVkOALRdDgCzXQ4AslkOALFRDgCwVQ4AryUOAK4hDgCtKQ4ArC0OAKsNDgCqCQ4AqQEOAKgFDgCnNQ4ApjEOAKU9DgCkOQ4AoyEOAKIlDgChNQ4AoDEOAIAFDgCBDQ4AggUOAIP1DwCEAQ4AhQEOAIYBDgCHAQ4AiAEOAIkBDgCKAQ4AiwEOAIwBDgCNAQ4AjgUOAI99DgCQBQ4AkQ0OAJIFDgCTHQ4AlAUOAJUNDgCWBQ4Alz0OAJgFDgCZDQ4AmgUOAJsdDgCcBQ4AnQ0OAJ4FDgCf/Q4AoAUOAKENDgCiBQ4Aox0OAKQFDgClDQ4ApgUOAKc9DgCoBQ4AqQ0OAKoFDgCrHQ4ArAUOAK0NDgCuBQ4Ar30OALAFDgCxDQ4AsgUOALMdDgC0BQ4AtQ0OALYFDgC3OQ4AuAkOALkJDgC6GQ4AuxkOALwJDgC9CQ4Avs0BAL/FAQCAPQIAgUUCAIJNAgCDRQIAhF0CAIVFAgCGTQIAh0UCAIh9AgCJRQIAik0CAItFAgCMXQIAjUUCAI5NAgCPRQIAkD0CAJFBAQCSQQEAk0EBAJRBAQCVQQEAlkEBAJdBAQCYQQEAmUEBAJpBAQCbQQEAnEEBAJ1BAQCeQQEAn0EBAKDBAQChwQEAosEBAKPBAQCkwQEApcEBAKaVDQCnxQEAqFkMAKm1DQCq9QEAq80BAKyRDQCt0QEArp0NAK+VDQCwqQEAsakBALL1DQCzvQEAtJENALWRDQC2rQEAt6UBALitDQC5mQEAurkNALu5DQC8OQ0AvTkNAL4hDQC/IQ0Ap1X8AEcWAIBMFgCAXxYAgGQWAICKFgCAlhYAgKIWAICxFgCAzhYAgNMWAID0EQCABRIAgIIWAICBAACAiwAAgJUAAICfAACAqQAAgLMAAID7DwCAABAAgAoQAIB7EACAgBAAgIoQAIDrEACA8BAAgB0RAIA5EQCAPhEAgEgRAIBXFQCAExYAgBgWAIAwFgCApxYAgKwWAIDEFgCA9g8AgA8QAICPEACAIhEAgN0SAIBFFQCANRYAgGkWAIDJFgCATREAgGoSAIClEgCAuBIAgBcUAIAjFACALxQAgJMTAICYEwCA1xMAgNwTAIADFACACBQAgG8SAIB0EgCAvRIAgIL5CwCD+QsAgO0LAIH5CwCGWQQAh1kEAIQtBACFWQQAiqUHAIutBwCIqQcAiXEEAI5JBACPSQQAjE0EAI2xBwCS1QcAk2UHAJB9BwCR3QcAlnkHAJdRCwCUwQcAlXkHAJptCwCbxQcAmGELAJnxBwCebQsAn1ULAJxtCwCdZQsAorELAKOxCwCgLQcAoaELAKbdCwCnzQsApKULAKU1BwCqxQsAq80LAKj1CwCpzQsArskLAK/JCwCsyQsArckLALJtBwCzTQsAsEkLALFJCwC2RQsAt00LALRVCwC1TQsAukkLALtJCwC4dQsAuUkLAL5JCwC/SQsAvEkLAL1JCwCAwQoAgcEKAILZCgCD2QoAhPkKAIX5CgCG6QoAh+kKAIjZCgCJHQUAihUFAIttBQCMdQUAjYUGAI5pBQCPaQUAkBkFAJEZBQCSIQUAkyEFAJQhBQCVIQUAlu0GAJdZBgCYaQYAmd0GAJp9BgCbdQYAnG0GAJ1VBgCexQYAn3EKAKAhBgChpQoAoi0GAKOxCgCkOQYApdkKAKZZBgCnHQoAqGUGAKltBgCqZQYAq1kKAKxJCgCt8QUArs0FAK8JBgCw4QYAsXkGALIZBgCzGQYAtAkGALUJBgC2OQYAtzkGALgJBgC5CQYAuhkGALsZBgC8CQYAvQkGAL75AwC/+QMAwhIAgMgRAIC8TQEAvUkBALqxCQC7sQkAuDUBALktAQC2XQkAtw0BALRdCQC1VQkAsv0FALOVCQCw8QUAsfkFAK5tAQCvdQEArAkBAK1lAQCqaQEAq2kBAKiRBQCpaQEApk0BAKdVAQCkTQEApUUBAKJtAQCjVQEAoG0BAKFlAQCejQEAn5UBAJyNAQCdhQEAmpEAAJuRAACYYQUAmWEFAJZRBQCXUQUAlEEFAJVBBQCSUQUAk1EFAJD5AQCRYQUAjvkBAI/5AQCMAQUAjfkBAIr9AQCL5QEAiP0BAIn1AQCG/QEAh8UBAIT9AQCF9QEAgv0BAIPlAQCA/QEAgfUBAPBQ/wDNEQCAnBEAgKERAIDkEQCA6REAgCwTAIAxEwCAZxMAgGwTAIB8EgCAiBIAgJsSAICgEgCASxIAgOISAIBdEwCAURAAgKkQAIDDEACAyhAAgNYQAID6EACAAREAgAgRAICHEQCAwREAgLoRAIAxEgCAHhIAgCUSAIBcEgCAVRIAgGMSAIDPEgCAJRMAgI0SAICBEgCAqhIAgLESAIBDEwCAVhMAgH4TAICFEwCAjBMAgK0TAIDCEwCAyRMAgO4TAID8EwCA9RMAgFIUAICDFACAihQAgBEVAIAfFQCAGBUAgPcUAIArFQCANxUAgIIVAICJFQCAmRUAgGEVAIBvFQCApRUAgKwVAIBoFQCAURYAgFgWAID5FQCAABYAgN8VAIDmFQCAKRYAgCIWAIC2FgCAdBAAgLcQAICwEACAlRn/AJQR/wCXKf8AlhH/AJEd/wCQHf8Akwn/AJIR/wCdFf8AnBX/AJ8V/wCeFf8AmRX/AJgR/wCbFf8AmhX/AKUJ/wCkDf8Apxn/AKYB/wChEf8AoOn/AKMd/wCiGf8ArT3/AKw5/wCvDf8Arg3/AKkl/wCoJf8AqyH/AKol/wC1df8AtHX/ALdx/wC2df8AsXn/ALBx/wCzdf8AsnX/AL0t/wC8Kf8Avz3/AL49/wC5Mf8AuEn/ALsx/wC6Of8AgNn+AIHZ/gCC6f4Ag+n+AIT1/gCF/f4AhvH+AIfx/gCIzf4AidX+AIrd/gCL1f4AjM3+AI01AQCOPQEAjzUBAOQQAIDdEACAkkUBAJNdAQCURQEAlU0BAJZFAQCXfQEAmEEBAJlBAQCaQQEAm0EBAJxBAQCdRQEAnk0BAJ9FAQCgvQEAocUAAKLNAACjjQMApJUDAKWdAwCmlQMAp40DAKi5AwCpuQMAqokDAKuJAwCsmQMArZkDAK6JAwCviQMAsPkDALH5AwCyiQMAs4kDALQB/gC1Df4AtpEDALeRAwC4sQMAubEDALqxAwC7sQMAvKkDAL2lAwC+mQMAv5kDABYRAIAPEQCAlREAgGQRAICOEQCAkQQHAD0SAIDWEgCAlBIAgEoTAIAFFQCAPhUAgJsWAICPFgCAvRYAgL8VAICRFACABxYAgNATAIDKFACA2BUAgLMVAID+FACAwxQAgGsRAICuEQCAdhUAgF4UAIBoEACARBIAgO0VAIAZEwCAdxQAgEgQAIA/EACAkBUAgOcSAID8EQCAtBMAgHEWAIDwEgCA9xIAgHIRAIAKEgCApgMAgBMjAIAXIwCAoAYAgMcAAIC1BgCAsSMAgLUjAIC/IQCAuyEAgOYHAIB+CQCAggkAgEcjAICtIwCAOyMAgD8jAIAjIwCAJyMAgCsjAICAaQEAgWkBAIJ5AQCDUQEAhKX8AIWZ/ACGbQEAh2UBAC8jAIAzIwCANyMAgN4HAIDiBwCA0QAAgNcAAICiAwCAqAMAgOAHAIDTAACA1QAAgL0GAIB5AACADRQAgH0AAICHAACAkQAAgBIUAICbAACAHhQAgKUAAIAqFACArwAAgDYUAIC5AACAOxQAgNUPAIBbEACAnhAAgKEQAIAxEQCAXBEAgKYRAIDSEQCA7hEAgPERAID5EQCAExIAgBYSAICwvQEAsUUBALJNAQCzRQEAtF0BALVFAQC2TQEAt0UBALh9AQC5RQEAuk0BALtFAQC8XQEAeRIAgMcSAIA2EwCAqAYAgLMGAIDMlLQAPBAAgHETAICdEwCApRMAgOETAIBAFACAbxQAgKIUAIDbFACAVBUAgNAVAIBEFgCAbhYAgJiNBgCZgbEAhxYAgN4UAIDjFACA6BQAgO0UAIDPAACAkNEGAJHRBgCS0QYAk9EGAJSpkgKVtQYAlr0GAJe1BgDZAACAswMAgOQHAICACQCAASMAgAUjAICHKQCAOyQAgHQkAICTJACApSQAgMokAIDJKACA5SgAgPcoAICOJgCAuCEGALkhBgC6IQYAuyEGALwhBgC9IQYAviEGAL95uACwIbEAsTUGALI9BgCzNQYAtCkGALVFsgC2TbIAtyEGAIC5uQCB+QcAgikGAIMpBgCEOQYAiykAgG8pAICHMQYAiBEGAIn5sACK9bAAi/GwAIztsACN7QcAjuEHAI/lBwCQ8QcAkfEHAJL1sAAvJACAlImTApXpBwCWnQcAl50HAJixBwCZ1bMAmt2zAJuxBwCckQcAnZEHAJ6RBwCfSQYAoLkGAKG5BgCiIbMAo80GAKSRAQClkQEApkGyADMkAICo5QYAqe0GAKrlBgCr/QYAhAkAgIcJAICQCQCAjQkAgLCVBgCxnQYAspUGAIoJAIC0sQYA8iEAgLa9BgC3tQYAuI0GALmVBgC6nQYAu5UGALyNBgC9dQYAvn0GAL91BgCCkaMCg5GjAoCFBQCBnaMChrmjAoeNowKEjaMChbWjAoqVowKLkaMCiLGjAomZowKOPQIAj6UFAIyNowKNMQIAktEFAJPRBQCQ2QUAkd0FAJbJBQCXzQUAlM0FAJXFBQCa/QUAm/kFAJjxBQCZ8QUAntEFAJ/VBQCc5QUAnd0FAKIlBQCjIQUAoCkFAKEpBQCmOQUApz0FAKQ9BQClNQUAqg0FAKsVBQCoAQUAqQEFAK4FBQCvDQUArAkFAK0JBQCyfQUAs3kFALBxBQCxcQUAtiUFALchBQC0ZQUAtSkFALoZBQC7HQUAuB0FALkVBQC+DQUAvwkFALwBBQC9AQUAhPwCAPUiAID6IQCAAiIAgPkiAID9IgCAkZQCAKQQDQCffBwAnmQIAJ08CADZzKACzMCFAsz8hQL2IQCA9egAgJmY1gD66ACAm/jWAP/oAICxNN0ABOkAgNmUoAKynPsApwUAAAnpAIDZhKAChywDAJE4AgAO6QCAjAwCAI20AgDwtAIAgqACABPpAIC++QAAkpACAJcdAAC5+AMAhAgCAPBAAgCRfAIAHekAgBjpAICljNYAIukAgI3IAgDwJAIAsGkAAJK4AgC4tAMAudgMAJa0AgCW/AIArz0AAJFUAgCvCCEAJ+kAgLpFAACRTAIALOkAgL1BAACWuAIArw0AAIsZDACKHQwAiREMAIgZDACPDQwAjgkMAI0FDACMBQwAgzEMAII1DACBOQwAgNELAIclDACGIQwAhS0MAIQtDACbaQwAmm0MAJlhDACYaQwAn30MAJ55DACddQwAnHUMAJNBDACSRQwAkUkMAJBxDACXVQwAllEMAJVdDACUXQwAq7kMAKq9DACpsQwAqLkMAK9J/gCuqQwAraUMAKylDACjkQwAopUMAKGZDACggQwAp4UMAKaBDACljQwApI0MALuZDAC6kQwAuZ0MALidDAC/2YgC8HwNALUcAgC8hQwAs6kMALKtDACxoQwAsDX+ALehDAC2uQwAtbUMALS1DACBwQsA8DgBAIM9CgCCUQ0AhV0KAIRdCgCHmQ0AhiUKAImVvwCIlb8Ai4G/AIoRCgCNcQoAjIULAI+9DQCOgbwAkbWdApCxvACTqZ0CkpGdApWtvACUrbwAl9W/AJbVvwCZxb8AmMW/AJuxnwKa0QsAnam+AJx1DQCfvQsAnnkNAKEBvwCg1QoAo5WxAKKFvQClgb0ApJm9AKeBvQCmqbEAqYm9AKiFvQCrmb0Aqom9AK3xvQCsjb0Ar+m9AK71vQCxIb8AsJ29ALNJyQCyrb0AtaG9ALS9vQC3ob0AtkmxALlpyQC4TbEAu8UKALrNsQC9wQoAvLEKAL8hCwC+dQ0AgP2+AIHVngKCZb4Ag8W+AISRvgCFnb4AhqW+AIeNvgCIrZECieW+AIopkgKLtb4AjBGSAo2VvgCOLbIAj8WeApDpvgCRsbUAkkGSApPxnwKU1b4AleW+AJbhvgCXTZICmGWSApl9kgKaub4Am7EIAJz9DgCdlQgAkRQDAJ/tDgCgFQ4AoT0IAKJ1CACjrQkApCUIAKUNDgCmBdYApwEOAKgBDgCpAQ4AqgEOAKsBDgCsAQ4ArQEOAK4BDgCvAQ4AsNUPALGV0wCyndMAs4XUALSJ0wC1idMAttnWALfZ1gC46dYAuenWALr51gC7+dYAvOnWAL3p1gC+udYAv7nWAIBJ1wCBSdcAglnXAINZ1wCESdcAhXXSAIZ90gCHddIAiE3SAIlV0gCKddcAi63XAIy11wCNvdcAjrXXAI9J0gCQOdIAkTnSAJLJ0gCTydIAlNnSAJXZ0gCWydIAl8nSAJj50gCZ+dIAmsnSAJvF0gCc4dcAnW0OAJ41DgCf4Q4AoNHZAKHB2wCiwdsAo93bAKTF2wClzdsApsXbAKf92wCoJdsAqWXbAKrN2wCrxdsArN3bAK3B2wCuwdsAr3nbALAJ2wCxCdsAshnbALMZ2wC0CdsAtbXYALbB3gC33d8AuOXfALn13wC6/d8Au63fALy13wC9pd8Avq3fAKQgBQCPFd4AjhXeAI0F3gCMBd4AixXeAIoV3gD+IQCAsCUAAIfd3ACG3dwAhd3cAITd3ACD2dwAgtXcADHpAIA26QCAP+kAgEjpAIBV6QCAnAXeAJtt3gCabd4AYukAgG/pAICXVd4AllXeAJVB3gCUSd4Ak1HeAJJp3gB86QCAiekAgKzpAICukQsArZkLAKyBCwCriQAAqp0LAKmhCwComQsAkukAgKZxCQClZQkApBEJAKNlmAKiDQkAoQ0JAKANCQC7vQAAtekAgL3BjAKf6QCAvx0AAL4RAAC9eQAAvF0AAIAFAADC6QCA0AoAgLMMAIBkDQCAag0AgHANAIB8DQCAhWgCAH8NAICH6AMAhiADAJ4tAQCfVQEAgg0AgIUNAICIDQCAlw0AgJ0NAICgDQCAow0AgCYiAIDNDQCA3A0AgJVAHQCURB4Al0gbAJYAGACRLAAAkFQBAJNYHwCSABwAnWASAJxkEwCfiBcAnmwRAJlwGQCYdBoAmwAQAJoAFAACDgCABQ4AgBQOAIAXDgCAIw4AgB4iAIA4DgCAOw4AgLgALAC5rC8AuqguAN0WAIAWFwCA4BYAgLoDAIC3AwCAygMAgO0EAICMBQCA3wUAgBoGAIBABgCAVwYAgHQGAICiuQEAo7kBAKCtAQChpQEAiwYAgDgBAICkgQEAPAEAgICBuACBDboAghW2AIMBugCEAboAhSG2AIYBugCHPboAiAm6AIkBugCKGboAixW6AIxxugCNfboAjmm6AI9lugCQobgAkSW6AJLJzgCTJboAlCG6AJXBtgCWIboAl/W2AJjpzgCZRbYAmrmaApsBugCcuboAnfW6AJ7xugCfwboAoBG6AKEJlQKiSboAo42WAqQJugCltZYCpjm6AKeJtgCoWZoCqQ26AKpdsQCrpZYCrA2bAq0xugCuCboArwW6ALDRlgKxwZYCstGWArMdugC0UbgAtd26ALbFtgC30boAuPG6ALnRtgC68boAu826ALzZugC90boAvsm6AL/FugCfCZcCnvGwAJ2huwCc9ZsCmwW3AJq1uwCZOZcCmIW7AJchlwKW5bsAzMS+AJS9uwCTjbsAkr27AJG5uwCQ9bsAjy27AI6VmwKNabcAjMXPAIv5twCKLbsAic23AIgtuwCHCbsAhuXPAIUJuwCEjbkAgym7AIIluwCBMbsAgD27AL8ptwC+/bsAvR23ALz9uwC7+bsAuhXPALn5uwC4fbkAt/m7ALb1uwC14bsAtO27ALOJuwCyhbsAsZ27ALCVuwCv4bsArt27AK39twCs3bsAq927AKrJtwCp0bsAqF25AKcxuwCm/ZcCpe2XAqT9lwKjSbsAokW7AKF9uwCgQZoCgJmkAIEliAKCqaQAgxmoAEABAICFvaQAhu2vAIcViAKInYUCiaGkAIqZpACLlaQAjCGIAo0xiAKOIYgCj+2kAJDBpgCRTaQAklWoAJNBpACUQaQAlWGoAJZBpACXfaQAmEmkAJlBpACaWaQAm1WkAJwxpACdPaQAnimkAJ8lpACgYaYAoeWkAKIJ0ACj5aQApOGkAKUBqACm4aQApzWoAKgp0ACphagAqnmEAqvBpACseaQArTWkAK4xpACvAaQAsFGkALFJiwKyCaQAs82IArRJpAC19YgCtnmkALfJqAC4GYQCuU2kALodrwC75YgCvE2FAr1xpAC+SaQAv0WkAIARiQKBAYkCghGJAoPdpQCEkacAhR2lAEQBAICHEaUAiDGlAIkRqQCKMaUAkQ0AgEgBAICNEaUAjgmlAI8FpQCQAaUAkQ2lAJIZpQCTFaUAlLGnAEwBAICW2dEAlzWlAJgRpQCZ8akAmhGlAJvFqQCc+dEAUAEAgJ6phQKfEaUAoEmlAKEFpQCiAaUAozGlAKQBpQClGYoCplmlAKediQKoOaUAqYWJAqoJpQCruakArEmFAq0dpQCuTa4Ar7WJArB9hAKxQaUAsnmlALN1pQC0wYkCtdGJArbBiQK3DaUAuGGnALntpQBUAQCAu+GlALzhpQC9wakAvuGlAFgBAICAKaYAgSGmAII5pgCDNaYAhFGmAFwBAICGSaYAzEyHAmABAIBkAQCAiqnSAItFpgCMQaYAjaGqAI5BpgCPlaoAkMnSAGgBAICSmYYCkyGmAJSZpgCV1aYAltGmAJfhpgCY8aYAmemJApqppgCbbYoCnOmmAJ1VigKe2aYAn2mqAKB5hgKhLaYAon2tAKOFigKkLYcCpRGmAKYppgCnJaYAqLGKAqmhigKqsYoCq32mAKwxpACtvaYArqWqAK+xpgCw0aYAsfGqALLRpgCz7aYAtPmmALXxpgC26aYAt+WmALihpgC5raYAurmmALu1pgC8EaQAvZWmAL550gC/laYAl7mnAJa1pwCVjacAlLGGApMZiwKS4awAkbGnAJDlhwKfLacAnjmrAGwBAICcraUAm+GnAJotiwKZPYsCmC2LAof9pwCGzacAhcmnAISFpwCDPacAgoWHAoF5qwCA1dMAj3WrAI7FpwCNSYsCjPWnAItxiwKKtacAifWIAojtpwC37acAtlWHArWpqwC0BdMAszmrALLtpwCxDasAsO2nAL+hiwK+ZacAvSWIAnABAIC7DacAdAEAgLk5pwC4dacAeAEAgKb1pwCl7acAfAEAgIABAICizacAhAEAgIgBAICviacArmXTAIwBAICsDaUAq6mnAKqlpwCpsacAkAEAgICRoACBiY8CgsmgAIMNjAKEiaAAhTWMAoa5oACHCawAiNmAAomNoACK3asAiyWMAoyNgQKNsaAAjomgAI+FoACQUYwCkUGMApJRjAKTnaAAlNGiAJVdoACWRawAl1GgAJhxoACZUawAmnGgAJtNoACcWaAAnVGgAJ5JoACfRaAAoMGgAKHNoACi2aAAo9WgAKRxogCl9aAAphnUAKf1oACo0aAAqTGsAKrRoACrBawArDnUAK2VrACuaYACr9GgALAJoACxRaAAskGgALNxoAC0QaAAtVmPArYZoAC33YwCuHmgALnFjAK6SaAAu/msALwJgAK9XaAAvg2rAL/1jAKAvYACgYGhAIK5oQCDtaEAhAGNAoURjQKGAY0Ch82hAIihowCJLaEAijWtAIshoQCMIaEAjQGtAI4hoQCPHaEAkGmhAJFhoQCSeaEAk3WhAJQRoQCVHaEAlgmhAJcFoQCYgaMAmQWhAJrp1QCbBaEAnAGhAJ3hrQCeAaEAn9WtAKAJ1QChpa0AolmBAqPhoQCkWaEApRWhAKYRoQCnIaEAqDGhAKkpjgKqaaEAq62NAqwpoQCtlY0CrhmhAK+prQCwOYECsW2hALI9qgCzxY0CtG2AArVRoQC2aaEAt2WhALjxjQK54Y0CuvGNArs9oQC8caMAvf2hAL7lrQC/8aEAgBGiAIExrgCCEaIAgy2iAIQ5ogCFMaIAhimiAIclogCIYaIAiW2iAIp5ogCLdaIAjNGgAI1VogCOudYAj1WiAJAxogCR0a4AkjGiAJPlrgCU2dYAlXWuAJaJggKXMaIAmKmiAJnlogCa4aIAm9GiAJzhogCd+Y0CnrmiAJ99jgKgGaIAoaWOAqIpogCjma4ApGmCAqU9ogCmbakAp5WOAqgdgwKpIaIAqhmiAKsVogCsoY4CrbGOAq6hjgKvbaIAsEGgALHNogCy1a4As8GiALTBogC14a4AtsGiALf9ogC4yaIAucGiALrZogC71aIAvLGiAL29ogC+qaIAv6WiAJMVrwCSpaMAkSmPApCVowCXGY8CluGoAJWxowCU5YMCm5mjAJqVowCZraMAmJGCAp/howCeLY8CnT2PApwtjwKD6a8Agj2jAIHdrwCAPaMAhz2jAIaFgwKFea8AhNXXAIvdowCK7aMAiemjAIilowCPcY8CjrWjAI31jAKM7aMAs+mjALIF1wCx6aMAsG2hALc5rwC27aMAtQ2vALTtowC7zaMAunWDArmJrwC4JdcAvw2jAL49owC9OaMAvHWjAKPNowCi2a8AocGjAKBNoQCn8aMAps2jAKXtrwCkzaMAq9mjAKrVowCpzaMAqMWjAK+powCupaMArbGjAKy9owC43YwCueWMArrxrQC78a0AvJmuAL2ZrgC+ua4Av7muALDZrQCx2a0AsqGuALOhrgC0ka4AtZGuALbNrQC3yYwCqMmuAKnJrgCq6a4Aq+muAKylrQCtoYwCroWMAq+9jAKgwa4AocGuAKKdrQCjmYwCpK2MAqWVjAKmga0Ap4GtAJh1rQCZcYwCmlWMApttjAKcaa0AnWmtAJ4RrgCfEa4AkH2MApFFjAKSUa0Ak1GtAJQ5rgCVOa4AlhmuAJcZrgCIwbwCiTm5AopRFQCLURUAYQ0AgJQBAICOLa0AjymMAoDdqwCBdRUAgqkBAIN5FQCE+bwChQW5Aob5ugKHCbkCbQ0AgJgBAIBzDQCAnAEAgHkNAIBLIwCAiw0AgNEGAIC+DQCAywcAgNANAIAPBwCACA4AgJcHAIAaDgCAnQcAgCYOAICGYAIAhYABAIRMOQCABwCAyAcAgE8HAIBSBwCAXQcAgJANAADhBgCAFSQAgOglAIA1LgCAiXg/AIgAPAC6LACA1i0AgD83AIAHKwCA0zAAgL8yAIAOLACAYC8AgKYrAICsMACA+isAgCc1AICbNwCAui0AgPIsAIBzMgCAEDAAgDwwAIAbOACAMDAAgAgwAIB/NACAuzQAgN4sAICvoAEAUzIAgKczAIASLACAPi0AgFM4AICPIwCAUyMAgMwouALNRLgCzIi4As38uALMkIQCzTSFAsywhQLNVIcCzBCCAs1QggLMoIICzYyCAswwgQLNJIECzBiBAs2EgQJdIwCAcSMAgJkjAIB7IwCAoyMAgGcjAICFIwCAZC8AgKo5AwCrOQMAziwAgNsqAIDMWLkCzaS6AqwZAwDTKgCAsHWvALFxjgKyVY4Cs22OArRprwC1aa8AthGsALcRrAC4AawAuQGsAOMqAIDP6QCALisAgEIrAIBKKwCAUisAgIJB2gCDra4AgCmsAIGtrgCGqa4Ah32iAISprgCFSaIAis0DAIs5xACIYdoAic2iAI6hAwCPoQMAjM0DAI3BAwCfua8AnrWvAJ2NrwCcsY4CmxmDAprhpACZsa8AmOWPApc1owCWha8AlQmDApS1rwCTMYMCkvWvAJG1gAKQra8Aj/2vAI7NrwCNya8AjIWvAIs9rwCKhY8CiXmjAIjV2wCHyaMAhh2vAIX9owCEHa8AgxmvAIL12wCBGa8AgJ2tAL+xFgC+qRYAvaEWALzBvgK7tRYAuom5ArmBuQK4sQoAt22sALa9AgC1iRYAtLEWALOpFgCysRYAsUUXALCtAgCv2bkCrs0CAK0xFwCszQIAqyUXAKodrACpKRcAqA0DAMyIgQLNSIECpQUXAKQFFwCjIa8Aou2DAqH9gwKg7YMCgAm4AoE9EQCCIQUAg/29AoT1qACFFb4ChiURAIfxiAKIoREAiaERAIoZBQCL0b0CjDG4Ao09vgKOsREAj7ERAJB5BQCRsb0CknWvAJPdEQCUEQUAlcERAJZRuAKXrb0CmGG+ApmRvQKaabgCm5G9ApyhBACdhRAAnrGrAJ+JEACggQUAoX0QAKKBBQCjlb4CpIEFAKVpEACmnREAp4URAKi9EQCphREAqrEFAKthqwCsnQ0Ara2+Aq6lvgKvmREAsI25ArHtEQCy5REAs/0RALT5pAC14REAtvkFALcxvQK4ya8AucmvALrhuAK71REAvNkFAL0FvQK+HagAv/2+AoA9EACB6YkCgokQAIOJEACEIQQAhem8AoYZuQKHFb8CiKkQAImpEACKEQQAi9m8AowNrgCNpRAAjnkEAI+pEACQSbkCkbW8ApJJvwKTubwClFG5ApWpvAKWiQUAl60RAJipqgCZkREAmmkEAJuVEQCceQQAnW2/Ap5pBACfgREAoIUQAKGdEACilRAAo60QAKSJBAClWaoAprUMAKeFvwKovb8CqYEQAKrluAKrhRAArJ0QAK2FEACukaUAr4kQALDhBACxKbwCsuGuALPhrgC02bkCte0QALbxBAC3LbwCuAWpALnlvwK61RAAuwGJArxxEAC9cRAAvskEAL8BvAKAAboCgQ28AoKBEwCDgRMAhCkHAIXhvwKGJa0Ah40TAIhhBwCJsRMAiiG6AovdvwKMMbwCjcG/Ao45ugKPwb8CkJEGAJG1EgCSgakAk7kSAJRRBwCVrRIAllEHAJdFvAKYcQcAmZkSAJptEwCbdRMAnG0TAJ1VEwCeYQcAn7GpAKCtDwChnbwCopW8AqOpEwCk3bsCpb0TAKa1EwCnrRMAqImmAKmREwCqiQcAq0G/AqyZrQCtma0ArrG6Aq+FEwCw6QcAsTW/ArItqgCzzbwCtO0TALU5igK2WRMAt1kTALjRBwC5Gb8Cuum6ArvlvAK8eRMAvXkTAL7BBwC/Cb8CngG9Ap/xvgKcAbsCnf2+AppRBgCbgRIAmCWsAJmNEgCWGQYAl9G+ApShEgCVoRIAkjG7ApM9vQKQCQYAkcG+Ao7BEgCPwRIAjHUSAI2hiwKKtasAi1W9AohxBgCJrb4Chmm7AoddEgCEQawAhUGsAIJRBgCDmb4CgFGnAIFJEgC+qawAv6msALypBgC9Yb4CurmnALuhEgC4tRIAua0SALbtugK3jRIAtLW9ArWJEgCynQ4As629ArChBgCxcagArt0SAK/lEgCszRIArdUSAKrBBgCrKRMAqNEGAKnFvQKm4QYApx0TAKQhqAClGRMAoiEHAKMFEwCg+bsCoQG+AoKhIwBWKwCAWisAgJMpAIDj6QCAh7EjAHorAIB+KwCAmisAgIsJJADU6QCAiWUkAI6NIwCPLSQAlykAgI0JJACSZSQAhisAgN7pAICRtSMAtisAgJf9IwCUrSMAvisAgBcrAICbeSQAxisAgJmRIwC56wCAn8EtAO4rAICdKdQAoiEjAJ8pAIAGLACAoR0jAAosAICnMSMApKEkABYsAICqiSQAoykAgKi5JACp5SQArg0jAK+tJACsiSQArYkkALLlJABOLACAsOkkALE1IwC2TSMAt30jALQtIwC1RSMAuvUkALv5JABSLACAuREjAL5BLQB6LACAvFUtAIYsAICADSUAgZUiAIKpIgCDoSIAhCklAIUpJQCGoSIAh7kiAIgxJQCJbSUAliwAgIsBJQCMASUAjQElAI6FIgCPJSUAkGElAJG9IgCSbSUAk/kiAJSlIgCVzSIAlsUiAJf1IgCY0SIAmZkiAJp9JQCbcSUAniwAgKIsAIDy6QCAozIAgLYsAIChFSIAoikiAKMhIgC+LACApaklAKYhIgCnOSIAqLElAKntJQD36QCAq4ElAKyBJQCtgSUArgUiAK+lJQCw4SUAsT0iALLtJQCzeSIAtCUiALVNIgC2RSIAt3UiALhRIgC5GSIAuv0lALvxJQDKLACA0iwAgNosAIACLQCAgLkvAIG5LwCCyS8Ag8kvAITZLwCF2S8AhskvAIfJLwCI+S8AifkvAIrZLwDuLACA+iwAgP4sAIAGLQCADi0AgJC1LwCRuS8AklkyAJNVMwCUYTMAlQEzAJYtMwCXtTMAmJ0zAJlxMACaSTAAm0kwAJw1MACdXTEAntUxAJ/JMQCgQTEAoUkxAKJZMQCjVTEApE0xAKV9MQCmZTEAp0k6AKilOwCpqTsAqr07AKuxmgCs0ZYArak7AK6dOwASLQCAsEGUALHNlgCy1ZoAs8GWALTBlgC14ZoAtsGWALf9lgC4yZYAucGWALrZlgC71ZYAvLGWAL29lgC+qZYAv6WWAMUAAACpTScAqiEnAKshJwCsIScArSEnAK6lIACvBScAGi0AgKG1IACiiSAAIi0AgKQJJwAmLQCANi0AgKeZIAA6LQCAubkgAEYtAIC7UScAai0AgG4tAIBSLQCAWi0AgLBBJwCxnSAAsk0nALYtAIC0hSAAte0gALblIAC31SAAiJEnAInNJwCKoScAi6EnAIyhJwCNoScAjiUgAI+FJwCArScAgTUgAIIJIACDASAAfi0AgIWJJwCGASAAhxkgAJhxIACZOSAAmt0nAJvRJwCcfS4AnYHXAJ5pLgCfaS4AkMEnAJEdIACSzScAk1kgAJQFIACVbSAAlmUgAJdVIACA+T0Agfk9AIJJPgCDST4AhFk+AIVZPgCGST4Ah0k+AIh5PgCCLQCAhi0AgHotAICOLQCAii0AgKItAID86QCAkB0+AJEtPgCoSBUAvi0AgMItAIDKLQCA3i0AgAEuAICiAAwAo2QPAKBoAwChbAAApgAUAKd0FwCkAAgApXwLAAHqAIAG6gCADS4AgBEuAIAVLgCACS4AgB0uAICnKQCAqykAgCUuAIAtLgCAC+oAgEkuAIBNLgCAWS4AgBDqAIBhLgCAZS4AgEQvAICvKQCAeS4AgJUuAICRLgCAGuoAgJ0uAIAf6gCAqS4AgKUuAICtLgCAvS4AgMEuAICzKQCAgG0/AIF5PwCCoT8Ag6E/AIShPwCFrcgAhq0/AIelPwCInT8AiSEFAIqVyACLJQUAjD0FAI0lBQCO+cgAjyUFAJBdBQCRZQUAkmEFAJN1BQCUFQUAlfU8AJYRBQCXDQUAmD0FAJkFBQCaAQUAmwnVAJwBBQCdgeUAngEFAJ/5BQCgCQUAoRUFAKJlPACjEQUApDUFAKXt1QCmXcgAp1XIAKjd1QCpYQUAqkEFAKtBwwCsRQUArXnIAK5FBQCvlQQAsLEEALGxBACyvQQAs7kEALSpBAC1qQQAtlkEALdNBAC4SQQAuRUEALodBAC7FQQAvA0EAL3ZBwC+yQcAv8kHAIA5BACBOQQAgrUEAIMtBQCEPQUAhSEFAIYtBQCHmdYAiBkFAIllBQCKYQUAi30FAIzlywCNYQUAjmEFAI9hBQCQdcsAkSkFAJL5BQCTaQIAlHkCAJV5AgCWaQIAl2kCAJhZAgCZWQIAmiUCAJs9AgCcJQIAnfE/AJ4hAgCfIQIAoAECAKEBAgCiAQIAowECAKQBAgClAQIApgECAKcBAgCoAQIAqQECAKoBAgCrBQIArB0CAK0FAgCuDQIAr2HAALB5AgCxeQIAsgkCALMJAgC0GQIAtdnlALYVAgC3DQIAuPXlALkxAgC6MQIAuzECALwRAgC9EQIAvhECAL8RAgCfpT4AnqU+AJ2VPgCclT4Am1XMAJqBPgCZiT4AmJE+AJdF9ACWrT4Ala0+AJTh5wCTvekAkrU+AJFNPgCQReQA0S4AgNkuAIDdLgCA4S4AgLcpAIAk6gCAuykAgAQvAIAILwCADC8AgOvrAIAu6gCA5zUAgIJh+wCBCT4AgBE+ADwvAIC/KQCAUeoAgCPrAIC7ZT4AumU+ALl1PgC4dT4At2HkALah8AC1TT4AtHE+ALNpPgCyYT4AsWk+ALCl7wCvDT4AronwAK3h9ACshfAAqxk+AKrJ9ACpbecAqBk+AKchPgCmWT4ApVE+AKRZPgCjQT4Aos3iAKFVPgCgVT4ATC8AgFQvAIDDKQCAaC8AgHgvAIB8LwCAhC8AgJQvAIDLKQCAxykAgDPqAICcLwCAsC8AgLQvAIDELwCA2C8AgNAvAIDULwCA3C8AgPAvAID0LwCADDAAgBQwAIAkMACAWAAAADgwAIBC6gCANDAAgCgwAIBAMACASDAAgFwwAIBH6gCAZDAAgFgwAIBQMACAzykAgGwwAIB0MACAfDAAgHAwAIDTKQCAlDAAgEzqAIDAMACAAjEAgN4wAIDfKQCA2ykAgNcpAICqKwCArisAgAYxAIDuMACAuzUAgEMqAIAaMQCALjEAgCYxAIBl6gCA4ykAgEIxAIA2MQCAXzEAgICJAQCBiQEAgpkBAIOZAQCEiQEAhYkBAIa5AQCHuQEAiN3EAImNAQCKhQEAi50BAIyFAQCNjQEAjoUBAI8BxwCQgQEAkYEBAJKBAQCTgQEAlIEBAJWBAQCWgQEAl4EBAJj1zACZsdkAmokBAJuJAQCcmQEAnZkBAJ6JAQCfiQEAoHkBAKF5AQCizccAo4kCAKSZAgClmQIApokCAKfh2wCotQIAqb0CAKq1AgCrjQIArJUCAK2dAgCulQIAr40CALD1AgCx/QIAsvUCALONAgC0lQIAtfXAALaRAgC3kQIAuLECALmxAgC6sQIAu7ECALyRAgC9kQIAvpECAL+RAgCYjQ0AmQUCAJoNAgCbBQIAnB0CAJ0FAgCeDQIAnwUCAJBt6QCRSQ0AkkUNAJNdDQCUsQ0AlbUNAJa9DQCXtQ0AiBENAIkRDQCKEQ0AixENAIwxDQCNMQ0AjsnPAI81DQCAkQ4AgZEOAIKRDgCDkQ4AhDENAIUxDQCGMQ0AhzENALjpAgC56QIAuvkCALv5AgC86QIAvekCAL4ZAgC/GQIAsFnHALGpzwCy5QIAs/0CALTlAgC17QIAtuUCALfZAgCoddoAqfECAKrxAgCrPccArO0CAK2VAgCunQIAr5UCAKD9AgChxQIAos0CAKMFxwCk2QIApdkCAKbJAgCnyQIAgAAAAG/qAIBrMQCASjEAgHMxAIB3MQCAezEAgH8xAICLMQCAdOoAgJMxAIDrKQCAnzEAgHnqAICjMQCA7ykAgK8xAIC7MQCAyzEAgH7qAIAV6gCAg+oAgOsxAICI6gCA9zEAgP8xAIDvMQCACzIAgBsyAIAjMgCALzIAgDMyAICN6gCAFzIAgEsyAIBPMgCA8ykAgF8yAICS6gCAQzIAgH8yAICX6gCAnOoAgIMyAICXMgCAjzIAgPcpAICbMgCAqzIAgKcyAICzMgCA2ekAgMMyAICh6gCAzzIAgKvqAIDjMgCAAzMAgLDqAIAXMwCAGzMAgLXqAIC66gCANzMAgEczAID7KQCASzMAgP8pAIBjMwCAZzMAgHMzAIB/MwCAAyoAgJczAIC/6gCAszMAgMTqAIAp6gCAzzMAgMnqAIDO6gCA0+oAgAcqAIALKgCA3eoAgNjqAIDi6gCA5+oAgA80AIATNACAHzQAgBcqAIAbKgCA8eoAgF4AAAAzNACAHyoAgPbqAID76gCAAOsAgKM0AIAjKgCArzQAgLM0AIAF6wCACusAgMs0AIAnKgCAD+sAgN80AIDjNACAKyoAgBTrAID/NACALyoAgA81AIAHNQCAFzUAgB7rAIAvNQCAMyoAgDs1AIBDNQCAUzUAgDcqAIAo6wCALesAgDsqAICADd8AgVUBAIJdAQCDVQEAhE0BAIV1AQCGfQEAh3kBAIgx3wCJod8AikEBAItBAQCMQQEAjUEBAI5FAQCP5cgAkFnfAJHFAQCSzQEAkwHCAJTZAQCV2QEAlt3AAJfNAQCY9QEAmWHdAJrxAQCb8QEAnNEBAJ3RAQCe3QEAn9UBAKAtAQChNQEAoj0BAKM1AQCkLQEApVUBAKZdAQCnVQEAqG0BAKl1AQCqfQEAq3UBAKxtAQCtVQEArl0BAK9VAQCwLQEAsTUBALI9AQCzNQEAtC0BALXRAgC20QIAt9ECALjxAgC58QIAuvXdALv1AgC87QIAvdUCAL7dAgC/1QIAntkFAJ/ZBQCc2QUAndkFAJrZBQCb2QUAmNkFAJnZBQCWmQUAl5kFAJTN3gCVmQUAkp0FAJOFBQCQnQUAkZUFAI7dBQCP5QUAjN0FAI3VBQCKwQUAi7XeAIjRBQCJ0QUAhuEFAIfhBQCEEQUAhREFAIIdxQCDAQUAgAkFAIExwAC+yQIAv8kCALzJAgC9yQIAuqkCALupAgC4ccgAuakCALaNAgC3lQIAtI0CALWFAgCyrQIAs5UCALBN3gCxpQIArtECAK/RAgCsxQIArcECAKrVAgCr3QIAqCUFAKndAgCmFQUApx0FAKQFBQClHQUAohUFAKMdBQCgGQUAoRHeAIAAAAAy6wCAazUAgDfrAIB3NQCAgzUAgDzrAIBB6wCAnzUAgEbrAICnNQCAVuoAgD8qAIC/NQCAwzUAgEcqAIDHNQCAIS4AgEvrAIBQ6wCAW+oAgGDqAIDrNQCAAzgAgEsqAIAXNgCAEzYAgBs2AIAmLACAHzYAgCM2AIAnNgCALzYAgFXrAIAzNgCARzYAgEs2AIA3NgCATzYAgGM2AIBDNgCAVzYAgFs2AIBfNgCAWusAgGTrAIBf6wCATyoAgH82AICDNgCAizYAgHjrAICPNgCAaesAgFMqAIBXKgCAbusAgHPrAIBbKgCArzYAgLc2AIC7NgCAxzYAgMM2AIDPNgCAyzYAgNM2AIDXNgCA3zYAgF8qAIDnNgCA6zYAgGMqAID7NgCAfesAgAs3AIAPNwCAZyoAgBs3AICbKQCAgusAgIfrAIBrKgCAbyoAgEc3AICM6wCAnzcAgKM3AIC7NwCAxzcAgJbrAIDo6QCAXQAAANM3AIDPNwCA2zcAgO3pAIDnNwCAm+sAgKDrAIAzOACAPzgAgEc4AICl6wCASzgAgHc4AICDOACAhzgAgH84AICTOACAlzgAgKrrAICjOACAcyoAgKs4AICv6wCAdyoAgOM4AICxLgCA+zgAgLTrAIC+6wCAeyoAgH8qAIAjOQCAw+sAgIMqAIDI6wCAgBkBAIEZAQCCKQEAgykBAIT99ACFPQEAhjUBAIctAQCIFQEAiR0BAIoVAQCLbQEAjHUBAI19AQCOdQEAj20BAJAJwwCRCcMAkkH5AJMZAQCUCQEAlQkBAJY5AQCXOQEAmAkBAJkJAQCaHQEAmxUBAJwNAQCd9QEAnv0BAJ8twwCgCQEAoQkBAKIZAQCjGQEApAkBAKUJAQCmOQEApzkBAKgJAQCpCQEAqhkBAKsZAQCsCQEArQkBAK55AQCveQEAsAkBALEJAQCyGQEAsxkBALQJAQC1CQEAtjkBALc5AQC4CQEAuQkBALoZAQC7GQEAvAkBAL0JAQC+kcMAv5XDAJyVHQCdnR0AnpUdAJ8lwgCYPdsAmZ0dAJqVHQCbjR0AlFkdAJVZHQCWqR0Al6kdAJBZHQCRWR0AkkkdAJNJHQCMGR0AjRkdAI4pHQCPKR0AiB0dAIkFHQCKDR0Aiy0VAIRdHQCFJR0Ahi0dAIclHQCAXR0AgUUdAIJNHQCDRR0AvIkCAL2JAgC+mQIAv5kCALhtHQC5lQIAup0CALupwAC0ZdcAtVUdALZdHQC3VR0AsFEdALFRHQCyUR0As1EdAKwRHQCtER0ArhEdAK8RHQCoER0AqREdAKoRHQCrER0ApFEdAKVRHQCmUR0Ap1EdAKBRHQChUR0AolEdAKNRHQCAeQAAgXkAAIKJAACDiQAAhJkAAIWZAACGiQAAh4kAAIi5AACJuQAAikXBAIuNAACMlQAAjZ0AAI6VAACPjQAAkPUAAJH9AACS9QAAk40AAJSVAACVnfsAlpEAAJeF+wCYrQAAmbUAAJq9AACbtQAAnJ37AJ2pAABDOQCAzesAgFs5AICHKgCAazkAgHc5AIB/OQCAhzkAgIsqAIDS6wCAtzkAgMM5AICPKgCAkyoAgMc5AIDX6wCAlyoAgNzrAIDh6wCA5usAgJsqAIAHOgCACzoAgBM6AIAbOgCA8OsAgLkAAAC4AAAAuwAAALoAAAC9AAAAvAAAAL8AAAC+AAAAACAAIMyBACDMgwAgzIQAIMyFACDMhgAgzIcAIMyIACDMiMyAACDMiMyBACDMiM2CACDMigAgzIsAIMyTACDMk8yAACDMk8yBACDMk82CACDMlAAgzJTMgAAgzJTMgQAgzJTNggAgzKcAIMyoACDMswAgzYIAIM2FACDZiwAg2YwAINmM2ZEAINmNACDZjdmRACDZjgAg2Y7ZkQAg2Y8AINmP2ZEAINmQACDZkNmRACDZkQAg2ZHZsAAg2ZIAIOOCmQAg44KaACEAISEAIT8AIgAjACQAJQAmACcAKAAoMSkAKDEwKQAoMTEpACgxMikAKDEzKQAoMTQpACgxNSkAKDE2KQAoMTcpACgxOCkAKDE5KQAoMikAKDIwKQAoMykAKDQpACg1KQAoNikAKDcpACg4KQAoOSkAKEEpAChCKQAoQykAKEQpAChFKQAoRikAKEcpAChIKQAoSSkAKEopAChLKQAoTCkAKE0pAChOKQAoTykAKFApAChRKQAoUikAKFMpAChUKQAoVSkAKFYpAChXKQAoWCkAKFkpAChaKQAoYSkAKGIpAChjKQAoZCkAKGUpAChmKQAoZykAKGgpAChpKQAoaikAKGspAChsKQAobSkAKG4pAChvKQAocCkAKHEpAChyKQAocykAKHQpACh1KQAodikAKHcpACh4KQAoeSkAKHopACjhhIApACjhhIIpACjhhIMpACjhhIUpACjhhIYpACjhhIcpACjhhIkpACjhhIspACjhhIwpACjhhI4pACjhhI8pACjhhJApACjhhJEpACjhhJIpACjkuIApACjkuIMpACjkuIkpACjkuZ0pACjkuowpACjkupQpACjku6MpACjkvIEpACjkvJEpACjlhaspACjlha0pACjlirQpACjljYEpACjljZQpACjlkI0pACjlkbwpACjlm5spACjlnJ8pACjlraYpACjml6UpACjmnIgpACjmnIkpACjmnKgpACjmoKopACjmsLQpACjngaspACjnibkpACjnm6MpACjnpL4pACjnpZ0pACjnpa0pACjoh6opACjoh7MpACjosqEpACjos4cpACjph5EpACjqsIApACjrgpgpACjri6QpACjrnbwpACjrp4gpACjrsJQpACjsgqwpACjslYQpACjsmKTsoIQpACjsmKTtm4QpACjsnpApACjso7wpACjssKgpACjsubQpACjtg4ApACjtjIwpACjtlZgpACkAKgArACwALQAuAC4uAC4uLgAvADAAMCwAMC4AMOKBhDMAMOeCuQAxADEsADEuADEwADEwLgAxMOaXpQAxMOaciAAxMOeCuQAxMQAxMS4AMTHml6UAMTHmnIgAMTHngrkAMTIAMTIuADEy5pelADEy5pyIADEy54K5ADEzADEzLgAxM+aXpQAxM+eCuQAxNAAxNC4AMTTml6UAMTTngrkAMTUAMTUuADE15pelADE154K5ADE2ADE2LgAxNuaXpQAxNueCuQAxNwAxNy4AMTfml6UAMTfngrkAMTgAMTguADE45pelADE454K5ADE5ADE5LgAxOeaXpQAxOeeCuQAx4oGEADHigYQxMAAx4oGEMgAx4oGEMwAx4oGENAAx4oGENQAx4oGENgAx4oGENwAx4oGEOAAx4oGEOQAx5pelADHmnIgAMeeCuQAyADIsADIuADIwADIwLgAyMOaXpQAyMOeCuQAyMQAyMeaXpQAyMeeCuQAyMgAyMuaXpQAyMueCuQAyMwAyM+aXpQAyM+eCuQAyNAAyNOaXpQAyNOeCuQAyNQAyNeaXpQAyNgAyNuaXpQAyNwAyN+aXpQAyOAAyOOaXpQAyOQAyOeaXpQAy4oGEMwAy4oGENQAy5pelADLmnIgAMueCuQAzADMsADMuADMwADMw5pelADMxADMx5pelADMyADMzADM0ADM1ADM2ADM3ADM4ADM5ADPigYQ0ADPigYQ1ADPigYQ4ADPml6UAM+aciAAz54K5ADQANCwANC4ANDAANDEANDIANDMANDQANDUANDYANDcANDgANDkANOKBhDUANOaXpQA05pyIADTngrkANQA1LAA1LgA1MAA14oGENgA14oGEOAA15pelADXmnIgANeeCuQA2ADYsADYuADbml6UANuaciAA254K5ADcANywANy4AN+KBhDgAN+aXpQA35pyIADfngrkAOAA4LAA4LgA45pelADjmnIgAOOeCuQA5ADksADkuADnml6UAOeaciAA554K5ADoAOjo9ADsAPAA9AD09AD09PQA+AD8APyEAPz8AQABBAEFVAEHiiJVtAEIAQnEAQwBDRABDby4AQ+KIlWtnAEQAREoARFoARHoARMW9AETFvgBFAEYARkFYAEcAR0IAR0h6AEdQYQBHeQBIAEhQAEhWAEhnAEh6AEkASUkASUlJAElKAElVAElWAElYAEoASwBLQgBLSwBLTQBMAExKAExURABMagBMwrcATQBNQgBNQwBNRABNSHoATVBhAE1SAE1WAE1XAE3OqQBOAE5KAE5qAE5vAE8AUABQSABQUE0AUFBWAFBSAFBURQBQYQBRAFIAUnMAUwBTRABTTQBTUwBTdgBUAFRFTABUSHoAVE0AVQBWAFZJAFZJSQBWSUlJAFbiiJVtAFcAV0MAV1oAV2IAWABYSQBYSUkAWQBaAFsAXABdAF4AXwBgAGEAYS5tLgBhL2MAYS9zAGHKvgBiAGJhcgBjAGMvbwBjL3UAY2FsAGNjAGNkAGNtAGNtMgBjbTMAZABkQgBkYQBkbABkbQBkbTIAZG0zAGR6AGTFvgBlAGVWAGVyZwBmAGZmAGZmaQBmZmwAZmkAZmwAZm0AZwBnYWwAaABoUGEAaGEAaQBpaQBpaWkAaWoAaW4AaXYAaXgAagBrAGtBAGtIegBrUGEAa1YAa1cAa2NhbABrZwBrbABrbQBrbTIAa20zAGt0AGvOqQBsAGxqAGxtAGxuAGxvZwBseABswrcAbQBtMgBtMwBtQQBtVgBtVwBtYgBtZwBtaWwAbWwAbW0AbW0yAG1tMwBtb2wAbXMAbeKIlXMAbeKIlXMyAG4AbkEAbkYAblYAblcAbmoAbm0AbnMAbwBvVgBwAHAubS4AcEEAcEYAcFYAcFcAcGMAcHMAcQByAHJhZAByYWTiiJVzAHJhZOKIlXMyAHMAc3IAc3QAdAB1AHYAdmkAdmlpAHZpaWkAdwB4AHhpAHhpaQB5AHoAewB8AH0AwqIAwqMAwqUAwqYAwqwAwrBDAMKwRgDCtwDDgADDgQDDggDDgwDDhADDhQDDhgDDhwDDiADDiQDDigDDiwDDjADDjQDDjgDDjwDDkQDDkgDDkwDDlADDlQDDlgDDmQDDmgDDmwDDnADDnQDDoADDoQDDogDDowDDpADDpQDDpwDDqADDqQDDqgDDqwDDrADDrQDDrgDDrwDDsADDsQDDsgDDswDDtADDtQDDtgDDuQDDugDDuwDDvADDvQDDvwDEgADEgQDEggDEgwDEhADEhQDEhgDEhwDEiADEiQDEigDEiwDEjADEjQDEjgDEjwDEkgDEkwDElADElQDElgDElwDEmADEmQDEmgDEmwDEnADEnQDEngDEnwDEoADEoQDEogDEowDEpADEpQDEpgDEpwDEqADEqQDEqgDEqwDErADErQDErgDErwDEsADEsQDEtADEtQDEtgDEtwDEuQDEugDEuwDEvADEvQDEvgDFgwDFhADFhQDFhgDFhwDFiADFiwDFjADFjQDFjgDFjwDFkADFkQDFkwDFlADFlQDFlgDFlwDFmADFmQDFmgDFmwDFnADFnQDFngDFnwDFoADFoQDFogDFowDFpADFpQDFqADFqQDFqgDFqwDFrADFrQDFrgDFrwDFsADFsQDFsgDFswDFtADFtQDFtgDFtwDFuADFuQDFugDFuwDFvADFvQDFvgDGjgDGkADGoADGoQDGqwDGrwDGsADHjQDHjgDHjwDHkADHkQDHkgDHkwDHlADHlQDHlgDHlwDHmADHmQDHmgDHmwDHnADHngDHnwDHoADHoQDHogDHowDHpgDHpwDHqADHqQDHqgDHqwDHrADHrQDHrgDHrwDHsADHtADHtQDHuADHuQDHugDHuwDHvADHvQDHvgDHvwDIgADIgQDIggDIgwDIhADIhQDIhgDIhwDIiADIiQDIigDIiwDIjADIjQDIjgDIjwDIkADIkQDIkgDIkwDIlADIlQDIlgDIlwDImADImQDImgDImwDIngDInwDIogDIpgDIpwDIqADIqQDIqgDIqwDIrADIrQDIrgDIrwDIsADIsQDIsgDIswDItwDJkADJkQDJkgDJlADJlQDJmQDJmwDJnADJnwDJoQDJowDJpQDJpgDJqADJqQDJqgDJqwDJrQDJrwDJsADJsQDJsgDJswDJtADJtQDJuADJuQDJuwDKgQDKggDKgwDKiQDKigDKiwDKjADKjQDKkADKkQDKkgDKlQDKnQDKnwDKuQDKvG4AzIAAzIEAzIjMgQDMkwDOhgDOiADOiQDOigDOjADOjgDOjwDOkADOkQDOkgDOkwDOlADOlQDOlgDOlwDOmADOmQDOmgDOmwDOnADOnQDOngDOnwDOoADOoQDOowDOpADOpQDOpgDOpwDOqADOqQDOqgDOqwDOrADOrQDOrgDOrwDOsADOsQDOsgDOswDOtADOtQDOtgDOtwDOuADOuQDOugDOuwDOvADOvEEAzrxGAM68VgDOvFcAzrxnAM68bADOvG0AzrxzAM69AM6+AM6/AM+AAM+BAM+CAM+DAM+EAM+FAM+GAM+HAM+IAM+JAM+KAM+LAM+MAM+NAM+OAM+cAM+dANCAANCBANCDANCHANCMANCNANCOANCZANC5ANC9ANGKANGMANGQANGRANGTANGXANGcANGdANGeANG2ANG3ANOBANOCANOQANORANOSANOTANOWANOXANOaANObANOcANOdANOeANOfANOiANOjANOkANOlANOmANOnANOqANOrANOsANOtANOuANOvANOwANOxANOyANOzANO0ANO1ANO4ANO5ANWl1oIA1bTVpQDVtNWrANW01a0A1bTVtgDVvtW2ANeQANeQ1rcA15DWuADXkNa8ANeQ15wA15EA15HWvADXkda/ANeSANeS1rwA15MA15PWvADXlADXlNa8ANeV1rkA15XWvADXlta8ANeY1rwA15nWtADXmda8ANea1rwA15sA15vWvADXm9a/ANecANec1rwA150A157WvADXoNa8ANeh1rwA16IA16PWvADXpNa8ANek1r8A16bWvADXp9a8ANeoANeo1rwA16nWvADXqda814EA16nWvNeCANep14EA16nXggDXqgDXqta8ANey1rcA2KEA2KIA2KMA2KQA2KUA2KYA2KbYpwDYptisANim2K0A2KbYrgDYptixANim2LIA2KbZhQDYptmGANim2YcA2KbZiADYptmJANim2YoA2KbbhgDYptuHANim24gA2KbbkADYptuVANinANin2YPYqNixANin2YTZhNmHANin2YsA2KfZtADYqADYqNisANio2K0A2KjYrdmKANio2K4A2KjYrtmKANio2LEA2KjYsgDYqNmFANio2YYA2KjZhwDYqNmJANio2YoA2KkA2KoA2KrYrADYqtis2YUA2KrYrNmJANiq2KzZigDYqtitANiq2K3YrADYqtit2YUA2KrYrgDYqtiu2YUA2KrYrtmJANiq2K7ZigDYqtixANiq2LIA2KrZhQDYqtmF2KwA2KrZhditANiq2YXYrgDYqtmF2YkA2KrZhdmKANiq2YYA2KrZhwDYqtmJANiq2YoA2KsA2KvYrADYq9ixANir2LIA2KvZhQDYq9mGANir2YcA2KvZiQDYq9mKANisANis2K0A2KzYrdmJANis2K3ZigDYrNmEINis2YTYp9mE2YcA2KzZhQDYrNmF2K0A2KzZhdmJANis2YXZigDYrNmJANis2YoA2K0A2K3YrADYrdis2YoA2K3ZhQDYrdmF2YkA2K3ZhdmKANit2YkA2K3ZigDYrgDYrtisANiu2K0A2K7ZhQDYrtmJANiu2YoA2K8A2LAA2LDZsADYsQDYsdiz2YjZhADYsdmwANix24zYp9mEANiyANizANiz2KwA2LPYrNitANiz2KzZiQDYs9itANiz2K3YrADYs9iuANiz2K7ZiQDYs9iu2YoA2LPYsQDYs9mFANiz2YXYrADYs9mF2K0A2LPZhdmFANiz2YcA2LPZiQDYs9mKANi0ANi02KwA2LTYrNmKANi02K0A2LTYrdmFANi02K3ZigDYtNiuANi02LEA2LTZhQDYtNmF2K4A2LTZhdmFANi02YcA2LTZiQDYtNmKANi1ANi12K0A2LXYrditANi12K3ZigDYtdiuANi12LEA2LXZhNi52YUA2LXZhNmJANi12YTZiSDYp9mE2YTZhyDYudmE2YrZhyDZiNiz2YTZhQDYtdmE25IA2LXZhQDYtdmF2YUA2LXZiQDYtdmKANi2ANi22KwA2LbYrQDYttit2YkA2LbYrdmKANi22K4A2LbYrtmFANi22LEA2LbZhQDYttmJANi22YoA2LcA2LfYrQDYt9mFANi32YXYrQDYt9mF2YUA2LfZhdmKANi32YkA2LfZigDYuADYuNmFANi5ANi52KwA2LnYrNmFANi52YTZitmHANi52YUA2LnZhdmFANi52YXZiQDYudmF2YoA2LnZiQDYudmKANi6ANi62KwA2LrZhQDYutmF2YUA2LrZhdmJANi62YXZigDYutmJANi62YoA2YDZiwDZgNmOANmA2Y7ZkQDZgNmPANmA2Y/ZkQDZgNmQANmA2ZDZkQDZgNmRANmA2ZIA2YEA2YHYrADZgditANmB2K4A2YHYrtmFANmB2YUA2YHZhdmKANmB2YkA2YHZigDZggDZgtitANmC2YTbkgDZgtmFANmC2YXYrQDZgtmF2YUA2YLZhdmKANmC2YkA2YLZigDZgwDZg9inANmD2KwA2YPYrQDZg9iuANmD2YQA2YPZhQDZg9mF2YUA2YPZhdmKANmD2YkA2YPZigDZhADZhNiiANmE2KMA2YTYpQDZhNinANmE2KwA2YTYrNisANmE2KzZhQDZhNis2YoA2YTYrQDZhNit2YUA2YTYrdmJANmE2K3ZigDZhNiuANmE2K7ZhQDZhNmFANmE2YXYrQDZhNmF2YoA2YTZhwDZhNmJANmE2YoA2YUA2YXYpwDZhdisANmF2KzYrQDZhdis2K4A2YXYrNmFANmF2KzZigDZhditANmF2K3YrADZhdit2YUA2YXYrdmF2K8A2YXYrdmKANmF2K4A2YXYrtisANmF2K7ZhQDZhdiu2YoA2YXZhQDZhdmF2YoA2YXZiQDZhdmKANmGANmG2KwA2YbYrNitANmG2KzZhQDZhtis2YkA2YbYrNmKANmG2K0A2YbYrdmFANmG2K3ZiQDZhtit2YoA2YbYrgDZhtixANmG2LIA2YbZhQDZhtmF2YkA2YbZhdmKANmG2YYA2YbZhwDZhtmJANmG2YoA2YcA2YfYrADZh9mFANmH2YXYrADZh9mF2YUA2YfZiQDZh9mKANmH2bAA2YgA2YjYs9mE2YUA2YjZtADZiQDZidmwANmKANmK2KwA2YrYrNmKANmK2K0A2YrYrdmKANmK2K4A2YrYsQDZitiyANmK2YUA2YrZhdmFANmK2YXZigDZitmGANmK2YcA2YrZiQDZitmKANmK2bQA2a4A2a8A2bEA2bkA2boA2bsA2b4A2b8A2oAA2oMA2oQA2oYA2ocA2ogA2owA2o0A2o4A2pEA2pgA2qEA2qQA2qYA2qkA2q0A2q8A2rEA2rMA2roA2rsA2r4A24AA24EA24IA24UA24YA24cA24fZtADbiADbiQDbiwDbjADbkADbkgDbkwDgpJXgpLwA4KSW4KS8AOCkl+CkvADgpJzgpLwA4KSh4KS8AOCkouCkvADgpKkA4KSr4KS8AOCkr+CkvADgpLEA4KS0AOCmoeCmvADgpqLgprwA4Kav4Ka8AOCniwDgp4wA4KiW4Ki8AOCol+CovADgqJzgqLwA4Kir4Ki8AOCosuCovADgqLjgqLwA4Kyh4Ky8AOCsouCsvADgrYgA4K2LAOCtjADgrpQA4K+KAOCviwDgr4wA4LGIAOCzgADgs4cA4LOIAOCzigDgs4sA4LWKAOC1iwDgtYwA4LeaAOC3nADgt50A4LeeAOC5jeC4sgDguqvgupkA4Lqr4LqhAOC7jeC6sgDgvIsA4L2A4L61AOC9guC+twDgvYzgvrcA4L2R4L63AOC9luC+twDgvZvgvrcA4L2x4L2yAOC9seC9tADgvbHgvoAA4L6Q4L61AOC+kuC+twDgvpzgvrcA4L6h4L63AOC+puC+twDgvqvgvrcA4L6y4L2x4L6AAOC+suC+gADgvrPgvbHgvoAA4L6z4L6AAOGApgDhg5wA4YSAAOGEgQDhhIIA4YSDAOGEhADhhIUA4YSGAOGEhwDhhIgA4YSJAOGEigDhhIsA4YSMAOGEjQDhhI4A4YSPAOGEkADhhJEA4YSSAOGElADhhJUA4YSaAOGEnADhhJ0A4YSeAOGEoADhhKEA4YSiAOGEowDhhKcA4YSpAOGEqwDhhKwA4YStAOGErgDhhK8A4YSyAOGEtgDhhYAA4YWHAOGFjADhhZcA4YWYAOGFmQDhhaAA4YWhAOGFogDhhaMA4YWkAOGFpQDhhaYA4YWnAOGFqADhhakA4YWqAOGFqwDhhawA4YWtAOGFrgDhha8A4YWwAOGFsQDhhbIA4YWzAOGFtADhhbUA4YaEAOGGhQDhhogA4YaRAOGGkgDhhpQA4YaeAOGGoQDhhqoA4YasAOGGrQDhhrAA4YaxAOGGsgDhhrMA4Ya0AOGGtQDhh4cA4YeIAOGHjADhh44A4YeTAOGHlwDhh5kA4YedAOGHnwDhh7EA4YeyAOGshgDhrIgA4ayKAOGsjADhrI4A4aySAOGsuwDhrL0A4a2AAOGtgQDhrYMA4bSCAOG0lgDhtJcA4bScAOG0nQDhtKUA4bW7AOG2hQDhuIAA4biBAOG4ggDhuIMA4biEAOG4hQDhuIYA4biHAOG4iADhuIkA4biKAOG4iwDhuIwA4biNAOG4jgDhuI8A4biQAOG4kQDhuJIA4biTAOG4lADhuJUA4biWAOG4lwDhuJgA4biZAOG4mgDhuJsA4bicAOG4nQDhuJ4A4bifAOG4oADhuKEA4biiAOG4owDhuKQA4bilAOG4pgDhuKcA4bioAOG4qQDhuKoA4birAOG4rADhuK0A4biuAOG4rwDhuLAA4bixAOG4sgDhuLMA4bi0AOG4tQDhuLYA4bi3AOG4uADhuLkA4bi6AOG4uwDhuLwA4bi9AOG4vgDhuL8A4bmAAOG5gQDhuYIA4bmDAOG5hADhuYUA4bmGAOG5hwDhuYgA4bmJAOG5igDhuYsA4bmMAOG5jQDhuY4A4bmPAOG5kADhuZEA4bmSAOG5kwDhuZQA4bmVAOG5lgDhuZcA4bmYAOG5mQDhuZoA4bmbAOG5nADhuZ0A4bmeAOG5nwDhuaAA4bmhAOG5ogDhuaMA4bmkAOG5pQDhuaYA4bmnAOG5qADhuakA4bmqAOG5qwDhuawA4bmtAOG5rgDhua8A4bmwAOG5sQDhubIA4bmzAOG5tADhubUA4bm2AOG5twDhubgA4bm5AOG5ugDhubsA4bm8AOG5vQDhub4A4bm/AOG6gADhuoEA4bqCAOG6gwDhuoQA4bqFAOG6hgDhuocA4bqIAOG6iQDhuooA4bqLAOG6jADhuo0A4bqOAOG6jwDhupAA4bqRAOG6kgDhupMA4bqUAOG6lQDhupYA4bqXAOG6mADhupkA4bqgAOG6oQDhuqIA4bqjAOG6pADhuqUA4bqmAOG6pwDhuqgA4bqpAOG6qgDhuqsA4bqsAOG6rQDhuq4A4bqvAOG6sADhurEA4bqyAOG6swDhurQA4bq1AOG6tgDhurcA4bq4AOG6uQDhuroA4bq7AOG6vADhur0A4bq+AOG6vwDhu4AA4buBAOG7ggDhu4MA4buEAOG7hQDhu4YA4buHAOG7iADhu4kA4buKAOG7iwDhu4wA4buNAOG7jgDhu48A4buQAOG7kQDhu5IA4buTAOG7lADhu5UA4buWAOG7lwDhu5gA4buZAOG7mgDhu5sA4bucAOG7nQDhu54A4bufAOG7oADhu6EA4buiAOG7owDhu6QA4bulAOG7pgDhu6cA4buoAOG7qQDhu6oA4burAOG7rADhu60A4buuAOG7rwDhu7AA4buxAOG7sgDhu7MA4bu0AOG7tQDhu7YA4bu3AOG7uADhu7kA4byAAOG8gQDhvIIA4byDAOG8hADhvIUA4byGAOG8hwDhvIgA4byJAOG8igDhvIsA4byMAOG8jQDhvI4A4byPAOG8kADhvJEA4bySAOG8kwDhvJQA4byVAOG8mADhvJkA4byaAOG8mwDhvJwA4bydAOG8oADhvKEA4byiAOG8owDhvKQA4bylAOG8pgDhvKcA4byoAOG8qQDhvKoA4byrAOG8rADhvK0A4byuAOG8rwDhvLAA4byxAOG8sgDhvLMA4by0AOG8tQDhvLYA4by3AOG8uADhvLkA4by6AOG8uwDhvLwA4by9AOG8vgDhvL8A4b2AAOG9gQDhvYIA4b2DAOG9hADhvYUA4b2IAOG9iQDhvYoA4b2LAOG9jADhvY0A4b2QAOG9kQDhvZIA4b2TAOG9lADhvZUA4b2WAOG9lwDhvZkA4b2bAOG9nQDhvZ8A4b2gAOG9oQDhvaIA4b2jAOG9pADhvaUA4b2mAOG9pwDhvagA4b2pAOG9qgDhvasA4b2sAOG9rQDhva4A4b2vAOG9sADhvbIA4b20AOG9tgDhvbgA4b26AOG9vADhvoAA4b6BAOG+ggDhvoMA4b6EAOG+hQDhvoYA4b6HAOG+iADhvokA4b6KAOG+iwDhvowA4b6NAOG+jgDhvo8A4b6QAOG+kQDhvpIA4b6TAOG+lADhvpUA4b6WAOG+lwDhvpgA4b6ZAOG+mgDhvpsA4b6cAOG+nQDhvp4A4b6fAOG+oADhvqEA4b6iAOG+owDhvqQA4b6lAOG+pgDhvqcA4b6oAOG+qQDhvqoA4b6rAOG+rADhvq0A4b6uAOG+rwDhvrAA4b6xAOG+sgDhvrMA4b60AOG+tgDhvrcA4b64AOG+uQDhvroA4b68AOG/ggDhv4MA4b+EAOG/hgDhv4cA4b+IAOG/igDhv4wA4b+QAOG/kQDhv5IA4b+WAOG/lwDhv5gA4b+ZAOG/mgDhv6AA4b+hAOG/ogDhv6QA4b+lAOG/pgDhv6cA4b+oAOG/qQDhv6oA4b+sAOG/sgDhv7MA4b+0AOG/tgDhv7cA4b+4AOG/ugDhv7wA4oCQAOKAkwDigJQA4oCy4oCyAOKAsuKAsuKAsgDigLLigLLigLLigLIA4oC14oC1AOKAteKAteKAtQDigqkA4oaQAOKGkQDihpIA4oaTAOKGmgDihpsA4oauAOKHjQDih44A4oePAOKIggDiiIQA4oiHAOKIiQDiiIwA4oiRAOKIkgDiiKQA4oimAOKIq+KIqwDiiKviiKviiKsA4oir4oir4oir4oirAOKIruKIrgDiiK7iiK7iiK4A4omBAOKJhADiiYcA4omJAOKJoADiiaIA4omtAOKJrgDiia8A4omwAOKJsQDiibQA4om1AOKJuADiibkA4oqAAOKKgQDiioQA4oqFAOKKiADiiokA4oqsAOKKrQDiiq4A4oqvAOKLoADii6EA4ouiAOKLowDii6oA4ourAOKLrADii60A4pSCAOKWoADil4sA4qaFAOKmhgDiq53MuADitaEA44CBAOOAggDjgIgA44CJAOOAigDjgIsA44CMAOOAjQDjgI4A44CPAOOAkADjgJEA44CSAOOAlADjgJRT44CVAOOAlOS4ieOAlQDjgJTkuozjgJUA44CU5Yud44CVAOOAlOWuieOAlQDjgJTmiZPjgJUA44CU5pWX44CVAOOAlOacrOOAlQDjgJTngrnjgJUA44CU55uX44CVAOOAlQDjgJYA44CXAOOBjADjgY4A44GQAOOBkgDjgZQA44GWAOOBmADjgZoA44GcAOOBngDjgaAA44GiAOOBpQDjgacA44GpAOOBsADjgbEA44GzAOOBtADjgbYA44G3AOOBuQDjgboA44G744GLAOOBvADjgb0A44KI44KKAOOClADjgpkA44KaAOOCngDjgqEA44KiAOOCouODkeODvOODiADjgqLjg6vjg5XjgqEA44Ki44Oz44Oa44KiAOOCouODvOODqwDjgqMA44KkAOOCpOODi+ODs+OCsADjgqTjg7Pjg4EA44KlAOOCpgDjgqbjgqnjg7MA44KnAOOCqADjgqjjgrnjgq/jg7zjg4kA44Ko44O844Kr44O8AOOCqQDjgqoA44Kq44Oz44K5AOOCquODvOODoADjgqsA44Kr44Kk44OqAOOCq+ODqeODg+ODiADjgqvjg63jg6rjg7wA44KsAOOCrOODreODswDjgqzjg7Pjg54A44KtAOOCreODpeODquODvADjgq3jg60A44Kt44Ot44Kw44Op44OgAOOCreODreODoeODvOODiOODqwDjgq3jg63jg6/jg4Pjg4gA44KuAOOCruOCrADjgq7jg4vjg7wA44Ku44Or44OA44O8AOOCrwDjgq/jg6vjgrzjgqTjg60A44Kv44Ot44O844ONAOOCsADjgrDjg6njg6AA44Kw44Op44Og44OI44OzAOOCsQDjgrHjg7zjgrkA44KyAOOCswDjgrPjgrMA44Kz44OIAOOCs+ODq+ODigDjgrPjg7zjg50A44K0AOOCtQDjgrXjgqTjgq/jg6sA44K144Oz44OB44O844OgAOOCtgDjgrcA44K344Oq44Oz44KwAOOCuADjgrkA44K6AOOCuwDjgrvjg7Pjg4EA44K744Oz44OIAOOCvADjgr0A44K+AOOCvwDjg4AA44OA44O844K5AOODgQDjg4IA44ODAOODhADjg4UA44OGAOODhwDjg4fjgrcA44OIAOODiOODswDjg4kA44OJ44OrAOODigDjg4rjg44A44OLAOODjADjg40A44OOAOODjuODg+ODiADjg48A44OP44Kk44OEAOODkADjg5Djg7zjg6zjg6sA44ORAOODkeODvOOCu+ODs+ODiADjg5Hjg7zjg4QA44OSAOODkwDjg5Pjg6sA44OUAOODlOOCouOCueODiOODqwDjg5Tjgq/jg6sA44OU44KzAOODlQDjg5XjgqHjg6njg4Pjg4kA44OV44Kj44O844OIAOODleODqeODswDjg5YA44OW44OD44K344Kn44OrAOODlwDjg5gA44OY44Kv44K/44O844OrAOODmOODq+ODhADjg5kA44OZ44O844K/AOODmgDjg5rjgr0A44Oa44OL44OSAOODmuODs+OCuQDjg5rjg7zjgrgA44ObAOODm+ODswDjg5vjg7zjg6sA44Ob44O844OzAOODnADjg5zjg6vjg4gA44OdAOODneOCpOODs+ODiADjg53jg7Pjg4kA44OeAOODnuOCpOOCr+ODrQDjg57jgqTjg6sA44Oe44OD44OPAOODnuODq+OCrwDjg57jg7Pjgrfjg6fjg7MA44OfAOODn+OCr+ODreODswDjg5/jg6oA44Of44Oq44OQ44O844OrAOODoADjg6EA44Oh44KsAOODoeOCrOODiOODswDjg6Hjg7zjg4jjg6sA44OiAOODowDjg6QA44Ok44O844OJAOODpOODvOODqwDjg6UA44OmAOODpuOCouODswDjg6cA44OoAOODqQDjg6oA44Oq44OD44OI44OrAOODquODqQDjg6sA44Or44OU44O8AOODq+ODvOODluODqwDjg6wA44Os44OgAOODrOODs+ODiOOCsuODswDjg60A44OvAOODr+ODg+ODiADjg7AA44OxAOODsgDjg7MA44O0AOODtwDjg7gA44O5AOODugDjg7sA44O8AOODvgDjkp4A45K5AOOSuwDjk58A45SVAOObrgDjm7wA456BAOOgrwDjoaIA46G8AOOjhwDjo6MA46ScAOOkugDjqK4A46msAOOrpADjrIgA46yZAOOtiQDjrp0A47CYAOOxjgDjtLMA47aWAOO6rADjurgA47ybAOO/vADkgIgA5ICYAOSAuQDkgYYA5IKWAOSDowDkhK8A5IiCAOSIpwDkiqAA5IyBAOSMtADkjZkA5I+VAOSPmQDkkIsA5JGrAOSUqwDklZ0A5JWhAOSVqwDkl5cA5Je5AOSYtQDkmr4A5JuHAOSmlQDkp6YA5KmuAOSptgDkqrIA5KyzAOSvjgDks44A5LOtAOSzuADktZYA5LiAAOS4gQDkuIMA5LiJAOS4igDkuIsA5LiNAOS4mQDkuKYA5LioAOS4rQDkuLIA5Li2AOS4uADkuLkA5Li9AOS4vwDkuYEA5LmZAOS5nQDkuoIA5LqFAOS6hgDkuowA5LqUAOS6oADkuqQA5LquAOS6ugDku4AA5LuMAOS7pADku6TlkowA5LyBAOS8kQDkvaAA5L6AAOS+hgDkvosA5L6uAOS+uwDkvr8A5YCCAOWAqwDlgboA5YKZAOWDjwDlg5oA5YOnAOWEqgDlhL8A5YWAAOWFhQDlhY0A5YWUAOWFpADlhaUA5YWnAOWFqADlhakA5YWrAOWFrQDlhbcA5YaAAOWGggDlho0A5YaSAOWGlQDlhpYA5YaXAOWGmQDlhqQA5YarAOWGrADlhrUA5Ya3AOWHiQDlh4wA5YecAOWHngDlh6AA5Ye1AOWIgADliIMA5YiHAOWIlwDliJ0A5YipAOWIugDliLsA5YmGAOWJjQDlibIA5Ym3AOWKiQDlipsA5YqjAOWKswDlirQA5YuHAOWLiQDli5IA5YueAOWLpADli7UA5Yu5AOWLugDljIUA5YyGAOWMlQDljJcA5YyaAOWMuADljLsA5Yy/AOWNgQDljYQA5Y2FAOWNiQDljZEA5Y2UAOWNmgDljZwA5Y2pAOWNsADljbMA5Y21AOWNvQDljb8A5Y6CAOWOtgDlj4MA5Y+IAOWPigDlj4wA5Y+fAOWPowDlj6UA5Y+rAOWPrwDlj7EA5Y+zAOWQhgDlkIgA5ZCNAOWQjwDlkJ0A5ZC4AOWQuQDlkYIA5ZGIAOWRqADlkp4A5ZKiAOWSvQDlk7YA5ZSQAOWVjwDllZMA5ZWVAOWVowDlloQA5ZaHAOWWmQDllp0A5ZarAOWWswDllrYA5ZeAAOWXggDll6IA5ZiGAOWZkQDlmagA5Zm0AOWblwDlm5sA5Zu5AOWclgDlnJcA5ZyfAOWcsADlnosA5Z+OAOWftADloI0A5aCxAOWgsgDloYAA5aGaAOWhngDloqgA5aKsAOWiswDlo5gA5aOfAOWjqwDlo64A5aOwAOWjsgDlo7cA5aSCAOWkhgDlpIoA5aSVAOWkmgDlpJwA5aSiAOWkpwDlpKfmraMA5aSpAOWlhADlpYgA5aWRAOWllADlpaIA5aWzAOWnmADlp6wA5aibAOWopwDlqaIA5ammAOWqtQDlrIgA5ayoAOWsvgDlrZAA5a2XAOWtpgDlroAA5a6FAOWulwDlr4MA5a+YAOWvpwDlr64A5a+zAOWvuADlr78A5bCGAOWwjwDlsKIA5bC4AOWwvwDlsaAA5bGiAOWxpADlsaUA5bGuAOWxsQDlso0A5bOAAOW0mQDltYMA5bWQAOW1qwDlta4A5bW8AOW2sgDltroA5bebAOW3oQDlt6IA5belAOW3pgDlt7EA5be9AOW3vgDluKgA5bi9AOW5qQDlubIA5bmz5oiQAOW5tADluboA5bm8AOW5vwDluqYA5bqwAOW6swDlurYA5buJAOW7igDlu5IA5buTAOW7mQDlu6wA5bu0AOW7vgDlvIQA5byLAOW8kwDlvKIA5b2QAOW9kwDlvaEA5b2iAOW9qQDlvasA5b2zAOW+iwDlvowA5b6XAOW+mgDlvqkA5b6tAOW/gwDlv40A5b+XAOW/tQDlv7kA5oCSAOaAnADmgbUA5oKBAOaClADmg4cA5oOYAOaDoQDmhIgA5oWEAOaFiADmhYwA5oWOAOaFoADmhagA5oW6AOaGjgDmhpAA5oakAOaGrwDmhrIA5oeeAOaHsgDmh7YA5oiAAOaIiADmiJAA5oibAOaIrgDmiLQA5oi2AOaJiwDmiZMA5omdAOaKlQDmirEA5ouJAOaLjwDmi5MA5ouUAOaLvADmi74A5oyHAOaMvQDmjZAA5o2VAOaNqADmjbsA5o6DAOaOoADmjqkA5o+EAOaPhQDmj6QA5pCcAOaQogDmkZIA5pGpAOaRtwDmkb4A5pKaAOaSnQDmk4QA5pSvAOaUtADmlY8A5pWWAOaVrADmlbgA5paHAOaWlwDmlpkA5pakAOaWsADmlrkA5peFAOaXoADml6IA5pejAOaXpQDmmI7msrsA5piTAOaYoADmmK3lkowA5pmJAOaZtADmmogA5pqRAOaanADmmrQA5puGAOabsADmm7QA5pu4AOacgADmnIgA5pyJAOaclwDmnJsA5pyhAOacqADmnY4A5p2TAOadlgDmnZ4A5p27AOaehQDmnpcA5p+zAOafugDmoJcA5qCfAOagqgDmoKrlvI/kvJrnpL4A5qGSAOaigQDmooUA5qKOAOaiqADmpJQA5qWCAOamowDmp6oA5qiCAOaokwDmqqgA5quTAOarmwDmrIQA5qygAOasoQDmrZQA5q2iAOatowDmrbIA5q23AOatuQDmrp8A5q6uAOauswDmrroA5q67AOaviwDmr40A5q+UAOavmwDmsI8A5rCUAOawtADmsY4A5rGnAOayiADmsr8A5rOMAOazjQDms6UA5rOoAOa0lgDmtJsA5rSeAOa0tADmtL4A5rWBAOa1qQDmtaoA5rW3AOa1uADmtoUA5reLAOa3mgDmt6oA5re5AOa4mgDmuK8A5rmuAOa6gADmupwA5rq6AOa7hwDmu4sA5ruRAOa7mwDmvI8A5ryUAOa8ogDmvKMA5r2uAOa/hgDmv6sA5r++AOeAmwDngJ4A54C5AOeBigDngasA54GwAOeBtwDngb0A54KZAOeCrQDng4gA54OZAOeEoQDnhYUA54WJAOeFrgDnhpwA54eOAOeHkADniJAA54ibAOeIqADniKoA54irAOeItQDniLYA54i7AOeIvwDniYcA54mQAOeJmQDniZsA54miAOeJuQDnioAA54qVAOeKrADniq8A54uAAOeLvADnjKoA5421AOeNugDnjoQA546HAOeOiQDnjosA546lAOeOsgDnj54A55CGAOeQiQDnkKIA55GHAOeRnADnkakA55GxAOeShQDnkokA55KYAOeTigDnk5wA55OmAOeUhgDnlJgA55SfAOeUpADnlKgA55SwAOeUsgDnlLMA55S3AOeUuwDnlL4A55WZAOeVpQDnlbAA55aLAOeWkgDnl6IA55iQAOeYnQDnmJ8A55mCAOeZqQDnmbYA55m9AOeargDnmr8A55uKAOebmwDnm6MA55unAOebrgDnm7QA55yBAOecngDnnJ8A552AAOedigDnnosA556nAOefmwDnn6IA55+zAOehjgDnoasA56KMAOeikQDno4oA56OMAOejuwDnpKoA56S6AOekvADnpL4A56WIAOeliQDnpZAA56WWAOelnQDnpZ4A56WlAOelvwDnpoEA56aNAOemjgDnpo8A56auAOemuADnpr4A56eKAOenmADnp6sA56icAOepgADnqYoA56mPAOeptADnqboA56qBAOeqsQDnq4sA56uuAOeruQDnrKAA566PAOevgADnr4YA56+JAOewvgDnsaAA57GzAOexuwDnspIA57K+AOezkgDns5YA57OjAOezpwDns6gA57O4AOe0gADntJAA57SiAOe0rwDntYIA57WbAOe1owDntqAA57a+AOe3hwDnt7QA57iCAOe4iQDnuLcA57mBAOe5hQDnvLYA57y+AOe9kQDnvbIA5725AOe9ugDnvoUA576KAOe+lQDnvpoA5769AOe/ugDogIEA6ICFAOiAjADogJIA6ICzAOiBhgDogaAA6IGvAOiBsADogb4A6IG/AOiCiQDogosA6IKtAOiCsgDohIMA6IS+AOiHmADoh6MA6IeoAOiHqgDoh60A6IezAOiHvADoiIEA6IiEAOiIjADoiJgA6IibAOiInwDoia4A6ImvAOiJsgDoibgA6Im5AOiKiwDoipEA6IqdAOiKsQDoirMA6Iq9AOiLpQDoi6YA6IydAOiMowDojLYA6I2SAOiNkwDojaMA6I6tAOiOvQDoj4kA6I+KAOiPjADoj5wA6I+nAOiPrwDoj7EA6JC9AOiRiQDokZcA6JOuAOiTsQDok7MA6JO8AOiUlgDolaQA6JeNAOiXugDomIYA6JiSAOiYrQDomL8A6JmNAOiZkADomZwA6JmnAOiZqQDomasA6JqIAOiaqQDom6IA6JyOAOicqADonasA6J25AOiehgDonroA6J+hAOiggQDooJ8A6KGAAOihjADooaAA6KGjAOijggDoo48A6KOXAOijngDoo6EA6KO4AOijugDopJAA6KWBAOilpADopb4A6KaGAOimiwDoppYA6KeSAOinowDoqIAA6KqgAOiqqgDoqr8A6KuLAOirkgDoq5YA6KutAOiruADoq74A6KyBAOisuQDorZgA6K6AAOiuigDosLcA6LGGAOixiADosZUA6LG4AOiynQDosqEA6LKpAOiyqwDos4EA6LOCAOizhwDos4gA6LOTAOi0iADotJsA6LWkAOi1sADotbcA6LazAOi2vADot4sA6LevAOi3sADouqsA6LuKAOi7lADovKYA6LyqAOi8uADovLsA6L2iAOi+mwDovp4A6L6wAOi+tQDovrYA6YCjAOmAuADpgYoA6YGpAOmBsgDpgbwA6YKPAOmCkQDpgpQA6YOOAOmDngDpg7EA6YO9AOmEkQDphJsA6YWJAOmFjQDphaoA6YaZAOmGtADph4YA6YeMAOmHjwDph5EA6Yi0AOmIuADpibYA6Ym8AOmLlwDpi5gA6YyEAOmNigDpj7kA6ZCVAOmVtwDploAA6ZaLAOmWrQDplrcA6ZicAOmYrgDpmYsA6ZmNAOmZtQDpmbgA6Zm8AOmahgDpmqMA6Zq2AOmatwDpmrgA6Zq5AOmbgwDpm6IA6ZujAOmbqADpm7YA6Zu3AOmcowDpnLIA6Z2IAOmdkQDpnZYA6Z2eAOmdogDpnakA6Z+LAOmfmwDpn6AA6Z+tAOmfswDpn78A6aCBAOmghQDpoIsA6aCYAOmgqQDpoLsA6aGeAOmiqADpo5sA6aOfAOmjogDpo68A6aO8AOmkqADppKkA6aaWAOmmmQDppqcA6aasAOmnggDpp7EA6ae+AOmpqgDpqqgA6auYAOmrnwDprJIA6aylAOmsrwDprLIA6ay8AOmtmgDpra8A6bGAAOmxlwDps6UA6bO9AOm1pwDptrQA6be6AOm4ngDpubUA6bm/AOm6lwDpup8A6bqlAOm6uwDpu4MA6buNAOm7jgDpu5EA6bu5AOm7vQDpu74A6byFAOm8jgDpvI8A6byTAOm8lgDpvKAA6by7AOm9gwDpvYoA6b2SAOm+jQDpvo4A6b6cAOm+nwDpvqAA6pynAOqdrwDqrLcA6q2SAOqwgADqsIEA6rCCAOqwgwDqsIQA6rCFAOqwhgDqsIcA6rCIAOqwiQDqsIoA6rCLAOqwjADqsI0A6rCOAOqwjwDqsJAA6rCRAOqwkgDqsJMA6rCUAOqwlQDqsJYA6rCXAOqwmADqsJkA6rCaAOqwmwDqsJwA6rCdAOqwngDqsJ8A6rCgAOqwoQDqsKIA6rCjAOqwpADqsKUA6rCmAOqwpwDqsKgA6rCpAOqwqgDqsKsA6rCsAOqwrQDqsK4A6rCvAOqwsADqsLEA6rCyAOqwswDqsLQA6rC1AOqwtgDqsLcA6rC4AOqwuQDqsLoA6rC7AOqwvADqsL0A6rC+AOqwvwDqsYAA6rGBAOqxggDqsYMA6rGEAOqxhQDqsYYA6rGHAOqxiADqsYkA6rGKAOqxiwDqsYwA6rGNAOqxjgDqsY8A6rGQAOqxkQDqsZIA6rGTAOqxlADqsZUA6rGWAOqxlwDqsZgA6rGZAOqxmgDqsZsA6rGcAOqxnQDqsZ4A6rGfAOqxoADqsaEA6rGiAOqxowDqsaQA6rGlAOqxpgDqsacA6rGoAOqxqQDqsaoA6rGrAOqxrADqsa0A6rGuAOqxrwDqsbAA6rGxAOqxsgDqsbMA6rG0AOqxtQDqsbYA6rG3AOqxuADqsbkA6rG6AOqxuwDqsbwA6rG9AOqxvgDqsb8A6rKAAOqygQDqsoIA6rKDAOqyhADqsoUA6rKGAOqyhwDqsogA6rKJAOqyigDqsosA6rKMAOqyjQDqso4A6rKPAOqykADqspEA6rKSAOqykwDqspQA6rKVAOqylgDqspcA6rKYAOqymQDqspoA6rKbAOqynADqsp0A6rKeAOqynwDqsqAA6rKhAOqyogDqsqMA6rKkAOqypQDqsqYA6rKnAOqyqADqsqkA6rKqAOqyqwDqsqwA6rKtAOqyrgDqsq8A6rKwAOqysQDqsrIA6rKzAOqytADqsrUA6rK2AOqytwDqsrgA6rK5AOqyugDqsrsA6rK8AOqyvQDqsr4A6rK/AOqzgADqs4EA6rOCAOqzgwDqs4QA6rOFAOqzhgDqs4cA6rOIAOqziQDqs4oA6rOLAOqzjADqs40A6rOOAOqzjwDqs5AA6rORAOqzkgDqs5MA6rOUAOqzlQDqs5YA6rOXAOqzmADqs5kA6rOaAOqzmwDqs5wA6rOdAOqzngDqs58A6rOgAOqzoQDqs6IA6rOjAOqzpADqs6UA6rOmAOqzpwDqs6gA6rOpAOqzqgDqs6sA6rOsAOqzrQDqs64A6rOvAOqzsADqs7EA6rOyAOqzswDqs7QA6rO1AOqztgDqs7cA6rO4AOqzuQDqs7oA6rO7AOqzvADqs70A6rO+AOqzvwDqtIAA6rSBAOq0ggDqtIMA6rSEAOq0hQDqtIYA6rSHAOq0iADqtIkA6rSKAOq0iwDqtIwA6rSNAOq0jgDqtI8A6rSQAOq0kQDqtJIA6rSTAOq0lADqtJUA6rSWAOq0lwDqtJgA6rSZAOq0mgDqtJsA6rScAOq0nQDqtJ4A6rSfAOq0oADqtKEA6rSiAOq0owDqtKQA6rSlAOq0pgDqtKcA6rSoAOq0qQDqtKoA6rSrAOq0rADqtK0A6rSuAOq0rwDqtLAA6rSxAOq0sgDqtLMA6rS0AOq0tQDqtLYA6rS3AOq0uADqtLkA6rS6AOq0uwDqtLwA6rS9AOq0vgDqtL8A6rWAAOq1gQDqtYIA6rWDAOq1hADqtYUA6rWGAOq1hwDqtYgA6rWJAOq1igDqtYsA6rWMAOq1jQDqtY4A6rWPAOq1kADqtZEA6rWSAOq1kwDqtZQA6rWVAOq1lgDqtZcA6rWYAOq1mQDqtZoA6rWbAOq1nADqtZ0A6rWeAOq1nwDqtaAA6rWhAOq1ogDqtaMA6rWkAOq1pQDqtaYA6rWnAOq1qADqtakA6rWqAOq1qwDqtawA6rWtAOq1rgDqta8A6rWwAOq1sQDqtbIA6rWzAOq1tADqtbUA6rW2AOq1twDqtbgA6rW5AOq1ugDqtbsA6rW8AOq1vQDqtb4A6rW/AOq2gADqtoEA6raCAOq2gwDqtoQA6raFAOq2hgDqtocA6raIAOq2iQDqtooA6raLAOq2jADqto0A6raOAOq2jwDqtpAA6raRAOq2kgDqtpMA6raUAOq2lQDqtpYA6raXAOq2mADqtpkA6raaAOq2mwDqtpwA6radAOq2ngDqtp8A6ragAOq2oQDqtqIA6rajAOq2pADqtqUA6ramAOq2pwDqtqgA6rapAOq2qgDqtqsA6rasAOq2rQDqtq4A6ravAOq2sADqtrEA6rayAOq2swDqtrQA6ra1AOq2tgDqtrcA6ra4AOq2uQDqtroA6ra7AOq2vADqtr0A6ra+AOq2vwDqt4AA6reBAOq3ggDqt4MA6reEAOq3hQDqt4YA6reHAOq3iADqt4kA6reKAOq3iwDqt4wA6reNAOq3jgDqt48A6reQAOq3kQDqt5IA6reTAOq3lADqt5UA6reWAOq3lwDqt5gA6reZAOq3mgDqt5sA6recAOq3nQDqt54A6refAOq3oADqt6EA6reiAOq3owDqt6QA6relAOq3pgDqt6cA6reoAOq3qQDqt6oA6rerAOq3rADqt60A6reuAOq3rwDqt7AA6rexAOq3sgDqt7MA6re0AOq3tQDqt7YA6re3AOq3uADqt7kA6re6AOq3uwDqt7wA6re9AOq3vgDqt78A6riAAOq4gQDquIIA6riDAOq4hADquIUA6riGAOq4hwDquIgA6riJAOq4igDquIsA6riMAOq4jQDquI4A6riPAOq4kADquJEA6riSAOq4kwDquJQA6riVAOq4lgDquJcA6riYAOq4mQDquJoA6ribAOq4nADquJ0A6rieAOq4nwDquKAA6rihAOq4ogDquKMA6rikAOq4pQDquKYA6rinAOq4qADquKkA6riqAOq4qwDquKwA6ritAOq4rgDquK8A6riwAOq4sQDquLIA6rizAOq4tADquLUA6ri2AOq4twDquLgA6ri5AOq4ugDquLsA6ri8AOq4vQDquL4A6ri/AOq5gADquYEA6rmCAOq5gwDquYQA6rmFAOq5hgDquYcA6rmIAOq5iQDquYoA6rmLAOq5jADquY0A6rmOAOq5jwDquZAA6rmRAOq5kgDquZMA6rmUAOq5lQDquZYA6rmXAOq5mADquZkA6rmaAOq5mwDquZwA6rmdAOq5ngDquZ8A6rmgAOq5oQDquaIA6rmjAOq5pADquaUA6rmmAOq5pwDquagA6rmpAOq5qgDquasA6rmsAOq5rQDqua4A6rmvAOq5sADqubEA6rmyAOq5swDqubQA6rm1AOq5tgDqubcA6rm4AOq5uQDquboA6rm7AOq5vADqub0A6rm+AOq5vwDquoAA6rqBAOq6ggDquoMA6rqEAOq6hQDquoYA6rqHAOq6iADquokA6rqKAOq6iwDquowA6rqNAOq6jgDquo8A6rqQAOq6kQDqupIA6rqTAOq6lADqupUA6rqWAOq6lwDqupgA6rqZAOq6mgDqupsA6rqcAOq6nQDqup4A6rqfAOq6oADquqEA6rqiAOq6owDquqQA6rqlAOq6pgDquqcA6rqoAOq6qQDquqoA6rqrAOq6rADquq0A6rquAOq6rwDqurAA6rqxAOq6sgDqurMA6rq0AOq6tQDqurYA6rq3AOq6uADqurkA6rq6AOq6uwDqurwA6rq9AOq6vgDqur8A6ruAAOq7gQDqu4IA6ruDAOq7hADqu4UA6ruGAOq7hwDqu4gA6ruJAOq7igDqu4sA6ruMAOq7jQDqu44A6ruPAOq7kADqu5EA6ruSAOq7kwDqu5QA6ruVAOq7lgDqu5cA6ruYAOq7mQDqu5oA6rubAOq7nADqu50A6rueAOq7nwDqu6AA6ruhAOq7ogDqu6MA6rukAOq7pQDqu6YA6runAOq7qADqu6kA6ruqAOq7qwDqu6wA6rutAOq7rgDqu68A6ruwAOq7sQDqu7IA6ruzAOq7tADqu7UA6ru2AOq7twDqu7gA6ru5AOq7ugDqu7sA6ru8AOq7vQDqu74A6ru/AOq8gADqvIEA6ryCAOq8gwDqvIQA6ryFAOq8hgDqvIcA6ryIAOq8iQDqvIoA6ryLAOq8jADqvI0A6ryOAOq8jwDqvJAA6ryRAOq8kgDqvJMA6ryUAOq8lQDqvJYA6ryXAOq8mADqvJkA6ryaAOq8mwDqvJwA6rydAOq8ngDqvJ8A6rygAOq8oQDqvKIA6ryjAOq8pADqvKUA6rymAOq8pwDqvKgA6rypAOq8qgDqvKsA6rysAOq8rQDqvK4A6ryvAOq8sADqvLEA6ryyAOq8swDqvLQA6ry1AOq8tgDqvLcA6ry4AOq8uQDqvLoA6ry7AOq8vADqvL0A6ry+AOq8vwDqvYAA6r2BAOq9ggDqvYMA6r2EAOq9hQDqvYYA6r2HAOq9iADqvYkA6r2KAOq9iwDqvYwA6r2NAOq9jgDqvY8A6r2QAOq9kQDqvZIA6r2TAOq9lADqvZUA6r2WAOq9lwDqvZgA6r2ZAOq9mgDqvZsA6r2cAOq9nQDqvZ4A6r2fAOq9oADqvaEA6r2iAOq9owDqvaQA6r2lAOq9pgDqvacA6r2oAOq9qQDqvaoA6r2rAOq9rADqva0A6r2uAOq9rwDqvbAA6r2xAOq9sgDqvbMA6r20AOq9tQDqvbYA6r23AOq9uADqvbkA6r26AOq9uwDqvbwA6r29AOq9vgDqvb8A6r6AAOq+gQDqvoIA6r6DAOq+hADqvoUA6r6GAOq+hwDqvogA6r6JAOq+igDqvosA6r6MAOq+jQDqvo4A6r6PAOq+kADqvpEA6r6SAOq+kwDqvpQA6r6VAOq+lgDqvpcA6r6YAOq+mQDqvpoA6r6bAOq+nADqvp0A6r6eAOq+nwDqvqAA6r6hAOq+ogDqvqMA6r6kAOq+pQDqvqYA6r6nAOq+qADqvqkA6r6qAOq+qwDqvqwA6r6tAOq+rgDqvq8A6r6wAOq+sQDqvrIA6r6zAOq+tADqvrUA6r62AOq+twDqvrgA6r65AOq+ugDqvrsA6r68AOq+vQDqvr4A6r6/AOq/gADqv4EA6r+CAOq/gwDqv4QA6r+FAOq/hgDqv4cA6r+IAOq/iQDqv4oA6r+LAOq/jADqv40A6r+OAOq/jwDqv5AA6r+RAOq/kgDqv5MA6r+UAOq/lQDqv5YA6r+XAOq/mADqv5kA6r+aAOq/mwDqv5wA6r+dAOq/ngDqv58A6r+gAOq/oQDqv6IA6r+jAOq/pADqv6UA6r+mAOq/pwDqv6gA6r+pAOq/qgDqv6sA6r+sAOq/rQDqv64A6r+vAOq/sADqv7EA6r+yAOq/swDqv7QA6r+1AOq/tgDqv7cA6r+4AOq/uQDqv7oA6r+7AOq/vADqv70A6r++AOq/vwDrgIAA64CBAOuAggDrgIMA64CEAOuAhQDrgIYA64CHAOuAiADrgIkA64CKAOuAiwDrgIwA64CNAOuAjgDrgI8A64CQAOuAkQDrgJIA64CTAOuAlADrgJUA64CWAOuAlwDrgJgA64CZAOuAmgDrgJsA64CcAOuAnQDrgJ4A64CfAOuAoADrgKEA64CiAOuAowDrgKQA64ClAOuApgDrgKcA64CoAOuAqQDrgKoA64CrAOuArADrgK0A64CuAOuArwDrgLAA64CxAOuAsgDrgLMA64C0AOuAtQDrgLYA64C3AOuAuADrgLkA64C6AOuAuwDrgLwA64C9AOuAvgDrgL8A64GAAOuBgQDrgYIA64GDAOuBhADrgYUA64GGAOuBhwDrgYgA64GJAOuBigDrgYsA64GMAOuBjQDrgY4A64GPAOuBkADrgZEA64GSAOuBkwDrgZQA64GVAOuBlgDrgZcA64GYAOuBmQDrgZoA64GbAOuBnADrgZ0A64GeAOuBnwDrgaAA64GhAOuBogDrgaMA64GkAOuBpQDrgaYA64GnAOuBqADrgakA64GqAOuBqwDrgawA64GtAOuBrgDrga8A64GwAOuBsQDrgbIA64GzAOuBtADrgbUA64G2AOuBtwDrgbgA64G5AOuBugDrgbsA64G8AOuBvQDrgb4A64G/AOuCgADrgoEA64KCAOuCgwDrgoQA64KFAOuChgDrgocA64KIAOuCiQDrgooA64KLAOuCjADrgo0A64KOAOuCjwDrgpAA64KRAOuCkgDrgpMA64KUAOuClQDrgpYA64KXAOuCmADrgpkA64KaAOuCmwDrgpwA64KdAOuCngDrgp8A64KgAOuCoQDrgqIA64KjAOuCpADrgqUA64KmAOuCpwDrgqgA64KpAOuCqgDrgqsA64KsAOuCrQDrgq4A64KvAOuCsADrgrEA64KyAOuCswDrgrQA64K1AOuCtgDrgrcA64K4AOuCuQDrgroA64K7AOuCvADrgr0A64K+AOuCvwDrg4AA64OBAOuDggDrg4MA64OEAOuDhQDrg4YA64OHAOuDiADrg4kA64OKAOuDiwDrg4wA64ONAOuDjgDrg48A64OQAOuDkQDrg5IA64OTAOuDlADrg5UA64OWAOuDlwDrg5gA64OZAOuDmgDrg5sA64OcAOuDnQDrg54A64OfAOuDoADrg6EA64OiAOuDowDrg6QA64OlAOuDpgDrg6cA64OoAOuDqQDrg6oA64OrAOuDrADrg60A64OuAOuDrwDrg7AA64OxAOuDsgDrg7MA64O0AOuDtQDrg7YA64O3AOuDuADrg7kA64O6AOuDuwDrg7wA64O9AOuDvgDrg78A64SAAOuEgQDrhIIA64SDAOuEhADrhIUA64SGAOuEhwDrhIgA64SJAOuEigDrhIsA64SMAOuEjQDrhI4A64SPAOuEkADrhJEA64SSAOuEkwDrhJQA64SVAOuElgDrhJcA64SYAOuEmQDrhJoA64SbAOuEnADrhJ0A64SeAOuEnwDrhKAA64ShAOuEogDrhKMA64SkAOuEpQDrhKYA64SnAOuEqADrhKkA64SqAOuEqwDrhKwA64StAOuErgDrhK8A64SwAOuEsQDrhLIA64SzAOuEtADrhLUA64S2AOuEtwDrhLgA64S5AOuEugDrhLsA64S8AOuEvQDrhL4A64S/AOuFgADrhYEA64WCAOuFgwDrhYQA64WFAOuFhgDrhYcA64WIAOuFiQDrhYoA64WLAOuFjADrhY0A64WOAOuFjwDrhZAA64WRAOuFkgDrhZMA64WUAOuFlQDrhZYA64WXAOuFmADrhZkA64WaAOuFmwDrhZwA64WdAOuFngDrhZ8A64WgAOuFoQDrhaIA64WjAOuFpADrhaUA64WmAOuFpwDrhagA64WpAOuFqgDrhasA64WsAOuFrQDrha4A64WvAOuFsADrhbEA64WyAOuFswDrhbQA64W1AOuFtgDrhbcA64W4AOuFuQDrhboA64W7AOuFvADrhb0A64W+AOuFvwDrhoAA64aBAOuGggDrhoMA64aEAOuGhQDrhoYA64aHAOuGiADrhokA64aKAOuGiwDrhowA64aNAOuGjgDrho8A64aQAOuGkQDrhpIA64aTAOuGlADrhpUA64aWAOuGlwDrhpgA64aZAOuGmgDrhpsA64acAOuGnQDrhp4A64afAOuGoADrhqEA64aiAOuGowDrhqQA64alAOuGpgDrhqcA64aoAOuGqQDrhqoA64arAOuGrADrhq0A64auAOuGrwDrhrAA64axAOuGsgDrhrMA64a0AOuGtQDrhrYA64a3AOuGuADrhrkA64a6AOuGuwDrhrwA64a9AOuGvgDrhr8A64eAAOuHgQDrh4IA64eDAOuHhADrh4UA64eGAOuHhwDrh4gA64eJAOuHigDrh4sA64eMAOuHjQDrh44A64ePAOuHkADrh5EA64eSAOuHkwDrh5QA64eVAOuHlgDrh5cA64eYAOuHmQDrh5oA64ebAOuHnADrh50A64eeAOuHnwDrh6AA64ehAOuHogDrh6MA64ekAOuHpQDrh6YA64enAOuHqADrh6kA64eqAOuHqwDrh6wA64etAOuHrgDrh68A64ewAOuHsQDrh7IA64ezAOuHtADrh7UA64e2AOuHtwDrh7gA64e5AOuHugDrh7sA64e8AOuHvQDrh74A64e/AOuIgADriIEA64iCAOuIgwDriIQA64iFAOuIhgDriIcA64iIAOuIiQDriIoA64iLAOuIjADriI0A64iOAOuIjwDriJAA64iRAOuIkgDriJMA64iUAOuIlQDriJYA64iXAOuImADriJkA64iaAOuImwDriJwA64idAOuIngDriJ8A64igAOuIoQDriKIA64ijAOuIpADriKUA64imAOuIpwDriKgA64ipAOuIqgDriKsA64isAOuIrQDriK4A64ivAOuIsADriLEA64iyAOuIswDriLQA64i1AOuItgDriLcA64i4AOuIuQDriLoA64i7AOuIvADriL0A64i+AOuIvwDriYAA64mBAOuJggDriYMA64mEAOuJhQDriYYA64mHAOuJiADriYkA64mKAOuJiwDriYwA64mNAOuJjgDriY8A64mQAOuJkQDriZIA64mTAOuJlADriZUA64mWAOuJlwDriZgA64mZAOuJmgDriZsA64mcAOuJnQDriZ4A64mfAOuJoADriaEA64miAOuJowDriaQA64mlAOuJpgDriacA64moAOuJqQDriaoA64mrAOuJrADria0A64muAOuJrwDribAA64mxAOuJsgDribMA64m0AOuJtQDribYA64m3AOuJuADribkA64m6AOuJuwDribwA64m9AOuJvgDrib8A64qAAOuKgQDrioIA64qDAOuKhADrioUA64qGAOuKhwDriogA64qJAOuKigDriosA64qMAOuKjQDrio4A64qPAOuKkADripEA64qSAOuKkwDripQA64qVAOuKlgDripcA64qYAOuKmQDripoA64qbAOuKnADrip0A64qeAOuKnwDriqAA64qhAOuKogDriqMA64qkAOuKpQDriqYA64qnAOuKqADriqkA64qqAOuKqwDriqwA64qtAOuKrgDriq8A64qwAOuKsQDrirIA64qzAOuKtADrirUA64q2AOuKtwDrirgA64q5AOuKugDrirsA64q8AOuKvQDrir4A64q/AOuLgADri4EA64uCAOuLgwDri4QA64uFAOuLhgDri4cA64uIAOuLiQDri4oA64uLAOuLjADri40A64uOAOuLjwDri5AA64uRAOuLkgDri5MA64uUAOuLlQDri5YA64uXAOuLmADri5kA64uaAOuLmwDri5wA64udAOuLngDri58A64ugAOuLoQDri6IA64ujAOuLpADri6UA64umAOuLpwDri6gA64upAOuLqgDri6sA64usAOuLrQDri64A64uvAOuLsADri7EA64uyAOuLswDri7QA64u1AOuLtgDri7cA64u4AOuLuQDri7oA64u7AOuLvADri70A64u+AOuLvwDrjIAA64yBAOuMggDrjIMA64yEAOuMhQDrjIYA64yHAOuMiADrjIkA64yKAOuMiwDrjIwA64yNAOuMjgDrjI8A64yQAOuMkQDrjJIA64yTAOuMlADrjJUA64yWAOuMlwDrjJgA64yZAOuMmgDrjJsA64ycAOuMnQDrjJ4A64yfAOuMoADrjKEA64yiAOuMowDrjKQA64ylAOuMpgDrjKcA64yoAOuMqQDrjKoA64yrAOuMrADrjK0A64yuAOuMrwDrjLAA64yxAOuMsgDrjLMA64y0AOuMtQDrjLYA64y3AOuMuADrjLkA64y6AOuMuwDrjLwA64y9AOuMvgDrjL8A642AAOuNgQDrjYIA642DAOuNhADrjYUA642GAOuNhwDrjYgA642JAOuNigDrjYsA642MAOuNjQDrjY4A642PAOuNkADrjZEA642SAOuNkwDrjZQA642VAOuNlgDrjZcA642YAOuNmQDrjZoA642bAOuNnADrjZ0A642eAOuNnwDrjaAA642hAOuNogDrjaMA642kAOuNpQDrjaYA642nAOuNqADrjakA642qAOuNqwDrjawA642tAOuNrgDrja8A642wAOuNsQDrjbIA642zAOuNtADrjbUA6422AOuNtwDrjbgA6425AOuNugDrjbsA6428AOuNvQDrjb4A642/AOuOgADrjoEA646CAOuOgwDrjoQA646FAOuOhgDrjocA646IAOuOiQDrjooA646LAOuOjADrjo0A646OAOuOjwDrjpAA646RAOuOkgDrjpMA646UAOuOlQDrjpYA646XAOuOmADrjpkA646aAOuOmwDrjpwA646dAOuOngDrjp8A646gAOuOoQDrjqIA646jAOuOpADrjqUA646mAOuOpwDrjqgA646pAOuOqgDrjqsA646sAOuOrQDrjq4A646vAOuOsADrjrEA646yAOuOswDrjrQA6461AOuOtgDrjrcA6464AOuOuQDrjroA6467AOuOvADrjr0A646+AOuOvwDrj4AA64+BAOuPggDrj4MA64+EAOuPhQDrj4YA64+HAOuPiADrj4kA64+KAOuPiwDrj4wA64+NAOuPjgDrj48A64+QAOuPkQDrj5IA64+TAOuPlADrj5UA64+WAOuPlwDrj5gA64+ZAOuPmgDrj5sA64+cAOuPnQDrj54A64+fAOuPoADrj6EA64+iAOuPowDrj6QA64+lAOuPpgDrj6cA64+oAOuPqQDrj6oA64+rAOuPrADrj60A64+uAOuPrwDrj7AA64+xAOuPsgDrj7MA64+0AOuPtQDrj7YA64+3AOuPuADrj7kA64+6AOuPuwDrj7wA64+9AOuPvgDrj78A65CAAOuQgQDrkIIA65CDAOuQhADrkIUA65CGAOuQhwDrkIgA65CJAOuQigDrkIsA65CMAOuQjQDrkI4A65CPAOuQkADrkJEA65CSAOuQkwDrkJQA65CVAOuQlgDrkJcA65CYAOuQmQDrkJoA65CbAOuQnADrkJ0A65CeAOuQnwDrkKAA65ChAOuQogDrkKMA65CkAOuQpQDrkKYA65CnAOuQqADrkKkA65CqAOuQqwDrkKwA65CtAOuQrgDrkK8A65CwAOuQsQDrkLIA65CzAOuQtADrkLUA65C2AOuQtwDrkLgA65C5AOuQugDrkLsA65C8AOuQvQDrkL4A65C/AOuRgADrkYEA65GCAOuRgwDrkYQA65GFAOuRhgDrkYcA65GIAOuRiQDrkYoA65GLAOuRjADrkY0A65GOAOuRjwDrkZAA65GRAOuRkgDrkZMA65GUAOuRlQDrkZYA65GXAOuRmADrkZkA65GaAOuRmwDrkZwA65GdAOuRngDrkZ8A65GgAOuRoQDrkaIA65GjAOuRpADrkaUA65GmAOuRpwDrkagA65GpAOuRqgDrkasA65GsAOuRrQDrka4A65GvAOuRsADrkbEA65GyAOuRswDrkbQA65G1AOuRtgDrkbcA65G4AOuRuQDrkboA65G7AOuRvADrkb0A65G+AOuRvwDrkoAA65KBAOuSggDrkoMA65KEAOuShQDrkoYA65KHAOuSiADrkokA65KKAOuSiwDrkowA65KNAOuSjgDrko8A65KQAOuSkQDrkpIA65KTAOuSlADrkpUA65KWAOuSlwDrkpgA65KZAOuSmgDrkpsA65KcAOuSnQDrkp4A65KfAOuSoADrkqEA65KiAOuSowDrkqQA65KlAOuSpgDrkqcA65KoAOuSqQDrkqoA65KrAOuSrADrkq0A65KuAOuSrwDrkrAA65KxAOuSsgDrkrMA65K0AOuStQDrkrYA65K3AOuSuADrkrkA65K6AOuSuwDrkrwA65K9AOuSvgDrkr8A65OAAOuTgQDrk4IA65ODAOuThADrk4UA65OGAOuThwDrk4gA65OJAOuTigDrk4sA65OMAOuTjQDrk44A65OPAOuTkADrk5EA65OSAOuTkwDrk5QA65OVAOuTlgDrk5cA65OYAOuTmQDrk5oA65ObAOuTnADrk50A65OeAOuTnwDrk6AA65OhAOuTogDrk6MA65OkAOuTpQDrk6YA65OnAOuTqADrk6kA65OqAOuTqwDrk6wA65OtAOuTrgDrk68A65OwAOuTsQDrk7IA65OzAOuTtADrk7UA65O2AOuTtwDrk7gA65O5AOuTugDrk7sA65O8AOuTvQDrk74A65O/AOuUgADrlIEA65SCAOuUgwDrlIQA65SFAOuUhgDrlIcA65SIAOuUiQDrlIoA65SLAOuUjADrlI0A65SOAOuUjwDrlJAA65SRAOuUkgDrlJMA65SUAOuUlQDrlJYA65SXAOuUmADrlJkA65SaAOuUmwDrlJwA65SdAOuUngDrlJ8A65SgAOuUoQDrlKIA65SjAOuUpADrlKUA65SmAOuUpwDrlKgA65SpAOuUqgDrlKsA65SsAOuUrQDrlK4A65SvAOuUsADrlLEA65SyAOuUswDrlLQA65S1AOuUtgDrlLcA65S4AOuUuQDrlLoA65S7AOuUvADrlL0A65S+AOuUvwDrlYAA65WBAOuVggDrlYMA65WEAOuVhQDrlYYA65WHAOuViADrlYkA65WKAOuViwDrlYwA65WNAOuVjgDrlY8A65WQAOuVkQDrlZIA65WTAOuVlADrlZUA65WWAOuVlwDrlZgA65WZAOuVmgDrlZsA65WcAOuVnQDrlZ4A65WfAOuVoADrlaEA65WiAOuVowDrlaQA65WlAOuVpgDrlacA65WoAOuVqQDrlaoA65WrAOuVrADrla0A65WuAOuVrwDrlbAA65WxAOuVsgDrlbMA65W0AOuVtQDrlbYA65W3AOuVuADrlbkA65W6AOuVuwDrlbwA65W9AOuVvgDrlb8A65aAAOuWgQDrloIA65aDAOuWhADrloUA65aGAOuWhwDrlogA65aJAOuWigDrlosA65aMAOuWjQDrlo4A65aPAOuWkADrlpEA65aSAOuWkwDrlpQA65aVAOuWlgDrlpcA65aYAOuWmQDrlpoA65abAOuWnADrlp0A65aeAOuWnwDrlqAA65ahAOuWogDrlqMA65akAOuWpQDrlqYA65anAOuWqADrlqkA65aqAOuWqwDrlqwA65atAOuWrgDrlq8A65awAOuWsQDrlrIA65azAOuWtADrlrUA65a2AOuWtwDrlrgA65a5AOuWugDrlrsA65a8AOuWvQDrlr4A65a/AOuXgADrl4EA65eCAOuXgwDrl4QA65eFAOuXhgDrl4cA65eIAOuXiQDrl4oA65eLAOuXjADrl40A65eOAOuXjwDrl5AA65eRAOuXkgDrl5MA65eUAOuXlQDrl5YA65eXAOuXmADrl5kA65eaAOuXmwDrl5wA65edAOuXngDrl58A65egAOuXoQDrl6IA65ejAOuXpADrl6UA65emAOuXpwDrl6gA65epAOuXqgDrl6sA65esAOuXrQDrl64A65evAOuXsADrl7EA65eyAOuXswDrl7QA65e1AOuXtgDrl7cA65e4AOuXuQDrl7oA65e7AOuXvADrl70A65e+AOuXvwDrmIAA65iBAOuYggDrmIMA65iEAOuYhQDrmIYA65iHAOuYiADrmIkA65iKAOuYiwDrmIwA65iNAOuYjgDrmI8A65iQAOuYkQDrmJIA65iTAOuYlADrmJUA65iWAOuYlwDrmJgA65iZAOuYmgDrmJsA65icAOuYnQDrmJ4A65ifAOuYoADrmKEA65iiAOuYowDrmKQA65ilAOuYpgDrmKcA65ioAOuYqQDrmKoA65irAOuYrADrmK0A65iuAOuYrwDrmLAA65ixAOuYsgDrmLMA65i0AOuYtQDrmLYA65i3AOuYuADrmLkA65i6AOuYuwDrmLwA65i9AOuYvgDrmL8A65mAAOuZgQDrmYIA65mDAOuZhADrmYUA65mGAOuZhwDrmYgA65mJAOuZigDrmYsA65mMAOuZjQDrmY4A65mPAOuZkADrmZEA65mSAOuZkwDrmZQA65mVAOuZlgDrmZcA65mYAOuZmQDrmZoA65mbAOuZnADrmZ0A65meAOuZnwDrmaAA65mhAOuZogDrmaMA65mkAOuZpQDrmaYA65mnAOuZqADrmakA65mqAOuZqwDrmawA65mtAOuZrgDrma8A65mwAOuZsQDrmbIA65mzAOuZtADrmbUA65m2AOuZtwDrmbgA65m5AOuZugDrmbsA65m8AOuZvQDrmb4A65m/AOuagADrmoEA65qCAOuagwDrmoQA65qFAOuahgDrmocA65qIAOuaiQDrmooA65qLAOuajADrmo0A65qOAOuajwDrmpAA65qRAOuakgDrmpMA65qUAOualQDrmpYA65qXAOuamADrmpkA65qaAOuamwDrmpwA65qdAOuangDrmp8A65qgAOuaoQDrmqIA65qjAOuapADrmqUA65qmAOuapwDrmqgA65qpAOuaqgDrmqsA65qsAOuarQDrmq4A65qvAOuasADrmrEA65qyAOuaswDrmrQA65q1AOuatgDrmrcA65q4AOuauQDrmroA65q7AOuavADrmr0A65q+AOuavwDrm4AA65uBAOubggDrm4MA65uEAOubhQDrm4YA65uHAOubiADrm4kA65uKAOubiwDrm4wA65uNAOubjgDrm48A65uQAOubkQDrm5IA65uTAOublADrm5UA65uWAOublwDrm5gA65uZAOubmgDrm5sA65ucAOubnQDrm54A65ufAOuboADrm6EA65uiAOubowDrm6QA65ulAOubpgDrm6cA65uoAOubqQDrm6oA65urAOubrADrm60A65uuAOubrwDrm7AA65uxAOubsgDrm7MA65u0AOubtQDrm7YA65u3AOubuADrm7kA65u6AOubuwDrm7wA65u9AOubvgDrm78A65yAAOucgQDrnIIA65yDAOuchADrnIUA65yGAOuchwDrnIgA65yJAOucigDrnIsA65yMAOucjQDrnI4A65yPAOuckADrnJEA65ySAOuckwDrnJQA65yVAOuclgDrnJcA65yYAOucmQDrnJoA65ybAOucnADrnJ0A65yeAOucnwDrnKAA65yhAOucogDrnKMA65ykAOucpQDrnKYA65ynAOucqADrnKkA65yqAOucqwDrnKwA65ytAOucrgDrnK8A65ywAOucsQDrnLIA65yzAOuctADrnLUA65y2AOuctwDrnLgA65y5AOucugDrnLsA65y8AOucvQDrnL4A65y/AOudgADrnYEA652CAOudgwDrnYQA652FAOudhgDrnYcA652IAOudiQDrnYoA652LAOudjADrnY0A652OAOudjwDrnZAA652RAOudkgDrnZMA652UAOudlQDrnZYA652XAOudmADrnZkA652aAOudmwDrnZwA652dAOudngDrnZ8A652gAOudoQDrnaIA652jAOudpADrnaUA652mAOudpwDrnagA652pAOudqgDrnasA652sAOudrQDrna4A652vAOudsADrnbEA652yAOudswDrnbQA6521AOudtgDrnbcA6524AOuduQDrnboA6527AOudvADrnb0A652+AOudvwDrnoAA656BAOueggDrnoMA656EAOuehQDrnoYA656HAOueiADrnokA656KAOueiwDrnowA656NAOuejgDrno8A656QAOuekQDrnpIA656TAOuelADrnpUA656WAOuelwDrnpgA656ZAOuemgDrnpsA656cAOuenQDrnp4A656fAOueoADrnqEA656iAOueowDrnqQA656lAOuepgDrnqcA656oAOueqQDrnqoA656rAOuerADrnq0A656uAOuerwDrnrAA656xAOuesgDrnrMA6560AOuetQDrnrYA6563AOueuADrnrkA6566AOueuwDrnrwA6569AOuevgDrnr8A65+AAOufgQDrn4IA65+DAOufhADrn4UA65+GAOufhwDrn4gA65+JAOufigDrn4sA65+MAOufjQDrn44A65+PAOufkADrn5EA65+SAOufkwDrn5QA65+VAOuflgDrn5cA65+YAOufmQDrn5oA65+bAOufnADrn50A65+eAOufnwDrn6AA65+hAOufogDrn6MA65+kAOufpQDrn6YA65+nAOufqADrn6kA65+qAOufqwDrn6wA65+tAOufrgDrn68A65+wAOufsQDrn7IA65+zAOuftADrn7UA65+2AOuftwDrn7gA65+5AOufugDrn7sA65+8AOufvQDrn74A65+/AOuggADroIEA66CCAOuggwDroIQA66CFAOughgDroIcA66CIAOugiQDroIoA66CLAOugjADroI0A66COAOugjwDroJAA66CRAOugkgDroJMA66CUAOuglQDroJYA66CXAOugmADroJkA66CaAOugmwDroJwA66CdAOugngDroJ8A66CgAOugoQDroKIA66CjAOugpADroKUA66CmAOugpwDroKgA66CpAOugqgDroKsA66CsAOugrQDroK4A66CvAOugsADroLEA66CyAOugswDroLQA66C1AOugtgDroLcA66C4AOuguQDroLoA66C7AOugvADroL0A66C+AOugvwDroYAA66GBAOuhggDroYMA66GEAOuhhQDroYYA66GHAOuhiADroYkA66GKAOuhiwDroYwA66GNAOuhjgDroY8A66GQAOuhkQDroZIA66GTAOuhlADroZUA66GWAOuhlwDroZgA66GZAOuhmgDroZsA66GcAOuhnQDroZ4A66GfAOuhoADroaEA66GiAOuhowDroaQA66GlAOuhpgDroacA66GoAOuhqQDroaoA66GrAOuhrADroa0A66GuAOuhrwDrobAA66GxAOuhsgDrobMA66G0AOuhtQDrobYA66G3AOuhuADrobkA66G6AOuhuwDrobwA66G9AOuhvgDrob8A66KAAOuigQDrooIA66KDAOuihADrooUA66KGAOuihwDroogA66KJAOuiigDroosA66KMAOuijQDroo4A66KPAOuikADropEA66KSAOuikwDropQA66KVAOuilgDropcA66KYAOuimQDropoA66KbAOuinADrop0A66KeAOuinwDroqAA66KhAOuiogDroqMA66KkAOuipQDroqYA66KnAOuiqADroqkA66KqAOuiqwDroqwA66KtAOuirgDroq8A66KwAOuisQDrorIA66KzAOuitADrorUA66K2AOuitwDrorgA66K5AOuiugDrorsA66K8AOuivQDror4A66K/AOujgADro4EA66OCAOujgwDro4QA66OFAOujhgDro4cA66OIAOujiQDro4oA66OLAOujjADro40A66OOAOujjwDro5AA66ORAOujkgDro5MA66OUAOujlQDro5YA66OXAOujmADro5kA66OaAOujmwDro5wA66OdAOujngDro58A66OgAOujoQDro6IA66OjAOujpADro6UA66OmAOujpwDro6gA66OpAOujqgDro6sA66OsAOujrQDro64A66OvAOujsADro7EA66OyAOujswDro7QA66O1AOujtgDro7cA66O4AOujuQDro7oA66O7AOujvADro70A66O+AOujvwDrpIAA66SBAOukggDrpIMA66SEAOukhQDrpIYA66SHAOukiADrpIkA66SKAOukiwDrpIwA66SNAOukjgDrpI8A66SQAOukkQDrpJIA66STAOuklADrpJUA66SWAOuklwDrpJgA66SZAOukmgDrpJsA66ScAOuknQDrpJ4A66SfAOukoADrpKEA66SiAOukowDrpKQA66SlAOukpgDrpKcA66SoAOukqQDrpKoA66SrAOukrADrpK0A66SuAOukrwDrpLAA66SxAOuksgDrpLMA66S0AOuktQDrpLYA66S3AOukuADrpLkA66S6AOukuwDrpLwA66S9AOukvgDrpL8A66WAAOulgQDrpYIA66WDAOulhADrpYUA66WGAOulhwDrpYgA66WJAOuligDrpYsA66WMAOuljQDrpY4A66WPAOulkADrpZEA66WSAOulkwDrpZQA66WVAOullgDrpZcA66WYAOulmQDrpZoA66WbAOulnADrpZ0A66WeAOulnwDrpaAA66WhAOulogDrpaMA66WkAOulpQDrpaYA66WnAOulqADrpakA66WqAOulqwDrpawA66WtAOulrgDrpa8A66WwAOulsQDrpbIA66WzAOultADrpbUA66W2AOultwDrpbgA66W5AOulugDrpbsA66W8AOulvQDrpb4A66W/AOumgADrpoEA66aCAOumgwDrpoQA66aFAOumhgDrpocA66aIAOumiQDrpooA66aLAOumjADrpo0A66aOAOumjwDrppAA66aRAOumkgDrppMA66aUAOumlQDrppYA66aXAOummADrppkA66aaAOummwDrppwA66adAOumngDrpp8A66agAOumoQDrpqIA66ajAOumpADrpqUA66amAOumpwDrpqgA66apAOumqgDrpqsA66asAOumrQDrpq4A66avAOumsADrprEA66ayAOumswDrprQA66a1AOumtgDrprcA66a4AOumuQDrproA66a7AOumvADrpr0A66a+AOumvwDrp4AA66eBAOunggDrp4MA66eEAOunhQDrp4YA66eHAOuniADrp4kA66eKAOuniwDrp4wA66eNAOunjgDrp48A66eQAOunkQDrp5IA66eTAOunlADrp5UA66eWAOunlwDrp5gA66eZAOunmgDrp5sA66ecAOunnQDrp54A66efAOunoADrp6EA66eiAOunowDrp6QA66elAOunpgDrp6cA66eoAOunqQDrp6oA66erAOunrADrp60A66euAOunrwDrp7AA66exAOunsgDrp7MA66e0AOuntQDrp7YA66e3AOunuADrp7kA66e6AOunuwDrp7wA66e9AOunvgDrp78A66iAAOuogQDrqIIA66iDAOuohADrqIUA66iGAOuohwDrqIgA66iJAOuoigDrqIsA66iMAOuojQDrqI4A66iPAOuokADrqJEA66iSAOuokwDrqJQA66iVAOuolgDrqJcA66iYAOuomQDrqJoA66ibAOuonADrqJ0A66ieAOuonwDrqKAA66ihAOuoogDrqKMA66ikAOuopQDrqKYA66inAOuoqADrqKkA66iqAOuoqwDrqKwA66itAOuorgDrqK8A66iwAOuosQDrqLIA66izAOuotADrqLUA66i2AOuotwDrqLgA66i5AOuougDrqLsA66i8AOuovQDrqL4A66i/AOupgADrqYEA66mCAOupgwDrqYQA66mFAOuphgDrqYcA66mIAOupiQDrqYoA66mLAOupjADrqY0A66mOAOupjwDrqZAA66mRAOupkgDrqZMA66mUAOuplQDrqZYA66mXAOupmADrqZkA66maAOupmwDrqZwA66mdAOupngDrqZ8A66mgAOupoQDrqaIA66mjAOuppADrqaUA66mmAOuppwDrqagA66mpAOupqgDrqasA66msAOuprQDrqa4A66mvAOupsADrqbEA66myAOupswDrqbQA66m1AOuptgDrqbcA66m4AOupuQDrqboA66m7AOupvADrqb0A66m+AOupvwDrqoAA66qBAOuqggDrqoMA66qEAOuqhQDrqoYA66qHAOuqiADrqokA66qKAOuqiwDrqowA66qNAOuqjgDrqo8A66qQAOuqkQDrqpIA66qTAOuqlADrqpUA66qWAOuqlwDrqpgA66qZAOuqmgDrqpsA66qcAOuqnQDrqp4A66qfAOuqoADrqqEA66qiAOuqowDrqqQA66qlAOuqpgDrqqcA66qoAOuqqQDrqqoA66qrAOuqrADrqq0A66quAOuqrwDrqrAA66qxAOuqsgDrqrMA66q0AOuqtQDrqrYA66q3AOuquADrqrkA66q6AOuquwDrqrwA66q9AOuqvgDrqr8A66uAAOurgQDrq4IA66uDAOurhADrq4UA66uGAOurhwDrq4gA66uJAOurigDrq4sA66uMAOurjQDrq44A66uPAOurkADrq5EA66uSAOurkwDrq5QA66uVAOurlgDrq5cA66uYAOurmQDrq5oA66ubAOurnADrq50A66ueAOurnwDrq6AA66uhAOurogDrq6MA66ukAOurpQDrq6YA66unAOurqADrq6kA66uqAOurqwDrq6wA66utAOurrgDrq68A66uwAOursQDrq7IA66uzAOurtADrq7UA66u2AOurtwDrq7gA66u5AOurugDrq7sA66u8AOurvQDrq74A66u/AOusgADrrIEA66yCAOusgwDrrIQA66yFAOushgDrrIcA66yIAOusiQDrrIoA66yLAOusjADrrI0A66yOAOusjwDrrJAA66yRAOuskgDrrJMA66yUAOuslQDrrJYA66yXAOusmADrrJkA66yaAOusmwDrrJwA66ydAOusngDrrJ8A66ygAOusoQDrrKIA66yjAOuspADrrKUA66ymAOuspwDrrKgA66ypAOusqgDrrKsA66ysAOusrQDrrK4A66yvAOussADrrLEA66yyAOusswDrrLQA66y1AOustgDrrLcA66y4AOusuQDrrLoA66y7AOusvADrrL0A66y+AOusvwDrrYAA662BAOutggDrrYMA662EAOuthQDrrYYA662HAOutiADrrYkA662KAOutiwDrrYwA662NAOutjgDrrY8A662QAOutkQDrrZIA662TAOutlADrrZUA662WAOutlwDrrZgA662ZAOutmgDrrZsA662cAOutnQDrrZ4A662fAOutoADrraEA662iAOutowDrraQA662lAOutpgDrracA662oAOutqQDrraoA662rAOutrADrra0A662uAOutrwDrrbAA662xAOutsgDrrbMA6620AOuttQDrrbYA6623AOutuADrrbkA6626AOutuwDrrbwA6629AOutvgDrrb8A666AAOuugQDrroIA666DAOuuhADrroUA666GAOuuhwDrrogA666JAOuuigDrrosA666MAOuujQDrro4A666PAOuukADrrpEA666SAOuukwDrrpQA666VAOuulgDrrpcA666YAOuumQDrrpoA666bAOuunADrrp0A666eAOuunwDrrqAA666hAOuuogDrrqMA666kAOuupQDrrqYA666nAOuuqADrrqkA666qAOuuqwDrrqwA666tAOuurgDrrq8A666wAOuusQDrrrIA666zAOuutADrrrUA6662AOuutwDrrrgA6665AOuuugDrrrsA6668AOuuvQDrrr4A666/AOuvgADrr4EA66+CAOuvgwDrr4QA66+FAOuvhgDrr4cA66+IAOuviQDrr4oA66+LAOuvjADrr40A66+OAOuvjwDrr5AA66+RAOuvkgDrr5MA66+UAOuvlQDrr5YA66+XAOuvmADrr5kA66+aAOuvmwDrr5wA66+dAOuvngDrr58A66+gAOuvoQDrr6IA66+jAOuvpADrr6UA66+mAOuvpwDrr6gA66+pAOuvqgDrr6sA66+sAOuvrQDrr64A66+vAOuvsADrr7EA66+yAOuvswDrr7QA66+1AOuvtgDrr7cA66+4AOuvuQDrr7oA66+7AOuvvADrr70A66++AOuvvwDrsIAA67CBAOuwggDrsIMA67CEAOuwhQDrsIYA67CHAOuwiADrsIkA67CKAOuwiwDrsIwA67CNAOuwjgDrsI8A67CQAOuwkQDrsJIA67CTAOuwlADrsJUA67CWAOuwlwDrsJgA67CZAOuwmgDrsJsA67CcAOuwnQDrsJ4A67CfAOuwoADrsKEA67CiAOuwowDrsKQA67ClAOuwpgDrsKcA67CoAOuwqQDrsKoA67CrAOuwrADrsK0A67CuAOuwrwDrsLAA67CxAOuwsgDrsLMA67C0AOuwtQDrsLYA67C3AOuwuADrsLkA67C6AOuwuwDrsLwA67C9AOuwvgDrsL8A67GAAOuxgQDrsYIA67GDAOuxhADrsYUA67GGAOuxhwDrsYgA67GJAOuxigDrsYsA67GMAOuxjQDrsY4A67GPAOuxkADrsZEA67GSAOuxkwDrsZQA67GVAOuxlgDrsZcA67GYAOuxmQDrsZoA67GbAOuxnADrsZ0A67GeAOuxnwDrsaAA67GhAOuxogDrsaMA67GkAOuxpQDrsaYA67GnAOuxqADrsakA67GqAOuxqwDrsawA67GtAOuxrgDrsa8A67GwAOuxsQDrsbIA67GzAOuxtADrsbUA67G2AOuxtwDrsbgA67G5AOuxugDrsbsA67G8AOuxvQDrsb4A67G/AOuygADrsoEA67KCAOuygwDrsoQA67KFAOuyhgDrsocA67KIAOuyiQDrsooA67KLAOuyjADrso0A67KOAOuyjwDrspAA67KRAOuykgDrspMA67KUAOuylQDrspYA67KXAOuymADrspkA67KaAOuymwDrspwA67KdAOuyngDrsp8A67KgAOuyoQDrsqIA67KjAOuypADrsqUA67KmAOuypwDrsqgA67KpAOuyqgDrsqsA67KsAOuyrQDrsq4A67KvAOuysADrsrEA67KyAOuyswDrsrQA67K1AOuytgDrsrcA67K4AOuyuQDrsroA67K7AOuyvADrsr0A67K+AOuyvwDrs4AA67OBAOuzggDrs4MA67OEAOuzhQDrs4YA67OHAOuziADrs4kA67OKAOuziwDrs4wA67ONAOuzjgDrs48A67OQAOuzkQDrs5IA67OTAOuzlADrs5UA67OWAOuzlwDrs5gA67OZAOuzmgDrs5sA67OcAOuznQDrs54A67OfAOuzoADrs6EA67OiAOuzowDrs6QA67OlAOuzpgDrs6cA67OoAOuzqQDrs6oA67OrAOuzrADrs60A67OuAOuzrwDrs7AA67OxAOuzsgDrs7MA67O0AOuztQDrs7YA67O3AOuzuADrs7kA67O6AOuzuwDrs7wA67O9AOuzvgDrs78A67SAAOu0gQDrtIIA67SDAOu0hADrtIUA67SGAOu0hwDrtIgA67SJAOu0igDrtIsA67SMAOu0jQDrtI4A67SPAOu0kADrtJEA67SSAOu0kwDrtJQA67SVAOu0lgDrtJcA67SYAOu0mQDrtJoA67SbAOu0nADrtJ0A67SeAOu0nwDrtKAA67ShAOu0ogDrtKMA67SkAOu0pQDrtKYA67SnAOu0qADrtKkA67SqAOu0qwDrtKwA67StAOu0rgDrtK8A67SwAOu0sQDrtLIA67SzAOu0tADrtLUA67S2AOu0twDrtLgA67S5AOu0ugDrtLsA67S8AOu0vQDrtL4A67S/AOu1gADrtYEA67WCAOu1gwDrtYQA67WFAOu1hgDrtYcA67WIAOu1iQDrtYoA67WLAOu1jADrtY0A67WOAOu1jwDrtZAA67WRAOu1kgDrtZMA67WUAOu1lQDrtZYA67WXAOu1mADrtZkA67WaAOu1mwDrtZwA67WdAOu1ngDrtZ8A67WgAOu1oQDrtaIA67WjAOu1pADrtaUA67WmAOu1pwDrtagA67WpAOu1qgDrtasA67WsAOu1rQDrta4A67WvAOu1sADrtbEA67WyAOu1swDrtbQA67W1AOu1tgDrtbcA67W4AOu1uQDrtboA67W7AOu1vADrtb0A67W+AOu1vwDrtoAA67aBAOu2ggDrtoMA67aEAOu2hQDrtoYA67aHAOu2iADrtokA67aKAOu2iwDrtowA67aNAOu2jgDrto8A67aQAOu2kQDrtpIA67aTAOu2lADrtpUA67aWAOu2lwDrtpgA67aZAOu2mgDrtpsA67acAOu2nQDrtp4A67afAOu2oADrtqEA67aiAOu2owDrtqQA67alAOu2pgDrtqcA67aoAOu2qQDrtqoA67arAOu2rADrtq0A67auAOu2rwDrtrAA67axAOu2sgDrtrMA67a0AOu2tQDrtrYA67a3AOu2uADrtrkA67a6AOu2uwDrtrwA67a9AOu2vgDrtr8A67eAAOu3gQDrt4IA67eDAOu3hADrt4UA67eGAOu3hwDrt4gA67eJAOu3igDrt4sA67eMAOu3jQDrt44A67ePAOu3kADrt5EA67eSAOu3kwDrt5QA67eVAOu3lgDrt5cA67eYAOu3mQDrt5oA67ebAOu3nADrt50A67eeAOu3nwDrt6AA67ehAOu3ogDrt6MA67ekAOu3pQDrt6YA67enAOu3qADrt6kA67eqAOu3qwDrt6wA67etAOu3rgDrt68A67ewAOu3sQDrt7IA67ezAOu3tADrt7UA67e2AOu3twDrt7gA67e5AOu3ugDrt7sA67e8AOu3vQDrt74A67e/AOu4gADruIEA67iCAOu4gwDruIQA67iFAOu4hgDruIcA67iIAOu4iQDruIoA67iLAOu4jADruI0A67iOAOu4jwDruJAA67iRAOu4kgDruJMA67iUAOu4lQDruJYA67iXAOu4mADruJkA67iaAOu4mwDruJwA67idAOu4ngDruJ8A67igAOu4oQDruKIA67ijAOu4pADruKUA67imAOu4pwDruKgA67ipAOu4qgDruKsA67isAOu4rQDruK4A67ivAOu4sADruLEA67iyAOu4swDruLQA67i1AOu4tgDruLcA67i4AOu4uQDruLoA67i7AOu4vADruL0A67i+AOu4vwDruYAA67mBAOu5ggDruYMA67mEAOu5hQDruYYA67mHAOu5iADruYkA67mKAOu5iwDruYwA67mNAOu5jgDruY8A67mQAOu5kQDruZIA67mTAOu5lADruZUA67mWAOu5lwDruZgA67mZAOu5mgDruZsA67mcAOu5nQDruZ4A67mfAOu5oADruaEA67miAOu5owDruaQA67mlAOu5pgDruacA67moAOu5qQDruaoA67mrAOu5rADrua0A67muAOu5rwDrubAA67mxAOu5sgDrubMA67m0AOu5tQDrubYA67m3AOu5uADrubkA67m6AOu5uwDrubwA67m9AOu5vgDrub8A67qAAOu6gQDruoIA67qDAOu6hADruoUA67qGAOu6hwDruogA67qJAOu6igDruosA67qMAOu6jQDruo4A67qPAOu6kADrupEA67qSAOu6kwDrupQA67qVAOu6lgDrupcA67qYAOu6mQDrupoA67qbAOu6nADrup0A67qeAOu6nwDruqAA67qhAOu6ogDruqMA67qkAOu6pQDruqYA67qnAOu6qADruqkA67qqAOu6qwDruqwA67qtAOu6rgDruq8A67qwAOu6sQDrurIA67qzAOu6tADrurUA67q2AOu6twDrurgA67q5AOu6ugDrursA67q8AOu6vQDrur4A67q/AOu7gADru4EA67uCAOu7gwDru4QA67uFAOu7hgDru4cA67uIAOu7iQDru4oA67uLAOu7jADru40A67uOAOu7jwDru5AA67uRAOu7kgDru5MA67uUAOu7lQDru5YA67uXAOu7mADru5kA67uaAOu7mwDru5wA67udAOu7ngDru58A67ugAOu7oQDru6IA67ujAOu7pADru6UA67umAOu7pwDru6gA67upAOu7qgDru6sA67usAOu7rQDru64A67uvAOu7sADru7EA67uyAOu7swDru7QA67u1AOu7tgDru7cA67u4AOu7uQDru7oA67u7AOu7vADru70A67u+AOu7vwDrvIAA67yBAOu8ggDrvIMA67yEAOu8hQDrvIYA67yHAOu8iADrvIkA67yKAOu8iwDrvIwA67yNAOu8jgDrvI8A67yQAOu8kQDrvJIA67yTAOu8lADrvJUA67yWAOu8lwDrvJgA67yZAOu8mgDrvJsA67ycAOu8nQDrvJ4A67yfAOu8oADrvKEA67yiAOu8owDrvKQA67ylAOu8pgDrvKcA67yoAOu8qQDrvKoA67yrAOu8rADrvK0A67yuAOu8rwDrvLAA67yxAOu8sgDrvLMA67y0AOu8tQDrvLYA67y3AOu8uADrvLkA67y6AOu8uwDrvLwA67y9AOu8vgDrvL8A672AAOu9gQDrvYIA672DAOu9hADrvYUA672GAOu9hwDrvYgA672JAOu9igDrvYsA672MAOu9jQDrvY4A672PAOu9kADrvZEA672SAOu9kwDrvZQA672VAOu9lgDrvZcA672YAOu9mQDrvZoA672bAOu9nADrvZ0A672eAOu9nwDrvaAA672hAOu9ogDrvaMA672kAOu9pQDrvaYA672nAOu9qADrvakA672qAOu9qwDrvawA672tAOu9rgDrva8A672wAOu9sQDrvbIA672zAOu9tADrvbUA6722AOu9twDrvbgA6725AOu9ugDrvbsA6728AOu9vQDrvb4A672/AOu+gADrvoEA676CAOu+gwDrvoQA676FAOu+hgDrvocA676IAOu+iQDrvooA676LAOu+jADrvo0A676OAOu+jwDrvpAA676RAOu+kgDrvpMA676UAOu+lQDrvpYA676XAOu+mADrvpkA676aAOu+mwDrvpwA676dAOu+ngDrvp8A676gAOu+oQDrvqIA676jAOu+pADrvqUA676mAOu+pwDrvqgA676pAOu+qgDrvqsA676sAOu+rQDrvq4A676vAOu+sADrvrEA676yAOu+swDrvrQA6761AOu+tgDrvrcA6764AOu+uQDrvroA6767AOu+vADrvr0A676+AOu+vwDrv4AA67+BAOu/ggDrv4MA67+EAOu/hQDrv4YA67+HAOu/iADrv4kA67+KAOu/iwDrv4wA67+NAOu/jgDrv48A67+QAOu/kQDrv5IA67+TAOu/lADrv5UA67+WAOu/lwDrv5gA67+ZAOu/mgDrv5sA67+cAOu/nQDrv54A67+fAOu/oADrv6EA67+iAOu/owDrv6QA67+lAOu/pgDrv6cA67+oAOu/qQDrv6oA67+rAOu/rADrv60A67+uAOu/rwDrv7AA67+xAOu/sgDrv7MA67+0AOu/tQDrv7YA67+3AOu/uADrv7kA67+6AOu/uwDrv7wA67+9AOu/vgDrv78A7ICAAOyAgQDsgIIA7ICDAOyAhADsgIUA7ICGAOyAhwDsgIgA7ICJAOyAigDsgIsA7ICMAOyAjQDsgI4A7ICPAOyAkADsgJEA7ICSAOyAkwDsgJQA7ICVAOyAlgDsgJcA7ICYAOyAmQDsgJoA7ICbAOyAnADsgJ0A7ICeAOyAnwDsgKAA7IChAOyAogDsgKMA7ICkAOyApQDsgKYA7ICnAOyAqADsgKkA7ICqAOyAqwDsgKwA7ICtAOyArgDsgK8A7ICwAOyAsQDsgLIA7ICzAOyAtADsgLUA7IC2AOyAtwDsgLgA7IC5AOyAugDsgLsA7IC8AOyAvQDsgL4A7IC/AOyBgADsgYEA7IGCAOyBgwDsgYQA7IGFAOyBhgDsgYcA7IGIAOyBiQDsgYoA7IGLAOyBjADsgY0A7IGOAOyBjwDsgZAA7IGRAOyBkgDsgZMA7IGUAOyBlQDsgZYA7IGXAOyBmADsgZkA7IGaAOyBmwDsgZwA7IGdAOyBngDsgZ8A7IGgAOyBoQDsgaIA7IGjAOyBpADsgaUA7IGmAOyBpwDsgagA7IGpAOyBqgDsgasA7IGsAOyBrQDsga4A7IGvAOyBsADsgbEA7IGyAOyBswDsgbQA7IG1AOyBtgDsgbcA7IG4AOyBuQDsgboA7IG7AOyBvADsgb0A7IG+AOyBvwDsgoAA7IKBAOyCggDsgoMA7IKEAOyChQDsgoYA7IKHAOyCiADsgokA7IKKAOyCiwDsgowA7IKNAOyCjgDsgo8A7IKQAOyCkQDsgpIA7IKTAOyClADsgpUA7IKWAOyClwDsgpgA7IKZAOyCmgDsgpsA7IKcAOyCnQDsgp4A7IKfAOyCoADsgqEA7IKiAOyCowDsgqQA7IKlAOyCpgDsgqcA7IKoAOyCqQDsgqoA7IKrAOyCrADsgq0A7IKuAOyCrwDsgrAA7IKxAOyCsgDsgrMA7IK0AOyCtQDsgrYA7IK3AOyCuADsgrkA7IK6AOyCuwDsgrwA7IK9AOyCvgDsgr8A7IOAAOyDgQDsg4IA7IODAOyDhADsg4UA7IOGAOyDhwDsg4gA7IOJAOyDigDsg4sA7IOMAOyDjQDsg44A7IOPAOyDkADsg5EA7IOSAOyDkwDsg5QA7IOVAOyDlgDsg5cA7IOYAOyDmQDsg5oA7IObAOyDnADsg50A7IOeAOyDnwDsg6AA7IOhAOyDogDsg6MA7IOkAOyDpQDsg6YA7IOnAOyDqADsg6kA7IOqAOyDqwDsg6wA7IOtAOyDrgDsg68A7IOwAOyDsQDsg7IA7IOzAOyDtADsg7UA7IO2AOyDtwDsg7gA7IO5AOyDugDsg7sA7IO8AOyDvQDsg74A7IO/AOyEgADshIEA7ISCAOyEgwDshIQA7ISFAOyEhgDshIcA7ISIAOyEiQDshIoA7ISLAOyEjADshI0A7ISOAOyEjwDshJAA7ISRAOyEkgDshJMA7ISUAOyElQDshJYA7ISXAOyEmADshJkA7ISaAOyEmwDshJwA7ISdAOyEngDshJ8A7ISgAOyEoQDshKIA7ISjAOyEpADshKUA7ISmAOyEpwDshKgA7ISpAOyEqgDshKsA7ISsAOyErQDshK4A7ISvAOyEsADshLEA7ISyAOyEswDshLQA7IS1AOyEtgDshLcA7IS4AOyEuQDshLoA7IS7AOyEvADshL0A7IS+AOyEvwDshYAA7IWBAOyFggDshYMA7IWEAOyFhQDshYYA7IWHAOyFiADshYkA7IWKAOyFiwDshYwA7IWNAOyFjgDshY8A7IWQAOyFkQDshZIA7IWTAOyFlADshZUA7IWWAOyFlwDshZgA7IWZAOyFmgDshZsA7IWcAOyFnQDshZ4A7IWfAOyFoADshaEA7IWiAOyFowDshaQA7IWlAOyFpgDshacA7IWoAOyFqQDshaoA7IWrAOyFrADsha0A7IWuAOyFrwDshbAA7IWxAOyFsgDshbMA7IW0AOyFtQDshbYA7IW3AOyFuADshbkA7IW6AOyFuwDshbwA7IW9AOyFvgDshb8A7IaAAOyGgQDshoIA7IaDAOyGhADshoUA7IaGAOyGhwDshogA7IaJAOyGigDshosA7IaMAOyGjQDsho4A7IaPAOyGkADshpEA7IaSAOyGkwDshpQA7IaVAOyGlgDshpcA7IaYAOyGmQDshpoA7IabAOyGnADshp0A7IaeAOyGnwDshqAA7IahAOyGogDshqMA7IakAOyGpQDshqYA7IanAOyGqADshqkA7IaqAOyGqwDshqwA7IatAOyGrgDshq8A7IawAOyGsQDshrIA7IazAOyGtADshrUA7Ia2AOyGtwDshrgA7Ia5AOyGugDshrsA7Ia8AOyGvQDshr4A7Ia/AOyHgADsh4EA7IeCAOyHgwDsh4QA7IeFAOyHhgDsh4cA7IeIAOyHiQDsh4oA7IeLAOyHjADsh40A7IeOAOyHjwDsh5AA7IeRAOyHkgDsh5MA7IeUAOyHlQDsh5YA7IeXAOyHmADsh5kA7IeaAOyHmwDsh5wA7IedAOyHngDsh58A7IegAOyHoQDsh6IA7IejAOyHpADsh6UA7IemAOyHpwDsh6gA7IepAOyHqgDsh6sA7IesAOyHrQDsh64A7IevAOyHsADsh7EA7IeyAOyHswDsh7QA7Ie1AOyHtgDsh7cA7Ie4AOyHuQDsh7oA7Ie7AOyHvADsh70A7Ie+AOyHvwDsiIAA7IiBAOyIggDsiIMA7IiEAOyIhQDsiIYA7IiHAOyIiADsiIkA7IiKAOyIiwDsiIwA7IiNAOyIjgDsiI8A7IiQAOyIkQDsiJIA7IiTAOyIlADsiJUA7IiWAOyIlwDsiJgA7IiZAOyImgDsiJsA7IicAOyInQDsiJ4A7IifAOyIoADsiKEA7IiiAOyIowDsiKQA7IilAOyIpgDsiKcA7IioAOyIqQDsiKoA7IirAOyIrADsiK0A7IiuAOyIrwDsiLAA7IixAOyIsgDsiLMA7Ii0AOyItQDsiLYA7Ii3AOyIuADsiLkA7Ii6AOyIuwDsiLwA7Ii9AOyIvgDsiL8A7ImAAOyJgQDsiYIA7ImDAOyJhADsiYUA7ImGAOyJhwDsiYgA7ImJAOyJigDsiYsA7ImMAOyJjQDsiY4A7ImPAOyJkADsiZEA7ImSAOyJkwDsiZQA7ImVAOyJlgDsiZcA7ImYAOyJmQDsiZoA7ImbAOyJnADsiZ0A7ImeAOyJnwDsiaAA7ImhAOyJogDsiaMA7ImkAOyJpQDsiaYA7ImnAOyJqADsiakA7ImqAOyJqwDsiawA7ImtAOyJrgDsia8A7ImwAOyJsQDsibIA7ImzAOyJtADsibUA7Im2AOyJtwDsibgA7Im5AOyJugDsibsA7Im8AOyJvQDsib4A7Im/AOyKgADsioEA7IqCAOyKgwDsioQA7IqFAOyKhgDsiocA7IqIAOyKiQDsiooA7IqLAOyKjADsio0A7IqOAOyKjwDsipAA7IqRAOyKkgDsipMA7IqUAOyKlQDsipYA7IqXAOyKmADsipkA7IqaAOyKmwDsipwA7IqdAOyKngDsip8A7IqgAOyKoQDsiqIA7IqjAOyKpADsiqUA7IqmAOyKpwDsiqgA7IqpAOyKqgDsiqsA7IqsAOyKrQDsiq4A7IqvAOyKsADsirEA7IqyAOyKswDsirQA7Iq1AOyKtgDsircA7Iq4AOyKuQDsiroA7Iq7AOyKvADsir0A7Iq+AOyKvwDsi4AA7IuBAOyLggDsi4MA7IuEAOyLhQDsi4YA7IuHAOyLiADsi4kA7IuKAOyLiwDsi4wA7IuNAOyLjgDsi48A7IuQAOyLkQDsi5IA7IuTAOyLlADsi5UA7IuWAOyLlwDsi5gA7IuZAOyLmgDsi5sA7IucAOyLnQDsi54A7IufAOyLoADsi6EA7IuiAOyLowDsi6QA7IulAOyLpgDsi6cA7IuoAOyLqQDsi6oA7IurAOyLrADsi60A7IuuAOyLrwDsi7AA7IuxAOyLsgDsi7MA7Iu0AOyLtQDsi7YA7Iu3AOyLuADsi7kA7Iu6AOyLuwDsi7wA7Iu9AOyLvgDsi78A7IyAAOyMgQDsjIIA7IyDAOyMhADsjIUA7IyGAOyMhwDsjIgA7IyJAOyMigDsjIsA7IyMAOyMjQDsjI4A7IyPAOyMkADsjJEA7IySAOyMkwDsjJQA7IyVAOyMlgDsjJcA7IyYAOyMmQDsjJoA7IybAOyMnADsjJ0A7IyeAOyMnwDsjKAA7IyhAOyMogDsjKMA7IykAOyMpQDsjKYA7IynAOyMqADsjKkA7IyqAOyMqwDsjKwA7IytAOyMrgDsjK8A7IywAOyMsQDsjLIA7IyzAOyMtADsjLUA7Iy2AOyMtwDsjLgA7Iy5AOyMugDsjLsA7Iy8AOyMvQDsjL4A7Iy/AOyNgADsjYEA7I2CAOyNgwDsjYQA7I2FAOyNhgDsjYcA7I2IAOyNiQDsjYoA7I2LAOyNjADsjY0A7I2OAOyNjwDsjZAA7I2RAOyNkgDsjZMA7I2UAOyNlQDsjZYA7I2XAOyNmADsjZkA7I2aAOyNmwDsjZwA7I2dAOyNngDsjZ8A7I2gAOyNoQDsjaIA7I2jAOyNpADsjaUA7I2mAOyNpwDsjagA7I2pAOyNqgDsjasA7I2sAOyNrQDsja4A7I2vAOyNsADsjbEA7I2yAOyNswDsjbQA7I21AOyNtgDsjbcA7I24AOyNuQDsjboA7I27AOyNvADsjb0A7I2+AOyNvwDsjoAA7I6BAOyOggDsjoMA7I6EAOyOhQDsjoYA7I6HAOyOiADsjokA7I6KAOyOiwDsjowA7I6NAOyOjgDsjo8A7I6QAOyOkQDsjpIA7I6TAOyOlADsjpUA7I6WAOyOlwDsjpgA7I6ZAOyOmgDsjpsA7I6cAOyOnQDsjp4A7I6fAOyOoADsjqEA7I6iAOyOowDsjqQA7I6lAOyOpgDsjqcA7I6oAOyOqQDsjqoA7I6rAOyOrADsjq0A7I6uAOyOrwDsjrAA7I6xAOyOsgDsjrMA7I60AOyOtQDsjrYA7I63AOyOuADsjrkA7I66AOyOuwDsjrwA7I69AOyOvgDsjr8A7I+AAOyPgQDsj4IA7I+DAOyPhADsj4UA7I+GAOyPhwDsj4gA7I+JAOyPigDsj4sA7I+MAOyPjQDsj44A7I+PAOyPkADsj5EA7I+SAOyPkwDsj5QA7I+VAOyPlgDsj5cA7I+YAOyPmQDsj5oA7I+bAOyPnADsj50A7I+eAOyPnwDsj6AA7I+hAOyPogDsj6MA7I+kAOyPpQDsj6YA7I+nAOyPqADsj6kA7I+qAOyPqwDsj6wA7I+tAOyPrgDsj68A7I+wAOyPsQDsj7IA7I+zAOyPtADsj7UA7I+2AOyPtwDsj7gA7I+5AOyPugDsj7sA7I+8AOyPvQDsj74A7I+/AOyQgADskIEA7JCCAOyQgwDskIQA7JCFAOyQhgDskIcA7JCIAOyQiQDskIoA7JCLAOyQjADskI0A7JCOAOyQjwDskJAA7JCRAOyQkgDskJMA7JCUAOyQlQDskJYA7JCXAOyQmADskJkA7JCaAOyQmwDskJwA7JCdAOyQngDskJ8A7JCgAOyQoQDskKIA7JCjAOyQpADskKUA7JCmAOyQpwDskKgA7JCpAOyQqgDskKsA7JCsAOyQrQDskK4A7JCvAOyQsADskLEA7JCyAOyQswDskLQA7JC1AOyQtgDskLcA7JC4AOyQuQDskLoA7JC7AOyQvADskL0A7JC+AOyQvwDskYAA7JGBAOyRggDskYMA7JGEAOyRhQDskYYA7JGHAOyRiADskYkA7JGKAOyRiwDskYwA7JGNAOyRjgDskY8A7JGQAOyRkQDskZIA7JGTAOyRlADskZUA7JGWAOyRlwDskZgA7JGZAOyRmgDskZsA7JGcAOyRnQDskZ4A7JGfAOyRoADskaEA7JGiAOyRowDskaQA7JGlAOyRpgDskacA7JGoAOyRqQDskaoA7JGrAOyRrADska0A7JGuAOyRrwDskbAA7JGxAOyRsgDskbMA7JG0AOyRtQDskbYA7JG3AOyRuADskbkA7JG6AOyRuwDskbwA7JG9AOyRvgDskb8A7JKAAOySgQDskoIA7JKDAOyShADskoUA7JKGAOyShwDskogA7JKJAOySigDskosA7JKMAOySjQDsko4A7JKPAOySkADskpEA7JKSAOySkwDskpQA7JKVAOySlgDskpcA7JKYAOySmQDskpoA7JKbAOySnADskp0A7JKeAOySnwDskqAA7JKhAOySogDskqMA7JKkAOySpQDskqYA7JKnAOySqADskqkA7JKqAOySqwDskqwA7JKtAOySrgDskq8A7JKwAOySsQDskrIA7JKzAOyStADskrUA7JK2AOyStwDskrgA7JK5AOySugDskrsA7JK8AOySvQDskr4A7JK/AOyTgADsk4EA7JOCAOyTgwDsk4QA7JOFAOyThgDsk4cA7JOIAOyTiQDsk4oA7JOLAOyTjADsk40A7JOOAOyTjwDsk5AA7JORAOyTkgDsk5MA7JOUAOyTlQDsk5YA7JOXAOyTmADsk5kA7JOaAOyTmwDsk5wA7JOdAOyTngDsk58A7JOgAOyToQDsk6IA7JOjAOyTpADsk6UA7JOmAOyTpwDsk6gA7JOpAOyTqgDsk6sA7JOsAOyTrQDsk64A7JOvAOyTsADsk7EA7JOyAOyTswDsk7QA7JO1AOyTtgDsk7cA7JO4AOyTuQDsk7oA7JO7AOyTvADsk70A7JO+AOyTvwDslIAA7JSBAOyUggDslIMA7JSEAOyUhQDslIYA7JSHAOyUiADslIkA7JSKAOyUiwDslIwA7JSNAOyUjgDslI8A7JSQAOyUkQDslJIA7JSTAOyUlADslJUA7JSWAOyUlwDslJgA7JSZAOyUmgDslJsA7JScAOyUnQDslJ4A7JSfAOyUoADslKEA7JSiAOyUowDslKQA7JSlAOyUpgDslKcA7JSoAOyUqQDslKoA7JSrAOyUrADslK0A7JSuAOyUrwDslLAA7JSxAOyUsgDslLMA7JS0AOyUtQDslLYA7JS3AOyUuADslLkA7JS6AOyUuwDslLwA7JS9AOyUvgDslL8A7JWAAOyVgQDslYIA7JWDAOyVhADslYUA7JWGAOyVhwDslYgA7JWJAOyVigDslYsA7JWMAOyVjQDslY4A7JWPAOyVkADslZEA7JWSAOyVkwDslZQA7JWVAOyVlgDslZcA7JWYAOyVmQDslZoA7JWbAOyVnADslZ0A7JWeAOyVnwDslaAA7JWhAOyVogDslaMA7JWkAOyVpQDslaYA7JWnAOyVqADslakA7JWqAOyVqwDslawA7JWtAOyVrgDsla8A7JWwAOyVsQDslbIA7JWzAOyVtADslbUA7JW2AOyVtwDslbgA7JW5AOyVugDslbsA7JW8AOyVvQDslb4A7JW/AOyWgADsloEA7JaCAOyWgwDsloQA7JaFAOyWhgDslocA7JaIAOyWiQDslooA7JaLAOyWjADslo0A7JaOAOyWjwDslpAA7JaRAOyWkgDslpMA7JaUAOyWlQDslpYA7JaXAOyWmADslpkA7JaaAOyWmwDslpwA7JadAOyWngDslp8A7JagAOyWoQDslqIA7JajAOyWpADslqUA7JamAOyWpwDslqgA7JapAOyWqgDslqsA7JasAOyWrQDslq4A7JavAOyWsADslrEA7JayAOyWswDslrQA7Ja1AOyWtgDslrcA7Ja4AOyWuQDslroA7Ja7AOyWvADslr0A7Ja+AOyWvwDsl4AA7JeBAOyXggDsl4MA7JeEAOyXhQDsl4YA7JeHAOyXiADsl4kA7JeKAOyXiwDsl4wA7JeNAOyXjgDsl48A7JeQAOyXkQDsl5IA7JeTAOyXlADsl5UA7JeWAOyXlwDsl5gA7JeZAOyXmgDsl5sA7JecAOyXnQDsl54A7JefAOyXoADsl6EA7JeiAOyXowDsl6QA7JelAOyXpgDsl6cA7JeoAOyXqQDsl6oA7JerAOyXrADsl60A7JeuAOyXrwDsl7AA7JexAOyXsgDsl7MA7Je0AOyXtQDsl7YA7Je3AOyXuADsl7kA7Je6AOyXuwDsl7wA7Je9AOyXvgDsl78A7JiAAOyYgQDsmIIA7JiDAOyYhADsmIUA7JiGAOyYhwDsmIgA7JiJAOyYigDsmIsA7JiMAOyYjQDsmI4A7JiPAOyYkADsmJEA7JiSAOyYkwDsmJQA7JiVAOyYlgDsmJcA7JiYAOyYmQDsmJoA7JibAOyYnADsmJ0A7JieAOyYnwDsmKAA7JihAOyYogDsmKMA7JikAOyYpQDsmKYA7JinAOyYqADsmKkA7JiqAOyYqwDsmKwA7JitAOyYrgDsmK8A7JiwAOyYsQDsmLIA7JizAOyYtADsmLUA7Ji2AOyYtwDsmLgA7Ji5AOyYugDsmLsA7Ji8AOyYvQDsmL4A7Ji/AOyZgADsmYEA7JmCAOyZgwDsmYQA7JmFAOyZhgDsmYcA7JmIAOyZiQDsmYoA7JmLAOyZjADsmY0A7JmOAOyZjwDsmZAA7JmRAOyZkgDsmZMA7JmUAOyZlQDsmZYA7JmXAOyZmADsmZkA7JmaAOyZmwDsmZwA7JmdAOyZngDsmZ8A7JmgAOyZoQDsmaIA7JmjAOyZpADsmaUA7JmmAOyZpwDsmagA7JmpAOyZqgDsmasA7JmsAOyZrQDsma4A7JmvAOyZsADsmbEA7JmyAOyZswDsmbQA7Jm1AOyZtgDsmbcA7Jm4AOyZuQDsmboA7Jm7AOyZvADsmb0A7Jm+AOyZvwDsmoAA7JqBAOyaggDsmoMA7JqEAOyahQDsmoYA7JqHAOyaiADsmokA7JqKAOyaiwDsmowA7JqNAOyajgDsmo8A7JqQAOyakQDsmpIA7JqTAOyalADsmpUA7JqWAOyalwDsmpgA7JqZAOyamgDsmpsA7JqcAOyanQDsmp4A7JqfAOyaoADsmqEA7JqiAOyaowDsmqQA7JqlAOyapgDsmqcA7JqoAOyaqQDsmqoA7JqrAOyarADsmq0A7JquAOyarwDsmrAA7JqxAOyasgDsmrMA7Jq0AOyatQDsmrYA7Jq3AOyauADsmrkA7Jq6AOyauwDsmrwA7Jq9AOyavgDsmr8A7JuAAOybgQDsm4IA7JuDAOybhADsm4UA7JuGAOybhwDsm4gA7JuJAOybigDsm4sA7JuMAOybjQDsm44A7JuPAOybkADsm5EA7JuSAOybkwDsm5QA7JuVAOyblgDsm5cA7JuYAOybmQDsm5oA7JubAOybnADsm50A7JueAOybnwDsm6AA7JuhAOybogDsm6MA7JukAOybpQDsm6YA7JunAOybqADsm6kA7JuqAOybqwDsm6wA7JutAOybrgDsm68A7JuwAOybsQDsm7IA7JuzAOybtADsm7UA7Ju2AOybtwDsm7gA7Ju5AOybugDsm7sA7Ju8AOybvQDsm74A7Ju/AOycgADsnIEA7JyCAOycgwDsnIQA7JyFAOychgDsnIcA7JyIAOyciQDsnIoA7JyLAOycjADsnI0A7JyOAOycjwDsnJAA7JyRAOyckgDsnJMA7JyUAOyclQDsnJYA7JyXAOycmADsnJkA7JyaAOycmwDsnJwA7JydAOycngDsnJ8A7JygAOycoQDsnKIA7JyjAOycpADsnKUA7JymAOycpwDsnKgA7JypAOycqgDsnKsA7JysAOycrQDsnK4A7JyvAOycsADsnLEA7JyyAOycswDsnLQA7Jy1AOyctgDsnLcA7Jy4AOycuQDsnLoA7Jy7AOycvADsnL0A7Jy+AOycvwDsnYAA7J2BAOydggDsnYMA7J2EAOydhQDsnYYA7J2HAOydiADsnYkA7J2KAOydiwDsnYwA7J2NAOydjgDsnY8A7J2QAOydkQDsnZIA7J2TAOydlADsnZUA7J2WAOydlwDsnZgA7J2ZAOydmgDsnZsA7J2cAOydnQDsnZ4A7J2fAOydoADsnaEA7J2iAOydowDsnaQA7J2lAOydpgDsnacA7J2oAOydqQDsnaoA7J2rAOydrADsna0A7J2uAOydrwDsnbAA7J2xAOydsgDsnbMA7J20AOydtQDsnbYA7J23AOyduADsnbkA7J26AOyduwDsnbwA7J29AOydvgDsnb8A7J6AAOyegQDsnoIA7J6DAOyehADsnoUA7J6GAOyehwDsnogA7J6JAOyeigDsnosA7J6MAOyejQDsno4A7J6PAOyekADsnpEA7J6SAOyekwDsnpQA7J6VAOyelgDsnpcA7J6YAOyemQDsnpoA7J6bAOyenADsnp0A7J6eAOyenwDsnqAA7J6hAOyeogDsnqMA7J6kAOyepQDsnqYA7J6nAOyeqADsnqkA7J6qAOyeqwDsnqwA7J6tAOyergDsnq8A7J6wAOyesQDsnrIA7J6zAOyetADsnrUA7J62AOyetwDsnrgA7J65AOyeugDsnrsA7J68AOyevQDsnr4A7J6/AOyfgADsn4EA7J+CAOyfgwDsn4QA7J+FAOyfhgDsn4cA7J+IAOyfiQDsn4oA7J+LAOyfjADsn40A7J+OAOyfjwDsn5AA7J+RAOyfkgDsn5MA7J+UAOyflQDsn5YA7J+XAOyfmADsn5kA7J+aAOyfmwDsn5wA7J+dAOyfngDsn58A7J+gAOyfoQDsn6IA7J+jAOyfpADsn6UA7J+mAOyfpwDsn6gA7J+pAOyfqgDsn6sA7J+sAOyfrQDsn64A7J+vAOyfsADsn7EA7J+yAOyfswDsn7QA7J+1AOyftgDsn7cA7J+4AOyfuQDsn7oA7J+7AOyfvADsn70A7J++AOyfvwDsoIAA7KCBAOygggDsoIMA7KCEAOyghQDsoIYA7KCHAOygiADsoIkA7KCKAOygiwDsoIwA7KCNAOygjgDsoI8A7KCQAOygkQDsoJIA7KCTAOyglADsoJUA7KCWAOyglwDsoJgA7KCZAOygmgDsoJsA7KCcAOygnQDsoJ4A7KCfAOygoADsoKEA7KCiAOygowDsoKQA7KClAOygpgDsoKcA7KCoAOygqQDsoKoA7KCrAOygrADsoK0A7KCuAOygrwDsoLAA7KCxAOygsgDsoLMA7KC0AOygtQDsoLYA7KC3AOyguADsoLkA7KC6AOyguwDsoLwA7KC9AOygvgDsoL8A7KGAAOyhgQDsoYIA7KGDAOyhhADsoYUA7KGGAOyhhwDsoYgA7KGJAOyhigDsoYsA7KGMAOyhjQDsoY4A7KGPAOyhkADsoZEA7KGSAOyhkwDsoZQA7KGVAOyhlgDsoZcA7KGYAOyhmQDsoZoA7KGbAOyhnADsoZ0A7KGeAOyhnwDsoaAA7KGhAOyhogDsoaMA7KGkAOyhpQDsoaYA7KGnAOyhqADsoakA7KGqAOyhqwDsoawA7KGtAOyhrgDsoa8A7KGwAOyhsQDsobIA7KGzAOyhtADsobUA7KG2AOyhtwDsobgA7KG5AOyhugDsobsA7KG8AOyhvQDsob4A7KG/AOyigADsooEA7KKCAOyigwDsooQA7KKFAOyihgDsoocA7KKIAOyiiQDsoooA7KKLAOyijADsoo0A7KKOAOyijwDsopAA7KKRAOyikgDsopMA7KKUAOyilQDsopYA7KKXAOyimADsopkA7KKaAOyimwDsopwA7KKdAOyingDsop8A7KKgAOyioQDsoqIA7KKjAOyipADsoqUA7KKmAOyipwDsoqgA7KKpAOyiqgDsoqsA7KKsAOyirQDsoq4A7KKvAOyisADsorEA7KKyAOyiswDsorQA7KK1AOyitgDsorcA7KK4AOyiuQDsoroA7KK7AOyivADsor0A7KK+AOyivwDso4AA7KOBAOyjggDso4MA7KOEAOyjhQDso4YA7KOHAOyjiADso4kA7KOKAOyjiwDso4wA7KONAOyjjgDso48A7KOQAOyjkQDso5IA7KOTAOyjlADso5UA7KOWAOyjlwDso5gA7KOZAOyjmgDso5sA7KOcAOyjnQDso54A7KOfAOyjoADso6EA7KOiAOyjowDso6QA7KOlAOyjpgDso6cA7KOoAOyjqQDso6oA7KOrAOyjrADso60A7KOuAOyjrwDso7AA7KOxAOyjsgDso7MA7KO0AOyjtQDso7YA7KO3AOyjuADso7kA7KO6AOyjuwDso7wA7KO87J2YAOyjvQDso74A7KO/AOykgADspIEA7KSCAOykgwDspIQA7KSFAOykhgDspIcA7KSIAOykiQDspIoA7KSLAOykjADspI0A7KSOAOykjwDspJAA7KSRAOykkgDspJMA7KSUAOyklQDspJYA7KSXAOykmADspJkA7KSaAOykmwDspJwA7KSdAOykngDspJ8A7KSgAOykoQDspKIA7KSjAOykpADspKUA7KSmAOykpwDspKgA7KSpAOykqgDspKsA7KSsAOykrQDspK4A7KSvAOyksADspLEA7KSyAOykswDspLQA7KS1AOyktgDspLcA7KS4AOykuQDspLoA7KS7AOykvADspL0A7KS+AOykvwDspYAA7KWBAOylggDspYMA7KWEAOylhQDspYYA7KWHAOyliADspYkA7KWKAOyliwDspYwA7KWNAOyljgDspY8A7KWQAOylkQDspZIA7KWTAOyllADspZUA7KWWAOyllwDspZgA7KWZAOylmgDspZsA7KWcAOylnQDspZ4A7KWfAOyloADspaEA7KWiAOylowDspaQA7KWlAOylpgDspacA7KWoAOylqQDspaoA7KWrAOylrADspa0A7KWuAOylrwDspbAA7KWxAOylsgDspbMA7KW0AOyltQDspbYA7KW3AOyluADspbkA7KW6AOyluwDspbwA7KW9AOylvgDspb8A7KaAAOymgQDspoIA7KaDAOymhADspoUA7KaGAOymhwDspogA7KaJAOymigDsposA7KaMAOymjQDspo4A7KaPAOymkADsppEA7KaSAOymkwDsppQA7KaVAOymlgDsppcA7KaYAOymmQDsppoA7KabAOymnADspp0A7KaeAOymnwDspqAA7KahAOymogDspqMA7KakAOympQDspqYA7KanAOymqADspqkA7KaqAOymqwDspqwA7KatAOymrgDspq8A7KawAOymsQDsprIA7KazAOymtADsprUA7Ka2AOymtwDsprgA7Ka5AOymugDsprsA7Ka8AOymvQDspr4A7Ka/AOyngADsp4EA7KeCAOyngwDsp4QA7KeFAOynhgDsp4cA7KeIAOyniQDsp4oA7KeLAOynjADsp40A7KeOAOynjwDsp5AA7KeRAOynkgDsp5MA7KeUAOynlQDsp5YA7KeXAOynmADsp5kA7KeaAOynmwDsp5wA7KedAOynngDsp58A7KegAOynoQDsp6IA7KejAOynpADsp6UA7KemAOynpwDsp6gA7KepAOynqgDsp6sA7KesAOynrQDsp64A7KevAOynsADsp7EA7KeyAOynswDsp7QA7Ke1AOyntgDsp7cA7Ke4AOynuQDsp7oA7Ke7AOynvADsp70A7Ke+AOynvwDsqIAA7KiBAOyoggDsqIMA7KiEAOyohQDsqIYA7KiHAOyoiADsqIkA7KiKAOyoiwDsqIwA7KiNAOyojgDsqI8A7KiQAOyokQDsqJIA7KiTAOyolADsqJUA7KiWAOyolwDsqJgA7KiZAOyomgDsqJsA7KicAOyonQDsqJ4A7KifAOyooADsqKEA7KiiAOyoowDsqKQA7KilAOyopgDsqKcA7KioAOyoqQDsqKoA7KirAOyorADsqK0A7KiuAOyorwDsqLAA7KixAOyosgDsqLMA7Ki0AOyotQDsqLYA7Ki3AOyouADsqLkA7Ki6AOyouwDsqLwA7Ki9AOyovgDsqL8A7KmAAOypgQDsqYIA7KmDAOyphADsqYUA7KmGAOyphwDsqYgA7KmJAOypigDsqYsA7KmMAOypjQDsqY4A7KmPAOypkADsqZEA7KmSAOypkwDsqZQA7KmVAOyplgDsqZcA7KmYAOypmQDsqZoA7KmbAOypnADsqZ0A7KmeAOypnwDsqaAA7KmhAOypogDsqaMA7KmkAOyppQDsqaYA7KmnAOypqADsqakA7KmqAOypqwDsqawA7KmtAOyprgDsqa8A7KmwAOypsQDsqbIA7KmzAOyptADsqbUA7Km2AOyptwDsqbgA7Km5AOypugDsqbsA7Km8AOypvQDsqb4A7Km/AOyqgADsqoEA7KqCAOyqgwDsqoQA7KqFAOyqhgDsqocA7KqIAOyqiQDsqooA7KqLAOyqjADsqo0A7KqOAOyqjwDsqpAA7KqRAOyqkgDsqpMA7KqUAOyqlQDsqpYA7KqXAOyqmADsqpkA7KqaAOyqmwDsqpwA7KqdAOyqngDsqp8A7KqgAOyqoQDsqqIA7KqjAOyqpADsqqUA7KqmAOyqpwDsqqgA7KqpAOyqqgDsqqsA7KqsAOyqrQDsqq4A7KqvAOyqsADsqrEA7KqyAOyqswDsqrQA7Kq1AOyqtgDsqrcA7Kq4AOyquQDsqroA7Kq7AOyqvADsqr0A7Kq+AOyqvwDsq4AA7KuBAOyrggDsq4MA7KuEAOyrhQDsq4YA7KuHAOyriADsq4kA7KuKAOyriwDsq4wA7KuNAOyrjgDsq48A7KuQAOyrkQDsq5IA7KuTAOyrlADsq5UA7KuWAOyrlwDsq5gA7KuZAOyrmgDsq5sA7KucAOyrnQDsq54A7KufAOyroADsq6EA7KuiAOyrowDsq6QA7KulAOyrpgDsq6cA7KuoAOyrqQDsq6oA7KurAOyrrADsq60A7KuuAOyrrwDsq7AA7KuxAOyrsgDsq7MA7Ku0AOyrtQDsq7YA7Ku3AOyruADsq7kA7Ku6AOyruwDsq7wA7Ku9AOyrvgDsq78A7KyAAOysgQDsrIIA7KyDAOyshADsrIUA7KyGAOyshwDsrIgA7KyJAOysigDsrIsA7KyMAOysjQDsrI4A7KyPAOyskADsrJEA7KySAOyskwDsrJQA7KyVAOyslgDsrJcA7KyYAOysmQDsrJoA7KybAOysnADsrJ0A7KyeAOysnwDsrKAA7KyhAOysogDsrKMA7KykAOyspQDsrKYA7KynAOysqADsrKkA7KyqAOysqwDsrKwA7KytAOysrgDsrK8A7KywAOyssQDsrLIA7KyzAOystADsrLUA7Ky2AOystwDsrLgA7Ky5AOysugDsrLsA7Ky8AOysvQDsrL4A7Ky/AOytgADsrYEA7K2CAOytgwDsrYQA7K2FAOythgDsrYcA7K2IAOytiQDsrYoA7K2LAOytjADsrY0A7K2OAOytjwDsrZAA7K2RAOytkgDsrZMA7K2UAOytlQDsrZYA7K2XAOytmADsrZkA7K2aAOytmwDsrZwA7K2dAOytngDsrZ8A7K2gAOytoQDsraIA7K2jAOytpADsraUA7K2mAOytpwDsragA7K2pAOytqgDsrasA7K2sAOytrQDsra4A7K2vAOytsADsrbEA7K2yAOytswDsrbQA7K21AOyttgDsrbcA7K24AOytuQDsrboA7K27AOytvADsrb0A7K2+AOytvwDsroAA7K6BAOyuggDsroMA7K6EAOyuhQDsroYA7K6HAOyuiADsrokA7K6KAOyuiwDsrowA7K6NAOyujgDsro8A7K6QAOyukQDsrpIA7K6TAOyulADsrpUA7K6WAOyulwDsrpgA7K6ZAOyumgDsrpsA7K6cAOyunQDsrp4A7K6fAOyuoADsrqEA7K6iAOyuowDsrqQA7K6lAOyupgDsrqcA7K6oAOyuqQDsrqoA7K6rAOyurADsrq0A7K6uAOyurwDsrrAA7K6xAOyusgDsrrMA7K60AOyutQDsrrYA7K63AOyuuADsrrkA7K66AOyuuwDsrrwA7K69AOyuvgDsrr8A7K+AAOyvgQDsr4IA7K+DAOyvhADsr4UA7K+GAOyvhwDsr4gA7K+JAOyvigDsr4sA7K+MAOyvjQDsr44A7K+PAOyvkADsr5EA7K+SAOyvkwDsr5QA7K+VAOyvlgDsr5cA7K+YAOyvmQDsr5oA7K+bAOyvnADsr50A7K+eAOyvnwDsr6AA7K+hAOyvogDsr6MA7K+kAOyvpQDsr6YA7K+nAOyvqADsr6kA7K+qAOyvqwDsr6wA7K+tAOyvrgDsr68A7K+wAOyvsQDsr7IA7K+zAOyvtADsr7UA7K+2AOyvtwDsr7gA7K+5AOyvugDsr7sA7K+8AOyvvQDsr74A7K+/AOywgADssIEA7LCCAOywgwDssIQA7LCFAOywhgDssIcA7LCIAOywiQDssIoA7LCLAOywjADssI0A7LCOAOywjwDssJAA7LCRAOywkgDssJMA7LCUAOywlQDssJYA7LCXAOywmADssJkA7LCaAOywmwDssJwA7LCdAOywngDssJ8A7LCgAOywoQDssKIA7LCjAOywpADssKUA7LCmAOywpwDssKgA7LCpAOywqgDssKsA7LCsAOywrQDssK4A7LCvAOywsADssLEA7LCyAOywswDssLQA7LC1AOywtgDssLcA7LC4AOywuOqzoADssLkA7LC6AOywuwDssLwA7LC9AOywvgDssL8A7LGAAOyxgQDssYIA7LGDAOyxhADssYUA7LGGAOyxhwDssYgA7LGJAOyxigDssYsA7LGMAOyxjQDssY4A7LGPAOyxkADssZEA7LGSAOyxkwDssZQA7LGVAOyxlgDssZcA7LGYAOyxmQDssZoA7LGbAOyxnADssZ0A7LGeAOyxnwDssaAA7LGhAOyxogDssaMA7LGkAOyxpQDssaYA7LGnAOyxqADssakA7LGqAOyxqwDssawA7LGtAOyxrgDssa8A7LGwAOyxsQDssbIA7LGzAOyxtADssbUA7LG2AOyxtwDssbgA7LG5AOyxugDssbsA7LG8AOyxvQDssb4A7LG/AOyygADssoEA7LKCAOyygwDssoQA7LKFAOyyhgDssocA7LKIAOyyiQDssooA7LKLAOyyjADsso0A7LKOAOyyjwDsspAA7LKRAOyykgDsspMA7LKUAOyylQDsspYA7LKXAOyymADsspkA7LKaAOyymwDsspwA7LKdAOyyngDssp8A7LKgAOyyoQDssqIA7LKjAOyypADssqUA7LKmAOyypwDssqgA7LKpAOyyqgDssqsA7LKsAOyyrQDssq4A7LKvAOyysADssrEA7LKyAOyyswDssrQA7LK1AOyytgDssrcA7LK4AOyyuQDssroA7LK7AOyyvADssr0A7LK+AOyyvwDss4AA7LOBAOyzggDss4MA7LOEAOyzhQDss4YA7LOHAOyziADss4kA7LOKAOyziwDss4wA7LONAOyzjgDss48A7LOQAOyzkQDss5IA7LOTAOyzlADss5UA7LOWAOyzlwDss5gA7LOZAOyzmgDss5sA7LOcAOyznQDss54A7LOfAOyzoADss6EA7LOiAOyzowDss6QA7LOlAOyzpgDss6cA7LOoAOyzqQDss6oA7LOrAOyzrADss60A7LOuAOyzrwDss7AA7LOxAOyzsgDss7MA7LO0AOyztQDss7YA7LO3AOyzuADss7kA7LO6AOyzuwDss7wA7LO9AOyzvgDss78A7LSAAOy0gQDstIIA7LSDAOy0hADstIUA7LSGAOy0hwDstIgA7LSJAOy0igDstIsA7LSMAOy0jQDstI4A7LSPAOy0kADstJEA7LSSAOy0kwDstJQA7LSVAOy0lgDstJcA7LSYAOy0mQDstJoA7LSbAOy0nADstJ0A7LSeAOy0nwDstKAA7LShAOy0ogDstKMA7LSkAOy0pQDstKYA7LSnAOy0qADstKkA7LSqAOy0qwDstKwA7LStAOy0rgDstK8A7LSwAOy0sQDstLIA7LSzAOy0tADstLUA7LS2AOy0twDstLgA7LS5AOy0ugDstLsA7LS8AOy0vQDstL4A7LS/AOy1gADstYEA7LWCAOy1gwDstYQA7LWFAOy1hgDstYcA7LWIAOy1iQDstYoA7LWLAOy1jADstY0A7LWOAOy1jwDstZAA7LWRAOy1kgDstZMA7LWUAOy1lQDstZYA7LWXAOy1mADstZkA7LWaAOy1mwDstZwA7LWdAOy1ngDstZ8A7LWgAOy1oQDstaIA7LWjAOy1pADstaUA7LWmAOy1pwDstagA7LWpAOy1qgDstasA7LWsAOy1rQDsta4A7LWvAOy1sADstbEA7LWyAOy1swDstbQA7LW1AOy1tgDstbcA7LW4AOy1uQDstboA7LW7AOy1vADstb0A7LW+AOy1vwDstoAA7LaBAOy2ggDstoMA7LaEAOy2hQDstoYA7LaHAOy2iADstokA7LaKAOy2iwDstowA7LaNAOy2jgDsto8A7LaQAOy2kQDstpIA7LaTAOy2lADstpUA7LaWAOy2lwDstpgA7LaZAOy2mgDstpsA7LacAOy2nQDstp4A7LafAOy2oADstqEA7LaiAOy2owDstqQA7LalAOy2pgDstqcA7LaoAOy2qQDstqoA7LarAOy2rADstq0A7LauAOy2rwDstrAA7LaxAOy2sgDstrMA7La0AOy2tQDstrYA7La3AOy2uADstrkA7La6AOy2uwDstrwA7La9AOy2vgDstr8A7LeAAOy3gQDst4IA7LeDAOy3hADst4UA7LeGAOy3hwDst4gA7LeJAOy3igDst4sA7LeMAOy3jQDst44A7LePAOy3kADst5EA7LeSAOy3kwDst5QA7LeVAOy3lgDst5cA7LeYAOy3mQDst5oA7LebAOy3nADst50A7LeeAOy3nwDst6AA7LehAOy3ogDst6MA7LekAOy3pQDst6YA7LenAOy3qADst6kA7LeqAOy3qwDst6wA7LetAOy3rgDst68A7LewAOy3sQDst7IA7LezAOy3tADst7UA7Le2AOy3twDst7gA7Le5AOy3ugDst7sA7Le8AOy3vQDst74A7Le/AOy4gADsuIEA7LiCAOy4gwDsuIQA7LiFAOy4hgDsuIcA7LiIAOy4iQDsuIoA7LiLAOy4jADsuI0A7LiOAOy4jwDsuJAA7LiRAOy4kgDsuJMA7LiUAOy4lQDsuJYA7LiXAOy4mADsuJkA7LiaAOy4mwDsuJwA7LidAOy4ngDsuJ8A7LigAOy4oQDsuKIA7LijAOy4pADsuKUA7LimAOy4pwDsuKgA7LipAOy4qgDsuKsA7LisAOy4rQDsuK4A7LivAOy4sADsuLEA7LiyAOy4swDsuLQA7Li1AOy4tgDsuLcA7Li4AOy4uQDsuLoA7Li7AOy4vADsuL0A7Li+AOy4vwDsuYAA7LmBAOy5ggDsuYMA7LmEAOy5hQDsuYYA7LmHAOy5iADsuYkA7LmKAOy5iwDsuYwA7LmNAOy5jgDsuY8A7LmQAOy5kQDsuZIA7LmTAOy5lADsuZUA7LmWAOy5lwDsuZgA7LmZAOy5mgDsuZsA7LmcAOy5nQDsuZ4A7LmfAOy5oADsuaEA7LmiAOy5owDsuaQA7LmlAOy5pgDsuacA7LmoAOy5qQDsuaoA7LmrAOy5rADsua0A7LmuAOy5rwDsubAA7LmxAOy5sgDsubMA7Lm0AOy5tQDsubYA7Lm3AOy5uADsubkA7Lm6AOy5uwDsubwA7Lm9AOy5vgDsub8A7LqAAOy6gQDsuoIA7LqDAOy6hADsuoUA7LqGAOy6hwDsuogA7LqJAOy6igDsuosA7LqMAOy6jQDsuo4A7LqPAOy6kADsupEA7LqSAOy6kwDsupQA7LqVAOy6lgDsupcA7LqYAOy6mQDsupoA7LqbAOy6nADsup0A7LqeAOy6nwDsuqAA7LqhAOy6ogDsuqMA7LqkAOy6pQDsuqYA7LqnAOy6qADsuqkA7LqqAOy6qwDsuqwA7LqtAOy6rgDsuq8A7LqwAOy6sQDsurIA7LqzAOy6tADsurUA7Lq2AOy6twDsurgA7Lq5AOy6ugDsursA7Lq8AOy6vQDsur4A7Lq/AOy7gADsu4EA7LuCAOy7gwDsu4QA7LuFAOy7hgDsu4cA7LuIAOy7iQDsu4oA7LuLAOy7jADsu40A7LuOAOy7jwDsu5AA7LuRAOy7kgDsu5MA7LuUAOy7lQDsu5YA7LuXAOy7mADsu5kA7LuaAOy7mwDsu5wA7LudAOy7ngDsu58A7LugAOy7oQDsu6IA7LujAOy7pADsu6UA7LumAOy7pwDsu6gA7LupAOy7qgDsu6sA7LusAOy7rQDsu64A7LuvAOy7sADsu7EA7LuyAOy7swDsu7QA7Lu1AOy7tgDsu7cA7Lu4AOy7uQDsu7oA7Lu7AOy7vADsu70A7Lu+AOy7vwDsvIAA7LyBAOy8ggDsvIMA7LyEAOy8hQDsvIYA7LyHAOy8iADsvIkA7LyKAOy8iwDsvIwA7LyNAOy8jgDsvI8A7LyQAOy8kQDsvJIA7LyTAOy8lADsvJUA7LyWAOy8lwDsvJgA7LyZAOy8mgDsvJsA7LycAOy8nQDsvJ4A7LyfAOy8oADsvKEA7LyiAOy8owDsvKQA7LylAOy8pgDsvKcA7LyoAOy8qQDsvKoA7LyrAOy8rADsvK0A7LyuAOy8rwDsvLAA7LyxAOy8sgDsvLMA7Ly0AOy8tQDsvLYA7Ly3AOy8uADsvLkA7Ly6AOy8uwDsvLwA7Ly9AOy8vgDsvL8A7L2AAOy9gQDsvYIA7L2DAOy9hADsvYUA7L2GAOy9hwDsvYgA7L2JAOy9igDsvYsA7L2MAOy9jQDsvY4A7L2PAOy9kADsvZEA7L2SAOy9kwDsvZQA7L2VAOy9lgDsvZcA7L2YAOy9mQDsvZoA7L2bAOy9nADsvZ0A7L2eAOy9nwDsvaAA7L2hAOy9ogDsvaMA7L2kAOy9pQDsvaYA7L2nAOy9qADsvakA7L2qAOy9qwDsvawA7L2tAOy9rgDsva8A7L2wAOy9sQDsvbIA7L2zAOy9tADsvbUA7L22AOy9twDsvbgA7L25AOy9ugDsvbsA7L28AOy9vQDsvb4A7L2/AOy+gADsvoEA7L6CAOy+gwDsvoQA7L6FAOy+hgDsvocA7L6IAOy+iQDsvooA7L6LAOy+jADsvo0A7L6OAOy+jwDsvpAA7L6RAOy+kgDsvpMA7L6UAOy+lQDsvpYA7L6XAOy+mADsvpkA7L6aAOy+mwDsvpwA7L6dAOy+ngDsvp8A7L6gAOy+oQDsvqIA7L6jAOy+pADsvqUA7L6mAOy+pwDsvqgA7L6pAOy+qgDsvqsA7L6sAOy+rQDsvq4A7L6vAOy+sADsvrEA7L6yAOy+swDsvrQA7L61AOy+tgDsvrcA7L64AOy+uQDsvroA7L67AOy+vADsvr0A7L6+AOy+vwDsv4AA7L+BAOy/ggDsv4MA7L+EAOy/hQDsv4YA7L+HAOy/iADsv4kA7L+KAOy/iwDsv4wA7L+NAOy/jgDsv48A7L+QAOy/kQDsv5IA7L+TAOy/lADsv5UA7L+WAOy/lwDsv5gA7L+ZAOy/mgDsv5sA7L+cAOy/nQDsv54A7L+fAOy/oADsv6EA7L+iAOy/owDsv6QA7L+lAOy/pgDsv6cA7L+oAOy/qQDsv6oA7L+rAOy/rADsv60A7L+uAOy/rwDsv7AA7L+xAOy/sgDsv7MA7L+0AOy/tQDsv7YA7L+3AOy/uADsv7kA7L+6AOy/uwDsv7wA7L+9AOy/vgDsv78A7YCAAO2AgQDtgIIA7YCDAO2AhADtgIUA7YCGAO2AhwDtgIgA7YCJAO2AigDtgIsA7YCMAO2AjQDtgI4A7YCPAO2AkADtgJEA7YCSAO2AkwDtgJQA7YCVAO2AlgDtgJcA7YCYAO2AmQDtgJoA7YCbAO2AnADtgJ0A7YCeAO2AnwDtgKAA7YChAO2AogDtgKMA7YCkAO2ApQDtgKYA7YCnAO2AqADtgKkA7YCqAO2AqwDtgKwA7YCtAO2ArgDtgK8A7YCwAO2AsQDtgLIA7YCzAO2AtADtgLUA7YC2AO2AtwDtgLgA7YC5AO2AugDtgLsA7YC8AO2AvQDtgL4A7YC/AO2BgADtgYEA7YGCAO2BgwDtgYQA7YGFAO2BhgDtgYcA7YGIAO2BiQDtgYoA7YGLAO2BjADtgY0A7YGOAO2BjwDtgZAA7YGRAO2BkgDtgZMA7YGUAO2BlQDtgZYA7YGXAO2BmADtgZkA7YGaAO2BmwDtgZwA7YGdAO2BngDtgZ8A7YGgAO2BoQDtgaIA7YGjAO2BpADtgaUA7YGmAO2BpwDtgagA7YGpAO2BqgDtgasA7YGsAO2BrQDtga4A7YGvAO2BsADtgbEA7YGyAO2BswDtgbQA7YG1AO2BtgDtgbcA7YG4AO2BuQDtgboA7YG7AO2BvADtgb0A7YG+AO2BvwDtgoAA7YKBAO2CggDtgoMA7YKEAO2ChQDtgoYA7YKHAO2CiADtgokA7YKKAO2CiwDtgowA7YKNAO2CjgDtgo8A7YKQAO2CkQDtgpIA7YKTAO2ClADtgpUA7YKWAO2ClwDtgpgA7YKZAO2CmgDtgpsA7YKcAO2CnQDtgp4A7YKfAO2CoADtgqEA7YKiAO2CowDtgqQA7YKlAO2CpgDtgqcA7YKoAO2CqQDtgqoA7YKrAO2CrADtgq0A7YKuAO2CrwDtgrAA7YKxAO2CsgDtgrMA7YK0AO2CtQDtgrYA7YK3AO2CuADtgrkA7YK6AO2CuwDtgrwA7YK9AO2CvgDtgr8A7YOAAO2DgQDtg4IA7YODAO2DhADtg4UA7YOGAO2DhwDtg4gA7YOJAO2DigDtg4sA7YOMAO2DjQDtg44A7YOPAO2DkADtg5EA7YOSAO2DkwDtg5QA7YOVAO2DlgDtg5cA7YOYAO2DmQDtg5oA7YObAO2DnADtg50A7YOeAO2DnwDtg6AA7YOhAO2DogDtg6MA7YOkAO2DpQDtg6YA7YOnAO2DqADtg6kA7YOqAO2DqwDtg6wA7YOtAO2DrgDtg68A7YOwAO2DsQDtg7IA7YOzAO2DtADtg7UA7YO2AO2DtwDtg7gA7YO5AO2DugDtg7sA7YO8AO2DvQDtg74A7YO/AO2EgADthIEA7YSCAO2EgwDthIQA7YSFAO2EhgDthIcA7YSIAO2EiQDthIoA7YSLAO2EjADthI0A7YSOAO2EjwDthJAA7YSRAO2EkgDthJMA7YSUAO2ElQDthJYA7YSXAO2EmADthJkA7YSaAO2EmwDthJwA7YSdAO2EngDthJ8A7YSgAO2EoQDthKIA7YSjAO2EpADthKUA7YSmAO2EpwDthKgA7YSpAO2EqgDthKsA7YSsAO2ErQDthK4A7YSvAO2EsADthLEA7YSyAO2EswDthLQA7YS1AO2EtgDthLcA7YS4AO2EuQDthLoA7YS7AO2EvADthL0A7YS+AO2EvwDthYAA7YWBAO2FggDthYMA7YWEAO2FhQDthYYA7YWHAO2FiADthYkA7YWKAO2FiwDthYwA7YWNAO2FjgDthY8A7YWQAO2FkQDthZIA7YWTAO2FlADthZUA7YWWAO2FlwDthZgA7YWZAO2FmgDthZsA7YWcAO2FnQDthZ4A7YWfAO2FoADthaEA7YWiAO2FowDthaQA7YWlAO2FpgDthacA7YWoAO2FqQDthaoA7YWrAO2FrADtha0A7YWuAO2FrwDthbAA7YWxAO2FsgDthbMA7YW0AO2FtQDthbYA7YW3AO2FuADthbkA7YW6AO2FuwDthbwA7YW9AO2FvgDthb8A7YaAAO2GgQDthoIA7YaDAO2GhADthoUA7YaGAO2GhwDthogA7YaJAO2GigDthosA7YaMAO2GjQDtho4A7YaPAO2GkADthpEA7YaSAO2GkwDthpQA7YaVAO2GlgDthpcA7YaYAO2GmQDthpoA7YabAO2GnADthp0A7YaeAO2GnwDthqAA7YahAO2GogDthqMA7YakAO2GpQDthqYA7YanAO2GqADthqkA7YaqAO2GqwDthqwA7YatAO2GrgDthq8A7YawAO2GsQDthrIA7YazAO2GtADthrUA7Ya2AO2GtwDthrgA7Ya5AO2GugDthrsA7Ya8AO2GvQDthr4A7Ya/AO2HgADth4EA7YeCAO2HgwDth4QA7YeFAO2HhgDth4cA7YeIAO2HiQDth4oA7YeLAO2HjADth40A7YeOAO2HjwDth5AA7YeRAO2HkgDth5MA7YeUAO2HlQDth5YA7YeXAO2HmADth5kA7YeaAO2HmwDth5wA7YedAO2HngDth58A7YegAO2HoQDth6IA7YejAO2HpADth6UA7YemAO2HpwDth6gA7YepAO2HqgDth6sA7YesAO2HrQDth64A7YevAO2HsADth7EA7YeyAO2HswDth7QA7Ye1AO2HtgDth7cA7Ye4AO2HuQDth7oA7Ye7AO2HvADth70A7Ye+AO2HvwDtiIAA7YiBAO2IggDtiIMA7YiEAO2IhQDtiIYA7YiHAO2IiADtiIkA7YiKAO2IiwDtiIwA7YiNAO2IjgDtiI8A7YiQAO2IkQDtiJIA7YiTAO2IlADtiJUA7YiWAO2IlwDtiJgA7YiZAO2ImgDtiJsA7YicAO2InQDtiJ4A7YifAO2IoADtiKEA7YiiAO2IowDtiKQA7YilAO2IpgDtiKcA7YioAO2IqQDtiKoA7YirAO2IrADtiK0A7YiuAO2IrwDtiLAA7YixAO2IsgDtiLMA7Yi0AO2ItQDtiLYA7Yi3AO2IuADtiLkA7Yi6AO2IuwDtiLwA7Yi9AO2IvgDtiL8A7YmAAO2JgQDtiYIA7YmDAO2JhADtiYUA7YmGAO2JhwDtiYgA7YmJAO2JigDtiYsA7YmMAO2JjQDtiY4A7YmPAO2JkADtiZEA7YmSAO2JkwDtiZQA7YmVAO2JlgDtiZcA7YmYAO2JmQDtiZoA7YmbAO2JnADtiZ0A7YmeAO2JnwDtiaAA7YmhAO2JogDtiaMA7YmkAO2JpQDtiaYA7YmnAO2JqADtiakA7YmqAO2JqwDtiawA7YmtAO2JrgDtia8A7YmwAO2JsQDtibIA7YmzAO2JtADtibUA7Ym2AO2JtwDtibgA7Ym5AO2JugDtibsA7Ym8AO2JvQDtib4A7Ym/AO2KgADtioEA7YqCAO2KgwDtioQA7YqFAO2KhgDtiocA7YqIAO2KiQDtiooA7YqLAO2KjADtio0A7YqOAO2KjwDtipAA7YqRAO2KkgDtipMA7YqUAO2KlQDtipYA7YqXAO2KmADtipkA7YqaAO2KmwDtipwA7YqdAO2KngDtip8A7YqgAO2KoQDtiqIA7YqjAO2KpADtiqUA7YqmAO2KpwDtiqgA7YqpAO2KqgDtiqsA7YqsAO2KrQDtiq4A7YqvAO2KsADtirEA7YqyAO2KswDtirQA7Yq1AO2KtgDtircA7Yq4AO2KuQDtiroA7Yq7AO2KvADtir0A7Yq+AO2KvwDti4AA7YuBAO2LggDti4MA7YuEAO2LhQDti4YA7YuHAO2LiADti4kA7YuKAO2LiwDti4wA7YuNAO2LjgDti48A7YuQAO2LkQDti5IA7YuTAO2LlADti5UA7YuWAO2LlwDti5gA7YuZAO2LmgDti5sA7YucAO2LnQDti54A7YufAO2LoADti6EA7YuiAO2LowDti6QA7YulAO2LpgDti6cA7YuoAO2LqQDti6oA7YurAO2LrADti60A7YuuAO2LrwDti7AA7YuxAO2LsgDti7MA7Yu0AO2LtQDti7YA7Yu3AO2LuADti7kA7Yu6AO2LuwDti7wA7Yu9AO2LvgDti78A7YyAAO2MgQDtjIIA7YyDAO2MhADtjIUA7YyGAO2MhwDtjIgA7YyJAO2MigDtjIsA7YyMAO2MjQDtjI4A7YyPAO2MkADtjJEA7YySAO2MkwDtjJQA7YyVAO2MlgDtjJcA7YyYAO2MmQDtjJoA7YybAO2MnADtjJ0A7YyeAO2MnwDtjKAA7YyhAO2MogDtjKMA7YykAO2MpQDtjKYA7YynAO2MqADtjKkA7YyqAO2MqwDtjKwA7YytAO2MrgDtjK8A7YywAO2MsQDtjLIA7YyzAO2MtADtjLUA7Yy2AO2MtwDtjLgA7Yy5AO2MugDtjLsA7Yy8AO2MvQDtjL4A7Yy/AO2NgADtjYEA7Y2CAO2NgwDtjYQA7Y2FAO2NhgDtjYcA7Y2IAO2NiQDtjYoA7Y2LAO2NjADtjY0A7Y2OAO2NjwDtjZAA7Y2RAO2NkgDtjZMA7Y2UAO2NlQDtjZYA7Y2XAO2NmADtjZkA7Y2aAO2NmwDtjZwA7Y2dAO2NngDtjZ8A7Y2gAO2NoQDtjaIA7Y2jAO2NpADtjaUA7Y2mAO2NpwDtjagA7Y2pAO2NqgDtjasA7Y2sAO2NrQDtja4A7Y2vAO2NsADtjbEA7Y2yAO2NswDtjbQA7Y21AO2NtgDtjbcA7Y24AO2NuQDtjboA7Y27AO2NvADtjb0A7Y2+AO2NvwDtjoAA7Y6BAO2OggDtjoMA7Y6EAO2OhQDtjoYA7Y6HAO2OiADtjokA7Y6KAO2OiwDtjowA7Y6NAO2OjgDtjo8A7Y6QAO2OkQDtjpIA7Y6TAO2OlADtjpUA7Y6WAO2OlwDtjpgA7Y6ZAO2OmgDtjpsA7Y6cAO2OnQDtjp4A7Y6fAO2OoADtjqEA7Y6iAO2OowDtjqQA7Y6lAO2OpgDtjqcA7Y6oAO2OqQDtjqoA7Y6rAO2OrADtjq0A7Y6uAO2OrwDtjrAA7Y6xAO2OsgDtjrMA7Y60AO2OtQDtjrYA7Y63AO2OuADtjrkA7Y66AO2OuwDtjrwA7Y69AO2OvgDtjr8A7Y+AAO2PgQDtj4IA7Y+DAO2PhADtj4UA7Y+GAO2PhwDtj4gA7Y+JAO2PigDtj4sA7Y+MAO2PjQDtj44A7Y+PAO2PkADtj5EA7Y+SAO2PkwDtj5QA7Y+VAO2PlgDtj5cA7Y+YAO2PmQDtj5oA7Y+bAO2PnADtj50A7Y+eAO2PnwDtj6AA7Y+hAO2PogDtj6MA7Y+kAO2PpQDtj6YA7Y+nAO2PqADtj6kA7Y+qAO2PqwDtj6wA7Y+tAO2PrgDtj68A7Y+wAO2PsQDtj7IA7Y+zAO2PtADtj7UA7Y+2AO2PtwDtj7gA7Y+5AO2PugDtj7sA7Y+8AO2PvQDtj74A7Y+/AO2QgADtkIEA7ZCCAO2QgwDtkIQA7ZCFAO2QhgDtkIcA7ZCIAO2QiQDtkIoA7ZCLAO2QjADtkI0A7ZCOAO2QjwDtkJAA7ZCRAO2QkgDtkJMA7ZCUAO2QlQDtkJYA7ZCXAO2QmADtkJkA7ZCaAO2QmwDtkJwA7ZCdAO2QngDtkJ8A7ZCgAO2QoQDtkKIA7ZCjAO2QpADtkKUA7ZCmAO2QpwDtkKgA7ZCpAO2QqgDtkKsA7ZCsAO2QrQDtkK4A7ZCvAO2QsADtkLEA7ZCyAO2QswDtkLQA7ZC1AO2QtgDtkLcA7ZC4AO2QuQDtkLoA7ZC7AO2QvADtkL0A7ZC+AO2QvwDtkYAA7ZGBAO2RggDtkYMA7ZGEAO2RhQDtkYYA7ZGHAO2RiADtkYkA7ZGKAO2RiwDtkYwA7ZGNAO2RjgDtkY8A7ZGQAO2RkQDtkZIA7ZGTAO2RlADtkZUA7ZGWAO2RlwDtkZgA7ZGZAO2RmgDtkZsA7ZGcAO2RnQDtkZ4A7ZGfAO2RoADtkaEA7ZGiAO2RowDtkaQA7ZGlAO2RpgDtkacA7ZGoAO2RqQDtkaoA7ZGrAO2RrADtka0A7ZGuAO2RrwDtkbAA7ZGxAO2RsgDtkbMA7ZG0AO2RtQDtkbYA7ZG3AO2RuADtkbkA7ZG6AO2RuwDtkbwA7ZG9AO2RvgDtkb8A7ZKAAO2SgQDtkoIA7ZKDAO2ShADtkoUA7ZKGAO2ShwDtkogA7ZKJAO2SigDtkosA7ZKMAO2SjQDtko4A7ZKPAO2SkADtkpEA7ZKSAO2SkwDtkpQA7ZKVAO2SlgDtkpcA7ZKYAO2SmQDtkpoA7ZKbAO2SnADtkp0A7ZKeAO2SnwDtkqAA7ZKhAO2SogDtkqMA7ZKkAO2SpQDtkqYA7ZKnAO2SqADtkqkA7ZKqAO2SqwDtkqwA7ZKtAO2SrgDtkq8A7ZKwAO2SsQDtkrIA7ZKzAO2StADtkrUA7ZK2AO2StwDtkrgA7ZK5AO2SugDtkrsA7ZK8AO2SvQDtkr4A7ZK/AO2TgADtk4EA7ZOCAO2TgwDtk4QA7ZOFAO2ThgDtk4cA7ZOIAO2TiQDtk4oA7ZOLAO2TjADtk40A7ZOOAO2TjwDtk5AA7ZORAO2TkgDtk5MA7ZOUAO2TlQDtk5YA7ZOXAO2TmADtk5kA7ZOaAO2TmwDtk5wA7ZOdAO2TngDtk58A7ZOgAO2ToQDtk6IA7ZOjAO2TpADtk6UA7ZOmAO2TpwDtk6gA7ZOpAO2TqgDtk6sA7ZOsAO2TrQDtk64A7ZOvAO2TsADtk7EA7ZOyAO2TswDtk7QA7ZO1AO2TtgDtk7cA7ZO4AO2TuQDtk7oA7ZO7AO2TvADtk70A7ZO+AO2TvwDtlIAA7ZSBAO2UggDtlIMA7ZSEAO2UhQDtlIYA7ZSHAO2UiADtlIkA7ZSKAO2UiwDtlIwA7ZSNAO2UjgDtlI8A7ZSQAO2UkQDtlJIA7ZSTAO2UlADtlJUA7ZSWAO2UlwDtlJgA7ZSZAO2UmgDtlJsA7ZScAO2UnQDtlJ4A7ZSfAO2UoADtlKEA7ZSiAO2UowDtlKQA7ZSlAO2UpgDtlKcA7ZSoAO2UqQDtlKoA7ZSrAO2UrADtlK0A7ZSuAO2UrwDtlLAA7ZSxAO2UsgDtlLMA7ZS0AO2UtQDtlLYA7ZS3AO2UuADtlLkA7ZS6AO2UuwDtlLwA7ZS9AO2UvgDtlL8A7ZWAAO2VgQDtlYIA7ZWDAO2VhADtlYUA7ZWGAO2VhwDtlYgA7ZWJAO2VigDtlYsA7ZWMAO2VjQDtlY4A7ZWPAO2VkADtlZEA7ZWSAO2VkwDtlZQA7ZWVAO2VlgDtlZcA7ZWYAO2VmQDtlZoA7ZWbAO2VnADtlZ0A7ZWeAO2VnwDtlaAA7ZWhAO2VogDtlaMA7ZWkAO2VpQDtlaYA7ZWnAO2VqADtlakA7ZWqAO2VqwDtlawA7ZWtAO2VrgDtla8A7ZWwAO2VsQDtlbIA7ZWzAO2VtADtlbUA7ZW2AO2VtwDtlbgA7ZW5AO2VugDtlbsA7ZW8AO2VvQDtlb4A7ZW/AO2WgADtloEA7ZaCAO2WgwDtloQA7ZaFAO2WhgDtlocA7ZaIAO2WiQDtlooA7ZaLAO2WjADtlo0A7ZaOAO2WjwDtlpAA7ZaRAO2WkgDtlpMA7ZaUAO2WlQDtlpYA7ZaXAO2WmADtlpkA7ZaaAO2WmwDtlpwA7ZadAO2WngDtlp8A7ZagAO2WoQDtlqIA7ZajAO2WpADtlqUA7ZamAO2WpwDtlqgA7ZapAO2WqgDtlqsA7ZasAO2WrQDtlq4A7ZavAO2WsADtlrEA7ZayAO2WswDtlrQA7Za1AO2WtgDtlrcA7Za4AO2WuQDtlroA7Za7AO2WvADtlr0A7Za+AO2WvwDtl4AA7ZeBAO2XggDtl4MA7ZeEAO2XhQDtl4YA7ZeHAO2XiADtl4kA7ZeKAO2XiwDtl4wA7ZeNAO2XjgDtl48A7ZeQAO2XkQDtl5IA7ZeTAO2XlADtl5UA7ZeWAO2XlwDtl5gA7ZeZAO2XmgDtl5sA7ZecAO2XnQDtl54A7ZefAO2XoADtl6EA7ZeiAO2XowDtl6QA7ZelAO2XpgDtl6cA7ZeoAO2XqQDtl6oA7ZerAO2XrADtl60A7ZeuAO2XrwDtl7AA7ZexAO2XsgDtl7MA7Ze0AO2XtQDtl7YA7Ze3AO2XuADtl7kA7Ze6AO2XuwDtl7wA7Ze9AO2XvgDtl78A7ZiAAO2YgQDtmIIA7ZiDAO2YhADtmIUA7ZiGAO2YhwDtmIgA7ZiJAO2YigDtmIsA7ZiMAO2YjQDtmI4A7ZiPAO2YkADtmJEA7ZiSAO2YkwDtmJQA7ZiVAO2YlgDtmJcA7ZiYAO2YmQDtmJoA7ZibAO2YnADtmJ0A7ZieAO2YnwDtmKAA7ZihAO2YogDtmKMA7ZikAO2YpQDtmKYA7ZinAO2YqADtmKkA7ZiqAO2YqwDtmKwA7ZitAO2YrgDtmK8A7ZiwAO2YsQDtmLIA7ZizAO2YtADtmLUA7Zi2AO2YtwDtmLgA7Zi5AO2YugDtmLsA7Zi8AO2YvQDtmL4A7Zi/AO2ZgADtmYEA7ZmCAO2ZgwDtmYQA7ZmFAO2ZhgDtmYcA7ZmIAO2ZiQDtmYoA7ZmLAO2ZjADtmY0A7ZmOAO2ZjwDtmZAA7ZmRAO2ZkgDtmZMA7ZmUAO2ZlQDtmZYA7ZmXAO2ZmADtmZkA7ZmaAO2ZmwDtmZwA7ZmdAO2ZngDtmZ8A7ZmgAO2ZoQDtmaIA7ZmjAO2ZpADtmaUA7ZmmAO2ZpwDtmagA7ZmpAO2ZqgDtmasA7ZmsAO2ZrQDtma4A7ZmvAO2ZsADtmbEA7ZmyAO2ZswDtmbQA7Zm1AO2ZtgDtmbcA7Zm4AO2ZuQDtmboA7Zm7AO2ZvADtmb0A7Zm+AO2ZvwDtmoAA7ZqBAO2aggDtmoMA7ZqEAO2ahQDtmoYA7ZqHAO2aiADtmokA7ZqKAO2aiwDtmowA7ZqNAO2ajgDtmo8A7ZqQAO2akQDtmpIA7ZqTAO2alADtmpUA7ZqWAO2alwDtmpgA7ZqZAO2amgDtmpsA7ZqcAO2anQDtmp4A7ZqfAO2aoADtmqEA7ZqiAO2aowDtmqQA7ZqlAO2apgDtmqcA7ZqoAO2aqQDtmqoA7ZqrAO2arADtmq0A7ZquAO2arwDtmrAA7ZqxAO2asgDtmrMA7Zq0AO2atQDtmrYA7Zq3AO2auADtmrkA7Zq6AO2auwDtmrwA7Zq9AO2avgDtmr8A7ZuAAO2bgQDtm4IA7ZuDAO2bhADtm4UA7ZuGAO2bhwDtm4gA7ZuJAO2bigDtm4sA7ZuMAO2bjQDtm44A7ZuPAO2bkADtm5EA7ZuSAO2bkwDtm5QA7ZuVAO2blgDtm5cA7ZuYAO2bmQDtm5oA7ZubAO2bnADtm50A7ZueAO2bnwDtm6AA7ZuhAO2bogDtm6MA7ZukAO2bpQDtm6YA7ZunAO2bqADtm6kA7ZuqAO2bqwDtm6wA7ZutAO2brgDtm68A7ZuwAO2bsQDtm7IA7ZuzAO2btADtm7UA7Zu2AO2btwDtm7gA7Zu5AO2bugDtm7sA7Zu8AO2bvQDtm74A7Zu/AO2cgADtnIEA7ZyCAO2cgwDtnIQA7ZyFAO2chgDtnIcA7ZyIAO2ciQDtnIoA7ZyLAO2cjADtnI0A7ZyOAO2cjwDtnJAA7ZyRAO2ckgDtnJMA7ZyUAO2clQDtnJYA7ZyXAO2cmADtnJkA7ZyaAO2cmwDtnJwA7ZydAO2cngDtnJ8A7ZygAO2coQDtnKIA7ZyjAO2cpADtnKUA7ZymAO2cpwDtnKgA7ZypAO2cqgDtnKsA7ZysAO2crQDtnK4A7ZyvAO2csADtnLEA7ZyyAO2cswDtnLQA7Zy1AO2ctgDtnLcA7Zy4AO2cuQDtnLoA7Zy7AO2cvADtnL0A7Zy+AO2cvwDtnYAA7Z2BAO2dggDtnYMA7Z2EAO2dhQDtnYYA7Z2HAO2diADtnYkA7Z2KAO2diwDtnYwA7Z2NAO2djgDtnY8A7Z2QAO2dkQDtnZIA7Z2TAO2dlADtnZUA7Z2WAO2dlwDtnZgA7Z2ZAO2dmgDtnZsA7Z2cAO2dnQDtnZ4A7Z2fAO2doADtnaEA7Z2iAO2dowDtnaQA7Z2lAO2dpgDtnacA7Z2oAO2dqQDtnaoA7Z2rAO2drADtna0A7Z2uAO2drwDtnbAA7Z2xAO2dsgDtnbMA7Z20AO2dtQDtnbYA7Z23AO2duADtnbkA7Z26AO2duwDtnbwA7Z29AO2dvgDtnb8A7Z6AAO2egQDtnoIA7Z6DAO2ehADtnoUA7Z6GAO2ehwDtnogA7Z6JAO2eigDtnosA7Z6MAO2ejQDtno4A7Z6PAO2ekADtnpEA7Z6SAO2ekwDtnpQA7Z6VAO2elgDtnpcA7Z6YAO2emQDtnpoA7Z6bAO2enADtnp0A7Z6eAO2enwDtnqAA7Z6hAO2eogDtnqMA8JGCmgDwkYKcAPCRgqsA8JGErgDwkYSvAPCRjYsA8JGNjADwkZK7APCRkrwA8JGSvgDwkZa6APCRlrsA8JGkuADwnYWX8J2FpQDwnYWY8J2FpQDwnYWY8J2FpfCdha4A8J2FmPCdhaXwnYWvAPCdhZjwnYWl8J2FsADwnYWY8J2FpfCdhbEA8J2FmPCdhaXwnYWyAPCdhrnwnYWlAPCdhrnwnYWl8J2FrgDwnYa58J2FpfCdha8A8J2GuvCdhaUA8J2GuvCdhaXwnYWuAPCdhrrwnYWl8J2FrwDwoISiAPCglJwA8KCUpQDwoJWLAPCgmLoA8KCghADwoKOeAPCgqKwA8KCtowDwoZOkAPChmqgA8KGbqgDwoaeIAPChrJgA8KG0iwDwobekAPCht6YA8KKGgwDwooafAPCijLEA8KKblADwoqGEAPCioYoA8KKsjADwoq+xAPCjgIoA8KOKuADwo42fAPCjjpMA8KOOnADwo4+DAPCjj5UA8KORrQDwo5qjAPCjoqcA8KOqjQDwo6u6APCjsrwA8KO0ngDwo7uRAPCjvZ4A8KO+jgDwpImjAPCki64A8KSOqwDwpJiIAPCknLUA8KSglADwpLC2APCkspIA8KS+oQDwpL64APClgYQA8KWDsgDwpYOzAPClhJkA8KWEswDwpYmJAPClkJ0A8KWYpgDwpZqaAPClm4UA8KWlvADwpaqnAPClrqsA8KWygADwpbOQAPClvoYA8KaHmgDwpoioAPCmiYcA8KaLmQDwpoy+APCmk5oA8KaUowDwppaoAPCmnqcA8KaetQDwpqy8APCmsLYA8KazlQDwprWrAPCmvKwA8Ka+sQDwp4OSAPCnj4oA8KeZpwDwp6KuAPCnpaYA8KeyqADwp7uTAPCnvK8A8KiXkgDwqJetAPConK4A8KivugDwqLW3APCphYUA8KmHnwDwqYiaAPCpkIoA8KmSlgDwqZa2APCprLAA8KqDjgDwqoSFAPCqiI4A8KqKkQDwqo6SAPCqmIAA" - }, - { - "type": "Replace", - "pattern": { - "Regex": " {2,}" - }, - "content": "▁" - } - ] - }, - "pre_tokenizer": { - "type": "Metaspace", - "replacement": "▁", - "prepend_scheme": "always", - "split": true - }, - "post_processor": null, - "decoder": { - "type": "Metaspace", - "replacement": "▁", - "prepend_scheme": "always", - "split": true - }, - "model": { - "type": "BPE", - "dropout": null, - "unk_token": "", - "continuing_subword_prefix": null, - "end_of_word_suffix": null, - "fuse_unk": true, - "byte_fallback": true, - "ignore_merges": false, - "vocab": { - "": 0, - "▁t": 1, - "▁th": 2, - "▁a": 3, - "in": 4, - "re": 5, - "▁the": 6, - "▁w": 7, - "▁s": 8, - "▁o": 9, - "er": 10, - "ou": 11, - "at": 12, - "nd": 13, - "it": 14, - "▁h": 15, - "▁c": 16, - "▁b": 17, - "is": 18, - "en": 19, - "on": 20, - "ing": 21, - "▁f": 22, - "▁to": 23, - "▁m": 24, - "es": 25, - "▁p": 26, - "or": 27, - "an": 28, - "▁d": 29, - "ll": 30, - "▁I": 31, - "ed": 32, - "▁and": 33, - "▁l": 34, - "▁of": 35, - "▁in": 36, - "▁y": 37, - "ar": 38, - "▁g": 39, - "▁you": 40, - "as": 41, - "om": 42, - "▁n": 43, - "ve": 44, - "▁that": 45, - "le": 46, - "ic": 47, - "us": 48, - "ow": 49, - "et": 50, - "al": 51, - "▁e": 52, - "ut": 53, - "▁it": 54, - "ot": 55, - "▁be": 56, - "▁T": 57, - "ion": 58, - "▁is": 59, - "▁wh": 60, - "▁re": 61, - "▁on": 62, - "▁we": 63, - "ent": 64, - "▁A": 65, - "ay": 66, - "▁ha": 67, - "▁Th": 68, - "id": 69, - "▁S": 70, - "ac": 71, - "gh": 72, - "ver": 73, - "ke": 74, - "▁for": 75, - "im": 76, - "ly": 77, - "ur": 78, - "ld": 79, - "▁he": 80, - "▁st": 81, - "all": 82, - "ro": 83, - "st": 84, - "se": 85, - "ct": 86, - "ith": 87, - "ir": 88, - "am": 89, - "▁this": 90, - "if": 91, - "▁W": 92, - "oo": 93, - "ri": 94, - "▁was": 95, - "ght": 96, - "▁u": 97, - "▁with": 98, - "ad": 99, - "ch": 100, - "▁se": 101, - "▁k": 102, - "▁an": 103, - "▁The": 104, - "▁li": 105, - "▁do": 106, - "▁B": 107, - "▁have": 108, - "▁as": 109, - "th": 110, - "▁are": 111, - "▁sh": 112, - "ust": 113, - "ce": 114, - "ally": 115, - "ill": 116, - "▁H": 117, - "▁j": 118, - "ter": 119, - "▁go": 120, - "▁And": 121, - "ation": 122, - "▁C": 123, - "▁so": 124, - "ome": 125, - "▁not": 126, - "op": 127, - "il": 128, - "ore": 129, - "▁ne": 130, - "▁can": 131, - "▁me": 132, - "▁at": 133, - "ould": 134, - "ant": 135, - "▁M": 136, - "▁like": 137, - "ere": 138, - "▁they": 139, - "ra": 140, - "ers": 141, - "▁ab": 142, - "▁de": 143, - "▁kn": 144, - "ge": 145, - "▁Y": 146, - "▁ch": 147, - "ul": 148, - "pp": 149, - "▁or": 150, - "▁al": 151, - "▁con": 152, - "▁com": 153, - "ess": 154, - "▁su": 155, - "out": 156, - "▁your": 157, - "▁So": 158, - "ate": 159, - "▁one": 160, - "▁all": 161, - "▁ex": 162, - "est": 163, - "▁fr": 164, - "▁just": 165, - "▁pro": 166, - "▁know": 167, - "▁O": 168, - "ain": 169, - "▁but": 170, - "ol": 171, - "ive": 172, - "▁v": 173, - "use": 174, - "very": 175, - "art": 176, - "qu": 177, - "▁my": 178, - "el": 179, - "▁N": 180, - "nt": 181, - "▁It": 182, - "▁what": 183, - "ab": 184, - "▁P": 185, - "▁wor": 186, - "▁out": 187, - "▁there": 188, - "▁up": 189, - "um": 190, - "▁from": 191, - "pe": 192, - "▁tw": 193, - "▁r": 194, - "and": 195, - "ight": 196, - "ort": 197, - "un": 198, - "▁L": 199, - "ist": 200, - "▁about": 201, - "ide": 202, - "ig": 203, - "ake": 204, - "▁D": 205, - "em": 206, - "os": 207, - "king": 208, - "rou": 209, - "ind": 210, - "our": 211, - "res": 212, - "▁We": 213, - "▁get": 214, - "▁E": 215, - "▁G": 216, - "ack": 217, - "▁le": 218, - "ity": 219, - "od": 220, - "▁F": 221, - "ard": 222, - "▁pl": 223, - "▁our": 224, - "▁int": 225, - "ment": 226, - "▁will": 227, - "ies": 228, - "▁by": 229, - "ink": 230, - "ca": 231, - "▁if": 232, - "red": 233, - "her": 234, - "ie": 235, - "▁us": 236, - "▁some": 237, - "▁don": 238, - "ven": 239, - "ood": 240, - "ast": 241, - "▁R": 242, - "▁his": 243, - "▁tim": 244, - "▁tr": 245, - "▁more": 246, - "ich": 247, - "ous": 248, - "ame": 249, - "▁going": 250, - "▁had": 251, - "▁them": 252, - "ook": 253, - "▁pe": 254, - "▁Wh": 255, - "▁You": 256, - "▁But": 257, - "ine": 258, - "▁here": 259, - "▁would": 260, - "cause": 261, - "right": 262, - "so": 263, - "ost": 264, - "ure": 265, - "▁has": 266, - "ect": 267, - "▁think": 268, - "▁fe": 269, - "ong": 270, - "▁see": 271, - "▁when": 272, - "▁who": 273, - "▁were": 274, - "▁really": 275, - "▁their": 276, - "▁want": 277, - "one": 278, - "ople": 279, - "▁then": 280, - "▁time": 281, - "▁sa": 282, - "ap": 283, - "▁te": 284, - "▁He": 285, - "▁ye": 286, - "ck": 287, - "▁her": 288, - "▁thing": 289, - "▁right": 290, - "▁which": 291, - "itt": 292, - "ice": 293, - "act": 294, - "▁people": 295, - "ty": 296, - "▁two": 297, - "▁J": 298, - "▁im": 299, - "ther": 300, - "ci": 301, - "ose": 302, - "▁cl": 303, - "▁qu": 304, - "▁man": 305, - "▁also": 306, - "ree": 307, - "▁en": 308, - "ud": 309, - "▁how": 310, - "reat": 311, - "ak": 312, - "hing": 313, - "ag": 314, - "▁any": 315, - "ff": 316, - "ace": 317, - "per": 318, - "▁because": 319, - "▁very": 320, - "own": 321, - "▁ad": 322, - "▁act": 323, - "▁been": 324, - "▁now": 325, - "▁ag": 326, - "▁into": 327, - "▁comp": 328, - "ars": 329, - "ions": 330, - "are": 331, - "ite": 332, - "iv": 333, - "▁these": 334, - "ays": 335, - "ep": 336, - "▁This": 337, - "▁she": 338, - "ans": 339, - "ah": 340, - "een": 341, - "▁over": 342, - "ry": 343, - "▁lo": 344, - "age": 345, - "▁pr": 346, - "▁sp": 347, - "ue": 348, - "▁co": 349, - "ick": 350, - "ber": 351, - "▁did": 352, - "ip": 353, - "ach": 354, - "▁back": 355, - "▁no": 356, - "▁cont": 357, - "▁other": 358, - "▁every": 359, - "pt": 360, - "▁need": 361, - "▁him": 362, - "▁U": 363, - "▁In": 364, - "▁work": 365, - "irst": 366, - "▁part": 367, - "▁look": 368, - "ittle": 369, - "ble": 370, - "iz": 371, - "▁un": 372, - "▁make": 373, - "omet": 374, - "nder": 375, - "ish": 376, - "na": 377, - "▁little": 378, - "▁off": 379, - "▁than": 380, - "▁got": 381, - "ually": 382, - "▁per": 383, - "▁good": 384, - "▁way": 385, - "▁could": 386, - "▁ac": 387, - "▁imp": 388, - "able": 389, - "▁where": 390, - "iff": 391, - "▁That": 392, - "▁res": 393, - "ount": 394, - "pl": 395, - "ance": 396, - "▁first": 397, - "▁ro": 398, - "▁pre": 399, - "ass": 400, - "▁say": 401, - "int": 402, - "ated": 403, - "ire": 404, - "uch": 405, - "ase": 406, - "▁somet": 407, - "ound": 408, - "▁down": 409, - "▁diff": 410, - "sel": 411, - "▁gu": 412, - "▁am": 413, - "ress": 414, - "▁lot": 415, - "ence": 416, - "▁dis": 417, - "orm": 418, - "ix": 419, - "▁po": 420, - "ving": 421, - "enty": 422, - "▁K": 423, - "▁spe": 424, - "und": 425, - "he": 426, - "▁much": 427, - "▁ar": 428, - "round": 429, - "▁app": 430, - "co": 431, - "ark": 432, - "▁new": 433, - "ater": 434, - "ult": 435, - "end": 436, - "▁even": 437, - "▁start": 438, - "ations": 439, - "rough": 440, - "ile": 441, - "fter": 442, - "▁well": 443, - "be": 444, - "▁They": 445, - "▁three": 446, - "ign": 447, - "ild": 448, - "▁said": 449, - "ough": 450, - "ang": 451, - "▁too": 452, - "ade": 453, - "▁bl": 454, - "ens": 455, - "▁inc": 456, - "ia": 457, - "▁those": 458, - "▁mo": 459, - "▁take": 460, - "▁through": 461, - "▁fl": 462, - "▁kind": 463, - "▁things": 464, - "▁bet": 465, - "▁only": 466, - "▁St": 467, - "▁let": 468, - "cess": 469, - "▁Ch": 470, - "ary": 471, - "vel": 472, - "▁If": 473, - "xt": 474, - "other": 475, - "av": 476, - "ical": 477, - "ord": 478, - "▁again": 479, - "▁something": 480, - "onna": 481, - "fore": 482, - "▁may": 483, - "ting": 484, - "▁bu": 485, - "▁differe": 486, - "urn": 487, - "▁gonna": 488, - "▁does": 489, - "uct": 490, - "og": 491, - "▁twenty": 492, - "▁gr": 493, - "▁Ye": 494, - "wn": 495, - "▁should": 496, - "▁comm": 497, - "ition": 498, - "▁under": 499, - "▁hel": 500, - "ory": 501, - "▁fo": 502, - "▁use": 503, - "igh": 504, - "ife": 505, - "▁actually": 506, - "▁tal": 507, - "▁call": 508, - "ents": 509, - "ious": 510, - "ull": 511, - "▁There": 512, - "▁Yeah": 513, - "▁most": 514, - "▁ke": 515, - "ors": 516, - "ved": 517, - "ys": 518, - "▁sc": 519, - "▁happ": 520, - "ope": 521, - "▁help": 522, - "atch": 523, - "▁What": 524, - "▁rem": 525, - "ple": 526, - "▁Now": 527, - "▁br": 528, - "ool": 529, - "oth": 530, - "▁four": 531, - "self": 532, - "▁str": 533, - "ne": 534, - "thing": 535, - "▁put": 536, - "ial": 537, - "▁great": 538, - "ail": 539, - "ub": 540, - "ning": 541, - "▁sm": 542, - "▁feel": 543, - "▁five": 544, - "ody": 545, - "undred": 546, - "iss": 547, - "ank": 548, - "get": 549, - "aking": 550, - "▁many": 551, - "▁hundred": 552, - "▁years": 553, - "▁being": 554, - "▁come": 555, - "▁mean": 556, - "ily": 557, - "▁different": 558, - "▁after": 559, - "▁ser": 560, - "▁show": 561, - "form": 562, - "ful": 563, - "oy": 564, - "▁six": 565, - "▁vide": 566, - "▁V": 567, - "▁its": 568, - "▁point": 569, - "▁day": 570, - "▁des": 571, - "ons": 572, - "▁bit": 573, - "▁bel": 574, - "▁before": 575, - "▁aw": 576, - "▁end": 577, - "▁Oh": 578, - "▁still": 579, - "ath": 580, - "▁long": 581, - "▁'": 582, - "ise": 583, - "ob": 584, - "day": 585, - "▁add": 586, - "ft": 587, - "ves": 588, - "ces": 589, - "ady": 590, - "▁cr": 591, - "▁around": 592, - "▁try": 593, - "les": 594, - "vers": 595, - "kay": 596, - "ian": 597, - "ates": 598, - "▁find": 599, - "ward": 600, - "▁As": 601, - "▁eight": 602, - "lic": 603, - "▁same": 604, - "▁pos": 605, - "▁em": 606, - "▁made": 607, - "▁supp": 608, - "▁life": 609, - "▁Be": 610, - "pect": 611, - "▁dec": 612, - "▁play": 613, - "ange": 614, - "▁att": 615, - "▁pers": 616, - "ways": 617, - "▁high": 618, - "▁hand": 619, - "▁next": 620, - "▁cons": 621, - "▁own": 622, - "▁inv": 623, - "ower": 624, - "▁ind": 625, - "ert": 626, - "ng": 627, - "ave": 628, - "▁year": 629, - "▁big": 630, - "ating": 631, - "▁world": 632, - "▁rel": 633, - "▁sure": 634, - "▁tra": 635, - "ew": 636, - "ered": 637, - "▁fin": 638, - "▁Well": 639, - "▁sl": 640, - "▁doing": 641, - "bs": 642, - "▁set": 643, - "▁rec": 644, - "ual": 645, - "cial": 646, - "▁ph": 647, - "erm": 648, - "▁love": 649, - "ph": 650, - "▁real": 651, - "▁last": 652, - "ict": 653, - "▁bo": 654, - "▁ra": 655, - "ible": 656, - "▁wr": 657, - "mer": 658, - "▁count": 659, - "ities": 660, - "▁always": 661, - "inet": 662, - "ments": 663, - "uc": 664, - "▁might": 665, - "▁inter": 666, - "▁video": 667, - "gin": 668, - "▁tell": 669, - "▁never": 670, - "vent": 671, - "▁import": 672, - "ied": 673, - "▁sy": 674, - "▁How": 675, - "ically": 676, - "ought": 677, - "▁thir": 678, - "▁rep": 679, - "ks": 680, - "ib": 681, - "▁fam": 682, - "ject": 683, - "▁bas": 684, - "▁She": 685, - "▁give": 686, - "akes": 687, - "▁ninet": 688, - "▁reg": 689, - "▁min": 690, - "▁op": 691, - "▁def": 692, - "▁didn": 693, - "te": 694, - "▁cour": 695, - "▁why": 696, - "▁ent": 697, - "▁place": 698, - "▁ins": 699, - "▁car": 700, - "ather": 701, - "▁person": 702, - "ular": 703, - "▁inst": 704, - "▁prod": 705, - "lect": 706, - "▁Al": 707, - "▁today": 708, - "▁bec": 709, - "▁sur": 710, - "▁All": 711, - "▁another": 712, - "▁bus": 713, - "▁keep": 714, - "ell": 715, - "ese": 716, - "riend": 717, - "▁quest": 718, - "▁talk": 719, - "als": 720, - "ings": 721, - "▁mon": 722, - "cond": 723, - "old": 724, - "▁acc": 725, - "▁la": 726, - "▁num": 727, - "ident": 728, - "▁che": 729, - "iness": 730, - "▁turn": 731, - "▁ear": 732, - "▁No": 733, - "ousand": 734, - "▁better": 735, - "ific": 736, - "▁loo": 737, - "▁gl": 738, - "oc": 739, - "▁important": 740, - "ited": 741, - "▁An": 742, - "▁thousand": 743, - "ility": 744, - "llow": 745, - "▁used": 746, - "▁gen": 747, - "▁sim": 748, - "li": 749, - "▁happen": 750, - "▁Un": 751, - "▁Let": 752, - "air": 753, - "ock": 754, - "ably": 755, - "gg": 756, - "▁watch": 757, - "▁For": 758, - "▁sw": 759, - "ren": 760, - "ute": 761, - "ever": 762, - "▁pol": 763, - "▁sch": 764, - "▁When": 765, - "▁such": 766, - "▁fif": 767, - "▁home": 768, - "▁cle": 769, - "▁contin": 770, - "ouse": 771, - "▁friend": 772, - "uring": 773, - "▁Okay": 774, - "gr": 775, - "▁able": 776, - "▁stud": 777, - "▁eff": 778, - "hip": 779, - "body": 780, - "▁top": 781, - "ness": 782, - "▁exper": 783, - "▁pret": 784, - "▁both": 785, - "▁done": 786, - "cri": 787, - "▁mark": 788, - "▁while": 789, - "▁old": 790, - "ros": 791, - "ont": 792, - "▁second": 793, - "ative": 794, - "▁thought": 795, - "▁best": 796, - "▁found": 797, - "iew": 798, - "▁belie": 799, - "▁each": 800, - "erest": 801, - "▁tri": 802, - "▁eas": 803, - "▁ca": 804, - "▁fact": 805, - "▁care": 806, - "▁fun": 807, - "atter": 808, - "ures": 809, - "▁head": 810, - "▁lear": 811, - "▁water": 812, - "▁hard": 813, - "▁few": 814, - "▁side": 815, - "ween": 816, - "▁exp": 817, - "▁away": 818, - "its": 819, - "▁ext": 820, - "lud": 821, - "▁run": 822, - "▁trans": 823, - "ince": 824, - "▁sk": 825, - "▁open": 826, - "cus": 827, - "▁between": 828, - "▁called": 829, - "▁wee": 830, - "▁pretty": 831, - "ason": 832, - "▁far": 833, - "ember": 834, - "omm": 835, - "▁interest": 836, - "any": 837, - "ner": 838, - "uff": 839, - "▁pres": 840, - "▁cur": 841, - "▁child": 842, - "ee": 843, - "▁toget": 844, - "▁together": 845, - "olog": 846, - "▁God": 847, - "ond": 848, - "▁char": 849, - "▁looking": 850, - "stem": 851, - "az": 852, - "cent": 853, - "▁ob": 854, - "▁ass": 855, - "land": 856, - "▁doesn": 857, - "▁business": 858, - "▁course": 859, - "▁ten": 860, - "ps": 861, - "arch": 862, - "ced": 863, - "ms": 864, - "ize": 865, - "nce": 866, - "▁ref": 867, - "▁name": 868, - "ross": 869, - "▁grow": 870, - "oney": 871, - "▁went": 872, - "ics": 873, - "teen": 874, - "▁cou": 875, - "▁prob": 876, - "▁ret": 877, - "▁guys": 878, - "▁came": 879, - "ash": 880, - "led": 881, - "▁Eur": 882, - "ues": 883, - "▁ide": 884, - "gan": 885, - "▁everything": 886, - "▁getting": 887, - "▁ask": 888, - "▁cor": 889, - "▁build": 890, - "▁sign": 891, - "▁small": 892, - "uck": 893, - "▁el": 894, - "▁col": 895, - "▁Is": 896, - "ational": 897, - "stand": 898, - "cy": 899, - "▁conf": 900, - "der": 901, - "▁bre": 902, - "▁cap": 903, - "▁mod": 904, - "ets": 905, - "ike": 906, - "▁number": 907, - "▁comple": 908, - "ertain": 909, - "▁ever": 910, - "▁coll": 911, - "▁hum": 912, - "▁Europe": 913, - "▁cre": 914, - "▁met": 915, - "▁exam": 916, - "▁move": 917, - "▁pass": 918, - "▁left": 919, - "▁system": 920, - "▁includ": 921, - "▁Thank": 922, - "cept": 923, - "▁wom": 924, - "▁product": 925, - "ten": 926, - "▁rest": 927, - "▁probably": 928, - "▁dri": 929, - "▁Do": 930, - "▁gener": 931, - "▁anything": 932, - "▁lar": 933, - "▁My": 934, - "▁school": 935, - "▁lead": 936, - "▁sub": 937, - "▁ty": 938, - "▁plan": 939, - "▁seem": 940, - "▁whole": 941, - "irect": 942, - "▁light": 943, - "▁must": 944, - "▁mom": 945, - "▁opp": 946, - "▁support": 947, - "▁family": 948, - "ices": 949, - "amp": 950, - "▁proble": 951, - "▁dr": 952, - "ready": 953, - "▁using": 954, - "ense": 955, - "▁prov": 956, - "ush": 957, - "ax": 958, - "▁power": 959, - "▁Re": 960, - "alth": 961, - "▁ev": 962, - "▁stand": 963, - "▁war": 964, - "ts": 965, - "▁": 966, - "e": 967, - "t": 968, - "o": 969, - "a": 970, - "n": 971, - "i": 972, - "s": 973, - "r": 974, - "h": 975, - "l": 976, - "d": 977, - "u": 978, - "c": 979, - "m": 980, - "y": 981, - "g": 982, - "w": 983, - "f": 984, - "p": 985, - ".": 986, - "b": 987, - ",": 988, - "v": 989, - "k": 990, - "'": 991, - "I": 992, - "T": 993, - "A": 994, - "S": 995, - "x": 996, - "W": 997, - "j": 998, - "B": 999, - "C": 1000, - "H": 1001, - "?": 1002, - "M": 1003, - "O": 1004, - "Y": 1005, - "N": 1006, - "P": 1007, - "E": 1008, - "q": 1009, - "L": 1010, - "D": 1011, - "z": 1012, - "G": 1013, - "F": 1014, - "R": 1015, - "!": 1016, - "J": 1017, - "U": 1018, - "K": 1019, - "V": 1020, - "Q": 1021, - "Z": 1022, - "X": 1023 - }, - "merges": [ - "▁ t", - "▁t h", - "▁ th", - "▁ a", - "i n", - "r e", - "▁th e", - "▁t he", - "▁ w", - "▁ s", - "▁ o", - "e r", - "o u", - "a t", - "n d", - "i t", - "▁ h", - "▁ c", - "▁ b", - "i s", - "e n", - "o n", - "in g", - "i ng", - "▁ f", - "▁t o", - "▁ m", - "e s", - "▁ p", - "o r", - "a n", - "▁ d", - "l l", - "▁ I", - "e d", - "▁an d", - "▁a nd", - "▁ and", - "▁ l", - "▁o f", - "▁ in", - "▁ y", - "a r", - "▁ g", - "▁y ou", - "a s", - "o m", - "▁ n", - "v e", - "▁th at", - "l e", - "i c", - "u s", - "o w", - "e t", - "a l", - "▁ e", - "u t", - "▁ it", - "o t", - "▁b e", - "▁ be", - "▁ T", - "i on", - "▁ is", - "▁w h", - "▁r e", - "▁ re", - "▁o n", - "▁ on", - "▁w e", - "en t", - "e nt", - "▁ A", - "a y", - "▁h a", - "▁T h", - "i d", - "▁ S", - "a c", - "g h", - "ve r", - "v er", - "k e", - "▁fo r", - "▁f or", - "i m", - "l y", - "u r", - "l d", - "▁h e", - "▁ he", - "▁s t", - "▁ st", - "al l", - "a ll", - "r o", - "s t", - "s e", - "c t", - "it h", - "i th", - "i r", - "a m", - "▁th is", - "i f", - "▁ W", - "o o", - "r i", - "▁w as", - "gh t", - "▁ u", - "▁w ith", - "a d", - "c h", - "▁s e", - "▁ se", - "▁ k", - "▁a n", - "▁ an", - "▁Th e", - "▁T he", - "▁l i", - "▁ li", - "▁d o", - "▁ B", - "▁ha ve", - "▁h ave", - "▁a s", - "▁ as", - "t h", - "▁ar e", - "▁a re", - "▁ are", - "▁s h", - "us t", - "u st", - "c e", - "all y", - "al ly", - "il l", - "i ll", - "▁ H", - "▁ j", - "te r", - "t er", - "▁g o", - "▁An d", - "▁A nd", - "at ion", - "▁ C", - "▁s o", - "▁ so", - "om e", - "▁no t", - "▁n ot", - "o p", - "i l", - "or e", - "o re", - "▁n e", - "▁ ne", - "▁ca n", - "▁c an", - "▁m e", - "▁a t", - "▁ at", - "ou ld", - "an t", - "a nt", - "▁ M", - "▁li ke", - "▁l ike", - "er e", - "e re", - "▁the y", - "r a", - "er s", - "▁a b", - "▁ ab", - "▁d e", - "▁k n", - "g e", - "▁ Y", - "▁c h", - "▁ ch", - "u l", - "p p", - "▁o r", - "▁ or", - "▁a l", - "▁ al", - "▁co n", - "▁c on", - "▁co m", - "▁c om", - "es s", - "▁s u", - "ou t", - "o ut", - "▁you r", - "▁y our", - "▁S o", - "at e", - "a te", - "▁on e", - "▁o ne", - "▁ one", - "▁al l", - "▁a ll", - "▁ all", - "▁e x", - "es t", - "e st", - "▁f r", - "▁j ust", - "▁pr o", - "▁p ro", - "▁kn ow", - "▁ O", - "a in", - "▁bu t", - "▁b ut", - "o l", - "iv e", - "i ve", - "▁ v", - "us e", - "u se", - "ver y", - "ve ry", - "ar t", - "q u", - "▁m y", - "e l", - "▁ N", - "n t", - "▁I t", - "▁wh at", - "a b", - "▁ P", - "▁w or", - "▁o ut", - "▁ out", - "▁the re", - "▁th ere", - "▁u p", - "u m", - "▁fr om", - "p e", - "▁t w", - "▁ r", - "an d", - "a nd", - "igh t", - "i ght", - "or t", - "u n", - "▁ L", - "is t", - "i st", - "▁ab out", - "id e", - "i g", - "ak e", - "a ke", - "▁ D", - "e m", - "o s", - "k ing", - "ro u", - "r ou", - "in d", - "i nd", - "ou r", - "o ur", - "re s", - "r es", - "▁W e", - "▁g et", - "▁ get", - "▁ E", - "▁ G", - "ac k", - "a ck", - "▁l e", - "▁ le", - "it y", - "i ty", - "o d", - "▁ F", - "ar d", - "▁p l", - "▁ pl", - "▁o ur", - "▁ our", - "▁in t", - "▁ int", - "m ent", - "▁w ill", - "ie s", - "i es", - "▁b y", - "in k", - "c a", - "▁ if", - "re d", - "r ed", - "he r", - "h er", - "i e", - "▁u s", - "▁ us", - "▁s ome", - "▁do n", - "▁d on", - "ve n", - "v en", - "oo d", - "o od", - "as t", - "a st", - "▁ R", - "▁h is", - "▁t im", - "▁t r", - "▁mo re", - "▁m ore", - "ic h", - "i ch", - "ou s", - "o us", - "am e", - "▁go ing", - "▁ha d", - "▁h ad", - "▁the m", - "▁th em", - "oo k", - "▁p e", - "▁ pe", - "▁W h", - "▁Y ou", - "▁B ut", - "in e", - "i ne", - "▁her e", - "▁he re", - "▁h ere", - "▁w ould", - "ca use", - "ri ght", - "r ight", - "s o", - "os t", - "o st", - "ur e", - "u re", - "▁ha s", - "▁h as", - "e ct", - "▁th ink", - "▁f e", - "on g", - "o ng", - "▁se e", - "▁s ee", - "▁wh en", - "▁wh o", - "▁we re", - "▁w ere", - "▁real ly", - "▁re ally", - "▁the ir", - "▁w ant", - "on e", - "o ne", - "op le", - "o ple", - "▁the n", - "▁th en", - "▁tim e", - "▁s a", - "a p", - "▁t e", - "▁ te", - "▁H e", - "▁y e", - "c k", - "▁he r", - "▁h er", - "▁ her", - "▁th ing", - "▁t hing", - "▁ thing", - "▁r ight", - "▁ right", - "▁wh ich", - "it t", - "ic e", - "i ce", - "ac t", - "a ct", - "▁pe ople", - "t y", - "▁tw o", - "▁ J", - "▁ im", - "th er", - "t her", - "c i", - "os e", - "o se", - "▁c l", - "▁ qu", - "▁m an", - "▁al so", - "re e", - "r ee", - "▁e n", - "▁ en", - "u d", - "▁h ow", - "re at", - "a k", - "h ing", - "a g", - "▁an y", - "▁ any", - "f f", - "ac e", - "a ce", - "pe r", - "p er", - "▁be cause", - "▁ very", - "ow n", - "o wn", - "▁a d", - "▁ ad", - "▁ac t", - "▁a ct", - "▁ act", - "▁be en", - "▁b een", - "▁no w", - "▁n ow", - "▁a g", - "▁ ag", - "▁int o", - "▁com p", - "ar s", - "ion s", - "i ons", - "ar e", - "a re", - "it e", - "i te", - "i v", - "▁the se", - "▁th ese", - "ay s", - "a ys", - "e p", - "▁Th is", - "▁sh e", - "▁s he", - "an s", - "a h", - "ee n", - "e en", - "▁o ver", - "r y", - "▁l o", - "ag e", - "a ge", - "▁p r", - "▁s p", - "u e", - "▁c o", - "▁ co", - "ic k", - "i ck", - "be r", - "b er", - "▁d id", - "i p", - "ac h", - "a ch", - "▁b ack", - "▁n o", - "▁con t", - "▁co nt", - "▁c ont", - "▁o ther", - "▁ other", - "▁ever y", - "▁e very", - "p t", - "▁ne ed", - "▁h im", - "▁ U", - "▁I n", - "▁wor k", - "ir st", - "▁p art", - "▁loo k", - "▁l ook", - "itt le", - "b le", - "i z", - "▁u n", - "▁ un", - "▁m ake", - "ome t", - "om et", - "nd er", - "n der", - "is h", - "n a", - "▁l ittle", - "▁of f", - "▁o ff", - "▁th an", - "▁go t", - "▁g ot", - "ual ly", - "u ally", - "▁pe r", - "▁p er", - "▁ per", - "▁go od", - "▁g ood", - "▁w ay", - "▁cou ld", - "▁c ould", - "▁a c", - "▁ ac", - "▁im p", - "ab le", - "a ble", - "▁wh ere", - "if f", - "i ff", - "▁Th at", - "▁re s", - "▁r es", - "▁ res", - "ou nt", - "p l", - "an ce", - "a nce", - "▁f irst", - "▁r o", - "▁ ro", - "▁pr e", - "▁p re", - "as s", - "▁sa y", - "▁s ay", - "in t", - "i nt", - "ate d", - "at ed", - "ir e", - "i re", - "uc h", - "u ch", - "as e", - "a se", - "▁some t", - "▁s omet", - "ou nd", - "o und", - "▁do wn", - "▁d own", - "▁d iff", - "se l", - "s el", - "▁g u", - "▁a m", - "▁ am", - "res s", - "r ess", - "▁lo t", - "▁l ot", - "en ce", - "e nce", - "▁d is", - "or m", - "i x", - "▁p o", - "v ing", - "ent y", - "en ty", - "▁ K", - "▁sp e", - "▁s pe", - "un d", - "u nd", - "h e", - "▁m uch", - "▁a r", - "▁ ar", - "rou nd", - "ro und", - "r ound", - "▁a pp", - "c o", - "ar k", - "▁ne w", - "▁n ew", - "ate r", - "at er", - "a ter", - "ul t", - "en d", - "e nd", - "▁ev en", - "▁e ven", - "▁st art", - "ation s", - "at ions", - "rou gh", - "r ough", - "il e", - "i le", - "ft er", - "f ter", - "▁we ll", - "▁w ell", - "b e", - "▁The y", - "▁th ree", - "ig n", - "il d", - "i ld", - "▁sa id", - "ou gh", - "an g", - "a ng", - "▁to o", - "▁t oo", - "ad e", - "▁b l", - "en s", - "▁in c", - "i a", - "▁th ose", - "▁m o", - "▁t ake", - "▁th rough", - "▁f l", - "▁k ind", - "▁thing s", - "▁th ings", - "▁be t", - "▁b et", - "▁on ly", - "▁S t", - "▁le t", - "▁l et", - "ces s", - "c ess", - "▁C h", - "ar y", - "a ry", - "ve l", - "v el", - "▁I f", - "x t", - "oth er", - "ot her", - "o ther", - "a v", - "ic al", - "or d", - "▁ag ain", - "▁somet hing", - "▁some thing", - "on na", - "f ore", - "▁m ay", - "t ing", - "▁b u", - "▁diff ere", - "ur n", - "▁g onna", - "▁do es", - "uc t", - "u ct", - "o g", - "▁tw enty", - "▁g r", - "▁ gr", - "▁Y e", - "w n", - "▁sh ould", - "▁com m", - "▁c omm", - "it ion", - "▁un der", - "▁u nder", - "▁he l", - "▁h el", - "or y", - "o ry", - "▁f o", - "▁us e", - "▁u se", - "▁ use", - "ig h", - "i gh", - "if e", - "▁act ually", - "▁t al", - "▁ca ll", - "▁c all", - "ent s", - "en ts", - "i ous", - "ul l", - "u ll", - "▁The re", - "▁Th ere", - "▁Ye ah", - "▁mo st", - "▁m ost", - "▁k e", - "▁ ke", - "or s", - "ve d", - "v ed", - "y s", - "▁s c", - "▁ha pp", - "op e", - "o pe", - "▁hel p", - "at ch", - "▁Wh at", - "▁re m", - "▁r em", - "pl e", - "p le", - "▁No w", - "▁N ow", - "▁b r", - "oo l", - "o ol", - "ot h", - "o th", - "▁fo ur", - "▁f our", - "sel f", - "▁st r", - "n e", - "th ing", - "t hing", - "▁p ut", - "ia l", - "i al", - "▁g reat", - "a il", - "u b", - "n ing", - "▁s m", - "▁fe el", - "▁f ive", - "od y", - "und red", - "is s", - "an k", - "ge t", - "g et", - "ak ing", - "a king", - "▁man y", - "▁m any", - "▁h undred", - "▁year s", - "▁ye ars", - "▁be ing", - "▁com e", - "▁c ome", - "▁me an", - "il y", - "i ly", - "▁differe nt", - "▁a fter", - "▁se r", - "▁s er", - "▁sh ow", - "f orm", - "f ul", - "o y", - "▁s ix", - "▁v ide", - "▁ V", - "▁it s", - "▁ its", - "▁po int", - "▁d ay", - "▁ day", - "▁de s", - "▁d es", - "on s", - "▁b it", - "▁be l", - "▁b el", - "▁be fore", - "▁a w", - "▁en d", - "▁e nd", - "▁ end", - "▁O h", - "▁st ill", - "at h", - "a th", - "▁lo ng", - "▁l ong", - "▁ '", - "is e", - "i se", - "o b", - "d ay", - "▁ad d", - "f t", - "ve s", - "v es", - "ce s", - "c es", - "ad y", - "▁c r", - "▁ar ound", - "▁a round", - "▁tr y", - "▁t ry", - "le s", - "l es", - "ver s", - "v ers", - "k ay", - "ia n", - "i an", - "ate s", - "at es", - "▁fin d", - "▁f ind", - "w ard", - "▁A s", - "▁e ight", - "li c", - "l ic", - "▁s ame", - "▁po s", - "▁p os", - "▁e m", - "▁ em", - "▁m ade", - "▁su pp", - "▁l ife", - "▁B e", - "pe ct", - "p ect", - "▁de c", - "▁pl ay", - "ang e", - "an ge", - "▁at t", - "▁per s", - "▁p ers", - "w ays", - "▁h igh", - "▁ha nd", - "▁h and", - "▁ne xt", - "▁con s", - "▁c ons", - "▁o wn", - "▁ own", - "▁in v", - "ow er", - "▁in d", - "▁ ind", - "er t", - "n g", - "av e", - "a ve", - "▁ye ar", - "▁b ig", - "at ing", - "a ting", - "▁wor ld", - "▁re l", - "▁r el", - "▁sur e", - "▁su re", - "▁s ure", - "▁tr a", - "▁t ra", - "e w", - "ere d", - "er ed", - "e red", - "▁f in", - "▁We ll", - "▁W ell", - "▁s l", - "▁do ing", - "b s", - "▁se t", - "▁s et", - "▁re c", - "u al", - "ci al", - "c ial", - "▁p h", - "▁ ph", - "er m", - "▁lo ve", - "p h", - "▁re al", - "▁la st", - "▁l ast", - "ic t", - "i ct", - "▁b o", - "▁r a", - "▁ ra", - "ib le", - "i ble", - "▁w r", - "m er", - "▁cou nt", - "▁c ount", - "it ies", - "▁al ways", - "ine t", - "in et", - "ment s", - "m ents", - "u c", - "▁m ight", - "▁int er", - "▁in ter", - "▁vide o", - "g in", - "▁te ll", - "▁t ell", - "▁ne ver", - "▁n ever", - "ven t", - "ve nt", - "v ent", - "▁imp ort", - "ie d", - "i ed", - "▁s y", - "▁H ow", - "ical ly", - "ic ally", - "ough t", - "ou ght", - "▁th ir", - "▁re p", - "▁r ep", - "k s", - "i b", - "▁f am", - "j ect", - "▁b as", - "▁S he", - "▁g ive", - "ake s", - "ak es", - "▁n inet", - "▁re g", - "▁m in", - "▁o p", - "▁ op", - "▁de f", - "▁did n", - "t e", - "▁cou r", - "▁co ur", - "▁c our", - "▁wh y", - "▁en t", - "▁e nt", - "▁ ent", - "▁pl ace", - "▁in s", - "▁ca r", - "▁c ar", - "ath er", - "at her", - "a ther", - "▁pers on", - "ul ar", - "▁ins t", - "▁in st", - "▁pro d", - "▁pr od", - "le ct", - "l ect", - "▁A l", - "▁to day", - "▁be c", - "▁su r", - "▁s ur", - "▁Al l", - "▁A ll", - "▁an other", - "▁bu s", - "▁b us", - "▁ke ep", - "el l", - "e ll", - "es e", - "e se", - "ri end", - "▁qu est", - "▁tal k", - "al s", - "ing s", - "▁mo n", - "▁m on", - "co nd", - "c ond", - "ol d", - "o ld", - "▁ac c", - "▁l a", - "▁n um", - "ide nt", - "id ent", - "▁ch e", - "▁c he", - "in ess", - "i ness", - "▁t urn", - "▁e ar", - "▁N o", - "ous and", - "▁bet ter", - "if ic", - "▁lo o", - "▁l oo", - "▁g l", - "o c", - "▁import ant", - "ite d", - "it ed", - "▁A n", - "▁th ousand", - "il ity", - "ll ow", - "▁use d", - "▁us ed", - "▁g en", - "▁s im", - "l i", - "▁happ en", - "▁U n", - "▁L et", - "a ir", - "oc k", - "o ck", - "ab ly", - "g g", - "▁w atch", - "▁F or", - "▁s w", - "re n", - "r en", - "ut e", - "u te", - "e ver", - "▁po l", - "▁p ol", - "▁sc h", - "▁s ch", - "▁Wh en", - "▁su ch", - "▁s uch", - "▁f if", - "▁h ome", - "▁cl e", - "▁c le", - "▁cont in", - "ous e", - "ou se", - "o use", - "▁f riend", - "ur ing", - "▁O kay", - "g r", - "▁ab le", - "▁a ble", - "▁ able", - "▁st ud", - "▁e ff", - "h ip", - "b ody", - "▁to p", - "▁t op", - "n ess", - "▁exp er", - "▁ex per", - "▁pre t", - "▁pr et", - "▁bo th", - "▁b oth", - "▁don e", - "▁do ne", - "▁d one", - "c ri", - "▁m ark", - "▁wh ile", - "▁o ld", - "▁ old", - "ro s", - "r os", - "on t", - "o nt", - "▁se cond", - "at ive", - "▁th ought", - "▁be st", - "▁b est", - "▁fo und", - "▁f ound", - "ie w", - "i ew", - "▁bel ie", - "▁e ach", - "ere st", - "er est", - "▁tr i", - "▁t ri", - "▁e as", - "▁c a", - "▁ ca", - "▁f act", - "▁car e", - "▁ca re", - "▁c are", - "▁f un", - "at ter", - "ure s", - "ur es", - "u res", - "▁he ad", - "▁le ar", - "▁w ater", - "▁h ard", - "▁fe w", - "▁f ew", - "▁s ide", - "w een", - "▁ex p", - "▁aw ay", - "it s", - "i ts", - "▁ex t", - "▁e xt", - "l ud", - "▁r un", - "▁tr ans", - "in ce", - "i nce", - "▁s k", - "▁op en", - "c us", - "▁bet ween", - "▁call ed", - "▁we e", - "▁w ee", - "▁pret ty", - "as on", - "▁f ar", - "em ber", - "om m", - "▁inter est", - "▁int erest", - "an y", - "ne r", - "n er", - "u ff", - "▁pre s", - "▁pr es", - "▁p res", - "▁c ur", - "▁ch ild", - "e e", - "▁to get", - "▁toget her", - "ol og", - "▁G od", - "on d", - "o nd", - "▁ch ar", - "▁look ing", - "▁loo king", - "st em", - "a z", - "ce nt", - "c ent", - "▁o b", - "▁ ob", - "▁as s", - "▁ ass", - "l and", - "▁does n", - "▁bus iness", - "▁cour se", - "▁te n", - "▁t en", - "▁ ten", - "p s", - "ar ch", - "ce d", - "c ed", - "m s", - "iz e", - "n ce", - "▁re f", - "▁n ame", - "ros s", - "▁gr ow", - "one y", - "▁we nt", - "▁w ent", - "ic s", - "te en", - "t een", - "▁co u", - "▁c ou", - "▁pro b", - "▁pr ob", - "▁re t", - "▁r et", - "▁gu ys", - "▁c ame", - "as h", - "le d", - "l ed", - "▁E ur", - "ue s", - "u es", - "▁ ide", - "g an", - "▁every thing", - "▁get ting", - "▁as k", - "▁co r", - "▁c or", - "▁bu ild", - "▁s ign", - "▁sm all", - "uc k", - "u ck", - "▁e l", - "▁ el", - "▁co l", - "▁c ol", - "▁I s", - "ation al", - "st and", - "c y", - "▁con f", - "d er", - "▁br e", - "▁b re", - "▁ca p", - "▁c ap", - "▁mo d", - "▁m od", - "et s", - "e ts", - "i ke", - "▁num ber", - "▁comp le", - "▁com ple", - "ert ain", - "▁ev er", - "▁e ver", - "▁ ever", - "▁col l", - "▁co ll", - "▁h um", - "▁Eur ope", - "▁cr e", - "▁c re", - "▁me t", - "▁m et", - "▁ex am", - "▁mo ve", - "▁p ass", - "▁le ft", - "▁sy stem", - "▁inc lud", - "▁Th ank", - "ce pt", - "▁w om", - "▁prod uct", - "te n", - "t en", - "▁res t", - "▁re st", - "▁r est", - "▁prob ably", - "▁dr i", - "▁d ri", - "▁D o", - "▁gen er", - "▁any thing", - "▁la r", - "▁l ar", - "▁M y", - "▁sch ool", - "▁le ad", - "▁su b", - "▁s ub", - "▁t y", - "▁ ty", - "▁pl an", - "▁see m", - "▁se em", - "▁who le", - "ire ct", - "ir ect", - "▁li ght", - "▁l ight", - "▁m ust", - "▁mo m", - "▁m om", - "▁op p", - "▁o pp", - "▁supp ort", - "▁fam ily", - "ice s", - "ic es", - "i ces", - "am p", - "▁prob le", - "▁pro ble", - "▁d r", - "re ady", - "▁us ing", - "ens e", - "en se", - "▁pro v", - "us h", - "a x", - "▁p ower", - "▁R e", - "al th", - "▁e v", - "▁st and", - "▁ stand", - "▁w ar", - "t s" - ] - } -} \ No newline at end of file diff --git a/tokenizer.model b/tokenizer.model deleted file mode 100644 index 3d4ed9e73fbf2fa4f8eb5bb8198e7d6659b7b871..0000000000000000000000000000000000000000 --- a/tokenizer.model +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:4b97fbc15390829e3b94069fdeccf0d053561d090e12d94789e4e71c531f93fc -size 251875 diff --git a/tokenizer_config.json b/tokenizer_config.json deleted file mode 100644 index d9eaa0fb393ab6e3ac6890fe99efb6191cded97c..0000000000000000000000000000000000000000 --- a/tokenizer_config.json +++ /dev/null @@ -1,27 +0,0 @@ -{ - "added_tokens_decoder": { - "0": { - "content": "", - "lstrip": false, - "normalized": false, - "rstrip": false, - "single_word": false, - "special": true - }, - "1024": { - "content": "", - "lstrip": false, - "normalized": false, - "rstrip": false, - "single_word": false, - "special": true - } - }, - "clean_up_tokenization_spaces": false, - "extra_special_tokens": {}, - "model_max_length": 1000000000000000019884624838656, - "pad_token": "", - "processor_class": "ParakeetProcessor", - "tokenizer_class": "PreTrainedTokenizer", - "unk_token": "" -} diff --git a/uv.lock b/uv.lock deleted file mode 100644 index 606b928b0c4b7a4e8b04de38fed541173ea5eb0d..0000000000000000000000000000000000000000 --- a/uv.lock +++ /dev/null @@ -1,5879 +0,0 @@ -version = 1 -requires-python = ">=3.10" -resolution-markers = [ - "python_full_version >= '3.14' and sys_platform == 'linux'", - "python_full_version == '3.13.*' and sys_platform == 'linux'", - "python_full_version >= '3.14' and sys_platform != 'linux'", - "python_full_version == '3.13.*' and sys_platform != 'linux'", - "python_full_version == '3.12.*' and sys_platform == 'linux'", - "python_full_version == '3.12.*' and sys_platform != 'linux'", - "python_full_version == '3.11.*' and sys_platform == 'linux'", - "python_full_version == '3.11.*' and sys_platform != 'linux'", - "python_full_version < '3.11' and sys_platform == 'linux'", - "python_full_version < '3.11' and sys_platform != 'linux'", -] - -[[package]] -name = "absl-py" -version = "2.3.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/10/2a/c93173ffa1b39c1d0395b7e842bbdc62e556ca9d8d3b5572926f3e4ca752/absl_py-2.3.1.tar.gz", hash = "sha256:a97820526f7fbfd2ec1bce83f3f25e3a14840dac0d8e02a0b71cd75db3f77fc9", size = 116588 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/8f/aa/ba0014cc4659328dc818a28827be78e6d97312ab0cb98105a770924dc11e/absl_py-2.3.1-py3-none-any.whl", hash = "sha256:eeecf07f0c2a93ace0772c92e596ace6d3d3996c042b2128459aaae2a76de11d", size = 135811 }, -] - -[[package]] -name = "accelerate" -version = "1.12.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "huggingface-hub" }, - { name = "numpy" }, - { name = "packaging" }, - { name = "psutil" }, - { name = "pyyaml" }, - { name = "safetensors" }, - { name = "torch" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/4a/8e/ac2a9566747a93f8be36ee08532eb0160558b07630a081a6056a9f89bf1d/accelerate-1.12.0.tar.gz", hash = "sha256:70988c352feb481887077d2ab845125024b2a137a5090d6d7a32b57d03a45df6", size = 398399 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/9f/d2/c581486aa6c4fbd7394c23c47b83fa1a919d34194e16944241daf9e762dd/accelerate-1.12.0-py3-none-any.whl", hash = "sha256:3e2091cd341423207e2f084a6654b1efcd250dc326f2a37d6dde446e07cabb11", size = 380935 }, -] - -[[package]] -name = "aiohappyeyeballs" -version = "2.6.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/26/30/f84a107a9c4331c14b2b586036f40965c128aa4fee4dda5d3d51cb14ad54/aiohappyeyeballs-2.6.1.tar.gz", hash = "sha256:c3f9d0113123803ccadfdf3f0faa505bc78e6a72d1cc4806cbd719826e943558", size = 22760 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/0f/15/5bf3b99495fb160b63f95972b81750f18f7f4e02ad051373b669d17d44f2/aiohappyeyeballs-2.6.1-py3-none-any.whl", hash = "sha256:f349ba8f4b75cb25c99c5c2d84e997e485204d2902a9597802b0371f09331fb8", size = 15265 }, -] - -[[package]] -name = "aiohttp" -version = "3.13.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "aiohappyeyeballs" }, - { name = "aiosignal" }, - { name = "async-timeout", marker = "python_full_version < '3.11'" }, - { name = "attrs" }, - { name = "frozenlist" }, - { name = "multidict" }, - { name = "propcache" }, - { name = "yarl" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/1c/ce/3b83ebba6b3207a7135e5fcaba49706f8a4b6008153b4e30540c982fae26/aiohttp-3.13.2.tar.gz", hash = "sha256:40176a52c186aefef6eb3cad2cdd30cd06e3afbe88fe8ab2af9c0b90f228daca", size = 7837994 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/6d/34/939730e66b716b76046dedfe0842995842fa906ccc4964bba414ff69e429/aiohttp-3.13.2-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:2372b15a5f62ed37789a6b383ff7344fc5b9f243999b0cd9b629d8bc5f5b4155", size = 736471 }, - { url = "https://files.pythonhosted.org/packages/fd/cf/dcbdf2df7f6ca72b0bb4c0b4509701f2d8942cf54e29ca197389c214c07f/aiohttp-3.13.2-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:e7f8659a48995edee7229522984bd1009c1213929c769c2daa80b40fe49a180c", size = 493985 }, - { url = "https://files.pythonhosted.org/packages/9d/87/71c8867e0a1d0882dcbc94af767784c3cb381c1c4db0943ab4aae4fed65e/aiohttp-3.13.2-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:939ced4a7add92296b0ad38892ce62b98c619288a081170695c6babe4f50e636", size = 489274 }, - { url = "https://files.pythonhosted.org/packages/38/0f/46c24e8dae237295eaadd113edd56dee96ef6462adf19b88592d44891dc5/aiohttp-3.13.2-cp310-cp310-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:6315fb6977f1d0dd41a107c527fee2ed5ab0550b7d885bc15fee20ccb17891da", size = 1668171 }, - { url = "https://files.pythonhosted.org/packages/eb/c6/4cdfb4440d0e28483681a48f69841fa5e39366347d66ef808cbdadddb20e/aiohttp-3.13.2-cp310-cp310-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:6e7352512f763f760baaed2637055c49134fd1d35b37c2dedfac35bfe5cf8725", size = 1636036 }, - { url = "https://files.pythonhosted.org/packages/84/37/8708cf678628216fb678ab327a4e1711c576d6673998f4f43e86e9ae90dd/aiohttp-3.13.2-cp310-cp310-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:e09a0a06348a2dd73e7213353c90d709502d9786219f69b731f6caa0efeb46f5", size = 1727975 }, - { url = "https://files.pythonhosted.org/packages/e6/2e/3ebfe12fdcb9b5f66e8a0a42dffcd7636844c8a018f261efb2419f68220b/aiohttp-3.13.2-cp310-cp310-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:a09a6d073fb5789456545bdee2474d14395792faa0527887f2f4ec1a486a59d3", size = 1815823 }, - { url = "https://files.pythonhosted.org/packages/a1/4f/ca2ef819488cbb41844c6cf92ca6dd15b9441e6207c58e5ae0e0fc8d70ad/aiohttp-3.13.2-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:b59d13c443f8e049d9e94099c7e412e34610f1f49be0f230ec656a10692a5802", size = 1669374 }, - { url = "https://files.pythonhosted.org/packages/f8/fe/1fe2e1179a0d91ce09c99069684aab619bf2ccde9b20bd6ca44f8837203e/aiohttp-3.13.2-cp310-cp310-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl", hash = "sha256:20db2d67985d71ca033443a1ba2001c4b5693fe09b0e29f6d9358a99d4d62a8a", size = 1555315 }, - { url = "https://files.pythonhosted.org/packages/5a/2b/f3781899b81c45d7cbc7140cddb8a3481c195e7cbff8e36374759d2ab5a5/aiohttp-3.13.2-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:960c2fc686ba27b535f9fd2b52d87ecd7e4fd1cf877f6a5cba8afb5b4a8bd204", size = 1639140 }, - { url = "https://files.pythonhosted.org/packages/72/27/c37e85cd3ece6f6c772e549bd5a253d0c122557b25855fb274224811e4f2/aiohttp-3.13.2-cp310-cp310-musllinux_1_2_armv7l.whl", hash = "sha256:6c00dbcf5f0d88796151e264a8eab23de2997c9303dd7c0bf622e23b24d3ce22", size = 1645496 }, - { url = "https://files.pythonhosted.org/packages/66/20/3af1ab663151bd3780b123e907761cdb86ec2c4e44b2d9b195ebc91fbe37/aiohttp-3.13.2-cp310-cp310-musllinux_1_2_ppc64le.whl", hash = "sha256:fed38a5edb7945f4d1bcabe2fcd05db4f6ec7e0e82560088b754f7e08d93772d", size = 1697625 }, - { url = "https://files.pythonhosted.org/packages/95/eb/ae5cab15efa365e13d56b31b0d085a62600298bf398a7986f8388f73b598/aiohttp-3.13.2-cp310-cp310-musllinux_1_2_riscv64.whl", hash = "sha256:b395bbca716c38bef3c764f187860e88c724b342c26275bc03e906142fc5964f", size = 1542025 }, - { url = "https://files.pythonhosted.org/packages/e9/2d/1683e8d67ec72d911397fe4e575688d2a9b8f6a6e03c8fdc9f3fd3d4c03f/aiohttp-3.13.2-cp310-cp310-musllinux_1_2_s390x.whl", hash = "sha256:204ffff2426c25dfda401ba08da85f9c59525cdc42bda26660463dd1cbcfec6f", size = 1714918 }, - { url = "https://files.pythonhosted.org/packages/99/a2/ffe8e0e1c57c5e542d47ffa1fcf95ef2b3ea573bf7c4d2ee877252431efc/aiohttp-3.13.2-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:05c4dd3c48fb5f15db31f57eb35374cb0c09afdde532e7fb70a75aede0ed30f6", size = 1656113 }, - { url = "https://files.pythonhosted.org/packages/0d/42/d511aff5c3a2b06c09d7d214f508a4ad8ac7799817f7c3d23e7336b5e896/aiohttp-3.13.2-cp310-cp310-win32.whl", hash = "sha256:e574a7d61cf10351d734bcddabbe15ede0eaa8a02070d85446875dc11189a251", size = 432290 }, - { url = "https://files.pythonhosted.org/packages/8b/ea/1c2eb7098b5bad4532994f2b7a8228d27674035c9b3234fe02c37469ef14/aiohttp-3.13.2-cp310-cp310-win_amd64.whl", hash = "sha256:364f55663085d658b8462a1c3f17b2b84a5c2e1ba858e1b79bff7b2e24ad1514", size = 455075 }, - { url = "https://files.pythonhosted.org/packages/35/74/b321e7d7ca762638cdf8cdeceb39755d9c745aff7a64c8789be96ddf6e96/aiohttp-3.13.2-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:4647d02df098f6434bafd7f32ad14942f05a9caa06c7016fdcc816f343997dd0", size = 743409 }, - { url = "https://files.pythonhosted.org/packages/99/3d/91524b905ec473beaf35158d17f82ef5a38033e5809fe8742e3657cdbb97/aiohttp-3.13.2-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:e3403f24bcb9c3b29113611c3c16a2a447c3953ecf86b79775e7be06f7ae7ccb", size = 497006 }, - { url = "https://files.pythonhosted.org/packages/eb/d3/7f68bc02a67716fe80f063e19adbd80a642e30682ce74071269e17d2dba1/aiohttp-3.13.2-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:43dff14e35aba17e3d6d5ba628858fb8cb51e30f44724a2d2f0c75be492c55e9", size = 493195 }, - { url = "https://files.pythonhosted.org/packages/98/31/913f774a4708775433b7375c4f867d58ba58ead833af96c8af3621a0d243/aiohttp-3.13.2-cp311-cp311-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:e2a9ea08e8c58bb17655630198833109227dea914cd20be660f52215f6de5613", size = 1747759 }, - { url = "https://files.pythonhosted.org/packages/e8/63/04efe156f4326f31c7c4a97144f82132c3bb21859b7bb84748d452ccc17c/aiohttp-3.13.2-cp311-cp311-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:53b07472f235eb80e826ad038c9d106c2f653584753f3ddab907c83f49eedead", size = 1704456 }, - { url = "https://files.pythonhosted.org/packages/8e/02/4e16154d8e0a9cf4ae76f692941fd52543bbb148f02f098ca73cab9b1c1b/aiohttp-3.13.2-cp311-cp311-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:e736c93e9c274fce6419af4aac199984d866e55f8a4cec9114671d0ea9688780", size = 1807572 }, - { url = "https://files.pythonhosted.org/packages/34/58/b0583defb38689e7f06798f0285b1ffb3a6fb371f38363ce5fd772112724/aiohttp-3.13.2-cp311-cp311-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:ff5e771f5dcbc81c64898c597a434f7682f2259e0cd666932a913d53d1341d1a", size = 1895954 }, - { url = "https://files.pythonhosted.org/packages/6b/f3/083907ee3437425b4e376aa58b2c915eb1a33703ec0dc30040f7ae3368c6/aiohttp-3.13.2-cp311-cp311-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:a3b6fb0c207cc661fa0bf8c66d8d9b657331ccc814f4719468af61034b478592", size = 1747092 }, - { url = "https://files.pythonhosted.org/packages/ac/61/98a47319b4e425cc134e05e5f3fc512bf9a04bf65aafd9fdcda5d57ec693/aiohttp-3.13.2-cp311-cp311-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl", hash = "sha256:97a0895a8e840ab3520e2288db7cace3a1981300d48babeb50e7425609e2e0ab", size = 1606815 }, - { url = "https://files.pythonhosted.org/packages/97/4b/e78b854d82f66bb974189135d31fce265dee0f5344f64dd0d345158a5973/aiohttp-3.13.2-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:9e8f8afb552297aca127c90cb840e9a1d4bfd6a10d7d8f2d9176e1acc69bad30", size = 1723789 }, - { url = "https://files.pythonhosted.org/packages/ed/fc/9d2ccc794fc9b9acd1379d625c3a8c64a45508b5091c546dea273a41929e/aiohttp-3.13.2-cp311-cp311-musllinux_1_2_armv7l.whl", hash = "sha256:ed2f9c7216e53c3df02264f25d824b079cc5914f9e2deba94155190ef648ee40", size = 1718104 }, - { url = "https://files.pythonhosted.org/packages/66/65/34564b8765ea5c7d79d23c9113135d1dd3609173da13084830f1507d56cf/aiohttp-3.13.2-cp311-cp311-musllinux_1_2_ppc64le.whl", hash = "sha256:99c5280a329d5fa18ef30fd10c793a190d996567667908bef8a7f81f8202b948", size = 1785584 }, - { url = "https://files.pythonhosted.org/packages/30/be/f6a7a426e02fc82781afd62016417b3948e2207426d90a0e478790d1c8a4/aiohttp-3.13.2-cp311-cp311-musllinux_1_2_riscv64.whl", hash = "sha256:2ca6ffef405fc9c09a746cb5d019c1672cd7f402542e379afc66b370833170cf", size = 1595126 }, - { url = "https://files.pythonhosted.org/packages/e5/c7/8e22d5d28f94f67d2af496f14a83b3c155d915d1fe53d94b66d425ec5b42/aiohttp-3.13.2-cp311-cp311-musllinux_1_2_s390x.whl", hash = "sha256:47f438b1a28e926c37632bff3c44df7d27c9b57aaf4e34b1def3c07111fdb782", size = 1800665 }, - { url = "https://files.pythonhosted.org/packages/d1/11/91133c8b68b1da9fc16555706aa7276fdf781ae2bb0876c838dd86b8116e/aiohttp-3.13.2-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:9acda8604a57bb60544e4646a4615c1866ee6c04a8edef9b8ee6fd1d8fa2ddc8", size = 1739532 }, - { url = "https://files.pythonhosted.org/packages/17/6b/3747644d26a998774b21a616016620293ddefa4d63af6286f389aedac844/aiohttp-3.13.2-cp311-cp311-win32.whl", hash = "sha256:868e195e39b24aaa930b063c08bb0c17924899c16c672a28a65afded9c46c6ec", size = 431876 }, - { url = "https://files.pythonhosted.org/packages/c3/63/688462108c1a00eb9f05765331c107f95ae86f6b197b865d29e930b7e462/aiohttp-3.13.2-cp311-cp311-win_amd64.whl", hash = "sha256:7fd19df530c292542636c2a9a85854fab93474396a52f1695e799186bbd7f24c", size = 456205 }, - { url = "https://files.pythonhosted.org/packages/29/9b/01f00e9856d0a73260e86dd8ed0c2234a466c5c1712ce1c281548df39777/aiohttp-3.13.2-cp312-cp312-macosx_10_13_universal2.whl", hash = "sha256:b1e56bab2e12b2b9ed300218c351ee2a3d8c8fdab5b1ec6193e11a817767e47b", size = 737623 }, - { url = "https://files.pythonhosted.org/packages/5a/1b/4be39c445e2b2bd0aab4ba736deb649fabf14f6757f405f0c9685019b9e9/aiohttp-3.13.2-cp312-cp312-macosx_10_13_x86_64.whl", hash = "sha256:364e25edaabd3d37b1db1f0cbcee8c73c9a3727bfa262b83e5e4cf3489a2a9dc", size = 492664 }, - { url = "https://files.pythonhosted.org/packages/28/66/d35dcfea8050e131cdd731dff36434390479b4045a8d0b9d7111b0a968f1/aiohttp-3.13.2-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:c5c94825f744694c4b8db20b71dba9a257cd2ba8e010a803042123f3a25d50d7", size = 491808 }, - { url = "https://files.pythonhosted.org/packages/00/29/8e4609b93e10a853b65f8291e64985de66d4f5848c5637cddc70e98f01f8/aiohttp-3.13.2-cp312-cp312-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:ba2715d842ffa787be87cbfce150d5e88c87a98e0b62e0f5aa489169a393dbbb", size = 1738863 }, - { url = "https://files.pythonhosted.org/packages/9d/fa/4ebdf4adcc0def75ced1a0d2d227577cd7b1b85beb7edad85fcc87693c75/aiohttp-3.13.2-cp312-cp312-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:585542825c4bc662221fb257889e011a5aa00f1ae4d75d1d246a5225289183e3", size = 1700586 }, - { url = "https://files.pythonhosted.org/packages/da/04/73f5f02ff348a3558763ff6abe99c223381b0bace05cd4530a0258e52597/aiohttp-3.13.2-cp312-cp312-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:39d02cb6025fe1aabca329c5632f48c9532a3dabccd859e7e2f110668972331f", size = 1768625 }, - { url = "https://files.pythonhosted.org/packages/f8/49/a825b79ffec124317265ca7d2344a86bcffeb960743487cb11988ffb3494/aiohttp-3.13.2-cp312-cp312-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:e67446b19e014d37342f7195f592a2a948141d15a312fe0e700c2fd2f03124f6", size = 1867281 }, - { url = "https://files.pythonhosted.org/packages/b9/48/adf56e05f81eac31edcfae45c90928f4ad50ef2e3ea72cb8376162a368f8/aiohttp-3.13.2-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:4356474ad6333e41ccefd39eae869ba15a6c5299c9c01dfdcfdd5c107be4363e", size = 1752431 }, - { url = "https://files.pythonhosted.org/packages/30/ab/593855356eead019a74e862f21523db09c27f12fd24af72dbc3555b9bfd9/aiohttp-3.13.2-cp312-cp312-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl", hash = "sha256:eeacf451c99b4525f700f078becff32c32ec327b10dcf31306a8a52d78166de7", size = 1562846 }, - { url = "https://files.pythonhosted.org/packages/39/0f/9f3d32271aa8dc35036e9668e31870a9d3b9542dd6b3e2c8a30931cb27ae/aiohttp-3.13.2-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:d8a9b889aeabd7a4e9af0b7f4ab5ad94d42e7ff679aaec6d0db21e3b639ad58d", size = 1699606 }, - { url = "https://files.pythonhosted.org/packages/2c/3c/52d2658c5699b6ef7692a3f7128b2d2d4d9775f2a68093f74bca06cf01e1/aiohttp-3.13.2-cp312-cp312-musllinux_1_2_armv7l.whl", hash = "sha256:fa89cb11bc71a63b69568d5b8a25c3ca25b6d54c15f907ca1c130d72f320b76b", size = 1720663 }, - { url = "https://files.pythonhosted.org/packages/9b/d4/8f8f3ff1fb7fb9e3f04fcad4e89d8a1cd8fc7d05de67e3de5b15b33008ff/aiohttp-3.13.2-cp312-cp312-musllinux_1_2_ppc64le.whl", hash = "sha256:8aa7c807df234f693fed0ecd507192fc97692e61fee5702cdc11155d2e5cadc8", size = 1737939 }, - { url = "https://files.pythonhosted.org/packages/03/d3/ddd348f8a27a634daae39a1b8e291ff19c77867af438af844bf8b7e3231b/aiohttp-3.13.2-cp312-cp312-musllinux_1_2_riscv64.whl", hash = "sha256:9eb3e33fdbe43f88c3c75fa608c25e7c47bbd80f48d012763cb67c47f39a7e16", size = 1555132 }, - { url = "https://files.pythonhosted.org/packages/39/b8/46790692dc46218406f94374903ba47552f2f9f90dad554eed61bfb7b64c/aiohttp-3.13.2-cp312-cp312-musllinux_1_2_s390x.whl", hash = "sha256:9434bc0d80076138ea986833156c5a48c9c7a8abb0c96039ddbb4afc93184169", size = 1764802 }, - { url = "https://files.pythonhosted.org/packages/ba/e4/19ce547b58ab2a385e5f0b8aa3db38674785085abcf79b6e0edd1632b12f/aiohttp-3.13.2-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:ff15c147b2ad66da1f2cbb0622313f2242d8e6e8f9b79b5206c84523a4473248", size = 1719512 }, - { url = "https://files.pythonhosted.org/packages/70/30/6355a737fed29dcb6dfdd48682d5790cb5eab050f7b4e01f49b121d3acad/aiohttp-3.13.2-cp312-cp312-win32.whl", hash = "sha256:27e569eb9d9e95dbd55c0fc3ec3a9335defbf1d8bc1d20171a49f3c4c607b93e", size = 426690 }, - { url = "https://files.pythonhosted.org/packages/0a/0d/b10ac09069973d112de6ef980c1f6bb31cb7dcd0bc363acbdad58f927873/aiohttp-3.13.2-cp312-cp312-win_amd64.whl", hash = "sha256:8709a0f05d59a71f33fd05c17fc11fcb8c30140506e13c2f5e8ee1b8964e1b45", size = 453465 }, - { url = "https://files.pythonhosted.org/packages/bf/78/7e90ca79e5aa39f9694dcfd74f4720782d3c6828113bb1f3197f7e7c4a56/aiohttp-3.13.2-cp313-cp313-macosx_10_13_universal2.whl", hash = "sha256:7519bdc7dfc1940d201651b52bf5e03f5503bda45ad6eacf64dda98be5b2b6be", size = 732139 }, - { url = "https://files.pythonhosted.org/packages/db/ed/1f59215ab6853fbaa5c8495fa6cbc39edfc93553426152b75d82a5f32b76/aiohttp-3.13.2-cp313-cp313-macosx_10_13_x86_64.whl", hash = "sha256:088912a78b4d4f547a1f19c099d5a506df17eacec3c6f4375e2831ec1d995742", size = 490082 }, - { url = "https://files.pythonhosted.org/packages/68/7b/fe0fe0f5e05e13629d893c760465173a15ad0039c0a5b0d0040995c8075e/aiohttp-3.13.2-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:5276807b9de9092af38ed23ce120539ab0ac955547b38563a9ba4f5b07b95293", size = 489035 }, - { url = "https://files.pythonhosted.org/packages/d2/04/db5279e38471b7ac801d7d36a57d1230feeee130bbe2a74f72731b23c2b1/aiohttp-3.13.2-cp313-cp313-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:1237c1375eaef0db4dcd7c2559f42e8af7b87ea7d295b118c60c36a6e61cb811", size = 1720387 }, - { url = "https://files.pythonhosted.org/packages/31/07/8ea4326bd7dae2bd59828f69d7fdc6e04523caa55e4a70f4a8725a7e4ed2/aiohttp-3.13.2-cp313-cp313-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:96581619c57419c3d7d78703d5b78c1e5e5fc0172d60f555bdebaced82ded19a", size = 1688314 }, - { url = "https://files.pythonhosted.org/packages/48/ab/3d98007b5b87ffd519d065225438cc3b668b2f245572a8cb53da5dd2b1bc/aiohttp-3.13.2-cp313-cp313-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:a2713a95b47374169409d18103366de1050fe0ea73db358fc7a7acb2880422d4", size = 1756317 }, - { url = "https://files.pythonhosted.org/packages/97/3d/801ca172b3d857fafb7b50c7c03f91b72b867a13abca982ed6b3081774ef/aiohttp-3.13.2-cp313-cp313-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:228a1cd556b3caca590e9511a89444925da87d35219a49ab5da0c36d2d943a6a", size = 1858539 }, - { url = "https://files.pythonhosted.org/packages/f7/0d/4764669bdf47bd472899b3d3db91fffbe925c8e3038ec591a2fd2ad6a14d/aiohttp-3.13.2-cp313-cp313-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:ac6cde5fba8d7d8c6ac963dbb0256a9854e9fafff52fbcc58fdf819357892c3e", size = 1739597 }, - { url = "https://files.pythonhosted.org/packages/c4/52/7bd3c6693da58ba16e657eb904a5b6decfc48ecd06e9ac098591653b1566/aiohttp-3.13.2-cp313-cp313-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl", hash = "sha256:f2bef8237544f4e42878c61cef4e2839fee6346dc60f5739f876a9c50be7fcdb", size = 1555006 }, - { url = "https://files.pythonhosted.org/packages/48/30/9586667acec5993b6f41d2ebcf96e97a1255a85f62f3c653110a5de4d346/aiohttp-3.13.2-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:16f15a4eac3bc2d76c45f7ebdd48a65d41b242eb6c31c2245463b40b34584ded", size = 1683220 }, - { url = "https://files.pythonhosted.org/packages/71/01/3afe4c96854cfd7b30d78333852e8e851dceaec1c40fd00fec90c6402dd2/aiohttp-3.13.2-cp313-cp313-musllinux_1_2_armv7l.whl", hash = "sha256:bb7fb776645af5cc58ab804c58d7eba545a97e047254a52ce89c157b5af6cd0b", size = 1712570 }, - { url = "https://files.pythonhosted.org/packages/11/2c/22799d8e720f4697a9e66fd9c02479e40a49de3de2f0bbe7f9f78a987808/aiohttp-3.13.2-cp313-cp313-musllinux_1_2_ppc64le.whl", hash = "sha256:e1b4951125ec10c70802f2cb09736c895861cd39fd9dcb35107b4dc8ae6220b8", size = 1733407 }, - { url = "https://files.pythonhosted.org/packages/34/cb/90f15dd029f07cebbd91f8238a8b363978b530cd128488085b5703683594/aiohttp-3.13.2-cp313-cp313-musllinux_1_2_riscv64.whl", hash = "sha256:550bf765101ae721ee1d37d8095f47b1f220650f85fe1af37a90ce75bab89d04", size = 1550093 }, - { url = "https://files.pythonhosted.org/packages/69/46/12dce9be9d3303ecbf4d30ad45a7683dc63d90733c2d9fe512be6716cd40/aiohttp-3.13.2-cp313-cp313-musllinux_1_2_s390x.whl", hash = "sha256:fe91b87fc295973096251e2d25a811388e7d8adf3bd2b97ef6ae78bc4ac6c476", size = 1758084 }, - { url = "https://files.pythonhosted.org/packages/f9/c8/0932b558da0c302ffd639fc6362a313b98fdf235dc417bc2493da8394df7/aiohttp-3.13.2-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:e0c8e31cfcc4592cb200160344b2fb6ae0f9e4effe06c644b5a125d4ae5ebe23", size = 1716987 }, - { url = "https://files.pythonhosted.org/packages/5d/8b/f5bd1a75003daed099baec373aed678f2e9b34f2ad40d85baa1368556396/aiohttp-3.13.2-cp313-cp313-win32.whl", hash = "sha256:0740f31a60848d6edb296a0df827473eede90c689b8f9f2a4cdde74889eb2254", size = 425859 }, - { url = "https://files.pythonhosted.org/packages/5d/28/a8a9fc6957b2cee8902414e41816b5ab5536ecf43c3b1843c10e82c559b2/aiohttp-3.13.2-cp313-cp313-win_amd64.whl", hash = "sha256:a88d13e7ca367394908f8a276b89d04a3652044612b9a408a0bb22a5ed976a1a", size = 452192 }, - { url = "https://files.pythonhosted.org/packages/9b/36/e2abae1bd815f01c957cbf7be817b3043304e1c87bad526292a0410fdcf9/aiohttp-3.13.2-cp314-cp314-macosx_10_13_universal2.whl", hash = "sha256:2475391c29230e063ef53a66669b7b691c9bfc3f1426a0f7bcdf1216bdbac38b", size = 735234 }, - { url = "https://files.pythonhosted.org/packages/ca/e3/1ee62dde9b335e4ed41db6bba02613295a0d5b41f74a783c142745a12763/aiohttp-3.13.2-cp314-cp314-macosx_10_13_x86_64.whl", hash = "sha256:f33c8748abef4d8717bb20e8fb1b3e07c6adacb7fd6beaae971a764cf5f30d61", size = 490733 }, - { url = "https://files.pythonhosted.org/packages/1a/aa/7a451b1d6a04e8d15a362af3e9b897de71d86feac3babf8894545d08d537/aiohttp-3.13.2-cp314-cp314-macosx_11_0_arm64.whl", hash = "sha256:ae32f24bbfb7dbb485a24b30b1149e2f200be94777232aeadba3eecece4d0aa4", size = 491303 }, - { url = "https://files.pythonhosted.org/packages/57/1e/209958dbb9b01174870f6a7538cd1f3f28274fdbc88a750c238e2c456295/aiohttp-3.13.2-cp314-cp314-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:5d7f02042c1f009ffb70067326ef183a047425bb2ff3bc434ead4dd4a4a66a2b", size = 1717965 }, - { url = "https://files.pythonhosted.org/packages/08/aa/6a01848d6432f241416bc4866cae8dc03f05a5a884d2311280f6a09c73d6/aiohttp-3.13.2-cp314-cp314-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:93655083005d71cd6c072cdab54c886e6570ad2c4592139c3fb967bfc19e4694", size = 1667221 }, - { url = "https://files.pythonhosted.org/packages/87/4f/36c1992432d31bbc789fa0b93c768d2e9047ec8c7177e5cd84ea85155f36/aiohttp-3.13.2-cp314-cp314-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:0db1e24b852f5f664cd728db140cf11ea0e82450471232a394b3d1a540b0f906", size = 1757178 }, - { url = "https://files.pythonhosted.org/packages/ac/b4/8e940dfb03b7e0f68a82b88fd182b9be0a65cb3f35612fe38c038c3112cf/aiohttp-3.13.2-cp314-cp314-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:b009194665bcd128e23eaddef362e745601afa4641930848af4c8559e88f18f9", size = 1838001 }, - { url = "https://files.pythonhosted.org/packages/d7/ef/39f3448795499c440ab66084a9db7d20ca7662e94305f175a80f5b7e0072/aiohttp-3.13.2-cp314-cp314-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:c038a8fdc8103cd51dbd986ecdce141473ffd9775a7a8057a6ed9c3653478011", size = 1716325 }, - { url = "https://files.pythonhosted.org/packages/d7/51/b311500ffc860b181c05d91c59a1313bdd05c82960fdd4035a15740d431e/aiohttp-3.13.2-cp314-cp314-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl", hash = "sha256:66bac29b95a00db411cd758fea0e4b9bdba6d549dfe333f9a945430f5f2cc5a6", size = 1547978 }, - { url = "https://files.pythonhosted.org/packages/31/64/b9d733296ef79815226dab8c586ff9e3df41c6aff2e16c06697b2d2e6775/aiohttp-3.13.2-cp314-cp314-musllinux_1_2_aarch64.whl", hash = "sha256:4ebf9cfc9ba24a74cf0718f04aac2a3bbe745902cc7c5ebc55c0f3b5777ef213", size = 1682042 }, - { url = "https://files.pythonhosted.org/packages/3f/30/43d3e0f9d6473a6db7d472104c4eff4417b1e9df01774cb930338806d36b/aiohttp-3.13.2-cp314-cp314-musllinux_1_2_armv7l.whl", hash = "sha256:a4b88ebe35ce54205c7074f7302bd08a4cb83256a3e0870c72d6f68a3aaf8e49", size = 1680085 }, - { url = "https://files.pythonhosted.org/packages/16/51/c709f352c911b1864cfd1087577760ced64b3e5bee2aa88b8c0c8e2e4972/aiohttp-3.13.2-cp314-cp314-musllinux_1_2_ppc64le.whl", hash = "sha256:98c4fb90bb82b70a4ed79ca35f656f4281885be076f3f970ce315402b53099ae", size = 1728238 }, - { url = "https://files.pythonhosted.org/packages/19/e2/19bd4c547092b773caeb48ff5ae4b1ae86756a0ee76c16727fcfd281404b/aiohttp-3.13.2-cp314-cp314-musllinux_1_2_riscv64.whl", hash = "sha256:ec7534e63ae0f3759df3a1ed4fa6bc8f75082a924b590619c0dd2f76d7043caa", size = 1544395 }, - { url = "https://files.pythonhosted.org/packages/cf/87/860f2803b27dfc5ed7be532832a3498e4919da61299b4a1f8eb89b8ff44d/aiohttp-3.13.2-cp314-cp314-musllinux_1_2_s390x.whl", hash = "sha256:5b927cf9b935a13e33644cbed6c8c4b2d0f25b713d838743f8fe7191b33829c4", size = 1742965 }, - { url = "https://files.pythonhosted.org/packages/67/7f/db2fc7618925e8c7a601094d5cbe539f732df4fb570740be88ed9e40e99a/aiohttp-3.13.2-cp314-cp314-musllinux_1_2_x86_64.whl", hash = "sha256:88d6c017966a78c5265d996c19cdb79235be5e6412268d7e2ce7dee339471b7a", size = 1697585 }, - { url = "https://files.pythonhosted.org/packages/0c/07/9127916cb09bb38284db5036036042b7b2c514c8ebaeee79da550c43a6d6/aiohttp-3.13.2-cp314-cp314-win32.whl", hash = "sha256:f7c183e786e299b5d6c49fb43a769f8eb8e04a2726a2bd5887b98b5cc2d67940", size = 431621 }, - { url = "https://files.pythonhosted.org/packages/fb/41/554a8a380df6d3a2bba8a7726429a23f4ac62aaf38de43bb6d6cde7b4d4d/aiohttp-3.13.2-cp314-cp314-win_amd64.whl", hash = "sha256:fe242cd381e0fb65758faf5ad96c2e460df6ee5b2de1072fe97e4127927e00b4", size = 457627 }, - { url = "https://files.pythonhosted.org/packages/c7/8e/3824ef98c039d3951cb65b9205a96dd2b20f22241ee17d89c5701557c826/aiohttp-3.13.2-cp314-cp314t-macosx_10_13_universal2.whl", hash = "sha256:f10d9c0b0188fe85398c61147bbd2a657d616c876863bfeff43376e0e3134673", size = 767360 }, - { url = "https://files.pythonhosted.org/packages/a4/0f/6a03e3fc7595421274fa34122c973bde2d89344f8a881b728fa8c774e4f1/aiohttp-3.13.2-cp314-cp314t-macosx_10_13_x86_64.whl", hash = "sha256:e7c952aefdf2460f4ae55c5e9c3e80aa72f706a6317e06020f80e96253b1accd", size = 504616 }, - { url = "https://files.pythonhosted.org/packages/c6/aa/ed341b670f1bc8a6f2c6a718353d13b9546e2cef3544f573c6a1ff0da711/aiohttp-3.13.2-cp314-cp314t-macosx_11_0_arm64.whl", hash = "sha256:c20423ce14771d98353d2e25e83591fa75dfa90a3c1848f3d7c68243b4fbded3", size = 509131 }, - { url = "https://files.pythonhosted.org/packages/7f/f0/c68dac234189dae5c4bbccc0f96ce0cc16b76632cfc3a08fff180045cfa4/aiohttp-3.13.2-cp314-cp314t-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:e96eb1a34396e9430c19d8338d2ec33015e4a87ef2b4449db94c22412e25ccdf", size = 1864168 }, - { url = "https://files.pythonhosted.org/packages/8f/65/75a9a76db8364b5d0e52a0c20eabc5d52297385d9af9c35335b924fafdee/aiohttp-3.13.2-cp314-cp314t-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:23fb0783bc1a33640036465019d3bba069942616a6a2353c6907d7fe1ccdaf4e", size = 1719200 }, - { url = "https://files.pythonhosted.org/packages/f5/55/8df2ed78d7f41d232f6bd3ff866b6f617026551aa1d07e2f03458f964575/aiohttp-3.13.2-cp314-cp314t-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:2e1a9bea6244a1d05a4e57c295d69e159a5c50d8ef16aa390948ee873478d9a5", size = 1843497 }, - { url = "https://files.pythonhosted.org/packages/e9/e0/94d7215e405c5a02ccb6a35c7a3a6cfff242f457a00196496935f700cde5/aiohttp-3.13.2-cp314-cp314t-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:0a3d54e822688b56e9f6b5816fb3de3a3a64660efac64e4c2dc435230ad23bad", size = 1935703 }, - { url = "https://files.pythonhosted.org/packages/0b/78/1eeb63c3f9b2d1015a4c02788fb543141aad0a03ae3f7a7b669b2483f8d4/aiohttp-3.13.2-cp314-cp314t-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:7a653d872afe9f33497215745da7a943d1dc15b728a9c8da1c3ac423af35178e", size = 1792738 }, - { url = "https://files.pythonhosted.org/packages/41/75/aaf1eea4c188e51538c04cc568040e3082db263a57086ea74a7d38c39e42/aiohttp-3.13.2-cp314-cp314t-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl", hash = "sha256:56d36e80d2003fa3fc0207fac644216d8532e9504a785ef9a8fd013f84a42c61", size = 1624061 }, - { url = "https://files.pythonhosted.org/packages/9b/c2/3b6034de81fbcc43de8aeb209073a2286dfb50b86e927b4efd81cf848197/aiohttp-3.13.2-cp314-cp314t-musllinux_1_2_aarch64.whl", hash = "sha256:78cd586d8331fb8e241c2dd6b2f4061778cc69e150514b39a9e28dd050475661", size = 1789201 }, - { url = "https://files.pythonhosted.org/packages/c9/38/c15dcf6d4d890217dae79d7213988f4e5fe6183d43893a9cf2fe9e84ca8d/aiohttp-3.13.2-cp314-cp314t-musllinux_1_2_armv7l.whl", hash = "sha256:20b10bbfbff766294fe99987f7bb3b74fdd2f1a2905f2562132641ad434dcf98", size = 1776868 }, - { url = "https://files.pythonhosted.org/packages/04/75/f74fd178ac81adf4f283a74847807ade5150e48feda6aef024403716c30c/aiohttp-3.13.2-cp314-cp314t-musllinux_1_2_ppc64le.whl", hash = "sha256:9ec49dff7e2b3c85cdeaa412e9d438f0ecd71676fde61ec57027dd392f00c693", size = 1790660 }, - { url = "https://files.pythonhosted.org/packages/e7/80/7368bd0d06b16b3aba358c16b919e9c46cf11587dc572091031b0e9e3ef0/aiohttp-3.13.2-cp314-cp314t-musllinux_1_2_riscv64.whl", hash = "sha256:94f05348c4406450f9d73d38efb41d669ad6cd90c7ee194810d0eefbfa875a7a", size = 1617548 }, - { url = "https://files.pythonhosted.org/packages/7d/4b/a6212790c50483cb3212e507378fbe26b5086d73941e1ec4b56a30439688/aiohttp-3.13.2-cp314-cp314t-musllinux_1_2_s390x.whl", hash = "sha256:fa4dcb605c6f82a80c7f95713c2b11c3b8e9893b3ebd2bc9bde93165ed6107be", size = 1817240 }, - { url = "https://files.pythonhosted.org/packages/ff/f7/ba5f0ba4ea8d8f3c32850912944532b933acbf0f3a75546b89269b9b7dde/aiohttp-3.13.2-cp314-cp314t-musllinux_1_2_x86_64.whl", hash = "sha256:cf00e5db968c3f67eccd2778574cf64d8b27d95b237770aa32400bd7a1ca4f6c", size = 1762334 }, - { url = "https://files.pythonhosted.org/packages/7e/83/1a5a1856574588b1cad63609ea9ad75b32a8353ac995d830bf5da9357364/aiohttp-3.13.2-cp314-cp314t-win32.whl", hash = "sha256:d23b5fe492b0805a50d3371e8a728a9134d8de5447dce4c885f5587294750734", size = 464685 }, - { url = "https://files.pythonhosted.org/packages/9f/4d/d22668674122c08f4d56972297c51a624e64b3ed1efaa40187607a7cb66e/aiohttp-3.13.2-cp314-cp314t-win_amd64.whl", hash = "sha256:ff0a7b0a82a7ab905cbda74006318d1b12e37c797eb1b0d4eb3e316cf47f658f", size = 498093 }, -] - -[[package]] -name = "aiosignal" -version = "1.4.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "frozenlist" }, - { name = "typing-extensions", marker = "python_full_version < '3.13'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/61/62/06741b579156360248d1ec624842ad0edf697050bbaf7c3e46394e106ad1/aiosignal-1.4.0.tar.gz", hash = "sha256:f47eecd9468083c2029cc99945502cb7708b082c232f9aca65da147157b251c7", size = 25007 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/fb/76/641ae371508676492379f16e2fa48f4e2c11741bd63c48be4b12a6b09cba/aiosignal-1.4.0-py3-none-any.whl", hash = "sha256:053243f8b92b990551949e63930a839ff0cf0b0ebbe0597b0f3fb19e1a0fe82e", size = 7490 }, -] - -[[package]] -name = "alembic" -version = "1.17.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "mako" }, - { name = "sqlalchemy" }, - { name = "tomli", marker = "python_full_version < '3.11'" }, - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/02/a6/74c8cadc2882977d80ad756a13857857dbcf9bd405bc80b662eb10651282/alembic-1.17.2.tar.gz", hash = "sha256:bbe9751705c5e0f14877f02d46c53d10885e377e3d90eda810a016f9baa19e8e", size = 1988064 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/ba/88/6237e97e3385b57b5f1528647addea5cc03d4d65d5979ab24327d41fb00d/alembic-1.17.2-py3-none-any.whl", hash = "sha256:f483dd1fe93f6c5d49217055e4d15b905b425b6af906746abb35b69c1996c4e6", size = 248554 }, -] - -[[package]] -name = "annotated-types" -version = "0.7.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/ee/67/531ea369ba64dcff5ec9c3402f9f51bf748cec26dde048a2f973a4eea7f5/annotated_types-0.7.0.tar.gz", hash = "sha256:aff07c09a53a08bc8cfccb9c85b05f1aa9a2a6f23728d790723543408344ce89", size = 16081 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/78/b6/6307fbef88d9b5ee7421e68d78a9f162e0da4900bc5f5793f6d3d0e34fb8/annotated_types-0.7.0-py3-none-any.whl", hash = "sha256:1f02e8b43a8fbbc3f3e0d4f0f4bfc8131bcb4eebe8849b8e5c773f3a1c582a53", size = 13643 }, -] - -[[package]] -name = "antlr4-python3-runtime" -version = "4.9.3" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/3e/38/7859ff46355f76f8d19459005ca000b6e7012f2f1ca597746cbcd1fbfe5e/antlr4-python3-runtime-4.9.3.tar.gz", hash = "sha256:f224469b4168294902bb1efa80a8bf7855f24c99aef99cbefc1bcd3cce77881b", size = 117034 } - -[[package]] -name = "anyio" -version = "4.12.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "exceptiongroup", marker = "python_full_version < '3.11'" }, - { name = "idna" }, - { name = "typing-extensions", marker = "python_full_version < '3.13'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/16/ce/8a777047513153587e5434fd752e89334ac33e379aa3497db860eeb60377/anyio-4.12.0.tar.gz", hash = "sha256:73c693b567b0c55130c104d0b43a9baf3aa6a31fc6110116509f27bf75e21ec0", size = 228266 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/7f/9c/36c5c37947ebfb8c7f22e0eb6e4d188ee2d53aa3880f3f2744fb894f0cb1/anyio-4.12.0-py3-none-any.whl", hash = "sha256:dad2376a628f98eeca4881fc56cd06affd18f659b17a747d3ff0307ced94b1bb", size = 113362 }, -] - -[[package]] -name = "asttokens" -version = "3.0.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/be/a5/8e3f9b6771b0b408517c82d97aed8f2036509bc247d46114925e32fe33f0/asttokens-3.0.1.tar.gz", hash = "sha256:71a4ee5de0bde6a31d64f6b13f2293ac190344478f081c3d1bccfcf5eacb0cb7", size = 62308 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/d2/39/e7eaf1799466a4aef85b6a4fe7bd175ad2b1c6345066aa33f1f58d4b18d0/asttokens-3.0.1-py3-none-any.whl", hash = "sha256:15a3ebc0f43c2d0a50eeafea25e19046c68398e487b9f1f5b517f7c0f40f976a", size = 27047 }, -] - -[[package]] -name = "async-timeout" -version = "5.0.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/a5/ae/136395dfbfe00dfc94da3f3e136d0b13f394cba8f4841120e34226265780/async_timeout-5.0.1.tar.gz", hash = "sha256:d9321a7a3d5a6a5e187e824d2fa0793ce379a202935782d555d6e9d2735677d3", size = 9274 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/fe/ba/e2081de779ca30d473f21f5b30e0e737c438205440784c7dfc81efc2b029/async_timeout-5.0.1-py3-none-any.whl", hash = "sha256:39e3809566ff85354557ec2398b55e096c8364bacac9405a7a1fa429e77fe76c", size = 6233 }, -] - -[[package]] -name = "attrs" -version = "25.4.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/6b/5c/685e6633917e101e5dcb62b9dd76946cbb57c26e133bae9e0cd36033c0a9/attrs-25.4.0.tar.gz", hash = "sha256:16d5969b87f0859ef33a48b35d55ac1be6e42ae49d5e853b597db70c35c57e11", size = 934251 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/3a/2a/7cc015f5b9f5db42b7d48157e23356022889fc354a2813c15934b7cb5c0e/attrs-25.4.0-py3-none-any.whl", hash = "sha256:adcf7e2a1fb3b36ac48d97835bb6d8ade15b8dcce26aba8bf1d14847b57a3373", size = 67615 }, -] - -[[package]] -name = "audioop-lts" -version = "0.2.2" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/38/53/946db57842a50b2da2e0c1e34bd37f36f5aadba1a929a3971c5d7841dbca/audioop_lts-0.2.2.tar.gz", hash = "sha256:64d0c62d88e67b98a1a5e71987b7aa7b5bcffc7dcee65b635823dbdd0a8dbbd0", size = 30686 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/de/d4/94d277ca941de5a507b07f0b592f199c22454eeaec8f008a286b3fbbacd6/audioop_lts-0.2.2-cp313-abi3-macosx_10_13_universal2.whl", hash = "sha256:fd3d4602dc64914d462924a08c1a9816435a2155d74f325853c1f1ac3b2d9800", size = 46523 }, - { url = "https://files.pythonhosted.org/packages/f8/5a/656d1c2da4b555920ce4177167bfeb8623d98765594af59702c8873f60ec/audioop_lts-0.2.2-cp313-abi3-macosx_10_13_x86_64.whl", hash = "sha256:550c114a8df0aafe9a05442a1162dfc8fec37e9af1d625ae6060fed6e756f303", size = 27455 }, - { url = "https://files.pythonhosted.org/packages/1b/83/ea581e364ce7b0d41456fb79d6ee0ad482beda61faf0cab20cbd4c63a541/audioop_lts-0.2.2-cp313-abi3-macosx_11_0_arm64.whl", hash = "sha256:9a13dc409f2564de15dd68be65b462ba0dde01b19663720c68c1140c782d1d75", size = 26997 }, - { url = "https://files.pythonhosted.org/packages/b8/3b/e8964210b5e216e5041593b7d33e97ee65967f17c282e8510d19c666dab4/audioop_lts-0.2.2-cp313-abi3-manylinux1_x86_64.manylinux_2_28_x86_64.manylinux_2_5_x86_64.whl", hash = "sha256:51c916108c56aa6e426ce611946f901badac950ee2ddaf302b7ed35d9958970d", size = 85844 }, - { url = "https://files.pythonhosted.org/packages/c7/2e/0a1c52faf10d51def20531a59ce4c706cb7952323b11709e10de324d6493/audioop_lts-0.2.2-cp313-abi3-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:47eba38322370347b1c47024defbd36374a211e8dd5b0dcbce7b34fdb6f8847b", size = 85056 }, - { url = "https://files.pythonhosted.org/packages/75/e8/cd95eef479656cb75ab05dfece8c1f8c395d17a7c651d88f8e6e291a63ab/audioop_lts-0.2.2-cp313-abi3-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:ba7c3a7e5f23e215cb271516197030c32aef2e754252c4c70a50aaff7031a2c8", size = 93892 }, - { url = "https://files.pythonhosted.org/packages/5c/1e/a0c42570b74f83efa5cca34905b3eef03f7ab09fe5637015df538a7f3345/audioop_lts-0.2.2-cp313-abi3-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:def246fe9e180626731b26e89816e79aae2276f825420a07b4a647abaa84becc", size = 96660 }, - { url = "https://files.pythonhosted.org/packages/50/d5/8a0ae607ca07dbb34027bac8db805498ee7bfecc05fd2c148cc1ed7646e7/audioop_lts-0.2.2-cp313-abi3-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl", hash = "sha256:e160bf9df356d841bb6c180eeeea1834085464626dc1b68fa4e1d59070affdc3", size = 79143 }, - { url = "https://files.pythonhosted.org/packages/12/17/0d28c46179e7910bfb0bb62760ccb33edb5de973052cb2230b662c14ca2e/audioop_lts-0.2.2-cp313-abi3-musllinux_1_2_aarch64.whl", hash = "sha256:4b4cd51a57b698b2d06cb9993b7ac8dfe89a3b2878e96bc7948e9f19ff51dba6", size = 84313 }, - { url = "https://files.pythonhosted.org/packages/84/ba/bd5d3806641564f2024e97ca98ea8f8811d4e01d9b9f9831474bc9e14f9e/audioop_lts-0.2.2-cp313-abi3-musllinux_1_2_ppc64le.whl", hash = "sha256:4a53aa7c16a60a6857e6b0b165261436396ef7293f8b5c9c828a3a203147ed4a", size = 93044 }, - { url = "https://files.pythonhosted.org/packages/f9/5e/435ce8d5642f1f7679540d1e73c1c42d933331c0976eb397d1717d7f01a3/audioop_lts-0.2.2-cp313-abi3-musllinux_1_2_riscv64.whl", hash = "sha256:3fc38008969796f0f689f1453722a0f463da1b8a6fbee11987830bfbb664f623", size = 78766 }, - { url = "https://files.pythonhosted.org/packages/ae/3b/b909e76b606cbfd53875693ec8c156e93e15a1366a012f0b7e4fb52d3c34/audioop_lts-0.2.2-cp313-abi3-musllinux_1_2_s390x.whl", hash = "sha256:15ab25dd3e620790f40e9ead897f91e79c0d3ce65fe193c8ed6c26cffdd24be7", size = 87640 }, - { url = "https://files.pythonhosted.org/packages/30/e7/8f1603b4572d79b775f2140d7952f200f5e6c62904585d08a01f0a70393a/audioop_lts-0.2.2-cp313-abi3-musllinux_1_2_x86_64.whl", hash = "sha256:03f061a1915538fd96272bac9551841859dbb2e3bf73ebe4a23ef043766f5449", size = 86052 }, - { url = "https://files.pythonhosted.org/packages/b5/96/c37846df657ccdda62ba1ae2b6534fa90e2e1b1742ca8dcf8ebd38c53801/audioop_lts-0.2.2-cp313-abi3-win32.whl", hash = "sha256:3bcddaaf6cc5935a300a8387c99f7a7fbbe212a11568ec6cf6e4bc458c048636", size = 26185 }, - { url = "https://files.pythonhosted.org/packages/34/a5/9d78fdb5b844a83da8a71226c7bdae7cc638861085fff7a1d707cb4823fa/audioop_lts-0.2.2-cp313-abi3-win_amd64.whl", hash = "sha256:a2c2a947fae7d1062ef08c4e369e0ba2086049a5e598fda41122535557012e9e", size = 30503 }, - { url = "https://files.pythonhosted.org/packages/34/25/20d8fde083123e90c61b51afb547bb0ea7e77bab50d98c0ab243d02a0e43/audioop_lts-0.2.2-cp313-abi3-win_arm64.whl", hash = "sha256:5f93a5db13927a37d2d09637ccca4b2b6b48c19cd9eda7b17a2e9f77edee6a6f", size = 24173 }, - { url = "https://files.pythonhosted.org/packages/58/a7/0a764f77b5c4ac58dc13c01a580f5d32ae8c74c92020b961556a43e26d02/audioop_lts-0.2.2-cp313-cp313t-macosx_10_13_universal2.whl", hash = "sha256:73f80bf4cd5d2ca7814da30a120de1f9408ee0619cc75da87d0641273d202a09", size = 47096 }, - { url = "https://files.pythonhosted.org/packages/aa/ed/ebebedde1a18848b085ad0fa54b66ceb95f1f94a3fc04f1cd1b5ccb0ed42/audioop_lts-0.2.2-cp313-cp313t-macosx_10_13_x86_64.whl", hash = "sha256:106753a83a25ee4d6f473f2be6b0966fc1c9af7e0017192f5531a3e7463dce58", size = 27748 }, - { url = "https://files.pythonhosted.org/packages/cb/6e/11ca8c21af79f15dbb1c7f8017952ee8c810c438ce4e2b25638dfef2b02c/audioop_lts-0.2.2-cp313-cp313t-macosx_11_0_arm64.whl", hash = "sha256:fbdd522624141e40948ab3e8cdae6e04c748d78710e9f0f8d4dae2750831de19", size = 27329 }, - { url = "https://files.pythonhosted.org/packages/84/52/0022f93d56d85eec5da6b9da6a958a1ef09e80c39f2cc0a590c6af81dcbb/audioop_lts-0.2.2-cp313-cp313t-manylinux1_x86_64.manylinux_2_28_x86_64.manylinux_2_5_x86_64.whl", hash = "sha256:143fad0311e8209ece30a8dbddab3b65ab419cbe8c0dde6e8828da25999be911", size = 92407 }, - { url = "https://files.pythonhosted.org/packages/87/1d/48a889855e67be8718adbc7a01f3c01d5743c325453a5e81cf3717664aad/audioop_lts-0.2.2-cp313-cp313t-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:dfbbc74ec68a0fd08cfec1f4b5e8cca3d3cd7de5501b01c4b5d209995033cde9", size = 91811 }, - { url = "https://files.pythonhosted.org/packages/98/a6/94b7213190e8077547ffae75e13ed05edc488653c85aa5c41472c297d295/audioop_lts-0.2.2-cp313-cp313t-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:cfcac6aa6f42397471e4943e0feb2244549db5c5d01efcd02725b96af417f3fe", size = 100470 }, - { url = "https://files.pythonhosted.org/packages/e9/e9/78450d7cb921ede0cfc33426d3a8023a3bda755883c95c868ee36db8d48d/audioop_lts-0.2.2-cp313-cp313t-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:752d76472d9804ac60f0078c79cdae8b956f293177acd2316cd1e15149aee132", size = 103878 }, - { url = "https://files.pythonhosted.org/packages/4f/e2/cd5439aad4f3e34ae1ee852025dc6aa8f67a82b97641e390bf7bd9891d3e/audioop_lts-0.2.2-cp313-cp313t-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl", hash = "sha256:83c381767e2cc10e93e40281a04852facc4cd9334550e0f392f72d1c0a9c5753", size = 84867 }, - { url = "https://files.pythonhosted.org/packages/68/4b/9d853e9076c43ebba0d411e8d2aa19061083349ac695a7d082540bad64d0/audioop_lts-0.2.2-cp313-cp313t-musllinux_1_2_aarch64.whl", hash = "sha256:c0022283e9556e0f3643b7c3c03f05063ca72b3063291834cca43234f20c60bb", size = 90001 }, - { url = "https://files.pythonhosted.org/packages/58/26/4bae7f9d2f116ed5593989d0e521d679b0d583973d203384679323d8fa85/audioop_lts-0.2.2-cp313-cp313t-musllinux_1_2_ppc64le.whl", hash = "sha256:a2d4f1513d63c795e82948e1305f31a6d530626e5f9f2605408b300ae6095093", size = 99046 }, - { url = "https://files.pythonhosted.org/packages/b2/67/a9f4fb3e250dda9e9046f8866e9fa7d52664f8985e445c6b4ad6dfb55641/audioop_lts-0.2.2-cp313-cp313t-musllinux_1_2_riscv64.whl", hash = "sha256:c9c8e68d8b4a56fda8c025e538e639f8c5953f5073886b596c93ec9b620055e7", size = 84788 }, - { url = "https://files.pythonhosted.org/packages/70/f7/3de86562db0121956148bcb0fe5b506615e3bcf6e63c4357a612b910765a/audioop_lts-0.2.2-cp313-cp313t-musllinux_1_2_s390x.whl", hash = "sha256:96f19de485a2925314f5020e85911fb447ff5fbef56e8c7c6927851b95533a1c", size = 94472 }, - { url = "https://files.pythonhosted.org/packages/f1/32/fd772bf9078ae1001207d2df1eef3da05bea611a87dd0e8217989b2848fa/audioop_lts-0.2.2-cp313-cp313t-musllinux_1_2_x86_64.whl", hash = "sha256:e541c3ef484852ef36545f66209444c48b28661e864ccadb29daddb6a4b8e5f5", size = 92279 }, - { url = "https://files.pythonhosted.org/packages/4f/41/affea7181592ab0ab560044632571a38edaf9130b84928177823fbf3176a/audioop_lts-0.2.2-cp313-cp313t-win32.whl", hash = "sha256:d5e73fa573e273e4f2e5ff96f9043858a5e9311e94ffefd88a3186a910c70917", size = 26568 }, - { url = "https://files.pythonhosted.org/packages/28/2b/0372842877016641db8fc54d5c88596b542eec2f8f6c20a36fb6612bf9ee/audioop_lts-0.2.2-cp313-cp313t-win_amd64.whl", hash = "sha256:9191d68659eda01e448188f60364c7763a7ca6653ed3f87ebb165822153a8547", size = 30942 }, - { url = "https://files.pythonhosted.org/packages/ee/ca/baf2b9cc7e96c179bb4a54f30fcd83e6ecb340031bde68f486403f943768/audioop_lts-0.2.2-cp313-cp313t-win_arm64.whl", hash = "sha256:c174e322bb5783c099aaf87faeb240c8d210686b04bd61dfd05a8e5a83d88969", size = 24603 }, - { url = "https://files.pythonhosted.org/packages/5c/73/413b5a2804091e2c7d5def1d618e4837f1cb82464e230f827226278556b7/audioop_lts-0.2.2-cp314-cp314t-macosx_10_13_universal2.whl", hash = "sha256:f9ee9b52f5f857fbaf9d605a360884f034c92c1c23021fb90b2e39b8e64bede6", size = 47104 }, - { url = "https://files.pythonhosted.org/packages/ae/8c/daa3308dc6593944410c2c68306a5e217f5c05b70a12e70228e7dd42dc5c/audioop_lts-0.2.2-cp314-cp314t-macosx_10_13_x86_64.whl", hash = "sha256:49ee1a41738a23e98d98b937a0638357a2477bc99e61b0f768a8f654f45d9b7a", size = 27754 }, - { url = "https://files.pythonhosted.org/packages/4e/86/c2e0f627168fcf61781a8f72cab06b228fe1da4b9fa4ab39cfb791b5836b/audioop_lts-0.2.2-cp314-cp314t-macosx_11_0_arm64.whl", hash = "sha256:5b00be98ccd0fc123dcfad31d50030d25fcf31488cde9e61692029cd7394733b", size = 27332 }, - { url = "https://files.pythonhosted.org/packages/c7/bd/35dce665255434f54e5307de39e31912a6f902d4572da7c37582809de14f/audioop_lts-0.2.2-cp314-cp314t-manylinux1_x86_64.manylinux_2_28_x86_64.manylinux_2_5_x86_64.whl", hash = "sha256:a6d2e0f9f7a69403e388894d4ca5ada5c47230716a03f2847cfc7bd1ecb589d6", size = 92396 }, - { url = "https://files.pythonhosted.org/packages/2d/d2/deeb9f51def1437b3afa35aeb729d577c04bcd89394cb56f9239a9f50b6f/audioop_lts-0.2.2-cp314-cp314t-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:f9b0b8a03ef474f56d1a842af1a2e01398b8f7654009823c6d9e0ecff4d5cfbf", size = 91811 }, - { url = "https://files.pythonhosted.org/packages/76/3b/09f8b35b227cee28cc8231e296a82759ed80c1a08e349811d69773c48426/audioop_lts-0.2.2-cp314-cp314t-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:2b267b70747d82125f1a021506565bdc5609a2b24bcb4773c16d79d2bb260bbd", size = 100483 }, - { url = "https://files.pythonhosted.org/packages/0b/15/05b48a935cf3b130c248bfdbdea71ce6437f5394ee8533e0edd7cfd93d5e/audioop_lts-0.2.2-cp314-cp314t-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:0337d658f9b81f4cd0fdb1f47635070cc084871a3d4646d9de74fdf4e7c3d24a", size = 103885 }, - { url = "https://files.pythonhosted.org/packages/83/80/186b7fce6d35b68d3d739f228dc31d60b3412105854edb975aa155a58339/audioop_lts-0.2.2-cp314-cp314t-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl", hash = "sha256:167d3b62586faef8b6b2275c3218796b12621a60e43f7e9d5845d627b9c9b80e", size = 84899 }, - { url = "https://files.pythonhosted.org/packages/49/89/c78cc5ac6cb5828f17514fb12966e299c850bc885e80f8ad94e38d450886/audioop_lts-0.2.2-cp314-cp314t-musllinux_1_2_aarch64.whl", hash = "sha256:0d9385e96f9f6da847f4d571ce3cb15b5091140edf3db97276872647ce37efd7", size = 89998 }, - { url = "https://files.pythonhosted.org/packages/4c/4b/6401888d0c010e586c2ca50fce4c903d70a6bb55928b16cfbdfd957a13da/audioop_lts-0.2.2-cp314-cp314t-musllinux_1_2_ppc64le.whl", hash = "sha256:48159d96962674eccdca9a3df280e864e8ac75e40a577cc97c5c42667ffabfc5", size = 99046 }, - { url = "https://files.pythonhosted.org/packages/de/f8/c874ca9bb447dae0e2ef2e231f6c4c2b0c39e31ae684d2420b0f9e97ee68/audioop_lts-0.2.2-cp314-cp314t-musllinux_1_2_riscv64.whl", hash = "sha256:8fefe5868cd082db1186f2837d64cfbfa78b548ea0d0543e9b28935ccce81ce9", size = 84843 }, - { url = "https://files.pythonhosted.org/packages/3e/c0/0323e66f3daebc13fd46b36b30c3be47e3fc4257eae44f1e77eb828c703f/audioop_lts-0.2.2-cp314-cp314t-musllinux_1_2_s390x.whl", hash = "sha256:58cf54380c3884fb49fdd37dfb7a772632b6701d28edd3e2904743c5e1773602", size = 94490 }, - { url = "https://files.pythonhosted.org/packages/98/6b/acc7734ac02d95ab791c10c3f17ffa3584ccb9ac5c18fd771c638ed6d1f5/audioop_lts-0.2.2-cp314-cp314t-musllinux_1_2_x86_64.whl", hash = "sha256:088327f00488cdeed296edd9215ca159f3a5a5034741465789cad403fcf4bec0", size = 92297 }, - { url = "https://files.pythonhosted.org/packages/13/c3/c3dc3f564ce6877ecd2a05f8d751b9b27a8c320c2533a98b0c86349778d0/audioop_lts-0.2.2-cp314-cp314t-win32.whl", hash = "sha256:068aa17a38b4e0e7de771c62c60bbca2455924b67a8814f3b0dee92b5820c0b3", size = 27331 }, - { url = "https://files.pythonhosted.org/packages/72/bb/b4608537e9ffcb86449091939d52d24a055216a36a8bf66b936af8c3e7ac/audioop_lts-0.2.2-cp314-cp314t-win_amd64.whl", hash = "sha256:a5bf613e96f49712073de86f20dbdd4014ca18efd4d34ed18c75bd808337851b", size = 31697 }, - { url = "https://files.pythonhosted.org/packages/f6/22/91616fe707a5c5510de2cac9b046a30defe7007ba8a0c04f9c08f27df312/audioop_lts-0.2.2-cp314-cp314t-win_arm64.whl", hash = "sha256:b492c3b040153e68b9fdaff5913305aaaba5bb433d8a7f73d5cf6a64ed3cc1dd", size = 25206 }, -] - -[[package]] -name = "audioread" -version = "3.1.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "standard-aifc", marker = "python_full_version >= '3.13'" }, - { name = "standard-sunau", marker = "python_full_version >= '3.13'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/a1/4a/874ecf9b472f998130c2b5e145dcdb9f6131e84786111489103b66772143/audioread-3.1.0.tar.gz", hash = "sha256:1c4ab2f2972764c896a8ac61ac53e261c8d29f0c6ccd652f84e18f08a4cab190", size = 20082 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/7e/16/fbe8e1e185a45042f7cd3a282def5bb8d95bb69ab9e9ef6a5368aa17e426/audioread-3.1.0-py3-none-any.whl", hash = "sha256:b30d1df6c5d3de5dcef0fb0e256f6ea17bdcf5f979408df0297d8a408e2971b4", size = 23143 }, -] - -[[package]] -name = "backports-datetime-fromisoformat" -version = "2.0.3" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/71/81/eff3184acb1d9dc3ce95a98b6f3c81a49b4be296e664db8e1c2eeabef3d9/backports_datetime_fromisoformat-2.0.3.tar.gz", hash = "sha256:b58edc8f517b66b397abc250ecc737969486703a66eb97e01e6d51291b1a139d", size = 23588 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/42/4b/d6b051ca4b3d76f23c2c436a9669f3be616b8cf6461a7e8061c7c4269642/backports_datetime_fromisoformat-2.0.3-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:5f681f638f10588fa3c101ee9ae2b63d3734713202ddfcfb6ec6cea0778a29d4", size = 27561 }, - { url = "https://files.pythonhosted.org/packages/6d/40/e39b0d471e55eb1b5c7c81edab605c02f71c786d59fb875f0a6f23318747/backports_datetime_fromisoformat-2.0.3-cp310-cp310-macosx_11_0_universal2.whl", hash = "sha256:cd681460e9142f1249408e5aee6d178c6d89b49e06d44913c8fdfb6defda8d1c", size = 34448 }, - { url = "https://files.pythonhosted.org/packages/f2/28/7a5c87c5561d14f1c9af979231fdf85d8f9fad7a95ff94e56d2205e2520a/backports_datetime_fromisoformat-2.0.3-cp310-cp310-macosx_11_0_x86_64.whl", hash = "sha256:ee68bc8735ae5058695b76d3bb2aee1d137c052a11c8303f1e966aa23b72b65b", size = 27093 }, - { url = "https://files.pythonhosted.org/packages/80/ba/f00296c5c4536967c7d1136107fdb91c48404fe769a4a6fd5ab045629af8/backports_datetime_fromisoformat-2.0.3-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8273fe7932db65d952a43e238318966eab9e49e8dd546550a41df12175cc2be4", size = 52836 }, - { url = "https://files.pythonhosted.org/packages/e3/92/bb1da57a069ddd601aee352a87262c7ae93467e66721d5762f59df5021a6/backports_datetime_fromisoformat-2.0.3-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:39d57ea50aa5a524bb239688adc1d1d824c31b6094ebd39aa164d6cadb85de22", size = 52798 }, - { url = "https://files.pythonhosted.org/packages/df/ef/b6cfd355982e817ccdb8d8d109f720cab6e06f900784b034b30efa8fa832/backports_datetime_fromisoformat-2.0.3-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:ac6272f87693e78209dc72e84cf9ab58052027733cd0721c55356d3c881791cf", size = 52891 }, - { url = "https://files.pythonhosted.org/packages/37/39/b13e3ae8a7c5d88b68a6e9248ffe7066534b0cfe504bf521963e61b6282d/backports_datetime_fromisoformat-2.0.3-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:44c497a71f80cd2bcfc26faae8857cf8e79388e3d5fbf79d2354b8c360547d58", size = 52955 }, - { url = "https://files.pythonhosted.org/packages/1e/e4/70cffa3ce1eb4f2ff0c0d6f5d56285aacead6bd3879b27a2ba57ab261172/backports_datetime_fromisoformat-2.0.3-cp310-cp310-win_amd64.whl", hash = "sha256:6335a4c9e8af329cb1ded5ab41a666e1448116161905a94e054f205aa6d263bc", size = 29323 }, - { url = "https://files.pythonhosted.org/packages/62/f5/5bc92030deadf34c365d908d4533709341fb05d0082db318774fdf1b2bcb/backports_datetime_fromisoformat-2.0.3-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:e2e4b66e017253cdbe5a1de49e0eecff3f66cd72bcb1229d7db6e6b1832c0443", size = 27626 }, - { url = "https://files.pythonhosted.org/packages/28/45/5885737d51f81dfcd0911dd5c16b510b249d4c4cf6f4a991176e0358a42a/backports_datetime_fromisoformat-2.0.3-cp311-cp311-macosx_11_0_universal2.whl", hash = "sha256:43e2d648e150777e13bbc2549cc960373e37bf65bd8a5d2e0cef40e16e5d8dd0", size = 34588 }, - { url = "https://files.pythonhosted.org/packages/bc/6d/bd74de70953f5dd3e768c8fc774af942af0ce9f211e7c38dd478fa7ea910/backports_datetime_fromisoformat-2.0.3-cp311-cp311-macosx_11_0_x86_64.whl", hash = "sha256:4ce6326fd86d5bae37813c7bf1543bae9e4c215ec6f5afe4c518be2635e2e005", size = 27162 }, - { url = "https://files.pythonhosted.org/packages/47/ba/1d14b097f13cce45b2b35db9898957578b7fcc984e79af3b35189e0d332f/backports_datetime_fromisoformat-2.0.3-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d7c8fac333bf860208fd522a5394369ee3c790d0aa4311f515fcc4b6c5ef8d75", size = 54482 }, - { url = "https://files.pythonhosted.org/packages/25/e9/a2a7927d053b6fa148b64b5e13ca741ca254c13edca99d8251e9a8a09cfe/backports_datetime_fromisoformat-2.0.3-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:24a4da5ab3aa0cc293dc0662a0c6d1da1a011dc1edcbc3122a288cfed13a0b45", size = 54362 }, - { url = "https://files.pythonhosted.org/packages/c1/99/394fb5e80131a7d58c49b89e78a61733a9994885804a0bb582416dd10c6f/backports_datetime_fromisoformat-2.0.3-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:58ea11e3bf912bd0a36b0519eae2c5b560b3cb972ea756e66b73fb9be460af01", size = 54162 }, - { url = "https://files.pythonhosted.org/packages/88/25/1940369de573c752889646d70b3fe8645e77b9e17984e72a554b9b51ffc4/backports_datetime_fromisoformat-2.0.3-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:8a375c7dbee4734318714a799b6c697223e4bbb57232af37fbfff88fb48a14c6", size = 54118 }, - { url = "https://files.pythonhosted.org/packages/b7/46/f275bf6c61683414acaf42b2df7286d68cfef03e98b45c168323d7707778/backports_datetime_fromisoformat-2.0.3-cp311-cp311-win_amd64.whl", hash = "sha256:ac677b1664c4585c2e014739f6678137c8336815406052349c85898206ec7061", size = 29329 }, - { url = "https://files.pythonhosted.org/packages/a2/0f/69bbdde2e1e57c09b5f01788804c50e68b29890aada999f2b1a40519def9/backports_datetime_fromisoformat-2.0.3-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:66ce47ee1ba91e146149cf40565c3d750ea1be94faf660ca733d8601e0848147", size = 27630 }, - { url = "https://files.pythonhosted.org/packages/d5/1d/1c84a50c673c87518b1adfeafcfd149991ed1f7aedc45d6e5eac2f7d19d7/backports_datetime_fromisoformat-2.0.3-cp312-cp312-macosx_11_0_universal2.whl", hash = "sha256:8b7e069910a66b3bba61df35b5f879e5253ff0821a70375b9daf06444d046fa4", size = 34707 }, - { url = "https://files.pythonhosted.org/packages/71/44/27eae384e7e045cda83f70b551d04b4a0b294f9822d32dea1cbf1592de59/backports_datetime_fromisoformat-2.0.3-cp312-cp312-macosx_11_0_x86_64.whl", hash = "sha256:a3b5d1d04a9e0f7b15aa1e647c750631a873b298cdd1255687bb68779fe8eb35", size = 27280 }, - { url = "https://files.pythonhosted.org/packages/a7/7a/a4075187eb6bbb1ff6beb7229db5f66d1070e6968abeb61e056fa51afa5e/backports_datetime_fromisoformat-2.0.3-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:ec1b95986430e789c076610aea704db20874f0781b8624f648ca9fb6ef67c6e1", size = 55094 }, - { url = "https://files.pythonhosted.org/packages/71/03/3fced4230c10af14aacadc195fe58e2ced91d011217b450c2e16a09a98c8/backports_datetime_fromisoformat-2.0.3-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ffe5f793db59e2f1d45ec35a1cf51404fdd69df9f6952a0c87c3060af4c00e32", size = 55605 }, - { url = "https://files.pythonhosted.org/packages/f6/0a/4b34a838c57bd16d3e5861ab963845e73a1041034651f7459e9935289cfd/backports_datetime_fromisoformat-2.0.3-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:620e8e73bd2595dfff1b4d256a12b67fce90ece3de87b38e1dde46b910f46f4d", size = 55353 }, - { url = "https://files.pythonhosted.org/packages/d9/68/07d13c6e98e1cad85606a876367ede2de46af859833a1da12c413c201d78/backports_datetime_fromisoformat-2.0.3-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:4cf9c0a985d68476c1cabd6385c691201dda2337d7453fb4da9679ce9f23f4e7", size = 55298 }, - { url = "https://files.pythonhosted.org/packages/60/33/45b4d5311f42360f9b900dea53ab2bb20a3d61d7f9b7c37ddfcb3962f86f/backports_datetime_fromisoformat-2.0.3-cp312-cp312-win_amd64.whl", hash = "sha256:d144868a73002e6e2e6fef72333e7b0129cecdd121aa8f1edba7107fd067255d", size = 29375 }, - { url = "https://files.pythonhosted.org/packages/be/03/7eaa9f9bf290395d57fd30d7f1f2f9dff60c06a31c237dc2beb477e8f899/backports_datetime_fromisoformat-2.0.3-pp310-pypy310_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:90e202e72a3d5aae673fcc8c9a4267d56b2f532beeb9173361293625fe4d2039", size = 28980 }, - { url = "https://files.pythonhosted.org/packages/47/80/a0ecf33446c7349e79f54cc532933780341d20cff0ee12b5bfdcaa47067e/backports_datetime_fromisoformat-2.0.3-pp310-pypy310_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2df98ef1b76f5a58bb493dda552259ba60c3a37557d848e039524203951c9f06", size = 28449 }, -] - -[[package]] -name = "braceexpand" -version = "0.1.7" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/54/93/badd4f5ccf25209f3fef2573073da9fe4a45a3da99fca2f800f942130c0f/braceexpand-0.1.7.tar.gz", hash = "sha256:e6e539bd20eaea53547472ff94f4fb5c3d3bf9d0a89388c4b56663aba765f705", size = 7777 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/fa/93/e8c04e80e82391a6e51f218ca49720f64236bc824e92152a2633b74cf7ab/braceexpand-0.1.7-py2.py3-none-any.whl", hash = "sha256:91332d53de7828103dcae5773fb43bc34950b0c8160e35e0f44c4427a3b85014", size = 5923 }, -] - -[[package]] -name = "cattrs" -version = "25.3.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "attrs" }, - { name = "exceptiongroup", marker = "python_full_version < '3.11'" }, - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/6e/00/2432bb2d445b39b5407f0a90e01b9a271475eea7caf913d7a86bcb956385/cattrs-25.3.0.tar.gz", hash = "sha256:1ac88d9e5eda10436c4517e390a4142d88638fe682c436c93db7ce4a277b884a", size = 509321 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/d8/2b/a40e1488fdfa02d3f9a653a61a5935ea08b3c2225ee818db6a76c7ba9695/cattrs-25.3.0-py3-none-any.whl", hash = "sha256:9896e84e0a5bf723bc7b4b68f4481785367ce07a8a02e7e9ee6eb2819bc306ff", size = 70738 }, -] - -[[package]] -name = "certifi" -version = "2025.11.12" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/a2/8c/58f469717fa48465e4a50c014a0400602d3c437d7c0c468e17ada824da3a/certifi-2025.11.12.tar.gz", hash = "sha256:d8ab5478f2ecd78af242878415affce761ca6bc54a22a27e026d7c25357c3316", size = 160538 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/70/7d/9bc192684cea499815ff478dfcdc13835ddf401365057044fb721ec6bddb/certifi-2025.11.12-py3-none-any.whl", hash = "sha256:97de8790030bbd5c2d96b7ec782fc2f7820ef8dba6db909ccf95449f2d062d4b", size = 159438 }, -] - -[[package]] -name = "cffi" -version = "2.0.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "pycparser", marker = "implementation_name != 'PyPy'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/eb/56/b1ba7935a17738ae8453301356628e8147c79dbb825bcbc73dc7401f9846/cffi-2.0.0.tar.gz", hash = "sha256:44d1b5909021139fe36001ae048dbdde8214afa20200eda0f64c068cac5d5529", size = 523588 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/93/d7/516d984057745a6cd96575eea814fe1edd6646ee6efd552fb7b0921dec83/cffi-2.0.0-cp310-cp310-macosx_10_13_x86_64.whl", hash = "sha256:0cf2d91ecc3fcc0625c2c530fe004f82c110405f101548512cce44322fa8ac44", size = 184283 }, - { url = "https://files.pythonhosted.org/packages/9e/84/ad6a0b408daa859246f57c03efd28e5dd1b33c21737c2db84cae8c237aa5/cffi-2.0.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:f73b96c41e3b2adedc34a7356e64c8eb96e03a3782b535e043a986276ce12a49", size = 180504 }, - { url = "https://files.pythonhosted.org/packages/50/bd/b1a6362b80628111e6653c961f987faa55262b4002fcec42308cad1db680/cffi-2.0.0-cp310-cp310-manylinux1_i686.manylinux2014_i686.manylinux_2_17_i686.manylinux_2_5_i686.whl", hash = "sha256:53f77cbe57044e88bbd5ed26ac1d0514d2acf0591dd6bb02a3ae37f76811b80c", size = 208811 }, - { url = "https://files.pythonhosted.org/packages/4f/27/6933a8b2562d7bd1fb595074cf99cc81fc3789f6a6c05cdabb46284a3188/cffi-2.0.0-cp310-cp310-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:3e837e369566884707ddaf85fc1744b47575005c0a229de3327f8f9a20f4efeb", size = 216402 }, - { url = "https://files.pythonhosted.org/packages/05/eb/b86f2a2645b62adcfff53b0dd97e8dfafb5c8aa864bd0d9a2c2049a0d551/cffi-2.0.0-cp310-cp310-manylinux2014_ppc64le.manylinux_2_17_ppc64le.whl", hash = "sha256:5eda85d6d1879e692d546a078b44251cdd08dd1cfb98dfb77b670c97cee49ea0", size = 203217 }, - { url = "https://files.pythonhosted.org/packages/9f/e0/6cbe77a53acf5acc7c08cc186c9928864bd7c005f9efd0d126884858a5fe/cffi-2.0.0-cp310-cp310-manylinux2014_s390x.manylinux_2_17_s390x.whl", hash = "sha256:9332088d75dc3241c702d852d4671613136d90fa6881da7d770a483fd05248b4", size = 203079 }, - { url = "https://files.pythonhosted.org/packages/98/29/9b366e70e243eb3d14a5cb488dfd3a0b6b2f1fb001a203f653b93ccfac88/cffi-2.0.0-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:fc7de24befaeae77ba923797c7c87834c73648a05a4bde34b3b7e5588973a453", size = 216475 }, - { url = "https://files.pythonhosted.org/packages/21/7a/13b24e70d2f90a322f2900c5d8e1f14fa7e2a6b3332b7309ba7b2ba51a5a/cffi-2.0.0-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:cf364028c016c03078a23b503f02058f1814320a56ad535686f90565636a9495", size = 218829 }, - { url = "https://files.pythonhosted.org/packages/60/99/c9dc110974c59cc981b1f5b66e1d8af8af764e00f0293266824d9c4254bc/cffi-2.0.0-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:e11e82b744887154b182fd3e7e8512418446501191994dbf9c9fc1f32cc8efd5", size = 211211 }, - { url = "https://files.pythonhosted.org/packages/49/72/ff2d12dbf21aca1b32a40ed792ee6b40f6dc3a9cf1644bd7ef6e95e0ac5e/cffi-2.0.0-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:8ea985900c5c95ce9db1745f7933eeef5d314f0565b27625d9a10ec9881e1bfb", size = 218036 }, - { url = "https://files.pythonhosted.org/packages/e2/cc/027d7fb82e58c48ea717149b03bcadcbdc293553edb283af792bd4bcbb3f/cffi-2.0.0-cp310-cp310-win32.whl", hash = "sha256:1f72fb8906754ac8a2cc3f9f5aaa298070652a0ffae577e0ea9bd480dc3c931a", size = 172184 }, - { url = "https://files.pythonhosted.org/packages/33/fa/072dd15ae27fbb4e06b437eb6e944e75b068deb09e2a2826039e49ee2045/cffi-2.0.0-cp310-cp310-win_amd64.whl", hash = "sha256:b18a3ed7d5b3bd8d9ef7a8cb226502c6bf8308df1525e1cc676c3680e7176739", size = 182790 }, - { url = "https://files.pythonhosted.org/packages/12/4a/3dfd5f7850cbf0d06dc84ba9aa00db766b52ca38d8b86e3a38314d52498c/cffi-2.0.0-cp311-cp311-macosx_10_13_x86_64.whl", hash = "sha256:b4c854ef3adc177950a8dfc81a86f5115d2abd545751a304c5bcf2c2c7283cfe", size = 184344 }, - { url = "https://files.pythonhosted.org/packages/4f/8b/f0e4c441227ba756aafbe78f117485b25bb26b1c059d01f137fa6d14896b/cffi-2.0.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:2de9a304e27f7596cd03d16f1b7c72219bd944e99cc52b84d0145aefb07cbd3c", size = 180560 }, - { url = "https://files.pythonhosted.org/packages/b1/b7/1200d354378ef52ec227395d95c2576330fd22a869f7a70e88e1447eb234/cffi-2.0.0-cp311-cp311-manylinux1_i686.manylinux2014_i686.manylinux_2_17_i686.manylinux_2_5_i686.whl", hash = "sha256:baf5215e0ab74c16e2dd324e8ec067ef59e41125d3eade2b863d294fd5035c92", size = 209613 }, - { url = "https://files.pythonhosted.org/packages/b8/56/6033f5e86e8cc9bb629f0077ba71679508bdf54a9a5e112a3c0b91870332/cffi-2.0.0-cp311-cp311-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:730cacb21e1bdff3ce90babf007d0a0917cc3e6492f336c2f0134101e0944f93", size = 216476 }, - { url = "https://files.pythonhosted.org/packages/dc/7f/55fecd70f7ece178db2f26128ec41430d8720f2d12ca97bf8f0a628207d5/cffi-2.0.0-cp311-cp311-manylinux2014_ppc64le.manylinux_2_17_ppc64le.whl", hash = "sha256:6824f87845e3396029f3820c206e459ccc91760e8fa24422f8b0c3d1731cbec5", size = 203374 }, - { url = "https://files.pythonhosted.org/packages/84/ef/a7b77c8bdc0f77adc3b46888f1ad54be8f3b7821697a7b89126e829e676a/cffi-2.0.0-cp311-cp311-manylinux2014_s390x.manylinux_2_17_s390x.whl", hash = "sha256:9de40a7b0323d889cf8d23d1ef214f565ab154443c42737dfe52ff82cf857664", size = 202597 }, - { url = "https://files.pythonhosted.org/packages/d7/91/500d892b2bf36529a75b77958edfcd5ad8e2ce4064ce2ecfeab2125d72d1/cffi-2.0.0-cp311-cp311-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:8941aaadaf67246224cee8c3803777eed332a19d909b47e29c9842ef1e79ac26", size = 215574 }, - { url = "https://files.pythonhosted.org/packages/44/64/58f6255b62b101093d5df22dcb752596066c7e89dd725e0afaed242a61be/cffi-2.0.0-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:a05d0c237b3349096d3981b727493e22147f934b20f6f125a3eba8f994bec4a9", size = 218971 }, - { url = "https://files.pythonhosted.org/packages/ab/49/fa72cebe2fd8a55fbe14956f9970fe8eb1ac59e5df042f603ef7c8ba0adc/cffi-2.0.0-cp311-cp311-musllinux_1_2_i686.whl", hash = "sha256:94698a9c5f91f9d138526b48fe26a199609544591f859c870d477351dc7b2414", size = 211972 }, - { url = "https://files.pythonhosted.org/packages/0b/28/dd0967a76aab36731b6ebfe64dec4e981aff7e0608f60c2d46b46982607d/cffi-2.0.0-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:5fed36fccc0612a53f1d4d9a816b50a36702c28a2aa880cb8a122b3466638743", size = 217078 }, - { url = "https://files.pythonhosted.org/packages/2b/c0/015b25184413d7ab0a410775fdb4a50fca20f5589b5dab1dbbfa3baad8ce/cffi-2.0.0-cp311-cp311-win32.whl", hash = "sha256:c649e3a33450ec82378822b3dad03cc228b8f5963c0c12fc3b1e0ab940f768a5", size = 172076 }, - { url = "https://files.pythonhosted.org/packages/ae/8f/dc5531155e7070361eb1b7e4c1a9d896d0cb21c49f807a6c03fd63fc877e/cffi-2.0.0-cp311-cp311-win_amd64.whl", hash = "sha256:66f011380d0e49ed280c789fbd08ff0d40968ee7b665575489afa95c98196ab5", size = 182820 }, - { url = "https://files.pythonhosted.org/packages/95/5c/1b493356429f9aecfd56bc171285a4c4ac8697f76e9bbbbb105e537853a1/cffi-2.0.0-cp311-cp311-win_arm64.whl", hash = "sha256:c6638687455baf640e37344fe26d37c404db8b80d037c3d29f58fe8d1c3b194d", size = 177635 }, - { url = "https://files.pythonhosted.org/packages/ea/47/4f61023ea636104d4f16ab488e268b93008c3d0bb76893b1b31db1f96802/cffi-2.0.0-cp312-cp312-macosx_10_13_x86_64.whl", hash = "sha256:6d02d6655b0e54f54c4ef0b94eb6be0607b70853c45ce98bd278dc7de718be5d", size = 185271 }, - { url = "https://files.pythonhosted.org/packages/df/a2/781b623f57358e360d62cdd7a8c681f074a71d445418a776eef0aadb4ab4/cffi-2.0.0-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:8eca2a813c1cb7ad4fb74d368c2ffbbb4789d377ee5bb8df98373c2cc0dee76c", size = 181048 }, - { url = "https://files.pythonhosted.org/packages/ff/df/a4f0fbd47331ceeba3d37c2e51e9dfc9722498becbeec2bd8bc856c9538a/cffi-2.0.0-cp312-cp312-manylinux1_i686.manylinux2014_i686.manylinux_2_17_i686.manylinux_2_5_i686.whl", hash = "sha256:21d1152871b019407d8ac3985f6775c079416c282e431a4da6afe7aefd2bccbe", size = 212529 }, - { url = "https://files.pythonhosted.org/packages/d5/72/12b5f8d3865bf0f87cf1404d8c374e7487dcf097a1c91c436e72e6badd83/cffi-2.0.0-cp312-cp312-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:b21e08af67b8a103c71a250401c78d5e0893beff75e28c53c98f4de42f774062", size = 220097 }, - { url = "https://files.pythonhosted.org/packages/c2/95/7a135d52a50dfa7c882ab0ac17e8dc11cec9d55d2c18dda414c051c5e69e/cffi-2.0.0-cp312-cp312-manylinux2014_ppc64le.manylinux_2_17_ppc64le.whl", hash = "sha256:1e3a615586f05fc4065a8b22b8152f0c1b00cdbc60596d187c2a74f9e3036e4e", size = 207983 }, - { url = "https://files.pythonhosted.org/packages/3a/c8/15cb9ada8895957ea171c62dc78ff3e99159ee7adb13c0123c001a2546c1/cffi-2.0.0-cp312-cp312-manylinux2014_s390x.manylinux_2_17_s390x.whl", hash = "sha256:81afed14892743bbe14dacb9e36d9e0e504cd204e0b165062c488942b9718037", size = 206519 }, - { url = "https://files.pythonhosted.org/packages/78/2d/7fa73dfa841b5ac06c7b8855cfc18622132e365f5b81d02230333ff26e9e/cffi-2.0.0-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:3e17ed538242334bf70832644a32a7aae3d83b57567f9fd60a26257e992b79ba", size = 219572 }, - { url = "https://files.pythonhosted.org/packages/07/e0/267e57e387b4ca276b90f0434ff88b2c2241ad72b16d31836adddfd6031b/cffi-2.0.0-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:3925dd22fa2b7699ed2617149842d2e6adde22b262fcbfada50e3d195e4b3a94", size = 222963 }, - { url = "https://files.pythonhosted.org/packages/b6/75/1f2747525e06f53efbd878f4d03bac5b859cbc11c633d0fb81432d98a795/cffi-2.0.0-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:2c8f814d84194c9ea681642fd164267891702542f028a15fc97d4674b6206187", size = 221361 }, - { url = "https://files.pythonhosted.org/packages/7b/2b/2b6435f76bfeb6bbf055596976da087377ede68df465419d192acf00c437/cffi-2.0.0-cp312-cp312-win32.whl", hash = "sha256:da902562c3e9c550df360bfa53c035b2f241fed6d9aef119048073680ace4a18", size = 172932 }, - { url = "https://files.pythonhosted.org/packages/f8/ed/13bd4418627013bec4ed6e54283b1959cf6db888048c7cf4b4c3b5b36002/cffi-2.0.0-cp312-cp312-win_amd64.whl", hash = "sha256:da68248800ad6320861f129cd9c1bf96ca849a2771a59e0344e88681905916f5", size = 183557 }, - { url = "https://files.pythonhosted.org/packages/95/31/9f7f93ad2f8eff1dbc1c3656d7ca5bfd8fb52c9d786b4dcf19b2d02217fa/cffi-2.0.0-cp312-cp312-win_arm64.whl", hash = "sha256:4671d9dd5ec934cb9a73e7ee9676f9362aba54f7f34910956b84d727b0d73fb6", size = 177762 }, - { url = "https://files.pythonhosted.org/packages/4b/8d/a0a47a0c9e413a658623d014e91e74a50cdd2c423f7ccfd44086ef767f90/cffi-2.0.0-cp313-cp313-macosx_10_13_x86_64.whl", hash = "sha256:00bdf7acc5f795150faa6957054fbbca2439db2f775ce831222b66f192f03beb", size = 185230 }, - { url = "https://files.pythonhosted.org/packages/4a/d2/a6c0296814556c68ee32009d9c2ad4f85f2707cdecfd7727951ec228005d/cffi-2.0.0-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:45d5e886156860dc35862657e1494b9bae8dfa63bf56796f2fb56e1679fc0bca", size = 181043 }, - { url = "https://files.pythonhosted.org/packages/b0/1e/d22cc63332bd59b06481ceaac49d6c507598642e2230f201649058a7e704/cffi-2.0.0-cp313-cp313-manylinux1_i686.manylinux2014_i686.manylinux_2_17_i686.manylinux_2_5_i686.whl", hash = "sha256:07b271772c100085dd28b74fa0cd81c8fb1a3ba18b21e03d7c27f3436a10606b", size = 212446 }, - { url = "https://files.pythonhosted.org/packages/a9/f5/a2c23eb03b61a0b8747f211eb716446c826ad66818ddc7810cc2cc19b3f2/cffi-2.0.0-cp313-cp313-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:d48a880098c96020b02d5a1f7d9251308510ce8858940e6fa99ece33f610838b", size = 220101 }, - { url = "https://files.pythonhosted.org/packages/f2/7f/e6647792fc5850d634695bc0e6ab4111ae88e89981d35ac269956605feba/cffi-2.0.0-cp313-cp313-manylinux2014_ppc64le.manylinux_2_17_ppc64le.whl", hash = "sha256:f93fd8e5c8c0a4aa1f424d6173f14a892044054871c771f8566e4008eaa359d2", size = 207948 }, - { url = "https://files.pythonhosted.org/packages/cb/1e/a5a1bd6f1fb30f22573f76533de12a00bf274abcdc55c8edab639078abb6/cffi-2.0.0-cp313-cp313-manylinux2014_s390x.manylinux_2_17_s390x.whl", hash = "sha256:dd4f05f54a52fb558f1ba9f528228066954fee3ebe629fc1660d874d040ae5a3", size = 206422 }, - { url = "https://files.pythonhosted.org/packages/98/df/0a1755e750013a2081e863e7cd37e0cdd02664372c754e5560099eb7aa44/cffi-2.0.0-cp313-cp313-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:c8d3b5532fc71b7a77c09192b4a5a200ea992702734a2e9279a37f2478236f26", size = 219499 }, - { url = "https://files.pythonhosted.org/packages/50/e1/a969e687fcf9ea58e6e2a928ad5e2dd88cc12f6f0ab477e9971f2309b57c/cffi-2.0.0-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:d9b29c1f0ae438d5ee9acb31cadee00a58c46cc9c0b2f9038c6b0b3470877a8c", size = 222928 }, - { url = "https://files.pythonhosted.org/packages/36/54/0362578dd2c9e557a28ac77698ed67323ed5b9775ca9d3fe73fe191bb5d8/cffi-2.0.0-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:6d50360be4546678fc1b79ffe7a66265e28667840010348dd69a314145807a1b", size = 221302 }, - { url = "https://files.pythonhosted.org/packages/eb/6d/bf9bda840d5f1dfdbf0feca87fbdb64a918a69bca42cfa0ba7b137c48cb8/cffi-2.0.0-cp313-cp313-win32.whl", hash = "sha256:74a03b9698e198d47562765773b4a8309919089150a0bb17d829ad7b44b60d27", size = 172909 }, - { url = "https://files.pythonhosted.org/packages/37/18/6519e1ee6f5a1e579e04b9ddb6f1676c17368a7aba48299c3759bbc3c8b3/cffi-2.0.0-cp313-cp313-win_amd64.whl", hash = "sha256:19f705ada2530c1167abacb171925dd886168931e0a7b78f5bffcae5c6b5be75", size = 183402 }, - { url = "https://files.pythonhosted.org/packages/cb/0e/02ceeec9a7d6ee63bb596121c2c8e9b3a9e150936f4fbef6ca1943e6137c/cffi-2.0.0-cp313-cp313-win_arm64.whl", hash = "sha256:256f80b80ca3853f90c21b23ee78cd008713787b1b1e93eae9f3d6a7134abd91", size = 177780 }, - { url = "https://files.pythonhosted.org/packages/92/c4/3ce07396253a83250ee98564f8d7e9789fab8e58858f35d07a9a2c78de9f/cffi-2.0.0-cp314-cp314-macosx_10_13_x86_64.whl", hash = "sha256:fc33c5141b55ed366cfaad382df24fe7dcbc686de5be719b207bb248e3053dc5", size = 185320 }, - { url = "https://files.pythonhosted.org/packages/59/dd/27e9fa567a23931c838c6b02d0764611c62290062a6d4e8ff7863daf9730/cffi-2.0.0-cp314-cp314-macosx_11_0_arm64.whl", hash = "sha256:c654de545946e0db659b3400168c9ad31b5d29593291482c43e3564effbcee13", size = 181487 }, - { url = "https://files.pythonhosted.org/packages/d6/43/0e822876f87ea8a4ef95442c3d766a06a51fc5298823f884ef87aaad168c/cffi-2.0.0-cp314-cp314-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:24b6f81f1983e6df8db3adc38562c83f7d4a0c36162885ec7f7b77c7dcbec97b", size = 220049 }, - { url = "https://files.pythonhosted.org/packages/b4/89/76799151d9c2d2d1ead63c2429da9ea9d7aac304603de0c6e8764e6e8e70/cffi-2.0.0-cp314-cp314-manylinux2014_ppc64le.manylinux_2_17_ppc64le.whl", hash = "sha256:12873ca6cb9b0f0d3a0da705d6086fe911591737a59f28b7936bdfed27c0d47c", size = 207793 }, - { url = "https://files.pythonhosted.org/packages/bb/dd/3465b14bb9e24ee24cb88c9e3730f6de63111fffe513492bf8c808a3547e/cffi-2.0.0-cp314-cp314-manylinux2014_s390x.manylinux_2_17_s390x.whl", hash = "sha256:d9b97165e8aed9272a6bb17c01e3cc5871a594a446ebedc996e2397a1c1ea8ef", size = 206300 }, - { url = "https://files.pythonhosted.org/packages/47/d9/d83e293854571c877a92da46fdec39158f8d7e68da75bf73581225d28e90/cffi-2.0.0-cp314-cp314-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:afb8db5439b81cf9c9d0c80404b60c3cc9c3add93e114dcae767f1477cb53775", size = 219244 }, - { url = "https://files.pythonhosted.org/packages/2b/0f/1f177e3683aead2bb00f7679a16451d302c436b5cbf2505f0ea8146ef59e/cffi-2.0.0-cp314-cp314-musllinux_1_2_aarch64.whl", hash = "sha256:737fe7d37e1a1bffe70bd5754ea763a62a066dc5913ca57e957824b72a85e205", size = 222828 }, - { url = "https://files.pythonhosted.org/packages/c6/0f/cafacebd4b040e3119dcb32fed8bdef8dfe94da653155f9d0b9dc660166e/cffi-2.0.0-cp314-cp314-musllinux_1_2_x86_64.whl", hash = "sha256:38100abb9d1b1435bc4cc340bb4489635dc2f0da7456590877030c9b3d40b0c1", size = 220926 }, - { url = "https://files.pythonhosted.org/packages/3e/aa/df335faa45b395396fcbc03de2dfcab242cd61a9900e914fe682a59170b1/cffi-2.0.0-cp314-cp314-win32.whl", hash = "sha256:087067fa8953339c723661eda6b54bc98c5625757ea62e95eb4898ad5e776e9f", size = 175328 }, - { url = "https://files.pythonhosted.org/packages/bb/92/882c2d30831744296ce713f0feb4c1cd30f346ef747b530b5318715cc367/cffi-2.0.0-cp314-cp314-win_amd64.whl", hash = "sha256:203a48d1fb583fc7d78a4c6655692963b860a417c0528492a6bc21f1aaefab25", size = 185650 }, - { url = "https://files.pythonhosted.org/packages/9f/2c/98ece204b9d35a7366b5b2c6539c350313ca13932143e79dc133ba757104/cffi-2.0.0-cp314-cp314-win_arm64.whl", hash = "sha256:dbd5c7a25a7cb98f5ca55d258b103a2054f859a46ae11aaf23134f9cc0d356ad", size = 180687 }, - { url = "https://files.pythonhosted.org/packages/3e/61/c768e4d548bfa607abcda77423448df8c471f25dbe64fb2ef6d555eae006/cffi-2.0.0-cp314-cp314t-macosx_10_13_x86_64.whl", hash = "sha256:9a67fc9e8eb39039280526379fb3a70023d77caec1852002b4da7e8b270c4dd9", size = 188773 }, - { url = "https://files.pythonhosted.org/packages/2c/ea/5f76bce7cf6fcd0ab1a1058b5af899bfbef198bea4d5686da88471ea0336/cffi-2.0.0-cp314-cp314t-macosx_11_0_arm64.whl", hash = "sha256:7a66c7204d8869299919db4d5069a82f1561581af12b11b3c9f48c584eb8743d", size = 185013 }, - { url = "https://files.pythonhosted.org/packages/be/b4/c56878d0d1755cf9caa54ba71e5d049479c52f9e4afc230f06822162ab2f/cffi-2.0.0-cp314-cp314t-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:7cc09976e8b56f8cebd752f7113ad07752461f48a58cbba644139015ac24954c", size = 221593 }, - { url = "https://files.pythonhosted.org/packages/e0/0d/eb704606dfe8033e7128df5e90fee946bbcb64a04fcdaa97321309004000/cffi-2.0.0-cp314-cp314t-manylinux2014_ppc64le.manylinux_2_17_ppc64le.whl", hash = "sha256:92b68146a71df78564e4ef48af17551a5ddd142e5190cdf2c5624d0c3ff5b2e8", size = 209354 }, - { url = "https://files.pythonhosted.org/packages/d8/19/3c435d727b368ca475fb8742ab97c9cb13a0de600ce86f62eab7fa3eea60/cffi-2.0.0-cp314-cp314t-manylinux2014_s390x.manylinux_2_17_s390x.whl", hash = "sha256:b1e74d11748e7e98e2f426ab176d4ed720a64412b6a15054378afdb71e0f37dc", size = 208480 }, - { url = "https://files.pythonhosted.org/packages/d0/44/681604464ed9541673e486521497406fadcc15b5217c3e326b061696899a/cffi-2.0.0-cp314-cp314t-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:28a3a209b96630bca57cce802da70c266eb08c6e97e5afd61a75611ee6c64592", size = 221584 }, - { url = "https://files.pythonhosted.org/packages/25/8e/342a504ff018a2825d395d44d63a767dd8ebc927ebda557fecdaca3ac33a/cffi-2.0.0-cp314-cp314t-musllinux_1_2_aarch64.whl", hash = "sha256:7553fb2090d71822f02c629afe6042c299edf91ba1bf94951165613553984512", size = 224443 }, - { url = "https://files.pythonhosted.org/packages/e1/5e/b666bacbbc60fbf415ba9988324a132c9a7a0448a9a8f125074671c0f2c3/cffi-2.0.0-cp314-cp314t-musllinux_1_2_x86_64.whl", hash = "sha256:6c6c373cfc5c83a975506110d17457138c8c63016b563cc9ed6e056a82f13ce4", size = 223437 }, - { url = "https://files.pythonhosted.org/packages/a0/1d/ec1a60bd1a10daa292d3cd6bb0b359a81607154fb8165f3ec95fe003b85c/cffi-2.0.0-cp314-cp314t-win32.whl", hash = "sha256:1fc9ea04857caf665289b7a75923f2c6ed559b8298a1b8c49e59f7dd95c8481e", size = 180487 }, - { url = "https://files.pythonhosted.org/packages/bf/41/4c1168c74fac325c0c8156f04b6749c8b6a8f405bbf91413ba088359f60d/cffi-2.0.0-cp314-cp314t-win_amd64.whl", hash = "sha256:d68b6cef7827e8641e8ef16f4494edda8b36104d79773a334beaa1e3521430f6", size = 191726 }, - { url = "https://files.pythonhosted.org/packages/ae/3a/dbeec9d1ee0844c679f6bb5d6ad4e9f198b1224f4e7a32825f47f6192b0c/cffi-2.0.0-cp314-cp314t-win_arm64.whl", hash = "sha256:0a1527a803f0a659de1af2e1fd700213caba79377e27e4693648c2923da066f9", size = 184195 }, -] - -[[package]] -name = "charset-normalizer" -version = "3.4.4" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/13/69/33ddede1939fdd074bce5434295f38fae7136463422fe4fd3e0e89b98062/charset_normalizer-3.4.4.tar.gz", hash = "sha256:94537985111c35f28720e43603b8e7b43a6ecfb2ce1d3058bbe955b73404e21a", size = 129418 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/1f/b8/6d51fc1d52cbd52cd4ccedd5b5b2f0f6a11bbf6765c782298b0f3e808541/charset_normalizer-3.4.4-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:e824f1492727fa856dd6eda4f7cee25f8518a12f3c4a56a74e8095695089cf6d", size = 209709 }, - { url = "https://files.pythonhosted.org/packages/5c/af/1f9d7f7faafe2ddfb6f72a2e07a548a629c61ad510fe60f9630309908fef/charset_normalizer-3.4.4-cp310-cp310-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:4bd5d4137d500351a30687c2d3971758aac9a19208fc110ccb9d7188fbe709e8", size = 148814 }, - { url = "https://files.pythonhosted.org/packages/79/3d/f2e3ac2bbc056ca0c204298ea4e3d9db9b4afe437812638759db2c976b5f/charset_normalizer-3.4.4-cp310-cp310-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:027f6de494925c0ab2a55eab46ae5129951638a49a34d87f4c3eda90f696b4ad", size = 144467 }, - { url = "https://files.pythonhosted.org/packages/ec/85/1bf997003815e60d57de7bd972c57dc6950446a3e4ccac43bc3070721856/charset_normalizer-3.4.4-cp310-cp310-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:f820802628d2694cb7e56db99213f930856014862f3fd943d290ea8438d07ca8", size = 162280 }, - { url = "https://files.pythonhosted.org/packages/3e/8e/6aa1952f56b192f54921c436b87f2aaf7c7a7c3d0d1a765547d64fd83c13/charset_normalizer-3.4.4-cp310-cp310-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:798d75d81754988d2565bff1b97ba5a44411867c0cf32b77a7e8f8d84796b10d", size = 159454 }, - { url = "https://files.pythonhosted.org/packages/36/3b/60cbd1f8e93aa25d1c669c649b7a655b0b5fb4c571858910ea9332678558/charset_normalizer-3.4.4-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:9d1bb833febdff5c8927f922386db610b49db6e0d4f4ee29601d71e7c2694313", size = 153609 }, - { url = "https://files.pythonhosted.org/packages/64/91/6a13396948b8fd3c4b4fd5bc74d045f5637d78c9675585e8e9fbe5636554/charset_normalizer-3.4.4-cp310-cp310-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl", hash = "sha256:9cd98cdc06614a2f768d2b7286d66805f94c48cde050acdbbb7db2600ab3197e", size = 151849 }, - { url = "https://files.pythonhosted.org/packages/b7/7a/59482e28b9981d105691e968c544cc0df3b7d6133152fb3dcdc8f135da7a/charset_normalizer-3.4.4-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:077fbb858e903c73f6c9db43374fd213b0b6a778106bc7032446a8e8b5b38b93", size = 151586 }, - { url = "https://files.pythonhosted.org/packages/92/59/f64ef6a1c4bdd2baf892b04cd78792ed8684fbc48d4c2afe467d96b4df57/charset_normalizer-3.4.4-cp310-cp310-musllinux_1_2_armv7l.whl", hash = "sha256:244bfb999c71b35de57821b8ea746b24e863398194a4014e4c76adc2bbdfeff0", size = 145290 }, - { url = "https://files.pythonhosted.org/packages/6b/63/3bf9f279ddfa641ffa1962b0db6a57a9c294361cc2f5fcac997049a00e9c/charset_normalizer-3.4.4-cp310-cp310-musllinux_1_2_ppc64le.whl", hash = "sha256:64b55f9dce520635f018f907ff1b0df1fdc31f2795a922fb49dd14fbcdf48c84", size = 163663 }, - { url = "https://files.pythonhosted.org/packages/ed/09/c9e38fc8fa9e0849b172b581fd9803bdf6e694041127933934184e19f8c3/charset_normalizer-3.4.4-cp310-cp310-musllinux_1_2_riscv64.whl", hash = "sha256:faa3a41b2b66b6e50f84ae4a68c64fcd0c44355741c6374813a800cd6695db9e", size = 151964 }, - { url = "https://files.pythonhosted.org/packages/d2/d1/d28b747e512d0da79d8b6a1ac18b7ab2ecfd81b2944c4c710e166d8dd09c/charset_normalizer-3.4.4-cp310-cp310-musllinux_1_2_s390x.whl", hash = "sha256:6515f3182dbe4ea06ced2d9e8666d97b46ef4c75e326b79bb624110f122551db", size = 161064 }, - { url = "https://files.pythonhosted.org/packages/bb/9a/31d62b611d901c3b9e5500c36aab0ff5eb442043fb3a1c254200d3d397d9/charset_normalizer-3.4.4-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:cc00f04ed596e9dc0da42ed17ac5e596c6ccba999ba6bd92b0e0aef2f170f2d6", size = 155015 }, - { url = "https://files.pythonhosted.org/packages/1f/f3/107e008fa2bff0c8b9319584174418e5e5285fef32f79d8ee6a430d0039c/charset_normalizer-3.4.4-cp310-cp310-win32.whl", hash = "sha256:f34be2938726fc13801220747472850852fe6b1ea75869a048d6f896838c896f", size = 99792 }, - { url = "https://files.pythonhosted.org/packages/eb/66/e396e8a408843337d7315bab30dbf106c38966f1819f123257f5520f8a96/charset_normalizer-3.4.4-cp310-cp310-win_amd64.whl", hash = "sha256:a61900df84c667873b292c3de315a786dd8dac506704dea57bc957bd31e22c7d", size = 107198 }, - { url = "https://files.pythonhosted.org/packages/b5/58/01b4f815bf0312704c267f2ccb6e5d42bcc7752340cd487bc9f8c3710597/charset_normalizer-3.4.4-cp310-cp310-win_arm64.whl", hash = "sha256:cead0978fc57397645f12578bfd2d5ea9138ea0fac82b2f63f7f7c6877986a69", size = 100262 }, - { url = "https://files.pythonhosted.org/packages/ed/27/c6491ff4954e58a10f69ad90aca8a1b6fe9c5d3c6f380907af3c37435b59/charset_normalizer-3.4.4-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:6e1fcf0720908f200cd21aa4e6750a48ff6ce4afe7ff5a79a90d5ed8a08296f8", size = 206988 }, - { url = "https://files.pythonhosted.org/packages/94/59/2e87300fe67ab820b5428580a53cad894272dbb97f38a7a814a2a1ac1011/charset_normalizer-3.4.4-cp311-cp311-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:5f819d5fe9234f9f82d75bdfa9aef3a3d72c4d24a6e57aeaebba32a704553aa0", size = 147324 }, - { url = "https://files.pythonhosted.org/packages/07/fb/0cf61dc84b2b088391830f6274cb57c82e4da8bbc2efeac8c025edb88772/charset_normalizer-3.4.4-cp311-cp311-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:a59cb51917aa591b1c4e6a43c132f0cdc3c76dbad6155df4e28ee626cc77a0a3", size = 142742 }, - { url = "https://files.pythonhosted.org/packages/62/8b/171935adf2312cd745d290ed93cf16cf0dfe320863ab7cbeeae1dcd6535f/charset_normalizer-3.4.4-cp311-cp311-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:8ef3c867360f88ac904fd3f5e1f902f13307af9052646963ee08ff4f131adafc", size = 160863 }, - { url = "https://files.pythonhosted.org/packages/09/73/ad875b192bda14f2173bfc1bc9a55e009808484a4b256748d931b6948442/charset_normalizer-3.4.4-cp311-cp311-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:d9e45d7faa48ee908174d8fe84854479ef838fc6a705c9315372eacbc2f02897", size = 157837 }, - { url = "https://files.pythonhosted.org/packages/6d/fc/de9cce525b2c5b94b47c70a4b4fb19f871b24995c728e957ee68ab1671ea/charset_normalizer-3.4.4-cp311-cp311-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:840c25fb618a231545cbab0564a799f101b63b9901f2569faecd6b222ac72381", size = 151550 }, - { url = "https://files.pythonhosted.org/packages/55/c2/43edd615fdfba8c6f2dfbd459b25a6b3b551f24ea21981e23fb768503ce1/charset_normalizer-3.4.4-cp311-cp311-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl", hash = "sha256:ca5862d5b3928c4940729dacc329aa9102900382fea192fc5e52eb69d6093815", size = 149162 }, - { url = "https://files.pythonhosted.org/packages/03/86/bde4ad8b4d0e9429a4e82c1e8f5c659993a9a863ad62c7df05cf7b678d75/charset_normalizer-3.4.4-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:d9c7f57c3d666a53421049053eaacdd14bbd0a528e2186fcb2e672effd053bb0", size = 150019 }, - { url = "https://files.pythonhosted.org/packages/1f/86/a151eb2af293a7e7bac3a739b81072585ce36ccfb4493039f49f1d3cae8c/charset_normalizer-3.4.4-cp311-cp311-musllinux_1_2_armv7l.whl", hash = "sha256:277e970e750505ed74c832b4bf75dac7476262ee2a013f5574dd49075879e161", size = 143310 }, - { url = "https://files.pythonhosted.org/packages/b5/fe/43dae6144a7e07b87478fdfc4dbe9efd5defb0e7ec29f5f58a55aeef7bf7/charset_normalizer-3.4.4-cp311-cp311-musllinux_1_2_ppc64le.whl", hash = "sha256:31fd66405eaf47bb62e8cd575dc621c56c668f27d46a61d975a249930dd5e2a4", size = 162022 }, - { url = "https://files.pythonhosted.org/packages/80/e6/7aab83774f5d2bca81f42ac58d04caf44f0cc2b65fc6db2b3b2e8a05f3b3/charset_normalizer-3.4.4-cp311-cp311-musllinux_1_2_riscv64.whl", hash = "sha256:0d3d8f15c07f86e9ff82319b3d9ef6f4bf907608f53fe9d92b28ea9ae3d1fd89", size = 149383 }, - { url = "https://files.pythonhosted.org/packages/4f/e8/b289173b4edae05c0dde07f69f8db476a0b511eac556dfe0d6bda3c43384/charset_normalizer-3.4.4-cp311-cp311-musllinux_1_2_s390x.whl", hash = "sha256:9f7fcd74d410a36883701fafa2482a6af2ff5ba96b9a620e9e0721e28ead5569", size = 159098 }, - { url = "https://files.pythonhosted.org/packages/d8/df/fe699727754cae3f8478493c7f45f777b17c3ef0600e28abfec8619eb49c/charset_normalizer-3.4.4-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:ebf3e58c7ec8a8bed6d66a75d7fb37b55e5015b03ceae72a8e7c74495551e224", size = 152991 }, - { url = "https://files.pythonhosted.org/packages/1a/86/584869fe4ddb6ffa3bd9f491b87a01568797fb9bd8933f557dba9771beaf/charset_normalizer-3.4.4-cp311-cp311-win32.whl", hash = "sha256:eecbc200c7fd5ddb9a7f16c7decb07b566c29fa2161a16cf67b8d068bd21690a", size = 99456 }, - { url = "https://files.pythonhosted.org/packages/65/f6/62fdd5feb60530f50f7e38b4f6a1d5203f4d16ff4f9f0952962c044e919a/charset_normalizer-3.4.4-cp311-cp311-win_amd64.whl", hash = "sha256:5ae497466c7901d54b639cf42d5b8c1b6a4fead55215500d2f486d34db48d016", size = 106978 }, - { url = "https://files.pythonhosted.org/packages/7a/9d/0710916e6c82948b3be62d9d398cb4fcf4e97b56d6a6aeccd66c4b2f2bd5/charset_normalizer-3.4.4-cp311-cp311-win_arm64.whl", hash = "sha256:65e2befcd84bc6f37095f5961e68a6f077bf44946771354a28ad434c2cce0ae1", size = 99969 }, - { url = "https://files.pythonhosted.org/packages/f3/85/1637cd4af66fa687396e757dec650f28025f2a2f5a5531a3208dc0ec43f2/charset_normalizer-3.4.4-cp312-cp312-macosx_10_13_universal2.whl", hash = "sha256:0a98e6759f854bd25a58a73fa88833fba3b7c491169f86ce1180c948ab3fd394", size = 208425 }, - { url = "https://files.pythonhosted.org/packages/9d/6a/04130023fef2a0d9c62d0bae2649b69f7b7d8d24ea5536feef50551029df/charset_normalizer-3.4.4-cp312-cp312-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:b5b290ccc2a263e8d185130284f8501e3e36c5e02750fc6b6bdeb2e9e96f1e25", size = 148162 }, - { url = "https://files.pythonhosted.org/packages/78/29/62328d79aa60da22c9e0b9a66539feae06ca0f5a4171ac4f7dc285b83688/charset_normalizer-3.4.4-cp312-cp312-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:74bb723680f9f7a6234dcf67aea57e708ec1fbdf5699fb91dfd6f511b0a320ef", size = 144558 }, - { url = "https://files.pythonhosted.org/packages/86/bb/b32194a4bf15b88403537c2e120b817c61cd4ecffa9b6876e941c3ee38fe/charset_normalizer-3.4.4-cp312-cp312-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:f1e34719c6ed0b92f418c7c780480b26b5d9c50349e9a9af7d76bf757530350d", size = 161497 }, - { url = "https://files.pythonhosted.org/packages/19/89/a54c82b253d5b9b111dc74aca196ba5ccfcca8242d0fb64146d4d3183ff1/charset_normalizer-3.4.4-cp312-cp312-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:2437418e20515acec67d86e12bf70056a33abdacb5cb1655042f6538d6b085a8", size = 159240 }, - { url = "https://files.pythonhosted.org/packages/c0/10/d20b513afe03acc89ec33948320a5544d31f21b05368436d580dec4e234d/charset_normalizer-3.4.4-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:11d694519d7f29d6cd09f6ac70028dba10f92f6cdd059096db198c283794ac86", size = 153471 }, - { url = "https://files.pythonhosted.org/packages/61/fa/fbf177b55bdd727010f9c0a3c49eefa1d10f960e5f09d1d887bf93c2e698/charset_normalizer-3.4.4-cp312-cp312-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl", hash = "sha256:ac1c4a689edcc530fc9d9aa11f5774b9e2f33f9a0c6a57864e90908f5208d30a", size = 150864 }, - { url = "https://files.pythonhosted.org/packages/05/12/9fbc6a4d39c0198adeebbde20b619790e9236557ca59fc40e0e3cebe6f40/charset_normalizer-3.4.4-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:21d142cc6c0ec30d2efee5068ca36c128a30b0f2c53c1c07bd78cb6bc1d3be5f", size = 150647 }, - { url = "https://files.pythonhosted.org/packages/ad/1f/6a9a593d52e3e8c5d2b167daf8c6b968808efb57ef4c210acb907c365bc4/charset_normalizer-3.4.4-cp312-cp312-musllinux_1_2_armv7l.whl", hash = "sha256:5dbe56a36425d26d6cfb40ce79c314a2e4dd6211d51d6d2191c00bed34f354cc", size = 145110 }, - { url = "https://files.pythonhosted.org/packages/30/42/9a52c609e72471b0fc54386dc63c3781a387bb4fe61c20231a4ebcd58bdd/charset_normalizer-3.4.4-cp312-cp312-musllinux_1_2_ppc64le.whl", hash = "sha256:5bfbb1b9acf3334612667b61bd3002196fe2a1eb4dd74d247e0f2a4d50ec9bbf", size = 162839 }, - { url = "https://files.pythonhosted.org/packages/c4/5b/c0682bbf9f11597073052628ddd38344a3d673fda35a36773f7d19344b23/charset_normalizer-3.4.4-cp312-cp312-musllinux_1_2_riscv64.whl", hash = "sha256:d055ec1e26e441f6187acf818b73564e6e6282709e9bcb5b63f5b23068356a15", size = 150667 }, - { url = "https://files.pythonhosted.org/packages/e4/24/a41afeab6f990cf2daf6cb8c67419b63b48cf518e4f56022230840c9bfb2/charset_normalizer-3.4.4-cp312-cp312-musllinux_1_2_s390x.whl", hash = "sha256:af2d8c67d8e573d6de5bc30cdb27e9b95e49115cd9baad5ddbd1a6207aaa82a9", size = 160535 }, - { url = "https://files.pythonhosted.org/packages/2a/e5/6a4ce77ed243c4a50a1fecca6aaaab419628c818a49434be428fe24c9957/charset_normalizer-3.4.4-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:780236ac706e66881f3b7f2f32dfe90507a09e67d1d454c762cf642e6e1586e0", size = 154816 }, - { url = "https://files.pythonhosted.org/packages/a8/ef/89297262b8092b312d29cdb2517cb1237e51db8ecef2e9af5edbe7b683b1/charset_normalizer-3.4.4-cp312-cp312-win32.whl", hash = "sha256:5833d2c39d8896e4e19b689ffc198f08ea58116bee26dea51e362ecc7cd3ed26", size = 99694 }, - { url = "https://files.pythonhosted.org/packages/3d/2d/1e5ed9dd3b3803994c155cd9aacb60c82c331bad84daf75bcb9c91b3295e/charset_normalizer-3.4.4-cp312-cp312-win_amd64.whl", hash = "sha256:a79cfe37875f822425b89a82333404539ae63dbdddf97f84dcbc3d339aae9525", size = 107131 }, - { url = "https://files.pythonhosted.org/packages/d0/d9/0ed4c7098a861482a7b6a95603edce4c0d9db2311af23da1fb2b75ec26fc/charset_normalizer-3.4.4-cp312-cp312-win_arm64.whl", hash = "sha256:376bec83a63b8021bb5c8ea75e21c4ccb86e7e45ca4eb81146091b56599b80c3", size = 100390 }, - { url = "https://files.pythonhosted.org/packages/97/45/4b3a1239bbacd321068ea6e7ac28875b03ab8bc0aa0966452db17cd36714/charset_normalizer-3.4.4-cp313-cp313-macosx_10_13_universal2.whl", hash = "sha256:e1f185f86a6f3403aa2420e815904c67b2f9ebc443f045edd0de921108345794", size = 208091 }, - { url = "https://files.pythonhosted.org/packages/7d/62/73a6d7450829655a35bb88a88fca7d736f9882a27eacdca2c6d505b57e2e/charset_normalizer-3.4.4-cp313-cp313-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:6b39f987ae8ccdf0d2642338faf2abb1862340facc796048b604ef14919e55ed", size = 147936 }, - { url = "https://files.pythonhosted.org/packages/89/c5/adb8c8b3d6625bef6d88b251bbb0d95f8205831b987631ab0c8bb5d937c2/charset_normalizer-3.4.4-cp313-cp313-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:3162d5d8ce1bb98dd51af660f2121c55d0fa541b46dff7bb9b9f86ea1d87de72", size = 144180 }, - { url = "https://files.pythonhosted.org/packages/91/ed/9706e4070682d1cc219050b6048bfd293ccf67b3d4f5a4f39207453d4b99/charset_normalizer-3.4.4-cp313-cp313-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:81d5eb2a312700f4ecaa977a8235b634ce853200e828fbadf3a9c50bab278328", size = 161346 }, - { url = "https://files.pythonhosted.org/packages/d5/0d/031f0d95e4972901a2f6f09ef055751805ff541511dc1252ba3ca1f80cf5/charset_normalizer-3.4.4-cp313-cp313-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:5bd2293095d766545ec1a8f612559f6b40abc0eb18bb2f5d1171872d34036ede", size = 158874 }, - { url = "https://files.pythonhosted.org/packages/f5/83/6ab5883f57c9c801ce5e5677242328aa45592be8a00644310a008d04f922/charset_normalizer-3.4.4-cp313-cp313-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:a8a8b89589086a25749f471e6a900d3f662d1d3b6e2e59dcecf787b1cc3a1894", size = 153076 }, - { url = "https://files.pythonhosted.org/packages/75/1e/5ff781ddf5260e387d6419959ee89ef13878229732732ee73cdae01800f2/charset_normalizer-3.4.4-cp313-cp313-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl", hash = "sha256:bc7637e2f80d8530ee4a78e878bce464f70087ce73cf7c1caf142416923b98f1", size = 150601 }, - { url = "https://files.pythonhosted.org/packages/d7/57/71be810965493d3510a6ca79b90c19e48696fb1ff964da319334b12677f0/charset_normalizer-3.4.4-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:f8bf04158c6b607d747e93949aa60618b61312fe647a6369f88ce2ff16043490", size = 150376 }, - { url = "https://files.pythonhosted.org/packages/e5/d5/c3d057a78c181d007014feb7e9f2e65905a6c4ef182c0ddf0de2924edd65/charset_normalizer-3.4.4-cp313-cp313-musllinux_1_2_armv7l.whl", hash = "sha256:554af85e960429cf30784dd47447d5125aaa3b99a6f0683589dbd27e2f45da44", size = 144825 }, - { url = "https://files.pythonhosted.org/packages/e6/8c/d0406294828d4976f275ffbe66f00266c4b3136b7506941d87c00cab5272/charset_normalizer-3.4.4-cp313-cp313-musllinux_1_2_ppc64le.whl", hash = "sha256:74018750915ee7ad843a774364e13a3db91682f26142baddf775342c3f5b1133", size = 162583 }, - { url = "https://files.pythonhosted.org/packages/d7/24/e2aa1f18c8f15c4c0e932d9287b8609dd30ad56dbe41d926bd846e22fb8d/charset_normalizer-3.4.4-cp313-cp313-musllinux_1_2_riscv64.whl", hash = "sha256:c0463276121fdee9c49b98908b3a89c39be45d86d1dbaa22957e38f6321d4ce3", size = 150366 }, - { url = "https://files.pythonhosted.org/packages/e4/5b/1e6160c7739aad1e2df054300cc618b06bf784a7a164b0f238360721ab86/charset_normalizer-3.4.4-cp313-cp313-musllinux_1_2_s390x.whl", hash = "sha256:362d61fd13843997c1c446760ef36f240cf81d3ebf74ac62652aebaf7838561e", size = 160300 }, - { url = "https://files.pythonhosted.org/packages/7a/10/f882167cd207fbdd743e55534d5d9620e095089d176d55cb22d5322f2afd/charset_normalizer-3.4.4-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:9a26f18905b8dd5d685d6d07b0cdf98a79f3c7a918906af7cc143ea2e164c8bc", size = 154465 }, - { url = "https://files.pythonhosted.org/packages/89/66/c7a9e1b7429be72123441bfdbaf2bc13faab3f90b933f664db506dea5915/charset_normalizer-3.4.4-cp313-cp313-win32.whl", hash = "sha256:9b35f4c90079ff2e2edc5b26c0c77925e5d2d255c42c74fdb70fb49b172726ac", size = 99404 }, - { url = "https://files.pythonhosted.org/packages/c4/26/b9924fa27db384bdcd97ab83b4f0a8058d96ad9626ead570674d5e737d90/charset_normalizer-3.4.4-cp313-cp313-win_amd64.whl", hash = "sha256:b435cba5f4f750aa6c0a0d92c541fb79f69a387c91e61f1795227e4ed9cece14", size = 107092 }, - { url = "https://files.pythonhosted.org/packages/af/8f/3ed4bfa0c0c72a7ca17f0380cd9e4dd842b09f664e780c13cff1dcf2ef1b/charset_normalizer-3.4.4-cp313-cp313-win_arm64.whl", hash = "sha256:542d2cee80be6f80247095cc36c418f7bddd14f4a6de45af91dfad36d817bba2", size = 100408 }, - { url = "https://files.pythonhosted.org/packages/2a/35/7051599bd493e62411d6ede36fd5af83a38f37c4767b92884df7301db25d/charset_normalizer-3.4.4-cp314-cp314-macosx_10_13_universal2.whl", hash = "sha256:da3326d9e65ef63a817ecbcc0df6e94463713b754fe293eaa03da99befb9a5bd", size = 207746 }, - { url = "https://files.pythonhosted.org/packages/10/9a/97c8d48ef10d6cd4fcead2415523221624bf58bcf68a802721a6bc807c8f/charset_normalizer-3.4.4-cp314-cp314-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:8af65f14dc14a79b924524b1e7fffe304517b2bff5a58bf64f30b98bbc5079eb", size = 147889 }, - { url = "https://files.pythonhosted.org/packages/10/bf/979224a919a1b606c82bd2c5fa49b5c6d5727aa47b4312bb27b1734f53cd/charset_normalizer-3.4.4-cp314-cp314-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:74664978bb272435107de04e36db5a9735e78232b85b77d45cfb38f758efd33e", size = 143641 }, - { url = "https://files.pythonhosted.org/packages/ba/33/0ad65587441fc730dc7bd90e9716b30b4702dc7b617e6ba4997dc8651495/charset_normalizer-3.4.4-cp314-cp314-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:752944c7ffbfdd10c074dc58ec2d5a8a4cd9493b314d367c14d24c17684ddd14", size = 160779 }, - { url = "https://files.pythonhosted.org/packages/67/ed/331d6b249259ee71ddea93f6f2f0a56cfebd46938bde6fcc6f7b9a3d0e09/charset_normalizer-3.4.4-cp314-cp314-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:d1f13550535ad8cff21b8d757a3257963e951d96e20ec82ab44bc64aeb62a191", size = 159035 }, - { url = "https://files.pythonhosted.org/packages/67/ff/f6b948ca32e4f2a4576aa129d8bed61f2e0543bf9f5f2b7fc3758ed005c9/charset_normalizer-3.4.4-cp314-cp314-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:ecaae4149d99b1c9e7b88bb03e3221956f68fd6d50be2ef061b2381b61d20838", size = 152542 }, - { url = "https://files.pythonhosted.org/packages/16/85/276033dcbcc369eb176594de22728541a925b2632f9716428c851b149e83/charset_normalizer-3.4.4-cp314-cp314-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl", hash = "sha256:cb6254dc36b47a990e59e1068afacdcd02958bdcce30bb50cc1700a8b9d624a6", size = 149524 }, - { url = "https://files.pythonhosted.org/packages/9e/f2/6a2a1f722b6aba37050e626530a46a68f74e63683947a8acff92569f979a/charset_normalizer-3.4.4-cp314-cp314-musllinux_1_2_aarch64.whl", hash = "sha256:c8ae8a0f02f57a6e61203a31428fa1d677cbe50c93622b4149d5c0f319c1d19e", size = 150395 }, - { url = "https://files.pythonhosted.org/packages/60/bb/2186cb2f2bbaea6338cad15ce23a67f9b0672929744381e28b0592676824/charset_normalizer-3.4.4-cp314-cp314-musllinux_1_2_armv7l.whl", hash = "sha256:47cc91b2f4dd2833fddaedd2893006b0106129d4b94fdb6af1f4ce5a9965577c", size = 143680 }, - { url = "https://files.pythonhosted.org/packages/7d/a5/bf6f13b772fbb2a90360eb620d52ed8f796f3c5caee8398c3b2eb7b1c60d/charset_normalizer-3.4.4-cp314-cp314-musllinux_1_2_ppc64le.whl", hash = "sha256:82004af6c302b5d3ab2cfc4cc5f29db16123b1a8417f2e25f9066f91d4411090", size = 162045 }, - { url = "https://files.pythonhosted.org/packages/df/c5/d1be898bf0dc3ef9030c3825e5d3b83f2c528d207d246cbabe245966808d/charset_normalizer-3.4.4-cp314-cp314-musllinux_1_2_riscv64.whl", hash = "sha256:2b7d8f6c26245217bd2ad053761201e9f9680f8ce52f0fcd8d0755aeae5b2152", size = 149687 }, - { url = "https://files.pythonhosted.org/packages/a5/42/90c1f7b9341eef50c8a1cb3f098ac43b0508413f33affd762855f67a410e/charset_normalizer-3.4.4-cp314-cp314-musllinux_1_2_s390x.whl", hash = "sha256:799a7a5e4fb2d5898c60b640fd4981d6a25f1c11790935a44ce38c54e985f828", size = 160014 }, - { url = "https://files.pythonhosted.org/packages/76/be/4d3ee471e8145d12795ab655ece37baed0929462a86e72372fd25859047c/charset_normalizer-3.4.4-cp314-cp314-musllinux_1_2_x86_64.whl", hash = "sha256:99ae2cffebb06e6c22bdc25801d7b30f503cc87dbd283479e7b606f70aff57ec", size = 154044 }, - { url = "https://files.pythonhosted.org/packages/b0/6f/8f7af07237c34a1defe7defc565a9bc1807762f672c0fde711a4b22bf9c0/charset_normalizer-3.4.4-cp314-cp314-win32.whl", hash = "sha256:f9d332f8c2a2fcbffe1378594431458ddbef721c1769d78e2cbc06280d8155f9", size = 99940 }, - { url = "https://files.pythonhosted.org/packages/4b/51/8ade005e5ca5b0d80fb4aff72a3775b325bdc3d27408c8113811a7cbe640/charset_normalizer-3.4.4-cp314-cp314-win_amd64.whl", hash = "sha256:8a6562c3700cce886c5be75ade4a5db4214fda19fede41d9792d100288d8f94c", size = 107104 }, - { url = "https://files.pythonhosted.org/packages/da/5f/6b8f83a55bb8278772c5ae54a577f3099025f9ade59d0136ac24a0df4bde/charset_normalizer-3.4.4-cp314-cp314-win_arm64.whl", hash = "sha256:de00632ca48df9daf77a2c65a484531649261ec9f25489917f09e455cb09ddb2", size = 100743 }, - { url = "https://files.pythonhosted.org/packages/0a/4c/925909008ed5a988ccbb72dcc897407e5d6d3bd72410d69e051fc0c14647/charset_normalizer-3.4.4-py3-none-any.whl", hash = "sha256:7a32c560861a02ff789ad905a2fe94e3f840803362c84fecf1851cb4cf3dc37f", size = 53402 }, -] - -[[package]] -name = "click" -version = "8.3.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "colorama", marker = "sys_platform == 'win32'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/3d/fa/656b739db8587d7b5dfa22e22ed02566950fbfbcdc20311993483657a5c0/click-8.3.1.tar.gz", hash = "sha256:12ff4785d337a1bb490bb7e9c2b1ee5da3112e94a8622f26a6c77f5d2fc6842a", size = 295065 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/98/78/01c019cdb5d6498122777c1a43056ebb3ebfeef2076d9d026bfe15583b2b/click-8.3.1-py3-none-any.whl", hash = "sha256:981153a64e25f12d547d3426c367a4857371575ee7ad18df2a6183ab0545b2a6", size = 108274 }, -] - -[[package]] -name = "cloudpickle" -version = "3.1.2" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/27/fb/576f067976d320f5f0114a8d9fa1215425441bb35627b1993e5afd8111e5/cloudpickle-3.1.2.tar.gz", hash = "sha256:7fda9eb655c9c230dab534f1983763de5835249750e85fbcef43aaa30a9a2414", size = 22330 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/88/39/799be3f2f0f38cc727ee3b4f1445fe6d5e4133064ec2e4115069418a5bb6/cloudpickle-3.1.2-py3-none-any.whl", hash = "sha256:9acb47f6afd73f60dc1df93bb801b472f05ff42fa6c84167d25cb206be1fbf4a", size = 22228 }, -] - -[[package]] -name = "colorama" -version = "0.4.6" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/d8/53/6f443c9a4a8358a93a6792e2acffb9d9d5cb0a5cfd8802644b7b1c9a02e4/colorama-0.4.6.tar.gz", hash = "sha256:08695f5cb7ed6e0531a20572697297273c47b8cae5a63ffc6d6ed5c201be6e44", size = 27697 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/d1/d6/3965ed04c63042e047cb6a3e6ed1a63a35087b6a609aa3a15ed8ac56c221/colorama-0.4.6-py2.py3-none-any.whl", hash = "sha256:4f1d9991f5acc0ca119f9d443620b77f9d6b33703e51011c16baf57afb285fc6", size = 25335 }, -] - -[[package]] -name = "colorlog" -version = "6.10.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "colorama", marker = "sys_platform == 'win32'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/a2/61/f083b5ac52e505dfc1c624eafbf8c7589a0d7f32daa398d2e7590efa5fda/colorlog-6.10.1.tar.gz", hash = "sha256:eb4ae5cb65fe7fec7773c2306061a8e63e02efc2c72eba9d27b0fa23c94f1321", size = 17162 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/6d/c1/e419ef3723a074172b68aaa89c9f3de486ed4c2399e2dbd8113a4fdcaf9e/colorlog-6.10.1-py3-none-any.whl", hash = "sha256:2d7e8348291948af66122cff006c9f8da6255d224e7cf8e37d8de2df3bad8c9c", size = 11743 }, -] - -[[package]] -name = "contourpy" -version = "1.3.2" -source = { registry = "https://pypi.org/simple" } -resolution-markers = [ - "python_full_version < '3.11' and sys_platform == 'linux'", - "python_full_version < '3.11' and sys_platform != 'linux'", -] -dependencies = [ - { name = "numpy", marker = "python_full_version < '3.11'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/66/54/eb9bfc647b19f2009dd5c7f5ec51c4e6ca831725f1aea7a993034f483147/contourpy-1.3.2.tar.gz", hash = "sha256:b6945942715a034c671b7fc54f9588126b0b8bf23db2696e3ca8328f3ff0ab54", size = 13466130 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/12/a3/da4153ec8fe25d263aa48c1a4cbde7f49b59af86f0b6f7862788c60da737/contourpy-1.3.2-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:ba38e3f9f330af820c4b27ceb4b9c7feee5fe0493ea53a8720f4792667465934", size = 268551 }, - { url = "https://files.pythonhosted.org/packages/2f/6c/330de89ae1087eb622bfca0177d32a7ece50c3ef07b28002de4757d9d875/contourpy-1.3.2-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:dc41ba0714aa2968d1f8674ec97504a8f7e334f48eeacebcaa6256213acb0989", size = 253399 }, - { url = "https://files.pythonhosted.org/packages/c1/bd/20c6726b1b7f81a8bee5271bed5c165f0a8e1f572578a9d27e2ccb763cb2/contourpy-1.3.2-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:9be002b31c558d1ddf1b9b415b162c603405414bacd6932d031c5b5a8b757f0d", size = 312061 }, - { url = "https://files.pythonhosted.org/packages/22/fc/a9665c88f8a2473f823cf1ec601de9e5375050f1958cbb356cdf06ef1ab6/contourpy-1.3.2-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:8d2e74acbcba3bfdb6d9d8384cdc4f9260cae86ed9beee8bd5f54fee49a430b9", size = 351956 }, - { url = "https://files.pythonhosted.org/packages/25/eb/9f0a0238f305ad8fb7ef42481020d6e20cf15e46be99a1fcf939546a177e/contourpy-1.3.2-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:e259bced5549ac64410162adc973c5e2fb77f04df4a439d00b478e57a0e65512", size = 320872 }, - { url = "https://files.pythonhosted.org/packages/32/5c/1ee32d1c7956923202f00cf8d2a14a62ed7517bdc0ee1e55301227fc273c/contourpy-1.3.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ad687a04bc802cbe8b9c399c07162a3c35e227e2daccf1668eb1f278cb698631", size = 325027 }, - { url = "https://files.pythonhosted.org/packages/83/bf/9baed89785ba743ef329c2b07fd0611d12bfecbedbdd3eeecf929d8d3b52/contourpy-1.3.2-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:cdd22595308f53ef2f891040ab2b93d79192513ffccbd7fe19be7aa773a5e09f", size = 1306641 }, - { url = "https://files.pythonhosted.org/packages/d4/cc/74e5e83d1e35de2d28bd97033426b450bc4fd96e092a1f7a63dc7369b55d/contourpy-1.3.2-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:b4f54d6a2defe9f257327b0f243612dd051cc43825587520b1bf74a31e2f6ef2", size = 1374075 }, - { url = "https://files.pythonhosted.org/packages/0c/42/17f3b798fd5e033b46a16f8d9fcb39f1aba051307f5ebf441bad1ecf78f8/contourpy-1.3.2-cp310-cp310-win32.whl", hash = "sha256:f939a054192ddc596e031e50bb13b657ce318cf13d264f095ce9db7dc6ae81c0", size = 177534 }, - { url = "https://files.pythonhosted.org/packages/54/ec/5162b8582f2c994721018d0c9ece9dc6ff769d298a8ac6b6a652c307e7df/contourpy-1.3.2-cp310-cp310-win_amd64.whl", hash = "sha256:c440093bbc8fc21c637c03bafcbef95ccd963bc6e0514ad887932c18ca2a759a", size = 221188 }, - { url = "https://files.pythonhosted.org/packages/b3/b9/ede788a0b56fc5b071639d06c33cb893f68b1178938f3425debebe2dab78/contourpy-1.3.2-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:6a37a2fb93d4df3fc4c0e363ea4d16f83195fc09c891bc8ce072b9d084853445", size = 269636 }, - { url = "https://files.pythonhosted.org/packages/e6/75/3469f011d64b8bbfa04f709bfc23e1dd71be54d05b1b083be9f5b22750d1/contourpy-1.3.2-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:b7cd50c38f500bbcc9b6a46643a40e0913673f869315d8e70de0438817cb7773", size = 254636 }, - { url = "https://files.pythonhosted.org/packages/8d/2f/95adb8dae08ce0ebca4fd8e7ad653159565d9739128b2d5977806656fcd2/contourpy-1.3.2-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d6658ccc7251a4433eebd89ed2672c2ed96fba367fd25ca9512aa92a4b46c4f1", size = 313053 }, - { url = "https://files.pythonhosted.org/packages/c3/a6/8ccf97a50f31adfa36917707fe39c9a0cbc24b3bbb58185577f119736cc9/contourpy-1.3.2-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:70771a461aaeb335df14deb6c97439973d253ae70660ca085eec25241137ef43", size = 352985 }, - { url = "https://files.pythonhosted.org/packages/1d/b6/7925ab9b77386143f39d9c3243fdd101621b4532eb126743201160ffa7e6/contourpy-1.3.2-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:65a887a6e8c4cd0897507d814b14c54a8c2e2aa4ac9f7686292f9769fcf9a6ab", size = 323750 }, - { url = "https://files.pythonhosted.org/packages/c2/f3/20c5d1ef4f4748e52d60771b8560cf00b69d5c6368b5c2e9311bcfa2a08b/contourpy-1.3.2-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3859783aefa2b8355697f16642695a5b9792e7a46ab86da1118a4a23a51a33d7", size = 326246 }, - { url = "https://files.pythonhosted.org/packages/8c/e5/9dae809e7e0b2d9d70c52b3d24cba134dd3dad979eb3e5e71f5df22ed1f5/contourpy-1.3.2-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:eab0f6db315fa4d70f1d8ab514e527f0366ec021ff853d7ed6a2d33605cf4b83", size = 1308728 }, - { url = "https://files.pythonhosted.org/packages/e2/4a/0058ba34aeea35c0b442ae61a4f4d4ca84d6df8f91309bc2d43bb8dd248f/contourpy-1.3.2-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:d91a3ccc7fea94ca0acab82ceb77f396d50a1f67412efe4c526f5d20264e6ecd", size = 1375762 }, - { url = "https://files.pythonhosted.org/packages/09/33/7174bdfc8b7767ef2c08ed81244762d93d5c579336fc0b51ca57b33d1b80/contourpy-1.3.2-cp311-cp311-win32.whl", hash = "sha256:1c48188778d4d2f3d48e4643fb15d8608b1d01e4b4d6b0548d9b336c28fc9b6f", size = 178196 }, - { url = "https://files.pythonhosted.org/packages/5e/fe/4029038b4e1c4485cef18e480b0e2cd2d755448bb071eb9977caac80b77b/contourpy-1.3.2-cp311-cp311-win_amd64.whl", hash = "sha256:5ebac872ba09cb8f2131c46b8739a7ff71de28a24c869bcad554477eb089a878", size = 222017 }, - { url = "https://files.pythonhosted.org/packages/34/f7/44785876384eff370c251d58fd65f6ad7f39adce4a093c934d4a67a7c6b6/contourpy-1.3.2-cp312-cp312-macosx_10_13_x86_64.whl", hash = "sha256:4caf2bcd2969402bf77edc4cb6034c7dd7c0803213b3523f111eb7460a51b8d2", size = 271580 }, - { url = "https://files.pythonhosted.org/packages/93/3b/0004767622a9826ea3d95f0e9d98cd8729015768075d61f9fea8eeca42a8/contourpy-1.3.2-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:82199cb78276249796419fe36b7386bd8d2cc3f28b3bc19fe2454fe2e26c4c15", size = 255530 }, - { url = "https://files.pythonhosted.org/packages/e7/bb/7bd49e1f4fa805772d9fd130e0d375554ebc771ed7172f48dfcd4ca61549/contourpy-1.3.2-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:106fab697af11456fcba3e352ad50effe493a90f893fca6c2ca5c033820cea92", size = 307688 }, - { url = "https://files.pythonhosted.org/packages/fc/97/e1d5dbbfa170725ef78357a9a0edc996b09ae4af170927ba8ce977e60a5f/contourpy-1.3.2-cp312-cp312-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:d14f12932a8d620e307f715857107b1d1845cc44fdb5da2bc8e850f5ceba9f87", size = 347331 }, - { url = "https://files.pythonhosted.org/packages/6f/66/e69e6e904f5ecf6901be3dd16e7e54d41b6ec6ae3405a535286d4418ffb4/contourpy-1.3.2-cp312-cp312-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:532fd26e715560721bb0d5fc7610fce279b3699b018600ab999d1be895b09415", size = 318963 }, - { url = "https://files.pythonhosted.org/packages/a8/32/b8a1c8965e4f72482ff2d1ac2cd670ce0b542f203c8e1d34e7c3e6925da7/contourpy-1.3.2-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:f26b383144cf2d2c29f01a1e8170f50dacf0eac02d64139dcd709a8ac4eb3cfe", size = 323681 }, - { url = "https://files.pythonhosted.org/packages/30/c6/12a7e6811d08757c7162a541ca4c5c6a34c0f4e98ef2b338791093518e40/contourpy-1.3.2-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:c49f73e61f1f774650a55d221803b101d966ca0c5a2d6d5e4320ec3997489441", size = 1308674 }, - { url = "https://files.pythonhosted.org/packages/2a/8a/bebe5a3f68b484d3a2b8ffaf84704b3e343ef1addea528132ef148e22b3b/contourpy-1.3.2-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:3d80b2c0300583228ac98d0a927a1ba6a2ba6b8a742463c564f1d419ee5b211e", size = 1380480 }, - { url = "https://files.pythonhosted.org/packages/34/db/fcd325f19b5978fb509a7d55e06d99f5f856294c1991097534360b307cf1/contourpy-1.3.2-cp312-cp312-win32.whl", hash = "sha256:90df94c89a91b7362e1142cbee7568f86514412ab8a2c0d0fca72d7e91b62912", size = 178489 }, - { url = "https://files.pythonhosted.org/packages/01/c8/fadd0b92ffa7b5eb5949bf340a63a4a496a6930a6c37a7ba0f12acb076d6/contourpy-1.3.2-cp312-cp312-win_amd64.whl", hash = "sha256:8c942a01d9163e2e5cfb05cb66110121b8d07ad438a17f9e766317bcb62abf73", size = 223042 }, - { url = "https://files.pythonhosted.org/packages/2e/61/5673f7e364b31e4e7ef6f61a4b5121c5f170f941895912f773d95270f3a2/contourpy-1.3.2-cp313-cp313-macosx_10_13_x86_64.whl", hash = "sha256:de39db2604ae755316cb5967728f4bea92685884b1e767b7c24e983ef5f771cb", size = 271630 }, - { url = "https://files.pythonhosted.org/packages/ff/66/a40badddd1223822c95798c55292844b7e871e50f6bfd9f158cb25e0bd39/contourpy-1.3.2-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:3f9e896f447c5c8618f1edb2bafa9a4030f22a575ec418ad70611450720b5b08", size = 255670 }, - { url = "https://files.pythonhosted.org/packages/1e/c7/cf9fdee8200805c9bc3b148f49cb9482a4e3ea2719e772602a425c9b09f8/contourpy-1.3.2-cp313-cp313-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:71e2bd4a1c4188f5c2b8d274da78faab884b59df20df63c34f74aa1813c4427c", size = 306694 }, - { url = "https://files.pythonhosted.org/packages/dd/e7/ccb9bec80e1ba121efbffad7f38021021cda5be87532ec16fd96533bb2e0/contourpy-1.3.2-cp313-cp313-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:de425af81b6cea33101ae95ece1f696af39446db9682a0b56daaa48cfc29f38f", size = 345986 }, - { url = "https://files.pythonhosted.org/packages/dc/49/ca13bb2da90391fa4219fdb23b078d6065ada886658ac7818e5441448b78/contourpy-1.3.2-cp313-cp313-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:977e98a0e0480d3fe292246417239d2d45435904afd6d7332d8455981c408b85", size = 318060 }, - { url = "https://files.pythonhosted.org/packages/c8/65/5245ce8c548a8422236c13ffcdcdada6a2a812c361e9e0c70548bb40b661/contourpy-1.3.2-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:434f0adf84911c924519d2b08fc10491dd282b20bdd3fa8f60fd816ea0b48841", size = 322747 }, - { url = "https://files.pythonhosted.org/packages/72/30/669b8eb48e0a01c660ead3752a25b44fdb2e5ebc13a55782f639170772f9/contourpy-1.3.2-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:c66c4906cdbc50e9cba65978823e6e00b45682eb09adbb78c9775b74eb222422", size = 1308895 }, - { url = "https://files.pythonhosted.org/packages/05/5a/b569f4250decee6e8d54498be7bdf29021a4c256e77fe8138c8319ef8eb3/contourpy-1.3.2-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:8b7fc0cd78ba2f4695fd0a6ad81a19e7e3ab825c31b577f384aa9d7817dc3bef", size = 1379098 }, - { url = "https://files.pythonhosted.org/packages/19/ba/b227c3886d120e60e41b28740ac3617b2f2b971b9f601c835661194579f1/contourpy-1.3.2-cp313-cp313-win32.whl", hash = "sha256:15ce6ab60957ca74cff444fe66d9045c1fd3e92c8936894ebd1f3eef2fff075f", size = 178535 }, - { url = "https://files.pythonhosted.org/packages/12/6e/2fed56cd47ca739b43e892707ae9a13790a486a3173be063681ca67d2262/contourpy-1.3.2-cp313-cp313-win_amd64.whl", hash = "sha256:e1578f7eafce927b168752ed7e22646dad6cd9bca673c60bff55889fa236ebf9", size = 223096 }, - { url = "https://files.pythonhosted.org/packages/54/4c/e76fe2a03014a7c767d79ea35c86a747e9325537a8b7627e0e5b3ba266b4/contourpy-1.3.2-cp313-cp313t-macosx_10_13_x86_64.whl", hash = "sha256:0475b1f6604896bc7c53bb070e355e9321e1bc0d381735421a2d2068ec56531f", size = 285090 }, - { url = "https://files.pythonhosted.org/packages/7b/e2/5aba47debd55d668e00baf9651b721e7733975dc9fc27264a62b0dd26eb8/contourpy-1.3.2-cp313-cp313t-macosx_11_0_arm64.whl", hash = "sha256:c85bb486e9be652314bb5b9e2e3b0d1b2e643d5eec4992c0fbe8ac71775da739", size = 268643 }, - { url = "https://files.pythonhosted.org/packages/a1/37/cd45f1f051fe6230f751cc5cdd2728bb3a203f5619510ef11e732109593c/contourpy-1.3.2-cp313-cp313t-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:745b57db7758f3ffc05a10254edd3182a2a83402a89c00957a8e8a22f5582823", size = 310443 }, - { url = "https://files.pythonhosted.org/packages/8b/a2/36ea6140c306c9ff6dd38e3bcec80b3b018474ef4d17eb68ceecd26675f4/contourpy-1.3.2-cp313-cp313t-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:970e9173dbd7eba9b4e01aab19215a48ee5dd3f43cef736eebde064a171f89a5", size = 349865 }, - { url = "https://files.pythonhosted.org/packages/95/b7/2fc76bc539693180488f7b6cc518da7acbbb9e3b931fd9280504128bf956/contourpy-1.3.2-cp313-cp313t-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:c6c4639a9c22230276b7bffb6a850dfc8258a2521305e1faefe804d006b2e532", size = 321162 }, - { url = "https://files.pythonhosted.org/packages/f4/10/76d4f778458b0aa83f96e59d65ece72a060bacb20cfbee46cf6cd5ceba41/contourpy-1.3.2-cp313-cp313t-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:cc829960f34ba36aad4302e78eabf3ef16a3a100863f0d4eeddf30e8a485a03b", size = 327355 }, - { url = "https://files.pythonhosted.org/packages/43/a3/10cf483ea683f9f8ab096c24bad3cce20e0d1dd9a4baa0e2093c1c962d9d/contourpy-1.3.2-cp313-cp313t-musllinux_1_2_aarch64.whl", hash = "sha256:d32530b534e986374fc19eaa77fcb87e8a99e5431499949b828312bdcd20ac52", size = 1307935 }, - { url = "https://files.pythonhosted.org/packages/78/73/69dd9a024444489e22d86108e7b913f3528f56cfc312b5c5727a44188471/contourpy-1.3.2-cp313-cp313t-musllinux_1_2_x86_64.whl", hash = "sha256:e298e7e70cf4eb179cc1077be1c725b5fd131ebc81181bf0c03525c8abc297fd", size = 1372168 }, - { url = "https://files.pythonhosted.org/packages/0f/1b/96d586ccf1b1a9d2004dd519b25fbf104a11589abfd05484ff12199cca21/contourpy-1.3.2-cp313-cp313t-win32.whl", hash = "sha256:d0e589ae0d55204991450bb5c23f571c64fe43adaa53f93fc902a84c96f52fe1", size = 189550 }, - { url = "https://files.pythonhosted.org/packages/b0/e6/6000d0094e8a5e32ad62591c8609e269febb6e4db83a1c75ff8868b42731/contourpy-1.3.2-cp313-cp313t-win_amd64.whl", hash = "sha256:78e9253c3de756b3f6a5174d024c4835acd59eb3f8e2ca13e775dbffe1558f69", size = 238214 }, - { url = "https://files.pythonhosted.org/packages/33/05/b26e3c6ecc05f349ee0013f0bb850a761016d89cec528a98193a48c34033/contourpy-1.3.2-pp310-pypy310_pp73-macosx_10_15_x86_64.whl", hash = "sha256:fd93cc7f3139b6dd7aab2f26a90dde0aa9fc264dbf70f6740d498a70b860b82c", size = 265681 }, - { url = "https://files.pythonhosted.org/packages/2b/25/ac07d6ad12affa7d1ffed11b77417d0a6308170f44ff20fa1d5aa6333f03/contourpy-1.3.2-pp310-pypy310_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:107ba8a6a7eec58bb475329e6d3b95deba9440667c4d62b9b6063942b61d7f16", size = 315101 }, - { url = "https://files.pythonhosted.org/packages/8f/4d/5bb3192bbe9d3f27e3061a6a8e7733c9120e203cb8515767d30973f71030/contourpy-1.3.2-pp310-pypy310_pp73-win_amd64.whl", hash = "sha256:ded1706ed0c1049224531b81128efbd5084598f18d8a2d9efae833edbd2b40ad", size = 220599 }, - { url = "https://files.pythonhosted.org/packages/ff/c0/91f1215d0d9f9f343e4773ba6c9b89e8c0cc7a64a6263f21139da639d848/contourpy-1.3.2-pp311-pypy311_pp73-macosx_10_15_x86_64.whl", hash = "sha256:5f5964cdad279256c084b69c3f412b7801e15356b16efa9d78aa974041903da0", size = 266807 }, - { url = "https://files.pythonhosted.org/packages/d4/79/6be7e90c955c0487e7712660d6cead01fa17bff98e0ea275737cc2bc8e71/contourpy-1.3.2-pp311-pypy311_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:49b65a95d642d4efa8f64ba12558fcb83407e58a2dfba9d796d77b63ccfcaff5", size = 318729 }, - { url = "https://files.pythonhosted.org/packages/87/68/7f46fb537958e87427d98a4074bcde4b67a70b04900cfc5ce29bc2f556c1/contourpy-1.3.2-pp311-pypy311_pp73-win_amd64.whl", hash = "sha256:8c5acb8dddb0752bf252e01a3035b21443158910ac16a3b0d20e7fed7d534ce5", size = 221791 }, -] - -[[package]] -name = "contourpy" -version = "1.3.3" -source = { registry = "https://pypi.org/simple" } -resolution-markers = [ - "python_full_version >= '3.14' and sys_platform == 'linux'", - "python_full_version == '3.13.*' and sys_platform == 'linux'", - "python_full_version >= '3.14' and sys_platform != 'linux'", - "python_full_version == '3.13.*' and sys_platform != 'linux'", - "python_full_version == '3.12.*' and sys_platform == 'linux'", - "python_full_version == '3.12.*' and sys_platform != 'linux'", - "python_full_version == '3.11.*' and sys_platform == 'linux'", - "python_full_version == '3.11.*' and sys_platform != 'linux'", -] -dependencies = [ - { name = "numpy", marker = "python_full_version >= '3.11'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/58/01/1253e6698a07380cd31a736d248a3f2a50a7c88779a1813da27503cadc2a/contourpy-1.3.3.tar.gz", hash = "sha256:083e12155b210502d0bca491432bb04d56dc3432f95a979b429f2848c3dbe880", size = 13466174 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/91/2e/c4390a31919d8a78b90e8ecf87cd4b4c4f05a5b48d05ec17db8e5404c6f4/contourpy-1.3.3-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:709a48ef9a690e1343202916450bc48b9e51c049b089c7f79a267b46cffcdaa1", size = 288773 }, - { url = "https://files.pythonhosted.org/packages/0d/44/c4b0b6095fef4dc9c420e041799591e3b63e9619e3044f7f4f6c21c0ab24/contourpy-1.3.3-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:23416f38bfd74d5d28ab8429cc4d63fa67d5068bd711a85edb1c3fb0c3e2f381", size = 270149 }, - { url = "https://files.pythonhosted.org/packages/30/2e/dd4ced42fefac8470661d7cb7e264808425e6c5d56d175291e93890cce09/contourpy-1.3.3-cp311-cp311-manylinux_2_26_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:929ddf8c4c7f348e4c0a5a3a714b5c8542ffaa8c22954862a46ca1813b667ee7", size = 329222 }, - { url = "https://files.pythonhosted.org/packages/f2/74/cc6ec2548e3d276c71389ea4802a774b7aa3558223b7bade3f25787fafc2/contourpy-1.3.3-cp311-cp311-manylinux_2_26_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:9e999574eddae35f1312c2b4b717b7885d4edd6cb46700e04f7f02db454e67c1", size = 377234 }, - { url = "https://files.pythonhosted.org/packages/03/b3/64ef723029f917410f75c09da54254c5f9ea90ef89b143ccadb09df14c15/contourpy-1.3.3-cp311-cp311-manylinux_2_26_s390x.manylinux_2_28_s390x.whl", hash = "sha256:0bf67e0e3f482cb69779dd3061b534eb35ac9b17f163d851e2a547d56dba0a3a", size = 380555 }, - { url = "https://files.pythonhosted.org/packages/5f/4b/6157f24ca425b89fe2eb7e7be642375711ab671135be21e6faa100f7448c/contourpy-1.3.3-cp311-cp311-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:51e79c1f7470158e838808d4a996fa9bac72c498e93d8ebe5119bc1e6becb0db", size = 355238 }, - { url = "https://files.pythonhosted.org/packages/98/56/f914f0dd678480708a04cfd2206e7c382533249bc5001eb9f58aa693e200/contourpy-1.3.3-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:598c3aaece21c503615fd59c92a3598b428b2f01bfb4b8ca9c4edeecc2438620", size = 1326218 }, - { url = "https://files.pythonhosted.org/packages/fb/d7/4a972334a0c971acd5172389671113ae82aa7527073980c38d5868ff1161/contourpy-1.3.3-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:322ab1c99b008dad206d406bb61d014cf0174df491ae9d9d0fac6a6fda4f977f", size = 1392867 }, - { url = "https://files.pythonhosted.org/packages/75/3e/f2cc6cd56dc8cff46b1a56232eabc6feea52720083ea71ab15523daab796/contourpy-1.3.3-cp311-cp311-win32.whl", hash = "sha256:fd907ae12cd483cd83e414b12941c632a969171bf90fc937d0c9f268a31cafff", size = 183677 }, - { url = "https://files.pythonhosted.org/packages/98/4b/9bd370b004b5c9d8045c6c33cf65bae018b27aca550a3f657cdc99acdbd8/contourpy-1.3.3-cp311-cp311-win_amd64.whl", hash = "sha256:3519428f6be58431c56581f1694ba8e50626f2dd550af225f82fb5f5814d2a42", size = 225234 }, - { url = "https://files.pythonhosted.org/packages/d9/b6/71771e02c2e004450c12b1120a5f488cad2e4d5b590b1af8bad060360fe4/contourpy-1.3.3-cp311-cp311-win_arm64.whl", hash = "sha256:15ff10bfada4bf92ec8b31c62bf7c1834c244019b4a33095a68000d7075df470", size = 193123 }, - { url = "https://files.pythonhosted.org/packages/be/45/adfee365d9ea3d853550b2e735f9d66366701c65db7855cd07621732ccfc/contourpy-1.3.3-cp312-cp312-macosx_10_13_x86_64.whl", hash = "sha256:b08a32ea2f8e42cf1d4be3169a98dd4be32bafe4f22b6c4cb4ba810fa9e5d2cb", size = 293419 }, - { url = "https://files.pythonhosted.org/packages/53/3e/405b59cfa13021a56bba395a6b3aca8cec012b45bf177b0eaf7a202cde2c/contourpy-1.3.3-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:556dba8fb6f5d8742f2923fe9457dbdd51e1049c4a43fd3986a0b14a1d815fc6", size = 273979 }, - { url = "https://files.pythonhosted.org/packages/d4/1c/a12359b9b2ca3a845e8f7f9ac08bdf776114eb931392fcad91743e2ea17b/contourpy-1.3.3-cp312-cp312-manylinux_2_26_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:92d9abc807cf7d0e047b95ca5d957cf4792fcd04e920ca70d48add15c1a90ea7", size = 332653 }, - { url = "https://files.pythonhosted.org/packages/63/12/897aeebfb475b7748ea67b61e045accdfcf0d971f8a588b67108ed7f5512/contourpy-1.3.3-cp312-cp312-manylinux_2_26_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:b2e8faa0ed68cb29af51edd8e24798bb661eac3bd9f65420c1887b6ca89987c8", size = 379536 }, - { url = "https://files.pythonhosted.org/packages/43/8a/a8c584b82deb248930ce069e71576fc09bd7174bbd35183b7943fb1064fd/contourpy-1.3.3-cp312-cp312-manylinux_2_26_s390x.manylinux_2_28_s390x.whl", hash = "sha256:626d60935cf668e70a5ce6ff184fd713e9683fb458898e4249b63be9e28286ea", size = 384397 }, - { url = "https://files.pythonhosted.org/packages/cc/8f/ec6289987824b29529d0dfda0d74a07cec60e54b9c92f3c9da4c0ac732de/contourpy-1.3.3-cp312-cp312-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:4d00e655fcef08aba35ec9610536bfe90267d7ab5ba944f7032549c55a146da1", size = 362601 }, - { url = "https://files.pythonhosted.org/packages/05/0a/a3fe3be3ee2dceb3e615ebb4df97ae6f3828aa915d3e10549ce016302bd1/contourpy-1.3.3-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:451e71b5a7d597379ef572de31eeb909a87246974d960049a9848c3bc6c41bf7", size = 1331288 }, - { url = "https://files.pythonhosted.org/packages/33/1d/acad9bd4e97f13f3e2b18a3977fe1b4a37ecf3d38d815333980c6c72e963/contourpy-1.3.3-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:459c1f020cd59fcfe6650180678a9993932d80d44ccde1fa1868977438f0b411", size = 1403386 }, - { url = "https://files.pythonhosted.org/packages/cf/8f/5847f44a7fddf859704217a99a23a4f6417b10e5ab1256a179264561540e/contourpy-1.3.3-cp312-cp312-win32.whl", hash = "sha256:023b44101dfe49d7d53932be418477dba359649246075c996866106da069af69", size = 185018 }, - { url = "https://files.pythonhosted.org/packages/19/e8/6026ed58a64563186a9ee3f29f41261fd1828f527dd93d33b60feca63352/contourpy-1.3.3-cp312-cp312-win_amd64.whl", hash = "sha256:8153b8bfc11e1e4d75bcb0bff1db232f9e10b274e0929de9d608027e0d34ff8b", size = 226567 }, - { url = "https://files.pythonhosted.org/packages/d1/e2/f05240d2c39a1ed228d8328a78b6f44cd695f7ef47beb3e684cf93604f86/contourpy-1.3.3-cp312-cp312-win_arm64.whl", hash = "sha256:07ce5ed73ecdc4a03ffe3e1b3e3c1166db35ae7584be76f65dbbe28a7791b0cc", size = 193655 }, - { url = "https://files.pythonhosted.org/packages/68/35/0167aad910bbdb9599272bd96d01a9ec6852f36b9455cf2ca67bd4cc2d23/contourpy-1.3.3-cp313-cp313-macosx_10_13_x86_64.whl", hash = "sha256:177fb367556747a686509d6fef71d221a4b198a3905fe824430e5ea0fda54eb5", size = 293257 }, - { url = "https://files.pythonhosted.org/packages/96/e4/7adcd9c8362745b2210728f209bfbcf7d91ba868a2c5f40d8b58f54c509b/contourpy-1.3.3-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:d002b6f00d73d69333dac9d0b8d5e84d9724ff9ef044fd63c5986e62b7c9e1b1", size = 274034 }, - { url = "https://files.pythonhosted.org/packages/73/23/90e31ceeed1de63058a02cb04b12f2de4b40e3bef5e082a7c18d9c8ae281/contourpy-1.3.3-cp313-cp313-manylinux_2_26_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:348ac1f5d4f1d66d3322420f01d42e43122f43616e0f194fc1c9f5d830c5b286", size = 334672 }, - { url = "https://files.pythonhosted.org/packages/ed/93/b43d8acbe67392e659e1d984700e79eb67e2acb2bd7f62012b583a7f1b55/contourpy-1.3.3-cp313-cp313-manylinux_2_26_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:655456777ff65c2c548b7c454af9c6f33f16c8884f11083244b5819cc214f1b5", size = 381234 }, - { url = "https://files.pythonhosted.org/packages/46/3b/bec82a3ea06f66711520f75a40c8fc0b113b2a75edb36aa633eb11c4f50f/contourpy-1.3.3-cp313-cp313-manylinux_2_26_s390x.manylinux_2_28_s390x.whl", hash = "sha256:644a6853d15b2512d67881586bd03f462c7ab755db95f16f14d7e238f2852c67", size = 385169 }, - { url = "https://files.pythonhosted.org/packages/4b/32/e0f13a1c5b0f8572d0ec6ae2f6c677b7991fafd95da523159c19eff0696a/contourpy-1.3.3-cp313-cp313-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:4debd64f124ca62069f313a9cb86656ff087786016d76927ae2cf37846b006c9", size = 362859 }, - { url = "https://files.pythonhosted.org/packages/33/71/e2a7945b7de4e58af42d708a219f3b2f4cff7386e6b6ab0a0fa0033c49a9/contourpy-1.3.3-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:a15459b0f4615b00bbd1e91f1b9e19b7e63aea7483d03d804186f278c0af2659", size = 1332062 }, - { url = "https://files.pythonhosted.org/packages/12/fc/4e87ac754220ccc0e807284f88e943d6d43b43843614f0a8afa469801db0/contourpy-1.3.3-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:ca0fdcd73925568ca027e0b17ab07aad764be4706d0a925b89227e447d9737b7", size = 1403932 }, - { url = "https://files.pythonhosted.org/packages/a6/2e/adc197a37443f934594112222ac1aa7dc9a98faf9c3842884df9a9d8751d/contourpy-1.3.3-cp313-cp313-win32.whl", hash = "sha256:b20c7c9a3bf701366556e1b1984ed2d0cedf999903c51311417cf5f591d8c78d", size = 185024 }, - { url = "https://files.pythonhosted.org/packages/18/0b/0098c214843213759692cc638fce7de5c289200a830e5035d1791d7a2338/contourpy-1.3.3-cp313-cp313-win_amd64.whl", hash = "sha256:1cadd8b8969f060ba45ed7c1b714fe69185812ab43bd6b86a9123fe8f99c3263", size = 226578 }, - { url = "https://files.pythonhosted.org/packages/8a/9a/2f6024a0c5995243cd63afdeb3651c984f0d2bc727fd98066d40e141ad73/contourpy-1.3.3-cp313-cp313-win_arm64.whl", hash = "sha256:fd914713266421b7536de2bfa8181aa8c699432b6763a0ea64195ebe28bff6a9", size = 193524 }, - { url = "https://files.pythonhosted.org/packages/c0/b3/f8a1a86bd3298513f500e5b1f5fd92b69896449f6cab6a146a5d52715479/contourpy-1.3.3-cp313-cp313t-macosx_10_13_x86_64.whl", hash = "sha256:88df9880d507169449d434c293467418b9f6cbe82edd19284aa0409e7fdb933d", size = 306730 }, - { url = "https://files.pythonhosted.org/packages/3f/11/4780db94ae62fc0c2053909b65dc3246bd7cecfc4f8a20d957ad43aa4ad8/contourpy-1.3.3-cp313-cp313t-macosx_11_0_arm64.whl", hash = "sha256:d06bb1f751ba5d417047db62bca3c8fde202b8c11fb50742ab3ab962c81e8216", size = 287897 }, - { url = "https://files.pythonhosted.org/packages/ae/15/e59f5f3ffdd6f3d4daa3e47114c53daabcb18574a26c21f03dc9e4e42ff0/contourpy-1.3.3-cp313-cp313t-manylinux_2_26_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:e4e6b05a45525357e382909a4c1600444e2a45b4795163d3b22669285591c1ae", size = 326751 }, - { url = "https://files.pythonhosted.org/packages/0f/81/03b45cfad088e4770b1dcf72ea78d3802d04200009fb364d18a493857210/contourpy-1.3.3-cp313-cp313t-manylinux_2_26_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:ab3074b48c4e2cf1a960e6bbeb7f04566bf36b1861d5c9d4d8ac04b82e38ba20", size = 375486 }, - { url = "https://files.pythonhosted.org/packages/0c/ba/49923366492ffbdd4486e970d421b289a670ae8cf539c1ea9a09822b371a/contourpy-1.3.3-cp313-cp313t-manylinux_2_26_s390x.manylinux_2_28_s390x.whl", hash = "sha256:6c3d53c796f8647d6deb1abe867daeb66dcc8a97e8455efa729516b997b8ed99", size = 388106 }, - { url = "https://files.pythonhosted.org/packages/9f/52/5b00ea89525f8f143651f9f03a0df371d3cbd2fccd21ca9b768c7a6500c2/contourpy-1.3.3-cp313-cp313t-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:50ed930df7289ff2a8d7afeb9603f8289e5704755c7e5c3bbd929c90c817164b", size = 352548 }, - { url = "https://files.pythonhosted.org/packages/32/1d/a209ec1a3a3452d490f6b14dd92e72280c99ae3d1e73da74f8277d4ee08f/contourpy-1.3.3-cp313-cp313t-musllinux_1_2_aarch64.whl", hash = "sha256:4feffb6537d64b84877da813a5c30f1422ea5739566abf0bd18065ac040e120a", size = 1322297 }, - { url = "https://files.pythonhosted.org/packages/bc/9e/46f0e8ebdd884ca0e8877e46a3f4e633f6c9c8c4f3f6e72be3fe075994aa/contourpy-1.3.3-cp313-cp313t-musllinux_1_2_x86_64.whl", hash = "sha256:2b7e9480ffe2b0cd2e787e4df64270e3a0440d9db8dc823312e2c940c167df7e", size = 1391023 }, - { url = "https://files.pythonhosted.org/packages/b9/70/f308384a3ae9cd2209e0849f33c913f658d3326900d0ff5d378d6a1422d2/contourpy-1.3.3-cp313-cp313t-win32.whl", hash = "sha256:283edd842a01e3dcd435b1c5116798d661378d83d36d337b8dde1d16a5fc9ba3", size = 196157 }, - { url = "https://files.pythonhosted.org/packages/b2/dd/880f890a6663b84d9e34a6f88cded89d78f0091e0045a284427cb6b18521/contourpy-1.3.3-cp313-cp313t-win_amd64.whl", hash = "sha256:87acf5963fc2b34825e5b6b048f40e3635dd547f590b04d2ab317c2619ef7ae8", size = 240570 }, - { url = "https://files.pythonhosted.org/packages/80/99/2adc7d8ffead633234817ef8e9a87115c8a11927a94478f6bb3d3f4d4f7d/contourpy-1.3.3-cp313-cp313t-win_arm64.whl", hash = "sha256:3c30273eb2a55024ff31ba7d052dde990d7d8e5450f4bbb6e913558b3d6c2301", size = 199713 }, - { url = "https://files.pythonhosted.org/packages/72/8b/4546f3ab60f78c514ffb7d01a0bd743f90de36f0019d1be84d0a708a580a/contourpy-1.3.3-cp314-cp314-macosx_10_13_x86_64.whl", hash = "sha256:fde6c716d51c04b1c25d0b90364d0be954624a0ee9d60e23e850e8d48353d07a", size = 292189 }, - { url = "https://files.pythonhosted.org/packages/fd/e1/3542a9cb596cadd76fcef413f19c79216e002623158befe6daa03dbfa88c/contourpy-1.3.3-cp314-cp314-macosx_11_0_arm64.whl", hash = "sha256:cbedb772ed74ff5be440fa8eee9bd49f64f6e3fc09436d9c7d8f1c287b121d77", size = 273251 }, - { url = "https://files.pythonhosted.org/packages/b1/71/f93e1e9471d189f79d0ce2497007731c1e6bf9ef6d1d61b911430c3db4e5/contourpy-1.3.3-cp314-cp314-manylinux_2_26_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:22e9b1bd7a9b1d652cd77388465dc358dafcd2e217d35552424aa4f996f524f5", size = 335810 }, - { url = "https://files.pythonhosted.org/packages/91/f9/e35f4c1c93f9275d4e38681a80506b5510e9327350c51f8d4a5a724d178c/contourpy-1.3.3-cp314-cp314-manylinux_2_26_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:a22738912262aa3e254e4f3cb079a95a67132fc5a063890e224393596902f5a4", size = 382871 }, - { url = "https://files.pythonhosted.org/packages/b5/71/47b512f936f66a0a900d81c396a7e60d73419868fba959c61efed7a8ab46/contourpy-1.3.3-cp314-cp314-manylinux_2_26_s390x.manylinux_2_28_s390x.whl", hash = "sha256:afe5a512f31ee6bd7d0dda52ec9864c984ca3d66664444f2d72e0dc4eb832e36", size = 386264 }, - { url = "https://files.pythonhosted.org/packages/04/5f/9ff93450ba96b09c7c2b3f81c94de31c89f92292f1380261bd7195bea4ea/contourpy-1.3.3-cp314-cp314-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:f64836de09927cba6f79dcd00fdd7d5329f3fccc633468507079c829ca4db4e3", size = 363819 }, - { url = "https://files.pythonhosted.org/packages/3e/a6/0b185d4cc480ee494945cde102cb0149ae830b5fa17bf855b95f2e70ad13/contourpy-1.3.3-cp314-cp314-musllinux_1_2_aarch64.whl", hash = "sha256:1fd43c3be4c8e5fd6e4f2baeae35ae18176cf2e5cced681cca908addf1cdd53b", size = 1333650 }, - { url = "https://files.pythonhosted.org/packages/43/d7/afdc95580ca56f30fbcd3060250f66cedbde69b4547028863abd8aa3b47e/contourpy-1.3.3-cp314-cp314-musllinux_1_2_x86_64.whl", hash = "sha256:6afc576f7b33cf00996e5c1102dc2a8f7cc89e39c0b55df93a0b78c1bd992b36", size = 1404833 }, - { url = "https://files.pythonhosted.org/packages/e2/e2/366af18a6d386f41132a48f033cbd2102e9b0cf6345d35ff0826cd984566/contourpy-1.3.3-cp314-cp314-win32.whl", hash = "sha256:66c8a43a4f7b8df8b71ee1840e4211a3c8d93b214b213f590e18a1beca458f7d", size = 189692 }, - { url = "https://files.pythonhosted.org/packages/7d/c2/57f54b03d0f22d4044b8afb9ca0e184f8b1afd57b4f735c2fa70883dc601/contourpy-1.3.3-cp314-cp314-win_amd64.whl", hash = "sha256:cf9022ef053f2694e31d630feaacb21ea24224be1c3ad0520b13d844274614fd", size = 232424 }, - { url = "https://files.pythonhosted.org/packages/18/79/a9416650df9b525737ab521aa181ccc42d56016d2123ddcb7b58e926a42c/contourpy-1.3.3-cp314-cp314-win_arm64.whl", hash = "sha256:95b181891b4c71de4bb404c6621e7e2390745f887f2a026b2d99e92c17892339", size = 198300 }, - { url = "https://files.pythonhosted.org/packages/1f/42/38c159a7d0f2b7b9c04c64ab317042bb6952b713ba875c1681529a2932fe/contourpy-1.3.3-cp314-cp314t-macosx_10_13_x86_64.whl", hash = "sha256:33c82d0138c0a062380332c861387650c82e4cf1747aaa6938b9b6516762e772", size = 306769 }, - { url = "https://files.pythonhosted.org/packages/c3/6c/26a8205f24bca10974e77460de68d3d7c63e282e23782f1239f226fcae6f/contourpy-1.3.3-cp314-cp314t-macosx_11_0_arm64.whl", hash = "sha256:ea37e7b45949df430fe649e5de8351c423430046a2af20b1c1961cae3afcda77", size = 287892 }, - { url = "https://files.pythonhosted.org/packages/66/06/8a475c8ab718ebfd7925661747dbb3c3ee9c82ac834ccb3570be49d129f4/contourpy-1.3.3-cp314-cp314t-manylinux_2_26_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:d304906ecc71672e9c89e87c4675dc5c2645e1f4269a5063b99b0bb29f232d13", size = 326748 }, - { url = "https://files.pythonhosted.org/packages/b4/a3/c5ca9f010a44c223f098fccd8b158bb1cb287378a31ac141f04730dc49be/contourpy-1.3.3-cp314-cp314t-manylinux_2_26_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:ca658cd1a680a5c9ea96dc61cdbae1e85c8f25849843aa799dfd3cb370ad4fbe", size = 375554 }, - { url = "https://files.pythonhosted.org/packages/80/5b/68bd33ae63fac658a4145088c1e894405e07584a316738710b636c6d0333/contourpy-1.3.3-cp314-cp314t-manylinux_2_26_s390x.manylinux_2_28_s390x.whl", hash = "sha256:ab2fd90904c503739a75b7c8c5c01160130ba67944a7b77bbf36ef8054576e7f", size = 388118 }, - { url = "https://files.pythonhosted.org/packages/40/52/4c285a6435940ae25d7410a6c36bda5145839bc3f0beb20c707cda18b9d2/contourpy-1.3.3-cp314-cp314t-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:b7301b89040075c30e5768810bc96a8e8d78085b47d8be6e4c3f5a0b4ed478a0", size = 352555 }, - { url = "https://files.pythonhosted.org/packages/24/ee/3e81e1dd174f5c7fefe50e85d0892de05ca4e26ef1c9a59c2a57e43b865a/contourpy-1.3.3-cp314-cp314t-musllinux_1_2_aarch64.whl", hash = "sha256:2a2a8b627d5cc6b7c41a4beff6c5ad5eb848c88255fda4a8745f7e901b32d8e4", size = 1322295 }, - { url = "https://files.pythonhosted.org/packages/3c/b2/6d913d4d04e14379de429057cd169e5e00f6c2af3bb13e1710bcbdb5da12/contourpy-1.3.3-cp314-cp314t-musllinux_1_2_x86_64.whl", hash = "sha256:fd6ec6be509c787f1caf6b247f0b1ca598bef13f4ddeaa126b7658215529ba0f", size = 1391027 }, - { url = "https://files.pythonhosted.org/packages/93/8a/68a4ec5c55a2971213d29a9374913f7e9f18581945a7a31d1a39b5d2dfe5/contourpy-1.3.3-cp314-cp314t-win32.whl", hash = "sha256:e74a9a0f5e3fff48fb5a7f2fd2b9b70a3fe014a67522f79b7cca4c0c7e43c9ae", size = 202428 }, - { url = "https://files.pythonhosted.org/packages/fa/96/fd9f641ffedc4fa3ace923af73b9d07e869496c9cc7a459103e6e978992f/contourpy-1.3.3-cp314-cp314t-win_amd64.whl", hash = "sha256:13b68d6a62db8eafaebb8039218921399baf6e47bf85006fd8529f2a08ef33fc", size = 250331 }, - { url = "https://files.pythonhosted.org/packages/ae/8c/469afb6465b853afff216f9528ffda78a915ff880ed58813ba4faf4ba0b6/contourpy-1.3.3-cp314-cp314t-win_arm64.whl", hash = "sha256:b7448cb5a725bb1e35ce88771b86fba35ef418952474492cf7c764059933ff8b", size = 203831 }, - { url = "https://files.pythonhosted.org/packages/a5/29/8dcfe16f0107943fa92388c23f6e05cff0ba58058c4c95b00280d4c75a14/contourpy-1.3.3-pp311-pypy311_pp73-macosx_10_15_x86_64.whl", hash = "sha256:cd5dfcaeb10f7b7f9dc8941717c6c2ade08f587be2226222c12b25f0483ed497", size = 278809 }, - { url = "https://files.pythonhosted.org/packages/85/a9/8b37ef4f7dafeb335daee3c8254645ef5725be4d9c6aa70b50ec46ef2f7e/contourpy-1.3.3-pp311-pypy311_pp73-macosx_11_0_arm64.whl", hash = "sha256:0c1fc238306b35f246d61a1d416a627348b5cf0648648a031e14bb8705fcdfe8", size = 261593 }, - { url = "https://files.pythonhosted.org/packages/0a/59/ebfb8c677c75605cc27f7122c90313fd2f375ff3c8d19a1694bda74aaa63/contourpy-1.3.3-pp311-pypy311_pp73-manylinux_2_26_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:70f9aad7de812d6541d29d2bbf8feb22ff7e1c299523db288004e3157ff4674e", size = 302202 }, - { url = "https://files.pythonhosted.org/packages/3c/37/21972a15834d90bfbfb009b9d004779bd5a07a0ec0234e5ba8f64d5736f4/contourpy-1.3.3-pp311-pypy311_pp73-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:5ed3657edf08512fc3fe81b510e35c2012fbd3081d2e26160f27ca28affec989", size = 329207 }, - { url = "https://files.pythonhosted.org/packages/0c/58/bd257695f39d05594ca4ad60df5bcb7e32247f9951fd09a9b8edb82d1daa/contourpy-1.3.3-pp311-pypy311_pp73-win_amd64.whl", hash = "sha256:3d1a3799d62d45c18bafd41c5fa05120b96a28079f2393af559b843d1a966a77", size = 225315 }, -] - -[[package]] -name = "coremltools" -version = "9.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "attrs" }, - { name = "cattrs" }, - { name = "numpy" }, - { name = "packaging" }, - { name = "protobuf" }, - { name = "pyaml" }, - { name = "sympy" }, - { name = "tqdm" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/79/e6/8cb11a246f61736e75b20488b9c3cf9c208f500c9a3f92d717dbf592348c/coremltools-9.0.tar.gz", hash = "sha256:4ff346b29c31c4b45acd19a20e0f0a1ac65180a96776e62f15bd5c46f4926687", size = 1656978 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/51/a5/c16527bac75d3c5f8abde9a0e65346587886e47fea18269d7157dab40333/coremltools-9.0-cp310-none-macosx_10_15_x86_64.whl", hash = "sha256:f3247ec310eb13ce3f0e98ff76747a238ff1bde31835a2a289c84e95fe93f6a9", size = 2788452 }, - { url = "https://files.pythonhosted.org/packages/dd/b5/34f15d9f43b5b70e27602752dfc6811d029e0ec61c311991be76c23022c0/coremltools-9.0-cp310-none-macosx_11_0_arm64.whl", hash = "sha256:e9692a53b8a18891c1a54e8871de4c59ed435c5016e734c8989298b03bdb50de", size = 2763550 }, - { url = "https://files.pythonhosted.org/packages/29/4e/4dfc48820739f00dc0e6d850ac8a612d8162c89de89125022adbf2b6ac00/coremltools-9.0-cp310-none-manylinux1_x86_64.whl", hash = "sha256:8e6765539e0c830ac39755e80cef9f8ff323a46c54dda051a0562a622931094b", size = 2308133 }, - { url = "https://files.pythonhosted.org/packages/87/ab/d1e06207fd68aab62b1b476fc4ccf1fb52f43fa1816d6ab02f13c38d7c2c/coremltools-9.0-cp311-none-macosx_10_15_x86_64.whl", hash = "sha256:e6e58143c5270c1a37872fef41f8c18c042d22fa38f0ad33b33250007d9e1186", size = 2793255 }, - { url = "https://files.pythonhosted.org/packages/27/b2/8ff944f25c0fc9e5ae9deef1707029784019b78a1df2a7d0b24d581be1a2/coremltools-9.0-cp311-none-macosx_11_0_arm64.whl", hash = "sha256:0e079fea3f13f96a30587c9f7375796ff61cad53f703bde53c56fbf1374813ed", size = 2767374 }, - { url = "https://files.pythonhosted.org/packages/54/fe/58fc966635ebe3b92b100fbec875d173addab62454c4438457fa3f427d1c/coremltools-9.0-cp311-none-manylinux1_x86_64.whl", hash = "sha256:e9080254a4b9d286e168f3b1bc8616edd5d48ab664c17870b85e496629a00e81", size = 2309379 }, - { url = "https://files.pythonhosted.org/packages/d5/7e/0746b42d39d903da2015d33b619319d84fc16a44e6ed68c1a4768ae27fc5/coremltools-9.0-cp312-none-macosx_10_15_x86_64.whl", hash = "sha256:35d6e972e254081e364e6c7763eae89df8cc775dbf53756ba1ca08a2bc22f018", size = 2792457 }, - { url = "https://files.pythonhosted.org/packages/44/ab/6231b83d770825803284453f1ff36e5f3ba0a5740fcedbd3ba7454e5d412/coremltools-9.0-cp312-none-macosx_11_0_arm64.whl", hash = "sha256:7079e8b6ff5a63f0e2c08eeeb8673e4eab8ca231d4b2eae4f7fb005e0d08a8cd", size = 2764416 }, - { url = "https://files.pythonhosted.org/packages/b5/87/add15e7b4537765bef9cb47ffbd6a5d48493e65181df9864afeffa13b99b/coremltools-9.0-cp312-none-manylinux1_x86_64.whl", hash = "sha256:99a101085a7919de9f1c18e514c17d2b3e6a06ad4f7a35aae9515ad47f5a843f", size = 2308591 }, - { url = "https://files.pythonhosted.org/packages/57/4c/925ad6d76ad5bb92c9dc64798dc13651050687a03b00c03abd216dfb3732/coremltools-9.0-cp313-none-macosx_10_15_x86_64.whl", hash = "sha256:c3965805df319d5f2755d0adfb8e28312db655be09be87bd00fad097b104be57", size = 2792787 }, - { url = "https://files.pythonhosted.org/packages/62/50/76d5a828d875ed8ad7392bf9294233261747de02f7415f51d4add8dc0acf/coremltools-9.0-cp313-none-macosx_11_0_arm64.whl", hash = "sha256:9f2f858beec7f5d486cd1a59aefb452d59347e236670b67db325795bf692f480", size = 2764608 }, - { url = "https://files.pythonhosted.org/packages/a7/9f/0e8ba95ae1a1f2c9a6460c7f95a722a28a3eeaa47aef9266ef454d2e3b8b/coremltools-9.0-cp313-none-manylinux1_x86_64.whl", hash = "sha256:0af02216767232ece83bc4ec5035d7bba3c53c27de11be1a32e8461b4025d866", size = 2308338 }, -] - -[[package]] -name = "ctc-segmentation" -version = "1.7.4" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "cython" }, - { name = "numpy" }, - { name = "setuptools" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/f4/ed/78dd82e0b022ed4cb047188c0b8cead62fa0c27700737fd34e3263d83882/ctc_segmentation-1.7.4.tar.gz", hash = "sha256:19d383ea5f22438ebb1699d72b22078b63f351a33fa50bedb19c14077ba6a116", size = 73584 } - -[[package]] -name = "cuda-bindings" -version = "13.1.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "cuda-pathfinder" }, -] -wheels = [ - { url = "https://files.pythonhosted.org/packages/60/63/579402b642f5b9b8ceb79e456b39b5771f27e132a8af3b140e54d69790fc/cuda_bindings-13.1.1-cp310-cp310-manylinux_2_24_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:4400370a83f1538e25ed4c18c34a0e9d5fad39741e282e69ce24d1479a11017d", size = 15777291 }, - { url = "https://files.pythonhosted.org/packages/df/6a/3a293cfb01cd4964444a0f75917b6edb1c31ea69d0230e329975da6991ba/cuda_bindings-13.1.1-cp310-cp310-manylinux_2_24_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:81f92500e2f6aec2dac00a5a1ce77d5aa77ea77b606dc484d951f1f2cc3eaa13", size = 16311623 }, - { url = "https://files.pythonhosted.org/packages/72/b8/a5860b9e70faa53658236dc61efc3ecc51846beff4a0b73de9151130ff98/cuda_bindings-13.1.1-cp310-cp310-win_amd64.whl", hash = "sha256:3f5bb8190267216f96597235252087accac4cbccefd1b60756cced114b2d6754", size = 15185932 }, - { url = "https://files.pythonhosted.org/packages/b0/58/b8d4c7c5fb29ba46088a7e78d1065484219f8fe41a08adc4a85b1ee56149/cuda_bindings-13.1.1-cp311-cp311-manylinux_2_24_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:a5f5a6ade0ad45096568bc4dd1eb3377b65884d29124338fe9a4353130ef6631", size = 15771605 }, - { url = "https://files.pythonhosted.org/packages/17/af/710403f76f2d608d483d87089465e1f666351641dbd73d19bd025e652bad/cuda_bindings-13.1.1-cp311-cp311-manylinux_2_24_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:9348f69b03b257f07159dd4c869615e139722c2bd81e96c66f6b8f77615efd82", size = 16338970 }, - { url = "https://files.pythonhosted.org/packages/64/1c/e7ea27d4cb7d07331c88e3bbed3cacc947d2237471801086c7447b3e195d/cuda_bindings-13.1.1-cp311-cp311-win_amd64.whl", hash = "sha256:ec33b84f4bd65a86a734427f2b9cb8f221bedab2c4cfb681488cabc82f1d64ab", size = 15210672 }, - { url = "https://files.pythonhosted.org/packages/53/3d/c8ed9d169843091f3f0d6b8218e826fd59520a37e0434c204feada597988/cuda_bindings-13.1.1-cp312-cp312-manylinux_2_24_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:1e75ad0cb863330df784236d289612d71ca855c013d19ae00e5693574abd6915", size = 15530160 }, - { url = "https://files.pythonhosted.org/packages/4a/8e/368295623ee43fba622909d780fbb6863efc1638dff55f67a0f04eac6470/cuda_bindings-13.1.1-cp312-cp312-manylinux_2_24_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:25785d1a3cdcd98f151240fd5efd025609319a6720a217dee2a929241749d488", size = 16110386 }, - { url = "https://files.pythonhosted.org/packages/60/1f/ecc4701ade3e85f091c625a920574527b9daf7fb354189fbfbc5516af6cd/cuda_bindings-13.1.1-cp312-cp312-win_amd64.whl", hash = "sha256:ccde9c95c0e953b31fe7731bb08da9d0a34b1770498df9a3c156fdfdbe3951ad", size = 15250028 }, - { url = "https://files.pythonhosted.org/packages/fe/c1/0ee8fd94bab7e23116e0e3da8c0902e299f3d9edc95f1d7d8ef894c897ed/cuda_bindings-13.1.1-cp313-cp313-manylinux_2_24_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:0c9822a57c8f952dc367aacd7c32fe4cb17371104383606f455ea74635bff4c7", size = 15421116 }, - { url = "https://files.pythonhosted.org/packages/f3/c2/f272fad414b96299e010dcbe510cf17fc25deaf3443e0fdb55020a8298a3/cuda_bindings-13.1.1-cp313-cp313-manylinux_2_24_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:5837f5ea422c5653626dcfe22e9ab68142cd19af9e67a226100f224cc25a1b99", size = 15940152 }, - { url = "https://files.pythonhosted.org/packages/2a/56/433093bec0121f031edb582ea3a72f71031e8fbebecaaf329809344da4c7/cuda_bindings-13.1.1-cp313-cp313-win_amd64.whl", hash = "sha256:9e4f348cd7a779657d51e6f71aac3965fb1738f40ff3bbe75265a3242fd6f29f", size = 15216463 }, - { url = "https://files.pythonhosted.org/packages/de/38/40416d037ed25db68f1dbd50e0232775a62d90c9f25af22b196c0a13b88c/cuda_bindings-13.1.1-cp314-cp314-manylinux_2_24_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:86258fe1b0d3998bea7f57dc891569e4996705b8dd00366e44c722d0a29b2090", size = 15498927 }, - { url = "https://files.pythonhosted.org/packages/ac/3f/f1f88b6cdb7d41ba076f8ff10edf6d3bd17e740da9a163544b43d6349653/cuda_bindings-13.1.1-cp314-cp314-manylinux_2_24_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:daf8468fd603b2724c2d16cbd499348c64916ed72b1d04643f1660ce13cd12ae", size = 15984539 }, - { url = "https://files.pythonhosted.org/packages/f6/33/7739cc5e9a3373df8e7dea9060528bee5f70cf6e28b9c14f765502816c71/cuda_bindings-13.1.1-cp314-cp314-win_amd64.whl", hash = "sha256:f2e079182014dbc162562b46467815272c14c7afe5b988978fa968728b0ac726", size = 15373212 }, - { url = "https://files.pythonhosted.org/packages/9e/0a/5c6d514e566ff86c4054bbbb6554bf49b9c55fefbc934eb456faecab53c9/cuda_bindings-13.1.1-cp314-cp314t-manylinux_2_24_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:d0cd96a6ec00a78235947bff9462b2139bc5b83ce8e297d865802f0b52d1e23d", size = 15403944 }, - { url = "https://files.pythonhosted.org/packages/0b/5b/319cfa491a685d4d4757aa24223b6dbc0976954afac42f49fc47290ba6a3/cuda_bindings-13.1.1-cp314-cp314t-manylinux_2_24_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:9ff465829c6c394c2b4047250324a19925cf8c44633345b2746a4741e07bf827", size = 15911462 }, - { url = "https://files.pythonhosted.org/packages/e3/5c/38b92080c5b6c4ddb09f0be2536123f81c7e9e1a89e4573f20cb00347ee3/cuda_bindings-13.1.1-cp314-cp314t-win_amd64.whl", hash = "sha256:8205eee6b8b458a2110c0384923ace206855d0f1b436fc1b145fcbaa1653b501", size = 16044390 }, -] - -[[package]] -name = "cuda-core" -version = "0.5.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "numpy" }, -] -wheels = [ - { url = "https://files.pythonhosted.org/packages/71/31/e02ef9bcf0e8174a4939b69e044b589652c9f62e2a11262451fdfa83c552/cuda_core-0.5.0-cp310-cp310-manylinux_2_24_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:a1c6bb0649933eb505811551cf6b8dd653550d5aec9eb2298e72b11484f37738", size = 17084960 }, - { url = "https://files.pythonhosted.org/packages/b8/42/670cae16d68247de3ecc3180f5a5886628506e0efc1edfa7be0df033058e/cuda_core-0.5.0-cp310-cp310-manylinux_2_24_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:48ee7b23383f0d448338018b871ad6e119c6c28d3e078ecc55b8522096a05b2d", size = 17162246 }, - { url = "https://files.pythonhosted.org/packages/29/4b/ee26fad6998e7bcb78f8253171e64b1c2b0a529bec7e394d407acd7dbff7/cuda_core-0.5.0-cp310-cp310-win_amd64.whl", hash = "sha256:6092ba5141b11ddeb24acbbeeeecb0a02aca849a14c1ebed01663ced8cc97024", size = 4464688 }, - { url = "https://files.pythonhosted.org/packages/6b/a6/46b0647c793734244405188c039f56da952aceb262a976f2c37cec95f1c5/cuda_core-0.5.0-cp311-cp311-manylinux_2_24_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:a11cda2a59e65fcfad88f54aacfebb8c17ae02b4cb45aa3185fa8496f551f40a", size = 17751118 }, - { url = "https://files.pythonhosted.org/packages/cd/4d/b2dae70e2bf67fa40a58acccaa531e7535f54af33459455626912c55a510/cuda_core-0.5.0-cp311-cp311-manylinux_2_24_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:665375681f67a594fb0070baf23fc11baa69935538e1b73f0f4fbdfb315fcb3a", size = 17846265 }, - { url = "https://files.pythonhosted.org/packages/19/45/a8f4836b2b26824dfc0896a5d9aa6fb601043298c2ca0caef4c587fb95f7/cuda_core-0.5.0-cp311-cp311-win_amd64.whl", hash = "sha256:84be77bf2f94c583b0c89f8501aae7c8aa9b3f5d1d2884fec6af5072a276dcef", size = 4468792 }, - { url = "https://files.pythonhosted.org/packages/69/b9/067d73aa16b74357f95af443d6e9963083a3fd803d653cbfccc2c1c699a8/cuda_core-0.5.0-cp312-cp312-manylinux_2_24_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:d4aea2b61803a579444f249e20608d1332ca85889b7eb74403fa37576706fa53", size = 17692006 }, - { url = "https://files.pythonhosted.org/packages/74/cd/9d6538c6c49d008fb6694ad7ece639cbb04e74da842f2fbf3fdd1401a5df/cuda_core-0.5.0-cp312-cp312-manylinux_2_24_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:cce7e1f6ea9ea2d17ef193aad816d83d1ffbd0033f251e50bc1b406561198886", size = 17917190 }, - { url = "https://files.pythonhosted.org/packages/26/5d/8ca7e4c24e5f1226595c1f9fe196f7de0682bc0417b456f05f63cf7b896c/cuda_core-0.5.0-cp312-cp312-win_amd64.whl", hash = "sha256:aa5aadb4d946663e42cd003abca82ef8a00b447f475d62c0b186cc2f1ac7c378", size = 4447714 }, - { url = "https://files.pythonhosted.org/packages/16/31/46dccd7b519f6a55184932b0aef68553949564090d1d4f97047b47987d53/cuda_core-0.5.0-cp313-cp313-manylinux_2_24_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:1f32981d32200082ae613ad4c34312fdf25e34a7f8f41930b6b26343d554d526", size = 17487254 }, - { url = "https://files.pythonhosted.org/packages/82/3a/35d49c64617c8133f9cd5af47186ea165618442ef728f2b7441f9a84dafe/cuda_core-0.5.0-cp313-cp313-manylinux_2_24_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:25ff8d368987279afa8277f125af88667f365db2e917c0c0e9767a91408e6746", size = 17718894 }, - { url = "https://files.pythonhosted.org/packages/70/c3/ba0d0d5ffa2c73c5e3d1bb05b58aeafd24549d63a1eea7902bc380f0750e/cuda_core-0.5.0-cp313-cp313-win_amd64.whl", hash = "sha256:d1a69812c9b27c3221ee85f396656967f7bfdcd10234a5fff198152cb36c11ff", size = 4433026 }, - { url = "https://files.pythonhosted.org/packages/88/37/3cf5224ab2fd5d467f8a78523b331eeaa6cb1aaa6188af3d068a502aeaa2/cuda_core-0.5.0-cp314-cp314-manylinux_2_24_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:345404d2aa88477397b51ebe90da02f3a4161cbfa839eab5c1a9c1aee7a8a516", size = 17486435 }, - { url = "https://files.pythonhosted.org/packages/49/82/95ba2f68567f5d20f6c542543a43923babf790bb64ba8f019222f7aca47a/cuda_core-0.5.0-cp314-cp314-manylinux_2_24_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:097664d1995e01cad18b24c7a4e272023b8da9c07abeed89184281af4f36f4d6", size = 17610782 }, - { url = "https://files.pythonhosted.org/packages/ad/2e/6a12ad5ea3d3229b538c38e0c5b6c5a55041117b56964602abc54d0393ee/cuda_core-0.5.0-cp314-cp314-win_amd64.whl", hash = "sha256:52337d91801dd92a414bfed0328ddef954eab98a353d47f6177e391e11466669", size = 4519797 }, - { url = "https://files.pythonhosted.org/packages/71/0a/3860d8c75cdbb881b5b573b5765060707c22e4cbf1bf3b0ca24ea347166a/cuda_core-0.5.0-cp314-cp314t-manylinux_2_24_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:6152f5d7826f73385753db06568c0aa0ba4404c1d641434ae9ebdbac4c198799", size = 18439866 }, - { url = "https://files.pythonhosted.org/packages/d8/fa/634906819d98076c1c39efb55a5ee6d195eae1a28d284c8e0f0e47795964/cuda_core-0.5.0-cp314-cp314t-manylinux_2_24_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:019fafaea5c232853491c544daeda9d368371f0b397a9f0951d663b682ecf8f1", size = 18245433 }, - { url = "https://files.pythonhosted.org/packages/28/88/610db2649aab2bbb377c12c5d59c4bbf56a489a3e5b77ccb43f309d9027b/cuda_core-0.5.0-cp314-cp314t-win_amd64.whl", hash = "sha256:83c8b487af9807fe1d49e20e0eee980aea61492a288534e35da6879f25cf71c8", size = 4842422 }, -] - -[[package]] -name = "cuda-pathfinder" -version = "1.3.3" -source = { registry = "https://pypi.org/simple" } -wheels = [ - { url = "https://files.pythonhosted.org/packages/0b/02/4dbe7568a42e46582248942f54dc64ad094769532adbe21e525e4edf7bc4/cuda_pathfinder-1.3.3-py3-none-any.whl", hash = "sha256:9984b664e404f7c134954a771be8775dfd6180ea1e1aef4a5a37d4be05d9bbb1", size = 27154 }, -] - -[[package]] -name = "cuda-python" -version = "13.1.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "cuda-bindings" }, - { name = "cuda-pathfinder" }, -] -wheels = [ - { url = "https://files.pythonhosted.org/packages/cd/08/b5e3b9822662d72d540d830531e3ab6a7cabbda3dd56175696aabccfeb76/cuda_python-13.1.1-py3-none-any.whl", hash = "sha256:944cc4fe6482673d28dd545797a28840945a1668739328fa2ad1e9be4f7050d9", size = 8038 }, -] - -[[package]] -name = "cycler" -version = "0.12.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/a9/95/a3dbbb5028f35eafb79008e7522a75244477d2838f38cbb722248dabc2a8/cycler-0.12.1.tar.gz", hash = "sha256:88bb128f02ba341da8ef447245a9e138fae777f6a23943da4540077d3601eb1c", size = 7615 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/e7/05/c19819d5e3d95294a6f5947fb9b9629efb316b96de511b418c53d245aae6/cycler-0.12.1-py3-none-any.whl", hash = "sha256:85cef7cff222d8644161529808465972e51340599459b8ac3ccbac5a854e0d30", size = 8321 }, -] - -[[package]] -name = "cython" -version = "3.2.3" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/39/e1/c0d92b1258722e1bc62a12e630c33f1f842fdab53fd8cd5de2f75c6449a9/cython-3.2.3.tar.gz", hash = "sha256:f13832412d633376ffc08d751cc18ed0d7d00a398a4065e2871db505258748a6", size = 3276650 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/11/77/71c2aef97648548116ca22197c191f8293178f9d4e939e2cb4cbe912619e/cython-3.2.3-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:55c0157a5940fbf0b054508207fe0fc5cc796d0532af492c0fa35b5b41a883f7", size = 2959265 }, - { url = "https://files.pythonhosted.org/packages/76/b8/bc06c6427dfe46164d36c0b35e45028d0427faac28d218e065da05edcce5/cython-3.2.3-cp310-cp310-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:51fd1a56d0fc682c05ecc44f11927dbe28dd2867c30148557b62d7d1017a13d8", size = 3368365 }, - { url = "https://files.pythonhosted.org/packages/c7/3e/7550e90ccd6493842dede63ac484181d4a254ed7332eaad01253ab789d36/cython-3.2.3-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:1309bdce06f767e8514377f44b3a5b9e5b91e58af1348010cca10b572e1852ad", size = 3536996 }, - { url = "https://files.pythonhosted.org/packages/33/94/df8d414d8fb3afd5a0350245ebc589e5bc25b655342ad7341e5cfc869cf5/cython-3.2.3-cp310-cp310-win_amd64.whl", hash = "sha256:6b6dd6b7aca8447b2a6779b314cc402f1e4990754507a88477e535b3c8b41ad1", size = 2765625 }, - { url = "https://files.pythonhosted.org/packages/c3/85/77315c92d29d782bee1b36e30b8d76ad1e731cb7ea0af17e285885f3bb68/cython-3.2.3-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:c041f7e338cca2422e0924716b04fabeda57636214324fc1941396acce99e7c7", size = 2951618 }, - { url = "https://files.pythonhosted.org/packages/cb/dd/a8209e0d424a0207ddb4a3097a97b667027af3cfada762d85f3bed08ccf8/cython-3.2.3-cp311-cp311-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:283262b8f902323ceb6ed3b643f275a2a963e7ab059f0714a467933383cbc56d", size = 3243636 }, - { url = "https://files.pythonhosted.org/packages/1f/2d/bc1927fd7174f7928b86cc9b83589d39592b9273c8b1d2295ca0c0071984/cython-3.2.3-cp311-cp311-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:22a624290c2883387b2c2cfb5224c15bff21432c6a2cf0c23ac8df3dcbd45e96", size = 3378528 }, - { url = "https://files.pythonhosted.org/packages/ad/10/5add6a6e1721f9c36b5d5b4f3b75fa7af43196e4f2a474921a7277e31b7a/cython-3.2.3-cp311-cp311-win_amd64.whl", hash = "sha256:26404441f733fd1cfb0dd9c45477f501437e7d51fad05bb402bd2feb4e127aa3", size = 2769341 }, - { url = "https://files.pythonhosted.org/packages/b4/14/d16282d17c9eb2f78ca9ccd5801fed22f6c3360f5a55dbcce3c93cc70352/cython-3.2.3-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:cf210228c15b5c625824d8e31d43b6fea25f9e13c81dac632f2f7d838e0229a5", size = 2968471 }, - { url = "https://files.pythonhosted.org/packages/d0/3c/46304a942dac5a636701c55f5b05ec00ad151e6722cd068fe3d0993349bb/cython-3.2.3-cp312-cp312-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:f5bf0cebeb4147e172a114437d3fce5a507595d8fdd821be792b1bb25c691514", size = 3223581 }, - { url = "https://files.pythonhosted.org/packages/29/ad/15da606d71f40bcf2c405f84ca3d4195cb252f4eaa2f551fe6b2e630ee7c/cython-3.2.3-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:d1f8700ba89c977438744f083890d87187f15709507a5489e0f6d682053b7fa0", size = 3391391 }, - { url = "https://files.pythonhosted.org/packages/51/9e/045b35eb678682edc3e2d57112cf5ac3581a9ef274eb220b638279195678/cython-3.2.3-cp312-cp312-win_amd64.whl", hash = "sha256:25732f3981a93407826297f4423206e5e22c3cfccfc74e37bf444453bbdc076f", size = 2756814 }, - { url = "https://files.pythonhosted.org/packages/d5/c2/35cedff7fcbc844e4e872c6719df5ece26551e14f37d76eb41c412d778c6/cython-3.2.3-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:1d097ad4686b58b8c03d760d08eca28f79878d404ef7452c49636170571654e0", size = 2959019 }, - { url = "https://files.pythonhosted.org/packages/44/1b/05787f71b4834a28b19a0a3edee44537c239924f9a7d96ea38ebba365e5c/cython-3.2.3-cp313-cp313-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:2a18f2e3bcd018416157d0a83446e29b4a31437ab79061fe5504c077e70389d0", size = 3212912 }, - { url = "https://files.pythonhosted.org/packages/48/fe/f5d560e3a2eb1891d55f465d17437179d9f5fbd4f46aebf2c00d01fa5e80/cython-3.2.3-cp313-cp313-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:73afc824896ffaf22bf8122d0a7107f0120e3188a353bdcfa92317fc0d9a87ce", size = 3375222 }, - { url = "https://files.pythonhosted.org/packages/3d/b9/dcf5a68ac2ef89424657b03f751ca799861db097fa83bd52068bed198120/cython-3.2.3-cp313-cp313-win_amd64.whl", hash = "sha256:9aa1a8abf3d8bb53cc19cfaa21c004afad8d4ccb17513f8aa11a788d1f525abd", size = 2754908 }, - { url = "https://files.pythonhosted.org/packages/5c/07/93c65fbee4ab419767b7e54937e91cacae5c71d2d1277cc882ea3b1ce777/cython-3.2.3-cp314-cp314-macosx_11_0_arm64.whl", hash = "sha256:80f20369d7aaf4e76cfef902025256918a5cc6eb0aed6d8783e4b1c563e4f6c4", size = 2969476 }, - { url = "https://files.pythonhosted.org/packages/00/ad/736b4cbcb42740608cae1315c790dd6a4419705545f0615af4074e267ea3/cython-3.2.3-cp314-cp314-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:60d19376252722241a3d3ec8a695c5cae4deb053486d2e5f9a40cb569a0cf984", size = 3258714 }, - { url = "https://files.pythonhosted.org/packages/a2/74/03c08a723a319640f0bb3eaca947e009caa2eb48957ff735bfd77b0be060/cython-3.2.3-cp314-cp314-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:e4293f1861480b397809a6f021a6c12e15e918feae1c7add80c99d07af206578", size = 3384940 }, - { url = "https://files.pythonhosted.org/packages/73/14/0871a0b407fa50257a79c57a608903ed50032c7619d9531451f7090a5ee3/cython-3.2.3-cp314-cp314-win_amd64.whl", hash = "sha256:84330e7c8bf220a82b633678b9f99e10227c8f4c406d67c5552449ab2afedef8", size = 2791923 }, - { url = "https://files.pythonhosted.org/packages/43/49/afe1e3df87a770861cf17ba39f4a91f6d22a2571010fc1890b3708360630/cython-3.2.3-cp39-abi3-macosx_10_9_x86_64.whl", hash = "sha256:74f482da8b605c61b4df6ff716d013f20131949cb2fa59b03e63abd36ef5bac0", size = 2874467 }, - { url = "https://files.pythonhosted.org/packages/c7/da/044f725a083e28fb4de5bd33d13ec13f0753734b6ae52d4bc07434610cc8/cython-3.2.3-cp39-abi3-manylinux1_i686.manylinux_2_28_i686.manylinux_2_5_i686.whl", hash = "sha256:0a75a04688875b275a6c875565e672325bae04327dd6ec2fc25aeb5c6cf82fce", size = 3211272 }, - { url = "https://files.pythonhosted.org/packages/95/14/af02ba6e2e03279f2ca2956e3024a44faed4c8496bda8170b663dc3ba6e8/cython-3.2.3-cp39-abi3-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:6b01b36c9eb1b68c25bddbeef7379f7bfc37f7c9afc044e71840ffab761a2dd0", size = 2856058 }, - { url = "https://files.pythonhosted.org/packages/69/16/d254359396c2f099ab154f89b2b35f5b8b0dd21a8102c2c96a7e00291434/cython-3.2.3-cp39-abi3-musllinux_1_2_aarch64.whl", hash = "sha256:3829f99d611412288f44ff543e9d2b5c0c83274998b2a6680bbe5cca3539c1fd", size = 2993276 }, - { url = "https://files.pythonhosted.org/packages/51/0e/1a071381923e896f751f8fbff2a01c5dc8860a8b9a90066f6ec8df561dc4/cython-3.2.3-cp39-abi3-musllinux_1_2_armv7l.whl", hash = "sha256:c2365a0c79ab9c0fa86d30a4a6ba7e37fc1be9537c48b79b9d63ee7e08bf2fef", size = 2890843 }, - { url = "https://files.pythonhosted.org/packages/f4/46/1e93e10766db988e6bb8e5c6f7e2e90b9e62f1ac8dee4c1a6cf1fc170773/cython-3.2.3-cp39-abi3-musllinux_1_2_i686.whl", hash = "sha256:3141734fb15f8b5e9402b9240f8da8336edecae91742b41c85678c31ab68f66d", size = 3225339 }, - { url = "https://files.pythonhosted.org/packages/d4/ae/c284b06ae6a9c95d5883bf8744d10466cf0df64cef041a4c80ccf9fd07bd/cython-3.2.3-cp39-abi3-musllinux_1_2_x86_64.whl", hash = "sha256:9a24cc653fad3adbd9cbaa638d80df3aa08a1fe27f62eb35850971c70be680df", size = 3114751 }, - { url = "https://files.pythonhosted.org/packages/c6/d6/7795a4775c70256217134195f06b07233cf17b00f8905d5b3d782208af64/cython-3.2.3-cp39-abi3-win32.whl", hash = "sha256:b39dff92db70cbd95528f3b81d70e06bd6d3fc9c1dd91321e4d3b999ece3bceb", size = 2435616 }, - { url = "https://files.pythonhosted.org/packages/18/9e/2a3edcb858ad74e6274448dccf32150c532bc6e423f112a71f65ff3b5680/cython-3.2.3-cp39-abi3-win_arm64.whl", hash = "sha256:18edc858e6a52de47fe03ffa97ea14dadf450e20069de0a8aef531006c4bbd93", size = 2440952 }, - { url = "https://files.pythonhosted.org/packages/e5/41/54fd429ff8147475fc24ca43246f85d78fb4e747c27f227e68f1594648f1/cython-3.2.3-py3-none-any.whl", hash = "sha256:06a1317097f540d3bb6c7b81ed58a0d8b9dbfa97abf39dfd4c22ee87a6c7241e", size = 1255561 }, -] - -[[package]] -name = "cytoolz" -version = "1.1.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "toolz" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/bd/d4/16916f3dc20a3f5455b63c35dcb260b3716f59ce27a93586804e70e431d5/cytoolz-1.1.0.tar.gz", hash = "sha256:13a7bf254c3c0d28b12e2290b82aed0f0977a4c2a2bf84854fcdc7796a29f3b0", size = 642510 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/a7/7a/3244e6e3587be9abfee3b1c320e43a279831b3c3a31fe5d08c1ee6193e6b/cytoolz-1.1.0-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:72d7043a88ea5e61ba9d17ea0d1c1eff10f645d7edfcc4e56a31ef78be287644", size = 1307813 }, - { url = "https://files.pythonhosted.org/packages/32/7e/eaf504ca59addce323ef4d4ffedc2913d83c121ec19f6419bc402f7702dc/cytoolz-1.1.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:d759e9ed421bacfeb456d47af8d734c057b9912b5f2441f95b27ca35e5efab07", size = 985777 }, - { url = "https://files.pythonhosted.org/packages/d4/a1/ec95443f0cf4cd0dbc574fa26ac85a0442d35f3b601a90a0e3dda077f614/cytoolz-1.1.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:fdb5be8fbcc0396141189022724155a4c1c93712ac4aef8c03829af0c2a816d7", size = 982865 }, - { url = "https://files.pythonhosted.org/packages/a7/1b/8503604b0c0534977363fb77d371019395dfa031a216f9b1d8729d1280e4/cytoolz-1.1.0-cp310-cp310-manylinux1_i686.manylinux_2_28_i686.manylinux_2_5_i686.whl", hash = "sha256:c8c0a513dc89bc05cc72893609118815bced5ef201f1a317b4cc3423b3a0e750", size = 2597969 }, - { url = "https://files.pythonhosted.org/packages/4e/e5/30748da06417cb2d4bc58e380b0c11d8c6539f4e289dc1e4f4b4fc248d0e/cytoolz-1.1.0-cp310-cp310-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:ce94db4f8ebe842c30c0ece42ff5de977c47859088c2c363dede5a68f6906484", size = 2692230 }, - { url = "https://files.pythonhosted.org/packages/d6/84/e06580b74deb97dfd3513e4e6b660c2dedc220c7653f5bd3e4f772f4d885/cytoolz-1.1.0-cp310-cp310-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:b622d4f54e370c853ded94a668f94fe72c6d70e06ac102f17a2746661c27ab52", size = 2565243 }, - { url = "https://files.pythonhosted.org/packages/91/5e/79c0122a34c33afcb5aaee1fec35be24fe16cecefb9bb8890f2908feae56/cytoolz-1.1.0-cp310-cp310-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:375a65baa5a5b4ff6a0c5ff17e170cf23312e4c710755771ca966144c24216b5", size = 2868602 }, - { url = "https://files.pythonhosted.org/packages/3f/84/404698ff02b32292db1e39cc4a2fbdabe15164b092cc364902984c3ce0f4/cytoolz-1.1.0-cp310-cp310-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:c0d51bcdb3203a062a78f66bbe33db5e3123048e24a5f0e1402422d79df8ee2d", size = 2905121 }, - { url = "https://files.pythonhosted.org/packages/9f/33/afad6593829ba73fc87b5ae64441e380fc937f79f24a1cda60d23cb99b8c/cytoolz-1.1.0-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:1010869529bb05dc9802b6d776a34ca1b6d48b9deec70ad5e2918ae175be5c2f", size = 2684382 }, - { url = "https://files.pythonhosted.org/packages/ce/86/7900013a82ca9c6cadbfb22bf50d0fbfc3b192915d2bdd9fab3f69a9afba/cytoolz-1.1.0-cp310-cp310-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl", hash = "sha256:11a8f2e83295bdb33f35454d6bafcb7845b03b5881dcaed66ecbd726c7f16772", size = 2518183 }, - { url = "https://files.pythonhosted.org/packages/c3/4b/acf9be2953fed6a6d795fb66de37c367915037a998a5b3d3b69476cf91fe/cytoolz-1.1.0-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:0499c5e0a8e688ed367a2e51cc13792ae8f08226c15f7d168589fc44b9b9cada", size = 2609368 }, - { url = "https://files.pythonhosted.org/packages/fd/ec/3e30455fd526f5cc37bd3dd2a0e2aafb803ae4d271e50ce53bfc30810053/cytoolz-1.1.0-cp310-cp310-musllinux_1_2_armv7l.whl", hash = "sha256:87d44e6033d4c5e95a7d39ba59b8e105ba1c29b1ccd1d215f26477cc1d64be39", size = 2561458 }, - { url = "https://files.pythonhosted.org/packages/49/27/e5815c85bb18cdf95780f9596dcfd76dee910a4d635a1924648cb8a636c6/cytoolz-1.1.0-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:a68cef396a7de237f7b97422a6a450dfb111722296ba217ba5b34551832f1f6e", size = 2578236 }, - { url = "https://files.pythonhosted.org/packages/17/db/588e266eff397670398ea335a809152e77b02ee92e0ec42091115b42f09b/cytoolz-1.1.0-cp310-cp310-musllinux_1_2_ppc64le.whl", hash = "sha256:06ad4c95b258141f138a93ebfdc1d76ac087afc1a82f1401100a1f44b44ba656", size = 2770523 }, - { url = "https://files.pythonhosted.org/packages/ab/ad/82be0b999c7a0a0b362cedfc183eb090b872fd42937af2d6e97d58bc70f8/cytoolz-1.1.0-cp310-cp310-musllinux_1_2_riscv64.whl", hash = "sha256:ada59a4b3c59d4ac7162e0ed08667ffa78abf48e975c8a9f9d5b9bc50720f4fd", size = 2512909 }, - { url = "https://files.pythonhosted.org/packages/25/21/45f07ab0339a20c518bc9006100922babc397ab7ea5ef40a395db83b9cdd/cytoolz-1.1.0-cp310-cp310-musllinux_1_2_s390x.whl", hash = "sha256:a8957bcaea1ba01327a9b219d2adb84144377684f51444253890dab500ca171f", size = 2755345 }, - { url = "https://files.pythonhosted.org/packages/8b/a7/e530bf2b304206f79b36d793caba1ff9448348713a41bb1ad0197714a0f2/cytoolz-1.1.0-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:6d8cdc299d67eb0f3b9ecdafeeb55eb3b7b7470e2d950ac34b05ed4c7a5572b8", size = 2617790 }, - { url = "https://files.pythonhosted.org/packages/9f/77/7f53092121d7431589344c7d65c3d43c4111547aafabb21d3ca9032d126c/cytoolz-1.1.0-cp310-cp310-win32.whl", hash = "sha256:d8e08464c5cdea4f6df31e84b11ed6bfd79cedb99fbcbfdc15eb9361a6053c5a", size = 900209 }, - { url = "https://files.pythonhosted.org/packages/84/e4/902578658303b9bc76b1704d3ed85e6d307d311bd9fa0b919581bea56e62/cytoolz-1.1.0-cp310-cp310-win_amd64.whl", hash = "sha256:7e49922a7ed54262d41960bf3b835a7700327bf79cff1e9bfc73d79021132ff8", size = 944802 }, - { url = "https://files.pythonhosted.org/packages/71/9f/56a7003617b4eabd8ddfb470aacc240425cbe6ddeb756adfbbaadaa175f1/cytoolz-1.1.0-cp310-cp310-win_arm64.whl", hash = "sha256:943a662d2e72ffc4438d43ab5a1de8d852237775a423236594a3b3e381b8032c", size = 904835 }, - { url = "https://files.pythonhosted.org/packages/69/82/edf1d0c32b6222f2c22e5618d6db855d44eb59f9b6f22436ff963c5d0a5c/cytoolz-1.1.0-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:dba8e5a8c6e3c789d27b0eb5e7ce5ed7d032a7a9aae17ca4ba5147b871f6e327", size = 1314345 }, - { url = "https://files.pythonhosted.org/packages/2d/b5/0e3c1edaa26c2bd9db90cba0ac62c85bbca84224c7ae1c2e0072c4ea64c5/cytoolz-1.1.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:44b31c05addb0889167a720123b3b497b28dd86f8a0aeaf3ae4ffa11e2c85d55", size = 989259 }, - { url = "https://files.pythonhosted.org/packages/09/aa/e2b2ee9fc684867e817640764ea5807f9d25aa1e7bdba02dd4b249aab0f7/cytoolz-1.1.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:653cb18c4fc5d8a8cfce2bce650aabcbe82957cd0536827367d10810566d5294", size = 986551 }, - { url = "https://files.pythonhosted.org/packages/39/9f/4e8ee41acf6674f10a9c2c9117b2f219429a5a0f09bba6135f34ca4f08a6/cytoolz-1.1.0-cp311-cp311-manylinux1_i686.manylinux_2_28_i686.manylinux_2_5_i686.whl", hash = "sha256:853a5b4806915020c890e1ce70cc056bbc1dd8bc44f2d74d555cccfd7aefba7d", size = 2688378 }, - { url = "https://files.pythonhosted.org/packages/78/94/ef006f3412bc22444d855a0fc9ecb81424237fb4e5c1a1f8f5fb79ac978f/cytoolz-1.1.0-cp311-cp311-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:c7b44e9de86bea013fe84fd8c399d6016bbb96c37c5290769e5c99460b9c53e5", size = 2798299 }, - { url = "https://files.pythonhosted.org/packages/df/aa/365953926ee8b4f2e07df7200c0d73632155908c8867af14b2d19cc9f1f7/cytoolz-1.1.0-cp311-cp311-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:098d628a801dc142e9740126be5624eb7aef1d732bc7a5719f60a2095547b485", size = 2639311 }, - { url = "https://files.pythonhosted.org/packages/7c/ee/62beaaee7df208f22590ad07ef8875519af49c52ca39d99460b14a00f15a/cytoolz-1.1.0-cp311-cp311-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:779ee4096ed7a82cffab89372ffc339631c285079dbf33dbe7aff1f6174985df", size = 2979532 }, - { url = "https://files.pythonhosted.org/packages/c5/04/2211251e450bed111ada1194dc42c461da9aea441de62a01e4085ea6de9f/cytoolz-1.1.0-cp311-cp311-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:f2ce18dd99533d077e9712f9faa852f389f560351b1efd2f2bdb193a95eddde2", size = 3018632 }, - { url = "https://files.pythonhosted.org/packages/ed/a2/4a3400e4d07d3916172bf74fede08020d7b4df01595d8a97f1e9507af5ae/cytoolz-1.1.0-cp311-cp311-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:ac266a34437812cf841cecbfe19f355ab9c3dd1ef231afc60415d40ff12a76e4", size = 2788579 }, - { url = "https://files.pythonhosted.org/packages/fe/82/bb88caa53a41f600e7763c517d50e2efbbe6427ea395716a92b83f44882a/cytoolz-1.1.0-cp311-cp311-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl", hash = "sha256:1920b9b9c13d60d0bb6cd14594b3bce0870022eccb430618c37156da5f2b7a55", size = 2593024 }, - { url = "https://files.pythonhosted.org/packages/09/a8/8b25e59570da16c7a0f173b8c6ec0aa6f3abd47fd385c007485acb459896/cytoolz-1.1.0-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:47caa376dafd2bdc29f8a250acf59c810ec9105cd6f7680b9a9d070aae8490ec", size = 2715304 }, - { url = "https://files.pythonhosted.org/packages/d4/56/faec7696f235521b926ffdf92c102f5b029f072d28e1020364e55b084820/cytoolz-1.1.0-cp311-cp311-musllinux_1_2_armv7l.whl", hash = "sha256:5ab2c97d8aaa522b038cca9187b1153347af22309e7c998b14750c6fdec7b1cb", size = 2654461 }, - { url = "https://files.pythonhosted.org/packages/aa/82/f790ed167c04b8d2a33bed30770a9b7066fc4f573321d797190e5f05685f/cytoolz-1.1.0-cp311-cp311-musllinux_1_2_i686.whl", hash = "sha256:4bce006121b120e8b359244ee140bb0b1093908efc8b739db8dbaa3f8fb42139", size = 2672077 }, - { url = "https://files.pythonhosted.org/packages/d9/b3/80b8183e7eee44f45bfa3cdd3ebdadf3dd43ffc686f96d442a6c4dded45d/cytoolz-1.1.0-cp311-cp311-musllinux_1_2_ppc64le.whl", hash = "sha256:7fc0f1e4e9bb384d26e73c6657bbc26abdae4ff66a95933c00f3d578be89181b", size = 2881589 }, - { url = "https://files.pythonhosted.org/packages/8f/05/ac5ba5ddb88a3ba7ecea4bf192194a838af564d22ea7a4812cbb6bd106ce/cytoolz-1.1.0-cp311-cp311-musllinux_1_2_riscv64.whl", hash = "sha256:dd3f894ff972da1994d06ac6157d74e40dda19eb31fe5e9b7863ca4278c3a167", size = 2589924 }, - { url = "https://files.pythonhosted.org/packages/8e/cd/100483cae3849d24351c8333a815dc6adaf3f04912486e59386d86d9db9a/cytoolz-1.1.0-cp311-cp311-musllinux_1_2_s390x.whl", hash = "sha256:0846f49cf8a4496bd42659040e68bd0484ce6af819709cae234938e039203ba0", size = 2868059 }, - { url = "https://files.pythonhosted.org/packages/34/6e/3a7c56b325772d39397fc3aafb4dc054273982097178b6c3917c6dad48de/cytoolz-1.1.0-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:16a3af394ade1973226d64bb2f9eb3336adbdea03ed5b134c1bbec5a3b20028e", size = 2721692 }, - { url = "https://files.pythonhosted.org/packages/fa/ca/9fdaee32c3bc769dfb7e7991d9499136afccea67e423d097b8fb3c5acbc1/cytoolz-1.1.0-cp311-cp311-win32.whl", hash = "sha256:b786c9c8aeab76cc2f76011e986f7321a23a56d985b77d14f155d5e5514ea781", size = 899349 }, - { url = "https://files.pythonhosted.org/packages/fd/04/2ab98edeea90311e4029e1643e43d2027b54da61453292d9ea51a103ee87/cytoolz-1.1.0-cp311-cp311-win_amd64.whl", hash = "sha256:ebf06d1c5344fb22fee71bf664234733e55db72d74988f2ecb7294b05e4db30c", size = 945831 }, - { url = "https://files.pythonhosted.org/packages/b4/8d/777d86ea6bcc68b0fc926b0ef8ab51819e2176b37aadea072aac949d5231/cytoolz-1.1.0-cp311-cp311-win_arm64.whl", hash = "sha256:b63f5f025fac893393b186e132e3e242de8ee7265d0cd3f5bdd4dda93f6616c9", size = 904076 }, - { url = "https://files.pythonhosted.org/packages/c6/ec/01426224f7acf60183d3921b25e1a8e71713d3d39cb464d64ac7aace6ea6/cytoolz-1.1.0-cp312-cp312-macosx_10_13_universal2.whl", hash = "sha256:99f8e134c9be11649342853ec8c90837af4089fc8ff1e8f9a024a57d1fa08514", size = 1327800 }, - { url = "https://files.pythonhosted.org/packages/b4/07/e07e8fedd332ac9626ad58bea31416dda19bfd14310731fa38b16a97e15f/cytoolz-1.1.0-cp312-cp312-macosx_10_13_x86_64.whl", hash = "sha256:0a6f44cf9319c30feb9a50aa513d777ef51efec16f31c404409e7deb8063df64", size = 997118 }, - { url = "https://files.pythonhosted.org/packages/ab/72/c0f766d63ed2f9ea8dc8e1628d385d99b41fb834ce17ac3669e3f91e115d/cytoolz-1.1.0-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:945580dc158c557172fca899a35a99a16fbcebf6db0c77cb6621084bc82189f9", size = 991169 }, - { url = "https://files.pythonhosted.org/packages/df/4b/1f757353d1bf33e56a7391ecc9bc49c1e529803b93a9d2f67fe5f92906fe/cytoolz-1.1.0-cp312-cp312-manylinux1_i686.manylinux_2_28_i686.manylinux_2_5_i686.whl", hash = "sha256:257905ec050d04f2f856854620d1e25556fd735064cebd81b460f54939b9f9d5", size = 2700680 }, - { url = "https://files.pythonhosted.org/packages/25/73/9b25bb7ed8d419b9d6ff2ae0b3d06694de79a3f98f5169a1293ff7ad3a3f/cytoolz-1.1.0-cp312-cp312-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:82779049f352fb3ab5e8c993ab45edbb6e02efb1f17f0b50f4972c706cc51d76", size = 2824951 }, - { url = "https://files.pythonhosted.org/packages/0c/93/9c787f7c909e75670fff467f2504725d06d8c3f51d6dfe22c55a08c8ccd4/cytoolz-1.1.0-cp312-cp312-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:7d3e405e435320e08c5a1633afaf285a392e2d9cef35c925d91e2a31dfd7a688", size = 2679635 }, - { url = "https://files.pythonhosted.org/packages/50/aa/9ee92c302cccf7a41a7311b325b51ebeff25d36c1f82bdc1bbe3f58dc947/cytoolz-1.1.0-cp312-cp312-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:923df8f5591e0d20543060c29909c149ab1963a7267037b39eee03a83dbc50a8", size = 2938352 }, - { url = "https://files.pythonhosted.org/packages/6a/a3/3b58c5c1692c3bacd65640d0d5c7267a7ebb76204f7507aec29de7063d2f/cytoolz-1.1.0-cp312-cp312-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:25db9e4862f22ea0ae2e56c8bec9fc9fd756b655ae13e8c7b5625d7ed1c582d4", size = 3022121 }, - { url = "https://files.pythonhosted.org/packages/e1/93/c647bc3334355088c57351a536c2d4a83dd45f7de591fab383975e45bff9/cytoolz-1.1.0-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:c7a98deb11ccd8e5d9f9441ef2ff3352aab52226a2b7d04756caaa53cd612363", size = 2857656 }, - { url = "https://files.pythonhosted.org/packages/b2/c2/43fea146bf4141deea959e19dcddf268c5ed759dec5c2ed4a6941d711933/cytoolz-1.1.0-cp312-cp312-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl", hash = "sha256:dce4ee9fc99104bc77efdea80f32ca5a650cd653bcc8a1d984a931153d3d9b58", size = 2551284 }, - { url = "https://files.pythonhosted.org/packages/6f/df/cdc7a81ce5cfcde7ef523143d545635fc37e80ccacce140ae58483a21da3/cytoolz-1.1.0-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:80d6da158f7d20c15819701bbda1c041f0944ede2f564f5c739b1bc80a9ffb8b", size = 2721673 }, - { url = "https://files.pythonhosted.org/packages/45/be/f8524bb9ad8812ad375e61238dcaa3177628234d1b908ad0b74e3657cafd/cytoolz-1.1.0-cp312-cp312-musllinux_1_2_armv7l.whl", hash = "sha256:3b5c5a192abda123ad45ef716ec9082b4cf7d95e9ada8291c5c2cc5558be858b", size = 2722884 }, - { url = "https://files.pythonhosted.org/packages/23/e6/6bb8e4f9c267ad42d1ff77b6d2e4984665505afae50a216290e1d7311431/cytoolz-1.1.0-cp312-cp312-musllinux_1_2_i686.whl", hash = "sha256:5b399ce7d967b1cb6280250818b786be652aa8ddffd3c0bb5c48c6220d945ab5", size = 2685486 }, - { url = "https://files.pythonhosted.org/packages/d7/dd/88619f9c8d2b682562c0c886bbb7c35720cb83fda2ac9a41bdd14073d9bd/cytoolz-1.1.0-cp312-cp312-musllinux_1_2_ppc64le.whl", hash = "sha256:e7e29a1a03f00b4322196cfe8e2c38da9a6c8d573566052c586df83aacc5663c", size = 2839661 }, - { url = "https://files.pythonhosted.org/packages/b8/8d/4478ebf471ee78dd496d254dc0f4ad729cd8e6ba8257de4f0a98a2838ef2/cytoolz-1.1.0-cp312-cp312-musllinux_1_2_riscv64.whl", hash = "sha256:5291b117d71652a817ec164e7011f18e6a51f8a352cc9a70ed5b976c51102fda", size = 2547095 }, - { url = "https://files.pythonhosted.org/packages/e6/68/f1dea33367b0b3f64e199c230a14a6b6f243c189020effafd31e970ca527/cytoolz-1.1.0-cp312-cp312-musllinux_1_2_s390x.whl", hash = "sha256:8caef62f846a9011676c51bda9189ae394cdd6bb17f2946ecaedc23243268320", size = 2870901 }, - { url = "https://files.pythonhosted.org/packages/4a/9a/33591c09dfe799b8fb692cf2ad383e2c41ab6593cc960b00d1fc8a145655/cytoolz-1.1.0-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:de425c5a8e3be7bb3a195e19191d28d9eb3c2038046064a92edc4505033ec9cb", size = 2765422 }, - { url = "https://files.pythonhosted.org/packages/60/2b/a8aa233c9416df87f004e57ae4280bd5e1f389b4943d179f01020c6ec629/cytoolz-1.1.0-cp312-cp312-win32.whl", hash = "sha256:296440a870e8d1f2e1d1edf98f60f1532b9d3ab8dfbd4b25ec08cd76311e79e5", size = 901933 }, - { url = "https://files.pythonhosted.org/packages/ad/33/4c9bdf8390dc01d2617c7f11930697157164a52259b6818ddfa2f94f89f4/cytoolz-1.1.0-cp312-cp312-win_amd64.whl", hash = "sha256:07156987f224c6dac59aa18fb8bf91e1412f5463961862716a3381bf429c8699", size = 947989 }, - { url = "https://files.pythonhosted.org/packages/35/ac/6e2708835875f5acb52318462ed296bf94ed0cb8c7cb70e62fbd03f709e3/cytoolz-1.1.0-cp312-cp312-win_arm64.whl", hash = "sha256:23e616b38f5b3160c7bb45b0f84a8f3deb4bd26b29fb2dfc716f241c738e27b8", size = 903913 }, - { url = "https://files.pythonhosted.org/packages/71/4a/b3ddb3ee44fe0045e95dd973746f93f033b6f92cce1fc3cbbe24b329943c/cytoolz-1.1.0-cp313-cp313-ios_13_0_arm64_iphoneos.whl", hash = "sha256:76c9b58555300be6dde87a41faf1f97966d79b9a678b7a526fcff75d28ef4945", size = 976728 }, - { url = "https://files.pythonhosted.org/packages/42/21/a3681434aa425875dd828bb515924b0f12c37a55c7d2bc5c0c5de3aeb0b4/cytoolz-1.1.0-cp313-cp313-ios_13_0_arm64_iphonesimulator.whl", hash = "sha256:d1d638b10d3144795655e9395566ce35807df09219fd7cacd9e6acbdef67946a", size = 986057 }, - { url = "https://files.pythonhosted.org/packages/d9/cb/efc1b29e211e0670a6953222afaac84dcbba5cb940b130c0e49858978040/cytoolz-1.1.0-cp313-cp313-ios_13_0_x86_64_iphonesimulator.whl", hash = "sha256:26801c1a165e84786a99e03c9c9973356caaca002d66727b761fb1042878ef06", size = 992632 }, - { url = "https://files.pythonhosted.org/packages/be/b0/e50621d21e939338c97faab651f58ea7fa32101226a91de79ecfb89d71e1/cytoolz-1.1.0-cp313-cp313-macosx_10_13_universal2.whl", hash = "sha256:2a9a464542912d3272f6dccc5142df057c71c6a5cbd30439389a732df401afb7", size = 1317534 }, - { url = "https://files.pythonhosted.org/packages/0d/6b/25aa9739b0235a5bc4c1ea293186bc6822a4c6607acfe1422423287e7400/cytoolz-1.1.0-cp313-cp313-macosx_10_13_x86_64.whl", hash = "sha256:ed6104fa942aa5784bf54f339563de637557e3443b105760bc4de8f16a7fc79b", size = 992336 }, - { url = "https://files.pythonhosted.org/packages/e1/53/5f4deb0ff958805309d135d899c764364c1e8a632ce4994bd7c45fb98df2/cytoolz-1.1.0-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:56161f0ab60dc4159ec343509abaf809dc88e85c7e420e354442c62e3e7cbb77", size = 986118 }, - { url = "https://files.pythonhosted.org/packages/1c/e3/f6255b76c8cc0debbe1c0779130777dc0434da6d9b28a90d9f76f8cb67cd/cytoolz-1.1.0-cp313-cp313-manylinux1_i686.manylinux_2_28_i686.manylinux_2_5_i686.whl", hash = "sha256:832bd36cc9123535f1945acf6921f8a2a15acc19cfe4065b1c9b985a28671886", size = 2679563 }, - { url = "https://files.pythonhosted.org/packages/59/8a/acc6e39a84e930522b965586ad3a36694f9bf247b23188ee0eb47b1c9ed1/cytoolz-1.1.0-cp313-cp313-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:1842636b6e034f229bf084c2bcdcfd36c8437e752eefd2c74ce9e2f10415cb6e", size = 2813020 }, - { url = "https://files.pythonhosted.org/packages/db/f5/0083608286ad1716eda7c41f868e85ac549f6fd6b7646993109fa0bdfd98/cytoolz-1.1.0-cp313-cp313-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:823df012ab90d2f2a0f92fea453528539bf71ac1879e518524cd0c86aa6df7b9", size = 2669312 }, - { url = "https://files.pythonhosted.org/packages/47/a8/d16080b575520fe5da00cede1ece4e0a4180ec23f88dcdc6a2f5a90a7f7f/cytoolz-1.1.0-cp313-cp313-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:2f1fcf9e7e7b3487883ff3f815abc35b89dcc45c4cf81c72b7ee457aa72d197b", size = 2922147 }, - { url = "https://files.pythonhosted.org/packages/7e/bc/716c9c1243701e58cad511eb3937fd550e645293c5ed1907639c5d66f194/cytoolz-1.1.0-cp313-cp313-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:4cdb3fa1772116827f263f25b0cdd44c663b6701346a56411960534a06c082de", size = 2981602 }, - { url = "https://files.pythonhosted.org/packages/14/bc/571b232996846b27f4ac0c957dc8bf60261e9b4d0d01c8d955e82329544e/cytoolz-1.1.0-cp313-cp313-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:d1b5c95041741b81430454db65183e133976f45ac3c03454cfa8147952568529", size = 2830103 }, - { url = "https://files.pythonhosted.org/packages/5b/55/c594afb46ecd78e4b7e1fb92c947ed041807875661ceda73baaf61baba4f/cytoolz-1.1.0-cp313-cp313-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl", hash = "sha256:b2079fd9f1a65f4c61e6278c8a6d4f85edf30c606df8d5b32f1add88cbbe2286", size = 2533802 }, - { url = "https://files.pythonhosted.org/packages/93/83/1edcf95832555a78fc43b975f3ebe8ceadcc9664dd47fd33747a14df5069/cytoolz-1.1.0-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:a92a320d72bef1c7e2d4c6d875125cf57fc38be45feb3fac1bfa64ea401f54a4", size = 2706071 }, - { url = "https://files.pythonhosted.org/packages/e2/df/035a408df87f25cfe3611557818b250126cd2281b2104cd88395de205583/cytoolz-1.1.0-cp313-cp313-musllinux_1_2_armv7l.whl", hash = "sha256:06d1c79aa51e6a92a90b0e456ebce2288f03dd6a76c7f582bfaa3eda7692e8a5", size = 2707575 }, - { url = "https://files.pythonhosted.org/packages/7a/a4/ef78e13e16e93bf695a9331321d75fbc834a088d941f1c19e6b63314e257/cytoolz-1.1.0-cp313-cp313-musllinux_1_2_i686.whl", hash = "sha256:e1d7be25f6971e986a52b6d3a0da28e1941850985417c35528f6823aef2cfec5", size = 2660486 }, - { url = "https://files.pythonhosted.org/packages/30/7a/2c3d60682b26058d435416c4e90d4a94db854de5be944dfd069ed1be648a/cytoolz-1.1.0-cp313-cp313-musllinux_1_2_ppc64le.whl", hash = "sha256:964b248edc31efc50a65e9eaa0c845718503823439d2fa5f8d2c7e974c2b5409", size = 2819605 }, - { url = "https://files.pythonhosted.org/packages/45/92/19b722a1d83cc443fbc0c16e0dc376f8a451437890d3d9ee370358cf0709/cytoolz-1.1.0-cp313-cp313-musllinux_1_2_riscv64.whl", hash = "sha256:c9ff2b3c57c79b65cb5be14a18c6fd4a06d5036fb3f33e973a9f70e9ac13ca28", size = 2533559 }, - { url = "https://files.pythonhosted.org/packages/1d/15/fa3b7891da51115204416f14192081d3dea0eaee091f123fdc1347de8dd1/cytoolz-1.1.0-cp313-cp313-musllinux_1_2_s390x.whl", hash = "sha256:22290b73086af600042d99f5ce52a43d4ad9872c382610413176e19fc1d4fd2d", size = 2839171 }, - { url = "https://files.pythonhosted.org/packages/46/40/d3519d5cd86eebebf1e8b7174ec32dfb6ecec67b48b0cfb92bf226659b5a/cytoolz-1.1.0-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:a2ade74fccd080ea793382968913ee38d7a35c921df435bbf0a6aeecf0d17574", size = 2743379 }, - { url = "https://files.pythonhosted.org/packages/93/e2/a9e7511f0a13fdbefa5bf73cf8e4763878140de9453fd3e50d6ac57b6be7/cytoolz-1.1.0-cp313-cp313-win32.whl", hash = "sha256:db5dbcfda1c00e937426cbf9bdc63c24ebbc358c3263bfcbc1ab4a88dc52aa8e", size = 900844 }, - { url = "https://files.pythonhosted.org/packages/d6/a4/fb7eb403c6a4c81e5a30363f34a71adcc8bf5292dc8ea32e2440aa5668f2/cytoolz-1.1.0-cp313-cp313-win_amd64.whl", hash = "sha256:9e2d3fe3b45c3eb7233746f7aca37789be3dceec3e07dcc406d3e045ea0f7bdc", size = 946461 }, - { url = "https://files.pythonhosted.org/packages/93/bb/1c8c33d353548d240bc6e8677ee8c3560ce5fa2f084e928facf7c35a6dcf/cytoolz-1.1.0-cp313-cp313-win_arm64.whl", hash = "sha256:32c559f95ff44a9ebcbd934acaa1e6dc8f3e6ffce4762a79a88528064873d6d5", size = 902673 }, - { url = "https://files.pythonhosted.org/packages/c4/ba/4a53acc60f59030fcaf48c7766e3c4c81bd997379425aa45b129396557b5/cytoolz-1.1.0-cp313-cp313t-macosx_10_13_universal2.whl", hash = "sha256:9e2cd93b28f667c5870a070ab2b8bb4397470a85c4b204f2454b0ad001cd1ca3", size = 1372336 }, - { url = "https://files.pythonhosted.org/packages/ac/90/f28fd8ad8319d8f5c8da69a2c29b8cf52a6d2c0161602d92b366d58926ab/cytoolz-1.1.0-cp313-cp313t-macosx_10_13_x86_64.whl", hash = "sha256:f494124e141a9361f31d79875fe7ea459a3be2b9dadd90480427c0c52a0943d4", size = 1011930 }, - { url = "https://files.pythonhosted.org/packages/c9/95/4561c4e0ad1c944f7673d6d916405d68080f10552cfc5d69a1cf2475a9a1/cytoolz-1.1.0-cp313-cp313t-macosx_11_0_arm64.whl", hash = "sha256:53a3262bf221f19437ed544bf8c0e1980c81ac8e2a53d87a9bc075dba943d36f", size = 1020610 }, - { url = "https://files.pythonhosted.org/packages/c3/14/b2e1ffa4995ec36e1372e243411ff36325e4e6d7ffa34eb4098f5357d176/cytoolz-1.1.0-cp313-cp313t-manylinux1_i686.manylinux_2_28_i686.manylinux_2_5_i686.whl", hash = "sha256:47663e57d3f3f124921f38055e86a1022d0844c444ede2e8f090d3bbf80deb65", size = 2917327 }, - { url = "https://files.pythonhosted.org/packages/4a/29/7cab6c609b4514ac84cca2f7dca6c509977a8fc16d27c3a50e97f105fa6a/cytoolz-1.1.0-cp313-cp313t-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:a5a8755c4104ee4e3d5ba434c543b5f85fdee6a1f1df33d93f518294da793a60", size = 3108951 }, - { url = "https://files.pythonhosted.org/packages/9a/71/1d1103b819458679277206ad07d78ca6b31c4bb88d6463fd193e19bfb270/cytoolz-1.1.0-cp313-cp313t-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:4d96ff3d381423af1b105295f97de86d1db51732c9566eb37378bab6670c5010", size = 2807149 }, - { url = "https://files.pythonhosted.org/packages/1a/d4/3d83a05a21e7d2ed2b9e6daf489999c29934b005de9190272b8a2e3735d0/cytoolz-1.1.0-cp313-cp313t-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:0ec96b3d537cdf47d4e76ded199f7440715f4c71029b45445cff92c1248808c2", size = 3111608 }, - { url = "https://files.pythonhosted.org/packages/51/88/96f68354c3d4af68de41f0db4fe41a23b96a50a4a416636cea325490cfeb/cytoolz-1.1.0-cp313-cp313t-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:208e2f2ef90a32b0acbff3303d90d89b13570a228d491d2e622a7883a3c68148", size = 3179373 }, - { url = "https://files.pythonhosted.org/packages/ce/50/ed87a5cd8e6f27ffbb64c39e9730e18ec66c37631db2888ae711909f10c9/cytoolz-1.1.0-cp313-cp313t-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:0d416a81bb0bd517558668e49d30a7475b5445f9bbafaab7dcf066f1e9adba36", size = 3003120 }, - { url = "https://files.pythonhosted.org/packages/d3/a7/acde155b050d6eaa8e9c7845c98fc5fb28501568e78e83ebbf44f8855274/cytoolz-1.1.0-cp313-cp313t-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl", hash = "sha256:f32e94c91ffe49af04835ee713ebd8e005c85ebe83e7e1fdcc00f27164c2d636", size = 2703225 }, - { url = "https://files.pythonhosted.org/packages/1b/b6/9d518597c5bdea626b61101e8d2ff94124787a42259dafd9f5fc396f346a/cytoolz-1.1.0-cp313-cp313t-musllinux_1_2_aarch64.whl", hash = "sha256:15d0c6405efc040499c46df44056a5c382f551a7624a41cf3e4c84a96b988a15", size = 2956033 }, - { url = "https://files.pythonhosted.org/packages/89/7a/93e5f860926165538c85e1c5e1670ad3424f158df810f8ccd269da652138/cytoolz-1.1.0-cp313-cp313t-musllinux_1_2_armv7l.whl", hash = "sha256:bf069c5381d757debae891401b88b3a346ba3a28ca45ba9251103b282463fad8", size = 2862950 }, - { url = "https://files.pythonhosted.org/packages/76/e6/99d6af00487bedc27597b54c9fcbfd5c833a69c6b7a9b9f0fff777bfc7aa/cytoolz-1.1.0-cp313-cp313t-musllinux_1_2_i686.whl", hash = "sha256:7d5cf15892e63411ec1bd67deff0e84317d974e6ab2cdfefdd4a7cea2989df66", size = 2861757 }, - { url = "https://files.pythonhosted.org/packages/71/ca/adfa1fb7949478135a37755cb8e88c20cd6b75c22a05f1128f05f3ab2c60/cytoolz-1.1.0-cp313-cp313t-musllinux_1_2_ppc64le.whl", hash = "sha256:3e3872c21170f8341656f8692f8939e8800dcee6549ad2474d4c817bdefd62cd", size = 2979049 }, - { url = "https://files.pythonhosted.org/packages/70/4c/7bf47a03a4497d500bc73d4204e2d907771a017fa4457741b2a1d7c09319/cytoolz-1.1.0-cp313-cp313t-musllinux_1_2_riscv64.whl", hash = "sha256:b9ddeff8e8fd65eb1fcefa61018100b2b627e759ea6ad275d2e2a93ffac147bf", size = 2699492 }, - { url = "https://files.pythonhosted.org/packages/7e/e7/3d034b0e4817314f07aa465d5864e9b8df9d25cb260a53dd84583e491558/cytoolz-1.1.0-cp313-cp313t-musllinux_1_2_s390x.whl", hash = "sha256:02feeeda93e1fa3b33414eb57c2b0aefd1db8f558dd33fdfcce664a0f86056e4", size = 2995646 }, - { url = "https://files.pythonhosted.org/packages/c1/62/be357181c71648d9fe1d1ce91cd42c63457dcf3c158e144416fd51dced83/cytoolz-1.1.0-cp313-cp313t-musllinux_1_2_x86_64.whl", hash = "sha256:d08154ad45349162b6c37f12d5d1b2e6eef338e657b85e1621e4e6a4a69d64cb", size = 2919481 }, - { url = "https://files.pythonhosted.org/packages/62/d5/bf5434fde726c4f80cb99912b2d8e0afa1587557e2a2d7e0315eb942f2de/cytoolz-1.1.0-cp313-cp313t-win32.whl", hash = "sha256:10ae4718a056948d73ca3e1bb9ab1f95f897ec1e362f829b9d37cc29ab566c60", size = 951595 }, - { url = "https://files.pythonhosted.org/packages/64/29/39c161e9204a9715321ddea698cbd0abc317e78522c7c642363c20589e71/cytoolz-1.1.0-cp313-cp313t-win_amd64.whl", hash = "sha256:1bb77bc6197e5cb19784b6a42bb0f8427e81737a630d9d7dda62ed31733f9e6c", size = 1004445 }, - { url = "https://files.pythonhosted.org/packages/e2/5a/7cbff5e9a689f558cb0bdf277f9562b2ac51acf7cd15e055b8c3efb0e1ef/cytoolz-1.1.0-cp313-cp313t-win_arm64.whl", hash = "sha256:563dda652c6ff52d215704fbe6b491879b78d7bbbb3a9524ec8e763483cb459f", size = 926207 }, - { url = "https://files.pythonhosted.org/packages/b7/e8/297a85ba700f437c01eba962428e6ab4572f6c3e68e8ff442ce5c9d3a496/cytoolz-1.1.0-cp314-cp314-ios_13_0_arm64_iphoneos.whl", hash = "sha256:d542cee7c7882d2a914a33dec4d3600416fb336734df979473249d4c53d207a1", size = 980613 }, - { url = "https://files.pythonhosted.org/packages/e8/d7/2b02c9d18e9cc263a0e22690f78080809f1eafe72f26b29ccc115d3bf5c8/cytoolz-1.1.0-cp314-cp314-ios_13_0_arm64_iphonesimulator.whl", hash = "sha256:31922849b701b0f24bb62e56eb2488dcd3aa6ae3057694bd6b3b7c4c2bc27c2f", size = 990476 }, - { url = "https://files.pythonhosted.org/packages/89/26/b6b159d2929310fca0eff8a4989cd4b1ecbdf7c46fdff46c7a20fcae55c8/cytoolz-1.1.0-cp314-cp314-ios_13_0_x86_64_iphonesimulator.whl", hash = "sha256:e68308d32afd31943314735c1335e4ab5696110e96b405f6bdb8f2a8dc771a16", size = 992712 }, - { url = "https://files.pythonhosted.org/packages/42/a0/f7c572aa151ed466b0fce4a327c3cc916d3ef3c82e341be59ea4b9bee9e4/cytoolz-1.1.0-cp314-cp314-macosx_10_15_universal2.whl", hash = "sha256:fc4bb48b3b866e1867f7c6411a4229e5b44be3989060663713e10efc24c9bd5f", size = 1322596 }, - { url = "https://files.pythonhosted.org/packages/72/7c/a55d035e20b77b6725e85c8f1a418b3a4c23967288b8b0c2d1a40f158cbe/cytoolz-1.1.0-cp314-cp314-macosx_10_15_x86_64.whl", hash = "sha256:456f77207d1445025d7ef262b8370a05492dcb1490cb428b0f3bf1bd744a89b0", size = 992825 }, - { url = "https://files.pythonhosted.org/packages/03/af/39d2d3db322136e12e9336a1f13bab51eab88b386bfb11f91d3faff8ba34/cytoolz-1.1.0-cp314-cp314-macosx_11_0_arm64.whl", hash = "sha256:174ebc71ebb20a9baeffce6ee07ee2cd913754325c93f99d767380d8317930f7", size = 990525 }, - { url = "https://files.pythonhosted.org/packages/a6/bd/65d7a869d307f9b10ad45c2c1cbb40b81a8d0ed1138fa17fd904f5c83298/cytoolz-1.1.0-cp314-cp314-manylinux1_i686.manylinux_2_28_i686.manylinux_2_5_i686.whl", hash = "sha256:8b3604fef602bcd53415055a4f68468339192fd17be39e687ae24f476d23d56e", size = 2672409 }, - { url = "https://files.pythonhosted.org/packages/2d/fb/74dfd844bfd67e810bd36e8e3903a143035447245828e7fcd7c81351d775/cytoolz-1.1.0-cp314-cp314-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:3604b959a01f64c366e7d10ec7634d5f5cfe10301e27a8f090f6eb3b2a628a18", size = 2808477 }, - { url = "https://files.pythonhosted.org/packages/d6/1f/587686c43e31c19241ec317da66438d093523921ea7749bbc65558a30df9/cytoolz-1.1.0-cp314-cp314-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:6db2127a3c1bc2f59f08010d2ae53a760771a9de2f67423ad8d400e9ba4276e8", size = 2636881 }, - { url = "https://files.pythonhosted.org/packages/bc/6d/90468cd34f77cb38a11af52c4dc6199efcc97a486395a21bef72e9b7602e/cytoolz-1.1.0-cp314-cp314-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:56584745ac647993a016a21bc76399113b7595e312f8d0a1b140c9fcf9b58a27", size = 2937315 }, - { url = "https://files.pythonhosted.org/packages/d9/50/7b92cd78c613b92e3509e6291d3fb7e0d72ebda999a8df806a96c40ca9ab/cytoolz-1.1.0-cp314-cp314-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:db2c4c3a7f7bd7e03bb1a236a125c8feb86c75802f4ecda6ecfaf946610b2930", size = 2959988 }, - { url = "https://files.pythonhosted.org/packages/44/d5/34b5a28a8d9bb329f984b4c2259407ca3f501d1abeb01bacea07937d85d1/cytoolz-1.1.0-cp314-cp314-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:48cb8a692111a285d2b9acd16d185428176bfbffa8a7c274308525fccd01dd42", size = 2795116 }, - { url = "https://files.pythonhosted.org/packages/f5/d9/5dd829e33273ec03bdc3c812e6c3281987ae2c5c91645582f6c331544a64/cytoolz-1.1.0-cp314-cp314-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl", hash = "sha256:d2f344ba5eb17dcf38ee37fdde726f69053f54927db8f8a1bed6ac61e5b1890d", size = 2535390 }, - { url = "https://files.pythonhosted.org/packages/87/1f/7f9c58068a8eec2183110df051bc6b69dd621143f84473eeb6dc1b32905a/cytoolz-1.1.0-cp314-cp314-musllinux_1_2_aarch64.whl", hash = "sha256:abf76b1c1abd031f098f293b6d90ee08bdaa45f8b5678430e331d991b82684b1", size = 2704834 }, - { url = "https://files.pythonhosted.org/packages/d2/90/667def5665333575d01a65fe3ec0ca31b897895f6e3bc1a42d6ea3659369/cytoolz-1.1.0-cp314-cp314-musllinux_1_2_armv7l.whl", hash = "sha256:ddf9a38a5b686091265ff45b53d142e44a538cd6c2e70610d3bc6be094219032", size = 2658441 }, - { url = "https://files.pythonhosted.org/packages/23/79/6615f9a14960bd29ac98b823777b6589357833f65cf1a11b5abc1587c120/cytoolz-1.1.0-cp314-cp314-musllinux_1_2_i686.whl", hash = "sha256:946786755274f07bb2be0400f28adb31d7d85a7c7001873c0a8e24a503428fb3", size = 2654766 }, - { url = "https://files.pythonhosted.org/packages/b0/99/be59c6e0ae02153ef10ae1ff0f380fb19d973c651b50cf829a731f6c9e79/cytoolz-1.1.0-cp314-cp314-musllinux_1_2_ppc64le.whl", hash = "sha256:d5b8f78b9fed79cf185ad4ddec099abeef45951bdcb416c5835ba05f0a1242c7", size = 2827649 }, - { url = "https://files.pythonhosted.org/packages/19/b7/854ddcf9f9618844108677c20d48f4611b5c636956adea0f0e85e027608f/cytoolz-1.1.0-cp314-cp314-musllinux_1_2_riscv64.whl", hash = "sha256:fccde6efefdbc02e676ccb352a2ccc8a8e929f59a1c6d3d60bb78e923a49ca44", size = 2533456 }, - { url = "https://files.pythonhosted.org/packages/45/66/bfe6fbb2bdcf03c8377c8c2f542576e15f3340c905a09d78a6cb3badd39a/cytoolz-1.1.0-cp314-cp314-musllinux_1_2_s390x.whl", hash = "sha256:717b7775313da5f51b0fbf50d865aa9c39cb241bd4cb605df3cf2246d6567397", size = 2826455 }, - { url = "https://files.pythonhosted.org/packages/c3/0c/cce4047bd927e95f59e73319c02c9bc86bd3d76392e0eb9e41a1147a479c/cytoolz-1.1.0-cp314-cp314-musllinux_1_2_x86_64.whl", hash = "sha256:5158744a09d0e0e4a4f82225e3a3c4ebf38f9ae74467aaa905467270e52f2794", size = 2714897 }, - { url = "https://files.pythonhosted.org/packages/ac/9a/061323bb289b565802bad14fb7ab59fcd8713105df142bcf4dd9ff64f8ac/cytoolz-1.1.0-cp314-cp314-win32.whl", hash = "sha256:1ed534bdbbf063b2bb28fca7d0f6723a3e5a72b086e7c7fe6d74ae8c3e4d00e2", size = 901490 }, - { url = "https://files.pythonhosted.org/packages/a3/20/1f3a733d710d2a25d6f10b463bef55ada52fe6392a5d233c8d770191f48a/cytoolz-1.1.0-cp314-cp314-win_amd64.whl", hash = "sha256:472c1c9a085f5ad973ec0ad7f0b9ba0969faea6f96c9e397f6293d386f3a25ec", size = 946730 }, - { url = "https://files.pythonhosted.org/packages/f2/22/2d657db4a5d1c10a152061800f812caba9ef20d7bd2406f51a5fd800c180/cytoolz-1.1.0-cp314-cp314-win_arm64.whl", hash = "sha256:a7ad7ca3386fa86bd301be3fa36e7f0acb024f412f665937955acfc8eb42deff", size = 905722 }, - { url = "https://files.pythonhosted.org/packages/19/97/b4a8c76796a9a8b9bc90c7992840fa1589a1af8e0426562dea4ce9b384a7/cytoolz-1.1.0-cp314-cp314t-macosx_10_15_universal2.whl", hash = "sha256:64b63ed4b71b1ba813300ad0f06b8aff19a12cf51116e0e4f1ed837cea4debcf", size = 1372606 }, - { url = "https://files.pythonhosted.org/packages/08/d4/a1bb1a32b454a2d650db8374ff3bf875ba0fc1c36e6446ec02a83b9140a1/cytoolz-1.1.0-cp314-cp314t-macosx_10_15_x86_64.whl", hash = "sha256:a60ba6f2ed9eb0003a737e1ee1e9fa2258e749da6477946008d4324efa25149f", size = 1012189 }, - { url = "https://files.pythonhosted.org/packages/21/4b/2f5cbbd81588918ee7dd70cffb66731608f578a9b72166aafa991071af7d/cytoolz-1.1.0-cp314-cp314t-macosx_11_0_arm64.whl", hash = "sha256:1aa58e2434d732241f7f051e6f17657e969a89971025e24578b5cbc6f1346485", size = 1020624 }, - { url = "https://files.pythonhosted.org/packages/f5/99/c4954dd86cd593cd776a038b36795a259b8b5c12cbab6363edf5f6d9c909/cytoolz-1.1.0-cp314-cp314t-manylinux1_i686.manylinux_2_28_i686.manylinux_2_5_i686.whl", hash = "sha256:6965af3fc7214645970e312deb9bd35a213a1eaabcfef4f39115e60bf2f76867", size = 2917016 }, - { url = "https://files.pythonhosted.org/packages/b2/7c/f1f70a17e272b433232bc8a27df97e46b202d6cc07e3b0d63f7f41ba0f2d/cytoolz-1.1.0-cp314-cp314t-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:ddd2863f321d67527d3b67a93000a378ad6f967056f68c06467fe011278a6d0e", size = 3107634 }, - { url = "https://files.pythonhosted.org/packages/8f/bd/c3226a57474b4aef1f90040510cba30d0decd3515fed48dc229b37c2f898/cytoolz-1.1.0-cp314-cp314t-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:4e6b428e9eb5126053c2ae0efa62512ff4b38ed3951f4d0888ca7005d63e56f5", size = 2806221 }, - { url = "https://files.pythonhosted.org/packages/c3/47/2f7bfe4aaa1e07dc9828bea228ed744faf73b26aee0c1bdf3b5520bf1909/cytoolz-1.1.0-cp314-cp314t-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:d758e5ef311d2671e0ae8c214c52e44617cf1e58bef8f022b547b9802a5a7f30", size = 3107671 }, - { url = "https://files.pythonhosted.org/packages/4d/12/6ff3b04fbd1369d0fcd5f8b5910ba6e427e33bf113754c4c35ec3f747924/cytoolz-1.1.0-cp314-cp314t-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:a95416eca473e6c1179b48d86adcf528b59c63ce78f4cb9934f2e413afa9b56b", size = 3176350 }, - { url = "https://files.pythonhosted.org/packages/e6/8c/6691d986b728e77b5d2872743ebcd962d37a2d0f7e9ad95a81b284fbf905/cytoolz-1.1.0-cp314-cp314t-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:36c8ede93525cf11e2cc787b7156e5cecd7340193ef800b816a16f1404a8dc6d", size = 3001173 }, - { url = "https://files.pythonhosted.org/packages/7a/cb/f59d83a5058e1198db5a1f04e4a124c94d60390e4fa89b6d2e38ee8288a0/cytoolz-1.1.0-cp314-cp314t-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl", hash = "sha256:c0c949755b6d8a649c5fbc888bc30915926f1b09fe42fea9f289e297c2f6ddd3", size = 2701374 }, - { url = "https://files.pythonhosted.org/packages/b7/f0/1ae6d28df503b0bdae094879da2072b8ba13db5919cd3798918761578411/cytoolz-1.1.0-cp314-cp314t-musllinux_1_2_aarch64.whl", hash = "sha256:e1b6d37545816905a76d9ed59fa4e332f929e879f062a39ea0f6f620405cdc27", size = 2953081 }, - { url = "https://files.pythonhosted.org/packages/f4/06/d86fe811c6222dc32d3e08f5d88d2be598a6055b4d0590e7c1428d55c386/cytoolz-1.1.0-cp314-cp314t-musllinux_1_2_armv7l.whl", hash = "sha256:05332112d4087904842b36954cd1d3fc0e463a2f4a7ef9477bd241427c593c3b", size = 2862228 }, - { url = "https://files.pythonhosted.org/packages/ae/32/978ef6f42623be44a0a03ae9de875ab54aa26c7e38c5c4cd505460b0927d/cytoolz-1.1.0-cp314-cp314t-musllinux_1_2_i686.whl", hash = "sha256:31538ca2fad2d688cbd962ccc3f1da847329e2258a52940f10a2ac0719e526be", size = 2861971 }, - { url = "https://files.pythonhosted.org/packages/ee/f7/74c69497e756b752b359925d1feef68b91df024a4124a823740f675dacd3/cytoolz-1.1.0-cp314-cp314t-musllinux_1_2_ppc64le.whl", hash = "sha256:747562aa70abf219ea16f07d50ac0157db856d447f7f498f592e097cbc77df0b", size = 2975304 }, - { url = "https://files.pythonhosted.org/packages/5b/2b/3ce0e6889a6491f3418ad4d84ae407b8456b02169a5a1f87990dbba7433b/cytoolz-1.1.0-cp314-cp314t-musllinux_1_2_riscv64.whl", hash = "sha256:3dc15c48b20c0f467e15e341e102896c8422dccf8efc6322def5c1b02f074629", size = 2697371 }, - { url = "https://files.pythonhosted.org/packages/15/87/c616577f0891d97860643c845f7221e95240aa589586de727e28a5eb6e52/cytoolz-1.1.0-cp314-cp314t-musllinux_1_2_s390x.whl", hash = "sha256:3c03137ee6103ba92d5d6ad6a510e86fded69cd67050bd8a1843f15283be17ac", size = 2992436 }, - { url = "https://files.pythonhosted.org/packages/e7/9f/490c81bffb3428ab1fa114051fbb5ba18aaa2e2fe4da5bf4170ca524e6b3/cytoolz-1.1.0-cp314-cp314t-musllinux_1_2_x86_64.whl", hash = "sha256:be8e298d88f88bd172b59912240558be3b7a04959375646e7fd4996401452941", size = 2917612 }, - { url = "https://files.pythonhosted.org/packages/66/35/0fec2769660ca6472bbf3317ab634675827bb706d193e3240aaf20eab961/cytoolz-1.1.0-cp314-cp314t-win32.whl", hash = "sha256:3d407140f5604a89578285d4aac7b18b8eafa055cf776e781aabb89c48738fad", size = 960842 }, - { url = "https://files.pythonhosted.org/packages/46/b4/b7ce3d3cd20337becfec978ecfa6d0ef64884d0cf32d44edfed8700914b9/cytoolz-1.1.0-cp314-cp314t-win_amd64.whl", hash = "sha256:56e5afb69eb6e1b3ffc34716ee5f92ffbdb5cb003b3a5ca4d4b0fe700e217162", size = 1020835 }, - { url = "https://files.pythonhosted.org/packages/2c/1f/0498009aa563a9c5d04f520aadc6e1c0942434d089d0b2f51ea986470f55/cytoolz-1.1.0-cp314-cp314t-win_arm64.whl", hash = "sha256:27b19b4a286b3ff52040efa42dbe403730aebe5fdfd2def704eb285e2125c63e", size = 927963 }, - { url = "https://files.pythonhosted.org/packages/84/32/0522207170294cf691112a93c70a8ef942f60fa9ff8e793b63b1f09cedc0/cytoolz-1.1.0-pp311-pypy311_pp73-macosx_10_15_x86_64.whl", hash = "sha256:f32e93a55681d782fc6af939f6df36509d65122423cbc930be39b141064adff8", size = 922014 }, - { url = "https://files.pythonhosted.org/packages/4c/49/9be2d24adaa18fa307ff14e3e43f02b2ae4b69c4ce51cee6889eb2114990/cytoolz-1.1.0-pp311-pypy311_pp73-macosx_11_0_arm64.whl", hash = "sha256:5d9bc596751cbda8073e65be02ca11706f00029768fbbbc81e11a8c290bb41aa", size = 918134 }, - { url = "https://files.pythonhosted.org/packages/5c/b3/6a76c3b94c6c87c72ea822e7e67405be6b649c2e37778eeac7c0c0c69de8/cytoolz-1.1.0-pp311-pypy311_pp73-manylinux1_i686.manylinux_2_28_i686.manylinux_2_5_i686.whl", hash = "sha256:9b16660d01c3931951fab49db422c627897c38c1a1f0393a97582004019a4887", size = 981970 }, - { url = "https://files.pythonhosted.org/packages/f6/8a/606e4c7ed14aa6a86aee6ca84a2cb804754dc6c4905b8f94e09e49f1ce60/cytoolz-1.1.0-pp311-pypy311_pp73-manylinux1_x86_64.manylinux_2_28_x86_64.manylinux_2_5_x86_64.whl", hash = "sha256:b7de5718e2113d4efccea3f06055758cdbc17388ecc3341ba4d1d812837d7c1a", size = 978877 }, - { url = "https://files.pythonhosted.org/packages/97/ec/ad474dcb1f6c1ebfdda3c2ad2edbb1af122a0e79c9ff2cb901ffb5f59662/cytoolz-1.1.0-pp311-pypy311_pp73-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:a12a2a1a6bc44099491c05a12039efa08cc33a3d0f8c7b0566185e085e139283", size = 964279 }, - { url = "https://files.pythonhosted.org/packages/68/8c/d245fd416c69d27d51f14d5ad62acc4ee5971088ee31c40ffe1cc109af68/cytoolz-1.1.0-pp311-pypy311_pp73-win_amd64.whl", hash = "sha256:047defa7f5f9a32f82373dbc3957289562e8a3fa58ae02ec8e4dca4f43a33a21", size = 916630 }, -] - -[[package]] -name = "datasets" -version = "4.4.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "dill" }, - { name = "filelock" }, - { name = "fsspec", extra = ["http"] }, - { name = "httpx" }, - { name = "huggingface-hub" }, - { name = "multiprocess" }, - { name = "numpy" }, - { name = "packaging" }, - { name = "pandas" }, - { name = "pyarrow" }, - { name = "pyyaml" }, - { name = "requests" }, - { name = "tqdm" }, - { name = "xxhash" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/c4/54/9359803da96bc65439a28fbb014dc2c90b7d4d8034a93b72362b0d40191f/datasets-4.4.2.tar.gz", hash = "sha256:9de16e415c4ba4713eac0493f7c7dc74f3aa21599297f00cc6ddab409cb7b24b", size = 586474 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/7b/b5/fefa518c809de7bced5cddb7c21c010da66fa2ae494bda96844a280cc6ce/datasets-4.4.2-py3-none-any.whl", hash = "sha256:6f5ef3417504d9cd663c71c1b90b9a494ff4c2076a2cd6a6e40ceee6ad95befc", size = 512268 }, -] - -[[package]] -name = "decorator" -version = "5.2.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/43/fa/6d96a0978d19e17b68d634497769987b16c8f4cd0a7a05048bec693caa6b/decorator-5.2.1.tar.gz", hash = "sha256:65f266143752f734b0a7cc83c46f4618af75b8c5911b00ccb61d0ac9b6da0360", size = 56711 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/4e/8c/f3147f5c4b73e7550fe5f9352eaa956ae838d5c51eb58e7a25b9f3e2643b/decorator-5.2.1-py3-none-any.whl", hash = "sha256:d316bb415a2d9e2d2b3abcc4084c6502fc09240e292cd76a76afc106a1c8e04a", size = 9190 }, -] - -[[package]] -name = "dill" -version = "0.4.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/12/80/630b4b88364e9a8c8c5797f4602d0f76ef820909ee32f0bacb9f90654042/dill-0.4.0.tar.gz", hash = "sha256:0633f1d2df477324f53a895b02c901fb961bdbf65a17122586ea7019292cbcf0", size = 186976 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/50/3d/9373ad9c56321fdab5b41197068e1d8c25883b3fea29dd361f9b55116869/dill-0.4.0-py3-none-any.whl", hash = "sha256:44f54bf6412c2c8464c14e8243eb163690a9800dbe2c367330883b19c7561049", size = 119668 }, -] - -[[package]] -name = "docopt" -version = "0.6.2" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/a2/55/8f8cab2afd404cf578136ef2cc5dfb50baa1761b68c9da1fb1e4eed343c9/docopt-0.6.2.tar.gz", hash = "sha256:49b3a825280bd66b3aa83585ef59c4a8c82f2c8a522dbe754a8bc8d08c85c491", size = 25901 } - -[[package]] -name = "editdistance" -version = "0.8.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/d5/18/9f4f975ca87a390832b1c22478f3702fcdf739f83211e24d054b7551270d/editdistance-0.8.1.tar.gz", hash = "sha256:d1cdf80a5d5014b0c9126a69a42ce55a457b457f6986ff69ca98e4fe4d2d8fed", size = 50006 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/09/c9/302658ce7f4c537a4e85cf578d11bbf7af120a712e1d78fedc6cb8823c65/editdistance-0.8.1-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:adeb705f32b93accc74960d227875abff150ee42d676e428536361fe5f8f5388", size = 106150 }, - { url = "https://files.pythonhosted.org/packages/45/80/0b3c7d2c0e183725986fea5dd2df11f0b4b46320e9a64f6077a121ab1f64/editdistance-0.8.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:3de77951b105d0972deec7684a0b3d1a9dee69c9b5d34f6e2acc0d76cd4a1c52", size = 80551 }, - { url = "https://files.pythonhosted.org/packages/b5/14/681460965c6a4a48321b07f88de2273d097fdca0491ff55db891aacbd291/editdistance-0.8.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:5e88efb052d45e924606c305cb833a80579dca3e8e4ff01309d50ba2c1c0bbd5", size = 79142 }, - { url = "https://files.pythonhosted.org/packages/ed/0d/abdbc8e394a9461cf2ae27c16564fadaa65f52bd242dd1582ae5e7736dc3/editdistance-0.8.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0247e7a1e9c66ea75211a97e725366bff19a52aac2c838ed5f90025630e976dd", size = 396768 }, - { url = "https://files.pythonhosted.org/packages/c2/fb/2940d26ebda12efd280ae939436f17ac482930d862df9e774cb8b771ab03/editdistance-0.8.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:67d143429a49ab552411505f550a0fb4285a1d4336e096804d233ec495ac20fc", size = 401846 }, - { url = "https://files.pythonhosted.org/packages/53/cc/c63d75c7f387d4df0645682c1ab8706c2dfe5c9c0c4999723ce9a3ba0853/editdistance-0.8.1-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ca9d3be2b10e5d44a950a4bd1e84bca9ebbecd364bce0cf5693bf8224c78eaef", size = 397543 }, - { url = "https://files.pythonhosted.org/packages/8e/38/bb0f734a7571e093184606b930734b12da5b6bff2635eba9312fe4536dd9/editdistance-0.8.1-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:5c72aa1df8535f2e2b3d8773a1a7da091bc1a7e52bb396e7e48d375ba687e7b2", size = 898934 }, - { url = "https://files.pythonhosted.org/packages/1c/9f/624fc7a09918f850a057465f02e86f269e139a457f48ff8cabfb12701756/editdistance-0.8.1-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:9a606c34a2a6cc190e4fffc856b36333cdcf1f1fab5b22bd3088e585c22d6ca0", size = 959637 }, - { url = "https://files.pythonhosted.org/packages/5e/5c/7fa6cc277f91c477ee370807d51c1826891dc6dfc307544223ce7f2687de/editdistance-0.8.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:5af173d442ffac33b7c7990132f97f88818a3abf4b21c0c702a7022df37c0c5c", size = 911024 }, - { url = "https://files.pythonhosted.org/packages/ad/97/556215f71184291155aee340a6d34f0676e7238fdfd10615b6b775ce25fe/editdistance-0.8.1-cp310-cp310-win32.whl", hash = "sha256:fd64b58f5a7b59afd9d75982aaeeacd2a98498bf472fa0360c122ffe6ea4c871", size = 80834 }, - { url = "https://files.pythonhosted.org/packages/c8/d1/7ec5f5cbb95838d0eff7f980a660c81acd1363d658f2f5d4ceba38877c5a/editdistance-0.8.1-cp310-cp310-win_amd64.whl", hash = "sha256:6c7c62c3cae45ca1fa01bb2722b297b9de1e3a244ac44cfba88bdcb488fe6aee", size = 79614 }, - { url = "https://files.pythonhosted.org/packages/e2/dc/d0c29fd52d8f9e795653ed2b838a2a48c739cdfff04ac5b79c6c0ecbdf79/editdistance-0.8.1-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:486105603a273d73d12a54f347dffa70ab281749d7c3879658b377bc49e4b98c", size = 106079 }, - { url = "https://files.pythonhosted.org/packages/b4/c6/75fa45d7b78fbea6fd894f4e48895a75bd3c83d4a9a6b57673881d74d3e0/editdistance-0.8.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:fad081f5f86a175c1a09a4e9e45b95c9349e454c21e181e842e01c85f1f536fc", size = 80580 }, - { url = "https://files.pythonhosted.org/packages/b7/a3/058d823b6285c3511dc94ed80620c3fb0c18b4aaa708f70ba71f3af28436/editdistance-0.8.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:8cb78e125f6759398885a775f5eed07c2bb72b2f86da43e674c6b6a3335b273b", size = 79087 }, - { url = "https://files.pythonhosted.org/packages/a0/3a/0b13c7864c93b1e9b9952bd2a33c5ef3c4fd1bf70a5fad6924789e70e5eb/editdistance-0.8.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3778ca60aa89def9144b70e330bcec5330c7da1d69cb28c612e90b84510a1d3d", size = 409296 }, - { url = "https://files.pythonhosted.org/packages/96/8a/db0fd79e8ddb9b5f86f274107c5d0a27ec4f2af88877df1f26c2c6d150cc/editdistance-0.8.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:fba945eaa0436cf40bc53d7e299dc537c7c71353379a095b7459ff4af910da33", size = 412913 }, - { url = "https://files.pythonhosted.org/packages/0d/d2/98be7112750ff17b436dd76f988f1e38570dcec0df8578ee19ef046f22fe/editdistance-0.8.1-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:877f2a0d801f32bc1a1878901ffb947b974361e849c66e314a7f1d786a446b58", size = 407430 }, - { url = "https://files.pythonhosted.org/packages/03/62/1815e3bf164910c47ba1948c8b5e937a40c7f9763b64e98fb6666b01dd06/editdistance-0.8.1-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:e79d351ca40a6ead5f3763253fd7521572ee0d3e5d42538630e56d10f48db481", size = 909217 }, - { url = "https://files.pythonhosted.org/packages/0c/d3/a832cea7b507a9be54e4ac3d1340fb66dca5f9c16c70bf38d5039e8fdede/editdistance-0.8.1-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:70ed382b3052a51161bad0149d4665003bf3b949fce0b01bf1253a4cc1a88239", size = 969407 }, - { url = "https://files.pythonhosted.org/packages/a3/b4/db291d2a3845cbf8047b4b5aad3b3e038a8a2994d87027b40e1a1b0f4b74/editdistance-0.8.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:a529bfb384c4000775d76739c4e64f73337f0f5a3784933b1321b577a62bed4e", size = 922112 }, - { url = "https://files.pythonhosted.org/packages/c4/26/7ddeacada4982d0b892a28897e21871d0f25bca165e3663e37c3a272808a/editdistance-0.8.1-cp311-cp311-win32.whl", hash = "sha256:b082232429e731f181af7f7d2bcf79da6ca8fadd04e9086c11e2973f7d330c81", size = 80799 }, - { url = "https://files.pythonhosted.org/packages/52/a1/778af8590b8b12f03f62eacc3c8744407ade9e3d69be6dabe38d0afbf2dd/editdistance-0.8.1-cp311-cp311-win_amd64.whl", hash = "sha256:cef1a4359252a49f2c4718e64e9d40027d9d951b289d045bdb278656e59f6af8", size = 79698 }, - { url = "https://files.pythonhosted.org/packages/cb/4c/7f195588949b4e72436dc7fc902632381f96e586af829685b56daebb38b8/editdistance-0.8.1-cp312-cp312-macosx_10_9_universal2.whl", hash = "sha256:b04af61b3fcdd287a07c15b6ae3b02af01c5e3e9c3aca76b8c1d13bd266b6f57", size = 106723 }, - { url = "https://files.pythonhosted.org/packages/8d/82/31dc1640d830cd7d36865098329f34e4dad3b77f31cfb9404b347e700196/editdistance-0.8.1-cp312-cp312-macosx_10_9_x86_64.whl", hash = "sha256:18fc8b6eaae01bfd9cf999af726c1e8dcf667d120e81aa7dbd515bea7427f62f", size = 80998 }, - { url = "https://files.pythonhosted.org/packages/ea/2a/6b823e71cef694d6f070a1d82be2842706fa193541aab8856a8f42044cd0/editdistance-0.8.1-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:6a87839450a5987028738d061ffa5ef6a68bac2ddc68c9147a8aae9806629c7f", size = 79248 }, - { url = "https://files.pythonhosted.org/packages/e1/31/bfb8e590f922089dc3471ed7828a6da2fc9453eba38c332efa9ee8749fd7/editdistance-0.8.1-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:24b5f9c9673c823d91b5973d0af8b39f883f414a55ade2b9d097138acd10f31e", size = 415262 }, - { url = "https://files.pythonhosted.org/packages/a9/c7/57423942b2f847cdbbb46494568d00cd8a45500904ea026f0aad6ca01bc7/editdistance-0.8.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c59248eabfad603f0fba47b0c263d5dc728fb01c2b6b50fb6ca187cec547fdb3", size = 418905 }, - { url = "https://files.pythonhosted.org/packages/1b/05/dfa4cdcce063596cbf0d7a32c46cd0f4fa70980311b7da64d35f33ad02a0/editdistance-0.8.1-cp312-cp312-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:84e239d88ff52821cf64023fabd06a1d9a07654f364b64bf1284577fd3a79d0e", size = 412511 }, - { url = "https://files.pythonhosted.org/packages/0e/14/39608ff724a9523f187c4e28926d78bc68f2798f74777ac6757981108345/editdistance-0.8.1-cp312-cp312-musllinux_1_1_aarch64.whl", hash = "sha256:2f7f71698f83e8c83839ac0d876a0f4ef996c86c5460aebd26d85568d4afd0db", size = 917293 }, - { url = "https://files.pythonhosted.org/packages/df/92/4a1c61d72da40dedfd0ff950fdc71ae83f478330c58a8bccfd776518bd67/editdistance-0.8.1-cp312-cp312-musllinux_1_1_i686.whl", hash = "sha256:04e229d6f4ce0c12abc9f4cd4023a5b5fa9620226e0207b119c3c2778b036250", size = 975580 }, - { url = "https://files.pythonhosted.org/packages/47/3d/9877566e724c8a37f2228a84ec5cbf66dbfd0673515baf68a0fe07caff40/editdistance-0.8.1-cp312-cp312-musllinux_1_1_x86_64.whl", hash = "sha256:e16721636da6d6b68a2c09eaced35a94f4a4a704ec09f45756d4fd5e128ed18d", size = 929121 }, - { url = "https://files.pythonhosted.org/packages/d2/f5/8c50757d198b8ca30ddb91e8b8f0247a8dca04ff2ec30755245f0ab1ff0c/editdistance-0.8.1-cp312-cp312-win32.whl", hash = "sha256:87533cf2ebc3777088d991947274cd7e1014b9c861a8aa65257bcdc0ee492526", size = 81039 }, - { url = "https://files.pythonhosted.org/packages/28/f0/65101e51dc7c850e7b7581a5d8fa8721a1d7479a0dca6c08386328e19882/editdistance-0.8.1-cp312-cp312-win_amd64.whl", hash = "sha256:09f01ed51746d90178af7dd7ea4ebb41497ef19f53c7f327e864421743dffb0a", size = 79853 }, - { url = "https://files.pythonhosted.org/packages/d4/4c/c9d02eeb47815d35f8d324b52f6704ea7beb032bcb209358cac44047d413/editdistance-0.8.1-pp310-pypy310_pp73-macosx_10_9_x86_64.whl", hash = "sha256:a4a90c6b03094c07358572027a8d0a13cca7450b1aa6caca98a5f1fa4f0b8961", size = 76455 }, - { url = "https://files.pythonhosted.org/packages/af/b0/2818fa6a24595dac069b0bfb9d05658406779a1ded8fd2b0c9066396cf99/editdistance-0.8.1-pp310-pypy310_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:510a4f9ced348a4fd89ae2e102357d4d801a771e29bb2bc2f130a1692193407f", size = 84104 }, - { url = "https://files.pythonhosted.org/packages/1f/d1/3d5e09bcf7fdb7aed705bf74047a8634bd2b8fd92177c25a2547e6dbadfb/editdistance-0.8.1-pp310-pypy310_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:4787fa7228ba6a34b430066d174320f011d605015baa7299c2c4911e6ea6bd46", size = 89058 }, - { url = "https://files.pythonhosted.org/packages/cd/88/fca5d7b1a1edf66ce1e5b6b60bff75842e6814b4f5facbdf4585d88c912d/editdistance-0.8.1-pp310-pypy310_pp73-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ee02601375073afccd6b4d811129ce1cb696d47db734784d8dbd1fddcea75447", size = 84635 }, - { url = "https://files.pythonhosted.org/packages/a9/91/0e6285bbe2358d81fd16313d30306b2d0036387348f7bc11d8c076ca3c72/editdistance-0.8.1-pp310-pypy310_pp73-win_amd64.whl", hash = "sha256:bc7ad9f9a20e6f351523de77c59249f005242e3f317b5de45d02c378d24f6531", size = 77389 }, -] - -[[package]] -name = "einops" -version = "0.8.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/e5/81/df4fbe24dff8ba3934af99044188e20a98ed441ad17a274539b74e82e126/einops-0.8.1.tar.gz", hash = "sha256:de5d960a7a761225532e0f1959e5315ebeafc0cd43394732f103ca44b9837e84", size = 54805 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/87/62/9773de14fe6c45c23649e98b83231fffd7b9892b6cf863251dc2afa73643/einops-0.8.1-py3-none-any.whl", hash = "sha256:919387eb55330f5757c6bea9165c5ff5cfe63a642682ea788a6d472576d81737", size = 64359 }, -] - -[[package]] -name = "exceptiongroup" -version = "1.3.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "typing-extensions", marker = "python_full_version < '3.11'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/50/79/66800aadf48771f6b62f7eb014e352e5d06856655206165d775e675a02c9/exceptiongroup-1.3.1.tar.gz", hash = "sha256:8b412432c6055b0b7d14c310000ae93352ed6754f70fa8f7c34141f91c4e3219", size = 30371 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/8a/0e/97c33bf5009bdbac74fd2beace167cab3f978feb69cc36f1ef79360d6c4e/exceptiongroup-1.3.1-py3-none-any.whl", hash = "sha256:a7a39a3bd276781e98394987d3a5701d0c4edffb633bb7a5144577f82c773598", size = 16740 }, -] - -[[package]] -name = "executing" -version = "2.2.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/cc/28/c14e053b6762b1044f34a13aab6859bbf40456d37d23aa286ac24cfd9a5d/executing-2.2.1.tar.gz", hash = "sha256:3632cc370565f6648cc328b32435bd120a1e4ebb20c77e3fdde9a13cd1e533c4", size = 1129488 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/c1/ea/53f2148663b321f21b5a606bd5f191517cf40b7072c0497d3c92c4a13b1e/executing-2.2.1-py2.py3-none-any.whl", hash = "sha256:760643d3452b4d777d295bb167ccc74c64a81df23fb5e08eff250c425a4b2017", size = 28317 }, -] - -[[package]] -name = "fiddle" -version = "0.3.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "absl-py" }, - { name = "graphviz" }, - { name = "libcst" }, - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/73/36/7a4fac76351619b36bbc7937abf59f7b601326dc4efc253b3c16819f782a/fiddle-0.3.0.tar.gz", hash = "sha256:5d083d3299a479868345513385a6c5546141bd92086c15d3dcbf8008a90075d3", size = 277884 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/3b/98/a38e949a91ff9e15874487fd8329ff53c25f3413c0cfc809eb6ff7eb7fa1/fiddle-0.3.0-py3-none-any.whl", hash = "sha256:f4824541c103a94a2f33f6c93eeddf6007c3a7300440087a95907f3e74362e61", size = 419830 }, -] - -[[package]] -name = "filelock" -version = "3.20.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/a7/23/ce7a1126827cedeb958fc043d61745754464eb56c5937c35bbf2b8e26f34/filelock-3.20.1.tar.gz", hash = "sha256:b8360948b351b80f420878d8516519a2204b07aefcdcfd24912a5d33127f188c", size = 19476 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/e3/7f/a1a97644e39e7316d850784c642093c99df1290a460df4ede27659056834/filelock-3.20.1-py3-none-any.whl", hash = "sha256:15d9e9a67306188a44baa72f569d2bfd803076269365fdea0934385da4dc361a", size = 16666 }, -] - -[[package]] -name = "fonttools" -version = "4.61.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/ec/ca/cf17b88a8df95691275a3d77dc0a5ad9907f328ae53acbe6795da1b2f5ed/fonttools-4.61.1.tar.gz", hash = "sha256:6675329885c44657f826ef01d9e4fb33b9158e9d93c537d84ad8399539bc6f69", size = 3565756 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/5b/94/8a28707adb00bed1bf22dac16ccafe60faf2ade353dcb32c3617ee917307/fonttools-4.61.1-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:7c7db70d57e5e1089a274cbb2b1fd635c9a24de809a231b154965d415d6c6d24", size = 2854799 }, - { url = "https://files.pythonhosted.org/packages/94/93/c2e682faaa5ee92034818d8f8a8145ae73eb83619600495dcf8503fa7771/fonttools-4.61.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:5fe9fd43882620017add5eabb781ebfbc6998ee49b35bd7f8f79af1f9f99a958", size = 2403032 }, - { url = "https://files.pythonhosted.org/packages/f1/62/1748f7e7e1ee41aa52279fd2e3a6d0733dc42a673b16932bad8e5d0c8b28/fonttools-4.61.1-cp310-cp310-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:d8db08051fc9e7d8bc622f2112511b8107d8f27cd89e2f64ec45e9825e8288da", size = 4897863 }, - { url = "https://files.pythonhosted.org/packages/69/69/4ca02ee367d2c98edcaeb83fc278d20972502ee071214ad9d8ca85e06080/fonttools-4.61.1-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:a76d4cb80f41ba94a6691264be76435e5f72f2cb3cab0b092a6212855f71c2f6", size = 4859076 }, - { url = "https://files.pythonhosted.org/packages/8c/f5/660f9e3cefa078861a7f099107c6d203b568a6227eef163dd173bfc56bdc/fonttools-4.61.1-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:a13fc8aeb24bad755eea8f7f9d409438eb94e82cf86b08fe77a03fbc8f6a96b1", size = 4875623 }, - { url = "https://files.pythonhosted.org/packages/63/d1/9d7c5091d2276ed47795c131c1bf9316c3c1ab2789c22e2f59e0572ccd38/fonttools-4.61.1-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:b846a1fcf8beadeb9ea4f44ec5bdde393e2f1569e17d700bfc49cd69bde75881", size = 4993327 }, - { url = "https://files.pythonhosted.org/packages/6f/2d/28def73837885ae32260d07660a052b99f0aa00454867d33745dfe49dbf0/fonttools-4.61.1-cp310-cp310-win32.whl", hash = "sha256:78a7d3ab09dc47ac1a363a493e6112d8cabed7ba7caad5f54dbe2f08676d1b47", size = 1502180 }, - { url = "https://files.pythonhosted.org/packages/63/fa/bfdc98abb4dd2bd491033e85e3ba69a2313c850e759a6daa014bc9433b0f/fonttools-4.61.1-cp310-cp310-win_amd64.whl", hash = "sha256:eff1ac3cc66c2ac7cda1e64b4e2f3ffef474b7335f92fc3833fc632d595fcee6", size = 1550654 }, - { url = "https://files.pythonhosted.org/packages/69/12/bf9f4eaa2fad039356cc627587e30ed008c03f1cebd3034376b5ee8d1d44/fonttools-4.61.1-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:c6604b735bb12fef8e0efd5578c9fb5d3d8532d5001ea13a19cddf295673ee09", size = 2852213 }, - { url = "https://files.pythonhosted.org/packages/ac/49/4138d1acb6261499bedde1c07f8c2605d1d8f9d77a151e5507fd3ef084b6/fonttools-4.61.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:5ce02f38a754f207f2f06557523cd39a06438ba3aafc0639c477ac409fc64e37", size = 2401689 }, - { url = "https://files.pythonhosted.org/packages/e5/fe/e6ce0fe20a40e03aef906af60aa87668696f9e4802fa283627d0b5ed777f/fonttools-4.61.1-cp311-cp311-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:77efb033d8d7ff233385f30c62c7c79271c8885d5c9657d967ede124671bbdfb", size = 5058809 }, - { url = "https://files.pythonhosted.org/packages/79/61/1ca198af22f7dd22c17ab86e9024ed3c06299cfdb08170640e9996d501a0/fonttools-4.61.1-cp311-cp311-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:75c1a6dfac6abd407634420c93864a1e274ebc1c7531346d9254c0d8f6ca00f9", size = 5036039 }, - { url = "https://files.pythonhosted.org/packages/99/cc/fa1801e408586b5fce4da9f5455af8d770f4fc57391cd5da7256bb364d38/fonttools-4.61.1-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:0de30bfe7745c0d1ffa2b0b7048fb7123ad0d71107e10ee090fa0b16b9452e87", size = 5034714 }, - { url = "https://files.pythonhosted.org/packages/bf/aa/b7aeafe65adb1b0a925f8f25725e09f078c635bc22754f3fecb7456955b0/fonttools-4.61.1-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:58b0ee0ab5b1fc9921eccfe11d1435added19d6494dde14e323f25ad2bc30c56", size = 5158648 }, - { url = "https://files.pythonhosted.org/packages/99/f9/08ea7a38663328881384c6e7777bbefc46fd7d282adfd87a7d2b84ec9d50/fonttools-4.61.1-cp311-cp311-win32.whl", hash = "sha256:f79b168428351d11e10c5aeb61a74e1851ec221081299f4cf56036a95431c43a", size = 2280681 }, - { url = "https://files.pythonhosted.org/packages/07/ad/37dd1ae5fa6e01612a1fbb954f0927681f282925a86e86198ccd7b15d515/fonttools-4.61.1-cp311-cp311-win_amd64.whl", hash = "sha256:fe2efccb324948a11dd09d22136fe2ac8a97d6c1347cf0b58a911dcd529f66b7", size = 2331951 }, - { url = "https://files.pythonhosted.org/packages/6f/16/7decaa24a1bd3a70c607b2e29f0adc6159f36a7e40eaba59846414765fd4/fonttools-4.61.1-cp312-cp312-macosx_10_13_universal2.whl", hash = "sha256:f3cb4a569029b9f291f88aafc927dd53683757e640081ca8c412781ea144565e", size = 2851593 }, - { url = "https://files.pythonhosted.org/packages/94/98/3c4cb97c64713a8cf499b3245c3bf9a2b8fd16a3e375feff2aed78f96259/fonttools-4.61.1-cp312-cp312-macosx_10_13_x86_64.whl", hash = "sha256:41a7170d042e8c0024703ed13b71893519a1a6d6e18e933e3ec7507a2c26a4b2", size = 2400231 }, - { url = "https://files.pythonhosted.org/packages/b7/37/82dbef0f6342eb01f54bca073ac1498433d6ce71e50c3c3282b655733b31/fonttools-4.61.1-cp312-cp312-manylinux1_x86_64.manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_5_x86_64.whl", hash = "sha256:10d88e55330e092940584774ee5e8a6971b01fc2f4d3466a1d6c158230880796", size = 4954103 }, - { url = "https://files.pythonhosted.org/packages/6c/44/f3aeac0fa98e7ad527f479e161aca6c3a1e47bb6996b053d45226fe37bf2/fonttools-4.61.1-cp312-cp312-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:15acc09befd16a0fb8a8f62bc147e1a82817542d72184acca9ce6e0aeda9fa6d", size = 5004295 }, - { url = "https://files.pythonhosted.org/packages/14/e8/7424ced75473983b964d09f6747fa09f054a6d656f60e9ac9324cf40c743/fonttools-4.61.1-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:e6bcdf33aec38d16508ce61fd81838f24c83c90a1d1b8c68982857038673d6b8", size = 4944109 }, - { url = "https://files.pythonhosted.org/packages/c8/8b/6391b257fa3d0b553d73e778f953a2f0154292a7a7a085e2374b111e5410/fonttools-4.61.1-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:5fade934607a523614726119164ff621e8c30e8fa1ffffbbd358662056ba69f0", size = 5093598 }, - { url = "https://files.pythonhosted.org/packages/d9/71/fd2ea96cdc512d92da5678a1c98c267ddd4d8c5130b76d0f7a80f9a9fde8/fonttools-4.61.1-cp312-cp312-win32.whl", hash = "sha256:75da8f28eff26defba42c52986de97b22106cb8f26515b7c22443ebc9c2d3261", size = 2269060 }, - { url = "https://files.pythonhosted.org/packages/80/3b/a3e81b71aed5a688e89dfe0e2694b26b78c7d7f39a5ffd8a7d75f54a12a8/fonttools-4.61.1-cp312-cp312-win_amd64.whl", hash = "sha256:497c31ce314219888c0e2fce5ad9178ca83fe5230b01a5006726cdf3ac9f24d9", size = 2319078 }, - { url = "https://files.pythonhosted.org/packages/4b/cf/00ba28b0990982530addb8dc3e9e6f2fa9cb5c20df2abdda7baa755e8fe1/fonttools-4.61.1-cp313-cp313-macosx_10_13_universal2.whl", hash = "sha256:8c56c488ab471628ff3bfa80964372fc13504ece601e0d97a78ee74126b2045c", size = 2846454 }, - { url = "https://files.pythonhosted.org/packages/5a/ca/468c9a8446a2103ae645d14fee3f610567b7042aba85031c1c65e3ef7471/fonttools-4.61.1-cp313-cp313-macosx_10_13_x86_64.whl", hash = "sha256:dc492779501fa723b04d0ab1f5be046797fee17d27700476edc7ee9ae535a61e", size = 2398191 }, - { url = "https://files.pythonhosted.org/packages/a3/4b/d67eedaed19def5967fade3297fed8161b25ba94699efc124b14fb68cdbc/fonttools-4.61.1-cp313-cp313-manylinux1_x86_64.manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_5_x86_64.whl", hash = "sha256:64102ca87e84261419c3747a0d20f396eb024bdbeb04c2bfb37e2891f5fadcb5", size = 4928410 }, - { url = "https://files.pythonhosted.org/packages/b0/8d/6fb3494dfe61a46258cd93d979cf4725ded4eb46c2a4ca35e4490d84daea/fonttools-4.61.1-cp313-cp313-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:4c1b526c8d3f615a7b1867f38a9410849c8f4aef078535742198e942fba0e9bd", size = 4984460 }, - { url = "https://files.pythonhosted.org/packages/f7/f1/a47f1d30b3dc00d75e7af762652d4cbc3dff5c2697a0dbd5203c81afd9c3/fonttools-4.61.1-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:41ed4b5ec103bd306bb68f81dc166e77409e5209443e5773cb4ed837bcc9b0d3", size = 4925800 }, - { url = "https://files.pythonhosted.org/packages/a7/01/e6ae64a0981076e8a66906fab01539799546181e32a37a0257b77e4aa88b/fonttools-4.61.1-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:b501c862d4901792adaec7c25b1ecc749e2662543f68bb194c42ba18d6eec98d", size = 5067859 }, - { url = "https://files.pythonhosted.org/packages/73/aa/28e40b8d6809a9b5075350a86779163f074d2b617c15d22343fce81918db/fonttools-4.61.1-cp313-cp313-win32.whl", hash = "sha256:4d7092bb38c53bbc78e9255a59158b150bcdc115a1e3b3ce0b5f267dc35dd63c", size = 2267821 }, - { url = "https://files.pythonhosted.org/packages/1a/59/453c06d1d83dc0951b69ef692d6b9f1846680342927df54e9a1ca91c6f90/fonttools-4.61.1-cp313-cp313-win_amd64.whl", hash = "sha256:21e7c8d76f62ab13c9472ccf74515ca5b9a761d1bde3265152a6dc58700d895b", size = 2318169 }, - { url = "https://files.pythonhosted.org/packages/32/8f/4e7bf82c0cbb738d3c2206c920ca34ca74ef9dabde779030145d28665104/fonttools-4.61.1-cp314-cp314-macosx_10_15_universal2.whl", hash = "sha256:fff4f534200a04b4a36e7ae3cb74493afe807b517a09e99cb4faa89a34ed6ecd", size = 2846094 }, - { url = "https://files.pythonhosted.org/packages/71/09/d44e45d0a4f3a651f23a1e9d42de43bc643cce2971b19e784cc67d823676/fonttools-4.61.1-cp314-cp314-macosx_10_15_x86_64.whl", hash = "sha256:d9203500f7c63545b4ce3799319fe4d9feb1a1b89b28d3cb5abd11b9dd64147e", size = 2396589 }, - { url = "https://files.pythonhosted.org/packages/89/18/58c64cafcf8eb677a99ef593121f719e6dcbdb7d1c594ae5a10d4997ca8a/fonttools-4.61.1-cp314-cp314-manylinux1_x86_64.manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_5_x86_64.whl", hash = "sha256:fa646ecec9528bef693415c79a86e733c70a4965dd938e9a226b0fc64c9d2e6c", size = 4877892 }, - { url = "https://files.pythonhosted.org/packages/8a/ec/9e6b38c7ba1e09eb51db849d5450f4c05b7e78481f662c3b79dbde6f3d04/fonttools-4.61.1-cp314-cp314-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:11f35ad7805edba3aac1a3710d104592df59f4b957e30108ae0ba6c10b11dd75", size = 4972884 }, - { url = "https://files.pythonhosted.org/packages/5e/87/b5339da8e0256734ba0dbbf5b6cdebb1dd79b01dc8c270989b7bcd465541/fonttools-4.61.1-cp314-cp314-musllinux_1_2_aarch64.whl", hash = "sha256:b931ae8f62db78861b0ff1ac017851764602288575d65b8e8ff1963fed419063", size = 4924405 }, - { url = "https://files.pythonhosted.org/packages/0b/47/e3409f1e1e69c073a3a6fd8cb886eb18c0bae0ee13db2c8d5e7f8495e8b7/fonttools-4.61.1-cp314-cp314-musllinux_1_2_x86_64.whl", hash = "sha256:b148b56f5de675ee16d45e769e69f87623a4944f7443850bf9a9376e628a89d2", size = 5035553 }, - { url = "https://files.pythonhosted.org/packages/bf/b6/1f6600161b1073a984294c6c031e1a56ebf95b6164249eecf30012bb2e38/fonttools-4.61.1-cp314-cp314-win32.whl", hash = "sha256:9b666a475a65f4e839d3d10473fad6d47e0a9db14a2f4a224029c5bfde58ad2c", size = 2271915 }, - { url = "https://files.pythonhosted.org/packages/52/7b/91e7b01e37cc8eb0e1f770d08305b3655e4f002fc160fb82b3390eabacf5/fonttools-4.61.1-cp314-cp314-win_amd64.whl", hash = "sha256:4f5686e1fe5fce75d82d93c47a438a25bf0d1319d2843a926f741140b2b16e0c", size = 2323487 }, - { url = "https://files.pythonhosted.org/packages/39/5c/908ad78e46c61c3e3ed70c3b58ff82ab48437faf84ec84f109592cabbd9f/fonttools-4.61.1-cp314-cp314t-macosx_10_15_universal2.whl", hash = "sha256:e76ce097e3c57c4bcb67c5aa24a0ecdbd9f74ea9219997a707a4061fbe2707aa", size = 2929571 }, - { url = "https://files.pythonhosted.org/packages/bd/41/975804132c6dea64cdbfbaa59f3518a21c137a10cccf962805b301ac6ab2/fonttools-4.61.1-cp314-cp314t-macosx_10_15_x86_64.whl", hash = "sha256:9cfef3ab326780c04d6646f68d4b4742aae222e8b8ea1d627c74e38afcbc9d91", size = 2435317 }, - { url = "https://files.pythonhosted.org/packages/b0/5a/aef2a0a8daf1ebaae4cfd83f84186d4a72ee08fd6a8451289fcd03ffa8a4/fonttools-4.61.1-cp314-cp314t-manylinux1_x86_64.manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_5_x86_64.whl", hash = "sha256:a75c301f96db737e1c5ed5fd7d77d9c34466de16095a266509e13da09751bd19", size = 4882124 }, - { url = "https://files.pythonhosted.org/packages/80/33/d6db3485b645b81cea538c9d1c9219d5805f0877fda18777add4671c5240/fonttools-4.61.1-cp314-cp314t-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:91669ccac46bbc1d09e9273546181919064e8df73488ea087dcac3e2968df9ba", size = 5100391 }, - { url = "https://files.pythonhosted.org/packages/6c/d6/675ba631454043c75fcf76f0ca5463eac8eb0666ea1d7badae5fea001155/fonttools-4.61.1-cp314-cp314t-musllinux_1_2_aarch64.whl", hash = "sha256:c33ab3ca9d3ccd581d58e989d67554e42d8d4ded94ab3ade3508455fe70e65f7", size = 4978800 }, - { url = "https://files.pythonhosted.org/packages/7f/33/d3ec753d547a8d2bdaedd390d4a814e8d5b45a093d558f025c6b990b554c/fonttools-4.61.1-cp314-cp314t-musllinux_1_2_x86_64.whl", hash = "sha256:664c5a68ec406f6b1547946683008576ef8b38275608e1cee6c061828171c118", size = 5006426 }, - { url = "https://files.pythonhosted.org/packages/b4/40/cc11f378b561a67bea850ab50063366a0d1dd3f6d0a30ce0f874b0ad5664/fonttools-4.61.1-cp314-cp314t-win32.whl", hash = "sha256:aed04cabe26f30c1647ef0e8fbb207516fd40fe9472e9439695f5c6998e60ac5", size = 2335377 }, - { url = "https://files.pythonhosted.org/packages/e4/ff/c9a2b66b39f8628531ea58b320d66d951267c98c6a38684daa8f50fb02f8/fonttools-4.61.1-cp314-cp314t-win_amd64.whl", hash = "sha256:2180f14c141d2f0f3da43f3a81bc8aa4684860f6b0e6f9e165a4831f24e6a23b", size = 2400613 }, - { url = "https://files.pythonhosted.org/packages/c7/4e/ce75a57ff3aebf6fc1f4e9d508b8e5810618a33d900ad6c19eb30b290b97/fonttools-4.61.1-py3-none-any.whl", hash = "sha256:17d2bf5d541add43822bcf0c43d7d847b160c9bb01d15d5007d84e2217aaa371", size = 1148996 }, -] - -[[package]] -name = "frozenlist" -version = "1.8.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/2d/f5/c831fac6cc817d26fd54c7eaccd04ef7e0288806943f7cc5bbf69f3ac1f0/frozenlist-1.8.0.tar.gz", hash = "sha256:3ede829ed8d842f6cd48fc7081d7a41001a56f1f38603f9d49bf3020d59a31ad", size = 45875 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/83/4a/557715d5047da48d54e659203b9335be7bfaafda2c3f627b7c47e0b3aaf3/frozenlist-1.8.0-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:b37f6d31b3dcea7deb5e9696e529a6aa4a898adc33db82da12e4c60a7c4d2011", size = 86230 }, - { url = "https://files.pythonhosted.org/packages/a2/fb/c85f9fed3ea8fe8740e5b46a59cc141c23b842eca617da8876cfce5f760e/frozenlist-1.8.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:ef2b7b394f208233e471abc541cc6991f907ffd47dc72584acee3147899d6565", size = 49621 }, - { url = "https://files.pythonhosted.org/packages/63/70/26ca3f06aace16f2352796b08704338d74b6d1a24ca38f2771afbb7ed915/frozenlist-1.8.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:a88f062f072d1589b7b46e951698950e7da00442fc1cacbe17e19e025dc327ad", size = 49889 }, - { url = "https://files.pythonhosted.org/packages/5d/ed/c7895fd2fde7f3ee70d248175f9b6cdf792fb741ab92dc59cd9ef3bd241b/frozenlist-1.8.0-cp310-cp310-manylinux1_x86_64.manylinux_2_28_x86_64.manylinux_2_5_x86_64.whl", hash = "sha256:f57fb59d9f385710aa7060e89410aeb5058b99e62f4d16b08b91986b9a2140c2", size = 219464 }, - { url = "https://files.pythonhosted.org/packages/6b/83/4d587dccbfca74cb8b810472392ad62bfa100bf8108c7223eb4c4fa2f7b3/frozenlist-1.8.0-cp310-cp310-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:799345ab092bee59f01a915620b5d014698547afd011e691a208637312db9186", size = 221649 }, - { url = "https://files.pythonhosted.org/packages/6a/c6/fd3b9cd046ec5fff9dab66831083bc2077006a874a2d3d9247dea93ddf7e/frozenlist-1.8.0-cp310-cp310-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:c23c3ff005322a6e16f71bf8692fcf4d5a304aaafe1e262c98c6d4adc7be863e", size = 219188 }, - { url = "https://files.pythonhosted.org/packages/ce/80/6693f55eb2e085fc8afb28cf611448fb5b90e98e068fa1d1b8d8e66e5c7d/frozenlist-1.8.0-cp310-cp310-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:8a76ea0f0b9dfa06f254ee06053d93a600865b3274358ca48a352ce4f0798450", size = 231748 }, - { url = "https://files.pythonhosted.org/packages/97/d6/e9459f7c5183854abd989ba384fe0cc1a0fb795a83c033f0571ec5933ca4/frozenlist-1.8.0-cp310-cp310-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:c7366fe1418a6133d5aa824ee53d406550110984de7637d65a178010f759c6ef", size = 236351 }, - { url = "https://files.pythonhosted.org/packages/97/92/24e97474b65c0262e9ecd076e826bfd1d3074adcc165a256e42e7b8a7249/frozenlist-1.8.0-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:13d23a45c4cebade99340c4165bd90eeb4a56c6d8a9d8aa49568cac19a6d0dc4", size = 218767 }, - { url = "https://files.pythonhosted.org/packages/ee/bf/dc394a097508f15abff383c5108cb8ad880d1f64a725ed3b90d5c2fbf0bb/frozenlist-1.8.0-cp310-cp310-musllinux_1_2_armv7l.whl", hash = "sha256:e4a3408834f65da56c83528fb52ce7911484f0d1eaf7b761fc66001db1646eff", size = 235887 }, - { url = "https://files.pythonhosted.org/packages/40/90/25b201b9c015dbc999a5baf475a257010471a1fa8c200c843fd4abbee725/frozenlist-1.8.0-cp310-cp310-musllinux_1_2_ppc64le.whl", hash = "sha256:42145cd2748ca39f32801dad54aeea10039da6f86e303659db90db1c4b614c8c", size = 228785 }, - { url = "https://files.pythonhosted.org/packages/84/f4/b5bc148df03082f05d2dd30c089e269acdbe251ac9a9cf4e727b2dbb8a3d/frozenlist-1.8.0-cp310-cp310-musllinux_1_2_s390x.whl", hash = "sha256:e2de870d16a7a53901e41b64ffdf26f2fbb8917b3e6ebf398098d72c5b20bd7f", size = 230312 }, - { url = "https://files.pythonhosted.org/packages/db/4b/87e95b5d15097c302430e647136b7d7ab2398a702390cf4c8601975709e7/frozenlist-1.8.0-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:20e63c9493d33ee48536600d1a5c95eefc870cd71e7ab037763d1fbb89cc51e7", size = 217650 }, - { url = "https://files.pythonhosted.org/packages/e5/70/78a0315d1fea97120591a83e0acd644da638c872f142fd72a6cebee825f3/frozenlist-1.8.0-cp310-cp310-win32.whl", hash = "sha256:adbeebaebae3526afc3c96fad434367cafbfd1b25d72369a9e5858453b1bb71a", size = 39659 }, - { url = "https://files.pythonhosted.org/packages/66/aa/3f04523fb189a00e147e60c5b2205126118f216b0aa908035c45336e27e4/frozenlist-1.8.0-cp310-cp310-win_amd64.whl", hash = "sha256:667c3777ca571e5dbeb76f331562ff98b957431df140b54c85fd4d52eea8d8f6", size = 43837 }, - { url = "https://files.pythonhosted.org/packages/39/75/1135feecdd7c336938bd55b4dc3b0dfc46d85b9be12ef2628574b28de776/frozenlist-1.8.0-cp310-cp310-win_arm64.whl", hash = "sha256:80f85f0a7cc86e7a54c46d99c9e1318ff01f4687c172ede30fd52d19d1da1c8e", size = 39989 }, - { url = "https://files.pythonhosted.org/packages/bc/03/077f869d540370db12165c0aa51640a873fb661d8b315d1d4d67b284d7ac/frozenlist-1.8.0-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:09474e9831bc2b2199fad6da3c14c7b0fbdd377cce9d3d77131be28906cb7d84", size = 86912 }, - { url = "https://files.pythonhosted.org/packages/df/b5/7610b6bd13e4ae77b96ba85abea1c8cb249683217ef09ac9e0ae93f25a91/frozenlist-1.8.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:17c883ab0ab67200b5f964d2b9ed6b00971917d5d8a92df149dc2c9779208ee9", size = 50046 }, - { url = "https://files.pythonhosted.org/packages/6e/ef/0e8f1fe32f8a53dd26bdd1f9347efe0778b0fddf62789ea683f4cc7d787d/frozenlist-1.8.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:fa47e444b8ba08fffd1c18e8cdb9a75db1b6a27f17507522834ad13ed5922b93", size = 50119 }, - { url = "https://files.pythonhosted.org/packages/11/b1/71a477adc7c36e5fb628245dfbdea2166feae310757dea848d02bd0689fd/frozenlist-1.8.0-cp311-cp311-manylinux1_x86_64.manylinux_2_28_x86_64.manylinux_2_5_x86_64.whl", hash = "sha256:2552f44204b744fba866e573be4c1f9048d6a324dfe14475103fd51613eb1d1f", size = 231067 }, - { url = "https://files.pythonhosted.org/packages/45/7e/afe40eca3a2dc19b9904c0f5d7edfe82b5304cb831391edec0ac04af94c2/frozenlist-1.8.0-cp311-cp311-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:957e7c38f250991e48a9a73e6423db1bb9dd14e722a10f6b8bb8e16a0f55f695", size = 233160 }, - { url = "https://files.pythonhosted.org/packages/a6/aa/7416eac95603ce428679d273255ffc7c998d4132cfae200103f164b108aa/frozenlist-1.8.0-cp311-cp311-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:8585e3bb2cdea02fc88ffa245069c36555557ad3609e83be0ec71f54fd4abb52", size = 228544 }, - { url = "https://files.pythonhosted.org/packages/8b/3d/2a2d1f683d55ac7e3875e4263d28410063e738384d3adc294f5ff3d7105e/frozenlist-1.8.0-cp311-cp311-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:edee74874ce20a373d62dc28b0b18b93f645633c2943fd90ee9d898550770581", size = 243797 }, - { url = "https://files.pythonhosted.org/packages/78/1e/2d5565b589e580c296d3bb54da08d206e797d941a83a6fdea42af23be79c/frozenlist-1.8.0-cp311-cp311-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:c9a63152fe95756b85f31186bddf42e4c02c6321207fd6601a1c89ebac4fe567", size = 247923 }, - { url = "https://files.pythonhosted.org/packages/aa/c3/65872fcf1d326a7f101ad4d86285c403c87be7d832b7470b77f6d2ed5ddc/frozenlist-1.8.0-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:b6db2185db9be0a04fecf2f241c70b63b1a242e2805be291855078f2b404dd6b", size = 230886 }, - { url = "https://files.pythonhosted.org/packages/a0/76/ac9ced601d62f6956f03cc794f9e04c81719509f85255abf96e2510f4265/frozenlist-1.8.0-cp311-cp311-musllinux_1_2_armv7l.whl", hash = "sha256:f4be2e3d8bc8aabd566f8d5b8ba7ecc09249d74ba3c9ed52e54dc23a293f0b92", size = 245731 }, - { url = "https://files.pythonhosted.org/packages/b9/49/ecccb5f2598daf0b4a1415497eba4c33c1e8ce07495eb07d2860c731b8d5/frozenlist-1.8.0-cp311-cp311-musllinux_1_2_ppc64le.whl", hash = "sha256:c8d1634419f39ea6f5c427ea2f90ca85126b54b50837f31497f3bf38266e853d", size = 241544 }, - { url = "https://files.pythonhosted.org/packages/53/4b/ddf24113323c0bbcc54cb38c8b8916f1da7165e07b8e24a717b4a12cbf10/frozenlist-1.8.0-cp311-cp311-musllinux_1_2_s390x.whl", hash = "sha256:1a7fa382a4a223773ed64242dbe1c9c326ec09457e6b8428efb4118c685c3dfd", size = 241806 }, - { url = "https://files.pythonhosted.org/packages/a7/fb/9b9a084d73c67175484ba2789a59f8eebebd0827d186a8102005ce41e1ba/frozenlist-1.8.0-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:11847b53d722050808926e785df837353bd4d75f1d494377e59b23594d834967", size = 229382 }, - { url = "https://files.pythonhosted.org/packages/95/a3/c8fb25aac55bf5e12dae5c5aa6a98f85d436c1dc658f21c3ac73f9fa95e5/frozenlist-1.8.0-cp311-cp311-win32.whl", hash = "sha256:27c6e8077956cf73eadd514be8fb04d77fc946a7fe9f7fe167648b0b9085cc25", size = 39647 }, - { url = "https://files.pythonhosted.org/packages/0a/f5/603d0d6a02cfd4c8f2a095a54672b3cf967ad688a60fb9faf04fc4887f65/frozenlist-1.8.0-cp311-cp311-win_amd64.whl", hash = "sha256:ac913f8403b36a2c8610bbfd25b8013488533e71e62b4b4adce9c86c8cea905b", size = 44064 }, - { url = "https://files.pythonhosted.org/packages/5d/16/c2c9ab44e181f043a86f9a8f84d5124b62dbcb3a02c0977ec72b9ac1d3e0/frozenlist-1.8.0-cp311-cp311-win_arm64.whl", hash = "sha256:d4d3214a0f8394edfa3e303136d0575eece0745ff2b47bd2cb2e66dd92d4351a", size = 39937 }, - { url = "https://files.pythonhosted.org/packages/69/29/948b9aa87e75820a38650af445d2ef2b6b8a6fab1a23b6bb9e4ef0be2d59/frozenlist-1.8.0-cp312-cp312-macosx_10_13_universal2.whl", hash = "sha256:78f7b9e5d6f2fdb88cdde9440dc147259b62b9d3b019924def9f6478be254ac1", size = 87782 }, - { url = "https://files.pythonhosted.org/packages/64/80/4f6e318ee2a7c0750ed724fa33a4bdf1eacdc5a39a7a24e818a773cd91af/frozenlist-1.8.0-cp312-cp312-macosx_10_13_x86_64.whl", hash = "sha256:229bf37d2e4acdaf808fd3f06e854a4a7a3661e871b10dc1f8f1896a3b05f18b", size = 50594 }, - { url = "https://files.pythonhosted.org/packages/2b/94/5c8a2b50a496b11dd519f4a24cb5496cf125681dd99e94c604ccdea9419a/frozenlist-1.8.0-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:f833670942247a14eafbb675458b4e61c82e002a148f49e68257b79296e865c4", size = 50448 }, - { url = "https://files.pythonhosted.org/packages/6a/bd/d91c5e39f490a49df14320f4e8c80161cfcce09f1e2cde1edd16a551abb3/frozenlist-1.8.0-cp312-cp312-manylinux1_x86_64.manylinux_2_28_x86_64.manylinux_2_5_x86_64.whl", hash = "sha256:494a5952b1c597ba44e0e78113a7266e656b9794eec897b19ead706bd7074383", size = 242411 }, - { url = "https://files.pythonhosted.org/packages/8f/83/f61505a05109ef3293dfb1ff594d13d64a2324ac3482be2cedc2be818256/frozenlist-1.8.0-cp312-cp312-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:96f423a119f4777a4a056b66ce11527366a8bb92f54e541ade21f2374433f6d4", size = 243014 }, - { url = "https://files.pythonhosted.org/packages/d8/cb/cb6c7b0f7d4023ddda30cf56b8b17494eb3a79e3fda666bf735f63118b35/frozenlist-1.8.0-cp312-cp312-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:3462dd9475af2025c31cc61be6652dfa25cbfb56cbbf52f4ccfe029f38decaf8", size = 234909 }, - { url = "https://files.pythonhosted.org/packages/31/c5/cd7a1f3b8b34af009fb17d4123c5a778b44ae2804e3ad6b86204255f9ec5/frozenlist-1.8.0-cp312-cp312-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:c4c800524c9cd9bac5166cd6f55285957fcfc907db323e193f2afcd4d9abd69b", size = 250049 }, - { url = "https://files.pythonhosted.org/packages/c0/01/2f95d3b416c584a1e7f0e1d6d31998c4a795f7544069ee2e0962a4b60740/frozenlist-1.8.0-cp312-cp312-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:d6a5df73acd3399d893dafc71663ad22534b5aa4f94e8a2fabfe856c3c1b6a52", size = 256485 }, - { url = "https://files.pythonhosted.org/packages/ce/03/024bf7720b3abaebcff6d0793d73c154237b85bdf67b7ed55e5e9596dc9a/frozenlist-1.8.0-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:405e8fe955c2280ce66428b3ca55e12b3c4e9c336fb2103a4937e891c69a4a29", size = 237619 }, - { url = "https://files.pythonhosted.org/packages/69/fa/f8abdfe7d76b731f5d8bd217827cf6764d4f1d9763407e42717b4bed50a0/frozenlist-1.8.0-cp312-cp312-musllinux_1_2_armv7l.whl", hash = "sha256:908bd3f6439f2fef9e85031b59fd4f1297af54415fb60e4254a95f75b3cab3f3", size = 250320 }, - { url = "https://files.pythonhosted.org/packages/f5/3c/b051329f718b463b22613e269ad72138cc256c540f78a6de89452803a47d/frozenlist-1.8.0-cp312-cp312-musllinux_1_2_ppc64le.whl", hash = "sha256:294e487f9ec720bd8ffcebc99d575f7eff3568a08a253d1ee1a0378754b74143", size = 246820 }, - { url = "https://files.pythonhosted.org/packages/0f/ae/58282e8f98e444b3f4dd42448ff36fa38bef29e40d40f330b22e7108f565/frozenlist-1.8.0-cp312-cp312-musllinux_1_2_s390x.whl", hash = "sha256:74c51543498289c0c43656701be6b077f4b265868fa7f8a8859c197006efb608", size = 250518 }, - { url = "https://files.pythonhosted.org/packages/8f/96/007e5944694d66123183845a106547a15944fbbb7154788cbf7272789536/frozenlist-1.8.0-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:776f352e8329135506a1d6bf16ac3f87bc25b28e765949282dcc627af36123aa", size = 239096 }, - { url = "https://files.pythonhosted.org/packages/66/bb/852b9d6db2fa40be96f29c0d1205c306288f0684df8fd26ca1951d461a56/frozenlist-1.8.0-cp312-cp312-win32.whl", hash = "sha256:433403ae80709741ce34038da08511d4a77062aa924baf411ef73d1146e74faf", size = 39985 }, - { url = "https://files.pythonhosted.org/packages/b8/af/38e51a553dd66eb064cdf193841f16f077585d4d28394c2fa6235cb41765/frozenlist-1.8.0-cp312-cp312-win_amd64.whl", hash = "sha256:34187385b08f866104f0c0617404c8eb08165ab1272e884abc89c112e9c00746", size = 44591 }, - { url = "https://files.pythonhosted.org/packages/a7/06/1dc65480ab147339fecc70797e9c2f69d9cea9cf38934ce08df070fdb9cb/frozenlist-1.8.0-cp312-cp312-win_arm64.whl", hash = "sha256:fe3c58d2f5db5fbd18c2987cba06d51b0529f52bc3a6cdc33d3f4eab725104bd", size = 40102 }, - { url = "https://files.pythonhosted.org/packages/2d/40/0832c31a37d60f60ed79e9dfb5a92e1e2af4f40a16a29abcc7992af9edff/frozenlist-1.8.0-cp313-cp313-macosx_10_13_universal2.whl", hash = "sha256:8d92f1a84bb12d9e56f818b3a746f3efba93c1b63c8387a73dde655e1e42282a", size = 85717 }, - { url = "https://files.pythonhosted.org/packages/30/ba/b0b3de23f40bc55a7057bd38434e25c34fa48e17f20ee273bbde5e0650f3/frozenlist-1.8.0-cp313-cp313-macosx_10_13_x86_64.whl", hash = "sha256:96153e77a591c8adc2ee805756c61f59fef4cf4073a9275ee86fe8cba41241f7", size = 49651 }, - { url = "https://files.pythonhosted.org/packages/0c/ab/6e5080ee374f875296c4243c381bbdef97a9ac39c6e3ce1d5f7d42cb78d6/frozenlist-1.8.0-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:f21f00a91358803399890ab167098c131ec2ddd5f8f5fd5fe9c9f2c6fcd91e40", size = 49417 }, - { url = "https://files.pythonhosted.org/packages/d5/4e/e4691508f9477ce67da2015d8c00acd751e6287739123113a9fca6f1604e/frozenlist-1.8.0-cp313-cp313-manylinux1_x86_64.manylinux_2_28_x86_64.manylinux_2_5_x86_64.whl", hash = "sha256:fb30f9626572a76dfe4293c7194a09fb1fe93ba94c7d4f720dfae3b646b45027", size = 234391 }, - { url = "https://files.pythonhosted.org/packages/40/76/c202df58e3acdf12969a7895fd6f3bc016c642e6726aa63bd3025e0fc71c/frozenlist-1.8.0-cp313-cp313-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:eaa352d7047a31d87dafcacbabe89df0aa506abb5b1b85a2fb91bc3faa02d822", size = 233048 }, - { url = "https://files.pythonhosted.org/packages/f9/c0/8746afb90f17b73ca5979c7a3958116e105ff796e718575175319b5bb4ce/frozenlist-1.8.0-cp313-cp313-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:03ae967b4e297f58f8c774c7eabcce57fe3c2434817d4385c50661845a058121", size = 226549 }, - { url = "https://files.pythonhosted.org/packages/7e/eb/4c7eefc718ff72f9b6c4893291abaae5fbc0c82226a32dcd8ef4f7a5dbef/frozenlist-1.8.0-cp313-cp313-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:f6292f1de555ffcc675941d65fffffb0a5bcd992905015f85d0592201793e0e5", size = 239833 }, - { url = "https://files.pythonhosted.org/packages/c2/4e/e5c02187cf704224f8b21bee886f3d713ca379535f16893233b9d672ea71/frozenlist-1.8.0-cp313-cp313-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:29548f9b5b5e3460ce7378144c3010363d8035cea44bc0bf02d57f5a685e084e", size = 245363 }, - { url = "https://files.pythonhosted.org/packages/1f/96/cb85ec608464472e82ad37a17f844889c36100eed57bea094518bf270692/frozenlist-1.8.0-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:ec3cc8c5d4084591b4237c0a272cc4f50a5b03396a47d9caaf76f5d7b38a4f11", size = 229314 }, - { url = "https://files.pythonhosted.org/packages/5d/6f/4ae69c550e4cee66b57887daeebe006fe985917c01d0fff9caab9883f6d0/frozenlist-1.8.0-cp313-cp313-musllinux_1_2_armv7l.whl", hash = "sha256:517279f58009d0b1f2e7c1b130b377a349405da3f7621ed6bfae50b10adf20c1", size = 243365 }, - { url = "https://files.pythonhosted.org/packages/7a/58/afd56de246cf11780a40a2c28dc7cbabbf06337cc8ddb1c780a2d97e88d8/frozenlist-1.8.0-cp313-cp313-musllinux_1_2_ppc64le.whl", hash = "sha256:db1e72ede2d0d7ccb213f218df6a078a9c09a7de257c2fe8fcef16d5925230b1", size = 237763 }, - { url = "https://files.pythonhosted.org/packages/cb/36/cdfaf6ed42e2644740d4a10452d8e97fa1c062e2a8006e4b09f1b5fd7d63/frozenlist-1.8.0-cp313-cp313-musllinux_1_2_s390x.whl", hash = "sha256:b4dec9482a65c54a5044486847b8a66bf10c9cb4926d42927ec4e8fd5db7fed8", size = 240110 }, - { url = "https://files.pythonhosted.org/packages/03/a8/9ea226fbefad669f11b52e864c55f0bd57d3c8d7eb07e9f2e9a0b39502e1/frozenlist-1.8.0-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:21900c48ae04d13d416f0e1e0c4d81f7931f73a9dfa0b7a8746fb2fe7dd970ed", size = 233717 }, - { url = "https://files.pythonhosted.org/packages/1e/0b/1b5531611e83ba7d13ccc9988967ea1b51186af64c42b7a7af465dcc9568/frozenlist-1.8.0-cp313-cp313-win32.whl", hash = "sha256:8b7b94a067d1c504ee0b16def57ad5738701e4ba10cec90529f13fa03c833496", size = 39628 }, - { url = "https://files.pythonhosted.org/packages/d8/cf/174c91dbc9cc49bc7b7aab74d8b734e974d1faa8f191c74af9b7e80848e6/frozenlist-1.8.0-cp313-cp313-win_amd64.whl", hash = "sha256:878be833caa6a3821caf85eb39c5ba92d28e85df26d57afb06b35b2efd937231", size = 43882 }, - { url = "https://files.pythonhosted.org/packages/c1/17/502cd212cbfa96eb1388614fe39a3fc9ab87dbbe042b66f97acb57474834/frozenlist-1.8.0-cp313-cp313-win_arm64.whl", hash = "sha256:44389d135b3ff43ba8cc89ff7f51f5a0bb6b63d829c8300f79a2fe4fe61bcc62", size = 39676 }, - { url = "https://files.pythonhosted.org/packages/d2/5c/3bbfaa920dfab09e76946a5d2833a7cbdf7b9b4a91c714666ac4855b88b4/frozenlist-1.8.0-cp313-cp313t-macosx_10_13_universal2.whl", hash = "sha256:e25ac20a2ef37e91c1b39938b591457666a0fa835c7783c3a8f33ea42870db94", size = 89235 }, - { url = "https://files.pythonhosted.org/packages/d2/d6/f03961ef72166cec1687e84e8925838442b615bd0b8854b54923ce5b7b8a/frozenlist-1.8.0-cp313-cp313t-macosx_10_13_x86_64.whl", hash = "sha256:07cdca25a91a4386d2e76ad992916a85038a9b97561bf7a3fd12d5d9ce31870c", size = 50742 }, - { url = "https://files.pythonhosted.org/packages/1e/bb/a6d12b7ba4c3337667d0e421f7181c82dda448ce4e7ad7ecd249a16fa806/frozenlist-1.8.0-cp313-cp313t-macosx_11_0_arm64.whl", hash = "sha256:4e0c11f2cc6717e0a741f84a527c52616140741cd812a50422f83dc31749fb52", size = 51725 }, - { url = "https://files.pythonhosted.org/packages/bc/71/d1fed0ffe2c2ccd70b43714c6cab0f4188f09f8a67a7914a6b46ee30f274/frozenlist-1.8.0-cp313-cp313t-manylinux1_x86_64.manylinux_2_28_x86_64.manylinux_2_5_x86_64.whl", hash = "sha256:b3210649ee28062ea6099cfda39e147fa1bc039583c8ee4481cb7811e2448c51", size = 284533 }, - { url = "https://files.pythonhosted.org/packages/c9/1f/fb1685a7b009d89f9bf78a42d94461bc06581f6e718c39344754a5d9bada/frozenlist-1.8.0-cp313-cp313t-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:581ef5194c48035a7de2aefc72ac6539823bb71508189e5de01d60c9dcd5fa65", size = 292506 }, - { url = "https://files.pythonhosted.org/packages/e6/3b/b991fe1612703f7e0d05c0cf734c1b77aaf7c7d321df4572e8d36e7048c8/frozenlist-1.8.0-cp313-cp313t-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:3ef2d026f16a2b1866e1d86fc4e1291e1ed8a387b2c333809419a2f8b3a77b82", size = 274161 }, - { url = "https://files.pythonhosted.org/packages/ca/ec/c5c618767bcdf66e88945ec0157d7f6c4a1322f1473392319b7a2501ded7/frozenlist-1.8.0-cp313-cp313t-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:5500ef82073f599ac84d888e3a8c1f77ac831183244bfd7f11eaa0289fb30714", size = 294676 }, - { url = "https://files.pythonhosted.org/packages/7c/ce/3934758637d8f8a88d11f0585d6495ef54b2044ed6ec84492a91fa3b27aa/frozenlist-1.8.0-cp313-cp313t-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:50066c3997d0091c411a66e710f4e11752251e6d2d73d70d8d5d4c76442a199d", size = 300638 }, - { url = "https://files.pythonhosted.org/packages/fc/4f/a7e4d0d467298f42de4b41cbc7ddaf19d3cfeabaf9ff97c20c6c7ee409f9/frozenlist-1.8.0-cp313-cp313t-musllinux_1_2_aarch64.whl", hash = "sha256:5c1c8e78426e59b3f8005e9b19f6ff46e5845895adbde20ece9218319eca6506", size = 283067 }, - { url = "https://files.pythonhosted.org/packages/dc/48/c7b163063d55a83772b268e6d1affb960771b0e203b632cfe09522d67ea5/frozenlist-1.8.0-cp313-cp313t-musllinux_1_2_armv7l.whl", hash = "sha256:eefdba20de0d938cec6a89bd4d70f346a03108a19b9df4248d3cf0d88f1b0f51", size = 292101 }, - { url = "https://files.pythonhosted.org/packages/9f/d0/2366d3c4ecdc2fd391e0afa6e11500bfba0ea772764d631bbf82f0136c9d/frozenlist-1.8.0-cp313-cp313t-musllinux_1_2_ppc64le.whl", hash = "sha256:cf253e0e1c3ceb4aaff6df637ce033ff6535fb8c70a764a8f46aafd3d6ab798e", size = 289901 }, - { url = "https://files.pythonhosted.org/packages/b8/94/daff920e82c1b70e3618a2ac39fbc01ae3e2ff6124e80739ce5d71c9b920/frozenlist-1.8.0-cp313-cp313t-musllinux_1_2_s390x.whl", hash = "sha256:032efa2674356903cd0261c4317a561a6850f3ac864a63fc1583147fb05a79b0", size = 289395 }, - { url = "https://files.pythonhosted.org/packages/e3/20/bba307ab4235a09fdcd3cc5508dbabd17c4634a1af4b96e0f69bfe551ebd/frozenlist-1.8.0-cp313-cp313t-musllinux_1_2_x86_64.whl", hash = "sha256:6da155091429aeba16851ecb10a9104a108bcd32f6c1642867eadaee401c1c41", size = 283659 }, - { url = "https://files.pythonhosted.org/packages/fd/00/04ca1c3a7a124b6de4f8a9a17cc2fcad138b4608e7a3fc5877804b8715d7/frozenlist-1.8.0-cp313-cp313t-win32.whl", hash = "sha256:0f96534f8bfebc1a394209427d0f8a63d343c9779cda6fc25e8e121b5fd8555b", size = 43492 }, - { url = "https://files.pythonhosted.org/packages/59/5e/c69f733a86a94ab10f68e496dc6b7e8bc078ebb415281d5698313e3af3a1/frozenlist-1.8.0-cp313-cp313t-win_amd64.whl", hash = "sha256:5d63a068f978fc69421fb0e6eb91a9603187527c86b7cd3f534a5b77a592b888", size = 48034 }, - { url = "https://files.pythonhosted.org/packages/16/6c/be9d79775d8abe79b05fa6d23da99ad6e7763a1d080fbae7290b286093fd/frozenlist-1.8.0-cp313-cp313t-win_arm64.whl", hash = "sha256:bf0a7e10b077bf5fb9380ad3ae8ce20ef919a6ad93b4552896419ac7e1d8e042", size = 41749 }, - { url = "https://files.pythonhosted.org/packages/f1/c8/85da824b7e7b9b6e7f7705b2ecaf9591ba6f79c1177f324c2735e41d36a2/frozenlist-1.8.0-cp314-cp314-macosx_10_13_universal2.whl", hash = "sha256:cee686f1f4cadeb2136007ddedd0aaf928ab95216e7691c63e50a8ec066336d0", size = 86127 }, - { url = "https://files.pythonhosted.org/packages/8e/e8/a1185e236ec66c20afd72399522f142c3724c785789255202d27ae992818/frozenlist-1.8.0-cp314-cp314-macosx_10_13_x86_64.whl", hash = "sha256:119fb2a1bd47307e899c2fac7f28e85b9a543864df47aa7ec9d3c1b4545f096f", size = 49698 }, - { url = "https://files.pythonhosted.org/packages/a1/93/72b1736d68f03fda5fdf0f2180fb6caaae3894f1b854d006ac61ecc727ee/frozenlist-1.8.0-cp314-cp314-macosx_11_0_arm64.whl", hash = "sha256:4970ece02dbc8c3a92fcc5228e36a3e933a01a999f7094ff7c23fbd2beeaa67c", size = 49749 }, - { url = "https://files.pythonhosted.org/packages/a7/b2/fabede9fafd976b991e9f1b9c8c873ed86f202889b864756f240ce6dd855/frozenlist-1.8.0-cp314-cp314-manylinux1_x86_64.manylinux_2_28_x86_64.manylinux_2_5_x86_64.whl", hash = "sha256:cba69cb73723c3f329622e34bdbf5ce1f80c21c290ff04256cff1cd3c2036ed2", size = 231298 }, - { url = "https://files.pythonhosted.org/packages/3a/3b/d9b1e0b0eed36e70477ffb8360c49c85c8ca8ef9700a4e6711f39a6e8b45/frozenlist-1.8.0-cp314-cp314-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:778a11b15673f6f1df23d9586f83c4846c471a8af693a22e066508b77d201ec8", size = 232015 }, - { url = "https://files.pythonhosted.org/packages/dc/94/be719d2766c1138148564a3960fc2c06eb688da592bdc25adcf856101be7/frozenlist-1.8.0-cp314-cp314-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:0325024fe97f94c41c08872db482cf8ac4800d80e79222c6b0b7b162d5b13686", size = 225038 }, - { url = "https://files.pythonhosted.org/packages/e4/09/6712b6c5465f083f52f50cf74167b92d4ea2f50e46a9eea0523d658454ae/frozenlist-1.8.0-cp314-cp314-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:97260ff46b207a82a7567b581ab4190bd4dfa09f4db8a8b49d1a958f6aa4940e", size = 240130 }, - { url = "https://files.pythonhosted.org/packages/f8/d4/cd065cdcf21550b54f3ce6a22e143ac9e4836ca42a0de1022da8498eac89/frozenlist-1.8.0-cp314-cp314-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:54b2077180eb7f83dd52c40b2750d0a9f175e06a42e3213ce047219de902717a", size = 242845 }, - { url = "https://files.pythonhosted.org/packages/62/c3/f57a5c8c70cd1ead3d5d5f776f89d33110b1addae0ab010ad774d9a44fb9/frozenlist-1.8.0-cp314-cp314-musllinux_1_2_aarch64.whl", hash = "sha256:2f05983daecab868a31e1da44462873306d3cbfd76d1f0b5b69c473d21dbb128", size = 229131 }, - { url = "https://files.pythonhosted.org/packages/6c/52/232476fe9cb64f0742f3fde2b7d26c1dac18b6d62071c74d4ded55e0ef94/frozenlist-1.8.0-cp314-cp314-musllinux_1_2_armv7l.whl", hash = "sha256:33f48f51a446114bc5d251fb2954ab0164d5be02ad3382abcbfe07e2531d650f", size = 240542 }, - { url = "https://files.pythonhosted.org/packages/5f/85/07bf3f5d0fb5414aee5f47d33c6f5c77bfe49aac680bfece33d4fdf6a246/frozenlist-1.8.0-cp314-cp314-musllinux_1_2_ppc64le.whl", hash = "sha256:154e55ec0655291b5dd1b8731c637ecdb50975a2ae70c606d100750a540082f7", size = 237308 }, - { url = "https://files.pythonhosted.org/packages/11/99/ae3a33d5befd41ac0ca2cc7fd3aa707c9c324de2e89db0e0f45db9a64c26/frozenlist-1.8.0-cp314-cp314-musllinux_1_2_s390x.whl", hash = "sha256:4314debad13beb564b708b4a496020e5306c7333fa9a3ab90374169a20ffab30", size = 238210 }, - { url = "https://files.pythonhosted.org/packages/b2/60/b1d2da22f4970e7a155f0adde9b1435712ece01b3cd45ba63702aea33938/frozenlist-1.8.0-cp314-cp314-musllinux_1_2_x86_64.whl", hash = "sha256:073f8bf8becba60aa931eb3bc420b217bb7d5b8f4750e6f8b3be7f3da85d38b7", size = 231972 }, - { url = "https://files.pythonhosted.org/packages/3f/ab/945b2f32de889993b9c9133216c068b7fcf257d8595a0ac420ac8677cab0/frozenlist-1.8.0-cp314-cp314-win32.whl", hash = "sha256:bac9c42ba2ac65ddc115d930c78d24ab8d4f465fd3fc473cdedfccadb9429806", size = 40536 }, - { url = "https://files.pythonhosted.org/packages/59/ad/9caa9b9c836d9ad6f067157a531ac48b7d36499f5036d4141ce78c230b1b/frozenlist-1.8.0-cp314-cp314-win_amd64.whl", hash = "sha256:3e0761f4d1a44f1d1a47996511752cf3dcec5bbdd9cc2b4fe595caf97754b7a0", size = 44330 }, - { url = "https://files.pythonhosted.org/packages/82/13/e6950121764f2676f43534c555249f57030150260aee9dcf7d64efda11dd/frozenlist-1.8.0-cp314-cp314-win_arm64.whl", hash = "sha256:d1eaff1d00c7751b7c6662e9c5ba6eb2c17a2306ba5e2a37f24ddf3cc953402b", size = 40627 }, - { url = "https://files.pythonhosted.org/packages/c0/c7/43200656ecc4e02d3f8bc248df68256cd9572b3f0017f0a0c4e93440ae23/frozenlist-1.8.0-cp314-cp314t-macosx_10_13_universal2.whl", hash = "sha256:d3bb933317c52d7ea5004a1c442eef86f426886fba134ef8cf4226ea6ee1821d", size = 89238 }, - { url = "https://files.pythonhosted.org/packages/d1/29/55c5f0689b9c0fb765055629f472c0de484dcaf0acee2f7707266ae3583c/frozenlist-1.8.0-cp314-cp314t-macosx_10_13_x86_64.whl", hash = "sha256:8009897cdef112072f93a0efdce29cd819e717fd2f649ee3016efd3cd885a7ed", size = 50738 }, - { url = "https://files.pythonhosted.org/packages/ba/7d/b7282a445956506fa11da8c2db7d276adcbf2b17d8bb8407a47685263f90/frozenlist-1.8.0-cp314-cp314t-macosx_11_0_arm64.whl", hash = "sha256:2c5dcbbc55383e5883246d11fd179782a9d07a986c40f49abe89ddf865913930", size = 51739 }, - { url = "https://files.pythonhosted.org/packages/62/1c/3d8622e60d0b767a5510d1d3cf21065b9db874696a51ea6d7a43180a259c/frozenlist-1.8.0-cp314-cp314t-manylinux1_x86_64.manylinux_2_28_x86_64.manylinux_2_5_x86_64.whl", hash = "sha256:39ecbc32f1390387d2aa4f5a995e465e9e2f79ba3adcac92d68e3e0afae6657c", size = 284186 }, - { url = "https://files.pythonhosted.org/packages/2d/14/aa36d5f85a89679a85a1d44cd7a6657e0b1c75f61e7cad987b203d2daca8/frozenlist-1.8.0-cp314-cp314t-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:92db2bf818d5cc8d9c1f1fc56b897662e24ea5adb36ad1f1d82875bd64e03c24", size = 292196 }, - { url = "https://files.pythonhosted.org/packages/05/23/6bde59eb55abd407d34f77d39a5126fb7b4f109a3f611d3929f14b700c66/frozenlist-1.8.0-cp314-cp314t-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:2dc43a022e555de94c3b68a4ef0b11c4f747d12c024a520c7101709a2144fb37", size = 273830 }, - { url = "https://files.pythonhosted.org/packages/d2/3f/22cff331bfad7a8afa616289000ba793347fcd7bc275f3b28ecea2a27909/frozenlist-1.8.0-cp314-cp314t-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:cb89a7f2de3602cfed448095bab3f178399646ab7c61454315089787df07733a", size = 294289 }, - { url = "https://files.pythonhosted.org/packages/a4/89/5b057c799de4838b6c69aa82b79705f2027615e01be996d2486a69ca99c4/frozenlist-1.8.0-cp314-cp314t-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:33139dc858c580ea50e7e60a1b0ea003efa1fd42e6ec7fdbad78fff65fad2fd2", size = 300318 }, - { url = "https://files.pythonhosted.org/packages/30/de/2c22ab3eb2a8af6d69dc799e48455813bab3690c760de58e1bf43b36da3e/frozenlist-1.8.0-cp314-cp314t-musllinux_1_2_aarch64.whl", hash = "sha256:168c0969a329b416119507ba30b9ea13688fafffac1b7822802537569a1cb0ef", size = 282814 }, - { url = "https://files.pythonhosted.org/packages/59/f7/970141a6a8dbd7f556d94977858cfb36fa9b66e0892c6dd780d2219d8cd8/frozenlist-1.8.0-cp314-cp314t-musllinux_1_2_armv7l.whl", hash = "sha256:28bd570e8e189d7f7b001966435f9dac6718324b5be2990ac496cf1ea9ddb7fe", size = 291762 }, - { url = "https://files.pythonhosted.org/packages/c1/15/ca1adae83a719f82df9116d66f5bb28bb95557b3951903d39135620ef157/frozenlist-1.8.0-cp314-cp314t-musllinux_1_2_ppc64le.whl", hash = "sha256:b2a095d45c5d46e5e79ba1e5b9cb787f541a8dee0433836cea4b96a2c439dcd8", size = 289470 }, - { url = "https://files.pythonhosted.org/packages/ac/83/dca6dc53bf657d371fbc88ddeb21b79891e747189c5de990b9dfff2ccba1/frozenlist-1.8.0-cp314-cp314t-musllinux_1_2_s390x.whl", hash = "sha256:eab8145831a0d56ec9c4139b6c3e594c7a83c2c8be25d5bcf2d86136a532287a", size = 289042 }, - { url = "https://files.pythonhosted.org/packages/96/52/abddd34ca99be142f354398700536c5bd315880ed0a213812bc491cff5e4/frozenlist-1.8.0-cp314-cp314t-musllinux_1_2_x86_64.whl", hash = "sha256:974b28cf63cc99dfb2188d8d222bc6843656188164848c4f679e63dae4b0708e", size = 283148 }, - { url = "https://files.pythonhosted.org/packages/af/d3/76bd4ed4317e7119c2b7f57c3f6934aba26d277acc6309f873341640e21f/frozenlist-1.8.0-cp314-cp314t-win32.whl", hash = "sha256:342c97bf697ac5480c0a7ec73cd700ecfa5a8a40ac923bd035484616efecc2df", size = 44676 }, - { url = "https://files.pythonhosted.org/packages/89/76/c615883b7b521ead2944bb3480398cbb07e12b7b4e4d073d3752eb721558/frozenlist-1.8.0-cp314-cp314t-win_amd64.whl", hash = "sha256:06be8f67f39c8b1dc671f5d83aaefd3358ae5cdcf8314552c57e7ed3e6475bdd", size = 49451 }, - { url = "https://files.pythonhosted.org/packages/e0/a3/5982da14e113d07b325230f95060e2169f5311b1017ea8af2a29b374c289/frozenlist-1.8.0-cp314-cp314t-win_arm64.whl", hash = "sha256:102e6314ca4da683dca92e3b1355490fed5f313b768500084fbe6371fddfdb79", size = 42507 }, - { url = "https://files.pythonhosted.org/packages/9a/9a/e35b4a917281c0b8419d4207f4334c8e8c5dbf4f3f5f9ada73958d937dcc/frozenlist-1.8.0-py3-none-any.whl", hash = "sha256:0c18a16eab41e82c295618a77502e17b195883241c563b00f0aa5106fc4eaa0d", size = 13409 }, -] - -[[package]] -name = "fsspec" -version = "2024.12.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/ee/11/de70dee31455c546fbc88301971ec03c328f3d1138cfba14263f651e9551/fsspec-2024.12.0.tar.gz", hash = "sha256:670700c977ed2fb51e0d9f9253177ed20cbde4a3e5c0283cc5385b5870c8533f", size = 291600 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/de/86/5486b0188d08aa643e127774a99bac51ffa6cf343e3deb0583956dca5b22/fsspec-2024.12.0-py3-none-any.whl", hash = "sha256:b520aed47ad9804237ff878b504267a3b0b441e97508bd6d2d8774e3db85cee2", size = 183862 }, -] - -[package.optional-dependencies] -http = [ - { name = "aiohttp" }, -] - -[[package]] -name = "future" -version = "1.0.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/a7/b2/4140c69c6a66432916b26158687e821ba631a4c9273c474343badf84d3ba/future-1.0.0.tar.gz", hash = "sha256:bd2968309307861edae1458a4f8a4f3598c03be43b97521076aebf5d94c07b05", size = 1228490 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/da/71/ae30dadffc90b9006d77af76b393cb9dfbfc9629f339fc1574a1c52e6806/future-1.0.0-py3-none-any.whl", hash = "sha256:929292d34f5872e70396626ef385ec22355a1fae8ad29e1a734c3e43f9fbc216", size = 491326 }, -] - -[[package]] -name = "gitdb" -version = "4.0.12" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "smmap" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/72/94/63b0fc47eb32792c7ba1fe1b694daec9a63620db1e313033d18140c2320a/gitdb-4.0.12.tar.gz", hash = "sha256:5ef71f855d191a3326fcfbc0d5da835f26b13fbcba60c32c21091c349ffdb571", size = 394684 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/a0/61/5c78b91c3143ed5c14207f463aecfc8f9dbb5092fb2869baf37c273b2705/gitdb-4.0.12-py3-none-any.whl", hash = "sha256:67073e15955400952c6565cc3e707c554a4eea2e428946f7a4c162fab9bd9bcf", size = 62794 }, -] - -[[package]] -name = "gitpython" -version = "3.1.45" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "gitdb" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/9a/c8/dd58967d119baab745caec2f9d853297cec1989ec1d63f677d3880632b88/gitpython-3.1.45.tar.gz", hash = "sha256:85b0ee964ceddf211c41b9f27a49086010a190fd8132a24e21f362a4b36a791c", size = 215076 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/01/61/d4b89fec821f72385526e1b9d9a3a0385dda4a72b206d28049e2c7cd39b8/gitpython-3.1.45-py3-none-any.whl", hash = "sha256:8908cb2e02fb3b93b7eb0f2827125cb699869470432cc885f019b8fd0fccff77", size = 208168 }, -] - -[[package]] -name = "graphviz" -version = "0.21" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/f8/b3/3ac91e9be6b761a4b30d66ff165e54439dcd48b83f4e20d644867215f6ca/graphviz-0.21.tar.gz", hash = "sha256:20743e7183be82aaaa8ad6c93f8893c923bd6658a04c32ee115edb3c8a835f78", size = 200434 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/91/4c/e0ce1ef95d4000ebc1c11801f9b944fa5910ecc15b5e351865763d8657f8/graphviz-0.21-py3-none-any.whl", hash = "sha256:54f33de9f4f911d7e84e4191749cac8cc5653f815b06738c54db9a15ab8b1e42", size = 47300 }, -] - -[[package]] -name = "greenlet" -version = "3.3.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/c7/e5/40dbda2736893e3e53d25838e0f19a2b417dfc122b9989c91918db30b5d3/greenlet-3.3.0.tar.gz", hash = "sha256:a82bb225a4e9e4d653dd2fb7b8b2d36e4fb25bc0165422a11e48b88e9e6f78fb", size = 190651 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/32/6a/33d1702184d94106d3cdd7bfb788e19723206fce152e303473ca3b946c7b/greenlet-3.3.0-cp310-cp310-macosx_11_0_universal2.whl", hash = "sha256:6f8496d434d5cb2dce025773ba5597f71f5410ae499d5dd9533e0653258cdb3d", size = 273658 }, - { url = "https://files.pythonhosted.org/packages/d6/b7/2b5805bbf1907c26e434f4e448cd8b696a0b71725204fa21a211ff0c04a7/greenlet-3.3.0-cp310-cp310-manylinux_2_24_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:b96dc7eef78fd404e022e165ec55327f935b9b52ff355b067eb4a0267fc1cffb", size = 574810 }, - { url = "https://files.pythonhosted.org/packages/94/38/343242ec12eddf3d8458c73f555c084359883d4ddc674240d9e61ec51fd6/greenlet-3.3.0-cp310-cp310-manylinux_2_24_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:73631cd5cccbcfe63e3f9492aaa664d278fda0ce5c3d43aeda8e77317e38efbd", size = 586248 }, - { url = "https://files.pythonhosted.org/packages/f0/d0/0ae86792fb212e4384041e0ef8e7bc66f59a54912ce407d26a966ed2914d/greenlet-3.3.0-cp310-cp310-manylinux_2_24_s390x.manylinux_2_28_s390x.whl", hash = "sha256:b299a0cb979f5d7197442dccc3aee67fce53500cd88951b7e6c35575701c980b", size = 597403 }, - { url = "https://files.pythonhosted.org/packages/b6/a8/15d0aa26c0036a15d2659175af00954aaaa5d0d66ba538345bd88013b4d7/greenlet-3.3.0-cp310-cp310-manylinux_2_24_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:7dee147740789a4632cace364816046e43310b59ff8fb79833ab043aefa72fd5", size = 586910 }, - { url = "https://files.pythonhosted.org/packages/e1/9b/68d5e3b7ccaba3907e5532cf8b9bf16f9ef5056a008f195a367db0ff32db/greenlet-3.3.0-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:39b28e339fc3c348427560494e28d8a6f3561c8d2bcf7d706e1c624ed8d822b9", size = 1547206 }, - { url = "https://files.pythonhosted.org/packages/66/bd/e3086ccedc61e49f91e2cfb5ffad9d8d62e5dc85e512a6200f096875b60c/greenlet-3.3.0-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:b3c374782c2935cc63b2a27ba8708471de4ad1abaa862ffdb1ef45a643ddbb7d", size = 1613359 }, - { url = "https://files.pythonhosted.org/packages/f4/6b/d4e73f5dfa888364bbf02efa85616c6714ae7c631c201349782e5b428925/greenlet-3.3.0-cp310-cp310-win_amd64.whl", hash = "sha256:b49e7ed51876b459bd645d83db257f0180e345d3f768a35a85437a24d5a49082", size = 300740 }, - { url = "https://files.pythonhosted.org/packages/1f/cb/48e964c452ca2b92175a9b2dca037a553036cb053ba69e284650ce755f13/greenlet-3.3.0-cp311-cp311-macosx_11_0_universal2.whl", hash = "sha256:e29f3018580e8412d6aaf5641bb7745d38c85228dacf51a73bd4e26ddf2a6a8e", size = 274908 }, - { url = "https://files.pythonhosted.org/packages/28/da/38d7bff4d0277b594ec557f479d65272a893f1f2a716cad91efeb8680953/greenlet-3.3.0-cp311-cp311-manylinux_2_24_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:a687205fb22794e838f947e2194c0566d3812966b41c78709554aa883183fb62", size = 577113 }, - { url = "https://files.pythonhosted.org/packages/3c/f2/89c5eb0faddc3ff014f1c04467d67dee0d1d334ab81fadbf3744847f8a8a/greenlet-3.3.0-cp311-cp311-manylinux_2_24_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:4243050a88ba61842186cb9e63c7dfa677ec146160b0efd73b855a3d9c7fcf32", size = 590338 }, - { url = "https://files.pythonhosted.org/packages/80/d7/db0a5085035d05134f8c089643da2b44cc9b80647c39e93129c5ef170d8f/greenlet-3.3.0-cp311-cp311-manylinux_2_24_s390x.manylinux_2_28_s390x.whl", hash = "sha256:670d0f94cd302d81796e37299bcd04b95d62403883b24225c6b5271466612f45", size = 601098 }, - { url = "https://files.pythonhosted.org/packages/dc/a6/e959a127b630a58e23529972dbc868c107f9d583b5a9f878fb858c46bc1a/greenlet-3.3.0-cp311-cp311-manylinux_2_24_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:6cb3a8ec3db4a3b0eb8a3c25436c2d49e3505821802074969db017b87bc6a948", size = 590206 }, - { url = "https://files.pythonhosted.org/packages/48/60/29035719feb91798693023608447283b266b12efc576ed013dd9442364bb/greenlet-3.3.0-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:2de5a0b09eab81fc6a382791b995b1ccf2b172a9fec934747a7a23d2ff291794", size = 1550668 }, - { url = "https://files.pythonhosted.org/packages/0a/5f/783a23754b691bfa86bd72c3033aa107490deac9b2ef190837b860996c9f/greenlet-3.3.0-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:4449a736606bd30f27f8e1ff4678ee193bc47f6ca810d705981cfffd6ce0d8c5", size = 1615483 }, - { url = "https://files.pythonhosted.org/packages/1d/d5/c339b3b4bc8198b7caa4f2bd9fd685ac9f29795816d8db112da3d04175bb/greenlet-3.3.0-cp311-cp311-win_amd64.whl", hash = "sha256:7652ee180d16d447a683c04e4c5f6441bae7ba7b17ffd9f6b3aff4605e9e6f71", size = 301164 }, - { url = "https://files.pythonhosted.org/packages/f8/0a/a3871375c7b9727edaeeea994bfff7c63ff7804c9829c19309ba2e058807/greenlet-3.3.0-cp312-cp312-macosx_11_0_universal2.whl", hash = "sha256:b01548f6e0b9e9784a2c99c5651e5dc89ffcbe870bc5fb2e5ef864e9cc6b5dcb", size = 276379 }, - { url = "https://files.pythonhosted.org/packages/43/ab/7ebfe34dce8b87be0d11dae91acbf76f7b8246bf9d6b319c741f99fa59c6/greenlet-3.3.0-cp312-cp312-manylinux_2_24_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:349345b770dc88f81506c6861d22a6ccd422207829d2c854ae2af8025af303e3", size = 597294 }, - { url = "https://files.pythonhosted.org/packages/a4/39/f1c8da50024feecd0793dbd5e08f526809b8ab5609224a2da40aad3a7641/greenlet-3.3.0-cp312-cp312-manylinux_2_24_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:e8e18ed6995e9e2c0b4ed264d2cf89260ab3ac7e13555b8032b25a74c6d18655", size = 607742 }, - { url = "https://files.pythonhosted.org/packages/77/cb/43692bcd5f7a0da6ec0ec6d58ee7cddb606d055ce94a62ac9b1aa481e969/greenlet-3.3.0-cp312-cp312-manylinux_2_24_s390x.manylinux_2_28_s390x.whl", hash = "sha256:c024b1e5696626890038e34f76140ed1daf858e37496d33f2af57f06189e70d7", size = 622297 }, - { url = "https://files.pythonhosted.org/packages/75/b0/6bde0b1011a60782108c01de5913c588cf51a839174538d266de15e4bf4d/greenlet-3.3.0-cp312-cp312-manylinux_2_24_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:047ab3df20ede6a57c35c14bf5200fcf04039d50f908270d3f9a7a82064f543b", size = 609885 }, - { url = "https://files.pythonhosted.org/packages/49/0e/49b46ac39f931f59f987b7cd9f34bfec8ef81d2a1e6e00682f55be5de9f4/greenlet-3.3.0-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:2d9ad37fc657b1102ec880e637cccf20191581f75c64087a549e66c57e1ceb53", size = 1567424 }, - { url = "https://files.pythonhosted.org/packages/05/f5/49a9ac2dff7f10091935def9165c90236d8f175afb27cbed38fb1d61ab6b/greenlet-3.3.0-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:83cd0e36932e0e7f36a64b732a6f60c2fc2df28c351bae79fbaf4f8092fe7614", size = 1636017 }, - { url = "https://files.pythonhosted.org/packages/6c/79/3912a94cf27ec503e51ba493692d6db1e3cd8ac7ac52b0b47c8e33d7f4f9/greenlet-3.3.0-cp312-cp312-win_amd64.whl", hash = "sha256:a7a34b13d43a6b78abf828a6d0e87d3385680eaf830cd60d20d52f249faabf39", size = 301964 }, - { url = "https://files.pythonhosted.org/packages/02/2f/28592176381b9ab2cafa12829ba7b472d177f3acc35d8fbcf3673d966fff/greenlet-3.3.0-cp313-cp313-macosx_11_0_universal2.whl", hash = "sha256:a1e41a81c7e2825822f4e068c48cb2196002362619e2d70b148f20a831c00739", size = 275140 }, - { url = "https://files.pythonhosted.org/packages/2c/80/fbe937bf81e9fca98c981fe499e59a3f45df2a04da0baa5c2be0dca0d329/greenlet-3.3.0-cp313-cp313-manylinux_2_24_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:9f515a47d02da4d30caaa85b69474cec77b7929b2e936ff7fb853d42f4bf8808", size = 599219 }, - { url = "https://files.pythonhosted.org/packages/c2/ff/7c985128f0514271b8268476af89aee6866df5eec04ac17dcfbc676213df/greenlet-3.3.0-cp313-cp313-manylinux_2_24_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:7d2d9fd66bfadf230b385fdc90426fcd6eb64db54b40c495b72ac0feb5766c54", size = 610211 }, - { url = "https://files.pythonhosted.org/packages/79/07/c47a82d881319ec18a4510bb30463ed6891f2ad2c1901ed5ec23d3de351f/greenlet-3.3.0-cp313-cp313-manylinux_2_24_s390x.manylinux_2_28_s390x.whl", hash = "sha256:30a6e28487a790417d036088b3bcb3f3ac7d8babaa7d0139edbaddebf3af9492", size = 624311 }, - { url = "https://files.pythonhosted.org/packages/fd/8e/424b8c6e78bd9837d14ff7df01a9829fc883ba2ab4ea787d4f848435f23f/greenlet-3.3.0-cp313-cp313-manylinux_2_24_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:087ea5e004437321508a8d6f20efc4cfec5e3c30118e1417ea96ed1d93950527", size = 612833 }, - { url = "https://files.pythonhosted.org/packages/b5/ba/56699ff9b7c76ca12f1cdc27a886d0f81f2189c3455ff9f65246780f713d/greenlet-3.3.0-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:ab97cf74045343f6c60a39913fa59710e4bd26a536ce7ab2397adf8b27e67c39", size = 1567256 }, - { url = "https://files.pythonhosted.org/packages/1e/37/f31136132967982d698c71a281a8901daf1a8fbab935dce7c0cf15f942cc/greenlet-3.3.0-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:5375d2e23184629112ca1ea89a53389dddbffcf417dad40125713d88eb5f96e8", size = 1636483 }, - { url = "https://files.pythonhosted.org/packages/7e/71/ba21c3fb8c5dce83b8c01f458a42e99ffdb1963aeec08fff5a18588d8fd7/greenlet-3.3.0-cp313-cp313-win_amd64.whl", hash = "sha256:9ee1942ea19550094033c35d25d20726e4f1c40d59545815e1128ac58d416d38", size = 301833 }, - { url = "https://files.pythonhosted.org/packages/d7/7c/f0a6d0ede2c7bf092d00bc83ad5bafb7e6ec9b4aab2fbdfa6f134dc73327/greenlet-3.3.0-cp314-cp314-macosx_11_0_universal2.whl", hash = "sha256:60c2ef0f578afb3c8d92ea07ad327f9a062547137afe91f38408f08aacab667f", size = 275671 }, - { url = "https://files.pythonhosted.org/packages/44/06/dac639ae1a50f5969d82d2e3dd9767d30d6dbdbab0e1a54010c8fe90263c/greenlet-3.3.0-cp314-cp314-manylinux_2_24_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:0a5d554d0712ba1de0a6c94c640f7aeba3f85b3a6e1f2899c11c2c0428da9365", size = 646360 }, - { url = "https://files.pythonhosted.org/packages/e0/94/0fb76fe6c5369fba9bf98529ada6f4c3a1adf19e406a47332245ef0eb357/greenlet-3.3.0-cp314-cp314-manylinux_2_24_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:3a898b1e9c5f7307ebbde4102908e6cbfcb9ea16284a3abe15cab996bee8b9b3", size = 658160 }, - { url = "https://files.pythonhosted.org/packages/93/79/d2c70cae6e823fac36c3bbc9077962105052b7ef81db2f01ec3b9bf17e2b/greenlet-3.3.0-cp314-cp314-manylinux_2_24_s390x.manylinux_2_28_s390x.whl", hash = "sha256:dcd2bdbd444ff340e8d6bdf54d2f206ccddbb3ccfdcd3c25bf4afaa7b8f0cf45", size = 671388 }, - { url = "https://files.pythonhosted.org/packages/b8/14/bab308fc2c1b5228c3224ec2bf928ce2e4d21d8046c161e44a2012b5203e/greenlet-3.3.0-cp314-cp314-manylinux_2_24_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:5773edda4dc00e173820722711d043799d3adb4f01731f40619e07ea2750b955", size = 660166 }, - { url = "https://files.pythonhosted.org/packages/4b/d2/91465d39164eaa0085177f61983d80ffe746c5a1860f009811d498e7259c/greenlet-3.3.0-cp314-cp314-musllinux_1_2_aarch64.whl", hash = "sha256:ac0549373982b36d5fd5d30beb8a7a33ee541ff98d2b502714a09f1169f31b55", size = 1615193 }, - { url = "https://files.pythonhosted.org/packages/42/1b/83d110a37044b92423084d52d5d5a3b3a73cafb51b547e6d7366ff62eff1/greenlet-3.3.0-cp314-cp314-musllinux_1_2_x86_64.whl", hash = "sha256:d198d2d977460358c3b3a4dc844f875d1adb33817f0613f663a656f463764ccc", size = 1683653 }, - { url = "https://files.pythonhosted.org/packages/7c/9a/9030e6f9aa8fd7808e9c31ba4c38f87c4f8ec324ee67431d181fe396d705/greenlet-3.3.0-cp314-cp314-win_amd64.whl", hash = "sha256:73f51dd0e0bdb596fb0417e475fa3c5e32d4c83638296e560086b8d7da7c4170", size = 305387 }, - { url = "https://files.pythonhosted.org/packages/a0/66/bd6317bc5932accf351fc19f177ffba53712a202f9df10587da8df257c7e/greenlet-3.3.0-cp314-cp314t-macosx_11_0_universal2.whl", hash = "sha256:d6ed6f85fae6cdfdb9ce04c9bf7a08d666cfcfb914e7d006f44f840b46741931", size = 282638 }, - { url = "https://files.pythonhosted.org/packages/30/cf/cc81cb030b40e738d6e69502ccbd0dd1bced0588e958f9e757945de24404/greenlet-3.3.0-cp314-cp314t-manylinux_2_24_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:d9125050fcf24554e69c4cacb086b87b3b55dc395a8b3ebe6487b045b2614388", size = 651145 }, - { url = "https://files.pythonhosted.org/packages/9c/ea/1020037b5ecfe95ca7df8d8549959baceb8186031da83d5ecceff8b08cd2/greenlet-3.3.0-cp314-cp314t-manylinux_2_24_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:87e63ccfa13c0a0f6234ed0add552af24cc67dd886731f2261e46e241608bee3", size = 654236 }, - { url = "https://files.pythonhosted.org/packages/69/cc/1e4bae2e45ca2fa55299f4e85854606a78ecc37fead20d69322f96000504/greenlet-3.3.0-cp314-cp314t-manylinux_2_24_s390x.manylinux_2_28_s390x.whl", hash = "sha256:2662433acbca297c9153a4023fe2161c8dcfdcc91f10433171cf7e7d94ba2221", size = 662506 }, - { url = "https://files.pythonhosted.org/packages/57/b9/f8025d71a6085c441a7eaff0fd928bbb275a6633773667023d19179fe815/greenlet-3.3.0-cp314-cp314t-manylinux_2_24_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:3c6e9b9c1527a78520357de498b0e709fb9e2f49c3a513afd5a249007261911b", size = 653783 }, - { url = "https://files.pythonhosted.org/packages/f6/c7/876a8c7a7485d5d6b5c6821201d542ef28be645aa024cfe1145b35c120c1/greenlet-3.3.0-cp314-cp314t-musllinux_1_2_aarch64.whl", hash = "sha256:286d093f95ec98fdd92fcb955003b8a3d054b4e2cab3e2707a5039e7b50520fd", size = 1614857 }, - { url = "https://files.pythonhosted.org/packages/4f/dc/041be1dff9f23dac5f48a43323cd0789cb798342011c19a248d9c9335536/greenlet-3.3.0-cp314-cp314t-musllinux_1_2_x86_64.whl", hash = "sha256:6c10513330af5b8ae16f023e8ddbfb486ab355d04467c4679c5cfe4659975dd9", size = 1676034 }, -] - -[[package]] -name = "grpcio" -version = "1.76.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/b6/e0/318c1ce3ae5a17894d5791e87aea147587c9e702f24122cc7a5c8bbaeeb1/grpcio-1.76.0.tar.gz", hash = "sha256:7be78388d6da1a25c0d5ec506523db58b18be22d9c37d8d3a32c08be4987bd73", size = 12785182 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/88/17/ff4795dc9a34b6aee6ec379f1b66438a3789cd1315aac0cbab60d92f74b3/grpcio-1.76.0-cp310-cp310-linux_armv7l.whl", hash = "sha256:65a20de41e85648e00305c1bb09a3598f840422e522277641145a32d42dcefcc", size = 5840037 }, - { url = "https://files.pythonhosted.org/packages/4e/ff/35f9b96e3fa2f12e1dcd58a4513a2e2294a001d64dec81677361b7040c9a/grpcio-1.76.0-cp310-cp310-macosx_11_0_universal2.whl", hash = "sha256:40ad3afe81676fd9ec6d9d406eda00933f218038433980aa19d401490e46ecde", size = 11836482 }, - { url = "https://files.pythonhosted.org/packages/3e/1c/8374990f9545e99462caacea5413ed783014b3b66ace49e35c533f07507b/grpcio-1.76.0-cp310-cp310-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:035d90bc79eaa4bed83f524331d55e35820725c9fbb00ffa1904d5550ed7ede3", size = 6407178 }, - { url = "https://files.pythonhosted.org/packages/1e/77/36fd7d7c75a6c12542c90a6d647a27935a1ecaad03e0ffdb7c42db6b04d2/grpcio-1.76.0-cp310-cp310-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:4215d3a102bd95e2e11b5395c78562967959824156af11fa93d18fdd18050990", size = 7075684 }, - { url = "https://files.pythonhosted.org/packages/38/f7/e3cdb252492278e004722306c5a8935eae91e64ea11f0af3437a7de2e2b7/grpcio-1.76.0-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:49ce47231818806067aea3324d4bf13825b658ad662d3b25fada0bdad9b8a6af", size = 6611133 }, - { url = "https://files.pythonhosted.org/packages/7e/20/340db7af162ccd20a0893b5f3c4a5d676af7b71105517e62279b5b61d95a/grpcio-1.76.0-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:8cc3309d8e08fd79089e13ed4819d0af72aa935dd8f435a195fd152796752ff2", size = 7195507 }, - { url = "https://files.pythonhosted.org/packages/10/f0/b2160addc1487bd8fa4810857a27132fb4ce35c1b330c2f3ac45d697b106/grpcio-1.76.0-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:971fd5a1d6e62e00d945423a567e42eb1fa678ba89072832185ca836a94daaa6", size = 8160651 }, - { url = "https://files.pythonhosted.org/packages/2c/2c/ac6f98aa113c6ef111b3f347854e99ebb7fb9d8f7bb3af1491d438f62af4/grpcio-1.76.0-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:9d9adda641db7207e800a7f089068f6f645959f2df27e870ee81d44701dd9db3", size = 7620568 }, - { url = "https://files.pythonhosted.org/packages/90/84/7852f7e087285e3ac17a2703bc4129fafee52d77c6c82af97d905566857e/grpcio-1.76.0-cp310-cp310-win32.whl", hash = "sha256:063065249d9e7e0782d03d2bca50787f53bd0fb89a67de9a7b521c4a01f1989b", size = 3998879 }, - { url = "https://files.pythonhosted.org/packages/10/30/d3d2adcbb6dd3ff59d6ac3df6ef830e02b437fb5c90990429fd180e52f30/grpcio-1.76.0-cp310-cp310-win_amd64.whl", hash = "sha256:a6ae758eb08088d36812dd5d9af7a9859c05b1e0f714470ea243694b49278e7b", size = 4706892 }, - { url = "https://files.pythonhosted.org/packages/a0/00/8163a1beeb6971f66b4bbe6ac9457b97948beba8dd2fc8e1281dce7f79ec/grpcio-1.76.0-cp311-cp311-linux_armv7l.whl", hash = "sha256:2e1743fbd7f5fa713a1b0a8ac8ebabf0ec980b5d8809ec358d488e273b9cf02a", size = 5843567 }, - { url = "https://files.pythonhosted.org/packages/10/c1/934202f5cf335e6d852530ce14ddb0fef21be612ba9ecbbcbd4d748ca32d/grpcio-1.76.0-cp311-cp311-macosx_11_0_universal2.whl", hash = "sha256:a8c2cf1209497cf659a667d7dea88985e834c24b7c3b605e6254cbb5076d985c", size = 11848017 }, - { url = "https://files.pythonhosted.org/packages/11/0b/8dec16b1863d74af6eb3543928600ec2195af49ca58b16334972f6775663/grpcio-1.76.0-cp311-cp311-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:08caea849a9d3c71a542827d6df9d5a69067b0a1efbea8a855633ff5d9571465", size = 6412027 }, - { url = "https://files.pythonhosted.org/packages/d7/64/7b9e6e7ab910bea9d46f2c090380bab274a0b91fb0a2fe9b0cd399fffa12/grpcio-1.76.0-cp311-cp311-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:f0e34c2079d47ae9f6188211db9e777c619a21d4faba6977774e8fa43b085e48", size = 7075913 }, - { url = "https://files.pythonhosted.org/packages/68/86/093c46e9546073cefa789bd76d44c5cb2abc824ca62af0c18be590ff13ba/grpcio-1.76.0-cp311-cp311-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:8843114c0cfce61b40ad48df65abcfc00d4dba82eae8718fab5352390848c5da", size = 6615417 }, - { url = "https://files.pythonhosted.org/packages/f7/b6/5709a3a68500a9c03da6fb71740dcdd5ef245e39266461a03f31a57036d8/grpcio-1.76.0-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:8eddfb4d203a237da6f3cc8a540dad0517d274b5a1e9e636fd8d2c79b5c1d397", size = 7199683 }, - { url = "https://files.pythonhosted.org/packages/91/d3/4b1f2bf16ed52ce0b508161df3a2d186e4935379a159a834cb4a7d687429/grpcio-1.76.0-cp311-cp311-musllinux_1_2_i686.whl", hash = "sha256:32483fe2aab2c3794101c2a159070584e5db11d0aa091b2c0ea9c4fc43d0d749", size = 8163109 }, - { url = "https://files.pythonhosted.org/packages/5c/61/d9043f95f5f4cf085ac5dd6137b469d41befb04bd80280952ffa2a4c3f12/grpcio-1.76.0-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:dcfe41187da8992c5f40aa8c5ec086fa3672834d2be57a32384c08d5a05b4c00", size = 7626676 }, - { url = "https://files.pythonhosted.org/packages/36/95/fd9a5152ca02d8881e4dd419cdd790e11805979f499a2e5b96488b85cf27/grpcio-1.76.0-cp311-cp311-win32.whl", hash = "sha256:2107b0c024d1b35f4083f11245c0e23846ae64d02f40b2b226684840260ed054", size = 3997688 }, - { url = "https://files.pythonhosted.org/packages/60/9c/5c359c8d4c9176cfa3c61ecd4efe5affe1f38d9bae81e81ac7186b4c9cc8/grpcio-1.76.0-cp311-cp311-win_amd64.whl", hash = "sha256:522175aba7af9113c48ec10cc471b9b9bd4f6ceb36aeb4544a8e2c80ed9d252d", size = 4709315 }, - { url = "https://files.pythonhosted.org/packages/bf/05/8e29121994b8d959ffa0afd28996d452f291b48cfc0875619de0bde2c50c/grpcio-1.76.0-cp312-cp312-linux_armv7l.whl", hash = "sha256:81fd9652b37b36f16138611c7e884eb82e0cec137c40d3ef7c3f9b3ed00f6ed8", size = 5799718 }, - { url = "https://files.pythonhosted.org/packages/d9/75/11d0e66b3cdf998c996489581bdad8900db79ebd83513e45c19548f1cba4/grpcio-1.76.0-cp312-cp312-macosx_11_0_universal2.whl", hash = "sha256:04bbe1bfe3a68bbfd4e52402ab7d4eb59d72d02647ae2042204326cf4bbad280", size = 11825627 }, - { url = "https://files.pythonhosted.org/packages/28/50/2f0aa0498bc188048f5d9504dcc5c2c24f2eb1a9337cd0fa09a61a2e75f0/grpcio-1.76.0-cp312-cp312-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:d388087771c837cdb6515539f43b9d4bf0b0f23593a24054ac16f7a960be16f4", size = 6359167 }, - { url = "https://files.pythonhosted.org/packages/66/e5/bbf0bb97d29ede1d59d6588af40018cfc345b17ce979b7b45424628dc8bb/grpcio-1.76.0-cp312-cp312-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:9f8f757bebaaea112c00dba718fc0d3260052ce714e25804a03f93f5d1c6cc11", size = 7044267 }, - { url = "https://files.pythonhosted.org/packages/f5/86/f6ec2164f743d9609691115ae8ece098c76b894ebe4f7c94a655c6b03e98/grpcio-1.76.0-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:980a846182ce88c4f2f7e2c22c56aefd515daeb36149d1c897f83cf57999e0b6", size = 6573963 }, - { url = "https://files.pythonhosted.org/packages/60/bc/8d9d0d8505feccfdf38a766d262c71e73639c165b311c9457208b56d92ae/grpcio-1.76.0-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:f92f88e6c033db65a5ae3d97905c8fea9c725b63e28d5a75cb73b49bda5024d8", size = 7164484 }, - { url = "https://files.pythonhosted.org/packages/67/e6/5d6c2fc10b95edf6df9b8f19cf10a34263b7fd48493936fffd5085521292/grpcio-1.76.0-cp312-cp312-musllinux_1_2_i686.whl", hash = "sha256:4baf3cbe2f0be3289eb68ac8ae771156971848bb8aaff60bad42005539431980", size = 8127777 }, - { url = "https://files.pythonhosted.org/packages/3f/c8/dce8ff21c86abe025efe304d9e31fdb0deaaa3b502b6a78141080f206da0/grpcio-1.76.0-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:615ba64c208aaceb5ec83bfdce7728b80bfeb8be97562944836a7a0a9647d882", size = 7594014 }, - { url = "https://files.pythonhosted.org/packages/e0/42/ad28191ebf983a5d0ecef90bab66baa5a6b18f2bfdef9d0a63b1973d9f75/grpcio-1.76.0-cp312-cp312-win32.whl", hash = "sha256:45d59a649a82df5718fd9527ce775fd66d1af35e6d31abdcdc906a49c6822958", size = 3984750 }, - { url = "https://files.pythonhosted.org/packages/9e/00/7bd478cbb851c04a48baccaa49b75abaa8e4122f7d86da797500cccdd771/grpcio-1.76.0-cp312-cp312-win_amd64.whl", hash = "sha256:c088e7a90b6017307f423efbb9d1ba97a22aa2170876223f9709e9d1de0b5347", size = 4704003 }, - { url = "https://files.pythonhosted.org/packages/fc/ed/71467ab770effc9e8cef5f2e7388beb2be26ed642d567697bb103a790c72/grpcio-1.76.0-cp313-cp313-linux_armv7l.whl", hash = "sha256:26ef06c73eb53267c2b319f43e6634c7556ea37672029241a056629af27c10e2", size = 5807716 }, - { url = "https://files.pythonhosted.org/packages/2c/85/c6ed56f9817fab03fa8a111ca91469941fb514e3e3ce6d793cb8f1e1347b/grpcio-1.76.0-cp313-cp313-macosx_11_0_universal2.whl", hash = "sha256:45e0111e73f43f735d70786557dc38141185072d7ff8dc1829d6a77ac1471468", size = 11821522 }, - { url = "https://files.pythonhosted.org/packages/ac/31/2b8a235ab40c39cbc141ef647f8a6eb7b0028f023015a4842933bc0d6831/grpcio-1.76.0-cp313-cp313-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:83d57312a58dcfe2a3a0f9d1389b299438909a02db60e2f2ea2ae2d8034909d3", size = 6362558 }, - { url = "https://files.pythonhosted.org/packages/bd/64/9784eab483358e08847498ee56faf8ff6ea8e0a4592568d9f68edc97e9e9/grpcio-1.76.0-cp313-cp313-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:3e2a27c89eb9ac3d81ec8835e12414d73536c6e620355d65102503064a4ed6eb", size = 7049990 }, - { url = "https://files.pythonhosted.org/packages/2b/94/8c12319a6369434e7a184b987e8e9f3b49a114c489b8315f029e24de4837/grpcio-1.76.0-cp313-cp313-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:61f69297cba3950a524f61c7c8ee12e55c486cb5f7db47ff9dcee33da6f0d3ae", size = 6575387 }, - { url = "https://files.pythonhosted.org/packages/15/0f/f12c32b03f731f4a6242f771f63039df182c8b8e2cf8075b245b409259d4/grpcio-1.76.0-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:6a15c17af8839b6801d554263c546c69c4d7718ad4321e3166175b37eaacca77", size = 7166668 }, - { url = "https://files.pythonhosted.org/packages/ff/2d/3ec9ce0c2b1d92dd59d1c3264aaec9f0f7c817d6e8ac683b97198a36ed5a/grpcio-1.76.0-cp313-cp313-musllinux_1_2_i686.whl", hash = "sha256:25a18e9810fbc7e7f03ec2516addc116a957f8cbb8cbc95ccc80faa072743d03", size = 8124928 }, - { url = "https://files.pythonhosted.org/packages/1a/74/fd3317be5672f4856bcdd1a9e7b5e17554692d3db9a3b273879dc02d657d/grpcio-1.76.0-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:931091142fd8cc14edccc0845a79248bc155425eee9a98b2db2ea4f00a235a42", size = 7589983 }, - { url = "https://files.pythonhosted.org/packages/45/bb/ca038cf420f405971f19821c8c15bcbc875505f6ffadafe9ffd77871dc4c/grpcio-1.76.0-cp313-cp313-win32.whl", hash = "sha256:5e8571632780e08526f118f74170ad8d50fb0a48c23a746bef2a6ebade3abd6f", size = 3984727 }, - { url = "https://files.pythonhosted.org/packages/41/80/84087dc56437ced7cdd4b13d7875e7439a52a261e3ab4e06488ba6173b0a/grpcio-1.76.0-cp313-cp313-win_amd64.whl", hash = "sha256:f9f7bd5faab55f47231ad8dba7787866b69f5e93bc306e3915606779bbfb4ba8", size = 4702799 }, - { url = "https://files.pythonhosted.org/packages/b4/46/39adac80de49d678e6e073b70204091e76631e03e94928b9ea4ecf0f6e0e/grpcio-1.76.0-cp314-cp314-linux_armv7l.whl", hash = "sha256:ff8a59ea85a1f2191a0ffcc61298c571bc566332f82e5f5be1b83c9d8e668a62", size = 5808417 }, - { url = "https://files.pythonhosted.org/packages/9c/f5/a4531f7fb8b4e2a60b94e39d5d924469b7a6988176b3422487be61fe2998/grpcio-1.76.0-cp314-cp314-macosx_11_0_universal2.whl", hash = "sha256:06c3d6b076e7b593905d04fdba6a0525711b3466f43b3400266f04ff735de0cd", size = 11828219 }, - { url = "https://files.pythonhosted.org/packages/4b/1c/de55d868ed7a8bd6acc6b1d6ddc4aa36d07a9f31d33c912c804adb1b971b/grpcio-1.76.0-cp314-cp314-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:fd5ef5932f6475c436c4a55e4336ebbe47bd3272be04964a03d316bbf4afbcbc", size = 6367826 }, - { url = "https://files.pythonhosted.org/packages/59/64/99e44c02b5adb0ad13ab3adc89cb33cb54bfa90c74770f2607eea629b86f/grpcio-1.76.0-cp314-cp314-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:b331680e46239e090f5b3cead313cc772f6caa7d0fc8de349337563125361a4a", size = 7049550 }, - { url = "https://files.pythonhosted.org/packages/43/28/40a5be3f9a86949b83e7d6a2ad6011d993cbe9b6bd27bea881f61c7788b6/grpcio-1.76.0-cp314-cp314-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:2229ae655ec4e8999599469559e97630185fdd53ae1e8997d147b7c9b2b72cba", size = 6575564 }, - { url = "https://files.pythonhosted.org/packages/4b/a9/1be18e6055b64467440208a8559afac243c66a8b904213af6f392dc2212f/grpcio-1.76.0-cp314-cp314-musllinux_1_2_aarch64.whl", hash = "sha256:490fa6d203992c47c7b9e4a9d39003a0c2bcc1c9aa3c058730884bbbb0ee9f09", size = 7176236 }, - { url = "https://files.pythonhosted.org/packages/0f/55/dba05d3fcc151ce6e81327541d2cc8394f442f6b350fead67401661bf041/grpcio-1.76.0-cp314-cp314-musllinux_1_2_i686.whl", hash = "sha256:479496325ce554792dba6548fae3df31a72cef7bad71ca2e12b0e58f9b336bfc", size = 8125795 }, - { url = "https://files.pythonhosted.org/packages/4a/45/122df922d05655f63930cf42c9e3f72ba20aadb26c100ee105cad4ce4257/grpcio-1.76.0-cp314-cp314-musllinux_1_2_x86_64.whl", hash = "sha256:1c9b93f79f48b03ada57ea24725d83a30284a012ec27eab2cf7e50a550cbbbcc", size = 7592214 }, - { url = "https://files.pythonhosted.org/packages/4a/6e/0b899b7f6b66e5af39e377055fb4a6675c9ee28431df5708139df2e93233/grpcio-1.76.0-cp314-cp314-win32.whl", hash = "sha256:747fa73efa9b8b1488a95d0ba1039c8e2dca0f741612d80415b1e1c560febf4e", size = 4062961 }, - { url = "https://files.pythonhosted.org/packages/19/41/0b430b01a2eb38ee887f88c1f07644a1df8e289353b78e82b37ef988fb64/grpcio-1.76.0-cp314-cp314-win_amd64.whl", hash = "sha256:922fa70ba549fce362d2e2871ab542082d66e2aaf0c19480ea453905b01f384e", size = 4834462 }, -] - -[[package]] -name = "h11" -version = "0.16.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/01/ee/02a2c011bdab74c6fb3c75474d40b3052059d95df7e73351460c8588d963/h11-0.16.0.tar.gz", hash = "sha256:4e35b956cf45792e4caa5885e69fba00bdbc6ffafbfa020300e549b208ee5ff1", size = 101250 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/04/4b/29cac41a4d98d144bf5f6d33995617b185d14b22401f75ca86f384e87ff1/h11-0.16.0-py3-none-any.whl", hash = "sha256:63cf8bbe7522de3bf65932fda1d9c2772064ffb3dae62d55932da54b31cb6c86", size = 37515 }, -] - -[[package]] -name = "hf-xet" -version = "1.2.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/5e/6e/0f11bacf08a67f7fb5ee09740f2ca54163863b07b70d579356e9222ce5d8/hf_xet-1.2.0.tar.gz", hash = "sha256:a8c27070ca547293b6890c4bf389f713f80e8c478631432962bb7f4bc0bd7d7f", size = 506020 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/9e/a5/85ef910a0aa034a2abcfadc360ab5ac6f6bc4e9112349bd40ca97551cff0/hf_xet-1.2.0-cp313-cp313t-macosx_10_12_x86_64.whl", hash = "sha256:ceeefcd1b7aed4956ae8499e2199607765fbd1c60510752003b6cc0b8413b649", size = 2861870 }, - { url = "https://files.pythonhosted.org/packages/ea/40/e2e0a7eb9a51fe8828ba2d47fe22a7e74914ea8a0db68a18c3aa7449c767/hf_xet-1.2.0-cp313-cp313t-macosx_11_0_arm64.whl", hash = "sha256:b70218dd548e9840224df5638fdc94bd033552963cfa97f9170829381179c813", size = 2717584 }, - { url = "https://files.pythonhosted.org/packages/a5/7d/daf7f8bc4594fdd59a8a596f9e3886133fdc68e675292218a5e4c1b7e834/hf_xet-1.2.0-cp313-cp313t-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:7d40b18769bb9a8bc82a9ede575ce1a44c75eb80e7375a01d76259089529b5dc", size = 3315004 }, - { url = "https://files.pythonhosted.org/packages/b1/ba/45ea2f605fbf6d81c8b21e4d970b168b18a53515923010c312c06cd83164/hf_xet-1.2.0-cp313-cp313t-manylinux_2_28_aarch64.whl", hash = "sha256:cd3a6027d59cfb60177c12d6424e31f4b5ff13d8e3a1247b3a584bf8977e6df5", size = 3222636 }, - { url = "https://files.pythonhosted.org/packages/4a/1d/04513e3cab8f29ab8c109d309ddd21a2705afab9d52f2ba1151e0c14f086/hf_xet-1.2.0-cp313-cp313t-musllinux_1_2_aarch64.whl", hash = "sha256:6de1fc44f58f6dd937956c8d304d8c2dea264c80680bcfa61ca4a15e7b76780f", size = 3408448 }, - { url = "https://files.pythonhosted.org/packages/f0/7c/60a2756d7feec7387db3a1176c632357632fbe7849fce576c5559d4520c7/hf_xet-1.2.0-cp313-cp313t-musllinux_1_2_x86_64.whl", hash = "sha256:f182f264ed2acd566c514e45da9f2119110e48a87a327ca271027904c70c5832", size = 3503401 }, - { url = "https://files.pythonhosted.org/packages/4e/64/48fffbd67fb418ab07451e4ce641a70de1c40c10a13e25325e24858ebe5a/hf_xet-1.2.0-cp313-cp313t-win_amd64.whl", hash = "sha256:293a7a3787e5c95d7be1857358a9130694a9c6021de3f27fa233f37267174382", size = 2900866 }, - { url = "https://files.pythonhosted.org/packages/e2/51/f7e2caae42f80af886db414d4e9885fac959330509089f97cccb339c6b87/hf_xet-1.2.0-cp314-cp314t-macosx_10_12_x86_64.whl", hash = "sha256:10bfab528b968c70e062607f663e21e34e2bba349e8038db546646875495179e", size = 2861861 }, - { url = "https://files.pythonhosted.org/packages/6e/1d/a641a88b69994f9371bd347f1dd35e5d1e2e2460a2e350c8d5165fc62005/hf_xet-1.2.0-cp314-cp314t-macosx_11_0_arm64.whl", hash = "sha256:2a212e842647b02eb6a911187dc878e79c4aa0aa397e88dd3b26761676e8c1f8", size = 2717699 }, - { url = "https://files.pythonhosted.org/packages/df/e0/e5e9bba7d15f0318955f7ec3f4af13f92e773fbb368c0b8008a5acbcb12f/hf_xet-1.2.0-cp314-cp314t-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:30e06daccb3a7d4c065f34fc26c14c74f4653069bb2b194e7f18f17cbe9939c0", size = 3314885 }, - { url = "https://files.pythonhosted.org/packages/21/90/b7fe5ff6f2b7b8cbdf1bd56145f863c90a5807d9758a549bf3d916aa4dec/hf_xet-1.2.0-cp314-cp314t-manylinux_2_28_aarch64.whl", hash = "sha256:29c8fc913a529ec0a91867ce3d119ac1aac966e098cf49501800c870328cc090", size = 3221550 }, - { url = "https://files.pythonhosted.org/packages/6f/cb/73f276f0a7ce46cc6a6ec7d6c7d61cbfe5f2e107123d9bbd0193c355f106/hf_xet-1.2.0-cp314-cp314t-musllinux_1_2_aarch64.whl", hash = "sha256:66e159cbfcfbb29f920db2c09ed8b660eb894640d284f102ada929b6e3dc410a", size = 3408010 }, - { url = "https://files.pythonhosted.org/packages/b8/1e/d642a12caa78171f4be64f7cd9c40e3ca5279d055d0873188a58c0f5fbb9/hf_xet-1.2.0-cp314-cp314t-musllinux_1_2_x86_64.whl", hash = "sha256:9c91d5ae931510107f148874e9e2de8a16052b6f1b3ca3c1b12f15ccb491390f", size = 3503264 }, - { url = "https://files.pythonhosted.org/packages/17/b5/33764714923fa1ff922770f7ed18c2daae034d21ae6e10dbf4347c854154/hf_xet-1.2.0-cp314-cp314t-win_amd64.whl", hash = "sha256:210d577732b519ac6ede149d2f2f34049d44e8622bf14eb3d63bbcd2d4b332dc", size = 2901071 }, - { url = "https://files.pythonhosted.org/packages/96/2d/22338486473df5923a9ab7107d375dbef9173c338ebef5098ef593d2b560/hf_xet-1.2.0-cp37-abi3-macosx_10_12_x86_64.whl", hash = "sha256:46740d4ac024a7ca9b22bebf77460ff43332868b661186a8e46c227fdae01848", size = 2866099 }, - { url = "https://files.pythonhosted.org/packages/7f/8c/c5becfa53234299bc2210ba314eaaae36c2875e0045809b82e40a9544f0c/hf_xet-1.2.0-cp37-abi3-macosx_11_0_arm64.whl", hash = "sha256:27df617a076420d8845bea087f59303da8be17ed7ec0cd7ee3b9b9f579dff0e4", size = 2722178 }, - { url = "https://files.pythonhosted.org/packages/9a/92/cf3ab0b652b082e66876d08da57fcc6fa2f0e6c70dfbbafbd470bb73eb47/hf_xet-1.2.0-cp37-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3651fd5bfe0281951b988c0facbe726aa5e347b103a675f49a3fa8144c7968fd", size = 3320214 }, - { url = "https://files.pythonhosted.org/packages/46/92/3f7ec4a1b6a65bf45b059b6d4a5d38988f63e193056de2f420137e3c3244/hf_xet-1.2.0-cp37-abi3-manylinux_2_28_aarch64.whl", hash = "sha256:d06fa97c8562fb3ee7a378dd9b51e343bc5bc8190254202c9771029152f5e08c", size = 3229054 }, - { url = "https://files.pythonhosted.org/packages/0b/dd/7ac658d54b9fb7999a0ccb07ad863b413cbaf5cf172f48ebcd9497ec7263/hf_xet-1.2.0-cp37-abi3-musllinux_1_2_aarch64.whl", hash = "sha256:4c1428c9ae73ec0939410ec73023c4f842927f39db09b063b9482dac5a3bb737", size = 3413812 }, - { url = "https://files.pythonhosted.org/packages/92/68/89ac4e5b12a9ff6286a12174c8538a5930e2ed662091dd2572bbe0a18c8a/hf_xet-1.2.0-cp37-abi3-musllinux_1_2_x86_64.whl", hash = "sha256:a55558084c16b09b5ed32ab9ed38421e2d87cf3f1f89815764d1177081b99865", size = 3508920 }, - { url = "https://files.pythonhosted.org/packages/cb/44/870d44b30e1dcfb6a65932e3e1506c103a8a5aea9103c337e7a53180322c/hf_xet-1.2.0-cp37-abi3-win_amd64.whl", hash = "sha256:e6584a52253f72c9f52f9e549d5895ca7a471608495c4ecaa6cc73dba2b24d69", size = 2905735 }, -] - -[[package]] -name = "httpcore" -version = "1.0.9" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "certifi" }, - { name = "h11" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/06/94/82699a10bca87a5556c9c59b5963f2d039dbd239f25bc2a63907a05a14cb/httpcore-1.0.9.tar.gz", hash = "sha256:6e34463af53fd2ab5d807f399a9b45ea31c3dfa2276f15a2c3f00afff6e176e8", size = 85484 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/7e/f5/f66802a942d491edb555dd61e3a9961140fd64c90bce1eafd741609d334d/httpcore-1.0.9-py3-none-any.whl", hash = "sha256:2d400746a40668fc9dec9810239072b40b4484b640a8c38fd654a024c7a1bf55", size = 78784 }, -] - -[[package]] -name = "httpx" -version = "0.28.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "anyio" }, - { name = "certifi" }, - { name = "httpcore" }, - { name = "idna" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/b1/df/48c586a5fe32a0f01324ee087459e112ebb7224f646c0b5023f5e79e9956/httpx-0.28.1.tar.gz", hash = "sha256:75e98c5f16b0f35b567856f597f06ff2270a374470a5c2392242528e3e3e42fc", size = 141406 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/2a/39/e50c7c3a983047577ee07d2a9e53faf5a69493943ec3f6a384bdc792deb2/httpx-0.28.1-py3-none-any.whl", hash = "sha256:d909fcccc110f8c7faf814ca82a9a4d816bc5a6dbfea25d6591d6985b8ba59ad", size = 73517 }, -] - -[[package]] -name = "huggingface-hub" -version = "0.36.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "filelock" }, - { name = "fsspec" }, - { name = "hf-xet", marker = "platform_machine == 'aarch64' or platform_machine == 'amd64' or platform_machine == 'arm64' or platform_machine == 'x86_64'" }, - { name = "packaging" }, - { name = "pyyaml" }, - { name = "requests" }, - { name = "tqdm" }, - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/98/63/4910c5fa9128fdadf6a9c5ac138e8b1b6cee4ca44bf7915bbfbce4e355ee/huggingface_hub-0.36.0.tar.gz", hash = "sha256:47b3f0e2539c39bf5cde015d63b72ec49baff67b6931c3d97f3f84532e2b8d25", size = 463358 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/cb/bd/1a875e0d592d447cbc02805fd3fe0f497714d6a2583f59d14fa9ebad96eb/huggingface_hub-0.36.0-py3-none-any.whl", hash = "sha256:7bcc9ad17d5b3f07b57c78e79d527102d08313caa278a641993acddcb894548d", size = 566094 }, -] - -[[package]] -name = "hydra-core" -version = "1.3.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "antlr4-python3-runtime" }, - { name = "omegaconf" }, - { name = "packaging" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/6d/8e/07e42bc434a847154083b315779b0a81d567154504624e181caf2c71cd98/hydra-core-1.3.2.tar.gz", hash = "sha256:8a878ed67216997c3e9d88a8e72e7b4767e81af37afb4ea3334b269a4390a824", size = 3263494 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/c6/50/e0edd38dcd63fb26a8547f13d28f7a008bc4a3fd4eb4ff030673f22ad41a/hydra_core-1.3.2-py3-none-any.whl", hash = "sha256:fa0238a9e31df3373b35b0bfb672c34cc92718d21f81311d8996a16de1141d8b", size = 154547 }, -] - -[[package]] -name = "idna" -version = "3.11" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/6f/6d/0703ccc57f3a7233505399edb88de3cbd678da106337b9fcde432b65ed60/idna-3.11.tar.gz", hash = "sha256:795dafcc9c04ed0c1fb032c2aa73654d8e8c5023a7df64a53f39190ada629902", size = 194582 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/0e/61/66938bbb5fc52dbdf84594873d5b51fb1f7c7794e9c0f5bd885f30bc507b/idna-3.11-py3-none-any.whl", hash = "sha256:771a87f49d9defaf64091e6e6fe9c18d4833f140bd19464795bc32d966ca37ea", size = 71008 }, -] - -[[package]] -name = "indic-numtowords" -version = "1.1.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/aa/46/683e92580d9c1752917d2f9ec2a44d2adc21cdfe4deeaa0fe87fc23dbea8/indic_numtowords-1.1.0.tar.gz", hash = "sha256:d1addc21444c332e05bfd8726af427960c096c2a16776d98bee4fbc36ade5d25", size = 44220 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/1e/16/b2b6491d95a15bda712163b4e27428f2cb1ac3a1b4fb59b140dbc76f6ce5/indic_numtowords-1.1.0-py3-none-any.whl", hash = "sha256:bf4b7b9e539323d9b00bc868caa2d9369170b8f3ac4d19619bf9c6cdc6f89572", size = 71635 }, -] - -[[package]] -name = "inflect" -version = "7.5.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "more-itertools" }, - { name = "typeguard" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/78/c6/943357d44a21fd995723d07ccaddd78023eace03c1846049a2645d4324a3/inflect-7.5.0.tar.gz", hash = "sha256:faf19801c3742ed5a05a8ce388e0d8fe1a07f8d095c82201eb904f5d27ad571f", size = 73751 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/8a/eb/427ed2b20a38a4ee29f24dbe4ae2dafab198674fe9a85e3d6adf9e5f5f41/inflect-7.5.0-py3-none-any.whl", hash = "sha256:2aea70e5e70c35d8350b8097396ec155ffd68def678c7ff97f51aa69c1d92344", size = 35197 }, -] - -[[package]] -name = "iniconfig" -version = "2.3.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/72/34/14ca021ce8e5dfedc35312d08ba8bf51fdd999c576889fc2c24cb97f4f10/iniconfig-2.3.0.tar.gz", hash = "sha256:c76315c77db068650d49c5b56314774a7804df16fee4402c1f19d6d15d8c4730", size = 20503 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/cb/b1/3846dd7f199d53cb17f49cba7e651e9ce294d8497c8c150530ed11865bb8/iniconfig-2.3.0-py3-none-any.whl", hash = "sha256:f631c04d2c48c52b84d0d0549c99ff3859c98df65b3101406327ecc7d53fbf12", size = 7484 }, -] - -[[package]] -name = "intervaltree" -version = "3.2.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "sortedcontainers" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/53/c3/b2afa612aa0373f3e6bb190e6de35f293b307d1537f109e3e25dbfcdf212/intervaltree-3.2.1.tar.gz", hash = "sha256:f3f7e8baeb7dd75b9f7a6d33cf3ec10025984a8e66e3016d537e52130c73cfe2", size = 1231531 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/83/7f/8a80a1c7c2ed05822b5a2b312d2995f30c533641f8198366ba2e26a7bb03/intervaltree-3.2.1-py2.py3-none-any.whl", hash = "sha256:a8a8381bbd35d48ceebee932c77ffc988492d22fb1d27d0ba1d74a7694eb8f0b", size = 25929 }, -] - -[[package]] -name = "ipython" -version = "8.37.0" -source = { registry = "https://pypi.org/simple" } -resolution-markers = [ - "python_full_version < '3.11' and sys_platform == 'linux'", - "python_full_version < '3.11' and sys_platform != 'linux'", -] -dependencies = [ - { name = "colorama", marker = "python_full_version < '3.11' and sys_platform == 'win32'" }, - { name = "decorator", marker = "python_full_version < '3.11'" }, - { name = "exceptiongroup", marker = "python_full_version < '3.11'" }, - { name = "jedi", marker = "python_full_version < '3.11'" }, - { name = "matplotlib-inline", marker = "python_full_version < '3.11'" }, - { name = "pexpect", marker = "python_full_version < '3.11' and sys_platform != 'emscripten' and sys_platform != 'win32'" }, - { name = "prompt-toolkit", marker = "python_full_version < '3.11'" }, - { name = "pygments", marker = "python_full_version < '3.11'" }, - { name = "stack-data", marker = "python_full_version < '3.11'" }, - { name = "traitlets", marker = "python_full_version < '3.11'" }, - { name = "typing-extensions", marker = "python_full_version < '3.11'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/85/31/10ac88f3357fc276dc8a64e8880c82e80e7459326ae1d0a211b40abf6665/ipython-8.37.0.tar.gz", hash = "sha256:ca815841e1a41a1e6b73a0b08f3038af9b2252564d01fc405356d34033012216", size = 5606088 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/91/d0/274fbf7b0b12643cbbc001ce13e6a5b1607ac4929d1b11c72460152c9fc3/ipython-8.37.0-py3-none-any.whl", hash = "sha256:ed87326596b878932dbcb171e3e698845434d8c61b8d8cd474bf663041a9dcf2", size = 831864 }, -] - -[[package]] -name = "ipython" -version = "9.8.0" -source = { registry = "https://pypi.org/simple" } -resolution-markers = [ - "python_full_version >= '3.14' and sys_platform == 'linux'", - "python_full_version == '3.13.*' and sys_platform == 'linux'", - "python_full_version >= '3.14' and sys_platform != 'linux'", - "python_full_version == '3.13.*' and sys_platform != 'linux'", - "python_full_version == '3.12.*' and sys_platform == 'linux'", - "python_full_version == '3.12.*' and sys_platform != 'linux'", - "python_full_version == '3.11.*' and sys_platform == 'linux'", - "python_full_version == '3.11.*' and sys_platform != 'linux'", -] -dependencies = [ - { name = "colorama", marker = "python_full_version >= '3.11' and sys_platform == 'win32'" }, - { name = "decorator", marker = "python_full_version >= '3.11'" }, - { name = "ipython-pygments-lexers", marker = "python_full_version >= '3.11'" }, - { name = "jedi", marker = "python_full_version >= '3.11'" }, - { name = "matplotlib-inline", marker = "python_full_version >= '3.11'" }, - { name = "pexpect", marker = "python_full_version >= '3.11' and sys_platform != 'emscripten' and sys_platform != 'win32'" }, - { name = "prompt-toolkit", marker = "python_full_version >= '3.11'" }, - { name = "pygments", marker = "python_full_version >= '3.11'" }, - { name = "stack-data", marker = "python_full_version >= '3.11'" }, - { name = "traitlets", marker = "python_full_version >= '3.11'" }, - { name = "typing-extensions", marker = "python_full_version == '3.11.*'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/12/51/a703c030f4928646d390b4971af4938a1b10c9dfce694f0d99a0bb073cb2/ipython-9.8.0.tar.gz", hash = "sha256:8e4ce129a627eb9dd221c41b1d2cdaed4ef7c9da8c17c63f6f578fe231141f83", size = 4424940 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/f1/df/8ee1c5dd1e3308b5d5b2f2dfea323bb2f3827da8d654abb6642051199049/ipython-9.8.0-py3-none-any.whl", hash = "sha256:ebe6d1d58d7d988fbf23ff8ff6d8e1622cfdb194daf4b7b73b792c4ec3b85385", size = 621374 }, -] - -[[package]] -name = "ipython-pygments-lexers" -version = "1.1.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "pygments", marker = "python_full_version >= '3.11'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/ef/4c/5dd1d8af08107f88c7f741ead7a40854b8ac24ddf9ae850afbcf698aa552/ipython_pygments_lexers-1.1.1.tar.gz", hash = "sha256:09c0138009e56b6854f9535736f4171d855c8c08a563a0dcd8022f78355c7e81", size = 8393 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/d9/33/1f075bf72b0b747cb3288d011319aaf64083cf2efef8354174e3ed4540e2/ipython_pygments_lexers-1.1.1-py3-none-any.whl", hash = "sha256:a9462224a505ade19a605f71f8fa63c2048833ce50abc86768a0d81d876dc81c", size = 8074 }, -] - -[[package]] -name = "jedi" -version = "0.19.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "parso" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/72/3a/79a912fbd4d8dd6fbb02bf69afd3bb72cf0c729bb3063c6f4498603db17a/jedi-0.19.2.tar.gz", hash = "sha256:4770dc3de41bde3966b02eb84fbcf557fb33cce26ad23da12c742fb50ecb11f0", size = 1231287 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/c0/5a/9cac0c82afec3d09ccd97c8b6502d48f165f9124db81b4bcb90b4af974ee/jedi-0.19.2-py2.py3-none-any.whl", hash = "sha256:a8ef22bde8490f57fe5c7681a3c83cb58874daf72b4784de3cce5b6ef6edb5b9", size = 1572278 }, -] - -[[package]] -name = "jinja2" -version = "3.1.6" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "markupsafe" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/df/bf/f7da0350254c0ed7c72f3e33cef02e048281fec7ecec5f032d4aac52226b/jinja2-3.1.6.tar.gz", hash = "sha256:0137fb05990d35f1275a587e9aee6d56da821fc83491a0fb838183be43f66d6d", size = 245115 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/62/a1/3d680cbfd5f4b8f15abc1d571870c5fc3e594bb582bc3b64ea099db13e56/jinja2-3.1.6-py3-none-any.whl", hash = "sha256:85ece4451f492d0c13c5dd7c13a64681a86afae63a5f347908daf103ce6d2f67", size = 134899 }, -] - -[[package]] -name = "jiwer" -version = "3.1.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "click" }, - { name = "rapidfuzz" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/85/3e/71b95cf0e2179fb5de8744a79fd36c8bd4e02e1803129a16d423884b6654/jiwer-3.1.0.tar.gz", hash = "sha256:dc492d09e570f1baba98c76aba09baf8e09c06e6808a4ba412dd4bde67fb79ac", size = 103187 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/ba/f4/35634d9eeff3b0bab51f5b9474ee569b1186bf29cf0d9d67b84acc80c53d/jiwer-3.1.0-py3-none-any.whl", hash = "sha256:5a14b5bba4692e1946ca3c6946435f7d90b1b526076ccb6c12be763e2146237d", size = 22303 }, -] - -[[package]] -name = "joblib" -version = "1.5.3" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/41/f2/d34e8b3a08a9cc79a50b2208a93dce981fe615b64d5a4d4abee421d898df/joblib-1.5.3.tar.gz", hash = "sha256:8561a3269e6801106863fd0d6d84bb737be9e7631e33aaed3fb9ce5953688da3", size = 331603 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/7b/91/984aca2ec129e2757d1e4e3c81c3fcda9d0f85b74670a094cc443d9ee949/joblib-1.5.3-py3-none-any.whl", hash = "sha256:5fc3c5039fc5ca8c0276333a188bbd59d6b7ab37fe6632daa76bc7f9ec18e713", size = 309071 }, -] - -[[package]] -name = "kaldi-python-io" -version = "1.2.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "numpy" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/80/45/e3e542ffa8970ebd782fcece35e2295de9c60e8c396c2c1a403410d1b24e/kaldi-python-io-1.2.2.tar.gz", hash = "sha256:4ebb4029c6c58296cc0abf96edff02832ba341d290ed37624a8d00105f0f7c00", size = 8814 } - -[[package]] -name = "kaldialign" -version = "0.9.1" -source = { registry = "https://pypi.org/simple" } -wheels = [ - { url = "https://files.pythonhosted.org/packages/d2/a0/f94eee8b64d7d20bac0725baf8221223e033a76387f9677845893dbcea4d/kaldialign-0.9.1-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:026f713cf18e272ef35602acb55294a92a7245ff94a3c45dcce0c3090f20d115", size = 113321 }, - { url = "https://files.pythonhosted.org/packages/99/d0/6791cf1e0aac7e2dcb4430397b48046654a1109dbe8ab40f92c0c3cd07c7/kaldialign-0.9.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8d75a3a1716b299225e8b6f6ec2f6300dccae0ee35cf5bd2b40e493a89f13bb8", size = 87360 }, - { url = "https://files.pythonhosted.org/packages/c7/b0/2d074f332743993f0938021d730fa1f6735f4312ada4245d4ad97295c8a1/kaldialign-0.9.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0a3f7af4ee8180d05cba148de79a6c81ea7f17822bb31707c1b00b9d9d5b5d50", size = 91767 }, - { url = "https://files.pythonhosted.org/packages/61/ec/d62225522e188bc70950f8c0a07a5c945018f39cb3660ce690c7b2e5ddd1/kaldialign-0.9.1-cp310-cp310-win32.whl", hash = "sha256:154725a816022632f166ea842c7becd7236f6d7dffc82f6549f36a5c940c8a1e", size = 61726 }, - { url = "https://files.pythonhosted.org/packages/85/57/d3e270229acd86207d06e1ed5ef0f5aba572877cdd671004f4cd6656b60d/kaldialign-0.9.1-cp310-cp310-win_amd64.whl", hash = "sha256:80c17a8723589291f68daece92f25c98bad21d49ecaf8d64b158ba343e79f5bd", size = 70277 }, - { url = "https://files.pythonhosted.org/packages/1d/6f/5d4ecd96842e5a6ba1288095daa35264e79d3e72fac326a7a2a80d1220f9/kaldialign-0.9.1-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:14991bef6400fe6f93f6ebbc5923b9baa5bb5f979baa66cb5e116496b030192d", size = 113339 }, - { url = "https://files.pythonhosted.org/packages/a5/d9/1effbcd3e7b8a5975d9661ea232f28d1694748676184e5ce21a61398a940/kaldialign-0.9.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:886386e06bee54929b5a56ac7693bcd22dd6568e257a0e15a58946eb5cb64bd9", size = 87375 }, - { url = "https://files.pythonhosted.org/packages/86/2c/6adf305326f48bac470985069036e2eb1ed198fcbd8a5ca04d8055c4373f/kaldialign-0.9.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:49a7eb2790acfa9e50409c4253fd785e0ecb0a23cc123bc97f1b06caf6382f8f", size = 91755 }, - { url = "https://files.pythonhosted.org/packages/5a/60/ddd1d78f85bd093d92eee78a4ae38dc778be8b73918a44153767c5d5c85d/kaldialign-0.9.1-cp311-cp311-win32.whl", hash = "sha256:5b03ee410c8829db325cc4b64a3cebb0614b9a500e41335a0e70604a227c5a29", size = 61841 }, - { url = "https://files.pythonhosted.org/packages/76/41/7b0d974b0104d232e69bd30350ce92d7de60ae65f33bbe92ad5d834d81aa/kaldialign-0.9.1-cp311-cp311-win_amd64.whl", hash = "sha256:c31e613c81fcbdd18496ed79066d44b1a44273eab223327c51e03964f5f561f2", size = 70082 }, - { url = "https://files.pythonhosted.org/packages/42/99/207f45fd2b02f431c40e620cc41b3c6af8078a192482668840df52dad592/kaldialign-0.9.1-cp312-cp312-macosx_10_9_universal2.whl", hash = "sha256:7112ff9f52c960825293cd3e8586094d700344c625cd4e75138408ace780402e", size = 115212 }, - { url = "https://files.pythonhosted.org/packages/f0/e4/b2c2b4dfb88677b1868ed9366ed415d93360c47615a53e189d375b986de4/kaldialign-0.9.1-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:cdaf85389756a1caf6ad95754480768195aea30e57f4ce741b6d385a84c6efd9", size = 87336 }, - { url = "https://files.pythonhosted.org/packages/a1/25/acca269c3b9e219d4611c7f9b966e884521b61c899750343b3548bc6393b/kaldialign-0.9.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e9dc8455832caf4316a7674676098e4e9cc2c0f50428c7105e24c1c767bebd6e", size = 91977 }, - { url = "https://files.pythonhosted.org/packages/87/56/4a9cd8468c6a3afd32ee29059e36e9b28dab41c37e49be979def88697166/kaldialign-0.9.1-cp312-cp312-win32.whl", hash = "sha256:c28620c352ecfb016b35579c3b52a4d049aba54c1e9a448d4bef88de1a8566c8", size = 62314 }, - { url = "https://files.pythonhosted.org/packages/83/b2/aac02ff83128f8068898ad37eb0077cecad0548d1207ec0ff5f61107f8e2/kaldialign-0.9.1-cp312-cp312-win_amd64.whl", hash = "sha256:c6090ef4ebdaf95fa78b46d7eda05b7367d8d889228a67696217487d4ceb783c", size = 70822 }, - { url = "https://files.pythonhosted.org/packages/da/32/a565d6828502782fc31d07d7676dc39adb16df5b64eec11f2225ac8a2d21/kaldialign-0.9.1-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9b6e8122ca31f1bab2690d225f74cf0493387c4e6794ff02d2425d2800117fd4", size = 92152 }, - { url = "https://files.pythonhosted.org/packages/c4/e1/940a24ce5f164fa53fcb07ff9889dbb2ad131784e3fab4085ad6f3e13b5c/kaldialign-0.9.1-cp313-cp313-win_amd64.whl", hash = "sha256:2bd9647cc76d294fa0e35cc27f2fb96f88077c5a03f07475fd80a7216f6b334b", size = 74749 }, -] - -[[package]] -name = "kiwisolver" -version = "1.4.9" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/5c/3c/85844f1b0feb11ee581ac23fe5fce65cd049a200c1446708cc1b7f922875/kiwisolver-1.4.9.tar.gz", hash = "sha256:c3b22c26c6fd6811b0ae8363b95ca8ce4ea3c202d3d0975b2914310ceb1bcc4d", size = 97564 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/c6/5d/8ce64e36d4e3aac5ca96996457dcf33e34e6051492399a3f1fec5657f30b/kiwisolver-1.4.9-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:b4b4d74bda2b8ebf4da5bd42af11d02d04428b2c32846e4c2c93219df8a7987b", size = 124159 }, - { url = "https://files.pythonhosted.org/packages/96/1e/22f63ec454874378175a5f435d6ea1363dd33fb2af832c6643e4ccea0dc8/kiwisolver-1.4.9-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:fb3b8132019ea572f4611d770991000d7f58127560c4889729248eb5852a102f", size = 66578 }, - { url = "https://files.pythonhosted.org/packages/41/4c/1925dcfff47a02d465121967b95151c82d11027d5ec5242771e580e731bd/kiwisolver-1.4.9-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:84fd60810829c27ae375114cd379da1fa65e6918e1da405f356a775d49a62bcf", size = 65312 }, - { url = "https://files.pythonhosted.org/packages/d4/42/0f333164e6307a0687d1eb9ad256215aae2f4bd5d28f4653d6cd319a3ba3/kiwisolver-1.4.9-cp310-cp310-manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:b78efa4c6e804ecdf727e580dbb9cba85624d2e1c6b5cb059c66290063bd99a9", size = 1628458 }, - { url = "https://files.pythonhosted.org/packages/86/b6/2dccb977d651943995a90bfe3495c2ab2ba5cd77093d9f2318a20c9a6f59/kiwisolver-1.4.9-cp310-cp310-manylinux_2_24_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:d4efec7bcf21671db6a3294ff301d2fc861c31faa3c8740d1a94689234d1b415", size = 1225640 }, - { url = "https://files.pythonhosted.org/packages/50/2b/362ebd3eec46c850ccf2bfe3e30f2fc4c008750011f38a850f088c56a1c6/kiwisolver-1.4.9-cp310-cp310-manylinux_2_24_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:90f47e70293fc3688b71271100a1a5453aa9944a81d27ff779c108372cf5567b", size = 1244074 }, - { url = "https://files.pythonhosted.org/packages/6f/bb/f09a1e66dab8984773d13184a10a29fe67125337649d26bdef547024ed6b/kiwisolver-1.4.9-cp310-cp310-manylinux_2_24_s390x.manylinux_2_28_s390x.whl", hash = "sha256:8fdca1def57a2e88ef339de1737a1449d6dbf5fab184c54a1fca01d541317154", size = 1293036 }, - { url = "https://files.pythonhosted.org/packages/ea/01/11ecf892f201cafda0f68fa59212edaea93e96c37884b747c181303fccd1/kiwisolver-1.4.9-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:9cf554f21be770f5111a1690d42313e140355e687e05cf82cb23d0a721a64a48", size = 2175310 }, - { url = "https://files.pythonhosted.org/packages/7f/5f/bfe11d5b934f500cc004314819ea92427e6e5462706a498c1d4fc052e08f/kiwisolver-1.4.9-cp310-cp310-musllinux_1_2_ppc64le.whl", hash = "sha256:fc1795ac5cd0510207482c3d1d3ed781143383b8cfd36f5c645f3897ce066220", size = 2270943 }, - { url = "https://files.pythonhosted.org/packages/3d/de/259f786bf71f1e03e73d87e2db1a9a3bcab64d7b4fd780167123161630ad/kiwisolver-1.4.9-cp310-cp310-musllinux_1_2_s390x.whl", hash = "sha256:ccd09f20ccdbbd341b21a67ab50a119b64a403b09288c27481575105283c1586", size = 2440488 }, - { url = "https://files.pythonhosted.org/packages/1b/76/c989c278faf037c4d3421ec07a5c452cd3e09545d6dae7f87c15f54e4edf/kiwisolver-1.4.9-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:540c7c72324d864406a009d72f5d6856f49693db95d1fbb46cf86febef873634", size = 2246787 }, - { url = "https://files.pythonhosted.org/packages/a2/55/c2898d84ca440852e560ca9f2a0d28e6e931ac0849b896d77231929900e7/kiwisolver-1.4.9-cp310-cp310-win_amd64.whl", hash = "sha256:ede8c6d533bc6601a47ad4046080d36b8fc99f81e6f1c17b0ac3c2dc91ac7611", size = 73730 }, - { url = "https://files.pythonhosted.org/packages/e8/09/486d6ac523dd33b80b368247f238125d027964cfacb45c654841e88fb2ae/kiwisolver-1.4.9-cp310-cp310-win_arm64.whl", hash = "sha256:7b4da0d01ac866a57dd61ac258c5607b4cd677f63abaec7b148354d2b2cdd536", size = 65036 }, - { url = "https://files.pythonhosted.org/packages/6f/ab/c80b0d5a9d8a1a65f4f815f2afff9798b12c3b9f31f1d304dd233dd920e2/kiwisolver-1.4.9-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:eb14a5da6dc7642b0f3a18f13654847cd8b7a2550e2645a5bda677862b03ba16", size = 124167 }, - { url = "https://files.pythonhosted.org/packages/a0/c0/27fe1a68a39cf62472a300e2879ffc13c0538546c359b86f149cc19f6ac3/kiwisolver-1.4.9-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:39a219e1c81ae3b103643d2aedb90f1ef22650deb266ff12a19e7773f3e5f089", size = 66579 }, - { url = "https://files.pythonhosted.org/packages/31/a2/a12a503ac1fd4943c50f9822678e8015a790a13b5490354c68afb8489814/kiwisolver-1.4.9-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:2405a7d98604b87f3fc28b1716783534b1b4b8510d8142adca34ee0bc3c87543", size = 65309 }, - { url = "https://files.pythonhosted.org/packages/66/e1/e533435c0be77c3f64040d68d7a657771194a63c279f55573188161e81ca/kiwisolver-1.4.9-cp311-cp311-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:dc1ae486f9abcef254b5618dfb4113dd49f94c68e3e027d03cf0143f3f772b61", size = 1435596 }, - { url = "https://files.pythonhosted.org/packages/67/1e/51b73c7347f9aabdc7215aa79e8b15299097dc2f8e67dee2b095faca9cb0/kiwisolver-1.4.9-cp311-cp311-manylinux_2_24_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:8a1f570ce4d62d718dce3f179ee78dac3b545ac16c0c04bb363b7607a949c0d1", size = 1246548 }, - { url = "https://files.pythonhosted.org/packages/21/aa/72a1c5d1e430294f2d32adb9542719cfb441b5da368d09d268c7757af46c/kiwisolver-1.4.9-cp311-cp311-manylinux_2_24_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:cb27e7b78d716c591e88e0a09a2139c6577865d7f2e152488c2cc6257f460872", size = 1263618 }, - { url = "https://files.pythonhosted.org/packages/a3/af/db1509a9e79dbf4c260ce0cfa3903ea8945f6240e9e59d1e4deb731b1a40/kiwisolver-1.4.9-cp311-cp311-manylinux_2_24_s390x.manylinux_2_28_s390x.whl", hash = "sha256:15163165efc2f627eb9687ea5f3a28137217d217ac4024893d753f46bce9de26", size = 1317437 }, - { url = "https://files.pythonhosted.org/packages/e0/f2/3ea5ee5d52abacdd12013a94130436e19969fa183faa1e7c7fbc89e9a42f/kiwisolver-1.4.9-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:bdee92c56a71d2b24c33a7d4c2856bd6419d017e08caa7802d2963870e315028", size = 2195742 }, - { url = "https://files.pythonhosted.org/packages/6f/9b/1efdd3013c2d9a2566aa6a337e9923a00590c516add9a1e89a768a3eb2fc/kiwisolver-1.4.9-cp311-cp311-musllinux_1_2_ppc64le.whl", hash = "sha256:412f287c55a6f54b0650bd9b6dce5aceddb95864a1a90c87af16979d37c89771", size = 2290810 }, - { url = "https://files.pythonhosted.org/packages/fb/e5/cfdc36109ae4e67361f9bc5b41323648cb24a01b9ade18784657e022e65f/kiwisolver-1.4.9-cp311-cp311-musllinux_1_2_s390x.whl", hash = "sha256:2c93f00dcba2eea70af2be5f11a830a742fe6b579a1d4e00f47760ef13be247a", size = 2461579 }, - { url = "https://files.pythonhosted.org/packages/62/86/b589e5e86c7610842213994cdea5add00960076bef4ae290c5fa68589cac/kiwisolver-1.4.9-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:f117e1a089d9411663a3207ba874f31be9ac8eaa5b533787024dc07aeb74f464", size = 2268071 }, - { url = "https://files.pythonhosted.org/packages/3b/c6/f8df8509fd1eee6c622febe54384a96cfaf4d43bf2ccec7a0cc17e4715c9/kiwisolver-1.4.9-cp311-cp311-win_amd64.whl", hash = "sha256:be6a04e6c79819c9a8c2373317d19a96048e5a3f90bec587787e86a1153883c2", size = 73840 }, - { url = "https://files.pythonhosted.org/packages/e2/2d/16e0581daafd147bc11ac53f032a2b45eabac897f42a338d0a13c1e5c436/kiwisolver-1.4.9-cp311-cp311-win_arm64.whl", hash = "sha256:0ae37737256ba2de764ddc12aed4956460277f00c4996d51a197e72f62f5eec7", size = 65159 }, - { url = "https://files.pythonhosted.org/packages/86/c9/13573a747838aeb1c76e3267620daa054f4152444d1f3d1a2324b78255b5/kiwisolver-1.4.9-cp312-cp312-macosx_10_13_universal2.whl", hash = "sha256:ac5a486ac389dddcc5bef4f365b6ae3ffff2c433324fb38dd35e3fab7c957999", size = 123686 }, - { url = "https://files.pythonhosted.org/packages/51/ea/2ecf727927f103ffd1739271ca19c424d0e65ea473fbaeea1c014aea93f6/kiwisolver-1.4.9-cp312-cp312-macosx_10_13_x86_64.whl", hash = "sha256:f2ba92255faa7309d06fe44c3a4a97efe1c8d640c2a79a5ef728b685762a6fd2", size = 66460 }, - { url = "https://files.pythonhosted.org/packages/5b/5a/51f5464373ce2aeb5194508298a508b6f21d3867f499556263c64c621914/kiwisolver-1.4.9-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:4a2899935e724dd1074cb568ce7ac0dce28b2cd6ab539c8e001a8578eb106d14", size = 64952 }, - { url = "https://files.pythonhosted.org/packages/70/90/6d240beb0f24b74371762873e9b7f499f1e02166a2d9c5801f4dbf8fa12e/kiwisolver-1.4.9-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:f6008a4919fdbc0b0097089f67a1eb55d950ed7e90ce2cc3e640abadd2757a04", size = 1474756 }, - { url = "https://files.pythonhosted.org/packages/12/42/f36816eaf465220f683fb711efdd1bbf7a7005a2473d0e4ed421389bd26c/kiwisolver-1.4.9-cp312-cp312-manylinux_2_24_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:67bb8b474b4181770f926f7b7d2f8c0248cbcb78b660fdd41a47054b28d2a752", size = 1276404 }, - { url = "https://files.pythonhosted.org/packages/2e/64/bc2de94800adc830c476dce44e9b40fd0809cddeef1fde9fcf0f73da301f/kiwisolver-1.4.9-cp312-cp312-manylinux_2_24_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:2327a4a30d3ee07d2fbe2e7933e8a37c591663b96ce42a00bc67461a87d7df77", size = 1294410 }, - { url = "https://files.pythonhosted.org/packages/5f/42/2dc82330a70aa8e55b6d395b11018045e58d0bb00834502bf11509f79091/kiwisolver-1.4.9-cp312-cp312-manylinux_2_24_s390x.manylinux_2_28_s390x.whl", hash = "sha256:7a08b491ec91b1d5053ac177afe5290adacf1f0f6307d771ccac5de30592d198", size = 1343631 }, - { url = "https://files.pythonhosted.org/packages/22/fd/f4c67a6ed1aab149ec5a8a401c323cee7a1cbe364381bb6c9c0d564e0e20/kiwisolver-1.4.9-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:d8fc5c867c22b828001b6a38d2eaeb88160bf5783c6cb4a5e440efc981ce286d", size = 2224963 }, - { url = "https://files.pythonhosted.org/packages/45/aa/76720bd4cb3713314677d9ec94dcc21ced3f1baf4830adde5bb9b2430a5f/kiwisolver-1.4.9-cp312-cp312-musllinux_1_2_ppc64le.whl", hash = "sha256:3b3115b2581ea35bb6d1f24a4c90af37e5d9b49dcff267eeed14c3893c5b86ab", size = 2321295 }, - { url = "https://files.pythonhosted.org/packages/80/19/d3ec0d9ab711242f56ae0dc2fc5d70e298bb4a1f9dfab44c027668c673a1/kiwisolver-1.4.9-cp312-cp312-musllinux_1_2_s390x.whl", hash = "sha256:858e4c22fb075920b96a291928cb7dea5644e94c0ee4fcd5af7e865655e4ccf2", size = 2487987 }, - { url = "https://files.pythonhosted.org/packages/39/e9/61e4813b2c97e86b6fdbd4dd824bf72d28bcd8d4849b8084a357bc0dd64d/kiwisolver-1.4.9-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:ed0fecd28cc62c54b262e3736f8bb2512d8dcfdc2bcf08be5f47f96bf405b145", size = 2291817 }, - { url = "https://files.pythonhosted.org/packages/a0/41/85d82b0291db7504da3c2defe35c9a8a5c9803a730f297bd823d11d5fb77/kiwisolver-1.4.9-cp312-cp312-win_amd64.whl", hash = "sha256:f68208a520c3d86ea51acf688a3e3002615a7f0238002cccc17affecc86a8a54", size = 73895 }, - { url = "https://files.pythonhosted.org/packages/e2/92/5f3068cf15ee5cb624a0c7596e67e2a0bb2adee33f71c379054a491d07da/kiwisolver-1.4.9-cp312-cp312-win_arm64.whl", hash = "sha256:2c1a4f57df73965f3f14df20b80ee29e6a7930a57d2d9e8491a25f676e197c60", size = 64992 }, - { url = "https://files.pythonhosted.org/packages/31/c1/c2686cda909742ab66c7388e9a1a8521a59eb89f8bcfbee28fc980d07e24/kiwisolver-1.4.9-cp313-cp313-macosx_10_13_universal2.whl", hash = "sha256:a5d0432ccf1c7ab14f9949eec60c5d1f924f17c037e9f8b33352fa05799359b8", size = 123681 }, - { url = "https://files.pythonhosted.org/packages/ca/f0/f44f50c9f5b1a1860261092e3bc91ecdc9acda848a8b8c6abfda4a24dd5c/kiwisolver-1.4.9-cp313-cp313-macosx_10_13_x86_64.whl", hash = "sha256:efb3a45b35622bb6c16dbfab491a8f5a391fe0e9d45ef32f4df85658232ca0e2", size = 66464 }, - { url = "https://files.pythonhosted.org/packages/2d/7a/9d90a151f558e29c3936b8a47ac770235f436f2120aca41a6d5f3d62ae8d/kiwisolver-1.4.9-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:1a12cf6398e8a0a001a059747a1cbf24705e18fe413bc22de7b3d15c67cffe3f", size = 64961 }, - { url = "https://files.pythonhosted.org/packages/e9/e9/f218a2cb3a9ffbe324ca29a9e399fa2d2866d7f348ec3a88df87fc248fc5/kiwisolver-1.4.9-cp313-cp313-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:b67e6efbf68e077dd71d1a6b37e43e1a99d0bff1a3d51867d45ee8908b931098", size = 1474607 }, - { url = "https://files.pythonhosted.org/packages/d9/28/aac26d4c882f14de59041636292bc838db8961373825df23b8eeb807e198/kiwisolver-1.4.9-cp313-cp313-manylinux_2_24_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:5656aa670507437af0207645273ccdfee4f14bacd7f7c67a4306d0dcaeaf6eed", size = 1276546 }, - { url = "https://files.pythonhosted.org/packages/8b/ad/8bfc1c93d4cc565e5069162f610ba2f48ff39b7de4b5b8d93f69f30c4bed/kiwisolver-1.4.9-cp313-cp313-manylinux_2_24_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:bfc08add558155345129c7803b3671cf195e6a56e7a12f3dde7c57d9b417f525", size = 1294482 }, - { url = "https://files.pythonhosted.org/packages/da/f1/6aca55ff798901d8ce403206d00e033191f63d82dd708a186e0ed2067e9c/kiwisolver-1.4.9-cp313-cp313-manylinux_2_24_s390x.manylinux_2_28_s390x.whl", hash = "sha256:40092754720b174e6ccf9e845d0d8c7d8e12c3d71e7fc35f55f3813e96376f78", size = 1343720 }, - { url = "https://files.pythonhosted.org/packages/d1/91/eed031876c595c81d90d0f6fc681ece250e14bf6998c3d7c419466b523b7/kiwisolver-1.4.9-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:497d05f29a1300d14e02e6441cf0f5ee81c1ff5a304b0d9fb77423974684e08b", size = 2224907 }, - { url = "https://files.pythonhosted.org/packages/e9/ec/4d1925f2e49617b9cca9c34bfa11adefad49d00db038e692a559454dfb2e/kiwisolver-1.4.9-cp313-cp313-musllinux_1_2_ppc64le.whl", hash = "sha256:bdd1a81a1860476eb41ac4bc1e07b3f07259e6d55bbf739b79c8aaedcf512799", size = 2321334 }, - { url = "https://files.pythonhosted.org/packages/43/cb/450cd4499356f68802750c6ddc18647b8ea01ffa28f50d20598e0befe6e9/kiwisolver-1.4.9-cp313-cp313-musllinux_1_2_s390x.whl", hash = "sha256:e6b93f13371d341afee3be9f7c5964e3fe61d5fa30f6a30eb49856935dfe4fc3", size = 2488313 }, - { url = "https://files.pythonhosted.org/packages/71/67/fc76242bd99f885651128a5d4fa6083e5524694b7c88b489b1b55fdc491d/kiwisolver-1.4.9-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:d75aa530ccfaa593da12834b86a0724f58bff12706659baa9227c2ccaa06264c", size = 2291970 }, - { url = "https://files.pythonhosted.org/packages/75/bd/f1a5d894000941739f2ae1b65a32892349423ad49c2e6d0771d0bad3fae4/kiwisolver-1.4.9-cp313-cp313-win_amd64.whl", hash = "sha256:dd0a578400839256df88c16abddf9ba14813ec5f21362e1fe65022e00c883d4d", size = 73894 }, - { url = "https://files.pythonhosted.org/packages/95/38/dce480814d25b99a391abbddadc78f7c117c6da34be68ca8b02d5848b424/kiwisolver-1.4.9-cp313-cp313-win_arm64.whl", hash = "sha256:d4188e73af84ca82468f09cadc5ac4db578109e52acb4518d8154698d3a87ca2", size = 64995 }, - { url = "https://files.pythonhosted.org/packages/e2/37/7d218ce5d92dadc5ebdd9070d903e0c7cf7edfe03f179433ac4d13ce659c/kiwisolver-1.4.9-cp313-cp313t-macosx_10_13_universal2.whl", hash = "sha256:5a0f2724dfd4e3b3ac5a82436a8e6fd16baa7d507117e4279b660fe8ca38a3a1", size = 126510 }, - { url = "https://files.pythonhosted.org/packages/23/b0/e85a2b48233daef4b648fb657ebbb6f8367696a2d9548a00b4ee0eb67803/kiwisolver-1.4.9-cp313-cp313t-macosx_10_13_x86_64.whl", hash = "sha256:1b11d6a633e4ed84fc0ddafd4ebfd8ea49b3f25082c04ad12b8315c11d504dc1", size = 67903 }, - { url = "https://files.pythonhosted.org/packages/44/98/f2425bc0113ad7de24da6bb4dae1343476e95e1d738be7c04d31a5d037fd/kiwisolver-1.4.9-cp313-cp313t-macosx_11_0_arm64.whl", hash = "sha256:61874cdb0a36016354853593cffc38e56fc9ca5aa97d2c05d3dcf6922cd55a11", size = 66402 }, - { url = "https://files.pythonhosted.org/packages/98/d8/594657886df9f34c4177cc353cc28ca7e6e5eb562d37ccc233bff43bbe2a/kiwisolver-1.4.9-cp313-cp313t-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:60c439763a969a6af93b4881db0eed8fadf93ee98e18cbc35bc8da868d0c4f0c", size = 1582135 }, - { url = "https://files.pythonhosted.org/packages/5c/c6/38a115b7170f8b306fc929e166340c24958347308ea3012c2b44e7e295db/kiwisolver-1.4.9-cp313-cp313t-manylinux_2_24_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:92a2f997387a1b79a75e7803aa7ded2cfbe2823852ccf1ba3bcf613b62ae3197", size = 1389409 }, - { url = "https://files.pythonhosted.org/packages/bf/3b/e04883dace81f24a568bcee6eb3001da4ba05114afa622ec9b6fafdc1f5e/kiwisolver-1.4.9-cp313-cp313t-manylinux_2_24_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:a31d512c812daea6d8b3be3b2bfcbeb091dbb09177706569bcfc6240dcf8b41c", size = 1401763 }, - { url = "https://files.pythonhosted.org/packages/9f/80/20ace48e33408947af49d7d15c341eaee69e4e0304aab4b7660e234d6288/kiwisolver-1.4.9-cp313-cp313t-manylinux_2_24_s390x.manylinux_2_28_s390x.whl", hash = "sha256:52a15b0f35dad39862d376df10c5230155243a2c1a436e39eb55623ccbd68185", size = 1453643 }, - { url = "https://files.pythonhosted.org/packages/64/31/6ce4380a4cd1f515bdda976a1e90e547ccd47b67a1546d63884463c92ca9/kiwisolver-1.4.9-cp313-cp313t-musllinux_1_2_aarch64.whl", hash = "sha256:a30fd6fdef1430fd9e1ba7b3398b5ee4e2887783917a687d86ba69985fb08748", size = 2330818 }, - { url = "https://files.pythonhosted.org/packages/fa/e9/3f3fcba3bcc7432c795b82646306e822f3fd74df0ee81f0fa067a1f95668/kiwisolver-1.4.9-cp313-cp313t-musllinux_1_2_ppc64le.whl", hash = "sha256:cc9617b46837c6468197b5945e196ee9ca43057bb7d9d1ae688101e4e1dddf64", size = 2419963 }, - { url = "https://files.pythonhosted.org/packages/99/43/7320c50e4133575c66e9f7dadead35ab22d7c012a3b09bb35647792b2a6d/kiwisolver-1.4.9-cp313-cp313t-musllinux_1_2_s390x.whl", hash = "sha256:0ab74e19f6a2b027ea4f845a78827969af45ce790e6cb3e1ebab71bdf9f215ff", size = 2594639 }, - { url = "https://files.pythonhosted.org/packages/65/d6/17ae4a270d4a987ef8a385b906d2bdfc9fce502d6dc0d3aea865b47f548c/kiwisolver-1.4.9-cp313-cp313t-musllinux_1_2_x86_64.whl", hash = "sha256:dba5ee5d3981160c28d5490f0d1b7ed730c22470ff7f6cc26cfcfaacb9896a07", size = 2391741 }, - { url = "https://files.pythonhosted.org/packages/2a/8f/8f6f491d595a9e5912971f3f863d81baddccc8a4d0c3749d6a0dd9ffc9df/kiwisolver-1.4.9-cp313-cp313t-win_arm64.whl", hash = "sha256:0749fd8f4218ad2e851e11cc4dc05c7cbc0cbc4267bdfdb31782e65aace4ee9c", size = 68646 }, - { url = "https://files.pythonhosted.org/packages/6b/32/6cc0fbc9c54d06c2969faa9c1d29f5751a2e51809dd55c69055e62d9b426/kiwisolver-1.4.9-cp314-cp314-macosx_10_13_universal2.whl", hash = "sha256:9928fe1eb816d11ae170885a74d074f57af3a0d65777ca47e9aeb854a1fba386", size = 123806 }, - { url = "https://files.pythonhosted.org/packages/b2/dd/2bfb1d4a4823d92e8cbb420fe024b8d2167f72079b3bb941207c42570bdf/kiwisolver-1.4.9-cp314-cp314-macosx_10_13_x86_64.whl", hash = "sha256:d0005b053977e7b43388ddec89fa567f43d4f6d5c2c0affe57de5ebf290dc552", size = 66605 }, - { url = "https://files.pythonhosted.org/packages/f7/69/00aafdb4e4509c2ca6064646cba9cd4b37933898f426756adb2cb92ebbed/kiwisolver-1.4.9-cp314-cp314-macosx_11_0_arm64.whl", hash = "sha256:2635d352d67458b66fd0667c14cb1d4145e9560d503219034a18a87e971ce4f3", size = 64925 }, - { url = "https://files.pythonhosted.org/packages/43/dc/51acc6791aa14e5cb6d8a2e28cefb0dc2886d8862795449d021334c0df20/kiwisolver-1.4.9-cp314-cp314-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:767c23ad1c58c9e827b649a9ab7809fd5fd9db266a9cf02b0e926ddc2c680d58", size = 1472414 }, - { url = "https://files.pythonhosted.org/packages/3d/bb/93fa64a81db304ac8a246f834d5094fae4b13baf53c839d6bb6e81177129/kiwisolver-1.4.9-cp314-cp314-manylinux_2_24_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:72d0eb9fba308b8311685c2268cf7d0a0639a6cd027d8128659f72bdd8a024b4", size = 1281272 }, - { url = "https://files.pythonhosted.org/packages/70/e6/6df102916960fb8d05069d4bd92d6d9a8202d5a3e2444494e7cd50f65b7a/kiwisolver-1.4.9-cp314-cp314-manylinux_2_24_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:f68e4f3eeca8fb22cc3d731f9715a13b652795ef657a13df1ad0c7dc0e9731df", size = 1298578 }, - { url = "https://files.pythonhosted.org/packages/7c/47/e142aaa612f5343736b087864dbaebc53ea8831453fb47e7521fa8658f30/kiwisolver-1.4.9-cp314-cp314-manylinux_2_24_s390x.manylinux_2_28_s390x.whl", hash = "sha256:d84cd4061ae292d8ac367b2c3fa3aad11cb8625a95d135fe93f286f914f3f5a6", size = 1345607 }, - { url = "https://files.pythonhosted.org/packages/54/89/d641a746194a0f4d1a3670fb900d0dbaa786fb98341056814bc3f058fa52/kiwisolver-1.4.9-cp314-cp314-musllinux_1_2_aarch64.whl", hash = "sha256:a60ea74330b91bd22a29638940d115df9dc00af5035a9a2a6ad9399ffb4ceca5", size = 2230150 }, - { url = "https://files.pythonhosted.org/packages/aa/6b/5ee1207198febdf16ac11f78c5ae40861b809cbe0e6d2a8d5b0b3044b199/kiwisolver-1.4.9-cp314-cp314-musllinux_1_2_ppc64le.whl", hash = "sha256:ce6a3a4e106cf35c2d9c4fa17c05ce0b180db622736845d4315519397a77beaf", size = 2325979 }, - { url = "https://files.pythonhosted.org/packages/fc/ff/b269eefd90f4ae14dcc74973d5a0f6d28d3b9bb1afd8c0340513afe6b39a/kiwisolver-1.4.9-cp314-cp314-musllinux_1_2_s390x.whl", hash = "sha256:77937e5e2a38a7b48eef0585114fe7930346993a88060d0bf886086d2aa49ef5", size = 2491456 }, - { url = "https://files.pythonhosted.org/packages/fc/d4/10303190bd4d30de547534601e259a4fbf014eed94aae3e5521129215086/kiwisolver-1.4.9-cp314-cp314-musllinux_1_2_x86_64.whl", hash = "sha256:24c175051354f4a28c5d6a31c93906dc653e2bf234e8a4bbfb964892078898ce", size = 2294621 }, - { url = "https://files.pythonhosted.org/packages/28/e0/a9a90416fce5c0be25742729c2ea52105d62eda6c4be4d803c2a7be1fa50/kiwisolver-1.4.9-cp314-cp314-win_amd64.whl", hash = "sha256:0763515d4df10edf6d06a3c19734e2566368980d21ebec439f33f9eb936c07b7", size = 75417 }, - { url = "https://files.pythonhosted.org/packages/1f/10/6949958215b7a9a264299a7db195564e87900f709db9245e4ebdd3c70779/kiwisolver-1.4.9-cp314-cp314-win_arm64.whl", hash = "sha256:0e4e2bf29574a6a7b7f6cb5fa69293b9f96c928949ac4a53ba3f525dffb87f9c", size = 66582 }, - { url = "https://files.pythonhosted.org/packages/ec/79/60e53067903d3bc5469b369fe0dfc6b3482e2133e85dae9daa9527535991/kiwisolver-1.4.9-cp314-cp314t-macosx_10_13_universal2.whl", hash = "sha256:d976bbb382b202f71c67f77b0ac11244021cfa3f7dfd9e562eefcea2df711548", size = 126514 }, - { url = "https://files.pythonhosted.org/packages/25/d1/4843d3e8d46b072c12a38c97c57fab4608d36e13fe47d47ee96b4d61ba6f/kiwisolver-1.4.9-cp314-cp314t-macosx_10_13_x86_64.whl", hash = "sha256:2489e4e5d7ef9a1c300a5e0196e43d9c739f066ef23270607d45aba368b91f2d", size = 67905 }, - { url = "https://files.pythonhosted.org/packages/8c/ae/29ffcbd239aea8b93108de1278271ae764dfc0d803a5693914975f200596/kiwisolver-1.4.9-cp314-cp314t-macosx_11_0_arm64.whl", hash = "sha256:e2ea9f7ab7fbf18fffb1b5434ce7c69a07582f7acc7717720f1d69f3e806f90c", size = 66399 }, - { url = "https://files.pythonhosted.org/packages/a1/ae/d7ba902aa604152c2ceba5d352d7b62106bedbccc8e95c3934d94472bfa3/kiwisolver-1.4.9-cp314-cp314t-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:b34e51affded8faee0dfdb705416153819d8ea9250bbbf7ea1b249bdeb5f1122", size = 1582197 }, - { url = "https://files.pythonhosted.org/packages/f2/41/27c70d427eddb8bc7e4f16420a20fefc6f480312122a59a959fdfe0445ad/kiwisolver-1.4.9-cp314-cp314t-manylinux_2_24_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:d8aacd3d4b33b772542b2e01beb50187536967b514b00003bdda7589722d2a64", size = 1390125 }, - { url = "https://files.pythonhosted.org/packages/41/42/b3799a12bafc76d962ad69083f8b43b12bf4fe78b097b12e105d75c9b8f1/kiwisolver-1.4.9-cp314-cp314t-manylinux_2_24_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:7cf974dd4e35fa315563ac99d6287a1024e4dc2077b8a7d7cd3d2fb65d283134", size = 1402612 }, - { url = "https://files.pythonhosted.org/packages/d2/b5/a210ea073ea1cfaca1bb5c55a62307d8252f531beb364e18aa1e0888b5a0/kiwisolver-1.4.9-cp314-cp314t-manylinux_2_24_s390x.manylinux_2_28_s390x.whl", hash = "sha256:85bd218b5ecfbee8c8a82e121802dcb519a86044c9c3b2e4aef02fa05c6da370", size = 1453990 }, - { url = "https://files.pythonhosted.org/packages/5f/ce/a829eb8c033e977d7ea03ed32fb3c1781b4fa0433fbadfff29e39c676f32/kiwisolver-1.4.9-cp314-cp314t-musllinux_1_2_aarch64.whl", hash = "sha256:0856e241c2d3df4efef7c04a1e46b1936b6120c9bcf36dd216e3acd84bc4fb21", size = 2331601 }, - { url = "https://files.pythonhosted.org/packages/e0/4b/b5e97eb142eb9cd0072dacfcdcd31b1c66dc7352b0f7c7255d339c0edf00/kiwisolver-1.4.9-cp314-cp314t-musllinux_1_2_ppc64le.whl", hash = "sha256:9af39d6551f97d31a4deebeac6f45b156f9755ddc59c07b402c148f5dbb6482a", size = 2422041 }, - { url = "https://files.pythonhosted.org/packages/40/be/8eb4cd53e1b85ba4edc3a9321666f12b83113a178845593307a3e7891f44/kiwisolver-1.4.9-cp314-cp314t-musllinux_1_2_s390x.whl", hash = "sha256:bb4ae2b57fc1d8cbd1cf7b1d9913803681ffa903e7488012be5b76dedf49297f", size = 2594897 }, - { url = "https://files.pythonhosted.org/packages/99/dd/841e9a66c4715477ea0abc78da039832fbb09dac5c35c58dc4c41a407b8a/kiwisolver-1.4.9-cp314-cp314t-musllinux_1_2_x86_64.whl", hash = "sha256:aedff62918805fb62d43a4aa2ecd4482c380dc76cd31bd7c8878588a61bd0369", size = 2391835 }, - { url = "https://files.pythonhosted.org/packages/0c/28/4b2e5c47a0da96896fdfdb006340ade064afa1e63675d01ea5ac222b6d52/kiwisolver-1.4.9-cp314-cp314t-win_amd64.whl", hash = "sha256:1fa333e8b2ce4d9660f2cda9c0e1b6bafcfb2457a9d259faa82289e73ec24891", size = 79988 }, - { url = "https://files.pythonhosted.org/packages/80/be/3578e8afd18c88cdf9cb4cffde75a96d2be38c5a903f1ed0ceec061bd09e/kiwisolver-1.4.9-cp314-cp314t-win_arm64.whl", hash = "sha256:4a48a2ce79d65d363597ef7b567ce3d14d68783d2b2263d98db3d9477805ba32", size = 70260 }, - { url = "https://files.pythonhosted.org/packages/a2/63/fde392691690f55b38d5dd7b3710f5353bf7a8e52de93a22968801ab8978/kiwisolver-1.4.9-pp310-pypy310_pp73-macosx_10_15_x86_64.whl", hash = "sha256:4d1d9e582ad4d63062d34077a9a1e9f3c34088a2ec5135b1f7190c07cf366527", size = 60183 }, - { url = "https://files.pythonhosted.org/packages/27/b1/6aad34edfdb7cced27f371866f211332bba215bfd918ad3322a58f480d8b/kiwisolver-1.4.9-pp310-pypy310_pp73-macosx_11_0_arm64.whl", hash = "sha256:deed0c7258ceb4c44ad5ec7d9918f9f14fd05b2be86378d86cf50e63d1e7b771", size = 58675 }, - { url = "https://files.pythonhosted.org/packages/9d/1a/23d855a702bb35a76faed5ae2ba3de57d323f48b1f6b17ee2176c4849463/kiwisolver-1.4.9-pp310-pypy310_pp73-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:0a590506f303f512dff6b7f75fd2fd18e16943efee932008fe7140e5fa91d80e", size = 80277 }, - { url = "https://files.pythonhosted.org/packages/5a/5b/5239e3c2b8fb5afa1e8508f721bb77325f740ab6994d963e61b2b7abcc1e/kiwisolver-1.4.9-pp310-pypy310_pp73-manylinux_2_24_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:e09c2279a4d01f099f52d5c4b3d9e208e91edcbd1a175c9662a8b16e000fece9", size = 77994 }, - { url = "https://files.pythonhosted.org/packages/f9/1c/5d4d468fb16f8410e596ed0eac02d2c68752aa7dc92997fe9d60a7147665/kiwisolver-1.4.9-pp310-pypy310_pp73-win_amd64.whl", hash = "sha256:c9e7cdf45d594ee04d5be1b24dd9d49f3d1590959b2271fb30b5ca2b262c00fb", size = 73744 }, - { url = "https://files.pythonhosted.org/packages/a3/0f/36d89194b5a32c054ce93e586d4049b6c2c22887b0eb229c61c68afd3078/kiwisolver-1.4.9-pp311-pypy311_pp73-macosx_10_15_x86_64.whl", hash = "sha256:720e05574713db64c356e86732c0f3c5252818d05f9df320f0ad8380641acea5", size = 60104 }, - { url = "https://files.pythonhosted.org/packages/52/ba/4ed75f59e4658fd21fe7dde1fee0ac397c678ec3befba3fe6482d987af87/kiwisolver-1.4.9-pp311-pypy311_pp73-macosx_11_0_arm64.whl", hash = "sha256:17680d737d5335b552994a2008fab4c851bcd7de33094a82067ef3a576ff02fa", size = 58592 }, - { url = "https://files.pythonhosted.org/packages/33/01/a8ea7c5ea32a9b45ceeaee051a04c8ed4320f5add3c51bfa20879b765b70/kiwisolver-1.4.9-pp311-pypy311_pp73-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:85b5352f94e490c028926ea567fc569c52ec79ce131dadb968d3853e809518c2", size = 80281 }, - { url = "https://files.pythonhosted.org/packages/da/e3/dbd2ecdce306f1d07a1aaf324817ee993aab7aee9db47ceac757deabafbe/kiwisolver-1.4.9-pp311-pypy311_pp73-manylinux_2_24_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:464415881e4801295659462c49461a24fb107c140de781d55518c4b80cb6790f", size = 78009 }, - { url = "https://files.pythonhosted.org/packages/da/e9/0d4add7873a73e462aeb45c036a2dead2562b825aa46ba326727b3f31016/kiwisolver-1.4.9-pp311-pypy311_pp73-win_amd64.whl", hash = "sha256:fb940820c63a9590d31d88b815e7a3aa5915cad3ce735ab45f0c730b39547de1", size = 73929 }, -] - -[[package]] -name = "lazy-loader" -version = "0.4" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "packaging" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/6f/6b/c875b30a1ba490860c93da4cabf479e03f584eba06fe5963f6f6644653d8/lazy_loader-0.4.tar.gz", hash = "sha256:47c75182589b91a4e1a85a136c074285a5ad4d9f39c63e0d7fb76391c4574cd1", size = 15431 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/83/60/d497a310bde3f01cb805196ac61b7ad6dc5dcf8dce66634dc34364b20b4f/lazy_loader-0.4-py3-none-any.whl", hash = "sha256:342aa8e14d543a154047afb4ba8ef17f5563baad3fc610d7b15b213b0f119efc", size = 12097 }, -] - -[[package]] -name = "lhotse" -version = "1.32.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "audioread" }, - { name = "click" }, - { name = "cytoolz" }, - { name = "intervaltree" }, - { name = "lilcom" }, - { name = "numpy" }, - { name = "packaging" }, - { name = "pyyaml" }, - { name = "soundfile" }, - { name = "tabulate" }, - { name = "torch" }, - { name = "tqdm" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/2e/fd/32baf46d238f029a22b2c1762fc717ebdb85515fb48bafa395d3de5da0f5/lhotse-1.32.1.tar.gz", hash = "sha256:8b0e946d1bd2c695b09df831ea612913f1a1f103b1aea36a4b43a8778be0a3d5", size = 674412 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/e6/90/9c8e26e4d56c93784503f53e68673d9bf6ea588a36f69aa48a20e99f94b0/lhotse-1.32.1-py3-none-any.whl", hash = "sha256:f2013832c568c146a0dbc76b922afa7776c13749a67c7e685f5988fac09472d0", size = 893069 }, -] - -[[package]] -name = "libcst" -version = "1.8.6" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "pyyaml", marker = "python_full_version != '3.13.*'" }, - { name = "pyyaml-ft", marker = "python_full_version == '3.13.*'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/de/cd/337df968b38d94c5aabd3e1b10630f047a2b345f6e1d4456bd9fe7417537/libcst-1.8.6.tar.gz", hash = "sha256:f729c37c9317126da9475bdd06a7208eb52fcbd180a6341648b45a56b4ba708b", size = 891354 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/c4/52/97d5454dee9d014821fe0c88f3dc0e83131b97dd074a4d49537056a75475/libcst-1.8.6-cp310-cp310-macosx_10_12_x86_64.whl", hash = "sha256:a20c5182af04332cc94d8520792befda06d73daf2865e6dddc5161c72ea92cb9", size = 2211698 }, - { url = "https://files.pythonhosted.org/packages/6c/a4/d1205985d378164687af3247a9c8f8bdb96278b0686ac98ab951bc6d336a/libcst-1.8.6-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:36473e47cb199b7e6531d653ee6ffed057de1d179301e6c67f651f3af0b499d6", size = 2093104 }, - { url = "https://files.pythonhosted.org/packages/9e/de/1338da681b7625b51e584922576d54f1b8db8fc7ff4dc79121afc5d4d2cd/libcst-1.8.6-cp310-cp310-manylinux_2_28_aarch64.whl", hash = "sha256:06fc56335a45d61b7c1b856bfab4587b84cfe31e9d6368f60bb3c9129d900f58", size = 2237419 }, - { url = "https://files.pythonhosted.org/packages/50/06/ee66f2d83b870534756e593d464d8b33b0914c224dff3a407e0f74dc04e0/libcst-1.8.6-cp310-cp310-manylinux_2_28_x86_64.whl", hash = "sha256:6b23d14a7fc0addd9795795763af26b185deb7c456b1e7cc4d5228e69dab5ce8", size = 2300820 }, - { url = "https://files.pythonhosted.org/packages/9c/ca/959088729de8e0eac8dd516e4fb8623d8d92bad539060fa85c9e94d418a5/libcst-1.8.6-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:16cfe0cfca5fd840e1fb2c30afb628b023d3085b30c3484a79b61eae9d6fe7ba", size = 2301201 }, - { url = "https://files.pythonhosted.org/packages/c2/4c/2a21a8c452436097dfe1da277f738c3517f3f728713f16d84b9a3d67ca8d/libcst-1.8.6-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:455f49a93aea4070132c30ebb6c07c2dea0ba6c1fde5ffde59fc45dbb9cfbe4b", size = 2408213 }, - { url = "https://files.pythonhosted.org/packages/3e/26/8f7b671fad38a515bb20b038718fd2221ab658299119ac9bcec56c2ced27/libcst-1.8.6-cp310-cp310-win_amd64.whl", hash = "sha256:72cca15800ffc00ba25788e4626189fe0bc5fe2a0c1cb4294bce2e4df21cc073", size = 2119189 }, - { url = "https://files.pythonhosted.org/packages/5b/bf/ffb23a48e27001165cc5c81c5d9b3d6583b21b7f5449109e03a0020b060c/libcst-1.8.6-cp310-cp310-win_arm64.whl", hash = "sha256:6cad63e3a26556b020b634d25a8703b605c0e0b491426b3e6b9e12ed20f09100", size = 2001736 }, - { url = "https://files.pythonhosted.org/packages/dc/15/95c2ecadc0fb4af8a7057ac2012a4c0ad5921b9ef1ace6c20006b56d3b5f/libcst-1.8.6-cp311-cp311-macosx_10_12_x86_64.whl", hash = "sha256:3649a813660fbffd7bc24d3f810b1f75ac98bd40d9d6f56d1f0ee38579021073", size = 2211289 }, - { url = "https://files.pythonhosted.org/packages/80/c3/7e1107acd5ed15cf60cc07c7bb64498a33042dc4821874aea3ec4942f3cd/libcst-1.8.6-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:0cbe17067055829607c5ba4afa46bfa4d0dd554c0b5a583546e690b7367a29b6", size = 2092927 }, - { url = "https://files.pythonhosted.org/packages/c1/ff/0d2be87f67e2841a4a37d35505e74b65991d30693295c46fc0380ace0454/libcst-1.8.6-cp311-cp311-manylinux_2_28_aarch64.whl", hash = "sha256:59a7e388c57d21d63722018978a8ddba7b176e3a99bd34b9b84a576ed53f2978", size = 2237002 }, - { url = "https://files.pythonhosted.org/packages/69/99/8c4a1b35c7894ccd7d33eae01ac8967122f43da41325223181ca7e4738fe/libcst-1.8.6-cp311-cp311-manylinux_2_28_x86_64.whl", hash = "sha256:b6c1248cc62952a3a005792b10cdef2a4e130847be9c74f33a7d617486f7e532", size = 2301048 }, - { url = "https://files.pythonhosted.org/packages/9b/8b/d1aa811eacf936cccfb386ae0585aa530ea1221ccf528d67144e041f5915/libcst-1.8.6-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:6421a930b028c5ef4a943b32a5a78b7f1bf15138214525a2088f11acbb7d3d64", size = 2300675 }, - { url = "https://files.pythonhosted.org/packages/c6/6b/7b65cd41f25a10c1fef2389ddc5c2b2cc23dc4d648083fa3e1aa7e0eeac2/libcst-1.8.6-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:6d8b67874f2188399a71a71731e1ba2d1a2c3173b7565d1cc7ffb32e8fbaba5b", size = 2407934 }, - { url = "https://files.pythonhosted.org/packages/c5/8b/401cfff374bb3b785adfad78f05225225767ee190997176b2a9da9ed9460/libcst-1.8.6-cp311-cp311-win_amd64.whl", hash = "sha256:b0d8c364c44ae343937f474b2e492c1040df96d94530377c2f9263fb77096e4f", size = 2119247 }, - { url = "https://files.pythonhosted.org/packages/f1/17/085f59eaa044b6ff6bc42148a5449df2b7f0ba567307de7782fe85c39ee2/libcst-1.8.6-cp311-cp311-win_arm64.whl", hash = "sha256:5dcaaebc835dfe5755bc85f9b186fb7e2895dda78e805e577fef1011d51d5a5c", size = 2001774 }, - { url = "https://files.pythonhosted.org/packages/0c/3c/93365c17da3d42b055a8edb0e1e99f1c60c776471db6c9b7f1ddf6a44b28/libcst-1.8.6-cp312-cp312-macosx_10_13_x86_64.whl", hash = "sha256:0c13d5bd3d8414a129e9dccaf0e5785108a4441e9b266e1e5e9d1f82d1b943c9", size = 2206166 }, - { url = "https://files.pythonhosted.org/packages/1d/cb/7530940e6ac50c6dd6022349721074e19309eb6aa296e942ede2213c1a19/libcst-1.8.6-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:f1472eeafd67cdb22544e59cf3bfc25d23dc94058a68cf41f6654ff4fcb92e09", size = 2083726 }, - { url = "https://files.pythonhosted.org/packages/1b/cf/7e5eaa8c8f2c54913160671575351d129170db757bb5e4b7faffed022271/libcst-1.8.6-cp312-cp312-manylinux_2_28_aarch64.whl", hash = "sha256:089c58e75cb142ec33738a1a4ea7760a28b40c078ab2fd26b270dac7d2633a4d", size = 2235755 }, - { url = "https://files.pythonhosted.org/packages/55/54/570ec2b0e9a3de0af9922e3bb1b69a5429beefbc753a7ea770a27ad308bd/libcst-1.8.6-cp312-cp312-manylinux_2_28_x86_64.whl", hash = "sha256:c9d7aeafb1b07d25a964b148c0dda9451efb47bbbf67756e16eeae65004b0eb5", size = 2301473 }, - { url = "https://files.pythonhosted.org/packages/11/4c/163457d1717cd12181c421a4cca493454bcabd143fc7e53313bc6a4ad82a/libcst-1.8.6-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:207481197afd328aa91d02670c15b48d0256e676ce1ad4bafb6dc2b593cc58f1", size = 2298899 }, - { url = "https://files.pythonhosted.org/packages/35/1d/317ddef3669883619ef3d3395ea583305f353ef4ad87d7a5ac1c39be38e3/libcst-1.8.6-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:375965f34cc6f09f5f809244d3ff9bd4f6cb6699f571121cebce53622e7e0b86", size = 2408239 }, - { url = "https://files.pythonhosted.org/packages/9a/a1/f47d8cccf74e212dd6044b9d6dbc223636508da99acff1d54786653196bc/libcst-1.8.6-cp312-cp312-win_amd64.whl", hash = "sha256:da95b38693b989eaa8d32e452e8261cfa77fe5babfef1d8d2ac25af8c4aa7e6d", size = 2119660 }, - { url = "https://files.pythonhosted.org/packages/19/d0/dd313bf6a7942cdf951828f07ecc1a7695263f385065edc75ef3016a3cb5/libcst-1.8.6-cp312-cp312-win_arm64.whl", hash = "sha256:bff00e1c766658adbd09a175267f8b2f7616e5ee70ce45db3d7c4ce6d9f6bec7", size = 1999824 }, - { url = "https://files.pythonhosted.org/packages/90/01/723cd467ec267e712480c772aacc5aa73f82370c9665162fd12c41b0065b/libcst-1.8.6-cp313-cp313-macosx_10_13_x86_64.whl", hash = "sha256:7445479ebe7d1aff0ee094ab5a1c7718e1ad78d33e3241e1a1ec65dcdbc22ffb", size = 2206386 }, - { url = "https://files.pythonhosted.org/packages/17/50/b944944f910f24c094f9b083f76f61e3985af5a376f5342a21e01e2d1a81/libcst-1.8.6-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:4fc3fef8a2c983e7abf5d633e1884c5dd6fa0dcb8f6e32035abd3d3803a3a196", size = 2083945 }, - { url = "https://files.pythonhosted.org/packages/36/a1/bd1b2b2b7f153d82301cdaddba787f4a9fc781816df6bdb295ca5f88b7cf/libcst-1.8.6-cp313-cp313-manylinux_2_28_aarch64.whl", hash = "sha256:1a3a5e4ee870907aa85a4076c914ae69066715a2741b821d9bf16f9579de1105", size = 2235818 }, - { url = "https://files.pythonhosted.org/packages/b9/ab/f5433988acc3b4d188c4bb154e57837df9488cc9ab551267cdeabd3bb5e7/libcst-1.8.6-cp313-cp313-manylinux_2_28_x86_64.whl", hash = "sha256:6609291c41f7ad0bac570bfca5af8fea1f4a27987d30a1fa8b67fe5e67e6c78d", size = 2301289 }, - { url = "https://files.pythonhosted.org/packages/5d/57/89f4ba7a6f1ac274eec9903a9e9174890d2198266eee8c00bc27eb45ecf7/libcst-1.8.6-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:25eaeae6567091443b5374b4c7d33a33636a2d58f5eda02135e96fc6c8807786", size = 2299230 }, - { url = "https://files.pythonhosted.org/packages/f2/36/0aa693bc24cce163a942df49d36bf47a7ed614a0cd5598eee2623bc31913/libcst-1.8.6-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:04030ea4d39d69a65873b1d4d877def1c3951a7ada1824242539e399b8763d30", size = 2408519 }, - { url = "https://files.pythonhosted.org/packages/db/18/6dd055b5f15afa640fb3304b2ee9df8b7f72e79513814dbd0a78638f4a0e/libcst-1.8.6-cp313-cp313-win_amd64.whl", hash = "sha256:8066f1b70f21a2961e96bedf48649f27dfd5ea68be5cd1bed3742b047f14acde", size = 2119853 }, - { url = "https://files.pythonhosted.org/packages/c9/ed/5ddb2a22f0b0abdd6dcffa40621ada1feaf252a15e5b2733a0a85dfd0429/libcst-1.8.6-cp313-cp313-win_arm64.whl", hash = "sha256:c188d06b583900e662cd791a3f962a8c96d3dfc9b36ea315be39e0a4c4792ebf", size = 1999808 }, - { url = "https://files.pythonhosted.org/packages/25/d3/72b2de2c40b97e1ef4a1a1db4e5e52163fc7e7740ffef3846d30bc0096b5/libcst-1.8.6-cp313-cp313t-macosx_10_13_x86_64.whl", hash = "sha256:c41c76e034a1094afed7057023b1d8967f968782433f7299cd170eaa01ec033e", size = 2190553 }, - { url = "https://files.pythonhosted.org/packages/0d/20/983b7b210ccc3ad94a82db54230e92599c4a11b9cfc7ce3bc97c1d2df75c/libcst-1.8.6-cp313-cp313t-macosx_11_0_arm64.whl", hash = "sha256:5432e785322aba3170352f6e72b32bea58d28abd141ac37cc9b0bf6b7c778f58", size = 2074717 }, - { url = "https://files.pythonhosted.org/packages/13/f2/9e01678fedc772e09672ed99930de7355757035780d65d59266fcee212b8/libcst-1.8.6-cp313-cp313t-manylinux_2_28_aarch64.whl", hash = "sha256:85b7025795b796dea5284d290ff69de5089fc8e989b25d6f6f15b6800be7167f", size = 2225834 }, - { url = "https://files.pythonhosted.org/packages/4a/0d/7bed847b5c8c365e9f1953da274edc87577042bee5a5af21fba63276e756/libcst-1.8.6-cp313-cp313t-manylinux_2_28_x86_64.whl", hash = "sha256:536567441182a62fb706e7aa954aca034827b19746832205953b2c725d254a93", size = 2287107 }, - { url = "https://files.pythonhosted.org/packages/02/f0/7e51fa84ade26c518bfbe7e2e4758b56d86a114c72d60309ac0d350426c4/libcst-1.8.6-cp313-cp313t-musllinux_1_2_aarch64.whl", hash = "sha256:2f04d3672bde1704f383a19e8f8331521abdbc1ed13abb349325a02ac56e5012", size = 2288672 }, - { url = "https://files.pythonhosted.org/packages/ad/cd/15762659a3f5799d36aab1bc2b7e732672722e249d7800e3c5f943b41250/libcst-1.8.6-cp313-cp313t-musllinux_1_2_x86_64.whl", hash = "sha256:7f04febcd70e1e67917be7de513c8d4749d2e09206798558d7fe632134426ea4", size = 2392661 }, - { url = "https://files.pythonhosted.org/packages/e4/6b/b7f9246c323910fcbe021241500f82e357521495dcfe419004dbb272c7cb/libcst-1.8.6-cp313-cp313t-win_amd64.whl", hash = "sha256:1dc3b897c8b0f7323412da3f4ad12b16b909150efc42238e19cbf19b561cc330", size = 2105068 }, - { url = "https://files.pythonhosted.org/packages/a6/0b/4fd40607bc4807ec2b93b054594373d7fa3d31bb983789901afcb9bcebe9/libcst-1.8.6-cp313-cp313t-win_arm64.whl", hash = "sha256:44f38139fa95e488db0f8976f9c7ca39a64d6bc09f2eceef260aa1f6da6a2e42", size = 1985181 }, - { url = "https://files.pythonhosted.org/packages/3a/60/4105441989e321f7ad0fd28ffccb83eb6aac0b7cfb0366dab855dcccfbe5/libcst-1.8.6-cp314-cp314-macosx_10_15_x86_64.whl", hash = "sha256:b188e626ce61de5ad1f95161b8557beb39253de4ec74fc9b1f25593324a0279c", size = 2204202 }, - { url = "https://files.pythonhosted.org/packages/67/2f/51a6f285c3a183e50cfe5269d4a533c21625aac2c8de5cdf2d41f079320d/libcst-1.8.6-cp314-cp314-macosx_11_0_arm64.whl", hash = "sha256:87e74f7d7dfcba9efa91127081e22331d7c42515f0a0ac6e81d4cf2c3ed14661", size = 2083581 }, - { url = "https://files.pythonhosted.org/packages/2f/64/921b1c19b638860af76cdb28bc81d430056592910b9478eea49e31a7f47a/libcst-1.8.6-cp314-cp314-manylinux_2_28_aarch64.whl", hash = "sha256:3a926a4b42015ee24ddfc8ae940c97bd99483d286b315b3ce82f3bafd9f53474", size = 2236495 }, - { url = "https://files.pythonhosted.org/packages/12/a8/b00592f9bede618cbb3df6ffe802fc65f1d1c03d48a10d353b108057d09c/libcst-1.8.6-cp314-cp314-manylinux_2_28_x86_64.whl", hash = "sha256:3f4fbb7f569e69fd9e89d9d9caa57ca42c577c28ed05062f96a8c207594e75b8", size = 2301466 }, - { url = "https://files.pythonhosted.org/packages/af/df/790d9002f31580fefd0aec2f373a0f5da99070e04c5e8b1c995d0104f303/libcst-1.8.6-cp314-cp314-musllinux_1_2_aarch64.whl", hash = "sha256:08bd63a8ce674be431260649e70fca1d43f1554f1591eac657f403ff8ef82c7a", size = 2300264 }, - { url = "https://files.pythonhosted.org/packages/21/de/dc3f10e65bab461be5de57850d2910a02c24c3ddb0da28f0e6e4133c3487/libcst-1.8.6-cp314-cp314-musllinux_1_2_x86_64.whl", hash = "sha256:e00e275d4ba95d4963431ea3e409aa407566a74ee2bf309a402f84fc744abe47", size = 2408572 }, - { url = "https://files.pythonhosted.org/packages/20/3b/35645157a7590891038b077db170d6dd04335cd2e82a63bdaa78c3297dfe/libcst-1.8.6-cp314-cp314-win_amd64.whl", hash = "sha256:fea5c7fa26556eedf277d4f72779c5ede45ac3018650721edd77fd37ccd4a2d4", size = 2193917 }, - { url = "https://files.pythonhosted.org/packages/b3/a2/1034a9ba7d3e82f2c2afaad84ba5180f601aed676d92b76325797ad60951/libcst-1.8.6-cp314-cp314-win_arm64.whl", hash = "sha256:bb9b4077bdf8857b2483879cbbf70f1073bc255b057ec5aac8a70d901bb838e9", size = 2078748 }, - { url = "https://files.pythonhosted.org/packages/95/a1/30bc61e8719f721a5562f77695e6154e9092d1bdf467aa35d0806dcd6cea/libcst-1.8.6-cp314-cp314t-macosx_10_15_x86_64.whl", hash = "sha256:55ec021a296960c92e5a33b8d93e8ad4182b0eab657021f45262510a58223de1", size = 2188980 }, - { url = "https://files.pythonhosted.org/packages/2c/14/c660204532407c5628e3b615015a902ed2d0b884b77714a6bdbe73350910/libcst-1.8.6-cp314-cp314t-macosx_11_0_arm64.whl", hash = "sha256:ba9ab2b012fbd53b36cafd8f4440a6b60e7e487cd8b87428e57336b7f38409a4", size = 2074828 }, - { url = "https://files.pythonhosted.org/packages/82/e2/c497c354943dff644749f177ee9737b09ed811b8fc842b05709a40fe0d1b/libcst-1.8.6-cp314-cp314t-manylinux_2_28_aarch64.whl", hash = "sha256:c0a0cc80aebd8aa15609dd4d330611cbc05e9b4216bcaeabba7189f99ef07c28", size = 2225568 }, - { url = "https://files.pythonhosted.org/packages/86/ef/45999676d07bd6d0eefa28109b4f97124db114e92f9e108de42ba46a8028/libcst-1.8.6-cp314-cp314t-manylinux_2_28_x86_64.whl", hash = "sha256:42a4f68121e2e9c29f49c97f6154e8527cd31021809cc4a941c7270aa64f41aa", size = 2286523 }, - { url = "https://files.pythonhosted.org/packages/f4/6c/517d8bf57d9f811862f4125358caaf8cd3320a01291b3af08f7b50719db4/libcst-1.8.6-cp314-cp314t-musllinux_1_2_aarch64.whl", hash = "sha256:8a434c521fadaf9680788b50d5c21f4048fa85ed19d7d70bd40549fbaeeecab1", size = 2288044 }, - { url = "https://files.pythonhosted.org/packages/83/ce/24d7d49478ffb61207f229239879845da40a374965874f5ee60f96b02ddb/libcst-1.8.6-cp314-cp314t-musllinux_1_2_x86_64.whl", hash = "sha256:6a65f844d813ab4ef351443badffa0ae358f98821561d19e18b3190f59e71996", size = 2392605 }, - { url = "https://files.pythonhosted.org/packages/39/c3/829092ead738b71e96a4e96896c96f276976e5a8a58b4473ed813d7c962b/libcst-1.8.6-cp314-cp314t-win_amd64.whl", hash = "sha256:bdb14bc4d4d83a57062fed2c5da93ecb426ff65b0dc02ddf3481040f5f074a82", size = 2181581 }, - { url = "https://files.pythonhosted.org/packages/98/6d/5d6a790a02eb0d9d36c4aed4f41b277497e6178900b2fa29c35353aa45ed/libcst-1.8.6-cp314-cp314t-win_arm64.whl", hash = "sha256:819c8081e2948635cab60c603e1bbdceccdfe19104a242530ad38a36222cb88f", size = 2065000 }, -] - -[[package]] -name = "librosa" -version = "0.11.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "audioread" }, - { name = "decorator" }, - { name = "joblib" }, - { name = "lazy-loader" }, - { name = "msgpack" }, - { name = "numba" }, - { name = "numpy" }, - { name = "pooch" }, - { name = "scikit-learn", version = "1.7.2", source = { registry = "https://pypi.org/simple" }, marker = "python_full_version < '3.11'" }, - { name = "scikit-learn", version = "1.8.0", source = { registry = "https://pypi.org/simple" }, marker = "python_full_version >= '3.11'" }, - { name = "scipy", version = "1.15.3", source = { registry = "https://pypi.org/simple" }, marker = "python_full_version < '3.11'" }, - { name = "scipy", version = "1.16.3", source = { registry = "https://pypi.org/simple" }, marker = "python_full_version >= '3.11'" }, - { name = "soundfile" }, - { name = "soxr" }, - { name = "standard-aifc", marker = "python_full_version >= '3.13'" }, - { name = "standard-sunau", marker = "python_full_version >= '3.13'" }, - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/64/36/360b5aafa0238e29758729e9486c6ed92a6f37fa403b7875e06c115cdf4a/librosa-0.11.0.tar.gz", hash = "sha256:f5ed951ca189b375bbe2e33b2abd7e040ceeee302b9bbaeeffdfddb8d0ace908", size = 327001 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/b5/ba/c63c5786dfee4c3417094c4b00966e61e4a63efecee22cb7b4c0387dda83/librosa-0.11.0-py3-none-any.whl", hash = "sha256:0b6415c4fd68bff4c29288abe67c6d80b587e0e1e2cfb0aad23e4559504a7fa1", size = 260749 }, -] - -[[package]] -name = "lightning" -version = "2.4.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "fsspec", extra = ["http"] }, - { name = "lightning-utilities" }, - { name = "packaging" }, - { name = "pytorch-lightning" }, - { name = "pyyaml" }, - { name = "torch" }, - { name = "torchmetrics" }, - { name = "tqdm" }, - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/56/d0/78ea244ac044cd4df15aa8294a50ff3561fb177e7e5ba788aaa542046cae/lightning-2.4.0.tar.gz", hash = "sha256:9156604cc56e4b2b603f34fa7f0fe5107375c8e6d85e74544b319a15faa9ed0e", size = 620632 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/a3/2c/85eaf42c983b0cd81bcda5876da2c8e2a9fd347908666ea9855724369171/lightning-2.4.0-py3-none-any.whl", hash = "sha256:560163af9711cf59055c448232c473150a299089efce0d2be3cc3288082d8768", size = 810971 }, -] - -[[package]] -name = "lightning-utilities" -version = "0.15.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "packaging" }, - { name = "setuptools" }, - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/b8/39/6fc58ca81492db047149b4b8fd385aa1bfb8c28cd7cacb0c7eb0c44d842f/lightning_utilities-0.15.2.tar.gz", hash = "sha256:cdf12f530214a63dacefd713f180d1ecf5d165338101617b4742e8f22c032e24", size = 31090 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/de/73/3d757cb3fc16f0f9794dd289bcd0c4a031d9cf54d8137d6b984b2d02edf3/lightning_utilities-0.15.2-py3-none-any.whl", hash = "sha256:ad3ab1703775044bbf880dbf7ddaaac899396c96315f3aa1779cec9d618a9841", size = 29431 }, -] - -[[package]] -name = "lilcom" -version = "1.8.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "numpy" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/8b/b5/97422825e61a2683dd39d78165fc470c1153b4a908dd5cd0711c0e9d262c/lilcom-1.8.1.tar.gz", hash = "sha256:69c62037c92e71e601ac3bb3ae19811f22ceffbdf58b0fdbf81cc6a0ec6fc3c5", size = 45813 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/e1/5c/51ece6a8f6ed39c27eb18bb1c09fe6508613bc67bcc5872caa788cdc9eef/lilcom-1.8.1-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:7e896dd6172011d320c26a4ca648edddf1e6b47bb1de749cf27ac096d73f63c9", size = 122436 }, - { url = "https://files.pythonhosted.org/packages/58/fd/8008d8b82965eeb2e35a46edb118e195d10bc424e3d80f2e3dafe3044e1f/lilcom-1.8.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:6b7c5488267c9c3c6ab49c3e3ec287031029d7de57f4c2307264dd75f5e77211", size = 86593 }, - { url = "https://files.pythonhosted.org/packages/7b/02/30e76ca2bfc3fe19adbcca254dde03ba48a6e623eb55df2752f4c06cfeb0/lilcom-1.8.1-cp310-cp310-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:7460e2dd481b38fc8fa1f0cc48c5aaa1b4206d0c27130c001bf23f1d4c99143b", size = 98968 }, - { url = "https://files.pythonhosted.org/packages/d7/cd/e3e45ef07c2e19f5718d7c9d33b40438b83a3397bdd4c70bd3d1bbf059ad/lilcom-1.8.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9bdeea4aa3d7cad9e38cac99ef012a61f843a87a661b9475f565787f73828ec5", size = 92337 }, - { url = "https://files.pythonhosted.org/packages/84/da/3b8a45be54f4d12a2d118c84d9d2aab81e5b75ae4a7b8469213a28b121dd/lilcom-1.8.1-cp310-cp310-win_amd64.whl", hash = "sha256:c254a74156ff8da2e43c8b891ed47eefcf0c1fee82a0c5984388052d97573a1b", size = 68524 }, - { url = "https://files.pythonhosted.org/packages/6c/b7/7044859b4b6cce0c4d9c2731198ad3970425dda1824a0cc64db94b379054/lilcom-1.8.1-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:13e7384675119f34504e44821d6baf8d9696dd2ed77ba54a392cd79d22c376f2", size = 125275 }, - { url = "https://files.pythonhosted.org/packages/76/15/78f1a44a2a98be5830e7d6a7f37e9b7b18475d3b65454adec13348bdf78b/lilcom-1.8.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:075b3155e37d2964ce47dc5d456c9afe3f5aace0f48e33236b0ca126e7f2d757", size = 88439 }, - { url = "https://files.pythonhosted.org/packages/5c/a5/3da24b27e60dce949675b2422bb58471de6268940ecf59d5238a3ad8ddb9/lilcom-1.8.1-cp311-cp311-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:9127afa12130c2421b6906c4ff77772f132f87b085404a1cf2486bbb4f37f1a5", size = 100600 }, - { url = "https://files.pythonhosted.org/packages/d5/cc/46f2bcd812a4039bd55633b10d5146466b295dea378d6c84facc7a55c8db/lilcom-1.8.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:defce157d05bf8bdef9e0c5d2346a9088a3b939ef653475a11e627b90802e501", size = 94252 }, - { url = "https://files.pythonhosted.org/packages/36/c6/49b67e55704eb4252966610d35dee6256425209b2ea9be667388d58cba62/lilcom-1.8.1-cp311-cp311-win_amd64.whl", hash = "sha256:af2585d979b8a92fd2f882a5e7320391e7131a98c47200c3668ef7bb7f061926", size = 69181 }, - { url = "https://files.pythonhosted.org/packages/78/1e/4bca706dc50230146b168d5c52403ec4e144650999fae2150a0e60ed7303/lilcom-1.8.1-cp312-cp312-macosx_10_13_universal2.whl", hash = "sha256:b0ace9d4f95199266a36d1144214ff2056679ddb44462f75326a0b9fbf296928", size = 124607 }, - { url = "https://files.pythonhosted.org/packages/20/80/0d07c944a09bfed9038121bf3d662e4f5cf56f7f8c2be3669269ceec429b/lilcom-1.8.1-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:834df2155ee7cad41cc5cc482bae764c89ebbaee623cb4788a3b7bbdcd2d22d3", size = 86841 }, - { url = "https://files.pythonhosted.org/packages/7c/0d/f8be338de1d45d6a6fc26a24d0ad9f6be0a94130b5b7687c2cc08a5be099/lilcom-1.8.1-cp312-cp312-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:23b45cb5b966999826eb1aa80154d392550902ef26374dc628a8c576d844d548", size = 98796 }, - { url = "https://files.pythonhosted.org/packages/e3/6e/7868a5a90e66fefa60642682f5551cb75e489c603f155367003b88363af3/lilcom-1.8.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:08ba2cf1cd9a175ec3b791ea727e981a1cd25a03d39c32264d651bc52e7e5c33", size = 92958 }, - { url = "https://files.pythonhosted.org/packages/db/94/9553e967a067e0579dd9be9658a428ae78aebde19d7c56d7478bf68cf48b/lilcom-1.8.1-cp312-cp312-win_amd64.whl", hash = "sha256:1eefd5b4735fca240aa35000dbe8a71e843fb00d64ccdc468c61eab03f23cf86", size = 69719 }, - { url = "https://files.pythonhosted.org/packages/c3/7f/bd4e63c9afea69f4130ef32d824530a62a628007012eb9cf9661509c60fe/lilcom-1.8.1-cp313-cp313-macosx_10_13_universal2.whl", hash = "sha256:a37956d17c9657177781391b5b4213f554737ae6bf8a5d37cd11465a221f1832", size = 124650 }, - { url = "https://files.pythonhosted.org/packages/1f/58/f1ce0b67a144a2f34828c73df8dd362500031ff07482eac8b985a93a8456/lilcom-1.8.1-cp313-cp313-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0c4bc1281e1d40163fc438abdd48ac71dbe04bb1e10ef62761b446339f763bb0", size = 86850 }, - { url = "https://files.pythonhosted.org/packages/bd/02/2c329139ef8615c689f4a402e46c7c9029110eab14ca4685aef7ecf0593f/lilcom-1.8.1-cp313-cp313-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:1267644b383a7b9f64ba5074b459e3f91635363fb36824c81205774f31b6e1c8", size = 98991 }, - { url = "https://files.pythonhosted.org/packages/c7/9a/16b944bf1d3c18637e759a059433537d5a54131decc2138a81d5834fc5a3/lilcom-1.8.1-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:bb8b1b76ab81b1c7953d25c80e7880d31c8fbc4b63da2c6375c6151332017da1", size = 93484 }, - { url = "https://files.pythonhosted.org/packages/56/5a/eee3e4afd3dcf3c23ff6e9899098abf94a661bc5eb474ca366c57142338f/lilcom-1.8.1-cp313-cp313-win_amd64.whl", hash = "sha256:5af35273eab932a6b84612bd1597863465078654f40b98a9f8e2d6f67d9f6ebc", size = 69719 }, -] - -[[package]] -name = "llvmlite" -version = "0.46.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/74/cd/08ae687ba099c7e3d21fe2ea536500563ef1943c5105bf6ab4ee3829f68e/llvmlite-0.46.0.tar.gz", hash = "sha256:227c9fd6d09dce2783c18b754b7cd9d9b3b3515210c46acc2d3c5badd9870ceb", size = 193456 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/3d/a4/3959e1c61c5ca9db7921e5fd115b344c29b9d57a5dadd87bef97963ca1a5/llvmlite-0.46.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:4323177e936d61ae0f73e653e2e614284d97d14d5dd12579adc92b6c2b0597b0", size = 37232766 }, - { url = "https://files.pythonhosted.org/packages/c2/a5/a4d916f1015106e1da876028606a8e87fd5d5c840f98c87bc2d5153b6a2f/llvmlite-0.46.0-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:0a2d461cb89537b7c20feb04c46c32e12d5ad4f0896c9dfc0f60336219ff248e", size = 56275176 }, - { url = "https://files.pythonhosted.org/packages/79/7f/a7f2028805dac8c1a6fae7bda4e739b7ebbcd45b29e15bf6d21556fcd3d5/llvmlite-0.46.0-cp310-cp310-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:b1f6595a35b7b39c3518b85a28bf18f45e075264e4b2dce3f0c2a4f232b4a910", size = 55128629 }, - { url = "https://files.pythonhosted.org/packages/b2/bc/4689e1ba0c073c196b594471eb21be0aa51d9e64b911728aa13cd85ef0ae/llvmlite-0.46.0-cp310-cp310-win_amd64.whl", hash = "sha256:e7a34d4aa6f9a97ee006b504be6d2b8cb7f755b80ab2f344dda1ef992f828559", size = 38138651 }, - { url = "https://files.pythonhosted.org/packages/7a/a1/2ad4b2367915faeebe8447f0a057861f646dbf5fbbb3561db42c65659cf3/llvmlite-0.46.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:82f3d39b16f19aa1a56d5fe625883a6ab600d5cc9ea8906cca70ce94cabba067", size = 37232766 }, - { url = "https://files.pythonhosted.org/packages/12/b5/99cf8772fdd846c07da4fd70f07812a3c8fd17ea2409522c946bb0f2b277/llvmlite-0.46.0-cp311-cp311-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:a3df43900119803bbc52720e758c76f316a9a0f34612a886862dfe0a5591a17e", size = 56275175 }, - { url = "https://files.pythonhosted.org/packages/38/f2/ed806f9c003563732da156139c45d970ee435bd0bfa5ed8de87ba972b452/llvmlite-0.46.0-cp311-cp311-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:de183fefc8022d21b0aa37fc3e90410bc3524aed8617f0ff76732fc6c3af5361", size = 55128630 }, - { url = "https://files.pythonhosted.org/packages/19/0c/8f5a37a65fc9b7b17408508145edd5f86263ad69c19d3574e818f533a0eb/llvmlite-0.46.0-cp311-cp311-win_amd64.whl", hash = "sha256:e8b10bc585c58bdffec9e0c309bb7d51be1f2f15e169a4b4d42f2389e431eb93", size = 38138652 }, - { url = "https://files.pythonhosted.org/packages/2b/f8/4db016a5e547d4e054ff2f3b99203d63a497465f81ab78ec8eb2ff7b2304/llvmlite-0.46.0-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:6b9588ad4c63b4f0175a3984b85494f0c927c6b001e3a246a3a7fb3920d9a137", size = 37232767 }, - { url = "https://files.pythonhosted.org/packages/aa/85/4890a7c14b4fa54400945cb52ac3cd88545bbdb973c440f98ca41591cdc5/llvmlite-0.46.0-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:3535bd2bb6a2d7ae4012681ac228e5132cdb75fefb1bcb24e33f2f3e0c865ed4", size = 56275176 }, - { url = "https://files.pythonhosted.org/packages/6a/07/3d31d39c1a1a08cd5337e78299fca77e6aebc07c059fbd0033e3edfab45c/llvmlite-0.46.0-cp312-cp312-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:4cbfd366e60ff87ea6cc62f50bc4cd800ebb13ed4c149466f50cf2163a473d1e", size = 55128630 }, - { url = "https://files.pythonhosted.org/packages/2a/6b/d139535d7590a1bba1ceb68751bef22fadaa5b815bbdf0e858e3875726b2/llvmlite-0.46.0-cp312-cp312-win_amd64.whl", hash = "sha256:398b39db462c39563a97b912d4f2866cd37cba60537975a09679b28fbbc0fb38", size = 38138940 }, - { url = "https://files.pythonhosted.org/packages/e6/ff/3eba7eb0aed4b6fca37125387cd417e8c458e750621fce56d2c541f67fa8/llvmlite-0.46.0-cp313-cp313-macosx_12_0_arm64.whl", hash = "sha256:30b60892d034bc560e0ec6654737aaa74e5ca327bd8114d82136aa071d611172", size = 37232767 }, - { url = "https://files.pythonhosted.org/packages/0e/54/737755c0a91558364b9200702c3c9c15d70ed63f9b98a2c32f1c2aa1f3ba/llvmlite-0.46.0-cp313-cp313-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:6cc19b051753368a9c9f31dc041299059ee91aceec81bd57b0e385e5d5bf1a54", size = 56275176 }, - { url = "https://files.pythonhosted.org/packages/e6/91/14f32e1d70905c1c0aa4e6609ab5d705c3183116ca02ac6df2091868413a/llvmlite-0.46.0-cp313-cp313-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:bca185892908f9ede48c0acd547fe4dc1bafefb8a4967d47db6cf664f9332d12", size = 55128629 }, - { url = "https://files.pythonhosted.org/packages/4a/a7/d526ae86708cea531935ae777b6dbcabe7db52718e6401e0fb9c5edea80e/llvmlite-0.46.0-cp313-cp313-win_amd64.whl", hash = "sha256:67438fd30e12349ebb054d86a5a1a57fd5e87d264d2451bcfafbbbaa25b82a35", size = 38138941 }, - { url = "https://files.pythonhosted.org/packages/95/ae/af0ffb724814cc2ea64445acad05f71cff5f799bb7efb22e47ee99340dbc/llvmlite-0.46.0-cp314-cp314-macosx_12_0_arm64.whl", hash = "sha256:d252edfb9f4ac1fcf20652258e3f102b26b03eef738dc8a6ffdab7d7d341d547", size = 37232768 }, - { url = "https://files.pythonhosted.org/packages/c9/19/5018e5352019be753b7b07f7759cdabb69ca5779fea2494be8839270df4c/llvmlite-0.46.0-cp314-cp314-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:379fdd1c59badeff8982cb47e4694a6143bec3bb49aa10a466e095410522064d", size = 56275173 }, - { url = "https://files.pythonhosted.org/packages/9f/c9/d57877759d707e84c082163c543853245f91b70c804115a5010532890f18/llvmlite-0.46.0-cp314-cp314-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:2e8cbfff7f6db0fa2c771ad24154e2a7e457c2444d7673e6de06b8b698c3b269", size = 55128628 }, - { url = "https://files.pythonhosted.org/packages/30/a8/e61a8c2b3cc7a597073d9cde1fcbb567e9d827f1db30c93cf80422eac70d/llvmlite-0.46.0-cp314-cp314-win_amd64.whl", hash = "sha256:7821eda3ec1f18050f981819756631d60b6d7ab1a6cf806d9efefbe3f4082d61", size = 39153056 }, -] - -[[package]] -name = "mako" -version = "1.3.10" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "markupsafe" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/9e/38/bd5b78a920a64d708fe6bc8e0a2c075e1389d53bef8413725c63ba041535/mako-1.3.10.tar.gz", hash = "sha256:99579a6f39583fa7e5630a28c3c1f440e4e97a414b80372649c0ce338da2ea28", size = 392474 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/87/fb/99f81ac72ae23375f22b7afdb7642aba97c00a713c217124420147681a2f/mako-1.3.10-py3-none-any.whl", hash = "sha256:baef24a52fc4fc514a0887ac600f9f1cff3d82c61d4d700a1fa84d597b88db59", size = 78509 }, -] - -[[package]] -name = "markdown" -version = "3.10" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/7d/ab/7dd27d9d863b3376fcf23a5a13cb5d024aed1db46f963f1b5735ae43b3be/markdown-3.10.tar.gz", hash = "sha256:37062d4f2aa4b2b6b32aefb80faa300f82cc790cb949a35b8caede34f2b68c0e", size = 364931 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/70/81/54e3ce63502cd085a0c556652a4e1b919c45a446bd1e5300e10c44c8c521/markdown-3.10-py3-none-any.whl", hash = "sha256:b5b99d6951e2e4948d939255596523444c0e677c669700b1d17aa4a8a464cb7c", size = 107678 }, -] - -[[package]] -name = "markdown-it-py" -version = "4.0.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "mdurl" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/5b/f5/4ec618ed16cc4f8fb3b701563655a69816155e79e24a17b651541804721d/markdown_it_py-4.0.0.tar.gz", hash = "sha256:cb0a2b4aa34f932c007117b194e945bd74e0ec24133ceb5bac59009cda1cb9f3", size = 73070 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/94/54/e7d793b573f298e1c9013b8c4dade17d481164aa517d1d7148619c2cedbf/markdown_it_py-4.0.0-py3-none-any.whl", hash = "sha256:87327c59b172c5011896038353a81343b6754500a08cd7a4973bb48c6d578147", size = 87321 }, -] - -[[package]] -name = "markupsafe" -version = "3.0.3" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/7e/99/7690b6d4034fffd95959cbe0c02de8deb3098cc577c67bb6a24fe5d7caa7/markupsafe-3.0.3.tar.gz", hash = "sha256:722695808f4b6457b320fdc131280796bdceb04ab50fe1795cd540799ebe1698", size = 80313 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/e8/4b/3541d44f3937ba468b75da9eebcae497dcf67adb65caa16760b0a6807ebb/markupsafe-3.0.3-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:2f981d352f04553a7171b8e44369f2af4055f888dfb147d55e42d29e29e74559", size = 11631 }, - { url = "https://files.pythonhosted.org/packages/98/1b/fbd8eed11021cabd9226c37342fa6ca4e8a98d8188a8d9b66740494960e4/markupsafe-3.0.3-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:e1c1493fb6e50ab01d20a22826e57520f1284df32f2d8601fdd90b6304601419", size = 12057 }, - { url = "https://files.pythonhosted.org/packages/40/01/e560d658dc0bb8ab762670ece35281dec7b6c1b33f5fbc09ebb57a185519/markupsafe-3.0.3-cp310-cp310-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:1ba88449deb3de88bd40044603fafffb7bc2b055d626a330323a9ed736661695", size = 22050 }, - { url = "https://files.pythonhosted.org/packages/af/cd/ce6e848bbf2c32314c9b237839119c5a564a59725b53157c856e90937b7a/markupsafe-3.0.3-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:f42d0984e947b8adf7dd6dde396e720934d12c506ce84eea8476409563607591", size = 20681 }, - { url = "https://files.pythonhosted.org/packages/c9/2a/b5c12c809f1c3045c4d580b035a743d12fcde53cf685dbc44660826308da/markupsafe-3.0.3-cp310-cp310-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl", hash = "sha256:c0c0b3ade1c0b13b936d7970b1d37a57acde9199dc2aecc4c336773e1d86049c", size = 20705 }, - { url = "https://files.pythonhosted.org/packages/cf/e3/9427a68c82728d0a88c50f890d0fc072a1484de2f3ac1ad0bfc1a7214fd5/markupsafe-3.0.3-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:0303439a41979d9e74d18ff5e2dd8c43ed6c6001fd40e5bf2e43f7bd9bbc523f", size = 21524 }, - { url = "https://files.pythonhosted.org/packages/bc/36/23578f29e9e582a4d0278e009b38081dbe363c5e7165113fad546918a232/markupsafe-3.0.3-cp310-cp310-musllinux_1_2_riscv64.whl", hash = "sha256:d2ee202e79d8ed691ceebae8e0486bd9a2cd4794cec4824e1c99b6f5009502f6", size = 20282 }, - { url = "https://files.pythonhosted.org/packages/56/21/dca11354e756ebd03e036bd8ad58d6d7168c80ce1fe5e75218e4945cbab7/markupsafe-3.0.3-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:177b5253b2834fe3678cb4a5f0059808258584c559193998be2601324fdeafb1", size = 20745 }, - { url = "https://files.pythonhosted.org/packages/87/99/faba9369a7ad6e4d10b6a5fbf71fa2a188fe4a593b15f0963b73859a1bbd/markupsafe-3.0.3-cp310-cp310-win32.whl", hash = "sha256:2a15a08b17dd94c53a1da0438822d70ebcd13f8c3a95abe3a9ef9f11a94830aa", size = 14571 }, - { url = "https://files.pythonhosted.org/packages/d6/25/55dc3ab959917602c96985cb1253efaa4ff42f71194bddeb61eb7278b8be/markupsafe-3.0.3-cp310-cp310-win_amd64.whl", hash = "sha256:c4ffb7ebf07cfe8931028e3e4c85f0357459a3f9f9490886198848f4fa002ec8", size = 15056 }, - { url = "https://files.pythonhosted.org/packages/d0/9e/0a02226640c255d1da0b8d12e24ac2aa6734da68bff14c05dd53b94a0fc3/markupsafe-3.0.3-cp310-cp310-win_arm64.whl", hash = "sha256:e2103a929dfa2fcaf9bb4e7c091983a49c9ac3b19c9061b6d5427dd7d14d81a1", size = 13932 }, - { url = "https://files.pythonhosted.org/packages/08/db/fefacb2136439fc8dd20e797950e749aa1f4997ed584c62cfb8ef7c2be0e/markupsafe-3.0.3-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:1cc7ea17a6824959616c525620e387f6dd30fec8cb44f649e31712db02123dad", size = 11631 }, - { url = "https://files.pythonhosted.org/packages/e1/2e/5898933336b61975ce9dc04decbc0a7f2fee78c30353c5efba7f2d6ff27a/markupsafe-3.0.3-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:4bd4cd07944443f5a265608cc6aab442e4f74dff8088b0dfc8238647b8f6ae9a", size = 12058 }, - { url = "https://files.pythonhosted.org/packages/1d/09/adf2df3699d87d1d8184038df46a9c80d78c0148492323f4693df54e17bb/markupsafe-3.0.3-cp311-cp311-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:6b5420a1d9450023228968e7e6a9ce57f65d148ab56d2313fcd589eee96a7a50", size = 24287 }, - { url = "https://files.pythonhosted.org/packages/30/ac/0273f6fcb5f42e314c6d8cd99effae6a5354604d461b8d392b5ec9530a54/markupsafe-3.0.3-cp311-cp311-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:0bf2a864d67e76e5c9a34dc26ec616a66b9888e25e7b9460e1c76d3293bd9dbf", size = 22940 }, - { url = "https://files.pythonhosted.org/packages/19/ae/31c1be199ef767124c042c6c3e904da327a2f7f0cd63a0337e1eca2967a8/markupsafe-3.0.3-cp311-cp311-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl", hash = "sha256:bc51efed119bc9cfdf792cdeaa4d67e8f6fcccab66ed4bfdd6bde3e59bfcbb2f", size = 21887 }, - { url = "https://files.pythonhosted.org/packages/b2/76/7edcab99d5349a4532a459e1fe64f0b0467a3365056ae550d3bcf3f79e1e/markupsafe-3.0.3-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:068f375c472b3e7acbe2d5318dea141359e6900156b5b2ba06a30b169086b91a", size = 23692 }, - { url = "https://files.pythonhosted.org/packages/a4/28/6e74cdd26d7514849143d69f0bf2399f929c37dc2b31e6829fd2045b2765/markupsafe-3.0.3-cp311-cp311-musllinux_1_2_riscv64.whl", hash = "sha256:7be7b61bb172e1ed687f1754f8e7484f1c8019780f6f6b0786e76bb01c2ae115", size = 21471 }, - { url = "https://files.pythonhosted.org/packages/62/7e/a145f36a5c2945673e590850a6f8014318d5577ed7e5920a4b3448e0865d/markupsafe-3.0.3-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:f9e130248f4462aaa8e2552d547f36ddadbeaa573879158d721bbd33dfe4743a", size = 22923 }, - { url = "https://files.pythonhosted.org/packages/0f/62/d9c46a7f5c9adbeeeda52f5b8d802e1094e9717705a645efc71b0913a0a8/markupsafe-3.0.3-cp311-cp311-win32.whl", hash = "sha256:0db14f5dafddbb6d9208827849fad01f1a2609380add406671a26386cdf15a19", size = 14572 }, - { url = "https://files.pythonhosted.org/packages/83/8a/4414c03d3f891739326e1783338e48fb49781cc915b2e0ee052aa490d586/markupsafe-3.0.3-cp311-cp311-win_amd64.whl", hash = "sha256:de8a88e63464af587c950061a5e6a67d3632e36df62b986892331d4620a35c01", size = 15077 }, - { url = "https://files.pythonhosted.org/packages/35/73/893072b42e6862f319b5207adc9ae06070f095b358655f077f69a35601f0/markupsafe-3.0.3-cp311-cp311-win_arm64.whl", hash = "sha256:3b562dd9e9ea93f13d53989d23a7e775fdfd1066c33494ff43f5418bc8c58a5c", size = 13876 }, - { url = "https://files.pythonhosted.org/packages/5a/72/147da192e38635ada20e0a2e1a51cf8823d2119ce8883f7053879c2199b5/markupsafe-3.0.3-cp312-cp312-macosx_10_13_x86_64.whl", hash = "sha256:d53197da72cc091b024dd97249dfc7794d6a56530370992a5e1a08983ad9230e", size = 11615 }, - { url = "https://files.pythonhosted.org/packages/9a/81/7e4e08678a1f98521201c3079f77db69fb552acd56067661f8c2f534a718/markupsafe-3.0.3-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:1872df69a4de6aead3491198eaf13810b565bdbeec3ae2dc8780f14458ec73ce", size = 12020 }, - { url = "https://files.pythonhosted.org/packages/1e/2c/799f4742efc39633a1b54a92eec4082e4f815314869865d876824c257c1e/markupsafe-3.0.3-cp312-cp312-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:3a7e8ae81ae39e62a41ec302f972ba6ae23a5c5396c8e60113e9066ef893da0d", size = 24332 }, - { url = "https://files.pythonhosted.org/packages/3c/2e/8d0c2ab90a8c1d9a24f0399058ab8519a3279d1bd4289511d74e909f060e/markupsafe-3.0.3-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:d6dd0be5b5b189d31db7cda48b91d7e0a9795f31430b7f271219ab30f1d3ac9d", size = 22947 }, - { url = "https://files.pythonhosted.org/packages/2c/54/887f3092a85238093a0b2154bd629c89444f395618842e8b0c41783898ea/markupsafe-3.0.3-cp312-cp312-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl", hash = "sha256:94c6f0bb423f739146aec64595853541634bde58b2135f27f61c1ffd1cd4d16a", size = 21962 }, - { url = "https://files.pythonhosted.org/packages/c9/2f/336b8c7b6f4a4d95e91119dc8521402461b74a485558d8f238a68312f11c/markupsafe-3.0.3-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:be8813b57049a7dc738189df53d69395eba14fb99345e0a5994914a3864c8a4b", size = 23760 }, - { url = "https://files.pythonhosted.org/packages/32/43/67935f2b7e4982ffb50a4d169b724d74b62a3964bc1a9a527f5ac4f1ee2b/markupsafe-3.0.3-cp312-cp312-musllinux_1_2_riscv64.whl", hash = "sha256:83891d0e9fb81a825d9a6d61e3f07550ca70a076484292a70fde82c4b807286f", size = 21529 }, - { url = "https://files.pythonhosted.org/packages/89/e0/4486f11e51bbba8b0c041098859e869e304d1c261e59244baa3d295d47b7/markupsafe-3.0.3-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:77f0643abe7495da77fb436f50f8dab76dbc6e5fd25d39589a0f1fe6548bfa2b", size = 23015 }, - { url = "https://files.pythonhosted.org/packages/2f/e1/78ee7a023dac597a5825441ebd17170785a9dab23de95d2c7508ade94e0e/markupsafe-3.0.3-cp312-cp312-win32.whl", hash = "sha256:d88b440e37a16e651bda4c7c2b930eb586fd15ca7406cb39e211fcff3bf3017d", size = 14540 }, - { url = "https://files.pythonhosted.org/packages/aa/5b/bec5aa9bbbb2c946ca2733ef9c4ca91c91b6a24580193e891b5f7dbe8e1e/markupsafe-3.0.3-cp312-cp312-win_amd64.whl", hash = "sha256:26a5784ded40c9e318cfc2bdb30fe164bdb8665ded9cd64d500a34fb42067b1c", size = 15105 }, - { url = "https://files.pythonhosted.org/packages/e5/f1/216fc1bbfd74011693a4fd837e7026152e89c4bcf3e77b6692fba9923123/markupsafe-3.0.3-cp312-cp312-win_arm64.whl", hash = "sha256:35add3b638a5d900e807944a078b51922212fb3dedb01633a8defc4b01a3c85f", size = 13906 }, - { url = "https://files.pythonhosted.org/packages/38/2f/907b9c7bbba283e68f20259574b13d005c121a0fa4c175f9bed27c4597ff/markupsafe-3.0.3-cp313-cp313-macosx_10_13_x86_64.whl", hash = "sha256:e1cf1972137e83c5d4c136c43ced9ac51d0e124706ee1c8aa8532c1287fa8795", size = 11622 }, - { url = "https://files.pythonhosted.org/packages/9c/d9/5f7756922cdd676869eca1c4e3c0cd0df60ed30199ffd775e319089cb3ed/markupsafe-3.0.3-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:116bb52f642a37c115f517494ea5feb03889e04df47eeff5b130b1808ce7c219", size = 12029 }, - { url = "https://files.pythonhosted.org/packages/00/07/575a68c754943058c78f30db02ee03a64b3c638586fba6a6dd56830b30a3/markupsafe-3.0.3-cp313-cp313-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:133a43e73a802c5562be9bbcd03d090aa5a1fe899db609c29e8c8d815c5f6de6", size = 24374 }, - { url = "https://files.pythonhosted.org/packages/a9/21/9b05698b46f218fc0e118e1f8168395c65c8a2c750ae2bab54fc4bd4e0e8/markupsafe-3.0.3-cp313-cp313-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:ccfcd093f13f0f0b7fdd0f198b90053bf7b2f02a3927a30e63f3ccc9df56b676", size = 22980 }, - { url = "https://files.pythonhosted.org/packages/7f/71/544260864f893f18b6827315b988c146b559391e6e7e8f7252839b1b846a/markupsafe-3.0.3-cp313-cp313-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl", hash = "sha256:509fa21c6deb7a7a273d629cf5ec029bc209d1a51178615ddf718f5918992ab9", size = 21990 }, - { url = "https://files.pythonhosted.org/packages/c2/28/b50fc2f74d1ad761af2f5dcce7492648b983d00a65b8c0e0cb457c82ebbe/markupsafe-3.0.3-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:a4afe79fb3de0b7097d81da19090f4df4f8d3a2b3adaa8764138aac2e44f3af1", size = 23784 }, - { url = "https://files.pythonhosted.org/packages/ed/76/104b2aa106a208da8b17a2fb72e033a5a9d7073c68f7e508b94916ed47a9/markupsafe-3.0.3-cp313-cp313-musllinux_1_2_riscv64.whl", hash = "sha256:795e7751525cae078558e679d646ae45574b47ed6e7771863fcc079a6171a0fc", size = 21588 }, - { url = "https://files.pythonhosted.org/packages/b5/99/16a5eb2d140087ebd97180d95249b00a03aa87e29cc224056274f2e45fd6/markupsafe-3.0.3-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:8485f406a96febb5140bfeca44a73e3ce5116b2501ac54fe953e488fb1d03b12", size = 23041 }, - { url = "https://files.pythonhosted.org/packages/19/bc/e7140ed90c5d61d77cea142eed9f9c303f4c4806f60a1044c13e3f1471d0/markupsafe-3.0.3-cp313-cp313-win32.whl", hash = "sha256:bdd37121970bfd8be76c5fb069c7751683bdf373db1ed6c010162b2a130248ed", size = 14543 }, - { url = "https://files.pythonhosted.org/packages/05/73/c4abe620b841b6b791f2edc248f556900667a5a1cf023a6646967ae98335/markupsafe-3.0.3-cp313-cp313-win_amd64.whl", hash = "sha256:9a1abfdc021a164803f4d485104931fb8f8c1efd55bc6b748d2f5774e78b62c5", size = 15113 }, - { url = "https://files.pythonhosted.org/packages/f0/3a/fa34a0f7cfef23cf9500d68cb7c32dd64ffd58a12b09225fb03dd37d5b80/markupsafe-3.0.3-cp313-cp313-win_arm64.whl", hash = "sha256:7e68f88e5b8799aa49c85cd116c932a1ac15caaa3f5db09087854d218359e485", size = 13911 }, - { url = "https://files.pythonhosted.org/packages/e4/d7/e05cd7efe43a88a17a37b3ae96e79a19e846f3f456fe79c57ca61356ef01/markupsafe-3.0.3-cp313-cp313t-macosx_10_13_x86_64.whl", hash = "sha256:218551f6df4868a8d527e3062d0fb968682fe92054e89978594c28e642c43a73", size = 11658 }, - { url = "https://files.pythonhosted.org/packages/99/9e/e412117548182ce2148bdeacdda3bb494260c0b0184360fe0d56389b523b/markupsafe-3.0.3-cp313-cp313t-macosx_11_0_arm64.whl", hash = "sha256:3524b778fe5cfb3452a09d31e7b5adefeea8c5be1d43c4f810ba09f2ceb29d37", size = 12066 }, - { url = "https://files.pythonhosted.org/packages/bc/e6/fa0ffcda717ef64a5108eaa7b4f5ed28d56122c9a6d70ab8b72f9f715c80/markupsafe-3.0.3-cp313-cp313t-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:4e885a3d1efa2eadc93c894a21770e4bc67899e3543680313b09f139e149ab19", size = 25639 }, - { url = "https://files.pythonhosted.org/packages/96/ec/2102e881fe9d25fc16cb4b25d5f5cde50970967ffa5dddafdb771237062d/markupsafe-3.0.3-cp313-cp313t-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:8709b08f4a89aa7586de0aadc8da56180242ee0ada3999749b183aa23df95025", size = 23569 }, - { url = "https://files.pythonhosted.org/packages/4b/30/6f2fce1f1f205fc9323255b216ca8a235b15860c34b6798f810f05828e32/markupsafe-3.0.3-cp313-cp313t-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl", hash = "sha256:b8512a91625c9b3da6f127803b166b629725e68af71f8184ae7e7d54686a56d6", size = 23284 }, - { url = "https://files.pythonhosted.org/packages/58/47/4a0ccea4ab9f5dcb6f79c0236d954acb382202721e704223a8aafa38b5c8/markupsafe-3.0.3-cp313-cp313t-musllinux_1_2_aarch64.whl", hash = "sha256:9b79b7a16f7fedff2495d684f2b59b0457c3b493778c9eed31111be64d58279f", size = 24801 }, - { url = "https://files.pythonhosted.org/packages/6a/70/3780e9b72180b6fecb83a4814d84c3bf4b4ae4bf0b19c27196104149734c/markupsafe-3.0.3-cp313-cp313t-musllinux_1_2_riscv64.whl", hash = "sha256:12c63dfb4a98206f045aa9563db46507995f7ef6d83b2f68eda65c307c6829eb", size = 22769 }, - { url = "https://files.pythonhosted.org/packages/98/c5/c03c7f4125180fc215220c035beac6b9cb684bc7a067c84fc69414d315f5/markupsafe-3.0.3-cp313-cp313t-musllinux_1_2_x86_64.whl", hash = "sha256:8f71bc33915be5186016f675cd83a1e08523649b0e33efdb898db577ef5bb009", size = 23642 }, - { url = "https://files.pythonhosted.org/packages/80/d6/2d1b89f6ca4bff1036499b1e29a1d02d282259f3681540e16563f27ebc23/markupsafe-3.0.3-cp313-cp313t-win32.whl", hash = "sha256:69c0b73548bc525c8cb9a251cddf1931d1db4d2258e9599c28c07ef3580ef354", size = 14612 }, - { url = "https://files.pythonhosted.org/packages/2b/98/e48a4bfba0a0ffcf9925fe2d69240bfaa19c6f7507b8cd09c70684a53c1e/markupsafe-3.0.3-cp313-cp313t-win_amd64.whl", hash = "sha256:1b4b79e8ebf6b55351f0d91fe80f893b4743f104bff22e90697db1590e47a218", size = 15200 }, - { url = "https://files.pythonhosted.org/packages/0e/72/e3cc540f351f316e9ed0f092757459afbc595824ca724cbc5a5d4263713f/markupsafe-3.0.3-cp313-cp313t-win_arm64.whl", hash = "sha256:ad2cf8aa28b8c020ab2fc8287b0f823d0a7d8630784c31e9ee5edea20f406287", size = 13973 }, - { url = "https://files.pythonhosted.org/packages/33/8a/8e42d4838cd89b7dde187011e97fe6c3af66d8c044997d2183fbd6d31352/markupsafe-3.0.3-cp314-cp314-macosx_10_13_x86_64.whl", hash = "sha256:eaa9599de571d72e2daf60164784109f19978b327a3910d3e9de8c97b5b70cfe", size = 11619 }, - { url = "https://files.pythonhosted.org/packages/b5/64/7660f8a4a8e53c924d0fa05dc3a55c9cee10bbd82b11c5afb27d44b096ce/markupsafe-3.0.3-cp314-cp314-macosx_11_0_arm64.whl", hash = "sha256:c47a551199eb8eb2121d4f0f15ae0f923d31350ab9280078d1e5f12b249e0026", size = 12029 }, - { url = "https://files.pythonhosted.org/packages/da/ef/e648bfd021127bef5fa12e1720ffed0c6cbb8310c8d9bea7266337ff06de/markupsafe-3.0.3-cp314-cp314-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:f34c41761022dd093b4b6896d4810782ffbabe30f2d443ff5f083e0cbbb8c737", size = 24408 }, - { url = "https://files.pythonhosted.org/packages/41/3c/a36c2450754618e62008bf7435ccb0f88053e07592e6028a34776213d877/markupsafe-3.0.3-cp314-cp314-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:457a69a9577064c05a97c41f4e65148652db078a3a509039e64d3467b9e7ef97", size = 23005 }, - { url = "https://files.pythonhosted.org/packages/bc/20/b7fdf89a8456b099837cd1dc21974632a02a999ec9bf7ca3e490aacd98e7/markupsafe-3.0.3-cp314-cp314-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl", hash = "sha256:e8afc3f2ccfa24215f8cb28dcf43f0113ac3c37c2f0f0806d8c70e4228c5cf4d", size = 22048 }, - { url = "https://files.pythonhosted.org/packages/9a/a7/591f592afdc734f47db08a75793a55d7fbcc6902a723ae4cfbab61010cc5/markupsafe-3.0.3-cp314-cp314-musllinux_1_2_aarch64.whl", hash = "sha256:ec15a59cf5af7be74194f7ab02d0f59a62bdcf1a537677ce67a2537c9b87fcda", size = 23821 }, - { url = "https://files.pythonhosted.org/packages/7d/33/45b24e4f44195b26521bc6f1a82197118f74df348556594bd2262bda1038/markupsafe-3.0.3-cp314-cp314-musllinux_1_2_riscv64.whl", hash = "sha256:0eb9ff8191e8498cca014656ae6b8d61f39da5f95b488805da4bb029cccbfbaf", size = 21606 }, - { url = "https://files.pythonhosted.org/packages/ff/0e/53dfaca23a69fbfbbf17a4b64072090e70717344c52eaaaa9c5ddff1e5f0/markupsafe-3.0.3-cp314-cp314-musllinux_1_2_x86_64.whl", hash = "sha256:2713baf880df847f2bece4230d4d094280f4e67b1e813eec43b4c0e144a34ffe", size = 23043 }, - { url = "https://files.pythonhosted.org/packages/46/11/f333a06fc16236d5238bfe74daccbca41459dcd8d1fa952e8fbd5dccfb70/markupsafe-3.0.3-cp314-cp314-win32.whl", hash = "sha256:729586769a26dbceff69f7a7dbbf59ab6572b99d94576a5592625d5b411576b9", size = 14747 }, - { url = "https://files.pythonhosted.org/packages/28/52/182836104b33b444e400b14f797212f720cbc9ed6ba34c800639d154e821/markupsafe-3.0.3-cp314-cp314-win_amd64.whl", hash = "sha256:bdc919ead48f234740ad807933cdf545180bfbe9342c2bb451556db2ed958581", size = 15341 }, - { url = "https://files.pythonhosted.org/packages/6f/18/acf23e91bd94fd7b3031558b1f013adfa21a8e407a3fdb32745538730382/markupsafe-3.0.3-cp314-cp314-win_arm64.whl", hash = "sha256:5a7d5dc5140555cf21a6fefbdbf8723f06fcd2f63ef108f2854de715e4422cb4", size = 14073 }, - { url = "https://files.pythonhosted.org/packages/3c/f0/57689aa4076e1b43b15fdfa646b04653969d50cf30c32a102762be2485da/markupsafe-3.0.3-cp314-cp314t-macosx_10_13_x86_64.whl", hash = "sha256:1353ef0c1b138e1907ae78e2f6c63ff67501122006b0f9abad68fda5f4ffc6ab", size = 11661 }, - { url = "https://files.pythonhosted.org/packages/89/c3/2e67a7ca217c6912985ec766c6393b636fb0c2344443ff9d91404dc4c79f/markupsafe-3.0.3-cp314-cp314t-macosx_11_0_arm64.whl", hash = "sha256:1085e7fbddd3be5f89cc898938f42c0b3c711fdcb37d75221de2666af647c175", size = 12069 }, - { url = "https://files.pythonhosted.org/packages/f0/00/be561dce4e6ca66b15276e184ce4b8aec61fe83662cce2f7d72bd3249d28/markupsafe-3.0.3-cp314-cp314t-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:1b52b4fb9df4eb9ae465f8d0c228a00624de2334f216f178a995ccdcf82c4634", size = 25670 }, - { url = "https://files.pythonhosted.org/packages/50/09/c419f6f5a92e5fadde27efd190eca90f05e1261b10dbd8cbcb39cd8ea1dc/markupsafe-3.0.3-cp314-cp314t-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:fed51ac40f757d41b7c48425901843666a6677e3e8eb0abcff09e4ba6e664f50", size = 23598 }, - { url = "https://files.pythonhosted.org/packages/22/44/a0681611106e0b2921b3033fc19bc53323e0b50bc70cffdd19f7d679bb66/markupsafe-3.0.3-cp314-cp314t-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl", hash = "sha256:f190daf01f13c72eac4efd5c430a8de82489d9cff23c364c3ea822545032993e", size = 23261 }, - { url = "https://files.pythonhosted.org/packages/5f/57/1b0b3f100259dc9fffe780cfb60d4be71375510e435efec3d116b6436d43/markupsafe-3.0.3-cp314-cp314t-musllinux_1_2_aarch64.whl", hash = "sha256:e56b7d45a839a697b5eb268c82a71bd8c7f6c94d6fd50c3d577fa39a9f1409f5", size = 24835 }, - { url = "https://files.pythonhosted.org/packages/26/6a/4bf6d0c97c4920f1597cc14dd720705eca0bf7c787aebc6bb4d1bead5388/markupsafe-3.0.3-cp314-cp314t-musllinux_1_2_riscv64.whl", hash = "sha256:f3e98bb3798ead92273dc0e5fd0f31ade220f59a266ffd8a4f6065e0a3ce0523", size = 22733 }, - { url = "https://files.pythonhosted.org/packages/14/c7/ca723101509b518797fedc2fdf79ba57f886b4aca8a7d31857ba3ee8281f/markupsafe-3.0.3-cp314-cp314t-musllinux_1_2_x86_64.whl", hash = "sha256:5678211cb9333a6468fb8d8be0305520aa073f50d17f089b5b4b477ea6e67fdc", size = 23672 }, - { url = "https://files.pythonhosted.org/packages/fb/df/5bd7a48c256faecd1d36edc13133e51397e41b73bb77e1a69deab746ebac/markupsafe-3.0.3-cp314-cp314t-win32.whl", hash = "sha256:915c04ba3851909ce68ccc2b8e2cd691618c4dc4c4232fb7982bca3f41fd8c3d", size = 14819 }, - { url = "https://files.pythonhosted.org/packages/1a/8a/0402ba61a2f16038b48b39bccca271134be00c5c9f0f623208399333c448/markupsafe-3.0.3-cp314-cp314t-win_amd64.whl", hash = "sha256:4faffd047e07c38848ce017e8725090413cd80cbc23d86e55c587bf979e579c9", size = 15426 }, - { url = "https://files.pythonhosted.org/packages/70/bc/6f1c2f612465f5fa89b95bead1f44dcb607670fd42891d8fdcd5d039f4f4/markupsafe-3.0.3-cp314-cp314t-win_arm64.whl", hash = "sha256:32001d6a8fc98c8cb5c947787c5d08b0a50663d139f1305bac5885d98d9b40fa", size = 14146 }, -] - -[[package]] -name = "marshmallow" -version = "4.1.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "backports-datetime-fromisoformat", marker = "python_full_version < '3.11'" }, - { name = "typing-extensions", marker = "python_full_version < '3.11'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/e9/e1/5edfd1edf05d3cc98415b0810ca45fa19d7dee6def0d0ec639eb4eb14e20/marshmallow-4.1.2.tar.gz", hash = "sha256:083f250643d2e75fd363f256aeb6b1af369a7513ad37647ce4a601f6966e3ba5", size = 220974 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/af/b6/66d1748fb45453e337c8a334dafed7b818e72ac9cf9d105a56e0cf21865f/marshmallow-4.1.2-py3-none-any.whl", hash = "sha256:a8cfa18bd8d0e5f7339e734edf84815fe8db1bdb57358c7ccc05472b746eeadc", size = 48360 }, -] - -[[package]] -name = "matplotlib" -version = "3.10.8" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "contourpy", version = "1.3.2", source = { registry = "https://pypi.org/simple" }, marker = "python_full_version < '3.11'" }, - { name = "contourpy", version = "1.3.3", source = { registry = "https://pypi.org/simple" }, marker = "python_full_version >= '3.11'" }, - { name = "cycler" }, - { name = "fonttools" }, - { name = "kiwisolver" }, - { name = "numpy" }, - { name = "packaging" }, - { name = "pillow" }, - { name = "pyparsing" }, - { name = "python-dateutil" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/8a/76/d3c6e3a13fe484ebe7718d14e269c9569c4eb0020a968a327acb3b9a8fe6/matplotlib-3.10.8.tar.gz", hash = "sha256:2299372c19d56bcd35cf05a2738308758d32b9eaed2371898d8f5bd33f084aa3", size = 34806269 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/58/be/a30bd917018ad220c400169fba298f2bb7003c8ccbc0c3e24ae2aacad1e8/matplotlib-3.10.8-cp310-cp310-macosx_10_12_x86_64.whl", hash = "sha256:00270d217d6b20d14b584c521f810d60c5c78406dc289859776550df837dcda7", size = 8239828 }, - { url = "https://files.pythonhosted.org/packages/58/27/ca01e043c4841078e82cf6e80a6993dfecd315c3d79f5f3153afbb8e1ec6/matplotlib-3.10.8-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:37b3c1cc42aa184b3f738cfa18c1c1d72fd496d85467a6cf7b807936d39aa656", size = 8128050 }, - { url = "https://files.pythonhosted.org/packages/cb/aa/7ab67f2b729ae6a91bcf9dcac0affb95fb8c56f7fd2b2af894ae0b0cf6fa/matplotlib-3.10.8-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:ee40c27c795bda6a5292e9cff9890189d32f7e3a0bf04e0e3c9430c4a00c37df", size = 8700452 }, - { url = "https://files.pythonhosted.org/packages/73/ae/2d5817b0acee3c49b7e7ccfbf5b273f284957cc8e270adf36375db353190/matplotlib-3.10.8-cp310-cp310-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:a48f2b74020919552ea25d222d5cc6af9ca3f4eb43a93e14d068457f545c2a17", size = 9534928 }, - { url = "https://files.pythonhosted.org/packages/c9/5b/8e66653e9f7c39cb2e5cab25fce4810daffa2bff02cbf5f3077cea9e942c/matplotlib-3.10.8-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:f254d118d14a7f99d616271d6c3c27922c092dac11112670b157798b89bf4933", size = 9586377 }, - { url = "https://files.pythonhosted.org/packages/e2/e2/fd0bbadf837f81edb0d208ba8f8cb552874c3b16e27cb91a31977d90875d/matplotlib-3.10.8-cp310-cp310-win_amd64.whl", hash = "sha256:f9b587c9c7274c1613a30afabf65a272114cd6cdbe67b3406f818c79d7ab2e2a", size = 8128127 }, - { url = "https://files.pythonhosted.org/packages/f8/86/de7e3a1cdcfc941483af70609edc06b83e7c8a0e0dc9ac325200a3f4d220/matplotlib-3.10.8-cp311-cp311-macosx_10_12_x86_64.whl", hash = "sha256:6be43b667360fef5c754dda5d25a32e6307a03c204f3c0fc5468b78fa87b4160", size = 8251215 }, - { url = "https://files.pythonhosted.org/packages/fd/14/baad3222f424b19ce6ad243c71de1ad9ec6b2e4eb1e458a48fdc6d120401/matplotlib-3.10.8-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:a2b336e2d91a3d7006864e0990c83b216fcdca64b5a6484912902cef87313d78", size = 8139625 }, - { url = "https://files.pythonhosted.org/packages/8f/a0/7024215e95d456de5883e6732e708d8187d9753a21d32f8ddb3befc0c445/matplotlib-3.10.8-cp311-cp311-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:efb30e3baaea72ce5928e32bab719ab4770099079d66726a62b11b1ef7273be4", size = 8712614 }, - { url = "https://files.pythonhosted.org/packages/5a/f4/b8347351da9a5b3f41e26cf547252d861f685c6867d179a7c9d60ad50189/matplotlib-3.10.8-cp311-cp311-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:d56a1efd5bfd61486c8bc968fa18734464556f0fb8e51690f4ac25d85cbbbbc2", size = 9540997 }, - { url = "https://files.pythonhosted.org/packages/9e/c0/c7b914e297efe0bc36917bf216b2acb91044b91e930e878ae12981e461e5/matplotlib-3.10.8-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:238b7ce5717600615c895050239ec955d91f321c209dd110db988500558e70d6", size = 9596825 }, - { url = "https://files.pythonhosted.org/packages/6f/d3/a4bbc01c237ab710a1f22b4da72f4ff6d77eb4c7735ea9811a94ae239067/matplotlib-3.10.8-cp311-cp311-win_amd64.whl", hash = "sha256:18821ace09c763ec93aef5eeff087ee493a24051936d7b9ebcad9662f66501f9", size = 8135090 }, - { url = "https://files.pythonhosted.org/packages/89/dd/a0b6588f102beab33ca6f5218b31725216577b2a24172f327eaf6417d5c9/matplotlib-3.10.8-cp311-cp311-win_arm64.whl", hash = "sha256:bab485bcf8b1c7d2060b4fcb6fc368a9e6f4cd754c9c2fea281f4be21df394a2", size = 8012377 }, - { url = "https://files.pythonhosted.org/packages/9e/67/f997cdcbb514012eb0d10cd2b4b332667997fb5ebe26b8d41d04962fa0e6/matplotlib-3.10.8-cp312-cp312-macosx_10_13_x86_64.whl", hash = "sha256:64fcc24778ca0404ce0cb7b6b77ae1f4c7231cdd60e6778f999ee05cbd581b9a", size = 8260453 }, - { url = "https://files.pythonhosted.org/packages/7e/65/07d5f5c7f7c994f12c768708bd2e17a4f01a2b0f44a1c9eccad872433e2e/matplotlib-3.10.8-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:b9a5ca4ac220a0cdd1ba6bcba3608547117d30468fefce49bb26f55c1a3d5c58", size = 8148321 }, - { url = "https://files.pythonhosted.org/packages/3e/f3/c5195b1ae57ef85339fd7285dfb603b22c8b4e79114bae5f4f0fcf688677/matplotlib-3.10.8-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:3ab4aabc72de4ff77b3ec33a6d78a68227bf1123465887f9905ba79184a1cc04", size = 8716944 }, - { url = "https://files.pythonhosted.org/packages/00/f9/7638f5cc82ec8a7aa005de48622eecc3ed7c9854b96ba15bd76b7fd27574/matplotlib-3.10.8-cp312-cp312-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:24d50994d8c5816ddc35411e50a86ab05f575e2530c02752e02538122613371f", size = 9550099 }, - { url = "https://files.pythonhosted.org/packages/57/61/78cd5920d35b29fd2a0fe894de8adf672ff52939d2e9b43cb83cd5ce1bc7/matplotlib-3.10.8-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:99eefd13c0dc3b3c1b4d561c1169e65fe47aab7b8158754d7c084088e2329466", size = 9613040 }, - { url = "https://files.pythonhosted.org/packages/30/4e/c10f171b6e2f44d9e3a2b96efa38b1677439d79c99357600a62cc1e9594e/matplotlib-3.10.8-cp312-cp312-win_amd64.whl", hash = "sha256:dd80ecb295460a5d9d260df63c43f4afbdd832d725a531f008dad1664f458adf", size = 8142717 }, - { url = "https://files.pythonhosted.org/packages/f1/76/934db220026b5fef85f45d51a738b91dea7d70207581063cd9bd8fafcf74/matplotlib-3.10.8-cp312-cp312-win_arm64.whl", hash = "sha256:3c624e43ed56313651bc18a47f838b60d7b8032ed348911c54906b130b20071b", size = 8012751 }, - { url = "https://files.pythonhosted.org/packages/3d/b9/15fd5541ef4f5b9a17eefd379356cf12175fe577424e7b1d80676516031a/matplotlib-3.10.8-cp313-cp313-macosx_10_13_x86_64.whl", hash = "sha256:3f2e409836d7f5ac2f1c013110a4d50b9f7edc26328c108915f9075d7d7a91b6", size = 8261076 }, - { url = "https://files.pythonhosted.org/packages/8d/a0/2ba3473c1b66b9c74dc7107c67e9008cb1782edbe896d4c899d39ae9cf78/matplotlib-3.10.8-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:56271f3dac49a88d7fca5060f004d9d22b865f743a12a23b1e937a0be4818ee1", size = 8148794 }, - { url = "https://files.pythonhosted.org/packages/75/97/a471f1c3eb1fd6f6c24a31a5858f443891d5127e63a7788678d14e249aea/matplotlib-3.10.8-cp313-cp313-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:a0a7f52498f72f13d4a25ea70f35f4cb60642b466cbb0a9be951b5bc3f45a486", size = 8718474 }, - { url = "https://files.pythonhosted.org/packages/01/be/cd478f4b66f48256f42927d0acbcd63a26a893136456cd079c0cc24fbabf/matplotlib-3.10.8-cp313-cp313-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:646d95230efb9ca614a7a594d4fcacde0ac61d25e37dd51710b36477594963ce", size = 9549637 }, - { url = "https://files.pythonhosted.org/packages/5d/7c/8dc289776eae5109e268c4fb92baf870678dc048a25d4ac903683b86d5bf/matplotlib-3.10.8-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:f89c151aab2e2e23cb3fe0acad1e8b82841fd265379c4cecd0f3fcb34c15e0f6", size = 9613678 }, - { url = "https://files.pythonhosted.org/packages/64/40/37612487cc8a437d4dd261b32ca21fe2d79510fe74af74e1f42becb1bdb8/matplotlib-3.10.8-cp313-cp313-win_amd64.whl", hash = "sha256:e8ea3e2d4066083e264e75c829078f9e149fa119d27e19acd503de65e0b13149", size = 8142686 }, - { url = "https://files.pythonhosted.org/packages/66/52/8d8a8730e968185514680c2a6625943f70269509c3dcfc0dcf7d75928cb8/matplotlib-3.10.8-cp313-cp313-win_arm64.whl", hash = "sha256:c108a1d6fa78a50646029cb6d49808ff0fc1330fda87fa6f6250c6b5369b6645", size = 8012917 }, - { url = "https://files.pythonhosted.org/packages/b5/27/51fe26e1062f298af5ef66343d8ef460e090a27fea73036c76c35821df04/matplotlib-3.10.8-cp313-cp313t-macosx_10_13_x86_64.whl", hash = "sha256:ad3d9833a64cf48cc4300f2b406c3d0f4f4724a91c0bd5640678a6ba7c102077", size = 8305679 }, - { url = "https://files.pythonhosted.org/packages/2c/1e/4de865bc591ac8e3062e835f42dd7fe7a93168d519557837f0e37513f629/matplotlib-3.10.8-cp313-cp313t-macosx_11_0_arm64.whl", hash = "sha256:eb3823f11823deade26ce3b9f40dcb4a213da7a670013929f31d5f5ed1055b22", size = 8198336 }, - { url = "https://files.pythonhosted.org/packages/c6/cb/2f7b6e75fb4dce87ef91f60cac4f6e34f4c145ab036a22318ec837971300/matplotlib-3.10.8-cp313-cp313t-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:d9050fee89a89ed57b4fb2c1bfac9a3d0c57a0d55aed95949eedbc42070fea39", size = 8731653 }, - { url = "https://files.pythonhosted.org/packages/46/b3/bd9c57d6ba670a37ab31fb87ec3e8691b947134b201f881665b28cc039ff/matplotlib-3.10.8-cp313-cp313t-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:b44d07310e404ba95f8c25aa5536f154c0a8ec473303535949e52eb71d0a1565", size = 9561356 }, - { url = "https://files.pythonhosted.org/packages/c0/3d/8b94a481456dfc9dfe6e39e93b5ab376e50998cddfd23f4ae3b431708f16/matplotlib-3.10.8-cp313-cp313t-musllinux_1_2_x86_64.whl", hash = "sha256:0a33deb84c15ede243aead39f77e990469fff93ad1521163305095b77b72ce4a", size = 9614000 }, - { url = "https://files.pythonhosted.org/packages/bd/cd/bc06149fe5585ba800b189a6a654a75f1f127e8aab02fd2be10df7fa500c/matplotlib-3.10.8-cp313-cp313t-win_amd64.whl", hash = "sha256:3a48a78d2786784cc2413e57397981fb45c79e968d99656706018d6e62e57958", size = 8220043 }, - { url = "https://files.pythonhosted.org/packages/e3/de/b22cf255abec916562cc04eef457c13e58a1990048de0c0c3604d082355e/matplotlib-3.10.8-cp313-cp313t-win_arm64.whl", hash = "sha256:15d30132718972c2c074cd14638c7f4592bd98719e2308bccea40e0538bc0cb5", size = 8062075 }, - { url = "https://files.pythonhosted.org/packages/3c/43/9c0ff7a2f11615e516c3b058e1e6e8f9614ddeca53faca06da267c48345d/matplotlib-3.10.8-cp314-cp314-macosx_10_13_x86_64.whl", hash = "sha256:b53285e65d4fa4c86399979e956235deb900be5baa7fc1218ea67fbfaeaadd6f", size = 8262481 }, - { url = "https://files.pythonhosted.org/packages/6f/ca/e8ae28649fcdf039fda5ef554b40a95f50592a3c47e6f7270c9561c12b07/matplotlib-3.10.8-cp314-cp314-macosx_11_0_arm64.whl", hash = "sha256:32f8dce744be5569bebe789e46727946041199030db8aeb2954d26013a0eb26b", size = 8151473 }, - { url = "https://files.pythonhosted.org/packages/f1/6f/009d129ae70b75e88cbe7e503a12a4c0670e08ed748a902c2568909e9eb5/matplotlib-3.10.8-cp314-cp314-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:4cf267add95b1c88300d96ca837833d4112756045364f5c734a2276038dae27d", size = 9553896 }, - { url = "https://files.pythonhosted.org/packages/f5/26/4221a741eb97967bc1fd5e4c52b9aa5a91b2f4ec05b59f6def4d820f9df9/matplotlib-3.10.8-cp314-cp314-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:2cf5bd12cecf46908f286d7838b2abc6c91cda506c0445b8223a7c19a00df008", size = 9824193 }, - { url = "https://files.pythonhosted.org/packages/1f/f3/3abf75f38605772cf48a9daf5821cd4f563472f38b4b828c6fba6fa6d06e/matplotlib-3.10.8-cp314-cp314-musllinux_1_2_x86_64.whl", hash = "sha256:41703cc95688f2516b480f7f339d8851a6035f18e100ee6a32bc0b8536a12a9c", size = 9615444 }, - { url = "https://files.pythonhosted.org/packages/93/a5/de89ac80f10b8dc615807ee1133cd99ac74082581196d4d9590bea10690d/matplotlib-3.10.8-cp314-cp314-win_amd64.whl", hash = "sha256:83d282364ea9f3e52363da262ce32a09dfe241e4080dcedda3c0db059d3c1f11", size = 8272719 }, - { url = "https://files.pythonhosted.org/packages/69/ce/b006495c19ccc0a137b48083168a37bd056392dee02f87dba0472f2797fe/matplotlib-3.10.8-cp314-cp314-win_arm64.whl", hash = "sha256:2c1998e92cd5999e295a731bcb2911c75f597d937341f3030cc24ef2733d78a8", size = 8144205 }, - { url = "https://files.pythonhosted.org/packages/68/d9/b31116a3a855bd313c6fcdb7226926d59b041f26061c6c5b1be66a08c826/matplotlib-3.10.8-cp314-cp314t-macosx_10_13_x86_64.whl", hash = "sha256:b5a2b97dbdc7d4f353ebf343744f1d1f1cca8aa8bfddb4262fcf4306c3761d50", size = 8305785 }, - { url = "https://files.pythonhosted.org/packages/1e/90/6effe8103f0272685767ba5f094f453784057072f49b393e3ea178fe70a5/matplotlib-3.10.8-cp314-cp314t-macosx_11_0_arm64.whl", hash = "sha256:3f5c3e4da343bba819f0234186b9004faba952cc420fbc522dc4e103c1985908", size = 8198361 }, - { url = "https://files.pythonhosted.org/packages/d7/65/a73188711bea603615fc0baecca1061429ac16940e2385433cc778a9d8e7/matplotlib-3.10.8-cp314-cp314t-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:5f62550b9a30afde8c1c3ae450e5eb547d579dd69b25c2fc7a1c67f934c1717a", size = 9561357 }, - { url = "https://files.pythonhosted.org/packages/f4/3d/b5c5d5d5be8ce63292567f0e2c43dde9953d3ed86ac2de0a72e93c8f07a1/matplotlib-3.10.8-cp314-cp314t-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:495672de149445ec1b772ff2c9ede9b769e3cb4f0d0aa7fa730d7f59e2d4e1c1", size = 9823610 }, - { url = "https://files.pythonhosted.org/packages/4d/4b/e7beb6bbd49f6bae727a12b270a2654d13c397576d25bd6786e47033300f/matplotlib-3.10.8-cp314-cp314t-musllinux_1_2_x86_64.whl", hash = "sha256:595ba4d8fe983b88f0eec8c26a241e16d6376fe1979086232f481f8f3f67494c", size = 9614011 }, - { url = "https://files.pythonhosted.org/packages/7c/e6/76f2813d31f032e65f6f797e3f2f6e4aab95b65015924b1c51370395c28a/matplotlib-3.10.8-cp314-cp314t-win_amd64.whl", hash = "sha256:25d380fe8b1dc32cf8f0b1b448470a77afb195438bafdf1d858bfb876f3edf7b", size = 8362801 }, - { url = "https://files.pythonhosted.org/packages/5d/49/d651878698a0b67f23aa28e17f45a6d6dd3d3f933fa29087fa4ce5947b5a/matplotlib-3.10.8-cp314-cp314t-win_arm64.whl", hash = "sha256:113bb52413ea508ce954a02c10ffd0d565f9c3bc7f2eddc27dfe1731e71c7b5f", size = 8192560 }, - { url = "https://files.pythonhosted.org/packages/f5/43/31d59500bb950b0d188e149a2e552040528c13d6e3d6e84d0cccac593dcd/matplotlib-3.10.8-pp310-pypy310_pp73-macosx_10_15_x86_64.whl", hash = "sha256:f97aeb209c3d2511443f8797e3e5a569aebb040d4f8bc79aa3ee78a8fb9e3dd8", size = 8237252 }, - { url = "https://files.pythonhosted.org/packages/0c/2c/615c09984f3c5f907f51c886538ad785cf72e0e11a3225de2c0f9442aecc/matplotlib-3.10.8-pp310-pypy310_pp73-macosx_11_0_arm64.whl", hash = "sha256:fb061f596dad3a0f52b60dc6a5dec4a0c300dec41e058a7efe09256188d170b7", size = 8124693 }, - { url = "https://files.pythonhosted.org/packages/91/e1/2757277a1c56041e1fc104b51a0f7b9a4afc8eb737865d63cababe30bc61/matplotlib-3.10.8-pp310-pypy310_pp73-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:12d90df9183093fcd479f4172ac26b322b1248b15729cb57f42f71f24c7e37a3", size = 8702205 }, - { url = "https://files.pythonhosted.org/packages/04/30/3afaa31c757f34b7725ab9d2ba8b48b5e89c2019c003e7d0ead143aabc5a/matplotlib-3.10.8-pp311-pypy311_pp73-macosx_10_15_x86_64.whl", hash = "sha256:6da7c2ce169267d0d066adcf63758f0604aa6c3eebf67458930f9d9b79ad1db1", size = 8249198 }, - { url = "https://files.pythonhosted.org/packages/48/2f/6334aec331f57485a642a7c8be03cb286f29111ae71c46c38b363230063c/matplotlib-3.10.8-pp311-pypy311_pp73-macosx_11_0_arm64.whl", hash = "sha256:9153c3292705be9f9c64498a8872118540c3f4123d1a1c840172edf262c8be4a", size = 8136817 }, - { url = "https://files.pythonhosted.org/packages/73/e4/6d6f14b2a759c622f191b2d67e9075a3f56aaccb3be4bb9bb6890030d0a0/matplotlib-3.10.8-pp311-pypy311_pp73-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:1ae029229a57cd1e8fe542485f27e7ca7b23aa9e8944ddb4985d0bc444f1eca2", size = 8713867 }, -] - -[[package]] -name = "matplotlib-inline" -version = "0.2.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "traitlets" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/c7/74/97e72a36efd4ae2bccb3463284300f8953f199b5ffbc04cbbb0ec78f74b1/matplotlib_inline-0.2.1.tar.gz", hash = "sha256:e1ee949c340d771fc39e241ea75683deb94762c8fa5f2927ec57c83c4dffa9fe", size = 8110 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/af/33/ee4519fa02ed11a94aef9559552f3b17bb863f2ecfe1a35dc7f548cde231/matplotlib_inline-0.2.1-py3-none-any.whl", hash = "sha256:d56ce5156ba6085e00a9d54fead6ed29a9c47e215cd1bba2e976ef39f5710a76", size = 9516 }, -] - -[[package]] -name = "mdurl" -version = "0.1.2" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/d6/54/cfe61301667036ec958cb99bd3efefba235e65cdeb9c84d24a8293ba1d90/mdurl-0.1.2.tar.gz", hash = "sha256:bb413d29f5eea38f31dd4754dd7377d4465116fb207585f97bf925588687c1ba", size = 8729 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/b3/38/89ba8ad64ae25be8de66a6d463314cf1eb366222074cfda9ee839c56a4b4/mdurl-0.1.2-py3-none-any.whl", hash = "sha256:84008a41e51615a49fc9966191ff91509e3c40b939176e643fd50a5c2196b8f8", size = 9979 }, -] - -[[package]] -name = "mediapy" -version = "1.1.6" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "ipython", version = "8.37.0", source = { registry = "https://pypi.org/simple" }, marker = "python_full_version < '3.11'" }, - { name = "ipython", version = "9.8.0", source = { registry = "https://pypi.org/simple" }, marker = "python_full_version >= '3.11'" }, - { name = "matplotlib" }, - { name = "numpy" }, - { name = "pillow" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/9e/b2/451be65c13d2d69b7601eded7ddd3f150884486715a9b3a705ffb08d0177/mediapy-1.1.6.tar.gz", hash = "sha256:9f44b760400964d8bea5121a213f94dc9a225d026d6a819901283a695e585634", size = 25459 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/e3/a0/0d55c59ea8a5f1b13b3eb931e1c0eab9c2a07322cad79cb51a596d2d2a5c/mediapy-1.1.6-py3-none-any.whl", hash = "sha256:c74370808b445666f95272bfdf0eb5707a43b7e05e5527f2dd0830e6892f976f", size = 24955 }, -] - -[[package]] -name = "ml-dtypes" -version = "0.4.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "numpy" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/fd/15/76f86faa0902836cc133939732f7611ace68cf54148487a99c539c272dc8/ml_dtypes-0.4.1.tar.gz", hash = "sha256:fad5f2de464fd09127e49b7fd1252b9006fb43d2edc1ff112d390c324af5ca7a", size = 692594 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/56/9e/76b84f77c7afee3b116dc8407903a2d5004ba3059a8f3dcdcfa6ebf33fff/ml_dtypes-0.4.1-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:1fe8b5b5e70cd67211db94b05cfd58dace592f24489b038dc6f9fe347d2e07d5", size = 397975 }, - { url = "https://files.pythonhosted.org/packages/03/7b/32650e1b2a2713a5923a0af2a8503d0d4a8fc99d1e1e0a1c40e996634460/ml_dtypes-0.4.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8c09a6d11d8475c2a9fd2bc0695628aec105f97cab3b3a3fb7c9660348ff7d24", size = 2182570 }, - { url = "https://files.pythonhosted.org/packages/16/86/a9f7569e7e4f5395f927de38a13b92efa73f809285d04f2923b291783dd2/ml_dtypes-0.4.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9f5e8f75fa371020dd30f9196e7d73babae2abd51cf59bdd56cb4f8de7e13354", size = 2160365 }, - { url = "https://files.pythonhosted.org/packages/04/1b/9a3afb437702503514f3934ec8d7904270edf013d28074f3e700e5dfbb0f/ml_dtypes-0.4.1-cp310-cp310-win_amd64.whl", hash = "sha256:15fdd922fea57e493844e5abb930b9c0bd0af217d9edd3724479fc3d7ce70e3f", size = 126633 }, - { url = "https://files.pythonhosted.org/packages/d1/76/9835c8609c29f2214359e88f29255fc4aad4ea0f613fb48aa8815ceda1b6/ml_dtypes-0.4.1-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:2d55b588116a7085d6e074cf0cdb1d6fa3875c059dddc4d2c94a4cc81c23e975", size = 397973 }, - { url = "https://files.pythonhosted.org/packages/7e/99/e68c56fac5de973007a10254b6e17a0362393724f40f66d5e4033f4962c2/ml_dtypes-0.4.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e138a9b7a48079c900ea969341a5754019a1ad17ae27ee330f7ebf43f23877f9", size = 2185134 }, - { url = "https://files.pythonhosted.org/packages/28/bc/6a2344338ea7b61cd7b46fb24ec459360a5a0903b57c55b156c1e46c644a/ml_dtypes-0.4.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:74c6cfb5cf78535b103fde9ea3ded8e9f16f75bc07789054edc7776abfb3d752", size = 2163661 }, - { url = "https://files.pythonhosted.org/packages/e8/d3/ddfd9878b223b3aa9a930c6100a99afca5cfab7ea703662e00323acb7568/ml_dtypes-0.4.1-cp311-cp311-win_amd64.whl", hash = "sha256:274cc7193dd73b35fb26bef6c5d40ae3eb258359ee71cd82f6e96a8c948bdaa6", size = 126727 }, - { url = "https://files.pythonhosted.org/packages/ba/1a/99e924f12e4b62139fbac87419698c65f956d58de0dbfa7c028fa5b096aa/ml_dtypes-0.4.1-cp312-cp312-macosx_10_9_universal2.whl", hash = "sha256:827d3ca2097085cf0355f8fdf092b888890bb1b1455f52801a2d7756f056f54b", size = 405077 }, - { url = "https://files.pythonhosted.org/packages/8f/8c/7b610bd500617854c8cc6ed7c8cfb9d48d6a5c21a1437a36a4b9bc8a3598/ml_dtypes-0.4.1-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:772426b08a6172a891274d581ce58ea2789cc8abc1c002a27223f314aaf894e7", size = 2181554 }, - { url = "https://files.pythonhosted.org/packages/c7/c6/f89620cecc0581dc1839e218c4315171312e46c62a62da6ace204bda91c0/ml_dtypes-0.4.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:126e7d679b8676d1a958f2651949fbfa182832c3cd08020d8facd94e4114f3e9", size = 2160488 }, - { url = "https://files.pythonhosted.org/packages/ae/11/a742d3c31b2cc8557a48efdde53427fd5f9caa2fa3c9c27d826e78a66f51/ml_dtypes-0.4.1-cp312-cp312-win_amd64.whl", hash = "sha256:df0fb650d5c582a9e72bb5bd96cfebb2cdb889d89daff621c8fbc60295eba66c", size = 127462 }, -] - -[[package]] -name = "more-itertools" -version = "10.8.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/ea/5d/38b681d3fce7a266dd9ab73c66959406d565b3e85f21d5e66e1181d93721/more_itertools-10.8.0.tar.gz", hash = "sha256:f638ddf8a1a0d134181275fb5d58b086ead7c6a72429ad725c67503f13ba30bd", size = 137431 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/a4/8e/469e5a4a2f5855992e425f3cb33804cc07bf18d48f2db061aec61ce50270/more_itertools-10.8.0-py3-none-any.whl", hash = "sha256:52d4362373dcf7c52546bc4af9a86ee7c4579df9a8dc268be0a2f949d376cc9b", size = 69667 }, -] - -[[package]] -name = "mpmath" -version = "1.3.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/e0/47/dd32fa426cc72114383ac549964eecb20ecfd886d1e5ccf5340b55b02f57/mpmath-1.3.0.tar.gz", hash = "sha256:7a28eb2a9774d00c7bc92411c19a89209d5da7c4c9a9e227be8330a23a25b91f", size = 508106 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/43/e3/7d92a15f894aa0c9c4b49b8ee9ac9850d6e63b03c9c32c0367a13ae62209/mpmath-1.3.0-py3-none-any.whl", hash = "sha256:a0b2b9fe80bbcd81a6647ff13108738cfb482d481d826cc0e02f5b35e5c88d2c", size = 536198 }, -] - -[[package]] -name = "msgpack" -version = "1.1.2" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/4d/f2/bfb55a6236ed8725a96b0aa3acbd0ec17588e6a2c3b62a93eb513ed8783f/msgpack-1.1.2.tar.gz", hash = "sha256:3b60763c1373dd60f398488069bcdc703cd08a711477b5d480eecc9f9626f47e", size = 173581 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/f5/a2/3b68a9e769db68668b25c6108444a35f9bd163bb848c0650d516761a59c0/msgpack-1.1.2-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:0051fffef5a37ca2cd16978ae4f0aef92f164df86823871b5162812bebecd8e2", size = 81318 }, - { url = "https://files.pythonhosted.org/packages/5b/e1/2b720cc341325c00be44e1ed59e7cfeae2678329fbf5aa68f5bda57fe728/msgpack-1.1.2-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:a605409040f2da88676e9c9e5853b3449ba8011973616189ea5ee55ddbc5bc87", size = 83786 }, - { url = "https://files.pythonhosted.org/packages/71/e5/c2241de64bfceac456b140737812a2ab310b10538a7b34a1d393b748e095/msgpack-1.1.2-cp310-cp310-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:8b696e83c9f1532b4af884045ba7f3aa741a63b2bc22617293a2c6a7c645f251", size = 398240 }, - { url = "https://files.pythonhosted.org/packages/b7/09/2a06956383c0fdebaef5aa9246e2356776f12ea6f2a44bd1368abf0e46c4/msgpack-1.1.2-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:365c0bbe981a27d8932da71af63ef86acc59ed5c01ad929e09a0b88c6294e28a", size = 406070 }, - { url = "https://files.pythonhosted.org/packages/0e/74/2957703f0e1ef20637d6aead4fbb314330c26f39aa046b348c7edcf6ca6b/msgpack-1.1.2-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:41d1a5d875680166d3ac5c38573896453bbbea7092936d2e107214daf43b1d4f", size = 393403 }, - { url = "https://files.pythonhosted.org/packages/a5/09/3bfc12aa90f77b37322fc33e7a8a7c29ba7c8edeadfa27664451801b9860/msgpack-1.1.2-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:354e81bcdebaab427c3df4281187edc765d5d76bfb3a7c125af9da7a27e8458f", size = 398947 }, - { url = "https://files.pythonhosted.org/packages/4b/4f/05fcebd3b4977cb3d840f7ef6b77c51f8582086de5e642f3fefee35c86fc/msgpack-1.1.2-cp310-cp310-win32.whl", hash = "sha256:e64c8d2f5e5d5fda7b842f55dec6133260ea8f53c4257d64494c534f306bf7a9", size = 64769 }, - { url = "https://files.pythonhosted.org/packages/d0/3e/b4547e3a34210956382eed1c85935fff7e0f9b98be3106b3745d7dec9c5e/msgpack-1.1.2-cp310-cp310-win_amd64.whl", hash = "sha256:db6192777d943bdaaafb6ba66d44bf65aa0e9c5616fa1d2da9bb08828c6b39aa", size = 71293 }, - { url = "https://files.pythonhosted.org/packages/2c/97/560d11202bcd537abca693fd85d81cebe2107ba17301de42b01ac1677b69/msgpack-1.1.2-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:2e86a607e558d22985d856948c12a3fa7b42efad264dca8a3ebbcfa2735d786c", size = 82271 }, - { url = "https://files.pythonhosted.org/packages/83/04/28a41024ccbd67467380b6fb440ae916c1e4f25e2cd4c63abe6835ac566e/msgpack-1.1.2-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:283ae72fc89da59aa004ba147e8fc2f766647b1251500182fac0350d8af299c0", size = 84914 }, - { url = "https://files.pythonhosted.org/packages/71/46/b817349db6886d79e57a966346cf0902a426375aadc1e8e7a86a75e22f19/msgpack-1.1.2-cp311-cp311-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:61c8aa3bd513d87c72ed0b37b53dd5c5a0f58f2ff9f26e1555d3bd7948fb7296", size = 416962 }, - { url = "https://files.pythonhosted.org/packages/da/e0/6cc2e852837cd6086fe7d8406af4294e66827a60a4cf60b86575a4a65ca8/msgpack-1.1.2-cp311-cp311-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:454e29e186285d2ebe65be34629fa0e8605202c60fbc7c4c650ccd41870896ef", size = 426183 }, - { url = "https://files.pythonhosted.org/packages/25/98/6a19f030b3d2ea906696cedd1eb251708e50a5891d0978b012cb6107234c/msgpack-1.1.2-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:7bc8813f88417599564fafa59fd6f95be417179f76b40325b500b3c98409757c", size = 411454 }, - { url = "https://files.pythonhosted.org/packages/b7/cd/9098fcb6adb32187a70b7ecaabf6339da50553351558f37600e53a4a2a23/msgpack-1.1.2-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:bafca952dc13907bdfdedfc6a5f579bf4f292bdd506fadb38389afa3ac5b208e", size = 422341 }, - { url = "https://files.pythonhosted.org/packages/e6/ae/270cecbcf36c1dc85ec086b33a51a4d7d08fc4f404bdbc15b582255d05ff/msgpack-1.1.2-cp311-cp311-win32.whl", hash = "sha256:602b6740e95ffc55bfb078172d279de3773d7b7db1f703b2f1323566b878b90e", size = 64747 }, - { url = "https://files.pythonhosted.org/packages/2a/79/309d0e637f6f37e83c711f547308b91af02b72d2326ddd860b966080ef29/msgpack-1.1.2-cp311-cp311-win_amd64.whl", hash = "sha256:d198d275222dc54244bf3327eb8cbe00307d220241d9cec4d306d49a44e85f68", size = 71633 }, - { url = "https://files.pythonhosted.org/packages/73/4d/7c4e2b3d9b1106cd0aa6cb56cc57c6267f59fa8bfab7d91df5adc802c847/msgpack-1.1.2-cp311-cp311-win_arm64.whl", hash = "sha256:86f8136dfa5c116365a8a651a7d7484b65b13339731dd6faebb9a0242151c406", size = 64755 }, - { url = "https://files.pythonhosted.org/packages/ad/bd/8b0d01c756203fbab65d265859749860682ccd2a59594609aeec3a144efa/msgpack-1.1.2-cp312-cp312-macosx_10_13_x86_64.whl", hash = "sha256:70a0dff9d1f8da25179ffcf880e10cf1aad55fdb63cd59c9a49a1b82290062aa", size = 81939 }, - { url = "https://files.pythonhosted.org/packages/34/68/ba4f155f793a74c1483d4bdef136e1023f7bcba557f0db4ef3db3c665cf1/msgpack-1.1.2-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:446abdd8b94b55c800ac34b102dffd2f6aa0ce643c55dfc017ad89347db3dbdb", size = 85064 }, - { url = "https://files.pythonhosted.org/packages/f2/60/a064b0345fc36c4c3d2c743c82d9100c40388d77f0b48b2f04d6041dbec1/msgpack-1.1.2-cp312-cp312-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:c63eea553c69ab05b6747901b97d620bb2a690633c77f23feb0c6a947a8a7b8f", size = 417131 }, - { url = "https://files.pythonhosted.org/packages/65/92/a5100f7185a800a5d29f8d14041f61475b9de465ffcc0f3b9fba606e4505/msgpack-1.1.2-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:372839311ccf6bdaf39b00b61288e0557916c3729529b301c52c2d88842add42", size = 427556 }, - { url = "https://files.pythonhosted.org/packages/f5/87/ffe21d1bf7d9991354ad93949286f643b2bb6ddbeab66373922b44c3b8cc/msgpack-1.1.2-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:2929af52106ca73fcb28576218476ffbb531a036c2adbcf54a3664de124303e9", size = 404920 }, - { url = "https://files.pythonhosted.org/packages/ff/41/8543ed2b8604f7c0d89ce066f42007faac1eaa7d79a81555f206a5cdb889/msgpack-1.1.2-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:be52a8fc79e45b0364210eef5234a7cf8d330836d0a64dfbb878efa903d84620", size = 415013 }, - { url = "https://files.pythonhosted.org/packages/41/0d/2ddfaa8b7e1cee6c490d46cb0a39742b19e2481600a7a0e96537e9c22f43/msgpack-1.1.2-cp312-cp312-win32.whl", hash = "sha256:1fff3d825d7859ac888b0fbda39a42d59193543920eda9d9bea44d958a878029", size = 65096 }, - { url = "https://files.pythonhosted.org/packages/8c/ec/d431eb7941fb55a31dd6ca3404d41fbb52d99172df2e7707754488390910/msgpack-1.1.2-cp312-cp312-win_amd64.whl", hash = "sha256:1de460f0403172cff81169a30b9a92b260cb809c4cb7e2fc79ae8d0510c78b6b", size = 72708 }, - { url = "https://files.pythonhosted.org/packages/c5/31/5b1a1f70eb0e87d1678e9624908f86317787b536060641d6798e3cf70ace/msgpack-1.1.2-cp312-cp312-win_arm64.whl", hash = "sha256:be5980f3ee0e6bd44f3a9e9dea01054f175b50c3e6cdb692bc9424c0bbb8bf69", size = 64119 }, - { url = "https://files.pythonhosted.org/packages/6b/31/b46518ecc604d7edf3a4f94cb3bf021fc62aa301f0cb849936968164ef23/msgpack-1.1.2-cp313-cp313-macosx_10_13_x86_64.whl", hash = "sha256:4efd7b5979ccb539c221a4c4e16aac1a533efc97f3b759bb5a5ac9f6d10383bf", size = 81212 }, - { url = "https://files.pythonhosted.org/packages/92/dc/c385f38f2c2433333345a82926c6bfa5ecfff3ef787201614317b58dd8be/msgpack-1.1.2-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:42eefe2c3e2af97ed470eec850facbe1b5ad1d6eacdbadc42ec98e7dcf68b4b7", size = 84315 }, - { url = "https://files.pythonhosted.org/packages/d3/68/93180dce57f684a61a88a45ed13047558ded2be46f03acb8dec6d7c513af/msgpack-1.1.2-cp313-cp313-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:1fdf7d83102bf09e7ce3357de96c59b627395352a4024f6e2458501f158bf999", size = 412721 }, - { url = "https://files.pythonhosted.org/packages/5d/ba/459f18c16f2b3fc1a1ca871f72f07d70c07bf768ad0a507a698b8052ac58/msgpack-1.1.2-cp313-cp313-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:fac4be746328f90caa3cd4bc67e6fe36ca2bf61d5c6eb6d895b6527e3f05071e", size = 424657 }, - { url = "https://files.pythonhosted.org/packages/38/f8/4398c46863b093252fe67368b44edc6c13b17f4e6b0e4929dbf0bdb13f23/msgpack-1.1.2-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:fffee09044073e69f2bad787071aeec727183e7580443dfeb8556cbf1978d162", size = 402668 }, - { url = "https://files.pythonhosted.org/packages/28/ce/698c1eff75626e4124b4d78e21cca0b4cc90043afb80a507626ea354ab52/msgpack-1.1.2-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:5928604de9b032bc17f5099496417f113c45bc6bc21b5c6920caf34b3c428794", size = 419040 }, - { url = "https://files.pythonhosted.org/packages/67/32/f3cd1667028424fa7001d82e10ee35386eea1408b93d399b09fb0aa7875f/msgpack-1.1.2-cp313-cp313-win32.whl", hash = "sha256:a7787d353595c7c7e145e2331abf8b7ff1e6673a6b974ded96e6d4ec09f00c8c", size = 65037 }, - { url = "https://files.pythonhosted.org/packages/74/07/1ed8277f8653c40ebc65985180b007879f6a836c525b3885dcc6448ae6cb/msgpack-1.1.2-cp313-cp313-win_amd64.whl", hash = "sha256:a465f0dceb8e13a487e54c07d04ae3ba131c7c5b95e2612596eafde1dccf64a9", size = 72631 }, - { url = "https://files.pythonhosted.org/packages/e5/db/0314e4e2db56ebcf450f277904ffd84a7988b9e5da8d0d61ab2d057df2b6/msgpack-1.1.2-cp313-cp313-win_arm64.whl", hash = "sha256:e69b39f8c0aa5ec24b57737ebee40be647035158f14ed4b40e6f150077e21a84", size = 64118 }, - { url = "https://files.pythonhosted.org/packages/22/71/201105712d0a2ff07b7873ed3c220292fb2ea5120603c00c4b634bcdafb3/msgpack-1.1.2-cp314-cp314-macosx_10_13_x86_64.whl", hash = "sha256:e23ce8d5f7aa6ea6d2a2b326b4ba46c985dbb204523759984430db7114f8aa00", size = 81127 }, - { url = "https://files.pythonhosted.org/packages/1b/9f/38ff9e57a2eade7bf9dfee5eae17f39fc0e998658050279cbb14d97d36d9/msgpack-1.1.2-cp314-cp314-macosx_11_0_arm64.whl", hash = "sha256:6c15b7d74c939ebe620dd8e559384be806204d73b4f9356320632d783d1f7939", size = 84981 }, - { url = "https://files.pythonhosted.org/packages/8e/a9/3536e385167b88c2cc8f4424c49e28d49a6fc35206d4a8060f136e71f94c/msgpack-1.1.2-cp314-cp314-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:99e2cb7b9031568a2a5c73aa077180f93dd2e95b4f8d3b8e14a73ae94a9e667e", size = 411885 }, - { url = "https://files.pythonhosted.org/packages/2f/40/dc34d1a8d5f1e51fc64640b62b191684da52ca469da9cd74e84936ffa4a6/msgpack-1.1.2-cp314-cp314-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:180759d89a057eab503cf62eeec0aa61c4ea1200dee709f3a8e9397dbb3b6931", size = 419658 }, - { url = "https://files.pythonhosted.org/packages/3b/ef/2b92e286366500a09a67e03496ee8b8ba00562797a52f3c117aa2b29514b/msgpack-1.1.2-cp314-cp314-musllinux_1_2_aarch64.whl", hash = "sha256:04fb995247a6e83830b62f0b07bf36540c213f6eac8e851166d8d86d83cbd014", size = 403290 }, - { url = "https://files.pythonhosted.org/packages/78/90/e0ea7990abea5764e4655b8177aa7c63cdfa89945b6e7641055800f6c16b/msgpack-1.1.2-cp314-cp314-musllinux_1_2_x86_64.whl", hash = "sha256:8e22ab046fa7ede9e36eeb4cfad44d46450f37bb05d5ec482b02868f451c95e2", size = 415234 }, - { url = "https://files.pythonhosted.org/packages/72/4e/9390aed5db983a2310818cd7d3ec0aecad45e1f7007e0cda79c79507bb0d/msgpack-1.1.2-cp314-cp314-win32.whl", hash = "sha256:80a0ff7d4abf5fecb995fcf235d4064b9a9a8a40a3ab80999e6ac1e30b702717", size = 66391 }, - { url = "https://files.pythonhosted.org/packages/6e/f1/abd09c2ae91228c5f3998dbd7f41353def9eac64253de3c8105efa2082f7/msgpack-1.1.2-cp314-cp314-win_amd64.whl", hash = "sha256:9ade919fac6a3e7260b7f64cea89df6bec59104987cbea34d34a2fa15d74310b", size = 73787 }, - { url = "https://files.pythonhosted.org/packages/6a/b0/9d9f667ab48b16ad4115c1935d94023b82b3198064cb84a123e97f7466c1/msgpack-1.1.2-cp314-cp314-win_arm64.whl", hash = "sha256:59415c6076b1e30e563eb732e23b994a61c159cec44deaf584e5cc1dd662f2af", size = 66453 }, - { url = "https://files.pythonhosted.org/packages/16/67/93f80545eb1792b61a217fa7f06d5e5cb9e0055bed867f43e2b8e012e137/msgpack-1.1.2-cp314-cp314t-macosx_10_13_x86_64.whl", hash = "sha256:897c478140877e5307760b0ea66e0932738879e7aa68144d9b78ea4c8302a84a", size = 85264 }, - { url = "https://files.pythonhosted.org/packages/87/1c/33c8a24959cf193966ef11a6f6a2995a65eb066bd681fd085afd519a57ce/msgpack-1.1.2-cp314-cp314t-macosx_11_0_arm64.whl", hash = "sha256:a668204fa43e6d02f89dbe79a30b0d67238d9ec4c5bd8a940fc3a004a47b721b", size = 89076 }, - { url = "https://files.pythonhosted.org/packages/fc/6b/62e85ff7193663fbea5c0254ef32f0c77134b4059f8da89b958beb7696f3/msgpack-1.1.2-cp314-cp314t-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:5559d03930d3aa0f3aacb4c42c776af1a2ace2611871c84a75afe436695e6245", size = 435242 }, - { url = "https://files.pythonhosted.org/packages/c1/47/5c74ecb4cc277cf09f64e913947871682ffa82b3b93c8dad68083112f412/msgpack-1.1.2-cp314-cp314t-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:70c5a7a9fea7f036b716191c29047374c10721c389c21e9ffafad04df8c52c90", size = 432509 }, - { url = "https://files.pythonhosted.org/packages/24/a4/e98ccdb56dc4e98c929a3f150de1799831c0a800583cde9fa022fa90602d/msgpack-1.1.2-cp314-cp314t-musllinux_1_2_aarch64.whl", hash = "sha256:f2cb069d8b981abc72b41aea1c580ce92d57c673ec61af4c500153a626cb9e20", size = 415957 }, - { url = "https://files.pythonhosted.org/packages/da/28/6951f7fb67bc0a4e184a6b38ab71a92d9ba58080b27a77d3e2fb0be5998f/msgpack-1.1.2-cp314-cp314t-musllinux_1_2_x86_64.whl", hash = "sha256:d62ce1f483f355f61adb5433ebfd8868c5f078d1a52d042b0a998682b4fa8c27", size = 422910 }, - { url = "https://files.pythonhosted.org/packages/f0/03/42106dcded51f0a0b5284d3ce30a671e7bd3f7318d122b2ead66ad289fed/msgpack-1.1.2-cp314-cp314t-win32.whl", hash = "sha256:1d1418482b1ee984625d88aa9585db570180c286d942da463533b238b98b812b", size = 75197 }, - { url = "https://files.pythonhosted.org/packages/15/86/d0071e94987f8db59d4eeb386ddc64d0bb9b10820a8d82bcd3e53eeb2da6/msgpack-1.1.2-cp314-cp314t-win_amd64.whl", hash = "sha256:5a46bf7e831d09470ad92dff02b8b1ac92175ca36b087f904a0519857c6be3ff", size = 85772 }, - { url = "https://files.pythonhosted.org/packages/81/f2/08ace4142eb281c12701fc3b93a10795e4d4dc7f753911d836675050f886/msgpack-1.1.2-cp314-cp314t-win_arm64.whl", hash = "sha256:d99ef64f349d5ec3293688e91486c5fdb925ed03807f64d98d205d2713c60b46", size = 70868 }, -] - -[[package]] -name = "multidict" -version = "6.7.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "typing-extensions", marker = "python_full_version < '3.11'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/80/1e/5492c365f222f907de1039b91f922b93fa4f764c713ee858d235495d8f50/multidict-6.7.0.tar.gz", hash = "sha256:c6e99d9a65ca282e578dfea819cfa9c0a62b2499d8677392e09feaf305e9e6f5", size = 101834 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/a9/63/7bdd4adc330abcca54c85728db2327130e49e52e8c3ce685cec44e0f2e9f/multidict-6.7.0-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:9f474ad5acda359c8758c8accc22032c6abe6dc87a8be2440d097785e27a9349", size = 77153 }, - { url = "https://files.pythonhosted.org/packages/3f/bb/b6c35ff175ed1a3142222b78455ee31be71a8396ed3ab5280fbe3ebe4e85/multidict-6.7.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:4b7a9db5a870f780220e931d0002bbfd88fb53aceb6293251e2c839415c1b20e", size = 44993 }, - { url = "https://files.pythonhosted.org/packages/e0/1f/064c77877c5fa6df6d346e68075c0f6998547afe952d6471b4c5f6a7345d/multidict-6.7.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:03ca744319864e92721195fa28c7a3b2bc7b686246b35e4078c1e4d0eb5466d3", size = 44607 }, - { url = "https://files.pythonhosted.org/packages/04/7a/bf6aa92065dd47f287690000b3d7d332edfccb2277634cadf6a810463c6a/multidict-6.7.0-cp310-cp310-manylinux1_i686.manylinux_2_28_i686.manylinux_2_5_i686.whl", hash = "sha256:f0e77e3c0008bc9316e662624535b88d360c3a5d3f81e15cf12c139a75250046", size = 241847 }, - { url = "https://files.pythonhosted.org/packages/94/39/297a8de920f76eda343e4ce05f3b489f0ab3f9504f2576dfb37b7c08ca08/multidict-6.7.0-cp310-cp310-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:08325c9e5367aa379a3496aa9a022fe8837ff22e00b94db256d3a1378c76ab32", size = 242616 }, - { url = "https://files.pythonhosted.org/packages/39/3a/d0eee2898cfd9d654aea6cb8c4addc2f9756e9a7e09391cfe55541f917f7/multidict-6.7.0-cp310-cp310-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:e2862408c99f84aa571ab462d25236ef9cb12a602ea959ba9c9009a54902fc73", size = 222333 }, - { url = "https://files.pythonhosted.org/packages/05/48/3b328851193c7a4240815b71eea165b49248867bbb6153a0aee227a0bb47/multidict-6.7.0-cp310-cp310-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:4d72a9a2d885f5c208b0cb91ff2ed43636bb7e345ec839ff64708e04f69a13cc", size = 253239 }, - { url = "https://files.pythonhosted.org/packages/b1/ca/0706a98c8d126a89245413225ca4a3fefc8435014de309cf8b30acb68841/multidict-6.7.0-cp310-cp310-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:478cc36476687bac1514d651cbbaa94b86b0732fb6855c60c673794c7dd2da62", size = 251618 }, - { url = "https://files.pythonhosted.org/packages/5e/4f/9c7992f245554d8b173f6f0a048ad24b3e645d883f096857ec2c0822b8bd/multidict-6.7.0-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:6843b28b0364dc605f21481c90fadb5f60d9123b442eb8a726bb74feef588a84", size = 241655 }, - { url = "https://files.pythonhosted.org/packages/31/79/26a85991ae67efd1c0b1fc2e0c275b8a6aceeb155a68861f63f87a798f16/multidict-6.7.0-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:23bfeee5316266e5ee2d625df2d2c602b829435fc3a235c2ba2131495706e4a0", size = 239245 }, - { url = "https://files.pythonhosted.org/packages/14/1e/75fa96394478930b79d0302eaf9a6c69f34005a1a5251ac8b9c336486ec9/multidict-6.7.0-cp310-cp310-musllinux_1_2_armv7l.whl", hash = "sha256:680878b9f3d45c31e1f730eef731f9b0bc1da456155688c6745ee84eb818e90e", size = 233523 }, - { url = "https://files.pythonhosted.org/packages/b2/5e/085544cb9f9c4ad2b5d97467c15f856df8d9bac410cffd5c43991a5d878b/multidict-6.7.0-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:eb866162ef2f45063acc7a53a88ef6fe8bf121d45c30ea3c9cd87ce7e191a8d4", size = 243129 }, - { url = "https://files.pythonhosted.org/packages/b9/c3/e9d9e2f20c9474e7a8fcef28f863c5cbd29bb5adce6b70cebe8bdad0039d/multidict-6.7.0-cp310-cp310-musllinux_1_2_ppc64le.whl", hash = "sha256:df0e3bf7993bdbeca5ac25aa859cf40d39019e015c9c91809ba7093967f7a648", size = 248999 }, - { url = "https://files.pythonhosted.org/packages/b5/3f/df171b6efa3239ae33b97b887e42671cd1d94d460614bfb2c30ffdab3b95/multidict-6.7.0-cp310-cp310-musllinux_1_2_s390x.whl", hash = "sha256:661709cdcd919a2ece2234f9bae7174e5220c80b034585d7d8a755632d3e2111", size = 243711 }, - { url = "https://files.pythonhosted.org/packages/3c/2f/9b5564888c4e14b9af64c54acf149263721a283aaf4aa0ae89b091d5d8c1/multidict-6.7.0-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:096f52730c3fb8ed419db2d44391932b63891b2c5ed14850a7e215c0ba9ade36", size = 237504 }, - { url = "https://files.pythonhosted.org/packages/6c/3a/0bd6ca0f7d96d790542d591c8c3354c1e1b6bfd2024d4d92dc3d87485ec7/multidict-6.7.0-cp310-cp310-win32.whl", hash = "sha256:afa8a2978ec65d2336305550535c9c4ff50ee527914328c8677b3973ade52b85", size = 41422 }, - { url = "https://files.pythonhosted.org/packages/00/35/f6a637ea2c75f0d3b7c7d41b1189189acff0d9deeb8b8f35536bb30f5e33/multidict-6.7.0-cp310-cp310-win_amd64.whl", hash = "sha256:b15b3afff74f707b9275d5ba6a91ae8f6429c3ffb29bbfd216b0b375a56f13d7", size = 46050 }, - { url = "https://files.pythonhosted.org/packages/e7/b8/f7bf8329b39893d02d9d95cf610c75885d12fc0f402b1c894e1c8e01c916/multidict-6.7.0-cp310-cp310-win_arm64.whl", hash = "sha256:4b73189894398d59131a66ff157837b1fafea9974be486d036bb3d32331fdbf0", size = 43153 }, - { url = "https://files.pythonhosted.org/packages/34/9e/5c727587644d67b2ed479041e4b1c58e30afc011e3d45d25bbe35781217c/multidict-6.7.0-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:4d409aa42a94c0b3fa617708ef5276dfe81012ba6753a0370fcc9d0195d0a1fc", size = 76604 }, - { url = "https://files.pythonhosted.org/packages/17/e4/67b5c27bd17c085a5ea8f1ec05b8a3e5cba0ca734bfcad5560fb129e70ca/multidict-6.7.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:14c9e076eede3b54c636f8ce1c9c252b5f057c62131211f0ceeec273810c9721", size = 44715 }, - { url = "https://files.pythonhosted.org/packages/4d/e1/866a5d77be6ea435711bef2a4291eed11032679b6b28b56b4776ab06ba3e/multidict-6.7.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:4c09703000a9d0fa3c3404b27041e574cc7f4df4c6563873246d0e11812a94b6", size = 44332 }, - { url = "https://files.pythonhosted.org/packages/31/61/0c2d50241ada71ff61a79518db85ada85fdabfcf395d5968dae1cbda04e5/multidict-6.7.0-cp311-cp311-manylinux1_i686.manylinux_2_28_i686.manylinux_2_5_i686.whl", hash = "sha256:a265acbb7bb33a3a2d626afbe756371dce0279e7b17f4f4eda406459c2b5ff1c", size = 245212 }, - { url = "https://files.pythonhosted.org/packages/ac/e0/919666a4e4b57fff1b57f279be1c9316e6cdc5de8a8b525d76f6598fefc7/multidict-6.7.0-cp311-cp311-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:51cb455de290ae462593e5b1cb1118c5c22ea7f0d3620d9940bf695cea5a4bd7", size = 246671 }, - { url = "https://files.pythonhosted.org/packages/a1/cc/d027d9c5a520f3321b65adea289b965e7bcbd2c34402663f482648c716ce/multidict-6.7.0-cp311-cp311-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:db99677b4457c7a5c5a949353e125ba72d62b35f74e26da141530fbb012218a7", size = 225491 }, - { url = "https://files.pythonhosted.org/packages/75/c4/bbd633980ce6155a28ff04e6a6492dd3335858394d7bb752d8b108708558/multidict-6.7.0-cp311-cp311-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:f470f68adc395e0183b92a2f4689264d1ea4b40504a24d9882c27375e6662bb9", size = 257322 }, - { url = "https://files.pythonhosted.org/packages/4c/6d/d622322d344f1f053eae47e033b0b3f965af01212de21b10bcf91be991fb/multidict-6.7.0-cp311-cp311-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:0db4956f82723cc1c270de9c6e799b4c341d327762ec78ef82bb962f79cc07d8", size = 254694 }, - { url = "https://files.pythonhosted.org/packages/a8/9f/78f8761c2705d4c6d7516faed63c0ebdac569f6db1bef95e0d5218fdc146/multidict-6.7.0-cp311-cp311-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:3e56d780c238f9e1ae66a22d2adf8d16f485381878250db8d496623cd38b22bd", size = 246715 }, - { url = "https://files.pythonhosted.org/packages/78/59/950818e04f91b9c2b95aab3d923d9eabd01689d0dcd889563988e9ea0fd8/multidict-6.7.0-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:9d14baca2ee12c1a64740d4531356ba50b82543017f3ad6de0deb943c5979abb", size = 243189 }, - { url = "https://files.pythonhosted.org/packages/7a/3d/77c79e1934cad2ee74991840f8a0110966d9599b3af95964c0cd79bb905b/multidict-6.7.0-cp311-cp311-musllinux_1_2_armv7l.whl", hash = "sha256:295a92a76188917c7f99cda95858c822f9e4aae5824246bba9b6b44004ddd0a6", size = 237845 }, - { url = "https://files.pythonhosted.org/packages/63/1b/834ce32a0a97a3b70f86437f685f880136677ac00d8bce0027e9fd9c2db7/multidict-6.7.0-cp311-cp311-musllinux_1_2_i686.whl", hash = "sha256:39f1719f57adbb767ef592a50ae5ebb794220d1188f9ca93de471336401c34d2", size = 246374 }, - { url = "https://files.pythonhosted.org/packages/23/ef/43d1c3ba205b5dec93dc97f3fba179dfa47910fc73aaaea4f7ceb41cec2a/multidict-6.7.0-cp311-cp311-musllinux_1_2_ppc64le.whl", hash = "sha256:0a13fb8e748dfc94749f622de065dd5c1def7e0d2216dba72b1d8069a389c6ff", size = 253345 }, - { url = "https://files.pythonhosted.org/packages/6b/03/eaf95bcc2d19ead522001f6a650ef32811aa9e3624ff0ad37c445c7a588c/multidict-6.7.0-cp311-cp311-musllinux_1_2_s390x.whl", hash = "sha256:e3aa16de190d29a0ea1b48253c57d99a68492c8dd8948638073ab9e74dc9410b", size = 246940 }, - { url = "https://files.pythonhosted.org/packages/e8/df/ec8a5fd66ea6cd6f525b1fcbb23511b033c3e9bc42b81384834ffa484a62/multidict-6.7.0-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:a048ce45dcdaaf1defb76b2e684f997fb5abf74437b6cb7b22ddad934a964e34", size = 242229 }, - { url = "https://files.pythonhosted.org/packages/8a/a2/59b405d59fd39ec86d1142630e9049243015a5f5291ba49cadf3c090c541/multidict-6.7.0-cp311-cp311-win32.whl", hash = "sha256:a90af66facec4cebe4181b9e62a68be65e45ac9b52b67de9eec118701856e7ff", size = 41308 }, - { url = "https://files.pythonhosted.org/packages/32/0f/13228f26f8b882c34da36efa776c3b7348455ec383bab4a66390e42963ae/multidict-6.7.0-cp311-cp311-win_amd64.whl", hash = "sha256:95b5ffa4349df2887518bb839409bcf22caa72d82beec453216802f475b23c81", size = 46037 }, - { url = "https://files.pythonhosted.org/packages/84/1f/68588e31b000535a3207fd3c909ebeec4fb36b52c442107499c18a896a2a/multidict-6.7.0-cp311-cp311-win_arm64.whl", hash = "sha256:329aa225b085b6f004a4955271a7ba9f1087e39dcb7e65f6284a988264a63912", size = 43023 }, - { url = "https://files.pythonhosted.org/packages/c2/9e/9f61ac18d9c8b475889f32ccfa91c9f59363480613fc807b6e3023d6f60b/multidict-6.7.0-cp312-cp312-macosx_10_13_universal2.whl", hash = "sha256:8a3862568a36d26e650a19bb5cbbba14b71789032aebc0423f8cc5f150730184", size = 76877 }, - { url = "https://files.pythonhosted.org/packages/38/6f/614f09a04e6184f8824268fce4bc925e9849edfa654ddd59f0b64508c595/multidict-6.7.0-cp312-cp312-macosx_10_13_x86_64.whl", hash = "sha256:960c60b5849b9b4f9dcc9bea6e3626143c252c74113df2c1540aebce70209b45", size = 45467 }, - { url = "https://files.pythonhosted.org/packages/b3/93/c4f67a436dd026f2e780c433277fff72be79152894d9fc36f44569cab1a6/multidict-6.7.0-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:2049be98fb57a31b4ccf870bf377af2504d4ae35646a19037ec271e4c07998aa", size = 43834 }, - { url = "https://files.pythonhosted.org/packages/7f/f5/013798161ca665e4a422afbc5e2d9e4070142a9ff8905e482139cd09e4d0/multidict-6.7.0-cp312-cp312-manylinux1_i686.manylinux_2_28_i686.manylinux_2_5_i686.whl", hash = "sha256:0934f3843a1860dd465d38895c17fce1f1cb37295149ab05cd1b9a03afacb2a7", size = 250545 }, - { url = "https://files.pythonhosted.org/packages/71/2f/91dbac13e0ba94669ea5119ba267c9a832f0cb65419aca75549fcf09a3dc/multidict-6.7.0-cp312-cp312-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:b3e34f3a1b8131ba06f1a73adab24f30934d148afcd5f5de9a73565a4404384e", size = 258305 }, - { url = "https://files.pythonhosted.org/packages/ef/b0/754038b26f6e04488b48ac621f779c341338d78503fb45403755af2df477/multidict-6.7.0-cp312-cp312-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:efbb54e98446892590dc2458c19c10344ee9a883a79b5cec4bc34d6656e8d546", size = 242363 }, - { url = "https://files.pythonhosted.org/packages/87/15/9da40b9336a7c9fa606c4cf2ed80a649dffeb42b905d4f63a1d7eb17d746/multidict-6.7.0-cp312-cp312-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:a35c5fc61d4f51eb045061e7967cfe3123d622cd500e8868e7c0c592a09fedc4", size = 268375 }, - { url = "https://files.pythonhosted.org/packages/82/72/c53fcade0cc94dfaad583105fd92b3a783af2091eddcb41a6d5a52474000/multidict-6.7.0-cp312-cp312-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:29fe6740ebccba4175af1b9b87bf553e9c15cd5868ee967e010efcf94e4fd0f1", size = 269346 }, - { url = "https://files.pythonhosted.org/packages/0d/e2/9baffdae21a76f77ef8447f1a05a96ec4bc0a24dae08767abc0a2fe680b8/multidict-6.7.0-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:123e2a72e20537add2f33a79e605f6191fba2afda4cbb876e35c1a7074298a7d", size = 256107 }, - { url = "https://files.pythonhosted.org/packages/3c/06/3f06f611087dc60d65ef775f1fb5aca7c6d61c6db4990e7cda0cef9b1651/multidict-6.7.0-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:b284e319754366c1aee2267a2036248b24eeb17ecd5dc16022095e747f2f4304", size = 253592 }, - { url = "https://files.pythonhosted.org/packages/20/24/54e804ec7945b6023b340c412ce9c3f81e91b3bf5fa5ce65558740141bee/multidict-6.7.0-cp312-cp312-musllinux_1_2_armv7l.whl", hash = "sha256:803d685de7be4303b5a657b76e2f6d1240e7e0a8aa2968ad5811fa2285553a12", size = 251024 }, - { url = "https://files.pythonhosted.org/packages/14/48/011cba467ea0b17ceb938315d219391d3e421dfd35928e5dbdc3f4ae76ef/multidict-6.7.0-cp312-cp312-musllinux_1_2_i686.whl", hash = "sha256:c04a328260dfd5db8c39538f999f02779012268f54614902d0afc775d44e0a62", size = 251484 }, - { url = "https://files.pythonhosted.org/packages/0d/2f/919258b43bb35b99fa127435cfb2d91798eb3a943396631ef43e3720dcf4/multidict-6.7.0-cp312-cp312-musllinux_1_2_ppc64le.whl", hash = "sha256:8a19cdb57cd3df4cd865849d93ee14920fb97224300c88501f16ecfa2604b4e0", size = 263579 }, - { url = "https://files.pythonhosted.org/packages/31/22/a0e884d86b5242b5a74cf08e876bdf299e413016b66e55511f7a804a366e/multidict-6.7.0-cp312-cp312-musllinux_1_2_s390x.whl", hash = "sha256:9b2fd74c52accced7e75de26023b7dccee62511a600e62311b918ec5c168fc2a", size = 259654 }, - { url = "https://files.pythonhosted.org/packages/b2/e5/17e10e1b5c5f5a40f2fcbb45953c9b215f8a4098003915e46a93f5fcaa8f/multidict-6.7.0-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:3e8bfdd0e487acf992407a140d2589fe598238eaeffa3da8448d63a63cd363f8", size = 251511 }, - { url = "https://files.pythonhosted.org/packages/e3/9a/201bb1e17e7af53139597069c375e7b0dcbd47594604f65c2d5359508566/multidict-6.7.0-cp312-cp312-win32.whl", hash = "sha256:dd32a49400a2c3d52088e120ee00c1e3576cbff7e10b98467962c74fdb762ed4", size = 41895 }, - { url = "https://files.pythonhosted.org/packages/46/e2/348cd32faad84eaf1d20cce80e2bb0ef8d312c55bca1f7fa9865e7770aaf/multidict-6.7.0-cp312-cp312-win_amd64.whl", hash = "sha256:92abb658ef2d7ef22ac9f8bb88e8b6c3e571671534e029359b6d9e845923eb1b", size = 46073 }, - { url = "https://files.pythonhosted.org/packages/25/ec/aad2613c1910dce907480e0c3aa306905830f25df2e54ccc9dea450cb5aa/multidict-6.7.0-cp312-cp312-win_arm64.whl", hash = "sha256:490dab541a6a642ce1a9d61a4781656b346a55c13038f0b1244653828e3a83ec", size = 43226 }, - { url = "https://files.pythonhosted.org/packages/d2/86/33272a544eeb36d66e4d9a920602d1a2f57d4ebea4ef3cdfe5a912574c95/multidict-6.7.0-cp313-cp313-macosx_10_13_universal2.whl", hash = "sha256:bee7c0588aa0076ce77c0ea5d19a68d76ad81fcd9fe8501003b9a24f9d4000f6", size = 76135 }, - { url = "https://files.pythonhosted.org/packages/91/1c/eb97db117a1ebe46d457a3d235a7b9d2e6dcab174f42d1b67663dd9e5371/multidict-6.7.0-cp313-cp313-macosx_10_13_x86_64.whl", hash = "sha256:7ef6b61cad77091056ce0e7ce69814ef72afacb150b7ac6a3e9470def2198159", size = 45117 }, - { url = "https://files.pythonhosted.org/packages/f1/d8/6c3442322e41fb1dd4de8bd67bfd11cd72352ac131f6368315617de752f1/multidict-6.7.0-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:9c0359b1ec12b1d6849c59f9d319610b7f20ef990a6d454ab151aa0e3b9f78ca", size = 43472 }, - { url = "https://files.pythonhosted.org/packages/75/3f/e2639e80325af0b6c6febdf8e57cc07043ff15f57fa1ef808f4ccb5ac4cd/multidict-6.7.0-cp313-cp313-manylinux1_i686.manylinux_2_28_i686.manylinux_2_5_i686.whl", hash = "sha256:cd240939f71c64bd658f186330603aac1a9a81bf6273f523fca63673cb7378a8", size = 249342 }, - { url = "https://files.pythonhosted.org/packages/5d/cc/84e0585f805cbeaa9cbdaa95f9a3d6aed745b9d25700623ac89a6ecff400/multidict-6.7.0-cp313-cp313-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:a60a4d75718a5efa473ebd5ab685786ba0c67b8381f781d1be14da49f1a2dc60", size = 257082 }, - { url = "https://files.pythonhosted.org/packages/b0/9c/ac851c107c92289acbbf5cfb485694084690c1b17e555f44952c26ddc5bd/multidict-6.7.0-cp313-cp313-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:53a42d364f323275126aff81fb67c5ca1b7a04fda0546245730a55c8c5f24bc4", size = 240704 }, - { url = "https://files.pythonhosted.org/packages/50/cc/5f93e99427248c09da95b62d64b25748a5f5c98c7c2ab09825a1d6af0e15/multidict-6.7.0-cp313-cp313-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:3b29b980d0ddbecb736735ee5bef69bb2ddca56eff603c86f3f29a1128299b4f", size = 266355 }, - { url = "https://files.pythonhosted.org/packages/ec/0c/2ec1d883ceb79c6f7f6d7ad90c919c898f5d1c6ea96d322751420211e072/multidict-6.7.0-cp313-cp313-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:f8a93b1c0ed2d04b97a5e9336fd2d33371b9a6e29ab7dd6503d63407c20ffbaf", size = 267259 }, - { url = "https://files.pythonhosted.org/packages/c6/2d/f0b184fa88d6630aa267680bdb8623fb69cb0d024b8c6f0d23f9a0f406d3/multidict-6.7.0-cp313-cp313-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:9ff96e8815eecacc6645da76c413eb3b3d34cfca256c70b16b286a687d013c32", size = 254903 }, - { url = "https://files.pythonhosted.org/packages/06/c9/11ea263ad0df7dfabcad404feb3c0dd40b131bc7f232d5537f2fb1356951/multidict-6.7.0-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:7516c579652f6a6be0e266aec0acd0db80829ca305c3d771ed898538804c2036", size = 252365 }, - { url = "https://files.pythonhosted.org/packages/41/88/d714b86ee2c17d6e09850c70c9d310abac3d808ab49dfa16b43aba9d53fd/multidict-6.7.0-cp313-cp313-musllinux_1_2_armv7l.whl", hash = "sha256:040f393368e63fb0f3330e70c26bfd336656bed925e5cbe17c9da839a6ab13ec", size = 250062 }, - { url = "https://files.pythonhosted.org/packages/15/fe/ad407bb9e818c2b31383f6131ca19ea7e35ce93cf1310fce69f12e89de75/multidict-6.7.0-cp313-cp313-musllinux_1_2_i686.whl", hash = "sha256:b3bc26a951007b1057a1c543af845f1c7e3e71cc240ed1ace7bf4484aa99196e", size = 249683 }, - { url = "https://files.pythonhosted.org/packages/8c/a4/a89abdb0229e533fb925e7c6e5c40201c2873efebc9abaf14046a4536ee6/multidict-6.7.0-cp313-cp313-musllinux_1_2_ppc64le.whl", hash = "sha256:7b022717c748dd1992a83e219587aabe45980d88969f01b316e78683e6285f64", size = 261254 }, - { url = "https://files.pythonhosted.org/packages/8d/aa/0e2b27bd88b40a4fb8dc53dd74eecac70edaa4c1dd0707eb2164da3675b3/multidict-6.7.0-cp313-cp313-musllinux_1_2_s390x.whl", hash = "sha256:9600082733859f00d79dee64effc7aef1beb26adb297416a4ad2116fd61374bd", size = 257967 }, - { url = "https://files.pythonhosted.org/packages/d0/8e/0c67b7120d5d5f6d874ed85a085f9dc770a7f9d8813e80f44a9fec820bb7/multidict-6.7.0-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:94218fcec4d72bc61df51c198d098ce2b378e0ccbac41ddbed5ef44092913288", size = 250085 }, - { url = "https://files.pythonhosted.org/packages/ba/55/b73e1d624ea4b8fd4dd07a3bb70f6e4c7c6c5d9d640a41c6ffe5cdbd2a55/multidict-6.7.0-cp313-cp313-win32.whl", hash = "sha256:a37bd74c3fa9d00be2d7b8eca074dc56bd8077ddd2917a839bd989612671ed17", size = 41713 }, - { url = "https://files.pythonhosted.org/packages/32/31/75c59e7d3b4205075b4c183fa4ca398a2daf2303ddf616b04ae6ef55cffe/multidict-6.7.0-cp313-cp313-win_amd64.whl", hash = "sha256:30d193c6cc6d559db42b6bcec8a5d395d34d60c9877a0b71ecd7c204fcf15390", size = 45915 }, - { url = "https://files.pythonhosted.org/packages/31/2a/8987831e811f1184c22bc2e45844934385363ee61c0a2dcfa8f71b87e608/multidict-6.7.0-cp313-cp313-win_arm64.whl", hash = "sha256:ea3334cabe4d41b7ccd01e4d349828678794edbc2d3ae97fc162a3312095092e", size = 43077 }, - { url = "https://files.pythonhosted.org/packages/e8/68/7b3a5170a382a340147337b300b9eb25a9ddb573bcdfff19c0fa3f31ffba/multidict-6.7.0-cp313-cp313t-macosx_10_13_universal2.whl", hash = "sha256:ad9ce259f50abd98a1ca0aa6e490b58c316a0fce0617f609723e40804add2c00", size = 83114 }, - { url = "https://files.pythonhosted.org/packages/55/5c/3fa2d07c84df4e302060f555bbf539310980362236ad49f50eeb0a1c1eb9/multidict-6.7.0-cp313-cp313t-macosx_10_13_x86_64.whl", hash = "sha256:07f5594ac6d084cbb5de2df218d78baf55ef150b91f0ff8a21cc7a2e3a5a58eb", size = 48442 }, - { url = "https://files.pythonhosted.org/packages/fc/56/67212d33239797f9bd91962bb899d72bb0f4c35a8652dcdb8ed049bef878/multidict-6.7.0-cp313-cp313t-macosx_11_0_arm64.whl", hash = "sha256:0591b48acf279821a579282444814a2d8d0af624ae0bc600aa4d1b920b6e924b", size = 46885 }, - { url = "https://files.pythonhosted.org/packages/46/d1/908f896224290350721597a61a69cd19b89ad8ee0ae1f38b3f5cd12ea2ac/multidict-6.7.0-cp313-cp313t-manylinux1_i686.manylinux_2_28_i686.manylinux_2_5_i686.whl", hash = "sha256:749a72584761531d2b9467cfbdfd29487ee21124c304c4b6cb760d8777b27f9c", size = 242588 }, - { url = "https://files.pythonhosted.org/packages/ab/67/8604288bbd68680eee0ab568fdcb56171d8b23a01bcd5cb0c8fedf6e5d99/multidict-6.7.0-cp313-cp313t-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:6b4c3d199f953acd5b446bf7c0de1fe25d94e09e79086f8dc2f48a11a129cdf1", size = 249966 }, - { url = "https://files.pythonhosted.org/packages/20/33/9228d76339f1ba51e3efef7da3ebd91964d3006217aae13211653193c3ff/multidict-6.7.0-cp313-cp313t-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:9fb0211dfc3b51efea2f349ec92c114d7754dd62c01f81c3e32b765b70c45c9b", size = 228618 }, - { url = "https://files.pythonhosted.org/packages/f8/2d/25d9b566d10cab1c42b3b9e5b11ef79c9111eaf4463b8c257a3bd89e0ead/multidict-6.7.0-cp313-cp313t-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:a027ec240fe73a8d6281872690b988eed307cd7d91b23998ff35ff577ca688b5", size = 257539 }, - { url = "https://files.pythonhosted.org/packages/b6/b1/8d1a965e6637fc33de3c0d8f414485c2b7e4af00f42cab3d84e7b955c222/multidict-6.7.0-cp313-cp313t-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:d1d964afecdf3a8288789df2f5751dc0a8261138c3768d9af117ed384e538fad", size = 256345 }, - { url = "https://files.pythonhosted.org/packages/ba/0c/06b5a8adbdeedada6f4fb8d8f193d44a347223b11939b42953eeb6530b6b/multidict-6.7.0-cp313-cp313t-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:caf53b15b1b7df9fbd0709aa01409000a2b4dd03a5f6f5cc548183c7c8f8b63c", size = 247934 }, - { url = "https://files.pythonhosted.org/packages/8f/31/b2491b5fe167ca044c6eb4b8f2c9f3b8a00b24c432c365358eadac5d7625/multidict-6.7.0-cp313-cp313t-musllinux_1_2_aarch64.whl", hash = "sha256:654030da3197d927f05a536a66186070e98765aa5142794c9904555d3a9d8fb5", size = 245243 }, - { url = "https://files.pythonhosted.org/packages/61/1a/982913957cb90406c8c94f53001abd9eafc271cb3e70ff6371590bec478e/multidict-6.7.0-cp313-cp313t-musllinux_1_2_armv7l.whl", hash = "sha256:2090d3718829d1e484706a2f525e50c892237b2bf9b17a79b059cb98cddc2f10", size = 235878 }, - { url = "https://files.pythonhosted.org/packages/be/c0/21435d804c1a1cf7a2608593f4d19bca5bcbd7a81a70b253fdd1c12af9c0/multidict-6.7.0-cp313-cp313t-musllinux_1_2_i686.whl", hash = "sha256:2d2cfeec3f6f45651b3d408c4acec0ebf3daa9bc8a112a084206f5db5d05b754", size = 243452 }, - { url = "https://files.pythonhosted.org/packages/54/0a/4349d540d4a883863191be6eb9a928846d4ec0ea007d3dcd36323bb058ac/multidict-6.7.0-cp313-cp313t-musllinux_1_2_ppc64le.whl", hash = "sha256:4ef089f985b8c194d341eb2c24ae6e7408c9a0e2e5658699c92f497437d88c3c", size = 252312 }, - { url = "https://files.pythonhosted.org/packages/26/64/d5416038dbda1488daf16b676e4dbfd9674dde10a0cc8f4fc2b502d8125d/multidict-6.7.0-cp313-cp313t-musllinux_1_2_s390x.whl", hash = "sha256:e93a0617cd16998784bf4414c7e40f17a35d2350e5c6f0bd900d3a8e02bd3762", size = 246935 }, - { url = "https://files.pythonhosted.org/packages/9f/8c/8290c50d14e49f35e0bd4abc25e1bc7711149ca9588ab7d04f886cdf03d9/multidict-6.7.0-cp313-cp313t-musllinux_1_2_x86_64.whl", hash = "sha256:f0feece2ef8ebc42ed9e2e8c78fc4aa3cf455733b507c09ef7406364c94376c6", size = 243385 }, - { url = "https://files.pythonhosted.org/packages/ef/a0/f83ae75e42d694b3fbad3e047670e511c138be747bc713cf1b10d5096416/multidict-6.7.0-cp313-cp313t-win32.whl", hash = "sha256:19a1d55338ec1be74ef62440ca9e04a2f001a04d0cc49a4983dc320ff0f3212d", size = 47777 }, - { url = "https://files.pythonhosted.org/packages/dc/80/9b174a92814a3830b7357307a792300f42c9e94664b01dee8e457551fa66/multidict-6.7.0-cp313-cp313t-win_amd64.whl", hash = "sha256:3da4fb467498df97e986af166b12d01f05d2e04f978a9c1c680ea1988e0bc4b6", size = 53104 }, - { url = "https://files.pythonhosted.org/packages/cc/28/04baeaf0428d95bb7a7bea0e691ba2f31394338ba424fb0679a9ed0f4c09/multidict-6.7.0-cp313-cp313t-win_arm64.whl", hash = "sha256:b4121773c49a0776461f4a904cdf6264c88e42218aaa8407e803ca8025872792", size = 45503 }, - { url = "https://files.pythonhosted.org/packages/e2/b1/3da6934455dd4b261d4c72f897e3a5728eba81db59959f3a639245891baa/multidict-6.7.0-cp314-cp314-macosx_10_13_universal2.whl", hash = "sha256:3bab1e4aff7adaa34410f93b1f8e57c4b36b9af0426a76003f441ee1d3c7e842", size = 75128 }, - { url = "https://files.pythonhosted.org/packages/14/2c/f069cab5b51d175a1a2cb4ccdf7a2c2dabd58aa5bd933fa036a8d15e2404/multidict-6.7.0-cp314-cp314-macosx_10_13_x86_64.whl", hash = "sha256:b8512bac933afc3e45fb2b18da8e59b78d4f408399a960339598374d4ae3b56b", size = 44410 }, - { url = "https://files.pythonhosted.org/packages/42/e2/64bb41266427af6642b6b128e8774ed84c11b80a90702c13ac0a86bb10cc/multidict-6.7.0-cp314-cp314-macosx_11_0_arm64.whl", hash = "sha256:79dcf9e477bc65414ebfea98ffd013cb39552b5ecd62908752e0e413d6d06e38", size = 43205 }, - { url = "https://files.pythonhosted.org/packages/02/68/6b086fef8a3f1a8541b9236c594f0c9245617c29841f2e0395d979485cde/multidict-6.7.0-cp314-cp314-manylinux1_i686.manylinux_2_28_i686.manylinux_2_5_i686.whl", hash = "sha256:31bae522710064b5cbeddaf2e9f32b1abab70ac6ac91d42572502299e9953128", size = 245084 }, - { url = "https://files.pythonhosted.org/packages/15/ee/f524093232007cd7a75c1d132df70f235cfd590a7c9eaccd7ff422ef4ae8/multidict-6.7.0-cp314-cp314-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:4a0df7ff02397bb63e2fd22af2c87dfa39e8c7f12947bc524dbdc528282c7e34", size = 252667 }, - { url = "https://files.pythonhosted.org/packages/02/a5/eeb3f43ab45878f1895118c3ef157a480db58ede3f248e29b5354139c2c9/multidict-6.7.0-cp314-cp314-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:7a0222514e8e4c514660e182d5156a415c13ef0aabbd71682fc714e327b95e99", size = 233590 }, - { url = "https://files.pythonhosted.org/packages/6a/1e/76d02f8270b97269d7e3dbd45644b1785bda457b474315f8cf999525a193/multidict-6.7.0-cp314-cp314-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:2397ab4daaf2698eb51a76721e98db21ce4f52339e535725de03ea962b5a3202", size = 264112 }, - { url = "https://files.pythonhosted.org/packages/76/0b/c28a70ecb58963847c2a8efe334904cd254812b10e535aefb3bcce513918/multidict-6.7.0-cp314-cp314-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:8891681594162635948a636c9fe0ff21746aeb3dd5463f6e25d9bea3a8a39ca1", size = 261194 }, - { url = "https://files.pythonhosted.org/packages/b4/63/2ab26e4209773223159b83aa32721b4021ffb08102f8ac7d689c943fded1/multidict-6.7.0-cp314-cp314-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:18706cc31dbf402a7945916dd5cddf160251b6dab8a2c5f3d6d5a55949f676b3", size = 248510 }, - { url = "https://files.pythonhosted.org/packages/93/cd/06c1fa8282af1d1c46fd55c10a7930af652afdce43999501d4d68664170c/multidict-6.7.0-cp314-cp314-musllinux_1_2_aarch64.whl", hash = "sha256:f844a1bbf1d207dd311a56f383f7eda2d0e134921d45751842d8235e7778965d", size = 248395 }, - { url = "https://files.pythonhosted.org/packages/99/ac/82cb419dd6b04ccf9e7e61befc00c77614fc8134362488b553402ecd55ce/multidict-6.7.0-cp314-cp314-musllinux_1_2_armv7l.whl", hash = "sha256:d4393e3581e84e5645506923816b9cc81f5609a778c7e7534054091acc64d1c6", size = 239520 }, - { url = "https://files.pythonhosted.org/packages/fa/f3/a0f9bf09493421bd8716a362e0cd1d244f5a6550f5beffdd6b47e885b331/multidict-6.7.0-cp314-cp314-musllinux_1_2_i686.whl", hash = "sha256:fbd18dc82d7bf274b37aa48d664534330af744e03bccf696d6f4c6042e7d19e7", size = 245479 }, - { url = "https://files.pythonhosted.org/packages/8d/01/476d38fc73a212843f43c852b0eee266b6971f0e28329c2184a8df90c376/multidict-6.7.0-cp314-cp314-musllinux_1_2_ppc64le.whl", hash = "sha256:b6234e14f9314731ec45c42fc4554b88133ad53a09092cc48a88e771c125dadb", size = 258903 }, - { url = "https://files.pythonhosted.org/packages/49/6d/23faeb0868adba613b817d0e69c5f15531b24d462af8012c4f6de4fa8dc3/multidict-6.7.0-cp314-cp314-musllinux_1_2_s390x.whl", hash = "sha256:08d4379f9744d8f78d98c8673c06e202ffa88296f009c71bbafe8a6bf847d01f", size = 252333 }, - { url = "https://files.pythonhosted.org/packages/1e/cc/48d02ac22b30fa247f7dad82866e4b1015431092f4ba6ebc7e77596e0b18/multidict-6.7.0-cp314-cp314-musllinux_1_2_x86_64.whl", hash = "sha256:9fe04da3f79387f450fd0061d4dd2e45a72749d31bf634aecc9e27f24fdc4b3f", size = 243411 }, - { url = "https://files.pythonhosted.org/packages/4a/03/29a8bf5a18abf1fe34535c88adbdfa88c9fb869b5a3b120692c64abe8284/multidict-6.7.0-cp314-cp314-win32.whl", hash = "sha256:fbafe31d191dfa7c4c51f7a6149c9fb7e914dcf9ffead27dcfd9f1ae382b3885", size = 40940 }, - { url = "https://files.pythonhosted.org/packages/82/16/7ed27b680791b939de138f906d5cf2b4657b0d45ca6f5dd6236fdddafb1a/multidict-6.7.0-cp314-cp314-win_amd64.whl", hash = "sha256:2f67396ec0310764b9222a1728ced1ab638f61aadc6226f17a71dd9324f9a99c", size = 45087 }, - { url = "https://files.pythonhosted.org/packages/cd/3c/e3e62eb35a1950292fe39315d3c89941e30a9d07d5d2df42965ab041da43/multidict-6.7.0-cp314-cp314-win_arm64.whl", hash = "sha256:ba672b26069957ee369cfa7fc180dde1fc6f176eaf1e6beaf61fbebbd3d9c000", size = 42368 }, - { url = "https://files.pythonhosted.org/packages/8b/40/cd499bd0dbc5f1136726db3153042a735fffd0d77268e2ee20d5f33c010f/multidict-6.7.0-cp314-cp314t-macosx_10_13_universal2.whl", hash = "sha256:c1dcc7524066fa918c6a27d61444d4ee7900ec635779058571f70d042d86ed63", size = 82326 }, - { url = "https://files.pythonhosted.org/packages/13/8a/18e031eca251c8df76daf0288e6790561806e439f5ce99a170b4af30676b/multidict-6.7.0-cp314-cp314t-macosx_10_13_x86_64.whl", hash = "sha256:27e0b36c2d388dc7b6ced3406671b401e84ad7eb0656b8f3a2f46ed0ce483718", size = 48065 }, - { url = "https://files.pythonhosted.org/packages/40/71/5e6701277470a87d234e433fb0a3a7deaf3bcd92566e421e7ae9776319de/multidict-6.7.0-cp314-cp314t-macosx_11_0_arm64.whl", hash = "sha256:2a7baa46a22e77f0988e3b23d4ede5513ebec1929e34ee9495be535662c0dfe2", size = 46475 }, - { url = "https://files.pythonhosted.org/packages/fe/6a/bab00cbab6d9cfb57afe1663318f72ec28289ea03fd4e8236bb78429893a/multidict-6.7.0-cp314-cp314t-manylinux1_i686.manylinux_2_28_i686.manylinux_2_5_i686.whl", hash = "sha256:7bf77f54997a9166a2f5675d1201520586439424c2511723a7312bdb4bcc034e", size = 239324 }, - { url = "https://files.pythonhosted.org/packages/2a/5f/8de95f629fc22a7769ade8b41028e3e5a822c1f8904f618d175945a81ad3/multidict-6.7.0-cp314-cp314t-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:e011555abada53f1578d63389610ac8a5400fc70ce71156b0aa30d326f1a5064", size = 246877 }, - { url = "https://files.pythonhosted.org/packages/23/b4/38881a960458f25b89e9f4a4fdcb02ac101cfa710190db6e5528841e67de/multidict-6.7.0-cp314-cp314t-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:28b37063541b897fd6a318007373930a75ca6d6ac7c940dbe14731ffdd8d498e", size = 225824 }, - { url = "https://files.pythonhosted.org/packages/1e/39/6566210c83f8a261575f18e7144736059f0c460b362e96e9cf797a24b8e7/multidict-6.7.0-cp314-cp314t-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:05047ada7a2fde2631a0ed706f1fd68b169a681dfe5e4cf0f8e4cb6618bbc2cd", size = 253558 }, - { url = "https://files.pythonhosted.org/packages/00/a3/67f18315100f64c269f46e6c0319fa87ba68f0f64f2b8e7fd7c72b913a0b/multidict-6.7.0-cp314-cp314t-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:716133f7d1d946a4e1b91b1756b23c088881e70ff180c24e864c26192ad7534a", size = 252339 }, - { url = "https://files.pythonhosted.org/packages/c8/2a/1cb77266afee2458d82f50da41beba02159b1d6b1f7973afc9a1cad1499b/multidict-6.7.0-cp314-cp314t-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:d1bed1b467ef657f2a0ae62844a607909ef1c6889562de5e1d505f74457d0b96", size = 244895 }, - { url = "https://files.pythonhosted.org/packages/dd/72/09fa7dd487f119b2eb9524946ddd36e2067c08510576d43ff68469563b3b/multidict-6.7.0-cp314-cp314t-musllinux_1_2_aarch64.whl", hash = "sha256:ca43bdfa5d37bd6aee89d85e1d0831fb86e25541be7e9d376ead1b28974f8e5e", size = 241862 }, - { url = "https://files.pythonhosted.org/packages/65/92/bc1f8bd0853d8669300f732c801974dfc3702c3eeadae2f60cef54dc69d7/multidict-6.7.0-cp314-cp314t-musllinux_1_2_armv7l.whl", hash = "sha256:44b546bd3eb645fd26fb949e43c02a25a2e632e2ca21a35e2e132c8105dc8599", size = 232376 }, - { url = "https://files.pythonhosted.org/packages/09/86/ac39399e5cb9d0c2ac8ef6e10a768e4d3bc933ac808d49c41f9dc23337eb/multidict-6.7.0-cp314-cp314t-musllinux_1_2_i686.whl", hash = "sha256:a6ef16328011d3f468e7ebc326f24c1445f001ca1dec335b2f8e66bed3006394", size = 240272 }, - { url = "https://files.pythonhosted.org/packages/3d/b6/fed5ac6b8563ec72df6cb1ea8dac6d17f0a4a1f65045f66b6d3bf1497c02/multidict-6.7.0-cp314-cp314t-musllinux_1_2_ppc64le.whl", hash = "sha256:5aa873cbc8e593d361ae65c68f85faadd755c3295ea2c12040ee146802f23b38", size = 248774 }, - { url = "https://files.pythonhosted.org/packages/6b/8d/b954d8c0dc132b68f760aefd45870978deec6818897389dace00fcde32ff/multidict-6.7.0-cp314-cp314t-musllinux_1_2_s390x.whl", hash = "sha256:3d7b6ccce016e29df4b7ca819659f516f0bc7a4b3efa3bb2012ba06431b044f9", size = 242731 }, - { url = "https://files.pythonhosted.org/packages/16/9d/a2dac7009125d3540c2f54e194829ea18ac53716c61b655d8ed300120b0f/multidict-6.7.0-cp314-cp314t-musllinux_1_2_x86_64.whl", hash = "sha256:171b73bd4ee683d307599b66793ac80981b06f069b62eea1c9e29c9241aa66b0", size = 240193 }, - { url = "https://files.pythonhosted.org/packages/39/ca/c05f144128ea232ae2178b008d5011d4e2cea86e4ee8c85c2631b1b94802/multidict-6.7.0-cp314-cp314t-win32.whl", hash = "sha256:b2d7f80c4e1fd010b07cb26820aae86b7e73b681ee4889684fb8d2d4537aab13", size = 48023 }, - { url = "https://files.pythonhosted.org/packages/ba/8f/0a60e501584145588be1af5cc829265701ba3c35a64aec8e07cbb71d39bb/multidict-6.7.0-cp314-cp314t-win_amd64.whl", hash = "sha256:09929cab6fcb68122776d575e03c6cc64ee0b8fca48d17e135474b042ce515cd", size = 53507 }, - { url = "https://files.pythonhosted.org/packages/7f/ae/3148b988a9c6239903e786eac19c889fab607c31d6efa7fb2147e5680f23/multidict-6.7.0-cp314-cp314t-win_arm64.whl", hash = "sha256:cc41db090ed742f32bd2d2c721861725e6109681eddf835d0a82bd3a5c382827", size = 44804 }, - { url = "https://files.pythonhosted.org/packages/b7/da/7d22601b625e241d4f23ef1ebff8acfc60da633c9e7e7922e24d10f592b3/multidict-6.7.0-py3-none-any.whl", hash = "sha256:394fc5c42a333c9ffc3e421a4c85e08580d990e08b99f6bf35b4132114c5dcb3", size = 12317 }, -] - -[[package]] -name = "multiprocess" -version = "0.70.18" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "dill" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/72/fd/2ae3826f5be24c6ed87266bc4e59c46ea5b059a103f3d7e7eb76a52aeecb/multiprocess-0.70.18.tar.gz", hash = "sha256:f9597128e6b3e67b23956da07cf3d2e5cba79e2f4e0fba8d7903636663ec6d0d", size = 1798503 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/c8/f8/7f9a8f08bf98cea1dfaa181e05cc8bbcb59cecf044b5a9ac3cce39f9c449/multiprocess-0.70.18-pp310-pypy310_pp73-macosx_10_15_x86_64.whl", hash = "sha256:25d4012dcaaf66b9e8e955f58482b42910c2ee526d532844d8bcf661bbc604df", size = 135083 }, - { url = "https://files.pythonhosted.org/packages/e5/03/b7b10dbfc17b2b3ce07d4d30b3ba8367d0ed32d6d46cd166e298f161dd46/multiprocess-0.70.18-pp310-pypy310_pp73-macosx_11_0_arm64.whl", hash = "sha256:06b19433de0d02afe5869aec8931dd5c01d99074664f806c73896b0d9e527213", size = 135128 }, - { url = "https://files.pythonhosted.org/packages/c1/a3/5f8d3b9690ea5580bee5868ab7d7e2cfca74b7e826b28192b40aa3881cdc/multiprocess-0.70.18-pp310-pypy310_pp73-manylinux_2_28_x86_64.whl", hash = "sha256:6fa1366f994373aaf2d4738b0f56e707caeaa05486e97a7f71ee0853823180c2", size = 135132 }, - { url = "https://files.pythonhosted.org/packages/55/4d/9af0d1279c84618bcd35bf5fd7e371657358c7b0a523e54a9cffb87461f8/multiprocess-0.70.18-pp311-pypy311_pp73-macosx_10_15_x86_64.whl", hash = "sha256:8b8940ae30139e04b076da6c5b83e9398585ebdf0f2ad3250673fef5b2ff06d6", size = 144695 }, - { url = "https://files.pythonhosted.org/packages/17/bf/87323e79dd0562474fad3373c21c66bc6c3c9963b68eb2a209deb4c8575e/multiprocess-0.70.18-pp311-pypy311_pp73-macosx_11_0_arm64.whl", hash = "sha256:0929ba95831adb938edbd5fb801ac45e705ecad9d100b3e653946b7716cb6bd3", size = 144742 }, - { url = "https://files.pythonhosted.org/packages/dd/74/cb8c831e58dc6d5cf450b17c7db87f14294a1df52eb391da948b5e0a0b94/multiprocess-0.70.18-pp311-pypy311_pp73-manylinux_2_28_x86_64.whl", hash = "sha256:4d77f8e4bfe6c6e2e661925bbf9aed4d5ade9a1c6502d5dfc10129b9d1141797", size = 144745 }, - { url = "https://files.pythonhosted.org/packages/ba/d8/0cba6cf51a1a31f20471fbc823a716170c73012ddc4fb85d706630ed6e8f/multiprocess-0.70.18-py310-none-any.whl", hash = "sha256:60c194974c31784019c1f459d984e8f33ee48f10fcf42c309ba97b30d9bd53ea", size = 134948 }, - { url = "https://files.pythonhosted.org/packages/4b/88/9039f2fed1012ef584751d4ceff9ab4a51e5ae264898f0b7cbf44340a859/multiprocess-0.70.18-py311-none-any.whl", hash = "sha256:5aa6eef98e691281b3ad923be2832bf1c55dd2c859acd73e5ec53a66aae06a1d", size = 144462 }, - { url = "https://files.pythonhosted.org/packages/bf/b6/5f922792be93b82ec6b5f270bbb1ef031fd0622847070bbcf9da816502cc/multiprocess-0.70.18-py312-none-any.whl", hash = "sha256:9b78f8e5024b573730bfb654783a13800c2c0f2dfc0c25e70b40d184d64adaa2", size = 150287 }, - { url = "https://files.pythonhosted.org/packages/ee/25/7d7e78e750bc1aecfaf0efbf826c69a791d2eeaf29cf20cba93ff4cced78/multiprocess-0.70.18-py313-none-any.whl", hash = "sha256:871743755f43ef57d7910a38433cfe41319e72be1bbd90b79c7a5ac523eb9334", size = 151917 }, - { url = "https://files.pythonhosted.org/packages/3b/c3/ca84c19bd14cdfc21c388fdcebf08b86a7a470ebc9f5c3c084fc2dbc50f7/multiprocess-0.70.18-py38-none-any.whl", hash = "sha256:dbf705e52a154fe5e90fb17b38f02556169557c2dd8bb084f2e06c2784d8279b", size = 132636 }, - { url = "https://files.pythonhosted.org/packages/6c/28/dd72947e59a6a8c856448a5e74da6201cb5502ddff644fbc790e4bd40b9a/multiprocess-0.70.18-py39-none-any.whl", hash = "sha256:e78ca805a72b1b810c690b6b4cc32579eba34f403094bbbae962b7b5bf9dfcb8", size = 133478 }, -] - -[[package]] -name = "nemo-toolkit" -version = "2.6.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "fsspec" }, - { name = "huggingface-hub" }, - { name = "numba", marker = "sys_platform == 'darwin'" }, - { name = "numba-cuda", extra = ["cu13"], marker = "sys_platform != 'darwin'" }, - { name = "numexpr" }, - { name = "numpy" }, - { name = "onnx" }, - { name = "protobuf" }, - { name = "python-dateutil" }, - { name = "ruamel-yaml" }, - { name = "scikit-learn", version = "1.7.2", source = { registry = "https://pypi.org/simple" }, marker = "python_full_version < '3.11'" }, - { name = "scikit-learn", version = "1.8.0", source = { registry = "https://pypi.org/simple" }, marker = "python_full_version >= '3.11'" }, - { name = "setuptools" }, - { name = "tensorboard" }, - { name = "text-unidecode" }, - { name = "torch" }, - { name = "tqdm" }, - { name = "wget" }, - { name = "wrapt" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/e0/e4/ba1e8c13088659b2d3876e7e9f36e7c3bca73d6c0b505c63f56831a591f1/nemo_toolkit-2.6.0.tar.gz", hash = "sha256:a01296c6615637787b800df8e9425af4f44edace6680877804424cf0c93a8f25", size = 4151390 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/3b/d0/dd254e3f1215fa5a8eb7d0c0c814664362817b95c7c1ed14a638829bb9c7/nemo_toolkit-2.6.0-py3-none-any.whl", hash = "sha256:0849541927e99fb701fc2b4565a54c42d011dc370cc3c3a6b5d8260f521a751f", size = 5899935 }, -] - -[package.optional-dependencies] -asr = [ - { name = "braceexpand" }, - { name = "cloudpickle" }, - { name = "ctc-segmentation" }, - { name = "datasets" }, - { name = "editdistance" }, - { name = "einops" }, - { name = "fiddle" }, - { name = "hydra-core" }, - { name = "inflect" }, - { name = "jiwer" }, - { name = "kaldi-python-io" }, - { name = "kaldialign" }, - { name = "lhotse" }, - { name = "librosa" }, - { name = "lightning" }, - { name = "marshmallow" }, - { name = "mediapy" }, - { name = "num2words" }, - { name = "numpy" }, - { name = "nv-one-logger-core" }, - { name = "nv-one-logger-pytorch-lightning-integration" }, - { name = "nv-one-logger-training-telemetry" }, - { name = "omegaconf" }, - { name = "optuna" }, - { name = "packaging" }, - { name = "pandas" }, - { name = "peft" }, - { name = "pyannote-core" }, - { name = "pyannote-metrics" }, - { name = "pydub" }, - { name = "pyloudnorm" }, - { name = "resampy" }, - { name = "ruamel-yaml" }, - { name = "sacremoses" }, - { name = "scipy", version = "1.15.3", source = { registry = "https://pypi.org/simple" }, marker = "python_full_version < '3.11'" }, - { name = "scipy", version = "1.16.3", source = { registry = "https://pypi.org/simple" }, marker = "python_full_version >= '3.11'" }, - { name = "sentencepiece" }, - { name = "soundfile" }, - { name = "sox" }, - { name = "torchmetrics" }, - { name = "transformers" }, - { name = "wandb" }, - { name = "webdataset" }, - { name = "whisper-normalizer" }, -] - -[[package]] -name = "networkx" -version = "3.4.2" -source = { registry = "https://pypi.org/simple" } -resolution-markers = [ - "python_full_version < '3.11' and sys_platform == 'linux'", - "python_full_version < '3.11' and sys_platform != 'linux'", -] -sdist = { url = "https://files.pythonhosted.org/packages/fd/1d/06475e1cd5264c0b870ea2cc6fdb3e37177c1e565c43f56ff17a10e3937f/networkx-3.4.2.tar.gz", hash = "sha256:307c3669428c5362aab27c8a1260aa8f47c4e91d3891f48be0141738d8d053e1", size = 2151368 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/b9/54/dd730b32ea14ea797530a4479b2ed46a6fb250f682a9cfb997e968bf0261/networkx-3.4.2-py3-none-any.whl", hash = "sha256:df5d4365b724cf81b8c6a7312509d0c22386097011ad1abe274afd5e9d3bbc5f", size = 1723263 }, -] - -[[package]] -name = "networkx" -version = "3.6.1" -source = { registry = "https://pypi.org/simple" } -resolution-markers = [ - "python_full_version >= '3.14' and sys_platform == 'linux'", - "python_full_version == '3.13.*' and sys_platform == 'linux'", - "python_full_version >= '3.14' and sys_platform != 'linux'", - "python_full_version == '3.13.*' and sys_platform != 'linux'", - "python_full_version == '3.12.*' and sys_platform == 'linux'", - "python_full_version == '3.12.*' and sys_platform != 'linux'", - "python_full_version == '3.11.*' and sys_platform == 'linux'", - "python_full_version == '3.11.*' and sys_platform != 'linux'", -] -sdist = { url = "https://files.pythonhosted.org/packages/6a/51/63fe664f3908c97be9d2e4f1158eb633317598cfa6e1fc14af5383f17512/networkx-3.6.1.tar.gz", hash = "sha256:26b7c357accc0c8cde558ad486283728b65b6a95d85ee1cd66bafab4c8168509", size = 2517025 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/9e/c9/b2622292ea83fbb4ec318f5b9ab867d0a28ab43c5717bb85b0a5f6b3b0a4/networkx-3.6.1-py3-none-any.whl", hash = "sha256:d47fbf302e7d9cbbb9e2555a0d267983d2aa476bac30e90dfbe5669bd57f3762", size = 2068504 }, -] - -[[package]] -name = "num2words" -version = "0.5.14" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "docopt" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/f6/58/ad645bd38b4b648eb2fc2ba1b909398e54eb0cbb6a7dbd2b4953e38c9621/num2words-0.5.14.tar.gz", hash = "sha256:b066ec18e56b6616a3b38086b5747daafbaa8868b226a36127e0451c0cf379c6", size = 218213 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/d6/5b/545e9267a1cc080c8a1be2746113a063e34bcdd0f5173fd665a5c13cb234/num2words-0.5.14-py3-none-any.whl", hash = "sha256:1c8e5b00142fc2966fd8d685001e36c4a9911e070d1b120e1beb721fa1edb33d", size = 163525 }, -] - -[[package]] -name = "numba" -version = "0.63.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "llvmlite" }, - { name = "numpy" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/dc/60/0145d479b2209bd8fdae5f44201eceb8ce5a23e0ed54c71f57db24618665/numba-0.63.1.tar.gz", hash = "sha256:b320aa675d0e3b17b40364935ea52a7b1c670c9037c39cf92c49502a75902f4b", size = 2761666 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/5e/ce/5283d4ffa568f795bb0fd61ee1f0efc0c6094b94209259167fc8d4276bde/numba-0.63.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:c6d6bf5bf00f7db629305caaec82a2ffb8abe2bf45eaad0d0738dc7de4113779", size = 2680810 }, - { url = "https://files.pythonhosted.org/packages/0f/72/a8bda517e26d912633b32626333339b7c769ea73a5c688365ea5f88fd07e/numba-0.63.1-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:08653d0dfc9cc9c4c9a8fba29ceb1f2d5340c3b86c4a7e5e07e42b643bc6a2f4", size = 3739735 }, - { url = "https://files.pythonhosted.org/packages/ca/17/1913b7c1173b2db30fb7a9696892a7c4c59aeee777a9af6859e9e01bac51/numba-0.63.1-cp310-cp310-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:f09eebf5650246ce2a4e9a8d38270e2d4b0b0ae978103bafb38ed7adc5ea906e", size = 3446707 }, - { url = "https://files.pythonhosted.org/packages/b4/77/703db56c3061e9fdad5e79c91452947fdeb2ec0bdfe4affe9b144e7025e0/numba-0.63.1-cp310-cp310-win_amd64.whl", hash = "sha256:f8bba17421d865d8c0f7be2142754ebce53e009daba41c44cf6909207d1a8d7d", size = 2747374 }, - { url = "https://files.pythonhosted.org/packages/70/90/5f8614c165d2e256fbc6c57028519db6f32e4982475a372bbe550ea0454c/numba-0.63.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:b33db00f18ccc790ee9911ce03fcdfe9d5124637d1ecc266f5ae0df06e02fec3", size = 2680501 }, - { url = "https://files.pythonhosted.org/packages/dc/9d/d0afc4cf915edd8eadd9b2ab5b696242886ee4f97720d9322650d66a88c6/numba-0.63.1-cp311-cp311-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:7d31ea186a78a7c0f6b1b2a3fe68057fdb291b045c52d86232b5383b6cf4fc25", size = 3744945 }, - { url = "https://files.pythonhosted.org/packages/05/a9/d82f38f2ab73f3be6f838a826b545b80339762ee8969c16a8bf1d39395a8/numba-0.63.1-cp311-cp311-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:ed3bb2fbdb651d6aac394388130a7001aab6f4541837123a4b4ab8b02716530c", size = 3450827 }, - { url = "https://files.pythonhosted.org/packages/18/3f/a9b106e93c5bd7434e65f044bae0d204e20aa7f7f85d72ceb872c7c04216/numba-0.63.1-cp311-cp311-win_amd64.whl", hash = "sha256:1ecbff7688f044b1601be70113e2fb1835367ee0b28ffa8f3adf3a05418c5c87", size = 2747262 }, - { url = "https://files.pythonhosted.org/packages/14/9c/c0974cd3d00ff70d30e8ff90522ba5fbb2bcee168a867d2321d8d0457676/numba-0.63.1-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:2819cd52afa5d8d04e057bdfd54367575105f8829350d8fb5e4066fb7591cc71", size = 2680981 }, - { url = "https://files.pythonhosted.org/packages/cb/70/ea2bc45205f206b7a24ee68a159f5097c9ca7e6466806e7c213587e0c2b1/numba-0.63.1-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:5cfd45dbd3d409e713b1ccfdc2ee72ca82006860254429f4ef01867fdba5845f", size = 3801656 }, - { url = "https://files.pythonhosted.org/packages/0d/82/4f4ba4fd0f99825cbf3cdefd682ca3678be1702b63362011de6e5f71f831/numba-0.63.1-cp312-cp312-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:69a599df6976c03b7ecf15d05302696f79f7e6d10d620367407517943355bcb0", size = 3501857 }, - { url = "https://files.pythonhosted.org/packages/af/fd/6540456efa90b5f6604a86ff50dabefb187e43557e9081adcad3be44f048/numba-0.63.1-cp312-cp312-win_amd64.whl", hash = "sha256:bbad8c63e4fc7eb3cdb2c2da52178e180419f7969f9a685f283b313a70b92af3", size = 2750282 }, - { url = "https://files.pythonhosted.org/packages/57/f7/e19e6eff445bec52dde5bed1ebb162925a8e6f988164f1ae4b3475a73680/numba-0.63.1-cp313-cp313-macosx_12_0_arm64.whl", hash = "sha256:0bd4fd820ef7442dcc07da184c3f54bb41d2bdb7b35bacf3448e73d081f730dc", size = 2680954 }, - { url = "https://files.pythonhosted.org/packages/e9/6c/1e222edba1e20e6b113912caa9b1665b5809433cbcb042dfd133c6f1fd38/numba-0.63.1-cp313-cp313-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:53de693abe4be3bd4dee38e1c55f01c55ff644a6a3696a3670589e6e4c39cde2", size = 3809736 }, - { url = "https://files.pythonhosted.org/packages/76/0a/590bad11a8b3feeac30a24d01198d46bdb76ad15c70d3a530691ce3cae58/numba-0.63.1-cp313-cp313-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:81227821a72a763c3d4ac290abbb4371d855b59fdf85d5af22a47c0e86bf8c7e", size = 3508854 }, - { url = "https://files.pythonhosted.org/packages/4e/f5/3800384a24eed1e4d524669cdbc0b9b8a628800bb1e90d7bd676e5f22581/numba-0.63.1-cp313-cp313-win_amd64.whl", hash = "sha256:eb227b07c2ac37b09432a9bda5142047a2d1055646e089d4a240a2643e508102", size = 2750228 }, - { url = "https://files.pythonhosted.org/packages/36/2f/53be2aa8a55ee2608ebe1231789cbb217f6ece7f5e1c685d2f0752e95a5b/numba-0.63.1-cp314-cp314-macosx_12_0_arm64.whl", hash = "sha256:f180883e5508940cc83de8a8bea37fc6dd20fbe4e5558d4659b8b9bef5ff4731", size = 2681153 }, - { url = "https://files.pythonhosted.org/packages/13/91/53e59c86759a0648282368d42ba732c29524a745fd555ed1fb1df83febbe/numba-0.63.1-cp314-cp314-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:f0938764afa82a47c0e895637a6c55547a42c9e1d35cac42285b1fa60a8b02bb", size = 3778718 }, - { url = "https://files.pythonhosted.org/packages/6c/0c/2be19eba50b0b7636f6d1f69dfb2825530537708a234ba1ff34afc640138/numba-0.63.1-cp314-cp314-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:f90a929fa5094e062d4e0368ede1f4497d5e40f800e80aa5222c4734236a2894", size = 3478712 }, - { url = "https://files.pythonhosted.org/packages/0d/5f/4d0c9e756732577a52211f31da13a3d943d185f7fb90723f56d79c696caa/numba-0.63.1-cp314-cp314-win_amd64.whl", hash = "sha256:8d6d5ce85f572ed4e1a135dbb8c0114538f9dd0e3657eeb0bb64ab204cbe2a8f", size = 2752161 }, -] - -[[package]] -name = "numba-cuda" -version = "0.23.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "cuda-bindings" }, - { name = "cuda-core" }, - { name = "numba" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/10/66/e7d4ed42eb11aa711ea88ad59197dea1e2866aec3542fee20f4dd831c83a/numba_cuda-0.23.0.tar.gz", hash = "sha256:b991cdedaf40b39232d701abf8949e6a4ae93083e1f55f4c9a3aa25e7557cd6c", size = 1342672 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/d9/4b/aa22714f06aa54c73070cd56e9239fd22bcee21bec2a26590289a5af3349/numba_cuda-0.23.0-cp310-cp310-manylinux_2_24_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:a64b597e8d83f7958bd1134f3b0acb30c7426c8c1f0e2e4086f7e27a4cf2768d", size = 1785082 }, - { url = "https://files.pythonhosted.org/packages/5e/fd/99f18e0fb8b71945e7fa3c86ee3699e6dbe5392336b202f97b2bd5c7c0fd/numba_cuda-0.23.0-cp310-cp310-manylinux_2_24_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:feabec56d5cfdaa1fc5b6d22d561ae9056f574b16d8c63d7b3c8dc39235b215a", size = 1783503 }, - { url = "https://files.pythonhosted.org/packages/94/68/d8df1131e089b387ed58af2003a55b522e9d7202a50aff3ea6dfe7627583/numba_cuda-0.23.0-cp310-cp310-win_amd64.whl", hash = "sha256:3731d1c3bf8838e213e2a5fa5a0851deb8477481c6ef57b1ec2ccd76dd663a63", size = 1618034 }, - { url = "https://files.pythonhosted.org/packages/f9/57/9d53908d007df56c857fc391f239a4b22885993a49f0dac5198ad21f8c1d/numba_cuda-0.23.0-cp311-cp311-manylinux_2_24_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:1cfac3483b3a178f894723fada29826f159fcb90d0bf49601b5b80d01cdfe289", size = 1788650 }, - { url = "https://files.pythonhosted.org/packages/b3/ed/60ae303c5d1516da8350d2548099630e7079451f707c35405df649286abd/numba_cuda-0.23.0-cp311-cp311-manylinux_2_24_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:402ef4f6e7e7930056acdad1528d3216a6c779f3e5ae3b6438a8bd99eb87d2b8", size = 1786929 }, - { url = "https://files.pythonhosted.org/packages/41/93/6c5f9ed36f108012e582db2baa831f6e2eb7622b3cb8a4725c139b007a8b/numba_cuda-0.23.0-cp311-cp311-win_amd64.whl", hash = "sha256:65cee8000b0ab7191c22c8ca5ec56e2792b222880ad602dbbf68747e6981a08e", size = 1618060 }, - { url = "https://files.pythonhosted.org/packages/fb/f0/28ffb06c194ef5487949fb03220a3874d4adfc60dbfcf5fb6e6f177f1844/numba_cuda-0.23.0-cp312-cp312-manylinux_2_24_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:4c942afadec6ddcc63828fdb569a0c3c850fbb7576a908e284f55f778e282df6", size = 1829061 }, - { url = "https://files.pythonhosted.org/packages/f1/18/e86d9be3c4b6afb7d94bd82c6c1c92cc1dd009b1a8771d97fc96ac7063f8/numba_cuda-0.23.0-cp312-cp312-manylinux_2_24_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:bee8f0f24bdf0383bb42fe20e3ace018079edc6dced8b850468a7dd07d5cd41f", size = 1827875 }, - { url = "https://files.pythonhosted.org/packages/39/f2/dd4f3b2e33d2e44c9121f417c7ff1d6818424be9b0ffc7e469881f8840f9/numba_cuda-0.23.0-cp312-cp312-win_amd64.whl", hash = "sha256:5cec578ad93f6f69a9a68b99ae2263e4610a87e78d246b81cea5dcb074440992", size = 1618238 }, - { url = "https://files.pythonhosted.org/packages/83/4c/5dd3c35064d5cd63d13fe8e8968a09dbc4a821ddd0dac5cc57ac7f30fa32/numba_cuda-0.23.0-cp313-cp313-manylinux_2_24_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:47554e7079b2ad29133dc0b3003004e3c24e97b5e8192fbb148b28a652da5a02", size = 1833651 }, - { url = "https://files.pythonhosted.org/packages/3a/8f/d18a1d5198ba88bea1dc310d863900f0d10aaff3fa8d5842fb9e4c0d39c9/numba_cuda-0.23.0-cp313-cp313-manylinux_2_24_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:c2f190ec01c9c2859e8c31978cd08c52849d72a4453f34fb59a1736a397e51e3", size = 1832709 }, - { url = "https://files.pythonhosted.org/packages/4e/40/80d8ef470e8aa1e8ed420b17372ec8d86e3547884ebe19a8fa235d1455f1/numba_cuda-0.23.0-cp313-cp313-win_amd64.whl", hash = "sha256:ca2ce5a1e6512ad0eb2666c4763f374ddebd6a2ef411843bcd8f5583e5525716", size = 1618229 }, -] - -[package.optional-dependencies] -cu13 = [ - { name = "cuda-bindings" }, - { name = "cuda-core" }, - { name = "cuda-python" }, - { name = "nvidia-cuda-cccl" }, - { name = "nvidia-cuda-nvrtc" }, - { name = "nvidia-cuda-runtime" }, - { name = "nvidia-nvjitlink" }, - { name = "nvidia-nvvm" }, -] - -[[package]] -name = "numexpr" -version = "2.13.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "numpy" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/8d/ca/c1217ae2c15c3284a9e219c269624f80fa1582622eb0400c711a26f84a43/numexpr-2.13.1.tar.gz", hash = "sha256:ecb722249c2d6ed7fefe8504bb17e056481a5f31233c23a7ee02085c3d661fa1", size = 119296 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/35/40/ec43ef49857b10111801e85b103f178d3d4473fa42ad3719fa059f55a257/numexpr-2.13.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:bdbc2b93ac59667f0ba725b24cd3b5559c300e91e179d09c74ebaf8c8961eef6", size = 162934 }, - { url = "https://files.pythonhosted.org/packages/4b/c4/cc0af2756065f1f97acf2237f6809ce72c0abfd31cc59e54e6f11a4fb1cb/numexpr-2.13.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:ad6b5dfc191c766e3ec89d2e3f956f7ef3181a1f8bf2bb00ec48fb3bf97b44ac", size = 151820 }, - { url = "https://files.pythonhosted.org/packages/25/36/59a71bd2cbd11ab8220474bd11a98cf4b4d65e90050ef68588e1caa40bb7/numexpr-2.13.1-cp310-cp310-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:a12dbd4c07a8303c6f01cdade531d75c9b4f5b8f72cbe5821d8f9197ee6fba47", size = 449129 }, - { url = "https://files.pythonhosted.org/packages/38/a6/f8ffb8519a20f3e58ad87a82a7ea31fbcd970dbac7d3c9531b5af4ece65a/numexpr-2.13.1-cp310-cp310-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:2de5c8ca2f25690d48e475d53a3524876164227cf4044743818f5704c28a8639", size = 439777 }, - { url = "https://files.pythonhosted.org/packages/0a/eb/290e3a871190e07d5de0ec1342cf38d2453b4b235f93a903b7d6bb969a7f/numexpr-2.13.1-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:533ec2d77fc059e3868e9798ef2f13ab57161517cd2e0c521bb33d1dc99068ca", size = 1413818 }, - { url = "https://files.pythonhosted.org/packages/ae/68/922980751260b62e451f5b21adaa63581ec2d7c06ef2ed9e356b8529fea8/numexpr-2.13.1-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:a75ddffc36f6b7a679fbc7df492685aed7e8888aec80ec2cd8e30f21fc019caa", size = 1462677 }, - { url = "https://files.pythonhosted.org/packages/1a/26/d111add556589fa8e37863fa89fac5ea914904982540eaf24adcf02994ae/numexpr-2.13.1-cp310-cp310-win32.whl", hash = "sha256:790af35095626ad2d02201c56ac2d49ae45fc95a02af85f40808752ed32ee103", size = 166606 }, - { url = "https://files.pythonhosted.org/packages/71/f5/f27ba83d134ce76708dec714e253665560e9e083425b6ff3d1d536b872e3/numexpr-2.13.1-cp310-cp310-win_amd64.whl", hash = "sha256:aadf3118b6ef87294277ffb77a9562970228341aaaa4b78de634a43ea8ea2c6e", size = 159891 }, - { url = "https://files.pythonhosted.org/packages/60/aa/734ccb5b2d62ddb8c903adf1be8bf668df7fd31f886f8a274203a8317a43/numexpr-2.13.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:bdf62745e072c670151c0705bddfe3f33c341dacb7eb255ddb1e8d2a257bfef5", size = 162936 }, - { url = "https://files.pythonhosted.org/packages/4b/bc/bc081354c99d896b5986bb6683bc7f36e221e1464d9b8a5d9c5ad7a29c13/numexpr-2.13.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:91cf0521d8fed3f804640c4a6d22b5d9813d7e64b32c38215de163c7f092f7cc", size = 151819 }, - { url = "https://files.pythonhosted.org/packages/bf/6d/c3a1c3c113a5cf72b431a9f4433511eb35f2063836ed1020f21781ca77aa/numexpr-2.13.1-cp311-cp311-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:58e2f111756fff63e27e495473d950e4c98bbebca55aa1572798b59110d6c84b", size = 450816 }, - { url = "https://files.pythonhosted.org/packages/77/45/634492e37e31c9db273b6f0d39a83759bfda58ea32690a892b6a5246cfc4/numexpr-2.13.1-cp311-cp311-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:5a5a37b74561ed8dbd5f9be182d94419fa53f452e2d7d3e8d6dbef35a20f19f7", size = 441502 }, - { url = "https://files.pythonhosted.org/packages/6a/04/cfd65881165fd800e0ea17985b03793a7a16488c1a93257b2cfa658bb73a/numexpr-2.13.1-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:78cb76676e63f02dcf507e3c563888018a68b6a2e2cd444628e09df270dfd0b2", size = 1415631 }, - { url = "https://files.pythonhosted.org/packages/c7/15/0d037d173c3cd0254fdf1cf148fa4aa79da10119a688cc2e1027de3e7cee/numexpr-2.13.1-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:d29b3351de4c43b56d2ef7f138ab7a8988e797291bcbbd56d545e4e7902f254a", size = 1464365 }, - { url = "https://files.pythonhosted.org/packages/73/4a/56d3aca7bea28f66d82f0b9577a632c2ad18834e9467e06fc3595ddc8c54/numexpr-2.13.1-cp311-cp311-win32.whl", hash = "sha256:912488ddbd500937bb6f4dfc010bdb3bf757a76e0b93db2f2c56db49ef6b9351", size = 166612 }, - { url = "https://files.pythonhosted.org/packages/52/62/5bd094657e051b1cb9e71f65ef4db733b50a24645f2380057fffc52aca6a/numexpr-2.13.1-cp311-cp311-win_amd64.whl", hash = "sha256:66d0292f3b9dc5faadb4dd8a89d733321ff01c9699aee0c3cdbf513c9505e39c", size = 159890 }, - { url = "https://files.pythonhosted.org/packages/b5/24/b87ad61f09132d92d92e93da8940055f1282ee30c913737ae977cebebab6/numexpr-2.13.1-cp312-cp312-macosx_10_13_x86_64.whl", hash = "sha256:6aa48c2f2bfa142dfe260441486452be8f70b5551c17bc846fccf76123d4a226", size = 162534 }, - { url = "https://files.pythonhosted.org/packages/91/b8/8ea90b2c64ef26b14866a38d13bb496195856b810c1a18a96cb89693b6af/numexpr-2.13.1-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:67a3dd8b51e94251f535a9a404f1ac939a3ebeb9398caad20ae9d0de37c6d3b3", size = 151938 }, - { url = "https://files.pythonhosted.org/packages/ab/65/4679408c4c61badbd12671920479918e2893c8488de8d5c7f801b3a5f57d/numexpr-2.13.1-cp312-cp312-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:ca152998d44ea30b45ad6b8a050ac4a9408b61a17508df87ad0d919335d79b44", size = 452166 }, - { url = "https://files.pythonhosted.org/packages/31/1b/11a1202f8b67dce8e119a9f6481d839b152cc0084940a146b52f8f38685b/numexpr-2.13.1-cp312-cp312-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:b4280c8f7cc024846be8fdd6582572bb0b6bad98fb2a68a367ef5e6e2e130d5f", size = 443123 }, - { url = "https://files.pythonhosted.org/packages/7b/5e/271bf56efac177abe6e5d5349365e460a2a4205a514c99e0b2203d827264/numexpr-2.13.1-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:b86e1daa4e27d6bf6304008ed4630a055babf863db2ec8f282b4058bbfe466bd", size = 1417039 }, - { url = "https://files.pythonhosted.org/packages/72/33/6b3164fdc553eceec901793f9df467a7b4151e21772514fc2a392f12c42f/numexpr-2.13.1-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:30d189fc52ee4a33b869a0592553cd2ed686c20cded21b2ddf347a4d143f1bea", size = 1465878 }, - { url = "https://files.pythonhosted.org/packages/f1/3e/037e9dc96f9681e7af694bf5abf699b137f1fccb8bb829c50505e98d60ba/numexpr-2.13.1-cp312-cp312-win32.whl", hash = "sha256:e926b59d385de2396935b362143ac2c282176875cf8ee7baba0a150b58421b5c", size = 166740 }, - { url = "https://files.pythonhosted.org/packages/b6/7e/92c01806608a3d1c88aabbda42e4849036200a5209af374bfa5c614aa5e5/numexpr-2.13.1-cp312-cp312-win_amd64.whl", hash = "sha256:8230a8f7cd4e6ba4022643c85e119aa4ca90412267ef20acdf1f54fb3136680d", size = 159987 }, - { url = "https://files.pythonhosted.org/packages/55/c8/eee9c3e78f856483b21d836b1db821451b91a1f3f249ead1cdc290fb4172/numexpr-2.13.1-cp313-cp313-macosx_10_13_x86_64.whl", hash = "sha256:0e4314ee477a2cfb9ecf4b15f2ef24bf7859f62b35de3caef297136ff25bb0b0", size = 162535 }, - { url = "https://files.pythonhosted.org/packages/a9/ed/aba137ba850fcac3f5e0c2e15b26420e00e93ab9a258757a4c1f2dca65de/numexpr-2.13.1-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:d82d088f67647861b61a7b0e0148fd7487000a20909d65734821dd27e0839a68", size = 151946 }, - { url = "https://files.pythonhosted.org/packages/8a/c9/13f421b2322c14062f9b22af9baf4c560c25ef2a9f7dd34a33f606c9cf6a/numexpr-2.13.1-cp313-cp313-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:c615b13976e6332336a052d5b03be1fed231bc1afe07699f4c7cc116c7c3092c", size = 455493 }, - { url = "https://files.pythonhosted.org/packages/bc/7d/3c5baf2bfe1c1504cbd3d993592e0e2596e83a61d6647e89fc8b38764496/numexpr-2.13.1-cp313-cp313-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:4874124bccc3c2462558ad2a75029bcc2d1c63ee4914b263bb06339e757efb85", size = 446051 }, - { url = "https://files.pythonhosted.org/packages/6c/be/702faf87d4e7eac4b69eda20a143c6d4f149ca9c5a990db9aed58fa55ad0/numexpr-2.13.1-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:0fc7b5b0f8d7ba6c81e948b1d967a56097194c894e4f57852ed8639fc653def2", size = 1417017 }, - { url = "https://files.pythonhosted.org/packages/8b/2c/c39be0f3e42afb2cb296d203d80d4dcf9a71d94be478ca4407e1a4cfe645/numexpr-2.13.1-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:e22104ab53f0933b5b522829149990cb74e0a8ec4b69ff0e6545eb4641b3f013", size = 1465833 }, - { url = "https://files.pythonhosted.org/packages/46/31/6fb1c5e450c09c6ba9808e27e7546e3c68ee4def4dfcbe9c9dc1cfc23d78/numexpr-2.13.1-cp313-cp313-win32.whl", hash = "sha256:824aea72663ec123e042341cea4a2a2b3c71f315e4bc58ee5035ffc7f945bd29", size = 166742 }, - { url = "https://files.pythonhosted.org/packages/57/dd/7b11419523a0eb20bb99c6c3134f44b760be956557eaf79cdb851360c4fe/numexpr-2.13.1-cp313-cp313-win_amd64.whl", hash = "sha256:9c7b1c3e9f398a5b062d9740c48ca454238bf1be433f0f75fe68619527bb7f1a", size = 159991 }, - { url = "https://files.pythonhosted.org/packages/5d/cd/e9d03848038d4c4b7237f46ebd8a8d3ee8fd5a87f44c87c487550a7bd637/numexpr-2.13.1-cp313-cp313t-macosx_10_13_x86_64.whl", hash = "sha256:366a7887c2bad86e6f64666e178886f606cf8e81a6871df450d19f0f83421501", size = 163275 }, - { url = "https://files.pythonhosted.org/packages/a7/c9/d63cbca11844247c87ad90d28428e3362de4c94d2589db9cc63b199e4a03/numexpr-2.13.1-cp313-cp313t-macosx_11_0_arm64.whl", hash = "sha256:33ff9f071d06aaa0276cb5e2369efd517fe155ea091e43790f1f8bfd85e64d29", size = 152647 }, - { url = "https://files.pythonhosted.org/packages/77/e4/71c393ddfcfacfe9a9afc1624a61a15804384c5bb72b78934bb2f96a380a/numexpr-2.13.1-cp313-cp313t-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:c29a204b1d35941c088ec39a79c2e83e382729e4066b4b1f882aa5f70bf929a8", size = 465611 }, - { url = "https://files.pythonhosted.org/packages/91/fd/d99652d4d99ff6606f8d4e39e52220351c3314d0216e8ee2ea6a2a12b652/numexpr-2.13.1-cp313-cp313t-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:40e02db74d66c5b0a81c925838f42ec2d58cc99b49cbaf682f06ac03d9ff4102", size = 456451 }, - { url = "https://files.pythonhosted.org/packages/98/2f/83dcc8b9d4edbc1814e552c090404bfa7e43dfcb7729a20df1d10281592b/numexpr-2.13.1-cp313-cp313t-musllinux_1_2_aarch64.whl", hash = "sha256:36bd9a2b9bda42506377c7510c61f76e08d50da77ffb86a7a15cc5d57c56bb0f", size = 1425799 }, - { url = "https://files.pythonhosted.org/packages/89/7f/90d9f4d5dfb7f033a8133dff6703245420113fb66babb5c465314680f9e1/numexpr-2.13.1-cp313-cp313t-musllinux_1_2_x86_64.whl", hash = "sha256:b9203651668a3994cf3fe52e079ff6be1c74bf775622edbc226e94f3d8ec8ec4", size = 1473868 }, - { url = "https://files.pythonhosted.org/packages/35/ed/5eacf6c584e1c5e8408f63ae0f909f85c6933b0a6aac730ce3c971a9dd60/numexpr-2.13.1-cp313-cp313t-win32.whl", hash = "sha256:b73774176b15fe88242e7ed174b5be5f2e3e830d2cd663234b1495628a30854c", size = 167412 }, - { url = "https://files.pythonhosted.org/packages/a7/63/1a3890f8c9bbac0c91ef04781bc765d23fbd964ef0f66b98637eace0c431/numexpr-2.13.1-cp313-cp313t-win_amd64.whl", hash = "sha256:b9e6228db24b7faa96fbb2beee55f90fc8b0fe167cf288f8481c53ff5e95865a", size = 160894 }, - { url = "https://files.pythonhosted.org/packages/47/f5/fa44066b3b41f6be89ad0ba778897f323c7939fb24a04ab559a577909a95/numexpr-2.13.1-cp314-cp314-macosx_10_13_x86_64.whl", hash = "sha256:cbadcbd2cf0822d595ccf5345c69478e9fe42d556b9823e6b0636a3efdf990f0", size = 162593 }, - { url = "https://files.pythonhosted.org/packages/e4/a1/c8bb07ebc37a3a65df5c0f280bac3f9b90f9cf4f94de18a0b0db6bcd5ddd/numexpr-2.13.1-cp314-cp314-macosx_11_0_arm64.whl", hash = "sha256:a189d514e8aa321ef1c650a2873000c08f843b3e3e66d69072005996ac25809c", size = 151986 }, - { url = "https://files.pythonhosted.org/packages/69/30/4adf5699154b65a9b6a80ed1a3d3e4ab915318d6be54dd77c840a9ca7546/numexpr-2.13.1-cp314-cp314-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:b6b01e9301bed8f89f6d561d79dcaa8731a75cc50efc072526cfbc07df74226c", size = 455718 }, - { url = "https://files.pythonhosted.org/packages/01/eb/39e056a2887e18cdeed1ffbf1dcd7cba2bd010ad8ac7d4db42c389f0e310/numexpr-2.13.1-cp314-cp314-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:d7749e8c0ff0bae41a534e56fab667e529f528645a0216bb64260773ae8cb697", size = 446008 }, - { url = "https://files.pythonhosted.org/packages/34/b8/f96d0bce9fa499f9fe07c439e6f389318e79f20eae5296db9cacb364e5e0/numexpr-2.13.1-cp314-cp314-musllinux_1_2_aarch64.whl", hash = "sha256:0b0f326542185c23fca53e10fee3c39bdadc8d69a03c613938afaf3eea31e77f", size = 1417260 }, - { url = "https://files.pythonhosted.org/packages/2c/3e/5f75fb72c8ad71148bf8a13f8c3860a26ec4c39ae08b1b8c48201ae8ba1b/numexpr-2.13.1-cp314-cp314-musllinux_1_2_x86_64.whl", hash = "sha256:33cc6d662a606cc5184c7faef1d7b176474a8c46b8b0d2df9ff0fa67ed56425f", size = 1465903 }, - { url = "https://files.pythonhosted.org/packages/50/93/a0578f726b39864f88ac259c70d7ee194ff9d223697c11fa9fb053dd4907/numexpr-2.13.1-cp314-cp314-win32.whl", hash = "sha256:71f442fd01ebfa77fce1bac37f671aed3c0d47a55e460beac54b89e767fbc0fa", size = 168583 }, - { url = "https://files.pythonhosted.org/packages/72/fe/ae6877a6cda902df19678ce6d5b56135f19b6a15d48eadbbdb64ba2daa24/numexpr-2.13.1-cp314-cp314-win_amd64.whl", hash = "sha256:208cd9422d87333e24deb2fe492941cd13b65dc8b9ce665de045a0be89e9a254", size = 162393 }, - { url = "https://files.pythonhosted.org/packages/b7/d9/70ee0e4098d31fbcc0b6d7d18bfc24ce0f3ea6f824e9c490ce4a9ea18336/numexpr-2.13.1-cp314-cp314t-macosx_10_13_x86_64.whl", hash = "sha256:37d31824b9c021078046bb2aa36aa1da23edaa7a6a8636ee998bf89a2f104722", size = 163277 }, - { url = "https://files.pythonhosted.org/packages/5e/24/fbf234d4dd154074d98519b10a44ed050ccbcd317f04fe24cbe1860d0e6b/numexpr-2.13.1-cp314-cp314t-macosx_11_0_arm64.whl", hash = "sha256:15cee07c74e4792993cd2ecd46c5683815e8758ac56e1d4d236d2c9eb9e8ae01", size = 152647 }, - { url = "https://files.pythonhosted.org/packages/d3/8e/2e4d64742f63d3932a62a96735e7b9140296b4e004e7cf2f8f9e227edf28/numexpr-2.13.1-cp314-cp314t-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:65cb46136f068ede2fc415c5f3d722f2c7dde3eda04ceafcfbcac03933f5d997", size = 465879 }, - { url = "https://files.pythonhosted.org/packages/40/06/3724d1e26cec148e2309a92376acf9f6aba506dee28e60b740acb4d90ef1/numexpr-2.13.1-cp314-cp314t-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:abc3c1601380c90659b9ac0241357c5788ab58de148f56c5f98adffe293c308c", size = 456726 }, - { url = "https://files.pythonhosted.org/packages/92/78/64441da9c97a2b62be60ced33ef686368af6eb1157e032ee77aca4261603/numexpr-2.13.1-cp314-cp314t-musllinux_1_2_aarch64.whl", hash = "sha256:2836e900377ce27e99c043a35e008bc911c51781cea47623612a4e498dfa9592", size = 1426003 }, - { url = "https://files.pythonhosted.org/packages/27/57/892857f8903f69e8f5e25332630215a32eb17a0b2535ed6d8d5ea3ba52e7/numexpr-2.13.1-cp314-cp314t-musllinux_1_2_x86_64.whl", hash = "sha256:f4e4c5b38bb5695fff119672c3462d9a36875256947bafb2df4117b3271fd6a3", size = 1473992 }, - { url = "https://files.pythonhosted.org/packages/6f/5c/c6b5163798fb3631da641361fde77c082e46f56bede50757353462058ef0/numexpr-2.13.1-cp314-cp314t-win32.whl", hash = "sha256:156591eb23684542fd53ca1cbefff872c47c429a200655ef7e59dd8c03eeeaef", size = 169242 }, - { url = "https://files.pythonhosted.org/packages/b4/13/61598a6c5802aefc74e113c3f1b89c49a71e76ebb8b179940560408fdaa3/numexpr-2.13.1-cp314-cp314t-win_amd64.whl", hash = "sha256:a2cc21b2d2e59db63006f190dbf20f5485dd846770870504ff2a72c8d0406e4e", size = 163406 }, -] - -[[package]] -name = "numpy" -version = "1.26.4" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/65/6e/09db70a523a96d25e115e71cc56a6f9031e7b8cd166c1ac8438307c14058/numpy-1.26.4.tar.gz", hash = "sha256:2a02aba9ed12e4ac4eb3ea9421c420301a0c6460d9830d74a9df87efa4912010", size = 15786129 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/a7/94/ace0fdea5241a27d13543ee117cbc65868e82213fb31a8eb7fe9ff23f313/numpy-1.26.4-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:9ff0f4f29c51e2803569d7a51c2304de5554655a60c5d776e35b4a41413830d0", size = 20631468 }, - { url = "https://files.pythonhosted.org/packages/20/f7/b24208eba89f9d1b58c1668bc6c8c4fd472b20c45573cb767f59d49fb0f6/numpy-1.26.4-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:2e4ee3380d6de9c9ec04745830fd9e2eccb3e6cf790d39d7b98ffd19b0dd754a", size = 13966411 }, - { url = "https://files.pythonhosted.org/packages/fc/a5/4beee6488160798683eed5bdb7eead455892c3b4e1f78d79d8d3f3b084ac/numpy-1.26.4-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d209d8969599b27ad20994c8e41936ee0964e6da07478d6c35016bc386b66ad4", size = 14219016 }, - { url = "https://files.pythonhosted.org/packages/4b/d7/ecf66c1cd12dc28b4040b15ab4d17b773b87fa9d29ca16125de01adb36cd/numpy-1.26.4-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ffa75af20b44f8dba823498024771d5ac50620e6915abac414251bd971b4529f", size = 18240889 }, - { url = "https://files.pythonhosted.org/packages/24/03/6f229fe3187546435c4f6f89f6d26c129d4f5bed40552899fcf1f0bf9e50/numpy-1.26.4-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:62b8e4b1e28009ef2846b4c7852046736bab361f7aeadeb6a5b89ebec3c7055a", size = 13876746 }, - { url = "https://files.pythonhosted.org/packages/39/fe/39ada9b094f01f5a35486577c848fe274e374bbf8d8f472e1423a0bbd26d/numpy-1.26.4-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:a4abb4f9001ad2858e7ac189089c42178fcce737e4169dc61321660f1a96c7d2", size = 18078620 }, - { url = "https://files.pythonhosted.org/packages/d5/ef/6ad11d51197aad206a9ad2286dc1aac6a378059e06e8cf22cd08ed4f20dc/numpy-1.26.4-cp310-cp310-win32.whl", hash = "sha256:bfe25acf8b437eb2a8b2d49d443800a5f18508cd811fea3181723922a8a82b07", size = 5972659 }, - { url = "https://files.pythonhosted.org/packages/19/77/538f202862b9183f54108557bfda67e17603fc560c384559e769321c9d92/numpy-1.26.4-cp310-cp310-win_amd64.whl", hash = "sha256:b97fe8060236edf3662adfc2c633f56a08ae30560c56310562cb4f95500022d5", size = 15808905 }, - { url = "https://files.pythonhosted.org/packages/11/57/baae43d14fe163fa0e4c47f307b6b2511ab8d7d30177c491960504252053/numpy-1.26.4-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:4c66707fabe114439db9068ee468c26bbdf909cac0fb58686a42a24de1760c71", size = 20630554 }, - { url = "https://files.pythonhosted.org/packages/1a/2e/151484f49fd03944c4a3ad9c418ed193cfd02724e138ac8a9505d056c582/numpy-1.26.4-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:edd8b5fe47dab091176d21bb6de568acdd906d1887a4584a15a9a96a1dca06ef", size = 13997127 }, - { url = "https://files.pythonhosted.org/packages/79/ae/7e5b85136806f9dadf4878bf73cf223fe5c2636818ba3ab1c585d0403164/numpy-1.26.4-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7ab55401287bfec946ced39700c053796e7cc0e3acbef09993a9ad2adba6ca6e", size = 14222994 }, - { url = "https://files.pythonhosted.org/packages/3a/d0/edc009c27b406c4f9cbc79274d6e46d634d139075492ad055e3d68445925/numpy-1.26.4-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:666dbfb6ec68962c033a450943ded891bed2d54e6755e35e5835d63f4f6931d5", size = 18252005 }, - { url = "https://files.pythonhosted.org/packages/09/bf/2b1aaf8f525f2923ff6cfcf134ae5e750e279ac65ebf386c75a0cf6da06a/numpy-1.26.4-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:96ff0b2ad353d8f990b63294c8986f1ec3cb19d749234014f4e7eb0112ceba5a", size = 13885297 }, - { url = "https://files.pythonhosted.org/packages/df/a0/4e0f14d847cfc2a633a1c8621d00724f3206cfeddeb66d35698c4e2cf3d2/numpy-1.26.4-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:60dedbb91afcbfdc9bc0b1f3f402804070deed7392c23eb7a7f07fa857868e8a", size = 18093567 }, - { url = "https://files.pythonhosted.org/packages/d2/b7/a734c733286e10a7f1a8ad1ae8c90f2d33bf604a96548e0a4a3a6739b468/numpy-1.26.4-cp311-cp311-win32.whl", hash = "sha256:1af303d6b2210eb850fcf03064d364652b7120803a0b872f5211f5234b399f20", size = 5968812 }, - { url = "https://files.pythonhosted.org/packages/3f/6b/5610004206cf7f8e7ad91c5a85a8c71b2f2f8051a0c0c4d5916b76d6cbb2/numpy-1.26.4-cp311-cp311-win_amd64.whl", hash = "sha256:cd25bcecc4974d09257ffcd1f098ee778f7834c3ad767fe5db785be9a4aa9cb2", size = 15811913 }, - { url = "https://files.pythonhosted.org/packages/95/12/8f2020a8e8b8383ac0177dc9570aad031a3beb12e38847f7129bacd96228/numpy-1.26.4-cp312-cp312-macosx_10_9_x86_64.whl", hash = "sha256:b3ce300f3644fb06443ee2222c2201dd3a89ea6040541412b8fa189341847218", size = 20335901 }, - { url = "https://files.pythonhosted.org/packages/75/5b/ca6c8bd14007e5ca171c7c03102d17b4f4e0ceb53957e8c44343a9546dcc/numpy-1.26.4-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:03a8c78d01d9781b28a6989f6fa1bb2c4f2d51201cf99d3dd875df6fbd96b23b", size = 13685868 }, - { url = "https://files.pythonhosted.org/packages/79/f8/97f10e6755e2a7d027ca783f63044d5b1bc1ae7acb12afe6a9b4286eac17/numpy-1.26.4-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:9fad7dcb1aac3c7f0584a5a8133e3a43eeb2fe127f47e3632d43d677c66c102b", size = 13925109 }, - { url = "https://files.pythonhosted.org/packages/0f/50/de23fde84e45f5c4fda2488c759b69990fd4512387a8632860f3ac9cd225/numpy-1.26.4-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:675d61ffbfa78604709862923189bad94014bef562cc35cf61d3a07bba02a7ed", size = 17950613 }, - { url = "https://files.pythonhosted.org/packages/4c/0c/9c603826b6465e82591e05ca230dfc13376da512b25ccd0894709b054ed0/numpy-1.26.4-cp312-cp312-musllinux_1_1_aarch64.whl", hash = "sha256:ab47dbe5cc8210f55aa58e4805fe224dac469cde56b9f731a4c098b91917159a", size = 13572172 }, - { url = "https://files.pythonhosted.org/packages/76/8c/2ba3902e1a0fc1c74962ea9bb33a534bb05984ad7ff9515bf8d07527cadd/numpy-1.26.4-cp312-cp312-musllinux_1_1_x86_64.whl", hash = "sha256:1dda2e7b4ec9dd512f84935c5f126c8bd8b9f2fc001e9f54af255e8c5f16b0e0", size = 17786643 }, - { url = "https://files.pythonhosted.org/packages/28/4a/46d9e65106879492374999e76eb85f87b15328e06bd1550668f79f7b18c6/numpy-1.26.4-cp312-cp312-win32.whl", hash = "sha256:50193e430acfc1346175fcbdaa28ffec49947a06918b7b92130744e81e640110", size = 5677803 }, - { url = "https://files.pythonhosted.org/packages/16/2e/86f24451c2d530c88daf997cb8d6ac622c1d40d19f5a031ed68a4b73a374/numpy-1.26.4-cp312-cp312-win_amd64.whl", hash = "sha256:08beddf13648eb95f8d867350f6a018a4be2e5ad54c8d8caed89ebca558b2818", size = 15517754 }, -] - -[[package]] -name = "nv-one-logger-core" -version = "2.3.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "overrides" }, - { name = "pydantic" }, - { name = "strenum" }, - { name = "toml" }, - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/3b/37/963095797035f371e0db6ea761f5aaccb624fc786af217115b423baeb0e2/nv_one_logger_core-2.3.1.tar.gz", hash = "sha256:cbb2f87604c78b96a302f32d87199902129d76153a73a20f8455a250b3246c1d", size = 52640 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/ee/c4/ea91554c4fcbff66057f667690101d7a4b965605741350ac661b03fa6c46/nv_one_logger_core-2.3.1-py3-none-any.whl", hash = "sha256:0c8b77bcdac4daa1ea913bf8d4afd2a057bd5526e3654ac39f67caba157341a6", size = 63066 }, -] - -[[package]] -name = "nv-one-logger-pytorch-lightning-integration" -version = "2.3.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "lightning" }, - { name = "nv-one-logger-core" }, - { name = "nv-one-logger-training-telemetry" }, - { name = "setuptools" }, - { name = "strenum" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/0c/d0/3475b7ab17d367362f650fb0419e8669f41e63c1018f4a8ac2fbecfd2e85/nv_one_logger_pytorch_lightning_integration-2.3.1.tar.gz", hash = "sha256:b32d99b6a8f02a16538bcade939b0a7edd7249e936aacefe336b5519447340c3", size = 10979 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/f0/56/01a55efb365b6864646b4ac941a1d9de66f024e880764510ba5a7a63f62c/nv_one_logger_pytorch_lightning_integration-2.3.1-py3-none-any.whl", hash = "sha256:f92904055fb0082516480cc1e3dd0bb6cedb2b033985ebfd4814b9cbf7da2cb2", size = 9822 }, -] - -[[package]] -name = "nv-one-logger-training-telemetry" -version = "2.3.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "nv-one-logger-core" }, - { name = "strenum" }, - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/c5/21/016fa067967734d52f1ccf5a2a37a1a65216f2d7053bc2b85872cce956ca/nv_one_logger_training_telemetry-2.3.1.tar.gz", hash = "sha256:8c67940ea71799afaf1f46df3ba2f52f93aea26321c6f1c1d54aae02efc2a4af", size = 44435 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/e5/15/97e6e4ddfe5fc35bcee74a45b7c33fb73abb83713c7dfa26420b971a86c3/nv_one_logger_training_telemetry-2.3.1-py3-none-any.whl", hash = "sha256:5319443829b59378a498c3c62ac98973e14f31be675c229ff2b14e2fe109aa0b", size = 44140 }, -] - -[[package]] -name = "nvidia-cublas-cu12" -version = "12.8.4.1" -source = { registry = "https://pypi.org/simple" } -wheels = [ - { url = "https://files.pythonhosted.org/packages/dc/61/e24b560ab2e2eaeb3c839129175fb330dfcfc29e5203196e5541a4c44682/nvidia_cublas_cu12-12.8.4.1-py3-none-manylinux_2_27_x86_64.whl", hash = "sha256:8ac4e771d5a348c551b2a426eda6193c19aa630236b418086020df5ba9667142", size = 594346921 }, -] - -[[package]] -name = "nvidia-cuda-cccl" -version = "13.1.78" -source = { registry = "https://pypi.org/simple" } -wheels = [ - { url = "https://files.pythonhosted.org/packages/2e/49/60587ce90ee6f681df7f3adceb3ea6a209aecee72dacd1caeb657e81d3c8/nvidia_cuda_cccl-13.1.78-py3-none-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:e5b5a2177135625619822d3707ede6e6d466761d47dbd71638da156ef3b43c0a", size = 3477655 }, - { url = "https://files.pythonhosted.org/packages/f2/f6/ffc31be6366411a227c50d10c822c8b34c0db80c2908c4a4d4dcb420661b/nvidia_cuda_cccl-13.1.78-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:24734b3dbb8ea7ee228ad4b19c8069ce862490513298e392641e077121e592c5", size = 3427730 }, - { url = "https://files.pythonhosted.org/packages/90/b9/d959b9de8320e43e29f7d8ce005a6c2d7de6a819ba66a431d8e215ace72f/nvidia_cuda_cccl-13.1.78-py3-none-win_amd64.whl", hash = "sha256:cecac91e456bf7ab82e0a417ef4d4517e1843d7ed720cf7f5aa13c597aa9e356", size = 3415700 }, -] - -[[package]] -name = "nvidia-cuda-cupti-cu12" -version = "12.8.90" -source = { registry = "https://pypi.org/simple" } -wheels = [ - { url = "https://files.pythonhosted.org/packages/f8/02/2adcaa145158bf1a8295d83591d22e4103dbfd821bcaf6f3f53151ca4ffa/nvidia_cuda_cupti_cu12-12.8.90-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:ea0cb07ebda26bb9b29ba82cda34849e73c166c18162d3913575b0c9db9a6182", size = 10248621 }, -] - -[[package]] -name = "nvidia-cuda-nvrtc" -version = "13.1.80" -source = { registry = "https://pypi.org/simple" } -wheels = [ - { url = "https://files.pythonhosted.org/packages/20/a9/9cca9907117247235afaface5dcce2dba139ab2eabb1fb37a5f4a4d867dd/nvidia_cuda_nvrtc-13.1.80-py3-none-manylinux2010_x86_64.manylinux_2_12_x86_64.whl", hash = "sha256:a3c2c104fe8cabaa289643cc1c999d7b1e614316eaee15c43747b14b9c8bcaad", size = 46570803 }, - { url = "https://files.pythonhosted.org/packages/a9/17/bc712515e9446eb0d3b7f4d20034beb6c176a1dcefcd98a3f4100a80b8a1/nvidia_cuda_nvrtc-13.1.80-py3-none-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:2df04c70712197e3d648d06896c13d1eb3a9af4948d94bd3fc3409368d9e8a75", size = 44318878 }, - { url = "https://files.pythonhosted.org/packages/7f/87/ad826e8f006acc78def3c4713d5ca41e9852f30c192a3181bd7623b29879/nvidia_cuda_nvrtc-13.1.80-py3-none-win_amd64.whl", hash = "sha256:3fc8f312b99195a49f9a81057a73187358d5e7280f983effc717a3eb3b93595c", size = 40564896 }, -] - -[[package]] -name = "nvidia-cuda-nvrtc-cu12" -version = "12.8.93" -source = { registry = "https://pypi.org/simple" } -wheels = [ - { url = "https://files.pythonhosted.org/packages/05/6b/32f747947df2da6994e999492ab306a903659555dddc0fbdeb9d71f75e52/nvidia_cuda_nvrtc_cu12-12.8.93-py3-none-manylinux2010_x86_64.manylinux_2_12_x86_64.whl", hash = "sha256:a7756528852ef889772a84c6cd89d41dfa74667e24cca16bb31f8f061e3e9994", size = 88040029 }, -] - -[[package]] -name = "nvidia-cuda-runtime" -version = "13.1.80" -source = { registry = "https://pypi.org/simple" } -wheels = [ - { url = "https://files.pythonhosted.org/packages/32/b7/8075a985c5a4a828d850f244266c67134f60b28d1106cd4abe6187bedcc3/nvidia_cuda_runtime-13.1.80-py3-none-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:5f1aa8fd779cae1a78aa8fdbaa5e4f245300baff120099d9b5ee84e3f4506a29", size = 2317343 }, - { url = "https://files.pythonhosted.org/packages/fe/60/03858bc3954b3263eedcff7626712c656b6b5a0d7d25bcb1fc1ebee9d4f1/nvidia_cuda_runtime-13.1.80-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:dac25b52e69050e5a7a84957e6df470a85b185f8f648a99a1ab1dfce027576cf", size = 2297816 }, - { url = "https://files.pythonhosted.org/packages/d7/c2/737165823dd0c7717d02faf892bd697f517dd017b9044e5d663f4e03b979/nvidia_cuda_runtime-13.1.80-py3-none-win_amd64.whl", hash = "sha256:00db7158108d3ec7fe213fa78e2640d664b5b1b6b85485007dde29ccaa6abef6", size = 3093656 }, -] - -[[package]] -name = "nvidia-cuda-runtime-cu12" -version = "12.8.90" -source = { registry = "https://pypi.org/simple" } -wheels = [ - { url = "https://files.pythonhosted.org/packages/0d/9b/a997b638fcd068ad6e4d53b8551a7d30fe8b404d6f1804abf1df69838932/nvidia_cuda_runtime_cu12-12.8.90-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:adade8dcbd0edf427b7204d480d6066d33902cab2a4707dcfc48a2d0fd44ab90", size = 954765 }, -] - -[[package]] -name = "nvidia-cudnn-cu12" -version = "9.10.2.21" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "nvidia-cublas-cu12" }, -] -wheels = [ - { url = "https://files.pythonhosted.org/packages/ba/51/e123d997aa098c61d029f76663dedbfb9bc8dcf8c60cbd6adbe42f76d049/nvidia_cudnn_cu12-9.10.2.21-py3-none-manylinux_2_27_x86_64.whl", hash = "sha256:949452be657fa16687d0930933f032835951ef0892b37d2d53824d1a84dc97a8", size = 706758467 }, -] - -[[package]] -name = "nvidia-cufft-cu12" -version = "11.3.3.83" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "nvidia-nvjitlink-cu12" }, -] -wheels = [ - { url = "https://files.pythonhosted.org/packages/1f/13/ee4e00f30e676b66ae65b4f08cb5bcbb8392c03f54f2d5413ea99a5d1c80/nvidia_cufft_cu12-11.3.3.83-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:4d2dd21ec0b88cf61b62e6b43564355e5222e4a3fb394cac0db101f2dd0d4f74", size = 193118695 }, -] - -[[package]] -name = "nvidia-cufile-cu12" -version = "1.13.1.3" -source = { registry = "https://pypi.org/simple" } -wheels = [ - { url = "https://files.pythonhosted.org/packages/bb/fe/1bcba1dfbfb8d01be8d93f07bfc502c93fa23afa6fd5ab3fc7c1df71038a/nvidia_cufile_cu12-1.13.1.3-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:1d069003be650e131b21c932ec3d8969c1715379251f8d23a1860554b1cb24fc", size = 1197834 }, -] - -[[package]] -name = "nvidia-curand-cu12" -version = "10.3.9.90" -source = { registry = "https://pypi.org/simple" } -wheels = [ - { url = "https://files.pythonhosted.org/packages/fb/aa/6584b56dc84ebe9cf93226a5cde4d99080c8e90ab40f0c27bda7a0f29aa1/nvidia_curand_cu12-10.3.9.90-py3-none-manylinux_2_27_x86_64.whl", hash = "sha256:b32331d4f4df5d6eefa0554c565b626c7216f87a06a4f56fab27c3b68a830ec9", size = 63619976 }, -] - -[[package]] -name = "nvidia-cusolver-cu12" -version = "11.7.3.90" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "nvidia-cublas-cu12" }, - { name = "nvidia-cusparse-cu12" }, - { name = "nvidia-nvjitlink-cu12" }, -] -wheels = [ - { url = "https://files.pythonhosted.org/packages/85/48/9a13d2975803e8cf2777d5ed57b87a0b6ca2cc795f9a4f59796a910bfb80/nvidia_cusolver_cu12-11.7.3.90-py3-none-manylinux_2_27_x86_64.whl", hash = "sha256:4376c11ad263152bd50ea295c05370360776f8c3427b30991df774f9fb26c450", size = 267506905 }, -] - -[[package]] -name = "nvidia-cusparse-cu12" -version = "12.5.8.93" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "nvidia-nvjitlink-cu12" }, -] -wheels = [ - { url = "https://files.pythonhosted.org/packages/c2/f5/e1854cb2f2bcd4280c44736c93550cc300ff4b8c95ebe370d0aa7d2b473d/nvidia_cusparse_cu12-12.5.8.93-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:1ec05d76bbbd8b61b06a80e1eaf8cf4959c3d4ce8e711b65ebd0443bb0ebb13b", size = 288216466 }, -] - -[[package]] -name = "nvidia-cusparselt-cu12" -version = "0.7.1" -source = { registry = "https://pypi.org/simple" } -wheels = [ - { url = "https://files.pythonhosted.org/packages/56/79/12978b96bd44274fe38b5dde5cfb660b1d114f70a65ef962bcbbed99b549/nvidia_cusparselt_cu12-0.7.1-py3-none-manylinux2014_x86_64.whl", hash = "sha256:f1bb701d6b930d5a7cea44c19ceb973311500847f81b634d802b7b539dc55623", size = 287193691 }, -] - -[[package]] -name = "nvidia-nccl-cu12" -version = "2.27.5" -source = { registry = "https://pypi.org/simple" } -wheels = [ - { url = "https://files.pythonhosted.org/packages/6e/89/f7a07dc961b60645dbbf42e80f2bc85ade7feb9a491b11a1e973aa00071f/nvidia_nccl_cu12-2.27.5-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:ad730cf15cb5d25fe849c6e6ca9eb5b76db16a80f13f425ac68d8e2e55624457", size = 322348229 }, -] - -[[package]] -name = "nvidia-nvjitlink" -version = "13.1.80" -source = { registry = "https://pypi.org/simple" } -wheels = [ - { url = "https://files.pythonhosted.org/packages/79/c9/1c1ee01649c7bbee2c232461f5c706f0ec7d9b5098a238fcc604b2f67672/nvidia_nvjitlink-13.1.80-py3-none-manylinux2010_x86_64.manylinux_2_12_x86_64.whl", hash = "sha256:8021a5146c336dafaee111ee78a1c665bc873161f51a5ae8989135b9d4a85434", size = 40936616 }, - { url = "https://files.pythonhosted.org/packages/35/4a/740252bec635d6e2d9ee40182906ea205b4b97afb169d82f4caf50bc7223/nvidia_nvjitlink-13.1.80-py3-none-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:a3b7866b52d8c959d4c8ba4e8572d9439b129caf81f6eea1410f2a9f82e6336d", size = 38857485 }, - { url = "https://files.pythonhosted.org/packages/20/e8/9b4b41d8169c080f6243bcaa74dcbfa9871ea7ba0e3efe65180f6c1f336f/nvidia_nvjitlink-13.1.80-py3-none-win_amd64.whl", hash = "sha256:8a1f07dc72078fe7700a1c7a8b1fc057c36fbbd70710a5a89f5e20464601c31f", size = 36865727 }, -] - -[[package]] -name = "nvidia-nvjitlink-cu12" -version = "12.8.93" -source = { registry = "https://pypi.org/simple" } -wheels = [ - { url = "https://files.pythonhosted.org/packages/f6/74/86a07f1d0f42998ca31312f998bd3b9a7eff7f52378f4f270c8679c77fb9/nvidia_nvjitlink_cu12-12.8.93-py3-none-manylinux2010_x86_64.manylinux_2_12_x86_64.whl", hash = "sha256:81ff63371a7ebd6e6451970684f916be2eab07321b73c9d244dc2b4da7f73b88", size = 39254836 }, -] - -[[package]] -name = "nvidia-nvshmem-cu12" -version = "3.3.20" -source = { registry = "https://pypi.org/simple" } -wheels = [ - { url = "https://files.pythonhosted.org/packages/3b/6c/99acb2f9eb85c29fc6f3a7ac4dccfd992e22666dd08a642b303311326a97/nvidia_nvshmem_cu12-3.3.20-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:d00f26d3f9b2e3c3065be895e3059d6479ea5c638a3f38c9fec49b1b9dd7c1e5", size = 124657145 }, -] - -[[package]] -name = "nvidia-nvtx-cu12" -version = "12.8.90" -source = { registry = "https://pypi.org/simple" } -wheels = [ - { url = "https://files.pythonhosted.org/packages/a2/eb/86626c1bbc2edb86323022371c39aa48df6fd8b0a1647bc274577f72e90b/nvidia_nvtx_cu12-12.8.90-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:5b17e2001cc0d751a5bc2c6ec6d26ad95913324a4adb86788c944f8ce9ba441f", size = 89954 }, -] - -[[package]] -name = "nvidia-nvvm" -version = "13.1.80" -source = { registry = "https://pypi.org/simple" } -wheels = [ - { url = "https://files.pythonhosted.org/packages/94/c4/050e882206afc18e409e6428403bf1b86155a8213051a4a07d2e8a0190dc/nvidia_nvvm-13.1.80-py3-none-manylinux2010_x86_64.manylinux_2_12_x86_64.whl", hash = "sha256:058519d3e290cad0ad4034addad3f72fef2782f383e33d3c5b38172ed8fdb144", size = 63386684 }, - { url = "https://files.pythonhosted.org/packages/6f/3a/31ab71b86983611364cd416cce61891a231b1385774f44800a01aedc1abe/nvidia_nvvm-13.1.80-py3-none-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:a3d29e47059b6535b91172262a8fc78f4d8f15168c0090cc42b09a2393425465", size = 61001764 }, - { url = "https://files.pythonhosted.org/packages/b2/8c/3f4302e50dea44870f15277b33c8d47cf6d6808c42aa0cda448e1eef0e16/nvidia_nvvm-13.1.80-py3-none-win_amd64.whl", hash = "sha256:aecefb4e9b4c6c0fb799db27ef32e7a3cf94d426c6689ff5679397a466af1769", size = 55413868 }, -] - -[[package]] -name = "omegaconf" -version = "2.3.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "antlr4-python3-runtime" }, - { name = "pyyaml" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/09/48/6388f1bb9da707110532cb70ec4d2822858ddfb44f1cdf1233c20a80ea4b/omegaconf-2.3.0.tar.gz", hash = "sha256:d5d4b6d29955cc50ad50c46dc269bcd92c6e00f5f90d23ab5fee7bfca4ba4cc7", size = 3298120 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/e3/94/1843518e420fa3ed6919835845df698c7e27e183cb997394e4a670973a65/omegaconf-2.3.0-py3-none-any.whl", hash = "sha256:7b4df175cdb08ba400f45cae3bdcae7ba8365db4d165fc65fd04b050ab63b46b", size = 79500 }, -] - -[[package]] -name = "onnx" -version = "1.19.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "ml-dtypes" }, - { name = "numpy" }, - { name = "protobuf" }, - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/5b/bf/b0a63ee9f3759dcd177b28c6f2cb22f2aecc6d9b3efecaabc298883caa5f/onnx-1.19.0.tar.gz", hash = "sha256:aa3f70b60f54a29015e41639298ace06adf1dd6b023b9b30f1bca91bb0db9473", size = 11949859 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/00/b3/8a6f3b05d18dffdc7c18839bd829587c826c8513f4bdbe21ddf37dacce50/onnx-1.19.0-cp310-cp310-macosx_12_0_universal2.whl", hash = "sha256:e927d745939d590f164e43c5aec7338c5a75855a15130ee795f492fc3a0fa565", size = 18310869 }, - { url = "https://files.pythonhosted.org/packages/b9/92/550d6155ab3f2c00e95add1726397c95b4b79d6eb4928d049ff591ad4c84/onnx-1.19.0-cp310-cp310-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:c6cdcb237c5c4202463bac50417c5a7f7092997a8469e8b7ffcd09f51de0f4a9", size = 18028144 }, - { url = "https://files.pythonhosted.org/packages/79/21/9bcc715ea6d9aab3f6c583bfc59504a14777e39e0591030e7345f4e40315/onnx-1.19.0-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:ed0b85a33deacb65baffe6ca4ce91adf2bb906fa2dee3856c3c94e163d2eb563", size = 18200923 }, - { url = "https://files.pythonhosted.org/packages/c8/90/3a6f0741ff22270e2f4b741f440ab68ba5525ebc94775cd6f2c01f531374/onnx-1.19.0-cp310-cp310-win32.whl", hash = "sha256:89a9cefe75547aec14a796352c2243e36793bbbcb642d8897118595ab0c2395b", size = 16332097 }, - { url = "https://files.pythonhosted.org/packages/4c/4c/ef61d359865712803d488672607023d36bfcd21fa008d8dc1d6ee8e8b23c/onnx-1.19.0-cp310-cp310-win_amd64.whl", hash = "sha256:a16a82bfdf4738691c0a6eda5293928645ab8b180ab033df84080817660b5e66", size = 16451402 }, - { url = "https://files.pythonhosted.org/packages/db/5c/b959b17608cfb6ccf6359b39fe56a5b0b7d965b3d6e6a3c0add90812c36e/onnx-1.19.0-cp311-cp311-macosx_12_0_universal2.whl", hash = "sha256:206f00c47b85b5c7af79671e3307147407991a17994c26974565aadc9e96e4e4", size = 18312580 }, - { url = "https://files.pythonhosted.org/packages/2c/ee/ac052bbbc832abe0debb784c2c57f9582444fb5f51d63c2967fd04432444/onnx-1.19.0-cp311-cp311-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:4d7bee94abaac28988b50da675ae99ef8dd3ce16210d591fbd0b214a5930beb3", size = 18029165 }, - { url = "https://files.pythonhosted.org/packages/5c/c9/8687ba0948d46fd61b04e3952af9237883bbf8f16d716e7ed27e688d73b8/onnx-1.19.0-cp311-cp311-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:7730b96b68c0c354bbc7857961bb4909b9aaa171360a8e3708d0a4c749aaadeb", size = 18202125 }, - { url = "https://files.pythonhosted.org/packages/e2/16/6249c013e81bd689f46f96c7236d7677f1af5dd9ef22746716b48f10e506/onnx-1.19.0-cp311-cp311-win32.whl", hash = "sha256:7cb7a3ad8059d1a0dfdc5e0a98f71837d82002e441f112825403b137227c2c97", size = 16332738 }, - { url = "https://files.pythonhosted.org/packages/6a/28/34a1e2166e418c6a78e5c82e66f409d9da9317832f11c647f7d4e23846a6/onnx-1.19.0-cp311-cp311-win_amd64.whl", hash = "sha256:d75452a9be868bd30c3ef6aa5991df89bbfe53d0d90b2325c5e730fbd91fff85", size = 16452303 }, - { url = "https://files.pythonhosted.org/packages/e6/b7/639664626e5ba8027860c4d2a639ee02b37e9c322215c921e9222513c3aa/onnx-1.19.0-cp311-cp311-win_arm64.whl", hash = "sha256:23c7959370d7b3236f821e609b0af7763cff7672a758e6c1fc877bac099e786b", size = 16425340 }, - { url = "https://files.pythonhosted.org/packages/0d/94/f56f6ca5e2f921b28c0f0476705eab56486b279f04e1d568ed64c14e7764/onnx-1.19.0-cp312-cp312-macosx_12_0_universal2.whl", hash = "sha256:61d94e6498ca636756f8f4ee2135708434601b2892b7c09536befb19bc8ca007", size = 18322331 }, - { url = "https://files.pythonhosted.org/packages/c8/00/8cc3f3c40b54b28f96923380f57c9176872e475face726f7d7a78bd74098/onnx-1.19.0-cp312-cp312-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:224473354462f005bae985c72028aaa5c85ab11de1b71d55b06fdadd64a667dd", size = 18027513 }, - { url = "https://files.pythonhosted.org/packages/61/90/17c4d2566fd0117a5e412688c9525f8950d467f477fbd574e6b32bc9cb8d/onnx-1.19.0-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:1ae475c85c89bc4d1f16571006fd21a3e7c0e258dd2c091f6e8aafb083d1ed9b", size = 18202278 }, - { url = "https://files.pythonhosted.org/packages/bc/6e/a9383d9cf6db4ac761a129b081e9fa5d0cd89aad43cf1e3fc6285b915c7d/onnx-1.19.0-cp312-cp312-win32.whl", hash = "sha256:323f6a96383a9cdb3960396cffea0a922593d221f3929b17312781e9f9b7fb9f", size = 16333080 }, - { url = "https://files.pythonhosted.org/packages/a7/2e/3ff480a8c1fa7939662bdc973e41914add2d4a1f2b8572a3c39c2e4982e5/onnx-1.19.0-cp312-cp312-win_amd64.whl", hash = "sha256:50220f3499a499b1a15e19451a678a58e22ad21b34edf2c844c6ef1d9febddc2", size = 16453927 }, - { url = "https://files.pythonhosted.org/packages/57/37/ad500945b1b5c154fe9d7b826b30816ebd629d10211ea82071b5bcc30aa4/onnx-1.19.0-cp312-cp312-win_arm64.whl", hash = "sha256:efb768299580b786e21abe504e1652ae6189f0beed02ab087cd841cb4bb37e43", size = 16426022 }, - { url = "https://files.pythonhosted.org/packages/be/29/d7b731f63d243f815d9256dce0dca3c151dcaa1ac59f73e6ee06c9afbe91/onnx-1.19.0-cp313-cp313-macosx_12_0_universal2.whl", hash = "sha256:9aed51a4b01acc9ea4e0fe522f34b2220d59e9b2a47f105ac8787c2e13ec5111", size = 18322412 }, - { url = "https://files.pythonhosted.org/packages/58/f5/d3106becb42cb374f0e17ff4c9933a97f1ee1d6a798c9452067f7d3ff61b/onnx-1.19.0-cp313-cp313-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:ce2cdc3eb518bb832668c4ea9aeeda01fbaa59d3e8e5dfaf7aa00f3d37119404", size = 18026565 }, - { url = "https://files.pythonhosted.org/packages/83/fa/b086d17bab3900754c7ffbabfb244f8e5e5da54a34dda2a27022aa2b373b/onnx-1.19.0-cp313-cp313-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:8b546bd7958734b6abcd40cfede3d025e9c274fd96334053a288ab11106bd0aa", size = 18202077 }, - { url = "https://files.pythonhosted.org/packages/35/f2/5e2dfb9d4cf873f091c3f3c6d151f071da4295f9893fbf880f107efe3447/onnx-1.19.0-cp313-cp313-win32.whl", hash = "sha256:03086bffa1cf5837430cf92f892ca0cd28c72758d8905578c2bf8ffaf86c6743", size = 16333198 }, - { url = "https://files.pythonhosted.org/packages/79/67/b3751a35c2522f62f313156959575619b8fa66aa883db3adda9d897d8eb2/onnx-1.19.0-cp313-cp313-win_amd64.whl", hash = "sha256:1715b51eb0ab65272e34ef51cb34696160204b003566cd8aced2ad20a8f95cb8", size = 16453836 }, - { url = "https://files.pythonhosted.org/packages/14/b9/1df85effc960fbbb90bb7bc36eb3907c676b104bc2f88bce022bcfdaef63/onnx-1.19.0-cp313-cp313-win_arm64.whl", hash = "sha256:6bf5acdb97a3ddd6e70747d50b371846c313952016d0c41133cbd8f61b71a8d5", size = 16425877 }, - { url = "https://files.pythonhosted.org/packages/23/2b/089174a1427be9149f37450f8959a558ba20f79fca506ba461d59379d3a1/onnx-1.19.0-cp313-cp313t-macosx_12_0_universal2.whl", hash = "sha256:46cf29adea63e68be0403c68de45ba1b6acc9bb9592c5ddc8c13675a7c71f2cb", size = 18348546 }, - { url = "https://files.pythonhosted.org/packages/c0/d6/3458f0e3a9dc7677675d45d7d6528cb84ad321c8670cc10c69b32c3e03da/onnx-1.19.0-cp313-cp313t-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:246f0de1345498d990a443d55a5b5af5101a3e25a05a2c3a5fe8b7bd7a7d0707", size = 18033067 }, - { url = "https://files.pythonhosted.org/packages/e4/16/6e4130e1b4b29465ee1fb07d04e8d6f382227615c28df8f607ba50909e2a/onnx-1.19.0-cp313-cp313t-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:ae0d163ffbc250007d984b8dd692a4e2e4506151236b50ca6e3560b612ccf9ff", size = 18205741 }, - { url = "https://files.pythonhosted.org/packages/fe/d8/f64d010fd024b2a2b11ce0c4ee179e4f8f6d4ccc95f8184961c894c22af1/onnx-1.19.0-cp313-cp313t-win_amd64.whl", hash = "sha256:7c151604c7cca6ae26161c55923a7b9b559df3344938f93ea0074d2d49e7fe78", size = 16453839 }, - { url = "https://files.pythonhosted.org/packages/67/ec/8761048eabef4dad55af4c002c672d139b9bd47c3616abaed642a1710063/onnx-1.19.0-cp314-cp314-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:236bc0e60d7c0f4159300da639953dd2564df1c195bce01caba172a712e75af4", size = 18027605 }, -] - -[[package]] -name = "optuna" -version = "4.6.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "alembic" }, - { name = "colorlog" }, - { name = "numpy" }, - { name = "packaging" }, - { name = "pyyaml" }, - { name = "sqlalchemy" }, - { name = "tqdm" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/6b/81/08f90f194eed78178064a9383432eca95611e2c5331e7b01e2418ce4b15a/optuna-4.6.0.tar.gz", hash = "sha256:89e38c2447c7f793a726617b8043f01e31f0bad54855040db17eb3b49404a369", size = 477444 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/58/de/3d8455b08cb6312f8cc46aacdf16c71d4d881a1db4a4140fc5ef31108422/optuna-4.6.0-py3-none-any.whl", hash = "sha256:4c3a9facdef2b2dd7e3e2a8ae3697effa70fae4056fcf3425cfc6f5a40feb069", size = 404708 }, -] - -[[package]] -name = "overrides" -version = "7.7.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/36/86/b585f53236dec60aba864e050778b25045f857e17f6e5ea0ae95fe80edd2/overrides-7.7.0.tar.gz", hash = "sha256:55158fa3d93b98cc75299b1e67078ad9003ca27945c76162c1c0766d6f91820a", size = 22812 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/2c/ab/fc8290c6a4c722e5514d80f62b2dc4c4df1a68a41d1364e625c35990fcf3/overrides-7.7.0-py3-none-any.whl", hash = "sha256:c7ed9d062f78b8e4c1a7b70bd8796b35ead4d9f510227ef9c5dc7626c60d7e49", size = 17832 }, -] - -[[package]] -name = "packaging" -version = "24.2" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/d0/63/68dbb6eb2de9cb10ee4c9c14a0148804425e13c4fb20d61cce69f53106da/packaging-24.2.tar.gz", hash = "sha256:c228a6dc5e932d346bc5739379109d49e8853dd8223571c7c5b55260edc0b97f", size = 163950 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/88/ef/eb23f262cca3c0c4eb7ab1933c3b1f03d021f2c48f54763065b6f0e321be/packaging-24.2-py3-none-any.whl", hash = "sha256:09abb1bccd265c01f4a3aa3f7a7db064b36514d2cba19a2f694fe6150451a759", size = 65451 }, -] - -[[package]] -name = "pandas" -version = "2.3.3" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "numpy" }, - { name = "python-dateutil" }, - { name = "pytz" }, - { name = "tzdata" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/33/01/d40b85317f86cf08d853a4f495195c73815fdf205eef3993821720274518/pandas-2.3.3.tar.gz", hash = "sha256:e05e1af93b977f7eafa636d043f9f94c7ee3ac81af99c13508215942e64c993b", size = 4495223 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/3d/f7/f425a00df4fcc22b292c6895c6831c0c8ae1d9fac1e024d16f98a9ce8749/pandas-2.3.3-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:376c6446ae31770764215a6c937f72d917f214b43560603cd60da6408f183b6c", size = 11555763 }, - { url = "https://files.pythonhosted.org/packages/13/4f/66d99628ff8ce7857aca52fed8f0066ce209f96be2fede6cef9f84e8d04f/pandas-2.3.3-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:e19d192383eab2f4ceb30b412b22ea30690c9e618f78870357ae1d682912015a", size = 10801217 }, - { url = "https://files.pythonhosted.org/packages/1d/03/3fc4a529a7710f890a239cc496fc6d50ad4a0995657dccc1d64695adb9f4/pandas-2.3.3-cp310-cp310-manylinux_2_24_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:5caf26f64126b6c7aec964f74266f435afef1c1b13da3b0636c7518a1fa3e2b1", size = 12148791 }, - { url = "https://files.pythonhosted.org/packages/40/a8/4dac1f8f8235e5d25b9955d02ff6f29396191d4e665d71122c3722ca83c5/pandas-2.3.3-cp310-cp310-manylinux_2_24_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:dd7478f1463441ae4ca7308a70e90b33470fa593429f9d4c578dd00d1fa78838", size = 12769373 }, - { url = "https://files.pythonhosted.org/packages/df/91/82cc5169b6b25440a7fc0ef3a694582418d875c8e3ebf796a6d6470aa578/pandas-2.3.3-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:4793891684806ae50d1288c9bae9330293ab4e083ccd1c5e383c34549c6e4250", size = 13200444 }, - { url = "https://files.pythonhosted.org/packages/10/ae/89b3283800ab58f7af2952704078555fa60c807fff764395bb57ea0b0dbd/pandas-2.3.3-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:28083c648d9a99a5dd035ec125d42439c6c1c525098c58af0fc38dd1a7a1b3d4", size = 13858459 }, - { url = "https://files.pythonhosted.org/packages/85/72/530900610650f54a35a19476eca5104f38555afccda1aa11a92ee14cb21d/pandas-2.3.3-cp310-cp310-win_amd64.whl", hash = "sha256:503cf027cf9940d2ceaa1a93cfb5f8c8c7e6e90720a2850378f0b3f3b1e06826", size = 11346086 }, - { url = "https://files.pythonhosted.org/packages/c1/fa/7ac648108144a095b4fb6aa3de1954689f7af60a14cf25583f4960ecb878/pandas-2.3.3-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:602b8615ebcc4a0c1751e71840428ddebeb142ec02c786e8ad6b1ce3c8dec523", size = 11578790 }, - { url = "https://files.pythonhosted.org/packages/9b/35/74442388c6cf008882d4d4bdfc4109be87e9b8b7ccd097ad1e7f006e2e95/pandas-2.3.3-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:8fe25fc7b623b0ef6b5009149627e34d2a4657e880948ec3c840e9402e5c1b45", size = 10833831 }, - { url = "https://files.pythonhosted.org/packages/fe/e4/de154cbfeee13383ad58d23017da99390b91d73f8c11856f2095e813201b/pandas-2.3.3-cp311-cp311-manylinux_2_24_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:b468d3dad6ff947df92dcb32ede5b7bd41a9b3cceef0a30ed925f6d01fb8fa66", size = 12199267 }, - { url = "https://files.pythonhosted.org/packages/bf/c9/63f8d545568d9ab91476b1818b4741f521646cbdd151c6efebf40d6de6f7/pandas-2.3.3-cp311-cp311-manylinux_2_24_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:b98560e98cb334799c0b07ca7967ac361a47326e9b4e5a7dfb5ab2b1c9d35a1b", size = 12789281 }, - { url = "https://files.pythonhosted.org/packages/f2/00/a5ac8c7a0e67fd1a6059e40aa08fa1c52cc00709077d2300e210c3ce0322/pandas-2.3.3-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:1d37b5848ba49824e5c30bedb9c830ab9b7751fd049bc7914533e01c65f79791", size = 13240453 }, - { url = "https://files.pythonhosted.org/packages/27/4d/5c23a5bc7bd209231618dd9e606ce076272c9bc4f12023a70e03a86b4067/pandas-2.3.3-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:db4301b2d1f926ae677a751eb2bd0e8c5f5319c9cb3f88b0becbbb0b07b34151", size = 13890361 }, - { url = "https://files.pythonhosted.org/packages/8e/59/712db1d7040520de7a4965df15b774348980e6df45c129b8c64d0dbe74ef/pandas-2.3.3-cp311-cp311-win_amd64.whl", hash = "sha256:f086f6fe114e19d92014a1966f43a3e62285109afe874f067f5abbdcbb10e59c", size = 11348702 }, - { url = "https://files.pythonhosted.org/packages/9c/fb/231d89e8637c808b997d172b18e9d4a4bc7bf31296196c260526055d1ea0/pandas-2.3.3-cp312-cp312-macosx_10_13_x86_64.whl", hash = "sha256:6d21f6d74eb1725c2efaa71a2bfc661a0689579b58e9c0ca58a739ff0b002b53", size = 11597846 }, - { url = "https://files.pythonhosted.org/packages/5c/bd/bf8064d9cfa214294356c2d6702b716d3cf3bb24be59287a6a21e24cae6b/pandas-2.3.3-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:3fd2f887589c7aa868e02632612ba39acb0b8948faf5cc58f0850e165bd46f35", size = 10729618 }, - { url = "https://files.pythonhosted.org/packages/57/56/cf2dbe1a3f5271370669475ead12ce77c61726ffd19a35546e31aa8edf4e/pandas-2.3.3-cp312-cp312-manylinux_2_24_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:ecaf1e12bdc03c86ad4a7ea848d66c685cb6851d807a26aa245ca3d2017a1908", size = 11737212 }, - { url = "https://files.pythonhosted.org/packages/e5/63/cd7d615331b328e287d8233ba9fdf191a9c2d11b6af0c7a59cfcec23de68/pandas-2.3.3-cp312-cp312-manylinux_2_24_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:b3d11d2fda7eb164ef27ffc14b4fcab16a80e1ce67e9f57e19ec0afaf715ba89", size = 12362693 }, - { url = "https://files.pythonhosted.org/packages/a6/de/8b1895b107277d52f2b42d3a6806e69cfef0d5cf1d0ba343470b9d8e0a04/pandas-2.3.3-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:a68e15f780eddf2b07d242e17a04aa187a7ee12b40b930bfdd78070556550e98", size = 12771002 }, - { url = "https://files.pythonhosted.org/packages/87/21/84072af3187a677c5893b170ba2c8fbe450a6ff911234916da889b698220/pandas-2.3.3-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:371a4ab48e950033bcf52b6527eccb564f52dc826c02afd9a1bc0ab731bba084", size = 13450971 }, - { url = "https://files.pythonhosted.org/packages/86/41/585a168330ff063014880a80d744219dbf1dd7a1c706e75ab3425a987384/pandas-2.3.3-cp312-cp312-win_amd64.whl", hash = "sha256:a16dcec078a01eeef8ee61bf64074b4e524a2a3f4b3be9326420cabe59c4778b", size = 10992722 }, - { url = "https://files.pythonhosted.org/packages/cd/4b/18b035ee18f97c1040d94debd8f2e737000ad70ccc8f5513f4eefad75f4b/pandas-2.3.3-cp313-cp313-macosx_10_13_x86_64.whl", hash = "sha256:56851a737e3470de7fa88e6131f41281ed440d29a9268dcbf0002da5ac366713", size = 11544671 }, - { url = "https://files.pythonhosted.org/packages/31/94/72fac03573102779920099bcac1c3b05975c2cb5f01eac609faf34bed1ca/pandas-2.3.3-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:bdcd9d1167f4885211e401b3036c0c8d9e274eee67ea8d0758a256d60704cfe8", size = 10680807 }, - { url = "https://files.pythonhosted.org/packages/16/87/9472cf4a487d848476865321de18cc8c920b8cab98453ab79dbbc98db63a/pandas-2.3.3-cp313-cp313-manylinux_2_24_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:e32e7cc9af0f1cc15548288a51a3b681cc2a219faa838e995f7dc53dbab1062d", size = 11709872 }, - { url = "https://files.pythonhosted.org/packages/15/07/284f757f63f8a8d69ed4472bfd85122bd086e637bf4ed09de572d575a693/pandas-2.3.3-cp313-cp313-manylinux_2_24_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:318d77e0e42a628c04dc56bcef4b40de67918f7041c2b061af1da41dcff670ac", size = 12306371 }, - { url = "https://files.pythonhosted.org/packages/33/81/a3afc88fca4aa925804a27d2676d22dcd2031c2ebe08aabd0ae55b9ff282/pandas-2.3.3-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:4e0a175408804d566144e170d0476b15d78458795bb18f1304fb94160cabf40c", size = 12765333 }, - { url = "https://files.pythonhosted.org/packages/8d/0f/b4d4ae743a83742f1153464cf1a8ecfafc3ac59722a0b5c8602310cb7158/pandas-2.3.3-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:93c2d9ab0fc11822b5eece72ec9587e172f63cff87c00b062f6e37448ced4493", size = 13418120 }, - { url = "https://files.pythonhosted.org/packages/4f/c7/e54682c96a895d0c808453269e0b5928a07a127a15704fedb643e9b0a4c8/pandas-2.3.3-cp313-cp313-win_amd64.whl", hash = "sha256:f8bfc0e12dc78f777f323f55c58649591b2cd0c43534e8355c51d3fede5f4dee", size = 10993991 }, - { url = "https://files.pythonhosted.org/packages/f9/ca/3f8d4f49740799189e1395812f3bf23b5e8fc7c190827d55a610da72ce55/pandas-2.3.3-cp313-cp313t-macosx_10_13_x86_64.whl", hash = "sha256:75ea25f9529fdec2d2e93a42c523962261e567d250b0013b16210e1d40d7c2e5", size = 12048227 }, - { url = "https://files.pythonhosted.org/packages/0e/5a/f43efec3e8c0cc92c4663ccad372dbdff72b60bdb56b2749f04aa1d07d7e/pandas-2.3.3-cp313-cp313t-macosx_11_0_arm64.whl", hash = "sha256:74ecdf1d301e812db96a465a525952f4dde225fdb6d8e5a521d47e1f42041e21", size = 11411056 }, - { url = "https://files.pythonhosted.org/packages/46/b1/85331edfc591208c9d1a63a06baa67b21d332e63b7a591a5ba42a10bb507/pandas-2.3.3-cp313-cp313t-manylinux_2_24_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:6435cb949cb34ec11cc9860246ccb2fdc9ecd742c12d3304989017d53f039a78", size = 11645189 }, - { url = "https://files.pythonhosted.org/packages/44/23/78d645adc35d94d1ac4f2a3c4112ab6f5b8999f4898b8cdf01252f8df4a9/pandas-2.3.3-cp313-cp313t-manylinux_2_24_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:900f47d8f20860de523a1ac881c4c36d65efcb2eb850e6948140fa781736e110", size = 12121912 }, - { url = "https://files.pythonhosted.org/packages/53/da/d10013df5e6aaef6b425aa0c32e1fc1f3e431e4bcabd420517dceadce354/pandas-2.3.3-cp313-cp313t-musllinux_1_2_aarch64.whl", hash = "sha256:a45c765238e2ed7d7c608fc5bc4a6f88b642f2f01e70c0c23d2224dd21829d86", size = 12712160 }, - { url = "https://files.pythonhosted.org/packages/bd/17/e756653095a083d8a37cbd816cb87148debcfcd920129b25f99dd8d04271/pandas-2.3.3-cp313-cp313t-musllinux_1_2_x86_64.whl", hash = "sha256:c4fc4c21971a1a9f4bdb4c73978c7f7256caa3e62b323f70d6cb80db583350bc", size = 13199233 }, - { url = "https://files.pythonhosted.org/packages/04/fd/74903979833db8390b73b3a8a7d30d146d710bd32703724dd9083950386f/pandas-2.3.3-cp314-cp314-macosx_10_13_x86_64.whl", hash = "sha256:ee15f284898e7b246df8087fc82b87b01686f98ee67d85a17b7ab44143a3a9a0", size = 11540635 }, - { url = "https://files.pythonhosted.org/packages/21/00/266d6b357ad5e6d3ad55093a7e8efc7dd245f5a842b584db9f30b0f0a287/pandas-2.3.3-cp314-cp314-macosx_11_0_arm64.whl", hash = "sha256:1611aedd912e1ff81ff41c745822980c49ce4a7907537be8692c8dbc31924593", size = 10759079 }, - { url = "https://files.pythonhosted.org/packages/ca/05/d01ef80a7a3a12b2f8bbf16daba1e17c98a2f039cbc8e2f77a2c5a63d382/pandas-2.3.3-cp314-cp314-manylinux_2_24_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:6d2cefc361461662ac48810cb14365a365ce864afe85ef1f447ff5a1e99ea81c", size = 11814049 }, - { url = "https://files.pythonhosted.org/packages/15/b2/0e62f78c0c5ba7e3d2c5945a82456f4fac76c480940f805e0b97fcbc2f65/pandas-2.3.3-cp314-cp314-manylinux_2_24_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:ee67acbbf05014ea6c763beb097e03cd629961c8a632075eeb34247120abcb4b", size = 12332638 }, - { url = "https://files.pythonhosted.org/packages/c5/33/dd70400631b62b9b29c3c93d2feee1d0964dc2bae2e5ad7a6c73a7f25325/pandas-2.3.3-cp314-cp314-musllinux_1_2_aarch64.whl", hash = "sha256:c46467899aaa4da076d5abc11084634e2d197e9460643dd455ac3db5856b24d6", size = 12886834 }, - { url = "https://files.pythonhosted.org/packages/d3/18/b5d48f55821228d0d2692b34fd5034bb185e854bdb592e9c640f6290e012/pandas-2.3.3-cp314-cp314-musllinux_1_2_x86_64.whl", hash = "sha256:6253c72c6a1d990a410bc7de641d34053364ef8bcd3126f7e7450125887dffe3", size = 13409925 }, - { url = "https://files.pythonhosted.org/packages/a6/3d/124ac75fcd0ecc09b8fdccb0246ef65e35b012030defb0e0eba2cbbbe948/pandas-2.3.3-cp314-cp314-win_amd64.whl", hash = "sha256:1b07204a219b3b7350abaae088f451860223a52cfb8a6c53358e7948735158e5", size = 11109071 }, - { url = "https://files.pythonhosted.org/packages/89/9c/0e21c895c38a157e0faa1fb64587a9226d6dd46452cac4532d80c3c4a244/pandas-2.3.3-cp314-cp314t-macosx_10_13_x86_64.whl", hash = "sha256:2462b1a365b6109d275250baaae7b760fd25c726aaca0054649286bcfbb3e8ec", size = 12048504 }, - { url = "https://files.pythonhosted.org/packages/d7/82/b69a1c95df796858777b68fbe6a81d37443a33319761d7c652ce77797475/pandas-2.3.3-cp314-cp314t-macosx_11_0_arm64.whl", hash = "sha256:0242fe9a49aa8b4d78a4fa03acb397a58833ef6199e9aa40a95f027bb3a1b6e7", size = 11410702 }, - { url = "https://files.pythonhosted.org/packages/f9/88/702bde3ba0a94b8c73a0181e05144b10f13f29ebfc2150c3a79062a8195d/pandas-2.3.3-cp314-cp314t-manylinux_2_24_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:a21d830e78df0a515db2b3d2f5570610f5e6bd2e27749770e8bb7b524b89b450", size = 11634535 }, - { url = "https://files.pythonhosted.org/packages/a4/1e/1bac1a839d12e6a82ec6cb40cda2edde64a2013a66963293696bbf31fbbb/pandas-2.3.3-cp314-cp314t-manylinux_2_24_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:2e3ebdb170b5ef78f19bfb71b0dc5dc58775032361fa188e814959b74d726dd5", size = 12121582 }, - { url = "https://files.pythonhosted.org/packages/44/91/483de934193e12a3b1d6ae7c8645d083ff88dec75f46e827562f1e4b4da6/pandas-2.3.3-cp314-cp314t-musllinux_1_2_aarch64.whl", hash = "sha256:d051c0e065b94b7a3cea50eb1ec32e912cd96dba41647eb24104b6c6c14c5788", size = 12699963 }, - { url = "https://files.pythonhosted.org/packages/70/44/5191d2e4026f86a2a109053e194d3ba7a31a2d10a9c2348368c63ed4e85a/pandas-2.3.3-cp314-cp314t-musllinux_1_2_x86_64.whl", hash = "sha256:3869faf4bd07b3b66a9f462417d0ca3a9df29a9f6abd5d0d0dbab15dac7abe87", size = 13202175 }, -] - -[[package]] -name = "parakeet-tdt-ctc-110m-coreml" -version = "1.0.0" -source = { editable = "." } -dependencies = [ - { name = "coremltools" }, - { name = "numpy" }, - { name = "scipy", version = "1.15.3", source = { registry = "https://pypi.org/simple" }, marker = "python_full_version < '3.11'" }, - { name = "scipy", version = "1.16.3", source = { registry = "https://pypi.org/simple" }, marker = "python_full_version >= '3.11'" }, -] - -[package.optional-dependencies] -audio = [ - { name = "librosa" }, -] -convert = [ - { name = "nemo-toolkit", extra = ["asr"] }, - { name = "torch" }, -] -dev = [ - { name = "pytest" }, - { name = "ruff" }, -] - -[package.metadata] -requires-dist = [ - { name = "coremltools", specifier = ">=7.0" }, - { name = "librosa", marker = "extra == 'audio'", specifier = ">=0.10.0" }, - { name = "nemo-toolkit", extras = ["asr"], marker = "extra == 'convert'", specifier = ">=1.20.0" }, - { name = "numpy", specifier = ">=1.24.0" }, - { name = "pytest", marker = "extra == 'dev'", specifier = ">=7.0.0" }, - { name = "ruff", marker = "extra == 'dev'", specifier = ">=0.1.0" }, - { name = "scipy", specifier = ">=1.10.0" }, - { name = "torch", marker = "extra == 'convert'", specifier = ">=2.0.0" }, -] - -[[package]] -name = "parso" -version = "0.8.5" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/d4/de/53e0bcf53d13e005bd8c92e7855142494f41171b34c2536b86187474184d/parso-0.8.5.tar.gz", hash = "sha256:034d7354a9a018bdce352f48b2a8a450f05e9d6ee85db84764e9b6bd96dafe5a", size = 401205 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/16/32/f8e3c85d1d5250232a5d3477a2a28cc291968ff175caeadaf3cc19ce0e4a/parso-0.8.5-py2.py3-none-any.whl", hash = "sha256:646204b5ee239c396d040b90f9e272e9a8017c630092bf59980beb62fd033887", size = 106668 }, -] - -[[package]] -name = "peft" -version = "0.18.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "accelerate" }, - { name = "huggingface-hub" }, - { name = "numpy" }, - { name = "packaging" }, - { name = "psutil" }, - { name = "pyyaml" }, - { name = "safetensors" }, - { name = "torch" }, - { name = "tqdm" }, - { name = "transformers" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/4b/0c/f2938db546ac7fc961ab5917cd50fcf5d0d70b406de93e3faccaa504e152/peft-0.18.0.tar.gz", hash = "sha256:c81c80b2056ab40c23d58ef25f74daab417ac653970718589a11a8af28218588", size = 634141 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/0f/55/481bf25613d40ef53534f664deba7b138fe566356b6ca10304e2b3b2529c/peft-0.18.0-py3-none-any.whl", hash = "sha256:624f69ca6393b765ccc6734adda7ca57d80b238f0900a42c357d8b67a03d62ff", size = 556427 }, -] - -[[package]] -name = "pexpect" -version = "4.9.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "ptyprocess" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/42/92/cc564bf6381ff43ce1f4d06852fc19a2f11d180f23dc32d9588bee2f149d/pexpect-4.9.0.tar.gz", hash = "sha256:ee7d41123f3c9911050ea2c2dac107568dc43b2d3b0c7557a33212c398ead30f", size = 166450 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/9e/c3/059298687310d527a58bb01f3b1965787ee3b40dce76752eda8b44e9a2c5/pexpect-4.9.0-py2.py3-none-any.whl", hash = "sha256:7236d1e080e4936be2dc3e326cec0af72acf9212a7e1d060210e70a47e253523", size = 63772 }, -] - -[[package]] -name = "pillow" -version = "12.0.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/5a/b0/cace85a1b0c9775a9f8f5d5423c8261c858760e2466c79b2dd184638b056/pillow-12.0.0.tar.gz", hash = "sha256:87d4f8125c9988bfbed67af47dd7a953e2fc7b0cc1e7800ec6d2080d490bb353", size = 47008828 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/5d/08/26e68b6b5da219c2a2cb7b563af008b53bb8e6b6fcb3fa40715fcdb2523a/pillow-12.0.0-cp310-cp310-macosx_10_10_x86_64.whl", hash = "sha256:3adfb466bbc544b926d50fe8f4a4e6abd8c6bffd28a26177594e6e9b2b76572b", size = 5289809 }, - { url = "https://files.pythonhosted.org/packages/cb/e9/4e58fb097fb74c7b4758a680aacd558810a417d1edaa7000142976ef9d2f/pillow-12.0.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:1ac11e8ea4f611c3c0147424eae514028b5e9077dd99ab91e1bd7bc33ff145e1", size = 4650606 }, - { url = "https://files.pythonhosted.org/packages/4b/e0/1fa492aa9f77b3bc6d471c468e62bfea1823056bf7e5e4f1914d7ab2565e/pillow-12.0.0-cp310-cp310-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:d49e2314c373f4c2b39446fb1a45ed333c850e09d0c59ac79b72eb3b95397363", size = 6221023 }, - { url = "https://files.pythonhosted.org/packages/c1/09/4de7cd03e33734ccd0c876f0251401f1314e819cbfd89a0fcb6e77927cc6/pillow-12.0.0-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:c7b2a63fd6d5246349f3d3f37b14430d73ee7e8173154461785e43036ffa96ca", size = 8024937 }, - { url = "https://files.pythonhosted.org/packages/2e/69/0688e7c1390666592876d9d474f5e135abb4acb39dcb583c4dc5490f1aff/pillow-12.0.0-cp310-cp310-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:d64317d2587c70324b79861babb9c09f71fbb780bad212018874b2c013d8600e", size = 6334139 }, - { url = "https://files.pythonhosted.org/packages/ed/1c/880921e98f525b9b44ce747ad1ea8f73fd7e992bafe3ca5e5644bf433dea/pillow-12.0.0-cp310-cp310-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:d77153e14b709fd8b8af6f66a3afbb9ed6e9fc5ccf0b6b7e1ced7b036a228782", size = 7026074 }, - { url = "https://files.pythonhosted.org/packages/28/03/96f718331b19b355610ef4ebdbbde3557c726513030665071fd025745671/pillow-12.0.0-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:32ed80ea8a90ee3e6fa08c21e2e091bba6eda8eccc83dbc34c95169507a91f10", size = 6448852 }, - { url = "https://files.pythonhosted.org/packages/3a/a0/6a193b3f0cc9437b122978d2c5cbce59510ccf9a5b48825096ed7472da2f/pillow-12.0.0-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:c828a1ae702fc712978bda0320ba1b9893d99be0badf2647f693cc01cf0f04fa", size = 7117058 }, - { url = "https://files.pythonhosted.org/packages/a7/c4/043192375eaa4463254e8e61f0e2ec9a846b983929a8d0a7122e0a6d6fff/pillow-12.0.0-cp310-cp310-win32.whl", hash = "sha256:bd87e140e45399c818fac4247880b9ce719e4783d767e030a883a970be632275", size = 6295431 }, - { url = "https://files.pythonhosted.org/packages/92/c6/c2f2fc7e56301c21827e689bb8b0b465f1b52878b57471a070678c0c33cd/pillow-12.0.0-cp310-cp310-win_amd64.whl", hash = "sha256:455247ac8a4cfb7b9bc45b7e432d10421aea9fc2e74d285ba4072688a74c2e9d", size = 7000412 }, - { url = "https://files.pythonhosted.org/packages/b2/d2/5f675067ba82da7a1c238a73b32e3fd78d67f9d9f80fbadd33a40b9c0481/pillow-12.0.0-cp310-cp310-win_arm64.whl", hash = "sha256:6ace95230bfb7cd79ef66caa064bbe2f2a1e63d93471c3a2e1f1348d9f22d6b7", size = 2435903 }, - { url = "https://files.pythonhosted.org/packages/0e/5a/a2f6773b64edb921a756eb0729068acad9fc5208a53f4a349396e9436721/pillow-12.0.0-cp311-cp311-macosx_10_10_x86_64.whl", hash = "sha256:0fd00cac9c03256c8b2ff58f162ebcd2587ad3e1f2e397eab718c47e24d231cc", size = 5289798 }, - { url = "https://files.pythonhosted.org/packages/2e/05/069b1f8a2e4b5a37493da6c5868531c3f77b85e716ad7a590ef87d58730d/pillow-12.0.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:a3475b96f5908b3b16c47533daaa87380c491357d197564e0ba34ae75c0f3257", size = 4650589 }, - { url = "https://files.pythonhosted.org/packages/61/e3/2c820d6e9a36432503ead175ae294f96861b07600a7156154a086ba7111a/pillow-12.0.0-cp311-cp311-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:110486b79f2d112cf6add83b28b627e369219388f64ef2f960fef9ebaf54c642", size = 6230472 }, - { url = "https://files.pythonhosted.org/packages/4f/89/63427f51c64209c5e23d4d52071c8d0f21024d3a8a487737caaf614a5795/pillow-12.0.0-cp311-cp311-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:5269cc1caeedb67e6f7269a42014f381f45e2e7cd42d834ede3c703a1d915fe3", size = 8033887 }, - { url = "https://files.pythonhosted.org/packages/f6/1b/c9711318d4901093c15840f268ad649459cd81984c9ec9887756cca049a5/pillow-12.0.0-cp311-cp311-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:aa5129de4e174daccbc59d0a3b6d20eaf24417d59851c07ebb37aeb02947987c", size = 6343964 }, - { url = "https://files.pythonhosted.org/packages/41/1e/db9470f2d030b4995083044cd8738cdd1bf773106819f6d8ba12597d5352/pillow-12.0.0-cp311-cp311-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:bee2a6db3a7242ea309aa7ee8e2780726fed67ff4e5b40169f2c940e7eb09227", size = 7034756 }, - { url = "https://files.pythonhosted.org/packages/cc/b0/6177a8bdd5ee4ed87cba2de5a3cc1db55ffbbec6176784ce5bb75aa96798/pillow-12.0.0-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:90387104ee8400a7b4598253b4c406f8958f59fcf983a6cea2b50d59f7d63d0b", size = 6458075 }, - { url = "https://files.pythonhosted.org/packages/bc/5e/61537aa6fa977922c6a03253a0e727e6e4a72381a80d63ad8eec350684f2/pillow-12.0.0-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:bc91a56697869546d1b8f0a3ff35224557ae7f881050e99f615e0119bf934b4e", size = 7125955 }, - { url = "https://files.pythonhosted.org/packages/1f/3d/d5033539344ee3cbd9a4d69e12e63ca3a44a739eb2d4c8da350a3d38edd7/pillow-12.0.0-cp311-cp311-win32.whl", hash = "sha256:27f95b12453d165099c84f8a8bfdfd46b9e4bda9e0e4b65f0635430027f55739", size = 6298440 }, - { url = "https://files.pythonhosted.org/packages/4d/42/aaca386de5cc8bd8a0254516957c1f265e3521c91515b16e286c662854c4/pillow-12.0.0-cp311-cp311-win_amd64.whl", hash = "sha256:b583dc9070312190192631373c6c8ed277254aa6e6084b74bdd0a6d3b221608e", size = 6999256 }, - { url = "https://files.pythonhosted.org/packages/ba/f1/9197c9c2d5708b785f631a6dfbfa8eb3fb9672837cb92ae9af812c13b4ed/pillow-12.0.0-cp311-cp311-win_arm64.whl", hash = "sha256:759de84a33be3b178a64c8ba28ad5c135900359e85fb662bc6e403ad4407791d", size = 2436025 }, - { url = "https://files.pythonhosted.org/packages/2c/90/4fcce2c22caf044e660a198d740e7fbc14395619e3cb1abad12192c0826c/pillow-12.0.0-cp312-cp312-macosx_10_13_x86_64.whl", hash = "sha256:53561a4ddc36facb432fae7a9d8afbfaf94795414f5cdc5fc52f28c1dca90371", size = 5249377 }, - { url = "https://files.pythonhosted.org/packages/fd/e0/ed960067543d080691d47d6938ebccbf3976a931c9567ab2fbfab983a5dd/pillow-12.0.0-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:71db6b4c1653045dacc1585c1b0d184004f0d7e694c7b34ac165ca70c0838082", size = 4650343 }, - { url = "https://files.pythonhosted.org/packages/e7/a1/f81fdeddcb99c044bf7d6faa47e12850f13cee0849537a7d27eeab5534d4/pillow-12.0.0-cp312-cp312-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:2fa5f0b6716fc88f11380b88b31fe591a06c6315e955c096c35715788b339e3f", size = 6232981 }, - { url = "https://files.pythonhosted.org/packages/88/e1/9098d3ce341a8750b55b0e00c03f1630d6178f38ac191c81c97a3b047b44/pillow-12.0.0-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:82240051c6ca513c616f7f9da06e871f61bfd7805f566275841af15015b8f98d", size = 8041399 }, - { url = "https://files.pythonhosted.org/packages/a7/62/a22e8d3b602ae8cc01446d0c57a54e982737f44b6f2e1e019a925143771d/pillow-12.0.0-cp312-cp312-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:55f818bd74fe2f11d4d7cbc65880a843c4075e0ac7226bc1a23261dbea531953", size = 6347740 }, - { url = "https://files.pythonhosted.org/packages/4f/87/424511bdcd02c8d7acf9f65caa09f291a519b16bd83c3fb3374b3d4ae951/pillow-12.0.0-cp312-cp312-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:b87843e225e74576437fd5b6a4c2205d422754f84a06942cfaf1dc32243e45a8", size = 7040201 }, - { url = "https://files.pythonhosted.org/packages/dc/4d/435c8ac688c54d11755aedfdd9f29c9eeddf68d150fe42d1d3dbd2365149/pillow-12.0.0-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:c607c90ba67533e1b2355b821fef6764d1dd2cbe26b8c1005ae84f7aea25ff79", size = 6462334 }, - { url = "https://files.pythonhosted.org/packages/2b/f2/ad34167a8059a59b8ad10bc5c72d4d9b35acc6b7c0877af8ac885b5f2044/pillow-12.0.0-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:21f241bdd5080a15bc86d3466a9f6074a9c2c2b314100dd896ac81ee6db2f1ba", size = 7134162 }, - { url = "https://files.pythonhosted.org/packages/0c/b1/a7391df6adacf0a5c2cf6ac1cf1fcc1369e7d439d28f637a847f8803beb3/pillow-12.0.0-cp312-cp312-win32.whl", hash = "sha256:dd333073e0cacdc3089525c7df7d39b211bcdf31fc2824e49d01c6b6187b07d0", size = 6298769 }, - { url = "https://files.pythonhosted.org/packages/a2/0b/d87733741526541c909bbf159e338dcace4f982daac6e5a8d6be225ca32d/pillow-12.0.0-cp312-cp312-win_amd64.whl", hash = "sha256:9fe611163f6303d1619bbcb653540a4d60f9e55e622d60a3108be0d5b441017a", size = 7001107 }, - { url = "https://files.pythonhosted.org/packages/bc/96/aaa61ce33cc98421fb6088af2a03be4157b1e7e0e87087c888e2370a7f45/pillow-12.0.0-cp312-cp312-win_arm64.whl", hash = "sha256:7dfb439562f234f7d57b1ac6bc8fe7f838a4bd49c79230e0f6a1da93e82f1fad", size = 2436012 }, - { url = "https://files.pythonhosted.org/packages/62/f2/de993bb2d21b33a98d031ecf6a978e4b61da207bef02f7b43093774c480d/pillow-12.0.0-cp313-cp313-ios_13_0_arm64_iphoneos.whl", hash = "sha256:0869154a2d0546545cde61d1789a6524319fc1897d9ee31218eae7a60ccc5643", size = 4045493 }, - { url = "https://files.pythonhosted.org/packages/0e/b6/bc8d0c4c9f6f111a783d045310945deb769b806d7574764234ffd50bc5ea/pillow-12.0.0-cp313-cp313-ios_13_0_arm64_iphonesimulator.whl", hash = "sha256:a7921c5a6d31b3d756ec980f2f47c0cfdbce0fc48c22a39347a895f41f4a6ea4", size = 4120461 }, - { url = "https://files.pythonhosted.org/packages/5d/57/d60d343709366a353dc56adb4ee1e7d8a2cc34e3fbc22905f4167cfec119/pillow-12.0.0-cp313-cp313-ios_13_0_x86_64_iphonesimulator.whl", hash = "sha256:1ee80a59f6ce048ae13cda1abf7fbd2a34ab9ee7d401c46be3ca685d1999a399", size = 3576912 }, - { url = "https://files.pythonhosted.org/packages/a4/a4/a0a31467e3f83b94d37568294b01d22b43ae3c5d85f2811769b9c66389dd/pillow-12.0.0-cp313-cp313-macosx_10_13_x86_64.whl", hash = "sha256:c50f36a62a22d350c96e49ad02d0da41dbd17ddc2e29750dbdba4323f85eb4a5", size = 5249132 }, - { url = "https://files.pythonhosted.org/packages/83/06/48eab21dd561de2914242711434c0c0eb992ed08ff3f6107a5f44527f5e9/pillow-12.0.0-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:5193fde9a5f23c331ea26d0cf171fbf67e3f247585f50c08b3e205c7aeb4589b", size = 4650099 }, - { url = "https://files.pythonhosted.org/packages/fc/bd/69ed99fd46a8dba7c1887156d3572fe4484e3f031405fcc5a92e31c04035/pillow-12.0.0-cp313-cp313-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:bde737cff1a975b70652b62d626f7785e0480918dece11e8fef3c0cf057351c3", size = 6230808 }, - { url = "https://files.pythonhosted.org/packages/ea/94/8fad659bcdbf86ed70099cb60ae40be6acca434bbc8c4c0d4ef356d7e0de/pillow-12.0.0-cp313-cp313-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:a6597ff2b61d121172f5844b53f21467f7082f5fb385a9a29c01414463f93b07", size = 8037804 }, - { url = "https://files.pythonhosted.org/packages/20/39/c685d05c06deecfd4e2d1950e9a908aa2ca8bc4e6c3b12d93b9cafbd7837/pillow-12.0.0-cp313-cp313-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:0b817e7035ea7f6b942c13aa03bb554fc44fea70838ea21f8eb31c638326584e", size = 6345553 }, - { url = "https://files.pythonhosted.org/packages/38/57/755dbd06530a27a5ed74f8cb0a7a44a21722ebf318edbe67ddbd7fb28f88/pillow-12.0.0-cp313-cp313-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:f4f1231b7dec408e8670264ce63e9c71409d9583dd21d32c163e25213ee2a344", size = 7037729 }, - { url = "https://files.pythonhosted.org/packages/ca/b6/7e94f4c41d238615674d06ed677c14883103dce1c52e4af16f000338cfd7/pillow-12.0.0-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:6e51b71417049ad6ab14c49608b4a24d8fb3fe605e5dfabfe523b58064dc3d27", size = 6459789 }, - { url = "https://files.pythonhosted.org/packages/9c/14/4448bb0b5e0f22dd865290536d20ec8a23b64e2d04280b89139f09a36bb6/pillow-12.0.0-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:d120c38a42c234dc9a8c5de7ceaaf899cf33561956acb4941653f8bdc657aa79", size = 7130917 }, - { url = "https://files.pythonhosted.org/packages/dd/ca/16c6926cc1c015845745d5c16c9358e24282f1e588237a4c36d2b30f182f/pillow-12.0.0-cp313-cp313-win32.whl", hash = "sha256:4cc6b3b2efff105c6a1656cfe59da4fdde2cda9af1c5e0b58529b24525d0a098", size = 6302391 }, - { url = "https://files.pythonhosted.org/packages/6d/2a/dd43dcfd6dae9b6a49ee28a8eedb98c7d5ff2de94a5d834565164667b97b/pillow-12.0.0-cp313-cp313-win_amd64.whl", hash = "sha256:4cf7fed4b4580601c4345ceb5d4cbf5a980d030fd5ad07c4d2ec589f95f09905", size = 7007477 }, - { url = "https://files.pythonhosted.org/packages/77/f0/72ea067f4b5ae5ead653053212af05ce3705807906ba3f3e8f58ddf617e6/pillow-12.0.0-cp313-cp313-win_arm64.whl", hash = "sha256:9f0b04c6b8584c2c193babcccc908b38ed29524b29dd464bc8801bf10d746a3a", size = 2435918 }, - { url = "https://files.pythonhosted.org/packages/f5/5e/9046b423735c21f0487ea6cb5b10f89ea8f8dfbe32576fe052b5ba9d4e5b/pillow-12.0.0-cp313-cp313t-macosx_10_13_x86_64.whl", hash = "sha256:7fa22993bac7b77b78cae22bad1e2a987ddf0d9015c63358032f84a53f23cdc3", size = 5251406 }, - { url = "https://files.pythonhosted.org/packages/12/66/982ceebcdb13c97270ef7a56c3969635b4ee7cd45227fa707c94719229c5/pillow-12.0.0-cp313-cp313t-macosx_11_0_arm64.whl", hash = "sha256:f135c702ac42262573fe9714dfe99c944b4ba307af5eb507abef1667e2cbbced", size = 4653218 }, - { url = "https://files.pythonhosted.org/packages/16/b3/81e625524688c31859450119bf12674619429cab3119eec0e30a7a1029cb/pillow-12.0.0-cp313-cp313t-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:c85de1136429c524e55cfa4e033b4a7940ac5c8ee4d9401cc2d1bf48154bbc7b", size = 6266564 }, - { url = "https://files.pythonhosted.org/packages/98/59/dfb38f2a41240d2408096e1a76c671d0a105a4a8471b1871c6902719450c/pillow-12.0.0-cp313-cp313t-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:38df9b4bfd3db902c9c2bd369bcacaf9d935b2fff73709429d95cc41554f7b3d", size = 8069260 }, - { url = "https://files.pythonhosted.org/packages/dc/3d/378dbea5cd1874b94c312425ca77b0f47776c78e0df2df751b820c8c1d6c/pillow-12.0.0-cp313-cp313t-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:7d87ef5795da03d742bf49439f9ca4d027cde49c82c5371ba52464aee266699a", size = 6379248 }, - { url = "https://files.pythonhosted.org/packages/84/b0/d525ef47d71590f1621510327acec75ae58c721dc071b17d8d652ca494d8/pillow-12.0.0-cp313-cp313t-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:aff9e4d82d082ff9513bdd6acd4f5bd359f5b2c870907d2b0a9c5e10d40c88fe", size = 7066043 }, - { url = "https://files.pythonhosted.org/packages/61/2c/aced60e9cf9d0cde341d54bf7932c9ffc33ddb4a1595798b3a5150c7ec4e/pillow-12.0.0-cp313-cp313t-musllinux_1_2_aarch64.whl", hash = "sha256:8d8ca2b210ada074d57fcee40c30446c9562e542fc46aedc19baf758a93532ee", size = 6490915 }, - { url = "https://files.pythonhosted.org/packages/ef/26/69dcb9b91f4e59f8f34b2332a4a0a951b44f547c4ed39d3e4dcfcff48f89/pillow-12.0.0-cp313-cp313t-musllinux_1_2_x86_64.whl", hash = "sha256:99a7f72fb6249302aa62245680754862a44179b545ded638cf1fef59befb57ef", size = 7157998 }, - { url = "https://files.pythonhosted.org/packages/61/2b/726235842220ca95fa441ddf55dd2382b52ab5b8d9c0596fe6b3f23dafe8/pillow-12.0.0-cp313-cp313t-win32.whl", hash = "sha256:4078242472387600b2ce8d93ade8899c12bf33fa89e55ec89fe126e9d6d5d9e9", size = 6306201 }, - { url = "https://files.pythonhosted.org/packages/c0/3d/2afaf4e840b2df71344ababf2f8edd75a705ce500e5dc1e7227808312ae1/pillow-12.0.0-cp313-cp313t-win_amd64.whl", hash = "sha256:2c54c1a783d6d60595d3514f0efe9b37c8808746a66920315bfd34a938d7994b", size = 7013165 }, - { url = "https://files.pythonhosted.org/packages/6f/75/3fa09aa5cf6ed04bee3fa575798ddf1ce0bace8edb47249c798077a81f7f/pillow-12.0.0-cp313-cp313t-win_arm64.whl", hash = "sha256:26d9f7d2b604cd23aba3e9faf795787456ac25634d82cd060556998e39c6fa47", size = 2437834 }, - { url = "https://files.pythonhosted.org/packages/54/2a/9a8c6ba2c2c07b71bec92cf63e03370ca5e5f5c5b119b742bcc0cde3f9c5/pillow-12.0.0-cp314-cp314-ios_13_0_arm64_iphoneos.whl", hash = "sha256:beeae3f27f62308f1ddbcfb0690bf44b10732f2ef43758f169d5e9303165d3f9", size = 4045531 }, - { url = "https://files.pythonhosted.org/packages/84/54/836fdbf1bfb3d66a59f0189ff0b9f5f666cee09c6188309300df04ad71fa/pillow-12.0.0-cp314-cp314-ios_13_0_arm64_iphonesimulator.whl", hash = "sha256:d4827615da15cd59784ce39d3388275ec093ae3ee8d7f0c089b76fa87af756c2", size = 4120554 }, - { url = "https://files.pythonhosted.org/packages/0d/cd/16aec9f0da4793e98e6b54778a5fbce4f375c6646fe662e80600b8797379/pillow-12.0.0-cp314-cp314-ios_13_0_x86_64_iphonesimulator.whl", hash = "sha256:3e42edad50b6909089750e65c91aa09aaf1e0a71310d383f11321b27c224ed8a", size = 3576812 }, - { url = "https://files.pythonhosted.org/packages/f6/b7/13957fda356dc46339298b351cae0d327704986337c3c69bb54628c88155/pillow-12.0.0-cp314-cp314-macosx_10_15_x86_64.whl", hash = "sha256:e5d8efac84c9afcb40914ab49ba063d94f5dbdf5066db4482c66a992f47a3a3b", size = 5252689 }, - { url = "https://files.pythonhosted.org/packages/fc/f5/eae31a306341d8f331f43edb2e9122c7661b975433de5e447939ae61c5da/pillow-12.0.0-cp314-cp314-macosx_11_0_arm64.whl", hash = "sha256:266cd5f2b63ff316d5a1bba46268e603c9caf5606d44f38c2873c380950576ad", size = 4650186 }, - { url = "https://files.pythonhosted.org/packages/86/62/2a88339aa40c4c77e79108facbd307d6091e2c0eb5b8d3cf4977cfca2fe6/pillow-12.0.0-cp314-cp314-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:58eea5ebe51504057dd95c5b77d21700b77615ab0243d8152793dc00eb4faf01", size = 6230308 }, - { url = "https://files.pythonhosted.org/packages/c7/33/5425a8992bcb32d1cb9fa3dd39a89e613d09a22f2c8083b7bf43c455f760/pillow-12.0.0-cp314-cp314-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:f13711b1a5ba512d647a0e4ba79280d3a9a045aaf7e0cc6fbe96b91d4cdf6b0c", size = 8039222 }, - { url = "https://files.pythonhosted.org/packages/d8/61/3f5d3b35c5728f37953d3eec5b5f3e77111949523bd2dd7f31a851e50690/pillow-12.0.0-cp314-cp314-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:6846bd2d116ff42cba6b646edf5bf61d37e5cbd256425fa089fee4ff5c07a99e", size = 6346657 }, - { url = "https://files.pythonhosted.org/packages/3a/be/ee90a3d79271227e0f0a33c453531efd6ed14b2e708596ba5dd9be948da3/pillow-12.0.0-cp314-cp314-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:c98fa880d695de164b4135a52fd2e9cd7b7c90a9d8ac5e9e443a24a95ef9248e", size = 7038482 }, - { url = "https://files.pythonhosted.org/packages/44/34/a16b6a4d1ad727de390e9bd9f19f5f669e079e5826ec0f329010ddea492f/pillow-12.0.0-cp314-cp314-musllinux_1_2_aarch64.whl", hash = "sha256:fa3ed2a29a9e9d2d488b4da81dcb54720ac3104a20bf0bd273f1e4648aff5af9", size = 6461416 }, - { url = "https://files.pythonhosted.org/packages/b6/39/1aa5850d2ade7d7ba9f54e4e4c17077244ff7a2d9e25998c38a29749eb3f/pillow-12.0.0-cp314-cp314-musllinux_1_2_x86_64.whl", hash = "sha256:d034140032870024e6b9892c692fe2968493790dd57208b2c37e3fb35f6df3ab", size = 7131584 }, - { url = "https://files.pythonhosted.org/packages/bf/db/4fae862f8fad0167073a7733973bfa955f47e2cac3dc3e3e6257d10fab4a/pillow-12.0.0-cp314-cp314-win32.whl", hash = "sha256:1b1b133e6e16105f524a8dec491e0586d072948ce15c9b914e41cdadd209052b", size = 6400621 }, - { url = "https://files.pythonhosted.org/packages/2b/24/b350c31543fb0107ab2599464d7e28e6f856027aadda995022e695313d94/pillow-12.0.0-cp314-cp314-win_amd64.whl", hash = "sha256:8dc232e39d409036af549c86f24aed8273a40ffa459981146829a324e0848b4b", size = 7142916 }, - { url = "https://files.pythonhosted.org/packages/0f/9b/0ba5a6fd9351793996ef7487c4fdbde8d3f5f75dbedc093bb598648fddf0/pillow-12.0.0-cp314-cp314-win_arm64.whl", hash = "sha256:d52610d51e265a51518692045e372a4c363056130d922a7351429ac9f27e70b0", size = 2523836 }, - { url = "https://files.pythonhosted.org/packages/f5/7a/ceee0840aebc579af529b523d530840338ecf63992395842e54edc805987/pillow-12.0.0-cp314-cp314t-macosx_10_15_x86_64.whl", hash = "sha256:1979f4566bb96c1e50a62d9831e2ea2d1211761e5662afc545fa766f996632f6", size = 5255092 }, - { url = "https://files.pythonhosted.org/packages/44/76/20776057b4bfd1aef4eeca992ebde0f53a4dce874f3ae693d0ec90a4f79b/pillow-12.0.0-cp314-cp314t-macosx_11_0_arm64.whl", hash = "sha256:b2e4b27a6e15b04832fe9bf292b94b5ca156016bbc1ea9c2c20098a0320d6cf6", size = 4653158 }, - { url = "https://files.pythonhosted.org/packages/82/3f/d9ff92ace07be8836b4e7e87e6a4c7a8318d47c2f1463ffcf121fc57d9cb/pillow-12.0.0-cp314-cp314t-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:fb3096c30df99fd01c7bf8e544f392103d0795b9f98ba71a8054bcbf56b255f1", size = 6267882 }, - { url = "https://files.pythonhosted.org/packages/9f/7a/4f7ff87f00d3ad33ba21af78bfcd2f032107710baf8280e3722ceec28cda/pillow-12.0.0-cp314-cp314t-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:7438839e9e053ef79f7112c881cef684013855016f928b168b81ed5835f3e75e", size = 8071001 }, - { url = "https://files.pythonhosted.org/packages/75/87/fcea108944a52dad8cca0715ae6247e271eb80459364a98518f1e4f480c1/pillow-12.0.0-cp314-cp314t-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:5d5c411a8eaa2299322b647cd932586b1427367fd3184ffbb8f7a219ea2041ca", size = 6380146 }, - { url = "https://files.pythonhosted.org/packages/91/52/0d31b5e571ef5fd111d2978b84603fce26aba1b6092f28e941cb46570745/pillow-12.0.0-cp314-cp314t-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:d7e091d464ac59d2c7ad8e7e08105eaf9dafbc3883fd7265ffccc2baad6ac925", size = 7067344 }, - { url = "https://files.pythonhosted.org/packages/7b/f4/2dd3d721f875f928d48e83bb30a434dee75a2531bca839bb996bb0aa5a91/pillow-12.0.0-cp314-cp314t-musllinux_1_2_aarch64.whl", hash = "sha256:792a2c0be4dcc18af9d4a2dfd8a11a17d5e25274a1062b0ec1c2d79c76f3e7f8", size = 6491864 }, - { url = "https://files.pythonhosted.org/packages/30/4b/667dfcf3d61fc309ba5a15b141845cece5915e39b99c1ceab0f34bf1d124/pillow-12.0.0-cp314-cp314t-musllinux_1_2_x86_64.whl", hash = "sha256:afbefa430092f71a9593a99ab6a4e7538bc9eabbf7bf94f91510d3503943edc4", size = 7158911 }, - { url = "https://files.pythonhosted.org/packages/a2/2f/16cabcc6426c32218ace36bf0d55955e813f2958afddbf1d391849fee9d1/pillow-12.0.0-cp314-cp314t-win32.whl", hash = "sha256:3830c769decf88f1289680a59d4f4c46c72573446352e2befec9a8512104fa52", size = 6408045 }, - { url = "https://files.pythonhosted.org/packages/35/73/e29aa0c9c666cf787628d3f0dcf379f4791fba79f4936d02f8b37165bdf8/pillow-12.0.0-cp314-cp314t-win_amd64.whl", hash = "sha256:905b0365b210c73afb0ebe9101a32572152dfd1c144c7e28968a331b9217b94a", size = 7148282 }, - { url = "https://files.pythonhosted.org/packages/c1/70/6b41bdcddf541b437bbb9f47f94d2db5d9ddef6c37ccab8c9107743748a4/pillow-12.0.0-cp314-cp314t-win_arm64.whl", hash = "sha256:99353a06902c2e43b43e8ff74ee65a7d90307d82370604746738a1e0661ccca7", size = 2525630 }, - { url = "https://files.pythonhosted.org/packages/1d/b3/582327e6c9f86d037b63beebe981425d6811104cb443e8193824ef1a2f27/pillow-12.0.0-pp311-pypy311_pp73-macosx_10_15_x86_64.whl", hash = "sha256:b22bd8c974942477156be55a768f7aa37c46904c175be4e158b6a86e3a6b7ca8", size = 5215068 }, - { url = "https://files.pythonhosted.org/packages/fd/d6/67748211d119f3b6540baf90f92fae73ae51d5217b171b0e8b5f7e5d558f/pillow-12.0.0-pp311-pypy311_pp73-macosx_11_0_arm64.whl", hash = "sha256:805ebf596939e48dbb2e4922a1d3852cfc25c38160751ce02da93058b48d252a", size = 4614994 }, - { url = "https://files.pythonhosted.org/packages/2d/e1/f8281e5d844c41872b273b9f2c34a4bf64ca08905668c8ae730eedc7c9fa/pillow-12.0.0-pp311-pypy311_pp73-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:cae81479f77420d217def5f54b5b9d279804d17e982e0f2fa19b1d1e14ab5197", size = 5246639 }, - { url = "https://files.pythonhosted.org/packages/94/5a/0d8ab8ffe8a102ff5df60d0de5af309015163bf710c7bb3e8311dd3b3ad0/pillow-12.0.0-pp311-pypy311_pp73-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:aeaefa96c768fc66818730b952a862235d68825c178f1b3ffd4efd7ad2edcb7c", size = 6986839 }, - { url = "https://files.pythonhosted.org/packages/20/2e/3434380e8110b76cd9eb00a363c484b050f949b4bbe84ba770bb8508a02c/pillow-12.0.0-pp311-pypy311_pp73-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:09f2d0abef9e4e2f349305a4f8cc784a8a6c2f58a8c4892eea13b10a943bd26e", size = 5313505 }, - { url = "https://files.pythonhosted.org/packages/57/ca/5a9d38900d9d74785141d6580950fe705de68af735ff6e727cb911b64740/pillow-12.0.0-pp311-pypy311_pp73-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:bdee52571a343d721fb2eb3b090a82d959ff37fc631e3f70422e0c2e029f3e76", size = 5963654 }, - { url = "https://files.pythonhosted.org/packages/95/7e/f896623c3c635a90537ac093c6a618ebe1a90d87206e42309cb5d98a1b9e/pillow-12.0.0-pp311-pypy311_pp73-win_amd64.whl", hash = "sha256:b290fd8aa38422444d4b50d579de197557f182ef1068b75f5aa8558638b8d0a5", size = 6997850 }, -] - -[[package]] -name = "platformdirs" -version = "4.5.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/cf/86/0248f086a84f01b37aaec0fa567b397df1a119f73c16f6c7a9aac73ea309/platformdirs-4.5.1.tar.gz", hash = "sha256:61d5cdcc6065745cdd94f0f878977f8de9437be93de97c1c12f853c9c0cdcbda", size = 21715 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/cb/28/3bfe2fa5a7b9c46fe7e13c97bda14c895fb10fa2ebf1d0abb90e0cea7ee1/platformdirs-4.5.1-py3-none-any.whl", hash = "sha256:d03afa3963c806a9bed9d5125c8f4cb2fdaf74a55ab60e5d59b3fde758104d31", size = 18731 }, -] - -[[package]] -name = "pluggy" -version = "1.6.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/f9/e2/3e91f31a7d2b083fe6ef3fa267035b518369d9511ffab804f839851d2779/pluggy-1.6.0.tar.gz", hash = "sha256:7dcc130b76258d33b90f61b658791dede3486c3e6bfb003ee5c9bfb396dd22f3", size = 69412 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/54/20/4d324d65cc6d9205fabedc306948156824eb9f0ee1633355a8f7ec5c66bf/pluggy-1.6.0-py3-none-any.whl", hash = "sha256:e920276dd6813095e9377c0bc5566d94c932c33b27a3e3945d8389c374dd4746", size = 20538 }, -] - -[[package]] -name = "pooch" -version = "1.8.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "packaging" }, - { name = "platformdirs" }, - { name = "requests" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/c6/77/b3d3e00c696c16cf99af81ef7b1f5fe73bd2a307abca41bd7605429fe6e5/pooch-1.8.2.tar.gz", hash = "sha256:76561f0de68a01da4df6af38e9955c4c9d1a5c90da73f7e40276a5728ec83d10", size = 59353 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/a8/87/77cc11c7a9ea9fd05503def69e3d18605852cd0d4b0d3b8f15bbeb3ef1d1/pooch-1.8.2-py3-none-any.whl", hash = "sha256:3529a57096f7198778a5ceefd5ac3ef0e4d06a6ddaf9fc2d609b806f25302c47", size = 64574 }, -] - -[[package]] -name = "prompt-toolkit" -version = "3.0.52" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "wcwidth" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/a1/96/06e01a7b38dce6fe1db213e061a4602dd6032a8a97ef6c1a862537732421/prompt_toolkit-3.0.52.tar.gz", hash = "sha256:28cde192929c8e7321de85de1ddbe736f1375148b02f2e17edd840042b1be855", size = 434198 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/84/03/0d3ce49e2505ae70cf43bc5bb3033955d2fc9f932163e84dc0779cc47f48/prompt_toolkit-3.0.52-py3-none-any.whl", hash = "sha256:9aac639a3bbd33284347de5ad8d68ecc044b91a762dc39b7c21095fcd6a19955", size = 391431 }, -] - -[[package]] -name = "propcache" -version = "0.4.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/9e/da/e9fc233cf63743258bff22b3dfa7ea5baef7b5bc324af47a0ad89b8ffc6f/propcache-0.4.1.tar.gz", hash = "sha256:f48107a8c637e80362555f37ecf49abe20370e557cc4ab374f04ec4423c97c3d", size = 46442 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/3c/0e/934b541323035566a9af292dba85a195f7b78179114f2c6ebb24551118a9/propcache-0.4.1-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:7c2d1fa3201efaf55d730400d945b5b3ab6e672e100ba0f9a409d950ab25d7db", size = 79534 }, - { url = "https://files.pythonhosted.org/packages/a1/6b/db0d03d96726d995dc7171286c6ba9d8d14251f37433890f88368951a44e/propcache-0.4.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:1eb2994229cc8ce7fe9b3db88f5465f5fd8651672840b2e426b88cdb1a30aac8", size = 45526 }, - { url = "https://files.pythonhosted.org/packages/e4/c3/82728404aea669e1600f304f2609cde9e665c18df5a11cdd57ed73c1dceb/propcache-0.4.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:66c1f011f45a3b33d7bcb22daed4b29c0c9e2224758b6be00686731e1b46f925", size = 47263 }, - { url = "https://files.pythonhosted.org/packages/df/1b/39313ddad2bf9187a1432654c38249bab4562ef535ef07f5eb6eb04d0b1b/propcache-0.4.1-cp310-cp310-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:9a52009f2adffe195d0b605c25ec929d26b36ef986ba85244891dee3b294df21", size = 201012 }, - { url = "https://files.pythonhosted.org/packages/5b/01/f1d0b57d136f294a142acf97f4ed58c8e5b974c21e543000968357115011/propcache-0.4.1-cp310-cp310-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:5d4e2366a9c7b837555cf02fb9be2e3167d333aff716332ef1b7c3a142ec40c5", size = 209491 }, - { url = "https://files.pythonhosted.org/packages/a1/c8/038d909c61c5bb039070b3fb02ad5cccdb1dde0d714792e251cdb17c9c05/propcache-0.4.1-cp310-cp310-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:9d2b6caef873b4f09e26ea7e33d65f42b944837563a47a94719cc3544319a0db", size = 215319 }, - { url = "https://files.pythonhosted.org/packages/08/57/8c87e93142b2c1fa2408e45695205a7ba05fb5db458c0bf5c06ba0e09ea6/propcache-0.4.1-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:2b16ec437a8c8a965ecf95739448dd938b5c7f56e67ea009f4300d8df05f32b7", size = 196856 }, - { url = "https://files.pythonhosted.org/packages/42/df/5615fec76aa561987a534759b3686008a288e73107faa49a8ae5795a9f7a/propcache-0.4.1-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:296f4c8ed03ca7476813fe666c9ea97869a8d7aec972618671b33a38a5182ef4", size = 193241 }, - { url = "https://files.pythonhosted.org/packages/d5/21/62949eb3a7a54afe8327011c90aca7e03547787a88fb8bd9726806482fea/propcache-0.4.1-cp310-cp310-musllinux_1_2_armv7l.whl", hash = "sha256:1f0978529a418ebd1f49dad413a2b68af33f85d5c5ca5c6ca2a3bed375a7ac60", size = 190552 }, - { url = "https://files.pythonhosted.org/packages/30/ee/ab4d727dd70806e5b4de96a798ae7ac6e4d42516f030ee60522474b6b332/propcache-0.4.1-cp310-cp310-musllinux_1_2_ppc64le.whl", hash = "sha256:fd138803047fb4c062b1c1dd95462f5209456bfab55c734458f15d11da288f8f", size = 200113 }, - { url = "https://files.pythonhosted.org/packages/8a/0b/38b46208e6711b016aa8966a3ac793eee0d05c7159d8342aa27fc0bc365e/propcache-0.4.1-cp310-cp310-musllinux_1_2_s390x.whl", hash = "sha256:8c9b3cbe4584636d72ff556d9036e0c9317fa27b3ac1f0f558e7e84d1c9c5900", size = 200778 }, - { url = "https://files.pythonhosted.org/packages/cf/81/5abec54355ed344476bee711e9f04815d4b00a311ab0535599204eecc257/propcache-0.4.1-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:f93243fdc5657247533273ac4f86ae106cc6445a0efacb9a1bfe982fcfefd90c", size = 193047 }, - { url = "https://files.pythonhosted.org/packages/ec/b6/1f237c04e32063cb034acd5f6ef34ef3a394f75502e72703545631ab1ef6/propcache-0.4.1-cp310-cp310-win32.whl", hash = "sha256:a0ee98db9c5f80785b266eb805016e36058ac72c51a064040f2bc43b61101cdb", size = 38093 }, - { url = "https://files.pythonhosted.org/packages/a6/67/354aac4e0603a15f76439caf0427781bcd6797f370377f75a642133bc954/propcache-0.4.1-cp310-cp310-win_amd64.whl", hash = "sha256:1cdb7988c4e5ac7f6d175a28a9aa0c94cb6f2ebe52756a3c0cda98d2809a9e37", size = 41638 }, - { url = "https://files.pythonhosted.org/packages/e0/e1/74e55b9fd1a4c209ff1a9a824bf6c8b3d1fc5a1ac3eabe23462637466785/propcache-0.4.1-cp310-cp310-win_arm64.whl", hash = "sha256:d82ad62b19645419fe79dd63b3f9253e15b30e955c0170e5cebc350c1844e581", size = 38229 }, - { url = "https://files.pythonhosted.org/packages/8c/d4/4e2c9aaf7ac2242b9358f98dccd8f90f2605402f5afeff6c578682c2c491/propcache-0.4.1-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:60a8fda9644b7dfd5dece8c61d8a85e271cb958075bfc4e01083c148b61a7caf", size = 80208 }, - { url = "https://files.pythonhosted.org/packages/c2/21/d7b68e911f9c8e18e4ae43bdbc1e1e9bbd971f8866eb81608947b6f585ff/propcache-0.4.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:c30b53e7e6bda1d547cabb47c825f3843a0a1a42b0496087bb58d8fedf9f41b5", size = 45777 }, - { url = "https://files.pythonhosted.org/packages/d3/1d/11605e99ac8ea9435651ee71ab4cb4bf03f0949586246476a25aadfec54a/propcache-0.4.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:6918ecbd897443087a3b7cd978d56546a812517dcaaca51b49526720571fa93e", size = 47647 }, - { url = "https://files.pythonhosted.org/packages/58/1a/3c62c127a8466c9c843bccb503d40a273e5cc69838805f322e2826509e0d/propcache-0.4.1-cp311-cp311-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:3d902a36df4e5989763425a8ab9e98cd8ad5c52c823b34ee7ef307fd50582566", size = 214929 }, - { url = "https://files.pythonhosted.org/packages/56/b9/8fa98f850960b367c4b8fe0592e7fc341daa7a9462e925228f10a60cf74f/propcache-0.4.1-cp311-cp311-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:a9695397f85973bb40427dedddf70d8dc4a44b22f1650dd4af9eedf443d45165", size = 221778 }, - { url = "https://files.pythonhosted.org/packages/46/a6/0ab4f660eb59649d14b3d3d65c439421cf2f87fe5dd68591cbe3c1e78a89/propcache-0.4.1-cp311-cp311-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:2bb07ffd7eaad486576430c89f9b215f9e4be68c4866a96e97db9e97fead85dc", size = 228144 }, - { url = "https://files.pythonhosted.org/packages/52/6a/57f43e054fb3d3a56ac9fc532bc684fc6169a26c75c353e65425b3e56eef/propcache-0.4.1-cp311-cp311-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:fd6f30fdcf9ae2a70abd34da54f18da086160e4d7d9251f81f3da0ff84fc5a48", size = 210030 }, - { url = "https://files.pythonhosted.org/packages/40/e2/27e6feebb5f6b8408fa29f5efbb765cd54c153ac77314d27e457a3e993b7/propcache-0.4.1-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:fc38cba02d1acba4e2869eef1a57a43dfbd3d49a59bf90dda7444ec2be6a5570", size = 208252 }, - { url = "https://files.pythonhosted.org/packages/9e/f8/91c27b22ccda1dbc7967f921c42825564fa5336a01ecd72eb78a9f4f53c2/propcache-0.4.1-cp311-cp311-musllinux_1_2_armv7l.whl", hash = "sha256:67fad6162281e80e882fb3ec355398cf72864a54069d060321f6cd0ade95fe85", size = 202064 }, - { url = "https://files.pythonhosted.org/packages/f2/26/7f00bd6bd1adba5aafe5f4a66390f243acab58eab24ff1a08bebb2ef9d40/propcache-0.4.1-cp311-cp311-musllinux_1_2_ppc64le.whl", hash = "sha256:f10207adf04d08bec185bae14d9606a1444715bc99180f9331c9c02093e1959e", size = 212429 }, - { url = "https://files.pythonhosted.org/packages/84/89/fd108ba7815c1117ddca79c228f3f8a15fc82a73bca8b142eb5de13b2785/propcache-0.4.1-cp311-cp311-musllinux_1_2_s390x.whl", hash = "sha256:e9b0d8d0845bbc4cfcdcbcdbf5086886bc8157aa963c31c777ceff7846c77757", size = 216727 }, - { url = "https://files.pythonhosted.org/packages/79/37/3ec3f7e3173e73f1d600495d8b545b53802cbf35506e5732dd8578db3724/propcache-0.4.1-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:981333cb2f4c1896a12f4ab92a9cc8f09ea664e9b7dbdc4eff74627af3a11c0f", size = 205097 }, - { url = "https://files.pythonhosted.org/packages/61/b0/b2631c19793f869d35f47d5a3a56fb19e9160d3c119f15ac7344fc3ccae7/propcache-0.4.1-cp311-cp311-win32.whl", hash = "sha256:f1d2f90aeec838a52f1c1a32fe9a619fefd5e411721a9117fbf82aea638fe8a1", size = 38084 }, - { url = "https://files.pythonhosted.org/packages/f4/78/6cce448e2098e9f3bfc91bb877f06aa24b6ccace872e39c53b2f707c4648/propcache-0.4.1-cp311-cp311-win_amd64.whl", hash = "sha256:364426a62660f3f699949ac8c621aad6977be7126c5807ce48c0aeb8e7333ea6", size = 41637 }, - { url = "https://files.pythonhosted.org/packages/9c/e9/754f180cccd7f51a39913782c74717c581b9cc8177ad0e949f4d51812383/propcache-0.4.1-cp311-cp311-win_arm64.whl", hash = "sha256:e53f3a38d3510c11953f3e6a33f205c6d1b001129f972805ca9b42fc308bc239", size = 38064 }, - { url = "https://files.pythonhosted.org/packages/a2/0f/f17b1b2b221d5ca28b4b876e8bb046ac40466513960646bda8e1853cdfa2/propcache-0.4.1-cp312-cp312-macosx_10_13_universal2.whl", hash = "sha256:e153e9cd40cc8945138822807139367f256f89c6810c2634a4f6902b52d3b4e2", size = 80061 }, - { url = "https://files.pythonhosted.org/packages/76/47/8ccf75935f51448ba9a16a71b783eb7ef6b9ee60f5d14c7f8a8a79fbeed7/propcache-0.4.1-cp312-cp312-macosx_10_13_x86_64.whl", hash = "sha256:cd547953428f7abb73c5ad82cbb32109566204260d98e41e5dfdc682eb7f8403", size = 46037 }, - { url = "https://files.pythonhosted.org/packages/0a/b6/5c9a0e42df4d00bfb4a3cbbe5cf9f54260300c88a0e9af1f47ca5ce17ac0/propcache-0.4.1-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:f048da1b4f243fc44f205dfd320933a951b8d89e0afd4c7cacc762a8b9165207", size = 47324 }, - { url = "https://files.pythonhosted.org/packages/9e/d3/6c7ee328b39a81ee877c962469f1e795f9db87f925251efeb0545e0020d0/propcache-0.4.1-cp312-cp312-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:ec17c65562a827bba85e3872ead335f95405ea1674860d96483a02f5c698fa72", size = 225505 }, - { url = "https://files.pythonhosted.org/packages/01/5d/1c53f4563490b1d06a684742cc6076ef944bc6457df6051b7d1a877c057b/propcache-0.4.1-cp312-cp312-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:405aac25c6394ef275dee4c709be43745d36674b223ba4eb7144bf4d691b7367", size = 230242 }, - { url = "https://files.pythonhosted.org/packages/20/e1/ce4620633b0e2422207c3cb774a0ee61cac13abc6217763a7b9e2e3f4a12/propcache-0.4.1-cp312-cp312-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:0013cb6f8dde4b2a2f66903b8ba740bdfe378c943c4377a200551ceb27f379e4", size = 238474 }, - { url = "https://files.pythonhosted.org/packages/46/4b/3aae6835b8e5f44ea6a68348ad90f78134047b503765087be2f9912140ea/propcache-0.4.1-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:15932ab57837c3368b024473a525e25d316d8353016e7cc0e5ba9eb343fbb1cf", size = 221575 }, - { url = "https://files.pythonhosted.org/packages/6e/a5/8a5e8678bcc9d3a1a15b9a29165640d64762d424a16af543f00629c87338/propcache-0.4.1-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:031dce78b9dc099f4c29785d9cf5577a3faf9ebf74ecbd3c856a7b92768c3df3", size = 216736 }, - { url = "https://files.pythonhosted.org/packages/f1/63/b7b215eddeac83ca1c6b934f89d09a625aa9ee4ba158338854c87210cc36/propcache-0.4.1-cp312-cp312-musllinux_1_2_armv7l.whl", hash = "sha256:ab08df6c9a035bee56e31af99be621526bd237bea9f32def431c656b29e41778", size = 213019 }, - { url = "https://files.pythonhosted.org/packages/57/74/f580099a58c8af587cac7ba19ee7cb418506342fbbe2d4a4401661cca886/propcache-0.4.1-cp312-cp312-musllinux_1_2_ppc64le.whl", hash = "sha256:4d7af63f9f93fe593afbf104c21b3b15868efb2c21d07d8732c0c4287e66b6a6", size = 220376 }, - { url = "https://files.pythonhosted.org/packages/c4/ee/542f1313aff7eaf19c2bb758c5d0560d2683dac001a1c96d0774af799843/propcache-0.4.1-cp312-cp312-musllinux_1_2_s390x.whl", hash = "sha256:cfc27c945f422e8b5071b6e93169679e4eb5bf73bbcbf1ba3ae3a83d2f78ebd9", size = 226988 }, - { url = "https://files.pythonhosted.org/packages/8f/18/9c6b015dd9c6930f6ce2229e1f02fb35298b847f2087ea2b436a5bfa7287/propcache-0.4.1-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:35c3277624a080cc6ec6f847cbbbb5b49affa3598c4535a0a4682a697aaa5c75", size = 215615 }, - { url = "https://files.pythonhosted.org/packages/80/9e/e7b85720b98c45a45e1fca6a177024934dc9bc5f4d5dd04207f216fc33ed/propcache-0.4.1-cp312-cp312-win32.whl", hash = "sha256:671538c2262dadb5ba6395e26c1731e1d52534bfe9ae56d0b5573ce539266aa8", size = 38066 }, - { url = "https://files.pythonhosted.org/packages/54/09/d19cff2a5aaac632ec8fc03737b223597b1e347416934c1b3a7df079784c/propcache-0.4.1-cp312-cp312-win_amd64.whl", hash = "sha256:cb2d222e72399fcf5890d1d5cc1060857b9b236adff2792ff48ca2dfd46c81db", size = 41655 }, - { url = "https://files.pythonhosted.org/packages/68/ab/6b5c191bb5de08036a8c697b265d4ca76148efb10fa162f14af14fb5f076/propcache-0.4.1-cp312-cp312-win_arm64.whl", hash = "sha256:204483131fb222bdaaeeea9f9e6c6ed0cac32731f75dfc1d4a567fc1926477c1", size = 37789 }, - { url = "https://files.pythonhosted.org/packages/bf/df/6d9c1b6ac12b003837dde8a10231a7344512186e87b36e855bef32241942/propcache-0.4.1-cp313-cp313-macosx_10_13_universal2.whl", hash = "sha256:43eedf29202c08550aac1d14e0ee619b0430aaef78f85864c1a892294fbc28cf", size = 77750 }, - { url = "https://files.pythonhosted.org/packages/8b/e8/677a0025e8a2acf07d3418a2e7ba529c9c33caf09d3c1f25513023c1db56/propcache-0.4.1-cp313-cp313-macosx_10_13_x86_64.whl", hash = "sha256:d62cdfcfd89ccb8de04e0eda998535c406bf5e060ffd56be6c586cbcc05b3311", size = 44780 }, - { url = "https://files.pythonhosted.org/packages/89/a4/92380f7ca60f99ebae761936bc48a72a639e8a47b29050615eef757cb2a7/propcache-0.4.1-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:cae65ad55793da34db5f54e4029b89d3b9b9490d8abe1b4c7ab5d4b8ec7ebf74", size = 46308 }, - { url = "https://files.pythonhosted.org/packages/2d/48/c5ac64dee5262044348d1d78a5f85dd1a57464a60d30daee946699963eb3/propcache-0.4.1-cp313-cp313-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:333ddb9031d2704a301ee3e506dc46b1fe5f294ec198ed6435ad5b6a085facfe", size = 208182 }, - { url = "https://files.pythonhosted.org/packages/c6/0c/cd762dd011a9287389a6a3eb43aa30207bde253610cca06824aeabfe9653/propcache-0.4.1-cp313-cp313-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:fd0858c20f078a32cf55f7e81473d96dcf3b93fd2ccdb3d40fdf54b8573df3af", size = 211215 }, - { url = "https://files.pythonhosted.org/packages/30/3e/49861e90233ba36890ae0ca4c660e95df565b2cd15d4a68556ab5865974e/propcache-0.4.1-cp313-cp313-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:678ae89ebc632c5c204c794f8dab2837c5f159aeb59e6ed0539500400577298c", size = 218112 }, - { url = "https://files.pythonhosted.org/packages/f1/8b/544bc867e24e1bd48f3118cecd3b05c694e160a168478fa28770f22fd094/propcache-0.4.1-cp313-cp313-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:d472aeb4fbf9865e0c6d622d7f4d54a4e101a89715d8904282bb5f9a2f476c3f", size = 204442 }, - { url = "https://files.pythonhosted.org/packages/50/a6/4282772fd016a76d3e5c0df58380a5ea64900afd836cec2c2f662d1b9bb3/propcache-0.4.1-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:4d3df5fa7e36b3225954fba85589da77a0fe6a53e3976de39caf04a0db4c36f1", size = 199398 }, - { url = "https://files.pythonhosted.org/packages/3e/ec/d8a7cd406ee1ddb705db2139f8a10a8a427100347bd698e7014351c7af09/propcache-0.4.1-cp313-cp313-musllinux_1_2_armv7l.whl", hash = "sha256:ee17f18d2498f2673e432faaa71698032b0127ebf23ae5974eeaf806c279df24", size = 196920 }, - { url = "https://files.pythonhosted.org/packages/f6/6c/f38ab64af3764f431e359f8baf9e0a21013e24329e8b85d2da32e8ed07ca/propcache-0.4.1-cp313-cp313-musllinux_1_2_ppc64le.whl", hash = "sha256:580e97762b950f993ae618e167e7be9256b8353c2dcd8b99ec100eb50f5286aa", size = 203748 }, - { url = "https://files.pythonhosted.org/packages/d6/e3/fa846bd70f6534d647886621388f0a265254d30e3ce47e5c8e6e27dbf153/propcache-0.4.1-cp313-cp313-musllinux_1_2_s390x.whl", hash = "sha256:501d20b891688eb8e7aa903021f0b72d5a55db40ffaab27edefd1027caaafa61", size = 205877 }, - { url = "https://files.pythonhosted.org/packages/e2/39/8163fc6f3133fea7b5f2827e8eba2029a0277ab2c5beee6c1db7b10fc23d/propcache-0.4.1-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:9a0bd56e5b100aef69bd8562b74b46254e7c8812918d3baa700c8a8009b0af66", size = 199437 }, - { url = "https://files.pythonhosted.org/packages/93/89/caa9089970ca49c7c01662bd0eeedfe85494e863e8043565aeb6472ce8fe/propcache-0.4.1-cp313-cp313-win32.whl", hash = "sha256:bcc9aaa5d80322bc2fb24bb7accb4a30f81e90ab8d6ba187aec0744bc302ad81", size = 37586 }, - { url = "https://files.pythonhosted.org/packages/f5/ab/f76ec3c3627c883215b5c8080debb4394ef5a7a29be811f786415fc1e6fd/propcache-0.4.1-cp313-cp313-win_amd64.whl", hash = "sha256:381914df18634f5494334d201e98245c0596067504b9372d8cf93f4bb23e025e", size = 40790 }, - { url = "https://files.pythonhosted.org/packages/59/1b/e71ae98235f8e2ba5004d8cb19765a74877abf189bc53fc0c80d799e56c3/propcache-0.4.1-cp313-cp313-win_arm64.whl", hash = "sha256:8873eb4460fd55333ea49b7d189749ecf6e55bf85080f11b1c4530ed3034cba1", size = 37158 }, - { url = "https://files.pythonhosted.org/packages/83/ce/a31bbdfc24ee0dcbba458c8175ed26089cf109a55bbe7b7640ed2470cfe9/propcache-0.4.1-cp313-cp313t-macosx_10_13_universal2.whl", hash = "sha256:92d1935ee1f8d7442da9c0c4fa7ac20d07e94064184811b685f5c4fada64553b", size = 81451 }, - { url = "https://files.pythonhosted.org/packages/25/9c/442a45a470a68456e710d96cacd3573ef26a1d0a60067e6a7d5e655621ed/propcache-0.4.1-cp313-cp313t-macosx_10_13_x86_64.whl", hash = "sha256:473c61b39e1460d386479b9b2f337da492042447c9b685f28be4f74d3529e566", size = 46374 }, - { url = "https://files.pythonhosted.org/packages/f4/bf/b1d5e21dbc3b2e889ea4327044fb16312a736d97640fb8b6aa3f9c7b3b65/propcache-0.4.1-cp313-cp313t-macosx_11_0_arm64.whl", hash = "sha256:c0ef0aaafc66fbd87842a3fe3902fd889825646bc21149eafe47be6072725835", size = 48396 }, - { url = "https://files.pythonhosted.org/packages/f4/04/5b4c54a103d480e978d3c8a76073502b18db0c4bc17ab91b3cb5092ad949/propcache-0.4.1-cp313-cp313t-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:f95393b4d66bfae908c3ca8d169d5f79cd65636ae15b5e7a4f6e67af675adb0e", size = 275950 }, - { url = "https://files.pythonhosted.org/packages/b4/c1/86f846827fb969c4b78b0af79bba1d1ea2156492e1b83dea8b8a6ae27395/propcache-0.4.1-cp313-cp313t-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:c07fda85708bc48578467e85099645167a955ba093be0a2dcba962195676e859", size = 273856 }, - { url = "https://files.pythonhosted.org/packages/36/1d/fc272a63c8d3bbad6878c336c7a7dea15e8f2d23a544bda43205dfa83ada/propcache-0.4.1-cp313-cp313t-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:af223b406d6d000830c6f65f1e6431783fc3f713ba3e6cc8c024d5ee96170a4b", size = 280420 }, - { url = "https://files.pythonhosted.org/packages/07/0c/01f2219d39f7e53d52e5173bcb09c976609ba30209912a0680adfb8c593a/propcache-0.4.1-cp313-cp313t-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:a78372c932c90ee474559c5ddfffd718238e8673c340dc21fe45c5b8b54559a0", size = 263254 }, - { url = "https://files.pythonhosted.org/packages/2d/18/cd28081658ce597898f0c4d174d4d0f3c5b6d4dc27ffafeef835c95eb359/propcache-0.4.1-cp313-cp313t-musllinux_1_2_aarch64.whl", hash = "sha256:564d9f0d4d9509e1a870c920a89b2fec951b44bf5ba7d537a9e7c1ccec2c18af", size = 261205 }, - { url = "https://files.pythonhosted.org/packages/7a/71/1f9e22eb8b8316701c2a19fa1f388c8a3185082607da8e406a803c9b954e/propcache-0.4.1-cp313-cp313t-musllinux_1_2_armv7l.whl", hash = "sha256:17612831fda0138059cc5546f4d12a2aacfb9e47068c06af35c400ba58ba7393", size = 247873 }, - { url = "https://files.pythonhosted.org/packages/4a/65/3d4b61f36af2b4eddba9def857959f1016a51066b4f1ce348e0cf7881f58/propcache-0.4.1-cp313-cp313t-musllinux_1_2_ppc64le.whl", hash = "sha256:41a89040cb10bd345b3c1a873b2bf36413d48da1def52f268a055f7398514874", size = 262739 }, - { url = "https://files.pythonhosted.org/packages/2a/42/26746ab087faa77c1c68079b228810436ccd9a5ce9ac85e2b7307195fd06/propcache-0.4.1-cp313-cp313t-musllinux_1_2_s390x.whl", hash = "sha256:e35b88984e7fa64aacecea39236cee32dd9bd8c55f57ba8a75cf2399553f9bd7", size = 263514 }, - { url = "https://files.pythonhosted.org/packages/94/13/630690fe201f5502d2403dd3cfd451ed8858fe3c738ee88d095ad2ff407b/propcache-0.4.1-cp313-cp313t-musllinux_1_2_x86_64.whl", hash = "sha256:6f8b465489f927b0df505cbe26ffbeed4d6d8a2bbc61ce90eb074ff129ef0ab1", size = 257781 }, - { url = "https://files.pythonhosted.org/packages/92/f7/1d4ec5841505f423469efbfc381d64b7b467438cd5a4bbcbb063f3b73d27/propcache-0.4.1-cp313-cp313t-win32.whl", hash = "sha256:2ad890caa1d928c7c2965b48f3a3815c853180831d0e5503d35cf00c472f4717", size = 41396 }, - { url = "https://files.pythonhosted.org/packages/48/f0/615c30622316496d2cbbc29f5985f7777d3ada70f23370608c1d3e081c1f/propcache-0.4.1-cp313-cp313t-win_amd64.whl", hash = "sha256:f7ee0e597f495cf415bcbd3da3caa3bd7e816b74d0d52b8145954c5e6fd3ff37", size = 44897 }, - { url = "https://files.pythonhosted.org/packages/fd/ca/6002e46eccbe0e33dcd4069ef32f7f1c9e243736e07adca37ae8c4830ec3/propcache-0.4.1-cp313-cp313t-win_arm64.whl", hash = "sha256:929d7cbe1f01bb7baffb33dc14eb5691c95831450a26354cd210a8155170c93a", size = 39789 }, - { url = "https://files.pythonhosted.org/packages/8e/5c/bca52d654a896f831b8256683457ceddd490ec18d9ec50e97dfd8fc726a8/propcache-0.4.1-cp314-cp314-macosx_10_13_universal2.whl", hash = "sha256:3f7124c9d820ba5548d431afb4632301acf965db49e666aa21c305cbe8c6de12", size = 78152 }, - { url = "https://files.pythonhosted.org/packages/65/9b/03b04e7d82a5f54fb16113d839f5ea1ede58a61e90edf515f6577c66fa8f/propcache-0.4.1-cp314-cp314-macosx_10_13_x86_64.whl", hash = "sha256:c0d4b719b7da33599dfe3b22d3db1ef789210a0597bc650b7cee9c77c2be8c5c", size = 44869 }, - { url = "https://files.pythonhosted.org/packages/b2/fa/89a8ef0468d5833a23fff277b143d0573897cf75bd56670a6d28126c7d68/propcache-0.4.1-cp314-cp314-macosx_11_0_arm64.whl", hash = "sha256:9f302f4783709a78240ebc311b793f123328716a60911d667e0c036bc5dcbded", size = 46596 }, - { url = "https://files.pythonhosted.org/packages/86/bd/47816020d337f4a746edc42fe8d53669965138f39ee117414c7d7a340cfe/propcache-0.4.1-cp314-cp314-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:c80ee5802e3fb9ea37938e7eecc307fb984837091d5fd262bb37238b1ae97641", size = 206981 }, - { url = "https://files.pythonhosted.org/packages/df/f6/c5fa1357cc9748510ee55f37173eb31bfde6d94e98ccd9e6f033f2fc06e1/propcache-0.4.1-cp314-cp314-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:ed5a841e8bb29a55fb8159ed526b26adc5bdd7e8bd7bf793ce647cb08656cdf4", size = 211490 }, - { url = "https://files.pythonhosted.org/packages/80/1e/e5889652a7c4a3846683401a48f0f2e5083ce0ec1a8a5221d8058fbd1adf/propcache-0.4.1-cp314-cp314-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:55c72fd6ea2da4c318e74ffdf93c4fe4e926051133657459131a95c846d16d44", size = 215371 }, - { url = "https://files.pythonhosted.org/packages/b2/f2/889ad4b2408f72fe1a4f6a19491177b30ea7bf1a0fd5f17050ca08cfc882/propcache-0.4.1-cp314-cp314-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:8326e144341460402713f91df60ade3c999d601e7eb5ff8f6f7862d54de0610d", size = 201424 }, - { url = "https://files.pythonhosted.org/packages/27/73/033d63069b57b0812c8bd19f311faebeceb6ba31b8f32b73432d12a0b826/propcache-0.4.1-cp314-cp314-musllinux_1_2_aarch64.whl", hash = "sha256:060b16ae65bc098da7f6d25bf359f1f31f688384858204fe5d652979e0015e5b", size = 197566 }, - { url = "https://files.pythonhosted.org/packages/dc/89/ce24f3dc182630b4e07aa6d15f0ff4b14ed4b9955fae95a0b54c58d66c05/propcache-0.4.1-cp314-cp314-musllinux_1_2_armv7l.whl", hash = "sha256:89eb3fa9524f7bec9de6e83cf3faed9d79bffa560672c118a96a171a6f55831e", size = 193130 }, - { url = "https://files.pythonhosted.org/packages/a9/24/ef0d5fd1a811fb5c609278d0209c9f10c35f20581fcc16f818da959fc5b4/propcache-0.4.1-cp314-cp314-musllinux_1_2_ppc64le.whl", hash = "sha256:dee69d7015dc235f526fe80a9c90d65eb0039103fe565776250881731f06349f", size = 202625 }, - { url = "https://files.pythonhosted.org/packages/f5/02/98ec20ff5546f68d673df2f7a69e8c0d076b5abd05ca882dc7ee3a83653d/propcache-0.4.1-cp314-cp314-musllinux_1_2_s390x.whl", hash = "sha256:5558992a00dfd54ccbc64a32726a3357ec93825a418a401f5cc67df0ac5d9e49", size = 204209 }, - { url = "https://files.pythonhosted.org/packages/a0/87/492694f76759b15f0467a2a93ab68d32859672b646aa8a04ce4864e7932d/propcache-0.4.1-cp314-cp314-musllinux_1_2_x86_64.whl", hash = "sha256:c9b822a577f560fbd9554812526831712c1436d2c046cedee4c3796d3543b144", size = 197797 }, - { url = "https://files.pythonhosted.org/packages/ee/36/66367de3575db1d2d3f3d177432bd14ee577a39d3f5d1b3d5df8afe3b6e2/propcache-0.4.1-cp314-cp314-win32.whl", hash = "sha256:ab4c29b49d560fe48b696cdcb127dd36e0bc2472548f3bf56cc5cb3da2b2984f", size = 38140 }, - { url = "https://files.pythonhosted.org/packages/0c/2a/a758b47de253636e1b8aef181c0b4f4f204bf0dd964914fb2af90a95b49b/propcache-0.4.1-cp314-cp314-win_amd64.whl", hash = "sha256:5a103c3eb905fcea0ab98be99c3a9a5ab2de60228aa5aceedc614c0281cf6153", size = 41257 }, - { url = "https://files.pythonhosted.org/packages/34/5e/63bd5896c3fec12edcbd6f12508d4890d23c265df28c74b175e1ef9f4f3b/propcache-0.4.1-cp314-cp314-win_arm64.whl", hash = "sha256:74c1fb26515153e482e00177a1ad654721bf9207da8a494a0c05e797ad27b992", size = 38097 }, - { url = "https://files.pythonhosted.org/packages/99/85/9ff785d787ccf9bbb3f3106f79884a130951436f58392000231b4c737c80/propcache-0.4.1-cp314-cp314t-macosx_10_13_universal2.whl", hash = "sha256:824e908bce90fb2743bd6b59db36eb4f45cd350a39637c9f73b1c1ea66f5b75f", size = 81455 }, - { url = "https://files.pythonhosted.org/packages/90/85/2431c10c8e7ddb1445c1f7c4b54d886e8ad20e3c6307e7218f05922cad67/propcache-0.4.1-cp314-cp314t-macosx_10_13_x86_64.whl", hash = "sha256:c2b5e7db5328427c57c8e8831abda175421b709672f6cfc3d630c3b7e2146393", size = 46372 }, - { url = "https://files.pythonhosted.org/packages/01/20/b0972d902472da9bcb683fa595099911f4d2e86e5683bcc45de60dd05dc3/propcache-0.4.1-cp314-cp314t-macosx_11_0_arm64.whl", hash = "sha256:6f6ff873ed40292cd4969ef5310179afd5db59fdf055897e282485043fc80ad0", size = 48411 }, - { url = "https://files.pythonhosted.org/packages/e2/e3/7dc89f4f21e8f99bad3d5ddb3a3389afcf9da4ac69e3deb2dcdc96e74169/propcache-0.4.1-cp314-cp314t-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:49a2dc67c154db2c1463013594c458881a069fcf98940e61a0569016a583020a", size = 275712 }, - { url = "https://files.pythonhosted.org/packages/20/67/89800c8352489b21a8047c773067644e3897f02ecbbd610f4d46b7f08612/propcache-0.4.1-cp314-cp314t-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:005f08e6a0529984491e37d8dbc3dd86f84bd78a8ceb5fa9a021f4c48d4984be", size = 273557 }, - { url = "https://files.pythonhosted.org/packages/e2/a1/b52b055c766a54ce6d9c16d9aca0cad8059acd9637cdf8aa0222f4a026ef/propcache-0.4.1-cp314-cp314t-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:5c3310452e0d31390da9035c348633b43d7e7feb2e37be252be6da45abd1abcc", size = 280015 }, - { url = "https://files.pythonhosted.org/packages/48/c8/33cee30bd890672c63743049f3c9e4be087e6780906bfc3ec58528be59c1/propcache-0.4.1-cp314-cp314t-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:4c3c70630930447f9ef1caac7728c8ad1c56bc5015338b20fed0d08ea2480b3a", size = 262880 }, - { url = "https://files.pythonhosted.org/packages/0c/b1/8f08a143b204b418285c88b83d00edbd61afbc2c6415ffafc8905da7038b/propcache-0.4.1-cp314-cp314t-musllinux_1_2_aarch64.whl", hash = "sha256:8e57061305815dfc910a3634dcf584f08168a8836e6999983569f51a8544cd89", size = 260938 }, - { url = "https://files.pythonhosted.org/packages/cf/12/96e4664c82ca2f31e1c8dff86afb867348979eb78d3cb8546a680287a1e9/propcache-0.4.1-cp314-cp314t-musllinux_1_2_armv7l.whl", hash = "sha256:521a463429ef54143092c11a77e04056dd00636f72e8c45b70aaa3140d639726", size = 247641 }, - { url = "https://files.pythonhosted.org/packages/18/ed/e7a9cfca28133386ba52278136d42209d3125db08d0a6395f0cba0c0285c/propcache-0.4.1-cp314-cp314t-musllinux_1_2_ppc64le.whl", hash = "sha256:120c964da3fdc75e3731aa392527136d4ad35868cc556fd09bb6d09172d9a367", size = 262510 }, - { url = "https://files.pythonhosted.org/packages/f5/76/16d8bf65e8845dd62b4e2b57444ab81f07f40caa5652b8969b87ddcf2ef6/propcache-0.4.1-cp314-cp314t-musllinux_1_2_s390x.whl", hash = "sha256:d8f353eb14ee3441ee844ade4277d560cdd68288838673273b978e3d6d2c8f36", size = 263161 }, - { url = "https://files.pythonhosted.org/packages/e7/70/c99e9edb5d91d5ad8a49fa3c1e8285ba64f1476782fed10ab251ff413ba1/propcache-0.4.1-cp314-cp314t-musllinux_1_2_x86_64.whl", hash = "sha256:ab2943be7c652f09638800905ee1bab2c544e537edb57d527997a24c13dc1455", size = 257393 }, - { url = "https://files.pythonhosted.org/packages/08/02/87b25304249a35c0915d236575bc3574a323f60b47939a2262b77632a3ee/propcache-0.4.1-cp314-cp314t-win32.whl", hash = "sha256:05674a162469f31358c30bcaa8883cb7829fa3110bf9c0991fe27d7896c42d85", size = 42546 }, - { url = "https://files.pythonhosted.org/packages/cb/ef/3c6ecf8b317aa982f309835e8f96987466123c6e596646d4e6a1dfcd080f/propcache-0.4.1-cp314-cp314t-win_amd64.whl", hash = "sha256:990f6b3e2a27d683cb7602ed6c86f15ee6b43b1194736f9baaeb93d0016633b1", size = 46259 }, - { url = "https://files.pythonhosted.org/packages/c4/2d/346e946d4951f37eca1e4f55be0f0174c52cd70720f84029b02f296f4a38/propcache-0.4.1-cp314-cp314t-win_arm64.whl", hash = "sha256:ecef2343af4cc68e05131e45024ba34f6095821988a9d0a02aa7c73fcc448aa9", size = 40428 }, - { url = "https://files.pythonhosted.org/packages/5b/5a/bc7b4a4ef808fa59a816c17b20c4bef6884daebbdf627ff2a161da67da19/propcache-0.4.1-py3-none-any.whl", hash = "sha256:af2a6052aeb6cf17d3e46ee169099044fd8224cbaf75c76a2ef596e8163e2237", size = 13305 }, -] - -[[package]] -name = "protobuf" -version = "5.29.5" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/43/29/d09e70352e4e88c9c7a198d5645d7277811448d76c23b00345670f7c8a38/protobuf-5.29.5.tar.gz", hash = "sha256:bc1463bafd4b0929216c35f437a8e28731a2b7fe3d98bb77a600efced5a15c84", size = 425226 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/5f/11/6e40e9fc5bba02988a214c07cf324595789ca7820160bfd1f8be96e48539/protobuf-5.29.5-cp310-abi3-win32.whl", hash = "sha256:3f1c6468a2cfd102ff4703976138844f78ebd1fb45f49011afc5139e9e283079", size = 422963 }, - { url = "https://files.pythonhosted.org/packages/81/7f/73cefb093e1a2a7c3ffd839e6f9fcafb7a427d300c7f8aef9c64405d8ac6/protobuf-5.29.5-cp310-abi3-win_amd64.whl", hash = "sha256:3f76e3a3675b4a4d867b52e4a5f5b78a2ef9565549d4037e06cf7b0942b1d3fc", size = 434818 }, - { url = "https://files.pythonhosted.org/packages/dd/73/10e1661c21f139f2c6ad9b23040ff36fee624310dc28fba20d33fdae124c/protobuf-5.29.5-cp38-abi3-macosx_10_9_universal2.whl", hash = "sha256:e38c5add5a311f2a6eb0340716ef9b039c1dfa428b28f25a7838ac329204a671", size = 418091 }, - { url = "https://files.pythonhosted.org/packages/6c/04/98f6f8cf5b07ab1294c13f34b4e69b3722bb609c5b701d6c169828f9f8aa/protobuf-5.29.5-cp38-abi3-manylinux2014_aarch64.whl", hash = "sha256:fa18533a299d7ab6c55a238bf8629311439995f2e7eca5caaff08663606e9015", size = 319824 }, - { url = "https://files.pythonhosted.org/packages/85/e4/07c80521879c2d15f321465ac24c70efe2381378c00bf5e56a0f4fbac8cd/protobuf-5.29.5-cp38-abi3-manylinux2014_x86_64.whl", hash = "sha256:63848923da3325e1bf7e9003d680ce6e14b07e55d0473253a690c3a8b8fd6e61", size = 319942 }, - { url = "https://files.pythonhosted.org/packages/7e/cc/7e77861000a0691aeea8f4566e5d3aa716f2b1dece4a24439437e41d3d25/protobuf-5.29.5-py3-none-any.whl", hash = "sha256:6cf42630262c59b2d8de33954443d94b746c952b01434fc58a417fdbd2e84bd5", size = 172823 }, -] - -[[package]] -name = "psutil" -version = "7.2.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/73/cb/09e5184fb5fc0358d110fc3ca7f6b1d033800734d34cac10f4136cfac10e/psutil-7.2.1.tar.gz", hash = "sha256:f7583aec590485b43ca601dd9cea0dcd65bd7bb21d30ef4ddbf4ea6b5ed1bdd3", size = 490253 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/77/8e/f0c242053a368c2aa89584ecd1b054a18683f13d6e5a318fc9ec36582c94/psutil-7.2.1-cp313-cp313t-macosx_10_13_x86_64.whl", hash = "sha256:ba9f33bb525b14c3ea563b2fd521a84d2fa214ec59e3e6a2858f78d0844dd60d", size = 129624 }, - { url = "https://files.pythonhosted.org/packages/26/97/a58a4968f8990617decee234258a2b4fc7cd9e35668387646c1963e69f26/psutil-7.2.1-cp313-cp313t-macosx_11_0_arm64.whl", hash = "sha256:81442dac7abfc2f4f4385ea9e12ddf5a796721c0f6133260687fec5c3780fa49", size = 130132 }, - { url = "https://files.pythonhosted.org/packages/db/6d/ed44901e830739af5f72a85fa7ec5ff1edea7f81bfbf4875e409007149bd/psutil-7.2.1-cp313-cp313t-manylinux2010_x86_64.manylinux_2_12_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:ea46c0d060491051d39f0d2cff4f98d5c72b288289f57a21556cc7d504db37fc", size = 180612 }, - { url = "https://files.pythonhosted.org/packages/c7/65/b628f8459bca4efbfae50d4bf3feaab803de9a160b9d5f3bd9295a33f0c2/psutil-7.2.1-cp313-cp313t-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:35630d5af80d5d0d49cfc4d64c1c13838baf6717a13effb35869a5919b854cdf", size = 183201 }, - { url = "https://files.pythonhosted.org/packages/fb/23/851cadc9764edcc18f0effe7d0bf69f727d4cf2442deb4a9f78d4e4f30f2/psutil-7.2.1-cp313-cp313t-win_amd64.whl", hash = "sha256:923f8653416604e356073e6e0bccbe7c09990acef442def2f5640dd0faa9689f", size = 139081 }, - { url = "https://files.pythonhosted.org/packages/59/82/d63e8494ec5758029f31c6cb06d7d161175d8281e91d011a4a441c8a43b5/psutil-7.2.1-cp313-cp313t-win_arm64.whl", hash = "sha256:cfbe6b40ca48019a51827f20d830887b3107a74a79b01ceb8cc8de4ccb17b672", size = 134767 }, - { url = "https://files.pythonhosted.org/packages/05/c2/5fb764bd61e40e1fe756a44bd4c21827228394c17414ade348e28f83cd79/psutil-7.2.1-cp314-cp314t-macosx_10_15_x86_64.whl", hash = "sha256:494c513ccc53225ae23eec7fe6e1482f1b8a44674241b54561f755a898650679", size = 129716 }, - { url = "https://files.pythonhosted.org/packages/c9/d2/935039c20e06f615d9ca6ca0ab756cf8408a19d298ffaa08666bc18dc805/psutil-7.2.1-cp314-cp314t-macosx_11_0_arm64.whl", hash = "sha256:3fce5f92c22b00cdefd1645aa58ab4877a01679e901555067b1bd77039aa589f", size = 130133 }, - { url = "https://files.pythonhosted.org/packages/77/69/19f1eb0e01d24c2b3eacbc2f78d3b5add8a89bf0bb69465bc8d563cc33de/psutil-7.2.1-cp314-cp314t-manylinux2010_x86_64.manylinux_2_12_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:93f3f7b0bb07711b49626e7940d6fe52aa9940ad86e8f7e74842e73189712129", size = 181518 }, - { url = "https://files.pythonhosted.org/packages/e1/6d/7e18b1b4fa13ad370787626c95887b027656ad4829c156bb6569d02f3262/psutil-7.2.1-cp314-cp314t-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:d34d2ca888208eea2b5c68186841336a7f5e0b990edec929be909353a202768a", size = 184348 }, - { url = "https://files.pythonhosted.org/packages/98/60/1672114392dd879586d60dd97896325df47d9a130ac7401318005aab28ec/psutil-7.2.1-cp314-cp314t-win_amd64.whl", hash = "sha256:2ceae842a78d1603753561132d5ad1b2f8a7979cb0c283f5b52fb4e6e14b1a79", size = 140400 }, - { url = "https://files.pythonhosted.org/packages/fb/7b/d0e9d4513c46e46897b46bcfc410d51fc65735837ea57a25170f298326e6/psutil-7.2.1-cp314-cp314t-win_arm64.whl", hash = "sha256:08a2f175e48a898c8eb8eace45ce01777f4785bc744c90aa2cc7f2fa5462a266", size = 135430 }, - { url = "https://files.pythonhosted.org/packages/c5/cf/5180eb8c8bdf6a503c6919f1da28328bd1e6b3b1b5b9d5b01ae64f019616/psutil-7.2.1-cp36-abi3-macosx_10_9_x86_64.whl", hash = "sha256:b2e953fcfaedcfbc952b44744f22d16575d3aa78eb4f51ae74165b4e96e55f42", size = 128137 }, - { url = "https://files.pythonhosted.org/packages/c5/2c/78e4a789306a92ade5000da4f5de3255202c534acdadc3aac7b5458fadef/psutil-7.2.1-cp36-abi3-macosx_11_0_arm64.whl", hash = "sha256:05cc68dbb8c174828624062e73078e7e35406f4ca2d0866c272c2410d8ef06d1", size = 128947 }, - { url = "https://files.pythonhosted.org/packages/29/f8/40e01c350ad9a2b3cb4e6adbcc8a83b17ee50dd5792102b6142385937db5/psutil-7.2.1-cp36-abi3-manylinux2010_x86_64.manylinux_2_12_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:5e38404ca2bb30ed7267a46c02f06ff842e92da3bb8c5bfdadbd35a5722314d8", size = 154694 }, - { url = "https://files.pythonhosted.org/packages/06/e4/b751cdf839c011a9714a783f120e6a86b7494eb70044d7d81a25a5cd295f/psutil-7.2.1-cp36-abi3-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:ab2b98c9fc19f13f59628d94df5cc4cc4844bc572467d113a8b517d634e362c6", size = 156136 }, - { url = "https://files.pythonhosted.org/packages/44/ad/bbf6595a8134ee1e94a4487af3f132cef7fce43aef4a93b49912a48c3af7/psutil-7.2.1-cp36-abi3-musllinux_1_2_aarch64.whl", hash = "sha256:f78baafb38436d5a128f837fab2d92c276dfb48af01a240b861ae02b2413ada8", size = 148108 }, - { url = "https://files.pythonhosted.org/packages/1c/15/dd6fd869753ce82ff64dcbc18356093471a5a5adf4f77ed1f805d473d859/psutil-7.2.1-cp36-abi3-musllinux_1_2_x86_64.whl", hash = "sha256:99a4cd17a5fdd1f3d014396502daa70b5ec21bf4ffe38393e152f8e449757d67", size = 147402 }, - { url = "https://files.pythonhosted.org/packages/34/68/d9317542e3f2b180c4306e3f45d3c922d7e86d8ce39f941bb9e2e9d8599e/psutil-7.2.1-cp37-abi3-win_amd64.whl", hash = "sha256:b1b0671619343aa71c20ff9767eced0483e4fc9e1f489d50923738caf6a03c17", size = 136938 }, - { url = "https://files.pythonhosted.org/packages/3e/73/2ce007f4198c80fcf2cb24c169884f833fe93fbc03d55d302627b094ee91/psutil-7.2.1-cp37-abi3-win_arm64.whl", hash = "sha256:0d67c1822c355aa6f7314d92018fb4268a76668a536f133599b91edd48759442", size = 133836 }, -] - -[[package]] -name = "ptyprocess" -version = "0.7.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/20/e5/16ff212c1e452235a90aeb09066144d0c5a6a8c0834397e03f5224495c4e/ptyprocess-0.7.0.tar.gz", hash = "sha256:5c5d0a3b48ceee0b48485e0c26037c0acd7d29765ca3fbb5cb3831d347423220", size = 70762 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/22/a6/858897256d0deac81a172289110f31629fc4cee19b6f01283303e18c8db3/ptyprocess-0.7.0-py2.py3-none-any.whl", hash = "sha256:4b41f3967fce3af57cc7e94b888626c18bf37a083e3651ca8feeb66d492fef35", size = 13993 }, -] - -[[package]] -name = "pure-eval" -version = "0.2.3" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/cd/05/0a34433a064256a578f1783a10da6df098ceaa4a57bbeaa96a6c0352786b/pure_eval-0.2.3.tar.gz", hash = "sha256:5f4e983f40564c576c7c8635ae88db5956bb2229d7e9237d03b3c0b0190eaf42", size = 19752 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/8e/37/efad0257dc6e593a18957422533ff0f87ede7c9c6ea010a2177d738fb82f/pure_eval-0.2.3-py3-none-any.whl", hash = "sha256:1db8e35b67b3d218d818ae653e27f06c3aa420901fa7b081ca98cbedc874e0d0", size = 11842 }, -] - -[[package]] -name = "pyaml" -version = "25.7.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "pyyaml" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/c4/01/41f63d66a801a561c9e335523516bd5f761bc43cc61f8b75918306bf2da8/pyaml-25.7.0.tar.gz", hash = "sha256:e113a64ec16881bf2b092e2beb84b7dcf1bd98096ad17f5f14e8fb782a75d99b", size = 29814 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/a8/ee/a878f2ad010cbccb311f947f0f2f09d38f613938ee28c34e60fceecc75a1/pyaml-25.7.0-py3-none-any.whl", hash = "sha256:ce5d7867cc2b455efdb9b0448324ff7b9f74d99f64650f12ca570102db6b985f", size = 26418 }, -] - -[[package]] -name = "pyannote-core" -version = "5.0.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "numpy" }, - { name = "scipy", version = "1.15.3", source = { registry = "https://pypi.org/simple" }, marker = "python_full_version < '3.11'" }, - { name = "scipy", version = "1.16.3", source = { registry = "https://pypi.org/simple" }, marker = "python_full_version >= '3.11'" }, - { name = "sortedcontainers" }, - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/65/03/feaf7534206f02c75baf151ce4b8c322b402a6f477c2be82f69d9269cbe6/pyannote.core-5.0.0.tar.gz", hash = "sha256:1a55bcc8bd680ba6be5fa53efa3b6f3d2cdd67144c07b6b4d8d66d5cb0d2096f", size = 59247 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/84/c4/370bc8ba66815a5832ece753a1009388bb07ea353d21c83f2d5a1a436f2c/pyannote.core-5.0.0-py3-none-any.whl", hash = "sha256:04920a6754492242ce0dc6017545595ab643870fe69a994f20c1a5f2da0544d0", size = 58475 }, -] - -[[package]] -name = "pyannote-database" -version = "5.1.3" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "pandas" }, - { name = "pyannote-core" }, - { name = "pyyaml" }, - { name = "typer" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/a9/ae/de36413d69a46be87cb612ebbcdc4eacbeebce3bc809124603e44a88fe26/pyannote.database-5.1.3.tar.gz", hash = "sha256:0eaf64c1cc506718de60d2d702f1359b1ae7ff252ee3e4799f1c5e378cd52c31", size = 49957 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/a1/64/92d51a3a05615ba58be8ba62a43f9f9f952d9f3646f7e4fb7826e5a3a24e/pyannote.database-5.1.3-py3-none-any.whl", hash = "sha256:37887844c7dfbcc075cb591eddc00aff45fae1ed905344e1f43e0090e63bd40a", size = 48127 }, -] - -[[package]] -name = "pyannote-metrics" -version = "3.2.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "docopt" }, - { name = "matplotlib" }, - { name = "numpy" }, - { name = "pandas" }, - { name = "pyannote-core" }, - { name = "pyannote-database" }, - { name = "scikit-learn", version = "1.7.2", source = { registry = "https://pypi.org/simple" }, marker = "python_full_version < '3.11'" }, - { name = "scikit-learn", version = "1.8.0", source = { registry = "https://pypi.org/simple" }, marker = "python_full_version >= '3.11'" }, - { name = "scipy", version = "1.15.3", source = { registry = "https://pypi.org/simple" }, marker = "python_full_version < '3.11'" }, - { name = "scipy", version = "1.16.3", source = { registry = "https://pypi.org/simple" }, marker = "python_full_version >= '3.11'" }, - { name = "sympy" }, - { name = "tabulate" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/39/2b/6c5f01d3c49aa1c160765946e23782ca6436ae8b9bc514b56319ff5f16e7/pyannote.metrics-3.2.1.tar.gz", hash = "sha256:08024255a3550e96a8e9da4f5f4af326886548480de891414567c8900920ee5c", size = 49086 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/6c/7d/035b370ab834b30e849fe9cd092b7bd7f321fcc4a2c56b84e96476b7ede5/pyannote.metrics-3.2.1-py3-none-any.whl", hash = "sha256:46be797cdade26c82773e5018659ae610145260069c7c5bf3d3c8a029ade8e22", size = 51386 }, -] - -[[package]] -name = "pyarrow" -version = "22.0.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/30/53/04a7fdc63e6056116c9ddc8b43bc28c12cdd181b85cbeadb79278475f3ae/pyarrow-22.0.0.tar.gz", hash = "sha256:3d600dc583260d845c7d8a6db540339dd883081925da2bd1c5cb808f720b3cd9", size = 1151151 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/d9/9b/cb3f7e0a345353def531ca879053e9ef6b9f38ed91aebcf68b09ba54dec0/pyarrow-22.0.0-cp310-cp310-macosx_12_0_arm64.whl", hash = "sha256:77718810bd3066158db1e95a63c160ad7ce08c6b0710bc656055033e39cdad88", size = 34223968 }, - { url = "https://files.pythonhosted.org/packages/6c/41/3184b8192a120306270c5307f105b70320fdaa592c99843c5ef78aaefdcf/pyarrow-22.0.0-cp310-cp310-macosx_12_0_x86_64.whl", hash = "sha256:44d2d26cda26d18f7af7db71453b7b783788322d756e81730acb98f24eb90ace", size = 35942085 }, - { url = "https://files.pythonhosted.org/packages/d9/3d/a1eab2f6f08001f9fb714b8ed5cfb045e2fe3e3e3c0c221f2c9ed1e6d67d/pyarrow-22.0.0-cp310-cp310-manylinux_2_28_aarch64.whl", hash = "sha256:b9d71701ce97c95480fecb0039ec5bb889e75f110da72005743451339262f4ce", size = 44964613 }, - { url = "https://files.pythonhosted.org/packages/46/46/a1d9c24baf21cfd9ce994ac820a24608decf2710521b29223d4334985127/pyarrow-22.0.0-cp310-cp310-manylinux_2_28_x86_64.whl", hash = "sha256:710624ab925dc2b05a6229d47f6f0dac1c1155e6ed559be7109f684eba048a48", size = 47627059 }, - { url = "https://files.pythonhosted.org/packages/3a/4c/f711acb13075c1391fd54bc17e078587672c575f8de2a6e62509af026dcf/pyarrow-22.0.0-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:f963ba8c3b0199f9d6b794c90ec77545e05eadc83973897a4523c9e8d84e9340", size = 47947043 }, - { url = "https://files.pythonhosted.org/packages/4e/70/1f3180dd7c2eab35c2aca2b29ace6c519f827dcd4cfeb8e0dca41612cf7a/pyarrow-22.0.0-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:bd0d42297ace400d8febe55f13fdf46e86754842b860c978dfec16f081e5c653", size = 50206505 }, - { url = "https://files.pythonhosted.org/packages/80/07/fea6578112c8c60ffde55883a571e4c4c6bc7049f119d6b09333b5cc6f73/pyarrow-22.0.0-cp310-cp310-win_amd64.whl", hash = "sha256:00626d9dc0f5ef3a75fe63fd68b9c7c8302d2b5bbc7f74ecaedba83447a24f84", size = 28101641 }, - { url = "https://files.pythonhosted.org/packages/2e/b7/18f611a8cdc43417f9394a3ccd3eace2f32183c08b9eddc3d17681819f37/pyarrow-22.0.0-cp311-cp311-macosx_12_0_arm64.whl", hash = "sha256:3e294c5eadfb93d78b0763e859a0c16d4051fc1c5231ae8956d61cb0b5666f5a", size = 34272022 }, - { url = "https://files.pythonhosted.org/packages/26/5c/f259e2526c67eb4b9e511741b19870a02363a47a35edbebc55c3178db22d/pyarrow-22.0.0-cp311-cp311-macosx_12_0_x86_64.whl", hash = "sha256:69763ab2445f632d90b504a815a2a033f74332997052b721002298ed6de40f2e", size = 35995834 }, - { url = "https://files.pythonhosted.org/packages/50/8d/281f0f9b9376d4b7f146913b26fac0aa2829cd1ee7e997f53a27411bbb92/pyarrow-22.0.0-cp311-cp311-manylinux_2_28_aarch64.whl", hash = "sha256:b41f37cabfe2463232684de44bad753d6be08a7a072f6a83447eeaf0e4d2a215", size = 45030348 }, - { url = "https://files.pythonhosted.org/packages/f5/e5/53c0a1c428f0976bf22f513d79c73000926cb00b9c138d8e02daf2102e18/pyarrow-22.0.0-cp311-cp311-manylinux_2_28_x86_64.whl", hash = "sha256:35ad0f0378c9359b3f297299c3309778bb03b8612f987399a0333a560b43862d", size = 47699480 }, - { url = "https://files.pythonhosted.org/packages/95/e1/9dbe4c465c3365959d183e6345d0a8d1dc5b02ca3f8db4760b3bc834cf25/pyarrow-22.0.0-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:8382ad21458075c2e66a82a29d650f963ce51c7708c7c0ff313a8c206c4fd5e8", size = 48011148 }, - { url = "https://files.pythonhosted.org/packages/c5/b4/7caf5d21930061444c3cf4fa7535c82faf5263e22ce43af7c2759ceb5b8b/pyarrow-22.0.0-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:1a812a5b727bc09c3d7ea072c4eebf657c2f7066155506ba31ebf4792f88f016", size = 50276964 }, - { url = "https://files.pythonhosted.org/packages/ae/f3/cec89bd99fa3abf826f14d4e53d3d11340ce6f6af4d14bdcd54cd83b6576/pyarrow-22.0.0-cp311-cp311-win_amd64.whl", hash = "sha256:ec5d40dd494882704fb876c16fa7261a69791e784ae34e6b5992e977bd2e238c", size = 28106517 }, - { url = "https://files.pythonhosted.org/packages/af/63/ba23862d69652f85b615ca14ad14f3bcfc5bf1b99ef3f0cd04ff93fdad5a/pyarrow-22.0.0-cp312-cp312-macosx_12_0_arm64.whl", hash = "sha256:bea79263d55c24a32b0d79c00a1c58bb2ee5f0757ed95656b01c0fb310c5af3d", size = 34211578 }, - { url = "https://files.pythonhosted.org/packages/b1/d0/f9ad86fe809efd2bcc8be32032fa72e8b0d112b01ae56a053006376c5930/pyarrow-22.0.0-cp312-cp312-macosx_12_0_x86_64.whl", hash = "sha256:12fe549c9b10ac98c91cf791d2945e878875d95508e1a5d14091a7aaa66d9cf8", size = 35989906 }, - { url = "https://files.pythonhosted.org/packages/b4/a8/f910afcb14630e64d673f15904ec27dd31f1e009b77033c365c84e8c1e1d/pyarrow-22.0.0-cp312-cp312-manylinux_2_28_aarch64.whl", hash = "sha256:334f900ff08ce0423407af97e6c26ad5d4e3b0763645559ece6fbf3747d6a8f5", size = 45021677 }, - { url = "https://files.pythonhosted.org/packages/13/95/aec81f781c75cd10554dc17a25849c720d54feafb6f7847690478dcf5ef8/pyarrow-22.0.0-cp312-cp312-manylinux_2_28_x86_64.whl", hash = "sha256:c6c791b09c57ed76a18b03f2631753a4960eefbbca80f846da8baefc6491fcfe", size = 47726315 }, - { url = "https://files.pythonhosted.org/packages/bb/d4/74ac9f7a54cfde12ee42734ea25d5a3c9a45db78f9def949307a92720d37/pyarrow-22.0.0-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:c3200cb41cdbc65156e5f8c908d739b0dfed57e890329413da2748d1a2cd1a4e", size = 47990906 }, - { url = "https://files.pythonhosted.org/packages/2e/71/fedf2499bf7a95062eafc989ace56572f3343432570e1c54e6599d5b88da/pyarrow-22.0.0-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:ac93252226cf288753d8b46280f4edf3433bf9508b6977f8dd8526b521a1bbb9", size = 50306783 }, - { url = "https://files.pythonhosted.org/packages/68/ed/b202abd5a5b78f519722f3d29063dda03c114711093c1995a33b8e2e0f4b/pyarrow-22.0.0-cp312-cp312-win_amd64.whl", hash = "sha256:44729980b6c50a5f2bfcc2668d36c569ce17f8b17bccaf470c4313dcbbf13c9d", size = 27972883 }, - { url = "https://files.pythonhosted.org/packages/a6/d6/d0fac16a2963002fc22c8fa75180a838737203d558f0ed3b564c4a54eef5/pyarrow-22.0.0-cp313-cp313-macosx_12_0_arm64.whl", hash = "sha256:e6e95176209257803a8b3d0394f21604e796dadb643d2f7ca21b66c9c0b30c9a", size = 34204629 }, - { url = "https://files.pythonhosted.org/packages/c6/9c/1d6357347fbae062ad3f17082f9ebc29cc733321e892c0d2085f42a2212b/pyarrow-22.0.0-cp313-cp313-macosx_12_0_x86_64.whl", hash = "sha256:001ea83a58024818826a9e3f89bf9310a114f7e26dfe404a4c32686f97bd7901", size = 35985783 }, - { url = "https://files.pythonhosted.org/packages/ff/c0/782344c2ce58afbea010150df07e3a2f5fdad299cd631697ae7bd3bac6e3/pyarrow-22.0.0-cp313-cp313-manylinux_2_28_aarch64.whl", hash = "sha256:ce20fe000754f477c8a9125543f1936ea5b8867c5406757c224d745ed033e691", size = 45020999 }, - { url = "https://files.pythonhosted.org/packages/1b/8b/5362443737a5307a7b67c1017c42cd104213189b4970bf607e05faf9c525/pyarrow-22.0.0-cp313-cp313-manylinux_2_28_x86_64.whl", hash = "sha256:e0a15757fccb38c410947df156f9749ae4a3c89b2393741a50521f39a8cf202a", size = 47724601 }, - { url = "https://files.pythonhosted.org/packages/69/4d/76e567a4fc2e190ee6072967cb4672b7d9249ac59ae65af2d7e3047afa3b/pyarrow-22.0.0-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:cedb9dd9358e4ea1d9bce3665ce0797f6adf97ff142c8e25b46ba9cdd508e9b6", size = 48001050 }, - { url = "https://files.pythonhosted.org/packages/01/5e/5653f0535d2a1aef8223cee9d92944cb6bccfee5cf1cd3f462d7cb022790/pyarrow-22.0.0-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:252be4a05f9d9185bb8c18e83764ebcfea7185076c07a7a662253af3a8c07941", size = 50307877 }, - { url = "https://files.pythonhosted.org/packages/2d/f8/1d0bd75bf9328a3b826e24a16e5517cd7f9fbf8d34a3184a4566ef5a7f29/pyarrow-22.0.0-cp313-cp313-win_amd64.whl", hash = "sha256:a4893d31e5ef780b6edcaf63122df0f8d321088bb0dee4c8c06eccb1ca28d145", size = 27977099 }, - { url = "https://files.pythonhosted.org/packages/90/81/db56870c997805bf2b0f6eeeb2d68458bf4654652dccdcf1bf7a42d80903/pyarrow-22.0.0-cp313-cp313t-macosx_12_0_arm64.whl", hash = "sha256:f7fe3dbe871294ba70d789be16b6e7e52b418311e166e0e3cba9522f0f437fb1", size = 34336685 }, - { url = "https://files.pythonhosted.org/packages/1c/98/0727947f199aba8a120f47dfc229eeb05df15bcd7a6f1b669e9f882afc58/pyarrow-22.0.0-cp313-cp313t-macosx_12_0_x86_64.whl", hash = "sha256:ba95112d15fd4f1105fb2402c4eab9068f0554435e9b7085924bcfaac2cc306f", size = 36032158 }, - { url = "https://files.pythonhosted.org/packages/96/b4/9babdef9c01720a0785945c7cf550e4acd0ebcd7bdd2e6f0aa7981fa85e2/pyarrow-22.0.0-cp313-cp313t-manylinux_2_28_aarch64.whl", hash = "sha256:c064e28361c05d72eed8e744c9605cbd6d2bb7481a511c74071fd9b24bc65d7d", size = 44892060 }, - { url = "https://files.pythonhosted.org/packages/f8/ca/2f8804edd6279f78a37062d813de3f16f29183874447ef6d1aadbb4efa0f/pyarrow-22.0.0-cp313-cp313t-manylinux_2_28_x86_64.whl", hash = "sha256:6f9762274496c244d951c819348afbcf212714902742225f649cf02823a6a10f", size = 47504395 }, - { url = "https://files.pythonhosted.org/packages/b9/f0/77aa5198fd3943682b2e4faaf179a674f0edea0d55d326d83cb2277d9363/pyarrow-22.0.0-cp313-cp313t-musllinux_1_2_aarch64.whl", hash = "sha256:a9d9ffdc2ab696f6b15b4d1f7cec6658e1d788124418cb30030afbae31c64746", size = 48066216 }, - { url = "https://files.pythonhosted.org/packages/79/87/a1937b6e78b2aff18b706d738c9e46ade5bfcf11b294e39c87706a0089ac/pyarrow-22.0.0-cp313-cp313t-musllinux_1_2_x86_64.whl", hash = "sha256:ec1a15968a9d80da01e1d30349b2b0d7cc91e96588ee324ce1b5228175043e95", size = 50288552 }, - { url = "https://files.pythonhosted.org/packages/60/ae/b5a5811e11f25788ccfdaa8f26b6791c9807119dffcf80514505527c384c/pyarrow-22.0.0-cp313-cp313t-win_amd64.whl", hash = "sha256:bba208d9c7decf9961998edf5c65e3ea4355d5818dd6cd0f6809bec1afb951cc", size = 28262504 }, - { url = "https://files.pythonhosted.org/packages/bd/b0/0fa4d28a8edb42b0a7144edd20befd04173ac79819547216f8a9f36f9e50/pyarrow-22.0.0-cp314-cp314-macosx_12_0_arm64.whl", hash = "sha256:9bddc2cade6561f6820d4cd73f99a0243532ad506bc510a75a5a65a522b2d74d", size = 34224062 }, - { url = "https://files.pythonhosted.org/packages/0f/a8/7a719076b3c1be0acef56a07220c586f25cd24de0e3f3102b438d18ae5df/pyarrow-22.0.0-cp314-cp314-macosx_12_0_x86_64.whl", hash = "sha256:e70ff90c64419709d38c8932ea9fe1cc98415c4f87ea8da81719e43f02534bc9", size = 35990057 }, - { url = "https://files.pythonhosted.org/packages/89/3c/359ed54c93b47fb6fe30ed16cdf50e3f0e8b9ccfb11b86218c3619ae50a8/pyarrow-22.0.0-cp314-cp314-manylinux_2_28_aarch64.whl", hash = "sha256:92843c305330aa94a36e706c16209cd4df274693e777ca47112617db7d0ef3d7", size = 45068002 }, - { url = "https://files.pythonhosted.org/packages/55/fc/4945896cc8638536ee787a3bd6ce7cec8ec9acf452d78ec39ab328efa0a1/pyarrow-22.0.0-cp314-cp314-manylinux_2_28_x86_64.whl", hash = "sha256:6dda1ddac033d27421c20d7a7943eec60be44e0db4e079f33cc5af3b8280ccde", size = 47737765 }, - { url = "https://files.pythonhosted.org/packages/cd/5e/7cb7edeb2abfaa1f79b5d5eb89432356155c8426f75d3753cbcb9592c0fd/pyarrow-22.0.0-cp314-cp314-musllinux_1_2_aarch64.whl", hash = "sha256:84378110dd9a6c06323b41b56e129c504d157d1a983ce8f5443761eb5256bafc", size = 48048139 }, - { url = "https://files.pythonhosted.org/packages/88/c6/546baa7c48185f5e9d6e59277c4b19f30f48c94d9dd938c2a80d4d6b067c/pyarrow-22.0.0-cp314-cp314-musllinux_1_2_x86_64.whl", hash = "sha256:854794239111d2b88b40b6ef92aa478024d1e5074f364033e73e21e3f76b25e0", size = 50314244 }, - { url = "https://files.pythonhosted.org/packages/3c/79/755ff2d145aafec8d347bf18f95e4e81c00127f06d080135dfc86aea417c/pyarrow-22.0.0-cp314-cp314-win_amd64.whl", hash = "sha256:b883fe6fd85adad7932b3271c38ac289c65b7337c2c132e9569f9d3940620730", size = 28757501 }, - { url = "https://files.pythonhosted.org/packages/0e/d2/237d75ac28ced3147912954e3c1a174df43a95f4f88e467809118a8165e0/pyarrow-22.0.0-cp314-cp314t-macosx_12_0_arm64.whl", hash = "sha256:7a820d8ae11facf32585507c11f04e3f38343c1e784c9b5a8b1da5c930547fe2", size = 34355506 }, - { url = "https://files.pythonhosted.org/packages/1e/2c/733dfffe6d3069740f98e57ff81007809067d68626c5faef293434d11bd6/pyarrow-22.0.0-cp314-cp314t-macosx_12_0_x86_64.whl", hash = "sha256:c6ec3675d98915bf1ec8b3c7986422682f7232ea76cad276f4c8abd5b7319b70", size = 36047312 }, - { url = "https://files.pythonhosted.org/packages/7c/2b/29d6e3782dc1f299727462c1543af357a0f2c1d3c160ce199950d9ca51eb/pyarrow-22.0.0-cp314-cp314t-manylinux_2_28_aarch64.whl", hash = "sha256:3e739edd001b04f654b166204fc7a9de896cf6007eaff33409ee9e50ceaff754", size = 45081609 }, - { url = "https://files.pythonhosted.org/packages/8d/42/aa9355ecc05997915af1b7b947a7f66c02dcaa927f3203b87871c114ba10/pyarrow-22.0.0-cp314-cp314t-manylinux_2_28_x86_64.whl", hash = "sha256:7388ac685cab5b279a41dfe0a6ccd99e4dbf322edfb63e02fc0443bf24134e91", size = 47703663 }, - { url = "https://files.pythonhosted.org/packages/ee/62/45abedde480168e83a1de005b7b7043fd553321c1e8c5a9a114425f64842/pyarrow-22.0.0-cp314-cp314t-musllinux_1_2_aarch64.whl", hash = "sha256:f633074f36dbc33d5c05b5dc75371e5660f1dbf9c8b1d95669def05e5425989c", size = 48066543 }, - { url = "https://files.pythonhosted.org/packages/84/e9/7878940a5b072e4f3bf998770acafeae13b267f9893af5f6d4ab3904b67e/pyarrow-22.0.0-cp314-cp314t-musllinux_1_2_x86_64.whl", hash = "sha256:4c19236ae2402a8663a2c8f21f1870a03cc57f0bef7e4b6eb3238cc82944de80", size = 50288838 }, - { url = "https://files.pythonhosted.org/packages/7b/03/f335d6c52b4a4761bcc83499789a1e2e16d9d201a58c327a9b5cc9a41bd9/pyarrow-22.0.0-cp314-cp314t-win_amd64.whl", hash = "sha256:0c34fe18094686194f204a3b1787a27456897d8a2d62caf84b61e8dfbc0252ae", size = 29185594 }, -] - -[[package]] -name = "pycparser" -version = "2.23" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/fe/cf/d2d3b9f5699fb1e4615c8e32ff220203e43b248e1dfcc6736ad9057731ca/pycparser-2.23.tar.gz", hash = "sha256:78816d4f24add8f10a06d6f05b4d424ad9e96cfebf68a4ddc99c65c0720d00c2", size = 173734 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/a0/e3/59cd50310fc9b59512193629e1984c1f95e5c8ae6e5d8c69532ccc65a7fe/pycparser-2.23-py3-none-any.whl", hash = "sha256:e5c6e8d3fbad53479cab09ac03729e0a9faf2bee3db8208a550daf5af81a5934", size = 118140 }, -] - -[[package]] -name = "pydantic" -version = "2.12.5" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "annotated-types" }, - { name = "pydantic-core" }, - { name = "typing-extensions" }, - { name = "typing-inspection" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/69/44/36f1a6e523abc58ae5f928898e4aca2e0ea509b5aa6f6f392a5d882be928/pydantic-2.12.5.tar.gz", hash = "sha256:4d351024c75c0f085a9febbb665ce8c0c6ec5d30e903bdb6394b7ede26aebb49", size = 821591 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/5a/87/b70ad306ebb6f9b585f114d0ac2137d792b48be34d732d60e597c2f8465a/pydantic-2.12.5-py3-none-any.whl", hash = "sha256:e561593fccf61e8a20fc46dfc2dfe075b8be7d0188df33f221ad1f0139180f9d", size = 463580 }, -] - -[[package]] -name = "pydantic-core" -version = "2.41.5" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/71/70/23b021c950c2addd24ec408e9ab05d59b035b39d97cdc1130e1bce647bb6/pydantic_core-2.41.5.tar.gz", hash = "sha256:08daa51ea16ad373ffd5e7606252cc32f07bc72b28284b6bc9c6df804816476e", size = 460952 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/c6/90/32c9941e728d564b411d574d8ee0cf09b12ec978cb22b294995bae5549a5/pydantic_core-2.41.5-cp310-cp310-macosx_10_12_x86_64.whl", hash = "sha256:77b63866ca88d804225eaa4af3e664c5faf3568cea95360d21f4725ab6e07146", size = 2107298 }, - { url = "https://files.pythonhosted.org/packages/fb/a8/61c96a77fe28993d9a6fb0f4127e05430a267b235a124545d79fea46dd65/pydantic_core-2.41.5-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:dfa8a0c812ac681395907e71e1274819dec685fec28273a28905df579ef137e2", size = 1901475 }, - { url = "https://files.pythonhosted.org/packages/5d/b6/338abf60225acc18cdc08b4faef592d0310923d19a87fba1faf05af5346e/pydantic_core-2.41.5-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5921a4d3ca3aee735d9fd163808f5e8dd6c6972101e4adbda9a4667908849b97", size = 1918815 }, - { url = "https://files.pythonhosted.org/packages/d1/1c/2ed0433e682983d8e8cba9c8d8ef274d4791ec6a6f24c58935b90e780e0a/pydantic_core-2.41.5-cp310-cp310-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:e25c479382d26a2a41b7ebea1043564a937db462816ea07afa8a44c0866d52f9", size = 2065567 }, - { url = "https://files.pythonhosted.org/packages/b3/24/cf84974ee7d6eae06b9e63289b7b8f6549d416b5c199ca2d7ce13bbcf619/pydantic_core-2.41.5-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:f547144f2966e1e16ae626d8ce72b4cfa0caedc7fa28052001c94fb2fcaa1c52", size = 2230442 }, - { url = "https://files.pythonhosted.org/packages/fd/21/4e287865504b3edc0136c89c9c09431be326168b1eb7841911cbc877a995/pydantic_core-2.41.5-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:6f52298fbd394f9ed112d56f3d11aabd0d5bd27beb3084cc3d8ad069483b8941", size = 2350956 }, - { url = "https://files.pythonhosted.org/packages/a8/76/7727ef2ffa4b62fcab916686a68a0426b9b790139720e1934e8ba797e238/pydantic_core-2.41.5-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:100baa204bb412b74fe285fb0f3a385256dad1d1879f0a5cb1499ed2e83d132a", size = 2068253 }, - { url = "https://files.pythonhosted.org/packages/d5/8c/a4abfc79604bcb4c748e18975c44f94f756f08fb04218d5cb87eb0d3a63e/pydantic_core-2.41.5-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:05a2c8852530ad2812cb7914dc61a1125dc4e06252ee98e5638a12da6cc6fb6c", size = 2177050 }, - { url = "https://files.pythonhosted.org/packages/67/b1/de2e9a9a79b480f9cb0b6e8b6ba4c50b18d4e89852426364c66aa82bb7b3/pydantic_core-2.41.5-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:29452c56df2ed968d18d7e21f4ab0ac55e71dc59524872f6fc57dcf4a3249ed2", size = 2147178 }, - { url = "https://files.pythonhosted.org/packages/16/c1/dfb33f837a47b20417500efaa0378adc6635b3c79e8369ff7a03c494b4ac/pydantic_core-2.41.5-cp310-cp310-musllinux_1_1_armv7l.whl", hash = "sha256:d5160812ea7a8a2ffbe233d8da666880cad0cbaf5d4de74ae15c313213d62556", size = 2341833 }, - { url = "https://files.pythonhosted.org/packages/47/36/00f398642a0f4b815a9a558c4f1dca1b4020a7d49562807d7bc9ff279a6c/pydantic_core-2.41.5-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:df3959765b553b9440adfd3c795617c352154e497a4eaf3752555cfb5da8fc49", size = 2321156 }, - { url = "https://files.pythonhosted.org/packages/7e/70/cad3acd89fde2010807354d978725ae111ddf6d0ea46d1ea1775b5c1bd0c/pydantic_core-2.41.5-cp310-cp310-win32.whl", hash = "sha256:1f8d33a7f4d5a7889e60dc39856d76d09333d8a6ed0f5f1190635cbec70ec4ba", size = 1989378 }, - { url = "https://files.pythonhosted.org/packages/76/92/d338652464c6c367e5608e4488201702cd1cbb0f33f7b6a85a60fe5f3720/pydantic_core-2.41.5-cp310-cp310-win_amd64.whl", hash = "sha256:62de39db01b8d593e45871af2af9e497295db8d73b085f6bfd0b18c83c70a8f9", size = 2013622 }, - { url = "https://files.pythonhosted.org/packages/e8/72/74a989dd9f2084b3d9530b0915fdda64ac48831c30dbf7c72a41a5232db8/pydantic_core-2.41.5-cp311-cp311-macosx_10_12_x86_64.whl", hash = "sha256:a3a52f6156e73e7ccb0f8cced536adccb7042be67cb45f9562e12b319c119da6", size = 2105873 }, - { url = "https://files.pythonhosted.org/packages/12/44/37e403fd9455708b3b942949e1d7febc02167662bf1a7da5b78ee1ea2842/pydantic_core-2.41.5-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:7f3bf998340c6d4b0c9a2f02d6a400e51f123b59565d74dc60d252ce888c260b", size = 1899826 }, - { url = "https://files.pythonhosted.org/packages/33/7f/1d5cab3ccf44c1935a359d51a8a2a9e1a654b744b5e7f80d41b88d501eec/pydantic_core-2.41.5-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:378bec5c66998815d224c9ca994f1e14c0c21cb95d2f52b6021cc0b2a58f2a5a", size = 1917869 }, - { url = "https://files.pythonhosted.org/packages/6e/6a/30d94a9674a7fe4f4744052ed6c5e083424510be1e93da5bc47569d11810/pydantic_core-2.41.5-cp311-cp311-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:e7b576130c69225432866fe2f4a469a85a54ade141d96fd396dffcf607b558f8", size = 2063890 }, - { url = "https://files.pythonhosted.org/packages/50/be/76e5d46203fcb2750e542f32e6c371ffa9b8ad17364cf94bb0818dbfb50c/pydantic_core-2.41.5-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:6cb58b9c66f7e4179a2d5e0f849c48eff5c1fca560994d6eb6543abf955a149e", size = 2229740 }, - { url = "https://files.pythonhosted.org/packages/d3/ee/fed784df0144793489f87db310a6bbf8118d7b630ed07aa180d6067e653a/pydantic_core-2.41.5-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:88942d3a3dff3afc8288c21e565e476fc278902ae4d6d134f1eeda118cc830b1", size = 2350021 }, - { url = "https://files.pythonhosted.org/packages/c8/be/8fed28dd0a180dca19e72c233cbf58efa36df055e5b9d90d64fd1740b828/pydantic_core-2.41.5-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:f31d95a179f8d64d90f6831d71fa93290893a33148d890ba15de25642c5d075b", size = 2066378 }, - { url = "https://files.pythonhosted.org/packages/b0/3b/698cf8ae1d536a010e05121b4958b1257f0b5522085e335360e53a6b1c8b/pydantic_core-2.41.5-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:c1df3d34aced70add6f867a8cf413e299177e0c22660cc767218373d0779487b", size = 2175761 }, - { url = "https://files.pythonhosted.org/packages/b8/ba/15d537423939553116dea94ce02f9c31be0fa9d0b806d427e0308ec17145/pydantic_core-2.41.5-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:4009935984bd36bd2c774e13f9a09563ce8de4abaa7226f5108262fa3e637284", size = 2146303 }, - { url = "https://files.pythonhosted.org/packages/58/7f/0de669bf37d206723795f9c90c82966726a2ab06c336deba4735b55af431/pydantic_core-2.41.5-cp311-cp311-musllinux_1_1_armv7l.whl", hash = "sha256:34a64bc3441dc1213096a20fe27e8e128bd3ff89921706e83c0b1ac971276594", size = 2340355 }, - { url = "https://files.pythonhosted.org/packages/e5/de/e7482c435b83d7e3c3ee5ee4451f6e8973cff0eb6007d2872ce6383f6398/pydantic_core-2.41.5-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:c9e19dd6e28fdcaa5a1de679aec4141f691023916427ef9bae8584f9c2fb3b0e", size = 2319875 }, - { url = "https://files.pythonhosted.org/packages/fe/e6/8c9e81bb6dd7560e33b9053351c29f30c8194b72f2d6932888581f503482/pydantic_core-2.41.5-cp311-cp311-win32.whl", hash = "sha256:2c010c6ded393148374c0f6f0bf89d206bf3217f201faa0635dcd56bd1520f6b", size = 1987549 }, - { url = "https://files.pythonhosted.org/packages/11/66/f14d1d978ea94d1bc21fc98fcf570f9542fe55bfcc40269d4e1a21c19bf7/pydantic_core-2.41.5-cp311-cp311-win_amd64.whl", hash = "sha256:76ee27c6e9c7f16f47db7a94157112a2f3a00e958bc626e2f4ee8bec5c328fbe", size = 2011305 }, - { url = "https://files.pythonhosted.org/packages/56/d8/0e271434e8efd03186c5386671328154ee349ff0354d83c74f5caaf096ed/pydantic_core-2.41.5-cp311-cp311-win_arm64.whl", hash = "sha256:4bc36bbc0b7584de96561184ad7f012478987882ebf9f9c389b23f432ea3d90f", size = 1972902 }, - { url = "https://files.pythonhosted.org/packages/5f/5d/5f6c63eebb5afee93bcaae4ce9a898f3373ca23df3ccaef086d0233a35a7/pydantic_core-2.41.5-cp312-cp312-macosx_10_12_x86_64.whl", hash = "sha256:f41a7489d32336dbf2199c8c0a215390a751c5b014c2c1c5366e817202e9cdf7", size = 2110990 }, - { url = "https://files.pythonhosted.org/packages/aa/32/9c2e8ccb57c01111e0fd091f236c7b371c1bccea0fa85247ac55b1e2b6b6/pydantic_core-2.41.5-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:070259a8818988b9a84a449a2a7337c7f430a22acc0859c6b110aa7212a6d9c0", size = 1896003 }, - { url = "https://files.pythonhosted.org/packages/68/b8/a01b53cb0e59139fbc9e4fda3e9724ede8de279097179be4ff31f1abb65a/pydantic_core-2.41.5-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e96cea19e34778f8d59fe40775a7a574d95816eb150850a85a7a4c8f4b94ac69", size = 1919200 }, - { url = "https://files.pythonhosted.org/packages/38/de/8c36b5198a29bdaade07b5985e80a233a5ac27137846f3bc2d3b40a47360/pydantic_core-2.41.5-cp312-cp312-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:ed2e99c456e3fadd05c991f8f437ef902e00eedf34320ba2b0842bd1c3ca3a75", size = 2052578 }, - { url = "https://files.pythonhosted.org/packages/00/b5/0e8e4b5b081eac6cb3dbb7e60a65907549a1ce035a724368c330112adfdd/pydantic_core-2.41.5-cp312-cp312-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:65840751b72fbfd82c3c640cff9284545342a4f1eb1586ad0636955b261b0b05", size = 2208504 }, - { url = "https://files.pythonhosted.org/packages/77/56/87a61aad59c7c5b9dc8caad5a41a5545cba3810c3e828708b3d7404f6cef/pydantic_core-2.41.5-cp312-cp312-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:e536c98a7626a98feb2d3eaf75944ef6f3dbee447e1f841eae16f2f0a72d8ddc", size = 2335816 }, - { url = "https://files.pythonhosted.org/packages/0d/76/941cc9f73529988688a665a5c0ecff1112b3d95ab48f81db5f7606f522d3/pydantic_core-2.41.5-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:eceb81a8d74f9267ef4081e246ffd6d129da5d87e37a77c9bde550cb04870c1c", size = 2075366 }, - { url = "https://files.pythonhosted.org/packages/d3/43/ebef01f69baa07a482844faaa0a591bad1ef129253ffd0cdaa9d8a7f72d3/pydantic_core-2.41.5-cp312-cp312-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:d38548150c39b74aeeb0ce8ee1d8e82696f4a4e16ddc6de7b1d8823f7de4b9b5", size = 2171698 }, - { url = "https://files.pythonhosted.org/packages/b1/87/41f3202e4193e3bacfc2c065fab7706ebe81af46a83d3e27605029c1f5a6/pydantic_core-2.41.5-cp312-cp312-musllinux_1_1_aarch64.whl", hash = "sha256:c23e27686783f60290e36827f9c626e63154b82b116d7fe9adba1fda36da706c", size = 2132603 }, - { url = "https://files.pythonhosted.org/packages/49/7d/4c00df99cb12070b6bccdef4a195255e6020a550d572768d92cc54dba91a/pydantic_core-2.41.5-cp312-cp312-musllinux_1_1_armv7l.whl", hash = "sha256:482c982f814460eabe1d3bb0adfdc583387bd4691ef00b90575ca0d2b6fe2294", size = 2329591 }, - { url = "https://files.pythonhosted.org/packages/cc/6a/ebf4b1d65d458f3cda6a7335d141305dfa19bdc61140a884d165a8a1bbc7/pydantic_core-2.41.5-cp312-cp312-musllinux_1_1_x86_64.whl", hash = "sha256:bfea2a5f0b4d8d43adf9d7b8bf019fb46fdd10a2e5cde477fbcb9d1fa08c68e1", size = 2319068 }, - { url = "https://files.pythonhosted.org/packages/49/3b/774f2b5cd4192d5ab75870ce4381fd89cf218af999515baf07e7206753f0/pydantic_core-2.41.5-cp312-cp312-win32.whl", hash = "sha256:b74557b16e390ec12dca509bce9264c3bbd128f8a2c376eaa68003d7f327276d", size = 1985908 }, - { url = "https://files.pythonhosted.org/packages/86/45/00173a033c801cacf67c190fef088789394feaf88a98a7035b0e40d53dc9/pydantic_core-2.41.5-cp312-cp312-win_amd64.whl", hash = "sha256:1962293292865bca8e54702b08a4f26da73adc83dd1fcf26fbc875b35d81c815", size = 2020145 }, - { url = "https://files.pythonhosted.org/packages/f9/22/91fbc821fa6d261b376a3f73809f907cec5ca6025642c463d3488aad22fb/pydantic_core-2.41.5-cp312-cp312-win_arm64.whl", hash = "sha256:1746d4a3d9a794cacae06a5eaaccb4b8643a131d45fbc9af23e353dc0a5ba5c3", size = 1976179 }, - { url = "https://files.pythonhosted.org/packages/87/06/8806241ff1f70d9939f9af039c6c35f2360cf16e93c2ca76f184e76b1564/pydantic_core-2.41.5-cp313-cp313-macosx_10_12_x86_64.whl", hash = "sha256:941103c9be18ac8daf7b7adca8228f8ed6bb7a1849020f643b3a14d15b1924d9", size = 2120403 }, - { url = "https://files.pythonhosted.org/packages/94/02/abfa0e0bda67faa65fef1c84971c7e45928e108fe24333c81f3bfe35d5f5/pydantic_core-2.41.5-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:112e305c3314f40c93998e567879e887a3160bb8689ef3d2c04b6cc62c33ac34", size = 1896206 }, - { url = "https://files.pythonhosted.org/packages/15/df/a4c740c0943e93e6500f9eb23f4ca7ec9bf71b19e608ae5b579678c8d02f/pydantic_core-2.41.5-cp313-cp313-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0cbaad15cb0c90aa221d43c00e77bb33c93e8d36e0bf74760cd00e732d10a6a0", size = 1919307 }, - { url = "https://files.pythonhosted.org/packages/9a/e3/6324802931ae1d123528988e0e86587c2072ac2e5394b4bc2bc34b61ff6e/pydantic_core-2.41.5-cp313-cp313-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:03ca43e12fab6023fc79d28ca6b39b05f794ad08ec2feccc59a339b02f2b3d33", size = 2063258 }, - { url = "https://files.pythonhosted.org/packages/c9/d4/2230d7151d4957dd79c3044ea26346c148c98fbf0ee6ebd41056f2d62ab5/pydantic_core-2.41.5-cp313-cp313-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:dc799088c08fa04e43144b164feb0c13f9a0bc40503f8df3e9fde58a3c0c101e", size = 2214917 }, - { url = "https://files.pythonhosted.org/packages/e6/9f/eaac5df17a3672fef0081b6c1bb0b82b33ee89aa5cec0d7b05f52fd4a1fa/pydantic_core-2.41.5-cp313-cp313-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:97aeba56665b4c3235a0e52b2c2f5ae9cd071b8a8310ad27bddb3f7fb30e9aa2", size = 2332186 }, - { url = "https://files.pythonhosted.org/packages/cf/4e/35a80cae583a37cf15604b44240e45c05e04e86f9cfd766623149297e971/pydantic_core-2.41.5-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:406bf18d345822d6c21366031003612b9c77b3e29ffdb0f612367352aab7d586", size = 2073164 }, - { url = "https://files.pythonhosted.org/packages/bf/e3/f6e262673c6140dd3305d144d032f7bd5f7497d3871c1428521f19f9efa2/pydantic_core-2.41.5-cp313-cp313-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:b93590ae81f7010dbe380cdeab6f515902ebcbefe0b9327cc4804d74e93ae69d", size = 2179146 }, - { url = "https://files.pythonhosted.org/packages/75/c7/20bd7fc05f0c6ea2056a4565c6f36f8968c0924f19b7d97bbfea55780e73/pydantic_core-2.41.5-cp313-cp313-musllinux_1_1_aarch64.whl", hash = "sha256:01a3d0ab748ee531f4ea6c3e48ad9dac84ddba4b0d82291f87248f2f9de8d740", size = 2137788 }, - { url = "https://files.pythonhosted.org/packages/3a/8d/34318ef985c45196e004bc46c6eab2eda437e744c124ef0dbe1ff2c9d06b/pydantic_core-2.41.5-cp313-cp313-musllinux_1_1_armv7l.whl", hash = "sha256:6561e94ba9dacc9c61bce40e2d6bdc3bfaa0259d3ff36ace3b1e6901936d2e3e", size = 2340133 }, - { url = "https://files.pythonhosted.org/packages/9c/59/013626bf8c78a5a5d9350d12e7697d3d4de951a75565496abd40ccd46bee/pydantic_core-2.41.5-cp313-cp313-musllinux_1_1_x86_64.whl", hash = "sha256:915c3d10f81bec3a74fbd4faebe8391013ba61e5a1a8d48c4455b923bdda7858", size = 2324852 }, - { url = "https://files.pythonhosted.org/packages/1a/d9/c248c103856f807ef70c18a4f986693a46a8ffe1602e5d361485da502d20/pydantic_core-2.41.5-cp313-cp313-win32.whl", hash = "sha256:650ae77860b45cfa6e2cdafc42618ceafab3a2d9a3811fcfbd3bbf8ac3c40d36", size = 1994679 }, - { url = "https://files.pythonhosted.org/packages/9e/8b/341991b158ddab181cff136acd2552c9f35bd30380422a639c0671e99a91/pydantic_core-2.41.5-cp313-cp313-win_amd64.whl", hash = "sha256:79ec52ec461e99e13791ec6508c722742ad745571f234ea6255bed38c6480f11", size = 2019766 }, - { url = "https://files.pythonhosted.org/packages/73/7d/f2f9db34af103bea3e09735bb40b021788a5e834c81eedb541991badf8f5/pydantic_core-2.41.5-cp313-cp313-win_arm64.whl", hash = "sha256:3f84d5c1b4ab906093bdc1ff10484838aca54ef08de4afa9de0f5f14d69639cd", size = 1981005 }, - { url = "https://files.pythonhosted.org/packages/ea/28/46b7c5c9635ae96ea0fbb779e271a38129df2550f763937659ee6c5dbc65/pydantic_core-2.41.5-cp314-cp314-macosx_10_12_x86_64.whl", hash = "sha256:3f37a19d7ebcdd20b96485056ba9e8b304e27d9904d233d7b1015db320e51f0a", size = 2119622 }, - { url = "https://files.pythonhosted.org/packages/74/1a/145646e5687e8d9a1e8d09acb278c8535ebe9e972e1f162ed338a622f193/pydantic_core-2.41.5-cp314-cp314-macosx_11_0_arm64.whl", hash = "sha256:1d1d9764366c73f996edd17abb6d9d7649a7eb690006ab6adbda117717099b14", size = 1891725 }, - { url = "https://files.pythonhosted.org/packages/23/04/e89c29e267b8060b40dca97bfc64a19b2a3cf99018167ea1677d96368273/pydantic_core-2.41.5-cp314-cp314-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:25e1c2af0fce638d5f1988b686f3b3ea8cd7de5f244ca147c777769e798a9cd1", size = 1915040 }, - { url = "https://files.pythonhosted.org/packages/84/a3/15a82ac7bd97992a82257f777b3583d3e84bdb06ba6858f745daa2ec8a85/pydantic_core-2.41.5-cp314-cp314-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:506d766a8727beef16b7adaeb8ee6217c64fc813646b424d0804d67c16eddb66", size = 2063691 }, - { url = "https://files.pythonhosted.org/packages/74/9b/0046701313c6ef08c0c1cf0e028c67c770a4e1275ca73131563c5f2a310a/pydantic_core-2.41.5-cp314-cp314-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:4819fa52133c9aa3c387b3328f25c1facc356491e6135b459f1de698ff64d869", size = 2213897 }, - { url = "https://files.pythonhosted.org/packages/8a/cd/6bac76ecd1b27e75a95ca3a9a559c643b3afcd2dd62086d4b7a32a18b169/pydantic_core-2.41.5-cp314-cp314-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:2b761d210c9ea91feda40d25b4efe82a1707da2ef62901466a42492c028553a2", size = 2333302 }, - { url = "https://files.pythonhosted.org/packages/4c/d2/ef2074dc020dd6e109611a8be4449b98cd25e1b9b8a303c2f0fca2f2bcf7/pydantic_core-2.41.5-cp314-cp314-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:22f0fb8c1c583a3b6f24df2470833b40207e907b90c928cc8d3594b76f874375", size = 2064877 }, - { url = "https://files.pythonhosted.org/packages/18/66/e9db17a9a763d72f03de903883c057b2592c09509ccfe468187f2a2eef29/pydantic_core-2.41.5-cp314-cp314-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:2782c870e99878c634505236d81e5443092fba820f0373997ff75f90f68cd553", size = 2180680 }, - { url = "https://files.pythonhosted.org/packages/d3/9e/3ce66cebb929f3ced22be85d4c2399b8e85b622db77dad36b73c5387f8f8/pydantic_core-2.41.5-cp314-cp314-musllinux_1_1_aarch64.whl", hash = "sha256:0177272f88ab8312479336e1d777f6b124537d47f2123f89cb37e0accea97f90", size = 2138960 }, - { url = "https://files.pythonhosted.org/packages/a6/62/205a998f4327d2079326b01abee48e502ea739d174f0a89295c481a2272e/pydantic_core-2.41.5-cp314-cp314-musllinux_1_1_armv7l.whl", hash = "sha256:63510af5e38f8955b8ee5687740d6ebf7c2a0886d15a6d65c32814613681bc07", size = 2339102 }, - { url = "https://files.pythonhosted.org/packages/3c/0d/f05e79471e889d74d3d88f5bd20d0ed189ad94c2423d81ff8d0000aab4ff/pydantic_core-2.41.5-cp314-cp314-musllinux_1_1_x86_64.whl", hash = "sha256:e56ba91f47764cc14f1daacd723e3e82d1a89d783f0f5afe9c364b8bb491ccdb", size = 2326039 }, - { url = "https://files.pythonhosted.org/packages/ec/e1/e08a6208bb100da7e0c4b288eed624a703f4d129bde2da475721a80cab32/pydantic_core-2.41.5-cp314-cp314-win32.whl", hash = "sha256:aec5cf2fd867b4ff45b9959f8b20ea3993fc93e63c7363fe6851424c8a7e7c23", size = 1995126 }, - { url = "https://files.pythonhosted.org/packages/48/5d/56ba7b24e9557f99c9237e29f5c09913c81eeb2f3217e40e922353668092/pydantic_core-2.41.5-cp314-cp314-win_amd64.whl", hash = "sha256:8e7c86f27c585ef37c35e56a96363ab8de4e549a95512445b85c96d3e2f7c1bf", size = 2015489 }, - { url = "https://files.pythonhosted.org/packages/4e/bb/f7a190991ec9e3e0ba22e4993d8755bbc4a32925c0b5b42775c03e8148f9/pydantic_core-2.41.5-cp314-cp314-win_arm64.whl", hash = "sha256:e672ba74fbc2dc8eea59fb6d4aed6845e6905fc2a8afe93175d94a83ba2a01a0", size = 1977288 }, - { url = "https://files.pythonhosted.org/packages/92/ed/77542d0c51538e32e15afe7899d79efce4b81eee631d99850edc2f5e9349/pydantic_core-2.41.5-cp314-cp314t-macosx_10_12_x86_64.whl", hash = "sha256:8566def80554c3faa0e65ac30ab0932b9e3a5cd7f8323764303d468e5c37595a", size = 2120255 }, - { url = "https://files.pythonhosted.org/packages/bb/3d/6913dde84d5be21e284439676168b28d8bbba5600d838b9dca99de0fad71/pydantic_core-2.41.5-cp314-cp314t-macosx_11_0_arm64.whl", hash = "sha256:b80aa5095cd3109962a298ce14110ae16b8c1aece8b72f9dafe81cf597ad80b3", size = 1863760 }, - { url = "https://files.pythonhosted.org/packages/5a/f0/e5e6b99d4191da102f2b0eb9687aaa7f5bea5d9964071a84effc3e40f997/pydantic_core-2.41.5-cp314-cp314t-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3006c3dd9ba34b0c094c544c6006cc79e87d8612999f1a5d43b769b89181f23c", size = 1878092 }, - { url = "https://files.pythonhosted.org/packages/71/48/36fb760642d568925953bcc8116455513d6e34c4beaa37544118c36aba6d/pydantic_core-2.41.5-cp314-cp314t-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:72f6c8b11857a856bcfa48c86f5368439f74453563f951e473514579d44aa612", size = 2053385 }, - { url = "https://files.pythonhosted.org/packages/20/25/92dc684dd8eb75a234bc1c764b4210cf2646479d54b47bf46061657292a8/pydantic_core-2.41.5-cp314-cp314t-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:5cb1b2f9742240e4bb26b652a5aeb840aa4b417c7748b6f8387927bc6e45e40d", size = 2218832 }, - { url = "https://files.pythonhosted.org/packages/e2/09/f53e0b05023d3e30357d82eb35835d0f6340ca344720a4599cd663dca599/pydantic_core-2.41.5-cp314-cp314t-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:bd3d54f38609ff308209bd43acea66061494157703364ae40c951f83ba99a1a9", size = 2327585 }, - { url = "https://files.pythonhosted.org/packages/aa/4e/2ae1aa85d6af35a39b236b1b1641de73f5a6ac4d5a7509f77b814885760c/pydantic_core-2.41.5-cp314-cp314t-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2ff4321e56e879ee8d2a879501c8e469414d948f4aba74a2d4593184eb326660", size = 2041078 }, - { url = "https://files.pythonhosted.org/packages/cd/13/2e215f17f0ef326fc72afe94776edb77525142c693767fc347ed6288728d/pydantic_core-2.41.5-cp314-cp314t-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:d0d2568a8c11bf8225044aa94409e21da0cb09dcdafe9ecd10250b2baad531a9", size = 2173914 }, - { url = "https://files.pythonhosted.org/packages/02/7a/f999a6dcbcd0e5660bc348a3991c8915ce6599f4f2c6ac22f01d7a10816c/pydantic_core-2.41.5-cp314-cp314t-musllinux_1_1_aarch64.whl", hash = "sha256:a39455728aabd58ceabb03c90e12f71fd30fa69615760a075b9fec596456ccc3", size = 2129560 }, - { url = "https://files.pythonhosted.org/packages/3a/b1/6c990ac65e3b4c079a4fb9f5b05f5b013afa0f4ed6780a3dd236d2cbdc64/pydantic_core-2.41.5-cp314-cp314t-musllinux_1_1_armv7l.whl", hash = "sha256:239edca560d05757817c13dc17c50766136d21f7cd0fac50295499ae24f90fdf", size = 2329244 }, - { url = "https://files.pythonhosted.org/packages/d9/02/3c562f3a51afd4d88fff8dffb1771b30cfdfd79befd9883ee094f5b6c0d8/pydantic_core-2.41.5-cp314-cp314t-musllinux_1_1_x86_64.whl", hash = "sha256:2a5e06546e19f24c6a96a129142a75cee553cc018ffee48a460059b1185f4470", size = 2331955 }, - { url = "https://files.pythonhosted.org/packages/5c/96/5fb7d8c3c17bc8c62fdb031c47d77a1af698f1d7a406b0f79aaa1338f9ad/pydantic_core-2.41.5-cp314-cp314t-win32.whl", hash = "sha256:b4ececa40ac28afa90871c2cc2b9ffd2ff0bf749380fbdf57d165fd23da353aa", size = 1988906 }, - { url = "https://files.pythonhosted.org/packages/22/ed/182129d83032702912c2e2d8bbe33c036f342cc735737064668585dac28f/pydantic_core-2.41.5-cp314-cp314t-win_amd64.whl", hash = "sha256:80aa89cad80b32a912a65332f64a4450ed00966111b6615ca6816153d3585a8c", size = 1981607 }, - { url = "https://files.pythonhosted.org/packages/9f/ed/068e41660b832bb0b1aa5b58011dea2a3fe0ba7861ff38c4d4904c1c1a99/pydantic_core-2.41.5-cp314-cp314t-win_arm64.whl", hash = "sha256:35b44f37a3199f771c3eaa53051bc8a70cd7b54f333531c59e29fd4db5d15008", size = 1974769 }, - { url = "https://files.pythonhosted.org/packages/e6/b0/1a2aa41e3b5a4ba11420aba2d091b2d17959c8d1519ece3627c371951e73/pydantic_core-2.41.5-pp310-pypy310_pp73-macosx_10_12_x86_64.whl", hash = "sha256:b5819cd790dbf0c5eb9f82c73c16b39a65dd6dd4d1439dcdea7816ec9adddab8", size = 2103351 }, - { url = "https://files.pythonhosted.org/packages/a4/ee/31b1f0020baaf6d091c87900ae05c6aeae101fa4e188e1613c80e4f1ea31/pydantic_core-2.41.5-pp310-pypy310_pp73-macosx_11_0_arm64.whl", hash = "sha256:5a4e67afbc95fa5c34cf27d9089bca7fcab4e51e57278d710320a70b956d1b9a", size = 1925363 }, - { url = "https://files.pythonhosted.org/packages/e1/89/ab8e86208467e467a80deaca4e434adac37b10a9d134cd2f99b28a01e483/pydantic_core-2.41.5-pp310-pypy310_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ece5c59f0ce7d001e017643d8d24da587ea1f74f6993467d85ae8a5ef9d4f42b", size = 2135615 }, - { url = "https://files.pythonhosted.org/packages/99/0a/99a53d06dd0348b2008f2f30884b34719c323f16c3be4e6cc1203b74a91d/pydantic_core-2.41.5-pp310-pypy310_pp73-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:16f80f7abe3351f8ea6858914ddc8c77e02578544a0ebc15b4c2e1a0e813b0b2", size = 2175369 }, - { url = "https://files.pythonhosted.org/packages/6d/94/30ca3b73c6d485b9bb0bc66e611cff4a7138ff9736b7e66bcf0852151636/pydantic_core-2.41.5-pp310-pypy310_pp73-musllinux_1_1_aarch64.whl", hash = "sha256:33cb885e759a705b426baada1fe68cbb0a2e68e34c5d0d0289a364cf01709093", size = 2144218 }, - { url = "https://files.pythonhosted.org/packages/87/57/31b4f8e12680b739a91f472b5671294236b82586889ef764b5fbc6669238/pydantic_core-2.41.5-pp310-pypy310_pp73-musllinux_1_1_armv7l.whl", hash = "sha256:c8d8b4eb992936023be7dee581270af5c6e0697a8559895f527f5b7105ecd36a", size = 2329951 }, - { url = "https://files.pythonhosted.org/packages/7d/73/3c2c8edef77b8f7310e6fb012dbc4b8551386ed575b9eb6fb2506e28a7eb/pydantic_core-2.41.5-pp310-pypy310_pp73-musllinux_1_1_x86_64.whl", hash = "sha256:242a206cd0318f95cd21bdacff3fcc3aab23e79bba5cac3db5a841c9ef9c6963", size = 2318428 }, - { url = "https://files.pythonhosted.org/packages/2f/02/8559b1f26ee0d502c74f9cca5c0d2fd97e967e083e006bbbb4e97f3a043a/pydantic_core-2.41.5-pp310-pypy310_pp73-win_amd64.whl", hash = "sha256:d3a978c4f57a597908b7e697229d996d77a6d3c94901e9edee593adada95ce1a", size = 2147009 }, - { url = "https://files.pythonhosted.org/packages/5f/9b/1b3f0e9f9305839d7e84912f9e8bfbd191ed1b1ef48083609f0dabde978c/pydantic_core-2.41.5-pp311-pypy311_pp73-macosx_10_12_x86_64.whl", hash = "sha256:b2379fa7ed44ddecb5bfe4e48577d752db9fc10be00a6b7446e9663ba143de26", size = 2101980 }, - { url = "https://files.pythonhosted.org/packages/a4/ed/d71fefcb4263df0da6a85b5d8a7508360f2f2e9b3bf5814be9c8bccdccc1/pydantic_core-2.41.5-pp311-pypy311_pp73-macosx_11_0_arm64.whl", hash = "sha256:266fb4cbf5e3cbd0b53669a6d1b039c45e3ce651fd5442eff4d07c2cc8d66808", size = 1923865 }, - { url = "https://files.pythonhosted.org/packages/ce/3a/626b38db460d675f873e4444b4bb030453bbe7b4ba55df821d026a0493c4/pydantic_core-2.41.5-pp311-pypy311_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:58133647260ea01e4d0500089a8c4f07bd7aa6ce109682b1426394988d8aaacc", size = 2134256 }, - { url = "https://files.pythonhosted.org/packages/83/d9/8412d7f06f616bbc053d30cb4e5f76786af3221462ad5eee1f202021eb4e/pydantic_core-2.41.5-pp311-pypy311_pp73-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:287dad91cfb551c363dc62899a80e9e14da1f0e2b6ebde82c806612ca2a13ef1", size = 2174762 }, - { url = "https://files.pythonhosted.org/packages/55/4c/162d906b8e3ba3a99354e20faa1b49a85206c47de97a639510a0e673f5da/pydantic_core-2.41.5-pp311-pypy311_pp73-musllinux_1_1_aarch64.whl", hash = "sha256:03b77d184b9eb40240ae9fd676ca364ce1085f203e1b1256f8ab9984dca80a84", size = 2143141 }, - { url = "https://files.pythonhosted.org/packages/1f/f2/f11dd73284122713f5f89fc940f370d035fa8e1e078d446b3313955157fe/pydantic_core-2.41.5-pp311-pypy311_pp73-musllinux_1_1_armv7l.whl", hash = "sha256:a668ce24de96165bb239160b3d854943128f4334822900534f2fe947930e5770", size = 2330317 }, - { url = "https://files.pythonhosted.org/packages/88/9d/b06ca6acfe4abb296110fb1273a4d848a0bfb2ff65f3ee92127b3244e16b/pydantic_core-2.41.5-pp311-pypy311_pp73-musllinux_1_1_x86_64.whl", hash = "sha256:f14f8f046c14563f8eb3f45f499cc658ab8d10072961e07225e507adb700e93f", size = 2316992 }, - { url = "https://files.pythonhosted.org/packages/36/c7/cfc8e811f061c841d7990b0201912c3556bfeb99cdcb7ed24adc8d6f8704/pydantic_core-2.41.5-pp311-pypy311_pp73-win_amd64.whl", hash = "sha256:56121965f7a4dc965bff783d70b907ddf3d57f6eba29b6d2e5dabfaf07799c51", size = 2145302 }, -] - -[[package]] -name = "pydub" -version = "0.25.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/fe/9a/e6bca0eed82db26562c73b5076539a4a08d3cffd19c3cc5913a3e61145fd/pydub-0.25.1.tar.gz", hash = "sha256:980a33ce9949cab2a569606b65674d748ecbca4f0796887fd6f46173a7b0d30f", size = 38326 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/a6/53/d78dc063216e62fc55f6b2eebb447f6a4b0a59f55c8406376f76bf959b08/pydub-0.25.1-py2.py3-none-any.whl", hash = "sha256:65617e33033874b59d87db603aa1ed450633288aefead953b30bded59cb599a6", size = 32327 }, -] - -[[package]] -name = "pygments" -version = "2.19.2" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/b0/77/a5b8c569bf593b0140bde72ea885a803b82086995367bf2037de0159d924/pygments-2.19.2.tar.gz", hash = "sha256:636cb2477cec7f8952536970bc533bc43743542f70392ae026374600add5b887", size = 4968631 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/c7/21/705964c7812476f378728bdf590ca4b771ec72385c533964653c68e86bdc/pygments-2.19.2-py3-none-any.whl", hash = "sha256:86540386c03d588bb81d44bc3928634ff26449851e99741617ecb9037ee5ec0b", size = 1225217 }, -] - -[[package]] -name = "pyloudnorm" -version = "0.1.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "future" }, - { name = "numpy" }, - { name = "scipy", version = "1.15.3", source = { registry = "https://pypi.org/simple" }, marker = "python_full_version < '3.11'" }, - { name = "scipy", version = "1.16.3", source = { registry = "https://pypi.org/simple" }, marker = "python_full_version >= '3.11'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/75/b5/39d59c44ecd828fabfdbd796b50a561e6543ca90ef440ab307374f107856/pyloudnorm-0.1.1.tar.gz", hash = "sha256:63cd4e197dea4e7795160ea08ed02d318091bce883e436a6dbc5963326b71e1e", size = 8588 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/58/f5/6724805521ab4e723a12182f92374031032aff28a8a89dc8505c52b79032/pyloudnorm-0.1.1-py3-none-any.whl", hash = "sha256:d7f12ebdd097a464d87ce2878fc4d942f15f8233e26cc03f33fefa226f869a14", size = 9636 }, -] - -[[package]] -name = "pyparsing" -version = "3.3.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/33/c1/1d9de9aeaa1b89b0186e5fe23294ff6517fce1bc69149185577cd31016b2/pyparsing-3.3.1.tar.gz", hash = "sha256:47fad0f17ac1e2cad3de3b458570fbc9b03560aa029ed5e16ee5554da9a2251c", size = 1550512 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/8b/40/2614036cdd416452f5bf98ec037f38a1afb17f327cb8e6b652d4729e0af8/pyparsing-3.3.1-py3-none-any.whl", hash = "sha256:023b5e7e5520ad96642e2c6db4cb683d3970bd640cdf7115049a6e9c3682df82", size = 121793 }, -] - -[[package]] -name = "pytest" -version = "9.0.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "colorama", marker = "sys_platform == 'win32'" }, - { name = "exceptiongroup", marker = "python_full_version < '3.11'" }, - { name = "iniconfig" }, - { name = "packaging" }, - { name = "pluggy" }, - { name = "pygments" }, - { name = "tomli", marker = "python_full_version < '3.11'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/d1/db/7ef3487e0fb0049ddb5ce41d3a49c235bf9ad299b6a25d5780a89f19230f/pytest-9.0.2.tar.gz", hash = "sha256:75186651a92bd89611d1d9fc20f0b4345fd827c41ccd5c299a868a05d70edf11", size = 1568901 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/3b/ab/b3226f0bd7cdcf710fbede2b3548584366da3b19b5021e74f5bde2a8fa3f/pytest-9.0.2-py3-none-any.whl", hash = "sha256:711ffd45bf766d5264d487b917733b453d917afd2b0ad65223959f59089f875b", size = 374801 }, -] - -[[package]] -name = "python-dateutil" -version = "2.9.0.post0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "six" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/66/c0/0c8b6ad9f17a802ee498c46e004a0eb49bc148f2fd230864601a86dcf6db/python-dateutil-2.9.0.post0.tar.gz", hash = "sha256:37dd54208da7e1cd875388217d5e00ebd4179249f90fb72437e91a35459a0ad3", size = 342432 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/ec/57/56b9bcc3c9c6a792fcbaf139543cee77261f3651ca9da0c93f5c1221264b/python_dateutil-2.9.0.post0-py2.py3-none-any.whl", hash = "sha256:a8b2bc7bffae282281c8140a97d3aa9c14da0b136dfe83f850eea9a5f7470427", size = 229892 }, -] - -[[package]] -name = "pytorch-lightning" -version = "2.6.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "fsspec", extra = ["http"] }, - { name = "lightning-utilities" }, - { name = "packaging" }, - { name = "pyyaml" }, - { name = "torch" }, - { name = "torchmetrics" }, - { name = "tqdm" }, - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/07/d7/e3963d9669758f93b07941f4e2e82a394eb3d0980e29baa4764f3bad6689/pytorch_lightning-2.6.0.tar.gz", hash = "sha256:25b0d4f05e1f33b72be0920c34d0465777fe5f623228f9d6252b4b0f685d7037", size = 658853 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/77/eb/cc6dbfe70d15318dbce82674b1e8057cef2634ca9f9121a16b8a06c630db/pytorch_lightning-2.6.0-py3-none-any.whl", hash = "sha256:ee72cff4b8c983ecfaae8599382544bd5236d9eb300adc7dd305f359195f4e79", size = 849476 }, -] - -[[package]] -name = "pytz" -version = "2025.2" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/f8/bf/abbd3cdfb8fbc7fb3d4d38d320f2441b1e7cbe29be4f23797b4a2b5d8aac/pytz-2025.2.tar.gz", hash = "sha256:360b9e3dbb49a209c21ad61809c7fb453643e048b38924c765813546746e81c3", size = 320884 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/81/c4/34e93fe5f5429d7570ec1fa436f1986fb1f00c3e0f43a589fe2bbcd22c3f/pytz-2025.2-py2.py3-none-any.whl", hash = "sha256:5ddf76296dd8c44c26eb8f4b6f35488f3ccbf6fbbd7adee0b7262d43f0ec2f00", size = 509225 }, -] - -[[package]] -name = "pyyaml" -version = "6.0.3" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/05/8e/961c0007c59b8dd7729d542c61a4d537767a59645b82a0b521206e1e25c2/pyyaml-6.0.3.tar.gz", hash = "sha256:d76623373421df22fb4cf8817020cbb7ef15c725b9d5e45f17e189bfc384190f", size = 130960 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/f4/a0/39350dd17dd6d6c6507025c0e53aef67a9293a6d37d3511f23ea510d5800/pyyaml-6.0.3-cp310-cp310-macosx_10_13_x86_64.whl", hash = "sha256:214ed4befebe12df36bcc8bc2b64b396ca31be9304b8f59e25c11cf94a4c033b", size = 184227 }, - { url = "https://files.pythonhosted.org/packages/05/14/52d505b5c59ce73244f59c7a50ecf47093ce4765f116cdb98286a71eeca2/pyyaml-6.0.3-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:02ea2dfa234451bbb8772601d7b8e426c2bfa197136796224e50e35a78777956", size = 174019 }, - { url = "https://files.pythonhosted.org/packages/43/f7/0e6a5ae5599c838c696adb4e6330a59f463265bfa1e116cfd1fbb0abaaae/pyyaml-6.0.3-cp310-cp310-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:b30236e45cf30d2b8e7b3e85881719e98507abed1011bf463a8fa23e9c3e98a8", size = 740646 }, - { url = "https://files.pythonhosted.org/packages/2f/3a/61b9db1d28f00f8fd0ae760459a5c4bf1b941baf714e207b6eb0657d2578/pyyaml-6.0.3-cp310-cp310-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:66291b10affd76d76f54fad28e22e51719ef9ba22b29e1d7d03d6777a9174198", size = 840793 }, - { url = "https://files.pythonhosted.org/packages/7a/1e/7acc4f0e74c4b3d9531e24739e0ab832a5edf40e64fbae1a9c01941cabd7/pyyaml-6.0.3-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:9c7708761fccb9397fe64bbc0395abcae8c4bf7b0eac081e12b809bf47700d0b", size = 770293 }, - { url = "https://files.pythonhosted.org/packages/8b/ef/abd085f06853af0cd59fa5f913d61a8eab65d7639ff2a658d18a25d6a89d/pyyaml-6.0.3-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:418cf3f2111bc80e0933b2cd8cd04f286338bb88bdc7bc8e6dd775ebde60b5e0", size = 732872 }, - { url = "https://files.pythonhosted.org/packages/1f/15/2bc9c8faf6450a8b3c9fc5448ed869c599c0a74ba2669772b1f3a0040180/pyyaml-6.0.3-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:5e0b74767e5f8c593e8c9b5912019159ed0533c70051e9cce3e8b6aa699fcd69", size = 758828 }, - { url = "https://files.pythonhosted.org/packages/a3/00/531e92e88c00f4333ce359e50c19b8d1de9fe8d581b1534e35ccfbc5f393/pyyaml-6.0.3-cp310-cp310-win32.whl", hash = "sha256:28c8d926f98f432f88adc23edf2e6d4921ac26fb084b028c733d01868d19007e", size = 142415 }, - { url = "https://files.pythonhosted.org/packages/2a/fa/926c003379b19fca39dd4634818b00dec6c62d87faf628d1394e137354d4/pyyaml-6.0.3-cp310-cp310-win_amd64.whl", hash = "sha256:bdb2c67c6c1390b63c6ff89f210c8fd09d9a1217a465701eac7316313c915e4c", size = 158561 }, - { url = "https://files.pythonhosted.org/packages/6d/16/a95b6757765b7b031c9374925bb718d55e0a9ba8a1b6a12d25962ea44347/pyyaml-6.0.3-cp311-cp311-macosx_10_13_x86_64.whl", hash = "sha256:44edc647873928551a01e7a563d7452ccdebee747728c1080d881d68af7b997e", size = 185826 }, - { url = "https://files.pythonhosted.org/packages/16/19/13de8e4377ed53079ee996e1ab0a9c33ec2faf808a4647b7b4c0d46dd239/pyyaml-6.0.3-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:652cb6edd41e718550aad172851962662ff2681490a8a711af6a4d288dd96824", size = 175577 }, - { url = "https://files.pythonhosted.org/packages/0c/62/d2eb46264d4b157dae1275b573017abec435397aa59cbcdab6fc978a8af4/pyyaml-6.0.3-cp311-cp311-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:10892704fc220243f5305762e276552a0395f7beb4dbf9b14ec8fd43b57f126c", size = 775556 }, - { url = "https://files.pythonhosted.org/packages/10/cb/16c3f2cf3266edd25aaa00d6c4350381c8b012ed6f5276675b9eba8d9ff4/pyyaml-6.0.3-cp311-cp311-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:850774a7879607d3a6f50d36d04f00ee69e7fc816450e5f7e58d7f17f1ae5c00", size = 882114 }, - { url = "https://files.pythonhosted.org/packages/71/60/917329f640924b18ff085ab889a11c763e0b573da888e8404ff486657602/pyyaml-6.0.3-cp311-cp311-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:b8bb0864c5a28024fac8a632c443c87c5aa6f215c0b126c449ae1a150412f31d", size = 806638 }, - { url = "https://files.pythonhosted.org/packages/dd/6f/529b0f316a9fd167281a6c3826b5583e6192dba792dd55e3203d3f8e655a/pyyaml-6.0.3-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:1d37d57ad971609cf3c53ba6a7e365e40660e3be0e5175fa9f2365a379d6095a", size = 767463 }, - { url = "https://files.pythonhosted.org/packages/f2/6a/b627b4e0c1dd03718543519ffb2f1deea4a1e6d42fbab8021936a4d22589/pyyaml-6.0.3-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:37503bfbfc9d2c40b344d06b2199cf0e96e97957ab1c1b546fd4f87e53e5d3e4", size = 794986 }, - { url = "https://files.pythonhosted.org/packages/45/91/47a6e1c42d9ee337c4839208f30d9f09caa9f720ec7582917b264defc875/pyyaml-6.0.3-cp311-cp311-win32.whl", hash = "sha256:8098f252adfa6c80ab48096053f512f2321f0b998f98150cea9bd23d83e1467b", size = 142543 }, - { url = "https://files.pythonhosted.org/packages/da/e3/ea007450a105ae919a72393cb06f122f288ef60bba2dc64b26e2646fa315/pyyaml-6.0.3-cp311-cp311-win_amd64.whl", hash = "sha256:9f3bfb4965eb874431221a3ff3fdcddc7e74e3b07799e0e84ca4a0f867d449bf", size = 158763 }, - { url = "https://files.pythonhosted.org/packages/d1/33/422b98d2195232ca1826284a76852ad5a86fe23e31b009c9886b2d0fb8b2/pyyaml-6.0.3-cp312-cp312-macosx_10_13_x86_64.whl", hash = "sha256:7f047e29dcae44602496db43be01ad42fc6f1cc0d8cd6c83d342306c32270196", size = 182063 }, - { url = "https://files.pythonhosted.org/packages/89/a0/6cf41a19a1f2f3feab0e9c0b74134aa2ce6849093d5517a0c550fe37a648/pyyaml-6.0.3-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:fc09d0aa354569bc501d4e787133afc08552722d3ab34836a80547331bb5d4a0", size = 173973 }, - { url = "https://files.pythonhosted.org/packages/ed/23/7a778b6bd0b9a8039df8b1b1d80e2e2ad78aa04171592c8a5c43a56a6af4/pyyaml-6.0.3-cp312-cp312-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:9149cad251584d5fb4981be1ecde53a1ca46c891a79788c0df828d2f166bda28", size = 775116 }, - { url = "https://files.pythonhosted.org/packages/65/30/d7353c338e12baef4ecc1b09e877c1970bd3382789c159b4f89d6a70dc09/pyyaml-6.0.3-cp312-cp312-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:5fdec68f91a0c6739b380c83b951e2c72ac0197ace422360e6d5a959d8d97b2c", size = 844011 }, - { url = "https://files.pythonhosted.org/packages/8b/9d/b3589d3877982d4f2329302ef98a8026e7f4443c765c46cfecc8858c6b4b/pyyaml-6.0.3-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:ba1cc08a7ccde2d2ec775841541641e4548226580ab850948cbfda66a1befcdc", size = 807870 }, - { url = "https://files.pythonhosted.org/packages/05/c0/b3be26a015601b822b97d9149ff8cb5ead58c66f981e04fedf4e762f4bd4/pyyaml-6.0.3-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:8dc52c23056b9ddd46818a57b78404882310fb473d63f17b07d5c40421e47f8e", size = 761089 }, - { url = "https://files.pythonhosted.org/packages/be/8e/98435a21d1d4b46590d5459a22d88128103f8da4c2d4cb8f14f2a96504e1/pyyaml-6.0.3-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:41715c910c881bc081f1e8872880d3c650acf13dfa8214bad49ed4cede7c34ea", size = 790181 }, - { url = "https://files.pythonhosted.org/packages/74/93/7baea19427dcfbe1e5a372d81473250b379f04b1bd3c4c5ff825e2327202/pyyaml-6.0.3-cp312-cp312-win32.whl", hash = "sha256:96b533f0e99f6579b3d4d4995707cf36df9100d67e0c8303a0c55b27b5f99bc5", size = 137658 }, - { url = "https://files.pythonhosted.org/packages/86/bf/899e81e4cce32febab4fb42bb97dcdf66bc135272882d1987881a4b519e9/pyyaml-6.0.3-cp312-cp312-win_amd64.whl", hash = "sha256:5fcd34e47f6e0b794d17de1b4ff496c00986e1c83f7ab2fb8fcfe9616ff7477b", size = 154003 }, - { url = "https://files.pythonhosted.org/packages/1a/08/67bd04656199bbb51dbed1439b7f27601dfb576fb864099c7ef0c3e55531/pyyaml-6.0.3-cp312-cp312-win_arm64.whl", hash = "sha256:64386e5e707d03a7e172c0701abfb7e10f0fb753ee1d773128192742712a98fd", size = 140344 }, - { url = "https://files.pythonhosted.org/packages/d1/11/0fd08f8192109f7169db964b5707a2f1e8b745d4e239b784a5a1dd80d1db/pyyaml-6.0.3-cp313-cp313-macosx_10_13_x86_64.whl", hash = "sha256:8da9669d359f02c0b91ccc01cac4a67f16afec0dac22c2ad09f46bee0697eba8", size = 181669 }, - { url = "https://files.pythonhosted.org/packages/b1/16/95309993f1d3748cd644e02e38b75d50cbc0d9561d21f390a76242ce073f/pyyaml-6.0.3-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:2283a07e2c21a2aa78d9c4442724ec1eb15f5e42a723b99cb3d822d48f5f7ad1", size = 173252 }, - { url = "https://files.pythonhosted.org/packages/50/31/b20f376d3f810b9b2371e72ef5adb33879b25edb7a6d072cb7ca0c486398/pyyaml-6.0.3-cp313-cp313-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:ee2922902c45ae8ccada2c5b501ab86c36525b883eff4255313a253a3160861c", size = 767081 }, - { url = "https://files.pythonhosted.org/packages/49/1e/a55ca81e949270d5d4432fbbd19dfea5321eda7c41a849d443dc92fd1ff7/pyyaml-6.0.3-cp313-cp313-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:a33284e20b78bd4a18c8c2282d549d10bc8408a2a7ff57653c0cf0b9be0afce5", size = 841159 }, - { url = "https://files.pythonhosted.org/packages/74/27/e5b8f34d02d9995b80abcef563ea1f8b56d20134d8f4e5e81733b1feceb2/pyyaml-6.0.3-cp313-cp313-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:0f29edc409a6392443abf94b9cf89ce99889a1dd5376d94316ae5145dfedd5d6", size = 801626 }, - { url = "https://files.pythonhosted.org/packages/f9/11/ba845c23988798f40e52ba45f34849aa8a1f2d4af4b798588010792ebad6/pyyaml-6.0.3-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:f7057c9a337546edc7973c0d3ba84ddcdf0daa14533c2065749c9075001090e6", size = 753613 }, - { url = "https://files.pythonhosted.org/packages/3d/e0/7966e1a7bfc0a45bf0a7fb6b98ea03fc9b8d84fa7f2229e9659680b69ee3/pyyaml-6.0.3-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:eda16858a3cab07b80edaf74336ece1f986ba330fdb8ee0d6c0d68fe82bc96be", size = 794115 }, - { url = "https://files.pythonhosted.org/packages/de/94/980b50a6531b3019e45ddeada0626d45fa85cbe22300844a7983285bed3b/pyyaml-6.0.3-cp313-cp313-win32.whl", hash = "sha256:d0eae10f8159e8fdad514efdc92d74fd8d682c933a6dd088030f3834bc8e6b26", size = 137427 }, - { url = "https://files.pythonhosted.org/packages/97/c9/39d5b874e8b28845e4ec2202b5da735d0199dbe5b8fb85f91398814a9a46/pyyaml-6.0.3-cp313-cp313-win_amd64.whl", hash = "sha256:79005a0d97d5ddabfeeea4cf676af11e647e41d81c9a7722a193022accdb6b7c", size = 154090 }, - { url = "https://files.pythonhosted.org/packages/73/e8/2bdf3ca2090f68bb3d75b44da7bbc71843b19c9f2b9cb9b0f4ab7a5a4329/pyyaml-6.0.3-cp313-cp313-win_arm64.whl", hash = "sha256:5498cd1645aa724a7c71c8f378eb29ebe23da2fc0d7a08071d89469bf1d2defb", size = 140246 }, - { url = "https://files.pythonhosted.org/packages/9d/8c/f4bd7f6465179953d3ac9bc44ac1a8a3e6122cf8ada906b4f96c60172d43/pyyaml-6.0.3-cp314-cp314-macosx_10_13_x86_64.whl", hash = "sha256:8d1fab6bb153a416f9aeb4b8763bc0f22a5586065f86f7664fc23339fc1c1fac", size = 181814 }, - { url = "https://files.pythonhosted.org/packages/bd/9c/4d95bb87eb2063d20db7b60faa3840c1b18025517ae857371c4dd55a6b3a/pyyaml-6.0.3-cp314-cp314-macosx_11_0_arm64.whl", hash = "sha256:34d5fcd24b8445fadc33f9cf348c1047101756fd760b4dacb5c3e99755703310", size = 173809 }, - { url = "https://files.pythonhosted.org/packages/92/b5/47e807c2623074914e29dabd16cbbdd4bf5e9b2db9f8090fa64411fc5382/pyyaml-6.0.3-cp314-cp314-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:501a031947e3a9025ed4405a168e6ef5ae3126c59f90ce0cd6f2bfc477be31b7", size = 766454 }, - { url = "https://files.pythonhosted.org/packages/02/9e/e5e9b168be58564121efb3de6859c452fccde0ab093d8438905899a3a483/pyyaml-6.0.3-cp314-cp314-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:b3bc83488de33889877a0f2543ade9f70c67d66d9ebb4ac959502e12de895788", size = 836355 }, - { url = "https://files.pythonhosted.org/packages/88/f9/16491d7ed2a919954993e48aa941b200f38040928474c9e85ea9e64222c3/pyyaml-6.0.3-cp314-cp314-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:c458b6d084f9b935061bc36216e8a69a7e293a2f1e68bf956dcd9e6cbcd143f5", size = 794175 }, - { url = "https://files.pythonhosted.org/packages/dd/3f/5989debef34dc6397317802b527dbbafb2b4760878a53d4166579111411e/pyyaml-6.0.3-cp314-cp314-musllinux_1_2_aarch64.whl", hash = "sha256:7c6610def4f163542a622a73fb39f534f8c101d690126992300bf3207eab9764", size = 755228 }, - { url = "https://files.pythonhosted.org/packages/d7/ce/af88a49043cd2e265be63d083fc75b27b6ed062f5f9fd6cdc223ad62f03e/pyyaml-6.0.3-cp314-cp314-musllinux_1_2_x86_64.whl", hash = "sha256:5190d403f121660ce8d1d2c1bb2ef1bd05b5f68533fc5c2ea899bd15f4399b35", size = 789194 }, - { url = "https://files.pythonhosted.org/packages/23/20/bb6982b26a40bb43951265ba29d4c246ef0ff59c9fdcdf0ed04e0687de4d/pyyaml-6.0.3-cp314-cp314-win_amd64.whl", hash = "sha256:4a2e8cebe2ff6ab7d1050ecd59c25d4c8bd7e6f400f5f82b96557ac0abafd0ac", size = 156429 }, - { url = "https://files.pythonhosted.org/packages/f4/f4/a4541072bb9422c8a883ab55255f918fa378ecf083f5b85e87fc2b4eda1b/pyyaml-6.0.3-cp314-cp314-win_arm64.whl", hash = "sha256:93dda82c9c22deb0a405ea4dc5f2d0cda384168e466364dec6255b293923b2f3", size = 143912 }, - { url = "https://files.pythonhosted.org/packages/7c/f9/07dd09ae774e4616edf6cda684ee78f97777bdd15847253637a6f052a62f/pyyaml-6.0.3-cp314-cp314t-macosx_10_13_x86_64.whl", hash = "sha256:02893d100e99e03eda1c8fd5c441d8c60103fd175728e23e431db1b589cf5ab3", size = 189108 }, - { url = "https://files.pythonhosted.org/packages/4e/78/8d08c9fb7ce09ad8c38ad533c1191cf27f7ae1effe5bb9400a46d9437fcf/pyyaml-6.0.3-cp314-cp314t-macosx_11_0_arm64.whl", hash = "sha256:c1ff362665ae507275af2853520967820d9124984e0f7466736aea23d8611fba", size = 183641 }, - { url = "https://files.pythonhosted.org/packages/7b/5b/3babb19104a46945cf816d047db2788bcaf8c94527a805610b0289a01c6b/pyyaml-6.0.3-cp314-cp314t-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:6adc77889b628398debc7b65c073bcb99c4a0237b248cacaf3fe8a557563ef6c", size = 831901 }, - { url = "https://files.pythonhosted.org/packages/8b/cc/dff0684d8dc44da4d22a13f35f073d558c268780ce3c6ba1b87055bb0b87/pyyaml-6.0.3-cp314-cp314t-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:a80cb027f6b349846a3bf6d73b5e95e782175e52f22108cfa17876aaeff93702", size = 861132 }, - { url = "https://files.pythonhosted.org/packages/b1/5e/f77dc6b9036943e285ba76b49e118d9ea929885becb0a29ba8a7c75e29fe/pyyaml-6.0.3-cp314-cp314t-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:00c4bdeba853cc34e7dd471f16b4114f4162dc03e6b7afcc2128711f0eca823c", size = 839261 }, - { url = "https://files.pythonhosted.org/packages/ce/88/a9db1376aa2a228197c58b37302f284b5617f56a5d959fd1763fb1675ce6/pyyaml-6.0.3-cp314-cp314t-musllinux_1_2_aarch64.whl", hash = "sha256:66e1674c3ef6f541c35191caae2d429b967b99e02040f5ba928632d9a7f0f065", size = 805272 }, - { url = "https://files.pythonhosted.org/packages/da/92/1446574745d74df0c92e6aa4a7b0b3130706a4142b2d1a5869f2eaa423c6/pyyaml-6.0.3-cp314-cp314t-musllinux_1_2_x86_64.whl", hash = "sha256:16249ee61e95f858e83976573de0f5b2893b3677ba71c9dd36b9cf8be9ac6d65", size = 829923 }, - { url = "https://files.pythonhosted.org/packages/f0/7a/1c7270340330e575b92f397352af856a8c06f230aa3e76f86b39d01b416a/pyyaml-6.0.3-cp314-cp314t-win_amd64.whl", hash = "sha256:4ad1906908f2f5ae4e5a8ddfce73c320c2a1429ec52eafd27138b7f1cbe341c9", size = 174062 }, - { url = "https://files.pythonhosted.org/packages/f1/12/de94a39c2ef588c7e6455cfbe7343d3b2dc9d6b6b2f40c4c6565744c873d/pyyaml-6.0.3-cp314-cp314t-win_arm64.whl", hash = "sha256:ebc55a14a21cb14062aa4162f906cd962b28e2e9ea38f9b4391244cd8de4ae0b", size = 149341 }, -] - -[[package]] -name = "pyyaml-ft" -version = "8.0.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/5e/eb/5a0d575de784f9a1f94e2b1288c6886f13f34185e13117ed530f32b6f8a8/pyyaml_ft-8.0.0.tar.gz", hash = "sha256:0c947dce03954c7b5d38869ed4878b2e6ff1d44b08a0d84dc83fdad205ae39ab", size = 141057 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/68/ba/a067369fe61a2e57fb38732562927d5bae088c73cb9bb5438736a9555b29/pyyaml_ft-8.0.0-cp313-cp313-macosx_10_13_x86_64.whl", hash = "sha256:8c1306282bc958bfda31237f900eb52c9bedf9b93a11f82e1aab004c9a5657a6", size = 187027 }, - { url = "https://files.pythonhosted.org/packages/ad/c5/a3d2020ce5ccfc6aede0d45bcb870298652ac0cf199f67714d250e0cdf39/pyyaml_ft-8.0.0-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:30c5f1751625786c19de751e3130fc345ebcba6a86f6bddd6e1285342f4bbb69", size = 176146 }, - { url = "https://files.pythonhosted.org/packages/e3/bb/23a9739291086ca0d3189eac7cd92b4d00e9fdc77d722ab610c35f9a82ba/pyyaml_ft-8.0.0-cp313-cp313-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3fa992481155ddda2e303fcc74c79c05eddcdbc907b888d3d9ce3ff3e2adcfb0", size = 746792 }, - { url = "https://files.pythonhosted.org/packages/5f/c2/e8825f4ff725b7e560d62a3609e31d735318068e1079539ebfde397ea03e/pyyaml_ft-8.0.0-cp313-cp313-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:cec6c92b4207004b62dfad1f0be321c9f04725e0f271c16247d8b39c3bf3ea42", size = 786772 }, - { url = "https://files.pythonhosted.org/packages/35/be/58a4dcae8854f2fdca9b28d9495298fd5571a50d8430b1c3033ec95d2d0e/pyyaml_ft-8.0.0-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:06237267dbcab70d4c0e9436d8f719f04a51123f0ca2694c00dd4b68c338e40b", size = 778723 }, - { url = "https://files.pythonhosted.org/packages/86/ed/fed0da92b5d5d7340a082e3802d84c6dc9d5fa142954404c41a544c1cb92/pyyaml_ft-8.0.0-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:8a7f332bc565817644cdb38ffe4739e44c3e18c55793f75dddb87630f03fc254", size = 758478 }, - { url = "https://files.pythonhosted.org/packages/f0/69/ac02afe286275980ecb2dcdc0156617389b7e0c0a3fcdedf155c67be2b80/pyyaml_ft-8.0.0-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:7d10175a746be65f6feb86224df5d6bc5c049ebf52b89a88cf1cd78af5a367a8", size = 799159 }, - { url = "https://files.pythonhosted.org/packages/4e/ac/c492a9da2e39abdff4c3094ec54acac9747743f36428281fb186a03fab76/pyyaml_ft-8.0.0-cp313-cp313-win_amd64.whl", hash = "sha256:58e1015098cf8d8aec82f360789c16283b88ca670fe4275ef6c48c5e30b22a96", size = 158779 }, - { url = "https://files.pythonhosted.org/packages/5d/9b/41998df3298960d7c67653669f37710fa2d568a5fc933ea24a6df60acaf6/pyyaml_ft-8.0.0-cp313-cp313t-macosx_10_13_x86_64.whl", hash = "sha256:e64fa5f3e2ceb790d50602b2fd4ec37abbd760a8c778e46354df647e7c5a4ebb", size = 191331 }, - { url = "https://files.pythonhosted.org/packages/0f/16/2710c252ee04cbd74d9562ebba709e5a284faeb8ada88fcda548c9191b47/pyyaml_ft-8.0.0-cp313-cp313t-macosx_11_0_arm64.whl", hash = "sha256:8d445bf6ea16bb93c37b42fdacfb2f94c8e92a79ba9e12768c96ecde867046d1", size = 182879 }, - { url = "https://files.pythonhosted.org/packages/9a/40/ae8163519d937fa7bfa457b6f78439cc6831a7c2b170e4f612f7eda71815/pyyaml_ft-8.0.0-cp313-cp313t-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8c56bb46b4fda34cbb92a9446a841da3982cdde6ea13de3fbd80db7eeeab8b49", size = 811277 }, - { url = "https://files.pythonhosted.org/packages/f9/66/28d82dbff7f87b96f0eeac79b7d972a96b4980c1e445eb6a857ba91eda00/pyyaml_ft-8.0.0-cp313-cp313t-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:dab0abb46eb1780da486f022dce034b952c8ae40753627b27a626d803926483b", size = 831650 }, - { url = "https://files.pythonhosted.org/packages/e8/df/161c4566facac7d75a9e182295c223060373d4116dead9cc53a265de60b9/pyyaml_ft-8.0.0-cp313-cp313t-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:bd48d639cab5ca50ad957b6dd632c7dd3ac02a1abe0e8196a3c24a52f5db3f7a", size = 815755 }, - { url = "https://files.pythonhosted.org/packages/05/10/f42c48fa5153204f42eaa945e8d1fd7c10d6296841dcb2447bf7da1be5c4/pyyaml_ft-8.0.0-cp313-cp313t-musllinux_1_2_aarch64.whl", hash = "sha256:052561b89d5b2a8e1289f326d060e794c21fa068aa11255fe71d65baf18a632e", size = 810403 }, - { url = "https://files.pythonhosted.org/packages/d5/d2/e369064aa51009eb9245399fd8ad2c562bd0bcd392a00be44b2a824ded7c/pyyaml_ft-8.0.0-cp313-cp313t-musllinux_1_2_x86_64.whl", hash = "sha256:3bb4b927929b0cb162fb1605392a321e3333e48ce616cdcfa04a839271373255", size = 835581 }, - { url = "https://files.pythonhosted.org/packages/c0/28/26534bed77109632a956977f60d8519049f545abc39215d086e33a61f1f2/pyyaml_ft-8.0.0-cp313-cp313t-win_amd64.whl", hash = "sha256:de04cfe9439565e32f178106c51dd6ca61afaa2907d143835d501d84703d3793", size = 171579 }, -] - -[[package]] -name = "rapidfuzz" -version = "3.14.3" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/d3/28/9d808fe62375b9aab5ba92fa9b29371297b067c2790b2d7cda648b1e2f8d/rapidfuzz-3.14.3.tar.gz", hash = "sha256:2491937177868bc4b1e469087601d53f925e8d270ccc21e07404b4b5814b7b5f", size = 57863900 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/69/d1/0efa42a602ed466d3ca1c462eed5d62015c3fd2a402199e2c4b87aa5aa25/rapidfuzz-3.14.3-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:b9fcd4d751a4fffa17aed1dde41647923c72c74af02459ad1222e3b0022da3a1", size = 1952376 }, - { url = "https://files.pythonhosted.org/packages/be/00/37a169bb28b23850a164e6624b1eb299e1ad73c9e7c218ee15744e68d628/rapidfuzz-3.14.3-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:4ad73afb688b36864a8d9b7344a9cf6da186c471e5790cbf541a635ee0f457f2", size = 1390903 }, - { url = "https://files.pythonhosted.org/packages/3c/91/b37207cbbdb6eaafac3da3f55ea85287b27745cb416e75e15769b7d8abe8/rapidfuzz-3.14.3-cp310-cp310-manylinux_2_26_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:c5fb2d978a601820d2cfd111e2c221a9a7bfdf84b41a3ccbb96ceef29f2f1ac7", size = 1385655 }, - { url = "https://files.pythonhosted.org/packages/f2/bb/ca53e518acf43430be61f23b9c5987bd1e01e74fcb7a9ee63e00f597aefb/rapidfuzz-3.14.3-cp310-cp310-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:1d83b8b712fa37e06d59f29a4b49e2e9e8635e908fbc21552fe4d1163db9d2a1", size = 3164708 }, - { url = "https://files.pythonhosted.org/packages/df/e1/7667bf2db3e52adb13cb933dd4a6a2efc66045d26fa150fc0feb64c26d61/rapidfuzz-3.14.3-cp310-cp310-manylinux_2_31_armv7l.whl", hash = "sha256:dc8c07801df5206b81ed6bd6c35cb520cf9b6c64b9b0d19d699f8633dc942897", size = 1221106 }, - { url = "https://files.pythonhosted.org/packages/05/8a/84d9f2d46a2c8eb2ccae81747c4901fa10fe4010aade2d57ce7b4b8e02ec/rapidfuzz-3.14.3-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:c71ce6d4231e5ef2e33caa952bfe671cb9fd42e2afb11952df9fad41d5c821f9", size = 2406048 }, - { url = "https://files.pythonhosted.org/packages/3c/a9/a0b7b7a1b81a020c034eb67c8e23b7e49f920004e295378de3046b0d99e1/rapidfuzz-3.14.3-cp310-cp310-musllinux_1_2_armv7l.whl", hash = "sha256:0e38828d1381a0cceb8a4831212b2f673d46f5129a1897b0451c883eaf4a1747", size = 2527020 }, - { url = "https://files.pythonhosted.org/packages/b4/bc/416df7d108b99b4942ba04dd4cf73c45c3aadb3ef003d95cad78b1d12eb9/rapidfuzz-3.14.3-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:da2a007434323904719158e50f3076a4dadb176ce43df28ed14610c773cc9825", size = 4273958 }, - { url = "https://files.pythonhosted.org/packages/81/d0/b81e041c17cd475002114e0ab8800e4305e60837882cb376a621e520d70f/rapidfuzz-3.14.3-cp310-cp310-win32.whl", hash = "sha256:fce3152f94afcfd12f3dd8cf51e48fa606e3cb56719bccebe3b401f43d0714f9", size = 1725043 }, - { url = "https://files.pythonhosted.org/packages/09/6b/64ad573337d81d64bc78a6a1df53a72a71d54d43d276ce0662c2e95a1f35/rapidfuzz-3.14.3-cp310-cp310-win_amd64.whl", hash = "sha256:37d3c653af15cd88592633e942f5407cb4c64184efab163c40fcebad05f25141", size = 1542273 }, - { url = "https://files.pythonhosted.org/packages/f4/5e/faf76e259bc15808bc0b86028f510215c3d755b6c3a3911113079485e561/rapidfuzz-3.14.3-cp310-cp310-win_arm64.whl", hash = "sha256:cc594bbcd3c62f647dfac66800f307beaee56b22aaba1c005e9c4c40ed733923", size = 814875 }, - { url = "https://files.pythonhosted.org/packages/76/25/5b0a33ad3332ee1213068c66f7c14e9e221be90bab434f0cb4defa9d6660/rapidfuzz-3.14.3-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:dea2d113e260a5da0c4003e0a5e9fdf24a9dc2bb9eaa43abd030a1e46ce7837d", size = 1953885 }, - { url = "https://files.pythonhosted.org/packages/2d/ab/f1181f500c32c8fcf7c966f5920c7e56b9b1d03193386d19c956505c312d/rapidfuzz-3.14.3-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:e6c31a4aa68cfa75d7eede8b0ed24b9e458447db604c2db53f358be9843d81d3", size = 1390200 }, - { url = "https://files.pythonhosted.org/packages/14/2a/0f2de974ececad873865c6bb3ea3ad07c976ac293d5025b2d73325aac1d4/rapidfuzz-3.14.3-cp311-cp311-manylinux_2_26_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:02821366d928e68ddcb567fed8723dad7ea3a979fada6283e6914d5858674850", size = 1389319 }, - { url = "https://files.pythonhosted.org/packages/ed/69/309d8f3a0bb3031fd9b667174cc4af56000645298af7c2931be5c3d14bb4/rapidfuzz-3.14.3-cp311-cp311-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:cfe8df315ab4e6db4e1be72c5170f8e66021acde22cd2f9d04d2058a9fd8162e", size = 3178495 }, - { url = "https://files.pythonhosted.org/packages/10/b7/f9c44a99269ea5bf6fd6a40b84e858414b6e241288b9f2b74af470d222b1/rapidfuzz-3.14.3-cp311-cp311-manylinux_2_31_armv7l.whl", hash = "sha256:769f31c60cd79420188fcdb3c823227fc4a6deb35cafec9d14045c7f6743acae", size = 1228443 }, - { url = "https://files.pythonhosted.org/packages/f2/0a/3b3137abac7f19c9220e14cd7ce993e35071a7655e7ef697785a3edfea1a/rapidfuzz-3.14.3-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:54fa03062124e73086dae66a3451c553c1e20a39c077fd704dc7154092c34c63", size = 2411998 }, - { url = "https://files.pythonhosted.org/packages/f3/b6/983805a844d44670eaae63831024cdc97ada4e9c62abc6b20703e81e7f9b/rapidfuzz-3.14.3-cp311-cp311-musllinux_1_2_armv7l.whl", hash = "sha256:834d1e818005ed0d4ae38f6b87b86fad9b0a74085467ece0727d20e15077c094", size = 2530120 }, - { url = "https://files.pythonhosted.org/packages/b4/cc/2c97beb2b1be2d7595d805682472f1b1b844111027d5ad89b65e16bdbaaa/rapidfuzz-3.14.3-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:948b00e8476a91f510dd1ec07272efc7d78c275d83b630455559671d4e33b678", size = 4283129 }, - { url = "https://files.pythonhosted.org/packages/4d/03/2f0e5e94941045aefe7eafab72320e61285c07b752df9884ce88d6b8b835/rapidfuzz-3.14.3-cp311-cp311-win32.whl", hash = "sha256:43d0305c36f504232f18ea04e55f2059bb89f169d3119c4ea96a0e15b59e2a91", size = 1724224 }, - { url = "https://files.pythonhosted.org/packages/cf/99/5fa23e204435803875daefda73fd61baeabc3c36b8fc0e34c1705aab8c7b/rapidfuzz-3.14.3-cp311-cp311-win_amd64.whl", hash = "sha256:ef6bf930b947bd0735c550683939a032090f1d688dfd8861d6b45307b96fd5c5", size = 1544259 }, - { url = "https://files.pythonhosted.org/packages/48/35/d657b85fcc615a42661b98ac90ce8e95bd32af474603a105643963749886/rapidfuzz-3.14.3-cp311-cp311-win_arm64.whl", hash = "sha256:f3eb0ff3b75d6fdccd40b55e7414bb859a1cda77c52762c9c82b85569f5088e7", size = 814734 }, - { url = "https://files.pythonhosted.org/packages/fa/8e/3c215e860b458cfbedb3ed73bc72e98eb7e0ed72f6b48099604a7a3260c2/rapidfuzz-3.14.3-cp312-cp312-macosx_10_13_x86_64.whl", hash = "sha256:685c93ea961d135893b5984a5a9851637d23767feabe414ec974f43babbd8226", size = 1945306 }, - { url = "https://files.pythonhosted.org/packages/36/d9/31b33512015c899f4a6e6af64df8dfe8acddf4c8b40a4b3e0e6e1bcd00e5/rapidfuzz-3.14.3-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:fa7c8f26f009f8c673fbfb443792f0cf8cf50c4e18121ff1e285b5e08a94fbdb", size = 1390788 }, - { url = "https://files.pythonhosted.org/packages/a9/67/2ee6f8de6e2081ccd560a571d9c9063184fe467f484a17fa90311a7f4a2e/rapidfuzz-3.14.3-cp312-cp312-manylinux_2_26_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:57f878330c8d361b2ce76cebb8e3e1dc827293b6abf404e67d53260d27b5d941", size = 1374580 }, - { url = "https://files.pythonhosted.org/packages/30/83/80d22997acd928eda7deadc19ccd15883904622396d6571e935993e0453a/rapidfuzz-3.14.3-cp312-cp312-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:6c5f545f454871e6af05753a0172849c82feaf0f521c5ca62ba09e1b382d6382", size = 3154947 }, - { url = "https://files.pythonhosted.org/packages/5b/cf/9f49831085a16384695f9fb096b99662f589e30b89b4a589a1ebc1a19d34/rapidfuzz-3.14.3-cp312-cp312-manylinux_2_31_armv7l.whl", hash = "sha256:07aa0b5d8863e3151e05026a28e0d924accf0a7a3b605da978f0359bb804df43", size = 1223872 }, - { url = "https://files.pythonhosted.org/packages/c8/0f/41ee8034e744b871c2e071ef0d360686f5ccfe5659f4fd96c3ec406b3c8b/rapidfuzz-3.14.3-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:73b07566bc7e010e7b5bd490fb04bb312e820970180df6b5655e9e6224c137db", size = 2392512 }, - { url = "https://files.pythonhosted.org/packages/da/86/280038b6b0c2ccec54fb957c732ad6b41cc1fd03b288d76545b9cf98343f/rapidfuzz-3.14.3-cp312-cp312-musllinux_1_2_armv7l.whl", hash = "sha256:6de00eb84c71476af7d3110cf25d8fe7c792d7f5fa86764ef0b4ca97e78ca3ed", size = 2521398 }, - { url = "https://files.pythonhosted.org/packages/fa/7b/05c26f939607dca0006505e3216248ae2de631e39ef94dd63dbbf0860021/rapidfuzz-3.14.3-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:d7843a1abf0091773a530636fdd2a49a41bcae22f9910b86b4f903e76ddc82dc", size = 4259416 }, - { url = "https://files.pythonhosted.org/packages/40/eb/9e3af4103d91788f81111af1b54a28de347cdbed8eaa6c91d5e98a889aab/rapidfuzz-3.14.3-cp312-cp312-win32.whl", hash = "sha256:dea97ac3ca18cd3ba8f3d04b5c1fe4aa60e58e8d9b7793d3bd595fdb04128d7a", size = 1709527 }, - { url = "https://files.pythonhosted.org/packages/b8/63/d06ecce90e2cf1747e29aeab9f823d21e5877a4c51b79720b2d3be7848f8/rapidfuzz-3.14.3-cp312-cp312-win_amd64.whl", hash = "sha256:b5100fd6bcee4d27f28f4e0a1c6b5127bc8ba7c2a9959cad9eab0bf4a7ab3329", size = 1538989 }, - { url = "https://files.pythonhosted.org/packages/fc/6d/beee32dcda64af8128aab3ace2ccb33d797ed58c434c6419eea015fec779/rapidfuzz-3.14.3-cp312-cp312-win_arm64.whl", hash = "sha256:4e49c9e992bc5fc873bd0fff7ef16a4405130ec42f2ce3d2b735ba5d3d4eb70f", size = 811161 }, - { url = "https://files.pythonhosted.org/packages/e4/4f/0d94d09646853bd26978cb3a7541b6233c5760687777fa97da8de0d9a6ac/rapidfuzz-3.14.3-cp313-cp313-macosx_10_13_x86_64.whl", hash = "sha256:dbcb726064b12f356bf10fffdb6db4b6dce5390b23627c08652b3f6e49aa56ae", size = 1939646 }, - { url = "https://files.pythonhosted.org/packages/b6/eb/f96aefc00f3bbdbab9c0657363ea8437a207d7545ac1c3789673e05d80bd/rapidfuzz-3.14.3-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:1704fc70d214294e554a2421b473779bcdeef715881c5e927dc0f11e1692a0ff", size = 1385512 }, - { url = "https://files.pythonhosted.org/packages/26/34/71c4f7749c12ee223dba90017a5947e8f03731a7cc9f489b662a8e9e643d/rapidfuzz-3.14.3-cp313-cp313-manylinux_2_26_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:cc65e72790ddfd310c2c8912b45106e3800fefe160b0c2ef4d6b6fec4e826457", size = 1373571 }, - { url = "https://files.pythonhosted.org/packages/32/00/ec8597a64f2be301ce1ee3290d067f49f6a7afb226b67d5f15b56d772ba5/rapidfuzz-3.14.3-cp313-cp313-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:43e38c1305cffae8472572a0584d4ffc2f130865586a81038ca3965301f7c97c", size = 3156759 }, - { url = "https://files.pythonhosted.org/packages/61/d5/b41eeb4930501cc899d5a9a7b5c9a33d85a670200d7e81658626dcc0ecc0/rapidfuzz-3.14.3-cp313-cp313-manylinux_2_31_armv7l.whl", hash = "sha256:e195a77d06c03c98b3fc06b8a28576ba824392ce40de8c708f96ce04849a052e", size = 1222067 }, - { url = "https://files.pythonhosted.org/packages/2a/7d/6d9abb4ffd1027c6ed837b425834f3bed8344472eb3a503ab55b3407c721/rapidfuzz-3.14.3-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:1b7ef2f4b8583a744338a18f12c69693c194fb6777c0e9ada98cd4d9e8f09d10", size = 2394775 }, - { url = "https://files.pythonhosted.org/packages/15/ce/4f3ab4c401c5a55364da1ffff8cc879fc97b4e5f4fa96033827da491a973/rapidfuzz-3.14.3-cp313-cp313-musllinux_1_2_armv7l.whl", hash = "sha256:a2135b138bcdcb4c3742d417f215ac2d8c2b87bde15b0feede231ae95f09ec41", size = 2526123 }, - { url = "https://files.pythonhosted.org/packages/c1/4b/54f804975376a328f57293bd817c12c9036171d15cf7292032e3f5820b2d/rapidfuzz-3.14.3-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:33a325ed0e8e1aa20c3e75f8ab057a7b248fdea7843c2a19ade0008906c14af0", size = 4262874 }, - { url = "https://files.pythonhosted.org/packages/e9/b6/958db27d8a29a50ee6edd45d33debd3ce732e7209183a72f57544cd5fe22/rapidfuzz-3.14.3-cp313-cp313-win32.whl", hash = "sha256:8383b6d0d92f6cd008f3c9216535be215a064b2cc890398a678b56e6d280cb63", size = 1707972 }, - { url = "https://files.pythonhosted.org/packages/07/75/fde1f334b0cec15b5946d9f84d73250fbfcc73c236b4bc1b25129d90876b/rapidfuzz-3.14.3-cp313-cp313-win_amd64.whl", hash = "sha256:e6b5e3036976f0fde888687d91be86d81f9ac5f7b02e218913c38285b756be6c", size = 1537011 }, - { url = "https://files.pythonhosted.org/packages/2e/d7/d83fe001ce599dc7ead57ba1debf923dc961b6bdce522b741e6b8c82f55c/rapidfuzz-3.14.3-cp313-cp313-win_arm64.whl", hash = "sha256:7ba009977601d8b0828bfac9a110b195b3e4e79b350dcfa48c11269a9f1918a0", size = 810744 }, - { url = "https://files.pythonhosted.org/packages/92/13/a486369e63ff3c1a58444d16b15c5feb943edd0e6c28a1d7d67cb8946b8f/rapidfuzz-3.14.3-cp313-cp313t-macosx_10_13_x86_64.whl", hash = "sha256:a0a28add871425c2fe94358c6300bbeb0bc2ed828ca003420ac6825408f5a424", size = 1967702 }, - { url = "https://files.pythonhosted.org/packages/f1/82/efad25e260b7810f01d6b69122685e355bed78c94a12784bac4e0beb2afb/rapidfuzz-3.14.3-cp313-cp313t-macosx_11_0_arm64.whl", hash = "sha256:010e12e2411a4854b0434f920e72b717c43f8ec48d57e7affe5c42ecfa05dd0e", size = 1410702 }, - { url = "https://files.pythonhosted.org/packages/ba/1a/34c977b860cde91082eae4a97ae503f43e0d84d4af301d857679b66f9869/rapidfuzz-3.14.3-cp313-cp313t-manylinux_2_26_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:5cfc3d57abd83c734d1714ec39c88a34dd69c85474918ebc21296f1e61eb5ca8", size = 1382337 }, - { url = "https://files.pythonhosted.org/packages/88/74/f50ea0e24a5880a9159e8fd256b84d8f4634c2f6b4f98028bdd31891d907/rapidfuzz-3.14.3-cp313-cp313t-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:89acb8cbb52904f763e5ac238083b9fc193bed8d1f03c80568b20e4cef43a519", size = 3165563 }, - { url = "https://files.pythonhosted.org/packages/e8/7a/e744359404d7737049c26099423fc54bcbf303de5d870d07d2fb1410f567/rapidfuzz-3.14.3-cp313-cp313t-manylinux_2_31_armv7l.whl", hash = "sha256:7d9af908c2f371bfb9c985bd134e295038e3031e666e4b2ade1e7cb7f5af2f1a", size = 1214727 }, - { url = "https://files.pythonhosted.org/packages/d3/2e/87adfe14ce75768ec6c2b8acd0e05e85e84be4be5e3d283cdae360afc4fe/rapidfuzz-3.14.3-cp313-cp313t-musllinux_1_2_aarch64.whl", hash = "sha256:1f1925619627f8798f8c3a391d81071336942e5fe8467bc3c567f982e7ce2897", size = 2403349 }, - { url = "https://files.pythonhosted.org/packages/70/17/6c0b2b2bff9c8b12e12624c07aa22e922b0c72a490f180fa9183d1ef2c75/rapidfuzz-3.14.3-cp313-cp313t-musllinux_1_2_armv7l.whl", hash = "sha256:152555187360978119e98ce3e8263d70dd0c40c7541193fc302e9b7125cf8f58", size = 2507596 }, - { url = "https://files.pythonhosted.org/packages/c3/d1/87852a7cbe4da7b962174c749a47433881a63a817d04f3e385ea9babcd9e/rapidfuzz-3.14.3-cp313-cp313t-musllinux_1_2_x86_64.whl", hash = "sha256:52619d25a09546b8db078981ca88939d72caa6b8701edd8b22e16482a38e799f", size = 4273595 }, - { url = "https://files.pythonhosted.org/packages/c1/ab/1d0354b7d1771a28fa7fe089bc23acec2bdd3756efa2419f463e3ed80e16/rapidfuzz-3.14.3-cp313-cp313t-win32.whl", hash = "sha256:489ce98a895c98cad284f0a47960c3e264c724cb4cfd47a1430fa091c0c25204", size = 1757773 }, - { url = "https://files.pythonhosted.org/packages/0b/0c/71ef356adc29e2bdf74cd284317b34a16b80258fa0e7e242dd92cc1e6d10/rapidfuzz-3.14.3-cp313-cp313t-win_amd64.whl", hash = "sha256:656e52b054d5b5c2524169240e50cfa080b04b1c613c5f90a2465e84888d6f15", size = 1576797 }, - { url = "https://files.pythonhosted.org/packages/fe/d2/0e64fc27bb08d4304aa3d11154eb5480bcf5d62d60140a7ee984dc07468a/rapidfuzz-3.14.3-cp313-cp313t-win_arm64.whl", hash = "sha256:c7e40c0a0af02ad6e57e89f62bef8604f55a04ecae90b0ceeda591bbf5923317", size = 829940 }, - { url = "https://files.pythonhosted.org/packages/32/6f/1b88aaeade83abc5418788f9e6b01efefcd1a69d65ded37d89cd1662be41/rapidfuzz-3.14.3-cp314-cp314-macosx_10_15_x86_64.whl", hash = "sha256:442125473b247227d3f2de807a11da6c08ccf536572d1be943f8e262bae7e4ea", size = 1942086 }, - { url = "https://files.pythonhosted.org/packages/a0/2c/b23861347436cb10f46c2bd425489ec462790faaa360a54a7ede5f78de88/rapidfuzz-3.14.3-cp314-cp314-macosx_11_0_arm64.whl", hash = "sha256:1ec0c8c0c3d4f97ced46b2e191e883f8c82dbbf6d5ebc1842366d7eff13cd5a6", size = 1386993 }, - { url = "https://files.pythonhosted.org/packages/83/86/5d72e2c060aa1fbdc1f7362d938f6b237dff91f5b9fc5dd7cc297e112250/rapidfuzz-3.14.3-cp314-cp314-manylinux_2_26_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:2dc37bc20272f388b8c3a4eba4febc6e77e50a8f450c472def4751e7678f55e4", size = 1379126 }, - { url = "https://files.pythonhosted.org/packages/c9/bc/ef2cee3e4d8b3fc22705ff519f0d487eecc756abdc7c25d53686689d6cf2/rapidfuzz-3.14.3-cp314-cp314-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:dee362e7e79bae940a5e2b3f6d09c6554db6a4e301cc68343886c08be99844f1", size = 3159304 }, - { url = "https://files.pythonhosted.org/packages/a0/36/dc5f2f62bbc7bc90be1f75eeaf49ed9502094bb19290dfb4747317b17f12/rapidfuzz-3.14.3-cp314-cp314-manylinux_2_31_armv7l.whl", hash = "sha256:4b39921df948388a863f0e267edf2c36302983459b021ab928d4b801cbe6a421", size = 1218207 }, - { url = "https://files.pythonhosted.org/packages/df/7e/8f4be75c1bc62f47edf2bbbe2370ee482fae655ebcc4718ac3827ead3904/rapidfuzz-3.14.3-cp314-cp314-musllinux_1_2_aarch64.whl", hash = "sha256:beda6aa9bc44d1d81242e7b291b446be352d3451f8217fcb068fc2933927d53b", size = 2401245 }, - { url = "https://files.pythonhosted.org/packages/05/38/f7c92759e1bb188dd05b80d11c630ba59b8d7856657baf454ff56059c2ab/rapidfuzz-3.14.3-cp314-cp314-musllinux_1_2_armv7l.whl", hash = "sha256:6a014ba09657abfcfeed64b7d09407acb29af436d7fc075b23a298a7e4a6b41c", size = 2518308 }, - { url = "https://files.pythonhosted.org/packages/c7/ac/85820f70fed5ecb5f1d9a55f1e1e2090ef62985ef41db289b5ac5ec56e28/rapidfuzz-3.14.3-cp314-cp314-musllinux_1_2_x86_64.whl", hash = "sha256:32eeafa3abce138bb725550c0e228fc7eaeec7059aa8093d9cbbec2b58c2371a", size = 4265011 }, - { url = "https://files.pythonhosted.org/packages/46/a9/616930721ea9835c918af7cde22bff17f9db3639b0c1a7f96684be7f5630/rapidfuzz-3.14.3-cp314-cp314-win32.whl", hash = "sha256:adb44d996fc610c7da8c5048775b21db60dd63b1548f078e95858c05c86876a3", size = 1742245 }, - { url = "https://files.pythonhosted.org/packages/06/8a/f2fa5e9635b1ccafda4accf0e38246003f69982d7c81f2faa150014525a4/rapidfuzz-3.14.3-cp314-cp314-win_amd64.whl", hash = "sha256:f3d15d8527e2b293e38ce6e437631af0708df29eafd7c9fc48210854c94472f9", size = 1584856 }, - { url = "https://files.pythonhosted.org/packages/ef/97/09e20663917678a6d60d8e0e29796db175b1165e2079830430342d5298be/rapidfuzz-3.14.3-cp314-cp314-win_arm64.whl", hash = "sha256:576e4b9012a67e0bf54fccb69a7b6c94d4e86a9540a62f1a5144977359133583", size = 833490 }, - { url = "https://files.pythonhosted.org/packages/03/1b/6b6084576ba87bf21877c77218a0c97ba98cb285b0c02eaaee3acd7c4513/rapidfuzz-3.14.3-cp314-cp314t-macosx_10_15_x86_64.whl", hash = "sha256:cec3c0da88562727dd5a5a364bd9efeb535400ff0bfb1443156dd139a1dd7b50", size = 1968658 }, - { url = "https://files.pythonhosted.org/packages/38/c0/fb02a0db80d95704b0a6469cc394e8c38501abf7e1c0b2afe3261d1510c2/rapidfuzz-3.14.3-cp314-cp314t-macosx_11_0_arm64.whl", hash = "sha256:d1fa009f8b1100e4880868137e7bf0501422898f7674f2adcd85d5a67f041296", size = 1410742 }, - { url = "https://files.pythonhosted.org/packages/a4/72/3fbf12819fc6afc8ec75a45204013b40979d068971e535a7f3512b05e765/rapidfuzz-3.14.3-cp314-cp314t-manylinux_2_26_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:1b86daa7419b5e8b180690efd1fdbac43ff19230803282521c5b5a9c83977655", size = 1382810 }, - { url = "https://files.pythonhosted.org/packages/0f/18/0f1991d59bb7eee28922a00f79d83eafa8c7bfb4e8edebf4af2a160e7196/rapidfuzz-3.14.3-cp314-cp314t-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:c7bd1816db05d6c5ffb3a4df0a2b7b56fb8c81ef584d08e37058afa217da91b1", size = 3166349 }, - { url = "https://files.pythonhosted.org/packages/0d/f0/baa958b1989c8f88c78bbb329e969440cf330b5a01a982669986495bb980/rapidfuzz-3.14.3-cp314-cp314t-manylinux_2_31_armv7l.whl", hash = "sha256:33da4bbaf44e9755b0ce192597f3bde7372fe2e381ab305f41b707a95ac57aa7", size = 1214994 }, - { url = "https://files.pythonhosted.org/packages/e4/a0/cd12ec71f9b2519a3954febc5740291cceabc64c87bc6433afcb36259f3b/rapidfuzz-3.14.3-cp314-cp314t-musllinux_1_2_aarch64.whl", hash = "sha256:3fecce764cf5a991ee2195a844196da840aba72029b2612f95ac68a8b74946bf", size = 2403919 }, - { url = "https://files.pythonhosted.org/packages/0b/ce/019bd2176c1644098eced4f0595cb4b3ef52e4941ac9a5854f209d0a6e16/rapidfuzz-3.14.3-cp314-cp314t-musllinux_1_2_armv7l.whl", hash = "sha256:ecd7453e02cf072258c3a6b8e930230d789d5d46cc849503729f9ce475d0e785", size = 2508346 }, - { url = "https://files.pythonhosted.org/packages/23/f8/be16c68e2c9e6c4f23e8f4adbb7bccc9483200087ed28ff76c5312da9b14/rapidfuzz-3.14.3-cp314-cp314t-musllinux_1_2_x86_64.whl", hash = "sha256:ea188aa00e9bcae8c8411f006a5f2f06c4607a02f24eab0d8dc58566aa911f35", size = 4274105 }, - { url = "https://files.pythonhosted.org/packages/a1/d1/5ab148e03f7e6ec8cd220ccf7af74d3aaa4de26dd96df58936beb7cba820/rapidfuzz-3.14.3-cp314-cp314t-win32.whl", hash = "sha256:7ccbf68100c170e9a0581accbe9291850936711548c6688ce3bfb897b8c589ad", size = 1793465 }, - { url = "https://files.pythonhosted.org/packages/cd/97/433b2d98e97abd9fff1c470a109b311669f44cdec8d0d5aa250aceaed1fb/rapidfuzz-3.14.3-cp314-cp314t-win_amd64.whl", hash = "sha256:9ec02e62ae765a318d6de38df609c57fc6dacc65c0ed1fd489036834fd8a620c", size = 1623491 }, - { url = "https://files.pythonhosted.org/packages/e2/f6/e2176eb94f94892441bce3ddc514c179facb65db245e7ce3356965595b19/rapidfuzz-3.14.3-cp314-cp314t-win_arm64.whl", hash = "sha256:e805e52322ae29aa945baf7168b6c898120fbc16d2b8f940b658a5e9e3999253", size = 851487 }, - { url = "https://files.pythonhosted.org/packages/c9/33/b5bd6475c7c27164b5becc9b0e3eb978f1e3640fea590dd3dced6006ee83/rapidfuzz-3.14.3-pp311-pypy311_pp73-macosx_10_15_x86_64.whl", hash = "sha256:7cf174b52cb3ef5d49e45d0a1133b7e7d0ecf770ed01f97ae9962c5c91d97d23", size = 1888499 }, - { url = "https://files.pythonhosted.org/packages/30/d2/89d65d4db4bb931beade9121bc71ad916b5fa9396e807d11b33731494e8e/rapidfuzz-3.14.3-pp311-pypy311_pp73-macosx_11_0_arm64.whl", hash = "sha256:442cba39957a008dfc5bdef21a9c3f4379e30ffb4e41b8555dbaf4887eca9300", size = 1336747 }, - { url = "https://files.pythonhosted.org/packages/85/33/cd87d92b23f0b06e8914a61cea6850c6d495ca027f669fab7a379041827a/rapidfuzz-3.14.3-pp311-pypy311_pp73-manylinux_2_26_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:1faa0f8f76ba75fd7b142c984947c280ef6558b5067af2ae9b8729b0a0f99ede", size = 1352187 }, - { url = "https://files.pythonhosted.org/packages/22/20/9d30b4a1ab26aac22fff17d21dec7e9089ccddfe25151d0a8bb57001dc3d/rapidfuzz-3.14.3-pp311-pypy311_pp73-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:1e6eefec45625c634926a9fd46c9e4f31118ac8f3156fff9494422cee45207e6", size = 3101472 }, - { url = "https://files.pythonhosted.org/packages/b1/ad/fa2d3e5c29a04ead7eaa731c7cd1f30f9ec3c77b3a578fdf90280797cbcb/rapidfuzz-3.14.3-pp311-pypy311_pp73-win_amd64.whl", hash = "sha256:56fefb4382bb12250f164250240b9dd7772e41c5c8ae976fd598a32292449cc5", size = 1511361 }, -] - -[[package]] -name = "regex" -version = "2025.11.3" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/cc/a9/546676f25e573a4cf00fe8e119b78a37b6a8fe2dc95cda877b30889c9c45/regex-2025.11.3.tar.gz", hash = "sha256:1fedc720f9bb2494ce31a58a1631f9c82df6a09b49c19517ea5cc280b4541e01", size = 414669 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/8a/d6/d788d52da01280a30a3f6268aef2aa71043bff359c618fea4c5b536654d5/regex-2025.11.3-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:2b441a4ae2c8049106e8b39973bfbddfb25a179dda2bdb99b0eeb60c40a6a3af", size = 488087 }, - { url = "https://files.pythonhosted.org/packages/69/39/abec3bd688ec9bbea3562de0fd764ff802976185f5ff22807bf0a2697992/regex-2025.11.3-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:2fa2eed3f76677777345d2f81ee89f5de2f5745910e805f7af7386a920fa7313", size = 290544 }, - { url = "https://files.pythonhosted.org/packages/39/b3/9a231475d5653e60002508f41205c61684bb2ffbf2401351ae2186897fc4/regex-2025.11.3-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:d8b4a27eebd684319bdf473d39f1d79eed36bf2cd34bd4465cdb4618d82b3d56", size = 288408 }, - { url = "https://files.pythonhosted.org/packages/c3/c5/1929a0491bd5ac2d1539a866768b88965fa8c405f3e16a8cef84313098d6/regex-2025.11.3-cp310-cp310-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:5cf77eac15bd264986c4a2c63353212c095b40f3affb2bc6b4ef80c4776c1a28", size = 781584 }, - { url = "https://files.pythonhosted.org/packages/ce/fd/16aa16cf5d497ef727ec966f74164fbe75d6516d3d58ac9aa989bc9cdaad/regex-2025.11.3-cp310-cp310-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:b7f9ee819f94c6abfa56ec7b1dbab586f41ebbdc0a57e6524bd5e7f487a878c7", size = 850733 }, - { url = "https://files.pythonhosted.org/packages/e6/49/3294b988855a221cb6565189edf5dc43239957427df2d81d4a6b15244f64/regex-2025.11.3-cp310-cp310-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:838441333bc90b829406d4a03cb4b8bf7656231b84358628b0406d803931ef32", size = 898691 }, - { url = "https://files.pythonhosted.org/packages/14/62/b56d29e70b03666193369bdbdedfdc23946dbe9f81dd78ce262c74d988ab/regex-2025.11.3-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:cfe6d3f0c9e3b7e8c0c694b24d25e677776f5ca26dce46fd6b0489f9c8339391", size = 791662 }, - { url = "https://files.pythonhosted.org/packages/15/fc/e4c31d061eced63fbf1ce9d853975f912c61a7d406ea14eda2dd355f48e7/regex-2025.11.3-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:2ab815eb8a96379a27c3b6157fcb127c8f59c36f043c1678110cea492868f1d5", size = 782587 }, - { url = "https://files.pythonhosted.org/packages/b2/bb/5e30c7394bcf63f0537121c23e796be67b55a8847c3956ae6068f4c70702/regex-2025.11.3-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:728a9d2d173a65b62bdc380b7932dd8e74ed4295279a8fe1021204ce210803e7", size = 774709 }, - { url = "https://files.pythonhosted.org/packages/c5/c4/fce773710af81b0cb37cb4ff0947e75d5d17dee304b93d940b87a67fc2f4/regex-2025.11.3-cp310-cp310-musllinux_1_2_ppc64le.whl", hash = "sha256:509dc827f89c15c66a0c216331260d777dd6c81e9a4e4f830e662b0bb296c313", size = 845773 }, - { url = "https://files.pythonhosted.org/packages/7b/5e/9466a7ec4b8ec282077095c6eb50a12a389d2e036581134d4919e8ca518c/regex-2025.11.3-cp310-cp310-musllinux_1_2_s390x.whl", hash = "sha256:849202cd789e5f3cf5dcc7822c34b502181b4824a65ff20ce82da5524e45e8e9", size = 836164 }, - { url = "https://files.pythonhosted.org/packages/95/18/82980a60e8ed1594eb3c89eb814fb276ef51b9af7caeab1340bfd8564af6/regex-2025.11.3-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:b6f78f98741dcc89607c16b1e9426ee46ce4bf31ac5e6b0d40e81c89f3481ea5", size = 779832 }, - { url = "https://files.pythonhosted.org/packages/03/cc/90ab0fdbe6dce064a42015433f9152710139fb04a8b81b4fb57a1cb63ffa/regex-2025.11.3-cp310-cp310-win32.whl", hash = "sha256:149eb0bba95231fb4f6d37c8f760ec9fa6fabf65bab555e128dde5f2475193ec", size = 265802 }, - { url = "https://files.pythonhosted.org/packages/34/9d/e9e8493a85f3b1ddc4a5014465f5c2b78c3ea1cbf238dcfde78956378041/regex-2025.11.3-cp310-cp310-win_amd64.whl", hash = "sha256:ee3a83ce492074c35a74cc76cf8235d49e77b757193a5365ff86e3f2f93db9fd", size = 277722 }, - { url = "https://files.pythonhosted.org/packages/15/c4/b54b24f553966564506dbf873a3e080aef47b356a3b39b5d5aba992b50db/regex-2025.11.3-cp310-cp310-win_arm64.whl", hash = "sha256:38af559ad934a7b35147716655d4a2f79fcef2d695ddfe06a06ba40ae631fa7e", size = 270289 }, - { url = "https://files.pythonhosted.org/packages/f7/90/4fb5056e5f03a7048abd2b11f598d464f0c167de4f2a51aa868c376b8c70/regex-2025.11.3-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:eadade04221641516fa25139273505a1c19f9bf97589a05bc4cfcd8b4a618031", size = 488081 }, - { url = "https://files.pythonhosted.org/packages/85/23/63e481293fac8b069d84fba0299b6666df720d875110efd0338406b5d360/regex-2025.11.3-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:feff9e54ec0dd3833d659257f5c3f5322a12eee58ffa360984b716f8b92983f4", size = 290554 }, - { url = "https://files.pythonhosted.org/packages/2b/9d/b101d0262ea293a0066b4522dfb722eb6a8785a8c3e084396a5f2c431a46/regex-2025.11.3-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:3b30bc921d50365775c09a7ed446359e5c0179e9e2512beec4a60cbcef6ddd50", size = 288407 }, - { url = "https://files.pythonhosted.org/packages/0c/64/79241c8209d5b7e00577ec9dca35cd493cc6be35b7d147eda367d6179f6d/regex-2025.11.3-cp311-cp311-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:f99be08cfead2020c7ca6e396c13543baea32343b7a9a5780c462e323bd8872f", size = 793418 }, - { url = "https://files.pythonhosted.org/packages/3d/e2/23cd5d3573901ce8f9757c92ca4db4d09600b865919b6d3e7f69f03b1afd/regex-2025.11.3-cp311-cp311-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:6dd329a1b61c0ee95ba95385fb0c07ea0d3fe1a21e1349fa2bec272636217118", size = 860448 }, - { url = "https://files.pythonhosted.org/packages/2a/4c/aecf31beeaa416d0ae4ecb852148d38db35391aac19c687b5d56aedf3a8b/regex-2025.11.3-cp311-cp311-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:4c5238d32f3c5269d9e87be0cf096437b7622b6920f5eac4fd202468aaeb34d2", size = 907139 }, - { url = "https://files.pythonhosted.org/packages/61/22/b8cb00df7d2b5e0875f60628594d44dba283e951b1ae17c12f99e332cc0a/regex-2025.11.3-cp311-cp311-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:10483eefbfb0adb18ee9474498c9a32fcf4e594fbca0543bb94c48bac6183e2e", size = 800439 }, - { url = "https://files.pythonhosted.org/packages/02/a8/c4b20330a5cdc7a8eb265f9ce593f389a6a88a0c5f280cf4d978f33966bc/regex-2025.11.3-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:78c2d02bb6e1da0720eedc0bad578049cad3f71050ef8cd065ecc87691bed2b0", size = 782965 }, - { url = "https://files.pythonhosted.org/packages/b4/4c/ae3e52988ae74af4b04d2af32fee4e8077f26e51b62ec2d12d246876bea2/regex-2025.11.3-cp311-cp311-musllinux_1_2_ppc64le.whl", hash = "sha256:e6b49cd2aad93a1790ce9cffb18964f6d3a4b0b3dbdbd5de094b65296fce6e58", size = 854398 }, - { url = "https://files.pythonhosted.org/packages/06/d1/a8b9cf45874eda14b2e275157ce3b304c87e10fb38d9fc26a6e14eb18227/regex-2025.11.3-cp311-cp311-musllinux_1_2_s390x.whl", hash = "sha256:885b26aa3ee56433b630502dc3d36ba78d186a00cc535d3806e6bfd9ed3c70ab", size = 845897 }, - { url = "https://files.pythonhosted.org/packages/ea/fe/1830eb0236be93d9b145e0bd8ab499f31602fe0999b1f19e99955aa8fe20/regex-2025.11.3-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:ddd76a9f58e6a00f8772e72cff8ebcff78e022be95edf018766707c730593e1e", size = 788906 }, - { url = "https://files.pythonhosted.org/packages/66/47/dc2577c1f95f188c1e13e2e69d8825a5ac582ac709942f8a03af42ed6e93/regex-2025.11.3-cp311-cp311-win32.whl", hash = "sha256:3e816cc9aac1cd3cc9a4ec4d860f06d40f994b5c7b4d03b93345f44e08cc68bf", size = 265812 }, - { url = "https://files.pythonhosted.org/packages/50/1e/15f08b2f82a9bbb510621ec9042547b54d11e83cb620643ebb54e4eb7d71/regex-2025.11.3-cp311-cp311-win_amd64.whl", hash = "sha256:087511f5c8b7dfbe3a03f5d5ad0c2a33861b1fc387f21f6f60825a44865a385a", size = 277737 }, - { url = "https://files.pythonhosted.org/packages/f4/fc/6500eb39f5f76c5e47a398df82e6b535a5e345f839581012a418b16f9cc3/regex-2025.11.3-cp311-cp311-win_arm64.whl", hash = "sha256:1ff0d190c7f68ae7769cd0313fe45820ba07ffebfddfaa89cc1eb70827ba0ddc", size = 270290 }, - { url = "https://files.pythonhosted.org/packages/e8/74/18f04cb53e58e3fb107439699bd8375cf5a835eec81084e0bddbd122e4c2/regex-2025.11.3-cp312-cp312-macosx_10_13_universal2.whl", hash = "sha256:bc8ab71e2e31b16e40868a40a69007bc305e1109bd4658eb6cad007e0bf67c41", size = 489312 }, - { url = "https://files.pythonhosted.org/packages/78/3f/37fcdd0d2b1e78909108a876580485ea37c91e1acf66d3bb8e736348f441/regex-2025.11.3-cp312-cp312-macosx_10_13_x86_64.whl", hash = "sha256:22b29dda7e1f7062a52359fca6e58e548e28c6686f205e780b02ad8ef710de36", size = 291256 }, - { url = "https://files.pythonhosted.org/packages/bf/26/0a575f58eb23b7ebd67a45fccbc02ac030b737b896b7e7a909ffe43ffd6a/regex-2025.11.3-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:3a91e4a29938bc1a082cc28fdea44be420bf2bebe2665343029723892eb073e1", size = 288921 }, - { url = "https://files.pythonhosted.org/packages/ea/98/6a8dff667d1af907150432cf5abc05a17ccd32c72a3615410d5365ac167a/regex-2025.11.3-cp312-cp312-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:08b884f4226602ad40c5d55f52bf91a9df30f513864e0054bad40c0e9cf1afb7", size = 798568 }, - { url = "https://files.pythonhosted.org/packages/64/15/92c1db4fa4e12733dd5a526c2dd2b6edcbfe13257e135fc0f6c57f34c173/regex-2025.11.3-cp312-cp312-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:3e0b11b2b2433d1c39c7c7a30e3f3d0aeeea44c2a8d0bae28f6b95f639927a69", size = 864165 }, - { url = "https://files.pythonhosted.org/packages/f9/e7/3ad7da8cdee1ce66c7cd37ab5ab05c463a86ffeb52b1a25fe7bd9293b36c/regex-2025.11.3-cp312-cp312-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:87eb52a81ef58c7ba4d45c3ca74e12aa4b4e77816f72ca25258a85b3ea96cb48", size = 912182 }, - { url = "https://files.pythonhosted.org/packages/84/bd/9ce9f629fcb714ffc2c3faf62b6766ecb7a585e1e885eb699bcf130a5209/regex-2025.11.3-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:a12ab1f5c29b4e93db518f5e3872116b7e9b1646c9f9f426f777b50d44a09e8c", size = 803501 }, - { url = "https://files.pythonhosted.org/packages/7c/0f/8dc2e4349d8e877283e6edd6c12bdcebc20f03744e86f197ab6e4492bf08/regex-2025.11.3-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:7521684c8c7c4f6e88e35ec89680ee1aa8358d3f09d27dfbdf62c446f5d4c695", size = 787842 }, - { url = "https://files.pythonhosted.org/packages/f9/73/cff02702960bc185164d5619c0c62a2f598a6abff6695d391b096237d4ab/regex-2025.11.3-cp312-cp312-musllinux_1_2_ppc64le.whl", hash = "sha256:7fe6e5440584e94cc4b3f5f4d98a25e29ca12dccf8873679a635638349831b98", size = 858519 }, - { url = "https://files.pythonhosted.org/packages/61/83/0e8d1ae71e15bc1dc36231c90b46ee35f9d52fab2e226b0e039e7ea9c10a/regex-2025.11.3-cp312-cp312-musllinux_1_2_s390x.whl", hash = "sha256:8e026094aa12b43f4fd74576714e987803a315c76edb6b098b9809db5de58f74", size = 850611 }, - { url = "https://files.pythonhosted.org/packages/c8/f5/70a5cdd781dcfaa12556f2955bf170cd603cb1c96a1827479f8faea2df97/regex-2025.11.3-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:435bbad13e57eb5606a68443af62bed3556de2f46deb9f7d4237bc2f1c9fb3a0", size = 789759 }, - { url = "https://files.pythonhosted.org/packages/59/9b/7c29be7903c318488983e7d97abcf8ebd3830e4c956c4c540005fcfb0462/regex-2025.11.3-cp312-cp312-win32.whl", hash = "sha256:3839967cf4dc4b985e1570fd8d91078f0c519f30491c60f9ac42a8db039be204", size = 266194 }, - { url = "https://files.pythonhosted.org/packages/1a/67/3b92df89f179d7c367be654ab5626ae311cb28f7d5c237b6bb976cd5fbbb/regex-2025.11.3-cp312-cp312-win_amd64.whl", hash = "sha256:e721d1b46e25c481dc5ded6f4b3f66c897c58d2e8cfdf77bbced84339108b0b9", size = 277069 }, - { url = "https://files.pythonhosted.org/packages/d7/55/85ba4c066fe5094d35b249c3ce8df0ba623cfd35afb22d6764f23a52a1c5/regex-2025.11.3-cp312-cp312-win_arm64.whl", hash = "sha256:64350685ff08b1d3a6fff33f45a9ca183dc1d58bbfe4981604e70ec9801bbc26", size = 270330 }, - { url = "https://files.pythonhosted.org/packages/e1/a7/dda24ebd49da46a197436ad96378f17df30ceb40e52e859fc42cac45b850/regex-2025.11.3-cp313-cp313-macosx_10_13_universal2.whl", hash = "sha256:c1e448051717a334891f2b9a620fe36776ebf3dd8ec46a0b877c8ae69575feb4", size = 489081 }, - { url = "https://files.pythonhosted.org/packages/19/22/af2dc751aacf88089836aa088a1a11c4f21a04707eb1b0478e8e8fb32847/regex-2025.11.3-cp313-cp313-macosx_10_13_x86_64.whl", hash = "sha256:9b5aca4d5dfd7fbfbfbdaf44850fcc7709a01146a797536a8f84952e940cca76", size = 291123 }, - { url = "https://files.pythonhosted.org/packages/a3/88/1a3ea5672f4b0a84802ee9891b86743438e7c04eb0b8f8c4e16a42375327/regex-2025.11.3-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:04d2765516395cf7dda331a244a3282c0f5ae96075f728629287dfa6f76ba70a", size = 288814 }, - { url = "https://files.pythonhosted.org/packages/fb/8c/f5987895bf42b8ddeea1b315c9fedcfe07cadee28b9c98cf50d00adcb14d/regex-2025.11.3-cp313-cp313-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:5d9903ca42bfeec4cebedba8022a7c97ad2aab22e09573ce9976ba01b65e4361", size = 798592 }, - { url = "https://files.pythonhosted.org/packages/99/2a/6591ebeede78203fa77ee46a1c36649e02df9eaa77a033d1ccdf2fcd5d4e/regex-2025.11.3-cp313-cp313-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:639431bdc89d6429f6721625e8129413980ccd62e9d3f496be618a41d205f160", size = 864122 }, - { url = "https://files.pythonhosted.org/packages/94/d6/be32a87cf28cf8ed064ff281cfbd49aefd90242a83e4b08b5a86b38e8eb4/regex-2025.11.3-cp313-cp313-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:f117efad42068f9715677c8523ed2be1518116d1c49b1dd17987716695181efe", size = 912272 }, - { url = "https://files.pythonhosted.org/packages/62/11/9bcef2d1445665b180ac7f230406ad80671f0fc2a6ffb93493b5dd8cd64c/regex-2025.11.3-cp313-cp313-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:4aecb6f461316adf9f1f0f6a4a1a3d79e045f9b71ec76055a791affa3b285850", size = 803497 }, - { url = "https://files.pythonhosted.org/packages/e5/a7/da0dc273d57f560399aa16d8a68ae7f9b57679476fc7ace46501d455fe84/regex-2025.11.3-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:3b3a5f320136873cc5561098dfab677eea139521cb9a9e8db98b7e64aef44cbc", size = 787892 }, - { url = "https://files.pythonhosted.org/packages/da/4b/732a0c5a9736a0b8d6d720d4945a2f1e6f38f87f48f3173559f53e8d5d82/regex-2025.11.3-cp313-cp313-musllinux_1_2_ppc64le.whl", hash = "sha256:75fa6f0056e7efb1f42a1c34e58be24072cb9e61a601340cc1196ae92326a4f9", size = 858462 }, - { url = "https://files.pythonhosted.org/packages/0c/f5/a2a03df27dc4c2d0c769220f5110ba8c4084b0bfa9ab0f9b4fcfa3d2b0fc/regex-2025.11.3-cp313-cp313-musllinux_1_2_s390x.whl", hash = "sha256:dbe6095001465294f13f1adcd3311e50dd84e5a71525f20a10bd16689c61ce0b", size = 850528 }, - { url = "https://files.pythonhosted.org/packages/d6/09/e1cd5bee3841c7f6eb37d95ca91cdee7100b8f88b81e41c2ef426910891a/regex-2025.11.3-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:454d9b4ae7881afbc25015b8627c16d88a597479b9dea82b8c6e7e2e07240dc7", size = 789866 }, - { url = "https://files.pythonhosted.org/packages/eb/51/702f5ea74e2a9c13d855a6a85b7f80c30f9e72a95493260193c07f3f8d74/regex-2025.11.3-cp313-cp313-win32.whl", hash = "sha256:28ba4d69171fc6e9896337d4fc63a43660002b7da53fc15ac992abcf3410917c", size = 266189 }, - { url = "https://files.pythonhosted.org/packages/8b/00/6e29bb314e271a743170e53649db0fdb8e8ff0b64b4f425f5602f4eb9014/regex-2025.11.3-cp313-cp313-win_amd64.whl", hash = "sha256:bac4200befe50c670c405dc33af26dad5a3b6b255dd6c000d92fe4629f9ed6a5", size = 277054 }, - { url = "https://files.pythonhosted.org/packages/25/f1/b156ff9f2ec9ac441710764dda95e4edaf5f36aca48246d1eea3f1fd96ec/regex-2025.11.3-cp313-cp313-win_arm64.whl", hash = "sha256:2292cd5a90dab247f9abe892ac584cb24f0f54680c73fcb4a7493c66c2bf2467", size = 270325 }, - { url = "https://files.pythonhosted.org/packages/20/28/fd0c63357caefe5680b8ea052131acbd7f456893b69cc2a90cc3e0dc90d4/regex-2025.11.3-cp313-cp313t-macosx_10_13_universal2.whl", hash = "sha256:1eb1ebf6822b756c723e09f5186473d93236c06c579d2cc0671a722d2ab14281", size = 491984 }, - { url = "https://files.pythonhosted.org/packages/df/ec/7014c15626ab46b902b3bcc4b28a7bae46d8f281fc7ea9c95e22fcaaa917/regex-2025.11.3-cp313-cp313t-macosx_10_13_x86_64.whl", hash = "sha256:1e00ec2970aab10dc5db34af535f21fcf32b4a31d99e34963419636e2f85ae39", size = 292673 }, - { url = "https://files.pythonhosted.org/packages/23/ab/3b952ff7239f20d05f1f99e9e20188513905f218c81d52fb5e78d2bf7634/regex-2025.11.3-cp313-cp313t-macosx_11_0_arm64.whl", hash = "sha256:a4cb042b615245d5ff9b3794f56be4138b5adc35a4166014d31d1814744148c7", size = 291029 }, - { url = "https://files.pythonhosted.org/packages/21/7e/3dc2749fc684f455f162dcafb8a187b559e2614f3826877d3844a131f37b/regex-2025.11.3-cp313-cp313t-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:44f264d4bf02f3176467d90b294d59bf1db9fe53c141ff772f27a8b456b2a9ed", size = 807437 }, - { url = "https://files.pythonhosted.org/packages/1b/0b/d529a85ab349c6a25d1ca783235b6e3eedf187247eab536797021f7126c6/regex-2025.11.3-cp313-cp313t-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:7be0277469bf3bd7a34a9c57c1b6a724532a0d235cd0dc4e7f4316f982c28b19", size = 873368 }, - { url = "https://files.pythonhosted.org/packages/7d/18/2d868155f8c9e3e9d8f9e10c64e9a9f496bb8f7e037a88a8bed26b435af6/regex-2025.11.3-cp313-cp313t-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:0d31e08426ff4b5b650f68839f5af51a92a5b51abd8554a60c2fbc7c71f25d0b", size = 914921 }, - { url = "https://files.pythonhosted.org/packages/2d/71/9d72ff0f354fa783fe2ba913c8734c3b433b86406117a8db4ea2bf1c7a2f/regex-2025.11.3-cp313-cp313t-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:e43586ce5bd28f9f285a6e729466841368c4a0353f6fd08d4ce4630843d3648a", size = 812708 }, - { url = "https://files.pythonhosted.org/packages/e7/19/ce4bf7f5575c97f82b6e804ffb5c4e940c62609ab2a0d9538d47a7fdf7d4/regex-2025.11.3-cp313-cp313t-musllinux_1_2_aarch64.whl", hash = "sha256:0f9397d561a4c16829d4e6ff75202c1c08b68a3bdbfe29dbfcdb31c9830907c6", size = 795472 }, - { url = "https://files.pythonhosted.org/packages/03/86/fd1063a176ffb7b2315f9a1b08d17b18118b28d9df163132615b835a26ee/regex-2025.11.3-cp313-cp313t-musllinux_1_2_ppc64le.whl", hash = "sha256:dd16e78eb18ffdb25ee33a0682d17912e8cc8a770e885aeee95020046128f1ce", size = 868341 }, - { url = "https://files.pythonhosted.org/packages/12/43/103fb2e9811205e7386366501bc866a164a0430c79dd59eac886a2822950/regex-2025.11.3-cp313-cp313t-musllinux_1_2_s390x.whl", hash = "sha256:ffcca5b9efe948ba0661e9df0fa50d2bc4b097c70b9810212d6b62f05d83b2dd", size = 854666 }, - { url = "https://files.pythonhosted.org/packages/7d/22/e392e53f3869b75804762c7c848bd2dd2abf2b70fb0e526f58724638bd35/regex-2025.11.3-cp313-cp313t-musllinux_1_2_x86_64.whl", hash = "sha256:c56b4d162ca2b43318ac671c65bd4d563e841a694ac70e1a976ac38fcf4ca1d2", size = 799473 }, - { url = "https://files.pythonhosted.org/packages/4f/f9/8bd6b656592f925b6845fcbb4d57603a3ac2fb2373344ffa1ed70aa6820a/regex-2025.11.3-cp313-cp313t-win32.whl", hash = "sha256:9ddc42e68114e161e51e272f667d640f97e84a2b9ef14b7477c53aac20c2d59a", size = 268792 }, - { url = "https://files.pythonhosted.org/packages/e5/87/0e7d603467775ff65cd2aeabf1b5b50cc1c3708556a8b849a2fa4dd1542b/regex-2025.11.3-cp313-cp313t-win_amd64.whl", hash = "sha256:7a7c7fdf755032ffdd72c77e3d8096bdcb0eb92e89e17571a196f03d88b11b3c", size = 280214 }, - { url = "https://files.pythonhosted.org/packages/8d/d0/2afc6f8e94e2b64bfb738a7c2b6387ac1699f09f032d363ed9447fd2bb57/regex-2025.11.3-cp313-cp313t-win_arm64.whl", hash = "sha256:df9eb838c44f570283712e7cff14c16329a9f0fb19ca492d21d4b7528ee6821e", size = 271469 }, - { url = "https://files.pythonhosted.org/packages/31/e9/f6e13de7e0983837f7b6d238ad9458800a874bf37c264f7923e63409944c/regex-2025.11.3-cp314-cp314-macosx_10_13_universal2.whl", hash = "sha256:9697a52e57576c83139d7c6f213d64485d3df5bf84807c35fa409e6c970801c6", size = 489089 }, - { url = "https://files.pythonhosted.org/packages/a3/5c/261f4a262f1fa65141c1b74b255988bd2fa020cc599e53b080667d591cfc/regex-2025.11.3-cp314-cp314-macosx_10_13_x86_64.whl", hash = "sha256:e18bc3f73bd41243c9b38a6d9f2366cd0e0137a9aebe2d8ff76c5b67d4c0a3f4", size = 291059 }, - { url = "https://files.pythonhosted.org/packages/8e/57/f14eeb7f072b0e9a5a090d1712741fd8f214ec193dba773cf5410108bb7d/regex-2025.11.3-cp314-cp314-macosx_11_0_arm64.whl", hash = "sha256:61a08bcb0ec14ff4e0ed2044aad948d0659604f824cbd50b55e30b0ec6f09c73", size = 288900 }, - { url = "https://files.pythonhosted.org/packages/3c/6b/1d650c45e99a9b327586739d926a1cd4e94666b1bd4af90428b36af66dc7/regex-2025.11.3-cp314-cp314-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:c9c30003b9347c24bcc210958c5d167b9e4f9be786cb380a7d32f14f9b84674f", size = 799010 }, - { url = "https://files.pythonhosted.org/packages/99/ee/d66dcbc6b628ce4e3f7f0cbbb84603aa2fc0ffc878babc857726b8aab2e9/regex-2025.11.3-cp314-cp314-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:4e1e592789704459900728d88d41a46fe3969b82ab62945560a31732ffc19a6d", size = 864893 }, - { url = "https://files.pythonhosted.org/packages/bf/2d/f238229f1caba7ac87a6c4153d79947fb0261415827ae0f77c304260c7d3/regex-2025.11.3-cp314-cp314-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:6538241f45eb5a25aa575dbba1069ad786f68a4f2773a29a2bd3dd1f9de787be", size = 911522 }, - { url = "https://files.pythonhosted.org/packages/bd/3d/22a4eaba214a917c80e04f6025d26143690f0419511e0116508e24b11c9b/regex-2025.11.3-cp314-cp314-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:bce22519c989bb72a7e6b36a199384c53db7722fe669ba891da75907fe3587db", size = 803272 }, - { url = "https://files.pythonhosted.org/packages/84/b1/03188f634a409353a84b5ef49754b97dbcc0c0f6fd6c8ede505a8960a0a4/regex-2025.11.3-cp314-cp314-musllinux_1_2_aarch64.whl", hash = "sha256:66d559b21d3640203ab9075797a55165d79017520685fb407b9234d72ab63c62", size = 787958 }, - { url = "https://files.pythonhosted.org/packages/99/6a/27d072f7fbf6fadd59c64d210305e1ff865cc3b78b526fd147db768c553b/regex-2025.11.3-cp314-cp314-musllinux_1_2_ppc64le.whl", hash = "sha256:669dcfb2e38f9e8c69507bace46f4889e3abbfd9b0c29719202883c0a603598f", size = 859289 }, - { url = "https://files.pythonhosted.org/packages/9a/70/1b3878f648e0b6abe023172dacb02157e685564853cc363d9961bcccde4e/regex-2025.11.3-cp314-cp314-musllinux_1_2_s390x.whl", hash = "sha256:32f74f35ff0f25a5021373ac61442edcb150731fbaa28286bbc8bb1582c89d02", size = 850026 }, - { url = "https://files.pythonhosted.org/packages/dd/d5/68e25559b526b8baab8e66839304ede68ff6727237a47727d240006bd0ff/regex-2025.11.3-cp314-cp314-musllinux_1_2_x86_64.whl", hash = "sha256:e6c7a21dffba883234baefe91bc3388e629779582038f75d2a5be918e250f0ed", size = 789499 }, - { url = "https://files.pythonhosted.org/packages/fc/df/43971264857140a350910d4e33df725e8c94dd9dee8d2e4729fa0d63d49e/regex-2025.11.3-cp314-cp314-win32.whl", hash = "sha256:795ea137b1d809eb6836b43748b12634291c0ed55ad50a7d72d21edf1cd565c4", size = 271604 }, - { url = "https://files.pythonhosted.org/packages/01/6f/9711b57dc6894a55faf80a4c1b5aa4f8649805cb9c7aef46f7d27e2b9206/regex-2025.11.3-cp314-cp314-win_amd64.whl", hash = "sha256:9f95fbaa0ee1610ec0fc6b26668e9917a582ba80c52cc6d9ada15e30aa9ab9ad", size = 280320 }, - { url = "https://files.pythonhosted.org/packages/f1/7e/f6eaa207d4377481f5e1775cdeb5a443b5a59b392d0065f3417d31d80f87/regex-2025.11.3-cp314-cp314-win_arm64.whl", hash = "sha256:dfec44d532be4c07088c3de2876130ff0fbeeacaa89a137decbbb5f665855a0f", size = 273372 }, - { url = "https://files.pythonhosted.org/packages/c3/06/49b198550ee0f5e4184271cee87ba4dfd9692c91ec55289e6282f0f86ccf/regex-2025.11.3-cp314-cp314t-macosx_10_13_universal2.whl", hash = "sha256:ba0d8a5d7f04f73ee7d01d974d47c5834f8a1b0224390e4fe7c12a3a92a78ecc", size = 491985 }, - { url = "https://files.pythonhosted.org/packages/ce/bf/abdafade008f0b1c9da10d934034cb670432d6cf6cbe38bbb53a1cfd6cf8/regex-2025.11.3-cp314-cp314t-macosx_10_13_x86_64.whl", hash = "sha256:442d86cf1cfe4faabf97db7d901ef58347efd004934da045c745e7b5bd57ac49", size = 292669 }, - { url = "https://files.pythonhosted.org/packages/f9/ef/0c357bb8edbd2ad8e273fcb9e1761bc37b8acbc6e1be050bebd6475f19c1/regex-2025.11.3-cp314-cp314t-macosx_11_0_arm64.whl", hash = "sha256:fd0a5e563c756de210bb964789b5abe4f114dacae9104a47e1a649b910361536", size = 291030 }, - { url = "https://files.pythonhosted.org/packages/79/06/edbb67257596649b8fb088d6aeacbcb248ac195714b18a65e018bf4c0b50/regex-2025.11.3-cp314-cp314t-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:bf3490bcbb985a1ae97b2ce9ad1c0f06a852d5b19dde9b07bdf25bf224248c95", size = 807674 }, - { url = "https://files.pythonhosted.org/packages/f4/d9/ad4deccfce0ea336296bd087f1a191543bb99ee1c53093dcd4c64d951d00/regex-2025.11.3-cp314-cp314t-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:3809988f0a8b8c9dcc0f92478d6501fac7200b9ec56aecf0ec21f4a2ec4b6009", size = 873451 }, - { url = "https://files.pythonhosted.org/packages/13/75/a55a4724c56ef13e3e04acaab29df26582f6978c000ac9cd6810ad1f341f/regex-2025.11.3-cp314-cp314t-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:f4ff94e58e84aedb9c9fce66d4ef9f27a190285b451420f297c9a09f2b9abee9", size = 914980 }, - { url = "https://files.pythonhosted.org/packages/67/1e/a1657ee15bd9116f70d4a530c736983eed997b361e20ecd8f5ca3759d5c5/regex-2025.11.3-cp314-cp314t-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:7eb542fd347ce61e1321b0a6b945d5701528dca0cd9759c2e3bb8bd57e47964d", size = 812852 }, - { url = "https://files.pythonhosted.org/packages/b8/6f/f7516dde5506a588a561d296b2d0044839de06035bb486b326065b4c101e/regex-2025.11.3-cp314-cp314t-musllinux_1_2_aarch64.whl", hash = "sha256:d6c2d5919075a1f2e413c00b056ea0c2f065b3f5fe83c3d07d325ab92dce51d6", size = 795566 }, - { url = "https://files.pythonhosted.org/packages/d9/dd/3d10b9e170cc16fb34cb2cef91513cf3df65f440b3366030631b2984a264/regex-2025.11.3-cp314-cp314t-musllinux_1_2_ppc64le.whl", hash = "sha256:3f8bf11a4827cc7ce5a53d4ef6cddd5ad25595d3c1435ef08f76825851343154", size = 868463 }, - { url = "https://files.pythonhosted.org/packages/f5/8e/935e6beff1695aa9085ff83195daccd72acc82c81793df480f34569330de/regex-2025.11.3-cp314-cp314t-musllinux_1_2_s390x.whl", hash = "sha256:22c12d837298651e5550ac1d964e4ff57c3f56965fc1812c90c9fb2028eaf267", size = 854694 }, - { url = "https://files.pythonhosted.org/packages/92/12/10650181a040978b2f5720a6a74d44f841371a3d984c2083fc1752e4acf6/regex-2025.11.3-cp314-cp314t-musllinux_1_2_x86_64.whl", hash = "sha256:62ba394a3dda9ad41c7c780f60f6e4a70988741415ae96f6d1bf6c239cf01379", size = 799691 }, - { url = "https://files.pythonhosted.org/packages/67/90/8f37138181c9a7690e7e4cb388debbd389342db3c7381d636d2875940752/regex-2025.11.3-cp314-cp314t-win32.whl", hash = "sha256:4bf146dca15cdd53224a1bf46d628bd7590e4a07fbb69e720d561aea43a32b38", size = 274583 }, - { url = "https://files.pythonhosted.org/packages/8f/cd/867f5ec442d56beb56f5f854f40abcfc75e11d10b11fdb1869dd39c63aaf/regex-2025.11.3-cp314-cp314t-win_amd64.whl", hash = "sha256:adad1a1bcf1c9e76346e091d22d23ac54ef28e1365117d99521631078dfec9de", size = 284286 }, - { url = "https://files.pythonhosted.org/packages/20/31/32c0c4610cbc070362bf1d2e4ea86d1ea29014d400a6d6c2486fcfd57766/regex-2025.11.3-cp314-cp314t-win_arm64.whl", hash = "sha256:c54f768482cef41e219720013cd05933b6f971d9562544d691c68699bf2b6801", size = 274741 }, -] - -[[package]] -name = "requests" -version = "2.32.5" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "certifi" }, - { name = "charset-normalizer" }, - { name = "idna" }, - { name = "urllib3" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/c9/74/b3ff8e6c8446842c3f5c837e9c3dfcfe2018ea6ecef224c710c85ef728f4/requests-2.32.5.tar.gz", hash = "sha256:dbba0bac56e100853db0ea71b82b4dfd5fe2bf6d3754a8893c3af500cec7d7cf", size = 134517 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/1e/db/4254e3eabe8020b458f1a747140d32277ec7a271daf1d235b70dc0b4e6e3/requests-2.32.5-py3-none-any.whl", hash = "sha256:2462f94637a34fd532264295e186976db0f5d453d1cdd31473c85a6a161affb6", size = 64738 }, -] - -[[package]] -name = "resampy" -version = "0.4.3" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "numba" }, - { name = "numpy" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/29/f1/34be702a69a5d272e844c98cee82351f880985cfbca0cc86378011078497/resampy-0.4.3.tar.gz", hash = "sha256:a0d1c28398f0e55994b739650afef4e3974115edbe96cd4bb81968425e916e47", size = 3080604 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/4d/b9/3b00ac340a1aab3389ebcc52c779914a44aadf7b0cb7a3bf053195735607/resampy-0.4.3-py3-none-any.whl", hash = "sha256:ad2ed64516b140a122d96704e32bc0f92b23f45419e8b8f478e5a05f83edcebd", size = 3076529 }, -] - -[[package]] -name = "rich" -version = "14.2.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "markdown-it-py" }, - { name = "pygments" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/fb/d2/8920e102050a0de7bfabeb4c4614a49248cf8d5d7a8d01885fbb24dc767a/rich-14.2.0.tar.gz", hash = "sha256:73ff50c7c0c1c77c8243079283f4edb376f0f6442433aecb8ce7e6d0b92d1fe4", size = 219990 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/25/7a/b0178788f8dc6cafce37a212c99565fa1fe7872c70c6c9c1e1a372d9d88f/rich-14.2.0-py3-none-any.whl", hash = "sha256:76bc51fe2e57d2b1be1f96c524b890b816e334ab4c1e45888799bfaab0021edd", size = 243393 }, -] - -[[package]] -name = "ruamel-yaml" -version = "0.19.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "ruamel-yaml-clibz", marker = "platform_python_implementation == 'CPython'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/0c/5d/8a1de57b5a11245c61c906d422cd1e66b6778e134a1c68823a451be5759c/ruamel_yaml-0.19.0.tar.gz", hash = "sha256:ff19233e1eb3e9301e7a3d437847713e361a80faace167639327efbe8c0e5f95", size = 142095 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/9b/3e/835d495068a4bb03419ce8c5464734ff6f3343df948e033cb5e5f81f7f08/ruamel_yaml-0.19.0-py3-none-any.whl", hash = "sha256:96ea8bafd9f3fdb0181ce3cc05e6ec02ce0a8788cbafa9b5a6e47c76fe26dfc6", size = 117777 }, -] - -[[package]] -name = "ruamel-yaml-clibz" -version = "0.3.4" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/8f/95/9bcc25e84703180c3941062796572e0fc73bd659086efdc4ef9b8af19e36/ruamel_yaml_clibz-0.3.4.tar.gz", hash = "sha256:e99077ac6aa4943af1000161a0cb793a379c5c8cd03ea8dd3803e0b58739b685", size = 231076 } - -[[package]] -name = "ruff" -version = "0.14.10" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/57/08/52232a877978dd8f9cf2aeddce3e611b40a63287dfca29b6b8da791f5e8d/ruff-0.14.10.tar.gz", hash = "sha256:9a2e830f075d1a42cd28420d7809ace390832a490ed0966fe373ba288e77aaf4", size = 5859763 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/60/01/933704d69f3f05ee16ef11406b78881733c186fe14b6a46b05cfcaf6d3b2/ruff-0.14.10-py3-none-linux_armv6l.whl", hash = "sha256:7a3ce585f2ade3e1f29ec1b92df13e3da262178df8c8bdf876f48fa0e8316c49", size = 13527080 }, - { url = "https://files.pythonhosted.org/packages/df/58/a0349197a7dfa603ffb7f5b0470391efa79ddc327c1e29c4851e85b09cc5/ruff-0.14.10-py3-none-macosx_10_12_x86_64.whl", hash = "sha256:674f9be9372907f7257c51f1d4fc902cb7cf014b9980152b802794317941f08f", size = 13797320 }, - { url = "https://files.pythonhosted.org/packages/7b/82/36be59f00a6082e38c23536df4e71cdbc6af8d7c707eade97fcad5c98235/ruff-0.14.10-py3-none-macosx_11_0_arm64.whl", hash = "sha256:d85713d522348837ef9df8efca33ccb8bd6fcfc86a2cde3ccb4bc9d28a18003d", size = 12918434 }, - { url = "https://files.pythonhosted.org/packages/a6/00/45c62a7f7e34da92a25804f813ebe05c88aa9e0c25e5cb5a7d23dd7450e3/ruff-0.14.10-py3-none-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:6987ebe0501ae4f4308d7d24e2d0fe3d7a98430f5adfd0f1fead050a740a3a77", size = 13371961 }, - { url = "https://files.pythonhosted.org/packages/40/31/a5906d60f0405f7e57045a70f2d57084a93ca7425f22e1d66904769d1628/ruff-0.14.10-py3-none-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:16a01dfb7b9e4eee556fbfd5392806b1b8550c9b4a9f6acd3dbe6812b193c70a", size = 13275629 }, - { url = "https://files.pythonhosted.org/packages/3e/60/61c0087df21894cf9d928dc04bcd4fb10e8b2e8dca7b1a276ba2155b2002/ruff-0.14.10-py3-none-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:7165d31a925b7a294465fa81be8c12a0e9b60fb02bf177e79067c867e71f8b1f", size = 14029234 }, - { url = "https://files.pythonhosted.org/packages/44/84/77d911bee3b92348b6e5dab5a0c898d87084ea03ac5dc708f46d88407def/ruff-0.14.10-py3-none-manylinux_2_17_ppc64.manylinux2014_ppc64.whl", hash = "sha256:c561695675b972effb0c0a45db233f2c816ff3da8dcfbe7dfc7eed625f218935", size = 15449890 }, - { url = "https://files.pythonhosted.org/packages/e9/36/480206eaefa24a7ec321582dda580443a8f0671fdbf6b1c80e9c3e93a16a/ruff-0.14.10-py3-none-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:4bb98fcbbc61725968893682fd4df8966a34611239c9fd07a1f6a07e7103d08e", size = 15123172 }, - { url = "https://files.pythonhosted.org/packages/5c/38/68e414156015ba80cef5473d57919d27dfb62ec804b96180bafdeaf0e090/ruff-0.14.10-py3-none-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:f24b47993a9d8cb858429e97bdf8544c78029f09b520af615c1d261bf827001d", size = 14460260 }, - { url = "https://files.pythonhosted.org/packages/b3/19/9e050c0dca8aba824d67cc0db69fb459c28d8cd3f6855b1405b3f29cc91d/ruff-0.14.10-py3-none-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:59aabd2e2c4fd614d2862e7939c34a532c04f1084476d6833dddef4afab87e9f", size = 14229978 }, - { url = "https://files.pythonhosted.org/packages/51/eb/e8dd1dd6e05b9e695aa9dd420f4577debdd0f87a5ff2fedda33c09e9be8c/ruff-0.14.10-py3-none-manylinux_2_31_riscv64.whl", hash = "sha256:213db2b2e44be8625002dbea33bb9c60c66ea2c07c084a00d55732689d697a7f", size = 14338036 }, - { url = "https://files.pythonhosted.org/packages/6a/12/f3e3a505db7c19303b70af370d137795fcfec136d670d5de5391e295c134/ruff-0.14.10-py3-none-musllinux_1_2_aarch64.whl", hash = "sha256:b914c40ab64865a17a9a5b67911d14df72346a634527240039eb3bd650e5979d", size = 13264051 }, - { url = "https://files.pythonhosted.org/packages/08/64/8c3a47eaccfef8ac20e0484e68e0772013eb85802f8a9f7603ca751eb166/ruff-0.14.10-py3-none-musllinux_1_2_armv7l.whl", hash = "sha256:1484983559f026788e3a5c07c81ef7d1e97c1c78ed03041a18f75df104c45405", size = 13283998 }, - { url = "https://files.pythonhosted.org/packages/12/84/534a5506f4074e5cc0529e5cd96cfc01bb480e460c7edf5af70d2bcae55e/ruff-0.14.10-py3-none-musllinux_1_2_i686.whl", hash = "sha256:c70427132db492d25f982fffc8d6c7535cc2fd2c83fc8888f05caaa248521e60", size = 13601891 }, - { url = "https://files.pythonhosted.org/packages/0d/1e/14c916087d8598917dbad9b2921d340f7884824ad6e9c55de948a93b106d/ruff-0.14.10-py3-none-musllinux_1_2_x86_64.whl", hash = "sha256:5bcf45b681e9f1ee6445d317ce1fa9d6cba9a6049542d1c3d5b5958986be8830", size = 14336660 }, - { url = "https://files.pythonhosted.org/packages/f2/1c/d7b67ab43f30013b47c12b42d1acd354c195351a3f7a1d67f59e54227ede/ruff-0.14.10-py3-none-win32.whl", hash = "sha256:104c49fc7ab73f3f3a758039adea978869a918f31b73280db175b43a2d9b51d6", size = 13196187 }, - { url = "https://files.pythonhosted.org/packages/fb/9c/896c862e13886fae2af961bef3e6312db9ebc6adc2b156fe95e615dee8c1/ruff-0.14.10-py3-none-win_amd64.whl", hash = "sha256:466297bd73638c6bdf06485683e812db1c00c7ac96d4ddd0294a338c62fdc154", size = 14661283 }, - { url = "https://files.pythonhosted.org/packages/74/31/b0e29d572670dca3674eeee78e418f20bdf97fa8aa9ea71380885e175ca0/ruff-0.14.10-py3-none-win_arm64.whl", hash = "sha256:e51d046cf6dda98a4633b8a8a771451107413b0f07183b2bef03f075599e44e6", size = 13729839 }, -] - -[[package]] -name = "sacremoses" -version = "0.1.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "click" }, - { name = "joblib" }, - { name = "regex" }, - { name = "tqdm" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/1d/51/fbdc4af4f6e85d26169e28be3763fe50ddfd0d4bf8b871422b0788dcc4d2/sacremoses-0.1.1.tar.gz", hash = "sha256:b6fd5d3a766b02154ed80b962ddca91e1fd25629c0978c7efba21ebccf663934", size = 883188 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/0b/f0/89ee2bc9da434bd78464f288fdb346bc2932f2ee80a90b2a4bbbac262c74/sacremoses-0.1.1-py3-none-any.whl", hash = "sha256:31e04c98b169bfd902144824d191825cd69220cdb4ae4bcf1ec58a7db5587b1a", size = 897476 }, -] - -[[package]] -name = "safetensors" -version = "0.7.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/29/9c/6e74567782559a63bd040a236edca26fd71bc7ba88de2ef35d75df3bca5e/safetensors-0.7.0.tar.gz", hash = "sha256:07663963b67e8bd9f0b8ad15bb9163606cd27cc5a1b96235a50d8369803b96b0", size = 200878 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/fa/47/aef6c06649039accf914afef490268e1067ed82be62bcfa5b7e886ad15e8/safetensors-0.7.0-cp38-abi3-macosx_10_12_x86_64.whl", hash = "sha256:c82f4d474cf725255d9e6acf17252991c3c8aac038d6ef363a4bf8be2f6db517", size = 467781 }, - { url = "https://files.pythonhosted.org/packages/e8/00/374c0c068e30cd31f1e1b46b4b5738168ec79e7689ca82ee93ddfea05109/safetensors-0.7.0-cp38-abi3-macosx_11_0_arm64.whl", hash = "sha256:94fd4858284736bb67a897a41608b5b0c2496c9bdb3bf2af1fa3409127f20d57", size = 447058 }, - { url = "https://files.pythonhosted.org/packages/f1/06/578ffed52c2296f93d7fd2d844cabfa92be51a587c38c8afbb8ae449ca89/safetensors-0.7.0-cp38-abi3-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e07d91d0c92a31200f25351f4acb2bc6aff7f48094e13ebb1d0fb995b54b6542", size = 491748 }, - { url = "https://files.pythonhosted.org/packages/ae/33/1debbbb70e4791dde185edb9413d1fe01619255abb64b300157d7f15dddd/safetensors-0.7.0-cp38-abi3-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:8469155f4cb518bafb4acf4865e8bb9d6804110d2d9bdcaa78564b9fd841e104", size = 503881 }, - { url = "https://files.pythonhosted.org/packages/8e/1c/40c2ca924d60792c3be509833df711b553c60effbd91da6f5284a83f7122/safetensors-0.7.0-cp38-abi3-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:54bef08bf00a2bff599982f6b08e8770e09cc012d7bba00783fc7ea38f1fb37d", size = 623463 }, - { url = "https://files.pythonhosted.org/packages/9b/3a/13784a9364bd43b0d61eef4bea2845039bc2030458b16594a1bd787ae26e/safetensors-0.7.0-cp38-abi3-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:42cb091236206bb2016d245c377ed383aa7f78691748f3bb6ee1bfa51ae2ce6a", size = 532855 }, - { url = "https://files.pythonhosted.org/packages/a0/60/429e9b1cb3fc651937727befe258ea24122d9663e4d5709a48c9cbfceecb/safetensors-0.7.0-cp38-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:dac7252938f0696ddea46f5e855dd3138444e82236e3be475f54929f0c510d48", size = 507152 }, - { url = "https://files.pythonhosted.org/packages/3c/a8/4b45e4e059270d17af60359713ffd83f97900d45a6afa73aaa0d737d48b6/safetensors-0.7.0-cp38-abi3-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:1d060c70284127fa805085d8f10fbd0962792aed71879d00864acda69dbab981", size = 541856 }, - { url = "https://files.pythonhosted.org/packages/06/87/d26d8407c44175d8ae164a95b5a62707fcc445f3c0c56108e37d98070a3d/safetensors-0.7.0-cp38-abi3-musllinux_1_2_aarch64.whl", hash = "sha256:cdab83a366799fa730f90a4ebb563e494f28e9e92c4819e556152ad55e43591b", size = 674060 }, - { url = "https://files.pythonhosted.org/packages/11/f5/57644a2ff08dc6325816ba7217e5095f17269dada2554b658442c66aed51/safetensors-0.7.0-cp38-abi3-musllinux_1_2_armv7l.whl", hash = "sha256:672132907fcad9f2aedcb705b2d7b3b93354a2aec1b2f706c4db852abe338f85", size = 771715 }, - { url = "https://files.pythonhosted.org/packages/86/31/17883e13a814bd278ae6e266b13282a01049b0c81341da7fd0e3e71a80a3/safetensors-0.7.0-cp38-abi3-musllinux_1_2_i686.whl", hash = "sha256:5d72abdb8a4d56d4020713724ba81dac065fedb7f3667151c4a637f1d3fb26c0", size = 714377 }, - { url = "https://files.pythonhosted.org/packages/4a/d8/0c8a7dc9b41dcac53c4cbf9df2b9c83e0e0097203de8b37a712b345c0be5/safetensors-0.7.0-cp38-abi3-musllinux_1_2_x86_64.whl", hash = "sha256:b0f6d66c1c538d5a94a73aa9ddca8ccc4227e6c9ff555322ea40bdd142391dd4", size = 677368 }, - { url = "https://files.pythonhosted.org/packages/05/e5/cb4b713c8a93469e3c5be7c3f8d77d307e65fe89673e731f5c2bfd0a9237/safetensors-0.7.0-cp38-abi3-win32.whl", hash = "sha256:c74af94bf3ac15ac4d0f2a7c7b4663a15f8c2ab15ed0fc7531ca61d0835eccba", size = 326423 }, - { url = "https://files.pythonhosted.org/packages/5d/e6/ec8471c8072382cb91233ba7267fd931219753bb43814cbc71757bfd4dab/safetensors-0.7.0-cp38-abi3-win_amd64.whl", hash = "sha256:d1239932053f56f3456f32eb9625590cc7582e905021f94636202a864d470755", size = 341380 }, - { url = "https://files.pythonhosted.org/packages/a7/6a/4d08d89a6fcbe905c5ae68b8b34f0791850882fc19782d0d02c65abbdf3b/safetensors-0.7.0-pp310-pypy310_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:f4729811a6640d019a4b7ba8638ee2fd21fa5ca8c7e7bdf0fed62068fcaac737", size = 492430 }, - { url = "https://files.pythonhosted.org/packages/dd/29/59ed8152b30f72c42d00d241e58eaca558ae9dbfa5695206e2e0f54c7063/safetensors-0.7.0-pp310-pypy310_pp73-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:12f49080303fa6bb424b362149a12949dfbbf1e06811a88f2307276b0c131afd", size = 503977 }, - { url = "https://files.pythonhosted.org/packages/d3/0b/4811bfec67fa260e791369b16dab105e4bae82686120554cc484064e22b4/safetensors-0.7.0-pp310-pypy310_pp73-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:0071bffba4150c2f46cae1432d31995d77acfd9f8db598b5d1a2ce67e8440ad2", size = 623890 }, - { url = "https://files.pythonhosted.org/packages/58/5b/632a58724221ef03d78ab65062e82a1010e1bef8e8e0b9d7c6d7b8044841/safetensors-0.7.0-pp310-pypy310_pp73-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:473b32699f4200e69801bf5abf93f1a4ecd432a70984df164fc22ccf39c4a6f3", size = 531885 }, -] - -[[package]] -name = "scikit-learn" -version = "1.7.2" -source = { registry = "https://pypi.org/simple" } -resolution-markers = [ - "python_full_version < '3.11' and sys_platform == 'linux'", - "python_full_version < '3.11' and sys_platform != 'linux'", -] -dependencies = [ - { name = "joblib", marker = "python_full_version < '3.11'" }, - { name = "numpy", marker = "python_full_version < '3.11'" }, - { name = "scipy", version = "1.15.3", source = { registry = "https://pypi.org/simple" }, marker = "python_full_version < '3.11'" }, - { name = "threadpoolctl", marker = "python_full_version < '3.11'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/98/c2/a7855e41c9d285dfe86dc50b250978105dce513d6e459ea66a6aeb0e1e0c/scikit_learn-1.7.2.tar.gz", hash = "sha256:20e9e49ecd130598f1ca38a1d85090e1a600147b9c02fa6f15d69cb53d968fda", size = 7193136 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/ba/3e/daed796fd69cce768b8788401cc464ea90b306fb196ae1ffed0b98182859/scikit_learn-1.7.2-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:6b33579c10a3081d076ab403df4a4190da4f4432d443521674637677dc91e61f", size = 9336221 }, - { url = "https://files.pythonhosted.org/packages/1c/ce/af9d99533b24c55ff4e18d9b7b4d9919bbc6cd8f22fe7a7be01519a347d5/scikit_learn-1.7.2-cp310-cp310-macosx_12_0_arm64.whl", hash = "sha256:36749fb62b3d961b1ce4fedf08fa57a1986cd409eff2d783bca5d4b9b5fce51c", size = 8653834 }, - { url = "https://files.pythonhosted.org/packages/58/0e/8c2a03d518fb6bd0b6b0d4b114c63d5f1db01ff0f9925d8eb10960d01c01/scikit_learn-1.7.2-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:7a58814265dfc52b3295b1900cfb5701589d30a8bb026c7540f1e9d3499d5ec8", size = 9660938 }, - { url = "https://files.pythonhosted.org/packages/2b/75/4311605069b5d220e7cf5adabb38535bd96f0079313cdbb04b291479b22a/scikit_learn-1.7.2-cp310-cp310-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:4a847fea807e278f821a0406ca01e387f97653e284ecbd9750e3ee7c90347f18", size = 9477818 }, - { url = "https://files.pythonhosted.org/packages/7f/9b/87961813c34adbca21a6b3f6b2bea344c43b30217a6d24cc437c6147f3e8/scikit_learn-1.7.2-cp310-cp310-win_amd64.whl", hash = "sha256:ca250e6836d10e6f402436d6463d6c0e4d8e0234cfb6a9a47835bd392b852ce5", size = 8886969 }, - { url = "https://files.pythonhosted.org/packages/43/83/564e141eef908a5863a54da8ca342a137f45a0bfb71d1d79704c9894c9d1/scikit_learn-1.7.2-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:c7509693451651cd7361d30ce4e86a1347493554f172b1c72a39300fa2aea79e", size = 9331967 }, - { url = "https://files.pythonhosted.org/packages/18/d6/ba863a4171ac9d7314c4d3fc251f015704a2caeee41ced89f321c049ed83/scikit_learn-1.7.2-cp311-cp311-macosx_12_0_arm64.whl", hash = "sha256:0486c8f827c2e7b64837c731c8feff72c0bd2b998067a8a9cbc10643c31f0fe1", size = 8648645 }, - { url = "https://files.pythonhosted.org/packages/ef/0e/97dbca66347b8cf0ea8b529e6bb9367e337ba2e8be0ef5c1a545232abfde/scikit_learn-1.7.2-cp311-cp311-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:89877e19a80c7b11a2891a27c21c4894fb18e2c2e077815bcade10d34287b20d", size = 9715424 }, - { url = "https://files.pythonhosted.org/packages/f7/32/1f3b22e3207e1d2c883a7e09abb956362e7d1bd2f14458c7de258a26ac15/scikit_learn-1.7.2-cp311-cp311-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:8da8bf89d4d79aaec192d2bda62f9b56ae4e5b4ef93b6a56b5de4977e375c1f1", size = 9509234 }, - { url = "https://files.pythonhosted.org/packages/9f/71/34ddbd21f1da67c7a768146968b4d0220ee6831e4bcbad3e03dd3eae88b6/scikit_learn-1.7.2-cp311-cp311-win_amd64.whl", hash = "sha256:9b7ed8d58725030568523e937c43e56bc01cadb478fc43c042a9aca1dacb3ba1", size = 8894244 }, - { url = "https://files.pythonhosted.org/packages/a7/aa/3996e2196075689afb9fce0410ebdb4a09099d7964d061d7213700204409/scikit_learn-1.7.2-cp312-cp312-macosx_10_13_x86_64.whl", hash = "sha256:8d91a97fa2b706943822398ab943cde71858a50245e31bc71dba62aab1d60a96", size = 9259818 }, - { url = "https://files.pythonhosted.org/packages/43/5d/779320063e88af9c4a7c2cf463ff11c21ac9c8bd730c4a294b0000b666c9/scikit_learn-1.7.2-cp312-cp312-macosx_12_0_arm64.whl", hash = "sha256:acbc0f5fd2edd3432a22c69bed78e837c70cf896cd7993d71d51ba6708507476", size = 8636997 }, - { url = "https://files.pythonhosted.org/packages/5c/d0/0c577d9325b05594fdd33aa970bf53fb673f051a45496842caee13cfd7fe/scikit_learn-1.7.2-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:e5bf3d930aee75a65478df91ac1225ff89cd28e9ac7bd1196853a9229b6adb0b", size = 9478381 }, - { url = "https://files.pythonhosted.org/packages/82/70/8bf44b933837ba8494ca0fc9a9ab60f1c13b062ad0197f60a56e2fc4c43e/scikit_learn-1.7.2-cp312-cp312-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:b4d6e9deed1a47aca9fe2f267ab8e8fe82ee20b4526b2c0cd9e135cea10feb44", size = 9300296 }, - { url = "https://files.pythonhosted.org/packages/c6/99/ed35197a158f1fdc2fe7c3680e9c70d0128f662e1fee4ed495f4b5e13db0/scikit_learn-1.7.2-cp312-cp312-win_amd64.whl", hash = "sha256:6088aa475f0785e01bcf8529f55280a3d7d298679f50c0bb70a2364a82d0b290", size = 8731256 }, - { url = "https://files.pythonhosted.org/packages/ae/93/a3038cb0293037fd335f77f31fe053b89c72f17b1c8908c576c29d953e84/scikit_learn-1.7.2-cp313-cp313-macosx_10_13_x86_64.whl", hash = "sha256:0b7dacaa05e5d76759fb071558a8b5130f4845166d88654a0f9bdf3eb57851b7", size = 9212382 }, - { url = "https://files.pythonhosted.org/packages/40/dd/9a88879b0c1104259136146e4742026b52df8540c39fec21a6383f8292c7/scikit_learn-1.7.2-cp313-cp313-macosx_12_0_arm64.whl", hash = "sha256:abebbd61ad9e1deed54cca45caea8ad5f79e1b93173dece40bb8e0c658dbe6fe", size = 8592042 }, - { url = "https://files.pythonhosted.org/packages/46/af/c5e286471b7d10871b811b72ae794ac5fe2989c0a2df07f0ec723030f5f5/scikit_learn-1.7.2-cp313-cp313-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:502c18e39849c0ea1a5d681af1dbcf15f6cce601aebb657aabbfe84133c1907f", size = 9434180 }, - { url = "https://files.pythonhosted.org/packages/f1/fd/df59faa53312d585023b2da27e866524ffb8faf87a68516c23896c718320/scikit_learn-1.7.2-cp313-cp313-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:7a4c328a71785382fe3fe676a9ecf2c86189249beff90bf85e22bdb7efaf9ae0", size = 9283660 }, - { url = "https://files.pythonhosted.org/packages/a7/c7/03000262759d7b6f38c836ff9d512f438a70d8a8ddae68ee80de72dcfb63/scikit_learn-1.7.2-cp313-cp313-win_amd64.whl", hash = "sha256:63a9afd6f7b229aad94618c01c252ce9e6fa97918c5ca19c9a17a087d819440c", size = 8702057 }, - { url = "https://files.pythonhosted.org/packages/55/87/ef5eb1f267084532c8e4aef98a28b6ffe7425acbfd64b5e2f2e066bc29b3/scikit_learn-1.7.2-cp313-cp313t-macosx_10_13_x86_64.whl", hash = "sha256:9acb6c5e867447b4e1390930e3944a005e2cb115922e693c08a323421a6966e8", size = 9558731 }, - { url = "https://files.pythonhosted.org/packages/93/f8/6c1e3fc14b10118068d7938878a9f3f4e6d7b74a8ddb1e5bed65159ccda8/scikit_learn-1.7.2-cp313-cp313t-macosx_12_0_arm64.whl", hash = "sha256:2a41e2a0ef45063e654152ec9d8bcfc39f7afce35b08902bfe290c2498a67a6a", size = 9038852 }, - { url = "https://files.pythonhosted.org/packages/83/87/066cafc896ee540c34becf95d30375fe5cbe93c3b75a0ee9aa852cd60021/scikit_learn-1.7.2-cp313-cp313t-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:98335fb98509b73385b3ab2bd0639b1f610541d3988ee675c670371d6a87aa7c", size = 9527094 }, - { url = "https://files.pythonhosted.org/packages/9c/2b/4903e1ccafa1f6453b1ab78413938c8800633988c838aa0be386cbb33072/scikit_learn-1.7.2-cp313-cp313t-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:191e5550980d45449126e23ed1d5e9e24b2c68329ee1f691a3987476e115e09c", size = 9367436 }, - { url = "https://files.pythonhosted.org/packages/b5/aa/8444be3cfb10451617ff9d177b3c190288f4563e6c50ff02728be67ad094/scikit_learn-1.7.2-cp313-cp313t-win_amd64.whl", hash = "sha256:57dc4deb1d3762c75d685507fbd0bc17160144b2f2ba4ccea5dc285ab0d0e973", size = 9275749 }, - { url = "https://files.pythonhosted.org/packages/d9/82/dee5acf66837852e8e68df6d8d3a6cb22d3df997b733b032f513d95205b7/scikit_learn-1.7.2-cp314-cp314-macosx_10_13_x86_64.whl", hash = "sha256:fa8f63940e29c82d1e67a45d5297bdebbcb585f5a5a50c4914cc2e852ab77f33", size = 9208906 }, - { url = "https://files.pythonhosted.org/packages/3c/30/9029e54e17b87cb7d50d51a5926429c683d5b4c1732f0507a6c3bed9bf65/scikit_learn-1.7.2-cp314-cp314-macosx_12_0_arm64.whl", hash = "sha256:f95dc55b7902b91331fa4e5845dd5bde0580c9cd9612b1b2791b7e80c3d32615", size = 8627836 }, - { url = "https://files.pythonhosted.org/packages/60/18/4a52c635c71b536879f4b971c2cedf32c35ee78f48367885ed8025d1f7ee/scikit_learn-1.7.2-cp314-cp314-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:9656e4a53e54578ad10a434dc1f993330568cfee176dff07112b8785fb413106", size = 9426236 }, - { url = "https://files.pythonhosted.org/packages/99/7e/290362f6ab582128c53445458a5befd471ed1ea37953d5bcf80604619250/scikit_learn-1.7.2-cp314-cp314-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:96dc05a854add0e50d3f47a1ef21a10a595016da5b007c7d9cd9d0bffd1fcc61", size = 9312593 }, - { url = "https://files.pythonhosted.org/packages/8e/87/24f541b6d62b1794939ae6422f8023703bbf6900378b2b34e0b4384dfefd/scikit_learn-1.7.2-cp314-cp314-win_amd64.whl", hash = "sha256:bb24510ed3f9f61476181e4db51ce801e2ba37541def12dc9333b946fc7a9cf8", size = 8820007 }, -] - -[[package]] -name = "scikit-learn" -version = "1.8.0" -source = { registry = "https://pypi.org/simple" } -resolution-markers = [ - "python_full_version >= '3.14' and sys_platform == 'linux'", - "python_full_version == '3.13.*' and sys_platform == 'linux'", - "python_full_version >= '3.14' and sys_platform != 'linux'", - "python_full_version == '3.13.*' and sys_platform != 'linux'", - "python_full_version == '3.12.*' and sys_platform == 'linux'", - "python_full_version == '3.12.*' and sys_platform != 'linux'", - "python_full_version == '3.11.*' and sys_platform == 'linux'", - "python_full_version == '3.11.*' and sys_platform != 'linux'", -] -dependencies = [ - { name = "joblib", marker = "python_full_version >= '3.11'" }, - { name = "numpy", marker = "python_full_version >= '3.11'" }, - { name = "scipy", version = "1.16.3", source = { registry = "https://pypi.org/simple" }, marker = "python_full_version >= '3.11'" }, - { name = "threadpoolctl", marker = "python_full_version >= '3.11'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/0e/d4/40988bf3b8e34feec1d0e6a051446b1f66225f8529b9309becaeef62b6c4/scikit_learn-1.8.0.tar.gz", hash = "sha256:9bccbb3b40e3de10351f8f5068e105d0f4083b1a65fa07b6634fbc401a6287fd", size = 7335585 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/c9/92/53ea2181da8ac6bf27170191028aee7251f8f841f8d3edbfdcaf2008fde9/scikit_learn-1.8.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:146b4d36f800c013d267b29168813f7a03a43ecd2895d04861f1240b564421da", size = 8595835 }, - { url = "https://files.pythonhosted.org/packages/01/18/d154dc1638803adf987910cdd07097d9c526663a55666a97c124d09fb96a/scikit_learn-1.8.0-cp311-cp311-macosx_12_0_arm64.whl", hash = "sha256:f984ca4b14914e6b4094c5d52a32ea16b49832c03bd17a110f004db3c223e8e1", size = 8080381 }, - { url = "https://files.pythonhosted.org/packages/8a/44/226142fcb7b7101e64fdee5f49dbe6288d4c7af8abf593237b70fca080a4/scikit_learn-1.8.0-cp311-cp311-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:5e30adb87f0cc81c7690a84f7932dd66be5bac57cfe16b91cb9151683a4a2d3b", size = 8799632 }, - { url = "https://files.pythonhosted.org/packages/36/4d/4a67f30778a45d542bbea5db2dbfa1e9e100bf9ba64aefe34215ba9f11f6/scikit_learn-1.8.0-cp311-cp311-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:ada8121bcb4dac28d930febc791a69f7cb1673c8495e5eee274190b73a4559c1", size = 9103788 }, - { url = "https://files.pythonhosted.org/packages/89/3c/45c352094cfa60050bcbb967b1faf246b22e93cb459f2f907b600f2ceda5/scikit_learn-1.8.0-cp311-cp311-win_amd64.whl", hash = "sha256:c57b1b610bd1f40ba43970e11ce62821c2e6569e4d74023db19c6b26f246cb3b", size = 8081706 }, - { url = "https://files.pythonhosted.org/packages/3d/46/5416595bb395757f754feb20c3d776553a386b661658fb21b7c814e89efe/scikit_learn-1.8.0-cp311-cp311-win_arm64.whl", hash = "sha256:2838551e011a64e3053ad7618dda9310175f7515f1742fa2d756f7c874c05961", size = 7688451 }, - { url = "https://files.pythonhosted.org/packages/90/74/e6a7cc4b820e95cc38cf36cd74d5aa2b42e8ffc2d21fe5a9a9c45c1c7630/scikit_learn-1.8.0-cp312-cp312-macosx_10_13_x86_64.whl", hash = "sha256:5fb63362b5a7ddab88e52b6dbb47dac3fd7dafeee740dc6c8d8a446ddedade8e", size = 8548242 }, - { url = "https://files.pythonhosted.org/packages/49/d8/9be608c6024d021041c7f0b3928d4749a706f4e2c3832bbede4fb4f58c95/scikit_learn-1.8.0-cp312-cp312-macosx_12_0_arm64.whl", hash = "sha256:5025ce924beccb28298246e589c691fe1b8c1c96507e6d27d12c5fadd85bfd76", size = 8079075 }, - { url = "https://files.pythonhosted.org/packages/dd/47/f187b4636ff80cc63f21cd40b7b2d177134acaa10f6bb73746130ee8c2e5/scikit_learn-1.8.0-cp312-cp312-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:4496bb2cf7a43ce1a2d7524a79e40bc5da45cf598dbf9545b7e8316ccba47bb4", size = 8660492 }, - { url = "https://files.pythonhosted.org/packages/97/74/b7a304feb2b49df9fafa9382d4d09061a96ee9a9449a7cbea7988dda0828/scikit_learn-1.8.0-cp312-cp312-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:a0bcfe4d0d14aec44921545fd2af2338c7471de9cb701f1da4c9d85906ab847a", size = 8931904 }, - { url = "https://files.pythonhosted.org/packages/9f/c4/0ab22726a04ede56f689476b760f98f8f46607caecff993017ac1b64aa5d/scikit_learn-1.8.0-cp312-cp312-win_amd64.whl", hash = "sha256:35c007dedb2ffe38fe3ee7d201ebac4a2deccd2408e8621d53067733e3c74809", size = 8019359 }, - { url = "https://files.pythonhosted.org/packages/24/90/344a67811cfd561d7335c1b96ca21455e7e472d281c3c279c4d3f2300236/scikit_learn-1.8.0-cp312-cp312-win_arm64.whl", hash = "sha256:8c497fff237d7b4e07e9ef1a640887fa4fb765647f86fbe00f969ff6280ce2bb", size = 7641898 }, - { url = "https://files.pythonhosted.org/packages/03/aa/e22e0768512ce9255eba34775be2e85c2048da73da1193e841707f8f039c/scikit_learn-1.8.0-cp313-cp313-macosx_10_13_x86_64.whl", hash = "sha256:0d6ae97234d5d7079dc0040990a6f7aeb97cb7fa7e8945f1999a429b23569e0a", size = 8513770 }, - { url = "https://files.pythonhosted.org/packages/58/37/31b83b2594105f61a381fc74ca19e8780ee923be2d496fcd8d2e1147bd99/scikit_learn-1.8.0-cp313-cp313-macosx_12_0_arm64.whl", hash = "sha256:edec98c5e7c128328124a029bceb09eda2d526997780fef8d65e9a69eead963e", size = 8044458 }, - { url = "https://files.pythonhosted.org/packages/2d/5a/3f1caed8765f33eabb723596666da4ebbf43d11e96550fb18bdec42b467b/scikit_learn-1.8.0-cp313-cp313-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:74b66d8689d52ed04c271e1329f0c61635bcaf5b926db9b12d58914cdc01fe57", size = 8610341 }, - { url = "https://files.pythonhosted.org/packages/38/cf/06896db3f71c75902a8e9943b444a56e727418f6b4b4a90c98c934f51ed4/scikit_learn-1.8.0-cp313-cp313-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:8fdf95767f989b0cfedb85f7ed8ca215d4be728031f56ff5a519ee1e3276dc2e", size = 8900022 }, - { url = "https://files.pythonhosted.org/packages/1c/f9/9b7563caf3ec8873e17a31401858efab6b39a882daf6c1bfa88879c0aa11/scikit_learn-1.8.0-cp313-cp313-win_amd64.whl", hash = "sha256:2de443b9373b3b615aec1bb57f9baa6bb3a9bd093f1269ba95c17d870422b271", size = 7989409 }, - { url = "https://files.pythonhosted.org/packages/49/bd/1f4001503650e72c4f6009ac0c4413cb17d2d601cef6f71c0453da2732fc/scikit_learn-1.8.0-cp313-cp313-win_arm64.whl", hash = "sha256:eddde82a035681427cbedded4e6eff5e57fa59216c2e3e90b10b19ab1d0a65c3", size = 7619760 }, - { url = "https://files.pythonhosted.org/packages/d2/7d/a630359fc9dcc95496588c8d8e3245cc8fd81980251079bc09c70d41d951/scikit_learn-1.8.0-cp313-cp313t-macosx_10_13_x86_64.whl", hash = "sha256:7cc267b6108f0a1499a734167282c00c4ebf61328566b55ef262d48e9849c735", size = 8826045 }, - { url = "https://files.pythonhosted.org/packages/cc/56/a0c86f6930cfcd1c7054a2bc417e26960bb88d32444fe7f71d5c2cfae891/scikit_learn-1.8.0-cp313-cp313t-macosx_12_0_arm64.whl", hash = "sha256:fe1c011a640a9f0791146011dfd3c7d9669785f9fed2b2a5f9e207536cf5c2fd", size = 8420324 }, - { url = "https://files.pythonhosted.org/packages/46/1e/05962ea1cebc1cf3876667ecb14c283ef755bf409993c5946ade3b77e303/scikit_learn-1.8.0-cp313-cp313t-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:72358cce49465d140cc4e7792015bb1f0296a9742d5622c67e31399b75468b9e", size = 8680651 }, - { url = "https://files.pythonhosted.org/packages/fe/56/a85473cd75f200c9759e3a5f0bcab2d116c92a8a02ee08ccd73b870f8bb4/scikit_learn-1.8.0-cp313-cp313t-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:80832434a6cc114f5219211eec13dcbc16c2bac0e31ef64c6d346cde3cf054cb", size = 8925045 }, - { url = "https://files.pythonhosted.org/packages/cc/b7/64d8cfa896c64435ae57f4917a548d7ac7a44762ff9802f75a79b77cb633/scikit_learn-1.8.0-cp313-cp313t-win_amd64.whl", hash = "sha256:ee787491dbfe082d9c3013f01f5991658b0f38aa8177e4cd4bf434c58f551702", size = 8507994 }, - { url = "https://files.pythonhosted.org/packages/5e/37/e192ea709551799379958b4c4771ec507347027bb7c942662c7fbeba31cb/scikit_learn-1.8.0-cp313-cp313t-win_arm64.whl", hash = "sha256:bf97c10a3f5a7543f9b88cbf488d33d175e9146115a451ae34568597ba33dcde", size = 7869518 }, - { url = "https://files.pythonhosted.org/packages/24/05/1af2c186174cc92dcab2233f327336058c077d38f6fe2aceb08e6ab4d509/scikit_learn-1.8.0-cp314-cp314-macosx_10_15_x86_64.whl", hash = "sha256:c22a2da7a198c28dd1a6e1136f19c830beab7fdca5b3e5c8bba8394f8a5c45b3", size = 8528667 }, - { url = "https://files.pythonhosted.org/packages/a8/25/01c0af38fe969473fb292bba9dc2b8f9b451f3112ff242c647fee3d0dfe7/scikit_learn-1.8.0-cp314-cp314-macosx_12_0_arm64.whl", hash = "sha256:6b595b07a03069a2b1740dc08c2299993850ea81cce4fe19b2421e0c970de6b7", size = 8066524 }, - { url = "https://files.pythonhosted.org/packages/be/ce/a0623350aa0b68647333940ee46fe45086c6060ec604874e38e9ab7d8e6c/scikit_learn-1.8.0-cp314-cp314-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:29ffc74089f3d5e87dfca4c2c8450f88bdc61b0fc6ed5d267f3988f19a1309f6", size = 8657133 }, - { url = "https://files.pythonhosted.org/packages/b8/cb/861b41341d6f1245e6ca80b1c1a8c4dfce43255b03df034429089ca2a2c5/scikit_learn-1.8.0-cp314-cp314-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:fb65db5d7531bccf3a4f6bec3462223bea71384e2cda41da0f10b7c292b9e7c4", size = 8923223 }, - { url = "https://files.pythonhosted.org/packages/76/18/a8def8f91b18cd1ba6e05dbe02540168cb24d47e8dcf69e8d00b7da42a08/scikit_learn-1.8.0-cp314-cp314-win_amd64.whl", hash = "sha256:56079a99c20d230e873ea40753102102734c5953366972a71d5cb39a32bc40c6", size = 8096518 }, - { url = "https://files.pythonhosted.org/packages/d1/77/482076a678458307f0deb44e29891d6022617b2a64c840c725495bee343f/scikit_learn-1.8.0-cp314-cp314-win_arm64.whl", hash = "sha256:3bad7565bc9cf37ce19a7c0d107742b320c1285df7aab1a6e2d28780df167242", size = 7754546 }, - { url = "https://files.pythonhosted.org/packages/2d/d1/ef294ca754826daa043b2a104e59960abfab4cf653891037d19dd5b6f3cf/scikit_learn-1.8.0-cp314-cp314t-macosx_10_15_x86_64.whl", hash = "sha256:4511be56637e46c25721e83d1a9cea9614e7badc7040c4d573d75fbe257d6fd7", size = 8848305 }, - { url = "https://files.pythonhosted.org/packages/5b/e2/b1f8b05138ee813b8e1a4149f2f0d289547e60851fd1bb268886915adbda/scikit_learn-1.8.0-cp314-cp314t-macosx_12_0_arm64.whl", hash = "sha256:a69525355a641bf8ef136a7fa447672fb54fe8d60cab5538d9eb7c6438543fb9", size = 8432257 }, - { url = "https://files.pythonhosted.org/packages/26/11/c32b2138a85dcb0c99f6afd13a70a951bfdff8a6ab42d8160522542fb647/scikit_learn-1.8.0-cp314-cp314t-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:c2656924ec73e5939c76ac4c8b026fc203b83d8900362eb2599d8aee80e4880f", size = 8678673 }, - { url = "https://files.pythonhosted.org/packages/c7/57/51f2384575bdec454f4fe4e7a919d696c9ebce914590abf3e52d47607ab8/scikit_learn-1.8.0-cp314-cp314t-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:15fc3b5d19cc2be65404786857f2e13c70c83dd4782676dd6814e3b89dc8f5b9", size = 8922467 }, - { url = "https://files.pythonhosted.org/packages/35/4d/748c9e2872637a57981a04adc038dacaa16ba8ca887b23e34953f0b3f742/scikit_learn-1.8.0-cp314-cp314t-win_amd64.whl", hash = "sha256:00d6f1d66fbcf4eba6e356e1420d33cc06c70a45bb1363cd6f6a8e4ebbbdece2", size = 8774395 }, - { url = "https://files.pythonhosted.org/packages/60/22/d7b2ebe4704a5e50790ba089d5c2ae308ab6bb852719e6c3bd4f04c3a363/scikit_learn-1.8.0-cp314-cp314t-win_arm64.whl", hash = "sha256:f28dd15c6bb0b66ba09728cf09fd8736c304be29409bd8445a080c1280619e8c", size = 8002647 }, -] - -[[package]] -name = "scipy" -version = "1.15.3" -source = { registry = "https://pypi.org/simple" } -resolution-markers = [ - "python_full_version < '3.11' and sys_platform == 'linux'", - "python_full_version < '3.11' and sys_platform != 'linux'", -] -dependencies = [ - { name = "numpy", marker = "python_full_version < '3.11'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/0f/37/6964b830433e654ec7485e45a00fc9a27cf868d622838f6b6d9c5ec0d532/scipy-1.15.3.tar.gz", hash = "sha256:eae3cf522bc7df64b42cad3925c876e1b0b6c35c1337c93e12c0f366f55b0eaf", size = 59419214 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/78/2f/4966032c5f8cc7e6a60f1b2e0ad686293b9474b65246b0c642e3ef3badd0/scipy-1.15.3-cp310-cp310-macosx_10_13_x86_64.whl", hash = "sha256:a345928c86d535060c9c2b25e71e87c39ab2f22fc96e9636bd74d1dbf9de448c", size = 38702770 }, - { url = "https://files.pythonhosted.org/packages/a0/6e/0c3bf90fae0e910c274db43304ebe25a6b391327f3f10b5dcc638c090795/scipy-1.15.3-cp310-cp310-macosx_12_0_arm64.whl", hash = "sha256:ad3432cb0f9ed87477a8d97f03b763fd1d57709f1bbde3c9369b1dff5503b253", size = 30094511 }, - { url = "https://files.pythonhosted.org/packages/ea/b1/4deb37252311c1acff7f101f6453f0440794f51b6eacb1aad4459a134081/scipy-1.15.3-cp310-cp310-macosx_14_0_arm64.whl", hash = "sha256:aef683a9ae6eb00728a542b796f52a5477b78252edede72b8327a886ab63293f", size = 22368151 }, - { url = "https://files.pythonhosted.org/packages/38/7d/f457626e3cd3c29b3a49ca115a304cebb8cc6f31b04678f03b216899d3c6/scipy-1.15.3-cp310-cp310-macosx_14_0_x86_64.whl", hash = "sha256:1c832e1bd78dea67d5c16f786681b28dd695a8cb1fb90af2e27580d3d0967e92", size = 25121732 }, - { url = "https://files.pythonhosted.org/packages/db/0a/92b1de4a7adc7a15dcf5bddc6e191f6f29ee663b30511ce20467ef9b82e4/scipy-1.15.3-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:263961f658ce2165bbd7b99fa5135195c3a12d9bef045345016b8b50c315cb82", size = 35547617 }, - { url = "https://files.pythonhosted.org/packages/8e/6d/41991e503e51fc1134502694c5fa7a1671501a17ffa12716a4a9151af3df/scipy-1.15.3-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9e2abc762b0811e09a0d3258abee2d98e0c703eee49464ce0069590846f31d40", size = 37662964 }, - { url = "https://files.pythonhosted.org/packages/25/e1/3df8f83cb15f3500478c889be8fb18700813b95e9e087328230b98d547ff/scipy-1.15.3-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:ed7284b21a7a0c8f1b6e5977ac05396c0d008b89e05498c8b7e8f4a1423bba0e", size = 37238749 }, - { url = "https://files.pythonhosted.org/packages/93/3e/b3257cf446f2a3533ed7809757039016b74cd6f38271de91682aa844cfc5/scipy-1.15.3-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:5380741e53df2c566f4d234b100a484b420af85deb39ea35a1cc1be84ff53a5c", size = 40022383 }, - { url = "https://files.pythonhosted.org/packages/d1/84/55bc4881973d3f79b479a5a2e2df61c8c9a04fcb986a213ac9c02cfb659b/scipy-1.15.3-cp310-cp310-win_amd64.whl", hash = "sha256:9d61e97b186a57350f6d6fd72640f9e99d5a4a2b8fbf4b9ee9a841eab327dc13", size = 41259201 }, - { url = "https://files.pythonhosted.org/packages/96/ab/5cc9f80f28f6a7dff646c5756e559823614a42b1939d86dd0ed550470210/scipy-1.15.3-cp311-cp311-macosx_10_13_x86_64.whl", hash = "sha256:993439ce220d25e3696d1b23b233dd010169b62f6456488567e830654ee37a6b", size = 38714255 }, - { url = "https://files.pythonhosted.org/packages/4a/4a/66ba30abe5ad1a3ad15bfb0b59d22174012e8056ff448cb1644deccbfed2/scipy-1.15.3-cp311-cp311-macosx_12_0_arm64.whl", hash = "sha256:34716e281f181a02341ddeaad584205bd2fd3c242063bd3423d61ac259ca7eba", size = 30111035 }, - { url = "https://files.pythonhosted.org/packages/4b/fa/a7e5b95afd80d24313307f03624acc65801846fa75599034f8ceb9e2cbf6/scipy-1.15.3-cp311-cp311-macosx_14_0_arm64.whl", hash = "sha256:3b0334816afb8b91dab859281b1b9786934392aa3d527cd847e41bb6f45bee65", size = 22384499 }, - { url = "https://files.pythonhosted.org/packages/17/99/f3aaddccf3588bb4aea70ba35328c204cadd89517a1612ecfda5b2dd9d7a/scipy-1.15.3-cp311-cp311-macosx_14_0_x86_64.whl", hash = "sha256:6db907c7368e3092e24919b5e31c76998b0ce1684d51a90943cb0ed1b4ffd6c1", size = 25152602 }, - { url = "https://files.pythonhosted.org/packages/56/c5/1032cdb565f146109212153339f9cb8b993701e9fe56b1c97699eee12586/scipy-1.15.3-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:721d6b4ef5dc82ca8968c25b111e307083d7ca9091bc38163fb89243e85e3889", size = 35503415 }, - { url = "https://files.pythonhosted.org/packages/bd/37/89f19c8c05505d0601ed5650156e50eb881ae3918786c8fd7262b4ee66d3/scipy-1.15.3-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:39cb9c62e471b1bb3750066ecc3a3f3052b37751c7c3dfd0fd7e48900ed52982", size = 37652622 }, - { url = "https://files.pythonhosted.org/packages/7e/31/be59513aa9695519b18e1851bb9e487de66f2d31f835201f1b42f5d4d475/scipy-1.15.3-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:795c46999bae845966368a3c013e0e00947932d68e235702b5c3f6ea799aa8c9", size = 37244796 }, - { url = "https://files.pythonhosted.org/packages/10/c0/4f5f3eeccc235632aab79b27a74a9130c6c35df358129f7ac8b29f562ac7/scipy-1.15.3-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:18aaacb735ab38b38db42cb01f6b92a2d0d4b6aabefeb07f02849e47f8fb3594", size = 40047684 }, - { url = "https://files.pythonhosted.org/packages/ab/a7/0ddaf514ce8a8714f6ed243a2b391b41dbb65251affe21ee3077ec45ea9a/scipy-1.15.3-cp311-cp311-win_amd64.whl", hash = "sha256:ae48a786a28412d744c62fd7816a4118ef97e5be0bee968ce8f0a2fba7acf3bb", size = 41246504 }, - { url = "https://files.pythonhosted.org/packages/37/4b/683aa044c4162e10ed7a7ea30527f2cbd92e6999c10a8ed8edb253836e9c/scipy-1.15.3-cp312-cp312-macosx_10_13_x86_64.whl", hash = "sha256:6ac6310fdbfb7aa6612408bd2f07295bcbd3fda00d2d702178434751fe48e019", size = 38766735 }, - { url = "https://files.pythonhosted.org/packages/7b/7e/f30be3d03de07f25dc0ec926d1681fed5c732d759ac8f51079708c79e680/scipy-1.15.3-cp312-cp312-macosx_12_0_arm64.whl", hash = "sha256:185cd3d6d05ca4b44a8f1595af87f9c372bb6acf9c808e99aa3e9aa03bd98cf6", size = 30173284 }, - { url = "https://files.pythonhosted.org/packages/07/9c/0ddb0d0abdabe0d181c1793db51f02cd59e4901da6f9f7848e1f96759f0d/scipy-1.15.3-cp312-cp312-macosx_14_0_arm64.whl", hash = "sha256:05dc6abcd105e1a29f95eada46d4a3f251743cfd7d3ae8ddb4088047f24ea477", size = 22446958 }, - { url = "https://files.pythonhosted.org/packages/af/43/0bce905a965f36c58ff80d8bea33f1f9351b05fad4beaad4eae34699b7a1/scipy-1.15.3-cp312-cp312-macosx_14_0_x86_64.whl", hash = "sha256:06efcba926324df1696931a57a176c80848ccd67ce6ad020c810736bfd58eb1c", size = 25242454 }, - { url = "https://files.pythonhosted.org/packages/56/30/a6f08f84ee5b7b28b4c597aca4cbe545535c39fe911845a96414700b64ba/scipy-1.15.3-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c05045d8b9bfd807ee1b9f38761993297b10b245f012b11b13b91ba8945f7e45", size = 35210199 }, - { url = "https://files.pythonhosted.org/packages/0b/1f/03f52c282437a168ee2c7c14a1a0d0781a9a4a8962d84ac05c06b4c5b555/scipy-1.15.3-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:271e3713e645149ea5ea3e97b57fdab61ce61333f97cfae392c28ba786f9bb49", size = 37309455 }, - { url = "https://files.pythonhosted.org/packages/89/b1/fbb53137f42c4bf630b1ffdfc2151a62d1d1b903b249f030d2b1c0280af8/scipy-1.15.3-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:6cfd56fc1a8e53f6e89ba3a7a7251f7396412d655bca2aa5611c8ec9a6784a1e", size = 36885140 }, - { url = "https://files.pythonhosted.org/packages/2e/2e/025e39e339f5090df1ff266d021892694dbb7e63568edcfe43f892fa381d/scipy-1.15.3-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:0ff17c0bb1cb32952c09217d8d1eed9b53d1463e5f1dd6052c7857f83127d539", size = 39710549 }, - { url = "https://files.pythonhosted.org/packages/e6/eb/3bf6ea8ab7f1503dca3a10df2e4b9c3f6b3316df07f6c0ded94b281c7101/scipy-1.15.3-cp312-cp312-win_amd64.whl", hash = "sha256:52092bc0472cfd17df49ff17e70624345efece4e1a12b23783a1ac59a1b728ed", size = 40966184 }, - { url = "https://files.pythonhosted.org/packages/73/18/ec27848c9baae6e0d6573eda6e01a602e5649ee72c27c3a8aad673ebecfd/scipy-1.15.3-cp313-cp313-macosx_10_13_x86_64.whl", hash = "sha256:2c620736bcc334782e24d173c0fdbb7590a0a436d2fdf39310a8902505008759", size = 38728256 }, - { url = "https://files.pythonhosted.org/packages/74/cd/1aef2184948728b4b6e21267d53b3339762c285a46a274ebb7863c9e4742/scipy-1.15.3-cp313-cp313-macosx_12_0_arm64.whl", hash = "sha256:7e11270a000969409d37ed399585ee530b9ef6aa99d50c019de4cb01e8e54e62", size = 30109540 }, - { url = "https://files.pythonhosted.org/packages/5b/d8/59e452c0a255ec352bd0a833537a3bc1bfb679944c4938ab375b0a6b3a3e/scipy-1.15.3-cp313-cp313-macosx_14_0_arm64.whl", hash = "sha256:8c9ed3ba2c8a2ce098163a9bdb26f891746d02136995df25227a20e71c396ebb", size = 22383115 }, - { url = "https://files.pythonhosted.org/packages/08/f5/456f56bbbfccf696263b47095291040655e3cbaf05d063bdc7c7517f32ac/scipy-1.15.3-cp313-cp313-macosx_14_0_x86_64.whl", hash = "sha256:0bdd905264c0c9cfa74a4772cdb2070171790381a5c4d312c973382fc6eaf730", size = 25163884 }, - { url = "https://files.pythonhosted.org/packages/a2/66/a9618b6a435a0f0c0b8a6d0a2efb32d4ec5a85f023c2b79d39512040355b/scipy-1.15.3-cp313-cp313-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:79167bba085c31f38603e11a267d862957cbb3ce018d8b38f79ac043bc92d825", size = 35174018 }, - { url = "https://files.pythonhosted.org/packages/b5/09/c5b6734a50ad4882432b6bb7c02baf757f5b2f256041da5df242e2d7e6b6/scipy-1.15.3-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c9deabd6d547aee2c9a81dee6cc96c6d7e9a9b1953f74850c179f91fdc729cb7", size = 37269716 }, - { url = "https://files.pythonhosted.org/packages/77/0a/eac00ff741f23bcabd352731ed9b8995a0a60ef57f5fd788d611d43d69a1/scipy-1.15.3-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:dde4fc32993071ac0c7dd2d82569e544f0bdaff66269cb475e0f369adad13f11", size = 36872342 }, - { url = "https://files.pythonhosted.org/packages/fe/54/4379be86dd74b6ad81551689107360d9a3e18f24d20767a2d5b9253a3f0a/scipy-1.15.3-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:f77f853d584e72e874d87357ad70f44b437331507d1c311457bed8ed2b956126", size = 39670869 }, - { url = "https://files.pythonhosted.org/packages/87/2e/892ad2862ba54f084ffe8cc4a22667eaf9c2bcec6d2bff1d15713c6c0703/scipy-1.15.3-cp313-cp313-win_amd64.whl", hash = "sha256:b90ab29d0c37ec9bf55424c064312930ca5f4bde15ee8619ee44e69319aab163", size = 40988851 }, - { url = "https://files.pythonhosted.org/packages/1b/e9/7a879c137f7e55b30d75d90ce3eb468197646bc7b443ac036ae3fe109055/scipy-1.15.3-cp313-cp313t-macosx_10_13_x86_64.whl", hash = "sha256:3ac07623267feb3ae308487c260ac684b32ea35fd81e12845039952f558047b8", size = 38863011 }, - { url = "https://files.pythonhosted.org/packages/51/d1/226a806bbd69f62ce5ef5f3ffadc35286e9fbc802f606a07eb83bf2359de/scipy-1.15.3-cp313-cp313t-macosx_12_0_arm64.whl", hash = "sha256:6487aa99c2a3d509a5227d9a5e889ff05830a06b2ce08ec30df6d79db5fcd5c5", size = 30266407 }, - { url = "https://files.pythonhosted.org/packages/e5/9b/f32d1d6093ab9eeabbd839b0f7619c62e46cc4b7b6dbf05b6e615bbd4400/scipy-1.15.3-cp313-cp313t-macosx_14_0_arm64.whl", hash = "sha256:50f9e62461c95d933d5c5ef4a1f2ebf9a2b4e83b0db374cb3f1de104d935922e", size = 22540030 }, - { url = "https://files.pythonhosted.org/packages/e7/29/c278f699b095c1a884f29fda126340fcc201461ee8bfea5c8bdb1c7c958b/scipy-1.15.3-cp313-cp313t-macosx_14_0_x86_64.whl", hash = "sha256:14ed70039d182f411ffc74789a16df3835e05dc469b898233a245cdfd7f162cb", size = 25218709 }, - { url = "https://files.pythonhosted.org/packages/24/18/9e5374b617aba742a990581373cd6b68a2945d65cc588482749ef2e64467/scipy-1.15.3-cp313-cp313t-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0a769105537aa07a69468a0eefcd121be52006db61cdd8cac8a0e68980bbb723", size = 34809045 }, - { url = "https://files.pythonhosted.org/packages/e1/fe/9c4361e7ba2927074360856db6135ef4904d505e9b3afbbcb073c4008328/scipy-1.15.3-cp313-cp313t-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9db984639887e3dffb3928d118145ffe40eff2fa40cb241a306ec57c219ebbbb", size = 36703062 }, - { url = "https://files.pythonhosted.org/packages/b7/8e/038ccfe29d272b30086b25a4960f757f97122cb2ec42e62b460d02fe98e9/scipy-1.15.3-cp313-cp313t-musllinux_1_2_aarch64.whl", hash = "sha256:40e54d5c7e7ebf1aa596c374c49fa3135f04648a0caabcb66c52884b943f02b4", size = 36393132 }, - { url = "https://files.pythonhosted.org/packages/10/7e/5c12285452970be5bdbe8352c619250b97ebf7917d7a9a9e96b8a8140f17/scipy-1.15.3-cp313-cp313t-musllinux_1_2_x86_64.whl", hash = "sha256:5e721fed53187e71d0ccf382b6bf977644c533e506c4d33c3fb24de89f5c3ed5", size = 38979503 }, - { url = "https://files.pythonhosted.org/packages/81/06/0a5e5349474e1cbc5757975b21bd4fad0e72ebf138c5592f191646154e06/scipy-1.15.3-cp313-cp313t-win_amd64.whl", hash = "sha256:76ad1fb5f8752eabf0fa02e4cc0336b4e8f021e2d5f061ed37d6d264db35e3ca", size = 40308097 }, -] - -[[package]] -name = "scipy" -version = "1.16.3" -source = { registry = "https://pypi.org/simple" } -resolution-markers = [ - "python_full_version >= '3.14' and sys_platform == 'linux'", - "python_full_version == '3.13.*' and sys_platform == 'linux'", - "python_full_version >= '3.14' and sys_platform != 'linux'", - "python_full_version == '3.13.*' and sys_platform != 'linux'", - "python_full_version == '3.12.*' and sys_platform == 'linux'", - "python_full_version == '3.12.*' and sys_platform != 'linux'", - "python_full_version == '3.11.*' and sys_platform == 'linux'", - "python_full_version == '3.11.*' and sys_platform != 'linux'", -] -dependencies = [ - { name = "numpy", marker = "python_full_version >= '3.11'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/0a/ca/d8ace4f98322d01abcd52d381134344bf7b431eba7ed8b42bdea5a3c2ac9/scipy-1.16.3.tar.gz", hash = "sha256:01e87659402762f43bd2fee13370553a17ada367d42e7487800bf2916535aecb", size = 30597883 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/9b/5f/6f37d7439de1455ce9c5a556b8d1db0979f03a796c030bafdf08d35b7bf9/scipy-1.16.3-cp311-cp311-macosx_10_14_x86_64.whl", hash = "sha256:40be6cf99e68b6c4321e9f8782e7d5ff8265af28ef2cd56e9c9b2638fa08ad97", size = 36630881 }, - { url = "https://files.pythonhosted.org/packages/7c/89/d70e9f628749b7e4db2aa4cd89735502ff3f08f7b9b27d2e799485987cd9/scipy-1.16.3-cp311-cp311-macosx_12_0_arm64.whl", hash = "sha256:8be1ca9170fcb6223cc7c27f4305d680ded114a1567c0bd2bfcbf947d1b17511", size = 28941012 }, - { url = "https://files.pythonhosted.org/packages/a8/a8/0e7a9a6872a923505dbdf6bb93451edcac120363131c19013044a1e7cb0c/scipy-1.16.3-cp311-cp311-macosx_14_0_arm64.whl", hash = "sha256:bea0a62734d20d67608660f69dcda23e7f90fb4ca20974ab80b6ed40df87a005", size = 20931935 }, - { url = "https://files.pythonhosted.org/packages/bd/c7/020fb72bd79ad798e4dbe53938543ecb96b3a9ac3fe274b7189e23e27353/scipy-1.16.3-cp311-cp311-macosx_14_0_x86_64.whl", hash = "sha256:2a207a6ce9c24f1951241f4693ede2d393f59c07abc159b2cb2be980820e01fb", size = 23534466 }, - { url = "https://files.pythonhosted.org/packages/be/a0/668c4609ce6dbf2f948e167836ccaf897f95fb63fa231c87da7558a374cd/scipy-1.16.3-cp311-cp311-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:532fb5ad6a87e9e9cd9c959b106b73145a03f04c7d57ea3e6f6bb60b86ab0876", size = 33593618 }, - { url = "https://files.pythonhosted.org/packages/ca/6e/8942461cf2636cdae083e3eb72622a7fbbfa5cf559c7d13ab250a5dbdc01/scipy-1.16.3-cp311-cp311-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:0151a0749efeaaab78711c78422d413c583b8cdd2011a3c1d6c794938ee9fdb2", size = 35899798 }, - { url = "https://files.pythonhosted.org/packages/79/e8/d0f33590364cdbd67f28ce79368b373889faa4ee959588beddf6daef9abe/scipy-1.16.3-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:b7180967113560cca57418a7bc719e30366b47959dd845a93206fbed693c867e", size = 36226154 }, - { url = "https://files.pythonhosted.org/packages/39/c1/1903de608c0c924a1749c590064e65810f8046e437aba6be365abc4f7557/scipy-1.16.3-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:deb3841c925eeddb6afc1e4e4a45e418d19ec7b87c5df177695224078e8ec733", size = 38878540 }, - { url = "https://files.pythonhosted.org/packages/f1/d0/22ec7036ba0b0a35bccb7f25ab407382ed34af0b111475eb301c16f8a2e5/scipy-1.16.3-cp311-cp311-win_amd64.whl", hash = "sha256:53c3844d527213631e886621df5695d35e4f6a75f620dca412bcd292f6b87d78", size = 38722107 }, - { url = "https://files.pythonhosted.org/packages/7b/60/8a00e5a524bb3bf8898db1650d350f50e6cffb9d7a491c561dc9826c7515/scipy-1.16.3-cp311-cp311-win_arm64.whl", hash = "sha256:9452781bd879b14b6f055b26643703551320aa8d79ae064a71df55c00286a184", size = 25506272 }, - { url = "https://files.pythonhosted.org/packages/40/41/5bf55c3f386b1643812f3a5674edf74b26184378ef0f3e7c7a09a7e2ca7f/scipy-1.16.3-cp312-cp312-macosx_10_14_x86_64.whl", hash = "sha256:81fc5827606858cf71446a5e98715ba0e11f0dbc83d71c7409d05486592a45d6", size = 36659043 }, - { url = "https://files.pythonhosted.org/packages/1e/0f/65582071948cfc45d43e9870bf7ca5f0e0684e165d7c9ef4e50d783073eb/scipy-1.16.3-cp312-cp312-macosx_12_0_arm64.whl", hash = "sha256:c97176013d404c7346bf57874eaac5187d969293bf40497140b0a2b2b7482e07", size = 28898986 }, - { url = "https://files.pythonhosted.org/packages/96/5e/36bf3f0ac298187d1ceadde9051177d6a4fe4d507e8f59067dc9dd39e650/scipy-1.16.3-cp312-cp312-macosx_14_0_arm64.whl", hash = "sha256:2b71d93c8a9936046866acebc915e2af2e292b883ed6e2cbe5c34beb094b82d9", size = 20889814 }, - { url = "https://files.pythonhosted.org/packages/80/35/178d9d0c35394d5d5211bbff7ac4f2986c5488b59506fef9e1de13ea28d3/scipy-1.16.3-cp312-cp312-macosx_14_0_x86_64.whl", hash = "sha256:3d4a07a8e785d80289dfe66b7c27d8634a773020742ec7187b85ccc4b0e7b686", size = 23565795 }, - { url = "https://files.pythonhosted.org/packages/fa/46/d1146ff536d034d02f83c8afc3c4bab2eddb634624d6529a8512f3afc9da/scipy-1.16.3-cp312-cp312-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:0553371015692a898e1aa858fed67a3576c34edefa6b7ebdb4e9dde49ce5c203", size = 33349476 }, - { url = "https://files.pythonhosted.org/packages/79/2e/415119c9ab3e62249e18c2b082c07aff907a273741b3f8160414b0e9193c/scipy-1.16.3-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:72d1717fd3b5e6ec747327ce9bda32d5463f472c9dce9f54499e81fbd50245a1", size = 35676692 }, - { url = "https://files.pythonhosted.org/packages/27/82/df26e44da78bf8d2aeaf7566082260cfa15955a5a6e96e6a29935b64132f/scipy-1.16.3-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:1fb2472e72e24d1530debe6ae078db70fb1605350c88a3d14bc401d6306dbffe", size = 36019345 }, - { url = "https://files.pythonhosted.org/packages/82/31/006cbb4b648ba379a95c87262c2855cd0d09453e500937f78b30f02fa1cd/scipy-1.16.3-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:c5192722cffe15f9329a3948c4b1db789fbb1f05c97899187dcf009b283aea70", size = 38678975 }, - { url = "https://files.pythonhosted.org/packages/c2/7f/acbd28c97e990b421af7d6d6cd416358c9c293fc958b8529e0bd5d2a2a19/scipy-1.16.3-cp312-cp312-win_amd64.whl", hash = "sha256:56edc65510d1331dae01ef9b658d428e33ed48b4f77b1d51caf479a0253f96dc", size = 38555926 }, - { url = "https://files.pythonhosted.org/packages/ce/69/c5c7807fd007dad4f48e0a5f2153038dc96e8725d3345b9ee31b2b7bed46/scipy-1.16.3-cp312-cp312-win_arm64.whl", hash = "sha256:a8a26c78ef223d3e30920ef759e25625a0ecdd0d60e5a8818b7513c3e5384cf2", size = 25463014 }, - { url = "https://files.pythonhosted.org/packages/72/f1/57e8327ab1508272029e27eeef34f2302ffc156b69e7e233e906c2a5c379/scipy-1.16.3-cp313-cp313-macosx_10_14_x86_64.whl", hash = "sha256:d2ec56337675e61b312179a1ad124f5f570c00f920cc75e1000025451b88241c", size = 36617856 }, - { url = "https://files.pythonhosted.org/packages/44/13/7e63cfba8a7452eb756306aa2fd9b37a29a323b672b964b4fdeded9a3f21/scipy-1.16.3-cp313-cp313-macosx_12_0_arm64.whl", hash = "sha256:16b8bc35a4cc24db80a0ec836a9286d0e31b2503cb2fd7ff7fb0e0374a97081d", size = 28874306 }, - { url = "https://files.pythonhosted.org/packages/15/65/3a9400efd0228a176e6ec3454b1fa998fbbb5a8defa1672c3f65706987db/scipy-1.16.3-cp313-cp313-macosx_14_0_arm64.whl", hash = "sha256:5803c5fadd29de0cf27fa08ccbfe7a9e5d741bf63e4ab1085437266f12460ff9", size = 20865371 }, - { url = "https://files.pythonhosted.org/packages/33/d7/eda09adf009a9fb81827194d4dd02d2e4bc752cef16737cc4ef065234031/scipy-1.16.3-cp313-cp313-macosx_14_0_x86_64.whl", hash = "sha256:b81c27fc41954319a943d43b20e07c40bdcd3ff7cf013f4fb86286faefe546c4", size = 23524877 }, - { url = "https://files.pythonhosted.org/packages/7d/6b/3f911e1ebc364cb81320223a3422aab7d26c9c7973109a9cd0f27c64c6c0/scipy-1.16.3-cp313-cp313-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:0c3b4dd3d9b08dbce0f3440032c52e9e2ab9f96ade2d3943313dfe51a7056959", size = 33342103 }, - { url = "https://files.pythonhosted.org/packages/21/f6/4bfb5695d8941e5c570a04d9fcd0d36bce7511b7d78e6e75c8f9791f82d0/scipy-1.16.3-cp313-cp313-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:7dc1360c06535ea6116a2220f760ae572db9f661aba2d88074fe30ec2aa1ff88", size = 35697297 }, - { url = "https://files.pythonhosted.org/packages/04/e1/6496dadbc80d8d896ff72511ecfe2316b50313bfc3ebf07a3f580f08bd8c/scipy-1.16.3-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:663b8d66a8748051c3ee9c96465fb417509315b99c71550fda2591d7dd634234", size = 36021756 }, - { url = "https://files.pythonhosted.org/packages/fe/bd/a8c7799e0136b987bda3e1b23d155bcb31aec68a4a472554df5f0937eef7/scipy-1.16.3-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:eab43fae33a0c39006a88096cd7b4f4ef545ea0447d250d5ac18202d40b6611d", size = 38696566 }, - { url = "https://files.pythonhosted.org/packages/cd/01/1204382461fcbfeb05b6161b594f4007e78b6eba9b375382f79153172b4d/scipy-1.16.3-cp313-cp313-win_amd64.whl", hash = "sha256:062246acacbe9f8210de8e751b16fc37458213f124bef161a5a02c7a39284304", size = 38529877 }, - { url = "https://files.pythonhosted.org/packages/7f/14/9d9fbcaa1260a94f4bb5b64ba9213ceb5d03cd88841fe9fd1ffd47a45b73/scipy-1.16.3-cp313-cp313-win_arm64.whl", hash = "sha256:50a3dbf286dbc7d84f176f9a1574c705f277cb6565069f88f60db9eafdbe3ee2", size = 25455366 }, - { url = "https://files.pythonhosted.org/packages/e2/a3/9ec205bd49f42d45d77f1730dbad9ccf146244c1647605cf834b3a8c4f36/scipy-1.16.3-cp313-cp313t-macosx_10_14_x86_64.whl", hash = "sha256:fb4b29f4cf8cc5a8d628bc8d8e26d12d7278cd1f219f22698a378c3d67db5e4b", size = 37027931 }, - { url = "https://files.pythonhosted.org/packages/25/06/ca9fd1f3a4589cbd825b1447e5db3a8ebb969c1eaf22c8579bd286f51b6d/scipy-1.16.3-cp313-cp313t-macosx_12_0_arm64.whl", hash = "sha256:8d09d72dc92742988b0e7750bddb8060b0c7079606c0d24a8cc8e9c9c11f9079", size = 29400081 }, - { url = "https://files.pythonhosted.org/packages/6a/56/933e68210d92657d93fb0e381683bc0e53a965048d7358ff5fbf9e6a1b17/scipy-1.16.3-cp313-cp313t-macosx_14_0_arm64.whl", hash = "sha256:03192a35e661470197556de24e7cb1330d84b35b94ead65c46ad6f16f6b28f2a", size = 21391244 }, - { url = "https://files.pythonhosted.org/packages/a8/7e/779845db03dc1418e215726329674b40576879b91814568757ff0014ad65/scipy-1.16.3-cp313-cp313t-macosx_14_0_x86_64.whl", hash = "sha256:57d01cb6f85e34f0946b33caa66e892aae072b64b034183f3d87c4025802a119", size = 23929753 }, - { url = "https://files.pythonhosted.org/packages/4c/4b/f756cf8161d5365dcdef9e5f460ab226c068211030a175d2fc7f3f41ca64/scipy-1.16.3-cp313-cp313t-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:96491a6a54e995f00a28a3c3badfff58fd093bf26cd5fb34a2188c8c756a3a2c", size = 33496912 }, - { url = "https://files.pythonhosted.org/packages/09/b5/222b1e49a58668f23839ca1542a6322bb095ab8d6590d4f71723869a6c2c/scipy-1.16.3-cp313-cp313t-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:cd13e354df9938598af2be05822c323e97132d5e6306b83a3b4ee6724c6e522e", size = 35802371 }, - { url = "https://files.pythonhosted.org/packages/c1/8d/5964ef68bb31829bde27611f8c9deeac13764589fe74a75390242b64ca44/scipy-1.16.3-cp313-cp313t-musllinux_1_2_aarch64.whl", hash = "sha256:63d3cdacb8a824a295191a723ee5e4ea7768ca5ca5f2838532d9f2e2b3ce2135", size = 36190477 }, - { url = "https://files.pythonhosted.org/packages/ab/f2/b31d75cb9b5fa4dd39a0a931ee9b33e7f6f36f23be5ef560bf72e0f92f32/scipy-1.16.3-cp313-cp313t-musllinux_1_2_x86_64.whl", hash = "sha256:e7efa2681ea410b10dde31a52b18b0154d66f2485328830e45fdf183af5aefc6", size = 38796678 }, - { url = "https://files.pythonhosted.org/packages/b4/1e/b3723d8ff64ab548c38d87055483714fefe6ee20e0189b62352b5e015bb1/scipy-1.16.3-cp313-cp313t-win_amd64.whl", hash = "sha256:2d1ae2cf0c350e7705168ff2429962a89ad90c2d49d1dd300686d8b2a5af22fc", size = 38640178 }, - { url = "https://files.pythonhosted.org/packages/8e/f3/d854ff38789aca9b0cc23008d607ced9de4f7ab14fa1ca4329f86b3758ca/scipy-1.16.3-cp313-cp313t-win_arm64.whl", hash = "sha256:0c623a54f7b79dd88ef56da19bc2873afec9673a48f3b85b18e4d402bdd29a5a", size = 25803246 }, - { url = "https://files.pythonhosted.org/packages/99/f6/99b10fd70f2d864c1e29a28bbcaa0c6340f9d8518396542d9ea3b4aaae15/scipy-1.16.3-cp314-cp314-macosx_10_14_x86_64.whl", hash = "sha256:875555ce62743e1d54f06cdf22c1e0bc47b91130ac40fe5d783b6dfa114beeb6", size = 36606469 }, - { url = "https://files.pythonhosted.org/packages/4d/74/043b54f2319f48ea940dd025779fa28ee360e6b95acb7cd188fad4391c6b/scipy-1.16.3-cp314-cp314-macosx_12_0_arm64.whl", hash = "sha256:bb61878c18a470021fb515a843dc7a76961a8daceaaaa8bad1332f1bf4b54657", size = 28872043 }, - { url = "https://files.pythonhosted.org/packages/4d/e1/24b7e50cc1c4ee6ffbcb1f27fe9f4c8b40e7911675f6d2d20955f41c6348/scipy-1.16.3-cp314-cp314-macosx_14_0_arm64.whl", hash = "sha256:f2622206f5559784fa5c4b53a950c3c7c1cf3e84ca1b9c4b6c03f062f289ca26", size = 20862952 }, - { url = "https://files.pythonhosted.org/packages/dd/3a/3e8c01a4d742b730df368e063787c6808597ccb38636ed821d10b39ca51b/scipy-1.16.3-cp314-cp314-macosx_14_0_x86_64.whl", hash = "sha256:7f68154688c515cdb541a31ef8eb66d8cd1050605be9dcd74199cbd22ac739bc", size = 23508512 }, - { url = "https://files.pythonhosted.org/packages/1f/60/c45a12b98ad591536bfe5330cb3cfe1850d7570259303563b1721564d458/scipy-1.16.3-cp314-cp314-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:8b3c820ddb80029fe9f43d61b81d8b488d3ef8ca010d15122b152db77dc94c22", size = 33413639 }, - { url = "https://files.pythonhosted.org/packages/71/bc/35957d88645476307e4839712642896689df442f3e53b0fa016ecf8a3357/scipy-1.16.3-cp314-cp314-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:d3837938ae715fc0fe3c39c0202de3a8853aff22ca66781ddc2ade7554b7e2cc", size = 35704729 }, - { url = "https://files.pythonhosted.org/packages/3b/15/89105e659041b1ca11c386e9995aefacd513a78493656e57789f9d9eab61/scipy-1.16.3-cp314-cp314-musllinux_1_2_aarch64.whl", hash = "sha256:aadd23f98f9cb069b3bd64ddc900c4d277778242e961751f77a8cb5c4b946fb0", size = 36086251 }, - { url = "https://files.pythonhosted.org/packages/1a/87/c0ea673ac9c6cc50b3da2196d860273bc7389aa69b64efa8493bdd25b093/scipy-1.16.3-cp314-cp314-musllinux_1_2_x86_64.whl", hash = "sha256:b7c5f1bda1354d6a19bc6af73a649f8285ca63ac6b52e64e658a5a11d4d69800", size = 38716681 }, - { url = "https://files.pythonhosted.org/packages/91/06/837893227b043fb9b0d13e4bd7586982d8136cb249ffb3492930dab905b8/scipy-1.16.3-cp314-cp314-win_amd64.whl", hash = "sha256:e5d42a9472e7579e473879a1990327830493a7047506d58d73fc429b84c1d49d", size = 39358423 }, - { url = "https://files.pythonhosted.org/packages/95/03/28bce0355e4d34a7c034727505a02d19548549e190bedd13a721e35380b7/scipy-1.16.3-cp314-cp314-win_arm64.whl", hash = "sha256:6020470b9d00245926f2d5bb93b119ca0340f0d564eb6fbaad843eaebf9d690f", size = 26135027 }, - { url = "https://files.pythonhosted.org/packages/b2/6f/69f1e2b682efe9de8fe9f91040f0cd32f13cfccba690512ba4c582b0bc29/scipy-1.16.3-cp314-cp314t-macosx_10_14_x86_64.whl", hash = "sha256:e1d27cbcb4602680a49d787d90664fa4974063ac9d4134813332a8c53dbe667c", size = 37028379 }, - { url = "https://files.pythonhosted.org/packages/7c/2d/e826f31624a5ebbab1cd93d30fd74349914753076ed0593e1d56a98c4fb4/scipy-1.16.3-cp314-cp314t-macosx_12_0_arm64.whl", hash = "sha256:9b9c9c07b6d56a35777a1b4cc8966118fb16cfd8daf6743867d17d36cfad2d40", size = 29400052 }, - { url = "https://files.pythonhosted.org/packages/69/27/d24feb80155f41fd1f156bf144e7e049b4e2b9dd06261a242905e3bc7a03/scipy-1.16.3-cp314-cp314t-macosx_14_0_arm64.whl", hash = "sha256:3a4c460301fb2cffb7f88528f30b3127742cff583603aa7dc964a52c463b385d", size = 21391183 }, - { url = "https://files.pythonhosted.org/packages/f8/d3/1b229e433074c5738a24277eca520a2319aac7465eea7310ea6ae0e98ae2/scipy-1.16.3-cp314-cp314t-macosx_14_0_x86_64.whl", hash = "sha256:f667a4542cc8917af1db06366d3f78a5c8e83badd56409f94d1eac8d8d9133fa", size = 23930174 }, - { url = "https://files.pythonhosted.org/packages/16/9d/d9e148b0ec680c0f042581a2be79a28a7ab66c0c4946697f9e7553ead337/scipy-1.16.3-cp314-cp314t-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:f379b54b77a597aa7ee5e697df0d66903e41b9c85a6dd7946159e356319158e8", size = 33497852 }, - { url = "https://files.pythonhosted.org/packages/2f/22/4e5f7561e4f98b7bea63cf3fd7934bff1e3182e9f1626b089a679914d5c8/scipy-1.16.3-cp314-cp314t-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:4aff59800a3b7f786b70bfd6ab551001cb553244988d7d6b8299cb1ea653b353", size = 35798595 }, - { url = "https://files.pythonhosted.org/packages/83/42/6644d714c179429fc7196857866f219fef25238319b650bb32dde7bf7a48/scipy-1.16.3-cp314-cp314t-musllinux_1_2_aarch64.whl", hash = "sha256:da7763f55885045036fabcebd80144b757d3db06ab0861415d1c3b7c69042146", size = 36186269 }, - { url = "https://files.pythonhosted.org/packages/ac/70/64b4d7ca92f9cf2e6fc6aaa2eecf80bb9b6b985043a9583f32f8177ea122/scipy-1.16.3-cp314-cp314t-musllinux_1_2_x86_64.whl", hash = "sha256:ffa6eea95283b2b8079b821dc11f50a17d0571c92b43e2b5b12764dc5f9b285d", size = 38802779 }, - { url = "https://files.pythonhosted.org/packages/61/82/8d0e39f62764cce5ffd5284131e109f07cf8955aef9ab8ed4e3aa5e30539/scipy-1.16.3-cp314-cp314t-win_amd64.whl", hash = "sha256:d9f48cafc7ce94cf9b15c6bffdc443a81a27bf7075cf2dcd5c8b40f85d10c4e7", size = 39471128 }, - { url = "https://files.pythonhosted.org/packages/64/47/a494741db7280eae6dc033510c319e34d42dd41b7ac0c7ead39354d1a2b5/scipy-1.16.3-cp314-cp314t-win_arm64.whl", hash = "sha256:21d9d6b197227a12dcbf9633320a4e34c6b0e51c57268df255a0942983bac562", size = 26464127 }, -] - -[[package]] -name = "sentencepiece" -version = "0.2.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/15/15/2e7a025fc62d764b151ae6d0f2a92f8081755ebe8d4a64099accc6f77ba6/sentencepiece-0.2.1.tar.gz", hash = "sha256:8138cec27c2f2282f4a34d9a016e3374cd40e5c6e9cb335063db66a0a3b71fad", size = 3228515 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/af/31/5b7cccb307b485db1a2372d6d2980b0a65d067f8be5ca943a103b4acd5b3/sentencepiece-0.2.1-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:e10fa50bdbaa5e2445dbd387979980d391760faf0ec99a09bd7780ff37eaec44", size = 1942557 }, - { url = "https://files.pythonhosted.org/packages/1f/41/0ac923a8e685ad290c5afc8ae55c5844977b8d75076fcc04302b9a324274/sentencepiece-0.2.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:2f27ae6deea72efdb6f361750c92f6c21fd0ad087445082770cc34015213c526", size = 1325384 }, - { url = "https://files.pythonhosted.org/packages/fc/ef/3751555d67daf9003384978f169d31c775cb5c7baf28633caaf1eb2b2b4d/sentencepiece-0.2.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:60937c959e6f44159fdd9f56fbdd302501f96114a5ba436829496d5f32d8de3f", size = 1253317 }, - { url = "https://files.pythonhosted.org/packages/46/a5/742c69b7bd144eb32b6e5fd50dbd8abbbc7a95fce2fe16e50156fa400e3b/sentencepiece-0.2.1-cp310-cp310-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:d8b1d91545578852f128650b8cce4ec20f93d39b378ff554ebe66290f2dabb92", size = 1316379 }, - { url = "https://files.pythonhosted.org/packages/c8/89/8deeafbba2871e8fa10f20f17447786f4ac38085925335728d360eaf4cae/sentencepiece-0.2.1-cp310-cp310-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:27e38eee653abc3d387862e67bc5c8b6f428cd604e688b85d29170b7e725c26c", size = 1387926 }, - { url = "https://files.pythonhosted.org/packages/c3/ca/67fe73005f0ab617c6a970b199754e28e524b6873aa7025224fad3cda252/sentencepiece-0.2.1-cp310-cp310-win32.whl", hash = "sha256:251874d720ac7f28024a168501f3c7bb15d1802245f6e66de565f18bbb9b5eaa", size = 999550 }, - { url = "https://files.pythonhosted.org/packages/6d/33/dc5b54042050d2dda4229c3ce1f862541c99966390b6aa20f54d520d2dc2/sentencepiece-0.2.1-cp310-cp310-win_amd64.whl", hash = "sha256:e52144670738b4b477fade6c2a9b6af71a8d0094514c9853ac9f6fc1fcfabae7", size = 1054613 }, - { url = "https://files.pythonhosted.org/packages/fa/19/1ea47f46ff97fe04422b78997da1a37cd632f414aae042d27a9009c5b733/sentencepiece-0.2.1-cp310-cp310-win_arm64.whl", hash = "sha256:9076430ac25dfa7147d9d05751dbc66a04bc1aaac371c07f84952979ea59f0d0", size = 1033884 }, - { url = "https://files.pythonhosted.org/packages/d8/15/46afbab00733d81788b64be430ca1b93011bb9388527958e26cc31832de5/sentencepiece-0.2.1-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:6356d0986b8b8dc351b943150fcd81a1c6e6e4d439772e8584c64230e58ca987", size = 1942560 }, - { url = "https://files.pythonhosted.org/packages/fa/79/7c01b8ef98a0567e9d84a4e7a910f8e7074fcbf398a5cd76f93f4b9316f9/sentencepiece-0.2.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:8f8ba89a3acb3dc1ae90f65ec1894b0b9596fdb98ab003ff38e058f898b39bc7", size = 1325385 }, - { url = "https://files.pythonhosted.org/packages/bb/88/2b41e07bd24f33dcf2f18ec3b74247aa4af3526bad8907b8727ea3caba03/sentencepiece-0.2.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:02593eca45440ef39247cee8c47322a34bdcc1d8ae83ad28ba5a899a2cf8d79a", size = 1253319 }, - { url = "https://files.pythonhosted.org/packages/a0/54/38a1af0c6210a3c6f95aa46d23d6640636d020fba7135cd0d9a84ada05a7/sentencepiece-0.2.1-cp311-cp311-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:0a0d15781a171d188b661ae4bde1d998c303f6bd8621498c50c671bd45a4798e", size = 1316162 }, - { url = "https://files.pythonhosted.org/packages/ef/66/fb191403ade791ad2c3c1e72fe8413e63781b08cfa3aa4c9dfc536d6e795/sentencepiece-0.2.1-cp311-cp311-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:4f5a3e0d9f445ed9d66c0fec47d4b23d12cfc858b407a03c194c1b26c2ac2a63", size = 1387785 }, - { url = "https://files.pythonhosted.org/packages/a9/2d/3bd9b08e70067b2124518b308db6a84a4f8901cc8a4317e2e4288cdd9b4d/sentencepiece-0.2.1-cp311-cp311-win32.whl", hash = "sha256:6d297a1748d429ba8534eebe5535448d78b8acc32d00a29b49acf28102eeb094", size = 999555 }, - { url = "https://files.pythonhosted.org/packages/32/b8/f709977f5fda195ae1ea24f24e7c581163b6f142b1005bc3d0bbfe4d7082/sentencepiece-0.2.1-cp311-cp311-win_amd64.whl", hash = "sha256:82d9ead6591015f009cb1be1cb1c015d5e6f04046dbb8c9588b931e869a29728", size = 1054617 }, - { url = "https://files.pythonhosted.org/packages/7a/40/a1fc23be23067da0f703709797b464e8a30a1c78cc8a687120cd58d4d509/sentencepiece-0.2.1-cp311-cp311-win_arm64.whl", hash = "sha256:39f8651bd10974eafb9834ce30d9bcf5b73e1fc798a7f7d2528f9820ca86e119", size = 1033877 }, - { url = "https://files.pythonhosted.org/packages/4a/be/32ce495aa1d0e0c323dcb1ba87096037358edee539cac5baf8755a6bd396/sentencepiece-0.2.1-cp312-cp312-macosx_10_13_universal2.whl", hash = "sha256:57cae326c8727de58c85977b175af132a7138d84c764635d7e71bbee7e774133", size = 1943152 }, - { url = "https://files.pythonhosted.org/packages/88/7e/ff23008899a58678e98c6ff592bf4d368eee5a71af96d0df6b38a039dd4f/sentencepiece-0.2.1-cp312-cp312-macosx_10_13_x86_64.whl", hash = "sha256:56dd39a3c4d6493db3cdca7e8cc68c6b633f0d4195495cbadfcf5af8a22d05a6", size = 1325651 }, - { url = "https://files.pythonhosted.org/packages/19/84/42eb3ce4796777a1b5d3699dfd4dca85113e68b637f194a6c8d786f16a04/sentencepiece-0.2.1-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:d9381351182ff9888cc80e41c632e7e274b106f450de33d67a9e8f6043da6f76", size = 1253645 }, - { url = "https://files.pythonhosted.org/packages/89/fa/d3d5ebcba3cb9e6d3775a096251860c41a6bc53a1b9461151df83fe93255/sentencepiece-0.2.1-cp312-cp312-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:99f955df238021bf11f0fc37cdb54fd5e5b5f7fd30ecc3d93fb48b6815437167", size = 1316273 }, - { url = "https://files.pythonhosted.org/packages/04/88/14f2f4a2b922d8b39be45bf63d79e6cd3a9b2f248b2fcb98a69b12af12f5/sentencepiece-0.2.1-cp312-cp312-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:0cdfecef430d985f1c2bcbfff3defd1d95dae876fbd0173376012d2d7d24044b", size = 1387881 }, - { url = "https://files.pythonhosted.org/packages/fd/b8/903e5ccb77b4ef140605d5d71b4f9e0ad95d456d6184688073ed11712809/sentencepiece-0.2.1-cp312-cp312-win32.whl", hash = "sha256:a483fd29a34c3e34c39ac5556b0a90942bec253d260235729e50976f5dba1068", size = 999540 }, - { url = "https://files.pythonhosted.org/packages/2d/81/92df5673c067148c2545b1bfe49adfd775bcc3a169a047f5a0e6575ddaca/sentencepiece-0.2.1-cp312-cp312-win_amd64.whl", hash = "sha256:4cdc7c36234fda305e85c32949c5211faaf8dd886096c7cea289ddc12a2d02de", size = 1054671 }, - { url = "https://files.pythonhosted.org/packages/fe/02/c5e3bc518655d714622bec87d83db9cdba1cd0619a4a04e2109751c4f47f/sentencepiece-0.2.1-cp312-cp312-win_arm64.whl", hash = "sha256:daeb5e9e9fcad012324807856113708614d534f596d5008638eb9b40112cd9e4", size = 1033923 }, - { url = "https://files.pythonhosted.org/packages/ba/4a/85fbe1706d4d04a7e826b53f327c4b80f849cf1c7b7c5e31a20a97d8f28b/sentencepiece-0.2.1-cp313-cp313-macosx_10_13_universal2.whl", hash = "sha256:dcd8161eee7b41aae57ded06272905dbd680a0a04b91edd0f64790c796b2f706", size = 1943150 }, - { url = "https://files.pythonhosted.org/packages/c2/83/4cfb393e287509fc2155480b9d184706ef8d9fa8cbf5505d02a5792bf220/sentencepiece-0.2.1-cp313-cp313-macosx_10_13_x86_64.whl", hash = "sha256:c6c8f42949f419ff8c7e9960dbadcfbc982d7b5efc2f6748210d3dd53a7de062", size = 1325651 }, - { url = "https://files.pythonhosted.org/packages/8d/de/5a007fb53b1ab0aafc69d11a5a3dd72a289d5a3e78dcf2c3a3d9b14ffe93/sentencepiece-0.2.1-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:097f3394e99456e9e4efba1737c3749d7e23563dd1588ce71a3d007f25475fff", size = 1253641 }, - { url = "https://files.pythonhosted.org/packages/2c/d2/f552be5928105588f4f4d66ee37dd4c61460d8097e62d0e2e0eec41bc61d/sentencepiece-0.2.1-cp313-cp313-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:d7b670879c370d350557edabadbad1f6561a9e6968126e6debca4029e5547820", size = 1316271 }, - { url = "https://files.pythonhosted.org/packages/96/df/0cfe748ace5485be740fed9476dee7877f109da32ed0d280312c94ec259f/sentencepiece-0.2.1-cp313-cp313-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:c7f0fd2f2693309e6628aeeb2e2faf6edd221134dfccac3308ca0de01f8dab47", size = 1387882 }, - { url = "https://files.pythonhosted.org/packages/ac/dd/f7774d42a881ced8e1739f393ab1e82ece39fc9abd4779e28050c2e975b5/sentencepiece-0.2.1-cp313-cp313-win32.whl", hash = "sha256:92b3816aa2339355fda2c8c4e021a5de92180b00aaccaf5e2808972e77a4b22f", size = 999541 }, - { url = "https://files.pythonhosted.org/packages/dd/e9/932b9eae6fd7019548321eee1ab8d5e3b3d1294df9d9a0c9ac517c7b636d/sentencepiece-0.2.1-cp313-cp313-win_amd64.whl", hash = "sha256:10ed3dab2044c47f7a2e7b4969b0c430420cdd45735d78c8f853191fa0e3148b", size = 1054669 }, - { url = "https://files.pythonhosted.org/packages/c9/3a/76488a00ea7d6931689cda28726a1447d66bf1a4837943489314593d5596/sentencepiece-0.2.1-cp313-cp313-win_arm64.whl", hash = "sha256:ac650534e2251083c5f75dde4ff28896ce7c8904133dc8fef42780f4d5588fcd", size = 1033922 }, - { url = "https://files.pythonhosted.org/packages/4a/b6/08fe2ce819e02ccb0296f4843e3f195764ce9829cbda61b7513f29b95718/sentencepiece-0.2.1-cp313-cp313t-macosx_10_13_universal2.whl", hash = "sha256:8dd4b477a7b069648d19363aad0cab9bad2f4e83b2d179be668efa672500dc94", size = 1946052 }, - { url = "https://files.pythonhosted.org/packages/ab/d9/1ea0e740591ff4c6fc2b6eb1d7510d02f3fb885093f19b2f3abd1363b402/sentencepiece-0.2.1-cp313-cp313t-macosx_10_13_x86_64.whl", hash = "sha256:0c0f672da370cc490e4c59d89e12289778310a0e71d176c541e4834759e1ae07", size = 1327408 }, - { url = "https://files.pythonhosted.org/packages/99/7e/1fb26e8a21613f6200e1ab88824d5d203714162cf2883248b517deb500b7/sentencepiece-0.2.1-cp313-cp313t-macosx_11_0_arm64.whl", hash = "sha256:ad8493bea8432dae8d6830365352350f3b4144415a1d09c4c8cb8d30cf3b6c3c", size = 1254857 }, - { url = "https://files.pythonhosted.org/packages/bc/85/c72fd1f3c7a6010544d6ae07f8ddb38b5e2a7e33bd4318f87266c0bbafbf/sentencepiece-0.2.1-cp313-cp313t-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:b81a24733726e3678d2db63619acc5a8dccd074f7aa7a54ecd5ca33ca6d2d596", size = 1315722 }, - { url = "https://files.pythonhosted.org/packages/4a/e8/661e5bd82a8aa641fd6c1020bd0e890ef73230a2b7215ddf9c8cd8e941c2/sentencepiece-0.2.1-cp313-cp313t-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:0a81799d0a68d618e89063fb423c3001a034c893069135ffe51fee439ae474d6", size = 1387452 }, - { url = "https://files.pythonhosted.org/packages/99/5e/ae66c361023a470afcbc1fbb8da722c72ea678a2fcd9a18f1a12598c7501/sentencepiece-0.2.1-cp313-cp313t-win32.whl", hash = "sha256:89a3ea015517c42c0341d0d962f3e6aaf2cf10d71b1932d475c44ba48d00aa2b", size = 1002501 }, - { url = "https://files.pythonhosted.org/packages/c1/03/d332828c4ff764e16c1b56c2c8f9a33488bbe796b53fb6b9c4205ddbf167/sentencepiece-0.2.1-cp313-cp313t-win_amd64.whl", hash = "sha256:33f068c9382dc2e7c228eedfd8163b52baa86bb92f50d0488bf2b7da7032e484", size = 1057555 }, - { url = "https://files.pythonhosted.org/packages/88/14/5aee0bf0864df9bd82bd59e7711362908e4935e3f9cdc1f57246b5d5c9b9/sentencepiece-0.2.1-cp313-cp313t-win_arm64.whl", hash = "sha256:b3616ad246f360e52c85781e47682d31abfb6554c779e42b65333d4b5f44ecc0", size = 1036042 }, - { url = "https://files.pythonhosted.org/packages/24/9c/89eb8b2052f720a612478baf11c8227dcf1dc28cd4ea4c0c19506b5af2a2/sentencepiece-0.2.1-cp314-cp314-macosx_10_13_universal2.whl", hash = "sha256:5d0350b686c320068702116276cfb26c066dc7e65cfef173980b11bb4d606719", size = 1943147 }, - { url = "https://files.pythonhosted.org/packages/82/0b/a1432bc87f97c2ace36386ca23e8bd3b91fb40581b5e6148d24b24186419/sentencepiece-0.2.1-cp314-cp314-macosx_10_13_x86_64.whl", hash = "sha256:c7f54a31cde6fa5cb030370566f68152a742f433f8d2be458463d06c208aef33", size = 1325624 }, - { url = "https://files.pythonhosted.org/packages/ea/99/bbe054ebb5a5039457c590e0a4156ed073fb0fe9ce4f7523404dd5b37463/sentencepiece-0.2.1-cp314-cp314-macosx_11_0_arm64.whl", hash = "sha256:c83b85ab2d6576607f31df77ff86f28182be4a8de6d175d2c33ca609925f5da1", size = 1253670 }, - { url = "https://files.pythonhosted.org/packages/19/ad/d5c7075f701bd97971d7c2ac2904f227566f51ef0838dfbdfdccb58cd212/sentencepiece-0.2.1-cp314-cp314-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:1855f57db07b51fb51ed6c9c452f570624d2b169b36f0f79ef71a6e6c618cd8b", size = 1316247 }, - { url = "https://files.pythonhosted.org/packages/fb/03/35fbe5f3d9a7435eebd0b473e09584bd3cc354ce118b960445b060d33781/sentencepiece-0.2.1-cp314-cp314-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:01e6912125cb45d3792f530a4d38f8e21bf884d6b4d4ade1b2de5cf7a8d2a52b", size = 1387894 }, - { url = "https://files.pythonhosted.org/packages/dc/aa/956ef729aafb6c8f9c443104c9636489093bb5c61d6b90fc27aa1a865574/sentencepiece-0.2.1-cp314-cp314-win32.whl", hash = "sha256:c415c9de1447e0a74ae3fdb2e52f967cb544113a3a5ce3a194df185cbc1f962f", size = 1096698 }, - { url = "https://files.pythonhosted.org/packages/b8/cb/fe400d8836952cc535c81a0ce47dc6875160e5fedb71d2d9ff0e9894c2a6/sentencepiece-0.2.1-cp314-cp314-win_amd64.whl", hash = "sha256:881b2e44b14fc19feade3cbed314be37de639fc415375cefaa5bc81a4be137fd", size = 1155115 }, - { url = "https://files.pythonhosted.org/packages/32/89/047921cf70f36c7b6b6390876b2399b3633ab73b8d0cb857e5a964238941/sentencepiece-0.2.1-cp314-cp314-win_arm64.whl", hash = "sha256:2005242a16d2dc3ac5fe18aa7667549134d37854823df4c4db244752453b78a8", size = 1133890 }, - { url = "https://files.pythonhosted.org/packages/a1/11/5b414b9fae6255b5fb1e22e2ed3dc3a72d3a694e5703910e640ac78346bb/sentencepiece-0.2.1-cp314-cp314t-macosx_10_13_universal2.whl", hash = "sha256:a19adcec27c524cb7069a1c741060add95f942d1cbf7ad0d104dffa0a7d28a2b", size = 1946081 }, - { url = "https://files.pythonhosted.org/packages/77/eb/7a5682bb25824db8545f8e5662e7f3e32d72a508fdce086029d89695106b/sentencepiece-0.2.1-cp314-cp314t-macosx_10_13_x86_64.whl", hash = "sha256:e37e4b4c4a11662b5db521def4e44d4d30ae69a1743241412a93ae40fdcab4bb", size = 1327406 }, - { url = "https://files.pythonhosted.org/packages/03/b0/811dae8fb9f2784e138785d481469788f2e0d0c109c5737372454415f55f/sentencepiece-0.2.1-cp314-cp314t-macosx_11_0_arm64.whl", hash = "sha256:477c81505db072b3ab627e7eab972ea1025331bd3a92bacbf798df2b75ea86ec", size = 1254846 }, - { url = "https://files.pythonhosted.org/packages/ef/23/195b2e7ec85ebb6a547969f60b723c7aca5a75800ece6cc3f41da872d14e/sentencepiece-0.2.1-cp314-cp314t-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:010f025a544ef770bb395091d57cb94deb9652d8972e0d09f71d85d5a0816c8c", size = 1315721 }, - { url = "https://files.pythonhosted.org/packages/7e/aa/553dbe4178b5f23eb28e59393dddd64186178b56b81d9b8d5c3ff1c28395/sentencepiece-0.2.1-cp314-cp314t-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:733e59ff1794d26db706cd41fc2d7ca5f6c64a820709cb801dc0ea31780d64ab", size = 1387458 }, - { url = "https://files.pythonhosted.org/packages/66/7c/08ff0012507297a4dd74a5420fdc0eb9e3e80f4e88cab1538d7f28db303d/sentencepiece-0.2.1-cp314-cp314t-win32.whl", hash = "sha256:d3233770f78e637dc8b1fda2cd7c3b99ec77e7505041934188a4e7fe751de3b0", size = 1099765 }, - { url = "https://files.pythonhosted.org/packages/91/d5/2a69e1ce15881beb9ddfc7e3f998322f5cedcd5e4d244cb74dade9441663/sentencepiece-0.2.1-cp314-cp314t-win_amd64.whl", hash = "sha256:5e4366c97b68218fd30ea72d70c525e6e78a6c0a88650f57ac4c43c63b234a9d", size = 1157807 }, - { url = "https://files.pythonhosted.org/packages/f3/16/54f611fcfc2d1c46cbe3ec4169780b2cfa7cf63708ef2b71611136db7513/sentencepiece-0.2.1-cp314-cp314t-win_arm64.whl", hash = "sha256:105e36e75cbac1292642045458e8da677b2342dcd33df503e640f0b457cb6751", size = 1136264 }, -] - -[[package]] -name = "sentry-sdk" -version = "2.48.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "certifi" }, - { name = "urllib3" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/40/f0/0e9dc590513d5e742d7799e2038df3a05167cba084c6ca4f3cdd75b55164/sentry_sdk-2.48.0.tar.gz", hash = "sha256:5213190977ff7fdff8a58b722fb807f8d5524a80488626ebeda1b5676c0c1473", size = 384828 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/4d/19/8d77f9992e5cbfcaa9133c3bf63b4fbbb051248802e1e803fed5c552fbb2/sentry_sdk-2.48.0-py2.py3-none-any.whl", hash = "sha256:6b12ac256769d41825d9b7518444e57fa35b5642df4c7c5e322af4d2c8721172", size = 414555 }, -] - -[[package]] -name = "setuptools" -version = "80.9.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/18/5d/3bf57dcd21979b887f014ea83c24ae194cfcd12b9e0fda66b957c69d1fca/setuptools-80.9.0.tar.gz", hash = "sha256:f36b47402ecde768dbfafc46e8e4207b4360c654f1f3bb84475f0a28628fb19c", size = 1319958 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/a3/dc/17031897dae0efacfea57dfd3a82fdd2a2aeb58e0ff71b77b87e44edc772/setuptools-80.9.0-py3-none-any.whl", hash = "sha256:062d34222ad13e0cc312a4c02d73f059e86a4acbfbdea8f8f76b28c99f306922", size = 1201486 }, -] - -[[package]] -name = "shellingham" -version = "1.5.4" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/58/15/8b3609fd3830ef7b27b655beb4b4e9c62313a4e8da8c676e142cc210d58e/shellingham-1.5.4.tar.gz", hash = "sha256:8dbca0739d487e5bd35ab3ca4b36e11c4078f3a234bfce294b0a0291363404de", size = 10310 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/e0/f9/0595336914c5619e5f28a1fb793285925a8cd4b432c9da0a987836c7f822/shellingham-1.5.4-py2.py3-none-any.whl", hash = "sha256:7ecfff8f2fd72616f7481040475a65b2bf8af90a56c89140852d1120324e8686", size = 9755 }, -] - -[[package]] -name = "six" -version = "1.17.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/94/e7/b2c673351809dca68a0e064b6af791aa332cf192da575fd474ed7d6f16a2/six-1.17.0.tar.gz", hash = "sha256:ff70335d468e7eb6ec65b95b99d3a2836546063f63acc5171de367e834932a81", size = 34031 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/b7/ce/149a00dd41f10bc29e5921b496af8b574d8413afcd5e30dfa0ed46c2cc5e/six-1.17.0-py2.py3-none-any.whl", hash = "sha256:4721f391ed90541fddacab5acf947aa0d3dc7d27b2e1e8eda2be8970586c3274", size = 11050 }, -] - -[[package]] -name = "smmap" -version = "5.0.2" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/44/cd/a040c4b3119bbe532e5b0732286f805445375489fceaec1f48306068ee3b/smmap-5.0.2.tar.gz", hash = "sha256:26ea65a03958fa0c8a1c7e8c7a58fdc77221b8910f6be2131affade476898ad5", size = 22329 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/04/be/d09147ad1ec7934636ad912901c5fd7667e1c858e19d355237db0d0cd5e4/smmap-5.0.2-py3-none-any.whl", hash = "sha256:b30115f0def7d7531d22a0fb6502488d879e75b260a9db4d0819cfb25403af5e", size = 24303 }, -] - -[[package]] -name = "sortedcontainers" -version = "2.4.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/e8/c4/ba2f8066cceb6f23394729afe52f3bf7adec04bf9ed2c820b39e19299111/sortedcontainers-2.4.0.tar.gz", hash = "sha256:25caa5a06cc30b6b83d11423433f65d1f9d76c4c6a0c90e3379eaa43b9bfdb88", size = 30594 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/32/46/9cb0e58b2deb7f82b84065f37f3bffeb12413f947f9388e4cac22c4621ce/sortedcontainers-2.4.0-py2.py3-none-any.whl", hash = "sha256:a163dcaede0f1c021485e957a39245190e74249897e2ae4b2aa38595db237ee0", size = 29575 }, -] - -[[package]] -name = "soundfile" -version = "0.13.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "cffi" }, - { name = "numpy" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/e1/41/9b873a8c055582859b239be17902a85339bec6a30ad162f98c9b0288a2cc/soundfile-0.13.1.tar.gz", hash = "sha256:b2c68dab1e30297317080a5b43df57e302584c49e2942defdde0acccc53f0e5b", size = 46156 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/64/28/e2a36573ccbcf3d57c00626a21fe51989380636e821b341d36ccca0c1c3a/soundfile-0.13.1-py2.py3-none-any.whl", hash = "sha256:a23c717560da2cf4c7b5ae1142514e0fd82d6bbd9dfc93a50423447142f2c445", size = 25751 }, - { url = "https://files.pythonhosted.org/packages/ea/ab/73e97a5b3cc46bba7ff8650a1504348fa1863a6f9d57d7001c6b67c5f20e/soundfile-0.13.1-py2.py3-none-macosx_10_9_x86_64.whl", hash = "sha256:82dc664d19831933fe59adad199bf3945ad06d84bc111a5b4c0d3089a5b9ec33", size = 1142250 }, - { url = "https://files.pythonhosted.org/packages/a0/e5/58fd1a8d7b26fc113af244f966ee3aecf03cb9293cb935daaddc1e455e18/soundfile-0.13.1-py2.py3-none-macosx_11_0_arm64.whl", hash = "sha256:743f12c12c4054921e15736c6be09ac26b3b3d603aef6fd69f9dde68748f2593", size = 1101406 }, - { url = "https://files.pythonhosted.org/packages/58/ae/c0e4a53d77cf6e9a04179535766b3321b0b9ced5f70522e4caf9329f0046/soundfile-0.13.1-py2.py3-none-manylinux_2_28_aarch64.whl", hash = "sha256:9c9e855f5a4d06ce4213f31918653ab7de0c5a8d8107cd2427e44b42df547deb", size = 1235729 }, - { url = "https://files.pythonhosted.org/packages/57/5e/70bdd9579b35003a489fc850b5047beeda26328053ebadc1fb60f320f7db/soundfile-0.13.1-py2.py3-none-manylinux_2_28_x86_64.whl", hash = "sha256:03267c4e493315294834a0870f31dbb3b28a95561b80b134f0bd3cf2d5f0e618", size = 1313646 }, - { url = "https://files.pythonhosted.org/packages/fe/df/8c11dc4dfceda14e3003bb81a0d0edcaaf0796dd7b4f826ea3e532146bba/soundfile-0.13.1-py2.py3-none-win32.whl", hash = "sha256:c734564fab7c5ddf8e9be5bf70bab68042cd17e9c214c06e365e20d64f9a69d5", size = 899881 }, - { url = "https://files.pythonhosted.org/packages/14/e9/6b761de83277f2f02ded7e7ea6f07828ec78e4b229b80e4ca55dd205b9dc/soundfile-0.13.1-py2.py3-none-win_amd64.whl", hash = "sha256:1e70a05a0626524a69e9f0f4dd2ec174b4e9567f4d8b6c11d38b5c289be36ee9", size = 1019162 }, -] - -[[package]] -name = "sox" -version = "1.5.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "numpy" }, - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/3d/a2/d8e0d8fd7abf509ead4a2cb0fb24e5758b5330166bf9223d5cb9f98a7e8d/sox-1.5.0.tar.gz", hash = "sha256:12c7be5bb1f548d891fe11e82c08cf5f1a1d74e225298f60082e5aeb2469ada0", size = 63905 } - -[[package]] -name = "soxr" -version = "1.0.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "numpy" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/42/7e/f4b461944662ad75036df65277d6130f9411002bfb79e9df7dff40a31db9/soxr-1.0.0.tar.gz", hash = "sha256:e07ee6c1d659bc6957034f4800c60cb8b98de798823e34d2a2bba1caa85a4509", size = 171415 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/1e/a7/11c36d71595b52fe84a220040ace679035953acf06b83bf2c7117c565d2c/soxr-1.0.0-cp310-cp310-macosx_10_14_x86_64.whl", hash = "sha256:b876a3156f67c76aef0cff1084eaf4088d9ca584bb569cb993f89a52ec5f399f", size = 206459 }, - { url = "https://files.pythonhosted.org/packages/43/5e/8962f2aeea7777d2a6e65a24a2b83c6aea1a28badeda027fd328f7f03bb7/soxr-1.0.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:4d3b957a7b0cc19ae6aa45d40b2181474e53a8dd00efd7bce6bcf4e60e020892", size = 164808 }, - { url = "https://files.pythonhosted.org/packages/fc/91/00384166f110a3888ea8efd44523ba7168dd2dc39e3e43c931cc2d069fa9/soxr-1.0.0-cp310-cp310-manylinux_2_26_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:b89685faedebc45af71f08f9957b61cc6143bc94ba43fe38e97067f81e272969", size = 208586 }, - { url = "https://files.pythonhosted.org/packages/75/34/e18f1003e242aabed44ed8902534814d3e64209e4d1d874f5b9b67d73cde/soxr-1.0.0-cp310-cp310-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:d255741b2f0084fd02d4a2ddd77cd495be9e7e7b6f9dba1c9494f86afefac65b", size = 242310 }, - { url = "https://files.pythonhosted.org/packages/61/9c/a1c5ed106b40cc1e2e12cd58831b7f1b61c5fbdb8eceeca4b3a0b0dbef6c/soxr-1.0.0-cp310-cp310-win_amd64.whl", hash = "sha256:158a4a9055958c4b95ef91dbbe280cabb00946b5423b25a9b0ce31bd9e0a271e", size = 173561 }, - { url = "https://files.pythonhosted.org/packages/65/ce/a3262bc8733d3a4ce5f660ed88c3d97f4b12658b0909e71334cba1721dcb/soxr-1.0.0-cp311-cp311-macosx_10_14_x86_64.whl", hash = "sha256:28e19d74a5ef45c0d7000f3c70ec1719e89077379df2a1215058914d9603d2d8", size = 206739 }, - { url = "https://files.pythonhosted.org/packages/64/dc/e8cbd100b652697cc9865dbed08832e7e135ff533f453eb6db9e6168d153/soxr-1.0.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:f8dc69fc18884e53b72f6141fdf9d80997edbb4fec9dc2942edcb63abbe0d023", size = 165233 }, - { url = "https://files.pythonhosted.org/packages/75/12/4b49611c9ba5e9fe6f807d0a83352516808e8e573f8b4e712fc0c17f3363/soxr-1.0.0-cp311-cp311-manylinux_2_26_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:3f15450e6f65f22f02fcd4c5a9219c873b1e583a73e232805ff160c759a6b586", size = 208867 }, - { url = "https://files.pythonhosted.org/packages/cc/70/92146ab970a3ef8c43ac160035b1e52fde5417f89adb10572f7e788d9596/soxr-1.0.0-cp311-cp311-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:1f73f57452f9df37b4de7a4052789fcbd474a5b28f38bba43278ae4b489d4384", size = 242633 }, - { url = "https://files.pythonhosted.org/packages/b5/a7/628479336206959463d08260bffed87905e7ba9e3bd83ca6b405a0736e94/soxr-1.0.0-cp311-cp311-win_amd64.whl", hash = "sha256:9f417c3d69236051cf5a1a7bad7c4bff04eb3d8fcaa24ac1cb06e26c8d48d8dc", size = 173814 }, - { url = "https://files.pythonhosted.org/packages/c5/c7/f92b81f1a151c13afb114f57799b86da9330bec844ea5a0d3fe6a8732678/soxr-1.0.0-cp312-abi3-macosx_10_14_x86_64.whl", hash = "sha256:abecf4e39017f3fadb5e051637c272ae5778d838e5c3926a35db36a53e3a607f", size = 205508 }, - { url = "https://files.pythonhosted.org/packages/ff/1d/c945fea9d83ea1f2be9d116b3674dbaef26ed090374a77c394b31e3b083b/soxr-1.0.0-cp312-abi3-macosx_11_0_arm64.whl", hash = "sha256:e973d487ee46aa8023ca00a139db6e09af053a37a032fe22f9ff0cc2e19c94b4", size = 163568 }, - { url = "https://files.pythonhosted.org/packages/b5/80/10640970998a1d2199bef6c4d92205f36968cddaf3e4d0e9fe35ddd405bd/soxr-1.0.0-cp312-abi3-manylinux_2_26_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:e8ce273cca101aff3d8c387db5a5a41001ba76ef1837883438d3c652507a9ccc", size = 204707 }, - { url = "https://files.pythonhosted.org/packages/b1/87/2726603c13c2126cb8ded9e57381b7377f4f0df6ba4408e1af5ddbfdc3dd/soxr-1.0.0-cp312-abi3-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:e8f2a69686f2856d37823bbb7b78c3d44904f311fe70ba49b893af11d6b6047b", size = 238032 }, - { url = "https://files.pythonhosted.org/packages/ce/04/530252227f4d0721a5524a936336485dfb429bb206a66baf8e470384f4a2/soxr-1.0.0-cp312-abi3-win_amd64.whl", hash = "sha256:2a3b77b115ae7c478eecdbd060ed4f61beda542dfb70639177ac263aceda42a2", size = 172070 }, - { url = "https://files.pythonhosted.org/packages/99/77/d3b3c25b4f1b1aa4a73f669355edcaee7a52179d0c50407697200a0e55b9/soxr-1.0.0-cp314-cp314t-macosx_10_14_x86_64.whl", hash = "sha256:392a5c70c04eb939c9c176bd6f654dec9a0eaa9ba33d8f1024ed63cf68cdba0a", size = 209509 }, - { url = "https://files.pythonhosted.org/packages/8a/ee/3ca73e18781bb2aff92b809f1c17c356dfb9a1870652004bd432e79afbfa/soxr-1.0.0-cp314-cp314t-macosx_11_0_arm64.whl", hash = "sha256:fdc41a1027ba46777186f26a8fba7893be913383414135577522da2fcc684490", size = 167690 }, - { url = "https://files.pythonhosted.org/packages/bd/f0/eea8b5f587a2531657dc5081d2543a5a845f271a3bea1c0fdee5cebde021/soxr-1.0.0-cp314-cp314t-manylinux_2_26_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:449acd1dfaf10f0ce6dfd75c7e2ef984890df94008765a6742dafb42061c1a24", size = 209541 }, - { url = "https://files.pythonhosted.org/packages/64/59/2430a48c705565eb09e78346950b586f253a11bd5313426ced3ecd9b0feb/soxr-1.0.0-cp314-cp314t-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:38b35c99e408b8f440c9376a5e1dd48014857cd977c117bdaa4304865ae0edd0", size = 243025 }, - { url = "https://files.pythonhosted.org/packages/3c/1b/f84a2570a74094e921bbad5450b2a22a85d58585916e131d9b98029c3e69/soxr-1.0.0-cp314-cp314t-win_amd64.whl", hash = "sha256:a39b519acca2364aa726b24a6fd55acf29e4c8909102e0b858c23013c38328e5", size = 184850 }, -] - -[[package]] -name = "sqlalchemy" -version = "2.0.45" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "greenlet", marker = "platform_machine == 'AMD64' or platform_machine == 'WIN32' or platform_machine == 'aarch64' or platform_machine == 'amd64' or platform_machine == 'ppc64le' or platform_machine == 'win32' or platform_machine == 'x86_64'" }, - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/be/f9/5e4491e5ccf42f5d9cfc663741d261b3e6e1683ae7812114e7636409fcc6/sqlalchemy-2.0.45.tar.gz", hash = "sha256:1632a4bda8d2d25703fdad6363058d882541bdaaee0e5e3ddfa0cd3229efce88", size = 9869912 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/fe/70/75b1387d72e2847220441166c5eb4e9846dd753895208c13e6d66523b2d9/sqlalchemy-2.0.45-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:c64772786d9eee72d4d3784c28f0a636af5b0a29f3fe26ff11f55efe90c0bd85", size = 2154148 }, - { url = "https://files.pythonhosted.org/packages/d8/a4/7805e02323c49cb9d1ae5cd4913b28c97103079765f520043f914fca4cb3/sqlalchemy-2.0.45-cp310-cp310-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:7ae64ebf7657395824a19bca98ab10eb9a3ecb026bf09524014f1bb81cb598d4", size = 3233051 }, - { url = "https://files.pythonhosted.org/packages/d7/ec/32ae09139f61bef3de3142e85c47abdee8db9a55af2bb438da54a4549263/sqlalchemy-2.0.45-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:0f02325709d1b1a1489f23a39b318e175a171497374149eae74d612634b234c0", size = 3232781 }, - { url = "https://files.pythonhosted.org/packages/ad/bd/bf7b869b6f5585eac34222e1cf4405f4ba8c3b85dd6b1af5d4ce8bca695f/sqlalchemy-2.0.45-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:d2c3684fca8a05f0ac1d9a21c1f4a266983a7ea9180efb80ffeb03861ecd01a0", size = 3182096 }, - { url = "https://files.pythonhosted.org/packages/21/6a/c219720a241bb8f35c88815ccc27761f5af7fdef04b987b0e8a2c1a6dcaa/sqlalchemy-2.0.45-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:040f6f0545b3b7da6b9317fc3e922c9a98fc7243b2a1b39f78390fc0942f7826", size = 3205109 }, - { url = "https://files.pythonhosted.org/packages/bd/c4/6ccf31b2bc925d5d95fab403ffd50d20d7c82b858cf1a4855664ca054dce/sqlalchemy-2.0.45-cp310-cp310-win32.whl", hash = "sha256:830d434d609fe7bfa47c425c445a8b37929f140a7a44cdaf77f6d34df3a7296a", size = 2114240 }, - { url = "https://files.pythonhosted.org/packages/de/29/a27a31fca07316def418db6f7c70ab14010506616a2decef1906050a0587/sqlalchemy-2.0.45-cp310-cp310-win_amd64.whl", hash = "sha256:0209d9753671b0da74da2cfbb9ecf9c02f72a759e4b018b3ab35f244c91842c7", size = 2137615 }, - { url = "https://files.pythonhosted.org/packages/a2/1c/769552a9d840065137272ebe86ffbb0bc92b0f1e0a68ee5266a225f8cd7b/sqlalchemy-2.0.45-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:2e90a344c644a4fa871eb01809c32096487928bd2038bf10f3e4515cb688cc56", size = 2153860 }, - { url = "https://files.pythonhosted.org/packages/f3/f8/9be54ff620e5b796ca7b44670ef58bc678095d51b0e89d6e3102ea468216/sqlalchemy-2.0.45-cp311-cp311-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:b8c8b41b97fba5f62349aa285654230296829672fc9939cd7f35aab246d1c08b", size = 3309379 }, - { url = "https://files.pythonhosted.org/packages/f6/2b/60ce3ee7a5ae172bfcd419ce23259bb874d2cddd44f67c5df3760a1e22f9/sqlalchemy-2.0.45-cp311-cp311-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:12c694ed6468333a090d2f60950e4250b928f457e4962389553d6ba5fe9951ac", size = 3309948 }, - { url = "https://files.pythonhosted.org/packages/a3/42/bac8d393f5db550e4e466d03d16daaafd2bad1f74e48c12673fb499a7fc1/sqlalchemy-2.0.45-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:f7d27a1d977a1cfef38a0e2e1ca86f09c4212666ce34e6ae542f3ed0a33bc606", size = 3261239 }, - { url = "https://files.pythonhosted.org/packages/6f/12/43dc70a0528c59842b04ea1c1ed176f072a9b383190eb015384dd102fb19/sqlalchemy-2.0.45-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:d62e47f5d8a50099b17e2bfc1b0c7d7ecd8ba6b46b1507b58cc4f05eefc3bb1c", size = 3284065 }, - { url = "https://files.pythonhosted.org/packages/cf/9c/563049cf761d9a2ec7bc489f7879e9d94e7b590496bea5bbee9ed7b4cc32/sqlalchemy-2.0.45-cp311-cp311-win32.whl", hash = "sha256:3c5f76216e7b85770d5bb5130ddd11ee89f4d52b11783674a662c7dd57018177", size = 2113480 }, - { url = "https://files.pythonhosted.org/packages/bc/fa/09d0a11fe9f15c7fa5c7f0dd26be3d235b0c0cbf2f9544f43bc42efc8a24/sqlalchemy-2.0.45-cp311-cp311-win_amd64.whl", hash = "sha256:a15b98adb7f277316f2c276c090259129ee4afca783495e212048daf846654b2", size = 2138407 }, - { url = "https://files.pythonhosted.org/packages/2d/c7/1900b56ce19bff1c26f39a4ce427faec7716c81ac792bfac8b6a9f3dca93/sqlalchemy-2.0.45-cp312-cp312-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:b3ee2aac15169fb0d45822983631466d60b762085bc4535cd39e66bea362df5f", size = 3333760 }, - { url = "https://files.pythonhosted.org/packages/0a/93/3be94d96bb442d0d9a60e55a6bb6e0958dd3457751c6f8502e56ef95fed0/sqlalchemy-2.0.45-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:ba547ac0b361ab4f1608afbc8432db669bd0819b3e12e29fb5fa9529a8bba81d", size = 3348268 }, - { url = "https://files.pythonhosted.org/packages/48/4b/f88ded696e61513595e4a9778f9d3f2bf7332cce4eb0c7cedaabddd6687b/sqlalchemy-2.0.45-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:215f0528b914e5c75ef2559f69dca86878a3beeb0c1be7279d77f18e8d180ed4", size = 3278144 }, - { url = "https://files.pythonhosted.org/packages/ed/6a/310ecb5657221f3e1bd5288ed83aa554923fb5da48d760a9f7622afeb065/sqlalchemy-2.0.45-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:107029bf4f43d076d4011f1afb74f7c3e2ea029ec82eb23d8527d5e909e97aa6", size = 3313907 }, - { url = "https://files.pythonhosted.org/packages/5c/39/69c0b4051079addd57c84a5bfb34920d87456dd4c90cf7ee0df6efafc8ff/sqlalchemy-2.0.45-cp312-cp312-win32.whl", hash = "sha256:0c9f6ada57b58420a2c0277ff853abe40b9e9449f8d7d231763c6bc30f5c4953", size = 2112182 }, - { url = "https://files.pythonhosted.org/packages/f7/4e/510db49dd89fc3a6e994bee51848c94c48c4a00dc905e8d0133c251f41a7/sqlalchemy-2.0.45-cp312-cp312-win_amd64.whl", hash = "sha256:8defe5737c6d2179c7997242d6473587c3beb52e557f5ef0187277009f73e5e1", size = 2139200 }, - { url = "https://files.pythonhosted.org/packages/6a/c8/7cc5221b47a54edc72a0140a1efa56e0a2730eefa4058d7ed0b4c4357ff8/sqlalchemy-2.0.45-cp313-cp313-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:fe187fc31a54d7fd90352f34e8c008cf3ad5d064d08fedd3de2e8df83eb4a1cf", size = 3277082 }, - { url = "https://files.pythonhosted.org/packages/0e/50/80a8d080ac7d3d321e5e5d420c9a522b0aa770ec7013ea91f9a8b7d36e4a/sqlalchemy-2.0.45-cp313-cp313-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:672c45cae53ba88e0dad74b9027dddd09ef6f441e927786b05bec75d949fbb2e", size = 3293131 }, - { url = "https://files.pythonhosted.org/packages/da/4c/13dab31266fc9904f7609a5dc308a2432a066141d65b857760c3bef97e69/sqlalchemy-2.0.45-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:470daea2c1ce73910f08caf10575676a37159a6d16c4da33d0033546bddebc9b", size = 3225389 }, - { url = "https://files.pythonhosted.org/packages/74/04/891b5c2e9f83589de202e7abaf24cd4e4fa59e1837d64d528829ad6cc107/sqlalchemy-2.0.45-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:9c6378449e0940476577047150fd09e242529b761dc887c9808a9a937fe990c8", size = 3266054 }, - { url = "https://files.pythonhosted.org/packages/f1/24/fc59e7f71b0948cdd4cff7a286210e86b0443ef1d18a23b0d83b87e4b1f7/sqlalchemy-2.0.45-cp313-cp313-win32.whl", hash = "sha256:4b6bec67ca45bc166c8729910bd2a87f1c0407ee955df110d78948f5b5827e8a", size = 2110299 }, - { url = "https://files.pythonhosted.org/packages/c0/c5/d17113020b2d43073412aeca09b60d2009442420372123b8d49cc253f8b8/sqlalchemy-2.0.45-cp313-cp313-win_amd64.whl", hash = "sha256:afbf47dc4de31fa38fd491f3705cac5307d21d4bb828a4f020ee59af412744ee", size = 2136264 }, - { url = "https://files.pythonhosted.org/packages/3d/8d/bb40a5d10e7a5f2195f235c0b2f2c79b0bf6e8f00c0c223130a4fbd2db09/sqlalchemy-2.0.45-cp313-cp313t-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:83d7009f40ce619d483d26ac1b757dfe3167b39921379a8bd1b596cf02dab4a6", size = 3521998 }, - { url = "https://files.pythonhosted.org/packages/75/a5/346128b0464886f036c039ea287b7332a410aa2d3fb0bb5d404cb8861635/sqlalchemy-2.0.45-cp313-cp313t-musllinux_1_2_x86_64.whl", hash = "sha256:d8a2ca754e5415cde2b656c27900b19d50ba076aa05ce66e2207623d3fe41f5a", size = 3473434 }, - { url = "https://files.pythonhosted.org/packages/cc/64/4e1913772646b060b025d3fc52ce91a58967fe58957df32b455de5a12b4f/sqlalchemy-2.0.45-cp314-cp314-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:7f46ec744e7f51275582e6a24326e10c49fbdd3fc99103e01376841213028774", size = 3272404 }, - { url = "https://files.pythonhosted.org/packages/b3/27/caf606ee924282fe4747ee4fd454b335a72a6e018f97eab5ff7f28199e16/sqlalchemy-2.0.45-cp314-cp314-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:883c600c345123c033c2f6caca18def08f1f7f4c3ebeb591a63b6fceffc95cce", size = 3277057 }, - { url = "https://files.pythonhosted.org/packages/85/d0/3d64218c9724e91f3d1574d12eb7ff8f19f937643815d8daf792046d88ab/sqlalchemy-2.0.45-cp314-cp314-musllinux_1_2_aarch64.whl", hash = "sha256:2c0b74aa79e2deade948fe8593654c8ef4228c44ba862bb7c9585c8e0db90f33", size = 3222279 }, - { url = "https://files.pythonhosted.org/packages/24/10/dd7688a81c5bc7690c2a3764d55a238c524cd1a5a19487928844cb247695/sqlalchemy-2.0.45-cp314-cp314-musllinux_1_2_x86_64.whl", hash = "sha256:8a420169cef179d4c9064365f42d779f1e5895ad26ca0c8b4c0233920973db74", size = 3244508 }, - { url = "https://files.pythonhosted.org/packages/aa/41/db75756ca49f777e029968d9c9fee338c7907c563267740c6d310a8e3f60/sqlalchemy-2.0.45-cp314-cp314-win32.whl", hash = "sha256:e50dcb81a5dfe4b7b4a4aa8f338116d127cb209559124f3694c70d6cd072b68f", size = 2113204 }, - { url = "https://files.pythonhosted.org/packages/89/a2/0e1590e9adb292b1d576dbcf67ff7df8cf55e56e78d2c927686d01080f4b/sqlalchemy-2.0.45-cp314-cp314-win_amd64.whl", hash = "sha256:4748601c8ea959e37e03d13dcda4a44837afcd1b21338e637f7c935b8da06177", size = 2138785 }, - { url = "https://files.pythonhosted.org/packages/42/39/f05f0ed54d451156bbed0e23eb0516bcad7cbb9f18b3bf219c786371b3f0/sqlalchemy-2.0.45-cp314-cp314t-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:cd337d3526ec5298f67d6a30bbbe4ed7e5e68862f0bf6dd21d289f8d37b7d60b", size = 3522029 }, - { url = "https://files.pythonhosted.org/packages/54/0f/d15398b98b65c2bce288d5ee3f7d0a81f77ab89d9456994d5c7cc8b2a9db/sqlalchemy-2.0.45-cp314-cp314t-musllinux_1_2_x86_64.whl", hash = "sha256:9a62b446b7d86a3909abbcd1cd3cc550a832f99c2bc37c5b22e1925438b9367b", size = 3475142 }, - { url = "https://files.pythonhosted.org/packages/bf/e1/3ccb13c643399d22289c6a9786c1a91e3dcbb68bce4beb44926ac2c557bf/sqlalchemy-2.0.45-py3-none-any.whl", hash = "sha256:5225a288e4c8cc2308dbdd874edad6e7d0fd38eac1e9e5f23503425c8eee20d0", size = 1936672 }, -] - -[[package]] -name = "stack-data" -version = "0.6.3" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "asttokens" }, - { name = "executing" }, - { name = "pure-eval" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/28/e3/55dcc2cfbc3ca9c29519eb6884dd1415ecb53b0e934862d3559ddcb7e20b/stack_data-0.6.3.tar.gz", hash = "sha256:836a778de4fec4dcd1dcd89ed8abff8a221f58308462e1c4aa2a3cf30148f0b9", size = 44707 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/f1/7b/ce1eafaf1a76852e2ec9b22edecf1daa58175c090266e9f6c64afcd81d91/stack_data-0.6.3-py3-none-any.whl", hash = "sha256:d5558e0c25a4cb0853cddad3d77da9891a08cb85dd9f9f91b9f8cd66e511e695", size = 24521 }, -] - -[[package]] -name = "standard-aifc" -version = "3.13.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "audioop-lts", marker = "python_full_version >= '3.13'" }, - { name = "standard-chunk", marker = "python_full_version >= '3.13'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/c4/53/6050dc3dde1671eb3db592c13b55a8005e5040131f7509cef0215212cb84/standard_aifc-3.13.0.tar.gz", hash = "sha256:64e249c7cb4b3daf2fdba4e95721f811bde8bdfc43ad9f936589b7bb2fae2e43", size = 15240 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/c3/52/5fbb203394cc852334d1575cc020f6bcec768d2265355984dfd361968f36/standard_aifc-3.13.0-py3-none-any.whl", hash = "sha256:f7ae09cc57de1224a0dd8e3eb8f73830be7c3d0bc485de4c1f82b4a7f645ac66", size = 10492 }, -] - -[[package]] -name = "standard-chunk" -version = "3.13.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/43/06/ce1bb165c1f111c7d23a1ad17204d67224baa69725bb6857a264db61beaf/standard_chunk-3.13.0.tar.gz", hash = "sha256:4ac345d37d7e686d2755e01836b8d98eda0d1a3ee90375e597ae43aaf064d654", size = 4672 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/7a/90/a5c1084d87767d787a6caba615aa50dc587229646308d9420c960cb5e4c0/standard_chunk-3.13.0-py3-none-any.whl", hash = "sha256:17880a26c285189c644bd5bd8f8ed2bdb795d216e3293e6dbe55bbd848e2982c", size = 4944 }, -] - -[[package]] -name = "standard-sunau" -version = "3.13.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "audioop-lts", marker = "python_full_version >= '3.13'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/66/e3/ce8d38cb2d70e05ffeddc28bb09bad77cfef979eb0a299c9117f7ed4e6a9/standard_sunau-3.13.0.tar.gz", hash = "sha256:b319a1ac95a09a2378a8442f403c66f4fd4b36616d6df6ae82b8e536ee790908", size = 9368 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/34/ae/e3707f6c1bc6f7aa0df600ba8075bfb8a19252140cd595335be60e25f9ee/standard_sunau-3.13.0-py3-none-any.whl", hash = "sha256:53af624a9529c41062f4c2fd33837f297f3baa196b0cfceffea6555654602622", size = 7364 }, -] - -[[package]] -name = "strenum" -version = "0.4.15" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/85/ad/430fb60d90e1d112a62ff57bdd1f286ec73a2a0331272febfddd21f330e1/StrEnum-0.4.15.tar.gz", hash = "sha256:878fb5ab705442070e4dd1929bb5e2249511c0bcf2b0eeacf3bcd80875c82eff", size = 23384 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/81/69/297302c5f5f59c862faa31e6cb9a4cd74721cd1e052b38e464c5b402df8b/StrEnum-0.4.15-py3-none-any.whl", hash = "sha256:a30cda4af7cc6b5bf52c8055bc4bf4b2b6b14a93b574626da33df53cf7740659", size = 8851 }, -] - -[[package]] -name = "sympy" -version = "1.14.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "mpmath" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/83/d3/803453b36afefb7c2bb238361cd4ae6125a569b4db67cd9e79846ba2d68c/sympy-1.14.0.tar.gz", hash = "sha256:d3d3fe8df1e5a0b42f0e7bdf50541697dbe7d23746e894990c030e2b05e72517", size = 7793921 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/a2/09/77d55d46fd61b4a135c444fc97158ef34a095e5681d0a6c10b75bf356191/sympy-1.14.0-py3-none-any.whl", hash = "sha256:e091cc3e99d2141a0ba2847328f5479b05d94a6635cb96148ccb3f34671bd8f5", size = 6299353 }, -] - -[[package]] -name = "tabulate" -version = "0.9.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/ec/fe/802052aecb21e3797b8f7902564ab6ea0d60ff8ca23952079064155d1ae1/tabulate-0.9.0.tar.gz", hash = "sha256:0095b12bf5966de529c0feb1fa08671671b3368eec77d7ef7ab114be2c068b3c", size = 81090 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/40/44/4a5f08c96eb108af5cb50b41f76142f0afa346dfa99d5296fe7202a11854/tabulate-0.9.0-py3-none-any.whl", hash = "sha256:024ca478df22e9340661486f85298cff5f6dcdba14f3813e8830015b9ed1948f", size = 35252 }, -] - -[[package]] -name = "tensorboard" -version = "2.20.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "absl-py" }, - { name = "grpcio" }, - { name = "markdown" }, - { name = "numpy" }, - { name = "packaging" }, - { name = "pillow" }, - { name = "protobuf" }, - { name = "setuptools" }, - { name = "tensorboard-data-server" }, - { name = "werkzeug" }, -] -wheels = [ - { url = "https://files.pythonhosted.org/packages/9c/d9/a5db55f88f258ac669a92858b70a714bbbd5acd993820b41ec4a96a4d77f/tensorboard-2.20.0-py3-none-any.whl", hash = "sha256:9dc9f978cb84c0723acf9a345d96c184f0293d18f166bb8d59ee098e6cfaaba6", size = 5525680 }, -] - -[[package]] -name = "tensorboard-data-server" -version = "0.7.2" -source = { registry = "https://pypi.org/simple" } -wheels = [ - { url = "https://files.pythonhosted.org/packages/7a/13/e503968fefabd4c6b2650af21e110aa8466fe21432cd7c43a84577a89438/tensorboard_data_server-0.7.2-py3-none-any.whl", hash = "sha256:7e0610d205889588983836ec05dc098e80f97b7e7bbff7e994ebb78f578d0ddb", size = 2356 }, - { url = "https://files.pythonhosted.org/packages/b7/85/dabeaf902892922777492e1d253bb7e1264cadce3cea932f7ff599e53fea/tensorboard_data_server-0.7.2-py3-none-macosx_10_9_x86_64.whl", hash = "sha256:9fe5d24221b29625dbc7328b0436ca7fc1c23de4acf4d272f1180856e32f9f60", size = 4823598 }, - { url = "https://files.pythonhosted.org/packages/73/c6/825dab04195756cf8ff2e12698f22513b3db2f64925bdd41671bfb33aaa5/tensorboard_data_server-0.7.2-py3-none-manylinux_2_31_x86_64.whl", hash = "sha256:ef687163c24185ae9754ed5650eb5bc4d84ff257aabdc33f0cc6f74d8ba54530", size = 6590363 }, -] - -[[package]] -name = "text-unidecode" -version = "1.3" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/ab/e2/e9a00f0ccb71718418230718b3d900e71a5d16e701a3dae079a21e9cd8f8/text-unidecode-1.3.tar.gz", hash = "sha256:bad6603bb14d279193107714b288be206cac565dfa49aa5b105294dd5c4aab93", size = 76885 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/a6/a5/c0b6468d3824fe3fde30dbb5e1f687b291608f9473681bbf7dabbf5a87d7/text_unidecode-1.3-py2.py3-none-any.whl", hash = "sha256:1311f10e8b895935241623731c2ba64f4c455287888b18189350b67134a822e8", size = 78154 }, -] - -[[package]] -name = "threadpoolctl" -version = "3.6.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/b7/4d/08c89e34946fce2aec4fbb45c9016efd5f4d7f24af8e5d93296e935631d8/threadpoolctl-3.6.0.tar.gz", hash = "sha256:8ab8b4aa3491d812b623328249fab5302a68d2d71745c8a4c719a2fcaba9f44e", size = 21274 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/32/d5/f9a850d79b0851d1d4ef6456097579a9005b31fea68726a4ae5f2d82ddd9/threadpoolctl-3.6.0-py3-none-any.whl", hash = "sha256:43a0b8fd5a2928500110039e43a5eed8480b918967083ea48dc3ab9f13c4a7fb", size = 18638 }, -] - -[[package]] -name = "tokenizers" -version = "0.21.4" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "huggingface-hub" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/c2/2f/402986d0823f8d7ca139d969af2917fefaa9b947d1fb32f6168c509f2492/tokenizers-0.21.4.tar.gz", hash = "sha256:fa23f85fbc9a02ec5c6978da172cdcbac23498c3ca9f3645c5c68740ac007880", size = 351253 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/98/c6/fdb6f72bf6454f52eb4a2510be7fb0f614e541a2554d6210e370d85efff4/tokenizers-0.21.4-cp39-abi3-macosx_10_12_x86_64.whl", hash = "sha256:2ccc10a7c3bcefe0f242867dc914fc1226ee44321eb618cfe3019b5df3400133", size = 2863987 }, - { url = "https://files.pythonhosted.org/packages/8d/a6/28975479e35ddc751dc1ddc97b9b69bf7fcf074db31548aab37f8116674c/tokenizers-0.21.4-cp39-abi3-macosx_11_0_arm64.whl", hash = "sha256:5e2f601a8e0cd5be5cc7506b20a79112370b9b3e9cb5f13f68ab11acd6ca7d60", size = 2732457 }, - { url = "https://files.pythonhosted.org/packages/aa/8f/24f39d7b5c726b7b0be95dca04f344df278a3fe3a4deb15a975d194cbb32/tokenizers-0.21.4-cp39-abi3-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:39b376f5a1aee67b4d29032ee85511bbd1b99007ec735f7f35c8a2eb104eade5", size = 3012624 }, - { url = "https://files.pythonhosted.org/packages/58/47/26358925717687a58cb74d7a508de96649544fad5778f0cd9827398dc499/tokenizers-0.21.4-cp39-abi3-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:2107ad649e2cda4488d41dfd031469e9da3fcbfd6183e74e4958fa729ffbf9c6", size = 2939681 }, - { url = "https://files.pythonhosted.org/packages/99/6f/cc300fea5db2ab5ddc2c8aea5757a27b89c84469899710c3aeddc1d39801/tokenizers-0.21.4-cp39-abi3-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3c73012da95afafdf235ba80047699df4384fdc481527448a078ffd00e45a7d9", size = 3247445 }, - { url = "https://files.pythonhosted.org/packages/be/bf/98cb4b9c3c4afd8be89cfa6423704337dc20b73eb4180397a6e0d456c334/tokenizers-0.21.4-cp39-abi3-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:f23186c40395fc390d27f519679a58023f368a0aad234af145e0f39ad1212732", size = 3428014 }, - { url = "https://files.pythonhosted.org/packages/75/c7/96c1cc780e6ca7f01a57c13235dd05b7bc1c0f3588512ebe9d1331b5f5ae/tokenizers-0.21.4-cp39-abi3-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:cc88bb34e23a54cc42713d6d98af5f1bf79c07653d24fe984d2d695ba2c922a2", size = 3193197 }, - { url = "https://files.pythonhosted.org/packages/f2/90/273b6c7ec78af547694eddeea9e05de771278bd20476525ab930cecaf7d8/tokenizers-0.21.4-cp39-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:51b7eabb104f46c1c50b486520555715457ae833d5aee9ff6ae853d1130506ff", size = 3115426 }, - { url = "https://files.pythonhosted.org/packages/91/43/c640d5a07e95f1cf9d2c92501f20a25f179ac53a4f71e1489a3dcfcc67ee/tokenizers-0.21.4-cp39-abi3-musllinux_1_2_aarch64.whl", hash = "sha256:714b05b2e1af1288bd1bc56ce496c4cebb64a20d158ee802887757791191e6e2", size = 9089127 }, - { url = "https://files.pythonhosted.org/packages/44/a1/dd23edd6271d4dca788e5200a807b49ec3e6987815cd9d0a07ad9c96c7c2/tokenizers-0.21.4-cp39-abi3-musllinux_1_2_armv7l.whl", hash = "sha256:1340ff877ceedfa937544b7d79f5b7becf33a4cfb58f89b3b49927004ef66f78", size = 9055243 }, - { url = "https://files.pythonhosted.org/packages/21/2b/b410d6e9021c4b7ddb57248304dc817c4d4970b73b6ee343674914701197/tokenizers-0.21.4-cp39-abi3-musllinux_1_2_i686.whl", hash = "sha256:3c1f4317576e465ac9ef0d165b247825a2a4078bcd01cba6b54b867bdf9fdd8b", size = 9298237 }, - { url = "https://files.pythonhosted.org/packages/b7/0a/42348c995c67e2e6e5c89ffb9cfd68507cbaeb84ff39c49ee6e0a6dd0fd2/tokenizers-0.21.4-cp39-abi3-musllinux_1_2_x86_64.whl", hash = "sha256:c212aa4e45ec0bb5274b16b6f31dd3f1c41944025c2358faaa5782c754e84c24", size = 9461980 }, - { url = "https://files.pythonhosted.org/packages/3d/d3/dacccd834404cd71b5c334882f3ba40331ad2120e69ded32cf5fda9a7436/tokenizers-0.21.4-cp39-abi3-win32.whl", hash = "sha256:6c42a930bc5f4c47f4ea775c91de47d27910881902b0f20e4990ebe045a415d0", size = 2329871 }, - { url = "https://files.pythonhosted.org/packages/41/f2/fd673d979185f5dcbac4be7d09461cbb99751554ffb6718d0013af8604cb/tokenizers-0.21.4-cp39-abi3-win_amd64.whl", hash = "sha256:475d807a5c3eb72c59ad9b5fcdb254f6e17f53dfcbb9903233b0dfa9c943b597", size = 2507568 }, -] - -[[package]] -name = "toml" -version = "0.10.2" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/be/ba/1f744cdc819428fc6b5084ec34d9b30660f6f9daaf70eead706e3203ec3c/toml-0.10.2.tar.gz", hash = "sha256:b3bda1d108d5dd99f4a20d24d9c348e91c4db7ab1b749200bded2f839ccbe68f", size = 22253 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/44/6f/7120676b6d73228c96e17f1f794d8ab046fc910d781c8d151120c3f1569e/toml-0.10.2-py2.py3-none-any.whl", hash = "sha256:806143ae5bfb6a3c6e736a764057db0e6a0e05e338b5630894a5f779cabb4f9b", size = 16588 }, -] - -[[package]] -name = "tomli" -version = "2.3.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/52/ed/3f73f72945444548f33eba9a87fc7a6e969915e7b1acc8260b30e1f76a2f/tomli-2.3.0.tar.gz", hash = "sha256:64be704a875d2a59753d80ee8a533c3fe183e3f06807ff7dc2232938ccb01549", size = 17392 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/b3/2e/299f62b401438d5fe1624119c723f5d877acc86a4c2492da405626665f12/tomli-2.3.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:88bd15eb972f3664f5ed4b57c1634a97153b4bac4479dcb6a495f41921eb7f45", size = 153236 }, - { url = "https://files.pythonhosted.org/packages/86/7f/d8fffe6a7aefdb61bced88fcb5e280cfd71e08939da5894161bd71bea022/tomli-2.3.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:883b1c0d6398a6a9d29b508c331fa56adbcdff647f6ace4dfca0f50e90dfd0ba", size = 148084 }, - { url = "https://files.pythonhosted.org/packages/47/5c/24935fb6a2ee63e86d80e4d3b58b222dafaf438c416752c8b58537c8b89a/tomli-2.3.0-cp311-cp311-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:d1381caf13ab9f300e30dd8feadb3de072aeb86f1d34a8569453ff32a7dea4bf", size = 234832 }, - { url = "https://files.pythonhosted.org/packages/89/da/75dfd804fc11e6612846758a23f13271b76d577e299592b4371a4ca4cd09/tomli-2.3.0-cp311-cp311-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:a0e285d2649b78c0d9027570d4da3425bdb49830a6156121360b3f8511ea3441", size = 242052 }, - { url = "https://files.pythonhosted.org/packages/70/8c/f48ac899f7b3ca7eb13af73bacbc93aec37f9c954df3c08ad96991c8c373/tomli-2.3.0-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:0a154a9ae14bfcf5d8917a59b51ffd5a3ac1fd149b71b47a3a104ca4edcfa845", size = 239555 }, - { url = "https://files.pythonhosted.org/packages/ba/28/72f8afd73f1d0e7829bfc093f4cb98ce0a40ffc0cc997009ee1ed94ba705/tomli-2.3.0-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:74bf8464ff93e413514fefd2be591c3b0b23231a77f901db1eb30d6f712fc42c", size = 245128 }, - { url = "https://files.pythonhosted.org/packages/b6/eb/a7679c8ac85208706d27436e8d421dfa39d4c914dcf5fa8083a9305f58d9/tomli-2.3.0-cp311-cp311-win32.whl", hash = "sha256:00b5f5d95bbfc7d12f91ad8c593a1659b6387b43f054104cda404be6bda62456", size = 96445 }, - { url = "https://files.pythonhosted.org/packages/0a/fe/3d3420c4cb1ad9cb462fb52967080575f15898da97e21cb6f1361d505383/tomli-2.3.0-cp311-cp311-win_amd64.whl", hash = "sha256:4dc4ce8483a5d429ab602f111a93a6ab1ed425eae3122032db7e9acf449451be", size = 107165 }, - { url = "https://files.pythonhosted.org/packages/ff/b7/40f36368fcabc518bb11c8f06379a0fd631985046c038aca08c6d6a43c6e/tomli-2.3.0-cp312-cp312-macosx_10_13_x86_64.whl", hash = "sha256:d7d86942e56ded512a594786a5ba0a5e521d02529b3826e7761a05138341a2ac", size = 154891 }, - { url = "https://files.pythonhosted.org/packages/f9/3f/d9dd692199e3b3aab2e4e4dd948abd0f790d9ded8cd10cbaae276a898434/tomli-2.3.0-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:73ee0b47d4dad1c5e996e3cd33b8a76a50167ae5f96a2607cbe8cc773506ab22", size = 148796 }, - { url = "https://files.pythonhosted.org/packages/60/83/59bff4996c2cf9f9387a0f5a3394629c7efa5ef16142076a23a90f1955fa/tomli-2.3.0-cp312-cp312-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:792262b94d5d0a466afb5bc63c7daa9d75520110971ee269152083270998316f", size = 242121 }, - { url = "https://files.pythonhosted.org/packages/45/e5/7c5119ff39de8693d6baab6c0b6dcb556d192c165596e9fc231ea1052041/tomli-2.3.0-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:4f195fe57ecceac95a66a75ac24d9d5fbc98ef0962e09b2eddec5d39375aae52", size = 250070 }, - { url = "https://files.pythonhosted.org/packages/45/12/ad5126d3a278f27e6701abde51d342aa78d06e27ce2bb596a01f7709a5a2/tomli-2.3.0-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:e31d432427dcbf4d86958c184b9bfd1e96b5b71f8eb17e6d02531f434fd335b8", size = 245859 }, - { url = "https://files.pythonhosted.org/packages/fb/a1/4d6865da6a71c603cfe6ad0e6556c73c76548557a8d658f9e3b142df245f/tomli-2.3.0-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:7b0882799624980785240ab732537fcfc372601015c00f7fc367c55308c186f6", size = 250296 }, - { url = "https://files.pythonhosted.org/packages/a0/b7/a7a7042715d55c9ba6e8b196d65d2cb662578b4d8cd17d882d45322b0d78/tomli-2.3.0-cp312-cp312-win32.whl", hash = "sha256:ff72b71b5d10d22ecb084d345fc26f42b5143c5533db5e2eaba7d2d335358876", size = 97124 }, - { url = "https://files.pythonhosted.org/packages/06/1e/f22f100db15a68b520664eb3328fb0ae4e90530887928558112c8d1f4515/tomli-2.3.0-cp312-cp312-win_amd64.whl", hash = "sha256:1cb4ed918939151a03f33d4242ccd0aa5f11b3547d0cf30f7c74a408a5b99878", size = 107698 }, - { url = "https://files.pythonhosted.org/packages/89/48/06ee6eabe4fdd9ecd48bf488f4ac783844fd777f547b8d1b61c11939974e/tomli-2.3.0-cp313-cp313-macosx_10_13_x86_64.whl", hash = "sha256:5192f562738228945d7b13d4930baffda67b69425a7f0da96d360b0a3888136b", size = 154819 }, - { url = "https://files.pythonhosted.org/packages/f1/01/88793757d54d8937015c75dcdfb673c65471945f6be98e6a0410fba167ed/tomli-2.3.0-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:be71c93a63d738597996be9528f4abe628d1adf5e6eb11607bc8fe1a510b5dae", size = 148766 }, - { url = "https://files.pythonhosted.org/packages/42/17/5e2c956f0144b812e7e107f94f1cc54af734eb17b5191c0bbfb72de5e93e/tomli-2.3.0-cp313-cp313-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:c4665508bcbac83a31ff8ab08f424b665200c0e1e645d2bd9ab3d3e557b6185b", size = 240771 }, - { url = "https://files.pythonhosted.org/packages/d5/f4/0fbd014909748706c01d16824eadb0307115f9562a15cbb012cd9b3512c5/tomli-2.3.0-cp313-cp313-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:4021923f97266babc6ccab9f5068642a0095faa0a51a246a6a02fccbb3514eaf", size = 248586 }, - { url = "https://files.pythonhosted.org/packages/30/77/fed85e114bde5e81ecf9bc5da0cc69f2914b38f4708c80ae67d0c10180c5/tomli-2.3.0-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:a4ea38c40145a357d513bffad0ed869f13c1773716cf71ccaa83b0fa0cc4e42f", size = 244792 }, - { url = "https://files.pythonhosted.org/packages/55/92/afed3d497f7c186dc71e6ee6d4fcb0acfa5f7d0a1a2878f8beae379ae0cc/tomli-2.3.0-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:ad805ea85eda330dbad64c7ea7a4556259665bdf9d2672f5dccc740eb9d3ca05", size = 248909 }, - { url = "https://files.pythonhosted.org/packages/f8/84/ef50c51b5a9472e7265ce1ffc7f24cd4023d289e109f669bdb1553f6a7c2/tomli-2.3.0-cp313-cp313-win32.whl", hash = "sha256:97d5eec30149fd3294270e889b4234023f2c69747e555a27bd708828353ab606", size = 96946 }, - { url = "https://files.pythonhosted.org/packages/b2/b7/718cd1da0884f281f95ccfa3a6cc572d30053cba64603f79d431d3c9b61b/tomli-2.3.0-cp313-cp313-win_amd64.whl", hash = "sha256:0c95ca56fbe89e065c6ead5b593ee64b84a26fca063b5d71a1122bf26e533999", size = 107705 }, - { url = "https://files.pythonhosted.org/packages/19/94/aeafa14a52e16163008060506fcb6aa1949d13548d13752171a755c65611/tomli-2.3.0-cp314-cp314-macosx_10_13_x86_64.whl", hash = "sha256:cebc6fe843e0733ee827a282aca4999b596241195f43b4cc371d64fc6639da9e", size = 154244 }, - { url = "https://files.pythonhosted.org/packages/db/e4/1e58409aa78eefa47ccd19779fc6f36787edbe7d4cd330eeeedb33a4515b/tomli-2.3.0-cp314-cp314-macosx_11_0_arm64.whl", hash = "sha256:4c2ef0244c75aba9355561272009d934953817c49f47d768070c3c94355c2aa3", size = 148637 }, - { url = "https://files.pythonhosted.org/packages/26/b6/d1eccb62f665e44359226811064596dd6a366ea1f985839c566cd61525ae/tomli-2.3.0-cp314-cp314-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:c22a8bf253bacc0cf11f35ad9808b6cb75ada2631c2d97c971122583b129afbc", size = 241925 }, - { url = "https://files.pythonhosted.org/packages/70/91/7cdab9a03e6d3d2bb11beae108da5bdc1c34bdeb06e21163482544ddcc90/tomli-2.3.0-cp314-cp314-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:0eea8cc5c5e9f89c9b90c4896a8deefc74f518db5927d0e0e8d4a80953d774d0", size = 249045 }, - { url = "https://files.pythonhosted.org/packages/15/1b/8c26874ed1f6e4f1fcfeb868db8a794cbe9f227299402db58cfcc858766c/tomli-2.3.0-cp314-cp314-musllinux_1_2_aarch64.whl", hash = "sha256:b74a0e59ec5d15127acdabd75ea17726ac4c5178ae51b85bfe39c4f8a278e879", size = 245835 }, - { url = "https://files.pythonhosted.org/packages/fd/42/8e3c6a9a4b1a1360c1a2a39f0b972cef2cc9ebd56025168c4137192a9321/tomli-2.3.0-cp314-cp314-musllinux_1_2_x86_64.whl", hash = "sha256:b5870b50c9db823c595983571d1296a6ff3e1b88f734a4c8f6fc6188397de005", size = 253109 }, - { url = "https://files.pythonhosted.org/packages/22/0c/b4da635000a71b5f80130937eeac12e686eefb376b8dee113b4a582bba42/tomli-2.3.0-cp314-cp314-win32.whl", hash = "sha256:feb0dacc61170ed7ab602d3d972a58f14ee3ee60494292d384649a3dc38ef463", size = 97930 }, - { url = "https://files.pythonhosted.org/packages/b9/74/cb1abc870a418ae99cd5c9547d6bce30701a954e0e721821df483ef7223c/tomli-2.3.0-cp314-cp314-win_amd64.whl", hash = "sha256:b273fcbd7fc64dc3600c098e39136522650c49bca95df2d11cf3b626422392c8", size = 107964 }, - { url = "https://files.pythonhosted.org/packages/54/78/5c46fff6432a712af9f792944f4fcd7067d8823157949f4e40c56b8b3c83/tomli-2.3.0-cp314-cp314t-macosx_10_13_x86_64.whl", hash = "sha256:940d56ee0410fa17ee1f12b817b37a4d4e4dc4d27340863cc67236c74f582e77", size = 163065 }, - { url = "https://files.pythonhosted.org/packages/39/67/f85d9bd23182f45eca8939cd2bc7050e1f90c41f4a2ecbbd5963a1d1c486/tomli-2.3.0-cp314-cp314t-macosx_11_0_arm64.whl", hash = "sha256:f85209946d1fe94416debbb88d00eb92ce9cd5266775424ff81bc959e001acaf", size = 159088 }, - { url = "https://files.pythonhosted.org/packages/26/5a/4b546a0405b9cc0659b399f12b6adb750757baf04250b148d3c5059fc4eb/tomli-2.3.0-cp314-cp314t-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:a56212bdcce682e56b0aaf79e869ba5d15a6163f88d5451cbde388d48b13f530", size = 268193 }, - { url = "https://files.pythonhosted.org/packages/42/4f/2c12a72ae22cf7b59a7fe75b3465b7aba40ea9145d026ba41cb382075b0e/tomli-2.3.0-cp314-cp314t-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:c5f3ffd1e098dfc032d4d3af5c0ac64f6d286d98bc148698356847b80fa4de1b", size = 275488 }, - { url = "https://files.pythonhosted.org/packages/92/04/a038d65dbe160c3aa5a624e93ad98111090f6804027d474ba9c37c8ae186/tomli-2.3.0-cp314-cp314t-musllinux_1_2_aarch64.whl", hash = "sha256:5e01decd096b1530d97d5d85cb4dff4af2d8347bd35686654a004f8dea20fc67", size = 272669 }, - { url = "https://files.pythonhosted.org/packages/be/2f/8b7c60a9d1612a7cbc39ffcca4f21a73bf368a80fc25bccf8253e2563267/tomli-2.3.0-cp314-cp314t-musllinux_1_2_x86_64.whl", hash = "sha256:8a35dd0e643bb2610f156cca8db95d213a90015c11fee76c946aa62b7ae7e02f", size = 279709 }, - { url = "https://files.pythonhosted.org/packages/7e/46/cc36c679f09f27ded940281c38607716c86cf8ba4a518d524e349c8b4874/tomli-2.3.0-cp314-cp314t-win32.whl", hash = "sha256:a1f7f282fe248311650081faafa5f4732bdbfef5d45fe3f2e702fbc6f2d496e0", size = 107563 }, - { url = "https://files.pythonhosted.org/packages/84/ff/426ca8683cf7b753614480484f6437f568fd2fda2edbdf57a2d3d8b27a0b/tomli-2.3.0-cp314-cp314t-win_amd64.whl", hash = "sha256:70a251f8d4ba2d9ac2542eecf008b3c8a9fc5c3f9f02c56a9d7952612be2fdba", size = 119756 }, - { url = "https://files.pythonhosted.org/packages/77/b8/0135fadc89e73be292b473cb820b4f5a08197779206b33191e801feeae40/tomli-2.3.0-py3-none-any.whl", hash = "sha256:e95b1af3c5b07d9e643909b5abbec77cd9f1217e6d0bca72b0234736b9fb1f1b", size = 14408 }, -] - -[[package]] -name = "toolz" -version = "1.1.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/11/d6/114b492226588d6ff54579d95847662fc69196bdeec318eb45393b24c192/toolz-1.1.0.tar.gz", hash = "sha256:27a5c770d068c110d9ed9323f24f1543e83b2f300a687b7891c1a6d56b697b5b", size = 52613 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/fb/12/5911ae3eeec47800503a238d971e51722ccea5feb8569b735184d5fcdbc0/toolz-1.1.0-py3-none-any.whl", hash = "sha256:15ccc861ac51c53696de0a5d6d4607f99c210739caf987b5d2054f3efed429d8", size = 58093 }, -] - -[[package]] -name = "torch" -version = "2.9.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "filelock" }, - { name = "fsspec" }, - { name = "jinja2" }, - { name = "networkx", version = "3.4.2", source = { registry = "https://pypi.org/simple" }, marker = "python_full_version < '3.11'" }, - { name = "networkx", version = "3.6.1", source = { registry = "https://pypi.org/simple" }, marker = "python_full_version >= '3.11'" }, - { name = "nvidia-cublas-cu12", marker = "platform_machine == 'x86_64' and sys_platform == 'linux'" }, - { name = "nvidia-cuda-cupti-cu12", marker = "platform_machine == 'x86_64' and sys_platform == 'linux'" }, - { name = "nvidia-cuda-nvrtc-cu12", marker = "platform_machine == 'x86_64' and sys_platform == 'linux'" }, - { name = "nvidia-cuda-runtime-cu12", marker = "platform_machine == 'x86_64' and sys_platform == 'linux'" }, - { name = "nvidia-cudnn-cu12", marker = "platform_machine == 'x86_64' and sys_platform == 'linux'" }, - { name = "nvidia-cufft-cu12", marker = "platform_machine == 'x86_64' and sys_platform == 'linux'" }, - { name = "nvidia-cufile-cu12", marker = "platform_machine == 'x86_64' and sys_platform == 'linux'" }, - { name = "nvidia-curand-cu12", marker = "platform_machine == 'x86_64' and sys_platform == 'linux'" }, - { name = "nvidia-cusolver-cu12", marker = "platform_machine == 'x86_64' and sys_platform == 'linux'" }, - { name = "nvidia-cusparse-cu12", marker = "platform_machine == 'x86_64' and sys_platform == 'linux'" }, - { name = "nvidia-cusparselt-cu12", marker = "platform_machine == 'x86_64' and sys_platform == 'linux'" }, - { name = "nvidia-nccl-cu12", marker = "platform_machine == 'x86_64' and sys_platform == 'linux'" }, - { name = "nvidia-nvjitlink-cu12", marker = "platform_machine == 'x86_64' and sys_platform == 'linux'" }, - { name = "nvidia-nvshmem-cu12", marker = "platform_machine == 'x86_64' and sys_platform == 'linux'" }, - { name = "nvidia-nvtx-cu12", marker = "platform_machine == 'x86_64' and sys_platform == 'linux'" }, - { name = "setuptools", marker = "python_full_version >= '3.12'" }, - { name = "sympy" }, - { name = "triton", marker = "platform_machine == 'x86_64' and sys_platform == 'linux'" }, - { name = "typing-extensions" }, -] -wheels = [ - { url = "https://files.pythonhosted.org/packages/5f/56/9577683b23072075ed2e40d725c52c2019d71a972fab8e083763da8e707e/torch-2.9.1-cp310-cp310-manylinux_2_28_aarch64.whl", hash = "sha256:1cc208435f6c379f9b8fdfd5ceb5be1e3b72a6bdf1cb46c0d2812aa73472db9e", size = 104207681 }, - { url = "https://files.pythonhosted.org/packages/38/45/be5a74f221df8f4b609b78ff79dc789b0cc9017624544ac4dd1c03973150/torch-2.9.1-cp310-cp310-manylinux_2_28_x86_64.whl", hash = "sha256:9fd35c68b3679378c11f5eb73220fdcb4e6f4592295277fbb657d31fd053237c", size = 899794036 }, - { url = "https://files.pythonhosted.org/packages/67/95/a581e8a382596b69385a44bab2733f1273d45c842f5d4a504c0edc3133b6/torch-2.9.1-cp310-cp310-win_amd64.whl", hash = "sha256:2af70e3be4a13becba4655d6cc07dcfec7ae844db6ac38d6c1dafeb245d17d65", size = 110969861 }, - { url = "https://files.pythonhosted.org/packages/ad/51/1756dc128d2bf6ea4e0a915cb89ea5e730315ff33d60c1ff56fd626ba3eb/torch-2.9.1-cp310-none-macosx_11_0_arm64.whl", hash = "sha256:a83b0e84cc375e3318a808d032510dde99d696a85fe9473fc8575612b63ae951", size = 74452222 }, - { url = "https://files.pythonhosted.org/packages/15/db/c064112ac0089af3d2f7a2b5bfbabf4aa407a78b74f87889e524b91c5402/torch-2.9.1-cp311-cp311-manylinux_2_28_aarch64.whl", hash = "sha256:62b3fd888277946918cba4478cf849303da5359f0fb4e3bfb86b0533ba2eaf8d", size = 104220430 }, - { url = "https://files.pythonhosted.org/packages/56/be/76eaa36c9cd032d3b01b001e2c5a05943df75f26211f68fae79e62f87734/torch-2.9.1-cp311-cp311-manylinux_2_28_x86_64.whl", hash = "sha256:d033ff0ac3f5400df862a51bdde9bad83561f3739ea0046e68f5401ebfa67c1b", size = 899821446 }, - { url = "https://files.pythonhosted.org/packages/47/cc/7a2949e38dfe3244c4df21f0e1c27bce8aedd6c604a587dd44fc21017cb4/torch-2.9.1-cp311-cp311-win_amd64.whl", hash = "sha256:0d06b30a9207b7c3516a9e0102114024755a07045f0c1d2f2a56b1819ac06bcb", size = 110973074 }, - { url = "https://files.pythonhosted.org/packages/1e/ce/7d251155a783fb2c1bb6837b2b7023c622a2070a0a72726ca1df47e7ea34/torch-2.9.1-cp311-none-macosx_11_0_arm64.whl", hash = "sha256:52347912d868653e1528b47cafaf79b285b98be3f4f35d5955389b1b95224475", size = 74463887 }, - { url = "https://files.pythonhosted.org/packages/0f/27/07c645c7673e73e53ded71705045d6cb5bae94c4b021b03aa8d03eee90ab/torch-2.9.1-cp312-cp312-manylinux_2_28_aarch64.whl", hash = "sha256:da5f6f4d7f4940a173e5572791af238cb0b9e21b1aab592bd8b26da4c99f1cd6", size = 104126592 }, - { url = "https://files.pythonhosted.org/packages/19/17/e377a460603132b00760511299fceba4102bd95db1a0ee788da21298ccff/torch-2.9.1-cp312-cp312-manylinux_2_28_x86_64.whl", hash = "sha256:27331cd902fb4322252657f3902adf1c4f6acad9dcad81d8df3ae14c7c4f07c4", size = 899742281 }, - { url = "https://files.pythonhosted.org/packages/b1/1a/64f5769025db846a82567fa5b7d21dba4558a7234ee631712ee4771c436c/torch-2.9.1-cp312-cp312-win_amd64.whl", hash = "sha256:81a285002d7b8cfd3fdf1b98aa8df138d41f1a8334fd9ea37511517cedf43083", size = 110940568 }, - { url = "https://files.pythonhosted.org/packages/6e/ab/07739fd776618e5882661d04c43f5b5586323e2f6a2d7d84aac20d8f20bd/torch-2.9.1-cp312-none-macosx_11_0_arm64.whl", hash = "sha256:c0d25d1d8e531b8343bea0ed811d5d528958f1dcbd37e7245bc686273177ad7e", size = 74479191 }, - { url = "https://files.pythonhosted.org/packages/20/60/8fc5e828d050bddfab469b3fe78e5ab9a7e53dda9c3bdc6a43d17ce99e63/torch-2.9.1-cp313-cp313-manylinux_2_28_aarch64.whl", hash = "sha256:c29455d2b910b98738131990394da3e50eea8291dfeb4b12de71ecf1fdeb21cb", size = 104135743 }, - { url = "https://files.pythonhosted.org/packages/f2/b7/6d3f80e6918213babddb2a37b46dbb14c15b14c5f473e347869a51f40e1f/torch-2.9.1-cp313-cp313-manylinux_2_28_x86_64.whl", hash = "sha256:524de44cd13931208ba2c4bde9ec7741fd4ae6bfd06409a604fc32f6520c2bc9", size = 899749493 }, - { url = "https://files.pythonhosted.org/packages/a6/47/c7843d69d6de8938c1cbb1eba426b1d48ddf375f101473d3e31a5fc52b74/torch-2.9.1-cp313-cp313-win_amd64.whl", hash = "sha256:545844cc16b3f91e08ce3b40e9c2d77012dd33a48d505aed34b7740ed627a1b2", size = 110944162 }, - { url = "https://files.pythonhosted.org/packages/28/0e/2a37247957e72c12151b33a01e4df651d9d155dd74d8cfcbfad15a79b44a/torch-2.9.1-cp313-cp313t-macosx_11_0_arm64.whl", hash = "sha256:5be4bf7496f1e3ffb1dd44b672adb1ac3f081f204c5ca81eba6442f5f634df8e", size = 74830751 }, - { url = "https://files.pythonhosted.org/packages/4b/f7/7a18745edcd7b9ca2381aa03353647bca8aace91683c4975f19ac233809d/torch-2.9.1-cp313-cp313t-manylinux_2_28_aarch64.whl", hash = "sha256:30a3e170a84894f3652434b56d59a64a2c11366b0ed5776fab33c2439396bf9a", size = 104142929 }, - { url = "https://files.pythonhosted.org/packages/f4/dd/f1c0d879f2863ef209e18823a988dc7a1bf40470750e3ebe927efdb9407f/torch-2.9.1-cp313-cp313t-manylinux_2_28_x86_64.whl", hash = "sha256:8301a7b431e51764629208d0edaa4f9e4c33e6df0f2f90b90e261d623df6a4e2", size = 899748978 }, - { url = "https://files.pythonhosted.org/packages/1f/9f/6986b83a53b4d043e36f3f898b798ab51f7f20fdf1a9b01a2720f445043d/torch-2.9.1-cp313-cp313t-win_amd64.whl", hash = "sha256:2e1c42c0ae92bf803a4b2409fdfed85e30f9027a66887f5e7dcdbc014c7531db", size = 111176995 }, - { url = "https://files.pythonhosted.org/packages/40/60/71c698b466dd01e65d0e9514b5405faae200c52a76901baf6906856f17e4/torch-2.9.1-cp313-none-macosx_11_0_arm64.whl", hash = "sha256:2c14b3da5df416cf9cb5efab83aa3056f5b8cd8620b8fde81b4987ecab730587", size = 74480347 }, - { url = "https://files.pythonhosted.org/packages/48/50/c4b5112546d0d13cc9eaa1c732b823d676a9f49ae8b6f97772f795874a03/torch-2.9.1-cp314-cp314-macosx_11_0_arm64.whl", hash = "sha256:1edee27a7c9897f4e0b7c14cfc2f3008c571921134522d5b9b5ec4ebbc69041a", size = 74433245 }, - { url = "https://files.pythonhosted.org/packages/81/c9/2628f408f0518b3bae49c95f5af3728b6ab498c8624ab1e03a43dd53d650/torch-2.9.1-cp314-cp314-manylinux_2_28_aarch64.whl", hash = "sha256:19d144d6b3e29921f1fc70503e9f2fc572cde6a5115c0c0de2f7ca8b1483e8b6", size = 104134804 }, - { url = "https://files.pythonhosted.org/packages/28/fc/5bc91d6d831ae41bf6e9e6da6468f25330522e92347c9156eb3f1cb95956/torch-2.9.1-cp314-cp314-manylinux_2_28_x86_64.whl", hash = "sha256:c432d04376f6d9767a9852ea0def7b47a7bbc8e7af3b16ac9cf9ce02b12851c9", size = 899747132 }, - { url = "https://files.pythonhosted.org/packages/63/5d/e8d4e009e52b6b2cf1684bde2a6be157b96fb873732542fb2a9a99e85a83/torch-2.9.1-cp314-cp314-win_amd64.whl", hash = "sha256:d187566a2cdc726fc80138c3cdb260970fab1c27e99f85452721f7759bbd554d", size = 110934845 }, - { url = "https://files.pythonhosted.org/packages/bd/b2/2d15a52516b2ea3f414643b8de68fa4cb220d3877ac8b1028c83dc8ca1c4/torch-2.9.1-cp314-cp314t-macosx_11_0_arm64.whl", hash = "sha256:cb10896a1f7fedaddbccc2017ce6ca9ecaaf990f0973bdfcf405439750118d2c", size = 74823558 }, - { url = "https://files.pythonhosted.org/packages/86/5c/5b2e5d84f5b9850cd1e71af07524d8cbb74cba19379800f1f9f7c997fc70/torch-2.9.1-cp314-cp314t-manylinux_2_28_aarch64.whl", hash = "sha256:0a2bd769944991c74acf0c4ef23603b9c777fdf7637f115605a4b2d8023110c7", size = 104145788 }, - { url = "https://files.pythonhosted.org/packages/a9/8c/3da60787bcf70add986c4ad485993026ac0ca74f2fc21410bc4eb1bb7695/torch-2.9.1-cp314-cp314t-manylinux_2_28_x86_64.whl", hash = "sha256:07c8a9660bc9414c39cac530ac83b1fb1b679d7155824144a40a54f4a47bfa73", size = 899735500 }, - { url = "https://files.pythonhosted.org/packages/db/2b/f7818f6ec88758dfd21da46b6cd46af9d1b3433e53ddbb19ad1e0da17f9b/torch-2.9.1-cp314-cp314t-win_amd64.whl", hash = "sha256:c88d3299ddeb2b35dcc31753305612db485ab6f1823e37fb29451c8b2732b87e", size = 111163659 }, -] - -[[package]] -name = "torchmetrics" -version = "1.8.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "lightning-utilities" }, - { name = "numpy" }, - { name = "packaging" }, - { name = "torch" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/85/2e/48a887a59ecc4a10ce9e8b35b3e3c5cef29d902c4eac143378526e7485cb/torchmetrics-1.8.2.tar.gz", hash = "sha256:cf64a901036bf107f17a524009eea7781c9c5315d130713aeca5747a686fe7a5", size = 580679 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/02/21/aa0f434434c48490f91b65962b1ce863fdcce63febc166ca9fe9d706c2b6/torchmetrics-1.8.2-py3-none-any.whl", hash = "sha256:08382fd96b923e39e904c4d570f3d49e2cc71ccabd2a94e0f895d1f0dac86242", size = 983161 }, -] - -[[package]] -name = "tqdm" -version = "4.67.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "colorama", marker = "sys_platform == 'win32'" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/a8/4b/29b4ef32e036bb34e4ab51796dd745cdba7ed47ad142a9f4a1eb8e0c744d/tqdm-4.67.1.tar.gz", hash = "sha256:f8aef9c52c08c13a65f30ea34f4e5aac3fd1a34959879d7e59e63027286627f2", size = 169737 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/d0/30/dc54f88dd4a2b5dc8a0279bdd7270e735851848b762aeb1c1184ed1f6b14/tqdm-4.67.1-py3-none-any.whl", hash = "sha256:26445eca388f82e72884e0d580d5464cd801a3ea01e63e5601bdff9ba6a48de2", size = 78540 }, -] - -[[package]] -name = "traitlets" -version = "5.14.3" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/eb/79/72064e6a701c2183016abbbfedaba506d81e30e232a68c9f0d6f6fcd1574/traitlets-5.14.3.tar.gz", hash = "sha256:9ed0579d3502c94b4b3732ac120375cda96f923114522847de4b3bb98b96b6b7", size = 161621 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/00/c0/8f5d070730d7836adc9c9b6408dec68c6ced86b304a9b26a14df072a6e8c/traitlets-5.14.3-py3-none-any.whl", hash = "sha256:b74e89e397b1ed28cc831db7aea759ba6640cb3de13090ca145426688ff1ac4f", size = 85359 }, -] - -[[package]] -name = "transformers" -version = "4.53.3" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "filelock" }, - { name = "huggingface-hub" }, - { name = "numpy" }, - { name = "packaging" }, - { name = "pyyaml" }, - { name = "regex" }, - { name = "requests" }, - { name = "safetensors" }, - { name = "tokenizers" }, - { name = "tqdm" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/f1/5c/49182918b58eaa0b4c954fd0e37c79fc299e5643e69d70089d0b0eb0cd9b/transformers-4.53.3.tar.gz", hash = "sha256:b2eda1a261de79b78b97f7888fe2005fc0c3fabf5dad33d52cc02983f9f675d8", size = 9197478 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/41/b1/d7520cc5cb69c825599042eb3a7c986fa9baa8a8d2dea9acd78e152c81e2/transformers-4.53.3-py3-none-any.whl", hash = "sha256:5aba81c92095806b6baf12df35d756cf23b66c356975fb2a7fa9e536138d7c75", size = 10826382 }, -] - -[[package]] -name = "triton" -version = "3.5.1" -source = { registry = "https://pypi.org/simple" } -wheels = [ - { url = "https://files.pythonhosted.org/packages/fd/6e/676ab5019b4dde8b9b7bab71245102fc02778ef3df48218b298686b9ffd6/triton-3.5.1-cp310-cp310-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:5fc53d849f879911ea13f4a877243afc513187bc7ee92d1f2c0f1ba3169e3c94", size = 170320692 }, - { url = "https://files.pythonhosted.org/packages/b0/72/ec90c3519eaf168f22cb1757ad412f3a2add4782ad3a92861c9ad135d886/triton-3.5.1-cp311-cp311-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:61413522a48add32302353fdbaaf92daaaab06f6b5e3229940d21b5207f47579", size = 170425802 }, - { url = "https://files.pythonhosted.org/packages/f2/50/9a8358d3ef58162c0a415d173cfb45b67de60176e1024f71fbc4d24c0b6d/triton-3.5.1-cp312-cp312-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:d2c6b915a03888ab931a9fd3e55ba36785e1fe70cbea0b40c6ef93b20fc85232", size = 170470207 }, - { url = "https://files.pythonhosted.org/packages/27/46/8c3bbb5b0a19313f50edcaa363b599e5a1a5ac9683ead82b9b80fe497c8d/triton-3.5.1-cp313-cp313-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:f3f4346b6ebbd4fad18773f5ba839114f4826037c9f2f34e0148894cd5dd3dba", size = 170470410 }, - { url = "https://files.pythonhosted.org/packages/37/92/e97fcc6b2c27cdb87ce5ee063d77f8f26f19f06916aa680464c8104ef0f6/triton-3.5.1-cp313-cp313t-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:0b4d2c70127fca6a23e247f9348b8adde979d2e7a20391bfbabaac6aebc7e6a8", size = 170579924 }, - { url = "https://files.pythonhosted.org/packages/a4/e6/c595c35e5c50c4bc56a7bac96493dad321e9e29b953b526bbbe20f9911d0/triton-3.5.1-cp314-cp314-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:d0637b1efb1db599a8e9dc960d53ab6e4637db7d4ab6630a0974705d77b14b60", size = 170480488 }, - { url = "https://files.pythonhosted.org/packages/16/b5/b0d3d8b901b6a04ca38df5e24c27e53afb15b93624d7fd7d658c7cd9352a/triton-3.5.1-cp314-cp314t-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:bac7f7d959ad0f48c0e97d6643a1cc0fd5786fe61cb1f83b537c6b2d54776478", size = 170582192 }, -] - -[[package]] -name = "typeguard" -version = "4.4.4" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/c7/68/71c1a15b5f65f40e91b65da23b8224dad41349894535a97f63a52e462196/typeguard-4.4.4.tar.gz", hash = "sha256:3a7fd2dffb705d4d0efaed4306a704c89b9dee850b688f060a8b1615a79e5f74", size = 75203 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/1b/a9/e3aee762739c1d7528da1c3e06d518503f8b6c439c35549b53735ba52ead/typeguard-4.4.4-py3-none-any.whl", hash = "sha256:b5f562281b6bfa1f5492470464730ef001646128b180769880468bd84b68b09e", size = 34874 }, -] - -[[package]] -name = "typer" -version = "0.21.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "click" }, - { name = "rich" }, - { name = "shellingham" }, - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/85/30/ff9ede605e3bd086b4dd842499814e128500621f7951ca1e5ce84bbf61b1/typer-0.21.0.tar.gz", hash = "sha256:c87c0d2b6eee3b49c5c64649ec92425492c14488096dfbc8a0c2799b2f6f9c53", size = 106781 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/e1/e4/5ebc1899d31d2b1601b32d21cfb4bba022ae6fce323d365f0448031b1660/typer-0.21.0-py3-none-any.whl", hash = "sha256:c79c01ca6b30af9fd48284058a7056ba0d3bf5cf10d0ff3d0c5b11b68c258ac6", size = 47109 }, -] - -[[package]] -name = "typing-extensions" -version = "4.15.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/72/94/1a15dd82efb362ac84269196e94cf00f187f7ed21c242792a923cdb1c61f/typing_extensions-4.15.0.tar.gz", hash = "sha256:0cea48d173cc12fa28ecabc3b837ea3cf6f38c6d1136f85cbaaf598984861466", size = 109391 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/18/67/36e9267722cc04a6b9f15c7f3441c2363321a3ea07da7ae0c0707beb2a9c/typing_extensions-4.15.0-py3-none-any.whl", hash = "sha256:f0fa19c6845758ab08074a0cfa8b7aecb71c999ca73d62883bc25cc018c4e548", size = 44614 }, -] - -[[package]] -name = "typing-inspection" -version = "0.4.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/55/e3/70399cb7dd41c10ac53367ae42139cf4b1ca5f36bb3dc6c9d33acdb43655/typing_inspection-0.4.2.tar.gz", hash = "sha256:ba561c48a67c5958007083d386c3295464928b01faa735ab8547c5692e87f464", size = 75949 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/dc/9b/47798a6c91d8bdb567fe2698fe81e0c6b7cb7ef4d13da4114b41d239f65d/typing_inspection-0.4.2-py3-none-any.whl", hash = "sha256:4ed1cacbdc298c220f1bd249ed5287caa16f34d44ef4e9c3d0cbad5b521545e7", size = 14611 }, -] - -[[package]] -name = "tzdata" -version = "2025.3" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/5e/a7/c202b344c5ca7daf398f3b8a477eeb205cf3b6f32e7ec3a6bac0629ca975/tzdata-2025.3.tar.gz", hash = "sha256:de39c2ca5dc7b0344f2eba86f49d614019d29f060fc4ebc8a417896a620b56a7", size = 196772 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/c7/b0/003792df09decd6849a5e39c28b513c06e84436a54440380862b5aeff25d/tzdata-2025.3-py2.py3-none-any.whl", hash = "sha256:06a47e5700f3081aab02b2e513160914ff0694bce9947d6b76ebd6bf57cfc5d1", size = 348521 }, -] - -[[package]] -name = "urllib3" -version = "2.6.2" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/1e/24/a2a2ed9addd907787d7aa0355ba36a6cadf1768b934c652ea78acbd59dcd/urllib3-2.6.2.tar.gz", hash = "sha256:016f9c98bb7e98085cb2b4b17b87d2c702975664e4f060c6532e64d1c1a5e797", size = 432930 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/6d/b9/4095b668ea3678bf6a0af005527f39de12fb026516fb3df17495a733b7f8/urllib3-2.6.2-py3-none-any.whl", hash = "sha256:ec21cddfe7724fc7cb4ba4bea7aa8e2ef36f607a4bab81aa6ce42a13dc3f03dd", size = 131182 }, -] - -[[package]] -name = "wandb" -version = "0.23.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "click" }, - { name = "gitpython" }, - { name = "packaging" }, - { name = "platformdirs" }, - { name = "protobuf" }, - { name = "pydantic" }, - { name = "pyyaml" }, - { name = "requests" }, - { name = "sentry-sdk" }, - { name = "typing-extensions" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/0a/cc/770ae3aa7ae44f6792f7ecb81c14c0e38b672deb35235719bb1006519487/wandb-0.23.1.tar.gz", hash = "sha256:f6fb1e3717949b29675a69359de0eeb01e67d3360d581947d5b3f98c273567d6", size = 44298053 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/12/0b/c3d7053dfd93fd259a63c7818d9c4ac2ba0642ff8dc8db98662ea0cf9cc0/wandb-0.23.1-py3-none-macosx_12_0_arm64.whl", hash = "sha256:358e15471d19b7d73fc464e37371c19d44d39e433252ac24df107aff993a286b", size = 21527293 }, - { url = "https://files.pythonhosted.org/packages/ee/9f/059420fa0cb6c511dc5c5a50184122b6aca7b178cb2aa210139e354020da/wandb-0.23.1-py3-none-macosx_12_0_x86_64.whl", hash = "sha256:110304407f4b38f163bdd50ed5c5225365e4df3092f13089c30171a75257b575", size = 22745926 }, - { url = "https://files.pythonhosted.org/packages/96/b6/fd465827c14c64d056d30b4c9fcf4dac889a6969dba64489a88fc4ffa333/wandb-0.23.1-py3-none-manylinux_2_28_aarch64.whl", hash = "sha256:6cc984cf85feb2f8ee0451d76bc9fb7f39da94956bb8183e30d26284cf203b65", size = 21212973 }, - { url = "https://files.pythonhosted.org/packages/5c/ee/9a8bb9a39cc1f09c3060456cc79565110226dc4099a719af5c63432da21d/wandb-0.23.1-py3-none-manylinux_2_28_x86_64.whl", hash = "sha256:67431cd3168d79fdb803e503bd669c577872ffd5dadfa86de733b3274b93088e", size = 22887885 }, - { url = "https://files.pythonhosted.org/packages/6d/4d/8d9e75add529142e037b05819cb3ab1005679272950128d69d218b7e5b2e/wandb-0.23.1-py3-none-musllinux_1_2_aarch64.whl", hash = "sha256:07be70c0baa97ea25fadc4a9d0097f7371eef6dcacc5ceb525c82491a31e9244", size = 21250967 }, - { url = "https://files.pythonhosted.org/packages/97/72/0b35cddc4e4168f03c759b96d9f671ad18aec8bdfdd84adfea7ecb3f5701/wandb-0.23.1-py3-none-musllinux_1_2_x86_64.whl", hash = "sha256:216c95b08e0a2ec6a6008373b056d597573d565e30b43a7a93c35a171485ee26", size = 22988382 }, - { url = "https://files.pythonhosted.org/packages/c0/6d/e78093d49d68afb26f5261a70fc7877c34c114af5c2ee0ab3b1af85f5e76/wandb-0.23.1-py3-none-win32.whl", hash = "sha256:fb5cf0f85692f758a5c36ab65fea96a1284126de64e836610f92ddbb26df5ded", size = 22150756 }, - { url = "https://files.pythonhosted.org/packages/05/27/4f13454b44c9eceaac3d6e4e4efa2230b6712d613ff9bf7df010eef4fd18/wandb-0.23.1-py3-none-win_amd64.whl", hash = "sha256:21c8c56e436eb707b7d54f705652e030d48e5cfcba24cf953823eb652e30e714", size = 22150760 }, - { url = "https://files.pythonhosted.org/packages/30/20/6c091d451e2a07689bfbfaeb7592d488011420e721de170884fedd68c644/wandb-0.23.1-py3-none-win_arm64.whl", hash = "sha256:8aee7f3bb573f2c0acf860f497ca9c684f9b35f2ca51011ba65af3d4592b77c1", size = 20137463 }, -] - -[[package]] -name = "wcwidth" -version = "0.2.14" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/24/30/6b0809f4510673dc723187aeaf24c7f5459922d01e2f794277a3dfb90345/wcwidth-0.2.14.tar.gz", hash = "sha256:4d478375d31bc5395a3c55c40ccdf3354688364cd61c4f6adacaa9215d0b3605", size = 102293 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/af/b5/123f13c975e9f27ab9c0770f514345bd406d0e8d3b7a0723af9d43f710af/wcwidth-0.2.14-py2.py3-none-any.whl", hash = "sha256:a7bb560c8aee30f9957e5f9895805edd20602f2d7f720186dfd906e82b4982e1", size = 37286 }, -] - -[[package]] -name = "webdataset" -version = "1.0.2" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "braceexpand" }, - { name = "numpy" }, - { name = "pyyaml" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/5a/3a/68800d92e065cf4750ebecf973b13979c0c929b439e1293012938862038d/webdataset-1.0.2.tar.gz", hash = "sha256:7f0498be827cfa46cc5430a58768a24e2c6a410676a61be1838f53d61afdaab4", size = 80090 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/d9/00/aca6beb3658dab4ed3dbb41a78e6e7f31342e0b41d28088f205525751601/webdataset-1.0.2-py3-none-any.whl", hash = "sha256:3dbfced32b25c0d199c6b9787937b6f85742bc3c84f652c846893075c1c082d9", size = 74956 }, -] - -[[package]] -name = "werkzeug" -version = "3.1.4" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "markupsafe" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/45/ea/b0f8eeb287f8df9066e56e831c7824ac6bab645dd6c7a8f4b2d767944f9b/werkzeug-3.1.4.tar.gz", hash = "sha256:cd3cd98b1b92dc3b7b3995038826c68097dcb16f9baa63abe35f20eafeb9fe5e", size = 864687 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/2f/f9/9e082990c2585c744734f85bec79b5dae5df9c974ffee58fe421652c8e91/werkzeug-3.1.4-py3-none-any.whl", hash = "sha256:2ad50fb9ed09cc3af22c54698351027ace879a0b60a3b5edf5730b2f7d876905", size = 224960 }, -] - -[[package]] -name = "wget" -version = "3.2" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/47/6a/62e288da7bcda82b935ff0c6cfe542970f04e29c756b0e147251b2fb251f/wget-3.2.zip", hash = "sha256:35e630eca2aa50ce998b9b1a127bb26b30dfee573702782aa982f875e3f16061", size = 10857 } - -[[package]] -name = "whisper-normalizer" -version = "0.1.12" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "indic-numtowords" }, - { name = "more-itertools" }, - { name = "regex" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/77/81/d4a23d67e9356f1c2d6fe9aa7e99f42078b5e3845b181412a5582f168af4/whisper_normalizer-0.1.12.tar.gz", hash = "sha256:484dcedbfeba2ee94cf9412d57ab1e66b847e91f80c15ffc4c6ab82ad5484b8c", size = 39630 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/07/d7/2bf42cb3f19da0aec48052a6e3bc3a592afc182fe98c011a0e0ae5fbe1f5/whisper_normalizer-0.1.12-py3-none-any.whl", hash = "sha256:2cd7276d2599c05147a50cf86d240e6cd27623f5ccfe8b20ccea6a518274989a", size = 36748 }, -] - -[[package]] -name = "wrapt" -version = "2.0.1" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/49/2a/6de8a50cb435b7f42c46126cf1a54b2aab81784e74c8595c8e025e8f36d3/wrapt-2.0.1.tar.gz", hash = "sha256:9c9c635e78497cacb81e84f8b11b23e0aacac7a136e73b8e5b2109a1d9fc468f", size = 82040 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/61/0d/12d8c803ed2ce4e5e7d5b9f5f602721f9dfef82c95959f3ce97fa584bb5c/wrapt-2.0.1-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:64b103acdaa53b7caf409e8d45d39a8442fe6dcfec6ba3f3d141e0cc2b5b4dbd", size = 77481 }, - { url = "https://files.pythonhosted.org/packages/05/3e/4364ebe221ebf2a44d9fc8695a19324692f7dd2795e64bd59090856ebf12/wrapt-2.0.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:91bcc576260a274b169c3098e9a3519fb01f2989f6d3d386ef9cbf8653de1374", size = 60692 }, - { url = "https://files.pythonhosted.org/packages/1f/ff/ae2a210022b521f86a8ddcdd6058d137c051003812b0388a5e9a03d3fe10/wrapt-2.0.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:ab594f346517010050126fcd822697b25a7031d815bb4fbc238ccbe568216489", size = 61574 }, - { url = "https://files.pythonhosted.org/packages/c6/93/5cf92edd99617095592af919cb81d4bff61c5dbbb70d3c92099425a8ec34/wrapt-2.0.1-cp310-cp310-manylinux1_x86_64.manylinux_2_28_x86_64.manylinux_2_5_x86_64.whl", hash = "sha256:36982b26f190f4d737f04a492a68accbfc6fa042c3f42326fdfbb6c5b7a20a31", size = 113688 }, - { url = "https://files.pythonhosted.org/packages/a0/0a/e38fc0cee1f146c9fb266d8ef96ca39fb14a9eef165383004019aa53f88a/wrapt-2.0.1-cp310-cp310-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:23097ed8bc4c93b7bf36fa2113c6c733c976316ce0ee2c816f64ca06102034ef", size = 115698 }, - { url = "https://files.pythonhosted.org/packages/b0/85/bef44ea018b3925fb0bcbe9112715f665e4d5309bd945191da814c314fd1/wrapt-2.0.1-cp310-cp310-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl", hash = "sha256:8bacfe6e001749a3b64db47bcf0341da757c95959f592823a93931a422395013", size = 112096 }, - { url = "https://files.pythonhosted.org/packages/7c/0b/733a2376e413117e497aa1a5b1b78e8f3a28c0e9537d26569f67d724c7c5/wrapt-2.0.1-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:8ec3303e8a81932171f455f792f8df500fc1a09f20069e5c16bd7049ab4e8e38", size = 114878 }, - { url = "https://files.pythonhosted.org/packages/da/03/d81dcb21bbf678fcda656495792b059f9d56677d119ca022169a12542bd0/wrapt-2.0.1-cp310-cp310-musllinux_1_2_riscv64.whl", hash = "sha256:3f373a4ab5dbc528a94334f9fe444395b23c2f5332adab9ff4ea82f5a9e33bc1", size = 111298 }, - { url = "https://files.pythonhosted.org/packages/c9/d5/5e623040e8056e1108b787020d56b9be93dbbf083bf2324d42cde80f3a19/wrapt-2.0.1-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:f49027b0b9503bf6c8cdc297ca55006b80c2f5dd36cecc72c6835ab6e10e8a25", size = 113361 }, - { url = "https://files.pythonhosted.org/packages/a1/f3/de535ccecede6960e28c7b722e5744846258111d6c9f071aa7578ea37ad3/wrapt-2.0.1-cp310-cp310-win32.whl", hash = "sha256:8330b42d769965e96e01fa14034b28a2a7600fbf7e8f0cc90ebb36d492c993e4", size = 58035 }, - { url = "https://files.pythonhosted.org/packages/21/15/39d3ca5428a70032c2ec8b1f1c9d24c32e497e7ed81aed887a4998905fcc/wrapt-2.0.1-cp310-cp310-win_amd64.whl", hash = "sha256:1218573502a8235bb8a7ecaed12736213b22dcde9feab115fa2989d42b5ded45", size = 60383 }, - { url = "https://files.pythonhosted.org/packages/43/c2/dfd23754b7f7a4dce07e08f4309c4e10a40046a83e9ae1800f2e6b18d7c1/wrapt-2.0.1-cp310-cp310-win_arm64.whl", hash = "sha256:eda8e4ecd662d48c28bb86be9e837c13e45c58b8300e43ba3c9b4fa9900302f7", size = 58894 }, - { url = "https://files.pythonhosted.org/packages/98/60/553997acf3939079dab022e37b67b1904b5b0cc235503226898ba573b10c/wrapt-2.0.1-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:0e17283f533a0d24d6e5429a7d11f250a58d28b4ae5186f8f47853e3e70d2590", size = 77480 }, - { url = "https://files.pythonhosted.org/packages/2d/50/e5b3d30895d77c52105c6d5cbf94d5b38e2a3dd4a53d22d246670da98f7c/wrapt-2.0.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:85df8d92158cb8f3965aecc27cf821461bb5f40b450b03facc5d9f0d4d6ddec6", size = 60690 }, - { url = "https://files.pythonhosted.org/packages/f0/40/660b2898703e5cbbb43db10cdefcc294274458c3ca4c68637c2b99371507/wrapt-2.0.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:c1be685ac7700c966b8610ccc63c3187a72e33cab53526a27b2a285a662cd4f7", size = 61578 }, - { url = "https://files.pythonhosted.org/packages/5b/36/825b44c8a10556957bc0c1d84c7b29a40e05fcf1873b6c40aa9dbe0bd972/wrapt-2.0.1-cp311-cp311-manylinux1_x86_64.manylinux_2_28_x86_64.manylinux_2_5_x86_64.whl", hash = "sha256:df0b6d3b95932809c5b3fecc18fda0f1e07452d05e2662a0b35548985f256e28", size = 114115 }, - { url = "https://files.pythonhosted.org/packages/83/73/0a5d14bb1599677304d3c613a55457d34c344e9b60eda8a737c2ead7619e/wrapt-2.0.1-cp311-cp311-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:4da7384b0e5d4cae05c97cd6f94faaf78cc8b0f791fc63af43436d98c4ab37bb", size = 116157 }, - { url = "https://files.pythonhosted.org/packages/01/22/1c158fe763dbf0a119f985d945711d288994fe5514c0646ebe0eb18b016d/wrapt-2.0.1-cp311-cp311-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl", hash = "sha256:ec65a78fbd9d6f083a15d7613b2800d5663dbb6bb96003899c834beaa68b242c", size = 112535 }, - { url = "https://files.pythonhosted.org/packages/5c/28/4f16861af67d6de4eae9927799b559c20ebdd4fe432e89ea7fe6fcd9d709/wrapt-2.0.1-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:7de3cc939be0e1174969f943f3b44e0d79b6f9a82198133a5b7fc6cc92882f16", size = 115404 }, - { url = "https://files.pythonhosted.org/packages/a0/8b/7960122e625fad908f189b59c4aae2d50916eb4098b0fb2819c5a177414f/wrapt-2.0.1-cp311-cp311-musllinux_1_2_riscv64.whl", hash = "sha256:fb1a5b72cbd751813adc02ef01ada0b0d05d3dcbc32976ce189a1279d80ad4a2", size = 111802 }, - { url = "https://files.pythonhosted.org/packages/3e/73/7881eee5ac31132a713ab19a22c9e5f1f7365c8b1df50abba5d45b781312/wrapt-2.0.1-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:3fa272ca34332581e00bf7773e993d4f632594eb2d1b0b162a9038df0fd971dd", size = 113837 }, - { url = "https://files.pythonhosted.org/packages/45/00/9499a3d14e636d1f7089339f96c4409bbc7544d0889f12264efa25502ae8/wrapt-2.0.1-cp311-cp311-win32.whl", hash = "sha256:fc007fdf480c77301ab1afdbb6ab22a5deee8885f3b1ed7afcb7e5e84a0e27be", size = 58028 }, - { url = "https://files.pythonhosted.org/packages/70/5d/8f3d7eea52f22638748f74b102e38fdf88cb57d08ddeb7827c476a20b01b/wrapt-2.0.1-cp311-cp311-win_amd64.whl", hash = "sha256:47434236c396d04875180171ee1f3815ca1eada05e24a1ee99546320d54d1d1b", size = 60385 }, - { url = "https://files.pythonhosted.org/packages/14/e2/32195e57a8209003587bbbad44d5922f13e0ced2a493bb46ca882c5b123d/wrapt-2.0.1-cp311-cp311-win_arm64.whl", hash = "sha256:837e31620e06b16030b1d126ed78e9383815cbac914693f54926d816d35d8edf", size = 58893 }, - { url = "https://files.pythonhosted.org/packages/cb/73/8cb252858dc8254baa0ce58ce382858e3a1cf616acebc497cb13374c95c6/wrapt-2.0.1-cp312-cp312-macosx_10_13_universal2.whl", hash = "sha256:1fdbb34da15450f2b1d735a0e969c24bdb8d8924892380126e2a293d9902078c", size = 78129 }, - { url = "https://files.pythonhosted.org/packages/19/42/44a0db2108526ee6e17a5ab72478061158f34b08b793df251d9fbb9a7eb4/wrapt-2.0.1-cp312-cp312-macosx_10_13_x86_64.whl", hash = "sha256:3d32794fe940b7000f0519904e247f902f0149edbe6316c710a8562fb6738841", size = 61205 }, - { url = "https://files.pythonhosted.org/packages/4d/8a/5b4b1e44b791c22046e90d9b175f9a7581a8cc7a0debbb930f81e6ae8e25/wrapt-2.0.1-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:386fb54d9cd903ee0012c09291336469eb7b244f7183d40dc3e86a16a4bace62", size = 61692 }, - { url = "https://files.pythonhosted.org/packages/11/53/3e794346c39f462bcf1f58ac0487ff9bdad02f9b6d5ee2dc84c72e0243b2/wrapt-2.0.1-cp312-cp312-manylinux1_x86_64.manylinux_2_28_x86_64.manylinux_2_5_x86_64.whl", hash = "sha256:7b219cb2182f230676308cdcacd428fa837987b89e4b7c5c9025088b8a6c9faf", size = 121492 }, - { url = "https://files.pythonhosted.org/packages/c6/7e/10b7b0e8841e684c8ca76b462a9091c45d62e8f2de9c4b1390b690eadf16/wrapt-2.0.1-cp312-cp312-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:641e94e789b5f6b4822bb8d8ebbdfc10f4e4eae7756d648b717d980f657a9eb9", size = 123064 }, - { url = "https://files.pythonhosted.org/packages/0e/d1/3c1e4321fc2f5ee7fd866b2d822aa89b84495f28676fd976c47327c5b6aa/wrapt-2.0.1-cp312-cp312-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl", hash = "sha256:fe21b118b9f58859b5ebaa4b130dee18669df4bd111daad082b7beb8799ad16b", size = 117403 }, - { url = "https://files.pythonhosted.org/packages/a4/b0/d2f0a413cf201c8c2466de08414a15420a25aa83f53e647b7255cc2fab5d/wrapt-2.0.1-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:17fb85fa4abc26a5184d93b3efd2dcc14deb4b09edcdb3535a536ad34f0b4dba", size = 121500 }, - { url = "https://files.pythonhosted.org/packages/bd/45/bddb11d28ca39970a41ed48a26d210505120f925918592283369219f83cc/wrapt-2.0.1-cp312-cp312-musllinux_1_2_riscv64.whl", hash = "sha256:b89ef9223d665ab255ae42cc282d27d69704d94be0deffc8b9d919179a609684", size = 116299 }, - { url = "https://files.pythonhosted.org/packages/81/af/34ba6dd570ef7a534e7eec0c25e2615c355602c52aba59413411c025a0cb/wrapt-2.0.1-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:a453257f19c31b31ba593c30d997d6e5be39e3b5ad9148c2af5a7314061c63eb", size = 120622 }, - { url = "https://files.pythonhosted.org/packages/e2/3e/693a13b4146646fb03254636f8bafd20c621955d27d65b15de07ab886187/wrapt-2.0.1-cp312-cp312-win32.whl", hash = "sha256:3e271346f01e9c8b1130a6a3b0e11908049fe5be2d365a5f402778049147e7e9", size = 58246 }, - { url = "https://files.pythonhosted.org/packages/a7/36/715ec5076f925a6be95f37917b66ebbeaa1372d1862c2ccd7a751574b068/wrapt-2.0.1-cp312-cp312-win_amd64.whl", hash = "sha256:2da620b31a90cdefa9cd0c2b661882329e2e19d1d7b9b920189956b76c564d75", size = 60492 }, - { url = "https://files.pythonhosted.org/packages/ef/3e/62451cd7d80f65cc125f2b426b25fbb6c514bf6f7011a0c3904fc8c8df90/wrapt-2.0.1-cp312-cp312-win_arm64.whl", hash = "sha256:aea9c7224c302bc8bfc892b908537f56c430802560e827b75ecbde81b604598b", size = 58987 }, - { url = "https://files.pythonhosted.org/packages/ad/fe/41af4c46b5e498c90fc87981ab2972fbd9f0bccda597adb99d3d3441b94b/wrapt-2.0.1-cp313-cp313-macosx_10_13_universal2.whl", hash = "sha256:47b0f8bafe90f7736151f61482c583c86b0693d80f075a58701dd1549b0010a9", size = 78132 }, - { url = "https://files.pythonhosted.org/packages/1c/92/d68895a984a5ebbbfb175512b0c0aad872354a4a2484fbd5552e9f275316/wrapt-2.0.1-cp313-cp313-macosx_10_13_x86_64.whl", hash = "sha256:cbeb0971e13b4bd81d34169ed57a6dda017328d1a22b62fda45e1d21dd06148f", size = 61211 }, - { url = "https://files.pythonhosted.org/packages/e8/26/ba83dc5ae7cf5aa2b02364a3d9cf74374b86169906a1f3ade9a2d03cf21c/wrapt-2.0.1-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:eb7cffe572ad0a141a7886a1d2efa5bef0bf7fe021deeea76b3ab334d2c38218", size = 61689 }, - { url = "https://files.pythonhosted.org/packages/cf/67/d7a7c276d874e5d26738c22444d466a3a64ed541f6ef35f740dbd865bab4/wrapt-2.0.1-cp313-cp313-manylinux1_x86_64.manylinux_2_28_x86_64.manylinux_2_5_x86_64.whl", hash = "sha256:c8d60527d1ecfc131426b10d93ab5d53e08a09c5fa0175f6b21b3252080c70a9", size = 121502 }, - { url = "https://files.pythonhosted.org/packages/0f/6b/806dbf6dd9579556aab22fc92908a876636e250f063f71548a8660382184/wrapt-2.0.1-cp313-cp313-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:c654eafb01afac55246053d67a4b9a984a3567c3808bb7df2f8de1c1caba2e1c", size = 123110 }, - { url = "https://files.pythonhosted.org/packages/e5/08/cdbb965fbe4c02c5233d185d070cabed2ecc1f1e47662854f95d77613f57/wrapt-2.0.1-cp313-cp313-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl", hash = "sha256:98d873ed6c8b4ee2418f7afce666751854d6d03e3c0ec2a399bb039cd2ae89db", size = 117434 }, - { url = "https://files.pythonhosted.org/packages/2d/d1/6aae2ce39db4cb5216302fa2e9577ad74424dfbe315bd6669725569e048c/wrapt-2.0.1-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:c9e850f5b7fc67af856ff054c71690d54fa940c3ef74209ad9f935b4f66a0233", size = 121533 }, - { url = "https://files.pythonhosted.org/packages/79/35/565abf57559fbe0a9155c29879ff43ce8bd28d2ca61033a3a3dd67b70794/wrapt-2.0.1-cp313-cp313-musllinux_1_2_riscv64.whl", hash = "sha256:e505629359cb5f751e16e30cf3f91a1d3ddb4552480c205947da415d597f7ac2", size = 116324 }, - { url = "https://files.pythonhosted.org/packages/e1/e0/53ff5e76587822ee33e560ad55876d858e384158272cd9947abdd4ad42ca/wrapt-2.0.1-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:2879af909312d0baf35f08edeea918ee3af7ab57c37fe47cb6a373c9f2749c7b", size = 120627 }, - { url = "https://files.pythonhosted.org/packages/7c/7b/38df30fd629fbd7612c407643c63e80e1c60bcc982e30ceeae163a9800e7/wrapt-2.0.1-cp313-cp313-win32.whl", hash = "sha256:d67956c676be5a24102c7407a71f4126d30de2a569a1c7871c9f3cabc94225d7", size = 58252 }, - { url = "https://files.pythonhosted.org/packages/85/64/d3954e836ea67c4d3ad5285e5c8fd9d362fd0a189a2db622df457b0f4f6a/wrapt-2.0.1-cp313-cp313-win_amd64.whl", hash = "sha256:9ca66b38dd642bf90c59b6738af8070747b610115a39af2498535f62b5cdc1c3", size = 60500 }, - { url = "https://files.pythonhosted.org/packages/89/4e/3c8b99ac93527cfab7f116089db120fef16aac96e5f6cdb724ddf286086d/wrapt-2.0.1-cp313-cp313-win_arm64.whl", hash = "sha256:5a4939eae35db6b6cec8e7aa0e833dcca0acad8231672c26c2a9ab7a0f8ac9c8", size = 58993 }, - { url = "https://files.pythonhosted.org/packages/f9/f4/eff2b7d711cae20d220780b9300faa05558660afb93f2ff5db61fe725b9a/wrapt-2.0.1-cp313-cp313t-macosx_10_13_universal2.whl", hash = "sha256:a52f93d95c8d38fed0669da2ebdb0b0376e895d84596a976c15a9eb45e3eccb3", size = 82028 }, - { url = "https://files.pythonhosted.org/packages/0c/67/cb945563f66fd0f61a999339460d950f4735c69f18f0a87ca586319b1778/wrapt-2.0.1-cp313-cp313t-macosx_10_13_x86_64.whl", hash = "sha256:4e54bbf554ee29fcceee24fa41c4d091398b911da6e7f5d7bffda963c9aed2e1", size = 62949 }, - { url = "https://files.pythonhosted.org/packages/ec/ca/f63e177f0bbe1e5cf5e8d9b74a286537cd709724384ff20860f8f6065904/wrapt-2.0.1-cp313-cp313t-macosx_11_0_arm64.whl", hash = "sha256:908f8c6c71557f4deaa280f55d0728c3bca0960e8c3dd5ceeeafb3c19942719d", size = 63681 }, - { url = "https://files.pythonhosted.org/packages/39/a1/1b88fcd21fd835dca48b556daef750952e917a2794fa20c025489e2e1f0f/wrapt-2.0.1-cp313-cp313t-manylinux1_x86_64.manylinux_2_28_x86_64.manylinux_2_5_x86_64.whl", hash = "sha256:e2f84e9af2060e3904a32cea9bb6db23ce3f91cfd90c6b426757cf7cc01c45c7", size = 152696 }, - { url = "https://files.pythonhosted.org/packages/62/1c/d9185500c1960d9f5f77b9c0b890b7fc62282b53af7ad1b6bd779157f714/wrapt-2.0.1-cp313-cp313t-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:e3612dc06b436968dfb9142c62e5dfa9eb5924f91120b3c8ff501ad878f90eb3", size = 158859 }, - { url = "https://files.pythonhosted.org/packages/91/60/5d796ed0f481ec003220c7878a1d6894652efe089853a208ea0838c13086/wrapt-2.0.1-cp313-cp313t-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl", hash = "sha256:6d2d947d266d99a1477cd005b23cbd09465276e302515e122df56bb9511aca1b", size = 146068 }, - { url = "https://files.pythonhosted.org/packages/04/f8/75282dd72f102ddbfba137e1e15ecba47b40acff32c08ae97edbf53f469e/wrapt-2.0.1-cp313-cp313t-musllinux_1_2_aarch64.whl", hash = "sha256:7d539241e87b650cbc4c3ac9f32c8d1ac8a54e510f6dca3f6ab60dcfd48c9b10", size = 155724 }, - { url = "https://files.pythonhosted.org/packages/5a/27/fe39c51d1b344caebb4a6a9372157bdb8d25b194b3561b52c8ffc40ac7d1/wrapt-2.0.1-cp313-cp313t-musllinux_1_2_riscv64.whl", hash = "sha256:4811e15d88ee62dbf5c77f2c3ff3932b1e3ac92323ba3912f51fc4016ce81ecf", size = 144413 }, - { url = "https://files.pythonhosted.org/packages/83/2b/9f6b643fe39d4505c7bf926d7c2595b7cb4b607c8c6b500e56c6b36ac238/wrapt-2.0.1-cp313-cp313t-musllinux_1_2_x86_64.whl", hash = "sha256:c1c91405fcf1d501fa5d55df21e58ea49e6b879ae829f1039faaf7e5e509b41e", size = 150325 }, - { url = "https://files.pythonhosted.org/packages/bb/b6/20ffcf2558596a7f58a2e69c89597128781f0b88e124bf5a4cadc05b8139/wrapt-2.0.1-cp313-cp313t-win32.whl", hash = "sha256:e76e3f91f864e89db8b8d2a8311d57df93f01ad6bb1e9b9976d1f2e83e18315c", size = 59943 }, - { url = "https://files.pythonhosted.org/packages/87/6a/0e56111cbb3320151eed5d3821ee1373be13e05b376ea0870711f18810c3/wrapt-2.0.1-cp313-cp313t-win_amd64.whl", hash = "sha256:83ce30937f0ba0d28818807b303a412440c4b63e39d3d8fc036a94764b728c92", size = 63240 }, - { url = "https://files.pythonhosted.org/packages/1d/54/5ab4c53ea1f7f7e5c3e7c1095db92932cc32fd62359d285486d00c2884c3/wrapt-2.0.1-cp313-cp313t-win_arm64.whl", hash = "sha256:4b55cacc57e1dc2d0991dbe74c6419ffd415fb66474a02335cb10efd1aa3f84f", size = 60416 }, - { url = "https://files.pythonhosted.org/packages/73/81/d08d83c102709258e7730d3cd25befd114c60e43ef3891d7e6877971c514/wrapt-2.0.1-cp314-cp314-macosx_10_13_universal2.whl", hash = "sha256:5e53b428f65ece6d9dad23cb87e64506392b720a0b45076c05354d27a13351a1", size = 78290 }, - { url = "https://files.pythonhosted.org/packages/f6/14/393afba2abb65677f313aa680ff0981e829626fed39b6a7e3ec807487790/wrapt-2.0.1-cp314-cp314-macosx_10_13_x86_64.whl", hash = "sha256:ad3ee9d0f254851c71780966eb417ef8e72117155cff04821ab9b60549694a55", size = 61255 }, - { url = "https://files.pythonhosted.org/packages/c4/10/a4a1f2fba205a9462e36e708ba37e5ac95f4987a0f1f8fd23f0bf1fc3b0f/wrapt-2.0.1-cp314-cp314-macosx_11_0_arm64.whl", hash = "sha256:d7b822c61ed04ee6ad64bc90d13368ad6eb094db54883b5dde2182f67a7f22c0", size = 61797 }, - { url = "https://files.pythonhosted.org/packages/12/db/99ba5c37cf1c4fad35349174f1e38bd8d992340afc1ff27f526729b98986/wrapt-2.0.1-cp314-cp314-manylinux1_x86_64.manylinux_2_28_x86_64.manylinux_2_5_x86_64.whl", hash = "sha256:7164a55f5e83a9a0b031d3ffab4d4e36bbec42e7025db560f225489fa929e509", size = 120470 }, - { url = "https://files.pythonhosted.org/packages/30/3f/a1c8d2411eb826d695fc3395a431757331582907a0ec59afce8fe8712473/wrapt-2.0.1-cp314-cp314-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:e60690ba71a57424c8d9ff28f8d006b7ad7772c22a4af432188572cd7fa004a1", size = 122851 }, - { url = "https://files.pythonhosted.org/packages/b3/8d/72c74a63f201768d6a04a8845c7976f86be6f5ff4d74996c272cefc8dafc/wrapt-2.0.1-cp314-cp314-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl", hash = "sha256:3cd1a4bd9a7a619922a8557e1318232e7269b5fb69d4ba97b04d20450a6bf970", size = 117433 }, - { url = "https://files.pythonhosted.org/packages/c7/5a/df37cf4042cb13b08256f8e27023e2f9b3d471d553376616591bb99bcb31/wrapt-2.0.1-cp314-cp314-musllinux_1_2_aarch64.whl", hash = "sha256:b4c2e3d777e38e913b8ce3a6257af72fb608f86a1df471cb1d4339755d0a807c", size = 121280 }, - { url = "https://files.pythonhosted.org/packages/54/34/40d6bc89349f9931e1186ceb3e5fbd61d307fef814f09fbbac98ada6a0c8/wrapt-2.0.1-cp314-cp314-musllinux_1_2_riscv64.whl", hash = "sha256:3d366aa598d69416b5afedf1faa539fac40c1d80a42f6b236c88c73a3c8f2d41", size = 116343 }, - { url = "https://files.pythonhosted.org/packages/70/66/81c3461adece09d20781dee17c2366fdf0cb8754738b521d221ca056d596/wrapt-2.0.1-cp314-cp314-musllinux_1_2_x86_64.whl", hash = "sha256:c235095d6d090aa903f1db61f892fffb779c1eaeb2a50e566b52001f7a0f66ed", size = 119650 }, - { url = "https://files.pythonhosted.org/packages/46/3a/d0146db8be8761a9e388cc9cc1c312b36d583950ec91696f19bbbb44af5a/wrapt-2.0.1-cp314-cp314-win32.whl", hash = "sha256:bfb5539005259f8127ea9c885bdc231978c06b7a980e63a8a61c8c4c979719d0", size = 58701 }, - { url = "https://files.pythonhosted.org/packages/1a/38/5359da9af7d64554be63e9046164bd4d8ff289a2dd365677d25ba3342c08/wrapt-2.0.1-cp314-cp314-win_amd64.whl", hash = "sha256:4ae879acc449caa9ed43fc36ba08392b9412ee67941748d31d94e3cedb36628c", size = 60947 }, - { url = "https://files.pythonhosted.org/packages/aa/3f/96db0619276a833842bf36343685fa04f987dd6e3037f314531a1e00492b/wrapt-2.0.1-cp314-cp314-win_arm64.whl", hash = "sha256:8639b843c9efd84675f1e100ed9e99538ebea7297b62c4b45a7042edb84db03e", size = 59359 }, - { url = "https://files.pythonhosted.org/packages/71/49/5f5d1e867bf2064bf3933bc6cf36ade23505f3902390e175e392173d36a2/wrapt-2.0.1-cp314-cp314t-macosx_10_13_universal2.whl", hash = "sha256:9219a1d946a9b32bb23ccae66bdb61e35c62773ce7ca6509ceea70f344656b7b", size = 82031 }, - { url = "https://files.pythonhosted.org/packages/2b/89/0009a218d88db66ceb83921e5685e820e2c61b59bbbb1324ba65342668bc/wrapt-2.0.1-cp314-cp314t-macosx_10_13_x86_64.whl", hash = "sha256:fa4184e74197af3adad3c889a1af95b53bb0466bced92ea99a0c014e48323eec", size = 62952 }, - { url = "https://files.pythonhosted.org/packages/ae/18/9b968e920dd05d6e44bcc918a046d02afea0fb31b2f1c80ee4020f377cbe/wrapt-2.0.1-cp314-cp314t-macosx_11_0_arm64.whl", hash = "sha256:c5ef2f2b8a53b7caee2f797ef166a390fef73979b15778a4a153e4b5fedce8fa", size = 63688 }, - { url = "https://files.pythonhosted.org/packages/a6/7d/78bdcb75826725885d9ea26c49a03071b10c4c92da93edda612910f150e4/wrapt-2.0.1-cp314-cp314t-manylinux1_x86_64.manylinux_2_28_x86_64.manylinux_2_5_x86_64.whl", hash = "sha256:e042d653a4745be832d5aa190ff80ee4f02c34b21f4b785745eceacd0907b815", size = 152706 }, - { url = "https://files.pythonhosted.org/packages/dd/77/cac1d46f47d32084a703df0d2d29d47e7eb2a7d19fa5cbca0e529ef57659/wrapt-2.0.1-cp314-cp314t-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:2afa23318136709c4b23d87d543b425c399887b4057936cd20386d5b1422b6fa", size = 158866 }, - { url = "https://files.pythonhosted.org/packages/8a/11/b521406daa2421508903bf8d5e8b929216ec2af04839db31c0a2c525eee0/wrapt-2.0.1-cp314-cp314t-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl", hash = "sha256:6c72328f668cf4c503ffcf9434c2b71fdd624345ced7941bc6693e61bbe36bef", size = 146148 }, - { url = "https://files.pythonhosted.org/packages/0c/c0/340b272bed297baa7c9ce0c98ef7017d9c035a17a6a71dce3184b8382da2/wrapt-2.0.1-cp314-cp314t-musllinux_1_2_aarch64.whl", hash = "sha256:3793ac154afb0e5b45d1233cb94d354ef7a983708cc3bb12563853b1d8d53747", size = 155737 }, - { url = "https://files.pythonhosted.org/packages/f3/93/bfcb1fb2bdf186e9c2883a4d1ab45ab099c79cbf8f4e70ea453811fa3ea7/wrapt-2.0.1-cp314-cp314t-musllinux_1_2_riscv64.whl", hash = "sha256:fec0d993ecba3991645b4857837277469c8cc4c554a7e24d064d1ca291cfb81f", size = 144451 }, - { url = "https://files.pythonhosted.org/packages/d2/6b/dca504fb18d971139d232652656180e3bd57120e1193d9a5899c3c0b7cdd/wrapt-2.0.1-cp314-cp314t-musllinux_1_2_x86_64.whl", hash = "sha256:949520bccc1fa227274da7d03bf238be15389cd94e32e4297b92337df9b7a349", size = 150353 }, - { url = "https://files.pythonhosted.org/packages/1d/f6/a1de4bd3653afdf91d250ca5c721ee51195df2b61a4603d4b373aa804d1d/wrapt-2.0.1-cp314-cp314t-win32.whl", hash = "sha256:be9e84e91d6497ba62594158d3d31ec0486c60055c49179edc51ee43d095f79c", size = 60609 }, - { url = "https://files.pythonhosted.org/packages/01/3a/07cd60a9d26fe73efead61c7830af975dfdba8537632d410462672e4432b/wrapt-2.0.1-cp314-cp314t-win_amd64.whl", hash = "sha256:61c4956171c7434634401db448371277d07032a81cc21c599c22953374781395", size = 64038 }, - { url = "https://files.pythonhosted.org/packages/41/99/8a06b8e17dddbf321325ae4eb12465804120f699cd1b8a355718300c62da/wrapt-2.0.1-cp314-cp314t-win_arm64.whl", hash = "sha256:35cdbd478607036fee40273be8ed54a451f5f23121bd9d4be515158f9498f7ad", size = 60634 }, - { url = "https://files.pythonhosted.org/packages/15/d1/b51471c11592ff9c012bd3e2f7334a6ff2f42a7aed2caffcf0bdddc9cb89/wrapt-2.0.1-py3-none-any.whl", hash = "sha256:4d2ce1bf1a48c5277d7969259232b57645aae5686dba1eaeade39442277afbca", size = 44046 }, -] - -[[package]] -name = "xxhash" -version = "3.6.0" -source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/02/84/30869e01909fb37a6cc7e18688ee8bf1e42d57e7e0777636bd47524c43c7/xxhash-3.6.0.tar.gz", hash = "sha256:f0162a78b13a0d7617b2845b90c763339d1f1d82bb04a4b07f4ab535cc5e05d6", size = 85160 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/34/ee/f9f1d656ad168681bb0f6b092372c1e533c4416b8069b1896a175c46e484/xxhash-3.6.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:87ff03d7e35c61435976554477a7f4cd1704c3596a89a8300d5ce7fc83874a71", size = 32845 }, - { url = "https://files.pythonhosted.org/packages/a3/b1/93508d9460b292c74a09b83d16750c52a0ead89c51eea9951cb97a60d959/xxhash-3.6.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:f572dfd3d0e2eb1a57511831cf6341242f5a9f8298a45862d085f5b93394a27d", size = 30807 }, - { url = "https://files.pythonhosted.org/packages/07/55/28c93a3662f2d200c70704efe74aab9640e824f8ce330d8d3943bf7c9b3c/xxhash-3.6.0-cp310-cp310-manylinux1_i686.manylinux_2_28_i686.manylinux_2_5_i686.whl", hash = "sha256:89952ea539566b9fed2bbd94e589672794b4286f342254fad28b149f9615fef8", size = 193786 }, - { url = "https://files.pythonhosted.org/packages/c1/96/fec0be9bb4b8f5d9c57d76380a366f31a1781fb802f76fc7cda6c84893c7/xxhash-3.6.0-cp310-cp310-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:48e6f2ffb07a50b52465a1032c3cf1f4a5683f944acaca8a134a2f23674c2058", size = 212830 }, - { url = "https://files.pythonhosted.org/packages/c4/a0/c706845ba77b9611f81fd2e93fad9859346b026e8445e76f8c6fd057cc6d/xxhash-3.6.0-cp310-cp310-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:b5b848ad6c16d308c3ac7ad4ba6bede80ed5df2ba8ed382f8932df63158dd4b2", size = 211606 }, - { url = "https://files.pythonhosted.org/packages/67/1e/164126a2999e5045f04a69257eea946c0dc3e86541b400d4385d646b53d7/xxhash-3.6.0-cp310-cp310-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:a034590a727b44dd8ac5914236a7b8504144447a9682586c3327e935f33ec8cc", size = 444872 }, - { url = "https://files.pythonhosted.org/packages/2d/4b/55ab404c56cd70a2cf5ecfe484838865d0fea5627365c6c8ca156bd09c8f/xxhash-3.6.0-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:8a8f1972e75ebdd161d7896743122834fe87378160c20e97f8b09166213bf8cc", size = 193217 }, - { url = "https://files.pythonhosted.org/packages/45/e6/52abf06bac316db33aa269091ae7311bd53cfc6f4b120ae77bac1b348091/xxhash-3.6.0-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:ee34327b187f002a596d7b167ebc59a1b729e963ce645964bbc050d2f1b73d07", size = 210139 }, - { url = "https://files.pythonhosted.org/packages/34/37/db94d490b8691236d356bc249c08819cbcef9273a1a30acf1254ff9ce157/xxhash-3.6.0-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:339f518c3c7a850dd033ab416ea25a692759dc7478a71131fe8869010d2b75e4", size = 197669 }, - { url = "https://files.pythonhosted.org/packages/b7/36/c4f219ef4a17a4f7a64ed3569bc2b5a9c8311abdb22249ac96093625b1a4/xxhash-3.6.0-cp310-cp310-musllinux_1_2_ppc64le.whl", hash = "sha256:bf48889c9630542d4709192578aebbd836177c9f7a4a2778a7d6340107c65f06", size = 210018 }, - { url = "https://files.pythonhosted.org/packages/fd/06/bfac889a374fc2fc439a69223d1750eed2e18a7db8514737ab630534fa08/xxhash-3.6.0-cp310-cp310-musllinux_1_2_s390x.whl", hash = "sha256:5576b002a56207f640636056b4160a378fe36a58db73ae5c27a7ec8db35f71d4", size = 413058 }, - { url = "https://files.pythonhosted.org/packages/c9/d1/555d8447e0dd32ad0930a249a522bb2e289f0d08b6b16204cfa42c1f5a0c/xxhash-3.6.0-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:af1f3278bd02814d6dedc5dec397993b549d6f16c19379721e5a1d31e132c49b", size = 190628 }, - { url = "https://files.pythonhosted.org/packages/d1/15/8751330b5186cedc4ed4b597989882ea05e0408b53fa47bcb46a6125bfc6/xxhash-3.6.0-cp310-cp310-win32.whl", hash = "sha256:aed058764db109dc9052720da65fafe84873b05eb8b07e5e653597951af57c3b", size = 30577 }, - { url = "https://files.pythonhosted.org/packages/bb/cc/53f87e8b5871a6eb2ff7e89c48c66093bda2be52315a8161ddc54ea550c4/xxhash-3.6.0-cp310-cp310-win_amd64.whl", hash = "sha256:e82da5670f2d0d98950317f82a0e4a0197150ff19a6df2ba40399c2a3b9ae5fb", size = 31487 }, - { url = "https://files.pythonhosted.org/packages/9f/00/60f9ea3bb697667a14314d7269956f58bf56bb73864f8f8d52a3c2535e9a/xxhash-3.6.0-cp310-cp310-win_arm64.whl", hash = "sha256:4a082ffff8c6ac07707fb6b671caf7c6e020c75226c561830b73d862060f281d", size = 27863 }, - { url = "https://files.pythonhosted.org/packages/17/d4/cc2f0400e9154df4b9964249da78ebd72f318e35ccc425e9f403c392f22a/xxhash-3.6.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:b47bbd8cf2d72797f3c2772eaaac0ded3d3af26481a26d7d7d41dc2d3c46b04a", size = 32844 }, - { url = "https://files.pythonhosted.org/packages/5e/ec/1cc11cd13e26ea8bc3cb4af4eaadd8d46d5014aebb67be3f71fb0b68802a/xxhash-3.6.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:2b6821e94346f96db75abaa6e255706fb06ebd530899ed76d32cd99f20dc52fa", size = 30809 }, - { url = "https://files.pythonhosted.org/packages/04/5f/19fe357ea348d98ca22f456f75a30ac0916b51c753e1f8b2e0e6fb884cce/xxhash-3.6.0-cp311-cp311-manylinux1_i686.manylinux_2_28_i686.manylinux_2_5_i686.whl", hash = "sha256:d0a9751f71a1a65ce3584e9cae4467651c7e70c9d31017fa57574583a4540248", size = 194665 }, - { url = "https://files.pythonhosted.org/packages/90/3b/d1f1a8f5442a5fd8beedae110c5af7604dc37349a8e16519c13c19a9a2de/xxhash-3.6.0-cp311-cp311-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:8b29ee68625ab37b04c0b40c3fafdf24d2f75ccd778333cfb698f65f6c463f62", size = 213550 }, - { url = "https://files.pythonhosted.org/packages/c4/ef/3a9b05eb527457d5db13a135a2ae1a26c80fecd624d20f3e8dcc4cb170f3/xxhash-3.6.0-cp311-cp311-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:6812c25fe0d6c36a46ccb002f40f27ac903bf18af9f6dd8f9669cb4d176ab18f", size = 212384 }, - { url = "https://files.pythonhosted.org/packages/0f/18/ccc194ee698c6c623acbf0f8c2969811a8a4b6185af5e824cd27b9e4fd3e/xxhash-3.6.0-cp311-cp311-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:4ccbff013972390b51a18ef1255ef5ac125c92dc9143b2d1909f59abc765540e", size = 445749 }, - { url = "https://files.pythonhosted.org/packages/a5/86/cf2c0321dc3940a7aa73076f4fd677a0fb3e405cb297ead7d864fd90847e/xxhash-3.6.0-cp311-cp311-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:297b7fbf86c82c550e12e8fb71968b3f033d27b874276ba3624ea868c11165a8", size = 193880 }, - { url = "https://files.pythonhosted.org/packages/82/fb/96213c8560e6f948a1ecc9a7613f8032b19ee45f747f4fca4eb31bb6d6ed/xxhash-3.6.0-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:dea26ae1eb293db089798d3973a5fc928a18fdd97cc8801226fae705b02b14b0", size = 210912 }, - { url = "https://files.pythonhosted.org/packages/40/aa/4395e669b0606a096d6788f40dbdf2b819d6773aa290c19e6e83cbfc312f/xxhash-3.6.0-cp311-cp311-musllinux_1_2_i686.whl", hash = "sha256:7a0b169aafb98f4284f73635a8e93f0735f9cbde17bd5ec332480484241aaa77", size = 198654 }, - { url = "https://files.pythonhosted.org/packages/67/74/b044fcd6b3d89e9b1b665924d85d3f400636c23590226feb1eb09e1176ce/xxhash-3.6.0-cp311-cp311-musllinux_1_2_ppc64le.whl", hash = "sha256:08d45aef063a4531b785cd72de4887766d01dc8f362a515693df349fdb825e0c", size = 210867 }, - { url = "https://files.pythonhosted.org/packages/bc/fd/3ce73bf753b08cb19daee1eb14aa0d7fe331f8da9c02dd95316ddfe5275e/xxhash-3.6.0-cp311-cp311-musllinux_1_2_s390x.whl", hash = "sha256:929142361a48ee07f09121fe9e96a84950e8d4df3bb298ca5d88061969f34d7b", size = 414012 }, - { url = "https://files.pythonhosted.org/packages/ba/b3/5a4241309217c5c876f156b10778f3ab3af7ba7e3259e6d5f5c7d0129eb2/xxhash-3.6.0-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:51312c768403d8540487dbbfb557454cfc55589bbde6424456951f7fcd4facb3", size = 191409 }, - { url = "https://files.pythonhosted.org/packages/c0/01/99bfbc15fb9abb9a72b088c1d95219fc4782b7d01fc835bd5744d66dd0b8/xxhash-3.6.0-cp311-cp311-win32.whl", hash = "sha256:d1927a69feddc24c987b337ce81ac15c4720955b667fe9b588e02254b80446fd", size = 30574 }, - { url = "https://files.pythonhosted.org/packages/65/79/9d24d7f53819fe301b231044ea362ce64e86c74f6e8c8e51320de248b3e5/xxhash-3.6.0-cp311-cp311-win_amd64.whl", hash = "sha256:26734cdc2d4ffe449b41d186bbeac416f704a482ed835d375a5c0cb02bc63fef", size = 31481 }, - { url = "https://files.pythonhosted.org/packages/30/4e/15cd0e3e8772071344eab2961ce83f6e485111fed8beb491a3f1ce100270/xxhash-3.6.0-cp311-cp311-win_arm64.whl", hash = "sha256:d72f67ef8bf36e05f5b6c65e8524f265bd61071471cd4cf1d36743ebeeeb06b7", size = 27861 }, - { url = "https://files.pythonhosted.org/packages/9a/07/d9412f3d7d462347e4511181dea65e47e0d0e16e26fbee2ea86a2aefb657/xxhash-3.6.0-cp312-cp312-macosx_10_13_x86_64.whl", hash = "sha256:01362c4331775398e7bb34e3ab403bc9ee9f7c497bc7dee6272114055277dd3c", size = 32744 }, - { url = "https://files.pythonhosted.org/packages/79/35/0429ee11d035fc33abe32dca1b2b69e8c18d236547b9a9b72c1929189b9a/xxhash-3.6.0-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:b7b2df81a23f8cb99656378e72501b2cb41b1827c0f5a86f87d6b06b69f9f204", size = 30816 }, - { url = "https://files.pythonhosted.org/packages/b7/f2/57eb99aa0f7d98624c0932c5b9a170e1806406cdbcdb510546634a1359e0/xxhash-3.6.0-cp312-cp312-manylinux1_i686.manylinux_2_28_i686.manylinux_2_5_i686.whl", hash = "sha256:dc94790144e66b14f67b10ac8ed75b39ca47536bf8800eb7c24b50271ea0c490", size = 194035 }, - { url = "https://files.pythonhosted.org/packages/4c/ed/6224ba353690d73af7a3f1c7cdb1fc1b002e38f783cb991ae338e1eb3d79/xxhash-3.6.0-cp312-cp312-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:93f107c673bccf0d592cdba077dedaf52fe7f42dcd7676eba1f6d6f0c3efffd2", size = 212914 }, - { url = "https://files.pythonhosted.org/packages/38/86/fb6b6130d8dd6b8942cc17ab4d90e223653a89aa32ad2776f8af7064ed13/xxhash-3.6.0-cp312-cp312-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:2aa5ee3444c25b69813663c9f8067dcfaa2e126dc55e8dddf40f4d1c25d7effa", size = 212163 }, - { url = "https://files.pythonhosted.org/packages/ee/dc/e84875682b0593e884ad73b2d40767b5790d417bde603cceb6878901d647/xxhash-3.6.0-cp312-cp312-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:f7f99123f0e1194fa59cc69ad46dbae2e07becec5df50a0509a808f90a0f03f0", size = 445411 }, - { url = "https://files.pythonhosted.org/packages/11/4f/426f91b96701ec2f37bb2b8cec664eff4f658a11f3fa9d94f0a887ea6d2b/xxhash-3.6.0-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:49e03e6fe2cac4a1bc64952dd250cf0dbc5ef4ebb7b8d96bce82e2de163c82a2", size = 193883 }, - { url = "https://files.pythonhosted.org/packages/53/5a/ddbb83eee8e28b778eacfc5a85c969673e4023cdeedcfcef61f36731610b/xxhash-3.6.0-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:bd17fede52a17a4f9a7bc4472a5867cb0b160deeb431795c0e4abe158bc784e9", size = 210392 }, - { url = "https://files.pythonhosted.org/packages/1e/c2/ff69efd07c8c074ccdf0a4f36fcdd3d27363665bcdf4ba399abebe643465/xxhash-3.6.0-cp312-cp312-musllinux_1_2_i686.whl", hash = "sha256:6fb5f5476bef678f69db04f2bd1efbed3030d2aba305b0fc1773645f187d6a4e", size = 197898 }, - { url = "https://files.pythonhosted.org/packages/58/ca/faa05ac19b3b622c7c9317ac3e23954187516298a091eb02c976d0d3dd45/xxhash-3.6.0-cp312-cp312-musllinux_1_2_ppc64le.whl", hash = "sha256:843b52f6d88071f87eba1631b684fcb4b2068cd2180a0224122fe4ef011a9374", size = 210655 }, - { url = "https://files.pythonhosted.org/packages/d4/7a/06aa7482345480cc0cb597f5c875b11a82c3953f534394f620b0be2f700c/xxhash-3.6.0-cp312-cp312-musllinux_1_2_s390x.whl", hash = "sha256:7d14a6cfaf03b1b6f5f9790f76880601ccc7896aff7ab9cd8978a939c1eb7e0d", size = 414001 }, - { url = "https://files.pythonhosted.org/packages/23/07/63ffb386cd47029aa2916b3d2f454e6cc5b9f5c5ada3790377d5430084e7/xxhash-3.6.0-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:418daf3db71e1413cfe211c2f9a528456936645c17f46b5204705581a45390ae", size = 191431 }, - { url = "https://files.pythonhosted.org/packages/0f/93/14fde614cadb4ddf5e7cebf8918b7e8fac5ae7861c1875964f17e678205c/xxhash-3.6.0-cp312-cp312-win32.whl", hash = "sha256:50fc255f39428a27299c20e280d6193d8b63b8ef8028995323bf834a026b4fbb", size = 30617 }, - { url = "https://files.pythonhosted.org/packages/13/5d/0d125536cbe7565a83d06e43783389ecae0c0f2ed037b48ede185de477c0/xxhash-3.6.0-cp312-cp312-win_amd64.whl", hash = "sha256:c0f2ab8c715630565ab8991b536ecded9416d615538be8ecddce43ccf26cbc7c", size = 31534 }, - { url = "https://files.pythonhosted.org/packages/54/85/6ec269b0952ec7e36ba019125982cf11d91256a778c7c3f98a4c5043d283/xxhash-3.6.0-cp312-cp312-win_arm64.whl", hash = "sha256:eae5c13f3bc455a3bbb68bdc513912dc7356de7e2280363ea235f71f54064829", size = 27876 }, - { url = "https://files.pythonhosted.org/packages/33/76/35d05267ac82f53ae9b0e554da7c5e281ee61f3cad44c743f0fcd354f211/xxhash-3.6.0-cp313-cp313-macosx_10_13_x86_64.whl", hash = "sha256:599e64ba7f67472481ceb6ee80fa3bd828fd61ba59fb11475572cc5ee52b89ec", size = 32738 }, - { url = "https://files.pythonhosted.org/packages/31/a8/3fbce1cd96534a95e35d5120637bf29b0d7f5d8fa2f6374e31b4156dd419/xxhash-3.6.0-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:7d8b8aaa30fca4f16f0c84a5c8d7ddee0e25250ec2796c973775373257dde8f1", size = 30821 }, - { url = "https://files.pythonhosted.org/packages/0c/ea/d387530ca7ecfa183cb358027f1833297c6ac6098223fd14f9782cd0015c/xxhash-3.6.0-cp313-cp313-manylinux1_i686.manylinux_2_28_i686.manylinux_2_5_i686.whl", hash = "sha256:d597acf8506d6e7101a4a44a5e428977a51c0fadbbfd3c39650cca9253f6e5a6", size = 194127 }, - { url = "https://files.pythonhosted.org/packages/ba/0c/71435dcb99874b09a43b8d7c54071e600a7481e42b3e3ce1eb5226a5711a/xxhash-3.6.0-cp313-cp313-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:858dc935963a33bc33490128edc1c12b0c14d9c7ebaa4e387a7869ecc4f3e263", size = 212975 }, - { url = "https://files.pythonhosted.org/packages/84/7a/c2b3d071e4bb4a90b7057228a99b10d51744878f4a8a6dd643c8bd897620/xxhash-3.6.0-cp313-cp313-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:ba284920194615cb8edf73bf52236ce2e1664ccd4a38fdb543506413529cc546", size = 212241 }, - { url = "https://files.pythonhosted.org/packages/81/5f/640b6eac0128e215f177df99eadcd0f1b7c42c274ab6a394a05059694c5a/xxhash-3.6.0-cp313-cp313-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:4b54219177f6c6674d5378bd862c6aedf64725f70dd29c472eaae154df1a2e89", size = 445471 }, - { url = "https://files.pythonhosted.org/packages/5e/1e/3c3d3ef071b051cc3abbe3721ffb8365033a172613c04af2da89d5548a87/xxhash-3.6.0-cp313-cp313-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:42c36dd7dbad2f5238950c377fcbf6811b1cdb1c444fab447960030cea60504d", size = 193936 }, - { url = "https://files.pythonhosted.org/packages/2c/bd/4a5f68381939219abfe1c22a9e3a5854a4f6f6f3c4983a87d255f21f2e5d/xxhash-3.6.0-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:f22927652cba98c44639ffdc7aaf35828dccf679b10b31c4ad72a5b530a18eb7", size = 210440 }, - { url = "https://files.pythonhosted.org/packages/eb/37/b80fe3d5cfb9faff01a02121a0f4d565eb7237e9e5fc66e73017e74dcd36/xxhash-3.6.0-cp313-cp313-musllinux_1_2_i686.whl", hash = "sha256:b45fad44d9c5c119e9c6fbf2e1c656a46dc68e280275007bbfd3d572b21426db", size = 197990 }, - { url = "https://files.pythonhosted.org/packages/d7/fd/2c0a00c97b9e18f72e1f240ad4e8f8a90fd9d408289ba9c7c495ed7dc05c/xxhash-3.6.0-cp313-cp313-musllinux_1_2_ppc64le.whl", hash = "sha256:6f2580ffab1a8b68ef2b901cde7e55fa8da5e4be0977c68f78fc80f3c143de42", size = 210689 }, - { url = "https://files.pythonhosted.org/packages/93/86/5dd8076a926b9a95db3206aba20d89a7fc14dd5aac16e5c4de4b56033140/xxhash-3.6.0-cp313-cp313-musllinux_1_2_s390x.whl", hash = "sha256:40c391dd3cd041ebc3ffe6f2c862f402e306eb571422e0aa918d8070ba31da11", size = 414068 }, - { url = "https://files.pythonhosted.org/packages/af/3c/0bb129170ee8f3650f08e993baee550a09593462a5cddd8e44d0011102b1/xxhash-3.6.0-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:f205badabde7aafd1a31e8ca2a3e5a763107a71c397c4481d6a804eb5063d8bd", size = 191495 }, - { url = "https://files.pythonhosted.org/packages/e9/3a/6797e0114c21d1725e2577508e24006fd7ff1d8c0c502d3b52e45c1771d8/xxhash-3.6.0-cp313-cp313-win32.whl", hash = "sha256:2577b276e060b73b73a53042ea5bd5203d3e6347ce0d09f98500f418a9fcf799", size = 30620 }, - { url = "https://files.pythonhosted.org/packages/86/15/9bc32671e9a38b413a76d24722a2bf8784a132c043063a8f5152d390b0f9/xxhash-3.6.0-cp313-cp313-win_amd64.whl", hash = "sha256:757320d45d2fbcce8f30c42a6b2f47862967aea7bf458b9625b4bbe7ee390392", size = 31542 }, - { url = "https://files.pythonhosted.org/packages/39/c5/cc01e4f6188656e56112d6a8e0dfe298a16934b8c47a247236549a3f7695/xxhash-3.6.0-cp313-cp313-win_arm64.whl", hash = "sha256:457b8f85dec5825eed7b69c11ae86834a018b8e3df5e77783c999663da2f96d6", size = 27880 }, - { url = "https://files.pythonhosted.org/packages/f3/30/25e5321c8732759e930c555176d37e24ab84365482d257c3b16362235212/xxhash-3.6.0-cp313-cp313t-macosx_10_13_x86_64.whl", hash = "sha256:a42e633d75cdad6d625434e3468126c73f13f7584545a9cf34e883aa1710e702", size = 32956 }, - { url = "https://files.pythonhosted.org/packages/9f/3c/0573299560d7d9f8ab1838f1efc021a280b5ae5ae2e849034ef3dee18810/xxhash-3.6.0-cp313-cp313t-macosx_11_0_arm64.whl", hash = "sha256:568a6d743219e717b07b4e03b0a828ce593833e498c3b64752e0f5df6bfe84db", size = 31072 }, - { url = "https://files.pythonhosted.org/packages/7a/1c/52d83a06e417cd9d4137722693424885cc9878249beb3a7c829e74bf7ce9/xxhash-3.6.0-cp313-cp313t-manylinux1_i686.manylinux_2_28_i686.manylinux_2_5_i686.whl", hash = "sha256:bec91b562d8012dae276af8025a55811b875baace6af510412a5e58e3121bc54", size = 196409 }, - { url = "https://files.pythonhosted.org/packages/e3/8e/c6d158d12a79bbd0b878f8355432075fc82759e356ab5a111463422a239b/xxhash-3.6.0-cp313-cp313t-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:78e7f2f4c521c30ad5e786fdd6bae89d47a32672a80195467b5de0480aa97b1f", size = 215736 }, - { url = "https://files.pythonhosted.org/packages/bc/68/c4c80614716345d55071a396cf03d06e34b5f4917a467faf43083c995155/xxhash-3.6.0-cp313-cp313t-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:3ed0df1b11a79856df5ffcab572cbd6b9627034c1c748c5566fa79df9048a7c5", size = 214833 }, - { url = "https://files.pythonhosted.org/packages/7e/e9/ae27c8ffec8b953efa84c7c4a6c6802c263d587b9fc0d6e7cea64e08c3af/xxhash-3.6.0-cp313-cp313t-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:0e4edbfc7d420925b0dd5e792478ed393d6e75ff8fc219a6546fb446b6a417b1", size = 448348 }, - { url = "https://files.pythonhosted.org/packages/d7/6b/33e21afb1b5b3f46b74b6bd1913639066af218d704cc0941404ca717fc57/xxhash-3.6.0-cp313-cp313t-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:fba27a198363a7ef87f8c0f6b171ec36b674fe9053742c58dd7e3201c1ab30ee", size = 196070 }, - { url = "https://files.pythonhosted.org/packages/96/b6/fcabd337bc5fa624e7203aa0fa7d0c49eed22f72e93229431752bddc83d9/xxhash-3.6.0-cp313-cp313t-musllinux_1_2_aarch64.whl", hash = "sha256:794fe9145fe60191c6532fa95063765529770edcdd67b3d537793e8004cabbfd", size = 212907 }, - { url = "https://files.pythonhosted.org/packages/4b/d3/9ee6160e644d660fcf176c5825e61411c7f62648728f69c79ba237250143/xxhash-3.6.0-cp313-cp313t-musllinux_1_2_i686.whl", hash = "sha256:6105ef7e62b5ac73a837778efc331a591d8442f8ef5c7e102376506cb4ae2729", size = 200839 }, - { url = "https://files.pythonhosted.org/packages/0d/98/e8de5baa5109394baf5118f5e72ab21a86387c4f89b0e77ef3e2f6b0327b/xxhash-3.6.0-cp313-cp313t-musllinux_1_2_ppc64le.whl", hash = "sha256:f01375c0e55395b814a679b3eea205db7919ac2af213f4a6682e01220e5fe292", size = 213304 }, - { url = "https://files.pythonhosted.org/packages/7b/1d/71056535dec5c3177eeb53e38e3d367dd1d16e024e63b1cee208d572a033/xxhash-3.6.0-cp313-cp313t-musllinux_1_2_s390x.whl", hash = "sha256:d706dca2d24d834a4661619dcacf51a75c16d65985718d6a7d73c1eeeb903ddf", size = 416930 }, - { url = "https://files.pythonhosted.org/packages/dc/6c/5cbde9de2cd967c322e651c65c543700b19e7ae3e0aae8ece3469bf9683d/xxhash-3.6.0-cp313-cp313t-musllinux_1_2_x86_64.whl", hash = "sha256:5f059d9faeacd49c0215d66f4056e1326c80503f51a1532ca336a385edadd033", size = 193787 }, - { url = "https://files.pythonhosted.org/packages/19/fa/0172e350361d61febcea941b0cc541d6e6c8d65d153e85f850a7b256ff8a/xxhash-3.6.0-cp313-cp313t-win32.whl", hash = "sha256:1244460adc3a9be84731d72b8e80625788e5815b68da3da8b83f78115a40a7ec", size = 30916 }, - { url = "https://files.pythonhosted.org/packages/ad/e6/e8cf858a2b19d6d45820f072eff1bea413910592ff17157cabc5f1227a16/xxhash-3.6.0-cp313-cp313t-win_amd64.whl", hash = "sha256:b1e420ef35c503869c4064f4a2f2b08ad6431ab7b229a05cce39d74268bca6b8", size = 31799 }, - { url = "https://files.pythonhosted.org/packages/56/15/064b197e855bfb7b343210e82490ae672f8bc7cdf3ddb02e92f64304ee8a/xxhash-3.6.0-cp313-cp313t-win_arm64.whl", hash = "sha256:ec44b73a4220623235f67a996c862049f375df3b1052d9899f40a6382c32d746", size = 28044 }, - { url = "https://files.pythonhosted.org/packages/7e/5e/0138bc4484ea9b897864d59fce9be9086030825bc778b76cb5a33a906d37/xxhash-3.6.0-cp314-cp314-macosx_10_13_x86_64.whl", hash = "sha256:a40a3d35b204b7cc7643cbcf8c9976d818cb47befcfac8bbefec8038ac363f3e", size = 32754 }, - { url = "https://files.pythonhosted.org/packages/18/d7/5dac2eb2ec75fd771957a13e5dda560efb2176d5203f39502a5fc571f899/xxhash-3.6.0-cp314-cp314-macosx_11_0_arm64.whl", hash = "sha256:a54844be970d3fc22630b32d515e79a90d0a3ddb2644d8d7402e3c4c8da61405", size = 30846 }, - { url = "https://files.pythonhosted.org/packages/fe/71/8bc5be2bb00deb5682e92e8da955ebe5fa982da13a69da5a40a4c8db12fb/xxhash-3.6.0-cp314-cp314-manylinux1_i686.manylinux_2_28_i686.manylinux_2_5_i686.whl", hash = "sha256:016e9190af8f0a4e3741343777710e3d5717427f175adfdc3e72508f59e2a7f3", size = 194343 }, - { url = "https://files.pythonhosted.org/packages/e7/3b/52badfb2aecec2c377ddf1ae75f55db3ba2d321c5e164f14461c90837ef3/xxhash-3.6.0-cp314-cp314-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:4f6f72232f849eb9d0141e2ebe2677ece15adfd0fa599bc058aad83c714bb2c6", size = 213074 }, - { url = "https://files.pythonhosted.org/packages/a2/2b/ae46b4e9b92e537fa30d03dbc19cdae57ed407e9c26d163895e968e3de85/xxhash-3.6.0-cp314-cp314-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:63275a8aba7865e44b1813d2177e0f5ea7eadad3dd063a21f7cf9afdc7054063", size = 212388 }, - { url = "https://files.pythonhosted.org/packages/f5/80/49f88d3afc724b4ac7fbd664c8452d6db51b49915be48c6982659e0e7942/xxhash-3.6.0-cp314-cp314-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:3cd01fa2aa00d8b017c97eb46b9a794fbdca53fc14f845f5a328c71254b0abb7", size = 445614 }, - { url = "https://files.pythonhosted.org/packages/ed/ba/603ce3961e339413543d8cd44f21f2c80e2a7c5cfe692a7b1f2cccf58f3c/xxhash-3.6.0-cp314-cp314-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:0226aa89035b62b6a86d3c68df4d7c1f47a342b8683da2b60cedcddb46c4d95b", size = 194024 }, - { url = "https://files.pythonhosted.org/packages/78/d1/8e225ff7113bf81545cfdcd79eef124a7b7064a0bba53605ff39590b95c2/xxhash-3.6.0-cp314-cp314-musllinux_1_2_aarch64.whl", hash = "sha256:c6e193e9f56e4ca4923c61238cdaced324f0feac782544eb4c6d55ad5cc99ddd", size = 210541 }, - { url = "https://files.pythonhosted.org/packages/6f/58/0f89d149f0bad89def1a8dd38feb50ccdeb643d9797ec84707091d4cb494/xxhash-3.6.0-cp314-cp314-musllinux_1_2_i686.whl", hash = "sha256:9176dcaddf4ca963d4deb93866d739a343c01c969231dbe21680e13a5d1a5bf0", size = 198305 }, - { url = "https://files.pythonhosted.org/packages/11/38/5eab81580703c4df93feb5f32ff8fa7fe1e2c51c1f183ee4e48d4bb9d3d7/xxhash-3.6.0-cp314-cp314-musllinux_1_2_ppc64le.whl", hash = "sha256:c1ce4009c97a752e682b897aa99aef84191077a9433eb237774689f14f8ec152", size = 210848 }, - { url = "https://files.pythonhosted.org/packages/5e/6b/953dc4b05c3ce678abca756416e4c130d2382f877a9c30a20d08ee6a77c0/xxhash-3.6.0-cp314-cp314-musllinux_1_2_s390x.whl", hash = "sha256:8cb2f4f679b01513b7adbb9b1b2f0f9cdc31b70007eaf9d59d0878809f385b11", size = 414142 }, - { url = "https://files.pythonhosted.org/packages/08/a9/238ec0d4e81a10eb5026d4a6972677cbc898ba6c8b9dbaec12ae001b1b35/xxhash-3.6.0-cp314-cp314-musllinux_1_2_x86_64.whl", hash = "sha256:653a91d7c2ab54a92c19ccf43508b6a555440b9be1bc8be553376778be7f20b5", size = 191547 }, - { url = "https://files.pythonhosted.org/packages/f1/ee/3cf8589e06c2164ac77c3bf0aa127012801128f1feebf2a079272da5737c/xxhash-3.6.0-cp314-cp314-win32.whl", hash = "sha256:a756fe893389483ee8c394d06b5ab765d96e68fbbfe6fde7aa17e11f5720559f", size = 31214 }, - { url = "https://files.pythonhosted.org/packages/02/5d/a19552fbc6ad4cb54ff953c3908bbc095f4a921bc569433d791f755186f1/xxhash-3.6.0-cp314-cp314-win_amd64.whl", hash = "sha256:39be8e4e142550ef69629c9cd71b88c90e9a5db703fecbcf265546d9536ca4ad", size = 32290 }, - { url = "https://files.pythonhosted.org/packages/b1/11/dafa0643bc30442c887b55baf8e73353a344ee89c1901b5a5c54a6c17d39/xxhash-3.6.0-cp314-cp314-win_arm64.whl", hash = "sha256:25915e6000338999236f1eb68a02a32c3275ac338628a7eaa5a269c401995679", size = 28795 }, - { url = "https://files.pythonhosted.org/packages/2c/db/0e99732ed7f64182aef4a6fb145e1a295558deec2a746265dcdec12d191e/xxhash-3.6.0-cp314-cp314t-macosx_10_13_x86_64.whl", hash = "sha256:c5294f596a9017ca5a3e3f8884c00b91ab2ad2933cf288f4923c3fd4346cf3d4", size = 32955 }, - { url = "https://files.pythonhosted.org/packages/55/f4/2a7c3c68e564a099becfa44bb3d398810cc0ff6749b0d3cb8ccb93f23c14/xxhash-3.6.0-cp314-cp314t-macosx_11_0_arm64.whl", hash = "sha256:1cf9dcc4ab9cff01dfbba78544297a3a01dafd60f3bde4e2bfd016cf7e4ddc67", size = 31072 }, - { url = "https://files.pythonhosted.org/packages/c6/d9/72a29cddc7250e8a5819dad5d466facb5dc4c802ce120645630149127e73/xxhash-3.6.0-cp314-cp314t-manylinux1_i686.manylinux_2_28_i686.manylinux_2_5_i686.whl", hash = "sha256:01262da8798422d0685f7cef03b2bd3f4f46511b02830861df548d7def4402ad", size = 196579 }, - { url = "https://files.pythonhosted.org/packages/63/93/b21590e1e381040e2ca305a884d89e1c345b347404f7780f07f2cdd47ef4/xxhash-3.6.0-cp314-cp314t-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:51a73fb7cb3a3ead9f7a8b583ffd9b8038e277cdb8cb87cf890e88b3456afa0b", size = 215854 }, - { url = "https://files.pythonhosted.org/packages/ce/b8/edab8a7d4fa14e924b29be877d54155dcbd8b80be85ea00d2be3413a9ed4/xxhash-3.6.0-cp314-cp314t-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:b9c6df83594f7df8f7f708ce5ebeacfc69f72c9fbaaababf6cf4758eaada0c9b", size = 214965 }, - { url = "https://files.pythonhosted.org/packages/27/67/dfa980ac7f0d509d54ea0d5a486d2bb4b80c3f1bb22b66e6a05d3efaf6c0/xxhash-3.6.0-cp314-cp314t-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:627f0af069b0ea56f312fd5189001c24578868643203bca1abbc2c52d3a6f3ca", size = 448484 }, - { url = "https://files.pythonhosted.org/packages/8c/63/8ffc2cc97e811c0ca5d00ab36604b3ea6f4254f20b7bc658ca825ce6c954/xxhash-3.6.0-cp314-cp314t-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:aa912c62f842dfd013c5f21a642c9c10cd9f4c4e943e0af83618b4a404d9091a", size = 196162 }, - { url = "https://files.pythonhosted.org/packages/4b/77/07f0e7a3edd11a6097e990f6e5b815b6592459cb16dae990d967693e6ea9/xxhash-3.6.0-cp314-cp314t-musllinux_1_2_aarch64.whl", hash = "sha256:b465afd7909db30168ab62afe40b2fcf79eedc0b89a6c0ab3123515dc0df8b99", size = 213007 }, - { url = "https://files.pythonhosted.org/packages/ae/d8/bc5fa0d152837117eb0bef6f83f956c509332ce133c91c63ce07ee7c4873/xxhash-3.6.0-cp314-cp314t-musllinux_1_2_i686.whl", hash = "sha256:a881851cf38b0a70e7c4d3ce81fc7afd86fbc2a024f4cfb2a97cf49ce04b75d3", size = 200956 }, - { url = "https://files.pythonhosted.org/packages/26/a5/d749334130de9411783873e9b98ecc46688dad5db64ca6e04b02acc8b473/xxhash-3.6.0-cp314-cp314t-musllinux_1_2_ppc64le.whl", hash = "sha256:9b3222c686a919a0f3253cfc12bb118b8b103506612253b5baeaac10d8027cf6", size = 213401 }, - { url = "https://files.pythonhosted.org/packages/89/72/abed959c956a4bfc72b58c0384bb7940663c678127538634d896b1195c10/xxhash-3.6.0-cp314-cp314t-musllinux_1_2_s390x.whl", hash = "sha256:c5aa639bc113e9286137cec8fadc20e9cd732b2cc385c0b7fa673b84fc1f2a93", size = 417083 }, - { url = "https://files.pythonhosted.org/packages/0c/b3/62fd2b586283b7d7d665fb98e266decadf31f058f1cf6c478741f68af0cb/xxhash-3.6.0-cp314-cp314t-musllinux_1_2_x86_64.whl", hash = "sha256:5c1343d49ac102799905e115aee590183c3921d475356cb24b4de29a4bc56518", size = 193913 }, - { url = "https://files.pythonhosted.org/packages/9a/9a/c19c42c5b3f5a4aad748a6d5b4f23df3bed7ee5445accc65a0fb3ff03953/xxhash-3.6.0-cp314-cp314t-win32.whl", hash = "sha256:5851f033c3030dd95c086b4a36a2683c2ff4a799b23af60977188b057e467119", size = 31586 }, - { url = "https://files.pythonhosted.org/packages/03/d6/4cc450345be9924fd5dc8c590ceda1db5b43a0a889587b0ae81a95511360/xxhash-3.6.0-cp314-cp314t-win_amd64.whl", hash = "sha256:0444e7967dac37569052d2409b00a8860c2135cff05502df4da80267d384849f", size = 32526 }, - { url = "https://files.pythonhosted.org/packages/0f/c9/7243eb3f9eaabd1a88a5a5acadf06df2d83b100c62684b7425c6a11bcaa8/xxhash-3.6.0-cp314-cp314t-win_arm64.whl", hash = "sha256:bb79b1e63f6fd84ec778a4b1916dfe0a7c3fdb986c06addd5db3a0d413819d95", size = 28898 }, - { url = "https://files.pythonhosted.org/packages/93/1e/8aec23647a34a249f62e2398c42955acd9b4c6ed5cf08cbea94dc46f78d2/xxhash-3.6.0-pp311-pypy311_pp73-macosx_10_15_x86_64.whl", hash = "sha256:0f7b7e2ec26c1666ad5fc9dbfa426a6a3367ceaf79db5dd76264659d509d73b0", size = 30662 }, - { url = "https://files.pythonhosted.org/packages/b8/0b/b14510b38ba91caf43006209db846a696ceea6a847a0c9ba0a5b1adc53d6/xxhash-3.6.0-pp311-pypy311_pp73-manylinux1_i686.manylinux_2_28_i686.manylinux_2_5_i686.whl", hash = "sha256:5dc1e14d14fa0f5789ec29a7062004b5933964bb9b02aae6622b8f530dc40296", size = 41056 }, - { url = "https://files.pythonhosted.org/packages/50/55/15a7b8a56590e66ccd374bbfa3f9ffc45b810886c8c3b614e3f90bd2367c/xxhash-3.6.0-pp311-pypy311_pp73-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:881b47fc47e051b37d94d13e7455131054b56749b91b508b0907eb07900d1c13", size = 36251 }, - { url = "https://files.pythonhosted.org/packages/62/b2/5ac99a041a29e58e95f907876b04f7067a0242cb85b5f39e726153981503/xxhash-3.6.0-pp311-pypy311_pp73-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:c6dc31591899f5e5666f04cc2e529e69b4072827085c1ef15294d91a004bc1bd", size = 32481 }, - { url = "https://files.pythonhosted.org/packages/7b/d9/8d95e906764a386a3d3b596f3c68bb63687dfca806373509f51ce8eea81f/xxhash-3.6.0-pp311-pypy311_pp73-win_amd64.whl", hash = "sha256:15e0dac10eb9309508bfc41f7f9deaa7755c69e35af835db9cb10751adebc35d", size = 31565 }, -] - -[[package]] -name = "yarl" -version = "1.22.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "idna" }, - { name = "multidict" }, - { name = "propcache" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/57/63/0c6ebca57330cd313f6102b16dd57ffaf3ec4c83403dcb45dbd15c6f3ea1/yarl-1.22.0.tar.gz", hash = "sha256:bebf8557577d4401ba8bd9ff33906f1376c877aa78d1fe216ad01b4d6745af71", size = 187169 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/d1/43/a2204825342f37c337f5edb6637040fa14e365b2fcc2346960201d457579/yarl-1.22.0-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:c7bd6683587567e5a49ee6e336e0612bec8329be1b7d4c8af5687dcdeb67ee1e", size = 140517 }, - { url = "https://files.pythonhosted.org/packages/44/6f/674f3e6f02266428c56f704cd2501c22f78e8b2eeb23f153117cc86fb28a/yarl-1.22.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:5cdac20da754f3a723cceea5b3448e1a2074866406adeb4ef35b469d089adb8f", size = 93495 }, - { url = "https://files.pythonhosted.org/packages/b8/12/5b274d8a0f30c07b91b2f02cba69152600b47830fcfb465c108880fcee9c/yarl-1.22.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:07a524d84df0c10f41e3ee918846e1974aba4ec017f990dc735aad487a0bdfdf", size = 94400 }, - { url = "https://files.pythonhosted.org/packages/e2/7f/df1b6949b1fa1aa9ff6de6e2631876ad4b73c4437822026e85d8acb56bb1/yarl-1.22.0-cp310-cp310-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:e1b329cb8146d7b736677a2440e422eadd775d1806a81db2d4cded80a48efc1a", size = 347545 }, - { url = "https://files.pythonhosted.org/packages/84/09/f92ed93bd6cd77872ab6c3462df45ca45cd058d8f1d0c9b4f54c1704429f/yarl-1.22.0-cp310-cp310-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:75976c6945d85dbb9ee6308cd7ff7b1fb9409380c82d6119bd778d8fcfe2931c", size = 319598 }, - { url = "https://files.pythonhosted.org/packages/c3/97/ac3f3feae7d522cf7ccec3d340bb0b2b61c56cb9767923df62a135092c6b/yarl-1.22.0-cp310-cp310-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:80ddf7a5f8c86cb3eb4bc9028b07bbbf1f08a96c5c0bc1244be5e8fefcb94147", size = 363893 }, - { url = "https://files.pythonhosted.org/packages/06/49/f3219097403b9c84a4d079b1d7bda62dd9b86d0d6e4428c02d46ab2c77fc/yarl-1.22.0-cp310-cp310-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:d332fc2e3c94dad927f2112395772a4e4fedbcf8f80efc21ed7cdfae4d574fdb", size = 371240 }, - { url = "https://files.pythonhosted.org/packages/35/9f/06b765d45c0e44e8ecf0fe15c9eacbbde342bb5b7561c46944f107bfb6c3/yarl-1.22.0-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:0cf71bf877efeac18b38d3930594c0948c82b64547c1cf420ba48722fe5509f6", size = 346965 }, - { url = "https://files.pythonhosted.org/packages/c5/69/599e7cea8d0fcb1694323b0db0dda317fa3162f7b90166faddecf532166f/yarl-1.22.0-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:663e1cadaddae26be034a6ab6072449a8426ddb03d500f43daf952b74553bba0", size = 342026 }, - { url = "https://files.pythonhosted.org/packages/95/6f/9dfd12c8bc90fea9eab39832ee32ea48f8e53d1256252a77b710c065c89f/yarl-1.22.0-cp310-cp310-musllinux_1_2_armv7l.whl", hash = "sha256:6dcbb0829c671f305be48a7227918cfcd11276c2d637a8033a99a02b67bf9eda", size = 335637 }, - { url = "https://files.pythonhosted.org/packages/57/2e/34c5b4eb9b07e16e873db5b182c71e5f06f9b5af388cdaa97736d79dd9a6/yarl-1.22.0-cp310-cp310-musllinux_1_2_ppc64le.whl", hash = "sha256:f0d97c18dfd9a9af4490631905a3f131a8e4c9e80a39353919e2cfed8f00aedc", size = 359082 }, - { url = "https://files.pythonhosted.org/packages/31/71/fa7e10fb772d273aa1f096ecb8ab8594117822f683bab7d2c5a89914c92a/yarl-1.22.0-cp310-cp310-musllinux_1_2_s390x.whl", hash = "sha256:437840083abe022c978470b942ff832c3940b2ad3734d424b7eaffcd07f76737", size = 357811 }, - { url = "https://files.pythonhosted.org/packages/26/da/11374c04e8e1184a6a03cf9c8f5688d3e5cec83ed6f31ad3481b3207f709/yarl-1.22.0-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:a899cbd98dce6f5d8de1aad31cb712ec0a530abc0a86bd6edaa47c1090138467", size = 351223 }, - { url = "https://files.pythonhosted.org/packages/82/8f/e2d01f161b0c034a30410e375e191a5d27608c1f8693bab1a08b089ca096/yarl-1.22.0-cp310-cp310-win32.whl", hash = "sha256:595697f68bd1f0c1c159fcb97b661fc9c3f5db46498043555d04805430e79bea", size = 82118 }, - { url = "https://files.pythonhosted.org/packages/62/46/94c76196642dbeae634c7a61ba3da88cd77bed875bf6e4a8bed037505aa6/yarl-1.22.0-cp310-cp310-win_amd64.whl", hash = "sha256:cb95a9b1adaa48e41815a55ae740cfda005758104049a640a398120bf02515ca", size = 86852 }, - { url = "https://files.pythonhosted.org/packages/af/af/7df4f179d3b1a6dcb9a4bd2ffbc67642746fcafdb62580e66876ce83fff4/yarl-1.22.0-cp310-cp310-win_arm64.whl", hash = "sha256:b85b982afde6df99ecc996990d4ad7ccbdbb70e2a4ba4de0aecde5922ba98a0b", size = 82012 }, - { url = "https://files.pythonhosted.org/packages/4d/27/5ab13fc84c76a0250afd3d26d5936349a35be56ce5785447d6c423b26d92/yarl-1.22.0-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:1ab72135b1f2db3fed3997d7e7dc1b80573c67138023852b6efb336a5eae6511", size = 141607 }, - { url = "https://files.pythonhosted.org/packages/6a/a1/d065d51d02dc02ce81501d476b9ed2229d9a990818332242a882d5d60340/yarl-1.22.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:669930400e375570189492dc8d8341301578e8493aec04aebc20d4717f899dd6", size = 94027 }, - { url = "https://files.pythonhosted.org/packages/c1/da/8da9f6a53f67b5106ffe902c6fa0164e10398d4e150d85838b82f424072a/yarl-1.22.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:792a2af6d58177ef7c19cbf0097aba92ca1b9cb3ffdd9c7470e156c8f9b5e028", size = 94963 }, - { url = "https://files.pythonhosted.org/packages/68/fe/2c1f674960c376e29cb0bec1249b117d11738db92a6ccc4a530b972648db/yarl-1.22.0-cp311-cp311-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:3ea66b1c11c9150f1372f69afb6b8116f2dd7286f38e14ea71a44eee9ec51b9d", size = 368406 }, - { url = "https://files.pythonhosted.org/packages/95/26/812a540e1c3c6418fec60e9bbd38e871eaba9545e94fa5eff8f4a8e28e1e/yarl-1.22.0-cp311-cp311-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:3e2daa88dc91870215961e96a039ec73e4937da13cf77ce17f9cad0c18df3503", size = 336581 }, - { url = "https://files.pythonhosted.org/packages/0b/f5/5777b19e26fdf98563985e481f8be3d8a39f8734147a6ebf459d0dab5a6b/yarl-1.22.0-cp311-cp311-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:ba440ae430c00eee41509353628600212112cd5018d5def7e9b05ea7ac34eb65", size = 388924 }, - { url = "https://files.pythonhosted.org/packages/86/08/24bd2477bd59c0bbd994fe1d93b126e0472e4e3df5a96a277b0a55309e89/yarl-1.22.0-cp311-cp311-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:e6438cc8f23a9c1478633d216b16104a586b9761db62bfacb6425bac0a36679e", size = 392890 }, - { url = "https://files.pythonhosted.org/packages/46/00/71b90ed48e895667ecfb1eaab27c1523ee2fa217433ed77a73b13205ca4b/yarl-1.22.0-cp311-cp311-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:4c52a6e78aef5cf47a98ef8e934755abf53953379b7d53e68b15ff4420e6683d", size = 365819 }, - { url = "https://files.pythonhosted.org/packages/30/2d/f715501cae832651d3282387c6a9236cd26bd00d0ff1e404b3dc52447884/yarl-1.22.0-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:3b06bcadaac49c70f4c88af4ffcfbe3dc155aab3163e75777818092478bcbbe7", size = 363601 }, - { url = "https://files.pythonhosted.org/packages/f8/f9/a678c992d78e394e7126ee0b0e4e71bd2775e4334d00a9278c06a6cce96a/yarl-1.22.0-cp311-cp311-musllinux_1_2_armv7l.whl", hash = "sha256:6944b2dc72c4d7f7052683487e3677456050ff77fcf5e6204e98caf785ad1967", size = 358072 }, - { url = "https://files.pythonhosted.org/packages/2c/d1/b49454411a60edb6fefdcad4f8e6dbba7d8019e3a508a1c5836cba6d0781/yarl-1.22.0-cp311-cp311-musllinux_1_2_ppc64le.whl", hash = "sha256:d5372ca1df0f91a86b047d1277c2aaf1edb32d78bbcefffc81b40ffd18f027ed", size = 385311 }, - { url = "https://files.pythonhosted.org/packages/87/e5/40d7a94debb8448c7771a916d1861d6609dddf7958dc381117e7ba36d9e8/yarl-1.22.0-cp311-cp311-musllinux_1_2_s390x.whl", hash = "sha256:51af598701f5299012b8416486b40fceef8c26fc87dc6d7d1f6fc30609ea0aa6", size = 381094 }, - { url = "https://files.pythonhosted.org/packages/35/d8/611cc282502381ad855448643e1ad0538957fc82ae83dfe7762c14069e14/yarl-1.22.0-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:b266bd01fedeffeeac01a79ae181719ff848a5a13ce10075adbefc8f1daee70e", size = 370944 }, - { url = "https://files.pythonhosted.org/packages/2d/df/fadd00fb1c90e1a5a8bd731fa3d3de2e165e5a3666a095b04e31b04d9cb6/yarl-1.22.0-cp311-cp311-win32.whl", hash = "sha256:a9b1ba5610a4e20f655258d5a1fdc7ebe3d837bb0e45b581398b99eb98b1f5ca", size = 81804 }, - { url = "https://files.pythonhosted.org/packages/b5/f7/149bb6f45f267cb5c074ac40c01c6b3ea6d8a620d34b337f6321928a1b4d/yarl-1.22.0-cp311-cp311-win_amd64.whl", hash = "sha256:078278b9b0b11568937d9509b589ee83ef98ed6d561dfe2020e24a9fd08eaa2b", size = 86858 }, - { url = "https://files.pythonhosted.org/packages/2b/13/88b78b93ad3f2f0b78e13bfaaa24d11cbc746e93fe76d8c06bf139615646/yarl-1.22.0-cp311-cp311-win_arm64.whl", hash = "sha256:b6a6f620cfe13ccec221fa312139135166e47ae169f8253f72a0abc0dae94376", size = 81637 }, - { url = "https://files.pythonhosted.org/packages/75/ff/46736024fee3429b80a165a732e38e5d5a238721e634ab41b040d49f8738/yarl-1.22.0-cp312-cp312-macosx_10_13_universal2.whl", hash = "sha256:e340382d1afa5d32b892b3ff062436d592ec3d692aeea3bef3a5cfe11bbf8c6f", size = 142000 }, - { url = "https://files.pythonhosted.org/packages/5a/9a/b312ed670df903145598914770eb12de1bac44599549b3360acc96878df8/yarl-1.22.0-cp312-cp312-macosx_10_13_x86_64.whl", hash = "sha256:f1e09112a2c31ffe8d80be1b0988fa6a18c5d5cad92a9ffbb1c04c91bfe52ad2", size = 94338 }, - { url = "https://files.pythonhosted.org/packages/ba/f5/0601483296f09c3c65e303d60c070a5c19fcdbc72daa061e96170785bc7d/yarl-1.22.0-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:939fe60db294c786f6b7c2d2e121576628468f65453d86b0fe36cb52f987bd74", size = 94909 }, - { url = "https://files.pythonhosted.org/packages/60/41/9a1fe0b73dbcefce72e46cf149b0e0a67612d60bfc90fb59c2b2efdfbd86/yarl-1.22.0-cp312-cp312-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:e1651bf8e0398574646744c1885a41198eba53dc8a9312b954073f845c90a8df", size = 372940 }, - { url = "https://files.pythonhosted.org/packages/17/7a/795cb6dfee561961c30b800f0ed616b923a2ec6258b5def2a00bf8231334/yarl-1.22.0-cp312-cp312-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:b8a0588521a26bf92a57a1705b77b8b59044cdceccac7151bd8d229e66b8dedb", size = 345825 }, - { url = "https://files.pythonhosted.org/packages/d7/93/a58f4d596d2be2ae7bab1a5846c4d270b894958845753b2c606d666744d3/yarl-1.22.0-cp312-cp312-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:42188e6a615c1a75bcaa6e150c3fe8f3e8680471a6b10150c5f7e83f47cc34d2", size = 386705 }, - { url = "https://files.pythonhosted.org/packages/61/92/682279d0e099d0e14d7fd2e176bd04f48de1484f56546a3e1313cd6c8e7c/yarl-1.22.0-cp312-cp312-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:f6d2cb59377d99718913ad9a151030d6f83ef420a2b8f521d94609ecc106ee82", size = 396518 }, - { url = "https://files.pythonhosted.org/packages/db/0f/0d52c98b8a885aeda831224b78f3be7ec2e1aa4a62091f9f9188c3c65b56/yarl-1.22.0-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:50678a3b71c751d58d7908edc96d332af328839eea883bb554a43f539101277a", size = 377267 }, - { url = "https://files.pythonhosted.org/packages/22/42/d2685e35908cbeaa6532c1fc73e89e7f2efb5d8a7df3959ea8e37177c5a3/yarl-1.22.0-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:1e8fbaa7cec507aa24ea27a01456e8dd4b6fab829059b69844bd348f2d467124", size = 365797 }, - { url = "https://files.pythonhosted.org/packages/a2/83/cf8c7bcc6355631762f7d8bdab920ad09b82efa6b722999dfb05afa6cfac/yarl-1.22.0-cp312-cp312-musllinux_1_2_armv7l.whl", hash = "sha256:433885ab5431bc3d3d4f2f9bd15bfa1614c522b0f1405d62c4f926ccd69d04fa", size = 365535 }, - { url = "https://files.pythonhosted.org/packages/25/e1/5302ff9b28f0c59cac913b91fe3f16c59a033887e57ce9ca5d41a3a94737/yarl-1.22.0-cp312-cp312-musllinux_1_2_ppc64le.whl", hash = "sha256:b790b39c7e9a4192dc2e201a282109ed2985a1ddbd5ac08dc56d0e121400a8f7", size = 382324 }, - { url = "https://files.pythonhosted.org/packages/bf/cd/4617eb60f032f19ae3a688dc990d8f0d89ee0ea378b61cac81ede3e52fae/yarl-1.22.0-cp312-cp312-musllinux_1_2_s390x.whl", hash = "sha256:31f0b53913220599446872d757257be5898019c85e7971599065bc55065dc99d", size = 383803 }, - { url = "https://files.pythonhosted.org/packages/59/65/afc6e62bb506a319ea67b694551dab4a7e6fb7bf604e9bd9f3e11d575fec/yarl-1.22.0-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:a49370e8f711daec68d09b821a34e1167792ee2d24d405cbc2387be4f158b520", size = 374220 }, - { url = "https://files.pythonhosted.org/packages/e7/3d/68bf18d50dc674b942daec86a9ba922d3113d8399b0e52b9897530442da2/yarl-1.22.0-cp312-cp312-win32.whl", hash = "sha256:70dfd4f241c04bd9239d53b17f11e6ab672b9f1420364af63e8531198e3f5fe8", size = 81589 }, - { url = "https://files.pythonhosted.org/packages/c8/9a/6ad1a9b37c2f72874f93e691b2e7ecb6137fb2b899983125db4204e47575/yarl-1.22.0-cp312-cp312-win_amd64.whl", hash = "sha256:8884d8b332a5e9b88e23f60bb166890009429391864c685e17bd73a9eda9105c", size = 87213 }, - { url = "https://files.pythonhosted.org/packages/44/c5/c21b562d1680a77634d748e30c653c3ca918beb35555cff24986fff54598/yarl-1.22.0-cp312-cp312-win_arm64.whl", hash = "sha256:ea70f61a47f3cc93bdf8b2f368ed359ef02a01ca6393916bc8ff877427181e74", size = 81330 }, - { url = "https://files.pythonhosted.org/packages/ea/f3/d67de7260456ee105dc1d162d43a019ecad6b91e2f51809d6cddaa56690e/yarl-1.22.0-cp313-cp313-macosx_10_13_universal2.whl", hash = "sha256:8dee9c25c74997f6a750cd317b8ca63545169c098faee42c84aa5e506c819b53", size = 139980 }, - { url = "https://files.pythonhosted.org/packages/01/88/04d98af0b47e0ef42597b9b28863b9060bb515524da0a65d5f4db160b2d5/yarl-1.22.0-cp313-cp313-macosx_10_13_x86_64.whl", hash = "sha256:01e73b85a5434f89fc4fe27dcda2aff08ddf35e4d47bbbea3bdcd25321af538a", size = 93424 }, - { url = "https://files.pythonhosted.org/packages/18/91/3274b215fd8442a03975ce6bee5fe6aa57a8326b29b9d3d56234a1dca244/yarl-1.22.0-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:22965c2af250d20c873cdbee8ff958fb809940aeb2e74ba5f20aaf6b7ac8c70c", size = 93821 }, - { url = "https://files.pythonhosted.org/packages/61/3a/caf4e25036db0f2da4ca22a353dfeb3c9d3c95d2761ebe9b14df8fc16eb0/yarl-1.22.0-cp313-cp313-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:b4f15793aa49793ec8d1c708ab7f9eded1aa72edc5174cae703651555ed1b601", size = 373243 }, - { url = "https://files.pythonhosted.org/packages/6e/9e/51a77ac7516e8e7803b06e01f74e78649c24ee1021eca3d6a739cb6ea49c/yarl-1.22.0-cp313-cp313-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:e5542339dcf2747135c5c85f68680353d5cb9ffd741c0f2e8d832d054d41f35a", size = 342361 }, - { url = "https://files.pythonhosted.org/packages/d4/f8/33b92454789dde8407f156c00303e9a891f1f51a0330b0fad7c909f87692/yarl-1.22.0-cp313-cp313-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:5c401e05ad47a75869c3ab3e35137f8468b846770587e70d71e11de797d113df", size = 387036 }, - { url = "https://files.pythonhosted.org/packages/d9/9a/c5db84ea024f76838220280f732970aa4ee154015d7f5c1bfb60a267af6f/yarl-1.22.0-cp313-cp313-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:243dda95d901c733f5b59214d28b0120893d91777cb8aa043e6ef059d3cddfe2", size = 397671 }, - { url = "https://files.pythonhosted.org/packages/11/c9/cd8538dc2e7727095e0c1d867bad1e40c98f37763e6d995c1939f5fdc7b1/yarl-1.22.0-cp313-cp313-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:bec03d0d388060058f5d291a813f21c011041938a441c593374da6077fe21b1b", size = 377059 }, - { url = "https://files.pythonhosted.org/packages/a1/b9/ab437b261702ced75122ed78a876a6dec0a1b0f5e17a4ac7a9a2482d8abe/yarl-1.22.0-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:b0748275abb8c1e1e09301ee3cf90c8a99678a4e92e4373705f2a2570d581273", size = 365356 }, - { url = "https://files.pythonhosted.org/packages/b2/9d/8e1ae6d1d008a9567877b08f0ce4077a29974c04c062dabdb923ed98e6fe/yarl-1.22.0-cp313-cp313-musllinux_1_2_armv7l.whl", hash = "sha256:47fdb18187e2a4e18fda2c25c05d8251a9e4a521edaed757fef033e7d8498d9a", size = 361331 }, - { url = "https://files.pythonhosted.org/packages/ca/5a/09b7be3905962f145b73beb468cdd53db8aa171cf18c80400a54c5b82846/yarl-1.22.0-cp313-cp313-musllinux_1_2_ppc64le.whl", hash = "sha256:c7044802eec4524fde550afc28edda0dd5784c4c45f0be151a2d3ba017daca7d", size = 382590 }, - { url = "https://files.pythonhosted.org/packages/aa/7f/59ec509abf90eda5048b0bc3e2d7b5099dffdb3e6b127019895ab9d5ef44/yarl-1.22.0-cp313-cp313-musllinux_1_2_s390x.whl", hash = "sha256:139718f35149ff544caba20fce6e8a2f71f1e39b92c700d8438a0b1d2a631a02", size = 385316 }, - { url = "https://files.pythonhosted.org/packages/e5/84/891158426bc8036bfdfd862fabd0e0fa25df4176ec793e447f4b85cf1be4/yarl-1.22.0-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:e1b51bebd221006d3d2f95fbe124b22b247136647ae5dcc8c7acafba66e5ee67", size = 374431 }, - { url = "https://files.pythonhosted.org/packages/bb/49/03da1580665baa8bef5e8ed34c6df2c2aca0a2f28bf397ed238cc1bbc6f2/yarl-1.22.0-cp313-cp313-win32.whl", hash = "sha256:d3e32536234a95f513bd374e93d717cf6b2231a791758de6c509e3653f234c95", size = 81555 }, - { url = "https://files.pythonhosted.org/packages/9a/ee/450914ae11b419eadd067c6183ae08381cfdfcb9798b90b2b713bbebddda/yarl-1.22.0-cp313-cp313-win_amd64.whl", hash = "sha256:47743b82b76d89a1d20b83e60d5c20314cbd5ba2befc9cda8f28300c4a08ed4d", size = 86965 }, - { url = "https://files.pythonhosted.org/packages/98/4d/264a01eae03b6cf629ad69bae94e3b0e5344741e929073678e84bf7a3e3b/yarl-1.22.0-cp313-cp313-win_arm64.whl", hash = "sha256:5d0fcda9608875f7d052eff120c7a5da474a6796fe4d83e152e0e4d42f6d1a9b", size = 81205 }, - { url = "https://files.pythonhosted.org/packages/88/fc/6908f062a2f77b5f9f6d69cecb1747260831ff206adcbc5b510aff88df91/yarl-1.22.0-cp313-cp313t-macosx_10_13_universal2.whl", hash = "sha256:719ae08b6972befcba4310e49edb1161a88cdd331e3a694b84466bd938a6ab10", size = 146209 }, - { url = "https://files.pythonhosted.org/packages/65/47/76594ae8eab26210b4867be6f49129861ad33da1f1ebdf7051e98492bf62/yarl-1.22.0-cp313-cp313t-macosx_10_13_x86_64.whl", hash = "sha256:47d8a5c446df1c4db9d21b49619ffdba90e77c89ec6e283f453856c74b50b9e3", size = 95966 }, - { url = "https://files.pythonhosted.org/packages/ab/ce/05e9828a49271ba6b5b038b15b3934e996980dd78abdfeb52a04cfb9467e/yarl-1.22.0-cp313-cp313t-macosx_11_0_arm64.whl", hash = "sha256:cfebc0ac8333520d2d0423cbbe43ae43c8838862ddb898f5ca68565e395516e9", size = 97312 }, - { url = "https://files.pythonhosted.org/packages/d1/c5/7dffad5e4f2265b29c9d7ec869c369e4223166e4f9206fc2243ee9eea727/yarl-1.22.0-cp313-cp313t-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:4398557cbf484207df000309235979c79c4356518fd5c99158c7d38203c4da4f", size = 361967 }, - { url = "https://files.pythonhosted.org/packages/50/b2/375b933c93a54bff7fc041e1a6ad2c0f6f733ffb0c6e642ce56ee3b39970/yarl-1.22.0-cp313-cp313t-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:2ca6fd72a8cd803be290d42f2dec5cdcd5299eeb93c2d929bf060ad9efaf5de0", size = 323949 }, - { url = "https://files.pythonhosted.org/packages/66/50/bfc2a29a1d78644c5a7220ce2f304f38248dc94124a326794e677634b6cf/yarl-1.22.0-cp313-cp313t-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:ca1f59c4e1ab6e72f0a23c13fca5430f889634166be85dbf1013683e49e3278e", size = 361818 }, - { url = "https://files.pythonhosted.org/packages/46/96/f3941a46af7d5d0f0498f86d71275696800ddcdd20426298e572b19b91ff/yarl-1.22.0-cp313-cp313t-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:6c5010a52015e7c70f86eb967db0f37f3c8bd503a695a49f8d45700144667708", size = 372626 }, - { url = "https://files.pythonhosted.org/packages/c1/42/8b27c83bb875cd89448e42cd627e0fb971fa1675c9ec546393d18826cb50/yarl-1.22.0-cp313-cp313t-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:9d7672ecf7557476642c88497c2f8d8542f8e36596e928e9bcba0e42e1e7d71f", size = 341129 }, - { url = "https://files.pythonhosted.org/packages/49/36/99ca3122201b382a3cf7cc937b95235b0ac944f7e9f2d5331d50821ed352/yarl-1.22.0-cp313-cp313t-musllinux_1_2_aarch64.whl", hash = "sha256:3b7c88eeef021579d600e50363e0b6ee4f7f6f728cd3486b9d0f3ee7b946398d", size = 346776 }, - { url = "https://files.pythonhosted.org/packages/85/b4/47328bf996acd01a4c16ef9dcd2f59c969f495073616586f78cd5f2efb99/yarl-1.22.0-cp313-cp313t-musllinux_1_2_armv7l.whl", hash = "sha256:f4afb5c34f2c6fecdcc182dfcfc6af6cccf1aa923eed4d6a12e9d96904e1a0d8", size = 334879 }, - { url = "https://files.pythonhosted.org/packages/c2/ad/b77d7b3f14a4283bffb8e92c6026496f6de49751c2f97d4352242bba3990/yarl-1.22.0-cp313-cp313t-musllinux_1_2_ppc64le.whl", hash = "sha256:59c189e3e99a59cf8d83cbb31d4db02d66cda5a1a4374e8a012b51255341abf5", size = 350996 }, - { url = "https://files.pythonhosted.org/packages/81/c8/06e1d69295792ba54d556f06686cbd6a7ce39c22307100e3fb4a2c0b0a1d/yarl-1.22.0-cp313-cp313t-musllinux_1_2_s390x.whl", hash = "sha256:5a3bf7f62a289fa90f1990422dc8dff5a458469ea71d1624585ec3a4c8d6960f", size = 356047 }, - { url = "https://files.pythonhosted.org/packages/4b/b8/4c0e9e9f597074b208d18cef227d83aac36184bfbc6eab204ea55783dbc5/yarl-1.22.0-cp313-cp313t-musllinux_1_2_x86_64.whl", hash = "sha256:de6b9a04c606978fdfe72666fa216ffcf2d1a9f6a381058d4378f8d7b1e5de62", size = 342947 }, - { url = "https://files.pythonhosted.org/packages/e0/e5/11f140a58bf4c6ad7aca69a892bff0ee638c31bea4206748fc0df4ebcb3a/yarl-1.22.0-cp313-cp313t-win32.whl", hash = "sha256:1834bb90991cc2999f10f97f5f01317f99b143284766d197e43cd5b45eb18d03", size = 86943 }, - { url = "https://files.pythonhosted.org/packages/31/74/8b74bae38ed7fe6793d0c15a0c8207bbb819cf287788459e5ed230996cdd/yarl-1.22.0-cp313-cp313t-win_amd64.whl", hash = "sha256:ff86011bd159a9d2dfc89c34cfd8aff12875980e3bd6a39ff097887520e60249", size = 93715 }, - { url = "https://files.pythonhosted.org/packages/69/66/991858aa4b5892d57aef7ee1ba6b4d01ec3b7eb3060795d34090a3ca3278/yarl-1.22.0-cp313-cp313t-win_arm64.whl", hash = "sha256:7861058d0582b847bc4e3a4a4c46828a410bca738673f35a29ba3ca5db0b473b", size = 83857 }, - { url = "https://files.pythonhosted.org/packages/46/b3/e20ef504049f1a1c54a814b4b9bed96d1ac0e0610c3b4da178f87209db05/yarl-1.22.0-cp314-cp314-macosx_10_13_universal2.whl", hash = "sha256:34b36c2c57124530884d89d50ed2c1478697ad7473efd59cfd479945c95650e4", size = 140520 }, - { url = "https://files.pythonhosted.org/packages/e4/04/3532d990fdbab02e5ede063676b5c4260e7f3abea2151099c2aa745acc4c/yarl-1.22.0-cp314-cp314-macosx_10_13_x86_64.whl", hash = "sha256:0dd9a702591ca2e543631c2a017e4a547e38a5c0f29eece37d9097e04a7ac683", size = 93504 }, - { url = "https://files.pythonhosted.org/packages/11/63/ff458113c5c2dac9a9719ac68ee7c947cb621432bcf28c9972b1c0e83938/yarl-1.22.0-cp314-cp314-macosx_11_0_arm64.whl", hash = "sha256:594fcab1032e2d2cc3321bb2e51271e7cd2b516c7d9aee780ece81b07ff8244b", size = 94282 }, - { url = "https://files.pythonhosted.org/packages/a7/bc/315a56aca762d44a6aaaf7ad253f04d996cb6b27bad34410f82d76ea8038/yarl-1.22.0-cp314-cp314-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:f3d7a87a78d46a2e3d5b72587ac14b4c16952dd0887dbb051451eceac774411e", size = 372080 }, - { url = "https://files.pythonhosted.org/packages/3f/3f/08e9b826ec2e099ea6e7c69a61272f4f6da62cb5b1b63590bb80ca2e4a40/yarl-1.22.0-cp314-cp314-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:852863707010316c973162e703bddabec35e8757e67fcb8ad58829de1ebc8590", size = 338696 }, - { url = "https://files.pythonhosted.org/packages/e3/9f/90360108e3b32bd76789088e99538febfea24a102380ae73827f62073543/yarl-1.22.0-cp314-cp314-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:131a085a53bfe839a477c0845acf21efc77457ba2bcf5899618136d64f3303a2", size = 387121 }, - { url = "https://files.pythonhosted.org/packages/98/92/ab8d4657bd5b46a38094cfaea498f18bb70ce6b63508fd7e909bd1f93066/yarl-1.22.0-cp314-cp314-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:078a8aefd263f4d4f923a9677b942b445a2be970ca24548a8102689a3a8ab8da", size = 394080 }, - { url = "https://files.pythonhosted.org/packages/f5/e7/d8c5a7752fef68205296201f8ec2bf718f5c805a7a7e9880576c67600658/yarl-1.22.0-cp314-cp314-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:bca03b91c323036913993ff5c738d0842fc9c60c4648e5c8d98331526df89784", size = 372661 }, - { url = "https://files.pythonhosted.org/packages/b6/2e/f4d26183c8db0bb82d491b072f3127fb8c381a6206a3a56332714b79b751/yarl-1.22.0-cp314-cp314-musllinux_1_2_aarch64.whl", hash = "sha256:68986a61557d37bb90d3051a45b91fa3d5c516d177dfc6dd6f2f436a07ff2b6b", size = 364645 }, - { url = "https://files.pythonhosted.org/packages/80/7c/428e5812e6b87cd00ee8e898328a62c95825bf37c7fa87f0b6bb2ad31304/yarl-1.22.0-cp314-cp314-musllinux_1_2_armv7l.whl", hash = "sha256:4792b262d585ff0dff6bcb787f8492e40698443ec982a3568c2096433660c694", size = 355361 }, - { url = "https://files.pythonhosted.org/packages/ec/2a/249405fd26776f8b13c067378ef4d7dd49c9098d1b6457cdd152a99e96a9/yarl-1.22.0-cp314-cp314-musllinux_1_2_ppc64le.whl", hash = "sha256:ebd4549b108d732dba1d4ace67614b9545b21ece30937a63a65dd34efa19732d", size = 381451 }, - { url = "https://files.pythonhosted.org/packages/67/a8/fb6b1adbe98cf1e2dd9fad71003d3a63a1bc22459c6e15f5714eb9323b93/yarl-1.22.0-cp314-cp314-musllinux_1_2_s390x.whl", hash = "sha256:f87ac53513d22240c7d59203f25cc3beac1e574c6cd681bbfd321987b69f95fd", size = 383814 }, - { url = "https://files.pythonhosted.org/packages/d9/f9/3aa2c0e480fb73e872ae2814c43bc1e734740bb0d54e8cb2a95925f98131/yarl-1.22.0-cp314-cp314-musllinux_1_2_x86_64.whl", hash = "sha256:22b029f2881599e2f1b06f8f1db2ee63bd309e2293ba2d566e008ba12778b8da", size = 370799 }, - { url = "https://files.pythonhosted.org/packages/50/3c/af9dba3b8b5eeb302f36f16f92791f3ea62e3f47763406abf6d5a4a3333b/yarl-1.22.0-cp314-cp314-win32.whl", hash = "sha256:6a635ea45ba4ea8238463b4f7d0e721bad669f80878b7bfd1f89266e2ae63da2", size = 82990 }, - { url = "https://files.pythonhosted.org/packages/ac/30/ac3a0c5bdc1d6efd1b41fa24d4897a4329b3b1e98de9449679dd327af4f0/yarl-1.22.0-cp314-cp314-win_amd64.whl", hash = "sha256:0d6e6885777af0f110b0e5d7e5dda8b704efed3894da26220b7f3d887b839a79", size = 88292 }, - { url = "https://files.pythonhosted.org/packages/df/0a/227ab4ff5b998a1b7410abc7b46c9b7a26b0ca9e86c34ba4b8d8bc7c63d5/yarl-1.22.0-cp314-cp314-win_arm64.whl", hash = "sha256:8218f4e98d3c10d683584cb40f0424f4b9fd6e95610232dd75e13743b070ee33", size = 82888 }, - { url = "https://files.pythonhosted.org/packages/06/5e/a15eb13db90abd87dfbefb9760c0f3f257ac42a5cac7e75dbc23bed97a9f/yarl-1.22.0-cp314-cp314t-macosx_10_13_universal2.whl", hash = "sha256:45c2842ff0e0d1b35a6bf1cd6c690939dacb617a70827f715232b2e0494d55d1", size = 146223 }, - { url = "https://files.pythonhosted.org/packages/18/82/9665c61910d4d84f41a5bf6837597c89e665fa88aa4941080704645932a9/yarl-1.22.0-cp314-cp314t-macosx_10_13_x86_64.whl", hash = "sha256:d947071e6ebcf2e2bee8fce76e10faca8f7a14808ca36a910263acaacef08eca", size = 95981 }, - { url = "https://files.pythonhosted.org/packages/5d/9a/2f65743589809af4d0a6d3aa749343c4b5f4c380cc24a8e94a3c6625a808/yarl-1.22.0-cp314-cp314t-macosx_11_0_arm64.whl", hash = "sha256:334b8721303e61b00019474cc103bdac3d7b1f65e91f0bfedeec2d56dfe74b53", size = 97303 }, - { url = "https://files.pythonhosted.org/packages/b0/ab/5b13d3e157505c43c3b43b5a776cbf7b24a02bc4cccc40314771197e3508/yarl-1.22.0-cp314-cp314t-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:1e7ce67c34138a058fd092f67d07a72b8e31ff0c9236e751957465a24b28910c", size = 361820 }, - { url = "https://files.pythonhosted.org/packages/fb/76/242a5ef4677615cf95330cfc1b4610e78184400699bdda0acb897ef5e49a/yarl-1.22.0-cp314-cp314t-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:d77e1b2c6d04711478cb1c4ab90db07f1609ccf06a287d5607fcd90dc9863acf", size = 323203 }, - { url = "https://files.pythonhosted.org/packages/8c/96/475509110d3f0153b43d06164cf4195c64d16999e0c7e2d8a099adcd6907/yarl-1.22.0-cp314-cp314t-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:c4647674b6150d2cae088fc07de2738a84b8bcedebef29802cf0b0a82ab6face", size = 363173 }, - { url = "https://files.pythonhosted.org/packages/c9/66/59db471aecfbd559a1fd48aedd954435558cd98c7d0da8b03cc6c140a32c/yarl-1.22.0-cp314-cp314t-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:efb07073be061c8f79d03d04139a80ba33cbd390ca8f0297aae9cce6411e4c6b", size = 373562 }, - { url = "https://files.pythonhosted.org/packages/03/1f/c5d94abc91557384719da10ff166b916107c1b45e4d0423a88457071dd88/yarl-1.22.0-cp314-cp314t-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:e51ac5435758ba97ad69617e13233da53908beccc6cfcd6c34bbed8dcbede486", size = 339828 }, - { url = "https://files.pythonhosted.org/packages/5f/97/aa6a143d3afba17b6465733681c70cf175af89f76ec8d9286e08437a7454/yarl-1.22.0-cp314-cp314t-musllinux_1_2_aarch64.whl", hash = "sha256:33e32a0dd0c8205efa8e83d04fc9f19313772b78522d1bdc7d9aed706bfd6138", size = 347551 }, - { url = "https://files.pythonhosted.org/packages/43/3c/45a2b6d80195959239a7b2a8810506d4eea5487dce61c2a3393e7fc3c52e/yarl-1.22.0-cp314-cp314t-musllinux_1_2_armv7l.whl", hash = "sha256:bf4a21e58b9cde0e401e683ebd00f6ed30a06d14e93f7c8fd059f8b6e8f87b6a", size = 334512 }, - { url = "https://files.pythonhosted.org/packages/86/a0/c2ab48d74599c7c84cb104ebd799c5813de252bea0f360ffc29d270c2caa/yarl-1.22.0-cp314-cp314t-musllinux_1_2_ppc64le.whl", hash = "sha256:e4b582bab49ac33c8deb97e058cd67c2c50dac0dd134874106d9c774fd272529", size = 352400 }, - { url = "https://files.pythonhosted.org/packages/32/75/f8919b2eafc929567d3d8411f72bdb1a2109c01caaab4ebfa5f8ffadc15b/yarl-1.22.0-cp314-cp314t-musllinux_1_2_s390x.whl", hash = "sha256:0b5bcc1a9c4839e7e30b7b30dd47fe5e7e44fb7054ec29b5bb8d526aa1041093", size = 357140 }, - { url = "https://files.pythonhosted.org/packages/cf/72/6a85bba382f22cf78add705d8c3731748397d986e197e53ecc7835e76de7/yarl-1.22.0-cp314-cp314t-musllinux_1_2_x86_64.whl", hash = "sha256:c0232bce2170103ec23c454e54a57008a9a72b5d1c3105dc2496750da8cfa47c", size = 341473 }, - { url = "https://files.pythonhosted.org/packages/35/18/55e6011f7c044dc80b98893060773cefcfdbf60dfefb8cb2f58b9bacbd83/yarl-1.22.0-cp314-cp314t-win32.whl", hash = "sha256:8009b3173bcd637be650922ac455946197d858b3630b6d8787aa9e5c4564533e", size = 89056 }, - { url = "https://files.pythonhosted.org/packages/f9/86/0f0dccb6e59a9e7f122c5afd43568b1d31b8ab7dda5f1b01fb5c7025c9a9/yarl-1.22.0-cp314-cp314t-win_amd64.whl", hash = "sha256:9fb17ea16e972c63d25d4a97f016d235c78dd2344820eb35bc034bc32012ee27", size = 96292 }, - { url = "https://files.pythonhosted.org/packages/48/b7/503c98092fb3b344a179579f55814b613c1fbb1c23b3ec14a7b008a66a6e/yarl-1.22.0-cp314-cp314t-win_arm64.whl", hash = "sha256:9f6d73c1436b934e3f01df1e1b21ff765cd1d28c77dfb9ace207f746d4610ee1", size = 85171 }, - { url = "https://files.pythonhosted.org/packages/73/ae/b48f95715333080afb75a4504487cbe142cae1268afc482d06692d605ae6/yarl-1.22.0-py3-none-any.whl", hash = "sha256:1380560bdba02b6b6c90de54133c81c9f2a453dee9912fe58c1dcced1edb7cff", size = 46814 }, -] diff --git a/vocab.json b/vocab.json index 100958bb7a5a7686628b02a908ce73a7157a299f..33b95831722563db5362641673b37885e25fb47a 100644 --- a/vocab.json +++ b/vocab.json @@ -1 +1 @@ -{"0": "", "1": "\u2581t", "2": "\u2581th", "3": "\u2581a", "4": "in", "5": "re", "6": "\u2581the", "7": "\u2581w", "8": "\u2581s", "9": "\u2581o", "10": "er", "11": "ou", "12": "at", "13": "nd", "14": "it", "15": "\u2581h", "16": "\u2581c", "17": "\u2581b", "18": "is", "19": "en", "20": "on", "21": "ing", "22": "\u2581f", "23": "\u2581to", "24": "\u2581m", "25": "es", "26": "\u2581p", "27": "or", "28": "an", "29": "\u2581d", "30": "ll", "31": "\u2581I", "32": "ed", "33": "\u2581and", "34": "\u2581l", "35": "\u2581of", "36": "\u2581in", "37": "\u2581y", "38": "ar", "39": "\u2581g", "40": "\u2581you", "41": "as", "42": "om", "43": "\u2581n", "44": "ve", "45": "\u2581that", "46": "le", "47": "ic", "48": "us", "49": "ow", "50": "et", "51": "al", "52": "\u2581e", "53": "ut", "54": "\u2581it", "55": "ot", "56": "\u2581be", "57": "\u2581T", "58": "ion", "59": "\u2581is", "60": "\u2581wh", "61": "\u2581re", "62": "\u2581on", "63": "\u2581we", "64": "ent", "65": "\u2581A", "66": "ay", "67": "\u2581ha", "68": "\u2581Th", "69": "id", "70": "\u2581S", "71": "ac", "72": "gh", "73": "ver", "74": "ke", "75": "\u2581for", "76": "im", "77": "ly", "78": "ur", "79": "ld", "80": "\u2581he", "81": "\u2581st", "82": "all", "83": "ro", "84": "st", "85": "se", "86": "ct", "87": "ith", "88": "ir", "89": "am", "90": "\u2581this", "91": "if", "92": "\u2581W", "93": "oo", "94": "ri", "95": "\u2581was", "96": "ght", "97": "\u2581u", "98": "\u2581with", "99": "ad", "100": "ch", "101": "\u2581se", "102": "\u2581k", "103": "\u2581an", "104": "\u2581The", "105": "\u2581li", "106": "\u2581do", "107": "\u2581B", "108": "\u2581have", "109": "\u2581as", "110": "th", "111": "\u2581are", "112": "\u2581sh", "113": "ust", "114": "ce", "115": "ally", "116": "ill", "117": "\u2581H", "118": "\u2581j", "119": "ter", "120": "\u2581go", "121": "\u2581And", "122": "ation", "123": "\u2581C", "124": "\u2581so", "125": "ome", "126": "\u2581not", "127": "op", "128": "il", "129": "ore", "130": "\u2581ne", "131": "\u2581can", "132": "\u2581me", "133": "\u2581at", "134": "ould", "135": "ant", "136": "\u2581M", "137": "\u2581like", "138": "ere", "139": "\u2581they", "140": "ra", "141": "ers", "142": "\u2581ab", "143": "\u2581de", "144": "\u2581kn", "145": "ge", "146": "\u2581Y", "147": "\u2581ch", "148": "ul", "149": "pp", "150": "\u2581or", "151": "\u2581al", "152": "\u2581con", "153": "\u2581com", "154": "ess", "155": "\u2581su", "156": "out", "157": "\u2581your", "158": "\u2581So", "159": "ate", "160": "\u2581one", "161": "\u2581all", "162": "\u2581ex", "163": "est", "164": "\u2581fr", "165": "\u2581just", "166": "\u2581pro", "167": "\u2581know", "168": "\u2581O", "169": "ain", "170": "\u2581but", "171": "ol", "172": "ive", "173": "\u2581v", "174": "use", "175": "very", "176": "art", "177": "qu", "178": "\u2581my", "179": "el", "180": "\u2581N", "181": "nt", "182": "\u2581It", "183": "\u2581what", "184": "ab", "185": "\u2581P", "186": "\u2581wor", "187": "\u2581out", "188": "\u2581there", "189": "\u2581up", "190": "um", "191": "\u2581from", "192": "pe", "193": "\u2581tw", "194": "\u2581r", "195": "and", "196": "ight", "197": "ort", "198": "un", "199": "\u2581L", "200": "ist", "201": "\u2581about", "202": "ide", "203": "ig", "204": "ake", "205": "\u2581D", "206": "em", "207": "os", "208": "king", "209": "rou", "210": "ind", "211": "our", "212": "res", "213": "\u2581We", "214": "\u2581get", "215": "\u2581E", "216": "\u2581G", "217": "ack", "218": "\u2581le", "219": "ity", "220": "od", "221": "\u2581F", "222": "ard", "223": "\u2581pl", "224": "\u2581our", "225": "\u2581int", "226": "ment", "227": "\u2581will", "228": "ies", "229": "\u2581by", "230": "ink", "231": "ca", "232": "\u2581if", "233": "red", "234": "her", "235": "ie", "236": "\u2581us", "237": "\u2581some", "238": "\u2581don", "239": "ven", "240": "ood", "241": "ast", "242": "\u2581R", "243": "\u2581his", "244": "\u2581tim", "245": "\u2581tr", "246": "\u2581more", "247": "ich", "248": "ous", "249": "ame", "250": "\u2581going", "251": "\u2581had", "252": "\u2581them", "253": "ook", "254": "\u2581pe", "255": "\u2581Wh", "256": "\u2581You", "257": "\u2581But", "258": "ine", "259": "\u2581here", "260": "\u2581would", "261": "cause", "262": "right", "263": "so", "264": "ost", "265": "ure", "266": "\u2581has", "267": "ect", "268": "\u2581think", "269": "\u2581fe", "270": "ong", "271": "\u2581see", "272": "\u2581when", "273": "\u2581who", "274": "\u2581were", "275": "\u2581really", "276": "\u2581their", "277": "\u2581want", "278": "one", "279": "ople", "280": "\u2581then", "281": "\u2581time", "282": "\u2581sa", "283": "ap", "284": "\u2581te", "285": "\u2581He", "286": "\u2581ye", "287": "ck", "288": "\u2581her", "289": "\u2581thing", "290": "\u2581right", "291": "\u2581which", "292": "itt", "293": "ice", "294": "act", "295": "\u2581people", "296": "ty", "297": "\u2581two", "298": "\u2581J", "299": "\u2581im", "300": "ther", "301": "ci", "302": "ose", "303": "\u2581cl", "304": "\u2581qu", "305": "\u2581man", "306": "\u2581also", "307": "ree", "308": "\u2581en", "309": "ud", "310": "\u2581how", "311": "reat", "312": "ak", "313": "hing", "314": "ag", "315": "\u2581any", "316": "ff", "317": "ace", "318": "per", "319": "\u2581because", "320": "\u2581very", "321": "own", "322": "\u2581ad", "323": "\u2581act", "324": "\u2581been", "325": "\u2581now", "326": "\u2581ag", "327": "\u2581into", "328": "\u2581comp", "329": "ars", "330": "ions", "331": "are", "332": "ite", "333": "iv", "334": "\u2581these", "335": "ays", "336": "ep", "337": "\u2581This", "338": "\u2581she", "339": "ans", "340": "ah", "341": "een", "342": "\u2581over", "343": "ry", "344": "\u2581lo", "345": "age", "346": "\u2581pr", "347": "\u2581sp", "348": "ue", "349": "\u2581co", "350": "ick", "351": "ber", "352": "\u2581did", "353": "ip", "354": "ach", "355": "\u2581back", "356": "\u2581no", "357": "\u2581cont", "358": "\u2581other", "359": "\u2581every", "360": "pt", "361": "\u2581need", "362": "\u2581him", "363": "\u2581U", "364": "\u2581In", "365": "\u2581work", "366": "irst", "367": "\u2581part", "368": "\u2581look", "369": "ittle", "370": "ble", "371": "iz", "372": "\u2581un", "373": "\u2581make", "374": "omet", "375": "nder", "376": "ish", "377": "na", "378": "\u2581little", "379": "\u2581off", "380": "\u2581than", "381": "\u2581got", "382": "ually", "383": "\u2581per", "384": "\u2581good", "385": "\u2581way", "386": "\u2581could", "387": "\u2581ac", "388": "\u2581imp", "389": "able", "390": "\u2581where", "391": "iff", "392": "\u2581That", "393": "\u2581res", "394": "ount", "395": "pl", "396": "ance", "397": "\u2581first", "398": "\u2581ro", "399": "\u2581pre", "400": "ass", "401": "\u2581say", "402": "int", "403": "ated", "404": "ire", "405": "uch", "406": "ase", "407": "\u2581somet", "408": "ound", "409": "\u2581down", "410": "\u2581diff", "411": "sel", "412": "\u2581gu", "413": "\u2581am", "414": "ress", "415": "\u2581lot", "416": "ence", "417": "\u2581dis", "418": "orm", "419": "ix", "420": "\u2581po", "421": "ving", "422": "enty", "423": "\u2581K", "424": "\u2581spe", "425": "und", "426": "he", "427": "\u2581much", "428": "\u2581ar", "429": "round", "430": "\u2581app", "431": "co", "432": "ark", "433": "\u2581new", "434": "ater", "435": "ult", "436": "end", "437": "\u2581even", "438": "\u2581start", "439": "ations", "440": "rough", "441": "ile", "442": "fter", "443": "\u2581well", "444": "be", "445": "\u2581They", "446": "\u2581three", "447": "ign", "448": "ild", "449": "\u2581said", "450": "ough", "451": "ang", "452": "\u2581too", "453": "ade", "454": "\u2581bl", "455": "ens", "456": "\u2581inc", "457": "ia", "458": "\u2581those", "459": "\u2581mo", "460": "\u2581take", "461": "\u2581through", "462": "\u2581fl", "463": "\u2581kind", "464": "\u2581things", "465": "\u2581bet", "466": "\u2581only", "467": "\u2581St", "468": "\u2581let", "469": "cess", "470": "\u2581Ch", "471": "ary", "472": "vel", "473": "\u2581If", "474": "xt", "475": "other", "476": "av", "477": "ical", "478": "ord", "479": "\u2581again", "480": "\u2581something", "481": "onna", "482": "fore", "483": "\u2581may", "484": "ting", "485": "\u2581bu", "486": "\u2581differe", "487": "urn", "488": "\u2581gonna", "489": "\u2581does", "490": "uct", "491": "og", "492": "\u2581twenty", "493": "\u2581gr", "494": "\u2581Ye", "495": "wn", "496": "\u2581should", "497": "\u2581comm", "498": "ition", "499": "\u2581under", "500": "\u2581hel", "501": "ory", "502": "\u2581fo", "503": "\u2581use", "504": "igh", "505": "ife", "506": "\u2581actually", "507": "\u2581tal", "508": "\u2581call", "509": "ents", "510": "ious", "511": "ull", "512": "\u2581There", "513": "\u2581Yeah", "514": "\u2581most", "515": "\u2581ke", "516": "ors", "517": "ved", "518": "ys", "519": "\u2581sc", "520": "\u2581happ", "521": "ope", "522": "\u2581help", "523": "atch", "524": "\u2581What", "525": "\u2581rem", "526": "ple", "527": "\u2581Now", "528": "\u2581br", "529": "ool", "530": "oth", "531": "\u2581four", "532": "self", "533": "\u2581str", "534": "ne", "535": "thing", "536": "\u2581put", "537": "ial", "538": "\u2581great", "539": "ail", "540": "ub", "541": "ning", "542": "\u2581sm", "543": "\u2581feel", "544": "\u2581five", "545": "ody", "546": "undred", "547": "iss", "548": "ank", "549": "get", "550": "aking", "551": "\u2581many", "552": "\u2581hundred", "553": "\u2581years", "554": "\u2581being", "555": "\u2581come", "556": "\u2581mean", "557": "ily", "558": "\u2581different", "559": "\u2581after", "560": "\u2581ser", "561": "\u2581show", "562": "form", "563": "ful", "564": "oy", "565": "\u2581six", "566": "\u2581vide", "567": "\u2581V", "568": "\u2581its", "569": "\u2581point", "570": "\u2581day", "571": "\u2581des", "572": "ons", "573": "\u2581bit", "574": "\u2581bel", "575": "\u2581before", "576": "\u2581aw", "577": "\u2581end", "578": "\u2581Oh", "579": "\u2581still", "580": "ath", "581": "\u2581long", "582": "\u2581'", "583": "ise", "584": "ob", "585": "day", "586": "\u2581add", "587": "ft", "588": "ves", "589": "ces", "590": "ady", "591": "\u2581cr", "592": "\u2581around", "593": "\u2581try", "594": "les", "595": "vers", "596": "kay", "597": "ian", "598": "ates", "599": "\u2581find", "600": "ward", "601": "\u2581As", "602": "\u2581eight", "603": "lic", "604": "\u2581same", "605": "\u2581pos", "606": "\u2581em", "607": "\u2581made", "608": "\u2581supp", "609": "\u2581life", "610": "\u2581Be", "611": "pect", "612": "\u2581dec", "613": "\u2581play", "614": "ange", "615": "\u2581att", "616": "\u2581pers", "617": "ways", "618": "\u2581high", "619": "\u2581hand", "620": "\u2581next", "621": "\u2581cons", "622": "\u2581own", "623": "\u2581inv", "624": "ower", "625": "\u2581ind", "626": "ert", "627": "ng", "628": "ave", "629": "\u2581year", "630": "\u2581big", "631": "ating", "632": "\u2581world", "633": "\u2581rel", "634": "\u2581sure", "635": "\u2581tra", "636": "ew", "637": "ered", "638": "\u2581fin", "639": "\u2581Well", "640": "\u2581sl", "641": "\u2581doing", "642": "bs", "643": "\u2581set", "644": "\u2581rec", "645": "ual", "646": "cial", "647": "\u2581ph", "648": "erm", "649": "\u2581love", "650": "ph", "651": "\u2581real", "652": "\u2581last", "653": "ict", "654": "\u2581bo", "655": "\u2581ra", "656": "ible", "657": "\u2581wr", "658": "mer", "659": "\u2581count", "660": "ities", "661": "\u2581always", "662": "inet", "663": "ments", "664": "uc", "665": "\u2581might", "666": "\u2581inter", "667": "\u2581video", "668": "gin", "669": "\u2581tell", "670": "\u2581never", "671": "vent", "672": "\u2581import", "673": "ied", "674": "\u2581sy", "675": "\u2581How", "676": "ically", "677": "ought", "678": "\u2581thir", "679": "\u2581rep", "680": "ks", "681": "ib", "682": "\u2581fam", "683": "ject", "684": "\u2581bas", "685": "\u2581She", "686": "\u2581give", "687": "akes", "688": "\u2581ninet", "689": "\u2581reg", "690": "\u2581min", "691": "\u2581op", "692": "\u2581def", "693": "\u2581didn", "694": "te", "695": "\u2581cour", "696": "\u2581why", "697": "\u2581ent", "698": "\u2581place", "699": "\u2581ins", "700": "\u2581car", "701": "ather", "702": "\u2581person", "703": "ular", "704": "\u2581inst", "705": "\u2581prod", "706": "lect", "707": "\u2581Al", "708": "\u2581today", "709": "\u2581bec", "710": "\u2581sur", "711": "\u2581All", "712": "\u2581another", "713": "\u2581bus", "714": "\u2581keep", "715": "ell", "716": "ese", "717": "riend", "718": "\u2581quest", "719": "\u2581talk", "720": "als", "721": "ings", "722": "\u2581mon", "723": "cond", "724": "old", "725": "\u2581acc", "726": "\u2581la", "727": "\u2581num", "728": "ident", "729": "\u2581che", "730": "iness", "731": "\u2581turn", "732": "\u2581ear", "733": "\u2581No", "734": "ousand", "735": "\u2581better", "736": "ific", "737": "\u2581loo", "738": "\u2581gl", "739": "oc", "740": "\u2581important", "741": "ited", "742": "\u2581An", "743": "\u2581thousand", "744": "ility", "745": "llow", "746": "\u2581used", "747": "\u2581gen", "748": "\u2581sim", "749": "li", "750": "\u2581happen", "751": "\u2581Un", "752": "\u2581Let", "753": "air", "754": "ock", "755": "ably", "756": "gg", "757": "\u2581watch", "758": "\u2581For", "759": "\u2581sw", "760": "ren", "761": "ute", "762": "ever", "763": "\u2581pol", "764": "\u2581sch", "765": "\u2581When", "766": "\u2581such", "767": "\u2581fif", "768": "\u2581home", "769": "\u2581cle", "770": "\u2581contin", "771": "ouse", "772": "\u2581friend", "773": "uring", "774": "\u2581Okay", "775": "gr", "776": "\u2581able", "777": "\u2581stud", "778": "\u2581eff", "779": "hip", "780": "body", "781": "\u2581top", "782": "ness", "783": "\u2581exper", "784": "\u2581pret", "785": "\u2581both", "786": "\u2581done", "787": "cri", "788": "\u2581mark", "789": "\u2581while", "790": "\u2581old", "791": "ros", "792": "ont", "793": "\u2581second", "794": "ative", "795": "\u2581thought", "796": "\u2581best", "797": "\u2581found", "798": "iew", "799": "\u2581belie", "800": "\u2581each", "801": "erest", "802": "\u2581tri", "803": "\u2581eas", "804": "\u2581ca", "805": "\u2581fact", "806": "\u2581care", "807": "\u2581fun", "808": "atter", "809": "ures", "810": "\u2581head", "811": "\u2581lear", "812": "\u2581water", "813": "\u2581hard", "814": "\u2581few", "815": "\u2581side", "816": "ween", "817": "\u2581exp", "818": "\u2581away", "819": "its", "820": "\u2581ext", "821": "lud", "822": "\u2581run", "823": "\u2581trans", "824": "ince", "825": "\u2581sk", "826": "\u2581open", "827": "cus", "828": "\u2581between", "829": "\u2581called", "830": "\u2581wee", "831": "\u2581pretty", "832": "ason", "833": "\u2581far", "834": "ember", "835": "omm", "836": "\u2581interest", "837": "any", "838": "ner", "839": "uff", "840": "\u2581pres", "841": "\u2581cur", "842": "\u2581child", "843": "ee", "844": "\u2581toget", "845": "\u2581together", "846": "olog", "847": "\u2581God", "848": "ond", "849": "\u2581char", "850": "\u2581looking", "851": "stem", "852": "az", "853": "cent", "854": "\u2581ob", "855": "\u2581ass", "856": "land", "857": "\u2581doesn", "858": "\u2581business", "859": "\u2581course", "860": "\u2581ten", "861": "ps", "862": "arch", "863": "ced", "864": "ms", "865": "ize", "866": "nce", "867": "\u2581ref", "868": "\u2581name", "869": "ross", "870": "\u2581grow", "871": "oney", "872": "\u2581went", "873": "ics", "874": "teen", "875": "\u2581cou", "876": "\u2581prob", "877": "\u2581ret", "878": "\u2581guys", "879": "\u2581came", "880": "ash", "881": "led", "882": "\u2581Eur", "883": "ues", "884": "\u2581ide", "885": "gan", "886": "\u2581everything", "887": "\u2581getting", "888": "\u2581ask", "889": "\u2581cor", "890": "\u2581build", "891": "\u2581sign", "892": "\u2581small", "893": "uck", "894": "\u2581el", "895": "\u2581col", "896": "\u2581Is", "897": "ational", "898": "stand", "899": "cy", "900": "\u2581conf", "901": "der", "902": "\u2581bre", "903": "\u2581cap", "904": "\u2581mod", "905": "ets", "906": "ike", "907": "\u2581number", "908": "\u2581comple", "909": "ertain", "910": "\u2581ever", "911": "\u2581coll", "912": "\u2581hum", "913": "\u2581Europe", "914": "\u2581cre", "915": "\u2581met", "916": "\u2581exam", "917": "\u2581move", "918": "\u2581pass", "919": "\u2581left", "920": "\u2581system", "921": "\u2581includ", "922": "\u2581Thank", "923": "cept", "924": "\u2581wom", "925": "\u2581product", "926": "ten", "927": "\u2581rest", "928": "\u2581probably", "929": "\u2581dri", "930": "\u2581Do", "931": "\u2581gener", "932": "\u2581anything", "933": "\u2581lar", "934": "\u2581My", "935": "\u2581school", "936": "\u2581lead", "937": "\u2581sub", "938": "\u2581ty", "939": "\u2581plan", "940": "\u2581seem", "941": "\u2581whole", "942": "irect", "943": "\u2581light", "944": "\u2581must", "945": "\u2581mom", "946": "\u2581opp", "947": "\u2581support", "948": "\u2581family", "949": "ices", "950": "amp", "951": "\u2581proble", "952": "\u2581dr", "953": "ready", "954": "\u2581using", "955": "ense", "956": "\u2581prov", "957": "ush", "958": "ax", "959": "\u2581power", "960": "\u2581Re", "961": "alth", "962": "\u2581ev", "963": "\u2581stand", "964": "\u2581war", "965": "ts", "966": "\u2581", "967": "e", "968": "t", "969": "o", "970": "a", "971": "n", "972": "i", "973": "s", "974": "r", "975": "h", "976": "l", "977": "d", "978": "u", "979": "c", "980": "m", "981": "y", "982": "g", "983": "w", "984": "f", "985": "p", "986": ".", "987": "b", "988": ",", "989": "v", "990": "k", "991": "'", "992": "I", "993": "T", "994": "A", "995": "S", "996": "x", "997": "W", "998": "j", "999": "B", "1000": "C", "1001": "H", "1002": "?", "1003": "M", "1004": "O", "1005": "Y", "1006": "N", "1007": "P", "1008": "E", "1009": "q", "1010": "L", "1011": "D", "1012": "z", "1013": "G", "1014": "F", "1015": "R", "1016": "!", "1017": "J", "1018": "U", "1019": "K", "1020": "V", "1021": "Q", "1022": "Z", "1023": "X"} \ No newline at end of file +{"0": "", "1": "▁t", "2": "▁th", "3": "▁a", "4": "in", "5": "re", "6": "▁the", "7": "▁w", "8": "▁s", "9": "▁o", "10": "er", "11": "ou", "12": "at", "13": "nd", "14": "it", "15": "▁h", "16": "▁c", "17": "▁b", "18": "is", "19": "en", "20": "on", "21": "ing", "22": "▁f", "23": "▁to", "24": "▁m", "25": "es", "26": "▁p", "27": "or", "28": "an", "29": "▁d", "30": "ll", "31": "▁I", "32": "ed", "33": "▁and", "34": "▁l", "35": "▁of", "36": "▁in", "37": "▁y", "38": "ar", "39": "▁g", "40": "▁you", "41": "as", "42": "om", "43": "▁n", "44": "ve", "45": "▁that", "46": "le", "47": "ic", "48": "us", "49": "ow", "50": "et", "51": "al", "52": "▁e", "53": "ut", "54": "▁it", "55": "ot", "56": "▁be", "57": "▁T", "58": "ion", "59": "▁is", "60": "▁wh", "61": "▁re", "62": "▁on", "63": "▁we", "64": "ent", "65": "▁A", "66": "ay", "67": "▁ha", "68": "▁Th", "69": "id", "70": "▁S", "71": "ac", "72": "gh", "73": "ver", "74": "ke", "75": "▁for", "76": "im", "77": "ly", "78": "ur", "79": "ld", "80": "▁he", "81": "▁st", "82": "all", "83": "ro", "84": "st", "85": "se", "86": "ct", "87": "ith", "88": "ir", "89": "am", "90": "▁this", "91": "if", "92": "▁W", "93": "oo", "94": "ri", "95": "▁was", "96": "ght", "97": "▁u", "98": "▁with", "99": "ad", "100": "ch", "101": "▁se", "102": "▁k", "103": "▁an", "104": "▁The", "105": "▁li", "106": "▁do", "107": "▁B", "108": "▁have", "109": "▁as", "110": "th", "111": "▁are", "112": "▁sh", "113": "ust", "114": "ce", "115": "ally", "116": "ill", "117": "▁H", "118": "▁j", "119": "ter", "120": "▁go", "121": "▁And", "122": "ation", "123": "▁C", "124": "▁so", "125": "ome", "126": "▁not", "127": "op", "128": "il", "129": "ore", "130": "▁ne", "131": "▁can", "132": "▁me", "133": "▁at", "134": "ould", "135": "ant", "136": "▁M", "137": "▁like", "138": "ere", "139": "▁they", "140": "ra", "141": "ers", "142": "▁ab", "143": "▁de", "144": "▁kn", "145": "ge", "146": "▁Y", "147": "▁ch", "148": "ul", "149": "pp", "150": "▁or", "151": "▁al", "152": "▁con", "153": "▁com", "154": "ess", "155": "▁su", "156": "out", "157": "▁your", "158": "▁So", "159": "ate", "160": "▁one", "161": "▁all", "162": "▁ex", "163": "est", "164": "▁fr", "165": "▁just", "166": "▁pro", "167": "▁know", "168": "▁O", "169": "ain", "170": "▁but", "171": "ol", "172": "ive", "173": "▁v", "174": "use", "175": "very", "176": "art", "177": "qu", "178": "▁my", "179": "el", "180": "▁N", "181": "nt", "182": "▁It", "183": "▁what", "184": "ab", "185": "▁P", "186": "▁wor", "187": "▁out", "188": "▁there", "189": "▁up", "190": "um", "191": "▁from", "192": "pe", "193": "▁tw", "194": "▁r", "195": "and", "196": "ight", "197": "ort", "198": "un", "199": "▁L", "200": "ist", "201": "▁about", "202": "ide", "203": "ig", "204": "ake", "205": "▁D", "206": "em", "207": "os", "208": "king", "209": "rou", "210": "ind", "211": "our", "212": "res", "213": "▁We", "214": "▁get", "215": "▁E", "216": "▁G", "217": "ack", "218": "▁le", "219": "ity", "220": "od", "221": "▁F", "222": "ard", "223": "▁pl", "224": "▁our", "225": "▁int", "226": "ment", "227": "▁will", "228": "ies", "229": "▁by", "230": "ink", "231": "ca", "232": "▁if", "233": "red", "234": "her", "235": "ie", "236": "▁us", "237": "▁some", "238": "▁don", "239": "ven", "240": "ood", "241": "ast", "242": "▁R", "243": "▁his", "244": "▁tim", "245": "▁tr", "246": "▁more", "247": "ich", "248": "ous", "249": "ame", "250": "▁going", "251": "▁had", "252": "▁them", "253": "ook", "254": "▁pe", "255": "▁Wh", "256": "▁You", "257": "▁But", "258": "ine", "259": "▁here", "260": "▁would", "261": "cause", "262": "right", "263": "so", "264": "ost", "265": "ure", "266": "▁has", "267": "ect", "268": "▁think", "269": "▁fe", "270": "ong", "271": "▁see", "272": "▁when", "273": "▁who", "274": "▁were", "275": "▁really", "276": "▁their", "277": "▁want", "278": "one", "279": "ople", "280": "▁then", "281": "▁time", "282": "▁sa", "283": "ap", "284": "▁te", "285": "▁He", "286": "▁ye", "287": "ck", "288": "▁her", "289": "▁thing", "290": "▁right", "291": "▁which", "292": "itt", "293": "ice", "294": "act", "295": "▁people", "296": "ty", "297": "▁two", "298": "▁J", "299": "▁im", "300": "ther", "301": "ci", "302": "ose", "303": "▁cl", "304": "▁qu", "305": "▁man", "306": "▁also", "307": "ree", "308": "▁en", "309": "ud", "310": "▁how", "311": "reat", "312": "ak", "313": "hing", "314": "ag", "315": "▁any", "316": "ff", "317": "ace", "318": "per", "319": "▁because", "320": "▁very", "321": "own", "322": "▁ad", "323": "▁act", "324": "▁been", "325": "▁now", "326": "▁ag", "327": "▁into", "328": "▁comp", "329": "ars", "330": "ions", "331": "are", "332": "ite", "333": "iv", "334": "▁these", "335": "ays", "336": "ep", "337": "▁This", "338": "▁she", "339": "ans", "340": "ah", "341": "een", "342": "▁over", "343": "ry", "344": "▁lo", "345": "age", "346": "▁pr", "347": "▁sp", "348": "ue", "349": "▁co", "350": "ick", "351": "ber", "352": "▁did", "353": "ip", "354": "ach", "355": "▁back", "356": "▁no", "357": "▁cont", "358": "▁other", "359": "▁every", "360": "pt", "361": "▁need", "362": "▁him", "363": "▁U", "364": "▁In", "365": "▁work", "366": "irst", "367": "▁part", "368": "▁look", "369": "ittle", "370": "ble", "371": "iz", "372": "▁un", "373": "▁make", "374": "omet", "375": "nder", "376": "ish", "377": "na", "378": "▁little", "379": "▁off", "380": "▁than", "381": "▁got", "382": "ually", "383": "▁per", "384": "▁good", "385": "▁way", "386": "▁could", "387": "▁ac", "388": "▁imp", "389": "able", "390": "▁where", "391": "iff", "392": "▁That", "393": "▁res", "394": "ount", "395": "pl", "396": "ance", "397": "▁first", "398": "▁ro", "399": "▁pre", "400": "ass", "401": "▁say", "402": "int", "403": "ated", "404": "ire", "405": "uch", "406": "ase", "407": "▁somet", "408": "ound", "409": "▁down", "410": "▁diff", "411": "sel", "412": "▁gu", "413": "▁am", "414": "ress", "415": "▁lot", "416": "ence", "417": "▁dis", "418": "orm", "419": "ix", "420": "▁po", "421": "ving", "422": "enty", "423": "▁K", "424": "▁spe", "425": "und", "426": "he", "427": "▁much", "428": "▁ar", "429": "round", "430": "▁app", "431": "co", "432": "ark", "433": "▁new", "434": "ater", "435": "ult", "436": "end", "437": "▁even", "438": "▁start", "439": "ations", "440": "rough", "441": "ile", "442": "fter", "443": "▁well", "444": "be", "445": "▁They", "446": "▁three", "447": "ign", "448": "ild", "449": "▁said", "450": "ough", "451": "ang", "452": "▁too", "453": "ade", "454": "▁bl", "455": "ens", "456": "▁inc", "457": "ia", "458": "▁those", "459": "▁mo", "460": "▁take", "461": "▁through", "462": "▁fl", "463": "▁kind", "464": "▁things", "465": "▁bet", "466": "▁only", "467": "▁St", "468": "▁let", "469": "cess", "470": "▁Ch", "471": "ary", "472": "vel", "473": "▁If", "474": "xt", "475": "other", "476": "av", "477": "ical", "478": "ord", "479": "▁again", "480": "▁something", "481": "onna", "482": "fore", "483": "▁may", "484": "ting", "485": "▁bu", "486": "▁differe", "487": "urn", "488": "▁gonna", "489": "▁does", "490": "uct", "491": "og", "492": "▁twenty", "493": "▁gr", "494": "▁Ye", "495": "wn", "496": "▁should", "497": "▁comm", "498": "ition", "499": "▁under", "500": "▁hel", "501": "ory", "502": "▁fo", "503": "▁use", "504": "igh", "505": "ife", "506": "▁actually", "507": "▁tal", "508": "▁call", "509": "ents", "510": "ious", "511": "ull", "512": "▁There", "513": "▁Yeah", "514": "▁most", "515": "▁ke", "516": "ors", "517": "ved", "518": "ys", "519": "▁sc", "520": "▁happ", "521": "ope", "522": "▁help", "523": "atch", "524": "▁What", "525": "▁rem", "526": "ple", "527": "▁Now", "528": "▁br", "529": "ool", "530": "oth", "531": "▁four", "532": "self", "533": "▁str", "534": "ne", "535": "thing", "536": "▁put", "537": "ial", "538": "▁great", "539": "ail", "540": "ub", "541": "ning", "542": "▁sm", "543": "▁feel", "544": "▁five", "545": "ody", "546": "undred", "547": "iss", "548": "ank", "549": "get", "550": "aking", "551": "▁many", "552": "▁hundred", "553": "▁years", "554": "▁being", "555": "▁come", "556": "▁mean", "557": "ily", "558": "▁different", "559": "▁after", "560": "▁ser", "561": "▁show", "562": "form", "563": "ful", "564": "oy", "565": "▁six", "566": "▁vide", "567": "▁V", "568": "▁its", "569": "▁point", "570": "▁day", "571": "▁des", "572": "ons", "573": "▁bit", "574": "▁bel", "575": "▁before", "576": "▁aw", "577": "▁end", "578": "▁Oh", "579": "▁still", "580": "ath", "581": "▁long", "582": "▁'", "583": "ise", "584": "ob", "585": "day", "586": "▁add", "587": "ft", "588": "ves", "589": "ces", "590": "ady", "591": "▁cr", "592": "▁around", "593": "▁try", "594": "les", "595": "vers", "596": "kay", "597": "ian", "598": "ates", "599": "▁find", "600": "ward", "601": "▁As", "602": "▁eight", "603": "lic", "604": "▁same", "605": "▁pos", "606": "▁em", "607": "▁made", "608": "▁supp", "609": "▁life", "610": "▁Be", "611": "pect", "612": "▁dec", "613": "▁play", "614": "ange", "615": "▁att", "616": "▁pers", "617": "ways", "618": "▁high", "619": "▁hand", "620": "▁next", "621": "▁cons", "622": "▁own", "623": "▁inv", "624": "ower", "625": "▁ind", "626": "ert", "627": "ng", "628": "ave", "629": "▁year", "630": "▁big", "631": "ating", "632": "▁world", "633": "▁rel", "634": "▁sure", "635": "▁tra", "636": "ew", "637": "ered", "638": "▁fin", "639": "▁Well", "640": "▁sl", "641": "▁doing", "642": "bs", "643": "▁set", "644": "▁rec", "645": "ual", "646": "cial", "647": "▁ph", "648": "erm", "649": "▁love", "650": "ph", "651": "▁real", "652": "▁last", "653": "ict", "654": "▁bo", "655": "▁ra", "656": "ible", "657": "▁wr", "658": "mer", "659": "▁count", "660": "ities", "661": "▁always", "662": "inet", "663": "ments", "664": "uc", "665": "▁might", "666": "▁inter", "667": "▁video", "668": "gin", "669": "▁tell", "670": "▁never", "671": "vent", "672": "▁import", "673": "ied", "674": "▁sy", "675": "▁How", "676": "ically", "677": "ought", "678": "▁thir", "679": "▁rep", "680": "ks", "681": "ib", "682": "▁fam", "683": "ject", "684": "▁bas", "685": "▁She", "686": "▁give", "687": "akes", "688": "▁ninet", "689": "▁reg", "690": "▁min", "691": "▁op", "692": "▁def", "693": "▁didn", "694": "te", "695": "▁cour", "696": "▁why", "697": "▁ent", "698": "▁place", "699": "▁ins", "700": "▁car", "701": "ather", "702": "▁person", "703": "ular", "704": "▁inst", "705": "▁prod", "706": "lect", "707": "▁Al", "708": "▁today", "709": "▁bec", "710": "▁sur", "711": "▁All", "712": "▁another", "713": "▁bus", "714": "▁keep", "715": "ell", "716": "ese", "717": "riend", "718": "▁quest", "719": "▁talk", "720": "als", "721": "ings", "722": "▁mon", "723": "cond", "724": "old", "725": "▁acc", "726": "▁la", "727": "▁num", "728": "ident", "729": "▁che", "730": "iness", "731": "▁turn", "732": "▁ear", "733": "▁No", "734": "ousand", "735": "▁better", "736": "ific", "737": "▁loo", "738": "▁gl", "739": "oc", "740": "▁important", "741": "ited", "742": "▁An", "743": "▁thousand", "744": "ility", "745": "llow", "746": "▁used", "747": "▁gen", "748": "▁sim", "749": "li", "750": "▁happen", "751": "▁Un", "752": "▁Let", "753": "air", "754": "ock", "755": "ably", "756": "gg", "757": "▁watch", "758": "▁For", "759": "▁sw", "760": "ren", "761": "ute", "762": "ever", "763": "▁pol", "764": "▁sch", "765": "▁When", "766": "▁such", "767": "▁fif", "768": "▁home", "769": "▁cle", "770": "▁contin", "771": "ouse", "772": "▁friend", "773": "uring", "774": "▁Okay", "775": "gr", "776": "▁able", "777": "▁stud", "778": "▁eff", "779": "hip", "780": "body", "781": "▁top", "782": "ness", "783": "▁exper", "784": "▁pret", "785": "▁both", "786": "▁done", "787": "cri", "788": "▁mark", "789": "▁while", "790": "▁old", "791": "ros", "792": "ont", "793": "▁second", "794": "ative", "795": "▁thought", "796": "▁best", "797": "▁found", "798": "iew", "799": "▁belie", "800": "▁each", "801": "erest", "802": "▁tri", "803": "▁eas", "804": "▁ca", "805": "▁fact", "806": "▁care", "807": "▁fun", "808": "atter", "809": "ures", "810": "▁head", "811": "▁lear", "812": "▁water", "813": "▁hard", "814": "▁few", "815": "▁side", "816": "ween", "817": "▁exp", "818": "▁away", "819": "its", "820": "▁ext", "821": "lud", "822": "▁run", "823": "▁trans", "824": "ince", "825": "▁sk", "826": "▁open", "827": "cus", "828": "▁between", "829": "▁called", "830": "▁wee", "831": "▁pretty", "832": "ason", "833": "▁far", "834": "ember", "835": "omm", "836": "▁interest", "837": "any", "838": "ner", "839": "uff", "840": "▁pres", "841": "▁cur", "842": "▁child", "843": "ee", "844": "▁toget", "845": "▁together", "846": "olog", "847": "▁God", "848": "ond", "849": "▁char", "850": "▁looking", "851": "stem", "852": "az", "853": "cent", "854": "▁ob", "855": "▁ass", "856": "land", "857": "▁doesn", "858": "▁business", "859": "▁course", "860": "▁ten", "861": "ps", "862": "arch", "863": "ced", "864": "ms", "865": "ize", "866": "nce", "867": "▁ref", "868": "▁name", "869": "ross", "870": "▁grow", "871": "oney", "872": "▁went", "873": "ics", "874": "teen", "875": "▁cou", "876": "▁prob", "877": "▁ret", "878": "▁guys", "879": "▁came", "880": "ash", "881": "led", "882": "▁Eur", "883": "ues", "884": "▁ide", "885": "gan", "886": "▁everything", "887": "▁getting", "888": "▁ask", "889": "▁cor", "890": "▁build", "891": "▁sign", "892": "▁small", "893": "uck", "894": "▁el", "895": "▁col", "896": "▁Is", "897": "ational", "898": "stand", "899": "cy", "900": "▁conf", "901": "der", "902": "▁bre", "903": "▁cap", "904": "▁mod", "905": "ets", "906": "ike", "907": "▁number", "908": "▁comple", "909": "ertain", "910": "▁ever", "911": "▁coll", "912": "▁hum", "913": "▁Europe", "914": "▁cre", "915": "▁met", "916": "▁exam", "917": "▁move", "918": "▁pass", "919": "▁left", "920": "▁system", "921": "▁includ", "922": "▁Thank", "923": "cept", "924": "▁wom", "925": "▁product", "926": "ten", "927": "▁rest", "928": "▁probably", "929": "▁dri", "930": "▁Do", "931": "▁gener", "932": "▁anything", "933": "▁lar", "934": "▁My", "935": "▁school", "936": "▁lead", "937": "▁sub", "938": "▁ty", "939": "▁plan", "940": "▁seem", "941": "▁whole", "942": "irect", "943": "▁light", "944": "▁must", "945": "▁mom", "946": "▁opp", "947": "▁support", "948": "▁family", "949": "ices", "950": "amp", "951": "▁proble", "952": "▁dr", "953": "ready", "954": "▁using", "955": "ense", "956": "▁prov", "957": "ush", "958": "ax", "959": "▁power", "960": "▁Re", "961": "alth", "962": "▁ev", "963": "▁stand", "964": "▁war", "965": "ts", "966": "▁", "967": "e", "968": "t", "969": "o", "970": "a", "971": "n", "972": "i", "973": "s", "974": "r", "975": "h", "976": "l", "977": "d", "978": "u", "979": "c", "980": "m", "981": "y", "982": "g", "983": "w", "984": "f", "985": "p", "986": ".", "987": "b", "988": ",", "989": "v", "990": "k", "991": "'", "992": "I", "993": "T", "994": "A", "995": "S", "996": "x", "997": "W", "998": "j", "999": "B", "1000": "C", "1001": "H", "1002": "?", "1003": "M", "1004": "O", "1005": "Y", "1006": "N", "1007": "P", "1008": "E", "1009": "q", "1010": "L", "1011": "D", "1012": "z", "1013": "G", "1014": "F", "1015": "R", "1016": "!", "1017": "J", "1018": "U", "1019": "K", "1020": "V", "1021": "Q", "1022": "Z", "1023": "X"} \ No newline at end of file

    KpkKQQyd;M13QlD(V~gG*REkqeXoqj -}i5ae1@oRp71T^^JVCsBh$RM13Q_SJXFhgJ|!_Ge!B4XNmG7UoOgzT*4cWBINGK&3Hb0A}>L? -7xFUX-pDJEw?$r!+!uKr@{Y(Gk#`a|Ko_t-K<t}a$dizJAvYuUMqYrtE%G -Ad-H?|Edytoj{P6~_LgbIUTI7$sPUMfgQRFXf5HDi>$bFG_L#{zS7I`Rg32zu9k$WLeMDC3|L+D4IC- -fsP6#9{u2>p0NStj%&uMql?R||c}>x4e!O+ufBb9DpiL++2<8+jn|vB*P_OL&7BDbgcP6zP#?i1f(wM -0)(?MxjWLyjY}1UMk|_4Q9EBkGxXEM_w!9;{afTh>yHU#7C~agz-mW`H_1e4@B;bJXFNT8`MZ4N1iC; -hjL%0=2uQ7n{5gLi>ZQ-W^e~}b6LK?hGsd@IxCQlPU_MrSMz) -egh>IE8a74m}@$863l8WUwV0=*?o=+C0&qAJtb}q*9W#CwW9z4d=aeK1S4n3sc<N5=A$DVuy+47bB-PL5V0A&s}CFkCo+SmOfn0-@?XY;^g=imVca_e+IKRQjSmCANAifnJ)q!ChMo -|k;;WXY&gCXDRT?sLZ6-SSXm!!ue5v4sntWbZ+3K=sa{hMahgrAy!N&4+w -4Y>FZiiiF^ta&YtlSQ6Q^;+2emKStO^R}6DY(UjTNV1US^DWRPhsVp0xe3|ljTy6ydQCB;Qa}0aolsf -U!fX#3tNPq2}e)!)pL$|%x3izBl}s8_Rd7k`;!RyeBtA(4*LvPZ;8lxe-bX&3-2Es>3RR)keir3=l;R -Q&Ob*x;iZdJ)GI%)9sQW|c@V*Tl7;aMF6TdQXW_EHo!gCbI^nlTIAX>9GD&WK{CXX!=y$_pySbmjd@&wrmCJc+KKfhAn -;^FzgQ6eh`An3{@2uCgUb$XZ`)A1gI@deZiCfuqV3Mqtw`9>Wc>1|c<>mG4dY(o&*~NPv*YxpnJL2iX -74;M=>*w*O%KCWx1gG|IW$i!CC4RJ%J=|||T;flX>x0Li<|H?<{3BiBN66QC9zRS`-?QZM=&>GL)5j| -0GZpn1?o>W2^UrJ-9_iE$tgMwpDfGL}E79{R0sYVK7sWWS@Jp;>TwmA3`6A?E-boy>D6d7Hg#09OGx9 -f)7a%V|UW9xt@)G2Sk(VL=3V8+c8syc;Uq@bte4nrv`9WduHO@~V^+hh~U4y&|c?j~4gg=m1A`ex}o5 -Z60dE^<$#Y|q4VqT{JGDp_oTh|H=6lkstDPB0h4_PD7Eug*+1Zc9EZA9w!mytC5R&NqK)2l#6*8F;6M)w}SGQkryKW -0(mj=9mq?OA3|P^{3+y>$ln$IME;5JC-Oa_JjlNm{`!maYT+;B;<_G)yh4-*`BCIz-c`)YMxy+P@E7u -Lg@2GA7ydy0x$p<_y}}>J_X~d@7n|9YBi}0gfqbX%2l6w*AIP5;{y<(Q{BfQ0_k=%?e<}Qd{H*W?@^3 -`>k)IajM}9(-ANi*uKWtxO9yJf;VqQ(BmidI&KOF6PNgQ7*0*eV%&Wyd33^AQ$Tj#5&SSl)oVKBNrR?G$6MlZ$d7fAH+Hou}($Z%=rf4FXZCBH4yo;$i=!9v92W)lDPgib#|{DEy;X2a||$aUUtxy@+)&87LR`y?MxsMR|}vA@n2PjJyQs|Vx0sIqmx>cKZ?8o`D@6VkS|8ACj9<(6LPUmN384dNBKPDV%>~bXA_9>ROF$^KSUmh`~dPqzLFMmv0pQK`!>=5$kA -12>+mbiO{cDhZBnORmjD<8L`eM66FQR#X1_X?k5rD%SCw<>#QQcfxj%BT?k^DeIp -m?pi;zbme+hXa^4-Fp$hQf9A{XnT3Xz+U^KscF^7WEG$Li(#C(e`Q`yoC40(OZDH^|oq&W-Z(Cg%&~@ -^YRj_rIKH$@L+4z6;M**q0+;KR92ckXvuzuKH|p`{F#`sk}>A`RV^DPU9`k;aN`OJAC{;S=P_zUy@~i -@bO}!oF5;LN@im>&U*MfjX}O|U2~>{LF^4GL~i@O&3;m9 -M*;+ZFP~PWo)jeh0U^l!wnl8Rh)=ykVMbFQ4y8ll_y1{fxndJKB%n87@4{Nx#LV{ib2RA^gqfK^*q+c -^b!g$>%K{>)!eNle64{=dCl(kn`vBoEdWde4b4&w_iRFXq4|m_&lglwujGiILgPlQEm@>9r{AK-{$jY ->9W80d{?G?eq<{89Wn1CLioD%Oxb=uKb9%?TUodxgiYp)85iw?&-*yeM?OE8E%!%!9?OwGA1%&y;ueL -RpKq=_TW){ZxSN2mhtCU}4|k2fSmB?=3V-CX_U+*LE_Vhtg}-gsA2@ghyS`+|-1T~5! -}|)0%xy09$>;s;a(uo{z%17fZyokq?mL{roy*JT^%u+b@%eX$J$$}^_9s>ax4CeWOZ*Jk9?rAn_QknX -?jM$*KXPSmcgbHq>U|RUShjCU@co}<;yASK`yXwSt2Yha%*L7YzwO)p(bo9~t_)1!_xY)NZ;bRfSo3( -=Q`b6=&BC#B-xEcK^@GEck4uyKZ0y>9zzq7#s#87Vxqs#CzCB+1^ON-KO6#fYoj+LU4KCULW$J(}DWC -gq{kh)*N4_xsd?qtM`~3?G_>FeG>E?ki7PfsqZFs;vL)=u8?(Nr9vuO6zj}M&mgpAe>DOE@8-nO(`?= -`wFSN-bq<-tb=`24W+!1{hWqv$|@uG_aK6E;_Uq@R0g`}X;twVmkkTY8uGkHo$4&R+wcT${9>4>tC0y -Xw8JTi3iX`}NIJ+Kk>k?Y_DNrwY8c-E7pISW{YE&@J;W=o(T(qla(!Y|g`9g-@;5_j&!qc;9g!Sbn{* -&enNuMY~sfUd}oZGpo>r%gZaPL>BB`0HwQpSak-_V4 -`qOUF4KJB6QGcHcc;KX>QcrXHbBk*_x0?*rcJ(DTXk1G6;&>&L7f?e)CcX#U_s!( -z=32m3$&LPL6IlUL^#2K_R8QOB8)AAY&z!lbx|=+m(8dwJ{sP*vEvKSAzmEVF<5&Fh`^oj7;K(%uj~E -ikuB-S62SuZx&@_|P*0W^F(B^?>h6H@*5M8wJu$t-gGzR?E{~y60gk6II*SQ;$=@Rih&lC`qy_4-}_7ZopaWA*jqr-_72n>uKDEnvg`@3^qB -t67pKiHoE#t4<(-(%s}~h_tUq2fMmOW!gzeXIa;NwI?hS9YP0X-a+r4TM+rIj*{`%REKip&C{;QAM7d -K2Ce0WMp*oz+OL)td)mNZlzzBsGm&$GKHNI%054T``4DY>wiAm^_LIQMidn8xm!E9{G)ZJcigE?y?QKWVe_av7S2xZwC@|;=VOENGM*o -vdE%2detX}TN{7emuP!L~>y0C`ek|2YeZFbVv)jFI?vfsOb?{WeJ}p&*#_hB{#r}@p2xNo&Fgc|h)s6SCDYbjni8CQ&9 -w2N&Zo=4rl;EaHZJ(3L)L+Q;j6D5dSRW9^^K*?9*+h`lo{=Vf(Jh`=T+aI1FDH`$!8am%NrIJ*7{y3{ --W}eSu1^Ky*;=3*oS{54qH*~cITvMw|fpvzjFuqB;Z^{bpA^PhsGWWe)CdOpX~cLt_*#k_Rjn-EsHAZAK>0JB8{_*L#T~pr -q?uP~5H_jXHX&5o8#G~}_eZzi_Xdlq~n}~aBH~iLp`<6Q=WYGD#)SJsb9O`~(TG@}kojn=z(pu{`UiF -)PnvvdabLZE;-n;krJ2r$`*L@uIL)Ei`7CbW{wEgdItNbAH#c(z+;rs2?IIl_foekf!YskpAUe0;>iwUMBfqVYQ-9N3mvL?N*sdQKR*h4$tnq8UY^TG?yk -o}SK#;^Kt>ebOF;-)uzv*?rFJ+65zpEPx0?Y84fqtB{0z4E-B6ppLsV|u!$j`?13|Im`^-**4#$N$=% -*J<|4qhYX7Lxx?+HvEyzAzI(keH!~NY8JIy|VMvXM=Z+550T(r`GE9?C;x@K0mYR7q^33&c>CDxy$$X>() -mnzI(#&u$%wm=I6(piff45eIe$ruAW0*o&Mq3^{?NKE*ihM*K{MR+@+IKXrkG7V!@1LaWD2?+jTvXLN_?5eVK6G-`o?-8WhecoeI -?H2qWx3>WZt0KD#vRljNZwX^^6}H(k0{q?G_L<(V^zn6U+sS~m^B#Rjt!%ieEQw~ -?hJIJ`xTmq}5mmRFSN3XmZQpkw=O=ZK5Ak(D+dkGUKatxsc>M9Ri;9O{{K_k3rsdl^OW&A$Z_w9Yo>B -$f{C4t}PxkFNU}w!;@7(#G`@um~ZQl9(hc~C3n&$Pw=c-emteN@Uu!7kK)|4g=9U0!seRZb`AKmPDz1 -gFEeoFd{(O;!4jy=0CnN9}l%hJ1fOEnc|J*vlA{l1wnWb@1CUf8WacPb_5efR$Bpa06N-_Y;;$NTI)o -q7-M8IZKFWBHb`H7l*lt|Xl~{`sea&KTR={djrbAKN^g_Jb+v&YW+aeA>RkcK_&&rcZ{wQ#AUM@W9G7 -`Mdg_W)m~1Vf!~0e)Zhoh!fo|7Hk^}8E=Hk!Y#`}y$a9>=#n`9T&*J^$9Jm{&~ -d(wAPGwANOcu;%)lPidwKEx=V!Q)n%^a)zvI{`4>vw{w0$4o%Jzk6l^Gl>*#MD72^2yXcM|6lc;cKLFzqqhc -l!*Q2yzEt{4qUa^^BqU|+?A>MB=JRbHI!BAE6RxMx0Qe9-iRJbm*g3|DUdjL -GkfJH+^*Uq5I3kF`~(=SznE`wmmA==*Xtlh02-!g$!g8ph8xFw{lAdX$B))*oZxJG*_w@Pi`^N#$c-G -x_m*jx*l$J;Q<#8*3R~Kjj3I-}NiQg3FsuGQEFH{D$%AKQOHLW9=yxeslP@EZxFm40YkQ(+t=8oni44 --(Xl0*RPI+Yb^}xEFUr~_|@wy({~_&VaazdGSn?-;Nc;|&N2P>=QFIkRKZYl-19t39~i^1qV8#iNvab -J>(=(HM?WysHEm@`YA-U>1P5JU@@tt4b-JAlOH@BGtP3CVotTc4k8gg>Us@0_T3eZ}D&B9+(7yO-;pT -at-miUbXjQ^sRiZW|-fzs31IgN_ms*WG=j*kX9^L)s!hX}WTb3-@_x`GR+EcI0-Eh~)9Br<2H1dY0N& -EZN=cb$)o~do>@ypfvfmzzQ-Oof;KIW?p30u5p_SI-@@rbvt4=qX2F80bSeDym$>dDl`&3bv&+Buop> --*kVRC$kHoBE~XdGsrTHoW(d-S1z_)~0reDPCjF){Y+f*`-OwS=y8H8an=dPNV(3?|?mZ?~K!aHaYNu -`;Hv#hv5fzb$b4OZU03vZVmTkYLCv{eRj_NY;E$UuNQiGjMaV{{qCP#de7HRnRl&U)q*r_Y~+(s>W%k -ma~Dj0bINzw+MbX7@j&<2?$ahcTlQGSLajFS=d$nZBuD#v@n0+aJV$72KD=@u;)f*dBlF+7+~$o1+Kd -HnjsNaSjP{j9-TsPrcCvQEoEY7i9$DJ^lTUhYTd&hzK7V6Fbg)S~a_#GvzPg&M{cwJj=}>yA_TJ0mt9 -53RcJ%2L;U9gNuAPwgL$^0qCTl}~vvlpb@?Pz_y#XKk&PdTd{X>_dXE$YQ!ws;_9deLo7uQOd0bo6Si~vh|siT5YM|9 -is_cW8`oCzz7aG)Im!T@NmR4SS+`RHw((C2jA0OeKl-{5DgneHxRz3tVXe8uZE53^Zj-;WB}<0DXoBG -fbsAANXaqTEIC=44818e%V4>soQ`@n#L^lGN;E7LU%HC?0+(pySkGYN} -j`J_CvA;rcu^A-Ynv$Mep4YU$!0r+*m5NKBe=r?PIAz2TBSw@p#jHX|fwN*X28Acm_9fd}pW`hz97;3 -j#h9(_a&EN?*c -_6DXyv(YdL%ey!}aGHZ5gyJS#5fYQ}92_)~Oy@c`;54{uAZn=e@J8Tl}4BU`@d}j>c}Krp-#2930wm{ -S3Hk0DU_ND+O&uiX3i@-+KU*F)J%4mnx>ITQcj`bglcOmG);zx6=x@Y61<5p+_^OX$;wBTmHbm&o3F8 -8a5_9-Gf(8-`0bTnK>ofke$P~j3RyT?CT?+hng0`_z}?kC)U~0M;f{gtiWGI3REf=866j-YhF%)?36- -A*NSb0phl22s$Aa-8Y+?ho`S|a$=fCJEjcK`D|MZB9pBP`UX%VV7(=Kh&ffv3m -+qnwW@T@?39RvXu9}L4bP23Y=%5Knyv0dNTgNbyl*xe(vDpxvAsHYSRUM+N~8&z$`E5J+yV4u%SE8>* -t+5&At5PgbYS~s+B_%Fl_cDE@&S42WURie)&zFVqcoj4rbOb~4c)pbbRBRm?)*)NM=pLwI{RD=wWljDN=2X{g=?; -BQ_NQ&+FuvSQ2YJGS?+~t@zPp!%b_bq}SAVmLH245FV}t?~iKTiiR!d6X5VnLfc7=*m&XmKPU!Cxej- -I3>=f)udk`I|uV?F&)n^~BC&5JL8wg=%mmia;JsBj{cRuA0 -SFh*`0d&DBdyfcsH*XX_-zb*Br^1PQgKoOEAF??=l9;J(mX`JriFZRE50}$Ch)dDODQ^$o~)b&L{+b%FeLgsZMva(+_ptgV+%<5^*5Wg@XNVpSL}<>X5qfu}b%F&w3X^*9V38pFLD~s12>lmA -o!Xp)LcUtUk>DjdC3516m9egfhJcy(AMjLy!P^Pk3guFk?QKmuA4*i*bxjo(2CuR`nW&U~ih+JPZ<-}q_l8PPvc%A6{JF#=62HS2>{Z# -Pn1ZERY8wY%f6YoK{$-?u06v#rT>~YQ`AYUzu>sk-duR;C6rx5Jdl+tzeu0Cp_3t2&)isb_z`ksBX(n -W`62lxa&J1|ZMg^tFZgk0b0(4mxv?xOD{cF5ZiU%4D2PUIf&&mR7rVq5GBClj;(*@tU9k1!tv$~x!oh -wB@Bs!?Z5EAbKz!S(WxBc2~EpCa!D`eUIZj&o-ejq|Xp4?<4C^$tFbsArg>jOuT}9{BjK=kl>s$7)v{ -$3XtbX+lO~9DkONt1Ua~2&t`eE&iWqSSFeHdfn8r7J%lxQ{``271a(#UG1c{u9Jf -z>1`9vpMss-*(iN~g9p_Ohgz;2{d-sRp~$0lutYAyCvSc0=g0M-a6cRZIxx;g@MS;Vce$KzwEvtq9Lf%#a*X@P2uL@=!8kn)bBvJ;SYS~w<`(dQix-6!Zx{hw)er~~S@)!+mqiqY<#li@%6Fzl -vA3*P)_%-PTsLLX!Ysb4#-#Zlby%54CvoJ?{hx6ap*Qp(*yT@wizH+VE_7~B0@Fn~=T69InFYx&TJ`! -(2=RKrY4gGygYw1J5=8kyY4uUpC*VfbTRNesck4dbi>^A8wBt-2TKL+B{xvI0LyqwO?^YZz5Cv6i>ZB -_^2kHAOw3`csOszx8S!Fx&xZk{3Bw++3ImM+sR=UnU -abvsge`%v3V2>EcM2#LI}@L(z|jBN26oaftULS`Z&Y+=N($_%UJ);%UV15&c5Y4~T;i#~@BboQ;@)n1{FuaRc -Hu#CH*^5Whw|hxk*7qunz5(a#$YcLcl2qmhq6?2p(5aXw;}B3-p=ioq672Vw2zC{tR_9CFD$-e8N!F< -Eo63=uX~+xzD7tCN0WpK^vXuiPhUO=8#~jgDZ53O~?sp3eA%Zl|~y)h -B>5$+m*^A)89{qS*^x&lXyp1jw1_Vqj9I_q}sFCyVB@O#^A~3cN&^t$Tst=NrT&TdzQ@@o^LZG|)8T_tZdk@Bli6;IVkKCNxy~_~Qgd=A<;@{`$w -c}V_MChkMBY%v=2)FPxsukZOlQSUiWB*bC$v(a!joBLleLV;8M1R0vnq@?8&c+w8B)B>0{@~QMto%So -Uqs^mV6>_Z?OgoYeJ?JLqv|)^jg}R?j75z~CB2+N6WE*BA&^{REiB(^Gi1{#(l9IY7hwn#`qOtc1PnfL@*>ZH6BgPAs7NJVqbbF@KTw6mMIs!1MOoNP4CpMHc5_m^}Q%LzmIR -TQ23j}ZS{8(=U3BZDMW(@{Gs)=q;J! -bC6NcNpMi`?IO>BCH^@n)lmwHorGM}k}rCG@*o4=+(B*TZJYnCcIldjw|ri>~iugploSPH#d4ew_Dbl -+m*iLUN_a&@fn^L6))alZ_TDD?ih@hK0o#@>tLu@~B&s6)Z8Pvhx^DDYA$9CM+VBkY}m(8Iz{d?+2L2 -Vrl|p&!OQKG+Zkz(D(kF*u_#M{qP{=4PMBQ~n#_~rxLAB_ib-<-<7nLlpjC<*8KebKInzf16+K73m}x -Lnif@!xLR|0_r9FXrU()?Xa?@3s5C`}n`TT_}$Kk2e!Z_H|x$Zhe --Gc=TOCkRLmruT|X+{6?n;dnBd;eklFvPbZ{^2e4Te^`8TN<{^Yc;;^!I#*#nI7;tdH+cbS#WaU$y=x -AmmjS?U3ZO;VGr_@4cKqDz4yNcAHw4@>h-<%{|txV=O$iXZ{qdXPB-?M=Fx`y@9}@<0lnZqymNlngMly2LY1+H5~|f0uEJ52Eu@UN63`rxAH<0NwlUyhJn{%K&TXx*zCp8Ud4f) -4j~nR1JVLRR*`vzJ#=s3~rJAX!<%*3V4B>RTcH8TrV~Cr)!FnrPu*f&GAug9!;% -8C05~;3wd<&?Stc;lt^kQUydkf?^%fjNm*JFp-76Hz3D`)>F^=xFQ7;uE(m9 -HkR7UsGFgglVM;gND-feZnH=K~`5WhN{+t+}ofg8mFs9zmmXQ^IFoJjY6(@WJ8xn2H~sQ>DSc@o`o!J -w)GY=HPllOaDTS*n}@{k3$-tuTs^Y6vfzM)%%KQx!yWdl~>^dqKX_X}v(b#Xu7QPS=?H2>L>1(!E6MN -CjYTsX^5gM}Om9FO|d-QV8*q66jt#dZ{9T?!lKVCDQdERNpLW=UFLnHrz7LP(!Fu@Qu#8v=V* -$m;6a*CnyR7z>Qi-5rGA*^pDa~8O!x3SFBLpO$ZzyKdJOyw`gM=fy>e4kO^?&`2318N$HJ9RpWxr>RU -G}FAY>^m{}XgC-#RJwNm?K0CEXfY@6f&gOR0P<-8=NWlu^XnVF_RzgqJ_X&MTGg(}Xku&wHBgNm)lK0 -mI;YR6j#Vpa#wdy73)g9$c5a;5?%n>D~wS7d}tV-;1i+7wBFXdMWZnLY|ZARdt)`UZC|-MG0-^$&&6R -TF>BLz;U3bZVMsNfPq{2d0l{5x0SY|WGU|zx|d$Eq<)p^HK_7lrTNuMwSXm%&i^$+@&I+EaK2LhIwA7 -`3ty*usGXBaxAAt;u#N8VSjRSwtpIt+4!Y(dNvhsK_3I_`PF~(>z|g?uvh@OIq*7z -sRbFL1YuZk6SPTxxUC&A)=5H+dCwof*^{VkKZK#IM^=+fRzB>|;W%!1-Hp5b{@Da;x~1z8cTyR#FLmf -_SxupuM?YaxeUnkQ*euP2pi)uIg&=C&=ldhDp@kBeWfzmwan@zUCTQZ}pPzQGOnk0``M`sroD0FF?J1 -P5U#wRQxq)64M=uot4u*LCX9@9u{!eob`~y7W0zJ>ps{+5H^_3(g0xpH{GQd --SAs6BL23P?2K49ZTx~FiRs=R^5hkm?)o=-`t+Dp7#k>A7h7UG5e0MGn1{tvW2JTKK==Ito(M}9p>`j -PJC2G_rzc)lTsNq~0gysGYJ=*OuYS9rUjljjs;0T;O$-79{h{l^7W@$YcGhwE~66YYo6+zS7M`jyh%O -8!Xgo&mOKj=T(h>>u5hNN#p}4hfXDtL>k{+A}hg9jUEz71zzMS5i?+= -mqZ=`O!SdRG39qiq!O^Py+rl^Hu*{s5}eoPrSg -6>9bD)7OQZne7Y>AQ(C0r$V%Oo4!4la7;l9BV>wgLXUxM}*I7EVeppIk!?uT|AI8=i61=ru9R4-h204 -pF}NC@1*^n>;oBC&oWWSB(iA)Ky7q1SK!krMRNX{u^K9mFdfMfKN7fukj|6?oGa(9@4z-{F)1Od1b<0 -j$$XB#Fu=NTd)jFhU~g{&3$lQ6kZR4U;7@4lpT7A}63f3!Mt}<$kG6T@1BTuhPw+ewhZ)KVSO?e$SiDGrej!; -EX_1IKy&tqoaGkyA)@0}P>c3cG?LT8N)d%_JO3+^>NoK&Ev>(lf_5tY{mPqWrB;x_aEV&}E*3JLlLxXwHxkxQVr@ezq#XNw<|$VjT^F|gnLJC*-xiQONCJRy+;2&Zd3?g38MDVM-~5? -zB_1xVKoYv8_yuA6KN_ZL-&bp2`>AYB)<39f&1eI^Z0MP7)w1F;(Ajfixe9t|IfNY`r7@I1s~M7nN?+ -C$gtgn&QkniJcAe>(c@&i0WEfm=NKvwFDID4>HdMTFgs(>+}aT;qGPG1v;WvYovHpn#@t1f=^_sNqzh ->xngh^Z_#zkgm^;m6UoU_6xa;7Q*Sizy&RY(|n6t2&YHcEd!!FprV!V>Q=(*S_yA#CH&S&1|II)O1P$ -#@Q@b5>3ei_D*k(c1}Aywqi{|;*^~5`ij+Eyz^K>4Iq2cw=JhJ}%p(r}cY*slc*W5k%hd)Dpoy)Tz{5s{J -|2oL&0dXGU`Z>?~gT=3^I{YaQ_=D*`L-Xex@^5ZleYlF624U2`>xZwM0roqUGy7LuUsZLz>UXMx@|$t -S21|LuOw1-bXGi~R-*w+89o -sS$jl1!gIoy5n-lO;=*xU})2q9U?w+cvUi&mPvszyA7ba^b=Sxt*Ko&syNwz+Co@C#{JG%nvlbo%RuV -=%b%MO3Tb44?z{Bebn6i&__QkfR>Q?&`-ZDcpJVv^x}ni&CP$MEjx5!)~s0{rA3e24!9t5>Y^FV&3hL -#Ke%$hNAqSaAP+Ua{L$W7RNuZeRR2TG%|HI}M;T$Q_yIy530xQvGr!3H-2`4F0&84 -Ewa1OgPv~Mt$B)#(mjL=6}&lB91kaX(yY>%rniz3ZGf$n#sJ2%_OC+nIvCsCd-#ECo5O3BY-}Wd{P72CEA&teg)4Y0A#I@1sNtgR4ei -DUPWC(C1lBbBk@3yL$h_v6B)@qPDQ;dz_BOvsjx`?<{VC}L>H9$X5J*2B(nmx3xsX1+k?sTEfvke`n; -`uzNM8ZzYao5SQ~Lfx`EP;lh4eN^zXsC34(UIJ^hY3lEu=pU>CZ#@OOXC&NdLQ2`l0>d`W6OleI8tA{ --5@)1um*;>mM=`trV}fE1Bhp4^UDtU-)ErDUc-UjVS8^DGDWs-~&q$A2+eQg!gs}GjrYy%nYEJQktUK -jh>>GQdyE|k!DgpqO{hx{(ENQVM0W&-|u_;?yl{bnRC|O>%G?A>u}Coj^p`?vX|G{^>4Bf?(eKn%^-z -7pQO;SISQRvqYy8W>AVX5k64Q_JSQ*(V;Me%;d!k3REA&3@J}=RdWJ7!_+1RYpW%U4;%3jv|dd{G!^Fs;^%1~(1Qw -q&_Nuf2H6xvv&(7q!ocn!l3WB3US|0u&}GkiY7n;8BThTp;PZ!{!-5|Ys|(Xx%Z!@Fz3z=6&Ydi(U$^coL%9zQxdDm -Xgf&Mvom^Lf6&&vilH1cna|0dQD!!W~_@-0nOtaO~})VI-p($4|KJk6pUlJ4RgKFD~e*0eo0E&W7OpA -G_RnuM4#a{V@Z>Cj^JY+!q}k5uMNt1KjT2wdh_IewU!LUJu(G%j%QPmDikR1C8?CL%gU5Fuue{Ac_V!Ut) -ALNt%jZQQsnUApiEjpq}F-ISmWh~NqC3VM9X00&eu^aG&c0coh{M(s({$M1WWOn>;)^M>2*$F(x{g73tFXkl{lE2e -)tI`G+Vz85YXAJY+~%*roFc385hqM|Eu1b{KyQOTg$+;g{x52*V{4N8NfuJ1x$K1+%)!*IY^du!3oC? -%Z)=LRd`jDAqMaQIGFzkzSog#Dp*(qg1Ccez-WzzYiZAHY%#=X{Y}1_~ERGQ4vE&x}8>y&#Z|G8R>Rf -iEi3r+)fi!E@|kL8(hnYMCJ9#=aw|no6Y5Nm0-7guH`&M=5v!LPZs$|PV)H1+@mOatdf@+p%i9@DVvu -?DCIBaD8s8br(i!9d5Dy_qt2Wmn~%>w|Gctp-8yCc`t^!luUAYalgJfbdF2(655E5T>mpb9xUxuD$N5 -G%=L&oF>`^}a@I#Rs96NeMIePS{a_rbK<>bke%6H#=r~LTCX^|V8J9kd`_19mOH9wKEk-zV|pz^Z;Jr -z+{%3!-Jn8tx_AqTo=InY^@&a_z>L_3s8RHe+J1IijYtZby?9KY*T@Ej0PsLSxdA7m(^ShZKn@X-uEm -Eq?w{7Qyj&+wZV{vC$D6zBevr~D^R`M;W{$oKMe>())&(sH|akx)19J^=v%cx-bwO}B2|-Fo)#;nlI@ -U94xf`}p&b+0(N_`?hUw>DQ@Kw;ONj*8k3%Jv-os0Iwdmwex7_6M%MS`?eVWPu@ -3l;cqwh>d?NuKigN&K6kX~&Sx3qPX5mG`*;7NZ!iCV0Dt^l9M9k1-@muNe@}lwUw3Y}*Q#ev=N&vtIw -{I%Y~rFS=%Km>*W(3gJ)HYPc#d15xJ1*UpKV3MP8b&7H>rAC<WEh|;n5^`T#@AX^=fTgew< -!WAmOmb!zZ1Zv6RIfDm>=6vkMoNh3P2V*wEp2c^L&GnM^A)?T#s7BZ_g=O^Jd395@Z -rPz0zbw};Q#Ks@6v0py+({5Ra8_6ypJD0PG5ihHC0zv)4qNC#C3q-909){W0(yaH*Q?Kb?er}oU%nMU -%q@|Mn*<>etv!|=heUsue9~=+O_N8RjXD#bk9BafOp>E;o*K5_iFJ!bLI?jF0-aruU<>Tu(7<|kfyb=-<(FUR#~*(@^YOKOk@ZCk`DP`_<*);+eR#N;+$F7F!&F7yz|aGq67D;sw!bCH -8nNjGiZPhfWE;_AS2+m<$I#!PlI53sFtzar);{CDiwv8H$L --a|Qu&u4!Lc?cSSFL(~Vf*!~MxZ^Wq|HT(yh>p+TIrt5Hp$DJ^c!U3tvEe(S@PkC!{Y0bo5)FT!XmAz -Nu!pEW)6n+~BA?fZ@(&YTym;}YE2Dl1fA;l_9r(*GWg+{u*a_?a`traRMB{6S!aipHg9ettu=j}WVH( -m85uLwCG-Dr8;Eu-9M62-s>8GEd%c}>nJLIHcOf}Jn51cdvvMt=bo5*kH)#J~;cuk)^eTK3e@?mH2EzQvZUB8@uf&Vy< -xBxi#8Fh<0#x@EXJ|~J|8X}m6@>A5B_5=49FNgonKmSY|XPR`-AiJcaIhiy^$At?QXy0q=$-F3D&WCe;$ffQC<8_RKmnTH5nFPJ4z=>bt{f&jYqQZJIsq`!5A_y4eOZ(v?TVQ4>Jw%j -K`mkI%$yh3>u_8!zT@ToB94GQ9sbY{MWq3_FpB2`LH?s+20HOXIslhJ_MVSbVyo21LDEubPM>yPrkY| -gi4g3Onkexk(@;nBz;UNN!zV?5cFCSWgS2PRAnh4G$#0t&<|lXjVP}XhoM$z5&>*` -z9nIN@qy=s3k`c6N(MT$NViXl;GY#{{(7E~(bm$i+4GB!c)Nhyul|4th>^bCs()7zbf^#KC@X+J{34Ff8PI^7N791d5Nl;#hI5Fw$QjUH&mBQqaz@Zbroq89Seb_HYiHBh)8 -Dqr*?h}%rm;+dq&>TG4Bs8i;}3r!^?&5Zk$oWx*zM$2c0(UR%m^vt}6=!-+uf(CVtG0>G`Tn>Ns_d@>}pM2mAyl@0g -z#H}eyroRw!P6mG&~kcD^r+;8R3*OJwnic2A?! -im17Kjt9kv0&p-@A9)WJPTFGX!5%MD0)iPuBB&Zh0rrh!&6MdGNxfM(}3olWBlbn^Zp0;vn~kzXT4qlTAq69DYDz`f}YmUA!z|@-M&|7_l -n_E$uw+Z8iYMF4e&|Qo^|Y#Udjxn7c#W;52hiXX?U7x_$%wt(&TV~D11_gD#v)`Wb^(X_)GmCHEL8}- -~jyc^70yKaHpdsT7LcYB7M2*MS5#>80}mhMB7~Uyh&xxCKnCsTr~Wh^(g=G5L%V$qybIj7#yD`a9{t8 -h`;YCzi+og<{W?5Oqw()Iz2u8UEm6Pgbn26PjBMmLl(UMHi;FC7zw8Ea5hGzWC<=XAwQ$az&F7}<|>+4Gi2?=C2n+b -6_D=UlU&YerESFaX!{rTsg3%`S0s5x4IH(~>P6>=fuqgW51!JhY8=F!Dpc)ekq-)PS=$5`vK=K{8cE8 -)L=`}R(p>-<^bKVZOsVKZjTNS{7^`bqYqLJkIlfhJCzNK>Xvp}+m@Z}j~0&(m|yJty$N`X2a#2E+oy2 -}uXmFrWec6ncnQh&lE=6Hfb!fd6@V*EIi9jxpbvW5Ax_lU~e@0s2>5e~lhJdR0tJ%u?tN>-rjA_YZyb -)mQCo!^Gz(J3E`$CJ5i}z4u;`BfRz2TS6azJ8*@Jq+UsRfIjFR);Y-QAPd+7Y+Bee%Rz0=#cuZe{?-l -11J|`ic}9#F@!Zuy6TwEOW>eY+-_3KBRTMJx(2W&&#L4)k>w4g($Fnl)=^*|KHl9(w2@nl)>d;5)A?g)V^x*a -heWZlJ@R7VJSF4nyw|qhTx9kH9(-zadA0>_AI1jmiJ{3-IV__kRG)wyVp?$T;%klTZEzJRf-A0nuU8l -7{~M`_qC23xpgHGvwZmqyzW@f4S$2wE>Q>2iOSg!QI{wv(;-4te=&u+<${@LH}8o^h=j6{Rnm%9UV=J -7A+ET01d!DG&GdrlXrlrhE -0UVZ(fwhbv&uZ@lq_paJ~Bz9HL(pdl$Ki6%^#Aau)QG70>_Yw#SOfj8^`{s+G!H--Kn*M=Ry24E8y7e -1JA|Bc64;qC3MWw{&n?AcSNIBP!bT8lsCRp&pFujp7W0w0zwac+UTiiJI}yovQq=m26Acr0P$ej@xRc -nbf&Y11YuEiDx>u&k_1z=7Ayr#kj+k8q4!#JKO~-&QKlx<_m4HzI#S{saHd_Lz_Fus0lm3*-%(Wjz(S -L|R&!@U3j?BEG@bK)`kYr}>O*kjb)aKjB(i+dyfGXH0OuK*3wFE&R9_zfS#v0J -PYpi_vo&Ftw5{2?2eL!hn$y@xJ==dhtd+{}OVd~brkTL0lA++(aeEjVJ2{o#inrWam#LFl)d@3P}M=n -`}t-w^|VyR46JUc3I^z+K?4UjK6J$%kGd27wN_-w*!*Ux#?4rUSg)vSkas_~MH~pQR0>gKoJukGu)rp -)<06z&^96O`A6V4(^TktK*_Nhf(J?Ey)C3r_%{|(1q(k8}Nm0!FPfdi96?~mTT1a-{B8j*uI4RbNpT* -*IzBsBKJIH{SLChZ;*qm$w6lDMW`j91_fGxJI`?~ex^nI)%!nkE!C1tkl!Q!!FSZ<(4kv6{_9`=5`Gs -vfW1QBfj{hk_0e=4aToZj_dnTBH17X^2Ec5PVg9OPVgRbko&uk1HQxO!`H%(@tQ)^HJI -k>jk~}fy4#YTUryiEz6j^#z6JOU8lmf~N5657P{*=DT?oa&XzXV8 -KPD9sL%(U_q^-1lK!{{OLL$+qjM%uA_3Djvj8~tLDvh9V@Q~{!3g(avnWg$2P8GJJ+$3>)69}40autx -Q^sJdf?df?{+lGH8kAc)6Uh~DlO!S*ze>14)^nJ>qUG=Ma6wT@Nx}vd2{X?$T2@UIXQU}=lkL8^KlzC -ZaYKHjmOtF>M6$lyn$%`QKHC$&Keu`f>Ber=S`xAKO{Q%qrjyl96)RLwr$%E=e6A&w&@1=GtN2BvA_M -6{nB^rmpj%o9bi_#1vk?CUKq{r@jvZ@<$E~To8Y@txXeqVAv1Fp1hnN{Vyk+bx+jWgg(|dYi+3UKp#< -~Lah^Z!Ux|IpjC2y#BpBlr!QW-_%U_8TjsMkBDXHh2@`vX*w4j2qI!PRUQs`Oz)c^ut{M$$eW)=(AN} -8O`Ug2b@*j1r)V=-=T+qs0^$FvA=YH(j(QGn7Z}~9I$T!UZ7!r9knKht -KNq?6ZS))kEk)Bwz1q*Us^QH*`HsQDsT|`SbeGX?LNPIE!e#V4EgU`87%5us1;#P6MN%QAF(fv`U+|j -qQ>Q#fAREjVy^KjeH`kruNaO-KzB;w)MmlPo%T|e@xZ~ur9#b4m^SH0WZ$heM0a4BUaQ2us>by+Fv)RYyx$rXIy -o%B`zG$gg)}T4|6~KQ&anw>oeFPa(;E4MP0iCpRQChI`x0L$0gkeS^RQynsK#IW*L7OP{!86V2(PO4pk8|H!cyEn1Yma^=c2@Zs5KpA~)&eiyM1{vWyke -n1y7x4P!^{n3x;)X{@%6K?uAE2Obj*3$flDcI-0o-TNd^%dJ<<6gRZElkP=vUqDvEY%k$jQl>{ph2Qz8x4CNRL1MxY!5I&(9Zo7jo?bp5qAq%X=Ha5BH -dihB5iN{j6Ip=6?eQ47i`|UaSKV>rt0MZHem;qeSS)D{>jFbS`WF$k1>$jf#2B6g$`mb9d!-Zhq{gc -8Ne?94!D39@Z0d2z;ox$ozE(*t5FF1U%h&Dl3uTmlWT6s892(3&RvANP_Q?_s -I;BoL1>Gc$#6z`d$+UlH;GuR%9_@__>fELS=`7uMz&gIXq+zym8}3TEBk1*eix_#GaF^DPH5;Vt -&jc<1cgx^vWC@wkPY7zyUgj8VB~Dwr<@j?-RV%xf|(k%^oQH416EtTwY!d{X50HSfx1kPp-A$Z%w?ev -BSQu$Wz8J$eC^Wpc)?av|YP)VW0H^k5S9|cNTVtwK4R -yl-D3rK6Bm&70mhdx}##Ktr8yBLZ^|#$$UoI6nt62x6U;a^3(mC)(m;ONq(##24gPRnp_j9bHSuj&Nc -AXAAc8y&vD-Z-vm1XuJZ3Ys<)eH&1`3GIL)GSF9DVRte;_(pts~wN|Xjfjjgc^A_hsi5%w)uf -f)xb&i3xIM!ZR`(fROb!QN-JCVa6rvl#aTWDCrAYXjtpL4_*>UQ#ib595R2w1a$e~=Smkh}+V+fxxD- -$K5JTo1V>a$T%9>@TenV_?k@$#_714?g&ysB^=Ife*EN^t2^+j95=1cSqiibpzH;i<8HTIi?&TvNB)R -ty?GZLEIyPx*B8&eR0plt9O-)ynpeu#+)0(z6$H96crAY&=o#+mGXz -~!Pi4iA^9J*W#h&DpX;;lrFUO10{p8Ie#ed-B3G053)G506Jj>@lK_i#He&alJ$stKqUE%+hj?*U+PJ#XLY -LRHbM;R6-9u@oXf$5pX-{|pn+M)pAim-GYd0u(9|4{mfTypuQSkNxe8aQXZd3vTo!$vgUTf>RQX`(bc -9ZK$jdJH@jN(!Drkiju(pMktf{QalFgr{@vn3F+@s<;w0o -5Hs`2B*)mf@>2Pp$Z4;05C;L7=Vo)f3>W$s$ -Kz!-qhZ*DN4oK6J_|j{;`2#O}grB6?qH9I|djPJtaEZ^X -Or;SX?N*g$Tkgp~!DLr*Ya@O29^=v$pMl*VD=DaNNl1JBQK35oRu#bRa$gJdP^DGcPq{8s088#&<^2tgK{TO~ -9o|0xmzx=zu0TjSgre4r~jMeB&?f&htO*|8LWA1mi%_VDL0*jDE&IV~|m6j5j74Q;Y@1LZi-DZmcv`8 -*7dAM#bc1(wG8GT2rhk(Uf7zH5Hh2rgBrIsoGR)sy8WSFSEuRXx5rz&57m=bFR6-tTUIJE6vsBT64Ww -v3OZDmOzWv5^G7cWLR=71s0v9+)`<&w$xhcElQDBk)|lHNLv(JlvtEelv`9#q$?^fsw}E5sx7K7;(%$ -@SOcwEYpgZVnqkeg7Fc!Ga%-iv+FEO^w<^V6#hT*4Vr_A3abj^sac*%zv97qhxU#sqxVE^ySSj%;(Ub -(1XiH*C5=$~la!U$IbR~*-VC-~0II*cz&9 -nOPyg@v3B+nhsv#0R<*^EIR<4{KZ7H@4TedCNmS-!l71}Dws>*80>dG3*@DSZ9)+0}yx6V%&q>I$W>r!; -tx;$N>&Z?`>Rq1MUb-D(fr`}udrw`Ic>f`k(`fPokzEE$~SLmzsHTpVzgWl8NZSXS$86pkwh7?1#A0lw!&@<(UdiR#Szk%2 -Z>jGc}kz&E94|bC5aG9B)oBXPfiPg=VX{!dzvpG1r+J%$^o+i=QRP5^0IIq*$^od6q(pm35+u^#R)8$ -(rC-6y(x_Y}SFoB5P4aQB_e*QC(3(ktfU6kL4Q4GEHH5=2;6lWv{SSS!-Cb4OUNJ+jA4?>XrIEst$YUv3Z56gETaB&G)?o9rd)xi -&LH0;{ygkL9ZO^k8+O75qdzHP$UT1Hxdpf)wevTkVq$A#u;>dR7ISL(CM}?!xQRAp{DB_{VqNvfVt(3 -E-RI`@Uvxa!Fb_B9!#IjaoutpTvbhdI^rLEdlYpb^@b}zfe9%$FvW9^Cd412D$-r))DJSbYwVk9R&`Zquf#HsCLvk>K#g{SE;5nuvA+bTbfv!QJPy?P^v2}FRd)CF0C!CFICFC$ -~0wxW!kdXvc$5CvfQ$QGF@4DS!G#uS#4Q;86MK?g#=jVrPJsFby{7lE>V}E%heU|lD%A4sjJr2>gsih --b=612kN!@Sbd^CL!YZJ(ChT&`bvGZzE)qaR}5YTjUmvWHN+Yc4H<@9LxDkOC^u9ZstvVBG!M}b!` -6yP)h>@6aWAK2ms3fSzG?_qApIw0001v0RS5S003}la4%nWWo~3|axY|Qb98KJVlQ_#G%jU$W#qkid= -yo-0NkDKPC6UaEDd1^L}?I-Xqbp5HfRdErCK_I0-8|}G#YV0WkNOJ2%1=lljhnZjylhoaW-dj)Ok9Nk -1QxA1hTM(uoyrg3Su<{61FTrQr|h}R(HbUy!UtZA -{a+jW^&3^y&-6+B_UkIl{_WSzS@hJhw55g5K3Vwa(`osSKJ(18a@u20q!k99Nqg#q%YUBh`nV_Z_hk9_p7ZcL_4v@9z4T1!`8hpr>gk4Ocky*SXDRQ$dwxmrBaHmxjOQ)%y#J -~EMOfdJeV*ktnI3;R))b%o{bRl3xZac)lVCEnK;UB@mvlG$IQUeUO~!kS$z;PHQ}2@^Ve;PDWrnw$;z -*&$EkB%mv;4$L8J?A27#+f~T -QAR(W-=|jx$yBv(-W(&=6Wr&pyli8HnSnqL-s{qg`2( -EIa=}ncm(7N+>$iJ<2+kYI3mBzC7`rffv|Ch1cm``K*$t7%J)#1okewKU1SbluT7>jh6#`5FbzGGQ`$ -Qa8nFa4j!viJ~(I1+6N!ZZjV~uL@FdRq&G%2St+mVTw -Z-?kM`vlK^pI+ljhs>VrOInk}WLK8NY9%($l_4pc(iZe-Cq=?;%oXQ|bHxY52PJj_+RgE^Fm(OYMeMq -B=~SpXl$+(vJn9cVk)>t;-VoWg(rQTC16hPOYSfiIx-K!u6Y{uXGMiP?1e~1QEVbOlU;O!PVrau9$X2 -S0f^^B7Svu>$d&dN*$#%cb`~4^-bklZl4Gylpfcnbj8W2`Rl3+B})g+w`+K9}Q9+HIgP08rpk-8p2j? -hE_zO+-oXu+YUf#qrTD~-{!eyb -sUz4S4*Qe!(Pc)@CYBPa$=m)6(c7oUDo9jm9E$oH%Tn{Z+GGV8|5^iYEmP-Cqv&qhNoD$7HX}Er=9(e -1HfKUE6S#ESXa($Kunmyo7o6+lG&Wus5{+X;gJ2Vg^@_WD;UL+eNdsYMc18a7*-y4<&4+8L&w!33wqJMt^Z? -GOJX-0D3E0|eErFqYbZo1KRkJ2UFb_-mm@WHzhn!Tilyj8w-z|trp|+E`f(MG8fCuFJ9!N!dZs|#Rgiq<#<-veKbmx1JH`-VLt1L -Z=87@(VZMrkaGw?R&wBSx^mq$oS4`cvLs5=+&47};g7QU%+B(}sQAt{h|OlLM_B$G_ay9!B -#8B}t?Uaw03;v^{}%+*iP?}bU<=Lc)MBXGPoP@$x4fo`JItoy?=UKXeY1~t*+D6xx3niaQ6PJv(CtP1RpB$zp=rpu!KaU6EP)-Eaab|_G|7pW+tic*`h%&_)U{oJNXn;Azi8G5BR -OtC;^w*@&Z2WP`SZ{7C9z!~D0(RiN9Wl~xc%uYocn4*?|<4uQ4eLAiT2I -Xoqyj$_2lt-oKO$YDTu!d;v@7I37vx&xd&n2JisyQaVT}qIjpWrcQ)=rRB}|@U^j7I7U0|?l+Umi0Od -+AP`a*{$B1m7`Y$J~=};)!56gyT*mdVC`yyom+mP{1^jQu`tRuYBL|r=%;;C4XL>2E;SyF`i=st^e`B -^>khV#zh6s|}S%KKrH{jo6*G)9~+RBYj8`k%vnU9V2S>`GmXkv$2sJ9*z!i5)8NWRI1H3gw9y-#v)p6 -JQBrAYpXiJ0syJCCK&yPjECQB8%tmi12lX&JwbK)?mjWuBbe#>jlbP*9UGaP~6w^OyRS{+G3Y}gUOMa -l=>L#9^??t;phcp^r1v*GOa*)apd*#(B6t1kPRBaitdAopocgvn)4~Vn?td3O@J7oyg?nuBa^T-q1;j -c1N;^?>}6=Qj$=}eMCUE2jncuT01PWrdinX#yIEa?NBAb8ah;UyVfB;`0vYiOwjdDGyn57gO`#+XgB)R{1AfC7OZKiik7wVpw^Qm}G~^ -&kV`Zdo>@C_7mein0^r1WB)v+zo}_YZllCP`(r~qIQNBL#;op>p5k&BNCjIe$mfrVI-B|Zp2(WQ9`3y -Q6AKtSPbQ1b1kaT-7}+M`z3X@ji-YWLKwh?qhPVPdLqcGSG<^u79E3{D<2^qV&72OxZ)YuKW<4JYt+DZf?^0Z8eg&x?RkNH$QP2ZFop%6|m~!l)9`w@B{f`v#azf&NlWjiaEWM%M! -|XKSUJMo0KHm%$hJqVc&XQc2j7m$h=~-&`XX -d88Gjo~Ed*5-F>e10#m=(AI54b1RXEhBr*W0Du4ZCVc1l)vq$BR^--xMXy`ct%*7kp;ehOc9hv<7D~*}R -K0Bg1|ZQ>HYN^UOAiXAdkE5!O@8FaVz!YyhPxvJRWT)h=BWo}VCvDW9MF*>_*pylVjS?87|xQEQ+Cp# -f#g-VRNapYn;jR3(NGjCTp%$)cm9RSx2T_DVlYkze3D#7HLuxfXc&{1Qut>O_PfspY<{;Yf9hvvaoFP -|{>+HBY4k<=4J^s4+i6Lb!b2@5C6)6P9<>^y{sK{Jy&%O2e0EGITb3RTjIFnsNAgMG3E$m_2xO;PdkG -n(ow{>hBZ4I?!BX7OD`6dSA{roA4^mTv3wjXo!8Q(xP@ra5Op+RpBCM@p^-S`tz!X|SgzQ^Ae-cvQ -2UdkzJ6PjnH`=D5wcTl9m{urqPh>>0M+FL?Wkk8ybjn_bz%x+Y!1ArK1EA7%^|V1)d(U%*lA@I^fm+F -+j)AZ*(48UMGeT^+WrmvR=}>}Nd#gtW_xE4kkRaEc6qqOra4#{WlMeqv&F0N*;Kb3liI02fylC#*$7||9Xj5b!^#QyP{SeuvT!iW^*ipc~!_;rbL#UVbtBJ|#EHU!%AR_vB8B{oP7a5jqwO -6S#bon0mnoMu1UpKV)nBh2s$m;?D|#`s;m<4-ikPbxYZzd!xJG@(3cAjn5FsxBH;O&V~w0}7N26nBB* -Hm3QMjWIsG0sGvDecmE(63RCU<&EhL0JvBs$k*yzBxY3qbukEeB`;?`1 -yz{Gz;K+YcO>sCNiEl&q$1`7>3-)a#FrL2(3xML1szW2sEI^F6mo=r)IWWP-r-YD$9^_CgkykB#VzV_ -(M(`@TkCKNHv`{NeRsFX -R5azw6eLqN^*ImGf_0P3q{R(CdNGQ}xE}KZ;D{yn{=z$9k}~Z0-m>CH$Tn8I{~rAtFQxC2I*v;<7XS% -tk!r$sh5Y>;stlV@s9KyN7)^&}WJDCG!OehtBW)Z|x(2NtsMI*rsaKY`CtBjc&0H@Ka;8#uwiHf+?)) -hZ-y2rkC{+CvN~Le<)2EK`6&(s68IAj}CIhN-Ky}z`vQSl+B0$^WI!wd{^)Y8RQCqm5lCt{Cur@WY1p -?=M9WA6huL1sRuAzv{kwQxVmV)!)FOVe#SLgVHPk<0>yA!*XxmgN&QQ4V`xay3-9S&B#*Y$FT`$2dOK -Z`=gs}eHPb%n7YGG}TBe9UWN(jeG(b3_mlfLE_Vg61Lm3_odV&vc-t`KV@ -Ikczud8@6Ji6cPo=U9%!Vs%X>cAFN+%n|cmS&NpWel;pODN<-=TX8nF6| -kwRJAv~B(I*#{XuIk@i`{`0WUzt^R*=C8hMsUBH!!~s9|{sEAg==k`DjwtLoG;s)3^+!eh17*(JEJ#= -~{>!FgNEb!XJ>D{7XE?g-Wx0p=slM!D($iC0{kEFZQETg{)f6!=}xBpf4|&hKHBzH0m7}0c7X0u3QwN -&Pc&IuB8aC!*T&!oVlg#fg9A;kx?QLutm_`O46%D7@bGv!z6s_BC!Nks2{Ae(xCe~&iUO9%WovTx5Ll -sGMiC#I7s5^TFCh4J{db#yy;_K`h*V;cT7)7B87z}$9mXImli9Mauju?3nk!6OY$Ju$R$=GQIa3==dZ -APOWWmHLK!MxeeUl)!qoF#w!`mk6n=i3&b~pg-Gwu{NOjx~)!n$lF0wj5;O|9Jx+1VVAe03l93e(R(^$5|=;;McHz -498(`Pe?-#h#tjd$5OxB#Qwnr>o)RI@emaOy|*g)Gg%nZV=08LV)=D^g58-X-04lFT$mSJSW44iG;$0 -Cr2mz&j8r!uvA<{+KrPSi!w7MO-la?=AYHV^oPgsIEQxJ$Lhnwkh{D>U*-=kOyg3wkDVqwbivQQMqpW -p!16Fc1+GrudI{-Bf}d^7Y*uf;wtqh&3PG6`*txgh=K30_^IC#UKiiQDGC3a4Zd32==6f1)JV@H$dM$ ->7TnC2>#m&Ak9uRq5{phgz^Ig?uGyg%ian^0T -5aSoT_b)!wiXyLb+Aa#**UI4?8r63==%8Mt7!vjyyUH*d~-if>2%~ly4nf6JauuCUT+uzlTT1I}J!aK -)7aM;t1n#QwkBL0NY!)lt -z{aXP{8W0W6!JQxP_DxtIk1?eJAbzUVfn9iGe0-Zfyu@L@(cgs-2Lswkp@Ez86A4Uv}MrYGFnj3F$L6 -!x+$D%VAiUjq&7lgrFQIZu$0#v3B5~f6NO-iO98{mK*Nf3XZL3~i+FCMea4S`B;DEa8L>zXz}iW8ju; -2af9_taA%ugMGPkK;okT48whL#+#~vQ%W6Lt|xP9J&zPL4waBFg&1Lef~KO+?s$v=PljF_mq9Of$zNO -9dnI)<$pVl{z`CYjZR$oM#4`pk$!eFZEY(DWxiyXLt9DWg~zT0SJ0VXgN25QNX@&Yhp4o^We~rfkQlz -RnMm4BK`N+O}MI@Q3H=&^ih*7C2<0&aUa0E?koaM!NR29ywo9x=OTV(5`C8=yR1V& -T26k$A)-V*qfE9hg1fSrk<(uL^o>3d@PKx?KN_0pX;;RNd{17rcFM%+wJa0!HX0=txw}S{QvTZ0|t21 -+fgtwJeuWV+sQ#v*xj0O_Ym)Tiecob@L-2Ra5EXV=uRQh5^`$mgA8Ce+UO_V*ZM)tVfW49N{PLw&dgr -cm3K#aaw0>0R7DLf5CayZ)@Zh~%=G_O3Zel#S~%kV=m@$DPULdl1S;-QCuxXhE`cOa1dJZz;Cx$Ksm&-##tHE+=AU*j21kwy@`2!2h1U!ECw*TBnw+GSr*i5%Ns>-%h2sMRcTp85|J@q+zfcc4 -FW0EL`n@S0xScxFPW^LG6-^$H0B{gw$W^D)jTG?&KH8LO-99@jijo7z1oc -C+k-?xyJSoywHy)ZH8Uba(hQx{IQ0PdRnC8-|)>(W%F`_x9L-tsX~YRI=0Qf#n&?bE!RpA_X-d>?^A>lT3kILo?&m#Y0eKN#S8YH!>Yne-uz>eFUbWqh)yW=nJ|hGpR -BGTfYl{4j4>f21_~yd51c5=RC;9fy|KE@r(yK35kO=ts{;28`rf)4AnRsKmtbK%t!STC_AJ?aTP$s4& -B2*@ID!OBncpUCqI>q1AMd2OsYKG6KgYd*mciCGf>*4=s7}&nNbGoxQC5sBNq3xiMNu0`%u%J)uQr4*=*m&7w~;CTW-xwVWpdk7It_>0ZaH22@Ys -ObBZdjtDc|RR>8@FcT2w#?fZejCAN|#h=;DRq8wJW~oE-vu|J-t%N((yZn|`R$tR<^H^F@#7ff!fvlK -^b-}=2g7V=}0LUKqF`*R6l3pK1j9!%o@1c7d)VF@bj}}qQEH9JtzkpW#P|6pyHYv6-@CK4oQj=6%!zZ -R@*|54CI~2*N%ZX1=7AC>AsLrDrw0$+-*+})A51c@af!GnLBXX|;#BK&=`{8bsPz6tqhkb{cE71sB?H -Y)T*Cx)|lAHiAvy`-FhF#9(aKP4l^6411l`bSaZkWNn1{ -%+zHMK-h=I#lmfrxGj#TH&wBR}92XziWAXy|)D=A1!WL$m$AuTl_3n@-uG^T3i4xE$VwwGH#dW0rB(c -xh*RPO3d3-mV&=XHGZl+$7|SLkoW+5!^grt)`l`LNS;702-=ZtaU!7~ipO)Mhe*X1KF$?>Ldg!!y$B_nQP1 -FRT0Tp?+kyP<-DcRb&w{YTQjl+2=?5+g7JeKCmf|5C@lp39!_e6iSgJapl#+G5Q2Gi^dW-|5`Y?*FYP -&!u0;5qKf(Eu}Mt!x{s1H?Rz|AaK(1W!OLy-$s^AmyU4&17~C!oNZrj0MHe(pxdcnfArRbLk(Vwjewz -AQx7et0{r{gukvzUPC)14bCbahw}e|Es5$2fPN&WMT>l4nUaSRRRumP_#v)#XnJ6Pi3AALLQ~l9Eew! -Lnl}_h`xZP9+-%X>%%GDl7r7Bh{};mp5VhN+Ay)b)+`T#U%T8-6snFS6vow0b0K?yg}Tnbo@VtV-R_- -6V3Q-fxjE$hA;Q@!1NAZyB5T18cU{IwJlrojRc+@uCo!S)pq!vwG|M*SVocyGb*%<`g@@H5pS@H49+` -_4d~=-ON}74sXuFS&S((yrB555`H4CEu(?W?tAa^j3DP^bDL8Kf|-E&Bcuv^V0wc~z7tzn3YRe -^55hjU0wo#^GFLIgr>KgyEOA^Qj!DI+A;phJOKGOj|AmN1`^yA_YgWTEIN~G<+Y5<6qFxhKgu-CYrj8 -x(~P9LUUCN*Y>GZ-{SdmEGni4ssED{Ri3|>gd0*D29mvx?ZKjct_B}WP5Fa2@E-oEA028tNfXdVxG6Y -8T;dOl{f6Gxo7wDVjX1UaP_~IjERqWQt-m@hsdt>8o*@_@N!^#;APVIsKx0<3PpOV6K#5+`w*rvds`QuDVeoB-{xqOfG;jy7`pUp8yy|cF;yH6u0t&N3#O+-$1EHCJ@N6gedoc$ybTaCZe#+jAD0wCfMjP;e(;CJ9D^v7CP0 -z^{@|Rj!JlPUzBbFZL)*8(NY-L(yC)0Sq(q3SCmBZTr=lf0uOC(i5ec~F}u<=cJ<0m)KdoQ3Xtf#a2lF=#&~r -34OV9#kq&wb*ZYeD+(NCbVIy4ZiLA2zxLmg_h!aWr|lviNmJ$0CQc5ZglD!&d34UutU18p6?&YI?Twr -8?&NgocdHW>#46G>pEbU)~aP=2*eXAhhxHzzDDs%`hd{pYp|+n-UIahSr7JOu924x*JB~CD&B8I(Nvv -nHK0oU=^^CMokY%C@JX}+kfp0mS#D{EoJ%)*$AI)0*K>hPjL$?qRq#Nqeu9sMei#iarx -vdvX};eq9+ --G*JNdcT|7EMET}&A{s);0z=~K$-HZZo&^d-*`7B-1;8+dvI-vO3xv{>NQ?4+D*1vl8ghfvfc>p=-!3S{NX{)V|MTjT*tA -Z+yG?ir)sk%5C45)jCIQN75=wrFZ*H?Z6lpGOJRjTZv()QNs6Pop+6!{=YE%BQr&|NScTL$yC!Ql=#F -g--p6kyIJFEEQZP%UeKs1|ScTi?hR5)rD)%>jW8L^ -U+)0g70Z^?CutTB@Zm6T5NowPh1Q{*!GUce*?gU)+l?$>S8a2^e* -N#g)K-|H~{w-z5*sY_6qV)SDt*FZ`bhu1F_WDfE+C49{hK{iH6}pQIo|mezOeGuWhr1T%Shu&_7u0=M -@&0FsrL?eMaFUb1(-BWmF6mfMAL3!i(X^HDiN=`2J~oE1Q$oYp9zeD65TChF%uH{+p1vdDVG#5T?Tc{ -~(VcEn(PB5UI2#X?CfQf?4G&LJi(Z&12szxbT|6ES_KP_he~3Mb0b^+x$lC7dGPpoA0Tp-PuE@QN2@8 -I%s}SDlIrBlP98*XIbOK_VJC;Wd=2MOy{YF@fdEcRk9nOUlj|q3ZiMZ7Ea~cnm5L6Ll?L -3D1#dD&c8B -c(lVRK_Gwm#=wn(flv<%$FKQ(=AWLNJooU2-@1ljbgMydx}uHf~Q0)aLQ -DH&BXbnpL|93a=4la(2W7wz39(DS?orZ>AhJTkfpzt4p|IV0eU%1RBA#?)!pWgv!_ZhT-QYN>VsCssVp!dbiT)Sm6lG -bf`-#{21{g8oP567G#6Y00sQ!5OP+jqX(hIhxIJhfk(zPjOu|X)dVXA3U`a%j(X9e@3}F^$pCzMqE -Tz7e%Gi;(QtvD(-;?cEHbQ6K95Jj4+7*QW7w0`^W=yPO{A(N}hC#%%9mVPbkOkPLR{}YM-S^>6}n_CM -2FzI{UAPtw=^HQ(WVLVOEgv+Xcew-PN_yTO#w+Sos-{dP^J~g$3e)@J|1Z1A4WC>n|}~=#!ovNgqV%Q -F}ctwl_~)G!IB1&9R6Eou`f&rnLUhEC&Fp8@eP0W1@y?|H(eO_(TiZerHzaqAL=J|9+$)>YcwLN{VSl -4NwG)afm+&&T*xnEvUm+Uz28geW+$P@~rL6%#84HBQ -|~bO{L2aOHX6!kt8NJN`tpSJn36hDZ3Rbx8V-z#b^XzQwc|1jdO#<^!?1vmWzPo}Gas*H%tQ`He8HyR -opP5umt1(#NFz>1y5YTtnX2z&$yq43EJ>rXlJFAb*DLO;tS!5hi&{GfF}iC5!G4O*nZJ7y@M#@JX_j) -oIUFj>9aYQls{lf?5E`Ct_=+<#C;D0sNpW$BPCg7+do2AGyINBERqU0-rm+w`{wX>bo?Px^Irj@b;y1^HB0GI)~B!K)mU~3z1YhFaI$HgUJu^$HV+F?}7odEZK)K`k2T5fQPeJ_z6 -VWi2a0W7iPAzUCQwou1ACy`7!*-#h*OoxJ90=9Jk6hE5~Ey!TJ_?lKS;UoJissmVX0CGY7WAt}e25+X -9d9E5zT7*}GiMa1qCW!utLm5w7Y}y3K(&V0JE%mi1&HK@817(a@_2*Yq;JHNaF+ZwWYTu4pCQj2w7Q^YanQ5nmvsGer2Ou~fn@e8Y>V -v_^}U*nZH0&Xu-p&391mRk7zDoFix9~MgE@rWTvy$pBL{Rs0=q>)L5e;I)Wa7qOlUwh@V*M5;DCw -1bU>3N$Ya0j}8H=#gEn2q!i9Ek&R*2mobJ8mZcu+5UE9zl|$%gMU)S%_DXGa!Nv#=^oyZpVG4OGC)Ee -H}u1G0v0nsKZn*q8Xm4z8G!6@U5;`hR3wJ;tqU3fyqax0E(`aNbpqmCEDx?#UYy&}eEZQ(zPb^GpfRuBZ9q+qc98OAA>iqM(eix$8&m~6)yj{C -BwAk1s(m#$y2cY(^7K|?ji31qu>>PgH?vhuLblpj_?ze&^Gy_|ZoPzJ*%iJ8WGNlGCq*qDY%+X@bm!h -*6Qw8^iae_A!J(vg57A-d -E_uv$p|FK}9q;3b3n=jr)~LR^g+u)1O7qL@HIN?TNvqV+G -Q7H@MQ<=@a4-G|r5>yzb>p6@AM#DQT<;n!_itYAga5bXG7kaON}%aY)Scb`#Yc&yfH~7HM>u`{Nvx(4 -HKPUDcG;2%%zVByS+X=_Ru^ZJE;Ol%>q75K5#Fdi3$jAKIN1wylQsJ-HKI_SGsr8H=f&O=@|#jXxRpn -;`;z?b=E4Lf+ASOXux^5=*8n{p=XGx_{03WKHGTSkbLnR_Ac8di6>W$pNk0y4oYVl8Zg@vaPheN -ztllXj_pM14^^--<%X_h#fqE$Y3qam$*8l7JC0Wj1k+$;y;1Ehh8@n+FNSI`RraNd?*TO}Oqf2-lwuv -e1^mrm$@J~Y0O#De(BDHH_>ZpmFqibXFSv}6>8v@@njvN8FTg? -{Lt9cMJ+%UF~f)GkA$<*IeGbXjQIU+)QOIF@EQfuI)wb*Phn+z>6@ipol-jNoiEh4_)*$b^*XR#DE`4 -a*tI<r_xB<~m-n6%&$@ii*ruf1+n9{;;QCX3bQ}&4~6RGL*x!goGBjVPK*ZW{Qx^vu{So -tdi3TQm}$4%UL5`l#qPcFtgi?WR4`lybZW>g)2AHj$`ga?PyWCl`|DevyX6M#XL2psAXz~Nl3d92%T+ -cE4As|g3vWf&QFW~4hmdV{c)ncqMn->5r(1u>Ijmkk44tKqE%F!Bts60fL3>!zh|Tz(-?FTU{Mq3#8w -k*U4j1n&A}f|}LVM_QUM2_JUx8cKCZ@O0vy(_2@hKzq+ad!bo&KdaZBcfHQ#^$E(9S5dPk4@K%|YbV) -eE2<|xuD4*F+Itv?<$TZlg(u5574!=%nU%N*SKp7&#&>g9;oxUmZqUI#sk`o!ny~|9${Uz$GvgIU@^T(W -YvI9txM7L^5ZmhJhFRAeaxD;lIp;2c6Sn)7~@he#gh(px7EY8@{^8I)bJQykleuHAG68EImZOq*aJ!A -(HZhCMUyDfog?Hg^|0E^Mm#xIU%Kxa*-Mu~jhB}Y$rjQml#6sbG5 -@ig&E>BplKDK`=&k6vd~BYdrQNWyLsIT3lT*KQ_^8e0cYTzBSOG-G#7uZz{~gAH+~jn$kIRg&nF1O-< -91L=5$VS^#*js6*d~6TQ2zUSgr8hTZeG)qLv~al7Xf*tW0A+KBJ@1Lrz%hTN&?2O;BB1oPo-dSYNOFplcJKJ2GW4Q<9YY%vG1(9fwXvzn$jL%3@Z`vGXO`_+`#M|Ydtb_LC&FUwRM;Y^N8_AbFfid;AeQjLS(Vb;$O(yOV -GrO-#jOJA?b_WL3+dhcqd;oH4dz6cp0)x<4CavDK)<|PHR&9OMuLk<7{E;dl86oFkeuixL$#5?nTg@( -Gw~#yW8P%tg7}AOK5^X!oz^#S{#}Mi3-jC*e64QKav-l}Cn4QIMz!ELne3lcEQk|i6*$TtOy4sDj)k1 -VOOz#~?ZOcKt`TMJU{~aUBhH^1P{>+HnOp*7FK!bn8ENbH*0Ph#9!AFuj*^j%@gqL5T93I!$ta+}n@` -S8Bvph~%mnE!|%6|QMR&LZCHiXif&jH1MrblVYQfghFODha3?!8wItZg99>71wGC-rT-nXBd->t{$DW`;cD9X{3|W~Imzvwm!rZxYt^7uJbpchicYu)%+ -nDz34JmOY}~jIp_vZSJP!C(z_PYiMAnPpP)&TE6t9p9rjnc`H7JL&KiRqm@%ud5Cf-o^uMwf?DrL1Y}{}OO|4qp)2y78t4u?dpq1_=ZOPP$=X;E(&#(ACZ4fIc%m0Y5K -5o3A5d;r=H`b!kQ3+-KMFm%9ooYs%L42UU5#@7l!~I`_#y-F23I$91%j9$|lwN>GyG{E7*+)8z-9WPK -4z&_Zve-0h&^G!cp1azEwvqIb6toS+7qshq<2SmF8X)C@ZM9R%-y)QIW;hDOV2QNoeh|VoYTvLgFunK -!g2w8BkY$Z$=!mIhC1%GRl;LYULMhg9m#-cyaOo1`asQ=36pmU?-Th?c1mP#J!Lb_FR)H1_^R;$_w -}|96Zp*nhB(lsnaayaz2iDJ9}1}oG38|+C@Q*7)e>#Nw{y8y -*H{l4SiAp2#e9Fu1zn(Yi78<1Ozd1mu&Nx!5Y- -WS)~0+xje^feR8axNjb1m^|?OrPt&^+8ToWnI&ByH7)7#_$5);VM498+qH)B5K#ZKf9*i3t$l= -ms6J`H)#4Rqn;WtN9w_8I9bqdwQTQX-b0>}UN!dX{gIg0R81&>gM0kT}W*fq&wKt;!vEw-4<1eRz9E6 -YRY^NGDs{M|W!IoNhcpy1WQ4^UK{Px|`16W%{Y{l>Qb5wREn!{Bbj3l@YM4*hZ@Fy-cOsN -;)Rqe6xsEFCX1Hb97~*i%pVX6h5i68?9oiJk0YHNQA*{37wVT+CLBQ#ziKI`X3j+P<&qrMG>;z&3prC -Ap->VztI&NjttofJORa~a~;#aE@-=#ZCOM&qwyZfIsfQ&oRLbqy4tSNd}u{ef=w{3W{7&#)a?ZM=CFy -wAp=9QF<@se^+hh9)WRW~Jgf`t}%yTz1HH+3S0rhCc?rvds6VF6`ho)o&LBjV4{P`H^T=Rrb&CtG$sC -Y0PtI~Wa*kbaFNpNO>eCcIlaw8)xk`8GR6@Ta)93Zb9V2&UY1*62OT=a0qkeYc1?E!bM9jG76?S@s5< -r_T^JSSCuLS-itVkKIyqS*J4+`gAz%>drPY=lf?!q_MkquPW1GS?X`kpoc(D>A@FP`IN~nU0#Hb9{Ew -R-D@(-58=Ze=;vceF5UT1327y~QvNAGPphCPo!&`ov+}LsYH~79vfNbTV^$Y>{EKV62zrh(uU&UuD&b -tcLTCh7G)eX~#5WOrX&RpMpGd<=-(lpCmjlifOA`ZU?yBPSI3-S(nzDdKWq4D~vX+g{w!Qh|O%!y -A>F^WM^80 -SGg$L@XBfRll9RC}tg8@+h#2=?ME^x~Z8K8FP0KTpLE`FNzEu~w(Vjw1u>M^inJX!oL~JOAp=jE -Huo>ZQJ&u|_*{UUZ)itncfGXehbqsG9C}yu!Xl56v{I?yrF)gcev()OV*&0Pb=16H@+Jz(xH}n+)R`s -4J>&@sQqRX**J-KCF9JFd8WIl?w@p#BA!ZZ_M2juj1R!W@&yT^|tGw3L$S=LDs --J2)*xi&JK#q@i;ao8dq^8h-g@mK039Tz%eqo>|(622^;l+O62xC%Nj{hNjhN=pUK`op2)4JEn#V*9+ -&6lU;Q$vA)*13eCQUg)`e>3GOwk6)*>DGY^W&PMxp9-_ZjGE^RparwxWIF|dk5d7>zIbULzT&e2rIpl_hsS -$JDSZFs33u__*N2sx8(v*79BYc&feE)T(dp>arzScfu)6Syz-|3{y%5jIjr2At|I4n|E1MsMWKEX$(~ -UtdQYxhm1U%kTc09MHZHrq`fn*A?DiL2pac(JQmGdgXQKyk+N(UQq`wi>23rt6%o~mZ@d)#_InE|E(X -uF&S5RZ=u!N72aF>Nv;Wh%a1~Zl16%I7D}p3+%xAjH;}2e?MJJDH0{RH>gUGQ#hJL{)` -9)FW`^U|_WjWXe5^WZ%YO||=?pY)x|ZM8AA0?^3_lhe{{cU16qO^LV!O%eXS?KqBFnMo=FhTrwD>K%r -2Gc|=v}`2S^#DTxq?e}$uh&j%dDo4HJcer&-D%YtYK*GIj6 -rG26=A6SqUc7dJtmNveSh|C~tp4*WfWprf>-e16KjmxW1OZA0^H6RMtSD8##qddC%j#ZOlb0))#0>YX4Fuo)>*ep`FZOMxNk@h@opSQQyIE0cQv7*x)%GW(o&=Q=H0(ctIl-5&9LkwW*f8d=pfm_tw|Hk#N& -XC+KAkS>(p@k3%Bm&#=ulaQ1LQ04G;2jV$aZ4p3)T^W;HAr6!+k2`R8e!91dP8uUz4YV@Lp^LW6hDTS -9&~CCDvfE-gYaGW<%1qw94k~DN4b60o7MP~PJ7{OVO_ -JY(roo_^qpI##c9g?LXmo@m5*1LQ);n+Yn55z&hBKI#SC|J7|1yQBOj`xLkVwTf)9kL7CchN0302 -RgBp1I(&Z2?g(3)L#La#ce(`6KN!;|2t31UA5eWqPlM@L;jK^ug8htld$svYsuQ)fW-ijpy4 -?pYlhmwz!>bkE+w<*U)&0w!|wLW4tNsHU_)_5k`-!%&*OCWFYMI_$Khp4qsGc@q_?#Q}^7sMhK{SK|^ -?~KKz6iRO+qv2+GmuJnAiCJO=D!RQfC&i|2QP{BEQ}*;iy-^+naGc0$JDT1#%kL^jdi?B+w$KQ_wM!G --Eut{B1rWxUXOJT$Tf^b%=-rV+Q8kNR=|bLCU`2q-$ZV!NQkFf4;&pOTDVA%Q!AfXIfOa!-A^{)UCsv -?|M!P2i;JpNmVb=IF;>9}}gpyyNk_LwJGRl}`{UT*VWs5Fa4){V3n7yHe9iB2zM!dJ=r2JSU15}U=70 -CCM9+Y#-5~AfS2PAzX6zRkw*5ezu`Q&mgo -tEM0TK?QGfh8$hg?FLUO714%=RdWQb%D7A0NU~m)ZU5J;Gm0yl1X?CQ&D$}Q1VI)`hv*GV$q!mC=O@X -t(vXq@eKP#U;ZW##PQ=JmO7B&RdO^Mj)y>>qD~wI`pLtD67*^p@7N?K(?Kr?^5JcS0lliEH~8J^vLU{ -Z*Nk_(#NvJN+6#WBhH6?u)DuIZfW#ck>E685`rTi$2>qbS_`wmKyhZU{POOQ -u-%0@?OUgOOx^GuA_+&=fAKq4jwY;~3dEgb+xG?<1hx_akalL-@{*4;1Zx-8(1w>7{k)CFVmyPsBtmy -+<(l2Lt|%;EDDqs8M~(C|I%h&TOM(dva}p>$UaUA!xlph20WZ*8{}0T~3A{a+^Id{`5d3)UBhK<^3%K -n8uIu+B0y>r+#;Xg0}5g>{Kjg$*TU43W3<^-T!%cz2(&RSyTCR=wK25$ep}C|O#x`Ik?J8=5KuxK=L2 -tm?nU5}6Z>^-RoGE?5Fr#omE6rMGAUVSYLe7&DaLtfnGQ>pkjcPtOw55A9MtY-=v7Ll!WznQqbd^Y1T -p{}k~7UcTtmohgNMOYc8-0?vi*Zh^|&wQPQ?6tX>Ctn2p}oy`&6toK@u23KWB;lGhzQWb7aIk`a09{e -8xav=fa6TJXQUm{RC0VZ!oC8g-zBs+|o9TD=v{3bpx2t(C92$;hFFk=d`9f5dAwub*{A`|2AS69Pk3x -`c|6o;A&;5=4HAlXlM-t<%MV)H9J5M2EwXKnz7N{X%oQa6pRAbW0$oP(c%SmSMpw1uF$#=E~yFgsGVY -)-}=9$u$B~wJM&`iw=Dp!7YZ*SbE92ZB3_=<`Ev`@gwRqHN0Ry`%AA!zE?tuNZyW1`CW>Rv@u~j9(4y()UU -TiX(M`nWgbM@_&a!WPx<{}0|;knUK;02E6&0y9J*pU7-a-x2U)r!TPNE(G-ao -6=fZ$kl2U9~NGsP-VR8QaT2`lq7po;15z?(f0Rx9x=#yd#CRXioK>Xt&e_R0r%ANOxDDqd!c_J-p)4z -1b>0j!J6L+cnbq7=d5GllkxxH;{5KfQU=C1wO!8v|U?uJ#wwU6CYD+laaj~Ah~xJCexP?_NlX6Fp{nO -3PczBzk__fS#n=omZZLh+5x*jO)%jNbetrkwR`rf`+%rGmP^WE;iYX1ccE5-1qzk} -Eb*ns!yg25w>&^{Nb0c;9<_18}3@E2{1TbE_AK-sDFnrH1%R -*QX*&awZNLPGlgVj5ut*Enk!yFiHUPsF_T%BTc$Jv|ly^s1pL)1hT{qVUw-aa)Y2V`Kp_U>m6Q@|$bC -TMBpEcE4}lcAb`O-a -729)Ca>(frYWJ|wKBlrWeioLKe9B7duDQQ9O&JZ8;e~8KNVz7ig4K7Zsdy1C;A3bz;8fh>1A{8>q8*I -ss*YYg?JsUa_g)L4ysNG!OzZWT=(U&O-_iNNm3E1Y&06^9T_) -jTN#+FKWl`0j^w!XKp-yb{gTo8qc-H(`!6$h@)reETdDEk0U;=5nde-gIa<~r0^4kl5Ob8+&N7sxsP7 -FLP-_9OczT2MK803k{Xf$^MsP;@v73!hvb<`=c7W&+mxO!lw5B{{$z>E^Z5kS|76c`EfGrRgTx}XZ^{ -u$Zn0q(s*c(PrQOQE*g7xbg5(c7$!HCD*g892Q;XkJk+UV1XGbfu45|?eCoUT-d~WiDp560nyrayMzH -wKlhjnVNd4$iiyzICKR>^~&YN>Ibl0l{V;`Mk;*$e_%cmXPWJ2v45cmo!1;*o6GCHdG$x~yLCbVl3um -K+gEX4BF9L0ErD@T69*$J)-8>cTHjZS2V!ER;C#;^}q$fPq=7UN?(cK?@5>uM_dQ#X65$U$rlMyyx;p -VHIeUA}K@)0BRS~NFpuEx0ZQaj{{_`^_0zVJ>kzUN+OHt+_k7Yc^p-{YUaI0YtJQO*(I8_An0w9Li3x -vtUJ6KZAQFhRK3@tRX%CLPm;$Vsoj=?nOGZ)BRCIZK98l7<5eyx{5x9JCxlnBpga|F*St_)`4k73k_1 -}jXM~bX5K;N85&I6o#8UZDrE{53@;;Uf7)i+JRj%ZL((_PxNc>QfU`qMOaaI>sVdF4(JljvwcNTcE9k -LzG%u4k@o0MKF(dvfZ2L=ZHP!I%82;{ph(!C;x3^w4KB6ra> -DCZTzmIZz5|07D$L@@-xZ3D9W>hFCiX|8B9B=5MG+Ym01mm=rU|c<3A%gU2t%Xk|ze$8D6hF&AecK$D -zXBbUJ*BO3d@hG(EgRuBsjv+LSl>`;5%+`~mn(dJcc(T2a1A`16c?MD5ARklySuYDV=|zm=j^_Y7MuV -(R3Av-wOYKax>SEIj(_eo9+;^PcaJc$2mR}oqB~&dcRLY~J6eV}j@|+|#m{!?J4+j1v>}qi3#u6)phl -_Vw4YUnre-Le*)Lc`OC5?^TeaPOLgk@StOJ1rRs^N6PEYAEfd5N$5)|*rN*78jh|l${0QsJ>^ziwMx* -q6X;BkiocZt-u@PEM?x;w?sQ8pE3c4BR!TYuq(;`;g6!Tg>i -6t-i0Hp+D-9%wSx0K%wrK)vj=m|2vSWCW9EVm3;9OE+LYCj%x?e_4GM2qZF@ppKzBoFrjoDGIbucr%1 -{8Yi*wdAlzJAg(o%*b^9eav)2-%Ny;aFVx!N}NCGCrOE!)R5i5q#DGpZCEDIIs#@3$~zQ7} -6?jQ|Uz=I$hof6i$EPQHfm?(RKDe)n4|5$I9R%IKefMDMcvjW&_mgR27Ec%=x9f&`7e;Yq<&87AkvL@OP4mUT!5XGf6P0Z*4x?j?Wll4e -3{uSu217z510=Fc4m{KOM`0vHE?wRJ=jwq;6Sc18!r%^6vJ<9pT)x(I{Iu@I(_mirE`%yt#S!rE;)s6 -J5V}T2&MgFI08xx$HmYGpY4=P3 -+jl>^C6VD0ZUqfOAl5A)V$n_cwJ{1i?H=-&1|-$+p3{4C0`j{nQGIaXG -u{r5F}N@n4;nbA|%YI5jSO(x?@Meg33oBpip{7Xf~p{xA0j|<%^T8!I`cB#{U#Xnh;A+i~GA<%Rw{i0 -BA9d+dwittOESA5=wzVNBRwYJiB9d_K~nEveP-81vEl< -)Y|(_|uyK-mC+z9(Kaj!;XHz$QbtK$gqtoLN<*Hd+2zqMksekbhBV{;G6{MyE-Mk;c_27eq}s)tN>fP -2yfx%9+LrOyH1}J<|oA-ww-P?!+W-x@f*qcosemz#CFmZjOsH7k-XxSVQAgzXFyW7V{ZP1ce+sOw{)5 -)-$CAEna!#~A39k%l3XmQ(ZDbGZF&;UZ}bj&BslnWS-d9(-qlxM2gJKs+xH*7rv)->>WPo|$8^~x6w^ -{s1gk<6p>$5co5hio?#F2g&6^hacI%PQbTc{mO8E`y+Xwg;AcD5XKH~O#pF*wZ`H!EkMF)_r>RX7z6! -(a@=1sXdoz{0^eEJYW$vb{($cKwqKF7HFjJt@CV3XPaXpThkFDxUHYmni=AOT0T^%*`Jm> -bjxc*8Qa^Vx9fNXZyWAfBd7DAN_$ooRR$Qhp3n0HNFk+;r5{0j$W&oe; -lt4%1<@@w{c$4x~9A#GvxddSL{(ky&29EOJ|HIz9z(rMUkK=oI4lf536vD@-s1!a>C{cXO$c)a|OrnB -XVSq3Q$zvL36w3z)bb%8^T`Rj;*UHMe)wQSE6JLO3zOo{%q^{PX(ZU{xWuE_9d+#$dAXMM`yZ8J3{y* -P);LKY4wf5d?ul+oGuN{Ut`P_iJ%kDu@k+*AIxvs$1IJA*4r~1S`0uZHrut6|}iqAq5R*@5V;TL**q1 -GF(al_lz_*rVy)gdS1WeFOsk=|uC8!r>=gcmwk8KJ&`;eY{9dL -4JL+YyW%iI^Vw+E{YGlgKgyj`6%mMbig7j&zLgB9%`uH&`3W_><-0!nrCHma!aWXxO*F^hEUd7ngp$q -Uh}3`;E=P+*AZf>H7d3BnoIN(zY%re(tWS=_IlxO^d^himl0I1_ED;;I{|w=p2l8D6H&P*)KqKo>M~U -PjM1+7x?;-Rgft3opymmOs<{#&DwCy$M=Pc@fTL_(f!23Z>wD2Um|5c|ap&z7Q~vB>hkH1kc@ql!U=F0(*eGczjY1_~H370-EMX#EBExgic_Ss_GvEd9AklmL~@|d{H>gQ -+phj;;uTSv36PvCQmE*@e`&_^d{{tPtLG+Vv?_`de%{W0Kxc9aL -w-u4@*$^aS<;ojTBB|J=V+U{$tjEMoCe!q7y2VI+GY66((&;E#se#v{-^G@HsgShS-hu)*f_i;SaB}L ->fMtK|3z8{;s>ECao)e|{O9Nq@u6xC8n=WXmzyF9;p3#autSo3$Wo7<#ehUD|58HOAaZuq$tkgI1MveJAE>$_k$vD)YT3d1Gv-Mtu5Woy5r2>*t4jSBB~eBV -F`;?RrL(5k*zQcP}{?~V!)A(P8W6-JM<(&Tr0wO!_i;pC`?y>y&~ZNy)=$?_)^3SR0@j|X8U|P{y6ID -nYEMx4UWO+6_hY8D|C&^yce;n8E~U|LLI7*_>uPAAEaZ_nw&w@v_D<0q -WeQTUWnaQM}PDzUqCuvUY8kc1NhwUsr4C($z`0?N(egi+x$X+8>ulfmU^gBlJQO(fC6Zn -4)jBd+P%H}|5h(bL+K~seh?LuJmJZmp@Ft1M?f6->GDs3X|GWE>pX~vrO@4C>&PrgCN9U{_lRTZ! -;f&IBHYx2&h>}2*=d}3YRO9s43mt+vOIPb?)V@V^X}5>k$KlUN{E5OJRpmvEqe`FLA$+XbxW^f$6^=T -CwZex^f30xb=_PzDH0>g|fh&FA$lY3cf0=!@Ho4J0Wjd}{+OL^T@U-sG>A0uLK6W~OKeoqCCw_Qt3g# -?B+GFtah;YU^K>H?sTXqI%-^Be!n7S`^VPfklEcNVFUUN;{RNz`VJV9xf -C{926P-)xH9A4R4<_8H&FmDn{vhroh-U}-&7xNLNXKg8r>JSYR^LW9Dpm_Pe4#}vG>zPyyxWmNcV=j+ -I_y(`hD;PP(K)PFf|wyG;!hq#9a9ZCL8kg^_r6UxD|+mmGzvaV_`US96tO6k?%;63j6-}1euSjT1>~VcZeaYhsehwRO!gh8q15RW%M_^;MFp_DNhbtNO|A5Q6l=e% -uZ;?+YkMeG76aN<8!&%psD{x7`d0dEzliYR8SELXe?8{WwcQUhUfBqOq0E%Le6Du04L0Bh-+S-U9w7O -k?}i_W^b6X2kE?*nY&9Sx}mU5vVT$YxFKWs!;vCBSg@*j@;w$2|Dle3hIDD-vi>FG95=r_N!qk0$>D~ -yl8dM02mG?A3~=oKEyFZSEFCz9JC8{0dt87)YTL%ar$6$!)mgs>N#Ba>ud&@T-#W?wnbI_SF{kCsick -EJsl-XRTTo3(`jpB#jhl|phx#Kq1Hcr@|l&P+R95lE62gFuc{i?$yINgm*|AA@zY*}f_>VX4X|0?B3* -_NT0h(cikt>sG)-Vb&Vw>@UY#KC&AnG*{mH?&IPl2bf}h|9R~t7r+G%E(EZ%3Ph(@n-ch$ -hqpy{`j#Da>VdQ>}f=o5o`nY}K4{E-X -^)4JVc(>`3&#ph=^ucnHw`2wB>MTomH7Gy5{t7wQDdE?UGobs;PeviQNdOOFN$Youp^a_+N?llzCTZ; -vhnhC>SM^?TOiHDiH$TlNOseYpv7e~wfE3b^2iE9iCfc6KZIP=Pf=21WFl>uZ9kDlEYyDV?xMb%1xK@ -hjxSqB$)z3(hDJR+}?4{QxKmu_8Ehb=Ad2XcgG8E*pKSSlSp$-_#*}F7n7v-ECGSvD63eD}~F476MOM -MYuhkNF3oG`fVtfgZL9EacqNH8y>S5HKWd$+l}#>Mww*GM;dRH{5~sIY%hxhAHJz%c*9?)CIGIuy5!= -<*IdLV)^d)jHu_dWZeTbaWz~xPY5Gz#$O=$3s_!7bhBWRQw7jLw&@wU2Ok4?p+sr_Rxk?P>EAELYs0D -DjTjOs-*T?CCPQuf`;FJmZTMS(|ca^cwyoRU~?>$5!6) --9$;Xrf{w4iMYSU^Za3E@y!>xW@mJt3x(1mK#!lbWo1y)&hx5xOBcQF&)*HOU9|TEanwCCUPgjY0*Z* -r0J~6}{Rc&~bEmA0L*`2%n?bqS)gP=N}Ku85u4%E5Z%dW~3z@k$0;#w&>cl+2wn3WH;@xkr#yW4#o^B -Lj|V_PK|c4xMLsf+t6M_dzLV%1#LS_bl3i~#tIQ*FjjQY4eFJcNVb!I*JYaA9{Gp -2!RX!6vw#fJ0lAE50U0XWy*vtu77*C^xC%)wpj3AORkkaUr<8FDcLDjc0wNw!{K4|k1$JSMJAg74P^1 -ZuK7cJSE}TW9RhnwU93Rr+DV2KUmzGb5B&}lQ17%U}Sxl^ac6*eMiN}Zh?Lz)`pq+?zSO`WD?b^_;74 -42dx@ZYu@y!oKdzvyLSzg0f3)9TCa#--H4_1@H+R9}f5+h6Yx3I(ETiSt0Gk-^=v(SNeO2{{BLLf2O||>2KFw&BAn#oqDO -L*S&_wYK;t90r-zw$3)#TxZ~3>Wylja?H9UK)l4lZ0aMH`haf6_l9jqM5E{KgP40qmh65Vbyv&o7xdJ9yUTqjbtkg>$9fc~R&a>e#s)+qzCJn8@4l?AE$*)Tu0Of&%gWPD39|> -#&Q5yQaUL(p^21BAobUgUo3tL^B5!xymeA;Y-u3N)URI1-Yx^14&oPG@JWRlp4Wb7H(a|@rHp(et8V3 -U=rh6o@{Z4lhp=&e=dyJJACC6ZM+zBn-K%7$RGr$PFGbU;R_F3)DSX1256x9x&tMat9-l(h%-LS>9`( -l@On(p)kw`lx6P+d$nN&?BCyxNTUm$mhpRA&RG;_zR2mf9T>KU-OSpiEtP(Q92uf=>c(}qFcpl(3qQfY)XO&S`g+s2Py=)@FymbX6-2 -8tB=273jXRbq+d9N+nm{Py+E96*2A4RWYxX~Qkg_KDcGrSnU^DsbGaTtQhoYYvmO8p{i`OBu^)q^uV6 -TRqA1OY7LNWq&?oL?)b7O8>8rc!lLfOCzY5_7nB?=SdwkeYhkP88y -25L*^pLo@V4SG`$ARm^^V}AH0Snw_WmXyz-WQixsR|ru%Ro2<5YBo_sJ!^K^a|YuR#b;^6x+#Eqky~D#Un -_qie9g!6E?>*61KdUs)KME)UpST_^F9rgatU&JkQpHW5KwEhHDl+qU*zZFG*7QnwWgTHVS0%jRJV4-D -(tW4mAp;kw)PkRM((q9n)YPi|T{9 -ho)*>qr}~Bkw{I)-87x{(4h#+o~l{FJfHvVq#w+613MPBJo)4FS^9z+)9^ts651jF3E$(8ndvEuEv|a -8^u@m;ribDa-9(l(bZ}$F%8q1UR)EdLqe@YEpUgWi1{<{q(C=8c)XNpuP_rV_pIXoO1Wpb%3Y(yQxsi -dYbh?}Z>Hr5<+zELy9UcWo(U$qmkesV2@9Bg0nlHjCA@@}aH_{v5Z6UJS#(Z^_4qW*q1^iVjY{!z8#} -Ble#nqW@gPG$b{G;XzRr+%@kNRRiO(<#jktv&2Ju0Lq>6VlWQkbKkSuWxL-NH^h9rr%GQ=uoGo(yhVs -mMQ7flSuThBxd!!g%lBEyj{Vyw-jG+rFcaAcbp%5ZvrpAW;a9*JGIQ5@@%c$VSFC9#d+coGNwvv4rs%agpJud0G6F;Yo-eVK}OU7WXmSfcQ>^qiSgJIfgGmd<(;~5Z}n~e8g)Qj% -sqnH4HC9+{*BB#B&*5gZL7LHz00e_$I_P4BvuyBEz>K9?S6Uh!18s9wHP&OX!f6Ds011VxGIM2N;q=r -VB6)^A*3LE!Fi(ut{Jxt;{AAY_4WDP0VHq*km%BSC~x}*rYR?$BX6IXxoC)@6onIg2&T!)Z=OvR1G#K -n9Xu#6ALy+m`ysfi3gh|W|P8flECIoW;2o5Xu#$rW;2Z07{F#5vk74~`C#)Jv-zcnH8)`M9J6U>Hf3P -*IJ5b<$khOW<^9ZZPm!x70?Qg^`FxS9Ndn6|nB}G-H{ZGbAmDxoj?tBLWUi}Gh;M9kkId)Y;W;utV1` -*j9Z9KyIRea&`3&J44*2`kgoztA%JYN$F#Lb_x9-1+J8BZ%#U+*KI>oHI@PMg%9E%LnZXmhXaA7x2!y -R8WxTo85>%#-`D!!Cdc@;l*saCxQo%mcK97JW0;~nGc^7P!~4b%P(i7tC3>@imFwKqFehG-bDwn}>w8 -gUV&0cPG~U=lu(PspQjLe6m&;6V)afyudmTKjq6UeZrjg+^c&!t)yH6jOZy9b?&w8rV6$kh`TOO=`c~P4}gX`rUM#@XryC2F2Lag=Y3qqk=K(RmcvK -OOMc(tijuPujjmS+4l}4)45|Y#U}3TMSw5I(2P-6ZAz)sIvEqV^K#48J>DCQUW^?5-`=Uod9rY-v{WN -WrqS9SpE5oq7pxAhE?REwmJh=YH%c=&0U7`C%&34!_{aC`gq79_qc`Hh*zj*hpfM*yge{Nfq0<-ai2V -^DnYzq_p!VjZWDdIq)yjJ2Unc_Pc{IP98&p$}dR>T=9e`#C!7TP=vHsZQ_-J0urp0^aU>7L?NDdJKBU -GopP7eAjmKNJnJ!#|iqM*}+G#(_6_<_R^^L9TG_&eZ15fvL+6#^lCQm8kHD+&SH-`g6~VESr)E -AK?Ckk1xCSH4MaatYL<@Q-=Crf6VY+QTsZ&HuZj1RNdo@;E^+uv@<>$b(`*F175**$x8X8p!zEO@~@| -dk%zB|(HU|+(et788Kx$~F-$CitPW2SXE(Se)F)RmP13vZnmyqclkg8iz3)|zq2}mIzS_#KhoKTN+>4 -QFULh^f*4b4_c%xO8#;a3v;ha?;m-lJM@t`A2I-(vVcHtl?u>W -TM#bqi`2^JxR`{-`D5W$df+AEaw(^=UB~2BC&cb9`|~H|54~Oo>qlR;f1bl_Kt$#E7j$jS?sA0h+2#R -TXxvQp6f?0uC%O3ZLQe5q&kdn7!~5cBiVk3goUXD50_ZVxVecBOaE72(F$)hjaIdBAJqW>QfI=u(ExaOpNdCoG -~$V2c$gTxJC;{Ad*O%D;aoF=a4-*pYQ!BbFAyb0OtXTkJNvcCO{JsTzxh>?>bwp05m!Tgq-My>m)aVg -$^>HhrM6k;pGqf!88N2zi&U~+nt<1@x1aUmwoWF=&PSN8Ju3Jh$(Qk>gp+3B43#+CPDf>M2u!o;7lz; -oE}fzNRW&NiIiMBx@#Kur*7?L3rG3)^9YHk%*_latXT|OsKSV1oN&^V79Wo4urom8Ot)}{LnsjlEcn| -<)mLWNh(30Mb2fK853mq*GEFOcQ{(kf?Ma+q3#WDeA&)%q-k`Xe*br^E+Chg=V)m_23OF`K8I^`x!oN -B{t@ZbIe@B_<0RoyCQ3P-}!V_(^tgy#fs8=)5#utSwsJ5UP+W~glPV&4IGd*#*oQDeo$%b3Hxedn-Z^ -Y7=-GJ)mLe7Wx&UZc1&hj^oCdzgaf%SuaFCrfN+dirbTL3=ep=$SG!(J7NNIp$oHw+JjK%rOs8oa}F8mH-?~Zz0eE{%6gbFi0aK -_d=Y|`y0|7O!aSZfSV7)+geUp(`khXt6wx*oB^r$pTHzAQFth{O^0-=3Rb$l9&Sq5p;_J -9x^sRI01GuUV^+mMryokHf@1^9N2e4EO&<3k%67+PfU5a>b94ncZFR_DY2`8Ofp6r;`ci|LY8-w>hRJnr=43ZqswKS -zo#sOdh%zG>pC&S=);)hLaCnp^M@4Lw)OFNVH#}i@`{BF%0P9x)@%naj~?y#zPmw<94o#;YDUE*1A0o -E$K@a1Kryw=Z|T@pt%Ya=emN^(`2ZW1oy`$PY=Kxt;Llw;evrGBpKD;og<^*^P?+AsA^YS+lx9gmk!E{a}YU?~yPV~bsDbjw_IM -E+9PJ}U+7iWq;GIbM2|A4xQUiD4H=AYcTK__M6p4#8E`1YL13eX4Z!V__CoV#Jl&cd&m#zsN* -ENv=Z<;tR6RPLgYFXFB7N%?BWVLIkX)Tmb?rowc_ZmEOO`P<{bB>3k&Xh=vsXk=aWc*I@Bw?HQPZr?X -dW6qv@RQtRk43F|BTYGt;IUycZvSo3!u$-k=a^SCE-`yo9YEzm&Ono}2fP$b#LA*NQQOA`9(ysuANO1=ENzh-$u%A!;AEu%GIa@P+W_f{em2=)Q#JT}+R~Zj1?hgeoyF0SY?nlp{c -`3qO@na_wJ|D=4``+Q;ZsJLYv6>eK1nUIsHv-pU5aN-m;+UItY;pwiaO=>iij#F834cwXAro&;$GyE6 -JIt{iG+<@>03qESmVXw(k%qRu+hJ<*;atDRuVhnKHAR@qg+^xjc*6ZAw>R|&gh1XWGsP*oHD*g$nT_~ -p!9Wvc4EKxpR}OjD}VLO(_0!h{zz%N>Te_Y6lbI@N6YJ>={v$^d7e)S^oY!*dn8!Pob3??Dq_Xh-cPv -a*SDOxXms15DwBzUtiS44?Th6-bRo1rsn+n(v$^>z6pUTB8qgI+KmUOk|~48tBq0@$e61FPGYjt -t9G)?!FDTEC-FX4C*h$(LUj0}d<2&Un3?!E$I}8rD&g?U^heC;at?x`0uV~={LDFXsQ!pKUDWKUI^vu -XsRmN2Jyb^oX)DiM&?f()Rn7Ya8q{2-I|AMM&>bJphL -5O^5l4?Sp@a-fn$?*eEToB)Dx;LzqH(uA|FY2PT@&IQWX;FkXSd`tvO=A7D?6|0i}srBb-VMivFPL8H -A9lZ^f-}z&zF4mzctcol-=`NF3~%%XPwGrZd76Q)UvH@JlG5I+(8F_fYV$6O~5%%9Tb8O=z3~w5dvnZ -whb#9!}|`m($G^G~u$!h&|N(#}OXx<;mlywup1GwusM9!dSqL2W5^Se^75uhN}7|x_=nTZz7hKveX|% -ZX6U-=YEAf6b*4MKtc%M9qH$k?7g=Ql)uM3xMYZ@fQyHgHE-hA52Ud>7syn8@tFyI*hX&amoN*8ypja^li|2#L~ -E73~c2IsHfp*>|2;32;k#$&RDI`se=)Q#yP^3T9#1x0FaHVjQ~H@on^x@!S-l3WtQ8Eoi0XmRS|X-9; -*9d5ez-=@Rq1nak;M~#k`B(rfcfjKA1L%VMq1EX^Z9wU?r!+UnNBx^%V2&dT9GKrQp -AF1$e=ZEnaRlNyFb8=t+NZ;BzI{CW7NV}uORF3Wv_W>(mQO8ogh1Dc#^Cl;u -)wJsz;pMk05*__;Qbw)Xv4@q1JyI}_CBs>Lf3u^^+5GZ=-TJYWTLgK;gpq2pk^cBn$Lw+r^=;O0;tGZB_dfswO -tFpBcbY*Uy8GjfZuWU0q`5Is_vxs6TPo&KgSTC_R|bew4X!-nC{; -3B3=HB0d?y8ekCrTOp|PfJ?!j>Mqi&f3^(ktwpsS(i36;F(DA`rjuCwu#7cViU0mW)phvD -X1hMjN18t&gax>v`@6OGNdyPN}SH(>VK!C9dG2PZqxpJ=S%mvoJ=@C&MsN@*S4x+R7P^(@BHyvuekAR -dqG3k?)Gjm*SjbTtgwf)i0oqovgVQ`EFXu6nn>_VJ7+mq2e07=KLEm*qyAv0h#k+&4{!(b=>Fr+a=y6I$jC$P++z}Kb%~O^^_pU;{#0rDpuX;_1*A&$HJNv -2LHlvOtapzz0GL*`TlKm|_hrT}1(9Z+tmsAAm!=e!8R)j#2oL$0tu$JDh!T%ED1aK0V!mtIK8V1~I($47J5XrQ#CW(l-5+oN&iTdJ(ef -HOPp!F3HkGXOkU#Hc3~_ozJUYQ$d%~CgLU&949|W*qIldP>Z|mw3QbX>*BS^=hwucIswQ)^-0zJji@; -$fDO#$4NIpYSwI+G^f%;GP+e8Y9k$WBq~!{`TA&t~2vyZ}^qhz59<+As1I3t?vi9dIQi{`cFFMoQ$G% -9gG{Ep5Z-(ZZ<=SX61cQ$hksiqol?I?n>!;|Q76;nPp^55tkFEIa4Qv4r4~#?!)29WH{Cc7%+&4v4b3ThJqLC6d#>r=ucRN^$f8B{+iC3={5AC3zi>>0)Qk -e&|J9Awl8}B*f)B`otQ+x`e9$9ZtipTDwM1u<0e(}AAT9D~$gIPixYC63YpJ$#r^O>hdnWwMOD13{X- -tRI9e&)Ed2H~>t=&z`>+-lA_hhFt>BkRqmf941lK*?*+Ssj4Lqp+2mH6<}$RrMJ1f7X=Xe7510aR>EZ -IGQ1r3XCQ0vSGL6ec_Zo;T#MKYw&oSX-W}MRc)XJZ~;X8CnEO09tY@>JnLukIWf0pP8tvj4HNXdoeu9VP&E~eVtpNP43yfZ_OpAB -&K3bSgj3#aPmcqKQh?y_ITlXs2oSWL-{s!ANMGPw~BiEZ~u>>6GJj$?8Tn`&cPfPF1WAPWqyUXI#Sp@ -p|QkaCp+4MClybR7Bb*H}%Q?`RUt{q82oh{m=|7Yrq -QjZ0br>p%(YMJBAy(`CM#jt_q`VUhkR)Mm`{TcQ9s{KM>BY}s8g&r8-OIeS|K(SC?CRQIpux=rbiXHz -6EDJ0QN)|FF=?^bc4@2Dcn@(515t{g?U`p=HnNE;rS_#gM{li~r;-Oez$YBl?xHtjKr#=s+t__7edsU -YJ378ekWhES9EEURtp1TT6F9<11v@XihGS&H -x*vmcD+ya9PBo+*RcVINI$3F8=+I&YfL_65I)qa%*|lk{5c!{1RFN7l&p((o) -lN(N0KC8N02zLL_Ys*s5~s8F6=|%pJ`H~`eWM&^|2-4m=tj{n4IaWm(sfkKLk&8Kd?+Py=!1`{zm?#L -kz>)-3-x^;le2cPV6PgF-rWZSdK{irI>FvSqVOI$SX-~Itj#$l=wIo@e5QZP&)!DCnj9<9a6s@+X}kG -ZRAVbbA{mdX4UBwNwZb2Whj(hjoONTV9Tacy?sP`^yiA7=`D0i0D`R@Pn+Rk?pTCg#I8o@LmY-#ol~W~a3Z>K6lo;H7ZZU6s^o -V$W4AO&e7DHeEoA~XsVaA8py2yLDey-jkJ;&!vg`t|kc=EGFqj|&i^Hi#fRMS&MuJ#zJn9_9zJ~Jez^ -K%cZ$b#`Pv>=?sCMY+hnghUcbYsrK#MzU*oZHSUc2}`8}2#Sa1Td%UcZ{tGx1uNGeq-*G%x5H+$k)IYXy;W7?)(6!NebO$8P#x$i=VA}h9}jAT(CtrDm&pS#@_M#VZg9lE -0*d$43O#SCb@gLe^jjuSipLA`t`>_RM1AH6NRanJoHDP4@tvi|_%0pHHGXs~+U!l+4v0ck6+ykZMtSY -NxT0){C0gNnn8N9)?v*6=5!+O+Pm(vIFm-u$9MKepGNnY7YhxkD2~-`=N&929-p)s)h*kY?26eltI-1 -&Fk@v~>L*8Fl!S^&m24*c_nO9Y9z1%%z8`=59Bz_~@IQg`(>-$@z{LvEzZ2Xq8P4l397SdbF_H1xOiM -kb@x0JD24SwF}+=KU&jb7opr>reO>fFQcDccjOs-mKRQ`j%*J!K2~x~D8|j^uGq*$`%O-BT8Dp7uiYp -0Z~;B*}G88Fu2H_mp|Dd&(+BuWlEWP5%P2gBO*(tA-@BPwV}nvW0i_eo@&QORu=7O!J-G2%V+eP6^G! -tNaGP<;CZO26{~y-B@D~-)_U`@v^eFafwK^W1p+x_V`xY<8C~5d8XYYhQ9!vJ)>)Y>0*cY_gStkd%U5 -3k%B7tY7x_xQYQHtgl1>t6?(bW&11^(uDml3A~P^SrRpG=P6x>L=~(Rh4&ZEAqhot3z6hbG9Dlxr)+U -3jr!S=a7S_Pv7v_tX?S4|kcR_?*`X`NA109w$Y90OE4!@do-dUn90`bIbynqv!uypT1mrLbaSYe?Wv! -UaOTYLD4XQ8vi%{+9xXK4B*M0~PmAh%PUhZB9=?m_5&Yj5%G!cRHkyjy#T5RwuvhV_;a-1{mV6_nzi{ -@$x=$anLw-JJ;0Ude>DS;Ve{qQmdGSLTFvmyHE|=7bItw)7TMJT%QSlxJcS4pb!h64I#JxLbg~-Xp!f -*Ywc%;(K+%rcaFM&B>5sR^$0*A>u3gl`{NYF+$%XOWkrsNxhrj3n>()Cl~ZmaFuK0i9G>&8f)d+M1sm -40=aSrQ@wBIDwlExqu|rXlsg2984G$TX2GlxQ#@iI|C*-(w-dtiD2LlQe4E2o4!`2Ci^I^7490S(<8T -p&g&c0+@KFw5;joFrk2w5>!=E_}jA3vvhhsUM#^GEJmvMMIhgnxMn9AXG9L8`M%Hic$2EXO7jYHii2A -6P{&!L^eat=3fxShj29DdH>B@RPJ^Y}TO!Qlc9mvOj~!#WOM;&2~_$2k0k!=E`+j$v>hhiVSvIh@Vm0 -uFEC(9Yoo4jVY!%;8fUzQo}^4nN}Xa}K}d@MjK#IUh%JsNrxChlL#8$>9SWKF47bhsQaz@$!E8 -*K0rarCFYh=V$+uu*i&sKIhIlwbr+FRl0!;}jacAsJjo@61l7H3bP5_@8;3dY2i{3KNh7n#IFiKUlFb -vzb;N{<23+gV -1K8rM`OevjEZ=PZHn{ZN`BgEu_j0K8uJIpc>v1_;o;TF`Z@_qX{tNQZBX2qcn@C-+ahXDPY;~TS}=c1 -OBZBn_Qq_1J`1*98AUF19QKUQgG8k_X6sY4dDvN?c84p*arePWcuhY7k+c7yM@Q8hS(gGKDw(R-Vz!s -x>@0`&=rDF91ku=noCBt2y9EK8GWt*x-9LOFHD_{#^eB7^iu*kS5tb0)IZBvHjP>4(>R(pkLXIm+82wOn&*|WSSxyrj3LfENn6jN?3hrF -sG1}lWk;pNMy7XoCtl}n1b6nSveQ}#ULoA^8Ob!E9*JoughbkcbF*jz4*(Q{5m6j`H -$&-F}dBRj7#T=T#To2jltDopeZ}RWZE8?wS0j-HPNBR}0^%BB3z?vV?bzPp|xuS=+ -$#M1UQpY~1I5?EY3XHwF4eU0}hzdxQ{^uc5l%e07^?Y*Wq-lP2fIQ?w+T@F5E%B8)Os7&A8-sjOWVzr -I+SXM9>Pzq|=yIHA?jhf}uurRq-@W_k9@=I(yrsQTdHwSQY<1frChqQ8YC^v_4a~L;o|CyQlb8{9q50 -E3~=16XAxX8>=+^pf|tGM|{7c)o8^tgGj96vV?<>rRKNx>QStIr3 -)MpVNZ!ii-Zv~ZrRq#85hU*!N-c`$oQ~Ighs7-b8%K(2ByNqx%_V|9W${%waaET|0C3X0=1v -OZD0;cNAmM1ipITZI9kdU_XD)63R=eoTfLrYvO-^_J^U@EEkZFJVR^cwfp1ScGi?LiJ{Hn+dYZFozF53+d@_BRo=is$4{~LgT2wT()Fn{4@ -QhWP;ZY;98Xm@f^Mu^o3CxfR;-cjuqqF2#ik4Yi0&zph%5LU)Irb7XF0@$fvV+#K+))aNo$Dy1^yWH> -X~Ao)wgM~zbG9wVRnC@WHb=J2#jv^A#ncG#Vq8ccX(LUfhqNI4YNUm9Fg%7K3$yJGTW&U`Guu|cG2lz -S!&d4t0lzG{#TE~XGzT-8ONy!GLI;Q0cBF&hF%0@cKj=<8ABC5o2}IDSh=cf_3d}8S-WmM`M>_CulS2~{axvH<(~n1* -x&lA>gBJyj|z`J#UK3h48OtS&+8BV$zj$1)9n9W@LXA@`TyM#pm6{F(WjdPDBOQufAG&O0spUmzn6df -3x_{Xlc(Ds&N9z;ck>_3uRGs8ZRiC#uBW_WLuFO{d&T-B_GgVk@E!i#Z)|H)WUFwK?{Jm6p^JtHbKD -XLgJnF$3l7<(63`Wh`VmjGXl_i^CP*@i4nE!5==cbk>2ZAXhVUv7@jMP(S=v;5^;ikNA0o`snp~Uwp2 -Ohxr$JR41h_@mG?`@$rjO{h_;3)9fm<^0VN939+>d9wqo7{zR!%;u3+9_{%);D=)vZyrLW;@V>MB&JE -?^zP=Dcb#*lqbbNeyYIzO#jvbSfy0{#|LnNuGsBKIgTV7KRp$J5ipP#Q(`j=U)tE|>5=;Ifc8{oz7GB -IFw08d1f)dj` -NWj{ULOD%T0hchY4UV+9^doFONpcq(FK(jA>`=@Fc6Clku0>svr?$%99`PAZ*IM1{2o -q@hHh{}$Dkkt%MfuT}S)qYol$nG8-J&8$g&L#;QviG&Galq>3}U7EmSx#sSy)hPnW`RDSlTC@Rrv*WO -Q|(G$D+>8vqO%?OmLLiCM+u`o?t0nIYtiLLq6T|i?M;n0$WLOk)_x!`*+7e7e -P+m||x|8WG(bItU-K92@ep|4S86<rsw0zuZqR{FC#?>b)neEBKv -jp`zgFN9`!Zq@}0L9+%X6`HUGkMxD)bfl|~XhAaIK9|L(TEG$`tgfVn)$+|LJ-#qC<{%p$~Sm34VSj$ -Ri(F*d^mLjWt_0_*0-!c$2E=<{w(41^27Fn1nQ;IA_C44Dm1h0J~T=h`hiya{g7U~x6uSoyOYxEu3=R -V)i>icPJ2@a@JrePSdqkhO3QTPyEBf{<(B?KsW+PX+u253@q$LsW@=L^UbM8a6nP49-=N!HNCIU|VFzz_#!s -{To9mEsWF!EB3Yj2ahm?XZ0+5^e8C2IYo8x`K%+A<*X6=5xd -gI`M;c=xZ^2pbrUzFyLZU1Sv=mZ2JyJtd@&v+3D1SN(gGkpz!!vC!zeHODKBKYOD7Yu9N -t}Wo>joBG=DeGXu2Rhc1TaIKcvf#L`{lllfyqSg^(Ta8hCi>ANUdle2D_S0KeJ>90_Zb%j?ytgnSBb4 -)>n{_^;4;R?6{52LfM1Ni>vGG?Y{HeYA{H6_J4?63QVG${`YX8o7X#U0Apu#6cu -P5l;MIoGJU8tOwq}`@tCo6~SI4SP@1P>R_VI4I}C_n7;#0)V9GLSG7eR8Q2(ZRSXVhGMxr84g4Owz%A -EdSeB4i735X5z@6?q7^2>3CZzQ%k9y|sMf?>HQ+L-Jp(GJd?h#P#5m4?ClfoLAT)_)65)O9A^RO%Zx# -ad?kjYQrJc=C1GP=llXf)-NPP#nT=}>zE%2o -aE3EuLr7>pVh1ad#0M5m>K?9U_7in)Z9B&*k`BsywB02yk3l?=^&g$zx -6nG98IC!xIFTYY7I_`CRl@qC#@$dqhC9{0?P%_}}?n39AhhLTX5yB$J0NxxHHkfs6pn;$K1nqiDT@v{ -5N1yHBpeM^qBzYpmTW!nF~pj5@+ATrn<2{JnfWOg7Ktmsb|7v(k=;sY6rN^`Y0jeW^-Qwg;BR<|s(@_ -}|Im$%moB04=Pf;{*0BK;->#w!B7FyHZt-qL+t5E+$toA)S1o_B9$p^gUG_>HoS+ -E2*!``u;RSAXIJKNFhV{fQz{34J5HP+uZJh9W0Hzd-v-c%dvKp{&4;wj@kmazS2h>Q!D^ynZ{4MIR9I -*asf+H(W{jr>UTB_apsnbQF^1;vv=pK2C~;@jx`l9`xs_td2wdg!DzFx!RUSnr0{`=-WH`w}l>I)B;n -x$)BxHS?!>*>gz>(6BR@epdbNQk0YQz0-6yo^|OX#1*ay&${I;!Gal+0vkOqT?37qGSCm5=mV;tgFd1eKBg1k-p*;GLVeav0YE(iH8D -Y05M&#xyMkMAaMksDkFc&_)3xv2Phmg_sOJsEJFJyG$&t$aXN8+Zy{5AOVb`1KQnqlr`?;6}o+(F2a5 -pKJm!S2ZS(Dr-SC5=*$?cbALpa1N`XYfD!@EQESx(|Pt!gITOy<(2Vp2DWsCYWI_CUHIsEx53lnr+Wt -OxA(PbM<*{$uhPgP4X49Z5E5EV3`efGi)O%xHf2!m$G5*hanAgG0Y=8jHi7UXBXIKm)I5-6fZBdFlzc -TGLUFXOAD44a|&rC?x;wvVm2kf*JOeESD0?hNJ>er#kF_;Dlf1#m*hGMEwc*>EjMHrq2H&y=i{1Wuem -7jHaiOK1+!M$Eomi-3vw+f`PnvdE7Dnr3o8yQ23t(-LYp*8k(IH9+|TptqNF8c1q+qSG7S>z#n=1|+v>%*Q!Aygq}0NRU$0+qgWe>sd>8ZdvyTM&vc+VwSE|Lb!p-6F-n^WQ#a4&iNYgjfE48H5qZBeR -eJd_`>$tX+(ODxpXW5D@Bi~S1LX+5PE_;!wzM&u+d>F#Y5A#igQ?UA?m--_NyRg1y6tTmGS{>E~m1;y3 -}qd!jE3sQrh+aM(*M51zId7pE9n491(JY;tMRSaJDenVQm!f@+A*5WCHvzg_3lX<|Cn5j=Vf5B-{fE; -%3dZSzLUyAn!c*u_)VCfJ?)RNj7uIotMYBJm1Ux1r<|5NCJ}&9_wQAKm25~^I|>C-F9_^lHYxBEkN5I6~fEsy>!?T#~ox(>ghjxa*rYhqwm?&nBSCKs;BEx^8jb2kIH -XeIH5=12_!iZb2Sy19@14Do!_)+u=Eh{`^Jz;oSaOZq{;h3OA?nFblZ-LJk*m`Zw~p3%Pv}!JUVM?lH -$R5;mSzKkmPrr@4Z|4cxtwyH{~K_ri-iN*?9r$9Q_S^7x+S_B(j`4srL7xqGBP`BKD0A<7^Wy((y8o5W|8*JY@%;Z>B -swPZ`GA)X?cTP+03vI{>J7uW`%|YFeC$aEEB@L>j|^(ol~cG4{O^9^#1ra1=vQ23P{ZM?f6%{{!`Hz7 -;m1!ueib?MxcKKc3@o8;4s -t+`{2z4mWYwz+nxCu1+z4c>|YtEI*uT*95pNEj&J-y`BE&|#sz~cd~hBtngf~)~J9wyd>XbdjL%0W6@V$Dp|1r$YXCkznuXm8@C$g$F)o0k#xVO>fCo5!5a5@wz -y}D6IxN>*&G`dx(lv~q)2^ZV!KidR4KOW$sNi5zhfPYH>o<}Iia{#|iV0;h(MkO*og8>Tg#)6-_06v? -@Xg&vU#kI^&A;1xnS^35QTm$bk@Q?87B#3vAf{X?D5xiMo-MzGHZSUsPmVJEyzcmu$588AKtyaQmuQbs2cU>( -Qr0yz2>7B&{(E4MKIAl#Y->4{QMcuO{nn*h%O*a5E&aD;eH#sqjio|nl3o&#Jc71FxVkppo1{a&gxM;Sggd07B7I$&-B?Q=Q6L-6K<|04jG -u7dmnZUuN#8T1{1X8?R*HA~NSfFG1YyN_uD_+16?8t@K)t2Qt{Ke}RQXI -O#=b|G_>H;Jfg~qYMDN=4Hq;;9~(k3vWK)&jI|)4oDN=n*l0!vOWc2|6R;J9H3(tZ^r;?UI*F0z8K)3 -Hz2=QrU1{ty9sa+;2Uo;8F&-mthX4h1Gse`w7uB>0<7E*?G?5m0B>#rz5|{C@bMJ>P|t~wA22=UCxHjo*FIouxMI*#L+l1Dj?a2NNF@B?m-@ -I1#6;<*CE5su?HLIcO~JU~9j5w7Dn?%{u&;|Sm4IKozrBm9Bm2&2DdIPN{4!EuC(IgSwb_+pw7?&9|U -dD!3(R36mE0Y1JFh(-Dc%qp722Zgb-a@bBX`$DUG9onW%L0nAEow|Uc-3U -ZT2{IQrXg6m$GYEMlES3cv%k0*2H%pr>xFD3;A1;pWSumUYFhyQZ2Y11b1^wUq1UAuOX#>PhS@y8#N% -a<<`{)w!ddIm!Wnu5ta2g`$rs(IC_!_;_K`tDMB@TFC&)~rDf-$@5A9hMGW0Ow{fE5Yy`gaz}eRcICu -A3S)Fn(6df2|wRSm-49lsx=+Tvk+Kn;^tL{m0cJ&_AZ$5Y!vXF=TQMNu`G2lD4me7UfU#lmDnGPNXA9o?Ul$W0Q%fuw+fttJCKfHCsa|fwke8K*#cA -C@$!RfbeNdKWH(hhcAI$YjJ{U4U3U(?^|xpwqP#DHT+@D< -kGOl^=;neGeHD~dA#26D#{6KsQP#n+@EK-nSH$L;g!mX0$y);{oz%?J4`}359PA{dvczbm`KcKGg-KB -A-Vnb+q=nlLqh|3_St92E3dplW&ER$J|bt%oN>vy6}K5|hkCnxhd-9Y9o9Rf*Ye&W>)-kAoxBw#WIYs -7-aC@C{vBr~NWzNs-~N>O8X)T*J#(`p{gSu#z?t;)^mp=1iGKr_xnk~Z^CaojOle*1n0IbY&m`-mC*F -BA9qH}58|kl?q;nT8Y$sHT?|}Yz*f1&7f0#tP`AdG|mHqhhSH<{*Jkn4;Df?y;JflX9B5`qXBtAZ#Or -JiTXti2mG#V-I=Fgu`7A;ytZoKhEvUKTEvMzTvS-yNZx%JjtNl{S|vDs|2{aCwpEqTl_ll<-W>Ez{g$ -)s|*p4_z}g={R*l80>?@?vu#1?@DCaA&F$X -FOihv61o0_MCPB8NGZJOUrOZWuO*VxCXr=lB~npQL27Gj$-Vd9OCEgiLGsW;50TBAHEe3F* -$_U+rrtFOLF+p;%dfcwGT?PSAwiM-k&k^TGklLH41koVtzpBz1UlpH^PoSZmuf_(PbXXLvxt>lC6CGy -1=Ur@PdZ*M0*e)kRe$3=;Bbaapl7cNj)!N+R?^cOJC$1|`>=(GbtZi1k)_v^s#0R72yX(G82IidG?MEYz# -kxt)6q;KveQpbx#>huU74&g^Yco@Lo`p|YWga=w3c$4{R2ww-`AA#_Hhw!^0{4ofB#v?rRYpGC`tuR< -=2!QZ`5FXl}4j2qc>i$HU0pT+s{Ou6_K?wgEgg@>Pp7bNBWDwNZ;V@>3gKB#nISi8DL5xHSyNO5>twf -rCCy{R3Or(2XBGQX3L^^uPBfJ{IUkBk$5PlhiFN5%RL-?%_{$&XNHiU16@DS&{pF;RE5Wd3|ek@S99w --z5g+Bp>CxOB~pzsAy=m?X@_Y)=Z<9vy9-X@U?_e$jAixTZmshg<5 -B+LaGSIwS1V88(I;BwEr&M#9lK@Sn!0PaI)&!$Nj>~c?0`Kpro#p~zl%)sT@!)PXk%;w@TbE;N5$y?< -cKQq&uX*TM4DqWg;jVXzvGSWG9fg;PoW4iKngL5=D9OLklgOt$=Ajw1b8d^8^J -MU7RYj9VoHkJJu`Ebc3yHnf8Xl>P02*+Mx8tU%oOx8FFC}^UxV%`GeKPE61O|@2LwzL6xMHEW{P=cGR -PVsp7hyc2o0NJ)&d`m9?nQ#N1efbazaY7vAZ*qAE2)TK{V>Go!Z0MgFdjvsGZuwnau6}n0q)&QseqAR -r$;|7)Uh!YsPcydwF{u%axOvJ;rmtJx-Cwa|;$MpyQEsmN~Am$J1pqNKJvBY_g@0S2pO$_P?wnlUra+ -0qvaOV~H&Kl8qV406y@*1LWa{A104H@(6kCvB${Pty}3>;pLZKrsKiaUw@sB6+Ucynmi2SjqNa2*tc& -VIdtd{9UGkf@)Y^<%m1&vGl7n(N)zylNCS4O*fTvwT00@^q#;pu2qZvMKva}AqG)S?0Kp`LgjFG|Eg~ -48r4>XZ3lLON1&AzyY@#AA1fqZ|h@cTfP!^HxpZ~j6H#~|Y1e%#MbB>pDdG)H^yZ7JMSC!q)o;`b<{r -mSjKm726^Yc#!O>A)V=uzkN>C?{CU&L9Y=UYy?>}*7Jb#%?t(Wzce>FJmv?dHkq=;k@s$r9&gS?$~-8 -=V1C=uDL_oJF!%{r4f){pt|A1UshrHs!B+e^b3bM(^*W_jlL(2k8Cd_5N9U|6;xWJ-z>|pZhIN`7KWQ -{~D**-|Medty*=CH<%yvuNGV@A|isn16nP#TD9P6)$4~;z2=&~sGQvpR<~BII<+FMt9os{tM$BE^%`~ -R=(X2X4T%Vk2zPD_t6w+brmG^Z(+jS-`s%;jSiM%=h~HmT_b>i8LPP7;kEmSv4|VmPdVjp;>T1E)*N^ -yp<;qu8saWw3^{&0P+LgbnR{y5oSH6ljBC3Y`sZvm-un5LYRVuRne+6IhAA0ut8dp`R60Z7Hz1E*A+^ -C=GEjQ`8?jL^Rf7YoH9uX04TG-WV)ejG^86I9e+~9Yk4xY`dUfrBJBd>LwD2jL${Ij(3X^`t0uICE^A --?($zh&^Kxg@kqJ6n!nzsm4C*R1H@^?zerO7R$Y`Rc%*LPA0yDwCgntX-u_6`nNUX$JnR1AhuaErt2| -fu;E68WsHO{-K=LLs#4*;vYJD_U!xB@2os{@Zh&U{`li}>U(w{IB;OE9(Nu&a^#@qrF*{k;)_M^zWeU -PprD}esHmt2KTg5^b@(aY8%i8{EBn_vofN^!1;&Z{I#y^JrPKW{rIJ-FK -q?RJLu~MuAFp?%XMwe+!>04x)bc%c7#9?-kdtw{G3~(zf -`S6svSo|8j(apl;Qc*%&+J8u7L8c8Y}p76*=`*>cI@4$sj0D0I -vC-B<|4tB -y5JAvQgVPD%Bh>M*HbMaK37f<*5%u9j)t+(Eqq({S0@FO3eefF8*!Pj4ZZO(xaolqS%$H -KxwGtmwB!4bZpJKXdB`|lh0;i2lkz=tzu?9*z#q30*{{Zo9dI(*B;z<=!6F;O>U5^4ONNW|MB;p;>i -t`)g=lgKN3MUEajR;api;{_pi!M}R->Zvtr)@-6Ne5Tq<_+fAWUT6+o!3SOdJFnsWS6_W)Caq9^ -D9@^b%IBDe1lY5$4R9~_hiP2U%}MR7>kE^_RYNbfBok*kYa6Xk;cmtTHCmM1l8)F@14VH`LpAI`%CSd -b%NEC~6D#M?6<#1$$cDuO9Y@ZPd-TE{(F;b`uSM=s96G7|-?rU{L+b*O79aX>h*Vwp>w1y8Yc3vswZ&6w -)vDD*`H+d8VOvVV16eOkzMwy!;}^IOJL4X4dsIik;Y*Q@io>mn!@>hnp7E2;7fOTw*I$1X^)u&8aIn+ -jQ4&u|!sF!0ld@&SESWpJle{n>TAu44V}N+@Tao*|5xHlV$erNusb8N}W}>VBr$ -}dosl%r(eFg`=J~zZBsZFh=HmSPe5VG87TXl>s3I4FKuwwaFACI3uk00nS{=o7Aec;^U0Q_qAWbO04< -kewOGH*z0$x$3;DGrQywVw|u4&4-vzwP$nVD%XstUhCtZhlwk{+38Ra8UY(u2B8oXqNe?B>2_d8~ImV -%OoB`=PVu;7jVEIlqOrikDYvdbX)N}(Mn!X9Ofzx*#mv>rymq~bid+IBy#^=pFU%gI(>dtpTWWEGdNg -%#wOKWX_omV5PtLw|DthL@dO7u1Mw(HM=UOkWg}b2;^8gD^LQ(Hd8pzrxV0QTv`@C5@Zs=?;?Vs&#lf -Y|o&5UTw$Nvjnt$l?&(ZfAILucZR?X-yM-F~pE@Sh(;+d>GX>Rq|A7j*6T{8UG3oHLETehqXFYH8Lfg9L>-Qt4n$ -KI3IU^nqq_%_DsG0kO#;_!M#OIbL)l`K#k0`>V{YLliWx0OkcwUeh3+soKK?PX-ISeZEJLHTNXfx*EY -V>I-~7^T6l_TI?9!jlQyz{L@mfE#@PZp#yFygNR||AHg(|JGA&oYbfdr%vhmDWb4r$)=Plo*+u)K(@c4o@l$V-$xGiUT7s#`t?-$@T~MRThl=t6Yx*mn -WZmQWh*&VDKpq9u^nwE!y&$tRL4*)+!Dw6$hiwiUT&u>T{Obq!-ho<@wYYc~)`AR2-gG9R8tlG&&*H0 -E$g&>xwa6+h4N%2YxI6ty;CJ4Gh5d)KgCt;}D2PS-6}&eM-Jw_kz4VslBWj+gw)q^?9*NpXc~-nCZvi -pDIV0kF}Kvi9Q?{CdN?z+*#+W?~MO_-#LHW0MFI`PQB-zdpf10q-+9K^bsA%$jC4_xaI3YJmH8=1?E^ -KjqMve#W4o@{ET0p14fzg5AehL4jnqc2YLDBmrd+NEHh!k1Q|Sda4`;L;ZYV(7}%sG8RgJt#i0Z{Q`) -!zI!&1}Wu4kisZ*zpJo3mRGI#D=!7rz$r^~>B17*^rNk*@~{PIgvW;0S5ZKdE -Ov7bwcxv+`+~AY-5ZWetn*zx^N-*SFKugt;RZkv+y@)(4cAW-n~?Wo~HPEL;8efQncwQE;-= -9y>Yx#ymfY15_|c*yU87aZ^f_z8;#ISe>pPmx1>A$x3?7%N-zfd80mnlt!pj4{|3W1!F2q!)(X3HTSx -zoMd|Cfs@FouiQ-mG!Bb_cwX{_171u4vT&!Lx&C()d}JJjyvu!F~Zw#zis3J*nt%uS-GY0;v^v>7vIm^ui1`*m4HjT$9!adA?kMh&S~ubyaZZD0Wobi+ -Nt!OlQjn8=h}8{7KftRAxJA2I5meK7|9`PC8aW!bZR)TuIN%$ -TDOKKP&{B_$cUYhGz&2^`Q1@B}vS2*iaN1b!H~$48?p)Fa3vd5;(g-hoRAL(%{BD`<4F^&f6l-OWl(P -2Ks#6HlB0&in7b-%NDc;!wYSeHk)jh~Wc1!`5~z9>5Fyw&qK2z!7~wN6?2reaC0Ja}V-o=OXKG=oa#? -vXnh~^yrV#(@vc_$?)OB4IjV(_}jH>C*8VrGw;I#aIg~`Y~D*QMtqIj11Cph2;AJVMNGx@YD@lEvu4e -vOO`B|Ob&4=@FSZ{WQlq+?*SJz=Rt!88U1bFzP-f6#7M)24GkaooO=TCC`q=gZrCGoggrp6u*dMu);A -O;+HK|SUl{lm?@3LYHVso6jzgc{eDh6%1N5P8sQO`Wh>wq#&Ye3O*_tzFj)5OqLvvmOH+q2m;eBFLg3 -L4vcY&{V>3Qe*9ix)2zkH=$tpx5g)_d#o=Q<2)Xht)?8SJ>C$0UU^gq46ZT^G-LmSIVYh4 -p8+ai@=)(B%<3(+j@zd@BY<6-Fb_QF-Kz_)Hn7|HRvuDqy#+Y@*6<1s=?8Wfg_+RNg4jimqEC~b@(GU9?*8_(xvjk3ojUXwmQrNZ(ExuZsHu7v -Hb(JnRyi|RQPvbFNWXk7u_+8JGLo{C(Nv@EOS42aXq*JFS3R01Q!dt#;5ZxQQpskA6QhsjQp$r9%u8f -vT(6APustPH@pWQY)=lJVT-UMAlV~&S;EKq`bnfR=dowGx!Q -0Ec`&-YaF&x%eNu_E<8o$u*97LYG;PDdrM9$4C7uV9URBh}}3N=Qh!N8|liwfQ`T&0}Z8+<$Uha?09Y -9v2zBTcpD_UyqGiFg=A^-V%A}Bav-C8(7MM0bFmaSg~R=&Fu!LP9MjfY0P<4?d|t!mwr&Y^qI!zIU3v -c(-@fVFx;#&OyiA%efIeJWYj|G^R@EVaz8#SviT>G=nZE7AIsSPCbi+Yy616sKixiSV(aJrnE#B#zd}u%TI?rBL<)W`SFbk&J3VI1nD}Yaro{mRu@m_N{t>&5FTz$6%Mz#X8Lt`Li_ -Z27>G9C-q^3=OhnhS+whQe~4xl+`islRbY<^Fk%f7_X(8T6E#7*u&zwE*9eEmLpWlIi~RJu5491M^Lv -%mYJTe}Fv3_0B6)$G8A_jQ00T8~NoipigJ^*E}hH#*7(pZaCN*Ip@fr-$qT9{sJ}OrO&5cuxGgMgh-mg(t -3-p*M?5e?{x1^Trq6eu#}M_NA`2~UV(wS3AA8PU#HiU@9+1~XQCd8Jkn#Lw=ve=UmD)iH=i4mXkak%S -a7!Y9k_pBFF3FV4FA`Sk2d`-dPUSUsf}BCq%KZ>h2DhearyTj(d#y|*X=HOZ2iHx{3j-xFkwRci!Z(y -hc1%u6Q6@Ous{QjTXPdl-;jP4b=6n>we8vd{sKBde}rD)sD#*J8Zg{8n%=Ej{?V<+AAdYW?SDVpA0RI -vw}U3w9%yl-=u_GBY!}lfpgz6OUtgc&(h2%Z6a9U%k$xB$MjmzFhjf1N%ene*^E332nBU!Jarf?^(}j -9Qh3mbh$3m}eoWhXd?=`x{dG?R}Cr+Fg|I$k@#nBrg9tRGaSKH?Xfdg7#kNAv1|JL*PS)C|J9$m6lqW -vev9zJ|{%J}i)lcB@p$&-!U!|vkyu>Z&c^gtHb+ud{eareh^VD~oFi2!*_YFpeZD{Fsz3Uv-@y3m;XN -_DZgmLAv(v%G;9Z%^$ahxU9d2fx{NRvrhQmB*PG53qfS{$JO>!tNgz7x#5^bhN3v(wk09Oq8dee%j=B -z+`QKwUzJ(+8i$WOg_qfOje9-Yn}zL@|ZL1Hd*|yRN24h3z1UdpEhmUG>VCd83rC185#Y1^yu+!WMrf -~_Sj>l4xE{pX=)cX_krddp}&2$5qbprWCm;Mbz4=o%8dUSG-z<2>b=PW@%8j&=q+iSCEdGsH+Yv;YYy -a%_2c&0N?o4kp3DA%ztW$*HU8K~FOj-&$BrFM{S_Dn3>aYOjyZDz#5jjCV@XnOG`7hfoD~1T@ijkYw*S<7Zw)IyU4xG{_LZ+#@PRo;IJ|9qD6~j)~s -2kR*Y?==45+{m$wT{oUQ-`yAg4{QZ@S_}*@4sc@=43@Pn1mB;q8X -U|@elaq5yZPbr*=g$4-ym|A^u{#SEEI6oZ^nc}*SIql*&6T;ixgE2zvR3Ek=No=n9|oUQr?=K}BnPJ#)8NYMH&CYmsSUI2+Gcox+wq^1Ux-B0k-!VNK(A&xwy -6d@y^VYc?lx$Aa+(d^zy4pU-o{M|Cd6HlatrYQNu6uxgG>?eB{Xd-`)H@<;Mlas_ggNX?aIC|vYf$;p -8o`DfpkGdh|W=dk8rGmrSzAQvb1BKITjBkydkc_%RpF%@uQw;1Fw#EY-}Yk*n9J@%jU)pV#Mkh4KQ_= -FF#&!Db+@>UaX5$_S}5o;3bl5Z?{ae`TcoTG!n0skI&-~rR;#)d(MT^q7x>8REwpCooCZYOUbcN&p!y -V;}bPLcUamzgtXns|_BMChx*OXMXm7B5)0(8T>CdKSmrAp1J>Q@>#B7CkwhAF#fZu_&XB?Aujn=t&Os -k3_To5zBvM4_l9%!pnQ_z1PqbI}IJ`mpmKay!DMdrc+(yq5QaoJvk;XjSuTPy7N~5P#{!obmf>SLkQHUk>2c*D5&twG+ujO3vD3p8PCKb<4077*HOWq@6R+0~Q)uG# -EWxbZRafE@Rtt4H>nDTsnx0NyJx|kj?sL-3E5_fQKkHg{?^vDbdS#mVKKMZWzrWsDo;7cEhU%-H`tCs -etcSUdaf?&MwSJtrI^DJI%}zrzBh9g~c`uT`$YAI&&|IIQD<-)x9HMDJW4%7X?CCiDoEmz4XMNqrTnS -}Tl_p@(M!04g_rL}{`zN=IPfu<*_~x+Cm8qzo)Fe1Dx#E|&u$)3`tDH?=fD~jQH->nM@3kAw&~KgOWWAkj(6SOt((5!$2#FxXl~ -uy$lT_+U2?nSZp2s%kP^%H2G|{Xm*;QH-;rOGe>}gkC)iWh)7;}k1~ -~;VE~{!*R^I%)g?Y>K*5+-@E6CfCw=1tGFLXhqt`ei`bkUU_)wNP}wNbj>6kRb(*IcNpuGMu5bmd*T_ -90!}@l^Iy)g3~0k4W7m#?!&m#na96sHd+d)icyH$}`=Q<(cnU=vnSr>)Gfj@a*vH@)UUvd5(J=Z)I;) -Z?HGiTh|-uZSIZncJOxbcJn^!?dwhT4)u=mKINU_o$k%@&i5|#F88kWZuAy-cX)Ssi@YdKkokLr-Li& -eP0w1MwIeGyJ2Ja@c1(7M>@L~eve)Kp%qhw_lyf}C$*r7QH8(idQm=!d-=n#Gb5nDN=8nqE%AKDZ=!t -$Czm5MptAh2;SjH)W-t@+=*f{`R{#p@p>fXxYKXdWLlFvE -EfCn?6Zrz+R1vO#2a0!-l@;T^MC(P)@jp=FXx|E;%~Aar9Fr*KmYuzXBTaTVfXFJm;3#GE_s`|vFn$e --^#uGj21rJ-2Heju9ny0)$Q_n@vyjE-iz0_cjDdsO(pJb-rn6_zk5YrSNzoV;{M@o@%=mch6Q{v7uTx -Qy&j;Pn|~IxnS~S0gxkoj6I&&FG2o< -5!Emb2{2XU`zFXRizw`N>##5cmU_@MEesmHDAhhx2rJyEPSyMxv1%^=LaS5`nJEY!n6$Z^;htN+dW1v -%$d3>!HJY`V4*3;Q)c6zotapw_YzH!S)@iLs{xvi3^Inx#Z7UC1c*93Q>Wch+L^2MyNGhIgwuInO**q8E4`GYw2Hhc3FBI|B+1|bu* -GIE6Vr?0@M?<)s7T_-dpV)2Rz4_PXcz!NvPdpw;Cyk^x7Ur+Rmz0S&QZTYs(h@KwE50qj5CJSrr|u`? -?$16i7nRn?ZDXt7&)m(V2GGQ+PSv5Kd`e{g`_!D2VY7R5AZsy&f(94!3lwfREeNzYq$is>}Z1zk@r*j -`O3nW{V%|jqIUyOb|px@90Sx1B$=t!pBxf;myP1s_6lPjH6>BZFZUx%CK_bm}^7=&Va8c2QJYICw=tj -?MRN#-K}c06vJUx9{atqA5K7Fk1ttc#vzKdoC3O!$4Cj_O&5-}l?{O8QR+^vx|sh3E>(o46o9o1p(GP -7M0A50@Nj}RZz(*i*{pIHsJ5_$qP9WA1rDH|PJ}dQ3* -1Ln;l00?Qz});q*b0w?R~VB^^1h=TvNH4~l+7kH;$7irq!Iu;1u`IDam?q?jP_ -VGE@e@P@H<#HIgi)1K4w1svm)Oj7oD^85*Sg0pr#?T$gLKY`$PUiss-u79XmHRa>c3eyW|c9 -4TZMKE_sA)g6!r{RcMO3eiGw(#x&7j84}8BHPNEi$qn_$Yj(LrN}zUwAt&g2+Bi^KkM=197ri!?3WA} -5-!%#qz4I`_=ID)2}~qPGTt`43co+6d{rdHK&Wr9br)gVjVoYB0FTH!IJ#JKRxhLiZ^Sr$LMx(|x!ir -UT9ibEP?@)?@Pv0iRX)#F64xq?UK -KDhiGsq~WH)f!dURandB-i|WuTGo8S!#yQcuwc)3e5S#8KQ%*P|I7l6NcZ2ZBDE*-@li%IS5_*pE;oA -F*N2Mk0aLYzccyQReD=cACOR{5l;wX5ep9p3;DXkK3b7#Rkf(6ygb#Q8AK3>2jZ(giifDB7QpBwl=W0 -FV{qPi}dqA^yOUqm1q@h`zjqjGAnUE`r;a5^=va!q*;l|3RO}R1`-4z(3x2r^8%1VoDvLgrv_Lih997 ->YxbJBMsNBT?0JV1)5jNJJ$6lvlR3z4IJ{`B!Y2)4_h@PY@{?*F`ajw*c3W719NSSGpCDa^^WC_@T38 -W*y6YqYJ$(b6eW`W)O&}I!-&WH6gvCsjPj)<0&KX<0gZ@}J=&dBs{Xg)|1(vDZKnaGHV~7+93Oe5qU5 -cGQYvDc6bqL$YN5w2JD#U%XX2ur>HH24Au+q*0&PcqsU4F78x08JhFfTY1RHc8o&6K^?ix1dl$#^4KLr8?o64kmEQaMu -Y!1sUlC?W?#aeL9!~hLTr#6T3e;Oms``JC-U_@V%N8^&R)n%f6@_=wkCZ^J>PG@tr|xp&?>2K>JI7e< -e-x4!*M6^9hcF#=c?G#4{kD*rap@LcR$pz6hdepbWd^OyGK`SPY{YtfCQ@| -Z!TJR?hc^Q@!^V%(}p=9D$1~29jg%C6~SVoh&NRQncxr|a=L}H`4Pk!eU2fO{xWm(=ow|Oa*bg^iz-9 -MBy;+VvqLUi{v@ZRdTO-hl0Yl@KQw;i>e=%4Ve#r_hU9x1*xrF -cVvf&;x2c0@=RstnSbKXKs05wJie-J)G^(l5O#GJSAS-WzuK}M&fqjsvC(uXBETS1Jl>9S;rvoUV+l(_*GS|5Rp2|12I<>NlT(6jNh2G$dmD;0BVFG~;&^rqPZS#T-c}TwLX -+DRBnTUmwXhI)0z6&G#IyH0QZzd1El}+P=P{Du!~WqIv{B<2DY;Fx>gQ31+aoII4nCoh&MrmaH(R= -%CT(D4@QxY>46aB`2Tf{E=Cd)IqTfO_I8y?=rT#n@*V=LGM3=U$cn|&wyBtyyBvn@j-@tO`O)XZ20rn -y{FL?@9Wru~_cUjd7(v^k=$1Q64G>tDO&b_3$79f9ibuSk8Gx7b^{o+0v_xrX^Jsv&LEzX%7{}3G`ValZiwtD#KM>li-$Ki6|BB|u -~@#oTP**2^W)9(p%Opd+`ak%T3&s>c(Zu;i9_i1;$eAn?_VFd3Nqea-N9k+-dx>@x9{%W-rnE%YU^(c -JGxnb{yS`{`PPANW_(vT4HtOMnw`}|&-uVDqKh8k_{)e}%uD^W(K&D;I&Oq(?HJjesdZ|TJ*bC!R-VR -|7p=I47IDjg-av4CJAOVtJbF&b2MuMS%R0>)hlv2YR&M2bI>V>c@s -)u5a)8?@x?3-9pOH`?w*D%C2#lZU(uG_Q8WEaiZqS>?)z;scQx#9Onr@Ac8(p^Mm@sDa-{?L#v%uAIV -sP=%JMpL$9lJ9=VAsYpd)<|@!w)FRX19|xB_vKTlH#q@-DrI^l3hFYncQ#@Jo6?9-2zbXmiF))D~G_L -I3$98JctW2(lA{%kD4z@+wl4ZG9VZszT6PR`waMIF0N{-31o$cytMOGF4RR*yx$OYpQh{aW#LD!+ae3 -CH^%rWC{H>km+d@r1#fE=;|ZXSHI#j={2_LJ3>whyz;UOxJ-u3z2V&l=UDREeRTrjE%v)X(>=5(yS1m -XXuFEU&4qo_h_Bf4_SB7LP3c`7ODSrv*UNLwJ3me07h0V^#L)E(w1RpSYk3uTe~= -36WMCWQ|1|KE$a;l{@)XtFCbZ1p>+LL4QTJlnRkK`{mbJ0zRHhglcz)!NjNaki%@Rry#l}qEbJUzQ2D3BoHU|1KkMTLQAUi_w_Q -v#cd;19tJRgtudMJ5Y1G8g!Wop~!8KvJ&$f#f91zRkTgZlG!37sd?8d^&;#k6%ggUbMC?Zh2z>bp% -j9Ha7Xw)iG2DOamD4xi?#ug|7f_j(d-10M!Ek03HAU0B~t=FJE? -LZe(wAFJx(RbaHPmUtei%X>?y-E^v8ulfh2JAP|P{eG0NCO&Z_8gQgyuRGVx)Y0QKQo1hdnK#h-YyDT -uJxR)~DPiLTC_GI)#?wGL7XdWcj`H$LEKxeS2!4X@}Qi4{cNDNjE%}a6d%a2{r#VYkx7$PH -5#Zzz;WeDFHkfXCWD=c^bL#G(ncD)!Gm_se!F#e{8_Ns)ST)DL(4rbu)gMq -x0|XQR000O8%K%wh9aOw&JOBUyKmY&$9smFUaA|NaUv_0~WN&gWWNCABa&IqRUu|J&ZeL$6aCu8B%Fk -8MOUW!QDau#K%q_?-DpANy%*^BB%1l#;kIzfYO^uINu(efCijN10$HyyKaVY=+h@+v&1prV>0|XQR00 -0O8%K%whN;++8E-e567Qp}j8~^|SaA|NaUv_0~WN&gWWNCABa&IqWX>)XPZ!U0o?R{%^+cvi7cmE0ub -MLC;%FLrPZLg+o$4%T;tEru3J8kDC8JUtGn>9u1kW?JCr~m!z2LJ*fD9L%WbIv+itBpwl8yg!N`vqX* --e2yos`CCM&F{muK~Cmt8eb(vmW*W%HGhkrYG^x)A$agvts{#wF!DDgC%%Dj>@vC3z%6!o< -fN6Tai|EMnm@uMuOw8+Jy_<`tQfk=Id`hN!q>tZDq$y($^EmoBT5NRdmX(okyGnLC)q`8 -637J2$T!Op@dqSRB%z!kz0Rd{5SSN?P_>lPbe>M3U-{K4xsu|lcrD8ux+9jdT%;8aWCiHW07{lF(mJW ->O{-J!-d+?%dl&HL@{)R2TwV$w_zi;XS?{VEz*U69(==bb5l=KN0CHa}%i?u9lht0DLy-lwCz46As#S -AXWoA_oX)VeX;8f%=2CM0{NC0CPL0}*lv%LijdYa3FQYJGRM8-snfNFrlYsIZrbI_MjjiZDOmJ*n8A) -&uB_0%|}S?^eWAhY;U++59h0VI=qCEHA0Lt#CU9Dd?~O9{xuNr20Ng8Ef&~;MFCu-% -OqJ^uVZlpRLE)=96<54q|qfrm_QRbO%_t*atfoLlxuo{vx`yO`y~FGm`UJ`X@V@rbIu?E1ELCwPvXy^ -n+srr8R7*^7FQBkD;A@=+5-qPY6E(&>x)3AvAdVz7DpmO!msVts}v^oo@($>{3*3>8qFG$J29^^>k@6}RQxmYjZum0X%U9N}pCV&hUnqi}RO307s{c%|qC0DxF4EtvWE>#hSFs4v*Z#-6$bv%Z5(Zl%v -#Seh!d($j|(N}5*YxGRrJ_BDQzd$=6dP;swbNg|!O0(HGUo9q5z11*BFv3BA>8z=)kZQdc0}s^I&_;| -tXX%Y>3Qvof%wg{8hR4rOMn8j;|K`c)40>5qF=*(t1QvsN*o&;M(LhA^Wj&4hTNpU9RI0drwR~&3o_M;~e=w -7AWVA{_?_|mj);2|{B_|N1pkjc?+;lhd;wj-HPG|M3$f=I9BK?>8FV -sK}xr{J%`5nFP86o1fC(tjJWEC1_9}!U15uq~huYS2&S~wPLCY&74#Ce!eilSd$nPBZ%&I9z?5|?luA=$zw>v`5{K~>G{{rP=d0R<+4mG+5DAM9m9KUiYS -xiuiZc@S;FXU;uG;Vt_|Av8v12w-WKIe#U=n&FRx|BmCwFAekwj4h|k1NO_4B+ru9(ubNU*tiB>wWNe -&V$R2{qq`hEi6@5GVDK!fRBrf3wfM!Du$El^w{G?)phs6FZ5hfwKi6`M>#jVgK4DQ-Rle?A$VU!0vfE -XpGk8jSc2jAs0X#z8!TDGrt-(P>)bIHVoK6unZSb-b5i$8vOkR1F9C1!p4B&!4vDaAf~k3a5 -5XkxgiPYjlNYyv%Kp{-S17+t{h~ZoZTJOv0wf_#uKdOMdhkUaj|D=b_1Zx`UtrLD3kqs -!!`)gHZ1Ux&-X`(S6_<1b0sYMF(j|CU+4^|qMato@f&)_wyHD*EgPYK-B-tL`yQuRv76Am3igT!8Q|( -X+gw0pEr5M&H8hb%&a#M>La@G|9&e67&}IV2cKF;k^o3Z@PEus&18!z;Lf_{FfN(SR`qSK$9XVXZ}k$ -ePp2ub7Bw7FwypNd-GM6Id8^n=}w9wgU(-+E7=zZU$Qsx0*VcU{@*LMv(c-1X&4Qm*tTK$hWauv&U@Hu7Q49fv0w}9ViY_=5PZuo5wcp>gzl+-sNW|h9^9O3DmRc -Hg^heS9lKf=%6Vem9mcu|Npw9@v8JSZe96y<^$`ZxB?aA@)FeIXfF1{tgw%xGEXwpkQ{8ORvGG|y$~IMM6o!~=wwV<@H4&wEmGBIEICYjbW<2VL -MJf_1{DaC{3lAU-x(W586mU+MXaE~8mizm)<~vhQDH -zXtJ7ts`lB&P*_Ah8yIvEZO2MAPAQuMm>?SG@VsT#)=m;9acr6$NRgo(V!nhK6jy=K%*W9U~8OyN@GG -&boEi2E2Mg-8VF07cSXvIXyy5!&u#-5%;OE_AjP=fcfH~d -uSRRdLgV`2Foz(=deDz8_OX1(Y1Xq%bl(-V4B1~dAL!u-zU>^q>hAH}ep2vhnoJf5E0o`m`I4LxdGTU -J&J=g-fNRdB;<#SS-^C>U43N35hAp2e@Z1?^~&{0ef^LK2O0y8t}epDE^B7DdfvzSfVbhgq?j(QKs8k -MWUl)g6^@KzvUM-NG?H+AXmAWKj54<`;!iAe}76OAHa*yk0=D(bqqSXZZO2k3Vn$cn@U^1b9!<84La@ -&dIp1l&64D7r@P2m{I)!o^7Aa0Q!L*T}Ld^pZL+n6SaGm$xMyjGySN(rp|~zOPdidlWJ)(+?jf`y*X3 -QiFWmH!gmDH`ih&TDbN=eFA2oCd9*P%?j1LCt{?g98qYCK@7Qmeg1^M<+8Rui4$ -3yPxDer>)SW*Xud7dS&obA>iZAs(R?STwc*(_d~{ekePeO7I2!$|IBn&8_!b@&d$>UZRp#jFg$4EPr4 -271&1pgZI(aQoFQ{elE2-2=!3(`cqK=+`nwiZgVh`oWbOmFQbt1=25ZpY~5~;q{iUQGN~%*>wR -y~m?6T3?D25uk)T3x7EH_LmJ8Is;-vFyZpL8I=4$igBm -gbhX$|ju-t0=B-IUeGO_=RS_@t$zzwn1rolYKrAJgKHnlD8KNl_PE!+sv-)c^LDM3BZDg=kqPTz+DjF -8oW-6Rh6wRo>b7x%n!xik=AF_75= -D!zlQ%9csg#Lr6RfOX+wBEk{d3CBzM|w-MwaiVt;Y^dKOpF+#kvqz{957g_I1`)~_iAYH$V02PT5fl -hfJcaQ-&Vj~rvogH6X{B!(1h#{+KnbO9YKa&yz!{P?KiT*hdnM!7+^rA2HhLv;%?I6!ZU+B5>WfLSJ1 -gic<6u^En+`Cb1S+YutFCA|QNid8MR`ngFnh`7?R5&G0qTm`s0pQFm@DPIC~0m2s?B#DBh~A%fxUP!ILry -?z-GnYHr`E;6}R@z(>Fj*Z~%)E1AZhhn;=k1IgtWQ1rB?N&*Mj=@|6w%>F< -z+7$ip-)L^i`m*PWMK{feHoMt%wn7!K0!=vbSXRO&UoNb9ibZ!JA4^8z#816Eq9CsXVP|EK -1NUDp17%FBeVjY7_D3TT%IaoJadj#9lW-u~snXa615iXG3TS9Zdkr+$>QrqBdcN4I?vXcXF^svviolT -g2V&&=yUMUJ{RBEL}51GsEIQv(4l@jAN-8T4jxI1HXdNSV4zw3f>rq+Pp&abney7ZxavKDDPW#lM5<|W{Kv#7Sg4cxuOjh29J$@dj -N^rp35_f&(5vws6lYLs_FYP(TkIX1x!b9ci~`yvgV9g5r4(yW&~4bB^1dmS2k`(po{Rd`_ans)pSc&G -2c+2Om(>NVBgx>ii3#(@=MYIS>cEETG5W4jfKQ`lLgnKdk=>E}y-puRd95?%!-XnM2UcBZffUbBQlqb -`=%WW*7GyP0yVUs(jeEKu57^}yRd;Amu9^b^tD#`UT!iXnn!2d8F5eNaiHECyk3Cyq{~L};i}KCBwi~&5 -l^uw`%!V#!}Eqzuuv~XO|zns*&4O64)ku)cPYj`uaa^`L0DeT#zRF23jKE&#VerCg2qY -j6;d0c(db=fzVBzX)-0e{n`08*70rjw&`jBG8nT`cL$mp;yUpSMxY?nBM{4WRhCUSwb5I?x8DV%|}Vq+ -RswTQ-a{-#A{3{FWKb5JLiWuZ;j;|(}fc9xR{IaT=eQ2aBRM%=+X{Pc -jXkp0?wHxIzk}UrNlQ1FQdC-|kEskvN|cf&b`lPo*s_hnxe3SFfwvD>1uPTrU=8oLrWDT%9f|Ukmpz} -hm?Juxj%V#XDC!iwGVL0(8dT*>5B$y)R3I))AYZ`9oxZSz7EJG`{XG~ejda%v8hHk2zFtfHlxr?B(V-K6T;+c`}mfknmUS -5hGyAJ&!o*C9KeWaX&SPi76?vJ)9?$_4qy4*kc2C!6|rj8$3-kDq0ct=#}50<~~z=j0z2OIis*~r%aU -1h`nR4(`vbV!&hJlXic7$bBl_kTI!&ryJ3PGDv6zYOrM4AA`U#tK{Yczf8~JWmvwD+=HgmFj7VN%m&r -Wmzm`S+BPzeEI@2eW@roCSCFX46$gvUIJV2`Fqo2!28Hvg8GN9!T|#YR5h`$Scp^Hc#O%k1{nOCCRtW -ltNg+a9GI9RO36yr79b2WT#mmd(NN+ce8m=7$4pP3_!ZG-JCbO&AHM#Te|T6oB -_Cnau}cJ~a8eUc5NU*RKYc*y|O|k&_~qLx&5nPCTAdgz`|ppLAw$JPsAY0)Eq1vT3=P{4%5@VxqXsF+ -CP$!g^sEhd(L_Gh!aA$uF@c^b4!*EBK}!dr+qDS -4BRCe@r4T+6QHyv~$JEGkuP?eP}-E5D3mmLOwo%f%hdXlHd-c*(#Fa)dLlYD?+YO~~KP_= -mv8>&mfuNVhFheb&jf5gVn#wB6W8W-8;CX~>Jy6FkklC69h`5(c9qLj0`7Qh^cObay-xar*>xOn|S82 -UcsV;BnK!)BIXOajC%#_KiqIjU)i8B&7y}n07Ht1?v@4sVQdE+Q8C~{`bCVK7fw56z(N#Gee%b@q5ru -pmQM(wmDcmXr8Q3RSF8xdsY7D)v2l;+cHHIrk2HZ*wyUtnw>CnwE%Q`T}55Is`xm}1U4KJ8fxUBz^(E -6yacrCmJH3DTpq3ZjGEDyN|q6&=#v2sa>fwHRT!ApJR`j_AANewE+p-bPF?9pxbgV^_Gcy=1;4}nbUiK}Ze%vUWmRW^AinTY -~g~k;Q1DHpTc^1;7MDjVr5?dS`T8$>CDyx8nriGGOFAV|HOPl_JgIaJ0SsT}+%N)Cy -S*6J^H5E@8Z>=sfs@1&51m$_ZCWZVZqW3_NuKBih -gK(>{B$lEqw6fK#s*b@Msbm5+A~*;RuMWI}J(L0sCOXl9)l^^i?9su5cIe^}>Q?waR0&)l1wmKEYt4H -!OVkH!n1T_80}KE44WgbtG5{L?T!lpTPh`(;3qbcRW2*?Gab3U*RN@Ny_duFqsid0U*f^S524nZUPIre)HSsal7HlJ0Y+9)nFCbX$f*n2t|kS*^k8E -eTlJ?~K-N?By{KXw?f91>&3n4=Ktz;79%Yb$B{S%)oxs)wxvI)>mQEDdo;@6$jhKA%9Sxc6^afhLp># -TLg%H}oOs@oW*m+~BNID{K15iX7hhSL!u?ac|re1X;k(y9jz_w!eK9$35Io0}m&f`6Ca(Z#h@*$ncM0 -pb+Dv-6*iNgJTus!zoc@J=kK0)mP;U5OFFm|0{DqOTGkDiUhgp(=FntYR#lFwa`pn>C_rWJO**FovNp -hV!cU%W+{4wAaG7=|g@9)=EVT;6DL%Lag5=(!UpV2IEy7wxw|l_lki^8|GW(KsQCev77hR{d? -I|gr0rcf<3LwlBWKpvT`mhBfxjd_GFQP_({thYxm8W8|G`f4sVapOf41T%q6yC*=O1ovdhS=0th2O%( -7Io{2KJBw|#6nRs{84SD6zBIGHA>Hi9hO{ik@F_spy)DhI)H+SLVP)CO{TJ`4acz!tHE+z7dHdDfQLt -jmc$ir&GPx_gG?-Czrq^(mr9Q2V<7uluwanup9<^3_dhXyy-poAwEi{w=QvN;=Ud6!lRFwl?LI5o8Q5 -CH>n<~(ldsXP&rb=p@ylU!gsG=S?RYO09ijKW~m#T{YI#|`;VW{i+?}L>Co9cA^z^5RHW_y~h*}UoKE -0N6WX!8y#^`n?&4Kef~adn9A0|{Q;)fKTr%JgOxqWdT{VQ$TpjXA;nl{oFtHtDns95>bO4oRegxuctP -p?NreH2zv=zqTcLzYJT+K`}w2sB`(wJE+cLji(PL25v(zDS`vv{K>!rQFcl;M(p30n$|)Qz)Iqls>Q5kJa6Ky-48oRpT -hQ;p$&jR1Isdz?Y4aRPdi?sbY?Y?&YK#;P{RQN41|9x)TZ`Zd*&iDXhk`GzsT^#Mw_q+ZQXj$6f -w8gY{J9Zg7gP8ud3RM;4~juGH*|;r>BPgyCcEg`1Phx=W>1C!IHGn*47T -n)B#4-jMa1D*by)eQF2Lo*Wv5?)kAjZ5vTMJZHS*67Of)G)LJ&t(!~s^d(IC7iO_0RZ$eO0NOK1Jq1e -#**`WC+(ZW*q1}ZQWwq$&4Cf^5DNoFS*!O1_?dYp4@y5XZ98)7nn-e)B4UNAGJuRk) -@8RMIO}f_Kh&hQD{wf0XnLi)i#TAeFrH&YF?Cvspimd0(buhhWbS`TVjy;7x_#IDi7Ul5g5qFwQ>izl -^`4f4?$+csc>*AMNiybb-b6zGt=EvfEV6HXC+lje}NHt@x@eR?Di_wzd?b*g)81@<6c}<<*KvqYWGb% -gx`lk;8T|sN8I3;UrOko$fpEj-PfQgox^=%`|qtxEBvzSsiy0WW9UGoqS>Mwc}oXgtzJVL#yLX7OHpd -_oi%MZ`WThWy9Na_t7gm9(p6YqgtUZZA_&_@Iu1!E@8!+F1On5SRh<9#;UAU?Ky08%ALAtG7N=<|L*; -JVe4Y!-D*1cp%8ffZ!h@{Tt}*!!K}nbw^1!t^v0#0kbxUqm -K$+8FcdklFWgZJjXlpzZ92G_l!Mjg3)`&GmdQ%8Klw$^RBkw3{}`AKJU5PqNnn+-TgL84kg2o$<^9v2HjfTBg&LR%dvC6Y -KA-9W&OZUm15*kDEWjxbnl*oqOl6;c*$Mz%e=Ml^SZo)T=#{IrKZ{EyOQHeGM5XCWT&48gI`#rokp8@Hb%SB;P019`pmby=*gt}Vdod|fdAxbr*4tB|k` -pF5o3Y1$Q^`_OR>O3Lv3;+uodd5kDmM=jg1TFcfSivMu7VTc?CPru6K8;WOF*YHSA>*vy4_M>3$sSIUc -2%NQsHSVH_xBhzAEnBi)gtHkV`WjsI8`l$`|(H9WOYWl41{WHn8exV;baAxx#eWN1}itg<>ippjP>eP -WiI1s*pY;O(p_UB;EmTcV-|2S<9j#FDa}#(kA2L27#Os^hJYDXt>=MA@gz#OemsD!t07YeRf{NT{^RE -6q@;%S_X<$=D)#oRyt+(BnWfEZ_Iwj406(a_{ELh2>6QD2W|PPcLhYe9t@>@GJ@p66IY@!-Sn+|ocF!Yy230XYVGBKq2(W5sp7Nqy)t^)>%Dy -PasOriKL;%6&7&@ASs_vYqTaQ(p1z{qxG#g0ZK`yV*b< -+7zjq@1%A96C7UFp&b=s2Fq?2?H;FL}s!pLBH{VU#w+nfxq%+-ZO<%FL-XtScZ;eezxVo}9(0CKIbU7 -nx_sB3+DMpN5I|IiEIZlH9FY*wHew{t|R$ttf`JCOAp>6TCyzSp>(Q`23~`2n8Yo+af>rS2#4(m7#VF -V1i$#B*w*+7)FVqgSnMkyLP-DfezsvK?%>?hG$~j4_7{ -XBd}1w9ZP!K=|p+n7iBqKuAv*@AA+p^W^x)ljl#LHo)$Q$BI^nKPAmTSz;5@+$kY(0j1$rURdd=bF>2 -+6--GdXV3VFKKtjvTBOh$i*)w3y1)hc9(Po2TOuvmYnh+6>1wl?HZ#?~{T+= -?p@FvsNyPK>5=%KA<^!uM;wW0}0-2hR^6mlB9c}=Eg4$2t~u3q#qkMuUUXxO$aU@B+eU){X!L+sU)w+IW}18Um@;y1-0e)%x0sJ}_I|r*A8gQ;4w!kWGyUINvtcJ`e3x)=NVcuoo15<-- -KaLnwX*t9f=!7hC7u2t>BD60kUJZq99yH!p)t@Y=|WxGw2;4*#F2s=-G$KaqzAf!NB%G{&u^oryUVnL -qAjvCp_SGg-f+@CavO{Kb_^z|rfKRe?$ZS_%g!*&d>P|9Jn|YIZPGpVi2Y5PNw9jqL3Q2D-+HsjMvdp -PtkN*&%Jto@>l9-KwCVZd?XOUs&VotA -%8^;$_0_)faYJ}i!!lnxDx8wmzbW?-Ss=~c&V&I#$vgtsGqgiH7zHCr|FW;d``H3h@EPUq}5JKL9985 -7v|J5*gzr@ytuasxP?(Y0JXU3Rhlz|VV;A@QO&3#_4D_&L9r-roRwfD|XuT}P3mFs3n+vz>qN*~^X?i#HfJZjQ})&n2h2{3`zD!ig -azJ9eMgnN`$rRkwUoorN~S9!2;-2Dh4h$D<>+;TeQ5bhVGO_$LQiD% -u{S_DL+iMDQaL#|$$?I%5n=gD!UzdpmrFKRl75>#74Ges@RTo}8XFy*>}Zq0of8_OrpnFr;cD^{Vyep -7;rKd6zZy6g&u-)CuR`q4O&V9+%<7)iwPeBexXBUS@5I-P2rZvaS%D{$0rUDi)7n;+~!l#A%sKaq;OG -?~|odDJY2oz${U93A_tDr$d96s37ZNsT|m}#H8|VXUYtIj9v89*Cvaf+b12rNY_Xuz5wsIr*}Cp3NdxVkc$^ih8C&;Xh|h+2`!nb0lzE9)gwV*6Psj!$nCY>N -Of*FRlxrmNNX{oxb&SIb+iXSG2HFaqhX)U>#c@~JTEUI36Pw6(SKvPa5ztVpXl2w%b?AY$q8`p+&fX^q89x%^mUa!ir7Q5Q7P@IDF$`g$@cWUZVOQ+n6aS5)<_f^h|(F+$zb<)0jX{rp>}NHUx8d9~6MI__A -kZp*W26{>g$(1-$7@tz|GA#9*wq*fcTa=-}X&s>o^e;}`h%!v-{C%j4R5G7wT@9^Zd$^K^*Wu%|^38I_0oRurOr>a8#4| -$4XQJDR+~P|jux$*Pt;^{q%GXgWA#1MlCFyqDaDkdEMX|HJ|3fd8jn?MorTrje*sWS0|XQR000O8%K% -wh000000ssI2000008vpc!Jc4cm4Z*nhbWNu+EOi4pUPE$oLZDh4l&2rl|5WeFnc5;em7$vcLYEKXeN -tjZk0)SR@q)8~o3`uwh$R59acOkh}>T@qP#NGY=c9*(3uB -#B=wxO>WzQ^JWV;k-4+9A9Lowh@$*9Oqk-2y7>(iByvcC;ji-v5e*M -?dt~F!}eUl)sCClw2Paa=*fZrE5VK9_-ZPS4)S1T1Fo&oa4cJLrjO~21hIqP6{OWV|<}&x!e(iC)?2> -;hdxMn(2h5^D%{Rc40PBgRWO*&y@gt?^Jy+3sJ!G0A?c~uXhI}*Bb+yx-5j!z^ekkRz}HXV``<@C$4e -wQ!4Q40OIFP39X@4Ah)|xA{e48UKuHL2C~Y_rjXS-0|bDoHmsCeX|T=ISt9*72K5TILggF$$(OQ}=8* -VS(p1E?sukm~F?$j3DksWk7FuS^n;lon4alYmcxLP)fcm1u-R;g; -XLpWas^rV{)jk%$VMZT;c=65NL+0Lw3Q`;zxr`ScHOa`5L`_{ul{dB4HE>?j&!N*pg<@6aWAK2ms3fSz7}a@w*=Z003kI000~S003}la4% -nWWo~3|axZCQZecH9UukY>bYEXCaCvo+u};J=42JhUMbrr-#Dv%w;9y|k1rV~*)NRA1c9o_lczdFx=_ -PliQ~dp@WBKbWkK$gjMWM%utrItu@ZbCE$McJB^_#V9?>3Z|bdpB52X1jUnam>mABp)|5V6VGZub -%e%0k=MiTG#~*VbEru>Ee$9h8%=2Q?5ZPmDwZuOKSVquozF}5%i`vTN3b}`DS|M(e<%$7y~8+Xp%WOS --VHgQoz>KD!EM!rE&Fv?Nj=VK)f))irKa`bB-K&&2(5H7-32(Ioi!6cRRKt20K}8{4A0hCTE|K1!e9< -daN+AezIse?G?}Wi%aUw1OML0GOm{karwK%>sfu1xb+yhEEJe>udP1f4UCJfR)t5;j-$!57A5cpJ1QY --O00;of09jign_y2m0{{SC4FCWc0001RX>c!Jc4cm4Z*nhbWNu+EV{c?-V=i!cFFk?wQMOFIxcir%@CYoN7(p+Rzi9K<*CdVMOpLp$ZNY4ZOlAsG)~ -M8w?C=eA*q*XN(_Km^dZ-uYF###f<@-5hl!~w`Lz!1x50#RchJxQIu(DOOWiI3*FSVE>L=9vZSKqwyy -DypYd*9NwuOT@rF~(tK&^b5T0MjTz7IB=BNh=@7A!VzUW`u+Ae@2^{_-6EK60*IsDz0`dw`nM(qQ8m<@|;;0%e(AVs)KS -BpAktSppI84RI8u{v?E3^1m&~&VcKNTk@>6tJ!x8q>~K+TOOiiV$_wllGC~dQ6?U=6ZcNjVNa~}?LeK -+y(u|}#;Eqjd2?;#JLH|jaITqIlJh;p_me?Vb9W~~#`#wiPCO$ed%dDWWA3^F=@1XP4l<`U1_j4(7Ze -1E5hxhJSoiaJ4=8x@7@IgC*j8Yg32mq-Jkn-1$+U05(y^n+2qH1n1!HmgK(O){sPaF6%A;dS^H(q>-0 -QHuJTz(~t$7a`?E>-1D4?q^iW$TUw}WP3rOvyoa2sk=8K`~)TK-1Wv-+wQYBW$)Eaz1t3cHdgg08>;i -^G!jo<8d?#;2l^Vc_kMqJjeyRtoxjGI`a3l#ue76cBaJLu1KC!5W*|SGFPfb!h7M^E3~;|Ay@$a>EY) -BDD9ng;}qurrS%{sJOS(6AwV$ifdQeYuA-&w?#t{<*GmJw!(|J=KFFexMJFNe$bLM`UOx+0|XQR000O -8%K%why<-UJ#{d8TJOKaz8UO$QaA|NaUv_0~WN&gWX=H9;FJo_QaA9;VaCudaJ&VII42F0A3ehQnTq& -K73+l#}{RxUTaqB -pWdarXtt#XBMtw!aHbv#_6qXOaUX3qC1MP6r1+YHL&09KEW|&AXd1l5A*!P2}xMF=ddZw4Nyx11QY-O -00;of09jkUIKChD3jhGOGXMY>0001RX>c!Jc4cm4Z*nhbWNu+EV{dY0E^v9(8flN)Huk%J1*?9s5Zmx -v?q-lFkj$}7cQPrGNl>g~7_>~=%E*#dl;+r=|9#&hDT$;gAK9ea6_9vDk?%f9hU+xRaxx1!3-V4V-!f -MIRm5SC%-Njgw6n;PHOZgSFkTY%`E8nqNlc?2`4Gb&9XV%Eeh%j`t(OkMIL~OjWW4%b4orh2&S?k$ru -Sh+*AxcUvoHo!(;%XplQZ(XGo=xov1poRY!N=f(9C;``+cu7Rb)=*NzSJkgHIV70EjewEgsn -8p*@D<+)D6;8v^t;kf5u;l{~qi`GxR~Vs2i*@SmOI)TDZWOXJjZ>sUf6HUmVjltJc*s@yRH78Tm+~LK -;`s`w-v7$wN$n70qanvkckOpm`E=#_}SI$$iOH$frEW!Za8B@I31g`7T}rbF(CgPW8nNU+X5>9Et2zZ -aR=a_*ni<{1RVU#7{_4mum=~a>fk|d?5S)9Il-2RS*gg6Rs|04^WJ?-iK#jPKU?8|KSb5v(^+QmurRE -r_%Ey7BBNvSG=6cm@^iH^bq32;vBRC3*>B{{7gniwLTQyD;ja86LOms%s>oq(R{XG?h%k~7~xEaL8pR -@l{Jkg*RxG%iVwfli6gDA;BM`M91FT#CvFQ4dU@Rh85kX+)tE7m)-{a_s%ow -3*8bMG&GNl}dG9FIe-Coyb%LODPKq{(4C5ACnhVw_DCq?sSaj|9@#1Tnrj~r;sR)NlPNh1WJ)#@&&Ej -oi>VmjTAP}#zW -#zK}GbhD_T%4^S_U0W4;iuTo|qpL9Ext>Qf<*W<%1Qu@T@483cVYUNG9r^-Q$%G*y1B*q@*NFntp0R8 -SMG`x$ilZ?G2$*)0#NX2)bSna3*(Xk}731QK1rR(=vS>Wk^XDE?0R!Px=vCDl~( -vu_XlCaBZvw~XlO#*dMo^1D2PqPAaCh0uVS)_AFXOOD>p5~7HA)kWbz07^I#bU184d(lrxSmh-30v~c -wMa5Bm6L}sUy(41*b=OwmY -{Ji!5&c+G@rj_T^q>{1%!ubxuq>88ouM!F@5ho{GqKxpf?Qn;dAZ&D>5;dV#aTvR>BKznJZo^PU3lIV -Mw@8wYcpd8zZThtGt^Ui6OBM8v95|*N5i&)}CU<(8*1g9cnbfYNzH}qCFSLfgks*3=7mLPT_7s?%w&5 -eRZg-2Y$8#(vPW=l2KpN(z;}nUE3_Ve65;_-A$XpA(5wC8MgD*2no^$#RYUaAJi+VAq5r -rRm*zev=JUPZ+x*odE-5rvP&8L8l@9y6}8Q5ET+=gid^B8KZX=8dpY@^XzSr+k{3dPFZ4YgkWE9J4(X -L3N9Q7EnKPc5GrnG$O&*SZI%!ydO~0VX*tR-w`n0wsY0?gr+ut+B?zEbBH|kEVjxeh>Mvu2!?IxP~lB -Qo6xXZ)7L>&#UUY#0IN_Oy84z5f_6cc#tu$VbYPG<-St?gZ;75r*RW9s=R5c}&tAkN*je*J~UuIROg( -F`9)Be3qXS`;tN#NhMYB>~f~1Prz1EI|xMSI=8XGZR81ahIfBxk$U(kF=4Cw9!|S<_J}4;|&QgcaTN2 -QrpP*L$i!NO`5e}$e|R^X&h90L|uE8o)XJFLGe47Gr22YQ`{O;t*eZfW(n( -SgI2k(e{wf63M6U@JhWw3Oykaq*4Acl^PHocuN`z$otVr~jE}wFKEAwuec??kZ)rHGu)MEDa{l(__VV -}JUmIxh?5TA(#8s1v9OrOsO@A$iZ>^gf-meMj1p9f!T6R^K(~T4l5ozdNvNj(iJ~dfPB5m-Qr@t>-H&$E&9O8HTfI>y{V8V -35k+;*({?pP%>Rh_bzW%#MS+Z8jn1+ujQq&AiaH|x|VvDe0NldWs&->!XJfX7(ySj))i#C8mA#TF-cLnD -k05b9p`mA2x9{HDwhqxv$OMe3>wM4LrY#B^OjCMtZbmxkP}u7&$+n$HQ#B36+L5AZ$v|OcK3vn2rX?hJV%OeTJt>JpXqvG#O}q1C4GFfqIg`(irsDQPwMT%(ytvU3%gjI{$ -M!+xyPoI0V|{=tFD6xYBFd{u>gEIu0w`DUFDOJl+`c~^;We>5Q`bVyeU@=NRcN+ -o)$U5+YWZ|4Cu}BSu(5F&v;W#CO}m|!y3Oms0iFy@_;AMB^hH42)4TQ+Y54dGLrXLK*QLY0J%4U>JBf%Z8&h(Jr06E2~0FTN3K4v0;;1voY=sc#YIUlXy& -8W+~mthSDSI+RaY}~TWgj**1Ds5U2VK#NT)pHLI74t;@EkVw)E^w*p6A1V_|B+M63u`;M*vGH{MQ!`z ->N=7s*B@G;*7~C@gpk_}Z}}QwLSLsI(C`dtexbgyiJ7?#uS)>Gq)5*7p$c)YGf#NcMc0#$h=D2W%P%b -kGWt`#PTE>Lh}kb(R@Ym>>RmKGY9n!jg(zSr&-TpT+dxShrbv%3(t%jguIAwXgn!#DGi!=Tu5oa(T1%*xGMtDD-O3n72I<0~)zs< -5Hzy#GZ=1QY-O00;of09jiGrvN+*IsgFTF#!M^0001RX>c!Jc4cm4Z*nhbWNu+EX=H9; -WMOn+E^v8uom;OZIdYxf=T|h^FESv@n|D%x9(eT?mLCMeZ-dsTEw?qE5oX#KLC}9U_sKZbwZ4)IC?vQ -tPiCyeNHRzUbN`F){_4N~^vCc1+fTp!@wr{DkX*T4SmFTVTUr9X3j_WJ%;fBelKe){{L|9 -GK){MC2&-}nFK4?q6+ci;c{cYpFxeE)m>-%ws3|3Zm>DsVEmD%=g;3ZDnAFM(?`I4PVBE(%wJhr-j~r -SN{>?gj4Qfw#hseXncJ*RWgu!<2g*il_D9*(;vy`i$$xdGM+Cx5v8Uvf{29z7_X&eZ=+Au8+7r+4ZRc -Z-#f^!|(}w8P5AZyiNgL?YSc;fNm^K)y{aRIA^v7J`W^LG2a_4V-f8u-)o8NtjAE<|n2TzG2Ehw*~D{P@>@`0@Y1V^aI5x -L2A|q0ifT;VpMh`*E10Z^s$#t9Zd{SUW~{kwn8*lxxit>V{tjCJ~w1nUgmh6(P -M;o5cubD|KmZNE+3qv|R85}x87ylGqA?VY;E$P>vbXxRU3|fp@Oj^uZELyBuY+CGE99o=O1f#9RBS -oito$__c*C}78e4X-j%GW7hr+l6Ab;{Q%U#EPX@^#ABDPOOAz4GPU -$1<<^7YEsE8n1egYpf^Hz?nre1q~0$~P$ApnQY!4azqt-=KVh@(s#2DBq}jqw$~P+C -sC=XHjmkGF->7_}@{P(jD&M4hlk!ciJI^3BRO3*T6S7NZuE7PA(M7ONJ8!i@*D=(QNM7`2$Rn6+56ShWZom9NGssVm>Ee24NK%6BN=p?rt(9m;no-= -Tbm@*T=|DBq!chw>fDcPPQBjHWW0N?|I6DK=+~g0eC~FjyH44fO&#hW)TI^aBaq>Ho` -HXPi#kZ0EmtxRj)MC{=lTlH+yvt7>)Elw>iEp9Cyx${Tp@>S7kQTe_@m#=R14qd)3CM_!8cj)qU -QTe_@m#>S;Hz?nre1q~0%J&_*d{umhF1h?tbb8inQTayY8fS}&yC}h;1go}JRo%O)dslVus_tFYy{o!+Rrjvy-c_5g+O;^eIJ -LO6xV5OwSJURJX@xee(54mIw7G6tzfJ46Y5g|k+mvt9`fXajP3yO5{Wh)NruEyke!JFh*ZS>Rzg_FMY -yEbu->&uB)t|fib637y`F7>om2X$RUHJ~>JCyHGzC-yAQ~i0WKTq}Nss23GpQrls)c1Gl`#Y8I)c1EP->L8K()wM>cj^1PlVmi9PX`;@Ic?A9K3YY)4%huv!PR-3mPxz)(6MsDpix3;2NYvHIrANA*>^?Sr)i%xYGCk| -XR$S+PDxVWIjsKunW%33U1tXgbZ?0V;+#i_-m#jUqe`Z_=;-f2Ug2(c%kT2XVIz7qEnqkr#g#HbrzlKEIQR$bgHxHRAMT0dS#+wi=u~IXs -m`KPokgcQi%xYGo$4$))me0^v*=W3(W%a&Q=KJBlUt(JU{sGy>aj^ZHmS!Z_1L5yo77{IdTdgUO;wBf -h9>pcG_iJI^3BROE8nbqv%aBOTcoUfv%aBO`RW*_=oqII<*Q?yQk1W -baY~6=RK7*|7Uf&CMJigqI^-!jriNT^1gGDC>i%tv{ofs@SF<5 -kBu;|2KnM!{ueT|^XRQei0m8tYKf+`w8m8I3a)V-Iw_tNTKTHQ;ld#QUbb?>Eom-1c8cPZbce3$aw%6 -BW@t$erg-O6_>->rPN^4-dJE8nerH8v|6n-z`C%2r3(>PSZ&>1YRZv;#WY0e!@E<(Hy7x9Y5E)mhW3v -!+#NO{*>zSDiJj4iDN|6oOS3i>odcS6wWwx>#Iwb-3#4aMjh}s>{JumxHS=2UlGVuDTptqeQt?XA!H; -B37M6tU8NWbr!M4tamP2ta@?Nvt7>)Elw>iEp9E!OII~!s2Vd=jTx%O3^hrgQZ;6%N%MUZ_e!pr*U$vvJ+R<0-=xbK1v}dnbYavS{Q;WKacE?qx9IJN6RlDP=GmuqhAgj(mR_%_fcE`2A$f3 -ol#ihlqMHwK}s(h>Rt;+Xf2D1E8j9N@ul=ivSrbTuBs?J~4`Kvm=&M{V9JgvHTT6OWX>f&kD#nY;br& -SkEt1g~ao!qNBxmR^^uj=Gp)yci8lY3Ps_o`0rRh`_cI=NT7x=L49>FO$7mDJh1so$9em`|hQE_tL(5Y2Uq~7PTwa*}ST=c~xigs?O$Boz1H{n^$!L7HPm!!sOi*D)2X4R%fU@oc$=>9He -KOuy29IZg}3PnZ_^dtrfFbJ)4*C)Gc@O$hwDcCFB^7235zy -H-lqO6j_S-y`lU6uMSf*G1}Ha8~N5tIoZukt%yv*}KZ#Rrap3cg^kTn%mQ-($_lbT1Q>$sO#I^_3iHZ -c6WWdJL*^8?xEE^w0?)q^bW1=q18R~{T=%L4z19k@6pirXy}VG^u-xkgJZt5C=Z(2d`-=bpUQXYME6w -sQzyEoHbGOpeX6%losORB{8OEOs#a6AnyS@Qt)^-<)%mA-`_xRfIrYvO_Dl$?9y`@zm&(4>&6ch{EUk -s5Zno6Tmb%$eH(M&*QgN1wvs9d=LEzHZY-wz^G&WlrjBQ=?*}AN=by;Tz{B6-<)ne14*u2#Vw>sh0C6 -XQSw?m6li%X00rfkjX+?uz!HE(lk-saZ4&8>NxTk|%z=521x+uZ8jTbtZH^;TNlTVwt`aVw>-G5^t-z -@syPN2~j2{T{8*qkY*aUW-ACQH%PZlC`MMEoyUt8C{D*i&KkBiz-fiX;BGkV}3Nx{Aix}(ZuqjiRGuN -85L$oTX&jTrD^9!)6S2kogYm*Kbm%aH0}Io+WFC`lA}{4N2f}TPL&*;Dmgk;a&)TXXrB4eJoBS@=123 -)kLH;l%`-o`oOEF9FO(dDG0%SlI^b9t`lWS%PvdM@ax6KSr#WKQyyWij%W2d&AoC}ldo<$qMT69D1(pg*l<--uw2zW8?I^+maDpJ!&Obgy5)`X(Q*L)0R92|1NaB<58xl -bKY)J#{{a30`~&z0@DJc0z(0V01pf&B5&R?gNAQo}AHhF@e+2&s{t^5m_($-M;2*(1f`0=41pW#96Zj -|aPvD=xKY@P&{{;RC{1f;m@K4~Mz(0Y12LBBH8T>Q&XYkM9pTR$ae+K^y{u%r;_-F9X;Ge-igMR`40{ -#X33-}lCFW_Imzkq)Mf5a0N@kFio`^vwEeiiyv=vSd%g?<(KRp{5CUxR)P`Zeg+pkIT24f-|kZ{Xj+z -kz=P{|5dI{2TZ;@bBQ?!M}rl2mcQK9sE1^cku7vuVX1!byJCJnkRJ24rPyWL0RLshWN4}-p6XsEo;2e -HDBtME6TWDXEtvc*JFIpHM`}OdB2JBLRsUzj`2j-*~42-D8t?iz8Uss*q>p4hW#1#BR=mKzhL3xmSJC -keFgRv*jHd*fqjMcY5ds-^wcr=EkoY{ei(mrj6eDSKa58@#v>i$k&f|5$9M#5L$?h78StM0{uBHs_)q -Yk;6K5Cfy6U-9=Hf8X)<9eV0K-Yr8v^e>}Bj}ARL^w_Acjr(uhf -AjC5-vNKkwQy52P;YsmyiwNsPt2FifcvBWo0wmj@XrbVoR}||m@k-^FPOTfbjt~4Ja6%N73GF -<@}-v$3&@ZSahUGU!p|6TCkwZR|$yx`9Z{=DGN3;w -*|&kO#%&|a<${u}%^_~U(88~oAUF0{7`?d>|?f50E_+dANX!2f_h-p|GN^ML;W{{#L9{15nRE{uQaeC -sW*ujgZLdE-3(UiVAfGVZ7SmB;#y$NG-P`o=9gls(D;<%lxu%djuQz6|>^?F0V>dl&3ouy?`U1$(vqx -aEv;LAj#bPzL`6{tNsU_%HBZ;J?9tgZ~Eq4gMSaH~4Sx-{8N&e}n%9{|){d{5SY-@ZaFC?T4Q@KY3o; -7r%7;=+Pc~w8tLpu}6FC(H?uW#~$smM|r)-;QyuG59qUsa>r)-;Qy -uG59qUsa>r)-;QyuG59qUsa>r)-;QyuG59qUsa>r)-;QyuG59qUsa>r)-;QyuG59qUsa>r)-;QyuG59 -qUsa>r)-;QyuG59qUsa>r)-;QyuG59qUsa>r)-;QyuG59qUsa>r)-;QyuG59qUsa>r)-;QyuG59qUsa ->r)-;Q~hIos^nkF$TzP@ILF?bj?Fi>jQ1H?=GczYu^*>nLr%wzoQ^Fy9eZ**Hsy5e%IVma)3Gn7V`EO -o&YX^|IlZzy)BdG5v`<(M=~xfB<%V)cd7wN|RzBFZ(y?vjmhrrZ=V1#)!yXFkd%5L}GW5_dA$Dw(>8; -{?<$3k5rt|r?47~^V5AYx0kG&dgfFJy|qrN)otE0X;>Z_x^x{l9!*ZIO*_9(+1oiFTNQ`tNBz+io)V? -Cu~{iI|4q+|W0V|}D!eWYVOqhmdzb8}3```Yooc8xFZ?GwrsWzF{-`c;Ipb$#KM9m*c%fO14Rp`1}JC -|8sl${po_@F!QX?w2Y(O#9{fG{d+_(*@4?@LzXyL0{vQ0b{XR#>Be&d9*7>3L -uP`a#G2{0NSTel;b6&A%dI4s^V%hYCJ#}xu69@znflOc$SOhi!<=nn2T;0SmEfj}f -6-*I>Yg+L?F39PT9r#G;9#R0Hh1RMcRfLs>2=r@DjKqgRNseU==4KxCsz#uRQESkP~#R0Hcx0`zGNPv -#J>A0JgZd$q*0b06IbECr2Q5~=Fydv<5NFWi&yhkBW2{hiL^NK-W@=q48*t~*zH0sf)N2g)E6Bq<&Sf -^nf74@+X@N3up*a@gtcmjbyB#;PX0);>&&Ea~Q4dEw9QAP2! -%+`MJskCL)WcB^M?F0C@YKVj#a*6yXjtBSe|`Z+z!N~DbdZDo -I!HplKqgQKR054aCol+10*k;VfXY6RdPM3GsRvThFF-va>h{D!pb}^VI)Oo85?BN_0lGta5g`3U`ib- -t`lwT=Y2i%^Z(4ZM^36>X0s2^>j}`h@p^p{%SfP&<`dFcl75Z4Ak1-H9m3nZO7xtXKfk*%i${BphpvZ -$F4~je}^2h|J>1e#76QHJpnvTUkp{9eHGNL~dZSt9HwiwYL#_MO&9p2e2UxfUc{bYv1<03tiqtz*qO)|_KCxIuvi1!6@xR+JkQ2uaZwpkjlHO+IDvDU(l` -d&^#uX}R6}HU64{+Zb|;aFMJg8AokV(2WOowTSwwagk)1_k-w+d@!e% -_O8INqnBioP2_9L?Wh-@z6AUWjJUVvVIY4IM&Cv}<4cxDTe@ks@j+2 -mw4Ihk&m>6V#pndx+yK9>0+XAF2n5gWdD*LF)KB_WQsPwT)AFK4SN*}BAu}YV!bSduTbB!%sW5d?iY&GPAUmy|41PTEfwrCjE -a2%Ig9G6=hms>6n2_yoUKp{{GGy>q#S+|{a+gZ1rb=z6DT-3S1ATS9m0;pSj%kct|%s+KUm};>dGx+_2aIF -OC})Cz%%8^~Hv5v0+<$ffnDFjeS8EfE|Zp)5AXdP}9SH?y$)@Y;q2(;ZV~92}4Blc>zb@n?DHzB7sC8 -6DR~KfkvPc7z8GPMPL&+03!8>v^3JvNJ}FvjkGk<(nw1qEseA^(bB}4ORTxXnoF!XL@2pqi`)eq0Z$+ -h(5Ku{(~E#3;0Xi*-`q42P^ZgWpvw3}<^|9gWPB#`4ZyI2$*#eEZm^#l?A8XmwZU#}uv?pKP$nCc3wQ -$G+=JVNF$lOoCQt})i!lZP7w7~Afk}Y-nvvOc0qn#+j(Rxi;i!kB9*%lA>fxw|qaKcWIO^f3hoc^jdN -}IgsE4B-o_cud;i-qG9-ex5>fxz}ryibqcI=?7`u84-P+UaQJ0|!%rI=e%s&#vJGU*&mXXhp -6sG0yXeU-dY(YGfo#J!Y+3t(wI8WRq#luaMCuW#N2DH@lwA&U;}@s|u)~?lJ{O1t5&=H8@UexDE#Ex0 -6Bq<0fkj{w!1i)3VJ$9UUBD9{+e)^TY%AGTvaMuW$+q)~{>>HC10$T};Zhd}1W>mfz1ZSZ*5Xvw;#Ah -+RMz5D*5Xvw;#Ah+RMz5D*5Xvw1qOjhU=i2^;4s}$4@W&5^>Ea~Q4dEw9QAP2!%+`MJskCL)WcB^M?D -<%aMZ(74^KTj_3+fgQx8u)JoWI@!&47WJv{aB)WcH`Pdz;K@YEwvk3c;F^$64>P>(=80`&;w638WxOC -Xm(E|FX!xv)X$Y*6|)4N4>z_EDVJUHF-vH^91$tlLOEBK3&WBT|n@Jreau)FV-kL_HGqNYo=yk3>BZ^ -+?nsQIAAD67@*bBTyk)FV@mOg%F7$kZcK55_S3sNv#Q4HrLaxcFVe#Sa@Uez|b(RhzW -O&c|B)U;E=P6<0D>@=*?^iByoCG3>2Q$ofco$*I!{LvYI^iKMiVcjo;8=Dz-7w`lEfk+?`$OI~ZMxYZ -I1SWw+U=tvHNBWNR9qBvLcckw~-;q9|iwjT>N4Ab^9oagvb!6+w)|0I#IZtw)Ts*mWa`CKN&$`8|(`G -*1Wfxw|qaKcWIO^f3h -oc^jdU)#LsfVW?o_cud;i-qG9-ex5>fxz}ryibqcH%dIahbs7Ih4fqDe$5vWI?9)WrU>Jg|% -pdNvG1nLo}N1z^odPM3GsYj$9k$PZ2v+*;qZ-C7f1~eOA)p!GqKqsKT1@}T=^p1;O1irb#69@z%fkYq -^CP>(=80`&;gBT$b(Jp%Oz)FV)jKs_S$h}0uek4Q -Zt^@!9XQjbVIBK3&WBT|oVoIg^JNIfF;h}0udk3>BZ^+?nsQIAAD67@*bBT^~lsCQ;$qNGWE#RBU6t|Ju>yk)FV@mOg%F7$kd}yk3u~P^(fS%P>(`A3iT+|qfn1RJqqEa~Q4dEw9QAP2!%+`MJskCL -)WcB^M?D<%aMZ(555^yk@rPsl;TV57#vhLHhhzNV7=Jj%ACB>dWBlP5e>lb;j`4?M{NWgXIL052@rPs -l;TV57#vhLHhhzNV7=Jj%ACB>dWBlP5e>lb;j`4?M{NWgXIL052@rPsl;TV57#vhLHhhzNV7=Jj%ACB ->dWBlP5e>lb;j`4?M{NWgXIL052@rPsl;TV57#vhLHhhzNV7=Jj%ACB>dWBlP5e>lb;j`4?M{NWgXIL -052@rPsl;TV57#vhLHhhzNV7=Jj%ACB>dWBlP5e>lb;j`4?M{NWgXIL052@rPsl;TV57#vhLHhhzNV7 -=Jj%ACB>dWBlP5e>lb;j`4?M{NcWdKPvU8)T2_5Nk3l^K^%&G+P>(@92K5-!V^EJlJtp;-)MHYQNj)a@nABrZk4Zfy^_b -LSQjbYJCiR%qV^WVvJr?y?)MHVPMLibvSkz-tk3~He^;pznQIADE7WG)vV^NPqJvQ~&)MHbRO+7aC*w -kZFk4-%`_1M&7Q;$tOHuc!lV^fdMHNJoP)o*_PH$VRU?|=H;&%gf7Z-4mXAAkJgcYpESKYm>J{eS*1{ -XhP@_5GiJ_a`6a=a2n*ef;zLPrpLL1HMB48qMxO%lH~y=Rr^T8a?#`ynpdE>R!%ZuW&FpDx3_?3KxT` -!p-2W@Gy8PybRt7KWv;+yPetX)NW^XJGI-H-A?UxX17zjo!RZwZfABowcDB9PVIJPw^zHp+3nSCZ+3g -N+ne29?e=E3SG&F0?bU8?c6+tk9L=et-I4bL2emtx-9haRW?)bQgBci_!rkDZ@HBWSybV4ANB_VvC># -w=3TK0h!uB>qH87e~QKgC|RaB{>NtINpWKt!SDp?gLt>R<~CRH$*f=R15SrsR(;$(IwwL6*JN$pN%cU -HTtHMs*rh0X4)c4xCYtKHe`&T4lyyR+Jz&F-vrXR|x2-P!CeYIiZai`rex?xJ=Vv%9F>#q2I>cQKFE7 -WR(73|5aV=CMUR_VKm%5KCb*u&RO846JHkHIJ?8vDGZ8Dpj+Wr0Gl2?A}f9-K=(+Rzov7P3d&gyQ|*a -=yavi?U7v{*^N$DI^Fc{YDqVT=*n`KutS9%))WkF3WhZWLz{wO1D8p)RH|iCEp^Rh4zW}!3kq&rCh)ep+*+60>T)ajZFRY|F1NiqTi=~+!ft)%x4rXQ_1> -oUR=c;^eN@uMYwY8rv5d+x8q27Uv>?Z8kmD`L@fzfK3v#YB0$Z!;r?Q++6Xut~+gNU8xwpa6n0^`SM0sQ -|@v%gSPg;TKk}_eX!O(Xloy=wGZ0b2W#zvw)VkV`=G6Tu+~0kYagt&58B!XYwfT13w+GLqXycLHk -_ugJ@SaNMRTjDQE#-UH)_-yE$WRL^+t<&qei{aqTZ-cZ?vd4YSbGo>Wv!pMvHo*M!nIZ-l&co&5@%za -x_Pd>d4We-l$P;H2;q3-?5n8MeR1Xk45b+X1Bf#v6$Ue?XG5bRl6-1j#VwG_Qc(Avqv_S%3|cGF>-7sRZ{~kbdF7>YG$`a)Ulb}P3>-GcUQYD(2g2tM+>y02HMdA?Wlov>}GdYySv% -l)$VR~Yy2H8{*D@d$8L6O-w=n{J=E@Db`Q0CnB7C|9%lDYyRF}eL+u`Bw+8Xig7`Sp?qPOMwR?JlgTm -3^q;NL4C|nJ03U`Bt!qZ@Ft8A2UyJUfv39G$pTqf*NVV4QJR9NeGqV_w{`kko#PPBd}YQGb$--+7qMC -*5=_B+w~ov8gzw0vy8|JJI@`sQpf~ekW?b6RqEg+V4c`ccS(?@tEC5?Y8bHYIhWEO -cHfW5|7z^)NUK9L>;O`>%5{4RiX`5q7GG}4ONm3Rgw)=k`7gp4ONm3Rgw)=k`7gp4ONm3Rgw)=k`7gp -4ONm3Rgw)=k`7gp4ONm3Rgw)=k`7gp4ONm3Rgw)=k`7gp4ONm3Rgw)=k`7gp4ONm3Rgw)=k`7gp4ONm -3Rgw)=k`7gp4ONm3Rgw)=k`7gp4ONm3Rgw)=k`7gp4ONm3Rgw)=k`7gp4ONm3Rg%pnCY?=8Ps8>~Ra& --WGf+uqppx~jN!!?DZEVsuHch))+t_4nY?`XcG)<;zVzVu2OJkbsElx-7{kcezom`#0!8*#w@wq%}n3 -(g^BIhEVK(CI~?{1}+GD46%tH)L&n@_FE*FL(J|^NB*I*zm1rr)v|HucD1ovG$~Bdtwunu -d41-qrp18T+L5v7E5X|SVOs5o53;+s_l7b)xM!x-_UgayV)a~hUG0kc+(iUnO03_SewnTwz?F4YNzS& -ty!4Zbf&V|Ol8x#$2PpdL18mc!S-EUG;{l=F+`igU{SDHhu*Y9x5?XA)84ghvqbx}cFYnTG_|u -0mPdB$V7m4s-FlM#8VWzcb|+GGYu39Lmfb(X_Mle>Z*Wl99;pjZJye6c!qecT@WZm}%tE(mm9C3K-R4 -iaPLy<8_UXFp(~~_i=_Ac!yAH^@jgGo{Y&VbXI^OB#v0Vo?-8{Cd$9D7Bt{&UXW4ngs-NN#&9^1`hyE -fF_Jho50+B~*TwcFb7uCwLcJhrRHcJtV-&11I!YH0H~%r%F)<}lYBuch##;852b=9)uYbC_!mbQMNpwV|!% -umz~0qv&D7<9p^iL;!5WAR8`TYMdWQ|)hB~qhFJD6)d02} -)v_&4~$U`03dg#|w*PO4%z^Se|-6QN&*PP~>Q=M&^vrTojX>+DioozarH#_EMrmoyF -RS^bHNR}ded)B}vW{`7Lo9QM)kN6U>?RrvFPni{I!mz^xErrNw(IGrPbY#!6`fp7Oy&%_HlH=!K( -6gbUc47SU4KT9!sP;p(@-A#!Vi(kH)dbQlpM84<1_{JaH*(H_Hk2FU<>c}?ueCh~e_037o{%{`q*W}JW8e23-XUosTcFK)RSj{79L4|3AH -<-U02H~DJUBf;rQt|M}P{4<=nN2E*+vu9}Hj<<9fKXIRF_nElQwEG}Q$3HE)3)->XvNs$VlFqJA$c22 -~92bzBX9HRn8t~7QN@T*sr*m*wU$t9453c>|E!*`PnLB%9UEaWtU9XF;Hu?JAUWNR+XX@0IpgB+Wj=O -@i$YTCbV1Icg;#?`Jb1B|%+{bh)1Nbu|a>6=RI!Vc7G)NeMvpk-qYDA;vV(6<5~1~1O7Qo&!MrM -v+(Bq!#j%G`!dn~&u9MEOpd-4r2tjOoxR;+x5( -(yVd^rbjFQ-%7Wkb_vU~H!;!l9bddLm+{I^^>;0KJ;LUJkFFw)UJtBSa>1gi}`HRmC(>u{pacA~p1H* -P^KXx!|XZE9A-OlXC9)|7AezZT^nf-Y0wln*&jbU56y|IsBd$S)K8MZh3v6EqYvmaX-wm18+mtlLeAD -bDrH~ZnA+nfE^&almIH~e-7vmYB8b};(`>!R!yP-`D3e?2l%Dg#FR%M?15l*^ey}+uEOv_;5$FADbk0GW)ShVkfg7@xV@IKlVxNWcFjD#7<^E+Toqd -er%Q47I$sLr8}AZh{tv|`>|VMXR{yMC3ZIZv0q|mvmYBKb~gL5V`5wTxv^zpXR{yu__la&W7EVgW|*w#pW4Oj$G(YO%zm`z+uHM8V1F_D3+yjuKl-n2{rg6|Y7-Rq2>Ywqk6jhJn*G>Tv8&l%VShFI5m) -SL_E*?n&3=qib~XDEx9?{5H`w3I{s#M-*^hW;KU3%Ti03wYZpU-GJr}z-cDLtZ`-V-{I84?!o2|J==+ -;fQ4m-Q)hPZKe(+yiYb~pP6>>po;_yAq{t5f1**{_bH2W9qUuOS;{mbmfcxo@Rf5CnmXKjqL_A>i1Uf9d* -U$B3f{TP?+W%hH!%QpMD<7J!u-14%`e(rhMWvRpI>7(WvHR%mbMKz<;5{D2pB06B6g~a4?J`s2u3L77bK^-eya2C;H{jjy0el*+TUBPU= -N3HoiI?sX_CNk`kGjidR=ZEdeX89@;}&mAp7)r67sD&?W_Sla45vNI@D2Ptc!U1U>~FBYnf(p+H?zON -{$}D)Ym)XBy|1$d*>|bX8hW*>@->`q1{TuDrw)PNP_v -W*A&(DMV9J`OkAz!{~vamsOJ@Y^JSb;CYm6z>x@GIl`XPoGf3E*v)^gR-aF4kDbBiQ)Bip=|$Glc}J@cIRsN3^iPl8bR=DRXahR{7O#h>}5@KQZ*xfwoy -FT=5yv)g?Ru0Qr%%`6SJkQwm(kzfl3!A;<{V-UNwEeGPgZv|UX9=u0QA?lVs-Lb>MQWai3zmUIaiJDp -GA-m;|-=hZ}43EH*;Td=_yaI2Aci_Wt+OrJbz|Vv0yxc9D{YYcY$?Q+CKbidr_9wGH!Tx0SC)l6N{sj -Az*`Hv4GW#>^&t`vy{n_l#us@sq8TMzhKg0fP_Gj3i&HfDgv)Nx@=QEDDNAN1ftK#}JY1o{%p&Hy`u*@?{h63os3 -JA>I7U}rEpBlNWPv!MMf)_xYWpT*kGg7&jm`&rO_7HdBX+RtL`XF>aE`)ckH_FMZ|(0&$cKMUH=V(lm -L9B$d{NBdc<{VZrdi?yEx?Psy}v!MMf)_xYWpT*kGg7&jm`&rO_7HdBX+RtL`XF>Z}toMfeimy#3);`ptX>-GrI~IWc6PI~!_IDYcIehkw+{Qe+23J*H -~Ty6?`F?{J;UxlaQ|WVAGrUp`(r;$8K%#GKEw1G&}W!E6ZTBgXF{K8`b_Baj4$pH{xi+bC;WWb{TJ@P -?EVY)U#8E3KFjo3&}W%G3;I0sf%k~(pWThUq93eJLf$>9|6Ag5uevVTEpMtJM1h!s`pS7Hyyb9%mym=sQ0(Lsoo3jW9?zr{V% -;g1K@ka{moA9!s^e&{XOdaZ6DS>5^shp-{I_fhwDFgr+MPlaNNi0WoTYyw+)#0sPD_?ezvi3{WI@zkE -(-hYraS9bbIP+p&h$k{vq7yn*Vdl&Tym@KlwlRn1G)L_qdO@`*_^P+kHIl^TeO`sHx+_J$4SwCp_D-s -xi-SEKlGtPvDW^qw*(DtmlMhM)OBiy7v!t;l#(>#QV5;;yfoJPb7PI;)!Y}mV__x_g1a{`NlUy{=CQ3Lm~ET30Gn#?PG#3%QN{DcdH=fvQN@p*z7@gH6D`-mE}rNCDS@I+ -hNdm$LVK(-TDJ$%{f-aVoQKfktI`Kjdks&=CBM1M{M{zWKuf`4J}N5zM8FYvq{l{~8Ps92?ZRpU|dt< -P6Ar>^=@g*OoG1fKEv(dXuP!rKX?#J!q5ydf9k0{c0EB#M_MrE>%N8%TDd@Wi84cl--ye}M$b3$*71> -*Lv6{yq+#cx3O9=9=tGUVe3B-r2}J{mJ`CtEO4CLGbgV+Joi}bLv57u$AN2z!LY%wo1GPa+6+`i_3kZ -7cg@tz2tNE$zlUw=_21G8H;TyyFH53$orKrfvL#*z>0>W -Ti+CN~hjsybaj+0>pSXE%yoo&Vu_MyGTN#}A7>nNH7&!5}%@pBff3^`8jom{Tc -4P6XoLsoZ3BK+oWrcKq_Pdp32vk6>9%)swTOmuxwqL!Sp61%i88Wph_6#Hh -IZIHwuL&y3GO_Qm^TgwD_!gzJuhripk^nDI>;!mNHnzj(-0r(+w@S!|yujnT2^-;DT6O{%cFw{mml3^ -QAleB?VmZ(*lX*6)@<7!-**Nbb{^IkO8;XOv#XU8M33F%PgMS}+t7s?KAX^OQDmr@m==wb($E9(gPG)JH={e}Q5r8c#g$M~i%h)kkA}Jp1S0NB%NKf8KL^Z0HTx3FK8?Al -L~wgT-*JG_vI%>7X4~HqltU$Jn_`|B99fiofteZ?ZoDZ$Eg?V!%}y+=r5`PcV|pG7e*shE-fWd|}+)tpVGzmNE_4tvHuMxID^V)DeY6aSy>w -D(2O6YdEh?~x~x{YCobM6weZCvtpFU>0ZC9?Iciu1yZ#qh<+T!2V)lL_h5F0QcB%qS*<&=i|v>eIHd4 -`=HD{VoRf?*xe)E+AqHdf<=QjFrE{LURVFv*7ZJSocLI9c^~l`CB50L;G!3B&k5|<`TXUPb+A=^9sfN -1cvaVOwBE_p$IqwG-Te_puk{eZ<2*_uv)0wcDDET`_R~Wj7o?ld+rsc%66S{? -RY{+q6Np{_*8|MkcF$F31lA^cQ%x6pV@=8nG|i`MH_={Y)X^k}ee`Uib=lkJ#E+DF>gwqD3vvp(vT -)q8@?YrWGnzp(hHU;RH&O9KQH000080LuVbTRLGxW+ejv0BQ#S0384T0B~t=FJE?LZe(wAFKJ|MVJ~T -JbaG*CXJvCPaCv=HQIFa%41VvgusCUIwNk*dKLt#XII7 -QC}GVSvS=l^et#l4gdb@VP;R*_0=|m9S8wEH9Fmx~xCks^xuWZ3fS24haE8>BRTjXW0=)>O_ol&yOK_ -y1x{X6(3CksXd4#`=MHIF1?CV)3S6$SJu#U9HCWs5Gbc!k$!5&p1(`FmaN|5PnB~9E4rt@&eTqUIhMpg}=QY!L}=u -dz{pN^~8Gnb%SKh?I|TH?wUy&a{~lLjo&`E80jt8Czh{_g2UDc2bWg47c(qvl=brle74Dr=bCCYgDVX7YIAt-Nbmpts=|Qh!?vmy -6>VJLZT0fX8!@6aWAK2ms3fSz7=A0006200000000>P003}la4%nWWo~3|axZCQZecH -Qc`kH$aAjlz08mQ<1QY-O00;of09jjQV9FlEsQ>`ErUL*S0001RX>c!Jc4cm4Z*nhbWNu+Eb#!wyHe_ -LRVJ>iaw7qGMTt}8A_&t9GD{E#-)m@9f&7XVrRM)%}C}4WNO!Z)Z>OziETdK5{q|B=9MgtUsK~ftjWo -o5Jtt6!sNr{UVE+<$NV*&`3!ck3f}-*)T0_x@tdU3c7n$3wS1a>wm!e)5Yok38_ey$}EO-M8O=>tEh} ->m#@R;7`~5{O3o_CG8?=pXLB-# -zo;M;`je@1N`&uC@N-KmOCY{lf9?Fmnja=V{PVxCZINh;MOz};Qqh))w -ytQ)MO#m_^+j7D+Dg$@iMCp_HKMImeI+|k^c~j-WTd(>Qhgbz&Wu!V -Myfj_)t|BI&sdH~)t|BI&sg)t^a6T$x0wF;{I&w#dbgv_KhDd%Lxhq6^*O8xwNdG!=cZd|%k$(!2@;Y))h*Z~+dqbqYj@%a_&2{Ad5NWR?4}@l>a~ -*jwM7(_dIYhjC9tsgJpNB)l%jc00@$&guh8cshRT`(OG){L)Vt+D$~$arlH%4#$1_( -TnQ>yg36U?$W?#lsy}nppSkMKT=i$J`ZHJknX7@HtAU@ZfuE~^pR0l2Tk6mMH*AblW2_nz)tIWrOf_~ -@W3C!|sD5woap}E}I7GbnL5tcHPkuDSy#H}(E`(Iim#XJW)AMhIn0Gv)5b@m~x@D#6{?hw|( -fp(9(-5sKN`2#f!$jSyOkL$&!%todA@3W0@_Gn)@qF?|2zmK@@@5Em_wbV)A>`%r$<7e+4&o=fLdbiF -pLlsp`9<*`-!H1X@A=8w;p_M@@X4=3$Pa=~-U%VUJosdH2>G$_$(|7M1L2dsA>>EFC+~)kUm|?+UI5h -?2>U_^17Uv%VIUj`Aq<3rA%ubOeh6V8d=Nqy2p@)!nyxh@btr@&sly=zNgW9xNa|<^K~kZ9sUfLjAp} -Vs51}PV)ys>laFx|4E!rq8+9*TSC_~jKL)9ol)hKn@sQz!1ciJcm)hO?@QQEOl#;H->exr<2qf}t4`m -wGt?JKK^=GU4vsL}ss{U+Mf3~VWTh*Vf>d#j7XVmek3i))FFxAgd$E%M|Hxj1$JL-4^^66uQseX?-UU_`_1YxTGqmEY`pFTyHl0($-O5@Wlgh_cskjFEG -NqIz&$FB&J@`xah=LnPXh#-#_2$S-NAdi;_lk$ilkCzFP@`xah?Sx5rM3Bd;gh_eCkVoiPMX{7e40(i -3RTN8k#E?hmP(`tnM+|v{&TSM+dBl)M=Y(4hkr$;d(!L(rk~6va -}|i5K+UzagC?Lk&^lC4G0uBS#5j8u~_&3^GIsWcm?dQl^Rbj6MzBpD0mgB}%+k^l9k&M2RvlQQ|$KPe -ZpSN|c$267RR(`)~9ih%tP&jxc4hBSaWJ3zw)-sw{Ve_`+u!2vZh3LUiG?aFrgV%92NjEqpl_a{Cd&3 -!jl?Nu?%3aN#quEU6se2rYa@mL-(~9D#+;$g*T|d?T#z8CjN0YBB^BJ|oML$x)Ax!e?YzGCArIQ22~2 -OC~iL!U>;|Wyz#GGRT80OD5%!K^|mTGAWNP5fJjk- -hctO;KJjk-hxIxr~JjejZ_(9Z#JjejZI6{;|9^uM4lJSHnhdjbnb0p&mQ4V>KRgm$8D2F^qnaVgrlw& -FVF=1Y9eMS~Th8?2ZTa3@hV&u|dR)N0>PjgzE6`(75G| -B^6i(fi(_pY$~5lna@zf(zOALOw~lvoD3*hmc9H^}Uc!hQwmj_w(;_?ntQ*`(C@9Bwf|_v*~0=+C_ai -oB9wW>6pR`@^i8|GJFyhm`>DhGK>-xm`-GKWH==%d{2M5mM}~w>ghsGrvk1SDS$%ixWf1U=VV<9X*za -Yy|OL5k|$G&)xFgcw-zkjinFgcw-y??QdFlm>-t$#tzb1j{<#w)TfNRdf;J!G-IDJjftPT0Uz1Xumy7*@{~9BN-$~ -vq!D($IodQ*}@=@d)uNnNg+w{J!;XLVVEeABz)AOH%Wy_GCpe2oBYPAH)D|UCqhy_mZW?PQvO6}^kPZ -M#~|fTgkEnfN%J_Q&@oW6Ac%VEa#moJ= -g){upfkiI9_tCEFi^?LQGNePYS>$6)(Ugp(4|3tXxh$Y(}gY7>N`n<7Z`(v>EC&DE~EZP1TZ2 -yUH@e)h6KL*=>A~Y+pWcy>V{hz-=n3MNnw40x{V~}76QNm&CEFi^?LQHkl~}UvVsR@+U&G5=&A(1}T3cBbJ1D -48r{L&|iv^29FB!W8y^i00{GwAz>a%!aN3H{&~2XijkE6B`IXk=3j<%LX5=xlcf4nIRcTEf0CahM<7` -9lcf43YaW9&KN+&-v1H9-u;wR8Lrc~?F8v4$w->}%ww)wlV%ZihK4Pre!esrp1W6J4FEN&GCqqvqF0p -_nf1$i#W`2zITbO_!%jH{Ko20rhv=Y~SfHX+z0${ktjh7b;--F>IKdvi)lWDGf%E9dZ@g2-0SJ~J}j#wzEAvj|NJL{Jg0G*x>W80NO~(|?Iyhppcqi#%uj{+kqJ0+iWnz7ucS_eZbky;{8X4dnSePz6*?P -8D7JpuJCp6k*cG6c5+s3CQ -nLAyJrs{60m}0{ML^%;QY@kQNQzKBN@}rjow*!A^yiF9CHCieSa_d5}oUC5!9;u|72Qzy*T&yO@ljd0iewIfnGNrkH!P4@aN3Jw~FSuHk3$mqoRLOMd -1o_gZ!v%Lz%l*d4l|CIVxRY9%4&+In4j0@>t=#JbJnrdm!JSARH-R^HI^=N^$>Sz)#7>9)TOxVf1Wwr -LaDA6(Ck#C9>2SfFNFFxk9(SIpyY9r7BhzQaJzzjYat -byw=(HE0V~Y&uU)}1$sFtitng|6HMuFj?XU30kPTTJ$qFZ}w>7`zJG>NrOE#+Y+UvKWjY`1+p9y15X~ -)arOh_E2;DXQi$kUX=n*Z?~*0kf*;+b%zk%DSI69%@@ju-WraD|kDa6S`;x6+Om_?b}cr=XqBgh8&f< -Hdd^RGcZu=QAO7lXko*I}@tU6cqHCFyNJT{J=O9D$x{t@tKfqOgnyHoC#HF3fB0mFv6TheqekRI*}>3 - -Qy(A7!7JD&+%mNdrv3th<+?DLt>l}uyIztELT!9kx1UCA`Y{0m*H6ioD)aN(B5n17+WmV%Ey6E4xx81 -pY=3R78)CPhBynb4IOLM*_zXKG`S(@!V`5mBe&eA;Z$?rf -iwZl0ug?C{i#xXGx*UA`~e&owKCSWg -UtXDU@ePq01D?6e*NvNud`~9tg#qC50{vQ>2K(oh5}{N_ijxca{`-DdmCK+gVcRvNA=Au-jQu=rZn>q -AbN(Qs}ZUMT$hrv!u{vqGgJD6lY1H%W4#9rD6mrDp8yxC$&;RsuWcy&XIdssl3M&6)4V;V_GSGAVsF= -IdVxW#Sf$?5OI#2(Msh#rYH_^j@-~nv7IUMP|uMAD)UfNFq-Gc_|$R^fXh5b#;2BZ04(M?GCsAO1K=; -uk@1oIWeWcC92paxcvS#2g&xcNbCNo(xu -$SjOkul{Q_W$_KVJ1^#Gw_z@LvAFKDY6+D%kv>ClF1a=3|!^;kPpdZifjg!@_e`|%4CXc27dBDoLx!i -)D$%ME}GNzdE>sxs&(7#$l}(KDni?hZp=PLt8hR`H}{gi^w#BYUUv}{xV -%ZAUZUS_6sVuSmme3Xtk_+vbjGesn9e(@G;bcsvfHVWVJ>s=^HQb6<(2&1R{pMBr1WB0npwu|dWotq+ -)L%PnJf;SsXpsv@^`!Pcf0D_da|r}Pga=j_0^p2HL^Unq!{|BsdHt=b60kRdby8M{a1E8ukZr+3kolR -Z&P>${B?!bz~4}K1AMo_Tj1{rTuP=d%a-?LNszuw5bnzeQeRj9mXV--p(}7J0neEWVSzpYX}dcTsj)g`&uuVpE7(1BNWy?$jEjRDkty|U{Kh07AOB7aw&;ZtDEbLH&=SKbbLx=7@=%U -XK{);d?-KBRt2`CElN-KR*@>?q{sz#>svqL6j;id4a7CtUNMWqO+s}*MG -uGhk7ZeU_@b)%^%f8E|`fl1|TZJ=K3S(2=-mTtJ&P#c~N|DR)B(ok0X>|%|bqaY5MIo(DA&-|Rq}3^; -)hVRaDWugYq}3@(`E4?=T1cx?NUKv+>X?f{TAiYjzbmayQOW+5>zAUI{UG&!A@zSD^?%Vyxyc|v*->k -=Qre1A+KN)zin62hRw->oDQ!h5ZAB?>*n@1~SGxs*D&lsdVTI=P -e;Gs{eQH>L5Q_q={(4;`FgJm@`dC|tjRlyFz}uGD;`+{jq=q|~J5E2ZWuC1X+clqD^tzg^1Wbfr8nr< -D0nr9AVd?8}iOPrWH+1-G)0vsfk&m4zIUvNBp($oG{e!<16{mZiM5Ql3gF6ERBZ^OrKAqLkXVl-jqH+ -P9SJgHo;!N~wKI$vjjY^$wNPl2u3bStYe(6{#^yA ->d8v#$x3pjRim&H75_1Xq%6sHFC*r1q<%_N%1dS4qFGlD=FeeYr| -%lS=w>m1GbqsZA=WO)9BPDydB>nb1_#a_q>Yp{iEvfJ*v(mGt{6sd=hajt7~oq&bM05w_27ssO6fhmYzv1J(F6Vu2=Wvw@D7XmZ_*6T{Xube}5rSxQ7DKD|E<#>=~DQh|3YFUu7Zlv7g9IE9Us^!r@wTvk>a_!U@F6!`8R}O{o- -A0}*-6X2_8d>+LkYKAm2g}o`1kKM@R^#;|Ft{i>p$`R`6 -={mR5UeGEWMbgNb)ySFENK4j8OV-F)*vQo6Msnee^oyHXS>DEQ;YW5~ITB_%HHHgc!o%E@Mn(!7IWt> -ng<9mKU6~8h*jkyC)*@-*%CU7GFqINedX -Bm>MnVLm|5l@ds*Ejd#($@_So7h$HUZ-R-WM8c9o;h%B^#4Ze+9n%CT^ow(>;f7D4glyTaggi?H~W;| -H!B57ReV1jd(e_Y>k0L^h8fRCwRw`X^-X;0ywNm!vpa)ato%{Fg$0ec4a)Y~EIq5-qCUTsgi?Jwc9P> -|DBX{B?1cBp=$s^#3k(mmI^`D|F@fZgrO&!=(|_(_##l`W!tuh9!;BR`Oc852LM>X^c8j*F_yk=tLcC -N|}G3>P2^C?u=`$aTAqu?#M+>)X^o%JIDYZ%uuQBj;>}L$+$;!gClEV>FGzYboHZHuJ5B*7Ljw!%4WJ -|S;sui^qXY*P2?&+isdptisiZEu37)KjM_x8tO^#zGQJYUGQbjDD+MX8)_qb+S;o?xan1ToYAtCWZD^AjYCC}NPc4y>9>|FlWUgqEXkERiV}SyP9hIwkFM1#|A*g -|Jln2I@rEp$SXshKm_LMN46tD5sYN@pb{Mh|VFlPpYaoxv79ORtraDb<>|D;NE)S)aEhN&Bd)JGI+NC -6Mbk$@S35^(mCOOh|}wnW+}#nOclRxmI4e^z&S^KC3C0Q!me@63BD?CVCL(l2UZd@+ga3Cgeo9WZqn} -oFKV8P}nugt9x3Oy^fS+PoGrVTTZ+Q3#Gbd`8mCgDytgxWd5aV*6$tbt8+PGdp9_AONNi4o<7s2r&UZ -}_h3JghN!RKyB{lZB2e^dX3^>A70v~ua$62W9 -UnTgtHKrxD(vJy?__=V5*_JP>6Ju0{#o$Xqo{LEm6j-sl-NDOA^FvSXmXa+NMC{=|D -!h;z)sHZ|nn!(KTjOIY;Xu}G{X$CVdG73+0hR)clP^M-uLq}Fss3ts^p~DQT{z@z23f#(Ih7M4yP+?{ -;6V~%{XbY4MxvB~y-`5y5@}=Xhs=~-O+$rc#BVQhjg^_Q#ThO6KzH|guRT%k(I|dzUv#7hC2uyYUE3YW>uw;FRiSq(#V%qR#j=_ODn6YH1 -egDRaF}KvTiJmd}(D>l}5fFGHT>YC(~A?k?$c!jeO~#uBtThrBiCFQhF{`X+%foa#f`f9b19Yh>lL^s -x%o=X?m{G{#GT^h$0=Kk7{g4zLhcmp`XyCtFbTnCyYsts>TsJw=?F)X_$r)P64luGJb0fRz8ZU<2k(=^S7Y}x&j -5rRF -wZmx7f61I&SQNh5@xO$Hu(^Z|3>cia?Nno4v{`>b -XcQ7C9}aH8V!xCW#*bOJr9wn8(GfGHNyZO3JaP!R3B-Ll4gTLVNElK>QlxeBPSXbjjU_tnvI-@QETKz -IM;0CL>8iv)y-V9krTO8jVy5HnvI;uF>YjyGuLe7L=JN!%bdApBPUYZjjVL$nvI-DCpS7g+Qddqd>*{ -4cjlUnocKui>5^yD-t#IPpyD2AMgX5TFYBqfW-|i#$az^(%{7}5z^Bg3s%oy;i~v4(UKUn!&1MAf+4H -itnrk*AfRCToG)9w{5kS_uk!J -Wm-;|krP`!SxwC~8#%G%%e0&fMR1rcpHBE{G9xFpd~z#x)O3xU*z)P5%%*GP#FkGUJ>!~krUe|c>s-THgaP7B+sC6%|=dapEL*1bd8+YKFO14T(glA+b4M#P1NK@PHdlKQoU<7a$@@=kEC(UMow -&>7uzl*?_>k&o70y&~55x#K!j`Y6M*xoX8608DClB_C8l3DiIKq}sR;qN(7 -y(Dv^2steQDagqjz5SDY)p28FbA?Ug=+@?$aY7bjT1E{ -VS%=XIaK$O$y7MPc1P|pi5iolaD?rS++i3sCO6>-+a0;p&^4RU%yvf}mE)StXlA=3Ps?%5W;CmbW?xn{`eC|ie0%c+8#juW?!ZREuENuJ9SHI0!I+b4N6k83t^V*4aZ=elMi -C$>-WfF9RuzRam_|fY@cKy@2F{woV2B8Ov+>|$%eM{^X|Y;k7yV2;eAsX#E83k&r}Eza#8%#oFBO)zPWvJqOG-94BiW7+DRJF -VUi)|zO|97QeL$n*|bqt*m%<|uC2My7YjD77YbGe?QvZDe|fY*QP_jTEik%+s1^#2lqQx3T#`(&McOO -3YE}a~qp4Bz4}J*u)&AKDV*?LelE32~o^Z7G@h8`IB;Q^)98>ko|L%QP3u)8_11s4d*{c2?uRrx`A|B -YpDM@N;zl~(+#BBS`z`7`;bx74W!{(z5S|9^)xUP`!QA$NCO-kV6*=_@eI%8_MA2|-^5c}mKKWI%zO;bX}ur5Mb6sb82Ov6-tyHNe)ky9_^$b4p7D -m}J;pP>Yd)N3ykUEf@r>`9@8=nBNZ(^TXwu -NgYR>VEX(<4{zT1^#k4>I&9F-I3sam`*}r)k~R=8~>zb1$gV;i;(()Z^69x^rt!k?Ultxi~jN;qVLR) -+rpjFy>|y4i|B5y~2?kJ2$Ivc#Crz6pr+{xsAZ59O-x!&b4t*?%kwk`#0&?;wC*?-lS)%oAhjblb&sE -(zC7XS&tjKcFlB<3EI-`~5*_xE -q|{l!hbzr4x!S2y|o`X=As+~oV)mA_wa4{r^Her%t*4TgNbg1uBv?NP8SB+%+Y8Lio>cl<;3YY^daUV -50#ddGR`VQBSZ;@%RU!CGc|EX9wSRx_CWRZpzAv7Q)9g=uq -+kg4T?K2w%9@d`8G`*-*)BuCQ%OFDdcg``($Mf!m(QV)E4!2+<8=>Mx&DS1bIY{6$9XwdS>Xx+=^dDD -4Y@FnQp*wwtt?jSn$fKvi{HY0`IuL1hBiOOOXeE=%5Hb9*KM`w&u2w{){VuvKWo#U&x-!k^>o?>{mJv -b(JK){V1LXDzs7VNFZ>LTeT;Smx6MlH2*YN7%qzRb^xr1Ye@$a??!U(O-~UILtUv6UjSiU;Q>68Md-^ -d2dmVCmvx2n_X{QGB^vYlr&QEd=G}=j-9Ybaxqn#6NyV7dXP};}Bqf8u%!N3bd5Zk(M&1hg*w{|Wus> --&+*NQ@3yNR)-R}HH>x#p#pte0+KNG)*Vj%1Zxv-dvd*M_Rfp^9X1%{jg@Bn%y@R{wD)Rw~prt3ZZpR -&1ARZg1RGk~ekD3BK@v`^y)k(-+8VRE)+ocX8_n?k``EPhTLf(fzf|yDop~pZp1fD!$^H&F~BPMu)2S -O591kH2?N*8SHtG-(s0oz4nUYPDSX%oh+x`Rvj*Dbh_phJ=ndR?wL)mJOs7Y4 -K!&Lo#*X)N}%9vk|WsLdp)@96hY|fb1q&>#G9_=$$T<`Ax$e356^P9rpkwg9H{Ts%-3cZCfKZ5_9F~8 -CZ3ogW+-UXC0$G9WQ0LQw)e)ScTmL%TF7MC?D7{yNjj-ZQKQ8`l -pFnSXT131GXJ)wvX4gODw!fx994+s&{T>aAEI8DUII2KL;ZLQYnM38k2jhZ5fU0 -c4WnCeggvrkA0upqrAZS-YK`cc{rsf)5@EmV{oI*yyQ^#VJwkIVLW0CUGv@mvEREqXRp%v+unYBJ(gR -WK72P9uaA#mI>LcP-m3HKsy_yU6x;sodv=O$8o(>Z-9jH4H`5EFdQbR;Y%OE`(At~dNa4qgYuh_ymKo -0XnF3;=u1_JTpmW)e;+EzSI66w}OSZc70u*%?(iK2q@50(mj21x2iA`i7k~odk`5AF4lTclADX)gUC!#5Z -z9*t7g}Ns}r%?7pG^Nn>WLi_`dNP|SR6Y45DHJ`KyA)cUe3cYRp8b>*T0Rmz10ycF#t+d*^o<{$k?0& -heIwC3Fx=cdQ1lV^4>Wzm9Ry7uaSuV!$J|59ivAj-77=fj=AI?EB^Ib>32sRQ>goMkQh|DY|CUUkp5V -WwD^SlZ-jWNn^k`3@o?X19FHrD{QAfZHVAWeKV9{N@4q(w=y$G=Auw}yddsy_?vX=iI7G2geVitYYgP -bGLX*~!!61~=gq+`)*JsDC<`;f$~d{ZBi_=!Lvh}#4TIo!@@0WsVmP)OlUfkFs(2^2E;sX!ruy9Ejf{ -F6W-fO`ar{=ZkCm}&P36f^C9fnugTAW+P-2L%fG{2kE>Ot#Ujz#I -{;NPC-(Lt6^8N1u^@foG5h;@7hgBU2>D5f_mnHIDeF6q#|v)21j?Njzc%u&T6vd=nBWxq+ldKIk!43raEc5&3R9%WucI -hUiu5|-{Zk~@(V{R#Ze3Wo-C=%Y(8?{vUiV)x=0^q*ylH*o&|O-4`SFaYZycgYk%`AM2Ake_#?)SZJY -%r$oo7t#<;OFozA+=1);DGZ)B45?cZwuF9@p3a&*K`0Zu7Xt9(W$twU-}{Yixt(aou?N@wmoLcplft{ -Nr(r&G0;~n;RAlcbJzSkL$+EkH3pOD8jvI*H&mwV#lq7$s9^?JFG1)hS{zTErEtR`W*_gW!IeD{Yna2yKx?hjLA9SZsG4^yEX3 -i<92b7LI}`R)%Z;>1;KLydgHZOaZd@_n9BBj0fMvO|r0 -!!zg{YUE2zKI91n-jqs=ohe~e0>Y70jZkBSW+6(QPaSR -1dI#8r^fQ18b3HGm+c**Lau{o1;{s6DIWc5(jY~;iahp -ZkNo1;qS53rSz)kEV71n&GnwokHpsB4Ct4zhic)k9q~))>RZ_9cJaU%=l#f{DuyjI1Zuhx!vgBnDO?~xg4^Z;m3L|}59)KjCUUki@@>LFSNx8Ufn9 -wIWN1xJ7N5Uu4~aP(K{v?zyqqjHsw=5eTJ39iyn84m3NB|fwTM}PH5SaTva#b>V`A)eh7pB=6j@9Ztj -aF&)9rugh|mOOh?e0I2j_-<2tcDR99OH+JyxIz4~A1lL+{Jr4lha1VwZo$zHH?j1=(GMRd%i4mYA3i~ -DbqkJu_#`QXHiPM?ZXuTwhZ@clcCz&S-4P=MJCpOSWmhbQO;Ji%ntP;bxXT*#F -^Xo=0H+hg&G>U<&gNw~(jMg8d)9OkP+E_J6pI6h;g7fA}VkI0rr2mgV -y3L*#F^9ay?qG|HHjRa<*Xqhwt)e2m3#Kk4GKY|KUFJ(M?(6;l8j)X>7_84>>Mq$`TLvlLuzX5)b#&g -4~oP9)83#6de8VBi2{Q*B*Yvzl_~-!;dH~*y0?;;bDr-w%9f|Jj^;48|Q{}N|Qq&-=h?}ZLxQ5cvQR< -6+#}w~%Xe`v&wW&kUfe6b7xI0VM^0+IK$b5yBQCy2QPLJ$5f|TM&47)Fi$a@W7vkbR@y -n1eXIHe?fjD}Utp@fVjy6%8ti|rbkr?gRdpO!ejGQT(J=(-t1^W(1kMVfMuEWt|?2BU0;pj2ZlU}}~$ -7#{sV!z>NGi556TH2${L`$^@mX3bKo(#61jD8iCr*^29@90^IWVZ;Kj-F+^kD%%3S)L2nhBA7VXBjr4 -jGpHai_q%md5SWcBH^Q#So(;tj$R^a)f5RIZDkvZK5MT5@_tunGq$#w|kAY0x>@tPf#lE#~OmLYIG-o*L}J1fVJ@%%;ndi)rBs;HPWevC&f! -r0@Nh=nnA#>X$Qr9(7({5r2SP-A@jI`;;G+T+(*Mkt^(-pO+xvF-6Lp4}*-G=7Um9YWmWxA~n>M`_Ht -kEW{lnDgpPP4O}3Y?+GUq*oOpUm>yM`(Pqk!6Is;ggL#CJ`Z@Y!uJ+@|`@#V*`Qm$@4s}P%n -J)BKtXrmru5`)j+NA$yT;j2%Aq{W={rn!YA9s7d1dB`_0q{pKNElk0APFJ0%9RrI+vIRi2@UrB7a^P` -@c6J`o-TLh6&(*wU4d?;G43h^$ZE=y3~`R?JhB=$~C_VC&n!S~4?wkg=YA(`j;$-wL7l2_5tux7h785xfc_@)J5# -H#T*}Cv;Y7Y^sV+=+x2JR284lQJk?UD?Xv)Ghir8T)_UxMd7yK1tDzuxJvbh -W7%qG`r=8BsPQy*$8hZ?3nTsb=weB4uEe8{1td|*0Nn|s>FUxgzl-2-(8%dl?Jk%@iDwX%@cB)G6H)S -pQQyS*J8DwL!nbNF1dAO8#MHZ80RhjF5-M<$)t+;?zcCM435YgHcU+pbkbq@RlteKok}!TQi&NFs30g -Y`6SOv$#vhHwRwn4)ci^HrNod_=&0EHrNnas>Bp=8*B)@35VjGlno(^keC{8gAJiAN=%Wr!Fn2$QEKEH`k@Xr@(sBS|+7eS|Zm>S|ofA`OZm>So`blI?D_kGiF^59F>qDm2p^) -$TFlymY$aj5cdmRe-t|z4(fqNdza!-PL9?XXRZ(_>e4Q9jjYhsGv4Q9g?TVg8U4JgK*nEH2v*^u>2O! -d3LY#2FAOzpeDZ0JiSrt;ljHuTvOQ}=E#%UZ=8*+`N9#MHbS%(7N7CpXTrRsr`sm<_pkT*%!XcSV -rtzDW?65Ta~x+`Z*+}(S#NZWd|7XFjeJ>em<>C#tT(zUNhL9L?gs0_fJ9;n-3``<OfIx#0x -Ze(?0PNt+Kabil{4L0(`H78SU3?0zK)VdpN3`1y%sdhKmNG9LZyBn+z-NeLHyc=u`>5|0Myc=u`7e$G -wdN)`v?lp=zToNay?A>5}=$w2DDUZZdsvFF*b7YRaoaNDO4!xY^@obK~+{h}y9C*2r-`N~@Im_eO9CkSyF4P@r6f -=)@bI|1mcAd;&lpDP3H09bSt8fTrV#@gqHn83>M_g`T!)p$>+`xv{9B;XSb)7lfas$snbF}3KHrldSf -1+y%x@MEy$Kwn{GN<4<6v=&u*wF%!JbXE%njDJczC-MWfk?i3i034TOmn0yPhdWq9fk+;1V>69~8mB8JkzS=mjkcn#ygCL&aEn`Rg*B8M9vOU{z+;2Y2s|kpJ)3?GZTS|NH0G3mo$Q{VSq>ak$kjzu)(^z#;$NAAVopkpI8Wi2 -jWr|9^i{jGqYd|M%xq|3m)2e^d1@AHrYUCk9_!o8K3M -H-_XV8-?<~`D(Wex4p`E0{Jh#uv6fW|Ke-!2s}3LC*+?%{-Yho1rGTSpMFZ&b|0j_D)uV3;9P%GNBPMSG`49K*5;){PdRhHGh -bhiV-=3obdSZ6W??;L2cigZSWnr3I4&Zrr|dxAxp7!e*?~AHBCMzEK;N}-SWnr3-UpAvddd#;E_fW)Q -+A;Dz>}~>n*+TAo`e;r9f)HgL=SoQI}ttPz3)WykaxZl?jb{wagZW56d4B^VndN}upl-R8HWSnu3=J$ -JdZ6?#&g0lWwb7fyDL%Lad#z(J2p%ili0Cg%9zBC4O7M>c5Ik3Cb46~lrf1NgDG>KF{56*&MzhuOquh -F8TImYJ~5+SzRn|N)XVpGlwR#nFJI>kGwS8*d|^hte4QuEsF$zvgBkVmbzU%|UcSx;X4K2qdBBW%`8x -lX(F$*M=(~eN9mXKboO}HH?cx`_lz+dKR?xBBTbsln$^^X;gD4a9M%**h2u~vrWrE&_L6kXvmr)ozjt -0a%LyfT708z&K4;>0>*C?J7gZgqlEu+Y;-Vx?sIus7g(-iBAL47%ImQmzTQjr{o!lBv85(M?-{8&aoeZ9pK2-Mf; -*)V(0pqS3<`g)2duGs^9kjDmcC?~J6kwgxq^I93j?zIg>=EcaNT--~+h`2D -+Cef2Iawr8o85d^$vVlbo0KP?4BZnGGX@=36u%Dr^ug7y?Z+C{jb7-qVBPN(4_46A}=###} -{dtDLcN%%1qhuMN($UjxTaDQ_=T+xTK=*{b)%=-}}LmioW+_CF7+5(lH}tpOB9kOBka?!Z@x+!Z@)5lWF{%Kf(-GN@zcRVCeFY7zPqKyvpvcC6`uwHx;BR6wl -9R-qO#f3M$wy%<%FDRg) -HGcxj9(_9AgHMjiWw8Mf=FWWO-OmK(L~7iM_1fNJ&&Gwczfp8diMFPu@)eqn|e&Zud>FvE*ql!IEBp| -qH|fqZA!BSclGg&AJMG?4EM1)<{x@||J7yfyOWwO(uF%N}rR( -Gasb}+eydh@EvXYSvLOcDtpb0`@k6~#Cq -zoz3!*vn;VO;XZbSGUG3C02Ze4)A~Uq!U0{AJi6VQ!l8E_bA~T}go9w*7J4eaXPr0vKD=$l@evRg75@ -pjDxFNUPpDdTC8(cZ_`n9>E26L2$7r%D{8AWDv-SE+Ux@P&gUQE*~omDuy)xComy6-J@_`12}!6~}5o -Vs%Faxal>O(HXf_b*Q?o-dIZzz3J5Ox8b%nyG8tvxt95A~SkFSbFW{PoT@f{?TzEwGsMwj!HyH5dz}T5B%R`8 -K#RN5(zPC^}?wKmMQvI5-r%x|mvqyfuF|5xn*(qlf^z4q?sVn7JN;*^5aAzjQoBEaf1>;aH8Q<%Ao-I -_8OkMfr4QzU1M$z+2Kjnt*y0_$`Gy8)x(tw(6Wiu}>b;7!jZ(!LIGp@FZMMskK>;+49fgI)}>D$LGsR -9;sQcV4hG{H$Zb&VTZ;ih(XjTuxgEnh=Rp(HVbYWsCpMD)2CRNIzcTU=v?)61%>(~f=C(t~&mq|wxmN -I|4=c#U|L8B4ERSCy>f(=@nDJub7?EDc>+B_(Nh>YCkRm7S#d)HMf}Rqk8z<;_5P^}6*Dm-X#s!oNz3 -=?kvw4e^a;IK8fXp0t|2;0CHSt?gwu^ea*`oOay6UrpQa8vdQkjHq4LTz9_4jHq|6DRcQ#@ziCs1WPh -AqINGSL*b`o>C`oI%L+<7%Z#W!$}VT!sSB2-CcDs1oE$L=yIcD3-7vIf;b9)k@ -fpmrTNboOC7Zhe)nqWKZ-Y><30)!Z?;&b{jG%BH3@qp@?M1A&VlCJ%>DsNOm1EDI(c-$fby6=b5wDb? -AL;({l8<%&dC|wZ|GeyD-G5&AvFtzOR>ZRZkX`Z92g74X9qOfjy8ylDZw(KmbfA~}Pi_s5z -;vjW`UAvDCrGV$Fg)PYfnMIXho^ixum_NWGzk(bZoiMBGzk(b?iOF;r93=*aB&kQd?ZM%SbTA>Vrp`; -GvO^0?VMiht-@Kr?$M}x_~Pb8I%30Jfbj2Y?+oX*T%Fr0;_V6YIK)Un8pr(t^uy@XoIw3Jx=(<99Ni; -8KaB1PkI8kQA4c~H&<~@#1c)Q4Zw-%`b*LXk_ej`}qX)yIbRFo&(ftCf&`ScXfl`l0`&5|he2 -d-{4_ku)q!5B_b3zxNh{Qg^)7{axjv%M3?Iu~jO2~`1PBEoKrh*Q1n32OC#x1DZrm@x60a*zoJ%Y~FV -@=x=%sp(09*VlD?21@+@?@3R|Up+sz5PYg%(B5#%&7qVpU)lj}@pF>wWW^o(m67c6e_J7w_$*E3_BVH -ttcV7w>%v#ewn)_40k_Rtm8uNZW9z{;oE-g1-#;8xQjRNZ+_$fJQG16p|L87xaT{!I8glzW}>0!&O^CLnL(VF7w+KFsqSSrh*vdcaHb=Wfi3ck%+gRpO1wg5r$Vq%m1gfTtb(RXAkWJ=o+|srNGZR -SGuwRSGuwRSGuwRSGuwRSGuwRSLHGRSLKHRSNgxD2gTShf<__&X1-@cby+lk?uP`pd#6M$kd2r?;%$s -lHG@FjY#$%@-<@Ff5_N~W&a^(BbNP#tc_UqAM!S0*?-8~h-LpFcO%yQ2XapLpO=5E`wt|I?mz5n(*1| -6O}hWQ{O{rwW7<*00*Z8b!56C2B9%2Hid34}B9(`A72V}P8(rN+XyUtKZwP+W;J@<jhw>pkVCbcx<*c6AjqLw -PF*9XFlOmcEhmta-)&_TCQp-h&FJ#2tit4J@~#GE*DgF|(fC#K88j5derE>BFChv7?y>Ml=Amxu8#hw3g*OqYjgO%Byvo|rBV^NJj* -yWIHM+sR1f+Rx67pS_)oRIdH(-1yns$w=ke&(4jXy`79yuKn!X_}SaZNafnk&W)eFos3kj{p{TM+1tr -T<=W5Aji0@pj8v}u?A-X-bZ&)1wVaHfy`4-+uKn!X_}SaZl;qma&W)eFolHrt{p{TM+1tsKuBxn=f>TCg>|%c_jBX!zrs3NyZbrZ{qJ6 -39W9TvaLt$rudt5p>$zW;E@vHG=q@izm$Qy8be9)qyz+=B^mr|doLEN}T26(L6YJRkjJ0o@1!W@ZGCy6DmE&DkG<`q -KHF%yy(lYBAWwI8STSx(~(1SyiJ(H?NC)q&^5zFc$HX-ywWbha;g9m6Se6>HVw7z!&>EBxMtmXwJI9z -n)T>LME<*1*@D!1bfcERcdxPqsrBeyj>YqPA-WjmhB(a6<}hQ?VSbE+2`3KA@O~dM+LiBzFyW`Hq7ft -fjMYjjhd?iuA+}y3xv5?vyQi1xt8>YF^?Lg1N}Va@nx*^F(}L~u)s_Bj-6Q>4_FunM6`G4ebCN|OL#u_Bdy&fbE;4;}Cf~crq^nyjD`UR8mUgFT<$IU%EIQYGgR -7;gC_AzT%Siurq<=e>eNx6ZxLWG|GS+XR=WuB_31)pu?8EtP4>=MgsEL!bwl1*==exb(e#fHh>bJq>l -P`(LD!Wo=m0cOSFU@H)o5P({4n>XB&Ez_k=Cqm3VYvr~qDJaw^1(}U+RWy#6o^AnBXu*m;-xulW;40M -r8#Y8GmFZcHnTY_BjQlhNZm}1d1+3Y*-Y+mX-=Ej%yKrT&1@z=yEF{KW};F`bK1=2@EjY5;&su9xPtnayP6OLN-HW}a;3w3*G}o?eF<`I7 -TqnloWG6Ju1GGhsH94_}%yVKx(SRGKqkHj^t~nloWGhudcyYUE2EeF;Wluq7;#>QKmc3pw^Bh>5`#^0 -!Ou!x(HK|Gq>FbFd}cgY8hrcMDNJCAf{jmaqtvLm}TSv_>pJaSXPQpI>4t#$XEt6G|{0gDvFlm)MLk* -g`Rd5`@QK3sFoZ0-S>_tdYTc47Sk9rNn-W!4`@zl!cLRxP8o_M!vM@C<`NBwk?H`FL}OYVdP5-wX!hs -rLaR;82LWWsF5$NOUlB?H!OPPP$OSjo|J`=Z+N7oLydfi=PC;$Ut&U)R!Q9M2LoaSCG@!*3JJa$ -dUFoNQSIwO%jr-^@a@pMaj2Kz;`^aBaHyByXeP8X4n>*oEg{A2P_Ku^PlvvqL%n<_bWE*7A>X$`f83# -7zE_WkFM@oZ3VmvaLcV)Kf6bwg?o>r{2=sS9EvjTABURNp_rjBgkF_Hy?n -=pN7{(i*T3{|H_Z-i?IhoX%8PS%sCG&TA#^yVCjGVZkQaj2K? -2Zr|Sw<)THQLPHCbHy4&xO3BL!lX-WGjhex -zU>;)8s9E$n-oVXYAJrb^ -`9SYBFmd6IT|KUdNNpSy*Z?F`={f`dvJOcMWevGwfZssq~GI0NsjjS1Rq3~4Eg?&*#L0=BdR>v_QCy+ --{koV?qBJ4bAZ-SmJzuB;m52u!2K^i#r7B6|Khvc&*1(?`*`ky`yUC80^I-Aqr9Rw2WY*_o{Tv_OL#K -o0Id(%(t-P5T+2Qoxc?ENpwlaYRX9`DJr|tJGijMXa;_c`eT9)TBl;Pd;XU!r&mJiB -&2aJzRI3VWjHtX+)(&w@d(d-RDK#gE#3;6;W?HdYCnx$XKu4HD{HpwNG6?#6#>dW4QJU$Lj9-3P2A6@ -|1^1tX9@~H4cGEJqXN|Meg0nLn=XFD_6QZA#&ZINXD&1YDnTvIGLM8xP>Zjzbw(*DwMs=TsL58I`zQu -A+{79f)u4t?vuBQ4P{aMaPC+H8#TS{eK_RHc?L50tt!eR%FwW^vc;*MibK$u?&EJbsO`{jsQ$?kw(J{ -8aDAY85lY0Yon#OOk21c2tiC}C{q-pZ9cq{1VCt1g$K-1z8w)-g0G<-UY>N^zkcO%bGlx7B$jhT#-7xNV|p{3jO>M&$8CUM0i}G6w_$8xYx^fw1?O1NU4F -pZwzRSZfnjn=WwK>?=mULgg!YGK;9t#>1PjE -hhc5_J<#4wJ!qLOm$)PJ`BC>0S%wF<+H@yZ}g_C97bJa!=*K8+3^5jES5(pzt=;SwtBFwmzXs`ldsLcJMpAigP7p8SO6id=s@mxbjl3Q9JjJ&v!!sZ3S$cxYM>_-@RksZ$hVdOPJT10ItIBQG9eTYxa~;+%LZXv$gcXM~YQn`j|bAlN&4TKqCJ?=Po`z_vlLyGzS|ZE)m}ZsOA3T)ha32Pa$`MNt0Po<^*GNFgldPgJXR5^9c~c!!g@ZzizI|42An;DM=_UyLYyM -sn1m44A!v=xbNk$NOc$E7YLEzyrBF_p0frrOg%OME7NGG2=6n1%*?FE9si|bjRAPB5{7zBYAXNkrrOb -}RjI3@_pL6rhQ;6>Ul}>oECmw;evb9F2?D>wZ)JkOoUB?P2)wwBLgEF2z>8wVh9L0bU -bYAb0*l0x0zqIw*cS)_FCGe;}8TMy~r9ELEzDgL_-!P2rM`T69j&lC5RyKh_`eU2m+6`i(iJ{wVmw -|g21CUnZ!a6c=V3wNyt}tk|tEUU-Ts8`vHG1Lbb{ZMW}W($KF0dwaSx3sCN7m>kWiz$F#Y~q43JK^ZZ -4qcD$2kC_=U4T|DX#s#Sg_Lbc;}*egS*R`D+g)sC5>E)c3!9&EW%5@mrN_+1{c2- -Pa@*Mw^4_eltGz5`!1_lL~F0EWk(3n+N*1M;2~Q3hn`!Q#UJR_l!JDlvsnV9%?snn*3p8m!g2NExw)HEU>T{RMXuS;x{EJ;O~$ -Q^WQ3-~cC(8Iz9!PH1ap6pe7F+Z%`VcIEeP$QQ?<<$LnI%i$e`L&S(Tl&nNT#ktc)-Dw9#kSeaPik)&#C+CsSCs}G?0OkUu$5pY9-@U{3``R%+D$qf__fH5cTs4*0UOMf{QPzySyrOQ%Jy&^;QK#)-NjqZ>&m9WR^p8T0yVPBnv~MXG@_bvtkms)z40*nzV94{iugT< ->hz+>OK@L=j;D8$q0&Yxi3E0eAq394FxJ=8Z5&?kANSW^+}8}bBA*{W~IWiqv;zM(afDFF2i5vit1(>D}bF=d#(p(U@Wv-AylX{M0UH? -%M@m6E<8x73tE`i8<+rqcj;~Q=6wT;H!ln<-rP^fOYf*RuM8~Syf(0Sa}30H)SI^k<~K*WorC$!QU -2x9Zt>>GySuboJR2)w2a9K?1{TkTy!VK1!HW3O9`$UFuKIDv`=I*q+ -Q8z+8w2xj!0PbvNXFMBNQ}A5nLEd0&21-Q5^ie0OVL?o=$-FFz`tjpe$7d4^)S?qHswSgv1QCteN9^~>w -jvoiyWXLk)Op6%s*d4{cbg=M>gxl^%hcd&T#!oZ?aOTd?DAyHx3?qKoVwSmQVHwG5p4SBB@ABJVTgZa -a-jCU}97!v(1uNQBQ<-CLW&9R(!Fuyr+{Vva{H}4u)ym@Y5@#c{Cta|glfyJ8_1{QA)d2di}UKv=td2 -L|v=8*RW_2#XC`OVF;euH>(=k8f6Z -1bmCaW8hCKJOTcUc>k{P|1S|uP$3i1!4y!fx@J)%Mg(#Z9ZV~(3dx9aY0|YzDjw=x&&&Zzw$*aGBDxLB^r8-PS)4&xXC&AjTRp4;{>7t!MlL_2b4r5M!-p{R8pc)<59yM*5*N8OT3$R+c{ugs+47! -;pa}jWALW9n2rrH~xY6u)gsR)Q1^;F1(Ar(dWXu=o@`5yoKlEoUd`xp^=d|+t5>u7 -oL>!exr6!D3aiig)e57}#j6!YpNm&3tUed7X7#yvHKWhf-H^9pgOHXe*&yU4N;U}50Lca+Gf}cZNKKS -kUKd=7tDY4{xV?5=^uvmtNA1|dyRvO&mGlyp -yN_#E}-khh|HO2g-&UxidiNd;AgDN=VE&p`AjvL -Pi~gLFvA)>MWm5c6KRo>P(fa#BvgBFK!Cln7EIrNe~ONC$JbAaE)<52Qv)&I6HE2eZRe8|Fa$cxhnq-IamGcSGLFpF(Wb!Q#7N -DU{RFn8=>E8%1rTUxDbZ^eY;>(i3-E&p_M_OQB4EV?6_LH!OuRfru_USlkUu@xHpdFtE5A@>aT~S(yo -L?EDt+%FJM6J%4^RM4;thYK`Zw?#BFHE6E@fx^WjMHV83ncY&gVQ10f^N5cO%LGF))|Br$<&PI#7dj= -MFL+VG=-Gzb0-S7_%s=G~mPu&gw;C*#>V_JS@%#6Aja_%F|1V(3`vU=+`2B|phP=fbt -4;h~%(2?U@5LOeP5fTWvD(D%#T=_m{9eqlI-W%`p?dS98q7@6{YL@q0DL -O#EKWF%!R6b8ISpFYY$$y&J_nR^jMi_h6G5px(=@B-K}8B&peJG6U4?HJJfw_L|H9HG5;LIe3DZ%m6i -eO=f_ay(Tk2&EB|P`MaByzq?)WyO9|nzZ;nW%5AEV8KC^48kqqq8>2>MfTBcdWCkcoq()|dqD1N>U$N -Jc-ip1J^jGe+V&xBmyan%2BQxOATbC)nvQFfNn>w>G8o_SWnPv3^zg1_JxDyOlof!vRkyv$R$vd&ytP -###zJNOx0o~;bxNEa13QzsQ6iX-cgo&>1RdN10vzb@2qo%kz^$ylrEA^gDIVecm5o#0dTY5 -E~(|wx{q*PbmW(28WVq$(9WdE6{-v&7f5cgXm9@Q`Or!Ek`V1l`FJ|Vofe1Z4{L*wbsoQmP8N0_W^sb -7m}o6qks70FYNn2P18r9?5s*{NTcisz|cFd5qV7fg<}l%|d9=|`A#p8g7`&ky8;E2h@=H(K?q!U53k( -NN4E>UT5zza9ot457NfaKic#t+_xwqMYIh)df?L0Cj#aN74I&YQ5lG}A>)_*mP@ -pMvuE!3dqFK=?cA|fBPq31tP*&!I&Qj3w)k#6UhzsP2^K*f5cdN9s5s=W!JGkV -XS){`&-6xyd+f*YCy^^tfhK&%`zv;pAf6l>Dwle6RH6%HNTeXfHYvG=GW5Z+$Qmz^<&0-hup%L?~p%d -EWeKZ55{`evH!_fA6+HBqX0@7Dt}%xWQ9VR~o2TUafCo>Ma~^Bu$UC-cgh=TGL9G0&f6w(bV=Co|(%yp0*hJb -$nc;!g>~9*F;qF~9%eKgdNywi+i-H`rA{rGy$3cK%Do{P_6a8S~@gKQrdX$KNm}YveRtKV$yee`C;(j -Q@o(KQjJT#{9_mKNzdzd*o&V3#%{Cg_qYV*z8_m4VBBoBbq$LO{}x|8k6u$I}RF?&rC&T8k2BLg-sfh5lkh18u`DAO`j<-nU^|T%D6aE=v-WZYB -*LJ+!C$F#VcsukT81r_Y{GGPrO%OTJEz+#W7iy7X^;e8}Z}6`PLv~e%F+VyZ#{7CFX3Q)3gfYLCNg4A -4K4Z)iGhN2~OvoAYK7EfdKSlbCdHK=$qK&-#XkF4qUVh}0wvk^vk>A-y-h1%=>a^#*8YeurkslefQfg -!GYLa)_A_?nO#{9_m31fa_+{Tz68Mia$N5&nD`H?|Ids`$%-Nl$689!yrj|?i<+ae$8pBVF6^&ZANYe -8jxTcknV$C#hl_cP{=%L9ZV`{_Z(y!`%|F)zP|81wRbm@zNEM;P<+`x#?ien0>JYx}kyIgTv7_x%+E_ -(cmZ7L}P5k(mbkkXe}t_($0AdUx&hW$dwMc4u~B7!q$oT_~zWin>u6N)$;^qsvGwjYgNz0J96M$z4$7 ->4EuWS7qh7#EI{VIH9r0em~aOWWS$iY_i{9YizRL-)L-#Y(<( -3#T40k3OE!~Xe&l#D5lWg$Pbn5;z+$}RDMf<}rLeD}I)+l%SJ4?mDeS8 -#i=h1u$P9`a9>4`IPwRYz)*{Q(I$wb*T%J~J?tt}#%hQUk9k -7^id0MfeLnShU5;=xSWCqm(dZJAO(t3VX* -oKelT2dp;SIinK>mKyGy)2|Tg40p~bLURBIy>m{#LaZ>{Ij3JCmKW}v(L)^87Vex;bm#zjd*_TUez2- -==bR$A2e98e=k#!aRew9@bQuD2eCM1tEY|z&oKqzC0PcK8A2d2(rQgmOZT42!S0~q2*jI790~Yx0=tC -WcR@heuJy`plGl~Nqu&!_Cj85CFu&++rt+207+pVy#PTN?^w{u3PZ7k&5Iiu4yR`KneQ8L;9tzqYk(# -Zy_-`hE(Q+`X&?ijG(Zs)9i!B}m#b5>_nEVbJ?tLsb18#`xp&c!0TowK_7#0tBevpOU5<#hx4Y3!WUB -`ex#?C7I%2XxceIjhq+^p>5oItQVb#?D!t#?eY+=d8{_eC^-R39IPT)CsF7ZEfg;Rdnv}gjIC%>4a5u -^67+Cbn@xARXSl6os&9Y6`hkhVHKT|Ix3|MXjbuC?XnQ6eycqejVgYtE!Cp~fUdjDk;{JiZ&ph!y^ThMyiRZ~X{yce~x -PRdOTqNvO*q&pj7c3?rTiJkH!B -o(nv#8b-Vjcw9G(_*~%ecr@Zp;Bk>K;$Gl!nK0s|z~e$;#4CZvrNW5U0*{M@5pM(@mkT4_3Oue1MtmX -gxHcH^PT+BMFyg(y^*ncaYw~x56|5jY2jJUA>Ry@NVabf?hxZ)adVgIdomObLa{#)^^bi{@Ix8 -iwez;j{$t$1cS;==x0@$7TNh5fhU`R9lW`)|c%%ZLm6Z^d)a5f}E~if5rCF6_S*&qGIC*ncaYiH>P0sUzcLKbHqh_U7Dum$IIyxIHP#{W)ZoU;yKiai^#ncPoqX$MDC?{>ObNlaxcX*q7fI -7dnukyjkt*1OYuxU;0x(lWaM5*&mtrDLV6Y%xfjy2$jH5ro<&CPh4d^kaxbK3k&$~LJ&TOo3+Y*86{9MtQ+m#}IuEeD3i1QhcD@EK~(RtOCB5tneyy{93H&=9Cb)|@# -D>|>bO4?oLRaZ&7>%8hJX?LAhT_x?V^Qx<)-F04dm9)FgtFDrE*Ll@d((XF1x=Pwz=T%opyX(B_Drt9 -}S6!8`dnN6z^Qo((-E}^7m9)Fgr>>HA*ZI`d((XE+x?0*@=TlcpyX$=FYR&+;TH0IZQCCZQ>lu)%`OM -wboGEe*+YLEhXB)J_^)_*To4TJ@y6&)EV)t|A$0zcBNspNi*XN1*9e+QiPu%bMxKbvVnG*TnA7Yw$L? -^HKhnQv_(J2@FtBl%R(->qh}r^R(agf<8Fz3a#vRk -C9=ui>9MWEDv!ImqA0P-;|}}V$#?E*o}$DmkGp!zE3wMst{(A9tn#>{$GZ}%JPuy_l6~`J;@`in34js -{J?`kiu*5=-@Zp+XDbZ!>F8OS%_y~Wk$(IsqKK@n@C?(c>+|k2Ti8UX0^k7wD%}2ZdcmyXa-_dn_i6t -N5p*7=EV$sJP&9{|U^l?YCZ6y|c+|^uKiA5iGbzN0r(Z^kqny~2O_wNzs#Hx?G44tfem;7K*G><;i^- -S4vC87i~b64|9C3u;;x;8Gs%G@RK4^HMT$x2{k{+_&J@G*a{*~$`Z%-?ISvJ_m*YhP+EuM~{QYhRL#* -z(G+N{sZ?jE`Z}jL#se8S9TSUT^sQMdI@l8e3H)K0l#ZRt2wYs)S})B`TH5C7!3|nqSoVj96b>e;2bN+klL?u_Zyz;ACZr_?NTN$tC^!{eX8Hi@aTC9x -Md8OOT`13Sb;`6h_=V$!)Hiw!9xjzvvO~K=_Nof2{!d}gs$5)fkP#c%HKT+8?bX{h+zDV3pmz|8)xj# -1pzrQ`6-;S5HCvZq^_{M~#+JalQO=OQL3&{a(>#A%aIiPJln_ox{Xj{)07m@?oju(!NxU|2X -wJ#(Wv>kV48gXg=c+0|wOZ)3-wm9=M;F{AGhkgcJ*Oubc&wy)ALLB=UaLs9pb3X&FXQ$%e&w%S0s5tp -E;F{AGM}G!fSD@nT&wy)ATO9rwa6Pvcr+)@q*X`o?&w%U7U7Y_JaLvDn13&|=c>!?(XuvfEBaQ$KxSl -16Ge85bCqm*7(17bXm^cMA;F{GC$AAW0&-2AOpaIu2eQ^+I!1Y{ToCF$h&25XLKm)F)^WrSffNO4hA= -UV8&22BF8o#aQf(xm}Z)?r)=sZv|?j$OZMWO;fI$lfP-z7A$E}@BaiHf>UR8U84rpWz?in>p%dN^us# -O~*-D*86@`3Y{ir%gq(*^qcYL*o4$?IkAnmx=qU#QllqKU#24?oZ6f23}!iiB-H=VioKtp7Q%kS_g*f -v|F6b#@L*xW1tN92_z6{5`GH!iz8}>JvEnb^JzZ*(|Z%H_Hmvxo*7f`TcY1nOVl?K3P_BPsn& -AW?4@w_-A5$bGS~qf^eO31>ri6^kT+;Z;??}G+gKYP~_ZRMb7P26p7C-cxEg3diroZ@q -78m7_QTiD=X$aUWz&8;==XB@13VUpZi;4eR;T^ct4y=4A*)7DI6V1LmErGPtWsP;S>JDN78Vezt4HXp -XZK_q~SV${<-J*XP(GEbGq9uTu+Q#^N#zEOMJd-`N->9p5I)@{UM>6XFlPdKH)!}S3a5Nm0`Hf?duc2 -*Yi~Zndg;oxX$k{6Ti32c>XLi?hj?o{io!WVOGxhc$CmGvLlUv+%IOFTh|ri6!1*oig5~fE^x&-1-uZ -rVw?g#7r0`a0`3H^7^i@Hfh)!-tHg|R>$+l`0$vGRF-`%m1+EyUfHwl?N7J|yIFNJv=2m>O;<5t15V+ -#90^SK+aajTH1+KWPfDZy!Tvk@K(z6to6>#ZUipvVP^en|?1zdWT;<5rRJxg&}0hgYoxU7In&r)1gz@ -=v?E-T>DvlN#VaOqi!%L=&kEX8F7T-sl8Spk>!S6o)WrTrC`6>w>P#bpIt+Fx;5S(V772~{s#=apQ|R -j6>Ca`E9h&(t+nt+JY{R^d9Y%<3%hJQ*J;Yp#Wa>%0=Gxyl=^^UADFj0AOJWUUh;Yt46#gzLQWs2g4x -*K`N^cs+4{&ntr_F_JZjnQB6%%bJG!Ta%diGy{MCZQ^;_guf-UpKzUj{(`R($QFFPVYtryZ87Kmwn)6 -6MWS+ABr4+tJ$rDx&SyT0g#RqKmX=gAXu@_g>8r{!~=* -}`=?66J|GdG6-4a>xhn$@#!5pCZAY7jt@oR=Cbrniq4viejF4KXcD32cMWj`i55~zTwZ~8%N*2Pppp# -*QtHOb;{=ZhBi|Q*SWui>xs|j9ExvxnwfpWInZ#OzJH&n9DK`1YTxoq9=FKBiFve)s8KHbnM!9xZ1Jh#nEd4S8M3KID9SOY8AcjowUE|S>Y4vz3*#U+4w}|QCUrz404qWzZ(xe4FL8GHWxg=7sBt`}xQnuBYy&Z!%n`mTfanSH6eq)N|U5SGsMY8&#X&EZT -*qk=Q0Uf^eO33hjbd((Qt)0dWK -2>Ri`wDWOsyMm~XA+3}ROJ!wqw(5hd_8)&PAl^+6Wsl4=RV`_Oz`)wo%=j-PZxg#9u8NsP8j!(N8!Ko;YkGmPcsUnWgyBWc$9F7mX8Ns -O*j!(Q9!6}U5BX34CqDFM1gB_;PrVtzDMaF9Zvu#Jb!yvdCtB6T2j7g~WY75In-QGs86 -SN!f|EVtvu{RlvgcnBaI$B7`ppPV_WTI}Cws={-;Cg7&-ehG5uEHvJRZ&Hh{vN79r1Xyp(7rT9(2Ux( -SVM4Ji5;jk4Nh{;_>J^M?4-)=ZMFnBEcv4o@ONG#zjIuc7bjE=+-PNO5SgyZN)Ea5yl5=%Icj>Hm9q$9C}Bk4#i;Y>ObOE{E{#1c-WBe8^ -I=}0W$Tsjgru&;_=`xiN}M$Bpwg;l6X9*OXBh1Es4j2v?L -x6#*%nE=t|=8;3|p7gQz4P50;X6JSa-y@!%(k$Ag?C9uH=ccsyuH;_=`liN}MGBpweol6X9*NaFF}A& -JL3N+xx4kdMUK!8{UY2kl6l9h@U^b`Xxl*}*mvX9v|toE^QBCI6DrjBhHSq>WH)Bs5;{8IH``LcQiUA -9*-`E#N*N8ka#@$8xoI4b3@|s=xj(l9&HVY$D^kq@pv>eBp#1$hQ#C1%8+Rd#N)ARop?O9tP_vNes$vU*sMNz{P9Bl1dh-fzjpEz5=+ -pO`h?0YA%giY@xmayZU#1giI}$7Cbs}V36m;fzcJ6WLmUDCz%%g(20vDgeB1 -gop?7opcC&#`*V_H(fgdZI2xZ57f07~;^Js|PFx)Q&WVeo**S4>bUG(4jyC7S#XF0??=1elv-tbY;_o -|)zwa#mzO(rI&f@Poi@)zI{=T#Ldz?E*oV~O7`_AI;JBz>XEdIW?`1{`C?|X~C?=AkmxA^P`(E?+$E)xYIP5Nb?O-VbgQW}%mNGC{%D|wN0RdY4{b2F;g -T>zu7Jolj{QY3@_rv5VB$ft?zaK3Aez5ra!Qk)R%;4|c%;4|c%;4|c%;4|c%;4|c%;4`G?@Su58`Qm< -8PvU-8PvU-8PvTy6k*x37{t7r8N|Gs8N|HH3}W781~Kn4gP3=jLCm|%Am&}hiCmW%oV?2nPTpk(C+{+ -YlXscH$vdd4#EUb7lXn@P$+*m*<6UOZ@h&&$c$XV=yvq$b-sJ`z?{b5VcaT4I%67TI$-CU(pyTah)U14zYt}r-xR~VeUD-2HF6$U5o3WJk(g~ -7=?-q0mnpG--dy)fu_R~U4>D-1f`6?A!Q=jH|z@8$**@8$**@8$**@8$**?|9dcaNQu`9ba@At{W`6g -J4N~bZ)TdZf>yXZf>yXZf>yX&KWGaB;P{{B_3ZIJi9pRjd(nyP~!2W!NiLb-iXIT3MC#7DU^78X>j -%8R5#-BkV1*amj+ia4s|0QUmAqHIL?iDd}R>!;vhHT@s&Z?izD2K$5#enFAi@b9$y)Ry*Rdwczk6L_T -s=c;_(nHiN`~*BpwgJl6ZV&Q1{}bHsbM>LEVcp+K9*326Zn^XCoe88`QlxpN)8YZSeQ5Hu!s28~nYi4 -gTKM27m8rgTHqWEQ!b027m8rgTHsR!QZ>u;O|{y@b|7U_9XYu!)#ou?v`g_+|Bz -|X+_?<=KcNU2cC6;)6XRN<>okjC^7R}#TG=Fc={Jll<_ZH3HTQq-f(fqx!{@(Q#*WX)Qe{XU9y~XwS7 -S|6YmUw(`as9oq{@(Q#;on{)0vM4;JA+ScLyz5&nZk_zxE0551Il{9qCO -gGKlc7U4e_gukB|gukB|gukB|gukB|gukB|gukB|gukB|gukB|gukB|gukB|gukB|i|{>mu4<3>GlTW -_GlTW_GlTW_Gh-3H&kX9{X9o4}GlTl~nL+*g%%J{#W-P+@nZf`2%vgl)Gh-3H&x}R*J~QM44{eord}h -c8J~QM4pK1Bvc(r~K$M=P=HG~PD8^VOojphA5H`EEA8|sA54RyljhC1PMV|l;N4GqEPhKAsCLqqVnp& -@ukwyfZWe?+gvWJ3~Y8&d?CNGc*M63=P3MLqqV+&=9;cGz1U -LmUz4~Gz9Mq4Z%A@L-5Yf5WF`u1n&(E!FxkP@ZQi6yf-ui?+p#XdqYF;-p~-dH#7wA4GqD2LqqV;Y>C -HvLqqW1&=9;gGz9Mr4Z)X&hTuy>L-3`cA^6hJ5PWH92);Bl1Ya5&f-emX!Iy@H;7db8@TH+4cxbl7<4 -Z$B@TH+4_|ni2d}(M1zA`ifUl|&Lue646yjnkrqXEO$8tR0v40Xa+hC1OZL!I!Ip-%Y9P$zt4s1v?2b -{BZ4!6cR{LqqVDp&|In&=7oOXb8SGGz4E88iKD44Z+ujhTv;ML-4hsA^6(R5PWTD2);Hn1Ya8(g0Bq? -!9xuuu~Zuxg0Bq?!Pka{;A=xe@QtA%_{Pu>d}C+`zA-ce4>kDjbYDUbIr#soKEna!y8t}Y-h1pbCxhM -79S%M8-g~OEazH1O-BW#)13KC4p6aR`(8*}`R8QrAPFB0ee(7Ykd#ayu=b`%EQ{9vUIvMVs>ZKge$#V -BpC*^=nrn{&5CO{7H*5o2re*F9otC*fgj(kAP--8YSLZDT*v!EYYnkgnt!2&uxt9H4=(WrYy=lVU4AI^U(cTQ -v-VD*+4AI^U(Vm3p@nZiZPJImDYX*OB20uh$X5Zcn{@x7!-VFZU4F16k{=xK|!3=(g!|Yu_9cHOvFjK -=|riQ^x4TG5)1~WAbW@>;)Oh-b<#4OC96SFXbP|U&%N-+tu5>hbp5_ -&NSh7y7?35F7iF$snek}(N}5}Gjyh7zJN35F7?F$snevM~vU61p*Q_Y&GLarYA9Fmd-1>M(Kl67n!{_ -Y(RrarY7eF>&`23NafA%goB|B}8K8I#7w3b3i7hKe(~XkeMdTtPm|TD@4o83ehq%Lo_!-G&e&uH$yZx -Lo_!-G&e&uH$yZxLo_!-G&e&uH$yZx?VFn+np@Rb35AyhaBc>0VFqwv25?~paA5{;VFqv^0+@X?3`~L -nS}zL*h`lTrp!Tw0fZWT10eUYB1_-_^7z)$CkbGG%K=Wn60MVBP15{rY43K?UFhKWZ!2scx1p|~{77U -PnSujBRW$q4rm$^FxUgqvlc$vFH;$`j*jhDGQL|(Z2@kzl+9Ml=Uck&H75;`-xLqcXw2@RRK9YkdIDx -o4X13S|>oEiMi41Q+@zcYj1nZXY+nFYTygWs9K@66zLX7GD6_`MnY-VAO|5Sm#ql%|0pHM3xV -*35zdVlxW{sLd=GAUCsMfZoi40fI9N1}M%f7%GcAE1@|vcZd4R+#T{Sb9d;!%-taXGk1pq%-kIkFmrc -kz;wK-%p43En7IygVCEc^rBRenf~gry6V_&k)@F#-W{B2ih}LF^)@F#-W{B2ih}LF^)@F#-W{5%-W+7 -UeAzGUuTALwSoAzzY5N*sF3hFQm;KmH##th)b4B*BL;KmH##th&_1Tgz(7?=cu5RpkR2oaeCgAkEPFb -ENu1cMNfNiYZznFK>?8d!+PBp8HkTc -eH`G~gsJGtGV7wtH$>a?|NG5LxIx=}fkdet7f{IMu5JY71hM*ynHv|cpydfyayqYOOc4u- -;H-y`kQELqqxvB~GL7nu=%k=CGb8_a{yaX&QcJaMSYp6K8dIL&lq24mm%=c{o&V)ct+p?Co%!wwVmqD -}H}Q&j1eB>HFzthq@ZM|4>n5_tRdP;kr-UpZLAditeiuwF5c@?0r?*2XuM1L=oGT|RcRm4DQ54h -(mtS5(B4;-P(Y`sy+4)q$4l&!I5ahU@01Ykt4TvZr;Ko4)n@^nQo?;zp9OTv3HMd;9?&T%+*ieWK&Py -5UzO(pozlX6Rh|cQ$_o!vZ5PlfF+5OpVLyWi^+46K0i8VOfhyhuI(g0mRlEmu@|*{%cn|30IS*7V8_> -yf9;o6ypp)l3P{n&dC;L8lnxK7-IvM!E6@pF%e((}OCjAvCw0*U}8;SC{)g*SvG7Tyq&S -a>@#fSnn@5RzDkLP%mk2qB3DA%r9rgbDc{U;(q>qj^=6P{(; -}0m%~u@A&Vnv266<^|#(6cqZ`c1kVNj7QqXF-yry0;5P~G1b&O)Uf^#NycD<|ifb$*4f -s0*uLZ7?OpT?a0e_d^t-#+S_(I_C6TB1nR4g2?x=-Q+)bP!{_-0*d)L6_KaGgQxOnR0sH|k7!mM%EzO -nR0sIqFP$mM%K#OnR0sJL*h&mM%Q%OnR0sJ?cz)mJ%caEyd -LG>bUEXalg@pXpemRjrN%y+4dV5&H?>MYhH3x -1YLUHizqzO~27oygtEB_PH64hwJp&Uzw4Zn~|8Ck(is2 -n46K9n~|8Ckyx0KSeTJmn2}f{>{XaCSeP+bm@!zGF<2yQTA0ztkIxF%&A2Q0h}_Rj51%LA*xdBhdE$4 -?OiIAS -DCY|%A9RgW`tH|gjQyRR*C$H)!}U1tjq|l%m}T_2(8Qrt<4Cn%?Pc{2(8U|R&B;HE4 -YBO4EGg@miTIjG@JMXwK02cWA*-_Z>QCsQ -(Ui3@yJyJwvPS(2}9`cWA}X<~y`zXiML1cF2Vu{?%iD@`nVTUh!l9fx*)|e(X;fJiX+{{*1xXTYl`%8 -9cq_$NnROr}zBWe`4_Tq96MU22XGLvH#5A=~X}WUl=^S>&N~pgWC!IZwy{c@b?+qP4Itba6iHSgTccD -|A4?b?EOC(4151C2E*PzWH9XgBL>6Xf5~9j`>z-bd;gfhu=h_G4151IgJJK#Vem8>EFZu8!{y^@WlXq -b8WL11-T+RuglhUaWT037R5j`iC=NqWjZ_0lmun0eh*i~6IDqN)+yP9t=MG@HJ$C@p?YRS(ZqFUSbbI -aqrrUD|Fx{Rzfa&(!0Zg~&4&Y==)lxWs>Gs?KOt8_K;y|>0huXnhg4VhNA5<@oUh_KV+|Pvgr?gz&?7i>W@i0a -gBV)z{Gu@{1ukjeV;ro__*&Ackpz5)R1AW{%0gu>SLAJ@#_2}j!qr2hO)Oq*3)(ULk_im`NS8?C)A?I -aK{uS%TwFoKXONoJJjSVz?2poa>&W+C&~9hPhP$EB7vtbU0r;Jz*B@SPb-=#z*CIwe$M`Siqh3vuMl{ -O)6FflmMp*}{#q^73UGz*Rx8Z{JVopJrE3J9Tzv10VhaO2`S|*ZS_>ZF$;nqws#WL#ot{%ep_bZwL -f$#5sr8eW2n>7QRD4c=CwuSRR#ZZOVejV@qZZ)F-s{iFJ4ao$zDnLXDy#KX_TiJg_b!vShU#kdq@wcz -JlT77Me!>EhP}^{5abf}kfRO?h4}jFyQ~z?Q4fWig17OTa(UC|&+wZHantJk#3#+gO?#(4<2O0^;7?x -SH+lIii`^&`h95r3)|^8{3=V_&sTg5Ri -c)yxaL1p34f@n#OL#!4?~rx6stCIf6M)|PQ0JSarU%HJb%M?stiq;xSy-3LsKR0uZ~6S@ml;O4n`fC# -G7y0#4l|*DbXx=q&Ew?@^@%D?wL&@QktIUvnCNK&A{((6Q0o4d;-uO^6TXO!(%7N{pby#cH0MV%C2g+ -eE_FHOYOD~;51mN-Sz>T1}U}MK7iBUq;}f}a2k}8AHZo4QoHQ~IF%1-w|xLv&IofnYMHz -_I)_2s-|(EBwV0M5T1K#yvz=U*o<_H|r)i9H)8C#n)2;Q??wXUo4GrN4#aN%#n(Ry!N(<|B+fJ -=ZQ@B{=He=i~Y02xE3~mb^E?KB#H-2xCLfwHIRIWW-_eb8#6u!Z7(c_RBFH{UW9SMi?eP7cYMpVVL|} -Oa+ZF_VV2RX*`P`aqtg6Wq%HnpO07Cjqqgh4trmSw>FP3?0q4o;YJ -uCej%P{2N?Xr>Hu3_b(7@g5y|Si5FCjFnmQ(y7{;B%5P!f3lM)+S4faR{#m_9>I3LMtE+L-gAsMde$>@3i_$q2*XYLEU1!{9 -dvjD8~rZ@fZs2euj}G^q3Ou#?N7~JHm#92L~cE -;<7?RQD8g&MkgH{Lis!YIbxxW)b)#{7U(TiAoJ`hfjn81n=2zOV~n^#MssrLg6@oDRkhy!9T!t!-8&%Pzxou@zx?ofHPxi*RuLDN?{>ZX5 -WrL{WXaN>_}K%C#4`7)~~)|-wmt%hGYS>tY7_x)LZCTzxoYfCD62feTuww?21^bHXtJmUGOc!h+tpD` -V^sK(768K%~u%?!B#Ej7-0yuXURy0EfVXqgwwztiS^m37Cb)bIDz9?$1kq!=V}16H$J)2`Wz`?u@7T? -jud*>h_Plg5O!j$&ylo>tr+W{k~t#wVyu5kz8jk{);}f52fH!W=gDLX+cDPXNuh@Q80#nFWhNsGPkM@ -^9_+|iKTZ4)TQb%c*rOqLy!kG%DE3aQpCRK3HczadCK(^QC)O9pV1)zM*UysS1P$-k7m2~q@&4e=57> -)=#h)W-4^8ja7YU1muJ`L_$w-N|_sh3QeSyCB>n--zu=o}kZP56B{XB69biQA2kr@wK-!E^ISp$0CuU -{Y>H=5tCUtoWY-bmLUC*v#*U|+vTMhl$4zJ8UAWH^F-{W2ln&_sXvF5wl>Mt}VZnVqAN{*ui-&`N*#b -K-4iroTEz>IbybU!5a#DH`gpuaa6BE%n#0kf8`o^*5*DMNcC<8GQF5c`azHzh;~UTI;W`lH`Qu`s=Ht -)UhR*3KG_l -edsh&jMJq>zRJfAGe~gu#Ode{lWh?CW84`qNwd`A1}+N2C7rM}*x#tN!&idw7&?A2O@K=pSoJS2WUGf -2;>n(MNOri5@LQ8_o46dJZAFXs$oe1G#9Tx&B0tXQGGZ`jh$_UpXukW4LBbjKExp!Vm`?Xl!|D>K+)s -WW@-g-%wn>F$R`Uu=^6T+b0|DC~RB5 -@SPQUw%$ld0t -s3E}bT_)pLAx!#_u2*ZZi2nu6uGWR*ZgqK5LxJCWiMTTq_=C5;(v?>Y2|nP&s?gvMZakqS;~FCT@;AD -^s3F4dK2CxOBK+Q~WT1ppesJS78S9{x2V7TnHN^6R8)x(ktA<)0aQ68K=Bp&HV>7_sRbm;)<$JFYr-W -R-%f=MQ<$FZgts$51K0yX7$mP4I$e1!0R((!SL~7{edvB1@5_a?=PswNu&3x}y#HP^9_dX* -r0?mB)8IraknIGI_1sXK-gPTwCNl0}9N6(Jm3~#zX#sjGG2REN)IK1iCr22v^e{k~|QVBqp-}{QxhS2 -5re#1U}a;MGX%qDQBo0mx5gD!t?^D^-#2=kjKNGL#;Ke%~?*>u34lPabX^Ro+Ntb{VZdy%{vDD#_BWS -oaGzd57X$QsK0?sH@)g)+Z6L#!s0`FBXgUJ0YWPBK&_MBdlg-wT50SxvRpl_2h)C51D@d8T035a&1NN -%;nGe)AQna3Id_Zjsauaeng@nbJd>-#kU$JGA-DQ=|q#f5FvLq&5@EJyF+cDEAxs*Rc6hWY!E}fAcgc -L4>e>^IbC1LfGFtO(s&%^*7J3AC5r3KmrxI{^nUSWrwc6d4~Np4E`+p;gi8P7l|`N*WYYuNuq|Xzqvq -M8M^+4tW>U{>u*$_g%O6qFOoPAy8cZzOKU`KCrgKF2>S;&w@C_uvcG$o{W(m2iMS*5_su2Zj?mvXmq| -?v{r%t;eRHt+3uFQd0si3Tmt@`#1%CGh;*U_^cV8egNGR|JH@_l39TNP(t<$6ig9g9*0(r5};5RRm`7 -Jc~&CA3qpuulmCjJZ!e)BS6iJ`%7USTk7{tB7gLW191C6g#f@S9h)%w0o*-@HbqhmhbmSIKY;34Ze$d -#MQX*GcKwCT0SV-ZwuZo&f26bBzo|kluG+B)HXl=n}pkj_I_~d1{sYYzVE(5>KBObyRVX&0L1srJLI)PeBb<>j5ZM -8H}8|g1o3_IA<3x_-w$qmM7$X4`{pC!#Zccj+r;05`u;It{h+>Ypf4Stgq*qu1!aHpOH!La+24Fh3U#6EUt^H~WqrEt348?eD)zrhd@&SHC0^M`-)|SIM9P -ZU5jl)3=1UfBQU{b3@$Uy-gen+WzisQn5nX-~SoO%tG3~{T#`?koI@qVLu&4zx_O^Iic)BfajeU2UJsT>4ZJk^jNbLKckk=+8_IFN`Plv?5yRGwIBQ -~Ii5 -ZU*?WRC`$f60Cr;`p7Hh|AE%7)7E{N0AkW!jif_D`@keA}?Cj*rT5v9pn6O^!4G`{bRlLmp}U9-2d61 -{OC`<_piVAqp_TN><~IE!Df&B;o(32<J -hMNgJRkCsKxmPHSjMNgMSkC#Qymq*W+$L}qAzC3!qJbJ!7dcHh*zC3!qJbJ!7dcGohz9M?QA~tOFd`0 -wpMf7|{^n6A1d`0wpMf800==tW+^Ub5@o5!9PJ>NWfzIpU~^XU2J(eurt=X25Xx#;;^^n5OQJ{N~Z^n -5OQJ{LWoi=NL#&*!7(^U?G9==pr~d_H@6aWAK2ms3fSzB1O -C_0D%003qJ000{R003}la4%nWWo~3|axZOjXK-O-YcE4jP+3V%M`e*gZ<|06h420qPd+7*gB_o$Hi}A -ZS4EAmWSm4fu3^BV!tQo=DdFFDmbg-)azV?yee=D?#@Nwua}K!LTw}djeY#n%Zf}w0={7~K%rC~BH@Y -#kDwVFruT7`8O-j?<&t?D~xFqeVLa!@w2xmfSRVlB_WPvAgo{h#jS>Y-Q=9kg@`b)GJO^;5E&?cbwBp -&Qhb4#Gtl57BL$UfV$(j{TwaGv&DT$1SVa`7;yphT|Zf|XPKUI-N?m4%l1v^m -k-%ymlwMmL7uKW>Xy8Rqu)I{340o$tN@kjB$%xlU>sEANM^xA;K6ZV-7Mfu9R}qn+Ena=<5M3#^z{!< -Qok2SV@hM57RZV@zNJN0ETYMCcKoq_JmNV$9Mb%l?F#G<*zEIdcFgvc!Jc4cm4Z*nhfb7yd2V{0#8UukY>bYEXCaCx0mO^=%}5WV|XjL -3yV$lGjtLh7Mv(+^c`quuPGConK%y)icRIGfGCudyMUZl!ixgkX%H-@G?7%)sv3HV18fh#cO(da-1(> -y33_h_Tk14mdv4*gI)7!ML{om^~hOdBV8~uzGNtV~-hZWQXKL*E4pHmOKf(hIE;Mzc1&kD8{cv;lC$K -zT}q)V@$R_KSzBptpMHy^)!P3AcRf5q&rP^uF3Zj!L}G1?4nbU#M>5LFB -nVLa*VF$Rw`McvW1e-_VhDAoo(1?<+?iX*8Q8Nq7%5V5&Q;9h9 -a7p&K|@=@6s?XP}JdOBT@Ah>+Jx5aneQP(o`gnAlhT3w6}x`w@}pNkQ0+ygcMeBa}zY^ZA*DYVhz6+A -X@h)iHMi)zKmShxC!gfdf{Xxm2}5pPRNg?!aFkhq7#;Gg}Ufv -y;I1UGNSPrBXL;KY3Zl6*>}BPhs`0TaZ$0KP)h>@6aWAK2ms3fSz9|hoztxW0021z001EX003}la4%n -WWo~3|axZOjXK-O-YcFMZV`Xr3X>V?GE^v8`Q_D`nFc7@wD^}%_N~Irw#03$gUMLcUYh~g!v2g6lkCy -&DYda59Kq3dn@yyPScQ>tK&cTa5u=b<*Z1Lcw(e<=j-c&%(VzIbm4xvqKVJtY)Bj8ALV3~Bt;nYcm5R -#nK@9QdsKW=x;&2S1mJ7@MB4#szA!FNo93vW4#5h?b<+8SUZ;A{e%nD*cew9L01Fr@&kHFlK`QtT83Q -ag`~ugBX|Rh6^c1tHKr1f}YWOvBX=#G2UK;J4zQdG;UjgRwpHO*B=%G8M4N{(cW7QyYC`N^Y_%tC&KB!#9$AFPH&)~er@bkd?cQi3d$O7KUUevIYHkNbx0iVjrAg|LtA% -8VFV%sBHq0DK7zkOF@6aWAK2ms3fSz9IGNB4OK003SV000^Q003}la4%nWWo~3|axZOjXK-O-Y -cFMZbS`jtrC4om+cpsX?q6|pu|Y~5RZfC*Hj-jM+72DIpv97{KvBrFOvl<}(iN#RN!R`M-BFZGTC(G9 -Lf$O$-k*DTOtxIfLc@${uCpcYWz}b;i!#-{c_EjO$~@<(&ZJOamjzq$nJ!m3@AZ0V&Xj^zclzpn#e?- -y(6{gU1L&u6#>oonhvB#f`1JerIuv|WaK(jI!0$9@dY1`Us+$!{|BP^F&xp+U9Fj!v&q)%bxf%dz9T> -G%DLvU7gw&9!OeoDn%7bcKCbYAQo`PkH1G(_3C}a^-UM@?eVM-q8ez-#`nH|9NUUT)BUfSB|H4<&jF` -S*7fav$|=y{jQ-&kID`WpX@5{}M->#`7_$`vnyFxr;1rg~Rcdkky2YHHKGajT;je7%)w#f3s#v?^AW> --Pm0FfFrO!{Vsve*)->4cqvW!(Ui^aceomS+zfm{T=5{R6w%Mn%tf#R| -40xYufZ=v|h}DK9|Io6s=-DK}uazI^i@_)4Y=2;z8bzTo=m0)AowhA7+Kt!fQn4bO$Svd!4r*|)2f5WoYR`Rg_*71uVY-;BT -Nf=H_=J|h4UW}C*0+)7*ip08=RF7^2r*V(T=1AN)^^lRuHo%~b7AQG`z_NI8q-17xoYBw`cYk12bo!;Hb6RbPqp?ja -81V{6~E(PwZ|eNxXsPeF}!8>xnwirO(Q~+99mQChF35$1IyuXEpMg<6gNIPdaLZHaSL$P4c=jCRk`3O -A?=QeZG@V0fvF>0o%k4a=)eraQDr>nfAce*)6?)n{L$O*^F!RM@!YfWy>Erau+x(cBkwwmtPNPf?;yU -5tLON&eRjBIrCz*%X8(?MErrt8ul9}VF21lz -fQ(dpqd(*{HqCaUi#ZcW84PW4FCY2SK`?BMga1UyYMVGE?D*^J!2-PO@swg*hXoCZS$uijsE_7w!$OJ -3OOg{se-AlyFiXs7~+K|?$Sc+!*|A@?a6s@;64nVjK5D5lceV!RPU}}*g`Pkg?3Fib3#TXhS$NlBp{?E3JUVTqyrtDN -=VlECUh-RQLW&+edrFbr1@t39H5=oEX;|;>VR{!w-0aTrKNEFv+J}qLJV)$nY=RwfjhLYgSBkC<*fr) -;C-&EjhFhCZ;Mp@6u2bT(C~aKt{1!LDft^|Z57(GnwyEYvyt2?mVjVeN8(0+)vc};`QjE7ms~rq7-pq -H47PhCRYfO9EZe`3AAMJ&dgv6@M3H6`R;@zsdu8P(`CuCYzayJ#br-fEY4`$izQqJmt1{MSDs#bCDU0 -M>MS382;R^1JeR)m&7iQ};M4^T@31QY-O00;of09jik9KmHa8vp=?egFU;0001RX>c!Jc4cm4Z*nh -fb7yd2V{0#FVQg$-VPk79aCzlD{d3z!uD|=Q*s9Z4DUnslPE*HI<;J$`tLKt9XFIR$^!3q{MA@umQb$ -s;eVzXA55RtZkdpFo*SV+3#3C1q1+Z8wu!~)_TCJ0KJ|D-ETeesh348lk-Q*cd?-ujqF3GC6O5rh|vA -gnm0Z^@0YikQi^P*y9RV*i!dU`CkW<`F-rg4>2>0QD^fqITh7LDNH5*`4_jNQa#TvbIIs&-jx@wj-5- -X}!~Ek3j;fB>3TDm6V74u(&AV#~M@|jl*t@L@bW3lj?muUvj1#ZmH1eS)=yAU@C?BAQ7gLw#)RNi9 -v`U;+J2yw)nW|B@dHif#XeU?Q%9tij(Dh-et!RRT0DRiM#A1FYe;%xG3_X%g&b7*^Ixxs9=QBYlX3rs -G$1g<09!Y9DN}FyT#Vl@lVmk-%j5G-F^IiboK`Ld3SUPF9!JH{PA~eOA*EM@fKbvSR3kL>aDMP3Ri=l-(>L(&j6WRzgk*fXwKY0EIe7Q#652d^cYbj -ky*hn;N=1VKNRCG1N{&k`cckl-C-!@&9DnWlx^m-p9qYv;=7M@$Wm1JkA6@-pm!Rb=Bi~e -~zS;?CMG>?9~;!zezIogabZdlN=@%umUVW5I)9xc4f_)R~rPJW|JIes48))86zW94A6$aDKO -TWppgW@17k_gqM5xa%#Lh3NozAN19Fj@>nycU!F+xf7dTZeU830Hr@>+ -${XY5jzpOnFLx!%e;6;HNtA;!#QR8g%y=hUE_-JDEL{GbXmq?6F2}wZJSPI4hap}xDG^MV8uGwXBCGh -&QzEPQ$g52=wFVe@^&Nm?4+Bi@y77n8x?n6-WgX=gFYI^ -USmQovV)3#_!OX}i0@aE0YC<7 -!z)-t1vq^G4j%U;pJG?vr_~L6cX_hC$C6CElO%Th&Z(qFtVukA*}ed}Z}NLxOimv5w9t_0z79D^<=M9 -3U7IBXD4EUkxMHk*0aRw7SENV46gz%l!VsF3Yp|BS|p&C1)nmMMw=GyXg$G`!jfIF%N|4X!J -q%Pz-Kjurcy@_Ii%;)`0I=NAZlvkZ*+%M;&78_KqvN2K}v!|JQmENTl*lLWFK&-r_IIWDCWZJ_KdY8n -|+Rl?FyF1&1o+T8)FXR0I3=aBzGPm>}%*p-!>-4arE6K1yN9k8O4&X9?9;a}n1dw!+f|Ba$KZAUP@Je -vGE0ig*nxGC)b*%_`>REziT4MOY%6wT&0+S6&2_zUnD4-=YN)|;j0mFbZ9Js|#99cd=Dl3?Kdah6bkO -8V|P+Sl#AU1WworuC+mS&%a7(l!+F*MoPNf0QUDv`l-|U1Nf#~H^%2kdYTs6X}u -D6=&!_>TjddHi$jqyuE8uVmr;HeMLtQ&vYl^4SD6E*mDFe=z;*#n -baCv~K`yxJ3E04eo4jw6(Ri|B&(ILFqK62u{z~T*HD_rIrq9B}Tn}qQ5F%e%RM9rC&ZoMKWOH`ynkK>sQM2G^^XLn}ft~wj8QFr^YGg>(u^x+l_ -Dv-Nh%EBH0s0(ON0y|9()y|GS2t4Su8sFTR(zB42(m? -3NNu4rwUn3Pt8k*(q$y~aCB~$~Yje#Jn>++=*jm8wBf5%gk(ViVsDS2WQGgEx9wx=+0en$P8xI-{ciELA -j?tS4ykm6Fk^*JW1%Xue2>@bv#hvvE?a4#c?sN@gSS-jEcB=&zsD$ui+G5i9%^$mDpcCQqO(1@XZ;i0 -vOLoeq;jVBHr=TaSoac*0I)8*|_c8~a1?iW#C3`Q7ma!Ye;>savl7nO*7P`x6m86Tc$oU2~Zj%IJ*8* -K|=nf2Nsi{YVxlamMGF9vm<@BxJHw|2;Gc-VC};XN5^a4F+{uea;TQiB;SI55ha0T}h&fiZj~fHI!KaN2Xy(E6QIM|wK|D!&HQ*ER -~`DWvB1pyy8kICn^^=K!odXvI?i&K{5A$yYfDv7&r7D-++!>b?+DCVEA*d~Ufm3M{B6rbnY7)v|lQmw -|@O4sLAP09@I)0lQxFhTg10ga%)i6kX#Z|H^3^voJRfQr+Bb>`&ptDpykyf3LT=+E~|VuP;IOV+mum?m?}ZU1FISn=0DlpuKC9tLZKE7ZDeU*g6Y8r1}J3_R6 -jpPjyW`|9B6I64@O*wZ06(@B|qjXvQ#1xz1-un+M1wNXLsMcK)u%&8n9Z`Rk)w&ipx@4*aJVp&tm*k&V55I+aO{X= -Pz|M>-m6SW!@(tGM#6`mbjf3VqQoQM90vwWWNS3F}Ni>)E}<4f7>iGhx7@K -&P80^>`OCt0|9y__Z#Qf6_2LPvf6t&bd43~Y0I~A!nlXlWR#Rv8YymzRY?2)*c6`wBYKIyP8+t(=h|L -;8qk^#n;K;SU#_=R#g%21N7~46C%5f{Y+VaQ$5Au@qP@AHvvS6hIV|+S{Q!kwom~Ops2C~Ksya`l4kb -#WBc-67kv-I^r@{~=>0kNqH)0WQI+Ic(CU$2U?zE*WQ|3)R^3-LwliU%>Ik@$Adoq;1YBiagD>#>3pTYwZLEb0zWVWKH9Aj_`hQK2x33i05WCqaYIN*Ze&`Y( -cVX@stQ2Nb>Huq?{VWNq$LRX!HVK{9{8058wLW=5ALFs!o}u10lzQ>eE9D+@9nJH`2^hm8_fHk#RJ51 -4`!E+K696Tq8{5^;Qv~m6Ay!!VmlA(D9*bb<>hUwL-BO?B`mk0zhbfiN^h`szKRERIkSZWQ~wvHbb$pmwtNhe$Gq6{ -&ft)NPuUv8ucz&4@?x^xviU$V=y(V0yCzE5xr-KAL!^x>93A&)}-#pHknPQ9Ew9`R(iPdMf!rMUZ1#M -yPy&H^c928FcphiO2ROlf59;4T|8He!QFWNraznxM$P-axF?IxU9pqfsnioLpS2on#6-{9F-OgfuTN}J3&Gq<$qT$#+PxMKP3UqZ@+O)GL$LI#=+ekRMN!ciYjrT-1e -gUps6p*HJ71{6$5Zi8)afl76_SG66au|fH|9s`DAlCUFF*dF?IoH|bN?|%gQGFsi2JcX#9&R``aUMCj -`RhuqKz!W4BGqjJXNpkTzOat-Z8QiZ**;k-W`Mq%z4Y1=>2zcA7wZOe9f>L8K9RRynt$_|9kH@k{<2Y -+u*;E$AN%ad9{<@j?T+LjqlcFtGFTKJ{0xe|csL&30}WH`1!EtU8AGk(oW60}I^_DWmaA -kwZ%in%lBU!(6L0NHCg|_$q%7m>1dBvDZ{{7oHID?U1A%YdJlD^#e;=(r$EWBXkxhXuDP%;5Rpf?L3%Sf#z!Y`uf#ogSNQ%KU=&`uK7QLLP^*n8_SML@zRRvass|EypqT6Q5D!uuTSE0Zq+!gc;6)%Dn*IT@1B-*;4Xlg`#prCk2Q -^}2BcVWao-0Ov-=dnzp==R;t}B(GddwA?y*r)r0?M+4O}Bzh@aOL$9Iud6B(VWO_*_03rbhCW0>_MuK -3ZfWhV4E;Q}X1XtdA3d8b=i*#d@rXHKRjx1F#9K9%1 -ug?APOZ5V`(z<5UP0~yS|Yut`c6=ngV0qJ)j#p3O-5G_P;Y<^<=Qjl=fD-NH+dz@0j^}|y4d}T1y?&N -QJ$@;inkr$5nLPt?=FuVGOqMd*WD|5vxLZym1p;CZa2T;LR|&3))g0{Gxp}}lD)e)W|u#lUVQO|ed`5 -$yTz#<5rg@Blm>T#R{&iHE@i+(I(sMWlmf_jE)5cv_S{DFof_EEk#jNY2b@|qPA3cQ-2=ald3FyaE4% -6(JvQx=Q)F5P7oF4t_kH8JV<}AIwdV|)jqwKLb(ps=|ib~vdy87M< -T&_@pWZ#E&oI@ueMR3D-HNL}@Or1>+e%>g22%Zr8=KzF@JZ}BZ!=&${Ba!i&U2!l37U_7KjVlCo?d`V -w<%Shib{a{<(*QvV>0~5|2IT#)7Z7=K(_+h$MoR0jMZ0q0rK0Wr!;ab5N3}F=Y6M6()&m!TMN7OB?*X(cM -wMO%}020|(`jYWp6lL*U5=Aw?7C?>^d7Wa3F0gx8uOJ=|wmiZcWO|Q=;w-*Lv%TIP&F2x|1zjR&3$am -NMSTA`;F0bF}n4L`4h6|hgi$`9qR*Y -9}NIO@}90nMi!%2y40ops2*ZAA8LGi1*jk?n1U?7K!}gXf!J+pEXc=;E&)V8XfouFGBp#;M)LhqoeJ> -%F>G>PE=g4(y9C=5+!o953rt4UdCxMw4M=^qRn?u8J`uSo@2CIbqFgFpUvg{3ZRhYc4u$yD9!_@ke(} -uUJzc9&neEGPK%~V>0XfXI0uApA@%IXB`e5UxNwH@~lCvjaNua$Lu#D54!U-kMsVELl>19x2txoJTQ$1SJ3-8Id6T;}C6Mjo~PWgq-_X$1C2aI* -$Nu>JxurWL`!n_!l(x%8~)_ZK4r8=(a4-5}9ZznRr=iafMfEzo(}_0ACFLZa{9YJFg|azI((Psf*+eq -rBE(%`wfN*fYIo$>v%aPOpld{W*G+&DFLLa?A78p#I{px1S<``GyKzjRjrYE4M+oBxO+CL}0vk0N2>biutZLnoqtD8vjB%+GS`H>Z -Vpu0qsErWTi41BZf^{N++6Z7B&%fT|U&qsL^6=~SLB4C@1t9LWgy|bM`}B2q_dLAxJ3}^N*Nv}3_QFS -Y>L6QZvqg@X=4!^hDQa7Dt@(l^)OXc={k@?6!;3!N0M+RP@CF9bX8RPQqizH0k3i2Os?o?s?+^0>x>D -L%({jYTdSd!`A@6-oAs>HUA)l-g@(6!YXSIw?mo5`dT6g8Xe7)c2Pxu&L-geo03^56Ptv7w!98wyP?` -tAH&6}|Z+@wE))kZ9!I0vrc7>V*$+6ZM~tZjya3oN4-UB9VXEnQnZ51O%l2xg&~hc4#uF<4Vq&tq?i+ -v!eYSv%aFc@>aC?^ETaY{&v$o+;nsI`Q}}Dm}a$Isc(R^Z`V8*>AFrLok+1Q -F%gq{o#N+jjXO3&92wz)YMLkLhMLACFsJ37UQeJjKK?s{M&hUbaV(uhM=3--6L-+HumTlD9+1q7>!m~ -cK>5t0)ocj*KlFpvK_xJLt(fQ&n-Gn?9Tv8rytDOxP)h>@6aWAK2ms3fSzF)k(!75F006}R0012T003 -}la4%nWWo~3|axZXUV{2h&X>MmPOi4pUPE$o~PRk0yFc7@wE9Tro|Do1`f{0LmAh8>=kY*vf)!#S0=% -u%LF*DD^3YBOdUkMGbej)g`=j=l!Y-OPeO9Z33DkP#KZoGCo-i^#JDoX02v^2Qjb{(htJoUq47%vl~{ -*=>v?lDTS%3(8>qA+PK2PfRobP!A3uIYcbcVAFT0|XQR000O8%K%whu(zWg=?wq?z9#?xBme*aaA|Na -Uv_0~WN&gWaA9L>VP|P>XD>`iLq$$gMJ_>5K|@GI-CA37+cp+{&#%B}W@qdfMoGKv?zVkWW2aR&l{1p -PnLa`!C}BeqEI?Y;{`x%^+$hOT_H{j()}o1ngLC=LcMkMQeCcP=%W|*82VJVpsaG$&`7dRiHl6r({rt^~gi{hY%JD8V_T5naYtcB(D*P1N;8MkOnrQv9jvMM-OcV#n@DIf_&X9m&eR^#sAM`!BcfZ^q6LvTt3)(^Qyg1i` -hp$Pe3>)mMS@8Fh3W$gfgU@Kjsc10nPxACkM{a3r_mT=;T -SOi*mV9_t*ih@9;I&Nwh8&4;ZlKUiUC+kA4)RfLO?e#1H2sJjV%Q)Yci`G2COY)p`fi7aRmoZPqOV{X -IKc(4q?s<@y$iwE7tux+W%VOi@lgphh?|{80Pb2=C@Ro4hcra>_xZko6^OW*SNgSf+QngI&& -jevZchr|L=i7vj`>-g+3#-ghM4R}HILsbhYsURsJ2v+cZCeT~29@zWfLDVyyV4d@dt)Y7O23Y++ZVz}~5BuFxB8(IZ_e@^)#ONs$e-@ -w*QfNeK!0XsZ+1l~c>MVga(J{4#q!tpa&5xh+0?!{`66BR^d+8*P5tH78wnBsZz@^#f-%;K45UKZE;2 -+OcYHIRVT02mkOg{^ETjt&;rbwobK&Rbt#7UV`($xvvg!KGv&`@7=BQWUW(-vj<7od6MagNqxWlprlD -S&<64qZ?G~atMc-J1Q(8CDlm(2D=|CAG4LnwQ{2u8H`M#AhR~w4LEK&I;!%tV@r7HY2+tDyi}UhQ!v8 -xtRWuZD6ZG#XhsLg+e-Txq0c6E6eyQ+de;&5Lnxw;$gme95y5J<0((o5=jwcRp28>mIvzV@;1?-RZa~ -Dx_GDAB2W7Sj;t43DA|wZ;^L!h&iio-Yr%>TsC^?L(Bl)(nMbjv-JaP|$HhbS-*5Yvw(m>mY0|9~H5@pb+)Zrmm6aK@m_5z%Tz -75XIPm9TRcD6J^f|31-5{&;;zGu&eC*2HQ>3fz!##h+ST_vXL6>&F^N111pk{+`17rJIG$=q}7z2wFU -GE9gWhV_T`Cx<*uO(Hl6Yn5#SFPNo4i3>~ujm5O=t~+JgE_@}|o9Ee`04Q0wMmS)>1BFXRBJFs7qZ2E -AhyCMF1=oluw-C1QlNbII0P(xbv-cl -2az<9K*4KKU!j{Acm>fG!g@43eBj`l+b9T9P=eFuqv(&n4;*t5G&FJ!dAh<5PZ?PAPxJ(ilN}Uxmc@w -GBl^HTX{^#c8WrPgF0}V2_C(`^2jAf8)35%K#t-F84dj3#ELv^o4JHPgJeZ&g~%W&v|6Ru%e85sC*t} -Zu<;ng?4XkB!)Yh+6P3}h=Qi7u2RDgIV;_b6I}dd%h7h -z5y2Ci(?$Kr2BgDm!Mw!^U@5mU6Pj)jlz0%5UI$e -W6cOi+jpDc$_pcm#1lpHoQ1pTyi4J%bEVt};Yup~~Pf(VU!dc1VSbPjX(Wr$(!;5wxQJedC9&o~`cgm -+x+8kbF-A*&EZ&?}3y%+b7Si3<54>%9&-!Uzw=TdP?x2tfO>5oB88FWmxP8OO(m3vW#pK>st1X-Tf(*cywb;21enY&a0&*d9 -ioLgB2u9r=@d~a~UO7%&m5P^UWkevkrZS@)td5pg4a6ShOGZG6Wp$Wa;a~FpAb#CU&-$Wj`m5xI`=pj -<`iVS6{j?XPZtm9-hAIpZUD*WAIzmrE~CMBruE%Y^+P_XqKI3hn2^;04TAt}>3~G_uwS%e}zh(`fI)l -@W`{?utx(L7O6_O-;?>@tIa6pU%ZpB6+xVHqA>4oZVJUs&Hi8I7=ofjd3&?%fwFQq!XAp(hxy!hN!2UV6^90=Eo%NYjxQ(9y~eo~ -K>lp85zEa?{JhvKH^I1{miNEjJ=}{=i}iZ3x?kRI#N9f-mcM&17OOwRKbNZ;pgWpxfS+(W>e5Lx;YyW -_NmB;VoKyy!Izp36rlN+(ox*`|bMOub}1P=jDgx{T~!U@0a -(h+YNtxVBuwaT&%%i4<8n5@$q5(@osa=)s}Ay8@yQn{d?F{)2##D%y3sY4Hxj7HNDkH&*{K5L>In7@s -|)co0sym(K+Bmc-#Qj+7YrlQ|j0zGpPG}R+`4<7p>>bEXUqI=I}X~6c>Haa;qD?Z=`TG`6obcTZ*ZFI?+W5k$|pXkmzGsD$EpQ&p7vEp$fS|F#_|+V+6XI4f8jS3j2~OvuZ?tqxb{i5o -x4#ylA+CRnK$>8b^d+W~Zc!4Vr%6#327V@tJ(I2(pE2P%-^n(0v`wXIXjgazFyrdN=v>;CkSLJrXvLJ -yqqXU_%ukL1lc6ieW|fDi#rfQLd-P-@9q5P_NH&dR{*XmuMjN9p9Kg4$8?39Bz6ST|zfem91QY-O00; -of09jj)Pzy)90ssI&1pojc0001RX>c!Jc4cm4Z*nhiVPk7yXK8L{FHA{8MNU&iE<#g8ol;M4n=ll=`% -^sa)JiDRcHB;mae^g+nQfBhNP#q#2(G}i>bKvslc?!dY0@|(5Bt5}zXx;AZ=-!Wdbn%uA$3FDx1jxeI -C!Xs7VOa47u(Kt4~-lAXdh?q#%V!l-$NXnVC<%DeCgT+A*yTbbndclhzRr4Aqq@x9-A(nJldfi-{9cJ -)0Dug^+#}H{CBe_C)e78ZR!A!Ks|Qw+>Iyeeb>VC=uiw6{#bkb=@6vv-K!m*g1pv-p(zfXI{$+Zz~|c -nKFXTF9jc+)c%lE&vAm> -dWS1stYol)N>a;`Eis{Tt3oSj~wTGz~{Zt(SWYhQ684jV=|)xH~T-B0h~$BhN#TO(-DWI}_ME3>1D1J -)7;G~#iKLe?@;(M~1skqdhOcxfa$)vS$V!+ZaIKJxLP?botjei^vQnkg9MBd? -g=RSw8=fwZ9T|add0{|jG|xYE8MJ@!xa263E^|&I6S_==;>-j)?-j%3P+y*aE;$pS;13+lP_<7kn415 -pun8|oN;hUk(g<1Hqw^icq6ninn#Ocx5v^iEQtbn1$RLKFz%Ju>u0o(ODH4rRZ3Ou+Ky%{}3ck+8ni -q@*H8P;=gy#3SE1^S!IMePZa#qoD*b%BCvAp?WlekF`E1>CC1lgBn+`Sf;?<7XXGC0S*Ain`nO9KQH0 -00080LuVbTYvn?TrmLv0PX<*03iSX0B~t=FJE?LZe(wAFK}UFYhh<;Zf7rFUukY>bYEXCaCuEpOH0Hs -5We?U4BhK)yB56&g1Wm0VG*I?wPc%VJD4URnJg6Xzne`~^fctl_n5~D-Y4=fkV1%62BAPvkJ_Vzpd~u -s34zhg9?Xy)PgeA?95ZbVALu+8U96y~wrdbtCfakm|K(KBA|~+ovOB!Chijltm@|g6ckIBSHJ;eGCHf -7E}FbmNa+OHlZR!UzfqX(LzaRISd1*ZIG69rnKAMXm)f8aln+0bDSybWYRtQq -^rmz#w)(viJc|O9KQH000080LuVbTMqBz -0ptV#0M-lu03iSX0B~t=FJE?LZe(wAFK}UFYhh<;Zf7rFWo%|4UKaCxm(+iu%95PjEIu#S -(#OMawvqfk(2ac`CwDh%;Dki%nV)Eb&k(Y2;(u3j7kMEaGc -u(S0t0^4<3Ocl!P%=irn9$5|SYSeDexzfR$>qR~b!Zc~eOQoVTYO_ISuy;UVKRW1z` -@_K=9Snwj3M}k0HVdiU>;jhqxdD^FR0g36VlMIrmgXX&nx`U2<-83sYtQM7_dm45ML&STIauf6GOZ#Ni6_IKSqa3Gv2O>{a^Hi;7O3OUd1&&^3T+CM(<4ebJir^#fvG_q_9(jM!IEP~ -?Q|XT!f+ItNjp6vS`S;9P!C2^H{4ec3_qybF(kbS{KF+%pw=Y!FZC+<;w)Q%=^NQCw4HsO%HotLie~8 -9C`*71c*fpMA`}9dB_|V_A4}b3uEtg^W?UM}cb7#L--MtY_0c@gf{=Y&dL7_%$lu62(N(|Nxb`k-@lj -4Hw>C{trvMs_!80EX!QiGcKWap4vq@rBM3}X~b8PZ|S`lXi3jfEMQCPJf4VQtjxxt-yW>#wcgA%tvo) -B%{oBUBNn<4DRCRFB%tg-s~T9Gh&#N}Tb~?%DN;RSm3B5<2H0GO-)-to7ZfKiLrBth9NDM(($VJ%-D< -Z_r>aFJ^`Bh&u*ed0e)gKTXL-JU&H^b|+NXH)$&;59MIyU3<^wez;_v5%6_~7lLQH%6_zh$WQx; -q$sa&~hRY{lzZyu&));hT8pTk*P8e%II66@J%tlz3=2DDj5{dZ{juhJ>7XV@wVy`^e4~r_ -)l0K|7s35m@~|331K=H{3o`xb;fv)H4;xg!38z~f2?tO^4%Knx)QuB{U(!XtMSS%k`z_l#th5xf@>(! -%}BU;6kA|kLY~4jTVMiM=u4Xjrf|QOfv>@_n9h~Ox{_&#xu41%)5=YSB=+O4V_9lgXcV4`g}lLuAy=k -*6KUgyUC+8)j*a0zZJ9=k_LQdGO4pK -Rg|iLsR)lRS_Qn6PXnQg2!4RO<(cPx+pXNNN+_cf=RZ(O0|XQR000O8%K%whm|;M{Hw^#)d?f$?BLDy -ZaA|NaUv_0~WN&gWaA9L>VP|P>XD?rEVQzVBX>N6RE^v9>TIqA!HWvS`zXGGov{F~5CHY9=soT1)oqA -TWJ#pG@$LpaX5|U8!Pzf-$HkpT;_&t$w#GtGFCq?&7)Ca -TwSYe5=CBCu=aQliC^j^w;pS>#Ee(=;lNq~%j4iYK)x7aw_+$YiN{3nD~s*7#8n@k%5j<2ubo>_Qh$m -pm5X+k72~5xWt8<{}A1&q8~?h!zVOi2}jd#c9Nq8sXt5t~vhB&d()5p1tP5N+y6jJNHlCy!KCiym)yC2E?v -ld7@>+t~Yv>CT!1pIbtehO3UDZMDZx~irp)n-2)d_Os|9mTcjDgUjy;I!+@Q{>-|uv#B+}^Pxw+~y-+ -Nezvh_|B&Pwv9^s;F$acP~WPR&o#Qw`J(*&ZbzVD8D3?H3s>;ruaed$^x2H%MFvw43=gf6Nx^w~%eg@ -ybFi=o*RPk~r#Hn_Su%Q8~#TX5zcf8|kbpNADDi-l-`Yz;qB`H@WWC!c5WYB3N`BFHtLN1|eF4s*09p -0O3r!ktBwa?PMP7Vw141q+3SVT+JiBZP=SC|e00ij1WTvy$Zr8LGpx&eA|ARbbV6t&Vng6}e2{g=wHX -&;td;yK*IA4EwvwUCtr)y{Wej7rjz+HHB2Pw*{C9cxS=}4T&z8RK}i+IO;n*0ViI-*F2DV!;X&&NAOP -frX45YLTO0GW=@i)cr!KcI2c=}G3T5Ec_&%iadHvyrD9*Pws#$~mQQ^VY{S8hyEjTi!!6U0hc6C4V;T -P+y13YTakwpq6@Mpum?9ihf+3&JWm_5}#?F7i^aQ^jO2p#Vb$-RM|~y}2ho7xjJ1B`h4zJ%{Ty*Ts__xq!2h103#(s~_qEJUKdr1oBppXv)ynkLI_T$xnCyK?D6uf_#WH -Uk1Md^Sj?$X)Kx%FrZ=iv?REY6ncbL(~aYJXs3Gq*g5Ee*grl3b}X09tBGTf|X9$Ly|tS#}%R(g(#Sn -2BKId)o~M<9=M5RhLoppx^`UoBgMs@!mC$S -C885Zj+51pCvun^B_g#Nxw`z%bk|S>SPCTnY~^yIPd}e&77P&SVT$sd|Y-ED~FQk!redvRu(O$_t0wY -an}8+o9<6XD%UB&gRg|0)_$`)37lXP&J{vpJp=;Ro2%~NCaz%NaGPomT2R!!P=n5d6qs1kd%ROrHm@y -k`{3Zbhg~kO}T<`!nR3O$cRaZoC{p@jMQk6ts|~s4>60=Fpq@S%A-qYI-i^Lxi)ptZs7jTqjbKz2o{r -YJl>lx_7~&HcpU6PW4i@TZ@qaon>;_*nN42oZ5J%h*v%@<;RAD$h7%=m%KFy*rrzXv-z(2Ig??019_* -?T@t}WtB~{%{V4l!y2P3v_eaB@)ahMG#g0VRdtq%@8&ua(qk|p=sG4g3AF{b!wA2~;?nDgM_uQ`b`@&=}cR -TTYlP`an&4weB#a%sbHyk2V>uZ$jOz{n3 -VINt6I+}!2L{*G{N2b)bMX4bK?gmc|BXh!p*_9}vXuw#z6N`nB@J(7^D`f+_^$R0UNBqZ^~$%GK~5}S#X1P|JzSf{F?$zV)a_leLC5H?7L7u -s^ZMM$7+o86)fw5#bjn09f{@CX|vN{s$8$DPB2D-v{lP)BM3O2ltTQiN3$!HQ?goDAOvT2KS;#a15ZC -Q+2(Q9+oV#^?;S1ZrDjc~V&-AP8Y(Ojd4Wt_zZ=RRW3-2_FZIm{8LgPDaQeG9ufc+o^6S1r`gQz{Cps -BXt-v5n)k<%mGSVk_BDw)={nz<209a-w2&s -88L(r!oUp>v-~KOfCIZ=48ic~IDj1sYMpG{t~p#0DIzxFXXOea&YoLty%JUM#4$ZEEA_psf_cjHbG^D -uhdpt_eYtAJ2*mu^LeY){=>(cv>^k3c?f2)24UxqJb9op -Q@i$nA`;^ZTYQRE2eGqbZlyl7SpF$@pUFeZ8iS%2{T^|z)Y`0jSR^YZQ=!+(2bi5M%E`lGrD4(n -zg*9K{B5bDuGYr4=sds^E8RnrDdM>I8V1rtn)R?#!pqi;*~5qn>|`$MgFrOz9h9LC}|dv$XU-WWVY=t -ZOBX|*$K2cw;d19F{hSy=9i>>p|bp(;p5;tz{w*%dqGl%89WTWw{@4=-D|b0!q`!KW@~`$nX*WE -+k(+Of^;M>@_Y1ma$>Ofj{GTwvRuje0zO%%?>7eBlB=L9+wlTY7S$j18(}_kQ_Ce&({1!ZIr#+aJVg; -g1ze8wdnlt-k)uyn0`)YNL~8_on*Na6_$Wj-W(?D4(z>eV>gbnN&Tq3^6y&bBLK%<^Dl8~_%jcT -F`olUNg&v*+QJ(a=U<&ozj_*n--%jrAysI*1kl`@aIf+fh#H7}&F0zEJ0?EGdp#1BK~`9n={4tN< -n-kwBZKEr~}i_2gBg{ -SLsC<)&}RKB`l&)MTHtHuj#VB*8emlkEgSc~%hsEOpTfn@5CJfGB%OxsAtZ6pVEBnJ-1WUs<;&_FQ({ -0CQe6ayAt^Q9VmbYNV(sxvEP$>H8F308#_Lc=QYQW;@dr98{1yD8^6wlX7~dZ>lWXv~R9O|7VCp4)CSh*CwvOn12(E!Odve#xB%4KymPx|Cv$1zu`(&HcTNL-+R-QWp|NY08u -Pdpn`Nzi-mEb>6hpW{hfv1e}&Lu#7TUi=xHfl2pBUnAaq|Q%R^0rfUn1*a;H0Q&>xuO+_`UAXmP0zFI -)~@v64my;&HxQ$W$GMj=Vx8l)XDs(ej}1DvAjd&~UgOeIn)+!J+PpVqU6JFkN{~WZ^B -uCU*!%LWfqOCWq29w?B@ElUv|;BL>8A*GR2>W}^?shPeM^^9W|iR^=f3Gm_!iJE#{9`QgZyK5+dy?!O -<|}RcfGdxwoN(2wTL -1211UfMOZvakHdVG$b$$gPl+nqwHL-bL}Q6Nt)0QsPW#yA~BT^B#rg7Icxa{mbMokhsSfG#nU -oIB?_n-ln65xk^hkAal2D=gBcCe`u2wb62lQe(s*YZn?=%xsviq3MYn{G>lZzb0WciTXJ*NNrtNUkBh -ikwP_{{v7<0|XQR000O8%K%wh>DV)-!2|#Rh6(@xBLDyZaA|NaUv_0~WN&gWaA9L>VP|P>XD?rEb#rW -NX>N6RE^v8`R&8(FHW2=jZuinbYwBJh$hK41qV^7y*~tpLff_wl9%KW4k5Kj2(z-f;?X -{MA2R8@5HyWf%l+^cFgV+hForkt}Y6e2r@VFOo2oGqT47;IsSQkl}rb*IR(?VG`z3(lSC5J(=k*PLbc --rrhWSDDuU3p`a8z(G|Z?4~V)>P=GHFt2)w?A|o>M;G*qZQuR8N3%Q*0()fdIlfx?;dr9LpVi=s%c~d -g{`jNy?H~!BuOeIoP+WYy3ogXpX}%;UnD@kR{Uk`0;KkFDD^l@k{TN*J7FD`3`#kGl9}FjyTy;{ -6l!#7JvIm9%c%Rh^hBprf@H={T#7On8rKC|f`gs`p~6@PD9BlzkmFF^OS5iXo7k241!}lASchW -y9)O~$?%9TAXQ1wOpcThHConfAQ(uJa->0^ -g&2bhu9z@-SHRom)RxOZkS@;+wvVP)HEbzR)1u$=X-eMXVv$TK3|CijTS4~p(^Yx(Dxa1{6fcU);&Pg -OJzVa_nw=FbybH&V5QnF?ha*QGrQd_;@HAaBNbNSwK{#lRJA+1QBIcmktjP-^2-=Q#H5`FY6?Pu7KR; -Ys%eCP4(Bbgh+9rLLk`X$=c1^n`=r2KK -bRX?J++S$XCh;+ZEAGK2)1d3D2x4z1X=-*lO<7T9f%yvxlJfiPSTN103tpA1Dry0)n4r5Qw~e&aP8emk7fLY=!)eMmq=n&dL8d~T`3dqEc2f-G -qnR9!G4fqNG@eC6+cukU!*a8sEMX&`W`G4rWV4B+-fYfe`(i&cg4#phawnu$p0;#l=_sn8!LK1jmzdm -cPmrD;5gNganI|BBqZ&UX>=~Bli7L({MLjXlh4}Yn9i+wLD6XkkPc{SF7XaUmeV^$*lqt&10lxAPP&^ -Yt7=~&@xniwEW=y+yDO7b{vqJAM8zU&f{0!z7{GXpo{veSQ*`(9;%TGr;ZXTb@Pr;s^eB@{P^hoG>Wl -7=lcgK?2U0Jpid0D36lARU1RxIe2gzv*pI*wl#mC=pd#l&zbS=8ff!RU_p%_5dW474%pq)=j{@6aWAK2ms3fSzFA4=io63008YL001BW003}la4%nWWo~3 -|axZXUV{2h&X>MmPUvOb^b7gWaaCyyITW{Mq7Jm1y;AkE!rK*8_HR^Vq)R`LV+F9Fa7YG8CmFYx?ENL -hyX=d8}?R$8UNJ^A1!CY2AVu|EAJlBthlpM$TfRoJ=jf#RKD8n=-6s23VAsI3VnYUe1R9BvJ&h00NH&}l9u*}eV2ea!pc68y9-`bjZ+_?0LBO>L0=P@SOhNF&?en -f5eebQK_(DuXI0ntT7~?1>pkOhrd`LDEN>gR1PZ&6Sjfu-F<_(YMER`*&7@#o(&0olE+joSPPQT{2$o ->G)K4SVnDCi14<)L*`>Fo=FZ{?6_L3S{PN8wFP<%1MHsvoPt<09SF3>Cc!=Nm^2bZ#rMS(Fz*Ru2;Nj -_P!(xfq?#6am`G9#<@#H%lOftYsc`RU%0vv5?{6&7Vpds|WI}pg13*7^b;ljCpGfBx>M}3ot~Q$lrOU -fFM7k*MG_djF%)VX&yZSmXSOJ?JDGBfM$cfKTjAIY63f{X6!sIX{dom%N@c6G6sV-gBq(OD<9D^Ci%# -KG>-)>2aYy_8VrXD_x6dx;97`OcR=)gUcE!p_mla0dO4je`J;#peyfcQHUw6SI5&9JG7Mlb$;%yaozE --B>(${WARFdUBgr@LS>e`c>Zs`<;6*RljGPRI5BQF7HV7ApEkSvK0rfjpKOvG=&keIG5e5aQby#Mp`= -~OwcviOd4zqwf^OaD(%pk?mV$*K^(pb)9{@ySSDBo!YDTfTbHcl|3_(Dvfo<-d5xT|G0CDnq^@9{7!I -CesQaOCpa2y)vQm(?6#wDB5UFK5X8yfOv|Dc^bT>M&|v-<484l-)MDq~>*u0e>*B6bh`Q9%@et{9{eg -Ru6`I5*6L*dZ%l}IQ8_k-r8>TYH3G=%x2yq@^E5 -&w5jip7uzzVANXm&DpAxA7+;MXK&8;EFw4)FDS&FwmdcU!Itq;+dqF&dreK{9Vaek!Zr90Sd&N?4mG@!P2e(ir@Lq -56{^XDAlds$h{VTn`?nIzBKqLg@Sahpfp|CEIXPdUR}1EeyOV$cQ(0P2j|Q2#lI$N^pt03vM*qi??lU -T`pbuUkhT`?bCu4wBjcr7H@8B|r>MBsSAisq0iu#SQI_Re9cc1a?S`CM2c~J6uv&;~COfPVWE#KDQfs -ZpEp#$4N~Hx0n=cz|PjV=y1v#S)JB{I3?s5haDj477b7u&F%L;b24$Vlek6@}0)C&-v;#cD(NdIB7nyz?dE>}f>(Bzpnh=kXM -H5SDR#BIA##?>%MSMsHx)M9V?)IP;Qw;JY$6=!udxe#XvF7po05ALV -B!#is%GB&*{|8c-*R|*v`J@eGky}uG}Pxorn%F)9cqfe9NYPy&QSIf!e^gqFRIiAj5ibn0yb8v?8a=V -Sbp%5o1P2y*OC{%*rBBsBLMqgNo^W`V@n-PkOl;(}P#wG6L$A<3OoIL3Tp4DzxERlZ+4 -y`i`=hNhOG6yjWps?clKAhpQ_55O@P_!6Zd=pJHRWw&3n}-Wh>EG4Qt52 -hij_e?ps5|M6InRYA29ZHf1!R{P3w^@8-+~iI5g7U$AP+RfU$Xar6v3YvFFObz?TW!P(0qo56*h#fiV -Q4pZCqb_G@!So-x{BMTz-#;mpF--m&=HKbjQS(lEZ%R4U2o)VAu?<@1zRh4?}9MR4Sy4E#2_A9pxHMb -qg=3}^jDgIoczWy%e=ZW+dLWYSHwFss5N|mQI7yKIs)%5R+{?hoR#XVaeh(uCE04TXT9=XBrdRefbGft1BP*yY)}w<#c>Ln*c=o`+BjSybo6E<#hf_^NguoL -rOx+$wNld5Yi5cut!rsz1)HOEXE<2WdP9O1ddl_g>5!Nwmnm3&&7SpO2dv>ZJBp`WxInZ@EsYcqX}H) -?W6oUfmO137g-`min10iDchPSGBntCHmSlvSxJ6S -dl5`mhxxte$=|^3fX~{RBTF_+tQM^ZB_lm7=$O9KQH000080LuVbTQH;Yl3)P<0G9;-0 -3-ka0B~t=FJE?LZe(wAFK}UFYhh<;Zf7rFb98cbV{~k96Yk4ECk4kPC_6=B(xtq)1U&%%D~o8lVz^&>Su3+`qh -3JayBiEsNn3jTKr2Gk8oMV&E$CWv{#UOHBehM8CLMzpb{r)p)f*?pGw-(ka7fR4+f2wd9fEyX+z?efA#v+R{pTpuAESdImGYnE~6oi*>7<7^)w`Dge6pfNiP1 -J1Bmd(3)N#}G7+I41@pbuY|l7}6%mLu8+mlk^nycvPtQxU=nWr?d4BcF(&i6@Nti)O6XWHGtPy{UT#E -GG9jBq%J-h+a0_V!=uLU{{TBL>pE)qb`r|)YB@O|Jim8J=wa&ooF|_&*|0;33n&m=(m_|-3alIc4H+v -kd91QY-O00;of09jjzt2n+j2LJ%R6aWAt0001RX>c!Jc4cm4Z*nhiVPk7yXK8L -{FJE+TYh`X}dS!AhaCx0rZFAa25dO}uxM(^Pfw3n2SX$CxA}35SxPaSEfhcEm*s8@zb0=aGzp6v4<>a(cT+s%RFkkHNH$K1e9&ln9U)qHpaHp>?S# -V3#sL-V_EcY3F)epHY4`{c==OposPSkJJK-RZOVk(^vVdE)u_Q;AfX235WE_HHdNd(0V*B%*D~(S-z> -=qasEc4YpCBZ$*CpOM5#|ssZS^v5of{G#(94w5qIa+;M&z#+PMQuaJcc_Kkr=)XTjuZ=ubN&j8@qkQ; -i;zJ*f6BX$FwxNL?<$=|>HZp?^LLE_)MyH2dgJ2UDj*gtKrRy$jA{@ZsYTezV}K$ -1J3?X)|r2DmMzDYHE&?8UD-76H-BNeu6@DwU=%r=bcMsQ5YF}XuAn;*Cs-9 -EQ@L@QKRLFLbfjF6>_hm5aw!Hr0^skAY+xc+Y_n_l|;!TG?SJhSO#=kbCh)eS2U?1nvHVT1Bh;pf2^_ -A%?alng$M#uLBat9#SSarv%*z0*J6oc^#hk6v)PY*paxPo{(M -D7c*X=Yv0j*`znn$`{_ERTmnL`@LZ>?49|;s^kf3M=K}p&z<>cu<({8rCBQZT^G5z<+DF}uc+Sv;i>G -us(D9WnJAlWUr!jPl5uRV`G+e~vuk^(z)kpMQX*)HLsn -a(s;bO6zsP?*uSFpu%9od3LP83z&Nj)I4(NzIOxD4kzM!|GI{rE5;MVb4j**j~mEm@Nndf6N=73zb)0>>v@=|jX;}9FACkzheb)=p -ntOY(>jt&fH9VAKY~w1EBq}i1^7l$lB3@qTb6p(qVS`x-QlW1suIXZIoGct=QMvPS;Uyoo7{ -~pQ+L(T@nL8l)Q)SxoHbpV6qkE4r==*;==^~?~5^924UQyB2!m;yX(}%V03&|DDN`*w+Ji!6SeUNKNV -ni!5z7QA%!3BFZpHLJv@G8YOw{C5frDJuo$6M(%avl0Af3`9H?e-1&zl;vK=ih&9-q>r@GaeW%I#m7r -&}z)#E$!!c-ByiW403G?-)!6a;2)Ud$M(yr4L^6Tm0TM#{?}q)uUfI_ -QpW%@GC}TYO(g<*^!~3;eNcV&ptG}h{a)}>Z2|HLKCo7)rwjrTSw=d7Y=3dsgya@D6k|yR`fpUo^0}n -eR&PJCy#nto}Bi9QUmm^6{juPtW~Ex$NvCOO9KQH000080LuVbTZ(}M@B$A201+<$03ZMW0B~t=FJE? -LZe(wAFK}UFYhh<;Zf7rTVRCC_a&so~z{<)loWX9ob_kb!TpFD2T --A81V>(03}^i&Ga4mhJBLmE&vkzI9cta_d_DV#bUAFyTBehUuTM~vRp9upVmBQd6tMmiImkvDk`Sef@ -x806q74@IVyPq|K1z08T*^4R3?kb9(x%b%os0H0{he14-O_`XF|~6`;&Lar)S4@!GgvrDYd~36jY*=D -$}ys0=|{3SRyxNF0^2&&^*gkG?}a_xnXgsaCx>UO2i9Qp0#*T{ZjJ@M|&M$M=`1=g_!t)tB-Yf&9(xtZDL_nCs3A4*^)&uio;uQFU>909$>8=A_vVzEEqrf7LBitM*`)1Bi=-?_VMEVZ&0NA2rG5B;`++>w3Hlom{u)g~aUwNJDShSzv({eKTV3h*sbHZFKK+INp= -Ft;84lI|;8M~A+Uru_P0W4M#7QqH|H07EHGZy4BxsFlK@ud>K)>$QphUMg^qfdW5{&W@}y?_7V^U3L- -w?adg?BnrA_T9njsKMiw91+8l7B=<=Or^EC)?!m?a4j&v8S)X`0@o~+LKXX9g>^|vmFY -)l=*jyLLKx~}Z)w=(rNI5C?a~Z92NQQ-hd{Z?Qek}^Ve#~5NR&rQ*th{pMuwgXzYqt=Hs%a;$ssuh=h -ubV2de_1YioBRx{BDv#ePz5^%}ljTtu|?2=W`_ggf}@|L7tXx?SN_^v$d2fWL$#$m_oQNGmcoK4a3L1 -=v4CIO-rjZa;3}{eaz=agRm#Ak}c+4P8b7fYtMIJ4d^jUq}@p{6$0nBt9Eh<@8*{MCM*cK5{1Tl1*O(qnbQ1c* -;K{Q1@-bGO*a>3OOo|(<>W27XgmYSCl#9tH(NCjR`&;JjknbjYo?8S#O5?f2x`>d!@05g2~oE53ORV| -r+;WzyMKldiYy%ovd?lBl4EWQ> -m>_t&2RnHffhd*qRw|AhPH|E|hthZQN#g&Uk*yx8RD3B(GCrSBRxuj)fUGP7_}hdenpIEImo-QdNDVkd5(si`g6G(cctBy=KqMJ2(VKO6 -_uS@Y2_?a@fceDit{6coYmg!OVpvi0A`w;&Xzr}OE8yB)B1(;_O-4z%`q9bXtp;Sq*X8n7RFJ> -|UDlR?J%q3#j|k(JTi{bopPnl)<=afJY3?ek6lJK|xowZb-gFyFEQ$(HnEm^XTh)bB-?g4%RS9va(%0 -llfPqw)WEkJqqObxbycq8<>n!htYtU8}gG4tbQKd{PCY@JdB@`<1925L0d;S*xcQc7oH<8gqV=OA_z< -@i*loNr&xj2BlQP-fnHYvbgz6M!H$oH$Yg1oN1sFO7qlHD!t6EG*u>i(wA96>5d2Y)+gMY@toXSDcnHf%)PK`P5u`>w;257uuw}P$t4HzZ -@)Cm^CjJ*9~5k=8*2{w@>YqV&90FQ*#Hs(x-+ad$2sZ4b60niO-C2|m~h5#vV3z`!_(c#<{#LHVusv+ -@G*I8-NQ6(=_Tgk8LBB9J*Ni!2Br4p0;E1``U6++gEg;R7$F5BK?veuY_N6x6i)+&bFLgHFclVym)ex -0Jyebf0kz(DrtvwUi>dDdz*q^A?jbSs^%DRjtdQDsFTuXC$ri0FWgZ&Y^5#KKu4eOe#p -)g=U^oz`jke2c76q3o&31PsV^{3&;)OfVMfxBJx{aJtg#U=vn&d40Y!ETn*$&f0G!ZtA`lh}1`z$>Rv -QkkOMRqklX_s02VMF2L*Hl4?Dxc4j*=g#jkewtIlx -Ypx<_eubSb`D6IJ4e31ap2kby6I5$p4mUC(sRdxhCO}|}1-UhwUd@MHM=FzjX{MXgCt0I_t0GFHAPZ& -7i%F>rEQL&tQ+GcW$eix~Wy#)XFW&K1NB9}{!O7aX`-B{;i)%Kp)y)3_~9?*l`1A+L6e&)WS=P0oB*$v@Bp;Nu-AS*GRu9fp6@}043u{B93iJDB(7 -|HmA<=nbg?D2g6`aAaTfBuWLdqTHu;x7hVMK&u-^xr%O1H`=cvL<)lbfz+)EXuXkrJBEZQHm1sQA)%j -SzY1#^8;Axg@t>|K4Yqum{lQ9_ErS10@|wBPd}fXw@Cqo6+_XbfEf)Ug9Dq?%2f(VY#=6OiNXkylW@S -MO-t()ZHN##p#%R$lagU%Zp>w7p>3_SsTCD`+@}3O-Mk*mXRs9! -Xba$bdMX;-6{XBjKKOazshl;-3NKB%;_Sd#aohMvfj1|8o@ -cU9({5_2KhShK4#v*<#rZ-n`g`JkLX{lZ^K=6;om7xp;0E?pB&@cUP=X~HJp$AkxCJm61n`L6D7J`*3LZz*P`i$hdny};M!Chf?YVY7p`5N3Y -EY|liN4fQaBiWQiJ$!%n{wZhqX=f=(HY^Gjq~^PF2&XOo0v9H=CKds|8Z&W$GQq3p`v9BVVebcrFAqkSQpt{FnyFXP$}X!GD^!n4U678{=zpd_X_vB -gzQh>&e)p|Ww`&=Jerp4KF|j5o{_c4_JQR6u@7S_KXvuP!>2A!B*H>>N9jNbzl&N%!ko^8!_=b5qzPtf6RoM!dSJX&izIwje6JRRaOfT_nMue@@t*ns=ALfbg~D -%-1*rbOs=qySJv;?V9U85ACH56OuTDHtHTVLQ63fow-AjVg+^%YGY8jX>C;w(=j^32C(5oW*V`h1PvT -Lh&>D>xK>UqG`AOF(2f1EL1{DRY44gCy7|skF&1W#hz#4PFQPJCn=6h}PqxUTo+vzanp=$k0ui#p&-U -!QtR(DNMZN~Ld^E6rZcit3N{1b53clDVfqjh@3Jsdh>4uXZM!A+XT?-zeEvn13V~mZu(v7-9;K6Tr6I -ElwdT_VS|J=dQYkc^`HG=*kZ429C#8!~UVZhgu89K3UF8PJs`aPPFJnHD@v$Uy^T8XusgeLOw`7-36=d(ZVejpPv`LEykJYF@6!QdsX4HtggWq;Wf|Q#hPUd-KN7s<+btuywO8!n}7T -50b!&o*cqPpQWISk&o#fg0)4TL1n*j;y~_|ITX+70wgMbjj8>S-rVBpuH3GX;SU!cpv%rh%yRn&O%E+ -e&@Gd+L{Eh)&nfwP(O9KQH000080LuVbTcMRVy;mgw0JDbx03iSX0B~t=FJE?LZe(wAFK}UFYhh<;Zf -7rTWprU=VRT_GaCyx=X>;2~mf!U&Fw_T;E((ckPb$kTO|8e4Gp_NqvYdR-sz{J%N{B#!i$gNU>;3Kfu -08=u^p%aO5|d!}>({Sezq=dDs$50W>9TFwN=>Ixwpy1}6D37aHc69}MZLEt&(tc(@_4CKx=yNERsQoU -m89Xbm8$FHQuz=2Iq-DZ_qsiJ{z^;#{mWu_|l+%{QW2mN -l~7a(Dj)cWIeRibS<0vbZcy}h&OSmqof*1smzaBwyNT=%ji(kRwLWW -}%`T&|KyRFu)8goZTRlMf5E#x?$>D%&-H_aE%3yjK0anOJ#J*Fm45`8t|VB -m5f-22Y|@R%9?aoN7+=BkqdIWi(MAU`2`~hYyR%%m+A&h*WPk%@wqM6dxVMdj{}}dpl=!l;jn#<|azj -GAmRX%{Lla*3>Frjt#ixDzR$@J=+?ZXGPsW|Ja?4#XvZG64gy*9x9DQJptCKn{$KJh3yIGz`g^bG}Gy -j@bG|*?}9Rl4*v$sQYcGf{6i>u&%tWy-PdViAww(sCB&^#>q>B=0Rbz2yh?%4D5B-yBY~Yxi6f>{_*5 -h-kS3poA6=aOQhgdn9|IKN;Rp!2$K3RyxPc9jzEVw+CQUN@HOX6bQdMOo{KUK@KS!|l@)URyxCzz?){ -1&Z_*-n&i5l4zg6C&fDnbmA`x^MnU}*X8fz`8W7QDZLJ&QzB=fRG-pZ%#_t7Y!I;`x>X=}3nT$ -bK?ggO?=ZKdNLdOH8az@g4C7E`XoWfSYLfi;c>)s5Wt{HPJhXuGNP52k7I#9p2)UsBKmgxOG+D03p*6(-iwMw_M4s?D1W#B|hTz=ly-QIo4OcYeCTok`0H%m53OE -4$gEtK>!riEgvz3O=s$7L+dSx;T2LR5YRWo6Pj;&7#|@5{c0?RW@%MnJg?dEJnk0t9uA8YxcA|L1Ekb -W{1?Sn-U3%oK4VB-r}@z7})b`Riwy^uvKoaGQbDb${N-c_&RA6vZR$GOmN14fN_`ms3(*Qupvxjibe? -SM@A4F#Xm-e(c6=^(Zk0_OfdOrDvHbsKBT?<*72&Kv((<;*%Ost)l`{_UAzN`_2eI}#&t -*q0RU1Wx-7v#Xxha!2^!RO3OE}!JnmQY -01)WB&Dcz+fumZGB85Q68Wz5=D*-GH3OgMT0c#k+rJx}Sw*2;`$3H-9!vJI0nnk;s!%!%a$}+D(=%s* -4f;4dr(2Gr-&+vIWKpQ}5={lr64+E=G}+}9%!XGD7F(tPEC75}0VMc;xo(nqmS@eTk5rE}UXT9Gl -{`QP?R|Bm(sY^S!8YHUzB@a8@%oqP^A{&CpS=SM?<0$C7H@;G`;9lXe&(!A&m(WS%$$S`Pit$hpn~5`6lp(G%u -@Y=H2)-^<%2;gmZhaJzN{sijOucL`aZy?ELsiz{!r-5&WR6j_;SMLfi1vYv?HRAUamTjwbMumu@6OBB -AV4rmJ)Y -jM4au1Ij900XHEW;hz=;0mB>NlOO^IX)oDi^~Th+DSu^iM&d3TzDlOr|sqRNB03YW<)U9yoE8z+J-S< -O>k87_{NAPi8pGLmSRpboC1`lnCJ)Ng9v@QJVyw$PNZ>(pxWzF5&7Zdzf{Trl-0#SBPMK{CskA-VPN7 -l2xub9n5ck+$zX^jNRo*BYW5o~B;b40rA0K%*e3_|m!tTuP5a_EOT*fQWg%dEmfVp)~>Ag0= -x*wNPw7WHqA5W;aZoT(~)UUJ;re8!vpbPbinL;KJ72OS}%tj6mZh&I}}rc2FD`^PZ8%?C-+=HV5J)8I -R!LKsyD-FJnL$aJ#KCa5EJ8r6u&P*l|b}ix&D5X$ZINT*JNACF&9Mn{t>=&66Sq$(jS}`y+9fr>@+^t -}$Z5d=a=%#F5rY8#6~2z)vNHPzJ;m1CPMOAk^#@$A=NqlHricRs*v=fsNKKt^k@k0NzR2+bxUSB{S~? -jv$P!ch&%BG)`;>R=h1phmh{cDHxpV|bZgCis1p=A8>R_Ieq= -Y1^IU^{478ZLi^_j!8iP_puanFxw38`bgHA9=i#U#1`r&pX{uN*yNOLyO4$;C;)eRw1Fx11WFmQQ;rl -l5R9XPjyY6+eocc*MVB6#TG)L~Zj^CTaxNkz5%s* -vb2d8|aX-`T)E_?2n97E9M*MO13=N?0%Gw0c9RD8VQMk3bJ*PGjY*wuS(ovR}uz;F#rI -(Fk5BmOPmw7Vxi1GFah%sttuzhXX@Ao*Yd_)H2OwFf!fo>p8W$0P4ZFJy-kVuUflbss=Rt1})U#s=l1 -~38DBE|BGa5s|ENV^9@*M(f=50#l)IHQ~x6JbzY_$9kJDbv?AZfTmUTs0i*y8hh$0O;xD35Q24Ks0y= -}2RxHYttOVS^Z9NuPvLI=Ro2pRknWms?SYm040>flA#^@5zGeaJkRD+ZI5F%hWzP^lIvRJZ$k*li4g( -Ynmm}w_;mLIpzn}|8qvDE_%jNycNuPHNYJx!dT_NDr!NogR1#C-dVQU!q3@mbZXyCyd`vm-=pQ?QwO6 -AZdT%mWMuZL|CtgzEss@1k)Sl+y~)`y#`J&!{=EIVu255cTW?n_?Ia^xROr7A6It2{st9K?HzIRQ@KZ -va+osHfNu|IeqoyELyh3g2V4@8Nz?)hEixj438SceTFgHqy|t%^|C6>Dgp>&l1|GpCW!VE6NJ_dL%J3 -0QIy9R_XZ$NEIi)ZM+CuW&r*h&)u>28bb(~L#C#};a&4}Haz&J~rXlqYYFAtWGy(QSxL6B8XQxCOH)f -jBV8i~pj$}~q6YUtBDrtC1SJ(tgl-yq5CRHj{EC+LV4!_a#NtP0GQ`{VgcY|V_oO`gDbrg0*NPN!C@|>{4;-@3J?Re@ShpBk^f}2+JYjNQc;QgO3x^b5nCu>UIKli8a}dp#Yzu50I)RPJUEvPh9-{_sOD~Gx#6%`{M@Z~y-4``E)IpEQJ&}}&A5pOpjmcf -%jYEnyOzsPKV2i%s2daF%s6mWDH{tlmWP7Z{pm1mKF-UvL+rc!zl{5i+o6BO%xlE{zhAO^{S)GEM7Fjx*jU+yMMz*u4QBiv+{hQnJ5 -GTcmQeqU)t2wi}%qp;+DQA~qs!)v6MR{3ZO^ue+CaB>jMI}MFpvXiVJ@2FtIY`bvA(~FWuSK0xXo=!u -TvY;-$8*6k*g)z28UVgIN-{+ERm4Np*K#6W6X -b#6=GeP)eCtrsg3D5iz7P8FF5ktXSl2DlsMSIKdO?&^eA?G;D6@*3?qR`pHvA96>yTps}+7>qv+!QAd -xQjH>B}kp=kQg2%u{m4Y7#xDssXP-Et0m@rZNXCKoDs57ibbpLrKYn2t$q*kH$J6!Xe52&BL$- -loR%@$HIWN*^`ai`y?FvZSLLcCy%QxQ?P%g} -u`goTb{2b5k#k~g>Lh@oX(7KmS~OGrpbfXt6xBVG($$4eYo94z8IVTOe;u#vON0>yjQz=kA|Mt%WDoJ -<&A2g#feiB?d@n*a1Dx#y#xsb7JIUGDFSm%b68k#pJ4n=*+%OYL;REsKy5cq7S|EZeTXquN?xGPZ&!BY -E*4NRFUxE4a%!b?R1swx##2zh8eAeqM%D>MRVV)dPjH9V_<{Dd@O(L^sFPFnWF1}^56fSBqufEX7}1r -w9IPtx1`yPi*#}A)J0TGZnq&UQxY%}#=Xr7|L4RV;sx3rG6&nty*KXtuAqzX9&}m%~)WDxWXd^waEia -71B0Jjo2U;N13c=;Wwb<=h0wm+m5)TlHh)%JtOz7L5T*vEqN(p+7jt-=&ejB!+6$bD -0V45L0wrdi{tYTU$Rj=pOei=(*w%qdr6maLf+(%NXjD{${0Ohq=14mBHdLW#i(b9Zi2^cDD5?`@r*{J -QAe~$ELXBkyfQL3$l-y3mnC}k_!OYpDbx|?B0R8 -Uy5@+c^CJ>zxAWK%Ipc5c3qUWTnGSQS1KuLk*@ylifbPl3NR%_M~ -!o8#Z+%&;`&Y{_l^ogIQ4qaL|jc0H3~VUV6z_G$AWy6v1{aj0d=EX3!VX*l5O#cP*q=E(WC$tAt%+|6 -vX0|+7m(o9fZ2`R%c7t6BaL?9TEi -5VIY}E|Gyr=T}r%7l2L!g%L9j_|z7ZmLGhOf`SLyJ)sCzM%4`tEUIv#%K9#AIJPBN^&Aen&I&n&f!kd -lT0v2Od_!d>82MK)#f3{{SwCnirWCCSl^q*D)&xBq(dX8pIKbShl)7n`%ea5@7Ew@(iX&lp39ZK{IzE -n$!c^$(x$bY6<^0EsdpW}qm^~cg)HKxS@dQgCMNwR*#k~9gqi_$%t7MQuCmhQp)MdL1UV4?XI{|+|!P$M^TY}vGQQ>1d{35zWp3&LB*Ov*#G__lRIS-p+!z<@Tr_xaozR=R3# -9c^8mx02m@fBG_5Ah*Y4*A6#Ka*^e}#gR)3VI(fD -xDR;$35S)mUWWG%qm>i_D;H0lUaFnoe8%sGQiq%DxtX4MW9SWNotf#$r>AIexNdwhQ**>JzHd;0TVK1H#I@d5N{A2}baJa~-tQy5-+BG&l4n+9@gs4JrpDz6SUX(|00C9?fuZF+< -ncvToIjgcOhIK8PX3IH6CxF4z77jJn~E9T!cifO~@o=Ur6Bn!w!%P#~G|z6qPe;&wsh^|{bYC8LI8^m -*baN_`-dLg7-yt`I@Wa+MZzobicjTR*N{qQqX%!Rc97wP9k8uJRytO8GK3JBMkmmP88%>o9bp^~GtS=KM*xQ@Fego#{oAd>EvxhH|OP`Pq -d%;V-``hnaN1g_>Zc#Z*E<(+A*4U>>jp>zs-y>ot}{5G(c`q(5U@5IX1>H`NrU9NjmE4Zw56q8tuQvD ->8V6LX?P{9fh^2@z#U?1=a8DQyGpZwgu{Q41t;z^J2$mlDEX!Sb1Tj1y@asy%3nB#L -X^4#&AOvoC~*SLa5D~GuT~HX_#7anbbKMt&}(_Yd?NP-W-;NG!^-egb5x8HwLsWIjh8kCfcJ3Z4%B7K --0K}^}L~D5^Px6;{XT~Ck4tXb3F+(*$mrbKqZj(Wo3ziq3&*6xz{x_7){J&Ugzgeip|A1Iz!U}77yvA -FX0)OW%_8a`84Qt+n#7LZ*kt%plEhYJ)n)K5!Ife&zI*0ovt6)nRrU4j~pI#dxsYKDytXP9=y{hr779 -mfMblDjUyLr5<8+Grdt06(m0|6J!BP=Vwl)Fz$9A<6E*Bl!YI%&6Blb}9Nzlm;qHk}8a(EB$8?pfhaU -&Dn+6k#uMEA$qfd7Iy)zSm9x-4AaOQ)DLfgFw -(%=V9aZ{dvJqoG4=WmxA4VWodsPEJ`%{kRG;*=OvEZNMvu0>I_HmwZ^B9Tu#u)3E=513yc@-ACahJTF -_pnbVt`K^4mGlrxxx(y8r!hZw|_|rAytQJxm51)+0MqK+xgrzPP#Iieh%eUd%NorTF? -%g9?!BW;;wNBRUSs0&ii9{ea2`81FXdW_ucc6QD&8B{wm&OQ9xv#NJ)0spqcxU&(y?^{@ -dlO+q9%G&8Y`sZ1`?z+gSgdH{ax}+u)(k!M(P|l!OBkWkpYX&yGQp(E8C|WxTwurxAPSGU>0}$$G;Y&X)Y>N6exOJ#T!j4I*5NAg?V -4R5}=WxhNle3*JH|dhDWCA)}t*4!f4b8S#2DwghBWofy)OtefjF&(i2=RM8o4o=blq@mnK^5(2Vz;HW -_{D!IE^kee8q;$315%KlKHrKW8bJY9GDfT?<)GU6I)+jl}lYB`WXFO55^%wOQYL9{S+*u2T^H5a#Py! -9J_sKiIh~pcitJ6SkvgY&zR}vC(>Vn9Lk}Nv9m#_ljgYcA$F{r^kHhgQznh+Z+IPN`3qPk8%KT?C);> -hXmfosC7DLS3l^(y-AK$5FvVFA#fh`@k817DMp||Pur#9QYL(E^gTin~}<5T;^*)h;$mt6wQ?${& -=J^1wR8(VG130^a8v|C*~UypG;(91r+nujc9->170U7iId4Is6;ot=|P6aL0RJ%4&n>Qt-TbS#c`BF; -eccNN{Ek{`iUU5@v9#mhhXYJ^;Clg7b>FD@exhM87@X?j -OY_zd_ri+s1prT-8}{`*^-k5sOb)y8SlZClhf8gtr!KnQP>tnW -L@cAHwpTwuGll5kTS+mLPfB&AKnvuMXO^D2#r0w0w{wZB_;E;tM{|8ZpGo?m;qeRE{!v)@r$5)SdWiS -t{4LynGWtq##R+8o2v=N3b@rFmm*K~2hjP*v}dik%6d7wvnz&Ax=06(GJs4No^c&-daDPPgyyeY*pAK -(Acl+LZDrvD`!JeS)-WHsotv_Ud>U5)>e8MCSBJuN^vM;O5~$`8U1#J7uby%|!}<5er~ZHmgpq6HtkR -a|ey$$ni~A_YVaWjM{mOW=F2+h0V}J(Sp+%MJ5poQaRa$J_gHxOH3NE5`GgABG{g-sEqvS8zSTe$$Ti -<5NM)p}u}yZPPP%%8dpQ-B^*UwOX3xMqrW#IsRbqQ$=N&F5Z0Gu+)9dl`43SVSG9q{}3aBof5H6KJnVX9%> -HXoYC8mr}&f(jLBUXREID^R^rNvD!aVIxV_F)putgR0o!YmW#bl7>u1=(h`Xe^Y|&}jep -2?%Q@unwq*o`c&Co~eM2~DDp@EZm&`%fGDIJi5scSE{=<^znRVK7EM*A?SfC=gvAJ3Jfl8P9b)UuOXvCE -Zv6gOrqkji!y~Y;DqT?96qec5qfEbKZo84lm(w|BqH{ytmzar -sKb^I}DYW39n!V39^p1f2$!tT=$F_Q0-&y3`@HhI__8nadAmdF%>ux_@kS#lncg%R;pG~yGrgq_=qlq -V_+-q36l41LW4ldwceG?4U1G}@MLec1gFfRrS?|}HCO{XDm+icq+gs9VvBSws4(iZwdCoCAeB9w*9^b -)`H1qcIFI}Jki3y!Y@2bPVYV -H~Nta)*Z6YJ%nFC2|{}N>^Y1v!OHQ;?0?tE;+oES##>TnT>1s6sM4T(lbk-Ai{JROGdVKIebUbOq$hT^4$09aYV_MRS&c+;=7Jgo%E?2fo8?VsD_L^NzbQkR -5W?gGTW7w_R{DoV@F@Y`b3ErJNy=eRJfuW={>-kimHw<+tMsbECPXw)djR&!F(`jPG8;ZpPIw*8FyJV -~D{3`i+#|dKkV(-3;cw7W1=c_;5!W+5vjT{G3UC#mK>&%}Y2q*u{C{?UY)>;%U8rr~|L{~_huUY%=R%fg=gGsnF^NDmp11cZ+mQIZWre$nx1-N8=lxe+xGUqnsQEi`hxcXc!J -c4cm4Z*nhiVPk7yXK8L{FK~G-ba`-PWCH+DO9KQH000080LuVbTk4G}yDbC&0D}tv03`qb0B~t=FJE? -LZe(wAFK}UFYhh<;Zf7rYWpQ@`Mw6izwKZYeP=97{6i~zcxbS -1|_i?U`M;1uvK4Bif@Dbjb~f#rRC2TT9clD+;IJcqSV3u^vfKWRC>c}Vqo}eHD^z}ZqT$sW{WFPiILp -0+gV~~Lt`l>-T8`3CB-T*O1{C_Y__VoF>on~T9mhg59w#8%Nv=^XFyMqB>sISDKj1p<4Ei%6qTZy$e} -ULD6>{g?J$+A5(PIGH}K1wUyGisQ$Et#Yx0o-KGsmo{9$+UD4tM878t92J|!s=*0^k0lY?niYnLbIFs --^6<(}=+?NumAA7E3!Ef)r-0s=~s)(2kJ``1)>8;sodWWj)ugYM%5K7Kg=Fo*YC!p^mdp;)bvDqDk58 -~X4@8>(a~3n7VjC){T2{b@@Up3914VXbyhD#uE{cFpZ6ocpCjVxcH_W&Dv?>Fs*SXPr?uMQpx>y%`jy -9oDsbyUnN2x~K3Lkpq8FO`vJ(^>oU?zGgkr=^Ntfm!}~vC05j>LJG@RYOt>IvAZn9a_(Z9!RvQ$tt5u -(@D}U+uyBv>!4eh|-IyNJX1y#vuW&ojm@WbstI)18raFZm&3oqOL|cM=x_r@f*MSf5VDerd4Yn{qN6wBD6 -ItuV1-BMU)+S~FNuCscDpzK&ipmWnCl2PzFB?*{OQQYqq~0WbxbhSTNDob+7D*sZmYTbi{# -yTP-;{nlu6gs!nj+8%N3>`o{#VK>|m6J5L_#J>HfVr1eUYI2*vUEFNa{Um^k-l1VF^(uA@=EIpF{e>u -GG?F$28cnwD_0yT{UE8r4m`4j9&>jQk5wO9KQH000080LuVbTO)hyab+R^0FRsi03!eZ0B~t=FJE?LZ -e(wAFK}UFYhh<;Zf7rZaAjj@W@%+|b1ras?LBLE+t!ia^($byXGPi&B00|KmTM|?61VF;shv~1$=Rc* -BuHFQs6l`QK+%ll{Pvxh`+k5A+1{pY^ud&&8o$+U@eO>)~0#UBuBko?u_=c5PRV -T2{H|abt=Sj0eS=DXyN9l|icn&WZzX9>$==EmFzE9zozm&Kf4j@^z -KtTy$mX=kS&j8z>*jS(;#&sZd^p|p(*1#I#K{VpvFT~$*)U@@bI|`No1I2>+J^#~Q98D`=wvqGE;7H> -5a>?-mIP%ejJt=~yip!&;ql{fe<5|V(nZl|s{vSr}4AIHC#BNAConJ+;IUoyp*|cdnWAa_a=L0?ve{K -{RzPBeD(2yFqvo7T*SlX;eo5tg?{`YCIV$bTjss(Ih9{m5$X^wKZzGh{Vwk=yMTV#N?ie@!SK_=2@NH -B*J1#J?o^Y$7(X)F!~8qurQDIy1=M{7v9!GBBUg$77=2IUlDoC1rl{QSTZ~>}-o{vCmlBZvN{U>-J7V>+1luiNPbk@Sa>nfw? -#1>&UC$Ux9N-gQ6$ltK2n*NQ=+Jtnf0@Ykst0If0%mQ- -5q&ihFK=5&4iM60TaJo*LYZIxwY_}toM)*o}gKS4Sb_*_tE{6SGEG3RAzz+751EQ6oi0)b%j$vu3iG^ -$n5`e%WSSv~=9ZBoTS+WAgX9Qq1i`d`cNE2+1KSPlJbRpOF8UR%sGQC5hby=u)e11mbjL7yX+!I_9?u$;h6Ca>!in -@kYyWb%i!Xz&mG{OuAB`n)YR5i3!l;)pD7@K*)bzyLqaTW6BV1n~fXecTejIb9X7)$qxxQGdhWMGE^k -Yr@o2QzQVj75XYfznHV^kZ4H6^r-lXAP`dzL1Wg#i2-Ed^ML?}e89y4mJX{pknLv`$TYBvSS1|biy^9 -w*hFDyZao|s_}>nx_=ch3m6t&MoGXhtD2|dQKrJrj#UlcCo6ISAs`TlBSb$9i)AhCDlE{#8F;r&8g@1 -xy?!=U!61vhx34sZd;dS~3g`iYxLDh*gXvo;)>n2^jh?;B^j1y5xhRNP=ra%EMH|JfOw7{6*7E^xd=XqQ{RirR7E`cG$tR#M -RO5Ar}^qLZ5Dyc{8h4HEeel05)9#cbBtbmBDh5c%Vu)bnMyG^0DfF`DBD!i%?#o%yR;9t0u4&9)KU42 -i~)bzy8?#*WmJj+|ab1<{-WGMp0ZGgMV8!cy+5*c&XcSKpA$7=WbLAe9%e5n4#xEJ6 -Y8;yX|=0~?sbgM=tQrcj{xpGSK!;mwzUNrxAwC%?OVcyI8fIVoFK#tA2X_)vJ~$AtenctE4YG3~G7KA -jXo_R)Y#oxMm+vlNbexDV>TQlRxs6p*m{d=HiEwAGxOy0BC8x>_yy$t|v}w=~lNj(q;OTH=i>4Jfn_j -}bAGA4zh5P-|~CaIU;Xt(o|A#-H2ZQf&VlV{+;5sG2G*me=X+6i&ihfOnI#HU57K -k8WqT^=xow$h7rgtOwz@_C=d}1GeK0uucT?-6Hq$#rE;tTR`)+9*9SLnPD`$*C}2G#9|i^xow# -!AP+Cnlb@uxa8T;W0Kwhw&_mAB?xqsa#hJD_8MgAs6s8@hjyA%zw{+pn@_F7iLLCR72|PiO%@=U+h1J -&2w$6$=j?TcU{$Yyip@m)4|3N7DG}^xNd{kA*KP=-%C9AkjBLQ-gUSk#PsIYP#_C~kKVhLJ?})fxwQ^z1bUF -R+Y!CP&lx6Ecu1lL;>i*_5Z3oe$mRqzUcFtKHc!~cFz(#+VX;FS6AJG8{dN-oF*Yx!^4o9;32#O@BF5 -xFMI&``)7at>gB6vV8y~8&;R)1<)5E@{q)r{K4#!}Ga6XUJH67MN4;Ux)5op>xi_NENI#j6@6lsWMPu -)yC-hOlnq~|)n9?6`P6Or?NUA%=Jh5O{Z@Vwst0ma#+AKbMf24ek<8fZ*tqK*Ob|Q}{P^8z}&u!}Yex -C!xoT8ba^^K^C2;}+BUz*k8pw~RygUZ&A_!uEI;RGXJiuR+x#l?aKdVWFc!RQaDT!xpb#9f7ntFxFK& -JI=-A`GQ5Lhq;a=%q9A9%v}~+{v#jc~C#u{t6AH8ak-@L -e)8RzTdpvJ+;i~^0UaYuZz>$Y7s!_Pj0%gOa>it6>V<)*!^%Fh;?<>s^Hswh4?`~2~5zqpUT;?;)1>g -nV=IXnIB?|%EAkH0wkf`F;$8p3TiOE!%1IS3NFr~vgk&)B6>BAxNIp+N}FVl}0=kTaEsy{ym{nujtD8 -sqR1Oq^p70rc+~u%<;`W?%!NZ^U*B0bBXBf*FzOwxS|d7%{DWpsfDT1;n9Oh;8r$6w5k4*Y< -g_bzR^@$2$vWT*=CR0_{AC}=osa*C}a4`QJvvkG=%%^Wz=e6866_)4m0)Z>`8U=22x4pZaGiHZW>gq2 -x?zN>yOdC(j9R}@Vb@!Q4lG`hrEhBS8?7JOg=!FAkLCHB}5*mLV+#kGvx>m)NZ8?wL(e4>f&|-0wz}92S8q5I -W$XLR_Qvz4LCw$KFm}1D|G5l7h3pO6V@4o{U;rs&2cn1@{ADs70mfDe<^|aeE4=|R9VAYscc!HMF(n^ -@Y?ZWkhSV!-WdbZDdA^z9~Dj;pWckkc5?jM4=S~^xT%2yHHSn9sa8K{IP*LH)ak5 -^Rs;30mhVfH0yMRG4TM&sbfC2^~SftILAmUTPoT< -sWapb8>)7?tLK3F-0t&Q1^CRwxvUatlR&c7%)~7;abn)be-JSbBXI|_mhe4HgBLlOrG16N5jt36 -uAWqN)U2+q^0zWvYuwOyFbG&sd7l4P<<6O04Kf>`A9jup~DB2((TXZ -OYfLQ$?_(_J;ifW$EkT#Ursi-Eyc9xF_UfJp}xDSe&N8j<9UtPJt%+Q9;$gzfUKMC -$L;yZ8Sk}QXHul@!PCjEN9z5ZIm>DkL5ggES%&pULBo_mbG|w}CJ;$;gnMtKi;>WcED#%|v+GKC3B$~ -6WKD91}h(O;UhvPhzhf5gZ$RMO)&AR?gw3bR(aH~A+N;8!}&Q#Sv+I-1o#x?qTL;rYt_S_{UIi}H~7M -VHa&vxL@Fe5xwRUsjwgOaPlbGVxyk~m@Pn+PWKvVU)|_fOKmKRX7&UGw%zPZ&75RiyjQ$G4c4k -mHvQ%PQ~d21bILgy?FULTB0WPT?R;CfV*tNPkJzfr -@dL_b;?a+cV!Mtl9AK%$CE}L7j-CX?!6KuddhjbH1C%`9O-aR&I|+zR)@>I?M+KUiPbE^sqFR85NToH -Z>W9o*PRmOdJ+y@tW8#G?vo0ZDlP8So@z^#+2S*a(FJofP0jCND64bzXd(8ErIO*ujh-P^`Tjkskqly -kQrUk>-CQ3TnG^=>^23VKC3wR~Kf-)e36&B{joO;<$0%<3H42@6rT7`H-<{lzTW|ugzW0+VVZ=KRqNN -lZ|BBYHUNmN%=X#mL7MwJvGhx#N{d4y^&=ssSRTE6s}`Yf3z(UZ@UGm&>-z*89+5qz9v?B8P>efX1ptxO#s1NHl@HA*tStLhu~ -^kWA-YvvcY+aF-ZmuE9>i~v5NMB}K#904pZB$VU4>&ozqrjdatv#r!7o(LWRjQIWkUv!Q0%7+uicGy# -H89!IK<`)!-+&)bnPjzrG^>Soxs&)z9P@rGHn_!xZC3~M?}sDMV@=1NG~obMWQEznFmAYQ9i@&sr_Io -ya@_6J^K{zZcDhaQ0Pi_QoTHo1q?O~_V{pGSlqGS$EV<8ORohv+U(>oW!grPelKKV3ySOuA&{LInf1; -JI#b%+3<{zDtz=vh1Upli%qHG)#sU{^KSL^cg`h%t5>yi&YHP6e8t~};JHe(qPF{fGUd=+iokFzzeOE -~D3`of8o*52d09@=21Q8I&tw0aQi$gd_?2W_o8`IV8ZaH_O%SJG+7u$IO_|-R1?8H?!9nW0I3mLyqf| -}~Xw{?T*(w|FLhuMm<*9~C3M8rpj1yNuE5g<91}w8Rk=jR( -3-?Q82?WyVk^F0Sj!NI5m`M9k{SMVR`Ah2}^#@vyXayzAW7S1Vw9O>TeIeTMupMVB$HQWWr$%ckm3=D -q)HzmpfVJ}|b&~Yu!L3%Vrjey*oZO76c&a)R?I9g393&1np80*ylWHdT1Pv0g5qCJFB@USST@?8otXv -cwSazmMm*gTk^{QBGAMbUbrTOwgCqF*Q;X2P^FAA+DlbJ)xwV)C&@IWxY4crR>Z>EOA+$_>sUGB)0Gn -ojf(E3;(VsUhlB+2DqGEo4hY?fk;5te?)G$($Is8_J}REjrTvgL5{`0?q4vz)2xh-VbESO5L;FW-JcM -;W%e1GnHjOZE{wizj5RkgIAb!?^Nwv7*YE --Ypi_k83TGlTgiB~m2?2{Dg;F1y>;;N?zof|MG6Vz_=cB3an}{5PWxI^lvrUeGb_1}IuDUN#2iVd{T1 -H&4VO#!RN&{5ksr{}mDmZ4%bE;jwx2mTvX^c?=XO3+1zc7zFQsuZ^P=9gEMx(DvE5Vv^ -3`>lLh&)@{)Uo0jibqrh(D65k<+U3%98>NMIw~VW0UaYeaYZaPX{dcYJ`Be2*O_PSWL)m6_e+*~ao&E -{>60F|EA?*xpQ;&Sb<-S)euBeW9vdlUf#G!&~#bEWKJq{PA;Jq_Je(H*PxdR|bM67m(J6I2LU -#-&W>#KyLf%Lj{j4QMEETqISQqwLv6G~XG$`i4(bpIqK`2Uix7-9;MHuQ -P>TbJDOGol|j1G|WXw1{6Lj5S}-e3EosYbv5tJl+;en!a($`>MiT -zMe^33;D+`6H;or-YJ8Mb|D`S`+s1l;4zbr#1@NGJ9nc5bK>_@z{XQ1I^3RC3*lPN%jF8MTnlL#JM2N -zP0-!Rc$*@=4|&}fj2g*LMAhOQYfWu4CkqP?bUJmoT~7-ln=V4??th4gTfyjQ(mq8KF-tq;J#V6JP(~S*??G6%-0?gn|ckUFP${uz~IZ7b%^nR|2o_VAYUxAkkD+y~dIUjwVoK7aF#JrZ>h^f2?Y53}`_#X -UwM(>i-#cVf0sRK^iZmZFg?GV&|scB@uS2+k3nJ#%(DD}<`Ct<9;gnRagf{oeHCQ2;PbI$G6ED9rOTW -ReRqqD)(uBKA>^QUyf)4!+8?V3L^z9`1yd?7}~Z#Vy=0!za!CUU1byc7u>x{BY1fGs>wt{qPRW3% -h1+I&#-Xk^O{`b6hUy?2BUBN34D12{@9hndd5)su&n!mc_uMMKBqNH-$X3M#GEqAZ*uy=FC%BGBnJxI -o7enXCVW@veO^-~fQVR?Xrlu=i08y-s(wrNGw-Tp66_9#dLuhpksU(58e?mOl^kQH!d0ZkdVO^Mg#9S -H5i}daQH3F1F>@?b0uK2Jxq-RmJ<7}Bpr9HS;OhC?@G{uiOYMe>JxsS+->_lSw)~pM)QX&CrcQ9yx!- -z>sr~+7bhsOrYu;giclgA^L^z?jlgEi*8Nk*o&_v7&RqfSG%x`hA|F`IfBrWY%5t12h`xPy!Ol*>NZW -*xf6qS{3aC4f=Mx}-%e&m*4&qXj4Tcax9hbiL1FJg{B_MBnq8SY$c0CT#S=|X(iJefT!AiXK8wF*fZz -A2L!=GPGgBH_B8hB&aJ2S8~BOJCyUX^lY!WC=SsC@rSgvPVE@=%&PTxISFTJQ*riwjb+cQ=RbCln^eFfX-ur>dZDB+cHJRADjZr$< -m$4dG@i&=Kg0qgToXXhsxC)IUlBtR7 -QHnV#ku{Qt~M20Iqd;@$14f}nPEuIhvy>R=W3xZA41V%&_n(lmpx3bk^85A5k=Dq#))&PKPFlXdO5RX -#HFaVQVdxfrTE_SEf^(uxL9u*=P1O`Zr2BTm_-`ut6g{v$emB2)lZ1!u?Is)FyE7#BY|Ks^R$w!yp0* -rKxbUdJ7JB*}J2)||W%l3{6gHZL+?JFIn)RwmBm1Aj=+FgFbLTiNNz+jcy|s~l~${r?> -oM*xl|dTcE%Oh)2uEwXe|f6qg;8O^>zo*ajar)RPJEEIC{f2BP!;VFJG9wcEv!jJK~Ps!Iwdl1vwGjP -I`8fvW32O{?}r*eTvx*)#h94+71QQ<0B4%v2VrNrv&rqYv3%WmG(M@i{eZd#O}Bq!QtOw%d#Eey^kV7 -CN2LRis=x|Uw%DZSU&DlTDJ?-U8?N%Sl%vyJ_ObQwRih!ggchu?Y-DLDIvQp0Y;~!-1<#Gy!eA*%JI; -pJZ+t(0leN#?)t7ew_{ih##pEZ#JaheC2Yqc+ABsH;81bN@0N)uVQ{u8(pnVmL|9*CWmjX!4Zo)lb84 -MRf6Fa}KE^kV_||Ak(sY7xprGzk)i3@KXdTGOGOt8YlBy-rk)-J+Dyr%YA7DW9LqUc)$qS+cZVk`OM} -<}*sdezj0GCp`PNUw={WcYzK5X^C$WefPhYR3idlhb%8tAbQ#h%?^5=qy^RW+PK(pGQrFh -Lp`e{D*GkTuuTA@LMk<83S;8)Q_(ZY6%|VvmsG^H*w!N{T4j9-U@L7!w#xOIbii%7 -veOL3glV_+D>vYrD(Mef3V1}$S*5vb&>f(aU<8U>NZL`_R*DPDx7R~%uE8jHk$Q`ELueF@N)^dJuEhm -~pNs|1_RCFp6@1qxW?ge>!Ip?8OoI9uVu(%tvPf^5a&GwP%IMpYh=X!Kp^_*}g3i6eybuD4!VfC5pP4 -D)G)Y+RV0KS{ox+!@$ciFJ+Ql-nmeB`p(gE7Aw57~FAUvDz{XaAdwkwfU_-?es6?r2@7OLM>%JE1 -DYU?I8hKsN=`*T7AGV(1!tm9B^A-7MfCK=H&ME%z{W7M-RRYTEuaFVf#f@~_u#Q(^7(a&3J2bWTQa6L -vrL9W+pcdOp>{Q_=_;211|E&YeQPVJ7L=p{-vGF@s9B -nAIMSGvX;reI!Js{ilEB=^L?|})o|L0@H{DY(a15ir?1QY-O00;of09jkTk -tBWy8UO&c!Jc4cm4Z*nhiVPk7yXK8L{FLYsNb1ras?L2F9+eVV#`77q2x)PX>uqb*R -M`w9gwv|M+j;)H6*XdLsNDQg5K>)!3vKXE2w_kV91A_-8Ih(z!t2&h};W5+G)6?(i!45mW5|S-N!WsO -Ht0-ZKSnyQxm{n=a3szonR`OykS+=B`qdZ!`pSP;ni2cM1DYA61!}f!{5sT88n!P%EW7-7ljB|SU_T= -^PyR&0GU`o%5EGrdkz@XrHCPkSQ8-TaW3bx4BdBRK1VqQifk-=cFEV4BV!(~-g1rI|e)_GQxFmRfcQ3 -rqrL<2r?&Ts&mtv(5QBM;w57s~zk@8|=FgpI_9 -A%qy -p^Wz^*e>jHU;n{cZe>e}{9esa%cEDn>C@+A_Bc$~ubiJi1csUK{9pIHWVr++f%Yleu5vcczy1^p%My3 --q#xC}WD6+I}DeUtP_SpvjKe!tV2H|u+e06djI;a9ViUWd0lq{t9o%`OYH)My0>?!nEB$1TtJX#$npq -@v6kC0spE_tf>&PYfZtyozIpbW8a#6Dedo&e|4(x$SEfSi&=1qb#{r2sJlaTDuxRU!*$5vo*+t4Lnq( -|{fS2I48o4FmC8d_=ZJcpx-UA~WFG92cL*fyI$7!cZnzDM3yS*@YKIa{*NLLIg_}uxz}5S-@|)H!|W~ -s{S!w0t}IgG7NpmlO?UR>2knmNdc<9GTrHR$i^?(yDa63Zuo-`0(VeA%ef1Z>=Ur#u+<*mLp%7r<-Hj -d8kx&Z%N@{l9HxRZ2Wr4t@oN-T36un)^MtQqAyF$u!1Lt=EcYe*B$9;9;q7&lR1r{@agcIWiZyotmsb -(!AHbyqkxUZFL@C)Uh=u%+!Q}kmjIBBFVLSp>5YT@eZ4fFp29Q^vX@LV!!@59W42i6hGPwZdhAJasd9 ->n)JBaj0N$dsc_#-@orlPz;-je{nL{+pR#2k=jS^+yP_!w1dRjOv3;RJ!(fE5|&zPLgTm132OrGU{v? -d6=uRZe3%Ami*4Fm%DAHSuz_2GtO>RtG5#D}#T0zrz5ddW}b|%6U<4EJr&12_kVC;88Og -ysL_|%WvBt!tjrQ1kJWVFjzkZjI~6LjPc*Bgk`WIAY0k|17DN`YMy0D4NEK;7(0-Cp|y!mJw~XbhIKP -%*mqfZf|e9{oX3MUR(_+8@2=aS!baT=iH)v6(To?OSX829M`uT;iV=CS1bC4IJ_s7{P@(y4Nwsk8YH# -a9(r=r?-PV>3y<&SfPwSpTm-I>})C&hawk#aBbbG=QyB+&heCgeGu)7=I+PT?lLx?-i9RY33!$Z6v7< -cS%>{i68@!d;+Ga$~43duuI4L43v`vh#CrDGQ*xy8^0e2U8Tve8F_raRRZl_4C)7<(1@9oVlK$ok#03s6^)(pd_Z@B<&_#MfE6^* -CpCD8&$R_`XfMfL$=X(7OXv{D97goLv8TEqPrD2SmcfN$ng^aEab)j-UD1lb7`<&wJ2QNUtt~EFyO?P -OQ5^d~T)&}1Nqbs|Lp1}#9cEc@`Q=bEMF`$ss65&5Ix2j00JRbS6cywCs9SOUa#uqv`D_a_Yw&@(da%(B4!G>jk$5?uOfS;A -`-4;`>Z>@?DJBTm6pdhBv(zSU5TXD<0d)dCOkmgqS>)nU={gJ1Kmy$mXQc!B;$UAhp`9Q!C+&sN$!3jXF5t3S$fxi}6BeEK -(6nl+X76+pzBzvN!C7EF -Z9{lgjMwdhu!wF!|2{{TH1OzQsPD`*XNRl372LR~_#PQ -LST;xxLlogjxrqk!s9lD0M$jsx(SJVAxUp@PJ-%c5ZM6!i(kZ)oV(VjeuRss%HVKTmR>)IQ?K}l$+hZ -#B+*es0~TXp7enC9-uRIzv8`h9!yHhgn(cJ%7)@f)*@HF%d-K=Jep@q$Kp?)0JAI1JSp_2-lG@50kxz -dw5W_WkPy>$v2x5MbiX2bxArH^dpxTYE0F4zkmh2UDeDk;Py{!IEy_@*m%wyng@YI6OP~_u~%6dYmSh -L-v%`w&N=`nz_pj0s6s*pV&*AF!t07 -n-~jlU*WDtCAvNer_T%fbG0oI2fcBhGOzQ6Bb%J@~Y3wh-K;G&lxx2M2cBgd#H3bhHf>?iV#J07*r5W -DcnsM)LkqiciHV6ny>v5yT(bx2~D!#``*6t$y8ClE@Y)X`rdS5f^Ge4s37wS*fc8msOJxs>6ClvZsa9 -m0qgGimzd#x?3o4j&7CTvK_Jye5B1#k?sdC?@s%kc}CEc`ZedAYqMTH{AGq`<7NAut-w1DNXp*vN+O6 -HNa%eGy6d9`qm4vCs+4WeXb1(qx1AqJ>xr%yWY0{lPJ@C8o40V7sx6EWtsQD4W`Df+vr2*v04QxtlB; -bc*XycM<7b6baoL-BL3wIYHP0A_5u^er+tE4K69P(b+M0Ckm#7lhFnQ;%OWe9e>XGN>CQA|e7qQLkLB3)+aFAUltDzqbU?06Y2F!D4^u0Sl6tJ2Zg+VAf1)|&cFJ34JI%;O5h#f&=?b=z>X+Y*b9Zuv-`~=^_T9);hc9z6-{g%* -}%5CHo2EzT=|Eik8^}O+0979W%B1fg9A5`Wx1n1Sc!dOc(U00h*%|zp{sd{z1p>+!{nGyg;5Pp23#)# -5GPJW5X-cwZ>E7NIOy8^n_Wd8rgg1t_Hau9*^!o5!8EnL)Hn>s^8J8?)dzc0ozI#8wcY^niO_B_JeTt?2hcFUW3Q*Fm -AGmO9}AxJ79!YG!unrC&#nTa64Gx~&k5nz8la-OSPVT7Tv+Ezgr1y3j|p;L*74=NfZ`}j!1dZ3@($I{ -Kmj5bjZ;AS__k~E%9_Qp5QpM}q!j_K`ql~#ewcG=-eN?PpM&g*D#Kx@$K97hFAp1RaOozBpt?u^~wK^ -U#$XHPK%a=YH&xzo7)qdI@Z{0eiwO;a!&T9CN=dSF22o9Xj$o_{snN(t2_#5NYhs)(Wlq3j0watvV7s -1(=S0Mf*XH;I_<>YktX1A}YmAt0cjD_sB%cqpsoQrz?fA~oyj807H0sCXl_12Ja* -0x?72ImxInIWN_MGW%jnwW?JKu_j51rAR?6I45qlQBW%cgFmqFvZ&@D+&&j|x! -!1*66HRCEmjtP=)xh@oBnCVTX`H3|ipsoh(`0ko*Wi=GESLwJmFr*F6Vhc1$!=#k$M1vkjSFl>=~>`M -d3=I+f96rJ@A(uCg+y90FZUZ7Nu;DFT#j=cakfc*|MOPnq6G}nF>$Y}?pJIx%J4J~LABsH -jz|sY~s?qjtl4^mHjkJFzAs}H=!HWZj0MiKlMpTw(?62*TmC&d*O^bYQe6P_x+)a2~iquNe`#?#~pvb -ZwE~DQYbREymvR(qC=4XS$AJ~X<*&lB{9zHmV8T--MhY=X7_g756H$`E?}zYtYt-P`Tb~F>M}uepn%_ -Q;NF!QSQFtQXNZ@$=)B5UerQ{RB*GYA4)@tM%KvVC}n9)6$>M^utg{Uozip7fmD0EJ@0Rf5=6V&{>GcKqx<97%Ok3;2n+Ldw;;FFB33IT -xquvmj?2Xvd{^B(4Q#zvsv_lg67ap!_f-AF(N`S5|ru*Kd&gzi6+rWbe@lV7823HsN-~h|hAjJ+lfB) -wF0ei<)ieZ6jEt+Rl$)@{b;0ktn`dWz^_QZr+R;{ksWG~qNYu{8}s;T^iqRl@%KjXY8+TJ&teu8Q&o_ -#(+d+F$o3u1liMa5dJx#q3AgRN7jQh``S$-cJ@ng%k`EHX07q%P<=E%ReWF@u;AE7eYmA4;FqlFA78N -?*%ZGAoD87lTLoY;NAlja2}h_h)EPO1^f?<4of;!_XYuajJ2{?V9paHWtuUl}%_{yD^2r3Yy)_$p{{~ -4h4^rrW@XxHckZH7<-Z?D5 -|u(Nq^CKf2=s5@z+r)YcX_dttOv+t_Yu|0P=c`gj;Gg$x8B*WX -XnS?hd&+vKt(FA-+zC4bdE<-_Po1pA6dp=77iWCot~T?ciWYqOM8YL1`0smqWo#R*o?Pp==#caW_rSS -B~rC_W#p_kmRF4%J`(Z$-jKbp15ya-TEPi6aLw;&Ay5Nk?E7fJkjJ3~9p%_y+H-^He;!GSJ#>qCwa$5 -LcMKF5k0LJ5j+IvHIYh!rp*FXlv=R6YLOvlmszjay(cC#0w7`p&Z<1d%!tDSjYU3WEf#2SNQVV5?neRme-`)*7O;zsI~vLlM8(ujE -D&(4S)pxbn}*W7(WufoLf6#RN20KU?o|L7SIkrGYI215P&&DIc>hnKH;G?m0ELM -UQK50hf@96RdmhuX$jT2>A0EQ9x4YyWYtkK1gzjR3wL?-`G1h$4iLi}^WFd%h13Ok@E#=-1?&`U -G|FB>G)S;TIz^hO0o(nLol{&7GBE-nrM&U@msFR8vGyC*(lV~dVopDV@6Jv=BfL)YFORiIPTL1@REsR -JJtz>MPmP~;G&@J!Ay*DW{S69($aFhLDFQDIs!Ss -4!x@rmWgp+jq_XO*!h>>Ys%UbrdH}otgdipue -AQ|75F;{iLFpS0$$I->v|eIr$B7-m*6=bA}bL|)Lk-LV{unmL>eIYC$S)>SWQez!!;!+N?1j6ItB9#S -P)tjrP?P1(UCB72^xhcY%o*w;5t|_)z?BBjXz2x_Y>^8T;T)?B`JCKI+ -i#?SyfNV}rlGpidW{wz2G)+1K}OD?ffh^`z+cXUs>>i99h7+9DHi+Z&5=5hlO_i8(pkOoZ$2iTrT~>; -UcA!R)`zqg7yrp4Cpuk-{oa5{sWes?tTNWmKfs*+;5gC8$_|rrXzhpDF~K&Fnm8Gty7+WG?W!SXF4Pq -%JO1huoWfA5Pq%gGwRYfJd`tv!l}!M$au#UzqcAF+Rp`PXwB_bd;{lvP3>HLF5yQ=*w9cqRc3DQ9DT% -Zp$%L4(hJ0y@17#y0doSW>&V$+CtOWtl>uD+R-9X^50MooyP~Jk;WgNtSV3t@(I??AojKAq`p0%y~69 -~5e&N3!L{?iLEDq4Jlu%LqtsHQt;arB8#Q1UD2LJOlboPjlP%#;0oS7{A`J)5Cabkt3o -wpJarDyGJ`#@W-5u)t-B9QC8X!z!Knh$hU|H0nYMFnJq(VD@cX&dYbz*>*xM*JlC0*?@nlN2rgmkVs7 -3FXb?I^F(8KkX2}wK**9%U$tgMeG{h@O%&f@pKHvD#J#W)nqt^NIf&DPd&@I>;PwXif%rf-WQ -5vH;d1gFA(7vP~tbVxPBkb`7HArM!+Y{Ffp=_rVMJ&xEK?oIgy(u|E3Ef6K!7(hUL?(m~PJf^QUt?yD>I1Of0lP4lSi%EWUnSn<9j*zL?u^(4m9 -2CR7S&(M3u>PTE2n6Pp62FbdfS^*u1A*~(Bm%KsyW}9n6i1+rsmjf+^E2O4lnAj05xmF9`L13Rb|_@b -QU>HbOa(vqWYBoy#^1N;td_@@Y-Qat7O8dWbX$9VO$qRGwJsr3}b7$1z8=J^+8x~aY?(7I$Ef2i_^62 -;Q*7(>{Mqr3e1|vu+)lCUANh+uhiG$y0z+O7??LRD^hP9VED_6Ys^+)QZnHsmTg4yIqqDj&mF3)6uLw -NjNBsaV$oiLQ;E^!4qbHKlh&~xi71agvb)et3tqypaNbvss_g1;^_Cq=5D@QtGuXRwfw@piT>&M&QJ2KV%nEIpY8H26iV8BGFo$ -ZTpPm!>@EUhYG{h>{13w*XSJBA`Qfztrh`Zf~f8SvXrU#3)+0`gG?BIdDC}>~y;bDy -KtEL1!@++8P)h>@6aWAK2ms3fSzEf@a+;C`006oZ0015U003}la4%nWWo~3|axZXUV{2h&X>MmPb#!T -Lb1rastypby(?$^fu3xbz%tQ*Z#z~WQ>P}KZ2%XF@fgu5AYE*HS&bC0-sduLsr|{pq_ae!16fg`OitW -?x-o8Bh?1~+@o@WB4Sxx~z(~9JfX9+C?O`$4M%0bR4NXi!i*i`>GElGkO&y83BZzvZTE1VrT2!{cXBG -qC~uAbI1AzV?a51(J0onKy^+Y1YQ#u<~wHhREm$wVd@U!lDzH3*dEuRsu+fkoLMzl9{N*ugNTc -l5jy|JhlQiCJnXBMR7y&EVapLvJ^mk&9lXVrmtvDiJ*@2q{{LXO~SFccdVTldwyj78E1tIKu8`%&L!g -uw0ld_OR}J-VqN##?h{ul_>SWwIT50EM?Lc1ki4SjoHOo^9Kiov*VUh==)IAHG(+`lQmJ%8fm(k+B`^ -dKb5`Xk=5ea9!E^-?5Fg1*f%mLWKUdt7>X{t^xke! -JI$D;3;#c_5U=cX+T(Qg+~saS98BbxI@wQPM;s{=F*FdneXc6G>xEXY>=6n$u8q1A5_UJKX)p8=nsT9 -u9twKG$FDXBZEnz52|5h;Tsdi@z>kygENSy*hV%Cr(JgiY(EX>AQow<3Suok6e^38gTVXq9bH*YaFp( ->XSj_{rO-p@cq5|3A*EB__jQQ -W74lO$g-(#V-s-QBYK{Q*0;rQbf(gp@|L|2Q~^7E{II#`Wg!;T`Q6a+Ud -fntFeTfXHhnS|J2C?{sp~PIy;7(@sA`QX#tvNRMw7$zZfLlDD0xsJ+%x)HSyd57G8T2=4?s1XAE@}~bqrFC{ZmYjC&Xe|1)rku=l}LRZVAB(QK{3N2q|{DMPWs*4-4BQ1{;tWk -{b5vIBD_@ze?2$~YRdu^WCe3_gKSXcGDFJWa6|;y0aGLwyjz=LMZSW`3hGEfMJl#(YuLE8yOnfrQjbl -WwDApXd=}wtXaqUVU_);dB*$HXrr-2*${0)t$Nz6@f7g#X)w9q;H@%JF>Zvnp)Av>P_QV1dL#Z8B$ba -1DQiy2AAwanXXY4f~+!lIWMduJiJFb6UXTLCfVXO)lw;7!ys*M^0?%zVR8#a|4KBy;7O9&=o`_WezLCvS_{NEB1S(nQ12ZLEmsI!69<6 -}5#{V|(c2LK1r##3D}wa#p;dA5|T3BOUq>(+Jel<|8~(M`wtMv9@gmw!=U>raL0jZRRy`nYGcB!-EV- -kQ_ljI*lr_I*EqyT=E2do`Y5_B-5o0M{!>-vo>~(?G?lQN2;c1z>L)zb_?3S4O -NE!F~aCW=S%StN7*}CY^BMk7{ntWu_aF5yBtlZwF7LI!I+QCZqMe&}t*?2eX9T6skT}4d0>TaAD?uNO -qGwlw%rFND2x60HSh`Xrf%)?<@&gDGA9VP|P>XD@bTa&u{KZZ2?n-5XtV+qm&vzXG8!UF;VYyTEdXUCd*}W^p1Id`2st -up|yeszk(AX(S}e=Yr)zE)>gV^yBT4hw$-1gZb=}kSfm7-VQqmhCbtIL|{Lhzqc>}I~Rf;emHq|e0qL -tCQRv>%(7h51`ecHW-87zxdwW(OtLUrEEACn7Kxn4i3)n2=k^|=>2z}T_Tu9BwU;Vo*+FSLwsKn -+m|s40z6;&EC>Pse79JhNpY5`{Jpp{2pV)V_xh1ZbqnpZ;FqQa@;)y(ZF -?5z6loAAc9?LU@`70V@*{o`%PV4MjZjaUnjkVNCN5)&`fc%F=&`e4a(Xm-V;HOUmL2T5G`ekL<$M;=t-YLokXpH(>0P*>C{x7#(!@I668q^ma$VFaXI$y8{||ZBd|Mt0wOPtutY`AH-Yz5$#T0@t>o360z$femIRi_=Cc{;3rpG^UDD4mRkR8+&Wqy|`-!+r -&7CI?h{0Hz`#}_8IVr|`x!m|!m+LLDx{^wlHCzHMs$&5S++_7ZR=ow)MB0zEjjlRRqXqm8EA(mvw2xf -^xo`YJiV-=?==V>Ue8SI_mt-&!npo$-8pVMnuvAK|fHeUm_N)rL{tg{ttoUpxecF))7So|&nRq!Z6Gi -9KXx!@6-t`m>RK!m$ -X;t+=ttESiTEQjd)UFP_;;P#Z2F4`^WTZj!!AB={)k#V>g4v1=&(l}%%%<1{;n?Y!c{8d)9*zRPlbe- -l&p0~Ps4d8Awoecp+J;#%6TIO8MMv!^)BEe%!1!S+MWX`h;=a(2biJ1lo9T5`5IzaBoGXumlC~6Ez5- -LMA|Lu?hSqP-;^T4_x_u+*OYx6T@LsEabE)S( -5*%h87E8?*p^-_Uamcey|L99+F|o*DX+bqqr;qUPE*AIr`4~RRaIiJ0)M~99u6!jZ*o -C*3Xos>h)+Q0?b>9-4q$4}A%j;tu#rfSOO_f*B4t_j>N*?9l>(SGLrb_T{(4`>A?Y_$50N<*gE;9i<$ -OQjyXH^96w(7txq0f5=Oquf0CEx$my1ph`Jb@KZETST3^Bh_txfxVU|Kko>G#Fb-cJ) -C0sJ`0N$u5fm|saTNXSI=Pd~&rprs7Lp_}qS*8t1JZB=t_=T}ZA1eR|BxLXthB5Mp#Z>6r2NEfm`UWz -la(2b~1oXfcRA3Vm2wWz(=?W-F#uWuH40v20O{7p14FXK(do5zAdA3R-HWie$0X$^Rz$$6s`dW$r)b_ -(q=f~IA#5O%T`SJ9pkH_!c;?v$N;Wr9mP7XR4FJR%bc^rxTJIu_-y=o3?Q|p#}0l?_+Mg#IGScQiU`LA%NdtxV$nmW>H>d~L{HGrJj~qRV`g@p~=>NfPqZO3leeX+*SCCNa-lDeDVJC*rU%o -UNfSP04I>2IYXNlnr+yaIvCxfZB?GZz=~GJ*bUV$41SSbfBjF(=NsPc?==qK8D(GDs;t6{sl -onE{K81Av+bYxCKU-pk$-HxOVVPW@ee1iD8XKnb5HZ0vZrqV8l8*K4Y(54Kc9rbgk}!p(QYXgNpoxkJuDBNTs2}QcLEMHFhA)*hWH1UK(>#j?$2jvWtQ!{%EI3Lk_m%Za<2Y$`yzcSkb -sKf|3g~EABoJUTPjkRWO@`izR$>_3F4}K(6F(05DFn2?XrCaTv_7X&jWeX_&OsP?3)37*ne!RuF2P0R -FZm>f$oBu{@tcq6G=fJ){tzSK6pk$P8fz&qXD~SI_zA9CWx_I@Rm?$dMEM>~~Y5&b$)-3o`y2Ab3@wK -#0vUb#5rtph}{dlm*-^jrIX+xy0D)hy}#@^)#fE&EyuX0%6vu@J8jT>b1vCRSJ=dtVwQ$MfuQ;$wVXI -z*ibUEdEjg|MtQSz8^bkRfhsZu;7<0PWvd;gN}O-825;-api$S_pGIEwOoUKZD#J~2F-~#%?o5|bHa+ -~o)qiO?Pi^M1pTGJt8G~C#hpm9CF-kDsoCkUiracN0RCdJYIA+4mQ)QZ+1YgAwe5U+UrHorIA8iT|EfKm)He=ONy4pQeLsCq$LG -5W?q%1`mV2!q=9!9}I@Dc>8J63<>vjd;K9Bfu!0}XJ7@UH{SD3ZKzYGob4SeWTp29asWeDW|$rzCkz6+2B*E!K-ME^BlzhTB}}gJA6_tv>`beT9;yywq5aAx*< -wiYiEy7LyJ@~!HnFK$zb!p^kvJ -205pH}Nv;k}?y^wuuy7o|hDo5y-EXq=Q`>-9Yk! -1>JQ}6J*h@sOSK{NNt{SLT9o9v^GZpWNH8&l^r&q6hu-WIrBJ7BHeLr-q2bvKz59RHh -aTFIogF={#8)>bX@rVjkAG*vIeH>d=Z`YjrXnsxdXiYX`EUN5Epv|vBSf4Y>i;J;*YbJ0DErJ{b~1xD -L3&tY4-HauOUUtV7iudgv8u2OynB^Hd#bIyxAI=bM{VADHiW3(g2OYo-z~RaSO@Zdpc9z40C~AdCsn}hb#llMojrSg-FBoi*o@hMH}t!z^aQW-Rv>=AiI!$5u?rymIE}=EzQ4k6K6X)H -ej=dEq}!1`VABoRUA5l84%>;L17sHbp>L`BsA|4S^25qmld&|=x*WU6du()N8m>27TIpxwN-5}@LfAf -=u5+Q_v-0+XZRdXStcM(qCaC0hKjUFus$%n+ilpYVDYh!$OB0aG#yQhhxXuAlJTuUhtK1#a;;eO=F!0 -?hjH!E`$8@Pe?>6F(DAxS%2A+2c@oek*FWfM=ahN^vlOs&( -LKd`z3se}zS#6P=v1HV8SQY&meGkiUpN+sFG!X2t4$;`0nzi5k7!By2g+{$=7rpK6pl7j);+ -r_{n=S}cDn57b`f=Z&<~H!yPQ9Abl-h~apv5$#=IWh#`L>X6z}sIJ(xV&R2QmN~UI_9{meBdikp}2^W -ysf?jP8iidf(Bn-P{))ZD4Zi=0-_=HT+qqSZ5e?VXjX}3^_N~09uRePDBAa&w$J>K(CaZ1F7pEn#2k; -6hR(Dq|trFonfZfUE-P@hvvBW%?K#LWx)Ot^9p;d -6;tqn&Z92VySyDAUTd6`$yje%j!l-ZBFR(*HJN(4z@UZUAhJmP$QumAZC<7G#evWH8;fra+*~6GUV3Z -&FB~Eb^V=m{5H*auj6}r*Xg9RJjdU|^@eDUNalVC3D>smrcG;eVYZ4Vh<@DGE67}_bybUd7{VcnMz5l -;wAj#)(4NtPz-N#P>~(@?9V8YaOoM`!vBz!D3vEe&YRIm8w^l9-t-7|mHI-KG$QXmXX{28|6d)b%Me* -OWsTn^B#PG|iMO2ey3^HrINI~iDCcgy8>u>YPNpS^lD3@Tvz>@DWMy6lKUOI$h9)(5x7PsBqUW;c>A=P|wuxOOm4TX77f2K4zHd%R`QMe -cBzWa&+H?B2f3iWLEN2UJMeURLqIF*?<3*W5GkuOCBJb7gZcUtei%X>?y-E^v9pSZ#0HMiTz6UopoXP^poqw%6it!r3^*qGRG@Ns#2YIm0KfsFlT --;u5>06(z`j?=wqMFP7{ihbzhiu(>Ylq;f`ct(Gmk(;`h1(`q*F!*r2N-9R -my$vP3ug-C-%a~i2>^Vh%q`m_Aym*!Li4>F!LjSlkdD2(r$Q5eYBNYhM&HquSjR?nqBrf$At-_KOC*5 -UmkrD{;4pMU)6?~O!DBjc3%bNN7B`AGj%#{5g+&7H}|Q+sS6^)fUjR568y7E;UUn(noT!J!TkQj(feu -n_uQ)+tpa;x#2w8yry6RD>~1BtZexvt+sm4PcF$rz@c)#AhUoQ9&r!D>D^jOV}h*);JF%X{eemBsrtv -gj2I>W>Q2H#>9K&7I0T^qRLVNvs8xxqtqc5MA?k(D7T|<85TMi!1~27Y-YfMZLQN%&BFP*4$!5vYPl!P9-n7AhEh4Rq!k=zb-t -HD$}22BS5c%^pa;}I#j}uknU)RT82e(X9;GFf$3IpnEY2-KG;&Mw5Eol!AwXd&3(rA0JS{e?fXg(qsY -#IsAp(&oZIwRd>*hvWcr@ytkFVQ9kNgo0hW)pG$Lo;O9^u`o)3rao=wFQq62o?H{EqtP)b72bfBU^oo -xI-$LvJ*q{*e63LD%;>92LKJ*1hWZy*Kn4CwqOQyN{#??6_}L7GS$4F~;^2NOB~?)km*A@q2cUT^F|FV@I=ix(PQw7XratbGLrLu=z%fADVTzquF -#^rGMGcvybz!MgVAu9vF=wX<&9zpPWIec68F*>ip146RsU;q`@Qm!PqY|7T;r-vbpc=6-KH#6lfhhvU -uRYk%a`sXg>Z?3VLk9}3v5px?JJIM?%X9CrC$WB>~xUbzDGjXfQ&-38i+kMD}x(BjNBN{YlR&pM=n;A -9kWVEcm5(=gRSuj#JnMMF7n+)+3~jfe9P$zSKBqxYFf0k$hOu$9&5s3n@nolV!hZLJa=$`s&t)|Nl0X -LFyX$z-0TnU<3Yg-cF$m; -(7+9%EVa!DCirrQ!<+_&hiH$Y7!wIPK#g(96SDpXfv>X2sjgLPz{^2%mfB36Yt5jalj?;yUsxKQ!=(> -Qa1Q9L_dX8`Pt%x$|X{~gvmGg3(p$HA^TrDK}7!?#6TWSn#LuSzFEN;afJv}|8FL{=!7k_5?}z3&TCzAt&hTQfQTA_Xh1#p=jji(ON%)TnVvF -GMFN70zujKZCU4y~KjJmk9jad5O-a<3NAWpPdbs=Ey#u427y%9!@lQG0P_uTuw1=D)<^#DsEEDRpG~c -B7nE}?i0Zxl#j?73_O?sWV9xeL}8k8qNun4TrI*>8bna$u{<=FtK%pSSIZ>A!Kzb5aywKdr?y?I*|i! -KX;mEOs>Att_yMraPCIn=v_rx`;M3Mka_(}i>T0dEe~w=9=1nBy>Vb|LwQe^MbKLn@Rp!?lt&NRd(xlqID(rYS!JGRb>#j=`4(bWLH?P$2I##v~Yo79UBTcR8m@D`&gK)YxJ;D~E>06^Xu}BF<%wfSu4V!Y`LXQiN7bgm}V>z$Yv98YW>RS&T0bdQf;M6rrbzDNc3w;U%gbn?o2KL$G1nXHu8Qn_5zgUyXU^+K}W|L -AeCEc?P(JfM*9&C64C!Xpm)Ua*juB)D|4_#sxS_*k>c%P!vB&i`ZSeQVhrDBQ~sH^a-OtCtVHPtohjz -!d5aE$0$CShZ!uRmYz$PfGV`|JV-A`*KgCeYNjZ+Bg-#wtxp!^(~e4b?(~QdSf?kZyZdMw=WfDN!P{s -{8n{dIkiKoHkl?xzsir?wBW`AL6gmknc~$x-9z2a;JyIqRGu2YbAER;aJI8S;~1S{zC7{?;x75mVbvY -nBggr9q0|9~sP8}=jXgi#+F8rz*%a8F+irF&xW#mq`?2;FQQ3EFSqkMh@6aWAK2ms3fSzF+NmX -FmQ000iX001Wd003}la4%nWWo~3|axZXUV{2h&X>MmPY-wX2j}|$VmSm3fWue&JdI!f -`RBeY${hds=W7yYX|a|Iaz`e}8- -Rn*B$@-(J%qul^n4EXRRgU?$t8X(Zw&&jHSOo>wxw`aIcqv(@tTKVJVdc+>y+&CO4*Ud?Bt$J=GGhk^!L|4UGiblsUK1NpTF0DL5OI?XmZqJSzVBWG4 -d{QD~0bz0&qB-1L`|kyhNCam{w#5hJKLd7=>vdVnct_N&$Gf!>E6A^NWN&3mw2Ap;_d=pe&^Y4RVqd8 -r8T_Su+){R&gLcL@(J<w{gg -f~ke?xex6Mlphtec^F@S5iQ`RWWQN$E-BTbtIQw{W|Fx8v -(6oZbZ7`Q*h!15FW_os;bFeHgc&`9jCI7A6PMPw%;7vok4fwL&{W#oFk9zrl_$S$A}8%NB*(8x~?GVC -k`;F&KO>&1HL0g$AKa+Ha0qTzBo-=H<{Hfn)r??=(j4;~E+mo$Lt1*Mb-1YEN6W&kYngCkZ4^!IZsP@XKfnbeN?WI!#p|=hP_VuE|& -5&nD5}@Au_wPllAO+wBp#@i!^A^R<1Sp8!b&ym^?nD{Rqb{a3PjjQ{n;#0URC?O8sn>xNeuUWk8W&Wp -An#$yW?r?%@S^bpv^|g(=yW?DKgOkKxNaK>06WT4z}s2?fbv{|A;PYhNx$t^2SW;N?>|GM&S5Z0s#vB{3K0RMIJP+Q{CzIB&P+6NFpuFB~DT03 -gR4N*(o8BPFb?aNg*y2B)$X}7P=)au42oWLQ-_Gt?N|xhkYXW)LDO@t3;5@)Lj{kdnauPkRN0u@H=?g -)GO$UA}70&JGHtvhk}YbaS&lY@1h)4iSA_k2tRjmRcY>2$E=I!!}~uh^BtGDVA4wQ)gSmr+tyP41A-veEqi$OiRqdT-tHs?~oeLpJ%z*;4k!A+v -{8)jgv$!wHHr(#5{w>5Db6(3L1@+1Z`E(xDfgLU0JN&SkzyyBCX0>dw#HN4DXg?^O%X~`z5Xon9ZNqF_` -Nu;)zXVh_}4$`oLaKuw!nipE&m;~zd!aK{|DbQ(fO7q7~D)caTsL}yS)t__g;?CKXnmgzA-o>4>`%bc -cwTAgwz=~4qwC}9ZyUTjuY?LoI6Ck$q;V$6?Hn+ks3|jc<8 -(X31$<@i{?=yae?u-Ud9*LJ88I#45oLG~BLSdQ=dg-HX~bBXs+RL+ez~z5xpQvs -&oS2B_YL%~0f*CNvt}yATWv#Rhk|&QSKIFrm?K>r9{*z9MifJd#Z5gf^w3AZM=Q|iJ;U7$bl3oQ1JDdu!=$ug7d;c&DXS -r`T~c0C>gzPcc|(fam{L$?k?ZGAo-yDZRaU{41^%T7avPWeSXKZ_9)O39OYMu|}j`DqL>^(wl)l@1=< -oQ|>;JNJwkislJPb|M^*NUNa9Xf}7LqZZsVfuVM(!F^NC@bkA|L(bsA(+0fLp|?)JKnFsD1?LG$k9t* -YmT2|Csj|)leKP@Y$m&dxX*J$ZtX*ooA)#F)-W4RLE|Lcal07LXk*wx)w)JL^u{`475d}}8FW@4Bq4A ->%i~*l{5Va?=#Ii>1c`OJ_GRbXG{kB8m62*QXo^2W}?$LsaoPa1v5t=w -mH~;h7P>z7CG~QRN_wGru5;v-HA-_?hL=MafQ{63tAmW-dlDjpH&Q1$#+nSaSxQ6TLPOAWOnDSIx*p2 -!OMf2=5D&;O7F}ypd;8rqnCxdR98Hba{fAb&XO=`S~k0*~ACwDeLrkoBnYTwE)g8H< -)eW+=!bv82y_(lDJ2X#|7x5^?s1`sVN>6DX%D6@bKwoIV1ce_wN?VSjI>Boa>?Ea+QoX9~hiK(gUFA+ -1nfxaJkdDSyKkWcEmnu@x|J_Qgun_IF-L^5By6m@wY!n;L6pbe%$1_DC+3#at+NnK@?7)aENAztHHrE -|boOViTg6}v_jTB!eo0S8J^zIn1fQb=P-=*mzqU>cFi`!!3G`QsnUY+1X@+_UnZyCNo2lKMXE^0oF03p}6)Qdt+e19TurrpVd!0;E816J{B>J#0tZKS`CwtLp5+%UxR_&3kRt>_ -r#+{WZRF%d##jHmr`mM;wCN^;_v|{L^S63t`gWmk#9J-KM|yMcv7&JqS)kazAc`09dc1d#6e&^@qK8de$1)*MY# -q1D9;V@FEWTNh7hs_C6734O9Np4%mr#Om7!}diaQZUwoHL$Dhz}2)s}BI6st>P3?E0LL7E^CJ1YQ>vgr6X3Pg9x?q*Y8^YE -TAcBs1LuFXjpoX(UfE$tfLMq}GDqxwj`CK9OvQ}gVo}!Zgs(xAlhal(uNVL>dQd`-sDotN!&|gJ{Uet#hLRWDrkXsX&)!Z_0=|x7ZAELKB!hg|=`km#zqNKL6T~(Ta&McV`<2~Q0~j#wIVw(>PxQoyLo>Ntcwm2k@kb{s^0 -*ljGkpoB>+IDB8#N5J#5#&ROOJc1wtzBdbbigOk2meL`-Q=f74-qbHQm{$2rc5CCtioQC1GxV)d?Lv8 -tB8uRnL+<+=Oj=k7Z{cfb3&``11&!1YFf~coLIZJ^r%E_mZe~L?TMU8y{)VQ&S}&PZL -KysrpCXc=BYy&|)<4yWAL>MyjPKd9CVVW9S`{toq)$A_fe7cpk*&e}9S;QXf|@0nO4_C7kO3*o{@%K< -v#qn6x1vjMk>f{|W?4&Nc(#kPNbjfXMhp{hd!RfMdZsvk(x0tTl^^h|>X=cE4{QjTncFl+s*Z-0(ypT -dnpRT1x}7Zz&al*89#BYu2G(u+??=w&zlfuPgpyg|&IWotX9Pc2SG$z(7 -2@G8+LmHVF=@i|^IJw=}`rX@kp1+!LNnM%4}Y>eySOGP$1vAu!1|(In440%n-0*O4V;P$RQL`Q{e;pG -d(dt1YiZErsBPwDC?Bl>h)mWY?Fns|0pXvQN=>BJQ%yxajK)*k~i{<&kp>srVOWm3-OA0SfsI=kRi?A -XN||6^Di+XkN>~`Oc^6A^4djw^K-QS7%VnRie3_W-hzu%$$f253mhP9X(+ZMZR=`j^WVfR%|8s=8Nl! -k3O)d_~z4$wXxjpAAP!Rgmy5i5_LQHBdZhwZwG(N66yHy$MtdsRLwUkP3t~-tECgLpZ=x|J;h7B~Bz8Rsw3jEv`+ic!QTwtik{e+0nH{QoMdvPV(a7KzojW;3qYlsL+}$-AS$w1O>ztzvRZnNIZ;!V)!}Ir?o?qM8ZP -D(}Z`|0u(6~SUKyQD@Z5SE+hI1#!VbtL{oV&XYBa82Fex37>Q$;EsI*r1tStbP+o!0^fkTTW;Ap~YKZ -IYDg*(D_bA}PpzFQY371S<;?q0oOK%pR2o*)h^4kE`TmoeU60w+o}}=Uoq0S#fhy|^c>F$ThEWs -puzfS&x?hy0;)G_tS^;zWD69OGrQ!~pug4cLK%mZh=PvW!2G-Yc(5xmiWayToc22bcLPH_b&Z#y)x1t -Vw9XWt{S_2IXP{8#Z(Nwa5{HGmI1|;;+juf$H)4Z4p;@gL0pmuDUFQM3)OMI+@@wS4NV~t?B|<9sKVCUHuz?kSNQgo2F+G0DQzG3 -8nD33nrUYG60$vBw>9hXU{~W8Ox$hx;-b^iGu8iG5_RLi0LsEkDM-owVY%p!iCQtTdajvWCxNYlRqr7 -Xe0nEPpJgTSAcgeCQ?V40bv}txAH~?AYkBdh!7oA0gEXhO%Rs~e6#?ZdmM(7#8+^Y2n6o&QNeME1B@2 -rN&4Uz$4*-@Nbq>f(ILPTag`@;RB*n8XxP)9r8TDZ5(Grh> -#(g8aV+X366ak%mn~QodqpiCPA?_2U@sHfnv>zvT(Tw)xh|n%<~-BbgrZ!>(GO;kTysh_S<5LBIz*V! -b5ZDFd_q~7zaqMKsFGhkWLFMwsQzjMt-7yH)+AePTx*C6ZFjla2uCVA>YSaG!g(N*)qt0B|f1A{|uw( -&Rh;>-nJG3;3}w<46OoMfk3_h5@EH#4%%r$b9w7BHj`jS`UToCSM)Mz|~GCR=-7r6v$M0Vgo8r{n~XmBSaDAY -)4%h_n_5GF0*U0H5t-jBg{u3@Gx-!*v058qWpY*nb9GOt*?)a9so5~XU$zU6%75U#HR_H`iKO=Xk41P -ykQK1yFaT~@yVi1ew`Lb9*iLb9ilLvJ+%JmzT-6!QFFH5zTf8A2LF>gD;>XzmatISqGe -H(HH%4*e(%4H3;(;}N=jaRBf4M10Y&rqda)#7(sz1}m2!1`Ui)2=jfSqj|z+3T&IUBxVLjcu#`@L#}~ -!6}@*w(R&>XF#MnfZCc$q1n*7{VT1~wocu(CkLPH6Ndagg6!?iBs)0aPY)Ue^B!DF4qgIS+PUBPpZAP -?moDenOfe8#+&Gb-6?~v8q7J7-KQRu6iMXULE=&)7PLi+|NuzqZz0|OM86}Qm01}N}7q#CNUO9h3SrG -*+Qs+KyzO<+f;Udw&AT4rj-eBE<4=k1!i-o|~mHtrkRxbND=KO3Mxz}y!4ivbE;C1|0)8lbJcGKXc>sjhj?+{;XX5J~d3 -u&FR?&T5e!tWPT`i-}P_|^xe$Q`F(<4ln>pmMX?LPuV#W@9SHQW(QBu$(b+Lms%ZUT?NCKEtf%9ECdt -UNwAxu$zwMC61HW4+fLe(i7U5^fTf43Wc2E-2f)rAG*o({#O{156n#XL;o$#2ZI1fl2UQ@1KrZ(kjOM -{wQ1~m40c;=GwkKfYn9K1^JR@MWupv57_^e?IZ*Fx`iXcS0m(lxXm&FEe+1#52#n*s*Qr5jL<008l0R -a8RN5<9F8gR7MCB6I{xhHEW~T?qJuK`gi+wZi{8mXd$}_v*JDqP_WkcX{(Qu;zkV|K{ec&d}UZs}ikb -xgS1-0>1ujhlKLM_D%KM4mnnl66H78=0g=L5qyc|ntyEL;vcJFKWi{xo|@}BE*q!amYn)17OQgO4bcK -{{N^-}hVjN5K?6B2cjDR84gU!OP*$j*N@N96|5^5Bd&4iFE4v0MgSZC5gP4JNyaN&0NbIlxi|$}=Gyy -g)s`##0yT)Bbcf~HZxG7FTK!M>@L;HqY;Z|i(zC -<`g3YISs?tt!257}D@wgxO3Jw{0WN~rf?j!a(&HNkp14yqWU@1T1mcG*LxL|(ahGjQEjK@_)81%ur}a -#m125ysg|Abps|JIn0QI7+^Wu*H1&5rW9G8E*m3XJ{gA0eQFNZ;A^T9 -Fl4_*ve3Tt}>u?Qp$mKCS~)Ng*Bk*h0rXiXcO$+|M$cQwg`ei0uS8k8|vYk5Y_^Et77T^b-`EqB!Ws{ -JJwnNhVuM#U0sqJd>`~3u-Nfq)@ge`a!p~xf$*>(*%$O^qa -oocg?hQ)RZjff{k6WS?@E+Y>=TRjCpb-h!)C5b=R%VB4qu42Cf0-93myF^4W{Wm+;0usXMC@Xp3D -CoyQd;Dx1X+38%?gO4|D-urKD+o*2gBj8jpD~3BZ!5LKL%b`iZ)A#n?u#I`!*QO -9>&!3JZ#!g9MA!Zo{u2POY1O-5l!?!m=|yUXlLQxH-82F_h)pYZDO<=Oz6nJR=sQW$CnCjj(saJII(F -(PmCoAMzwHoKVDZuUPdh+b6Y*bmb^0jzT)udZkiX^mdny^?`nMJGqP<%FaCx=7XrA~#Ts-mLrSb3;8V -BDomTvlgt$qU#?JMD5Tag_|;$5y@AozL_x`E%Cz<{4c2{iK+3^T(4mLQMDWILhOnqQq-Q}cFyb@6q6* -Un6Q1dXYyZ6Q`NV42oN}saDETNA^ -o}r%TM#dNY{PV|Ip7OxG5w+9X)a@B+ae7!D*YfiTyjzI+Du|UkUUW -a`dYLNQ?ixN5awLzwxbRWfR@m#n{qz$)>qN-wI0 -PqnW6UVoaylewe}R9V>N{2K)6>C?Ne$t5M!pX;TqGG=eg2W8}D7UP;J0pQ3*JYd}f%^B0C!?Q$ek2)( -rOk=@C3t(hdbXX%iT&3h%l11M}!_L!_Csoak+Kt0lkEXsOgREpwpwovL6)fqMO?`f9w -x4WI)L1>whHEA#C&(K~zRb*TIZ@jI9pIDt?48RC!3dy=s>I~1&>k0kIzR{v~~mXnU5_KA!-bga~oOCO -v6U?=0@_VW51&2i8l7`YoAKJoY<>p;{+nwm%QTuTF}cr-K0H|?WHwf1v8(M}f0H@D|=@alFHJ;mndn1 -5b4KLig+WhCJ-sa%6s5c7vzWVw~1y5@}j5~l>pJ>g2>cU8gWLrnuej!@N8ZM}3g;%Zf2y|SNH*JZ`>$ -Twg|TP4v>#(MHM{^?yxalDIOx`~y6f?@Q`ef6ez@3JeEpcQ>$9-#%sXFKL$c1yJKYt{#|R^pu_rUmRZ3u9_}8hs3;(XAS*h| -gCgu+=sVR1vJgm&{K9UxP%Jh7*8CP1y5%a@8%&ws5h=1fMDZojZ=PHuFiD;{(*|&iachq!t6=O1RW+F7#_WkYH*~w8R((+p=@<`Ot1+Ub6&!?)9FMz*^o;m2SuR{{c`-0|XQR000O8%K%wh9#ch%P -XYh{{{sL3ApigXaA|NaUv_0~WN&gWaBF8@a%FRGb#h~6b1zIuLq$$gMO{%#kDD+MzWY~ -XR%5!GR~$s5-`CP!4Wo%_TP7gWM7q#&^*4^xU|#7h2Z=ME7Yi& -J%8-ZqTBe=QP`#Hj7L!4Yg90@Kp!`zaGxAEXE?xBEiHo0>j138}qCLo^)Tzu;~`r9Sebv_L)^a-?%s8 -4gmW}s$cM?ztaAAVxN4g)bb^u1*LaJTG`%9RcpA17LLZ@s^KP2vLU=4nJpyI@+6CtDQ%*d|hk);m88{ -zlOnKemn%g`bQW9MvcP*DhI>crf~iR__+*0j?jf`WS`2M(-#4I2h`>k=38#Oe7y7KD@)*U;VV`{U7rQ -^)msL+L{B`0%QfHRxdIYDXe&f^(ujbG@05Pzzxj{a|E>oUL4_FJz^Pzwp-3UOzlQ^S`R!4<0am3SwFRPY?RDu&?C% -TMedP)h>@6aWAK2ms3fSzCEian4qP001NA0RSZc003}la4%nWWo~3|axZXeXJ2w?y-E^vA5z3p}zN0KP`pHEQ@jxPWSEK0V!$ITwvBTKZKw{3|>lI`};9H0mk$!Y>sa1}rbtNdH}`?I7e7R&i2SvPrt-->CmN}iRQ{qM?aC{rf4SH<)yxvkdoS& -~7Ubv|2{v#ea{1_9_|v)mjcv!cn)=lNb<*HyhYt!DXQ?Oon7Vttf|)ZG*6zL9 -PjK!K07#vSeOe7Sn8=Tom&h23D@JqJ$yd7OShGg!fAZJxXAZw)9T3KA#me%t2LelGRmS=ha1; -ByX>BQ^D&?lD}I*;l@2Ivqj$Qz`R~%H#tHmYN)xof`4i#*W~aQ$44DWCX@ZiWRyIDK|g{Sgcmn?jkTa -d*d3}Vb)cU`1ATc>!J@2aA*v>&64f%V-HI-TB+NS=9@f;AYdT@`b72jCRue@{-?-VEt#wy2N(?!M%W?9IujT28%Io1TL-2yjT=8L) -nHJ+DZRjEV&NBqQqbclcSU5lP4d;bY-)=hT+2;Ce3=etm+jFZZf&dSL0c>$_B&9WM`4DuBsUg7 -sj=yX6t#*BV59&O5go;Bg;7;e!dFK0uxq=c$>aW-{6d-$5%GHaVgG9xOM@;$Xs)0}C*L -VZwA}&<(`#or@X)n9{(@aZz4W$-gCoPe#dS!-FK*g@yS!J4n7 -edGg60;Meo3e0rT+0Q#&;nHd<7)NXbX_|#be*z|f`u8Ku|gj;gZJK4ZYmB}aRXShQUn-Kp1HGW>htju -~t+o`HC--^n7LqO(xZNDM?X?VQw6>(8s45%AGPDb?Q8(9B3oAZws#dO8rj`1n%1OD}GwnPx}W4%Ip@& -^!FKmYU5^S>Ow{F)z~)YYm2(#XG_XY)BWrMhH)y -@m08pGp5Jm|{fB(az9}AMhB_SpDR0=D}@NmoOwdpAD?P`F^o3fhtvV`*L}Gxn31>y9!I59QzuU3>JN1 -ek}!m)3(cJD;{ -6F0v_*?!4YG2Fdsh3gI^Auh4l@8K3OHp)<{vxVexb0HVWgKHne}Od8mq=hZuybB0+p2Hcu3ICuzi3ceS&#sy7FO4FROZHI=G!B+Mh&ne(E&-l38|ygq(=G=B5l$;qqNZ>Utzo|ysy9lOb#)Fn31@W)G%@-5*-;ekS3q`ip+3uEnF1k!sVlR2iqqPp< -G;bFGp)wB$o@Xv6#0BMh8`oO#|N{zT8^jHVv(1!RsU8M1R&_yJ?aW=4Qk8NNPXFM?!!@>?Zf2n7frLyoBgLBfBfkuFsL!@G$!AJ-uNvXZ{rt09qZzJj -WGS!-kaViIj>-=9rpBN0I*9nfV(L&=;_Y*We%tQ45l4m(TQ^k#YgrCIgLH*xmvw5`0R|2fp>r-D6bJX -cFP<{r>DKw4##_Eqa^HTSbJ+_?u|dD5gegqj=ea!zJ&4uM0MPUvNJw9K>%L45Tt?y6}}bL$**$!TF?j -1;+q1fn(4~U7~jlq^4U=dwDttZ&I+mlCxNu_IWU*!c;ne5*^1lDk;g>Fka<6@kC0)8^K9Yj;_WHfPmW -7K8(gP;EmkGx%~$3KXd1;!VeOQ2>Su(X^N$X&u!O0En^Qz0*kyMVxCh8 -Z1Hgv78e8@NNjz8^Dh3ydLb4qy|Z -BBSrWn+{|GfEP@Wsi$HQ7b}`QVF#(JbV1R;TQY`p8S;Z*WsgK54o@OcxXGwuz%naW+0a#`6+$G> -l)n`@cs;Yqht{Q2^>n^GrWB=%LBlkMV`30DDs-_?L4ka-oD*=y)G(sR|J6jXXM;<4JknHs^2zuJt`R%>JF -v87S4lZ|0NZ4_84PXV#uUEJHzPQ>Mhr~qRX{!FIS{y^`MF-s5No)Q*Cd}h@kaf{_8OOJUCpYgoa`J;sLF9=I*dLIK#yu#L@YILQ{-yJE3k(1r!>7U`?mdVEov -0ioxs68@WFI}W-?$KVjJb8Na{rUNzv!ZddQ8tVl#HGTDlWmmW6fHr~_P;e7;B}u- --)0{v7F+%w(5vnOlfspE$;S_K`-5})x_$a}}{t+xox%__RBoTX5uS+C|k5DLD0dG5pn(GFIhEU$f|EL -N917?Rr#WW8zAfSz9ehb7fY-faREY{76<$+$Grt2E6$}3$5r}NbtS&HKeUOu>S&I-DTR8y$a(7r?VuR -=m?S2rZmuqlfbi^6OX`8Y!_$M0&xWI9kzUr-dO7%HIJD428u2|+btQ~2pIz!IEvnxEIA%40zz*S_9s?QydwZCiicI6 -ntSP2W3JnvSH!S;|DGqNxN0vWX63-7WrYLda{m>U=>yPB;x*Rvtx>&Bv6WXo9lj&!=0rQ;rA8Kq&jp_ -3Ic#$pJ4^c2{o}xfggJ^8eWRPRVx7cI2mb4z@EdbBm)(aHr&5FA9Jbwo-+O3kbN}lc}N99dXS0!HFvN -T1q44DUdS1lZ@cWxVSkTLPrmJj!OXF|fJC4G9IPOQ?V{uP=Z(_U74k3&}Zbt2(2-Hb$|Ud4;D@x -n4Rsg12b-MYf)=SUO@yb5a+J3=jm>RdgX`83l2BIU{0Ujo1MEfd}-)jmO+izRgOYZgw0mBB-LSzvhv< -n7Vnkq6*i*oZ!%IGBt%PJg>?NAfl^SyDzKNS15ci??fASgp+7HU}scA-%Wln!W1gyvA=E(fdL_bj?h2 -eg^a5tBLiw{=8}0J!duwZ83*hbZk)hrrF%p(+k&o!#=CVGLB9!l5p8EQ3pxFvxVDpv-3zy5ms;` -q&51Mkb@m(O1R=skM#>buv^kGw}DFnsmhTkqkEqwkNtd3B1FRl{fLU1F}uUE3RlY(hF;IEihz-(Ofg=>9o0a+}q -&fMcx6xjAFuP+eq&-K)=x9O6Yl+t2JBJ0NreO9&k|{_MOVa+RECO-}sVEMFp{RZcg=5^(*5gvAZrma! -(^*nror@w$-DFVxM-T$bP|Tq~$2<1u{02?f66F|z12RZ8*p;6wre0UiY152OOmP?L=k!p&YGVpgkX#h -ptq8sSBHp5nPusu3~d(eYQu+({^#0&UkjLN(@U$T4eY4CU^e?U&tHG!f*ehF|tc -nV+!^Y5^umN{Jf?J6J?UIDxw>psR;1gHG;$ypaTXwN}Y!`TAt>Di8|j@x1@Mwp36@g&JsgGZ0Z_ySnq -p@W0dPYNwKGeihxW@3E4IRH?rQ*S-~i{vk=O@k)*Ilj=fnYl0bh(wJ{=Q4fU!4~-8G!S8UNh5M2x25$*VRQutL4agQ6%`9 -@Dh$Gq;ut0Kclq-nR6jFV6;qKgITeeU*|L)!6qz~70w~NfoedupNc~i)tJ|Y-m)E2z|*b&+okJwGVo+ -(?49^OBAfo8Qu^3(jO&qq4%K;e|xex#-VIKaiuA@@AVy}8z((t>S>)aV2X)zPcHHd -%{qEU9&}mUrK`wx9PE1@aYaK~` -vQ23J_#gfDCk9QP54!@u7c+5CFd9tH9ad@!zROFa329bB4dT@|LMi$HhoK~^1?r4c9HAQ($-8St*ggmLSv!RK>FexJP~v{Rf9xi2|EdE_wU9K%u1t8p#eAYss?&RiWj0Zt!}X%?Y>|X?Q1h&1 -nJIh0F4Wwv^oW4+N_-uro#lctMKHqjqJ?myWyESi_KI6H7puBgyG(+J2>*X)}lB{8kOP=OjjMMwgyOn -^}8V+2MEX$n@S%KN!-dmnp0JctB5%CFZm`AhK}V{1_V;K^+529Vk}5TGX{YLMBeR`K5MBxikV!-xP4Em3P92rBtV{Pgk+iOIZ>+paMhxfTcSYTK -4fHkpgNv)7R|9&`KOwcvO8VGGy@Lm&-G-i2Rq!+fb*2~ITqjBLh -0=Yic;i*e>gm-$tSNn!-TVGYb(_~G&-({Sf4Q6@CqYlfuvBO;-dveCA5h_Cpo|g&PAT0&oyT1YH<2nL -Vwk6UqC1~@ykx@%%wPZd*W@`XhUOcFCZ|jT1O``aGTEx5L(_Y_;1;Z_D)W5SkX;V&8PjV%*uE*s^*cI -%NT1;!2xR64dlxVF#ih1Hwbr*gj44(T6sUm2mOG}&p+=D<#Fr^hLf9yXpaY}8=_~AQDSMRZfZnQJ8Cf -$Z1AAkQR2g>ar)S8q%`+goy`+amWIex@pi0lEk!I;fv%Qa+9%_>3Ucp57Z!&$%71N -2fj-LSr>M8hiw>oQRz3Ab;+@Pwy6VJWVt{%(u;oXA!HYjssu>&vS|yeLEm5eP@UIvx72whAz`YmuoQ{ -4{$!{3)eBdxP}R5LSkJytOtL@oEP^c4;pK0>2%Nu2D6n?Dd#DejumK%rF#qbU2ZTe(LA~n(Lt}kYyAM -&o`Smy(}zRQn3l&#DH_Jzt|D67xQ_mdY$8<=Y_BTEs!=~<46LeG23@ibOYpUo;pJFl8A#Z3MHL-NG4d -*ye^A(o8;9Sw0D)!nf-3Ygq4iEDH4)BNTC9 -yfllk^X1@(lKPP%pb!kuMe=Qlw%=LJ)vKCoJM<^IMr}c9qJUv`A0a3nnjzNo3cyrc75BwBh<~E@9@VJ -{xw=cPFB~I&!I0*pQNAkqQy}ipkRLX -Aef)FVGi4qi66a_^mEvCmvGLLEP~e?eWUWrY`T#(jUAQZY2kom?258+%*rDBTjfk*WB_zHa6fomPEmzxwmW)hLcO=(~v75UtWI=MtUfed`{0p%(_pHB2*3ubVz -}nYS<@##)9XvO3-$5r`=HGs;#-t0sc5heA{}qkMtgOUE5=@oC*9LclJo-KI$;+K5144SbFsmy@CyG8#$ZURl^(XT5Hx{PZKeX2Z(h -bb-VuBRyMxF;x~KGQKoqBbD(aI%~NR|ns)wKFpwJi%IH(#>ee1vy{hdWq-d5jhCtGd9IlgCSNW@Sbqf -D1*TV!kC42O9tDVQ^qKvttVE$)&Q8Xx$RjI;?D$YOcHH}Rb3<#g2}Qpx`0iH#i3y;2R4I4((fpC57X8 -d2srn37~QMZB8!it?lHe)!%;F74NQ66H#hjRv#uTOTHHFpK^qFPzsK(hPS|G|LF9IwdkX{bctv=1EZd -8w{8A8#+;36*UJ^ik@ptBL_s6(c_mxYJB1tlU$BcY54#UeowD;7-0HIhi-LhAS`9%N~85>t`(Fyxo2p -v>B1@Mb1=hIYp&3@xkUWrqnrKAG?3(wL4hp@F}`0F+Nwgpg`S0XR&=z6pfHmQm>WxTQj?IeKu5sk67H -A`~6>VmG{=ku(*rf=d%noyJO-2W!eW^`LeLkFNHbMm3MDpqV1gM11OV!4FYNHCKF2Sq^GK#9Q_XcB&5 -cQV0_&_*a~=#A^j^a}l7@eY9OTD2%v7#+ZyAv-%{MtMIJIBp`V3kzV4!@fw80l9=tCPrk*;D?jRFy6c -veMUQg@7hcAjm*6500)Q=)N>&)R8LGJt!_BuTIm4)7%P7gYq}J5q3{)|0Q=}&K1EkOH&G1I^bgJX=S4 -nqa0x8B&D9%L{mC{Cbr6Hr#vUcX&PIK><$*{SBNDn4#bx*9*WD@j-iE4agY@7Vsj4SZNjE}L?{qi3_Q1l7=O$K_79 -@nwqA!3VopERSvaF9$Lz-SL91mRZ>505Sb- -0rS#}IY&~>e=Dw0K6`ir>1gBmI*kmBG@I%wTp*BJzqY&EzJH6=san>%h8DHm{#yJAom*?~*^1>Qb&&I -22Jk8)l$Iwn)aHgV=Y`6nja$&|JLRnupVO}2ERd98Y8Zk=^2w{fm)n;>{nOtoK^sljPBDCN@MQ#CE^1 -jH>EMqX8HywtBCp89(;**mL>otp81>WKyrgNX3kv -d$J{-9uMq!X|Xr@`82f+D6UaXbBaY*J~-21OS1DFpu4%x^0Yc)Ig7=tx)of`=U~+DTYkpRl>>g{2&78 -p(x9YX4}fQVM2o>njNEfs{*g` -0?qu1R-{Yz_i}A0A&gHGCwXmov6MnV1Gpb&%WPnX3ZD%na?}!AzaF!xD5VxQBn|`(kYM -^AqHlV^d{B^f6DV`a93{=vls7J`S^J%qE;zL0)5X~Rpzv@tWaBq9_K&biY8eEnxYhmFlID{VnsgMcs0 -@UrOjfQe<33pte(P;A*!@kertM6Hf+}2V7ql+%QP99#y}**C97hjXvAHhj{wxAt2+fbRJ+g?UE}oKSY -*U(HxUfF!To?Dp9=Rq{aWO?xWya2S9m62)C}w6PV@kK$v*BQT#ftoQF=yuq;L!YF&-Rz6FXy#3MRTIB -iIQ61He>=B-*nyI%j3|&^a?QOvivWGIhEQjZLDV2(bxZ6EbPW1XF8Er6b$=J9%(N^KNTi6wKnPWv@q2B9FKy)`&+hjR207 -L8paY;Pv1iF4`Dlu=Ed&tNgxRU6cwF~;KGQ5`QbS=*Kp}7YZW=fF)HH+|IC&nO7WN@*1_K5oT0xU>Vz -BHD+}hlLK&3|iWZo&Q-2v_EmlQmkb#&?8FzHf^_qw+5^>Xh+(oy<$isS#z4X**Ini$jAt)c -oN=t*m(HH;{D#km4^}W4J|xYOV`{9HB(j$BJ(%iY?ZnG{!!v;n_vYsfmI#5wLm&OLL*cxGfjxKs9(eO -fSX~&Bm3ZVL|cJzVV~%pD8F_^ca4mOJY?$71~e5SROTKHJs|1ab1V)(s<`O?Xa49P+AFRN{W-JFSgHXF{v7Q|!oDyGiIGR}>5@VpsWf`$8p*r$pekgf?rMYF -em>;(TNANxr?sU(ek*cSFmG(-G6PaTwA25(>z2f;K`Ct+BY5nbz3N%qOVkB8HQ`F6XERn#ckk -;<8RWMvz*hE{*7x>!ggvjZ3?9!&Pg2j^yn;NAYl?L_qi<3*_^hfOIi~WN1aW=4}8v!1%?u5`RFMYkNy -FqmaMDmJD>$EdfRDjoL6+V@rs*^ze~D1TJ5JOvH2rO^M_T(!=(NB{rq{fgP^oU*P8|LFPhxn(+H`5CwmVJO4mCZBJHET@|2+18c--g9LR&`4ek;^-!7QIN@F2?9&IZzgK -?FRY=>NX0!Q-)XgY6V7LXhOU^>#oeYUoPh-pFq`cWyD-KkM^Cv?7=yfq+}lH*Oka~=6c%eHBX0~XeyS -*Ms%yiCk+ql>K$7Xtl(Mj3Vc^M7g{8#q?O;E;bAQ4S%dFXmE|ZDI1f2ZZEuIJf6^r|w&u40+Wu9}yf+ -;lM3uB_#aQJFHUDpF0D~Vhwq0bE^U|%?B|+^^zgSCxHhQw3WUz@TQj70mgcPZYlIaY{*HWWZ3h@S1UXI+h}lD9gd+^ -jV_17Q+PAH$TkUvMdNVqlf?%0n1Aw#4f^503;N!J#nm~roJ5T<7YHoeiSXJ-=9|lPozLBNF)`$I7jSc -O-l35V4J#J&xan|DHh53qRi|-BsXKeG)N2>gHMo-4ALM3*A?4}fv;xjI4)Bl`VxXT5PY(|Eo}LAuD)N -3o8$2=sir5Ia*P&Eomt+ZKfvq_7^PFVM$|wYwYlA|8MF^^z1xpUmUhQX!V1hD0)re#zV%s1^Oe!+~pX -m9Khlr|cZ{?lz=Y8cr((l{)m&xe4v1h|kay)n`DmxsT!D?MNoo1!6;~) -j1Ns)zu!E%?4McpOZs>%Pgzq;7LnXiUrvWgx4%4bUNmn*>p37p|Y!_$t$P3b59wR3xFjnnM!V%G5Y2r -29AUfL>$B7AhmX4EKCxqE!!S(1RgPN6k|j1K7KQDJdYP(Y^3qh101MeV71oGm32z=_E}{zzWI9*KTvi -Y>`A)08CfhTd}L;VttlgOi_L+N2O-E-M{SguXlI21IwvaA5+nw|5skw3q6#N7pQl0TGm;Wod^eK -q0(5oV9sqr-hh41=KYB3pKD#h5}=mks)Xlw}ZT2qmm(Yqlb!;ns+{T5m{%b`u}&z2-!J5YnM%$D{>%a -nW}Ucbaa9oN2^bV&&>iMyVE=v&N$f(eb7#W?MiZS9!6n$^9}C?@~?S79e`fA21PG4j6uDb)3~pw5&qh -Q$ZNq(a=ze&&KSX9c_a4)@>8)pEk!HdDem#^{u`eIc$hGcktrYHR!K{1GTD>W?S!O8!wP*TF`!VMH95 -Buf5T`HSVvfWy7%dcm&a(R|Fb`ET*Balx0j3r(CTgE&SH?>N(c=RsA$VQy(<=~)w39>Lo5}O&x(5P9Vs+*!6AYip#Xh6FV35&u`_r$P+GaJB+e^Pmdfck>Me1o`qj`#^xZFvzT>_!$)(%^6rmKW -pQxlVI&xv%sUZ-r8U|SSZJ-5+q;}WQAO5R6+y6aU1j2P-YVJ5@`0oCV@xr15stWQ~$PC!^ji@QYRYAA -NZY^Mxvk!Ku46fL;xjdJ2rjHqWwZRE%eN{P|yIFL^_TJY$LKhfbD7;KETWnWk$0Tk(x1oC`i<5hFxMh -WAL1kCD5)k$(O&n|7gNy^|z{P=AyE9qpwD~#~F;y4%tHmT+_mzYVyvkDrqu`Q);YF1wUDq4RlCKph%R -PO;ywa)8c1F$9^?-2n!NCp#c&L(I`g4>9^^C!a*C!utc;FoR-azl$Q6l#Z5%>a|sR`l{MF~?v>emd6k{#D-s3*wsyfz8wo -n?4WFj|VEe)8i-Wy6-(FK#1rsycK!BZNC|#Ew&HZonSoZOuVt(BCTK9a0YukRTYdKnbmCdSKJQ1xB(` -3l-?#U=uWIbiIuAOIOsQ`Ga%mB|@_l}QSGS5_Na=~$`th|i_qI&BXxFIUbT{oIj4I^Sl$k#(Q(;>fSE -3_L7&vshHTW*354()Ar`gB86US^>nZPqdb0qmU@m66cpazb&Z_6;`1&j{P+j -GWXDAEkRo4#Hc$f%qHAf_6jEY@4kLJ&kZ7~suxBAEKJ;O%XCn$P7=6PjFa2XfXl@7R8>`YHrN)gZQg) -;eQG47=&~1`MKk{uwT!wSs==I2W^=7N!Xw3YXT*jSwow_JECk)wVU|YQZ+K4J#nng!vZ0pMw -ECYU^5`_n147QdzLp341(AGmXSfr1}$7IZqOs0-QM}!N_Ujl<|0nSz)HlTtEd_1M2xMaT_DNSxRXc7) -9HlKIe3EOLiv5B?D$uY-vZ;D(|BlI;7T2{Drezw63J#j+UG={C0l{8(>c+@#@s#!;<_Z?(s8#-}uJiE -!Djf27$ogOGDjrUUQnK>78q>QMoI^S<~%j7~MR(#P`lp1!2`)rcMt+7%f)aT|2fUSx&xN|Vf-WPzp(&{V|OI8mg2jj?v)rL@8iDg(K})Fl2Kyzg`7&;zdd!RHX -kgOEmQXpdFH4TeL^pC${3I%kp@D;Be*Y!Z7F=|jT#tOu&cE@+{;LRutxlVjw3k|H!^wrsAdh!ql^BhX -VUNrr#%*g^hbseWNS%3!)RppX{bb7Eaxk|i9zJe?(jY~)U(kwI7B1bcY6Y-F@R)NC!27D~tz=i*`A^r_84urQoE -r(^TT}A+sqxojRce$XMki<5N432~rUdG|%;lX%d0G21u8?vZVch(bv0h=$qOkwK!pQL~2PbNcWakr%} -eu7T?{8+8Yq*EohM;VHm^8wJOee}YW&JQ_M2pjn#QQK5&3l0d?6{pr(CXe70c@(NtkPQ3{rx#!~tr_+ -NeI~7i&2|8tkipw=s6FBzXLOyzUbG`+say~(^xy*A$f4|0|0Rm^25J*1?gQQ+OV=HdBa8&+*#(IoAg6 -!i=zx{P?N>YT3tuPWK_IlHOQnbVbigdViBrT`nX0)^6}K38Cv2KhABqZt!qsw|af^0$#d%n6#=P9Evq -{goh!(x~!d*iVlUMt`#SRZO@b1&e^$jO0nhAtK%g>RIN?u07MRk+k69-;XVYBpB5u_%)6S1AQUMfuno -WCF{1v0(o(H0y!D7g0NeM>WhYE(qkqKz(hA3^_QCk{FVw-l3T!~+Y=!IrcL#cr>Ce{12~7pO(U9JeO# -amJQ^HX>8kl~xTkE@RgPx+3;DpLw_HEym8=RwL1guztigFeaR^H%zQ!uUuM;9R#5C9kcwcAigE6n%i4 -ddnm4LR=z>07vf-|TT(jy@6eS5vm~HTOI|hwMX>Qp^Cqv026Z36b*^cRGZs5Jnq~FSA=(?&m@7ei2jB -}Jf(Wv7tdTjUAHzOLoaTa}2&j`d42feJldGWad`iFuEh^>_2pf_~17b&GyLR_Kt(5CN;OQ+YtIcso!O -KicBD&h(A#dbt=NQ_7u)c-sbn9dogj3Gb8N<5Dei{U*i0`dKwjFEeSZ#-@s|Kcg1Syh)1|P?50sSH*P -KJgKF+DB@=}0lFAvHxW2g*S7zs0AuGh@g<(=dOf -A8KcaGSAxUe|Bj9Fc;GM$7l1HQ8Ik@G#nhgM;HcVc-5dMnynQhmqv_YD0Up$uUcaf8xBxIVBaJXv$(| -a&AMn?hv|5R9NjE6aSZV)L?*V5{fLL{l?SI4sBQ4J92C?~=v4uQ+ZB_CX -LO$zNceqD);P4yQ^U>X>}a8f-(_gul`N8Sx@w0ep+n~O`$4c9tbeC$)*@UA -mbnTttr*AcJBf$iuNj*RkvjCvoGrrWl`em)^##U>28{8(6fWKKJ@T=;EDDdcF$?+ym$j*j??9@@1I2f -jq}&pqdSkv|E@W+6~fa|z3}$c!%oqA9gVNp^}8#0sn+UGM5=dFWOgGdHLwq<5>kzT@ -{qhP(hD -OfL3oUuQ@54(EZ)7uER^jYDb{gi5Ra4P2J8xE*Cdl7e0#1POHE8MHJ+@#T$7k061FiuF%sEIOk~lJhc -QpxqdSBDHSvr@*r|0{{JhQs`$AzuI?R@Jm&0Io_CiI5Kf~6+K*vV=$iNTBro%{0dPo>++sFlIDO!pC6 -at182YV6#Yc?xZ!d?Z-MnNv4!ZJS_A2r>Xx(+{9yg^0P`969w(+UvPQEE#j`5~ -nU1KXFCwR=m!LG>g<XZlSB)seiZbe3x_jngk3V$l0L2icXD?H3N1`Pv7jW^-Z;1Mru -Q{*H`uZg^hL?fcjSHJsh;H$;Pb=5hN6%}NtwtNVp>5=jL2t59ZZPPq0|glXV@tEusZ^-HG4Pp0;p;0u -Mu3)A972#IrX{I#^h45EvAsKbTMGa+184x%<%WM_5)1sCQ)Jo-SOvvLn9-~1ay^^mlA|8=cYf*<4h8Kli-dOeEF3|(oRGergNknOQlkUl|>dlVAm1DQoM0wtHGm37`_?v%?-V3so| -1?P&)X^!|@20jAeXw6Uqq!IFdkDJb3G;`E=lfc%k|qc5W+yt+D}mv#L*^0*bX@7|~PioE*H+vy@|1<3 -);Bk5urw5taR-@iaOZhmj1>lr!>$68Mvu7fy-|5iXw697G*4&s_$Ua+&xpG_3P`;fzF9=GI|&32E0&c -sZR{ocIiO(ST)Bs09)~;LxDZmH&IOM4-PLYOy5xoltl?Ao9HwK=u*}BPP)#WopPZU8uS5nyyY{XWK$E -L!E$GM8NA3PJtfx=kl`O=*xszrVhttX^UL&Z76n*P9~?2Ic1c)jy8RfNS4I(wi;zM4*|cBq#+Ia-e9E -1(gP&bN0LPee_S(veVLai_%eBs`Jky}H)BGMa=lF8=AYMF%%G>MUviWf5UZ@F6PF>FM67=U0yEIi{XL -dEPWpR5A^M}_5%qjaBFgdk294Oqobz#P_r*Z9+)^?fxx2cC6%LS(bagJW#@VzW -)1ruFmWH&kIZ?iWsY_}OEx%0W9~KFu^F%Nd|6@aK@^2lYWyZ(zv&eQQz(pWE2)gLHB3T{ei_i+4T|Q| -Wn!fOi#!5?kgqC6*KqJ=xlGin)q?Wo!@G?m{K9MliT0U+z@KYT&2rxNh`Nl8DOz{Hb3JcV?tftnj8*z(eEh?!w| -`D(&i%(=iR~E!FLDEP`i? -;P_cd^h;r`5G$Fn0v_sB{UzRFPv}T6lNY}6=^3!tVw}B+_#>$HBnt1uG#Hx76b(lfNxsdd$rZ3gIZ9P -c3S7%t%&Ic~lWY@f)QnI2ZqsqTm~ykavThaGoX>9oG^;%CrWVT}#indW>DZ1}FTeRQ$u2NTs -pQU5mF-yk48&$lL@w9&gyh{pG8-+yav6(w0%`KNHu)@QGq><78Lby~Iwe$#-SEB?kDWsi56@{SZjolj>jL<3WIm(<>-wYmNx0wn%U9cq?WR33p(DF1aj;gXjAJVY*#$JH76_ -T~)+gv#+#T=3{bH0R*O_!y%~Aa*F~Ox1;MLtR+q$9qS|CFc^Ciy4iUC@MM%8;V~y@J2m(LFB%$YaEvx -;{bUl-yU8R<@kUHc@n^MG;lo=5#l*@$vv_0`ScAi3N9*QlO8RW3(uQxa@C^SjY;S%(j`jSbLS7%&HsS -+L84q*4)MjiYQ)z4&V`h|QhfF5Nrj-++Ff@6_XNed)gpB8$+x%wJy36JW!~C<68R7sD5yF7UM6MI|=x -fNFSuu69e4ejz=cj6!x`q!`q^nGwY{t~fbyjC^-^qO!(XPg9xFR{S*_}f?;=L9XIuBk}D>z{`G8Kf=P -P!5@#hf#YC)X+u?ba7@1*TtCTZ;|@vfj46u%;c@ECo!bX}oKM9a)#14Y<2s#B|x;<970t*)q~#lp|Rd -CmfJ&_re=Z$aWkmpGk7$YK?-)6ij27=FYh`w2U_vB9j2`Ed=cD~sE(fU2|b> -0do$Y5pL;E4aWj@FXnAc{m}f)HM{F|EUnk^cqf3?e4rr0<3=LcIr_F;tP04w_@d#$dq}zX -D>^O2`IeE)1U2+*R7iJ3Fwmo|UUM`f%(F_-+)6k>;YPckB7LG=Ui0B7a)FtN11%`r-|I(bqs~(}K?nW -Q|9rOhAKBjDpX~iHJ$u|AVFi}pdV@56GV(VVD*e6xc#R=Dzx053BCK5(V1kgT?N`crzg0a@sRE7L?)Q -F|u1NXg5?HEYCe>Uote2;C2$rcH^b?$~wBNF7+CI6aLo&Z(z-Kr#Rab16?F4^i0<;i?n@seYOi -V~nS}E9E)%uc|zRy4ZJmJ7W25_XU{*(0MPuAz_arYhZr*ar*d+$Yj9tU@9Is30L`iPAO^Uv%CA8Qt4Dmf9`FQ`{E?~X&)xzlcSSQKYcR(>e)Bn -eEID8Us874w1`H9QPhOxm1z$m)FD4T?cuN9nZkAQ7!VMra77(piD4q)Xndi|HHNq~ZJ7<+C -S-j4~}oBWc35Mr(2hhEC6v|id9YO;3Gdyl0*@LawpcprLZr+I+n@BIjNQ8Ed`QAkCz7V?u~>TlcjroE -0zd#YAPC7-&~{h!iSomADaQ?%FfXgP5KER6rfN0!8Nx#wwh-?M7mXEj$Qc}>Johdi9fxZ3)+&hCAv?e -WBn`x3KfX#HAwN>-IK0orzoTXWmW1$QL5$f_bmoB^_zH`8p%4A;U<4XkE33r=4xMXAg*0eGG`O5=e&@Xt-2?^F-Sp1o -H=A;my|Yn16#>apD=!Y*l=5_5lyK~Y&!zLZo3a)SpKwBJWj+Iqfi{iFpiHM0 -buaIy)@j0h)71j;m#yr{r+(1+l#dAEM)*4_R)73x_L1OE+j|0>TgiJt23Ut -=CA8YQKh7OIes-T;mcaKf{Q4?&Z&wUPClfYR5fNT>CGQrgYP|)s!LyBU(_h}c5L>j7VA^ -NJ21p-Mtif>Loq-VJQq;J5qtcIkF5tY>JlsGjCrb!L?Q$6)91COD~JH4>$m#+3;L>~Cq1N}tbtIa2fj -T?%N#{(E`ULwRx{vZ)~k!X&-yJZl<&#{-@l+zjvsh84V6#F9EO)&;z`ygB0$fK6cq+Y_=}58P=8fd>+ -^ZuTvZik_oGy9z=u)RP?P~%$8hV7vDz$iU$qbKfHlh+c=`A+>0>9}e-QEMN;FUJKlF{Nm^D}abkPivU -v_k*$H?R~Q5@}X;gjhmmpN5C_3$U$W>LR}`|oa7>!JJGj#D?hZ7)i@_|V_vVvp0+oq!fZ2k+Zw-+%bJ --Y=d2zwt^0*3tQ7J^r>S3&s@s}qFzeFOBz=&zhvkv%c`qGOxBvs#& -A%4Hm1abM$O!+?^U0&u>sn7kYPMQ8-=|ap(rDt_J;vE&!+9eEj?s@6|k;pcHEy~8|BoD90#=ZUskJQbHB@H -k?k+a-K7{t(e0W;YZ;q_b!n7y`AlwUbl@vPe$_AE{qLWtXQ7Xg^eu9N{10{T!bFZla}9k>zT?CNIzqN -O(b$EdXL_6VPq6X?rS$-tRIp_-fzFT`g(J?U$x%TzY+~e_duC=BF#jr_UJJ`ujzXdCEXtj9MZesg -y{;;Vx@s~V_35wVlqGaJXDQO*!m%WVGre*0+0@>|WPJ373@a@t84wLutGXvoQ@o{2P>?5nG((nMdaR? -FsKf4_mwEmyFIU}&3VmKN21UhbnUezV$NUSB$siS%lyXHkqPiu1Ob$90(bO+G7cHQ1+(A*uPrZ)V0Q8wL$-)s`k&bI$gb61yG5l5Ik`YAqXZ)-T$NRvBQR21=Mijqh&5}9XSVh$6^2tUp`>40*OR~PV_%q5KDl)3t5{+>&I+9^z^OzC -W-x6cp$Pw4NL^Q^pP>4Aiu;3Dl2^?$*-1L5BjRhMc4dJ>s2UuOCX=jfSo>(fZOIE$pTnxH`zFbrwWzaqTImw@ub1&U8YJZ+8TuCA859)kPEe~TVA!XD0Ec -QyLTE9D_w{B^;0gnxm>fSRJkd2kjNr)NCzfEwX%S$u3fj3e+DD%XbyJyKwQLon7TzO~2^S0bey|!MW0 -YPdjaq*DZCpLBrdjlBGf!+w-iy)c~U@(RW&rOI_5k#T+^*M%FZ8(`6x{&bx&l{#{VSYk}VAUa|@w5Z4IdL?jC_2&Py>eRYQ-))2D$2;R|!_WtsASmpY>|FBbujB=XPT;5KvP$e`oTmRn)X~GM(A>@PUyitP_vd+f#A?MeBLnHbMu!W5W`lHx;bfSos@x#`cT*A6Ze -@yabN(sM=`kw`x^Wr>RWEf;LM-Ld5_|aiRn?BHP#SSe8e3VQ%L07P^Qa4HJ6j&@*oBN&v?1v=bs)0=- -&qI@05`)_SS#_2jK!UAX;*kST0v}d#Ugs0*j98Rcd0lW!8WjENuHa^87YASrF(^+#J_2Omp^%h -XRlg%9BW+FuSlUZ62omiK?{1WqV{!yMffwhqEYSp1-FWVa`bJFD?>}(+Yn@S7;5c-ahV{e%sv8@22*a -UxOY4^WKZ@nISy^1*VsUmrQZ4nq%@-9~it6_m9)-T4S79&A6ZrRHnV15}`MQ{!+~^7iCOEK(__)B)JC -P_Rur9~1PHCRbCgSo1*mYA>B%mTX{>_^cj$0D#8e;sUniQ0QgZ4bYt3R&%(L1E* -kGihfw1l%wl*msFbOeh+!aqiNe$7(eTCkN(D`FLG?^-E|1i9zWiiKMPj-@!yC+;jCD)X2i5UFVBShYv -hpFyzvqUE!b6tyHkg22uN=c!!Nb`h$hn?_ud>n+B#mqfXMN3?#=Nn$`_eY5gsI5yO?4pifz8Z`^#m7# -oU|T(?X)%MqUzC5^FB}q836wrvQBs{ojvK@>Ozw#&Vz*EX9|!tJm{3#7o9VT?rU~+9GI+h|1~1VshZ1 -%w>@_8{ZdffH;+eZwEih>5L2>~@Svhio18ebxNV~cr9<<1(5$vDAi`M33cd>A9>UqHr9;=jJQ|O4;d9 -Zhv7ytT=c?Bo -!a6c2hNoAkiq;IaT}5R=&CzXRulg1JhqJyAQ>@fF98zp0~^17HTj{6o(CkLU>=~>gj{8~$w-balUhMMWa1O -20hqHr_Zk+ZwZ#q%Od~4l`w2$z -(`$Eb$nM=X%Q5X~Y>5-RB6f6X+tG8>1eHUlvg^S8d4w)Ua9-TzoMaD!Khee4LNEwt?V)V{Ru#^Em(HH -g7te;49_e5OdIo`=WEm8f=M8AL9+E_JWO+l?%wp_4ei)(&x@HZwtuea1r{2k+;s$_s0kYlQYDyro?0g -Jo^0!gU4DUuUtW -ajRxJIe}GIB1a{~~DJYYPzoc@INGwhgw-jlK4pl!CVnxFb;&sG5_YDn_T}s{X4rQA8fqC>f|tWWNT$% -u4Cdcc#7@i9dHO#O3Lx(ce{9WL<*3Z7|A4xo9Y$wu2PUne)pr#C&#LsGakCiX_#;RgVrUpNhvrht7uxd`pl2MGjj=h)=NKZ|MFau>^(I=2k{JcTKD$THgiKBImM(WK@h={v^0%R$Z@~^ -cY3QbzQ9PQxi@M0m8PGC6bqszCQg4fv8z&EX0;UU#s5UN)hA|j`-<3nZUf@$1UQdE+r*$vli%rz%+Fp82&;0Hx9&ykwNhUC!7 -v~LG8KP@|-!>Z`MyYb!0Vv`z;`b+|py;!<)0;TMMc7F_(nA3gy$#7LwP}g@4d$-P1CqB9oY+(=BtL;+ -pwEw#GfuA*PP50c%T8=h4u^I>)RuywfWv%@Jx?7=UoNmXf(b(~mK0E)?8@R)nS{|)ub|Ke-RAU6Y*XiO~uQ?eWk&)NuLR}PB8-`EP{V}~5henVuWZNrO}yulN&O10QFk1c -${;_vu1uEQXtvuB~)snx4(E%>_PmV{Vflak8=kviY>Sd*h<;5lykQjCXsZMJ0DzV{nS>$W+jhmxV+u= -(wtqj4+1s0LxA)f7&pj;hjc4MFwv?oPR4-X~XZ++#4|3qDZGSvE(-$;n^XYKpCwcwW9kr|Sci3&t9<9 -r|td_Hk_ct7_kXF>y5Mx@fHe>YQ}XWP8D$nL1*JG2t*26nkjkF{jKaH&CV`IX}pmob#!TocXKd;s{eH -i61vXlGN*zRYT>$kjd)Q)Uc{jYzRU#Hrez^fM_#D4j@1{Lo`Iql$`%@9OeyrE{(t`O|I1T5waO^BW`A&!&#L#{PSXcJvMVbn(0ZPNF_4G!cW -=Mi`^=ZVv*mtA?jo;bDS*;+* -`z9B$QGh#LV-un9u{!3jV7*RHuegoKL*+`MtQxF8X%*jT?Ye_hIX~YEuu|t%h1-=AD%ur^I$-AovxPJ -q$VYn8ntm~TXV>M==P49#qUbAoXuz9vRQFF2M$7lfU$jx`Lph& -jF*F~(X^p2;zy>{s*xy{JzyatGdt#z6h&a6=+}nq*7uV?ByP7y}D0;G$|AP%N*?Z|zKs*)9wJJsrP28 -zm7?2d!Q~Np1ck!L%olx)U}pDg7|ZWO9#LBKeAhc5{q06R%%3TL=Ci$8Xvh6mi=GhU*<0#7hPqM1m(w -^H5TGL_5glX29BVaQ>@oH6r6!WK-RqMUXZ!rnMw8$*D(t=rze?qG=8W2e#R`S-&D797&pG`zi?=CKKP -i8hbdF0QMM=oFi}N?=nNq`Swcp|RK777jq6K?e(f -iWFQ;HFO>|!5Q@DLCYixXo_agz;~-5>roqlAYz*#SCjA7T`OX3;`}$Zc3=L=%yojtp -#1h9p#~gUE!d<u>49cU^#i1pTZis^26T{9`5?(P-$rTw4OndaheV4K`9VUY=a -Bx7(XY3k6Ux_ipFy-2PkY`|MOsE>n^K9=LNa^kV#p=GR3*FP9AC#c2p`&FW{Zey^*2s74a{{7#=Q6f1 -V#a1HKU)>d5g#E~riy9XAntwODdtyv4RkH*X~Wr=sG?=ph#2C9kDi)n=!Aew^M%^b8+~*;8fi*Xn_a~ -nfgJ`CL|j%oA@`VW+n$hA-A@dCtvcdS4DU8>_wSfCixt9|GRU`Z3*Hqh!kl`NOjK*5nJSLZJ2V4yAy| -0q&Tn!mgk}$&kfL*sHTnU>m(`2Hc_TQfmOS^lI~xnd*5Yc)F%XN8I>r_m^i8rxBi|4$Z*aXwz-=eLc0 -61zjn^G346qxZRjz^DnhMHcsA -461bto_jCeH+1?2H6TZLy4+8`r+{1@V;lM^d7AHbeN_M+2Tld~Rh|KG}{9?Q1$|Gwaa=9Ev5{BLg96p -Cd?l#2$hp=3V5N1~2D$DPbZI^OBAjbVJ6s%R&3h -90mRH1SAwIL_s0-KFqy}2eiDt1VD3OKaUqnBXR578-jl;kiw}dI`Xaq!}d)g63-yToybDRR`s!sQaTn -9Um!+(`R333`~gzbHcTT!Rj(v^NaI{+?Onz?H9n*%1Fu6|+v<~l%WH{mTjdYJxqxCjG2v~iCW&rk6b` -NlpYq=$u#I%e7uTXiJ4vau#*fHhYvKZtov`yP(0`dpE@R(&6F=6+Ayf5mYS4Ab-JASrd1nPMR7oKpE>jNv211BM`Fmo+(3p}e@juy&<@6nZ5|8tp8OjF8&q^TtN?c6UB<(ghwiaD_C!vEyZoC -ytTcNuEsSJ8K}Nz{GmOX@xH`0chpU>_}V6JUK2Vw#bd$tIEX -;cm)EtFhN+r@$Y?zF&sN5v8~q~+FmZb1qGuMNxG=@HE*Son`9CBP75`{b&eRcf2SwXlxCJx@A3lq!nItmh^Orz?2lMB>Hb$~7DY;{8_**AN(jwG3MR>qMF4Xguo*U~kkxUEGYeA{% -!dN>9C}Z&<5ez~E3b^an46ubklnR*>v}7Ea$%q{3w5w+riku)!s~RZ&vMN>l>7sVBTPo|d39pQr{WK! -HW$t*X70D{-5vQ6I2x8UB5}~-~#t&6}oz?6B2Vi%9|M3%|sk^^_`bj)WNT+eRG@9y0-Bpi?s0z>sM0c!x3TTZtvAnh*skD_;ojFSZKI4E0s{J~QL3yMg6`*eKE6d@p(?S}mX+MQ8>} -5U_*o9==yi{;vKooTNXM4kLRux&dX1U2vZF$b}^tV9-DglEr3}#6EreWSA&uWcG}g?V7_E+||2Ds%xV -a$Gzg{+xvX8uc`yi-NIg#!20rQ&$hK^hSF2l|CTxDC>9bJO~$9!b2NlJ{CH?$)-9H4HU(}T283Z61`! -UCTt@8llN2a&8IB3)JeqNke7mMtOmEOw9LCMje-7K!hItUU!gDNI)p%Ot_N5I;)FTM8gw_0x#E`wV`F -vZIjmY)R0n+3NPd^?F%|fUvfP14+D`gCI4SG9T&iO -ZHqBXdI)6j -|KbNK??V_PyIqM-y_%i32p;`4d_uY8JKD@Ge%^HEUGmvpl{TRSOKQJL;ZXbN+HW7e=46+D%@Slq=ja) -kcI?AVtvPGR$4n%7g8$Xb%>vw?&y95_S)jw2-|Gq0`PNfUyO{-6H$2Kdh%(&*u-vcqM0#BE9!A5(tg~&D#>6S;Hjd`sF^nb3f<*sjP>n8T#u~_Zo!Qtwm}=ynb)9D2g$JhkceNN*%D*%fWMA*27YjL ->*Z{C~fk{-C>l&^;>_VVsONYe7(9BkOg>e*E&=3aA98Wiww*~2Krml;_p|vT97if=x^yYPTxyTM+;8U -DFPh9Z4Dmh~~@a!kA-W>lm`S{6`&(K^ac}^{iP!K!4oh5@$hDM^@h@@eGYC@U-z=v|xq{(-flAYd~A! -X;}{5lvWz?5I(!B66X701W;&`&T -wp#NIrITE)a1x>zz~76I!Bi<3`1>F*L19w>A0PII_i+^mJmhcZph2fH%2#-Jd&u@+aFJk#?SwJ^!zW! -FTXwzNwm4;To$XV^*P+>7yHZ2>Z&UD7n|i~pHiPR`%nMy@gJVvS%c0fpw6b&A3pi?U#XJYCHRZbrA5* -2*k}(2)4AvF*$0ib(iNaS3JxDE+_Q^yIpsUzyk0Z7!6JKykuwfm;)d<5w(WL0D1+T|+E6N@On-qH^%k -V`zYrTXI_!`U&uW8z9Bnd+de{KE{TLHUO577QuGC%fYWT*dzO1!nPfP|{$~G(FG3!fZv49Cv%q<6nbi -dl2{)N+J0CmpW*3rdAWe2o1p~gsg`Ub=6miiyaIqM1V0`83&H}+IiK^u&qD+pg5x}oe#>mN0>)ubdj|V@c@bB=^zx0MF2aSQL8NYn?ZBUq(!mP>Qi<5uDMn6CO -&%Lw9LxqHU_zFl$`whKaZty+1ImSy^7)XIXPag*X#=qDJzRDly*hsJ?{r>3nm#^L&0bzk3$6vpE_4?@ -fvo}Y;-^d898zBmyPhhfF7HMCxrn%0=_yy%=fSuV4W6KsO6=rsf6zt6Z#C-|X!p$$-b+`0yBKQ$8ZsZ -LvIM7A%j)C@!SZJV(n<aeSTe$iVvYjsjr0fu?{wK8be -?)IR28dT8n=tl455+_gSijrSC}D(Cdd}ypFtV%h{Zp8C1JkWR-*_CgX$id5W038qgK!T08qBS!G%0|4 -7>gSvy9ul@5;78$cjf`B&Q!WXMN~Xf284rLpGN#JygW^P7B$d+b5XP4|v;;>LHH3k!)xtx@wD+pyl}s!>g(dD%T7RML%_M3BNQ -Bz)>g4tbtI%@!X=YH2^yj2_*cf@?HXS@&^m>BCnkY%!?Bt%sRBTqSv5@xS1Xi%rlwyM0<{e&H64<~Bl -&z2W-7$b$@Qg)j@xX@TxeHbsV%x9YErxnMhUOPZt_Deu{6-UJDI>fo3LG(-ANO%(VlNYaV?srU+4J}P -sa%%mDDxJ!VbA;a05SQjh{fpCtXOQq>BRrHL@tz%5vDEVO*IeUk7(GkBB%u5E*$TDQGWQvn{bOD+W*- -EGHX|mI!nl5v^g0QiX)4!-#;QQuFY%Kc+|HaUZxjUeYI9;^7)yxV<3N`D{JS2a*A6qz)V3CIUa*nv-| -a*!j*PhEC?4V5l7((7+TI_)y!~w+aOP;Xr%ImVaEsyYA431U4;`>t*nGy?aAel -4G*oQVjX66}G_jpOXm?2;1E3s3gx(o9Du(VGjAK8T=p(!Yk6m>LH(cN}^>Igp;DeiA_K&GsIw8N<)Y} -9!L+xFJgz+K!_9yAFjiP9W;3th!IYgiE`UAWCD>sXq+LaFX;f%(WTJZBE;XO#pN*CB8Uah0j`l2Aawi -A<5{29f0Dy2jYe6*^3I0y>8!EQzlYezCWhf!$c -zl3J!8eBG;jkq0Sd;zo)j-4`c1EUf2e`+4})lYjHOCNtR4Bj0^$}&ax{Pd^W6fR(4fdhDerI37>GQ%$ --V0dt!|8$x@`ss{H{7c{5pz%*MhZr-EltZkkr=n=P>xSZ6@thzCG&qF`oJYvCtQO`d$g)KA`V7Gs!6p|Ea9^$7inUTYGj80E!3a?n9y#C7hd$BuyK}w>ThrJ( -+l0dotm6nlpI80tYf56HJdEKG5gsoP-{TMSSTUz-+Xl-Ub;WCBwBlBa_9tp+H+jne-_ll1u~NyG+;m3 -5V6{MM$Ne|CttDUq#3*BPHJo$`huKDY>}eP{xEd9GAbr1Q(}${@zy+H_c^`UgMAcnUcpaD`ej=rEkZl -MWRl{aJYCH(+?gt{**(NTW8GOnnly=IH5g%Crje?@uglyk9@DgKf_>OJJKkZv;I50OX^kvZZO4%pxi>tlHbT~NrH|RvZw0jws7K<31)5Xv=a>>G!f^#f%8NulByjY_>7>2J>Q)PxUG7tmX9e -{l*Z*W{HK9yDf7I)(a9)WQTVRuE=3;05<|5DQ$d4Kkspbh$np($t!c^Zv9WA3GGpJYp(zf-J(XBzvIp -(96y!NCQqKNh!3ii<)Ff~JqHAY7(6(LCTRrW~)x5A+D*^AI#Z)d;Q%Imaw2gY3n+_2jZ%uys<|VvQ}(mBR6!y_Jn-zU4Ov%va7`?Y3?4U>m3M%sHAS&|zZsIiOwo|}Y{K(R)o!C(pb$QYz0mAM -8)V$K$g?KJj)`WuV3T;^8Cqq3*eNeI!p%j9Er9so>@ZugD*vz^+LzkEHW=u$dh?ftb8# -``*A;@0-LZab0t<#goCtpLqpZY#6gjN`>lejI=M?0>&{t%{1OemgwtCVqhV05SN{*I#Q-@i6YO8Vr(D -TxRw`)-PbS`XWu}dt8?EJ?=7N8!^-l2DXHOx%zsz(ljf{CqR=WS%6Ufy)WWJ6Y -CAIYXVmlOz!9a~e9c3SAHL!N;fLDzhqTi41%PZz>ZPuVUMYjH@ -o-lDQ!mc2H!!c54D8Sh(1&3MeyHZaFCz1CXukNb5r0*rEX*Z|dJap8{A2>&tL_}2UD-pgvgT^KSa#I( -r&5v)RP|+n@=3l%&*EbY$X4X#b-sx_jW1V7YmvtHV?_+61qJk!VS0_LPQ5C=9GL7eI5S?=xr<0ljFo? -(>4>r~fRcaXI_9?rH~%$o54!G0=UI-LI;0%n7bwp2!4)~KY(&^IRfJ|4HiA{<-0XRo -v5D#DHq5>qt)d99>)Y7~B=r`XUAw+yXp%~@+`)*Oy?CKjA=r+<3|?7c+TU!mKo(I((sEBpI~hmx&O4W -QzkT1{!W5kQtdSOZr(-(Qz=3}{7`;mq{THxeQsO_D*8=IMyTU9}_90I+OBT}GZu{A3L$W;U$Rz+gA<= -rw$CXI5jN4i8I@1Qpyh{O-(1wh|ml{BMiJ^kUrwZDMO3J@=_%BtMpRN@sNF1*EhW;BM48OL#u7BAc}} -xz~EyVqMi7iE9Xwb19aJNX=X>NeLMd9&i}Hi=HvUJ4oKQrtrhCt$A1O1o=A{dp4s`oOw8C*U{7z8BpE -;L^0=X?>#G2haZyn*xrYpC{|ra1%(K)Rp`Qu&ZF}|m>#tCTde23P)YSZid{ -kKOk6Lr%lfJ8vEh>_!g>y*d)Jzi2M?g(M2robktR{kldv -5bk`7F(wDEq<;}*O*n;xXRUv)Bn=4>7AN{Ejq_y++iJyM&+k$WM{qKz|GP%kr)(zbC_hCryKX{P>;fw --uZW0bfPYs -rH>na2;aGc3zga&jbN;13Ke!rZtF;o-cQscA8W<;g~E%g;D&jG3lkOB@h%g6A0{+!fPMDeQxvExH2cD -dq0jS;wN7=#J1P$)?QmG#sK1-NWHlT#%nSN~WQ}g_|K|kG06^YYN6z&P1M`kRemK@xa)+WI%v27PEQ+ -Aii`I>zXaDB&(vx?0LAeMd+xE!8xcxAlId50TPf!&ul>+7W94?2St*7?xukhlkM9+XT`i3+w$@S=cm1 -&3+8;J^nm}64uGzCs2ae$#Qtj!N{h(p<|E1|9(9n{)(+Lkl8a#V4W9 -||2(C|0(H$p~VYg;$I7YD?%LgaAfe85lvH#gIpjpmx?A*6}$8#>dXE)|ZkWhXiE(hQxSuB^!?eZf+Ot_@;Th7e&KV7kz_^8z7)Q1MYEormGQtZy+9?u1${msv@c5Dx> -qZ#4vY^3XGr?m4DYw4n_q<37BLEijKwgT+sT+v>jz>`ll|_rrkGPxk?4m%2 -KIwln0G>To{a&K0=v0m~Z1N+hyqB+RcFKBVu^j2W_HRCLq|>oG!Yax#W<l}WgzG_xQFJVUg2;Xx{wEE!n-^MO|w>YA0use5Dz -|ThU&oWz_+%Io4ITGo`+lfHCrd3lQS40A$P&p81+Yq3obyN*dCYC!LzY#{l{v_dZ%k{WznLfVF -W%&y+amFL5!@LrqtL?yUZYc!O83cNzpMsu8j~TJsT*W5`!6a)w)_;%3BOpP=9mdn*{D-ugCB-lu-oRP -oe9gcd1*!L_&?ORrL{z8@z`PXgP$!CJVx!jBI=z>tBBnV*MT^MM&IDNkBBUpOw^TVGGX;5i8yd$gog* -lb4TVk*c?h%9MPeSVpRIQ}>ggu}ybiV$gIYocBS^r2A0C2hMNu(yGod9oKF+r=>%>k5HVP5? -pf^t;?N)#ZrNA51(uZMZy72=)YNZOeI4uTySCeaO9V7f}Ic3~}4GQ609;Z5K&xoeZ~1q8XIlmgd@PVA -K{Q`kw0Bp=|TO7-h6bkYt_7_L#i4) -)HTPgGC~a9-$JPr6Qpom}$S21&47De76O^lFtPzq!<2|_f&_#PutJOIdxCQ%pD}LjG)u|4;|uYEsXY8 -gwTE4hU{7o -bT{3w3iHi&6?okM-tbT^W^Z6g2xVYNBIdpJ&^ej;i{X;f-k&OA3$aS)HA*=N59Y4nx_L@h)?Bz1R(LF -4wA(_yMg$b=xZf-HkB5Zrgl!o&8GyrkbIv4~no8?gSDmvyV06NDeQaRmh_uD}Uc*>K%p;tzR3Nja19i -gPHMyV?pC)bF&@^WT!h@`RE!H^dBoRTV-l4L-__{+;C6z>7LUKN)p!&=rwRij>-gjs6*gIK+P!YOW7f -HS(f;Ydas&fmCaM=41OwyjICGD`yt9G|oP;#hYC{J=V}Z8^@VhxQ>&hUYON=RlPVtIZq2bGH!Z7^$(H -hOhJrl_ntlV9EI~V{z$GvZ12WtH|{Dq30tTTQx7&3(Q&$Ed_}NWQd$x_mD;rm3|l+5*2y4SgZQeX6Ve -vq~Bgtv*KdIXg{E+iFWm8##yPs%S;CpVpV90IFbr=F=LD%8fYf$;>?J8(Xh|nh}mTM+k~}3=qia^587KHj$YUkf -HD{UY{lqkaT)&1)|fcyqNnA)8@l5@Q6kX(5Fmq1M?|^LS?j+kArR8yVn5j-}!A@LihJcWKi*q!7)BLM -0%@ANNskLrAg7&EMcYj#JJ66OJG=KYCp#uxYIu+Xn{(5D_1+qLc6^}Tafly71fHB$ZeW6!Au>v9wX?_ -LyDJfZI4@k -8-xU9EbywTmwsC~N`&S^;WCSWCV%d{6bw_O#iPLE6SV|;MCPQ;T@J(h!5ez^|QCI8yx3cBc -N2h``<9UibF)o_&^W@l^&YsH-0;qv%b%bGVn~ia6iKYqCo~4T4w4TMgQV)Pr`Z8n5UUizaR&Afol+JM -_VIzDkJ#C}cQCG1M3fEl^@K7;}0fssDNCgJ)6%E(iQGOmkiZ9eFv3mzEkIdYDAK@i!xnzhDyy-5CVSo -_WSxLxL$^98e%GqTEg!)?KOGY`F#3QKyStdXA|mRItZpOZ#-k?aAK#g>1hX{4Kyu#_ZkEmu2*rnIiyG -h0D#|mn9qaz9a}Kp+R};3s98V-H;3w(svzbTHQQf=NfZs9&Z)*#SGfJj%th9E#3T1X}w<};Dq#fT30O -)mv}9~-wtBw{k>g2cTE)jf$tqeH<#y^Z&b=NFB9h(FEWJC5{r^BHIMoK2gfg8|Kas3?DEl`{>kWqmTGF|I0kM-vBI^)yo;|sfAV -5E3&jVI3H_X+PbBuB)vPy9~^aP=r+^J}xYN;)dyGM)AVRAG7-iDu1+p8h>vmL5%J?zxX`&AB57ng`eS -v(09F&^on?~Gj>2cbr5_q-w846pAp6k~A}!5!~zedzQq#qHb%(=>?0rj5NnnX-8d|L=Ae#WrFj_9^V| -j{Wxb_F_59tz9J!cvkJ<`)8-gRTbuh8L=ir#t!9TRE84HB#5(sv(xH{4v6FV>d8nHP@Bdd5zj>We~$M -S#xt4&xEii9_5jkE2c?b>!HD5?V3I`l84^TIXqCDwAZ{O6i=$AWe32ma>O{xM@p1ApnM@yEe)a10t3O -RMbZK&#&FfTNw-a%d^4ygeWse*3s9NW2P#2#n6X#}@$nCy8F{DQ1Xg$p -%Wz)n_annTGUxCEEU5Ky_?hX)0R`X+F -E2a#m3Z-+8?qG372sqz;n(azY>L%$Df!A@9+)p1E06}HZr6S3aF|I<9<)VRsvzM2LWi5zduo=WdCqnY -ZPl-?Je_xNh!3~_(Qb_bIu#!nGq5Vx9k$OH#|gT2=vbQex>8tu{QgXou*u)^B5jd=twojmF!k}_U4v) -Wege@=kC{Eo0g;^Knp(xkPRt?%mrWCIrYmT7Tj-bybXcaJ7=`zDw2qKNB4SEg5d6#=%scg;ah8sC*AoQ`As@+ -azsnnJn4*luFE&EEFw&NIzRJyDgcrJK`id)KxjysXJm9A;CC|>0tKSztF?vg6et{B58URyp3Mr6TV)YPgIq3dlb<_yO8V3G-J! -MJ5{8q0Wym6C8*tj(FyX!B=8g+a{Nj#Fo2lS#Yo%yTe(EYJkZ%ea}R75e(tZ#F;X@oe#t)Z|FHqXReU -=a5C%%d_`C^xoa{&LhwuzrQ^HGfnm;&Zhl^w-Y;@o7AHD{hFBGEFGbh>Lp% -mjjilgE@_>cs)-(ArUU=QPRawyMbxBpPltacsJvzC90@n@r$UivC<)4>+lR(3F097j#xU77hH}y*a2y -9tRDx|H(B2Rw}>aWQn^40MYw8OZkNS6iKY91-NeMI0er(``x}o4&q#B9X@#GJAx`aKaomn_>}o>U -zU}2V40+r%QrHno892D;85d*wjQXN(j;B`dEmD4uI`|M)>?TXElDCJ=j-7&NSIq>?*9T7+v>SVwAJp# -3*|cqHuPNjpqvC2#htVQPzUpT<9PsYQPcG^4!p9`Dj571-gaXMw3GtT3KJ5xLFZ$4KzkZzX-aU^=m3~T~|SAbe;j+cvfhKjL -|N|;=0ajj%hZgsvl#ZmOXY&_l+HCKGARk@j;=uk<&nnn?VWPh3_GpY9vnbVZnTVmz}9fSo<8RWq=A45 -UkX*>1|U#n;U!btUJ{!kv(lown~yI+GnRDeqY@zQv062$qw$SEWvr9{D*FD6~GibPVHuQ9;Xm4Z}H=2 -QWgFg*wTNu!H;ber;8+kq-(noY75IudiBQqVyoImqV -?+*t-15LeKse%cUA>gM+t!zb7v*5vwh6)_>Qn -ou`Ep3)i7JD6xkH~oqE*v8K`yY5l3eWH$q6hpiTZv)&G=_KDSiop3+QC9aJP{oZq%cur{Q3g{m^xEW{ -bJ{}(ZzxLXoI9tE&SRqRA~bD!No*A`_gZ#!H}n%}8l;>kUF!ox-ytbJA6YiRB$|GmqpaWI^}NucFzvm -mh_fBxy>rk7s7zq-2o$$)XV>o2dl;}2U{>pLi=#k(VWIqJxPB-bk+4SVQXW;~|&dZ+3MypE6#6UKq5` -@x)R@LK!G2)(WSqDi-#^r#J0x4+E)_{}#BKOYR!+@o;~*!1Z5$KzA?I`@unC|pQtL(Ky0;0|J{rZ^l( -EYF?HB97wi9m4SHFGslkm~ZiLha(MWmBDo4{iKIy!QoKSz(JA7gUAN5f&nvggK#dKOo>e5R$rbBggqM -vW6DL!*R)7w{cuIIlP7ed80kifYTIQc7t)@lGau>$Zf3srH^JIx@ZFhX!YwDb4}zzn6?HXIxtRmPpaW33Cj -;AesAfojC4sZBZ0h<=+Eh!u_mn^jLuaGGVGv@^y193Mb@dIaKlDpRUBn41WsVj=QN6_<7429JV)rV)< -e2ILB_S+d-(J=7%8c~qAuZBJ;4S^=Z293gW%N5jY~?8cw>P-7$m|FAhKAmteP -ELumz6A-{*F|H|g0F+m@Q#uESV40UvD#;8!lFF+&6yjvSMG9S3i5;_WGOR)fDosZ$LKA96nZEFNO#yi -eB@{?jr9KiN7OQGr+J@uuW%u -++;VD}&AC{Ir66>t&G-M%`tD8Xd2oTI%tB{V{^?VC=iw{n2Ou;Ch -unSP^Hv@rUYRHOx^biZx2aWrN@tsFxYW*YHcw2k*%o%1RFmlY0XmdI}Jvp2|9GR7J2imNyZKe%WAF&e -r$8{M$pxFK^AaZh~0UihN;+`(R=fLkv4r^Yxzx}o9rqsgv`D#agXK8WrTp*b)^kB+KXn@2~iCbIoN=2 -U474rV5QU86iy)>y|2ND)L`&MUGD@K8=&ySSCoa$Fg#acUKa>_~7Pu=>V6c_Aq^1OZ!238J^6<-}+hy -XFLN0a6MoK~e*@H6{d1Onwi%{Hq^Mxd}E -eWF_uG|q*{8>0L|8|iTMB(~gS4O_#|#Fv@heuS;r1<{0LAd5`=z5m*5wI*>z?ww917>N;<9jXz3HShT -!3I|b2*_6CZPP_*Z(UZCSj!2m*vCt%Ad?u2s7rC{)r+jHFseL33ihGjtzh*@?&J!0gG&)XRv(&|-eVX -Bti)Y@I6-8j>;%AM>tkx)lXq>;9a=2W)iOOd1(UZx-?znlG-XR4A3 -E3Ycel)7LlmN4O?wh)U;n^QM8e+o4^v5VI3W5q`kUV+G}Db2e#~b=wa-Xman$wk+lpew!=$8ZMAoAlj -v7k!JrG(@$h!q5)!Jg -|eL+L4-K8XC$xCqzMHr{h&K%~fp!2fDw~nIur0I&IZs$Ik7|f)wR8w6ScB-a2UWu8rp$rH>*d`f<9D+ -q46&Y276>wmzuB=*72u6iL=`QF$%o+e%D;RW*aq-ZiY1WLjkRqe&(~?ZV%o>X=Z=Ldbk#y^`xwJ`vB) ->WF_JS37jmLr=DatL4erapNii(n)!hu5=P8X=S_s!Fcn_RWCjJzIXNyP}JY(41$qhb5zMDARN=t6OE} -AlyZ|);;HI^$2oOkC*83b5DN+&Y^h&RDij!sK<>L;5p2vwNzf=ZsU`9=lBM#Qe3Zt|W^ALoays%QFSE -dfcW329jf}V+p3SI+4fhP^!t?EvRAN?J^QAX!E2((k266`yvJnnd;?2syfP&c@$zBA+dnLBx_W!Mz4o -l*H15ir?1QY-O00;of09jkC)*D@e0ssI(1ONaa0001RX>c!Jc4cm4Z*nhiY+-a}Z*py9X>xNfOi4pUP -E$o)QcG{!FbuxyR}eY1fY)_B?xxsEBg`+zPO}^xTaGn2Rv;b^_yJ&`$4P;ozDn=q`fZX$`i -Ac4*u|M-VJD(<=}Z0d^l6GnwfC4YVlEa54-?<82=XLF2t`O@n;U+4g+YQ)mL|^=8x_?gH9dR5! -%peirF;Gcq#)y}dob#e~70132x2Gc9pMV5W9FcjV#ic{E3JQy}=f5zmnH>@flPjNoWH)6-Yau}xa8}hOD1xq2ab7Po2wPkDrqm&-$~QP*TsB&f>~u#b3G&) -S>a-j-$O$wAN0Yp+x>WV|31DZ-47K%k|)s4QTahfevu`xa#f656w8jUf&jmy13rvpE!=>eZa$;W`vV< -v)=zPA5ecSpek1NmMa0Wf>ISPG7$+Zc!l*Mf~~CA#ij3H*;wpuq;*`rQf`n)MSo&>@qRo>JgL{~c^^5BFvp02385fF9MKpjIJp5ZO(PsYj!E*Ir6|A?9Dop0*7v -**2!U5q&QmrV3~n%=X4n0I#Dg?~T*!$fn|1P9i-$O3045>G7`#D3R8JV;O(=hHe*FcK1-#+ihT9=#fg9P?&Ro?_h-K=(S4X=-F~6WIu>S)-yA81TSS@AqnUhP5qy6U1 -JBY(+6!^uxCd!}HV_Jtuy -NNC@jgTCMl$sN;DKdDn`_$6Xl#61+XX$T<|U6)c3!+?|u{)EB#2U|Qt%Sg!*{9OzADE}KWJTF8O;4{R -1-NzT5{2+pi?d^l-uchOX^~Pt%m*?+~{p0b~xVN8=br3=hdwGPmFkYUYUuj -0+CXQm|m8R?u?@mrn1+NHugL-j!^7j2-z5S|}{81$0KB-8lw@K$<5@ZrqhuKxg0`h=#-}71NUlN7#Re}9fBW@iF3AigFlveCI>X%X1kZ-xD5Df5JemO*aD0n^ -=JdvCsw7wfWv4eV_;v?r$bhg`tR47TF5EM4Cx#%`YB?JU92rB@mvJmL@8$Lhh6Ki9Be7~_|=E^4WpE0Pi@GRe&lToV$BmR!<76&Y{eTpb?1I4F*Vn^D5#`j5V2#@VtRbI1 -!DiiwcvL>`AJq%f|=Ei=|vbp~G9A)MlVZMEnt&evc2HEG&QT3&I!oWz2+CbM*7E>UNh#qtzqO2WWV*!Z66ebJtWu{&XvfoM1EvLdVK7A --MJY*flD8|UUD3=z^nZqE>GOL6RQ1~cJY|?}WA;=Y8L8t6*;511N}djtJC90r=1}>_sI&oS=1RsP-z+ -NKy{5a@Tnx)RPPY-3Yu{>?2sB|CkUL8RM~+GSP`V#h{&}sUlsVQO-vGGF-kbgQP~Jv<6*UC{D -EhF -uE3bSh0ZEH%`)pkx;O$83&av(*&9&9LMgqZntw)2X5;Zp$A3+Rt2?(5z~5 -YI$(BTpT7 -X+c3sjwP>pIGlSMI(yV2Y^PJ#DEnFRxvL45!$z4Na=hCDshubbZK1d~89gS6KT8_zP(6 -$xB&Dlh{S0!B25M7$Kgdw^ZZgn$EWvzv93#}+i;Z`-mQq@uhx6lGr2)C+!Lrt>?ZlUs`3~p83m6Em+x -Wy5mir`kBe?w7Q0i4TEyPcmp@O(v=**vivWg@Vx{LLZLA+t-!+?MV92djNk!1R3>MZPbB`ExO*aq*40 -UeT#%4yuhN{HV#sM=_hE;)|D>Zom21LY3CpB-(H#1($dLz(_~Idqmz17r$K8lmM>7WP#etV+a~P|jV>*7+sIBnjHJur|bDx5t1BOG|IthHZ}pp<41rU};oI&JsfYy(2qIomJ_f?b7;W- -snBe(>#fDekY9s?NT^^N8{3(bIw>F(1H;Cn8u)m0AoFd1kyQ7G8uI7yv7^97JKa!hP5AL>M>YlLVRLu -&)7S{R;0MsArf5(#&-(6%WkI&G*W*hI$+6H*o~H)$DdRl-GEY -+6>qwHB4uH3l*SUaV>DfXhS4QsFS?8-cm2%$1eWe+___NDy;QClV@mg6X~tCz%^UcDPH*P*oZpWQ(oR -q1A_94h^jB(fagDPhVW&3di)*Fmd4YknHU%x>SyA-javUo?{-ShB{ymq@iWP#u`aN6y5s}%}9L!K8VDiOr<5S_(bD1;f~3M#FmGv|6uD{V9hlNu> -|Gq0an<9IXjX>VHfQlAgM$eU#mV41SrD>NQ)v6C4j$}BSl -Go6gR=qQ}qyWt{gO(xLVNFuiSn8O@jk>KZcOwD`Mi}hBPA5-54LsGK?OEjqjcBO9>salMD& -O&-sZ)N_Ro11E_2kcX>+621PYcCcgtRT45mWP9~yzCl8n?!4MkP1|@Fx^A8SfC!;LarWNojaSFt7X?h -+G_%&nd*?DMz$8I@0Kni{>`}pRKCI1BB+Rl2+j|PL(K;|fXe| -Q5fshI1g4z8I73wR@&!@z8nM7n*UY!nix{gZda)GEOa#m{!`9A0$)H@HOGiIBUpQqQ{RHfVMcZA3cEH -I$@Bp1(yzcj(EyAN^>uW6clNn)3HZ3%nd<`afASDb#&|ET0A{~Es&ppiamu~8On-*qe8+MHsw=Ci~Xz -qoT-`ESQzs`%ytVUkM)Yk{#rcPwy8~Si=)$3izZEEDf4GjoyYJV!cDFUP86L$o=PvCk7tVxZ`pCyv|4 -Xrn_8ybJ2b?ePeO=(7cBCi#Xn^HfO-Zbo~_EOv1H460yeUUIlXKTvW{mNj~9=bZHS^S{Y!OZt>R|i;h -yXs(0KaGH?!DGz{s5E{3*`g!}7x0()wZ1dMcar`|Ui7;yZS=ikp0yW!MW>vRK@pJA^Syp+(Sb`01V~l -kM;IE}dpJUf1P8jJ)n2l6#*kt62d=gE&^u&ZEa9IS5wZOWCI;o4c}lI;me409mlE6ZF(Y^sExpE=qq|d) -x1RHH>15L{a;!YUAWCCa;G5B>6t%krkM9kIiqwOBAL(VfbpF$3HxE)P?Euf}Y0a!EwL=@*e6jB(#EZ-b9K}DRKZ4c6B3vy=im7J&5pfvF<77Qc@AQC~I;o -@4F^aeW>P+*WHq6o4yBr%xQ)VH2V3pn7hu8X#;Bcgn#T)Py3B>e(B(y1lZu@mjIJa+W=S5qWE1wVJ)j -sss=%PXe>m|!di4rW$YxvL48F5}Fw@~5V_b*nO!K$|isVOK*E$kzL~_xYZ}r6BmaKo#o#U>oFgk_SGT -#5mXxFQjlfybyFfrcqM7OcmdUnxoc&F=NUlcQIJve#E=saQA3$_xbMUzaBsR`+ta$JvO8R&CUt;@0pb -o&hne@$`e9%2G~KbpfqO3APQGBQ`~K$H6oQfy;)M+Fd-6lLj`BrmF22%oE#T2CCy5&AU)w!QtJ3l(*Q -?#YAr4093H6^3R;e)WHWYKE2SfNF>lysZ1F5bT#KLAt8_N*dG)<3KAq^#MDC -JPO!l_at4vLBTqMQd@3TrxU7*`6GQB`b4C(VVQZqfl+ejcONgB{kAsP)h>@6aWAK2ms3fSzBUzjPzy)002QG0018V003}la4%n -WWo~3|axZXfVRUA1a&2U3a&s?XaA_`Zd97M&Z`(E${qA2uq+jg8Rg$gf!)S{oX}1O|x?l}Z6h$C25*@ -RVM2)2C8rYBDy?jWNpYkI}VNp8dJv_X;Jdt~Ha&k!(BNsQ<(aFh4uQ!n*CviNfbR}6F6P}ksYC?-bXs -Wp=RIlgm31!jrKJl=N%c~)@m6kNoInz^-T1-jxG~@RkcLPoPK+j8FJox*!UvJ*WS0CPA{dN86PY5(=- -x>7?{Fr`;$TF4~LQz2AS~E$tkPWaM;3G#s*6a0>j4DNLGO8y+=I^*vK)w(E1|veAkV5=RN95+?KWDGb -o|(}=o(nfKa*vwX5ui6I-6$mG5=aTjnX+K$4#?BDX2zp>EL=B#eI*&yj8Iar+43sSqh5_1l~mHaS*_2Ek+ -ExyqZKf}U><1O%g}m}4A8ZfVf8v0m}@P=f^{*#y1fjoSHb{YYZy@Fx@ibfmnz*j;n*vM8?<3EDF;6m>+2IFp0&1C`1+g+`MUKFr~NS$U -LXX-Fv8OvFrF&=PYMT%EA2YedDH8lCdRFfIC@CwSF5e6#r9G;r{Hiv+;3!+eFcq)y~(CZ!)$4SAZtO~ --5lwnk$kj*1M6%V`{lQ03x}omC!F!IHj0uCdWJ+*{ -4*~X1fy*@C&dPGZkjTl!69~7MKTciOB;)NriCgJ3s}|Ga(;=o32U&@^MNjCOl(mXlIUni3vR8S%wo=^ -CDp>36?n%54<3K#aMl746|V5^;GLpjZRNBOQsMq%6TG%5)&OkW}dR*X{Ao#AeL0lPm|=u^XJcAzUqJZ -9G#yHBC5)tZ^!Xqld5&W0A1_rt?6e<1%oq*%%(h^8V57;S6GZL$Rwi=m>`Zdu9-_0hl81<^suB|}NtiPG6G~GmKw&U$#Kc=8TOmT*LSQr(cu -}jy%-E`dA2C+67%0VV^Y@K=A-|wQGRDH3u<+R&uqGK$3|#9z0>7P)?gsU9;2xe0O~EEV4m${#nmiXOGL`YW;DzOTb_Y|mBMbsZcl+V|{AD`wd~`lO*EULCTO7brv6rI -dz^cz+h)eg&_1PGVxSkdOrrb|#_qd>dSk7Z=@N@a(;NiE%vOtWtHo5sI6%xNaa|J}xteUl5CbkVY@A7 -p-4i??nxX^s(b8%*yT9M;@UVZgaoz;J4W_6$>e_c}t5`XJRb5;QqfOn-&Hi^N*d>#CZd(g~?s$E$jK^ -5TaJ`He%R2#f)yPH~k%gm1b5yG2(V6J_t{Y6>|3HFs`LvK@n)sVAbC@Q04!E%?aOKa4>mf# -L+KD0Peu$Ro#2MN*;9R$cd?#~+=?tjGjSyYq1PBl7?&k -1v(F((v#jv-rVNCc1L4;3G50E5(zJ1Pu1r`M}hp7G>9Lu;?RmfWHGZ&%0Izhdy#?Jd{JX^4RtQwKH1b -gqF@@<7`&V)b|KoiZM|w)3MU7w9UcpbXH7yEt{4JS<5DL3tjoR(|S*V8e_0oq60{+aSxlN>cA`T6rE| -AQ+;2lynWcB)V_>%PN(gJ0^iYp2%&5Y2AbmfCyWKTBE@NDMNIFrr=p!O#Fm!;4SZR6)1OLu -$t^Tao2fZ-)0LRBH+#}*oYZDg*6(!XWF60*lxl0WCBK{l>YQE91MQnzjp -5qN%R7>;yu2sblUHDXHkbE(dTR(*9PF5L*Rgi@<`&PRc^gd*s89Nx?bEG(dHXu%z=y-UH&%C -LFBcpL(Ix_HZLGo@Wy?yUqYJ^Tg1(OQ!sBz3rlTjKMmS1UhOQJm&2>l`546OKP|fOZ8p(vY#V=iu+Fu -2?WAwBk#-{-R@_4=ht;?X>DF%AM*DwIO9KQH000080LuVbTd|cKBGCi@0J0bW03iSX0B~t=FJE?LZe( -wAFK}#ObY^dIZDeV3b1!XSV{daVaCzNWU2oeq6n)pPAhZCp0asdHjV@T1v`w(4HIO`P2m*?cu57N7sF -Ac2W7vNCE=ftYoW!P9EZgcATjIGN=aBbaOJNw6Ed6v9g<%*3xfCU#G+*ka1WE}nYaumZRV6gjTvRFun -tMV;HPegsQmvJ<*6W&A^VahEm)To-dj9tGr;Cs88~+ijdAmL6RzNQ -ovDc)1ZlyPBA154Xvf>w85?HiyImz2myl{9IVgfRh+Y7(A`nr7M4p&G^3hO6fwg=mibC8g3!E=`#;O! --PIfhUQ`ZHMOr*;aDyI{9z0kT=(>Tsx1%>g8v=J7iqKuJ+WO5xdnJ)MW;+|>zju&vrDmDjmFFq9*t=3 -Cc*p(}r*i4GKWMzq|zQ=?hwgUWowqf@8PPEWvagxx;W_`&%BmJcUB63SC!Tc�(o6GLZME|Y*w_RYt -$f5l8F;w1yWhk2Q3?9X9<-uCk-CB(#pfWhd)W0{jq1){P>}kMrysbR~h7nprjeojH1txf+8Q2S0_Yi= -?+I@K9zt+zM2vP?GDJAXbu_C!U|+uO$bEuh$P9M@rPy;&F=Hm?fTWSHfJ+FY7~l(8p27^oPP^SC+oJ4BDB#LymwgRb -6p`ijrT_i#-a&iyfN~5_cMn-09y+HFjUZF=UR#P1|W^yh{xRY?dsW@{LsRf?&iQTs@)vW;cyeP;W`mZ -u@4WRQ=OUU^{N8@PGEKs))TfQCyl5$ySZnn?kTzA-$gm6WmVs`~(6YRWMqXe(I6enAOOzIL;jjb7sB_ -r0=D;pR0()Kc_8MZYiSU3$Gw=F0BET=Z~@?@1VoTF15B;n48J-W$0zI*Xn3&}svU6lS7FJ -vZZQ_sjRd{j9InMY{Cy{jKb%V8_~}qLw~s?bF2d+=xbQ%kZ^X=9Bk7LNbZck3 -Uek0EJPOgTf>)z1qw#K6>Q+#~!@a|xH_c<#3;(FExQ&n5J_Vju-*B)xWs4QK8Z>k16Y(k$J6ZJwv*##RqhdmT{JcxItI5>%)7vldid{<1ymiAXafcG$+dT -i*|J$E~C{^{A#9h)QEKkV?%{^AgXONelNQGHcKSc(f)1X{>k# -!k2^t`%$hDiT|Ne&+IMMGkF6YlDokVtp!k@wWxh7!J$3tz=4NUfwLG$K`txJOW)zqp#Rp -h#Wk3h(s<(n2j9y+YV5=w`sDGY8a;aG=ug-!_xp$J7BC0`N?B1*YW^X*H@x<3b~p)cgTDY!O9KQH000 -080LuVbTL1t600IC20000003iSX0B~t=FJE?LZe(wAFK}#ObY^dIZDeV3b1!gtE_8WtWn=>YP)h>@6a -WAK2ms3fSzAWTQblvP+<5 -&y2g0>gMF(u_?yy?$ujOQw$PB%X^ck0oE*@p&j%1SQlI!2*yUle_!dZx;Zc)`N1Crc-^vUo%W+V*bxt>cFhJ?>DAvK|CDm2wxq^i?#zX_FpQcubtNk*k&GG{$9zg?K$ELBG -JG~u^7!3BJ^f9R)-r<45m;lD0U!lO?oM?amOf7~VK;*x<vUM6SVeE4a#iixola+*P^FBxZo>k-143SrRQ!h?k -c-ck`+wd4vyp@@#`nI}BKXYH`Qt8mU?c{!hzFvljEw$y{E@t;pe>hVkg+lOajJEu4uU|zHy-n`jRZfI -j73b{X_5ItPJ&y@w1;H1FMDQH(2*(aPxUN$Hwt}g%^|FaB!r>@iimqAC66!;fEzp|Bxi~|a6Kg~z2{P --GnQ&{PbH_f2~)li!5-->gKq?k{*V$XnZW|SqVoi)lQ~nnW<{EabfS1s71$$JTD4ra#`A9Eyl!d@e)xMPMaHClDpCuKD#M6s8iLHOVo7YWOlesrSRTjtI*>N|2=Wv$^F%a{8eyhX6JM-byRUyP6v0z{iXDK -R3f#W9;CQyG+hhD0<`qD}#Jp|F)&oq@buu|KF$2-z!gvgyPnS(#HVrCDg)R|I#od@Deg?$qG^&2G^#7 -)uj8?chFM5G;+ -Jl1vrxhOC3|SZEGN2DbrgWaY|FYal{#>&CVpq(RPy^3=fCfi@H+UH#wCcux}oF##M<}v%?hr0m_x$Q8 -*l+ij3+ZFv?Bh4#%dbq>~7%Ig;mXL{&qrFn%I;{`~;qY6`$GlFEwUsRFUD-h9=2VI -F`(`EY={9fop6(wZcX -#6y}08np-H=-UEvY6vZ3`zSZ-?f>QEvw3;`C^Eb$L>8PA+E+mU|vC(H4m -%ZinM%L%SJ@pHGiZ`V|aT#u@~(x#HuN01RzA{m+neGxg7|2B$}*_C~_G)Ub2jwuxb>Tc(92-Y_Xn20s -n^7nPi_Qr5tbcF?}*M()_!;0b^>z+Zx$E|19Gyw`R+zq~pho}Qf`q(v-lbZTB7ps)WR8GAoC4*IZH;` -=$%@CGftCJ*M~!I{uCPVLmMk~ZP&MUZYuhGiK^)n)qodxmHGnNS+fC(h0v~u0^Z ->HBI883~(caU2;2zOyx;ZEEhv_MsDf&4hxGp5kvE-#F(1~(-j*jAbRHJf;Cl|(Fm`gD0pi!PpHI}cEX -}PXh1DZ4E5l3CDV$GrPy<0lnH8;t|;=tTDG`)Ew -%XnOQ|U;jn_WYy*m9C6Wa{Gk*rMXi`wv~))e5~h8=X()7U3{97#F7Fv!hx3cSx|kmC0&wVo}8OXVC6B -6l16|W-ZwxEvxMq<0x!pUZN@sSk(ADF$0DaX^eKIQP$`d1j(I!}hD{8}`b0pj^H|Vjc!rGLWq!Sa3J -{68)Puk}K8D*Ic7?d1uz?nCp=RT|QLe5K0{m(V)4;tr2&|Q1c|R5orFYcM=6R6unX=E)>QB_zJ1EisS -IR;DGQt7ip!4R)sZo1}DYr2YKyp?h6Hby(~ARHV0-4a6X@zbt(a0$a`kWD(pM1z$fok6$2)5hozX?$293NhBwNjLVc!NNDcq+jg9=1*+iD|&+s!0RRy&1D@5x%Tjcd{fuc1ro+ -Xr(zak{2JRe$h&dj{@Qz8M*J)C3F0@w3oz()_&vmfnAyUwng!YzPBMTwW?gT`jhh#aj-d@&8TwY^P8d -SYgdu2qZ3D_YV7J@3>HGmuO9KQH000080LuVbTMO*&ls*9f0KftO03!eZ0B~t=FJE?LZe(wAFK}#ObY -^dIZDeV3b1!yfa&u{KZZ2?nby7`FgFqC#zhCjv9;(SodT1{sJyleq{fJT%(~u2?ZCzbM>=FrE?=z+uZJ<&VCs>#_T{Y0tpTi8Mxjj7Sj8U?_4 -q7>&i!8IT~Jt|$#dzr%$mg-A|M8c-j)HMJ_OUpaDGow?h;i9kXvjKe!07 -~CP#PJOT9yx?~QdB9sT|zSDGS0$-V~9FX3X&qt)^$Ae(Y;T?*DxJBEd`F+YFMDNGa;0I7k0|XQR000O8%K%wht(cJU%?SVijVAyABLDyZaA|NaUv_0~WN&gWa -BN|8W^ZzBWNC79FL!BfWN&wKE^v9h8f{bJNcua!qK~;PVGAQ7x~sJ_bq?T;H7cCwOijsDXhQ=zLpqyo -RBmqm`#n#02NFmijLuqC5z^25dv{YT7C*cGfIkVVSS%I_!;nviZ4ak%8Zz4^?j+!$B-Ho0q|)WSC=|> -)E}|zf6B=XDkXz}Jtr)4)jv>G4Jul_WwEgHDdddH)3$iqp~4aMpE1~4 -8baZU;Z{uG|WJ&!p`K#Sfx5s3!%JbLf3LZRSzR0v{}EiKV5Yf@+e{On(Ln^%PUoOgX=brtpa@UI}`0So0!eH*YLx+1jE-8RrH3FZyUWcM9}La3u -99x?cL5;Djd|3rrf;6Om~a7Lud1{g!=Q>0lT%o-G -LQcaeTZSNHp*1UX5JEz89CdjrL54W3OiqL-?n1L-BnTT*Ov&zl$aX{XaXy*VusNK7N7t&}GYi>4(&HRBK&`^pJVwVeZ=Oq^rzM^=Zp#Li5* -9tN5WIs}G%G7J}KXS!=24dmIsod{UI`#OIYscuZT0A5D+wS-~WA%sso4z*fF3-%xouQR5E*>wI-gH;tpSCl?&k_{>E~VRVBojhEON&XR1aBT_hlr8z;S9C*& -j0N2hRl=El=BXKtAUZL}{h+gI}p-8R`W>Lw)a&>iJ8g$7<*G6gn5Do8(3l@n^m#Z5r1_1Mb~WGiNdx{jdDUdeREi?Tk&bbo&t*w$n!kcBkGHiyX+?B8A+6p+XB76yR)5fQNuVtb9J -pBkR}{R=)4kbRJHu1xW2iY5_6)LT=#bau}yBFAEO%f!E*&^EOW+?h_8FGb$o8NsisYbKSJYEm;o?WYA -u&SdGduDk84(+fQDrowe{Dx+Yx*X0t?UZ%?BZTGyfF_D;fkUYqnMjCqK9y -}$BH0FzPsMY;=)9y6SuUn0q_Vt%ds)bWux|6NPI)nC_=L|P&5@hW!wwlQj-6hw62JHvAgRL_A&kycK) -QFbL#%oNwGNvW_d5rw}sUrqhC=^hc2_)ZBGmmtORSpo?NhqMp?!QvEiik9$YTCC=Yn?&52^R-0y(s=;$5WXy|31y5*7>JmTG< -9r?8Att89p_OFmRhhUZ|8@V^O0>OUK7rYw5u+FlYpp1h_cYG%sVXc_95CTki*$VQM2u-n?yAk^n_m%#ZTqF`=8ZpKhn;2*LPwsrT^ -c_5NGW8eL23T;l3dnTks&UYp%7x5kiEe3k{qXapX#qa`(%4kYM>Px2Xv=orTL+q!J!QzSO?W#(8&(mp -d|EdyS#BOHrQ(c-x>gVYrdaB9OM;d%G;XcX4SZyOXC}qPhY7E3a%n7(jUWJO~J}G`X_^sY-)sCyX`?Z -6!-TL9t;qLMNQGNH|xKTf@9k%L6zrFvyN+h1K5K=QT&$Sd#lT8Xv8K7HmiCHv-b%&Jd)v~&v(mMlHa6ndyBa` -8{GjESO#Px+_bP!Q|g^T<`JR-DS84;n46y2q+DP^^z%{A=caAVb~{l2bHt>Pz2%fa!Qda_?%Q%Cj>)> -giHb#-}p=L|%SHWZ<_F@IY5QOoz!ArLbWYtB8VzKb=)hP5H=ZTG9Sx|R&X1yWM&wWx$7(s&hBvS20N{ -@+)t$M8jWQv;91iav68fDIWW3OMXz-AAU=W~H{pcv>np4{BET17w@72_2Y8K9|romAJtwM8B$5P&uRK -8-i&v3h6-q@MVngtTy_(#1Dxp$ruLlu{#>?-oqCMHpjw>976qBiHC!dQzq|7?eI_)V+YT%k;h3qC$4M -NYJF~cSgz$au;FHO1{3%Ul?Ii&C^2e$1@e9_0j8yHCX|x33Q;tYhvM3`xcoX(dyMDincNVVec;?a;LLRoU5d$wFWDiZNV?eWeA!1@7qcf`Yhc4vbXA&Bc*ibL5A5jIM2q0e5ZS&Y3(p;e -J%VSP)h>@6aWAK2ms3fSz8MLAIrf4001)u000~S003}la4%nWWo~3|axZXsXKiI}baO9ENkc_WQ$?*( -&5qhI5Wf2?0uXip|F#8QY;+ktWniCHW`j1t3Y-@ap))v|4+UV6&JmyBF{W1A`gQruFt}pt<;i8%B5-jR9KgTu23bdo+_Np~oXclrc=0;cB`mA<}eq^n=}9ymd -V^eFvA(fbC&)m!UQBWM}(kc!9kePGbVk)*rwP@#`+0oLp!3wrv7H0?lCHY=)Ed-gIyt+|_ni_(S8P+P -ym--LvhVpmlv`!_XLp$jSJxgnWeeibinzS1GOQ44f{bM+3e=i2<+Kxoc?MA|e32^R_h!cI^m9#0n7iW -PSIW5+rPojXjxR`Zop0beF)Z0<_j$P}Tnz0l%yvO6S_k$@IQ?UCw9B-C=KlQ{&BGo8$PFaa>FQy_b)U -bj}qhvCtbT8DL#WvE~`eV7>*NGe||Xm3*0N$VHJcd^5Dn@U+yD&ufj3h3{E3WaH7D*)ia}2}+R!pWj{?a)wxZXy-1=nTZS_nxfR8(r7)&- -SN)v^+bL9mr^l@^q*ST;p=WCXTmr3RJLqImC;h)v04kbZ9w<_ro>=LHKy(PYLYOLcJbR;3sm$}1A0Dw -c9AFbjrWsN5zuOvS#}cnO=3(G^{ywTFKsf-6aD$yPx!#-QrC(p=Y?!BU7U=B$`p^OUJ?PzV*nsufE}M -m3GZ2!Tmq8|!(kc+8BKnn_t#nu~IR>)l`~sFC7m7Q+=~u!?Vl3AqiRA%mEG0-K!SxeR&6q-cmhVPdJi -zlMAmpeAn~p=8T~FIkzg&=mn?!xfw0s<;Xlc%;SUBWE2ghZUhJ63e>>o5V!|UjWV4JjlLX#?>o+dnPf -mG{40?B|iaBO9KQH000080LuVbTaA*H9PM -tBUtcb8dA(NKZrer>eb-k^od-z;O{WH0paT4mxK3@rxgb`H0!1ON$f2|qxn%ZYS$*iY_v|i7UE-uKMF -j|;w7X}{%$zyPv$L}i;qzg)s##eRe*E2iMn_8|4GA(kWyKS%ST9mK8~bLw4P$G -$P#Dcy`5#wah=mU`=JCmE2d`-gmL#_szv0R&kwsM0o0BY!|1GF*Wx=F{LY82YH+9`8s_@$(acouv)k1G -v>q*rjR?E1yZM5a0gE1}V5(x@lCC|%mYK07>3bsO4NJFYzvoOA}pPZO8{RnDd(AcHEL5zC5wH$^Bbwx3s0F-p=@;c!^0}^Lg#Dw$QKbJ{Y$nlc(uCfq0bRB2At(hyVE|=8C~~}bouh-59gOJFTeK^f$I0 -eYFX3?YR>fPm-+4cj}K{t!AP(&JfqhSKgIR?yX%{G4>v(Yoe0y>8GNDCO--^vql3yGtcE7<-Y4{WXW8 -7nAB`4^N7e>HUM%Q}#+T`LKc<&s{9QD8ltf&BoUC@ab(#yyPj>8&qPSu8_i@EfP23nfm&VJ{>Yg43`ZOeES6w1Ba&% -?7nf>Dp->#vuW+w-|-V$iY8KIYvrXY!~swWSXT%{H)#$vgl)dcF6OOjQA<%*3x{}@{QJw-k*Izx9qeu -7A^|UfV+7o?ER?5g4l36Y!4l5Wt*Z;s%UiD>n;t^`_6vk6+eaA3@2rY+34#u3A>Rz2uwn5{wvL%M=exr7fD=SwFmTysYaeG{Ny6%jjw%%0WGbg#!DI!+DGSr*_X(B_-rzr8;D`? -JKEjxs_)fh*~~F1RB-;*NmZOCNXnC+9YB=^jIP5oMbB``H83jLW&0CV>}b(^Y;Wyl%O_!=Wx|lX#VGe -FtT{br3uXm^)`S!4|%vcSy;%s^fnm4i;6n04KmOpaObw(5NwPjHW;U*g|MlNV+p1$OxX*`_g231U-PH -Co(vQIXxj02J>Luk0;9O_yC(LbCSdtBOOKgABK7_v2L&jO?_dm4+{P4Y#*!-Q2#dGT@9X{@B@PX-INc -P#o-0fHOeG+`3z#>DrD^9rvL)$u!S=~V7t*;4;Di`FdR4JKHB=BpMbEXMQ$j5~t&XO2R4GVA -Jeda*r6d^-9EP)h>@6aWAK2ms3fSzB(M#1QHL003bD001BW003}la4%nWWo~3|axZXsXKiI}baO9XUu -|J&ZeL$6aCu!%yKciU4BYhWzk_y1G>rpA|)8Y;dycsesLY3wf43Eb}bCheH7B -5k4+DAgt9Lctm*o{(64@Ect9R$Og5?0@mfzynoqkwr}NjdZ9(C<={C>Vk#z;HIBds&yLQ-r=;cLgt|4 -3U*k5a9i&FMkjLzd#bCvBFKRiSB`rpLF;2DXPHD>U+@K}cV>p=FwL4Ad+>fOj*(zm!P)h>@6aWAK2ms -3fSzE_rd?u9w006cI0018V003}la4%nWWo~3|axZXsXKiI}baO9bZ*FsMY-KKRd8Jg_Zqq;zefL+4R; -uhYX|U5aQeE`{N)eS1APP?*N;dW+Sz_--_L{~;@a^oycI*&TB=*H~IcH{eX4Zrd2h`b?s}#A+hmqj9N -YEI69URA4--s-w5+ad^3}dJWAPTuH1M0ok- -^V<)f9)>nd!Xo53hVbg)we`QdzrB0^a0g3~L5z~JBp*6N-NL~X(RbHjXtW^%P`8csXO$sxEtHrf;!#VBq(Cw-{catS(v=zw)rhF1+>&WYrcR=V*- -2u?+oZB;mC!0pTC28_SJY!A)ZSJl)=<~;xMouIz;Sk}jC0>HJE0md08=6`4o__BwR%!{Dy+`DmGL%Kep}V#f+yL1V -dPbOSupFqzw3ElE4P>h7)gbjb8kF*!EzK+UsFYbJBB_2D)n8yf*S`529+_MNwe_d~@eB3IMT(#!utgy -;4@0LJvBlZLb*RLe8o;ShPmQb_bl|ofni;uyONXGoQTsNAeou_AT9EoG>heV*^(Ie^&X%U=Nf@Xh-B| -f72#wGB9A@WO4lTEr@aQH`A~%k!n5EXlAJ!9M0MmpbIhL0(X{pXM926VQW3juvrK=bxn ->pJhH3WKb%);sYF;TlDO?O-6lx<0T1TQ@B2SygO05nx+}BsU9+?zfM6_?5MX -=2BTAKa$@^vh(?T2J~wV+X`!=lR}gRmAm^p{mtCv51F?mg4uBcqH0QXZuE_iW5L_ffQLxDeVWPUpu~1 -(Qn>7naDI07`Skqk{1a>r&i7Y{4+eP^gGPjlhx>0of3Fh{Q;H00jVXLxU-|d%x0ly<>uV^aqR@tmQYW -)H)iaw>CcvAfJTRY&jDaOo{Hs0_zb|9RbE18jRtImn~WryvNjTV0LVsxR}v -$QeB|KnngvzSj{VBL|Jqbio5X*u#DlkkFj8^a#P0GB84rrwlPFpJb75gWBXiY&e+Cq=h17&qkZ<*FH2 -G;E4kL-;~z^hnPXVD8*l8!!XJsMHgxObScS*XoV}PmSh55Bo(hApeGSMdTe~#MrVnfh8Gl34`BDbq-%EhR9fA8n4>+5cV(9unOc -}U>?KSBpG{hn!!thf;wFKH)vbPYO4L`PapBb|!i)&p8I0)?9^qyjGK2k`v*`<0|XQR000O8%K%whWi@i;;|2f#CJ_JtApigXaA|NaUv_0~WN&gWaCv -8KWo~qHFJ^CYZDDkDWpXZXd6idNZ`(E$e)q2++y^;@EfgI#48}4b%^I{o*A7dHVJHHNmMDu0MQS9K)O -+Y}-#LdAB_;0A`N6O$p4)f6bJ23STrl`_$3rW$cX@4W%e@ztj=wzlEa?0dcY>v>tW?~&JUrT#raM}BX -#AF44nNt;7cYL;ynONUd$vC^-w5_wG|0T*?V33u*u}@2Hy?h#&Vw2yOQD_EvRBu?gy-+xUcP>R{Tl7E -N_Z|6*jz#Y3)yzYdRCfBly0%8t!Y^&yL@*qbY*OwqU*g>EZ#l{zuOz5cBb?HA>#Z|2{#p@limKf)4Ua -n#cs!y+U?jm`;=V`kYu-u#i9~5+bLu2`Ys5q==*F-5z3lLC$;ljm+=Y2>1}+BA1i!s^;4r!&th_iPQW -Imf+atKu4T;frP24o0_dE1b0<*B0{rj?((u{G3?^a6Tw{7wv4-CZ21i!zdk*a7R=CdS3I(?0=(hNhfL7wOu=zXuE`p9^Fglj!jQ8H8^X-;V3L;utZ8rB@^ -s(1-uKOPnwNwI)d{yR8)3cIm)*)4$*J{Vg85{EuAeZT>7Zu`;$%Sgk8Qe!QFY*~uV -<5j4!J$JGk)JBU%LCG$P#1p_h5CCoT4tRtf?ICBFVT2n>hO=wB;`%*dpQKm7Q{F46*Jb>*Jj>a8<55+ -G1uODHjt-3mg>9wgDp4hmp4GilQ^$l-j-no$1-NK>-}N3mTH^gBt4v84t(nY4o}@Wo)pkmFflRA6fE- -l42A>82q?~l{S2&${;^`=@!ST8Vf>OtDrWI3|$Ea(roH!dI@Y2t)eOsrnpk5hd4&cRz=py{$ekc6Vq9i>me?Q1!UrA1$^8-Wj -sK;d`Db{`;cD=TPLQ)sd25~Q)C(Dc9`&$jFea!jBSU1pqJBQ8zfa=n3mS@FFJ4}%LNb52e -rWfx3Ih|Sc<4GYBXwfwsm3_Q|a))E;=)QN=}GsnWT7_`g_hh`Kf#p_U#y`xrI_qyB`IgW{RJB^C7B6t -~FAC9Tp6K>WTZ{kO`^tz`H@qWD&qdXie3ZB5?s{C+BsO5Q%TXdf85D*d_9TF3zVJjt$jwzoN7Q~v$+mzv!G(SnXQtm89oK -+D{9ws!O)A(*~}arqPgtR?bJda2@K~TLmg&lbx9h)o+|)4ud+iA*rJ`X;T9ItY?6B-7L`TAyAD`~6-2 -cg_a6C=V4Dw#iRYwjCjQvvtoQZi$L!lenl#6Z;+fr^>ll`oq+(?GxPihy7r73MJkPSP5S`{yD|}}XvYS^HLq3QipwiW*WxmeGvg|H=UaXLC -Q>g1p^C^+33V%D>4t%%+~OgRTZ(wJ^L( -1p@{bi>EK&@M!TbP)h>@6aWAK2ms3fSzA{MhFpmy000$=0012T003}la4%nWWo~3|axZXsXKiI}baO9 -kWq4(BE^vA6J!x+nN0#6HD=HQYHXVwVoC5>yCigOr -?Pcd^(8;!y0WSFc{ZdUqA?-@kuP!L!|6ZggH((@fv!lD_)S;JHWNeIHj^Q4_Vw;;PcslwZtN#dcSwms -j;nJvATITp1r8b|cU1dJj=|yUliU{;ry+ -AL+xv-j(ToscY#xU+(643_}n1t&0yjH$9(~Wl>JlTYXbc)N7y+{PFjzwAR%&Ug?RNzs1&XciVjsn>x* -WRF*}N&5JES#J|$Kc7J4PC4apqdT-(ndLHL-wyWYAG5N8nn-T`TY;;x46TK>ueM6V}Qr`ePW!D1!#sT -i#o6qAco6psmIv>42ju}nV=x2Oy|9W;)13aCC-~OO8eHqtjekorOrt&bTr$t@^$*^^rud*i5{1NDv(z -naqWmz=a3O_d$fWi;;Ra&*6C3+n<8FjfX%7ofyX|Cs`9$nnKCm&%BrJk;e%{I++G@3sA%lmvZhAHYC= -mXH_n|O;8jh3VLH+ud4X1Rp_)--y(KK|2a91P13HC{vGwSIr{{=_u?>hWM}Xhp4|mutP$hoGCbr{C^r -89+V#7NG(G=0YIh5u`*_!`iGiy1pusD2q4CBvv;wbuv8xirgcqplFd^!e)(XVBPVI2E8w$<}ECPg8nh -LjhnjI#Fyy`*kz|yMgCEjARrbC8FG#*y;vw75%yeO7X*h9vEN@`rK>9yXV>wrg5?4>&eb|A;#z%-vjz -o~L4rQO#MFGA>+AVE+N@yH=3gVk^$1zCLyVqAgfxPTU`}Rmhghq*M*9c1Xi#Bp7hpdHb~T_%x;erY~l()&tVLm^mIFj>?Ym`>)Z{?xxR -@v+pKT->8pP)8nudZwZ&}&iV7fC$j!)vAbhI2Y*w&ary!;@vL5Ug)2ak*c$ugFlnH@J10qXUU&VO>OH -r&qZNf~%+(Fh=3!3IaiMT5bCW{3OmTvQGv9N;R4SoY4W7H}Hvo68n!9PGMpgO7)X=8+*=G&%Lk%;J-d -RsPnJoRA%ejVV9bI$POTByhlJPsSk9*~MN%;^3jn;jib2EF02)yRu)O>DTMH*kmt|b;yeXuxJ%N37V#o_C-6imAAW8{3sv@ZV8 -XzR1eajX&0)}pY#z=fstf?RI*92Gis@cT(F%m9b!bE`sH`I$-pQ%?xt}U1+3N37e0m8NfGrP&te>AW> -2?#Y%zK&&qB5p}lRumr$jArW);GGx~RA9a7u -6P!pnPEdISB`xu8w8q1`0BV0t>C&8PSV1B9pTyFc7|b^ysZx|hq70Y^aJqS6#{|%zmp`V)4l(u; -z&^2{-~If;ZhT7~v6o|9#z!C+(>!H)_O9YJ@*^4s64=lUei3uiH@Zd!KoP&$!k|tjlNrbfi0imgq|_KtV0X1aK4a5}c8O@3pEdAA`cWGb2-Kmj>gthLQP{EqMo3cRzoOht4J_ -It;r0mUt}b=1OZ0pYHmuk+QGwwy7-Opic(P-Lv<4@r(%E{V9zGNk%=5~Ohl3B{;u&wCoIjwdDwk=ED# -TiFK&uaWhn_yXylgmzVZl73%~Lf?w$w;HuhSC#6~hkH5ZH{@xr2v>0Pb9LUM(D=qcQp%R=~Ef)imldV -<=X}N=#R_w}G9dU_n_E(KBYASSpg*ke0ZIW`V_16-~LK01kQyJMfZo_x|>TmIoV92&6Y+>Or`dX{PaZ3l410>iUtr2>%(HaTvA~b&l$b92;UpR_w+A^d$dtYfYPQEUFJ2jWm% -gV=Q4$F;GgdXPjsxe`2vXnVzuowOE|u$0_-xX)Q1+(`4c-4GV(YQKUTzWu1Xt0C>lh3*&)yTaxeuoo> -80fop-txD$M5{96kw-E4G%QNU~`u>jA~a3DPM+t-h8t@_1)=%N#9t|iobUB=sr`OIhl(3zeVM;ZclMQ -L)05M_V7K>F@d7kJtE`hs{7mcN0)`%ifIH?dN@#sT0e?iL$OZ0r*4+89w<$s`iz$jM=B -{}#9YC%%p;gi-QUZx6NS6H&N3$Y@LQkFm5dQOyWQIwGhBi;{%A7uA;x&H~K*Ocpo0=#O_RIe1Jn2eu8 -U?zai1`Kxi?)%p-#hfoU;CcMhkAqDuO8w%Xrkoncm6jkd^46)6Gox%eMvr13BsIK*7;o^PvCewonn*_ -vj5}zZ?v?)%s(R_aS6b2mx1vOJ$e#51dM1&dx{ct;7tflxcf0hfjqtfcJbt_|CD -hGa^^Yprfg+oCS@>TRM`svM7s7Ln5!$jt$iT4^16TiJp;Zi6u6r>S2MIFvXPRYoqM5mpAi_&7?zb_k{ -^09A*b8JB5`_E*W1~nM*?KJ);fWmUxDT-%F9P!kWee#fq4Yu>eUC5`N-+yYjj#M%UNHp{Te$ji_OH51 -mu2@nCvGrZhhclup;BKuFz(P08H2J@4*B6oo71d4i{6|6lP3V3fSpcywW|aV4fxL?}r|WfJ7xy6iuGM -?&auUJ2L3a!C(cvh~()>I;2KH_G1ajhPv)kD -zN4TrhKTfEl;Z7ReDh~@hn9S_RwA86TWXQaPzxxyLvP^d*T_9zqd%@20%AJ{l!->H*f0GzwPrjg&N3( -dNQtbJ?go+oCZ^7bOpFkhY#S=@@4#5TNi<^vkwR0UIc76XRL*{s&d8GEaHf<0n@rCtC93j=hy#`F~^x -!A;VpVyvs?OBf$q`|ayhLuWw#FqL{04hakhx!_wEVd`cu!|bJn64`*$M7b0q%^PHKk&h=X=Cw< -8mQ)0VSF`10xMiL0?mq^CD5h&7r)@ryeZvm$fWgalpzVUp&5kU%!fpQ(bOAV+z~>)PLlf8m3bhPq^A -NoWfb&h+Yd-TvZCk{Sf~|ms=^E6&Mf~+7@Ot^LPPGa*?c3JDY=)eI-VH78?f#ADZ#d7PjEI#cs&D_*4 -eSm~3SkV;F`dW{QbE3c*cKIq`b|!NYxa<9gN!c;#lSr`DP7sO3G`aJkTN(eB9Cw*f|HBIZ@#N%H!&LJ -7}yVk=$rT^-Jp;M&4;;FU}>OIKl`#7{i>*+8<;vFbK2Grwd(95`nRuEi{lX-Q^&dy{u|M=F?97rl?8B -{yvo@$iCr>9r16hrSW%V=X?aPTdKKre3&r&EQ!oTX60z|b7X2D5IF>kz1yxH@5Cd6`_A{rH92`W+!H| -`Ddu8)Qw#)&u9-JKmfINq7y!wEMb9Ne?YFql+TS*oKx92$WTri)T>DVt&x@Y=K>OK2{WVHPp5ZI<3$Y ->}^6%K@*!fysUGD6saKiRjp3fKz174M#gb5W!5c-!v*ICvH%A=@}m7DH3}%qfYqPKvn$5V?;B -{Gk^A53Qs{0xKhr)ba2706L^W)ZE>Ozvd{Abv&>p -5892(=PS2~d$xQKA42m5@2e+FBu+ZAED9xoJdNw$m?v6%1cpOZ3R?SwI1TGwCUrIyP-WFuvyJx_jS)8 -KrK^mE9fT~=gofeT6|7_x#nHq3BM_WTcFZWdmuQITN9)ixRx>F_0ra?<;IGoaM!8x)ZxEX95U7;tCVD -SJeOE0eMFY-^7Zj$09);@#_{X}i)mDEK@kl1FV4FzKHjt{FfvXeUEoGoO}&1}7)UR*d{dR`#?DS&|(X -BfWe64nFZx}p%eLr3+cLW$+@^z{}V6Bq^tauL35Wy?4~i6*4PCtnpD&L&`^uAyuPZjD3o179<$BI%;hQLw#vE6yC!spEPg6xG05G&;y9L`xyJXqcmM8NnFQNv5(VZmPmnXPJ5$c*tlQ&vxm -k|n7tW_nec@A-a=>J86OjiAC7@SLs*aD@GiWZg4<(9_3`iRL`w-~Gp-^}PnLhbW_OP-xS~Fy;l#$oMt -j22lwQ=E;Rg8d!eBdpXLsn)rUA+sp1kZZ!{|=~}_Gc5MG%ybRRYfn6+tcz?_9|oo5p -agHKcT!BXSdbHA5^yAw$q6~*5u>r}bZQpO3*$+AxIVb!!^jY;2C2tRQZlFVy5lf)s;0=_f;pi_#-S>& -Xjbw)3sR{I&T$i{6pw@C35Mb-joDF3@k$}Kw3a#_ZC&%)t0Biy0jYb#q%n*4H0P+)N@8-Rq?#lPh6zl -rF&d#s$w$R7hd7cN8565Fwao-YD)nrF9u^4z(p1sG2Z2~@O?-fSU$Mrh^D9c){YZ7CF~zJfjXY5cUaA -EL3M7e+ncA*txd`R#Wj*J`1s(uFFDyKwLFq&izxC8IH01Aqr?-)`3`~A6TM7nC#6`wzWVLpj9sL~PtU -k~ICl_VwO0nJ!mj1>t0vwni9EnLi^zWuu@Hs+o{fx@h{ASP54i0D{kt`z&Ml?~}SCcTXDYWg?Si@2}obcyx&~29N*2_hdX8!h8=>HtmhjgPkN^-y*E5wgArd~va>_;H -8##A|t;b|6yC{9h~s{oHi*l-C`7q- -q0J)WNQxWa8usn%^Di3j8pw9dPwbH@>%yB-e{%T(596#UM(o2~e%%Q&y_WC|6gbHO?Cqm$X*Q3 -(W^?~DQ?XuoB`1Ym+SKI`=63~66AYyQ|K17m*1UBXw~9p3SW7WoJu(vLm_vBv?A$R0vS?5X&yRj`_k5#;l6d9CZbKmpkdceg4X8XB6;HX>{R7-kxPmTgXXM%>XZ)_^!Vx~K}D117zBNU -|E;t5vg&dr^#92wwxkm5vdLoHW>n9O`w?rqK~&~7r&T!>;+=+Q6=ixhg^V7swp$=%$20yv)lGEVX79* --Fpu$+7`SKVgd5MCKJJ|=PvkDM!we1{w?oI0;fH=DiZYafXmd&fbI*c#)*~JPVc2Am9H^^XASPrVexe -Rvl)sMJ^r06$RKOS%tu#}@{)riOt1>3;uF!WF|7Lb^8=Fr!m&z(8z -=TcyA;)tpTzQSt_+()QtVT%oFQW;gcGgl}*uw5u#G6}T*l<>H)aM)fE5q&-2sc -zraLDxZ{~OoB#*(?_ZaN5!ZYe`jf%B1B -9j3$pvnu(a{KdOUC9v|8%YF>J12OSpdi6%;2_-6$b4Vp?qND0_#L!f6{0Bip~e)5`;V_U^+2tzX%8KLRdp62f|%~N28Y(-Juts8wPbl%=*_!*-yH6ytO^uvF-YjYpCt)!2&5{vY5KD*FL*G$~lH7 -}c4D1*o43I=^uPC&lIDq|({zq1CrF7X-^TD~mzhRjf=K|OFo1Do*k!N;uKFOGXI+d!h#bq`q+R~o@QT -RAhpJ13@@0%k7W!ai38QiUDvDZUGoL;ozOIX(-FUM$JKHkqYof8KA=Hx9I?fmh@=U6j?jM*iH&y}X~u -b&HjpQ&;>HJ~AUAubc$x@Eqr+M&X>GR2fEMoUT#r?i}+Dp0CUN(t%kl03%h9EZJd;e9(x!~s+5_A@$( -ncO5O-2z=@H6{j)BHOPeGI|8RZQ5f@v`J|j^b~<12qfh|#pg?1UTUM4iPK$`iKbf)UjFEI`%t -K2qF^T0s;FN^G#XI=d-qeAZrfZB74l|d^(3m0SolguPupUc -g}<|#e9to0GZNX?Rsw7K~2dm&x|;tP-Iu1&{Wp0{od0`N9xsdIR%m$^7HXmpnZBNd}hr&30$sHFiEoA -0Juc!t-4Be8JDZ8RC7wq2_jgql1~oc5mXMpuLsbIhLX5)ZU`d1j6G+7^O`n~sZ=P&|TE=dQZXro(+A_tZ2k&phxGF#)^LuBTC{_=Eqq4yl6rJ -ZDU5Et&l`XX=fMvO!?Oj=TI032R&Jf>rb#d?j0QS7~R*LWhPs`<$7QfI|=KkK4x0@Jz+&)t&WwtZ;#+ -i&^(aZp2zYMO7#d|a*Y&r~hU50%Z$@G1gP2Vt<47&a`%xv405}V>2E6t7FPy5)NO -4R@r#mT}!@pc6b4MGGURVw%ir8VV10w8>4%}Nx^cigE(&8WIYL|W4`azzZC8bc5*dA@Cv!aP_Zhcl(A -^|++2~1DwKP_yzD6#`O0yk_Q>txU%VEOVnTd{89H_LWXVux=<@>`dQ?a9^4*KK&tE)$^-ONJlMoa5(D -seHJ&mtA!kf$GvZ7QXL#q!C_o;f08-foEibVzQubv3X|ky$_l1Q3jHU}%b8@RHFbRTElRqIpr7!^mQW5ncwIkDk3UBt*r -SSbR`&Ly{fO5y6zTETUF1?c*@0a@agf0)@p3m%3lXc>R`EDk?%HmiB{E9Q%^gDjw>Ki*M* -&4$>Fdd5oD45u_=-5v4!bygC($LVPkFnCE@4May^ -0ag1Sn((1h`2V`E9YAyh(;!If58i!bdfx_Z>e@?D=YxSwZ;FyXcSNU#Yuc&_A{6b`ouEVmyDH<;=QVw -1_7k`IL$&2c#=_a<4=aug8qV>+_DBIMh%_+8h6*ND?{Ozr51BE7On}udbx;d>!1491I;FoAA<{nG(YB -Kxp9&D2kfS^UYj12wYjJq-pNVh6?i`)k -#u`CTgb{f7(O06jO+jVX>B>^N*GhhRCv!u9UVo|}H4QY_~c%JodqTY%dT<*1_s_tEe6soM4v1QGY7WP -1j!4i#w(^rI6xmZQh&yEX^!dt66yc86ElJbdZK!2*s0V@dEi+=#as{1B@h7;7IF53W0i!UL;*1U$Rtt -l0+)8O3-G08I>=xbLdTVOSCHXaakM?q4w%&f|68IT+1+&l?P9ewE+=FuuGTylU}=`)i~92gS{@as&X5 -I%_l8;YCPv<}Jec>8yJm*MDxevMr!)e+>h_L(!4{d_J!po=K~iA*cob{WJ^f7`j>L0LMG1H~;D0BGr% -gk@`Dd{Yq{X*;>dSS!U!Z4^O{0vD8!+2=X<-qs2Y2#4ra)*RkBIZ;n{zUQz9)8vz=ORLNyX=b$vW4D5 -ZrXK9acYd(AZ0+S;cg));+_t7>vKx$)Z7(?dH3v;NSr+)pPvZ9NDhWeng?38i&&)!I?H=_KYmL=w(YV -IA)TvTVzcJ{9Uvt968Jo1ZF?E|=Ycd#yl`x&naJa8$OOqhE{@TlA2^S9RSGS)C>-CFF#UcbadouD -nBQ}d9KinsCLn8sy@}6YeKDR}#7h3szZ=O|+!R=rznRQP>g>zd8(-rDGf((8un0S6=K%2-u!4?6y|1L -*Lj@Ix8K=*&MS7mX%y=R4wcY0`#Lxj(7@Wc41@FA-ZiF5~-26Qc7pVba5{pDcS&tOG2>eDCh-adc*YG -%(>@o@oE_d?7s{@LHrHLcVt0nlMRvxCv;)iCmbqNY!Ic -zp-zVRMAlVymFk}qsYwk?&?)~8DGX6HrT`wf_)xRO5 -s5)@Fv<_ghAfa#r8Q83tu@HbgQ`QV!!tc$(D1&{wPlkW|^_+YOA9GR;JhJGvGzElECL>s`>8EF5&s=( -QAwu@d0&4Nc!65!OwdN#$Hahbt6FA$v84+g$)s=;PV3WBd)GkLL((#uZw}Dh&k%%4TXgbfJOZN5tWlorY!#U$XhYt$MCpz#lUbwwp^ -$Ied~I+{g_x9z>+u`w&K^U5JxV*%BrsU!PXI-L5gu{@vqX8m1>>q4S=kuZh)TUU*e!I~W+lAaS|#`HU -v?XJLzQr0!Y@TU_)qqtFeQ>zJULy@;6dK6D#9;LVnF;2#)~))U30eLMtg)qm=;xYr)~%zgrvaXSDRXz -O-pFR%1Wa{ugHSf;(&w_oYQtGsW853R0=zWr1dYbuj!U79dd9Tr}i(9;{^Wf}f1P-U*LNH_c|7wcmBT -Q5N9y{~|7iRfHcU@t46>m>Nvg77{G{!T8G*lTplWfR=N+%{p?spsdp?BMo$If9jNF`2t3!QGM&UQ15# -ra4}76nRl(FeKssXVtEnZsHsIDS$RhtPAgu&w7r5YBc4iutm%rET*^^Mjubq`N>3`Oix-~NqvH*Yr4@ ->{HNtkm-I`YW=hjmk%2+81@LqaKtvl)n|4ozDn~HOe*x -Sd#%u4P~i%n$~{~jjaVTY~8T-^pkB+Uxz@G3NM5t&w|!uvrwJ5}QvdePG@)TbSH -HU`}qeP}T=-pWsZ?>cR21UjzFS6JuJXrcH*s48OulE=(;R{caL4>8Lx#ac2yiXkrKz%Ix;=-@w1$B@6aWAK2ms3fSzG0}oo(;|008X+001BW003}la4%nWWo~3|axZXsXKiI}baO9lZ -)9a`X>MgMaCv=FVQbqk5dE%SabzE2cZo^Dx`C+|+AgrcXiL5AOPU!aJ|_Z8_9Ug$F!tMfmTE6)*9k1i -x_9qAoo+A~FrZO4TZzh=NLD$PLZLhT^L#LOd=5s)xu_oc9OI;jqe0pyQ!0vM0s$L$RO979bsKm6d=95c4)MFtw2UwUG3{&`nsO4Qb5ZHB!6)FCOHi~*Wjs@ -(hXPcDLvdhOXtM^S9zz{*Mt02)I`D%cCYg2h|Z@aEn8&jQdoBo_?xuct^tMez~y)vk9%+J&vbQkRbqO -rkZrT}Q1V3Rzo`e6qNNuXt+M87i$SkCg7LKd!uXI)z!{XNoQw%y5dYtoS&CJ7l=3fs&pfh&|0^mmdjw -ER0|XQR000O8%K%whe-4dud;$OfV+Q~L9smFUaA|NaUv_0~WN&gWaCv8KWo~qHFK}#iXK8LOaCxm$QE -%EX5Pr|EII#z)#4<*;X^X@I>!^mrpaLeXs;bB&2Uv5Q$Tow#^tbO~ry)Q~*J*w_-<`kv?z?lN(eQvym -ah_&M*ErAthltkhr{Dp=S`UiE)lpisFo~3-^B&)P?bR)1kPEUyQ+ju(f3s(PdCav>sdQV{dlY1yz_BqUdXk{eO=^0#3X3Mc6Ety$t41zpwv8-y66DR8uy*p_gik -(z(_Y1bGZU7m|mA!>RcMOJKwK1xxk}dt*_*6ZOeOBSf!R+?1g+rN1Jk_p?lV=U7RA;^c+kOrCF}{YHb -3zdNiTi>3n?K?Q}oDBDW$`dt{I$?Ls=}|CoQ9OlQ7>5j-N{Fn}+!uP#0sUJb^xfz>fa!?@7C*PsPnsx -$$fWT`TOFZ_gJF*B@m0hLjCn5DF;+Ve((pMz;Q7|*A-;mu??o>M=~qup$J*OTe@{(OEWbH8X!Z#KUj4 -W7+(BO=V<=q#XBk@7r`b_Y0J!aDWVN~N1Dlq|uEz-nY2nm`WHE^Lf1Va+rVRPLQ_LsO%`$TZ8>#d!l` --0uG{Su_~4P9oR`3=+3A^WVhGE^)dn##aNVbPp@t7cBZ!IQ=S3q1h-|680Sr$E#QC%BG{Bq1W& -yBBuHk#WsyTQy4LW@l$juL~5Mn*M`N$z!s^s0>d1@e6lE){^#e&uh7&c(T2PGi*jQ0-{a8zsnG3FtCY -IfSUV#-qN3jEbWirOHm01Th-(PGNLk!^s_uR(UsLsJiC2TE&Ouoq|#0-VKm>!JbDG+1v>vTMjTZL`%ENO}_oL4SK^NJ`d9a+#ym=gQ -Fc^4%zw&&@c&=3_`GJ2aN^MsEYyUj6{E$PE9ttV+F#~-}G#q%&WumzPbQPKS%6DKQ;mW>{kR^J16!`( -tg8mUoqvhTNONB<)6qJ@|phBA>+7{(mio;Py@K}Aq-DL9f_-=Cd6Ku|yn8POr8#GU{5h%{#;jic4|9o -16HcTqEJXbt|UlzaG{SWUS=N}hy18c{1B2vsbKpH%em4(!Z+uO?`!~empaJ5ZJw_2#MO!TqZ+7?+>ZYIl3rPRhb-QaVj@w=Ux&|nA&Zkb#~wyeLU}Rv7Rp;A3n`thRH(=U5b?ZlEuq-`S)_U94}WZ7 -I?qCe_A}R=O5;O%s)P(?Ingk2|Ul+@jVj$FJEtuiBw$sso7LPWH^yrpJWa!W-x?0XQ)maKTWdDE`clp -wp`M778>+^FapNHe+rS;!2;Z7E9UZm&1#E^S-Y|q5)sd;K=EW%g->~25fs2|2Gjl|MV_8v>ij!d8kG*6LVc5x8q*-9S(DlCPiE0N=6P1?^Y -=4eX?Q%1R;U&KK4JkZ>j?Bix!kjUY($8|-liivrCO9JM>hDF1YPlYZ#0e~aw@NV(9miWxsO+kB1|_LD ->MNVsjSoLW}n)SI%&eBL}GSm`O1c;ICeT-grdj6C97iw@D6iQ{ksA_b2uh19I^sO5+0VEayQDf@bBB` -DODFhv*%MoXtJdsLu1c+0ROo6;RkLEnD&uAr -6yL{mkWl1xKI%l_nhH$H{B~@^eTawlnUGFJ$8Q92uz2(rm)Sc7T?ozB`0vxS6!oPlE+dSQXP&AY!)um -2*Ol)=MBCREVp+lzPC>l?DGFfBec!g~WQ8R8WFN~`T<8~|PrO~+|jrPl;`3{*s;?CYdK9)5m2;IVhs} -#B}HG%Xg@CI{kPpDMXx#I=J?&h&&h%2vMXqqT`un;`v^_MnmAi~3GFY$JLbh-NGs*e$#Im_KMOf?%Ds -@uJ4E*HIfJE#cgs-vN4COUoxO;TQ0-g-P-qS=aPtZpHG_BZ=1QY-O00;of09jj$!VtEA1ONaO3;+Nh0001RX>c!Jc4c -m4Z*nhid1q~9Zgg`mb7Ns{Ze?;VaCy~MOK;pZ5WerPm>Mv47uYI>f%dSd130mZ6bO>Q>zs;0;!3j;V~ -P|=D)ypie|u+0z4noyJys`aUi{{p2X%3AF#~#M+1y3_``_N)f7sl^&RC+-3#nW(yP!5;uB -;~>X0utYgmZAOi;ufUVt4V-YA&NX+4>o6vYYX_2#A3Rk1AOhKLGJzDKTd!4DQBM`cDu+SzpRc`7=D#A4y -xgVAzme_bZNK>h$Pt%1%wOfT`SFlr}bT?NWF;keK2GrJ~4yNlkKNv}Rt|TSyD$}Y0wtX_iv15q7j^~S -3`E&UMo3p1g&ZH1tz)d*yRFNvwrLRa&FAKI=9vV3>tWRl>$mHr!bygU_koXW(P*FiWX+9W@hl*l|bV{5e`|;9&L*mw-EhEMZp$B%8fM*TEAb==Za-Xni -eKc)8&Yej?ACBM(G;z8!d9%G-?l7XOs=Qt+ke>8wif%*$A@~aSpC*7s)f*FPPIJT-+ISdQO|jv?M0yW -dffa#x+A@Klt*>5*3OfIg5%&3=B|qBkVv`82NEKET=AA55P2{4Jt|q;; -JX^HWK(VV9Lm5{?bthHw==}GJ6l)`fyuMePQ1K}l(z73lT>>X%InVkurPx6|EZ(ZjYM7A$Klk{*YBJu -V0eVZ`0>ti9q%2@3Q5BRFFS8EuSD{*Di~_dCC=92Gap6c^$ef&AKb` -%d;S?-lumZwQDY?iX; -@m+L>(R^@JvxV+$#z?4L7*++UZ08Z^v!A!^gEfsOw(*-$rqgP0h4%e;+XrJM34X+|lhYy_@RzwK&4tE -2#pqaqrbQ`uY^5w?}XvP7m<5f7K9(zE;+e~!?4bi`EvS4VX?8@AH^hShhYB8 -rGv}j)n)N|4wtYvQ_(d1C#5O9X;EXEu1<}Jgfr6PyN#da>>p4|0|XQR000O8%K%wh&Pxrm%?JPhEgt{ -?ApigXaA|NaUv_0~WN&gWaCv8KWo~qHFLQ8cX>NFBcyumsdEHrUZ{s!+{_bBv7+6qBz1lSQg2Ff-u(x -q~3naV6CcPfu7+OZ6V|8UoC(?O2!~N|y!xxF75={nMC@+<)l`IP+a -`drUnzJE9Wa6L!!B;}H2IoKRS7)IHqDr-r^vlxw5b-7XQOPD8SrYKV`x>J#W&G8LOARgSx{UhU#mzj{ -BXvJ!QJFJDYlIK~?6PZ15Evkbbp?IW>HI})MJdKrx_%V~~SY|Q@XGWu|kLN#L&d)!-BQruN9Yqn@l<7 -9-(#){H6u)8v|YiWSa*{3+95h*=Q3Ad -Vtj!1`#wu+SHtyCIdfn|wU3py%!eo$! -}{38^%!bd9C$9arqEqR<;y}!78;PnGg#L*FHR&>k&t991bEEAl}pAxPPHK1AX!1635)(?fdv1NL&F>X -OgTg^y3;E3X%C@Fz}wnge#e#l|gQd|IDL(RxD8PMB;hQP7qwodqrx*89#mjak(&q6O%fX4H*=0%$a2am!_4#?u~n*vBAy -W8fhCRQ42U&|b(#^H4XMGUBAyD5M?YOac!5>kXXQTI)i6~GZin)6hbzf5S+Mp=O8`=KoOIMgfQ_?h6E -mKe#i;tNt)YeO+pWC=*c9LrFZRZ;u^5i9GcX$*|(@rG3u#MV|ORa35yUT>EHJ*7INv2vnOptE4HY(oU -G%(*v)C!>=(#$5$ESlc=eyWoISgu(8ZbLK1hd525)dmWG`b4mM`(a)+hAaYhv!;%08t(W -Js0!QsNW1-o9vC@!JXPSlM@6UPJ|!KA;qbrV$338B%?WmW|3a0sbdT3M$SD<8+!Ea>sng{?+BFBe?;>Fzw&BtV8{0Ct(6O0xwmE6tnOIXxUjrP(y<96D8}<|| -;6A9!$Az9=h6-vhfbi030DyJ`JTP=G1yd!5%Uz$nM0%_NtTCt!P7D=!umpm*P{*#B4cq`>mo@Ne64&= -ZZ*qbjpdzmFEJ26kqO-}Kr{Q5K)I>iRc+V}o0}=_zVBh3}YM}G*8 -;;+*Qd=>rrtRzXcvj$3g7k+gCf9j(*8y6W2lUgZ(9aj$LD-^@HFj#<+22-SPnTZqX=7l*(7ZnhGjXi< -6UTv;o)D}u_Wz5<8;^3kovG?AVSN;(1yZ89e6EB%M%@A5p+Rr<%5U1( -|X#ps1S_A=E8^BZvF7smcsfCNm#Dd;fK%G1L7S&7kXgXDCj;fTR7Oa08KHs&^ -f-EHz6A^o(WYU1|Il5a$JqCG9fSl+f+n9cUsfltCDNrYF*9_neHS$U?sx&DhTmyKmMSR7euE?Y+vEOB -2D5LlYqh3>X{@k1EGf?^&P`BNIJUx-8;6LZvd~U`-jUE{^aPgiemZm=Ic+&I46 -1-2^wuRFp2UVr8;Xz{OoKtDYv9E>#|eE`lajDHoO4_wD_{I68n>^|(h5dK;sm^c5&XMy=fWK%$W#rhi -lHS`fKfPAyW;(g#fco%=-Swx0k7y8`vT^uhD;PQV1Z7!6WZ+(Cz7kd*EnxhYJgr^4$%C-tXw{a7GVWi -^bJ4isC?;`*z=SyKY`VUY`0|XQR000O8%K%wh?Pqs``UU_16d3>j9RL6TaA|NaUv_0~WN&gWaCv8KWo -~qHFLQKxY-KKRd6igOZ`(E$e)q57;s@KCW4K9^6bAZ`F5OmO>xw3O7>q!XCCcVPiyB46F@o&3?;Ktv> -f%e(pf-KJ^PLM1k4T=Job(7h*><_6smdq0+H$IY`mgr%^lzWSoJB-qh*y}SN;bN&4$xcmFt=i6SdPi|-u7F>~am}M+o5|N -NF&DrzwGf{$k|Me4z1Q!y_1^7F_$Vc*`{p^!Cl#lxB0A&jvMvsFT8BFyv7*`@mV-#E^O~F!9ni?Gx`l -2o6Tx^WmRYGgB1uv+xbd~s$b-0A+LuainjJ9EEtS(u~aElUpf&Z+ePf{V*A@}U+DJ=uN&`-@V4Y1)Rt -#FOl3pz=<09EKo(v9#FkR)xnmj)#0f>g;sBffmIVv0KJTJ#E);#EjYGfAUR(U|mIwU{!&qDkq-j7S=; -Y2FJexUM0K>?)&vt&Xa}*b^h#CfJo$Q(PvvDls3jkYZDcZLXUXF2QBB`GBmD885{y*FFN4Hg|HJXkW- -p5txcAtFNo(yof5aUxYrbmz1gueLyWOekuB(s`gs-Sj(ZGpQ-9pz(PuVb%H~?<9-JBsdL<4+(z!^miqufiXFm#O4t_;TRp#23B -!y{7HMC28KApV-<6yl%Xe$FLW(#reX7?>IaB!Lq0X0I3%Qk`M0iar9${Q5thbZRLd-5YWBR}cEJAZJd -2fY@*&Jsr05uhc8)OV`E9EDQl8>Uvn!BZ#)BZAe?ltdwq2BYJ152Z}uW&TUZb11sxWFrMMWYcTSgp(I -k`a!;)R5cNpf`JLa>ZwLAuf2M)Vs-B;z*e3Cqgx0b?_ts@z`jP_9Z8sqqba9HQZg073E&Xr+1tE)1L?7T7N8Jp!-U~}{xsd|cV(E|Y#ZkiCkFa~HbeY|;={I -HPQK=rdBk9J@^+z}RE=J#9{n>5L?y8#>7RxHE_cg;#a%5l?(T>n8V3&>Ey!Gj6hv->$L6)|fwewQb9B -34rNovW0Kb-e*9co?_l~E!15cbFSqXy2k>N@VMh4cEPhdXcKY$%S<1-r{hY+I#qg$=43X`4)fmy>Li%3&)THJ~@FfUi||;4D{m(PL-uT7snvn -Li8VA@nJ)_rbnyIaRmU@733dzwf_)0l7VO*>w+kReE7vRMH|PC1PfUP$7dy2gx2E?~rRdy|&0L;;InK -?#)KDHj_mqYzKdsen|35nkD>wi&SHrrxv-y|cw{u?-8lF<$TBAcVo`hrDO@7Q@ORO!iY9>fj5j3atE=ZGD{Vt3M^}9!gN`LOM0p -&E@6?&hB=Js5ippg99RPZeU5k$WAyrVq4z&fO9KQH000080LuVbTU8^;g@FeE08$tL0384T0B~t=FJE -?LZe(wAFK~HhZDnqBb1!snYh`XOaCxm*TW{mI6@J&RATkehTCXrW!1lqw43OJw2F)Zv>~>KM1`aLJGH -ZzxL@J3?Y4g6wF?jGvY;TM^z4~-~xW#(=v5pN|Aw3 -z95&&Yxn%_+ttne@`h|xLvn7IDB;dCuqH{el4(uWD^z3!O3qAMm;8L;3}p@6L_U*yCAm{F=UbxdQka? -M<<{fyvuU>g{PDBjkSgb}gRp+@btg*ho(K)!#!5a4@q$uT*Iec^1fB;Ey-@s0opwV7(uTkAM)TS6P{U -2zNPKwIf4vq`nDyG4FeiUvO`!*td;sx8a*p9487oVo+nP6NcdQ5E@l>}P7hE*ez|DheG9|&;QukNAuy -Syb*9{+`JK!OK0L=n}QT7avA;Mk=VligxHPR`-0_?}4h!mO>U -KH%jUlafX-VoY#Yk=A`14`hQ=SSSN(B15-bcWz91Id>ItHWX0dm@fMllBKcz{4A(WwFpY61J?UE%DdD+fzx`hCBJUdS)y_fQF-d})7kX{uwe}TYX7KOaDdm}e=DYT -jWAUnKIhMEQXN!G)#=^>o7l8UcKilKa&%Q77H=&L5(}rdHS7_WC+*H*Nf{lPyct}YLY2_s%&%6Ne$ZH -xX(bdTS=#A0VUNFenWbM%-Bd(N$4rfqd=O!V6?`*)@#Ukkw_sf?rW2rOh+?rmic2jbCER_LK(*X6p!0 -@1p&!k)*zIO$V4?BbGPwi43*~ZF-v;5#^D7)pn3#AM)WXEK|^Z(%tXXX1o2>dUQb6x$M>G_}vMjCr}1 -QSD_JYMzFqn~Py8^|?KqC$%3)XiJaTn1P8BfinvNoj76oDX@1+X-gvcUF(4PM`h29=%h*C<$c?>}jzR -z^zxyC2_MMJU}LZkZ)UZP%VKb3yi}Cj`z(8w0s8i$KKIG5I|znN>_uoIPSp)6eTdJ!WsaXBQ+!xm)QZ -z1kd(vEQQHIWXy2v@h!CSWVyPyzrJ|7Ua$V~c(Vjs|8=Cnf=qoVoqM`Kf0%|@9DX0WPZUgFJtV3dhdPhOtG#+VN&pXJs&KBQirl5511$rFRddtwU{dAv=fY%&i+}KvK2!Qzrh)pfxCAya*;G^bku1 -++i`+5X?f=Q%#;bWudz8L{Aw<4C)z3JS8%DorrPDjM07u{TXs-t8>;AJaHBVtseauWeYOsAWA2)x34i -qlMH0Wcmv3sgfO#826FA2B8$0&k>&J{7y0N)Z6xGG34Q;DLMy{EAaZ3`thK;^6w{eV<=YUrV#Ng0s5C?1-&rFVw9c4&UShygcF>nB3jDoP8#c6`z7$d -5H{YigN1%uTPb2eEb;;$M;biUU9Nsm}YZgf@#O0kt*H))XsrC~62y(IAino|V%|hTM3};4U`!Sa4O$f -Z&v1qVqU&FiUhnJ#4yT(N-FGoO?#o43tF|du3LVpT@Q~ic&9s+N9!}G}XRPRFIHwlK+>Lij>ix4=K3{ -<~`z$aetcx0LoxU}z-F=Kum0T0#YR6_*nzS!FVa+*DTNwK)}GFrhAw+Vq?OT}UX1Fg0p@HT2Q0(T)09 -bQCKrIm;R(>oQW1r=ooe8}P>4Xsq?p6>l`s!Db3qr;_tL_{6EiW($K7q1HX!pz5v{gkP{rI*{@@u`#;I)-95$ -q+xDb?x@zxnSf;o+nWO-hkAOE-y^B%2;A3EXhNVR_0^Y(aSfrTKVmAUpFM{5FI_?!>1bk>s-JWuT7HR -~9aiOYXzeevC*3KaNGN8kj!tzXl^l7C0>YJNYkAO9KQH000080LuVbTM -rfW*}zNy0O)XPX>MmOaCz;$YqK85k>~fGpW=*U2 -f$r{uBxuS(U$G8cEURm8ZBcuig1XK>AI5GApr&ejv&k0&+e*Uc4a*WJd{X^qGo0V#L2Epc6HsKy8JWq -M?d<}J9q8h7r*`F^YZ-ltM|Tq{_NH3m(QMmLhyh4%D;CA{}7f}cQ5K)`0TU0Up{}9UgUC@J`FEJdR<; -h%v~xkUx#PU?^1YOK6&x-x9{=tFJF~>S6{rm`#ii(pSE=F8q1Z$lkl7HvfOcVE1G@n -1`NeZBvFdhx|?Uq1Wf)7S6c{cJezoH5%IXWZJ8?Qef@Rm|P%5FBKYl2MdH=I#s -XTvG-oN`lKKxG-|MS29+28%*!@r}5xwH!S?A3ek{HRs*J3mWzKfQbO^o=o3p1!g5>6^Va$5|h>s47t* -KmF2douNP0PHpNm_lVupMvvm2erXR=o_^W#8=vr~ow|U>5wjUQn}=@VX>7IBPHR1%@Q6=;X^gSf#5`Q --a>W!{BEz|eE8Bf}qRxP;bl-5f>z(G@Dcf_(_N7_&;|Y(m(}O}T#{!MeUdSc1;+b_Zov!;69!IA~#B= -ltSx@2#%5ma}D^ -y<0y%gu+q4^1q{{8c`cuwL&uZlBT=t2akoH2-}*UOnG%b7OlDV_kWcrr2h5D)KUK2|(Ig)}S7!5FBAM -hWW7!IHCJd(MLEISY%0!n^EHG8UEy?Nt#x84rGgp5h5m>LRRsadqN3puD+Cj<%k?u+!GzfSHO5)L5XZ -0#y~Lst}2lIttWLpmGAW5|VtDUJKiPTDz`iRs -=^fXkYCZg0tq$Z-&L|j+sAs*gY`N~GdA~g|ZoWz6AP!m!5eq@BilZbO8=p_3mX`Ms|Mpk4T#UNP4(I1i&DA6AhB}tSdNl8+MEAQeV5qd?EUXdtIss>7ZBubU(6`8VS%9bfer -XYDGZ+Yz#S(aR9a-C(0%d}&r9rMZ8^U0ghK@*fbW@e(Cgp1tg6vfl~NzNpkdp+bL$&>9|#Z&o~On!wv -T^8OiOMkz}999@)g}JvJ!hCGd6DqGT)0Tsrl@n%qTM4`#Je2nWU82w&CGcJhPMO#>iY-l#)zzY|86A@m3~_jG?U-j-)O&H8?^+s`40ujFRo$zvC*{95!*c2sVC{w7aukq)Kuc5ZPvmQiIL8mC`ghDXDYI -}_GGH`rAl9_wuq<4S1&Q180q7c9M=B3+bHkT?`r?$v2_6nDfOaVOlAjv2bq?k>Lf8dr8aXXsS2# -T4girXaQ8sni73x)^Ud<>T$q%CK1W2JPWH?Th4vT3aj~s6FuTi`E|9<1DnN=ri+{C&Q;S=gGH%XOY?- -c=!QokM7wRq+wg|)B_Li?EK<}y|#J{hp@3?4hKK`frWE8SkDBegm5rHwny$X3tO+Ja56KsN9mLS6&U5 -=OauutBf1UF9ub(?+5->Wy*==(Fo(Xz3W3#42sw0Gd4~mGqnGD^Ry>v%D>1$DLUcP6H+!k}z*G2UEO; -2HfnJ#R&`Z-Ec&N|z2%XZznHjVOrndILLkTPMOO>ou7M_hB;z^zMm3X5W3a^K$uS#l64%WT_X=AuIae -H*Ge2S;9oZ5gU#axKtx1~MUA2OY{2cEvHsb$-ti3r(6_CeVCY$m9dzL)G(+BTHE7F$2I&_g_Dr`coJm1O$JZM2A|%Eq((rQvE#bGeqNt}$36rN3>ugvOQipnJf1G+}dU^gaW34V<6 -UsnyzIhO6&|hjT=`_M>{amQh_+=is)}+tqcdYZTm_>e5NBIrZfdHBd|IW@&YN;oOps(W-X=uOxSH!(xyeeRqP+btZU-P%f81GPke3VsjXj;Zv{tg -VNuCoz89(D~6Q-Ox&pOU|E~f_iO>UXOC61Gii87^H!Y& -9!qaT++%Y`fJFcxW9SK?hw>&e+=Z$#9MqjB|A!b=1VPjm$?K3^DCl19uHPP7ORXP&-X?#^DO1S0v3+H -rg6le_gY6qgOQNFMw1C{gC{bO@#pw6DnnTKl|87^rz!xi -;IS<@_I@l%R3>RQyLuN1eb>lm&@8J8Ir)RpvQZ`HNTKuyu>+g{)HX}DI+*s9-KXU0}dX4Pa?O=g|Rl; -PSm!A%p~W`dh0xW(bxm2uaP>{(q}oriw-oTU%?y`Z%nlrc`mqUq?YSG31fqbrr>)WFkg;HiP3YG5?Om -DH8hr4v-u394FM1~;b$Zu)VZU8@a~vs}HrIz6pv&cW%lGc>0L>O#0WuYs%6z*Pe`&8JR!b<3m8no|Rd -W@#%LsO$c!tHH`m%W!I7)-34|t%uRfA6lYA*D}qifjS!Np_xC9$$4mCC>p3AI_HN@{m`lPb?#Q9cc-p ->2A=AQ!==-BJ%c6`HJfMDz%)6>BBT9IkC~+E8bwY^#W{MXYm1dDA1;hKc5MrPHBontz{8d&>bwF|30J -ds=M`9DxeDyGHQE+#yx!q1MqN$=uQZ$Mm>8Hk0&N|En)xaxo5M*UyKxD`354c_Kxq2s`a$SuxN6L(rp -}q=B@N{L+~0pC@8^T}b6=4>ZO_zCY#Z%Y9a%p0^UpeKc=G=~eG}KGZ*qqy)vf&<-O;aPt+CJrx{-@)(Tcklx!unv9)Kj<6_9HZ@hq40Ap&>jagKohiIg4=$r(x`)a -=mzBy+j6aMJ>T(s$Mc<)?{Lp=0diPi3wDJcG=~M(VnHz24gL*&00DM`e}jL6e}jL6e}{jEe}^BuLkC% -PN&&d7YYS0I!2P{FU_B@+_h{`ws^DhqN%RRA#R3ztC$$jxTp*1AQy`7NU3}XBLP;s|IZzfhQdLpXjVF8o0%@ve{vSc4@h6S5p3%M-h5~PDIR7&9kg-jOsxiILX49W|Y7b ->h3)(R{A*wK$2{n*iu9sStRj~)Hk(T^Sd*wK$2{n*iu9sStRj~)Hk(T}-tYwIxh>Bo+K?C8gie(dPSj -(+Uu$BusN7LX@xg9uGfO6KUcj&AEbDbZ6MJ=M`uoflTX8tw$y(T}D&>dr!e+tIJM0O$`;ci`z3uG7sz -i!E6YdYKpVDjwd=2hLdMFd4`i`IC%=}DX?cad5Z0UANn1B%HDU_jRMp0AQ3$Sh`UhZ0UCK~)H9g;NjhM7JR` -~jyz;;RJ>$wVt~}$)1HSTLV?4vkGpsxfsaknz=R>s}RxLX-_E8}iu+^vkem2tN+?pDU#3R2dreFkm6 -TR8` -v@LTu+d=v}6g&#xIpu`^grBLjJ%|PwR07d~=s1boKVT|XZQi_6yRqi!=K>?4OKvOl>&c(zrbJMFYp)m3;Y1T3aFn_;RpLuz@?Q6KVYZA -;3xo%+Yhn8lLRwF2d+*$^vk-4XA!KDc<5Jk6%XB|ZsMUY)LlF@Yn4G*X~lYq=fo4lLxa>Pp1yo(Q~** -@Vn9Fvm}n2;1;1CRY5^>?#{!iYDzz4gp~h;J1S$y#%B}Xgan=MVAR|`8j&-nO<=U1K9(qCr0@Vkcv%t -6aQs+BWC`4bYBoN;^#bO0z#mWZ%RSJ3WZ+qoUCjx@MI{6suTVb{u -6!>Sm}iSg#Uyev{VZ42lxa0;HN0i%k@_ptU!h_PZP3SqlV1Rr`>fk2*sng*hON(;jV{m?xG3xSfjIskohr+o`ynirc -9;kgs8<;&v)-r{Z=hZl~gQDh_mPqwoUxvPpb1JewYyWhJGHx0yZZr8mtzP!HM*Z*(RQkIr%HFKbe -AeU`Q}NYffoXa2KWQ~0sbHmJi;U4k?_8B36Y5+^t+ahRTB^8jJtTEfXp~gh-L#_wg|`!{1pJ)TE~H=@ -e@Pf6{z9!3>=a~CM5+xq&n`)4h^)5aM_9;7jJr$nY|rnMA?sX-t6_3I{Q^G1r)uJ`(EI39ZmzsB}w*i -a>i3#GdgQ@(ogF7c-9-|?2BfuuV62i)y%EqHZ-q#oS4^foSWA}^RbRY)4ax+X4W+Ik}YrJz&~$J)7-` -_XWnKlY~z45Z-=J2g>h}MrTKae>e4HSt>_I#otG5@-qT#$3>;tIml^2GaIJc8vFo+ZzR$qpP*<3NBlC -Nhfyt@M&%je%_Q9YOGP9T`qtL&EF%sU(}875+6DB@(k2TqKKCE;UNfDZva#S-& -+w~mcEMN+tooFn`y{{Z86-21V+WSF?>H8-^ika&;)&pjGeP}OXfa?@+oMldrmxw87sM|+xUiW%pKZJR -LI^HcW-NirB5=^=M8mAdTq&>beU%u1=7+OaUn_KH3ZQwpR_zbZoRwti0t~u?DLn=MIfl#@6(-S~*O1x -jInI%{$;nD>IBoA2FD_Bb@bt}fe8tHS;Pm#8latE1x7HjRh@6CyXCUSurGi7^pwhcV5**J -pOc8Ou@y|MM)io~;X=60Yz5bgoyhc^i0qvrsCIJla)RzRCr6K`2o0S`2%U#WA##Lto+8G`S>hQkF-Bq#c -RJ7NXS?+1gtpAj)asxDZOMr8B_cS$P}vxvMO2gPBL1Qx^w -kc}L?+!`YiJd`P;V{M%^Fo#aWD1a_0GR@0DM02jB -9rIrk?31RVsPD{v_>Z9?Pm}Vr;}L`>OnuK2V#G6E@vEPd7NPxIax3qM;T;9+N1yC8cv@Ko9Xu8ms-TmO#c -b$m~lKDBALFRw<3#UT^CVoG>E4s=HxT3H)^BA%f}oab8@EPI6ibDeC9ZMa&nB}det=Q#Ir&V@$^r-M7 -*3_I8KM02#vYSH?tg8_P*EAqvBbi5`!R^2;sL!uNzNCv0JamMmfXbm)>rN#Bhwc$=S+Q@KlNEXZLoO` -q&K%93v8h9AfTr?!zWfAdX!E5s%cw&PR8taF!2tm%Qy9nC-})9g@u9QGno)Pdiu~9m0I{e%mPWjz=J- -Pd4_DGoXWWt5fpXLGz7DQ=J?vo^rf&_KDt}lIF=t-YE&39KHofAlP2jK{6G(AhVVE+A{ZA5eILNoPN`TZPEMGN-tSUw*)pjYTM2U&Sh%g^>j1yrZhm5kzSzNq6}pQjKrgX5*C|6I9Q{l?9DO6(9uZF! -4@n>Wu(i=FK?uhg5su?5=9JC<*>@e`l+FLiX~lVnj9p)fwRO0}^zBchyLeckizM4g($d=lE2SCHgwk6r*-BZV&O(f12IJ~D -3w&M+E1#nW>xx*OVq7$@f*PCSQrU`rCy^{NCJfSFU5yhri$PTu#LczWemnL}zv$~L--2U)ggk;}7)ud -zIn>E+e5Zla@y1Pa#c>xE2raCPRA<;;*AFJ@J^PPOpJ&wb~#}=`B7j>cdB5X0=$eFu)gNf -GbA%Ge9e_+U{;sPwivI0>-^BFfyWF~fhjiw-8u5WnYJ -Detj7L*jC;dI1Zkg>zILh(l*mn{H$4LK3A24T$dR*V1}YoNSwQhuT5bCqsKI6-)7)8h2J!=&tb!Ls7p -nOjzy!V5Z<8zSI3gnHRa&+NDU;E)_J=lsw)BuAR7%) -rqa!Hfp6Xg>po+iQDb%Kw6lQ_L0KRt}I2x!OHOy1hpaPj8%{0pp!;7=PQ9m?KOwdAex#rj7! -1cAw-V7YV879p%(*VtyJOlH9?)D}62&_Za#Z>K`b@t9W@YH=|kHcZ3;w0vnfht{PR6xb&1TtkY-O^KTh5E?E`vg3R%*D*PdO))iuS_P7OR -~pvug+I8)CmLT0AOl~IW;qp$bA}9>*NZ8q-~#xL6z=8<$5i2p6*8CZr;)tsNV==4N;)5B{OI^Lp+25s@s6YcqW9es1Fqo?e%7{^| -(5q8WRA7NBYoSbc;T!u?;Xv!-^HUfET8<(O@AXnqeB4JxN*J1mUbP`Z-;CAJa%lq3>8t@x(NtyuR@x-U -;eO~O=Zjr(qSI46z`<$|4@Wu`HBL2&(xlW={<5e1w@-fc^i7|Y-?VbQ=|4elI8nA7YqlK-6mYZw9Bs$ -|w;Yy3Ws6cm%bBu8C!w|QTlg*f+(KFuhxUQpf%5P>__>|z>!eQMckny-9sC>?x8Hv#{9Kd(Rce$9;Sk -$$1!g@cZ4c<0<7CU(tmRJ3qH7Zvld)VZ*_D*h{bxD(wp>_Q&hIT(FP5v9_L~fae}jL6e}kW!Ez8lb-Q -Y(>TwmsO3jYqj=v1_Te=Xo&3;5T{xwqxq8<16_n*>pWLy+#^g00*fSgw1lT=!V+3aw~kwCJWo1x$Yk3 -}0~R=*+VyWwhLe*dVWPK5jt=TQuicxf!v*gWHUZhvRVzKG;SPZ-{&=3dWuk(X3|y0$ViWSrEXM+iuG> -q~#jYf&jK4fGr4MD;=?r%R(->7Ps7Ww>cV0%3PFNF3K%;a<;-*r99xcE%)J;%R1YSk@~oX0$-<(JNme -zk30IfqmMiKxTB9d`naQyJNmezk30IfqmMiKxTB9d`naQyJNmezk30IfqmMiKxTB9d`nVGmMk -8)=-EzGoI0-YIqz?j1nJt&3upicJza01)1lI@ZTxY|&BVer%ECR=BE5;zce99I*ah_XBDZD>wVgJWtIVdYk)T|5`mP6XpZQ55>clZ<$gr)(!Fq9)EVlb8}xt*1V2Ih#ZTdGtXTLh{M_dh3%`Zm!p}uf7 -gO#Oeh0sU-@(r%V{!01_!syW_!syW_!s!m2U!;Q7x+E=+-(&PzlWde#V()RDf}M(75)`|E}shZSh#L1 --0u|>Jr)!_cH!er;osorma}Z|gET5!!<7wwZsGcpu~YbW__=T^JN!HR=+H!`tmXdLa(`@5DQmetwp<@ -uG|Dy#{|WyI{|Wy|I5%V^5FQ8*q{DSy;W`fuvyH;fRcVRvgYhmA{s=!ErLZ0=T%8uKPD_Fx%(vjON`g -PZ&&sVN_%r+&{tQ1?tR=&r;pcj?WcUmGC}}SR{sMo2pKH^?yg=E7#uDi-*hUiu^@I{-UlD0@Ef;o2+2PgII!S@oNt}askvsbx`2p+*>(tB#jawSo{OT{Q=_s0C9f+!8`cq3I7TI2|stO>4YC -Oph+NmC;S2a0Dpj=>sWv(jlv(`=VmrV_#^xgelBTKgg?R`;pe_KCHNEk34X3_Q-VLipWx>fH)Z%U{26 -{ObW?^u!=K^L@E7l(2c#~_S@)fYL%^OSM1}5D=#EIieh8i&9x8FC`gW>sm+CwD%t>ki&H!g1wE%yBKZwHF04tZX?Ex2 -e$a_RvsFTY~y-b{kcx^?idMBrvK0i>`W*%Req;~;_UedK)phDN@ai5;g(?(}xbp<~nDfCKP5uQ5Epyna!f=Ym8=dHLB>wOhq?FHR9+kZeA4}zD{uXYE -`q&qNQ9H{m?RX@C*fAdDZG{PEFIR-X1?8^5dj4uX?hT*F$qWRSK<&^k4HV+c<{Jo1RtY%_;DGnP_-Lu -5S~$zKxTwyoLGtP;ipBG7}l%Q%`@z^j+O~TzFpI6gB;4;vF$RaB})}iur*Px#E*!gJ1+fD^@nV*y|8v>t%R@Ahtt -6UJpqCtb!ahf=tEm1+PcX7b83CC>B~GS&pb!i1EUT0x3;mivl_5D0u!zUZcRGqkN3Kif)Vc0MRK(?S# -&UB;`l)&F~v3f-C(&ywn1}kBtg2l+9fdS$KXAO_nlcK9HqOGKI=uH`?QX3S^T)WeSxg1DOnDvE-6DsS -(hWL4CN%Vx}Y`UtQ$`tm=&DY)7|Ke*k#|VQ1K4m_;JkW0)1OM1x4BVJT#>4b-OjwI30mPfctN>9e2FG@>;* -L{y3++iCGvMb;2RY=3)UY2@wvb+df0In+@jMwfCLK_nK8s4+XMNGA_~w8I?;>Zt-xdTx{}{Xqc@z!jG -s|+sB7L;n>Yja8K^h=Lx*+O3aS_KeqP<{eSN&S#8)M*w9OSMY?NdPKd~dv*m}%*BuO -+5##W*IaQ%k2>n-UgBC+%a>iF22O3L>kzDKuJQJ!xR2&^>WVWksY~Ap8m7ybu3mDu;^+svqIsIrs%u0 -av&j!_a>XRspooQ_+6--2HTqNqrwR&4n#UAW5vo>2-;aqaA2&DFs*F_Sl$ujH&ym|QH5T`i09;$utHi -C28@X&%cU-nZcU*Jn>a0~kUnG=kk>KSB+N19vuq!lK>~pYrGezuoFv>R(6Wt^UHQ^?*MvG`z;m4vq(5 -@lp*DC=2l^^O=eyCS|sMqAhgF+4tOa)ro9tcj!uRdbU?I#IoJ0$MgLoDF&1G#Z5x&c;gj{_=FRH!5YZ -*F!`9K$Zj>~hbtbK1L?%gEaT?XMhGWT8+ea#*qGhW3!dt>W3Cr+9crgoF;1J{{mP&4E0~QAGZUL)H^< -k8_d5ml%FyIdFQBvXUT{1Lr2|F%;aKhnR%pMXAPzrz9W9`A+oDYDoblbj7lV8etZt)XUX9N -Osce1$OhB)dU|8EjJQzrmtFge2K&sMBavL~Qt26)9FHs&P2vHjAt$(d1#4 -Q#>MdbFmc`pJJ1X^G73gw#+8NHHeXzU6!2=ddb`HHDqH9+E+l3^+#WiilC(3Iltc}8=S7#eH9PtXZBq -@=!)5+7`j5LW{_FSgDKa1@mjfsEi`s)VFz5fL)Cv2uwzl -16i7V?}f;hPkQa{KXw#BW1I_N5^-o$cOmcYAsDswrX2Rd&>q~Eg1vKT3}m_FzaR=4a2uoz>bC@YH8OB -8t*%nr7duq<^^Y$I}e?Qufo9aP&yj<%zm0!K8*#3B(@mdY%JMVdJw6_X@0iO!Nl!wScYjmjdL19Ta4? -QuX+y2R;tl2hh{+8L2Tu~|G-B&5Y8l}Mo&jwlUEJPgv?&Qv+6#3~a)GY%O$CkFLg7TZf4I=4qgKIO!7 -$~hvUAlRa655%Goc?_6i3tN?6&MJZ*$aNr_a3D}3KPv4(P`VcR$c2b(!yzwX;Hd4LSjwYHXZ9}j?k+VQt1nJcv(yIZhn~jRerHC|0l&w -N!tB{cCD9E#Pv|HajAZ$@&4}p!PBGM6Ev_NEYD%!4UapF<)6t)hOyDN&e->3bVU9Ei>HI5UEBCyRwV5 -{YxpjuA5%@=icbscXNac-@e{cvs~lGQAQ)9*Od5z(A#T0@u6 -s$#rtt|o4}sDX-E$i+;X7jvVeI|miZXuBi*;<9PVv@4+-Hm_10%{c+dqb>ZZd!pG?UN^0 -(jJ&y~B@5kv`e`^tm5sEGk_N82e_d5Wvt89X^E%U9HO)0o&OA7`MKiZ4qp-_bO+bWM$Z)fxyMa}Ywh9Bl*?wVXLT -cS)+8nZBfH1`{2S~#aYjdPm8n7Ha8U?avKhcl~dn!%X3>Y$sg5ESKF>n2($)9t*tFCy2n*vhpi4f8!@SgrI^mS -@tu|ohLw)C+tqsoJr`=OxkWyP&Kf)Q%UCrb;B?G4=QyRGTBvgv`OLF-39JEMkfz+Z%n@6@lQI_p!!06HzENjU@61F5--$ZbAL$%r=g33o`jVF@1r@4V>s?DUB?_BUr}dd;(lql7%rkJ_66aBM+Z69>&Jh^(^(~Iq^R8yh@-) -$LDra+^eGGmk2XR*P2z6nQj(OPx69YJf8LvMMB<|(jK?Li24_a7<_6K_zIm=gr1zk7#XlpyxNM;!)$V -SbEb+ypC6wlYx=jQ#l^IPfpy_7BWpRYIVhwDx2>u>t1<_#Bar>M|6Mc0jJ*@XY^5bO`wqQtPbIz*y>3fJ0ciP4S -=i>@_(Hr^1@w7P^b`U}{$RMViS>g`*mOzfiKoQE2HGiq4E94aYd4z*@@gemWaTM9?U}YIbuh$P_NZl_rm9TTvPZ>Un5 -bn>YW%~f88b6g9vT&`F;$n)trL$;<~;Im&anwGs#UP-su?q=wx+dXiPa`HT}SR}5$#a*g;u4eMYG7|+ -|>Bis&=;G5Y*hJ2+&r>ayZY+vFwJ$(zf)}z(oz$9BCRC8m(E9+7>BkY(APR&p_pjlXKj#)VT$2&Ms;5 -0hox7`#?r-zQc;PGb5V2$(iX`e?{0O>2BQ4hG7V;pA$z=(D{he~b-TDWa -qnG(>P?TXm$w&*&+PPggMruJzoIdE84J}{`fU_H4@Abr4P&AU&Z{`&V!Wf9s>zc#>v{pg_2S6tozG+g -?xf1~}QO9z}ebeKAHm^yTrI&_$}S;|5eXwN+do^J0qxsDZ1qr3%lZ^t6}3E1%j?05opJOMkR`>oNQ2U -Nc`>Y(rgc02()vLtD=r+oIw(V?W=p`_fQq}&zxZ8Ad9OK3QGGaSbCTb+d(Z8Zc$N9je<~xjT;MgJ+jcLP_F)M(Jg=y5&d{MbRMK9SC=a>UamL-GNna>6xBh|JnV~-#q# -8|M=w5fB4CJPyVO(pZw4N+mrwOY5(^o*!H)#{BHecxI2(9;m386yF644wFkF&2Ws=VfLRfGSh*ot(b? -D@T*#qzr9HUHL#0i7^y`QX{rsC1wTCF^-!{3h!&t+{J&34|Y7g%4u+VP!wKnq0hb85H-^Qw8lT`yoAW -p6d=~f1HGbU8g_{RMTxnqR4J%FenRc^lF2WmxYU^qnuFy=nnMz -X==#wLootlmuB(5@b;lWKYp`b)p>|GerHt(Z2{s1R(N3&gp|3 -&5ORM$UQJ;cMfu*BOdUUZ?#6b&xCNCA -orE78A>sRM_p1Bc1D{t(?&vWD)XdFd)?x{9iPp6|(WU?JF+`#^6tBUc0{&juyCiTY==4(eUJQIrxHQS{1{K((S%$nw4@ -RH}}NcM%On7D_JaHkZb83o{PI+tjIQd{oQNy6GxaZGWGt?&q9W08{mjqC!m2wckA8GGX8~y24Fe5Nbm -4oPny>rK&|iQ*|?GCOAkn9d*m%8WFJ>yrIZ)zxQbe!M~(-V^UCS`uE(Ql(ryc*KW&pb}TG_c5HNo?oif~n5H<%!-;8%6VntYoA49U6ep%B&Md1?R;^W -5NvxreG(w8cA;srd47NF5j!sZRXjs4-egdp4={*#^g%(Z=CwkxV!eu5Z^f9VaF{)Gfk*RDADn@l`5l( -czVN|DLPN#BISy`pjB8yFfR0HoKwpFX>#@;$M>qGWC|3h2T^)nV!uC7QFqrb4NqKYi0ID!={@mU2o6j -?2S%VDi0c7Api>LQNm%ys;5piz#WfQ`j)Uw>9HgVG*M#6>3caZB1ZX?-pF;Xlik>i_Ik1P5poTU!MH -0|KERWm6KY3fy7cybKHAm>-?5)SdOF+&xEkg4+s=83v^)D|=h=@lWxJ7%Yq(XhmAP5V$)^CcFW3|r# -*Tqo9t9Yx#*r`l3)S37v_p8F)+*dw*v#w#yaarL=>DT6L)m+M#(=oG$)2J9H=uVat97kgBl$JZ)0gAi -wY!UTVIz2NRrhE&C2eu%I>hsyfu1C?;jUJTT+>ltfm^OnoUrf)BB>X@f(?Uwd1g>yvZKsQL*Y&(WOUo -_KoC+L%IFd{?7JST7MGhdO(RJ)6-$Xw(Yt^<5BdExI4qO|!!krFyB1<0GYem`kMTE>F5lM7Zv(3t -=*pTI(VY%1#S3+0Waser`a#qWRu%cq5qGF_Cw6lu9M`d29f-^*&N~2c{FN!2aMHJVZf+#~B%ObZf%PB -WChI3!yPxe4@v$8t^62*GlMmnu%b#_=eb!e9B4@Lu$w+=g}4z(+dilHWUM^<(kST1!~E_GNgbyzNSs8 -(@S4vEmA;?R%MsF;58V){uAoQ3O<5i2M9SXdI>BJBYi8gm_5MH=GPW@ywIC4K#yg7u@O2LJkZrj0j^{_#%WUK+wU>fmRnV>S(G?NaRrCyV -tBAR -ZwJnw@MJ@~q4O3*Y!$6v=^3jTcsVZ|8a?!1JW%#JN1vctQvqMYPUzygZ0I6N#}~KDRfEJwVk8;I&!4pavh>RHg0J&H;({k10i%=Q5DmHqs* -n%z9u7=89hh_qyWqt7Y|wsjS;rJvGq9TB=eIqWL90xyiukh2&FzL>P0dROJ7z_(csrcqINd5>0gJbnY -ZyZ}^X?G^HP;q!pSKr!cM^0l01Wq6t^@@bD`)y4Pt0WcFgV8(hr&&_F8Rh7-9lACrY^J=aMQc#q^ -)K!_S!VG2|WUD_Cb_#ylp(ry?WtG{DJH6%_nVU$fybN4So_gA{-nQhh%#FpsmgF!c55le(2!-rrRa}_ -EL}i3Smy1K%qt2SEX5fUXS!UolmQyCoz@gi*{amZHa-&AZO`3sfWZdK#IGU(h)#zI7Fx*t9TD)pCpSD -cn@Z+SRy7{^rJGScl3>>#p)vYsdwZpZ|z;UEeXT#KGh?UlP*;@M{>g6gka4bdYR%hV2SFclXW!(o3u< -pkU%p;$(2h?Mza<0cL<}r{$^_V3(##u(4I(_SLDCao?2Lz{1`PSxCJM!~G=h)KMbJk}+03NPsM}BB`Z -LXlMY2!Yq5&Jk38sblN-a0m?a^@L0#zCF;`&|`yTY&xQx}b0U;ppBk)Z4G`f5lGwuC2@Zf&^6rqM))r -*kto7Ww;G)5$%;X&+^CV#q-zCo_|>`{LSA!`PIi5Xvc)ndL812w>8TRtL@*?^wlNdb6LRJ#KTtPp1tU -y+Df)xUsY;#Pmk4Q;F)FIu(Nz4JLC_!ZxD05H%fsn<6E~DdcTXH(H}76bWa57vg#zDrpu7iu-zn{dy+ -`E!bkT_IQ5`Q_~Y{>mk82j!bxO{h6yKRi@uo&rw2ulM3U$p1WSZJE}8G|%Z8KKZMwvh9u -ghr(gtzthmAaun+qLwW8~>A!6rnC{)~2?L}qG8wSOa-sRv9&UAVUU0#tnYd$t=zMZP77bx-2{0mD)xa -|^Z$qV!Egw4OXw@%-=E@|4(|x@5f`lJ>vkzwn6A|ATg_NctW`@G@E@pZ-hOuDX;H`~iIglz^rXtpxXB -jWYPZt<@_$qRHSJ%Dp{kBMTw8d%#MO3A^+n{ONYHE;%|;sUY9EuTnt_Yu%_+c+1KbY2Kne-!i!+2r$v -V@JH=$U9x!(X=V^`*=3Cj1r5U*7fIiPnBF(db*snWJKO2HqyvW@h$$U-UmL_Nxb>cD2(|Mi;)A4oi4W -ex!2TudeBF~+{Hld7u~T*{#CQ*>`;Y~&dqR#kNZyAe9xt644_FJkWFhZ~KYp#m<7G`uAi|d&F(eYdR& -$}qpjiwWSuw5oZ`xYet>mTREBi6~Jv(Ce^ffLbt+!wp{ySM18$j1y_=hcwA?`UWjNus;#*pCrdJALsq -(U!qW0%zD4owmShT|R%{GL4MEhv1LD7$Bs?2nls8`7iqRHNXzAm-o3IGG~R(V#Lcl_6z1ER|jIrFe!I -DW0za*2OmMg94^v*6bdj{ULEu(SLH8H4~xFdrdFFL`c0F};Gv&38q6(#0Us3-{__gbs4ae? -_&XyezL1pZ@^u=h1J%DJ=USol)TD5u_kL^&fM8Y>_g)yIg(6XQAj9zb<;2;S4tDAzJ+y%aX$DX1L;N# -)R^pNt-Q->bNzFf% -zMj9UdJ7+!%7~y5-%%xVi-?s#62V@O_N-r(f3V8bGI;TMju#w8}{tQTApY~x^Ey)JR&@OD;T4a)L*fd -cS%xT2J%Fm(tQJY;z4)P$zp7X)Lho`?%Bk<kZH7&}VFnM+eWP}wcS+bTjY}eJEuzzZvTZv -lXx*a##1%RRTi?H6TGCk%ytae{bga9uzgYk^29}*7|Pr1N>3DeJZ|mYo -bpB|1=YQ4~azG(>Vi8!a7E_r`-#v3^2yr;@{4^Mw=4 -J9Qw==xM3&AuS#EWNyEMCBuhQcl=;B4OJb;gF?dhbcCob#=o%j!y8qO-$m);G?}RCA8(B(>c_vis*@W -ti_@awgW}aw!O4xm#RqIFJ}8HMYjE*Km&sHQd84-E#tNgV8@aKQxJ&x#ywUNx|2TePwj*&mrwBv(%Nv4&@heX{UH2wHBdP{W5xzR@Q-H -b$j7riB_t>i)B{O@LCa%$|Huit1uxl!Y(uRov#;J#tXt88CmnDU+g{c!zY)+!&;R`8$+%ZHV^+}O2Lr -7m*?(VWY0J)i;MTUfxn)xAH&aK6?orUnja>-&0U^z|}1%hkmHR%~g0osygz?XU_9yQdviUEjavoLf@} -s|HS`_kV6%2Wov`=pt?|ExPL6YA#Lv`?GprH6MQts|Tt9piNb3q2NJd1Zi@P^|}F56hUzepdv3KjA?dJ(sYYj9C*ckLD(ok*n2#HD -k*{y&o@Ibo}!T91hR21q8=T{p}>~T4Pf8^FcR~90T}t%g;IP7pD2(-a(;_X{Bwd4A3zq0daOV{{=UJ4 -4`2mJOcZl{&rm?Y^F0La1Ic&c`ThbxzATF61&a19;doEK4q%6V`9oBa{K;TC!OGo($C3VhMBE<(!2uE -DR_txq8+fw;djqw0AL;g8!EFjoGf=Q+1ENO!UjmgzUYHNQ9+dU~^i1;o5S&bYHmcZ5y5M3?)Zn8t<^q -jj)yN3K%K$4e7&kJa=G|bH)d*4vHFf=-1`n#)QTz1Y59sVK3N7pPt!D>Ank?Q!lRf=9d> -EcTh7ywo`&L*mg@2(PXu!X)ohA0Jm?|u@v=!ONf$I`OU=cwTnI&!4Y}eJ5UUsoCb#>vo93Za?%7ucYB -%SHo;kuBaZjS^N#YAF3pswQS1sNc(VfMLu{#}B+z7Em#Zv?SbM7@)k-Tn6PEfKE$LE&3pgV;L14PxtG -KS&F})kZIP8?jYxKw2k=tspUc=UXDSj2OTWGj|4%mbqe3Z-s1mI53|37O;z?53E`of`@tl%XI+7I*pj -VbL$PT3sbpo0lO}9c}{>`H;69mU5WfrKs$9ftE71k>xD^Q#d-naLM>Z+#0RimP*K1Yz;#8;V?;5yfw| -fc#V$t#1SaPC{s1~G+@iz$4}=bLe|8KQ6mZnPG9=iyMs?i-ay#7{1V^xOh76_%*j -&6WZHzr^;dg>FrK}UTT^poO3emKli9{7X5G?M8dP|^()N$tzt0E`sq<|pLo26c1;E7CsU^~&!-tkEs3 -Xz2R5k04UhMK=f{HQ9QDAi4ny1%&`TgbTVw`%HsHH)tQ#gn19`qplk(L{435xkxJ$t*w{yK?Kn8eIbC -<`t1$u=OKj8cf{?CG_!&iza#KQQ&(WkJq*oNng1#=G!MaLZjmu=>I|BJ6B+a80l^HHLYDrWAYc9pGZE -)sVJ6=r%;aGJ#yzaXL{MnpUttx0g;o5Zu!=|TypzJ~u2JfvM?XuCescHdXBGM!`U3i@)1Q9^{Txd8&t -F5o?DQXa_a6$BcYl$gy!8wCKY-^0!asoX1HSdaGpO+Jtq!~Xc_{a+K>lW0sM31Pi -qfA3}a{KdbN$1U>RZ -iNmwYIG*^P~SG_#p{s`}>4>9czGS$Y0Ay?gi5pWfLSe#P~4+Nb;TqbK&)AK71bzVzwK$1h&y_SO^o?) -`VJk9X#lfyd)L^Xs47rFZXfmd{?52j6`!eDOtjo*zF>{hj&JS6}!VW}4+Ecd-|_vD_#wMf8ayHQ6K#-{Qk-B;_ustdZ(rJ@yE}@=jG#%dye`tefhd -RIUc=x7hc`fFJ^IiLgjgfQq?b7{|x!@yOP?c&0Svh9xfaoy?;60tIt{mx2faqr;pwpJ?h2ld%X8pe{? -$_fAnq}EVCQ7g7;>58J>Sq9-B8vSDBkW)RmO??$sG*)Av#F)7y%B@2a!kKECNytx2R;_69L;a$859~-G97GH{ES!yj4xB6_Wh+|h0kug^OvvsoZZ^}^YA)-+G^z{wP(3?m}m7_dHJdr`-9)U{`AH3yXC#x=1lKBdo}qUzg7 -DAp^t6`&95KGdrRu^qd*0``>m4HVRxJA*ABi%zj^cxMqnFvzj^uWb$Ki!?n;lC`2+B(%@id+p80tzqy -G|6AbI_e-2Z);W#~UE@7|ww_3LdmYtg^&Q%+0lZ^G}NK7acB(M@iyAH1*TSLpPjXyC(SE?Idvpe%vPDSI=HN{{sm!pS}Ce{|``00|XQR000O8%K%wh+ljQdZw&wd -FDC#18~^|SaA|NaUv_0~WN&gWaCv8KWo~qHFLiWjY%Xwly&7qA8b|g!zoH|jc4lN4kaBF6SW?+KV$~{ -=%a)Sem1Jm!1|}M&ad(e^wY|T6-+SG20aEP56_PSTAMd_LTU%QV2A|n-JQb3*OBIhNrppfVN4DGPeBavb?0&}vOJ*j5y>d915q#QYS_t<1>hk5q>o=hVlf;oowd -k@JZ;tH#+3Df&`I}=ll8Oz5;c)^uw}9qGJe^5pm=cYVl2aB-=2kAn!kk|Gs`zXsRHM<*GoZOQ<7U!Rg -6+YLFp|?*oCsfe*X^4dcIyR=p^VIOCVGjKciC*iNzWr1pM7<7ad`RfSI0N{sZT5Hnz#P@y!D^nt$XOS -{&Bl)xBP=HgM;+Uo-jE>squuxDSLVO^VxTg0DxG`q!`OER^7Z-1igNEVb4SYTG!>0#GpTB -VkHsTH%i)qpuiYVED^G0t!yE%XZ#KjFFBftTuD@wT5>=6Gwm*zxf>Cpd$Cz&`_N~)ksRDlJ&=CKw-7L -S-r1(PGPFsr@NKG`rHrk-J=I2|(1rrbo6I2|)W5|SM`gcqb5!r>n;B0)b(Z1jCOi8T}XCY#4e!UlpVO -9>b`hXt!+Hd8VZS|hTOns`KTkWw@@21b34tmWi$HO|mO7~n~zVUbiyL?$e#Xbi=OjfLq&6A8Z317yBM -sGEp6~AW9002-DFsXR!rx?^`cpph0d1j<06q=tLL>o8Ql -r%AK+5C+jaC)8s<_l-v5D_RvTTA%Kt1q(|4}J|eg+waP9z~RV7o>VtQSCm0Hu*phG6d^U(zgsbu@M{g -L`13aEoilL=%N7)leva#CU`rvJS8`Cx`@74j!4rX=(I2+-!ugwh79j%?+<&kvX${hA^Fdf=|mR=1Ar| -HWP4uM~e+>g?;ubO79Z|F}WWfAhss!JE0*RPIDM_QVlt5=+Ym`2{??0xtS5WnvAF4zf_s%18oon`~96 -h2*2N73T+>Z+f0YngZeVX4df71or55UEnsx))l}N=QN9Vx!~M~y-zQi;MLQrC9i;^nLuz$~68st8i#2 -l3!Rh^6`)IjA#hYpdjw>SBd@)YV`aOi^P6Xv`a@}v7$D#}e&KEWmZ`Tc1vHz-L3QwMD)C<;fj}VzY -p}?E9I$H8d`w){W&a9!l@Sok;j;qZbIrP#UAz5i{B+?BfC&Xr!;m*wb|7Mb&*RQmsPGTizfsFI=>oXA -xl)C|F2UYNs9(}Nmoj@?+?1lm;N=xYE%RQDy0WDDqUP}D=qn5%iS_4M!spd6WyZirpC>VE3)n4S4&a8 -(C$SNF#v?(&wX9q!X(CaO2cYs}NHReHxo~E(Spvk`Lq*y^^P_|H_zUd+f#3_b*215q5OzZ1!w)t_U%K -fd9ZgGUunq%v@nG;mg5qjs=$~U?J@F;AF42#+vKopq>D!U* -FOXq?D==EM&^NAtIXjffXkn(0)Z$fDG62Q{5U<06)YY6#ADfw#A;-jt0fix@hq;0F%0_R*IghC;?&+C -JurV#+2WQ9*3TCskuQ-R5~Pk8||B$ID)Qw0cK4{!RPX>1bl-VhY9M9L~1 -bl-zyNA_6Bs0}R_QXEW$<28Y7Kc -+D!I)d>-lrN-bXD=g_+2qU%i;CQcuKygH_wQP170di^pQ&s8FDh!-#+Hq*vAi!AS-#RB{+ALrm_e}CVG&e~uopRufvTD!H%hWF&FcBWN(5T|XC-UC38K#Co|XC -M9DoA72B9C)QlisI)}fLbWK;kRELW=my*<>OGx((2Q}w!bu3FfC=YMB9R}Nc~POvkKz$HZax=H^wd!Td1(y`mxe59U&*6ys%Qpy9NH9RypRHgS=P8i`N4-@W!OlX08)<(0gWGSIv?{;t3^J -VP=)GVkr!$h&Dk<;>m+w0cO?P{1mMgCU`U-E_3R>>}oN#xb(<;Z*cp4LD5H=z%P`XQPQKMc6}0InDW2 -Vc8kW-%o#X$NxYW@rt?Cz%BK$O0?cjWiXJt43o{4AI>B;iR#eGu(|LMet4b=)++*aS6i596+XE_bxOb -bW&fDZ8&q&a$E3%6EJULrU`p>_2TUGkZrZv?O(q+Y`2dtkJ!I(B+ejTvo|!CYPXNix4fDQOhFfGllok -mWw`;wPf@CMiLPllm&6{MhGQkOnZHw;&{egL0(C>`kR<23+l7Tz)##LitoAl(=mFJG^hCZi6pMEP+lL -AD3?#E0VJ-|uT0F87~ex#H@dXhBsg`ZV99HLQE>YnCd=>+Vjc)48o%NPg#GM -lkN&uksBmu3F51u`xGd#$)A2rNVxqnvkkBgl`8eo}Rih;$5JW9T$+1H3wAJKwWGY;?0~ -TlA)4F5O;UJKFsL0zlVoO^lg!+wJ*$9y+o@smARAu-=X)oo~X~a8$+XcNW9 -HM;&OI{0IsMw$llkv>0{V;@?sqUNvUP29+U~NAWQB0*ckoy{Q_8Amt>Stki6L0Hm=Kg8`>pEAP1M%>b*=Y#3Lp -)o?fFaf;>E@HZ7RhMswF4ALrx^RD^mdH$PCnbtpsd0~2!U@Uy&lmb+s4&xtA08S%s>#5L!A0s9#4oQ$ -c60c@}%{{kG2eRPCeq#oTDaF=vryNxbA!zVn9l_(TXqt2!-8HBB6B=65I$=14_CekPYd6GI4%r&i35j5GWz>@g-^g(RnZBo73XLxm2l^7>$-Ic-n@0ts^vYvaSmjF -qeq0$6!xZRlImWDQd_QmzLQr|ew*y~331+I6#FnJLKsI9*3Tm}Cpbz~E7i6rHO7h$stAR(V+W&^-P>> -XCV7xJsi$Vogx{124b!E65AZBZ+_ZgBPHjfIO~}B{Dh4MwmVz&JRTq^TQ&v)A`lljz7aul0i!nH2x1z -O9KQH000080LuVbTQuD)n71DQ0LYyH04M+e0B~t=FJE?LZe(wAFK~HhZDnqBb1!CTY;Mt -BUtcb8dBt6AciXtJ{+?gKrXM74U0vIG*~EwLIoZ6eyH1?iPB(3mS`YmTQ^U)nt)owJa-r@qftYmcII;m^Yb>k|M7~ny2|PGM -FeUiVw{uS~Pi5r$t^xBA-Wdxe!fOYh+vZ#ME<`6q}neU9RfG=yCToI@sI0|NDczgL~2JCaPC5dew3k) -l#f?qe{x?(d&!vPtVRLrkX5GWM0X`=)3bL=Kjg?XD -(#_=Zlg#HuIoYb9E?)+30htim_%o?Zp -u7*S>!TG7g1gSEyd>Kc5AljBI+xQfTL%}Cl^o8&P~u>wg?~O@Kc!sCYlvg!rnzU?7cw+mkBYZPnOM)}BDy||u3Ormyc!y5Z_Dyp)X -7RQtSL+s+YqzLv8Ia{1_LW0E`Y}aT!SY1(s*aL{Ry|mJJ==k{pdGGnyBp!Mtm&dTU|&)-V-fZLSy+1l -W#TX5Xisp$+vRC#m?mP6~&CqnQ%kwyobioHVlD&E-%E=gy6CH(_7H@&`D*wqiT7DesC_MS}b++NK7x$ -c6~)+I)RG1Xp+bcINzbUJXKYw8m7@y?R=vYiWgl -~tE~vr>6q({N}||^Yy#e&!@81yK+#{cV_d2bb$L0iWVYaOP<>2uOa@DC3+e_rviinxTWB{LbqXI1cvky)I| -BWTpuEzK`m4oIUqCbJk5bAA*`i`+0u_Uc1MyN$z#cYfCdcLI-E92#pLgEPMicNwaHH|)zekgwiZma;C -B4T)R*BfiG1Rl89T&bR+cGw`P=B|G_d7c~y#h`2X_b=6YYB-7m~w;_{aT)2_yIQ)tJk(e8R|8o0J?ggC! -h&x&jyKs+vTaE)3;%9VtF(gb2er4dzBH4uTpPv5;%L}q*D^=Z-?t_(diPFbo0S+*n>d79ulLmF&TJpW4gky` -p0M&3;J^kf#d__haK{GBW#%A(0e5Y{C+HkDRSGsV8UWn00iT01+_wSF@)qD*#uAtkMPa}L8;}+umLA% -G4`8-_EeRW75+{nlfSC^|-DbY!IgKE!UEl!u+mVOT!dMAGa2p -svZva2m92#l?)t&DUd#{l-W#_!xZlAk(NFS0qzmNswh7AOkbsDT;hvTy@a{0cbz -1Yi9@qpuH%f&}>{X(9{7_=n0W^19OTbC)sTB{m?=!X4EFIAFr!F=D0 -O^6Q%q87l&xycCF9tEQv^*noZ3v3my`2M6~ -_J2-5c0BGdnF@^vLOG{TnTYwprp{lBMJ#0(g()p#j1qW~`5P*&pI|?>a7(i!#KDwm90H -W4J1Ui(1hb($KFgSWs3QdTD;1oCpQ=l#V(M1CbeV0*iV30rTQPcS7(i5gYSo%&=pcqVnxG1#QI?bDE& -@}c5OW@8M-$)UXKAKf#w0axxKr$^kAXo;XR!0=GDcm8Q;=Ew)W&`NZ5-9iqcS$=q!X5#TKqji2wYR13 -6165U0QdHBcS%djk4@_BMeh@z{$t?^$Jifef>m7AsVBh_)HF7wOuQ7n6*N>P0lmu~IndaW3n&vU}2S4nwk8+NV?TVBQ`V-P@a!9%$~cBnv=iu#&lJ#DQ&z#-*Ci{n! -9Hcb}Gi)7WQO<#H)K!!r1f0GtR6pp!W(Z%ZE>&?kOn5I{%lxr+t{&}S&Qiv|YJfnS;GA-FZuVhsnVEw -un^rUe7&^SBN$g_`QXIXBPZyych|X(359Gk$9*DGb;q)L(fsb|_7)*hfMiFwLhdf_$U@!$@- -i0F%uqhCaE2_*f^1}e?aRnb7+YUs1+KE6EBu#XiKqj!iPy2_$KT_Ny6q3RP!4!yE(|l3i_+;3Z6cCu3 -IUs#p;Y45np^#*d+&AgqaNz!ZIzemZ1PBHLh_$YyC~HOn1L*i+Bj(ZzAZitv-1uf4_V3X2XT7O!d}AP -*8Rue_d7j4sfIbWcnNvZ>`4~X7w8(_F;qKp`QW?;smTjws2W10OlM}uS(g`Adf@<;9{wqFXURSZs_!dWe -`rlT2x7sZ4sCP0m!CgV=OCF(s{m(dcS!%dIL+ur*EWdt_1^VCIB)dl7<=#IB)=(vXps(0w4z5<$$?dm -i8nq)>3DDKbse{x>vV`-qDIb%d+^xupam4-%6s#7%azuupq>I;`+)(#DYi)QRc=$7(zTVUKCB~mu0a> -A?YWVFc?DOa9pL=0a+FY6e2&#+$+n1*j_AE#PCCiZN$pa0KMgW=hqnH+Fod@L^B`l?$>Z;XROsWa_&5lmNmTsDFk`g!e+WZJx*4B@8KdIA2rHQy|I09Aq|b -~`!w}Me#IM2-+AeusEsXy5cN2PA)No`pbzsxx^6aFE8B5C4#U#`K3F##1Ja}sXM -vE5b7_l+{q<|P=EQ8JGsOV>Mv*Rwc2=$i_?&K0fsKI3JSt{mVypa)}|-UrKj!i6PWqDtB^;AZ$*lb|;q@Lj9$2CzlvP{pFK8xx^6aFIVp55<{r -JT)UG?459vV<4!Iyg!;>8cXEj#)L-<{;^Em*$J*jpEiU%VjyhHs$9j3Op9)p7USRBpP$jqPizx_0sFF -X2WuZ!TBPem55V|HgJ`b8DeI^^f4%!%d%<(E{FzhkMtDwQq6&Uep&|ugz -3$?49Ct9kGk3!U5IGJSKG9K^KUMv^Rj06u7sm97MgiYK@;gdx{XtYf85EYidZx^rHKm?)h>BrWR?jF` -d_)}4|rOar63RN=MiDu-W!XBjsWnpuDCt+ErS)K)DVRL;aVOgkI+AFCY5VpcQ=1aUa3{#OG)jw$?jdr*lW+Bgj4a_F?=Y2%#_K$JT;xm%*?V-?6>q+ofdoAeAio?l -1_#G9|j~yTu>~%VDgA-%eRr3TH3VDlZI!cw>(ggqHCtH^y3hQ4kczYap^ -Lh{(GB=LhYcPaGezP-x$CByR0C7u%qA-Lix!)aN0K&Eh=%`Q>hS1I;^*^$4Y!MLaJp#N+ -W<9c5{q74Fben+Ktbg}HxLKMt_QOrd#P3$wdmz;JyS4lt2zxN}(J2jru!)F%$ABS&VHs?{XTTtYDml% -$<=YVUSnxOq;UkMY4J7dFN#g4ugbjJBk4`#Rd>UcSLlCxs08nx23{#}F2Ii+AqEeuXzT0Z-2ZWeVUGpXl5F2&4z?2l -WaG}46%d5Y^r~ge0UKlfQae?KAyi|t23p0Nkg!+I)E)|fJeeB5*%QG-g>9n{&)padVY8#F&`^}U-cW` -iZ00I=(!mh+P&jwe!4NihM>{GYBp$vov&4HUxbkGSslpaAJe|(O-|nk0=r43*g}4az$ZW3fG~6T8)-T -*{iK$Q*I}JhD1`BZ!^h&zLLVOy6uuT@?!j&i4RH!4HeHOeB=9PtGvcq;U_TqsXgQ&0-spm~7?O`t)zb -cEsNSwU|{aEj~;EiKc$!We&r@zpbwJ?MxiYnz@mh34n?#0-vfNl(iP%7Qd3xibH)(bJ(d%^txf^EJKZ -+Bl9VvKFS5O4Qi7=}m3+;2E(>sh`T))hLy}l7J3&3|KOdy4`cRW@J-yY7j!e{I{{dHn)bANF$_~- -n=!P&U=Z1OkTD?7<_Y;xK*zF^PI6rovvHyiPK)W>4TS6&f -x+HnLKj5fG+An^Ns);qQ{z+_>4(!fRtRTHd&JBCsJe!l*qhe2nppE#huhi=jBix3-AQz7{@N*m`8K7M -WvQjk1e7RLSeW%aaHD>?u!He3m(%y2j}!HOs5uqh~fuxGJ4<1bB*oO?T8~lX%B3>=jLsG_@~Uf|4mD& -FB1uGYsLIB+OHYlvO%7*#)gn#ZTls&3HeC8KaO@lP^WttTWL#hY5QJa~j90V(r@^V{h*cdl@X3SD9mD -4Kzc-7Je2Rr;i~==I`}o1&{kSgxc5&bnK;ahjrqAs~XpuA!L^9`945gwu~XvU;Z{LGVh4*+k1@4%*h# -|GN-NoHyMPusCcuBeKvQYs01KgXYq4)BwD__T#Jp`X{SFq4G?rIp53TDbq(%W=cxVs>zwAO7;8^+?2j -%+=Q*nTNP73@Z7Md%0m~w;$X%8K4KzpJ|QmpR#))e0(|umhwUNPk3 -}T|=Nx`8CDj3=QC3{ZG69S_W3bt5`X7ndq_sYKSe+%f-i@wSX|mFFt)<8-&|+EDpuSq2!n=j4102=mu -3mh=VIp#MKwx{ip`0uyN_*-Q!~THNSCEo*N3^PQC)@K4BQY17T5+PQGpD8&3R^;}++=duq;nZMv=#W7 -My|sQ;f^{V5o%1hl3Yg{YWYUf>DGebn%03i>rnmd6*zR#VOPnoe6Xr+QI9LC`(NOShFd6acCYpo{H)FLz0haw{{VpzC;#MMe>JHh`(*LS;yb!@RH$OrCDFrZ~7zKQ5GY7Ns4C9-&IrP0`*~ -A*mK>XuE`KsEj_lwAHUb`L-od2wgfLXq -Mzk`m9?lcS7w&(=#$Ms>j1B-kPyKO}wwwW_UNEri4F) -bnl}Y)8DX>24##x-|qF$YG=!FPOV_{>*p`6BMnJw>&KP2cwF?)X_d}0`Qs`Dr`d=kG-{~D4B~J*KcT# -A{$HvGQ%LO)pG3aoTvaul{_;;q>c@>ayuGdM#qpCTCr_g%rx!aDTz9k(5G-BQDl0-&o9?%*%^ -DXxQ#Ev>f~nH`ppx`K>^kIE+PpfY?nbkwwsNJGtaid)lbYTy|C&zQ`Z^(_v*YLAU;O^n>G8{p8~NgRUv)yURVJ)+d3N@4(-;WnEF7neP+9iuz -Dx==5*Dp40OHj46QJp&x%mc3(`q)jNAD9!A-Er$A)>4U9VbdLBdo5nx5X}mf -TmSy?dg^HTYWO0(4@LmQ4v*1C~N3R;}@VQPwV?{=smG8lVy>Q~bNUDn$LB~+SYz<UEPsJ0Wk*Me{W?Uq-y$P=l?0J<~X(RpnKDJ5LliH?q< -ecmxv-6Vt7kfET8e?dCyf+|v)Y^P4p?`VWaASRr&(<=~B$grio4?1~ZjYLe&4L4om2NjBbcNC4_zi*; -(N8c?~5?IZ-*%u8wcH!u$Bq#_LjD5+Y%Y}RH>S2E2B+W0IWhY$zQD}mOL+J&|wdZMa)e|?2D#^u3V=hrYXl#)M|1oRc{)hRBH!6e`s^Kd0A{sie?6 -<(`gR*Jxb4t#Uhf>b>Ha9b?yd~+GImlw-z_(Nyc_60`s-dnH-v-P-7X$@ABGYl`v9BujSm#dafa_FcX -t)>Z2$1@=*=Js7gyfq=o_{DVwGaO?TSatxi+id1w(WwMqvwmi@`WHe*qLf1T;S!R}L;(H`o047Y7{Z3 -t2C?b;oWx)Q}YRWM^>GUzu!PTzW~${xN4DxUDs0!AZ440__e>aXk;^KD19y`-x=f;$jfzAvu8w(=X?N -cui%sSfw0f#$c-nOql;TTiE|s%e`*S0+BUPw~LccUwo$mm#B$^eDNSSe2=F`KqnISDP%YwYS?oKGwzF -?!9kYrxUtuK600Z&s2$`b_bB-{avkVjo5K?!EIfsmy)`h+hn@9rFftGRHPjVyU*J`+x-ocbabhnIHsr -1UdVDe6wCCzN=@!pGHGfx_-WfkKkccz;TxLpw_<><$F#TdYGF+7Kkmq1Xa43zt@^T?9WQdp?sZLL-TZ -gv&M-0Fb~d@+9ewe%H;}NzXeS62yY4>HSP91L+SCVmYt^JB3J5^?|6(q~+}*nM?D*v3>Dl@AIH!^7;c -Br=7dN9rUCU#*a3Pk)Tucr(Y9FEN$IH4DdqtrwxP4gFR_1zQ*%s9kpCE&p7S*K(o*=Zv`*D9;^}wuWT -iisAjtltUY@{ut{c)|Fw=G9Ky)D@6aWAK2ms3fSz9&QKakD_000RU001ih -003}la4%nWWo~3|axZXsXKiI}baO9eZ*py6baZ8Mb1z?CX>MtBUtcb8dBs>uZ`?KzzUx=8o5O0;t>7A -H4-5FH~*z1^_0?v))M9sT;#@zL=wY`bUnUJL@P;JY<5La@`1*S~+byh -@yyO6Ectv0-nneslY8|2R8;e|3&WvO-ucD>!ojH;Ck}){SM2=-$lUbgy<_i~dKcK-v60DjU6H+9Wl%_ -ss8Yxe -oX>2?gVFfA;q&d&Q=O^s(iLqBKN7aQvD3HO^ny6Gj7DZ7ZN|u%?C65Ucmuk)KD!pB8NL6tBrLb+I*yn -ozh-wH{rmgErih`kDBprOtya?vcVco5hg~z`vNip-pd^j{jsraf8gau-u3w*Si|T#~x~lZM -ZTOw<0FxSG};tY3~IkL=9Yyz;DhS6Ht01js0Tcoi5r+kkgj) -td*_@Xo1&UYdp{tl4|5`u}g19*E0ZcM%NM0SDv|I0*O?PdHdNuMC_5&;!~Ak62GQ2ir}dbpyN!lAsgd --CO1m#_<7qI8Z6kaO)H$iws8cQ0*=8ioJic0Xs?KU1Sc4*B(ZthM@Q)dMzsPg)~Td!w?OaV+fn%fmf~ -Q)%4gzx|^SGdw5Q6zX@aa5DvtBbURpYFGW=x+R&eJ+pCub4#0Eg^Hr#;licrlHxOTN8k#BQmE16bZkKpP4CrRiVxK#QQn-c>Pdd{(t@qEPci##&xl`u$0)LrOL-9 -zvWeS`7xBCmwE!9bb4kw11EJO@=IPKkNVUcc70m0ATZ47|iR6gg!qSGiysj)%4Yy}rqki%w()$$Cqx5 -Kr8)dr&YV=@}|2t!qKNvQ%FZb`2RSz@%Y!5{e`xuFNZxF%SmaUI8mPE_)FDB_X^UE{$Lt@g?76quQE` -ifZOi7M%B4Lc4&<4w|73GbnVqY!>*=f~&6XkQW`Q2ZuNfN7}^YVx$WlN>CsmK02{KG*63B1p)E#;4AO -i=L&h;6~-X&@#o3ntTk5edfzR+4FHzSdAbJ$s(@u0dmZ#%-RK8d_`ea%_i4gCG4ZYT50Ze3yE$k*n -?>HEa#0V+p5C{mbkhF_N~ZyYy2g!X6&F-fm(`+{@qH*J9r;ZwP}^YUNXxBEXJ!WV_FT8=z53XC%NAl0 -53tkari3`p4b}*#_^xpGto2}u8+jS~7Kz9I -qBgKPo7$tm7fYqTK+|o-3lkC)s)~K`bFb0DNS@9!Dadyjm`0fb7T408h)|CX+mcjnqVin~~ -Jpzy_z7ill-rXCI&r&})a7@aUZOzti*>BrE5LT-26*xH?DZp{D?Y-$<0poGY{pKm7Iua?+8@%FWr_hV2ugScKFrwhtOrON*Ag6o2v+ -g?>yHD^n1iE~*u#P@@gZwJFx5ZhWW&b*`F8U_2x|HLr!eao|f_FPjm%>R$2a$Qjrxll-X3vo#8cT8W0l6Umu8s3N4Iz=Kta_*9v|6(( -I|Dg|0NJ}Gl>c>`?jg6>E2eq&T$4T!b#2GNeW8e8f?|Y{?d|bQv^@*-~j32NEJ<08}LkLa-{Suk~6j< -f9wSw}ozd;n~s&F3(%-$rg!ap;XjV8#b^MS-EB -~6fKkM$NlhpqNIsLo)KKL(CO9KQH000080LuVbTaJqT%pe5-08kJB04o3h0B~t=FJE?LZe(wAFK~HhZ -DnqBb1!CZa&2LBbY*gLFJEn8aByjEXD)Dg%~#uQ<2Dd|&sPj~ADk_+E?RW=p*JAT)r;I<8@)WOD2cM! -NF+;ANra;R-kG5!%Sjq1T_9+HfEUe>=Nt}a=(n(62UCflGEtTdETxuC1ertHfX|$=!GDdDpDSStRZ2g -$&3NZ69A3i356qz-;ubklyErWx9e_Bs#j?T*%yX*zm`Gx$G>Ye6vAQhT^nVFi;3+4(UnR?^S&|vC={mc8Zxp% -?w@<_>dw}$ru*Abawh|iIA~@v`S7{swonsK|06sM30Ox6_%Iz>$Xhld)7w2aTGETILcVD?jK;LYF6du -SCzJnzB(OaQ^P7(-WiMMOgKrRbWmtl{!%n%}O4tOY1XQiPVhIdMkr**P*+{3HkWm&+j6V?t)a&^<1z; -0S2S;;Cuq79iQU~EuEzvFAYb{Kg$n9m_Ah4*}6Q#@4}=x(!j<~@Jw{Ks!P`KGgjX9TRvNFKCh?FPdq4 -i!JQN9X0AHHsm9e`qhhC~@C7oSvUhWA(Dj0o2{RF;^ESuYO}b+2zXz>4e|Me%FU7g)@ -;V>VQs2;f5A1u(j77&l@JG^=ru9yd)9?#nvZn(G>n9n?A|tDK|W;3q1k>vFrsbcJT-T)6AnZii;cxME -|1!oG0cjnRiD_!$KPU;*hbxsk2)AL_W*1%`tc0o(MKu~mPw`a6Y+;SYmnO -0Jsd=#z#{DzP)VS3y`P|6}n@x+azaAs2+;l-#1r#=hMs8}v=O`0u^Bebxtyj}kxBD97mavPlef*>{?l -ZL|szhmsMS8lR8K8`pm#*Cq4G^QHQ#aAdz5kl*9mAYN1JpsD8N@Lwrc`e -F5l$vswKNiwexM=z9hmYIE3v1(;Ool^BCmWv;q&KB(N7Ot~ME5V|)B9C)k3%8`T6{R*BJZP7w-*?CH} -OVzB4_HFI78Aw>p3O&7w6vv?JD$>2=QonOXn1S5rh@a*jQ9Z$ri?~ao423a~QH^^H9j=t2mT!L%)G|U -oIKv?p~pT*AZdiYROyu=a}kO*ih!1kJUh&c0|pW%Y^&=e?!ZMO!#NKw+3}Sy0q$bgMY1?_&9gQIz`VV -l130>=o}f!-#h;RP)h>@6aWAK2ms3fSz7>tTrEcp002`j001Wd003}la4%nWWo~3|axZXsXKiI}baO9 -kWq4(Bb1z?CX>MtBUtcb8dF5JdkK4Er{_bDF?tQSF^QoI0t|(k=i(Hzt2jnh6l0yqD7P4$pYhqxtUcy0M~>;mYhyxU~CPb~i<}%k6qQDH^>K+N5=Ex5AIDawDsJCns*SR&|`tS<@3KkfAjW^bvwP0rK)6Da4Bt8s8SMh87<3JZBfYP4ia_QUaBe(C{(q{+@s9v>Gq| -qRb^QCtjvx1N!u5?tu}LU*-B%wja-Al$>iCqA74LxC(gtN*1`1k-fngE?MvEVx(N5~c0QQ$uF2POm9M -Wxb4SnpqL(tSx`}l2Z5aMInPgdBmRSZa&Zi_NK3=xkXQZ6zTtp9H2`t7i!cuWVYdWXj&;L!`VN17_n3 -g&Zvsnd1X?h{P7Ns&aaXOuaTPKgwmAKhTI7GuRRxn4m9_#lGGu7c$snvh8_hVU#TR(o$aJlD -HOMF4Q&QXHl__&86J~za5o;Xxo>0oOvC2*GgU7asW_ls@dN2tbr7b0^x);<-Oc1#XGa|_=+?1uzBsD| -lCCRyk!fHM1#DZsZ*36lHmc`WP;wN2o|DR92)lDzl0Vu{eMY9sH3D%~5QZLbUX|Mwf66b7){5)FK%PO -{+bF4{em;{b%F9kAcBJSnjT-%(zYg&0=F*#;evdJx&)P=BZUCK@n;Cv^R%fuZrFBQU?aeff&#>~Y|?K -t?D*>XuVNT64hz8S{y9}4lvbMi?$(uk#WPp3Hjn?3FAxJnj7P;v%?;KT$ue -Q#KJhjI1%mo+^)}PRi~&$lBjV(smCznR}Iwqg67rdj9nqCy_5XMlK=Z6;H>*+=?583`WC`X!RpOhNxD -pg{Y0VR7eF-1(S*=HJ2@=hM`Cynqcg)!7j!wKd@%DWS;SrD7C(7>rRIeejo`CoUa_3r^xW{PE`ix&d% -K0p-OIzA%j`>@ef-A-!O3&s<&uPs|G7;#1^>w*sAGh6b9}!9h@MzKAqz55T3n0CX5XvB*7$OxqYgsY+ -(^gUb!T06~LF5Kc~1kA6vK98AD}k4sqm`agOy7;ib9)`Gd5)oI~qYY&2-OBotOlW{A1)B>tulYnmKBF -BbCsda)eflym@4maKQ}z0=z|j>3E+B6$s9rYD}J;u>;Sy2>{jxmhg4lk~}fm4ibk!pOFnp~2=!G3~Y{ -ExS!WxRt+YKDv#kvro7F&oF+sYg+0X*(5WX?UxU2D7Z;L0!RJ$f6QwRk4}Dc^!SkreW_CWv^=nM?|XC -c=#j21?Ccgd%q#B%YsvZb!9M9yW#pHscn1&i`6l>jCEjh}N`PZqlnC;DE!UI`UDZ|UHPY4w2SwQfXG -2Zla#-6mkg>SY%@uMYli8>SOLenH^4zNRmNk{b2SJRP^O;Y4zYmak4p;5A1YMU_c`i0u`Km8E-yg(Yw -_=TeVz5XRd%6ekzo{H?Uf`mJaYZDmd8&~-$aDPTxJ6s7qp0w+VZe@z4>XH?kvI3)rjbQk_OMb_bqk(X -a-FwEIzmov&h24$|7aCl%kR{*RYizF;t^1MPa~H88?qELG!-&KDh8yh6RN$a8@(#!&iK{^Q6*|Dln1A -K_(X5NUpc+YU2UNcwI=wC(CC%M)ETkY`I-wUeuC}RQ&*AqbswuI>nC&3A4XYtx|UUByVGVhosou$p*s -=vq(~c?ZxU2fW5TnDcVw1E$?FJFiDl;!OK0R9tGm^)i+3|a3>$AXa`W4q8G2P4@u4I4(X&!GC061?IP -}M9&wYeqsq7tXecn;Kh|c@8BB@s>P4@#psiLaK`~A2?iL0tnltKcOwH2@4K5rU|NHB+{xi4YC`$%e}r -RaPM@d0bYF+JhI(h=Qw-gEu|Uk_wKGP^fhB!3E%#=bLenyWM<`in^42SYTu8zY~kxDIehfg)FA~$O%b4m|JMy -^JVXuFt>KnUz#B)9Na&WGiF0R71$DeacbAUdw|J18W8?t95LNAWV+UH^9a!~ExkKPE?gECHvqR+YNic -){#!%xBW<(-MAyWxsN)9nc#8oey2cd_t90)l4BC;4W4Z;I11ME~L0PTb<8o6q3rJ2Q1-}Sc^u~l}t(fHIX+DdUxJP=_5XM=t@Z|C_4wL-)7a-yn2M;r`T2NfEQ*u$2kS|hA^zcEA#zPHj<+lMy3! -{X_qX|RuuB=tm3)9(ayaXQF$AyEKv-Q;_d5u=KGYw_trc2{;v8ef+!k|{1CQeEZ^h=z%KOhXwOnHms) -%MKrfy{EEkN0e|j;1q5EVhccq)ASVN-n0NX>qgbq{B|QPbdc?ofx0fUnAf!4$6RT)&Jc=R6saUpwn76 -MGzz$&N(W6CXzeygn+`%1>sNg3BtPfzjFVQDjS~)yAz0}~0a+Yh5$x65b`eZ%A?a2YE -oOlh+zTS#{wTiXH8Dp?xrWZQn318jcTX~38Es0du=nOkBnNRAf{8AkiJ2t5SOT9_O=RFQlvgq~Az3mN}$*-@%ieIp= -@dZ7NPijttGf+6a;&k-YEO*iC`bA7%`uYMF&?KbH(UEb{bmPAOTUK1EF64|CRzwXuN5i`F7=#SoVelgVVIZX4-CgID&Q -pTUobqwegIyvAyJdX7HpW^gIUiub*PY5GopP6|N6EAmrs$@idMsFP-AH2;SpQzS#GCK&O(h$t=OX}Ec -Za>B~R(KV}K!^XI#W9O!g1PkcXjFGzK4Ph7tE&*3Yu_WDMO7JpV#Uzf4r)Tf-+T?4gWrUz2!#j$)?fr -q8yss4>HajB7^;B5Luiu06@m0|*ETw6Vgb(Qo#oi`#}~MzpRTy@x1p+t)zcRc-I=`{dL)KvTmwA5ID* -Ndka|xXejpG-HOWNfCUht9oC;1~S}y4Z*#3pbM01|8AOO5rrXz41+V#2+l6%$ky^1#ursf$i2%u$&Ex -RTmZ%KuQ?}*R2UbY?nbGagXmC`lkx9SoR!3qXeu2Z~{aVK)abh@uu{;dfhjw4YA#Oux2CdN>etN`_?m -O2X^(-a6inf4>o`EaCrJ{ve{zhDF>SjK1WH}(CIXkRFJ+S=mxQA=1N8Xh4+eYa7c#m9?@^Ck<`Bl)@( -$+g;v?&HcrHO76tP+0d_*3pkF2lQ~Jf+Wa-q@6aWAK2ms3fSzCw9=Kmr;001H{0RSfe003}la4%nWWo~3|axZXsXKiI}baO9kWq4(Bb1z? -QVQ_G1Zf7oVd98ihaw9j==KVYcI;l`4QN6KmvG; -qm@3)Vz0P=tYifx1Jzav^qB#`(LC=?1v0Es_NPKUm!lD7Tza -04jS<%4R+FRa(}oIGxsAQdLR!Yu)s^RWvnJ`ORyr@_HqzoC93bmcyW_$zWky(YKK`Vnv -|gS!9vL=98i!?lw)r(L@BZ)Qu!>gI@ -=$u2L$!c-@ilR<=0<-$NFYVY=)U4BJtlslYvm9j8vmbH+lcu=2n>td{Hv-qGa@`n)UbyKKb(4k*HNAqP>G&2t~!JRs3N*u{RsA*e@KL|8T`akA-?G{xxzM4b5P -UP9RW?udMa+HIL1(uG1tSk!CKoTaBU*2}4@U%!^cJ&d713;>CysSMLpXinfzyXGl -&=%$kbBoGGBQdG9?X;h|u=pu`h*tC#8QU-vSn&KfuH>x|8MuI39k}b|JFDFS?HC>(o5Ja?j<%aAy6Ei -$b6xpwtN|4Iafc{n8Q!sn^-FM&o-4D6wPT@FO&IHt%1yiU94wzqnYB2}@j6J165C^D~RY2 -Nk4%5d^ox=D&!OdC}FO>B+T-$0e!oP2{-WjI_*6cBm0x)=W;F8l4Htz4AB{0+Tz5Zj;;bG$=*uG(IYx -TB}U?WRbfw;@r4b|DjRZQ(i8ffG6{hIO8(!}~ic2JlD{>aY&0cIcbS*x4!?f^f=jB@v}>3eh7C-v(u) -c!|1cdfh;eK~7kj(K)DRME6|Bn|qw(ZSKoys1S)brJf>ZEhdvZD`494gwqxKJ6+>%g}=4=C3OsI-<17 -H&Qs3L&d(B6MD{*QlhgodvB?<*{xn@KQ;b1u%?KX67HG1_=3OP-{m>$emLd#G((sMymidenIf6^db-w -IuA9!{S5TXDh{eQZNtfB|oXmonW^T7>yn0C1??jmz7mb$z)&!&0v -++$>Fp~(u?)(Y10rjcUnlpuF|EcvT=?u8$q9)fmdYV^G(^{lH<9C+4Vp-d2xV)CDg3>MpO3`D)J>b$T -}o95kzbzN6G3@=d6F-OqRfy^%S54xC*N)IE?2X!DOeV0fI=vnM`vFiM_J~tUlJ3jj~wf$d!m?_|M=$Y -=gX^$@wa$41H_Q0i#jsYNv>Dh3q2YwVkZJ#rB6`QR=)mBsK!2DsOy=+NUmw-*vi_IXiqVuOc?EiU2*# -K@YbVcM9b76cu2PLvPLoMsi=YGm!yZHq*b=B09tm2!cN01i+n#b+w{&d81va6;;wQWMtPm1O~5euYN* -$d@2NBZq99-rP385%epy|dQ=#i3Kf+6s6wrY%*vMMuT-{-bDM_e_VLLU7a;I7w7fhNq<6zgJMc~ -2Qk<#5ZQ|9I5*mIVRc1UYi~t!)*{wj9`uwZz^h)18!>`X9_tq3dLKj5=D(DN`j-8|If(sCGehRDN9yd -D|6rGUjbHKdV4W3Q7UQ9u%*>2k$w64zmt`?_W -!;}ja!ku#!74?{~dq@4ZyW&vspFh??92_(%>!_uVEXp(e(THU`|99{z;2jAVKx(Dlfk(GZTD98Qsg!v -9O)^cMF#X3D8@zX?W!-n$5@a|vB_TQD)1*My;6N8Xc!H}k?l69`byJgW=uzJp{w%8e#Q`QU}4;yh>5z -yHUGwp|3Cb%k;7VUSvuQ5t6{JZ5g8>&bne4?z4AT>5%BLzyh5GfLH2wSijf4(@4!!8R2PfJSxZmV2>; -sP$~#p36yzrm7NFV4(e7{tt`eqFrZ!Af -MD4Qa9>oq->imybbXv(t|7id6JDvtBnUvDuCf|Q|SXfjgxY?o=* -uyMV~Qy@=do;&oX@3;|k+un(Y~cTFE_U)IT6lU!&1A6j-=PptHRAXY72I~NjY-7jpv%e*@!m(Svhr~3{Oo#icDlGcpS*`v)^u4dg&mG*2bIM(hctuz4Cfd0pfc|{aezt%90tmt%h --xUQ!0o67LtS}gM>#sMqK6fx~TKW4L-t=$KZrai*HX6lgnXXDBcbd4BAQU4kpa~`1XmY;E^FG;p#x95#n1#i& -TOd+qKv=a!zbe2aW*8Lj+aNB@ -Sxj*WYc}uOvnU7-k#f5c!L=DSo64=AGXlcV1K6`+R+YFlv -vwMb`~+nm*;rs(__PD+RR*6u;aqM4cP57_dwt?M9+M(K3K$2?BEg!UB$i%9Genzex&EKo~4b(E@ggEh -J0JV5C_#+acC2PhUomK^&qrx#iASlVE);x2Zpaap{9biV>k^hwLKo^^@N-;D}FJxm7~ha9kWF->d5A! -`3XD@^zQztEQ`xSbLJ#!F~Sj?A?{g0CYClt!#)CUE)(bKf4N^qDgxspaXM24VvwA0(_3`Gjitt@Si6N -I!&_tr^pewSuuNnue+5(O^Erx-}VPA_d^-B-e$GR4{IXaRRgY*M|8sw?Zk#>A%mL^ghb=c&Hp!@3{7o60tJ5>G1vv}oix;RMG#MS_XMqo6 -LquxR8V(UD8BEb)n(ojlwISjhe%?|6=u3cQ>)W6mzIYKBR$sg@rd3(OqwJP#>sv0puqieM?Z-IOt?0oO#X*JS|Jq=?jrFNF(G*)+#mz6H>8+sB -S4#BmUmLr|g(>B<+5c`2C$*S&nKeE|jN6;YPpj9)}*+r8M1|(7gD`OR~)P%%A5zCupjz -_FiWzH&_g&%L5=AE&Cs=kX{t{Z8>@G#&NRpLFpI?t5dP|iqTW(9s27=Kf(q)z*2cMIf|Kn_>>+ -=wwu-sCtZnvEzlNK()E#6fDs@FGp54e7?xugd5ArsCrl4NU*Kxn0Ou6bfm`~3t$f({pSr<2t9Sm~Zt} -?EE_Oe#g|$Kej3l8x)0Yy&Ip$|O9K~e}t%=UQ)Kb_l=h&)Z~G9n7U>?|KlWq{9$UL5clzOb6P4N -?f5`N*071->|De3W^kqinRU8*&JG&&cK?=~6Y1->{YyE4!n?%1!3DDNQTU;ubU1pr!3d{}EGK2OYcY46Dc&xY5Z58F3eUF)#9E -k<{l+-FLA=IOB;RDN8)T1Q|>)hY9aK%<^RlJeU~+`320ukiSQhMKg419{CN}s|n`dX`=Zj3ixiVS%!r -cWgGi3bkkXoLNp|&tvy~ab*UPA&JM%W%Otq7C2KqH`HyD37rDCco7l;oTOUL@Ss#|+8if}Q_I+}kQ{% -Bn^S;X?M|c^h2W3Ras3Ie?o{g-Aglg$k(CJED)`#}~*&1-yuJr5RgJC5nz4+ua9js{0Acnz)*mZHSm}P_j -5s2v)k7R6BQ&;5+6Cyn@#g`{r=29JC6fJhzc-HH=v|kjng+8(|7zwMLAtp*}#kWy2u^}HAUGA3hgCw4 -wL6oQBF}gXO+~cAImIR{EAGqmAk&yyLWNYztxVpR$RZ|vrh-iLLA>@2{vdP-95+J~eygHBF{{;3+o!) -}mpfqjn=|X@nkPe)VbdWV{4w(V@b=y~VN}TlNrjJin4?Pk)l@GnD--S0NSs@LMLWrJxs^O0?Iq9s&E&82H!il*%57Q;M -wEs_k^loVie3QvlB=k-WMPGe|^Y}4YwBuX5paU4Dro5(2~-^h17Qp$0bnlrJ( -p;MXDWY$yC)2oV9s;oex3e|B7dyZz5%7bWQefQ{;0MZtS(iQ@yEkc}D8jfD+#;y)uPh&BFG$CG0%b75 -2uJAI{15*W^A1t^D^@`YrZ<1QwW=L7ZfR|VN-7_R;Qt1|FcAR&s^lvX;%1SxPs`%>m3_!c3Xz70Oj2Y -1z912X+fDdl$L>X8hshAypR5wJ>VF2DzA)f{b`|2#ri+P+*6XTggb^y{RiKgwj#XyBmb_9>j!;pW@?m -x)E>R~tQE;CFN47gupj>Ty<$R(-M*5C`@hBt?DUAFh@GMEsmdJ1p!^)*{x=UdQ -@r}PiGJ5jQd`AHXat}z`(WQNMH1BOqjvtkhMoT_-ub^`?@cRww^phk!vBb@!b1SB7kv`>ZG$Fe`z8`|Wz$+*qBJm)ge-%ENjwKyr;$TQT7|J92nf`4#fzIO+#+S8+9bXAi)^^Ql~?E -1H<2Igl}(E}0A-aBMXvY0wIt|J&8Q&~)Fic-NN7l6tx7hm8xl%h?)-5)rWV+B9Xnj4uQ13UIFi#<&0k -GJ`mgKjVijpZUg?w;1Aa9EGW0L!UxYjDUl(d~!1ssy9V3S*u*z?KA(pXd`M+$Ez2~+eg;Pa$sPR1@?L -r_aP=*8H7+AzOyylSIcB`^+GcT4!SP`WIt}%scz^|LGk6edXnWPpA35^33<=%0Zw6}IKv4rlD`gO6(B -WLz2kRTmMm{rk?6);F6(C77ZROT+$qA|vlaJGK`&U@HaUw?}=@ON|)@tQcVDHT0`%0Q^Ko5%X0S~WY| -4FyuD6EPoGxWBDvTQJ@2jOPtfbAF2s(b_I>iDov)zRqL2O|u;rHjKSyQma8Uee<(mQOqS^*rATxeY4U -jE#Odty@MDl-f*!(vxklGOE>iTa~yf@cm;_8j7WorG@%W_8kkE!uy0Q12TiJl*BUMAji%R0E6sZ*D2; -pl$S=&BSZ~)pD1mV?aSPrWjb&>b@0cV$-o+tZ*gA9F6iEWKh*-0&Zk_p6EmQe@4aYdeuJqt9J7~2Xy> -Dqru;Q&t<%XJ%Irny5Lba1O-aBpt{87o9MeP3VwStY=Mgq_+gKB~UF6&sgUbxcw=5&Ei89cfvb}EO1+ -J#@1M6N2lNuxbItxZx(+lM(z$=lP~IeRiC60V;Z<=DM8Z*ZfGK_PN_50+jt*>Ox0x_8!Kn`Y0q%L({= -E9$FjHuoH@vfCOYw*`H|ssKr5v~`Fc6z(2!TJ`+_D((7J)uTpMaB`mX3GBhuvWL+tzJuOSa -lJ)c8ir|%~3?u+Ox?20VCC2XAPA^L)3Yi0;gr7J$E9QiA_pF@7WxJT91vJH)&xX_4SbBOo1h15%19_YaAX7!t#rOjTtMX9WE -Z-v#;(!G!#X58!v&kQ4reAT7H-yA8Wv3^5^CtDyLgD5z+g8IQa>@g4q?>Cyib%nuj9pu;0dR3>gkvnr -W`S6$*>b|U85q;`tf5?1L~><@m)f#)Q$+;llExFL_xzCylPX!9d?KZD>5NT{8IKpp?=gJ2QtOV$=4h6 -Qbm89O=5g-AP2z2CU6<{ksgbW2SDl6j0H4;VMNyOI`sMx>YG_qZjs!w9_!3yE|hCxH*q)@j#dbj0%}? -lQ6=mM?qA=`{{HVNuC?JJ@Hf~zzHO_0-;bWMu_{AI>e@WUu2W)Tu_C0ZCHYxl3_Q1t_Be3$VQK<@t4D ->$jws*Kh;k!Gln|I8<=gZyA%ovtUmfbe#TvxPgNrqJOBXJNG4|o|Hi@-A(-sXQ0SHMYisk}TAmBKlTG -#k8>D1?#`&}-skQ>2cZA3B?UdqE%BdLkgnV7Bv*;9gJv=0;ybPk?L>~Vn&dis2u;30vH^z^w&!Y9(@- -V?D$PhyeyoXT4!4IZ;Mydtde?JaPW^c5!-%y@X_YkcTL2bkW4F(sWxaY;pZCsPESMQ~(AsN_||(X&}w5n@Vf0o;t#25Al(s5U -w=65qXYk4^`afg$RZYf4lRlFFNKP=xnwxBBhO0=?AjxYH)KhUmD7&6v-dnAwle+#g=pRN}^#UBDI1-t#Vgltb^ME83X^-rn6&9rt^Cv0ZfQqMm!l}7-H^Qat+vjChM1J4BU1%VPdH~5PEJGbI=|~UDXI(C1U9lD5+v^ft -dlka7mt8Gz;DzVa>oSc!E;Q6d3LC~RQ`+{G3YT6K`{l{y`MVq6UVNEF*QodiSt%?I)|Mvdu^Q5=Q%vM -Q*Rkv2#Tr(_FW0yhP7L#Fs*fOBZz~@J8Ll=?OCry~6m|dWsz}MPD#%XYL`(r0S^V!Ax(=665$~|s^e1 -p9y_{JW-@BPL(*f>ZKL_uxqMU!V<@_s^v&JQd;;+H&dY5&q7c#fK{qp7#k!-H%Y7Nt9Wf)wsh={RR)s -yLoxv8P)a6h}(&_D#*KrN=KmyrXYqM=^u5{zDjxF{#EtQCP}?XrpWmWq>MBY+^Vg}C6e^2L3Kb)N@lU -3%5T`puQ!kjvZiN!hIN_3hkt<)=JPH%V_z%^r4_0*|3pveX?x?A|Qr)`A29ngg_W7Jd#y3WAQkFvg(4 -MKSJ*m?HFqByWfL;ri8Lp2RN2Z>!uK@3X7%BnfCDV#~f6wxR3G9~M`Umr`uaNbb)d0SJ~s8M^?L2{;U -t9OGK~hf{Q|yiL~bQ@N#OHsej%FHTbDk+yRh6<86(uoQ#ft}jhywfAy!ROs2xu!SzFie0_H_Yg!-86= -U~v0ZN)KEmiL`WWRgK@4#Xp9pX<9~18h(gy#dCvw7>;b=N1y!(lxR?-cCNWt^GZQxm8>l_PgeZK;q^g -`$WKAD8!M9qw($8o;cr-eo$d+PkSEy<$7+24UcqYZak=K(>)+I`G6(sPOM$y%Z;#ueekg#0~)J$iyH$ -{Q}DZbJtPTP3E>La88C7!gqsF57Wc1aW(^A6G}@+wC2|aWxSb1DqwHP -_Y`?wS&y6ci|0<0e^IkIgWrOeiQj*In11~5WJJK;Qef)Nh_xK;u*!5F6&;9|IZ?|IE*3ndcZ@Kmdwf>ANLVj4hd5uL{uS9`3;JrqcJ$iwV3s1 -lVgldtu_d#~W+KkQy*NgM&8$sQ~q$O^rs+IwSU6&;N@uKPem9r|vg{u^SYLF=`<$<4|N|Rkiu3{ -K0QS(WB!nB(~MSB-DgANOsB00oo-7gNJ`?k3L!C-DBQO<_TU$wN-lB;K}BxD-@-nH9~K`F -$p*VTkbg0_VY3i^!D{ZzxGybmR^lM`P9byK=4)rJ87$d?W95R@`D57j-3s{Ixt8;h}! -ct{I1GmkRd+%b5L(o4+6VufP5M-&0@h|EfDOToJ?+hBew!rTzY7iDkxfB|mBu;TPTIY#3S%5X}icxCS?Zq!ChUP&L`0E51q~0dE+u^Oq8u0n>UpG1OtwZl> -NZ@^}|^REsMxGpwsd4H7ZoNd7lU{Sm6QXNga7-5#2g*MVC$x!~F6^MdBg~Ld0Yu`0dnFAaeRH75EEl; -AZXnq5tBQR;fchl%fpc2{)x8z;H^ -a6^jxR;IDjy@yFX4XDtE>0dfL=bjvrA<-;)O_wa$R(1O%`eK7GSgM%iD>OWYk$zbhD-9jO-eSbQFpQz -(r%)VKnO|vVnoEuCQ*!Kn_5FLIgA%?QkE-jy6hG{-lqDgc^NPMjlgk7i6Fq=>ar7#x+^%Fe1Rm#H_&( -*|qvprUARu!A*26ZGfK)<)IFHtVlMI&?53oQmDp2dJIyLA$Z)^=jPmbAAMg=Rn&RfCE&!u30tfxUAdq -M(W@#WT=6}w(`S)x@+z|G18qPQbx{~#9XtkVz&yskynYpnUD<|jBmmMfcv?^mWnZ*Or(w`RgJ5;J`Sb -_Oq_mbH`$h?c=8;!qS4FEpq4@$8YL$4?TU_3>@qQzC_wGFUEVI|nP!iyp2G2DclAnEe;W{UC@QaIgiv -^tUkJN!Yo>A&Ide#)}F0UWtw<2P^Mt&NefmeSP=T;_;RihTe@ITxe$*0k`q*TTAam%WqpNR$2J -M(+YXU$XWcbyNik{oKrfoa8S&lP$=;aI -NpM== -I*)!qsvNgnT<8395X%y}M=NMoieymyuT5(nuYT2b{$YH~#wk_Q#Uc0J*z`Egt=U2*0Sw#Uuy14zu~7J -9T^43k*Hu+))u{rbj!E~4eS;zkr=r~h9@`zr9pfa4gK-M*I6NGNa{>XahqbZeV@Qm^nw-5PVB%LVsZO -#SyBr-m4YYvNLW!dOrMQpW0xLdSTF(6c>aI)DiOFHYks3}jXHS|5vlCommzszaLYHmi_^Hwymz%s3HZ0j1Z~>Bx@srtOoJ~X6)gq$|Oa|o@RuWBEayh) -OFL;$G?p%Oq0)Ayp2s$iwS5QmQtn-tDNJ6o+{R-Xl0UudJ27tB8#;N&xc<(ALm`(c1;>R$#SjU%&@BF -V+$F$d6r-yu(@oo2sa}*?6*nmO`{H6qinR81P@u`&mAKkI@NNxVmED2_U--{tb*-t*1$aP(meK7POlP -LkPPY}CSz7sGz|`6?{@ZYcMaTC?{QJ9fU}?$)@b$qLLts)xV8&v)=sVzB&`%ymnf0+@;X(Br&;#xz{G -q0iljwy(x_85cwI6tPBhxb9FW2VNgD6&3H1AXpk4s2XpuB}6$Jrm#kRs9*3BB#-A-%}o0i$Vo%VtwS* -aF5Xqo~x+Exr*99S5bSXDsC2+k(PV^Rzf2OND7u?Yw)aJs|`=($=5#fuvLJFQz21@Er8yfE?@z)whV+ -z=!g`=8r<;mNlY(+kGWzLZwkfiw2aR{eTZ4;;bmYeGwJr+H+qv-c^7@T7+-sUVo)PQ(=!NwIoe^WgD< -eEwG5)ER5Y^?O9eNoYpg!c%!$89M7QO~i4w35(iy;-m)q-b^nQ~+9e#NZ3L(-Huz5ir%~5oi6Z_ejaigb!~nl|FT+Q{DHwcR)Vrg0Us=gdZM~{ZB!XLd&t5Cb4>wxf6jj$mZ -c3_cTB{3D8jYEna;_+_Prz7Djl$~-ipWM;7x2;**b)nO+JG-RVsbMquHVPLI@Z; -um4I9a%fgCfTI5~LBL^Vv4pu9K*;nnpoiRGaGP%5sY!N0gT#jbte#PA}mj5e6P#S6>9OXS6As|^sjx_ -s>ub}(@hTlG4k?{xb9qrAD`~-P>d02RMPC8&XMfD&^A!fUI7xWGO;Ka*jkEU&8EqzMY7cW9>El23fVG -(60(wfx%q~1nqta+?Q6t7!wivJL(z4er)5VN;|JB+cK(y7Y3G}?bDtZuzKC-l{ -rXCM$ENSnK8d}!(ygU{6e8K2Sn0(E6W0(6!`Ei$>LJzGs@jp_QQ!mV4Hv};3DQ8tBm0XlgS@! -in3L#*|_&C>#oZJNo(q_MJ~1(KuB`aDT1do%RmV`(R9AzB+0BW7TTF|+_Z&*f9 -o)7$HMz7UNd%z?LLys0t9;-0BBj$Ic0Wj)z^0FR@UU^u+4jPzpI;yt8R1aW|xw(!#>_XBrG)VtHP>VYcxd-(mHQ+!W -E$x#(W5m8|jZh5vyOjS+h0_OS<7s{I8Ns`Ke(^$*)YUS`s;W~jT`BY8N{T(J@sM>zbgkqc6X=1y9dE$ -J76GdR4PWkRBXcZf8iWu#?2mR=fs~!xJH(m0a3tn=uIyTf0HtoN@{mXa%{=_S{k;DIa?AdobF+ImK?` -`Vsz)8db)tDWHy2DDw9YcW` -l?;iQcGL5>%EIDE`qU|qVCO6zQ{I2$t|V`XfYR}g7sy%;yiEa*eSS92_W?|=3hJ0zYB=0yM9>5p8RU{ -bO=TSB!?grAp3KP4hc9W09E}#H?E0mD4Z-ptKLn~FvM>n-6GxY@ed{?JTgt=r^{oPM7~g7(|h;+Hlf$*QM^@;pXlK&jtBgdcgNfl>jV*92RX8Rf-7QmS>acG13%R_?n -U^9hMdG?9bHy?^gQm6hm-eCdKg7#&W>snO&hnd%fWX^64^7fCaHx)LMI1`N=^x^z6Stxqz4$y`;mK(|_J1E3~O%Z;;Z$xS0f`NSjLRhL=GP -BrhWk5@PfLVhm$x!l_)TX$;*}jsSqG~&A8MoXnyjH;h?;G6hJ0D@WP6hR>HV1u=X~? -dg7jMJ$o5`x_COnuoS?{xA^s#f09;P67;EBx!c^DTgQ2~P?hR7a(&M;^Jz_7p5ZQ?9G$TEJBGyiZH#& -GM(gf(nFv~8@5J#J4{2K4}vbO^n5&!6Yd(H^hFNXW=duBfMbZlAGBGb9(7*n4blF3xFUmZc(hNxL;H5 -SzR_!PzC`0;;?m4>VkKs1HPh>X=>6`7%jkPfYL8Yr=V;hA(?}XhMd39qrC(6brOE>@^Ej+P75NKBN0T -Nd4*YEktKc3g(SKvEwun4-Ql?e-$2O0Y1#)pY6{w`3QS~RiAVvUZdSU7p~dim*n}02(;vFF^2LsHevI --FcsQ&131XEA)c#zd9wUy(awH8{}FyKtq)NkPFR~?-rfGVn4g}%LhDKrj54`=`x>5j{Y8QIrTjK*qMz -GB!`9g^v99~2tI>W*Rmp6>MTo{4mPEt_svqj^7o^zLZ~N@I`fZ=bu6}7z2M?;98^k^lzl^_tw(wi0gjFnKq-c{4#_fUV7?=oFrIH8hu*OPGEhb$r`B1=(V8!m(*o!uwkMP>|cMv0Y?(f692+^|N!Fp9K^DN6F*L -MXZ2~KP_0x_?C2QvOblYJ$$j-!J@dKaDS2R#nZVeFal*c69uLE2X&&STs-?RKNzE}6Ob<9KW%9fHzz9 -vI@0#~%EryXHRjIEz>t*MgJYNKVihJNuL$4YMGRnN2Mx_qUNv7J7f!*2En -68Q1!YycvF=1RoDZkPMMp`3)!VHvx~(3zxGi&tmwNQI?XT|k3Q_bs6>1N3F%L&sJx1Ls)B;P6NPe5Ia -MpI5Hn)b;0?3cxclP}QLsNeXq+Xx}}#YWln@@H?EPv<;*q8Wl>P!seW>&4q6VTe7-*jR5OH&s^NvQ5P -<=KNUK!Cp8l(1$g9;QVTawwsk|ejl>KNK_kZjsb=WX*ufK~20)f3x5X(IB@z# -1{cvcxp81Q&P5Ui*o_Uy+V$VkhlMx0~ZmVBPXfxgA-W>`5!UP|4D8^%3eQuBCP7vY?lB$`rBfyDMN5( -``WwyAZ8f1gO-7=ZNaI=p#SBT>%$m+`_K`>cw}~1uPJT8Kzk_<{f!s7K0-3P{d*)J+bnDj;z@lSg3~x -jleI>z_?ChtXK3dcr80*gom6@UGCXyGmm2w*Qu_sWwWVI?%XAZjqhFIDkCFy#eej74qnYs%)Y$}Z-Lpy3LI^PLQw?wmLA+ -M98+Rzd+yV&Wx#7H#`#|KOrs@k46lbtRa$-3nej5wW>IJ*f9sdZlAJ@+;(emXMD_Dx=q(ECJpG<84nB -}WB;Rz52%{%83JwB&~rPv&FjaaYh+a|W|SiYWwp)Rc!yYmzV(s -{Oy&`^EAMsWN->Ni7I;(21cw+(oK8?=KvM;-Uabb-FB#plvd@wx;+%3fXOZ{AdR|D47-&~7-FlV?KwE -$?VCD3bl{_ON&vFI2FMT$L;;s|n#8XBQ( -YMe$QvliT!g3?$P!?&!(bZjE>b!^qrBrYGUU*_N$n!dVh?htG=tqIMEq$gSyKkZi6uk^6}ic)$d4OYG -eHL7kknM=@Nmn73c5i~9K3hH7djP3xOxLA)T)rK-gJ=4j~}((8;Wppr@)~v$oT?$;#9sM>F5RVFKJlB -8WD`RnR#0*v?C}m3as&0nO_1XOTxd_>fg*K9ddbeWU+a*i?qyj;4#ojfCig<2W?FgjBRMy$3fWP#ahF --wQbO~i?W1+33dtsXVkU~*P$1A>=86&l3IKuv;shp2E`+NYYf|etBUe5kMz*46Ba8}81RfN0s_g1_;q -`1yAZEl4`rffn33Z|VVn`N2r9yejGo(MQXjvJb{wTlQkxkP8WAWeC3VLl_af4Ng;-@XLYR9|iOh>CXf -~yHOnp%fU3C^;8rq#`TJLW=Z)7iu3%_naXtL -;!#}N<9qVC1L7je5wy>*a=8=y8exxVZUy=Dt2EcUJo0{f>PqnS*CroV;tUy+lT3bUJ3WEtA-7$7HV{DT2iS0WU|dNmCYCbibx -BwMBBYM&bY=HDjcvj@(+NOj5uALN|jt>zX11XS(ngQdzuHh&@?`JK7dayNSJ}RN1rui>Du9E{+t}nF~u@@&@ybw;ItoztE(b*X78`jU*yXosV9(&9tyx~p5;pbgV`>S7)!wu8~y_7MsR{!X_ -M$(s5J1IpOQgOS!rR@!;t{6N8JMOXS%QSRx}oE5xdHuOj4@^5=IAPS)xp8zPb}*yhnsAhJGl2DC%uw! -}I&ZpRQh5Sc_nmF0M6tqe9G8i}<{zfIA6he3&J -!tTUPphFPp^cjldu4Q9MR7Z!%7;gh7UIR`0^c -fBetEv_xb2jY$dyQBWg@?+vThp}>?CQ -ziA2vLn9OF+i%Hs=E-i(f9^5ZYRRp+!>eYQPkXW>@v!>j`1>S)L7?l!yHQk01%)0cWZn`H9fP_0jim6 -Nm(AMVcyPe+6U;~oli_uMvkFmPfVQt2-4Zj*!cTw~R)-S#6de0T3+ZRq0h7$L~ -J5f(}am_ynvtq`d>>m^C7S{Ko?U8@goz`TpzcG}~f(mUE~3Ji -XF>8MJROhsN$Th#rptAejQyF0yDOuj`BoryykLBwvZK!+(wiMiU~u&WKlMF>Vg7%d;ofo?$W$PH-!+i -pPn&<$wsy8*pRW6KGL4)0Pdr|_}b(|N@1(O*Tz2sW^qbzMsBe+Ugz=yb4&$B3X_V8KeAD^_9|59^zKT -eoezhko2CSR?73e>KSauR6QSlRF$g9WmO{(vW}dhW3u8%FpFGi57|8EtKF>qb8jvrGyFR=689g3)4lm -U$~hRtZ+thVnioZl6DQEK|FD$QlRhMJJpwAc|Ad?UX>=1gId`IgB&RZ^cP-eLZlgSPs2*7;s@*jERWc -G4t(JUCrs~!?vhm#`#^kbkR)&pO?x|e*`%)i@xBmd1uyh!&r8l33SA?2)wvBXclifY&+d0HF9k1f^t6 -pG8Ib4$Z#kF(Lnth?Io4HHT|Sdm;?uSBhOnWFgu9I%w5W@-$QQ?^&g^;`Ndlx@LE~t`4wq%`HeK`(oe -Y8?11LyR_A(eeW{P6uBFoBWbK%Q(hd%7yh~hk^7k6$u+SaVOnaU=84(^A!DM^3X2B1$FhkD;Sbl2s1-yD8`1?2#kXnK|$XTKpqANqph;N$lU_FBIS;OJK7e?IIo0(&kSN1Yv}s -uoois^1Q9?)a{tGWVn-b?@SP5Niu*YiIov&=wF!uFA0hWFa$KQ -PJqyd;8s%_J}yN&iJhN}3bEPh#4OKd;#?#f$gjB{fsZwxmW!&Siq-hkNk59P-3YezwyB -d5Ra)9VsWyo?5qZzq|(%HqXkMt%fs-PHMdB7t?@6`LD2bk&PvOKmtKD7LE{K1o=TA}x00c}|ZmO(G1ATIE%2%~{PLVc1+66asWe#$vgbRM6M}2yFnwio2%uCC9s7M6{Wj8?mya4C -A7}P&}Ehe%ok@$`5{n;Vi10T63F^l-P3~F0508b|i;9_r}c}%*x05ijTyoKgbSV$hzV5FlN{`DYQYEc -*Q4qALb6jcPQTv!B0``Sg6)WnRR%$mm#?-z*HXPBozq)`q^6K4nA&)IWWtdPCW%O)*8@A@&y!;M@Y -nd849VPvSq5gK^lKCY<~G~cdlFYP}+Ondx&Tpsjpf3d?mx%L)bPw(4UYvLoOhO#ql%1!|)doV&02_*E -Pt8a4MO{}=_g4^R5)n8$Izln8?O^EROCnN}GL$?EH%5U(HjPYVORLYIVMhGy)hRR2{0Lc7y_ANZgcSx -+6_p75Q_E#6n1hPE6z0H~L?HdkG3*Uez@5mNXv`_*gKYC|0rF2X#u5w-vE@K -LHE#79p2??9S4sTz?@{UM;PKb^1%bJas{hl)%jEDdA17Pvij9#`&oE^70~b$f1GpMh`7DxkPvO#3d{d -bSUfClcG;+MF?-^KTj5$BrEE*Zytm+gtzw3k7AgSH+muy`CWRlX)QWY8VpH-lR@$WKs*L5?u$O!bncA -1BnQAQFs*dAmRL(}IARUVyT3HQ+{GTqd1{_|IDYd>)0NqzSxtUPI(sT?a`z5SO;8$|wKt(>qhGq%<@2 -WwR>*(paM>IfRIeQ1tUrA7bzIK7Z@#vlbo-;f9gW?Uf64mv!xJb -d`cY^aQW(dc9Vhye_4$<#%Az#sTq!Ve#5!e~Q~grVuTB*I$<=cL|<2e_^fRe7=1uVUmX-e|-H!e6-^n -WpM8%d};vXn&@wN#2!dX}2qLG%|D9gfW<Nn0iRXgvC{|8JD10#Lk$A&5Ya3 -C95ch(3mlPQz?zxEx$#5I|K%I*4cWkLV}$49`ac-n>D;dMyAsAX_4`9aAAy)IJKmv9tS?W9DS+Bd5jlfGC*vT3SU@sP*sJ_Ac5UJ#tTy};7dAhNh18G -!I_gJnRWoU`G75kbYLwHP#XlAkbN5)4_g#@rvRoDGdQ!?sM^dY2bcT`ff_+kp15E$x5C0!fO9KQH000 -080LuVbTd0s@2XZR_0A11m03`qb0B~t=FJE?LZe(wAFK~HhZDnqBb1!UVcx7^PFK~HuXm4&VaCz-LZF -}3ek>B$xxK6i{Y?-m0YkPH*sCjX&xn?(cl5DqkmLgLWBymjF$=wfjh)borl~h& -!j-9yeZIJggXt_ecAn~>b98X<$Ni&&qdz#KD<_#TfF$eUB+Q$49l@CM_=k6Yd-dk6E5U?;&mzH&ohNU -f$?q>;JU#jT?Fm(6%n~mMq0Aj<+nr#(;Bn%dv#TYK$9L{bVm@~k!NMJ0uxQLW+a#u%%LaX)&PUkJ* -JuJ130JF~QtXL+W` -xCl0FBZUni7NmnoW!1 -NTN)RTlz18kvJ>aQoF&j7k~R|HOZP7Ka}x1#pvHT1j4H^5ocHUp-dCifM%B<8kD=y3ya(>#K}(W2Frc -9#^U2QXM~J$Y!FY%?n0y-KFMJ-sJdm|Z+OF$%9K{@A5(74mbC1)OL-_*6)fkEdesaH4fY**&-?sbh{&KJ1>GV$b?;bdNgWdM&! -Tz5Id#lskL8sU5yJo!8YZcMjI-P|t2ffpi{r~!>zdsFlEWGsc+d)fD3+{_3Vex{8ovx)V5nJlEeEigM -W#wzyQ?7ddW2r82tBq~xz+V@it9JkN^nHIY*y}^hccJcB>?B}-KQO$k -ac6+A}9zE>8A0RgOoI&qy?*LFZ-s|+bh)N0(u15RQhzmeeuYyu&*(WRdO8CR+NWhE&GkVwgIpr{u(#S -R_Km^lBIP+8_&WJe+FGg;_BVg)@<2gxC!T}%^@YV0PcE0`Cj|O{C3rUyVN-{{(?+P=p``=j=q`4ip4pW1eY{kG9EmM%rT`(m7T)nGv#Dp -ISIOBGLLGM^Vigp-|v>6X+n-K1VE=9{`2LZVB95Zya7p5zd)2FA@^&l7W~PwGxy&nGgA|qQtwDZF7NJ -!;GA?LaAxY6eB$+Gct40 -lFlPIzCt4MHxlLOB`04X#`Fey$k~Gd&AM=SV<~(T~mq-mUhqG8N3sW;P4R;!hl3X#sv&6e8gKfe2oyZ5_Rw(@ZQ26 -TbzmulL(?yi~eEkF$|=mRo94C8vn_R;=Ht<%~qwfYlfs?8DzK1ufLQb7rAU_vk{^^W^u1yvE9J!Enjp -0d@HNdZo40_!K7rECCcA+bG3%2o#x7hvl_y13+1omxw=PBP0JB>05b3 --gRPc=}h(|vnq1mHc)@iX~&b*B&H!)Kk8~|J{?s^G+X1hTd`u1O!(e{Ry@i0=SizFPIk;{G!8>+HLN4 -0xf;S&jUqN~9Tyv*;?@yi!vjnbJlb!hn$yYifhL&G#LLDcBFmK2@tJtom>yI+Z=sdVjl6{>+It#%d32 -=(lqMcAT8-k>XK@MY+Uk_DQ-pv0;ZNOIgc$ypb+0msr=svW++1!sZ6vc=L4YENi#7YT4E#n -@4omU=zB(dANG`aP{E9>d~Xs@x#^W{x5s?cE5f2;L-2juUuU0>EjH`f^hKh({a1g{mUt)GtMz=qHMGZ -#0qAD6^~aDPdfdP0o(l%q}V38w07B$r5dF{2mcaC|*f9B+&XTw;s8(gp_t_r -j1bX$P3LSFX2sv(whilP9r-&ZjDbuh8U-XH*xiPdvFR=IH((bHC94_wU7Q!eSQr>>dJR0()1uvt%9?+ -dxIM+r^H@I(XtL@{zLX{AB3`2?kgwh@q?Z!36(?-h4FnEJrb|E8K<0BouIPzX00oh!iK2hk(CP8L9fz -xr}R7&(^y*r3yUvR879;5v#(~U|c#^^P;2%)3r!$LX2QBA%jplQqJ4^M%Jj^J2vZtaxzY+$ad%)*md1<-qYjmUmm{y{pzp3bozr&gB5|;B -kn)*V$V-dM+ecOu^~?`@DM{DX9OZ@B^COp`Q3FEUIHK^hnh>rlp~n*v6t^*-`}H-$DNotx;S1R3}fsi9DKMrF`0_ME|#HkJ%^z93#zg^8a+cBxn`&> -iI;2Yw=-Dta-OkQl5cP;Qq@_reQpFeiX9+15$Oy+pC%xeCSn11gtMWXcjWu1^W(-a -MbS@u@vNo3lnxqR!b9f?%Pg*^}^Fx5ro&$oDoQ&GN2jBEv9H>om9-d>kPvl^;EFvY0wqW68brOcbLcp -5xd!D>N^2L)aHeUImkZVg5mit<;bj)k!(rU9F$1p(yG6J6z@UTITNoWD)l)nLN2EdbC4 -p4Hc>--`aR%WDj@}==7-vC@KZ9;%QpqNJTZMY^>Ujp?B)SM<9%0+_cV>ytaqt14*T^wyEBZV|%_5^Fq -=$$r^OH*{m%VtE!GA$~trnL2Y`}Vv&*eD`cz2n_fHT*EgKs15#ba7M#i^9jU -isX{zdG$6A!k7Y{wX6m4nnh!1o2|E;)*HQS$l$$UQZuokOW&js9?a)iWZVGzHsOmG3HC`~)!AyjC6$B ->f{dB~u0+D-#&?QR?^c1m^F~=p)bGn8KSU#BGcS=d8ZQG>x_rj)Z!OkTs*}h_mA@^;-vO%=sNcK}oYB -G`V1OFYe|TY%i!#JlEq9ae2!ZpjFg+tOaYPgy|!L@Eqg_ARUp2! -YdeB&yhOnS%V(KMr~e*$Ik-PzbTAho>bo2#$h}UgfRJv!d;0!yo0gC(^{4VVkc@_3+hMG|08|k`<6il -qwq1VS&g9x?1zujO6zgl`)S-nI4v3rKmnQn%hMtJt}}sV6F`k9ZtR3^BhP{iD}~u1YIbVe$J8^h$0Euh(I3%7g?O7<~o4WtbS<9kOy_Uj3~(}^>7lib<+$M3fnpy82uio4z#X(%54ZjF$ -*Rc%m5lRPs2RQ>p_|6P-w^_IsUYy@6VWjUJEHQ(!&&=$9RUR0HpCO9A|a5f`bJ_med#lK>~|&v+0*P_ -^=lJu=2IT~#+i8n^mEOAbv22LZEho*IqfG_5$@}ZhHY{5kyT-JdBUbgV!tiafy -rr2NyadAB1TBi=8UjJli%Y+vfQ5QF3EH;ZZkQC=FrBzlIX=rR57fUY1a^i^%g0TWJ2uS2(TolK+D2`u -I&((~g^%^)EH~4Z*b!2VU>|N3{^c?s9VHR?~!}+FZE}})_R#QRej*TF#sD)HVn|t%q=M%3lpA!|B2Pp9Y=1JjW~Fntt>|r#~zte`4F@I5@rhr%LYaPaf|-N23xm -*WcJ>W3%#~Wb$AqiClNO7I8|0nl=QRFlV;3{;YZ}{_f(r3Leoo+{ -eYA1ItP%1Z`8C9Xe2SF6FiWCfWh63*^gW4&RGPup*Rh(1^oxiZzBiggadNTKZw*%cXmx3D?~g~H -Ry0?9JnY^l!>wXFt%a_|Yrp+{|D(Ix>wj9MW?}#FY4w8{|4&PaHOMdi2h6%B-f;1;i~R=EuY*=&ZI*CpYV(G9q -iJ$6EWH)=o+Pu(tL)s`f3a{mzZoUUjFv_NZI=YsCGAsJB)A>!W@n%qMkZzq2Xw>j|~mNl_VQO{MCnxw -R=$>*=qtQi`M0VaiUV+hgh`5oM>n?Gbeoc>0u^7JHEwUI{it#{zBqQ#C9uIbWevU0CT!rMzXYeV+`&J -7@^rae(6cg(@i-+L7v@koWZX@ZjJ8-HDh!IGl&3Uy+ipk)zF>jU2g`ro)jfT#U#9XPyAbjz!RVl*Oci -n8nvf=?!`(xs%~-f`=9fSN?qDp@nmS>+~prym%%feT*nwBH@I=IRlvJp96r+yo*enq{__s1yltfX3(I -TOBRNWj!O_3pW(dZ2zMhES4TESBm=3NwR92>9qw=vh_B2pblT|Rg>p7{zo0T~J}i#vd|0%;uosH*___ -s$6C3)QxK)1%U)WtjEu`O!)xW5*db63z?O3yKVV-g=!}sf%wUkV5Hnz7}Yi9qO0llSrzNi8HmoY&?_M -8O%xy2e42K8WdO5<7v;g5{oig9gIakpS=+t|7rHoa|Z+D%&GN*h#w`xUHhWvO4<-d45sOIqBjhJG2F+ -ge+};>W5%=8aj^thb@5+RRkpfpATxIh)(kXxo`8tW29ZGphNjhBUXL+F#J<$EAi;?Z?HAv<2-dMgJpssI~o1L{lwyyY-!Tg=9(-OwRsfNNquExk<6QhDEv|BYY$X_&A32bF|P|d!HkpZk^0WGcVG~&?*mg|% -9T(339-M0GNwF=zEkS{1>-=<=BRdd@4H*c+})NN44Mta*;&?UwHt@XBTKt*YE3nS;}>2WJjTc`e2sC9 -3o!`-B_ZRP*XI@}F9+t$Ge-B+ktuEO1*vaMCHtzBHFl3hvsB~-Nc@Qi(T1^=WM_FT_e9yWG*8>mP_#! -|M8^r}%su8^``pBh!EH`1gwG~w3kQ188Lm6fO6m)v#)1!>KSYE5hV4wK%nuVB1&=I-{+TG{sF6P0pXj -qTI^(v{6yqpS_s`0#lDe}{wqa<_GR`qtKhZd>izUT9JK+6uElhq!l4?qWz0t5bh?>4|~@ZiCizx7FxS -Pd9)%G;70#8oD>8IlW=4B?NoRYx?-B>PK5PV`{4@(2(~o!2H2uR>l=iGBty-%b<8iRj(McCF`+-S66LUZn;h;wSM@1N>Vv+_;0Wnt(WWQp>pRXc!@kPUT6^gg^r4?Hk+ -hWl7Er_E^fCa@@s%U5B195At5*C>X>>lt+vJqm7pjW!G+CsHLwCEq!`Cj|)aPQlv$L*~=nL?&Op>aLs -k3YH%q!vWxs8HZ`_ZFEPN(he!Yth~PPWw@g6?aW8Sqxayn>{fBhQ4T@oaVCBNf$|w^%?mx(;)vZr54P -Sd=$R;(>wOORADEMFNu3x!#TcO9>$41AN0Yw#@Wu6ecGLfJMTF!U&E?{TUJicA=U -f7{18?Y+m#7`uR#}%a0JKuQ& -Fp#?lkps!tN3y0!nsJu2jMoo{@y=3bxCn1P%w_f@)oDZ0{j_Z9A*qp+7}+MQGim->#fTRgh#ZW*UiH( -L{-ms<`@8(tS6XU%PZMYrPu0X5ES{;>EC%F)+y#P#L?lcg@%yq@yp+G}d -3PMHjR>6p&*$X-8MGib?V+3yiB2Jkzc_r_@VT%xu`T`e-I2+dRP5s-D5cQ7~Xxbe&%c6R{TPJ9n6SE2 -WPAx=5B;9ILETV`3usC%TDQhG1sr&AwAyIavtfVP|7~af7l}p2&-p$hQk~7|ifW-icu2Pn1FqEfiVw<@L)GgoTYka#NQ-vv|bvwo|TM6e{Xf#D-7`v5&5d{4#C0ijf9S -Wqsaq3f)QgZ^s>z>IEpi|kO%F@Vf%NU5RAK9y6DgIrxEN7|cN^(1B9vU90dR2;-e#vPoPy^Ved=N$JR -veuzF=+~yHR3mAwNSwOwyH#@t*X&!tBQ2!=HE7nU9kGtQ^Ic%4J9Y!=4P%IujOG(79$rIjk5CkI(=S6 -))ww;(B7&)b3>fylr3-OUVLu9=;wV~IpZ&`!!{^Y8CsRvLw7_Q0)N?MzAW0)TC-v0)yRW-lU8s$#`}^Hy2oX)A{%x-A|qN=`ZgGTM(^bdFOsd^=LG*x( -?<)qAb6c;{8*3IgtQfjM$cUQk|>)8@YVI4IDB@_QL8mrV@188N$_WTjL~S -nda@2bn`}}r82+C2LH;kfdpwv1xHO{!DmYavb=ved7!_%31OgNy0`G2=jFMzx<#y})BgLyPOofeb<(` -9RFZG>KF#IEbyHl-E{QeDEJaTzI(1P*|Ao81B0bj}o6kR*nT&i{O(FbyNL9H+pjqX?L+R=hD@`ws>uZ{MUnr^}+Wl@CZ!24k@+X;VLmM~sL9x;ni&d|s*YQEI= -#gKG4MF*kxGEO4)vmLLZ41LIDH}PSGs0Uqq=>{Lc#qjdr&!1h^rzsXw1KRqdyV@^WvC~Eb#4~uZXOIw1M3ZYWj728M;OdKGDM -tm=hvDlR}58fo4aPXfm=+0@jN?uzIyR|b@Gz_ygiXqxpwOG8m)tPoejg6ciKP?E8ec{yrBxJd}HI5&K -s%#x4FX6=DeW_vDTTxm$}r@>b#)}cgyP?ZO$92K(BMLW6^oT>#LykaO?2>4=-Q-=GWo>v0sO>ABXCjTA*Lj(?i*_LlxZio*qh`9jYL=@$_JN -cBn%7_we-4(3xM%o5Nbm&fXf+&emERO{t~aY@T_mR+p8eU(4pQzx+Z5mz!EsTA|Wrh>ac==?J8N6|8T -@xb3k5p#~(YmA431L4>Cqoq}N>-v1e2Fq<4N_cc~1QLbRp=g~?!4K?#yvK2SyX*aX?@U(q=_u*jmyZS -A!QrkB=39yyby+#r1HhOTQ#cXwbjfw`N)$hlxLRk$>1Dv6~-FhAK5)GOAnieJUR%HkFD_Y^+;1x()TG -+P#Rqi>;cJbEUbks0Twzp9|J3|J88u?O(VP^*XgvC2(359l5XLul(eLZjIKnO#Pong8wqJ8X8Y!X3v= -@amEX21k1gveI5t2j71~1Y3$ACIM6~bnx@`Vj@~$pB2t--NU!W^rdalm?}~=2S`T1sIpl|T_5=OwgZ1 -=2`)cSDkg6CL?leed>BtTEy>Y#w`X8@zj4v9G!F$$=2dYw8&^Z`w;-iLAa_RYI&BBV3UgB&36Vuvd6{ -a@Uc70bq+iY83J|#d*-I%(GN1swwo5P$6?shO!dlawY-l*5v>PV}r=iKVql&nhzcB|92Ca|U2ZhZE&M -O8PdXtkc@ik{(mb@yziNe!VtYNKX_)Un|4YKeS%BMpnwZ+`df?|1k5`|iEt2m62R_usF6`Ae6G)F`f) -t9FAlJ5aG7qm<4_jB}9eu)|OKE$iWh^{7IfEo8?6@Wkv+Yfd|Z7XhP<{CeljR{zZTwz~wDZ@SyOCRc% -4_Si8Xfzft*d5=i`nl_f&y>CDU>n~xsxXu3g=bWkKhi9_8l-F{7vZcu -ad{IF~0r9wL>&(5*}0yp8|I6UmhAc~+A64u_|!cDn6y(yh=*|*H|ldlcDlm|(6mWK+=^sI4Qo}GC$J1 -Va~(Xx@hcJ9B%!+3W7~?@C -rceHjS4H@A!0vLswwjnR*ty8(x(07Rkks_Qrpm-)^2k_wyoq+vvQz?^PN@YjAd0gJ6P?MR@OWDQGnU{ -YCUWfp^b=6Tk1=!l#av-#^frPtR(Y6HNX`a?kxksR;0#It%M#nE25eesxGWZouN=C%XVkER;RRmv{yi -Id0*E^I-q&iR)%%wZHJQQ5GxaHyR%ls5;yONF8-}HQtI|mzQ!GS|9;yiRax-n_B%dv=i&4B06u>qU%w -iM^~~!lxga;qGK(wn$J!uEx@3r%&!_O;RT#*4`bp=k}>h#?Y+( -T{BeV1<7j05y}1S)Oq6AQ(izW&`xuz~*vKR~l|r7xA{cJY^Xt|9to6@$l~_|MTOkH_zS{SLIrS9 -0?w#QlRmCKb@zc5^tc=QpS{miIz3P5FG@Y@*ciJ56JQ}jN&nuI(|yv@sz~lkw3MrsRPCeGKtg*8+zo) -+q(>eN(Ju_SMW-%OT0|-_hR4YA`i_9&|+Vxk|{J(L6GqB&`};#->16T5Gvdugl|w8uxdm@ekV#}T;)xmMu^GEm(GX^mxmO^Eky@;bV(yK7wHUy~^N%MUbsH3IReE-ui~L^6I{sE65-JuNV7x+*|^g1}H?7vgHu0IDS0jWbwIkQmcSyaqIQn<>3wS -nZ_ixL$)}cyWbabxd>G9M)7C8DawTF(vumGL{1n-xw;~7=>wDrpex$u;}~Zs+@_IK}nwgPHr!DmiyfGjB#QQFstN&9(RmQhE|73yP> -Hl|XcuaO79&V=mFEh}wVjqWpFKj;3G(+PGl80E2d}o;a8(NM6{aj1aM<7TY0MW$^eCRvv9eg#BOcR4@+ZKb+cRC28RtMEq`GPMM$e0(amkUaZga(P)w1&sr6U38aIvc?2fwC)Lq -cNWkmG8^~?$jsY&DD_yyqKt`1O@^*RY|ied=7)ju(JF^2~DpIARM4EkV4m8>hP6*z -f(CSjDd0(Oz#=6Di-7sbCMi5{V ->O;HBx$eFR9&h#@Zo@U_&{VnUo<4QijN-i~^iY+Q4w=O?B -~e^gSJCFoCD4u`$@qTF^ytw98zGriNGSs^}$#u -P67Clugm=|%jkvH+in)d6wJ*av-9NNV~ptHg6jz9K0MK^U!>jIOP7E+jf&VW&4I@?(=oJ!m-u!Zg&B2 --w&dV{+M-D4{pulPsjqNE4oKBamkw70?R+0~$!TALQ3SJwo-I50TFb6-(taWro0xl>96xk}19;)zdE7 -7~vMD3i3$O0R(?eSsv1FiFpjw9FYS|BQ1enc$yhp$iZ`=#`Xma6&@=J(C5Jy9A+p{Wd0Xfw_appcFBT -KDk(kNomtLke#zQT0Xb{v{Y8ABZY7HJQh{ZdM##lL&|zccKD6YzA9a4HeO;;Z(seG-NCAE4^ee|=#&) -HbY91 -OB=>KcUhx_c1X_tixRcol+Ex?F{bLLO5&Ekq5K*eUe}`JmFt@6 -aWAK2ms3fSzCf`CZc2m008<2001Wd003}la4%nWWo~3|axZXsXKiI}baO9rba`xLb1z?CX>MtBUtcb8 -d7V^EZ{s!)z57>8wFk>sN5}?f4+3&*w+LUm-+@K8pMpVz+0)~n58oC^C?j}|(qIa|FFuF;FSj@I`^6kerJ+E=An2bF!5@`c?V;v%k{wVMN+ -&}zwP-3r$oI9<7D_IQEPTqkHs4NdEEj#^jbO&yEBl9PWbsZ;$3(Vgb{ZbSOw-&!a@*tenu0au5FHfuxa9@}Hx}q#V68D{7K;R)48r;x -YJj!|5|orEM*cm!n|+0Acs?vYtx`BFFIS1F1-J1y8BdlMpH}b@#%VgH+s@0eo5F!gbd4t5;@-rQAHL& -TX;DiiK;r>5Vdy)wiXL}_sX`m%$Tn@}T3W+iX#^}wo^wJw$^xwW+1)&Qe8}z|K0lqKj2+A=>?ZF68Wm -W=s~^VhGD*GjqHZ9*p>4DT)+yoncJAe5TC;Vsx}=2o?qwo5owdEBgQC6`qi&oeSU6B=!7NLEE`@7q-H -p;;To7xF6K5%Y&yO;kHaB*9o -{DRTS~TIOC}^sO2q%_D=5?5M_$lS}tj3dibY7%J -uo?~s{*2m7ZR{Et-@X(20;xf8XjVQ=7;au#P5625U)BP9yBZ7~w_&i>5PY0s{(UeT5nriDfr3rD>H^V -qYlXoi-4@nnt!a=7VO)Rrk$GrzndY;h^-!vBgf7@ZYpQ$o+4gK(TY&c*q2L}`4%#bur!(maOM9r*swF -ZjqjYKC$aQH^qEgbU}OrG>Z>GF=xt^*Uzsrg%u)PXzphJ;W$`|rUCx!#rk0#Hi>1QY-O00;of09jjm{ -_Gta0{{T<3;+Np0001RX>c!Jc4cm4Z*nhid1q~9Zgg`mb98xZWpgiIZDDY5X>MmOaCwcET~FIE6o&8n -D=h7zAVGJ(X&10EhQvy>qHXM^IEkBBb?nF=g)r^E&mW27QgS*j@avrSdDFx`N#3CgqeYE1%`sywDhVe -NYf_-BMJkuH(dxZDxUO-tp^T_m)6ORE(EJW9?(Wgm{N^5AKA^7`i}^<+4M&e!Jfm}hdjBJ$Txo)f*<` -Z1f4IF`p-$p#`?I2&j0a>-4)*^OP)LS(0Z_r<(I^N0v)F71aMIk@9 -UWB)Ga)wyShBoxgn+1rh&kC{!}NGU5ezP1kO?nwZWspmVDNmD1AkTHT#eJjl^xrQf&3yV;RS22C2p1N -G2kX{I_tXZy)N?2qlfv&W7Du#5r%3oz|{lZsR6#O=2&jW;j&s6*5K%p>dIv2-86}iW|xWQBRRq3bAlW -Vx}{|WY?U{Xm?QS0Zb~7?BdD!9z8e#aZgeM>AcRRRuaf+;=|y7_3@k -zVfO`UZ}9haJ7dN>gB68eb0?ol0OLSs`+O4hD}$Iq=^x<(YVzV^xi3$Y8SfMpZ88)jUa)z%7x|--s{_ -%nAeVFNOWvD+~!E7?$*f6!4m=-sUs<$>Ea5xFd@?l6SbID~+Z7+@R2tN=1)POJ{pQ8XvhxXNlX^&)&E -~clg-V`JlLAL(<{3-D6$?UzysK9p8(DwBIiPXsezJc!Jc4 -cm4Z*nhid2n)XYGq?|UubV{YjZD5Nkc_WQ$<};OK;*Z5Wf3Yj5xJY)O}o5+M7vCu^M9Jn6ey0Vu)8EM -zK@Gzu%caTXm&K(Y(ILj7@7{B@N`RwtcYhfWaeUsXt!_cWfg()Q|A<`;T8>T7RknymaIB%?^yM?Qn7- -INyT{(AvT7uW%fyKH4UM!(c7=1JrFb9Bl&8L)Bm5Y=?jYejh8>yZ#6j)JVm!Gqwm5{2^Yd!D4d*RS3R -z72-kT>+xj!SViJHxXuQ6h^T#5-X~^{6G>yMj=3JF_FjUEi>)6c3^v5U)r66N>+5c8NW*)%b0>Fapz+ -CnU2IQhmM=x--m=W*Y;&?eA0g5Hl2lYx@cIur8o%?}W4*D*wJsC)AIscz~a>1SGKyss!w=7 -KB#>cj6Dy?KIPPT-n6<*Vd6T)PC??=PzWf_I=~XQTWZ6!BVyNZ+7x>J8$nJO1(*-1DrpvdM|~x> -N?ol`@)dd^^8*Qb_Svk5fAUG4$gj<`2M%~g#1eZrCOLR*8*e-Ypq^oCNh}u6609{TWOYRV*ooeFU$_q -0(h~57g=Nph(BvB$`X`ja<$H-z?>}7e3QxI8Rj@wD16BhpD=zo!8mS7hk*dOI3=?#z&By3gmLdVJWl0NJXthchtr -I8I1cGygxS`Oh;mbv^h2l*qJ^&#TJVz`%Px0k^PwXF1O9KQH000080LuVbTbLREok0Nr0J8!B04D$d0 -B~t=FJE?LZe(wAFK~Hqa&Ky7V{~6=Z*OaJFJE72ZfSI1UoLQYb&}mn!!Q)Z-}6^o;)~)e>Z9O`bGRXX -pbilkA#Lw8E=yuQiu>=qZP(#k*M~yRIo$ltC7knDFP%dlAX+G-*5Fpo+b=9)7VT1>&w>Wn64 -TU7W&q!&vz;Wx~>(bCL~u6#qxoA$f&Goof79V=_o?VsZoI;2%n7UCVeM`z2*^PLR7R8nj{2V0zb>nvr -|qgfi%i{GlB0G;Daq^Vk)Us*{OfvJE;%9^Y#|#0(&ZKAbR&m?u@Z*K&rIc1>gX#(RnN+(E;UT7yz9sW -mC{AkPy&GVn)E90$t6{n3``tfQtsE=Mg+_xStwq)&#pHAJdlk{|KM4Urc!Jc4cm4Z*nhid2n)XYGq?|UubV{YjZDOX>D+9E^vA6TWxRKI1>KuU%^v9oL#8 -GVS&RfjP>HuHn+|8+5~NS#Ucn~TA~v+vZxg$yGGIfelsLRk&rcisDYvmH5?A-<(V0p;;TZG75N -}VKH~bjkj*!_wO=c?DwaHBqh%>pBoTSVK2{lD5cAxOtY}WJSgEOmj0##^Hzaf+@z23$9|X?{m88Hgnn -Ko#q7(^}vLSkV{`~aac}S9LmfXY%&9dgZsH>u`^t@_Y@cgQobN4?B*K%}!$R3Z`CND(9%1o7MxRp)GD;A@-j>Y1`lz2SBxR) -V8YfOn6c>MTgL(41KVY}hrc%oqW7->TiY=V%$o<$0*Wfep9Jpqb)~f;+4$0EtZe%MNO| -@uN1hsHa#$hQA2~e`CE^{qMCTKcd;=xo<>|Yet|`yxB4gv@j|o$nHy#9D@t$WHSuj#kE?G*zV -bOsWdM-(5o{I{fYd=MbTv1#LaYKxgwce#Yi|4#@JejSc_mNV1Ce>Sizcn!82ZMBTo^Ly8&3+{d1U5VdBi&j8AA -a3RL!cOHvnr9h1n~#V8u~L4xD5g#R|zfa1o)xJ$EC&^pr#F!w7iB|N=_7m1v#G_mp%Nygw;lq>;pl$h -XFe^*{r1NCHw`zWtb^Np>W*zwKN+n${<@wH^fK}+_#-Qc?%7-<(jEie_p{)Tq?if)~aF5r)vXK5@`E~ -{?z@I2)St(b&S3QsG&@xA-?qhn~lyDhFE2maAItKv2Z#5xJ+fZPS>2Y~z@0NxQEc6|=YG2%hESJxuN$ -iYN}u7axDl)w+7s}(-dsye^s65U0T>9Wp~>D-}cbN`&wIR>Xes5EL;)_cg;ARJT~W*bo=itSjt37F2` -qE$EyM0JsZ9mgKU#N=wb_!v|hq8?}P53zeagMUpX__gNjPid5AAbPiI(13su0nn%7SXWcGHijnu!_1+ -VA2tYT1cWMQx^!&plZ$c{Q@~fnVJt;mCM>Q>#!TT*qIo3XLmQd!zg6RFm`^i61wt`#A`^hT1_}_gjq; -8GX-cLc0TpBUdc*`YEDM(KC5I=F^(hc$084A8o@)odrNBZWTO-H_Aj(9du*=c>70k(r-VnhJ8A0CY;( -_K^g^R4N7+JGY;$*Sxh^FOttp^q|^?Jl5Fc@26r#w|KN&ZK(nyE}P4se1bt6D1jZ?+T|>j9tG2s*RuO -^@K3WMZjX5!(r9N;RVyzy@O6@^)-F(VQNU_u!!vbPGxZ$HoM{5;=JPl7CdG{%i|ym6PwHe|3^Oc`L|< -87{eXf`m%NjIISwu~35gYb9aN3GMOcXx-$R0+2%&VpuIGCK$C=$)Ua@(UQ^~`fr3cJ+K4=|*#zv6G$^O?wGB -00o}htgvwX@+mOv-Cy<1spyc1@3>RMeOm<7Af?p&ZMrKrr|&jxN8x-n}*!moeDVlBl|PpY{f=0zqr!13>9F%;wo_cY!$aTsvyb3zIq^41=BW+A|)u -EAHQ}jGnk;LP1wQQFbk`Cw1pu-m{PDYXr!RYMc#VM|{Aly~n^vZlTh#1OCcG^t~}{o_}{ -v)o={)sb5*@hNAK3K1Qe@Z_d3g&|xk+2TNutV%E}8#IQ|-oRn>efO?_iX;Zu#$7&r!bj`G -<=9@x2$-5QsE|Qc`UT0`WiQG{_IdljT63P6g~7VHYgLbL@xo`J;-!3Y$?p%L$>C@ei -F>MYYw!ZkhuswF#CE2ip$+@t6ktJO#Zf=;%C{cCGNaJ=<;_5JCw|#^M#LD0H{VWr}_MGglA(^Ufmc0*XT^y*+EK^Z(|-p9i9tF -*&G(!uZgBYt~IA>>U{3RDsWa3p)FNm)C53dS;TL!U&p -%iL2>F3palLebf0?G4f>`9ENdEp{=-SwXK@e<2j+Cl1amqg{+!R@5S-ouIn~221Diu|b)ehsLg-WNa> -mW{PsV#8vbUW*_^6rw4&Znz+%6eb@$4f+2?B-(aect*wa@uZ0Mpwv?ZJHB(a?Nfb#jtpW0a^ZD_~jf! ->7UM-vAJ+&cOa?YJuN9{7{?2UqhzzLm@j42hebX@b7VJezH}E2WbR*krktpq=dlqq~9Sop7r~Fg2K3-b?)&zYGO^}uSTMzU_dY~q5ceX)3V0kUk@7V^m==I -m{gbrcpSJ4c0Q~$TN=<8^Ux`6PYosnfe&>Z<-rBcLiXpTH?x>34!bL8>*>KdinBj;bRS5kL9zluglNj -iAwd$)GRk>f$y1|`>T=oxNd2lepb^V$CFq;CQ5O-lEGYt3~ZT;JP1o;ePMaz~|3D0iMbKqwCfq_dBq+ -uCeZ{&f^!Sr!U{ePv4#V_j&yE=-JET*Uv%X{ymq^tz`kF -d$&#YUT-lM>?AAn{lwmE|F(fA@5O+yTFm(D==_Dc{^i|sY3n@IAea}oqH}YZWBa_Z-B`V1Ql-A=O3yq -m>Pl%!s~I$+zkK#+o!4>NX`z0LYg|F14b(`46u~6KTk_DF&AyxXCB!4V8q&UMg0F%(H$eOqaZf`pVCl -{1f&P^XeG{a)T+zR7)1}{W4Px9Q#LtE17q&4{m18Z*Wa2n3)E((%KVI;I7ni$to(2g$*O+;K^(S=HtS -ZWx`<_y3Cv)dp8_M@a@OIV90>F=hx_x&d7`WP^Z+{NHVA1lhZ!thx2==a)izfPZFE;S*)1c;du0O13^ -XAN~oz?^y!)vYkrn;~B%=$pGe-b*-tcMRnw%w^U<#6VuoZWhJU-N0_nrX3h+F*h&4-LF%qd4irgVQep -{YSi}S#UY}KTt~p1QY-O00;of09jiA00002000000000a0001RX>c!Jc4cm4Z*nhid2n)XYGq?|UubV -{YjZDfc`kH$aAjlz08mQ<1QY-O00;of09jkIX1_ld0RR9k0ssIr0001RX>c!Jc4cm4Z*nhid2n)XYGq -?|UubV{YjZDOX>MO|a&Kd0b8|0WUukY>bYEXCaCx0ju};J=4Bhh;R#>hQ>cWCa>; -U$NQ7L~m)MpHQ&sAuoBXR0a&KRb*HPpXMRR(wv0(+0hRm!d;W{Lnr^$5RbWfkx7@e)o9Z<2&g -RMRm(w4$Vg2)HJ90Msx)@cioS3_h5gn&^E^4D0npg7xChjaEBf7NAOz{P5a6;uVa6^`Q|s~3s6e~1QY --O00;of09jiwC3Alr4FCYRF8}~G0001RX>c!Jc4cm4Z*nhid2n)XYGq?|UubV{YjZDOX>MO|a&Kd0b8 -|0WX>MO|a&Kd0b8{|mdEHuVa~ro2{?1>4o+o3{rOtWz(2QDUMzQ7EP5d&lJegEc&nFTTVIL3I1CBDI@ -qh2`;`LrgNw%Bm)RT!w-eR%XefGgF4oQ-nu7qOvU%8R*`2zm3oT=ubs^y$3#k6GeyeODnady#&V!vaxK3sG{G*O%&$1hX<<^C_FlQrEJ`7NQ2(<8mWvS+04h*(R? -=eo=5x!suW2>ep8%FaL4;=rnu!^5svjvmf6)`}uM9V_Ty#OAfA^pqi1 --rAv8bzW7xR&PIdt7w@J -F&7QIc4@^)e26+Y&6)(r;adMSR887E@A*uYUZVPWkwc>XXwvdcnQwAYU@DD|Kg@Zhg)HrK0JxJ}D9mzj28YBh(aR1ynBhhVB9lg-=fm5CfrNCMHmzfj&A -fVPpoX~QG+{6iE`~#muHo*gQz!JC)C)(!$&u2F-+)H+r?th21`Og+88L@0;BQXwZ!ki=HG4g#rMvGC> -e$6N+XPBMb&}@3i^*Gr}kcqLASFM3(1A}5NPR?l*&_`xqoJ|>|fwO>Dzz?++7@=`;2y>7)#3nX1c<~< -W3#x(Wi1I+!+=G$FsD59x=1l9@+i>FUjRR} -$fw*nurtpcoSGd2~a;$>Ju55}d<0gs$am}qm_QJ+*)tM3u4`F`(Q9=Hqn -F8}RuW%31%IkPCV@9F9flYJlthV~E@i#Wbq3RpCm~EOmSJv?Bl65JuvjuU5$nbzO`^02YXt=6TGxSnN -r73b#(3v6N%b7Vu#imoJtgt^@hYekvg8hh1ZZuR5QH;O~;JARTG&m=?1;@^M;he;dAw_^EG -?(UfquFdW6tk}9j^=Y&FKFMC1Y=Yof=$=^3cAYTZLu(}^89{;-hn)QlBd -hE@cv9ogryH&1e>$RfL*R-dIH*BXM#SRM8_C(_i+X-8z0t;s~|91nmn#v%Tt%XpP@d5*}6qlLe8f!BZ -oc)YWAIv66PR@(Hny6}SJNfJ!mq{g|E%*`YU=r}f;!gf+6=|f*Z&c|%qvDe7NE`g|PSs|eX5+#O#e2_ -i#-6A4d)Gzgi#XUR`(N|1v+ECVGZt8|>1B@cdO$V}k9+DGlPz&`tPm;_C=KOR=K9J}FC;!3H%)3n5f# -3Tabsq(%L)RYu8Wr62=9s4k*Pjc`VZiH7WF<-$x4Bw)Dn~rg)zy4wozj5ie2stbl!P -r?(dwVq7c6@d)xd9-zh_O*|F9ZYXp_r;)aFoWRb*JL+VchH;~2iTUX^u4hABxyqYy!L5NAVo>gjFrgGSyW>pa2z!HMTN -=eD^AWvHXi9a-p_8p_1a%EFO{^dFLmcj4d9NyeUuv?Ma2W9)!kx3Fz^?r^9#TTq%wPJGDa@z9~>+T-NWRijs -B#c_)^p%uZ4KfV!upEX;`yaje_>{wb;_peJ=1(W-{s>6J+?nkPl~VUa$L%HTB&sd>82^?Wt1le2$4{O -z~%KOfk?zXc;6CZ^+V_1EhT;zBtcLSH-pl -miV9?-zU^krpY`x^7<2g!iCu`2PqL{&8jLMMh@nlY1D0Ka}YKPjLbPm93GHg@}LdaM6%612Co%ZBG3L -}z9%P_Aa0xozGg7$>C!uZSO@yeBO@q*|wYHQ#{s8!=gWhQB7~p;>0n%~3Y>>CF+U!FmKAvkg -;CMeGM-kRN(Xhn%Z0Qr?3SXADl0uaJ&A?4I6(+ -^x;c_AU$AdA#WHc%?_4P=->GbUj;FrqiUR_))VepC9_5ictanYly#6KP+NW03HLBbDaPr!K%X`n^_zg7FOXD1~k?ZWUbSg`?YbIvWzi^1crIfQpDgYKMm!7%Bv*RH7Uys1sz)WQ-=dA)|Z@y-4~hm}#)on^j0Bso~Dv! -lIK8e7vE#wcmmWjP+(rMyt%$(djHc@W6xGA|wJ_?wv2Q$>-0Obu6l6L$p)zhsieUMEe+$K=o)(!!|L< -z6h_KK9wgsX2?K&+QGs@q-7!$TN4%_`F3c{p$vJHo%8wD5c{$n&Gr=_<)qB5fb8os|RG=t9ZI)Vddj{ -o7To1jJ4Cy*qxtjshvAM9p5z{IoaL>-8Z2X$vl$m#33tT7mcoN;^^plu~Rfrt^~gi{hY%JD8V_T5naYtcB(D*P1 -N;8MkOnrQv9jvMM-OcV#n@DIf_&X9m&eR^#sAM`!BcfZ^q6LvTt3)(^Qyg1i`hp$Pe3>)mMS@8Fh3W$ -gfgU@K -jsc10nPxACkM{a3r_mT=;TSOi*mV9_t*ih@9;I&N -wh8&4;ZlKUiUC+kA4)RfLO?e#1H2sJjV%Q)Yci`G2COY)p`fi7aRmoZPqOV{XIKc(4q?s<@y$iwE7tux+W%VOi@lgphh?|{80Pb2=C@Ro4hcra>_xZko6^OW*SNgSf+QngI&&jevZchr|L=i7vj`>-g -+3#-ghM4R}HILsbhYsURsJ2v -+cZCeT~29@zWfLDVyyV4d@dt)Y7O23Y++ZVz}~5BuFxB8(IZ_e@^)#ONs$e-@w*QfNeK!0XsZ+1l~c> -MVga(J{4#q!tpa&5xh+0?!{`66BR^d+8*P5tH78wnBsZz@^#f-%;K45UKZE;2+OcYHIRVT02mkOg{^E -Tjt&;rbwobK&Rbt#7UV`($xvvg!KGv&`@7=BQWUW(-vj<7od6MagNqxWlprlDS&<64qZ?G~atMc-J1Q -(8CDlm(2D=|CAG4LnwQ{2u8H`M#AhR~w4LEK&I;!%tV@r7HY2+tDyi}UhQ!v8xtRWuZD6ZG#XhsLg+e --Txq0c6E6eyQ+de;&5Lnxw;$gme95y5J<0((o5=jwcRp28>mIvzV@;1?-RZa~Dx_GDAB2W7Sj;t43DA -|wZ;^L!h&iio-Yr%>TsC^?L(Bl)(nMbjv-JaP|$HhbS-*5Yvw(m>mY0|9~H5@pb+)Zrmm6aK@m_5z%Tz75XIPm9TRcD6J^f|31 --5{&;;zGu&eC*2HQ>3fz!##h+ST_vXL6>&F^N111pk{+`17rJIG$=q}7z2wFUGE9gWhV_T`Cx<*uO(H -l6Yn5#SFPNo4i3>~ujm5O=t~+JgE_@}|o9Ee`04Q0wMmS)>1BFXRBJFs7qZ2EAhyCMF1=oluw-C1QlNbII0P(xbv-cl2az<9K*4KKU!j{Acm> -fG!g@43eBj`l+b9T9P=eFuqv(&n4;*t5G&FJ!dAh<5PZ?PAPxJ(ilN}Uxmc@wGBl^HTX{^#c8WrPgF0 -}V2_C(`^2jAf8)35%K#t-F84dj3#ELv^o4JHPgJeZ&g~%W&v|6Ru%e85sC*t}Zu<;ng?4XkB!)Yh+6P3}h=Qi7u2RDgIV;_b6I}dd%h7hz5y2Ci(?$Kr2BgDm!M -w!^U@5mU6Pj)jlz0%5UI$eW6cOi+jpDc$_pcm#1l -pHoQ1pTyi4J%bEVt};Yup~~Pf(VU!dc1VSbPjX(Wr$(!;5wxQJedC9&o~`cgm+x+8kbF-A*&EZ&?}3y -%+b7Si3<54>%9&-!Uzw=TdP?x2tfO>5oB88FWmxP8OO(m3vW#pK>st1X-Tf(*cywb;21enY&a0&*d9ioLgB2u9r=@d~a~UO7 -%&m5P^UWkevkrZS@)td5pg4a6ShOGZG6Wp$Wa;a~FpAb#CU&-$Wj`m5xI`=pj<`iVS6{j?XPZtm9-hA -IpZUD*WAIzmrE~CMBruE%Y^+P_XqKI3hn2^;04TAt}>3~G_uwS%e}zh(`fI)l@W`{?utx(L7O6_O-;? ->@tIa6pU%ZpB6+xVHqA>4oZVJUs&Hi8I7=ofjd3&?%fwFQq!XAp(hxy!hN!2UV6^90=Eo%NYjxQ(9y~eo~K>lp85zEa?{JhvKH^I -1{miNEjJ=}{=i}iZ3x?kRI#N9f-mcM&17OOwRKbNZ;pgWpxfS+(W>e5Lx;YyW_NmB;VoKyy!Izp36rl -N+(ox*`|bMOub}1P=jDgx{T~!U@0a(h+YNtxVBuwaT&%%i4 -<8n5@$q5(@osa=)s}Ay8@yQn{d?F{)2##D%y3sY4Hxj7HNDkH&*{K5L>In7@s|)co0sym(K+Bmc-#Qj -+7YrlQ|j0zGpPG}R+`4<7p>>bEXUqI=I}X~6c>Haa;qD?Z=`TG`6obcTZ*ZF -I?+W5k$|pXkmzGsD$EpQ&p7vEp$fS|F#_|+V+6XI4f8jS3j2~OvuZ?tqxb{i5ox4#ylA+CRnK$>8b^d+ -W~Zc!4Vr%6#327V@tJ(I2(pE2P%-^n(0v`wXIXjgazFyrdN=v>;CkSLJrXvLJyqqXU_%ukL1lc6ieW| -fDi#rfQLd-P-@9q5P_NH&dR{*XmuMjN9p^H=``P)h>@6aWAK2ms3fSz8Jt)Z=Fd006-g001BW003}la -4%nWWo~3|axZdaadl;LbaO9XUukY>bYEXCaCx;@ZExE+68^4VF{^yA45&)%6y1Z-1n9NtwR_#PNOIc` -MWLl7#^zQQwIr3SchURp`;a0f*>U!Iz+I5U(r|{HdFF*PJ|dz#kB*3r^L&bD_eQ1EUL{hp}!WdqYsOiJjtYw -|jtyHvbH(*_IQLl6s{heu1&_6D2sMT!67f}RN&d<+DY#ONyDStvBD?00G*@~NFvNopCi{s<_`}?eM=V -Vf?CevOb(}tN==Y=e}CC*Py!rF4uY3B5&M#+_8n+-Ihiq)$Y3Qer~Dl23SQ;e0nQ7w<6EXxSSe0YC-G -w9r~Z6#R=6-!Z=6waJuyMjGRX4siOrr)AbqBe=j4FAnI -;vbRoD0J)|V&J}s7l96fRSkr=Q!Y-&3-f+q9(VMw>CT2xwnor;^&NBw^c=@EZPkKWk$$4vgjLn9_;P~6lWz40kh78JkO$NQOIVi#AI8w5_IBt9hYy`DoOk!Oz?plLOgA_yiL=-R88NN-fJ86%-5Gs~OMVxBCOV>@ -xDvP9z=S-jCk82nl&mU*h(m8BOy+V3{MQn(BsN^9y^O4@EmFjy3Uu)oPGXv7&$3w;epUhe}jrL!} -dFG6@fFo`wRJ4t*3o4f`uExi5VHcob+?oQjVU!+P2rNVw-wOBs{%TKChAQ*a@on-Trp3^ftob2O~oTh{lF`v@&GfECd+L^q+75~x+5&)9LR&B -eAq1a(9t8a*1dZbNRoPcPxw{7?e|()HG`3VQ3#qQ;eJStjEY9q$&rH&9_?XQ-!*u^_o!N)Pzom~~YN( -!3s36@12q%|;OThI9ib(H_LpR!f$F*IU$C<~@|Wa8V7qb5lBUnU9`b(|a}IPHx)(N;sOJZ9Q!@ha}a9mr|ZTX1$zU$hlU|s^~jjPb;q=6MYeH%Ga79$ihMkQEIDSq?fI)` -K8lwLzGJ{`pO``|2urJanFSO!&^twcy!g@!^NNSUaAL6NtB%G}E*(}A&XLB;^=N3m06hr^#d$`9+%gN -3f&<(qGm=rK&Y>*;8e6c3U~lU~afrF)5d5a><|US_GsJ<}p|4IFJQcqc6@n;<#R`ef;kQGT8 -C^B%6fRQMR>r2y%4W$L30k2}A-~z36sSm-vM1ikwl3qL7KXvd3wK6@o&4G*2D>+%d>D)Rs6eg2Fm0Q9 -n?rte>@m;SmA*V@Ee1-3O1JyYH>e2%b=iul99|3dg^%9Cq&SdDt_vj`#skrhiUYw2Q;lBXp5bLE!O_% -PfvGlVU-AlVOlp%Vr8S1vlw{{o-zwn)b3FdH_~*ymM<>V*ioFSF-MgLQ-9um-zT4?p-Afet@U-6&y~88YB+qNMforE3D@z}$$tY}-smqo(XZPoy5|$>=FKzv9nv)V6HrS71QY-O00;of09jjpxozqM0RRBA0RR9a0001RX>c!J -c4cm4Z*nhkWpQ<7b98erUte}*a&u{KZeL$6aCv2qL2JV>42AFc6(T)sm#CgY=@<+)b}8&qMt3VDZlX4 -(iJjTbSVw;R>^fZ;!#pIt_aWgKV1`!#{+%;SW-#6YjF}gXaDKioed^m#J#ph-nFW;?Ym~P`O_HK;(zR -3+@Cd1)kAbYSskdvgm65%ZTE%dEa*bB7+HBUa*9KNYCzQp&4yKJv<6Lha=JU%UG$d-*P^~MAQU~^vjQ -!ooG3U4l_HK}EeKy$_<2h#wZ;HX6G!cCzrN}k#M$&z!*N{3Pr|TU(9V2A|gxVYe*vg8Ooxb?oi>jfFB -MVWb`;(D}#>MdEGQz)Wl{iitwgX@Cd)U4+#_sm{l0kIR+*q06^-sKrSfJE#tlNWS-GXAlHT<4=dinkW_0qL#c&ofV~Ff1f;zU1) -VDax~zNg>7K#WM+ku25iACaE*}B%2k-dpj9@oX<**lV7o$O1>{(x%t1-Od#kZRLTz#9S1kz&ENj#TLg -NU|4e3@jw<8vi!=kE{wxE%*tyUF8-68b*8z)LDWNNHtEd4t9G@GU$lgTuhr=KV1Njgu`x8IftoRLCH6 -(xpx6RfY+@ycBLw_eHX*RO)n3Ho9=KVbg{nHHBX7+)=});;z2`UUne+t0Y{5p6b|EWgYbJuy5P(HKr! -932zv<2a${Pftqv!x4KXNov0s(6ebQnix(P9k&=N?@@#gh5ccS<+uf_Qpv5jL*ETT-iF<03@9^|)8AR -ZO(sMzid#`?uMdQlz@bBe+I+{%jw$1}T<0Yk1B!H;8vUMin!2CUU#-!gwA{}FAzHO|Dta9uX^GHzGi- -J;gD=2oCUWc-y8@}hy^O?+%yQjCuq*_bGk(MjOcw8EGjOtX%BqqEicIL8Fc=Ndv|q6WVT3f6%MAA^cr -+({^!`!KQW-i-yNRX2&CVtRfy}KO#7HhVBU*0Q=~C0i?2U(b$5@_y%r>debE{*1lRXY?LaW -WS7c;bnQIK@ZiDNTOM_-pqqyz4(CyuIShv$+ynX)F171?!}|yo!z=irDRBiERHEsOHy_`x$JMhe&9)vik)Qkb)yf7L;?+<(e -G}+gTY|0Qh(X1x~UVH%1xuHdhf68&)(koDyzl1Ot-lbi?V2BR@9P{p=Z@ -Te52T4sgXlSL}l&>JBjq(Uv4tSsteZvb=I%hm*|tdoswR_142F0ND~!~;>3e~}aM!@oayYTh-so2{=CRStZ*l&$Ib=iTbcm8yZvSPKS9}tMqkQ=0{akR%3C<^E%Jgs@yj8@ILT -KF$h~wf~=2`lqlBr@a!IS%d8tw`BoRkZRkN+g|t>R_U^PlC1H>#;_y+^Oq&7s{G@A> -PhY|4c}(%$GklYqV-ZXHhC&-(y*18ev9WDHM$_KI5F8`Rm^iJOi2(GOq)w6T_jyoAj#D}qNq^P{oY;C -g1*eFk0n;cS~MHvsfXa0dG8ncOba-8ye%RjbdBl~lO54u9}Nb4DV!&C0x(&1Ib}X834UHyRJgqFe(Ln -l{h`__+e+8VeYLCSV(uSDCW7RUbeoESi~qJZs9?FVX<*qt8x?3;B?r1$+=ZBBFZx<|KG{wpgjP3?A7g -jtcp=M_=Ya$Qg!EEb%8G5>@UID3hap3Fz7`nr-D!a|lR0h6Y8QHQBX#nJpUK7I&+Kl9xw69K1a_pB)~ -|Uc7zv^5kgt`r!OWo{7Qnui0j!(isZNn%`WihBwksqkF81MVZ0|@~ahLQ1{wYY6Ijms}WslmMv$6QmI -P$ZLMzx&pfq=kczXEL8Eu4Srah(*aZK+C-IF~63*0+{G8P_pa6j3dI2LyxJwWO1Xeh08~{ -7#Gqum$<;e(JH`(0yZU)j69;z!6W))+9)WDDt7?yr;P~5(^KRVxz2D}fe64nc=lIt3f^6`@$dBz -0cJw%a}S{wn+;GGd=L;V=3b;LwZtN=~oehI%z^6~~KaHNL>0_bdvZH8gMdms)Li?TxLjK*@gtwdTbK(5^=A#~SzA!F@ -OfWZ-jgZk-a{gnSEZMbdJX6x9CD&78SfDtzAf`y7V@ ->kY^J)t}5V`s}>QYcW^o!tn6M@;D!FY9BpNMk)3+Ukl?GivzZS4*7V$Xt@5bm7lN7PIRbCZiU)bCplo -?nQ$LPCT^1I!jkUhG5+OETkRJ`#UkDWJoOh?CwDS?aOLUF-s+@f1M%5ePjmnFnrU){tgKU_kuWdIZ2s -K+~ygBup4uDsh=zJ`GNAv&rd;M{bGFMJ-eCRoQ@5nS;;*icHz%vmCy_AV6Cr0E|Q2o*GEX1UR$+&K7_{w -XTypj-%xkMgz1^8hSEp6bP~THp`pGuwB^N$S~psV@vND2L`c>lzyoSRmuD@Z!L^lE#*Q{9~+7R2d*a1 -3s6QuoCU`!P&*Pl{Bee3)M8FvBd>v!(yBwOx`7fw5h7mWO_t~ECDdC`Ubnd83t%lk2EuJTS#-uz65s1 -Dz-bHv$W9UKv*A<%eb;)`hE`B3co8HPa>aH%hmPgapvns}Q5Rml1l^TH(AM%po&Bm968y;4+qDPL8;m -58yhc>hda!!{4|>ng`6g=a2B%*OC6Og9Y^P;b`@{-)<&DfD$S}Hc$I -2gfm?V492j+f%L+zRwP#U)VNHkXwVgrdA=iogeH#w#H*v9MUW-6%bhbShdqXqqNKQeIq;bAMsz4NG0~<3l@L?PJO}JfoaW-<7nlScI7l^oGA -84NjnI+p6MGtc5E2Rsfk5}EWE5aF;aznewSB?yPZJ@f_n$SZZRdzG(2Z_A>gu7l%e2-%yIL|2PQ9IJO -@U_Sg)4^b@yEd!W9I6qgE#W2@=`V6iZ&{CF|>E{5m{-1F**D)WOo9;j>)kdk+{I$s9ty-hWkGG1J&U` -=!w8DZ@dTMP!4&%X5SdRfM11q3_=esIU!SQIBJu6j9=>4rod?XV~&Z4blgFThA!HqXTf=7(my6oHyE -k~KNc(E(IekWrm+Gu_85lCWDY=Z%ZzR0#pFEak&}I(*~-*tF;w4Z#eq2os+nS+Xy?N$QFssH^E7+u9NWf-o=cI16Md{3cZ4P?MN_@Xlmp#9Tz1^=j20dWk!WQ(vX@lVfNg~2U0cKmJRA`uit -tGF)KnfYgw>)@iPhy6O?Aq*rZ}U_Za7T}cUsu3_6Wmga9jj5hTS4I;lMa?Fs4{{ -^8<#E?kcbEAga5hv@d(6KJAA53Y$c*>Z#8|K@+3t4PON>naAkg@I}C5<1VHSgxU4;<5RoayB_Y!u`&y9UA`82R+|T8GHFq!q3^imJNo5^kqu;%iAX>jR!x1AHLKx(@@#{aPDC=GNGy3fgr#YmGaM4s(S -2tI@8Seq00ci?8nJ6;Q#-}Dr=IfEe2JSl=I>G4Q4b>vI@Ed8wp+c&?LtNm-|sTcxtGq&hM3k5zkFxH8 -ihM=OECI0`1Q|;{kAc#5#>b!F)zZA)}gb-sOrJQ^Es3s3h+O1=1#Ulp2J;n05!+heD -gMeYmMMj}LlsX&+3r>W(-#&$Zha%gAKy6zdqG0kKZRp|KTAjK~) -l!!MX(U@rWlbWU^G4Ns(1wD2O{<;#yJ}m8-Ni7YoulC9`Wb -KaAp`Osn%ARv7977nJ)@X?EF$r<92{B>US_ZdKt9{7L4%#B3rND>XC{xa81w+kxIBgEYa%|HxqXV;x7 -LcRnCvPj=)z2x``>{GhORUJCMOR~gNCvGYInuVe^>9gXU}q6jYHqeYyx%1_#q!;EaE3!2=dDasD8kez -4c~QHrx##kiQ4FV&H$%Eiq%^a5~a~q0r+S;NP%_LjWA`}#JFD8){6mk6>d9}>F2%(^2hmO -#^FO(mQ3XuCh|(5GESqm(2b>(`+_3M`Q9V|NV6`#SpFpf -TP%Jv;w^kJRAuKnGcCZd}^dPDm9fS(x(U=}m5skCYwFOR5S?AA6t!_9MQZB9YB}h`KhI9ETepzRQB7t -C6!69A?j2#%A2qMqU_x7}&-~gmqtgG24pusBHw%~3B}1p;IaTxO4N8aXqUNhsleC -`?SVpqYOkG;G~l({)r~Fos3O(?llhGVYU1T!eM*e(fx#lx{&5rpP^z#>g-XN-v@HvEkn7T4GX}vGR$O -{qR8q}Lm8B*v1Cu~L01=GpA8FheQ_Lsybxal-J+Pp3^a;t;u#lVsa1ce8*@7bDD_LABCpl`YpiET_Ta -_l_hqCh2gv83QG43u-12$vDXZ#Q}hz%eUkQTUw+6DtEPJsM1^?#pVO!7Eobiz{T#r%yr5 -La+ZA{^%DNyNO-)i^{ZEDB&^U~ib1zg`j`d?l~5a$EIR%?5h*JQUT60z;-sHmFUqLf?X=V$n{h219?f -CL*6GNm^Rrq9;xk0Q-WNS?yNU{#_IM%3(2327KeaYUF4A(6}g)t#&25_&xgqo}^K%p+=0^dw9jppf -)D51qma+8ns?nw!!oLon(RBJgEa)g=6J0#wiM353v9#WFRD120{E8O@5USmawW!}a+9VT=E{$PSNJ;a -~wYBQ9&w4z|b_fIGS2tO+}%o@!sVzHz|%u7p`RvG`a>-?T%ItxSXDfy}>BYoT$W^C7%mcw+kWT96 -l8ibzlNX$|m-8TQ6#&179$Qc-fC0^-=A#T}K$ty6gDjDe4h@WY$!^qjzzG3P;v)ipH+gDcs+3?rq}$h -w9VsTAlms-=Vi(9=(2Zba-%n^l~EJ(&0?K?lt;VAmA=OV>LLNp2}LmWizT4Pj)Q*qe37lk)}*6;raAg -^4+uKnWt9oVMP#W$A*Fhb$^55Z&8e3>jz-^x{`~}@s)QvMRs73_qc9uL2ZUsrqP63?P!uW9qYbU<<}h -)jJc2|!7m{#$CN#2k<102xq+<#VnK)uoaf_x}-AHSh%!d#AKI~?1186v6sJkn*=1 -!Mt%xFpmZh~)U4W=~VY0t}36t8+@rsfCO}*SV)UdzV=K21A{O8jr1bvNXI#boi3_Rix+9*_YFEx5#!G -C6q`}LXTTc@@UwfyfIuB-b$w1z`0+FKj&^VRx^Pq6#9Lj>^g<85P_T%z#Jhn;i(tdg_S->sth&ZoZ2Y -LvyKF=TPeM>0X=>4Omt>fyGkD9S!Iu|n!*oOe@Wi(8Fw7Vu~4H&ZoVQGP_|L@*viW~HU;pytS&{!`IMaD2jlOfnQ&9DcL8LQOHYUnt#3e03gZ#It*1 -T?hWKhniM0{rE$`=z^`27lS7EhIhch|)@yGHb(b=9uB)!OVfAp{AYZa?!TOZW14AolkMqX~-smn9xV! -^(f;mv*U(1HK&Wf25*I49aYV5BDFAJ`Gw){ce<-@}g*oQCxyi=B~~PuN`V@0Gr51VJ6emRxz}MS;ys9 -K#Q37N7-zuFW5pcC-A+f7-w9?^|Ry&s|Gx=Z6U$JR>)Ix?`-#cOLjx9mV8BrPE+Y-`l=et)fge+q9